diff --git a/.gitattributes b/.gitattributes
index 22d43eccc054154fdd49daa00f949a70d8c4c2fe..b255de9dd7d1ad2d650c0205fed4acadfeb0b660 100644
--- a/.gitattributes
+++ b/.gitattributes
@@ -679,3 +679,10 @@ runs/l2r50-i2-fulle-lm/checkpoint-72/tokenizer.json filter=lfs diff=lfs merge=lf
runs/l2r50-i2-fulle-lm/checkpoint-7200/tokenizer.json filter=lfs diff=lfs merge=lfs -text
runs/l2r50-i2-fulle-lm/checkpoint-864/tokenizer.json filter=lfs diff=lfs merge=lfs -text
runs/mask15-l2r50-fulle-lm/checkpoint-7200/tokenizer.json filter=lfs diff=lfs merge=lfs -text
+runs/l2r50-i2-fulle-lm/checkpoint-7488/tokenizer.json filter=lfs diff=lfs merge=lfs -text
+runs/mask15-l2r50-i2-fulle-lm/checkpoint-1152/tokenizer.json filter=lfs diff=lfs merge=lfs -text
+runs/mask15-l2r50-i2-fulle-lm/checkpoint-1440/tokenizer.json filter=lfs diff=lfs merge=lfs -text
+runs/mask15-l2r50-i2-fulle-lm/checkpoint-288/tokenizer.json filter=lfs diff=lfs merge=lfs -text
+runs/mask15-l2r50-i2-fulle-lm/checkpoint-576/tokenizer.json filter=lfs diff=lfs merge=lfs -text
+runs/mask15-l2r50-i2-fulle-lm/checkpoint-72/tokenizer.json filter=lfs diff=lfs merge=lfs -text
+runs/mask15-l2r50-i2-fulle-lm/checkpoint-864/tokenizer.json filter=lfs diff=lfs merge=lfs -text
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-1152/chat_template.jinja b/runs/l2r50-i2-fulle-lm/checkpoint-1152/chat_template.jinja
new file mode 100644
index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-1152/chat_template.jinja
@@ -0,0 +1,85 @@
+{%- if tools %}
+ {{- '<|im_start|>system\n' }}
+ {%- if messages[0].role == 'system' %}
+ {{- messages[0].content + '\n\n' }}
+ {%- endif %}
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }}
+ {%- for tool in tools %}
+ {{- "\n" }}
+ {{- tool | tojson }}
+ {%- endfor %}
+ {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }}
+{%- else %}
+ {%- if messages[0].role == 'system' %}
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
+ {%- endif %}
+{%- endif %}
+{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
+{%- for message in messages[::-1] %}
+ {%- set index = (messages|length - 1) - loop.index0 %}
+ {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %}
+ {%- set ns.multi_step_tool = false %}
+ {%- set ns.last_query_index = index %}
+ {%- endif %}
+{%- endfor %}
+{%- for message in messages %}
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
+ {%- elif message.role == "assistant" %}
+ {%- set content = message.content %}
+ {%- set reasoning_content = '' %}
+ {%- if message.reasoning_content is defined and message.reasoning_content is not none %}
+ {%- set reasoning_content = message.reasoning_content %}
+ {%- else %}
+ {%- if '' in message.content %}
+ {%- set content = message.content.split('')[-1].lstrip('\n') %}
+ {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %}
+ {%- endif %}
+ {%- endif %}
+ {%- if loop.index0 > ns.last_query_index %}
+ {%- if loop.last or (not loop.last and reasoning_content) %}
+ {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- if message.tool_calls %}
+ {%- for tool_call in message.tool_calls %}
+ {%- if (loop.first and content) or (not loop.first) %}
+ {{- '\n' }}
+ {%- endif %}
+ {%- if tool_call.function %}
+ {%- set tool_call = tool_call.function %}
+ {%- endif %}
+ {{- '\n{"name": "' }}
+ {{- tool_call.name }}
+ {{- '", "arguments": ' }}
+ {%- if tool_call.arguments is string %}
+ {{- tool_call.arguments }}
+ {%- else %}
+ {{- tool_call.arguments | tojson }}
+ {%- endif %}
+ {{- '}\n' }}
+ {%- endfor %}
+ {%- endif %}
+ {{- '<|im_end|>\n' }}
+ {%- elif message.role == "tool" %}
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
+ {{- '<|im_start|>user' }}
+ {%- endif %}
+ {{- '\n\n' }}
+ {{- message.content }}
+ {{- '\n' }}
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
+ {{- '<|im_end|>\n' }}
+ {%- endif %}
+ {%- endif %}
+{%- endfor %}
+{%- if add_generation_prompt %}
+ {{- '<|im_start|>assistant\n' }}
+ {%- if enable_thinking is defined and enable_thinking is false %}
+ {{- '\n\n\n\n' }}
+ {%- endif %}
+{%- endif %}
\ No newline at end of file
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-1152/config.json b/runs/l2r50-i2-fulle-lm/checkpoint-1152/config.json
new file mode 100644
index 0000000000000000000000000000000000000000..22f250953c6ce8fc91ee4411856dd73041a415e2
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-1152/config.json
@@ -0,0 +1,32 @@
+{
+ "architectures": [
+ "LlamaForCausalLM"
+ ],
+ "attention_bias": false,
+ "attention_dropout": 0.0,
+ "bos_token_id": null,
+ "dtype": "float32",
+ "eos_token_id": 151645,
+ "head_dim": 128,
+ "hidden_act": "silu",
+ "hidden_size": 512,
+ "initializer_range": 0.02,
+ "intermediate_size": 1536,
+ "max_position_embeddings": 2048,
+ "mlp_bias": false,
+ "model_type": "llama",
+ "num_attention_heads": 4,
+ "num_hidden_layers": 20,
+ "num_key_value_heads": 4,
+ "pad_token_id": 151645,
+ "pretraining_tp": 1,
+ "rms_norm_eps": 1e-06,
+ "rope_parameters": {
+ "rope_theta": 10000.0,
+ "rope_type": "default"
+ },
+ "tie_word_embeddings": true,
+ "transformers_version": "5.5.0",
+ "use_cache": false,
+ "vocab_size": 151673
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-1152/generation_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-1152/generation_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-1152/generation_config.json
@@ -0,0 +1,11 @@
+{
+ "_from_model_config": true,
+ "eos_token_id": [
+ 151645
+ ],
+ "output_attentions": false,
+ "output_hidden_states": false,
+ "pad_token_id": 151645,
+ "transformers_version": "5.5.0",
+ "use_cache": true
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-1152/model.safetensors b/runs/l2r50-i2-fulle-lm/checkpoint-1152/model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..eab74536f5292607bb166ad567c81f2a71d212e5
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-1152/model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:7456ca1104a53851cb33f03ad5729d8c58e31c480e5e1fd319c864031c1d4024
+size 583360328
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-1152/optimizer.pt b/runs/l2r50-i2-fulle-lm/checkpoint-1152/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..3b99eafc944be08e97bfa7b25ec548e04f179dfa
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-1152/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:86273e83eba0e7d0a005a4d73b6f3b0c73fbe7a5f68231fced6a0b61d3ad406b
+size 1166833530
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-1152/rng_state_0.pth b/runs/l2r50-i2-fulle-lm/checkpoint-1152/rng_state_0.pth
new file mode 100644
index 0000000000000000000000000000000000000000..f953b3d2601997ff26348cfe65bdf6f072da1241
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-1152/rng_state_0.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:62b735ad70e5179210c3231e43080bd2451d7d358b5efe5db5997cbd75fd394d
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-1152/rng_state_1.pth b/runs/l2r50-i2-fulle-lm/checkpoint-1152/rng_state_1.pth
new file mode 100644
index 0000000000000000000000000000000000000000..90eeeb9606172a80974fd4d514e5c336ea10bede
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-1152/rng_state_1.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:4e39f26de4d91b53c07b4529fc28b0d8db87f95f5cd65b50c9d4f7b36f41aac1
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-1152/rng_state_2.pth b/runs/l2r50-i2-fulle-lm/checkpoint-1152/rng_state_2.pth
new file mode 100644
index 0000000000000000000000000000000000000000..41b5a0aaa0603136050870ac83e05be64379af70
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-1152/rng_state_2.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:d09a5602d03fd240123f234cec1b33801653989554e35eede4963f7b24092f96
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-1152/rng_state_3.pth b/runs/l2r50-i2-fulle-lm/checkpoint-1152/rng_state_3.pth
new file mode 100644
index 0000000000000000000000000000000000000000..5fa2d1f78ea851b45326b8bc87837827bc74fe42
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-1152/rng_state_3.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:751dd23154d8ff44e1735c7b0fd261bc89701be5ab5e929dbb27efa8d599b231
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-1152/scheduler.pt b/runs/l2r50-i2-fulle-lm/checkpoint-1152/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..630fbc73d70608487940c0c34e65caee42415237
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-1152/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:6a67914066d1aebda73626cb9d7d4cbc46ee7df332e5ae85a86cc87be8e519c8
+size 1064
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-1152/tokenizer.json b/runs/l2r50-i2-fulle-lm/checkpoint-1152/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..53786016f070e11184813f3c83101dc5e93e392e
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-1152/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:c3c63fda1646a555448e8f03662542e04b11d103721319e4d9b267d7da03ceb8
+size 11424463
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-1152/tokenizer_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-1152/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..98e2fb9b2cc54785dc9d871b8e503f8b8f2defe7
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-1152/tokenizer_config.json
@@ -0,0 +1,20 @@
+{
+ "add_prefix_space": false,
+ "backend": "tokenizers",
+ "bos_token": null,
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "<|im_end|>",
+ "errors": "replace",
+ "extra_special_tokens": [
+ "<|l2r_pred|>",
+ "<|r2l_pred|>",
+ "<|next_1_pred|>",
+ "<|next_2_pred|>"
+ ],
+ "is_local": false,
+ "model_max_length": 131072,
+ "pad_token": "<|im_end|>",
+ "split_special_tokens": false,
+ "tokenizer_class": "Qwen2Tokenizer",
+ "unk_token": null
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-1152/trainer_state.json b/runs/l2r50-i2-fulle-lm/checkpoint-1152/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..122680c53242df36b46a84d812209adda9132756
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-1152/trainer_state.json
@@ -0,0 +1,8218 @@
+{
+ "best_global_step": null,
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 16.0,
+ "eval_steps": 500,
+ "global_step": 1152,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.013973799126637555,
+ "grad_norm": 0.1739506870508194,
+ "learning_rate": 0.0,
+ "loss": 11.99299430847168,
+ "step": 1
+ },
+ {
+ "epoch": 0.02794759825327511,
+ "grad_norm": 0.17216597497463226,
+ "learning_rate": 5.999999999999999e-06,
+ "loss": 11.988704681396484,
+ "step": 2
+ },
+ {
+ "epoch": 0.04192139737991266,
+ "grad_norm": 0.17459715902805328,
+ "learning_rate": 1.1999999999999999e-05,
+ "loss": 11.945581436157227,
+ "step": 3
+ },
+ {
+ "epoch": 0.05589519650655022,
+ "grad_norm": 0.17577414214611053,
+ "learning_rate": 1.7999999999999997e-05,
+ "loss": 11.859772682189941,
+ "step": 4
+ },
+ {
+ "epoch": 0.06986899563318777,
+ "grad_norm": 0.16163651645183563,
+ "learning_rate": 2.3999999999999997e-05,
+ "loss": 11.744415283203125,
+ "step": 5
+ },
+ {
+ "epoch": 0.08384279475982533,
+ "grad_norm": 0.14409932494163513,
+ "learning_rate": 2.9999999999999997e-05,
+ "loss": 11.618099212646484,
+ "step": 6
+ },
+ {
+ "epoch": 0.09781659388646288,
+ "grad_norm": 0.13079848885536194,
+ "learning_rate": 3.5999999999999994e-05,
+ "loss": 11.505935668945312,
+ "step": 7
+ },
+ {
+ "epoch": 0.11179039301310044,
+ "grad_norm": 0.11964613944292068,
+ "learning_rate": 4.2e-05,
+ "loss": 11.416532516479492,
+ "step": 8
+ },
+ {
+ "epoch": 0.125764192139738,
+ "grad_norm": 0.11345142871141434,
+ "learning_rate": 4.7999999999999994e-05,
+ "loss": 11.33206558227539,
+ "step": 9
+ },
+ {
+ "epoch": 0.13973799126637554,
+ "grad_norm": 0.11033952981233597,
+ "learning_rate": 5.399999999999999e-05,
+ "loss": 11.259052276611328,
+ "step": 10
+ },
+ {
+ "epoch": 0.1537117903930131,
+ "grad_norm": 0.10776489973068237,
+ "learning_rate": 5.9999999999999995e-05,
+ "loss": 11.203384399414062,
+ "step": 11
+ },
+ {
+ "epoch": 0.16768558951965065,
+ "grad_norm": 0.10721946507692337,
+ "learning_rate": 6.599999999999999e-05,
+ "loss": 11.153295516967773,
+ "step": 12
+ },
+ {
+ "epoch": 0.18165938864628822,
+ "grad_norm": 0.10673430562019348,
+ "learning_rate": 7.199999999999999e-05,
+ "loss": 11.110200881958008,
+ "step": 13
+ },
+ {
+ "epoch": 0.19563318777292577,
+ "grad_norm": 0.10630254447460175,
+ "learning_rate": 7.8e-05,
+ "loss": 11.067477226257324,
+ "step": 14
+ },
+ {
+ "epoch": 0.2096069868995633,
+ "grad_norm": 0.10631615668535233,
+ "learning_rate": 8.4e-05,
+ "loss": 11.022941589355469,
+ "step": 15
+ },
+ {
+ "epoch": 0.22358078602620088,
+ "grad_norm": 0.10609443485736847,
+ "learning_rate": 8.999999999999999e-05,
+ "loss": 10.970746040344238,
+ "step": 16
+ },
+ {
+ "epoch": 0.23755458515283842,
+ "grad_norm": 0.10648232698440552,
+ "learning_rate": 9.599999999999999e-05,
+ "loss": 10.91267204284668,
+ "step": 17
+ },
+ {
+ "epoch": 0.251528384279476,
+ "grad_norm": 0.10584589838981628,
+ "learning_rate": 0.000102,
+ "loss": 10.854547500610352,
+ "step": 18
+ },
+ {
+ "epoch": 0.26550218340611353,
+ "grad_norm": 0.10591720044612885,
+ "learning_rate": 0.00010799999999999998,
+ "loss": 10.788917541503906,
+ "step": 19
+ },
+ {
+ "epoch": 0.2794759825327511,
+ "grad_norm": 0.10611672699451447,
+ "learning_rate": 0.00011399999999999999,
+ "loss": 10.717556953430176,
+ "step": 20
+ },
+ {
+ "epoch": 0.2934497816593886,
+ "grad_norm": 0.10455359518527985,
+ "learning_rate": 0.00011999999999999999,
+ "loss": 10.650186538696289,
+ "step": 21
+ },
+ {
+ "epoch": 0.3074235807860262,
+ "grad_norm": 0.1047525629401207,
+ "learning_rate": 0.00012599999999999997,
+ "loss": 10.56808090209961,
+ "step": 22
+ },
+ {
+ "epoch": 0.32139737991266376,
+ "grad_norm": 0.10511605441570282,
+ "learning_rate": 0.00013199999999999998,
+ "loss": 10.484028816223145,
+ "step": 23
+ },
+ {
+ "epoch": 0.3353711790393013,
+ "grad_norm": 0.10467177629470825,
+ "learning_rate": 0.000138,
+ "loss": 10.399947166442871,
+ "step": 24
+ },
+ {
+ "epoch": 0.34934497816593885,
+ "grad_norm": 0.10432136058807373,
+ "learning_rate": 0.00014399999999999998,
+ "loss": 10.312714576721191,
+ "step": 25
+ },
+ {
+ "epoch": 0.36331877729257644,
+ "grad_norm": 0.10527262091636658,
+ "learning_rate": 0.00015,
+ "loss": 10.212261199951172,
+ "step": 26
+ },
+ {
+ "epoch": 0.377292576419214,
+ "grad_norm": 0.103610560297966,
+ "learning_rate": 0.000156,
+ "loss": 10.132034301757812,
+ "step": 27
+ },
+ {
+ "epoch": 0.39126637554585153,
+ "grad_norm": 0.10553059726953506,
+ "learning_rate": 0.000162,
+ "loss": 10.020434379577637,
+ "step": 28
+ },
+ {
+ "epoch": 0.4052401746724891,
+ "grad_norm": 0.1037299782037735,
+ "learning_rate": 0.000168,
+ "loss": 9.934466361999512,
+ "step": 29
+ },
+ {
+ "epoch": 0.4192139737991266,
+ "grad_norm": 0.1042739674448967,
+ "learning_rate": 0.00017399999999999997,
+ "loss": 9.822539329528809,
+ "step": 30
+ },
+ {
+ "epoch": 0.4331877729257642,
+ "grad_norm": 0.10300865024328232,
+ "learning_rate": 0.00017999999999999998,
+ "loss": 9.734994888305664,
+ "step": 31
+ },
+ {
+ "epoch": 0.44716157205240176,
+ "grad_norm": 0.10291654616594315,
+ "learning_rate": 0.000186,
+ "loss": 9.624505996704102,
+ "step": 32
+ },
+ {
+ "epoch": 0.4611353711790393,
+ "grad_norm": 0.10210851579904556,
+ "learning_rate": 0.00019199999999999998,
+ "loss": 9.518205642700195,
+ "step": 33
+ },
+ {
+ "epoch": 0.47510917030567684,
+ "grad_norm": 0.10222529619932175,
+ "learning_rate": 0.000198,
+ "loss": 9.412222862243652,
+ "step": 34
+ },
+ {
+ "epoch": 0.4890829694323144,
+ "grad_norm": 0.09992019087076187,
+ "learning_rate": 0.000204,
+ "loss": 9.326566696166992,
+ "step": 35
+ },
+ {
+ "epoch": 0.503056768558952,
+ "grad_norm": 0.09931084513664246,
+ "learning_rate": 0.00020999999999999998,
+ "loss": 9.210111618041992,
+ "step": 36
+ },
+ {
+ "epoch": 0.5170305676855895,
+ "grad_norm": 0.09928256273269653,
+ "learning_rate": 0.00021599999999999996,
+ "loss": 9.100493431091309,
+ "step": 37
+ },
+ {
+ "epoch": 0.5310043668122271,
+ "grad_norm": 0.09843816608190536,
+ "learning_rate": 0.00022199999999999998,
+ "loss": 8.996543884277344,
+ "step": 38
+ },
+ {
+ "epoch": 0.5449781659388646,
+ "grad_norm": 0.09825374186038971,
+ "learning_rate": 0.00022799999999999999,
+ "loss": 8.888991355895996,
+ "step": 39
+ },
+ {
+ "epoch": 0.5589519650655022,
+ "grad_norm": 0.09602385759353638,
+ "learning_rate": 0.000234,
+ "loss": 8.802907943725586,
+ "step": 40
+ },
+ {
+ "epoch": 0.5729257641921397,
+ "grad_norm": 0.09414064139127731,
+ "learning_rate": 0.00023999999999999998,
+ "loss": 8.715423583984375,
+ "step": 41
+ },
+ {
+ "epoch": 0.5868995633187772,
+ "grad_norm": 0.09273893386125565,
+ "learning_rate": 0.00024599999999999996,
+ "loss": 8.61374282836914,
+ "step": 42
+ },
+ {
+ "epoch": 0.6008733624454149,
+ "grad_norm": 0.09079599380493164,
+ "learning_rate": 0.00025199999999999995,
+ "loss": 8.502904891967773,
+ "step": 43
+ },
+ {
+ "epoch": 0.6148471615720524,
+ "grad_norm": 0.08605825901031494,
+ "learning_rate": 0.000258,
+ "loss": 8.453008651733398,
+ "step": 44
+ },
+ {
+ "epoch": 0.62882096069869,
+ "grad_norm": 0.08338996022939682,
+ "learning_rate": 0.00026399999999999997,
+ "loss": 8.365163803100586,
+ "step": 45
+ },
+ {
+ "epoch": 0.6427947598253275,
+ "grad_norm": 0.08150867372751236,
+ "learning_rate": 0.00027,
+ "loss": 8.285372734069824,
+ "step": 46
+ },
+ {
+ "epoch": 0.6567685589519651,
+ "grad_norm": 0.07794807851314545,
+ "learning_rate": 0.000276,
+ "loss": 8.185811996459961,
+ "step": 47
+ },
+ {
+ "epoch": 0.6707423580786026,
+ "grad_norm": 0.07592720538377762,
+ "learning_rate": 0.00028199999999999997,
+ "loss": 8.098482131958008,
+ "step": 48
+ },
+ {
+ "epoch": 0.6847161572052402,
+ "grad_norm": 0.0709897130727768,
+ "learning_rate": 0.00028799999999999995,
+ "loss": 8.042091369628906,
+ "step": 49
+ },
+ {
+ "epoch": 0.6986899563318777,
+ "grad_norm": 0.06582305580377579,
+ "learning_rate": 0.000294,
+ "loss": 7.972844123840332,
+ "step": 50
+ },
+ {
+ "epoch": 0.7126637554585152,
+ "grad_norm": 0.059452351182699203,
+ "learning_rate": 0.0003,
+ "loss": 7.924579620361328,
+ "step": 51
+ },
+ {
+ "epoch": 0.7266375545851529,
+ "grad_norm": 0.052426718175411224,
+ "learning_rate": 0.00030599999999999996,
+ "loss": 7.876336097717285,
+ "step": 52
+ },
+ {
+ "epoch": 0.7406113537117904,
+ "grad_norm": 0.04768150672316551,
+ "learning_rate": 0.000312,
+ "loss": 7.8102707862854,
+ "step": 53
+ },
+ {
+ "epoch": 0.754585152838428,
+ "grad_norm": 0.03970075398683548,
+ "learning_rate": 0.000318,
+ "loss": 7.791913986206055,
+ "step": 54
+ },
+ {
+ "epoch": 0.7685589519650655,
+ "grad_norm": 0.03559752553701401,
+ "learning_rate": 0.000324,
+ "loss": 7.756960868835449,
+ "step": 55
+ },
+ {
+ "epoch": 0.7825327510917031,
+ "grad_norm": 0.029321353882551193,
+ "learning_rate": 0.00033,
+ "loss": 7.724878787994385,
+ "step": 56
+ },
+ {
+ "epoch": 0.7965065502183406,
+ "grad_norm": 0.023485427722334862,
+ "learning_rate": 0.000336,
+ "loss": 7.72553014755249,
+ "step": 57
+ },
+ {
+ "epoch": 0.8104803493449781,
+ "grad_norm": 0.019319474697113037,
+ "learning_rate": 0.00034199999999999996,
+ "loss": 7.709493160247803,
+ "step": 58
+ },
+ {
+ "epoch": 0.8244541484716157,
+ "grad_norm": 0.016705691814422607,
+ "learning_rate": 0.00034799999999999995,
+ "loss": 7.6873297691345215,
+ "step": 59
+ },
+ {
+ "epoch": 0.8384279475982532,
+ "grad_norm": 0.017667734995484352,
+ "learning_rate": 0.00035399999999999993,
+ "loss": 7.677059173583984,
+ "step": 60
+ },
+ {
+ "epoch": 0.8524017467248908,
+ "grad_norm": 0.01671992614865303,
+ "learning_rate": 0.00035999999999999997,
+ "loss": 7.643512725830078,
+ "step": 61
+ },
+ {
+ "epoch": 0.8663755458515284,
+ "grad_norm": 0.012288172729313374,
+ "learning_rate": 0.00036599999999999995,
+ "loss": 7.658343315124512,
+ "step": 62
+ },
+ {
+ "epoch": 0.880349344978166,
+ "grad_norm": 0.012004776857793331,
+ "learning_rate": 0.000372,
+ "loss": 7.649759292602539,
+ "step": 63
+ },
+ {
+ "epoch": 0.8943231441048035,
+ "grad_norm": 0.012487445957958698,
+ "learning_rate": 0.00037799999999999997,
+ "loss": 7.631503105163574,
+ "step": 64
+ },
+ {
+ "epoch": 0.9082969432314411,
+ "grad_norm": 0.01489241048693657,
+ "learning_rate": 0.00038399999999999996,
+ "loss": 7.647110462188721,
+ "step": 65
+ },
+ {
+ "epoch": 0.9222707423580786,
+ "grad_norm": 0.01428980939090252,
+ "learning_rate": 0.00039,
+ "loss": 7.636983871459961,
+ "step": 66
+ },
+ {
+ "epoch": 0.9362445414847161,
+ "grad_norm": 0.010711144655942917,
+ "learning_rate": 0.000396,
+ "loss": 7.621130466461182,
+ "step": 67
+ },
+ {
+ "epoch": 0.9502183406113537,
+ "grad_norm": 0.010178886353969574,
+ "learning_rate": 0.000402,
+ "loss": 7.593924522399902,
+ "step": 68
+ },
+ {
+ "epoch": 0.9641921397379912,
+ "grad_norm": 0.010604874230921268,
+ "learning_rate": 0.000408,
+ "loss": 7.6015496253967285,
+ "step": 69
+ },
+ {
+ "epoch": 0.9781659388646288,
+ "grad_norm": 0.009751426056027412,
+ "learning_rate": 0.0004139999999999999,
+ "loss": 7.571964263916016,
+ "step": 70
+ },
+ {
+ "epoch": 0.9921397379912664,
+ "grad_norm": 0.009816481731832027,
+ "learning_rate": 0.00041999999999999996,
+ "loss": 7.57344913482666,
+ "step": 71
+ },
+ {
+ "epoch": 1.0,
+ "grad_norm": 0.008665637113153934,
+ "learning_rate": 0.00042599999999999995,
+ "loss": 7.579050064086914,
+ "step": 72
+ },
+ {
+ "epoch": 1.0,
+ "eval_loss": 7.581204891204834,
+ "eval_runtime": 59.0091,
+ "eval_samples_per_second": 41.383,
+ "eval_steps_per_second": 1.305,
+ "step": 72
+ },
+ {
+ "epoch": 1.0139737991266375,
+ "grad_norm": 0.009684201329946518,
+ "learning_rate": 0.00043199999999999993,
+ "loss": 7.568594932556152,
+ "step": 73
+ },
+ {
+ "epoch": 1.027947598253275,
+ "grad_norm": 0.025783710181713104,
+ "learning_rate": 0.00043799999999999997,
+ "loss": 7.578948497772217,
+ "step": 74
+ },
+ {
+ "epoch": 1.0419213973799126,
+ "grad_norm": 0.0327819399535656,
+ "learning_rate": 0.00044399999999999995,
+ "loss": 7.548173904418945,
+ "step": 75
+ },
+ {
+ "epoch": 1.0558951965065502,
+ "grad_norm": 0.021604498848319054,
+ "learning_rate": 0.00045,
+ "loss": 7.552953720092773,
+ "step": 76
+ },
+ {
+ "epoch": 1.0698689956331877,
+ "grad_norm": 0.010230105370283127,
+ "learning_rate": 0.00045599999999999997,
+ "loss": 7.5213212966918945,
+ "step": 77
+ },
+ {
+ "epoch": 1.0838427947598253,
+ "grad_norm": 0.023315589874982834,
+ "learning_rate": 0.00046199999999999995,
+ "loss": 7.501640319824219,
+ "step": 78
+ },
+ {
+ "epoch": 1.0978165938864628,
+ "grad_norm": 0.01109944935888052,
+ "learning_rate": 0.000468,
+ "loss": 7.539519786834717,
+ "step": 79
+ },
+ {
+ "epoch": 1.1117903930131003,
+ "grad_norm": 0.026093758642673492,
+ "learning_rate": 0.000474,
+ "loss": 7.525928974151611,
+ "step": 80
+ },
+ {
+ "epoch": 1.125764192139738,
+ "grad_norm": 0.03274895250797272,
+ "learning_rate": 0.00047999999999999996,
+ "loss": 7.535602569580078,
+ "step": 81
+ },
+ {
+ "epoch": 1.1397379912663754,
+ "grad_norm": 0.01543602254241705,
+ "learning_rate": 0.000486,
+ "loss": 7.513683795928955,
+ "step": 82
+ },
+ {
+ "epoch": 1.1537117903930132,
+ "grad_norm": 0.028377249836921692,
+ "learning_rate": 0.0004919999999999999,
+ "loss": 7.49384880065918,
+ "step": 83
+ },
+ {
+ "epoch": 1.1676855895196507,
+ "grad_norm": 0.04121481627225876,
+ "learning_rate": 0.000498,
+ "loss": 7.488447189331055,
+ "step": 84
+ },
+ {
+ "epoch": 1.1816593886462883,
+ "grad_norm": 0.013282302767038345,
+ "learning_rate": 0.0005039999999999999,
+ "loss": 7.463010787963867,
+ "step": 85
+ },
+ {
+ "epoch": 1.1956331877729258,
+ "grad_norm": 0.03200364485383034,
+ "learning_rate": 0.0005099999999999999,
+ "loss": 7.473515033721924,
+ "step": 86
+ },
+ {
+ "epoch": 1.2096069868995634,
+ "grad_norm": 0.020768029615283012,
+ "learning_rate": 0.000516,
+ "loss": 7.464688777923584,
+ "step": 87
+ },
+ {
+ "epoch": 1.223580786026201,
+ "grad_norm": 0.021715788170695305,
+ "learning_rate": 0.000522,
+ "loss": 7.4835028648376465,
+ "step": 88
+ },
+ {
+ "epoch": 1.2375545851528384,
+ "grad_norm": 0.026255754753947258,
+ "learning_rate": 0.0005279999999999999,
+ "loss": 7.443392753601074,
+ "step": 89
+ },
+ {
+ "epoch": 1.251528384279476,
+ "grad_norm": 0.022325951606035233,
+ "learning_rate": 0.000534,
+ "loss": 7.457373142242432,
+ "step": 90
+ },
+ {
+ "epoch": 1.2655021834061135,
+ "grad_norm": 0.009270387701690197,
+ "learning_rate": 0.00054,
+ "loss": 7.455986976623535,
+ "step": 91
+ },
+ {
+ "epoch": 1.279475982532751,
+ "grad_norm": 0.02369544468820095,
+ "learning_rate": 0.0005459999999999999,
+ "loss": 7.442216873168945,
+ "step": 92
+ },
+ {
+ "epoch": 1.2934497816593886,
+ "grad_norm": 0.008474506437778473,
+ "learning_rate": 0.000552,
+ "loss": 7.4294915199279785,
+ "step": 93
+ },
+ {
+ "epoch": 1.3074235807860262,
+ "grad_norm": 0.017043249681591988,
+ "learning_rate": 0.000558,
+ "loss": 7.406932830810547,
+ "step": 94
+ },
+ {
+ "epoch": 1.3213973799126637,
+ "grad_norm": 0.011813412420451641,
+ "learning_rate": 0.0005639999999999999,
+ "loss": 7.415900230407715,
+ "step": 95
+ },
+ {
+ "epoch": 1.3353711790393012,
+ "grad_norm": 0.01086883433163166,
+ "learning_rate": 0.00057,
+ "loss": 7.452998638153076,
+ "step": 96
+ },
+ {
+ "epoch": 1.3493449781659388,
+ "grad_norm": 0.014700951054692268,
+ "learning_rate": 0.0005759999999999999,
+ "loss": 7.410249710083008,
+ "step": 97
+ },
+ {
+ "epoch": 1.3633187772925766,
+ "grad_norm": 0.014512976631522179,
+ "learning_rate": 0.0005819999999999999,
+ "loss": 7.429132461547852,
+ "step": 98
+ },
+ {
+ "epoch": 1.3772925764192139,
+ "grad_norm": 0.012519920244812965,
+ "learning_rate": 0.000588,
+ "loss": 7.430817604064941,
+ "step": 99
+ },
+ {
+ "epoch": 1.3912663755458516,
+ "grad_norm": 0.007132664322853088,
+ "learning_rate": 0.0005939999999999999,
+ "loss": 7.405234336853027,
+ "step": 100
+ },
+ {
+ "epoch": 1.405240174672489,
+ "grad_norm": 0.019489388912916183,
+ "learning_rate": 0.0006,
+ "loss": 7.427791595458984,
+ "step": 101
+ },
+ {
+ "epoch": 1.4192139737991267,
+ "grad_norm": 0.027252664789557457,
+ "learning_rate": 0.0006,
+ "loss": 7.41761589050293,
+ "step": 102
+ },
+ {
+ "epoch": 1.4331877729257643,
+ "grad_norm": 0.02501637488603592,
+ "learning_rate": 0.0006,
+ "loss": 7.416735649108887,
+ "step": 103
+ },
+ {
+ "epoch": 1.4471615720524018,
+ "grad_norm": 0.00989772193133831,
+ "learning_rate": 0.0006,
+ "loss": 7.3756632804870605,
+ "step": 104
+ },
+ {
+ "epoch": 1.4611353711790394,
+ "grad_norm": 0.01622949168086052,
+ "learning_rate": 0.0006,
+ "loss": 7.400811672210693,
+ "step": 105
+ },
+ {
+ "epoch": 1.475109170305677,
+ "grad_norm": 0.012407763861119747,
+ "learning_rate": 0.0006,
+ "loss": 7.37675666809082,
+ "step": 106
+ },
+ {
+ "epoch": 1.4890829694323144,
+ "grad_norm": 0.010353565216064453,
+ "learning_rate": 0.0006,
+ "loss": 7.4013142585754395,
+ "step": 107
+ },
+ {
+ "epoch": 1.503056768558952,
+ "grad_norm": 0.0120553532615304,
+ "learning_rate": 0.0006,
+ "loss": 7.379873275756836,
+ "step": 108
+ },
+ {
+ "epoch": 1.5170305676855895,
+ "grad_norm": 0.01610225811600685,
+ "learning_rate": 0.0006,
+ "loss": 7.371631145477295,
+ "step": 109
+ },
+ {
+ "epoch": 1.531004366812227,
+ "grad_norm": 0.01159879844635725,
+ "learning_rate": 0.0006,
+ "loss": 7.389007568359375,
+ "step": 110
+ },
+ {
+ "epoch": 1.5449781659388646,
+ "grad_norm": 0.008013816550374031,
+ "learning_rate": 0.0006,
+ "loss": 7.395920753479004,
+ "step": 111
+ },
+ {
+ "epoch": 1.5589519650655022,
+ "grad_norm": 0.016334809362888336,
+ "learning_rate": 0.0006,
+ "loss": 7.393121719360352,
+ "step": 112
+ },
+ {
+ "epoch": 1.5729257641921397,
+ "grad_norm": 0.018821721896529198,
+ "learning_rate": 0.0006,
+ "loss": 7.367543697357178,
+ "step": 113
+ },
+ {
+ "epoch": 1.5868995633187772,
+ "grad_norm": 0.014088994823396206,
+ "learning_rate": 0.0006,
+ "loss": 7.385537624359131,
+ "step": 114
+ },
+ {
+ "epoch": 1.600873362445415,
+ "grad_norm": 0.005864050704985857,
+ "learning_rate": 0.0006,
+ "loss": 7.365346908569336,
+ "step": 115
+ },
+ {
+ "epoch": 1.6148471615720523,
+ "grad_norm": 0.011745430529117584,
+ "learning_rate": 0.0006,
+ "loss": 7.390227317810059,
+ "step": 116
+ },
+ {
+ "epoch": 1.62882096069869,
+ "grad_norm": 0.017205573618412018,
+ "learning_rate": 0.0006,
+ "loss": 7.3558549880981445,
+ "step": 117
+ },
+ {
+ "epoch": 1.6427947598253274,
+ "grad_norm": 0.021256346255540848,
+ "learning_rate": 0.0006,
+ "loss": 7.377346992492676,
+ "step": 118
+ },
+ {
+ "epoch": 1.6567685589519652,
+ "grad_norm": 0.019335204735398293,
+ "learning_rate": 0.0006,
+ "loss": 7.346194267272949,
+ "step": 119
+ },
+ {
+ "epoch": 1.6707423580786025,
+ "grad_norm": 0.008134003728628159,
+ "learning_rate": 0.0006,
+ "loss": 7.3063578605651855,
+ "step": 120
+ },
+ {
+ "epoch": 1.6847161572052403,
+ "grad_norm": 0.017613831907510757,
+ "learning_rate": 0.0006,
+ "loss": 7.338171005249023,
+ "step": 121
+ },
+ {
+ "epoch": 1.6986899563318776,
+ "grad_norm": 0.00890734139829874,
+ "learning_rate": 0.0006,
+ "loss": 7.355384826660156,
+ "step": 122
+ },
+ {
+ "epoch": 1.7126637554585153,
+ "grad_norm": 0.013448155485093594,
+ "learning_rate": 0.0006,
+ "loss": 7.344296455383301,
+ "step": 123
+ },
+ {
+ "epoch": 1.726637554585153,
+ "grad_norm": 0.009241633117198944,
+ "learning_rate": 0.0006,
+ "loss": 7.3454060554504395,
+ "step": 124
+ },
+ {
+ "epoch": 1.7406113537117904,
+ "grad_norm": 0.018561311066150665,
+ "learning_rate": 0.0006,
+ "loss": 7.32348108291626,
+ "step": 125
+ },
+ {
+ "epoch": 1.754585152838428,
+ "grad_norm": 0.025667704641819,
+ "learning_rate": 0.0006,
+ "loss": 7.356142520904541,
+ "step": 126
+ },
+ {
+ "epoch": 1.7685589519650655,
+ "grad_norm": 0.036232151091098785,
+ "learning_rate": 0.0006,
+ "loss": 7.363539695739746,
+ "step": 127
+ },
+ {
+ "epoch": 1.782532751091703,
+ "grad_norm": 0.059483714401721954,
+ "learning_rate": 0.0006,
+ "loss": 7.402331352233887,
+ "step": 128
+ },
+ {
+ "epoch": 1.7965065502183406,
+ "grad_norm": 0.05365009233355522,
+ "learning_rate": 0.0006,
+ "loss": 7.40457820892334,
+ "step": 129
+ },
+ {
+ "epoch": 1.8104803493449781,
+ "grad_norm": 0.02738415263593197,
+ "learning_rate": 0.0006,
+ "loss": 7.3756208419799805,
+ "step": 130
+ },
+ {
+ "epoch": 1.8244541484716157,
+ "grad_norm": 0.035017091780900955,
+ "learning_rate": 0.0006,
+ "loss": 7.379263877868652,
+ "step": 131
+ },
+ {
+ "epoch": 1.8384279475982532,
+ "grad_norm": 0.02819863148033619,
+ "learning_rate": 0.0006,
+ "loss": 7.327673435211182,
+ "step": 132
+ },
+ {
+ "epoch": 1.8524017467248908,
+ "grad_norm": 0.017754238098859787,
+ "learning_rate": 0.0006,
+ "loss": 7.333420276641846,
+ "step": 133
+ },
+ {
+ "epoch": 1.8663755458515285,
+ "grad_norm": 0.024685733020305634,
+ "learning_rate": 0.0006,
+ "loss": 7.34493350982666,
+ "step": 134
+ },
+ {
+ "epoch": 1.8803493449781659,
+ "grad_norm": 0.019426532089710236,
+ "learning_rate": 0.0006,
+ "loss": 7.328448295593262,
+ "step": 135
+ },
+ {
+ "epoch": 1.8943231441048036,
+ "grad_norm": 0.02106391079723835,
+ "learning_rate": 0.0006,
+ "loss": 7.327917098999023,
+ "step": 136
+ },
+ {
+ "epoch": 1.908296943231441,
+ "grad_norm": 0.02056879736483097,
+ "learning_rate": 0.0006,
+ "loss": 7.3311991691589355,
+ "step": 137
+ },
+ {
+ "epoch": 1.9222707423580787,
+ "grad_norm": 0.016865141689777374,
+ "learning_rate": 0.0006,
+ "loss": 7.312766075134277,
+ "step": 138
+ },
+ {
+ "epoch": 1.936244541484716,
+ "grad_norm": 0.017549855634570122,
+ "learning_rate": 0.0006,
+ "loss": 7.31191349029541,
+ "step": 139
+ },
+ {
+ "epoch": 1.9502183406113538,
+ "grad_norm": 0.015930157154798508,
+ "learning_rate": 0.0006,
+ "loss": 7.330901622772217,
+ "step": 140
+ },
+ {
+ "epoch": 1.9641921397379911,
+ "grad_norm": 0.013283558189868927,
+ "learning_rate": 0.0006,
+ "loss": 7.298318386077881,
+ "step": 141
+ },
+ {
+ "epoch": 1.9781659388646289,
+ "grad_norm": 0.012023469433188438,
+ "learning_rate": 0.0006,
+ "loss": 7.282337188720703,
+ "step": 142
+ },
+ {
+ "epoch": 1.9921397379912664,
+ "grad_norm": 0.01459511648863554,
+ "learning_rate": 0.0006,
+ "loss": 7.282924175262451,
+ "step": 143
+ },
+ {
+ "epoch": 2.0,
+ "grad_norm": 0.017474235966801643,
+ "learning_rate": 0.0006,
+ "loss": 7.254333972930908,
+ "step": 144
+ },
+ {
+ "epoch": 2.0,
+ "eval_loss": 7.318708896636963,
+ "eval_runtime": 59.6555,
+ "eval_samples_per_second": 40.935,
+ "eval_steps_per_second": 1.291,
+ "step": 144
+ },
+ {
+ "epoch": 2.0139737991266378,
+ "grad_norm": 0.019757507368922234,
+ "learning_rate": 0.0006,
+ "loss": 7.30288553237915,
+ "step": 145
+ },
+ {
+ "epoch": 2.027947598253275,
+ "grad_norm": 0.014571248553693295,
+ "learning_rate": 0.0006,
+ "loss": 7.274681568145752,
+ "step": 146
+ },
+ {
+ "epoch": 2.041921397379913,
+ "grad_norm": 0.015744198113679886,
+ "learning_rate": 0.0006,
+ "loss": 7.278632640838623,
+ "step": 147
+ },
+ {
+ "epoch": 2.05589519650655,
+ "grad_norm": 0.028048858046531677,
+ "learning_rate": 0.0006,
+ "loss": 7.26558780670166,
+ "step": 148
+ },
+ {
+ "epoch": 2.069868995633188,
+ "grad_norm": 0.06547307223081589,
+ "learning_rate": 0.0006,
+ "loss": 7.32567834854126,
+ "step": 149
+ },
+ {
+ "epoch": 2.0838427947598253,
+ "grad_norm": 0.05969763547182083,
+ "learning_rate": 0.0006,
+ "loss": 7.395836353302002,
+ "step": 150
+ },
+ {
+ "epoch": 2.097816593886463,
+ "grad_norm": 0.032574281096458435,
+ "learning_rate": 0.0006,
+ "loss": 7.318150997161865,
+ "step": 151
+ },
+ {
+ "epoch": 2.1117903930131003,
+ "grad_norm": 0.029452962800860405,
+ "learning_rate": 0.0006,
+ "loss": 7.308066368103027,
+ "step": 152
+ },
+ {
+ "epoch": 2.125764192139738,
+ "grad_norm": 0.03161991387605667,
+ "learning_rate": 0.0006,
+ "loss": 7.306290626525879,
+ "step": 153
+ },
+ {
+ "epoch": 2.1397379912663754,
+ "grad_norm": 0.027096295729279518,
+ "learning_rate": 0.0006,
+ "loss": 7.297796249389648,
+ "step": 154
+ },
+ {
+ "epoch": 2.153711790393013,
+ "grad_norm": 0.028657056391239166,
+ "learning_rate": 0.0006,
+ "loss": 7.317424774169922,
+ "step": 155
+ },
+ {
+ "epoch": 2.1676855895196505,
+ "grad_norm": 0.02046995982527733,
+ "learning_rate": 0.0006,
+ "loss": 7.297986030578613,
+ "step": 156
+ },
+ {
+ "epoch": 2.1816593886462883,
+ "grad_norm": 0.02220369316637516,
+ "learning_rate": 0.0006,
+ "loss": 7.268237590789795,
+ "step": 157
+ },
+ {
+ "epoch": 2.1956331877729256,
+ "grad_norm": 0.018381357192993164,
+ "learning_rate": 0.0006,
+ "loss": 7.259415626525879,
+ "step": 158
+ },
+ {
+ "epoch": 2.2096069868995634,
+ "grad_norm": 0.0204413291066885,
+ "learning_rate": 0.0006,
+ "loss": 7.258173942565918,
+ "step": 159
+ },
+ {
+ "epoch": 2.2235807860262007,
+ "grad_norm": 0.015874288976192474,
+ "learning_rate": 0.0006,
+ "loss": 7.281834602355957,
+ "step": 160
+ },
+ {
+ "epoch": 2.2375545851528384,
+ "grad_norm": 0.01634068600833416,
+ "learning_rate": 0.0006,
+ "loss": 7.2469868659973145,
+ "step": 161
+ },
+ {
+ "epoch": 2.251528384279476,
+ "grad_norm": 0.01728747971355915,
+ "learning_rate": 0.0006,
+ "loss": 7.280995845794678,
+ "step": 162
+ },
+ {
+ "epoch": 2.2655021834061135,
+ "grad_norm": 0.016341188922524452,
+ "learning_rate": 0.0006,
+ "loss": 7.273357391357422,
+ "step": 163
+ },
+ {
+ "epoch": 2.279475982532751,
+ "grad_norm": 0.016125807538628578,
+ "learning_rate": 0.0006,
+ "loss": 7.273510932922363,
+ "step": 164
+ },
+ {
+ "epoch": 2.2934497816593886,
+ "grad_norm": 0.01629287749528885,
+ "learning_rate": 0.0006,
+ "loss": 7.267029762268066,
+ "step": 165
+ },
+ {
+ "epoch": 2.3074235807860264,
+ "grad_norm": 0.015296131372451782,
+ "learning_rate": 0.0006,
+ "loss": 7.211989402770996,
+ "step": 166
+ },
+ {
+ "epoch": 2.3213973799126637,
+ "grad_norm": 0.01709570363163948,
+ "learning_rate": 0.0006,
+ "loss": 7.242851734161377,
+ "step": 167
+ },
+ {
+ "epoch": 2.3353711790393015,
+ "grad_norm": 0.012552388943731785,
+ "learning_rate": 0.0006,
+ "loss": 7.241696834564209,
+ "step": 168
+ },
+ {
+ "epoch": 2.349344978165939,
+ "grad_norm": 0.014571291394531727,
+ "learning_rate": 0.0006,
+ "loss": 7.270681381225586,
+ "step": 169
+ },
+ {
+ "epoch": 2.3633187772925766,
+ "grad_norm": 0.013389437459409237,
+ "learning_rate": 0.0006,
+ "loss": 7.191600799560547,
+ "step": 170
+ },
+ {
+ "epoch": 2.377292576419214,
+ "grad_norm": 0.011895314790308475,
+ "learning_rate": 0.0006,
+ "loss": 7.213610649108887,
+ "step": 171
+ },
+ {
+ "epoch": 2.3912663755458516,
+ "grad_norm": 0.01345642190426588,
+ "learning_rate": 0.0006,
+ "loss": 7.193787574768066,
+ "step": 172
+ },
+ {
+ "epoch": 2.405240174672489,
+ "grad_norm": 0.0091794328764081,
+ "learning_rate": 0.0006,
+ "loss": 7.233364105224609,
+ "step": 173
+ },
+ {
+ "epoch": 2.4192139737991267,
+ "grad_norm": 0.013767086900770664,
+ "learning_rate": 0.0006,
+ "loss": 7.217272758483887,
+ "step": 174
+ },
+ {
+ "epoch": 2.433187772925764,
+ "grad_norm": 0.015134960412979126,
+ "learning_rate": 0.0006,
+ "loss": 7.192791938781738,
+ "step": 175
+ },
+ {
+ "epoch": 2.447161572052402,
+ "grad_norm": 0.01790648326277733,
+ "learning_rate": 0.0006,
+ "loss": 7.229467391967773,
+ "step": 176
+ },
+ {
+ "epoch": 2.461135371179039,
+ "grad_norm": 0.015362377278506756,
+ "learning_rate": 0.0006,
+ "loss": 7.180346488952637,
+ "step": 177
+ },
+ {
+ "epoch": 2.475109170305677,
+ "grad_norm": 0.010505498386919498,
+ "learning_rate": 0.0006,
+ "loss": 7.188833236694336,
+ "step": 178
+ },
+ {
+ "epoch": 2.489082969432314,
+ "grad_norm": 0.009129747748374939,
+ "learning_rate": 0.0006,
+ "loss": 7.203153133392334,
+ "step": 179
+ },
+ {
+ "epoch": 2.503056768558952,
+ "grad_norm": 0.01147883478552103,
+ "learning_rate": 0.0006,
+ "loss": 7.163824558258057,
+ "step": 180
+ },
+ {
+ "epoch": 2.5170305676855893,
+ "grad_norm": 0.011338942684233189,
+ "learning_rate": 0.0006,
+ "loss": 7.16287899017334,
+ "step": 181
+ },
+ {
+ "epoch": 2.531004366812227,
+ "grad_norm": 0.013277675025165081,
+ "learning_rate": 0.0006,
+ "loss": 7.155905723571777,
+ "step": 182
+ },
+ {
+ "epoch": 2.544978165938865,
+ "grad_norm": 0.015269347466528416,
+ "learning_rate": 0.0006,
+ "loss": 7.189371109008789,
+ "step": 183
+ },
+ {
+ "epoch": 2.558951965065502,
+ "grad_norm": 0.017905596643686295,
+ "learning_rate": 0.0006,
+ "loss": 7.177547931671143,
+ "step": 184
+ },
+ {
+ "epoch": 2.5729257641921395,
+ "grad_norm": 0.024431610479950905,
+ "learning_rate": 0.0006,
+ "loss": 7.192169189453125,
+ "step": 185
+ },
+ {
+ "epoch": 2.5868995633187772,
+ "grad_norm": 0.03656961768865585,
+ "learning_rate": 0.0006,
+ "loss": 7.124387741088867,
+ "step": 186
+ },
+ {
+ "epoch": 2.600873362445415,
+ "grad_norm": 0.04337048903107643,
+ "learning_rate": 0.0006,
+ "loss": 7.140231609344482,
+ "step": 187
+ },
+ {
+ "epoch": 2.6148471615720523,
+ "grad_norm": 0.04047228768467903,
+ "learning_rate": 0.0006,
+ "loss": 7.171927452087402,
+ "step": 188
+ },
+ {
+ "epoch": 2.62882096069869,
+ "grad_norm": 0.07825770974159241,
+ "learning_rate": 0.0006,
+ "loss": 7.207631587982178,
+ "step": 189
+ },
+ {
+ "epoch": 2.6427947598253274,
+ "grad_norm": 0.07118063420057297,
+ "learning_rate": 0.0006,
+ "loss": 7.209488391876221,
+ "step": 190
+ },
+ {
+ "epoch": 2.656768558951965,
+ "grad_norm": 0.05153293535113335,
+ "learning_rate": 0.0006,
+ "loss": 7.14793062210083,
+ "step": 191
+ },
+ {
+ "epoch": 2.6707423580786025,
+ "grad_norm": 0.03318839520215988,
+ "learning_rate": 0.0006,
+ "loss": 7.163010597229004,
+ "step": 192
+ },
+ {
+ "epoch": 2.6847161572052403,
+ "grad_norm": 0.029333539307117462,
+ "learning_rate": 0.0006,
+ "loss": 7.159955978393555,
+ "step": 193
+ },
+ {
+ "epoch": 2.6986899563318776,
+ "grad_norm": 0.029405880719423294,
+ "learning_rate": 0.0006,
+ "loss": 7.136178493499756,
+ "step": 194
+ },
+ {
+ "epoch": 2.7126637554585153,
+ "grad_norm": 0.037815988063812256,
+ "learning_rate": 0.0006,
+ "loss": 7.183186054229736,
+ "step": 195
+ },
+ {
+ "epoch": 2.726637554585153,
+ "grad_norm": 0.020307250320911407,
+ "learning_rate": 0.0006,
+ "loss": 7.133164882659912,
+ "step": 196
+ },
+ {
+ "epoch": 2.7406113537117904,
+ "grad_norm": 0.026048069819808006,
+ "learning_rate": 0.0006,
+ "loss": 7.124977111816406,
+ "step": 197
+ },
+ {
+ "epoch": 2.7545851528384278,
+ "grad_norm": 0.027316724881529808,
+ "learning_rate": 0.0006,
+ "loss": 7.07265567779541,
+ "step": 198
+ },
+ {
+ "epoch": 2.7685589519650655,
+ "grad_norm": 0.021629009395837784,
+ "learning_rate": 0.0006,
+ "loss": 7.150020599365234,
+ "step": 199
+ },
+ {
+ "epoch": 2.7825327510917033,
+ "grad_norm": 0.01605929434299469,
+ "learning_rate": 0.0006,
+ "loss": 7.10919189453125,
+ "step": 200
+ },
+ {
+ "epoch": 2.7965065502183406,
+ "grad_norm": 0.020800089463591576,
+ "learning_rate": 0.0006,
+ "loss": 7.132045269012451,
+ "step": 201
+ },
+ {
+ "epoch": 2.810480349344978,
+ "grad_norm": 0.019225774332880974,
+ "learning_rate": 0.0006,
+ "loss": 7.128364562988281,
+ "step": 202
+ },
+ {
+ "epoch": 2.8244541484716157,
+ "grad_norm": 0.01561372634023428,
+ "learning_rate": 0.0006,
+ "loss": 7.083669662475586,
+ "step": 203
+ },
+ {
+ "epoch": 2.8384279475982535,
+ "grad_norm": 0.021969519555568695,
+ "learning_rate": 0.0006,
+ "loss": 7.050841331481934,
+ "step": 204
+ },
+ {
+ "epoch": 2.8524017467248908,
+ "grad_norm": 0.021453695371747017,
+ "learning_rate": 0.0006,
+ "loss": 7.097085475921631,
+ "step": 205
+ },
+ {
+ "epoch": 2.8663755458515285,
+ "grad_norm": 0.022836022078990936,
+ "learning_rate": 0.0006,
+ "loss": 7.10405969619751,
+ "step": 206
+ },
+ {
+ "epoch": 2.880349344978166,
+ "grad_norm": 0.02430851384997368,
+ "learning_rate": 0.0006,
+ "loss": 7.045181751251221,
+ "step": 207
+ },
+ {
+ "epoch": 2.8943231441048036,
+ "grad_norm": 0.020166993141174316,
+ "learning_rate": 0.0006,
+ "loss": 7.068183898925781,
+ "step": 208
+ },
+ {
+ "epoch": 2.908296943231441,
+ "grad_norm": 0.0118505684658885,
+ "learning_rate": 0.0006,
+ "loss": 7.07132625579834,
+ "step": 209
+ },
+ {
+ "epoch": 2.9222707423580787,
+ "grad_norm": 0.023772427812218666,
+ "learning_rate": 0.0006,
+ "loss": 7.055238723754883,
+ "step": 210
+ },
+ {
+ "epoch": 2.936244541484716,
+ "grad_norm": 0.02350712940096855,
+ "learning_rate": 0.0006,
+ "loss": 7.055898666381836,
+ "step": 211
+ },
+ {
+ "epoch": 2.950218340611354,
+ "grad_norm": 0.017244907096028328,
+ "learning_rate": 0.0006,
+ "loss": 7.0099711418151855,
+ "step": 212
+ },
+ {
+ "epoch": 2.964192139737991,
+ "grad_norm": 0.021565575152635574,
+ "learning_rate": 0.0006,
+ "loss": 6.991751670837402,
+ "step": 213
+ },
+ {
+ "epoch": 2.978165938864629,
+ "grad_norm": 0.03241860866546631,
+ "learning_rate": 0.0006,
+ "loss": 7.005980491638184,
+ "step": 214
+ },
+ {
+ "epoch": 2.992139737991266,
+ "grad_norm": 0.04613726586103439,
+ "learning_rate": 0.0006,
+ "loss": 7.027859687805176,
+ "step": 215
+ },
+ {
+ "epoch": 3.0,
+ "grad_norm": 0.06404894590377808,
+ "learning_rate": 0.0006,
+ "loss": 7.123725891113281,
+ "step": 216
+ },
+ {
+ "epoch": 3.0,
+ "eval_loss": 7.070647716522217,
+ "eval_runtime": 58.5801,
+ "eval_samples_per_second": 41.687,
+ "eval_steps_per_second": 1.314,
+ "step": 216
+ },
+ {
+ "epoch": 3.0139737991266378,
+ "grad_norm": 0.05542432889342308,
+ "learning_rate": 0.0006,
+ "loss": 7.078299522399902,
+ "step": 217
+ },
+ {
+ "epoch": 3.027947598253275,
+ "grad_norm": 0.026484526693820953,
+ "learning_rate": 0.0006,
+ "loss": 6.961159706115723,
+ "step": 218
+ },
+ {
+ "epoch": 3.041921397379913,
+ "grad_norm": 0.028276391327381134,
+ "learning_rate": 0.0006,
+ "loss": 7.019961833953857,
+ "step": 219
+ },
+ {
+ "epoch": 3.05589519650655,
+ "grad_norm": 0.024486595764756203,
+ "learning_rate": 0.0006,
+ "loss": 7.039677619934082,
+ "step": 220
+ },
+ {
+ "epoch": 3.069868995633188,
+ "grad_norm": 0.017620893195271492,
+ "learning_rate": 0.0006,
+ "loss": 6.9270782470703125,
+ "step": 221
+ },
+ {
+ "epoch": 3.0838427947598253,
+ "grad_norm": 0.02547302283346653,
+ "learning_rate": 0.0006,
+ "loss": 6.924233436584473,
+ "step": 222
+ },
+ {
+ "epoch": 3.097816593886463,
+ "grad_norm": 0.02539009042084217,
+ "learning_rate": 0.0006,
+ "loss": 6.973979949951172,
+ "step": 223
+ },
+ {
+ "epoch": 3.1117903930131003,
+ "grad_norm": 0.03256227448582649,
+ "learning_rate": 0.0006,
+ "loss": 6.933725357055664,
+ "step": 224
+ },
+ {
+ "epoch": 3.125764192139738,
+ "grad_norm": 0.04651800915598869,
+ "learning_rate": 0.0006,
+ "loss": 6.951044082641602,
+ "step": 225
+ },
+ {
+ "epoch": 3.1397379912663754,
+ "grad_norm": 0.04822975769639015,
+ "learning_rate": 0.0006,
+ "loss": 6.9614410400390625,
+ "step": 226
+ },
+ {
+ "epoch": 3.153711790393013,
+ "grad_norm": 0.031165020540356636,
+ "learning_rate": 0.0006,
+ "loss": 6.91193151473999,
+ "step": 227
+ },
+ {
+ "epoch": 3.1676855895196505,
+ "grad_norm": 0.029754292219877243,
+ "learning_rate": 0.0006,
+ "loss": 6.91802978515625,
+ "step": 228
+ },
+ {
+ "epoch": 3.1816593886462883,
+ "grad_norm": 0.02193152718245983,
+ "learning_rate": 0.0006,
+ "loss": 6.928090572357178,
+ "step": 229
+ },
+ {
+ "epoch": 3.1956331877729256,
+ "grad_norm": 0.02428470179438591,
+ "learning_rate": 0.0006,
+ "loss": 6.90396785736084,
+ "step": 230
+ },
+ {
+ "epoch": 3.2096069868995634,
+ "grad_norm": 0.023374345153570175,
+ "learning_rate": 0.0006,
+ "loss": 6.885858535766602,
+ "step": 231
+ },
+ {
+ "epoch": 3.2235807860262007,
+ "grad_norm": 0.023405000567436218,
+ "learning_rate": 0.0006,
+ "loss": 6.8855767250061035,
+ "step": 232
+ },
+ {
+ "epoch": 3.2375545851528384,
+ "grad_norm": 0.017277134582400322,
+ "learning_rate": 0.0006,
+ "loss": 6.901449203491211,
+ "step": 233
+ },
+ {
+ "epoch": 3.251528384279476,
+ "grad_norm": 0.019496750086545944,
+ "learning_rate": 0.0006,
+ "loss": 6.885687828063965,
+ "step": 234
+ },
+ {
+ "epoch": 3.2655021834061135,
+ "grad_norm": 0.016893276944756508,
+ "learning_rate": 0.0006,
+ "loss": 6.872369766235352,
+ "step": 235
+ },
+ {
+ "epoch": 3.279475982532751,
+ "grad_norm": 0.02058715932071209,
+ "learning_rate": 0.0006,
+ "loss": 6.823280334472656,
+ "step": 236
+ },
+ {
+ "epoch": 3.2934497816593886,
+ "grad_norm": 0.015530755743384361,
+ "learning_rate": 0.0006,
+ "loss": 6.860499858856201,
+ "step": 237
+ },
+ {
+ "epoch": 3.3074235807860264,
+ "grad_norm": 0.01934926211833954,
+ "learning_rate": 0.0006,
+ "loss": 6.817448616027832,
+ "step": 238
+ },
+ {
+ "epoch": 3.3213973799126637,
+ "grad_norm": 0.017413552850484848,
+ "learning_rate": 0.0006,
+ "loss": 6.881955146789551,
+ "step": 239
+ },
+ {
+ "epoch": 3.3353711790393015,
+ "grad_norm": 0.026478875428438187,
+ "learning_rate": 0.0006,
+ "loss": 6.827404499053955,
+ "step": 240
+ },
+ {
+ "epoch": 3.349344978165939,
+ "grad_norm": 0.032611701637506485,
+ "learning_rate": 0.0006,
+ "loss": 6.78285551071167,
+ "step": 241
+ },
+ {
+ "epoch": 3.3633187772925766,
+ "grad_norm": 0.041900306940078735,
+ "learning_rate": 0.0006,
+ "loss": 6.783053874969482,
+ "step": 242
+ },
+ {
+ "epoch": 3.377292576419214,
+ "grad_norm": 0.04149497672915459,
+ "learning_rate": 0.0006,
+ "loss": 6.745220184326172,
+ "step": 243
+ },
+ {
+ "epoch": 3.3912663755458516,
+ "grad_norm": 0.023949826136231422,
+ "learning_rate": 0.0006,
+ "loss": 6.841533184051514,
+ "step": 244
+ },
+ {
+ "epoch": 3.405240174672489,
+ "grad_norm": 0.028601842001080513,
+ "learning_rate": 0.0006,
+ "loss": 6.852852821350098,
+ "step": 245
+ },
+ {
+ "epoch": 3.4192139737991267,
+ "grad_norm": 0.027365155518054962,
+ "learning_rate": 0.0006,
+ "loss": 6.74376106262207,
+ "step": 246
+ },
+ {
+ "epoch": 3.433187772925764,
+ "grad_norm": 0.02856568619608879,
+ "learning_rate": 0.0006,
+ "loss": 6.74906063079834,
+ "step": 247
+ },
+ {
+ "epoch": 3.447161572052402,
+ "grad_norm": 0.019172416999936104,
+ "learning_rate": 0.0006,
+ "loss": 6.809206008911133,
+ "step": 248
+ },
+ {
+ "epoch": 3.461135371179039,
+ "grad_norm": 0.02161746844649315,
+ "learning_rate": 0.0006,
+ "loss": 6.789888381958008,
+ "step": 249
+ },
+ {
+ "epoch": 3.475109170305677,
+ "grad_norm": 0.020273666828870773,
+ "learning_rate": 0.0006,
+ "loss": 6.73512077331543,
+ "step": 250
+ },
+ {
+ "epoch": 3.489082969432314,
+ "grad_norm": 0.022329283878207207,
+ "learning_rate": 0.0006,
+ "loss": 6.826148509979248,
+ "step": 251
+ },
+ {
+ "epoch": 3.503056768558952,
+ "grad_norm": 0.02284284122288227,
+ "learning_rate": 0.0006,
+ "loss": 6.748521327972412,
+ "step": 252
+ },
+ {
+ "epoch": 3.5170305676855893,
+ "grad_norm": 0.026368247345089912,
+ "learning_rate": 0.0006,
+ "loss": 6.645879745483398,
+ "step": 253
+ },
+ {
+ "epoch": 3.531004366812227,
+ "grad_norm": 0.03627244383096695,
+ "learning_rate": 0.0006,
+ "loss": 6.666064739227295,
+ "step": 254
+ },
+ {
+ "epoch": 3.544978165938865,
+ "grad_norm": 0.04140935093164444,
+ "learning_rate": 0.0006,
+ "loss": 6.712275981903076,
+ "step": 255
+ },
+ {
+ "epoch": 3.558951965065502,
+ "grad_norm": 0.0390462800860405,
+ "learning_rate": 0.0006,
+ "loss": 6.665900230407715,
+ "step": 256
+ },
+ {
+ "epoch": 3.5729257641921395,
+ "grad_norm": 0.05178583785891533,
+ "learning_rate": 0.0006,
+ "loss": 6.753298759460449,
+ "step": 257
+ },
+ {
+ "epoch": 3.5868995633187772,
+ "grad_norm": 0.054328061640262604,
+ "learning_rate": 0.0006,
+ "loss": 6.732633113861084,
+ "step": 258
+ },
+ {
+ "epoch": 3.600873362445415,
+ "grad_norm": 0.04590460658073425,
+ "learning_rate": 0.0006,
+ "loss": 6.65250301361084,
+ "step": 259
+ },
+ {
+ "epoch": 3.6148471615720523,
+ "grad_norm": 0.044331666082143784,
+ "learning_rate": 0.0006,
+ "loss": 6.68654203414917,
+ "step": 260
+ },
+ {
+ "epoch": 3.62882096069869,
+ "grad_norm": 0.05308730527758598,
+ "learning_rate": 0.0006,
+ "loss": 6.642482757568359,
+ "step": 261
+ },
+ {
+ "epoch": 3.6427947598253274,
+ "grad_norm": 0.03999016433954239,
+ "learning_rate": 0.0006,
+ "loss": 6.625227451324463,
+ "step": 262
+ },
+ {
+ "epoch": 3.656768558951965,
+ "grad_norm": 0.04088086634874344,
+ "learning_rate": 0.0006,
+ "loss": 6.67623233795166,
+ "step": 263
+ },
+ {
+ "epoch": 3.6707423580786025,
+ "grad_norm": 0.03615636005997658,
+ "learning_rate": 0.0006,
+ "loss": 6.6869001388549805,
+ "step": 264
+ },
+ {
+ "epoch": 3.6847161572052403,
+ "grad_norm": 0.02817360684275627,
+ "learning_rate": 0.0006,
+ "loss": 6.674601078033447,
+ "step": 265
+ },
+ {
+ "epoch": 3.6986899563318776,
+ "grad_norm": 0.032586123794317245,
+ "learning_rate": 0.0006,
+ "loss": 6.612994194030762,
+ "step": 266
+ },
+ {
+ "epoch": 3.7126637554585153,
+ "grad_norm": 0.031117867678403854,
+ "learning_rate": 0.0006,
+ "loss": 6.6244354248046875,
+ "step": 267
+ },
+ {
+ "epoch": 3.726637554585153,
+ "grad_norm": 0.028330687433481216,
+ "learning_rate": 0.0006,
+ "loss": 6.669557094573975,
+ "step": 268
+ },
+ {
+ "epoch": 3.7406113537117904,
+ "grad_norm": 0.021489202976226807,
+ "learning_rate": 0.0006,
+ "loss": 6.632035732269287,
+ "step": 269
+ },
+ {
+ "epoch": 3.7545851528384278,
+ "grad_norm": 0.024269424378871918,
+ "learning_rate": 0.0006,
+ "loss": 6.660867691040039,
+ "step": 270
+ },
+ {
+ "epoch": 3.7685589519650655,
+ "grad_norm": 0.02237873338162899,
+ "learning_rate": 0.0006,
+ "loss": 6.602587699890137,
+ "step": 271
+ },
+ {
+ "epoch": 3.7825327510917033,
+ "grad_norm": 0.026503929868340492,
+ "learning_rate": 0.0006,
+ "loss": 6.599200248718262,
+ "step": 272
+ },
+ {
+ "epoch": 3.7965065502183406,
+ "grad_norm": 0.03687124326825142,
+ "learning_rate": 0.0006,
+ "loss": 6.567296504974365,
+ "step": 273
+ },
+ {
+ "epoch": 3.810480349344978,
+ "grad_norm": 0.04708952456712723,
+ "learning_rate": 0.0006,
+ "loss": 6.621265411376953,
+ "step": 274
+ },
+ {
+ "epoch": 3.8244541484716157,
+ "grad_norm": 0.041846614331007004,
+ "learning_rate": 0.0006,
+ "loss": 6.550536155700684,
+ "step": 275
+ },
+ {
+ "epoch": 3.8384279475982535,
+ "grad_norm": 0.031249675899744034,
+ "learning_rate": 0.0006,
+ "loss": 6.54278564453125,
+ "step": 276
+ },
+ {
+ "epoch": 3.8524017467248908,
+ "grad_norm": 0.023082256317138672,
+ "learning_rate": 0.0006,
+ "loss": 6.544787406921387,
+ "step": 277
+ },
+ {
+ "epoch": 3.8663755458515285,
+ "grad_norm": 0.020995570346713066,
+ "learning_rate": 0.0006,
+ "loss": 6.518143653869629,
+ "step": 278
+ },
+ {
+ "epoch": 3.880349344978166,
+ "grad_norm": 0.020839324221014977,
+ "learning_rate": 0.0006,
+ "loss": 6.546667098999023,
+ "step": 279
+ },
+ {
+ "epoch": 3.8943231441048036,
+ "grad_norm": 0.018377432599663734,
+ "learning_rate": 0.0006,
+ "loss": 6.478307723999023,
+ "step": 280
+ },
+ {
+ "epoch": 3.908296943231441,
+ "grad_norm": 0.021713724359869957,
+ "learning_rate": 0.0006,
+ "loss": 6.485574245452881,
+ "step": 281
+ },
+ {
+ "epoch": 3.9222707423580787,
+ "grad_norm": 0.01359301246702671,
+ "learning_rate": 0.0006,
+ "loss": 6.570268630981445,
+ "step": 282
+ },
+ {
+ "epoch": 3.936244541484716,
+ "grad_norm": 0.016233332455158234,
+ "learning_rate": 0.0006,
+ "loss": 6.57150936126709,
+ "step": 283
+ },
+ {
+ "epoch": 3.950218340611354,
+ "grad_norm": 0.020687608048319817,
+ "learning_rate": 0.0006,
+ "loss": 6.527956008911133,
+ "step": 284
+ },
+ {
+ "epoch": 3.964192139737991,
+ "grad_norm": 0.02032964862883091,
+ "learning_rate": 0.0006,
+ "loss": 6.468169689178467,
+ "step": 285
+ },
+ {
+ "epoch": 3.978165938864629,
+ "grad_norm": 0.016676288098096848,
+ "learning_rate": 0.0006,
+ "loss": 6.448418617248535,
+ "step": 286
+ },
+ {
+ "epoch": 3.992139737991266,
+ "grad_norm": 0.0168539360165596,
+ "learning_rate": 0.0006,
+ "loss": 6.461180686950684,
+ "step": 287
+ },
+ {
+ "epoch": 4.0,
+ "grad_norm": 0.02666233666241169,
+ "learning_rate": 0.0006,
+ "loss": 6.344893932342529,
+ "step": 288
+ },
+ {
+ "epoch": 4.0,
+ "eval_loss": 6.5384979248046875,
+ "eval_runtime": 58.814,
+ "eval_samples_per_second": 41.521,
+ "eval_steps_per_second": 1.309,
+ "step": 288
+ },
+ {
+ "epoch": 4.013973799126638,
+ "grad_norm": 0.04154708981513977,
+ "learning_rate": 0.0006,
+ "loss": 6.41855525970459,
+ "step": 289
+ },
+ {
+ "epoch": 4.0279475982532755,
+ "grad_norm": 0.04895668104290962,
+ "learning_rate": 0.0006,
+ "loss": 6.499302864074707,
+ "step": 290
+ },
+ {
+ "epoch": 4.041921397379912,
+ "grad_norm": 0.034464482218027115,
+ "learning_rate": 0.0006,
+ "loss": 6.438611030578613,
+ "step": 291
+ },
+ {
+ "epoch": 4.05589519650655,
+ "grad_norm": 0.06845773756504059,
+ "learning_rate": 0.0006,
+ "loss": 6.538361072540283,
+ "step": 292
+ },
+ {
+ "epoch": 4.069868995633188,
+ "grad_norm": 0.053768858313560486,
+ "learning_rate": 0.0006,
+ "loss": 6.527524948120117,
+ "step": 293
+ },
+ {
+ "epoch": 4.083842794759826,
+ "grad_norm": 0.039037931710481644,
+ "learning_rate": 0.0006,
+ "loss": 6.503847122192383,
+ "step": 294
+ },
+ {
+ "epoch": 4.097816593886463,
+ "grad_norm": 0.03962196782231331,
+ "learning_rate": 0.0006,
+ "loss": 6.475178241729736,
+ "step": 295
+ },
+ {
+ "epoch": 4.1117903930131,
+ "grad_norm": 0.049101535230875015,
+ "learning_rate": 0.0006,
+ "loss": 6.474587917327881,
+ "step": 296
+ },
+ {
+ "epoch": 4.125764192139738,
+ "grad_norm": 0.02830282226204872,
+ "learning_rate": 0.0006,
+ "loss": 6.45065975189209,
+ "step": 297
+ },
+ {
+ "epoch": 4.139737991266376,
+ "grad_norm": 0.023153482005000114,
+ "learning_rate": 0.0006,
+ "loss": 6.5218706130981445,
+ "step": 298
+ },
+ {
+ "epoch": 4.153711790393013,
+ "grad_norm": 0.025064002722501755,
+ "learning_rate": 0.0006,
+ "loss": 6.456475257873535,
+ "step": 299
+ },
+ {
+ "epoch": 4.1676855895196505,
+ "grad_norm": 0.02357092685997486,
+ "learning_rate": 0.0006,
+ "loss": 6.485952377319336,
+ "step": 300
+ },
+ {
+ "epoch": 4.181659388646288,
+ "grad_norm": 0.021832305938005447,
+ "learning_rate": 0.0006,
+ "loss": 6.425506591796875,
+ "step": 301
+ },
+ {
+ "epoch": 4.195633187772926,
+ "grad_norm": 0.0227900929749012,
+ "learning_rate": 0.0006,
+ "loss": 6.439155578613281,
+ "step": 302
+ },
+ {
+ "epoch": 4.209606986899563,
+ "grad_norm": 0.015961607918143272,
+ "learning_rate": 0.0006,
+ "loss": 6.460862636566162,
+ "step": 303
+ },
+ {
+ "epoch": 4.223580786026201,
+ "grad_norm": 0.017876390367746353,
+ "learning_rate": 0.0006,
+ "loss": 6.44564151763916,
+ "step": 304
+ },
+ {
+ "epoch": 4.2375545851528384,
+ "grad_norm": 0.013703204691410065,
+ "learning_rate": 0.0006,
+ "loss": 6.418414115905762,
+ "step": 305
+ },
+ {
+ "epoch": 4.251528384279476,
+ "grad_norm": 0.015240306034684181,
+ "learning_rate": 0.0006,
+ "loss": 6.411238670349121,
+ "step": 306
+ },
+ {
+ "epoch": 4.265502183406113,
+ "grad_norm": 0.01587662845849991,
+ "learning_rate": 0.0006,
+ "loss": 6.395642280578613,
+ "step": 307
+ },
+ {
+ "epoch": 4.279475982532751,
+ "grad_norm": 0.014650700613856316,
+ "learning_rate": 0.0006,
+ "loss": 6.302793025970459,
+ "step": 308
+ },
+ {
+ "epoch": 4.293449781659389,
+ "grad_norm": 0.016533225774765015,
+ "learning_rate": 0.0006,
+ "loss": 6.380588054656982,
+ "step": 309
+ },
+ {
+ "epoch": 4.307423580786026,
+ "grad_norm": 0.019963741302490234,
+ "learning_rate": 0.0006,
+ "loss": 6.345336437225342,
+ "step": 310
+ },
+ {
+ "epoch": 4.321397379912664,
+ "grad_norm": 0.02237936295568943,
+ "learning_rate": 0.0006,
+ "loss": 6.387770652770996,
+ "step": 311
+ },
+ {
+ "epoch": 4.335371179039301,
+ "grad_norm": 0.026658328250050545,
+ "learning_rate": 0.0006,
+ "loss": 6.323662757873535,
+ "step": 312
+ },
+ {
+ "epoch": 4.349344978165939,
+ "grad_norm": 0.02852284163236618,
+ "learning_rate": 0.0006,
+ "loss": 6.382204055786133,
+ "step": 313
+ },
+ {
+ "epoch": 4.3633187772925766,
+ "grad_norm": 0.028815090656280518,
+ "learning_rate": 0.0006,
+ "loss": 6.402390480041504,
+ "step": 314
+ },
+ {
+ "epoch": 4.377292576419214,
+ "grad_norm": 0.029393529519438744,
+ "learning_rate": 0.0006,
+ "loss": 6.280278205871582,
+ "step": 315
+ },
+ {
+ "epoch": 4.391266375545851,
+ "grad_norm": 0.02186041697859764,
+ "learning_rate": 0.0006,
+ "loss": 6.302707672119141,
+ "step": 316
+ },
+ {
+ "epoch": 4.405240174672489,
+ "grad_norm": 0.02071218006312847,
+ "learning_rate": 0.0006,
+ "loss": 6.3365020751953125,
+ "step": 317
+ },
+ {
+ "epoch": 4.419213973799127,
+ "grad_norm": 0.02779117226600647,
+ "learning_rate": 0.0006,
+ "loss": 6.252157211303711,
+ "step": 318
+ },
+ {
+ "epoch": 4.4331877729257645,
+ "grad_norm": 0.03470654785633087,
+ "learning_rate": 0.0006,
+ "loss": 6.319070339202881,
+ "step": 319
+ },
+ {
+ "epoch": 4.447161572052401,
+ "grad_norm": 0.037067804485559464,
+ "learning_rate": 0.0006,
+ "loss": 6.314116477966309,
+ "step": 320
+ },
+ {
+ "epoch": 4.461135371179039,
+ "grad_norm": 0.032182756811380386,
+ "learning_rate": 0.0006,
+ "loss": 6.267298698425293,
+ "step": 321
+ },
+ {
+ "epoch": 4.475109170305677,
+ "grad_norm": 0.026305217295885086,
+ "learning_rate": 0.0006,
+ "loss": 6.331688404083252,
+ "step": 322
+ },
+ {
+ "epoch": 4.489082969432315,
+ "grad_norm": 0.027228357270359993,
+ "learning_rate": 0.0006,
+ "loss": 6.31139612197876,
+ "step": 323
+ },
+ {
+ "epoch": 4.503056768558952,
+ "grad_norm": 0.025781169533729553,
+ "learning_rate": 0.0006,
+ "loss": 6.194684982299805,
+ "step": 324
+ },
+ {
+ "epoch": 4.517030567685589,
+ "grad_norm": 0.029186977073550224,
+ "learning_rate": 0.0006,
+ "loss": 6.31306266784668,
+ "step": 325
+ },
+ {
+ "epoch": 4.531004366812227,
+ "grad_norm": 0.023608777672052383,
+ "learning_rate": 0.0006,
+ "loss": 6.285243034362793,
+ "step": 326
+ },
+ {
+ "epoch": 4.544978165938865,
+ "grad_norm": 0.02267594076693058,
+ "learning_rate": 0.0006,
+ "loss": 6.30342960357666,
+ "step": 327
+ },
+ {
+ "epoch": 4.558951965065502,
+ "grad_norm": 0.01926310732960701,
+ "learning_rate": 0.0006,
+ "loss": 6.280606269836426,
+ "step": 328
+ },
+ {
+ "epoch": 4.5729257641921395,
+ "grad_norm": 0.026484837755560875,
+ "learning_rate": 0.0006,
+ "loss": 6.237980365753174,
+ "step": 329
+ },
+ {
+ "epoch": 4.586899563318777,
+ "grad_norm": 0.026228854432702065,
+ "learning_rate": 0.0006,
+ "loss": 6.3018293380737305,
+ "step": 330
+ },
+ {
+ "epoch": 4.600873362445415,
+ "grad_norm": 0.022096967324614525,
+ "learning_rate": 0.0006,
+ "loss": 6.253863334655762,
+ "step": 331
+ },
+ {
+ "epoch": 4.614847161572053,
+ "grad_norm": 0.027223890647292137,
+ "learning_rate": 0.0006,
+ "loss": 6.166882038116455,
+ "step": 332
+ },
+ {
+ "epoch": 4.62882096069869,
+ "grad_norm": 0.03160124272108078,
+ "learning_rate": 0.0006,
+ "loss": 6.230309963226318,
+ "step": 333
+ },
+ {
+ "epoch": 4.642794759825327,
+ "grad_norm": 0.04110539332032204,
+ "learning_rate": 0.0006,
+ "loss": 6.2334818840026855,
+ "step": 334
+ },
+ {
+ "epoch": 4.656768558951965,
+ "grad_norm": 0.04355933889746666,
+ "learning_rate": 0.0006,
+ "loss": 6.160956382751465,
+ "step": 335
+ },
+ {
+ "epoch": 4.670742358078603,
+ "grad_norm": 0.03611086308956146,
+ "learning_rate": 0.0006,
+ "loss": 6.23399543762207,
+ "step": 336
+ },
+ {
+ "epoch": 4.68471615720524,
+ "grad_norm": 0.051811400800943375,
+ "learning_rate": 0.0006,
+ "loss": 6.241855144500732,
+ "step": 337
+ },
+ {
+ "epoch": 4.698689956331878,
+ "grad_norm": 0.057434193789958954,
+ "learning_rate": 0.0006,
+ "loss": 6.220993995666504,
+ "step": 338
+ },
+ {
+ "epoch": 4.712663755458515,
+ "grad_norm": 0.04420953616499901,
+ "learning_rate": 0.0006,
+ "loss": 6.300461292266846,
+ "step": 339
+ },
+ {
+ "epoch": 4.726637554585153,
+ "grad_norm": 0.03515457361936569,
+ "learning_rate": 0.0006,
+ "loss": 6.269349098205566,
+ "step": 340
+ },
+ {
+ "epoch": 4.74061135371179,
+ "grad_norm": 0.03222600743174553,
+ "learning_rate": 0.0006,
+ "loss": 6.248613357543945,
+ "step": 341
+ },
+ {
+ "epoch": 4.754585152838428,
+ "grad_norm": 0.0384046845138073,
+ "learning_rate": 0.0006,
+ "loss": 6.266979217529297,
+ "step": 342
+ },
+ {
+ "epoch": 4.7685589519650655,
+ "grad_norm": 0.0514867827296257,
+ "learning_rate": 0.0006,
+ "loss": 6.234103202819824,
+ "step": 343
+ },
+ {
+ "epoch": 4.782532751091703,
+ "grad_norm": 0.050747793167829514,
+ "learning_rate": 0.0006,
+ "loss": 6.322582244873047,
+ "step": 344
+ },
+ {
+ "epoch": 4.796506550218341,
+ "grad_norm": 0.040878988802433014,
+ "learning_rate": 0.0006,
+ "loss": 6.212509632110596,
+ "step": 345
+ },
+ {
+ "epoch": 4.810480349344978,
+ "grad_norm": 0.034655820578336716,
+ "learning_rate": 0.0006,
+ "loss": 6.265879154205322,
+ "step": 346
+ },
+ {
+ "epoch": 4.824454148471616,
+ "grad_norm": 0.04202224686741829,
+ "learning_rate": 0.0006,
+ "loss": 6.2896409034729,
+ "step": 347
+ },
+ {
+ "epoch": 4.8384279475982535,
+ "grad_norm": 0.03632277995347977,
+ "learning_rate": 0.0006,
+ "loss": 6.253917217254639,
+ "step": 348
+ },
+ {
+ "epoch": 4.85240174672489,
+ "grad_norm": 0.028645765036344528,
+ "learning_rate": 0.0006,
+ "loss": 6.207965850830078,
+ "step": 349
+ },
+ {
+ "epoch": 4.866375545851528,
+ "grad_norm": 0.025305170565843582,
+ "learning_rate": 0.0006,
+ "loss": 6.168068885803223,
+ "step": 350
+ },
+ {
+ "epoch": 4.880349344978166,
+ "grad_norm": 0.022137803956866264,
+ "learning_rate": 0.0006,
+ "loss": 6.235283374786377,
+ "step": 351
+ },
+ {
+ "epoch": 4.894323144104804,
+ "grad_norm": 0.01814538985490799,
+ "learning_rate": 0.0006,
+ "loss": 6.259641647338867,
+ "step": 352
+ },
+ {
+ "epoch": 4.908296943231441,
+ "grad_norm": 0.018465086817741394,
+ "learning_rate": 0.0006,
+ "loss": 6.111738681793213,
+ "step": 353
+ },
+ {
+ "epoch": 4.922270742358078,
+ "grad_norm": 0.018638933077454567,
+ "learning_rate": 0.0006,
+ "loss": 6.225642204284668,
+ "step": 354
+ },
+ {
+ "epoch": 4.936244541484716,
+ "grad_norm": 0.016461260616779327,
+ "learning_rate": 0.0006,
+ "loss": 6.140647888183594,
+ "step": 355
+ },
+ {
+ "epoch": 4.950218340611354,
+ "grad_norm": 0.015458385460078716,
+ "learning_rate": 0.0006,
+ "loss": 6.155224800109863,
+ "step": 356
+ },
+ {
+ "epoch": 4.964192139737992,
+ "grad_norm": 0.014934048056602478,
+ "learning_rate": 0.0006,
+ "loss": 6.101555824279785,
+ "step": 357
+ },
+ {
+ "epoch": 4.978165938864628,
+ "grad_norm": 0.013979545794427395,
+ "learning_rate": 0.0006,
+ "loss": 6.161718845367432,
+ "step": 358
+ },
+ {
+ "epoch": 4.992139737991266,
+ "grad_norm": 0.01640394888818264,
+ "learning_rate": 0.0006,
+ "loss": 6.055768966674805,
+ "step": 359
+ },
+ {
+ "epoch": 5.0,
+ "grad_norm": 0.015135680325329304,
+ "learning_rate": 0.0006,
+ "loss": 6.169695854187012,
+ "step": 360
+ },
+ {
+ "epoch": 5.0,
+ "eval_loss": 6.16408109664917,
+ "eval_runtime": 59.3649,
+ "eval_samples_per_second": 41.135,
+ "eval_steps_per_second": 1.297,
+ "step": 360
+ },
+ {
+ "epoch": 5.013973799126638,
+ "grad_norm": 0.012849048711359501,
+ "learning_rate": 0.0006,
+ "loss": 6.055830001831055,
+ "step": 361
+ },
+ {
+ "epoch": 5.0279475982532755,
+ "grad_norm": 0.018169576302170753,
+ "learning_rate": 0.0006,
+ "loss": 6.082655429840088,
+ "step": 362
+ },
+ {
+ "epoch": 5.041921397379912,
+ "grad_norm": 0.024293430149555206,
+ "learning_rate": 0.0006,
+ "loss": 6.109032154083252,
+ "step": 363
+ },
+ {
+ "epoch": 5.05589519650655,
+ "grad_norm": 0.03266787901520729,
+ "learning_rate": 0.0006,
+ "loss": 6.030613899230957,
+ "step": 364
+ },
+ {
+ "epoch": 5.069868995633188,
+ "grad_norm": 0.03572266176342964,
+ "learning_rate": 0.0006,
+ "loss": 6.050390720367432,
+ "step": 365
+ },
+ {
+ "epoch": 5.083842794759826,
+ "grad_norm": 0.032625213265419006,
+ "learning_rate": 0.0006,
+ "loss": 6.169313430786133,
+ "step": 366
+ },
+ {
+ "epoch": 5.097816593886463,
+ "grad_norm": 0.04051872715353966,
+ "learning_rate": 0.0006,
+ "loss": 6.10398006439209,
+ "step": 367
+ },
+ {
+ "epoch": 5.1117903930131,
+ "grad_norm": 0.0351213738322258,
+ "learning_rate": 0.0006,
+ "loss": 6.121973514556885,
+ "step": 368
+ },
+ {
+ "epoch": 5.125764192139738,
+ "grad_norm": 0.03523759916424751,
+ "learning_rate": 0.0006,
+ "loss": 6.0626654624938965,
+ "step": 369
+ },
+ {
+ "epoch": 5.139737991266376,
+ "grad_norm": 0.02706182189285755,
+ "learning_rate": 0.0006,
+ "loss": 6.049354076385498,
+ "step": 370
+ },
+ {
+ "epoch": 5.153711790393013,
+ "grad_norm": 0.024212589487433434,
+ "learning_rate": 0.0006,
+ "loss": 6.034826278686523,
+ "step": 371
+ },
+ {
+ "epoch": 5.1676855895196505,
+ "grad_norm": 0.024798082187771797,
+ "learning_rate": 0.0006,
+ "loss": 6.004058837890625,
+ "step": 372
+ },
+ {
+ "epoch": 5.181659388646288,
+ "grad_norm": 0.02325567416846752,
+ "learning_rate": 0.0006,
+ "loss": 5.986461639404297,
+ "step": 373
+ },
+ {
+ "epoch": 5.195633187772926,
+ "grad_norm": 0.019060570746660233,
+ "learning_rate": 0.0006,
+ "loss": 6.059736251831055,
+ "step": 374
+ },
+ {
+ "epoch": 5.209606986899563,
+ "grad_norm": 0.016822976991534233,
+ "learning_rate": 0.0006,
+ "loss": 6.072957515716553,
+ "step": 375
+ },
+ {
+ "epoch": 5.223580786026201,
+ "grad_norm": 0.017218658700585365,
+ "learning_rate": 0.0006,
+ "loss": 6.031739234924316,
+ "step": 376
+ },
+ {
+ "epoch": 5.2375545851528384,
+ "grad_norm": 0.014889383688569069,
+ "learning_rate": 0.0006,
+ "loss": 6.025674343109131,
+ "step": 377
+ },
+ {
+ "epoch": 5.251528384279476,
+ "grad_norm": 0.01708107627928257,
+ "learning_rate": 0.0006,
+ "loss": 6.090451240539551,
+ "step": 378
+ },
+ {
+ "epoch": 5.265502183406113,
+ "grad_norm": 0.020260317251086235,
+ "learning_rate": 0.0006,
+ "loss": 6.094257354736328,
+ "step": 379
+ },
+ {
+ "epoch": 5.279475982532751,
+ "grad_norm": 0.020110400393605232,
+ "learning_rate": 0.0006,
+ "loss": 6.021188735961914,
+ "step": 380
+ },
+ {
+ "epoch": 5.293449781659389,
+ "grad_norm": 0.02027830481529236,
+ "learning_rate": 0.0006,
+ "loss": 6.022156238555908,
+ "step": 381
+ },
+ {
+ "epoch": 5.307423580786026,
+ "grad_norm": 0.022747062146663666,
+ "learning_rate": 0.0006,
+ "loss": 6.011836051940918,
+ "step": 382
+ },
+ {
+ "epoch": 5.321397379912664,
+ "grad_norm": 0.01990230567753315,
+ "learning_rate": 0.0006,
+ "loss": 6.003054618835449,
+ "step": 383
+ },
+ {
+ "epoch": 5.335371179039301,
+ "grad_norm": 0.01596238650381565,
+ "learning_rate": 0.0006,
+ "loss": 5.991410732269287,
+ "step": 384
+ },
+ {
+ "epoch": 5.349344978165939,
+ "grad_norm": 0.015847105532884598,
+ "learning_rate": 0.0006,
+ "loss": 6.100622177124023,
+ "step": 385
+ },
+ {
+ "epoch": 5.3633187772925766,
+ "grad_norm": 0.016179397702217102,
+ "learning_rate": 0.0006,
+ "loss": 6.032659530639648,
+ "step": 386
+ },
+ {
+ "epoch": 5.377292576419214,
+ "grad_norm": 0.018256602808833122,
+ "learning_rate": 0.0006,
+ "loss": 5.961983680725098,
+ "step": 387
+ },
+ {
+ "epoch": 5.391266375545851,
+ "grad_norm": 0.02805912122130394,
+ "learning_rate": 0.0006,
+ "loss": 5.983541965484619,
+ "step": 388
+ },
+ {
+ "epoch": 5.405240174672489,
+ "grad_norm": 0.039082255214452744,
+ "learning_rate": 0.0006,
+ "loss": 6.021474361419678,
+ "step": 389
+ },
+ {
+ "epoch": 5.419213973799127,
+ "grad_norm": 0.036757659167051315,
+ "learning_rate": 0.0006,
+ "loss": 6.005046367645264,
+ "step": 390
+ },
+ {
+ "epoch": 5.4331877729257645,
+ "grad_norm": 0.035277366638183594,
+ "learning_rate": 0.0006,
+ "loss": 6.036296844482422,
+ "step": 391
+ },
+ {
+ "epoch": 5.447161572052401,
+ "grad_norm": 0.034404635429382324,
+ "learning_rate": 0.0006,
+ "loss": 5.994539260864258,
+ "step": 392
+ },
+ {
+ "epoch": 5.461135371179039,
+ "grad_norm": 0.0377206988632679,
+ "learning_rate": 0.0006,
+ "loss": 5.941534519195557,
+ "step": 393
+ },
+ {
+ "epoch": 5.475109170305677,
+ "grad_norm": 0.0389922633767128,
+ "learning_rate": 0.0006,
+ "loss": 5.983644485473633,
+ "step": 394
+ },
+ {
+ "epoch": 5.489082969432315,
+ "grad_norm": 0.04176778718829155,
+ "learning_rate": 0.0006,
+ "loss": 6.030922889709473,
+ "step": 395
+ },
+ {
+ "epoch": 5.503056768558952,
+ "grad_norm": 0.0337153784930706,
+ "learning_rate": 0.0006,
+ "loss": 5.961367130279541,
+ "step": 396
+ },
+ {
+ "epoch": 5.517030567685589,
+ "grad_norm": 0.036116816103458405,
+ "learning_rate": 0.0006,
+ "loss": 5.819280624389648,
+ "step": 397
+ },
+ {
+ "epoch": 5.531004366812227,
+ "grad_norm": 0.030724041163921356,
+ "learning_rate": 0.0006,
+ "loss": 5.906380653381348,
+ "step": 398
+ },
+ {
+ "epoch": 5.544978165938865,
+ "grad_norm": 0.030264202505350113,
+ "learning_rate": 0.0006,
+ "loss": 5.975833415985107,
+ "step": 399
+ },
+ {
+ "epoch": 5.558951965065502,
+ "grad_norm": 0.032096318900585175,
+ "learning_rate": 0.0006,
+ "loss": 6.001348972320557,
+ "step": 400
+ },
+ {
+ "epoch": 5.5729257641921395,
+ "grad_norm": 0.030579356476664543,
+ "learning_rate": 0.0006,
+ "loss": 5.928768634796143,
+ "step": 401
+ },
+ {
+ "epoch": 5.586899563318777,
+ "grad_norm": 0.028241973370313644,
+ "learning_rate": 0.0006,
+ "loss": 5.92582893371582,
+ "step": 402
+ },
+ {
+ "epoch": 5.600873362445415,
+ "grad_norm": 0.023086953908205032,
+ "learning_rate": 0.0006,
+ "loss": 5.892926216125488,
+ "step": 403
+ },
+ {
+ "epoch": 5.614847161572053,
+ "grad_norm": 0.02547992393374443,
+ "learning_rate": 0.0006,
+ "loss": 5.903195381164551,
+ "step": 404
+ },
+ {
+ "epoch": 5.62882096069869,
+ "grad_norm": 0.023089831694960594,
+ "learning_rate": 0.0006,
+ "loss": 5.912033557891846,
+ "step": 405
+ },
+ {
+ "epoch": 5.642794759825327,
+ "grad_norm": 0.025446368381381035,
+ "learning_rate": 0.0006,
+ "loss": 5.926138401031494,
+ "step": 406
+ },
+ {
+ "epoch": 5.656768558951965,
+ "grad_norm": 0.03748747706413269,
+ "learning_rate": 0.0006,
+ "loss": 5.913451194763184,
+ "step": 407
+ },
+ {
+ "epoch": 5.670742358078603,
+ "grad_norm": 0.03493810072541237,
+ "learning_rate": 0.0006,
+ "loss": 5.97633695602417,
+ "step": 408
+ },
+ {
+ "epoch": 5.68471615720524,
+ "grad_norm": 0.023131275549530983,
+ "learning_rate": 0.0006,
+ "loss": 5.899571418762207,
+ "step": 409
+ },
+ {
+ "epoch": 5.698689956331878,
+ "grad_norm": 0.02986014075577259,
+ "learning_rate": 0.0006,
+ "loss": 5.894903182983398,
+ "step": 410
+ },
+ {
+ "epoch": 5.712663755458515,
+ "grad_norm": 0.030171144753694534,
+ "learning_rate": 0.0006,
+ "loss": 5.876595497131348,
+ "step": 411
+ },
+ {
+ "epoch": 5.726637554585153,
+ "grad_norm": 0.029377546161413193,
+ "learning_rate": 0.0006,
+ "loss": 5.911637306213379,
+ "step": 412
+ },
+ {
+ "epoch": 5.74061135371179,
+ "grad_norm": 0.029829490929841995,
+ "learning_rate": 0.0006,
+ "loss": 5.819629669189453,
+ "step": 413
+ },
+ {
+ "epoch": 5.754585152838428,
+ "grad_norm": 0.024139299988746643,
+ "learning_rate": 0.0006,
+ "loss": 5.81728458404541,
+ "step": 414
+ },
+ {
+ "epoch": 5.7685589519650655,
+ "grad_norm": 0.021453432738780975,
+ "learning_rate": 0.0006,
+ "loss": 5.987326145172119,
+ "step": 415
+ },
+ {
+ "epoch": 5.782532751091703,
+ "grad_norm": 0.017942246049642563,
+ "learning_rate": 0.0006,
+ "loss": 5.893008232116699,
+ "step": 416
+ },
+ {
+ "epoch": 5.796506550218341,
+ "grad_norm": 0.019724637269973755,
+ "learning_rate": 0.0006,
+ "loss": 5.912441253662109,
+ "step": 417
+ },
+ {
+ "epoch": 5.810480349344978,
+ "grad_norm": 0.018597912043333054,
+ "learning_rate": 0.0006,
+ "loss": 5.860394477844238,
+ "step": 418
+ },
+ {
+ "epoch": 5.824454148471616,
+ "grad_norm": 0.016232412308454514,
+ "learning_rate": 0.0006,
+ "loss": 5.94991397857666,
+ "step": 419
+ },
+ {
+ "epoch": 5.8384279475982535,
+ "grad_norm": 0.01667204685509205,
+ "learning_rate": 0.0006,
+ "loss": 5.8506364822387695,
+ "step": 420
+ },
+ {
+ "epoch": 5.85240174672489,
+ "grad_norm": 0.014232151210308075,
+ "learning_rate": 0.0006,
+ "loss": 5.928424835205078,
+ "step": 421
+ },
+ {
+ "epoch": 5.866375545851528,
+ "grad_norm": 0.01570476032793522,
+ "learning_rate": 0.0006,
+ "loss": 5.953692436218262,
+ "step": 422
+ },
+ {
+ "epoch": 5.880349344978166,
+ "grad_norm": 0.01586179807782173,
+ "learning_rate": 0.0006,
+ "loss": 5.820014476776123,
+ "step": 423
+ },
+ {
+ "epoch": 5.894323144104804,
+ "grad_norm": 0.014729145914316177,
+ "learning_rate": 0.0006,
+ "loss": 5.860154628753662,
+ "step": 424
+ },
+ {
+ "epoch": 5.908296943231441,
+ "grad_norm": 0.013111459091305733,
+ "learning_rate": 0.0006,
+ "loss": 5.85988187789917,
+ "step": 425
+ },
+ {
+ "epoch": 5.922270742358078,
+ "grad_norm": 0.012937922962009907,
+ "learning_rate": 0.0006,
+ "loss": 5.756265640258789,
+ "step": 426
+ },
+ {
+ "epoch": 5.936244541484716,
+ "grad_norm": 0.01377453189343214,
+ "learning_rate": 0.0006,
+ "loss": 5.8333048820495605,
+ "step": 427
+ },
+ {
+ "epoch": 5.950218340611354,
+ "grad_norm": 0.014054795727133751,
+ "learning_rate": 0.0006,
+ "loss": 5.803333759307861,
+ "step": 428
+ },
+ {
+ "epoch": 5.964192139737992,
+ "grad_norm": 0.01674959622323513,
+ "learning_rate": 0.0006,
+ "loss": 5.8718485832214355,
+ "step": 429
+ },
+ {
+ "epoch": 5.978165938864628,
+ "grad_norm": 0.018733017146587372,
+ "learning_rate": 0.0006,
+ "loss": 5.875979900360107,
+ "step": 430
+ },
+ {
+ "epoch": 5.992139737991266,
+ "grad_norm": 0.02088983915746212,
+ "learning_rate": 0.0006,
+ "loss": 5.763634204864502,
+ "step": 431
+ },
+ {
+ "epoch": 6.0,
+ "grad_norm": 0.026705985888838768,
+ "learning_rate": 0.0006,
+ "loss": 5.6922478675842285,
+ "step": 432
+ },
+ {
+ "epoch": 6.0,
+ "eval_loss": 5.847958087921143,
+ "eval_runtime": 58.5451,
+ "eval_samples_per_second": 41.711,
+ "eval_steps_per_second": 1.315,
+ "step": 432
+ },
+ {
+ "epoch": 6.013973799126638,
+ "grad_norm": 0.03181544691324234,
+ "learning_rate": 0.0006,
+ "loss": 5.786368370056152,
+ "step": 433
+ },
+ {
+ "epoch": 6.0279475982532755,
+ "grad_norm": 0.03238112851977348,
+ "learning_rate": 0.0006,
+ "loss": 5.810310363769531,
+ "step": 434
+ },
+ {
+ "epoch": 6.041921397379912,
+ "grad_norm": 0.03916756808757782,
+ "learning_rate": 0.0006,
+ "loss": 5.7761993408203125,
+ "step": 435
+ },
+ {
+ "epoch": 6.05589519650655,
+ "grad_norm": 0.044009730219841,
+ "learning_rate": 0.0006,
+ "loss": 5.898112773895264,
+ "step": 436
+ },
+ {
+ "epoch": 6.069868995633188,
+ "grad_norm": 0.045235246419906616,
+ "learning_rate": 0.0006,
+ "loss": 5.876371383666992,
+ "step": 437
+ },
+ {
+ "epoch": 6.083842794759826,
+ "grad_norm": 0.050218384712934494,
+ "learning_rate": 0.0006,
+ "loss": 5.805103302001953,
+ "step": 438
+ },
+ {
+ "epoch": 6.097816593886463,
+ "grad_norm": 0.05444490164518356,
+ "learning_rate": 0.0006,
+ "loss": 5.826424598693848,
+ "step": 439
+ },
+ {
+ "epoch": 6.1117903930131,
+ "grad_norm": 0.04890982061624527,
+ "learning_rate": 0.0006,
+ "loss": 5.7735395431518555,
+ "step": 440
+ },
+ {
+ "epoch": 6.125764192139738,
+ "grad_norm": 0.05478639155626297,
+ "learning_rate": 0.0006,
+ "loss": 5.823677062988281,
+ "step": 441
+ },
+ {
+ "epoch": 6.139737991266376,
+ "grad_norm": 0.057562313973903656,
+ "learning_rate": 0.0006,
+ "loss": 5.889334678649902,
+ "step": 442
+ },
+ {
+ "epoch": 6.153711790393013,
+ "grad_norm": 0.06447703391313553,
+ "learning_rate": 0.0006,
+ "loss": 5.84621524810791,
+ "step": 443
+ },
+ {
+ "epoch": 6.1676855895196505,
+ "grad_norm": 0.048861872404813766,
+ "learning_rate": 0.0006,
+ "loss": 5.931595802307129,
+ "step": 444
+ },
+ {
+ "epoch": 6.181659388646288,
+ "grad_norm": 0.05521339178085327,
+ "learning_rate": 0.0006,
+ "loss": 5.8229780197143555,
+ "step": 445
+ },
+ {
+ "epoch": 6.195633187772926,
+ "grad_norm": 0.053666990250349045,
+ "learning_rate": 0.0006,
+ "loss": 5.879191875457764,
+ "step": 446
+ },
+ {
+ "epoch": 6.209606986899563,
+ "grad_norm": 0.0451025515794754,
+ "learning_rate": 0.0006,
+ "loss": 5.801628112792969,
+ "step": 447
+ },
+ {
+ "epoch": 6.223580786026201,
+ "grad_norm": 0.04291655868291855,
+ "learning_rate": 0.0006,
+ "loss": 5.914680480957031,
+ "step": 448
+ },
+ {
+ "epoch": 6.2375545851528384,
+ "grad_norm": 0.035102490335702896,
+ "learning_rate": 0.0006,
+ "loss": 5.843682765960693,
+ "step": 449
+ },
+ {
+ "epoch": 6.251528384279476,
+ "grad_norm": 0.03403995931148529,
+ "learning_rate": 0.0006,
+ "loss": 5.827154159545898,
+ "step": 450
+ },
+ {
+ "epoch": 6.265502183406113,
+ "grad_norm": 0.03444375470280647,
+ "learning_rate": 0.0006,
+ "loss": 5.800136566162109,
+ "step": 451
+ },
+ {
+ "epoch": 6.279475982532751,
+ "grad_norm": 0.03165159001946449,
+ "learning_rate": 0.0006,
+ "loss": 5.906252384185791,
+ "step": 452
+ },
+ {
+ "epoch": 6.293449781659389,
+ "grad_norm": 0.026153093203902245,
+ "learning_rate": 0.0006,
+ "loss": 5.7423200607299805,
+ "step": 453
+ },
+ {
+ "epoch": 6.307423580786026,
+ "grad_norm": 0.024357657879590988,
+ "learning_rate": 0.0006,
+ "loss": 5.701364517211914,
+ "step": 454
+ },
+ {
+ "epoch": 6.321397379912664,
+ "grad_norm": 0.021508049219846725,
+ "learning_rate": 0.0006,
+ "loss": 5.8328094482421875,
+ "step": 455
+ },
+ {
+ "epoch": 6.335371179039301,
+ "grad_norm": 0.017313921824097633,
+ "learning_rate": 0.0006,
+ "loss": 5.792145252227783,
+ "step": 456
+ },
+ {
+ "epoch": 6.349344978165939,
+ "grad_norm": 0.018563397228717804,
+ "learning_rate": 0.0006,
+ "loss": 5.732672214508057,
+ "step": 457
+ },
+ {
+ "epoch": 6.3633187772925766,
+ "grad_norm": 0.016568679362535477,
+ "learning_rate": 0.0006,
+ "loss": 5.81948184967041,
+ "step": 458
+ },
+ {
+ "epoch": 6.377292576419214,
+ "grad_norm": 0.014133770950138569,
+ "learning_rate": 0.0006,
+ "loss": 5.734982490539551,
+ "step": 459
+ },
+ {
+ "epoch": 6.391266375545851,
+ "grad_norm": 0.014114447869360447,
+ "learning_rate": 0.0006,
+ "loss": 5.70073127746582,
+ "step": 460
+ },
+ {
+ "epoch": 6.405240174672489,
+ "grad_norm": 0.012907393276691437,
+ "learning_rate": 0.0006,
+ "loss": 5.758626937866211,
+ "step": 461
+ },
+ {
+ "epoch": 6.419213973799127,
+ "grad_norm": 0.012673444114625454,
+ "learning_rate": 0.0006,
+ "loss": 5.775138854980469,
+ "step": 462
+ },
+ {
+ "epoch": 6.4331877729257645,
+ "grad_norm": 0.01311410591006279,
+ "learning_rate": 0.0006,
+ "loss": 5.70920991897583,
+ "step": 463
+ },
+ {
+ "epoch": 6.447161572052401,
+ "grad_norm": 0.012935544364154339,
+ "learning_rate": 0.0006,
+ "loss": 5.781479835510254,
+ "step": 464
+ },
+ {
+ "epoch": 6.461135371179039,
+ "grad_norm": 0.01439738366752863,
+ "learning_rate": 0.0006,
+ "loss": 5.691315650939941,
+ "step": 465
+ },
+ {
+ "epoch": 6.475109170305677,
+ "grad_norm": 0.01205186452716589,
+ "learning_rate": 0.0006,
+ "loss": 5.662499904632568,
+ "step": 466
+ },
+ {
+ "epoch": 6.489082969432315,
+ "grad_norm": 0.011575652286410332,
+ "learning_rate": 0.0006,
+ "loss": 5.6271772384643555,
+ "step": 467
+ },
+ {
+ "epoch": 6.503056768558952,
+ "grad_norm": 0.011672502383589745,
+ "learning_rate": 0.0006,
+ "loss": 5.761662006378174,
+ "step": 468
+ },
+ {
+ "epoch": 6.517030567685589,
+ "grad_norm": 0.011541751213371754,
+ "learning_rate": 0.0006,
+ "loss": 5.762078285217285,
+ "step": 469
+ },
+ {
+ "epoch": 6.531004366812227,
+ "grad_norm": 0.011396365240216255,
+ "learning_rate": 0.0006,
+ "loss": 5.67873477935791,
+ "step": 470
+ },
+ {
+ "epoch": 6.544978165938865,
+ "grad_norm": 0.010688499547541142,
+ "learning_rate": 0.0006,
+ "loss": 5.717051982879639,
+ "step": 471
+ },
+ {
+ "epoch": 6.558951965065502,
+ "grad_norm": 0.012224317528307438,
+ "learning_rate": 0.0006,
+ "loss": 5.707948684692383,
+ "step": 472
+ },
+ {
+ "epoch": 6.5729257641921395,
+ "grad_norm": 0.011856825090944767,
+ "learning_rate": 0.0006,
+ "loss": 5.70878791809082,
+ "step": 473
+ },
+ {
+ "epoch": 6.586899563318777,
+ "grad_norm": 0.01199064590036869,
+ "learning_rate": 0.0006,
+ "loss": 5.708697319030762,
+ "step": 474
+ },
+ {
+ "epoch": 6.600873362445415,
+ "grad_norm": 0.01219446212053299,
+ "learning_rate": 0.0006,
+ "loss": 5.6061577796936035,
+ "step": 475
+ },
+ {
+ "epoch": 6.614847161572053,
+ "grad_norm": 0.013397484086453915,
+ "learning_rate": 0.0006,
+ "loss": 5.624789714813232,
+ "step": 476
+ },
+ {
+ "epoch": 6.62882096069869,
+ "grad_norm": 0.01483136136084795,
+ "learning_rate": 0.0006,
+ "loss": 5.721141338348389,
+ "step": 477
+ },
+ {
+ "epoch": 6.642794759825327,
+ "grad_norm": 0.0194488987326622,
+ "learning_rate": 0.0006,
+ "loss": 5.604279518127441,
+ "step": 478
+ },
+ {
+ "epoch": 6.656768558951965,
+ "grad_norm": 0.02079896256327629,
+ "learning_rate": 0.0006,
+ "loss": 5.586322784423828,
+ "step": 479
+ },
+ {
+ "epoch": 6.670742358078603,
+ "grad_norm": 0.018996229395270348,
+ "learning_rate": 0.0006,
+ "loss": 5.5987772941589355,
+ "step": 480
+ },
+ {
+ "epoch": 6.68471615720524,
+ "grad_norm": 0.015779603272676468,
+ "learning_rate": 0.0006,
+ "loss": 5.588602066040039,
+ "step": 481
+ },
+ {
+ "epoch": 6.698689956331878,
+ "grad_norm": 0.015322973020374775,
+ "learning_rate": 0.0006,
+ "loss": 5.686467170715332,
+ "step": 482
+ },
+ {
+ "epoch": 6.712663755458515,
+ "grad_norm": 0.015282213687896729,
+ "learning_rate": 0.0006,
+ "loss": 5.664676666259766,
+ "step": 483
+ },
+ {
+ "epoch": 6.726637554585153,
+ "grad_norm": 0.016589272767305374,
+ "learning_rate": 0.0006,
+ "loss": 5.666739463806152,
+ "step": 484
+ },
+ {
+ "epoch": 6.74061135371179,
+ "grad_norm": 0.01811421848833561,
+ "learning_rate": 0.0006,
+ "loss": 5.647593021392822,
+ "step": 485
+ },
+ {
+ "epoch": 6.754585152838428,
+ "grad_norm": 0.019652029499411583,
+ "learning_rate": 0.0006,
+ "loss": 5.628726959228516,
+ "step": 486
+ },
+ {
+ "epoch": 6.7685589519650655,
+ "grad_norm": 0.02118665538728237,
+ "learning_rate": 0.0006,
+ "loss": 5.6580705642700195,
+ "step": 487
+ },
+ {
+ "epoch": 6.782532751091703,
+ "grad_norm": 0.02237832546234131,
+ "learning_rate": 0.0006,
+ "loss": 5.635931968688965,
+ "step": 488
+ },
+ {
+ "epoch": 6.796506550218341,
+ "grad_norm": 0.023897631093859673,
+ "learning_rate": 0.0006,
+ "loss": 5.517949104309082,
+ "step": 489
+ },
+ {
+ "epoch": 6.810480349344978,
+ "grad_norm": 0.02442227490246296,
+ "learning_rate": 0.0006,
+ "loss": 5.600021839141846,
+ "step": 490
+ },
+ {
+ "epoch": 6.824454148471616,
+ "grad_norm": 0.024675287306308746,
+ "learning_rate": 0.0006,
+ "loss": 5.6639909744262695,
+ "step": 491
+ },
+ {
+ "epoch": 6.8384279475982535,
+ "grad_norm": 0.030372392386198044,
+ "learning_rate": 0.0006,
+ "loss": 5.582888603210449,
+ "step": 492
+ },
+ {
+ "epoch": 6.85240174672489,
+ "grad_norm": 0.03237690031528473,
+ "learning_rate": 0.0006,
+ "loss": 5.573896884918213,
+ "step": 493
+ },
+ {
+ "epoch": 6.866375545851528,
+ "grad_norm": 0.03511589393019676,
+ "learning_rate": 0.0006,
+ "loss": 5.565878868103027,
+ "step": 494
+ },
+ {
+ "epoch": 6.880349344978166,
+ "grad_norm": 0.03883938118815422,
+ "learning_rate": 0.0006,
+ "loss": 5.58540678024292,
+ "step": 495
+ },
+ {
+ "epoch": 6.894323144104804,
+ "grad_norm": 0.04175502806901932,
+ "learning_rate": 0.0006,
+ "loss": 5.6095194816589355,
+ "step": 496
+ },
+ {
+ "epoch": 6.908296943231441,
+ "grad_norm": 0.04013441503047943,
+ "learning_rate": 0.0006,
+ "loss": 5.654401779174805,
+ "step": 497
+ },
+ {
+ "epoch": 6.922270742358078,
+ "grad_norm": 0.04618770629167557,
+ "learning_rate": 0.0006,
+ "loss": 5.6483473777771,
+ "step": 498
+ },
+ {
+ "epoch": 6.936244541484716,
+ "grad_norm": 0.04507308453321457,
+ "learning_rate": 0.0006,
+ "loss": 5.5856218338012695,
+ "step": 499
+ },
+ {
+ "epoch": 6.950218340611354,
+ "grad_norm": 0.03151383996009827,
+ "learning_rate": 0.0006,
+ "loss": 5.594086647033691,
+ "step": 500
+ },
+ {
+ "epoch": 6.964192139737992,
+ "grad_norm": 0.026883797720074654,
+ "learning_rate": 0.0006,
+ "loss": 5.607676029205322,
+ "step": 501
+ },
+ {
+ "epoch": 6.978165938864628,
+ "grad_norm": 0.02981836162507534,
+ "learning_rate": 0.0006,
+ "loss": 5.563666820526123,
+ "step": 502
+ },
+ {
+ "epoch": 6.992139737991266,
+ "grad_norm": 0.03233477845788002,
+ "learning_rate": 0.0006,
+ "loss": 5.667543411254883,
+ "step": 503
+ },
+ {
+ "epoch": 7.0,
+ "grad_norm": 0.029927456751465797,
+ "learning_rate": 0.0006,
+ "loss": 5.733482837677002,
+ "step": 504
+ },
+ {
+ "epoch": 7.0,
+ "eval_loss": 5.679076671600342,
+ "eval_runtime": 59.5965,
+ "eval_samples_per_second": 40.976,
+ "eval_steps_per_second": 1.292,
+ "step": 504
+ },
+ {
+ "epoch": 7.013973799126638,
+ "grad_norm": 0.027081597596406937,
+ "learning_rate": 0.0006,
+ "loss": 5.634487152099609,
+ "step": 505
+ },
+ {
+ "epoch": 7.0279475982532755,
+ "grad_norm": 0.031959421932697296,
+ "learning_rate": 0.0006,
+ "loss": 5.643294811248779,
+ "step": 506
+ },
+ {
+ "epoch": 7.041921397379912,
+ "grad_norm": 0.02802063524723053,
+ "learning_rate": 0.0006,
+ "loss": 5.573239326477051,
+ "step": 507
+ },
+ {
+ "epoch": 7.05589519650655,
+ "grad_norm": 0.032579705119132996,
+ "learning_rate": 0.0006,
+ "loss": 5.627150058746338,
+ "step": 508
+ },
+ {
+ "epoch": 7.069868995633188,
+ "grad_norm": 0.02503928542137146,
+ "learning_rate": 0.0006,
+ "loss": 5.6045308113098145,
+ "step": 509
+ },
+ {
+ "epoch": 7.083842794759826,
+ "grad_norm": 0.02421317622065544,
+ "learning_rate": 0.0006,
+ "loss": 5.522153377532959,
+ "step": 510
+ },
+ {
+ "epoch": 7.097816593886463,
+ "grad_norm": 0.021983426064252853,
+ "learning_rate": 0.0006,
+ "loss": 5.632939338684082,
+ "step": 511
+ },
+ {
+ "epoch": 7.1117903930131,
+ "grad_norm": 0.021933183073997498,
+ "learning_rate": 0.0006,
+ "loss": 5.592185974121094,
+ "step": 512
+ },
+ {
+ "epoch": 7.125764192139738,
+ "grad_norm": 0.02287031151354313,
+ "learning_rate": 0.0006,
+ "loss": 5.613700866699219,
+ "step": 513
+ },
+ {
+ "epoch": 7.139737991266376,
+ "grad_norm": 0.02281602844595909,
+ "learning_rate": 0.0006,
+ "loss": 5.551383018493652,
+ "step": 514
+ },
+ {
+ "epoch": 7.153711790393013,
+ "grad_norm": 0.021050360053777695,
+ "learning_rate": 0.0006,
+ "loss": 5.55275821685791,
+ "step": 515
+ },
+ {
+ "epoch": 7.1676855895196505,
+ "grad_norm": 0.018299926072359085,
+ "learning_rate": 0.0006,
+ "loss": 5.528225421905518,
+ "step": 516
+ },
+ {
+ "epoch": 7.181659388646288,
+ "grad_norm": 0.02024853602051735,
+ "learning_rate": 0.0006,
+ "loss": 5.522453784942627,
+ "step": 517
+ },
+ {
+ "epoch": 7.195633187772926,
+ "grad_norm": 0.02135239914059639,
+ "learning_rate": 0.0006,
+ "loss": 5.527522563934326,
+ "step": 518
+ },
+ {
+ "epoch": 7.209606986899563,
+ "grad_norm": 0.023394184187054634,
+ "learning_rate": 0.0006,
+ "loss": 5.4535017013549805,
+ "step": 519
+ },
+ {
+ "epoch": 7.223580786026201,
+ "grad_norm": 0.025952080264687538,
+ "learning_rate": 0.0006,
+ "loss": 5.5374674797058105,
+ "step": 520
+ },
+ {
+ "epoch": 7.2375545851528384,
+ "grad_norm": 0.022287718951702118,
+ "learning_rate": 0.0006,
+ "loss": 5.493753433227539,
+ "step": 521
+ },
+ {
+ "epoch": 7.251528384279476,
+ "grad_norm": 0.02018447406589985,
+ "learning_rate": 0.0006,
+ "loss": 5.510575294494629,
+ "step": 522
+ },
+ {
+ "epoch": 7.265502183406113,
+ "grad_norm": 0.02050507813692093,
+ "learning_rate": 0.0006,
+ "loss": 5.466026782989502,
+ "step": 523
+ },
+ {
+ "epoch": 7.279475982532751,
+ "grad_norm": 0.023688802495598793,
+ "learning_rate": 0.0006,
+ "loss": 5.485815525054932,
+ "step": 524
+ },
+ {
+ "epoch": 7.293449781659389,
+ "grad_norm": 0.021078843623399734,
+ "learning_rate": 0.0006,
+ "loss": 5.415581703186035,
+ "step": 525
+ },
+ {
+ "epoch": 7.307423580786026,
+ "grad_norm": 0.018266011029481888,
+ "learning_rate": 0.0006,
+ "loss": 5.402815818786621,
+ "step": 526
+ },
+ {
+ "epoch": 7.321397379912664,
+ "grad_norm": 0.019693493843078613,
+ "learning_rate": 0.0006,
+ "loss": 5.504674911499023,
+ "step": 527
+ },
+ {
+ "epoch": 7.335371179039301,
+ "grad_norm": 0.02031373605132103,
+ "learning_rate": 0.0006,
+ "loss": 5.503837585449219,
+ "step": 528
+ },
+ {
+ "epoch": 7.349344978165939,
+ "grad_norm": 0.017636626958847046,
+ "learning_rate": 0.0006,
+ "loss": 5.530580997467041,
+ "step": 529
+ },
+ {
+ "epoch": 7.3633187772925766,
+ "grad_norm": 0.01787244901061058,
+ "learning_rate": 0.0006,
+ "loss": 5.450218677520752,
+ "step": 530
+ },
+ {
+ "epoch": 7.377292576419214,
+ "grad_norm": 0.0170669574290514,
+ "learning_rate": 0.0006,
+ "loss": 5.481570243835449,
+ "step": 531
+ },
+ {
+ "epoch": 7.391266375545851,
+ "grad_norm": 0.01802165061235428,
+ "learning_rate": 0.0006,
+ "loss": 5.386394500732422,
+ "step": 532
+ },
+ {
+ "epoch": 7.405240174672489,
+ "grad_norm": 0.022949472069740295,
+ "learning_rate": 0.0006,
+ "loss": 5.424929618835449,
+ "step": 533
+ },
+ {
+ "epoch": 7.419213973799127,
+ "grad_norm": 0.034213390201330185,
+ "learning_rate": 0.0006,
+ "loss": 5.346663475036621,
+ "step": 534
+ },
+ {
+ "epoch": 7.4331877729257645,
+ "grad_norm": 0.03847593814134598,
+ "learning_rate": 0.0006,
+ "loss": 5.418482780456543,
+ "step": 535
+ },
+ {
+ "epoch": 7.447161572052401,
+ "grad_norm": 0.029393676668405533,
+ "learning_rate": 0.0006,
+ "loss": 5.456090927124023,
+ "step": 536
+ },
+ {
+ "epoch": 7.461135371179039,
+ "grad_norm": 0.03388667106628418,
+ "learning_rate": 0.0006,
+ "loss": 5.572983741760254,
+ "step": 537
+ },
+ {
+ "epoch": 7.475109170305677,
+ "grad_norm": 0.039690613746643066,
+ "learning_rate": 0.0006,
+ "loss": 5.414067268371582,
+ "step": 538
+ },
+ {
+ "epoch": 7.489082969432315,
+ "grad_norm": 0.03630776330828667,
+ "learning_rate": 0.0006,
+ "loss": 5.523739814758301,
+ "step": 539
+ },
+ {
+ "epoch": 7.503056768558952,
+ "grad_norm": 0.03356870263814926,
+ "learning_rate": 0.0006,
+ "loss": 5.444767951965332,
+ "step": 540
+ },
+ {
+ "epoch": 7.517030567685589,
+ "grad_norm": 0.030812304466962814,
+ "learning_rate": 0.0006,
+ "loss": 5.461244106292725,
+ "step": 541
+ },
+ {
+ "epoch": 7.531004366812227,
+ "grad_norm": 0.03319217637181282,
+ "learning_rate": 0.0006,
+ "loss": 5.502161026000977,
+ "step": 542
+ },
+ {
+ "epoch": 7.544978165938865,
+ "grad_norm": 0.032997481524944305,
+ "learning_rate": 0.0006,
+ "loss": 5.4162278175354,
+ "step": 543
+ },
+ {
+ "epoch": 7.558951965065502,
+ "grad_norm": 0.03364962339401245,
+ "learning_rate": 0.0006,
+ "loss": 5.415736198425293,
+ "step": 544
+ },
+ {
+ "epoch": 7.5729257641921395,
+ "grad_norm": 0.0344221405684948,
+ "learning_rate": 0.0006,
+ "loss": 5.541967868804932,
+ "step": 545
+ },
+ {
+ "epoch": 7.586899563318777,
+ "grad_norm": 0.029609503224492073,
+ "learning_rate": 0.0006,
+ "loss": 5.440614223480225,
+ "step": 546
+ },
+ {
+ "epoch": 7.600873362445415,
+ "grad_norm": 0.029055926948785782,
+ "learning_rate": 0.0006,
+ "loss": 5.462865352630615,
+ "step": 547
+ },
+ {
+ "epoch": 7.614847161572053,
+ "grad_norm": 0.02658848837018013,
+ "learning_rate": 0.0006,
+ "loss": 5.444467544555664,
+ "step": 548
+ },
+ {
+ "epoch": 7.62882096069869,
+ "grad_norm": 0.026276404038071632,
+ "learning_rate": 0.0006,
+ "loss": 5.462278842926025,
+ "step": 549
+ },
+ {
+ "epoch": 7.642794759825327,
+ "grad_norm": 0.0286889486014843,
+ "learning_rate": 0.0006,
+ "loss": 5.4849534034729,
+ "step": 550
+ },
+ {
+ "epoch": 7.656768558951965,
+ "grad_norm": 0.026849817484617233,
+ "learning_rate": 0.0006,
+ "loss": 5.431473731994629,
+ "step": 551
+ },
+ {
+ "epoch": 7.670742358078603,
+ "grad_norm": 0.02312396466732025,
+ "learning_rate": 0.0006,
+ "loss": 5.45462703704834,
+ "step": 552
+ },
+ {
+ "epoch": 7.68471615720524,
+ "grad_norm": 0.026197485625743866,
+ "learning_rate": 0.0006,
+ "loss": 5.452552318572998,
+ "step": 553
+ },
+ {
+ "epoch": 7.698689956331878,
+ "grad_norm": 0.02747255191206932,
+ "learning_rate": 0.0006,
+ "loss": 5.447169780731201,
+ "step": 554
+ },
+ {
+ "epoch": 7.712663755458515,
+ "grad_norm": 0.028123432770371437,
+ "learning_rate": 0.0006,
+ "loss": 5.394400596618652,
+ "step": 555
+ },
+ {
+ "epoch": 7.726637554585153,
+ "grad_norm": 0.02599870041012764,
+ "learning_rate": 0.0006,
+ "loss": 5.398341178894043,
+ "step": 556
+ },
+ {
+ "epoch": 7.74061135371179,
+ "grad_norm": 0.022171657532453537,
+ "learning_rate": 0.0006,
+ "loss": 5.368820667266846,
+ "step": 557
+ },
+ {
+ "epoch": 7.754585152838428,
+ "grad_norm": 0.022309480234980583,
+ "learning_rate": 0.0006,
+ "loss": 5.416772365570068,
+ "step": 558
+ },
+ {
+ "epoch": 7.7685589519650655,
+ "grad_norm": 0.024074165150523186,
+ "learning_rate": 0.0006,
+ "loss": 5.359002113342285,
+ "step": 559
+ },
+ {
+ "epoch": 7.782532751091703,
+ "grad_norm": 0.02636653557419777,
+ "learning_rate": 0.0006,
+ "loss": 5.484368324279785,
+ "step": 560
+ },
+ {
+ "epoch": 7.796506550218341,
+ "grad_norm": 0.02203752100467682,
+ "learning_rate": 0.0006,
+ "loss": 5.375498294830322,
+ "step": 561
+ },
+ {
+ "epoch": 7.810480349344978,
+ "grad_norm": 0.019924819469451904,
+ "learning_rate": 0.0006,
+ "loss": 5.334723472595215,
+ "step": 562
+ },
+ {
+ "epoch": 7.824454148471616,
+ "grad_norm": 0.018755966797471046,
+ "learning_rate": 0.0006,
+ "loss": 5.348781108856201,
+ "step": 563
+ },
+ {
+ "epoch": 7.8384279475982535,
+ "grad_norm": 0.016747845336794853,
+ "learning_rate": 0.0006,
+ "loss": 5.433079719543457,
+ "step": 564
+ },
+ {
+ "epoch": 7.85240174672489,
+ "grad_norm": 0.0170395877212286,
+ "learning_rate": 0.0006,
+ "loss": 5.346794128417969,
+ "step": 565
+ },
+ {
+ "epoch": 7.866375545851528,
+ "grad_norm": 0.016074176877737045,
+ "learning_rate": 0.0006,
+ "loss": 5.3277740478515625,
+ "step": 566
+ },
+ {
+ "epoch": 7.880349344978166,
+ "grad_norm": 0.014119806699454784,
+ "learning_rate": 0.0006,
+ "loss": 5.483551979064941,
+ "step": 567
+ },
+ {
+ "epoch": 7.894323144104804,
+ "grad_norm": 0.015270394273102283,
+ "learning_rate": 0.0006,
+ "loss": 5.378519535064697,
+ "step": 568
+ },
+ {
+ "epoch": 7.908296943231441,
+ "grad_norm": 0.015677539631724358,
+ "learning_rate": 0.0006,
+ "loss": 5.290210247039795,
+ "step": 569
+ },
+ {
+ "epoch": 7.922270742358078,
+ "grad_norm": 0.015930423513054848,
+ "learning_rate": 0.0006,
+ "loss": 5.427360534667969,
+ "step": 570
+ },
+ {
+ "epoch": 7.936244541484716,
+ "grad_norm": 0.016641564667224884,
+ "learning_rate": 0.0006,
+ "loss": 5.301599979400635,
+ "step": 571
+ },
+ {
+ "epoch": 7.950218340611354,
+ "grad_norm": 0.018293552100658417,
+ "learning_rate": 0.0006,
+ "loss": 5.277889251708984,
+ "step": 572
+ },
+ {
+ "epoch": 7.964192139737992,
+ "grad_norm": 0.017618799582123756,
+ "learning_rate": 0.0006,
+ "loss": 5.297194480895996,
+ "step": 573
+ },
+ {
+ "epoch": 7.978165938864628,
+ "grad_norm": 0.015034242533147335,
+ "learning_rate": 0.0006,
+ "loss": 5.42428731918335,
+ "step": 574
+ },
+ {
+ "epoch": 7.992139737991266,
+ "grad_norm": 0.01617511734366417,
+ "learning_rate": 0.0006,
+ "loss": 5.420044898986816,
+ "step": 575
+ },
+ {
+ "epoch": 8.0,
+ "grad_norm": 0.01837257295846939,
+ "learning_rate": 0.0006,
+ "loss": 5.240387916564941,
+ "step": 576
+ },
+ {
+ "epoch": 8.0,
+ "eval_loss": 5.423932075500488,
+ "eval_runtime": 58.0841,
+ "eval_samples_per_second": 42.042,
+ "eval_steps_per_second": 1.326,
+ "step": 576
+ },
+ {
+ "epoch": 8.013973799126637,
+ "grad_norm": 0.020953809842467308,
+ "learning_rate": 0.0006,
+ "loss": 5.4320783615112305,
+ "step": 577
+ },
+ {
+ "epoch": 8.027947598253276,
+ "grad_norm": 0.0227745920419693,
+ "learning_rate": 0.0006,
+ "loss": 5.310197830200195,
+ "step": 578
+ },
+ {
+ "epoch": 8.041921397379912,
+ "grad_norm": 0.021372603252530098,
+ "learning_rate": 0.0006,
+ "loss": 5.339468002319336,
+ "step": 579
+ },
+ {
+ "epoch": 8.055895196506551,
+ "grad_norm": 0.022401731461286545,
+ "learning_rate": 0.0006,
+ "loss": 5.279829978942871,
+ "step": 580
+ },
+ {
+ "epoch": 8.069868995633188,
+ "grad_norm": 0.023942379280924797,
+ "learning_rate": 0.0006,
+ "loss": 5.372323989868164,
+ "step": 581
+ },
+ {
+ "epoch": 8.083842794759825,
+ "grad_norm": 0.024380534887313843,
+ "learning_rate": 0.0006,
+ "loss": 5.253486633300781,
+ "step": 582
+ },
+ {
+ "epoch": 8.097816593886463,
+ "grad_norm": 0.03246625140309334,
+ "learning_rate": 0.0006,
+ "loss": 5.305774211883545,
+ "step": 583
+ },
+ {
+ "epoch": 8.1117903930131,
+ "grad_norm": 0.030399736016988754,
+ "learning_rate": 0.0006,
+ "loss": 5.370976448059082,
+ "step": 584
+ },
+ {
+ "epoch": 8.125764192139737,
+ "grad_norm": 0.027148913592100143,
+ "learning_rate": 0.0006,
+ "loss": 5.368010520935059,
+ "step": 585
+ },
+ {
+ "epoch": 8.139737991266376,
+ "grad_norm": 0.02925540879368782,
+ "learning_rate": 0.0006,
+ "loss": 5.303013801574707,
+ "step": 586
+ },
+ {
+ "epoch": 8.153711790393013,
+ "grad_norm": 0.02915453538298607,
+ "learning_rate": 0.0006,
+ "loss": 5.329667091369629,
+ "step": 587
+ },
+ {
+ "epoch": 8.167685589519651,
+ "grad_norm": 0.031732626259326935,
+ "learning_rate": 0.0006,
+ "loss": 5.3634934425354,
+ "step": 588
+ },
+ {
+ "epoch": 8.181659388646288,
+ "grad_norm": 0.036166731268167496,
+ "learning_rate": 0.0006,
+ "loss": 5.223980903625488,
+ "step": 589
+ },
+ {
+ "epoch": 8.195633187772925,
+ "grad_norm": 0.035654108971357346,
+ "learning_rate": 0.0006,
+ "loss": 5.361059188842773,
+ "step": 590
+ },
+ {
+ "epoch": 8.209606986899564,
+ "grad_norm": 0.0365324430167675,
+ "learning_rate": 0.0006,
+ "loss": 5.336151599884033,
+ "step": 591
+ },
+ {
+ "epoch": 8.2235807860262,
+ "grad_norm": 0.03578880429267883,
+ "learning_rate": 0.0006,
+ "loss": 5.434076309204102,
+ "step": 592
+ },
+ {
+ "epoch": 8.237554585152838,
+ "grad_norm": 0.035734012722969055,
+ "learning_rate": 0.0006,
+ "loss": 5.330893516540527,
+ "step": 593
+ },
+ {
+ "epoch": 8.251528384279476,
+ "grad_norm": 0.03196857124567032,
+ "learning_rate": 0.0006,
+ "loss": 5.306009769439697,
+ "step": 594
+ },
+ {
+ "epoch": 8.265502183406113,
+ "grad_norm": 0.032042086124420166,
+ "learning_rate": 0.0006,
+ "loss": 5.343267917633057,
+ "step": 595
+ },
+ {
+ "epoch": 8.279475982532752,
+ "grad_norm": 0.03160746395587921,
+ "learning_rate": 0.0006,
+ "loss": 5.2353715896606445,
+ "step": 596
+ },
+ {
+ "epoch": 8.293449781659389,
+ "grad_norm": 0.026689818128943443,
+ "learning_rate": 0.0006,
+ "loss": 5.254042625427246,
+ "step": 597
+ },
+ {
+ "epoch": 8.307423580786025,
+ "grad_norm": 0.02880188450217247,
+ "learning_rate": 0.0006,
+ "loss": 5.326833248138428,
+ "step": 598
+ },
+ {
+ "epoch": 8.321397379912664,
+ "grad_norm": 0.027516381815075874,
+ "learning_rate": 0.0006,
+ "loss": 5.342336177825928,
+ "step": 599
+ },
+ {
+ "epoch": 8.335371179039301,
+ "grad_norm": 0.027875030413269997,
+ "learning_rate": 0.0006,
+ "loss": 5.200719833374023,
+ "step": 600
+ },
+ {
+ "epoch": 8.34934497816594,
+ "grad_norm": 0.0268265288323164,
+ "learning_rate": 0.0006,
+ "loss": 5.249449729919434,
+ "step": 601
+ },
+ {
+ "epoch": 8.363318777292577,
+ "grad_norm": 0.024824608117341995,
+ "learning_rate": 0.0006,
+ "loss": 5.318952560424805,
+ "step": 602
+ },
+ {
+ "epoch": 8.377292576419213,
+ "grad_norm": 0.022990627214312553,
+ "learning_rate": 0.0006,
+ "loss": 5.244993209838867,
+ "step": 603
+ },
+ {
+ "epoch": 8.391266375545852,
+ "grad_norm": 0.022804604843258858,
+ "learning_rate": 0.0006,
+ "loss": 5.3236846923828125,
+ "step": 604
+ },
+ {
+ "epoch": 8.405240174672489,
+ "grad_norm": 0.02419872209429741,
+ "learning_rate": 0.0006,
+ "loss": 5.255486488342285,
+ "step": 605
+ },
+ {
+ "epoch": 8.419213973799126,
+ "grad_norm": 0.021952059119939804,
+ "learning_rate": 0.0006,
+ "loss": 5.3511552810668945,
+ "step": 606
+ },
+ {
+ "epoch": 8.433187772925764,
+ "grad_norm": 0.022375497967004776,
+ "learning_rate": 0.0006,
+ "loss": 5.294790267944336,
+ "step": 607
+ },
+ {
+ "epoch": 8.447161572052401,
+ "grad_norm": 0.019242815673351288,
+ "learning_rate": 0.0006,
+ "loss": 5.252618789672852,
+ "step": 608
+ },
+ {
+ "epoch": 8.46113537117904,
+ "grad_norm": 0.0173486340790987,
+ "learning_rate": 0.0006,
+ "loss": 5.30147647857666,
+ "step": 609
+ },
+ {
+ "epoch": 8.475109170305677,
+ "grad_norm": 0.01612604409456253,
+ "learning_rate": 0.0006,
+ "loss": 5.23960542678833,
+ "step": 610
+ },
+ {
+ "epoch": 8.489082969432314,
+ "grad_norm": 0.014293976128101349,
+ "learning_rate": 0.0006,
+ "loss": 5.119932651519775,
+ "step": 611
+ },
+ {
+ "epoch": 8.503056768558952,
+ "grad_norm": 0.014453536830842495,
+ "learning_rate": 0.0006,
+ "loss": 5.166794776916504,
+ "step": 612
+ },
+ {
+ "epoch": 8.51703056768559,
+ "grad_norm": 0.014685769565403461,
+ "learning_rate": 0.0006,
+ "loss": 5.264822959899902,
+ "step": 613
+ },
+ {
+ "epoch": 8.531004366812226,
+ "grad_norm": 0.01427740603685379,
+ "learning_rate": 0.0006,
+ "loss": 5.278433799743652,
+ "step": 614
+ },
+ {
+ "epoch": 8.544978165938865,
+ "grad_norm": 0.014864951372146606,
+ "learning_rate": 0.0006,
+ "loss": 5.214512348175049,
+ "step": 615
+ },
+ {
+ "epoch": 8.558951965065502,
+ "grad_norm": 0.01565164513885975,
+ "learning_rate": 0.0006,
+ "loss": 5.183066368103027,
+ "step": 616
+ },
+ {
+ "epoch": 8.57292576419214,
+ "grad_norm": 0.01779816672205925,
+ "learning_rate": 0.0006,
+ "loss": 5.163185119628906,
+ "step": 617
+ },
+ {
+ "epoch": 8.586899563318777,
+ "grad_norm": 0.017254827544093132,
+ "learning_rate": 0.0006,
+ "loss": 5.243555068969727,
+ "step": 618
+ },
+ {
+ "epoch": 8.600873362445414,
+ "grad_norm": 0.01650584116578102,
+ "learning_rate": 0.0006,
+ "loss": 5.234884738922119,
+ "step": 619
+ },
+ {
+ "epoch": 8.614847161572053,
+ "grad_norm": 0.017021115869283676,
+ "learning_rate": 0.0006,
+ "loss": 5.12271785736084,
+ "step": 620
+ },
+ {
+ "epoch": 8.62882096069869,
+ "grad_norm": 0.01773759163916111,
+ "learning_rate": 0.0006,
+ "loss": 5.242153167724609,
+ "step": 621
+ },
+ {
+ "epoch": 8.642794759825328,
+ "grad_norm": 0.015679042786359787,
+ "learning_rate": 0.0006,
+ "loss": 5.170779228210449,
+ "step": 622
+ },
+ {
+ "epoch": 8.656768558951965,
+ "grad_norm": 0.013760549947619438,
+ "learning_rate": 0.0006,
+ "loss": 5.238644599914551,
+ "step": 623
+ },
+ {
+ "epoch": 8.670742358078602,
+ "grad_norm": 0.014571522362530231,
+ "learning_rate": 0.0006,
+ "loss": 5.177056312561035,
+ "step": 624
+ },
+ {
+ "epoch": 8.68471615720524,
+ "grad_norm": 0.016209015622735023,
+ "learning_rate": 0.0006,
+ "loss": 5.163750648498535,
+ "step": 625
+ },
+ {
+ "epoch": 8.698689956331878,
+ "grad_norm": 0.016813941299915314,
+ "learning_rate": 0.0006,
+ "loss": 5.183428764343262,
+ "step": 626
+ },
+ {
+ "epoch": 8.712663755458514,
+ "grad_norm": 0.020985357463359833,
+ "learning_rate": 0.0006,
+ "loss": 5.229465007781982,
+ "step": 627
+ },
+ {
+ "epoch": 8.726637554585153,
+ "grad_norm": 0.02679632417857647,
+ "learning_rate": 0.0006,
+ "loss": 5.257368087768555,
+ "step": 628
+ },
+ {
+ "epoch": 8.74061135371179,
+ "grad_norm": 0.025756070390343666,
+ "learning_rate": 0.0006,
+ "loss": 5.253736972808838,
+ "step": 629
+ },
+ {
+ "epoch": 8.754585152838429,
+ "grad_norm": 0.02643490768969059,
+ "learning_rate": 0.0006,
+ "loss": 5.228633403778076,
+ "step": 630
+ },
+ {
+ "epoch": 8.768558951965066,
+ "grad_norm": 0.03303210437297821,
+ "learning_rate": 0.0006,
+ "loss": 5.163540363311768,
+ "step": 631
+ },
+ {
+ "epoch": 8.782532751091702,
+ "grad_norm": 0.030432431027293205,
+ "learning_rate": 0.0006,
+ "loss": 5.243169784545898,
+ "step": 632
+ },
+ {
+ "epoch": 8.796506550218341,
+ "grad_norm": 0.028631288558244705,
+ "learning_rate": 0.0006,
+ "loss": 5.166398048400879,
+ "step": 633
+ },
+ {
+ "epoch": 8.810480349344978,
+ "grad_norm": 0.026188310235738754,
+ "learning_rate": 0.0006,
+ "loss": 5.128122329711914,
+ "step": 634
+ },
+ {
+ "epoch": 8.824454148471617,
+ "grad_norm": 0.025308528915047646,
+ "learning_rate": 0.0006,
+ "loss": 5.211284637451172,
+ "step": 635
+ },
+ {
+ "epoch": 8.838427947598253,
+ "grad_norm": 0.020657729357481003,
+ "learning_rate": 0.0006,
+ "loss": 5.161575794219971,
+ "step": 636
+ },
+ {
+ "epoch": 8.85240174672489,
+ "grad_norm": 0.021901234984397888,
+ "learning_rate": 0.0006,
+ "loss": 5.20050048828125,
+ "step": 637
+ },
+ {
+ "epoch": 8.866375545851529,
+ "grad_norm": 0.021219318732619286,
+ "learning_rate": 0.0006,
+ "loss": 5.211699485778809,
+ "step": 638
+ },
+ {
+ "epoch": 8.880349344978166,
+ "grad_norm": 0.0200974028557539,
+ "learning_rate": 0.0006,
+ "loss": 5.092503547668457,
+ "step": 639
+ },
+ {
+ "epoch": 8.894323144104803,
+ "grad_norm": 0.023804882541298866,
+ "learning_rate": 0.0006,
+ "loss": 5.216068267822266,
+ "step": 640
+ },
+ {
+ "epoch": 8.908296943231441,
+ "grad_norm": 0.026088010519742966,
+ "learning_rate": 0.0006,
+ "loss": 5.155592918395996,
+ "step": 641
+ },
+ {
+ "epoch": 8.922270742358078,
+ "grad_norm": 0.02601276896893978,
+ "learning_rate": 0.0006,
+ "loss": 5.197238922119141,
+ "step": 642
+ },
+ {
+ "epoch": 8.936244541484717,
+ "grad_norm": 0.020387211814522743,
+ "learning_rate": 0.0006,
+ "loss": 5.1955437660217285,
+ "step": 643
+ },
+ {
+ "epoch": 8.950218340611354,
+ "grad_norm": 0.019214622676372528,
+ "learning_rate": 0.0006,
+ "loss": 5.206346035003662,
+ "step": 644
+ },
+ {
+ "epoch": 8.96419213973799,
+ "grad_norm": 0.019674314185976982,
+ "learning_rate": 0.0006,
+ "loss": 5.219857215881348,
+ "step": 645
+ },
+ {
+ "epoch": 8.97816593886463,
+ "grad_norm": 0.020318234339356422,
+ "learning_rate": 0.0006,
+ "loss": 5.191132545471191,
+ "step": 646
+ },
+ {
+ "epoch": 8.992139737991266,
+ "grad_norm": 0.021428676322102547,
+ "learning_rate": 0.0006,
+ "loss": 5.214695930480957,
+ "step": 647
+ },
+ {
+ "epoch": 9.0,
+ "grad_norm": 0.023235054686665535,
+ "learning_rate": 0.0006,
+ "loss": 5.229091644287109,
+ "step": 648
+ },
+ {
+ "epoch": 9.0,
+ "eval_loss": 5.289593696594238,
+ "eval_runtime": 56.9256,
+ "eval_samples_per_second": 42.898,
+ "eval_steps_per_second": 1.353,
+ "step": 648
+ },
+ {
+ "epoch": 9.013973799126637,
+ "grad_norm": 0.02320289984345436,
+ "learning_rate": 0.0006,
+ "loss": 4.983669281005859,
+ "step": 649
+ },
+ {
+ "epoch": 9.027947598253276,
+ "grad_norm": 0.018521282821893692,
+ "learning_rate": 0.0006,
+ "loss": 5.200575828552246,
+ "step": 650
+ },
+ {
+ "epoch": 9.041921397379912,
+ "grad_norm": 0.0180169939994812,
+ "learning_rate": 0.0006,
+ "loss": 5.148033618927002,
+ "step": 651
+ },
+ {
+ "epoch": 9.055895196506551,
+ "grad_norm": 0.019969860091805458,
+ "learning_rate": 0.0006,
+ "loss": 5.099125385284424,
+ "step": 652
+ },
+ {
+ "epoch": 9.069868995633188,
+ "grad_norm": 0.01643305830657482,
+ "learning_rate": 0.0006,
+ "loss": 5.092347145080566,
+ "step": 653
+ },
+ {
+ "epoch": 9.083842794759825,
+ "grad_norm": 0.016022363677620888,
+ "learning_rate": 0.0006,
+ "loss": 5.129279613494873,
+ "step": 654
+ },
+ {
+ "epoch": 9.097816593886463,
+ "grad_norm": 0.01657041162252426,
+ "learning_rate": 0.0006,
+ "loss": 5.159562587738037,
+ "step": 655
+ },
+ {
+ "epoch": 9.1117903930131,
+ "grad_norm": 0.017924228683114052,
+ "learning_rate": 0.0006,
+ "loss": 5.165497779846191,
+ "step": 656
+ },
+ {
+ "epoch": 9.125764192139737,
+ "grad_norm": 0.021592361852526665,
+ "learning_rate": 0.0006,
+ "loss": 5.204775810241699,
+ "step": 657
+ },
+ {
+ "epoch": 9.139737991266376,
+ "grad_norm": 0.026924120262265205,
+ "learning_rate": 0.0006,
+ "loss": 5.128296852111816,
+ "step": 658
+ },
+ {
+ "epoch": 9.153711790393013,
+ "grad_norm": 0.023156899958848953,
+ "learning_rate": 0.0006,
+ "loss": 5.1255340576171875,
+ "step": 659
+ },
+ {
+ "epoch": 9.167685589519651,
+ "grad_norm": 0.020670367404818535,
+ "learning_rate": 0.0006,
+ "loss": 5.128479957580566,
+ "step": 660
+ },
+ {
+ "epoch": 9.181659388646288,
+ "grad_norm": 0.025447173044085503,
+ "learning_rate": 0.0006,
+ "loss": 5.154821872711182,
+ "step": 661
+ },
+ {
+ "epoch": 9.195633187772925,
+ "grad_norm": 0.027940068393945694,
+ "learning_rate": 0.0006,
+ "loss": 5.143466472625732,
+ "step": 662
+ },
+ {
+ "epoch": 9.209606986899564,
+ "grad_norm": 0.026847844943404198,
+ "learning_rate": 0.0006,
+ "loss": 5.045665740966797,
+ "step": 663
+ },
+ {
+ "epoch": 9.2235807860262,
+ "grad_norm": 0.02340601570904255,
+ "learning_rate": 0.0006,
+ "loss": 5.21933126449585,
+ "step": 664
+ },
+ {
+ "epoch": 9.237554585152838,
+ "grad_norm": 0.02340371161699295,
+ "learning_rate": 0.0006,
+ "loss": 4.942949295043945,
+ "step": 665
+ },
+ {
+ "epoch": 9.251528384279476,
+ "grad_norm": 0.02221992425620556,
+ "learning_rate": 0.0006,
+ "loss": 5.197099208831787,
+ "step": 666
+ },
+ {
+ "epoch": 9.265502183406113,
+ "grad_norm": 0.023508677259087563,
+ "learning_rate": 0.0006,
+ "loss": 5.022897720336914,
+ "step": 667
+ },
+ {
+ "epoch": 9.279475982532752,
+ "grad_norm": 0.026120394468307495,
+ "learning_rate": 0.0006,
+ "loss": 5.087725639343262,
+ "step": 668
+ },
+ {
+ "epoch": 9.293449781659389,
+ "grad_norm": 0.027273228392004967,
+ "learning_rate": 0.0006,
+ "loss": 5.090963840484619,
+ "step": 669
+ },
+ {
+ "epoch": 9.307423580786025,
+ "grad_norm": 0.03241586685180664,
+ "learning_rate": 0.0006,
+ "loss": 5.047842979431152,
+ "step": 670
+ },
+ {
+ "epoch": 9.321397379912664,
+ "grad_norm": 0.030245667323470116,
+ "learning_rate": 0.0006,
+ "loss": 5.104803085327148,
+ "step": 671
+ },
+ {
+ "epoch": 9.335371179039301,
+ "grad_norm": 0.027698364108800888,
+ "learning_rate": 0.0006,
+ "loss": 5.0644097328186035,
+ "step": 672
+ },
+ {
+ "epoch": 9.34934497816594,
+ "grad_norm": 0.029692554846405983,
+ "learning_rate": 0.0006,
+ "loss": 5.111942291259766,
+ "step": 673
+ },
+ {
+ "epoch": 9.363318777292577,
+ "grad_norm": 0.03328656405210495,
+ "learning_rate": 0.0006,
+ "loss": 5.1574554443359375,
+ "step": 674
+ },
+ {
+ "epoch": 9.377292576419213,
+ "grad_norm": 0.031938180327415466,
+ "learning_rate": 0.0006,
+ "loss": 5.190371990203857,
+ "step": 675
+ },
+ {
+ "epoch": 9.391266375545852,
+ "grad_norm": 0.024858945980668068,
+ "learning_rate": 0.0006,
+ "loss": 5.121420860290527,
+ "step": 676
+ },
+ {
+ "epoch": 9.405240174672489,
+ "grad_norm": 0.023033203557133675,
+ "learning_rate": 0.0006,
+ "loss": 5.0974345207214355,
+ "step": 677
+ },
+ {
+ "epoch": 9.419213973799126,
+ "grad_norm": 0.022083545103669167,
+ "learning_rate": 0.0006,
+ "loss": 5.046319961547852,
+ "step": 678
+ },
+ {
+ "epoch": 9.433187772925764,
+ "grad_norm": 0.0228437427431345,
+ "learning_rate": 0.0006,
+ "loss": 5.119963645935059,
+ "step": 679
+ },
+ {
+ "epoch": 9.447161572052401,
+ "grad_norm": 0.0242351982742548,
+ "learning_rate": 0.0006,
+ "loss": 5.0444560050964355,
+ "step": 680
+ },
+ {
+ "epoch": 9.46113537117904,
+ "grad_norm": 0.02401185780763626,
+ "learning_rate": 0.0006,
+ "loss": 5.063238620758057,
+ "step": 681
+ },
+ {
+ "epoch": 9.475109170305677,
+ "grad_norm": 0.02560959942638874,
+ "learning_rate": 0.0006,
+ "loss": 5.035224914550781,
+ "step": 682
+ },
+ {
+ "epoch": 9.489082969432314,
+ "grad_norm": 0.024985626339912415,
+ "learning_rate": 0.0006,
+ "loss": 5.049284934997559,
+ "step": 683
+ },
+ {
+ "epoch": 9.503056768558952,
+ "grad_norm": 0.023178115487098694,
+ "learning_rate": 0.0006,
+ "loss": 5.2266621589660645,
+ "step": 684
+ },
+ {
+ "epoch": 9.51703056768559,
+ "grad_norm": 0.0215512253344059,
+ "learning_rate": 0.0006,
+ "loss": 5.1875715255737305,
+ "step": 685
+ },
+ {
+ "epoch": 9.531004366812226,
+ "grad_norm": 0.01662837713956833,
+ "learning_rate": 0.0006,
+ "loss": 5.137012481689453,
+ "step": 686
+ },
+ {
+ "epoch": 9.544978165938865,
+ "grad_norm": 0.01952764391899109,
+ "learning_rate": 0.0006,
+ "loss": 5.072500228881836,
+ "step": 687
+ },
+ {
+ "epoch": 9.558951965065502,
+ "grad_norm": 0.017521077767014503,
+ "learning_rate": 0.0006,
+ "loss": 5.045711994171143,
+ "step": 688
+ },
+ {
+ "epoch": 9.57292576419214,
+ "grad_norm": 0.017355265095829964,
+ "learning_rate": 0.0006,
+ "loss": 5.02881383895874,
+ "step": 689
+ },
+ {
+ "epoch": 9.586899563318777,
+ "grad_norm": 0.015587719157338142,
+ "learning_rate": 0.0006,
+ "loss": 5.132787704467773,
+ "step": 690
+ },
+ {
+ "epoch": 9.600873362445414,
+ "grad_norm": 0.017469845712184906,
+ "learning_rate": 0.0006,
+ "loss": 5.105447292327881,
+ "step": 691
+ },
+ {
+ "epoch": 9.614847161572053,
+ "grad_norm": 0.019630100578069687,
+ "learning_rate": 0.0006,
+ "loss": 5.032464981079102,
+ "step": 692
+ },
+ {
+ "epoch": 9.62882096069869,
+ "grad_norm": 0.01784994639456272,
+ "learning_rate": 0.0006,
+ "loss": 5.081345558166504,
+ "step": 693
+ },
+ {
+ "epoch": 9.642794759825328,
+ "grad_norm": 0.01872754842042923,
+ "learning_rate": 0.0006,
+ "loss": 4.9772114753723145,
+ "step": 694
+ },
+ {
+ "epoch": 9.656768558951965,
+ "grad_norm": 0.01999032311141491,
+ "learning_rate": 0.0006,
+ "loss": 5.089740753173828,
+ "step": 695
+ },
+ {
+ "epoch": 9.670742358078602,
+ "grad_norm": 0.01710767112672329,
+ "learning_rate": 0.0006,
+ "loss": 4.9869890213012695,
+ "step": 696
+ },
+ {
+ "epoch": 9.68471615720524,
+ "grad_norm": 0.015144161880016327,
+ "learning_rate": 0.0006,
+ "loss": 5.022564888000488,
+ "step": 697
+ },
+ {
+ "epoch": 9.698689956331878,
+ "grad_norm": 0.01618543453514576,
+ "learning_rate": 0.0006,
+ "loss": 5.035921573638916,
+ "step": 698
+ },
+ {
+ "epoch": 9.712663755458514,
+ "grad_norm": 0.014653387479484081,
+ "learning_rate": 0.0006,
+ "loss": 4.914456367492676,
+ "step": 699
+ },
+ {
+ "epoch": 9.726637554585153,
+ "grad_norm": 0.013434977270662785,
+ "learning_rate": 0.0006,
+ "loss": 5.063764572143555,
+ "step": 700
+ },
+ {
+ "epoch": 9.74061135371179,
+ "grad_norm": 0.015128864906728268,
+ "learning_rate": 0.0006,
+ "loss": 5.052367687225342,
+ "step": 701
+ },
+ {
+ "epoch": 9.754585152838429,
+ "grad_norm": 0.014210768043994904,
+ "learning_rate": 0.0006,
+ "loss": 5.016489028930664,
+ "step": 702
+ },
+ {
+ "epoch": 9.768558951965066,
+ "grad_norm": 0.015191680751740932,
+ "learning_rate": 0.0006,
+ "loss": 4.990176200866699,
+ "step": 703
+ },
+ {
+ "epoch": 9.782532751091702,
+ "grad_norm": 0.018285127356648445,
+ "learning_rate": 0.0006,
+ "loss": 5.015164375305176,
+ "step": 704
+ },
+ {
+ "epoch": 9.796506550218341,
+ "grad_norm": 0.0204079058021307,
+ "learning_rate": 0.0006,
+ "loss": 5.091142654418945,
+ "step": 705
+ },
+ {
+ "epoch": 9.810480349344978,
+ "grad_norm": 0.01980586163699627,
+ "learning_rate": 0.0006,
+ "loss": 5.03350830078125,
+ "step": 706
+ },
+ {
+ "epoch": 9.824454148471617,
+ "grad_norm": 0.02048729732632637,
+ "learning_rate": 0.0006,
+ "loss": 5.03611421585083,
+ "step": 707
+ },
+ {
+ "epoch": 9.838427947598253,
+ "grad_norm": 0.02058163844048977,
+ "learning_rate": 0.0006,
+ "loss": 4.9613447189331055,
+ "step": 708
+ },
+ {
+ "epoch": 9.85240174672489,
+ "grad_norm": 0.018004924058914185,
+ "learning_rate": 0.0006,
+ "loss": 4.966541767120361,
+ "step": 709
+ },
+ {
+ "epoch": 9.866375545851529,
+ "grad_norm": 0.0198152307420969,
+ "learning_rate": 0.0006,
+ "loss": 5.01750373840332,
+ "step": 710
+ },
+ {
+ "epoch": 9.880349344978166,
+ "grad_norm": 0.018532969057559967,
+ "learning_rate": 0.0006,
+ "loss": 5.097424507141113,
+ "step": 711
+ },
+ {
+ "epoch": 9.894323144104803,
+ "grad_norm": 0.018939971923828125,
+ "learning_rate": 0.0006,
+ "loss": 5.177223205566406,
+ "step": 712
+ },
+ {
+ "epoch": 9.908296943231441,
+ "grad_norm": 0.017969170585274696,
+ "learning_rate": 0.0006,
+ "loss": 5.118013858795166,
+ "step": 713
+ },
+ {
+ "epoch": 9.922270742358078,
+ "grad_norm": 0.018021270632743835,
+ "learning_rate": 0.0006,
+ "loss": 5.034787178039551,
+ "step": 714
+ },
+ {
+ "epoch": 9.936244541484717,
+ "grad_norm": 0.017009710893034935,
+ "learning_rate": 0.0006,
+ "loss": 5.003920555114746,
+ "step": 715
+ },
+ {
+ "epoch": 9.950218340611354,
+ "grad_norm": 0.017722049728035927,
+ "learning_rate": 0.0006,
+ "loss": 5.031939506530762,
+ "step": 716
+ },
+ {
+ "epoch": 9.96419213973799,
+ "grad_norm": 0.017134087160229683,
+ "learning_rate": 0.0006,
+ "loss": 5.093474388122559,
+ "step": 717
+ },
+ {
+ "epoch": 9.97816593886463,
+ "grad_norm": 0.018511613830924034,
+ "learning_rate": 0.0006,
+ "loss": 5.078604698181152,
+ "step": 718
+ },
+ {
+ "epoch": 9.992139737991266,
+ "grad_norm": 0.021206334233283997,
+ "learning_rate": 0.0006,
+ "loss": 5.052572727203369,
+ "step": 719
+ },
+ {
+ "epoch": 10.0,
+ "grad_norm": 0.02239062264561653,
+ "learning_rate": 0.0006,
+ "loss": 5.063547134399414,
+ "step": 720
+ },
+ {
+ "epoch": 10.0,
+ "eval_loss": 5.173044681549072,
+ "eval_runtime": 56.7287,
+ "eval_samples_per_second": 43.047,
+ "eval_steps_per_second": 1.357,
+ "step": 720
+ },
+ {
+ "epoch": 10.013973799126637,
+ "grad_norm": 0.020757969468832016,
+ "learning_rate": 0.0006,
+ "loss": 5.045648574829102,
+ "step": 721
+ },
+ {
+ "epoch": 10.027947598253276,
+ "grad_norm": 0.022697702050209045,
+ "learning_rate": 0.0006,
+ "loss": 4.993833065032959,
+ "step": 722
+ },
+ {
+ "epoch": 10.041921397379912,
+ "grad_norm": 0.02434331737458706,
+ "learning_rate": 0.0006,
+ "loss": 5.073219299316406,
+ "step": 723
+ },
+ {
+ "epoch": 10.055895196506551,
+ "grad_norm": 0.025499660521745682,
+ "learning_rate": 0.0006,
+ "loss": 5.019072532653809,
+ "step": 724
+ },
+ {
+ "epoch": 10.069868995633188,
+ "grad_norm": 0.029281053692102432,
+ "learning_rate": 0.0006,
+ "loss": 4.959851264953613,
+ "step": 725
+ },
+ {
+ "epoch": 10.083842794759825,
+ "grad_norm": 0.033878110349178314,
+ "learning_rate": 0.0006,
+ "loss": 5.132717609405518,
+ "step": 726
+ },
+ {
+ "epoch": 10.097816593886463,
+ "grad_norm": 0.03385764732956886,
+ "learning_rate": 0.0006,
+ "loss": 5.0482892990112305,
+ "step": 727
+ },
+ {
+ "epoch": 10.1117903930131,
+ "grad_norm": 0.03466648980975151,
+ "learning_rate": 0.0006,
+ "loss": 4.946855545043945,
+ "step": 728
+ },
+ {
+ "epoch": 10.125764192139737,
+ "grad_norm": 0.028546305373311043,
+ "learning_rate": 0.0006,
+ "loss": 5.027774333953857,
+ "step": 729
+ },
+ {
+ "epoch": 10.139737991266376,
+ "grad_norm": 0.030575869604945183,
+ "learning_rate": 0.0006,
+ "loss": 4.9768476486206055,
+ "step": 730
+ },
+ {
+ "epoch": 10.153711790393013,
+ "grad_norm": 0.03614303469657898,
+ "learning_rate": 0.0006,
+ "loss": 5.078113555908203,
+ "step": 731
+ },
+ {
+ "epoch": 10.167685589519651,
+ "grad_norm": 0.040991660207509995,
+ "learning_rate": 0.0006,
+ "loss": 5.033185958862305,
+ "step": 732
+ },
+ {
+ "epoch": 10.181659388646288,
+ "grad_norm": 0.04551699757575989,
+ "learning_rate": 0.0006,
+ "loss": 5.0820746421813965,
+ "step": 733
+ },
+ {
+ "epoch": 10.195633187772925,
+ "grad_norm": 0.040193989872932434,
+ "learning_rate": 0.0006,
+ "loss": 5.059700965881348,
+ "step": 734
+ },
+ {
+ "epoch": 10.209606986899564,
+ "grad_norm": 0.035851605236530304,
+ "learning_rate": 0.0006,
+ "loss": 5.142467498779297,
+ "step": 735
+ },
+ {
+ "epoch": 10.2235807860262,
+ "grad_norm": 0.034200169146060944,
+ "learning_rate": 0.0006,
+ "loss": 5.072312355041504,
+ "step": 736
+ },
+ {
+ "epoch": 10.237554585152838,
+ "grad_norm": 0.033621110022068024,
+ "learning_rate": 0.0006,
+ "loss": 5.08037805557251,
+ "step": 737
+ },
+ {
+ "epoch": 10.251528384279476,
+ "grad_norm": 0.037372369319200516,
+ "learning_rate": 0.0006,
+ "loss": 4.998425483703613,
+ "step": 738
+ },
+ {
+ "epoch": 10.265502183406113,
+ "grad_norm": 0.037044707685709,
+ "learning_rate": 0.0006,
+ "loss": 4.923993110656738,
+ "step": 739
+ },
+ {
+ "epoch": 10.279475982532752,
+ "grad_norm": 0.030454808846116066,
+ "learning_rate": 0.0006,
+ "loss": 5.0889997482299805,
+ "step": 740
+ },
+ {
+ "epoch": 10.293449781659389,
+ "grad_norm": 0.030969638377428055,
+ "learning_rate": 0.0006,
+ "loss": 4.984112739562988,
+ "step": 741
+ },
+ {
+ "epoch": 10.307423580786025,
+ "grad_norm": 0.0276983380317688,
+ "learning_rate": 0.0006,
+ "loss": 4.955412864685059,
+ "step": 742
+ },
+ {
+ "epoch": 10.321397379912664,
+ "grad_norm": 0.02716052532196045,
+ "learning_rate": 0.0006,
+ "loss": 5.0098876953125,
+ "step": 743
+ },
+ {
+ "epoch": 10.335371179039301,
+ "grad_norm": 0.02403552085161209,
+ "learning_rate": 0.0006,
+ "loss": 5.165160655975342,
+ "step": 744
+ },
+ {
+ "epoch": 10.34934497816594,
+ "grad_norm": 0.0257862601429224,
+ "learning_rate": 0.0006,
+ "loss": 5.001547813415527,
+ "step": 745
+ },
+ {
+ "epoch": 10.363318777292577,
+ "grad_norm": 0.02300378680229187,
+ "learning_rate": 0.0006,
+ "loss": 5.038888931274414,
+ "step": 746
+ },
+ {
+ "epoch": 10.377292576419213,
+ "grad_norm": 0.019766854122281075,
+ "learning_rate": 0.0006,
+ "loss": 4.924291610717773,
+ "step": 747
+ },
+ {
+ "epoch": 10.391266375545852,
+ "grad_norm": 0.01888395845890045,
+ "learning_rate": 0.0006,
+ "loss": 5.059971332550049,
+ "step": 748
+ },
+ {
+ "epoch": 10.405240174672489,
+ "grad_norm": 0.015577499754726887,
+ "learning_rate": 0.0006,
+ "loss": 4.987686634063721,
+ "step": 749
+ },
+ {
+ "epoch": 10.419213973799126,
+ "grad_norm": 0.016012346372008324,
+ "learning_rate": 0.0006,
+ "loss": 5.015157699584961,
+ "step": 750
+ },
+ {
+ "epoch": 10.433187772925764,
+ "grad_norm": 0.016277998685836792,
+ "learning_rate": 0.0006,
+ "loss": 5.004931449890137,
+ "step": 751
+ },
+ {
+ "epoch": 10.447161572052401,
+ "grad_norm": 0.017504561692476273,
+ "learning_rate": 0.0006,
+ "loss": 4.902827739715576,
+ "step": 752
+ },
+ {
+ "epoch": 10.46113537117904,
+ "grad_norm": 0.01569022797048092,
+ "learning_rate": 0.0006,
+ "loss": 4.954863548278809,
+ "step": 753
+ },
+ {
+ "epoch": 10.475109170305677,
+ "grad_norm": 0.014114422723650932,
+ "learning_rate": 0.0006,
+ "loss": 5.00656795501709,
+ "step": 754
+ },
+ {
+ "epoch": 10.489082969432314,
+ "grad_norm": 0.01573832333087921,
+ "learning_rate": 0.0006,
+ "loss": 4.978353023529053,
+ "step": 755
+ },
+ {
+ "epoch": 10.503056768558952,
+ "grad_norm": 0.016480036079883575,
+ "learning_rate": 0.0006,
+ "loss": 4.953327655792236,
+ "step": 756
+ },
+ {
+ "epoch": 10.51703056768559,
+ "grad_norm": 0.014008025638759136,
+ "learning_rate": 0.0006,
+ "loss": 5.050840377807617,
+ "step": 757
+ },
+ {
+ "epoch": 10.531004366812226,
+ "grad_norm": 0.013147998601198196,
+ "learning_rate": 0.0006,
+ "loss": 4.974964141845703,
+ "step": 758
+ },
+ {
+ "epoch": 10.544978165938865,
+ "grad_norm": 0.013807603158056736,
+ "learning_rate": 0.0006,
+ "loss": 4.927907943725586,
+ "step": 759
+ },
+ {
+ "epoch": 10.558951965065502,
+ "grad_norm": 0.013555224984884262,
+ "learning_rate": 0.0006,
+ "loss": 4.980690956115723,
+ "step": 760
+ },
+ {
+ "epoch": 10.57292576419214,
+ "grad_norm": 0.013045408762991428,
+ "learning_rate": 0.0006,
+ "loss": 4.9436116218566895,
+ "step": 761
+ },
+ {
+ "epoch": 10.586899563318777,
+ "grad_norm": 0.012858862057328224,
+ "learning_rate": 0.0006,
+ "loss": 4.9595770835876465,
+ "step": 762
+ },
+ {
+ "epoch": 10.600873362445414,
+ "grad_norm": 0.01468253880739212,
+ "learning_rate": 0.0006,
+ "loss": 4.90725040435791,
+ "step": 763
+ },
+ {
+ "epoch": 10.614847161572053,
+ "grad_norm": 0.013224679045379162,
+ "learning_rate": 0.0006,
+ "loss": 4.877615928649902,
+ "step": 764
+ },
+ {
+ "epoch": 10.62882096069869,
+ "grad_norm": 0.013555348850786686,
+ "learning_rate": 0.0006,
+ "loss": 4.9456610679626465,
+ "step": 765
+ },
+ {
+ "epoch": 10.642794759825328,
+ "grad_norm": 0.014387160539627075,
+ "learning_rate": 0.0006,
+ "loss": 4.9233717918396,
+ "step": 766
+ },
+ {
+ "epoch": 10.656768558951965,
+ "grad_norm": 0.014354088343679905,
+ "learning_rate": 0.0006,
+ "loss": 4.983511924743652,
+ "step": 767
+ },
+ {
+ "epoch": 10.670742358078602,
+ "grad_norm": 0.01332700252532959,
+ "learning_rate": 0.0006,
+ "loss": 4.934416770935059,
+ "step": 768
+ },
+ {
+ "epoch": 10.68471615720524,
+ "grad_norm": 0.011543313041329384,
+ "learning_rate": 0.0006,
+ "loss": 4.96388053894043,
+ "step": 769
+ },
+ {
+ "epoch": 10.698689956331878,
+ "grad_norm": 0.012330878525972366,
+ "learning_rate": 0.0006,
+ "loss": 4.946396350860596,
+ "step": 770
+ },
+ {
+ "epoch": 10.712663755458514,
+ "grad_norm": 0.01379779726266861,
+ "learning_rate": 0.0006,
+ "loss": 4.981019020080566,
+ "step": 771
+ },
+ {
+ "epoch": 10.726637554585153,
+ "grad_norm": 0.014612431637942791,
+ "learning_rate": 0.0006,
+ "loss": 4.894567966461182,
+ "step": 772
+ },
+ {
+ "epoch": 10.74061135371179,
+ "grad_norm": 0.013062899932265282,
+ "learning_rate": 0.0006,
+ "loss": 5.033038139343262,
+ "step": 773
+ },
+ {
+ "epoch": 10.754585152838429,
+ "grad_norm": 0.013898049481213093,
+ "learning_rate": 0.0006,
+ "loss": 4.823636054992676,
+ "step": 774
+ },
+ {
+ "epoch": 10.768558951965066,
+ "grad_norm": 0.01502019353210926,
+ "learning_rate": 0.0006,
+ "loss": 5.020156383514404,
+ "step": 775
+ },
+ {
+ "epoch": 10.782532751091702,
+ "grad_norm": 0.013810204342007637,
+ "learning_rate": 0.0006,
+ "loss": 4.935647010803223,
+ "step": 776
+ },
+ {
+ "epoch": 10.796506550218341,
+ "grad_norm": 0.013616513460874557,
+ "learning_rate": 0.0006,
+ "loss": 4.901486396789551,
+ "step": 777
+ },
+ {
+ "epoch": 10.810480349344978,
+ "grad_norm": 0.013601024635136127,
+ "learning_rate": 0.0006,
+ "loss": 4.931873321533203,
+ "step": 778
+ },
+ {
+ "epoch": 10.824454148471617,
+ "grad_norm": 0.013668350875377655,
+ "learning_rate": 0.0006,
+ "loss": 4.937256813049316,
+ "step": 779
+ },
+ {
+ "epoch": 10.838427947598253,
+ "grad_norm": 0.01581823080778122,
+ "learning_rate": 0.0006,
+ "loss": 4.9611406326293945,
+ "step": 780
+ },
+ {
+ "epoch": 10.85240174672489,
+ "grad_norm": 0.019478484988212585,
+ "learning_rate": 0.0006,
+ "loss": 4.9212236404418945,
+ "step": 781
+ },
+ {
+ "epoch": 10.866375545851529,
+ "grad_norm": 0.026921333745121956,
+ "learning_rate": 0.0006,
+ "loss": 4.88501501083374,
+ "step": 782
+ },
+ {
+ "epoch": 10.880349344978166,
+ "grad_norm": 0.03155914694070816,
+ "learning_rate": 0.0006,
+ "loss": 4.989967346191406,
+ "step": 783
+ },
+ {
+ "epoch": 10.894323144104803,
+ "grad_norm": 0.029689429327845573,
+ "learning_rate": 0.0006,
+ "loss": 4.936000823974609,
+ "step": 784
+ },
+ {
+ "epoch": 10.908296943231441,
+ "grad_norm": 0.026626063510775566,
+ "learning_rate": 0.0006,
+ "loss": 4.984785079956055,
+ "step": 785
+ },
+ {
+ "epoch": 10.922270742358078,
+ "grad_norm": 0.026168212294578552,
+ "learning_rate": 0.0006,
+ "loss": 4.908282279968262,
+ "step": 786
+ },
+ {
+ "epoch": 10.936244541484717,
+ "grad_norm": 0.02828158251941204,
+ "learning_rate": 0.0006,
+ "loss": 4.926390647888184,
+ "step": 787
+ },
+ {
+ "epoch": 10.950218340611354,
+ "grad_norm": 0.02649078331887722,
+ "learning_rate": 0.0006,
+ "loss": 5.00400972366333,
+ "step": 788
+ },
+ {
+ "epoch": 10.96419213973799,
+ "grad_norm": 0.0256856270134449,
+ "learning_rate": 0.0006,
+ "loss": 5.0317487716674805,
+ "step": 789
+ },
+ {
+ "epoch": 10.97816593886463,
+ "grad_norm": 0.024728883057832718,
+ "learning_rate": 0.0006,
+ "loss": 4.9299635887146,
+ "step": 790
+ },
+ {
+ "epoch": 10.992139737991266,
+ "grad_norm": 0.026241116225719452,
+ "learning_rate": 0.0006,
+ "loss": 4.947319984436035,
+ "step": 791
+ },
+ {
+ "epoch": 11.0,
+ "grad_norm": 0.023526743054389954,
+ "learning_rate": 0.0006,
+ "loss": 4.94429349899292,
+ "step": 792
+ },
+ {
+ "epoch": 11.0,
+ "eval_loss": 5.104345798492432,
+ "eval_runtime": 57.0955,
+ "eval_samples_per_second": 42.77,
+ "eval_steps_per_second": 1.349,
+ "step": 792
+ },
+ {
+ "epoch": 11.013973799126637,
+ "grad_norm": 0.0216788612306118,
+ "learning_rate": 0.0006,
+ "loss": 4.881950855255127,
+ "step": 793
+ },
+ {
+ "epoch": 11.027947598253276,
+ "grad_norm": 0.01830463856458664,
+ "learning_rate": 0.0006,
+ "loss": 4.971016883850098,
+ "step": 794
+ },
+ {
+ "epoch": 11.041921397379912,
+ "grad_norm": 0.02204521745443344,
+ "learning_rate": 0.0006,
+ "loss": 4.894623756408691,
+ "step": 795
+ },
+ {
+ "epoch": 11.055895196506551,
+ "grad_norm": 0.0210255216807127,
+ "learning_rate": 0.0006,
+ "loss": 4.834009170532227,
+ "step": 796
+ },
+ {
+ "epoch": 11.069868995633188,
+ "grad_norm": 0.019797448068857193,
+ "learning_rate": 0.0006,
+ "loss": 5.086714744567871,
+ "step": 797
+ },
+ {
+ "epoch": 11.083842794759825,
+ "grad_norm": 0.02036641724407673,
+ "learning_rate": 0.0006,
+ "loss": 4.920056343078613,
+ "step": 798
+ },
+ {
+ "epoch": 11.097816593886463,
+ "grad_norm": 0.01966066285967827,
+ "learning_rate": 0.0006,
+ "loss": 4.879859447479248,
+ "step": 799
+ },
+ {
+ "epoch": 11.1117903930131,
+ "grad_norm": 0.0202149897813797,
+ "learning_rate": 0.0006,
+ "loss": 4.933640480041504,
+ "step": 800
+ },
+ {
+ "epoch": 11.125764192139737,
+ "grad_norm": 0.01893039606511593,
+ "learning_rate": 0.0006,
+ "loss": 4.901970863342285,
+ "step": 801
+ },
+ {
+ "epoch": 11.139737991266376,
+ "grad_norm": 0.016040155664086342,
+ "learning_rate": 0.0006,
+ "loss": 4.905491828918457,
+ "step": 802
+ },
+ {
+ "epoch": 11.153711790393013,
+ "grad_norm": 0.01720350608229637,
+ "learning_rate": 0.0006,
+ "loss": 4.798816680908203,
+ "step": 803
+ },
+ {
+ "epoch": 11.167685589519651,
+ "grad_norm": 0.017002852633595467,
+ "learning_rate": 0.0006,
+ "loss": 4.851859092712402,
+ "step": 804
+ },
+ {
+ "epoch": 11.181659388646288,
+ "grad_norm": 0.015594680793583393,
+ "learning_rate": 0.0006,
+ "loss": 4.90328311920166,
+ "step": 805
+ },
+ {
+ "epoch": 11.195633187772925,
+ "grad_norm": 0.014904462732374668,
+ "learning_rate": 0.0006,
+ "loss": 4.904215335845947,
+ "step": 806
+ },
+ {
+ "epoch": 11.209606986899564,
+ "grad_norm": 0.01366228237748146,
+ "learning_rate": 0.0006,
+ "loss": 4.9813055992126465,
+ "step": 807
+ },
+ {
+ "epoch": 11.2235807860262,
+ "grad_norm": 0.014719638973474503,
+ "learning_rate": 0.0006,
+ "loss": 4.927306175231934,
+ "step": 808
+ },
+ {
+ "epoch": 11.237554585152838,
+ "grad_norm": 0.015016036108136177,
+ "learning_rate": 0.0006,
+ "loss": 4.874215126037598,
+ "step": 809
+ },
+ {
+ "epoch": 11.251528384279476,
+ "grad_norm": 0.014925424009561539,
+ "learning_rate": 0.0006,
+ "loss": 4.838257789611816,
+ "step": 810
+ },
+ {
+ "epoch": 11.265502183406113,
+ "grad_norm": 0.014335823245346546,
+ "learning_rate": 0.0006,
+ "loss": 4.940608024597168,
+ "step": 811
+ },
+ {
+ "epoch": 11.279475982532752,
+ "grad_norm": 0.014987512491643429,
+ "learning_rate": 0.0006,
+ "loss": 4.831531524658203,
+ "step": 812
+ },
+ {
+ "epoch": 11.293449781659389,
+ "grad_norm": 0.016087457537651062,
+ "learning_rate": 0.0006,
+ "loss": 4.863227367401123,
+ "step": 813
+ },
+ {
+ "epoch": 11.307423580786025,
+ "grad_norm": 0.01671300269663334,
+ "learning_rate": 0.0006,
+ "loss": 4.8535661697387695,
+ "step": 814
+ },
+ {
+ "epoch": 11.321397379912664,
+ "grad_norm": 0.015229142270982265,
+ "learning_rate": 0.0006,
+ "loss": 4.948635101318359,
+ "step": 815
+ },
+ {
+ "epoch": 11.335371179039301,
+ "grad_norm": 0.01532587967813015,
+ "learning_rate": 0.0006,
+ "loss": 4.9035797119140625,
+ "step": 816
+ },
+ {
+ "epoch": 11.34934497816594,
+ "grad_norm": 0.015715476125478745,
+ "learning_rate": 0.0006,
+ "loss": 4.894709587097168,
+ "step": 817
+ },
+ {
+ "epoch": 11.363318777292577,
+ "grad_norm": 0.014156977646052837,
+ "learning_rate": 0.0006,
+ "loss": 4.960206985473633,
+ "step": 818
+ },
+ {
+ "epoch": 11.377292576419213,
+ "grad_norm": 0.017076566815376282,
+ "learning_rate": 0.0006,
+ "loss": 5.005850791931152,
+ "step": 819
+ },
+ {
+ "epoch": 11.391266375545852,
+ "grad_norm": 0.0172084029763937,
+ "learning_rate": 0.0006,
+ "loss": 4.903355598449707,
+ "step": 820
+ },
+ {
+ "epoch": 11.405240174672489,
+ "grad_norm": 0.0180258359760046,
+ "learning_rate": 0.0006,
+ "loss": 4.842667579650879,
+ "step": 821
+ },
+ {
+ "epoch": 11.419213973799126,
+ "grad_norm": 0.022130120545625687,
+ "learning_rate": 0.0006,
+ "loss": 4.799970626831055,
+ "step": 822
+ },
+ {
+ "epoch": 11.433187772925764,
+ "grad_norm": 0.02384897693991661,
+ "learning_rate": 0.0006,
+ "loss": 4.860746383666992,
+ "step": 823
+ },
+ {
+ "epoch": 11.447161572052401,
+ "grad_norm": 0.021881572902202606,
+ "learning_rate": 0.0006,
+ "loss": 4.918107032775879,
+ "step": 824
+ },
+ {
+ "epoch": 11.46113537117904,
+ "grad_norm": 0.023085080087184906,
+ "learning_rate": 0.0006,
+ "loss": 4.878946304321289,
+ "step": 825
+ },
+ {
+ "epoch": 11.475109170305677,
+ "grad_norm": 0.02291012555360794,
+ "learning_rate": 0.0006,
+ "loss": 4.958887100219727,
+ "step": 826
+ },
+ {
+ "epoch": 11.489082969432314,
+ "grad_norm": 0.0199968870729208,
+ "learning_rate": 0.0006,
+ "loss": 4.80955696105957,
+ "step": 827
+ },
+ {
+ "epoch": 11.503056768558952,
+ "grad_norm": 0.018702229484915733,
+ "learning_rate": 0.0006,
+ "loss": 5.0276408195495605,
+ "step": 828
+ },
+ {
+ "epoch": 11.51703056768559,
+ "grad_norm": 0.017787311226129532,
+ "learning_rate": 0.0006,
+ "loss": 4.96018123626709,
+ "step": 829
+ },
+ {
+ "epoch": 11.531004366812226,
+ "grad_norm": 0.01704501546919346,
+ "learning_rate": 0.0006,
+ "loss": 4.914553165435791,
+ "step": 830
+ },
+ {
+ "epoch": 11.544978165938865,
+ "grad_norm": 0.017519677057862282,
+ "learning_rate": 0.0006,
+ "loss": 4.798098087310791,
+ "step": 831
+ },
+ {
+ "epoch": 11.558951965065502,
+ "grad_norm": 0.018511991947889328,
+ "learning_rate": 0.0006,
+ "loss": 4.817785263061523,
+ "step": 832
+ },
+ {
+ "epoch": 11.57292576419214,
+ "grad_norm": 0.01876644790172577,
+ "learning_rate": 0.0006,
+ "loss": 4.818995475769043,
+ "step": 833
+ },
+ {
+ "epoch": 11.586899563318777,
+ "grad_norm": 0.01753568835556507,
+ "learning_rate": 0.0006,
+ "loss": 4.811938285827637,
+ "step": 834
+ },
+ {
+ "epoch": 11.600873362445414,
+ "grad_norm": 0.017046531662344933,
+ "learning_rate": 0.0006,
+ "loss": 4.914674282073975,
+ "step": 835
+ },
+ {
+ "epoch": 11.614847161572053,
+ "grad_norm": 0.01802094094455242,
+ "learning_rate": 0.0006,
+ "loss": 4.993284225463867,
+ "step": 836
+ },
+ {
+ "epoch": 11.62882096069869,
+ "grad_norm": 0.019493764266371727,
+ "learning_rate": 0.0006,
+ "loss": 4.895628452301025,
+ "step": 837
+ },
+ {
+ "epoch": 11.642794759825328,
+ "grad_norm": 0.019461210817098618,
+ "learning_rate": 0.0006,
+ "loss": 4.869269371032715,
+ "step": 838
+ },
+ {
+ "epoch": 11.656768558951965,
+ "grad_norm": 0.01981567218899727,
+ "learning_rate": 0.0006,
+ "loss": 4.863052845001221,
+ "step": 839
+ },
+ {
+ "epoch": 11.670742358078602,
+ "grad_norm": 0.019894491881132126,
+ "learning_rate": 0.0006,
+ "loss": 4.8775787353515625,
+ "step": 840
+ },
+ {
+ "epoch": 11.68471615720524,
+ "grad_norm": 0.01875786855816841,
+ "learning_rate": 0.0006,
+ "loss": 4.817763328552246,
+ "step": 841
+ },
+ {
+ "epoch": 11.698689956331878,
+ "grad_norm": 0.020926695317029953,
+ "learning_rate": 0.0006,
+ "loss": 4.7470526695251465,
+ "step": 842
+ },
+ {
+ "epoch": 11.712663755458514,
+ "grad_norm": 0.021477103233337402,
+ "learning_rate": 0.0006,
+ "loss": 4.762682914733887,
+ "step": 843
+ },
+ {
+ "epoch": 11.726637554585153,
+ "grad_norm": 0.020628711208701134,
+ "learning_rate": 0.0006,
+ "loss": 4.807985782623291,
+ "step": 844
+ },
+ {
+ "epoch": 11.74061135371179,
+ "grad_norm": 0.021489612758159637,
+ "learning_rate": 0.0006,
+ "loss": 4.843091011047363,
+ "step": 845
+ },
+ {
+ "epoch": 11.754585152838429,
+ "grad_norm": 0.021194491535425186,
+ "learning_rate": 0.0006,
+ "loss": 4.8300251960754395,
+ "step": 846
+ },
+ {
+ "epoch": 11.768558951965066,
+ "grad_norm": 0.01833350397646427,
+ "learning_rate": 0.0006,
+ "loss": 5.000202178955078,
+ "step": 847
+ },
+ {
+ "epoch": 11.782532751091702,
+ "grad_norm": 0.018096931278705597,
+ "learning_rate": 0.0006,
+ "loss": 4.880648136138916,
+ "step": 848
+ },
+ {
+ "epoch": 11.796506550218341,
+ "grad_norm": 0.0197161752730608,
+ "learning_rate": 0.0006,
+ "loss": 4.788316249847412,
+ "step": 849
+ },
+ {
+ "epoch": 11.810480349344978,
+ "grad_norm": 0.0203181654214859,
+ "learning_rate": 0.0006,
+ "loss": 4.870944976806641,
+ "step": 850
+ },
+ {
+ "epoch": 11.824454148471617,
+ "grad_norm": 0.01908455230295658,
+ "learning_rate": 0.0006,
+ "loss": 4.890144348144531,
+ "step": 851
+ },
+ {
+ "epoch": 11.838427947598253,
+ "grad_norm": 0.021281801164150238,
+ "learning_rate": 0.0006,
+ "loss": 4.89578914642334,
+ "step": 852
+ },
+ {
+ "epoch": 11.85240174672489,
+ "grad_norm": 0.02124273031949997,
+ "learning_rate": 0.0006,
+ "loss": 4.883749008178711,
+ "step": 853
+ },
+ {
+ "epoch": 11.866375545851529,
+ "grad_norm": 0.020969398319721222,
+ "learning_rate": 0.0006,
+ "loss": 4.809142589569092,
+ "step": 854
+ },
+ {
+ "epoch": 11.880349344978166,
+ "grad_norm": 0.01967657171189785,
+ "learning_rate": 0.0006,
+ "loss": 4.821864604949951,
+ "step": 855
+ },
+ {
+ "epoch": 11.894323144104803,
+ "grad_norm": 0.020887333899736404,
+ "learning_rate": 0.0006,
+ "loss": 4.917985916137695,
+ "step": 856
+ },
+ {
+ "epoch": 11.908296943231441,
+ "grad_norm": 0.019203083589673042,
+ "learning_rate": 0.0006,
+ "loss": 4.824960708618164,
+ "step": 857
+ },
+ {
+ "epoch": 11.922270742358078,
+ "grad_norm": 0.020162401720881462,
+ "learning_rate": 0.0006,
+ "loss": 4.772617816925049,
+ "step": 858
+ },
+ {
+ "epoch": 11.936244541484717,
+ "grad_norm": 0.02251487225294113,
+ "learning_rate": 0.0006,
+ "loss": 4.735418319702148,
+ "step": 859
+ },
+ {
+ "epoch": 11.950218340611354,
+ "grad_norm": 0.020313527435064316,
+ "learning_rate": 0.0006,
+ "loss": 4.939233779907227,
+ "step": 860
+ },
+ {
+ "epoch": 11.96419213973799,
+ "grad_norm": 0.021635599434375763,
+ "learning_rate": 0.0006,
+ "loss": 4.861807346343994,
+ "step": 861
+ },
+ {
+ "epoch": 11.97816593886463,
+ "grad_norm": 0.02285311557352543,
+ "learning_rate": 0.0006,
+ "loss": 4.932732582092285,
+ "step": 862
+ },
+ {
+ "epoch": 11.992139737991266,
+ "grad_norm": 0.02072535641491413,
+ "learning_rate": 0.0006,
+ "loss": 4.796767711639404,
+ "step": 863
+ },
+ {
+ "epoch": 12.0,
+ "grad_norm": 0.020588304847478867,
+ "learning_rate": 0.0006,
+ "loss": 4.997537612915039,
+ "step": 864
+ },
+ {
+ "epoch": 12.0,
+ "eval_loss": 5.044651985168457,
+ "eval_runtime": 56.9145,
+ "eval_samples_per_second": 42.906,
+ "eval_steps_per_second": 1.353,
+ "step": 864
+ },
+ {
+ "epoch": 12.013973799126637,
+ "grad_norm": 0.017597166821360588,
+ "learning_rate": 0.0006,
+ "loss": 4.830524444580078,
+ "step": 865
+ },
+ {
+ "epoch": 12.027947598253276,
+ "grad_norm": 0.01716788113117218,
+ "learning_rate": 0.0006,
+ "loss": 4.894325256347656,
+ "step": 866
+ },
+ {
+ "epoch": 12.041921397379912,
+ "grad_norm": 0.017712457105517387,
+ "learning_rate": 0.0006,
+ "loss": 4.868297576904297,
+ "step": 867
+ },
+ {
+ "epoch": 12.055895196506551,
+ "grad_norm": 0.021486390382051468,
+ "learning_rate": 0.0006,
+ "loss": 4.922049522399902,
+ "step": 868
+ },
+ {
+ "epoch": 12.069868995633188,
+ "grad_norm": 0.024560654535889626,
+ "learning_rate": 0.0006,
+ "loss": 4.734241008758545,
+ "step": 869
+ },
+ {
+ "epoch": 12.083842794759825,
+ "grad_norm": 0.025870252400636673,
+ "learning_rate": 0.0006,
+ "loss": 4.868566989898682,
+ "step": 870
+ },
+ {
+ "epoch": 12.097816593886463,
+ "grad_norm": 0.0277389045804739,
+ "learning_rate": 0.0006,
+ "loss": 4.755006790161133,
+ "step": 871
+ },
+ {
+ "epoch": 12.1117903930131,
+ "grad_norm": 0.029100844636559486,
+ "learning_rate": 0.0006,
+ "loss": 4.909667015075684,
+ "step": 872
+ },
+ {
+ "epoch": 12.125764192139737,
+ "grad_norm": 0.031833067536354065,
+ "learning_rate": 0.0006,
+ "loss": 4.858674049377441,
+ "step": 873
+ },
+ {
+ "epoch": 12.139737991266376,
+ "grad_norm": 0.032709237188100815,
+ "learning_rate": 0.0006,
+ "loss": 4.823696136474609,
+ "step": 874
+ },
+ {
+ "epoch": 12.153711790393013,
+ "grad_norm": 0.02995380386710167,
+ "learning_rate": 0.0006,
+ "loss": 4.850235462188721,
+ "step": 875
+ },
+ {
+ "epoch": 12.167685589519651,
+ "grad_norm": 0.034098610281944275,
+ "learning_rate": 0.0006,
+ "loss": 4.8169989585876465,
+ "step": 876
+ },
+ {
+ "epoch": 12.181659388646288,
+ "grad_norm": 0.03377080336213112,
+ "learning_rate": 0.0006,
+ "loss": 4.90738582611084,
+ "step": 877
+ },
+ {
+ "epoch": 12.195633187772925,
+ "grad_norm": 0.029780825600028038,
+ "learning_rate": 0.0006,
+ "loss": 4.811163902282715,
+ "step": 878
+ },
+ {
+ "epoch": 12.209606986899564,
+ "grad_norm": 0.03118024580180645,
+ "learning_rate": 0.0006,
+ "loss": 4.861286640167236,
+ "step": 879
+ },
+ {
+ "epoch": 12.2235807860262,
+ "grad_norm": 0.033052004873752594,
+ "learning_rate": 0.0006,
+ "loss": 4.865021228790283,
+ "step": 880
+ },
+ {
+ "epoch": 12.237554585152838,
+ "grad_norm": 0.029070032760500908,
+ "learning_rate": 0.0006,
+ "loss": 4.947303771972656,
+ "step": 881
+ },
+ {
+ "epoch": 12.251528384279476,
+ "grad_norm": 0.028355715796351433,
+ "learning_rate": 0.0006,
+ "loss": 4.804648399353027,
+ "step": 882
+ },
+ {
+ "epoch": 12.265502183406113,
+ "grad_norm": 0.031257227063179016,
+ "learning_rate": 0.0006,
+ "loss": 4.757718086242676,
+ "step": 883
+ },
+ {
+ "epoch": 12.279475982532752,
+ "grad_norm": 0.031438373029232025,
+ "learning_rate": 0.0006,
+ "loss": 4.851268768310547,
+ "step": 884
+ },
+ {
+ "epoch": 12.293449781659389,
+ "grad_norm": 0.030492138117551804,
+ "learning_rate": 0.0006,
+ "loss": 4.7891035079956055,
+ "step": 885
+ },
+ {
+ "epoch": 12.307423580786025,
+ "grad_norm": 0.026115991175174713,
+ "learning_rate": 0.0006,
+ "loss": 4.86474609375,
+ "step": 886
+ },
+ {
+ "epoch": 12.321397379912664,
+ "grad_norm": 0.024184465408325195,
+ "learning_rate": 0.0006,
+ "loss": 4.860456943511963,
+ "step": 887
+ },
+ {
+ "epoch": 12.335371179039301,
+ "grad_norm": 0.025437982752919197,
+ "learning_rate": 0.0006,
+ "loss": 4.806251049041748,
+ "step": 888
+ },
+ {
+ "epoch": 12.34934497816594,
+ "grad_norm": 0.022390231490135193,
+ "learning_rate": 0.0006,
+ "loss": 4.807086944580078,
+ "step": 889
+ },
+ {
+ "epoch": 12.363318777292577,
+ "grad_norm": 0.020174294710159302,
+ "learning_rate": 0.0006,
+ "loss": 4.78292179107666,
+ "step": 890
+ },
+ {
+ "epoch": 12.377292576419213,
+ "grad_norm": 0.019734324887394905,
+ "learning_rate": 0.0006,
+ "loss": 4.784509658813477,
+ "step": 891
+ },
+ {
+ "epoch": 12.391266375545852,
+ "grad_norm": 0.016733523458242416,
+ "learning_rate": 0.0006,
+ "loss": 4.811755180358887,
+ "step": 892
+ },
+ {
+ "epoch": 12.405240174672489,
+ "grad_norm": 0.01570146717131138,
+ "learning_rate": 0.0006,
+ "loss": 4.720088005065918,
+ "step": 893
+ },
+ {
+ "epoch": 12.419213973799126,
+ "grad_norm": 0.01601393334567547,
+ "learning_rate": 0.0006,
+ "loss": 4.783543109893799,
+ "step": 894
+ },
+ {
+ "epoch": 12.433187772925764,
+ "grad_norm": 0.014035487547516823,
+ "learning_rate": 0.0006,
+ "loss": 4.848340034484863,
+ "step": 895
+ },
+ {
+ "epoch": 12.447161572052401,
+ "grad_norm": 0.015248487703502178,
+ "learning_rate": 0.0006,
+ "loss": 4.773287773132324,
+ "step": 896
+ },
+ {
+ "epoch": 12.46113537117904,
+ "grad_norm": 0.01532268151640892,
+ "learning_rate": 0.0006,
+ "loss": 4.799646854400635,
+ "step": 897
+ },
+ {
+ "epoch": 12.475109170305677,
+ "grad_norm": 0.013982534408569336,
+ "learning_rate": 0.0006,
+ "loss": 4.8090500831604,
+ "step": 898
+ },
+ {
+ "epoch": 12.489082969432314,
+ "grad_norm": 0.014081164263188839,
+ "learning_rate": 0.0006,
+ "loss": 4.751888751983643,
+ "step": 899
+ },
+ {
+ "epoch": 12.503056768558952,
+ "grad_norm": 0.01380133256316185,
+ "learning_rate": 0.0006,
+ "loss": 4.78144645690918,
+ "step": 900
+ },
+ {
+ "epoch": 12.51703056768559,
+ "grad_norm": 0.013555269688367844,
+ "learning_rate": 0.0006,
+ "loss": 4.779257774353027,
+ "step": 901
+ },
+ {
+ "epoch": 12.531004366812226,
+ "grad_norm": 0.013075701892375946,
+ "learning_rate": 0.0006,
+ "loss": 4.820705413818359,
+ "step": 902
+ },
+ {
+ "epoch": 12.544978165938865,
+ "grad_norm": 0.01276316400617361,
+ "learning_rate": 0.0006,
+ "loss": 4.836733818054199,
+ "step": 903
+ },
+ {
+ "epoch": 12.558951965065502,
+ "grad_norm": 0.013033892959356308,
+ "learning_rate": 0.0006,
+ "loss": 4.764178276062012,
+ "step": 904
+ },
+ {
+ "epoch": 12.57292576419214,
+ "grad_norm": 0.013996980153024197,
+ "learning_rate": 0.0006,
+ "loss": 4.767983436584473,
+ "step": 905
+ },
+ {
+ "epoch": 12.586899563318777,
+ "grad_norm": 0.015222184360027313,
+ "learning_rate": 0.0006,
+ "loss": 4.759721755981445,
+ "step": 906
+ },
+ {
+ "epoch": 12.600873362445414,
+ "grad_norm": 0.01623941957950592,
+ "learning_rate": 0.0006,
+ "loss": 4.849811553955078,
+ "step": 907
+ },
+ {
+ "epoch": 12.614847161572053,
+ "grad_norm": 0.015711182728409767,
+ "learning_rate": 0.0006,
+ "loss": 4.7829084396362305,
+ "step": 908
+ },
+ {
+ "epoch": 12.62882096069869,
+ "grad_norm": 0.014137467369437218,
+ "learning_rate": 0.0006,
+ "loss": 4.73321533203125,
+ "step": 909
+ },
+ {
+ "epoch": 12.642794759825328,
+ "grad_norm": 0.012738462537527084,
+ "learning_rate": 0.0006,
+ "loss": 4.870052337646484,
+ "step": 910
+ },
+ {
+ "epoch": 12.656768558951965,
+ "grad_norm": 0.013683130033314228,
+ "learning_rate": 0.0006,
+ "loss": 4.82658052444458,
+ "step": 911
+ },
+ {
+ "epoch": 12.670742358078602,
+ "grad_norm": 0.014409830793738365,
+ "learning_rate": 0.0006,
+ "loss": 4.707494735717773,
+ "step": 912
+ },
+ {
+ "epoch": 12.68471615720524,
+ "grad_norm": 0.013806809671223164,
+ "learning_rate": 0.0006,
+ "loss": 4.778928279876709,
+ "step": 913
+ },
+ {
+ "epoch": 12.698689956331878,
+ "grad_norm": 0.012857912108302116,
+ "learning_rate": 0.0006,
+ "loss": 4.890717506408691,
+ "step": 914
+ },
+ {
+ "epoch": 12.712663755458514,
+ "grad_norm": 0.013828758150339127,
+ "learning_rate": 0.0006,
+ "loss": 4.8357439041137695,
+ "step": 915
+ },
+ {
+ "epoch": 12.726637554585153,
+ "grad_norm": 0.01478351280093193,
+ "learning_rate": 0.0006,
+ "loss": 4.869986534118652,
+ "step": 916
+ },
+ {
+ "epoch": 12.74061135371179,
+ "grad_norm": 0.01326421182602644,
+ "learning_rate": 0.0006,
+ "loss": 4.830646514892578,
+ "step": 917
+ },
+ {
+ "epoch": 12.754585152838429,
+ "grad_norm": 0.013335433788597584,
+ "learning_rate": 0.0006,
+ "loss": 4.7786149978637695,
+ "step": 918
+ },
+ {
+ "epoch": 12.768558951965066,
+ "grad_norm": 0.014261147007346153,
+ "learning_rate": 0.0006,
+ "loss": 4.788693428039551,
+ "step": 919
+ },
+ {
+ "epoch": 12.782532751091702,
+ "grad_norm": 0.014274783432483673,
+ "learning_rate": 0.0006,
+ "loss": 4.77295446395874,
+ "step": 920
+ },
+ {
+ "epoch": 12.796506550218341,
+ "grad_norm": 0.015074674971401691,
+ "learning_rate": 0.0006,
+ "loss": 4.830754280090332,
+ "step": 921
+ },
+ {
+ "epoch": 12.810480349344978,
+ "grad_norm": 0.015599515289068222,
+ "learning_rate": 0.0006,
+ "loss": 4.837759017944336,
+ "step": 922
+ },
+ {
+ "epoch": 12.824454148471617,
+ "grad_norm": 0.014595242217183113,
+ "learning_rate": 0.0006,
+ "loss": 4.882965564727783,
+ "step": 923
+ },
+ {
+ "epoch": 12.838427947598253,
+ "grad_norm": 0.014951467514038086,
+ "learning_rate": 0.0006,
+ "loss": 4.709748268127441,
+ "step": 924
+ },
+ {
+ "epoch": 12.85240174672489,
+ "grad_norm": 0.01539295818656683,
+ "learning_rate": 0.0006,
+ "loss": 4.791160583496094,
+ "step": 925
+ },
+ {
+ "epoch": 12.866375545851529,
+ "grad_norm": 0.015987906605005264,
+ "learning_rate": 0.0006,
+ "loss": 4.661876678466797,
+ "step": 926
+ },
+ {
+ "epoch": 12.880349344978166,
+ "grad_norm": 0.016527341678738594,
+ "learning_rate": 0.0006,
+ "loss": 4.721952438354492,
+ "step": 927
+ },
+ {
+ "epoch": 12.894323144104803,
+ "grad_norm": 0.01637161523103714,
+ "learning_rate": 0.0006,
+ "loss": 4.704446792602539,
+ "step": 928
+ },
+ {
+ "epoch": 12.908296943231441,
+ "grad_norm": 0.015221050009131432,
+ "learning_rate": 0.0006,
+ "loss": 4.771907806396484,
+ "step": 929
+ },
+ {
+ "epoch": 12.922270742358078,
+ "grad_norm": 0.016242166981101036,
+ "learning_rate": 0.0006,
+ "loss": 4.806267738342285,
+ "step": 930
+ },
+ {
+ "epoch": 12.936244541484717,
+ "grad_norm": 0.01628357358276844,
+ "learning_rate": 0.0006,
+ "loss": 4.7252326011657715,
+ "step": 931
+ },
+ {
+ "epoch": 12.950218340611354,
+ "grad_norm": 0.017428133636713028,
+ "learning_rate": 0.0006,
+ "loss": 4.765689849853516,
+ "step": 932
+ },
+ {
+ "epoch": 12.96419213973799,
+ "grad_norm": 0.01870061084628105,
+ "learning_rate": 0.0006,
+ "loss": 4.814488887786865,
+ "step": 933
+ },
+ {
+ "epoch": 12.97816593886463,
+ "grad_norm": 0.019796060398221016,
+ "learning_rate": 0.0006,
+ "loss": 4.739185810089111,
+ "step": 934
+ },
+ {
+ "epoch": 12.992139737991266,
+ "grad_norm": 0.01985820196568966,
+ "learning_rate": 0.0006,
+ "loss": 4.810608863830566,
+ "step": 935
+ },
+ {
+ "epoch": 13.0,
+ "grad_norm": 0.020834477618336678,
+ "learning_rate": 0.0006,
+ "loss": 4.727858066558838,
+ "step": 936
+ },
+ {
+ "epoch": 13.0,
+ "eval_loss": 4.979116916656494,
+ "eval_runtime": 57.2877,
+ "eval_samples_per_second": 42.627,
+ "eval_steps_per_second": 1.344,
+ "step": 936
+ },
+ {
+ "epoch": 13.013973799126637,
+ "grad_norm": 0.02025519125163555,
+ "learning_rate": 0.0006,
+ "loss": 4.628875732421875,
+ "step": 937
+ },
+ {
+ "epoch": 13.027947598253276,
+ "grad_norm": 0.019596368074417114,
+ "learning_rate": 0.0006,
+ "loss": 4.694699287414551,
+ "step": 938
+ },
+ {
+ "epoch": 13.041921397379912,
+ "grad_norm": 0.020494934171438217,
+ "learning_rate": 0.0006,
+ "loss": 4.691821098327637,
+ "step": 939
+ },
+ {
+ "epoch": 13.055895196506551,
+ "grad_norm": 0.020121760666370392,
+ "learning_rate": 0.0006,
+ "loss": 4.71478271484375,
+ "step": 940
+ },
+ {
+ "epoch": 13.069868995633188,
+ "grad_norm": 0.021091405302286148,
+ "learning_rate": 0.0006,
+ "loss": 4.695915222167969,
+ "step": 941
+ },
+ {
+ "epoch": 13.083842794759825,
+ "grad_norm": 0.019787423312664032,
+ "learning_rate": 0.0006,
+ "loss": 4.8652448654174805,
+ "step": 942
+ },
+ {
+ "epoch": 13.097816593886463,
+ "grad_norm": 0.0185833927243948,
+ "learning_rate": 0.0006,
+ "loss": 4.7065887451171875,
+ "step": 943
+ },
+ {
+ "epoch": 13.1117903930131,
+ "grad_norm": 0.021220475435256958,
+ "learning_rate": 0.0006,
+ "loss": 4.741232395172119,
+ "step": 944
+ },
+ {
+ "epoch": 13.125764192139737,
+ "grad_norm": 0.02136622555553913,
+ "learning_rate": 0.0006,
+ "loss": 4.705967903137207,
+ "step": 945
+ },
+ {
+ "epoch": 13.139737991266376,
+ "grad_norm": 0.0191587433218956,
+ "learning_rate": 0.0006,
+ "loss": 4.758667945861816,
+ "step": 946
+ },
+ {
+ "epoch": 13.153711790393013,
+ "grad_norm": 0.018290022388100624,
+ "learning_rate": 0.0006,
+ "loss": 4.760666847229004,
+ "step": 947
+ },
+ {
+ "epoch": 13.167685589519651,
+ "grad_norm": 0.021704409271478653,
+ "learning_rate": 0.0006,
+ "loss": 4.750555992126465,
+ "step": 948
+ },
+ {
+ "epoch": 13.181659388646288,
+ "grad_norm": 0.021668918430805206,
+ "learning_rate": 0.0006,
+ "loss": 4.840903282165527,
+ "step": 949
+ },
+ {
+ "epoch": 13.195633187772925,
+ "grad_norm": 0.019497141242027283,
+ "learning_rate": 0.0006,
+ "loss": 4.738969326019287,
+ "step": 950
+ },
+ {
+ "epoch": 13.209606986899564,
+ "grad_norm": 0.018500596284866333,
+ "learning_rate": 0.0006,
+ "loss": 4.75154972076416,
+ "step": 951
+ },
+ {
+ "epoch": 13.2235807860262,
+ "grad_norm": 0.01872239261865616,
+ "learning_rate": 0.0006,
+ "loss": 4.811738014221191,
+ "step": 952
+ },
+ {
+ "epoch": 13.237554585152838,
+ "grad_norm": 0.020321547985076904,
+ "learning_rate": 0.0006,
+ "loss": 4.815384387969971,
+ "step": 953
+ },
+ {
+ "epoch": 13.251528384279476,
+ "grad_norm": 0.02283575013279915,
+ "learning_rate": 0.0006,
+ "loss": 4.74582052230835,
+ "step": 954
+ },
+ {
+ "epoch": 13.265502183406113,
+ "grad_norm": 0.02474398910999298,
+ "learning_rate": 0.0006,
+ "loss": 4.738375186920166,
+ "step": 955
+ },
+ {
+ "epoch": 13.279475982532752,
+ "grad_norm": 0.02065764181315899,
+ "learning_rate": 0.0006,
+ "loss": 4.746875286102295,
+ "step": 956
+ },
+ {
+ "epoch": 13.293449781659389,
+ "grad_norm": 0.020534520968794823,
+ "learning_rate": 0.0006,
+ "loss": 4.656914710998535,
+ "step": 957
+ },
+ {
+ "epoch": 13.307423580786025,
+ "grad_norm": 0.0206378772854805,
+ "learning_rate": 0.0006,
+ "loss": 4.842615604400635,
+ "step": 958
+ },
+ {
+ "epoch": 13.321397379912664,
+ "grad_norm": 0.017935654148459435,
+ "learning_rate": 0.0006,
+ "loss": 4.717827796936035,
+ "step": 959
+ },
+ {
+ "epoch": 13.335371179039301,
+ "grad_norm": 0.016958363354206085,
+ "learning_rate": 0.0006,
+ "loss": 4.681070327758789,
+ "step": 960
+ },
+ {
+ "epoch": 13.34934497816594,
+ "grad_norm": 0.015735412016510963,
+ "learning_rate": 0.0006,
+ "loss": 4.7957048416137695,
+ "step": 961
+ },
+ {
+ "epoch": 13.363318777292577,
+ "grad_norm": 0.01563699170947075,
+ "learning_rate": 0.0006,
+ "loss": 4.701192378997803,
+ "step": 962
+ },
+ {
+ "epoch": 13.377292576419213,
+ "grad_norm": 0.015207760035991669,
+ "learning_rate": 0.0006,
+ "loss": 4.73256254196167,
+ "step": 963
+ },
+ {
+ "epoch": 13.391266375545852,
+ "grad_norm": 0.015396883711218834,
+ "learning_rate": 0.0006,
+ "loss": 4.76412296295166,
+ "step": 964
+ },
+ {
+ "epoch": 13.405240174672489,
+ "grad_norm": 0.015690583735704422,
+ "learning_rate": 0.0006,
+ "loss": 4.791157245635986,
+ "step": 965
+ },
+ {
+ "epoch": 13.419213973799126,
+ "grad_norm": 0.01755724661052227,
+ "learning_rate": 0.0006,
+ "loss": 4.845666885375977,
+ "step": 966
+ },
+ {
+ "epoch": 13.433187772925764,
+ "grad_norm": 0.019443422555923462,
+ "learning_rate": 0.0006,
+ "loss": 4.5902299880981445,
+ "step": 967
+ },
+ {
+ "epoch": 13.447161572052401,
+ "grad_norm": 0.02118338644504547,
+ "learning_rate": 0.0006,
+ "loss": 4.740627288818359,
+ "step": 968
+ },
+ {
+ "epoch": 13.46113537117904,
+ "grad_norm": 0.02174839936196804,
+ "learning_rate": 0.0006,
+ "loss": 4.814560890197754,
+ "step": 969
+ },
+ {
+ "epoch": 13.475109170305677,
+ "grad_norm": 0.022860374301671982,
+ "learning_rate": 0.0006,
+ "loss": 4.66440486907959,
+ "step": 970
+ },
+ {
+ "epoch": 13.489082969432314,
+ "grad_norm": 0.024742677807807922,
+ "learning_rate": 0.0006,
+ "loss": 4.7451653480529785,
+ "step": 971
+ },
+ {
+ "epoch": 13.503056768558952,
+ "grad_norm": 0.027068907395005226,
+ "learning_rate": 0.0006,
+ "loss": 4.802584171295166,
+ "step": 972
+ },
+ {
+ "epoch": 13.51703056768559,
+ "grad_norm": 0.026316625997424126,
+ "learning_rate": 0.0006,
+ "loss": 4.842397689819336,
+ "step": 973
+ },
+ {
+ "epoch": 13.531004366812226,
+ "grad_norm": 0.024610962718725204,
+ "learning_rate": 0.0006,
+ "loss": 4.650721073150635,
+ "step": 974
+ },
+ {
+ "epoch": 13.544978165938865,
+ "grad_norm": 0.024122603237628937,
+ "learning_rate": 0.0006,
+ "loss": 4.6984357833862305,
+ "step": 975
+ },
+ {
+ "epoch": 13.558951965065502,
+ "grad_norm": 0.022699708119034767,
+ "learning_rate": 0.0006,
+ "loss": 4.793007850646973,
+ "step": 976
+ },
+ {
+ "epoch": 13.57292576419214,
+ "grad_norm": 0.021706534549593925,
+ "learning_rate": 0.0006,
+ "loss": 4.691362380981445,
+ "step": 977
+ },
+ {
+ "epoch": 13.586899563318777,
+ "grad_norm": 0.021363291889429092,
+ "learning_rate": 0.0006,
+ "loss": 4.76762580871582,
+ "step": 978
+ },
+ {
+ "epoch": 13.600873362445414,
+ "grad_norm": 0.023476677015423775,
+ "learning_rate": 0.0006,
+ "loss": 4.777202606201172,
+ "step": 979
+ },
+ {
+ "epoch": 13.614847161572053,
+ "grad_norm": 0.02457100711762905,
+ "learning_rate": 0.0006,
+ "loss": 4.590703010559082,
+ "step": 980
+ },
+ {
+ "epoch": 13.62882096069869,
+ "grad_norm": 0.02469099499285221,
+ "learning_rate": 0.0006,
+ "loss": 4.640185356140137,
+ "step": 981
+ },
+ {
+ "epoch": 13.642794759825328,
+ "grad_norm": 0.024914603680372238,
+ "learning_rate": 0.0006,
+ "loss": 4.772201061248779,
+ "step": 982
+ },
+ {
+ "epoch": 13.656768558951965,
+ "grad_norm": 0.025133173912763596,
+ "learning_rate": 0.0006,
+ "loss": 4.743337154388428,
+ "step": 983
+ },
+ {
+ "epoch": 13.670742358078602,
+ "grad_norm": 0.02034137211740017,
+ "learning_rate": 0.0006,
+ "loss": 4.752127647399902,
+ "step": 984
+ },
+ {
+ "epoch": 13.68471615720524,
+ "grad_norm": 0.019645661115646362,
+ "learning_rate": 0.0006,
+ "loss": 4.838453769683838,
+ "step": 985
+ },
+ {
+ "epoch": 13.698689956331878,
+ "grad_norm": 0.01900297962129116,
+ "learning_rate": 0.0006,
+ "loss": 4.804868221282959,
+ "step": 986
+ },
+ {
+ "epoch": 13.712663755458514,
+ "grad_norm": 0.020292341709136963,
+ "learning_rate": 0.0006,
+ "loss": 4.609743118286133,
+ "step": 987
+ },
+ {
+ "epoch": 13.726637554585153,
+ "grad_norm": 0.017014775425195694,
+ "learning_rate": 0.0006,
+ "loss": 4.740612983703613,
+ "step": 988
+ },
+ {
+ "epoch": 13.74061135371179,
+ "grad_norm": 0.015199844725430012,
+ "learning_rate": 0.0006,
+ "loss": 4.719701766967773,
+ "step": 989
+ },
+ {
+ "epoch": 13.754585152838429,
+ "grad_norm": 0.015125414356589317,
+ "learning_rate": 0.0006,
+ "loss": 4.686114311218262,
+ "step": 990
+ },
+ {
+ "epoch": 13.768558951965066,
+ "grad_norm": 0.014779540710151196,
+ "learning_rate": 0.0006,
+ "loss": 4.674347877502441,
+ "step": 991
+ },
+ {
+ "epoch": 13.782532751091702,
+ "grad_norm": 0.015255163423717022,
+ "learning_rate": 0.0006,
+ "loss": 4.6309309005737305,
+ "step": 992
+ },
+ {
+ "epoch": 13.796506550218341,
+ "grad_norm": 0.013692197389900684,
+ "learning_rate": 0.0006,
+ "loss": 4.698497295379639,
+ "step": 993
+ },
+ {
+ "epoch": 13.810480349344978,
+ "grad_norm": 0.014529623091220856,
+ "learning_rate": 0.0006,
+ "loss": 4.628443717956543,
+ "step": 994
+ },
+ {
+ "epoch": 13.824454148471617,
+ "grad_norm": 0.01576659083366394,
+ "learning_rate": 0.0006,
+ "loss": 4.788015842437744,
+ "step": 995
+ },
+ {
+ "epoch": 13.838427947598253,
+ "grad_norm": 0.015834277495741844,
+ "learning_rate": 0.0006,
+ "loss": 4.672703742980957,
+ "step": 996
+ },
+ {
+ "epoch": 13.85240174672489,
+ "grad_norm": 0.015539503656327724,
+ "learning_rate": 0.0006,
+ "loss": 4.731503486633301,
+ "step": 997
+ },
+ {
+ "epoch": 13.866375545851529,
+ "grad_norm": 0.013463746756315231,
+ "learning_rate": 0.0006,
+ "loss": 4.7006330490112305,
+ "step": 998
+ },
+ {
+ "epoch": 13.880349344978166,
+ "grad_norm": 0.016377726569771767,
+ "learning_rate": 0.0006,
+ "loss": 4.636074066162109,
+ "step": 999
+ },
+ {
+ "epoch": 13.894323144104803,
+ "grad_norm": 0.016660034656524658,
+ "learning_rate": 0.0006,
+ "loss": 4.742693901062012,
+ "step": 1000
+ },
+ {
+ "epoch": 13.908296943231441,
+ "grad_norm": 0.014902938157320023,
+ "learning_rate": 0.0006,
+ "loss": 4.769811630249023,
+ "step": 1001
+ },
+ {
+ "epoch": 13.922270742358078,
+ "grad_norm": 0.014883476309478283,
+ "learning_rate": 0.0006,
+ "loss": 4.731159210205078,
+ "step": 1002
+ },
+ {
+ "epoch": 13.936244541484717,
+ "grad_norm": 0.014494217932224274,
+ "learning_rate": 0.0006,
+ "loss": 4.659485816955566,
+ "step": 1003
+ },
+ {
+ "epoch": 13.950218340611354,
+ "grad_norm": 0.015426991507411003,
+ "learning_rate": 0.0006,
+ "loss": 4.755229949951172,
+ "step": 1004
+ },
+ {
+ "epoch": 13.96419213973799,
+ "grad_norm": 0.016648586839437485,
+ "learning_rate": 0.0006,
+ "loss": 4.74111270904541,
+ "step": 1005
+ },
+ {
+ "epoch": 13.97816593886463,
+ "grad_norm": 0.016980541869997978,
+ "learning_rate": 0.0006,
+ "loss": 4.74072265625,
+ "step": 1006
+ },
+ {
+ "epoch": 13.992139737991266,
+ "grad_norm": 0.018070152029395103,
+ "learning_rate": 0.0006,
+ "loss": 4.732975006103516,
+ "step": 1007
+ },
+ {
+ "epoch": 14.0,
+ "grad_norm": 0.019910665228962898,
+ "learning_rate": 0.0006,
+ "loss": 4.634697914123535,
+ "step": 1008
+ },
+ {
+ "epoch": 14.0,
+ "eval_loss": 4.850640296936035,
+ "eval_runtime": 57.3039,
+ "eval_samples_per_second": 42.615,
+ "eval_steps_per_second": 1.344,
+ "step": 1008
+ },
+ {
+ "epoch": 14.013973799126637,
+ "grad_norm": 0.01980053074657917,
+ "learning_rate": 0.0006,
+ "loss": 4.628744602203369,
+ "step": 1009
+ },
+ {
+ "epoch": 14.027947598253276,
+ "grad_norm": 0.02497623674571514,
+ "learning_rate": 0.0006,
+ "loss": 4.75930118560791,
+ "step": 1010
+ },
+ {
+ "epoch": 14.041921397379912,
+ "grad_norm": 0.02794097363948822,
+ "learning_rate": 0.0006,
+ "loss": 4.691340446472168,
+ "step": 1011
+ },
+ {
+ "epoch": 14.055895196506551,
+ "grad_norm": 0.023725535720586777,
+ "learning_rate": 0.0006,
+ "loss": 4.733664512634277,
+ "step": 1012
+ },
+ {
+ "epoch": 14.069868995633188,
+ "grad_norm": 0.022966474294662476,
+ "learning_rate": 0.0006,
+ "loss": 4.664808750152588,
+ "step": 1013
+ },
+ {
+ "epoch": 14.083842794759825,
+ "grad_norm": 0.02530563250184059,
+ "learning_rate": 0.0006,
+ "loss": 4.8047285079956055,
+ "step": 1014
+ },
+ {
+ "epoch": 14.097816593886463,
+ "grad_norm": 0.02491082064807415,
+ "learning_rate": 0.0006,
+ "loss": 4.667956352233887,
+ "step": 1015
+ },
+ {
+ "epoch": 14.1117903930131,
+ "grad_norm": 0.022110294550657272,
+ "learning_rate": 0.0006,
+ "loss": 4.666140556335449,
+ "step": 1016
+ },
+ {
+ "epoch": 14.125764192139737,
+ "grad_norm": 0.02065849117934704,
+ "learning_rate": 0.0006,
+ "loss": 4.581960678100586,
+ "step": 1017
+ },
+ {
+ "epoch": 14.139737991266376,
+ "grad_norm": 0.01941298507153988,
+ "learning_rate": 0.0006,
+ "loss": 4.614226341247559,
+ "step": 1018
+ },
+ {
+ "epoch": 14.153711790393013,
+ "grad_norm": 0.022441856563091278,
+ "learning_rate": 0.0006,
+ "loss": 4.638352394104004,
+ "step": 1019
+ },
+ {
+ "epoch": 14.167685589519651,
+ "grad_norm": 0.021519694477319717,
+ "learning_rate": 0.0006,
+ "loss": 4.700160980224609,
+ "step": 1020
+ },
+ {
+ "epoch": 14.181659388646288,
+ "grad_norm": 0.018956486135721207,
+ "learning_rate": 0.0006,
+ "loss": 4.649593353271484,
+ "step": 1021
+ },
+ {
+ "epoch": 14.195633187772925,
+ "grad_norm": 0.016681140288710594,
+ "learning_rate": 0.0006,
+ "loss": 4.603952407836914,
+ "step": 1022
+ },
+ {
+ "epoch": 14.209606986899564,
+ "grad_norm": 0.01788809522986412,
+ "learning_rate": 0.0006,
+ "loss": 4.669160842895508,
+ "step": 1023
+ },
+ {
+ "epoch": 14.2235807860262,
+ "grad_norm": 0.016047228127717972,
+ "learning_rate": 0.0006,
+ "loss": 4.6367950439453125,
+ "step": 1024
+ },
+ {
+ "epoch": 14.237554585152838,
+ "grad_norm": 0.01614239625632763,
+ "learning_rate": 0.0006,
+ "loss": 4.664529323577881,
+ "step": 1025
+ },
+ {
+ "epoch": 14.251528384279476,
+ "grad_norm": 0.016362549737095833,
+ "learning_rate": 0.0006,
+ "loss": 4.702373504638672,
+ "step": 1026
+ },
+ {
+ "epoch": 14.265502183406113,
+ "grad_norm": 0.015800409018993378,
+ "learning_rate": 0.0006,
+ "loss": 4.7594428062438965,
+ "step": 1027
+ },
+ {
+ "epoch": 14.279475982532752,
+ "grad_norm": 0.016730139032006264,
+ "learning_rate": 0.0006,
+ "loss": 4.677244186401367,
+ "step": 1028
+ },
+ {
+ "epoch": 14.293449781659389,
+ "grad_norm": 0.01661759801208973,
+ "learning_rate": 0.0006,
+ "loss": 4.633679389953613,
+ "step": 1029
+ },
+ {
+ "epoch": 14.307423580786025,
+ "grad_norm": 0.017673347145318985,
+ "learning_rate": 0.0006,
+ "loss": 4.649107456207275,
+ "step": 1030
+ },
+ {
+ "epoch": 14.321397379912664,
+ "grad_norm": 0.019243352115154266,
+ "learning_rate": 0.0006,
+ "loss": 4.611257553100586,
+ "step": 1031
+ },
+ {
+ "epoch": 14.335371179039301,
+ "grad_norm": 0.019817164167761803,
+ "learning_rate": 0.0006,
+ "loss": 4.588086128234863,
+ "step": 1032
+ },
+ {
+ "epoch": 14.34934497816594,
+ "grad_norm": 0.022292211651802063,
+ "learning_rate": 0.0006,
+ "loss": 4.783066749572754,
+ "step": 1033
+ },
+ {
+ "epoch": 14.363318777292577,
+ "grad_norm": 0.02140096202492714,
+ "learning_rate": 0.0006,
+ "loss": 4.604718208312988,
+ "step": 1034
+ },
+ {
+ "epoch": 14.377292576419213,
+ "grad_norm": 0.0187817495316267,
+ "learning_rate": 0.0006,
+ "loss": 4.721773147583008,
+ "step": 1035
+ },
+ {
+ "epoch": 14.391266375545852,
+ "grad_norm": 0.019879210740327835,
+ "learning_rate": 0.0006,
+ "loss": 4.789286136627197,
+ "step": 1036
+ },
+ {
+ "epoch": 14.405240174672489,
+ "grad_norm": 0.01885785162448883,
+ "learning_rate": 0.0006,
+ "loss": 4.669743537902832,
+ "step": 1037
+ },
+ {
+ "epoch": 14.419213973799126,
+ "grad_norm": 0.018324270844459534,
+ "learning_rate": 0.0006,
+ "loss": 4.65901517868042,
+ "step": 1038
+ },
+ {
+ "epoch": 14.433187772925764,
+ "grad_norm": 0.017664821818470955,
+ "learning_rate": 0.0006,
+ "loss": 4.667839527130127,
+ "step": 1039
+ },
+ {
+ "epoch": 14.447161572052401,
+ "grad_norm": 0.015702905133366585,
+ "learning_rate": 0.0006,
+ "loss": 4.64700984954834,
+ "step": 1040
+ },
+ {
+ "epoch": 14.46113537117904,
+ "grad_norm": 0.016674194484949112,
+ "learning_rate": 0.0006,
+ "loss": 4.694093227386475,
+ "step": 1041
+ },
+ {
+ "epoch": 14.475109170305677,
+ "grad_norm": 0.016869477927684784,
+ "learning_rate": 0.0006,
+ "loss": 4.619544982910156,
+ "step": 1042
+ },
+ {
+ "epoch": 14.489082969432314,
+ "grad_norm": 0.016557540744543076,
+ "learning_rate": 0.0006,
+ "loss": 4.568667411804199,
+ "step": 1043
+ },
+ {
+ "epoch": 14.503056768558952,
+ "grad_norm": 0.014500945806503296,
+ "learning_rate": 0.0006,
+ "loss": 4.561741828918457,
+ "step": 1044
+ },
+ {
+ "epoch": 14.51703056768559,
+ "grad_norm": 0.014228739775717258,
+ "learning_rate": 0.0006,
+ "loss": 4.654277324676514,
+ "step": 1045
+ },
+ {
+ "epoch": 14.531004366812226,
+ "grad_norm": 0.013894238509237766,
+ "learning_rate": 0.0006,
+ "loss": 4.5867719650268555,
+ "step": 1046
+ },
+ {
+ "epoch": 14.544978165938865,
+ "grad_norm": 0.01550068985670805,
+ "learning_rate": 0.0006,
+ "loss": 4.61168098449707,
+ "step": 1047
+ },
+ {
+ "epoch": 14.558951965065502,
+ "grad_norm": 0.015940772369503975,
+ "learning_rate": 0.0006,
+ "loss": 4.691247940063477,
+ "step": 1048
+ },
+ {
+ "epoch": 14.57292576419214,
+ "grad_norm": 0.01714535616338253,
+ "learning_rate": 0.0006,
+ "loss": 4.61183500289917,
+ "step": 1049
+ },
+ {
+ "epoch": 14.586899563318777,
+ "grad_norm": 0.017925038933753967,
+ "learning_rate": 0.0006,
+ "loss": 4.719022750854492,
+ "step": 1050
+ },
+ {
+ "epoch": 14.600873362445414,
+ "grad_norm": 0.018141552805900574,
+ "learning_rate": 0.0006,
+ "loss": 4.760281562805176,
+ "step": 1051
+ },
+ {
+ "epoch": 14.614847161572053,
+ "grad_norm": 0.017131328582763672,
+ "learning_rate": 0.0006,
+ "loss": 4.709680080413818,
+ "step": 1052
+ },
+ {
+ "epoch": 14.62882096069869,
+ "grad_norm": 0.018730048090219498,
+ "learning_rate": 0.0006,
+ "loss": 4.5994720458984375,
+ "step": 1053
+ },
+ {
+ "epoch": 14.642794759825328,
+ "grad_norm": 0.018894298002123833,
+ "learning_rate": 0.0006,
+ "loss": 4.669070720672607,
+ "step": 1054
+ },
+ {
+ "epoch": 14.656768558951965,
+ "grad_norm": 0.017643198370933533,
+ "learning_rate": 0.0006,
+ "loss": 4.641790390014648,
+ "step": 1055
+ },
+ {
+ "epoch": 14.670742358078602,
+ "grad_norm": 0.017647244036197662,
+ "learning_rate": 0.0006,
+ "loss": 4.653899192810059,
+ "step": 1056
+ },
+ {
+ "epoch": 14.68471615720524,
+ "grad_norm": 0.017304198816418648,
+ "learning_rate": 0.0006,
+ "loss": 4.7487473487854,
+ "step": 1057
+ },
+ {
+ "epoch": 14.698689956331878,
+ "grad_norm": 0.01638447679579258,
+ "learning_rate": 0.0006,
+ "loss": 4.71945858001709,
+ "step": 1058
+ },
+ {
+ "epoch": 14.712663755458514,
+ "grad_norm": 0.015892071649432182,
+ "learning_rate": 0.0006,
+ "loss": 4.744903564453125,
+ "step": 1059
+ },
+ {
+ "epoch": 14.726637554585153,
+ "grad_norm": 0.017420614138245583,
+ "learning_rate": 0.0006,
+ "loss": 4.781563758850098,
+ "step": 1060
+ },
+ {
+ "epoch": 14.74061135371179,
+ "grad_norm": 0.016498520970344543,
+ "learning_rate": 0.0006,
+ "loss": 4.622779846191406,
+ "step": 1061
+ },
+ {
+ "epoch": 14.754585152838429,
+ "grad_norm": 0.017482127994298935,
+ "learning_rate": 0.0006,
+ "loss": 4.62825870513916,
+ "step": 1062
+ },
+ {
+ "epoch": 14.768558951965066,
+ "grad_norm": 0.019018379971385002,
+ "learning_rate": 0.0006,
+ "loss": 4.693194389343262,
+ "step": 1063
+ },
+ {
+ "epoch": 14.782532751091702,
+ "grad_norm": 0.01993914321064949,
+ "learning_rate": 0.0006,
+ "loss": 4.742903709411621,
+ "step": 1064
+ },
+ {
+ "epoch": 14.796506550218341,
+ "grad_norm": 0.02301335707306862,
+ "learning_rate": 0.0006,
+ "loss": 4.613251686096191,
+ "step": 1065
+ },
+ {
+ "epoch": 14.810480349344978,
+ "grad_norm": 0.024088222533464432,
+ "learning_rate": 0.0006,
+ "loss": 4.594305992126465,
+ "step": 1066
+ },
+ {
+ "epoch": 14.824454148471617,
+ "grad_norm": 0.02493320405483246,
+ "learning_rate": 0.0006,
+ "loss": 4.700588226318359,
+ "step": 1067
+ },
+ {
+ "epoch": 14.838427947598253,
+ "grad_norm": 0.02501937560737133,
+ "learning_rate": 0.0006,
+ "loss": 4.601742744445801,
+ "step": 1068
+ },
+ {
+ "epoch": 14.85240174672489,
+ "grad_norm": 0.023490751162171364,
+ "learning_rate": 0.0006,
+ "loss": 4.589211463928223,
+ "step": 1069
+ },
+ {
+ "epoch": 14.866375545851529,
+ "grad_norm": 0.019695665687322617,
+ "learning_rate": 0.0006,
+ "loss": 4.635081768035889,
+ "step": 1070
+ },
+ {
+ "epoch": 14.880349344978166,
+ "grad_norm": 0.019276980310678482,
+ "learning_rate": 0.0006,
+ "loss": 4.737002372741699,
+ "step": 1071
+ },
+ {
+ "epoch": 14.894323144104803,
+ "grad_norm": 0.022430511191487312,
+ "learning_rate": 0.0006,
+ "loss": 4.808295726776123,
+ "step": 1072
+ },
+ {
+ "epoch": 14.908296943231441,
+ "grad_norm": 0.021543916314840317,
+ "learning_rate": 0.0006,
+ "loss": 4.559475421905518,
+ "step": 1073
+ },
+ {
+ "epoch": 14.922270742358078,
+ "grad_norm": 0.019885241985321045,
+ "learning_rate": 0.0006,
+ "loss": 4.647948741912842,
+ "step": 1074
+ },
+ {
+ "epoch": 14.936244541484717,
+ "grad_norm": 0.02115175500512123,
+ "learning_rate": 0.0006,
+ "loss": 4.680505752563477,
+ "step": 1075
+ },
+ {
+ "epoch": 14.950218340611354,
+ "grad_norm": 0.02156151458621025,
+ "learning_rate": 0.0006,
+ "loss": 4.673477649688721,
+ "step": 1076
+ },
+ {
+ "epoch": 14.96419213973799,
+ "grad_norm": 0.02184601128101349,
+ "learning_rate": 0.0006,
+ "loss": 4.644402027130127,
+ "step": 1077
+ },
+ {
+ "epoch": 14.97816593886463,
+ "grad_norm": 0.0211983360350132,
+ "learning_rate": 0.0006,
+ "loss": 4.821558475494385,
+ "step": 1078
+ },
+ {
+ "epoch": 14.992139737991266,
+ "grad_norm": 0.019636567682027817,
+ "learning_rate": 0.0006,
+ "loss": 4.578868865966797,
+ "step": 1079
+ },
+ {
+ "epoch": 15.0,
+ "grad_norm": 0.019997350871562958,
+ "learning_rate": 0.0006,
+ "loss": 4.691307544708252,
+ "step": 1080
+ },
+ {
+ "epoch": 15.0,
+ "eval_loss": 4.872478008270264,
+ "eval_runtime": 56.8513,
+ "eval_samples_per_second": 42.954,
+ "eval_steps_per_second": 1.354,
+ "step": 1080
+ },
+ {
+ "epoch": 15.013973799126637,
+ "grad_norm": 0.018830662593245506,
+ "learning_rate": 0.0006,
+ "loss": 4.761779308319092,
+ "step": 1081
+ },
+ {
+ "epoch": 15.027947598253276,
+ "grad_norm": 0.016876481473445892,
+ "learning_rate": 0.0006,
+ "loss": 4.61148738861084,
+ "step": 1082
+ },
+ {
+ "epoch": 15.041921397379912,
+ "grad_norm": 0.015550863929092884,
+ "learning_rate": 0.0006,
+ "loss": 4.676124095916748,
+ "step": 1083
+ },
+ {
+ "epoch": 15.055895196506551,
+ "grad_norm": 0.017178896814584732,
+ "learning_rate": 0.0006,
+ "loss": 4.588935852050781,
+ "step": 1084
+ },
+ {
+ "epoch": 15.069868995633188,
+ "grad_norm": 0.015465447679162025,
+ "learning_rate": 0.0006,
+ "loss": 4.605499267578125,
+ "step": 1085
+ },
+ {
+ "epoch": 15.083842794759825,
+ "grad_norm": 0.01606195978820324,
+ "learning_rate": 0.0006,
+ "loss": 4.723373889923096,
+ "step": 1086
+ },
+ {
+ "epoch": 15.097816593886463,
+ "grad_norm": 0.016521936282515526,
+ "learning_rate": 0.0006,
+ "loss": 4.646608352661133,
+ "step": 1087
+ },
+ {
+ "epoch": 15.1117903930131,
+ "grad_norm": 0.01589205674827099,
+ "learning_rate": 0.0006,
+ "loss": 4.498138427734375,
+ "step": 1088
+ },
+ {
+ "epoch": 15.125764192139737,
+ "grad_norm": 0.01524856686592102,
+ "learning_rate": 0.0006,
+ "loss": 4.594562530517578,
+ "step": 1089
+ },
+ {
+ "epoch": 15.139737991266376,
+ "grad_norm": 0.014714641496539116,
+ "learning_rate": 0.0006,
+ "loss": 4.630744934082031,
+ "step": 1090
+ },
+ {
+ "epoch": 15.153711790393013,
+ "grad_norm": 0.015333331190049648,
+ "learning_rate": 0.0006,
+ "loss": 4.682388782501221,
+ "step": 1091
+ },
+ {
+ "epoch": 15.167685589519651,
+ "grad_norm": 0.01550073828548193,
+ "learning_rate": 0.0006,
+ "loss": 4.692652702331543,
+ "step": 1092
+ },
+ {
+ "epoch": 15.181659388646288,
+ "grad_norm": 0.014783318154513836,
+ "learning_rate": 0.0006,
+ "loss": 4.6223835945129395,
+ "step": 1093
+ },
+ {
+ "epoch": 15.195633187772925,
+ "grad_norm": 0.014372751116752625,
+ "learning_rate": 0.0006,
+ "loss": 4.52687406539917,
+ "step": 1094
+ },
+ {
+ "epoch": 15.209606986899564,
+ "grad_norm": 0.014779130928218365,
+ "learning_rate": 0.0006,
+ "loss": 4.637009620666504,
+ "step": 1095
+ },
+ {
+ "epoch": 15.2235807860262,
+ "grad_norm": 0.015697073191404343,
+ "learning_rate": 0.0006,
+ "loss": 4.682981014251709,
+ "step": 1096
+ },
+ {
+ "epoch": 15.237554585152838,
+ "grad_norm": 0.016523664817214012,
+ "learning_rate": 0.0006,
+ "loss": 4.705722808837891,
+ "step": 1097
+ },
+ {
+ "epoch": 15.251528384279476,
+ "grad_norm": 0.017633996903896332,
+ "learning_rate": 0.0006,
+ "loss": 4.574686050415039,
+ "step": 1098
+ },
+ {
+ "epoch": 15.265502183406113,
+ "grad_norm": 0.019955076277256012,
+ "learning_rate": 0.0006,
+ "loss": 4.634450912475586,
+ "step": 1099
+ },
+ {
+ "epoch": 15.279475982532752,
+ "grad_norm": 0.02132618986070156,
+ "learning_rate": 0.0006,
+ "loss": 4.56110143661499,
+ "step": 1100
+ },
+ {
+ "epoch": 15.293449781659389,
+ "grad_norm": 0.0207593385130167,
+ "learning_rate": 0.0006,
+ "loss": 4.629396438598633,
+ "step": 1101
+ },
+ {
+ "epoch": 15.307423580786025,
+ "grad_norm": 0.017882540822029114,
+ "learning_rate": 0.0006,
+ "loss": 4.730963230133057,
+ "step": 1102
+ },
+ {
+ "epoch": 15.321397379912664,
+ "grad_norm": 0.01917876861989498,
+ "learning_rate": 0.0006,
+ "loss": 4.64585542678833,
+ "step": 1103
+ },
+ {
+ "epoch": 15.335371179039301,
+ "grad_norm": 0.01943974196910858,
+ "learning_rate": 0.0006,
+ "loss": 4.656181335449219,
+ "step": 1104
+ },
+ {
+ "epoch": 15.34934497816594,
+ "grad_norm": 0.019904715940356255,
+ "learning_rate": 0.0006,
+ "loss": 4.619329452514648,
+ "step": 1105
+ },
+ {
+ "epoch": 15.363318777292577,
+ "grad_norm": 0.019174346700310707,
+ "learning_rate": 0.0006,
+ "loss": 4.647892475128174,
+ "step": 1106
+ },
+ {
+ "epoch": 15.377292576419213,
+ "grad_norm": 0.018244758248329163,
+ "learning_rate": 0.0006,
+ "loss": 4.56166934967041,
+ "step": 1107
+ },
+ {
+ "epoch": 15.391266375545852,
+ "grad_norm": 0.018914205953478813,
+ "learning_rate": 0.0006,
+ "loss": 4.648063659667969,
+ "step": 1108
+ },
+ {
+ "epoch": 15.405240174672489,
+ "grad_norm": 0.01961071416735649,
+ "learning_rate": 0.0006,
+ "loss": 4.665525436401367,
+ "step": 1109
+ },
+ {
+ "epoch": 15.419213973799126,
+ "grad_norm": 0.01841311901807785,
+ "learning_rate": 0.0006,
+ "loss": 4.611342430114746,
+ "step": 1110
+ },
+ {
+ "epoch": 15.433187772925764,
+ "grad_norm": 0.01863427273929119,
+ "learning_rate": 0.0006,
+ "loss": 4.664256572723389,
+ "step": 1111
+ },
+ {
+ "epoch": 15.447161572052401,
+ "grad_norm": 0.020506877452135086,
+ "learning_rate": 0.0006,
+ "loss": 4.631237506866455,
+ "step": 1112
+ },
+ {
+ "epoch": 15.46113537117904,
+ "grad_norm": 0.02223833277821541,
+ "learning_rate": 0.0006,
+ "loss": 4.668654441833496,
+ "step": 1113
+ },
+ {
+ "epoch": 15.475109170305677,
+ "grad_norm": 0.023336702957749367,
+ "learning_rate": 0.0006,
+ "loss": 4.562833309173584,
+ "step": 1114
+ },
+ {
+ "epoch": 15.489082969432314,
+ "grad_norm": 0.023052413016557693,
+ "learning_rate": 0.0006,
+ "loss": 4.621532917022705,
+ "step": 1115
+ },
+ {
+ "epoch": 15.503056768558952,
+ "grad_norm": 0.023420924320816994,
+ "learning_rate": 0.0006,
+ "loss": 4.665465354919434,
+ "step": 1116
+ },
+ {
+ "epoch": 15.51703056768559,
+ "grad_norm": 0.024071281775832176,
+ "learning_rate": 0.0006,
+ "loss": 4.595980644226074,
+ "step": 1117
+ },
+ {
+ "epoch": 15.531004366812226,
+ "grad_norm": 0.023406682536005974,
+ "learning_rate": 0.0006,
+ "loss": 4.642635345458984,
+ "step": 1118
+ },
+ {
+ "epoch": 15.544978165938865,
+ "grad_norm": 0.02138647809624672,
+ "learning_rate": 0.0006,
+ "loss": 4.704513072967529,
+ "step": 1119
+ },
+ {
+ "epoch": 15.558951965065502,
+ "grad_norm": 0.0221384447067976,
+ "learning_rate": 0.0006,
+ "loss": 4.567503452301025,
+ "step": 1120
+ },
+ {
+ "epoch": 15.57292576419214,
+ "grad_norm": 0.020879851654171944,
+ "learning_rate": 0.0006,
+ "loss": 4.664057731628418,
+ "step": 1121
+ },
+ {
+ "epoch": 15.586899563318777,
+ "grad_norm": 0.021478967741131783,
+ "learning_rate": 0.0006,
+ "loss": 4.539627552032471,
+ "step": 1122
+ },
+ {
+ "epoch": 15.600873362445414,
+ "grad_norm": 0.022748656570911407,
+ "learning_rate": 0.0006,
+ "loss": 4.5654497146606445,
+ "step": 1123
+ },
+ {
+ "epoch": 15.614847161572053,
+ "grad_norm": 0.021870248019695282,
+ "learning_rate": 0.0006,
+ "loss": 4.56899356842041,
+ "step": 1124
+ },
+ {
+ "epoch": 15.62882096069869,
+ "grad_norm": 0.021711355075240135,
+ "learning_rate": 0.0006,
+ "loss": 4.657808303833008,
+ "step": 1125
+ },
+ {
+ "epoch": 15.642794759825328,
+ "grad_norm": 0.022810909897089005,
+ "learning_rate": 0.0006,
+ "loss": 4.6526689529418945,
+ "step": 1126
+ },
+ {
+ "epoch": 15.656768558951965,
+ "grad_norm": 0.022561343386769295,
+ "learning_rate": 0.0006,
+ "loss": 4.534887790679932,
+ "step": 1127
+ },
+ {
+ "epoch": 15.670742358078602,
+ "grad_norm": 0.019909802824258804,
+ "learning_rate": 0.0006,
+ "loss": 4.600790023803711,
+ "step": 1128
+ },
+ {
+ "epoch": 15.68471615720524,
+ "grad_norm": 0.020092325285077095,
+ "learning_rate": 0.0006,
+ "loss": 4.773830413818359,
+ "step": 1129
+ },
+ {
+ "epoch": 15.698689956331878,
+ "grad_norm": 0.018834445625543594,
+ "learning_rate": 0.0006,
+ "loss": 4.628868103027344,
+ "step": 1130
+ },
+ {
+ "epoch": 15.712663755458514,
+ "grad_norm": 0.017552798613905907,
+ "learning_rate": 0.0006,
+ "loss": 4.647500038146973,
+ "step": 1131
+ },
+ {
+ "epoch": 15.726637554585153,
+ "grad_norm": 0.01751251146197319,
+ "learning_rate": 0.0006,
+ "loss": 4.590545654296875,
+ "step": 1132
+ },
+ {
+ "epoch": 15.74061135371179,
+ "grad_norm": 0.01699099875986576,
+ "learning_rate": 0.0006,
+ "loss": 4.509699821472168,
+ "step": 1133
+ },
+ {
+ "epoch": 15.754585152838429,
+ "grad_norm": 0.015808649361133575,
+ "learning_rate": 0.0006,
+ "loss": 4.606992244720459,
+ "step": 1134
+ },
+ {
+ "epoch": 15.768558951965066,
+ "grad_norm": 0.01655443385243416,
+ "learning_rate": 0.0006,
+ "loss": 4.585219860076904,
+ "step": 1135
+ },
+ {
+ "epoch": 15.782532751091702,
+ "grad_norm": 0.01708163693547249,
+ "learning_rate": 0.0006,
+ "loss": 4.668147087097168,
+ "step": 1136
+ },
+ {
+ "epoch": 15.796506550218341,
+ "grad_norm": 0.017161913216114044,
+ "learning_rate": 0.0006,
+ "loss": 4.7170634269714355,
+ "step": 1137
+ },
+ {
+ "epoch": 15.810480349344978,
+ "grad_norm": 0.017262674868106842,
+ "learning_rate": 0.0006,
+ "loss": 4.572132110595703,
+ "step": 1138
+ },
+ {
+ "epoch": 15.824454148471617,
+ "grad_norm": 0.017713576555252075,
+ "learning_rate": 0.0006,
+ "loss": 4.702538967132568,
+ "step": 1139
+ },
+ {
+ "epoch": 15.838427947598253,
+ "grad_norm": 0.018730733543634415,
+ "learning_rate": 0.0006,
+ "loss": 4.629721641540527,
+ "step": 1140
+ },
+ {
+ "epoch": 15.85240174672489,
+ "grad_norm": 0.017960187047719955,
+ "learning_rate": 0.0006,
+ "loss": 4.661039352416992,
+ "step": 1141
+ },
+ {
+ "epoch": 15.866375545851529,
+ "grad_norm": 0.01750079356133938,
+ "learning_rate": 0.0006,
+ "loss": 4.503710746765137,
+ "step": 1142
+ },
+ {
+ "epoch": 15.880349344978166,
+ "grad_norm": 0.016630422323942184,
+ "learning_rate": 0.0006,
+ "loss": 4.6582136154174805,
+ "step": 1143
+ },
+ {
+ "epoch": 15.894323144104803,
+ "grad_norm": 0.015896832570433617,
+ "learning_rate": 0.0006,
+ "loss": 4.596693992614746,
+ "step": 1144
+ },
+ {
+ "epoch": 15.908296943231441,
+ "grad_norm": 0.01566407084465027,
+ "learning_rate": 0.0006,
+ "loss": 4.6337199211120605,
+ "step": 1145
+ },
+ {
+ "epoch": 15.922270742358078,
+ "grad_norm": 0.016823330894112587,
+ "learning_rate": 0.0006,
+ "loss": 4.686994552612305,
+ "step": 1146
+ },
+ {
+ "epoch": 15.936244541484717,
+ "grad_norm": 0.016208255663514137,
+ "learning_rate": 0.0006,
+ "loss": 4.6205549240112305,
+ "step": 1147
+ },
+ {
+ "epoch": 15.950218340611354,
+ "grad_norm": 0.0171707421541214,
+ "learning_rate": 0.0006,
+ "loss": 4.582968711853027,
+ "step": 1148
+ },
+ {
+ "epoch": 15.96419213973799,
+ "grad_norm": 0.017009446397423744,
+ "learning_rate": 0.0006,
+ "loss": 4.510274887084961,
+ "step": 1149
+ },
+ {
+ "epoch": 15.97816593886463,
+ "grad_norm": 0.01565811224281788,
+ "learning_rate": 0.0006,
+ "loss": 4.55834436416626,
+ "step": 1150
+ },
+ {
+ "epoch": 15.992139737991266,
+ "grad_norm": 0.015276037156581879,
+ "learning_rate": 0.0006,
+ "loss": 4.537870407104492,
+ "step": 1151
+ },
+ {
+ "epoch": 16.0,
+ "grad_norm": 0.017305882647633553,
+ "learning_rate": 0.0006,
+ "loss": 4.611305236816406,
+ "step": 1152
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 28800,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 400,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": true,
+ "should_log": false,
+ "should_save": false,
+ "should_training_stop": false
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 4.9113813781315584e+17,
+ "train_batch_size": 8,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-1152/training_args.bin b/runs/l2r50-i2-fulle-lm/checkpoint-1152/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..cad232831c222cc7f8b938b8bbe42f61e7d89c3a
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-1152/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:f7c40e3bfa2f5ac1450ca1243e3f2a895d44a06ed5995e6aefd97f78d229d8c8
+size 4792
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-1440/chat_template.jinja b/runs/l2r50-i2-fulle-lm/checkpoint-1440/chat_template.jinja
new file mode 100644
index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-1440/chat_template.jinja
@@ -0,0 +1,85 @@
+{%- if tools %}
+ {{- '<|im_start|>system\n' }}
+ {%- if messages[0].role == 'system' %}
+ {{- messages[0].content + '\n\n' }}
+ {%- endif %}
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }}
+ {%- for tool in tools %}
+ {{- "\n" }}
+ {{- tool | tojson }}
+ {%- endfor %}
+ {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }}
+{%- else %}
+ {%- if messages[0].role == 'system' %}
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
+ {%- endif %}
+{%- endif %}
+{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
+{%- for message in messages[::-1] %}
+ {%- set index = (messages|length - 1) - loop.index0 %}
+ {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %}
+ {%- set ns.multi_step_tool = false %}
+ {%- set ns.last_query_index = index %}
+ {%- endif %}
+{%- endfor %}
+{%- for message in messages %}
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
+ {%- elif message.role == "assistant" %}
+ {%- set content = message.content %}
+ {%- set reasoning_content = '' %}
+ {%- if message.reasoning_content is defined and message.reasoning_content is not none %}
+ {%- set reasoning_content = message.reasoning_content %}
+ {%- else %}
+ {%- if '' in message.content %}
+ {%- set content = message.content.split('')[-1].lstrip('\n') %}
+ {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %}
+ {%- endif %}
+ {%- endif %}
+ {%- if loop.index0 > ns.last_query_index %}
+ {%- if loop.last or (not loop.last and reasoning_content) %}
+ {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- if message.tool_calls %}
+ {%- for tool_call in message.tool_calls %}
+ {%- if (loop.first and content) or (not loop.first) %}
+ {{- '\n' }}
+ {%- endif %}
+ {%- if tool_call.function %}
+ {%- set tool_call = tool_call.function %}
+ {%- endif %}
+ {{- '\n{"name": "' }}
+ {{- tool_call.name }}
+ {{- '", "arguments": ' }}
+ {%- if tool_call.arguments is string %}
+ {{- tool_call.arguments }}
+ {%- else %}
+ {{- tool_call.arguments | tojson }}
+ {%- endif %}
+ {{- '}\n' }}
+ {%- endfor %}
+ {%- endif %}
+ {{- '<|im_end|>\n' }}
+ {%- elif message.role == "tool" %}
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
+ {{- '<|im_start|>user' }}
+ {%- endif %}
+ {{- '\n\n' }}
+ {{- message.content }}
+ {{- '\n' }}
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
+ {{- '<|im_end|>\n' }}
+ {%- endif %}
+ {%- endif %}
+{%- endfor %}
+{%- if add_generation_prompt %}
+ {{- '<|im_start|>assistant\n' }}
+ {%- if enable_thinking is defined and enable_thinking is false %}
+ {{- '\n\n\n\n' }}
+ {%- endif %}
+{%- endif %}
\ No newline at end of file
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-1440/config.json b/runs/l2r50-i2-fulle-lm/checkpoint-1440/config.json
new file mode 100644
index 0000000000000000000000000000000000000000..22f250953c6ce8fc91ee4411856dd73041a415e2
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-1440/config.json
@@ -0,0 +1,32 @@
+{
+ "architectures": [
+ "LlamaForCausalLM"
+ ],
+ "attention_bias": false,
+ "attention_dropout": 0.0,
+ "bos_token_id": null,
+ "dtype": "float32",
+ "eos_token_id": 151645,
+ "head_dim": 128,
+ "hidden_act": "silu",
+ "hidden_size": 512,
+ "initializer_range": 0.02,
+ "intermediate_size": 1536,
+ "max_position_embeddings": 2048,
+ "mlp_bias": false,
+ "model_type": "llama",
+ "num_attention_heads": 4,
+ "num_hidden_layers": 20,
+ "num_key_value_heads": 4,
+ "pad_token_id": 151645,
+ "pretraining_tp": 1,
+ "rms_norm_eps": 1e-06,
+ "rope_parameters": {
+ "rope_theta": 10000.0,
+ "rope_type": "default"
+ },
+ "tie_word_embeddings": true,
+ "transformers_version": "5.5.0",
+ "use_cache": false,
+ "vocab_size": 151673
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-1440/generation_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-1440/generation_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-1440/generation_config.json
@@ -0,0 +1,11 @@
+{
+ "_from_model_config": true,
+ "eos_token_id": [
+ 151645
+ ],
+ "output_attentions": false,
+ "output_hidden_states": false,
+ "pad_token_id": 151645,
+ "transformers_version": "5.5.0",
+ "use_cache": true
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-1440/model.safetensors b/runs/l2r50-i2-fulle-lm/checkpoint-1440/model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..a95eb51a71cd11b85f9e052497cbed691f196824
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-1440/model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:0a7ef37b8c5a1dc5e759d2999fc4d9b087b2eaa5bc0fb5009e3ea232ff9a271c
+size 583360328
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-1440/optimizer.pt b/runs/l2r50-i2-fulle-lm/checkpoint-1440/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..42d3088c631310994a23360d6e906b1f73b7bc60
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-1440/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:bd15e0e248cb1e32f608a1690428b2e487cf4635fdee6ef078225c433a16402b
+size 1166833530
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-1440/rng_state_0.pth b/runs/l2r50-i2-fulle-lm/checkpoint-1440/rng_state_0.pth
new file mode 100644
index 0000000000000000000000000000000000000000..fa6594ba5f910ea66238338f849a8e953e567e1a
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-1440/rng_state_0.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:ff0d1c60dbaf626d9ffd2cf5f49d2a345d6f5acba8013a96c4d1b9b99d743250
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-1440/rng_state_1.pth b/runs/l2r50-i2-fulle-lm/checkpoint-1440/rng_state_1.pth
new file mode 100644
index 0000000000000000000000000000000000000000..e062f681cfc898c2151a607d123b890be40e76d8
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-1440/rng_state_1.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:92360cbd2f45549906da13b2089de2060a693f31beb68deff404d66ba4058b9f
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-1440/rng_state_2.pth b/runs/l2r50-i2-fulle-lm/checkpoint-1440/rng_state_2.pth
new file mode 100644
index 0000000000000000000000000000000000000000..cbf904db0a94ff0866429f1dee2aad9896198c54
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-1440/rng_state_2.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:43ae0c839fafd8f4cae880ebeae7081f01518492971a5a9301e81c43373bafae
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-1440/rng_state_3.pth b/runs/l2r50-i2-fulle-lm/checkpoint-1440/rng_state_3.pth
new file mode 100644
index 0000000000000000000000000000000000000000..322e7e0906e04eea8aa972bd0c5855ca1ca3fa97
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-1440/rng_state_3.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:a970baceb23ab09484fef82bc4cb2e61052c409e73fd0fd60914013c019d38d6
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-1440/scheduler.pt b/runs/l2r50-i2-fulle-lm/checkpoint-1440/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..0d34fe10b069b50b5fecbd4a0affbd8f653b7b9b
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-1440/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:8a84bbc84bb0dd489a2d72fd441cc0b8482e9e06d7cde3ac44b50a389883ce86
+size 1064
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-1440/tokenizer.json b/runs/l2r50-i2-fulle-lm/checkpoint-1440/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..53786016f070e11184813f3c83101dc5e93e392e
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-1440/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:c3c63fda1646a555448e8f03662542e04b11d103721319e4d9b267d7da03ceb8
+size 11424463
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-1440/tokenizer_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-1440/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..98e2fb9b2cc54785dc9d871b8e503f8b8f2defe7
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-1440/tokenizer_config.json
@@ -0,0 +1,20 @@
+{
+ "add_prefix_space": false,
+ "backend": "tokenizers",
+ "bos_token": null,
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "<|im_end|>",
+ "errors": "replace",
+ "extra_special_tokens": [
+ "<|l2r_pred|>",
+ "<|r2l_pred|>",
+ "<|next_1_pred|>",
+ "<|next_2_pred|>"
+ ],
+ "is_local": false,
+ "model_max_length": 131072,
+ "pad_token": "<|im_end|>",
+ "split_special_tokens": false,
+ "tokenizer_class": "Qwen2Tokenizer",
+ "unk_token": null
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-1440/trainer_state.json b/runs/l2r50-i2-fulle-lm/checkpoint-1440/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..4e979d33cc004ab839d678dd86839e9af319496f
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-1440/trainer_state.json
@@ -0,0 +1,10266 @@
+{
+ "best_global_step": null,
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 20.0,
+ "eval_steps": 500,
+ "global_step": 1440,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.013973799126637555,
+ "grad_norm": 0.1739506870508194,
+ "learning_rate": 0.0,
+ "loss": 11.99299430847168,
+ "step": 1
+ },
+ {
+ "epoch": 0.02794759825327511,
+ "grad_norm": 0.17216597497463226,
+ "learning_rate": 5.999999999999999e-06,
+ "loss": 11.988704681396484,
+ "step": 2
+ },
+ {
+ "epoch": 0.04192139737991266,
+ "grad_norm": 0.17459715902805328,
+ "learning_rate": 1.1999999999999999e-05,
+ "loss": 11.945581436157227,
+ "step": 3
+ },
+ {
+ "epoch": 0.05589519650655022,
+ "grad_norm": 0.17577414214611053,
+ "learning_rate": 1.7999999999999997e-05,
+ "loss": 11.859772682189941,
+ "step": 4
+ },
+ {
+ "epoch": 0.06986899563318777,
+ "grad_norm": 0.16163651645183563,
+ "learning_rate": 2.3999999999999997e-05,
+ "loss": 11.744415283203125,
+ "step": 5
+ },
+ {
+ "epoch": 0.08384279475982533,
+ "grad_norm": 0.14409932494163513,
+ "learning_rate": 2.9999999999999997e-05,
+ "loss": 11.618099212646484,
+ "step": 6
+ },
+ {
+ "epoch": 0.09781659388646288,
+ "grad_norm": 0.13079848885536194,
+ "learning_rate": 3.5999999999999994e-05,
+ "loss": 11.505935668945312,
+ "step": 7
+ },
+ {
+ "epoch": 0.11179039301310044,
+ "grad_norm": 0.11964613944292068,
+ "learning_rate": 4.2e-05,
+ "loss": 11.416532516479492,
+ "step": 8
+ },
+ {
+ "epoch": 0.125764192139738,
+ "grad_norm": 0.11345142871141434,
+ "learning_rate": 4.7999999999999994e-05,
+ "loss": 11.33206558227539,
+ "step": 9
+ },
+ {
+ "epoch": 0.13973799126637554,
+ "grad_norm": 0.11033952981233597,
+ "learning_rate": 5.399999999999999e-05,
+ "loss": 11.259052276611328,
+ "step": 10
+ },
+ {
+ "epoch": 0.1537117903930131,
+ "grad_norm": 0.10776489973068237,
+ "learning_rate": 5.9999999999999995e-05,
+ "loss": 11.203384399414062,
+ "step": 11
+ },
+ {
+ "epoch": 0.16768558951965065,
+ "grad_norm": 0.10721946507692337,
+ "learning_rate": 6.599999999999999e-05,
+ "loss": 11.153295516967773,
+ "step": 12
+ },
+ {
+ "epoch": 0.18165938864628822,
+ "grad_norm": 0.10673430562019348,
+ "learning_rate": 7.199999999999999e-05,
+ "loss": 11.110200881958008,
+ "step": 13
+ },
+ {
+ "epoch": 0.19563318777292577,
+ "grad_norm": 0.10630254447460175,
+ "learning_rate": 7.8e-05,
+ "loss": 11.067477226257324,
+ "step": 14
+ },
+ {
+ "epoch": 0.2096069868995633,
+ "grad_norm": 0.10631615668535233,
+ "learning_rate": 8.4e-05,
+ "loss": 11.022941589355469,
+ "step": 15
+ },
+ {
+ "epoch": 0.22358078602620088,
+ "grad_norm": 0.10609443485736847,
+ "learning_rate": 8.999999999999999e-05,
+ "loss": 10.970746040344238,
+ "step": 16
+ },
+ {
+ "epoch": 0.23755458515283842,
+ "grad_norm": 0.10648232698440552,
+ "learning_rate": 9.599999999999999e-05,
+ "loss": 10.91267204284668,
+ "step": 17
+ },
+ {
+ "epoch": 0.251528384279476,
+ "grad_norm": 0.10584589838981628,
+ "learning_rate": 0.000102,
+ "loss": 10.854547500610352,
+ "step": 18
+ },
+ {
+ "epoch": 0.26550218340611353,
+ "grad_norm": 0.10591720044612885,
+ "learning_rate": 0.00010799999999999998,
+ "loss": 10.788917541503906,
+ "step": 19
+ },
+ {
+ "epoch": 0.2794759825327511,
+ "grad_norm": 0.10611672699451447,
+ "learning_rate": 0.00011399999999999999,
+ "loss": 10.717556953430176,
+ "step": 20
+ },
+ {
+ "epoch": 0.2934497816593886,
+ "grad_norm": 0.10455359518527985,
+ "learning_rate": 0.00011999999999999999,
+ "loss": 10.650186538696289,
+ "step": 21
+ },
+ {
+ "epoch": 0.3074235807860262,
+ "grad_norm": 0.1047525629401207,
+ "learning_rate": 0.00012599999999999997,
+ "loss": 10.56808090209961,
+ "step": 22
+ },
+ {
+ "epoch": 0.32139737991266376,
+ "grad_norm": 0.10511605441570282,
+ "learning_rate": 0.00013199999999999998,
+ "loss": 10.484028816223145,
+ "step": 23
+ },
+ {
+ "epoch": 0.3353711790393013,
+ "grad_norm": 0.10467177629470825,
+ "learning_rate": 0.000138,
+ "loss": 10.399947166442871,
+ "step": 24
+ },
+ {
+ "epoch": 0.34934497816593885,
+ "grad_norm": 0.10432136058807373,
+ "learning_rate": 0.00014399999999999998,
+ "loss": 10.312714576721191,
+ "step": 25
+ },
+ {
+ "epoch": 0.36331877729257644,
+ "grad_norm": 0.10527262091636658,
+ "learning_rate": 0.00015,
+ "loss": 10.212261199951172,
+ "step": 26
+ },
+ {
+ "epoch": 0.377292576419214,
+ "grad_norm": 0.103610560297966,
+ "learning_rate": 0.000156,
+ "loss": 10.132034301757812,
+ "step": 27
+ },
+ {
+ "epoch": 0.39126637554585153,
+ "grad_norm": 0.10553059726953506,
+ "learning_rate": 0.000162,
+ "loss": 10.020434379577637,
+ "step": 28
+ },
+ {
+ "epoch": 0.4052401746724891,
+ "grad_norm": 0.1037299782037735,
+ "learning_rate": 0.000168,
+ "loss": 9.934466361999512,
+ "step": 29
+ },
+ {
+ "epoch": 0.4192139737991266,
+ "grad_norm": 0.1042739674448967,
+ "learning_rate": 0.00017399999999999997,
+ "loss": 9.822539329528809,
+ "step": 30
+ },
+ {
+ "epoch": 0.4331877729257642,
+ "grad_norm": 0.10300865024328232,
+ "learning_rate": 0.00017999999999999998,
+ "loss": 9.734994888305664,
+ "step": 31
+ },
+ {
+ "epoch": 0.44716157205240176,
+ "grad_norm": 0.10291654616594315,
+ "learning_rate": 0.000186,
+ "loss": 9.624505996704102,
+ "step": 32
+ },
+ {
+ "epoch": 0.4611353711790393,
+ "grad_norm": 0.10210851579904556,
+ "learning_rate": 0.00019199999999999998,
+ "loss": 9.518205642700195,
+ "step": 33
+ },
+ {
+ "epoch": 0.47510917030567684,
+ "grad_norm": 0.10222529619932175,
+ "learning_rate": 0.000198,
+ "loss": 9.412222862243652,
+ "step": 34
+ },
+ {
+ "epoch": 0.4890829694323144,
+ "grad_norm": 0.09992019087076187,
+ "learning_rate": 0.000204,
+ "loss": 9.326566696166992,
+ "step": 35
+ },
+ {
+ "epoch": 0.503056768558952,
+ "grad_norm": 0.09931084513664246,
+ "learning_rate": 0.00020999999999999998,
+ "loss": 9.210111618041992,
+ "step": 36
+ },
+ {
+ "epoch": 0.5170305676855895,
+ "grad_norm": 0.09928256273269653,
+ "learning_rate": 0.00021599999999999996,
+ "loss": 9.100493431091309,
+ "step": 37
+ },
+ {
+ "epoch": 0.5310043668122271,
+ "grad_norm": 0.09843816608190536,
+ "learning_rate": 0.00022199999999999998,
+ "loss": 8.996543884277344,
+ "step": 38
+ },
+ {
+ "epoch": 0.5449781659388646,
+ "grad_norm": 0.09825374186038971,
+ "learning_rate": 0.00022799999999999999,
+ "loss": 8.888991355895996,
+ "step": 39
+ },
+ {
+ "epoch": 0.5589519650655022,
+ "grad_norm": 0.09602385759353638,
+ "learning_rate": 0.000234,
+ "loss": 8.802907943725586,
+ "step": 40
+ },
+ {
+ "epoch": 0.5729257641921397,
+ "grad_norm": 0.09414064139127731,
+ "learning_rate": 0.00023999999999999998,
+ "loss": 8.715423583984375,
+ "step": 41
+ },
+ {
+ "epoch": 0.5868995633187772,
+ "grad_norm": 0.09273893386125565,
+ "learning_rate": 0.00024599999999999996,
+ "loss": 8.61374282836914,
+ "step": 42
+ },
+ {
+ "epoch": 0.6008733624454149,
+ "grad_norm": 0.09079599380493164,
+ "learning_rate": 0.00025199999999999995,
+ "loss": 8.502904891967773,
+ "step": 43
+ },
+ {
+ "epoch": 0.6148471615720524,
+ "grad_norm": 0.08605825901031494,
+ "learning_rate": 0.000258,
+ "loss": 8.453008651733398,
+ "step": 44
+ },
+ {
+ "epoch": 0.62882096069869,
+ "grad_norm": 0.08338996022939682,
+ "learning_rate": 0.00026399999999999997,
+ "loss": 8.365163803100586,
+ "step": 45
+ },
+ {
+ "epoch": 0.6427947598253275,
+ "grad_norm": 0.08150867372751236,
+ "learning_rate": 0.00027,
+ "loss": 8.285372734069824,
+ "step": 46
+ },
+ {
+ "epoch": 0.6567685589519651,
+ "grad_norm": 0.07794807851314545,
+ "learning_rate": 0.000276,
+ "loss": 8.185811996459961,
+ "step": 47
+ },
+ {
+ "epoch": 0.6707423580786026,
+ "grad_norm": 0.07592720538377762,
+ "learning_rate": 0.00028199999999999997,
+ "loss": 8.098482131958008,
+ "step": 48
+ },
+ {
+ "epoch": 0.6847161572052402,
+ "grad_norm": 0.0709897130727768,
+ "learning_rate": 0.00028799999999999995,
+ "loss": 8.042091369628906,
+ "step": 49
+ },
+ {
+ "epoch": 0.6986899563318777,
+ "grad_norm": 0.06582305580377579,
+ "learning_rate": 0.000294,
+ "loss": 7.972844123840332,
+ "step": 50
+ },
+ {
+ "epoch": 0.7126637554585152,
+ "grad_norm": 0.059452351182699203,
+ "learning_rate": 0.0003,
+ "loss": 7.924579620361328,
+ "step": 51
+ },
+ {
+ "epoch": 0.7266375545851529,
+ "grad_norm": 0.052426718175411224,
+ "learning_rate": 0.00030599999999999996,
+ "loss": 7.876336097717285,
+ "step": 52
+ },
+ {
+ "epoch": 0.7406113537117904,
+ "grad_norm": 0.04768150672316551,
+ "learning_rate": 0.000312,
+ "loss": 7.8102707862854,
+ "step": 53
+ },
+ {
+ "epoch": 0.754585152838428,
+ "grad_norm": 0.03970075398683548,
+ "learning_rate": 0.000318,
+ "loss": 7.791913986206055,
+ "step": 54
+ },
+ {
+ "epoch": 0.7685589519650655,
+ "grad_norm": 0.03559752553701401,
+ "learning_rate": 0.000324,
+ "loss": 7.756960868835449,
+ "step": 55
+ },
+ {
+ "epoch": 0.7825327510917031,
+ "grad_norm": 0.029321353882551193,
+ "learning_rate": 0.00033,
+ "loss": 7.724878787994385,
+ "step": 56
+ },
+ {
+ "epoch": 0.7965065502183406,
+ "grad_norm": 0.023485427722334862,
+ "learning_rate": 0.000336,
+ "loss": 7.72553014755249,
+ "step": 57
+ },
+ {
+ "epoch": 0.8104803493449781,
+ "grad_norm": 0.019319474697113037,
+ "learning_rate": 0.00034199999999999996,
+ "loss": 7.709493160247803,
+ "step": 58
+ },
+ {
+ "epoch": 0.8244541484716157,
+ "grad_norm": 0.016705691814422607,
+ "learning_rate": 0.00034799999999999995,
+ "loss": 7.6873297691345215,
+ "step": 59
+ },
+ {
+ "epoch": 0.8384279475982532,
+ "grad_norm": 0.017667734995484352,
+ "learning_rate": 0.00035399999999999993,
+ "loss": 7.677059173583984,
+ "step": 60
+ },
+ {
+ "epoch": 0.8524017467248908,
+ "grad_norm": 0.01671992614865303,
+ "learning_rate": 0.00035999999999999997,
+ "loss": 7.643512725830078,
+ "step": 61
+ },
+ {
+ "epoch": 0.8663755458515284,
+ "grad_norm": 0.012288172729313374,
+ "learning_rate": 0.00036599999999999995,
+ "loss": 7.658343315124512,
+ "step": 62
+ },
+ {
+ "epoch": 0.880349344978166,
+ "grad_norm": 0.012004776857793331,
+ "learning_rate": 0.000372,
+ "loss": 7.649759292602539,
+ "step": 63
+ },
+ {
+ "epoch": 0.8943231441048035,
+ "grad_norm": 0.012487445957958698,
+ "learning_rate": 0.00037799999999999997,
+ "loss": 7.631503105163574,
+ "step": 64
+ },
+ {
+ "epoch": 0.9082969432314411,
+ "grad_norm": 0.01489241048693657,
+ "learning_rate": 0.00038399999999999996,
+ "loss": 7.647110462188721,
+ "step": 65
+ },
+ {
+ "epoch": 0.9222707423580786,
+ "grad_norm": 0.01428980939090252,
+ "learning_rate": 0.00039,
+ "loss": 7.636983871459961,
+ "step": 66
+ },
+ {
+ "epoch": 0.9362445414847161,
+ "grad_norm": 0.010711144655942917,
+ "learning_rate": 0.000396,
+ "loss": 7.621130466461182,
+ "step": 67
+ },
+ {
+ "epoch": 0.9502183406113537,
+ "grad_norm": 0.010178886353969574,
+ "learning_rate": 0.000402,
+ "loss": 7.593924522399902,
+ "step": 68
+ },
+ {
+ "epoch": 0.9641921397379912,
+ "grad_norm": 0.010604874230921268,
+ "learning_rate": 0.000408,
+ "loss": 7.6015496253967285,
+ "step": 69
+ },
+ {
+ "epoch": 0.9781659388646288,
+ "grad_norm": 0.009751426056027412,
+ "learning_rate": 0.0004139999999999999,
+ "loss": 7.571964263916016,
+ "step": 70
+ },
+ {
+ "epoch": 0.9921397379912664,
+ "grad_norm": 0.009816481731832027,
+ "learning_rate": 0.00041999999999999996,
+ "loss": 7.57344913482666,
+ "step": 71
+ },
+ {
+ "epoch": 1.0,
+ "grad_norm": 0.008665637113153934,
+ "learning_rate": 0.00042599999999999995,
+ "loss": 7.579050064086914,
+ "step": 72
+ },
+ {
+ "epoch": 1.0,
+ "eval_loss": 7.581204891204834,
+ "eval_runtime": 59.0091,
+ "eval_samples_per_second": 41.383,
+ "eval_steps_per_second": 1.305,
+ "step": 72
+ },
+ {
+ "epoch": 1.0139737991266375,
+ "grad_norm": 0.009684201329946518,
+ "learning_rate": 0.00043199999999999993,
+ "loss": 7.568594932556152,
+ "step": 73
+ },
+ {
+ "epoch": 1.027947598253275,
+ "grad_norm": 0.025783710181713104,
+ "learning_rate": 0.00043799999999999997,
+ "loss": 7.578948497772217,
+ "step": 74
+ },
+ {
+ "epoch": 1.0419213973799126,
+ "grad_norm": 0.0327819399535656,
+ "learning_rate": 0.00044399999999999995,
+ "loss": 7.548173904418945,
+ "step": 75
+ },
+ {
+ "epoch": 1.0558951965065502,
+ "grad_norm": 0.021604498848319054,
+ "learning_rate": 0.00045,
+ "loss": 7.552953720092773,
+ "step": 76
+ },
+ {
+ "epoch": 1.0698689956331877,
+ "grad_norm": 0.010230105370283127,
+ "learning_rate": 0.00045599999999999997,
+ "loss": 7.5213212966918945,
+ "step": 77
+ },
+ {
+ "epoch": 1.0838427947598253,
+ "grad_norm": 0.023315589874982834,
+ "learning_rate": 0.00046199999999999995,
+ "loss": 7.501640319824219,
+ "step": 78
+ },
+ {
+ "epoch": 1.0978165938864628,
+ "grad_norm": 0.01109944935888052,
+ "learning_rate": 0.000468,
+ "loss": 7.539519786834717,
+ "step": 79
+ },
+ {
+ "epoch": 1.1117903930131003,
+ "grad_norm": 0.026093758642673492,
+ "learning_rate": 0.000474,
+ "loss": 7.525928974151611,
+ "step": 80
+ },
+ {
+ "epoch": 1.125764192139738,
+ "grad_norm": 0.03274895250797272,
+ "learning_rate": 0.00047999999999999996,
+ "loss": 7.535602569580078,
+ "step": 81
+ },
+ {
+ "epoch": 1.1397379912663754,
+ "grad_norm": 0.01543602254241705,
+ "learning_rate": 0.000486,
+ "loss": 7.513683795928955,
+ "step": 82
+ },
+ {
+ "epoch": 1.1537117903930132,
+ "grad_norm": 0.028377249836921692,
+ "learning_rate": 0.0004919999999999999,
+ "loss": 7.49384880065918,
+ "step": 83
+ },
+ {
+ "epoch": 1.1676855895196507,
+ "grad_norm": 0.04121481627225876,
+ "learning_rate": 0.000498,
+ "loss": 7.488447189331055,
+ "step": 84
+ },
+ {
+ "epoch": 1.1816593886462883,
+ "grad_norm": 0.013282302767038345,
+ "learning_rate": 0.0005039999999999999,
+ "loss": 7.463010787963867,
+ "step": 85
+ },
+ {
+ "epoch": 1.1956331877729258,
+ "grad_norm": 0.03200364485383034,
+ "learning_rate": 0.0005099999999999999,
+ "loss": 7.473515033721924,
+ "step": 86
+ },
+ {
+ "epoch": 1.2096069868995634,
+ "grad_norm": 0.020768029615283012,
+ "learning_rate": 0.000516,
+ "loss": 7.464688777923584,
+ "step": 87
+ },
+ {
+ "epoch": 1.223580786026201,
+ "grad_norm": 0.021715788170695305,
+ "learning_rate": 0.000522,
+ "loss": 7.4835028648376465,
+ "step": 88
+ },
+ {
+ "epoch": 1.2375545851528384,
+ "grad_norm": 0.026255754753947258,
+ "learning_rate": 0.0005279999999999999,
+ "loss": 7.443392753601074,
+ "step": 89
+ },
+ {
+ "epoch": 1.251528384279476,
+ "grad_norm": 0.022325951606035233,
+ "learning_rate": 0.000534,
+ "loss": 7.457373142242432,
+ "step": 90
+ },
+ {
+ "epoch": 1.2655021834061135,
+ "grad_norm": 0.009270387701690197,
+ "learning_rate": 0.00054,
+ "loss": 7.455986976623535,
+ "step": 91
+ },
+ {
+ "epoch": 1.279475982532751,
+ "grad_norm": 0.02369544468820095,
+ "learning_rate": 0.0005459999999999999,
+ "loss": 7.442216873168945,
+ "step": 92
+ },
+ {
+ "epoch": 1.2934497816593886,
+ "grad_norm": 0.008474506437778473,
+ "learning_rate": 0.000552,
+ "loss": 7.4294915199279785,
+ "step": 93
+ },
+ {
+ "epoch": 1.3074235807860262,
+ "grad_norm": 0.017043249681591988,
+ "learning_rate": 0.000558,
+ "loss": 7.406932830810547,
+ "step": 94
+ },
+ {
+ "epoch": 1.3213973799126637,
+ "grad_norm": 0.011813412420451641,
+ "learning_rate": 0.0005639999999999999,
+ "loss": 7.415900230407715,
+ "step": 95
+ },
+ {
+ "epoch": 1.3353711790393012,
+ "grad_norm": 0.01086883433163166,
+ "learning_rate": 0.00057,
+ "loss": 7.452998638153076,
+ "step": 96
+ },
+ {
+ "epoch": 1.3493449781659388,
+ "grad_norm": 0.014700951054692268,
+ "learning_rate": 0.0005759999999999999,
+ "loss": 7.410249710083008,
+ "step": 97
+ },
+ {
+ "epoch": 1.3633187772925766,
+ "grad_norm": 0.014512976631522179,
+ "learning_rate": 0.0005819999999999999,
+ "loss": 7.429132461547852,
+ "step": 98
+ },
+ {
+ "epoch": 1.3772925764192139,
+ "grad_norm": 0.012519920244812965,
+ "learning_rate": 0.000588,
+ "loss": 7.430817604064941,
+ "step": 99
+ },
+ {
+ "epoch": 1.3912663755458516,
+ "grad_norm": 0.007132664322853088,
+ "learning_rate": 0.0005939999999999999,
+ "loss": 7.405234336853027,
+ "step": 100
+ },
+ {
+ "epoch": 1.405240174672489,
+ "grad_norm": 0.019489388912916183,
+ "learning_rate": 0.0006,
+ "loss": 7.427791595458984,
+ "step": 101
+ },
+ {
+ "epoch": 1.4192139737991267,
+ "grad_norm": 0.027252664789557457,
+ "learning_rate": 0.0006,
+ "loss": 7.41761589050293,
+ "step": 102
+ },
+ {
+ "epoch": 1.4331877729257643,
+ "grad_norm": 0.02501637488603592,
+ "learning_rate": 0.0006,
+ "loss": 7.416735649108887,
+ "step": 103
+ },
+ {
+ "epoch": 1.4471615720524018,
+ "grad_norm": 0.00989772193133831,
+ "learning_rate": 0.0006,
+ "loss": 7.3756632804870605,
+ "step": 104
+ },
+ {
+ "epoch": 1.4611353711790394,
+ "grad_norm": 0.01622949168086052,
+ "learning_rate": 0.0006,
+ "loss": 7.400811672210693,
+ "step": 105
+ },
+ {
+ "epoch": 1.475109170305677,
+ "grad_norm": 0.012407763861119747,
+ "learning_rate": 0.0006,
+ "loss": 7.37675666809082,
+ "step": 106
+ },
+ {
+ "epoch": 1.4890829694323144,
+ "grad_norm": 0.010353565216064453,
+ "learning_rate": 0.0006,
+ "loss": 7.4013142585754395,
+ "step": 107
+ },
+ {
+ "epoch": 1.503056768558952,
+ "grad_norm": 0.0120553532615304,
+ "learning_rate": 0.0006,
+ "loss": 7.379873275756836,
+ "step": 108
+ },
+ {
+ "epoch": 1.5170305676855895,
+ "grad_norm": 0.01610225811600685,
+ "learning_rate": 0.0006,
+ "loss": 7.371631145477295,
+ "step": 109
+ },
+ {
+ "epoch": 1.531004366812227,
+ "grad_norm": 0.01159879844635725,
+ "learning_rate": 0.0006,
+ "loss": 7.389007568359375,
+ "step": 110
+ },
+ {
+ "epoch": 1.5449781659388646,
+ "grad_norm": 0.008013816550374031,
+ "learning_rate": 0.0006,
+ "loss": 7.395920753479004,
+ "step": 111
+ },
+ {
+ "epoch": 1.5589519650655022,
+ "grad_norm": 0.016334809362888336,
+ "learning_rate": 0.0006,
+ "loss": 7.393121719360352,
+ "step": 112
+ },
+ {
+ "epoch": 1.5729257641921397,
+ "grad_norm": 0.018821721896529198,
+ "learning_rate": 0.0006,
+ "loss": 7.367543697357178,
+ "step": 113
+ },
+ {
+ "epoch": 1.5868995633187772,
+ "grad_norm": 0.014088994823396206,
+ "learning_rate": 0.0006,
+ "loss": 7.385537624359131,
+ "step": 114
+ },
+ {
+ "epoch": 1.600873362445415,
+ "grad_norm": 0.005864050704985857,
+ "learning_rate": 0.0006,
+ "loss": 7.365346908569336,
+ "step": 115
+ },
+ {
+ "epoch": 1.6148471615720523,
+ "grad_norm": 0.011745430529117584,
+ "learning_rate": 0.0006,
+ "loss": 7.390227317810059,
+ "step": 116
+ },
+ {
+ "epoch": 1.62882096069869,
+ "grad_norm": 0.017205573618412018,
+ "learning_rate": 0.0006,
+ "loss": 7.3558549880981445,
+ "step": 117
+ },
+ {
+ "epoch": 1.6427947598253274,
+ "grad_norm": 0.021256346255540848,
+ "learning_rate": 0.0006,
+ "loss": 7.377346992492676,
+ "step": 118
+ },
+ {
+ "epoch": 1.6567685589519652,
+ "grad_norm": 0.019335204735398293,
+ "learning_rate": 0.0006,
+ "loss": 7.346194267272949,
+ "step": 119
+ },
+ {
+ "epoch": 1.6707423580786025,
+ "grad_norm": 0.008134003728628159,
+ "learning_rate": 0.0006,
+ "loss": 7.3063578605651855,
+ "step": 120
+ },
+ {
+ "epoch": 1.6847161572052403,
+ "grad_norm": 0.017613831907510757,
+ "learning_rate": 0.0006,
+ "loss": 7.338171005249023,
+ "step": 121
+ },
+ {
+ "epoch": 1.6986899563318776,
+ "grad_norm": 0.00890734139829874,
+ "learning_rate": 0.0006,
+ "loss": 7.355384826660156,
+ "step": 122
+ },
+ {
+ "epoch": 1.7126637554585153,
+ "grad_norm": 0.013448155485093594,
+ "learning_rate": 0.0006,
+ "loss": 7.344296455383301,
+ "step": 123
+ },
+ {
+ "epoch": 1.726637554585153,
+ "grad_norm": 0.009241633117198944,
+ "learning_rate": 0.0006,
+ "loss": 7.3454060554504395,
+ "step": 124
+ },
+ {
+ "epoch": 1.7406113537117904,
+ "grad_norm": 0.018561311066150665,
+ "learning_rate": 0.0006,
+ "loss": 7.32348108291626,
+ "step": 125
+ },
+ {
+ "epoch": 1.754585152838428,
+ "grad_norm": 0.025667704641819,
+ "learning_rate": 0.0006,
+ "loss": 7.356142520904541,
+ "step": 126
+ },
+ {
+ "epoch": 1.7685589519650655,
+ "grad_norm": 0.036232151091098785,
+ "learning_rate": 0.0006,
+ "loss": 7.363539695739746,
+ "step": 127
+ },
+ {
+ "epoch": 1.782532751091703,
+ "grad_norm": 0.059483714401721954,
+ "learning_rate": 0.0006,
+ "loss": 7.402331352233887,
+ "step": 128
+ },
+ {
+ "epoch": 1.7965065502183406,
+ "grad_norm": 0.05365009233355522,
+ "learning_rate": 0.0006,
+ "loss": 7.40457820892334,
+ "step": 129
+ },
+ {
+ "epoch": 1.8104803493449781,
+ "grad_norm": 0.02738415263593197,
+ "learning_rate": 0.0006,
+ "loss": 7.3756208419799805,
+ "step": 130
+ },
+ {
+ "epoch": 1.8244541484716157,
+ "grad_norm": 0.035017091780900955,
+ "learning_rate": 0.0006,
+ "loss": 7.379263877868652,
+ "step": 131
+ },
+ {
+ "epoch": 1.8384279475982532,
+ "grad_norm": 0.02819863148033619,
+ "learning_rate": 0.0006,
+ "loss": 7.327673435211182,
+ "step": 132
+ },
+ {
+ "epoch": 1.8524017467248908,
+ "grad_norm": 0.017754238098859787,
+ "learning_rate": 0.0006,
+ "loss": 7.333420276641846,
+ "step": 133
+ },
+ {
+ "epoch": 1.8663755458515285,
+ "grad_norm": 0.024685733020305634,
+ "learning_rate": 0.0006,
+ "loss": 7.34493350982666,
+ "step": 134
+ },
+ {
+ "epoch": 1.8803493449781659,
+ "grad_norm": 0.019426532089710236,
+ "learning_rate": 0.0006,
+ "loss": 7.328448295593262,
+ "step": 135
+ },
+ {
+ "epoch": 1.8943231441048036,
+ "grad_norm": 0.02106391079723835,
+ "learning_rate": 0.0006,
+ "loss": 7.327917098999023,
+ "step": 136
+ },
+ {
+ "epoch": 1.908296943231441,
+ "grad_norm": 0.02056879736483097,
+ "learning_rate": 0.0006,
+ "loss": 7.3311991691589355,
+ "step": 137
+ },
+ {
+ "epoch": 1.9222707423580787,
+ "grad_norm": 0.016865141689777374,
+ "learning_rate": 0.0006,
+ "loss": 7.312766075134277,
+ "step": 138
+ },
+ {
+ "epoch": 1.936244541484716,
+ "grad_norm": 0.017549855634570122,
+ "learning_rate": 0.0006,
+ "loss": 7.31191349029541,
+ "step": 139
+ },
+ {
+ "epoch": 1.9502183406113538,
+ "grad_norm": 0.015930157154798508,
+ "learning_rate": 0.0006,
+ "loss": 7.330901622772217,
+ "step": 140
+ },
+ {
+ "epoch": 1.9641921397379911,
+ "grad_norm": 0.013283558189868927,
+ "learning_rate": 0.0006,
+ "loss": 7.298318386077881,
+ "step": 141
+ },
+ {
+ "epoch": 1.9781659388646289,
+ "grad_norm": 0.012023469433188438,
+ "learning_rate": 0.0006,
+ "loss": 7.282337188720703,
+ "step": 142
+ },
+ {
+ "epoch": 1.9921397379912664,
+ "grad_norm": 0.01459511648863554,
+ "learning_rate": 0.0006,
+ "loss": 7.282924175262451,
+ "step": 143
+ },
+ {
+ "epoch": 2.0,
+ "grad_norm": 0.017474235966801643,
+ "learning_rate": 0.0006,
+ "loss": 7.254333972930908,
+ "step": 144
+ },
+ {
+ "epoch": 2.0,
+ "eval_loss": 7.318708896636963,
+ "eval_runtime": 59.6555,
+ "eval_samples_per_second": 40.935,
+ "eval_steps_per_second": 1.291,
+ "step": 144
+ },
+ {
+ "epoch": 2.0139737991266378,
+ "grad_norm": 0.019757507368922234,
+ "learning_rate": 0.0006,
+ "loss": 7.30288553237915,
+ "step": 145
+ },
+ {
+ "epoch": 2.027947598253275,
+ "grad_norm": 0.014571248553693295,
+ "learning_rate": 0.0006,
+ "loss": 7.274681568145752,
+ "step": 146
+ },
+ {
+ "epoch": 2.041921397379913,
+ "grad_norm": 0.015744198113679886,
+ "learning_rate": 0.0006,
+ "loss": 7.278632640838623,
+ "step": 147
+ },
+ {
+ "epoch": 2.05589519650655,
+ "grad_norm": 0.028048858046531677,
+ "learning_rate": 0.0006,
+ "loss": 7.26558780670166,
+ "step": 148
+ },
+ {
+ "epoch": 2.069868995633188,
+ "grad_norm": 0.06547307223081589,
+ "learning_rate": 0.0006,
+ "loss": 7.32567834854126,
+ "step": 149
+ },
+ {
+ "epoch": 2.0838427947598253,
+ "grad_norm": 0.05969763547182083,
+ "learning_rate": 0.0006,
+ "loss": 7.395836353302002,
+ "step": 150
+ },
+ {
+ "epoch": 2.097816593886463,
+ "grad_norm": 0.032574281096458435,
+ "learning_rate": 0.0006,
+ "loss": 7.318150997161865,
+ "step": 151
+ },
+ {
+ "epoch": 2.1117903930131003,
+ "grad_norm": 0.029452962800860405,
+ "learning_rate": 0.0006,
+ "loss": 7.308066368103027,
+ "step": 152
+ },
+ {
+ "epoch": 2.125764192139738,
+ "grad_norm": 0.03161991387605667,
+ "learning_rate": 0.0006,
+ "loss": 7.306290626525879,
+ "step": 153
+ },
+ {
+ "epoch": 2.1397379912663754,
+ "grad_norm": 0.027096295729279518,
+ "learning_rate": 0.0006,
+ "loss": 7.297796249389648,
+ "step": 154
+ },
+ {
+ "epoch": 2.153711790393013,
+ "grad_norm": 0.028657056391239166,
+ "learning_rate": 0.0006,
+ "loss": 7.317424774169922,
+ "step": 155
+ },
+ {
+ "epoch": 2.1676855895196505,
+ "grad_norm": 0.02046995982527733,
+ "learning_rate": 0.0006,
+ "loss": 7.297986030578613,
+ "step": 156
+ },
+ {
+ "epoch": 2.1816593886462883,
+ "grad_norm": 0.02220369316637516,
+ "learning_rate": 0.0006,
+ "loss": 7.268237590789795,
+ "step": 157
+ },
+ {
+ "epoch": 2.1956331877729256,
+ "grad_norm": 0.018381357192993164,
+ "learning_rate": 0.0006,
+ "loss": 7.259415626525879,
+ "step": 158
+ },
+ {
+ "epoch": 2.2096069868995634,
+ "grad_norm": 0.0204413291066885,
+ "learning_rate": 0.0006,
+ "loss": 7.258173942565918,
+ "step": 159
+ },
+ {
+ "epoch": 2.2235807860262007,
+ "grad_norm": 0.015874288976192474,
+ "learning_rate": 0.0006,
+ "loss": 7.281834602355957,
+ "step": 160
+ },
+ {
+ "epoch": 2.2375545851528384,
+ "grad_norm": 0.01634068600833416,
+ "learning_rate": 0.0006,
+ "loss": 7.2469868659973145,
+ "step": 161
+ },
+ {
+ "epoch": 2.251528384279476,
+ "grad_norm": 0.01728747971355915,
+ "learning_rate": 0.0006,
+ "loss": 7.280995845794678,
+ "step": 162
+ },
+ {
+ "epoch": 2.2655021834061135,
+ "grad_norm": 0.016341188922524452,
+ "learning_rate": 0.0006,
+ "loss": 7.273357391357422,
+ "step": 163
+ },
+ {
+ "epoch": 2.279475982532751,
+ "grad_norm": 0.016125807538628578,
+ "learning_rate": 0.0006,
+ "loss": 7.273510932922363,
+ "step": 164
+ },
+ {
+ "epoch": 2.2934497816593886,
+ "grad_norm": 0.01629287749528885,
+ "learning_rate": 0.0006,
+ "loss": 7.267029762268066,
+ "step": 165
+ },
+ {
+ "epoch": 2.3074235807860264,
+ "grad_norm": 0.015296131372451782,
+ "learning_rate": 0.0006,
+ "loss": 7.211989402770996,
+ "step": 166
+ },
+ {
+ "epoch": 2.3213973799126637,
+ "grad_norm": 0.01709570363163948,
+ "learning_rate": 0.0006,
+ "loss": 7.242851734161377,
+ "step": 167
+ },
+ {
+ "epoch": 2.3353711790393015,
+ "grad_norm": 0.012552388943731785,
+ "learning_rate": 0.0006,
+ "loss": 7.241696834564209,
+ "step": 168
+ },
+ {
+ "epoch": 2.349344978165939,
+ "grad_norm": 0.014571291394531727,
+ "learning_rate": 0.0006,
+ "loss": 7.270681381225586,
+ "step": 169
+ },
+ {
+ "epoch": 2.3633187772925766,
+ "grad_norm": 0.013389437459409237,
+ "learning_rate": 0.0006,
+ "loss": 7.191600799560547,
+ "step": 170
+ },
+ {
+ "epoch": 2.377292576419214,
+ "grad_norm": 0.011895314790308475,
+ "learning_rate": 0.0006,
+ "loss": 7.213610649108887,
+ "step": 171
+ },
+ {
+ "epoch": 2.3912663755458516,
+ "grad_norm": 0.01345642190426588,
+ "learning_rate": 0.0006,
+ "loss": 7.193787574768066,
+ "step": 172
+ },
+ {
+ "epoch": 2.405240174672489,
+ "grad_norm": 0.0091794328764081,
+ "learning_rate": 0.0006,
+ "loss": 7.233364105224609,
+ "step": 173
+ },
+ {
+ "epoch": 2.4192139737991267,
+ "grad_norm": 0.013767086900770664,
+ "learning_rate": 0.0006,
+ "loss": 7.217272758483887,
+ "step": 174
+ },
+ {
+ "epoch": 2.433187772925764,
+ "grad_norm": 0.015134960412979126,
+ "learning_rate": 0.0006,
+ "loss": 7.192791938781738,
+ "step": 175
+ },
+ {
+ "epoch": 2.447161572052402,
+ "grad_norm": 0.01790648326277733,
+ "learning_rate": 0.0006,
+ "loss": 7.229467391967773,
+ "step": 176
+ },
+ {
+ "epoch": 2.461135371179039,
+ "grad_norm": 0.015362377278506756,
+ "learning_rate": 0.0006,
+ "loss": 7.180346488952637,
+ "step": 177
+ },
+ {
+ "epoch": 2.475109170305677,
+ "grad_norm": 0.010505498386919498,
+ "learning_rate": 0.0006,
+ "loss": 7.188833236694336,
+ "step": 178
+ },
+ {
+ "epoch": 2.489082969432314,
+ "grad_norm": 0.009129747748374939,
+ "learning_rate": 0.0006,
+ "loss": 7.203153133392334,
+ "step": 179
+ },
+ {
+ "epoch": 2.503056768558952,
+ "grad_norm": 0.01147883478552103,
+ "learning_rate": 0.0006,
+ "loss": 7.163824558258057,
+ "step": 180
+ },
+ {
+ "epoch": 2.5170305676855893,
+ "grad_norm": 0.011338942684233189,
+ "learning_rate": 0.0006,
+ "loss": 7.16287899017334,
+ "step": 181
+ },
+ {
+ "epoch": 2.531004366812227,
+ "grad_norm": 0.013277675025165081,
+ "learning_rate": 0.0006,
+ "loss": 7.155905723571777,
+ "step": 182
+ },
+ {
+ "epoch": 2.544978165938865,
+ "grad_norm": 0.015269347466528416,
+ "learning_rate": 0.0006,
+ "loss": 7.189371109008789,
+ "step": 183
+ },
+ {
+ "epoch": 2.558951965065502,
+ "grad_norm": 0.017905596643686295,
+ "learning_rate": 0.0006,
+ "loss": 7.177547931671143,
+ "step": 184
+ },
+ {
+ "epoch": 2.5729257641921395,
+ "grad_norm": 0.024431610479950905,
+ "learning_rate": 0.0006,
+ "loss": 7.192169189453125,
+ "step": 185
+ },
+ {
+ "epoch": 2.5868995633187772,
+ "grad_norm": 0.03656961768865585,
+ "learning_rate": 0.0006,
+ "loss": 7.124387741088867,
+ "step": 186
+ },
+ {
+ "epoch": 2.600873362445415,
+ "grad_norm": 0.04337048903107643,
+ "learning_rate": 0.0006,
+ "loss": 7.140231609344482,
+ "step": 187
+ },
+ {
+ "epoch": 2.6148471615720523,
+ "grad_norm": 0.04047228768467903,
+ "learning_rate": 0.0006,
+ "loss": 7.171927452087402,
+ "step": 188
+ },
+ {
+ "epoch": 2.62882096069869,
+ "grad_norm": 0.07825770974159241,
+ "learning_rate": 0.0006,
+ "loss": 7.207631587982178,
+ "step": 189
+ },
+ {
+ "epoch": 2.6427947598253274,
+ "grad_norm": 0.07118063420057297,
+ "learning_rate": 0.0006,
+ "loss": 7.209488391876221,
+ "step": 190
+ },
+ {
+ "epoch": 2.656768558951965,
+ "grad_norm": 0.05153293535113335,
+ "learning_rate": 0.0006,
+ "loss": 7.14793062210083,
+ "step": 191
+ },
+ {
+ "epoch": 2.6707423580786025,
+ "grad_norm": 0.03318839520215988,
+ "learning_rate": 0.0006,
+ "loss": 7.163010597229004,
+ "step": 192
+ },
+ {
+ "epoch": 2.6847161572052403,
+ "grad_norm": 0.029333539307117462,
+ "learning_rate": 0.0006,
+ "loss": 7.159955978393555,
+ "step": 193
+ },
+ {
+ "epoch": 2.6986899563318776,
+ "grad_norm": 0.029405880719423294,
+ "learning_rate": 0.0006,
+ "loss": 7.136178493499756,
+ "step": 194
+ },
+ {
+ "epoch": 2.7126637554585153,
+ "grad_norm": 0.037815988063812256,
+ "learning_rate": 0.0006,
+ "loss": 7.183186054229736,
+ "step": 195
+ },
+ {
+ "epoch": 2.726637554585153,
+ "grad_norm": 0.020307250320911407,
+ "learning_rate": 0.0006,
+ "loss": 7.133164882659912,
+ "step": 196
+ },
+ {
+ "epoch": 2.7406113537117904,
+ "grad_norm": 0.026048069819808006,
+ "learning_rate": 0.0006,
+ "loss": 7.124977111816406,
+ "step": 197
+ },
+ {
+ "epoch": 2.7545851528384278,
+ "grad_norm": 0.027316724881529808,
+ "learning_rate": 0.0006,
+ "loss": 7.07265567779541,
+ "step": 198
+ },
+ {
+ "epoch": 2.7685589519650655,
+ "grad_norm": 0.021629009395837784,
+ "learning_rate": 0.0006,
+ "loss": 7.150020599365234,
+ "step": 199
+ },
+ {
+ "epoch": 2.7825327510917033,
+ "grad_norm": 0.01605929434299469,
+ "learning_rate": 0.0006,
+ "loss": 7.10919189453125,
+ "step": 200
+ },
+ {
+ "epoch": 2.7965065502183406,
+ "grad_norm": 0.020800089463591576,
+ "learning_rate": 0.0006,
+ "loss": 7.132045269012451,
+ "step": 201
+ },
+ {
+ "epoch": 2.810480349344978,
+ "grad_norm": 0.019225774332880974,
+ "learning_rate": 0.0006,
+ "loss": 7.128364562988281,
+ "step": 202
+ },
+ {
+ "epoch": 2.8244541484716157,
+ "grad_norm": 0.01561372634023428,
+ "learning_rate": 0.0006,
+ "loss": 7.083669662475586,
+ "step": 203
+ },
+ {
+ "epoch": 2.8384279475982535,
+ "grad_norm": 0.021969519555568695,
+ "learning_rate": 0.0006,
+ "loss": 7.050841331481934,
+ "step": 204
+ },
+ {
+ "epoch": 2.8524017467248908,
+ "grad_norm": 0.021453695371747017,
+ "learning_rate": 0.0006,
+ "loss": 7.097085475921631,
+ "step": 205
+ },
+ {
+ "epoch": 2.8663755458515285,
+ "grad_norm": 0.022836022078990936,
+ "learning_rate": 0.0006,
+ "loss": 7.10405969619751,
+ "step": 206
+ },
+ {
+ "epoch": 2.880349344978166,
+ "grad_norm": 0.02430851384997368,
+ "learning_rate": 0.0006,
+ "loss": 7.045181751251221,
+ "step": 207
+ },
+ {
+ "epoch": 2.8943231441048036,
+ "grad_norm": 0.020166993141174316,
+ "learning_rate": 0.0006,
+ "loss": 7.068183898925781,
+ "step": 208
+ },
+ {
+ "epoch": 2.908296943231441,
+ "grad_norm": 0.0118505684658885,
+ "learning_rate": 0.0006,
+ "loss": 7.07132625579834,
+ "step": 209
+ },
+ {
+ "epoch": 2.9222707423580787,
+ "grad_norm": 0.023772427812218666,
+ "learning_rate": 0.0006,
+ "loss": 7.055238723754883,
+ "step": 210
+ },
+ {
+ "epoch": 2.936244541484716,
+ "grad_norm": 0.02350712940096855,
+ "learning_rate": 0.0006,
+ "loss": 7.055898666381836,
+ "step": 211
+ },
+ {
+ "epoch": 2.950218340611354,
+ "grad_norm": 0.017244907096028328,
+ "learning_rate": 0.0006,
+ "loss": 7.0099711418151855,
+ "step": 212
+ },
+ {
+ "epoch": 2.964192139737991,
+ "grad_norm": 0.021565575152635574,
+ "learning_rate": 0.0006,
+ "loss": 6.991751670837402,
+ "step": 213
+ },
+ {
+ "epoch": 2.978165938864629,
+ "grad_norm": 0.03241860866546631,
+ "learning_rate": 0.0006,
+ "loss": 7.005980491638184,
+ "step": 214
+ },
+ {
+ "epoch": 2.992139737991266,
+ "grad_norm": 0.04613726586103439,
+ "learning_rate": 0.0006,
+ "loss": 7.027859687805176,
+ "step": 215
+ },
+ {
+ "epoch": 3.0,
+ "grad_norm": 0.06404894590377808,
+ "learning_rate": 0.0006,
+ "loss": 7.123725891113281,
+ "step": 216
+ },
+ {
+ "epoch": 3.0,
+ "eval_loss": 7.070647716522217,
+ "eval_runtime": 58.5801,
+ "eval_samples_per_second": 41.687,
+ "eval_steps_per_second": 1.314,
+ "step": 216
+ },
+ {
+ "epoch": 3.0139737991266378,
+ "grad_norm": 0.05542432889342308,
+ "learning_rate": 0.0006,
+ "loss": 7.078299522399902,
+ "step": 217
+ },
+ {
+ "epoch": 3.027947598253275,
+ "grad_norm": 0.026484526693820953,
+ "learning_rate": 0.0006,
+ "loss": 6.961159706115723,
+ "step": 218
+ },
+ {
+ "epoch": 3.041921397379913,
+ "grad_norm": 0.028276391327381134,
+ "learning_rate": 0.0006,
+ "loss": 7.019961833953857,
+ "step": 219
+ },
+ {
+ "epoch": 3.05589519650655,
+ "grad_norm": 0.024486595764756203,
+ "learning_rate": 0.0006,
+ "loss": 7.039677619934082,
+ "step": 220
+ },
+ {
+ "epoch": 3.069868995633188,
+ "grad_norm": 0.017620893195271492,
+ "learning_rate": 0.0006,
+ "loss": 6.9270782470703125,
+ "step": 221
+ },
+ {
+ "epoch": 3.0838427947598253,
+ "grad_norm": 0.02547302283346653,
+ "learning_rate": 0.0006,
+ "loss": 6.924233436584473,
+ "step": 222
+ },
+ {
+ "epoch": 3.097816593886463,
+ "grad_norm": 0.02539009042084217,
+ "learning_rate": 0.0006,
+ "loss": 6.973979949951172,
+ "step": 223
+ },
+ {
+ "epoch": 3.1117903930131003,
+ "grad_norm": 0.03256227448582649,
+ "learning_rate": 0.0006,
+ "loss": 6.933725357055664,
+ "step": 224
+ },
+ {
+ "epoch": 3.125764192139738,
+ "grad_norm": 0.04651800915598869,
+ "learning_rate": 0.0006,
+ "loss": 6.951044082641602,
+ "step": 225
+ },
+ {
+ "epoch": 3.1397379912663754,
+ "grad_norm": 0.04822975769639015,
+ "learning_rate": 0.0006,
+ "loss": 6.9614410400390625,
+ "step": 226
+ },
+ {
+ "epoch": 3.153711790393013,
+ "grad_norm": 0.031165020540356636,
+ "learning_rate": 0.0006,
+ "loss": 6.91193151473999,
+ "step": 227
+ },
+ {
+ "epoch": 3.1676855895196505,
+ "grad_norm": 0.029754292219877243,
+ "learning_rate": 0.0006,
+ "loss": 6.91802978515625,
+ "step": 228
+ },
+ {
+ "epoch": 3.1816593886462883,
+ "grad_norm": 0.02193152718245983,
+ "learning_rate": 0.0006,
+ "loss": 6.928090572357178,
+ "step": 229
+ },
+ {
+ "epoch": 3.1956331877729256,
+ "grad_norm": 0.02428470179438591,
+ "learning_rate": 0.0006,
+ "loss": 6.90396785736084,
+ "step": 230
+ },
+ {
+ "epoch": 3.2096069868995634,
+ "grad_norm": 0.023374345153570175,
+ "learning_rate": 0.0006,
+ "loss": 6.885858535766602,
+ "step": 231
+ },
+ {
+ "epoch": 3.2235807860262007,
+ "grad_norm": 0.023405000567436218,
+ "learning_rate": 0.0006,
+ "loss": 6.8855767250061035,
+ "step": 232
+ },
+ {
+ "epoch": 3.2375545851528384,
+ "grad_norm": 0.017277134582400322,
+ "learning_rate": 0.0006,
+ "loss": 6.901449203491211,
+ "step": 233
+ },
+ {
+ "epoch": 3.251528384279476,
+ "grad_norm": 0.019496750086545944,
+ "learning_rate": 0.0006,
+ "loss": 6.885687828063965,
+ "step": 234
+ },
+ {
+ "epoch": 3.2655021834061135,
+ "grad_norm": 0.016893276944756508,
+ "learning_rate": 0.0006,
+ "loss": 6.872369766235352,
+ "step": 235
+ },
+ {
+ "epoch": 3.279475982532751,
+ "grad_norm": 0.02058715932071209,
+ "learning_rate": 0.0006,
+ "loss": 6.823280334472656,
+ "step": 236
+ },
+ {
+ "epoch": 3.2934497816593886,
+ "grad_norm": 0.015530755743384361,
+ "learning_rate": 0.0006,
+ "loss": 6.860499858856201,
+ "step": 237
+ },
+ {
+ "epoch": 3.3074235807860264,
+ "grad_norm": 0.01934926211833954,
+ "learning_rate": 0.0006,
+ "loss": 6.817448616027832,
+ "step": 238
+ },
+ {
+ "epoch": 3.3213973799126637,
+ "grad_norm": 0.017413552850484848,
+ "learning_rate": 0.0006,
+ "loss": 6.881955146789551,
+ "step": 239
+ },
+ {
+ "epoch": 3.3353711790393015,
+ "grad_norm": 0.026478875428438187,
+ "learning_rate": 0.0006,
+ "loss": 6.827404499053955,
+ "step": 240
+ },
+ {
+ "epoch": 3.349344978165939,
+ "grad_norm": 0.032611701637506485,
+ "learning_rate": 0.0006,
+ "loss": 6.78285551071167,
+ "step": 241
+ },
+ {
+ "epoch": 3.3633187772925766,
+ "grad_norm": 0.041900306940078735,
+ "learning_rate": 0.0006,
+ "loss": 6.783053874969482,
+ "step": 242
+ },
+ {
+ "epoch": 3.377292576419214,
+ "grad_norm": 0.04149497672915459,
+ "learning_rate": 0.0006,
+ "loss": 6.745220184326172,
+ "step": 243
+ },
+ {
+ "epoch": 3.3912663755458516,
+ "grad_norm": 0.023949826136231422,
+ "learning_rate": 0.0006,
+ "loss": 6.841533184051514,
+ "step": 244
+ },
+ {
+ "epoch": 3.405240174672489,
+ "grad_norm": 0.028601842001080513,
+ "learning_rate": 0.0006,
+ "loss": 6.852852821350098,
+ "step": 245
+ },
+ {
+ "epoch": 3.4192139737991267,
+ "grad_norm": 0.027365155518054962,
+ "learning_rate": 0.0006,
+ "loss": 6.74376106262207,
+ "step": 246
+ },
+ {
+ "epoch": 3.433187772925764,
+ "grad_norm": 0.02856568619608879,
+ "learning_rate": 0.0006,
+ "loss": 6.74906063079834,
+ "step": 247
+ },
+ {
+ "epoch": 3.447161572052402,
+ "grad_norm": 0.019172416999936104,
+ "learning_rate": 0.0006,
+ "loss": 6.809206008911133,
+ "step": 248
+ },
+ {
+ "epoch": 3.461135371179039,
+ "grad_norm": 0.02161746844649315,
+ "learning_rate": 0.0006,
+ "loss": 6.789888381958008,
+ "step": 249
+ },
+ {
+ "epoch": 3.475109170305677,
+ "grad_norm": 0.020273666828870773,
+ "learning_rate": 0.0006,
+ "loss": 6.73512077331543,
+ "step": 250
+ },
+ {
+ "epoch": 3.489082969432314,
+ "grad_norm": 0.022329283878207207,
+ "learning_rate": 0.0006,
+ "loss": 6.826148509979248,
+ "step": 251
+ },
+ {
+ "epoch": 3.503056768558952,
+ "grad_norm": 0.02284284122288227,
+ "learning_rate": 0.0006,
+ "loss": 6.748521327972412,
+ "step": 252
+ },
+ {
+ "epoch": 3.5170305676855893,
+ "grad_norm": 0.026368247345089912,
+ "learning_rate": 0.0006,
+ "loss": 6.645879745483398,
+ "step": 253
+ },
+ {
+ "epoch": 3.531004366812227,
+ "grad_norm": 0.03627244383096695,
+ "learning_rate": 0.0006,
+ "loss": 6.666064739227295,
+ "step": 254
+ },
+ {
+ "epoch": 3.544978165938865,
+ "grad_norm": 0.04140935093164444,
+ "learning_rate": 0.0006,
+ "loss": 6.712275981903076,
+ "step": 255
+ },
+ {
+ "epoch": 3.558951965065502,
+ "grad_norm": 0.0390462800860405,
+ "learning_rate": 0.0006,
+ "loss": 6.665900230407715,
+ "step": 256
+ },
+ {
+ "epoch": 3.5729257641921395,
+ "grad_norm": 0.05178583785891533,
+ "learning_rate": 0.0006,
+ "loss": 6.753298759460449,
+ "step": 257
+ },
+ {
+ "epoch": 3.5868995633187772,
+ "grad_norm": 0.054328061640262604,
+ "learning_rate": 0.0006,
+ "loss": 6.732633113861084,
+ "step": 258
+ },
+ {
+ "epoch": 3.600873362445415,
+ "grad_norm": 0.04590460658073425,
+ "learning_rate": 0.0006,
+ "loss": 6.65250301361084,
+ "step": 259
+ },
+ {
+ "epoch": 3.6148471615720523,
+ "grad_norm": 0.044331666082143784,
+ "learning_rate": 0.0006,
+ "loss": 6.68654203414917,
+ "step": 260
+ },
+ {
+ "epoch": 3.62882096069869,
+ "grad_norm": 0.05308730527758598,
+ "learning_rate": 0.0006,
+ "loss": 6.642482757568359,
+ "step": 261
+ },
+ {
+ "epoch": 3.6427947598253274,
+ "grad_norm": 0.03999016433954239,
+ "learning_rate": 0.0006,
+ "loss": 6.625227451324463,
+ "step": 262
+ },
+ {
+ "epoch": 3.656768558951965,
+ "grad_norm": 0.04088086634874344,
+ "learning_rate": 0.0006,
+ "loss": 6.67623233795166,
+ "step": 263
+ },
+ {
+ "epoch": 3.6707423580786025,
+ "grad_norm": 0.03615636005997658,
+ "learning_rate": 0.0006,
+ "loss": 6.6869001388549805,
+ "step": 264
+ },
+ {
+ "epoch": 3.6847161572052403,
+ "grad_norm": 0.02817360684275627,
+ "learning_rate": 0.0006,
+ "loss": 6.674601078033447,
+ "step": 265
+ },
+ {
+ "epoch": 3.6986899563318776,
+ "grad_norm": 0.032586123794317245,
+ "learning_rate": 0.0006,
+ "loss": 6.612994194030762,
+ "step": 266
+ },
+ {
+ "epoch": 3.7126637554585153,
+ "grad_norm": 0.031117867678403854,
+ "learning_rate": 0.0006,
+ "loss": 6.6244354248046875,
+ "step": 267
+ },
+ {
+ "epoch": 3.726637554585153,
+ "grad_norm": 0.028330687433481216,
+ "learning_rate": 0.0006,
+ "loss": 6.669557094573975,
+ "step": 268
+ },
+ {
+ "epoch": 3.7406113537117904,
+ "grad_norm": 0.021489202976226807,
+ "learning_rate": 0.0006,
+ "loss": 6.632035732269287,
+ "step": 269
+ },
+ {
+ "epoch": 3.7545851528384278,
+ "grad_norm": 0.024269424378871918,
+ "learning_rate": 0.0006,
+ "loss": 6.660867691040039,
+ "step": 270
+ },
+ {
+ "epoch": 3.7685589519650655,
+ "grad_norm": 0.02237873338162899,
+ "learning_rate": 0.0006,
+ "loss": 6.602587699890137,
+ "step": 271
+ },
+ {
+ "epoch": 3.7825327510917033,
+ "grad_norm": 0.026503929868340492,
+ "learning_rate": 0.0006,
+ "loss": 6.599200248718262,
+ "step": 272
+ },
+ {
+ "epoch": 3.7965065502183406,
+ "grad_norm": 0.03687124326825142,
+ "learning_rate": 0.0006,
+ "loss": 6.567296504974365,
+ "step": 273
+ },
+ {
+ "epoch": 3.810480349344978,
+ "grad_norm": 0.04708952456712723,
+ "learning_rate": 0.0006,
+ "loss": 6.621265411376953,
+ "step": 274
+ },
+ {
+ "epoch": 3.8244541484716157,
+ "grad_norm": 0.041846614331007004,
+ "learning_rate": 0.0006,
+ "loss": 6.550536155700684,
+ "step": 275
+ },
+ {
+ "epoch": 3.8384279475982535,
+ "grad_norm": 0.031249675899744034,
+ "learning_rate": 0.0006,
+ "loss": 6.54278564453125,
+ "step": 276
+ },
+ {
+ "epoch": 3.8524017467248908,
+ "grad_norm": 0.023082256317138672,
+ "learning_rate": 0.0006,
+ "loss": 6.544787406921387,
+ "step": 277
+ },
+ {
+ "epoch": 3.8663755458515285,
+ "grad_norm": 0.020995570346713066,
+ "learning_rate": 0.0006,
+ "loss": 6.518143653869629,
+ "step": 278
+ },
+ {
+ "epoch": 3.880349344978166,
+ "grad_norm": 0.020839324221014977,
+ "learning_rate": 0.0006,
+ "loss": 6.546667098999023,
+ "step": 279
+ },
+ {
+ "epoch": 3.8943231441048036,
+ "grad_norm": 0.018377432599663734,
+ "learning_rate": 0.0006,
+ "loss": 6.478307723999023,
+ "step": 280
+ },
+ {
+ "epoch": 3.908296943231441,
+ "grad_norm": 0.021713724359869957,
+ "learning_rate": 0.0006,
+ "loss": 6.485574245452881,
+ "step": 281
+ },
+ {
+ "epoch": 3.9222707423580787,
+ "grad_norm": 0.01359301246702671,
+ "learning_rate": 0.0006,
+ "loss": 6.570268630981445,
+ "step": 282
+ },
+ {
+ "epoch": 3.936244541484716,
+ "grad_norm": 0.016233332455158234,
+ "learning_rate": 0.0006,
+ "loss": 6.57150936126709,
+ "step": 283
+ },
+ {
+ "epoch": 3.950218340611354,
+ "grad_norm": 0.020687608048319817,
+ "learning_rate": 0.0006,
+ "loss": 6.527956008911133,
+ "step": 284
+ },
+ {
+ "epoch": 3.964192139737991,
+ "grad_norm": 0.02032964862883091,
+ "learning_rate": 0.0006,
+ "loss": 6.468169689178467,
+ "step": 285
+ },
+ {
+ "epoch": 3.978165938864629,
+ "grad_norm": 0.016676288098096848,
+ "learning_rate": 0.0006,
+ "loss": 6.448418617248535,
+ "step": 286
+ },
+ {
+ "epoch": 3.992139737991266,
+ "grad_norm": 0.0168539360165596,
+ "learning_rate": 0.0006,
+ "loss": 6.461180686950684,
+ "step": 287
+ },
+ {
+ "epoch": 4.0,
+ "grad_norm": 0.02666233666241169,
+ "learning_rate": 0.0006,
+ "loss": 6.344893932342529,
+ "step": 288
+ },
+ {
+ "epoch": 4.0,
+ "eval_loss": 6.5384979248046875,
+ "eval_runtime": 58.814,
+ "eval_samples_per_second": 41.521,
+ "eval_steps_per_second": 1.309,
+ "step": 288
+ },
+ {
+ "epoch": 4.013973799126638,
+ "grad_norm": 0.04154708981513977,
+ "learning_rate": 0.0006,
+ "loss": 6.41855525970459,
+ "step": 289
+ },
+ {
+ "epoch": 4.0279475982532755,
+ "grad_norm": 0.04895668104290962,
+ "learning_rate": 0.0006,
+ "loss": 6.499302864074707,
+ "step": 290
+ },
+ {
+ "epoch": 4.041921397379912,
+ "grad_norm": 0.034464482218027115,
+ "learning_rate": 0.0006,
+ "loss": 6.438611030578613,
+ "step": 291
+ },
+ {
+ "epoch": 4.05589519650655,
+ "grad_norm": 0.06845773756504059,
+ "learning_rate": 0.0006,
+ "loss": 6.538361072540283,
+ "step": 292
+ },
+ {
+ "epoch": 4.069868995633188,
+ "grad_norm": 0.053768858313560486,
+ "learning_rate": 0.0006,
+ "loss": 6.527524948120117,
+ "step": 293
+ },
+ {
+ "epoch": 4.083842794759826,
+ "grad_norm": 0.039037931710481644,
+ "learning_rate": 0.0006,
+ "loss": 6.503847122192383,
+ "step": 294
+ },
+ {
+ "epoch": 4.097816593886463,
+ "grad_norm": 0.03962196782231331,
+ "learning_rate": 0.0006,
+ "loss": 6.475178241729736,
+ "step": 295
+ },
+ {
+ "epoch": 4.1117903930131,
+ "grad_norm": 0.049101535230875015,
+ "learning_rate": 0.0006,
+ "loss": 6.474587917327881,
+ "step": 296
+ },
+ {
+ "epoch": 4.125764192139738,
+ "grad_norm": 0.02830282226204872,
+ "learning_rate": 0.0006,
+ "loss": 6.45065975189209,
+ "step": 297
+ },
+ {
+ "epoch": 4.139737991266376,
+ "grad_norm": 0.023153482005000114,
+ "learning_rate": 0.0006,
+ "loss": 6.5218706130981445,
+ "step": 298
+ },
+ {
+ "epoch": 4.153711790393013,
+ "grad_norm": 0.025064002722501755,
+ "learning_rate": 0.0006,
+ "loss": 6.456475257873535,
+ "step": 299
+ },
+ {
+ "epoch": 4.1676855895196505,
+ "grad_norm": 0.02357092685997486,
+ "learning_rate": 0.0006,
+ "loss": 6.485952377319336,
+ "step": 300
+ },
+ {
+ "epoch": 4.181659388646288,
+ "grad_norm": 0.021832305938005447,
+ "learning_rate": 0.0006,
+ "loss": 6.425506591796875,
+ "step": 301
+ },
+ {
+ "epoch": 4.195633187772926,
+ "grad_norm": 0.0227900929749012,
+ "learning_rate": 0.0006,
+ "loss": 6.439155578613281,
+ "step": 302
+ },
+ {
+ "epoch": 4.209606986899563,
+ "grad_norm": 0.015961607918143272,
+ "learning_rate": 0.0006,
+ "loss": 6.460862636566162,
+ "step": 303
+ },
+ {
+ "epoch": 4.223580786026201,
+ "grad_norm": 0.017876390367746353,
+ "learning_rate": 0.0006,
+ "loss": 6.44564151763916,
+ "step": 304
+ },
+ {
+ "epoch": 4.2375545851528384,
+ "grad_norm": 0.013703204691410065,
+ "learning_rate": 0.0006,
+ "loss": 6.418414115905762,
+ "step": 305
+ },
+ {
+ "epoch": 4.251528384279476,
+ "grad_norm": 0.015240306034684181,
+ "learning_rate": 0.0006,
+ "loss": 6.411238670349121,
+ "step": 306
+ },
+ {
+ "epoch": 4.265502183406113,
+ "grad_norm": 0.01587662845849991,
+ "learning_rate": 0.0006,
+ "loss": 6.395642280578613,
+ "step": 307
+ },
+ {
+ "epoch": 4.279475982532751,
+ "grad_norm": 0.014650700613856316,
+ "learning_rate": 0.0006,
+ "loss": 6.302793025970459,
+ "step": 308
+ },
+ {
+ "epoch": 4.293449781659389,
+ "grad_norm": 0.016533225774765015,
+ "learning_rate": 0.0006,
+ "loss": 6.380588054656982,
+ "step": 309
+ },
+ {
+ "epoch": 4.307423580786026,
+ "grad_norm": 0.019963741302490234,
+ "learning_rate": 0.0006,
+ "loss": 6.345336437225342,
+ "step": 310
+ },
+ {
+ "epoch": 4.321397379912664,
+ "grad_norm": 0.02237936295568943,
+ "learning_rate": 0.0006,
+ "loss": 6.387770652770996,
+ "step": 311
+ },
+ {
+ "epoch": 4.335371179039301,
+ "grad_norm": 0.026658328250050545,
+ "learning_rate": 0.0006,
+ "loss": 6.323662757873535,
+ "step": 312
+ },
+ {
+ "epoch": 4.349344978165939,
+ "grad_norm": 0.02852284163236618,
+ "learning_rate": 0.0006,
+ "loss": 6.382204055786133,
+ "step": 313
+ },
+ {
+ "epoch": 4.3633187772925766,
+ "grad_norm": 0.028815090656280518,
+ "learning_rate": 0.0006,
+ "loss": 6.402390480041504,
+ "step": 314
+ },
+ {
+ "epoch": 4.377292576419214,
+ "grad_norm": 0.029393529519438744,
+ "learning_rate": 0.0006,
+ "loss": 6.280278205871582,
+ "step": 315
+ },
+ {
+ "epoch": 4.391266375545851,
+ "grad_norm": 0.02186041697859764,
+ "learning_rate": 0.0006,
+ "loss": 6.302707672119141,
+ "step": 316
+ },
+ {
+ "epoch": 4.405240174672489,
+ "grad_norm": 0.02071218006312847,
+ "learning_rate": 0.0006,
+ "loss": 6.3365020751953125,
+ "step": 317
+ },
+ {
+ "epoch": 4.419213973799127,
+ "grad_norm": 0.02779117226600647,
+ "learning_rate": 0.0006,
+ "loss": 6.252157211303711,
+ "step": 318
+ },
+ {
+ "epoch": 4.4331877729257645,
+ "grad_norm": 0.03470654785633087,
+ "learning_rate": 0.0006,
+ "loss": 6.319070339202881,
+ "step": 319
+ },
+ {
+ "epoch": 4.447161572052401,
+ "grad_norm": 0.037067804485559464,
+ "learning_rate": 0.0006,
+ "loss": 6.314116477966309,
+ "step": 320
+ },
+ {
+ "epoch": 4.461135371179039,
+ "grad_norm": 0.032182756811380386,
+ "learning_rate": 0.0006,
+ "loss": 6.267298698425293,
+ "step": 321
+ },
+ {
+ "epoch": 4.475109170305677,
+ "grad_norm": 0.026305217295885086,
+ "learning_rate": 0.0006,
+ "loss": 6.331688404083252,
+ "step": 322
+ },
+ {
+ "epoch": 4.489082969432315,
+ "grad_norm": 0.027228357270359993,
+ "learning_rate": 0.0006,
+ "loss": 6.31139612197876,
+ "step": 323
+ },
+ {
+ "epoch": 4.503056768558952,
+ "grad_norm": 0.025781169533729553,
+ "learning_rate": 0.0006,
+ "loss": 6.194684982299805,
+ "step": 324
+ },
+ {
+ "epoch": 4.517030567685589,
+ "grad_norm": 0.029186977073550224,
+ "learning_rate": 0.0006,
+ "loss": 6.31306266784668,
+ "step": 325
+ },
+ {
+ "epoch": 4.531004366812227,
+ "grad_norm": 0.023608777672052383,
+ "learning_rate": 0.0006,
+ "loss": 6.285243034362793,
+ "step": 326
+ },
+ {
+ "epoch": 4.544978165938865,
+ "grad_norm": 0.02267594076693058,
+ "learning_rate": 0.0006,
+ "loss": 6.30342960357666,
+ "step": 327
+ },
+ {
+ "epoch": 4.558951965065502,
+ "grad_norm": 0.01926310732960701,
+ "learning_rate": 0.0006,
+ "loss": 6.280606269836426,
+ "step": 328
+ },
+ {
+ "epoch": 4.5729257641921395,
+ "grad_norm": 0.026484837755560875,
+ "learning_rate": 0.0006,
+ "loss": 6.237980365753174,
+ "step": 329
+ },
+ {
+ "epoch": 4.586899563318777,
+ "grad_norm": 0.026228854432702065,
+ "learning_rate": 0.0006,
+ "loss": 6.3018293380737305,
+ "step": 330
+ },
+ {
+ "epoch": 4.600873362445415,
+ "grad_norm": 0.022096967324614525,
+ "learning_rate": 0.0006,
+ "loss": 6.253863334655762,
+ "step": 331
+ },
+ {
+ "epoch": 4.614847161572053,
+ "grad_norm": 0.027223890647292137,
+ "learning_rate": 0.0006,
+ "loss": 6.166882038116455,
+ "step": 332
+ },
+ {
+ "epoch": 4.62882096069869,
+ "grad_norm": 0.03160124272108078,
+ "learning_rate": 0.0006,
+ "loss": 6.230309963226318,
+ "step": 333
+ },
+ {
+ "epoch": 4.642794759825327,
+ "grad_norm": 0.04110539332032204,
+ "learning_rate": 0.0006,
+ "loss": 6.2334818840026855,
+ "step": 334
+ },
+ {
+ "epoch": 4.656768558951965,
+ "grad_norm": 0.04355933889746666,
+ "learning_rate": 0.0006,
+ "loss": 6.160956382751465,
+ "step": 335
+ },
+ {
+ "epoch": 4.670742358078603,
+ "grad_norm": 0.03611086308956146,
+ "learning_rate": 0.0006,
+ "loss": 6.23399543762207,
+ "step": 336
+ },
+ {
+ "epoch": 4.68471615720524,
+ "grad_norm": 0.051811400800943375,
+ "learning_rate": 0.0006,
+ "loss": 6.241855144500732,
+ "step": 337
+ },
+ {
+ "epoch": 4.698689956331878,
+ "grad_norm": 0.057434193789958954,
+ "learning_rate": 0.0006,
+ "loss": 6.220993995666504,
+ "step": 338
+ },
+ {
+ "epoch": 4.712663755458515,
+ "grad_norm": 0.04420953616499901,
+ "learning_rate": 0.0006,
+ "loss": 6.300461292266846,
+ "step": 339
+ },
+ {
+ "epoch": 4.726637554585153,
+ "grad_norm": 0.03515457361936569,
+ "learning_rate": 0.0006,
+ "loss": 6.269349098205566,
+ "step": 340
+ },
+ {
+ "epoch": 4.74061135371179,
+ "grad_norm": 0.03222600743174553,
+ "learning_rate": 0.0006,
+ "loss": 6.248613357543945,
+ "step": 341
+ },
+ {
+ "epoch": 4.754585152838428,
+ "grad_norm": 0.0384046845138073,
+ "learning_rate": 0.0006,
+ "loss": 6.266979217529297,
+ "step": 342
+ },
+ {
+ "epoch": 4.7685589519650655,
+ "grad_norm": 0.0514867827296257,
+ "learning_rate": 0.0006,
+ "loss": 6.234103202819824,
+ "step": 343
+ },
+ {
+ "epoch": 4.782532751091703,
+ "grad_norm": 0.050747793167829514,
+ "learning_rate": 0.0006,
+ "loss": 6.322582244873047,
+ "step": 344
+ },
+ {
+ "epoch": 4.796506550218341,
+ "grad_norm": 0.040878988802433014,
+ "learning_rate": 0.0006,
+ "loss": 6.212509632110596,
+ "step": 345
+ },
+ {
+ "epoch": 4.810480349344978,
+ "grad_norm": 0.034655820578336716,
+ "learning_rate": 0.0006,
+ "loss": 6.265879154205322,
+ "step": 346
+ },
+ {
+ "epoch": 4.824454148471616,
+ "grad_norm": 0.04202224686741829,
+ "learning_rate": 0.0006,
+ "loss": 6.2896409034729,
+ "step": 347
+ },
+ {
+ "epoch": 4.8384279475982535,
+ "grad_norm": 0.03632277995347977,
+ "learning_rate": 0.0006,
+ "loss": 6.253917217254639,
+ "step": 348
+ },
+ {
+ "epoch": 4.85240174672489,
+ "grad_norm": 0.028645765036344528,
+ "learning_rate": 0.0006,
+ "loss": 6.207965850830078,
+ "step": 349
+ },
+ {
+ "epoch": 4.866375545851528,
+ "grad_norm": 0.025305170565843582,
+ "learning_rate": 0.0006,
+ "loss": 6.168068885803223,
+ "step": 350
+ },
+ {
+ "epoch": 4.880349344978166,
+ "grad_norm": 0.022137803956866264,
+ "learning_rate": 0.0006,
+ "loss": 6.235283374786377,
+ "step": 351
+ },
+ {
+ "epoch": 4.894323144104804,
+ "grad_norm": 0.01814538985490799,
+ "learning_rate": 0.0006,
+ "loss": 6.259641647338867,
+ "step": 352
+ },
+ {
+ "epoch": 4.908296943231441,
+ "grad_norm": 0.018465086817741394,
+ "learning_rate": 0.0006,
+ "loss": 6.111738681793213,
+ "step": 353
+ },
+ {
+ "epoch": 4.922270742358078,
+ "grad_norm": 0.018638933077454567,
+ "learning_rate": 0.0006,
+ "loss": 6.225642204284668,
+ "step": 354
+ },
+ {
+ "epoch": 4.936244541484716,
+ "grad_norm": 0.016461260616779327,
+ "learning_rate": 0.0006,
+ "loss": 6.140647888183594,
+ "step": 355
+ },
+ {
+ "epoch": 4.950218340611354,
+ "grad_norm": 0.015458385460078716,
+ "learning_rate": 0.0006,
+ "loss": 6.155224800109863,
+ "step": 356
+ },
+ {
+ "epoch": 4.964192139737992,
+ "grad_norm": 0.014934048056602478,
+ "learning_rate": 0.0006,
+ "loss": 6.101555824279785,
+ "step": 357
+ },
+ {
+ "epoch": 4.978165938864628,
+ "grad_norm": 0.013979545794427395,
+ "learning_rate": 0.0006,
+ "loss": 6.161718845367432,
+ "step": 358
+ },
+ {
+ "epoch": 4.992139737991266,
+ "grad_norm": 0.01640394888818264,
+ "learning_rate": 0.0006,
+ "loss": 6.055768966674805,
+ "step": 359
+ },
+ {
+ "epoch": 5.0,
+ "grad_norm": 0.015135680325329304,
+ "learning_rate": 0.0006,
+ "loss": 6.169695854187012,
+ "step": 360
+ },
+ {
+ "epoch": 5.0,
+ "eval_loss": 6.16408109664917,
+ "eval_runtime": 59.3649,
+ "eval_samples_per_second": 41.135,
+ "eval_steps_per_second": 1.297,
+ "step": 360
+ },
+ {
+ "epoch": 5.013973799126638,
+ "grad_norm": 0.012849048711359501,
+ "learning_rate": 0.0006,
+ "loss": 6.055830001831055,
+ "step": 361
+ },
+ {
+ "epoch": 5.0279475982532755,
+ "grad_norm": 0.018169576302170753,
+ "learning_rate": 0.0006,
+ "loss": 6.082655429840088,
+ "step": 362
+ },
+ {
+ "epoch": 5.041921397379912,
+ "grad_norm": 0.024293430149555206,
+ "learning_rate": 0.0006,
+ "loss": 6.109032154083252,
+ "step": 363
+ },
+ {
+ "epoch": 5.05589519650655,
+ "grad_norm": 0.03266787901520729,
+ "learning_rate": 0.0006,
+ "loss": 6.030613899230957,
+ "step": 364
+ },
+ {
+ "epoch": 5.069868995633188,
+ "grad_norm": 0.03572266176342964,
+ "learning_rate": 0.0006,
+ "loss": 6.050390720367432,
+ "step": 365
+ },
+ {
+ "epoch": 5.083842794759826,
+ "grad_norm": 0.032625213265419006,
+ "learning_rate": 0.0006,
+ "loss": 6.169313430786133,
+ "step": 366
+ },
+ {
+ "epoch": 5.097816593886463,
+ "grad_norm": 0.04051872715353966,
+ "learning_rate": 0.0006,
+ "loss": 6.10398006439209,
+ "step": 367
+ },
+ {
+ "epoch": 5.1117903930131,
+ "grad_norm": 0.0351213738322258,
+ "learning_rate": 0.0006,
+ "loss": 6.121973514556885,
+ "step": 368
+ },
+ {
+ "epoch": 5.125764192139738,
+ "grad_norm": 0.03523759916424751,
+ "learning_rate": 0.0006,
+ "loss": 6.0626654624938965,
+ "step": 369
+ },
+ {
+ "epoch": 5.139737991266376,
+ "grad_norm": 0.02706182189285755,
+ "learning_rate": 0.0006,
+ "loss": 6.049354076385498,
+ "step": 370
+ },
+ {
+ "epoch": 5.153711790393013,
+ "grad_norm": 0.024212589487433434,
+ "learning_rate": 0.0006,
+ "loss": 6.034826278686523,
+ "step": 371
+ },
+ {
+ "epoch": 5.1676855895196505,
+ "grad_norm": 0.024798082187771797,
+ "learning_rate": 0.0006,
+ "loss": 6.004058837890625,
+ "step": 372
+ },
+ {
+ "epoch": 5.181659388646288,
+ "grad_norm": 0.02325567416846752,
+ "learning_rate": 0.0006,
+ "loss": 5.986461639404297,
+ "step": 373
+ },
+ {
+ "epoch": 5.195633187772926,
+ "grad_norm": 0.019060570746660233,
+ "learning_rate": 0.0006,
+ "loss": 6.059736251831055,
+ "step": 374
+ },
+ {
+ "epoch": 5.209606986899563,
+ "grad_norm": 0.016822976991534233,
+ "learning_rate": 0.0006,
+ "loss": 6.072957515716553,
+ "step": 375
+ },
+ {
+ "epoch": 5.223580786026201,
+ "grad_norm": 0.017218658700585365,
+ "learning_rate": 0.0006,
+ "loss": 6.031739234924316,
+ "step": 376
+ },
+ {
+ "epoch": 5.2375545851528384,
+ "grad_norm": 0.014889383688569069,
+ "learning_rate": 0.0006,
+ "loss": 6.025674343109131,
+ "step": 377
+ },
+ {
+ "epoch": 5.251528384279476,
+ "grad_norm": 0.01708107627928257,
+ "learning_rate": 0.0006,
+ "loss": 6.090451240539551,
+ "step": 378
+ },
+ {
+ "epoch": 5.265502183406113,
+ "grad_norm": 0.020260317251086235,
+ "learning_rate": 0.0006,
+ "loss": 6.094257354736328,
+ "step": 379
+ },
+ {
+ "epoch": 5.279475982532751,
+ "grad_norm": 0.020110400393605232,
+ "learning_rate": 0.0006,
+ "loss": 6.021188735961914,
+ "step": 380
+ },
+ {
+ "epoch": 5.293449781659389,
+ "grad_norm": 0.02027830481529236,
+ "learning_rate": 0.0006,
+ "loss": 6.022156238555908,
+ "step": 381
+ },
+ {
+ "epoch": 5.307423580786026,
+ "grad_norm": 0.022747062146663666,
+ "learning_rate": 0.0006,
+ "loss": 6.011836051940918,
+ "step": 382
+ },
+ {
+ "epoch": 5.321397379912664,
+ "grad_norm": 0.01990230567753315,
+ "learning_rate": 0.0006,
+ "loss": 6.003054618835449,
+ "step": 383
+ },
+ {
+ "epoch": 5.335371179039301,
+ "grad_norm": 0.01596238650381565,
+ "learning_rate": 0.0006,
+ "loss": 5.991410732269287,
+ "step": 384
+ },
+ {
+ "epoch": 5.349344978165939,
+ "grad_norm": 0.015847105532884598,
+ "learning_rate": 0.0006,
+ "loss": 6.100622177124023,
+ "step": 385
+ },
+ {
+ "epoch": 5.3633187772925766,
+ "grad_norm": 0.016179397702217102,
+ "learning_rate": 0.0006,
+ "loss": 6.032659530639648,
+ "step": 386
+ },
+ {
+ "epoch": 5.377292576419214,
+ "grad_norm": 0.018256602808833122,
+ "learning_rate": 0.0006,
+ "loss": 5.961983680725098,
+ "step": 387
+ },
+ {
+ "epoch": 5.391266375545851,
+ "grad_norm": 0.02805912122130394,
+ "learning_rate": 0.0006,
+ "loss": 5.983541965484619,
+ "step": 388
+ },
+ {
+ "epoch": 5.405240174672489,
+ "grad_norm": 0.039082255214452744,
+ "learning_rate": 0.0006,
+ "loss": 6.021474361419678,
+ "step": 389
+ },
+ {
+ "epoch": 5.419213973799127,
+ "grad_norm": 0.036757659167051315,
+ "learning_rate": 0.0006,
+ "loss": 6.005046367645264,
+ "step": 390
+ },
+ {
+ "epoch": 5.4331877729257645,
+ "grad_norm": 0.035277366638183594,
+ "learning_rate": 0.0006,
+ "loss": 6.036296844482422,
+ "step": 391
+ },
+ {
+ "epoch": 5.447161572052401,
+ "grad_norm": 0.034404635429382324,
+ "learning_rate": 0.0006,
+ "loss": 5.994539260864258,
+ "step": 392
+ },
+ {
+ "epoch": 5.461135371179039,
+ "grad_norm": 0.0377206988632679,
+ "learning_rate": 0.0006,
+ "loss": 5.941534519195557,
+ "step": 393
+ },
+ {
+ "epoch": 5.475109170305677,
+ "grad_norm": 0.0389922633767128,
+ "learning_rate": 0.0006,
+ "loss": 5.983644485473633,
+ "step": 394
+ },
+ {
+ "epoch": 5.489082969432315,
+ "grad_norm": 0.04176778718829155,
+ "learning_rate": 0.0006,
+ "loss": 6.030922889709473,
+ "step": 395
+ },
+ {
+ "epoch": 5.503056768558952,
+ "grad_norm": 0.0337153784930706,
+ "learning_rate": 0.0006,
+ "loss": 5.961367130279541,
+ "step": 396
+ },
+ {
+ "epoch": 5.517030567685589,
+ "grad_norm": 0.036116816103458405,
+ "learning_rate": 0.0006,
+ "loss": 5.819280624389648,
+ "step": 397
+ },
+ {
+ "epoch": 5.531004366812227,
+ "grad_norm": 0.030724041163921356,
+ "learning_rate": 0.0006,
+ "loss": 5.906380653381348,
+ "step": 398
+ },
+ {
+ "epoch": 5.544978165938865,
+ "grad_norm": 0.030264202505350113,
+ "learning_rate": 0.0006,
+ "loss": 5.975833415985107,
+ "step": 399
+ },
+ {
+ "epoch": 5.558951965065502,
+ "grad_norm": 0.032096318900585175,
+ "learning_rate": 0.0006,
+ "loss": 6.001348972320557,
+ "step": 400
+ },
+ {
+ "epoch": 5.5729257641921395,
+ "grad_norm": 0.030579356476664543,
+ "learning_rate": 0.0006,
+ "loss": 5.928768634796143,
+ "step": 401
+ },
+ {
+ "epoch": 5.586899563318777,
+ "grad_norm": 0.028241973370313644,
+ "learning_rate": 0.0006,
+ "loss": 5.92582893371582,
+ "step": 402
+ },
+ {
+ "epoch": 5.600873362445415,
+ "grad_norm": 0.023086953908205032,
+ "learning_rate": 0.0006,
+ "loss": 5.892926216125488,
+ "step": 403
+ },
+ {
+ "epoch": 5.614847161572053,
+ "grad_norm": 0.02547992393374443,
+ "learning_rate": 0.0006,
+ "loss": 5.903195381164551,
+ "step": 404
+ },
+ {
+ "epoch": 5.62882096069869,
+ "grad_norm": 0.023089831694960594,
+ "learning_rate": 0.0006,
+ "loss": 5.912033557891846,
+ "step": 405
+ },
+ {
+ "epoch": 5.642794759825327,
+ "grad_norm": 0.025446368381381035,
+ "learning_rate": 0.0006,
+ "loss": 5.926138401031494,
+ "step": 406
+ },
+ {
+ "epoch": 5.656768558951965,
+ "grad_norm": 0.03748747706413269,
+ "learning_rate": 0.0006,
+ "loss": 5.913451194763184,
+ "step": 407
+ },
+ {
+ "epoch": 5.670742358078603,
+ "grad_norm": 0.03493810072541237,
+ "learning_rate": 0.0006,
+ "loss": 5.97633695602417,
+ "step": 408
+ },
+ {
+ "epoch": 5.68471615720524,
+ "grad_norm": 0.023131275549530983,
+ "learning_rate": 0.0006,
+ "loss": 5.899571418762207,
+ "step": 409
+ },
+ {
+ "epoch": 5.698689956331878,
+ "grad_norm": 0.02986014075577259,
+ "learning_rate": 0.0006,
+ "loss": 5.894903182983398,
+ "step": 410
+ },
+ {
+ "epoch": 5.712663755458515,
+ "grad_norm": 0.030171144753694534,
+ "learning_rate": 0.0006,
+ "loss": 5.876595497131348,
+ "step": 411
+ },
+ {
+ "epoch": 5.726637554585153,
+ "grad_norm": 0.029377546161413193,
+ "learning_rate": 0.0006,
+ "loss": 5.911637306213379,
+ "step": 412
+ },
+ {
+ "epoch": 5.74061135371179,
+ "grad_norm": 0.029829490929841995,
+ "learning_rate": 0.0006,
+ "loss": 5.819629669189453,
+ "step": 413
+ },
+ {
+ "epoch": 5.754585152838428,
+ "grad_norm": 0.024139299988746643,
+ "learning_rate": 0.0006,
+ "loss": 5.81728458404541,
+ "step": 414
+ },
+ {
+ "epoch": 5.7685589519650655,
+ "grad_norm": 0.021453432738780975,
+ "learning_rate": 0.0006,
+ "loss": 5.987326145172119,
+ "step": 415
+ },
+ {
+ "epoch": 5.782532751091703,
+ "grad_norm": 0.017942246049642563,
+ "learning_rate": 0.0006,
+ "loss": 5.893008232116699,
+ "step": 416
+ },
+ {
+ "epoch": 5.796506550218341,
+ "grad_norm": 0.019724637269973755,
+ "learning_rate": 0.0006,
+ "loss": 5.912441253662109,
+ "step": 417
+ },
+ {
+ "epoch": 5.810480349344978,
+ "grad_norm": 0.018597912043333054,
+ "learning_rate": 0.0006,
+ "loss": 5.860394477844238,
+ "step": 418
+ },
+ {
+ "epoch": 5.824454148471616,
+ "grad_norm": 0.016232412308454514,
+ "learning_rate": 0.0006,
+ "loss": 5.94991397857666,
+ "step": 419
+ },
+ {
+ "epoch": 5.8384279475982535,
+ "grad_norm": 0.01667204685509205,
+ "learning_rate": 0.0006,
+ "loss": 5.8506364822387695,
+ "step": 420
+ },
+ {
+ "epoch": 5.85240174672489,
+ "grad_norm": 0.014232151210308075,
+ "learning_rate": 0.0006,
+ "loss": 5.928424835205078,
+ "step": 421
+ },
+ {
+ "epoch": 5.866375545851528,
+ "grad_norm": 0.01570476032793522,
+ "learning_rate": 0.0006,
+ "loss": 5.953692436218262,
+ "step": 422
+ },
+ {
+ "epoch": 5.880349344978166,
+ "grad_norm": 0.01586179807782173,
+ "learning_rate": 0.0006,
+ "loss": 5.820014476776123,
+ "step": 423
+ },
+ {
+ "epoch": 5.894323144104804,
+ "grad_norm": 0.014729145914316177,
+ "learning_rate": 0.0006,
+ "loss": 5.860154628753662,
+ "step": 424
+ },
+ {
+ "epoch": 5.908296943231441,
+ "grad_norm": 0.013111459091305733,
+ "learning_rate": 0.0006,
+ "loss": 5.85988187789917,
+ "step": 425
+ },
+ {
+ "epoch": 5.922270742358078,
+ "grad_norm": 0.012937922962009907,
+ "learning_rate": 0.0006,
+ "loss": 5.756265640258789,
+ "step": 426
+ },
+ {
+ "epoch": 5.936244541484716,
+ "grad_norm": 0.01377453189343214,
+ "learning_rate": 0.0006,
+ "loss": 5.8333048820495605,
+ "step": 427
+ },
+ {
+ "epoch": 5.950218340611354,
+ "grad_norm": 0.014054795727133751,
+ "learning_rate": 0.0006,
+ "loss": 5.803333759307861,
+ "step": 428
+ },
+ {
+ "epoch": 5.964192139737992,
+ "grad_norm": 0.01674959622323513,
+ "learning_rate": 0.0006,
+ "loss": 5.8718485832214355,
+ "step": 429
+ },
+ {
+ "epoch": 5.978165938864628,
+ "grad_norm": 0.018733017146587372,
+ "learning_rate": 0.0006,
+ "loss": 5.875979900360107,
+ "step": 430
+ },
+ {
+ "epoch": 5.992139737991266,
+ "grad_norm": 0.02088983915746212,
+ "learning_rate": 0.0006,
+ "loss": 5.763634204864502,
+ "step": 431
+ },
+ {
+ "epoch": 6.0,
+ "grad_norm": 0.026705985888838768,
+ "learning_rate": 0.0006,
+ "loss": 5.6922478675842285,
+ "step": 432
+ },
+ {
+ "epoch": 6.0,
+ "eval_loss": 5.847958087921143,
+ "eval_runtime": 58.5451,
+ "eval_samples_per_second": 41.711,
+ "eval_steps_per_second": 1.315,
+ "step": 432
+ },
+ {
+ "epoch": 6.013973799126638,
+ "grad_norm": 0.03181544691324234,
+ "learning_rate": 0.0006,
+ "loss": 5.786368370056152,
+ "step": 433
+ },
+ {
+ "epoch": 6.0279475982532755,
+ "grad_norm": 0.03238112851977348,
+ "learning_rate": 0.0006,
+ "loss": 5.810310363769531,
+ "step": 434
+ },
+ {
+ "epoch": 6.041921397379912,
+ "grad_norm": 0.03916756808757782,
+ "learning_rate": 0.0006,
+ "loss": 5.7761993408203125,
+ "step": 435
+ },
+ {
+ "epoch": 6.05589519650655,
+ "grad_norm": 0.044009730219841,
+ "learning_rate": 0.0006,
+ "loss": 5.898112773895264,
+ "step": 436
+ },
+ {
+ "epoch": 6.069868995633188,
+ "grad_norm": 0.045235246419906616,
+ "learning_rate": 0.0006,
+ "loss": 5.876371383666992,
+ "step": 437
+ },
+ {
+ "epoch": 6.083842794759826,
+ "grad_norm": 0.050218384712934494,
+ "learning_rate": 0.0006,
+ "loss": 5.805103302001953,
+ "step": 438
+ },
+ {
+ "epoch": 6.097816593886463,
+ "grad_norm": 0.05444490164518356,
+ "learning_rate": 0.0006,
+ "loss": 5.826424598693848,
+ "step": 439
+ },
+ {
+ "epoch": 6.1117903930131,
+ "grad_norm": 0.04890982061624527,
+ "learning_rate": 0.0006,
+ "loss": 5.7735395431518555,
+ "step": 440
+ },
+ {
+ "epoch": 6.125764192139738,
+ "grad_norm": 0.05478639155626297,
+ "learning_rate": 0.0006,
+ "loss": 5.823677062988281,
+ "step": 441
+ },
+ {
+ "epoch": 6.139737991266376,
+ "grad_norm": 0.057562313973903656,
+ "learning_rate": 0.0006,
+ "loss": 5.889334678649902,
+ "step": 442
+ },
+ {
+ "epoch": 6.153711790393013,
+ "grad_norm": 0.06447703391313553,
+ "learning_rate": 0.0006,
+ "loss": 5.84621524810791,
+ "step": 443
+ },
+ {
+ "epoch": 6.1676855895196505,
+ "grad_norm": 0.048861872404813766,
+ "learning_rate": 0.0006,
+ "loss": 5.931595802307129,
+ "step": 444
+ },
+ {
+ "epoch": 6.181659388646288,
+ "grad_norm": 0.05521339178085327,
+ "learning_rate": 0.0006,
+ "loss": 5.8229780197143555,
+ "step": 445
+ },
+ {
+ "epoch": 6.195633187772926,
+ "grad_norm": 0.053666990250349045,
+ "learning_rate": 0.0006,
+ "loss": 5.879191875457764,
+ "step": 446
+ },
+ {
+ "epoch": 6.209606986899563,
+ "grad_norm": 0.0451025515794754,
+ "learning_rate": 0.0006,
+ "loss": 5.801628112792969,
+ "step": 447
+ },
+ {
+ "epoch": 6.223580786026201,
+ "grad_norm": 0.04291655868291855,
+ "learning_rate": 0.0006,
+ "loss": 5.914680480957031,
+ "step": 448
+ },
+ {
+ "epoch": 6.2375545851528384,
+ "grad_norm": 0.035102490335702896,
+ "learning_rate": 0.0006,
+ "loss": 5.843682765960693,
+ "step": 449
+ },
+ {
+ "epoch": 6.251528384279476,
+ "grad_norm": 0.03403995931148529,
+ "learning_rate": 0.0006,
+ "loss": 5.827154159545898,
+ "step": 450
+ },
+ {
+ "epoch": 6.265502183406113,
+ "grad_norm": 0.03444375470280647,
+ "learning_rate": 0.0006,
+ "loss": 5.800136566162109,
+ "step": 451
+ },
+ {
+ "epoch": 6.279475982532751,
+ "grad_norm": 0.03165159001946449,
+ "learning_rate": 0.0006,
+ "loss": 5.906252384185791,
+ "step": 452
+ },
+ {
+ "epoch": 6.293449781659389,
+ "grad_norm": 0.026153093203902245,
+ "learning_rate": 0.0006,
+ "loss": 5.7423200607299805,
+ "step": 453
+ },
+ {
+ "epoch": 6.307423580786026,
+ "grad_norm": 0.024357657879590988,
+ "learning_rate": 0.0006,
+ "loss": 5.701364517211914,
+ "step": 454
+ },
+ {
+ "epoch": 6.321397379912664,
+ "grad_norm": 0.021508049219846725,
+ "learning_rate": 0.0006,
+ "loss": 5.8328094482421875,
+ "step": 455
+ },
+ {
+ "epoch": 6.335371179039301,
+ "grad_norm": 0.017313921824097633,
+ "learning_rate": 0.0006,
+ "loss": 5.792145252227783,
+ "step": 456
+ },
+ {
+ "epoch": 6.349344978165939,
+ "grad_norm": 0.018563397228717804,
+ "learning_rate": 0.0006,
+ "loss": 5.732672214508057,
+ "step": 457
+ },
+ {
+ "epoch": 6.3633187772925766,
+ "grad_norm": 0.016568679362535477,
+ "learning_rate": 0.0006,
+ "loss": 5.81948184967041,
+ "step": 458
+ },
+ {
+ "epoch": 6.377292576419214,
+ "grad_norm": 0.014133770950138569,
+ "learning_rate": 0.0006,
+ "loss": 5.734982490539551,
+ "step": 459
+ },
+ {
+ "epoch": 6.391266375545851,
+ "grad_norm": 0.014114447869360447,
+ "learning_rate": 0.0006,
+ "loss": 5.70073127746582,
+ "step": 460
+ },
+ {
+ "epoch": 6.405240174672489,
+ "grad_norm": 0.012907393276691437,
+ "learning_rate": 0.0006,
+ "loss": 5.758626937866211,
+ "step": 461
+ },
+ {
+ "epoch": 6.419213973799127,
+ "grad_norm": 0.012673444114625454,
+ "learning_rate": 0.0006,
+ "loss": 5.775138854980469,
+ "step": 462
+ },
+ {
+ "epoch": 6.4331877729257645,
+ "grad_norm": 0.01311410591006279,
+ "learning_rate": 0.0006,
+ "loss": 5.70920991897583,
+ "step": 463
+ },
+ {
+ "epoch": 6.447161572052401,
+ "grad_norm": 0.012935544364154339,
+ "learning_rate": 0.0006,
+ "loss": 5.781479835510254,
+ "step": 464
+ },
+ {
+ "epoch": 6.461135371179039,
+ "grad_norm": 0.01439738366752863,
+ "learning_rate": 0.0006,
+ "loss": 5.691315650939941,
+ "step": 465
+ },
+ {
+ "epoch": 6.475109170305677,
+ "grad_norm": 0.01205186452716589,
+ "learning_rate": 0.0006,
+ "loss": 5.662499904632568,
+ "step": 466
+ },
+ {
+ "epoch": 6.489082969432315,
+ "grad_norm": 0.011575652286410332,
+ "learning_rate": 0.0006,
+ "loss": 5.6271772384643555,
+ "step": 467
+ },
+ {
+ "epoch": 6.503056768558952,
+ "grad_norm": 0.011672502383589745,
+ "learning_rate": 0.0006,
+ "loss": 5.761662006378174,
+ "step": 468
+ },
+ {
+ "epoch": 6.517030567685589,
+ "grad_norm": 0.011541751213371754,
+ "learning_rate": 0.0006,
+ "loss": 5.762078285217285,
+ "step": 469
+ },
+ {
+ "epoch": 6.531004366812227,
+ "grad_norm": 0.011396365240216255,
+ "learning_rate": 0.0006,
+ "loss": 5.67873477935791,
+ "step": 470
+ },
+ {
+ "epoch": 6.544978165938865,
+ "grad_norm": 0.010688499547541142,
+ "learning_rate": 0.0006,
+ "loss": 5.717051982879639,
+ "step": 471
+ },
+ {
+ "epoch": 6.558951965065502,
+ "grad_norm": 0.012224317528307438,
+ "learning_rate": 0.0006,
+ "loss": 5.707948684692383,
+ "step": 472
+ },
+ {
+ "epoch": 6.5729257641921395,
+ "grad_norm": 0.011856825090944767,
+ "learning_rate": 0.0006,
+ "loss": 5.70878791809082,
+ "step": 473
+ },
+ {
+ "epoch": 6.586899563318777,
+ "grad_norm": 0.01199064590036869,
+ "learning_rate": 0.0006,
+ "loss": 5.708697319030762,
+ "step": 474
+ },
+ {
+ "epoch": 6.600873362445415,
+ "grad_norm": 0.01219446212053299,
+ "learning_rate": 0.0006,
+ "loss": 5.6061577796936035,
+ "step": 475
+ },
+ {
+ "epoch": 6.614847161572053,
+ "grad_norm": 0.013397484086453915,
+ "learning_rate": 0.0006,
+ "loss": 5.624789714813232,
+ "step": 476
+ },
+ {
+ "epoch": 6.62882096069869,
+ "grad_norm": 0.01483136136084795,
+ "learning_rate": 0.0006,
+ "loss": 5.721141338348389,
+ "step": 477
+ },
+ {
+ "epoch": 6.642794759825327,
+ "grad_norm": 0.0194488987326622,
+ "learning_rate": 0.0006,
+ "loss": 5.604279518127441,
+ "step": 478
+ },
+ {
+ "epoch": 6.656768558951965,
+ "grad_norm": 0.02079896256327629,
+ "learning_rate": 0.0006,
+ "loss": 5.586322784423828,
+ "step": 479
+ },
+ {
+ "epoch": 6.670742358078603,
+ "grad_norm": 0.018996229395270348,
+ "learning_rate": 0.0006,
+ "loss": 5.5987772941589355,
+ "step": 480
+ },
+ {
+ "epoch": 6.68471615720524,
+ "grad_norm": 0.015779603272676468,
+ "learning_rate": 0.0006,
+ "loss": 5.588602066040039,
+ "step": 481
+ },
+ {
+ "epoch": 6.698689956331878,
+ "grad_norm": 0.015322973020374775,
+ "learning_rate": 0.0006,
+ "loss": 5.686467170715332,
+ "step": 482
+ },
+ {
+ "epoch": 6.712663755458515,
+ "grad_norm": 0.015282213687896729,
+ "learning_rate": 0.0006,
+ "loss": 5.664676666259766,
+ "step": 483
+ },
+ {
+ "epoch": 6.726637554585153,
+ "grad_norm": 0.016589272767305374,
+ "learning_rate": 0.0006,
+ "loss": 5.666739463806152,
+ "step": 484
+ },
+ {
+ "epoch": 6.74061135371179,
+ "grad_norm": 0.01811421848833561,
+ "learning_rate": 0.0006,
+ "loss": 5.647593021392822,
+ "step": 485
+ },
+ {
+ "epoch": 6.754585152838428,
+ "grad_norm": 0.019652029499411583,
+ "learning_rate": 0.0006,
+ "loss": 5.628726959228516,
+ "step": 486
+ },
+ {
+ "epoch": 6.7685589519650655,
+ "grad_norm": 0.02118665538728237,
+ "learning_rate": 0.0006,
+ "loss": 5.6580705642700195,
+ "step": 487
+ },
+ {
+ "epoch": 6.782532751091703,
+ "grad_norm": 0.02237832546234131,
+ "learning_rate": 0.0006,
+ "loss": 5.635931968688965,
+ "step": 488
+ },
+ {
+ "epoch": 6.796506550218341,
+ "grad_norm": 0.023897631093859673,
+ "learning_rate": 0.0006,
+ "loss": 5.517949104309082,
+ "step": 489
+ },
+ {
+ "epoch": 6.810480349344978,
+ "grad_norm": 0.02442227490246296,
+ "learning_rate": 0.0006,
+ "loss": 5.600021839141846,
+ "step": 490
+ },
+ {
+ "epoch": 6.824454148471616,
+ "grad_norm": 0.024675287306308746,
+ "learning_rate": 0.0006,
+ "loss": 5.6639909744262695,
+ "step": 491
+ },
+ {
+ "epoch": 6.8384279475982535,
+ "grad_norm": 0.030372392386198044,
+ "learning_rate": 0.0006,
+ "loss": 5.582888603210449,
+ "step": 492
+ },
+ {
+ "epoch": 6.85240174672489,
+ "grad_norm": 0.03237690031528473,
+ "learning_rate": 0.0006,
+ "loss": 5.573896884918213,
+ "step": 493
+ },
+ {
+ "epoch": 6.866375545851528,
+ "grad_norm": 0.03511589393019676,
+ "learning_rate": 0.0006,
+ "loss": 5.565878868103027,
+ "step": 494
+ },
+ {
+ "epoch": 6.880349344978166,
+ "grad_norm": 0.03883938118815422,
+ "learning_rate": 0.0006,
+ "loss": 5.58540678024292,
+ "step": 495
+ },
+ {
+ "epoch": 6.894323144104804,
+ "grad_norm": 0.04175502806901932,
+ "learning_rate": 0.0006,
+ "loss": 5.6095194816589355,
+ "step": 496
+ },
+ {
+ "epoch": 6.908296943231441,
+ "grad_norm": 0.04013441503047943,
+ "learning_rate": 0.0006,
+ "loss": 5.654401779174805,
+ "step": 497
+ },
+ {
+ "epoch": 6.922270742358078,
+ "grad_norm": 0.04618770629167557,
+ "learning_rate": 0.0006,
+ "loss": 5.6483473777771,
+ "step": 498
+ },
+ {
+ "epoch": 6.936244541484716,
+ "grad_norm": 0.04507308453321457,
+ "learning_rate": 0.0006,
+ "loss": 5.5856218338012695,
+ "step": 499
+ },
+ {
+ "epoch": 6.950218340611354,
+ "grad_norm": 0.03151383996009827,
+ "learning_rate": 0.0006,
+ "loss": 5.594086647033691,
+ "step": 500
+ },
+ {
+ "epoch": 6.964192139737992,
+ "grad_norm": 0.026883797720074654,
+ "learning_rate": 0.0006,
+ "loss": 5.607676029205322,
+ "step": 501
+ },
+ {
+ "epoch": 6.978165938864628,
+ "grad_norm": 0.02981836162507534,
+ "learning_rate": 0.0006,
+ "loss": 5.563666820526123,
+ "step": 502
+ },
+ {
+ "epoch": 6.992139737991266,
+ "grad_norm": 0.03233477845788002,
+ "learning_rate": 0.0006,
+ "loss": 5.667543411254883,
+ "step": 503
+ },
+ {
+ "epoch": 7.0,
+ "grad_norm": 0.029927456751465797,
+ "learning_rate": 0.0006,
+ "loss": 5.733482837677002,
+ "step": 504
+ },
+ {
+ "epoch": 7.0,
+ "eval_loss": 5.679076671600342,
+ "eval_runtime": 59.5965,
+ "eval_samples_per_second": 40.976,
+ "eval_steps_per_second": 1.292,
+ "step": 504
+ },
+ {
+ "epoch": 7.013973799126638,
+ "grad_norm": 0.027081597596406937,
+ "learning_rate": 0.0006,
+ "loss": 5.634487152099609,
+ "step": 505
+ },
+ {
+ "epoch": 7.0279475982532755,
+ "grad_norm": 0.031959421932697296,
+ "learning_rate": 0.0006,
+ "loss": 5.643294811248779,
+ "step": 506
+ },
+ {
+ "epoch": 7.041921397379912,
+ "grad_norm": 0.02802063524723053,
+ "learning_rate": 0.0006,
+ "loss": 5.573239326477051,
+ "step": 507
+ },
+ {
+ "epoch": 7.05589519650655,
+ "grad_norm": 0.032579705119132996,
+ "learning_rate": 0.0006,
+ "loss": 5.627150058746338,
+ "step": 508
+ },
+ {
+ "epoch": 7.069868995633188,
+ "grad_norm": 0.02503928542137146,
+ "learning_rate": 0.0006,
+ "loss": 5.6045308113098145,
+ "step": 509
+ },
+ {
+ "epoch": 7.083842794759826,
+ "grad_norm": 0.02421317622065544,
+ "learning_rate": 0.0006,
+ "loss": 5.522153377532959,
+ "step": 510
+ },
+ {
+ "epoch": 7.097816593886463,
+ "grad_norm": 0.021983426064252853,
+ "learning_rate": 0.0006,
+ "loss": 5.632939338684082,
+ "step": 511
+ },
+ {
+ "epoch": 7.1117903930131,
+ "grad_norm": 0.021933183073997498,
+ "learning_rate": 0.0006,
+ "loss": 5.592185974121094,
+ "step": 512
+ },
+ {
+ "epoch": 7.125764192139738,
+ "grad_norm": 0.02287031151354313,
+ "learning_rate": 0.0006,
+ "loss": 5.613700866699219,
+ "step": 513
+ },
+ {
+ "epoch": 7.139737991266376,
+ "grad_norm": 0.02281602844595909,
+ "learning_rate": 0.0006,
+ "loss": 5.551383018493652,
+ "step": 514
+ },
+ {
+ "epoch": 7.153711790393013,
+ "grad_norm": 0.021050360053777695,
+ "learning_rate": 0.0006,
+ "loss": 5.55275821685791,
+ "step": 515
+ },
+ {
+ "epoch": 7.1676855895196505,
+ "grad_norm": 0.018299926072359085,
+ "learning_rate": 0.0006,
+ "loss": 5.528225421905518,
+ "step": 516
+ },
+ {
+ "epoch": 7.181659388646288,
+ "grad_norm": 0.02024853602051735,
+ "learning_rate": 0.0006,
+ "loss": 5.522453784942627,
+ "step": 517
+ },
+ {
+ "epoch": 7.195633187772926,
+ "grad_norm": 0.02135239914059639,
+ "learning_rate": 0.0006,
+ "loss": 5.527522563934326,
+ "step": 518
+ },
+ {
+ "epoch": 7.209606986899563,
+ "grad_norm": 0.023394184187054634,
+ "learning_rate": 0.0006,
+ "loss": 5.4535017013549805,
+ "step": 519
+ },
+ {
+ "epoch": 7.223580786026201,
+ "grad_norm": 0.025952080264687538,
+ "learning_rate": 0.0006,
+ "loss": 5.5374674797058105,
+ "step": 520
+ },
+ {
+ "epoch": 7.2375545851528384,
+ "grad_norm": 0.022287718951702118,
+ "learning_rate": 0.0006,
+ "loss": 5.493753433227539,
+ "step": 521
+ },
+ {
+ "epoch": 7.251528384279476,
+ "grad_norm": 0.02018447406589985,
+ "learning_rate": 0.0006,
+ "loss": 5.510575294494629,
+ "step": 522
+ },
+ {
+ "epoch": 7.265502183406113,
+ "grad_norm": 0.02050507813692093,
+ "learning_rate": 0.0006,
+ "loss": 5.466026782989502,
+ "step": 523
+ },
+ {
+ "epoch": 7.279475982532751,
+ "grad_norm": 0.023688802495598793,
+ "learning_rate": 0.0006,
+ "loss": 5.485815525054932,
+ "step": 524
+ },
+ {
+ "epoch": 7.293449781659389,
+ "grad_norm": 0.021078843623399734,
+ "learning_rate": 0.0006,
+ "loss": 5.415581703186035,
+ "step": 525
+ },
+ {
+ "epoch": 7.307423580786026,
+ "grad_norm": 0.018266011029481888,
+ "learning_rate": 0.0006,
+ "loss": 5.402815818786621,
+ "step": 526
+ },
+ {
+ "epoch": 7.321397379912664,
+ "grad_norm": 0.019693493843078613,
+ "learning_rate": 0.0006,
+ "loss": 5.504674911499023,
+ "step": 527
+ },
+ {
+ "epoch": 7.335371179039301,
+ "grad_norm": 0.02031373605132103,
+ "learning_rate": 0.0006,
+ "loss": 5.503837585449219,
+ "step": 528
+ },
+ {
+ "epoch": 7.349344978165939,
+ "grad_norm": 0.017636626958847046,
+ "learning_rate": 0.0006,
+ "loss": 5.530580997467041,
+ "step": 529
+ },
+ {
+ "epoch": 7.3633187772925766,
+ "grad_norm": 0.01787244901061058,
+ "learning_rate": 0.0006,
+ "loss": 5.450218677520752,
+ "step": 530
+ },
+ {
+ "epoch": 7.377292576419214,
+ "grad_norm": 0.0170669574290514,
+ "learning_rate": 0.0006,
+ "loss": 5.481570243835449,
+ "step": 531
+ },
+ {
+ "epoch": 7.391266375545851,
+ "grad_norm": 0.01802165061235428,
+ "learning_rate": 0.0006,
+ "loss": 5.386394500732422,
+ "step": 532
+ },
+ {
+ "epoch": 7.405240174672489,
+ "grad_norm": 0.022949472069740295,
+ "learning_rate": 0.0006,
+ "loss": 5.424929618835449,
+ "step": 533
+ },
+ {
+ "epoch": 7.419213973799127,
+ "grad_norm": 0.034213390201330185,
+ "learning_rate": 0.0006,
+ "loss": 5.346663475036621,
+ "step": 534
+ },
+ {
+ "epoch": 7.4331877729257645,
+ "grad_norm": 0.03847593814134598,
+ "learning_rate": 0.0006,
+ "loss": 5.418482780456543,
+ "step": 535
+ },
+ {
+ "epoch": 7.447161572052401,
+ "grad_norm": 0.029393676668405533,
+ "learning_rate": 0.0006,
+ "loss": 5.456090927124023,
+ "step": 536
+ },
+ {
+ "epoch": 7.461135371179039,
+ "grad_norm": 0.03388667106628418,
+ "learning_rate": 0.0006,
+ "loss": 5.572983741760254,
+ "step": 537
+ },
+ {
+ "epoch": 7.475109170305677,
+ "grad_norm": 0.039690613746643066,
+ "learning_rate": 0.0006,
+ "loss": 5.414067268371582,
+ "step": 538
+ },
+ {
+ "epoch": 7.489082969432315,
+ "grad_norm": 0.03630776330828667,
+ "learning_rate": 0.0006,
+ "loss": 5.523739814758301,
+ "step": 539
+ },
+ {
+ "epoch": 7.503056768558952,
+ "grad_norm": 0.03356870263814926,
+ "learning_rate": 0.0006,
+ "loss": 5.444767951965332,
+ "step": 540
+ },
+ {
+ "epoch": 7.517030567685589,
+ "grad_norm": 0.030812304466962814,
+ "learning_rate": 0.0006,
+ "loss": 5.461244106292725,
+ "step": 541
+ },
+ {
+ "epoch": 7.531004366812227,
+ "grad_norm": 0.03319217637181282,
+ "learning_rate": 0.0006,
+ "loss": 5.502161026000977,
+ "step": 542
+ },
+ {
+ "epoch": 7.544978165938865,
+ "grad_norm": 0.032997481524944305,
+ "learning_rate": 0.0006,
+ "loss": 5.4162278175354,
+ "step": 543
+ },
+ {
+ "epoch": 7.558951965065502,
+ "grad_norm": 0.03364962339401245,
+ "learning_rate": 0.0006,
+ "loss": 5.415736198425293,
+ "step": 544
+ },
+ {
+ "epoch": 7.5729257641921395,
+ "grad_norm": 0.0344221405684948,
+ "learning_rate": 0.0006,
+ "loss": 5.541967868804932,
+ "step": 545
+ },
+ {
+ "epoch": 7.586899563318777,
+ "grad_norm": 0.029609503224492073,
+ "learning_rate": 0.0006,
+ "loss": 5.440614223480225,
+ "step": 546
+ },
+ {
+ "epoch": 7.600873362445415,
+ "grad_norm": 0.029055926948785782,
+ "learning_rate": 0.0006,
+ "loss": 5.462865352630615,
+ "step": 547
+ },
+ {
+ "epoch": 7.614847161572053,
+ "grad_norm": 0.02658848837018013,
+ "learning_rate": 0.0006,
+ "loss": 5.444467544555664,
+ "step": 548
+ },
+ {
+ "epoch": 7.62882096069869,
+ "grad_norm": 0.026276404038071632,
+ "learning_rate": 0.0006,
+ "loss": 5.462278842926025,
+ "step": 549
+ },
+ {
+ "epoch": 7.642794759825327,
+ "grad_norm": 0.0286889486014843,
+ "learning_rate": 0.0006,
+ "loss": 5.4849534034729,
+ "step": 550
+ },
+ {
+ "epoch": 7.656768558951965,
+ "grad_norm": 0.026849817484617233,
+ "learning_rate": 0.0006,
+ "loss": 5.431473731994629,
+ "step": 551
+ },
+ {
+ "epoch": 7.670742358078603,
+ "grad_norm": 0.02312396466732025,
+ "learning_rate": 0.0006,
+ "loss": 5.45462703704834,
+ "step": 552
+ },
+ {
+ "epoch": 7.68471615720524,
+ "grad_norm": 0.026197485625743866,
+ "learning_rate": 0.0006,
+ "loss": 5.452552318572998,
+ "step": 553
+ },
+ {
+ "epoch": 7.698689956331878,
+ "grad_norm": 0.02747255191206932,
+ "learning_rate": 0.0006,
+ "loss": 5.447169780731201,
+ "step": 554
+ },
+ {
+ "epoch": 7.712663755458515,
+ "grad_norm": 0.028123432770371437,
+ "learning_rate": 0.0006,
+ "loss": 5.394400596618652,
+ "step": 555
+ },
+ {
+ "epoch": 7.726637554585153,
+ "grad_norm": 0.02599870041012764,
+ "learning_rate": 0.0006,
+ "loss": 5.398341178894043,
+ "step": 556
+ },
+ {
+ "epoch": 7.74061135371179,
+ "grad_norm": 0.022171657532453537,
+ "learning_rate": 0.0006,
+ "loss": 5.368820667266846,
+ "step": 557
+ },
+ {
+ "epoch": 7.754585152838428,
+ "grad_norm": 0.022309480234980583,
+ "learning_rate": 0.0006,
+ "loss": 5.416772365570068,
+ "step": 558
+ },
+ {
+ "epoch": 7.7685589519650655,
+ "grad_norm": 0.024074165150523186,
+ "learning_rate": 0.0006,
+ "loss": 5.359002113342285,
+ "step": 559
+ },
+ {
+ "epoch": 7.782532751091703,
+ "grad_norm": 0.02636653557419777,
+ "learning_rate": 0.0006,
+ "loss": 5.484368324279785,
+ "step": 560
+ },
+ {
+ "epoch": 7.796506550218341,
+ "grad_norm": 0.02203752100467682,
+ "learning_rate": 0.0006,
+ "loss": 5.375498294830322,
+ "step": 561
+ },
+ {
+ "epoch": 7.810480349344978,
+ "grad_norm": 0.019924819469451904,
+ "learning_rate": 0.0006,
+ "loss": 5.334723472595215,
+ "step": 562
+ },
+ {
+ "epoch": 7.824454148471616,
+ "grad_norm": 0.018755966797471046,
+ "learning_rate": 0.0006,
+ "loss": 5.348781108856201,
+ "step": 563
+ },
+ {
+ "epoch": 7.8384279475982535,
+ "grad_norm": 0.016747845336794853,
+ "learning_rate": 0.0006,
+ "loss": 5.433079719543457,
+ "step": 564
+ },
+ {
+ "epoch": 7.85240174672489,
+ "grad_norm": 0.0170395877212286,
+ "learning_rate": 0.0006,
+ "loss": 5.346794128417969,
+ "step": 565
+ },
+ {
+ "epoch": 7.866375545851528,
+ "grad_norm": 0.016074176877737045,
+ "learning_rate": 0.0006,
+ "loss": 5.3277740478515625,
+ "step": 566
+ },
+ {
+ "epoch": 7.880349344978166,
+ "grad_norm": 0.014119806699454784,
+ "learning_rate": 0.0006,
+ "loss": 5.483551979064941,
+ "step": 567
+ },
+ {
+ "epoch": 7.894323144104804,
+ "grad_norm": 0.015270394273102283,
+ "learning_rate": 0.0006,
+ "loss": 5.378519535064697,
+ "step": 568
+ },
+ {
+ "epoch": 7.908296943231441,
+ "grad_norm": 0.015677539631724358,
+ "learning_rate": 0.0006,
+ "loss": 5.290210247039795,
+ "step": 569
+ },
+ {
+ "epoch": 7.922270742358078,
+ "grad_norm": 0.015930423513054848,
+ "learning_rate": 0.0006,
+ "loss": 5.427360534667969,
+ "step": 570
+ },
+ {
+ "epoch": 7.936244541484716,
+ "grad_norm": 0.016641564667224884,
+ "learning_rate": 0.0006,
+ "loss": 5.301599979400635,
+ "step": 571
+ },
+ {
+ "epoch": 7.950218340611354,
+ "grad_norm": 0.018293552100658417,
+ "learning_rate": 0.0006,
+ "loss": 5.277889251708984,
+ "step": 572
+ },
+ {
+ "epoch": 7.964192139737992,
+ "grad_norm": 0.017618799582123756,
+ "learning_rate": 0.0006,
+ "loss": 5.297194480895996,
+ "step": 573
+ },
+ {
+ "epoch": 7.978165938864628,
+ "grad_norm": 0.015034242533147335,
+ "learning_rate": 0.0006,
+ "loss": 5.42428731918335,
+ "step": 574
+ },
+ {
+ "epoch": 7.992139737991266,
+ "grad_norm": 0.01617511734366417,
+ "learning_rate": 0.0006,
+ "loss": 5.420044898986816,
+ "step": 575
+ },
+ {
+ "epoch": 8.0,
+ "grad_norm": 0.01837257295846939,
+ "learning_rate": 0.0006,
+ "loss": 5.240387916564941,
+ "step": 576
+ },
+ {
+ "epoch": 8.0,
+ "eval_loss": 5.423932075500488,
+ "eval_runtime": 58.0841,
+ "eval_samples_per_second": 42.042,
+ "eval_steps_per_second": 1.326,
+ "step": 576
+ },
+ {
+ "epoch": 8.013973799126637,
+ "grad_norm": 0.020953809842467308,
+ "learning_rate": 0.0006,
+ "loss": 5.4320783615112305,
+ "step": 577
+ },
+ {
+ "epoch": 8.027947598253276,
+ "grad_norm": 0.0227745920419693,
+ "learning_rate": 0.0006,
+ "loss": 5.310197830200195,
+ "step": 578
+ },
+ {
+ "epoch": 8.041921397379912,
+ "grad_norm": 0.021372603252530098,
+ "learning_rate": 0.0006,
+ "loss": 5.339468002319336,
+ "step": 579
+ },
+ {
+ "epoch": 8.055895196506551,
+ "grad_norm": 0.022401731461286545,
+ "learning_rate": 0.0006,
+ "loss": 5.279829978942871,
+ "step": 580
+ },
+ {
+ "epoch": 8.069868995633188,
+ "grad_norm": 0.023942379280924797,
+ "learning_rate": 0.0006,
+ "loss": 5.372323989868164,
+ "step": 581
+ },
+ {
+ "epoch": 8.083842794759825,
+ "grad_norm": 0.024380534887313843,
+ "learning_rate": 0.0006,
+ "loss": 5.253486633300781,
+ "step": 582
+ },
+ {
+ "epoch": 8.097816593886463,
+ "grad_norm": 0.03246625140309334,
+ "learning_rate": 0.0006,
+ "loss": 5.305774211883545,
+ "step": 583
+ },
+ {
+ "epoch": 8.1117903930131,
+ "grad_norm": 0.030399736016988754,
+ "learning_rate": 0.0006,
+ "loss": 5.370976448059082,
+ "step": 584
+ },
+ {
+ "epoch": 8.125764192139737,
+ "grad_norm": 0.027148913592100143,
+ "learning_rate": 0.0006,
+ "loss": 5.368010520935059,
+ "step": 585
+ },
+ {
+ "epoch": 8.139737991266376,
+ "grad_norm": 0.02925540879368782,
+ "learning_rate": 0.0006,
+ "loss": 5.303013801574707,
+ "step": 586
+ },
+ {
+ "epoch": 8.153711790393013,
+ "grad_norm": 0.02915453538298607,
+ "learning_rate": 0.0006,
+ "loss": 5.329667091369629,
+ "step": 587
+ },
+ {
+ "epoch": 8.167685589519651,
+ "grad_norm": 0.031732626259326935,
+ "learning_rate": 0.0006,
+ "loss": 5.3634934425354,
+ "step": 588
+ },
+ {
+ "epoch": 8.181659388646288,
+ "grad_norm": 0.036166731268167496,
+ "learning_rate": 0.0006,
+ "loss": 5.223980903625488,
+ "step": 589
+ },
+ {
+ "epoch": 8.195633187772925,
+ "grad_norm": 0.035654108971357346,
+ "learning_rate": 0.0006,
+ "loss": 5.361059188842773,
+ "step": 590
+ },
+ {
+ "epoch": 8.209606986899564,
+ "grad_norm": 0.0365324430167675,
+ "learning_rate": 0.0006,
+ "loss": 5.336151599884033,
+ "step": 591
+ },
+ {
+ "epoch": 8.2235807860262,
+ "grad_norm": 0.03578880429267883,
+ "learning_rate": 0.0006,
+ "loss": 5.434076309204102,
+ "step": 592
+ },
+ {
+ "epoch": 8.237554585152838,
+ "grad_norm": 0.035734012722969055,
+ "learning_rate": 0.0006,
+ "loss": 5.330893516540527,
+ "step": 593
+ },
+ {
+ "epoch": 8.251528384279476,
+ "grad_norm": 0.03196857124567032,
+ "learning_rate": 0.0006,
+ "loss": 5.306009769439697,
+ "step": 594
+ },
+ {
+ "epoch": 8.265502183406113,
+ "grad_norm": 0.032042086124420166,
+ "learning_rate": 0.0006,
+ "loss": 5.343267917633057,
+ "step": 595
+ },
+ {
+ "epoch": 8.279475982532752,
+ "grad_norm": 0.03160746395587921,
+ "learning_rate": 0.0006,
+ "loss": 5.2353715896606445,
+ "step": 596
+ },
+ {
+ "epoch": 8.293449781659389,
+ "grad_norm": 0.026689818128943443,
+ "learning_rate": 0.0006,
+ "loss": 5.254042625427246,
+ "step": 597
+ },
+ {
+ "epoch": 8.307423580786025,
+ "grad_norm": 0.02880188450217247,
+ "learning_rate": 0.0006,
+ "loss": 5.326833248138428,
+ "step": 598
+ },
+ {
+ "epoch": 8.321397379912664,
+ "grad_norm": 0.027516381815075874,
+ "learning_rate": 0.0006,
+ "loss": 5.342336177825928,
+ "step": 599
+ },
+ {
+ "epoch": 8.335371179039301,
+ "grad_norm": 0.027875030413269997,
+ "learning_rate": 0.0006,
+ "loss": 5.200719833374023,
+ "step": 600
+ },
+ {
+ "epoch": 8.34934497816594,
+ "grad_norm": 0.0268265288323164,
+ "learning_rate": 0.0006,
+ "loss": 5.249449729919434,
+ "step": 601
+ },
+ {
+ "epoch": 8.363318777292577,
+ "grad_norm": 0.024824608117341995,
+ "learning_rate": 0.0006,
+ "loss": 5.318952560424805,
+ "step": 602
+ },
+ {
+ "epoch": 8.377292576419213,
+ "grad_norm": 0.022990627214312553,
+ "learning_rate": 0.0006,
+ "loss": 5.244993209838867,
+ "step": 603
+ },
+ {
+ "epoch": 8.391266375545852,
+ "grad_norm": 0.022804604843258858,
+ "learning_rate": 0.0006,
+ "loss": 5.3236846923828125,
+ "step": 604
+ },
+ {
+ "epoch": 8.405240174672489,
+ "grad_norm": 0.02419872209429741,
+ "learning_rate": 0.0006,
+ "loss": 5.255486488342285,
+ "step": 605
+ },
+ {
+ "epoch": 8.419213973799126,
+ "grad_norm": 0.021952059119939804,
+ "learning_rate": 0.0006,
+ "loss": 5.3511552810668945,
+ "step": 606
+ },
+ {
+ "epoch": 8.433187772925764,
+ "grad_norm": 0.022375497967004776,
+ "learning_rate": 0.0006,
+ "loss": 5.294790267944336,
+ "step": 607
+ },
+ {
+ "epoch": 8.447161572052401,
+ "grad_norm": 0.019242815673351288,
+ "learning_rate": 0.0006,
+ "loss": 5.252618789672852,
+ "step": 608
+ },
+ {
+ "epoch": 8.46113537117904,
+ "grad_norm": 0.0173486340790987,
+ "learning_rate": 0.0006,
+ "loss": 5.30147647857666,
+ "step": 609
+ },
+ {
+ "epoch": 8.475109170305677,
+ "grad_norm": 0.01612604409456253,
+ "learning_rate": 0.0006,
+ "loss": 5.23960542678833,
+ "step": 610
+ },
+ {
+ "epoch": 8.489082969432314,
+ "grad_norm": 0.014293976128101349,
+ "learning_rate": 0.0006,
+ "loss": 5.119932651519775,
+ "step": 611
+ },
+ {
+ "epoch": 8.503056768558952,
+ "grad_norm": 0.014453536830842495,
+ "learning_rate": 0.0006,
+ "loss": 5.166794776916504,
+ "step": 612
+ },
+ {
+ "epoch": 8.51703056768559,
+ "grad_norm": 0.014685769565403461,
+ "learning_rate": 0.0006,
+ "loss": 5.264822959899902,
+ "step": 613
+ },
+ {
+ "epoch": 8.531004366812226,
+ "grad_norm": 0.01427740603685379,
+ "learning_rate": 0.0006,
+ "loss": 5.278433799743652,
+ "step": 614
+ },
+ {
+ "epoch": 8.544978165938865,
+ "grad_norm": 0.014864951372146606,
+ "learning_rate": 0.0006,
+ "loss": 5.214512348175049,
+ "step": 615
+ },
+ {
+ "epoch": 8.558951965065502,
+ "grad_norm": 0.01565164513885975,
+ "learning_rate": 0.0006,
+ "loss": 5.183066368103027,
+ "step": 616
+ },
+ {
+ "epoch": 8.57292576419214,
+ "grad_norm": 0.01779816672205925,
+ "learning_rate": 0.0006,
+ "loss": 5.163185119628906,
+ "step": 617
+ },
+ {
+ "epoch": 8.586899563318777,
+ "grad_norm": 0.017254827544093132,
+ "learning_rate": 0.0006,
+ "loss": 5.243555068969727,
+ "step": 618
+ },
+ {
+ "epoch": 8.600873362445414,
+ "grad_norm": 0.01650584116578102,
+ "learning_rate": 0.0006,
+ "loss": 5.234884738922119,
+ "step": 619
+ },
+ {
+ "epoch": 8.614847161572053,
+ "grad_norm": 0.017021115869283676,
+ "learning_rate": 0.0006,
+ "loss": 5.12271785736084,
+ "step": 620
+ },
+ {
+ "epoch": 8.62882096069869,
+ "grad_norm": 0.01773759163916111,
+ "learning_rate": 0.0006,
+ "loss": 5.242153167724609,
+ "step": 621
+ },
+ {
+ "epoch": 8.642794759825328,
+ "grad_norm": 0.015679042786359787,
+ "learning_rate": 0.0006,
+ "loss": 5.170779228210449,
+ "step": 622
+ },
+ {
+ "epoch": 8.656768558951965,
+ "grad_norm": 0.013760549947619438,
+ "learning_rate": 0.0006,
+ "loss": 5.238644599914551,
+ "step": 623
+ },
+ {
+ "epoch": 8.670742358078602,
+ "grad_norm": 0.014571522362530231,
+ "learning_rate": 0.0006,
+ "loss": 5.177056312561035,
+ "step": 624
+ },
+ {
+ "epoch": 8.68471615720524,
+ "grad_norm": 0.016209015622735023,
+ "learning_rate": 0.0006,
+ "loss": 5.163750648498535,
+ "step": 625
+ },
+ {
+ "epoch": 8.698689956331878,
+ "grad_norm": 0.016813941299915314,
+ "learning_rate": 0.0006,
+ "loss": 5.183428764343262,
+ "step": 626
+ },
+ {
+ "epoch": 8.712663755458514,
+ "grad_norm": 0.020985357463359833,
+ "learning_rate": 0.0006,
+ "loss": 5.229465007781982,
+ "step": 627
+ },
+ {
+ "epoch": 8.726637554585153,
+ "grad_norm": 0.02679632417857647,
+ "learning_rate": 0.0006,
+ "loss": 5.257368087768555,
+ "step": 628
+ },
+ {
+ "epoch": 8.74061135371179,
+ "grad_norm": 0.025756070390343666,
+ "learning_rate": 0.0006,
+ "loss": 5.253736972808838,
+ "step": 629
+ },
+ {
+ "epoch": 8.754585152838429,
+ "grad_norm": 0.02643490768969059,
+ "learning_rate": 0.0006,
+ "loss": 5.228633403778076,
+ "step": 630
+ },
+ {
+ "epoch": 8.768558951965066,
+ "grad_norm": 0.03303210437297821,
+ "learning_rate": 0.0006,
+ "loss": 5.163540363311768,
+ "step": 631
+ },
+ {
+ "epoch": 8.782532751091702,
+ "grad_norm": 0.030432431027293205,
+ "learning_rate": 0.0006,
+ "loss": 5.243169784545898,
+ "step": 632
+ },
+ {
+ "epoch": 8.796506550218341,
+ "grad_norm": 0.028631288558244705,
+ "learning_rate": 0.0006,
+ "loss": 5.166398048400879,
+ "step": 633
+ },
+ {
+ "epoch": 8.810480349344978,
+ "grad_norm": 0.026188310235738754,
+ "learning_rate": 0.0006,
+ "loss": 5.128122329711914,
+ "step": 634
+ },
+ {
+ "epoch": 8.824454148471617,
+ "grad_norm": 0.025308528915047646,
+ "learning_rate": 0.0006,
+ "loss": 5.211284637451172,
+ "step": 635
+ },
+ {
+ "epoch": 8.838427947598253,
+ "grad_norm": 0.020657729357481003,
+ "learning_rate": 0.0006,
+ "loss": 5.161575794219971,
+ "step": 636
+ },
+ {
+ "epoch": 8.85240174672489,
+ "grad_norm": 0.021901234984397888,
+ "learning_rate": 0.0006,
+ "loss": 5.20050048828125,
+ "step": 637
+ },
+ {
+ "epoch": 8.866375545851529,
+ "grad_norm": 0.021219318732619286,
+ "learning_rate": 0.0006,
+ "loss": 5.211699485778809,
+ "step": 638
+ },
+ {
+ "epoch": 8.880349344978166,
+ "grad_norm": 0.0200974028557539,
+ "learning_rate": 0.0006,
+ "loss": 5.092503547668457,
+ "step": 639
+ },
+ {
+ "epoch": 8.894323144104803,
+ "grad_norm": 0.023804882541298866,
+ "learning_rate": 0.0006,
+ "loss": 5.216068267822266,
+ "step": 640
+ },
+ {
+ "epoch": 8.908296943231441,
+ "grad_norm": 0.026088010519742966,
+ "learning_rate": 0.0006,
+ "loss": 5.155592918395996,
+ "step": 641
+ },
+ {
+ "epoch": 8.922270742358078,
+ "grad_norm": 0.02601276896893978,
+ "learning_rate": 0.0006,
+ "loss": 5.197238922119141,
+ "step": 642
+ },
+ {
+ "epoch": 8.936244541484717,
+ "grad_norm": 0.020387211814522743,
+ "learning_rate": 0.0006,
+ "loss": 5.1955437660217285,
+ "step": 643
+ },
+ {
+ "epoch": 8.950218340611354,
+ "grad_norm": 0.019214622676372528,
+ "learning_rate": 0.0006,
+ "loss": 5.206346035003662,
+ "step": 644
+ },
+ {
+ "epoch": 8.96419213973799,
+ "grad_norm": 0.019674314185976982,
+ "learning_rate": 0.0006,
+ "loss": 5.219857215881348,
+ "step": 645
+ },
+ {
+ "epoch": 8.97816593886463,
+ "grad_norm": 0.020318234339356422,
+ "learning_rate": 0.0006,
+ "loss": 5.191132545471191,
+ "step": 646
+ },
+ {
+ "epoch": 8.992139737991266,
+ "grad_norm": 0.021428676322102547,
+ "learning_rate": 0.0006,
+ "loss": 5.214695930480957,
+ "step": 647
+ },
+ {
+ "epoch": 9.0,
+ "grad_norm": 0.023235054686665535,
+ "learning_rate": 0.0006,
+ "loss": 5.229091644287109,
+ "step": 648
+ },
+ {
+ "epoch": 9.0,
+ "eval_loss": 5.289593696594238,
+ "eval_runtime": 56.9256,
+ "eval_samples_per_second": 42.898,
+ "eval_steps_per_second": 1.353,
+ "step": 648
+ },
+ {
+ "epoch": 9.013973799126637,
+ "grad_norm": 0.02320289984345436,
+ "learning_rate": 0.0006,
+ "loss": 4.983669281005859,
+ "step": 649
+ },
+ {
+ "epoch": 9.027947598253276,
+ "grad_norm": 0.018521282821893692,
+ "learning_rate": 0.0006,
+ "loss": 5.200575828552246,
+ "step": 650
+ },
+ {
+ "epoch": 9.041921397379912,
+ "grad_norm": 0.0180169939994812,
+ "learning_rate": 0.0006,
+ "loss": 5.148033618927002,
+ "step": 651
+ },
+ {
+ "epoch": 9.055895196506551,
+ "grad_norm": 0.019969860091805458,
+ "learning_rate": 0.0006,
+ "loss": 5.099125385284424,
+ "step": 652
+ },
+ {
+ "epoch": 9.069868995633188,
+ "grad_norm": 0.01643305830657482,
+ "learning_rate": 0.0006,
+ "loss": 5.092347145080566,
+ "step": 653
+ },
+ {
+ "epoch": 9.083842794759825,
+ "grad_norm": 0.016022363677620888,
+ "learning_rate": 0.0006,
+ "loss": 5.129279613494873,
+ "step": 654
+ },
+ {
+ "epoch": 9.097816593886463,
+ "grad_norm": 0.01657041162252426,
+ "learning_rate": 0.0006,
+ "loss": 5.159562587738037,
+ "step": 655
+ },
+ {
+ "epoch": 9.1117903930131,
+ "grad_norm": 0.017924228683114052,
+ "learning_rate": 0.0006,
+ "loss": 5.165497779846191,
+ "step": 656
+ },
+ {
+ "epoch": 9.125764192139737,
+ "grad_norm": 0.021592361852526665,
+ "learning_rate": 0.0006,
+ "loss": 5.204775810241699,
+ "step": 657
+ },
+ {
+ "epoch": 9.139737991266376,
+ "grad_norm": 0.026924120262265205,
+ "learning_rate": 0.0006,
+ "loss": 5.128296852111816,
+ "step": 658
+ },
+ {
+ "epoch": 9.153711790393013,
+ "grad_norm": 0.023156899958848953,
+ "learning_rate": 0.0006,
+ "loss": 5.1255340576171875,
+ "step": 659
+ },
+ {
+ "epoch": 9.167685589519651,
+ "grad_norm": 0.020670367404818535,
+ "learning_rate": 0.0006,
+ "loss": 5.128479957580566,
+ "step": 660
+ },
+ {
+ "epoch": 9.181659388646288,
+ "grad_norm": 0.025447173044085503,
+ "learning_rate": 0.0006,
+ "loss": 5.154821872711182,
+ "step": 661
+ },
+ {
+ "epoch": 9.195633187772925,
+ "grad_norm": 0.027940068393945694,
+ "learning_rate": 0.0006,
+ "loss": 5.143466472625732,
+ "step": 662
+ },
+ {
+ "epoch": 9.209606986899564,
+ "grad_norm": 0.026847844943404198,
+ "learning_rate": 0.0006,
+ "loss": 5.045665740966797,
+ "step": 663
+ },
+ {
+ "epoch": 9.2235807860262,
+ "grad_norm": 0.02340601570904255,
+ "learning_rate": 0.0006,
+ "loss": 5.21933126449585,
+ "step": 664
+ },
+ {
+ "epoch": 9.237554585152838,
+ "grad_norm": 0.02340371161699295,
+ "learning_rate": 0.0006,
+ "loss": 4.942949295043945,
+ "step": 665
+ },
+ {
+ "epoch": 9.251528384279476,
+ "grad_norm": 0.02221992425620556,
+ "learning_rate": 0.0006,
+ "loss": 5.197099208831787,
+ "step": 666
+ },
+ {
+ "epoch": 9.265502183406113,
+ "grad_norm": 0.023508677259087563,
+ "learning_rate": 0.0006,
+ "loss": 5.022897720336914,
+ "step": 667
+ },
+ {
+ "epoch": 9.279475982532752,
+ "grad_norm": 0.026120394468307495,
+ "learning_rate": 0.0006,
+ "loss": 5.087725639343262,
+ "step": 668
+ },
+ {
+ "epoch": 9.293449781659389,
+ "grad_norm": 0.027273228392004967,
+ "learning_rate": 0.0006,
+ "loss": 5.090963840484619,
+ "step": 669
+ },
+ {
+ "epoch": 9.307423580786025,
+ "grad_norm": 0.03241586685180664,
+ "learning_rate": 0.0006,
+ "loss": 5.047842979431152,
+ "step": 670
+ },
+ {
+ "epoch": 9.321397379912664,
+ "grad_norm": 0.030245667323470116,
+ "learning_rate": 0.0006,
+ "loss": 5.104803085327148,
+ "step": 671
+ },
+ {
+ "epoch": 9.335371179039301,
+ "grad_norm": 0.027698364108800888,
+ "learning_rate": 0.0006,
+ "loss": 5.0644097328186035,
+ "step": 672
+ },
+ {
+ "epoch": 9.34934497816594,
+ "grad_norm": 0.029692554846405983,
+ "learning_rate": 0.0006,
+ "loss": 5.111942291259766,
+ "step": 673
+ },
+ {
+ "epoch": 9.363318777292577,
+ "grad_norm": 0.03328656405210495,
+ "learning_rate": 0.0006,
+ "loss": 5.1574554443359375,
+ "step": 674
+ },
+ {
+ "epoch": 9.377292576419213,
+ "grad_norm": 0.031938180327415466,
+ "learning_rate": 0.0006,
+ "loss": 5.190371990203857,
+ "step": 675
+ },
+ {
+ "epoch": 9.391266375545852,
+ "grad_norm": 0.024858945980668068,
+ "learning_rate": 0.0006,
+ "loss": 5.121420860290527,
+ "step": 676
+ },
+ {
+ "epoch": 9.405240174672489,
+ "grad_norm": 0.023033203557133675,
+ "learning_rate": 0.0006,
+ "loss": 5.0974345207214355,
+ "step": 677
+ },
+ {
+ "epoch": 9.419213973799126,
+ "grad_norm": 0.022083545103669167,
+ "learning_rate": 0.0006,
+ "loss": 5.046319961547852,
+ "step": 678
+ },
+ {
+ "epoch": 9.433187772925764,
+ "grad_norm": 0.0228437427431345,
+ "learning_rate": 0.0006,
+ "loss": 5.119963645935059,
+ "step": 679
+ },
+ {
+ "epoch": 9.447161572052401,
+ "grad_norm": 0.0242351982742548,
+ "learning_rate": 0.0006,
+ "loss": 5.0444560050964355,
+ "step": 680
+ },
+ {
+ "epoch": 9.46113537117904,
+ "grad_norm": 0.02401185780763626,
+ "learning_rate": 0.0006,
+ "loss": 5.063238620758057,
+ "step": 681
+ },
+ {
+ "epoch": 9.475109170305677,
+ "grad_norm": 0.02560959942638874,
+ "learning_rate": 0.0006,
+ "loss": 5.035224914550781,
+ "step": 682
+ },
+ {
+ "epoch": 9.489082969432314,
+ "grad_norm": 0.024985626339912415,
+ "learning_rate": 0.0006,
+ "loss": 5.049284934997559,
+ "step": 683
+ },
+ {
+ "epoch": 9.503056768558952,
+ "grad_norm": 0.023178115487098694,
+ "learning_rate": 0.0006,
+ "loss": 5.2266621589660645,
+ "step": 684
+ },
+ {
+ "epoch": 9.51703056768559,
+ "grad_norm": 0.0215512253344059,
+ "learning_rate": 0.0006,
+ "loss": 5.1875715255737305,
+ "step": 685
+ },
+ {
+ "epoch": 9.531004366812226,
+ "grad_norm": 0.01662837713956833,
+ "learning_rate": 0.0006,
+ "loss": 5.137012481689453,
+ "step": 686
+ },
+ {
+ "epoch": 9.544978165938865,
+ "grad_norm": 0.01952764391899109,
+ "learning_rate": 0.0006,
+ "loss": 5.072500228881836,
+ "step": 687
+ },
+ {
+ "epoch": 9.558951965065502,
+ "grad_norm": 0.017521077767014503,
+ "learning_rate": 0.0006,
+ "loss": 5.045711994171143,
+ "step": 688
+ },
+ {
+ "epoch": 9.57292576419214,
+ "grad_norm": 0.017355265095829964,
+ "learning_rate": 0.0006,
+ "loss": 5.02881383895874,
+ "step": 689
+ },
+ {
+ "epoch": 9.586899563318777,
+ "grad_norm": 0.015587719157338142,
+ "learning_rate": 0.0006,
+ "loss": 5.132787704467773,
+ "step": 690
+ },
+ {
+ "epoch": 9.600873362445414,
+ "grad_norm": 0.017469845712184906,
+ "learning_rate": 0.0006,
+ "loss": 5.105447292327881,
+ "step": 691
+ },
+ {
+ "epoch": 9.614847161572053,
+ "grad_norm": 0.019630100578069687,
+ "learning_rate": 0.0006,
+ "loss": 5.032464981079102,
+ "step": 692
+ },
+ {
+ "epoch": 9.62882096069869,
+ "grad_norm": 0.01784994639456272,
+ "learning_rate": 0.0006,
+ "loss": 5.081345558166504,
+ "step": 693
+ },
+ {
+ "epoch": 9.642794759825328,
+ "grad_norm": 0.01872754842042923,
+ "learning_rate": 0.0006,
+ "loss": 4.9772114753723145,
+ "step": 694
+ },
+ {
+ "epoch": 9.656768558951965,
+ "grad_norm": 0.01999032311141491,
+ "learning_rate": 0.0006,
+ "loss": 5.089740753173828,
+ "step": 695
+ },
+ {
+ "epoch": 9.670742358078602,
+ "grad_norm": 0.01710767112672329,
+ "learning_rate": 0.0006,
+ "loss": 4.9869890213012695,
+ "step": 696
+ },
+ {
+ "epoch": 9.68471615720524,
+ "grad_norm": 0.015144161880016327,
+ "learning_rate": 0.0006,
+ "loss": 5.022564888000488,
+ "step": 697
+ },
+ {
+ "epoch": 9.698689956331878,
+ "grad_norm": 0.01618543453514576,
+ "learning_rate": 0.0006,
+ "loss": 5.035921573638916,
+ "step": 698
+ },
+ {
+ "epoch": 9.712663755458514,
+ "grad_norm": 0.014653387479484081,
+ "learning_rate": 0.0006,
+ "loss": 4.914456367492676,
+ "step": 699
+ },
+ {
+ "epoch": 9.726637554585153,
+ "grad_norm": 0.013434977270662785,
+ "learning_rate": 0.0006,
+ "loss": 5.063764572143555,
+ "step": 700
+ },
+ {
+ "epoch": 9.74061135371179,
+ "grad_norm": 0.015128864906728268,
+ "learning_rate": 0.0006,
+ "loss": 5.052367687225342,
+ "step": 701
+ },
+ {
+ "epoch": 9.754585152838429,
+ "grad_norm": 0.014210768043994904,
+ "learning_rate": 0.0006,
+ "loss": 5.016489028930664,
+ "step": 702
+ },
+ {
+ "epoch": 9.768558951965066,
+ "grad_norm": 0.015191680751740932,
+ "learning_rate": 0.0006,
+ "loss": 4.990176200866699,
+ "step": 703
+ },
+ {
+ "epoch": 9.782532751091702,
+ "grad_norm": 0.018285127356648445,
+ "learning_rate": 0.0006,
+ "loss": 5.015164375305176,
+ "step": 704
+ },
+ {
+ "epoch": 9.796506550218341,
+ "grad_norm": 0.0204079058021307,
+ "learning_rate": 0.0006,
+ "loss": 5.091142654418945,
+ "step": 705
+ },
+ {
+ "epoch": 9.810480349344978,
+ "grad_norm": 0.01980586163699627,
+ "learning_rate": 0.0006,
+ "loss": 5.03350830078125,
+ "step": 706
+ },
+ {
+ "epoch": 9.824454148471617,
+ "grad_norm": 0.02048729732632637,
+ "learning_rate": 0.0006,
+ "loss": 5.03611421585083,
+ "step": 707
+ },
+ {
+ "epoch": 9.838427947598253,
+ "grad_norm": 0.02058163844048977,
+ "learning_rate": 0.0006,
+ "loss": 4.9613447189331055,
+ "step": 708
+ },
+ {
+ "epoch": 9.85240174672489,
+ "grad_norm": 0.018004924058914185,
+ "learning_rate": 0.0006,
+ "loss": 4.966541767120361,
+ "step": 709
+ },
+ {
+ "epoch": 9.866375545851529,
+ "grad_norm": 0.0198152307420969,
+ "learning_rate": 0.0006,
+ "loss": 5.01750373840332,
+ "step": 710
+ },
+ {
+ "epoch": 9.880349344978166,
+ "grad_norm": 0.018532969057559967,
+ "learning_rate": 0.0006,
+ "loss": 5.097424507141113,
+ "step": 711
+ },
+ {
+ "epoch": 9.894323144104803,
+ "grad_norm": 0.018939971923828125,
+ "learning_rate": 0.0006,
+ "loss": 5.177223205566406,
+ "step": 712
+ },
+ {
+ "epoch": 9.908296943231441,
+ "grad_norm": 0.017969170585274696,
+ "learning_rate": 0.0006,
+ "loss": 5.118013858795166,
+ "step": 713
+ },
+ {
+ "epoch": 9.922270742358078,
+ "grad_norm": 0.018021270632743835,
+ "learning_rate": 0.0006,
+ "loss": 5.034787178039551,
+ "step": 714
+ },
+ {
+ "epoch": 9.936244541484717,
+ "grad_norm": 0.017009710893034935,
+ "learning_rate": 0.0006,
+ "loss": 5.003920555114746,
+ "step": 715
+ },
+ {
+ "epoch": 9.950218340611354,
+ "grad_norm": 0.017722049728035927,
+ "learning_rate": 0.0006,
+ "loss": 5.031939506530762,
+ "step": 716
+ },
+ {
+ "epoch": 9.96419213973799,
+ "grad_norm": 0.017134087160229683,
+ "learning_rate": 0.0006,
+ "loss": 5.093474388122559,
+ "step": 717
+ },
+ {
+ "epoch": 9.97816593886463,
+ "grad_norm": 0.018511613830924034,
+ "learning_rate": 0.0006,
+ "loss": 5.078604698181152,
+ "step": 718
+ },
+ {
+ "epoch": 9.992139737991266,
+ "grad_norm": 0.021206334233283997,
+ "learning_rate": 0.0006,
+ "loss": 5.052572727203369,
+ "step": 719
+ },
+ {
+ "epoch": 10.0,
+ "grad_norm": 0.02239062264561653,
+ "learning_rate": 0.0006,
+ "loss": 5.063547134399414,
+ "step": 720
+ },
+ {
+ "epoch": 10.0,
+ "eval_loss": 5.173044681549072,
+ "eval_runtime": 56.7287,
+ "eval_samples_per_second": 43.047,
+ "eval_steps_per_second": 1.357,
+ "step": 720
+ },
+ {
+ "epoch": 10.013973799126637,
+ "grad_norm": 0.020757969468832016,
+ "learning_rate": 0.0006,
+ "loss": 5.045648574829102,
+ "step": 721
+ },
+ {
+ "epoch": 10.027947598253276,
+ "grad_norm": 0.022697702050209045,
+ "learning_rate": 0.0006,
+ "loss": 4.993833065032959,
+ "step": 722
+ },
+ {
+ "epoch": 10.041921397379912,
+ "grad_norm": 0.02434331737458706,
+ "learning_rate": 0.0006,
+ "loss": 5.073219299316406,
+ "step": 723
+ },
+ {
+ "epoch": 10.055895196506551,
+ "grad_norm": 0.025499660521745682,
+ "learning_rate": 0.0006,
+ "loss": 5.019072532653809,
+ "step": 724
+ },
+ {
+ "epoch": 10.069868995633188,
+ "grad_norm": 0.029281053692102432,
+ "learning_rate": 0.0006,
+ "loss": 4.959851264953613,
+ "step": 725
+ },
+ {
+ "epoch": 10.083842794759825,
+ "grad_norm": 0.033878110349178314,
+ "learning_rate": 0.0006,
+ "loss": 5.132717609405518,
+ "step": 726
+ },
+ {
+ "epoch": 10.097816593886463,
+ "grad_norm": 0.03385764732956886,
+ "learning_rate": 0.0006,
+ "loss": 5.0482892990112305,
+ "step": 727
+ },
+ {
+ "epoch": 10.1117903930131,
+ "grad_norm": 0.03466648980975151,
+ "learning_rate": 0.0006,
+ "loss": 4.946855545043945,
+ "step": 728
+ },
+ {
+ "epoch": 10.125764192139737,
+ "grad_norm": 0.028546305373311043,
+ "learning_rate": 0.0006,
+ "loss": 5.027774333953857,
+ "step": 729
+ },
+ {
+ "epoch": 10.139737991266376,
+ "grad_norm": 0.030575869604945183,
+ "learning_rate": 0.0006,
+ "loss": 4.9768476486206055,
+ "step": 730
+ },
+ {
+ "epoch": 10.153711790393013,
+ "grad_norm": 0.03614303469657898,
+ "learning_rate": 0.0006,
+ "loss": 5.078113555908203,
+ "step": 731
+ },
+ {
+ "epoch": 10.167685589519651,
+ "grad_norm": 0.040991660207509995,
+ "learning_rate": 0.0006,
+ "loss": 5.033185958862305,
+ "step": 732
+ },
+ {
+ "epoch": 10.181659388646288,
+ "grad_norm": 0.04551699757575989,
+ "learning_rate": 0.0006,
+ "loss": 5.0820746421813965,
+ "step": 733
+ },
+ {
+ "epoch": 10.195633187772925,
+ "grad_norm": 0.040193989872932434,
+ "learning_rate": 0.0006,
+ "loss": 5.059700965881348,
+ "step": 734
+ },
+ {
+ "epoch": 10.209606986899564,
+ "grad_norm": 0.035851605236530304,
+ "learning_rate": 0.0006,
+ "loss": 5.142467498779297,
+ "step": 735
+ },
+ {
+ "epoch": 10.2235807860262,
+ "grad_norm": 0.034200169146060944,
+ "learning_rate": 0.0006,
+ "loss": 5.072312355041504,
+ "step": 736
+ },
+ {
+ "epoch": 10.237554585152838,
+ "grad_norm": 0.033621110022068024,
+ "learning_rate": 0.0006,
+ "loss": 5.08037805557251,
+ "step": 737
+ },
+ {
+ "epoch": 10.251528384279476,
+ "grad_norm": 0.037372369319200516,
+ "learning_rate": 0.0006,
+ "loss": 4.998425483703613,
+ "step": 738
+ },
+ {
+ "epoch": 10.265502183406113,
+ "grad_norm": 0.037044707685709,
+ "learning_rate": 0.0006,
+ "loss": 4.923993110656738,
+ "step": 739
+ },
+ {
+ "epoch": 10.279475982532752,
+ "grad_norm": 0.030454808846116066,
+ "learning_rate": 0.0006,
+ "loss": 5.0889997482299805,
+ "step": 740
+ },
+ {
+ "epoch": 10.293449781659389,
+ "grad_norm": 0.030969638377428055,
+ "learning_rate": 0.0006,
+ "loss": 4.984112739562988,
+ "step": 741
+ },
+ {
+ "epoch": 10.307423580786025,
+ "grad_norm": 0.0276983380317688,
+ "learning_rate": 0.0006,
+ "loss": 4.955412864685059,
+ "step": 742
+ },
+ {
+ "epoch": 10.321397379912664,
+ "grad_norm": 0.02716052532196045,
+ "learning_rate": 0.0006,
+ "loss": 5.0098876953125,
+ "step": 743
+ },
+ {
+ "epoch": 10.335371179039301,
+ "grad_norm": 0.02403552085161209,
+ "learning_rate": 0.0006,
+ "loss": 5.165160655975342,
+ "step": 744
+ },
+ {
+ "epoch": 10.34934497816594,
+ "grad_norm": 0.0257862601429224,
+ "learning_rate": 0.0006,
+ "loss": 5.001547813415527,
+ "step": 745
+ },
+ {
+ "epoch": 10.363318777292577,
+ "grad_norm": 0.02300378680229187,
+ "learning_rate": 0.0006,
+ "loss": 5.038888931274414,
+ "step": 746
+ },
+ {
+ "epoch": 10.377292576419213,
+ "grad_norm": 0.019766854122281075,
+ "learning_rate": 0.0006,
+ "loss": 4.924291610717773,
+ "step": 747
+ },
+ {
+ "epoch": 10.391266375545852,
+ "grad_norm": 0.01888395845890045,
+ "learning_rate": 0.0006,
+ "loss": 5.059971332550049,
+ "step": 748
+ },
+ {
+ "epoch": 10.405240174672489,
+ "grad_norm": 0.015577499754726887,
+ "learning_rate": 0.0006,
+ "loss": 4.987686634063721,
+ "step": 749
+ },
+ {
+ "epoch": 10.419213973799126,
+ "grad_norm": 0.016012346372008324,
+ "learning_rate": 0.0006,
+ "loss": 5.015157699584961,
+ "step": 750
+ },
+ {
+ "epoch": 10.433187772925764,
+ "grad_norm": 0.016277998685836792,
+ "learning_rate": 0.0006,
+ "loss": 5.004931449890137,
+ "step": 751
+ },
+ {
+ "epoch": 10.447161572052401,
+ "grad_norm": 0.017504561692476273,
+ "learning_rate": 0.0006,
+ "loss": 4.902827739715576,
+ "step": 752
+ },
+ {
+ "epoch": 10.46113537117904,
+ "grad_norm": 0.01569022797048092,
+ "learning_rate": 0.0006,
+ "loss": 4.954863548278809,
+ "step": 753
+ },
+ {
+ "epoch": 10.475109170305677,
+ "grad_norm": 0.014114422723650932,
+ "learning_rate": 0.0006,
+ "loss": 5.00656795501709,
+ "step": 754
+ },
+ {
+ "epoch": 10.489082969432314,
+ "grad_norm": 0.01573832333087921,
+ "learning_rate": 0.0006,
+ "loss": 4.978353023529053,
+ "step": 755
+ },
+ {
+ "epoch": 10.503056768558952,
+ "grad_norm": 0.016480036079883575,
+ "learning_rate": 0.0006,
+ "loss": 4.953327655792236,
+ "step": 756
+ },
+ {
+ "epoch": 10.51703056768559,
+ "grad_norm": 0.014008025638759136,
+ "learning_rate": 0.0006,
+ "loss": 5.050840377807617,
+ "step": 757
+ },
+ {
+ "epoch": 10.531004366812226,
+ "grad_norm": 0.013147998601198196,
+ "learning_rate": 0.0006,
+ "loss": 4.974964141845703,
+ "step": 758
+ },
+ {
+ "epoch": 10.544978165938865,
+ "grad_norm": 0.013807603158056736,
+ "learning_rate": 0.0006,
+ "loss": 4.927907943725586,
+ "step": 759
+ },
+ {
+ "epoch": 10.558951965065502,
+ "grad_norm": 0.013555224984884262,
+ "learning_rate": 0.0006,
+ "loss": 4.980690956115723,
+ "step": 760
+ },
+ {
+ "epoch": 10.57292576419214,
+ "grad_norm": 0.013045408762991428,
+ "learning_rate": 0.0006,
+ "loss": 4.9436116218566895,
+ "step": 761
+ },
+ {
+ "epoch": 10.586899563318777,
+ "grad_norm": 0.012858862057328224,
+ "learning_rate": 0.0006,
+ "loss": 4.9595770835876465,
+ "step": 762
+ },
+ {
+ "epoch": 10.600873362445414,
+ "grad_norm": 0.01468253880739212,
+ "learning_rate": 0.0006,
+ "loss": 4.90725040435791,
+ "step": 763
+ },
+ {
+ "epoch": 10.614847161572053,
+ "grad_norm": 0.013224679045379162,
+ "learning_rate": 0.0006,
+ "loss": 4.877615928649902,
+ "step": 764
+ },
+ {
+ "epoch": 10.62882096069869,
+ "grad_norm": 0.013555348850786686,
+ "learning_rate": 0.0006,
+ "loss": 4.9456610679626465,
+ "step": 765
+ },
+ {
+ "epoch": 10.642794759825328,
+ "grad_norm": 0.014387160539627075,
+ "learning_rate": 0.0006,
+ "loss": 4.9233717918396,
+ "step": 766
+ },
+ {
+ "epoch": 10.656768558951965,
+ "grad_norm": 0.014354088343679905,
+ "learning_rate": 0.0006,
+ "loss": 4.983511924743652,
+ "step": 767
+ },
+ {
+ "epoch": 10.670742358078602,
+ "grad_norm": 0.01332700252532959,
+ "learning_rate": 0.0006,
+ "loss": 4.934416770935059,
+ "step": 768
+ },
+ {
+ "epoch": 10.68471615720524,
+ "grad_norm": 0.011543313041329384,
+ "learning_rate": 0.0006,
+ "loss": 4.96388053894043,
+ "step": 769
+ },
+ {
+ "epoch": 10.698689956331878,
+ "grad_norm": 0.012330878525972366,
+ "learning_rate": 0.0006,
+ "loss": 4.946396350860596,
+ "step": 770
+ },
+ {
+ "epoch": 10.712663755458514,
+ "grad_norm": 0.01379779726266861,
+ "learning_rate": 0.0006,
+ "loss": 4.981019020080566,
+ "step": 771
+ },
+ {
+ "epoch": 10.726637554585153,
+ "grad_norm": 0.014612431637942791,
+ "learning_rate": 0.0006,
+ "loss": 4.894567966461182,
+ "step": 772
+ },
+ {
+ "epoch": 10.74061135371179,
+ "grad_norm": 0.013062899932265282,
+ "learning_rate": 0.0006,
+ "loss": 5.033038139343262,
+ "step": 773
+ },
+ {
+ "epoch": 10.754585152838429,
+ "grad_norm": 0.013898049481213093,
+ "learning_rate": 0.0006,
+ "loss": 4.823636054992676,
+ "step": 774
+ },
+ {
+ "epoch": 10.768558951965066,
+ "grad_norm": 0.01502019353210926,
+ "learning_rate": 0.0006,
+ "loss": 5.020156383514404,
+ "step": 775
+ },
+ {
+ "epoch": 10.782532751091702,
+ "grad_norm": 0.013810204342007637,
+ "learning_rate": 0.0006,
+ "loss": 4.935647010803223,
+ "step": 776
+ },
+ {
+ "epoch": 10.796506550218341,
+ "grad_norm": 0.013616513460874557,
+ "learning_rate": 0.0006,
+ "loss": 4.901486396789551,
+ "step": 777
+ },
+ {
+ "epoch": 10.810480349344978,
+ "grad_norm": 0.013601024635136127,
+ "learning_rate": 0.0006,
+ "loss": 4.931873321533203,
+ "step": 778
+ },
+ {
+ "epoch": 10.824454148471617,
+ "grad_norm": 0.013668350875377655,
+ "learning_rate": 0.0006,
+ "loss": 4.937256813049316,
+ "step": 779
+ },
+ {
+ "epoch": 10.838427947598253,
+ "grad_norm": 0.01581823080778122,
+ "learning_rate": 0.0006,
+ "loss": 4.9611406326293945,
+ "step": 780
+ },
+ {
+ "epoch": 10.85240174672489,
+ "grad_norm": 0.019478484988212585,
+ "learning_rate": 0.0006,
+ "loss": 4.9212236404418945,
+ "step": 781
+ },
+ {
+ "epoch": 10.866375545851529,
+ "grad_norm": 0.026921333745121956,
+ "learning_rate": 0.0006,
+ "loss": 4.88501501083374,
+ "step": 782
+ },
+ {
+ "epoch": 10.880349344978166,
+ "grad_norm": 0.03155914694070816,
+ "learning_rate": 0.0006,
+ "loss": 4.989967346191406,
+ "step": 783
+ },
+ {
+ "epoch": 10.894323144104803,
+ "grad_norm": 0.029689429327845573,
+ "learning_rate": 0.0006,
+ "loss": 4.936000823974609,
+ "step": 784
+ },
+ {
+ "epoch": 10.908296943231441,
+ "grad_norm": 0.026626063510775566,
+ "learning_rate": 0.0006,
+ "loss": 4.984785079956055,
+ "step": 785
+ },
+ {
+ "epoch": 10.922270742358078,
+ "grad_norm": 0.026168212294578552,
+ "learning_rate": 0.0006,
+ "loss": 4.908282279968262,
+ "step": 786
+ },
+ {
+ "epoch": 10.936244541484717,
+ "grad_norm": 0.02828158251941204,
+ "learning_rate": 0.0006,
+ "loss": 4.926390647888184,
+ "step": 787
+ },
+ {
+ "epoch": 10.950218340611354,
+ "grad_norm": 0.02649078331887722,
+ "learning_rate": 0.0006,
+ "loss": 5.00400972366333,
+ "step": 788
+ },
+ {
+ "epoch": 10.96419213973799,
+ "grad_norm": 0.0256856270134449,
+ "learning_rate": 0.0006,
+ "loss": 5.0317487716674805,
+ "step": 789
+ },
+ {
+ "epoch": 10.97816593886463,
+ "grad_norm": 0.024728883057832718,
+ "learning_rate": 0.0006,
+ "loss": 4.9299635887146,
+ "step": 790
+ },
+ {
+ "epoch": 10.992139737991266,
+ "grad_norm": 0.026241116225719452,
+ "learning_rate": 0.0006,
+ "loss": 4.947319984436035,
+ "step": 791
+ },
+ {
+ "epoch": 11.0,
+ "grad_norm": 0.023526743054389954,
+ "learning_rate": 0.0006,
+ "loss": 4.94429349899292,
+ "step": 792
+ },
+ {
+ "epoch": 11.0,
+ "eval_loss": 5.104345798492432,
+ "eval_runtime": 57.0955,
+ "eval_samples_per_second": 42.77,
+ "eval_steps_per_second": 1.349,
+ "step": 792
+ },
+ {
+ "epoch": 11.013973799126637,
+ "grad_norm": 0.0216788612306118,
+ "learning_rate": 0.0006,
+ "loss": 4.881950855255127,
+ "step": 793
+ },
+ {
+ "epoch": 11.027947598253276,
+ "grad_norm": 0.01830463856458664,
+ "learning_rate": 0.0006,
+ "loss": 4.971016883850098,
+ "step": 794
+ },
+ {
+ "epoch": 11.041921397379912,
+ "grad_norm": 0.02204521745443344,
+ "learning_rate": 0.0006,
+ "loss": 4.894623756408691,
+ "step": 795
+ },
+ {
+ "epoch": 11.055895196506551,
+ "grad_norm": 0.0210255216807127,
+ "learning_rate": 0.0006,
+ "loss": 4.834009170532227,
+ "step": 796
+ },
+ {
+ "epoch": 11.069868995633188,
+ "grad_norm": 0.019797448068857193,
+ "learning_rate": 0.0006,
+ "loss": 5.086714744567871,
+ "step": 797
+ },
+ {
+ "epoch": 11.083842794759825,
+ "grad_norm": 0.02036641724407673,
+ "learning_rate": 0.0006,
+ "loss": 4.920056343078613,
+ "step": 798
+ },
+ {
+ "epoch": 11.097816593886463,
+ "grad_norm": 0.01966066285967827,
+ "learning_rate": 0.0006,
+ "loss": 4.879859447479248,
+ "step": 799
+ },
+ {
+ "epoch": 11.1117903930131,
+ "grad_norm": 0.0202149897813797,
+ "learning_rate": 0.0006,
+ "loss": 4.933640480041504,
+ "step": 800
+ },
+ {
+ "epoch": 11.125764192139737,
+ "grad_norm": 0.01893039606511593,
+ "learning_rate": 0.0006,
+ "loss": 4.901970863342285,
+ "step": 801
+ },
+ {
+ "epoch": 11.139737991266376,
+ "grad_norm": 0.016040155664086342,
+ "learning_rate": 0.0006,
+ "loss": 4.905491828918457,
+ "step": 802
+ },
+ {
+ "epoch": 11.153711790393013,
+ "grad_norm": 0.01720350608229637,
+ "learning_rate": 0.0006,
+ "loss": 4.798816680908203,
+ "step": 803
+ },
+ {
+ "epoch": 11.167685589519651,
+ "grad_norm": 0.017002852633595467,
+ "learning_rate": 0.0006,
+ "loss": 4.851859092712402,
+ "step": 804
+ },
+ {
+ "epoch": 11.181659388646288,
+ "grad_norm": 0.015594680793583393,
+ "learning_rate": 0.0006,
+ "loss": 4.90328311920166,
+ "step": 805
+ },
+ {
+ "epoch": 11.195633187772925,
+ "grad_norm": 0.014904462732374668,
+ "learning_rate": 0.0006,
+ "loss": 4.904215335845947,
+ "step": 806
+ },
+ {
+ "epoch": 11.209606986899564,
+ "grad_norm": 0.01366228237748146,
+ "learning_rate": 0.0006,
+ "loss": 4.9813055992126465,
+ "step": 807
+ },
+ {
+ "epoch": 11.2235807860262,
+ "grad_norm": 0.014719638973474503,
+ "learning_rate": 0.0006,
+ "loss": 4.927306175231934,
+ "step": 808
+ },
+ {
+ "epoch": 11.237554585152838,
+ "grad_norm": 0.015016036108136177,
+ "learning_rate": 0.0006,
+ "loss": 4.874215126037598,
+ "step": 809
+ },
+ {
+ "epoch": 11.251528384279476,
+ "grad_norm": 0.014925424009561539,
+ "learning_rate": 0.0006,
+ "loss": 4.838257789611816,
+ "step": 810
+ },
+ {
+ "epoch": 11.265502183406113,
+ "grad_norm": 0.014335823245346546,
+ "learning_rate": 0.0006,
+ "loss": 4.940608024597168,
+ "step": 811
+ },
+ {
+ "epoch": 11.279475982532752,
+ "grad_norm": 0.014987512491643429,
+ "learning_rate": 0.0006,
+ "loss": 4.831531524658203,
+ "step": 812
+ },
+ {
+ "epoch": 11.293449781659389,
+ "grad_norm": 0.016087457537651062,
+ "learning_rate": 0.0006,
+ "loss": 4.863227367401123,
+ "step": 813
+ },
+ {
+ "epoch": 11.307423580786025,
+ "grad_norm": 0.01671300269663334,
+ "learning_rate": 0.0006,
+ "loss": 4.8535661697387695,
+ "step": 814
+ },
+ {
+ "epoch": 11.321397379912664,
+ "grad_norm": 0.015229142270982265,
+ "learning_rate": 0.0006,
+ "loss": 4.948635101318359,
+ "step": 815
+ },
+ {
+ "epoch": 11.335371179039301,
+ "grad_norm": 0.01532587967813015,
+ "learning_rate": 0.0006,
+ "loss": 4.9035797119140625,
+ "step": 816
+ },
+ {
+ "epoch": 11.34934497816594,
+ "grad_norm": 0.015715476125478745,
+ "learning_rate": 0.0006,
+ "loss": 4.894709587097168,
+ "step": 817
+ },
+ {
+ "epoch": 11.363318777292577,
+ "grad_norm": 0.014156977646052837,
+ "learning_rate": 0.0006,
+ "loss": 4.960206985473633,
+ "step": 818
+ },
+ {
+ "epoch": 11.377292576419213,
+ "grad_norm": 0.017076566815376282,
+ "learning_rate": 0.0006,
+ "loss": 5.005850791931152,
+ "step": 819
+ },
+ {
+ "epoch": 11.391266375545852,
+ "grad_norm": 0.0172084029763937,
+ "learning_rate": 0.0006,
+ "loss": 4.903355598449707,
+ "step": 820
+ },
+ {
+ "epoch": 11.405240174672489,
+ "grad_norm": 0.0180258359760046,
+ "learning_rate": 0.0006,
+ "loss": 4.842667579650879,
+ "step": 821
+ },
+ {
+ "epoch": 11.419213973799126,
+ "grad_norm": 0.022130120545625687,
+ "learning_rate": 0.0006,
+ "loss": 4.799970626831055,
+ "step": 822
+ },
+ {
+ "epoch": 11.433187772925764,
+ "grad_norm": 0.02384897693991661,
+ "learning_rate": 0.0006,
+ "loss": 4.860746383666992,
+ "step": 823
+ },
+ {
+ "epoch": 11.447161572052401,
+ "grad_norm": 0.021881572902202606,
+ "learning_rate": 0.0006,
+ "loss": 4.918107032775879,
+ "step": 824
+ },
+ {
+ "epoch": 11.46113537117904,
+ "grad_norm": 0.023085080087184906,
+ "learning_rate": 0.0006,
+ "loss": 4.878946304321289,
+ "step": 825
+ },
+ {
+ "epoch": 11.475109170305677,
+ "grad_norm": 0.02291012555360794,
+ "learning_rate": 0.0006,
+ "loss": 4.958887100219727,
+ "step": 826
+ },
+ {
+ "epoch": 11.489082969432314,
+ "grad_norm": 0.0199968870729208,
+ "learning_rate": 0.0006,
+ "loss": 4.80955696105957,
+ "step": 827
+ },
+ {
+ "epoch": 11.503056768558952,
+ "grad_norm": 0.018702229484915733,
+ "learning_rate": 0.0006,
+ "loss": 5.0276408195495605,
+ "step": 828
+ },
+ {
+ "epoch": 11.51703056768559,
+ "grad_norm": 0.017787311226129532,
+ "learning_rate": 0.0006,
+ "loss": 4.96018123626709,
+ "step": 829
+ },
+ {
+ "epoch": 11.531004366812226,
+ "grad_norm": 0.01704501546919346,
+ "learning_rate": 0.0006,
+ "loss": 4.914553165435791,
+ "step": 830
+ },
+ {
+ "epoch": 11.544978165938865,
+ "grad_norm": 0.017519677057862282,
+ "learning_rate": 0.0006,
+ "loss": 4.798098087310791,
+ "step": 831
+ },
+ {
+ "epoch": 11.558951965065502,
+ "grad_norm": 0.018511991947889328,
+ "learning_rate": 0.0006,
+ "loss": 4.817785263061523,
+ "step": 832
+ },
+ {
+ "epoch": 11.57292576419214,
+ "grad_norm": 0.01876644790172577,
+ "learning_rate": 0.0006,
+ "loss": 4.818995475769043,
+ "step": 833
+ },
+ {
+ "epoch": 11.586899563318777,
+ "grad_norm": 0.01753568835556507,
+ "learning_rate": 0.0006,
+ "loss": 4.811938285827637,
+ "step": 834
+ },
+ {
+ "epoch": 11.600873362445414,
+ "grad_norm": 0.017046531662344933,
+ "learning_rate": 0.0006,
+ "loss": 4.914674282073975,
+ "step": 835
+ },
+ {
+ "epoch": 11.614847161572053,
+ "grad_norm": 0.01802094094455242,
+ "learning_rate": 0.0006,
+ "loss": 4.993284225463867,
+ "step": 836
+ },
+ {
+ "epoch": 11.62882096069869,
+ "grad_norm": 0.019493764266371727,
+ "learning_rate": 0.0006,
+ "loss": 4.895628452301025,
+ "step": 837
+ },
+ {
+ "epoch": 11.642794759825328,
+ "grad_norm": 0.019461210817098618,
+ "learning_rate": 0.0006,
+ "loss": 4.869269371032715,
+ "step": 838
+ },
+ {
+ "epoch": 11.656768558951965,
+ "grad_norm": 0.01981567218899727,
+ "learning_rate": 0.0006,
+ "loss": 4.863052845001221,
+ "step": 839
+ },
+ {
+ "epoch": 11.670742358078602,
+ "grad_norm": 0.019894491881132126,
+ "learning_rate": 0.0006,
+ "loss": 4.8775787353515625,
+ "step": 840
+ },
+ {
+ "epoch": 11.68471615720524,
+ "grad_norm": 0.01875786855816841,
+ "learning_rate": 0.0006,
+ "loss": 4.817763328552246,
+ "step": 841
+ },
+ {
+ "epoch": 11.698689956331878,
+ "grad_norm": 0.020926695317029953,
+ "learning_rate": 0.0006,
+ "loss": 4.7470526695251465,
+ "step": 842
+ },
+ {
+ "epoch": 11.712663755458514,
+ "grad_norm": 0.021477103233337402,
+ "learning_rate": 0.0006,
+ "loss": 4.762682914733887,
+ "step": 843
+ },
+ {
+ "epoch": 11.726637554585153,
+ "grad_norm": 0.020628711208701134,
+ "learning_rate": 0.0006,
+ "loss": 4.807985782623291,
+ "step": 844
+ },
+ {
+ "epoch": 11.74061135371179,
+ "grad_norm": 0.021489612758159637,
+ "learning_rate": 0.0006,
+ "loss": 4.843091011047363,
+ "step": 845
+ },
+ {
+ "epoch": 11.754585152838429,
+ "grad_norm": 0.021194491535425186,
+ "learning_rate": 0.0006,
+ "loss": 4.8300251960754395,
+ "step": 846
+ },
+ {
+ "epoch": 11.768558951965066,
+ "grad_norm": 0.01833350397646427,
+ "learning_rate": 0.0006,
+ "loss": 5.000202178955078,
+ "step": 847
+ },
+ {
+ "epoch": 11.782532751091702,
+ "grad_norm": 0.018096931278705597,
+ "learning_rate": 0.0006,
+ "loss": 4.880648136138916,
+ "step": 848
+ },
+ {
+ "epoch": 11.796506550218341,
+ "grad_norm": 0.0197161752730608,
+ "learning_rate": 0.0006,
+ "loss": 4.788316249847412,
+ "step": 849
+ },
+ {
+ "epoch": 11.810480349344978,
+ "grad_norm": 0.0203181654214859,
+ "learning_rate": 0.0006,
+ "loss": 4.870944976806641,
+ "step": 850
+ },
+ {
+ "epoch": 11.824454148471617,
+ "grad_norm": 0.01908455230295658,
+ "learning_rate": 0.0006,
+ "loss": 4.890144348144531,
+ "step": 851
+ },
+ {
+ "epoch": 11.838427947598253,
+ "grad_norm": 0.021281801164150238,
+ "learning_rate": 0.0006,
+ "loss": 4.89578914642334,
+ "step": 852
+ },
+ {
+ "epoch": 11.85240174672489,
+ "grad_norm": 0.02124273031949997,
+ "learning_rate": 0.0006,
+ "loss": 4.883749008178711,
+ "step": 853
+ },
+ {
+ "epoch": 11.866375545851529,
+ "grad_norm": 0.020969398319721222,
+ "learning_rate": 0.0006,
+ "loss": 4.809142589569092,
+ "step": 854
+ },
+ {
+ "epoch": 11.880349344978166,
+ "grad_norm": 0.01967657171189785,
+ "learning_rate": 0.0006,
+ "loss": 4.821864604949951,
+ "step": 855
+ },
+ {
+ "epoch": 11.894323144104803,
+ "grad_norm": 0.020887333899736404,
+ "learning_rate": 0.0006,
+ "loss": 4.917985916137695,
+ "step": 856
+ },
+ {
+ "epoch": 11.908296943231441,
+ "grad_norm": 0.019203083589673042,
+ "learning_rate": 0.0006,
+ "loss": 4.824960708618164,
+ "step": 857
+ },
+ {
+ "epoch": 11.922270742358078,
+ "grad_norm": 0.020162401720881462,
+ "learning_rate": 0.0006,
+ "loss": 4.772617816925049,
+ "step": 858
+ },
+ {
+ "epoch": 11.936244541484717,
+ "grad_norm": 0.02251487225294113,
+ "learning_rate": 0.0006,
+ "loss": 4.735418319702148,
+ "step": 859
+ },
+ {
+ "epoch": 11.950218340611354,
+ "grad_norm": 0.020313527435064316,
+ "learning_rate": 0.0006,
+ "loss": 4.939233779907227,
+ "step": 860
+ },
+ {
+ "epoch": 11.96419213973799,
+ "grad_norm": 0.021635599434375763,
+ "learning_rate": 0.0006,
+ "loss": 4.861807346343994,
+ "step": 861
+ },
+ {
+ "epoch": 11.97816593886463,
+ "grad_norm": 0.02285311557352543,
+ "learning_rate": 0.0006,
+ "loss": 4.932732582092285,
+ "step": 862
+ },
+ {
+ "epoch": 11.992139737991266,
+ "grad_norm": 0.02072535641491413,
+ "learning_rate": 0.0006,
+ "loss": 4.796767711639404,
+ "step": 863
+ },
+ {
+ "epoch": 12.0,
+ "grad_norm": 0.020588304847478867,
+ "learning_rate": 0.0006,
+ "loss": 4.997537612915039,
+ "step": 864
+ },
+ {
+ "epoch": 12.0,
+ "eval_loss": 5.044651985168457,
+ "eval_runtime": 56.9145,
+ "eval_samples_per_second": 42.906,
+ "eval_steps_per_second": 1.353,
+ "step": 864
+ },
+ {
+ "epoch": 12.013973799126637,
+ "grad_norm": 0.017597166821360588,
+ "learning_rate": 0.0006,
+ "loss": 4.830524444580078,
+ "step": 865
+ },
+ {
+ "epoch": 12.027947598253276,
+ "grad_norm": 0.01716788113117218,
+ "learning_rate": 0.0006,
+ "loss": 4.894325256347656,
+ "step": 866
+ },
+ {
+ "epoch": 12.041921397379912,
+ "grad_norm": 0.017712457105517387,
+ "learning_rate": 0.0006,
+ "loss": 4.868297576904297,
+ "step": 867
+ },
+ {
+ "epoch": 12.055895196506551,
+ "grad_norm": 0.021486390382051468,
+ "learning_rate": 0.0006,
+ "loss": 4.922049522399902,
+ "step": 868
+ },
+ {
+ "epoch": 12.069868995633188,
+ "grad_norm": 0.024560654535889626,
+ "learning_rate": 0.0006,
+ "loss": 4.734241008758545,
+ "step": 869
+ },
+ {
+ "epoch": 12.083842794759825,
+ "grad_norm": 0.025870252400636673,
+ "learning_rate": 0.0006,
+ "loss": 4.868566989898682,
+ "step": 870
+ },
+ {
+ "epoch": 12.097816593886463,
+ "grad_norm": 0.0277389045804739,
+ "learning_rate": 0.0006,
+ "loss": 4.755006790161133,
+ "step": 871
+ },
+ {
+ "epoch": 12.1117903930131,
+ "grad_norm": 0.029100844636559486,
+ "learning_rate": 0.0006,
+ "loss": 4.909667015075684,
+ "step": 872
+ },
+ {
+ "epoch": 12.125764192139737,
+ "grad_norm": 0.031833067536354065,
+ "learning_rate": 0.0006,
+ "loss": 4.858674049377441,
+ "step": 873
+ },
+ {
+ "epoch": 12.139737991266376,
+ "grad_norm": 0.032709237188100815,
+ "learning_rate": 0.0006,
+ "loss": 4.823696136474609,
+ "step": 874
+ },
+ {
+ "epoch": 12.153711790393013,
+ "grad_norm": 0.02995380386710167,
+ "learning_rate": 0.0006,
+ "loss": 4.850235462188721,
+ "step": 875
+ },
+ {
+ "epoch": 12.167685589519651,
+ "grad_norm": 0.034098610281944275,
+ "learning_rate": 0.0006,
+ "loss": 4.8169989585876465,
+ "step": 876
+ },
+ {
+ "epoch": 12.181659388646288,
+ "grad_norm": 0.03377080336213112,
+ "learning_rate": 0.0006,
+ "loss": 4.90738582611084,
+ "step": 877
+ },
+ {
+ "epoch": 12.195633187772925,
+ "grad_norm": 0.029780825600028038,
+ "learning_rate": 0.0006,
+ "loss": 4.811163902282715,
+ "step": 878
+ },
+ {
+ "epoch": 12.209606986899564,
+ "grad_norm": 0.03118024580180645,
+ "learning_rate": 0.0006,
+ "loss": 4.861286640167236,
+ "step": 879
+ },
+ {
+ "epoch": 12.2235807860262,
+ "grad_norm": 0.033052004873752594,
+ "learning_rate": 0.0006,
+ "loss": 4.865021228790283,
+ "step": 880
+ },
+ {
+ "epoch": 12.237554585152838,
+ "grad_norm": 0.029070032760500908,
+ "learning_rate": 0.0006,
+ "loss": 4.947303771972656,
+ "step": 881
+ },
+ {
+ "epoch": 12.251528384279476,
+ "grad_norm": 0.028355715796351433,
+ "learning_rate": 0.0006,
+ "loss": 4.804648399353027,
+ "step": 882
+ },
+ {
+ "epoch": 12.265502183406113,
+ "grad_norm": 0.031257227063179016,
+ "learning_rate": 0.0006,
+ "loss": 4.757718086242676,
+ "step": 883
+ },
+ {
+ "epoch": 12.279475982532752,
+ "grad_norm": 0.031438373029232025,
+ "learning_rate": 0.0006,
+ "loss": 4.851268768310547,
+ "step": 884
+ },
+ {
+ "epoch": 12.293449781659389,
+ "grad_norm": 0.030492138117551804,
+ "learning_rate": 0.0006,
+ "loss": 4.7891035079956055,
+ "step": 885
+ },
+ {
+ "epoch": 12.307423580786025,
+ "grad_norm": 0.026115991175174713,
+ "learning_rate": 0.0006,
+ "loss": 4.86474609375,
+ "step": 886
+ },
+ {
+ "epoch": 12.321397379912664,
+ "grad_norm": 0.024184465408325195,
+ "learning_rate": 0.0006,
+ "loss": 4.860456943511963,
+ "step": 887
+ },
+ {
+ "epoch": 12.335371179039301,
+ "grad_norm": 0.025437982752919197,
+ "learning_rate": 0.0006,
+ "loss": 4.806251049041748,
+ "step": 888
+ },
+ {
+ "epoch": 12.34934497816594,
+ "grad_norm": 0.022390231490135193,
+ "learning_rate": 0.0006,
+ "loss": 4.807086944580078,
+ "step": 889
+ },
+ {
+ "epoch": 12.363318777292577,
+ "grad_norm": 0.020174294710159302,
+ "learning_rate": 0.0006,
+ "loss": 4.78292179107666,
+ "step": 890
+ },
+ {
+ "epoch": 12.377292576419213,
+ "grad_norm": 0.019734324887394905,
+ "learning_rate": 0.0006,
+ "loss": 4.784509658813477,
+ "step": 891
+ },
+ {
+ "epoch": 12.391266375545852,
+ "grad_norm": 0.016733523458242416,
+ "learning_rate": 0.0006,
+ "loss": 4.811755180358887,
+ "step": 892
+ },
+ {
+ "epoch": 12.405240174672489,
+ "grad_norm": 0.01570146717131138,
+ "learning_rate": 0.0006,
+ "loss": 4.720088005065918,
+ "step": 893
+ },
+ {
+ "epoch": 12.419213973799126,
+ "grad_norm": 0.01601393334567547,
+ "learning_rate": 0.0006,
+ "loss": 4.783543109893799,
+ "step": 894
+ },
+ {
+ "epoch": 12.433187772925764,
+ "grad_norm": 0.014035487547516823,
+ "learning_rate": 0.0006,
+ "loss": 4.848340034484863,
+ "step": 895
+ },
+ {
+ "epoch": 12.447161572052401,
+ "grad_norm": 0.015248487703502178,
+ "learning_rate": 0.0006,
+ "loss": 4.773287773132324,
+ "step": 896
+ },
+ {
+ "epoch": 12.46113537117904,
+ "grad_norm": 0.01532268151640892,
+ "learning_rate": 0.0006,
+ "loss": 4.799646854400635,
+ "step": 897
+ },
+ {
+ "epoch": 12.475109170305677,
+ "grad_norm": 0.013982534408569336,
+ "learning_rate": 0.0006,
+ "loss": 4.8090500831604,
+ "step": 898
+ },
+ {
+ "epoch": 12.489082969432314,
+ "grad_norm": 0.014081164263188839,
+ "learning_rate": 0.0006,
+ "loss": 4.751888751983643,
+ "step": 899
+ },
+ {
+ "epoch": 12.503056768558952,
+ "grad_norm": 0.01380133256316185,
+ "learning_rate": 0.0006,
+ "loss": 4.78144645690918,
+ "step": 900
+ },
+ {
+ "epoch": 12.51703056768559,
+ "grad_norm": 0.013555269688367844,
+ "learning_rate": 0.0006,
+ "loss": 4.779257774353027,
+ "step": 901
+ },
+ {
+ "epoch": 12.531004366812226,
+ "grad_norm": 0.013075701892375946,
+ "learning_rate": 0.0006,
+ "loss": 4.820705413818359,
+ "step": 902
+ },
+ {
+ "epoch": 12.544978165938865,
+ "grad_norm": 0.01276316400617361,
+ "learning_rate": 0.0006,
+ "loss": 4.836733818054199,
+ "step": 903
+ },
+ {
+ "epoch": 12.558951965065502,
+ "grad_norm": 0.013033892959356308,
+ "learning_rate": 0.0006,
+ "loss": 4.764178276062012,
+ "step": 904
+ },
+ {
+ "epoch": 12.57292576419214,
+ "grad_norm": 0.013996980153024197,
+ "learning_rate": 0.0006,
+ "loss": 4.767983436584473,
+ "step": 905
+ },
+ {
+ "epoch": 12.586899563318777,
+ "grad_norm": 0.015222184360027313,
+ "learning_rate": 0.0006,
+ "loss": 4.759721755981445,
+ "step": 906
+ },
+ {
+ "epoch": 12.600873362445414,
+ "grad_norm": 0.01623941957950592,
+ "learning_rate": 0.0006,
+ "loss": 4.849811553955078,
+ "step": 907
+ },
+ {
+ "epoch": 12.614847161572053,
+ "grad_norm": 0.015711182728409767,
+ "learning_rate": 0.0006,
+ "loss": 4.7829084396362305,
+ "step": 908
+ },
+ {
+ "epoch": 12.62882096069869,
+ "grad_norm": 0.014137467369437218,
+ "learning_rate": 0.0006,
+ "loss": 4.73321533203125,
+ "step": 909
+ },
+ {
+ "epoch": 12.642794759825328,
+ "grad_norm": 0.012738462537527084,
+ "learning_rate": 0.0006,
+ "loss": 4.870052337646484,
+ "step": 910
+ },
+ {
+ "epoch": 12.656768558951965,
+ "grad_norm": 0.013683130033314228,
+ "learning_rate": 0.0006,
+ "loss": 4.82658052444458,
+ "step": 911
+ },
+ {
+ "epoch": 12.670742358078602,
+ "grad_norm": 0.014409830793738365,
+ "learning_rate": 0.0006,
+ "loss": 4.707494735717773,
+ "step": 912
+ },
+ {
+ "epoch": 12.68471615720524,
+ "grad_norm": 0.013806809671223164,
+ "learning_rate": 0.0006,
+ "loss": 4.778928279876709,
+ "step": 913
+ },
+ {
+ "epoch": 12.698689956331878,
+ "grad_norm": 0.012857912108302116,
+ "learning_rate": 0.0006,
+ "loss": 4.890717506408691,
+ "step": 914
+ },
+ {
+ "epoch": 12.712663755458514,
+ "grad_norm": 0.013828758150339127,
+ "learning_rate": 0.0006,
+ "loss": 4.8357439041137695,
+ "step": 915
+ },
+ {
+ "epoch": 12.726637554585153,
+ "grad_norm": 0.01478351280093193,
+ "learning_rate": 0.0006,
+ "loss": 4.869986534118652,
+ "step": 916
+ },
+ {
+ "epoch": 12.74061135371179,
+ "grad_norm": 0.01326421182602644,
+ "learning_rate": 0.0006,
+ "loss": 4.830646514892578,
+ "step": 917
+ },
+ {
+ "epoch": 12.754585152838429,
+ "grad_norm": 0.013335433788597584,
+ "learning_rate": 0.0006,
+ "loss": 4.7786149978637695,
+ "step": 918
+ },
+ {
+ "epoch": 12.768558951965066,
+ "grad_norm": 0.014261147007346153,
+ "learning_rate": 0.0006,
+ "loss": 4.788693428039551,
+ "step": 919
+ },
+ {
+ "epoch": 12.782532751091702,
+ "grad_norm": 0.014274783432483673,
+ "learning_rate": 0.0006,
+ "loss": 4.77295446395874,
+ "step": 920
+ },
+ {
+ "epoch": 12.796506550218341,
+ "grad_norm": 0.015074674971401691,
+ "learning_rate": 0.0006,
+ "loss": 4.830754280090332,
+ "step": 921
+ },
+ {
+ "epoch": 12.810480349344978,
+ "grad_norm": 0.015599515289068222,
+ "learning_rate": 0.0006,
+ "loss": 4.837759017944336,
+ "step": 922
+ },
+ {
+ "epoch": 12.824454148471617,
+ "grad_norm": 0.014595242217183113,
+ "learning_rate": 0.0006,
+ "loss": 4.882965564727783,
+ "step": 923
+ },
+ {
+ "epoch": 12.838427947598253,
+ "grad_norm": 0.014951467514038086,
+ "learning_rate": 0.0006,
+ "loss": 4.709748268127441,
+ "step": 924
+ },
+ {
+ "epoch": 12.85240174672489,
+ "grad_norm": 0.01539295818656683,
+ "learning_rate": 0.0006,
+ "loss": 4.791160583496094,
+ "step": 925
+ },
+ {
+ "epoch": 12.866375545851529,
+ "grad_norm": 0.015987906605005264,
+ "learning_rate": 0.0006,
+ "loss": 4.661876678466797,
+ "step": 926
+ },
+ {
+ "epoch": 12.880349344978166,
+ "grad_norm": 0.016527341678738594,
+ "learning_rate": 0.0006,
+ "loss": 4.721952438354492,
+ "step": 927
+ },
+ {
+ "epoch": 12.894323144104803,
+ "grad_norm": 0.01637161523103714,
+ "learning_rate": 0.0006,
+ "loss": 4.704446792602539,
+ "step": 928
+ },
+ {
+ "epoch": 12.908296943231441,
+ "grad_norm": 0.015221050009131432,
+ "learning_rate": 0.0006,
+ "loss": 4.771907806396484,
+ "step": 929
+ },
+ {
+ "epoch": 12.922270742358078,
+ "grad_norm": 0.016242166981101036,
+ "learning_rate": 0.0006,
+ "loss": 4.806267738342285,
+ "step": 930
+ },
+ {
+ "epoch": 12.936244541484717,
+ "grad_norm": 0.01628357358276844,
+ "learning_rate": 0.0006,
+ "loss": 4.7252326011657715,
+ "step": 931
+ },
+ {
+ "epoch": 12.950218340611354,
+ "grad_norm": 0.017428133636713028,
+ "learning_rate": 0.0006,
+ "loss": 4.765689849853516,
+ "step": 932
+ },
+ {
+ "epoch": 12.96419213973799,
+ "grad_norm": 0.01870061084628105,
+ "learning_rate": 0.0006,
+ "loss": 4.814488887786865,
+ "step": 933
+ },
+ {
+ "epoch": 12.97816593886463,
+ "grad_norm": 0.019796060398221016,
+ "learning_rate": 0.0006,
+ "loss": 4.739185810089111,
+ "step": 934
+ },
+ {
+ "epoch": 12.992139737991266,
+ "grad_norm": 0.01985820196568966,
+ "learning_rate": 0.0006,
+ "loss": 4.810608863830566,
+ "step": 935
+ },
+ {
+ "epoch": 13.0,
+ "grad_norm": 0.020834477618336678,
+ "learning_rate": 0.0006,
+ "loss": 4.727858066558838,
+ "step": 936
+ },
+ {
+ "epoch": 13.0,
+ "eval_loss": 4.979116916656494,
+ "eval_runtime": 57.2877,
+ "eval_samples_per_second": 42.627,
+ "eval_steps_per_second": 1.344,
+ "step": 936
+ },
+ {
+ "epoch": 13.013973799126637,
+ "grad_norm": 0.02025519125163555,
+ "learning_rate": 0.0006,
+ "loss": 4.628875732421875,
+ "step": 937
+ },
+ {
+ "epoch": 13.027947598253276,
+ "grad_norm": 0.019596368074417114,
+ "learning_rate": 0.0006,
+ "loss": 4.694699287414551,
+ "step": 938
+ },
+ {
+ "epoch": 13.041921397379912,
+ "grad_norm": 0.020494934171438217,
+ "learning_rate": 0.0006,
+ "loss": 4.691821098327637,
+ "step": 939
+ },
+ {
+ "epoch": 13.055895196506551,
+ "grad_norm": 0.020121760666370392,
+ "learning_rate": 0.0006,
+ "loss": 4.71478271484375,
+ "step": 940
+ },
+ {
+ "epoch": 13.069868995633188,
+ "grad_norm": 0.021091405302286148,
+ "learning_rate": 0.0006,
+ "loss": 4.695915222167969,
+ "step": 941
+ },
+ {
+ "epoch": 13.083842794759825,
+ "grad_norm": 0.019787423312664032,
+ "learning_rate": 0.0006,
+ "loss": 4.8652448654174805,
+ "step": 942
+ },
+ {
+ "epoch": 13.097816593886463,
+ "grad_norm": 0.0185833927243948,
+ "learning_rate": 0.0006,
+ "loss": 4.7065887451171875,
+ "step": 943
+ },
+ {
+ "epoch": 13.1117903930131,
+ "grad_norm": 0.021220475435256958,
+ "learning_rate": 0.0006,
+ "loss": 4.741232395172119,
+ "step": 944
+ },
+ {
+ "epoch": 13.125764192139737,
+ "grad_norm": 0.02136622555553913,
+ "learning_rate": 0.0006,
+ "loss": 4.705967903137207,
+ "step": 945
+ },
+ {
+ "epoch": 13.139737991266376,
+ "grad_norm": 0.0191587433218956,
+ "learning_rate": 0.0006,
+ "loss": 4.758667945861816,
+ "step": 946
+ },
+ {
+ "epoch": 13.153711790393013,
+ "grad_norm": 0.018290022388100624,
+ "learning_rate": 0.0006,
+ "loss": 4.760666847229004,
+ "step": 947
+ },
+ {
+ "epoch": 13.167685589519651,
+ "grad_norm": 0.021704409271478653,
+ "learning_rate": 0.0006,
+ "loss": 4.750555992126465,
+ "step": 948
+ },
+ {
+ "epoch": 13.181659388646288,
+ "grad_norm": 0.021668918430805206,
+ "learning_rate": 0.0006,
+ "loss": 4.840903282165527,
+ "step": 949
+ },
+ {
+ "epoch": 13.195633187772925,
+ "grad_norm": 0.019497141242027283,
+ "learning_rate": 0.0006,
+ "loss": 4.738969326019287,
+ "step": 950
+ },
+ {
+ "epoch": 13.209606986899564,
+ "grad_norm": 0.018500596284866333,
+ "learning_rate": 0.0006,
+ "loss": 4.75154972076416,
+ "step": 951
+ },
+ {
+ "epoch": 13.2235807860262,
+ "grad_norm": 0.01872239261865616,
+ "learning_rate": 0.0006,
+ "loss": 4.811738014221191,
+ "step": 952
+ },
+ {
+ "epoch": 13.237554585152838,
+ "grad_norm": 0.020321547985076904,
+ "learning_rate": 0.0006,
+ "loss": 4.815384387969971,
+ "step": 953
+ },
+ {
+ "epoch": 13.251528384279476,
+ "grad_norm": 0.02283575013279915,
+ "learning_rate": 0.0006,
+ "loss": 4.74582052230835,
+ "step": 954
+ },
+ {
+ "epoch": 13.265502183406113,
+ "grad_norm": 0.02474398910999298,
+ "learning_rate": 0.0006,
+ "loss": 4.738375186920166,
+ "step": 955
+ },
+ {
+ "epoch": 13.279475982532752,
+ "grad_norm": 0.02065764181315899,
+ "learning_rate": 0.0006,
+ "loss": 4.746875286102295,
+ "step": 956
+ },
+ {
+ "epoch": 13.293449781659389,
+ "grad_norm": 0.020534520968794823,
+ "learning_rate": 0.0006,
+ "loss": 4.656914710998535,
+ "step": 957
+ },
+ {
+ "epoch": 13.307423580786025,
+ "grad_norm": 0.0206378772854805,
+ "learning_rate": 0.0006,
+ "loss": 4.842615604400635,
+ "step": 958
+ },
+ {
+ "epoch": 13.321397379912664,
+ "grad_norm": 0.017935654148459435,
+ "learning_rate": 0.0006,
+ "loss": 4.717827796936035,
+ "step": 959
+ },
+ {
+ "epoch": 13.335371179039301,
+ "grad_norm": 0.016958363354206085,
+ "learning_rate": 0.0006,
+ "loss": 4.681070327758789,
+ "step": 960
+ },
+ {
+ "epoch": 13.34934497816594,
+ "grad_norm": 0.015735412016510963,
+ "learning_rate": 0.0006,
+ "loss": 4.7957048416137695,
+ "step": 961
+ },
+ {
+ "epoch": 13.363318777292577,
+ "grad_norm": 0.01563699170947075,
+ "learning_rate": 0.0006,
+ "loss": 4.701192378997803,
+ "step": 962
+ },
+ {
+ "epoch": 13.377292576419213,
+ "grad_norm": 0.015207760035991669,
+ "learning_rate": 0.0006,
+ "loss": 4.73256254196167,
+ "step": 963
+ },
+ {
+ "epoch": 13.391266375545852,
+ "grad_norm": 0.015396883711218834,
+ "learning_rate": 0.0006,
+ "loss": 4.76412296295166,
+ "step": 964
+ },
+ {
+ "epoch": 13.405240174672489,
+ "grad_norm": 0.015690583735704422,
+ "learning_rate": 0.0006,
+ "loss": 4.791157245635986,
+ "step": 965
+ },
+ {
+ "epoch": 13.419213973799126,
+ "grad_norm": 0.01755724661052227,
+ "learning_rate": 0.0006,
+ "loss": 4.845666885375977,
+ "step": 966
+ },
+ {
+ "epoch": 13.433187772925764,
+ "grad_norm": 0.019443422555923462,
+ "learning_rate": 0.0006,
+ "loss": 4.5902299880981445,
+ "step": 967
+ },
+ {
+ "epoch": 13.447161572052401,
+ "grad_norm": 0.02118338644504547,
+ "learning_rate": 0.0006,
+ "loss": 4.740627288818359,
+ "step": 968
+ },
+ {
+ "epoch": 13.46113537117904,
+ "grad_norm": 0.02174839936196804,
+ "learning_rate": 0.0006,
+ "loss": 4.814560890197754,
+ "step": 969
+ },
+ {
+ "epoch": 13.475109170305677,
+ "grad_norm": 0.022860374301671982,
+ "learning_rate": 0.0006,
+ "loss": 4.66440486907959,
+ "step": 970
+ },
+ {
+ "epoch": 13.489082969432314,
+ "grad_norm": 0.024742677807807922,
+ "learning_rate": 0.0006,
+ "loss": 4.7451653480529785,
+ "step": 971
+ },
+ {
+ "epoch": 13.503056768558952,
+ "grad_norm": 0.027068907395005226,
+ "learning_rate": 0.0006,
+ "loss": 4.802584171295166,
+ "step": 972
+ },
+ {
+ "epoch": 13.51703056768559,
+ "grad_norm": 0.026316625997424126,
+ "learning_rate": 0.0006,
+ "loss": 4.842397689819336,
+ "step": 973
+ },
+ {
+ "epoch": 13.531004366812226,
+ "grad_norm": 0.024610962718725204,
+ "learning_rate": 0.0006,
+ "loss": 4.650721073150635,
+ "step": 974
+ },
+ {
+ "epoch": 13.544978165938865,
+ "grad_norm": 0.024122603237628937,
+ "learning_rate": 0.0006,
+ "loss": 4.6984357833862305,
+ "step": 975
+ },
+ {
+ "epoch": 13.558951965065502,
+ "grad_norm": 0.022699708119034767,
+ "learning_rate": 0.0006,
+ "loss": 4.793007850646973,
+ "step": 976
+ },
+ {
+ "epoch": 13.57292576419214,
+ "grad_norm": 0.021706534549593925,
+ "learning_rate": 0.0006,
+ "loss": 4.691362380981445,
+ "step": 977
+ },
+ {
+ "epoch": 13.586899563318777,
+ "grad_norm": 0.021363291889429092,
+ "learning_rate": 0.0006,
+ "loss": 4.76762580871582,
+ "step": 978
+ },
+ {
+ "epoch": 13.600873362445414,
+ "grad_norm": 0.023476677015423775,
+ "learning_rate": 0.0006,
+ "loss": 4.777202606201172,
+ "step": 979
+ },
+ {
+ "epoch": 13.614847161572053,
+ "grad_norm": 0.02457100711762905,
+ "learning_rate": 0.0006,
+ "loss": 4.590703010559082,
+ "step": 980
+ },
+ {
+ "epoch": 13.62882096069869,
+ "grad_norm": 0.02469099499285221,
+ "learning_rate": 0.0006,
+ "loss": 4.640185356140137,
+ "step": 981
+ },
+ {
+ "epoch": 13.642794759825328,
+ "grad_norm": 0.024914603680372238,
+ "learning_rate": 0.0006,
+ "loss": 4.772201061248779,
+ "step": 982
+ },
+ {
+ "epoch": 13.656768558951965,
+ "grad_norm": 0.025133173912763596,
+ "learning_rate": 0.0006,
+ "loss": 4.743337154388428,
+ "step": 983
+ },
+ {
+ "epoch": 13.670742358078602,
+ "grad_norm": 0.02034137211740017,
+ "learning_rate": 0.0006,
+ "loss": 4.752127647399902,
+ "step": 984
+ },
+ {
+ "epoch": 13.68471615720524,
+ "grad_norm": 0.019645661115646362,
+ "learning_rate": 0.0006,
+ "loss": 4.838453769683838,
+ "step": 985
+ },
+ {
+ "epoch": 13.698689956331878,
+ "grad_norm": 0.01900297962129116,
+ "learning_rate": 0.0006,
+ "loss": 4.804868221282959,
+ "step": 986
+ },
+ {
+ "epoch": 13.712663755458514,
+ "grad_norm": 0.020292341709136963,
+ "learning_rate": 0.0006,
+ "loss": 4.609743118286133,
+ "step": 987
+ },
+ {
+ "epoch": 13.726637554585153,
+ "grad_norm": 0.017014775425195694,
+ "learning_rate": 0.0006,
+ "loss": 4.740612983703613,
+ "step": 988
+ },
+ {
+ "epoch": 13.74061135371179,
+ "grad_norm": 0.015199844725430012,
+ "learning_rate": 0.0006,
+ "loss": 4.719701766967773,
+ "step": 989
+ },
+ {
+ "epoch": 13.754585152838429,
+ "grad_norm": 0.015125414356589317,
+ "learning_rate": 0.0006,
+ "loss": 4.686114311218262,
+ "step": 990
+ },
+ {
+ "epoch": 13.768558951965066,
+ "grad_norm": 0.014779540710151196,
+ "learning_rate": 0.0006,
+ "loss": 4.674347877502441,
+ "step": 991
+ },
+ {
+ "epoch": 13.782532751091702,
+ "grad_norm": 0.015255163423717022,
+ "learning_rate": 0.0006,
+ "loss": 4.6309309005737305,
+ "step": 992
+ },
+ {
+ "epoch": 13.796506550218341,
+ "grad_norm": 0.013692197389900684,
+ "learning_rate": 0.0006,
+ "loss": 4.698497295379639,
+ "step": 993
+ },
+ {
+ "epoch": 13.810480349344978,
+ "grad_norm": 0.014529623091220856,
+ "learning_rate": 0.0006,
+ "loss": 4.628443717956543,
+ "step": 994
+ },
+ {
+ "epoch": 13.824454148471617,
+ "grad_norm": 0.01576659083366394,
+ "learning_rate": 0.0006,
+ "loss": 4.788015842437744,
+ "step": 995
+ },
+ {
+ "epoch": 13.838427947598253,
+ "grad_norm": 0.015834277495741844,
+ "learning_rate": 0.0006,
+ "loss": 4.672703742980957,
+ "step": 996
+ },
+ {
+ "epoch": 13.85240174672489,
+ "grad_norm": 0.015539503656327724,
+ "learning_rate": 0.0006,
+ "loss": 4.731503486633301,
+ "step": 997
+ },
+ {
+ "epoch": 13.866375545851529,
+ "grad_norm": 0.013463746756315231,
+ "learning_rate": 0.0006,
+ "loss": 4.7006330490112305,
+ "step": 998
+ },
+ {
+ "epoch": 13.880349344978166,
+ "grad_norm": 0.016377726569771767,
+ "learning_rate": 0.0006,
+ "loss": 4.636074066162109,
+ "step": 999
+ },
+ {
+ "epoch": 13.894323144104803,
+ "grad_norm": 0.016660034656524658,
+ "learning_rate": 0.0006,
+ "loss": 4.742693901062012,
+ "step": 1000
+ },
+ {
+ "epoch": 13.908296943231441,
+ "grad_norm": 0.014902938157320023,
+ "learning_rate": 0.0006,
+ "loss": 4.769811630249023,
+ "step": 1001
+ },
+ {
+ "epoch": 13.922270742358078,
+ "grad_norm": 0.014883476309478283,
+ "learning_rate": 0.0006,
+ "loss": 4.731159210205078,
+ "step": 1002
+ },
+ {
+ "epoch": 13.936244541484717,
+ "grad_norm": 0.014494217932224274,
+ "learning_rate": 0.0006,
+ "loss": 4.659485816955566,
+ "step": 1003
+ },
+ {
+ "epoch": 13.950218340611354,
+ "grad_norm": 0.015426991507411003,
+ "learning_rate": 0.0006,
+ "loss": 4.755229949951172,
+ "step": 1004
+ },
+ {
+ "epoch": 13.96419213973799,
+ "grad_norm": 0.016648586839437485,
+ "learning_rate": 0.0006,
+ "loss": 4.74111270904541,
+ "step": 1005
+ },
+ {
+ "epoch": 13.97816593886463,
+ "grad_norm": 0.016980541869997978,
+ "learning_rate": 0.0006,
+ "loss": 4.74072265625,
+ "step": 1006
+ },
+ {
+ "epoch": 13.992139737991266,
+ "grad_norm": 0.018070152029395103,
+ "learning_rate": 0.0006,
+ "loss": 4.732975006103516,
+ "step": 1007
+ },
+ {
+ "epoch": 14.0,
+ "grad_norm": 0.019910665228962898,
+ "learning_rate": 0.0006,
+ "loss": 4.634697914123535,
+ "step": 1008
+ },
+ {
+ "epoch": 14.0,
+ "eval_loss": 4.850640296936035,
+ "eval_runtime": 57.3039,
+ "eval_samples_per_second": 42.615,
+ "eval_steps_per_second": 1.344,
+ "step": 1008
+ },
+ {
+ "epoch": 14.013973799126637,
+ "grad_norm": 0.01980053074657917,
+ "learning_rate": 0.0006,
+ "loss": 4.628744602203369,
+ "step": 1009
+ },
+ {
+ "epoch": 14.027947598253276,
+ "grad_norm": 0.02497623674571514,
+ "learning_rate": 0.0006,
+ "loss": 4.75930118560791,
+ "step": 1010
+ },
+ {
+ "epoch": 14.041921397379912,
+ "grad_norm": 0.02794097363948822,
+ "learning_rate": 0.0006,
+ "loss": 4.691340446472168,
+ "step": 1011
+ },
+ {
+ "epoch": 14.055895196506551,
+ "grad_norm": 0.023725535720586777,
+ "learning_rate": 0.0006,
+ "loss": 4.733664512634277,
+ "step": 1012
+ },
+ {
+ "epoch": 14.069868995633188,
+ "grad_norm": 0.022966474294662476,
+ "learning_rate": 0.0006,
+ "loss": 4.664808750152588,
+ "step": 1013
+ },
+ {
+ "epoch": 14.083842794759825,
+ "grad_norm": 0.02530563250184059,
+ "learning_rate": 0.0006,
+ "loss": 4.8047285079956055,
+ "step": 1014
+ },
+ {
+ "epoch": 14.097816593886463,
+ "grad_norm": 0.02491082064807415,
+ "learning_rate": 0.0006,
+ "loss": 4.667956352233887,
+ "step": 1015
+ },
+ {
+ "epoch": 14.1117903930131,
+ "grad_norm": 0.022110294550657272,
+ "learning_rate": 0.0006,
+ "loss": 4.666140556335449,
+ "step": 1016
+ },
+ {
+ "epoch": 14.125764192139737,
+ "grad_norm": 0.02065849117934704,
+ "learning_rate": 0.0006,
+ "loss": 4.581960678100586,
+ "step": 1017
+ },
+ {
+ "epoch": 14.139737991266376,
+ "grad_norm": 0.01941298507153988,
+ "learning_rate": 0.0006,
+ "loss": 4.614226341247559,
+ "step": 1018
+ },
+ {
+ "epoch": 14.153711790393013,
+ "grad_norm": 0.022441856563091278,
+ "learning_rate": 0.0006,
+ "loss": 4.638352394104004,
+ "step": 1019
+ },
+ {
+ "epoch": 14.167685589519651,
+ "grad_norm": 0.021519694477319717,
+ "learning_rate": 0.0006,
+ "loss": 4.700160980224609,
+ "step": 1020
+ },
+ {
+ "epoch": 14.181659388646288,
+ "grad_norm": 0.018956486135721207,
+ "learning_rate": 0.0006,
+ "loss": 4.649593353271484,
+ "step": 1021
+ },
+ {
+ "epoch": 14.195633187772925,
+ "grad_norm": 0.016681140288710594,
+ "learning_rate": 0.0006,
+ "loss": 4.603952407836914,
+ "step": 1022
+ },
+ {
+ "epoch": 14.209606986899564,
+ "grad_norm": 0.01788809522986412,
+ "learning_rate": 0.0006,
+ "loss": 4.669160842895508,
+ "step": 1023
+ },
+ {
+ "epoch": 14.2235807860262,
+ "grad_norm": 0.016047228127717972,
+ "learning_rate": 0.0006,
+ "loss": 4.6367950439453125,
+ "step": 1024
+ },
+ {
+ "epoch": 14.237554585152838,
+ "grad_norm": 0.01614239625632763,
+ "learning_rate": 0.0006,
+ "loss": 4.664529323577881,
+ "step": 1025
+ },
+ {
+ "epoch": 14.251528384279476,
+ "grad_norm": 0.016362549737095833,
+ "learning_rate": 0.0006,
+ "loss": 4.702373504638672,
+ "step": 1026
+ },
+ {
+ "epoch": 14.265502183406113,
+ "grad_norm": 0.015800409018993378,
+ "learning_rate": 0.0006,
+ "loss": 4.7594428062438965,
+ "step": 1027
+ },
+ {
+ "epoch": 14.279475982532752,
+ "grad_norm": 0.016730139032006264,
+ "learning_rate": 0.0006,
+ "loss": 4.677244186401367,
+ "step": 1028
+ },
+ {
+ "epoch": 14.293449781659389,
+ "grad_norm": 0.01661759801208973,
+ "learning_rate": 0.0006,
+ "loss": 4.633679389953613,
+ "step": 1029
+ },
+ {
+ "epoch": 14.307423580786025,
+ "grad_norm": 0.017673347145318985,
+ "learning_rate": 0.0006,
+ "loss": 4.649107456207275,
+ "step": 1030
+ },
+ {
+ "epoch": 14.321397379912664,
+ "grad_norm": 0.019243352115154266,
+ "learning_rate": 0.0006,
+ "loss": 4.611257553100586,
+ "step": 1031
+ },
+ {
+ "epoch": 14.335371179039301,
+ "grad_norm": 0.019817164167761803,
+ "learning_rate": 0.0006,
+ "loss": 4.588086128234863,
+ "step": 1032
+ },
+ {
+ "epoch": 14.34934497816594,
+ "grad_norm": 0.022292211651802063,
+ "learning_rate": 0.0006,
+ "loss": 4.783066749572754,
+ "step": 1033
+ },
+ {
+ "epoch": 14.363318777292577,
+ "grad_norm": 0.02140096202492714,
+ "learning_rate": 0.0006,
+ "loss": 4.604718208312988,
+ "step": 1034
+ },
+ {
+ "epoch": 14.377292576419213,
+ "grad_norm": 0.0187817495316267,
+ "learning_rate": 0.0006,
+ "loss": 4.721773147583008,
+ "step": 1035
+ },
+ {
+ "epoch": 14.391266375545852,
+ "grad_norm": 0.019879210740327835,
+ "learning_rate": 0.0006,
+ "loss": 4.789286136627197,
+ "step": 1036
+ },
+ {
+ "epoch": 14.405240174672489,
+ "grad_norm": 0.01885785162448883,
+ "learning_rate": 0.0006,
+ "loss": 4.669743537902832,
+ "step": 1037
+ },
+ {
+ "epoch": 14.419213973799126,
+ "grad_norm": 0.018324270844459534,
+ "learning_rate": 0.0006,
+ "loss": 4.65901517868042,
+ "step": 1038
+ },
+ {
+ "epoch": 14.433187772925764,
+ "grad_norm": 0.017664821818470955,
+ "learning_rate": 0.0006,
+ "loss": 4.667839527130127,
+ "step": 1039
+ },
+ {
+ "epoch": 14.447161572052401,
+ "grad_norm": 0.015702905133366585,
+ "learning_rate": 0.0006,
+ "loss": 4.64700984954834,
+ "step": 1040
+ },
+ {
+ "epoch": 14.46113537117904,
+ "grad_norm": 0.016674194484949112,
+ "learning_rate": 0.0006,
+ "loss": 4.694093227386475,
+ "step": 1041
+ },
+ {
+ "epoch": 14.475109170305677,
+ "grad_norm": 0.016869477927684784,
+ "learning_rate": 0.0006,
+ "loss": 4.619544982910156,
+ "step": 1042
+ },
+ {
+ "epoch": 14.489082969432314,
+ "grad_norm": 0.016557540744543076,
+ "learning_rate": 0.0006,
+ "loss": 4.568667411804199,
+ "step": 1043
+ },
+ {
+ "epoch": 14.503056768558952,
+ "grad_norm": 0.014500945806503296,
+ "learning_rate": 0.0006,
+ "loss": 4.561741828918457,
+ "step": 1044
+ },
+ {
+ "epoch": 14.51703056768559,
+ "grad_norm": 0.014228739775717258,
+ "learning_rate": 0.0006,
+ "loss": 4.654277324676514,
+ "step": 1045
+ },
+ {
+ "epoch": 14.531004366812226,
+ "grad_norm": 0.013894238509237766,
+ "learning_rate": 0.0006,
+ "loss": 4.5867719650268555,
+ "step": 1046
+ },
+ {
+ "epoch": 14.544978165938865,
+ "grad_norm": 0.01550068985670805,
+ "learning_rate": 0.0006,
+ "loss": 4.61168098449707,
+ "step": 1047
+ },
+ {
+ "epoch": 14.558951965065502,
+ "grad_norm": 0.015940772369503975,
+ "learning_rate": 0.0006,
+ "loss": 4.691247940063477,
+ "step": 1048
+ },
+ {
+ "epoch": 14.57292576419214,
+ "grad_norm": 0.01714535616338253,
+ "learning_rate": 0.0006,
+ "loss": 4.61183500289917,
+ "step": 1049
+ },
+ {
+ "epoch": 14.586899563318777,
+ "grad_norm": 0.017925038933753967,
+ "learning_rate": 0.0006,
+ "loss": 4.719022750854492,
+ "step": 1050
+ },
+ {
+ "epoch": 14.600873362445414,
+ "grad_norm": 0.018141552805900574,
+ "learning_rate": 0.0006,
+ "loss": 4.760281562805176,
+ "step": 1051
+ },
+ {
+ "epoch": 14.614847161572053,
+ "grad_norm": 0.017131328582763672,
+ "learning_rate": 0.0006,
+ "loss": 4.709680080413818,
+ "step": 1052
+ },
+ {
+ "epoch": 14.62882096069869,
+ "grad_norm": 0.018730048090219498,
+ "learning_rate": 0.0006,
+ "loss": 4.5994720458984375,
+ "step": 1053
+ },
+ {
+ "epoch": 14.642794759825328,
+ "grad_norm": 0.018894298002123833,
+ "learning_rate": 0.0006,
+ "loss": 4.669070720672607,
+ "step": 1054
+ },
+ {
+ "epoch": 14.656768558951965,
+ "grad_norm": 0.017643198370933533,
+ "learning_rate": 0.0006,
+ "loss": 4.641790390014648,
+ "step": 1055
+ },
+ {
+ "epoch": 14.670742358078602,
+ "grad_norm": 0.017647244036197662,
+ "learning_rate": 0.0006,
+ "loss": 4.653899192810059,
+ "step": 1056
+ },
+ {
+ "epoch": 14.68471615720524,
+ "grad_norm": 0.017304198816418648,
+ "learning_rate": 0.0006,
+ "loss": 4.7487473487854,
+ "step": 1057
+ },
+ {
+ "epoch": 14.698689956331878,
+ "grad_norm": 0.01638447679579258,
+ "learning_rate": 0.0006,
+ "loss": 4.71945858001709,
+ "step": 1058
+ },
+ {
+ "epoch": 14.712663755458514,
+ "grad_norm": 0.015892071649432182,
+ "learning_rate": 0.0006,
+ "loss": 4.744903564453125,
+ "step": 1059
+ },
+ {
+ "epoch": 14.726637554585153,
+ "grad_norm": 0.017420614138245583,
+ "learning_rate": 0.0006,
+ "loss": 4.781563758850098,
+ "step": 1060
+ },
+ {
+ "epoch": 14.74061135371179,
+ "grad_norm": 0.016498520970344543,
+ "learning_rate": 0.0006,
+ "loss": 4.622779846191406,
+ "step": 1061
+ },
+ {
+ "epoch": 14.754585152838429,
+ "grad_norm": 0.017482127994298935,
+ "learning_rate": 0.0006,
+ "loss": 4.62825870513916,
+ "step": 1062
+ },
+ {
+ "epoch": 14.768558951965066,
+ "grad_norm": 0.019018379971385002,
+ "learning_rate": 0.0006,
+ "loss": 4.693194389343262,
+ "step": 1063
+ },
+ {
+ "epoch": 14.782532751091702,
+ "grad_norm": 0.01993914321064949,
+ "learning_rate": 0.0006,
+ "loss": 4.742903709411621,
+ "step": 1064
+ },
+ {
+ "epoch": 14.796506550218341,
+ "grad_norm": 0.02301335707306862,
+ "learning_rate": 0.0006,
+ "loss": 4.613251686096191,
+ "step": 1065
+ },
+ {
+ "epoch": 14.810480349344978,
+ "grad_norm": 0.024088222533464432,
+ "learning_rate": 0.0006,
+ "loss": 4.594305992126465,
+ "step": 1066
+ },
+ {
+ "epoch": 14.824454148471617,
+ "grad_norm": 0.02493320405483246,
+ "learning_rate": 0.0006,
+ "loss": 4.700588226318359,
+ "step": 1067
+ },
+ {
+ "epoch": 14.838427947598253,
+ "grad_norm": 0.02501937560737133,
+ "learning_rate": 0.0006,
+ "loss": 4.601742744445801,
+ "step": 1068
+ },
+ {
+ "epoch": 14.85240174672489,
+ "grad_norm": 0.023490751162171364,
+ "learning_rate": 0.0006,
+ "loss": 4.589211463928223,
+ "step": 1069
+ },
+ {
+ "epoch": 14.866375545851529,
+ "grad_norm": 0.019695665687322617,
+ "learning_rate": 0.0006,
+ "loss": 4.635081768035889,
+ "step": 1070
+ },
+ {
+ "epoch": 14.880349344978166,
+ "grad_norm": 0.019276980310678482,
+ "learning_rate": 0.0006,
+ "loss": 4.737002372741699,
+ "step": 1071
+ },
+ {
+ "epoch": 14.894323144104803,
+ "grad_norm": 0.022430511191487312,
+ "learning_rate": 0.0006,
+ "loss": 4.808295726776123,
+ "step": 1072
+ },
+ {
+ "epoch": 14.908296943231441,
+ "grad_norm": 0.021543916314840317,
+ "learning_rate": 0.0006,
+ "loss": 4.559475421905518,
+ "step": 1073
+ },
+ {
+ "epoch": 14.922270742358078,
+ "grad_norm": 0.019885241985321045,
+ "learning_rate": 0.0006,
+ "loss": 4.647948741912842,
+ "step": 1074
+ },
+ {
+ "epoch": 14.936244541484717,
+ "grad_norm": 0.02115175500512123,
+ "learning_rate": 0.0006,
+ "loss": 4.680505752563477,
+ "step": 1075
+ },
+ {
+ "epoch": 14.950218340611354,
+ "grad_norm": 0.02156151458621025,
+ "learning_rate": 0.0006,
+ "loss": 4.673477649688721,
+ "step": 1076
+ },
+ {
+ "epoch": 14.96419213973799,
+ "grad_norm": 0.02184601128101349,
+ "learning_rate": 0.0006,
+ "loss": 4.644402027130127,
+ "step": 1077
+ },
+ {
+ "epoch": 14.97816593886463,
+ "grad_norm": 0.0211983360350132,
+ "learning_rate": 0.0006,
+ "loss": 4.821558475494385,
+ "step": 1078
+ },
+ {
+ "epoch": 14.992139737991266,
+ "grad_norm": 0.019636567682027817,
+ "learning_rate": 0.0006,
+ "loss": 4.578868865966797,
+ "step": 1079
+ },
+ {
+ "epoch": 15.0,
+ "grad_norm": 0.019997350871562958,
+ "learning_rate": 0.0006,
+ "loss": 4.691307544708252,
+ "step": 1080
+ },
+ {
+ "epoch": 15.0,
+ "eval_loss": 4.872478008270264,
+ "eval_runtime": 56.8513,
+ "eval_samples_per_second": 42.954,
+ "eval_steps_per_second": 1.354,
+ "step": 1080
+ },
+ {
+ "epoch": 15.013973799126637,
+ "grad_norm": 0.018830662593245506,
+ "learning_rate": 0.0006,
+ "loss": 4.761779308319092,
+ "step": 1081
+ },
+ {
+ "epoch": 15.027947598253276,
+ "grad_norm": 0.016876481473445892,
+ "learning_rate": 0.0006,
+ "loss": 4.61148738861084,
+ "step": 1082
+ },
+ {
+ "epoch": 15.041921397379912,
+ "grad_norm": 0.015550863929092884,
+ "learning_rate": 0.0006,
+ "loss": 4.676124095916748,
+ "step": 1083
+ },
+ {
+ "epoch": 15.055895196506551,
+ "grad_norm": 0.017178896814584732,
+ "learning_rate": 0.0006,
+ "loss": 4.588935852050781,
+ "step": 1084
+ },
+ {
+ "epoch": 15.069868995633188,
+ "grad_norm": 0.015465447679162025,
+ "learning_rate": 0.0006,
+ "loss": 4.605499267578125,
+ "step": 1085
+ },
+ {
+ "epoch": 15.083842794759825,
+ "grad_norm": 0.01606195978820324,
+ "learning_rate": 0.0006,
+ "loss": 4.723373889923096,
+ "step": 1086
+ },
+ {
+ "epoch": 15.097816593886463,
+ "grad_norm": 0.016521936282515526,
+ "learning_rate": 0.0006,
+ "loss": 4.646608352661133,
+ "step": 1087
+ },
+ {
+ "epoch": 15.1117903930131,
+ "grad_norm": 0.01589205674827099,
+ "learning_rate": 0.0006,
+ "loss": 4.498138427734375,
+ "step": 1088
+ },
+ {
+ "epoch": 15.125764192139737,
+ "grad_norm": 0.01524856686592102,
+ "learning_rate": 0.0006,
+ "loss": 4.594562530517578,
+ "step": 1089
+ },
+ {
+ "epoch": 15.139737991266376,
+ "grad_norm": 0.014714641496539116,
+ "learning_rate": 0.0006,
+ "loss": 4.630744934082031,
+ "step": 1090
+ },
+ {
+ "epoch": 15.153711790393013,
+ "grad_norm": 0.015333331190049648,
+ "learning_rate": 0.0006,
+ "loss": 4.682388782501221,
+ "step": 1091
+ },
+ {
+ "epoch": 15.167685589519651,
+ "grad_norm": 0.01550073828548193,
+ "learning_rate": 0.0006,
+ "loss": 4.692652702331543,
+ "step": 1092
+ },
+ {
+ "epoch": 15.181659388646288,
+ "grad_norm": 0.014783318154513836,
+ "learning_rate": 0.0006,
+ "loss": 4.6223835945129395,
+ "step": 1093
+ },
+ {
+ "epoch": 15.195633187772925,
+ "grad_norm": 0.014372751116752625,
+ "learning_rate": 0.0006,
+ "loss": 4.52687406539917,
+ "step": 1094
+ },
+ {
+ "epoch": 15.209606986899564,
+ "grad_norm": 0.014779130928218365,
+ "learning_rate": 0.0006,
+ "loss": 4.637009620666504,
+ "step": 1095
+ },
+ {
+ "epoch": 15.2235807860262,
+ "grad_norm": 0.015697073191404343,
+ "learning_rate": 0.0006,
+ "loss": 4.682981014251709,
+ "step": 1096
+ },
+ {
+ "epoch": 15.237554585152838,
+ "grad_norm": 0.016523664817214012,
+ "learning_rate": 0.0006,
+ "loss": 4.705722808837891,
+ "step": 1097
+ },
+ {
+ "epoch": 15.251528384279476,
+ "grad_norm": 0.017633996903896332,
+ "learning_rate": 0.0006,
+ "loss": 4.574686050415039,
+ "step": 1098
+ },
+ {
+ "epoch": 15.265502183406113,
+ "grad_norm": 0.019955076277256012,
+ "learning_rate": 0.0006,
+ "loss": 4.634450912475586,
+ "step": 1099
+ },
+ {
+ "epoch": 15.279475982532752,
+ "grad_norm": 0.02132618986070156,
+ "learning_rate": 0.0006,
+ "loss": 4.56110143661499,
+ "step": 1100
+ },
+ {
+ "epoch": 15.293449781659389,
+ "grad_norm": 0.0207593385130167,
+ "learning_rate": 0.0006,
+ "loss": 4.629396438598633,
+ "step": 1101
+ },
+ {
+ "epoch": 15.307423580786025,
+ "grad_norm": 0.017882540822029114,
+ "learning_rate": 0.0006,
+ "loss": 4.730963230133057,
+ "step": 1102
+ },
+ {
+ "epoch": 15.321397379912664,
+ "grad_norm": 0.01917876861989498,
+ "learning_rate": 0.0006,
+ "loss": 4.64585542678833,
+ "step": 1103
+ },
+ {
+ "epoch": 15.335371179039301,
+ "grad_norm": 0.01943974196910858,
+ "learning_rate": 0.0006,
+ "loss": 4.656181335449219,
+ "step": 1104
+ },
+ {
+ "epoch": 15.34934497816594,
+ "grad_norm": 0.019904715940356255,
+ "learning_rate": 0.0006,
+ "loss": 4.619329452514648,
+ "step": 1105
+ },
+ {
+ "epoch": 15.363318777292577,
+ "grad_norm": 0.019174346700310707,
+ "learning_rate": 0.0006,
+ "loss": 4.647892475128174,
+ "step": 1106
+ },
+ {
+ "epoch": 15.377292576419213,
+ "grad_norm": 0.018244758248329163,
+ "learning_rate": 0.0006,
+ "loss": 4.56166934967041,
+ "step": 1107
+ },
+ {
+ "epoch": 15.391266375545852,
+ "grad_norm": 0.018914205953478813,
+ "learning_rate": 0.0006,
+ "loss": 4.648063659667969,
+ "step": 1108
+ },
+ {
+ "epoch": 15.405240174672489,
+ "grad_norm": 0.01961071416735649,
+ "learning_rate": 0.0006,
+ "loss": 4.665525436401367,
+ "step": 1109
+ },
+ {
+ "epoch": 15.419213973799126,
+ "grad_norm": 0.01841311901807785,
+ "learning_rate": 0.0006,
+ "loss": 4.611342430114746,
+ "step": 1110
+ },
+ {
+ "epoch": 15.433187772925764,
+ "grad_norm": 0.01863427273929119,
+ "learning_rate": 0.0006,
+ "loss": 4.664256572723389,
+ "step": 1111
+ },
+ {
+ "epoch": 15.447161572052401,
+ "grad_norm": 0.020506877452135086,
+ "learning_rate": 0.0006,
+ "loss": 4.631237506866455,
+ "step": 1112
+ },
+ {
+ "epoch": 15.46113537117904,
+ "grad_norm": 0.02223833277821541,
+ "learning_rate": 0.0006,
+ "loss": 4.668654441833496,
+ "step": 1113
+ },
+ {
+ "epoch": 15.475109170305677,
+ "grad_norm": 0.023336702957749367,
+ "learning_rate": 0.0006,
+ "loss": 4.562833309173584,
+ "step": 1114
+ },
+ {
+ "epoch": 15.489082969432314,
+ "grad_norm": 0.023052413016557693,
+ "learning_rate": 0.0006,
+ "loss": 4.621532917022705,
+ "step": 1115
+ },
+ {
+ "epoch": 15.503056768558952,
+ "grad_norm": 0.023420924320816994,
+ "learning_rate": 0.0006,
+ "loss": 4.665465354919434,
+ "step": 1116
+ },
+ {
+ "epoch": 15.51703056768559,
+ "grad_norm": 0.024071281775832176,
+ "learning_rate": 0.0006,
+ "loss": 4.595980644226074,
+ "step": 1117
+ },
+ {
+ "epoch": 15.531004366812226,
+ "grad_norm": 0.023406682536005974,
+ "learning_rate": 0.0006,
+ "loss": 4.642635345458984,
+ "step": 1118
+ },
+ {
+ "epoch": 15.544978165938865,
+ "grad_norm": 0.02138647809624672,
+ "learning_rate": 0.0006,
+ "loss": 4.704513072967529,
+ "step": 1119
+ },
+ {
+ "epoch": 15.558951965065502,
+ "grad_norm": 0.0221384447067976,
+ "learning_rate": 0.0006,
+ "loss": 4.567503452301025,
+ "step": 1120
+ },
+ {
+ "epoch": 15.57292576419214,
+ "grad_norm": 0.020879851654171944,
+ "learning_rate": 0.0006,
+ "loss": 4.664057731628418,
+ "step": 1121
+ },
+ {
+ "epoch": 15.586899563318777,
+ "grad_norm": 0.021478967741131783,
+ "learning_rate": 0.0006,
+ "loss": 4.539627552032471,
+ "step": 1122
+ },
+ {
+ "epoch": 15.600873362445414,
+ "grad_norm": 0.022748656570911407,
+ "learning_rate": 0.0006,
+ "loss": 4.5654497146606445,
+ "step": 1123
+ },
+ {
+ "epoch": 15.614847161572053,
+ "grad_norm": 0.021870248019695282,
+ "learning_rate": 0.0006,
+ "loss": 4.56899356842041,
+ "step": 1124
+ },
+ {
+ "epoch": 15.62882096069869,
+ "grad_norm": 0.021711355075240135,
+ "learning_rate": 0.0006,
+ "loss": 4.657808303833008,
+ "step": 1125
+ },
+ {
+ "epoch": 15.642794759825328,
+ "grad_norm": 0.022810909897089005,
+ "learning_rate": 0.0006,
+ "loss": 4.6526689529418945,
+ "step": 1126
+ },
+ {
+ "epoch": 15.656768558951965,
+ "grad_norm": 0.022561343386769295,
+ "learning_rate": 0.0006,
+ "loss": 4.534887790679932,
+ "step": 1127
+ },
+ {
+ "epoch": 15.670742358078602,
+ "grad_norm": 0.019909802824258804,
+ "learning_rate": 0.0006,
+ "loss": 4.600790023803711,
+ "step": 1128
+ },
+ {
+ "epoch": 15.68471615720524,
+ "grad_norm": 0.020092325285077095,
+ "learning_rate": 0.0006,
+ "loss": 4.773830413818359,
+ "step": 1129
+ },
+ {
+ "epoch": 15.698689956331878,
+ "grad_norm": 0.018834445625543594,
+ "learning_rate": 0.0006,
+ "loss": 4.628868103027344,
+ "step": 1130
+ },
+ {
+ "epoch": 15.712663755458514,
+ "grad_norm": 0.017552798613905907,
+ "learning_rate": 0.0006,
+ "loss": 4.647500038146973,
+ "step": 1131
+ },
+ {
+ "epoch": 15.726637554585153,
+ "grad_norm": 0.01751251146197319,
+ "learning_rate": 0.0006,
+ "loss": 4.590545654296875,
+ "step": 1132
+ },
+ {
+ "epoch": 15.74061135371179,
+ "grad_norm": 0.01699099875986576,
+ "learning_rate": 0.0006,
+ "loss": 4.509699821472168,
+ "step": 1133
+ },
+ {
+ "epoch": 15.754585152838429,
+ "grad_norm": 0.015808649361133575,
+ "learning_rate": 0.0006,
+ "loss": 4.606992244720459,
+ "step": 1134
+ },
+ {
+ "epoch": 15.768558951965066,
+ "grad_norm": 0.01655443385243416,
+ "learning_rate": 0.0006,
+ "loss": 4.585219860076904,
+ "step": 1135
+ },
+ {
+ "epoch": 15.782532751091702,
+ "grad_norm": 0.01708163693547249,
+ "learning_rate": 0.0006,
+ "loss": 4.668147087097168,
+ "step": 1136
+ },
+ {
+ "epoch": 15.796506550218341,
+ "grad_norm": 0.017161913216114044,
+ "learning_rate": 0.0006,
+ "loss": 4.7170634269714355,
+ "step": 1137
+ },
+ {
+ "epoch": 15.810480349344978,
+ "grad_norm": 0.017262674868106842,
+ "learning_rate": 0.0006,
+ "loss": 4.572132110595703,
+ "step": 1138
+ },
+ {
+ "epoch": 15.824454148471617,
+ "grad_norm": 0.017713576555252075,
+ "learning_rate": 0.0006,
+ "loss": 4.702538967132568,
+ "step": 1139
+ },
+ {
+ "epoch": 15.838427947598253,
+ "grad_norm": 0.018730733543634415,
+ "learning_rate": 0.0006,
+ "loss": 4.629721641540527,
+ "step": 1140
+ },
+ {
+ "epoch": 15.85240174672489,
+ "grad_norm": 0.017960187047719955,
+ "learning_rate": 0.0006,
+ "loss": 4.661039352416992,
+ "step": 1141
+ },
+ {
+ "epoch": 15.866375545851529,
+ "grad_norm": 0.01750079356133938,
+ "learning_rate": 0.0006,
+ "loss": 4.503710746765137,
+ "step": 1142
+ },
+ {
+ "epoch": 15.880349344978166,
+ "grad_norm": 0.016630422323942184,
+ "learning_rate": 0.0006,
+ "loss": 4.6582136154174805,
+ "step": 1143
+ },
+ {
+ "epoch": 15.894323144104803,
+ "grad_norm": 0.015896832570433617,
+ "learning_rate": 0.0006,
+ "loss": 4.596693992614746,
+ "step": 1144
+ },
+ {
+ "epoch": 15.908296943231441,
+ "grad_norm": 0.01566407084465027,
+ "learning_rate": 0.0006,
+ "loss": 4.6337199211120605,
+ "step": 1145
+ },
+ {
+ "epoch": 15.922270742358078,
+ "grad_norm": 0.016823330894112587,
+ "learning_rate": 0.0006,
+ "loss": 4.686994552612305,
+ "step": 1146
+ },
+ {
+ "epoch": 15.936244541484717,
+ "grad_norm": 0.016208255663514137,
+ "learning_rate": 0.0006,
+ "loss": 4.6205549240112305,
+ "step": 1147
+ },
+ {
+ "epoch": 15.950218340611354,
+ "grad_norm": 0.0171707421541214,
+ "learning_rate": 0.0006,
+ "loss": 4.582968711853027,
+ "step": 1148
+ },
+ {
+ "epoch": 15.96419213973799,
+ "grad_norm": 0.017009446397423744,
+ "learning_rate": 0.0006,
+ "loss": 4.510274887084961,
+ "step": 1149
+ },
+ {
+ "epoch": 15.97816593886463,
+ "grad_norm": 0.01565811224281788,
+ "learning_rate": 0.0006,
+ "loss": 4.55834436416626,
+ "step": 1150
+ },
+ {
+ "epoch": 15.992139737991266,
+ "grad_norm": 0.015276037156581879,
+ "learning_rate": 0.0006,
+ "loss": 4.537870407104492,
+ "step": 1151
+ },
+ {
+ "epoch": 16.0,
+ "grad_norm": 0.017305882647633553,
+ "learning_rate": 0.0006,
+ "loss": 4.611305236816406,
+ "step": 1152
+ },
+ {
+ "epoch": 16.0,
+ "eval_loss": 4.775513172149658,
+ "eval_runtime": 56.7341,
+ "eval_samples_per_second": 43.043,
+ "eval_steps_per_second": 1.357,
+ "step": 1152
+ },
+ {
+ "epoch": 16.01397379912664,
+ "grad_norm": 0.016752436757087708,
+ "learning_rate": 0.0006,
+ "loss": 4.541312217712402,
+ "step": 1153
+ },
+ {
+ "epoch": 16.027947598253274,
+ "grad_norm": 0.01643134094774723,
+ "learning_rate": 0.0006,
+ "loss": 4.585402965545654,
+ "step": 1154
+ },
+ {
+ "epoch": 16.041921397379912,
+ "grad_norm": 0.01888720504939556,
+ "learning_rate": 0.0006,
+ "loss": 4.597825050354004,
+ "step": 1155
+ },
+ {
+ "epoch": 16.05589519650655,
+ "grad_norm": 0.022044632583856583,
+ "learning_rate": 0.0006,
+ "loss": 4.532772064208984,
+ "step": 1156
+ },
+ {
+ "epoch": 16.069868995633186,
+ "grad_norm": 0.022986987605690956,
+ "learning_rate": 0.0006,
+ "loss": 4.551351547241211,
+ "step": 1157
+ },
+ {
+ "epoch": 16.083842794759825,
+ "grad_norm": 0.020116791129112244,
+ "learning_rate": 0.0006,
+ "loss": 4.484149932861328,
+ "step": 1158
+ },
+ {
+ "epoch": 16.097816593886463,
+ "grad_norm": 0.021442247554659843,
+ "learning_rate": 0.0006,
+ "loss": 4.515316486358643,
+ "step": 1159
+ },
+ {
+ "epoch": 16.111790393013102,
+ "grad_norm": 0.020974772050976753,
+ "learning_rate": 0.0006,
+ "loss": 4.51606559753418,
+ "step": 1160
+ },
+ {
+ "epoch": 16.125764192139737,
+ "grad_norm": 0.021274864673614502,
+ "learning_rate": 0.0006,
+ "loss": 4.587329864501953,
+ "step": 1161
+ },
+ {
+ "epoch": 16.139737991266376,
+ "grad_norm": 0.022799784317612648,
+ "learning_rate": 0.0006,
+ "loss": 4.568149089813232,
+ "step": 1162
+ },
+ {
+ "epoch": 16.153711790393015,
+ "grad_norm": 0.021784933283925056,
+ "learning_rate": 0.0006,
+ "loss": 4.515336990356445,
+ "step": 1163
+ },
+ {
+ "epoch": 16.16768558951965,
+ "grad_norm": 0.024261271581053734,
+ "learning_rate": 0.0006,
+ "loss": 4.5127081871032715,
+ "step": 1164
+ },
+ {
+ "epoch": 16.18165938864629,
+ "grad_norm": 0.02248375490307808,
+ "learning_rate": 0.0006,
+ "loss": 4.754251003265381,
+ "step": 1165
+ },
+ {
+ "epoch": 16.195633187772927,
+ "grad_norm": 0.02136492356657982,
+ "learning_rate": 0.0006,
+ "loss": 4.58193826675415,
+ "step": 1166
+ },
+ {
+ "epoch": 16.209606986899562,
+ "grad_norm": 0.020727066323161125,
+ "learning_rate": 0.0006,
+ "loss": 4.564209938049316,
+ "step": 1167
+ },
+ {
+ "epoch": 16.2235807860262,
+ "grad_norm": 0.017789119854569435,
+ "learning_rate": 0.0006,
+ "loss": 4.698680877685547,
+ "step": 1168
+ },
+ {
+ "epoch": 16.23755458515284,
+ "grad_norm": 0.017603034153580666,
+ "learning_rate": 0.0006,
+ "loss": 4.501171588897705,
+ "step": 1169
+ },
+ {
+ "epoch": 16.251528384279474,
+ "grad_norm": 0.017345862463116646,
+ "learning_rate": 0.0006,
+ "loss": 4.593590259552002,
+ "step": 1170
+ },
+ {
+ "epoch": 16.265502183406113,
+ "grad_norm": 0.018047425895929337,
+ "learning_rate": 0.0006,
+ "loss": 4.597467422485352,
+ "step": 1171
+ },
+ {
+ "epoch": 16.27947598253275,
+ "grad_norm": 0.01947222836315632,
+ "learning_rate": 0.0006,
+ "loss": 4.5072221755981445,
+ "step": 1172
+ },
+ {
+ "epoch": 16.29344978165939,
+ "grad_norm": 0.019598500803112984,
+ "learning_rate": 0.0006,
+ "loss": 4.601351737976074,
+ "step": 1173
+ },
+ {
+ "epoch": 16.307423580786025,
+ "grad_norm": 0.020002514123916626,
+ "learning_rate": 0.0006,
+ "loss": 4.580431938171387,
+ "step": 1174
+ },
+ {
+ "epoch": 16.321397379912664,
+ "grad_norm": 0.0216530729085207,
+ "learning_rate": 0.0006,
+ "loss": 4.424681663513184,
+ "step": 1175
+ },
+ {
+ "epoch": 16.335371179039303,
+ "grad_norm": 0.01949235238134861,
+ "learning_rate": 0.0006,
+ "loss": 4.751883029937744,
+ "step": 1176
+ },
+ {
+ "epoch": 16.349344978165938,
+ "grad_norm": 0.017779016867280006,
+ "learning_rate": 0.0006,
+ "loss": 4.541490077972412,
+ "step": 1177
+ },
+ {
+ "epoch": 16.363318777292577,
+ "grad_norm": 0.01909121684730053,
+ "learning_rate": 0.0006,
+ "loss": 4.594141006469727,
+ "step": 1178
+ },
+ {
+ "epoch": 16.377292576419215,
+ "grad_norm": 0.019894849509000778,
+ "learning_rate": 0.0006,
+ "loss": 4.555070400238037,
+ "step": 1179
+ },
+ {
+ "epoch": 16.39126637554585,
+ "grad_norm": 0.01993032731115818,
+ "learning_rate": 0.0006,
+ "loss": 4.521401405334473,
+ "step": 1180
+ },
+ {
+ "epoch": 16.40524017467249,
+ "grad_norm": 0.0210788045078516,
+ "learning_rate": 0.0006,
+ "loss": 4.672842979431152,
+ "step": 1181
+ },
+ {
+ "epoch": 16.419213973799128,
+ "grad_norm": 0.02274598926305771,
+ "learning_rate": 0.0006,
+ "loss": 4.629504203796387,
+ "step": 1182
+ },
+ {
+ "epoch": 16.433187772925763,
+ "grad_norm": 0.020448120310902596,
+ "learning_rate": 0.0006,
+ "loss": 4.548209190368652,
+ "step": 1183
+ },
+ {
+ "epoch": 16.4471615720524,
+ "grad_norm": 0.017565643414855003,
+ "learning_rate": 0.0006,
+ "loss": 4.529054641723633,
+ "step": 1184
+ },
+ {
+ "epoch": 16.46113537117904,
+ "grad_norm": 0.019504237920045853,
+ "learning_rate": 0.0006,
+ "loss": 4.48182487487793,
+ "step": 1185
+ },
+ {
+ "epoch": 16.475109170305675,
+ "grad_norm": 0.016823742538690567,
+ "learning_rate": 0.0006,
+ "loss": 4.608696937561035,
+ "step": 1186
+ },
+ {
+ "epoch": 16.489082969432314,
+ "grad_norm": 0.01602894812822342,
+ "learning_rate": 0.0006,
+ "loss": 4.620062351226807,
+ "step": 1187
+ },
+ {
+ "epoch": 16.503056768558952,
+ "grad_norm": 0.016615642234683037,
+ "learning_rate": 0.0006,
+ "loss": 4.69378662109375,
+ "step": 1188
+ },
+ {
+ "epoch": 16.51703056768559,
+ "grad_norm": 0.018576730042696,
+ "learning_rate": 0.0006,
+ "loss": 4.537041664123535,
+ "step": 1189
+ },
+ {
+ "epoch": 16.531004366812226,
+ "grad_norm": 0.018611151725053787,
+ "learning_rate": 0.0006,
+ "loss": 4.496169090270996,
+ "step": 1190
+ },
+ {
+ "epoch": 16.544978165938865,
+ "grad_norm": 0.018622413277626038,
+ "learning_rate": 0.0006,
+ "loss": 4.557358264923096,
+ "step": 1191
+ },
+ {
+ "epoch": 16.558951965065503,
+ "grad_norm": 0.01823544315993786,
+ "learning_rate": 0.0006,
+ "loss": 4.730792999267578,
+ "step": 1192
+ },
+ {
+ "epoch": 16.57292576419214,
+ "grad_norm": 0.018796846270561218,
+ "learning_rate": 0.0006,
+ "loss": 4.781791687011719,
+ "step": 1193
+ },
+ {
+ "epoch": 16.586899563318777,
+ "grad_norm": 0.01728292554616928,
+ "learning_rate": 0.0006,
+ "loss": 4.515257835388184,
+ "step": 1194
+ },
+ {
+ "epoch": 16.600873362445416,
+ "grad_norm": 0.01699589006602764,
+ "learning_rate": 0.0006,
+ "loss": 4.5996904373168945,
+ "step": 1195
+ },
+ {
+ "epoch": 16.61484716157205,
+ "grad_norm": 0.015905970707535744,
+ "learning_rate": 0.0006,
+ "loss": 4.601301193237305,
+ "step": 1196
+ },
+ {
+ "epoch": 16.62882096069869,
+ "grad_norm": 0.01338571310043335,
+ "learning_rate": 0.0006,
+ "loss": 4.557708740234375,
+ "step": 1197
+ },
+ {
+ "epoch": 16.64279475982533,
+ "grad_norm": 0.015298855490982533,
+ "learning_rate": 0.0006,
+ "loss": 4.542287349700928,
+ "step": 1198
+ },
+ {
+ "epoch": 16.656768558951963,
+ "grad_norm": 0.014753367751836777,
+ "learning_rate": 0.0006,
+ "loss": 4.595787048339844,
+ "step": 1199
+ },
+ {
+ "epoch": 16.670742358078602,
+ "grad_norm": 0.015527483075857162,
+ "learning_rate": 0.0006,
+ "loss": 4.607529163360596,
+ "step": 1200
+ },
+ {
+ "epoch": 16.68471615720524,
+ "grad_norm": 0.015042162500321865,
+ "learning_rate": 0.0006,
+ "loss": 4.686626434326172,
+ "step": 1201
+ },
+ {
+ "epoch": 16.69868995633188,
+ "grad_norm": 0.0159393772482872,
+ "learning_rate": 0.0006,
+ "loss": 4.47656774520874,
+ "step": 1202
+ },
+ {
+ "epoch": 16.712663755458514,
+ "grad_norm": 0.01532787922769785,
+ "learning_rate": 0.0006,
+ "loss": 4.55678653717041,
+ "step": 1203
+ },
+ {
+ "epoch": 16.726637554585153,
+ "grad_norm": 0.016332164406776428,
+ "learning_rate": 0.0006,
+ "loss": 4.51441764831543,
+ "step": 1204
+ },
+ {
+ "epoch": 16.74061135371179,
+ "grad_norm": 0.018901726230978966,
+ "learning_rate": 0.0006,
+ "loss": 4.57227897644043,
+ "step": 1205
+ },
+ {
+ "epoch": 16.754585152838427,
+ "grad_norm": 0.018589433282613754,
+ "learning_rate": 0.0006,
+ "loss": 4.550070762634277,
+ "step": 1206
+ },
+ {
+ "epoch": 16.768558951965066,
+ "grad_norm": 0.01634034886956215,
+ "learning_rate": 0.0006,
+ "loss": 4.586910724639893,
+ "step": 1207
+ },
+ {
+ "epoch": 16.782532751091704,
+ "grad_norm": 0.017599433660507202,
+ "learning_rate": 0.0006,
+ "loss": 4.5891923904418945,
+ "step": 1208
+ },
+ {
+ "epoch": 16.79650655021834,
+ "grad_norm": 0.01878446340560913,
+ "learning_rate": 0.0006,
+ "loss": 4.57227087020874,
+ "step": 1209
+ },
+ {
+ "epoch": 16.810480349344978,
+ "grad_norm": 0.01803469844162464,
+ "learning_rate": 0.0006,
+ "loss": 4.583845138549805,
+ "step": 1210
+ },
+ {
+ "epoch": 16.824454148471617,
+ "grad_norm": 0.018505144864320755,
+ "learning_rate": 0.0006,
+ "loss": 4.561151504516602,
+ "step": 1211
+ },
+ {
+ "epoch": 16.83842794759825,
+ "grad_norm": 0.01920081488788128,
+ "learning_rate": 0.0006,
+ "loss": 4.506834030151367,
+ "step": 1212
+ },
+ {
+ "epoch": 16.85240174672489,
+ "grad_norm": 0.0177712831646204,
+ "learning_rate": 0.0006,
+ "loss": 4.557195663452148,
+ "step": 1213
+ },
+ {
+ "epoch": 16.86637554585153,
+ "grad_norm": 0.019113918766379356,
+ "learning_rate": 0.0006,
+ "loss": 4.651227951049805,
+ "step": 1214
+ },
+ {
+ "epoch": 16.880349344978168,
+ "grad_norm": 0.02012539468705654,
+ "learning_rate": 0.0006,
+ "loss": 4.4952239990234375,
+ "step": 1215
+ },
+ {
+ "epoch": 16.894323144104803,
+ "grad_norm": 0.019993646070361137,
+ "learning_rate": 0.0006,
+ "loss": 4.517433166503906,
+ "step": 1216
+ },
+ {
+ "epoch": 16.90829694323144,
+ "grad_norm": 0.020041441544890404,
+ "learning_rate": 0.0006,
+ "loss": 4.489407539367676,
+ "step": 1217
+ },
+ {
+ "epoch": 16.92227074235808,
+ "grad_norm": 0.0183611661195755,
+ "learning_rate": 0.0006,
+ "loss": 4.5435638427734375,
+ "step": 1218
+ },
+ {
+ "epoch": 16.936244541484715,
+ "grad_norm": 0.019138429313898087,
+ "learning_rate": 0.0006,
+ "loss": 4.545949459075928,
+ "step": 1219
+ },
+ {
+ "epoch": 16.950218340611354,
+ "grad_norm": 0.018554266542196274,
+ "learning_rate": 0.0006,
+ "loss": 4.579620361328125,
+ "step": 1220
+ },
+ {
+ "epoch": 16.964192139737992,
+ "grad_norm": 0.0193545650690794,
+ "learning_rate": 0.0006,
+ "loss": 4.588255405426025,
+ "step": 1221
+ },
+ {
+ "epoch": 16.978165938864628,
+ "grad_norm": 0.019389165565371513,
+ "learning_rate": 0.0006,
+ "loss": 4.520698547363281,
+ "step": 1222
+ },
+ {
+ "epoch": 16.992139737991266,
+ "grad_norm": 0.016601379960775375,
+ "learning_rate": 0.0006,
+ "loss": 4.4809064865112305,
+ "step": 1223
+ },
+ {
+ "epoch": 17.0,
+ "grad_norm": 0.01771303080022335,
+ "learning_rate": 0.0006,
+ "loss": 4.721122741699219,
+ "step": 1224
+ },
+ {
+ "epoch": 17.0,
+ "eval_loss": 4.7903666496276855,
+ "eval_runtime": 56.9826,
+ "eval_samples_per_second": 42.855,
+ "eval_steps_per_second": 1.351,
+ "step": 1224
+ },
+ {
+ "epoch": 17.01397379912664,
+ "grad_norm": 0.017720289528369904,
+ "learning_rate": 0.0006,
+ "loss": 4.523834705352783,
+ "step": 1225
+ },
+ {
+ "epoch": 17.027947598253274,
+ "grad_norm": 0.020137161016464233,
+ "learning_rate": 0.0006,
+ "loss": 4.574061870574951,
+ "step": 1226
+ },
+ {
+ "epoch": 17.041921397379912,
+ "grad_norm": 0.02002314105629921,
+ "learning_rate": 0.0006,
+ "loss": 4.512407302856445,
+ "step": 1227
+ },
+ {
+ "epoch": 17.05589519650655,
+ "grad_norm": 0.019628183916211128,
+ "learning_rate": 0.0006,
+ "loss": 4.502533435821533,
+ "step": 1228
+ },
+ {
+ "epoch": 17.069868995633186,
+ "grad_norm": 0.019808458164334297,
+ "learning_rate": 0.0006,
+ "loss": 4.463156700134277,
+ "step": 1229
+ },
+ {
+ "epoch": 17.083842794759825,
+ "grad_norm": 0.021250786259770393,
+ "learning_rate": 0.0006,
+ "loss": 4.622682094573975,
+ "step": 1230
+ },
+ {
+ "epoch": 17.097816593886463,
+ "grad_norm": 0.022031694650650024,
+ "learning_rate": 0.0006,
+ "loss": 4.451882839202881,
+ "step": 1231
+ },
+ {
+ "epoch": 17.111790393013102,
+ "grad_norm": 0.01898987777531147,
+ "learning_rate": 0.0006,
+ "loss": 4.529331684112549,
+ "step": 1232
+ },
+ {
+ "epoch": 17.125764192139737,
+ "grad_norm": 0.018749834969639778,
+ "learning_rate": 0.0006,
+ "loss": 4.435815334320068,
+ "step": 1233
+ },
+ {
+ "epoch": 17.139737991266376,
+ "grad_norm": 0.01896674931049347,
+ "learning_rate": 0.0006,
+ "loss": 4.484029293060303,
+ "step": 1234
+ },
+ {
+ "epoch": 17.153711790393015,
+ "grad_norm": 0.018286997452378273,
+ "learning_rate": 0.0006,
+ "loss": 4.523510932922363,
+ "step": 1235
+ },
+ {
+ "epoch": 17.16768558951965,
+ "grad_norm": 0.017273059114813805,
+ "learning_rate": 0.0006,
+ "loss": 4.451767444610596,
+ "step": 1236
+ },
+ {
+ "epoch": 17.18165938864629,
+ "grad_norm": 0.016499150544404984,
+ "learning_rate": 0.0006,
+ "loss": 4.5568647384643555,
+ "step": 1237
+ },
+ {
+ "epoch": 17.195633187772927,
+ "grad_norm": 0.015934469178318977,
+ "learning_rate": 0.0006,
+ "loss": 4.4639997482299805,
+ "step": 1238
+ },
+ {
+ "epoch": 17.209606986899562,
+ "grad_norm": 0.01714209094643593,
+ "learning_rate": 0.0006,
+ "loss": 4.547614097595215,
+ "step": 1239
+ },
+ {
+ "epoch": 17.2235807860262,
+ "grad_norm": 0.019167495891451836,
+ "learning_rate": 0.0006,
+ "loss": 4.571097373962402,
+ "step": 1240
+ },
+ {
+ "epoch": 17.23755458515284,
+ "grad_norm": 0.018276996910572052,
+ "learning_rate": 0.0006,
+ "loss": 4.489764213562012,
+ "step": 1241
+ },
+ {
+ "epoch": 17.251528384279474,
+ "grad_norm": 0.021296866238117218,
+ "learning_rate": 0.0006,
+ "loss": 4.526913642883301,
+ "step": 1242
+ },
+ {
+ "epoch": 17.265502183406113,
+ "grad_norm": 0.022606629878282547,
+ "learning_rate": 0.0006,
+ "loss": 4.521624565124512,
+ "step": 1243
+ },
+ {
+ "epoch": 17.27947598253275,
+ "grad_norm": 0.021204711869359016,
+ "learning_rate": 0.0006,
+ "loss": 4.500864028930664,
+ "step": 1244
+ },
+ {
+ "epoch": 17.29344978165939,
+ "grad_norm": 0.01988849975168705,
+ "learning_rate": 0.0006,
+ "loss": 4.45162296295166,
+ "step": 1245
+ },
+ {
+ "epoch": 17.307423580786025,
+ "grad_norm": 0.020268384367227554,
+ "learning_rate": 0.0006,
+ "loss": 4.536198616027832,
+ "step": 1246
+ },
+ {
+ "epoch": 17.321397379912664,
+ "grad_norm": 0.020185677334666252,
+ "learning_rate": 0.0006,
+ "loss": 4.440262794494629,
+ "step": 1247
+ },
+ {
+ "epoch": 17.335371179039303,
+ "grad_norm": 0.020106853917241096,
+ "learning_rate": 0.0006,
+ "loss": 4.475907325744629,
+ "step": 1248
+ },
+ {
+ "epoch": 17.349344978165938,
+ "grad_norm": 0.018883297219872475,
+ "learning_rate": 0.0006,
+ "loss": 4.640661716461182,
+ "step": 1249
+ },
+ {
+ "epoch": 17.363318777292577,
+ "grad_norm": 0.01852620206773281,
+ "learning_rate": 0.0006,
+ "loss": 4.578243255615234,
+ "step": 1250
+ },
+ {
+ "epoch": 17.377292576419215,
+ "grad_norm": 0.01886451430618763,
+ "learning_rate": 0.0006,
+ "loss": 4.546026229858398,
+ "step": 1251
+ },
+ {
+ "epoch": 17.39126637554585,
+ "grad_norm": 0.01777452416718006,
+ "learning_rate": 0.0006,
+ "loss": 4.521542549133301,
+ "step": 1252
+ },
+ {
+ "epoch": 17.40524017467249,
+ "grad_norm": 0.016626447439193726,
+ "learning_rate": 0.0006,
+ "loss": 4.52629280090332,
+ "step": 1253
+ },
+ {
+ "epoch": 17.419213973799128,
+ "grad_norm": 0.01599477231502533,
+ "learning_rate": 0.0006,
+ "loss": 4.4796037673950195,
+ "step": 1254
+ },
+ {
+ "epoch": 17.433187772925763,
+ "grad_norm": 0.017021458595991135,
+ "learning_rate": 0.0006,
+ "loss": 4.468813419342041,
+ "step": 1255
+ },
+ {
+ "epoch": 17.4471615720524,
+ "grad_norm": 0.016925333067774773,
+ "learning_rate": 0.0006,
+ "loss": 4.584898948669434,
+ "step": 1256
+ },
+ {
+ "epoch": 17.46113537117904,
+ "grad_norm": 0.01714998669922352,
+ "learning_rate": 0.0006,
+ "loss": 4.523544788360596,
+ "step": 1257
+ },
+ {
+ "epoch": 17.475109170305675,
+ "grad_norm": 0.01802288182079792,
+ "learning_rate": 0.0006,
+ "loss": 4.595581531524658,
+ "step": 1258
+ },
+ {
+ "epoch": 17.489082969432314,
+ "grad_norm": 0.0180289875715971,
+ "learning_rate": 0.0006,
+ "loss": 4.394440650939941,
+ "step": 1259
+ },
+ {
+ "epoch": 17.503056768558952,
+ "grad_norm": 0.018504245206713676,
+ "learning_rate": 0.0006,
+ "loss": 4.4794416427612305,
+ "step": 1260
+ },
+ {
+ "epoch": 17.51703056768559,
+ "grad_norm": 0.019973233342170715,
+ "learning_rate": 0.0006,
+ "loss": 4.4241437911987305,
+ "step": 1261
+ },
+ {
+ "epoch": 17.531004366812226,
+ "grad_norm": 0.020063083618879318,
+ "learning_rate": 0.0006,
+ "loss": 4.48068380355835,
+ "step": 1262
+ },
+ {
+ "epoch": 17.544978165938865,
+ "grad_norm": 0.02066640742123127,
+ "learning_rate": 0.0006,
+ "loss": 4.628993034362793,
+ "step": 1263
+ },
+ {
+ "epoch": 17.558951965065503,
+ "grad_norm": 0.02206416428089142,
+ "learning_rate": 0.0006,
+ "loss": 4.564790725708008,
+ "step": 1264
+ },
+ {
+ "epoch": 17.57292576419214,
+ "grad_norm": 0.020640335977077484,
+ "learning_rate": 0.0006,
+ "loss": 4.578705787658691,
+ "step": 1265
+ },
+ {
+ "epoch": 17.586899563318777,
+ "grad_norm": 0.019093405455350876,
+ "learning_rate": 0.0006,
+ "loss": 4.5023603439331055,
+ "step": 1266
+ },
+ {
+ "epoch": 17.600873362445416,
+ "grad_norm": 0.019459193572402,
+ "learning_rate": 0.0006,
+ "loss": 4.594531059265137,
+ "step": 1267
+ },
+ {
+ "epoch": 17.61484716157205,
+ "grad_norm": 0.02021171525120735,
+ "learning_rate": 0.0006,
+ "loss": 4.770299911499023,
+ "step": 1268
+ },
+ {
+ "epoch": 17.62882096069869,
+ "grad_norm": 0.021161191165447235,
+ "learning_rate": 0.0006,
+ "loss": 4.5380401611328125,
+ "step": 1269
+ },
+ {
+ "epoch": 17.64279475982533,
+ "grad_norm": 0.021013403311371803,
+ "learning_rate": 0.0006,
+ "loss": 4.455524921417236,
+ "step": 1270
+ },
+ {
+ "epoch": 17.656768558951963,
+ "grad_norm": 0.020336592569947243,
+ "learning_rate": 0.0006,
+ "loss": 4.533672332763672,
+ "step": 1271
+ },
+ {
+ "epoch": 17.670742358078602,
+ "grad_norm": 0.020159434527158737,
+ "learning_rate": 0.0006,
+ "loss": 4.596151351928711,
+ "step": 1272
+ },
+ {
+ "epoch": 17.68471615720524,
+ "grad_norm": 0.020432960242033005,
+ "learning_rate": 0.0006,
+ "loss": 4.55531644821167,
+ "step": 1273
+ },
+ {
+ "epoch": 17.69868995633188,
+ "grad_norm": 0.025466229766607285,
+ "learning_rate": 0.0006,
+ "loss": 4.566912651062012,
+ "step": 1274
+ },
+ {
+ "epoch": 17.712663755458514,
+ "grad_norm": 0.026537558063864708,
+ "learning_rate": 0.0006,
+ "loss": 4.526003837585449,
+ "step": 1275
+ },
+ {
+ "epoch": 17.726637554585153,
+ "grad_norm": 0.024397607892751694,
+ "learning_rate": 0.0006,
+ "loss": 4.541755676269531,
+ "step": 1276
+ },
+ {
+ "epoch": 17.74061135371179,
+ "grad_norm": 0.022062312811613083,
+ "learning_rate": 0.0006,
+ "loss": 4.6532793045043945,
+ "step": 1277
+ },
+ {
+ "epoch": 17.754585152838427,
+ "grad_norm": 0.02204282023012638,
+ "learning_rate": 0.0006,
+ "loss": 4.582825660705566,
+ "step": 1278
+ },
+ {
+ "epoch": 17.768558951965066,
+ "grad_norm": 0.020844416692852974,
+ "learning_rate": 0.0006,
+ "loss": 4.522336483001709,
+ "step": 1279
+ },
+ {
+ "epoch": 17.782532751091704,
+ "grad_norm": 0.019815703853964806,
+ "learning_rate": 0.0006,
+ "loss": 4.607398509979248,
+ "step": 1280
+ },
+ {
+ "epoch": 17.79650655021834,
+ "grad_norm": 0.016676954925060272,
+ "learning_rate": 0.0006,
+ "loss": 4.446206092834473,
+ "step": 1281
+ },
+ {
+ "epoch": 17.810480349344978,
+ "grad_norm": 0.0166452769190073,
+ "learning_rate": 0.0006,
+ "loss": 4.505931854248047,
+ "step": 1282
+ },
+ {
+ "epoch": 17.824454148471617,
+ "grad_norm": 0.015928907319903374,
+ "learning_rate": 0.0006,
+ "loss": 4.455267429351807,
+ "step": 1283
+ },
+ {
+ "epoch": 17.83842794759825,
+ "grad_norm": 0.015219560824334621,
+ "learning_rate": 0.0006,
+ "loss": 4.5623674392700195,
+ "step": 1284
+ },
+ {
+ "epoch": 17.85240174672489,
+ "grad_norm": 0.015125355683267117,
+ "learning_rate": 0.0006,
+ "loss": 4.554104804992676,
+ "step": 1285
+ },
+ {
+ "epoch": 17.86637554585153,
+ "grad_norm": 0.015612718649208546,
+ "learning_rate": 0.0006,
+ "loss": 4.484847068786621,
+ "step": 1286
+ },
+ {
+ "epoch": 17.880349344978168,
+ "grad_norm": 0.015215662308037281,
+ "learning_rate": 0.0006,
+ "loss": 4.4802093505859375,
+ "step": 1287
+ },
+ {
+ "epoch": 17.894323144104803,
+ "grad_norm": 0.01511684525758028,
+ "learning_rate": 0.0006,
+ "loss": 4.537298679351807,
+ "step": 1288
+ },
+ {
+ "epoch": 17.90829694323144,
+ "grad_norm": 0.013841322623193264,
+ "learning_rate": 0.0006,
+ "loss": 4.427488327026367,
+ "step": 1289
+ },
+ {
+ "epoch": 17.92227074235808,
+ "grad_norm": 0.015115964226424694,
+ "learning_rate": 0.0006,
+ "loss": 4.4547119140625,
+ "step": 1290
+ },
+ {
+ "epoch": 17.936244541484715,
+ "grad_norm": 0.016411006450653076,
+ "learning_rate": 0.0006,
+ "loss": 4.52994966506958,
+ "step": 1291
+ },
+ {
+ "epoch": 17.950218340611354,
+ "grad_norm": 0.017313152551651,
+ "learning_rate": 0.0006,
+ "loss": 4.5337653160095215,
+ "step": 1292
+ },
+ {
+ "epoch": 17.964192139737992,
+ "grad_norm": 0.016559796407818794,
+ "learning_rate": 0.0006,
+ "loss": 4.5277910232543945,
+ "step": 1293
+ },
+ {
+ "epoch": 17.978165938864628,
+ "grad_norm": 0.015854761004447937,
+ "learning_rate": 0.0006,
+ "loss": 4.6334309577941895,
+ "step": 1294
+ },
+ {
+ "epoch": 17.992139737991266,
+ "grad_norm": 0.015658238902688026,
+ "learning_rate": 0.0006,
+ "loss": 4.613372802734375,
+ "step": 1295
+ },
+ {
+ "epoch": 18.0,
+ "grad_norm": 0.0169155802577734,
+ "learning_rate": 0.0006,
+ "loss": 4.527356147766113,
+ "step": 1296
+ },
+ {
+ "epoch": 18.0,
+ "eval_loss": 4.764543056488037,
+ "eval_runtime": 56.5998,
+ "eval_samples_per_second": 43.145,
+ "eval_steps_per_second": 1.36,
+ "step": 1296
+ },
+ {
+ "epoch": 18.01397379912664,
+ "grad_norm": 0.015348963439464569,
+ "learning_rate": 0.0006,
+ "loss": 4.466423988342285,
+ "step": 1297
+ },
+ {
+ "epoch": 18.027947598253274,
+ "grad_norm": 0.01768588088452816,
+ "learning_rate": 0.0006,
+ "loss": 4.545933246612549,
+ "step": 1298
+ },
+ {
+ "epoch": 18.041921397379912,
+ "grad_norm": 0.017381226643919945,
+ "learning_rate": 0.0006,
+ "loss": 4.494060039520264,
+ "step": 1299
+ },
+ {
+ "epoch": 18.05589519650655,
+ "grad_norm": 0.019002335146069527,
+ "learning_rate": 0.0006,
+ "loss": 4.395200729370117,
+ "step": 1300
+ },
+ {
+ "epoch": 18.069868995633186,
+ "grad_norm": 0.01827932707965374,
+ "learning_rate": 0.0006,
+ "loss": 4.424848556518555,
+ "step": 1301
+ },
+ {
+ "epoch": 18.083842794759825,
+ "grad_norm": 0.016383031383156776,
+ "learning_rate": 0.0006,
+ "loss": 4.494722366333008,
+ "step": 1302
+ },
+ {
+ "epoch": 18.097816593886463,
+ "grad_norm": 0.016265157610177994,
+ "learning_rate": 0.0006,
+ "loss": 4.416522979736328,
+ "step": 1303
+ },
+ {
+ "epoch": 18.111790393013102,
+ "grad_norm": 0.01702386513352394,
+ "learning_rate": 0.0006,
+ "loss": 4.600103378295898,
+ "step": 1304
+ },
+ {
+ "epoch": 18.125764192139737,
+ "grad_norm": 0.018495498225092888,
+ "learning_rate": 0.0006,
+ "loss": 4.588401794433594,
+ "step": 1305
+ },
+ {
+ "epoch": 18.139737991266376,
+ "grad_norm": 0.017921684309840202,
+ "learning_rate": 0.0006,
+ "loss": 4.478075981140137,
+ "step": 1306
+ },
+ {
+ "epoch": 18.153711790393015,
+ "grad_norm": 0.01801278069615364,
+ "learning_rate": 0.0006,
+ "loss": 4.413186550140381,
+ "step": 1307
+ },
+ {
+ "epoch": 18.16768558951965,
+ "grad_norm": 0.017211275175213814,
+ "learning_rate": 0.0006,
+ "loss": 4.461341857910156,
+ "step": 1308
+ },
+ {
+ "epoch": 18.18165938864629,
+ "grad_norm": 0.017589328810572624,
+ "learning_rate": 0.0006,
+ "loss": 4.546969413757324,
+ "step": 1309
+ },
+ {
+ "epoch": 18.195633187772927,
+ "grad_norm": 0.01797698810696602,
+ "learning_rate": 0.0006,
+ "loss": 4.530797004699707,
+ "step": 1310
+ },
+ {
+ "epoch": 18.209606986899562,
+ "grad_norm": 0.020408619195222855,
+ "learning_rate": 0.0006,
+ "loss": 4.62999153137207,
+ "step": 1311
+ },
+ {
+ "epoch": 18.2235807860262,
+ "grad_norm": 0.01994038000702858,
+ "learning_rate": 0.0006,
+ "loss": 4.586811065673828,
+ "step": 1312
+ },
+ {
+ "epoch": 18.23755458515284,
+ "grad_norm": 0.019381340593099594,
+ "learning_rate": 0.0006,
+ "loss": 4.4769368171691895,
+ "step": 1313
+ },
+ {
+ "epoch": 18.251528384279474,
+ "grad_norm": 0.020484555512666702,
+ "learning_rate": 0.0006,
+ "loss": 4.524302959442139,
+ "step": 1314
+ },
+ {
+ "epoch": 18.265502183406113,
+ "grad_norm": 0.02426416426897049,
+ "learning_rate": 0.0006,
+ "loss": 4.561705589294434,
+ "step": 1315
+ },
+ {
+ "epoch": 18.27947598253275,
+ "grad_norm": 0.024748101830482483,
+ "learning_rate": 0.0006,
+ "loss": 4.4654316902160645,
+ "step": 1316
+ },
+ {
+ "epoch": 18.29344978165939,
+ "grad_norm": 0.020703328773379326,
+ "learning_rate": 0.0006,
+ "loss": 4.53033447265625,
+ "step": 1317
+ },
+ {
+ "epoch": 18.307423580786025,
+ "grad_norm": 0.01765458658337593,
+ "learning_rate": 0.0006,
+ "loss": 4.5154266357421875,
+ "step": 1318
+ },
+ {
+ "epoch": 18.321397379912664,
+ "grad_norm": 0.018646420910954475,
+ "learning_rate": 0.0006,
+ "loss": 4.5224714279174805,
+ "step": 1319
+ },
+ {
+ "epoch": 18.335371179039303,
+ "grad_norm": 0.01907474733889103,
+ "learning_rate": 0.0006,
+ "loss": 4.5055437088012695,
+ "step": 1320
+ },
+ {
+ "epoch": 18.349344978165938,
+ "grad_norm": 0.018179846927523613,
+ "learning_rate": 0.0006,
+ "loss": 4.462161064147949,
+ "step": 1321
+ },
+ {
+ "epoch": 18.363318777292577,
+ "grad_norm": 0.016697878018021584,
+ "learning_rate": 0.0006,
+ "loss": 4.473718643188477,
+ "step": 1322
+ },
+ {
+ "epoch": 18.377292576419215,
+ "grad_norm": 0.017504019662737846,
+ "learning_rate": 0.0006,
+ "loss": 4.474348068237305,
+ "step": 1323
+ },
+ {
+ "epoch": 18.39126637554585,
+ "grad_norm": 0.018889913335442543,
+ "learning_rate": 0.0006,
+ "loss": 4.4531636238098145,
+ "step": 1324
+ },
+ {
+ "epoch": 18.40524017467249,
+ "grad_norm": 0.019167251884937286,
+ "learning_rate": 0.0006,
+ "loss": 4.4278364181518555,
+ "step": 1325
+ },
+ {
+ "epoch": 18.419213973799128,
+ "grad_norm": 0.015895741060376167,
+ "learning_rate": 0.0006,
+ "loss": 4.474592208862305,
+ "step": 1326
+ },
+ {
+ "epoch": 18.433187772925763,
+ "grad_norm": 0.01531882956624031,
+ "learning_rate": 0.0006,
+ "loss": 4.403757572174072,
+ "step": 1327
+ },
+ {
+ "epoch": 18.4471615720524,
+ "grad_norm": 0.01565849967300892,
+ "learning_rate": 0.0006,
+ "loss": 4.507664680480957,
+ "step": 1328
+ },
+ {
+ "epoch": 18.46113537117904,
+ "grad_norm": 0.01602315716445446,
+ "learning_rate": 0.0006,
+ "loss": 4.5648908615112305,
+ "step": 1329
+ },
+ {
+ "epoch": 18.475109170305675,
+ "grad_norm": 0.016467366367578506,
+ "learning_rate": 0.0006,
+ "loss": 4.428762912750244,
+ "step": 1330
+ },
+ {
+ "epoch": 18.489082969432314,
+ "grad_norm": 0.01604127697646618,
+ "learning_rate": 0.0006,
+ "loss": 4.527771472930908,
+ "step": 1331
+ },
+ {
+ "epoch": 18.503056768558952,
+ "grad_norm": 0.018825586885213852,
+ "learning_rate": 0.0006,
+ "loss": 4.485386848449707,
+ "step": 1332
+ },
+ {
+ "epoch": 18.51703056768559,
+ "grad_norm": 0.019258007407188416,
+ "learning_rate": 0.0006,
+ "loss": 4.414972305297852,
+ "step": 1333
+ },
+ {
+ "epoch": 18.531004366812226,
+ "grad_norm": 0.018192021176218987,
+ "learning_rate": 0.0006,
+ "loss": 4.412166595458984,
+ "step": 1334
+ },
+ {
+ "epoch": 18.544978165938865,
+ "grad_norm": 0.016120698302984238,
+ "learning_rate": 0.0006,
+ "loss": 4.484317779541016,
+ "step": 1335
+ },
+ {
+ "epoch": 18.558951965065503,
+ "grad_norm": 0.016496622934937477,
+ "learning_rate": 0.0006,
+ "loss": 4.514986515045166,
+ "step": 1336
+ },
+ {
+ "epoch": 18.57292576419214,
+ "grad_norm": 0.01795116998255253,
+ "learning_rate": 0.0006,
+ "loss": 4.41473388671875,
+ "step": 1337
+ },
+ {
+ "epoch": 18.586899563318777,
+ "grad_norm": 0.019062276929616928,
+ "learning_rate": 0.0006,
+ "loss": 4.601996898651123,
+ "step": 1338
+ },
+ {
+ "epoch": 18.600873362445416,
+ "grad_norm": 0.017174631357192993,
+ "learning_rate": 0.0006,
+ "loss": 4.571218490600586,
+ "step": 1339
+ },
+ {
+ "epoch": 18.61484716157205,
+ "grad_norm": 0.014015606604516506,
+ "learning_rate": 0.0006,
+ "loss": 4.526606559753418,
+ "step": 1340
+ },
+ {
+ "epoch": 18.62882096069869,
+ "grad_norm": 0.015358510427176952,
+ "learning_rate": 0.0006,
+ "loss": 4.604286193847656,
+ "step": 1341
+ },
+ {
+ "epoch": 18.64279475982533,
+ "grad_norm": 0.01619391329586506,
+ "learning_rate": 0.0006,
+ "loss": 4.560883522033691,
+ "step": 1342
+ },
+ {
+ "epoch": 18.656768558951963,
+ "grad_norm": 0.015972964465618134,
+ "learning_rate": 0.0006,
+ "loss": 4.479755401611328,
+ "step": 1343
+ },
+ {
+ "epoch": 18.670742358078602,
+ "grad_norm": 0.016861025243997574,
+ "learning_rate": 0.0006,
+ "loss": 4.474471092224121,
+ "step": 1344
+ },
+ {
+ "epoch": 18.68471615720524,
+ "grad_norm": 0.01822078227996826,
+ "learning_rate": 0.0006,
+ "loss": 4.509489059448242,
+ "step": 1345
+ },
+ {
+ "epoch": 18.69868995633188,
+ "grad_norm": 0.019415007904171944,
+ "learning_rate": 0.0006,
+ "loss": 4.508941173553467,
+ "step": 1346
+ },
+ {
+ "epoch": 18.712663755458514,
+ "grad_norm": 0.019856126978993416,
+ "learning_rate": 0.0006,
+ "loss": 4.446670055389404,
+ "step": 1347
+ },
+ {
+ "epoch": 18.726637554585153,
+ "grad_norm": 0.01842520199716091,
+ "learning_rate": 0.0006,
+ "loss": 4.524422645568848,
+ "step": 1348
+ },
+ {
+ "epoch": 18.74061135371179,
+ "grad_norm": 0.017196480184793472,
+ "learning_rate": 0.0006,
+ "loss": 4.498466968536377,
+ "step": 1349
+ },
+ {
+ "epoch": 18.754585152838427,
+ "grad_norm": 0.01823010854423046,
+ "learning_rate": 0.0006,
+ "loss": 4.534684181213379,
+ "step": 1350
+ },
+ {
+ "epoch": 18.768558951965066,
+ "grad_norm": 0.020398065447807312,
+ "learning_rate": 0.0006,
+ "loss": 4.456494331359863,
+ "step": 1351
+ },
+ {
+ "epoch": 18.782532751091704,
+ "grad_norm": 0.022992262616753578,
+ "learning_rate": 0.0006,
+ "loss": 4.476374626159668,
+ "step": 1352
+ },
+ {
+ "epoch": 18.79650655021834,
+ "grad_norm": 0.02235420234501362,
+ "learning_rate": 0.0006,
+ "loss": 4.587776184082031,
+ "step": 1353
+ },
+ {
+ "epoch": 18.810480349344978,
+ "grad_norm": 0.02093626745045185,
+ "learning_rate": 0.0006,
+ "loss": 4.424172401428223,
+ "step": 1354
+ },
+ {
+ "epoch": 18.824454148471617,
+ "grad_norm": 0.01997302658855915,
+ "learning_rate": 0.0006,
+ "loss": 4.448066711425781,
+ "step": 1355
+ },
+ {
+ "epoch": 18.83842794759825,
+ "grad_norm": 0.018265459686517715,
+ "learning_rate": 0.0006,
+ "loss": 4.436610698699951,
+ "step": 1356
+ },
+ {
+ "epoch": 18.85240174672489,
+ "grad_norm": 0.01673833839595318,
+ "learning_rate": 0.0006,
+ "loss": 4.5502214431762695,
+ "step": 1357
+ },
+ {
+ "epoch": 18.86637554585153,
+ "grad_norm": 0.017908833920955658,
+ "learning_rate": 0.0006,
+ "loss": 4.550530433654785,
+ "step": 1358
+ },
+ {
+ "epoch": 18.880349344978168,
+ "grad_norm": 0.01796654611825943,
+ "learning_rate": 0.0006,
+ "loss": 4.531463146209717,
+ "step": 1359
+ },
+ {
+ "epoch": 18.894323144104803,
+ "grad_norm": 0.01646183803677559,
+ "learning_rate": 0.0006,
+ "loss": 4.498582363128662,
+ "step": 1360
+ },
+ {
+ "epoch": 18.90829694323144,
+ "grad_norm": 0.015430327504873276,
+ "learning_rate": 0.0006,
+ "loss": 4.52571439743042,
+ "step": 1361
+ },
+ {
+ "epoch": 18.92227074235808,
+ "grad_norm": 0.016527952626347542,
+ "learning_rate": 0.0006,
+ "loss": 4.547013759613037,
+ "step": 1362
+ },
+ {
+ "epoch": 18.936244541484715,
+ "grad_norm": 0.016263660043478012,
+ "learning_rate": 0.0006,
+ "loss": 4.468754291534424,
+ "step": 1363
+ },
+ {
+ "epoch": 18.950218340611354,
+ "grad_norm": 0.016347210854291916,
+ "learning_rate": 0.0006,
+ "loss": 4.450741767883301,
+ "step": 1364
+ },
+ {
+ "epoch": 18.964192139737992,
+ "grad_norm": 0.01701398566365242,
+ "learning_rate": 0.0006,
+ "loss": 4.534899711608887,
+ "step": 1365
+ },
+ {
+ "epoch": 18.978165938864628,
+ "grad_norm": 0.017693081870675087,
+ "learning_rate": 0.0006,
+ "loss": 4.460553169250488,
+ "step": 1366
+ },
+ {
+ "epoch": 18.992139737991266,
+ "grad_norm": 0.018746308982372284,
+ "learning_rate": 0.0006,
+ "loss": 4.538792610168457,
+ "step": 1367
+ },
+ {
+ "epoch": 19.0,
+ "grad_norm": 0.020099416375160217,
+ "learning_rate": 0.0006,
+ "loss": 4.449617385864258,
+ "step": 1368
+ },
+ {
+ "epoch": 19.0,
+ "eval_loss": 4.7756571769714355,
+ "eval_runtime": 57.1804,
+ "eval_samples_per_second": 42.707,
+ "eval_steps_per_second": 1.347,
+ "step": 1368
+ },
+ {
+ "epoch": 19.01397379912664,
+ "grad_norm": 0.018726933747529984,
+ "learning_rate": 0.0006,
+ "loss": 4.446059703826904,
+ "step": 1369
+ },
+ {
+ "epoch": 19.027947598253274,
+ "grad_norm": 0.01956087350845337,
+ "learning_rate": 0.0006,
+ "loss": 4.4520039558410645,
+ "step": 1370
+ },
+ {
+ "epoch": 19.041921397379912,
+ "grad_norm": 0.020092811435461044,
+ "learning_rate": 0.0006,
+ "loss": 4.543343544006348,
+ "step": 1371
+ },
+ {
+ "epoch": 19.05589519650655,
+ "grad_norm": 0.01843622885644436,
+ "learning_rate": 0.0006,
+ "loss": 4.535579204559326,
+ "step": 1372
+ },
+ {
+ "epoch": 19.069868995633186,
+ "grad_norm": 0.016989488154649734,
+ "learning_rate": 0.0006,
+ "loss": 4.52316951751709,
+ "step": 1373
+ },
+ {
+ "epoch": 19.083842794759825,
+ "grad_norm": 0.01805422455072403,
+ "learning_rate": 0.0006,
+ "loss": 4.5393266677856445,
+ "step": 1374
+ },
+ {
+ "epoch": 19.097816593886463,
+ "grad_norm": 0.01874629408121109,
+ "learning_rate": 0.0006,
+ "loss": 4.422059059143066,
+ "step": 1375
+ },
+ {
+ "epoch": 19.111790393013102,
+ "grad_norm": 0.01804221048951149,
+ "learning_rate": 0.0006,
+ "loss": 4.50508451461792,
+ "step": 1376
+ },
+ {
+ "epoch": 19.125764192139737,
+ "grad_norm": 0.01785474829375744,
+ "learning_rate": 0.0006,
+ "loss": 4.412599563598633,
+ "step": 1377
+ },
+ {
+ "epoch": 19.139737991266376,
+ "grad_norm": 0.01971791312098503,
+ "learning_rate": 0.0006,
+ "loss": 4.518318176269531,
+ "step": 1378
+ },
+ {
+ "epoch": 19.153711790393015,
+ "grad_norm": 0.02061633951961994,
+ "learning_rate": 0.0006,
+ "loss": 4.332393169403076,
+ "step": 1379
+ },
+ {
+ "epoch": 19.16768558951965,
+ "grad_norm": 0.020563429221510887,
+ "learning_rate": 0.0006,
+ "loss": 4.445863723754883,
+ "step": 1380
+ },
+ {
+ "epoch": 19.18165938864629,
+ "grad_norm": 0.019995713606476784,
+ "learning_rate": 0.0006,
+ "loss": 4.497312068939209,
+ "step": 1381
+ },
+ {
+ "epoch": 19.195633187772927,
+ "grad_norm": 0.019823070615530014,
+ "learning_rate": 0.0006,
+ "loss": 4.475085258483887,
+ "step": 1382
+ },
+ {
+ "epoch": 19.209606986899562,
+ "grad_norm": 0.019039355218410492,
+ "learning_rate": 0.0006,
+ "loss": 4.50432014465332,
+ "step": 1383
+ },
+ {
+ "epoch": 19.2235807860262,
+ "grad_norm": 0.017876973375678062,
+ "learning_rate": 0.0006,
+ "loss": 4.384788513183594,
+ "step": 1384
+ },
+ {
+ "epoch": 19.23755458515284,
+ "grad_norm": 0.018697699531912804,
+ "learning_rate": 0.0006,
+ "loss": 4.410778045654297,
+ "step": 1385
+ },
+ {
+ "epoch": 19.251528384279474,
+ "grad_norm": 0.018826806917786598,
+ "learning_rate": 0.0006,
+ "loss": 4.524034023284912,
+ "step": 1386
+ },
+ {
+ "epoch": 19.265502183406113,
+ "grad_norm": 0.01824226602911949,
+ "learning_rate": 0.0006,
+ "loss": 4.526393890380859,
+ "step": 1387
+ },
+ {
+ "epoch": 19.27947598253275,
+ "grad_norm": 0.01753208599984646,
+ "learning_rate": 0.0006,
+ "loss": 4.493703365325928,
+ "step": 1388
+ },
+ {
+ "epoch": 19.29344978165939,
+ "grad_norm": 0.01742432825267315,
+ "learning_rate": 0.0006,
+ "loss": 4.392831802368164,
+ "step": 1389
+ },
+ {
+ "epoch": 19.307423580786025,
+ "grad_norm": 0.017099356278777122,
+ "learning_rate": 0.0006,
+ "loss": 4.578668594360352,
+ "step": 1390
+ },
+ {
+ "epoch": 19.321397379912664,
+ "grad_norm": 0.017946023494005203,
+ "learning_rate": 0.0006,
+ "loss": 4.456718444824219,
+ "step": 1391
+ },
+ {
+ "epoch": 19.335371179039303,
+ "grad_norm": 0.017705943435430527,
+ "learning_rate": 0.0006,
+ "loss": 4.553682327270508,
+ "step": 1392
+ },
+ {
+ "epoch": 19.349344978165938,
+ "grad_norm": 0.016647523269057274,
+ "learning_rate": 0.0006,
+ "loss": 4.525016784667969,
+ "step": 1393
+ },
+ {
+ "epoch": 19.363318777292577,
+ "grad_norm": 0.015407416969537735,
+ "learning_rate": 0.0006,
+ "loss": 4.500231742858887,
+ "step": 1394
+ },
+ {
+ "epoch": 19.377292576419215,
+ "grad_norm": 0.017264019697904587,
+ "learning_rate": 0.0006,
+ "loss": 4.4615349769592285,
+ "step": 1395
+ },
+ {
+ "epoch": 19.39126637554585,
+ "grad_norm": 0.01871555484831333,
+ "learning_rate": 0.0006,
+ "loss": 4.281026840209961,
+ "step": 1396
+ },
+ {
+ "epoch": 19.40524017467249,
+ "grad_norm": 0.017361650243401527,
+ "learning_rate": 0.0006,
+ "loss": 4.479382514953613,
+ "step": 1397
+ },
+ {
+ "epoch": 19.419213973799128,
+ "grad_norm": 0.019781362265348434,
+ "learning_rate": 0.0006,
+ "loss": 4.402320861816406,
+ "step": 1398
+ },
+ {
+ "epoch": 19.433187772925763,
+ "grad_norm": 0.019043298438191414,
+ "learning_rate": 0.0006,
+ "loss": 4.476983070373535,
+ "step": 1399
+ },
+ {
+ "epoch": 19.4471615720524,
+ "grad_norm": 0.017137985676527023,
+ "learning_rate": 0.0006,
+ "loss": 4.418689250946045,
+ "step": 1400
+ },
+ {
+ "epoch": 19.46113537117904,
+ "grad_norm": 0.016481833532452583,
+ "learning_rate": 0.0006,
+ "loss": 4.579145431518555,
+ "step": 1401
+ },
+ {
+ "epoch": 19.475109170305675,
+ "grad_norm": 0.016507714986801147,
+ "learning_rate": 0.0006,
+ "loss": 4.519158363342285,
+ "step": 1402
+ },
+ {
+ "epoch": 19.489082969432314,
+ "grad_norm": 0.018092872574925423,
+ "learning_rate": 0.0006,
+ "loss": 4.513748645782471,
+ "step": 1403
+ },
+ {
+ "epoch": 19.503056768558952,
+ "grad_norm": 0.018153440207242966,
+ "learning_rate": 0.0006,
+ "loss": 4.645384311676025,
+ "step": 1404
+ },
+ {
+ "epoch": 19.51703056768559,
+ "grad_norm": 0.01657119207084179,
+ "learning_rate": 0.0006,
+ "loss": 4.399757385253906,
+ "step": 1405
+ },
+ {
+ "epoch": 19.531004366812226,
+ "grad_norm": 0.0156280305236578,
+ "learning_rate": 0.0006,
+ "loss": 4.467836380004883,
+ "step": 1406
+ },
+ {
+ "epoch": 19.544978165938865,
+ "grad_norm": 0.016838403418660164,
+ "learning_rate": 0.0006,
+ "loss": 4.529176235198975,
+ "step": 1407
+ },
+ {
+ "epoch": 19.558951965065503,
+ "grad_norm": 0.01743236929178238,
+ "learning_rate": 0.0006,
+ "loss": 4.440032482147217,
+ "step": 1408
+ },
+ {
+ "epoch": 19.57292576419214,
+ "grad_norm": 0.01553610060364008,
+ "learning_rate": 0.0006,
+ "loss": 4.425588607788086,
+ "step": 1409
+ },
+ {
+ "epoch": 19.586899563318777,
+ "grad_norm": 0.017259271815419197,
+ "learning_rate": 0.0006,
+ "loss": 4.405452728271484,
+ "step": 1410
+ },
+ {
+ "epoch": 19.600873362445416,
+ "grad_norm": 0.015568481758236885,
+ "learning_rate": 0.0006,
+ "loss": 4.444279193878174,
+ "step": 1411
+ },
+ {
+ "epoch": 19.61484716157205,
+ "grad_norm": 0.015331248752772808,
+ "learning_rate": 0.0006,
+ "loss": 4.451846122741699,
+ "step": 1412
+ },
+ {
+ "epoch": 19.62882096069869,
+ "grad_norm": 0.015667378902435303,
+ "learning_rate": 0.0006,
+ "loss": 4.533931255340576,
+ "step": 1413
+ },
+ {
+ "epoch": 19.64279475982533,
+ "grad_norm": 0.016177602112293243,
+ "learning_rate": 0.0006,
+ "loss": 4.467313289642334,
+ "step": 1414
+ },
+ {
+ "epoch": 19.656768558951963,
+ "grad_norm": 0.01733129471540451,
+ "learning_rate": 0.0006,
+ "loss": 4.496893882751465,
+ "step": 1415
+ },
+ {
+ "epoch": 19.670742358078602,
+ "grad_norm": 0.016539594158530235,
+ "learning_rate": 0.0006,
+ "loss": 4.401961326599121,
+ "step": 1416
+ },
+ {
+ "epoch": 19.68471615720524,
+ "grad_norm": 0.016058551147580147,
+ "learning_rate": 0.0006,
+ "loss": 4.503375053405762,
+ "step": 1417
+ },
+ {
+ "epoch": 19.69868995633188,
+ "grad_norm": 0.01715777814388275,
+ "learning_rate": 0.0006,
+ "loss": 4.3808369636535645,
+ "step": 1418
+ },
+ {
+ "epoch": 19.712663755458514,
+ "grad_norm": 0.01681119203567505,
+ "learning_rate": 0.0006,
+ "loss": 4.5469560623168945,
+ "step": 1419
+ },
+ {
+ "epoch": 19.726637554585153,
+ "grad_norm": 0.01634550653398037,
+ "learning_rate": 0.0006,
+ "loss": 4.328197956085205,
+ "step": 1420
+ },
+ {
+ "epoch": 19.74061135371179,
+ "grad_norm": 0.01549347210675478,
+ "learning_rate": 0.0006,
+ "loss": 4.534621238708496,
+ "step": 1421
+ },
+ {
+ "epoch": 19.754585152838427,
+ "grad_norm": 0.01654665172100067,
+ "learning_rate": 0.0006,
+ "loss": 4.4724225997924805,
+ "step": 1422
+ },
+ {
+ "epoch": 19.768558951965066,
+ "grad_norm": 0.017119025811553,
+ "learning_rate": 0.0006,
+ "loss": 4.3897247314453125,
+ "step": 1423
+ },
+ {
+ "epoch": 19.782532751091704,
+ "grad_norm": 0.01603223942220211,
+ "learning_rate": 0.0006,
+ "loss": 4.412731170654297,
+ "step": 1424
+ },
+ {
+ "epoch": 19.79650655021834,
+ "grad_norm": 0.016774067655205727,
+ "learning_rate": 0.0006,
+ "loss": 4.54707145690918,
+ "step": 1425
+ },
+ {
+ "epoch": 19.810480349344978,
+ "grad_norm": 0.016453025862574577,
+ "learning_rate": 0.0006,
+ "loss": 4.462501049041748,
+ "step": 1426
+ },
+ {
+ "epoch": 19.824454148471617,
+ "grad_norm": 0.01597629487514496,
+ "learning_rate": 0.0006,
+ "loss": 4.229758262634277,
+ "step": 1427
+ },
+ {
+ "epoch": 19.83842794759825,
+ "grad_norm": 0.017999034374952316,
+ "learning_rate": 0.0006,
+ "loss": 4.389263153076172,
+ "step": 1428
+ },
+ {
+ "epoch": 19.85240174672489,
+ "grad_norm": 0.02165898121893406,
+ "learning_rate": 0.0006,
+ "loss": 4.342130661010742,
+ "step": 1429
+ },
+ {
+ "epoch": 19.86637554585153,
+ "grad_norm": 0.02222822792828083,
+ "learning_rate": 0.0006,
+ "loss": 4.492661476135254,
+ "step": 1430
+ },
+ {
+ "epoch": 19.880349344978168,
+ "grad_norm": 0.019378066062927246,
+ "learning_rate": 0.0006,
+ "loss": 4.4614105224609375,
+ "step": 1431
+ },
+ {
+ "epoch": 19.894323144104803,
+ "grad_norm": 0.019908083602786064,
+ "learning_rate": 0.0006,
+ "loss": 4.370929718017578,
+ "step": 1432
+ },
+ {
+ "epoch": 19.90829694323144,
+ "grad_norm": 0.022296471521258354,
+ "learning_rate": 0.0006,
+ "loss": 4.438320159912109,
+ "step": 1433
+ },
+ {
+ "epoch": 19.92227074235808,
+ "grad_norm": 0.021810343489050865,
+ "learning_rate": 0.0006,
+ "loss": 4.462067127227783,
+ "step": 1434
+ },
+ {
+ "epoch": 19.936244541484715,
+ "grad_norm": 0.02129209227859974,
+ "learning_rate": 0.0006,
+ "loss": 4.541261196136475,
+ "step": 1435
+ },
+ {
+ "epoch": 19.950218340611354,
+ "grad_norm": 0.02079629711806774,
+ "learning_rate": 0.0006,
+ "loss": 4.36712646484375,
+ "step": 1436
+ },
+ {
+ "epoch": 19.964192139737992,
+ "grad_norm": 0.022501740604639053,
+ "learning_rate": 0.0006,
+ "loss": 4.5279083251953125,
+ "step": 1437
+ },
+ {
+ "epoch": 19.978165938864628,
+ "grad_norm": 0.019121676683425903,
+ "learning_rate": 0.0006,
+ "loss": 4.477789878845215,
+ "step": 1438
+ },
+ {
+ "epoch": 19.992139737991266,
+ "grad_norm": 0.017429588362574577,
+ "learning_rate": 0.0006,
+ "loss": 4.4023847579956055,
+ "step": 1439
+ },
+ {
+ "epoch": 20.0,
+ "grad_norm": 0.018857665359973907,
+ "learning_rate": 0.0006,
+ "loss": 4.494743347167969,
+ "step": 1440
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 28800,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 400,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": true,
+ "should_log": false,
+ "should_save": false,
+ "should_training_stop": false
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 6.139226722664448e+17,
+ "train_batch_size": 8,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-1440/training_args.bin b/runs/l2r50-i2-fulle-lm/checkpoint-1440/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..cad232831c222cc7f8b938b8bbe42f61e7d89c3a
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-1440/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:f7c40e3bfa2f5ac1450ca1243e3f2a895d44a06ed5995e6aefd97f78d229d8c8
+size 4792
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-1728/chat_template.jinja b/runs/l2r50-i2-fulle-lm/checkpoint-1728/chat_template.jinja
new file mode 100644
index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-1728/chat_template.jinja
@@ -0,0 +1,85 @@
+{%- if tools %}
+ {{- '<|im_start|>system\n' }}
+ {%- if messages[0].role == 'system' %}
+ {{- messages[0].content + '\n\n' }}
+ {%- endif %}
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }}
+ {%- for tool in tools %}
+ {{- "\n" }}
+ {{- tool | tojson }}
+ {%- endfor %}
+ {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }}
+{%- else %}
+ {%- if messages[0].role == 'system' %}
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
+ {%- endif %}
+{%- endif %}
+{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
+{%- for message in messages[::-1] %}
+ {%- set index = (messages|length - 1) - loop.index0 %}
+ {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %}
+ {%- set ns.multi_step_tool = false %}
+ {%- set ns.last_query_index = index %}
+ {%- endif %}
+{%- endfor %}
+{%- for message in messages %}
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
+ {%- elif message.role == "assistant" %}
+ {%- set content = message.content %}
+ {%- set reasoning_content = '' %}
+ {%- if message.reasoning_content is defined and message.reasoning_content is not none %}
+ {%- set reasoning_content = message.reasoning_content %}
+ {%- else %}
+ {%- if '' in message.content %}
+ {%- set content = message.content.split('')[-1].lstrip('\n') %}
+ {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %}
+ {%- endif %}
+ {%- endif %}
+ {%- if loop.index0 > ns.last_query_index %}
+ {%- if loop.last or (not loop.last and reasoning_content) %}
+ {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- if message.tool_calls %}
+ {%- for tool_call in message.tool_calls %}
+ {%- if (loop.first and content) or (not loop.first) %}
+ {{- '\n' }}
+ {%- endif %}
+ {%- if tool_call.function %}
+ {%- set tool_call = tool_call.function %}
+ {%- endif %}
+ {{- '\n{"name": "' }}
+ {{- tool_call.name }}
+ {{- '", "arguments": ' }}
+ {%- if tool_call.arguments is string %}
+ {{- tool_call.arguments }}
+ {%- else %}
+ {{- tool_call.arguments | tojson }}
+ {%- endif %}
+ {{- '}\n' }}
+ {%- endfor %}
+ {%- endif %}
+ {{- '<|im_end|>\n' }}
+ {%- elif message.role == "tool" %}
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
+ {{- '<|im_start|>user' }}
+ {%- endif %}
+ {{- '\n\n' }}
+ {{- message.content }}
+ {{- '\n' }}
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
+ {{- '<|im_end|>\n' }}
+ {%- endif %}
+ {%- endif %}
+{%- endfor %}
+{%- if add_generation_prompt %}
+ {{- '<|im_start|>assistant\n' }}
+ {%- if enable_thinking is defined and enable_thinking is false %}
+ {{- '\n\n\n\n' }}
+ {%- endif %}
+{%- endif %}
\ No newline at end of file
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-1728/config.json b/runs/l2r50-i2-fulle-lm/checkpoint-1728/config.json
new file mode 100644
index 0000000000000000000000000000000000000000..22f250953c6ce8fc91ee4411856dd73041a415e2
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-1728/config.json
@@ -0,0 +1,32 @@
+{
+ "architectures": [
+ "LlamaForCausalLM"
+ ],
+ "attention_bias": false,
+ "attention_dropout": 0.0,
+ "bos_token_id": null,
+ "dtype": "float32",
+ "eos_token_id": 151645,
+ "head_dim": 128,
+ "hidden_act": "silu",
+ "hidden_size": 512,
+ "initializer_range": 0.02,
+ "intermediate_size": 1536,
+ "max_position_embeddings": 2048,
+ "mlp_bias": false,
+ "model_type": "llama",
+ "num_attention_heads": 4,
+ "num_hidden_layers": 20,
+ "num_key_value_heads": 4,
+ "pad_token_id": 151645,
+ "pretraining_tp": 1,
+ "rms_norm_eps": 1e-06,
+ "rope_parameters": {
+ "rope_theta": 10000.0,
+ "rope_type": "default"
+ },
+ "tie_word_embeddings": true,
+ "transformers_version": "5.5.0",
+ "use_cache": false,
+ "vocab_size": 151673
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-1728/generation_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-1728/generation_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-1728/generation_config.json
@@ -0,0 +1,11 @@
+{
+ "_from_model_config": true,
+ "eos_token_id": [
+ 151645
+ ],
+ "output_attentions": false,
+ "output_hidden_states": false,
+ "pad_token_id": 151645,
+ "transformers_version": "5.5.0",
+ "use_cache": true
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-1728/model.safetensors b/runs/l2r50-i2-fulle-lm/checkpoint-1728/model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..5b232e6b5d30199e8cae2325e41456d50419a6ca
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-1728/model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:0f24d233851bba30c3d5184941fd624e9c0099141a75a2ab4bc1f1d4f6ce606e
+size 583360328
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-1728/optimizer.pt b/runs/l2r50-i2-fulle-lm/checkpoint-1728/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..c0378d08c4b5c930a5bf4d85c245e568ad70b9cd
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-1728/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:8f6584b3817173853f997d68041230ab13a91932ef5f6e3e446e4c989d7f672f
+size 1166833530
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-1728/rng_state_0.pth b/runs/l2r50-i2-fulle-lm/checkpoint-1728/rng_state_0.pth
new file mode 100644
index 0000000000000000000000000000000000000000..441c22c7739a4a330ab7b7bc71a1ac30ae88d070
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-1728/rng_state_0.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:9d7091aa49cc705439119e0681a55221108bcd330a7cddbab2f08ea063d2bc57
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-1728/rng_state_1.pth b/runs/l2r50-i2-fulle-lm/checkpoint-1728/rng_state_1.pth
new file mode 100644
index 0000000000000000000000000000000000000000..3db45464d5c2e27b3ea96dc60d48c27b1d6456be
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-1728/rng_state_1.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:3628985308248fe16453703204feca526901cbada0027336a1d40b0a847018d9
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-1728/rng_state_2.pth b/runs/l2r50-i2-fulle-lm/checkpoint-1728/rng_state_2.pth
new file mode 100644
index 0000000000000000000000000000000000000000..723c5a12484ed7058823a51ed4d22b3a8e90093d
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-1728/rng_state_2.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:1b345f134315d19b00d8eca0e14d68627acead75dcd96900a83ea3573c76bc71
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-1728/rng_state_3.pth b/runs/l2r50-i2-fulle-lm/checkpoint-1728/rng_state_3.pth
new file mode 100644
index 0000000000000000000000000000000000000000..c213a79c2b53b3b860b2a41aa1f723a98c3e9d28
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-1728/rng_state_3.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:4fd22b557bc4e909e131aafc4f0279bca66927adbdaae3b726b651f76b7208d6
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-1728/scheduler.pt b/runs/l2r50-i2-fulle-lm/checkpoint-1728/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..baba916c0cec91178026a4cf9d1f610779561fa9
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-1728/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:d6a6171dda32af3a9caef592bd2cfcdef1224bfed4ddda14026d355f2f43e8f1
+size 1064
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-1728/tokenizer.json b/runs/l2r50-i2-fulle-lm/checkpoint-1728/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..53786016f070e11184813f3c83101dc5e93e392e
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-1728/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:c3c63fda1646a555448e8f03662542e04b11d103721319e4d9b267d7da03ceb8
+size 11424463
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-1728/tokenizer_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-1728/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..98e2fb9b2cc54785dc9d871b8e503f8b8f2defe7
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-1728/tokenizer_config.json
@@ -0,0 +1,20 @@
+{
+ "add_prefix_space": false,
+ "backend": "tokenizers",
+ "bos_token": null,
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "<|im_end|>",
+ "errors": "replace",
+ "extra_special_tokens": [
+ "<|l2r_pred|>",
+ "<|r2l_pred|>",
+ "<|next_1_pred|>",
+ "<|next_2_pred|>"
+ ],
+ "is_local": false,
+ "model_max_length": 131072,
+ "pad_token": "<|im_end|>",
+ "split_special_tokens": false,
+ "tokenizer_class": "Qwen2Tokenizer",
+ "unk_token": null
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-1728/trainer_state.json b/runs/l2r50-i2-fulle-lm/checkpoint-1728/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..236c964fbfdd98802b3f174c21121bb546dc11df
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-1728/trainer_state.json
@@ -0,0 +1,12314 @@
+{
+ "best_global_step": null,
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 24.0,
+ "eval_steps": 500,
+ "global_step": 1728,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.013973799126637555,
+ "grad_norm": 0.1739506870508194,
+ "learning_rate": 0.0,
+ "loss": 11.99299430847168,
+ "step": 1
+ },
+ {
+ "epoch": 0.02794759825327511,
+ "grad_norm": 0.17216597497463226,
+ "learning_rate": 5.999999999999999e-06,
+ "loss": 11.988704681396484,
+ "step": 2
+ },
+ {
+ "epoch": 0.04192139737991266,
+ "grad_norm": 0.17459715902805328,
+ "learning_rate": 1.1999999999999999e-05,
+ "loss": 11.945581436157227,
+ "step": 3
+ },
+ {
+ "epoch": 0.05589519650655022,
+ "grad_norm": 0.17577414214611053,
+ "learning_rate": 1.7999999999999997e-05,
+ "loss": 11.859772682189941,
+ "step": 4
+ },
+ {
+ "epoch": 0.06986899563318777,
+ "grad_norm": 0.16163651645183563,
+ "learning_rate": 2.3999999999999997e-05,
+ "loss": 11.744415283203125,
+ "step": 5
+ },
+ {
+ "epoch": 0.08384279475982533,
+ "grad_norm": 0.14409932494163513,
+ "learning_rate": 2.9999999999999997e-05,
+ "loss": 11.618099212646484,
+ "step": 6
+ },
+ {
+ "epoch": 0.09781659388646288,
+ "grad_norm": 0.13079848885536194,
+ "learning_rate": 3.5999999999999994e-05,
+ "loss": 11.505935668945312,
+ "step": 7
+ },
+ {
+ "epoch": 0.11179039301310044,
+ "grad_norm": 0.11964613944292068,
+ "learning_rate": 4.2e-05,
+ "loss": 11.416532516479492,
+ "step": 8
+ },
+ {
+ "epoch": 0.125764192139738,
+ "grad_norm": 0.11345142871141434,
+ "learning_rate": 4.7999999999999994e-05,
+ "loss": 11.33206558227539,
+ "step": 9
+ },
+ {
+ "epoch": 0.13973799126637554,
+ "grad_norm": 0.11033952981233597,
+ "learning_rate": 5.399999999999999e-05,
+ "loss": 11.259052276611328,
+ "step": 10
+ },
+ {
+ "epoch": 0.1537117903930131,
+ "grad_norm": 0.10776489973068237,
+ "learning_rate": 5.9999999999999995e-05,
+ "loss": 11.203384399414062,
+ "step": 11
+ },
+ {
+ "epoch": 0.16768558951965065,
+ "grad_norm": 0.10721946507692337,
+ "learning_rate": 6.599999999999999e-05,
+ "loss": 11.153295516967773,
+ "step": 12
+ },
+ {
+ "epoch": 0.18165938864628822,
+ "grad_norm": 0.10673430562019348,
+ "learning_rate": 7.199999999999999e-05,
+ "loss": 11.110200881958008,
+ "step": 13
+ },
+ {
+ "epoch": 0.19563318777292577,
+ "grad_norm": 0.10630254447460175,
+ "learning_rate": 7.8e-05,
+ "loss": 11.067477226257324,
+ "step": 14
+ },
+ {
+ "epoch": 0.2096069868995633,
+ "grad_norm": 0.10631615668535233,
+ "learning_rate": 8.4e-05,
+ "loss": 11.022941589355469,
+ "step": 15
+ },
+ {
+ "epoch": 0.22358078602620088,
+ "grad_norm": 0.10609443485736847,
+ "learning_rate": 8.999999999999999e-05,
+ "loss": 10.970746040344238,
+ "step": 16
+ },
+ {
+ "epoch": 0.23755458515283842,
+ "grad_norm": 0.10648232698440552,
+ "learning_rate": 9.599999999999999e-05,
+ "loss": 10.91267204284668,
+ "step": 17
+ },
+ {
+ "epoch": 0.251528384279476,
+ "grad_norm": 0.10584589838981628,
+ "learning_rate": 0.000102,
+ "loss": 10.854547500610352,
+ "step": 18
+ },
+ {
+ "epoch": 0.26550218340611353,
+ "grad_norm": 0.10591720044612885,
+ "learning_rate": 0.00010799999999999998,
+ "loss": 10.788917541503906,
+ "step": 19
+ },
+ {
+ "epoch": 0.2794759825327511,
+ "grad_norm": 0.10611672699451447,
+ "learning_rate": 0.00011399999999999999,
+ "loss": 10.717556953430176,
+ "step": 20
+ },
+ {
+ "epoch": 0.2934497816593886,
+ "grad_norm": 0.10455359518527985,
+ "learning_rate": 0.00011999999999999999,
+ "loss": 10.650186538696289,
+ "step": 21
+ },
+ {
+ "epoch": 0.3074235807860262,
+ "grad_norm": 0.1047525629401207,
+ "learning_rate": 0.00012599999999999997,
+ "loss": 10.56808090209961,
+ "step": 22
+ },
+ {
+ "epoch": 0.32139737991266376,
+ "grad_norm": 0.10511605441570282,
+ "learning_rate": 0.00013199999999999998,
+ "loss": 10.484028816223145,
+ "step": 23
+ },
+ {
+ "epoch": 0.3353711790393013,
+ "grad_norm": 0.10467177629470825,
+ "learning_rate": 0.000138,
+ "loss": 10.399947166442871,
+ "step": 24
+ },
+ {
+ "epoch": 0.34934497816593885,
+ "grad_norm": 0.10432136058807373,
+ "learning_rate": 0.00014399999999999998,
+ "loss": 10.312714576721191,
+ "step": 25
+ },
+ {
+ "epoch": 0.36331877729257644,
+ "grad_norm": 0.10527262091636658,
+ "learning_rate": 0.00015,
+ "loss": 10.212261199951172,
+ "step": 26
+ },
+ {
+ "epoch": 0.377292576419214,
+ "grad_norm": 0.103610560297966,
+ "learning_rate": 0.000156,
+ "loss": 10.132034301757812,
+ "step": 27
+ },
+ {
+ "epoch": 0.39126637554585153,
+ "grad_norm": 0.10553059726953506,
+ "learning_rate": 0.000162,
+ "loss": 10.020434379577637,
+ "step": 28
+ },
+ {
+ "epoch": 0.4052401746724891,
+ "grad_norm": 0.1037299782037735,
+ "learning_rate": 0.000168,
+ "loss": 9.934466361999512,
+ "step": 29
+ },
+ {
+ "epoch": 0.4192139737991266,
+ "grad_norm": 0.1042739674448967,
+ "learning_rate": 0.00017399999999999997,
+ "loss": 9.822539329528809,
+ "step": 30
+ },
+ {
+ "epoch": 0.4331877729257642,
+ "grad_norm": 0.10300865024328232,
+ "learning_rate": 0.00017999999999999998,
+ "loss": 9.734994888305664,
+ "step": 31
+ },
+ {
+ "epoch": 0.44716157205240176,
+ "grad_norm": 0.10291654616594315,
+ "learning_rate": 0.000186,
+ "loss": 9.624505996704102,
+ "step": 32
+ },
+ {
+ "epoch": 0.4611353711790393,
+ "grad_norm": 0.10210851579904556,
+ "learning_rate": 0.00019199999999999998,
+ "loss": 9.518205642700195,
+ "step": 33
+ },
+ {
+ "epoch": 0.47510917030567684,
+ "grad_norm": 0.10222529619932175,
+ "learning_rate": 0.000198,
+ "loss": 9.412222862243652,
+ "step": 34
+ },
+ {
+ "epoch": 0.4890829694323144,
+ "grad_norm": 0.09992019087076187,
+ "learning_rate": 0.000204,
+ "loss": 9.326566696166992,
+ "step": 35
+ },
+ {
+ "epoch": 0.503056768558952,
+ "grad_norm": 0.09931084513664246,
+ "learning_rate": 0.00020999999999999998,
+ "loss": 9.210111618041992,
+ "step": 36
+ },
+ {
+ "epoch": 0.5170305676855895,
+ "grad_norm": 0.09928256273269653,
+ "learning_rate": 0.00021599999999999996,
+ "loss": 9.100493431091309,
+ "step": 37
+ },
+ {
+ "epoch": 0.5310043668122271,
+ "grad_norm": 0.09843816608190536,
+ "learning_rate": 0.00022199999999999998,
+ "loss": 8.996543884277344,
+ "step": 38
+ },
+ {
+ "epoch": 0.5449781659388646,
+ "grad_norm": 0.09825374186038971,
+ "learning_rate": 0.00022799999999999999,
+ "loss": 8.888991355895996,
+ "step": 39
+ },
+ {
+ "epoch": 0.5589519650655022,
+ "grad_norm": 0.09602385759353638,
+ "learning_rate": 0.000234,
+ "loss": 8.802907943725586,
+ "step": 40
+ },
+ {
+ "epoch": 0.5729257641921397,
+ "grad_norm": 0.09414064139127731,
+ "learning_rate": 0.00023999999999999998,
+ "loss": 8.715423583984375,
+ "step": 41
+ },
+ {
+ "epoch": 0.5868995633187772,
+ "grad_norm": 0.09273893386125565,
+ "learning_rate": 0.00024599999999999996,
+ "loss": 8.61374282836914,
+ "step": 42
+ },
+ {
+ "epoch": 0.6008733624454149,
+ "grad_norm": 0.09079599380493164,
+ "learning_rate": 0.00025199999999999995,
+ "loss": 8.502904891967773,
+ "step": 43
+ },
+ {
+ "epoch": 0.6148471615720524,
+ "grad_norm": 0.08605825901031494,
+ "learning_rate": 0.000258,
+ "loss": 8.453008651733398,
+ "step": 44
+ },
+ {
+ "epoch": 0.62882096069869,
+ "grad_norm": 0.08338996022939682,
+ "learning_rate": 0.00026399999999999997,
+ "loss": 8.365163803100586,
+ "step": 45
+ },
+ {
+ "epoch": 0.6427947598253275,
+ "grad_norm": 0.08150867372751236,
+ "learning_rate": 0.00027,
+ "loss": 8.285372734069824,
+ "step": 46
+ },
+ {
+ "epoch": 0.6567685589519651,
+ "grad_norm": 0.07794807851314545,
+ "learning_rate": 0.000276,
+ "loss": 8.185811996459961,
+ "step": 47
+ },
+ {
+ "epoch": 0.6707423580786026,
+ "grad_norm": 0.07592720538377762,
+ "learning_rate": 0.00028199999999999997,
+ "loss": 8.098482131958008,
+ "step": 48
+ },
+ {
+ "epoch": 0.6847161572052402,
+ "grad_norm": 0.0709897130727768,
+ "learning_rate": 0.00028799999999999995,
+ "loss": 8.042091369628906,
+ "step": 49
+ },
+ {
+ "epoch": 0.6986899563318777,
+ "grad_norm": 0.06582305580377579,
+ "learning_rate": 0.000294,
+ "loss": 7.972844123840332,
+ "step": 50
+ },
+ {
+ "epoch": 0.7126637554585152,
+ "grad_norm": 0.059452351182699203,
+ "learning_rate": 0.0003,
+ "loss": 7.924579620361328,
+ "step": 51
+ },
+ {
+ "epoch": 0.7266375545851529,
+ "grad_norm": 0.052426718175411224,
+ "learning_rate": 0.00030599999999999996,
+ "loss": 7.876336097717285,
+ "step": 52
+ },
+ {
+ "epoch": 0.7406113537117904,
+ "grad_norm": 0.04768150672316551,
+ "learning_rate": 0.000312,
+ "loss": 7.8102707862854,
+ "step": 53
+ },
+ {
+ "epoch": 0.754585152838428,
+ "grad_norm": 0.03970075398683548,
+ "learning_rate": 0.000318,
+ "loss": 7.791913986206055,
+ "step": 54
+ },
+ {
+ "epoch": 0.7685589519650655,
+ "grad_norm": 0.03559752553701401,
+ "learning_rate": 0.000324,
+ "loss": 7.756960868835449,
+ "step": 55
+ },
+ {
+ "epoch": 0.7825327510917031,
+ "grad_norm": 0.029321353882551193,
+ "learning_rate": 0.00033,
+ "loss": 7.724878787994385,
+ "step": 56
+ },
+ {
+ "epoch": 0.7965065502183406,
+ "grad_norm": 0.023485427722334862,
+ "learning_rate": 0.000336,
+ "loss": 7.72553014755249,
+ "step": 57
+ },
+ {
+ "epoch": 0.8104803493449781,
+ "grad_norm": 0.019319474697113037,
+ "learning_rate": 0.00034199999999999996,
+ "loss": 7.709493160247803,
+ "step": 58
+ },
+ {
+ "epoch": 0.8244541484716157,
+ "grad_norm": 0.016705691814422607,
+ "learning_rate": 0.00034799999999999995,
+ "loss": 7.6873297691345215,
+ "step": 59
+ },
+ {
+ "epoch": 0.8384279475982532,
+ "grad_norm": 0.017667734995484352,
+ "learning_rate": 0.00035399999999999993,
+ "loss": 7.677059173583984,
+ "step": 60
+ },
+ {
+ "epoch": 0.8524017467248908,
+ "grad_norm": 0.01671992614865303,
+ "learning_rate": 0.00035999999999999997,
+ "loss": 7.643512725830078,
+ "step": 61
+ },
+ {
+ "epoch": 0.8663755458515284,
+ "grad_norm": 0.012288172729313374,
+ "learning_rate": 0.00036599999999999995,
+ "loss": 7.658343315124512,
+ "step": 62
+ },
+ {
+ "epoch": 0.880349344978166,
+ "grad_norm": 0.012004776857793331,
+ "learning_rate": 0.000372,
+ "loss": 7.649759292602539,
+ "step": 63
+ },
+ {
+ "epoch": 0.8943231441048035,
+ "grad_norm": 0.012487445957958698,
+ "learning_rate": 0.00037799999999999997,
+ "loss": 7.631503105163574,
+ "step": 64
+ },
+ {
+ "epoch": 0.9082969432314411,
+ "grad_norm": 0.01489241048693657,
+ "learning_rate": 0.00038399999999999996,
+ "loss": 7.647110462188721,
+ "step": 65
+ },
+ {
+ "epoch": 0.9222707423580786,
+ "grad_norm": 0.01428980939090252,
+ "learning_rate": 0.00039,
+ "loss": 7.636983871459961,
+ "step": 66
+ },
+ {
+ "epoch": 0.9362445414847161,
+ "grad_norm": 0.010711144655942917,
+ "learning_rate": 0.000396,
+ "loss": 7.621130466461182,
+ "step": 67
+ },
+ {
+ "epoch": 0.9502183406113537,
+ "grad_norm": 0.010178886353969574,
+ "learning_rate": 0.000402,
+ "loss": 7.593924522399902,
+ "step": 68
+ },
+ {
+ "epoch": 0.9641921397379912,
+ "grad_norm": 0.010604874230921268,
+ "learning_rate": 0.000408,
+ "loss": 7.6015496253967285,
+ "step": 69
+ },
+ {
+ "epoch": 0.9781659388646288,
+ "grad_norm": 0.009751426056027412,
+ "learning_rate": 0.0004139999999999999,
+ "loss": 7.571964263916016,
+ "step": 70
+ },
+ {
+ "epoch": 0.9921397379912664,
+ "grad_norm": 0.009816481731832027,
+ "learning_rate": 0.00041999999999999996,
+ "loss": 7.57344913482666,
+ "step": 71
+ },
+ {
+ "epoch": 1.0,
+ "grad_norm": 0.008665637113153934,
+ "learning_rate": 0.00042599999999999995,
+ "loss": 7.579050064086914,
+ "step": 72
+ },
+ {
+ "epoch": 1.0,
+ "eval_loss": 7.581204891204834,
+ "eval_runtime": 59.0091,
+ "eval_samples_per_second": 41.383,
+ "eval_steps_per_second": 1.305,
+ "step": 72
+ },
+ {
+ "epoch": 1.0139737991266375,
+ "grad_norm": 0.009684201329946518,
+ "learning_rate": 0.00043199999999999993,
+ "loss": 7.568594932556152,
+ "step": 73
+ },
+ {
+ "epoch": 1.027947598253275,
+ "grad_norm": 0.025783710181713104,
+ "learning_rate": 0.00043799999999999997,
+ "loss": 7.578948497772217,
+ "step": 74
+ },
+ {
+ "epoch": 1.0419213973799126,
+ "grad_norm": 0.0327819399535656,
+ "learning_rate": 0.00044399999999999995,
+ "loss": 7.548173904418945,
+ "step": 75
+ },
+ {
+ "epoch": 1.0558951965065502,
+ "grad_norm": 0.021604498848319054,
+ "learning_rate": 0.00045,
+ "loss": 7.552953720092773,
+ "step": 76
+ },
+ {
+ "epoch": 1.0698689956331877,
+ "grad_norm": 0.010230105370283127,
+ "learning_rate": 0.00045599999999999997,
+ "loss": 7.5213212966918945,
+ "step": 77
+ },
+ {
+ "epoch": 1.0838427947598253,
+ "grad_norm": 0.023315589874982834,
+ "learning_rate": 0.00046199999999999995,
+ "loss": 7.501640319824219,
+ "step": 78
+ },
+ {
+ "epoch": 1.0978165938864628,
+ "grad_norm": 0.01109944935888052,
+ "learning_rate": 0.000468,
+ "loss": 7.539519786834717,
+ "step": 79
+ },
+ {
+ "epoch": 1.1117903930131003,
+ "grad_norm": 0.026093758642673492,
+ "learning_rate": 0.000474,
+ "loss": 7.525928974151611,
+ "step": 80
+ },
+ {
+ "epoch": 1.125764192139738,
+ "grad_norm": 0.03274895250797272,
+ "learning_rate": 0.00047999999999999996,
+ "loss": 7.535602569580078,
+ "step": 81
+ },
+ {
+ "epoch": 1.1397379912663754,
+ "grad_norm": 0.01543602254241705,
+ "learning_rate": 0.000486,
+ "loss": 7.513683795928955,
+ "step": 82
+ },
+ {
+ "epoch": 1.1537117903930132,
+ "grad_norm": 0.028377249836921692,
+ "learning_rate": 0.0004919999999999999,
+ "loss": 7.49384880065918,
+ "step": 83
+ },
+ {
+ "epoch": 1.1676855895196507,
+ "grad_norm": 0.04121481627225876,
+ "learning_rate": 0.000498,
+ "loss": 7.488447189331055,
+ "step": 84
+ },
+ {
+ "epoch": 1.1816593886462883,
+ "grad_norm": 0.013282302767038345,
+ "learning_rate": 0.0005039999999999999,
+ "loss": 7.463010787963867,
+ "step": 85
+ },
+ {
+ "epoch": 1.1956331877729258,
+ "grad_norm": 0.03200364485383034,
+ "learning_rate": 0.0005099999999999999,
+ "loss": 7.473515033721924,
+ "step": 86
+ },
+ {
+ "epoch": 1.2096069868995634,
+ "grad_norm": 0.020768029615283012,
+ "learning_rate": 0.000516,
+ "loss": 7.464688777923584,
+ "step": 87
+ },
+ {
+ "epoch": 1.223580786026201,
+ "grad_norm": 0.021715788170695305,
+ "learning_rate": 0.000522,
+ "loss": 7.4835028648376465,
+ "step": 88
+ },
+ {
+ "epoch": 1.2375545851528384,
+ "grad_norm": 0.026255754753947258,
+ "learning_rate": 0.0005279999999999999,
+ "loss": 7.443392753601074,
+ "step": 89
+ },
+ {
+ "epoch": 1.251528384279476,
+ "grad_norm": 0.022325951606035233,
+ "learning_rate": 0.000534,
+ "loss": 7.457373142242432,
+ "step": 90
+ },
+ {
+ "epoch": 1.2655021834061135,
+ "grad_norm": 0.009270387701690197,
+ "learning_rate": 0.00054,
+ "loss": 7.455986976623535,
+ "step": 91
+ },
+ {
+ "epoch": 1.279475982532751,
+ "grad_norm": 0.02369544468820095,
+ "learning_rate": 0.0005459999999999999,
+ "loss": 7.442216873168945,
+ "step": 92
+ },
+ {
+ "epoch": 1.2934497816593886,
+ "grad_norm": 0.008474506437778473,
+ "learning_rate": 0.000552,
+ "loss": 7.4294915199279785,
+ "step": 93
+ },
+ {
+ "epoch": 1.3074235807860262,
+ "grad_norm": 0.017043249681591988,
+ "learning_rate": 0.000558,
+ "loss": 7.406932830810547,
+ "step": 94
+ },
+ {
+ "epoch": 1.3213973799126637,
+ "grad_norm": 0.011813412420451641,
+ "learning_rate": 0.0005639999999999999,
+ "loss": 7.415900230407715,
+ "step": 95
+ },
+ {
+ "epoch": 1.3353711790393012,
+ "grad_norm": 0.01086883433163166,
+ "learning_rate": 0.00057,
+ "loss": 7.452998638153076,
+ "step": 96
+ },
+ {
+ "epoch": 1.3493449781659388,
+ "grad_norm": 0.014700951054692268,
+ "learning_rate": 0.0005759999999999999,
+ "loss": 7.410249710083008,
+ "step": 97
+ },
+ {
+ "epoch": 1.3633187772925766,
+ "grad_norm": 0.014512976631522179,
+ "learning_rate": 0.0005819999999999999,
+ "loss": 7.429132461547852,
+ "step": 98
+ },
+ {
+ "epoch": 1.3772925764192139,
+ "grad_norm": 0.012519920244812965,
+ "learning_rate": 0.000588,
+ "loss": 7.430817604064941,
+ "step": 99
+ },
+ {
+ "epoch": 1.3912663755458516,
+ "grad_norm": 0.007132664322853088,
+ "learning_rate": 0.0005939999999999999,
+ "loss": 7.405234336853027,
+ "step": 100
+ },
+ {
+ "epoch": 1.405240174672489,
+ "grad_norm": 0.019489388912916183,
+ "learning_rate": 0.0006,
+ "loss": 7.427791595458984,
+ "step": 101
+ },
+ {
+ "epoch": 1.4192139737991267,
+ "grad_norm": 0.027252664789557457,
+ "learning_rate": 0.0006,
+ "loss": 7.41761589050293,
+ "step": 102
+ },
+ {
+ "epoch": 1.4331877729257643,
+ "grad_norm": 0.02501637488603592,
+ "learning_rate": 0.0006,
+ "loss": 7.416735649108887,
+ "step": 103
+ },
+ {
+ "epoch": 1.4471615720524018,
+ "grad_norm": 0.00989772193133831,
+ "learning_rate": 0.0006,
+ "loss": 7.3756632804870605,
+ "step": 104
+ },
+ {
+ "epoch": 1.4611353711790394,
+ "grad_norm": 0.01622949168086052,
+ "learning_rate": 0.0006,
+ "loss": 7.400811672210693,
+ "step": 105
+ },
+ {
+ "epoch": 1.475109170305677,
+ "grad_norm": 0.012407763861119747,
+ "learning_rate": 0.0006,
+ "loss": 7.37675666809082,
+ "step": 106
+ },
+ {
+ "epoch": 1.4890829694323144,
+ "grad_norm": 0.010353565216064453,
+ "learning_rate": 0.0006,
+ "loss": 7.4013142585754395,
+ "step": 107
+ },
+ {
+ "epoch": 1.503056768558952,
+ "grad_norm": 0.0120553532615304,
+ "learning_rate": 0.0006,
+ "loss": 7.379873275756836,
+ "step": 108
+ },
+ {
+ "epoch": 1.5170305676855895,
+ "grad_norm": 0.01610225811600685,
+ "learning_rate": 0.0006,
+ "loss": 7.371631145477295,
+ "step": 109
+ },
+ {
+ "epoch": 1.531004366812227,
+ "grad_norm": 0.01159879844635725,
+ "learning_rate": 0.0006,
+ "loss": 7.389007568359375,
+ "step": 110
+ },
+ {
+ "epoch": 1.5449781659388646,
+ "grad_norm": 0.008013816550374031,
+ "learning_rate": 0.0006,
+ "loss": 7.395920753479004,
+ "step": 111
+ },
+ {
+ "epoch": 1.5589519650655022,
+ "grad_norm": 0.016334809362888336,
+ "learning_rate": 0.0006,
+ "loss": 7.393121719360352,
+ "step": 112
+ },
+ {
+ "epoch": 1.5729257641921397,
+ "grad_norm": 0.018821721896529198,
+ "learning_rate": 0.0006,
+ "loss": 7.367543697357178,
+ "step": 113
+ },
+ {
+ "epoch": 1.5868995633187772,
+ "grad_norm": 0.014088994823396206,
+ "learning_rate": 0.0006,
+ "loss": 7.385537624359131,
+ "step": 114
+ },
+ {
+ "epoch": 1.600873362445415,
+ "grad_norm": 0.005864050704985857,
+ "learning_rate": 0.0006,
+ "loss": 7.365346908569336,
+ "step": 115
+ },
+ {
+ "epoch": 1.6148471615720523,
+ "grad_norm": 0.011745430529117584,
+ "learning_rate": 0.0006,
+ "loss": 7.390227317810059,
+ "step": 116
+ },
+ {
+ "epoch": 1.62882096069869,
+ "grad_norm": 0.017205573618412018,
+ "learning_rate": 0.0006,
+ "loss": 7.3558549880981445,
+ "step": 117
+ },
+ {
+ "epoch": 1.6427947598253274,
+ "grad_norm": 0.021256346255540848,
+ "learning_rate": 0.0006,
+ "loss": 7.377346992492676,
+ "step": 118
+ },
+ {
+ "epoch": 1.6567685589519652,
+ "grad_norm": 0.019335204735398293,
+ "learning_rate": 0.0006,
+ "loss": 7.346194267272949,
+ "step": 119
+ },
+ {
+ "epoch": 1.6707423580786025,
+ "grad_norm": 0.008134003728628159,
+ "learning_rate": 0.0006,
+ "loss": 7.3063578605651855,
+ "step": 120
+ },
+ {
+ "epoch": 1.6847161572052403,
+ "grad_norm": 0.017613831907510757,
+ "learning_rate": 0.0006,
+ "loss": 7.338171005249023,
+ "step": 121
+ },
+ {
+ "epoch": 1.6986899563318776,
+ "grad_norm": 0.00890734139829874,
+ "learning_rate": 0.0006,
+ "loss": 7.355384826660156,
+ "step": 122
+ },
+ {
+ "epoch": 1.7126637554585153,
+ "grad_norm": 0.013448155485093594,
+ "learning_rate": 0.0006,
+ "loss": 7.344296455383301,
+ "step": 123
+ },
+ {
+ "epoch": 1.726637554585153,
+ "grad_norm": 0.009241633117198944,
+ "learning_rate": 0.0006,
+ "loss": 7.3454060554504395,
+ "step": 124
+ },
+ {
+ "epoch": 1.7406113537117904,
+ "grad_norm": 0.018561311066150665,
+ "learning_rate": 0.0006,
+ "loss": 7.32348108291626,
+ "step": 125
+ },
+ {
+ "epoch": 1.754585152838428,
+ "grad_norm": 0.025667704641819,
+ "learning_rate": 0.0006,
+ "loss": 7.356142520904541,
+ "step": 126
+ },
+ {
+ "epoch": 1.7685589519650655,
+ "grad_norm": 0.036232151091098785,
+ "learning_rate": 0.0006,
+ "loss": 7.363539695739746,
+ "step": 127
+ },
+ {
+ "epoch": 1.782532751091703,
+ "grad_norm": 0.059483714401721954,
+ "learning_rate": 0.0006,
+ "loss": 7.402331352233887,
+ "step": 128
+ },
+ {
+ "epoch": 1.7965065502183406,
+ "grad_norm": 0.05365009233355522,
+ "learning_rate": 0.0006,
+ "loss": 7.40457820892334,
+ "step": 129
+ },
+ {
+ "epoch": 1.8104803493449781,
+ "grad_norm": 0.02738415263593197,
+ "learning_rate": 0.0006,
+ "loss": 7.3756208419799805,
+ "step": 130
+ },
+ {
+ "epoch": 1.8244541484716157,
+ "grad_norm": 0.035017091780900955,
+ "learning_rate": 0.0006,
+ "loss": 7.379263877868652,
+ "step": 131
+ },
+ {
+ "epoch": 1.8384279475982532,
+ "grad_norm": 0.02819863148033619,
+ "learning_rate": 0.0006,
+ "loss": 7.327673435211182,
+ "step": 132
+ },
+ {
+ "epoch": 1.8524017467248908,
+ "grad_norm": 0.017754238098859787,
+ "learning_rate": 0.0006,
+ "loss": 7.333420276641846,
+ "step": 133
+ },
+ {
+ "epoch": 1.8663755458515285,
+ "grad_norm": 0.024685733020305634,
+ "learning_rate": 0.0006,
+ "loss": 7.34493350982666,
+ "step": 134
+ },
+ {
+ "epoch": 1.8803493449781659,
+ "grad_norm": 0.019426532089710236,
+ "learning_rate": 0.0006,
+ "loss": 7.328448295593262,
+ "step": 135
+ },
+ {
+ "epoch": 1.8943231441048036,
+ "grad_norm": 0.02106391079723835,
+ "learning_rate": 0.0006,
+ "loss": 7.327917098999023,
+ "step": 136
+ },
+ {
+ "epoch": 1.908296943231441,
+ "grad_norm": 0.02056879736483097,
+ "learning_rate": 0.0006,
+ "loss": 7.3311991691589355,
+ "step": 137
+ },
+ {
+ "epoch": 1.9222707423580787,
+ "grad_norm": 0.016865141689777374,
+ "learning_rate": 0.0006,
+ "loss": 7.312766075134277,
+ "step": 138
+ },
+ {
+ "epoch": 1.936244541484716,
+ "grad_norm": 0.017549855634570122,
+ "learning_rate": 0.0006,
+ "loss": 7.31191349029541,
+ "step": 139
+ },
+ {
+ "epoch": 1.9502183406113538,
+ "grad_norm": 0.015930157154798508,
+ "learning_rate": 0.0006,
+ "loss": 7.330901622772217,
+ "step": 140
+ },
+ {
+ "epoch": 1.9641921397379911,
+ "grad_norm": 0.013283558189868927,
+ "learning_rate": 0.0006,
+ "loss": 7.298318386077881,
+ "step": 141
+ },
+ {
+ "epoch": 1.9781659388646289,
+ "grad_norm": 0.012023469433188438,
+ "learning_rate": 0.0006,
+ "loss": 7.282337188720703,
+ "step": 142
+ },
+ {
+ "epoch": 1.9921397379912664,
+ "grad_norm": 0.01459511648863554,
+ "learning_rate": 0.0006,
+ "loss": 7.282924175262451,
+ "step": 143
+ },
+ {
+ "epoch": 2.0,
+ "grad_norm": 0.017474235966801643,
+ "learning_rate": 0.0006,
+ "loss": 7.254333972930908,
+ "step": 144
+ },
+ {
+ "epoch": 2.0,
+ "eval_loss": 7.318708896636963,
+ "eval_runtime": 59.6555,
+ "eval_samples_per_second": 40.935,
+ "eval_steps_per_second": 1.291,
+ "step": 144
+ },
+ {
+ "epoch": 2.0139737991266378,
+ "grad_norm": 0.019757507368922234,
+ "learning_rate": 0.0006,
+ "loss": 7.30288553237915,
+ "step": 145
+ },
+ {
+ "epoch": 2.027947598253275,
+ "grad_norm": 0.014571248553693295,
+ "learning_rate": 0.0006,
+ "loss": 7.274681568145752,
+ "step": 146
+ },
+ {
+ "epoch": 2.041921397379913,
+ "grad_norm": 0.015744198113679886,
+ "learning_rate": 0.0006,
+ "loss": 7.278632640838623,
+ "step": 147
+ },
+ {
+ "epoch": 2.05589519650655,
+ "grad_norm": 0.028048858046531677,
+ "learning_rate": 0.0006,
+ "loss": 7.26558780670166,
+ "step": 148
+ },
+ {
+ "epoch": 2.069868995633188,
+ "grad_norm": 0.06547307223081589,
+ "learning_rate": 0.0006,
+ "loss": 7.32567834854126,
+ "step": 149
+ },
+ {
+ "epoch": 2.0838427947598253,
+ "grad_norm": 0.05969763547182083,
+ "learning_rate": 0.0006,
+ "loss": 7.395836353302002,
+ "step": 150
+ },
+ {
+ "epoch": 2.097816593886463,
+ "grad_norm": 0.032574281096458435,
+ "learning_rate": 0.0006,
+ "loss": 7.318150997161865,
+ "step": 151
+ },
+ {
+ "epoch": 2.1117903930131003,
+ "grad_norm": 0.029452962800860405,
+ "learning_rate": 0.0006,
+ "loss": 7.308066368103027,
+ "step": 152
+ },
+ {
+ "epoch": 2.125764192139738,
+ "grad_norm": 0.03161991387605667,
+ "learning_rate": 0.0006,
+ "loss": 7.306290626525879,
+ "step": 153
+ },
+ {
+ "epoch": 2.1397379912663754,
+ "grad_norm": 0.027096295729279518,
+ "learning_rate": 0.0006,
+ "loss": 7.297796249389648,
+ "step": 154
+ },
+ {
+ "epoch": 2.153711790393013,
+ "grad_norm": 0.028657056391239166,
+ "learning_rate": 0.0006,
+ "loss": 7.317424774169922,
+ "step": 155
+ },
+ {
+ "epoch": 2.1676855895196505,
+ "grad_norm": 0.02046995982527733,
+ "learning_rate": 0.0006,
+ "loss": 7.297986030578613,
+ "step": 156
+ },
+ {
+ "epoch": 2.1816593886462883,
+ "grad_norm": 0.02220369316637516,
+ "learning_rate": 0.0006,
+ "loss": 7.268237590789795,
+ "step": 157
+ },
+ {
+ "epoch": 2.1956331877729256,
+ "grad_norm": 0.018381357192993164,
+ "learning_rate": 0.0006,
+ "loss": 7.259415626525879,
+ "step": 158
+ },
+ {
+ "epoch": 2.2096069868995634,
+ "grad_norm": 0.0204413291066885,
+ "learning_rate": 0.0006,
+ "loss": 7.258173942565918,
+ "step": 159
+ },
+ {
+ "epoch": 2.2235807860262007,
+ "grad_norm": 0.015874288976192474,
+ "learning_rate": 0.0006,
+ "loss": 7.281834602355957,
+ "step": 160
+ },
+ {
+ "epoch": 2.2375545851528384,
+ "grad_norm": 0.01634068600833416,
+ "learning_rate": 0.0006,
+ "loss": 7.2469868659973145,
+ "step": 161
+ },
+ {
+ "epoch": 2.251528384279476,
+ "grad_norm": 0.01728747971355915,
+ "learning_rate": 0.0006,
+ "loss": 7.280995845794678,
+ "step": 162
+ },
+ {
+ "epoch": 2.2655021834061135,
+ "grad_norm": 0.016341188922524452,
+ "learning_rate": 0.0006,
+ "loss": 7.273357391357422,
+ "step": 163
+ },
+ {
+ "epoch": 2.279475982532751,
+ "grad_norm": 0.016125807538628578,
+ "learning_rate": 0.0006,
+ "loss": 7.273510932922363,
+ "step": 164
+ },
+ {
+ "epoch": 2.2934497816593886,
+ "grad_norm": 0.01629287749528885,
+ "learning_rate": 0.0006,
+ "loss": 7.267029762268066,
+ "step": 165
+ },
+ {
+ "epoch": 2.3074235807860264,
+ "grad_norm": 0.015296131372451782,
+ "learning_rate": 0.0006,
+ "loss": 7.211989402770996,
+ "step": 166
+ },
+ {
+ "epoch": 2.3213973799126637,
+ "grad_norm": 0.01709570363163948,
+ "learning_rate": 0.0006,
+ "loss": 7.242851734161377,
+ "step": 167
+ },
+ {
+ "epoch": 2.3353711790393015,
+ "grad_norm": 0.012552388943731785,
+ "learning_rate": 0.0006,
+ "loss": 7.241696834564209,
+ "step": 168
+ },
+ {
+ "epoch": 2.349344978165939,
+ "grad_norm": 0.014571291394531727,
+ "learning_rate": 0.0006,
+ "loss": 7.270681381225586,
+ "step": 169
+ },
+ {
+ "epoch": 2.3633187772925766,
+ "grad_norm": 0.013389437459409237,
+ "learning_rate": 0.0006,
+ "loss": 7.191600799560547,
+ "step": 170
+ },
+ {
+ "epoch": 2.377292576419214,
+ "grad_norm": 0.011895314790308475,
+ "learning_rate": 0.0006,
+ "loss": 7.213610649108887,
+ "step": 171
+ },
+ {
+ "epoch": 2.3912663755458516,
+ "grad_norm": 0.01345642190426588,
+ "learning_rate": 0.0006,
+ "loss": 7.193787574768066,
+ "step": 172
+ },
+ {
+ "epoch": 2.405240174672489,
+ "grad_norm": 0.0091794328764081,
+ "learning_rate": 0.0006,
+ "loss": 7.233364105224609,
+ "step": 173
+ },
+ {
+ "epoch": 2.4192139737991267,
+ "grad_norm": 0.013767086900770664,
+ "learning_rate": 0.0006,
+ "loss": 7.217272758483887,
+ "step": 174
+ },
+ {
+ "epoch": 2.433187772925764,
+ "grad_norm": 0.015134960412979126,
+ "learning_rate": 0.0006,
+ "loss": 7.192791938781738,
+ "step": 175
+ },
+ {
+ "epoch": 2.447161572052402,
+ "grad_norm": 0.01790648326277733,
+ "learning_rate": 0.0006,
+ "loss": 7.229467391967773,
+ "step": 176
+ },
+ {
+ "epoch": 2.461135371179039,
+ "grad_norm": 0.015362377278506756,
+ "learning_rate": 0.0006,
+ "loss": 7.180346488952637,
+ "step": 177
+ },
+ {
+ "epoch": 2.475109170305677,
+ "grad_norm": 0.010505498386919498,
+ "learning_rate": 0.0006,
+ "loss": 7.188833236694336,
+ "step": 178
+ },
+ {
+ "epoch": 2.489082969432314,
+ "grad_norm": 0.009129747748374939,
+ "learning_rate": 0.0006,
+ "loss": 7.203153133392334,
+ "step": 179
+ },
+ {
+ "epoch": 2.503056768558952,
+ "grad_norm": 0.01147883478552103,
+ "learning_rate": 0.0006,
+ "loss": 7.163824558258057,
+ "step": 180
+ },
+ {
+ "epoch": 2.5170305676855893,
+ "grad_norm": 0.011338942684233189,
+ "learning_rate": 0.0006,
+ "loss": 7.16287899017334,
+ "step": 181
+ },
+ {
+ "epoch": 2.531004366812227,
+ "grad_norm": 0.013277675025165081,
+ "learning_rate": 0.0006,
+ "loss": 7.155905723571777,
+ "step": 182
+ },
+ {
+ "epoch": 2.544978165938865,
+ "grad_norm": 0.015269347466528416,
+ "learning_rate": 0.0006,
+ "loss": 7.189371109008789,
+ "step": 183
+ },
+ {
+ "epoch": 2.558951965065502,
+ "grad_norm": 0.017905596643686295,
+ "learning_rate": 0.0006,
+ "loss": 7.177547931671143,
+ "step": 184
+ },
+ {
+ "epoch": 2.5729257641921395,
+ "grad_norm": 0.024431610479950905,
+ "learning_rate": 0.0006,
+ "loss": 7.192169189453125,
+ "step": 185
+ },
+ {
+ "epoch": 2.5868995633187772,
+ "grad_norm": 0.03656961768865585,
+ "learning_rate": 0.0006,
+ "loss": 7.124387741088867,
+ "step": 186
+ },
+ {
+ "epoch": 2.600873362445415,
+ "grad_norm": 0.04337048903107643,
+ "learning_rate": 0.0006,
+ "loss": 7.140231609344482,
+ "step": 187
+ },
+ {
+ "epoch": 2.6148471615720523,
+ "grad_norm": 0.04047228768467903,
+ "learning_rate": 0.0006,
+ "loss": 7.171927452087402,
+ "step": 188
+ },
+ {
+ "epoch": 2.62882096069869,
+ "grad_norm": 0.07825770974159241,
+ "learning_rate": 0.0006,
+ "loss": 7.207631587982178,
+ "step": 189
+ },
+ {
+ "epoch": 2.6427947598253274,
+ "grad_norm": 0.07118063420057297,
+ "learning_rate": 0.0006,
+ "loss": 7.209488391876221,
+ "step": 190
+ },
+ {
+ "epoch": 2.656768558951965,
+ "grad_norm": 0.05153293535113335,
+ "learning_rate": 0.0006,
+ "loss": 7.14793062210083,
+ "step": 191
+ },
+ {
+ "epoch": 2.6707423580786025,
+ "grad_norm": 0.03318839520215988,
+ "learning_rate": 0.0006,
+ "loss": 7.163010597229004,
+ "step": 192
+ },
+ {
+ "epoch": 2.6847161572052403,
+ "grad_norm": 0.029333539307117462,
+ "learning_rate": 0.0006,
+ "loss": 7.159955978393555,
+ "step": 193
+ },
+ {
+ "epoch": 2.6986899563318776,
+ "grad_norm": 0.029405880719423294,
+ "learning_rate": 0.0006,
+ "loss": 7.136178493499756,
+ "step": 194
+ },
+ {
+ "epoch": 2.7126637554585153,
+ "grad_norm": 0.037815988063812256,
+ "learning_rate": 0.0006,
+ "loss": 7.183186054229736,
+ "step": 195
+ },
+ {
+ "epoch": 2.726637554585153,
+ "grad_norm": 0.020307250320911407,
+ "learning_rate": 0.0006,
+ "loss": 7.133164882659912,
+ "step": 196
+ },
+ {
+ "epoch": 2.7406113537117904,
+ "grad_norm": 0.026048069819808006,
+ "learning_rate": 0.0006,
+ "loss": 7.124977111816406,
+ "step": 197
+ },
+ {
+ "epoch": 2.7545851528384278,
+ "grad_norm": 0.027316724881529808,
+ "learning_rate": 0.0006,
+ "loss": 7.07265567779541,
+ "step": 198
+ },
+ {
+ "epoch": 2.7685589519650655,
+ "grad_norm": 0.021629009395837784,
+ "learning_rate": 0.0006,
+ "loss": 7.150020599365234,
+ "step": 199
+ },
+ {
+ "epoch": 2.7825327510917033,
+ "grad_norm": 0.01605929434299469,
+ "learning_rate": 0.0006,
+ "loss": 7.10919189453125,
+ "step": 200
+ },
+ {
+ "epoch": 2.7965065502183406,
+ "grad_norm": 0.020800089463591576,
+ "learning_rate": 0.0006,
+ "loss": 7.132045269012451,
+ "step": 201
+ },
+ {
+ "epoch": 2.810480349344978,
+ "grad_norm": 0.019225774332880974,
+ "learning_rate": 0.0006,
+ "loss": 7.128364562988281,
+ "step": 202
+ },
+ {
+ "epoch": 2.8244541484716157,
+ "grad_norm": 0.01561372634023428,
+ "learning_rate": 0.0006,
+ "loss": 7.083669662475586,
+ "step": 203
+ },
+ {
+ "epoch": 2.8384279475982535,
+ "grad_norm": 0.021969519555568695,
+ "learning_rate": 0.0006,
+ "loss": 7.050841331481934,
+ "step": 204
+ },
+ {
+ "epoch": 2.8524017467248908,
+ "grad_norm": 0.021453695371747017,
+ "learning_rate": 0.0006,
+ "loss": 7.097085475921631,
+ "step": 205
+ },
+ {
+ "epoch": 2.8663755458515285,
+ "grad_norm": 0.022836022078990936,
+ "learning_rate": 0.0006,
+ "loss": 7.10405969619751,
+ "step": 206
+ },
+ {
+ "epoch": 2.880349344978166,
+ "grad_norm": 0.02430851384997368,
+ "learning_rate": 0.0006,
+ "loss": 7.045181751251221,
+ "step": 207
+ },
+ {
+ "epoch": 2.8943231441048036,
+ "grad_norm": 0.020166993141174316,
+ "learning_rate": 0.0006,
+ "loss": 7.068183898925781,
+ "step": 208
+ },
+ {
+ "epoch": 2.908296943231441,
+ "grad_norm": 0.0118505684658885,
+ "learning_rate": 0.0006,
+ "loss": 7.07132625579834,
+ "step": 209
+ },
+ {
+ "epoch": 2.9222707423580787,
+ "grad_norm": 0.023772427812218666,
+ "learning_rate": 0.0006,
+ "loss": 7.055238723754883,
+ "step": 210
+ },
+ {
+ "epoch": 2.936244541484716,
+ "grad_norm": 0.02350712940096855,
+ "learning_rate": 0.0006,
+ "loss": 7.055898666381836,
+ "step": 211
+ },
+ {
+ "epoch": 2.950218340611354,
+ "grad_norm": 0.017244907096028328,
+ "learning_rate": 0.0006,
+ "loss": 7.0099711418151855,
+ "step": 212
+ },
+ {
+ "epoch": 2.964192139737991,
+ "grad_norm": 0.021565575152635574,
+ "learning_rate": 0.0006,
+ "loss": 6.991751670837402,
+ "step": 213
+ },
+ {
+ "epoch": 2.978165938864629,
+ "grad_norm": 0.03241860866546631,
+ "learning_rate": 0.0006,
+ "loss": 7.005980491638184,
+ "step": 214
+ },
+ {
+ "epoch": 2.992139737991266,
+ "grad_norm": 0.04613726586103439,
+ "learning_rate": 0.0006,
+ "loss": 7.027859687805176,
+ "step": 215
+ },
+ {
+ "epoch": 3.0,
+ "grad_norm": 0.06404894590377808,
+ "learning_rate": 0.0006,
+ "loss": 7.123725891113281,
+ "step": 216
+ },
+ {
+ "epoch": 3.0,
+ "eval_loss": 7.070647716522217,
+ "eval_runtime": 58.5801,
+ "eval_samples_per_second": 41.687,
+ "eval_steps_per_second": 1.314,
+ "step": 216
+ },
+ {
+ "epoch": 3.0139737991266378,
+ "grad_norm": 0.05542432889342308,
+ "learning_rate": 0.0006,
+ "loss": 7.078299522399902,
+ "step": 217
+ },
+ {
+ "epoch": 3.027947598253275,
+ "grad_norm": 0.026484526693820953,
+ "learning_rate": 0.0006,
+ "loss": 6.961159706115723,
+ "step": 218
+ },
+ {
+ "epoch": 3.041921397379913,
+ "grad_norm": 0.028276391327381134,
+ "learning_rate": 0.0006,
+ "loss": 7.019961833953857,
+ "step": 219
+ },
+ {
+ "epoch": 3.05589519650655,
+ "grad_norm": 0.024486595764756203,
+ "learning_rate": 0.0006,
+ "loss": 7.039677619934082,
+ "step": 220
+ },
+ {
+ "epoch": 3.069868995633188,
+ "grad_norm": 0.017620893195271492,
+ "learning_rate": 0.0006,
+ "loss": 6.9270782470703125,
+ "step": 221
+ },
+ {
+ "epoch": 3.0838427947598253,
+ "grad_norm": 0.02547302283346653,
+ "learning_rate": 0.0006,
+ "loss": 6.924233436584473,
+ "step": 222
+ },
+ {
+ "epoch": 3.097816593886463,
+ "grad_norm": 0.02539009042084217,
+ "learning_rate": 0.0006,
+ "loss": 6.973979949951172,
+ "step": 223
+ },
+ {
+ "epoch": 3.1117903930131003,
+ "grad_norm": 0.03256227448582649,
+ "learning_rate": 0.0006,
+ "loss": 6.933725357055664,
+ "step": 224
+ },
+ {
+ "epoch": 3.125764192139738,
+ "grad_norm": 0.04651800915598869,
+ "learning_rate": 0.0006,
+ "loss": 6.951044082641602,
+ "step": 225
+ },
+ {
+ "epoch": 3.1397379912663754,
+ "grad_norm": 0.04822975769639015,
+ "learning_rate": 0.0006,
+ "loss": 6.9614410400390625,
+ "step": 226
+ },
+ {
+ "epoch": 3.153711790393013,
+ "grad_norm": 0.031165020540356636,
+ "learning_rate": 0.0006,
+ "loss": 6.91193151473999,
+ "step": 227
+ },
+ {
+ "epoch": 3.1676855895196505,
+ "grad_norm": 0.029754292219877243,
+ "learning_rate": 0.0006,
+ "loss": 6.91802978515625,
+ "step": 228
+ },
+ {
+ "epoch": 3.1816593886462883,
+ "grad_norm": 0.02193152718245983,
+ "learning_rate": 0.0006,
+ "loss": 6.928090572357178,
+ "step": 229
+ },
+ {
+ "epoch": 3.1956331877729256,
+ "grad_norm": 0.02428470179438591,
+ "learning_rate": 0.0006,
+ "loss": 6.90396785736084,
+ "step": 230
+ },
+ {
+ "epoch": 3.2096069868995634,
+ "grad_norm": 0.023374345153570175,
+ "learning_rate": 0.0006,
+ "loss": 6.885858535766602,
+ "step": 231
+ },
+ {
+ "epoch": 3.2235807860262007,
+ "grad_norm": 0.023405000567436218,
+ "learning_rate": 0.0006,
+ "loss": 6.8855767250061035,
+ "step": 232
+ },
+ {
+ "epoch": 3.2375545851528384,
+ "grad_norm": 0.017277134582400322,
+ "learning_rate": 0.0006,
+ "loss": 6.901449203491211,
+ "step": 233
+ },
+ {
+ "epoch": 3.251528384279476,
+ "grad_norm": 0.019496750086545944,
+ "learning_rate": 0.0006,
+ "loss": 6.885687828063965,
+ "step": 234
+ },
+ {
+ "epoch": 3.2655021834061135,
+ "grad_norm": 0.016893276944756508,
+ "learning_rate": 0.0006,
+ "loss": 6.872369766235352,
+ "step": 235
+ },
+ {
+ "epoch": 3.279475982532751,
+ "grad_norm": 0.02058715932071209,
+ "learning_rate": 0.0006,
+ "loss": 6.823280334472656,
+ "step": 236
+ },
+ {
+ "epoch": 3.2934497816593886,
+ "grad_norm": 0.015530755743384361,
+ "learning_rate": 0.0006,
+ "loss": 6.860499858856201,
+ "step": 237
+ },
+ {
+ "epoch": 3.3074235807860264,
+ "grad_norm": 0.01934926211833954,
+ "learning_rate": 0.0006,
+ "loss": 6.817448616027832,
+ "step": 238
+ },
+ {
+ "epoch": 3.3213973799126637,
+ "grad_norm": 0.017413552850484848,
+ "learning_rate": 0.0006,
+ "loss": 6.881955146789551,
+ "step": 239
+ },
+ {
+ "epoch": 3.3353711790393015,
+ "grad_norm": 0.026478875428438187,
+ "learning_rate": 0.0006,
+ "loss": 6.827404499053955,
+ "step": 240
+ },
+ {
+ "epoch": 3.349344978165939,
+ "grad_norm": 0.032611701637506485,
+ "learning_rate": 0.0006,
+ "loss": 6.78285551071167,
+ "step": 241
+ },
+ {
+ "epoch": 3.3633187772925766,
+ "grad_norm": 0.041900306940078735,
+ "learning_rate": 0.0006,
+ "loss": 6.783053874969482,
+ "step": 242
+ },
+ {
+ "epoch": 3.377292576419214,
+ "grad_norm": 0.04149497672915459,
+ "learning_rate": 0.0006,
+ "loss": 6.745220184326172,
+ "step": 243
+ },
+ {
+ "epoch": 3.3912663755458516,
+ "grad_norm": 0.023949826136231422,
+ "learning_rate": 0.0006,
+ "loss": 6.841533184051514,
+ "step": 244
+ },
+ {
+ "epoch": 3.405240174672489,
+ "grad_norm": 0.028601842001080513,
+ "learning_rate": 0.0006,
+ "loss": 6.852852821350098,
+ "step": 245
+ },
+ {
+ "epoch": 3.4192139737991267,
+ "grad_norm": 0.027365155518054962,
+ "learning_rate": 0.0006,
+ "loss": 6.74376106262207,
+ "step": 246
+ },
+ {
+ "epoch": 3.433187772925764,
+ "grad_norm": 0.02856568619608879,
+ "learning_rate": 0.0006,
+ "loss": 6.74906063079834,
+ "step": 247
+ },
+ {
+ "epoch": 3.447161572052402,
+ "grad_norm": 0.019172416999936104,
+ "learning_rate": 0.0006,
+ "loss": 6.809206008911133,
+ "step": 248
+ },
+ {
+ "epoch": 3.461135371179039,
+ "grad_norm": 0.02161746844649315,
+ "learning_rate": 0.0006,
+ "loss": 6.789888381958008,
+ "step": 249
+ },
+ {
+ "epoch": 3.475109170305677,
+ "grad_norm": 0.020273666828870773,
+ "learning_rate": 0.0006,
+ "loss": 6.73512077331543,
+ "step": 250
+ },
+ {
+ "epoch": 3.489082969432314,
+ "grad_norm": 0.022329283878207207,
+ "learning_rate": 0.0006,
+ "loss": 6.826148509979248,
+ "step": 251
+ },
+ {
+ "epoch": 3.503056768558952,
+ "grad_norm": 0.02284284122288227,
+ "learning_rate": 0.0006,
+ "loss": 6.748521327972412,
+ "step": 252
+ },
+ {
+ "epoch": 3.5170305676855893,
+ "grad_norm": 0.026368247345089912,
+ "learning_rate": 0.0006,
+ "loss": 6.645879745483398,
+ "step": 253
+ },
+ {
+ "epoch": 3.531004366812227,
+ "grad_norm": 0.03627244383096695,
+ "learning_rate": 0.0006,
+ "loss": 6.666064739227295,
+ "step": 254
+ },
+ {
+ "epoch": 3.544978165938865,
+ "grad_norm": 0.04140935093164444,
+ "learning_rate": 0.0006,
+ "loss": 6.712275981903076,
+ "step": 255
+ },
+ {
+ "epoch": 3.558951965065502,
+ "grad_norm": 0.0390462800860405,
+ "learning_rate": 0.0006,
+ "loss": 6.665900230407715,
+ "step": 256
+ },
+ {
+ "epoch": 3.5729257641921395,
+ "grad_norm": 0.05178583785891533,
+ "learning_rate": 0.0006,
+ "loss": 6.753298759460449,
+ "step": 257
+ },
+ {
+ "epoch": 3.5868995633187772,
+ "grad_norm": 0.054328061640262604,
+ "learning_rate": 0.0006,
+ "loss": 6.732633113861084,
+ "step": 258
+ },
+ {
+ "epoch": 3.600873362445415,
+ "grad_norm": 0.04590460658073425,
+ "learning_rate": 0.0006,
+ "loss": 6.65250301361084,
+ "step": 259
+ },
+ {
+ "epoch": 3.6148471615720523,
+ "grad_norm": 0.044331666082143784,
+ "learning_rate": 0.0006,
+ "loss": 6.68654203414917,
+ "step": 260
+ },
+ {
+ "epoch": 3.62882096069869,
+ "grad_norm": 0.05308730527758598,
+ "learning_rate": 0.0006,
+ "loss": 6.642482757568359,
+ "step": 261
+ },
+ {
+ "epoch": 3.6427947598253274,
+ "grad_norm": 0.03999016433954239,
+ "learning_rate": 0.0006,
+ "loss": 6.625227451324463,
+ "step": 262
+ },
+ {
+ "epoch": 3.656768558951965,
+ "grad_norm": 0.04088086634874344,
+ "learning_rate": 0.0006,
+ "loss": 6.67623233795166,
+ "step": 263
+ },
+ {
+ "epoch": 3.6707423580786025,
+ "grad_norm": 0.03615636005997658,
+ "learning_rate": 0.0006,
+ "loss": 6.6869001388549805,
+ "step": 264
+ },
+ {
+ "epoch": 3.6847161572052403,
+ "grad_norm": 0.02817360684275627,
+ "learning_rate": 0.0006,
+ "loss": 6.674601078033447,
+ "step": 265
+ },
+ {
+ "epoch": 3.6986899563318776,
+ "grad_norm": 0.032586123794317245,
+ "learning_rate": 0.0006,
+ "loss": 6.612994194030762,
+ "step": 266
+ },
+ {
+ "epoch": 3.7126637554585153,
+ "grad_norm": 0.031117867678403854,
+ "learning_rate": 0.0006,
+ "loss": 6.6244354248046875,
+ "step": 267
+ },
+ {
+ "epoch": 3.726637554585153,
+ "grad_norm": 0.028330687433481216,
+ "learning_rate": 0.0006,
+ "loss": 6.669557094573975,
+ "step": 268
+ },
+ {
+ "epoch": 3.7406113537117904,
+ "grad_norm": 0.021489202976226807,
+ "learning_rate": 0.0006,
+ "loss": 6.632035732269287,
+ "step": 269
+ },
+ {
+ "epoch": 3.7545851528384278,
+ "grad_norm": 0.024269424378871918,
+ "learning_rate": 0.0006,
+ "loss": 6.660867691040039,
+ "step": 270
+ },
+ {
+ "epoch": 3.7685589519650655,
+ "grad_norm": 0.02237873338162899,
+ "learning_rate": 0.0006,
+ "loss": 6.602587699890137,
+ "step": 271
+ },
+ {
+ "epoch": 3.7825327510917033,
+ "grad_norm": 0.026503929868340492,
+ "learning_rate": 0.0006,
+ "loss": 6.599200248718262,
+ "step": 272
+ },
+ {
+ "epoch": 3.7965065502183406,
+ "grad_norm": 0.03687124326825142,
+ "learning_rate": 0.0006,
+ "loss": 6.567296504974365,
+ "step": 273
+ },
+ {
+ "epoch": 3.810480349344978,
+ "grad_norm": 0.04708952456712723,
+ "learning_rate": 0.0006,
+ "loss": 6.621265411376953,
+ "step": 274
+ },
+ {
+ "epoch": 3.8244541484716157,
+ "grad_norm": 0.041846614331007004,
+ "learning_rate": 0.0006,
+ "loss": 6.550536155700684,
+ "step": 275
+ },
+ {
+ "epoch": 3.8384279475982535,
+ "grad_norm": 0.031249675899744034,
+ "learning_rate": 0.0006,
+ "loss": 6.54278564453125,
+ "step": 276
+ },
+ {
+ "epoch": 3.8524017467248908,
+ "grad_norm": 0.023082256317138672,
+ "learning_rate": 0.0006,
+ "loss": 6.544787406921387,
+ "step": 277
+ },
+ {
+ "epoch": 3.8663755458515285,
+ "grad_norm": 0.020995570346713066,
+ "learning_rate": 0.0006,
+ "loss": 6.518143653869629,
+ "step": 278
+ },
+ {
+ "epoch": 3.880349344978166,
+ "grad_norm": 0.020839324221014977,
+ "learning_rate": 0.0006,
+ "loss": 6.546667098999023,
+ "step": 279
+ },
+ {
+ "epoch": 3.8943231441048036,
+ "grad_norm": 0.018377432599663734,
+ "learning_rate": 0.0006,
+ "loss": 6.478307723999023,
+ "step": 280
+ },
+ {
+ "epoch": 3.908296943231441,
+ "grad_norm": 0.021713724359869957,
+ "learning_rate": 0.0006,
+ "loss": 6.485574245452881,
+ "step": 281
+ },
+ {
+ "epoch": 3.9222707423580787,
+ "grad_norm": 0.01359301246702671,
+ "learning_rate": 0.0006,
+ "loss": 6.570268630981445,
+ "step": 282
+ },
+ {
+ "epoch": 3.936244541484716,
+ "grad_norm": 0.016233332455158234,
+ "learning_rate": 0.0006,
+ "loss": 6.57150936126709,
+ "step": 283
+ },
+ {
+ "epoch": 3.950218340611354,
+ "grad_norm": 0.020687608048319817,
+ "learning_rate": 0.0006,
+ "loss": 6.527956008911133,
+ "step": 284
+ },
+ {
+ "epoch": 3.964192139737991,
+ "grad_norm": 0.02032964862883091,
+ "learning_rate": 0.0006,
+ "loss": 6.468169689178467,
+ "step": 285
+ },
+ {
+ "epoch": 3.978165938864629,
+ "grad_norm": 0.016676288098096848,
+ "learning_rate": 0.0006,
+ "loss": 6.448418617248535,
+ "step": 286
+ },
+ {
+ "epoch": 3.992139737991266,
+ "grad_norm": 0.0168539360165596,
+ "learning_rate": 0.0006,
+ "loss": 6.461180686950684,
+ "step": 287
+ },
+ {
+ "epoch": 4.0,
+ "grad_norm": 0.02666233666241169,
+ "learning_rate": 0.0006,
+ "loss": 6.344893932342529,
+ "step": 288
+ },
+ {
+ "epoch": 4.0,
+ "eval_loss": 6.5384979248046875,
+ "eval_runtime": 58.814,
+ "eval_samples_per_second": 41.521,
+ "eval_steps_per_second": 1.309,
+ "step": 288
+ },
+ {
+ "epoch": 4.013973799126638,
+ "grad_norm": 0.04154708981513977,
+ "learning_rate": 0.0006,
+ "loss": 6.41855525970459,
+ "step": 289
+ },
+ {
+ "epoch": 4.0279475982532755,
+ "grad_norm": 0.04895668104290962,
+ "learning_rate": 0.0006,
+ "loss": 6.499302864074707,
+ "step": 290
+ },
+ {
+ "epoch": 4.041921397379912,
+ "grad_norm": 0.034464482218027115,
+ "learning_rate": 0.0006,
+ "loss": 6.438611030578613,
+ "step": 291
+ },
+ {
+ "epoch": 4.05589519650655,
+ "grad_norm": 0.06845773756504059,
+ "learning_rate": 0.0006,
+ "loss": 6.538361072540283,
+ "step": 292
+ },
+ {
+ "epoch": 4.069868995633188,
+ "grad_norm": 0.053768858313560486,
+ "learning_rate": 0.0006,
+ "loss": 6.527524948120117,
+ "step": 293
+ },
+ {
+ "epoch": 4.083842794759826,
+ "grad_norm": 0.039037931710481644,
+ "learning_rate": 0.0006,
+ "loss": 6.503847122192383,
+ "step": 294
+ },
+ {
+ "epoch": 4.097816593886463,
+ "grad_norm": 0.03962196782231331,
+ "learning_rate": 0.0006,
+ "loss": 6.475178241729736,
+ "step": 295
+ },
+ {
+ "epoch": 4.1117903930131,
+ "grad_norm": 0.049101535230875015,
+ "learning_rate": 0.0006,
+ "loss": 6.474587917327881,
+ "step": 296
+ },
+ {
+ "epoch": 4.125764192139738,
+ "grad_norm": 0.02830282226204872,
+ "learning_rate": 0.0006,
+ "loss": 6.45065975189209,
+ "step": 297
+ },
+ {
+ "epoch": 4.139737991266376,
+ "grad_norm": 0.023153482005000114,
+ "learning_rate": 0.0006,
+ "loss": 6.5218706130981445,
+ "step": 298
+ },
+ {
+ "epoch": 4.153711790393013,
+ "grad_norm": 0.025064002722501755,
+ "learning_rate": 0.0006,
+ "loss": 6.456475257873535,
+ "step": 299
+ },
+ {
+ "epoch": 4.1676855895196505,
+ "grad_norm": 0.02357092685997486,
+ "learning_rate": 0.0006,
+ "loss": 6.485952377319336,
+ "step": 300
+ },
+ {
+ "epoch": 4.181659388646288,
+ "grad_norm": 0.021832305938005447,
+ "learning_rate": 0.0006,
+ "loss": 6.425506591796875,
+ "step": 301
+ },
+ {
+ "epoch": 4.195633187772926,
+ "grad_norm": 0.0227900929749012,
+ "learning_rate": 0.0006,
+ "loss": 6.439155578613281,
+ "step": 302
+ },
+ {
+ "epoch": 4.209606986899563,
+ "grad_norm": 0.015961607918143272,
+ "learning_rate": 0.0006,
+ "loss": 6.460862636566162,
+ "step": 303
+ },
+ {
+ "epoch": 4.223580786026201,
+ "grad_norm": 0.017876390367746353,
+ "learning_rate": 0.0006,
+ "loss": 6.44564151763916,
+ "step": 304
+ },
+ {
+ "epoch": 4.2375545851528384,
+ "grad_norm": 0.013703204691410065,
+ "learning_rate": 0.0006,
+ "loss": 6.418414115905762,
+ "step": 305
+ },
+ {
+ "epoch": 4.251528384279476,
+ "grad_norm": 0.015240306034684181,
+ "learning_rate": 0.0006,
+ "loss": 6.411238670349121,
+ "step": 306
+ },
+ {
+ "epoch": 4.265502183406113,
+ "grad_norm": 0.01587662845849991,
+ "learning_rate": 0.0006,
+ "loss": 6.395642280578613,
+ "step": 307
+ },
+ {
+ "epoch": 4.279475982532751,
+ "grad_norm": 0.014650700613856316,
+ "learning_rate": 0.0006,
+ "loss": 6.302793025970459,
+ "step": 308
+ },
+ {
+ "epoch": 4.293449781659389,
+ "grad_norm": 0.016533225774765015,
+ "learning_rate": 0.0006,
+ "loss": 6.380588054656982,
+ "step": 309
+ },
+ {
+ "epoch": 4.307423580786026,
+ "grad_norm": 0.019963741302490234,
+ "learning_rate": 0.0006,
+ "loss": 6.345336437225342,
+ "step": 310
+ },
+ {
+ "epoch": 4.321397379912664,
+ "grad_norm": 0.02237936295568943,
+ "learning_rate": 0.0006,
+ "loss": 6.387770652770996,
+ "step": 311
+ },
+ {
+ "epoch": 4.335371179039301,
+ "grad_norm": 0.026658328250050545,
+ "learning_rate": 0.0006,
+ "loss": 6.323662757873535,
+ "step": 312
+ },
+ {
+ "epoch": 4.349344978165939,
+ "grad_norm": 0.02852284163236618,
+ "learning_rate": 0.0006,
+ "loss": 6.382204055786133,
+ "step": 313
+ },
+ {
+ "epoch": 4.3633187772925766,
+ "grad_norm": 0.028815090656280518,
+ "learning_rate": 0.0006,
+ "loss": 6.402390480041504,
+ "step": 314
+ },
+ {
+ "epoch": 4.377292576419214,
+ "grad_norm": 0.029393529519438744,
+ "learning_rate": 0.0006,
+ "loss": 6.280278205871582,
+ "step": 315
+ },
+ {
+ "epoch": 4.391266375545851,
+ "grad_norm": 0.02186041697859764,
+ "learning_rate": 0.0006,
+ "loss": 6.302707672119141,
+ "step": 316
+ },
+ {
+ "epoch": 4.405240174672489,
+ "grad_norm": 0.02071218006312847,
+ "learning_rate": 0.0006,
+ "loss": 6.3365020751953125,
+ "step": 317
+ },
+ {
+ "epoch": 4.419213973799127,
+ "grad_norm": 0.02779117226600647,
+ "learning_rate": 0.0006,
+ "loss": 6.252157211303711,
+ "step": 318
+ },
+ {
+ "epoch": 4.4331877729257645,
+ "grad_norm": 0.03470654785633087,
+ "learning_rate": 0.0006,
+ "loss": 6.319070339202881,
+ "step": 319
+ },
+ {
+ "epoch": 4.447161572052401,
+ "grad_norm": 0.037067804485559464,
+ "learning_rate": 0.0006,
+ "loss": 6.314116477966309,
+ "step": 320
+ },
+ {
+ "epoch": 4.461135371179039,
+ "grad_norm": 0.032182756811380386,
+ "learning_rate": 0.0006,
+ "loss": 6.267298698425293,
+ "step": 321
+ },
+ {
+ "epoch": 4.475109170305677,
+ "grad_norm": 0.026305217295885086,
+ "learning_rate": 0.0006,
+ "loss": 6.331688404083252,
+ "step": 322
+ },
+ {
+ "epoch": 4.489082969432315,
+ "grad_norm": 0.027228357270359993,
+ "learning_rate": 0.0006,
+ "loss": 6.31139612197876,
+ "step": 323
+ },
+ {
+ "epoch": 4.503056768558952,
+ "grad_norm": 0.025781169533729553,
+ "learning_rate": 0.0006,
+ "loss": 6.194684982299805,
+ "step": 324
+ },
+ {
+ "epoch": 4.517030567685589,
+ "grad_norm": 0.029186977073550224,
+ "learning_rate": 0.0006,
+ "loss": 6.31306266784668,
+ "step": 325
+ },
+ {
+ "epoch": 4.531004366812227,
+ "grad_norm": 0.023608777672052383,
+ "learning_rate": 0.0006,
+ "loss": 6.285243034362793,
+ "step": 326
+ },
+ {
+ "epoch": 4.544978165938865,
+ "grad_norm": 0.02267594076693058,
+ "learning_rate": 0.0006,
+ "loss": 6.30342960357666,
+ "step": 327
+ },
+ {
+ "epoch": 4.558951965065502,
+ "grad_norm": 0.01926310732960701,
+ "learning_rate": 0.0006,
+ "loss": 6.280606269836426,
+ "step": 328
+ },
+ {
+ "epoch": 4.5729257641921395,
+ "grad_norm": 0.026484837755560875,
+ "learning_rate": 0.0006,
+ "loss": 6.237980365753174,
+ "step": 329
+ },
+ {
+ "epoch": 4.586899563318777,
+ "grad_norm": 0.026228854432702065,
+ "learning_rate": 0.0006,
+ "loss": 6.3018293380737305,
+ "step": 330
+ },
+ {
+ "epoch": 4.600873362445415,
+ "grad_norm": 0.022096967324614525,
+ "learning_rate": 0.0006,
+ "loss": 6.253863334655762,
+ "step": 331
+ },
+ {
+ "epoch": 4.614847161572053,
+ "grad_norm": 0.027223890647292137,
+ "learning_rate": 0.0006,
+ "loss": 6.166882038116455,
+ "step": 332
+ },
+ {
+ "epoch": 4.62882096069869,
+ "grad_norm": 0.03160124272108078,
+ "learning_rate": 0.0006,
+ "loss": 6.230309963226318,
+ "step": 333
+ },
+ {
+ "epoch": 4.642794759825327,
+ "grad_norm": 0.04110539332032204,
+ "learning_rate": 0.0006,
+ "loss": 6.2334818840026855,
+ "step": 334
+ },
+ {
+ "epoch": 4.656768558951965,
+ "grad_norm": 0.04355933889746666,
+ "learning_rate": 0.0006,
+ "loss": 6.160956382751465,
+ "step": 335
+ },
+ {
+ "epoch": 4.670742358078603,
+ "grad_norm": 0.03611086308956146,
+ "learning_rate": 0.0006,
+ "loss": 6.23399543762207,
+ "step": 336
+ },
+ {
+ "epoch": 4.68471615720524,
+ "grad_norm": 0.051811400800943375,
+ "learning_rate": 0.0006,
+ "loss": 6.241855144500732,
+ "step": 337
+ },
+ {
+ "epoch": 4.698689956331878,
+ "grad_norm": 0.057434193789958954,
+ "learning_rate": 0.0006,
+ "loss": 6.220993995666504,
+ "step": 338
+ },
+ {
+ "epoch": 4.712663755458515,
+ "grad_norm": 0.04420953616499901,
+ "learning_rate": 0.0006,
+ "loss": 6.300461292266846,
+ "step": 339
+ },
+ {
+ "epoch": 4.726637554585153,
+ "grad_norm": 0.03515457361936569,
+ "learning_rate": 0.0006,
+ "loss": 6.269349098205566,
+ "step": 340
+ },
+ {
+ "epoch": 4.74061135371179,
+ "grad_norm": 0.03222600743174553,
+ "learning_rate": 0.0006,
+ "loss": 6.248613357543945,
+ "step": 341
+ },
+ {
+ "epoch": 4.754585152838428,
+ "grad_norm": 0.0384046845138073,
+ "learning_rate": 0.0006,
+ "loss": 6.266979217529297,
+ "step": 342
+ },
+ {
+ "epoch": 4.7685589519650655,
+ "grad_norm": 0.0514867827296257,
+ "learning_rate": 0.0006,
+ "loss": 6.234103202819824,
+ "step": 343
+ },
+ {
+ "epoch": 4.782532751091703,
+ "grad_norm": 0.050747793167829514,
+ "learning_rate": 0.0006,
+ "loss": 6.322582244873047,
+ "step": 344
+ },
+ {
+ "epoch": 4.796506550218341,
+ "grad_norm": 0.040878988802433014,
+ "learning_rate": 0.0006,
+ "loss": 6.212509632110596,
+ "step": 345
+ },
+ {
+ "epoch": 4.810480349344978,
+ "grad_norm": 0.034655820578336716,
+ "learning_rate": 0.0006,
+ "loss": 6.265879154205322,
+ "step": 346
+ },
+ {
+ "epoch": 4.824454148471616,
+ "grad_norm": 0.04202224686741829,
+ "learning_rate": 0.0006,
+ "loss": 6.2896409034729,
+ "step": 347
+ },
+ {
+ "epoch": 4.8384279475982535,
+ "grad_norm": 0.03632277995347977,
+ "learning_rate": 0.0006,
+ "loss": 6.253917217254639,
+ "step": 348
+ },
+ {
+ "epoch": 4.85240174672489,
+ "grad_norm": 0.028645765036344528,
+ "learning_rate": 0.0006,
+ "loss": 6.207965850830078,
+ "step": 349
+ },
+ {
+ "epoch": 4.866375545851528,
+ "grad_norm": 0.025305170565843582,
+ "learning_rate": 0.0006,
+ "loss": 6.168068885803223,
+ "step": 350
+ },
+ {
+ "epoch": 4.880349344978166,
+ "grad_norm": 0.022137803956866264,
+ "learning_rate": 0.0006,
+ "loss": 6.235283374786377,
+ "step": 351
+ },
+ {
+ "epoch": 4.894323144104804,
+ "grad_norm": 0.01814538985490799,
+ "learning_rate": 0.0006,
+ "loss": 6.259641647338867,
+ "step": 352
+ },
+ {
+ "epoch": 4.908296943231441,
+ "grad_norm": 0.018465086817741394,
+ "learning_rate": 0.0006,
+ "loss": 6.111738681793213,
+ "step": 353
+ },
+ {
+ "epoch": 4.922270742358078,
+ "grad_norm": 0.018638933077454567,
+ "learning_rate": 0.0006,
+ "loss": 6.225642204284668,
+ "step": 354
+ },
+ {
+ "epoch": 4.936244541484716,
+ "grad_norm": 0.016461260616779327,
+ "learning_rate": 0.0006,
+ "loss": 6.140647888183594,
+ "step": 355
+ },
+ {
+ "epoch": 4.950218340611354,
+ "grad_norm": 0.015458385460078716,
+ "learning_rate": 0.0006,
+ "loss": 6.155224800109863,
+ "step": 356
+ },
+ {
+ "epoch": 4.964192139737992,
+ "grad_norm": 0.014934048056602478,
+ "learning_rate": 0.0006,
+ "loss": 6.101555824279785,
+ "step": 357
+ },
+ {
+ "epoch": 4.978165938864628,
+ "grad_norm": 0.013979545794427395,
+ "learning_rate": 0.0006,
+ "loss": 6.161718845367432,
+ "step": 358
+ },
+ {
+ "epoch": 4.992139737991266,
+ "grad_norm": 0.01640394888818264,
+ "learning_rate": 0.0006,
+ "loss": 6.055768966674805,
+ "step": 359
+ },
+ {
+ "epoch": 5.0,
+ "grad_norm": 0.015135680325329304,
+ "learning_rate": 0.0006,
+ "loss": 6.169695854187012,
+ "step": 360
+ },
+ {
+ "epoch": 5.0,
+ "eval_loss": 6.16408109664917,
+ "eval_runtime": 59.3649,
+ "eval_samples_per_second": 41.135,
+ "eval_steps_per_second": 1.297,
+ "step": 360
+ },
+ {
+ "epoch": 5.013973799126638,
+ "grad_norm": 0.012849048711359501,
+ "learning_rate": 0.0006,
+ "loss": 6.055830001831055,
+ "step": 361
+ },
+ {
+ "epoch": 5.0279475982532755,
+ "grad_norm": 0.018169576302170753,
+ "learning_rate": 0.0006,
+ "loss": 6.082655429840088,
+ "step": 362
+ },
+ {
+ "epoch": 5.041921397379912,
+ "grad_norm": 0.024293430149555206,
+ "learning_rate": 0.0006,
+ "loss": 6.109032154083252,
+ "step": 363
+ },
+ {
+ "epoch": 5.05589519650655,
+ "grad_norm": 0.03266787901520729,
+ "learning_rate": 0.0006,
+ "loss": 6.030613899230957,
+ "step": 364
+ },
+ {
+ "epoch": 5.069868995633188,
+ "grad_norm": 0.03572266176342964,
+ "learning_rate": 0.0006,
+ "loss": 6.050390720367432,
+ "step": 365
+ },
+ {
+ "epoch": 5.083842794759826,
+ "grad_norm": 0.032625213265419006,
+ "learning_rate": 0.0006,
+ "loss": 6.169313430786133,
+ "step": 366
+ },
+ {
+ "epoch": 5.097816593886463,
+ "grad_norm": 0.04051872715353966,
+ "learning_rate": 0.0006,
+ "loss": 6.10398006439209,
+ "step": 367
+ },
+ {
+ "epoch": 5.1117903930131,
+ "grad_norm": 0.0351213738322258,
+ "learning_rate": 0.0006,
+ "loss": 6.121973514556885,
+ "step": 368
+ },
+ {
+ "epoch": 5.125764192139738,
+ "grad_norm": 0.03523759916424751,
+ "learning_rate": 0.0006,
+ "loss": 6.0626654624938965,
+ "step": 369
+ },
+ {
+ "epoch": 5.139737991266376,
+ "grad_norm": 0.02706182189285755,
+ "learning_rate": 0.0006,
+ "loss": 6.049354076385498,
+ "step": 370
+ },
+ {
+ "epoch": 5.153711790393013,
+ "grad_norm": 0.024212589487433434,
+ "learning_rate": 0.0006,
+ "loss": 6.034826278686523,
+ "step": 371
+ },
+ {
+ "epoch": 5.1676855895196505,
+ "grad_norm": 0.024798082187771797,
+ "learning_rate": 0.0006,
+ "loss": 6.004058837890625,
+ "step": 372
+ },
+ {
+ "epoch": 5.181659388646288,
+ "grad_norm": 0.02325567416846752,
+ "learning_rate": 0.0006,
+ "loss": 5.986461639404297,
+ "step": 373
+ },
+ {
+ "epoch": 5.195633187772926,
+ "grad_norm": 0.019060570746660233,
+ "learning_rate": 0.0006,
+ "loss": 6.059736251831055,
+ "step": 374
+ },
+ {
+ "epoch": 5.209606986899563,
+ "grad_norm": 0.016822976991534233,
+ "learning_rate": 0.0006,
+ "loss": 6.072957515716553,
+ "step": 375
+ },
+ {
+ "epoch": 5.223580786026201,
+ "grad_norm": 0.017218658700585365,
+ "learning_rate": 0.0006,
+ "loss": 6.031739234924316,
+ "step": 376
+ },
+ {
+ "epoch": 5.2375545851528384,
+ "grad_norm": 0.014889383688569069,
+ "learning_rate": 0.0006,
+ "loss": 6.025674343109131,
+ "step": 377
+ },
+ {
+ "epoch": 5.251528384279476,
+ "grad_norm": 0.01708107627928257,
+ "learning_rate": 0.0006,
+ "loss": 6.090451240539551,
+ "step": 378
+ },
+ {
+ "epoch": 5.265502183406113,
+ "grad_norm": 0.020260317251086235,
+ "learning_rate": 0.0006,
+ "loss": 6.094257354736328,
+ "step": 379
+ },
+ {
+ "epoch": 5.279475982532751,
+ "grad_norm": 0.020110400393605232,
+ "learning_rate": 0.0006,
+ "loss": 6.021188735961914,
+ "step": 380
+ },
+ {
+ "epoch": 5.293449781659389,
+ "grad_norm": 0.02027830481529236,
+ "learning_rate": 0.0006,
+ "loss": 6.022156238555908,
+ "step": 381
+ },
+ {
+ "epoch": 5.307423580786026,
+ "grad_norm": 0.022747062146663666,
+ "learning_rate": 0.0006,
+ "loss": 6.011836051940918,
+ "step": 382
+ },
+ {
+ "epoch": 5.321397379912664,
+ "grad_norm": 0.01990230567753315,
+ "learning_rate": 0.0006,
+ "loss": 6.003054618835449,
+ "step": 383
+ },
+ {
+ "epoch": 5.335371179039301,
+ "grad_norm": 0.01596238650381565,
+ "learning_rate": 0.0006,
+ "loss": 5.991410732269287,
+ "step": 384
+ },
+ {
+ "epoch": 5.349344978165939,
+ "grad_norm": 0.015847105532884598,
+ "learning_rate": 0.0006,
+ "loss": 6.100622177124023,
+ "step": 385
+ },
+ {
+ "epoch": 5.3633187772925766,
+ "grad_norm": 0.016179397702217102,
+ "learning_rate": 0.0006,
+ "loss": 6.032659530639648,
+ "step": 386
+ },
+ {
+ "epoch": 5.377292576419214,
+ "grad_norm": 0.018256602808833122,
+ "learning_rate": 0.0006,
+ "loss": 5.961983680725098,
+ "step": 387
+ },
+ {
+ "epoch": 5.391266375545851,
+ "grad_norm": 0.02805912122130394,
+ "learning_rate": 0.0006,
+ "loss": 5.983541965484619,
+ "step": 388
+ },
+ {
+ "epoch": 5.405240174672489,
+ "grad_norm": 0.039082255214452744,
+ "learning_rate": 0.0006,
+ "loss": 6.021474361419678,
+ "step": 389
+ },
+ {
+ "epoch": 5.419213973799127,
+ "grad_norm": 0.036757659167051315,
+ "learning_rate": 0.0006,
+ "loss": 6.005046367645264,
+ "step": 390
+ },
+ {
+ "epoch": 5.4331877729257645,
+ "grad_norm": 0.035277366638183594,
+ "learning_rate": 0.0006,
+ "loss": 6.036296844482422,
+ "step": 391
+ },
+ {
+ "epoch": 5.447161572052401,
+ "grad_norm": 0.034404635429382324,
+ "learning_rate": 0.0006,
+ "loss": 5.994539260864258,
+ "step": 392
+ },
+ {
+ "epoch": 5.461135371179039,
+ "grad_norm": 0.0377206988632679,
+ "learning_rate": 0.0006,
+ "loss": 5.941534519195557,
+ "step": 393
+ },
+ {
+ "epoch": 5.475109170305677,
+ "grad_norm": 0.0389922633767128,
+ "learning_rate": 0.0006,
+ "loss": 5.983644485473633,
+ "step": 394
+ },
+ {
+ "epoch": 5.489082969432315,
+ "grad_norm": 0.04176778718829155,
+ "learning_rate": 0.0006,
+ "loss": 6.030922889709473,
+ "step": 395
+ },
+ {
+ "epoch": 5.503056768558952,
+ "grad_norm": 0.0337153784930706,
+ "learning_rate": 0.0006,
+ "loss": 5.961367130279541,
+ "step": 396
+ },
+ {
+ "epoch": 5.517030567685589,
+ "grad_norm": 0.036116816103458405,
+ "learning_rate": 0.0006,
+ "loss": 5.819280624389648,
+ "step": 397
+ },
+ {
+ "epoch": 5.531004366812227,
+ "grad_norm": 0.030724041163921356,
+ "learning_rate": 0.0006,
+ "loss": 5.906380653381348,
+ "step": 398
+ },
+ {
+ "epoch": 5.544978165938865,
+ "grad_norm": 0.030264202505350113,
+ "learning_rate": 0.0006,
+ "loss": 5.975833415985107,
+ "step": 399
+ },
+ {
+ "epoch": 5.558951965065502,
+ "grad_norm": 0.032096318900585175,
+ "learning_rate": 0.0006,
+ "loss": 6.001348972320557,
+ "step": 400
+ },
+ {
+ "epoch": 5.5729257641921395,
+ "grad_norm": 0.030579356476664543,
+ "learning_rate": 0.0006,
+ "loss": 5.928768634796143,
+ "step": 401
+ },
+ {
+ "epoch": 5.586899563318777,
+ "grad_norm": 0.028241973370313644,
+ "learning_rate": 0.0006,
+ "loss": 5.92582893371582,
+ "step": 402
+ },
+ {
+ "epoch": 5.600873362445415,
+ "grad_norm": 0.023086953908205032,
+ "learning_rate": 0.0006,
+ "loss": 5.892926216125488,
+ "step": 403
+ },
+ {
+ "epoch": 5.614847161572053,
+ "grad_norm": 0.02547992393374443,
+ "learning_rate": 0.0006,
+ "loss": 5.903195381164551,
+ "step": 404
+ },
+ {
+ "epoch": 5.62882096069869,
+ "grad_norm": 0.023089831694960594,
+ "learning_rate": 0.0006,
+ "loss": 5.912033557891846,
+ "step": 405
+ },
+ {
+ "epoch": 5.642794759825327,
+ "grad_norm": 0.025446368381381035,
+ "learning_rate": 0.0006,
+ "loss": 5.926138401031494,
+ "step": 406
+ },
+ {
+ "epoch": 5.656768558951965,
+ "grad_norm": 0.03748747706413269,
+ "learning_rate": 0.0006,
+ "loss": 5.913451194763184,
+ "step": 407
+ },
+ {
+ "epoch": 5.670742358078603,
+ "grad_norm": 0.03493810072541237,
+ "learning_rate": 0.0006,
+ "loss": 5.97633695602417,
+ "step": 408
+ },
+ {
+ "epoch": 5.68471615720524,
+ "grad_norm": 0.023131275549530983,
+ "learning_rate": 0.0006,
+ "loss": 5.899571418762207,
+ "step": 409
+ },
+ {
+ "epoch": 5.698689956331878,
+ "grad_norm": 0.02986014075577259,
+ "learning_rate": 0.0006,
+ "loss": 5.894903182983398,
+ "step": 410
+ },
+ {
+ "epoch": 5.712663755458515,
+ "grad_norm": 0.030171144753694534,
+ "learning_rate": 0.0006,
+ "loss": 5.876595497131348,
+ "step": 411
+ },
+ {
+ "epoch": 5.726637554585153,
+ "grad_norm": 0.029377546161413193,
+ "learning_rate": 0.0006,
+ "loss": 5.911637306213379,
+ "step": 412
+ },
+ {
+ "epoch": 5.74061135371179,
+ "grad_norm": 0.029829490929841995,
+ "learning_rate": 0.0006,
+ "loss": 5.819629669189453,
+ "step": 413
+ },
+ {
+ "epoch": 5.754585152838428,
+ "grad_norm": 0.024139299988746643,
+ "learning_rate": 0.0006,
+ "loss": 5.81728458404541,
+ "step": 414
+ },
+ {
+ "epoch": 5.7685589519650655,
+ "grad_norm": 0.021453432738780975,
+ "learning_rate": 0.0006,
+ "loss": 5.987326145172119,
+ "step": 415
+ },
+ {
+ "epoch": 5.782532751091703,
+ "grad_norm": 0.017942246049642563,
+ "learning_rate": 0.0006,
+ "loss": 5.893008232116699,
+ "step": 416
+ },
+ {
+ "epoch": 5.796506550218341,
+ "grad_norm": 0.019724637269973755,
+ "learning_rate": 0.0006,
+ "loss": 5.912441253662109,
+ "step": 417
+ },
+ {
+ "epoch": 5.810480349344978,
+ "grad_norm": 0.018597912043333054,
+ "learning_rate": 0.0006,
+ "loss": 5.860394477844238,
+ "step": 418
+ },
+ {
+ "epoch": 5.824454148471616,
+ "grad_norm": 0.016232412308454514,
+ "learning_rate": 0.0006,
+ "loss": 5.94991397857666,
+ "step": 419
+ },
+ {
+ "epoch": 5.8384279475982535,
+ "grad_norm": 0.01667204685509205,
+ "learning_rate": 0.0006,
+ "loss": 5.8506364822387695,
+ "step": 420
+ },
+ {
+ "epoch": 5.85240174672489,
+ "grad_norm": 0.014232151210308075,
+ "learning_rate": 0.0006,
+ "loss": 5.928424835205078,
+ "step": 421
+ },
+ {
+ "epoch": 5.866375545851528,
+ "grad_norm": 0.01570476032793522,
+ "learning_rate": 0.0006,
+ "loss": 5.953692436218262,
+ "step": 422
+ },
+ {
+ "epoch": 5.880349344978166,
+ "grad_norm": 0.01586179807782173,
+ "learning_rate": 0.0006,
+ "loss": 5.820014476776123,
+ "step": 423
+ },
+ {
+ "epoch": 5.894323144104804,
+ "grad_norm": 0.014729145914316177,
+ "learning_rate": 0.0006,
+ "loss": 5.860154628753662,
+ "step": 424
+ },
+ {
+ "epoch": 5.908296943231441,
+ "grad_norm": 0.013111459091305733,
+ "learning_rate": 0.0006,
+ "loss": 5.85988187789917,
+ "step": 425
+ },
+ {
+ "epoch": 5.922270742358078,
+ "grad_norm": 0.012937922962009907,
+ "learning_rate": 0.0006,
+ "loss": 5.756265640258789,
+ "step": 426
+ },
+ {
+ "epoch": 5.936244541484716,
+ "grad_norm": 0.01377453189343214,
+ "learning_rate": 0.0006,
+ "loss": 5.8333048820495605,
+ "step": 427
+ },
+ {
+ "epoch": 5.950218340611354,
+ "grad_norm": 0.014054795727133751,
+ "learning_rate": 0.0006,
+ "loss": 5.803333759307861,
+ "step": 428
+ },
+ {
+ "epoch": 5.964192139737992,
+ "grad_norm": 0.01674959622323513,
+ "learning_rate": 0.0006,
+ "loss": 5.8718485832214355,
+ "step": 429
+ },
+ {
+ "epoch": 5.978165938864628,
+ "grad_norm": 0.018733017146587372,
+ "learning_rate": 0.0006,
+ "loss": 5.875979900360107,
+ "step": 430
+ },
+ {
+ "epoch": 5.992139737991266,
+ "grad_norm": 0.02088983915746212,
+ "learning_rate": 0.0006,
+ "loss": 5.763634204864502,
+ "step": 431
+ },
+ {
+ "epoch": 6.0,
+ "grad_norm": 0.026705985888838768,
+ "learning_rate": 0.0006,
+ "loss": 5.6922478675842285,
+ "step": 432
+ },
+ {
+ "epoch": 6.0,
+ "eval_loss": 5.847958087921143,
+ "eval_runtime": 58.5451,
+ "eval_samples_per_second": 41.711,
+ "eval_steps_per_second": 1.315,
+ "step": 432
+ },
+ {
+ "epoch": 6.013973799126638,
+ "grad_norm": 0.03181544691324234,
+ "learning_rate": 0.0006,
+ "loss": 5.786368370056152,
+ "step": 433
+ },
+ {
+ "epoch": 6.0279475982532755,
+ "grad_norm": 0.03238112851977348,
+ "learning_rate": 0.0006,
+ "loss": 5.810310363769531,
+ "step": 434
+ },
+ {
+ "epoch": 6.041921397379912,
+ "grad_norm": 0.03916756808757782,
+ "learning_rate": 0.0006,
+ "loss": 5.7761993408203125,
+ "step": 435
+ },
+ {
+ "epoch": 6.05589519650655,
+ "grad_norm": 0.044009730219841,
+ "learning_rate": 0.0006,
+ "loss": 5.898112773895264,
+ "step": 436
+ },
+ {
+ "epoch": 6.069868995633188,
+ "grad_norm": 0.045235246419906616,
+ "learning_rate": 0.0006,
+ "loss": 5.876371383666992,
+ "step": 437
+ },
+ {
+ "epoch": 6.083842794759826,
+ "grad_norm": 0.050218384712934494,
+ "learning_rate": 0.0006,
+ "loss": 5.805103302001953,
+ "step": 438
+ },
+ {
+ "epoch": 6.097816593886463,
+ "grad_norm": 0.05444490164518356,
+ "learning_rate": 0.0006,
+ "loss": 5.826424598693848,
+ "step": 439
+ },
+ {
+ "epoch": 6.1117903930131,
+ "grad_norm": 0.04890982061624527,
+ "learning_rate": 0.0006,
+ "loss": 5.7735395431518555,
+ "step": 440
+ },
+ {
+ "epoch": 6.125764192139738,
+ "grad_norm": 0.05478639155626297,
+ "learning_rate": 0.0006,
+ "loss": 5.823677062988281,
+ "step": 441
+ },
+ {
+ "epoch": 6.139737991266376,
+ "grad_norm": 0.057562313973903656,
+ "learning_rate": 0.0006,
+ "loss": 5.889334678649902,
+ "step": 442
+ },
+ {
+ "epoch": 6.153711790393013,
+ "grad_norm": 0.06447703391313553,
+ "learning_rate": 0.0006,
+ "loss": 5.84621524810791,
+ "step": 443
+ },
+ {
+ "epoch": 6.1676855895196505,
+ "grad_norm": 0.048861872404813766,
+ "learning_rate": 0.0006,
+ "loss": 5.931595802307129,
+ "step": 444
+ },
+ {
+ "epoch": 6.181659388646288,
+ "grad_norm": 0.05521339178085327,
+ "learning_rate": 0.0006,
+ "loss": 5.8229780197143555,
+ "step": 445
+ },
+ {
+ "epoch": 6.195633187772926,
+ "grad_norm": 0.053666990250349045,
+ "learning_rate": 0.0006,
+ "loss": 5.879191875457764,
+ "step": 446
+ },
+ {
+ "epoch": 6.209606986899563,
+ "grad_norm": 0.0451025515794754,
+ "learning_rate": 0.0006,
+ "loss": 5.801628112792969,
+ "step": 447
+ },
+ {
+ "epoch": 6.223580786026201,
+ "grad_norm": 0.04291655868291855,
+ "learning_rate": 0.0006,
+ "loss": 5.914680480957031,
+ "step": 448
+ },
+ {
+ "epoch": 6.2375545851528384,
+ "grad_norm": 0.035102490335702896,
+ "learning_rate": 0.0006,
+ "loss": 5.843682765960693,
+ "step": 449
+ },
+ {
+ "epoch": 6.251528384279476,
+ "grad_norm": 0.03403995931148529,
+ "learning_rate": 0.0006,
+ "loss": 5.827154159545898,
+ "step": 450
+ },
+ {
+ "epoch": 6.265502183406113,
+ "grad_norm": 0.03444375470280647,
+ "learning_rate": 0.0006,
+ "loss": 5.800136566162109,
+ "step": 451
+ },
+ {
+ "epoch": 6.279475982532751,
+ "grad_norm": 0.03165159001946449,
+ "learning_rate": 0.0006,
+ "loss": 5.906252384185791,
+ "step": 452
+ },
+ {
+ "epoch": 6.293449781659389,
+ "grad_norm": 0.026153093203902245,
+ "learning_rate": 0.0006,
+ "loss": 5.7423200607299805,
+ "step": 453
+ },
+ {
+ "epoch": 6.307423580786026,
+ "grad_norm": 0.024357657879590988,
+ "learning_rate": 0.0006,
+ "loss": 5.701364517211914,
+ "step": 454
+ },
+ {
+ "epoch": 6.321397379912664,
+ "grad_norm": 0.021508049219846725,
+ "learning_rate": 0.0006,
+ "loss": 5.8328094482421875,
+ "step": 455
+ },
+ {
+ "epoch": 6.335371179039301,
+ "grad_norm": 0.017313921824097633,
+ "learning_rate": 0.0006,
+ "loss": 5.792145252227783,
+ "step": 456
+ },
+ {
+ "epoch": 6.349344978165939,
+ "grad_norm": 0.018563397228717804,
+ "learning_rate": 0.0006,
+ "loss": 5.732672214508057,
+ "step": 457
+ },
+ {
+ "epoch": 6.3633187772925766,
+ "grad_norm": 0.016568679362535477,
+ "learning_rate": 0.0006,
+ "loss": 5.81948184967041,
+ "step": 458
+ },
+ {
+ "epoch": 6.377292576419214,
+ "grad_norm": 0.014133770950138569,
+ "learning_rate": 0.0006,
+ "loss": 5.734982490539551,
+ "step": 459
+ },
+ {
+ "epoch": 6.391266375545851,
+ "grad_norm": 0.014114447869360447,
+ "learning_rate": 0.0006,
+ "loss": 5.70073127746582,
+ "step": 460
+ },
+ {
+ "epoch": 6.405240174672489,
+ "grad_norm": 0.012907393276691437,
+ "learning_rate": 0.0006,
+ "loss": 5.758626937866211,
+ "step": 461
+ },
+ {
+ "epoch": 6.419213973799127,
+ "grad_norm": 0.012673444114625454,
+ "learning_rate": 0.0006,
+ "loss": 5.775138854980469,
+ "step": 462
+ },
+ {
+ "epoch": 6.4331877729257645,
+ "grad_norm": 0.01311410591006279,
+ "learning_rate": 0.0006,
+ "loss": 5.70920991897583,
+ "step": 463
+ },
+ {
+ "epoch": 6.447161572052401,
+ "grad_norm": 0.012935544364154339,
+ "learning_rate": 0.0006,
+ "loss": 5.781479835510254,
+ "step": 464
+ },
+ {
+ "epoch": 6.461135371179039,
+ "grad_norm": 0.01439738366752863,
+ "learning_rate": 0.0006,
+ "loss": 5.691315650939941,
+ "step": 465
+ },
+ {
+ "epoch": 6.475109170305677,
+ "grad_norm": 0.01205186452716589,
+ "learning_rate": 0.0006,
+ "loss": 5.662499904632568,
+ "step": 466
+ },
+ {
+ "epoch": 6.489082969432315,
+ "grad_norm": 0.011575652286410332,
+ "learning_rate": 0.0006,
+ "loss": 5.6271772384643555,
+ "step": 467
+ },
+ {
+ "epoch": 6.503056768558952,
+ "grad_norm": 0.011672502383589745,
+ "learning_rate": 0.0006,
+ "loss": 5.761662006378174,
+ "step": 468
+ },
+ {
+ "epoch": 6.517030567685589,
+ "grad_norm": 0.011541751213371754,
+ "learning_rate": 0.0006,
+ "loss": 5.762078285217285,
+ "step": 469
+ },
+ {
+ "epoch": 6.531004366812227,
+ "grad_norm": 0.011396365240216255,
+ "learning_rate": 0.0006,
+ "loss": 5.67873477935791,
+ "step": 470
+ },
+ {
+ "epoch": 6.544978165938865,
+ "grad_norm": 0.010688499547541142,
+ "learning_rate": 0.0006,
+ "loss": 5.717051982879639,
+ "step": 471
+ },
+ {
+ "epoch": 6.558951965065502,
+ "grad_norm": 0.012224317528307438,
+ "learning_rate": 0.0006,
+ "loss": 5.707948684692383,
+ "step": 472
+ },
+ {
+ "epoch": 6.5729257641921395,
+ "grad_norm": 0.011856825090944767,
+ "learning_rate": 0.0006,
+ "loss": 5.70878791809082,
+ "step": 473
+ },
+ {
+ "epoch": 6.586899563318777,
+ "grad_norm": 0.01199064590036869,
+ "learning_rate": 0.0006,
+ "loss": 5.708697319030762,
+ "step": 474
+ },
+ {
+ "epoch": 6.600873362445415,
+ "grad_norm": 0.01219446212053299,
+ "learning_rate": 0.0006,
+ "loss": 5.6061577796936035,
+ "step": 475
+ },
+ {
+ "epoch": 6.614847161572053,
+ "grad_norm": 0.013397484086453915,
+ "learning_rate": 0.0006,
+ "loss": 5.624789714813232,
+ "step": 476
+ },
+ {
+ "epoch": 6.62882096069869,
+ "grad_norm": 0.01483136136084795,
+ "learning_rate": 0.0006,
+ "loss": 5.721141338348389,
+ "step": 477
+ },
+ {
+ "epoch": 6.642794759825327,
+ "grad_norm": 0.0194488987326622,
+ "learning_rate": 0.0006,
+ "loss": 5.604279518127441,
+ "step": 478
+ },
+ {
+ "epoch": 6.656768558951965,
+ "grad_norm": 0.02079896256327629,
+ "learning_rate": 0.0006,
+ "loss": 5.586322784423828,
+ "step": 479
+ },
+ {
+ "epoch": 6.670742358078603,
+ "grad_norm": 0.018996229395270348,
+ "learning_rate": 0.0006,
+ "loss": 5.5987772941589355,
+ "step": 480
+ },
+ {
+ "epoch": 6.68471615720524,
+ "grad_norm": 0.015779603272676468,
+ "learning_rate": 0.0006,
+ "loss": 5.588602066040039,
+ "step": 481
+ },
+ {
+ "epoch": 6.698689956331878,
+ "grad_norm": 0.015322973020374775,
+ "learning_rate": 0.0006,
+ "loss": 5.686467170715332,
+ "step": 482
+ },
+ {
+ "epoch": 6.712663755458515,
+ "grad_norm": 0.015282213687896729,
+ "learning_rate": 0.0006,
+ "loss": 5.664676666259766,
+ "step": 483
+ },
+ {
+ "epoch": 6.726637554585153,
+ "grad_norm": 0.016589272767305374,
+ "learning_rate": 0.0006,
+ "loss": 5.666739463806152,
+ "step": 484
+ },
+ {
+ "epoch": 6.74061135371179,
+ "grad_norm": 0.01811421848833561,
+ "learning_rate": 0.0006,
+ "loss": 5.647593021392822,
+ "step": 485
+ },
+ {
+ "epoch": 6.754585152838428,
+ "grad_norm": 0.019652029499411583,
+ "learning_rate": 0.0006,
+ "loss": 5.628726959228516,
+ "step": 486
+ },
+ {
+ "epoch": 6.7685589519650655,
+ "grad_norm": 0.02118665538728237,
+ "learning_rate": 0.0006,
+ "loss": 5.6580705642700195,
+ "step": 487
+ },
+ {
+ "epoch": 6.782532751091703,
+ "grad_norm": 0.02237832546234131,
+ "learning_rate": 0.0006,
+ "loss": 5.635931968688965,
+ "step": 488
+ },
+ {
+ "epoch": 6.796506550218341,
+ "grad_norm": 0.023897631093859673,
+ "learning_rate": 0.0006,
+ "loss": 5.517949104309082,
+ "step": 489
+ },
+ {
+ "epoch": 6.810480349344978,
+ "grad_norm": 0.02442227490246296,
+ "learning_rate": 0.0006,
+ "loss": 5.600021839141846,
+ "step": 490
+ },
+ {
+ "epoch": 6.824454148471616,
+ "grad_norm": 0.024675287306308746,
+ "learning_rate": 0.0006,
+ "loss": 5.6639909744262695,
+ "step": 491
+ },
+ {
+ "epoch": 6.8384279475982535,
+ "grad_norm": 0.030372392386198044,
+ "learning_rate": 0.0006,
+ "loss": 5.582888603210449,
+ "step": 492
+ },
+ {
+ "epoch": 6.85240174672489,
+ "grad_norm": 0.03237690031528473,
+ "learning_rate": 0.0006,
+ "loss": 5.573896884918213,
+ "step": 493
+ },
+ {
+ "epoch": 6.866375545851528,
+ "grad_norm": 0.03511589393019676,
+ "learning_rate": 0.0006,
+ "loss": 5.565878868103027,
+ "step": 494
+ },
+ {
+ "epoch": 6.880349344978166,
+ "grad_norm": 0.03883938118815422,
+ "learning_rate": 0.0006,
+ "loss": 5.58540678024292,
+ "step": 495
+ },
+ {
+ "epoch": 6.894323144104804,
+ "grad_norm": 0.04175502806901932,
+ "learning_rate": 0.0006,
+ "loss": 5.6095194816589355,
+ "step": 496
+ },
+ {
+ "epoch": 6.908296943231441,
+ "grad_norm": 0.04013441503047943,
+ "learning_rate": 0.0006,
+ "loss": 5.654401779174805,
+ "step": 497
+ },
+ {
+ "epoch": 6.922270742358078,
+ "grad_norm": 0.04618770629167557,
+ "learning_rate": 0.0006,
+ "loss": 5.6483473777771,
+ "step": 498
+ },
+ {
+ "epoch": 6.936244541484716,
+ "grad_norm": 0.04507308453321457,
+ "learning_rate": 0.0006,
+ "loss": 5.5856218338012695,
+ "step": 499
+ },
+ {
+ "epoch": 6.950218340611354,
+ "grad_norm": 0.03151383996009827,
+ "learning_rate": 0.0006,
+ "loss": 5.594086647033691,
+ "step": 500
+ },
+ {
+ "epoch": 6.964192139737992,
+ "grad_norm": 0.026883797720074654,
+ "learning_rate": 0.0006,
+ "loss": 5.607676029205322,
+ "step": 501
+ },
+ {
+ "epoch": 6.978165938864628,
+ "grad_norm": 0.02981836162507534,
+ "learning_rate": 0.0006,
+ "loss": 5.563666820526123,
+ "step": 502
+ },
+ {
+ "epoch": 6.992139737991266,
+ "grad_norm": 0.03233477845788002,
+ "learning_rate": 0.0006,
+ "loss": 5.667543411254883,
+ "step": 503
+ },
+ {
+ "epoch": 7.0,
+ "grad_norm": 0.029927456751465797,
+ "learning_rate": 0.0006,
+ "loss": 5.733482837677002,
+ "step": 504
+ },
+ {
+ "epoch": 7.0,
+ "eval_loss": 5.679076671600342,
+ "eval_runtime": 59.5965,
+ "eval_samples_per_second": 40.976,
+ "eval_steps_per_second": 1.292,
+ "step": 504
+ },
+ {
+ "epoch": 7.013973799126638,
+ "grad_norm": 0.027081597596406937,
+ "learning_rate": 0.0006,
+ "loss": 5.634487152099609,
+ "step": 505
+ },
+ {
+ "epoch": 7.0279475982532755,
+ "grad_norm": 0.031959421932697296,
+ "learning_rate": 0.0006,
+ "loss": 5.643294811248779,
+ "step": 506
+ },
+ {
+ "epoch": 7.041921397379912,
+ "grad_norm": 0.02802063524723053,
+ "learning_rate": 0.0006,
+ "loss": 5.573239326477051,
+ "step": 507
+ },
+ {
+ "epoch": 7.05589519650655,
+ "grad_norm": 0.032579705119132996,
+ "learning_rate": 0.0006,
+ "loss": 5.627150058746338,
+ "step": 508
+ },
+ {
+ "epoch": 7.069868995633188,
+ "grad_norm": 0.02503928542137146,
+ "learning_rate": 0.0006,
+ "loss": 5.6045308113098145,
+ "step": 509
+ },
+ {
+ "epoch": 7.083842794759826,
+ "grad_norm": 0.02421317622065544,
+ "learning_rate": 0.0006,
+ "loss": 5.522153377532959,
+ "step": 510
+ },
+ {
+ "epoch": 7.097816593886463,
+ "grad_norm": 0.021983426064252853,
+ "learning_rate": 0.0006,
+ "loss": 5.632939338684082,
+ "step": 511
+ },
+ {
+ "epoch": 7.1117903930131,
+ "grad_norm": 0.021933183073997498,
+ "learning_rate": 0.0006,
+ "loss": 5.592185974121094,
+ "step": 512
+ },
+ {
+ "epoch": 7.125764192139738,
+ "grad_norm": 0.02287031151354313,
+ "learning_rate": 0.0006,
+ "loss": 5.613700866699219,
+ "step": 513
+ },
+ {
+ "epoch": 7.139737991266376,
+ "grad_norm": 0.02281602844595909,
+ "learning_rate": 0.0006,
+ "loss": 5.551383018493652,
+ "step": 514
+ },
+ {
+ "epoch": 7.153711790393013,
+ "grad_norm": 0.021050360053777695,
+ "learning_rate": 0.0006,
+ "loss": 5.55275821685791,
+ "step": 515
+ },
+ {
+ "epoch": 7.1676855895196505,
+ "grad_norm": 0.018299926072359085,
+ "learning_rate": 0.0006,
+ "loss": 5.528225421905518,
+ "step": 516
+ },
+ {
+ "epoch": 7.181659388646288,
+ "grad_norm": 0.02024853602051735,
+ "learning_rate": 0.0006,
+ "loss": 5.522453784942627,
+ "step": 517
+ },
+ {
+ "epoch": 7.195633187772926,
+ "grad_norm": 0.02135239914059639,
+ "learning_rate": 0.0006,
+ "loss": 5.527522563934326,
+ "step": 518
+ },
+ {
+ "epoch": 7.209606986899563,
+ "grad_norm": 0.023394184187054634,
+ "learning_rate": 0.0006,
+ "loss": 5.4535017013549805,
+ "step": 519
+ },
+ {
+ "epoch": 7.223580786026201,
+ "grad_norm": 0.025952080264687538,
+ "learning_rate": 0.0006,
+ "loss": 5.5374674797058105,
+ "step": 520
+ },
+ {
+ "epoch": 7.2375545851528384,
+ "grad_norm": 0.022287718951702118,
+ "learning_rate": 0.0006,
+ "loss": 5.493753433227539,
+ "step": 521
+ },
+ {
+ "epoch": 7.251528384279476,
+ "grad_norm": 0.02018447406589985,
+ "learning_rate": 0.0006,
+ "loss": 5.510575294494629,
+ "step": 522
+ },
+ {
+ "epoch": 7.265502183406113,
+ "grad_norm": 0.02050507813692093,
+ "learning_rate": 0.0006,
+ "loss": 5.466026782989502,
+ "step": 523
+ },
+ {
+ "epoch": 7.279475982532751,
+ "grad_norm": 0.023688802495598793,
+ "learning_rate": 0.0006,
+ "loss": 5.485815525054932,
+ "step": 524
+ },
+ {
+ "epoch": 7.293449781659389,
+ "grad_norm": 0.021078843623399734,
+ "learning_rate": 0.0006,
+ "loss": 5.415581703186035,
+ "step": 525
+ },
+ {
+ "epoch": 7.307423580786026,
+ "grad_norm": 0.018266011029481888,
+ "learning_rate": 0.0006,
+ "loss": 5.402815818786621,
+ "step": 526
+ },
+ {
+ "epoch": 7.321397379912664,
+ "grad_norm": 0.019693493843078613,
+ "learning_rate": 0.0006,
+ "loss": 5.504674911499023,
+ "step": 527
+ },
+ {
+ "epoch": 7.335371179039301,
+ "grad_norm": 0.02031373605132103,
+ "learning_rate": 0.0006,
+ "loss": 5.503837585449219,
+ "step": 528
+ },
+ {
+ "epoch": 7.349344978165939,
+ "grad_norm": 0.017636626958847046,
+ "learning_rate": 0.0006,
+ "loss": 5.530580997467041,
+ "step": 529
+ },
+ {
+ "epoch": 7.3633187772925766,
+ "grad_norm": 0.01787244901061058,
+ "learning_rate": 0.0006,
+ "loss": 5.450218677520752,
+ "step": 530
+ },
+ {
+ "epoch": 7.377292576419214,
+ "grad_norm": 0.0170669574290514,
+ "learning_rate": 0.0006,
+ "loss": 5.481570243835449,
+ "step": 531
+ },
+ {
+ "epoch": 7.391266375545851,
+ "grad_norm": 0.01802165061235428,
+ "learning_rate": 0.0006,
+ "loss": 5.386394500732422,
+ "step": 532
+ },
+ {
+ "epoch": 7.405240174672489,
+ "grad_norm": 0.022949472069740295,
+ "learning_rate": 0.0006,
+ "loss": 5.424929618835449,
+ "step": 533
+ },
+ {
+ "epoch": 7.419213973799127,
+ "grad_norm": 0.034213390201330185,
+ "learning_rate": 0.0006,
+ "loss": 5.346663475036621,
+ "step": 534
+ },
+ {
+ "epoch": 7.4331877729257645,
+ "grad_norm": 0.03847593814134598,
+ "learning_rate": 0.0006,
+ "loss": 5.418482780456543,
+ "step": 535
+ },
+ {
+ "epoch": 7.447161572052401,
+ "grad_norm": 0.029393676668405533,
+ "learning_rate": 0.0006,
+ "loss": 5.456090927124023,
+ "step": 536
+ },
+ {
+ "epoch": 7.461135371179039,
+ "grad_norm": 0.03388667106628418,
+ "learning_rate": 0.0006,
+ "loss": 5.572983741760254,
+ "step": 537
+ },
+ {
+ "epoch": 7.475109170305677,
+ "grad_norm": 0.039690613746643066,
+ "learning_rate": 0.0006,
+ "loss": 5.414067268371582,
+ "step": 538
+ },
+ {
+ "epoch": 7.489082969432315,
+ "grad_norm": 0.03630776330828667,
+ "learning_rate": 0.0006,
+ "loss": 5.523739814758301,
+ "step": 539
+ },
+ {
+ "epoch": 7.503056768558952,
+ "grad_norm": 0.03356870263814926,
+ "learning_rate": 0.0006,
+ "loss": 5.444767951965332,
+ "step": 540
+ },
+ {
+ "epoch": 7.517030567685589,
+ "grad_norm": 0.030812304466962814,
+ "learning_rate": 0.0006,
+ "loss": 5.461244106292725,
+ "step": 541
+ },
+ {
+ "epoch": 7.531004366812227,
+ "grad_norm": 0.03319217637181282,
+ "learning_rate": 0.0006,
+ "loss": 5.502161026000977,
+ "step": 542
+ },
+ {
+ "epoch": 7.544978165938865,
+ "grad_norm": 0.032997481524944305,
+ "learning_rate": 0.0006,
+ "loss": 5.4162278175354,
+ "step": 543
+ },
+ {
+ "epoch": 7.558951965065502,
+ "grad_norm": 0.03364962339401245,
+ "learning_rate": 0.0006,
+ "loss": 5.415736198425293,
+ "step": 544
+ },
+ {
+ "epoch": 7.5729257641921395,
+ "grad_norm": 0.0344221405684948,
+ "learning_rate": 0.0006,
+ "loss": 5.541967868804932,
+ "step": 545
+ },
+ {
+ "epoch": 7.586899563318777,
+ "grad_norm": 0.029609503224492073,
+ "learning_rate": 0.0006,
+ "loss": 5.440614223480225,
+ "step": 546
+ },
+ {
+ "epoch": 7.600873362445415,
+ "grad_norm": 0.029055926948785782,
+ "learning_rate": 0.0006,
+ "loss": 5.462865352630615,
+ "step": 547
+ },
+ {
+ "epoch": 7.614847161572053,
+ "grad_norm": 0.02658848837018013,
+ "learning_rate": 0.0006,
+ "loss": 5.444467544555664,
+ "step": 548
+ },
+ {
+ "epoch": 7.62882096069869,
+ "grad_norm": 0.026276404038071632,
+ "learning_rate": 0.0006,
+ "loss": 5.462278842926025,
+ "step": 549
+ },
+ {
+ "epoch": 7.642794759825327,
+ "grad_norm": 0.0286889486014843,
+ "learning_rate": 0.0006,
+ "loss": 5.4849534034729,
+ "step": 550
+ },
+ {
+ "epoch": 7.656768558951965,
+ "grad_norm": 0.026849817484617233,
+ "learning_rate": 0.0006,
+ "loss": 5.431473731994629,
+ "step": 551
+ },
+ {
+ "epoch": 7.670742358078603,
+ "grad_norm": 0.02312396466732025,
+ "learning_rate": 0.0006,
+ "loss": 5.45462703704834,
+ "step": 552
+ },
+ {
+ "epoch": 7.68471615720524,
+ "grad_norm": 0.026197485625743866,
+ "learning_rate": 0.0006,
+ "loss": 5.452552318572998,
+ "step": 553
+ },
+ {
+ "epoch": 7.698689956331878,
+ "grad_norm": 0.02747255191206932,
+ "learning_rate": 0.0006,
+ "loss": 5.447169780731201,
+ "step": 554
+ },
+ {
+ "epoch": 7.712663755458515,
+ "grad_norm": 0.028123432770371437,
+ "learning_rate": 0.0006,
+ "loss": 5.394400596618652,
+ "step": 555
+ },
+ {
+ "epoch": 7.726637554585153,
+ "grad_norm": 0.02599870041012764,
+ "learning_rate": 0.0006,
+ "loss": 5.398341178894043,
+ "step": 556
+ },
+ {
+ "epoch": 7.74061135371179,
+ "grad_norm": 0.022171657532453537,
+ "learning_rate": 0.0006,
+ "loss": 5.368820667266846,
+ "step": 557
+ },
+ {
+ "epoch": 7.754585152838428,
+ "grad_norm": 0.022309480234980583,
+ "learning_rate": 0.0006,
+ "loss": 5.416772365570068,
+ "step": 558
+ },
+ {
+ "epoch": 7.7685589519650655,
+ "grad_norm": 0.024074165150523186,
+ "learning_rate": 0.0006,
+ "loss": 5.359002113342285,
+ "step": 559
+ },
+ {
+ "epoch": 7.782532751091703,
+ "grad_norm": 0.02636653557419777,
+ "learning_rate": 0.0006,
+ "loss": 5.484368324279785,
+ "step": 560
+ },
+ {
+ "epoch": 7.796506550218341,
+ "grad_norm": 0.02203752100467682,
+ "learning_rate": 0.0006,
+ "loss": 5.375498294830322,
+ "step": 561
+ },
+ {
+ "epoch": 7.810480349344978,
+ "grad_norm": 0.019924819469451904,
+ "learning_rate": 0.0006,
+ "loss": 5.334723472595215,
+ "step": 562
+ },
+ {
+ "epoch": 7.824454148471616,
+ "grad_norm": 0.018755966797471046,
+ "learning_rate": 0.0006,
+ "loss": 5.348781108856201,
+ "step": 563
+ },
+ {
+ "epoch": 7.8384279475982535,
+ "grad_norm": 0.016747845336794853,
+ "learning_rate": 0.0006,
+ "loss": 5.433079719543457,
+ "step": 564
+ },
+ {
+ "epoch": 7.85240174672489,
+ "grad_norm": 0.0170395877212286,
+ "learning_rate": 0.0006,
+ "loss": 5.346794128417969,
+ "step": 565
+ },
+ {
+ "epoch": 7.866375545851528,
+ "grad_norm": 0.016074176877737045,
+ "learning_rate": 0.0006,
+ "loss": 5.3277740478515625,
+ "step": 566
+ },
+ {
+ "epoch": 7.880349344978166,
+ "grad_norm": 0.014119806699454784,
+ "learning_rate": 0.0006,
+ "loss": 5.483551979064941,
+ "step": 567
+ },
+ {
+ "epoch": 7.894323144104804,
+ "grad_norm": 0.015270394273102283,
+ "learning_rate": 0.0006,
+ "loss": 5.378519535064697,
+ "step": 568
+ },
+ {
+ "epoch": 7.908296943231441,
+ "grad_norm": 0.015677539631724358,
+ "learning_rate": 0.0006,
+ "loss": 5.290210247039795,
+ "step": 569
+ },
+ {
+ "epoch": 7.922270742358078,
+ "grad_norm": 0.015930423513054848,
+ "learning_rate": 0.0006,
+ "loss": 5.427360534667969,
+ "step": 570
+ },
+ {
+ "epoch": 7.936244541484716,
+ "grad_norm": 0.016641564667224884,
+ "learning_rate": 0.0006,
+ "loss": 5.301599979400635,
+ "step": 571
+ },
+ {
+ "epoch": 7.950218340611354,
+ "grad_norm": 0.018293552100658417,
+ "learning_rate": 0.0006,
+ "loss": 5.277889251708984,
+ "step": 572
+ },
+ {
+ "epoch": 7.964192139737992,
+ "grad_norm": 0.017618799582123756,
+ "learning_rate": 0.0006,
+ "loss": 5.297194480895996,
+ "step": 573
+ },
+ {
+ "epoch": 7.978165938864628,
+ "grad_norm": 0.015034242533147335,
+ "learning_rate": 0.0006,
+ "loss": 5.42428731918335,
+ "step": 574
+ },
+ {
+ "epoch": 7.992139737991266,
+ "grad_norm": 0.01617511734366417,
+ "learning_rate": 0.0006,
+ "loss": 5.420044898986816,
+ "step": 575
+ },
+ {
+ "epoch": 8.0,
+ "grad_norm": 0.01837257295846939,
+ "learning_rate": 0.0006,
+ "loss": 5.240387916564941,
+ "step": 576
+ },
+ {
+ "epoch": 8.0,
+ "eval_loss": 5.423932075500488,
+ "eval_runtime": 58.0841,
+ "eval_samples_per_second": 42.042,
+ "eval_steps_per_second": 1.326,
+ "step": 576
+ },
+ {
+ "epoch": 8.013973799126637,
+ "grad_norm": 0.020953809842467308,
+ "learning_rate": 0.0006,
+ "loss": 5.4320783615112305,
+ "step": 577
+ },
+ {
+ "epoch": 8.027947598253276,
+ "grad_norm": 0.0227745920419693,
+ "learning_rate": 0.0006,
+ "loss": 5.310197830200195,
+ "step": 578
+ },
+ {
+ "epoch": 8.041921397379912,
+ "grad_norm": 0.021372603252530098,
+ "learning_rate": 0.0006,
+ "loss": 5.339468002319336,
+ "step": 579
+ },
+ {
+ "epoch": 8.055895196506551,
+ "grad_norm": 0.022401731461286545,
+ "learning_rate": 0.0006,
+ "loss": 5.279829978942871,
+ "step": 580
+ },
+ {
+ "epoch": 8.069868995633188,
+ "grad_norm": 0.023942379280924797,
+ "learning_rate": 0.0006,
+ "loss": 5.372323989868164,
+ "step": 581
+ },
+ {
+ "epoch": 8.083842794759825,
+ "grad_norm": 0.024380534887313843,
+ "learning_rate": 0.0006,
+ "loss": 5.253486633300781,
+ "step": 582
+ },
+ {
+ "epoch": 8.097816593886463,
+ "grad_norm": 0.03246625140309334,
+ "learning_rate": 0.0006,
+ "loss": 5.305774211883545,
+ "step": 583
+ },
+ {
+ "epoch": 8.1117903930131,
+ "grad_norm": 0.030399736016988754,
+ "learning_rate": 0.0006,
+ "loss": 5.370976448059082,
+ "step": 584
+ },
+ {
+ "epoch": 8.125764192139737,
+ "grad_norm": 0.027148913592100143,
+ "learning_rate": 0.0006,
+ "loss": 5.368010520935059,
+ "step": 585
+ },
+ {
+ "epoch": 8.139737991266376,
+ "grad_norm": 0.02925540879368782,
+ "learning_rate": 0.0006,
+ "loss": 5.303013801574707,
+ "step": 586
+ },
+ {
+ "epoch": 8.153711790393013,
+ "grad_norm": 0.02915453538298607,
+ "learning_rate": 0.0006,
+ "loss": 5.329667091369629,
+ "step": 587
+ },
+ {
+ "epoch": 8.167685589519651,
+ "grad_norm": 0.031732626259326935,
+ "learning_rate": 0.0006,
+ "loss": 5.3634934425354,
+ "step": 588
+ },
+ {
+ "epoch": 8.181659388646288,
+ "grad_norm": 0.036166731268167496,
+ "learning_rate": 0.0006,
+ "loss": 5.223980903625488,
+ "step": 589
+ },
+ {
+ "epoch": 8.195633187772925,
+ "grad_norm": 0.035654108971357346,
+ "learning_rate": 0.0006,
+ "loss": 5.361059188842773,
+ "step": 590
+ },
+ {
+ "epoch": 8.209606986899564,
+ "grad_norm": 0.0365324430167675,
+ "learning_rate": 0.0006,
+ "loss": 5.336151599884033,
+ "step": 591
+ },
+ {
+ "epoch": 8.2235807860262,
+ "grad_norm": 0.03578880429267883,
+ "learning_rate": 0.0006,
+ "loss": 5.434076309204102,
+ "step": 592
+ },
+ {
+ "epoch": 8.237554585152838,
+ "grad_norm": 0.035734012722969055,
+ "learning_rate": 0.0006,
+ "loss": 5.330893516540527,
+ "step": 593
+ },
+ {
+ "epoch": 8.251528384279476,
+ "grad_norm": 0.03196857124567032,
+ "learning_rate": 0.0006,
+ "loss": 5.306009769439697,
+ "step": 594
+ },
+ {
+ "epoch": 8.265502183406113,
+ "grad_norm": 0.032042086124420166,
+ "learning_rate": 0.0006,
+ "loss": 5.343267917633057,
+ "step": 595
+ },
+ {
+ "epoch": 8.279475982532752,
+ "grad_norm": 0.03160746395587921,
+ "learning_rate": 0.0006,
+ "loss": 5.2353715896606445,
+ "step": 596
+ },
+ {
+ "epoch": 8.293449781659389,
+ "grad_norm": 0.026689818128943443,
+ "learning_rate": 0.0006,
+ "loss": 5.254042625427246,
+ "step": 597
+ },
+ {
+ "epoch": 8.307423580786025,
+ "grad_norm": 0.02880188450217247,
+ "learning_rate": 0.0006,
+ "loss": 5.326833248138428,
+ "step": 598
+ },
+ {
+ "epoch": 8.321397379912664,
+ "grad_norm": 0.027516381815075874,
+ "learning_rate": 0.0006,
+ "loss": 5.342336177825928,
+ "step": 599
+ },
+ {
+ "epoch": 8.335371179039301,
+ "grad_norm": 0.027875030413269997,
+ "learning_rate": 0.0006,
+ "loss": 5.200719833374023,
+ "step": 600
+ },
+ {
+ "epoch": 8.34934497816594,
+ "grad_norm": 0.0268265288323164,
+ "learning_rate": 0.0006,
+ "loss": 5.249449729919434,
+ "step": 601
+ },
+ {
+ "epoch": 8.363318777292577,
+ "grad_norm": 0.024824608117341995,
+ "learning_rate": 0.0006,
+ "loss": 5.318952560424805,
+ "step": 602
+ },
+ {
+ "epoch": 8.377292576419213,
+ "grad_norm": 0.022990627214312553,
+ "learning_rate": 0.0006,
+ "loss": 5.244993209838867,
+ "step": 603
+ },
+ {
+ "epoch": 8.391266375545852,
+ "grad_norm": 0.022804604843258858,
+ "learning_rate": 0.0006,
+ "loss": 5.3236846923828125,
+ "step": 604
+ },
+ {
+ "epoch": 8.405240174672489,
+ "grad_norm": 0.02419872209429741,
+ "learning_rate": 0.0006,
+ "loss": 5.255486488342285,
+ "step": 605
+ },
+ {
+ "epoch": 8.419213973799126,
+ "grad_norm": 0.021952059119939804,
+ "learning_rate": 0.0006,
+ "loss": 5.3511552810668945,
+ "step": 606
+ },
+ {
+ "epoch": 8.433187772925764,
+ "grad_norm": 0.022375497967004776,
+ "learning_rate": 0.0006,
+ "loss": 5.294790267944336,
+ "step": 607
+ },
+ {
+ "epoch": 8.447161572052401,
+ "grad_norm": 0.019242815673351288,
+ "learning_rate": 0.0006,
+ "loss": 5.252618789672852,
+ "step": 608
+ },
+ {
+ "epoch": 8.46113537117904,
+ "grad_norm": 0.0173486340790987,
+ "learning_rate": 0.0006,
+ "loss": 5.30147647857666,
+ "step": 609
+ },
+ {
+ "epoch": 8.475109170305677,
+ "grad_norm": 0.01612604409456253,
+ "learning_rate": 0.0006,
+ "loss": 5.23960542678833,
+ "step": 610
+ },
+ {
+ "epoch": 8.489082969432314,
+ "grad_norm": 0.014293976128101349,
+ "learning_rate": 0.0006,
+ "loss": 5.119932651519775,
+ "step": 611
+ },
+ {
+ "epoch": 8.503056768558952,
+ "grad_norm": 0.014453536830842495,
+ "learning_rate": 0.0006,
+ "loss": 5.166794776916504,
+ "step": 612
+ },
+ {
+ "epoch": 8.51703056768559,
+ "grad_norm": 0.014685769565403461,
+ "learning_rate": 0.0006,
+ "loss": 5.264822959899902,
+ "step": 613
+ },
+ {
+ "epoch": 8.531004366812226,
+ "grad_norm": 0.01427740603685379,
+ "learning_rate": 0.0006,
+ "loss": 5.278433799743652,
+ "step": 614
+ },
+ {
+ "epoch": 8.544978165938865,
+ "grad_norm": 0.014864951372146606,
+ "learning_rate": 0.0006,
+ "loss": 5.214512348175049,
+ "step": 615
+ },
+ {
+ "epoch": 8.558951965065502,
+ "grad_norm": 0.01565164513885975,
+ "learning_rate": 0.0006,
+ "loss": 5.183066368103027,
+ "step": 616
+ },
+ {
+ "epoch": 8.57292576419214,
+ "grad_norm": 0.01779816672205925,
+ "learning_rate": 0.0006,
+ "loss": 5.163185119628906,
+ "step": 617
+ },
+ {
+ "epoch": 8.586899563318777,
+ "grad_norm": 0.017254827544093132,
+ "learning_rate": 0.0006,
+ "loss": 5.243555068969727,
+ "step": 618
+ },
+ {
+ "epoch": 8.600873362445414,
+ "grad_norm": 0.01650584116578102,
+ "learning_rate": 0.0006,
+ "loss": 5.234884738922119,
+ "step": 619
+ },
+ {
+ "epoch": 8.614847161572053,
+ "grad_norm": 0.017021115869283676,
+ "learning_rate": 0.0006,
+ "loss": 5.12271785736084,
+ "step": 620
+ },
+ {
+ "epoch": 8.62882096069869,
+ "grad_norm": 0.01773759163916111,
+ "learning_rate": 0.0006,
+ "loss": 5.242153167724609,
+ "step": 621
+ },
+ {
+ "epoch": 8.642794759825328,
+ "grad_norm": 0.015679042786359787,
+ "learning_rate": 0.0006,
+ "loss": 5.170779228210449,
+ "step": 622
+ },
+ {
+ "epoch": 8.656768558951965,
+ "grad_norm": 0.013760549947619438,
+ "learning_rate": 0.0006,
+ "loss": 5.238644599914551,
+ "step": 623
+ },
+ {
+ "epoch": 8.670742358078602,
+ "grad_norm": 0.014571522362530231,
+ "learning_rate": 0.0006,
+ "loss": 5.177056312561035,
+ "step": 624
+ },
+ {
+ "epoch": 8.68471615720524,
+ "grad_norm": 0.016209015622735023,
+ "learning_rate": 0.0006,
+ "loss": 5.163750648498535,
+ "step": 625
+ },
+ {
+ "epoch": 8.698689956331878,
+ "grad_norm": 0.016813941299915314,
+ "learning_rate": 0.0006,
+ "loss": 5.183428764343262,
+ "step": 626
+ },
+ {
+ "epoch": 8.712663755458514,
+ "grad_norm": 0.020985357463359833,
+ "learning_rate": 0.0006,
+ "loss": 5.229465007781982,
+ "step": 627
+ },
+ {
+ "epoch": 8.726637554585153,
+ "grad_norm": 0.02679632417857647,
+ "learning_rate": 0.0006,
+ "loss": 5.257368087768555,
+ "step": 628
+ },
+ {
+ "epoch": 8.74061135371179,
+ "grad_norm": 0.025756070390343666,
+ "learning_rate": 0.0006,
+ "loss": 5.253736972808838,
+ "step": 629
+ },
+ {
+ "epoch": 8.754585152838429,
+ "grad_norm": 0.02643490768969059,
+ "learning_rate": 0.0006,
+ "loss": 5.228633403778076,
+ "step": 630
+ },
+ {
+ "epoch": 8.768558951965066,
+ "grad_norm": 0.03303210437297821,
+ "learning_rate": 0.0006,
+ "loss": 5.163540363311768,
+ "step": 631
+ },
+ {
+ "epoch": 8.782532751091702,
+ "grad_norm": 0.030432431027293205,
+ "learning_rate": 0.0006,
+ "loss": 5.243169784545898,
+ "step": 632
+ },
+ {
+ "epoch": 8.796506550218341,
+ "grad_norm": 0.028631288558244705,
+ "learning_rate": 0.0006,
+ "loss": 5.166398048400879,
+ "step": 633
+ },
+ {
+ "epoch": 8.810480349344978,
+ "grad_norm": 0.026188310235738754,
+ "learning_rate": 0.0006,
+ "loss": 5.128122329711914,
+ "step": 634
+ },
+ {
+ "epoch": 8.824454148471617,
+ "grad_norm": 0.025308528915047646,
+ "learning_rate": 0.0006,
+ "loss": 5.211284637451172,
+ "step": 635
+ },
+ {
+ "epoch": 8.838427947598253,
+ "grad_norm": 0.020657729357481003,
+ "learning_rate": 0.0006,
+ "loss": 5.161575794219971,
+ "step": 636
+ },
+ {
+ "epoch": 8.85240174672489,
+ "grad_norm": 0.021901234984397888,
+ "learning_rate": 0.0006,
+ "loss": 5.20050048828125,
+ "step": 637
+ },
+ {
+ "epoch": 8.866375545851529,
+ "grad_norm": 0.021219318732619286,
+ "learning_rate": 0.0006,
+ "loss": 5.211699485778809,
+ "step": 638
+ },
+ {
+ "epoch": 8.880349344978166,
+ "grad_norm": 0.0200974028557539,
+ "learning_rate": 0.0006,
+ "loss": 5.092503547668457,
+ "step": 639
+ },
+ {
+ "epoch": 8.894323144104803,
+ "grad_norm": 0.023804882541298866,
+ "learning_rate": 0.0006,
+ "loss": 5.216068267822266,
+ "step": 640
+ },
+ {
+ "epoch": 8.908296943231441,
+ "grad_norm": 0.026088010519742966,
+ "learning_rate": 0.0006,
+ "loss": 5.155592918395996,
+ "step": 641
+ },
+ {
+ "epoch": 8.922270742358078,
+ "grad_norm": 0.02601276896893978,
+ "learning_rate": 0.0006,
+ "loss": 5.197238922119141,
+ "step": 642
+ },
+ {
+ "epoch": 8.936244541484717,
+ "grad_norm": 0.020387211814522743,
+ "learning_rate": 0.0006,
+ "loss": 5.1955437660217285,
+ "step": 643
+ },
+ {
+ "epoch": 8.950218340611354,
+ "grad_norm": 0.019214622676372528,
+ "learning_rate": 0.0006,
+ "loss": 5.206346035003662,
+ "step": 644
+ },
+ {
+ "epoch": 8.96419213973799,
+ "grad_norm": 0.019674314185976982,
+ "learning_rate": 0.0006,
+ "loss": 5.219857215881348,
+ "step": 645
+ },
+ {
+ "epoch": 8.97816593886463,
+ "grad_norm": 0.020318234339356422,
+ "learning_rate": 0.0006,
+ "loss": 5.191132545471191,
+ "step": 646
+ },
+ {
+ "epoch": 8.992139737991266,
+ "grad_norm": 0.021428676322102547,
+ "learning_rate": 0.0006,
+ "loss": 5.214695930480957,
+ "step": 647
+ },
+ {
+ "epoch": 9.0,
+ "grad_norm": 0.023235054686665535,
+ "learning_rate": 0.0006,
+ "loss": 5.229091644287109,
+ "step": 648
+ },
+ {
+ "epoch": 9.0,
+ "eval_loss": 5.289593696594238,
+ "eval_runtime": 56.9256,
+ "eval_samples_per_second": 42.898,
+ "eval_steps_per_second": 1.353,
+ "step": 648
+ },
+ {
+ "epoch": 9.013973799126637,
+ "grad_norm": 0.02320289984345436,
+ "learning_rate": 0.0006,
+ "loss": 4.983669281005859,
+ "step": 649
+ },
+ {
+ "epoch": 9.027947598253276,
+ "grad_norm": 0.018521282821893692,
+ "learning_rate": 0.0006,
+ "loss": 5.200575828552246,
+ "step": 650
+ },
+ {
+ "epoch": 9.041921397379912,
+ "grad_norm": 0.0180169939994812,
+ "learning_rate": 0.0006,
+ "loss": 5.148033618927002,
+ "step": 651
+ },
+ {
+ "epoch": 9.055895196506551,
+ "grad_norm": 0.019969860091805458,
+ "learning_rate": 0.0006,
+ "loss": 5.099125385284424,
+ "step": 652
+ },
+ {
+ "epoch": 9.069868995633188,
+ "grad_norm": 0.01643305830657482,
+ "learning_rate": 0.0006,
+ "loss": 5.092347145080566,
+ "step": 653
+ },
+ {
+ "epoch": 9.083842794759825,
+ "grad_norm": 0.016022363677620888,
+ "learning_rate": 0.0006,
+ "loss": 5.129279613494873,
+ "step": 654
+ },
+ {
+ "epoch": 9.097816593886463,
+ "grad_norm": 0.01657041162252426,
+ "learning_rate": 0.0006,
+ "loss": 5.159562587738037,
+ "step": 655
+ },
+ {
+ "epoch": 9.1117903930131,
+ "grad_norm": 0.017924228683114052,
+ "learning_rate": 0.0006,
+ "loss": 5.165497779846191,
+ "step": 656
+ },
+ {
+ "epoch": 9.125764192139737,
+ "grad_norm": 0.021592361852526665,
+ "learning_rate": 0.0006,
+ "loss": 5.204775810241699,
+ "step": 657
+ },
+ {
+ "epoch": 9.139737991266376,
+ "grad_norm": 0.026924120262265205,
+ "learning_rate": 0.0006,
+ "loss": 5.128296852111816,
+ "step": 658
+ },
+ {
+ "epoch": 9.153711790393013,
+ "grad_norm": 0.023156899958848953,
+ "learning_rate": 0.0006,
+ "loss": 5.1255340576171875,
+ "step": 659
+ },
+ {
+ "epoch": 9.167685589519651,
+ "grad_norm": 0.020670367404818535,
+ "learning_rate": 0.0006,
+ "loss": 5.128479957580566,
+ "step": 660
+ },
+ {
+ "epoch": 9.181659388646288,
+ "grad_norm": 0.025447173044085503,
+ "learning_rate": 0.0006,
+ "loss": 5.154821872711182,
+ "step": 661
+ },
+ {
+ "epoch": 9.195633187772925,
+ "grad_norm": 0.027940068393945694,
+ "learning_rate": 0.0006,
+ "loss": 5.143466472625732,
+ "step": 662
+ },
+ {
+ "epoch": 9.209606986899564,
+ "grad_norm": 0.026847844943404198,
+ "learning_rate": 0.0006,
+ "loss": 5.045665740966797,
+ "step": 663
+ },
+ {
+ "epoch": 9.2235807860262,
+ "grad_norm": 0.02340601570904255,
+ "learning_rate": 0.0006,
+ "loss": 5.21933126449585,
+ "step": 664
+ },
+ {
+ "epoch": 9.237554585152838,
+ "grad_norm": 0.02340371161699295,
+ "learning_rate": 0.0006,
+ "loss": 4.942949295043945,
+ "step": 665
+ },
+ {
+ "epoch": 9.251528384279476,
+ "grad_norm": 0.02221992425620556,
+ "learning_rate": 0.0006,
+ "loss": 5.197099208831787,
+ "step": 666
+ },
+ {
+ "epoch": 9.265502183406113,
+ "grad_norm": 0.023508677259087563,
+ "learning_rate": 0.0006,
+ "loss": 5.022897720336914,
+ "step": 667
+ },
+ {
+ "epoch": 9.279475982532752,
+ "grad_norm": 0.026120394468307495,
+ "learning_rate": 0.0006,
+ "loss": 5.087725639343262,
+ "step": 668
+ },
+ {
+ "epoch": 9.293449781659389,
+ "grad_norm": 0.027273228392004967,
+ "learning_rate": 0.0006,
+ "loss": 5.090963840484619,
+ "step": 669
+ },
+ {
+ "epoch": 9.307423580786025,
+ "grad_norm": 0.03241586685180664,
+ "learning_rate": 0.0006,
+ "loss": 5.047842979431152,
+ "step": 670
+ },
+ {
+ "epoch": 9.321397379912664,
+ "grad_norm": 0.030245667323470116,
+ "learning_rate": 0.0006,
+ "loss": 5.104803085327148,
+ "step": 671
+ },
+ {
+ "epoch": 9.335371179039301,
+ "grad_norm": 0.027698364108800888,
+ "learning_rate": 0.0006,
+ "loss": 5.0644097328186035,
+ "step": 672
+ },
+ {
+ "epoch": 9.34934497816594,
+ "grad_norm": 0.029692554846405983,
+ "learning_rate": 0.0006,
+ "loss": 5.111942291259766,
+ "step": 673
+ },
+ {
+ "epoch": 9.363318777292577,
+ "grad_norm": 0.03328656405210495,
+ "learning_rate": 0.0006,
+ "loss": 5.1574554443359375,
+ "step": 674
+ },
+ {
+ "epoch": 9.377292576419213,
+ "grad_norm": 0.031938180327415466,
+ "learning_rate": 0.0006,
+ "loss": 5.190371990203857,
+ "step": 675
+ },
+ {
+ "epoch": 9.391266375545852,
+ "grad_norm": 0.024858945980668068,
+ "learning_rate": 0.0006,
+ "loss": 5.121420860290527,
+ "step": 676
+ },
+ {
+ "epoch": 9.405240174672489,
+ "grad_norm": 0.023033203557133675,
+ "learning_rate": 0.0006,
+ "loss": 5.0974345207214355,
+ "step": 677
+ },
+ {
+ "epoch": 9.419213973799126,
+ "grad_norm": 0.022083545103669167,
+ "learning_rate": 0.0006,
+ "loss": 5.046319961547852,
+ "step": 678
+ },
+ {
+ "epoch": 9.433187772925764,
+ "grad_norm": 0.0228437427431345,
+ "learning_rate": 0.0006,
+ "loss": 5.119963645935059,
+ "step": 679
+ },
+ {
+ "epoch": 9.447161572052401,
+ "grad_norm": 0.0242351982742548,
+ "learning_rate": 0.0006,
+ "loss": 5.0444560050964355,
+ "step": 680
+ },
+ {
+ "epoch": 9.46113537117904,
+ "grad_norm": 0.02401185780763626,
+ "learning_rate": 0.0006,
+ "loss": 5.063238620758057,
+ "step": 681
+ },
+ {
+ "epoch": 9.475109170305677,
+ "grad_norm": 0.02560959942638874,
+ "learning_rate": 0.0006,
+ "loss": 5.035224914550781,
+ "step": 682
+ },
+ {
+ "epoch": 9.489082969432314,
+ "grad_norm": 0.024985626339912415,
+ "learning_rate": 0.0006,
+ "loss": 5.049284934997559,
+ "step": 683
+ },
+ {
+ "epoch": 9.503056768558952,
+ "grad_norm": 0.023178115487098694,
+ "learning_rate": 0.0006,
+ "loss": 5.2266621589660645,
+ "step": 684
+ },
+ {
+ "epoch": 9.51703056768559,
+ "grad_norm": 0.0215512253344059,
+ "learning_rate": 0.0006,
+ "loss": 5.1875715255737305,
+ "step": 685
+ },
+ {
+ "epoch": 9.531004366812226,
+ "grad_norm": 0.01662837713956833,
+ "learning_rate": 0.0006,
+ "loss": 5.137012481689453,
+ "step": 686
+ },
+ {
+ "epoch": 9.544978165938865,
+ "grad_norm": 0.01952764391899109,
+ "learning_rate": 0.0006,
+ "loss": 5.072500228881836,
+ "step": 687
+ },
+ {
+ "epoch": 9.558951965065502,
+ "grad_norm": 0.017521077767014503,
+ "learning_rate": 0.0006,
+ "loss": 5.045711994171143,
+ "step": 688
+ },
+ {
+ "epoch": 9.57292576419214,
+ "grad_norm": 0.017355265095829964,
+ "learning_rate": 0.0006,
+ "loss": 5.02881383895874,
+ "step": 689
+ },
+ {
+ "epoch": 9.586899563318777,
+ "grad_norm": 0.015587719157338142,
+ "learning_rate": 0.0006,
+ "loss": 5.132787704467773,
+ "step": 690
+ },
+ {
+ "epoch": 9.600873362445414,
+ "grad_norm": 0.017469845712184906,
+ "learning_rate": 0.0006,
+ "loss": 5.105447292327881,
+ "step": 691
+ },
+ {
+ "epoch": 9.614847161572053,
+ "grad_norm": 0.019630100578069687,
+ "learning_rate": 0.0006,
+ "loss": 5.032464981079102,
+ "step": 692
+ },
+ {
+ "epoch": 9.62882096069869,
+ "grad_norm": 0.01784994639456272,
+ "learning_rate": 0.0006,
+ "loss": 5.081345558166504,
+ "step": 693
+ },
+ {
+ "epoch": 9.642794759825328,
+ "grad_norm": 0.01872754842042923,
+ "learning_rate": 0.0006,
+ "loss": 4.9772114753723145,
+ "step": 694
+ },
+ {
+ "epoch": 9.656768558951965,
+ "grad_norm": 0.01999032311141491,
+ "learning_rate": 0.0006,
+ "loss": 5.089740753173828,
+ "step": 695
+ },
+ {
+ "epoch": 9.670742358078602,
+ "grad_norm": 0.01710767112672329,
+ "learning_rate": 0.0006,
+ "loss": 4.9869890213012695,
+ "step": 696
+ },
+ {
+ "epoch": 9.68471615720524,
+ "grad_norm": 0.015144161880016327,
+ "learning_rate": 0.0006,
+ "loss": 5.022564888000488,
+ "step": 697
+ },
+ {
+ "epoch": 9.698689956331878,
+ "grad_norm": 0.01618543453514576,
+ "learning_rate": 0.0006,
+ "loss": 5.035921573638916,
+ "step": 698
+ },
+ {
+ "epoch": 9.712663755458514,
+ "grad_norm": 0.014653387479484081,
+ "learning_rate": 0.0006,
+ "loss": 4.914456367492676,
+ "step": 699
+ },
+ {
+ "epoch": 9.726637554585153,
+ "grad_norm": 0.013434977270662785,
+ "learning_rate": 0.0006,
+ "loss": 5.063764572143555,
+ "step": 700
+ },
+ {
+ "epoch": 9.74061135371179,
+ "grad_norm": 0.015128864906728268,
+ "learning_rate": 0.0006,
+ "loss": 5.052367687225342,
+ "step": 701
+ },
+ {
+ "epoch": 9.754585152838429,
+ "grad_norm": 0.014210768043994904,
+ "learning_rate": 0.0006,
+ "loss": 5.016489028930664,
+ "step": 702
+ },
+ {
+ "epoch": 9.768558951965066,
+ "grad_norm": 0.015191680751740932,
+ "learning_rate": 0.0006,
+ "loss": 4.990176200866699,
+ "step": 703
+ },
+ {
+ "epoch": 9.782532751091702,
+ "grad_norm": 0.018285127356648445,
+ "learning_rate": 0.0006,
+ "loss": 5.015164375305176,
+ "step": 704
+ },
+ {
+ "epoch": 9.796506550218341,
+ "grad_norm": 0.0204079058021307,
+ "learning_rate": 0.0006,
+ "loss": 5.091142654418945,
+ "step": 705
+ },
+ {
+ "epoch": 9.810480349344978,
+ "grad_norm": 0.01980586163699627,
+ "learning_rate": 0.0006,
+ "loss": 5.03350830078125,
+ "step": 706
+ },
+ {
+ "epoch": 9.824454148471617,
+ "grad_norm": 0.02048729732632637,
+ "learning_rate": 0.0006,
+ "loss": 5.03611421585083,
+ "step": 707
+ },
+ {
+ "epoch": 9.838427947598253,
+ "grad_norm": 0.02058163844048977,
+ "learning_rate": 0.0006,
+ "loss": 4.9613447189331055,
+ "step": 708
+ },
+ {
+ "epoch": 9.85240174672489,
+ "grad_norm": 0.018004924058914185,
+ "learning_rate": 0.0006,
+ "loss": 4.966541767120361,
+ "step": 709
+ },
+ {
+ "epoch": 9.866375545851529,
+ "grad_norm": 0.0198152307420969,
+ "learning_rate": 0.0006,
+ "loss": 5.01750373840332,
+ "step": 710
+ },
+ {
+ "epoch": 9.880349344978166,
+ "grad_norm": 0.018532969057559967,
+ "learning_rate": 0.0006,
+ "loss": 5.097424507141113,
+ "step": 711
+ },
+ {
+ "epoch": 9.894323144104803,
+ "grad_norm": 0.018939971923828125,
+ "learning_rate": 0.0006,
+ "loss": 5.177223205566406,
+ "step": 712
+ },
+ {
+ "epoch": 9.908296943231441,
+ "grad_norm": 0.017969170585274696,
+ "learning_rate": 0.0006,
+ "loss": 5.118013858795166,
+ "step": 713
+ },
+ {
+ "epoch": 9.922270742358078,
+ "grad_norm": 0.018021270632743835,
+ "learning_rate": 0.0006,
+ "loss": 5.034787178039551,
+ "step": 714
+ },
+ {
+ "epoch": 9.936244541484717,
+ "grad_norm": 0.017009710893034935,
+ "learning_rate": 0.0006,
+ "loss": 5.003920555114746,
+ "step": 715
+ },
+ {
+ "epoch": 9.950218340611354,
+ "grad_norm": 0.017722049728035927,
+ "learning_rate": 0.0006,
+ "loss": 5.031939506530762,
+ "step": 716
+ },
+ {
+ "epoch": 9.96419213973799,
+ "grad_norm": 0.017134087160229683,
+ "learning_rate": 0.0006,
+ "loss": 5.093474388122559,
+ "step": 717
+ },
+ {
+ "epoch": 9.97816593886463,
+ "grad_norm": 0.018511613830924034,
+ "learning_rate": 0.0006,
+ "loss": 5.078604698181152,
+ "step": 718
+ },
+ {
+ "epoch": 9.992139737991266,
+ "grad_norm": 0.021206334233283997,
+ "learning_rate": 0.0006,
+ "loss": 5.052572727203369,
+ "step": 719
+ },
+ {
+ "epoch": 10.0,
+ "grad_norm": 0.02239062264561653,
+ "learning_rate": 0.0006,
+ "loss": 5.063547134399414,
+ "step": 720
+ },
+ {
+ "epoch": 10.0,
+ "eval_loss": 5.173044681549072,
+ "eval_runtime": 56.7287,
+ "eval_samples_per_second": 43.047,
+ "eval_steps_per_second": 1.357,
+ "step": 720
+ },
+ {
+ "epoch": 10.013973799126637,
+ "grad_norm": 0.020757969468832016,
+ "learning_rate": 0.0006,
+ "loss": 5.045648574829102,
+ "step": 721
+ },
+ {
+ "epoch": 10.027947598253276,
+ "grad_norm": 0.022697702050209045,
+ "learning_rate": 0.0006,
+ "loss": 4.993833065032959,
+ "step": 722
+ },
+ {
+ "epoch": 10.041921397379912,
+ "grad_norm": 0.02434331737458706,
+ "learning_rate": 0.0006,
+ "loss": 5.073219299316406,
+ "step": 723
+ },
+ {
+ "epoch": 10.055895196506551,
+ "grad_norm": 0.025499660521745682,
+ "learning_rate": 0.0006,
+ "loss": 5.019072532653809,
+ "step": 724
+ },
+ {
+ "epoch": 10.069868995633188,
+ "grad_norm": 0.029281053692102432,
+ "learning_rate": 0.0006,
+ "loss": 4.959851264953613,
+ "step": 725
+ },
+ {
+ "epoch": 10.083842794759825,
+ "grad_norm": 0.033878110349178314,
+ "learning_rate": 0.0006,
+ "loss": 5.132717609405518,
+ "step": 726
+ },
+ {
+ "epoch": 10.097816593886463,
+ "grad_norm": 0.03385764732956886,
+ "learning_rate": 0.0006,
+ "loss": 5.0482892990112305,
+ "step": 727
+ },
+ {
+ "epoch": 10.1117903930131,
+ "grad_norm": 0.03466648980975151,
+ "learning_rate": 0.0006,
+ "loss": 4.946855545043945,
+ "step": 728
+ },
+ {
+ "epoch": 10.125764192139737,
+ "grad_norm": 0.028546305373311043,
+ "learning_rate": 0.0006,
+ "loss": 5.027774333953857,
+ "step": 729
+ },
+ {
+ "epoch": 10.139737991266376,
+ "grad_norm": 0.030575869604945183,
+ "learning_rate": 0.0006,
+ "loss": 4.9768476486206055,
+ "step": 730
+ },
+ {
+ "epoch": 10.153711790393013,
+ "grad_norm": 0.03614303469657898,
+ "learning_rate": 0.0006,
+ "loss": 5.078113555908203,
+ "step": 731
+ },
+ {
+ "epoch": 10.167685589519651,
+ "grad_norm": 0.040991660207509995,
+ "learning_rate": 0.0006,
+ "loss": 5.033185958862305,
+ "step": 732
+ },
+ {
+ "epoch": 10.181659388646288,
+ "grad_norm": 0.04551699757575989,
+ "learning_rate": 0.0006,
+ "loss": 5.0820746421813965,
+ "step": 733
+ },
+ {
+ "epoch": 10.195633187772925,
+ "grad_norm": 0.040193989872932434,
+ "learning_rate": 0.0006,
+ "loss": 5.059700965881348,
+ "step": 734
+ },
+ {
+ "epoch": 10.209606986899564,
+ "grad_norm": 0.035851605236530304,
+ "learning_rate": 0.0006,
+ "loss": 5.142467498779297,
+ "step": 735
+ },
+ {
+ "epoch": 10.2235807860262,
+ "grad_norm": 0.034200169146060944,
+ "learning_rate": 0.0006,
+ "loss": 5.072312355041504,
+ "step": 736
+ },
+ {
+ "epoch": 10.237554585152838,
+ "grad_norm": 0.033621110022068024,
+ "learning_rate": 0.0006,
+ "loss": 5.08037805557251,
+ "step": 737
+ },
+ {
+ "epoch": 10.251528384279476,
+ "grad_norm": 0.037372369319200516,
+ "learning_rate": 0.0006,
+ "loss": 4.998425483703613,
+ "step": 738
+ },
+ {
+ "epoch": 10.265502183406113,
+ "grad_norm": 0.037044707685709,
+ "learning_rate": 0.0006,
+ "loss": 4.923993110656738,
+ "step": 739
+ },
+ {
+ "epoch": 10.279475982532752,
+ "grad_norm": 0.030454808846116066,
+ "learning_rate": 0.0006,
+ "loss": 5.0889997482299805,
+ "step": 740
+ },
+ {
+ "epoch": 10.293449781659389,
+ "grad_norm": 0.030969638377428055,
+ "learning_rate": 0.0006,
+ "loss": 4.984112739562988,
+ "step": 741
+ },
+ {
+ "epoch": 10.307423580786025,
+ "grad_norm": 0.0276983380317688,
+ "learning_rate": 0.0006,
+ "loss": 4.955412864685059,
+ "step": 742
+ },
+ {
+ "epoch": 10.321397379912664,
+ "grad_norm": 0.02716052532196045,
+ "learning_rate": 0.0006,
+ "loss": 5.0098876953125,
+ "step": 743
+ },
+ {
+ "epoch": 10.335371179039301,
+ "grad_norm": 0.02403552085161209,
+ "learning_rate": 0.0006,
+ "loss": 5.165160655975342,
+ "step": 744
+ },
+ {
+ "epoch": 10.34934497816594,
+ "grad_norm": 0.0257862601429224,
+ "learning_rate": 0.0006,
+ "loss": 5.001547813415527,
+ "step": 745
+ },
+ {
+ "epoch": 10.363318777292577,
+ "grad_norm": 0.02300378680229187,
+ "learning_rate": 0.0006,
+ "loss": 5.038888931274414,
+ "step": 746
+ },
+ {
+ "epoch": 10.377292576419213,
+ "grad_norm": 0.019766854122281075,
+ "learning_rate": 0.0006,
+ "loss": 4.924291610717773,
+ "step": 747
+ },
+ {
+ "epoch": 10.391266375545852,
+ "grad_norm": 0.01888395845890045,
+ "learning_rate": 0.0006,
+ "loss": 5.059971332550049,
+ "step": 748
+ },
+ {
+ "epoch": 10.405240174672489,
+ "grad_norm": 0.015577499754726887,
+ "learning_rate": 0.0006,
+ "loss": 4.987686634063721,
+ "step": 749
+ },
+ {
+ "epoch": 10.419213973799126,
+ "grad_norm": 0.016012346372008324,
+ "learning_rate": 0.0006,
+ "loss": 5.015157699584961,
+ "step": 750
+ },
+ {
+ "epoch": 10.433187772925764,
+ "grad_norm": 0.016277998685836792,
+ "learning_rate": 0.0006,
+ "loss": 5.004931449890137,
+ "step": 751
+ },
+ {
+ "epoch": 10.447161572052401,
+ "grad_norm": 0.017504561692476273,
+ "learning_rate": 0.0006,
+ "loss": 4.902827739715576,
+ "step": 752
+ },
+ {
+ "epoch": 10.46113537117904,
+ "grad_norm": 0.01569022797048092,
+ "learning_rate": 0.0006,
+ "loss": 4.954863548278809,
+ "step": 753
+ },
+ {
+ "epoch": 10.475109170305677,
+ "grad_norm": 0.014114422723650932,
+ "learning_rate": 0.0006,
+ "loss": 5.00656795501709,
+ "step": 754
+ },
+ {
+ "epoch": 10.489082969432314,
+ "grad_norm": 0.01573832333087921,
+ "learning_rate": 0.0006,
+ "loss": 4.978353023529053,
+ "step": 755
+ },
+ {
+ "epoch": 10.503056768558952,
+ "grad_norm": 0.016480036079883575,
+ "learning_rate": 0.0006,
+ "loss": 4.953327655792236,
+ "step": 756
+ },
+ {
+ "epoch": 10.51703056768559,
+ "grad_norm": 0.014008025638759136,
+ "learning_rate": 0.0006,
+ "loss": 5.050840377807617,
+ "step": 757
+ },
+ {
+ "epoch": 10.531004366812226,
+ "grad_norm": 0.013147998601198196,
+ "learning_rate": 0.0006,
+ "loss": 4.974964141845703,
+ "step": 758
+ },
+ {
+ "epoch": 10.544978165938865,
+ "grad_norm": 0.013807603158056736,
+ "learning_rate": 0.0006,
+ "loss": 4.927907943725586,
+ "step": 759
+ },
+ {
+ "epoch": 10.558951965065502,
+ "grad_norm": 0.013555224984884262,
+ "learning_rate": 0.0006,
+ "loss": 4.980690956115723,
+ "step": 760
+ },
+ {
+ "epoch": 10.57292576419214,
+ "grad_norm": 0.013045408762991428,
+ "learning_rate": 0.0006,
+ "loss": 4.9436116218566895,
+ "step": 761
+ },
+ {
+ "epoch": 10.586899563318777,
+ "grad_norm": 0.012858862057328224,
+ "learning_rate": 0.0006,
+ "loss": 4.9595770835876465,
+ "step": 762
+ },
+ {
+ "epoch": 10.600873362445414,
+ "grad_norm": 0.01468253880739212,
+ "learning_rate": 0.0006,
+ "loss": 4.90725040435791,
+ "step": 763
+ },
+ {
+ "epoch": 10.614847161572053,
+ "grad_norm": 0.013224679045379162,
+ "learning_rate": 0.0006,
+ "loss": 4.877615928649902,
+ "step": 764
+ },
+ {
+ "epoch": 10.62882096069869,
+ "grad_norm": 0.013555348850786686,
+ "learning_rate": 0.0006,
+ "loss": 4.9456610679626465,
+ "step": 765
+ },
+ {
+ "epoch": 10.642794759825328,
+ "grad_norm": 0.014387160539627075,
+ "learning_rate": 0.0006,
+ "loss": 4.9233717918396,
+ "step": 766
+ },
+ {
+ "epoch": 10.656768558951965,
+ "grad_norm": 0.014354088343679905,
+ "learning_rate": 0.0006,
+ "loss": 4.983511924743652,
+ "step": 767
+ },
+ {
+ "epoch": 10.670742358078602,
+ "grad_norm": 0.01332700252532959,
+ "learning_rate": 0.0006,
+ "loss": 4.934416770935059,
+ "step": 768
+ },
+ {
+ "epoch": 10.68471615720524,
+ "grad_norm": 0.011543313041329384,
+ "learning_rate": 0.0006,
+ "loss": 4.96388053894043,
+ "step": 769
+ },
+ {
+ "epoch": 10.698689956331878,
+ "grad_norm": 0.012330878525972366,
+ "learning_rate": 0.0006,
+ "loss": 4.946396350860596,
+ "step": 770
+ },
+ {
+ "epoch": 10.712663755458514,
+ "grad_norm": 0.01379779726266861,
+ "learning_rate": 0.0006,
+ "loss": 4.981019020080566,
+ "step": 771
+ },
+ {
+ "epoch": 10.726637554585153,
+ "grad_norm": 0.014612431637942791,
+ "learning_rate": 0.0006,
+ "loss": 4.894567966461182,
+ "step": 772
+ },
+ {
+ "epoch": 10.74061135371179,
+ "grad_norm": 0.013062899932265282,
+ "learning_rate": 0.0006,
+ "loss": 5.033038139343262,
+ "step": 773
+ },
+ {
+ "epoch": 10.754585152838429,
+ "grad_norm": 0.013898049481213093,
+ "learning_rate": 0.0006,
+ "loss": 4.823636054992676,
+ "step": 774
+ },
+ {
+ "epoch": 10.768558951965066,
+ "grad_norm": 0.01502019353210926,
+ "learning_rate": 0.0006,
+ "loss": 5.020156383514404,
+ "step": 775
+ },
+ {
+ "epoch": 10.782532751091702,
+ "grad_norm": 0.013810204342007637,
+ "learning_rate": 0.0006,
+ "loss": 4.935647010803223,
+ "step": 776
+ },
+ {
+ "epoch": 10.796506550218341,
+ "grad_norm": 0.013616513460874557,
+ "learning_rate": 0.0006,
+ "loss": 4.901486396789551,
+ "step": 777
+ },
+ {
+ "epoch": 10.810480349344978,
+ "grad_norm": 0.013601024635136127,
+ "learning_rate": 0.0006,
+ "loss": 4.931873321533203,
+ "step": 778
+ },
+ {
+ "epoch": 10.824454148471617,
+ "grad_norm": 0.013668350875377655,
+ "learning_rate": 0.0006,
+ "loss": 4.937256813049316,
+ "step": 779
+ },
+ {
+ "epoch": 10.838427947598253,
+ "grad_norm": 0.01581823080778122,
+ "learning_rate": 0.0006,
+ "loss": 4.9611406326293945,
+ "step": 780
+ },
+ {
+ "epoch": 10.85240174672489,
+ "grad_norm": 0.019478484988212585,
+ "learning_rate": 0.0006,
+ "loss": 4.9212236404418945,
+ "step": 781
+ },
+ {
+ "epoch": 10.866375545851529,
+ "grad_norm": 0.026921333745121956,
+ "learning_rate": 0.0006,
+ "loss": 4.88501501083374,
+ "step": 782
+ },
+ {
+ "epoch": 10.880349344978166,
+ "grad_norm": 0.03155914694070816,
+ "learning_rate": 0.0006,
+ "loss": 4.989967346191406,
+ "step": 783
+ },
+ {
+ "epoch": 10.894323144104803,
+ "grad_norm": 0.029689429327845573,
+ "learning_rate": 0.0006,
+ "loss": 4.936000823974609,
+ "step": 784
+ },
+ {
+ "epoch": 10.908296943231441,
+ "grad_norm": 0.026626063510775566,
+ "learning_rate": 0.0006,
+ "loss": 4.984785079956055,
+ "step": 785
+ },
+ {
+ "epoch": 10.922270742358078,
+ "grad_norm": 0.026168212294578552,
+ "learning_rate": 0.0006,
+ "loss": 4.908282279968262,
+ "step": 786
+ },
+ {
+ "epoch": 10.936244541484717,
+ "grad_norm": 0.02828158251941204,
+ "learning_rate": 0.0006,
+ "loss": 4.926390647888184,
+ "step": 787
+ },
+ {
+ "epoch": 10.950218340611354,
+ "grad_norm": 0.02649078331887722,
+ "learning_rate": 0.0006,
+ "loss": 5.00400972366333,
+ "step": 788
+ },
+ {
+ "epoch": 10.96419213973799,
+ "grad_norm": 0.0256856270134449,
+ "learning_rate": 0.0006,
+ "loss": 5.0317487716674805,
+ "step": 789
+ },
+ {
+ "epoch": 10.97816593886463,
+ "grad_norm": 0.024728883057832718,
+ "learning_rate": 0.0006,
+ "loss": 4.9299635887146,
+ "step": 790
+ },
+ {
+ "epoch": 10.992139737991266,
+ "grad_norm": 0.026241116225719452,
+ "learning_rate": 0.0006,
+ "loss": 4.947319984436035,
+ "step": 791
+ },
+ {
+ "epoch": 11.0,
+ "grad_norm": 0.023526743054389954,
+ "learning_rate": 0.0006,
+ "loss": 4.94429349899292,
+ "step": 792
+ },
+ {
+ "epoch": 11.0,
+ "eval_loss": 5.104345798492432,
+ "eval_runtime": 57.0955,
+ "eval_samples_per_second": 42.77,
+ "eval_steps_per_second": 1.349,
+ "step": 792
+ },
+ {
+ "epoch": 11.013973799126637,
+ "grad_norm": 0.0216788612306118,
+ "learning_rate": 0.0006,
+ "loss": 4.881950855255127,
+ "step": 793
+ },
+ {
+ "epoch": 11.027947598253276,
+ "grad_norm": 0.01830463856458664,
+ "learning_rate": 0.0006,
+ "loss": 4.971016883850098,
+ "step": 794
+ },
+ {
+ "epoch": 11.041921397379912,
+ "grad_norm": 0.02204521745443344,
+ "learning_rate": 0.0006,
+ "loss": 4.894623756408691,
+ "step": 795
+ },
+ {
+ "epoch": 11.055895196506551,
+ "grad_norm": 0.0210255216807127,
+ "learning_rate": 0.0006,
+ "loss": 4.834009170532227,
+ "step": 796
+ },
+ {
+ "epoch": 11.069868995633188,
+ "grad_norm": 0.019797448068857193,
+ "learning_rate": 0.0006,
+ "loss": 5.086714744567871,
+ "step": 797
+ },
+ {
+ "epoch": 11.083842794759825,
+ "grad_norm": 0.02036641724407673,
+ "learning_rate": 0.0006,
+ "loss": 4.920056343078613,
+ "step": 798
+ },
+ {
+ "epoch": 11.097816593886463,
+ "grad_norm": 0.01966066285967827,
+ "learning_rate": 0.0006,
+ "loss": 4.879859447479248,
+ "step": 799
+ },
+ {
+ "epoch": 11.1117903930131,
+ "grad_norm": 0.0202149897813797,
+ "learning_rate": 0.0006,
+ "loss": 4.933640480041504,
+ "step": 800
+ },
+ {
+ "epoch": 11.125764192139737,
+ "grad_norm": 0.01893039606511593,
+ "learning_rate": 0.0006,
+ "loss": 4.901970863342285,
+ "step": 801
+ },
+ {
+ "epoch": 11.139737991266376,
+ "grad_norm": 0.016040155664086342,
+ "learning_rate": 0.0006,
+ "loss": 4.905491828918457,
+ "step": 802
+ },
+ {
+ "epoch": 11.153711790393013,
+ "grad_norm": 0.01720350608229637,
+ "learning_rate": 0.0006,
+ "loss": 4.798816680908203,
+ "step": 803
+ },
+ {
+ "epoch": 11.167685589519651,
+ "grad_norm": 0.017002852633595467,
+ "learning_rate": 0.0006,
+ "loss": 4.851859092712402,
+ "step": 804
+ },
+ {
+ "epoch": 11.181659388646288,
+ "grad_norm": 0.015594680793583393,
+ "learning_rate": 0.0006,
+ "loss": 4.90328311920166,
+ "step": 805
+ },
+ {
+ "epoch": 11.195633187772925,
+ "grad_norm": 0.014904462732374668,
+ "learning_rate": 0.0006,
+ "loss": 4.904215335845947,
+ "step": 806
+ },
+ {
+ "epoch": 11.209606986899564,
+ "grad_norm": 0.01366228237748146,
+ "learning_rate": 0.0006,
+ "loss": 4.9813055992126465,
+ "step": 807
+ },
+ {
+ "epoch": 11.2235807860262,
+ "grad_norm": 0.014719638973474503,
+ "learning_rate": 0.0006,
+ "loss": 4.927306175231934,
+ "step": 808
+ },
+ {
+ "epoch": 11.237554585152838,
+ "grad_norm": 0.015016036108136177,
+ "learning_rate": 0.0006,
+ "loss": 4.874215126037598,
+ "step": 809
+ },
+ {
+ "epoch": 11.251528384279476,
+ "grad_norm": 0.014925424009561539,
+ "learning_rate": 0.0006,
+ "loss": 4.838257789611816,
+ "step": 810
+ },
+ {
+ "epoch": 11.265502183406113,
+ "grad_norm": 0.014335823245346546,
+ "learning_rate": 0.0006,
+ "loss": 4.940608024597168,
+ "step": 811
+ },
+ {
+ "epoch": 11.279475982532752,
+ "grad_norm": 0.014987512491643429,
+ "learning_rate": 0.0006,
+ "loss": 4.831531524658203,
+ "step": 812
+ },
+ {
+ "epoch": 11.293449781659389,
+ "grad_norm": 0.016087457537651062,
+ "learning_rate": 0.0006,
+ "loss": 4.863227367401123,
+ "step": 813
+ },
+ {
+ "epoch": 11.307423580786025,
+ "grad_norm": 0.01671300269663334,
+ "learning_rate": 0.0006,
+ "loss": 4.8535661697387695,
+ "step": 814
+ },
+ {
+ "epoch": 11.321397379912664,
+ "grad_norm": 0.015229142270982265,
+ "learning_rate": 0.0006,
+ "loss": 4.948635101318359,
+ "step": 815
+ },
+ {
+ "epoch": 11.335371179039301,
+ "grad_norm": 0.01532587967813015,
+ "learning_rate": 0.0006,
+ "loss": 4.9035797119140625,
+ "step": 816
+ },
+ {
+ "epoch": 11.34934497816594,
+ "grad_norm": 0.015715476125478745,
+ "learning_rate": 0.0006,
+ "loss": 4.894709587097168,
+ "step": 817
+ },
+ {
+ "epoch": 11.363318777292577,
+ "grad_norm": 0.014156977646052837,
+ "learning_rate": 0.0006,
+ "loss": 4.960206985473633,
+ "step": 818
+ },
+ {
+ "epoch": 11.377292576419213,
+ "grad_norm": 0.017076566815376282,
+ "learning_rate": 0.0006,
+ "loss": 5.005850791931152,
+ "step": 819
+ },
+ {
+ "epoch": 11.391266375545852,
+ "grad_norm": 0.0172084029763937,
+ "learning_rate": 0.0006,
+ "loss": 4.903355598449707,
+ "step": 820
+ },
+ {
+ "epoch": 11.405240174672489,
+ "grad_norm": 0.0180258359760046,
+ "learning_rate": 0.0006,
+ "loss": 4.842667579650879,
+ "step": 821
+ },
+ {
+ "epoch": 11.419213973799126,
+ "grad_norm": 0.022130120545625687,
+ "learning_rate": 0.0006,
+ "loss": 4.799970626831055,
+ "step": 822
+ },
+ {
+ "epoch": 11.433187772925764,
+ "grad_norm": 0.02384897693991661,
+ "learning_rate": 0.0006,
+ "loss": 4.860746383666992,
+ "step": 823
+ },
+ {
+ "epoch": 11.447161572052401,
+ "grad_norm": 0.021881572902202606,
+ "learning_rate": 0.0006,
+ "loss": 4.918107032775879,
+ "step": 824
+ },
+ {
+ "epoch": 11.46113537117904,
+ "grad_norm": 0.023085080087184906,
+ "learning_rate": 0.0006,
+ "loss": 4.878946304321289,
+ "step": 825
+ },
+ {
+ "epoch": 11.475109170305677,
+ "grad_norm": 0.02291012555360794,
+ "learning_rate": 0.0006,
+ "loss": 4.958887100219727,
+ "step": 826
+ },
+ {
+ "epoch": 11.489082969432314,
+ "grad_norm": 0.0199968870729208,
+ "learning_rate": 0.0006,
+ "loss": 4.80955696105957,
+ "step": 827
+ },
+ {
+ "epoch": 11.503056768558952,
+ "grad_norm": 0.018702229484915733,
+ "learning_rate": 0.0006,
+ "loss": 5.0276408195495605,
+ "step": 828
+ },
+ {
+ "epoch": 11.51703056768559,
+ "grad_norm": 0.017787311226129532,
+ "learning_rate": 0.0006,
+ "loss": 4.96018123626709,
+ "step": 829
+ },
+ {
+ "epoch": 11.531004366812226,
+ "grad_norm": 0.01704501546919346,
+ "learning_rate": 0.0006,
+ "loss": 4.914553165435791,
+ "step": 830
+ },
+ {
+ "epoch": 11.544978165938865,
+ "grad_norm": 0.017519677057862282,
+ "learning_rate": 0.0006,
+ "loss": 4.798098087310791,
+ "step": 831
+ },
+ {
+ "epoch": 11.558951965065502,
+ "grad_norm": 0.018511991947889328,
+ "learning_rate": 0.0006,
+ "loss": 4.817785263061523,
+ "step": 832
+ },
+ {
+ "epoch": 11.57292576419214,
+ "grad_norm": 0.01876644790172577,
+ "learning_rate": 0.0006,
+ "loss": 4.818995475769043,
+ "step": 833
+ },
+ {
+ "epoch": 11.586899563318777,
+ "grad_norm": 0.01753568835556507,
+ "learning_rate": 0.0006,
+ "loss": 4.811938285827637,
+ "step": 834
+ },
+ {
+ "epoch": 11.600873362445414,
+ "grad_norm": 0.017046531662344933,
+ "learning_rate": 0.0006,
+ "loss": 4.914674282073975,
+ "step": 835
+ },
+ {
+ "epoch": 11.614847161572053,
+ "grad_norm": 0.01802094094455242,
+ "learning_rate": 0.0006,
+ "loss": 4.993284225463867,
+ "step": 836
+ },
+ {
+ "epoch": 11.62882096069869,
+ "grad_norm": 0.019493764266371727,
+ "learning_rate": 0.0006,
+ "loss": 4.895628452301025,
+ "step": 837
+ },
+ {
+ "epoch": 11.642794759825328,
+ "grad_norm": 0.019461210817098618,
+ "learning_rate": 0.0006,
+ "loss": 4.869269371032715,
+ "step": 838
+ },
+ {
+ "epoch": 11.656768558951965,
+ "grad_norm": 0.01981567218899727,
+ "learning_rate": 0.0006,
+ "loss": 4.863052845001221,
+ "step": 839
+ },
+ {
+ "epoch": 11.670742358078602,
+ "grad_norm": 0.019894491881132126,
+ "learning_rate": 0.0006,
+ "loss": 4.8775787353515625,
+ "step": 840
+ },
+ {
+ "epoch": 11.68471615720524,
+ "grad_norm": 0.01875786855816841,
+ "learning_rate": 0.0006,
+ "loss": 4.817763328552246,
+ "step": 841
+ },
+ {
+ "epoch": 11.698689956331878,
+ "grad_norm": 0.020926695317029953,
+ "learning_rate": 0.0006,
+ "loss": 4.7470526695251465,
+ "step": 842
+ },
+ {
+ "epoch": 11.712663755458514,
+ "grad_norm": 0.021477103233337402,
+ "learning_rate": 0.0006,
+ "loss": 4.762682914733887,
+ "step": 843
+ },
+ {
+ "epoch": 11.726637554585153,
+ "grad_norm": 0.020628711208701134,
+ "learning_rate": 0.0006,
+ "loss": 4.807985782623291,
+ "step": 844
+ },
+ {
+ "epoch": 11.74061135371179,
+ "grad_norm": 0.021489612758159637,
+ "learning_rate": 0.0006,
+ "loss": 4.843091011047363,
+ "step": 845
+ },
+ {
+ "epoch": 11.754585152838429,
+ "grad_norm": 0.021194491535425186,
+ "learning_rate": 0.0006,
+ "loss": 4.8300251960754395,
+ "step": 846
+ },
+ {
+ "epoch": 11.768558951965066,
+ "grad_norm": 0.01833350397646427,
+ "learning_rate": 0.0006,
+ "loss": 5.000202178955078,
+ "step": 847
+ },
+ {
+ "epoch": 11.782532751091702,
+ "grad_norm": 0.018096931278705597,
+ "learning_rate": 0.0006,
+ "loss": 4.880648136138916,
+ "step": 848
+ },
+ {
+ "epoch": 11.796506550218341,
+ "grad_norm": 0.0197161752730608,
+ "learning_rate": 0.0006,
+ "loss": 4.788316249847412,
+ "step": 849
+ },
+ {
+ "epoch": 11.810480349344978,
+ "grad_norm": 0.0203181654214859,
+ "learning_rate": 0.0006,
+ "loss": 4.870944976806641,
+ "step": 850
+ },
+ {
+ "epoch": 11.824454148471617,
+ "grad_norm": 0.01908455230295658,
+ "learning_rate": 0.0006,
+ "loss": 4.890144348144531,
+ "step": 851
+ },
+ {
+ "epoch": 11.838427947598253,
+ "grad_norm": 0.021281801164150238,
+ "learning_rate": 0.0006,
+ "loss": 4.89578914642334,
+ "step": 852
+ },
+ {
+ "epoch": 11.85240174672489,
+ "grad_norm": 0.02124273031949997,
+ "learning_rate": 0.0006,
+ "loss": 4.883749008178711,
+ "step": 853
+ },
+ {
+ "epoch": 11.866375545851529,
+ "grad_norm": 0.020969398319721222,
+ "learning_rate": 0.0006,
+ "loss": 4.809142589569092,
+ "step": 854
+ },
+ {
+ "epoch": 11.880349344978166,
+ "grad_norm": 0.01967657171189785,
+ "learning_rate": 0.0006,
+ "loss": 4.821864604949951,
+ "step": 855
+ },
+ {
+ "epoch": 11.894323144104803,
+ "grad_norm": 0.020887333899736404,
+ "learning_rate": 0.0006,
+ "loss": 4.917985916137695,
+ "step": 856
+ },
+ {
+ "epoch": 11.908296943231441,
+ "grad_norm": 0.019203083589673042,
+ "learning_rate": 0.0006,
+ "loss": 4.824960708618164,
+ "step": 857
+ },
+ {
+ "epoch": 11.922270742358078,
+ "grad_norm": 0.020162401720881462,
+ "learning_rate": 0.0006,
+ "loss": 4.772617816925049,
+ "step": 858
+ },
+ {
+ "epoch": 11.936244541484717,
+ "grad_norm": 0.02251487225294113,
+ "learning_rate": 0.0006,
+ "loss": 4.735418319702148,
+ "step": 859
+ },
+ {
+ "epoch": 11.950218340611354,
+ "grad_norm": 0.020313527435064316,
+ "learning_rate": 0.0006,
+ "loss": 4.939233779907227,
+ "step": 860
+ },
+ {
+ "epoch": 11.96419213973799,
+ "grad_norm": 0.021635599434375763,
+ "learning_rate": 0.0006,
+ "loss": 4.861807346343994,
+ "step": 861
+ },
+ {
+ "epoch": 11.97816593886463,
+ "grad_norm": 0.02285311557352543,
+ "learning_rate": 0.0006,
+ "loss": 4.932732582092285,
+ "step": 862
+ },
+ {
+ "epoch": 11.992139737991266,
+ "grad_norm": 0.02072535641491413,
+ "learning_rate": 0.0006,
+ "loss": 4.796767711639404,
+ "step": 863
+ },
+ {
+ "epoch": 12.0,
+ "grad_norm": 0.020588304847478867,
+ "learning_rate": 0.0006,
+ "loss": 4.997537612915039,
+ "step": 864
+ },
+ {
+ "epoch": 12.0,
+ "eval_loss": 5.044651985168457,
+ "eval_runtime": 56.9145,
+ "eval_samples_per_second": 42.906,
+ "eval_steps_per_second": 1.353,
+ "step": 864
+ },
+ {
+ "epoch": 12.013973799126637,
+ "grad_norm": 0.017597166821360588,
+ "learning_rate": 0.0006,
+ "loss": 4.830524444580078,
+ "step": 865
+ },
+ {
+ "epoch": 12.027947598253276,
+ "grad_norm": 0.01716788113117218,
+ "learning_rate": 0.0006,
+ "loss": 4.894325256347656,
+ "step": 866
+ },
+ {
+ "epoch": 12.041921397379912,
+ "grad_norm": 0.017712457105517387,
+ "learning_rate": 0.0006,
+ "loss": 4.868297576904297,
+ "step": 867
+ },
+ {
+ "epoch": 12.055895196506551,
+ "grad_norm": 0.021486390382051468,
+ "learning_rate": 0.0006,
+ "loss": 4.922049522399902,
+ "step": 868
+ },
+ {
+ "epoch": 12.069868995633188,
+ "grad_norm": 0.024560654535889626,
+ "learning_rate": 0.0006,
+ "loss": 4.734241008758545,
+ "step": 869
+ },
+ {
+ "epoch": 12.083842794759825,
+ "grad_norm": 0.025870252400636673,
+ "learning_rate": 0.0006,
+ "loss": 4.868566989898682,
+ "step": 870
+ },
+ {
+ "epoch": 12.097816593886463,
+ "grad_norm": 0.0277389045804739,
+ "learning_rate": 0.0006,
+ "loss": 4.755006790161133,
+ "step": 871
+ },
+ {
+ "epoch": 12.1117903930131,
+ "grad_norm": 0.029100844636559486,
+ "learning_rate": 0.0006,
+ "loss": 4.909667015075684,
+ "step": 872
+ },
+ {
+ "epoch": 12.125764192139737,
+ "grad_norm": 0.031833067536354065,
+ "learning_rate": 0.0006,
+ "loss": 4.858674049377441,
+ "step": 873
+ },
+ {
+ "epoch": 12.139737991266376,
+ "grad_norm": 0.032709237188100815,
+ "learning_rate": 0.0006,
+ "loss": 4.823696136474609,
+ "step": 874
+ },
+ {
+ "epoch": 12.153711790393013,
+ "grad_norm": 0.02995380386710167,
+ "learning_rate": 0.0006,
+ "loss": 4.850235462188721,
+ "step": 875
+ },
+ {
+ "epoch": 12.167685589519651,
+ "grad_norm": 0.034098610281944275,
+ "learning_rate": 0.0006,
+ "loss": 4.8169989585876465,
+ "step": 876
+ },
+ {
+ "epoch": 12.181659388646288,
+ "grad_norm": 0.03377080336213112,
+ "learning_rate": 0.0006,
+ "loss": 4.90738582611084,
+ "step": 877
+ },
+ {
+ "epoch": 12.195633187772925,
+ "grad_norm": 0.029780825600028038,
+ "learning_rate": 0.0006,
+ "loss": 4.811163902282715,
+ "step": 878
+ },
+ {
+ "epoch": 12.209606986899564,
+ "grad_norm": 0.03118024580180645,
+ "learning_rate": 0.0006,
+ "loss": 4.861286640167236,
+ "step": 879
+ },
+ {
+ "epoch": 12.2235807860262,
+ "grad_norm": 0.033052004873752594,
+ "learning_rate": 0.0006,
+ "loss": 4.865021228790283,
+ "step": 880
+ },
+ {
+ "epoch": 12.237554585152838,
+ "grad_norm": 0.029070032760500908,
+ "learning_rate": 0.0006,
+ "loss": 4.947303771972656,
+ "step": 881
+ },
+ {
+ "epoch": 12.251528384279476,
+ "grad_norm": 0.028355715796351433,
+ "learning_rate": 0.0006,
+ "loss": 4.804648399353027,
+ "step": 882
+ },
+ {
+ "epoch": 12.265502183406113,
+ "grad_norm": 0.031257227063179016,
+ "learning_rate": 0.0006,
+ "loss": 4.757718086242676,
+ "step": 883
+ },
+ {
+ "epoch": 12.279475982532752,
+ "grad_norm": 0.031438373029232025,
+ "learning_rate": 0.0006,
+ "loss": 4.851268768310547,
+ "step": 884
+ },
+ {
+ "epoch": 12.293449781659389,
+ "grad_norm": 0.030492138117551804,
+ "learning_rate": 0.0006,
+ "loss": 4.7891035079956055,
+ "step": 885
+ },
+ {
+ "epoch": 12.307423580786025,
+ "grad_norm": 0.026115991175174713,
+ "learning_rate": 0.0006,
+ "loss": 4.86474609375,
+ "step": 886
+ },
+ {
+ "epoch": 12.321397379912664,
+ "grad_norm": 0.024184465408325195,
+ "learning_rate": 0.0006,
+ "loss": 4.860456943511963,
+ "step": 887
+ },
+ {
+ "epoch": 12.335371179039301,
+ "grad_norm": 0.025437982752919197,
+ "learning_rate": 0.0006,
+ "loss": 4.806251049041748,
+ "step": 888
+ },
+ {
+ "epoch": 12.34934497816594,
+ "grad_norm": 0.022390231490135193,
+ "learning_rate": 0.0006,
+ "loss": 4.807086944580078,
+ "step": 889
+ },
+ {
+ "epoch": 12.363318777292577,
+ "grad_norm": 0.020174294710159302,
+ "learning_rate": 0.0006,
+ "loss": 4.78292179107666,
+ "step": 890
+ },
+ {
+ "epoch": 12.377292576419213,
+ "grad_norm": 0.019734324887394905,
+ "learning_rate": 0.0006,
+ "loss": 4.784509658813477,
+ "step": 891
+ },
+ {
+ "epoch": 12.391266375545852,
+ "grad_norm": 0.016733523458242416,
+ "learning_rate": 0.0006,
+ "loss": 4.811755180358887,
+ "step": 892
+ },
+ {
+ "epoch": 12.405240174672489,
+ "grad_norm": 0.01570146717131138,
+ "learning_rate": 0.0006,
+ "loss": 4.720088005065918,
+ "step": 893
+ },
+ {
+ "epoch": 12.419213973799126,
+ "grad_norm": 0.01601393334567547,
+ "learning_rate": 0.0006,
+ "loss": 4.783543109893799,
+ "step": 894
+ },
+ {
+ "epoch": 12.433187772925764,
+ "grad_norm": 0.014035487547516823,
+ "learning_rate": 0.0006,
+ "loss": 4.848340034484863,
+ "step": 895
+ },
+ {
+ "epoch": 12.447161572052401,
+ "grad_norm": 0.015248487703502178,
+ "learning_rate": 0.0006,
+ "loss": 4.773287773132324,
+ "step": 896
+ },
+ {
+ "epoch": 12.46113537117904,
+ "grad_norm": 0.01532268151640892,
+ "learning_rate": 0.0006,
+ "loss": 4.799646854400635,
+ "step": 897
+ },
+ {
+ "epoch": 12.475109170305677,
+ "grad_norm": 0.013982534408569336,
+ "learning_rate": 0.0006,
+ "loss": 4.8090500831604,
+ "step": 898
+ },
+ {
+ "epoch": 12.489082969432314,
+ "grad_norm": 0.014081164263188839,
+ "learning_rate": 0.0006,
+ "loss": 4.751888751983643,
+ "step": 899
+ },
+ {
+ "epoch": 12.503056768558952,
+ "grad_norm": 0.01380133256316185,
+ "learning_rate": 0.0006,
+ "loss": 4.78144645690918,
+ "step": 900
+ },
+ {
+ "epoch": 12.51703056768559,
+ "grad_norm": 0.013555269688367844,
+ "learning_rate": 0.0006,
+ "loss": 4.779257774353027,
+ "step": 901
+ },
+ {
+ "epoch": 12.531004366812226,
+ "grad_norm": 0.013075701892375946,
+ "learning_rate": 0.0006,
+ "loss": 4.820705413818359,
+ "step": 902
+ },
+ {
+ "epoch": 12.544978165938865,
+ "grad_norm": 0.01276316400617361,
+ "learning_rate": 0.0006,
+ "loss": 4.836733818054199,
+ "step": 903
+ },
+ {
+ "epoch": 12.558951965065502,
+ "grad_norm": 0.013033892959356308,
+ "learning_rate": 0.0006,
+ "loss": 4.764178276062012,
+ "step": 904
+ },
+ {
+ "epoch": 12.57292576419214,
+ "grad_norm": 0.013996980153024197,
+ "learning_rate": 0.0006,
+ "loss": 4.767983436584473,
+ "step": 905
+ },
+ {
+ "epoch": 12.586899563318777,
+ "grad_norm": 0.015222184360027313,
+ "learning_rate": 0.0006,
+ "loss": 4.759721755981445,
+ "step": 906
+ },
+ {
+ "epoch": 12.600873362445414,
+ "grad_norm": 0.01623941957950592,
+ "learning_rate": 0.0006,
+ "loss": 4.849811553955078,
+ "step": 907
+ },
+ {
+ "epoch": 12.614847161572053,
+ "grad_norm": 0.015711182728409767,
+ "learning_rate": 0.0006,
+ "loss": 4.7829084396362305,
+ "step": 908
+ },
+ {
+ "epoch": 12.62882096069869,
+ "grad_norm": 0.014137467369437218,
+ "learning_rate": 0.0006,
+ "loss": 4.73321533203125,
+ "step": 909
+ },
+ {
+ "epoch": 12.642794759825328,
+ "grad_norm": 0.012738462537527084,
+ "learning_rate": 0.0006,
+ "loss": 4.870052337646484,
+ "step": 910
+ },
+ {
+ "epoch": 12.656768558951965,
+ "grad_norm": 0.013683130033314228,
+ "learning_rate": 0.0006,
+ "loss": 4.82658052444458,
+ "step": 911
+ },
+ {
+ "epoch": 12.670742358078602,
+ "grad_norm": 0.014409830793738365,
+ "learning_rate": 0.0006,
+ "loss": 4.707494735717773,
+ "step": 912
+ },
+ {
+ "epoch": 12.68471615720524,
+ "grad_norm": 0.013806809671223164,
+ "learning_rate": 0.0006,
+ "loss": 4.778928279876709,
+ "step": 913
+ },
+ {
+ "epoch": 12.698689956331878,
+ "grad_norm": 0.012857912108302116,
+ "learning_rate": 0.0006,
+ "loss": 4.890717506408691,
+ "step": 914
+ },
+ {
+ "epoch": 12.712663755458514,
+ "grad_norm": 0.013828758150339127,
+ "learning_rate": 0.0006,
+ "loss": 4.8357439041137695,
+ "step": 915
+ },
+ {
+ "epoch": 12.726637554585153,
+ "grad_norm": 0.01478351280093193,
+ "learning_rate": 0.0006,
+ "loss": 4.869986534118652,
+ "step": 916
+ },
+ {
+ "epoch": 12.74061135371179,
+ "grad_norm": 0.01326421182602644,
+ "learning_rate": 0.0006,
+ "loss": 4.830646514892578,
+ "step": 917
+ },
+ {
+ "epoch": 12.754585152838429,
+ "grad_norm": 0.013335433788597584,
+ "learning_rate": 0.0006,
+ "loss": 4.7786149978637695,
+ "step": 918
+ },
+ {
+ "epoch": 12.768558951965066,
+ "grad_norm": 0.014261147007346153,
+ "learning_rate": 0.0006,
+ "loss": 4.788693428039551,
+ "step": 919
+ },
+ {
+ "epoch": 12.782532751091702,
+ "grad_norm": 0.014274783432483673,
+ "learning_rate": 0.0006,
+ "loss": 4.77295446395874,
+ "step": 920
+ },
+ {
+ "epoch": 12.796506550218341,
+ "grad_norm": 0.015074674971401691,
+ "learning_rate": 0.0006,
+ "loss": 4.830754280090332,
+ "step": 921
+ },
+ {
+ "epoch": 12.810480349344978,
+ "grad_norm": 0.015599515289068222,
+ "learning_rate": 0.0006,
+ "loss": 4.837759017944336,
+ "step": 922
+ },
+ {
+ "epoch": 12.824454148471617,
+ "grad_norm": 0.014595242217183113,
+ "learning_rate": 0.0006,
+ "loss": 4.882965564727783,
+ "step": 923
+ },
+ {
+ "epoch": 12.838427947598253,
+ "grad_norm": 0.014951467514038086,
+ "learning_rate": 0.0006,
+ "loss": 4.709748268127441,
+ "step": 924
+ },
+ {
+ "epoch": 12.85240174672489,
+ "grad_norm": 0.01539295818656683,
+ "learning_rate": 0.0006,
+ "loss": 4.791160583496094,
+ "step": 925
+ },
+ {
+ "epoch": 12.866375545851529,
+ "grad_norm": 0.015987906605005264,
+ "learning_rate": 0.0006,
+ "loss": 4.661876678466797,
+ "step": 926
+ },
+ {
+ "epoch": 12.880349344978166,
+ "grad_norm": 0.016527341678738594,
+ "learning_rate": 0.0006,
+ "loss": 4.721952438354492,
+ "step": 927
+ },
+ {
+ "epoch": 12.894323144104803,
+ "grad_norm": 0.01637161523103714,
+ "learning_rate": 0.0006,
+ "loss": 4.704446792602539,
+ "step": 928
+ },
+ {
+ "epoch": 12.908296943231441,
+ "grad_norm": 0.015221050009131432,
+ "learning_rate": 0.0006,
+ "loss": 4.771907806396484,
+ "step": 929
+ },
+ {
+ "epoch": 12.922270742358078,
+ "grad_norm": 0.016242166981101036,
+ "learning_rate": 0.0006,
+ "loss": 4.806267738342285,
+ "step": 930
+ },
+ {
+ "epoch": 12.936244541484717,
+ "grad_norm": 0.01628357358276844,
+ "learning_rate": 0.0006,
+ "loss": 4.7252326011657715,
+ "step": 931
+ },
+ {
+ "epoch": 12.950218340611354,
+ "grad_norm": 0.017428133636713028,
+ "learning_rate": 0.0006,
+ "loss": 4.765689849853516,
+ "step": 932
+ },
+ {
+ "epoch": 12.96419213973799,
+ "grad_norm": 0.01870061084628105,
+ "learning_rate": 0.0006,
+ "loss": 4.814488887786865,
+ "step": 933
+ },
+ {
+ "epoch": 12.97816593886463,
+ "grad_norm": 0.019796060398221016,
+ "learning_rate": 0.0006,
+ "loss": 4.739185810089111,
+ "step": 934
+ },
+ {
+ "epoch": 12.992139737991266,
+ "grad_norm": 0.01985820196568966,
+ "learning_rate": 0.0006,
+ "loss": 4.810608863830566,
+ "step": 935
+ },
+ {
+ "epoch": 13.0,
+ "grad_norm": 0.020834477618336678,
+ "learning_rate": 0.0006,
+ "loss": 4.727858066558838,
+ "step": 936
+ },
+ {
+ "epoch": 13.0,
+ "eval_loss": 4.979116916656494,
+ "eval_runtime": 57.2877,
+ "eval_samples_per_second": 42.627,
+ "eval_steps_per_second": 1.344,
+ "step": 936
+ },
+ {
+ "epoch": 13.013973799126637,
+ "grad_norm": 0.02025519125163555,
+ "learning_rate": 0.0006,
+ "loss": 4.628875732421875,
+ "step": 937
+ },
+ {
+ "epoch": 13.027947598253276,
+ "grad_norm": 0.019596368074417114,
+ "learning_rate": 0.0006,
+ "loss": 4.694699287414551,
+ "step": 938
+ },
+ {
+ "epoch": 13.041921397379912,
+ "grad_norm": 0.020494934171438217,
+ "learning_rate": 0.0006,
+ "loss": 4.691821098327637,
+ "step": 939
+ },
+ {
+ "epoch": 13.055895196506551,
+ "grad_norm": 0.020121760666370392,
+ "learning_rate": 0.0006,
+ "loss": 4.71478271484375,
+ "step": 940
+ },
+ {
+ "epoch": 13.069868995633188,
+ "grad_norm": 0.021091405302286148,
+ "learning_rate": 0.0006,
+ "loss": 4.695915222167969,
+ "step": 941
+ },
+ {
+ "epoch": 13.083842794759825,
+ "grad_norm": 0.019787423312664032,
+ "learning_rate": 0.0006,
+ "loss": 4.8652448654174805,
+ "step": 942
+ },
+ {
+ "epoch": 13.097816593886463,
+ "grad_norm": 0.0185833927243948,
+ "learning_rate": 0.0006,
+ "loss": 4.7065887451171875,
+ "step": 943
+ },
+ {
+ "epoch": 13.1117903930131,
+ "grad_norm": 0.021220475435256958,
+ "learning_rate": 0.0006,
+ "loss": 4.741232395172119,
+ "step": 944
+ },
+ {
+ "epoch": 13.125764192139737,
+ "grad_norm": 0.02136622555553913,
+ "learning_rate": 0.0006,
+ "loss": 4.705967903137207,
+ "step": 945
+ },
+ {
+ "epoch": 13.139737991266376,
+ "grad_norm": 0.0191587433218956,
+ "learning_rate": 0.0006,
+ "loss": 4.758667945861816,
+ "step": 946
+ },
+ {
+ "epoch": 13.153711790393013,
+ "grad_norm": 0.018290022388100624,
+ "learning_rate": 0.0006,
+ "loss": 4.760666847229004,
+ "step": 947
+ },
+ {
+ "epoch": 13.167685589519651,
+ "grad_norm": 0.021704409271478653,
+ "learning_rate": 0.0006,
+ "loss": 4.750555992126465,
+ "step": 948
+ },
+ {
+ "epoch": 13.181659388646288,
+ "grad_norm": 0.021668918430805206,
+ "learning_rate": 0.0006,
+ "loss": 4.840903282165527,
+ "step": 949
+ },
+ {
+ "epoch": 13.195633187772925,
+ "grad_norm": 0.019497141242027283,
+ "learning_rate": 0.0006,
+ "loss": 4.738969326019287,
+ "step": 950
+ },
+ {
+ "epoch": 13.209606986899564,
+ "grad_norm": 0.018500596284866333,
+ "learning_rate": 0.0006,
+ "loss": 4.75154972076416,
+ "step": 951
+ },
+ {
+ "epoch": 13.2235807860262,
+ "grad_norm": 0.01872239261865616,
+ "learning_rate": 0.0006,
+ "loss": 4.811738014221191,
+ "step": 952
+ },
+ {
+ "epoch": 13.237554585152838,
+ "grad_norm": 0.020321547985076904,
+ "learning_rate": 0.0006,
+ "loss": 4.815384387969971,
+ "step": 953
+ },
+ {
+ "epoch": 13.251528384279476,
+ "grad_norm": 0.02283575013279915,
+ "learning_rate": 0.0006,
+ "loss": 4.74582052230835,
+ "step": 954
+ },
+ {
+ "epoch": 13.265502183406113,
+ "grad_norm": 0.02474398910999298,
+ "learning_rate": 0.0006,
+ "loss": 4.738375186920166,
+ "step": 955
+ },
+ {
+ "epoch": 13.279475982532752,
+ "grad_norm": 0.02065764181315899,
+ "learning_rate": 0.0006,
+ "loss": 4.746875286102295,
+ "step": 956
+ },
+ {
+ "epoch": 13.293449781659389,
+ "grad_norm": 0.020534520968794823,
+ "learning_rate": 0.0006,
+ "loss": 4.656914710998535,
+ "step": 957
+ },
+ {
+ "epoch": 13.307423580786025,
+ "grad_norm": 0.0206378772854805,
+ "learning_rate": 0.0006,
+ "loss": 4.842615604400635,
+ "step": 958
+ },
+ {
+ "epoch": 13.321397379912664,
+ "grad_norm": 0.017935654148459435,
+ "learning_rate": 0.0006,
+ "loss": 4.717827796936035,
+ "step": 959
+ },
+ {
+ "epoch": 13.335371179039301,
+ "grad_norm": 0.016958363354206085,
+ "learning_rate": 0.0006,
+ "loss": 4.681070327758789,
+ "step": 960
+ },
+ {
+ "epoch": 13.34934497816594,
+ "grad_norm": 0.015735412016510963,
+ "learning_rate": 0.0006,
+ "loss": 4.7957048416137695,
+ "step": 961
+ },
+ {
+ "epoch": 13.363318777292577,
+ "grad_norm": 0.01563699170947075,
+ "learning_rate": 0.0006,
+ "loss": 4.701192378997803,
+ "step": 962
+ },
+ {
+ "epoch": 13.377292576419213,
+ "grad_norm": 0.015207760035991669,
+ "learning_rate": 0.0006,
+ "loss": 4.73256254196167,
+ "step": 963
+ },
+ {
+ "epoch": 13.391266375545852,
+ "grad_norm": 0.015396883711218834,
+ "learning_rate": 0.0006,
+ "loss": 4.76412296295166,
+ "step": 964
+ },
+ {
+ "epoch": 13.405240174672489,
+ "grad_norm": 0.015690583735704422,
+ "learning_rate": 0.0006,
+ "loss": 4.791157245635986,
+ "step": 965
+ },
+ {
+ "epoch": 13.419213973799126,
+ "grad_norm": 0.01755724661052227,
+ "learning_rate": 0.0006,
+ "loss": 4.845666885375977,
+ "step": 966
+ },
+ {
+ "epoch": 13.433187772925764,
+ "grad_norm": 0.019443422555923462,
+ "learning_rate": 0.0006,
+ "loss": 4.5902299880981445,
+ "step": 967
+ },
+ {
+ "epoch": 13.447161572052401,
+ "grad_norm": 0.02118338644504547,
+ "learning_rate": 0.0006,
+ "loss": 4.740627288818359,
+ "step": 968
+ },
+ {
+ "epoch": 13.46113537117904,
+ "grad_norm": 0.02174839936196804,
+ "learning_rate": 0.0006,
+ "loss": 4.814560890197754,
+ "step": 969
+ },
+ {
+ "epoch": 13.475109170305677,
+ "grad_norm": 0.022860374301671982,
+ "learning_rate": 0.0006,
+ "loss": 4.66440486907959,
+ "step": 970
+ },
+ {
+ "epoch": 13.489082969432314,
+ "grad_norm": 0.024742677807807922,
+ "learning_rate": 0.0006,
+ "loss": 4.7451653480529785,
+ "step": 971
+ },
+ {
+ "epoch": 13.503056768558952,
+ "grad_norm": 0.027068907395005226,
+ "learning_rate": 0.0006,
+ "loss": 4.802584171295166,
+ "step": 972
+ },
+ {
+ "epoch": 13.51703056768559,
+ "grad_norm": 0.026316625997424126,
+ "learning_rate": 0.0006,
+ "loss": 4.842397689819336,
+ "step": 973
+ },
+ {
+ "epoch": 13.531004366812226,
+ "grad_norm": 0.024610962718725204,
+ "learning_rate": 0.0006,
+ "loss": 4.650721073150635,
+ "step": 974
+ },
+ {
+ "epoch": 13.544978165938865,
+ "grad_norm": 0.024122603237628937,
+ "learning_rate": 0.0006,
+ "loss": 4.6984357833862305,
+ "step": 975
+ },
+ {
+ "epoch": 13.558951965065502,
+ "grad_norm": 0.022699708119034767,
+ "learning_rate": 0.0006,
+ "loss": 4.793007850646973,
+ "step": 976
+ },
+ {
+ "epoch": 13.57292576419214,
+ "grad_norm": 0.021706534549593925,
+ "learning_rate": 0.0006,
+ "loss": 4.691362380981445,
+ "step": 977
+ },
+ {
+ "epoch": 13.586899563318777,
+ "grad_norm": 0.021363291889429092,
+ "learning_rate": 0.0006,
+ "loss": 4.76762580871582,
+ "step": 978
+ },
+ {
+ "epoch": 13.600873362445414,
+ "grad_norm": 0.023476677015423775,
+ "learning_rate": 0.0006,
+ "loss": 4.777202606201172,
+ "step": 979
+ },
+ {
+ "epoch": 13.614847161572053,
+ "grad_norm": 0.02457100711762905,
+ "learning_rate": 0.0006,
+ "loss": 4.590703010559082,
+ "step": 980
+ },
+ {
+ "epoch": 13.62882096069869,
+ "grad_norm": 0.02469099499285221,
+ "learning_rate": 0.0006,
+ "loss": 4.640185356140137,
+ "step": 981
+ },
+ {
+ "epoch": 13.642794759825328,
+ "grad_norm": 0.024914603680372238,
+ "learning_rate": 0.0006,
+ "loss": 4.772201061248779,
+ "step": 982
+ },
+ {
+ "epoch": 13.656768558951965,
+ "grad_norm": 0.025133173912763596,
+ "learning_rate": 0.0006,
+ "loss": 4.743337154388428,
+ "step": 983
+ },
+ {
+ "epoch": 13.670742358078602,
+ "grad_norm": 0.02034137211740017,
+ "learning_rate": 0.0006,
+ "loss": 4.752127647399902,
+ "step": 984
+ },
+ {
+ "epoch": 13.68471615720524,
+ "grad_norm": 0.019645661115646362,
+ "learning_rate": 0.0006,
+ "loss": 4.838453769683838,
+ "step": 985
+ },
+ {
+ "epoch": 13.698689956331878,
+ "grad_norm": 0.01900297962129116,
+ "learning_rate": 0.0006,
+ "loss": 4.804868221282959,
+ "step": 986
+ },
+ {
+ "epoch": 13.712663755458514,
+ "grad_norm": 0.020292341709136963,
+ "learning_rate": 0.0006,
+ "loss": 4.609743118286133,
+ "step": 987
+ },
+ {
+ "epoch": 13.726637554585153,
+ "grad_norm": 0.017014775425195694,
+ "learning_rate": 0.0006,
+ "loss": 4.740612983703613,
+ "step": 988
+ },
+ {
+ "epoch": 13.74061135371179,
+ "grad_norm": 0.015199844725430012,
+ "learning_rate": 0.0006,
+ "loss": 4.719701766967773,
+ "step": 989
+ },
+ {
+ "epoch": 13.754585152838429,
+ "grad_norm": 0.015125414356589317,
+ "learning_rate": 0.0006,
+ "loss": 4.686114311218262,
+ "step": 990
+ },
+ {
+ "epoch": 13.768558951965066,
+ "grad_norm": 0.014779540710151196,
+ "learning_rate": 0.0006,
+ "loss": 4.674347877502441,
+ "step": 991
+ },
+ {
+ "epoch": 13.782532751091702,
+ "grad_norm": 0.015255163423717022,
+ "learning_rate": 0.0006,
+ "loss": 4.6309309005737305,
+ "step": 992
+ },
+ {
+ "epoch": 13.796506550218341,
+ "grad_norm": 0.013692197389900684,
+ "learning_rate": 0.0006,
+ "loss": 4.698497295379639,
+ "step": 993
+ },
+ {
+ "epoch": 13.810480349344978,
+ "grad_norm": 0.014529623091220856,
+ "learning_rate": 0.0006,
+ "loss": 4.628443717956543,
+ "step": 994
+ },
+ {
+ "epoch": 13.824454148471617,
+ "grad_norm": 0.01576659083366394,
+ "learning_rate": 0.0006,
+ "loss": 4.788015842437744,
+ "step": 995
+ },
+ {
+ "epoch": 13.838427947598253,
+ "grad_norm": 0.015834277495741844,
+ "learning_rate": 0.0006,
+ "loss": 4.672703742980957,
+ "step": 996
+ },
+ {
+ "epoch": 13.85240174672489,
+ "grad_norm": 0.015539503656327724,
+ "learning_rate": 0.0006,
+ "loss": 4.731503486633301,
+ "step": 997
+ },
+ {
+ "epoch": 13.866375545851529,
+ "grad_norm": 0.013463746756315231,
+ "learning_rate": 0.0006,
+ "loss": 4.7006330490112305,
+ "step": 998
+ },
+ {
+ "epoch": 13.880349344978166,
+ "grad_norm": 0.016377726569771767,
+ "learning_rate": 0.0006,
+ "loss": 4.636074066162109,
+ "step": 999
+ },
+ {
+ "epoch": 13.894323144104803,
+ "grad_norm": 0.016660034656524658,
+ "learning_rate": 0.0006,
+ "loss": 4.742693901062012,
+ "step": 1000
+ },
+ {
+ "epoch": 13.908296943231441,
+ "grad_norm": 0.014902938157320023,
+ "learning_rate": 0.0006,
+ "loss": 4.769811630249023,
+ "step": 1001
+ },
+ {
+ "epoch": 13.922270742358078,
+ "grad_norm": 0.014883476309478283,
+ "learning_rate": 0.0006,
+ "loss": 4.731159210205078,
+ "step": 1002
+ },
+ {
+ "epoch": 13.936244541484717,
+ "grad_norm": 0.014494217932224274,
+ "learning_rate": 0.0006,
+ "loss": 4.659485816955566,
+ "step": 1003
+ },
+ {
+ "epoch": 13.950218340611354,
+ "grad_norm": 0.015426991507411003,
+ "learning_rate": 0.0006,
+ "loss": 4.755229949951172,
+ "step": 1004
+ },
+ {
+ "epoch": 13.96419213973799,
+ "grad_norm": 0.016648586839437485,
+ "learning_rate": 0.0006,
+ "loss": 4.74111270904541,
+ "step": 1005
+ },
+ {
+ "epoch": 13.97816593886463,
+ "grad_norm": 0.016980541869997978,
+ "learning_rate": 0.0006,
+ "loss": 4.74072265625,
+ "step": 1006
+ },
+ {
+ "epoch": 13.992139737991266,
+ "grad_norm": 0.018070152029395103,
+ "learning_rate": 0.0006,
+ "loss": 4.732975006103516,
+ "step": 1007
+ },
+ {
+ "epoch": 14.0,
+ "grad_norm": 0.019910665228962898,
+ "learning_rate": 0.0006,
+ "loss": 4.634697914123535,
+ "step": 1008
+ },
+ {
+ "epoch": 14.0,
+ "eval_loss": 4.850640296936035,
+ "eval_runtime": 57.3039,
+ "eval_samples_per_second": 42.615,
+ "eval_steps_per_second": 1.344,
+ "step": 1008
+ },
+ {
+ "epoch": 14.013973799126637,
+ "grad_norm": 0.01980053074657917,
+ "learning_rate": 0.0006,
+ "loss": 4.628744602203369,
+ "step": 1009
+ },
+ {
+ "epoch": 14.027947598253276,
+ "grad_norm": 0.02497623674571514,
+ "learning_rate": 0.0006,
+ "loss": 4.75930118560791,
+ "step": 1010
+ },
+ {
+ "epoch": 14.041921397379912,
+ "grad_norm": 0.02794097363948822,
+ "learning_rate": 0.0006,
+ "loss": 4.691340446472168,
+ "step": 1011
+ },
+ {
+ "epoch": 14.055895196506551,
+ "grad_norm": 0.023725535720586777,
+ "learning_rate": 0.0006,
+ "loss": 4.733664512634277,
+ "step": 1012
+ },
+ {
+ "epoch": 14.069868995633188,
+ "grad_norm": 0.022966474294662476,
+ "learning_rate": 0.0006,
+ "loss": 4.664808750152588,
+ "step": 1013
+ },
+ {
+ "epoch": 14.083842794759825,
+ "grad_norm": 0.02530563250184059,
+ "learning_rate": 0.0006,
+ "loss": 4.8047285079956055,
+ "step": 1014
+ },
+ {
+ "epoch": 14.097816593886463,
+ "grad_norm": 0.02491082064807415,
+ "learning_rate": 0.0006,
+ "loss": 4.667956352233887,
+ "step": 1015
+ },
+ {
+ "epoch": 14.1117903930131,
+ "grad_norm": 0.022110294550657272,
+ "learning_rate": 0.0006,
+ "loss": 4.666140556335449,
+ "step": 1016
+ },
+ {
+ "epoch": 14.125764192139737,
+ "grad_norm": 0.02065849117934704,
+ "learning_rate": 0.0006,
+ "loss": 4.581960678100586,
+ "step": 1017
+ },
+ {
+ "epoch": 14.139737991266376,
+ "grad_norm": 0.01941298507153988,
+ "learning_rate": 0.0006,
+ "loss": 4.614226341247559,
+ "step": 1018
+ },
+ {
+ "epoch": 14.153711790393013,
+ "grad_norm": 0.022441856563091278,
+ "learning_rate": 0.0006,
+ "loss": 4.638352394104004,
+ "step": 1019
+ },
+ {
+ "epoch": 14.167685589519651,
+ "grad_norm": 0.021519694477319717,
+ "learning_rate": 0.0006,
+ "loss": 4.700160980224609,
+ "step": 1020
+ },
+ {
+ "epoch": 14.181659388646288,
+ "grad_norm": 0.018956486135721207,
+ "learning_rate": 0.0006,
+ "loss": 4.649593353271484,
+ "step": 1021
+ },
+ {
+ "epoch": 14.195633187772925,
+ "grad_norm": 0.016681140288710594,
+ "learning_rate": 0.0006,
+ "loss": 4.603952407836914,
+ "step": 1022
+ },
+ {
+ "epoch": 14.209606986899564,
+ "grad_norm": 0.01788809522986412,
+ "learning_rate": 0.0006,
+ "loss": 4.669160842895508,
+ "step": 1023
+ },
+ {
+ "epoch": 14.2235807860262,
+ "grad_norm": 0.016047228127717972,
+ "learning_rate": 0.0006,
+ "loss": 4.6367950439453125,
+ "step": 1024
+ },
+ {
+ "epoch": 14.237554585152838,
+ "grad_norm": 0.01614239625632763,
+ "learning_rate": 0.0006,
+ "loss": 4.664529323577881,
+ "step": 1025
+ },
+ {
+ "epoch": 14.251528384279476,
+ "grad_norm": 0.016362549737095833,
+ "learning_rate": 0.0006,
+ "loss": 4.702373504638672,
+ "step": 1026
+ },
+ {
+ "epoch": 14.265502183406113,
+ "grad_norm": 0.015800409018993378,
+ "learning_rate": 0.0006,
+ "loss": 4.7594428062438965,
+ "step": 1027
+ },
+ {
+ "epoch": 14.279475982532752,
+ "grad_norm": 0.016730139032006264,
+ "learning_rate": 0.0006,
+ "loss": 4.677244186401367,
+ "step": 1028
+ },
+ {
+ "epoch": 14.293449781659389,
+ "grad_norm": 0.01661759801208973,
+ "learning_rate": 0.0006,
+ "loss": 4.633679389953613,
+ "step": 1029
+ },
+ {
+ "epoch": 14.307423580786025,
+ "grad_norm": 0.017673347145318985,
+ "learning_rate": 0.0006,
+ "loss": 4.649107456207275,
+ "step": 1030
+ },
+ {
+ "epoch": 14.321397379912664,
+ "grad_norm": 0.019243352115154266,
+ "learning_rate": 0.0006,
+ "loss": 4.611257553100586,
+ "step": 1031
+ },
+ {
+ "epoch": 14.335371179039301,
+ "grad_norm": 0.019817164167761803,
+ "learning_rate": 0.0006,
+ "loss": 4.588086128234863,
+ "step": 1032
+ },
+ {
+ "epoch": 14.34934497816594,
+ "grad_norm": 0.022292211651802063,
+ "learning_rate": 0.0006,
+ "loss": 4.783066749572754,
+ "step": 1033
+ },
+ {
+ "epoch": 14.363318777292577,
+ "grad_norm": 0.02140096202492714,
+ "learning_rate": 0.0006,
+ "loss": 4.604718208312988,
+ "step": 1034
+ },
+ {
+ "epoch": 14.377292576419213,
+ "grad_norm": 0.0187817495316267,
+ "learning_rate": 0.0006,
+ "loss": 4.721773147583008,
+ "step": 1035
+ },
+ {
+ "epoch": 14.391266375545852,
+ "grad_norm": 0.019879210740327835,
+ "learning_rate": 0.0006,
+ "loss": 4.789286136627197,
+ "step": 1036
+ },
+ {
+ "epoch": 14.405240174672489,
+ "grad_norm": 0.01885785162448883,
+ "learning_rate": 0.0006,
+ "loss": 4.669743537902832,
+ "step": 1037
+ },
+ {
+ "epoch": 14.419213973799126,
+ "grad_norm": 0.018324270844459534,
+ "learning_rate": 0.0006,
+ "loss": 4.65901517868042,
+ "step": 1038
+ },
+ {
+ "epoch": 14.433187772925764,
+ "grad_norm": 0.017664821818470955,
+ "learning_rate": 0.0006,
+ "loss": 4.667839527130127,
+ "step": 1039
+ },
+ {
+ "epoch": 14.447161572052401,
+ "grad_norm": 0.015702905133366585,
+ "learning_rate": 0.0006,
+ "loss": 4.64700984954834,
+ "step": 1040
+ },
+ {
+ "epoch": 14.46113537117904,
+ "grad_norm": 0.016674194484949112,
+ "learning_rate": 0.0006,
+ "loss": 4.694093227386475,
+ "step": 1041
+ },
+ {
+ "epoch": 14.475109170305677,
+ "grad_norm": 0.016869477927684784,
+ "learning_rate": 0.0006,
+ "loss": 4.619544982910156,
+ "step": 1042
+ },
+ {
+ "epoch": 14.489082969432314,
+ "grad_norm": 0.016557540744543076,
+ "learning_rate": 0.0006,
+ "loss": 4.568667411804199,
+ "step": 1043
+ },
+ {
+ "epoch": 14.503056768558952,
+ "grad_norm": 0.014500945806503296,
+ "learning_rate": 0.0006,
+ "loss": 4.561741828918457,
+ "step": 1044
+ },
+ {
+ "epoch": 14.51703056768559,
+ "grad_norm": 0.014228739775717258,
+ "learning_rate": 0.0006,
+ "loss": 4.654277324676514,
+ "step": 1045
+ },
+ {
+ "epoch": 14.531004366812226,
+ "grad_norm": 0.013894238509237766,
+ "learning_rate": 0.0006,
+ "loss": 4.5867719650268555,
+ "step": 1046
+ },
+ {
+ "epoch": 14.544978165938865,
+ "grad_norm": 0.01550068985670805,
+ "learning_rate": 0.0006,
+ "loss": 4.61168098449707,
+ "step": 1047
+ },
+ {
+ "epoch": 14.558951965065502,
+ "grad_norm": 0.015940772369503975,
+ "learning_rate": 0.0006,
+ "loss": 4.691247940063477,
+ "step": 1048
+ },
+ {
+ "epoch": 14.57292576419214,
+ "grad_norm": 0.01714535616338253,
+ "learning_rate": 0.0006,
+ "loss": 4.61183500289917,
+ "step": 1049
+ },
+ {
+ "epoch": 14.586899563318777,
+ "grad_norm": 0.017925038933753967,
+ "learning_rate": 0.0006,
+ "loss": 4.719022750854492,
+ "step": 1050
+ },
+ {
+ "epoch": 14.600873362445414,
+ "grad_norm": 0.018141552805900574,
+ "learning_rate": 0.0006,
+ "loss": 4.760281562805176,
+ "step": 1051
+ },
+ {
+ "epoch": 14.614847161572053,
+ "grad_norm": 0.017131328582763672,
+ "learning_rate": 0.0006,
+ "loss": 4.709680080413818,
+ "step": 1052
+ },
+ {
+ "epoch": 14.62882096069869,
+ "grad_norm": 0.018730048090219498,
+ "learning_rate": 0.0006,
+ "loss": 4.5994720458984375,
+ "step": 1053
+ },
+ {
+ "epoch": 14.642794759825328,
+ "grad_norm": 0.018894298002123833,
+ "learning_rate": 0.0006,
+ "loss": 4.669070720672607,
+ "step": 1054
+ },
+ {
+ "epoch": 14.656768558951965,
+ "grad_norm": 0.017643198370933533,
+ "learning_rate": 0.0006,
+ "loss": 4.641790390014648,
+ "step": 1055
+ },
+ {
+ "epoch": 14.670742358078602,
+ "grad_norm": 0.017647244036197662,
+ "learning_rate": 0.0006,
+ "loss": 4.653899192810059,
+ "step": 1056
+ },
+ {
+ "epoch": 14.68471615720524,
+ "grad_norm": 0.017304198816418648,
+ "learning_rate": 0.0006,
+ "loss": 4.7487473487854,
+ "step": 1057
+ },
+ {
+ "epoch": 14.698689956331878,
+ "grad_norm": 0.01638447679579258,
+ "learning_rate": 0.0006,
+ "loss": 4.71945858001709,
+ "step": 1058
+ },
+ {
+ "epoch": 14.712663755458514,
+ "grad_norm": 0.015892071649432182,
+ "learning_rate": 0.0006,
+ "loss": 4.744903564453125,
+ "step": 1059
+ },
+ {
+ "epoch": 14.726637554585153,
+ "grad_norm": 0.017420614138245583,
+ "learning_rate": 0.0006,
+ "loss": 4.781563758850098,
+ "step": 1060
+ },
+ {
+ "epoch": 14.74061135371179,
+ "grad_norm": 0.016498520970344543,
+ "learning_rate": 0.0006,
+ "loss": 4.622779846191406,
+ "step": 1061
+ },
+ {
+ "epoch": 14.754585152838429,
+ "grad_norm": 0.017482127994298935,
+ "learning_rate": 0.0006,
+ "loss": 4.62825870513916,
+ "step": 1062
+ },
+ {
+ "epoch": 14.768558951965066,
+ "grad_norm": 0.019018379971385002,
+ "learning_rate": 0.0006,
+ "loss": 4.693194389343262,
+ "step": 1063
+ },
+ {
+ "epoch": 14.782532751091702,
+ "grad_norm": 0.01993914321064949,
+ "learning_rate": 0.0006,
+ "loss": 4.742903709411621,
+ "step": 1064
+ },
+ {
+ "epoch": 14.796506550218341,
+ "grad_norm": 0.02301335707306862,
+ "learning_rate": 0.0006,
+ "loss": 4.613251686096191,
+ "step": 1065
+ },
+ {
+ "epoch": 14.810480349344978,
+ "grad_norm": 0.024088222533464432,
+ "learning_rate": 0.0006,
+ "loss": 4.594305992126465,
+ "step": 1066
+ },
+ {
+ "epoch": 14.824454148471617,
+ "grad_norm": 0.02493320405483246,
+ "learning_rate": 0.0006,
+ "loss": 4.700588226318359,
+ "step": 1067
+ },
+ {
+ "epoch": 14.838427947598253,
+ "grad_norm": 0.02501937560737133,
+ "learning_rate": 0.0006,
+ "loss": 4.601742744445801,
+ "step": 1068
+ },
+ {
+ "epoch": 14.85240174672489,
+ "grad_norm": 0.023490751162171364,
+ "learning_rate": 0.0006,
+ "loss": 4.589211463928223,
+ "step": 1069
+ },
+ {
+ "epoch": 14.866375545851529,
+ "grad_norm": 0.019695665687322617,
+ "learning_rate": 0.0006,
+ "loss": 4.635081768035889,
+ "step": 1070
+ },
+ {
+ "epoch": 14.880349344978166,
+ "grad_norm": 0.019276980310678482,
+ "learning_rate": 0.0006,
+ "loss": 4.737002372741699,
+ "step": 1071
+ },
+ {
+ "epoch": 14.894323144104803,
+ "grad_norm": 0.022430511191487312,
+ "learning_rate": 0.0006,
+ "loss": 4.808295726776123,
+ "step": 1072
+ },
+ {
+ "epoch": 14.908296943231441,
+ "grad_norm": 0.021543916314840317,
+ "learning_rate": 0.0006,
+ "loss": 4.559475421905518,
+ "step": 1073
+ },
+ {
+ "epoch": 14.922270742358078,
+ "grad_norm": 0.019885241985321045,
+ "learning_rate": 0.0006,
+ "loss": 4.647948741912842,
+ "step": 1074
+ },
+ {
+ "epoch": 14.936244541484717,
+ "grad_norm": 0.02115175500512123,
+ "learning_rate": 0.0006,
+ "loss": 4.680505752563477,
+ "step": 1075
+ },
+ {
+ "epoch": 14.950218340611354,
+ "grad_norm": 0.02156151458621025,
+ "learning_rate": 0.0006,
+ "loss": 4.673477649688721,
+ "step": 1076
+ },
+ {
+ "epoch": 14.96419213973799,
+ "grad_norm": 0.02184601128101349,
+ "learning_rate": 0.0006,
+ "loss": 4.644402027130127,
+ "step": 1077
+ },
+ {
+ "epoch": 14.97816593886463,
+ "grad_norm": 0.0211983360350132,
+ "learning_rate": 0.0006,
+ "loss": 4.821558475494385,
+ "step": 1078
+ },
+ {
+ "epoch": 14.992139737991266,
+ "grad_norm": 0.019636567682027817,
+ "learning_rate": 0.0006,
+ "loss": 4.578868865966797,
+ "step": 1079
+ },
+ {
+ "epoch": 15.0,
+ "grad_norm": 0.019997350871562958,
+ "learning_rate": 0.0006,
+ "loss": 4.691307544708252,
+ "step": 1080
+ },
+ {
+ "epoch": 15.0,
+ "eval_loss": 4.872478008270264,
+ "eval_runtime": 56.8513,
+ "eval_samples_per_second": 42.954,
+ "eval_steps_per_second": 1.354,
+ "step": 1080
+ },
+ {
+ "epoch": 15.013973799126637,
+ "grad_norm": 0.018830662593245506,
+ "learning_rate": 0.0006,
+ "loss": 4.761779308319092,
+ "step": 1081
+ },
+ {
+ "epoch": 15.027947598253276,
+ "grad_norm": 0.016876481473445892,
+ "learning_rate": 0.0006,
+ "loss": 4.61148738861084,
+ "step": 1082
+ },
+ {
+ "epoch": 15.041921397379912,
+ "grad_norm": 0.015550863929092884,
+ "learning_rate": 0.0006,
+ "loss": 4.676124095916748,
+ "step": 1083
+ },
+ {
+ "epoch": 15.055895196506551,
+ "grad_norm": 0.017178896814584732,
+ "learning_rate": 0.0006,
+ "loss": 4.588935852050781,
+ "step": 1084
+ },
+ {
+ "epoch": 15.069868995633188,
+ "grad_norm": 0.015465447679162025,
+ "learning_rate": 0.0006,
+ "loss": 4.605499267578125,
+ "step": 1085
+ },
+ {
+ "epoch": 15.083842794759825,
+ "grad_norm": 0.01606195978820324,
+ "learning_rate": 0.0006,
+ "loss": 4.723373889923096,
+ "step": 1086
+ },
+ {
+ "epoch": 15.097816593886463,
+ "grad_norm": 0.016521936282515526,
+ "learning_rate": 0.0006,
+ "loss": 4.646608352661133,
+ "step": 1087
+ },
+ {
+ "epoch": 15.1117903930131,
+ "grad_norm": 0.01589205674827099,
+ "learning_rate": 0.0006,
+ "loss": 4.498138427734375,
+ "step": 1088
+ },
+ {
+ "epoch": 15.125764192139737,
+ "grad_norm": 0.01524856686592102,
+ "learning_rate": 0.0006,
+ "loss": 4.594562530517578,
+ "step": 1089
+ },
+ {
+ "epoch": 15.139737991266376,
+ "grad_norm": 0.014714641496539116,
+ "learning_rate": 0.0006,
+ "loss": 4.630744934082031,
+ "step": 1090
+ },
+ {
+ "epoch": 15.153711790393013,
+ "grad_norm": 0.015333331190049648,
+ "learning_rate": 0.0006,
+ "loss": 4.682388782501221,
+ "step": 1091
+ },
+ {
+ "epoch": 15.167685589519651,
+ "grad_norm": 0.01550073828548193,
+ "learning_rate": 0.0006,
+ "loss": 4.692652702331543,
+ "step": 1092
+ },
+ {
+ "epoch": 15.181659388646288,
+ "grad_norm": 0.014783318154513836,
+ "learning_rate": 0.0006,
+ "loss": 4.6223835945129395,
+ "step": 1093
+ },
+ {
+ "epoch": 15.195633187772925,
+ "grad_norm": 0.014372751116752625,
+ "learning_rate": 0.0006,
+ "loss": 4.52687406539917,
+ "step": 1094
+ },
+ {
+ "epoch": 15.209606986899564,
+ "grad_norm": 0.014779130928218365,
+ "learning_rate": 0.0006,
+ "loss": 4.637009620666504,
+ "step": 1095
+ },
+ {
+ "epoch": 15.2235807860262,
+ "grad_norm": 0.015697073191404343,
+ "learning_rate": 0.0006,
+ "loss": 4.682981014251709,
+ "step": 1096
+ },
+ {
+ "epoch": 15.237554585152838,
+ "grad_norm": 0.016523664817214012,
+ "learning_rate": 0.0006,
+ "loss": 4.705722808837891,
+ "step": 1097
+ },
+ {
+ "epoch": 15.251528384279476,
+ "grad_norm": 0.017633996903896332,
+ "learning_rate": 0.0006,
+ "loss": 4.574686050415039,
+ "step": 1098
+ },
+ {
+ "epoch": 15.265502183406113,
+ "grad_norm": 0.019955076277256012,
+ "learning_rate": 0.0006,
+ "loss": 4.634450912475586,
+ "step": 1099
+ },
+ {
+ "epoch": 15.279475982532752,
+ "grad_norm": 0.02132618986070156,
+ "learning_rate": 0.0006,
+ "loss": 4.56110143661499,
+ "step": 1100
+ },
+ {
+ "epoch": 15.293449781659389,
+ "grad_norm": 0.0207593385130167,
+ "learning_rate": 0.0006,
+ "loss": 4.629396438598633,
+ "step": 1101
+ },
+ {
+ "epoch": 15.307423580786025,
+ "grad_norm": 0.017882540822029114,
+ "learning_rate": 0.0006,
+ "loss": 4.730963230133057,
+ "step": 1102
+ },
+ {
+ "epoch": 15.321397379912664,
+ "grad_norm": 0.01917876861989498,
+ "learning_rate": 0.0006,
+ "loss": 4.64585542678833,
+ "step": 1103
+ },
+ {
+ "epoch": 15.335371179039301,
+ "grad_norm": 0.01943974196910858,
+ "learning_rate": 0.0006,
+ "loss": 4.656181335449219,
+ "step": 1104
+ },
+ {
+ "epoch": 15.34934497816594,
+ "grad_norm": 0.019904715940356255,
+ "learning_rate": 0.0006,
+ "loss": 4.619329452514648,
+ "step": 1105
+ },
+ {
+ "epoch": 15.363318777292577,
+ "grad_norm": 0.019174346700310707,
+ "learning_rate": 0.0006,
+ "loss": 4.647892475128174,
+ "step": 1106
+ },
+ {
+ "epoch": 15.377292576419213,
+ "grad_norm": 0.018244758248329163,
+ "learning_rate": 0.0006,
+ "loss": 4.56166934967041,
+ "step": 1107
+ },
+ {
+ "epoch": 15.391266375545852,
+ "grad_norm": 0.018914205953478813,
+ "learning_rate": 0.0006,
+ "loss": 4.648063659667969,
+ "step": 1108
+ },
+ {
+ "epoch": 15.405240174672489,
+ "grad_norm": 0.01961071416735649,
+ "learning_rate": 0.0006,
+ "loss": 4.665525436401367,
+ "step": 1109
+ },
+ {
+ "epoch": 15.419213973799126,
+ "grad_norm": 0.01841311901807785,
+ "learning_rate": 0.0006,
+ "loss": 4.611342430114746,
+ "step": 1110
+ },
+ {
+ "epoch": 15.433187772925764,
+ "grad_norm": 0.01863427273929119,
+ "learning_rate": 0.0006,
+ "loss": 4.664256572723389,
+ "step": 1111
+ },
+ {
+ "epoch": 15.447161572052401,
+ "grad_norm": 0.020506877452135086,
+ "learning_rate": 0.0006,
+ "loss": 4.631237506866455,
+ "step": 1112
+ },
+ {
+ "epoch": 15.46113537117904,
+ "grad_norm": 0.02223833277821541,
+ "learning_rate": 0.0006,
+ "loss": 4.668654441833496,
+ "step": 1113
+ },
+ {
+ "epoch": 15.475109170305677,
+ "grad_norm": 0.023336702957749367,
+ "learning_rate": 0.0006,
+ "loss": 4.562833309173584,
+ "step": 1114
+ },
+ {
+ "epoch": 15.489082969432314,
+ "grad_norm": 0.023052413016557693,
+ "learning_rate": 0.0006,
+ "loss": 4.621532917022705,
+ "step": 1115
+ },
+ {
+ "epoch": 15.503056768558952,
+ "grad_norm": 0.023420924320816994,
+ "learning_rate": 0.0006,
+ "loss": 4.665465354919434,
+ "step": 1116
+ },
+ {
+ "epoch": 15.51703056768559,
+ "grad_norm": 0.024071281775832176,
+ "learning_rate": 0.0006,
+ "loss": 4.595980644226074,
+ "step": 1117
+ },
+ {
+ "epoch": 15.531004366812226,
+ "grad_norm": 0.023406682536005974,
+ "learning_rate": 0.0006,
+ "loss": 4.642635345458984,
+ "step": 1118
+ },
+ {
+ "epoch": 15.544978165938865,
+ "grad_norm": 0.02138647809624672,
+ "learning_rate": 0.0006,
+ "loss": 4.704513072967529,
+ "step": 1119
+ },
+ {
+ "epoch": 15.558951965065502,
+ "grad_norm": 0.0221384447067976,
+ "learning_rate": 0.0006,
+ "loss": 4.567503452301025,
+ "step": 1120
+ },
+ {
+ "epoch": 15.57292576419214,
+ "grad_norm": 0.020879851654171944,
+ "learning_rate": 0.0006,
+ "loss": 4.664057731628418,
+ "step": 1121
+ },
+ {
+ "epoch": 15.586899563318777,
+ "grad_norm": 0.021478967741131783,
+ "learning_rate": 0.0006,
+ "loss": 4.539627552032471,
+ "step": 1122
+ },
+ {
+ "epoch": 15.600873362445414,
+ "grad_norm": 0.022748656570911407,
+ "learning_rate": 0.0006,
+ "loss": 4.5654497146606445,
+ "step": 1123
+ },
+ {
+ "epoch": 15.614847161572053,
+ "grad_norm": 0.021870248019695282,
+ "learning_rate": 0.0006,
+ "loss": 4.56899356842041,
+ "step": 1124
+ },
+ {
+ "epoch": 15.62882096069869,
+ "grad_norm": 0.021711355075240135,
+ "learning_rate": 0.0006,
+ "loss": 4.657808303833008,
+ "step": 1125
+ },
+ {
+ "epoch": 15.642794759825328,
+ "grad_norm": 0.022810909897089005,
+ "learning_rate": 0.0006,
+ "loss": 4.6526689529418945,
+ "step": 1126
+ },
+ {
+ "epoch": 15.656768558951965,
+ "grad_norm": 0.022561343386769295,
+ "learning_rate": 0.0006,
+ "loss": 4.534887790679932,
+ "step": 1127
+ },
+ {
+ "epoch": 15.670742358078602,
+ "grad_norm": 0.019909802824258804,
+ "learning_rate": 0.0006,
+ "loss": 4.600790023803711,
+ "step": 1128
+ },
+ {
+ "epoch": 15.68471615720524,
+ "grad_norm": 0.020092325285077095,
+ "learning_rate": 0.0006,
+ "loss": 4.773830413818359,
+ "step": 1129
+ },
+ {
+ "epoch": 15.698689956331878,
+ "grad_norm": 0.018834445625543594,
+ "learning_rate": 0.0006,
+ "loss": 4.628868103027344,
+ "step": 1130
+ },
+ {
+ "epoch": 15.712663755458514,
+ "grad_norm": 0.017552798613905907,
+ "learning_rate": 0.0006,
+ "loss": 4.647500038146973,
+ "step": 1131
+ },
+ {
+ "epoch": 15.726637554585153,
+ "grad_norm": 0.01751251146197319,
+ "learning_rate": 0.0006,
+ "loss": 4.590545654296875,
+ "step": 1132
+ },
+ {
+ "epoch": 15.74061135371179,
+ "grad_norm": 0.01699099875986576,
+ "learning_rate": 0.0006,
+ "loss": 4.509699821472168,
+ "step": 1133
+ },
+ {
+ "epoch": 15.754585152838429,
+ "grad_norm": 0.015808649361133575,
+ "learning_rate": 0.0006,
+ "loss": 4.606992244720459,
+ "step": 1134
+ },
+ {
+ "epoch": 15.768558951965066,
+ "grad_norm": 0.01655443385243416,
+ "learning_rate": 0.0006,
+ "loss": 4.585219860076904,
+ "step": 1135
+ },
+ {
+ "epoch": 15.782532751091702,
+ "grad_norm": 0.01708163693547249,
+ "learning_rate": 0.0006,
+ "loss": 4.668147087097168,
+ "step": 1136
+ },
+ {
+ "epoch": 15.796506550218341,
+ "grad_norm": 0.017161913216114044,
+ "learning_rate": 0.0006,
+ "loss": 4.7170634269714355,
+ "step": 1137
+ },
+ {
+ "epoch": 15.810480349344978,
+ "grad_norm": 0.017262674868106842,
+ "learning_rate": 0.0006,
+ "loss": 4.572132110595703,
+ "step": 1138
+ },
+ {
+ "epoch": 15.824454148471617,
+ "grad_norm": 0.017713576555252075,
+ "learning_rate": 0.0006,
+ "loss": 4.702538967132568,
+ "step": 1139
+ },
+ {
+ "epoch": 15.838427947598253,
+ "grad_norm": 0.018730733543634415,
+ "learning_rate": 0.0006,
+ "loss": 4.629721641540527,
+ "step": 1140
+ },
+ {
+ "epoch": 15.85240174672489,
+ "grad_norm": 0.017960187047719955,
+ "learning_rate": 0.0006,
+ "loss": 4.661039352416992,
+ "step": 1141
+ },
+ {
+ "epoch": 15.866375545851529,
+ "grad_norm": 0.01750079356133938,
+ "learning_rate": 0.0006,
+ "loss": 4.503710746765137,
+ "step": 1142
+ },
+ {
+ "epoch": 15.880349344978166,
+ "grad_norm": 0.016630422323942184,
+ "learning_rate": 0.0006,
+ "loss": 4.6582136154174805,
+ "step": 1143
+ },
+ {
+ "epoch": 15.894323144104803,
+ "grad_norm": 0.015896832570433617,
+ "learning_rate": 0.0006,
+ "loss": 4.596693992614746,
+ "step": 1144
+ },
+ {
+ "epoch": 15.908296943231441,
+ "grad_norm": 0.01566407084465027,
+ "learning_rate": 0.0006,
+ "loss": 4.6337199211120605,
+ "step": 1145
+ },
+ {
+ "epoch": 15.922270742358078,
+ "grad_norm": 0.016823330894112587,
+ "learning_rate": 0.0006,
+ "loss": 4.686994552612305,
+ "step": 1146
+ },
+ {
+ "epoch": 15.936244541484717,
+ "grad_norm": 0.016208255663514137,
+ "learning_rate": 0.0006,
+ "loss": 4.6205549240112305,
+ "step": 1147
+ },
+ {
+ "epoch": 15.950218340611354,
+ "grad_norm": 0.0171707421541214,
+ "learning_rate": 0.0006,
+ "loss": 4.582968711853027,
+ "step": 1148
+ },
+ {
+ "epoch": 15.96419213973799,
+ "grad_norm": 0.017009446397423744,
+ "learning_rate": 0.0006,
+ "loss": 4.510274887084961,
+ "step": 1149
+ },
+ {
+ "epoch": 15.97816593886463,
+ "grad_norm": 0.01565811224281788,
+ "learning_rate": 0.0006,
+ "loss": 4.55834436416626,
+ "step": 1150
+ },
+ {
+ "epoch": 15.992139737991266,
+ "grad_norm": 0.015276037156581879,
+ "learning_rate": 0.0006,
+ "loss": 4.537870407104492,
+ "step": 1151
+ },
+ {
+ "epoch": 16.0,
+ "grad_norm": 0.017305882647633553,
+ "learning_rate": 0.0006,
+ "loss": 4.611305236816406,
+ "step": 1152
+ },
+ {
+ "epoch": 16.0,
+ "eval_loss": 4.775513172149658,
+ "eval_runtime": 56.7341,
+ "eval_samples_per_second": 43.043,
+ "eval_steps_per_second": 1.357,
+ "step": 1152
+ },
+ {
+ "epoch": 16.01397379912664,
+ "grad_norm": 0.016752436757087708,
+ "learning_rate": 0.0006,
+ "loss": 4.541312217712402,
+ "step": 1153
+ },
+ {
+ "epoch": 16.027947598253274,
+ "grad_norm": 0.01643134094774723,
+ "learning_rate": 0.0006,
+ "loss": 4.585402965545654,
+ "step": 1154
+ },
+ {
+ "epoch": 16.041921397379912,
+ "grad_norm": 0.01888720504939556,
+ "learning_rate": 0.0006,
+ "loss": 4.597825050354004,
+ "step": 1155
+ },
+ {
+ "epoch": 16.05589519650655,
+ "grad_norm": 0.022044632583856583,
+ "learning_rate": 0.0006,
+ "loss": 4.532772064208984,
+ "step": 1156
+ },
+ {
+ "epoch": 16.069868995633186,
+ "grad_norm": 0.022986987605690956,
+ "learning_rate": 0.0006,
+ "loss": 4.551351547241211,
+ "step": 1157
+ },
+ {
+ "epoch": 16.083842794759825,
+ "grad_norm": 0.020116791129112244,
+ "learning_rate": 0.0006,
+ "loss": 4.484149932861328,
+ "step": 1158
+ },
+ {
+ "epoch": 16.097816593886463,
+ "grad_norm": 0.021442247554659843,
+ "learning_rate": 0.0006,
+ "loss": 4.515316486358643,
+ "step": 1159
+ },
+ {
+ "epoch": 16.111790393013102,
+ "grad_norm": 0.020974772050976753,
+ "learning_rate": 0.0006,
+ "loss": 4.51606559753418,
+ "step": 1160
+ },
+ {
+ "epoch": 16.125764192139737,
+ "grad_norm": 0.021274864673614502,
+ "learning_rate": 0.0006,
+ "loss": 4.587329864501953,
+ "step": 1161
+ },
+ {
+ "epoch": 16.139737991266376,
+ "grad_norm": 0.022799784317612648,
+ "learning_rate": 0.0006,
+ "loss": 4.568149089813232,
+ "step": 1162
+ },
+ {
+ "epoch": 16.153711790393015,
+ "grad_norm": 0.021784933283925056,
+ "learning_rate": 0.0006,
+ "loss": 4.515336990356445,
+ "step": 1163
+ },
+ {
+ "epoch": 16.16768558951965,
+ "grad_norm": 0.024261271581053734,
+ "learning_rate": 0.0006,
+ "loss": 4.5127081871032715,
+ "step": 1164
+ },
+ {
+ "epoch": 16.18165938864629,
+ "grad_norm": 0.02248375490307808,
+ "learning_rate": 0.0006,
+ "loss": 4.754251003265381,
+ "step": 1165
+ },
+ {
+ "epoch": 16.195633187772927,
+ "grad_norm": 0.02136492356657982,
+ "learning_rate": 0.0006,
+ "loss": 4.58193826675415,
+ "step": 1166
+ },
+ {
+ "epoch": 16.209606986899562,
+ "grad_norm": 0.020727066323161125,
+ "learning_rate": 0.0006,
+ "loss": 4.564209938049316,
+ "step": 1167
+ },
+ {
+ "epoch": 16.2235807860262,
+ "grad_norm": 0.017789119854569435,
+ "learning_rate": 0.0006,
+ "loss": 4.698680877685547,
+ "step": 1168
+ },
+ {
+ "epoch": 16.23755458515284,
+ "grad_norm": 0.017603034153580666,
+ "learning_rate": 0.0006,
+ "loss": 4.501171588897705,
+ "step": 1169
+ },
+ {
+ "epoch": 16.251528384279474,
+ "grad_norm": 0.017345862463116646,
+ "learning_rate": 0.0006,
+ "loss": 4.593590259552002,
+ "step": 1170
+ },
+ {
+ "epoch": 16.265502183406113,
+ "grad_norm": 0.018047425895929337,
+ "learning_rate": 0.0006,
+ "loss": 4.597467422485352,
+ "step": 1171
+ },
+ {
+ "epoch": 16.27947598253275,
+ "grad_norm": 0.01947222836315632,
+ "learning_rate": 0.0006,
+ "loss": 4.5072221755981445,
+ "step": 1172
+ },
+ {
+ "epoch": 16.29344978165939,
+ "grad_norm": 0.019598500803112984,
+ "learning_rate": 0.0006,
+ "loss": 4.601351737976074,
+ "step": 1173
+ },
+ {
+ "epoch": 16.307423580786025,
+ "grad_norm": 0.020002514123916626,
+ "learning_rate": 0.0006,
+ "loss": 4.580431938171387,
+ "step": 1174
+ },
+ {
+ "epoch": 16.321397379912664,
+ "grad_norm": 0.0216530729085207,
+ "learning_rate": 0.0006,
+ "loss": 4.424681663513184,
+ "step": 1175
+ },
+ {
+ "epoch": 16.335371179039303,
+ "grad_norm": 0.01949235238134861,
+ "learning_rate": 0.0006,
+ "loss": 4.751883029937744,
+ "step": 1176
+ },
+ {
+ "epoch": 16.349344978165938,
+ "grad_norm": 0.017779016867280006,
+ "learning_rate": 0.0006,
+ "loss": 4.541490077972412,
+ "step": 1177
+ },
+ {
+ "epoch": 16.363318777292577,
+ "grad_norm": 0.01909121684730053,
+ "learning_rate": 0.0006,
+ "loss": 4.594141006469727,
+ "step": 1178
+ },
+ {
+ "epoch": 16.377292576419215,
+ "grad_norm": 0.019894849509000778,
+ "learning_rate": 0.0006,
+ "loss": 4.555070400238037,
+ "step": 1179
+ },
+ {
+ "epoch": 16.39126637554585,
+ "grad_norm": 0.01993032731115818,
+ "learning_rate": 0.0006,
+ "loss": 4.521401405334473,
+ "step": 1180
+ },
+ {
+ "epoch": 16.40524017467249,
+ "grad_norm": 0.0210788045078516,
+ "learning_rate": 0.0006,
+ "loss": 4.672842979431152,
+ "step": 1181
+ },
+ {
+ "epoch": 16.419213973799128,
+ "grad_norm": 0.02274598926305771,
+ "learning_rate": 0.0006,
+ "loss": 4.629504203796387,
+ "step": 1182
+ },
+ {
+ "epoch": 16.433187772925763,
+ "grad_norm": 0.020448120310902596,
+ "learning_rate": 0.0006,
+ "loss": 4.548209190368652,
+ "step": 1183
+ },
+ {
+ "epoch": 16.4471615720524,
+ "grad_norm": 0.017565643414855003,
+ "learning_rate": 0.0006,
+ "loss": 4.529054641723633,
+ "step": 1184
+ },
+ {
+ "epoch": 16.46113537117904,
+ "grad_norm": 0.019504237920045853,
+ "learning_rate": 0.0006,
+ "loss": 4.48182487487793,
+ "step": 1185
+ },
+ {
+ "epoch": 16.475109170305675,
+ "grad_norm": 0.016823742538690567,
+ "learning_rate": 0.0006,
+ "loss": 4.608696937561035,
+ "step": 1186
+ },
+ {
+ "epoch": 16.489082969432314,
+ "grad_norm": 0.01602894812822342,
+ "learning_rate": 0.0006,
+ "loss": 4.620062351226807,
+ "step": 1187
+ },
+ {
+ "epoch": 16.503056768558952,
+ "grad_norm": 0.016615642234683037,
+ "learning_rate": 0.0006,
+ "loss": 4.69378662109375,
+ "step": 1188
+ },
+ {
+ "epoch": 16.51703056768559,
+ "grad_norm": 0.018576730042696,
+ "learning_rate": 0.0006,
+ "loss": 4.537041664123535,
+ "step": 1189
+ },
+ {
+ "epoch": 16.531004366812226,
+ "grad_norm": 0.018611151725053787,
+ "learning_rate": 0.0006,
+ "loss": 4.496169090270996,
+ "step": 1190
+ },
+ {
+ "epoch": 16.544978165938865,
+ "grad_norm": 0.018622413277626038,
+ "learning_rate": 0.0006,
+ "loss": 4.557358264923096,
+ "step": 1191
+ },
+ {
+ "epoch": 16.558951965065503,
+ "grad_norm": 0.01823544315993786,
+ "learning_rate": 0.0006,
+ "loss": 4.730792999267578,
+ "step": 1192
+ },
+ {
+ "epoch": 16.57292576419214,
+ "grad_norm": 0.018796846270561218,
+ "learning_rate": 0.0006,
+ "loss": 4.781791687011719,
+ "step": 1193
+ },
+ {
+ "epoch": 16.586899563318777,
+ "grad_norm": 0.01728292554616928,
+ "learning_rate": 0.0006,
+ "loss": 4.515257835388184,
+ "step": 1194
+ },
+ {
+ "epoch": 16.600873362445416,
+ "grad_norm": 0.01699589006602764,
+ "learning_rate": 0.0006,
+ "loss": 4.5996904373168945,
+ "step": 1195
+ },
+ {
+ "epoch": 16.61484716157205,
+ "grad_norm": 0.015905970707535744,
+ "learning_rate": 0.0006,
+ "loss": 4.601301193237305,
+ "step": 1196
+ },
+ {
+ "epoch": 16.62882096069869,
+ "grad_norm": 0.01338571310043335,
+ "learning_rate": 0.0006,
+ "loss": 4.557708740234375,
+ "step": 1197
+ },
+ {
+ "epoch": 16.64279475982533,
+ "grad_norm": 0.015298855490982533,
+ "learning_rate": 0.0006,
+ "loss": 4.542287349700928,
+ "step": 1198
+ },
+ {
+ "epoch": 16.656768558951963,
+ "grad_norm": 0.014753367751836777,
+ "learning_rate": 0.0006,
+ "loss": 4.595787048339844,
+ "step": 1199
+ },
+ {
+ "epoch": 16.670742358078602,
+ "grad_norm": 0.015527483075857162,
+ "learning_rate": 0.0006,
+ "loss": 4.607529163360596,
+ "step": 1200
+ },
+ {
+ "epoch": 16.68471615720524,
+ "grad_norm": 0.015042162500321865,
+ "learning_rate": 0.0006,
+ "loss": 4.686626434326172,
+ "step": 1201
+ },
+ {
+ "epoch": 16.69868995633188,
+ "grad_norm": 0.0159393772482872,
+ "learning_rate": 0.0006,
+ "loss": 4.47656774520874,
+ "step": 1202
+ },
+ {
+ "epoch": 16.712663755458514,
+ "grad_norm": 0.01532787922769785,
+ "learning_rate": 0.0006,
+ "loss": 4.55678653717041,
+ "step": 1203
+ },
+ {
+ "epoch": 16.726637554585153,
+ "grad_norm": 0.016332164406776428,
+ "learning_rate": 0.0006,
+ "loss": 4.51441764831543,
+ "step": 1204
+ },
+ {
+ "epoch": 16.74061135371179,
+ "grad_norm": 0.018901726230978966,
+ "learning_rate": 0.0006,
+ "loss": 4.57227897644043,
+ "step": 1205
+ },
+ {
+ "epoch": 16.754585152838427,
+ "grad_norm": 0.018589433282613754,
+ "learning_rate": 0.0006,
+ "loss": 4.550070762634277,
+ "step": 1206
+ },
+ {
+ "epoch": 16.768558951965066,
+ "grad_norm": 0.01634034886956215,
+ "learning_rate": 0.0006,
+ "loss": 4.586910724639893,
+ "step": 1207
+ },
+ {
+ "epoch": 16.782532751091704,
+ "grad_norm": 0.017599433660507202,
+ "learning_rate": 0.0006,
+ "loss": 4.5891923904418945,
+ "step": 1208
+ },
+ {
+ "epoch": 16.79650655021834,
+ "grad_norm": 0.01878446340560913,
+ "learning_rate": 0.0006,
+ "loss": 4.57227087020874,
+ "step": 1209
+ },
+ {
+ "epoch": 16.810480349344978,
+ "grad_norm": 0.01803469844162464,
+ "learning_rate": 0.0006,
+ "loss": 4.583845138549805,
+ "step": 1210
+ },
+ {
+ "epoch": 16.824454148471617,
+ "grad_norm": 0.018505144864320755,
+ "learning_rate": 0.0006,
+ "loss": 4.561151504516602,
+ "step": 1211
+ },
+ {
+ "epoch": 16.83842794759825,
+ "grad_norm": 0.01920081488788128,
+ "learning_rate": 0.0006,
+ "loss": 4.506834030151367,
+ "step": 1212
+ },
+ {
+ "epoch": 16.85240174672489,
+ "grad_norm": 0.0177712831646204,
+ "learning_rate": 0.0006,
+ "loss": 4.557195663452148,
+ "step": 1213
+ },
+ {
+ "epoch": 16.86637554585153,
+ "grad_norm": 0.019113918766379356,
+ "learning_rate": 0.0006,
+ "loss": 4.651227951049805,
+ "step": 1214
+ },
+ {
+ "epoch": 16.880349344978168,
+ "grad_norm": 0.02012539468705654,
+ "learning_rate": 0.0006,
+ "loss": 4.4952239990234375,
+ "step": 1215
+ },
+ {
+ "epoch": 16.894323144104803,
+ "grad_norm": 0.019993646070361137,
+ "learning_rate": 0.0006,
+ "loss": 4.517433166503906,
+ "step": 1216
+ },
+ {
+ "epoch": 16.90829694323144,
+ "grad_norm": 0.020041441544890404,
+ "learning_rate": 0.0006,
+ "loss": 4.489407539367676,
+ "step": 1217
+ },
+ {
+ "epoch": 16.92227074235808,
+ "grad_norm": 0.0183611661195755,
+ "learning_rate": 0.0006,
+ "loss": 4.5435638427734375,
+ "step": 1218
+ },
+ {
+ "epoch": 16.936244541484715,
+ "grad_norm": 0.019138429313898087,
+ "learning_rate": 0.0006,
+ "loss": 4.545949459075928,
+ "step": 1219
+ },
+ {
+ "epoch": 16.950218340611354,
+ "grad_norm": 0.018554266542196274,
+ "learning_rate": 0.0006,
+ "loss": 4.579620361328125,
+ "step": 1220
+ },
+ {
+ "epoch": 16.964192139737992,
+ "grad_norm": 0.0193545650690794,
+ "learning_rate": 0.0006,
+ "loss": 4.588255405426025,
+ "step": 1221
+ },
+ {
+ "epoch": 16.978165938864628,
+ "grad_norm": 0.019389165565371513,
+ "learning_rate": 0.0006,
+ "loss": 4.520698547363281,
+ "step": 1222
+ },
+ {
+ "epoch": 16.992139737991266,
+ "grad_norm": 0.016601379960775375,
+ "learning_rate": 0.0006,
+ "loss": 4.4809064865112305,
+ "step": 1223
+ },
+ {
+ "epoch": 17.0,
+ "grad_norm": 0.01771303080022335,
+ "learning_rate": 0.0006,
+ "loss": 4.721122741699219,
+ "step": 1224
+ },
+ {
+ "epoch": 17.0,
+ "eval_loss": 4.7903666496276855,
+ "eval_runtime": 56.9826,
+ "eval_samples_per_second": 42.855,
+ "eval_steps_per_second": 1.351,
+ "step": 1224
+ },
+ {
+ "epoch": 17.01397379912664,
+ "grad_norm": 0.017720289528369904,
+ "learning_rate": 0.0006,
+ "loss": 4.523834705352783,
+ "step": 1225
+ },
+ {
+ "epoch": 17.027947598253274,
+ "grad_norm": 0.020137161016464233,
+ "learning_rate": 0.0006,
+ "loss": 4.574061870574951,
+ "step": 1226
+ },
+ {
+ "epoch": 17.041921397379912,
+ "grad_norm": 0.02002314105629921,
+ "learning_rate": 0.0006,
+ "loss": 4.512407302856445,
+ "step": 1227
+ },
+ {
+ "epoch": 17.05589519650655,
+ "grad_norm": 0.019628183916211128,
+ "learning_rate": 0.0006,
+ "loss": 4.502533435821533,
+ "step": 1228
+ },
+ {
+ "epoch": 17.069868995633186,
+ "grad_norm": 0.019808458164334297,
+ "learning_rate": 0.0006,
+ "loss": 4.463156700134277,
+ "step": 1229
+ },
+ {
+ "epoch": 17.083842794759825,
+ "grad_norm": 0.021250786259770393,
+ "learning_rate": 0.0006,
+ "loss": 4.622682094573975,
+ "step": 1230
+ },
+ {
+ "epoch": 17.097816593886463,
+ "grad_norm": 0.022031694650650024,
+ "learning_rate": 0.0006,
+ "loss": 4.451882839202881,
+ "step": 1231
+ },
+ {
+ "epoch": 17.111790393013102,
+ "grad_norm": 0.01898987777531147,
+ "learning_rate": 0.0006,
+ "loss": 4.529331684112549,
+ "step": 1232
+ },
+ {
+ "epoch": 17.125764192139737,
+ "grad_norm": 0.018749834969639778,
+ "learning_rate": 0.0006,
+ "loss": 4.435815334320068,
+ "step": 1233
+ },
+ {
+ "epoch": 17.139737991266376,
+ "grad_norm": 0.01896674931049347,
+ "learning_rate": 0.0006,
+ "loss": 4.484029293060303,
+ "step": 1234
+ },
+ {
+ "epoch": 17.153711790393015,
+ "grad_norm": 0.018286997452378273,
+ "learning_rate": 0.0006,
+ "loss": 4.523510932922363,
+ "step": 1235
+ },
+ {
+ "epoch": 17.16768558951965,
+ "grad_norm": 0.017273059114813805,
+ "learning_rate": 0.0006,
+ "loss": 4.451767444610596,
+ "step": 1236
+ },
+ {
+ "epoch": 17.18165938864629,
+ "grad_norm": 0.016499150544404984,
+ "learning_rate": 0.0006,
+ "loss": 4.5568647384643555,
+ "step": 1237
+ },
+ {
+ "epoch": 17.195633187772927,
+ "grad_norm": 0.015934469178318977,
+ "learning_rate": 0.0006,
+ "loss": 4.4639997482299805,
+ "step": 1238
+ },
+ {
+ "epoch": 17.209606986899562,
+ "grad_norm": 0.01714209094643593,
+ "learning_rate": 0.0006,
+ "loss": 4.547614097595215,
+ "step": 1239
+ },
+ {
+ "epoch": 17.2235807860262,
+ "grad_norm": 0.019167495891451836,
+ "learning_rate": 0.0006,
+ "loss": 4.571097373962402,
+ "step": 1240
+ },
+ {
+ "epoch": 17.23755458515284,
+ "grad_norm": 0.018276996910572052,
+ "learning_rate": 0.0006,
+ "loss": 4.489764213562012,
+ "step": 1241
+ },
+ {
+ "epoch": 17.251528384279474,
+ "grad_norm": 0.021296866238117218,
+ "learning_rate": 0.0006,
+ "loss": 4.526913642883301,
+ "step": 1242
+ },
+ {
+ "epoch": 17.265502183406113,
+ "grad_norm": 0.022606629878282547,
+ "learning_rate": 0.0006,
+ "loss": 4.521624565124512,
+ "step": 1243
+ },
+ {
+ "epoch": 17.27947598253275,
+ "grad_norm": 0.021204711869359016,
+ "learning_rate": 0.0006,
+ "loss": 4.500864028930664,
+ "step": 1244
+ },
+ {
+ "epoch": 17.29344978165939,
+ "grad_norm": 0.01988849975168705,
+ "learning_rate": 0.0006,
+ "loss": 4.45162296295166,
+ "step": 1245
+ },
+ {
+ "epoch": 17.307423580786025,
+ "grad_norm": 0.020268384367227554,
+ "learning_rate": 0.0006,
+ "loss": 4.536198616027832,
+ "step": 1246
+ },
+ {
+ "epoch": 17.321397379912664,
+ "grad_norm": 0.020185677334666252,
+ "learning_rate": 0.0006,
+ "loss": 4.440262794494629,
+ "step": 1247
+ },
+ {
+ "epoch": 17.335371179039303,
+ "grad_norm": 0.020106853917241096,
+ "learning_rate": 0.0006,
+ "loss": 4.475907325744629,
+ "step": 1248
+ },
+ {
+ "epoch": 17.349344978165938,
+ "grad_norm": 0.018883297219872475,
+ "learning_rate": 0.0006,
+ "loss": 4.640661716461182,
+ "step": 1249
+ },
+ {
+ "epoch": 17.363318777292577,
+ "grad_norm": 0.01852620206773281,
+ "learning_rate": 0.0006,
+ "loss": 4.578243255615234,
+ "step": 1250
+ },
+ {
+ "epoch": 17.377292576419215,
+ "grad_norm": 0.01886451430618763,
+ "learning_rate": 0.0006,
+ "loss": 4.546026229858398,
+ "step": 1251
+ },
+ {
+ "epoch": 17.39126637554585,
+ "grad_norm": 0.01777452416718006,
+ "learning_rate": 0.0006,
+ "loss": 4.521542549133301,
+ "step": 1252
+ },
+ {
+ "epoch": 17.40524017467249,
+ "grad_norm": 0.016626447439193726,
+ "learning_rate": 0.0006,
+ "loss": 4.52629280090332,
+ "step": 1253
+ },
+ {
+ "epoch": 17.419213973799128,
+ "grad_norm": 0.01599477231502533,
+ "learning_rate": 0.0006,
+ "loss": 4.4796037673950195,
+ "step": 1254
+ },
+ {
+ "epoch": 17.433187772925763,
+ "grad_norm": 0.017021458595991135,
+ "learning_rate": 0.0006,
+ "loss": 4.468813419342041,
+ "step": 1255
+ },
+ {
+ "epoch": 17.4471615720524,
+ "grad_norm": 0.016925333067774773,
+ "learning_rate": 0.0006,
+ "loss": 4.584898948669434,
+ "step": 1256
+ },
+ {
+ "epoch": 17.46113537117904,
+ "grad_norm": 0.01714998669922352,
+ "learning_rate": 0.0006,
+ "loss": 4.523544788360596,
+ "step": 1257
+ },
+ {
+ "epoch": 17.475109170305675,
+ "grad_norm": 0.01802288182079792,
+ "learning_rate": 0.0006,
+ "loss": 4.595581531524658,
+ "step": 1258
+ },
+ {
+ "epoch": 17.489082969432314,
+ "grad_norm": 0.0180289875715971,
+ "learning_rate": 0.0006,
+ "loss": 4.394440650939941,
+ "step": 1259
+ },
+ {
+ "epoch": 17.503056768558952,
+ "grad_norm": 0.018504245206713676,
+ "learning_rate": 0.0006,
+ "loss": 4.4794416427612305,
+ "step": 1260
+ },
+ {
+ "epoch": 17.51703056768559,
+ "grad_norm": 0.019973233342170715,
+ "learning_rate": 0.0006,
+ "loss": 4.4241437911987305,
+ "step": 1261
+ },
+ {
+ "epoch": 17.531004366812226,
+ "grad_norm": 0.020063083618879318,
+ "learning_rate": 0.0006,
+ "loss": 4.48068380355835,
+ "step": 1262
+ },
+ {
+ "epoch": 17.544978165938865,
+ "grad_norm": 0.02066640742123127,
+ "learning_rate": 0.0006,
+ "loss": 4.628993034362793,
+ "step": 1263
+ },
+ {
+ "epoch": 17.558951965065503,
+ "grad_norm": 0.02206416428089142,
+ "learning_rate": 0.0006,
+ "loss": 4.564790725708008,
+ "step": 1264
+ },
+ {
+ "epoch": 17.57292576419214,
+ "grad_norm": 0.020640335977077484,
+ "learning_rate": 0.0006,
+ "loss": 4.578705787658691,
+ "step": 1265
+ },
+ {
+ "epoch": 17.586899563318777,
+ "grad_norm": 0.019093405455350876,
+ "learning_rate": 0.0006,
+ "loss": 4.5023603439331055,
+ "step": 1266
+ },
+ {
+ "epoch": 17.600873362445416,
+ "grad_norm": 0.019459193572402,
+ "learning_rate": 0.0006,
+ "loss": 4.594531059265137,
+ "step": 1267
+ },
+ {
+ "epoch": 17.61484716157205,
+ "grad_norm": 0.02021171525120735,
+ "learning_rate": 0.0006,
+ "loss": 4.770299911499023,
+ "step": 1268
+ },
+ {
+ "epoch": 17.62882096069869,
+ "grad_norm": 0.021161191165447235,
+ "learning_rate": 0.0006,
+ "loss": 4.5380401611328125,
+ "step": 1269
+ },
+ {
+ "epoch": 17.64279475982533,
+ "grad_norm": 0.021013403311371803,
+ "learning_rate": 0.0006,
+ "loss": 4.455524921417236,
+ "step": 1270
+ },
+ {
+ "epoch": 17.656768558951963,
+ "grad_norm": 0.020336592569947243,
+ "learning_rate": 0.0006,
+ "loss": 4.533672332763672,
+ "step": 1271
+ },
+ {
+ "epoch": 17.670742358078602,
+ "grad_norm": 0.020159434527158737,
+ "learning_rate": 0.0006,
+ "loss": 4.596151351928711,
+ "step": 1272
+ },
+ {
+ "epoch": 17.68471615720524,
+ "grad_norm": 0.020432960242033005,
+ "learning_rate": 0.0006,
+ "loss": 4.55531644821167,
+ "step": 1273
+ },
+ {
+ "epoch": 17.69868995633188,
+ "grad_norm": 0.025466229766607285,
+ "learning_rate": 0.0006,
+ "loss": 4.566912651062012,
+ "step": 1274
+ },
+ {
+ "epoch": 17.712663755458514,
+ "grad_norm": 0.026537558063864708,
+ "learning_rate": 0.0006,
+ "loss": 4.526003837585449,
+ "step": 1275
+ },
+ {
+ "epoch": 17.726637554585153,
+ "grad_norm": 0.024397607892751694,
+ "learning_rate": 0.0006,
+ "loss": 4.541755676269531,
+ "step": 1276
+ },
+ {
+ "epoch": 17.74061135371179,
+ "grad_norm": 0.022062312811613083,
+ "learning_rate": 0.0006,
+ "loss": 4.6532793045043945,
+ "step": 1277
+ },
+ {
+ "epoch": 17.754585152838427,
+ "grad_norm": 0.02204282023012638,
+ "learning_rate": 0.0006,
+ "loss": 4.582825660705566,
+ "step": 1278
+ },
+ {
+ "epoch": 17.768558951965066,
+ "grad_norm": 0.020844416692852974,
+ "learning_rate": 0.0006,
+ "loss": 4.522336483001709,
+ "step": 1279
+ },
+ {
+ "epoch": 17.782532751091704,
+ "grad_norm": 0.019815703853964806,
+ "learning_rate": 0.0006,
+ "loss": 4.607398509979248,
+ "step": 1280
+ },
+ {
+ "epoch": 17.79650655021834,
+ "grad_norm": 0.016676954925060272,
+ "learning_rate": 0.0006,
+ "loss": 4.446206092834473,
+ "step": 1281
+ },
+ {
+ "epoch": 17.810480349344978,
+ "grad_norm": 0.0166452769190073,
+ "learning_rate": 0.0006,
+ "loss": 4.505931854248047,
+ "step": 1282
+ },
+ {
+ "epoch": 17.824454148471617,
+ "grad_norm": 0.015928907319903374,
+ "learning_rate": 0.0006,
+ "loss": 4.455267429351807,
+ "step": 1283
+ },
+ {
+ "epoch": 17.83842794759825,
+ "grad_norm": 0.015219560824334621,
+ "learning_rate": 0.0006,
+ "loss": 4.5623674392700195,
+ "step": 1284
+ },
+ {
+ "epoch": 17.85240174672489,
+ "grad_norm": 0.015125355683267117,
+ "learning_rate": 0.0006,
+ "loss": 4.554104804992676,
+ "step": 1285
+ },
+ {
+ "epoch": 17.86637554585153,
+ "grad_norm": 0.015612718649208546,
+ "learning_rate": 0.0006,
+ "loss": 4.484847068786621,
+ "step": 1286
+ },
+ {
+ "epoch": 17.880349344978168,
+ "grad_norm": 0.015215662308037281,
+ "learning_rate": 0.0006,
+ "loss": 4.4802093505859375,
+ "step": 1287
+ },
+ {
+ "epoch": 17.894323144104803,
+ "grad_norm": 0.01511684525758028,
+ "learning_rate": 0.0006,
+ "loss": 4.537298679351807,
+ "step": 1288
+ },
+ {
+ "epoch": 17.90829694323144,
+ "grad_norm": 0.013841322623193264,
+ "learning_rate": 0.0006,
+ "loss": 4.427488327026367,
+ "step": 1289
+ },
+ {
+ "epoch": 17.92227074235808,
+ "grad_norm": 0.015115964226424694,
+ "learning_rate": 0.0006,
+ "loss": 4.4547119140625,
+ "step": 1290
+ },
+ {
+ "epoch": 17.936244541484715,
+ "grad_norm": 0.016411006450653076,
+ "learning_rate": 0.0006,
+ "loss": 4.52994966506958,
+ "step": 1291
+ },
+ {
+ "epoch": 17.950218340611354,
+ "grad_norm": 0.017313152551651,
+ "learning_rate": 0.0006,
+ "loss": 4.5337653160095215,
+ "step": 1292
+ },
+ {
+ "epoch": 17.964192139737992,
+ "grad_norm": 0.016559796407818794,
+ "learning_rate": 0.0006,
+ "loss": 4.5277910232543945,
+ "step": 1293
+ },
+ {
+ "epoch": 17.978165938864628,
+ "grad_norm": 0.015854761004447937,
+ "learning_rate": 0.0006,
+ "loss": 4.6334309577941895,
+ "step": 1294
+ },
+ {
+ "epoch": 17.992139737991266,
+ "grad_norm": 0.015658238902688026,
+ "learning_rate": 0.0006,
+ "loss": 4.613372802734375,
+ "step": 1295
+ },
+ {
+ "epoch": 18.0,
+ "grad_norm": 0.0169155802577734,
+ "learning_rate": 0.0006,
+ "loss": 4.527356147766113,
+ "step": 1296
+ },
+ {
+ "epoch": 18.0,
+ "eval_loss": 4.764543056488037,
+ "eval_runtime": 56.5998,
+ "eval_samples_per_second": 43.145,
+ "eval_steps_per_second": 1.36,
+ "step": 1296
+ },
+ {
+ "epoch": 18.01397379912664,
+ "grad_norm": 0.015348963439464569,
+ "learning_rate": 0.0006,
+ "loss": 4.466423988342285,
+ "step": 1297
+ },
+ {
+ "epoch": 18.027947598253274,
+ "grad_norm": 0.01768588088452816,
+ "learning_rate": 0.0006,
+ "loss": 4.545933246612549,
+ "step": 1298
+ },
+ {
+ "epoch": 18.041921397379912,
+ "grad_norm": 0.017381226643919945,
+ "learning_rate": 0.0006,
+ "loss": 4.494060039520264,
+ "step": 1299
+ },
+ {
+ "epoch": 18.05589519650655,
+ "grad_norm": 0.019002335146069527,
+ "learning_rate": 0.0006,
+ "loss": 4.395200729370117,
+ "step": 1300
+ },
+ {
+ "epoch": 18.069868995633186,
+ "grad_norm": 0.01827932707965374,
+ "learning_rate": 0.0006,
+ "loss": 4.424848556518555,
+ "step": 1301
+ },
+ {
+ "epoch": 18.083842794759825,
+ "grad_norm": 0.016383031383156776,
+ "learning_rate": 0.0006,
+ "loss": 4.494722366333008,
+ "step": 1302
+ },
+ {
+ "epoch": 18.097816593886463,
+ "grad_norm": 0.016265157610177994,
+ "learning_rate": 0.0006,
+ "loss": 4.416522979736328,
+ "step": 1303
+ },
+ {
+ "epoch": 18.111790393013102,
+ "grad_norm": 0.01702386513352394,
+ "learning_rate": 0.0006,
+ "loss": 4.600103378295898,
+ "step": 1304
+ },
+ {
+ "epoch": 18.125764192139737,
+ "grad_norm": 0.018495498225092888,
+ "learning_rate": 0.0006,
+ "loss": 4.588401794433594,
+ "step": 1305
+ },
+ {
+ "epoch": 18.139737991266376,
+ "grad_norm": 0.017921684309840202,
+ "learning_rate": 0.0006,
+ "loss": 4.478075981140137,
+ "step": 1306
+ },
+ {
+ "epoch": 18.153711790393015,
+ "grad_norm": 0.01801278069615364,
+ "learning_rate": 0.0006,
+ "loss": 4.413186550140381,
+ "step": 1307
+ },
+ {
+ "epoch": 18.16768558951965,
+ "grad_norm": 0.017211275175213814,
+ "learning_rate": 0.0006,
+ "loss": 4.461341857910156,
+ "step": 1308
+ },
+ {
+ "epoch": 18.18165938864629,
+ "grad_norm": 0.017589328810572624,
+ "learning_rate": 0.0006,
+ "loss": 4.546969413757324,
+ "step": 1309
+ },
+ {
+ "epoch": 18.195633187772927,
+ "grad_norm": 0.01797698810696602,
+ "learning_rate": 0.0006,
+ "loss": 4.530797004699707,
+ "step": 1310
+ },
+ {
+ "epoch": 18.209606986899562,
+ "grad_norm": 0.020408619195222855,
+ "learning_rate": 0.0006,
+ "loss": 4.62999153137207,
+ "step": 1311
+ },
+ {
+ "epoch": 18.2235807860262,
+ "grad_norm": 0.01994038000702858,
+ "learning_rate": 0.0006,
+ "loss": 4.586811065673828,
+ "step": 1312
+ },
+ {
+ "epoch": 18.23755458515284,
+ "grad_norm": 0.019381340593099594,
+ "learning_rate": 0.0006,
+ "loss": 4.4769368171691895,
+ "step": 1313
+ },
+ {
+ "epoch": 18.251528384279474,
+ "grad_norm": 0.020484555512666702,
+ "learning_rate": 0.0006,
+ "loss": 4.524302959442139,
+ "step": 1314
+ },
+ {
+ "epoch": 18.265502183406113,
+ "grad_norm": 0.02426416426897049,
+ "learning_rate": 0.0006,
+ "loss": 4.561705589294434,
+ "step": 1315
+ },
+ {
+ "epoch": 18.27947598253275,
+ "grad_norm": 0.024748101830482483,
+ "learning_rate": 0.0006,
+ "loss": 4.4654316902160645,
+ "step": 1316
+ },
+ {
+ "epoch": 18.29344978165939,
+ "grad_norm": 0.020703328773379326,
+ "learning_rate": 0.0006,
+ "loss": 4.53033447265625,
+ "step": 1317
+ },
+ {
+ "epoch": 18.307423580786025,
+ "grad_norm": 0.01765458658337593,
+ "learning_rate": 0.0006,
+ "loss": 4.5154266357421875,
+ "step": 1318
+ },
+ {
+ "epoch": 18.321397379912664,
+ "grad_norm": 0.018646420910954475,
+ "learning_rate": 0.0006,
+ "loss": 4.5224714279174805,
+ "step": 1319
+ },
+ {
+ "epoch": 18.335371179039303,
+ "grad_norm": 0.01907474733889103,
+ "learning_rate": 0.0006,
+ "loss": 4.5055437088012695,
+ "step": 1320
+ },
+ {
+ "epoch": 18.349344978165938,
+ "grad_norm": 0.018179846927523613,
+ "learning_rate": 0.0006,
+ "loss": 4.462161064147949,
+ "step": 1321
+ },
+ {
+ "epoch": 18.363318777292577,
+ "grad_norm": 0.016697878018021584,
+ "learning_rate": 0.0006,
+ "loss": 4.473718643188477,
+ "step": 1322
+ },
+ {
+ "epoch": 18.377292576419215,
+ "grad_norm": 0.017504019662737846,
+ "learning_rate": 0.0006,
+ "loss": 4.474348068237305,
+ "step": 1323
+ },
+ {
+ "epoch": 18.39126637554585,
+ "grad_norm": 0.018889913335442543,
+ "learning_rate": 0.0006,
+ "loss": 4.4531636238098145,
+ "step": 1324
+ },
+ {
+ "epoch": 18.40524017467249,
+ "grad_norm": 0.019167251884937286,
+ "learning_rate": 0.0006,
+ "loss": 4.4278364181518555,
+ "step": 1325
+ },
+ {
+ "epoch": 18.419213973799128,
+ "grad_norm": 0.015895741060376167,
+ "learning_rate": 0.0006,
+ "loss": 4.474592208862305,
+ "step": 1326
+ },
+ {
+ "epoch": 18.433187772925763,
+ "grad_norm": 0.01531882956624031,
+ "learning_rate": 0.0006,
+ "loss": 4.403757572174072,
+ "step": 1327
+ },
+ {
+ "epoch": 18.4471615720524,
+ "grad_norm": 0.01565849967300892,
+ "learning_rate": 0.0006,
+ "loss": 4.507664680480957,
+ "step": 1328
+ },
+ {
+ "epoch": 18.46113537117904,
+ "grad_norm": 0.01602315716445446,
+ "learning_rate": 0.0006,
+ "loss": 4.5648908615112305,
+ "step": 1329
+ },
+ {
+ "epoch": 18.475109170305675,
+ "grad_norm": 0.016467366367578506,
+ "learning_rate": 0.0006,
+ "loss": 4.428762912750244,
+ "step": 1330
+ },
+ {
+ "epoch": 18.489082969432314,
+ "grad_norm": 0.01604127697646618,
+ "learning_rate": 0.0006,
+ "loss": 4.527771472930908,
+ "step": 1331
+ },
+ {
+ "epoch": 18.503056768558952,
+ "grad_norm": 0.018825586885213852,
+ "learning_rate": 0.0006,
+ "loss": 4.485386848449707,
+ "step": 1332
+ },
+ {
+ "epoch": 18.51703056768559,
+ "grad_norm": 0.019258007407188416,
+ "learning_rate": 0.0006,
+ "loss": 4.414972305297852,
+ "step": 1333
+ },
+ {
+ "epoch": 18.531004366812226,
+ "grad_norm": 0.018192021176218987,
+ "learning_rate": 0.0006,
+ "loss": 4.412166595458984,
+ "step": 1334
+ },
+ {
+ "epoch": 18.544978165938865,
+ "grad_norm": 0.016120698302984238,
+ "learning_rate": 0.0006,
+ "loss": 4.484317779541016,
+ "step": 1335
+ },
+ {
+ "epoch": 18.558951965065503,
+ "grad_norm": 0.016496622934937477,
+ "learning_rate": 0.0006,
+ "loss": 4.514986515045166,
+ "step": 1336
+ },
+ {
+ "epoch": 18.57292576419214,
+ "grad_norm": 0.01795116998255253,
+ "learning_rate": 0.0006,
+ "loss": 4.41473388671875,
+ "step": 1337
+ },
+ {
+ "epoch": 18.586899563318777,
+ "grad_norm": 0.019062276929616928,
+ "learning_rate": 0.0006,
+ "loss": 4.601996898651123,
+ "step": 1338
+ },
+ {
+ "epoch": 18.600873362445416,
+ "grad_norm": 0.017174631357192993,
+ "learning_rate": 0.0006,
+ "loss": 4.571218490600586,
+ "step": 1339
+ },
+ {
+ "epoch": 18.61484716157205,
+ "grad_norm": 0.014015606604516506,
+ "learning_rate": 0.0006,
+ "loss": 4.526606559753418,
+ "step": 1340
+ },
+ {
+ "epoch": 18.62882096069869,
+ "grad_norm": 0.015358510427176952,
+ "learning_rate": 0.0006,
+ "loss": 4.604286193847656,
+ "step": 1341
+ },
+ {
+ "epoch": 18.64279475982533,
+ "grad_norm": 0.01619391329586506,
+ "learning_rate": 0.0006,
+ "loss": 4.560883522033691,
+ "step": 1342
+ },
+ {
+ "epoch": 18.656768558951963,
+ "grad_norm": 0.015972964465618134,
+ "learning_rate": 0.0006,
+ "loss": 4.479755401611328,
+ "step": 1343
+ },
+ {
+ "epoch": 18.670742358078602,
+ "grad_norm": 0.016861025243997574,
+ "learning_rate": 0.0006,
+ "loss": 4.474471092224121,
+ "step": 1344
+ },
+ {
+ "epoch": 18.68471615720524,
+ "grad_norm": 0.01822078227996826,
+ "learning_rate": 0.0006,
+ "loss": 4.509489059448242,
+ "step": 1345
+ },
+ {
+ "epoch": 18.69868995633188,
+ "grad_norm": 0.019415007904171944,
+ "learning_rate": 0.0006,
+ "loss": 4.508941173553467,
+ "step": 1346
+ },
+ {
+ "epoch": 18.712663755458514,
+ "grad_norm": 0.019856126978993416,
+ "learning_rate": 0.0006,
+ "loss": 4.446670055389404,
+ "step": 1347
+ },
+ {
+ "epoch": 18.726637554585153,
+ "grad_norm": 0.01842520199716091,
+ "learning_rate": 0.0006,
+ "loss": 4.524422645568848,
+ "step": 1348
+ },
+ {
+ "epoch": 18.74061135371179,
+ "grad_norm": 0.017196480184793472,
+ "learning_rate": 0.0006,
+ "loss": 4.498466968536377,
+ "step": 1349
+ },
+ {
+ "epoch": 18.754585152838427,
+ "grad_norm": 0.01823010854423046,
+ "learning_rate": 0.0006,
+ "loss": 4.534684181213379,
+ "step": 1350
+ },
+ {
+ "epoch": 18.768558951965066,
+ "grad_norm": 0.020398065447807312,
+ "learning_rate": 0.0006,
+ "loss": 4.456494331359863,
+ "step": 1351
+ },
+ {
+ "epoch": 18.782532751091704,
+ "grad_norm": 0.022992262616753578,
+ "learning_rate": 0.0006,
+ "loss": 4.476374626159668,
+ "step": 1352
+ },
+ {
+ "epoch": 18.79650655021834,
+ "grad_norm": 0.02235420234501362,
+ "learning_rate": 0.0006,
+ "loss": 4.587776184082031,
+ "step": 1353
+ },
+ {
+ "epoch": 18.810480349344978,
+ "grad_norm": 0.02093626745045185,
+ "learning_rate": 0.0006,
+ "loss": 4.424172401428223,
+ "step": 1354
+ },
+ {
+ "epoch": 18.824454148471617,
+ "grad_norm": 0.01997302658855915,
+ "learning_rate": 0.0006,
+ "loss": 4.448066711425781,
+ "step": 1355
+ },
+ {
+ "epoch": 18.83842794759825,
+ "grad_norm": 0.018265459686517715,
+ "learning_rate": 0.0006,
+ "loss": 4.436610698699951,
+ "step": 1356
+ },
+ {
+ "epoch": 18.85240174672489,
+ "grad_norm": 0.01673833839595318,
+ "learning_rate": 0.0006,
+ "loss": 4.5502214431762695,
+ "step": 1357
+ },
+ {
+ "epoch": 18.86637554585153,
+ "grad_norm": 0.017908833920955658,
+ "learning_rate": 0.0006,
+ "loss": 4.550530433654785,
+ "step": 1358
+ },
+ {
+ "epoch": 18.880349344978168,
+ "grad_norm": 0.01796654611825943,
+ "learning_rate": 0.0006,
+ "loss": 4.531463146209717,
+ "step": 1359
+ },
+ {
+ "epoch": 18.894323144104803,
+ "grad_norm": 0.01646183803677559,
+ "learning_rate": 0.0006,
+ "loss": 4.498582363128662,
+ "step": 1360
+ },
+ {
+ "epoch": 18.90829694323144,
+ "grad_norm": 0.015430327504873276,
+ "learning_rate": 0.0006,
+ "loss": 4.52571439743042,
+ "step": 1361
+ },
+ {
+ "epoch": 18.92227074235808,
+ "grad_norm": 0.016527952626347542,
+ "learning_rate": 0.0006,
+ "loss": 4.547013759613037,
+ "step": 1362
+ },
+ {
+ "epoch": 18.936244541484715,
+ "grad_norm": 0.016263660043478012,
+ "learning_rate": 0.0006,
+ "loss": 4.468754291534424,
+ "step": 1363
+ },
+ {
+ "epoch": 18.950218340611354,
+ "grad_norm": 0.016347210854291916,
+ "learning_rate": 0.0006,
+ "loss": 4.450741767883301,
+ "step": 1364
+ },
+ {
+ "epoch": 18.964192139737992,
+ "grad_norm": 0.01701398566365242,
+ "learning_rate": 0.0006,
+ "loss": 4.534899711608887,
+ "step": 1365
+ },
+ {
+ "epoch": 18.978165938864628,
+ "grad_norm": 0.017693081870675087,
+ "learning_rate": 0.0006,
+ "loss": 4.460553169250488,
+ "step": 1366
+ },
+ {
+ "epoch": 18.992139737991266,
+ "grad_norm": 0.018746308982372284,
+ "learning_rate": 0.0006,
+ "loss": 4.538792610168457,
+ "step": 1367
+ },
+ {
+ "epoch": 19.0,
+ "grad_norm": 0.020099416375160217,
+ "learning_rate": 0.0006,
+ "loss": 4.449617385864258,
+ "step": 1368
+ },
+ {
+ "epoch": 19.0,
+ "eval_loss": 4.7756571769714355,
+ "eval_runtime": 57.1804,
+ "eval_samples_per_second": 42.707,
+ "eval_steps_per_second": 1.347,
+ "step": 1368
+ },
+ {
+ "epoch": 19.01397379912664,
+ "grad_norm": 0.018726933747529984,
+ "learning_rate": 0.0006,
+ "loss": 4.446059703826904,
+ "step": 1369
+ },
+ {
+ "epoch": 19.027947598253274,
+ "grad_norm": 0.01956087350845337,
+ "learning_rate": 0.0006,
+ "loss": 4.4520039558410645,
+ "step": 1370
+ },
+ {
+ "epoch": 19.041921397379912,
+ "grad_norm": 0.020092811435461044,
+ "learning_rate": 0.0006,
+ "loss": 4.543343544006348,
+ "step": 1371
+ },
+ {
+ "epoch": 19.05589519650655,
+ "grad_norm": 0.01843622885644436,
+ "learning_rate": 0.0006,
+ "loss": 4.535579204559326,
+ "step": 1372
+ },
+ {
+ "epoch": 19.069868995633186,
+ "grad_norm": 0.016989488154649734,
+ "learning_rate": 0.0006,
+ "loss": 4.52316951751709,
+ "step": 1373
+ },
+ {
+ "epoch": 19.083842794759825,
+ "grad_norm": 0.01805422455072403,
+ "learning_rate": 0.0006,
+ "loss": 4.5393266677856445,
+ "step": 1374
+ },
+ {
+ "epoch": 19.097816593886463,
+ "grad_norm": 0.01874629408121109,
+ "learning_rate": 0.0006,
+ "loss": 4.422059059143066,
+ "step": 1375
+ },
+ {
+ "epoch": 19.111790393013102,
+ "grad_norm": 0.01804221048951149,
+ "learning_rate": 0.0006,
+ "loss": 4.50508451461792,
+ "step": 1376
+ },
+ {
+ "epoch": 19.125764192139737,
+ "grad_norm": 0.01785474829375744,
+ "learning_rate": 0.0006,
+ "loss": 4.412599563598633,
+ "step": 1377
+ },
+ {
+ "epoch": 19.139737991266376,
+ "grad_norm": 0.01971791312098503,
+ "learning_rate": 0.0006,
+ "loss": 4.518318176269531,
+ "step": 1378
+ },
+ {
+ "epoch": 19.153711790393015,
+ "grad_norm": 0.02061633951961994,
+ "learning_rate": 0.0006,
+ "loss": 4.332393169403076,
+ "step": 1379
+ },
+ {
+ "epoch": 19.16768558951965,
+ "grad_norm": 0.020563429221510887,
+ "learning_rate": 0.0006,
+ "loss": 4.445863723754883,
+ "step": 1380
+ },
+ {
+ "epoch": 19.18165938864629,
+ "grad_norm": 0.019995713606476784,
+ "learning_rate": 0.0006,
+ "loss": 4.497312068939209,
+ "step": 1381
+ },
+ {
+ "epoch": 19.195633187772927,
+ "grad_norm": 0.019823070615530014,
+ "learning_rate": 0.0006,
+ "loss": 4.475085258483887,
+ "step": 1382
+ },
+ {
+ "epoch": 19.209606986899562,
+ "grad_norm": 0.019039355218410492,
+ "learning_rate": 0.0006,
+ "loss": 4.50432014465332,
+ "step": 1383
+ },
+ {
+ "epoch": 19.2235807860262,
+ "grad_norm": 0.017876973375678062,
+ "learning_rate": 0.0006,
+ "loss": 4.384788513183594,
+ "step": 1384
+ },
+ {
+ "epoch": 19.23755458515284,
+ "grad_norm": 0.018697699531912804,
+ "learning_rate": 0.0006,
+ "loss": 4.410778045654297,
+ "step": 1385
+ },
+ {
+ "epoch": 19.251528384279474,
+ "grad_norm": 0.018826806917786598,
+ "learning_rate": 0.0006,
+ "loss": 4.524034023284912,
+ "step": 1386
+ },
+ {
+ "epoch": 19.265502183406113,
+ "grad_norm": 0.01824226602911949,
+ "learning_rate": 0.0006,
+ "loss": 4.526393890380859,
+ "step": 1387
+ },
+ {
+ "epoch": 19.27947598253275,
+ "grad_norm": 0.01753208599984646,
+ "learning_rate": 0.0006,
+ "loss": 4.493703365325928,
+ "step": 1388
+ },
+ {
+ "epoch": 19.29344978165939,
+ "grad_norm": 0.01742432825267315,
+ "learning_rate": 0.0006,
+ "loss": 4.392831802368164,
+ "step": 1389
+ },
+ {
+ "epoch": 19.307423580786025,
+ "grad_norm": 0.017099356278777122,
+ "learning_rate": 0.0006,
+ "loss": 4.578668594360352,
+ "step": 1390
+ },
+ {
+ "epoch": 19.321397379912664,
+ "grad_norm": 0.017946023494005203,
+ "learning_rate": 0.0006,
+ "loss": 4.456718444824219,
+ "step": 1391
+ },
+ {
+ "epoch": 19.335371179039303,
+ "grad_norm": 0.017705943435430527,
+ "learning_rate": 0.0006,
+ "loss": 4.553682327270508,
+ "step": 1392
+ },
+ {
+ "epoch": 19.349344978165938,
+ "grad_norm": 0.016647523269057274,
+ "learning_rate": 0.0006,
+ "loss": 4.525016784667969,
+ "step": 1393
+ },
+ {
+ "epoch": 19.363318777292577,
+ "grad_norm": 0.015407416969537735,
+ "learning_rate": 0.0006,
+ "loss": 4.500231742858887,
+ "step": 1394
+ },
+ {
+ "epoch": 19.377292576419215,
+ "grad_norm": 0.017264019697904587,
+ "learning_rate": 0.0006,
+ "loss": 4.4615349769592285,
+ "step": 1395
+ },
+ {
+ "epoch": 19.39126637554585,
+ "grad_norm": 0.01871555484831333,
+ "learning_rate": 0.0006,
+ "loss": 4.281026840209961,
+ "step": 1396
+ },
+ {
+ "epoch": 19.40524017467249,
+ "grad_norm": 0.017361650243401527,
+ "learning_rate": 0.0006,
+ "loss": 4.479382514953613,
+ "step": 1397
+ },
+ {
+ "epoch": 19.419213973799128,
+ "grad_norm": 0.019781362265348434,
+ "learning_rate": 0.0006,
+ "loss": 4.402320861816406,
+ "step": 1398
+ },
+ {
+ "epoch": 19.433187772925763,
+ "grad_norm": 0.019043298438191414,
+ "learning_rate": 0.0006,
+ "loss": 4.476983070373535,
+ "step": 1399
+ },
+ {
+ "epoch": 19.4471615720524,
+ "grad_norm": 0.017137985676527023,
+ "learning_rate": 0.0006,
+ "loss": 4.418689250946045,
+ "step": 1400
+ },
+ {
+ "epoch": 19.46113537117904,
+ "grad_norm": 0.016481833532452583,
+ "learning_rate": 0.0006,
+ "loss": 4.579145431518555,
+ "step": 1401
+ },
+ {
+ "epoch": 19.475109170305675,
+ "grad_norm": 0.016507714986801147,
+ "learning_rate": 0.0006,
+ "loss": 4.519158363342285,
+ "step": 1402
+ },
+ {
+ "epoch": 19.489082969432314,
+ "grad_norm": 0.018092872574925423,
+ "learning_rate": 0.0006,
+ "loss": 4.513748645782471,
+ "step": 1403
+ },
+ {
+ "epoch": 19.503056768558952,
+ "grad_norm": 0.018153440207242966,
+ "learning_rate": 0.0006,
+ "loss": 4.645384311676025,
+ "step": 1404
+ },
+ {
+ "epoch": 19.51703056768559,
+ "grad_norm": 0.01657119207084179,
+ "learning_rate": 0.0006,
+ "loss": 4.399757385253906,
+ "step": 1405
+ },
+ {
+ "epoch": 19.531004366812226,
+ "grad_norm": 0.0156280305236578,
+ "learning_rate": 0.0006,
+ "loss": 4.467836380004883,
+ "step": 1406
+ },
+ {
+ "epoch": 19.544978165938865,
+ "grad_norm": 0.016838403418660164,
+ "learning_rate": 0.0006,
+ "loss": 4.529176235198975,
+ "step": 1407
+ },
+ {
+ "epoch": 19.558951965065503,
+ "grad_norm": 0.01743236929178238,
+ "learning_rate": 0.0006,
+ "loss": 4.440032482147217,
+ "step": 1408
+ },
+ {
+ "epoch": 19.57292576419214,
+ "grad_norm": 0.01553610060364008,
+ "learning_rate": 0.0006,
+ "loss": 4.425588607788086,
+ "step": 1409
+ },
+ {
+ "epoch": 19.586899563318777,
+ "grad_norm": 0.017259271815419197,
+ "learning_rate": 0.0006,
+ "loss": 4.405452728271484,
+ "step": 1410
+ },
+ {
+ "epoch": 19.600873362445416,
+ "grad_norm": 0.015568481758236885,
+ "learning_rate": 0.0006,
+ "loss": 4.444279193878174,
+ "step": 1411
+ },
+ {
+ "epoch": 19.61484716157205,
+ "grad_norm": 0.015331248752772808,
+ "learning_rate": 0.0006,
+ "loss": 4.451846122741699,
+ "step": 1412
+ },
+ {
+ "epoch": 19.62882096069869,
+ "grad_norm": 0.015667378902435303,
+ "learning_rate": 0.0006,
+ "loss": 4.533931255340576,
+ "step": 1413
+ },
+ {
+ "epoch": 19.64279475982533,
+ "grad_norm": 0.016177602112293243,
+ "learning_rate": 0.0006,
+ "loss": 4.467313289642334,
+ "step": 1414
+ },
+ {
+ "epoch": 19.656768558951963,
+ "grad_norm": 0.01733129471540451,
+ "learning_rate": 0.0006,
+ "loss": 4.496893882751465,
+ "step": 1415
+ },
+ {
+ "epoch": 19.670742358078602,
+ "grad_norm": 0.016539594158530235,
+ "learning_rate": 0.0006,
+ "loss": 4.401961326599121,
+ "step": 1416
+ },
+ {
+ "epoch": 19.68471615720524,
+ "grad_norm": 0.016058551147580147,
+ "learning_rate": 0.0006,
+ "loss": 4.503375053405762,
+ "step": 1417
+ },
+ {
+ "epoch": 19.69868995633188,
+ "grad_norm": 0.01715777814388275,
+ "learning_rate": 0.0006,
+ "loss": 4.3808369636535645,
+ "step": 1418
+ },
+ {
+ "epoch": 19.712663755458514,
+ "grad_norm": 0.01681119203567505,
+ "learning_rate": 0.0006,
+ "loss": 4.5469560623168945,
+ "step": 1419
+ },
+ {
+ "epoch": 19.726637554585153,
+ "grad_norm": 0.01634550653398037,
+ "learning_rate": 0.0006,
+ "loss": 4.328197956085205,
+ "step": 1420
+ },
+ {
+ "epoch": 19.74061135371179,
+ "grad_norm": 0.01549347210675478,
+ "learning_rate": 0.0006,
+ "loss": 4.534621238708496,
+ "step": 1421
+ },
+ {
+ "epoch": 19.754585152838427,
+ "grad_norm": 0.01654665172100067,
+ "learning_rate": 0.0006,
+ "loss": 4.4724225997924805,
+ "step": 1422
+ },
+ {
+ "epoch": 19.768558951965066,
+ "grad_norm": 0.017119025811553,
+ "learning_rate": 0.0006,
+ "loss": 4.3897247314453125,
+ "step": 1423
+ },
+ {
+ "epoch": 19.782532751091704,
+ "grad_norm": 0.01603223942220211,
+ "learning_rate": 0.0006,
+ "loss": 4.412731170654297,
+ "step": 1424
+ },
+ {
+ "epoch": 19.79650655021834,
+ "grad_norm": 0.016774067655205727,
+ "learning_rate": 0.0006,
+ "loss": 4.54707145690918,
+ "step": 1425
+ },
+ {
+ "epoch": 19.810480349344978,
+ "grad_norm": 0.016453025862574577,
+ "learning_rate": 0.0006,
+ "loss": 4.462501049041748,
+ "step": 1426
+ },
+ {
+ "epoch": 19.824454148471617,
+ "grad_norm": 0.01597629487514496,
+ "learning_rate": 0.0006,
+ "loss": 4.229758262634277,
+ "step": 1427
+ },
+ {
+ "epoch": 19.83842794759825,
+ "grad_norm": 0.017999034374952316,
+ "learning_rate": 0.0006,
+ "loss": 4.389263153076172,
+ "step": 1428
+ },
+ {
+ "epoch": 19.85240174672489,
+ "grad_norm": 0.02165898121893406,
+ "learning_rate": 0.0006,
+ "loss": 4.342130661010742,
+ "step": 1429
+ },
+ {
+ "epoch": 19.86637554585153,
+ "grad_norm": 0.02222822792828083,
+ "learning_rate": 0.0006,
+ "loss": 4.492661476135254,
+ "step": 1430
+ },
+ {
+ "epoch": 19.880349344978168,
+ "grad_norm": 0.019378066062927246,
+ "learning_rate": 0.0006,
+ "loss": 4.4614105224609375,
+ "step": 1431
+ },
+ {
+ "epoch": 19.894323144104803,
+ "grad_norm": 0.019908083602786064,
+ "learning_rate": 0.0006,
+ "loss": 4.370929718017578,
+ "step": 1432
+ },
+ {
+ "epoch": 19.90829694323144,
+ "grad_norm": 0.022296471521258354,
+ "learning_rate": 0.0006,
+ "loss": 4.438320159912109,
+ "step": 1433
+ },
+ {
+ "epoch": 19.92227074235808,
+ "grad_norm": 0.021810343489050865,
+ "learning_rate": 0.0006,
+ "loss": 4.462067127227783,
+ "step": 1434
+ },
+ {
+ "epoch": 19.936244541484715,
+ "grad_norm": 0.02129209227859974,
+ "learning_rate": 0.0006,
+ "loss": 4.541261196136475,
+ "step": 1435
+ },
+ {
+ "epoch": 19.950218340611354,
+ "grad_norm": 0.02079629711806774,
+ "learning_rate": 0.0006,
+ "loss": 4.36712646484375,
+ "step": 1436
+ },
+ {
+ "epoch": 19.964192139737992,
+ "grad_norm": 0.022501740604639053,
+ "learning_rate": 0.0006,
+ "loss": 4.5279083251953125,
+ "step": 1437
+ },
+ {
+ "epoch": 19.978165938864628,
+ "grad_norm": 0.019121676683425903,
+ "learning_rate": 0.0006,
+ "loss": 4.477789878845215,
+ "step": 1438
+ },
+ {
+ "epoch": 19.992139737991266,
+ "grad_norm": 0.017429588362574577,
+ "learning_rate": 0.0006,
+ "loss": 4.4023847579956055,
+ "step": 1439
+ },
+ {
+ "epoch": 20.0,
+ "grad_norm": 0.018857665359973907,
+ "learning_rate": 0.0006,
+ "loss": 4.494743347167969,
+ "step": 1440
+ },
+ {
+ "epoch": 20.0,
+ "eval_loss": 4.694555759429932,
+ "eval_runtime": 56.4146,
+ "eval_samples_per_second": 43.287,
+ "eval_steps_per_second": 1.365,
+ "step": 1440
+ },
+ {
+ "epoch": 20.01397379912664,
+ "grad_norm": 0.016505861654877663,
+ "learning_rate": 0.0006,
+ "loss": 4.506229877471924,
+ "step": 1441
+ },
+ {
+ "epoch": 20.027947598253274,
+ "grad_norm": 0.017571842297911644,
+ "learning_rate": 0.0006,
+ "loss": 4.384073257446289,
+ "step": 1442
+ },
+ {
+ "epoch": 20.041921397379912,
+ "grad_norm": 0.019398365169763565,
+ "learning_rate": 0.0006,
+ "loss": 4.482177734375,
+ "step": 1443
+ },
+ {
+ "epoch": 20.05589519650655,
+ "grad_norm": 0.02011062018573284,
+ "learning_rate": 0.0006,
+ "loss": 4.440773963928223,
+ "step": 1444
+ },
+ {
+ "epoch": 20.069868995633186,
+ "grad_norm": 0.019026953727006912,
+ "learning_rate": 0.0006,
+ "loss": 4.507168769836426,
+ "step": 1445
+ },
+ {
+ "epoch": 20.083842794759825,
+ "grad_norm": 0.020506031811237335,
+ "learning_rate": 0.0006,
+ "loss": 4.469079971313477,
+ "step": 1446
+ },
+ {
+ "epoch": 20.097816593886463,
+ "grad_norm": 0.019608162343502045,
+ "learning_rate": 0.0006,
+ "loss": 4.395927429199219,
+ "step": 1447
+ },
+ {
+ "epoch": 20.111790393013102,
+ "grad_norm": 0.019419124349951744,
+ "learning_rate": 0.0006,
+ "loss": 4.484235763549805,
+ "step": 1448
+ },
+ {
+ "epoch": 20.125764192139737,
+ "grad_norm": 0.021775128319859505,
+ "learning_rate": 0.0006,
+ "loss": 4.333173751831055,
+ "step": 1449
+ },
+ {
+ "epoch": 20.139737991266376,
+ "grad_norm": 0.025512272492051125,
+ "learning_rate": 0.0006,
+ "loss": 4.460362434387207,
+ "step": 1450
+ },
+ {
+ "epoch": 20.153711790393015,
+ "grad_norm": 0.025157401338219643,
+ "learning_rate": 0.0006,
+ "loss": 4.4361395835876465,
+ "step": 1451
+ },
+ {
+ "epoch": 20.16768558951965,
+ "grad_norm": 0.019862616434693336,
+ "learning_rate": 0.0006,
+ "loss": 4.426807403564453,
+ "step": 1452
+ },
+ {
+ "epoch": 20.18165938864629,
+ "grad_norm": 0.01932152360677719,
+ "learning_rate": 0.0006,
+ "loss": 4.375677108764648,
+ "step": 1453
+ },
+ {
+ "epoch": 20.195633187772927,
+ "grad_norm": 0.01760999858379364,
+ "learning_rate": 0.0006,
+ "loss": 4.42259407043457,
+ "step": 1454
+ },
+ {
+ "epoch": 20.209606986899562,
+ "grad_norm": 0.016998112201690674,
+ "learning_rate": 0.0006,
+ "loss": 4.481280326843262,
+ "step": 1455
+ },
+ {
+ "epoch": 20.2235807860262,
+ "grad_norm": 0.017199190333485603,
+ "learning_rate": 0.0006,
+ "loss": 4.496712684631348,
+ "step": 1456
+ },
+ {
+ "epoch": 20.23755458515284,
+ "grad_norm": 0.016833679750561714,
+ "learning_rate": 0.0006,
+ "loss": 4.40736722946167,
+ "step": 1457
+ },
+ {
+ "epoch": 20.251528384279474,
+ "grad_norm": 0.017546730116009712,
+ "learning_rate": 0.0006,
+ "loss": 4.4915266036987305,
+ "step": 1458
+ },
+ {
+ "epoch": 20.265502183406113,
+ "grad_norm": 0.016595078632235527,
+ "learning_rate": 0.0006,
+ "loss": 4.460815906524658,
+ "step": 1459
+ },
+ {
+ "epoch": 20.27947598253275,
+ "grad_norm": 0.017728568986058235,
+ "learning_rate": 0.0006,
+ "loss": 4.346048831939697,
+ "step": 1460
+ },
+ {
+ "epoch": 20.29344978165939,
+ "grad_norm": 0.01849197782576084,
+ "learning_rate": 0.0006,
+ "loss": 4.399681091308594,
+ "step": 1461
+ },
+ {
+ "epoch": 20.307423580786025,
+ "grad_norm": 0.018336588516831398,
+ "learning_rate": 0.0006,
+ "loss": 4.414183616638184,
+ "step": 1462
+ },
+ {
+ "epoch": 20.321397379912664,
+ "grad_norm": 0.018086519092321396,
+ "learning_rate": 0.0006,
+ "loss": 4.346586227416992,
+ "step": 1463
+ },
+ {
+ "epoch": 20.335371179039303,
+ "grad_norm": 0.019697299227118492,
+ "learning_rate": 0.0006,
+ "loss": 4.4420952796936035,
+ "step": 1464
+ },
+ {
+ "epoch": 20.349344978165938,
+ "grad_norm": 0.02198074571788311,
+ "learning_rate": 0.0006,
+ "loss": 4.588729381561279,
+ "step": 1465
+ },
+ {
+ "epoch": 20.363318777292577,
+ "grad_norm": 0.022813329473137856,
+ "learning_rate": 0.0006,
+ "loss": 4.49729585647583,
+ "step": 1466
+ },
+ {
+ "epoch": 20.377292576419215,
+ "grad_norm": 0.022006649523973465,
+ "learning_rate": 0.0006,
+ "loss": 4.475641250610352,
+ "step": 1467
+ },
+ {
+ "epoch": 20.39126637554585,
+ "grad_norm": 0.023188438266515732,
+ "learning_rate": 0.0006,
+ "loss": 4.404322624206543,
+ "step": 1468
+ },
+ {
+ "epoch": 20.40524017467249,
+ "grad_norm": 0.022226877510547638,
+ "learning_rate": 0.0006,
+ "loss": 4.50904655456543,
+ "step": 1469
+ },
+ {
+ "epoch": 20.419213973799128,
+ "grad_norm": 0.021082637831568718,
+ "learning_rate": 0.0006,
+ "loss": 4.339011192321777,
+ "step": 1470
+ },
+ {
+ "epoch": 20.433187772925763,
+ "grad_norm": 0.02013542130589485,
+ "learning_rate": 0.0006,
+ "loss": 4.516942024230957,
+ "step": 1471
+ },
+ {
+ "epoch": 20.4471615720524,
+ "grad_norm": 0.01898968778550625,
+ "learning_rate": 0.0006,
+ "loss": 4.315804481506348,
+ "step": 1472
+ },
+ {
+ "epoch": 20.46113537117904,
+ "grad_norm": 0.019450880587100983,
+ "learning_rate": 0.0006,
+ "loss": 4.4794921875,
+ "step": 1473
+ },
+ {
+ "epoch": 20.475109170305675,
+ "grad_norm": 0.018577802926301956,
+ "learning_rate": 0.0006,
+ "loss": 4.422323703765869,
+ "step": 1474
+ },
+ {
+ "epoch": 20.489082969432314,
+ "grad_norm": 0.01778412237763405,
+ "learning_rate": 0.0006,
+ "loss": 4.429789066314697,
+ "step": 1475
+ },
+ {
+ "epoch": 20.503056768558952,
+ "grad_norm": 0.017134476453065872,
+ "learning_rate": 0.0006,
+ "loss": 4.416459083557129,
+ "step": 1476
+ },
+ {
+ "epoch": 20.51703056768559,
+ "grad_norm": 0.017153557389974594,
+ "learning_rate": 0.0006,
+ "loss": 4.446170806884766,
+ "step": 1477
+ },
+ {
+ "epoch": 20.531004366812226,
+ "grad_norm": 0.014998532831668854,
+ "learning_rate": 0.0006,
+ "loss": 4.419447898864746,
+ "step": 1478
+ },
+ {
+ "epoch": 20.544978165938865,
+ "grad_norm": 0.015877850353717804,
+ "learning_rate": 0.0006,
+ "loss": 4.312338829040527,
+ "step": 1479
+ },
+ {
+ "epoch": 20.558951965065503,
+ "grad_norm": 0.01569489948451519,
+ "learning_rate": 0.0006,
+ "loss": 4.572905540466309,
+ "step": 1480
+ },
+ {
+ "epoch": 20.57292576419214,
+ "grad_norm": 0.01591474376618862,
+ "learning_rate": 0.0006,
+ "loss": 4.414907455444336,
+ "step": 1481
+ },
+ {
+ "epoch": 20.586899563318777,
+ "grad_norm": 0.015395903028547764,
+ "learning_rate": 0.0006,
+ "loss": 4.549131393432617,
+ "step": 1482
+ },
+ {
+ "epoch": 20.600873362445416,
+ "grad_norm": 0.015715358778834343,
+ "learning_rate": 0.0006,
+ "loss": 4.530194282531738,
+ "step": 1483
+ },
+ {
+ "epoch": 20.61484716157205,
+ "grad_norm": 0.016042975708842278,
+ "learning_rate": 0.0006,
+ "loss": 4.322149276733398,
+ "step": 1484
+ },
+ {
+ "epoch": 20.62882096069869,
+ "grad_norm": 0.01664186641573906,
+ "learning_rate": 0.0006,
+ "loss": 4.424866676330566,
+ "step": 1485
+ },
+ {
+ "epoch": 20.64279475982533,
+ "grad_norm": 0.01631534844636917,
+ "learning_rate": 0.0006,
+ "loss": 4.441752910614014,
+ "step": 1486
+ },
+ {
+ "epoch": 20.656768558951963,
+ "grad_norm": 0.014057058840990067,
+ "learning_rate": 0.0006,
+ "loss": 4.399982452392578,
+ "step": 1487
+ },
+ {
+ "epoch": 20.670742358078602,
+ "grad_norm": 0.015076273120939732,
+ "learning_rate": 0.0006,
+ "loss": 4.441324234008789,
+ "step": 1488
+ },
+ {
+ "epoch": 20.68471615720524,
+ "grad_norm": 0.014138277620077133,
+ "learning_rate": 0.0006,
+ "loss": 4.4312543869018555,
+ "step": 1489
+ },
+ {
+ "epoch": 20.69868995633188,
+ "grad_norm": 0.013762440532445908,
+ "learning_rate": 0.0006,
+ "loss": 4.387807846069336,
+ "step": 1490
+ },
+ {
+ "epoch": 20.712663755458514,
+ "grad_norm": 0.015058411285281181,
+ "learning_rate": 0.0006,
+ "loss": 4.594747543334961,
+ "step": 1491
+ },
+ {
+ "epoch": 20.726637554585153,
+ "grad_norm": 0.013827123679220676,
+ "learning_rate": 0.0006,
+ "loss": 4.434604167938232,
+ "step": 1492
+ },
+ {
+ "epoch": 20.74061135371179,
+ "grad_norm": 0.015315295197069645,
+ "learning_rate": 0.0006,
+ "loss": 4.588066101074219,
+ "step": 1493
+ },
+ {
+ "epoch": 20.754585152838427,
+ "grad_norm": 0.018678732216358185,
+ "learning_rate": 0.0006,
+ "loss": 4.3514204025268555,
+ "step": 1494
+ },
+ {
+ "epoch": 20.768558951965066,
+ "grad_norm": 0.019521350041031837,
+ "learning_rate": 0.0006,
+ "loss": 4.437658786773682,
+ "step": 1495
+ },
+ {
+ "epoch": 20.782532751091704,
+ "grad_norm": 0.018363403156399727,
+ "learning_rate": 0.0006,
+ "loss": 4.495002269744873,
+ "step": 1496
+ },
+ {
+ "epoch": 20.79650655021834,
+ "grad_norm": 0.016830725595355034,
+ "learning_rate": 0.0006,
+ "loss": 4.402471542358398,
+ "step": 1497
+ },
+ {
+ "epoch": 20.810480349344978,
+ "grad_norm": 0.01722072623670101,
+ "learning_rate": 0.0006,
+ "loss": 4.437653541564941,
+ "step": 1498
+ },
+ {
+ "epoch": 20.824454148471617,
+ "grad_norm": 0.017459599301218987,
+ "learning_rate": 0.0006,
+ "loss": 4.366603851318359,
+ "step": 1499
+ },
+ {
+ "epoch": 20.83842794759825,
+ "grad_norm": 0.018902219831943512,
+ "learning_rate": 0.0006,
+ "loss": 4.501561164855957,
+ "step": 1500
+ },
+ {
+ "epoch": 20.85240174672489,
+ "grad_norm": 0.020814096555113792,
+ "learning_rate": 0.0006,
+ "loss": 4.535295486450195,
+ "step": 1501
+ },
+ {
+ "epoch": 20.86637554585153,
+ "grad_norm": 0.018990976735949516,
+ "learning_rate": 0.0006,
+ "loss": 4.333580017089844,
+ "step": 1502
+ },
+ {
+ "epoch": 20.880349344978168,
+ "grad_norm": 0.01837880350649357,
+ "learning_rate": 0.0006,
+ "loss": 4.313388824462891,
+ "step": 1503
+ },
+ {
+ "epoch": 20.894323144104803,
+ "grad_norm": 0.019820360466837883,
+ "learning_rate": 0.0006,
+ "loss": 4.388157844543457,
+ "step": 1504
+ },
+ {
+ "epoch": 20.90829694323144,
+ "grad_norm": 0.020181143656373024,
+ "learning_rate": 0.0006,
+ "loss": 4.572224140167236,
+ "step": 1505
+ },
+ {
+ "epoch": 20.92227074235808,
+ "grad_norm": 0.021206077188253403,
+ "learning_rate": 0.0006,
+ "loss": 4.3127970695495605,
+ "step": 1506
+ },
+ {
+ "epoch": 20.936244541484715,
+ "grad_norm": 0.017962060868740082,
+ "learning_rate": 0.0006,
+ "loss": 4.384734630584717,
+ "step": 1507
+ },
+ {
+ "epoch": 20.950218340611354,
+ "grad_norm": 0.01737220585346222,
+ "learning_rate": 0.0006,
+ "loss": 4.4075751304626465,
+ "step": 1508
+ },
+ {
+ "epoch": 20.964192139737992,
+ "grad_norm": 0.01839268207550049,
+ "learning_rate": 0.0006,
+ "loss": 4.553779125213623,
+ "step": 1509
+ },
+ {
+ "epoch": 20.978165938864628,
+ "grad_norm": 0.020102640613913536,
+ "learning_rate": 0.0006,
+ "loss": 4.399398326873779,
+ "step": 1510
+ },
+ {
+ "epoch": 20.992139737991266,
+ "grad_norm": 0.018831463530659676,
+ "learning_rate": 0.0006,
+ "loss": 4.358499526977539,
+ "step": 1511
+ },
+ {
+ "epoch": 21.0,
+ "grad_norm": 0.018254734575748444,
+ "learning_rate": 0.0006,
+ "loss": 4.384119510650635,
+ "step": 1512
+ },
+ {
+ "epoch": 21.0,
+ "eval_loss": 4.7524237632751465,
+ "eval_runtime": 57.8812,
+ "eval_samples_per_second": 42.19,
+ "eval_steps_per_second": 1.33,
+ "step": 1512
+ },
+ {
+ "epoch": 21.01397379912664,
+ "grad_norm": 0.019728587940335274,
+ "learning_rate": 0.0006,
+ "loss": 4.32363224029541,
+ "step": 1513
+ },
+ {
+ "epoch": 21.027947598253274,
+ "grad_norm": 0.02216625027358532,
+ "learning_rate": 0.0006,
+ "loss": 4.436988353729248,
+ "step": 1514
+ },
+ {
+ "epoch": 21.041921397379912,
+ "grad_norm": 0.02227606810629368,
+ "learning_rate": 0.0006,
+ "loss": 4.406063556671143,
+ "step": 1515
+ },
+ {
+ "epoch": 21.05589519650655,
+ "grad_norm": 0.020187009125947952,
+ "learning_rate": 0.0006,
+ "loss": 4.324402332305908,
+ "step": 1516
+ },
+ {
+ "epoch": 21.069868995633186,
+ "grad_norm": 0.02095157466828823,
+ "learning_rate": 0.0006,
+ "loss": 4.456544876098633,
+ "step": 1517
+ },
+ {
+ "epoch": 21.083842794759825,
+ "grad_norm": 0.02146013453602791,
+ "learning_rate": 0.0006,
+ "loss": 4.368387222290039,
+ "step": 1518
+ },
+ {
+ "epoch": 21.097816593886463,
+ "grad_norm": 0.02162035182118416,
+ "learning_rate": 0.0006,
+ "loss": 4.48838996887207,
+ "step": 1519
+ },
+ {
+ "epoch": 21.111790393013102,
+ "grad_norm": 0.021745676174759865,
+ "learning_rate": 0.0006,
+ "loss": 4.30778694152832,
+ "step": 1520
+ },
+ {
+ "epoch": 21.125764192139737,
+ "grad_norm": 0.02173588052392006,
+ "learning_rate": 0.0006,
+ "loss": 4.433149814605713,
+ "step": 1521
+ },
+ {
+ "epoch": 21.139737991266376,
+ "grad_norm": 0.020950648933649063,
+ "learning_rate": 0.0006,
+ "loss": 4.520089149475098,
+ "step": 1522
+ },
+ {
+ "epoch": 21.153711790393015,
+ "grad_norm": 0.020684752613306046,
+ "learning_rate": 0.0006,
+ "loss": 4.372437477111816,
+ "step": 1523
+ },
+ {
+ "epoch": 21.16768558951965,
+ "grad_norm": 0.019472869113087654,
+ "learning_rate": 0.0006,
+ "loss": 4.437015533447266,
+ "step": 1524
+ },
+ {
+ "epoch": 21.18165938864629,
+ "grad_norm": 0.01846962980926037,
+ "learning_rate": 0.0006,
+ "loss": 4.449531555175781,
+ "step": 1525
+ },
+ {
+ "epoch": 21.195633187772927,
+ "grad_norm": 0.017377803102135658,
+ "learning_rate": 0.0006,
+ "loss": 4.458279132843018,
+ "step": 1526
+ },
+ {
+ "epoch": 21.209606986899562,
+ "grad_norm": 0.016002200543880463,
+ "learning_rate": 0.0006,
+ "loss": 4.266297817230225,
+ "step": 1527
+ },
+ {
+ "epoch": 21.2235807860262,
+ "grad_norm": 0.015575903467833996,
+ "learning_rate": 0.0006,
+ "loss": 4.342632293701172,
+ "step": 1528
+ },
+ {
+ "epoch": 21.23755458515284,
+ "grad_norm": 0.015634698793292046,
+ "learning_rate": 0.0006,
+ "loss": 4.338623523712158,
+ "step": 1529
+ },
+ {
+ "epoch": 21.251528384279474,
+ "grad_norm": 0.01449581328779459,
+ "learning_rate": 0.0006,
+ "loss": 4.354330062866211,
+ "step": 1530
+ },
+ {
+ "epoch": 21.265502183406113,
+ "grad_norm": 0.01487616915255785,
+ "learning_rate": 0.0006,
+ "loss": 4.486835479736328,
+ "step": 1531
+ },
+ {
+ "epoch": 21.27947598253275,
+ "grad_norm": 0.016346555203199387,
+ "learning_rate": 0.0006,
+ "loss": 4.407098770141602,
+ "step": 1532
+ },
+ {
+ "epoch": 21.29344978165939,
+ "grad_norm": 0.01682182401418686,
+ "learning_rate": 0.0006,
+ "loss": 4.436809539794922,
+ "step": 1533
+ },
+ {
+ "epoch": 21.307423580786025,
+ "grad_norm": 0.017704004421830177,
+ "learning_rate": 0.0006,
+ "loss": 4.3854780197143555,
+ "step": 1534
+ },
+ {
+ "epoch": 21.321397379912664,
+ "grad_norm": 0.018437420949339867,
+ "learning_rate": 0.0006,
+ "loss": 4.413283824920654,
+ "step": 1535
+ },
+ {
+ "epoch": 21.335371179039303,
+ "grad_norm": 0.016911081969738007,
+ "learning_rate": 0.0006,
+ "loss": 4.357306003570557,
+ "step": 1536
+ },
+ {
+ "epoch": 21.349344978165938,
+ "grad_norm": 0.01642223633825779,
+ "learning_rate": 0.0006,
+ "loss": 4.422322750091553,
+ "step": 1537
+ },
+ {
+ "epoch": 21.363318777292577,
+ "grad_norm": 0.01780877821147442,
+ "learning_rate": 0.0006,
+ "loss": 4.387620449066162,
+ "step": 1538
+ },
+ {
+ "epoch": 21.377292576419215,
+ "grad_norm": 0.01653975248336792,
+ "learning_rate": 0.0006,
+ "loss": 4.327267646789551,
+ "step": 1539
+ },
+ {
+ "epoch": 21.39126637554585,
+ "grad_norm": 0.01572156324982643,
+ "learning_rate": 0.0006,
+ "loss": 4.475979804992676,
+ "step": 1540
+ },
+ {
+ "epoch": 21.40524017467249,
+ "grad_norm": 0.017156066372990608,
+ "learning_rate": 0.0006,
+ "loss": 4.4011125564575195,
+ "step": 1541
+ },
+ {
+ "epoch": 21.419213973799128,
+ "grad_norm": 0.01788829080760479,
+ "learning_rate": 0.0006,
+ "loss": 4.4174089431762695,
+ "step": 1542
+ },
+ {
+ "epoch": 21.433187772925763,
+ "grad_norm": 0.01728985086083412,
+ "learning_rate": 0.0006,
+ "loss": 4.429244041442871,
+ "step": 1543
+ },
+ {
+ "epoch": 21.4471615720524,
+ "grad_norm": 0.016534466296434402,
+ "learning_rate": 0.0006,
+ "loss": 4.326068878173828,
+ "step": 1544
+ },
+ {
+ "epoch": 21.46113537117904,
+ "grad_norm": 0.016582584008574486,
+ "learning_rate": 0.0006,
+ "loss": 4.2975969314575195,
+ "step": 1545
+ },
+ {
+ "epoch": 21.475109170305675,
+ "grad_norm": 0.018034178763628006,
+ "learning_rate": 0.0006,
+ "loss": 4.501319408416748,
+ "step": 1546
+ },
+ {
+ "epoch": 21.489082969432314,
+ "grad_norm": 0.017827702686190605,
+ "learning_rate": 0.0006,
+ "loss": 4.502284526824951,
+ "step": 1547
+ },
+ {
+ "epoch": 21.503056768558952,
+ "grad_norm": 0.017182299867272377,
+ "learning_rate": 0.0006,
+ "loss": 4.380828857421875,
+ "step": 1548
+ },
+ {
+ "epoch": 21.51703056768559,
+ "grad_norm": 0.019101588055491447,
+ "learning_rate": 0.0006,
+ "loss": 4.31564998626709,
+ "step": 1549
+ },
+ {
+ "epoch": 21.531004366812226,
+ "grad_norm": 0.0204677302390337,
+ "learning_rate": 0.0006,
+ "loss": 4.3840131759643555,
+ "step": 1550
+ },
+ {
+ "epoch": 21.544978165938865,
+ "grad_norm": 0.0188993439078331,
+ "learning_rate": 0.0006,
+ "loss": 4.4257731437683105,
+ "step": 1551
+ },
+ {
+ "epoch": 21.558951965065503,
+ "grad_norm": 0.01730230636894703,
+ "learning_rate": 0.0006,
+ "loss": 4.445132255554199,
+ "step": 1552
+ },
+ {
+ "epoch": 21.57292576419214,
+ "grad_norm": 0.01801574043929577,
+ "learning_rate": 0.0006,
+ "loss": 4.427064895629883,
+ "step": 1553
+ },
+ {
+ "epoch": 21.586899563318777,
+ "grad_norm": 0.018417565152049065,
+ "learning_rate": 0.0006,
+ "loss": 4.319595813751221,
+ "step": 1554
+ },
+ {
+ "epoch": 21.600873362445416,
+ "grad_norm": 0.01683160290122032,
+ "learning_rate": 0.0006,
+ "loss": 4.349218368530273,
+ "step": 1555
+ },
+ {
+ "epoch": 21.61484716157205,
+ "grad_norm": 0.016973106190562248,
+ "learning_rate": 0.0006,
+ "loss": 4.375990390777588,
+ "step": 1556
+ },
+ {
+ "epoch": 21.62882096069869,
+ "grad_norm": 0.015688583254814148,
+ "learning_rate": 0.0006,
+ "loss": 4.408170223236084,
+ "step": 1557
+ },
+ {
+ "epoch": 21.64279475982533,
+ "grad_norm": 0.01573043316602707,
+ "learning_rate": 0.0006,
+ "loss": 4.416255950927734,
+ "step": 1558
+ },
+ {
+ "epoch": 21.656768558951963,
+ "grad_norm": 0.01744293048977852,
+ "learning_rate": 0.0006,
+ "loss": 4.477141380310059,
+ "step": 1559
+ },
+ {
+ "epoch": 21.670742358078602,
+ "grad_norm": 0.017503513023257256,
+ "learning_rate": 0.0006,
+ "loss": 4.3367486000061035,
+ "step": 1560
+ },
+ {
+ "epoch": 21.68471615720524,
+ "grad_norm": 0.016592150554060936,
+ "learning_rate": 0.0006,
+ "loss": 4.314785480499268,
+ "step": 1561
+ },
+ {
+ "epoch": 21.69868995633188,
+ "grad_norm": 0.015774084255099297,
+ "learning_rate": 0.0006,
+ "loss": 4.448209285736084,
+ "step": 1562
+ },
+ {
+ "epoch": 21.712663755458514,
+ "grad_norm": 0.015779150649905205,
+ "learning_rate": 0.0006,
+ "loss": 4.392690658569336,
+ "step": 1563
+ },
+ {
+ "epoch": 21.726637554585153,
+ "grad_norm": 0.015392648056149483,
+ "learning_rate": 0.0006,
+ "loss": 4.437741756439209,
+ "step": 1564
+ },
+ {
+ "epoch": 21.74061135371179,
+ "grad_norm": 0.01530836895108223,
+ "learning_rate": 0.0006,
+ "loss": 4.424595832824707,
+ "step": 1565
+ },
+ {
+ "epoch": 21.754585152838427,
+ "grad_norm": 0.015600779093801975,
+ "learning_rate": 0.0006,
+ "loss": 4.335417747497559,
+ "step": 1566
+ },
+ {
+ "epoch": 21.768558951965066,
+ "grad_norm": 0.014646518044173717,
+ "learning_rate": 0.0006,
+ "loss": 4.279081344604492,
+ "step": 1567
+ },
+ {
+ "epoch": 21.782532751091704,
+ "grad_norm": 0.014132864773273468,
+ "learning_rate": 0.0006,
+ "loss": 4.345950126647949,
+ "step": 1568
+ },
+ {
+ "epoch": 21.79650655021834,
+ "grad_norm": 0.01425724383443594,
+ "learning_rate": 0.0006,
+ "loss": 4.277588844299316,
+ "step": 1569
+ },
+ {
+ "epoch": 21.810480349344978,
+ "grad_norm": 0.015396572649478912,
+ "learning_rate": 0.0006,
+ "loss": 4.326196670532227,
+ "step": 1570
+ },
+ {
+ "epoch": 21.824454148471617,
+ "grad_norm": 0.01683669537305832,
+ "learning_rate": 0.0006,
+ "loss": 4.3294172286987305,
+ "step": 1571
+ },
+ {
+ "epoch": 21.83842794759825,
+ "grad_norm": 0.016365647315979004,
+ "learning_rate": 0.0006,
+ "loss": 4.523321628570557,
+ "step": 1572
+ },
+ {
+ "epoch": 21.85240174672489,
+ "grad_norm": 0.017142774537205696,
+ "learning_rate": 0.0006,
+ "loss": 4.475842475891113,
+ "step": 1573
+ },
+ {
+ "epoch": 21.86637554585153,
+ "grad_norm": 0.015950839966535568,
+ "learning_rate": 0.0006,
+ "loss": 4.397187232971191,
+ "step": 1574
+ },
+ {
+ "epoch": 21.880349344978168,
+ "grad_norm": 0.016180362552404404,
+ "learning_rate": 0.0006,
+ "loss": 4.414778709411621,
+ "step": 1575
+ },
+ {
+ "epoch": 21.894323144104803,
+ "grad_norm": 0.017615556716918945,
+ "learning_rate": 0.0006,
+ "loss": 4.348567962646484,
+ "step": 1576
+ },
+ {
+ "epoch": 21.90829694323144,
+ "grad_norm": 0.01896277628839016,
+ "learning_rate": 0.0006,
+ "loss": 4.431762218475342,
+ "step": 1577
+ },
+ {
+ "epoch": 21.92227074235808,
+ "grad_norm": 0.01990971714258194,
+ "learning_rate": 0.0006,
+ "loss": 4.431285858154297,
+ "step": 1578
+ },
+ {
+ "epoch": 21.936244541484715,
+ "grad_norm": 0.0201258547604084,
+ "learning_rate": 0.0006,
+ "loss": 4.309872627258301,
+ "step": 1579
+ },
+ {
+ "epoch": 21.950218340611354,
+ "grad_norm": 0.019620198756456375,
+ "learning_rate": 0.0006,
+ "loss": 4.484044551849365,
+ "step": 1580
+ },
+ {
+ "epoch": 21.964192139737992,
+ "grad_norm": 0.01771123707294464,
+ "learning_rate": 0.0006,
+ "loss": 4.368093013763428,
+ "step": 1581
+ },
+ {
+ "epoch": 21.978165938864628,
+ "grad_norm": 0.017501654103398323,
+ "learning_rate": 0.0006,
+ "loss": 4.372651100158691,
+ "step": 1582
+ },
+ {
+ "epoch": 21.992139737991266,
+ "grad_norm": 0.01693608984351158,
+ "learning_rate": 0.0006,
+ "loss": 4.379167556762695,
+ "step": 1583
+ },
+ {
+ "epoch": 22.0,
+ "grad_norm": 0.02002849616110325,
+ "learning_rate": 0.0006,
+ "loss": 4.329627990722656,
+ "step": 1584
+ },
+ {
+ "epoch": 22.0,
+ "eval_loss": 4.723696231842041,
+ "eval_runtime": 56.1758,
+ "eval_samples_per_second": 43.471,
+ "eval_steps_per_second": 1.371,
+ "step": 1584
+ },
+ {
+ "epoch": 22.01397379912664,
+ "grad_norm": 0.02030305378139019,
+ "learning_rate": 0.0006,
+ "loss": 4.355566024780273,
+ "step": 1585
+ },
+ {
+ "epoch": 22.027947598253274,
+ "grad_norm": 0.018638934940099716,
+ "learning_rate": 0.0006,
+ "loss": 4.314009189605713,
+ "step": 1586
+ },
+ {
+ "epoch": 22.041921397379912,
+ "grad_norm": 0.019917072728276253,
+ "learning_rate": 0.0006,
+ "loss": 4.367525577545166,
+ "step": 1587
+ },
+ {
+ "epoch": 22.05589519650655,
+ "grad_norm": 0.020785167813301086,
+ "learning_rate": 0.0006,
+ "loss": 4.375561714172363,
+ "step": 1588
+ },
+ {
+ "epoch": 22.069868995633186,
+ "grad_norm": 0.021214453503489494,
+ "learning_rate": 0.0006,
+ "loss": 4.418305397033691,
+ "step": 1589
+ },
+ {
+ "epoch": 22.083842794759825,
+ "grad_norm": 0.019051678478717804,
+ "learning_rate": 0.0006,
+ "loss": 4.215126991271973,
+ "step": 1590
+ },
+ {
+ "epoch": 22.097816593886463,
+ "grad_norm": 0.02323366142809391,
+ "learning_rate": 0.0006,
+ "loss": 4.306807994842529,
+ "step": 1591
+ },
+ {
+ "epoch": 22.111790393013102,
+ "grad_norm": 0.026186738163232803,
+ "learning_rate": 0.0006,
+ "loss": 4.408508777618408,
+ "step": 1592
+ },
+ {
+ "epoch": 22.125764192139737,
+ "grad_norm": 0.027361227199435234,
+ "learning_rate": 0.0006,
+ "loss": 4.329322814941406,
+ "step": 1593
+ },
+ {
+ "epoch": 22.139737991266376,
+ "grad_norm": 0.026564646512269974,
+ "learning_rate": 0.0006,
+ "loss": 4.346642017364502,
+ "step": 1594
+ },
+ {
+ "epoch": 22.153711790393015,
+ "grad_norm": 0.025073667988181114,
+ "learning_rate": 0.0006,
+ "loss": 4.3984198570251465,
+ "step": 1595
+ },
+ {
+ "epoch": 22.16768558951965,
+ "grad_norm": 0.02347794733941555,
+ "learning_rate": 0.0006,
+ "loss": 4.352538108825684,
+ "step": 1596
+ },
+ {
+ "epoch": 22.18165938864629,
+ "grad_norm": 0.02203468047082424,
+ "learning_rate": 0.0006,
+ "loss": 4.298148155212402,
+ "step": 1597
+ },
+ {
+ "epoch": 22.195633187772927,
+ "grad_norm": 0.020621957257390022,
+ "learning_rate": 0.0006,
+ "loss": 4.480309963226318,
+ "step": 1598
+ },
+ {
+ "epoch": 22.209606986899562,
+ "grad_norm": 0.023710055276751518,
+ "learning_rate": 0.0006,
+ "loss": 4.339317321777344,
+ "step": 1599
+ },
+ {
+ "epoch": 22.2235807860262,
+ "grad_norm": 0.0233482476323843,
+ "learning_rate": 0.0006,
+ "loss": 4.436896324157715,
+ "step": 1600
+ },
+ {
+ "epoch": 22.23755458515284,
+ "grad_norm": 0.020122403278946877,
+ "learning_rate": 0.0006,
+ "loss": 4.236002445220947,
+ "step": 1601
+ },
+ {
+ "epoch": 22.251528384279474,
+ "grad_norm": 0.01957266591489315,
+ "learning_rate": 0.0006,
+ "loss": 4.497328281402588,
+ "step": 1602
+ },
+ {
+ "epoch": 22.265502183406113,
+ "grad_norm": 0.018964895978569984,
+ "learning_rate": 0.0006,
+ "loss": 4.36360502243042,
+ "step": 1603
+ },
+ {
+ "epoch": 22.27947598253275,
+ "grad_norm": 0.018902769312262535,
+ "learning_rate": 0.0006,
+ "loss": 4.427093505859375,
+ "step": 1604
+ },
+ {
+ "epoch": 22.29344978165939,
+ "grad_norm": 0.018918119370937347,
+ "learning_rate": 0.0006,
+ "loss": 4.294950485229492,
+ "step": 1605
+ },
+ {
+ "epoch": 22.307423580786025,
+ "grad_norm": 0.019315775483846664,
+ "learning_rate": 0.0006,
+ "loss": 4.427952289581299,
+ "step": 1606
+ },
+ {
+ "epoch": 22.321397379912664,
+ "grad_norm": 0.018350522965192795,
+ "learning_rate": 0.0006,
+ "loss": 4.272945404052734,
+ "step": 1607
+ },
+ {
+ "epoch": 22.335371179039303,
+ "grad_norm": 0.01759534701704979,
+ "learning_rate": 0.0006,
+ "loss": 4.348023891448975,
+ "step": 1608
+ },
+ {
+ "epoch": 22.349344978165938,
+ "grad_norm": 0.01646057888865471,
+ "learning_rate": 0.0006,
+ "loss": 4.278518199920654,
+ "step": 1609
+ },
+ {
+ "epoch": 22.363318777292577,
+ "grad_norm": 0.015875060111284256,
+ "learning_rate": 0.0006,
+ "loss": 4.423527240753174,
+ "step": 1610
+ },
+ {
+ "epoch": 22.377292576419215,
+ "grad_norm": 0.01638529635965824,
+ "learning_rate": 0.0006,
+ "loss": 4.321723937988281,
+ "step": 1611
+ },
+ {
+ "epoch": 22.39126637554585,
+ "grad_norm": 0.016942061483860016,
+ "learning_rate": 0.0006,
+ "loss": 4.346164703369141,
+ "step": 1612
+ },
+ {
+ "epoch": 22.40524017467249,
+ "grad_norm": 0.015209621749818325,
+ "learning_rate": 0.0006,
+ "loss": 4.434340000152588,
+ "step": 1613
+ },
+ {
+ "epoch": 22.419213973799128,
+ "grad_norm": 0.014423009008169174,
+ "learning_rate": 0.0006,
+ "loss": 4.342560768127441,
+ "step": 1614
+ },
+ {
+ "epoch": 22.433187772925763,
+ "grad_norm": 0.014369525946676731,
+ "learning_rate": 0.0006,
+ "loss": 4.296586990356445,
+ "step": 1615
+ },
+ {
+ "epoch": 22.4471615720524,
+ "grad_norm": 0.01392540242522955,
+ "learning_rate": 0.0006,
+ "loss": 4.433176040649414,
+ "step": 1616
+ },
+ {
+ "epoch": 22.46113537117904,
+ "grad_norm": 0.013206899166107178,
+ "learning_rate": 0.0006,
+ "loss": 4.32962703704834,
+ "step": 1617
+ },
+ {
+ "epoch": 22.475109170305675,
+ "grad_norm": 0.013427077792584896,
+ "learning_rate": 0.0006,
+ "loss": 4.420437335968018,
+ "step": 1618
+ },
+ {
+ "epoch": 22.489082969432314,
+ "grad_norm": 0.014194196090102196,
+ "learning_rate": 0.0006,
+ "loss": 4.410478591918945,
+ "step": 1619
+ },
+ {
+ "epoch": 22.503056768558952,
+ "grad_norm": 0.01362364087253809,
+ "learning_rate": 0.0006,
+ "loss": 4.338932037353516,
+ "step": 1620
+ },
+ {
+ "epoch": 22.51703056768559,
+ "grad_norm": 0.014299589209258556,
+ "learning_rate": 0.0006,
+ "loss": 4.34089994430542,
+ "step": 1621
+ },
+ {
+ "epoch": 22.531004366812226,
+ "grad_norm": 0.014061295427381992,
+ "learning_rate": 0.0006,
+ "loss": 4.29201602935791,
+ "step": 1622
+ },
+ {
+ "epoch": 22.544978165938865,
+ "grad_norm": 0.013945229351520538,
+ "learning_rate": 0.0006,
+ "loss": 4.326052188873291,
+ "step": 1623
+ },
+ {
+ "epoch": 22.558951965065503,
+ "grad_norm": 0.01387068536132574,
+ "learning_rate": 0.0006,
+ "loss": 4.393521308898926,
+ "step": 1624
+ },
+ {
+ "epoch": 22.57292576419214,
+ "grad_norm": 0.013931898400187492,
+ "learning_rate": 0.0006,
+ "loss": 4.3739752769470215,
+ "step": 1625
+ },
+ {
+ "epoch": 22.586899563318777,
+ "grad_norm": 0.01497740764170885,
+ "learning_rate": 0.0006,
+ "loss": 4.3991618156433105,
+ "step": 1626
+ },
+ {
+ "epoch": 22.600873362445416,
+ "grad_norm": 0.015842905268073082,
+ "learning_rate": 0.0006,
+ "loss": 4.405340671539307,
+ "step": 1627
+ },
+ {
+ "epoch": 22.61484716157205,
+ "grad_norm": 0.01580633595585823,
+ "learning_rate": 0.0006,
+ "loss": 4.280102729797363,
+ "step": 1628
+ },
+ {
+ "epoch": 22.62882096069869,
+ "grad_norm": 0.016567382961511612,
+ "learning_rate": 0.0006,
+ "loss": 4.407657146453857,
+ "step": 1629
+ },
+ {
+ "epoch": 22.64279475982533,
+ "grad_norm": 0.01648622378706932,
+ "learning_rate": 0.0006,
+ "loss": 4.392590522766113,
+ "step": 1630
+ },
+ {
+ "epoch": 22.656768558951963,
+ "grad_norm": 0.017066307365894318,
+ "learning_rate": 0.0006,
+ "loss": 4.324862003326416,
+ "step": 1631
+ },
+ {
+ "epoch": 22.670742358078602,
+ "grad_norm": 0.01637062057852745,
+ "learning_rate": 0.0006,
+ "loss": 4.34598445892334,
+ "step": 1632
+ },
+ {
+ "epoch": 22.68471615720524,
+ "grad_norm": 0.018038960173726082,
+ "learning_rate": 0.0006,
+ "loss": 4.320711612701416,
+ "step": 1633
+ },
+ {
+ "epoch": 22.69868995633188,
+ "grad_norm": 0.018096428364515305,
+ "learning_rate": 0.0006,
+ "loss": 4.536527156829834,
+ "step": 1634
+ },
+ {
+ "epoch": 22.712663755458514,
+ "grad_norm": 0.017889291048049927,
+ "learning_rate": 0.0006,
+ "loss": 4.388131141662598,
+ "step": 1635
+ },
+ {
+ "epoch": 22.726637554585153,
+ "grad_norm": 0.018771521747112274,
+ "learning_rate": 0.0006,
+ "loss": 4.357668876647949,
+ "step": 1636
+ },
+ {
+ "epoch": 22.74061135371179,
+ "grad_norm": 0.02100873365998268,
+ "learning_rate": 0.0006,
+ "loss": 4.293972015380859,
+ "step": 1637
+ },
+ {
+ "epoch": 22.754585152838427,
+ "grad_norm": 0.017715785652399063,
+ "learning_rate": 0.0006,
+ "loss": 4.349849224090576,
+ "step": 1638
+ },
+ {
+ "epoch": 22.768558951965066,
+ "grad_norm": 0.015546333976089954,
+ "learning_rate": 0.0006,
+ "loss": 4.311728000640869,
+ "step": 1639
+ },
+ {
+ "epoch": 22.782532751091704,
+ "grad_norm": 0.01605990156531334,
+ "learning_rate": 0.0006,
+ "loss": 4.499123573303223,
+ "step": 1640
+ },
+ {
+ "epoch": 22.79650655021834,
+ "grad_norm": 0.016642142087221146,
+ "learning_rate": 0.0006,
+ "loss": 4.4319257736206055,
+ "step": 1641
+ },
+ {
+ "epoch": 22.810480349344978,
+ "grad_norm": 0.016120387241244316,
+ "learning_rate": 0.0006,
+ "loss": 4.4373650550842285,
+ "step": 1642
+ },
+ {
+ "epoch": 22.824454148471617,
+ "grad_norm": 0.015174890868365765,
+ "learning_rate": 0.0006,
+ "loss": 4.314082145690918,
+ "step": 1643
+ },
+ {
+ "epoch": 22.83842794759825,
+ "grad_norm": 0.01551714539527893,
+ "learning_rate": 0.0006,
+ "loss": 4.360363960266113,
+ "step": 1644
+ },
+ {
+ "epoch": 22.85240174672489,
+ "grad_norm": 0.014966742135584354,
+ "learning_rate": 0.0006,
+ "loss": 4.346253395080566,
+ "step": 1645
+ },
+ {
+ "epoch": 22.86637554585153,
+ "grad_norm": 0.014288028702139854,
+ "learning_rate": 0.0006,
+ "loss": 4.464366436004639,
+ "step": 1646
+ },
+ {
+ "epoch": 22.880349344978168,
+ "grad_norm": 0.015291682444512844,
+ "learning_rate": 0.0006,
+ "loss": 4.504101753234863,
+ "step": 1647
+ },
+ {
+ "epoch": 22.894323144104803,
+ "grad_norm": 0.014534426853060722,
+ "learning_rate": 0.0006,
+ "loss": 4.462172031402588,
+ "step": 1648
+ },
+ {
+ "epoch": 22.90829694323144,
+ "grad_norm": 0.014129819348454475,
+ "learning_rate": 0.0006,
+ "loss": 4.350126266479492,
+ "step": 1649
+ },
+ {
+ "epoch": 22.92227074235808,
+ "grad_norm": 0.014932668767869473,
+ "learning_rate": 0.0006,
+ "loss": 4.374392032623291,
+ "step": 1650
+ },
+ {
+ "epoch": 22.936244541484715,
+ "grad_norm": 0.014647853560745716,
+ "learning_rate": 0.0006,
+ "loss": 4.252264499664307,
+ "step": 1651
+ },
+ {
+ "epoch": 22.950218340611354,
+ "grad_norm": 0.013553360477089882,
+ "learning_rate": 0.0006,
+ "loss": 4.379729270935059,
+ "step": 1652
+ },
+ {
+ "epoch": 22.964192139737992,
+ "grad_norm": 0.014606194570660591,
+ "learning_rate": 0.0006,
+ "loss": 4.363372802734375,
+ "step": 1653
+ },
+ {
+ "epoch": 22.978165938864628,
+ "grad_norm": 0.015000631101429462,
+ "learning_rate": 0.0006,
+ "loss": 4.408048629760742,
+ "step": 1654
+ },
+ {
+ "epoch": 22.992139737991266,
+ "grad_norm": 0.015370816923677921,
+ "learning_rate": 0.0006,
+ "loss": 4.297451019287109,
+ "step": 1655
+ },
+ {
+ "epoch": 23.0,
+ "grad_norm": 0.01709144562482834,
+ "learning_rate": 0.0006,
+ "loss": 4.294776916503906,
+ "step": 1656
+ },
+ {
+ "epoch": 23.0,
+ "eval_loss": 4.678618907928467,
+ "eval_runtime": 56.1668,
+ "eval_samples_per_second": 43.478,
+ "eval_steps_per_second": 1.371,
+ "step": 1656
+ },
+ {
+ "epoch": 23.01397379912664,
+ "grad_norm": 0.01727290451526642,
+ "learning_rate": 0.0006,
+ "loss": 4.311701774597168,
+ "step": 1657
+ },
+ {
+ "epoch": 23.027947598253274,
+ "grad_norm": 0.020645759999752045,
+ "learning_rate": 0.0006,
+ "loss": 4.429897308349609,
+ "step": 1658
+ },
+ {
+ "epoch": 23.041921397379912,
+ "grad_norm": 0.02204226516187191,
+ "learning_rate": 0.0006,
+ "loss": 4.365151882171631,
+ "step": 1659
+ },
+ {
+ "epoch": 23.05589519650655,
+ "grad_norm": 0.023069269955158234,
+ "learning_rate": 0.0006,
+ "loss": 4.3710832595825195,
+ "step": 1660
+ },
+ {
+ "epoch": 23.069868995633186,
+ "grad_norm": 0.024151179939508438,
+ "learning_rate": 0.0006,
+ "loss": 4.403108596801758,
+ "step": 1661
+ },
+ {
+ "epoch": 23.083842794759825,
+ "grad_norm": 0.02511233650147915,
+ "learning_rate": 0.0006,
+ "loss": 4.414128303527832,
+ "step": 1662
+ },
+ {
+ "epoch": 23.097816593886463,
+ "grad_norm": 0.025802113115787506,
+ "learning_rate": 0.0006,
+ "loss": 4.388265609741211,
+ "step": 1663
+ },
+ {
+ "epoch": 23.111790393013102,
+ "grad_norm": 0.02377348765730858,
+ "learning_rate": 0.0006,
+ "loss": 4.357505798339844,
+ "step": 1664
+ },
+ {
+ "epoch": 23.125764192139737,
+ "grad_norm": 0.020876651629805565,
+ "learning_rate": 0.0006,
+ "loss": 4.23043966293335,
+ "step": 1665
+ },
+ {
+ "epoch": 23.139737991266376,
+ "grad_norm": 0.01959727331995964,
+ "learning_rate": 0.0006,
+ "loss": 4.280970096588135,
+ "step": 1666
+ },
+ {
+ "epoch": 23.153711790393015,
+ "grad_norm": 0.018678616732358932,
+ "learning_rate": 0.0006,
+ "loss": 4.346716403961182,
+ "step": 1667
+ },
+ {
+ "epoch": 23.16768558951965,
+ "grad_norm": 0.02108944021165371,
+ "learning_rate": 0.0006,
+ "loss": 4.277904510498047,
+ "step": 1668
+ },
+ {
+ "epoch": 23.18165938864629,
+ "grad_norm": 0.020221566781401634,
+ "learning_rate": 0.0006,
+ "loss": 4.367693901062012,
+ "step": 1669
+ },
+ {
+ "epoch": 23.195633187772927,
+ "grad_norm": 0.019599711522459984,
+ "learning_rate": 0.0006,
+ "loss": 4.257092475891113,
+ "step": 1670
+ },
+ {
+ "epoch": 23.209606986899562,
+ "grad_norm": 0.019833361729979515,
+ "learning_rate": 0.0006,
+ "loss": 4.540737628936768,
+ "step": 1671
+ },
+ {
+ "epoch": 23.2235807860262,
+ "grad_norm": 0.020352095365524292,
+ "learning_rate": 0.0006,
+ "loss": 4.360923767089844,
+ "step": 1672
+ },
+ {
+ "epoch": 23.23755458515284,
+ "grad_norm": 0.01939748041331768,
+ "learning_rate": 0.0006,
+ "loss": 4.393110275268555,
+ "step": 1673
+ },
+ {
+ "epoch": 23.251528384279474,
+ "grad_norm": 0.01884354278445244,
+ "learning_rate": 0.0006,
+ "loss": 4.383938789367676,
+ "step": 1674
+ },
+ {
+ "epoch": 23.265502183406113,
+ "grad_norm": 0.02008470520377159,
+ "learning_rate": 0.0006,
+ "loss": 4.346344947814941,
+ "step": 1675
+ },
+ {
+ "epoch": 23.27947598253275,
+ "grad_norm": 0.019065195694565773,
+ "learning_rate": 0.0006,
+ "loss": 4.36611795425415,
+ "step": 1676
+ },
+ {
+ "epoch": 23.29344978165939,
+ "grad_norm": 0.01938541606068611,
+ "learning_rate": 0.0006,
+ "loss": 4.4009599685668945,
+ "step": 1677
+ },
+ {
+ "epoch": 23.307423580786025,
+ "grad_norm": 0.019044723361730576,
+ "learning_rate": 0.0006,
+ "loss": 4.340169906616211,
+ "step": 1678
+ },
+ {
+ "epoch": 23.321397379912664,
+ "grad_norm": 0.018850378692150116,
+ "learning_rate": 0.0006,
+ "loss": 4.37674617767334,
+ "step": 1679
+ },
+ {
+ "epoch": 23.335371179039303,
+ "grad_norm": 0.019076095893979073,
+ "learning_rate": 0.0006,
+ "loss": 4.226874351501465,
+ "step": 1680
+ },
+ {
+ "epoch": 23.349344978165938,
+ "grad_norm": 0.017543038353323936,
+ "learning_rate": 0.0006,
+ "loss": 4.337751388549805,
+ "step": 1681
+ },
+ {
+ "epoch": 23.363318777292577,
+ "grad_norm": 0.017170218750834465,
+ "learning_rate": 0.0006,
+ "loss": 4.282512664794922,
+ "step": 1682
+ },
+ {
+ "epoch": 23.377292576419215,
+ "grad_norm": 0.01574796624481678,
+ "learning_rate": 0.0006,
+ "loss": 4.30638313293457,
+ "step": 1683
+ },
+ {
+ "epoch": 23.39126637554585,
+ "grad_norm": 0.015160846523940563,
+ "learning_rate": 0.0006,
+ "loss": 4.336716175079346,
+ "step": 1684
+ },
+ {
+ "epoch": 23.40524017467249,
+ "grad_norm": 0.01534526702016592,
+ "learning_rate": 0.0006,
+ "loss": 4.2174787521362305,
+ "step": 1685
+ },
+ {
+ "epoch": 23.419213973799128,
+ "grad_norm": 0.015289999544620514,
+ "learning_rate": 0.0006,
+ "loss": 4.272546291351318,
+ "step": 1686
+ },
+ {
+ "epoch": 23.433187772925763,
+ "grad_norm": 0.014894185587763786,
+ "learning_rate": 0.0006,
+ "loss": 4.296045303344727,
+ "step": 1687
+ },
+ {
+ "epoch": 23.4471615720524,
+ "grad_norm": 0.014869502745568752,
+ "learning_rate": 0.0006,
+ "loss": 4.425461769104004,
+ "step": 1688
+ },
+ {
+ "epoch": 23.46113537117904,
+ "grad_norm": 0.01474483497440815,
+ "learning_rate": 0.0006,
+ "loss": 4.361867427825928,
+ "step": 1689
+ },
+ {
+ "epoch": 23.475109170305675,
+ "grad_norm": 0.015299245715141296,
+ "learning_rate": 0.0006,
+ "loss": 4.364523410797119,
+ "step": 1690
+ },
+ {
+ "epoch": 23.489082969432314,
+ "grad_norm": 0.015201340429484844,
+ "learning_rate": 0.0006,
+ "loss": 4.285182952880859,
+ "step": 1691
+ },
+ {
+ "epoch": 23.503056768558952,
+ "grad_norm": 0.016057293862104416,
+ "learning_rate": 0.0006,
+ "loss": 4.284873962402344,
+ "step": 1692
+ },
+ {
+ "epoch": 23.51703056768559,
+ "grad_norm": 0.015871062874794006,
+ "learning_rate": 0.0006,
+ "loss": 4.326470851898193,
+ "step": 1693
+ },
+ {
+ "epoch": 23.531004366812226,
+ "grad_norm": 0.01586288772523403,
+ "learning_rate": 0.0006,
+ "loss": 4.324883460998535,
+ "step": 1694
+ },
+ {
+ "epoch": 23.544978165938865,
+ "grad_norm": 0.016092827543616295,
+ "learning_rate": 0.0006,
+ "loss": 4.3258867263793945,
+ "step": 1695
+ },
+ {
+ "epoch": 23.558951965065503,
+ "grad_norm": 0.01582016795873642,
+ "learning_rate": 0.0006,
+ "loss": 4.358892440795898,
+ "step": 1696
+ },
+ {
+ "epoch": 23.57292576419214,
+ "grad_norm": 0.01688452623784542,
+ "learning_rate": 0.0006,
+ "loss": 4.280842304229736,
+ "step": 1697
+ },
+ {
+ "epoch": 23.586899563318777,
+ "grad_norm": 0.015865886583924294,
+ "learning_rate": 0.0006,
+ "loss": 4.245567321777344,
+ "step": 1698
+ },
+ {
+ "epoch": 23.600873362445416,
+ "grad_norm": 0.015303662046790123,
+ "learning_rate": 0.0006,
+ "loss": 4.383112907409668,
+ "step": 1699
+ },
+ {
+ "epoch": 23.61484716157205,
+ "grad_norm": 0.015649516135454178,
+ "learning_rate": 0.0006,
+ "loss": 4.519082069396973,
+ "step": 1700
+ },
+ {
+ "epoch": 23.62882096069869,
+ "grad_norm": 0.0167120099067688,
+ "learning_rate": 0.0006,
+ "loss": 4.447847366333008,
+ "step": 1701
+ },
+ {
+ "epoch": 23.64279475982533,
+ "grad_norm": 0.016368716955184937,
+ "learning_rate": 0.0006,
+ "loss": 4.434186935424805,
+ "step": 1702
+ },
+ {
+ "epoch": 23.656768558951963,
+ "grad_norm": 0.017379503697156906,
+ "learning_rate": 0.0006,
+ "loss": 4.317594528198242,
+ "step": 1703
+ },
+ {
+ "epoch": 23.670742358078602,
+ "grad_norm": 0.018573811277747154,
+ "learning_rate": 0.0006,
+ "loss": 4.205793857574463,
+ "step": 1704
+ },
+ {
+ "epoch": 23.68471615720524,
+ "grad_norm": 0.01749090477824211,
+ "learning_rate": 0.0006,
+ "loss": 4.3985700607299805,
+ "step": 1705
+ },
+ {
+ "epoch": 23.69868995633188,
+ "grad_norm": 0.0165668074041605,
+ "learning_rate": 0.0006,
+ "loss": 4.329226970672607,
+ "step": 1706
+ },
+ {
+ "epoch": 23.712663755458514,
+ "grad_norm": 0.016913846135139465,
+ "learning_rate": 0.0006,
+ "loss": 4.388485908508301,
+ "step": 1707
+ },
+ {
+ "epoch": 23.726637554585153,
+ "grad_norm": 0.016813859343528748,
+ "learning_rate": 0.0006,
+ "loss": 4.3395915031433105,
+ "step": 1708
+ },
+ {
+ "epoch": 23.74061135371179,
+ "grad_norm": 0.01676975190639496,
+ "learning_rate": 0.0006,
+ "loss": 4.254745960235596,
+ "step": 1709
+ },
+ {
+ "epoch": 23.754585152838427,
+ "grad_norm": 0.016370350494980812,
+ "learning_rate": 0.0006,
+ "loss": 4.2612223625183105,
+ "step": 1710
+ },
+ {
+ "epoch": 23.768558951965066,
+ "grad_norm": 0.01696198247373104,
+ "learning_rate": 0.0006,
+ "loss": 4.327524662017822,
+ "step": 1711
+ },
+ {
+ "epoch": 23.782532751091704,
+ "grad_norm": 0.0189108457416296,
+ "learning_rate": 0.0006,
+ "loss": 4.325839519500732,
+ "step": 1712
+ },
+ {
+ "epoch": 23.79650655021834,
+ "grad_norm": 0.01862351782619953,
+ "learning_rate": 0.0006,
+ "loss": 4.333662509918213,
+ "step": 1713
+ },
+ {
+ "epoch": 23.810480349344978,
+ "grad_norm": 0.01655014418065548,
+ "learning_rate": 0.0006,
+ "loss": 4.39305305480957,
+ "step": 1714
+ },
+ {
+ "epoch": 23.824454148471617,
+ "grad_norm": 0.015899376943707466,
+ "learning_rate": 0.0006,
+ "loss": 4.390053749084473,
+ "step": 1715
+ },
+ {
+ "epoch": 23.83842794759825,
+ "grad_norm": 0.016519056633114815,
+ "learning_rate": 0.0006,
+ "loss": 4.255903244018555,
+ "step": 1716
+ },
+ {
+ "epoch": 23.85240174672489,
+ "grad_norm": 0.0167539119720459,
+ "learning_rate": 0.0006,
+ "loss": 4.255727767944336,
+ "step": 1717
+ },
+ {
+ "epoch": 23.86637554585153,
+ "grad_norm": 0.0171353816986084,
+ "learning_rate": 0.0006,
+ "loss": 4.376740455627441,
+ "step": 1718
+ },
+ {
+ "epoch": 23.880349344978168,
+ "grad_norm": 0.016719117760658264,
+ "learning_rate": 0.0006,
+ "loss": 4.3376688957214355,
+ "step": 1719
+ },
+ {
+ "epoch": 23.894323144104803,
+ "grad_norm": 0.01561494916677475,
+ "learning_rate": 0.0006,
+ "loss": 4.3345746994018555,
+ "step": 1720
+ },
+ {
+ "epoch": 23.90829694323144,
+ "grad_norm": 0.016303174197673798,
+ "learning_rate": 0.0006,
+ "loss": 4.332371234893799,
+ "step": 1721
+ },
+ {
+ "epoch": 23.92227074235808,
+ "grad_norm": 0.016722865402698517,
+ "learning_rate": 0.0006,
+ "loss": 4.267125129699707,
+ "step": 1722
+ },
+ {
+ "epoch": 23.936244541484715,
+ "grad_norm": 0.017072124406695366,
+ "learning_rate": 0.0006,
+ "loss": 4.322751998901367,
+ "step": 1723
+ },
+ {
+ "epoch": 23.950218340611354,
+ "grad_norm": 0.01699932850897312,
+ "learning_rate": 0.0006,
+ "loss": 4.378422737121582,
+ "step": 1724
+ },
+ {
+ "epoch": 23.964192139737992,
+ "grad_norm": 0.01742720976471901,
+ "learning_rate": 0.0006,
+ "loss": 4.345083236694336,
+ "step": 1725
+ },
+ {
+ "epoch": 23.978165938864628,
+ "grad_norm": 0.018797501921653748,
+ "learning_rate": 0.0006,
+ "loss": 4.290443420410156,
+ "step": 1726
+ },
+ {
+ "epoch": 23.992139737991266,
+ "grad_norm": 0.018090683966875076,
+ "learning_rate": 0.0006,
+ "loss": 4.377580642700195,
+ "step": 1727
+ },
+ {
+ "epoch": 24.0,
+ "grad_norm": 0.017305459827184677,
+ "learning_rate": 0.0006,
+ "loss": 4.364426612854004,
+ "step": 1728
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 28800,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 400,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": true,
+ "should_log": false,
+ "should_save": false,
+ "should_training_stop": false
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 7.367072067197338e+17,
+ "train_batch_size": 8,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-1728/training_args.bin b/runs/l2r50-i2-fulle-lm/checkpoint-1728/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..cad232831c222cc7f8b938b8bbe42f61e7d89c3a
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-1728/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:f7c40e3bfa2f5ac1450ca1243e3f2a895d44a06ed5995e6aefd97f78d229d8c8
+size 4792
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2016/chat_template.jinja b/runs/l2r50-i2-fulle-lm/checkpoint-2016/chat_template.jinja
new file mode 100644
index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-2016/chat_template.jinja
@@ -0,0 +1,85 @@
+{%- if tools %}
+ {{- '<|im_start|>system\n' }}
+ {%- if messages[0].role == 'system' %}
+ {{- messages[0].content + '\n\n' }}
+ {%- endif %}
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }}
+ {%- for tool in tools %}
+ {{- "\n" }}
+ {{- tool | tojson }}
+ {%- endfor %}
+ {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }}
+{%- else %}
+ {%- if messages[0].role == 'system' %}
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
+ {%- endif %}
+{%- endif %}
+{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
+{%- for message in messages[::-1] %}
+ {%- set index = (messages|length - 1) - loop.index0 %}
+ {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %}
+ {%- set ns.multi_step_tool = false %}
+ {%- set ns.last_query_index = index %}
+ {%- endif %}
+{%- endfor %}
+{%- for message in messages %}
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
+ {%- elif message.role == "assistant" %}
+ {%- set content = message.content %}
+ {%- set reasoning_content = '' %}
+ {%- if message.reasoning_content is defined and message.reasoning_content is not none %}
+ {%- set reasoning_content = message.reasoning_content %}
+ {%- else %}
+ {%- if '' in message.content %}
+ {%- set content = message.content.split('')[-1].lstrip('\n') %}
+ {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %}
+ {%- endif %}
+ {%- endif %}
+ {%- if loop.index0 > ns.last_query_index %}
+ {%- if loop.last or (not loop.last and reasoning_content) %}
+ {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- if message.tool_calls %}
+ {%- for tool_call in message.tool_calls %}
+ {%- if (loop.first and content) or (not loop.first) %}
+ {{- '\n' }}
+ {%- endif %}
+ {%- if tool_call.function %}
+ {%- set tool_call = tool_call.function %}
+ {%- endif %}
+ {{- '\n{"name": "' }}
+ {{- tool_call.name }}
+ {{- '", "arguments": ' }}
+ {%- if tool_call.arguments is string %}
+ {{- tool_call.arguments }}
+ {%- else %}
+ {{- tool_call.arguments | tojson }}
+ {%- endif %}
+ {{- '}\n' }}
+ {%- endfor %}
+ {%- endif %}
+ {{- '<|im_end|>\n' }}
+ {%- elif message.role == "tool" %}
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
+ {{- '<|im_start|>user' }}
+ {%- endif %}
+ {{- '\n\n' }}
+ {{- message.content }}
+ {{- '\n' }}
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
+ {{- '<|im_end|>\n' }}
+ {%- endif %}
+ {%- endif %}
+{%- endfor %}
+{%- if add_generation_prompt %}
+ {{- '<|im_start|>assistant\n' }}
+ {%- if enable_thinking is defined and enable_thinking is false %}
+ {{- '\n\n\n\n' }}
+ {%- endif %}
+{%- endif %}
\ No newline at end of file
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2016/config.json b/runs/l2r50-i2-fulle-lm/checkpoint-2016/config.json
new file mode 100644
index 0000000000000000000000000000000000000000..22f250953c6ce8fc91ee4411856dd73041a415e2
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-2016/config.json
@@ -0,0 +1,32 @@
+{
+ "architectures": [
+ "LlamaForCausalLM"
+ ],
+ "attention_bias": false,
+ "attention_dropout": 0.0,
+ "bos_token_id": null,
+ "dtype": "float32",
+ "eos_token_id": 151645,
+ "head_dim": 128,
+ "hidden_act": "silu",
+ "hidden_size": 512,
+ "initializer_range": 0.02,
+ "intermediate_size": 1536,
+ "max_position_embeddings": 2048,
+ "mlp_bias": false,
+ "model_type": "llama",
+ "num_attention_heads": 4,
+ "num_hidden_layers": 20,
+ "num_key_value_heads": 4,
+ "pad_token_id": 151645,
+ "pretraining_tp": 1,
+ "rms_norm_eps": 1e-06,
+ "rope_parameters": {
+ "rope_theta": 10000.0,
+ "rope_type": "default"
+ },
+ "tie_word_embeddings": true,
+ "transformers_version": "5.5.0",
+ "use_cache": false,
+ "vocab_size": 151673
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2016/generation_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-2016/generation_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-2016/generation_config.json
@@ -0,0 +1,11 @@
+{
+ "_from_model_config": true,
+ "eos_token_id": [
+ 151645
+ ],
+ "output_attentions": false,
+ "output_hidden_states": false,
+ "pad_token_id": 151645,
+ "transformers_version": "5.5.0",
+ "use_cache": true
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2016/model.safetensors b/runs/l2r50-i2-fulle-lm/checkpoint-2016/model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..609c1b93849f9cb071c4d3b6968fa93a69562b4f
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-2016/model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:e00c75ff786934715e62f92203ed4883abf95a65b5ffb07ef440b9fb7088a976
+size 583360328
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2016/optimizer.pt b/runs/l2r50-i2-fulle-lm/checkpoint-2016/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..622ef6b55f2312aa6e7055620c9d6684fbea5193
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-2016/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:d2ef370b20e344a29bc7d95c4e6a4be06765f0073271e325dee5354c06500cc7
+size 1166833530
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2016/rng_state_0.pth b/runs/l2r50-i2-fulle-lm/checkpoint-2016/rng_state_0.pth
new file mode 100644
index 0000000000000000000000000000000000000000..fe1a4bdbccb862d60c07f9b4d54d6913589aa4a8
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-2016/rng_state_0.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:e8dcc5baf9a16093438252563827f7b6a020c20d8bdd59a2434185bca6b10d55
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2016/rng_state_1.pth b/runs/l2r50-i2-fulle-lm/checkpoint-2016/rng_state_1.pth
new file mode 100644
index 0000000000000000000000000000000000000000..454a4ec443a59d30815357e097257bfe87492944
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-2016/rng_state_1.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:0d1f4cbabe5810040153ebf881570d58f8d5b569da22e6703989d9622df57a2b
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2016/rng_state_2.pth b/runs/l2r50-i2-fulle-lm/checkpoint-2016/rng_state_2.pth
new file mode 100644
index 0000000000000000000000000000000000000000..7281baf097ae7539af2086a75ab2f4aabf11a454
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-2016/rng_state_2.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:467348aa2353e1b0a97dac51f9a54d62cebca24058072cc53462eca16b886251
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2016/rng_state_3.pth b/runs/l2r50-i2-fulle-lm/checkpoint-2016/rng_state_3.pth
new file mode 100644
index 0000000000000000000000000000000000000000..37625fad4a311eebea8762bceb0d3b43cd6fbd3d
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-2016/rng_state_3.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:958522b68e2cb9ce69ad44e4ae6a2ea4c6ce0bc22433642dce3b33306f5be3fc
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2016/scheduler.pt b/runs/l2r50-i2-fulle-lm/checkpoint-2016/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..b92e75de04edb55fd4d6a1778589870a5bc904d0
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-2016/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:71f52a3ae3412b2053eb61b18dd446a6b3125f82036fb38a407b55123d55fdf7
+size 1064
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2016/tokenizer.json b/runs/l2r50-i2-fulle-lm/checkpoint-2016/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..53786016f070e11184813f3c83101dc5e93e392e
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-2016/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:c3c63fda1646a555448e8f03662542e04b11d103721319e4d9b267d7da03ceb8
+size 11424463
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2016/tokenizer_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-2016/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..98e2fb9b2cc54785dc9d871b8e503f8b8f2defe7
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-2016/tokenizer_config.json
@@ -0,0 +1,20 @@
+{
+ "add_prefix_space": false,
+ "backend": "tokenizers",
+ "bos_token": null,
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "<|im_end|>",
+ "errors": "replace",
+ "extra_special_tokens": [
+ "<|l2r_pred|>",
+ "<|r2l_pred|>",
+ "<|next_1_pred|>",
+ "<|next_2_pred|>"
+ ],
+ "is_local": false,
+ "model_max_length": 131072,
+ "pad_token": "<|im_end|>",
+ "split_special_tokens": false,
+ "tokenizer_class": "Qwen2Tokenizer",
+ "unk_token": null
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2016/trainer_state.json b/runs/l2r50-i2-fulle-lm/checkpoint-2016/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..09b7a92a89f2dfae040dd44f3d79c6234258eeda
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-2016/trainer_state.json
@@ -0,0 +1,14362 @@
+{
+ "best_global_step": null,
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 28.0,
+ "eval_steps": 500,
+ "global_step": 2016,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.013973799126637555,
+ "grad_norm": 0.1739506870508194,
+ "learning_rate": 0.0,
+ "loss": 11.99299430847168,
+ "step": 1
+ },
+ {
+ "epoch": 0.02794759825327511,
+ "grad_norm": 0.17216597497463226,
+ "learning_rate": 5.999999999999999e-06,
+ "loss": 11.988704681396484,
+ "step": 2
+ },
+ {
+ "epoch": 0.04192139737991266,
+ "grad_norm": 0.17459715902805328,
+ "learning_rate": 1.1999999999999999e-05,
+ "loss": 11.945581436157227,
+ "step": 3
+ },
+ {
+ "epoch": 0.05589519650655022,
+ "grad_norm": 0.17577414214611053,
+ "learning_rate": 1.7999999999999997e-05,
+ "loss": 11.859772682189941,
+ "step": 4
+ },
+ {
+ "epoch": 0.06986899563318777,
+ "grad_norm": 0.16163651645183563,
+ "learning_rate": 2.3999999999999997e-05,
+ "loss": 11.744415283203125,
+ "step": 5
+ },
+ {
+ "epoch": 0.08384279475982533,
+ "grad_norm": 0.14409932494163513,
+ "learning_rate": 2.9999999999999997e-05,
+ "loss": 11.618099212646484,
+ "step": 6
+ },
+ {
+ "epoch": 0.09781659388646288,
+ "grad_norm": 0.13079848885536194,
+ "learning_rate": 3.5999999999999994e-05,
+ "loss": 11.505935668945312,
+ "step": 7
+ },
+ {
+ "epoch": 0.11179039301310044,
+ "grad_norm": 0.11964613944292068,
+ "learning_rate": 4.2e-05,
+ "loss": 11.416532516479492,
+ "step": 8
+ },
+ {
+ "epoch": 0.125764192139738,
+ "grad_norm": 0.11345142871141434,
+ "learning_rate": 4.7999999999999994e-05,
+ "loss": 11.33206558227539,
+ "step": 9
+ },
+ {
+ "epoch": 0.13973799126637554,
+ "grad_norm": 0.11033952981233597,
+ "learning_rate": 5.399999999999999e-05,
+ "loss": 11.259052276611328,
+ "step": 10
+ },
+ {
+ "epoch": 0.1537117903930131,
+ "grad_norm": 0.10776489973068237,
+ "learning_rate": 5.9999999999999995e-05,
+ "loss": 11.203384399414062,
+ "step": 11
+ },
+ {
+ "epoch": 0.16768558951965065,
+ "grad_norm": 0.10721946507692337,
+ "learning_rate": 6.599999999999999e-05,
+ "loss": 11.153295516967773,
+ "step": 12
+ },
+ {
+ "epoch": 0.18165938864628822,
+ "grad_norm": 0.10673430562019348,
+ "learning_rate": 7.199999999999999e-05,
+ "loss": 11.110200881958008,
+ "step": 13
+ },
+ {
+ "epoch": 0.19563318777292577,
+ "grad_norm": 0.10630254447460175,
+ "learning_rate": 7.8e-05,
+ "loss": 11.067477226257324,
+ "step": 14
+ },
+ {
+ "epoch": 0.2096069868995633,
+ "grad_norm": 0.10631615668535233,
+ "learning_rate": 8.4e-05,
+ "loss": 11.022941589355469,
+ "step": 15
+ },
+ {
+ "epoch": 0.22358078602620088,
+ "grad_norm": 0.10609443485736847,
+ "learning_rate": 8.999999999999999e-05,
+ "loss": 10.970746040344238,
+ "step": 16
+ },
+ {
+ "epoch": 0.23755458515283842,
+ "grad_norm": 0.10648232698440552,
+ "learning_rate": 9.599999999999999e-05,
+ "loss": 10.91267204284668,
+ "step": 17
+ },
+ {
+ "epoch": 0.251528384279476,
+ "grad_norm": 0.10584589838981628,
+ "learning_rate": 0.000102,
+ "loss": 10.854547500610352,
+ "step": 18
+ },
+ {
+ "epoch": 0.26550218340611353,
+ "grad_norm": 0.10591720044612885,
+ "learning_rate": 0.00010799999999999998,
+ "loss": 10.788917541503906,
+ "step": 19
+ },
+ {
+ "epoch": 0.2794759825327511,
+ "grad_norm": 0.10611672699451447,
+ "learning_rate": 0.00011399999999999999,
+ "loss": 10.717556953430176,
+ "step": 20
+ },
+ {
+ "epoch": 0.2934497816593886,
+ "grad_norm": 0.10455359518527985,
+ "learning_rate": 0.00011999999999999999,
+ "loss": 10.650186538696289,
+ "step": 21
+ },
+ {
+ "epoch": 0.3074235807860262,
+ "grad_norm": 0.1047525629401207,
+ "learning_rate": 0.00012599999999999997,
+ "loss": 10.56808090209961,
+ "step": 22
+ },
+ {
+ "epoch": 0.32139737991266376,
+ "grad_norm": 0.10511605441570282,
+ "learning_rate": 0.00013199999999999998,
+ "loss": 10.484028816223145,
+ "step": 23
+ },
+ {
+ "epoch": 0.3353711790393013,
+ "grad_norm": 0.10467177629470825,
+ "learning_rate": 0.000138,
+ "loss": 10.399947166442871,
+ "step": 24
+ },
+ {
+ "epoch": 0.34934497816593885,
+ "grad_norm": 0.10432136058807373,
+ "learning_rate": 0.00014399999999999998,
+ "loss": 10.312714576721191,
+ "step": 25
+ },
+ {
+ "epoch": 0.36331877729257644,
+ "grad_norm": 0.10527262091636658,
+ "learning_rate": 0.00015,
+ "loss": 10.212261199951172,
+ "step": 26
+ },
+ {
+ "epoch": 0.377292576419214,
+ "grad_norm": 0.103610560297966,
+ "learning_rate": 0.000156,
+ "loss": 10.132034301757812,
+ "step": 27
+ },
+ {
+ "epoch": 0.39126637554585153,
+ "grad_norm": 0.10553059726953506,
+ "learning_rate": 0.000162,
+ "loss": 10.020434379577637,
+ "step": 28
+ },
+ {
+ "epoch": 0.4052401746724891,
+ "grad_norm": 0.1037299782037735,
+ "learning_rate": 0.000168,
+ "loss": 9.934466361999512,
+ "step": 29
+ },
+ {
+ "epoch": 0.4192139737991266,
+ "grad_norm": 0.1042739674448967,
+ "learning_rate": 0.00017399999999999997,
+ "loss": 9.822539329528809,
+ "step": 30
+ },
+ {
+ "epoch": 0.4331877729257642,
+ "grad_norm": 0.10300865024328232,
+ "learning_rate": 0.00017999999999999998,
+ "loss": 9.734994888305664,
+ "step": 31
+ },
+ {
+ "epoch": 0.44716157205240176,
+ "grad_norm": 0.10291654616594315,
+ "learning_rate": 0.000186,
+ "loss": 9.624505996704102,
+ "step": 32
+ },
+ {
+ "epoch": 0.4611353711790393,
+ "grad_norm": 0.10210851579904556,
+ "learning_rate": 0.00019199999999999998,
+ "loss": 9.518205642700195,
+ "step": 33
+ },
+ {
+ "epoch": 0.47510917030567684,
+ "grad_norm": 0.10222529619932175,
+ "learning_rate": 0.000198,
+ "loss": 9.412222862243652,
+ "step": 34
+ },
+ {
+ "epoch": 0.4890829694323144,
+ "grad_norm": 0.09992019087076187,
+ "learning_rate": 0.000204,
+ "loss": 9.326566696166992,
+ "step": 35
+ },
+ {
+ "epoch": 0.503056768558952,
+ "grad_norm": 0.09931084513664246,
+ "learning_rate": 0.00020999999999999998,
+ "loss": 9.210111618041992,
+ "step": 36
+ },
+ {
+ "epoch": 0.5170305676855895,
+ "grad_norm": 0.09928256273269653,
+ "learning_rate": 0.00021599999999999996,
+ "loss": 9.100493431091309,
+ "step": 37
+ },
+ {
+ "epoch": 0.5310043668122271,
+ "grad_norm": 0.09843816608190536,
+ "learning_rate": 0.00022199999999999998,
+ "loss": 8.996543884277344,
+ "step": 38
+ },
+ {
+ "epoch": 0.5449781659388646,
+ "grad_norm": 0.09825374186038971,
+ "learning_rate": 0.00022799999999999999,
+ "loss": 8.888991355895996,
+ "step": 39
+ },
+ {
+ "epoch": 0.5589519650655022,
+ "grad_norm": 0.09602385759353638,
+ "learning_rate": 0.000234,
+ "loss": 8.802907943725586,
+ "step": 40
+ },
+ {
+ "epoch": 0.5729257641921397,
+ "grad_norm": 0.09414064139127731,
+ "learning_rate": 0.00023999999999999998,
+ "loss": 8.715423583984375,
+ "step": 41
+ },
+ {
+ "epoch": 0.5868995633187772,
+ "grad_norm": 0.09273893386125565,
+ "learning_rate": 0.00024599999999999996,
+ "loss": 8.61374282836914,
+ "step": 42
+ },
+ {
+ "epoch": 0.6008733624454149,
+ "grad_norm": 0.09079599380493164,
+ "learning_rate": 0.00025199999999999995,
+ "loss": 8.502904891967773,
+ "step": 43
+ },
+ {
+ "epoch": 0.6148471615720524,
+ "grad_norm": 0.08605825901031494,
+ "learning_rate": 0.000258,
+ "loss": 8.453008651733398,
+ "step": 44
+ },
+ {
+ "epoch": 0.62882096069869,
+ "grad_norm": 0.08338996022939682,
+ "learning_rate": 0.00026399999999999997,
+ "loss": 8.365163803100586,
+ "step": 45
+ },
+ {
+ "epoch": 0.6427947598253275,
+ "grad_norm": 0.08150867372751236,
+ "learning_rate": 0.00027,
+ "loss": 8.285372734069824,
+ "step": 46
+ },
+ {
+ "epoch": 0.6567685589519651,
+ "grad_norm": 0.07794807851314545,
+ "learning_rate": 0.000276,
+ "loss": 8.185811996459961,
+ "step": 47
+ },
+ {
+ "epoch": 0.6707423580786026,
+ "grad_norm": 0.07592720538377762,
+ "learning_rate": 0.00028199999999999997,
+ "loss": 8.098482131958008,
+ "step": 48
+ },
+ {
+ "epoch": 0.6847161572052402,
+ "grad_norm": 0.0709897130727768,
+ "learning_rate": 0.00028799999999999995,
+ "loss": 8.042091369628906,
+ "step": 49
+ },
+ {
+ "epoch": 0.6986899563318777,
+ "grad_norm": 0.06582305580377579,
+ "learning_rate": 0.000294,
+ "loss": 7.972844123840332,
+ "step": 50
+ },
+ {
+ "epoch": 0.7126637554585152,
+ "grad_norm": 0.059452351182699203,
+ "learning_rate": 0.0003,
+ "loss": 7.924579620361328,
+ "step": 51
+ },
+ {
+ "epoch": 0.7266375545851529,
+ "grad_norm": 0.052426718175411224,
+ "learning_rate": 0.00030599999999999996,
+ "loss": 7.876336097717285,
+ "step": 52
+ },
+ {
+ "epoch": 0.7406113537117904,
+ "grad_norm": 0.04768150672316551,
+ "learning_rate": 0.000312,
+ "loss": 7.8102707862854,
+ "step": 53
+ },
+ {
+ "epoch": 0.754585152838428,
+ "grad_norm": 0.03970075398683548,
+ "learning_rate": 0.000318,
+ "loss": 7.791913986206055,
+ "step": 54
+ },
+ {
+ "epoch": 0.7685589519650655,
+ "grad_norm": 0.03559752553701401,
+ "learning_rate": 0.000324,
+ "loss": 7.756960868835449,
+ "step": 55
+ },
+ {
+ "epoch": 0.7825327510917031,
+ "grad_norm": 0.029321353882551193,
+ "learning_rate": 0.00033,
+ "loss": 7.724878787994385,
+ "step": 56
+ },
+ {
+ "epoch": 0.7965065502183406,
+ "grad_norm": 0.023485427722334862,
+ "learning_rate": 0.000336,
+ "loss": 7.72553014755249,
+ "step": 57
+ },
+ {
+ "epoch": 0.8104803493449781,
+ "grad_norm": 0.019319474697113037,
+ "learning_rate": 0.00034199999999999996,
+ "loss": 7.709493160247803,
+ "step": 58
+ },
+ {
+ "epoch": 0.8244541484716157,
+ "grad_norm": 0.016705691814422607,
+ "learning_rate": 0.00034799999999999995,
+ "loss": 7.6873297691345215,
+ "step": 59
+ },
+ {
+ "epoch": 0.8384279475982532,
+ "grad_norm": 0.017667734995484352,
+ "learning_rate": 0.00035399999999999993,
+ "loss": 7.677059173583984,
+ "step": 60
+ },
+ {
+ "epoch": 0.8524017467248908,
+ "grad_norm": 0.01671992614865303,
+ "learning_rate": 0.00035999999999999997,
+ "loss": 7.643512725830078,
+ "step": 61
+ },
+ {
+ "epoch": 0.8663755458515284,
+ "grad_norm": 0.012288172729313374,
+ "learning_rate": 0.00036599999999999995,
+ "loss": 7.658343315124512,
+ "step": 62
+ },
+ {
+ "epoch": 0.880349344978166,
+ "grad_norm": 0.012004776857793331,
+ "learning_rate": 0.000372,
+ "loss": 7.649759292602539,
+ "step": 63
+ },
+ {
+ "epoch": 0.8943231441048035,
+ "grad_norm": 0.012487445957958698,
+ "learning_rate": 0.00037799999999999997,
+ "loss": 7.631503105163574,
+ "step": 64
+ },
+ {
+ "epoch": 0.9082969432314411,
+ "grad_norm": 0.01489241048693657,
+ "learning_rate": 0.00038399999999999996,
+ "loss": 7.647110462188721,
+ "step": 65
+ },
+ {
+ "epoch": 0.9222707423580786,
+ "grad_norm": 0.01428980939090252,
+ "learning_rate": 0.00039,
+ "loss": 7.636983871459961,
+ "step": 66
+ },
+ {
+ "epoch": 0.9362445414847161,
+ "grad_norm": 0.010711144655942917,
+ "learning_rate": 0.000396,
+ "loss": 7.621130466461182,
+ "step": 67
+ },
+ {
+ "epoch": 0.9502183406113537,
+ "grad_norm": 0.010178886353969574,
+ "learning_rate": 0.000402,
+ "loss": 7.593924522399902,
+ "step": 68
+ },
+ {
+ "epoch": 0.9641921397379912,
+ "grad_norm": 0.010604874230921268,
+ "learning_rate": 0.000408,
+ "loss": 7.6015496253967285,
+ "step": 69
+ },
+ {
+ "epoch": 0.9781659388646288,
+ "grad_norm": 0.009751426056027412,
+ "learning_rate": 0.0004139999999999999,
+ "loss": 7.571964263916016,
+ "step": 70
+ },
+ {
+ "epoch": 0.9921397379912664,
+ "grad_norm": 0.009816481731832027,
+ "learning_rate": 0.00041999999999999996,
+ "loss": 7.57344913482666,
+ "step": 71
+ },
+ {
+ "epoch": 1.0,
+ "grad_norm": 0.008665637113153934,
+ "learning_rate": 0.00042599999999999995,
+ "loss": 7.579050064086914,
+ "step": 72
+ },
+ {
+ "epoch": 1.0,
+ "eval_loss": 7.581204891204834,
+ "eval_runtime": 59.0091,
+ "eval_samples_per_second": 41.383,
+ "eval_steps_per_second": 1.305,
+ "step": 72
+ },
+ {
+ "epoch": 1.0139737991266375,
+ "grad_norm": 0.009684201329946518,
+ "learning_rate": 0.00043199999999999993,
+ "loss": 7.568594932556152,
+ "step": 73
+ },
+ {
+ "epoch": 1.027947598253275,
+ "grad_norm": 0.025783710181713104,
+ "learning_rate": 0.00043799999999999997,
+ "loss": 7.578948497772217,
+ "step": 74
+ },
+ {
+ "epoch": 1.0419213973799126,
+ "grad_norm": 0.0327819399535656,
+ "learning_rate": 0.00044399999999999995,
+ "loss": 7.548173904418945,
+ "step": 75
+ },
+ {
+ "epoch": 1.0558951965065502,
+ "grad_norm": 0.021604498848319054,
+ "learning_rate": 0.00045,
+ "loss": 7.552953720092773,
+ "step": 76
+ },
+ {
+ "epoch": 1.0698689956331877,
+ "grad_norm": 0.010230105370283127,
+ "learning_rate": 0.00045599999999999997,
+ "loss": 7.5213212966918945,
+ "step": 77
+ },
+ {
+ "epoch": 1.0838427947598253,
+ "grad_norm": 0.023315589874982834,
+ "learning_rate": 0.00046199999999999995,
+ "loss": 7.501640319824219,
+ "step": 78
+ },
+ {
+ "epoch": 1.0978165938864628,
+ "grad_norm": 0.01109944935888052,
+ "learning_rate": 0.000468,
+ "loss": 7.539519786834717,
+ "step": 79
+ },
+ {
+ "epoch": 1.1117903930131003,
+ "grad_norm": 0.026093758642673492,
+ "learning_rate": 0.000474,
+ "loss": 7.525928974151611,
+ "step": 80
+ },
+ {
+ "epoch": 1.125764192139738,
+ "grad_norm": 0.03274895250797272,
+ "learning_rate": 0.00047999999999999996,
+ "loss": 7.535602569580078,
+ "step": 81
+ },
+ {
+ "epoch": 1.1397379912663754,
+ "grad_norm": 0.01543602254241705,
+ "learning_rate": 0.000486,
+ "loss": 7.513683795928955,
+ "step": 82
+ },
+ {
+ "epoch": 1.1537117903930132,
+ "grad_norm": 0.028377249836921692,
+ "learning_rate": 0.0004919999999999999,
+ "loss": 7.49384880065918,
+ "step": 83
+ },
+ {
+ "epoch": 1.1676855895196507,
+ "grad_norm": 0.04121481627225876,
+ "learning_rate": 0.000498,
+ "loss": 7.488447189331055,
+ "step": 84
+ },
+ {
+ "epoch": 1.1816593886462883,
+ "grad_norm": 0.013282302767038345,
+ "learning_rate": 0.0005039999999999999,
+ "loss": 7.463010787963867,
+ "step": 85
+ },
+ {
+ "epoch": 1.1956331877729258,
+ "grad_norm": 0.03200364485383034,
+ "learning_rate": 0.0005099999999999999,
+ "loss": 7.473515033721924,
+ "step": 86
+ },
+ {
+ "epoch": 1.2096069868995634,
+ "grad_norm": 0.020768029615283012,
+ "learning_rate": 0.000516,
+ "loss": 7.464688777923584,
+ "step": 87
+ },
+ {
+ "epoch": 1.223580786026201,
+ "grad_norm": 0.021715788170695305,
+ "learning_rate": 0.000522,
+ "loss": 7.4835028648376465,
+ "step": 88
+ },
+ {
+ "epoch": 1.2375545851528384,
+ "grad_norm": 0.026255754753947258,
+ "learning_rate": 0.0005279999999999999,
+ "loss": 7.443392753601074,
+ "step": 89
+ },
+ {
+ "epoch": 1.251528384279476,
+ "grad_norm": 0.022325951606035233,
+ "learning_rate": 0.000534,
+ "loss": 7.457373142242432,
+ "step": 90
+ },
+ {
+ "epoch": 1.2655021834061135,
+ "grad_norm": 0.009270387701690197,
+ "learning_rate": 0.00054,
+ "loss": 7.455986976623535,
+ "step": 91
+ },
+ {
+ "epoch": 1.279475982532751,
+ "grad_norm": 0.02369544468820095,
+ "learning_rate": 0.0005459999999999999,
+ "loss": 7.442216873168945,
+ "step": 92
+ },
+ {
+ "epoch": 1.2934497816593886,
+ "grad_norm": 0.008474506437778473,
+ "learning_rate": 0.000552,
+ "loss": 7.4294915199279785,
+ "step": 93
+ },
+ {
+ "epoch": 1.3074235807860262,
+ "grad_norm": 0.017043249681591988,
+ "learning_rate": 0.000558,
+ "loss": 7.406932830810547,
+ "step": 94
+ },
+ {
+ "epoch": 1.3213973799126637,
+ "grad_norm": 0.011813412420451641,
+ "learning_rate": 0.0005639999999999999,
+ "loss": 7.415900230407715,
+ "step": 95
+ },
+ {
+ "epoch": 1.3353711790393012,
+ "grad_norm": 0.01086883433163166,
+ "learning_rate": 0.00057,
+ "loss": 7.452998638153076,
+ "step": 96
+ },
+ {
+ "epoch": 1.3493449781659388,
+ "grad_norm": 0.014700951054692268,
+ "learning_rate": 0.0005759999999999999,
+ "loss": 7.410249710083008,
+ "step": 97
+ },
+ {
+ "epoch": 1.3633187772925766,
+ "grad_norm": 0.014512976631522179,
+ "learning_rate": 0.0005819999999999999,
+ "loss": 7.429132461547852,
+ "step": 98
+ },
+ {
+ "epoch": 1.3772925764192139,
+ "grad_norm": 0.012519920244812965,
+ "learning_rate": 0.000588,
+ "loss": 7.430817604064941,
+ "step": 99
+ },
+ {
+ "epoch": 1.3912663755458516,
+ "grad_norm": 0.007132664322853088,
+ "learning_rate": 0.0005939999999999999,
+ "loss": 7.405234336853027,
+ "step": 100
+ },
+ {
+ "epoch": 1.405240174672489,
+ "grad_norm": 0.019489388912916183,
+ "learning_rate": 0.0006,
+ "loss": 7.427791595458984,
+ "step": 101
+ },
+ {
+ "epoch": 1.4192139737991267,
+ "grad_norm": 0.027252664789557457,
+ "learning_rate": 0.0006,
+ "loss": 7.41761589050293,
+ "step": 102
+ },
+ {
+ "epoch": 1.4331877729257643,
+ "grad_norm": 0.02501637488603592,
+ "learning_rate": 0.0006,
+ "loss": 7.416735649108887,
+ "step": 103
+ },
+ {
+ "epoch": 1.4471615720524018,
+ "grad_norm": 0.00989772193133831,
+ "learning_rate": 0.0006,
+ "loss": 7.3756632804870605,
+ "step": 104
+ },
+ {
+ "epoch": 1.4611353711790394,
+ "grad_norm": 0.01622949168086052,
+ "learning_rate": 0.0006,
+ "loss": 7.400811672210693,
+ "step": 105
+ },
+ {
+ "epoch": 1.475109170305677,
+ "grad_norm": 0.012407763861119747,
+ "learning_rate": 0.0006,
+ "loss": 7.37675666809082,
+ "step": 106
+ },
+ {
+ "epoch": 1.4890829694323144,
+ "grad_norm": 0.010353565216064453,
+ "learning_rate": 0.0006,
+ "loss": 7.4013142585754395,
+ "step": 107
+ },
+ {
+ "epoch": 1.503056768558952,
+ "grad_norm": 0.0120553532615304,
+ "learning_rate": 0.0006,
+ "loss": 7.379873275756836,
+ "step": 108
+ },
+ {
+ "epoch": 1.5170305676855895,
+ "grad_norm": 0.01610225811600685,
+ "learning_rate": 0.0006,
+ "loss": 7.371631145477295,
+ "step": 109
+ },
+ {
+ "epoch": 1.531004366812227,
+ "grad_norm": 0.01159879844635725,
+ "learning_rate": 0.0006,
+ "loss": 7.389007568359375,
+ "step": 110
+ },
+ {
+ "epoch": 1.5449781659388646,
+ "grad_norm": 0.008013816550374031,
+ "learning_rate": 0.0006,
+ "loss": 7.395920753479004,
+ "step": 111
+ },
+ {
+ "epoch": 1.5589519650655022,
+ "grad_norm": 0.016334809362888336,
+ "learning_rate": 0.0006,
+ "loss": 7.393121719360352,
+ "step": 112
+ },
+ {
+ "epoch": 1.5729257641921397,
+ "grad_norm": 0.018821721896529198,
+ "learning_rate": 0.0006,
+ "loss": 7.367543697357178,
+ "step": 113
+ },
+ {
+ "epoch": 1.5868995633187772,
+ "grad_norm": 0.014088994823396206,
+ "learning_rate": 0.0006,
+ "loss": 7.385537624359131,
+ "step": 114
+ },
+ {
+ "epoch": 1.600873362445415,
+ "grad_norm": 0.005864050704985857,
+ "learning_rate": 0.0006,
+ "loss": 7.365346908569336,
+ "step": 115
+ },
+ {
+ "epoch": 1.6148471615720523,
+ "grad_norm": 0.011745430529117584,
+ "learning_rate": 0.0006,
+ "loss": 7.390227317810059,
+ "step": 116
+ },
+ {
+ "epoch": 1.62882096069869,
+ "grad_norm": 0.017205573618412018,
+ "learning_rate": 0.0006,
+ "loss": 7.3558549880981445,
+ "step": 117
+ },
+ {
+ "epoch": 1.6427947598253274,
+ "grad_norm": 0.021256346255540848,
+ "learning_rate": 0.0006,
+ "loss": 7.377346992492676,
+ "step": 118
+ },
+ {
+ "epoch": 1.6567685589519652,
+ "grad_norm": 0.019335204735398293,
+ "learning_rate": 0.0006,
+ "loss": 7.346194267272949,
+ "step": 119
+ },
+ {
+ "epoch": 1.6707423580786025,
+ "grad_norm": 0.008134003728628159,
+ "learning_rate": 0.0006,
+ "loss": 7.3063578605651855,
+ "step": 120
+ },
+ {
+ "epoch": 1.6847161572052403,
+ "grad_norm": 0.017613831907510757,
+ "learning_rate": 0.0006,
+ "loss": 7.338171005249023,
+ "step": 121
+ },
+ {
+ "epoch": 1.6986899563318776,
+ "grad_norm": 0.00890734139829874,
+ "learning_rate": 0.0006,
+ "loss": 7.355384826660156,
+ "step": 122
+ },
+ {
+ "epoch": 1.7126637554585153,
+ "grad_norm": 0.013448155485093594,
+ "learning_rate": 0.0006,
+ "loss": 7.344296455383301,
+ "step": 123
+ },
+ {
+ "epoch": 1.726637554585153,
+ "grad_norm": 0.009241633117198944,
+ "learning_rate": 0.0006,
+ "loss": 7.3454060554504395,
+ "step": 124
+ },
+ {
+ "epoch": 1.7406113537117904,
+ "grad_norm": 0.018561311066150665,
+ "learning_rate": 0.0006,
+ "loss": 7.32348108291626,
+ "step": 125
+ },
+ {
+ "epoch": 1.754585152838428,
+ "grad_norm": 0.025667704641819,
+ "learning_rate": 0.0006,
+ "loss": 7.356142520904541,
+ "step": 126
+ },
+ {
+ "epoch": 1.7685589519650655,
+ "grad_norm": 0.036232151091098785,
+ "learning_rate": 0.0006,
+ "loss": 7.363539695739746,
+ "step": 127
+ },
+ {
+ "epoch": 1.782532751091703,
+ "grad_norm": 0.059483714401721954,
+ "learning_rate": 0.0006,
+ "loss": 7.402331352233887,
+ "step": 128
+ },
+ {
+ "epoch": 1.7965065502183406,
+ "grad_norm": 0.05365009233355522,
+ "learning_rate": 0.0006,
+ "loss": 7.40457820892334,
+ "step": 129
+ },
+ {
+ "epoch": 1.8104803493449781,
+ "grad_norm": 0.02738415263593197,
+ "learning_rate": 0.0006,
+ "loss": 7.3756208419799805,
+ "step": 130
+ },
+ {
+ "epoch": 1.8244541484716157,
+ "grad_norm": 0.035017091780900955,
+ "learning_rate": 0.0006,
+ "loss": 7.379263877868652,
+ "step": 131
+ },
+ {
+ "epoch": 1.8384279475982532,
+ "grad_norm": 0.02819863148033619,
+ "learning_rate": 0.0006,
+ "loss": 7.327673435211182,
+ "step": 132
+ },
+ {
+ "epoch": 1.8524017467248908,
+ "grad_norm": 0.017754238098859787,
+ "learning_rate": 0.0006,
+ "loss": 7.333420276641846,
+ "step": 133
+ },
+ {
+ "epoch": 1.8663755458515285,
+ "grad_norm": 0.024685733020305634,
+ "learning_rate": 0.0006,
+ "loss": 7.34493350982666,
+ "step": 134
+ },
+ {
+ "epoch": 1.8803493449781659,
+ "grad_norm": 0.019426532089710236,
+ "learning_rate": 0.0006,
+ "loss": 7.328448295593262,
+ "step": 135
+ },
+ {
+ "epoch": 1.8943231441048036,
+ "grad_norm": 0.02106391079723835,
+ "learning_rate": 0.0006,
+ "loss": 7.327917098999023,
+ "step": 136
+ },
+ {
+ "epoch": 1.908296943231441,
+ "grad_norm": 0.02056879736483097,
+ "learning_rate": 0.0006,
+ "loss": 7.3311991691589355,
+ "step": 137
+ },
+ {
+ "epoch": 1.9222707423580787,
+ "grad_norm": 0.016865141689777374,
+ "learning_rate": 0.0006,
+ "loss": 7.312766075134277,
+ "step": 138
+ },
+ {
+ "epoch": 1.936244541484716,
+ "grad_norm": 0.017549855634570122,
+ "learning_rate": 0.0006,
+ "loss": 7.31191349029541,
+ "step": 139
+ },
+ {
+ "epoch": 1.9502183406113538,
+ "grad_norm": 0.015930157154798508,
+ "learning_rate": 0.0006,
+ "loss": 7.330901622772217,
+ "step": 140
+ },
+ {
+ "epoch": 1.9641921397379911,
+ "grad_norm": 0.013283558189868927,
+ "learning_rate": 0.0006,
+ "loss": 7.298318386077881,
+ "step": 141
+ },
+ {
+ "epoch": 1.9781659388646289,
+ "grad_norm": 0.012023469433188438,
+ "learning_rate": 0.0006,
+ "loss": 7.282337188720703,
+ "step": 142
+ },
+ {
+ "epoch": 1.9921397379912664,
+ "grad_norm": 0.01459511648863554,
+ "learning_rate": 0.0006,
+ "loss": 7.282924175262451,
+ "step": 143
+ },
+ {
+ "epoch": 2.0,
+ "grad_norm": 0.017474235966801643,
+ "learning_rate": 0.0006,
+ "loss": 7.254333972930908,
+ "step": 144
+ },
+ {
+ "epoch": 2.0,
+ "eval_loss": 7.318708896636963,
+ "eval_runtime": 59.6555,
+ "eval_samples_per_second": 40.935,
+ "eval_steps_per_second": 1.291,
+ "step": 144
+ },
+ {
+ "epoch": 2.0139737991266378,
+ "grad_norm": 0.019757507368922234,
+ "learning_rate": 0.0006,
+ "loss": 7.30288553237915,
+ "step": 145
+ },
+ {
+ "epoch": 2.027947598253275,
+ "grad_norm": 0.014571248553693295,
+ "learning_rate": 0.0006,
+ "loss": 7.274681568145752,
+ "step": 146
+ },
+ {
+ "epoch": 2.041921397379913,
+ "grad_norm": 0.015744198113679886,
+ "learning_rate": 0.0006,
+ "loss": 7.278632640838623,
+ "step": 147
+ },
+ {
+ "epoch": 2.05589519650655,
+ "grad_norm": 0.028048858046531677,
+ "learning_rate": 0.0006,
+ "loss": 7.26558780670166,
+ "step": 148
+ },
+ {
+ "epoch": 2.069868995633188,
+ "grad_norm": 0.06547307223081589,
+ "learning_rate": 0.0006,
+ "loss": 7.32567834854126,
+ "step": 149
+ },
+ {
+ "epoch": 2.0838427947598253,
+ "grad_norm": 0.05969763547182083,
+ "learning_rate": 0.0006,
+ "loss": 7.395836353302002,
+ "step": 150
+ },
+ {
+ "epoch": 2.097816593886463,
+ "grad_norm": 0.032574281096458435,
+ "learning_rate": 0.0006,
+ "loss": 7.318150997161865,
+ "step": 151
+ },
+ {
+ "epoch": 2.1117903930131003,
+ "grad_norm": 0.029452962800860405,
+ "learning_rate": 0.0006,
+ "loss": 7.308066368103027,
+ "step": 152
+ },
+ {
+ "epoch": 2.125764192139738,
+ "grad_norm": 0.03161991387605667,
+ "learning_rate": 0.0006,
+ "loss": 7.306290626525879,
+ "step": 153
+ },
+ {
+ "epoch": 2.1397379912663754,
+ "grad_norm": 0.027096295729279518,
+ "learning_rate": 0.0006,
+ "loss": 7.297796249389648,
+ "step": 154
+ },
+ {
+ "epoch": 2.153711790393013,
+ "grad_norm": 0.028657056391239166,
+ "learning_rate": 0.0006,
+ "loss": 7.317424774169922,
+ "step": 155
+ },
+ {
+ "epoch": 2.1676855895196505,
+ "grad_norm": 0.02046995982527733,
+ "learning_rate": 0.0006,
+ "loss": 7.297986030578613,
+ "step": 156
+ },
+ {
+ "epoch": 2.1816593886462883,
+ "grad_norm": 0.02220369316637516,
+ "learning_rate": 0.0006,
+ "loss": 7.268237590789795,
+ "step": 157
+ },
+ {
+ "epoch": 2.1956331877729256,
+ "grad_norm": 0.018381357192993164,
+ "learning_rate": 0.0006,
+ "loss": 7.259415626525879,
+ "step": 158
+ },
+ {
+ "epoch": 2.2096069868995634,
+ "grad_norm": 0.0204413291066885,
+ "learning_rate": 0.0006,
+ "loss": 7.258173942565918,
+ "step": 159
+ },
+ {
+ "epoch": 2.2235807860262007,
+ "grad_norm": 0.015874288976192474,
+ "learning_rate": 0.0006,
+ "loss": 7.281834602355957,
+ "step": 160
+ },
+ {
+ "epoch": 2.2375545851528384,
+ "grad_norm": 0.01634068600833416,
+ "learning_rate": 0.0006,
+ "loss": 7.2469868659973145,
+ "step": 161
+ },
+ {
+ "epoch": 2.251528384279476,
+ "grad_norm": 0.01728747971355915,
+ "learning_rate": 0.0006,
+ "loss": 7.280995845794678,
+ "step": 162
+ },
+ {
+ "epoch": 2.2655021834061135,
+ "grad_norm": 0.016341188922524452,
+ "learning_rate": 0.0006,
+ "loss": 7.273357391357422,
+ "step": 163
+ },
+ {
+ "epoch": 2.279475982532751,
+ "grad_norm": 0.016125807538628578,
+ "learning_rate": 0.0006,
+ "loss": 7.273510932922363,
+ "step": 164
+ },
+ {
+ "epoch": 2.2934497816593886,
+ "grad_norm": 0.01629287749528885,
+ "learning_rate": 0.0006,
+ "loss": 7.267029762268066,
+ "step": 165
+ },
+ {
+ "epoch": 2.3074235807860264,
+ "grad_norm": 0.015296131372451782,
+ "learning_rate": 0.0006,
+ "loss": 7.211989402770996,
+ "step": 166
+ },
+ {
+ "epoch": 2.3213973799126637,
+ "grad_norm": 0.01709570363163948,
+ "learning_rate": 0.0006,
+ "loss": 7.242851734161377,
+ "step": 167
+ },
+ {
+ "epoch": 2.3353711790393015,
+ "grad_norm": 0.012552388943731785,
+ "learning_rate": 0.0006,
+ "loss": 7.241696834564209,
+ "step": 168
+ },
+ {
+ "epoch": 2.349344978165939,
+ "grad_norm": 0.014571291394531727,
+ "learning_rate": 0.0006,
+ "loss": 7.270681381225586,
+ "step": 169
+ },
+ {
+ "epoch": 2.3633187772925766,
+ "grad_norm": 0.013389437459409237,
+ "learning_rate": 0.0006,
+ "loss": 7.191600799560547,
+ "step": 170
+ },
+ {
+ "epoch": 2.377292576419214,
+ "grad_norm": 0.011895314790308475,
+ "learning_rate": 0.0006,
+ "loss": 7.213610649108887,
+ "step": 171
+ },
+ {
+ "epoch": 2.3912663755458516,
+ "grad_norm": 0.01345642190426588,
+ "learning_rate": 0.0006,
+ "loss": 7.193787574768066,
+ "step": 172
+ },
+ {
+ "epoch": 2.405240174672489,
+ "grad_norm": 0.0091794328764081,
+ "learning_rate": 0.0006,
+ "loss": 7.233364105224609,
+ "step": 173
+ },
+ {
+ "epoch": 2.4192139737991267,
+ "grad_norm": 0.013767086900770664,
+ "learning_rate": 0.0006,
+ "loss": 7.217272758483887,
+ "step": 174
+ },
+ {
+ "epoch": 2.433187772925764,
+ "grad_norm": 0.015134960412979126,
+ "learning_rate": 0.0006,
+ "loss": 7.192791938781738,
+ "step": 175
+ },
+ {
+ "epoch": 2.447161572052402,
+ "grad_norm": 0.01790648326277733,
+ "learning_rate": 0.0006,
+ "loss": 7.229467391967773,
+ "step": 176
+ },
+ {
+ "epoch": 2.461135371179039,
+ "grad_norm": 0.015362377278506756,
+ "learning_rate": 0.0006,
+ "loss": 7.180346488952637,
+ "step": 177
+ },
+ {
+ "epoch": 2.475109170305677,
+ "grad_norm": 0.010505498386919498,
+ "learning_rate": 0.0006,
+ "loss": 7.188833236694336,
+ "step": 178
+ },
+ {
+ "epoch": 2.489082969432314,
+ "grad_norm": 0.009129747748374939,
+ "learning_rate": 0.0006,
+ "loss": 7.203153133392334,
+ "step": 179
+ },
+ {
+ "epoch": 2.503056768558952,
+ "grad_norm": 0.01147883478552103,
+ "learning_rate": 0.0006,
+ "loss": 7.163824558258057,
+ "step": 180
+ },
+ {
+ "epoch": 2.5170305676855893,
+ "grad_norm": 0.011338942684233189,
+ "learning_rate": 0.0006,
+ "loss": 7.16287899017334,
+ "step": 181
+ },
+ {
+ "epoch": 2.531004366812227,
+ "grad_norm": 0.013277675025165081,
+ "learning_rate": 0.0006,
+ "loss": 7.155905723571777,
+ "step": 182
+ },
+ {
+ "epoch": 2.544978165938865,
+ "grad_norm": 0.015269347466528416,
+ "learning_rate": 0.0006,
+ "loss": 7.189371109008789,
+ "step": 183
+ },
+ {
+ "epoch": 2.558951965065502,
+ "grad_norm": 0.017905596643686295,
+ "learning_rate": 0.0006,
+ "loss": 7.177547931671143,
+ "step": 184
+ },
+ {
+ "epoch": 2.5729257641921395,
+ "grad_norm": 0.024431610479950905,
+ "learning_rate": 0.0006,
+ "loss": 7.192169189453125,
+ "step": 185
+ },
+ {
+ "epoch": 2.5868995633187772,
+ "grad_norm": 0.03656961768865585,
+ "learning_rate": 0.0006,
+ "loss": 7.124387741088867,
+ "step": 186
+ },
+ {
+ "epoch": 2.600873362445415,
+ "grad_norm": 0.04337048903107643,
+ "learning_rate": 0.0006,
+ "loss": 7.140231609344482,
+ "step": 187
+ },
+ {
+ "epoch": 2.6148471615720523,
+ "grad_norm": 0.04047228768467903,
+ "learning_rate": 0.0006,
+ "loss": 7.171927452087402,
+ "step": 188
+ },
+ {
+ "epoch": 2.62882096069869,
+ "grad_norm": 0.07825770974159241,
+ "learning_rate": 0.0006,
+ "loss": 7.207631587982178,
+ "step": 189
+ },
+ {
+ "epoch": 2.6427947598253274,
+ "grad_norm": 0.07118063420057297,
+ "learning_rate": 0.0006,
+ "loss": 7.209488391876221,
+ "step": 190
+ },
+ {
+ "epoch": 2.656768558951965,
+ "grad_norm": 0.05153293535113335,
+ "learning_rate": 0.0006,
+ "loss": 7.14793062210083,
+ "step": 191
+ },
+ {
+ "epoch": 2.6707423580786025,
+ "grad_norm": 0.03318839520215988,
+ "learning_rate": 0.0006,
+ "loss": 7.163010597229004,
+ "step": 192
+ },
+ {
+ "epoch": 2.6847161572052403,
+ "grad_norm": 0.029333539307117462,
+ "learning_rate": 0.0006,
+ "loss": 7.159955978393555,
+ "step": 193
+ },
+ {
+ "epoch": 2.6986899563318776,
+ "grad_norm": 0.029405880719423294,
+ "learning_rate": 0.0006,
+ "loss": 7.136178493499756,
+ "step": 194
+ },
+ {
+ "epoch": 2.7126637554585153,
+ "grad_norm": 0.037815988063812256,
+ "learning_rate": 0.0006,
+ "loss": 7.183186054229736,
+ "step": 195
+ },
+ {
+ "epoch": 2.726637554585153,
+ "grad_norm": 0.020307250320911407,
+ "learning_rate": 0.0006,
+ "loss": 7.133164882659912,
+ "step": 196
+ },
+ {
+ "epoch": 2.7406113537117904,
+ "grad_norm": 0.026048069819808006,
+ "learning_rate": 0.0006,
+ "loss": 7.124977111816406,
+ "step": 197
+ },
+ {
+ "epoch": 2.7545851528384278,
+ "grad_norm": 0.027316724881529808,
+ "learning_rate": 0.0006,
+ "loss": 7.07265567779541,
+ "step": 198
+ },
+ {
+ "epoch": 2.7685589519650655,
+ "grad_norm": 0.021629009395837784,
+ "learning_rate": 0.0006,
+ "loss": 7.150020599365234,
+ "step": 199
+ },
+ {
+ "epoch": 2.7825327510917033,
+ "grad_norm": 0.01605929434299469,
+ "learning_rate": 0.0006,
+ "loss": 7.10919189453125,
+ "step": 200
+ },
+ {
+ "epoch": 2.7965065502183406,
+ "grad_norm": 0.020800089463591576,
+ "learning_rate": 0.0006,
+ "loss": 7.132045269012451,
+ "step": 201
+ },
+ {
+ "epoch": 2.810480349344978,
+ "grad_norm": 0.019225774332880974,
+ "learning_rate": 0.0006,
+ "loss": 7.128364562988281,
+ "step": 202
+ },
+ {
+ "epoch": 2.8244541484716157,
+ "grad_norm": 0.01561372634023428,
+ "learning_rate": 0.0006,
+ "loss": 7.083669662475586,
+ "step": 203
+ },
+ {
+ "epoch": 2.8384279475982535,
+ "grad_norm": 0.021969519555568695,
+ "learning_rate": 0.0006,
+ "loss": 7.050841331481934,
+ "step": 204
+ },
+ {
+ "epoch": 2.8524017467248908,
+ "grad_norm": 0.021453695371747017,
+ "learning_rate": 0.0006,
+ "loss": 7.097085475921631,
+ "step": 205
+ },
+ {
+ "epoch": 2.8663755458515285,
+ "grad_norm": 0.022836022078990936,
+ "learning_rate": 0.0006,
+ "loss": 7.10405969619751,
+ "step": 206
+ },
+ {
+ "epoch": 2.880349344978166,
+ "grad_norm": 0.02430851384997368,
+ "learning_rate": 0.0006,
+ "loss": 7.045181751251221,
+ "step": 207
+ },
+ {
+ "epoch": 2.8943231441048036,
+ "grad_norm": 0.020166993141174316,
+ "learning_rate": 0.0006,
+ "loss": 7.068183898925781,
+ "step": 208
+ },
+ {
+ "epoch": 2.908296943231441,
+ "grad_norm": 0.0118505684658885,
+ "learning_rate": 0.0006,
+ "loss": 7.07132625579834,
+ "step": 209
+ },
+ {
+ "epoch": 2.9222707423580787,
+ "grad_norm": 0.023772427812218666,
+ "learning_rate": 0.0006,
+ "loss": 7.055238723754883,
+ "step": 210
+ },
+ {
+ "epoch": 2.936244541484716,
+ "grad_norm": 0.02350712940096855,
+ "learning_rate": 0.0006,
+ "loss": 7.055898666381836,
+ "step": 211
+ },
+ {
+ "epoch": 2.950218340611354,
+ "grad_norm": 0.017244907096028328,
+ "learning_rate": 0.0006,
+ "loss": 7.0099711418151855,
+ "step": 212
+ },
+ {
+ "epoch": 2.964192139737991,
+ "grad_norm": 0.021565575152635574,
+ "learning_rate": 0.0006,
+ "loss": 6.991751670837402,
+ "step": 213
+ },
+ {
+ "epoch": 2.978165938864629,
+ "grad_norm": 0.03241860866546631,
+ "learning_rate": 0.0006,
+ "loss": 7.005980491638184,
+ "step": 214
+ },
+ {
+ "epoch": 2.992139737991266,
+ "grad_norm": 0.04613726586103439,
+ "learning_rate": 0.0006,
+ "loss": 7.027859687805176,
+ "step": 215
+ },
+ {
+ "epoch": 3.0,
+ "grad_norm": 0.06404894590377808,
+ "learning_rate": 0.0006,
+ "loss": 7.123725891113281,
+ "step": 216
+ },
+ {
+ "epoch": 3.0,
+ "eval_loss": 7.070647716522217,
+ "eval_runtime": 58.5801,
+ "eval_samples_per_second": 41.687,
+ "eval_steps_per_second": 1.314,
+ "step": 216
+ },
+ {
+ "epoch": 3.0139737991266378,
+ "grad_norm": 0.05542432889342308,
+ "learning_rate": 0.0006,
+ "loss": 7.078299522399902,
+ "step": 217
+ },
+ {
+ "epoch": 3.027947598253275,
+ "grad_norm": 0.026484526693820953,
+ "learning_rate": 0.0006,
+ "loss": 6.961159706115723,
+ "step": 218
+ },
+ {
+ "epoch": 3.041921397379913,
+ "grad_norm": 0.028276391327381134,
+ "learning_rate": 0.0006,
+ "loss": 7.019961833953857,
+ "step": 219
+ },
+ {
+ "epoch": 3.05589519650655,
+ "grad_norm": 0.024486595764756203,
+ "learning_rate": 0.0006,
+ "loss": 7.039677619934082,
+ "step": 220
+ },
+ {
+ "epoch": 3.069868995633188,
+ "grad_norm": 0.017620893195271492,
+ "learning_rate": 0.0006,
+ "loss": 6.9270782470703125,
+ "step": 221
+ },
+ {
+ "epoch": 3.0838427947598253,
+ "grad_norm": 0.02547302283346653,
+ "learning_rate": 0.0006,
+ "loss": 6.924233436584473,
+ "step": 222
+ },
+ {
+ "epoch": 3.097816593886463,
+ "grad_norm": 0.02539009042084217,
+ "learning_rate": 0.0006,
+ "loss": 6.973979949951172,
+ "step": 223
+ },
+ {
+ "epoch": 3.1117903930131003,
+ "grad_norm": 0.03256227448582649,
+ "learning_rate": 0.0006,
+ "loss": 6.933725357055664,
+ "step": 224
+ },
+ {
+ "epoch": 3.125764192139738,
+ "grad_norm": 0.04651800915598869,
+ "learning_rate": 0.0006,
+ "loss": 6.951044082641602,
+ "step": 225
+ },
+ {
+ "epoch": 3.1397379912663754,
+ "grad_norm": 0.04822975769639015,
+ "learning_rate": 0.0006,
+ "loss": 6.9614410400390625,
+ "step": 226
+ },
+ {
+ "epoch": 3.153711790393013,
+ "grad_norm": 0.031165020540356636,
+ "learning_rate": 0.0006,
+ "loss": 6.91193151473999,
+ "step": 227
+ },
+ {
+ "epoch": 3.1676855895196505,
+ "grad_norm": 0.029754292219877243,
+ "learning_rate": 0.0006,
+ "loss": 6.91802978515625,
+ "step": 228
+ },
+ {
+ "epoch": 3.1816593886462883,
+ "grad_norm": 0.02193152718245983,
+ "learning_rate": 0.0006,
+ "loss": 6.928090572357178,
+ "step": 229
+ },
+ {
+ "epoch": 3.1956331877729256,
+ "grad_norm": 0.02428470179438591,
+ "learning_rate": 0.0006,
+ "loss": 6.90396785736084,
+ "step": 230
+ },
+ {
+ "epoch": 3.2096069868995634,
+ "grad_norm": 0.023374345153570175,
+ "learning_rate": 0.0006,
+ "loss": 6.885858535766602,
+ "step": 231
+ },
+ {
+ "epoch": 3.2235807860262007,
+ "grad_norm": 0.023405000567436218,
+ "learning_rate": 0.0006,
+ "loss": 6.8855767250061035,
+ "step": 232
+ },
+ {
+ "epoch": 3.2375545851528384,
+ "grad_norm": 0.017277134582400322,
+ "learning_rate": 0.0006,
+ "loss": 6.901449203491211,
+ "step": 233
+ },
+ {
+ "epoch": 3.251528384279476,
+ "grad_norm": 0.019496750086545944,
+ "learning_rate": 0.0006,
+ "loss": 6.885687828063965,
+ "step": 234
+ },
+ {
+ "epoch": 3.2655021834061135,
+ "grad_norm": 0.016893276944756508,
+ "learning_rate": 0.0006,
+ "loss": 6.872369766235352,
+ "step": 235
+ },
+ {
+ "epoch": 3.279475982532751,
+ "grad_norm": 0.02058715932071209,
+ "learning_rate": 0.0006,
+ "loss": 6.823280334472656,
+ "step": 236
+ },
+ {
+ "epoch": 3.2934497816593886,
+ "grad_norm": 0.015530755743384361,
+ "learning_rate": 0.0006,
+ "loss": 6.860499858856201,
+ "step": 237
+ },
+ {
+ "epoch": 3.3074235807860264,
+ "grad_norm": 0.01934926211833954,
+ "learning_rate": 0.0006,
+ "loss": 6.817448616027832,
+ "step": 238
+ },
+ {
+ "epoch": 3.3213973799126637,
+ "grad_norm": 0.017413552850484848,
+ "learning_rate": 0.0006,
+ "loss": 6.881955146789551,
+ "step": 239
+ },
+ {
+ "epoch": 3.3353711790393015,
+ "grad_norm": 0.026478875428438187,
+ "learning_rate": 0.0006,
+ "loss": 6.827404499053955,
+ "step": 240
+ },
+ {
+ "epoch": 3.349344978165939,
+ "grad_norm": 0.032611701637506485,
+ "learning_rate": 0.0006,
+ "loss": 6.78285551071167,
+ "step": 241
+ },
+ {
+ "epoch": 3.3633187772925766,
+ "grad_norm": 0.041900306940078735,
+ "learning_rate": 0.0006,
+ "loss": 6.783053874969482,
+ "step": 242
+ },
+ {
+ "epoch": 3.377292576419214,
+ "grad_norm": 0.04149497672915459,
+ "learning_rate": 0.0006,
+ "loss": 6.745220184326172,
+ "step": 243
+ },
+ {
+ "epoch": 3.3912663755458516,
+ "grad_norm": 0.023949826136231422,
+ "learning_rate": 0.0006,
+ "loss": 6.841533184051514,
+ "step": 244
+ },
+ {
+ "epoch": 3.405240174672489,
+ "grad_norm": 0.028601842001080513,
+ "learning_rate": 0.0006,
+ "loss": 6.852852821350098,
+ "step": 245
+ },
+ {
+ "epoch": 3.4192139737991267,
+ "grad_norm": 0.027365155518054962,
+ "learning_rate": 0.0006,
+ "loss": 6.74376106262207,
+ "step": 246
+ },
+ {
+ "epoch": 3.433187772925764,
+ "grad_norm": 0.02856568619608879,
+ "learning_rate": 0.0006,
+ "loss": 6.74906063079834,
+ "step": 247
+ },
+ {
+ "epoch": 3.447161572052402,
+ "grad_norm": 0.019172416999936104,
+ "learning_rate": 0.0006,
+ "loss": 6.809206008911133,
+ "step": 248
+ },
+ {
+ "epoch": 3.461135371179039,
+ "grad_norm": 0.02161746844649315,
+ "learning_rate": 0.0006,
+ "loss": 6.789888381958008,
+ "step": 249
+ },
+ {
+ "epoch": 3.475109170305677,
+ "grad_norm": 0.020273666828870773,
+ "learning_rate": 0.0006,
+ "loss": 6.73512077331543,
+ "step": 250
+ },
+ {
+ "epoch": 3.489082969432314,
+ "grad_norm": 0.022329283878207207,
+ "learning_rate": 0.0006,
+ "loss": 6.826148509979248,
+ "step": 251
+ },
+ {
+ "epoch": 3.503056768558952,
+ "grad_norm": 0.02284284122288227,
+ "learning_rate": 0.0006,
+ "loss": 6.748521327972412,
+ "step": 252
+ },
+ {
+ "epoch": 3.5170305676855893,
+ "grad_norm": 0.026368247345089912,
+ "learning_rate": 0.0006,
+ "loss": 6.645879745483398,
+ "step": 253
+ },
+ {
+ "epoch": 3.531004366812227,
+ "grad_norm": 0.03627244383096695,
+ "learning_rate": 0.0006,
+ "loss": 6.666064739227295,
+ "step": 254
+ },
+ {
+ "epoch": 3.544978165938865,
+ "grad_norm": 0.04140935093164444,
+ "learning_rate": 0.0006,
+ "loss": 6.712275981903076,
+ "step": 255
+ },
+ {
+ "epoch": 3.558951965065502,
+ "grad_norm": 0.0390462800860405,
+ "learning_rate": 0.0006,
+ "loss": 6.665900230407715,
+ "step": 256
+ },
+ {
+ "epoch": 3.5729257641921395,
+ "grad_norm": 0.05178583785891533,
+ "learning_rate": 0.0006,
+ "loss": 6.753298759460449,
+ "step": 257
+ },
+ {
+ "epoch": 3.5868995633187772,
+ "grad_norm": 0.054328061640262604,
+ "learning_rate": 0.0006,
+ "loss": 6.732633113861084,
+ "step": 258
+ },
+ {
+ "epoch": 3.600873362445415,
+ "grad_norm": 0.04590460658073425,
+ "learning_rate": 0.0006,
+ "loss": 6.65250301361084,
+ "step": 259
+ },
+ {
+ "epoch": 3.6148471615720523,
+ "grad_norm": 0.044331666082143784,
+ "learning_rate": 0.0006,
+ "loss": 6.68654203414917,
+ "step": 260
+ },
+ {
+ "epoch": 3.62882096069869,
+ "grad_norm": 0.05308730527758598,
+ "learning_rate": 0.0006,
+ "loss": 6.642482757568359,
+ "step": 261
+ },
+ {
+ "epoch": 3.6427947598253274,
+ "grad_norm": 0.03999016433954239,
+ "learning_rate": 0.0006,
+ "loss": 6.625227451324463,
+ "step": 262
+ },
+ {
+ "epoch": 3.656768558951965,
+ "grad_norm": 0.04088086634874344,
+ "learning_rate": 0.0006,
+ "loss": 6.67623233795166,
+ "step": 263
+ },
+ {
+ "epoch": 3.6707423580786025,
+ "grad_norm": 0.03615636005997658,
+ "learning_rate": 0.0006,
+ "loss": 6.6869001388549805,
+ "step": 264
+ },
+ {
+ "epoch": 3.6847161572052403,
+ "grad_norm": 0.02817360684275627,
+ "learning_rate": 0.0006,
+ "loss": 6.674601078033447,
+ "step": 265
+ },
+ {
+ "epoch": 3.6986899563318776,
+ "grad_norm": 0.032586123794317245,
+ "learning_rate": 0.0006,
+ "loss": 6.612994194030762,
+ "step": 266
+ },
+ {
+ "epoch": 3.7126637554585153,
+ "grad_norm": 0.031117867678403854,
+ "learning_rate": 0.0006,
+ "loss": 6.6244354248046875,
+ "step": 267
+ },
+ {
+ "epoch": 3.726637554585153,
+ "grad_norm": 0.028330687433481216,
+ "learning_rate": 0.0006,
+ "loss": 6.669557094573975,
+ "step": 268
+ },
+ {
+ "epoch": 3.7406113537117904,
+ "grad_norm": 0.021489202976226807,
+ "learning_rate": 0.0006,
+ "loss": 6.632035732269287,
+ "step": 269
+ },
+ {
+ "epoch": 3.7545851528384278,
+ "grad_norm": 0.024269424378871918,
+ "learning_rate": 0.0006,
+ "loss": 6.660867691040039,
+ "step": 270
+ },
+ {
+ "epoch": 3.7685589519650655,
+ "grad_norm": 0.02237873338162899,
+ "learning_rate": 0.0006,
+ "loss": 6.602587699890137,
+ "step": 271
+ },
+ {
+ "epoch": 3.7825327510917033,
+ "grad_norm": 0.026503929868340492,
+ "learning_rate": 0.0006,
+ "loss": 6.599200248718262,
+ "step": 272
+ },
+ {
+ "epoch": 3.7965065502183406,
+ "grad_norm": 0.03687124326825142,
+ "learning_rate": 0.0006,
+ "loss": 6.567296504974365,
+ "step": 273
+ },
+ {
+ "epoch": 3.810480349344978,
+ "grad_norm": 0.04708952456712723,
+ "learning_rate": 0.0006,
+ "loss": 6.621265411376953,
+ "step": 274
+ },
+ {
+ "epoch": 3.8244541484716157,
+ "grad_norm": 0.041846614331007004,
+ "learning_rate": 0.0006,
+ "loss": 6.550536155700684,
+ "step": 275
+ },
+ {
+ "epoch": 3.8384279475982535,
+ "grad_norm": 0.031249675899744034,
+ "learning_rate": 0.0006,
+ "loss": 6.54278564453125,
+ "step": 276
+ },
+ {
+ "epoch": 3.8524017467248908,
+ "grad_norm": 0.023082256317138672,
+ "learning_rate": 0.0006,
+ "loss": 6.544787406921387,
+ "step": 277
+ },
+ {
+ "epoch": 3.8663755458515285,
+ "grad_norm": 0.020995570346713066,
+ "learning_rate": 0.0006,
+ "loss": 6.518143653869629,
+ "step": 278
+ },
+ {
+ "epoch": 3.880349344978166,
+ "grad_norm": 0.020839324221014977,
+ "learning_rate": 0.0006,
+ "loss": 6.546667098999023,
+ "step": 279
+ },
+ {
+ "epoch": 3.8943231441048036,
+ "grad_norm": 0.018377432599663734,
+ "learning_rate": 0.0006,
+ "loss": 6.478307723999023,
+ "step": 280
+ },
+ {
+ "epoch": 3.908296943231441,
+ "grad_norm": 0.021713724359869957,
+ "learning_rate": 0.0006,
+ "loss": 6.485574245452881,
+ "step": 281
+ },
+ {
+ "epoch": 3.9222707423580787,
+ "grad_norm": 0.01359301246702671,
+ "learning_rate": 0.0006,
+ "loss": 6.570268630981445,
+ "step": 282
+ },
+ {
+ "epoch": 3.936244541484716,
+ "grad_norm": 0.016233332455158234,
+ "learning_rate": 0.0006,
+ "loss": 6.57150936126709,
+ "step": 283
+ },
+ {
+ "epoch": 3.950218340611354,
+ "grad_norm": 0.020687608048319817,
+ "learning_rate": 0.0006,
+ "loss": 6.527956008911133,
+ "step": 284
+ },
+ {
+ "epoch": 3.964192139737991,
+ "grad_norm": 0.02032964862883091,
+ "learning_rate": 0.0006,
+ "loss": 6.468169689178467,
+ "step": 285
+ },
+ {
+ "epoch": 3.978165938864629,
+ "grad_norm": 0.016676288098096848,
+ "learning_rate": 0.0006,
+ "loss": 6.448418617248535,
+ "step": 286
+ },
+ {
+ "epoch": 3.992139737991266,
+ "grad_norm": 0.0168539360165596,
+ "learning_rate": 0.0006,
+ "loss": 6.461180686950684,
+ "step": 287
+ },
+ {
+ "epoch": 4.0,
+ "grad_norm": 0.02666233666241169,
+ "learning_rate": 0.0006,
+ "loss": 6.344893932342529,
+ "step": 288
+ },
+ {
+ "epoch": 4.0,
+ "eval_loss": 6.5384979248046875,
+ "eval_runtime": 58.814,
+ "eval_samples_per_second": 41.521,
+ "eval_steps_per_second": 1.309,
+ "step": 288
+ },
+ {
+ "epoch": 4.013973799126638,
+ "grad_norm": 0.04154708981513977,
+ "learning_rate": 0.0006,
+ "loss": 6.41855525970459,
+ "step": 289
+ },
+ {
+ "epoch": 4.0279475982532755,
+ "grad_norm": 0.04895668104290962,
+ "learning_rate": 0.0006,
+ "loss": 6.499302864074707,
+ "step": 290
+ },
+ {
+ "epoch": 4.041921397379912,
+ "grad_norm": 0.034464482218027115,
+ "learning_rate": 0.0006,
+ "loss": 6.438611030578613,
+ "step": 291
+ },
+ {
+ "epoch": 4.05589519650655,
+ "grad_norm": 0.06845773756504059,
+ "learning_rate": 0.0006,
+ "loss": 6.538361072540283,
+ "step": 292
+ },
+ {
+ "epoch": 4.069868995633188,
+ "grad_norm": 0.053768858313560486,
+ "learning_rate": 0.0006,
+ "loss": 6.527524948120117,
+ "step": 293
+ },
+ {
+ "epoch": 4.083842794759826,
+ "grad_norm": 0.039037931710481644,
+ "learning_rate": 0.0006,
+ "loss": 6.503847122192383,
+ "step": 294
+ },
+ {
+ "epoch": 4.097816593886463,
+ "grad_norm": 0.03962196782231331,
+ "learning_rate": 0.0006,
+ "loss": 6.475178241729736,
+ "step": 295
+ },
+ {
+ "epoch": 4.1117903930131,
+ "grad_norm": 0.049101535230875015,
+ "learning_rate": 0.0006,
+ "loss": 6.474587917327881,
+ "step": 296
+ },
+ {
+ "epoch": 4.125764192139738,
+ "grad_norm": 0.02830282226204872,
+ "learning_rate": 0.0006,
+ "loss": 6.45065975189209,
+ "step": 297
+ },
+ {
+ "epoch": 4.139737991266376,
+ "grad_norm": 0.023153482005000114,
+ "learning_rate": 0.0006,
+ "loss": 6.5218706130981445,
+ "step": 298
+ },
+ {
+ "epoch": 4.153711790393013,
+ "grad_norm": 0.025064002722501755,
+ "learning_rate": 0.0006,
+ "loss": 6.456475257873535,
+ "step": 299
+ },
+ {
+ "epoch": 4.1676855895196505,
+ "grad_norm": 0.02357092685997486,
+ "learning_rate": 0.0006,
+ "loss": 6.485952377319336,
+ "step": 300
+ },
+ {
+ "epoch": 4.181659388646288,
+ "grad_norm": 0.021832305938005447,
+ "learning_rate": 0.0006,
+ "loss": 6.425506591796875,
+ "step": 301
+ },
+ {
+ "epoch": 4.195633187772926,
+ "grad_norm": 0.0227900929749012,
+ "learning_rate": 0.0006,
+ "loss": 6.439155578613281,
+ "step": 302
+ },
+ {
+ "epoch": 4.209606986899563,
+ "grad_norm": 0.015961607918143272,
+ "learning_rate": 0.0006,
+ "loss": 6.460862636566162,
+ "step": 303
+ },
+ {
+ "epoch": 4.223580786026201,
+ "grad_norm": 0.017876390367746353,
+ "learning_rate": 0.0006,
+ "loss": 6.44564151763916,
+ "step": 304
+ },
+ {
+ "epoch": 4.2375545851528384,
+ "grad_norm": 0.013703204691410065,
+ "learning_rate": 0.0006,
+ "loss": 6.418414115905762,
+ "step": 305
+ },
+ {
+ "epoch": 4.251528384279476,
+ "grad_norm": 0.015240306034684181,
+ "learning_rate": 0.0006,
+ "loss": 6.411238670349121,
+ "step": 306
+ },
+ {
+ "epoch": 4.265502183406113,
+ "grad_norm": 0.01587662845849991,
+ "learning_rate": 0.0006,
+ "loss": 6.395642280578613,
+ "step": 307
+ },
+ {
+ "epoch": 4.279475982532751,
+ "grad_norm": 0.014650700613856316,
+ "learning_rate": 0.0006,
+ "loss": 6.302793025970459,
+ "step": 308
+ },
+ {
+ "epoch": 4.293449781659389,
+ "grad_norm": 0.016533225774765015,
+ "learning_rate": 0.0006,
+ "loss": 6.380588054656982,
+ "step": 309
+ },
+ {
+ "epoch": 4.307423580786026,
+ "grad_norm": 0.019963741302490234,
+ "learning_rate": 0.0006,
+ "loss": 6.345336437225342,
+ "step": 310
+ },
+ {
+ "epoch": 4.321397379912664,
+ "grad_norm": 0.02237936295568943,
+ "learning_rate": 0.0006,
+ "loss": 6.387770652770996,
+ "step": 311
+ },
+ {
+ "epoch": 4.335371179039301,
+ "grad_norm": 0.026658328250050545,
+ "learning_rate": 0.0006,
+ "loss": 6.323662757873535,
+ "step": 312
+ },
+ {
+ "epoch": 4.349344978165939,
+ "grad_norm": 0.02852284163236618,
+ "learning_rate": 0.0006,
+ "loss": 6.382204055786133,
+ "step": 313
+ },
+ {
+ "epoch": 4.3633187772925766,
+ "grad_norm": 0.028815090656280518,
+ "learning_rate": 0.0006,
+ "loss": 6.402390480041504,
+ "step": 314
+ },
+ {
+ "epoch": 4.377292576419214,
+ "grad_norm": 0.029393529519438744,
+ "learning_rate": 0.0006,
+ "loss": 6.280278205871582,
+ "step": 315
+ },
+ {
+ "epoch": 4.391266375545851,
+ "grad_norm": 0.02186041697859764,
+ "learning_rate": 0.0006,
+ "loss": 6.302707672119141,
+ "step": 316
+ },
+ {
+ "epoch": 4.405240174672489,
+ "grad_norm": 0.02071218006312847,
+ "learning_rate": 0.0006,
+ "loss": 6.3365020751953125,
+ "step": 317
+ },
+ {
+ "epoch": 4.419213973799127,
+ "grad_norm": 0.02779117226600647,
+ "learning_rate": 0.0006,
+ "loss": 6.252157211303711,
+ "step": 318
+ },
+ {
+ "epoch": 4.4331877729257645,
+ "grad_norm": 0.03470654785633087,
+ "learning_rate": 0.0006,
+ "loss": 6.319070339202881,
+ "step": 319
+ },
+ {
+ "epoch": 4.447161572052401,
+ "grad_norm": 0.037067804485559464,
+ "learning_rate": 0.0006,
+ "loss": 6.314116477966309,
+ "step": 320
+ },
+ {
+ "epoch": 4.461135371179039,
+ "grad_norm": 0.032182756811380386,
+ "learning_rate": 0.0006,
+ "loss": 6.267298698425293,
+ "step": 321
+ },
+ {
+ "epoch": 4.475109170305677,
+ "grad_norm": 0.026305217295885086,
+ "learning_rate": 0.0006,
+ "loss": 6.331688404083252,
+ "step": 322
+ },
+ {
+ "epoch": 4.489082969432315,
+ "grad_norm": 0.027228357270359993,
+ "learning_rate": 0.0006,
+ "loss": 6.31139612197876,
+ "step": 323
+ },
+ {
+ "epoch": 4.503056768558952,
+ "grad_norm": 0.025781169533729553,
+ "learning_rate": 0.0006,
+ "loss": 6.194684982299805,
+ "step": 324
+ },
+ {
+ "epoch": 4.517030567685589,
+ "grad_norm": 0.029186977073550224,
+ "learning_rate": 0.0006,
+ "loss": 6.31306266784668,
+ "step": 325
+ },
+ {
+ "epoch": 4.531004366812227,
+ "grad_norm": 0.023608777672052383,
+ "learning_rate": 0.0006,
+ "loss": 6.285243034362793,
+ "step": 326
+ },
+ {
+ "epoch": 4.544978165938865,
+ "grad_norm": 0.02267594076693058,
+ "learning_rate": 0.0006,
+ "loss": 6.30342960357666,
+ "step": 327
+ },
+ {
+ "epoch": 4.558951965065502,
+ "grad_norm": 0.01926310732960701,
+ "learning_rate": 0.0006,
+ "loss": 6.280606269836426,
+ "step": 328
+ },
+ {
+ "epoch": 4.5729257641921395,
+ "grad_norm": 0.026484837755560875,
+ "learning_rate": 0.0006,
+ "loss": 6.237980365753174,
+ "step": 329
+ },
+ {
+ "epoch": 4.586899563318777,
+ "grad_norm": 0.026228854432702065,
+ "learning_rate": 0.0006,
+ "loss": 6.3018293380737305,
+ "step": 330
+ },
+ {
+ "epoch": 4.600873362445415,
+ "grad_norm": 0.022096967324614525,
+ "learning_rate": 0.0006,
+ "loss": 6.253863334655762,
+ "step": 331
+ },
+ {
+ "epoch": 4.614847161572053,
+ "grad_norm": 0.027223890647292137,
+ "learning_rate": 0.0006,
+ "loss": 6.166882038116455,
+ "step": 332
+ },
+ {
+ "epoch": 4.62882096069869,
+ "grad_norm": 0.03160124272108078,
+ "learning_rate": 0.0006,
+ "loss": 6.230309963226318,
+ "step": 333
+ },
+ {
+ "epoch": 4.642794759825327,
+ "grad_norm": 0.04110539332032204,
+ "learning_rate": 0.0006,
+ "loss": 6.2334818840026855,
+ "step": 334
+ },
+ {
+ "epoch": 4.656768558951965,
+ "grad_norm": 0.04355933889746666,
+ "learning_rate": 0.0006,
+ "loss": 6.160956382751465,
+ "step": 335
+ },
+ {
+ "epoch": 4.670742358078603,
+ "grad_norm": 0.03611086308956146,
+ "learning_rate": 0.0006,
+ "loss": 6.23399543762207,
+ "step": 336
+ },
+ {
+ "epoch": 4.68471615720524,
+ "grad_norm": 0.051811400800943375,
+ "learning_rate": 0.0006,
+ "loss": 6.241855144500732,
+ "step": 337
+ },
+ {
+ "epoch": 4.698689956331878,
+ "grad_norm": 0.057434193789958954,
+ "learning_rate": 0.0006,
+ "loss": 6.220993995666504,
+ "step": 338
+ },
+ {
+ "epoch": 4.712663755458515,
+ "grad_norm": 0.04420953616499901,
+ "learning_rate": 0.0006,
+ "loss": 6.300461292266846,
+ "step": 339
+ },
+ {
+ "epoch": 4.726637554585153,
+ "grad_norm": 0.03515457361936569,
+ "learning_rate": 0.0006,
+ "loss": 6.269349098205566,
+ "step": 340
+ },
+ {
+ "epoch": 4.74061135371179,
+ "grad_norm": 0.03222600743174553,
+ "learning_rate": 0.0006,
+ "loss": 6.248613357543945,
+ "step": 341
+ },
+ {
+ "epoch": 4.754585152838428,
+ "grad_norm": 0.0384046845138073,
+ "learning_rate": 0.0006,
+ "loss": 6.266979217529297,
+ "step": 342
+ },
+ {
+ "epoch": 4.7685589519650655,
+ "grad_norm": 0.0514867827296257,
+ "learning_rate": 0.0006,
+ "loss": 6.234103202819824,
+ "step": 343
+ },
+ {
+ "epoch": 4.782532751091703,
+ "grad_norm": 0.050747793167829514,
+ "learning_rate": 0.0006,
+ "loss": 6.322582244873047,
+ "step": 344
+ },
+ {
+ "epoch": 4.796506550218341,
+ "grad_norm": 0.040878988802433014,
+ "learning_rate": 0.0006,
+ "loss": 6.212509632110596,
+ "step": 345
+ },
+ {
+ "epoch": 4.810480349344978,
+ "grad_norm": 0.034655820578336716,
+ "learning_rate": 0.0006,
+ "loss": 6.265879154205322,
+ "step": 346
+ },
+ {
+ "epoch": 4.824454148471616,
+ "grad_norm": 0.04202224686741829,
+ "learning_rate": 0.0006,
+ "loss": 6.2896409034729,
+ "step": 347
+ },
+ {
+ "epoch": 4.8384279475982535,
+ "grad_norm": 0.03632277995347977,
+ "learning_rate": 0.0006,
+ "loss": 6.253917217254639,
+ "step": 348
+ },
+ {
+ "epoch": 4.85240174672489,
+ "grad_norm": 0.028645765036344528,
+ "learning_rate": 0.0006,
+ "loss": 6.207965850830078,
+ "step": 349
+ },
+ {
+ "epoch": 4.866375545851528,
+ "grad_norm": 0.025305170565843582,
+ "learning_rate": 0.0006,
+ "loss": 6.168068885803223,
+ "step": 350
+ },
+ {
+ "epoch": 4.880349344978166,
+ "grad_norm": 0.022137803956866264,
+ "learning_rate": 0.0006,
+ "loss": 6.235283374786377,
+ "step": 351
+ },
+ {
+ "epoch": 4.894323144104804,
+ "grad_norm": 0.01814538985490799,
+ "learning_rate": 0.0006,
+ "loss": 6.259641647338867,
+ "step": 352
+ },
+ {
+ "epoch": 4.908296943231441,
+ "grad_norm": 0.018465086817741394,
+ "learning_rate": 0.0006,
+ "loss": 6.111738681793213,
+ "step": 353
+ },
+ {
+ "epoch": 4.922270742358078,
+ "grad_norm": 0.018638933077454567,
+ "learning_rate": 0.0006,
+ "loss": 6.225642204284668,
+ "step": 354
+ },
+ {
+ "epoch": 4.936244541484716,
+ "grad_norm": 0.016461260616779327,
+ "learning_rate": 0.0006,
+ "loss": 6.140647888183594,
+ "step": 355
+ },
+ {
+ "epoch": 4.950218340611354,
+ "grad_norm": 0.015458385460078716,
+ "learning_rate": 0.0006,
+ "loss": 6.155224800109863,
+ "step": 356
+ },
+ {
+ "epoch": 4.964192139737992,
+ "grad_norm": 0.014934048056602478,
+ "learning_rate": 0.0006,
+ "loss": 6.101555824279785,
+ "step": 357
+ },
+ {
+ "epoch": 4.978165938864628,
+ "grad_norm": 0.013979545794427395,
+ "learning_rate": 0.0006,
+ "loss": 6.161718845367432,
+ "step": 358
+ },
+ {
+ "epoch": 4.992139737991266,
+ "grad_norm": 0.01640394888818264,
+ "learning_rate": 0.0006,
+ "loss": 6.055768966674805,
+ "step": 359
+ },
+ {
+ "epoch": 5.0,
+ "grad_norm": 0.015135680325329304,
+ "learning_rate": 0.0006,
+ "loss": 6.169695854187012,
+ "step": 360
+ },
+ {
+ "epoch": 5.0,
+ "eval_loss": 6.16408109664917,
+ "eval_runtime": 59.3649,
+ "eval_samples_per_second": 41.135,
+ "eval_steps_per_second": 1.297,
+ "step": 360
+ },
+ {
+ "epoch": 5.013973799126638,
+ "grad_norm": 0.012849048711359501,
+ "learning_rate": 0.0006,
+ "loss": 6.055830001831055,
+ "step": 361
+ },
+ {
+ "epoch": 5.0279475982532755,
+ "grad_norm": 0.018169576302170753,
+ "learning_rate": 0.0006,
+ "loss": 6.082655429840088,
+ "step": 362
+ },
+ {
+ "epoch": 5.041921397379912,
+ "grad_norm": 0.024293430149555206,
+ "learning_rate": 0.0006,
+ "loss": 6.109032154083252,
+ "step": 363
+ },
+ {
+ "epoch": 5.05589519650655,
+ "grad_norm": 0.03266787901520729,
+ "learning_rate": 0.0006,
+ "loss": 6.030613899230957,
+ "step": 364
+ },
+ {
+ "epoch": 5.069868995633188,
+ "grad_norm": 0.03572266176342964,
+ "learning_rate": 0.0006,
+ "loss": 6.050390720367432,
+ "step": 365
+ },
+ {
+ "epoch": 5.083842794759826,
+ "grad_norm": 0.032625213265419006,
+ "learning_rate": 0.0006,
+ "loss": 6.169313430786133,
+ "step": 366
+ },
+ {
+ "epoch": 5.097816593886463,
+ "grad_norm": 0.04051872715353966,
+ "learning_rate": 0.0006,
+ "loss": 6.10398006439209,
+ "step": 367
+ },
+ {
+ "epoch": 5.1117903930131,
+ "grad_norm": 0.0351213738322258,
+ "learning_rate": 0.0006,
+ "loss": 6.121973514556885,
+ "step": 368
+ },
+ {
+ "epoch": 5.125764192139738,
+ "grad_norm": 0.03523759916424751,
+ "learning_rate": 0.0006,
+ "loss": 6.0626654624938965,
+ "step": 369
+ },
+ {
+ "epoch": 5.139737991266376,
+ "grad_norm": 0.02706182189285755,
+ "learning_rate": 0.0006,
+ "loss": 6.049354076385498,
+ "step": 370
+ },
+ {
+ "epoch": 5.153711790393013,
+ "grad_norm": 0.024212589487433434,
+ "learning_rate": 0.0006,
+ "loss": 6.034826278686523,
+ "step": 371
+ },
+ {
+ "epoch": 5.1676855895196505,
+ "grad_norm": 0.024798082187771797,
+ "learning_rate": 0.0006,
+ "loss": 6.004058837890625,
+ "step": 372
+ },
+ {
+ "epoch": 5.181659388646288,
+ "grad_norm": 0.02325567416846752,
+ "learning_rate": 0.0006,
+ "loss": 5.986461639404297,
+ "step": 373
+ },
+ {
+ "epoch": 5.195633187772926,
+ "grad_norm": 0.019060570746660233,
+ "learning_rate": 0.0006,
+ "loss": 6.059736251831055,
+ "step": 374
+ },
+ {
+ "epoch": 5.209606986899563,
+ "grad_norm": 0.016822976991534233,
+ "learning_rate": 0.0006,
+ "loss": 6.072957515716553,
+ "step": 375
+ },
+ {
+ "epoch": 5.223580786026201,
+ "grad_norm": 0.017218658700585365,
+ "learning_rate": 0.0006,
+ "loss": 6.031739234924316,
+ "step": 376
+ },
+ {
+ "epoch": 5.2375545851528384,
+ "grad_norm": 0.014889383688569069,
+ "learning_rate": 0.0006,
+ "loss": 6.025674343109131,
+ "step": 377
+ },
+ {
+ "epoch": 5.251528384279476,
+ "grad_norm": 0.01708107627928257,
+ "learning_rate": 0.0006,
+ "loss": 6.090451240539551,
+ "step": 378
+ },
+ {
+ "epoch": 5.265502183406113,
+ "grad_norm": 0.020260317251086235,
+ "learning_rate": 0.0006,
+ "loss": 6.094257354736328,
+ "step": 379
+ },
+ {
+ "epoch": 5.279475982532751,
+ "grad_norm": 0.020110400393605232,
+ "learning_rate": 0.0006,
+ "loss": 6.021188735961914,
+ "step": 380
+ },
+ {
+ "epoch": 5.293449781659389,
+ "grad_norm": 0.02027830481529236,
+ "learning_rate": 0.0006,
+ "loss": 6.022156238555908,
+ "step": 381
+ },
+ {
+ "epoch": 5.307423580786026,
+ "grad_norm": 0.022747062146663666,
+ "learning_rate": 0.0006,
+ "loss": 6.011836051940918,
+ "step": 382
+ },
+ {
+ "epoch": 5.321397379912664,
+ "grad_norm": 0.01990230567753315,
+ "learning_rate": 0.0006,
+ "loss": 6.003054618835449,
+ "step": 383
+ },
+ {
+ "epoch": 5.335371179039301,
+ "grad_norm": 0.01596238650381565,
+ "learning_rate": 0.0006,
+ "loss": 5.991410732269287,
+ "step": 384
+ },
+ {
+ "epoch": 5.349344978165939,
+ "grad_norm": 0.015847105532884598,
+ "learning_rate": 0.0006,
+ "loss": 6.100622177124023,
+ "step": 385
+ },
+ {
+ "epoch": 5.3633187772925766,
+ "grad_norm": 0.016179397702217102,
+ "learning_rate": 0.0006,
+ "loss": 6.032659530639648,
+ "step": 386
+ },
+ {
+ "epoch": 5.377292576419214,
+ "grad_norm": 0.018256602808833122,
+ "learning_rate": 0.0006,
+ "loss": 5.961983680725098,
+ "step": 387
+ },
+ {
+ "epoch": 5.391266375545851,
+ "grad_norm": 0.02805912122130394,
+ "learning_rate": 0.0006,
+ "loss": 5.983541965484619,
+ "step": 388
+ },
+ {
+ "epoch": 5.405240174672489,
+ "grad_norm": 0.039082255214452744,
+ "learning_rate": 0.0006,
+ "loss": 6.021474361419678,
+ "step": 389
+ },
+ {
+ "epoch": 5.419213973799127,
+ "grad_norm": 0.036757659167051315,
+ "learning_rate": 0.0006,
+ "loss": 6.005046367645264,
+ "step": 390
+ },
+ {
+ "epoch": 5.4331877729257645,
+ "grad_norm": 0.035277366638183594,
+ "learning_rate": 0.0006,
+ "loss": 6.036296844482422,
+ "step": 391
+ },
+ {
+ "epoch": 5.447161572052401,
+ "grad_norm": 0.034404635429382324,
+ "learning_rate": 0.0006,
+ "loss": 5.994539260864258,
+ "step": 392
+ },
+ {
+ "epoch": 5.461135371179039,
+ "grad_norm": 0.0377206988632679,
+ "learning_rate": 0.0006,
+ "loss": 5.941534519195557,
+ "step": 393
+ },
+ {
+ "epoch": 5.475109170305677,
+ "grad_norm": 0.0389922633767128,
+ "learning_rate": 0.0006,
+ "loss": 5.983644485473633,
+ "step": 394
+ },
+ {
+ "epoch": 5.489082969432315,
+ "grad_norm": 0.04176778718829155,
+ "learning_rate": 0.0006,
+ "loss": 6.030922889709473,
+ "step": 395
+ },
+ {
+ "epoch": 5.503056768558952,
+ "grad_norm": 0.0337153784930706,
+ "learning_rate": 0.0006,
+ "loss": 5.961367130279541,
+ "step": 396
+ },
+ {
+ "epoch": 5.517030567685589,
+ "grad_norm": 0.036116816103458405,
+ "learning_rate": 0.0006,
+ "loss": 5.819280624389648,
+ "step": 397
+ },
+ {
+ "epoch": 5.531004366812227,
+ "grad_norm": 0.030724041163921356,
+ "learning_rate": 0.0006,
+ "loss": 5.906380653381348,
+ "step": 398
+ },
+ {
+ "epoch": 5.544978165938865,
+ "grad_norm": 0.030264202505350113,
+ "learning_rate": 0.0006,
+ "loss": 5.975833415985107,
+ "step": 399
+ },
+ {
+ "epoch": 5.558951965065502,
+ "grad_norm": 0.032096318900585175,
+ "learning_rate": 0.0006,
+ "loss": 6.001348972320557,
+ "step": 400
+ },
+ {
+ "epoch": 5.5729257641921395,
+ "grad_norm": 0.030579356476664543,
+ "learning_rate": 0.0006,
+ "loss": 5.928768634796143,
+ "step": 401
+ },
+ {
+ "epoch": 5.586899563318777,
+ "grad_norm": 0.028241973370313644,
+ "learning_rate": 0.0006,
+ "loss": 5.92582893371582,
+ "step": 402
+ },
+ {
+ "epoch": 5.600873362445415,
+ "grad_norm": 0.023086953908205032,
+ "learning_rate": 0.0006,
+ "loss": 5.892926216125488,
+ "step": 403
+ },
+ {
+ "epoch": 5.614847161572053,
+ "grad_norm": 0.02547992393374443,
+ "learning_rate": 0.0006,
+ "loss": 5.903195381164551,
+ "step": 404
+ },
+ {
+ "epoch": 5.62882096069869,
+ "grad_norm": 0.023089831694960594,
+ "learning_rate": 0.0006,
+ "loss": 5.912033557891846,
+ "step": 405
+ },
+ {
+ "epoch": 5.642794759825327,
+ "grad_norm": 0.025446368381381035,
+ "learning_rate": 0.0006,
+ "loss": 5.926138401031494,
+ "step": 406
+ },
+ {
+ "epoch": 5.656768558951965,
+ "grad_norm": 0.03748747706413269,
+ "learning_rate": 0.0006,
+ "loss": 5.913451194763184,
+ "step": 407
+ },
+ {
+ "epoch": 5.670742358078603,
+ "grad_norm": 0.03493810072541237,
+ "learning_rate": 0.0006,
+ "loss": 5.97633695602417,
+ "step": 408
+ },
+ {
+ "epoch": 5.68471615720524,
+ "grad_norm": 0.023131275549530983,
+ "learning_rate": 0.0006,
+ "loss": 5.899571418762207,
+ "step": 409
+ },
+ {
+ "epoch": 5.698689956331878,
+ "grad_norm": 0.02986014075577259,
+ "learning_rate": 0.0006,
+ "loss": 5.894903182983398,
+ "step": 410
+ },
+ {
+ "epoch": 5.712663755458515,
+ "grad_norm": 0.030171144753694534,
+ "learning_rate": 0.0006,
+ "loss": 5.876595497131348,
+ "step": 411
+ },
+ {
+ "epoch": 5.726637554585153,
+ "grad_norm": 0.029377546161413193,
+ "learning_rate": 0.0006,
+ "loss": 5.911637306213379,
+ "step": 412
+ },
+ {
+ "epoch": 5.74061135371179,
+ "grad_norm": 0.029829490929841995,
+ "learning_rate": 0.0006,
+ "loss": 5.819629669189453,
+ "step": 413
+ },
+ {
+ "epoch": 5.754585152838428,
+ "grad_norm": 0.024139299988746643,
+ "learning_rate": 0.0006,
+ "loss": 5.81728458404541,
+ "step": 414
+ },
+ {
+ "epoch": 5.7685589519650655,
+ "grad_norm": 0.021453432738780975,
+ "learning_rate": 0.0006,
+ "loss": 5.987326145172119,
+ "step": 415
+ },
+ {
+ "epoch": 5.782532751091703,
+ "grad_norm": 0.017942246049642563,
+ "learning_rate": 0.0006,
+ "loss": 5.893008232116699,
+ "step": 416
+ },
+ {
+ "epoch": 5.796506550218341,
+ "grad_norm": 0.019724637269973755,
+ "learning_rate": 0.0006,
+ "loss": 5.912441253662109,
+ "step": 417
+ },
+ {
+ "epoch": 5.810480349344978,
+ "grad_norm": 0.018597912043333054,
+ "learning_rate": 0.0006,
+ "loss": 5.860394477844238,
+ "step": 418
+ },
+ {
+ "epoch": 5.824454148471616,
+ "grad_norm": 0.016232412308454514,
+ "learning_rate": 0.0006,
+ "loss": 5.94991397857666,
+ "step": 419
+ },
+ {
+ "epoch": 5.8384279475982535,
+ "grad_norm": 0.01667204685509205,
+ "learning_rate": 0.0006,
+ "loss": 5.8506364822387695,
+ "step": 420
+ },
+ {
+ "epoch": 5.85240174672489,
+ "grad_norm": 0.014232151210308075,
+ "learning_rate": 0.0006,
+ "loss": 5.928424835205078,
+ "step": 421
+ },
+ {
+ "epoch": 5.866375545851528,
+ "grad_norm": 0.01570476032793522,
+ "learning_rate": 0.0006,
+ "loss": 5.953692436218262,
+ "step": 422
+ },
+ {
+ "epoch": 5.880349344978166,
+ "grad_norm": 0.01586179807782173,
+ "learning_rate": 0.0006,
+ "loss": 5.820014476776123,
+ "step": 423
+ },
+ {
+ "epoch": 5.894323144104804,
+ "grad_norm": 0.014729145914316177,
+ "learning_rate": 0.0006,
+ "loss": 5.860154628753662,
+ "step": 424
+ },
+ {
+ "epoch": 5.908296943231441,
+ "grad_norm": 0.013111459091305733,
+ "learning_rate": 0.0006,
+ "loss": 5.85988187789917,
+ "step": 425
+ },
+ {
+ "epoch": 5.922270742358078,
+ "grad_norm": 0.012937922962009907,
+ "learning_rate": 0.0006,
+ "loss": 5.756265640258789,
+ "step": 426
+ },
+ {
+ "epoch": 5.936244541484716,
+ "grad_norm": 0.01377453189343214,
+ "learning_rate": 0.0006,
+ "loss": 5.8333048820495605,
+ "step": 427
+ },
+ {
+ "epoch": 5.950218340611354,
+ "grad_norm": 0.014054795727133751,
+ "learning_rate": 0.0006,
+ "loss": 5.803333759307861,
+ "step": 428
+ },
+ {
+ "epoch": 5.964192139737992,
+ "grad_norm": 0.01674959622323513,
+ "learning_rate": 0.0006,
+ "loss": 5.8718485832214355,
+ "step": 429
+ },
+ {
+ "epoch": 5.978165938864628,
+ "grad_norm": 0.018733017146587372,
+ "learning_rate": 0.0006,
+ "loss": 5.875979900360107,
+ "step": 430
+ },
+ {
+ "epoch": 5.992139737991266,
+ "grad_norm": 0.02088983915746212,
+ "learning_rate": 0.0006,
+ "loss": 5.763634204864502,
+ "step": 431
+ },
+ {
+ "epoch": 6.0,
+ "grad_norm": 0.026705985888838768,
+ "learning_rate": 0.0006,
+ "loss": 5.6922478675842285,
+ "step": 432
+ },
+ {
+ "epoch": 6.0,
+ "eval_loss": 5.847958087921143,
+ "eval_runtime": 58.5451,
+ "eval_samples_per_second": 41.711,
+ "eval_steps_per_second": 1.315,
+ "step": 432
+ },
+ {
+ "epoch": 6.013973799126638,
+ "grad_norm": 0.03181544691324234,
+ "learning_rate": 0.0006,
+ "loss": 5.786368370056152,
+ "step": 433
+ },
+ {
+ "epoch": 6.0279475982532755,
+ "grad_norm": 0.03238112851977348,
+ "learning_rate": 0.0006,
+ "loss": 5.810310363769531,
+ "step": 434
+ },
+ {
+ "epoch": 6.041921397379912,
+ "grad_norm": 0.03916756808757782,
+ "learning_rate": 0.0006,
+ "loss": 5.7761993408203125,
+ "step": 435
+ },
+ {
+ "epoch": 6.05589519650655,
+ "grad_norm": 0.044009730219841,
+ "learning_rate": 0.0006,
+ "loss": 5.898112773895264,
+ "step": 436
+ },
+ {
+ "epoch": 6.069868995633188,
+ "grad_norm": 0.045235246419906616,
+ "learning_rate": 0.0006,
+ "loss": 5.876371383666992,
+ "step": 437
+ },
+ {
+ "epoch": 6.083842794759826,
+ "grad_norm": 0.050218384712934494,
+ "learning_rate": 0.0006,
+ "loss": 5.805103302001953,
+ "step": 438
+ },
+ {
+ "epoch": 6.097816593886463,
+ "grad_norm": 0.05444490164518356,
+ "learning_rate": 0.0006,
+ "loss": 5.826424598693848,
+ "step": 439
+ },
+ {
+ "epoch": 6.1117903930131,
+ "grad_norm": 0.04890982061624527,
+ "learning_rate": 0.0006,
+ "loss": 5.7735395431518555,
+ "step": 440
+ },
+ {
+ "epoch": 6.125764192139738,
+ "grad_norm": 0.05478639155626297,
+ "learning_rate": 0.0006,
+ "loss": 5.823677062988281,
+ "step": 441
+ },
+ {
+ "epoch": 6.139737991266376,
+ "grad_norm": 0.057562313973903656,
+ "learning_rate": 0.0006,
+ "loss": 5.889334678649902,
+ "step": 442
+ },
+ {
+ "epoch": 6.153711790393013,
+ "grad_norm": 0.06447703391313553,
+ "learning_rate": 0.0006,
+ "loss": 5.84621524810791,
+ "step": 443
+ },
+ {
+ "epoch": 6.1676855895196505,
+ "grad_norm": 0.048861872404813766,
+ "learning_rate": 0.0006,
+ "loss": 5.931595802307129,
+ "step": 444
+ },
+ {
+ "epoch": 6.181659388646288,
+ "grad_norm": 0.05521339178085327,
+ "learning_rate": 0.0006,
+ "loss": 5.8229780197143555,
+ "step": 445
+ },
+ {
+ "epoch": 6.195633187772926,
+ "grad_norm": 0.053666990250349045,
+ "learning_rate": 0.0006,
+ "loss": 5.879191875457764,
+ "step": 446
+ },
+ {
+ "epoch": 6.209606986899563,
+ "grad_norm": 0.0451025515794754,
+ "learning_rate": 0.0006,
+ "loss": 5.801628112792969,
+ "step": 447
+ },
+ {
+ "epoch": 6.223580786026201,
+ "grad_norm": 0.04291655868291855,
+ "learning_rate": 0.0006,
+ "loss": 5.914680480957031,
+ "step": 448
+ },
+ {
+ "epoch": 6.2375545851528384,
+ "grad_norm": 0.035102490335702896,
+ "learning_rate": 0.0006,
+ "loss": 5.843682765960693,
+ "step": 449
+ },
+ {
+ "epoch": 6.251528384279476,
+ "grad_norm": 0.03403995931148529,
+ "learning_rate": 0.0006,
+ "loss": 5.827154159545898,
+ "step": 450
+ },
+ {
+ "epoch": 6.265502183406113,
+ "grad_norm": 0.03444375470280647,
+ "learning_rate": 0.0006,
+ "loss": 5.800136566162109,
+ "step": 451
+ },
+ {
+ "epoch": 6.279475982532751,
+ "grad_norm": 0.03165159001946449,
+ "learning_rate": 0.0006,
+ "loss": 5.906252384185791,
+ "step": 452
+ },
+ {
+ "epoch": 6.293449781659389,
+ "grad_norm": 0.026153093203902245,
+ "learning_rate": 0.0006,
+ "loss": 5.7423200607299805,
+ "step": 453
+ },
+ {
+ "epoch": 6.307423580786026,
+ "grad_norm": 0.024357657879590988,
+ "learning_rate": 0.0006,
+ "loss": 5.701364517211914,
+ "step": 454
+ },
+ {
+ "epoch": 6.321397379912664,
+ "grad_norm": 0.021508049219846725,
+ "learning_rate": 0.0006,
+ "loss": 5.8328094482421875,
+ "step": 455
+ },
+ {
+ "epoch": 6.335371179039301,
+ "grad_norm": 0.017313921824097633,
+ "learning_rate": 0.0006,
+ "loss": 5.792145252227783,
+ "step": 456
+ },
+ {
+ "epoch": 6.349344978165939,
+ "grad_norm": 0.018563397228717804,
+ "learning_rate": 0.0006,
+ "loss": 5.732672214508057,
+ "step": 457
+ },
+ {
+ "epoch": 6.3633187772925766,
+ "grad_norm": 0.016568679362535477,
+ "learning_rate": 0.0006,
+ "loss": 5.81948184967041,
+ "step": 458
+ },
+ {
+ "epoch": 6.377292576419214,
+ "grad_norm": 0.014133770950138569,
+ "learning_rate": 0.0006,
+ "loss": 5.734982490539551,
+ "step": 459
+ },
+ {
+ "epoch": 6.391266375545851,
+ "grad_norm": 0.014114447869360447,
+ "learning_rate": 0.0006,
+ "loss": 5.70073127746582,
+ "step": 460
+ },
+ {
+ "epoch": 6.405240174672489,
+ "grad_norm": 0.012907393276691437,
+ "learning_rate": 0.0006,
+ "loss": 5.758626937866211,
+ "step": 461
+ },
+ {
+ "epoch": 6.419213973799127,
+ "grad_norm": 0.012673444114625454,
+ "learning_rate": 0.0006,
+ "loss": 5.775138854980469,
+ "step": 462
+ },
+ {
+ "epoch": 6.4331877729257645,
+ "grad_norm": 0.01311410591006279,
+ "learning_rate": 0.0006,
+ "loss": 5.70920991897583,
+ "step": 463
+ },
+ {
+ "epoch": 6.447161572052401,
+ "grad_norm": 0.012935544364154339,
+ "learning_rate": 0.0006,
+ "loss": 5.781479835510254,
+ "step": 464
+ },
+ {
+ "epoch": 6.461135371179039,
+ "grad_norm": 0.01439738366752863,
+ "learning_rate": 0.0006,
+ "loss": 5.691315650939941,
+ "step": 465
+ },
+ {
+ "epoch": 6.475109170305677,
+ "grad_norm": 0.01205186452716589,
+ "learning_rate": 0.0006,
+ "loss": 5.662499904632568,
+ "step": 466
+ },
+ {
+ "epoch": 6.489082969432315,
+ "grad_norm": 0.011575652286410332,
+ "learning_rate": 0.0006,
+ "loss": 5.6271772384643555,
+ "step": 467
+ },
+ {
+ "epoch": 6.503056768558952,
+ "grad_norm": 0.011672502383589745,
+ "learning_rate": 0.0006,
+ "loss": 5.761662006378174,
+ "step": 468
+ },
+ {
+ "epoch": 6.517030567685589,
+ "grad_norm": 0.011541751213371754,
+ "learning_rate": 0.0006,
+ "loss": 5.762078285217285,
+ "step": 469
+ },
+ {
+ "epoch": 6.531004366812227,
+ "grad_norm": 0.011396365240216255,
+ "learning_rate": 0.0006,
+ "loss": 5.67873477935791,
+ "step": 470
+ },
+ {
+ "epoch": 6.544978165938865,
+ "grad_norm": 0.010688499547541142,
+ "learning_rate": 0.0006,
+ "loss": 5.717051982879639,
+ "step": 471
+ },
+ {
+ "epoch": 6.558951965065502,
+ "grad_norm": 0.012224317528307438,
+ "learning_rate": 0.0006,
+ "loss": 5.707948684692383,
+ "step": 472
+ },
+ {
+ "epoch": 6.5729257641921395,
+ "grad_norm": 0.011856825090944767,
+ "learning_rate": 0.0006,
+ "loss": 5.70878791809082,
+ "step": 473
+ },
+ {
+ "epoch": 6.586899563318777,
+ "grad_norm": 0.01199064590036869,
+ "learning_rate": 0.0006,
+ "loss": 5.708697319030762,
+ "step": 474
+ },
+ {
+ "epoch": 6.600873362445415,
+ "grad_norm": 0.01219446212053299,
+ "learning_rate": 0.0006,
+ "loss": 5.6061577796936035,
+ "step": 475
+ },
+ {
+ "epoch": 6.614847161572053,
+ "grad_norm": 0.013397484086453915,
+ "learning_rate": 0.0006,
+ "loss": 5.624789714813232,
+ "step": 476
+ },
+ {
+ "epoch": 6.62882096069869,
+ "grad_norm": 0.01483136136084795,
+ "learning_rate": 0.0006,
+ "loss": 5.721141338348389,
+ "step": 477
+ },
+ {
+ "epoch": 6.642794759825327,
+ "grad_norm": 0.0194488987326622,
+ "learning_rate": 0.0006,
+ "loss": 5.604279518127441,
+ "step": 478
+ },
+ {
+ "epoch": 6.656768558951965,
+ "grad_norm": 0.02079896256327629,
+ "learning_rate": 0.0006,
+ "loss": 5.586322784423828,
+ "step": 479
+ },
+ {
+ "epoch": 6.670742358078603,
+ "grad_norm": 0.018996229395270348,
+ "learning_rate": 0.0006,
+ "loss": 5.5987772941589355,
+ "step": 480
+ },
+ {
+ "epoch": 6.68471615720524,
+ "grad_norm": 0.015779603272676468,
+ "learning_rate": 0.0006,
+ "loss": 5.588602066040039,
+ "step": 481
+ },
+ {
+ "epoch": 6.698689956331878,
+ "grad_norm": 0.015322973020374775,
+ "learning_rate": 0.0006,
+ "loss": 5.686467170715332,
+ "step": 482
+ },
+ {
+ "epoch": 6.712663755458515,
+ "grad_norm": 0.015282213687896729,
+ "learning_rate": 0.0006,
+ "loss": 5.664676666259766,
+ "step": 483
+ },
+ {
+ "epoch": 6.726637554585153,
+ "grad_norm": 0.016589272767305374,
+ "learning_rate": 0.0006,
+ "loss": 5.666739463806152,
+ "step": 484
+ },
+ {
+ "epoch": 6.74061135371179,
+ "grad_norm": 0.01811421848833561,
+ "learning_rate": 0.0006,
+ "loss": 5.647593021392822,
+ "step": 485
+ },
+ {
+ "epoch": 6.754585152838428,
+ "grad_norm": 0.019652029499411583,
+ "learning_rate": 0.0006,
+ "loss": 5.628726959228516,
+ "step": 486
+ },
+ {
+ "epoch": 6.7685589519650655,
+ "grad_norm": 0.02118665538728237,
+ "learning_rate": 0.0006,
+ "loss": 5.6580705642700195,
+ "step": 487
+ },
+ {
+ "epoch": 6.782532751091703,
+ "grad_norm": 0.02237832546234131,
+ "learning_rate": 0.0006,
+ "loss": 5.635931968688965,
+ "step": 488
+ },
+ {
+ "epoch": 6.796506550218341,
+ "grad_norm": 0.023897631093859673,
+ "learning_rate": 0.0006,
+ "loss": 5.517949104309082,
+ "step": 489
+ },
+ {
+ "epoch": 6.810480349344978,
+ "grad_norm": 0.02442227490246296,
+ "learning_rate": 0.0006,
+ "loss": 5.600021839141846,
+ "step": 490
+ },
+ {
+ "epoch": 6.824454148471616,
+ "grad_norm": 0.024675287306308746,
+ "learning_rate": 0.0006,
+ "loss": 5.6639909744262695,
+ "step": 491
+ },
+ {
+ "epoch": 6.8384279475982535,
+ "grad_norm": 0.030372392386198044,
+ "learning_rate": 0.0006,
+ "loss": 5.582888603210449,
+ "step": 492
+ },
+ {
+ "epoch": 6.85240174672489,
+ "grad_norm": 0.03237690031528473,
+ "learning_rate": 0.0006,
+ "loss": 5.573896884918213,
+ "step": 493
+ },
+ {
+ "epoch": 6.866375545851528,
+ "grad_norm": 0.03511589393019676,
+ "learning_rate": 0.0006,
+ "loss": 5.565878868103027,
+ "step": 494
+ },
+ {
+ "epoch": 6.880349344978166,
+ "grad_norm": 0.03883938118815422,
+ "learning_rate": 0.0006,
+ "loss": 5.58540678024292,
+ "step": 495
+ },
+ {
+ "epoch": 6.894323144104804,
+ "grad_norm": 0.04175502806901932,
+ "learning_rate": 0.0006,
+ "loss": 5.6095194816589355,
+ "step": 496
+ },
+ {
+ "epoch": 6.908296943231441,
+ "grad_norm": 0.04013441503047943,
+ "learning_rate": 0.0006,
+ "loss": 5.654401779174805,
+ "step": 497
+ },
+ {
+ "epoch": 6.922270742358078,
+ "grad_norm": 0.04618770629167557,
+ "learning_rate": 0.0006,
+ "loss": 5.6483473777771,
+ "step": 498
+ },
+ {
+ "epoch": 6.936244541484716,
+ "grad_norm": 0.04507308453321457,
+ "learning_rate": 0.0006,
+ "loss": 5.5856218338012695,
+ "step": 499
+ },
+ {
+ "epoch": 6.950218340611354,
+ "grad_norm": 0.03151383996009827,
+ "learning_rate": 0.0006,
+ "loss": 5.594086647033691,
+ "step": 500
+ },
+ {
+ "epoch": 6.964192139737992,
+ "grad_norm": 0.026883797720074654,
+ "learning_rate": 0.0006,
+ "loss": 5.607676029205322,
+ "step": 501
+ },
+ {
+ "epoch": 6.978165938864628,
+ "grad_norm": 0.02981836162507534,
+ "learning_rate": 0.0006,
+ "loss": 5.563666820526123,
+ "step": 502
+ },
+ {
+ "epoch": 6.992139737991266,
+ "grad_norm": 0.03233477845788002,
+ "learning_rate": 0.0006,
+ "loss": 5.667543411254883,
+ "step": 503
+ },
+ {
+ "epoch": 7.0,
+ "grad_norm": 0.029927456751465797,
+ "learning_rate": 0.0006,
+ "loss": 5.733482837677002,
+ "step": 504
+ },
+ {
+ "epoch": 7.0,
+ "eval_loss": 5.679076671600342,
+ "eval_runtime": 59.5965,
+ "eval_samples_per_second": 40.976,
+ "eval_steps_per_second": 1.292,
+ "step": 504
+ },
+ {
+ "epoch": 7.013973799126638,
+ "grad_norm": 0.027081597596406937,
+ "learning_rate": 0.0006,
+ "loss": 5.634487152099609,
+ "step": 505
+ },
+ {
+ "epoch": 7.0279475982532755,
+ "grad_norm": 0.031959421932697296,
+ "learning_rate": 0.0006,
+ "loss": 5.643294811248779,
+ "step": 506
+ },
+ {
+ "epoch": 7.041921397379912,
+ "grad_norm": 0.02802063524723053,
+ "learning_rate": 0.0006,
+ "loss": 5.573239326477051,
+ "step": 507
+ },
+ {
+ "epoch": 7.05589519650655,
+ "grad_norm": 0.032579705119132996,
+ "learning_rate": 0.0006,
+ "loss": 5.627150058746338,
+ "step": 508
+ },
+ {
+ "epoch": 7.069868995633188,
+ "grad_norm": 0.02503928542137146,
+ "learning_rate": 0.0006,
+ "loss": 5.6045308113098145,
+ "step": 509
+ },
+ {
+ "epoch": 7.083842794759826,
+ "grad_norm": 0.02421317622065544,
+ "learning_rate": 0.0006,
+ "loss": 5.522153377532959,
+ "step": 510
+ },
+ {
+ "epoch": 7.097816593886463,
+ "grad_norm": 0.021983426064252853,
+ "learning_rate": 0.0006,
+ "loss": 5.632939338684082,
+ "step": 511
+ },
+ {
+ "epoch": 7.1117903930131,
+ "grad_norm": 0.021933183073997498,
+ "learning_rate": 0.0006,
+ "loss": 5.592185974121094,
+ "step": 512
+ },
+ {
+ "epoch": 7.125764192139738,
+ "grad_norm": 0.02287031151354313,
+ "learning_rate": 0.0006,
+ "loss": 5.613700866699219,
+ "step": 513
+ },
+ {
+ "epoch": 7.139737991266376,
+ "grad_norm": 0.02281602844595909,
+ "learning_rate": 0.0006,
+ "loss": 5.551383018493652,
+ "step": 514
+ },
+ {
+ "epoch": 7.153711790393013,
+ "grad_norm": 0.021050360053777695,
+ "learning_rate": 0.0006,
+ "loss": 5.55275821685791,
+ "step": 515
+ },
+ {
+ "epoch": 7.1676855895196505,
+ "grad_norm": 0.018299926072359085,
+ "learning_rate": 0.0006,
+ "loss": 5.528225421905518,
+ "step": 516
+ },
+ {
+ "epoch": 7.181659388646288,
+ "grad_norm": 0.02024853602051735,
+ "learning_rate": 0.0006,
+ "loss": 5.522453784942627,
+ "step": 517
+ },
+ {
+ "epoch": 7.195633187772926,
+ "grad_norm": 0.02135239914059639,
+ "learning_rate": 0.0006,
+ "loss": 5.527522563934326,
+ "step": 518
+ },
+ {
+ "epoch": 7.209606986899563,
+ "grad_norm": 0.023394184187054634,
+ "learning_rate": 0.0006,
+ "loss": 5.4535017013549805,
+ "step": 519
+ },
+ {
+ "epoch": 7.223580786026201,
+ "grad_norm": 0.025952080264687538,
+ "learning_rate": 0.0006,
+ "loss": 5.5374674797058105,
+ "step": 520
+ },
+ {
+ "epoch": 7.2375545851528384,
+ "grad_norm": 0.022287718951702118,
+ "learning_rate": 0.0006,
+ "loss": 5.493753433227539,
+ "step": 521
+ },
+ {
+ "epoch": 7.251528384279476,
+ "grad_norm": 0.02018447406589985,
+ "learning_rate": 0.0006,
+ "loss": 5.510575294494629,
+ "step": 522
+ },
+ {
+ "epoch": 7.265502183406113,
+ "grad_norm": 0.02050507813692093,
+ "learning_rate": 0.0006,
+ "loss": 5.466026782989502,
+ "step": 523
+ },
+ {
+ "epoch": 7.279475982532751,
+ "grad_norm": 0.023688802495598793,
+ "learning_rate": 0.0006,
+ "loss": 5.485815525054932,
+ "step": 524
+ },
+ {
+ "epoch": 7.293449781659389,
+ "grad_norm": 0.021078843623399734,
+ "learning_rate": 0.0006,
+ "loss": 5.415581703186035,
+ "step": 525
+ },
+ {
+ "epoch": 7.307423580786026,
+ "grad_norm": 0.018266011029481888,
+ "learning_rate": 0.0006,
+ "loss": 5.402815818786621,
+ "step": 526
+ },
+ {
+ "epoch": 7.321397379912664,
+ "grad_norm": 0.019693493843078613,
+ "learning_rate": 0.0006,
+ "loss": 5.504674911499023,
+ "step": 527
+ },
+ {
+ "epoch": 7.335371179039301,
+ "grad_norm": 0.02031373605132103,
+ "learning_rate": 0.0006,
+ "loss": 5.503837585449219,
+ "step": 528
+ },
+ {
+ "epoch": 7.349344978165939,
+ "grad_norm": 0.017636626958847046,
+ "learning_rate": 0.0006,
+ "loss": 5.530580997467041,
+ "step": 529
+ },
+ {
+ "epoch": 7.3633187772925766,
+ "grad_norm": 0.01787244901061058,
+ "learning_rate": 0.0006,
+ "loss": 5.450218677520752,
+ "step": 530
+ },
+ {
+ "epoch": 7.377292576419214,
+ "grad_norm": 0.0170669574290514,
+ "learning_rate": 0.0006,
+ "loss": 5.481570243835449,
+ "step": 531
+ },
+ {
+ "epoch": 7.391266375545851,
+ "grad_norm": 0.01802165061235428,
+ "learning_rate": 0.0006,
+ "loss": 5.386394500732422,
+ "step": 532
+ },
+ {
+ "epoch": 7.405240174672489,
+ "grad_norm": 0.022949472069740295,
+ "learning_rate": 0.0006,
+ "loss": 5.424929618835449,
+ "step": 533
+ },
+ {
+ "epoch": 7.419213973799127,
+ "grad_norm": 0.034213390201330185,
+ "learning_rate": 0.0006,
+ "loss": 5.346663475036621,
+ "step": 534
+ },
+ {
+ "epoch": 7.4331877729257645,
+ "grad_norm": 0.03847593814134598,
+ "learning_rate": 0.0006,
+ "loss": 5.418482780456543,
+ "step": 535
+ },
+ {
+ "epoch": 7.447161572052401,
+ "grad_norm": 0.029393676668405533,
+ "learning_rate": 0.0006,
+ "loss": 5.456090927124023,
+ "step": 536
+ },
+ {
+ "epoch": 7.461135371179039,
+ "grad_norm": 0.03388667106628418,
+ "learning_rate": 0.0006,
+ "loss": 5.572983741760254,
+ "step": 537
+ },
+ {
+ "epoch": 7.475109170305677,
+ "grad_norm": 0.039690613746643066,
+ "learning_rate": 0.0006,
+ "loss": 5.414067268371582,
+ "step": 538
+ },
+ {
+ "epoch": 7.489082969432315,
+ "grad_norm": 0.03630776330828667,
+ "learning_rate": 0.0006,
+ "loss": 5.523739814758301,
+ "step": 539
+ },
+ {
+ "epoch": 7.503056768558952,
+ "grad_norm": 0.03356870263814926,
+ "learning_rate": 0.0006,
+ "loss": 5.444767951965332,
+ "step": 540
+ },
+ {
+ "epoch": 7.517030567685589,
+ "grad_norm": 0.030812304466962814,
+ "learning_rate": 0.0006,
+ "loss": 5.461244106292725,
+ "step": 541
+ },
+ {
+ "epoch": 7.531004366812227,
+ "grad_norm": 0.03319217637181282,
+ "learning_rate": 0.0006,
+ "loss": 5.502161026000977,
+ "step": 542
+ },
+ {
+ "epoch": 7.544978165938865,
+ "grad_norm": 0.032997481524944305,
+ "learning_rate": 0.0006,
+ "loss": 5.4162278175354,
+ "step": 543
+ },
+ {
+ "epoch": 7.558951965065502,
+ "grad_norm": 0.03364962339401245,
+ "learning_rate": 0.0006,
+ "loss": 5.415736198425293,
+ "step": 544
+ },
+ {
+ "epoch": 7.5729257641921395,
+ "grad_norm": 0.0344221405684948,
+ "learning_rate": 0.0006,
+ "loss": 5.541967868804932,
+ "step": 545
+ },
+ {
+ "epoch": 7.586899563318777,
+ "grad_norm": 0.029609503224492073,
+ "learning_rate": 0.0006,
+ "loss": 5.440614223480225,
+ "step": 546
+ },
+ {
+ "epoch": 7.600873362445415,
+ "grad_norm": 0.029055926948785782,
+ "learning_rate": 0.0006,
+ "loss": 5.462865352630615,
+ "step": 547
+ },
+ {
+ "epoch": 7.614847161572053,
+ "grad_norm": 0.02658848837018013,
+ "learning_rate": 0.0006,
+ "loss": 5.444467544555664,
+ "step": 548
+ },
+ {
+ "epoch": 7.62882096069869,
+ "grad_norm": 0.026276404038071632,
+ "learning_rate": 0.0006,
+ "loss": 5.462278842926025,
+ "step": 549
+ },
+ {
+ "epoch": 7.642794759825327,
+ "grad_norm": 0.0286889486014843,
+ "learning_rate": 0.0006,
+ "loss": 5.4849534034729,
+ "step": 550
+ },
+ {
+ "epoch": 7.656768558951965,
+ "grad_norm": 0.026849817484617233,
+ "learning_rate": 0.0006,
+ "loss": 5.431473731994629,
+ "step": 551
+ },
+ {
+ "epoch": 7.670742358078603,
+ "grad_norm": 0.02312396466732025,
+ "learning_rate": 0.0006,
+ "loss": 5.45462703704834,
+ "step": 552
+ },
+ {
+ "epoch": 7.68471615720524,
+ "grad_norm": 0.026197485625743866,
+ "learning_rate": 0.0006,
+ "loss": 5.452552318572998,
+ "step": 553
+ },
+ {
+ "epoch": 7.698689956331878,
+ "grad_norm": 0.02747255191206932,
+ "learning_rate": 0.0006,
+ "loss": 5.447169780731201,
+ "step": 554
+ },
+ {
+ "epoch": 7.712663755458515,
+ "grad_norm": 0.028123432770371437,
+ "learning_rate": 0.0006,
+ "loss": 5.394400596618652,
+ "step": 555
+ },
+ {
+ "epoch": 7.726637554585153,
+ "grad_norm": 0.02599870041012764,
+ "learning_rate": 0.0006,
+ "loss": 5.398341178894043,
+ "step": 556
+ },
+ {
+ "epoch": 7.74061135371179,
+ "grad_norm": 0.022171657532453537,
+ "learning_rate": 0.0006,
+ "loss": 5.368820667266846,
+ "step": 557
+ },
+ {
+ "epoch": 7.754585152838428,
+ "grad_norm": 0.022309480234980583,
+ "learning_rate": 0.0006,
+ "loss": 5.416772365570068,
+ "step": 558
+ },
+ {
+ "epoch": 7.7685589519650655,
+ "grad_norm": 0.024074165150523186,
+ "learning_rate": 0.0006,
+ "loss": 5.359002113342285,
+ "step": 559
+ },
+ {
+ "epoch": 7.782532751091703,
+ "grad_norm": 0.02636653557419777,
+ "learning_rate": 0.0006,
+ "loss": 5.484368324279785,
+ "step": 560
+ },
+ {
+ "epoch": 7.796506550218341,
+ "grad_norm": 0.02203752100467682,
+ "learning_rate": 0.0006,
+ "loss": 5.375498294830322,
+ "step": 561
+ },
+ {
+ "epoch": 7.810480349344978,
+ "grad_norm": 0.019924819469451904,
+ "learning_rate": 0.0006,
+ "loss": 5.334723472595215,
+ "step": 562
+ },
+ {
+ "epoch": 7.824454148471616,
+ "grad_norm": 0.018755966797471046,
+ "learning_rate": 0.0006,
+ "loss": 5.348781108856201,
+ "step": 563
+ },
+ {
+ "epoch": 7.8384279475982535,
+ "grad_norm": 0.016747845336794853,
+ "learning_rate": 0.0006,
+ "loss": 5.433079719543457,
+ "step": 564
+ },
+ {
+ "epoch": 7.85240174672489,
+ "grad_norm": 0.0170395877212286,
+ "learning_rate": 0.0006,
+ "loss": 5.346794128417969,
+ "step": 565
+ },
+ {
+ "epoch": 7.866375545851528,
+ "grad_norm": 0.016074176877737045,
+ "learning_rate": 0.0006,
+ "loss": 5.3277740478515625,
+ "step": 566
+ },
+ {
+ "epoch": 7.880349344978166,
+ "grad_norm": 0.014119806699454784,
+ "learning_rate": 0.0006,
+ "loss": 5.483551979064941,
+ "step": 567
+ },
+ {
+ "epoch": 7.894323144104804,
+ "grad_norm": 0.015270394273102283,
+ "learning_rate": 0.0006,
+ "loss": 5.378519535064697,
+ "step": 568
+ },
+ {
+ "epoch": 7.908296943231441,
+ "grad_norm": 0.015677539631724358,
+ "learning_rate": 0.0006,
+ "loss": 5.290210247039795,
+ "step": 569
+ },
+ {
+ "epoch": 7.922270742358078,
+ "grad_norm": 0.015930423513054848,
+ "learning_rate": 0.0006,
+ "loss": 5.427360534667969,
+ "step": 570
+ },
+ {
+ "epoch": 7.936244541484716,
+ "grad_norm": 0.016641564667224884,
+ "learning_rate": 0.0006,
+ "loss": 5.301599979400635,
+ "step": 571
+ },
+ {
+ "epoch": 7.950218340611354,
+ "grad_norm": 0.018293552100658417,
+ "learning_rate": 0.0006,
+ "loss": 5.277889251708984,
+ "step": 572
+ },
+ {
+ "epoch": 7.964192139737992,
+ "grad_norm": 0.017618799582123756,
+ "learning_rate": 0.0006,
+ "loss": 5.297194480895996,
+ "step": 573
+ },
+ {
+ "epoch": 7.978165938864628,
+ "grad_norm": 0.015034242533147335,
+ "learning_rate": 0.0006,
+ "loss": 5.42428731918335,
+ "step": 574
+ },
+ {
+ "epoch": 7.992139737991266,
+ "grad_norm": 0.01617511734366417,
+ "learning_rate": 0.0006,
+ "loss": 5.420044898986816,
+ "step": 575
+ },
+ {
+ "epoch": 8.0,
+ "grad_norm": 0.01837257295846939,
+ "learning_rate": 0.0006,
+ "loss": 5.240387916564941,
+ "step": 576
+ },
+ {
+ "epoch": 8.0,
+ "eval_loss": 5.423932075500488,
+ "eval_runtime": 58.0841,
+ "eval_samples_per_second": 42.042,
+ "eval_steps_per_second": 1.326,
+ "step": 576
+ },
+ {
+ "epoch": 8.013973799126637,
+ "grad_norm": 0.020953809842467308,
+ "learning_rate": 0.0006,
+ "loss": 5.4320783615112305,
+ "step": 577
+ },
+ {
+ "epoch": 8.027947598253276,
+ "grad_norm": 0.0227745920419693,
+ "learning_rate": 0.0006,
+ "loss": 5.310197830200195,
+ "step": 578
+ },
+ {
+ "epoch": 8.041921397379912,
+ "grad_norm": 0.021372603252530098,
+ "learning_rate": 0.0006,
+ "loss": 5.339468002319336,
+ "step": 579
+ },
+ {
+ "epoch": 8.055895196506551,
+ "grad_norm": 0.022401731461286545,
+ "learning_rate": 0.0006,
+ "loss": 5.279829978942871,
+ "step": 580
+ },
+ {
+ "epoch": 8.069868995633188,
+ "grad_norm": 0.023942379280924797,
+ "learning_rate": 0.0006,
+ "loss": 5.372323989868164,
+ "step": 581
+ },
+ {
+ "epoch": 8.083842794759825,
+ "grad_norm": 0.024380534887313843,
+ "learning_rate": 0.0006,
+ "loss": 5.253486633300781,
+ "step": 582
+ },
+ {
+ "epoch": 8.097816593886463,
+ "grad_norm": 0.03246625140309334,
+ "learning_rate": 0.0006,
+ "loss": 5.305774211883545,
+ "step": 583
+ },
+ {
+ "epoch": 8.1117903930131,
+ "grad_norm": 0.030399736016988754,
+ "learning_rate": 0.0006,
+ "loss": 5.370976448059082,
+ "step": 584
+ },
+ {
+ "epoch": 8.125764192139737,
+ "grad_norm": 0.027148913592100143,
+ "learning_rate": 0.0006,
+ "loss": 5.368010520935059,
+ "step": 585
+ },
+ {
+ "epoch": 8.139737991266376,
+ "grad_norm": 0.02925540879368782,
+ "learning_rate": 0.0006,
+ "loss": 5.303013801574707,
+ "step": 586
+ },
+ {
+ "epoch": 8.153711790393013,
+ "grad_norm": 0.02915453538298607,
+ "learning_rate": 0.0006,
+ "loss": 5.329667091369629,
+ "step": 587
+ },
+ {
+ "epoch": 8.167685589519651,
+ "grad_norm": 0.031732626259326935,
+ "learning_rate": 0.0006,
+ "loss": 5.3634934425354,
+ "step": 588
+ },
+ {
+ "epoch": 8.181659388646288,
+ "grad_norm": 0.036166731268167496,
+ "learning_rate": 0.0006,
+ "loss": 5.223980903625488,
+ "step": 589
+ },
+ {
+ "epoch": 8.195633187772925,
+ "grad_norm": 0.035654108971357346,
+ "learning_rate": 0.0006,
+ "loss": 5.361059188842773,
+ "step": 590
+ },
+ {
+ "epoch": 8.209606986899564,
+ "grad_norm": 0.0365324430167675,
+ "learning_rate": 0.0006,
+ "loss": 5.336151599884033,
+ "step": 591
+ },
+ {
+ "epoch": 8.2235807860262,
+ "grad_norm": 0.03578880429267883,
+ "learning_rate": 0.0006,
+ "loss": 5.434076309204102,
+ "step": 592
+ },
+ {
+ "epoch": 8.237554585152838,
+ "grad_norm": 0.035734012722969055,
+ "learning_rate": 0.0006,
+ "loss": 5.330893516540527,
+ "step": 593
+ },
+ {
+ "epoch": 8.251528384279476,
+ "grad_norm": 0.03196857124567032,
+ "learning_rate": 0.0006,
+ "loss": 5.306009769439697,
+ "step": 594
+ },
+ {
+ "epoch": 8.265502183406113,
+ "grad_norm": 0.032042086124420166,
+ "learning_rate": 0.0006,
+ "loss": 5.343267917633057,
+ "step": 595
+ },
+ {
+ "epoch": 8.279475982532752,
+ "grad_norm": 0.03160746395587921,
+ "learning_rate": 0.0006,
+ "loss": 5.2353715896606445,
+ "step": 596
+ },
+ {
+ "epoch": 8.293449781659389,
+ "grad_norm": 0.026689818128943443,
+ "learning_rate": 0.0006,
+ "loss": 5.254042625427246,
+ "step": 597
+ },
+ {
+ "epoch": 8.307423580786025,
+ "grad_norm": 0.02880188450217247,
+ "learning_rate": 0.0006,
+ "loss": 5.326833248138428,
+ "step": 598
+ },
+ {
+ "epoch": 8.321397379912664,
+ "grad_norm": 0.027516381815075874,
+ "learning_rate": 0.0006,
+ "loss": 5.342336177825928,
+ "step": 599
+ },
+ {
+ "epoch": 8.335371179039301,
+ "grad_norm": 0.027875030413269997,
+ "learning_rate": 0.0006,
+ "loss": 5.200719833374023,
+ "step": 600
+ },
+ {
+ "epoch": 8.34934497816594,
+ "grad_norm": 0.0268265288323164,
+ "learning_rate": 0.0006,
+ "loss": 5.249449729919434,
+ "step": 601
+ },
+ {
+ "epoch": 8.363318777292577,
+ "grad_norm": 0.024824608117341995,
+ "learning_rate": 0.0006,
+ "loss": 5.318952560424805,
+ "step": 602
+ },
+ {
+ "epoch": 8.377292576419213,
+ "grad_norm": 0.022990627214312553,
+ "learning_rate": 0.0006,
+ "loss": 5.244993209838867,
+ "step": 603
+ },
+ {
+ "epoch": 8.391266375545852,
+ "grad_norm": 0.022804604843258858,
+ "learning_rate": 0.0006,
+ "loss": 5.3236846923828125,
+ "step": 604
+ },
+ {
+ "epoch": 8.405240174672489,
+ "grad_norm": 0.02419872209429741,
+ "learning_rate": 0.0006,
+ "loss": 5.255486488342285,
+ "step": 605
+ },
+ {
+ "epoch": 8.419213973799126,
+ "grad_norm": 0.021952059119939804,
+ "learning_rate": 0.0006,
+ "loss": 5.3511552810668945,
+ "step": 606
+ },
+ {
+ "epoch": 8.433187772925764,
+ "grad_norm": 0.022375497967004776,
+ "learning_rate": 0.0006,
+ "loss": 5.294790267944336,
+ "step": 607
+ },
+ {
+ "epoch": 8.447161572052401,
+ "grad_norm": 0.019242815673351288,
+ "learning_rate": 0.0006,
+ "loss": 5.252618789672852,
+ "step": 608
+ },
+ {
+ "epoch": 8.46113537117904,
+ "grad_norm": 0.0173486340790987,
+ "learning_rate": 0.0006,
+ "loss": 5.30147647857666,
+ "step": 609
+ },
+ {
+ "epoch": 8.475109170305677,
+ "grad_norm": 0.01612604409456253,
+ "learning_rate": 0.0006,
+ "loss": 5.23960542678833,
+ "step": 610
+ },
+ {
+ "epoch": 8.489082969432314,
+ "grad_norm": 0.014293976128101349,
+ "learning_rate": 0.0006,
+ "loss": 5.119932651519775,
+ "step": 611
+ },
+ {
+ "epoch": 8.503056768558952,
+ "grad_norm": 0.014453536830842495,
+ "learning_rate": 0.0006,
+ "loss": 5.166794776916504,
+ "step": 612
+ },
+ {
+ "epoch": 8.51703056768559,
+ "grad_norm": 0.014685769565403461,
+ "learning_rate": 0.0006,
+ "loss": 5.264822959899902,
+ "step": 613
+ },
+ {
+ "epoch": 8.531004366812226,
+ "grad_norm": 0.01427740603685379,
+ "learning_rate": 0.0006,
+ "loss": 5.278433799743652,
+ "step": 614
+ },
+ {
+ "epoch": 8.544978165938865,
+ "grad_norm": 0.014864951372146606,
+ "learning_rate": 0.0006,
+ "loss": 5.214512348175049,
+ "step": 615
+ },
+ {
+ "epoch": 8.558951965065502,
+ "grad_norm": 0.01565164513885975,
+ "learning_rate": 0.0006,
+ "loss": 5.183066368103027,
+ "step": 616
+ },
+ {
+ "epoch": 8.57292576419214,
+ "grad_norm": 0.01779816672205925,
+ "learning_rate": 0.0006,
+ "loss": 5.163185119628906,
+ "step": 617
+ },
+ {
+ "epoch": 8.586899563318777,
+ "grad_norm": 0.017254827544093132,
+ "learning_rate": 0.0006,
+ "loss": 5.243555068969727,
+ "step": 618
+ },
+ {
+ "epoch": 8.600873362445414,
+ "grad_norm": 0.01650584116578102,
+ "learning_rate": 0.0006,
+ "loss": 5.234884738922119,
+ "step": 619
+ },
+ {
+ "epoch": 8.614847161572053,
+ "grad_norm": 0.017021115869283676,
+ "learning_rate": 0.0006,
+ "loss": 5.12271785736084,
+ "step": 620
+ },
+ {
+ "epoch": 8.62882096069869,
+ "grad_norm": 0.01773759163916111,
+ "learning_rate": 0.0006,
+ "loss": 5.242153167724609,
+ "step": 621
+ },
+ {
+ "epoch": 8.642794759825328,
+ "grad_norm": 0.015679042786359787,
+ "learning_rate": 0.0006,
+ "loss": 5.170779228210449,
+ "step": 622
+ },
+ {
+ "epoch": 8.656768558951965,
+ "grad_norm": 0.013760549947619438,
+ "learning_rate": 0.0006,
+ "loss": 5.238644599914551,
+ "step": 623
+ },
+ {
+ "epoch": 8.670742358078602,
+ "grad_norm": 0.014571522362530231,
+ "learning_rate": 0.0006,
+ "loss": 5.177056312561035,
+ "step": 624
+ },
+ {
+ "epoch": 8.68471615720524,
+ "grad_norm": 0.016209015622735023,
+ "learning_rate": 0.0006,
+ "loss": 5.163750648498535,
+ "step": 625
+ },
+ {
+ "epoch": 8.698689956331878,
+ "grad_norm": 0.016813941299915314,
+ "learning_rate": 0.0006,
+ "loss": 5.183428764343262,
+ "step": 626
+ },
+ {
+ "epoch": 8.712663755458514,
+ "grad_norm": 0.020985357463359833,
+ "learning_rate": 0.0006,
+ "loss": 5.229465007781982,
+ "step": 627
+ },
+ {
+ "epoch": 8.726637554585153,
+ "grad_norm": 0.02679632417857647,
+ "learning_rate": 0.0006,
+ "loss": 5.257368087768555,
+ "step": 628
+ },
+ {
+ "epoch": 8.74061135371179,
+ "grad_norm": 0.025756070390343666,
+ "learning_rate": 0.0006,
+ "loss": 5.253736972808838,
+ "step": 629
+ },
+ {
+ "epoch": 8.754585152838429,
+ "grad_norm": 0.02643490768969059,
+ "learning_rate": 0.0006,
+ "loss": 5.228633403778076,
+ "step": 630
+ },
+ {
+ "epoch": 8.768558951965066,
+ "grad_norm": 0.03303210437297821,
+ "learning_rate": 0.0006,
+ "loss": 5.163540363311768,
+ "step": 631
+ },
+ {
+ "epoch": 8.782532751091702,
+ "grad_norm": 0.030432431027293205,
+ "learning_rate": 0.0006,
+ "loss": 5.243169784545898,
+ "step": 632
+ },
+ {
+ "epoch": 8.796506550218341,
+ "grad_norm": 0.028631288558244705,
+ "learning_rate": 0.0006,
+ "loss": 5.166398048400879,
+ "step": 633
+ },
+ {
+ "epoch": 8.810480349344978,
+ "grad_norm": 0.026188310235738754,
+ "learning_rate": 0.0006,
+ "loss": 5.128122329711914,
+ "step": 634
+ },
+ {
+ "epoch": 8.824454148471617,
+ "grad_norm": 0.025308528915047646,
+ "learning_rate": 0.0006,
+ "loss": 5.211284637451172,
+ "step": 635
+ },
+ {
+ "epoch": 8.838427947598253,
+ "grad_norm": 0.020657729357481003,
+ "learning_rate": 0.0006,
+ "loss": 5.161575794219971,
+ "step": 636
+ },
+ {
+ "epoch": 8.85240174672489,
+ "grad_norm": 0.021901234984397888,
+ "learning_rate": 0.0006,
+ "loss": 5.20050048828125,
+ "step": 637
+ },
+ {
+ "epoch": 8.866375545851529,
+ "grad_norm": 0.021219318732619286,
+ "learning_rate": 0.0006,
+ "loss": 5.211699485778809,
+ "step": 638
+ },
+ {
+ "epoch": 8.880349344978166,
+ "grad_norm": 0.0200974028557539,
+ "learning_rate": 0.0006,
+ "loss": 5.092503547668457,
+ "step": 639
+ },
+ {
+ "epoch": 8.894323144104803,
+ "grad_norm": 0.023804882541298866,
+ "learning_rate": 0.0006,
+ "loss": 5.216068267822266,
+ "step": 640
+ },
+ {
+ "epoch": 8.908296943231441,
+ "grad_norm": 0.026088010519742966,
+ "learning_rate": 0.0006,
+ "loss": 5.155592918395996,
+ "step": 641
+ },
+ {
+ "epoch": 8.922270742358078,
+ "grad_norm": 0.02601276896893978,
+ "learning_rate": 0.0006,
+ "loss": 5.197238922119141,
+ "step": 642
+ },
+ {
+ "epoch": 8.936244541484717,
+ "grad_norm": 0.020387211814522743,
+ "learning_rate": 0.0006,
+ "loss": 5.1955437660217285,
+ "step": 643
+ },
+ {
+ "epoch": 8.950218340611354,
+ "grad_norm": 0.019214622676372528,
+ "learning_rate": 0.0006,
+ "loss": 5.206346035003662,
+ "step": 644
+ },
+ {
+ "epoch": 8.96419213973799,
+ "grad_norm": 0.019674314185976982,
+ "learning_rate": 0.0006,
+ "loss": 5.219857215881348,
+ "step": 645
+ },
+ {
+ "epoch": 8.97816593886463,
+ "grad_norm": 0.020318234339356422,
+ "learning_rate": 0.0006,
+ "loss": 5.191132545471191,
+ "step": 646
+ },
+ {
+ "epoch": 8.992139737991266,
+ "grad_norm": 0.021428676322102547,
+ "learning_rate": 0.0006,
+ "loss": 5.214695930480957,
+ "step": 647
+ },
+ {
+ "epoch": 9.0,
+ "grad_norm": 0.023235054686665535,
+ "learning_rate": 0.0006,
+ "loss": 5.229091644287109,
+ "step": 648
+ },
+ {
+ "epoch": 9.0,
+ "eval_loss": 5.289593696594238,
+ "eval_runtime": 56.9256,
+ "eval_samples_per_second": 42.898,
+ "eval_steps_per_second": 1.353,
+ "step": 648
+ },
+ {
+ "epoch": 9.013973799126637,
+ "grad_norm": 0.02320289984345436,
+ "learning_rate": 0.0006,
+ "loss": 4.983669281005859,
+ "step": 649
+ },
+ {
+ "epoch": 9.027947598253276,
+ "grad_norm": 0.018521282821893692,
+ "learning_rate": 0.0006,
+ "loss": 5.200575828552246,
+ "step": 650
+ },
+ {
+ "epoch": 9.041921397379912,
+ "grad_norm": 0.0180169939994812,
+ "learning_rate": 0.0006,
+ "loss": 5.148033618927002,
+ "step": 651
+ },
+ {
+ "epoch": 9.055895196506551,
+ "grad_norm": 0.019969860091805458,
+ "learning_rate": 0.0006,
+ "loss": 5.099125385284424,
+ "step": 652
+ },
+ {
+ "epoch": 9.069868995633188,
+ "grad_norm": 0.01643305830657482,
+ "learning_rate": 0.0006,
+ "loss": 5.092347145080566,
+ "step": 653
+ },
+ {
+ "epoch": 9.083842794759825,
+ "grad_norm": 0.016022363677620888,
+ "learning_rate": 0.0006,
+ "loss": 5.129279613494873,
+ "step": 654
+ },
+ {
+ "epoch": 9.097816593886463,
+ "grad_norm": 0.01657041162252426,
+ "learning_rate": 0.0006,
+ "loss": 5.159562587738037,
+ "step": 655
+ },
+ {
+ "epoch": 9.1117903930131,
+ "grad_norm": 0.017924228683114052,
+ "learning_rate": 0.0006,
+ "loss": 5.165497779846191,
+ "step": 656
+ },
+ {
+ "epoch": 9.125764192139737,
+ "grad_norm": 0.021592361852526665,
+ "learning_rate": 0.0006,
+ "loss": 5.204775810241699,
+ "step": 657
+ },
+ {
+ "epoch": 9.139737991266376,
+ "grad_norm": 0.026924120262265205,
+ "learning_rate": 0.0006,
+ "loss": 5.128296852111816,
+ "step": 658
+ },
+ {
+ "epoch": 9.153711790393013,
+ "grad_norm": 0.023156899958848953,
+ "learning_rate": 0.0006,
+ "loss": 5.1255340576171875,
+ "step": 659
+ },
+ {
+ "epoch": 9.167685589519651,
+ "grad_norm": 0.020670367404818535,
+ "learning_rate": 0.0006,
+ "loss": 5.128479957580566,
+ "step": 660
+ },
+ {
+ "epoch": 9.181659388646288,
+ "grad_norm": 0.025447173044085503,
+ "learning_rate": 0.0006,
+ "loss": 5.154821872711182,
+ "step": 661
+ },
+ {
+ "epoch": 9.195633187772925,
+ "grad_norm": 0.027940068393945694,
+ "learning_rate": 0.0006,
+ "loss": 5.143466472625732,
+ "step": 662
+ },
+ {
+ "epoch": 9.209606986899564,
+ "grad_norm": 0.026847844943404198,
+ "learning_rate": 0.0006,
+ "loss": 5.045665740966797,
+ "step": 663
+ },
+ {
+ "epoch": 9.2235807860262,
+ "grad_norm": 0.02340601570904255,
+ "learning_rate": 0.0006,
+ "loss": 5.21933126449585,
+ "step": 664
+ },
+ {
+ "epoch": 9.237554585152838,
+ "grad_norm": 0.02340371161699295,
+ "learning_rate": 0.0006,
+ "loss": 4.942949295043945,
+ "step": 665
+ },
+ {
+ "epoch": 9.251528384279476,
+ "grad_norm": 0.02221992425620556,
+ "learning_rate": 0.0006,
+ "loss": 5.197099208831787,
+ "step": 666
+ },
+ {
+ "epoch": 9.265502183406113,
+ "grad_norm": 0.023508677259087563,
+ "learning_rate": 0.0006,
+ "loss": 5.022897720336914,
+ "step": 667
+ },
+ {
+ "epoch": 9.279475982532752,
+ "grad_norm": 0.026120394468307495,
+ "learning_rate": 0.0006,
+ "loss": 5.087725639343262,
+ "step": 668
+ },
+ {
+ "epoch": 9.293449781659389,
+ "grad_norm": 0.027273228392004967,
+ "learning_rate": 0.0006,
+ "loss": 5.090963840484619,
+ "step": 669
+ },
+ {
+ "epoch": 9.307423580786025,
+ "grad_norm": 0.03241586685180664,
+ "learning_rate": 0.0006,
+ "loss": 5.047842979431152,
+ "step": 670
+ },
+ {
+ "epoch": 9.321397379912664,
+ "grad_norm": 0.030245667323470116,
+ "learning_rate": 0.0006,
+ "loss": 5.104803085327148,
+ "step": 671
+ },
+ {
+ "epoch": 9.335371179039301,
+ "grad_norm": 0.027698364108800888,
+ "learning_rate": 0.0006,
+ "loss": 5.0644097328186035,
+ "step": 672
+ },
+ {
+ "epoch": 9.34934497816594,
+ "grad_norm": 0.029692554846405983,
+ "learning_rate": 0.0006,
+ "loss": 5.111942291259766,
+ "step": 673
+ },
+ {
+ "epoch": 9.363318777292577,
+ "grad_norm": 0.03328656405210495,
+ "learning_rate": 0.0006,
+ "loss": 5.1574554443359375,
+ "step": 674
+ },
+ {
+ "epoch": 9.377292576419213,
+ "grad_norm": 0.031938180327415466,
+ "learning_rate": 0.0006,
+ "loss": 5.190371990203857,
+ "step": 675
+ },
+ {
+ "epoch": 9.391266375545852,
+ "grad_norm": 0.024858945980668068,
+ "learning_rate": 0.0006,
+ "loss": 5.121420860290527,
+ "step": 676
+ },
+ {
+ "epoch": 9.405240174672489,
+ "grad_norm": 0.023033203557133675,
+ "learning_rate": 0.0006,
+ "loss": 5.0974345207214355,
+ "step": 677
+ },
+ {
+ "epoch": 9.419213973799126,
+ "grad_norm": 0.022083545103669167,
+ "learning_rate": 0.0006,
+ "loss": 5.046319961547852,
+ "step": 678
+ },
+ {
+ "epoch": 9.433187772925764,
+ "grad_norm": 0.0228437427431345,
+ "learning_rate": 0.0006,
+ "loss": 5.119963645935059,
+ "step": 679
+ },
+ {
+ "epoch": 9.447161572052401,
+ "grad_norm": 0.0242351982742548,
+ "learning_rate": 0.0006,
+ "loss": 5.0444560050964355,
+ "step": 680
+ },
+ {
+ "epoch": 9.46113537117904,
+ "grad_norm": 0.02401185780763626,
+ "learning_rate": 0.0006,
+ "loss": 5.063238620758057,
+ "step": 681
+ },
+ {
+ "epoch": 9.475109170305677,
+ "grad_norm": 0.02560959942638874,
+ "learning_rate": 0.0006,
+ "loss": 5.035224914550781,
+ "step": 682
+ },
+ {
+ "epoch": 9.489082969432314,
+ "grad_norm": 0.024985626339912415,
+ "learning_rate": 0.0006,
+ "loss": 5.049284934997559,
+ "step": 683
+ },
+ {
+ "epoch": 9.503056768558952,
+ "grad_norm": 0.023178115487098694,
+ "learning_rate": 0.0006,
+ "loss": 5.2266621589660645,
+ "step": 684
+ },
+ {
+ "epoch": 9.51703056768559,
+ "grad_norm": 0.0215512253344059,
+ "learning_rate": 0.0006,
+ "loss": 5.1875715255737305,
+ "step": 685
+ },
+ {
+ "epoch": 9.531004366812226,
+ "grad_norm": 0.01662837713956833,
+ "learning_rate": 0.0006,
+ "loss": 5.137012481689453,
+ "step": 686
+ },
+ {
+ "epoch": 9.544978165938865,
+ "grad_norm": 0.01952764391899109,
+ "learning_rate": 0.0006,
+ "loss": 5.072500228881836,
+ "step": 687
+ },
+ {
+ "epoch": 9.558951965065502,
+ "grad_norm": 0.017521077767014503,
+ "learning_rate": 0.0006,
+ "loss": 5.045711994171143,
+ "step": 688
+ },
+ {
+ "epoch": 9.57292576419214,
+ "grad_norm": 0.017355265095829964,
+ "learning_rate": 0.0006,
+ "loss": 5.02881383895874,
+ "step": 689
+ },
+ {
+ "epoch": 9.586899563318777,
+ "grad_norm": 0.015587719157338142,
+ "learning_rate": 0.0006,
+ "loss": 5.132787704467773,
+ "step": 690
+ },
+ {
+ "epoch": 9.600873362445414,
+ "grad_norm": 0.017469845712184906,
+ "learning_rate": 0.0006,
+ "loss": 5.105447292327881,
+ "step": 691
+ },
+ {
+ "epoch": 9.614847161572053,
+ "grad_norm": 0.019630100578069687,
+ "learning_rate": 0.0006,
+ "loss": 5.032464981079102,
+ "step": 692
+ },
+ {
+ "epoch": 9.62882096069869,
+ "grad_norm": 0.01784994639456272,
+ "learning_rate": 0.0006,
+ "loss": 5.081345558166504,
+ "step": 693
+ },
+ {
+ "epoch": 9.642794759825328,
+ "grad_norm": 0.01872754842042923,
+ "learning_rate": 0.0006,
+ "loss": 4.9772114753723145,
+ "step": 694
+ },
+ {
+ "epoch": 9.656768558951965,
+ "grad_norm": 0.01999032311141491,
+ "learning_rate": 0.0006,
+ "loss": 5.089740753173828,
+ "step": 695
+ },
+ {
+ "epoch": 9.670742358078602,
+ "grad_norm": 0.01710767112672329,
+ "learning_rate": 0.0006,
+ "loss": 4.9869890213012695,
+ "step": 696
+ },
+ {
+ "epoch": 9.68471615720524,
+ "grad_norm": 0.015144161880016327,
+ "learning_rate": 0.0006,
+ "loss": 5.022564888000488,
+ "step": 697
+ },
+ {
+ "epoch": 9.698689956331878,
+ "grad_norm": 0.01618543453514576,
+ "learning_rate": 0.0006,
+ "loss": 5.035921573638916,
+ "step": 698
+ },
+ {
+ "epoch": 9.712663755458514,
+ "grad_norm": 0.014653387479484081,
+ "learning_rate": 0.0006,
+ "loss": 4.914456367492676,
+ "step": 699
+ },
+ {
+ "epoch": 9.726637554585153,
+ "grad_norm": 0.013434977270662785,
+ "learning_rate": 0.0006,
+ "loss": 5.063764572143555,
+ "step": 700
+ },
+ {
+ "epoch": 9.74061135371179,
+ "grad_norm": 0.015128864906728268,
+ "learning_rate": 0.0006,
+ "loss": 5.052367687225342,
+ "step": 701
+ },
+ {
+ "epoch": 9.754585152838429,
+ "grad_norm": 0.014210768043994904,
+ "learning_rate": 0.0006,
+ "loss": 5.016489028930664,
+ "step": 702
+ },
+ {
+ "epoch": 9.768558951965066,
+ "grad_norm": 0.015191680751740932,
+ "learning_rate": 0.0006,
+ "loss": 4.990176200866699,
+ "step": 703
+ },
+ {
+ "epoch": 9.782532751091702,
+ "grad_norm": 0.018285127356648445,
+ "learning_rate": 0.0006,
+ "loss": 5.015164375305176,
+ "step": 704
+ },
+ {
+ "epoch": 9.796506550218341,
+ "grad_norm": 0.0204079058021307,
+ "learning_rate": 0.0006,
+ "loss": 5.091142654418945,
+ "step": 705
+ },
+ {
+ "epoch": 9.810480349344978,
+ "grad_norm": 0.01980586163699627,
+ "learning_rate": 0.0006,
+ "loss": 5.03350830078125,
+ "step": 706
+ },
+ {
+ "epoch": 9.824454148471617,
+ "grad_norm": 0.02048729732632637,
+ "learning_rate": 0.0006,
+ "loss": 5.03611421585083,
+ "step": 707
+ },
+ {
+ "epoch": 9.838427947598253,
+ "grad_norm": 0.02058163844048977,
+ "learning_rate": 0.0006,
+ "loss": 4.9613447189331055,
+ "step": 708
+ },
+ {
+ "epoch": 9.85240174672489,
+ "grad_norm": 0.018004924058914185,
+ "learning_rate": 0.0006,
+ "loss": 4.966541767120361,
+ "step": 709
+ },
+ {
+ "epoch": 9.866375545851529,
+ "grad_norm": 0.0198152307420969,
+ "learning_rate": 0.0006,
+ "loss": 5.01750373840332,
+ "step": 710
+ },
+ {
+ "epoch": 9.880349344978166,
+ "grad_norm": 0.018532969057559967,
+ "learning_rate": 0.0006,
+ "loss": 5.097424507141113,
+ "step": 711
+ },
+ {
+ "epoch": 9.894323144104803,
+ "grad_norm": 0.018939971923828125,
+ "learning_rate": 0.0006,
+ "loss": 5.177223205566406,
+ "step": 712
+ },
+ {
+ "epoch": 9.908296943231441,
+ "grad_norm": 0.017969170585274696,
+ "learning_rate": 0.0006,
+ "loss": 5.118013858795166,
+ "step": 713
+ },
+ {
+ "epoch": 9.922270742358078,
+ "grad_norm": 0.018021270632743835,
+ "learning_rate": 0.0006,
+ "loss": 5.034787178039551,
+ "step": 714
+ },
+ {
+ "epoch": 9.936244541484717,
+ "grad_norm": 0.017009710893034935,
+ "learning_rate": 0.0006,
+ "loss": 5.003920555114746,
+ "step": 715
+ },
+ {
+ "epoch": 9.950218340611354,
+ "grad_norm": 0.017722049728035927,
+ "learning_rate": 0.0006,
+ "loss": 5.031939506530762,
+ "step": 716
+ },
+ {
+ "epoch": 9.96419213973799,
+ "grad_norm": 0.017134087160229683,
+ "learning_rate": 0.0006,
+ "loss": 5.093474388122559,
+ "step": 717
+ },
+ {
+ "epoch": 9.97816593886463,
+ "grad_norm": 0.018511613830924034,
+ "learning_rate": 0.0006,
+ "loss": 5.078604698181152,
+ "step": 718
+ },
+ {
+ "epoch": 9.992139737991266,
+ "grad_norm": 0.021206334233283997,
+ "learning_rate": 0.0006,
+ "loss": 5.052572727203369,
+ "step": 719
+ },
+ {
+ "epoch": 10.0,
+ "grad_norm": 0.02239062264561653,
+ "learning_rate": 0.0006,
+ "loss": 5.063547134399414,
+ "step": 720
+ },
+ {
+ "epoch": 10.0,
+ "eval_loss": 5.173044681549072,
+ "eval_runtime": 56.7287,
+ "eval_samples_per_second": 43.047,
+ "eval_steps_per_second": 1.357,
+ "step": 720
+ },
+ {
+ "epoch": 10.013973799126637,
+ "grad_norm": 0.020757969468832016,
+ "learning_rate": 0.0006,
+ "loss": 5.045648574829102,
+ "step": 721
+ },
+ {
+ "epoch": 10.027947598253276,
+ "grad_norm": 0.022697702050209045,
+ "learning_rate": 0.0006,
+ "loss": 4.993833065032959,
+ "step": 722
+ },
+ {
+ "epoch": 10.041921397379912,
+ "grad_norm": 0.02434331737458706,
+ "learning_rate": 0.0006,
+ "loss": 5.073219299316406,
+ "step": 723
+ },
+ {
+ "epoch": 10.055895196506551,
+ "grad_norm": 0.025499660521745682,
+ "learning_rate": 0.0006,
+ "loss": 5.019072532653809,
+ "step": 724
+ },
+ {
+ "epoch": 10.069868995633188,
+ "grad_norm": 0.029281053692102432,
+ "learning_rate": 0.0006,
+ "loss": 4.959851264953613,
+ "step": 725
+ },
+ {
+ "epoch": 10.083842794759825,
+ "grad_norm": 0.033878110349178314,
+ "learning_rate": 0.0006,
+ "loss": 5.132717609405518,
+ "step": 726
+ },
+ {
+ "epoch": 10.097816593886463,
+ "grad_norm": 0.03385764732956886,
+ "learning_rate": 0.0006,
+ "loss": 5.0482892990112305,
+ "step": 727
+ },
+ {
+ "epoch": 10.1117903930131,
+ "grad_norm": 0.03466648980975151,
+ "learning_rate": 0.0006,
+ "loss": 4.946855545043945,
+ "step": 728
+ },
+ {
+ "epoch": 10.125764192139737,
+ "grad_norm": 0.028546305373311043,
+ "learning_rate": 0.0006,
+ "loss": 5.027774333953857,
+ "step": 729
+ },
+ {
+ "epoch": 10.139737991266376,
+ "grad_norm": 0.030575869604945183,
+ "learning_rate": 0.0006,
+ "loss": 4.9768476486206055,
+ "step": 730
+ },
+ {
+ "epoch": 10.153711790393013,
+ "grad_norm": 0.03614303469657898,
+ "learning_rate": 0.0006,
+ "loss": 5.078113555908203,
+ "step": 731
+ },
+ {
+ "epoch": 10.167685589519651,
+ "grad_norm": 0.040991660207509995,
+ "learning_rate": 0.0006,
+ "loss": 5.033185958862305,
+ "step": 732
+ },
+ {
+ "epoch": 10.181659388646288,
+ "grad_norm": 0.04551699757575989,
+ "learning_rate": 0.0006,
+ "loss": 5.0820746421813965,
+ "step": 733
+ },
+ {
+ "epoch": 10.195633187772925,
+ "grad_norm": 0.040193989872932434,
+ "learning_rate": 0.0006,
+ "loss": 5.059700965881348,
+ "step": 734
+ },
+ {
+ "epoch": 10.209606986899564,
+ "grad_norm": 0.035851605236530304,
+ "learning_rate": 0.0006,
+ "loss": 5.142467498779297,
+ "step": 735
+ },
+ {
+ "epoch": 10.2235807860262,
+ "grad_norm": 0.034200169146060944,
+ "learning_rate": 0.0006,
+ "loss": 5.072312355041504,
+ "step": 736
+ },
+ {
+ "epoch": 10.237554585152838,
+ "grad_norm": 0.033621110022068024,
+ "learning_rate": 0.0006,
+ "loss": 5.08037805557251,
+ "step": 737
+ },
+ {
+ "epoch": 10.251528384279476,
+ "grad_norm": 0.037372369319200516,
+ "learning_rate": 0.0006,
+ "loss": 4.998425483703613,
+ "step": 738
+ },
+ {
+ "epoch": 10.265502183406113,
+ "grad_norm": 0.037044707685709,
+ "learning_rate": 0.0006,
+ "loss": 4.923993110656738,
+ "step": 739
+ },
+ {
+ "epoch": 10.279475982532752,
+ "grad_norm": 0.030454808846116066,
+ "learning_rate": 0.0006,
+ "loss": 5.0889997482299805,
+ "step": 740
+ },
+ {
+ "epoch": 10.293449781659389,
+ "grad_norm": 0.030969638377428055,
+ "learning_rate": 0.0006,
+ "loss": 4.984112739562988,
+ "step": 741
+ },
+ {
+ "epoch": 10.307423580786025,
+ "grad_norm": 0.0276983380317688,
+ "learning_rate": 0.0006,
+ "loss": 4.955412864685059,
+ "step": 742
+ },
+ {
+ "epoch": 10.321397379912664,
+ "grad_norm": 0.02716052532196045,
+ "learning_rate": 0.0006,
+ "loss": 5.0098876953125,
+ "step": 743
+ },
+ {
+ "epoch": 10.335371179039301,
+ "grad_norm": 0.02403552085161209,
+ "learning_rate": 0.0006,
+ "loss": 5.165160655975342,
+ "step": 744
+ },
+ {
+ "epoch": 10.34934497816594,
+ "grad_norm": 0.0257862601429224,
+ "learning_rate": 0.0006,
+ "loss": 5.001547813415527,
+ "step": 745
+ },
+ {
+ "epoch": 10.363318777292577,
+ "grad_norm": 0.02300378680229187,
+ "learning_rate": 0.0006,
+ "loss": 5.038888931274414,
+ "step": 746
+ },
+ {
+ "epoch": 10.377292576419213,
+ "grad_norm": 0.019766854122281075,
+ "learning_rate": 0.0006,
+ "loss": 4.924291610717773,
+ "step": 747
+ },
+ {
+ "epoch": 10.391266375545852,
+ "grad_norm": 0.01888395845890045,
+ "learning_rate": 0.0006,
+ "loss": 5.059971332550049,
+ "step": 748
+ },
+ {
+ "epoch": 10.405240174672489,
+ "grad_norm": 0.015577499754726887,
+ "learning_rate": 0.0006,
+ "loss": 4.987686634063721,
+ "step": 749
+ },
+ {
+ "epoch": 10.419213973799126,
+ "grad_norm": 0.016012346372008324,
+ "learning_rate": 0.0006,
+ "loss": 5.015157699584961,
+ "step": 750
+ },
+ {
+ "epoch": 10.433187772925764,
+ "grad_norm": 0.016277998685836792,
+ "learning_rate": 0.0006,
+ "loss": 5.004931449890137,
+ "step": 751
+ },
+ {
+ "epoch": 10.447161572052401,
+ "grad_norm": 0.017504561692476273,
+ "learning_rate": 0.0006,
+ "loss": 4.902827739715576,
+ "step": 752
+ },
+ {
+ "epoch": 10.46113537117904,
+ "grad_norm": 0.01569022797048092,
+ "learning_rate": 0.0006,
+ "loss": 4.954863548278809,
+ "step": 753
+ },
+ {
+ "epoch": 10.475109170305677,
+ "grad_norm": 0.014114422723650932,
+ "learning_rate": 0.0006,
+ "loss": 5.00656795501709,
+ "step": 754
+ },
+ {
+ "epoch": 10.489082969432314,
+ "grad_norm": 0.01573832333087921,
+ "learning_rate": 0.0006,
+ "loss": 4.978353023529053,
+ "step": 755
+ },
+ {
+ "epoch": 10.503056768558952,
+ "grad_norm": 0.016480036079883575,
+ "learning_rate": 0.0006,
+ "loss": 4.953327655792236,
+ "step": 756
+ },
+ {
+ "epoch": 10.51703056768559,
+ "grad_norm": 0.014008025638759136,
+ "learning_rate": 0.0006,
+ "loss": 5.050840377807617,
+ "step": 757
+ },
+ {
+ "epoch": 10.531004366812226,
+ "grad_norm": 0.013147998601198196,
+ "learning_rate": 0.0006,
+ "loss": 4.974964141845703,
+ "step": 758
+ },
+ {
+ "epoch": 10.544978165938865,
+ "grad_norm": 0.013807603158056736,
+ "learning_rate": 0.0006,
+ "loss": 4.927907943725586,
+ "step": 759
+ },
+ {
+ "epoch": 10.558951965065502,
+ "grad_norm": 0.013555224984884262,
+ "learning_rate": 0.0006,
+ "loss": 4.980690956115723,
+ "step": 760
+ },
+ {
+ "epoch": 10.57292576419214,
+ "grad_norm": 0.013045408762991428,
+ "learning_rate": 0.0006,
+ "loss": 4.9436116218566895,
+ "step": 761
+ },
+ {
+ "epoch": 10.586899563318777,
+ "grad_norm": 0.012858862057328224,
+ "learning_rate": 0.0006,
+ "loss": 4.9595770835876465,
+ "step": 762
+ },
+ {
+ "epoch": 10.600873362445414,
+ "grad_norm": 0.01468253880739212,
+ "learning_rate": 0.0006,
+ "loss": 4.90725040435791,
+ "step": 763
+ },
+ {
+ "epoch": 10.614847161572053,
+ "grad_norm": 0.013224679045379162,
+ "learning_rate": 0.0006,
+ "loss": 4.877615928649902,
+ "step": 764
+ },
+ {
+ "epoch": 10.62882096069869,
+ "grad_norm": 0.013555348850786686,
+ "learning_rate": 0.0006,
+ "loss": 4.9456610679626465,
+ "step": 765
+ },
+ {
+ "epoch": 10.642794759825328,
+ "grad_norm": 0.014387160539627075,
+ "learning_rate": 0.0006,
+ "loss": 4.9233717918396,
+ "step": 766
+ },
+ {
+ "epoch": 10.656768558951965,
+ "grad_norm": 0.014354088343679905,
+ "learning_rate": 0.0006,
+ "loss": 4.983511924743652,
+ "step": 767
+ },
+ {
+ "epoch": 10.670742358078602,
+ "grad_norm": 0.01332700252532959,
+ "learning_rate": 0.0006,
+ "loss": 4.934416770935059,
+ "step": 768
+ },
+ {
+ "epoch": 10.68471615720524,
+ "grad_norm": 0.011543313041329384,
+ "learning_rate": 0.0006,
+ "loss": 4.96388053894043,
+ "step": 769
+ },
+ {
+ "epoch": 10.698689956331878,
+ "grad_norm": 0.012330878525972366,
+ "learning_rate": 0.0006,
+ "loss": 4.946396350860596,
+ "step": 770
+ },
+ {
+ "epoch": 10.712663755458514,
+ "grad_norm": 0.01379779726266861,
+ "learning_rate": 0.0006,
+ "loss": 4.981019020080566,
+ "step": 771
+ },
+ {
+ "epoch": 10.726637554585153,
+ "grad_norm": 0.014612431637942791,
+ "learning_rate": 0.0006,
+ "loss": 4.894567966461182,
+ "step": 772
+ },
+ {
+ "epoch": 10.74061135371179,
+ "grad_norm": 0.013062899932265282,
+ "learning_rate": 0.0006,
+ "loss": 5.033038139343262,
+ "step": 773
+ },
+ {
+ "epoch": 10.754585152838429,
+ "grad_norm": 0.013898049481213093,
+ "learning_rate": 0.0006,
+ "loss": 4.823636054992676,
+ "step": 774
+ },
+ {
+ "epoch": 10.768558951965066,
+ "grad_norm": 0.01502019353210926,
+ "learning_rate": 0.0006,
+ "loss": 5.020156383514404,
+ "step": 775
+ },
+ {
+ "epoch": 10.782532751091702,
+ "grad_norm": 0.013810204342007637,
+ "learning_rate": 0.0006,
+ "loss": 4.935647010803223,
+ "step": 776
+ },
+ {
+ "epoch": 10.796506550218341,
+ "grad_norm": 0.013616513460874557,
+ "learning_rate": 0.0006,
+ "loss": 4.901486396789551,
+ "step": 777
+ },
+ {
+ "epoch": 10.810480349344978,
+ "grad_norm": 0.013601024635136127,
+ "learning_rate": 0.0006,
+ "loss": 4.931873321533203,
+ "step": 778
+ },
+ {
+ "epoch": 10.824454148471617,
+ "grad_norm": 0.013668350875377655,
+ "learning_rate": 0.0006,
+ "loss": 4.937256813049316,
+ "step": 779
+ },
+ {
+ "epoch": 10.838427947598253,
+ "grad_norm": 0.01581823080778122,
+ "learning_rate": 0.0006,
+ "loss": 4.9611406326293945,
+ "step": 780
+ },
+ {
+ "epoch": 10.85240174672489,
+ "grad_norm": 0.019478484988212585,
+ "learning_rate": 0.0006,
+ "loss": 4.9212236404418945,
+ "step": 781
+ },
+ {
+ "epoch": 10.866375545851529,
+ "grad_norm": 0.026921333745121956,
+ "learning_rate": 0.0006,
+ "loss": 4.88501501083374,
+ "step": 782
+ },
+ {
+ "epoch": 10.880349344978166,
+ "grad_norm": 0.03155914694070816,
+ "learning_rate": 0.0006,
+ "loss": 4.989967346191406,
+ "step": 783
+ },
+ {
+ "epoch": 10.894323144104803,
+ "grad_norm": 0.029689429327845573,
+ "learning_rate": 0.0006,
+ "loss": 4.936000823974609,
+ "step": 784
+ },
+ {
+ "epoch": 10.908296943231441,
+ "grad_norm": 0.026626063510775566,
+ "learning_rate": 0.0006,
+ "loss": 4.984785079956055,
+ "step": 785
+ },
+ {
+ "epoch": 10.922270742358078,
+ "grad_norm": 0.026168212294578552,
+ "learning_rate": 0.0006,
+ "loss": 4.908282279968262,
+ "step": 786
+ },
+ {
+ "epoch": 10.936244541484717,
+ "grad_norm": 0.02828158251941204,
+ "learning_rate": 0.0006,
+ "loss": 4.926390647888184,
+ "step": 787
+ },
+ {
+ "epoch": 10.950218340611354,
+ "grad_norm": 0.02649078331887722,
+ "learning_rate": 0.0006,
+ "loss": 5.00400972366333,
+ "step": 788
+ },
+ {
+ "epoch": 10.96419213973799,
+ "grad_norm": 0.0256856270134449,
+ "learning_rate": 0.0006,
+ "loss": 5.0317487716674805,
+ "step": 789
+ },
+ {
+ "epoch": 10.97816593886463,
+ "grad_norm": 0.024728883057832718,
+ "learning_rate": 0.0006,
+ "loss": 4.9299635887146,
+ "step": 790
+ },
+ {
+ "epoch": 10.992139737991266,
+ "grad_norm": 0.026241116225719452,
+ "learning_rate": 0.0006,
+ "loss": 4.947319984436035,
+ "step": 791
+ },
+ {
+ "epoch": 11.0,
+ "grad_norm": 0.023526743054389954,
+ "learning_rate": 0.0006,
+ "loss": 4.94429349899292,
+ "step": 792
+ },
+ {
+ "epoch": 11.0,
+ "eval_loss": 5.104345798492432,
+ "eval_runtime": 57.0955,
+ "eval_samples_per_second": 42.77,
+ "eval_steps_per_second": 1.349,
+ "step": 792
+ },
+ {
+ "epoch": 11.013973799126637,
+ "grad_norm": 0.0216788612306118,
+ "learning_rate": 0.0006,
+ "loss": 4.881950855255127,
+ "step": 793
+ },
+ {
+ "epoch": 11.027947598253276,
+ "grad_norm": 0.01830463856458664,
+ "learning_rate": 0.0006,
+ "loss": 4.971016883850098,
+ "step": 794
+ },
+ {
+ "epoch": 11.041921397379912,
+ "grad_norm": 0.02204521745443344,
+ "learning_rate": 0.0006,
+ "loss": 4.894623756408691,
+ "step": 795
+ },
+ {
+ "epoch": 11.055895196506551,
+ "grad_norm": 0.0210255216807127,
+ "learning_rate": 0.0006,
+ "loss": 4.834009170532227,
+ "step": 796
+ },
+ {
+ "epoch": 11.069868995633188,
+ "grad_norm": 0.019797448068857193,
+ "learning_rate": 0.0006,
+ "loss": 5.086714744567871,
+ "step": 797
+ },
+ {
+ "epoch": 11.083842794759825,
+ "grad_norm": 0.02036641724407673,
+ "learning_rate": 0.0006,
+ "loss": 4.920056343078613,
+ "step": 798
+ },
+ {
+ "epoch": 11.097816593886463,
+ "grad_norm": 0.01966066285967827,
+ "learning_rate": 0.0006,
+ "loss": 4.879859447479248,
+ "step": 799
+ },
+ {
+ "epoch": 11.1117903930131,
+ "grad_norm": 0.0202149897813797,
+ "learning_rate": 0.0006,
+ "loss": 4.933640480041504,
+ "step": 800
+ },
+ {
+ "epoch": 11.125764192139737,
+ "grad_norm": 0.01893039606511593,
+ "learning_rate": 0.0006,
+ "loss": 4.901970863342285,
+ "step": 801
+ },
+ {
+ "epoch": 11.139737991266376,
+ "grad_norm": 0.016040155664086342,
+ "learning_rate": 0.0006,
+ "loss": 4.905491828918457,
+ "step": 802
+ },
+ {
+ "epoch": 11.153711790393013,
+ "grad_norm": 0.01720350608229637,
+ "learning_rate": 0.0006,
+ "loss": 4.798816680908203,
+ "step": 803
+ },
+ {
+ "epoch": 11.167685589519651,
+ "grad_norm": 0.017002852633595467,
+ "learning_rate": 0.0006,
+ "loss": 4.851859092712402,
+ "step": 804
+ },
+ {
+ "epoch": 11.181659388646288,
+ "grad_norm": 0.015594680793583393,
+ "learning_rate": 0.0006,
+ "loss": 4.90328311920166,
+ "step": 805
+ },
+ {
+ "epoch": 11.195633187772925,
+ "grad_norm": 0.014904462732374668,
+ "learning_rate": 0.0006,
+ "loss": 4.904215335845947,
+ "step": 806
+ },
+ {
+ "epoch": 11.209606986899564,
+ "grad_norm": 0.01366228237748146,
+ "learning_rate": 0.0006,
+ "loss": 4.9813055992126465,
+ "step": 807
+ },
+ {
+ "epoch": 11.2235807860262,
+ "grad_norm": 0.014719638973474503,
+ "learning_rate": 0.0006,
+ "loss": 4.927306175231934,
+ "step": 808
+ },
+ {
+ "epoch": 11.237554585152838,
+ "grad_norm": 0.015016036108136177,
+ "learning_rate": 0.0006,
+ "loss": 4.874215126037598,
+ "step": 809
+ },
+ {
+ "epoch": 11.251528384279476,
+ "grad_norm": 0.014925424009561539,
+ "learning_rate": 0.0006,
+ "loss": 4.838257789611816,
+ "step": 810
+ },
+ {
+ "epoch": 11.265502183406113,
+ "grad_norm": 0.014335823245346546,
+ "learning_rate": 0.0006,
+ "loss": 4.940608024597168,
+ "step": 811
+ },
+ {
+ "epoch": 11.279475982532752,
+ "grad_norm": 0.014987512491643429,
+ "learning_rate": 0.0006,
+ "loss": 4.831531524658203,
+ "step": 812
+ },
+ {
+ "epoch": 11.293449781659389,
+ "grad_norm": 0.016087457537651062,
+ "learning_rate": 0.0006,
+ "loss": 4.863227367401123,
+ "step": 813
+ },
+ {
+ "epoch": 11.307423580786025,
+ "grad_norm": 0.01671300269663334,
+ "learning_rate": 0.0006,
+ "loss": 4.8535661697387695,
+ "step": 814
+ },
+ {
+ "epoch": 11.321397379912664,
+ "grad_norm": 0.015229142270982265,
+ "learning_rate": 0.0006,
+ "loss": 4.948635101318359,
+ "step": 815
+ },
+ {
+ "epoch": 11.335371179039301,
+ "grad_norm": 0.01532587967813015,
+ "learning_rate": 0.0006,
+ "loss": 4.9035797119140625,
+ "step": 816
+ },
+ {
+ "epoch": 11.34934497816594,
+ "grad_norm": 0.015715476125478745,
+ "learning_rate": 0.0006,
+ "loss": 4.894709587097168,
+ "step": 817
+ },
+ {
+ "epoch": 11.363318777292577,
+ "grad_norm": 0.014156977646052837,
+ "learning_rate": 0.0006,
+ "loss": 4.960206985473633,
+ "step": 818
+ },
+ {
+ "epoch": 11.377292576419213,
+ "grad_norm": 0.017076566815376282,
+ "learning_rate": 0.0006,
+ "loss": 5.005850791931152,
+ "step": 819
+ },
+ {
+ "epoch": 11.391266375545852,
+ "grad_norm": 0.0172084029763937,
+ "learning_rate": 0.0006,
+ "loss": 4.903355598449707,
+ "step": 820
+ },
+ {
+ "epoch": 11.405240174672489,
+ "grad_norm": 0.0180258359760046,
+ "learning_rate": 0.0006,
+ "loss": 4.842667579650879,
+ "step": 821
+ },
+ {
+ "epoch": 11.419213973799126,
+ "grad_norm": 0.022130120545625687,
+ "learning_rate": 0.0006,
+ "loss": 4.799970626831055,
+ "step": 822
+ },
+ {
+ "epoch": 11.433187772925764,
+ "grad_norm": 0.02384897693991661,
+ "learning_rate": 0.0006,
+ "loss": 4.860746383666992,
+ "step": 823
+ },
+ {
+ "epoch": 11.447161572052401,
+ "grad_norm": 0.021881572902202606,
+ "learning_rate": 0.0006,
+ "loss": 4.918107032775879,
+ "step": 824
+ },
+ {
+ "epoch": 11.46113537117904,
+ "grad_norm": 0.023085080087184906,
+ "learning_rate": 0.0006,
+ "loss": 4.878946304321289,
+ "step": 825
+ },
+ {
+ "epoch": 11.475109170305677,
+ "grad_norm": 0.02291012555360794,
+ "learning_rate": 0.0006,
+ "loss": 4.958887100219727,
+ "step": 826
+ },
+ {
+ "epoch": 11.489082969432314,
+ "grad_norm": 0.0199968870729208,
+ "learning_rate": 0.0006,
+ "loss": 4.80955696105957,
+ "step": 827
+ },
+ {
+ "epoch": 11.503056768558952,
+ "grad_norm": 0.018702229484915733,
+ "learning_rate": 0.0006,
+ "loss": 5.0276408195495605,
+ "step": 828
+ },
+ {
+ "epoch": 11.51703056768559,
+ "grad_norm": 0.017787311226129532,
+ "learning_rate": 0.0006,
+ "loss": 4.96018123626709,
+ "step": 829
+ },
+ {
+ "epoch": 11.531004366812226,
+ "grad_norm": 0.01704501546919346,
+ "learning_rate": 0.0006,
+ "loss": 4.914553165435791,
+ "step": 830
+ },
+ {
+ "epoch": 11.544978165938865,
+ "grad_norm": 0.017519677057862282,
+ "learning_rate": 0.0006,
+ "loss": 4.798098087310791,
+ "step": 831
+ },
+ {
+ "epoch": 11.558951965065502,
+ "grad_norm": 0.018511991947889328,
+ "learning_rate": 0.0006,
+ "loss": 4.817785263061523,
+ "step": 832
+ },
+ {
+ "epoch": 11.57292576419214,
+ "grad_norm": 0.01876644790172577,
+ "learning_rate": 0.0006,
+ "loss": 4.818995475769043,
+ "step": 833
+ },
+ {
+ "epoch": 11.586899563318777,
+ "grad_norm": 0.01753568835556507,
+ "learning_rate": 0.0006,
+ "loss": 4.811938285827637,
+ "step": 834
+ },
+ {
+ "epoch": 11.600873362445414,
+ "grad_norm": 0.017046531662344933,
+ "learning_rate": 0.0006,
+ "loss": 4.914674282073975,
+ "step": 835
+ },
+ {
+ "epoch": 11.614847161572053,
+ "grad_norm": 0.01802094094455242,
+ "learning_rate": 0.0006,
+ "loss": 4.993284225463867,
+ "step": 836
+ },
+ {
+ "epoch": 11.62882096069869,
+ "grad_norm": 0.019493764266371727,
+ "learning_rate": 0.0006,
+ "loss": 4.895628452301025,
+ "step": 837
+ },
+ {
+ "epoch": 11.642794759825328,
+ "grad_norm": 0.019461210817098618,
+ "learning_rate": 0.0006,
+ "loss": 4.869269371032715,
+ "step": 838
+ },
+ {
+ "epoch": 11.656768558951965,
+ "grad_norm": 0.01981567218899727,
+ "learning_rate": 0.0006,
+ "loss": 4.863052845001221,
+ "step": 839
+ },
+ {
+ "epoch": 11.670742358078602,
+ "grad_norm": 0.019894491881132126,
+ "learning_rate": 0.0006,
+ "loss": 4.8775787353515625,
+ "step": 840
+ },
+ {
+ "epoch": 11.68471615720524,
+ "grad_norm": 0.01875786855816841,
+ "learning_rate": 0.0006,
+ "loss": 4.817763328552246,
+ "step": 841
+ },
+ {
+ "epoch": 11.698689956331878,
+ "grad_norm": 0.020926695317029953,
+ "learning_rate": 0.0006,
+ "loss": 4.7470526695251465,
+ "step": 842
+ },
+ {
+ "epoch": 11.712663755458514,
+ "grad_norm": 0.021477103233337402,
+ "learning_rate": 0.0006,
+ "loss": 4.762682914733887,
+ "step": 843
+ },
+ {
+ "epoch": 11.726637554585153,
+ "grad_norm": 0.020628711208701134,
+ "learning_rate": 0.0006,
+ "loss": 4.807985782623291,
+ "step": 844
+ },
+ {
+ "epoch": 11.74061135371179,
+ "grad_norm": 0.021489612758159637,
+ "learning_rate": 0.0006,
+ "loss": 4.843091011047363,
+ "step": 845
+ },
+ {
+ "epoch": 11.754585152838429,
+ "grad_norm": 0.021194491535425186,
+ "learning_rate": 0.0006,
+ "loss": 4.8300251960754395,
+ "step": 846
+ },
+ {
+ "epoch": 11.768558951965066,
+ "grad_norm": 0.01833350397646427,
+ "learning_rate": 0.0006,
+ "loss": 5.000202178955078,
+ "step": 847
+ },
+ {
+ "epoch": 11.782532751091702,
+ "grad_norm": 0.018096931278705597,
+ "learning_rate": 0.0006,
+ "loss": 4.880648136138916,
+ "step": 848
+ },
+ {
+ "epoch": 11.796506550218341,
+ "grad_norm": 0.0197161752730608,
+ "learning_rate": 0.0006,
+ "loss": 4.788316249847412,
+ "step": 849
+ },
+ {
+ "epoch": 11.810480349344978,
+ "grad_norm": 0.0203181654214859,
+ "learning_rate": 0.0006,
+ "loss": 4.870944976806641,
+ "step": 850
+ },
+ {
+ "epoch": 11.824454148471617,
+ "grad_norm": 0.01908455230295658,
+ "learning_rate": 0.0006,
+ "loss": 4.890144348144531,
+ "step": 851
+ },
+ {
+ "epoch": 11.838427947598253,
+ "grad_norm": 0.021281801164150238,
+ "learning_rate": 0.0006,
+ "loss": 4.89578914642334,
+ "step": 852
+ },
+ {
+ "epoch": 11.85240174672489,
+ "grad_norm": 0.02124273031949997,
+ "learning_rate": 0.0006,
+ "loss": 4.883749008178711,
+ "step": 853
+ },
+ {
+ "epoch": 11.866375545851529,
+ "grad_norm": 0.020969398319721222,
+ "learning_rate": 0.0006,
+ "loss": 4.809142589569092,
+ "step": 854
+ },
+ {
+ "epoch": 11.880349344978166,
+ "grad_norm": 0.01967657171189785,
+ "learning_rate": 0.0006,
+ "loss": 4.821864604949951,
+ "step": 855
+ },
+ {
+ "epoch": 11.894323144104803,
+ "grad_norm": 0.020887333899736404,
+ "learning_rate": 0.0006,
+ "loss": 4.917985916137695,
+ "step": 856
+ },
+ {
+ "epoch": 11.908296943231441,
+ "grad_norm": 0.019203083589673042,
+ "learning_rate": 0.0006,
+ "loss": 4.824960708618164,
+ "step": 857
+ },
+ {
+ "epoch": 11.922270742358078,
+ "grad_norm": 0.020162401720881462,
+ "learning_rate": 0.0006,
+ "loss": 4.772617816925049,
+ "step": 858
+ },
+ {
+ "epoch": 11.936244541484717,
+ "grad_norm": 0.02251487225294113,
+ "learning_rate": 0.0006,
+ "loss": 4.735418319702148,
+ "step": 859
+ },
+ {
+ "epoch": 11.950218340611354,
+ "grad_norm": 0.020313527435064316,
+ "learning_rate": 0.0006,
+ "loss": 4.939233779907227,
+ "step": 860
+ },
+ {
+ "epoch": 11.96419213973799,
+ "grad_norm": 0.021635599434375763,
+ "learning_rate": 0.0006,
+ "loss": 4.861807346343994,
+ "step": 861
+ },
+ {
+ "epoch": 11.97816593886463,
+ "grad_norm": 0.02285311557352543,
+ "learning_rate": 0.0006,
+ "loss": 4.932732582092285,
+ "step": 862
+ },
+ {
+ "epoch": 11.992139737991266,
+ "grad_norm": 0.02072535641491413,
+ "learning_rate": 0.0006,
+ "loss": 4.796767711639404,
+ "step": 863
+ },
+ {
+ "epoch": 12.0,
+ "grad_norm": 0.020588304847478867,
+ "learning_rate": 0.0006,
+ "loss": 4.997537612915039,
+ "step": 864
+ },
+ {
+ "epoch": 12.0,
+ "eval_loss": 5.044651985168457,
+ "eval_runtime": 56.9145,
+ "eval_samples_per_second": 42.906,
+ "eval_steps_per_second": 1.353,
+ "step": 864
+ },
+ {
+ "epoch": 12.013973799126637,
+ "grad_norm": 0.017597166821360588,
+ "learning_rate": 0.0006,
+ "loss": 4.830524444580078,
+ "step": 865
+ },
+ {
+ "epoch": 12.027947598253276,
+ "grad_norm": 0.01716788113117218,
+ "learning_rate": 0.0006,
+ "loss": 4.894325256347656,
+ "step": 866
+ },
+ {
+ "epoch": 12.041921397379912,
+ "grad_norm": 0.017712457105517387,
+ "learning_rate": 0.0006,
+ "loss": 4.868297576904297,
+ "step": 867
+ },
+ {
+ "epoch": 12.055895196506551,
+ "grad_norm": 0.021486390382051468,
+ "learning_rate": 0.0006,
+ "loss": 4.922049522399902,
+ "step": 868
+ },
+ {
+ "epoch": 12.069868995633188,
+ "grad_norm": 0.024560654535889626,
+ "learning_rate": 0.0006,
+ "loss": 4.734241008758545,
+ "step": 869
+ },
+ {
+ "epoch": 12.083842794759825,
+ "grad_norm": 0.025870252400636673,
+ "learning_rate": 0.0006,
+ "loss": 4.868566989898682,
+ "step": 870
+ },
+ {
+ "epoch": 12.097816593886463,
+ "grad_norm": 0.0277389045804739,
+ "learning_rate": 0.0006,
+ "loss": 4.755006790161133,
+ "step": 871
+ },
+ {
+ "epoch": 12.1117903930131,
+ "grad_norm": 0.029100844636559486,
+ "learning_rate": 0.0006,
+ "loss": 4.909667015075684,
+ "step": 872
+ },
+ {
+ "epoch": 12.125764192139737,
+ "grad_norm": 0.031833067536354065,
+ "learning_rate": 0.0006,
+ "loss": 4.858674049377441,
+ "step": 873
+ },
+ {
+ "epoch": 12.139737991266376,
+ "grad_norm": 0.032709237188100815,
+ "learning_rate": 0.0006,
+ "loss": 4.823696136474609,
+ "step": 874
+ },
+ {
+ "epoch": 12.153711790393013,
+ "grad_norm": 0.02995380386710167,
+ "learning_rate": 0.0006,
+ "loss": 4.850235462188721,
+ "step": 875
+ },
+ {
+ "epoch": 12.167685589519651,
+ "grad_norm": 0.034098610281944275,
+ "learning_rate": 0.0006,
+ "loss": 4.8169989585876465,
+ "step": 876
+ },
+ {
+ "epoch": 12.181659388646288,
+ "grad_norm": 0.03377080336213112,
+ "learning_rate": 0.0006,
+ "loss": 4.90738582611084,
+ "step": 877
+ },
+ {
+ "epoch": 12.195633187772925,
+ "grad_norm": 0.029780825600028038,
+ "learning_rate": 0.0006,
+ "loss": 4.811163902282715,
+ "step": 878
+ },
+ {
+ "epoch": 12.209606986899564,
+ "grad_norm": 0.03118024580180645,
+ "learning_rate": 0.0006,
+ "loss": 4.861286640167236,
+ "step": 879
+ },
+ {
+ "epoch": 12.2235807860262,
+ "grad_norm": 0.033052004873752594,
+ "learning_rate": 0.0006,
+ "loss": 4.865021228790283,
+ "step": 880
+ },
+ {
+ "epoch": 12.237554585152838,
+ "grad_norm": 0.029070032760500908,
+ "learning_rate": 0.0006,
+ "loss": 4.947303771972656,
+ "step": 881
+ },
+ {
+ "epoch": 12.251528384279476,
+ "grad_norm": 0.028355715796351433,
+ "learning_rate": 0.0006,
+ "loss": 4.804648399353027,
+ "step": 882
+ },
+ {
+ "epoch": 12.265502183406113,
+ "grad_norm": 0.031257227063179016,
+ "learning_rate": 0.0006,
+ "loss": 4.757718086242676,
+ "step": 883
+ },
+ {
+ "epoch": 12.279475982532752,
+ "grad_norm": 0.031438373029232025,
+ "learning_rate": 0.0006,
+ "loss": 4.851268768310547,
+ "step": 884
+ },
+ {
+ "epoch": 12.293449781659389,
+ "grad_norm": 0.030492138117551804,
+ "learning_rate": 0.0006,
+ "loss": 4.7891035079956055,
+ "step": 885
+ },
+ {
+ "epoch": 12.307423580786025,
+ "grad_norm": 0.026115991175174713,
+ "learning_rate": 0.0006,
+ "loss": 4.86474609375,
+ "step": 886
+ },
+ {
+ "epoch": 12.321397379912664,
+ "grad_norm": 0.024184465408325195,
+ "learning_rate": 0.0006,
+ "loss": 4.860456943511963,
+ "step": 887
+ },
+ {
+ "epoch": 12.335371179039301,
+ "grad_norm": 0.025437982752919197,
+ "learning_rate": 0.0006,
+ "loss": 4.806251049041748,
+ "step": 888
+ },
+ {
+ "epoch": 12.34934497816594,
+ "grad_norm": 0.022390231490135193,
+ "learning_rate": 0.0006,
+ "loss": 4.807086944580078,
+ "step": 889
+ },
+ {
+ "epoch": 12.363318777292577,
+ "grad_norm": 0.020174294710159302,
+ "learning_rate": 0.0006,
+ "loss": 4.78292179107666,
+ "step": 890
+ },
+ {
+ "epoch": 12.377292576419213,
+ "grad_norm": 0.019734324887394905,
+ "learning_rate": 0.0006,
+ "loss": 4.784509658813477,
+ "step": 891
+ },
+ {
+ "epoch": 12.391266375545852,
+ "grad_norm": 0.016733523458242416,
+ "learning_rate": 0.0006,
+ "loss": 4.811755180358887,
+ "step": 892
+ },
+ {
+ "epoch": 12.405240174672489,
+ "grad_norm": 0.01570146717131138,
+ "learning_rate": 0.0006,
+ "loss": 4.720088005065918,
+ "step": 893
+ },
+ {
+ "epoch": 12.419213973799126,
+ "grad_norm": 0.01601393334567547,
+ "learning_rate": 0.0006,
+ "loss": 4.783543109893799,
+ "step": 894
+ },
+ {
+ "epoch": 12.433187772925764,
+ "grad_norm": 0.014035487547516823,
+ "learning_rate": 0.0006,
+ "loss": 4.848340034484863,
+ "step": 895
+ },
+ {
+ "epoch": 12.447161572052401,
+ "grad_norm": 0.015248487703502178,
+ "learning_rate": 0.0006,
+ "loss": 4.773287773132324,
+ "step": 896
+ },
+ {
+ "epoch": 12.46113537117904,
+ "grad_norm": 0.01532268151640892,
+ "learning_rate": 0.0006,
+ "loss": 4.799646854400635,
+ "step": 897
+ },
+ {
+ "epoch": 12.475109170305677,
+ "grad_norm": 0.013982534408569336,
+ "learning_rate": 0.0006,
+ "loss": 4.8090500831604,
+ "step": 898
+ },
+ {
+ "epoch": 12.489082969432314,
+ "grad_norm": 0.014081164263188839,
+ "learning_rate": 0.0006,
+ "loss": 4.751888751983643,
+ "step": 899
+ },
+ {
+ "epoch": 12.503056768558952,
+ "grad_norm": 0.01380133256316185,
+ "learning_rate": 0.0006,
+ "loss": 4.78144645690918,
+ "step": 900
+ },
+ {
+ "epoch": 12.51703056768559,
+ "grad_norm": 0.013555269688367844,
+ "learning_rate": 0.0006,
+ "loss": 4.779257774353027,
+ "step": 901
+ },
+ {
+ "epoch": 12.531004366812226,
+ "grad_norm": 0.013075701892375946,
+ "learning_rate": 0.0006,
+ "loss": 4.820705413818359,
+ "step": 902
+ },
+ {
+ "epoch": 12.544978165938865,
+ "grad_norm": 0.01276316400617361,
+ "learning_rate": 0.0006,
+ "loss": 4.836733818054199,
+ "step": 903
+ },
+ {
+ "epoch": 12.558951965065502,
+ "grad_norm": 0.013033892959356308,
+ "learning_rate": 0.0006,
+ "loss": 4.764178276062012,
+ "step": 904
+ },
+ {
+ "epoch": 12.57292576419214,
+ "grad_norm": 0.013996980153024197,
+ "learning_rate": 0.0006,
+ "loss": 4.767983436584473,
+ "step": 905
+ },
+ {
+ "epoch": 12.586899563318777,
+ "grad_norm": 0.015222184360027313,
+ "learning_rate": 0.0006,
+ "loss": 4.759721755981445,
+ "step": 906
+ },
+ {
+ "epoch": 12.600873362445414,
+ "grad_norm": 0.01623941957950592,
+ "learning_rate": 0.0006,
+ "loss": 4.849811553955078,
+ "step": 907
+ },
+ {
+ "epoch": 12.614847161572053,
+ "grad_norm": 0.015711182728409767,
+ "learning_rate": 0.0006,
+ "loss": 4.7829084396362305,
+ "step": 908
+ },
+ {
+ "epoch": 12.62882096069869,
+ "grad_norm": 0.014137467369437218,
+ "learning_rate": 0.0006,
+ "loss": 4.73321533203125,
+ "step": 909
+ },
+ {
+ "epoch": 12.642794759825328,
+ "grad_norm": 0.012738462537527084,
+ "learning_rate": 0.0006,
+ "loss": 4.870052337646484,
+ "step": 910
+ },
+ {
+ "epoch": 12.656768558951965,
+ "grad_norm": 0.013683130033314228,
+ "learning_rate": 0.0006,
+ "loss": 4.82658052444458,
+ "step": 911
+ },
+ {
+ "epoch": 12.670742358078602,
+ "grad_norm": 0.014409830793738365,
+ "learning_rate": 0.0006,
+ "loss": 4.707494735717773,
+ "step": 912
+ },
+ {
+ "epoch": 12.68471615720524,
+ "grad_norm": 0.013806809671223164,
+ "learning_rate": 0.0006,
+ "loss": 4.778928279876709,
+ "step": 913
+ },
+ {
+ "epoch": 12.698689956331878,
+ "grad_norm": 0.012857912108302116,
+ "learning_rate": 0.0006,
+ "loss": 4.890717506408691,
+ "step": 914
+ },
+ {
+ "epoch": 12.712663755458514,
+ "grad_norm": 0.013828758150339127,
+ "learning_rate": 0.0006,
+ "loss": 4.8357439041137695,
+ "step": 915
+ },
+ {
+ "epoch": 12.726637554585153,
+ "grad_norm": 0.01478351280093193,
+ "learning_rate": 0.0006,
+ "loss": 4.869986534118652,
+ "step": 916
+ },
+ {
+ "epoch": 12.74061135371179,
+ "grad_norm": 0.01326421182602644,
+ "learning_rate": 0.0006,
+ "loss": 4.830646514892578,
+ "step": 917
+ },
+ {
+ "epoch": 12.754585152838429,
+ "grad_norm": 0.013335433788597584,
+ "learning_rate": 0.0006,
+ "loss": 4.7786149978637695,
+ "step": 918
+ },
+ {
+ "epoch": 12.768558951965066,
+ "grad_norm": 0.014261147007346153,
+ "learning_rate": 0.0006,
+ "loss": 4.788693428039551,
+ "step": 919
+ },
+ {
+ "epoch": 12.782532751091702,
+ "grad_norm": 0.014274783432483673,
+ "learning_rate": 0.0006,
+ "loss": 4.77295446395874,
+ "step": 920
+ },
+ {
+ "epoch": 12.796506550218341,
+ "grad_norm": 0.015074674971401691,
+ "learning_rate": 0.0006,
+ "loss": 4.830754280090332,
+ "step": 921
+ },
+ {
+ "epoch": 12.810480349344978,
+ "grad_norm": 0.015599515289068222,
+ "learning_rate": 0.0006,
+ "loss": 4.837759017944336,
+ "step": 922
+ },
+ {
+ "epoch": 12.824454148471617,
+ "grad_norm": 0.014595242217183113,
+ "learning_rate": 0.0006,
+ "loss": 4.882965564727783,
+ "step": 923
+ },
+ {
+ "epoch": 12.838427947598253,
+ "grad_norm": 0.014951467514038086,
+ "learning_rate": 0.0006,
+ "loss": 4.709748268127441,
+ "step": 924
+ },
+ {
+ "epoch": 12.85240174672489,
+ "grad_norm": 0.01539295818656683,
+ "learning_rate": 0.0006,
+ "loss": 4.791160583496094,
+ "step": 925
+ },
+ {
+ "epoch": 12.866375545851529,
+ "grad_norm": 0.015987906605005264,
+ "learning_rate": 0.0006,
+ "loss": 4.661876678466797,
+ "step": 926
+ },
+ {
+ "epoch": 12.880349344978166,
+ "grad_norm": 0.016527341678738594,
+ "learning_rate": 0.0006,
+ "loss": 4.721952438354492,
+ "step": 927
+ },
+ {
+ "epoch": 12.894323144104803,
+ "grad_norm": 0.01637161523103714,
+ "learning_rate": 0.0006,
+ "loss": 4.704446792602539,
+ "step": 928
+ },
+ {
+ "epoch": 12.908296943231441,
+ "grad_norm": 0.015221050009131432,
+ "learning_rate": 0.0006,
+ "loss": 4.771907806396484,
+ "step": 929
+ },
+ {
+ "epoch": 12.922270742358078,
+ "grad_norm": 0.016242166981101036,
+ "learning_rate": 0.0006,
+ "loss": 4.806267738342285,
+ "step": 930
+ },
+ {
+ "epoch": 12.936244541484717,
+ "grad_norm": 0.01628357358276844,
+ "learning_rate": 0.0006,
+ "loss": 4.7252326011657715,
+ "step": 931
+ },
+ {
+ "epoch": 12.950218340611354,
+ "grad_norm": 0.017428133636713028,
+ "learning_rate": 0.0006,
+ "loss": 4.765689849853516,
+ "step": 932
+ },
+ {
+ "epoch": 12.96419213973799,
+ "grad_norm": 0.01870061084628105,
+ "learning_rate": 0.0006,
+ "loss": 4.814488887786865,
+ "step": 933
+ },
+ {
+ "epoch": 12.97816593886463,
+ "grad_norm": 0.019796060398221016,
+ "learning_rate": 0.0006,
+ "loss": 4.739185810089111,
+ "step": 934
+ },
+ {
+ "epoch": 12.992139737991266,
+ "grad_norm": 0.01985820196568966,
+ "learning_rate": 0.0006,
+ "loss": 4.810608863830566,
+ "step": 935
+ },
+ {
+ "epoch": 13.0,
+ "grad_norm": 0.020834477618336678,
+ "learning_rate": 0.0006,
+ "loss": 4.727858066558838,
+ "step": 936
+ },
+ {
+ "epoch": 13.0,
+ "eval_loss": 4.979116916656494,
+ "eval_runtime": 57.2877,
+ "eval_samples_per_second": 42.627,
+ "eval_steps_per_second": 1.344,
+ "step": 936
+ },
+ {
+ "epoch": 13.013973799126637,
+ "grad_norm": 0.02025519125163555,
+ "learning_rate": 0.0006,
+ "loss": 4.628875732421875,
+ "step": 937
+ },
+ {
+ "epoch": 13.027947598253276,
+ "grad_norm": 0.019596368074417114,
+ "learning_rate": 0.0006,
+ "loss": 4.694699287414551,
+ "step": 938
+ },
+ {
+ "epoch": 13.041921397379912,
+ "grad_norm": 0.020494934171438217,
+ "learning_rate": 0.0006,
+ "loss": 4.691821098327637,
+ "step": 939
+ },
+ {
+ "epoch": 13.055895196506551,
+ "grad_norm": 0.020121760666370392,
+ "learning_rate": 0.0006,
+ "loss": 4.71478271484375,
+ "step": 940
+ },
+ {
+ "epoch": 13.069868995633188,
+ "grad_norm": 0.021091405302286148,
+ "learning_rate": 0.0006,
+ "loss": 4.695915222167969,
+ "step": 941
+ },
+ {
+ "epoch": 13.083842794759825,
+ "grad_norm": 0.019787423312664032,
+ "learning_rate": 0.0006,
+ "loss": 4.8652448654174805,
+ "step": 942
+ },
+ {
+ "epoch": 13.097816593886463,
+ "grad_norm": 0.0185833927243948,
+ "learning_rate": 0.0006,
+ "loss": 4.7065887451171875,
+ "step": 943
+ },
+ {
+ "epoch": 13.1117903930131,
+ "grad_norm": 0.021220475435256958,
+ "learning_rate": 0.0006,
+ "loss": 4.741232395172119,
+ "step": 944
+ },
+ {
+ "epoch": 13.125764192139737,
+ "grad_norm": 0.02136622555553913,
+ "learning_rate": 0.0006,
+ "loss": 4.705967903137207,
+ "step": 945
+ },
+ {
+ "epoch": 13.139737991266376,
+ "grad_norm": 0.0191587433218956,
+ "learning_rate": 0.0006,
+ "loss": 4.758667945861816,
+ "step": 946
+ },
+ {
+ "epoch": 13.153711790393013,
+ "grad_norm": 0.018290022388100624,
+ "learning_rate": 0.0006,
+ "loss": 4.760666847229004,
+ "step": 947
+ },
+ {
+ "epoch": 13.167685589519651,
+ "grad_norm": 0.021704409271478653,
+ "learning_rate": 0.0006,
+ "loss": 4.750555992126465,
+ "step": 948
+ },
+ {
+ "epoch": 13.181659388646288,
+ "grad_norm": 0.021668918430805206,
+ "learning_rate": 0.0006,
+ "loss": 4.840903282165527,
+ "step": 949
+ },
+ {
+ "epoch": 13.195633187772925,
+ "grad_norm": 0.019497141242027283,
+ "learning_rate": 0.0006,
+ "loss": 4.738969326019287,
+ "step": 950
+ },
+ {
+ "epoch": 13.209606986899564,
+ "grad_norm": 0.018500596284866333,
+ "learning_rate": 0.0006,
+ "loss": 4.75154972076416,
+ "step": 951
+ },
+ {
+ "epoch": 13.2235807860262,
+ "grad_norm": 0.01872239261865616,
+ "learning_rate": 0.0006,
+ "loss": 4.811738014221191,
+ "step": 952
+ },
+ {
+ "epoch": 13.237554585152838,
+ "grad_norm": 0.020321547985076904,
+ "learning_rate": 0.0006,
+ "loss": 4.815384387969971,
+ "step": 953
+ },
+ {
+ "epoch": 13.251528384279476,
+ "grad_norm": 0.02283575013279915,
+ "learning_rate": 0.0006,
+ "loss": 4.74582052230835,
+ "step": 954
+ },
+ {
+ "epoch": 13.265502183406113,
+ "grad_norm": 0.02474398910999298,
+ "learning_rate": 0.0006,
+ "loss": 4.738375186920166,
+ "step": 955
+ },
+ {
+ "epoch": 13.279475982532752,
+ "grad_norm": 0.02065764181315899,
+ "learning_rate": 0.0006,
+ "loss": 4.746875286102295,
+ "step": 956
+ },
+ {
+ "epoch": 13.293449781659389,
+ "grad_norm": 0.020534520968794823,
+ "learning_rate": 0.0006,
+ "loss": 4.656914710998535,
+ "step": 957
+ },
+ {
+ "epoch": 13.307423580786025,
+ "grad_norm": 0.0206378772854805,
+ "learning_rate": 0.0006,
+ "loss": 4.842615604400635,
+ "step": 958
+ },
+ {
+ "epoch": 13.321397379912664,
+ "grad_norm": 0.017935654148459435,
+ "learning_rate": 0.0006,
+ "loss": 4.717827796936035,
+ "step": 959
+ },
+ {
+ "epoch": 13.335371179039301,
+ "grad_norm": 0.016958363354206085,
+ "learning_rate": 0.0006,
+ "loss": 4.681070327758789,
+ "step": 960
+ },
+ {
+ "epoch": 13.34934497816594,
+ "grad_norm": 0.015735412016510963,
+ "learning_rate": 0.0006,
+ "loss": 4.7957048416137695,
+ "step": 961
+ },
+ {
+ "epoch": 13.363318777292577,
+ "grad_norm": 0.01563699170947075,
+ "learning_rate": 0.0006,
+ "loss": 4.701192378997803,
+ "step": 962
+ },
+ {
+ "epoch": 13.377292576419213,
+ "grad_norm": 0.015207760035991669,
+ "learning_rate": 0.0006,
+ "loss": 4.73256254196167,
+ "step": 963
+ },
+ {
+ "epoch": 13.391266375545852,
+ "grad_norm": 0.015396883711218834,
+ "learning_rate": 0.0006,
+ "loss": 4.76412296295166,
+ "step": 964
+ },
+ {
+ "epoch": 13.405240174672489,
+ "grad_norm": 0.015690583735704422,
+ "learning_rate": 0.0006,
+ "loss": 4.791157245635986,
+ "step": 965
+ },
+ {
+ "epoch": 13.419213973799126,
+ "grad_norm": 0.01755724661052227,
+ "learning_rate": 0.0006,
+ "loss": 4.845666885375977,
+ "step": 966
+ },
+ {
+ "epoch": 13.433187772925764,
+ "grad_norm": 0.019443422555923462,
+ "learning_rate": 0.0006,
+ "loss": 4.5902299880981445,
+ "step": 967
+ },
+ {
+ "epoch": 13.447161572052401,
+ "grad_norm": 0.02118338644504547,
+ "learning_rate": 0.0006,
+ "loss": 4.740627288818359,
+ "step": 968
+ },
+ {
+ "epoch": 13.46113537117904,
+ "grad_norm": 0.02174839936196804,
+ "learning_rate": 0.0006,
+ "loss": 4.814560890197754,
+ "step": 969
+ },
+ {
+ "epoch": 13.475109170305677,
+ "grad_norm": 0.022860374301671982,
+ "learning_rate": 0.0006,
+ "loss": 4.66440486907959,
+ "step": 970
+ },
+ {
+ "epoch": 13.489082969432314,
+ "grad_norm": 0.024742677807807922,
+ "learning_rate": 0.0006,
+ "loss": 4.7451653480529785,
+ "step": 971
+ },
+ {
+ "epoch": 13.503056768558952,
+ "grad_norm": 0.027068907395005226,
+ "learning_rate": 0.0006,
+ "loss": 4.802584171295166,
+ "step": 972
+ },
+ {
+ "epoch": 13.51703056768559,
+ "grad_norm": 0.026316625997424126,
+ "learning_rate": 0.0006,
+ "loss": 4.842397689819336,
+ "step": 973
+ },
+ {
+ "epoch": 13.531004366812226,
+ "grad_norm": 0.024610962718725204,
+ "learning_rate": 0.0006,
+ "loss": 4.650721073150635,
+ "step": 974
+ },
+ {
+ "epoch": 13.544978165938865,
+ "grad_norm": 0.024122603237628937,
+ "learning_rate": 0.0006,
+ "loss": 4.6984357833862305,
+ "step": 975
+ },
+ {
+ "epoch": 13.558951965065502,
+ "grad_norm": 0.022699708119034767,
+ "learning_rate": 0.0006,
+ "loss": 4.793007850646973,
+ "step": 976
+ },
+ {
+ "epoch": 13.57292576419214,
+ "grad_norm": 0.021706534549593925,
+ "learning_rate": 0.0006,
+ "loss": 4.691362380981445,
+ "step": 977
+ },
+ {
+ "epoch": 13.586899563318777,
+ "grad_norm": 0.021363291889429092,
+ "learning_rate": 0.0006,
+ "loss": 4.76762580871582,
+ "step": 978
+ },
+ {
+ "epoch": 13.600873362445414,
+ "grad_norm": 0.023476677015423775,
+ "learning_rate": 0.0006,
+ "loss": 4.777202606201172,
+ "step": 979
+ },
+ {
+ "epoch": 13.614847161572053,
+ "grad_norm": 0.02457100711762905,
+ "learning_rate": 0.0006,
+ "loss": 4.590703010559082,
+ "step": 980
+ },
+ {
+ "epoch": 13.62882096069869,
+ "grad_norm": 0.02469099499285221,
+ "learning_rate": 0.0006,
+ "loss": 4.640185356140137,
+ "step": 981
+ },
+ {
+ "epoch": 13.642794759825328,
+ "grad_norm": 0.024914603680372238,
+ "learning_rate": 0.0006,
+ "loss": 4.772201061248779,
+ "step": 982
+ },
+ {
+ "epoch": 13.656768558951965,
+ "grad_norm": 0.025133173912763596,
+ "learning_rate": 0.0006,
+ "loss": 4.743337154388428,
+ "step": 983
+ },
+ {
+ "epoch": 13.670742358078602,
+ "grad_norm": 0.02034137211740017,
+ "learning_rate": 0.0006,
+ "loss": 4.752127647399902,
+ "step": 984
+ },
+ {
+ "epoch": 13.68471615720524,
+ "grad_norm": 0.019645661115646362,
+ "learning_rate": 0.0006,
+ "loss": 4.838453769683838,
+ "step": 985
+ },
+ {
+ "epoch": 13.698689956331878,
+ "grad_norm": 0.01900297962129116,
+ "learning_rate": 0.0006,
+ "loss": 4.804868221282959,
+ "step": 986
+ },
+ {
+ "epoch": 13.712663755458514,
+ "grad_norm": 0.020292341709136963,
+ "learning_rate": 0.0006,
+ "loss": 4.609743118286133,
+ "step": 987
+ },
+ {
+ "epoch": 13.726637554585153,
+ "grad_norm": 0.017014775425195694,
+ "learning_rate": 0.0006,
+ "loss": 4.740612983703613,
+ "step": 988
+ },
+ {
+ "epoch": 13.74061135371179,
+ "grad_norm": 0.015199844725430012,
+ "learning_rate": 0.0006,
+ "loss": 4.719701766967773,
+ "step": 989
+ },
+ {
+ "epoch": 13.754585152838429,
+ "grad_norm": 0.015125414356589317,
+ "learning_rate": 0.0006,
+ "loss": 4.686114311218262,
+ "step": 990
+ },
+ {
+ "epoch": 13.768558951965066,
+ "grad_norm": 0.014779540710151196,
+ "learning_rate": 0.0006,
+ "loss": 4.674347877502441,
+ "step": 991
+ },
+ {
+ "epoch": 13.782532751091702,
+ "grad_norm": 0.015255163423717022,
+ "learning_rate": 0.0006,
+ "loss": 4.6309309005737305,
+ "step": 992
+ },
+ {
+ "epoch": 13.796506550218341,
+ "grad_norm": 0.013692197389900684,
+ "learning_rate": 0.0006,
+ "loss": 4.698497295379639,
+ "step": 993
+ },
+ {
+ "epoch": 13.810480349344978,
+ "grad_norm": 0.014529623091220856,
+ "learning_rate": 0.0006,
+ "loss": 4.628443717956543,
+ "step": 994
+ },
+ {
+ "epoch": 13.824454148471617,
+ "grad_norm": 0.01576659083366394,
+ "learning_rate": 0.0006,
+ "loss": 4.788015842437744,
+ "step": 995
+ },
+ {
+ "epoch": 13.838427947598253,
+ "grad_norm": 0.015834277495741844,
+ "learning_rate": 0.0006,
+ "loss": 4.672703742980957,
+ "step": 996
+ },
+ {
+ "epoch": 13.85240174672489,
+ "grad_norm": 0.015539503656327724,
+ "learning_rate": 0.0006,
+ "loss": 4.731503486633301,
+ "step": 997
+ },
+ {
+ "epoch": 13.866375545851529,
+ "grad_norm": 0.013463746756315231,
+ "learning_rate": 0.0006,
+ "loss": 4.7006330490112305,
+ "step": 998
+ },
+ {
+ "epoch": 13.880349344978166,
+ "grad_norm": 0.016377726569771767,
+ "learning_rate": 0.0006,
+ "loss": 4.636074066162109,
+ "step": 999
+ },
+ {
+ "epoch": 13.894323144104803,
+ "grad_norm": 0.016660034656524658,
+ "learning_rate": 0.0006,
+ "loss": 4.742693901062012,
+ "step": 1000
+ },
+ {
+ "epoch": 13.908296943231441,
+ "grad_norm": 0.014902938157320023,
+ "learning_rate": 0.0006,
+ "loss": 4.769811630249023,
+ "step": 1001
+ },
+ {
+ "epoch": 13.922270742358078,
+ "grad_norm": 0.014883476309478283,
+ "learning_rate": 0.0006,
+ "loss": 4.731159210205078,
+ "step": 1002
+ },
+ {
+ "epoch": 13.936244541484717,
+ "grad_norm": 0.014494217932224274,
+ "learning_rate": 0.0006,
+ "loss": 4.659485816955566,
+ "step": 1003
+ },
+ {
+ "epoch": 13.950218340611354,
+ "grad_norm": 0.015426991507411003,
+ "learning_rate": 0.0006,
+ "loss": 4.755229949951172,
+ "step": 1004
+ },
+ {
+ "epoch": 13.96419213973799,
+ "grad_norm": 0.016648586839437485,
+ "learning_rate": 0.0006,
+ "loss": 4.74111270904541,
+ "step": 1005
+ },
+ {
+ "epoch": 13.97816593886463,
+ "grad_norm": 0.016980541869997978,
+ "learning_rate": 0.0006,
+ "loss": 4.74072265625,
+ "step": 1006
+ },
+ {
+ "epoch": 13.992139737991266,
+ "grad_norm": 0.018070152029395103,
+ "learning_rate": 0.0006,
+ "loss": 4.732975006103516,
+ "step": 1007
+ },
+ {
+ "epoch": 14.0,
+ "grad_norm": 0.019910665228962898,
+ "learning_rate": 0.0006,
+ "loss": 4.634697914123535,
+ "step": 1008
+ },
+ {
+ "epoch": 14.0,
+ "eval_loss": 4.850640296936035,
+ "eval_runtime": 57.3039,
+ "eval_samples_per_second": 42.615,
+ "eval_steps_per_second": 1.344,
+ "step": 1008
+ },
+ {
+ "epoch": 14.013973799126637,
+ "grad_norm": 0.01980053074657917,
+ "learning_rate": 0.0006,
+ "loss": 4.628744602203369,
+ "step": 1009
+ },
+ {
+ "epoch": 14.027947598253276,
+ "grad_norm": 0.02497623674571514,
+ "learning_rate": 0.0006,
+ "loss": 4.75930118560791,
+ "step": 1010
+ },
+ {
+ "epoch": 14.041921397379912,
+ "grad_norm": 0.02794097363948822,
+ "learning_rate": 0.0006,
+ "loss": 4.691340446472168,
+ "step": 1011
+ },
+ {
+ "epoch": 14.055895196506551,
+ "grad_norm": 0.023725535720586777,
+ "learning_rate": 0.0006,
+ "loss": 4.733664512634277,
+ "step": 1012
+ },
+ {
+ "epoch": 14.069868995633188,
+ "grad_norm": 0.022966474294662476,
+ "learning_rate": 0.0006,
+ "loss": 4.664808750152588,
+ "step": 1013
+ },
+ {
+ "epoch": 14.083842794759825,
+ "grad_norm": 0.02530563250184059,
+ "learning_rate": 0.0006,
+ "loss": 4.8047285079956055,
+ "step": 1014
+ },
+ {
+ "epoch": 14.097816593886463,
+ "grad_norm": 0.02491082064807415,
+ "learning_rate": 0.0006,
+ "loss": 4.667956352233887,
+ "step": 1015
+ },
+ {
+ "epoch": 14.1117903930131,
+ "grad_norm": 0.022110294550657272,
+ "learning_rate": 0.0006,
+ "loss": 4.666140556335449,
+ "step": 1016
+ },
+ {
+ "epoch": 14.125764192139737,
+ "grad_norm": 0.02065849117934704,
+ "learning_rate": 0.0006,
+ "loss": 4.581960678100586,
+ "step": 1017
+ },
+ {
+ "epoch": 14.139737991266376,
+ "grad_norm": 0.01941298507153988,
+ "learning_rate": 0.0006,
+ "loss": 4.614226341247559,
+ "step": 1018
+ },
+ {
+ "epoch": 14.153711790393013,
+ "grad_norm": 0.022441856563091278,
+ "learning_rate": 0.0006,
+ "loss": 4.638352394104004,
+ "step": 1019
+ },
+ {
+ "epoch": 14.167685589519651,
+ "grad_norm": 0.021519694477319717,
+ "learning_rate": 0.0006,
+ "loss": 4.700160980224609,
+ "step": 1020
+ },
+ {
+ "epoch": 14.181659388646288,
+ "grad_norm": 0.018956486135721207,
+ "learning_rate": 0.0006,
+ "loss": 4.649593353271484,
+ "step": 1021
+ },
+ {
+ "epoch": 14.195633187772925,
+ "grad_norm": 0.016681140288710594,
+ "learning_rate": 0.0006,
+ "loss": 4.603952407836914,
+ "step": 1022
+ },
+ {
+ "epoch": 14.209606986899564,
+ "grad_norm": 0.01788809522986412,
+ "learning_rate": 0.0006,
+ "loss": 4.669160842895508,
+ "step": 1023
+ },
+ {
+ "epoch": 14.2235807860262,
+ "grad_norm": 0.016047228127717972,
+ "learning_rate": 0.0006,
+ "loss": 4.6367950439453125,
+ "step": 1024
+ },
+ {
+ "epoch": 14.237554585152838,
+ "grad_norm": 0.01614239625632763,
+ "learning_rate": 0.0006,
+ "loss": 4.664529323577881,
+ "step": 1025
+ },
+ {
+ "epoch": 14.251528384279476,
+ "grad_norm": 0.016362549737095833,
+ "learning_rate": 0.0006,
+ "loss": 4.702373504638672,
+ "step": 1026
+ },
+ {
+ "epoch": 14.265502183406113,
+ "grad_norm": 0.015800409018993378,
+ "learning_rate": 0.0006,
+ "loss": 4.7594428062438965,
+ "step": 1027
+ },
+ {
+ "epoch": 14.279475982532752,
+ "grad_norm": 0.016730139032006264,
+ "learning_rate": 0.0006,
+ "loss": 4.677244186401367,
+ "step": 1028
+ },
+ {
+ "epoch": 14.293449781659389,
+ "grad_norm": 0.01661759801208973,
+ "learning_rate": 0.0006,
+ "loss": 4.633679389953613,
+ "step": 1029
+ },
+ {
+ "epoch": 14.307423580786025,
+ "grad_norm": 0.017673347145318985,
+ "learning_rate": 0.0006,
+ "loss": 4.649107456207275,
+ "step": 1030
+ },
+ {
+ "epoch": 14.321397379912664,
+ "grad_norm": 0.019243352115154266,
+ "learning_rate": 0.0006,
+ "loss": 4.611257553100586,
+ "step": 1031
+ },
+ {
+ "epoch": 14.335371179039301,
+ "grad_norm": 0.019817164167761803,
+ "learning_rate": 0.0006,
+ "loss": 4.588086128234863,
+ "step": 1032
+ },
+ {
+ "epoch": 14.34934497816594,
+ "grad_norm": 0.022292211651802063,
+ "learning_rate": 0.0006,
+ "loss": 4.783066749572754,
+ "step": 1033
+ },
+ {
+ "epoch": 14.363318777292577,
+ "grad_norm": 0.02140096202492714,
+ "learning_rate": 0.0006,
+ "loss": 4.604718208312988,
+ "step": 1034
+ },
+ {
+ "epoch": 14.377292576419213,
+ "grad_norm": 0.0187817495316267,
+ "learning_rate": 0.0006,
+ "loss": 4.721773147583008,
+ "step": 1035
+ },
+ {
+ "epoch": 14.391266375545852,
+ "grad_norm": 0.019879210740327835,
+ "learning_rate": 0.0006,
+ "loss": 4.789286136627197,
+ "step": 1036
+ },
+ {
+ "epoch": 14.405240174672489,
+ "grad_norm": 0.01885785162448883,
+ "learning_rate": 0.0006,
+ "loss": 4.669743537902832,
+ "step": 1037
+ },
+ {
+ "epoch": 14.419213973799126,
+ "grad_norm": 0.018324270844459534,
+ "learning_rate": 0.0006,
+ "loss": 4.65901517868042,
+ "step": 1038
+ },
+ {
+ "epoch": 14.433187772925764,
+ "grad_norm": 0.017664821818470955,
+ "learning_rate": 0.0006,
+ "loss": 4.667839527130127,
+ "step": 1039
+ },
+ {
+ "epoch": 14.447161572052401,
+ "grad_norm": 0.015702905133366585,
+ "learning_rate": 0.0006,
+ "loss": 4.64700984954834,
+ "step": 1040
+ },
+ {
+ "epoch": 14.46113537117904,
+ "grad_norm": 0.016674194484949112,
+ "learning_rate": 0.0006,
+ "loss": 4.694093227386475,
+ "step": 1041
+ },
+ {
+ "epoch": 14.475109170305677,
+ "grad_norm": 0.016869477927684784,
+ "learning_rate": 0.0006,
+ "loss": 4.619544982910156,
+ "step": 1042
+ },
+ {
+ "epoch": 14.489082969432314,
+ "grad_norm": 0.016557540744543076,
+ "learning_rate": 0.0006,
+ "loss": 4.568667411804199,
+ "step": 1043
+ },
+ {
+ "epoch": 14.503056768558952,
+ "grad_norm": 0.014500945806503296,
+ "learning_rate": 0.0006,
+ "loss": 4.561741828918457,
+ "step": 1044
+ },
+ {
+ "epoch": 14.51703056768559,
+ "grad_norm": 0.014228739775717258,
+ "learning_rate": 0.0006,
+ "loss": 4.654277324676514,
+ "step": 1045
+ },
+ {
+ "epoch": 14.531004366812226,
+ "grad_norm": 0.013894238509237766,
+ "learning_rate": 0.0006,
+ "loss": 4.5867719650268555,
+ "step": 1046
+ },
+ {
+ "epoch": 14.544978165938865,
+ "grad_norm": 0.01550068985670805,
+ "learning_rate": 0.0006,
+ "loss": 4.61168098449707,
+ "step": 1047
+ },
+ {
+ "epoch": 14.558951965065502,
+ "grad_norm": 0.015940772369503975,
+ "learning_rate": 0.0006,
+ "loss": 4.691247940063477,
+ "step": 1048
+ },
+ {
+ "epoch": 14.57292576419214,
+ "grad_norm": 0.01714535616338253,
+ "learning_rate": 0.0006,
+ "loss": 4.61183500289917,
+ "step": 1049
+ },
+ {
+ "epoch": 14.586899563318777,
+ "grad_norm": 0.017925038933753967,
+ "learning_rate": 0.0006,
+ "loss": 4.719022750854492,
+ "step": 1050
+ },
+ {
+ "epoch": 14.600873362445414,
+ "grad_norm": 0.018141552805900574,
+ "learning_rate": 0.0006,
+ "loss": 4.760281562805176,
+ "step": 1051
+ },
+ {
+ "epoch": 14.614847161572053,
+ "grad_norm": 0.017131328582763672,
+ "learning_rate": 0.0006,
+ "loss": 4.709680080413818,
+ "step": 1052
+ },
+ {
+ "epoch": 14.62882096069869,
+ "grad_norm": 0.018730048090219498,
+ "learning_rate": 0.0006,
+ "loss": 4.5994720458984375,
+ "step": 1053
+ },
+ {
+ "epoch": 14.642794759825328,
+ "grad_norm": 0.018894298002123833,
+ "learning_rate": 0.0006,
+ "loss": 4.669070720672607,
+ "step": 1054
+ },
+ {
+ "epoch": 14.656768558951965,
+ "grad_norm": 0.017643198370933533,
+ "learning_rate": 0.0006,
+ "loss": 4.641790390014648,
+ "step": 1055
+ },
+ {
+ "epoch": 14.670742358078602,
+ "grad_norm": 0.017647244036197662,
+ "learning_rate": 0.0006,
+ "loss": 4.653899192810059,
+ "step": 1056
+ },
+ {
+ "epoch": 14.68471615720524,
+ "grad_norm": 0.017304198816418648,
+ "learning_rate": 0.0006,
+ "loss": 4.7487473487854,
+ "step": 1057
+ },
+ {
+ "epoch": 14.698689956331878,
+ "grad_norm": 0.01638447679579258,
+ "learning_rate": 0.0006,
+ "loss": 4.71945858001709,
+ "step": 1058
+ },
+ {
+ "epoch": 14.712663755458514,
+ "grad_norm": 0.015892071649432182,
+ "learning_rate": 0.0006,
+ "loss": 4.744903564453125,
+ "step": 1059
+ },
+ {
+ "epoch": 14.726637554585153,
+ "grad_norm": 0.017420614138245583,
+ "learning_rate": 0.0006,
+ "loss": 4.781563758850098,
+ "step": 1060
+ },
+ {
+ "epoch": 14.74061135371179,
+ "grad_norm": 0.016498520970344543,
+ "learning_rate": 0.0006,
+ "loss": 4.622779846191406,
+ "step": 1061
+ },
+ {
+ "epoch": 14.754585152838429,
+ "grad_norm": 0.017482127994298935,
+ "learning_rate": 0.0006,
+ "loss": 4.62825870513916,
+ "step": 1062
+ },
+ {
+ "epoch": 14.768558951965066,
+ "grad_norm": 0.019018379971385002,
+ "learning_rate": 0.0006,
+ "loss": 4.693194389343262,
+ "step": 1063
+ },
+ {
+ "epoch": 14.782532751091702,
+ "grad_norm": 0.01993914321064949,
+ "learning_rate": 0.0006,
+ "loss": 4.742903709411621,
+ "step": 1064
+ },
+ {
+ "epoch": 14.796506550218341,
+ "grad_norm": 0.02301335707306862,
+ "learning_rate": 0.0006,
+ "loss": 4.613251686096191,
+ "step": 1065
+ },
+ {
+ "epoch": 14.810480349344978,
+ "grad_norm": 0.024088222533464432,
+ "learning_rate": 0.0006,
+ "loss": 4.594305992126465,
+ "step": 1066
+ },
+ {
+ "epoch": 14.824454148471617,
+ "grad_norm": 0.02493320405483246,
+ "learning_rate": 0.0006,
+ "loss": 4.700588226318359,
+ "step": 1067
+ },
+ {
+ "epoch": 14.838427947598253,
+ "grad_norm": 0.02501937560737133,
+ "learning_rate": 0.0006,
+ "loss": 4.601742744445801,
+ "step": 1068
+ },
+ {
+ "epoch": 14.85240174672489,
+ "grad_norm": 0.023490751162171364,
+ "learning_rate": 0.0006,
+ "loss": 4.589211463928223,
+ "step": 1069
+ },
+ {
+ "epoch": 14.866375545851529,
+ "grad_norm": 0.019695665687322617,
+ "learning_rate": 0.0006,
+ "loss": 4.635081768035889,
+ "step": 1070
+ },
+ {
+ "epoch": 14.880349344978166,
+ "grad_norm": 0.019276980310678482,
+ "learning_rate": 0.0006,
+ "loss": 4.737002372741699,
+ "step": 1071
+ },
+ {
+ "epoch": 14.894323144104803,
+ "grad_norm": 0.022430511191487312,
+ "learning_rate": 0.0006,
+ "loss": 4.808295726776123,
+ "step": 1072
+ },
+ {
+ "epoch": 14.908296943231441,
+ "grad_norm": 0.021543916314840317,
+ "learning_rate": 0.0006,
+ "loss": 4.559475421905518,
+ "step": 1073
+ },
+ {
+ "epoch": 14.922270742358078,
+ "grad_norm": 0.019885241985321045,
+ "learning_rate": 0.0006,
+ "loss": 4.647948741912842,
+ "step": 1074
+ },
+ {
+ "epoch": 14.936244541484717,
+ "grad_norm": 0.02115175500512123,
+ "learning_rate": 0.0006,
+ "loss": 4.680505752563477,
+ "step": 1075
+ },
+ {
+ "epoch": 14.950218340611354,
+ "grad_norm": 0.02156151458621025,
+ "learning_rate": 0.0006,
+ "loss": 4.673477649688721,
+ "step": 1076
+ },
+ {
+ "epoch": 14.96419213973799,
+ "grad_norm": 0.02184601128101349,
+ "learning_rate": 0.0006,
+ "loss": 4.644402027130127,
+ "step": 1077
+ },
+ {
+ "epoch": 14.97816593886463,
+ "grad_norm": 0.0211983360350132,
+ "learning_rate": 0.0006,
+ "loss": 4.821558475494385,
+ "step": 1078
+ },
+ {
+ "epoch": 14.992139737991266,
+ "grad_norm": 0.019636567682027817,
+ "learning_rate": 0.0006,
+ "loss": 4.578868865966797,
+ "step": 1079
+ },
+ {
+ "epoch": 15.0,
+ "grad_norm": 0.019997350871562958,
+ "learning_rate": 0.0006,
+ "loss": 4.691307544708252,
+ "step": 1080
+ },
+ {
+ "epoch": 15.0,
+ "eval_loss": 4.872478008270264,
+ "eval_runtime": 56.8513,
+ "eval_samples_per_second": 42.954,
+ "eval_steps_per_second": 1.354,
+ "step": 1080
+ },
+ {
+ "epoch": 15.013973799126637,
+ "grad_norm": 0.018830662593245506,
+ "learning_rate": 0.0006,
+ "loss": 4.761779308319092,
+ "step": 1081
+ },
+ {
+ "epoch": 15.027947598253276,
+ "grad_norm": 0.016876481473445892,
+ "learning_rate": 0.0006,
+ "loss": 4.61148738861084,
+ "step": 1082
+ },
+ {
+ "epoch": 15.041921397379912,
+ "grad_norm": 0.015550863929092884,
+ "learning_rate": 0.0006,
+ "loss": 4.676124095916748,
+ "step": 1083
+ },
+ {
+ "epoch": 15.055895196506551,
+ "grad_norm": 0.017178896814584732,
+ "learning_rate": 0.0006,
+ "loss": 4.588935852050781,
+ "step": 1084
+ },
+ {
+ "epoch": 15.069868995633188,
+ "grad_norm": 0.015465447679162025,
+ "learning_rate": 0.0006,
+ "loss": 4.605499267578125,
+ "step": 1085
+ },
+ {
+ "epoch": 15.083842794759825,
+ "grad_norm": 0.01606195978820324,
+ "learning_rate": 0.0006,
+ "loss": 4.723373889923096,
+ "step": 1086
+ },
+ {
+ "epoch": 15.097816593886463,
+ "grad_norm": 0.016521936282515526,
+ "learning_rate": 0.0006,
+ "loss": 4.646608352661133,
+ "step": 1087
+ },
+ {
+ "epoch": 15.1117903930131,
+ "grad_norm": 0.01589205674827099,
+ "learning_rate": 0.0006,
+ "loss": 4.498138427734375,
+ "step": 1088
+ },
+ {
+ "epoch": 15.125764192139737,
+ "grad_norm": 0.01524856686592102,
+ "learning_rate": 0.0006,
+ "loss": 4.594562530517578,
+ "step": 1089
+ },
+ {
+ "epoch": 15.139737991266376,
+ "grad_norm": 0.014714641496539116,
+ "learning_rate": 0.0006,
+ "loss": 4.630744934082031,
+ "step": 1090
+ },
+ {
+ "epoch": 15.153711790393013,
+ "grad_norm": 0.015333331190049648,
+ "learning_rate": 0.0006,
+ "loss": 4.682388782501221,
+ "step": 1091
+ },
+ {
+ "epoch": 15.167685589519651,
+ "grad_norm": 0.01550073828548193,
+ "learning_rate": 0.0006,
+ "loss": 4.692652702331543,
+ "step": 1092
+ },
+ {
+ "epoch": 15.181659388646288,
+ "grad_norm": 0.014783318154513836,
+ "learning_rate": 0.0006,
+ "loss": 4.6223835945129395,
+ "step": 1093
+ },
+ {
+ "epoch": 15.195633187772925,
+ "grad_norm": 0.014372751116752625,
+ "learning_rate": 0.0006,
+ "loss": 4.52687406539917,
+ "step": 1094
+ },
+ {
+ "epoch": 15.209606986899564,
+ "grad_norm": 0.014779130928218365,
+ "learning_rate": 0.0006,
+ "loss": 4.637009620666504,
+ "step": 1095
+ },
+ {
+ "epoch": 15.2235807860262,
+ "grad_norm": 0.015697073191404343,
+ "learning_rate": 0.0006,
+ "loss": 4.682981014251709,
+ "step": 1096
+ },
+ {
+ "epoch": 15.237554585152838,
+ "grad_norm": 0.016523664817214012,
+ "learning_rate": 0.0006,
+ "loss": 4.705722808837891,
+ "step": 1097
+ },
+ {
+ "epoch": 15.251528384279476,
+ "grad_norm": 0.017633996903896332,
+ "learning_rate": 0.0006,
+ "loss": 4.574686050415039,
+ "step": 1098
+ },
+ {
+ "epoch": 15.265502183406113,
+ "grad_norm": 0.019955076277256012,
+ "learning_rate": 0.0006,
+ "loss": 4.634450912475586,
+ "step": 1099
+ },
+ {
+ "epoch": 15.279475982532752,
+ "grad_norm": 0.02132618986070156,
+ "learning_rate": 0.0006,
+ "loss": 4.56110143661499,
+ "step": 1100
+ },
+ {
+ "epoch": 15.293449781659389,
+ "grad_norm": 0.0207593385130167,
+ "learning_rate": 0.0006,
+ "loss": 4.629396438598633,
+ "step": 1101
+ },
+ {
+ "epoch": 15.307423580786025,
+ "grad_norm": 0.017882540822029114,
+ "learning_rate": 0.0006,
+ "loss": 4.730963230133057,
+ "step": 1102
+ },
+ {
+ "epoch": 15.321397379912664,
+ "grad_norm": 0.01917876861989498,
+ "learning_rate": 0.0006,
+ "loss": 4.64585542678833,
+ "step": 1103
+ },
+ {
+ "epoch": 15.335371179039301,
+ "grad_norm": 0.01943974196910858,
+ "learning_rate": 0.0006,
+ "loss": 4.656181335449219,
+ "step": 1104
+ },
+ {
+ "epoch": 15.34934497816594,
+ "grad_norm": 0.019904715940356255,
+ "learning_rate": 0.0006,
+ "loss": 4.619329452514648,
+ "step": 1105
+ },
+ {
+ "epoch": 15.363318777292577,
+ "grad_norm": 0.019174346700310707,
+ "learning_rate": 0.0006,
+ "loss": 4.647892475128174,
+ "step": 1106
+ },
+ {
+ "epoch": 15.377292576419213,
+ "grad_norm": 0.018244758248329163,
+ "learning_rate": 0.0006,
+ "loss": 4.56166934967041,
+ "step": 1107
+ },
+ {
+ "epoch": 15.391266375545852,
+ "grad_norm": 0.018914205953478813,
+ "learning_rate": 0.0006,
+ "loss": 4.648063659667969,
+ "step": 1108
+ },
+ {
+ "epoch": 15.405240174672489,
+ "grad_norm": 0.01961071416735649,
+ "learning_rate": 0.0006,
+ "loss": 4.665525436401367,
+ "step": 1109
+ },
+ {
+ "epoch": 15.419213973799126,
+ "grad_norm": 0.01841311901807785,
+ "learning_rate": 0.0006,
+ "loss": 4.611342430114746,
+ "step": 1110
+ },
+ {
+ "epoch": 15.433187772925764,
+ "grad_norm": 0.01863427273929119,
+ "learning_rate": 0.0006,
+ "loss": 4.664256572723389,
+ "step": 1111
+ },
+ {
+ "epoch": 15.447161572052401,
+ "grad_norm": 0.020506877452135086,
+ "learning_rate": 0.0006,
+ "loss": 4.631237506866455,
+ "step": 1112
+ },
+ {
+ "epoch": 15.46113537117904,
+ "grad_norm": 0.02223833277821541,
+ "learning_rate": 0.0006,
+ "loss": 4.668654441833496,
+ "step": 1113
+ },
+ {
+ "epoch": 15.475109170305677,
+ "grad_norm": 0.023336702957749367,
+ "learning_rate": 0.0006,
+ "loss": 4.562833309173584,
+ "step": 1114
+ },
+ {
+ "epoch": 15.489082969432314,
+ "grad_norm": 0.023052413016557693,
+ "learning_rate": 0.0006,
+ "loss": 4.621532917022705,
+ "step": 1115
+ },
+ {
+ "epoch": 15.503056768558952,
+ "grad_norm": 0.023420924320816994,
+ "learning_rate": 0.0006,
+ "loss": 4.665465354919434,
+ "step": 1116
+ },
+ {
+ "epoch": 15.51703056768559,
+ "grad_norm": 0.024071281775832176,
+ "learning_rate": 0.0006,
+ "loss": 4.595980644226074,
+ "step": 1117
+ },
+ {
+ "epoch": 15.531004366812226,
+ "grad_norm": 0.023406682536005974,
+ "learning_rate": 0.0006,
+ "loss": 4.642635345458984,
+ "step": 1118
+ },
+ {
+ "epoch": 15.544978165938865,
+ "grad_norm": 0.02138647809624672,
+ "learning_rate": 0.0006,
+ "loss": 4.704513072967529,
+ "step": 1119
+ },
+ {
+ "epoch": 15.558951965065502,
+ "grad_norm": 0.0221384447067976,
+ "learning_rate": 0.0006,
+ "loss": 4.567503452301025,
+ "step": 1120
+ },
+ {
+ "epoch": 15.57292576419214,
+ "grad_norm": 0.020879851654171944,
+ "learning_rate": 0.0006,
+ "loss": 4.664057731628418,
+ "step": 1121
+ },
+ {
+ "epoch": 15.586899563318777,
+ "grad_norm": 0.021478967741131783,
+ "learning_rate": 0.0006,
+ "loss": 4.539627552032471,
+ "step": 1122
+ },
+ {
+ "epoch": 15.600873362445414,
+ "grad_norm": 0.022748656570911407,
+ "learning_rate": 0.0006,
+ "loss": 4.5654497146606445,
+ "step": 1123
+ },
+ {
+ "epoch": 15.614847161572053,
+ "grad_norm": 0.021870248019695282,
+ "learning_rate": 0.0006,
+ "loss": 4.56899356842041,
+ "step": 1124
+ },
+ {
+ "epoch": 15.62882096069869,
+ "grad_norm": 0.021711355075240135,
+ "learning_rate": 0.0006,
+ "loss": 4.657808303833008,
+ "step": 1125
+ },
+ {
+ "epoch": 15.642794759825328,
+ "grad_norm": 0.022810909897089005,
+ "learning_rate": 0.0006,
+ "loss": 4.6526689529418945,
+ "step": 1126
+ },
+ {
+ "epoch": 15.656768558951965,
+ "grad_norm": 0.022561343386769295,
+ "learning_rate": 0.0006,
+ "loss": 4.534887790679932,
+ "step": 1127
+ },
+ {
+ "epoch": 15.670742358078602,
+ "grad_norm": 0.019909802824258804,
+ "learning_rate": 0.0006,
+ "loss": 4.600790023803711,
+ "step": 1128
+ },
+ {
+ "epoch": 15.68471615720524,
+ "grad_norm": 0.020092325285077095,
+ "learning_rate": 0.0006,
+ "loss": 4.773830413818359,
+ "step": 1129
+ },
+ {
+ "epoch": 15.698689956331878,
+ "grad_norm": 0.018834445625543594,
+ "learning_rate": 0.0006,
+ "loss": 4.628868103027344,
+ "step": 1130
+ },
+ {
+ "epoch": 15.712663755458514,
+ "grad_norm": 0.017552798613905907,
+ "learning_rate": 0.0006,
+ "loss": 4.647500038146973,
+ "step": 1131
+ },
+ {
+ "epoch": 15.726637554585153,
+ "grad_norm": 0.01751251146197319,
+ "learning_rate": 0.0006,
+ "loss": 4.590545654296875,
+ "step": 1132
+ },
+ {
+ "epoch": 15.74061135371179,
+ "grad_norm": 0.01699099875986576,
+ "learning_rate": 0.0006,
+ "loss": 4.509699821472168,
+ "step": 1133
+ },
+ {
+ "epoch": 15.754585152838429,
+ "grad_norm": 0.015808649361133575,
+ "learning_rate": 0.0006,
+ "loss": 4.606992244720459,
+ "step": 1134
+ },
+ {
+ "epoch": 15.768558951965066,
+ "grad_norm": 0.01655443385243416,
+ "learning_rate": 0.0006,
+ "loss": 4.585219860076904,
+ "step": 1135
+ },
+ {
+ "epoch": 15.782532751091702,
+ "grad_norm": 0.01708163693547249,
+ "learning_rate": 0.0006,
+ "loss": 4.668147087097168,
+ "step": 1136
+ },
+ {
+ "epoch": 15.796506550218341,
+ "grad_norm": 0.017161913216114044,
+ "learning_rate": 0.0006,
+ "loss": 4.7170634269714355,
+ "step": 1137
+ },
+ {
+ "epoch": 15.810480349344978,
+ "grad_norm": 0.017262674868106842,
+ "learning_rate": 0.0006,
+ "loss": 4.572132110595703,
+ "step": 1138
+ },
+ {
+ "epoch": 15.824454148471617,
+ "grad_norm": 0.017713576555252075,
+ "learning_rate": 0.0006,
+ "loss": 4.702538967132568,
+ "step": 1139
+ },
+ {
+ "epoch": 15.838427947598253,
+ "grad_norm": 0.018730733543634415,
+ "learning_rate": 0.0006,
+ "loss": 4.629721641540527,
+ "step": 1140
+ },
+ {
+ "epoch": 15.85240174672489,
+ "grad_norm": 0.017960187047719955,
+ "learning_rate": 0.0006,
+ "loss": 4.661039352416992,
+ "step": 1141
+ },
+ {
+ "epoch": 15.866375545851529,
+ "grad_norm": 0.01750079356133938,
+ "learning_rate": 0.0006,
+ "loss": 4.503710746765137,
+ "step": 1142
+ },
+ {
+ "epoch": 15.880349344978166,
+ "grad_norm": 0.016630422323942184,
+ "learning_rate": 0.0006,
+ "loss": 4.6582136154174805,
+ "step": 1143
+ },
+ {
+ "epoch": 15.894323144104803,
+ "grad_norm": 0.015896832570433617,
+ "learning_rate": 0.0006,
+ "loss": 4.596693992614746,
+ "step": 1144
+ },
+ {
+ "epoch": 15.908296943231441,
+ "grad_norm": 0.01566407084465027,
+ "learning_rate": 0.0006,
+ "loss": 4.6337199211120605,
+ "step": 1145
+ },
+ {
+ "epoch": 15.922270742358078,
+ "grad_norm": 0.016823330894112587,
+ "learning_rate": 0.0006,
+ "loss": 4.686994552612305,
+ "step": 1146
+ },
+ {
+ "epoch": 15.936244541484717,
+ "grad_norm": 0.016208255663514137,
+ "learning_rate": 0.0006,
+ "loss": 4.6205549240112305,
+ "step": 1147
+ },
+ {
+ "epoch": 15.950218340611354,
+ "grad_norm": 0.0171707421541214,
+ "learning_rate": 0.0006,
+ "loss": 4.582968711853027,
+ "step": 1148
+ },
+ {
+ "epoch": 15.96419213973799,
+ "grad_norm": 0.017009446397423744,
+ "learning_rate": 0.0006,
+ "loss": 4.510274887084961,
+ "step": 1149
+ },
+ {
+ "epoch": 15.97816593886463,
+ "grad_norm": 0.01565811224281788,
+ "learning_rate": 0.0006,
+ "loss": 4.55834436416626,
+ "step": 1150
+ },
+ {
+ "epoch": 15.992139737991266,
+ "grad_norm": 0.015276037156581879,
+ "learning_rate": 0.0006,
+ "loss": 4.537870407104492,
+ "step": 1151
+ },
+ {
+ "epoch": 16.0,
+ "grad_norm": 0.017305882647633553,
+ "learning_rate": 0.0006,
+ "loss": 4.611305236816406,
+ "step": 1152
+ },
+ {
+ "epoch": 16.0,
+ "eval_loss": 4.775513172149658,
+ "eval_runtime": 56.7341,
+ "eval_samples_per_second": 43.043,
+ "eval_steps_per_second": 1.357,
+ "step": 1152
+ },
+ {
+ "epoch": 16.01397379912664,
+ "grad_norm": 0.016752436757087708,
+ "learning_rate": 0.0006,
+ "loss": 4.541312217712402,
+ "step": 1153
+ },
+ {
+ "epoch": 16.027947598253274,
+ "grad_norm": 0.01643134094774723,
+ "learning_rate": 0.0006,
+ "loss": 4.585402965545654,
+ "step": 1154
+ },
+ {
+ "epoch": 16.041921397379912,
+ "grad_norm": 0.01888720504939556,
+ "learning_rate": 0.0006,
+ "loss": 4.597825050354004,
+ "step": 1155
+ },
+ {
+ "epoch": 16.05589519650655,
+ "grad_norm": 0.022044632583856583,
+ "learning_rate": 0.0006,
+ "loss": 4.532772064208984,
+ "step": 1156
+ },
+ {
+ "epoch": 16.069868995633186,
+ "grad_norm": 0.022986987605690956,
+ "learning_rate": 0.0006,
+ "loss": 4.551351547241211,
+ "step": 1157
+ },
+ {
+ "epoch": 16.083842794759825,
+ "grad_norm": 0.020116791129112244,
+ "learning_rate": 0.0006,
+ "loss": 4.484149932861328,
+ "step": 1158
+ },
+ {
+ "epoch": 16.097816593886463,
+ "grad_norm": 0.021442247554659843,
+ "learning_rate": 0.0006,
+ "loss": 4.515316486358643,
+ "step": 1159
+ },
+ {
+ "epoch": 16.111790393013102,
+ "grad_norm": 0.020974772050976753,
+ "learning_rate": 0.0006,
+ "loss": 4.51606559753418,
+ "step": 1160
+ },
+ {
+ "epoch": 16.125764192139737,
+ "grad_norm": 0.021274864673614502,
+ "learning_rate": 0.0006,
+ "loss": 4.587329864501953,
+ "step": 1161
+ },
+ {
+ "epoch": 16.139737991266376,
+ "grad_norm": 0.022799784317612648,
+ "learning_rate": 0.0006,
+ "loss": 4.568149089813232,
+ "step": 1162
+ },
+ {
+ "epoch": 16.153711790393015,
+ "grad_norm": 0.021784933283925056,
+ "learning_rate": 0.0006,
+ "loss": 4.515336990356445,
+ "step": 1163
+ },
+ {
+ "epoch": 16.16768558951965,
+ "grad_norm": 0.024261271581053734,
+ "learning_rate": 0.0006,
+ "loss": 4.5127081871032715,
+ "step": 1164
+ },
+ {
+ "epoch": 16.18165938864629,
+ "grad_norm": 0.02248375490307808,
+ "learning_rate": 0.0006,
+ "loss": 4.754251003265381,
+ "step": 1165
+ },
+ {
+ "epoch": 16.195633187772927,
+ "grad_norm": 0.02136492356657982,
+ "learning_rate": 0.0006,
+ "loss": 4.58193826675415,
+ "step": 1166
+ },
+ {
+ "epoch": 16.209606986899562,
+ "grad_norm": 0.020727066323161125,
+ "learning_rate": 0.0006,
+ "loss": 4.564209938049316,
+ "step": 1167
+ },
+ {
+ "epoch": 16.2235807860262,
+ "grad_norm": 0.017789119854569435,
+ "learning_rate": 0.0006,
+ "loss": 4.698680877685547,
+ "step": 1168
+ },
+ {
+ "epoch": 16.23755458515284,
+ "grad_norm": 0.017603034153580666,
+ "learning_rate": 0.0006,
+ "loss": 4.501171588897705,
+ "step": 1169
+ },
+ {
+ "epoch": 16.251528384279474,
+ "grad_norm": 0.017345862463116646,
+ "learning_rate": 0.0006,
+ "loss": 4.593590259552002,
+ "step": 1170
+ },
+ {
+ "epoch": 16.265502183406113,
+ "grad_norm": 0.018047425895929337,
+ "learning_rate": 0.0006,
+ "loss": 4.597467422485352,
+ "step": 1171
+ },
+ {
+ "epoch": 16.27947598253275,
+ "grad_norm": 0.01947222836315632,
+ "learning_rate": 0.0006,
+ "loss": 4.5072221755981445,
+ "step": 1172
+ },
+ {
+ "epoch": 16.29344978165939,
+ "grad_norm": 0.019598500803112984,
+ "learning_rate": 0.0006,
+ "loss": 4.601351737976074,
+ "step": 1173
+ },
+ {
+ "epoch": 16.307423580786025,
+ "grad_norm": 0.020002514123916626,
+ "learning_rate": 0.0006,
+ "loss": 4.580431938171387,
+ "step": 1174
+ },
+ {
+ "epoch": 16.321397379912664,
+ "grad_norm": 0.0216530729085207,
+ "learning_rate": 0.0006,
+ "loss": 4.424681663513184,
+ "step": 1175
+ },
+ {
+ "epoch": 16.335371179039303,
+ "grad_norm": 0.01949235238134861,
+ "learning_rate": 0.0006,
+ "loss": 4.751883029937744,
+ "step": 1176
+ },
+ {
+ "epoch": 16.349344978165938,
+ "grad_norm": 0.017779016867280006,
+ "learning_rate": 0.0006,
+ "loss": 4.541490077972412,
+ "step": 1177
+ },
+ {
+ "epoch": 16.363318777292577,
+ "grad_norm": 0.01909121684730053,
+ "learning_rate": 0.0006,
+ "loss": 4.594141006469727,
+ "step": 1178
+ },
+ {
+ "epoch": 16.377292576419215,
+ "grad_norm": 0.019894849509000778,
+ "learning_rate": 0.0006,
+ "loss": 4.555070400238037,
+ "step": 1179
+ },
+ {
+ "epoch": 16.39126637554585,
+ "grad_norm": 0.01993032731115818,
+ "learning_rate": 0.0006,
+ "loss": 4.521401405334473,
+ "step": 1180
+ },
+ {
+ "epoch": 16.40524017467249,
+ "grad_norm": 0.0210788045078516,
+ "learning_rate": 0.0006,
+ "loss": 4.672842979431152,
+ "step": 1181
+ },
+ {
+ "epoch": 16.419213973799128,
+ "grad_norm": 0.02274598926305771,
+ "learning_rate": 0.0006,
+ "loss": 4.629504203796387,
+ "step": 1182
+ },
+ {
+ "epoch": 16.433187772925763,
+ "grad_norm": 0.020448120310902596,
+ "learning_rate": 0.0006,
+ "loss": 4.548209190368652,
+ "step": 1183
+ },
+ {
+ "epoch": 16.4471615720524,
+ "grad_norm": 0.017565643414855003,
+ "learning_rate": 0.0006,
+ "loss": 4.529054641723633,
+ "step": 1184
+ },
+ {
+ "epoch": 16.46113537117904,
+ "grad_norm": 0.019504237920045853,
+ "learning_rate": 0.0006,
+ "loss": 4.48182487487793,
+ "step": 1185
+ },
+ {
+ "epoch": 16.475109170305675,
+ "grad_norm": 0.016823742538690567,
+ "learning_rate": 0.0006,
+ "loss": 4.608696937561035,
+ "step": 1186
+ },
+ {
+ "epoch": 16.489082969432314,
+ "grad_norm": 0.01602894812822342,
+ "learning_rate": 0.0006,
+ "loss": 4.620062351226807,
+ "step": 1187
+ },
+ {
+ "epoch": 16.503056768558952,
+ "grad_norm": 0.016615642234683037,
+ "learning_rate": 0.0006,
+ "loss": 4.69378662109375,
+ "step": 1188
+ },
+ {
+ "epoch": 16.51703056768559,
+ "grad_norm": 0.018576730042696,
+ "learning_rate": 0.0006,
+ "loss": 4.537041664123535,
+ "step": 1189
+ },
+ {
+ "epoch": 16.531004366812226,
+ "grad_norm": 0.018611151725053787,
+ "learning_rate": 0.0006,
+ "loss": 4.496169090270996,
+ "step": 1190
+ },
+ {
+ "epoch": 16.544978165938865,
+ "grad_norm": 0.018622413277626038,
+ "learning_rate": 0.0006,
+ "loss": 4.557358264923096,
+ "step": 1191
+ },
+ {
+ "epoch": 16.558951965065503,
+ "grad_norm": 0.01823544315993786,
+ "learning_rate": 0.0006,
+ "loss": 4.730792999267578,
+ "step": 1192
+ },
+ {
+ "epoch": 16.57292576419214,
+ "grad_norm": 0.018796846270561218,
+ "learning_rate": 0.0006,
+ "loss": 4.781791687011719,
+ "step": 1193
+ },
+ {
+ "epoch": 16.586899563318777,
+ "grad_norm": 0.01728292554616928,
+ "learning_rate": 0.0006,
+ "loss": 4.515257835388184,
+ "step": 1194
+ },
+ {
+ "epoch": 16.600873362445416,
+ "grad_norm": 0.01699589006602764,
+ "learning_rate": 0.0006,
+ "loss": 4.5996904373168945,
+ "step": 1195
+ },
+ {
+ "epoch": 16.61484716157205,
+ "grad_norm": 0.015905970707535744,
+ "learning_rate": 0.0006,
+ "loss": 4.601301193237305,
+ "step": 1196
+ },
+ {
+ "epoch": 16.62882096069869,
+ "grad_norm": 0.01338571310043335,
+ "learning_rate": 0.0006,
+ "loss": 4.557708740234375,
+ "step": 1197
+ },
+ {
+ "epoch": 16.64279475982533,
+ "grad_norm": 0.015298855490982533,
+ "learning_rate": 0.0006,
+ "loss": 4.542287349700928,
+ "step": 1198
+ },
+ {
+ "epoch": 16.656768558951963,
+ "grad_norm": 0.014753367751836777,
+ "learning_rate": 0.0006,
+ "loss": 4.595787048339844,
+ "step": 1199
+ },
+ {
+ "epoch": 16.670742358078602,
+ "grad_norm": 0.015527483075857162,
+ "learning_rate": 0.0006,
+ "loss": 4.607529163360596,
+ "step": 1200
+ },
+ {
+ "epoch": 16.68471615720524,
+ "grad_norm": 0.015042162500321865,
+ "learning_rate": 0.0006,
+ "loss": 4.686626434326172,
+ "step": 1201
+ },
+ {
+ "epoch": 16.69868995633188,
+ "grad_norm": 0.0159393772482872,
+ "learning_rate": 0.0006,
+ "loss": 4.47656774520874,
+ "step": 1202
+ },
+ {
+ "epoch": 16.712663755458514,
+ "grad_norm": 0.01532787922769785,
+ "learning_rate": 0.0006,
+ "loss": 4.55678653717041,
+ "step": 1203
+ },
+ {
+ "epoch": 16.726637554585153,
+ "grad_norm": 0.016332164406776428,
+ "learning_rate": 0.0006,
+ "loss": 4.51441764831543,
+ "step": 1204
+ },
+ {
+ "epoch": 16.74061135371179,
+ "grad_norm": 0.018901726230978966,
+ "learning_rate": 0.0006,
+ "loss": 4.57227897644043,
+ "step": 1205
+ },
+ {
+ "epoch": 16.754585152838427,
+ "grad_norm": 0.018589433282613754,
+ "learning_rate": 0.0006,
+ "loss": 4.550070762634277,
+ "step": 1206
+ },
+ {
+ "epoch": 16.768558951965066,
+ "grad_norm": 0.01634034886956215,
+ "learning_rate": 0.0006,
+ "loss": 4.586910724639893,
+ "step": 1207
+ },
+ {
+ "epoch": 16.782532751091704,
+ "grad_norm": 0.017599433660507202,
+ "learning_rate": 0.0006,
+ "loss": 4.5891923904418945,
+ "step": 1208
+ },
+ {
+ "epoch": 16.79650655021834,
+ "grad_norm": 0.01878446340560913,
+ "learning_rate": 0.0006,
+ "loss": 4.57227087020874,
+ "step": 1209
+ },
+ {
+ "epoch": 16.810480349344978,
+ "grad_norm": 0.01803469844162464,
+ "learning_rate": 0.0006,
+ "loss": 4.583845138549805,
+ "step": 1210
+ },
+ {
+ "epoch": 16.824454148471617,
+ "grad_norm": 0.018505144864320755,
+ "learning_rate": 0.0006,
+ "loss": 4.561151504516602,
+ "step": 1211
+ },
+ {
+ "epoch": 16.83842794759825,
+ "grad_norm": 0.01920081488788128,
+ "learning_rate": 0.0006,
+ "loss": 4.506834030151367,
+ "step": 1212
+ },
+ {
+ "epoch": 16.85240174672489,
+ "grad_norm": 0.0177712831646204,
+ "learning_rate": 0.0006,
+ "loss": 4.557195663452148,
+ "step": 1213
+ },
+ {
+ "epoch": 16.86637554585153,
+ "grad_norm": 0.019113918766379356,
+ "learning_rate": 0.0006,
+ "loss": 4.651227951049805,
+ "step": 1214
+ },
+ {
+ "epoch": 16.880349344978168,
+ "grad_norm": 0.02012539468705654,
+ "learning_rate": 0.0006,
+ "loss": 4.4952239990234375,
+ "step": 1215
+ },
+ {
+ "epoch": 16.894323144104803,
+ "grad_norm": 0.019993646070361137,
+ "learning_rate": 0.0006,
+ "loss": 4.517433166503906,
+ "step": 1216
+ },
+ {
+ "epoch": 16.90829694323144,
+ "grad_norm": 0.020041441544890404,
+ "learning_rate": 0.0006,
+ "loss": 4.489407539367676,
+ "step": 1217
+ },
+ {
+ "epoch": 16.92227074235808,
+ "grad_norm": 0.0183611661195755,
+ "learning_rate": 0.0006,
+ "loss": 4.5435638427734375,
+ "step": 1218
+ },
+ {
+ "epoch": 16.936244541484715,
+ "grad_norm": 0.019138429313898087,
+ "learning_rate": 0.0006,
+ "loss": 4.545949459075928,
+ "step": 1219
+ },
+ {
+ "epoch": 16.950218340611354,
+ "grad_norm": 0.018554266542196274,
+ "learning_rate": 0.0006,
+ "loss": 4.579620361328125,
+ "step": 1220
+ },
+ {
+ "epoch": 16.964192139737992,
+ "grad_norm": 0.0193545650690794,
+ "learning_rate": 0.0006,
+ "loss": 4.588255405426025,
+ "step": 1221
+ },
+ {
+ "epoch": 16.978165938864628,
+ "grad_norm": 0.019389165565371513,
+ "learning_rate": 0.0006,
+ "loss": 4.520698547363281,
+ "step": 1222
+ },
+ {
+ "epoch": 16.992139737991266,
+ "grad_norm": 0.016601379960775375,
+ "learning_rate": 0.0006,
+ "loss": 4.4809064865112305,
+ "step": 1223
+ },
+ {
+ "epoch": 17.0,
+ "grad_norm": 0.01771303080022335,
+ "learning_rate": 0.0006,
+ "loss": 4.721122741699219,
+ "step": 1224
+ },
+ {
+ "epoch": 17.0,
+ "eval_loss": 4.7903666496276855,
+ "eval_runtime": 56.9826,
+ "eval_samples_per_second": 42.855,
+ "eval_steps_per_second": 1.351,
+ "step": 1224
+ },
+ {
+ "epoch": 17.01397379912664,
+ "grad_norm": 0.017720289528369904,
+ "learning_rate": 0.0006,
+ "loss": 4.523834705352783,
+ "step": 1225
+ },
+ {
+ "epoch": 17.027947598253274,
+ "grad_norm": 0.020137161016464233,
+ "learning_rate": 0.0006,
+ "loss": 4.574061870574951,
+ "step": 1226
+ },
+ {
+ "epoch": 17.041921397379912,
+ "grad_norm": 0.02002314105629921,
+ "learning_rate": 0.0006,
+ "loss": 4.512407302856445,
+ "step": 1227
+ },
+ {
+ "epoch": 17.05589519650655,
+ "grad_norm": 0.019628183916211128,
+ "learning_rate": 0.0006,
+ "loss": 4.502533435821533,
+ "step": 1228
+ },
+ {
+ "epoch": 17.069868995633186,
+ "grad_norm": 0.019808458164334297,
+ "learning_rate": 0.0006,
+ "loss": 4.463156700134277,
+ "step": 1229
+ },
+ {
+ "epoch": 17.083842794759825,
+ "grad_norm": 0.021250786259770393,
+ "learning_rate": 0.0006,
+ "loss": 4.622682094573975,
+ "step": 1230
+ },
+ {
+ "epoch": 17.097816593886463,
+ "grad_norm": 0.022031694650650024,
+ "learning_rate": 0.0006,
+ "loss": 4.451882839202881,
+ "step": 1231
+ },
+ {
+ "epoch": 17.111790393013102,
+ "grad_norm": 0.01898987777531147,
+ "learning_rate": 0.0006,
+ "loss": 4.529331684112549,
+ "step": 1232
+ },
+ {
+ "epoch": 17.125764192139737,
+ "grad_norm": 0.018749834969639778,
+ "learning_rate": 0.0006,
+ "loss": 4.435815334320068,
+ "step": 1233
+ },
+ {
+ "epoch": 17.139737991266376,
+ "grad_norm": 0.01896674931049347,
+ "learning_rate": 0.0006,
+ "loss": 4.484029293060303,
+ "step": 1234
+ },
+ {
+ "epoch": 17.153711790393015,
+ "grad_norm": 0.018286997452378273,
+ "learning_rate": 0.0006,
+ "loss": 4.523510932922363,
+ "step": 1235
+ },
+ {
+ "epoch": 17.16768558951965,
+ "grad_norm": 0.017273059114813805,
+ "learning_rate": 0.0006,
+ "loss": 4.451767444610596,
+ "step": 1236
+ },
+ {
+ "epoch": 17.18165938864629,
+ "grad_norm": 0.016499150544404984,
+ "learning_rate": 0.0006,
+ "loss": 4.5568647384643555,
+ "step": 1237
+ },
+ {
+ "epoch": 17.195633187772927,
+ "grad_norm": 0.015934469178318977,
+ "learning_rate": 0.0006,
+ "loss": 4.4639997482299805,
+ "step": 1238
+ },
+ {
+ "epoch": 17.209606986899562,
+ "grad_norm": 0.01714209094643593,
+ "learning_rate": 0.0006,
+ "loss": 4.547614097595215,
+ "step": 1239
+ },
+ {
+ "epoch": 17.2235807860262,
+ "grad_norm": 0.019167495891451836,
+ "learning_rate": 0.0006,
+ "loss": 4.571097373962402,
+ "step": 1240
+ },
+ {
+ "epoch": 17.23755458515284,
+ "grad_norm": 0.018276996910572052,
+ "learning_rate": 0.0006,
+ "loss": 4.489764213562012,
+ "step": 1241
+ },
+ {
+ "epoch": 17.251528384279474,
+ "grad_norm": 0.021296866238117218,
+ "learning_rate": 0.0006,
+ "loss": 4.526913642883301,
+ "step": 1242
+ },
+ {
+ "epoch": 17.265502183406113,
+ "grad_norm": 0.022606629878282547,
+ "learning_rate": 0.0006,
+ "loss": 4.521624565124512,
+ "step": 1243
+ },
+ {
+ "epoch": 17.27947598253275,
+ "grad_norm": 0.021204711869359016,
+ "learning_rate": 0.0006,
+ "loss": 4.500864028930664,
+ "step": 1244
+ },
+ {
+ "epoch": 17.29344978165939,
+ "grad_norm": 0.01988849975168705,
+ "learning_rate": 0.0006,
+ "loss": 4.45162296295166,
+ "step": 1245
+ },
+ {
+ "epoch": 17.307423580786025,
+ "grad_norm": 0.020268384367227554,
+ "learning_rate": 0.0006,
+ "loss": 4.536198616027832,
+ "step": 1246
+ },
+ {
+ "epoch": 17.321397379912664,
+ "grad_norm": 0.020185677334666252,
+ "learning_rate": 0.0006,
+ "loss": 4.440262794494629,
+ "step": 1247
+ },
+ {
+ "epoch": 17.335371179039303,
+ "grad_norm": 0.020106853917241096,
+ "learning_rate": 0.0006,
+ "loss": 4.475907325744629,
+ "step": 1248
+ },
+ {
+ "epoch": 17.349344978165938,
+ "grad_norm": 0.018883297219872475,
+ "learning_rate": 0.0006,
+ "loss": 4.640661716461182,
+ "step": 1249
+ },
+ {
+ "epoch": 17.363318777292577,
+ "grad_norm": 0.01852620206773281,
+ "learning_rate": 0.0006,
+ "loss": 4.578243255615234,
+ "step": 1250
+ },
+ {
+ "epoch": 17.377292576419215,
+ "grad_norm": 0.01886451430618763,
+ "learning_rate": 0.0006,
+ "loss": 4.546026229858398,
+ "step": 1251
+ },
+ {
+ "epoch": 17.39126637554585,
+ "grad_norm": 0.01777452416718006,
+ "learning_rate": 0.0006,
+ "loss": 4.521542549133301,
+ "step": 1252
+ },
+ {
+ "epoch": 17.40524017467249,
+ "grad_norm": 0.016626447439193726,
+ "learning_rate": 0.0006,
+ "loss": 4.52629280090332,
+ "step": 1253
+ },
+ {
+ "epoch": 17.419213973799128,
+ "grad_norm": 0.01599477231502533,
+ "learning_rate": 0.0006,
+ "loss": 4.4796037673950195,
+ "step": 1254
+ },
+ {
+ "epoch": 17.433187772925763,
+ "grad_norm": 0.017021458595991135,
+ "learning_rate": 0.0006,
+ "loss": 4.468813419342041,
+ "step": 1255
+ },
+ {
+ "epoch": 17.4471615720524,
+ "grad_norm": 0.016925333067774773,
+ "learning_rate": 0.0006,
+ "loss": 4.584898948669434,
+ "step": 1256
+ },
+ {
+ "epoch": 17.46113537117904,
+ "grad_norm": 0.01714998669922352,
+ "learning_rate": 0.0006,
+ "loss": 4.523544788360596,
+ "step": 1257
+ },
+ {
+ "epoch": 17.475109170305675,
+ "grad_norm": 0.01802288182079792,
+ "learning_rate": 0.0006,
+ "loss": 4.595581531524658,
+ "step": 1258
+ },
+ {
+ "epoch": 17.489082969432314,
+ "grad_norm": 0.0180289875715971,
+ "learning_rate": 0.0006,
+ "loss": 4.394440650939941,
+ "step": 1259
+ },
+ {
+ "epoch": 17.503056768558952,
+ "grad_norm": 0.018504245206713676,
+ "learning_rate": 0.0006,
+ "loss": 4.4794416427612305,
+ "step": 1260
+ },
+ {
+ "epoch": 17.51703056768559,
+ "grad_norm": 0.019973233342170715,
+ "learning_rate": 0.0006,
+ "loss": 4.4241437911987305,
+ "step": 1261
+ },
+ {
+ "epoch": 17.531004366812226,
+ "grad_norm": 0.020063083618879318,
+ "learning_rate": 0.0006,
+ "loss": 4.48068380355835,
+ "step": 1262
+ },
+ {
+ "epoch": 17.544978165938865,
+ "grad_norm": 0.02066640742123127,
+ "learning_rate": 0.0006,
+ "loss": 4.628993034362793,
+ "step": 1263
+ },
+ {
+ "epoch": 17.558951965065503,
+ "grad_norm": 0.02206416428089142,
+ "learning_rate": 0.0006,
+ "loss": 4.564790725708008,
+ "step": 1264
+ },
+ {
+ "epoch": 17.57292576419214,
+ "grad_norm": 0.020640335977077484,
+ "learning_rate": 0.0006,
+ "loss": 4.578705787658691,
+ "step": 1265
+ },
+ {
+ "epoch": 17.586899563318777,
+ "grad_norm": 0.019093405455350876,
+ "learning_rate": 0.0006,
+ "loss": 4.5023603439331055,
+ "step": 1266
+ },
+ {
+ "epoch": 17.600873362445416,
+ "grad_norm": 0.019459193572402,
+ "learning_rate": 0.0006,
+ "loss": 4.594531059265137,
+ "step": 1267
+ },
+ {
+ "epoch": 17.61484716157205,
+ "grad_norm": 0.02021171525120735,
+ "learning_rate": 0.0006,
+ "loss": 4.770299911499023,
+ "step": 1268
+ },
+ {
+ "epoch": 17.62882096069869,
+ "grad_norm": 0.021161191165447235,
+ "learning_rate": 0.0006,
+ "loss": 4.5380401611328125,
+ "step": 1269
+ },
+ {
+ "epoch": 17.64279475982533,
+ "grad_norm": 0.021013403311371803,
+ "learning_rate": 0.0006,
+ "loss": 4.455524921417236,
+ "step": 1270
+ },
+ {
+ "epoch": 17.656768558951963,
+ "grad_norm": 0.020336592569947243,
+ "learning_rate": 0.0006,
+ "loss": 4.533672332763672,
+ "step": 1271
+ },
+ {
+ "epoch": 17.670742358078602,
+ "grad_norm": 0.020159434527158737,
+ "learning_rate": 0.0006,
+ "loss": 4.596151351928711,
+ "step": 1272
+ },
+ {
+ "epoch": 17.68471615720524,
+ "grad_norm": 0.020432960242033005,
+ "learning_rate": 0.0006,
+ "loss": 4.55531644821167,
+ "step": 1273
+ },
+ {
+ "epoch": 17.69868995633188,
+ "grad_norm": 0.025466229766607285,
+ "learning_rate": 0.0006,
+ "loss": 4.566912651062012,
+ "step": 1274
+ },
+ {
+ "epoch": 17.712663755458514,
+ "grad_norm": 0.026537558063864708,
+ "learning_rate": 0.0006,
+ "loss": 4.526003837585449,
+ "step": 1275
+ },
+ {
+ "epoch": 17.726637554585153,
+ "grad_norm": 0.024397607892751694,
+ "learning_rate": 0.0006,
+ "loss": 4.541755676269531,
+ "step": 1276
+ },
+ {
+ "epoch": 17.74061135371179,
+ "grad_norm": 0.022062312811613083,
+ "learning_rate": 0.0006,
+ "loss": 4.6532793045043945,
+ "step": 1277
+ },
+ {
+ "epoch": 17.754585152838427,
+ "grad_norm": 0.02204282023012638,
+ "learning_rate": 0.0006,
+ "loss": 4.582825660705566,
+ "step": 1278
+ },
+ {
+ "epoch": 17.768558951965066,
+ "grad_norm": 0.020844416692852974,
+ "learning_rate": 0.0006,
+ "loss": 4.522336483001709,
+ "step": 1279
+ },
+ {
+ "epoch": 17.782532751091704,
+ "grad_norm": 0.019815703853964806,
+ "learning_rate": 0.0006,
+ "loss": 4.607398509979248,
+ "step": 1280
+ },
+ {
+ "epoch": 17.79650655021834,
+ "grad_norm": 0.016676954925060272,
+ "learning_rate": 0.0006,
+ "loss": 4.446206092834473,
+ "step": 1281
+ },
+ {
+ "epoch": 17.810480349344978,
+ "grad_norm": 0.0166452769190073,
+ "learning_rate": 0.0006,
+ "loss": 4.505931854248047,
+ "step": 1282
+ },
+ {
+ "epoch": 17.824454148471617,
+ "grad_norm": 0.015928907319903374,
+ "learning_rate": 0.0006,
+ "loss": 4.455267429351807,
+ "step": 1283
+ },
+ {
+ "epoch": 17.83842794759825,
+ "grad_norm": 0.015219560824334621,
+ "learning_rate": 0.0006,
+ "loss": 4.5623674392700195,
+ "step": 1284
+ },
+ {
+ "epoch": 17.85240174672489,
+ "grad_norm": 0.015125355683267117,
+ "learning_rate": 0.0006,
+ "loss": 4.554104804992676,
+ "step": 1285
+ },
+ {
+ "epoch": 17.86637554585153,
+ "grad_norm": 0.015612718649208546,
+ "learning_rate": 0.0006,
+ "loss": 4.484847068786621,
+ "step": 1286
+ },
+ {
+ "epoch": 17.880349344978168,
+ "grad_norm": 0.015215662308037281,
+ "learning_rate": 0.0006,
+ "loss": 4.4802093505859375,
+ "step": 1287
+ },
+ {
+ "epoch": 17.894323144104803,
+ "grad_norm": 0.01511684525758028,
+ "learning_rate": 0.0006,
+ "loss": 4.537298679351807,
+ "step": 1288
+ },
+ {
+ "epoch": 17.90829694323144,
+ "grad_norm": 0.013841322623193264,
+ "learning_rate": 0.0006,
+ "loss": 4.427488327026367,
+ "step": 1289
+ },
+ {
+ "epoch": 17.92227074235808,
+ "grad_norm": 0.015115964226424694,
+ "learning_rate": 0.0006,
+ "loss": 4.4547119140625,
+ "step": 1290
+ },
+ {
+ "epoch": 17.936244541484715,
+ "grad_norm": 0.016411006450653076,
+ "learning_rate": 0.0006,
+ "loss": 4.52994966506958,
+ "step": 1291
+ },
+ {
+ "epoch": 17.950218340611354,
+ "grad_norm": 0.017313152551651,
+ "learning_rate": 0.0006,
+ "loss": 4.5337653160095215,
+ "step": 1292
+ },
+ {
+ "epoch": 17.964192139737992,
+ "grad_norm": 0.016559796407818794,
+ "learning_rate": 0.0006,
+ "loss": 4.5277910232543945,
+ "step": 1293
+ },
+ {
+ "epoch": 17.978165938864628,
+ "grad_norm": 0.015854761004447937,
+ "learning_rate": 0.0006,
+ "loss": 4.6334309577941895,
+ "step": 1294
+ },
+ {
+ "epoch": 17.992139737991266,
+ "grad_norm": 0.015658238902688026,
+ "learning_rate": 0.0006,
+ "loss": 4.613372802734375,
+ "step": 1295
+ },
+ {
+ "epoch": 18.0,
+ "grad_norm": 0.0169155802577734,
+ "learning_rate": 0.0006,
+ "loss": 4.527356147766113,
+ "step": 1296
+ },
+ {
+ "epoch": 18.0,
+ "eval_loss": 4.764543056488037,
+ "eval_runtime": 56.5998,
+ "eval_samples_per_second": 43.145,
+ "eval_steps_per_second": 1.36,
+ "step": 1296
+ },
+ {
+ "epoch": 18.01397379912664,
+ "grad_norm": 0.015348963439464569,
+ "learning_rate": 0.0006,
+ "loss": 4.466423988342285,
+ "step": 1297
+ },
+ {
+ "epoch": 18.027947598253274,
+ "grad_norm": 0.01768588088452816,
+ "learning_rate": 0.0006,
+ "loss": 4.545933246612549,
+ "step": 1298
+ },
+ {
+ "epoch": 18.041921397379912,
+ "grad_norm": 0.017381226643919945,
+ "learning_rate": 0.0006,
+ "loss": 4.494060039520264,
+ "step": 1299
+ },
+ {
+ "epoch": 18.05589519650655,
+ "grad_norm": 0.019002335146069527,
+ "learning_rate": 0.0006,
+ "loss": 4.395200729370117,
+ "step": 1300
+ },
+ {
+ "epoch": 18.069868995633186,
+ "grad_norm": 0.01827932707965374,
+ "learning_rate": 0.0006,
+ "loss": 4.424848556518555,
+ "step": 1301
+ },
+ {
+ "epoch": 18.083842794759825,
+ "grad_norm": 0.016383031383156776,
+ "learning_rate": 0.0006,
+ "loss": 4.494722366333008,
+ "step": 1302
+ },
+ {
+ "epoch": 18.097816593886463,
+ "grad_norm": 0.016265157610177994,
+ "learning_rate": 0.0006,
+ "loss": 4.416522979736328,
+ "step": 1303
+ },
+ {
+ "epoch": 18.111790393013102,
+ "grad_norm": 0.01702386513352394,
+ "learning_rate": 0.0006,
+ "loss": 4.600103378295898,
+ "step": 1304
+ },
+ {
+ "epoch": 18.125764192139737,
+ "grad_norm": 0.018495498225092888,
+ "learning_rate": 0.0006,
+ "loss": 4.588401794433594,
+ "step": 1305
+ },
+ {
+ "epoch": 18.139737991266376,
+ "grad_norm": 0.017921684309840202,
+ "learning_rate": 0.0006,
+ "loss": 4.478075981140137,
+ "step": 1306
+ },
+ {
+ "epoch": 18.153711790393015,
+ "grad_norm": 0.01801278069615364,
+ "learning_rate": 0.0006,
+ "loss": 4.413186550140381,
+ "step": 1307
+ },
+ {
+ "epoch": 18.16768558951965,
+ "grad_norm": 0.017211275175213814,
+ "learning_rate": 0.0006,
+ "loss": 4.461341857910156,
+ "step": 1308
+ },
+ {
+ "epoch": 18.18165938864629,
+ "grad_norm": 0.017589328810572624,
+ "learning_rate": 0.0006,
+ "loss": 4.546969413757324,
+ "step": 1309
+ },
+ {
+ "epoch": 18.195633187772927,
+ "grad_norm": 0.01797698810696602,
+ "learning_rate": 0.0006,
+ "loss": 4.530797004699707,
+ "step": 1310
+ },
+ {
+ "epoch": 18.209606986899562,
+ "grad_norm": 0.020408619195222855,
+ "learning_rate": 0.0006,
+ "loss": 4.62999153137207,
+ "step": 1311
+ },
+ {
+ "epoch": 18.2235807860262,
+ "grad_norm": 0.01994038000702858,
+ "learning_rate": 0.0006,
+ "loss": 4.586811065673828,
+ "step": 1312
+ },
+ {
+ "epoch": 18.23755458515284,
+ "grad_norm": 0.019381340593099594,
+ "learning_rate": 0.0006,
+ "loss": 4.4769368171691895,
+ "step": 1313
+ },
+ {
+ "epoch": 18.251528384279474,
+ "grad_norm": 0.020484555512666702,
+ "learning_rate": 0.0006,
+ "loss": 4.524302959442139,
+ "step": 1314
+ },
+ {
+ "epoch": 18.265502183406113,
+ "grad_norm": 0.02426416426897049,
+ "learning_rate": 0.0006,
+ "loss": 4.561705589294434,
+ "step": 1315
+ },
+ {
+ "epoch": 18.27947598253275,
+ "grad_norm": 0.024748101830482483,
+ "learning_rate": 0.0006,
+ "loss": 4.4654316902160645,
+ "step": 1316
+ },
+ {
+ "epoch": 18.29344978165939,
+ "grad_norm": 0.020703328773379326,
+ "learning_rate": 0.0006,
+ "loss": 4.53033447265625,
+ "step": 1317
+ },
+ {
+ "epoch": 18.307423580786025,
+ "grad_norm": 0.01765458658337593,
+ "learning_rate": 0.0006,
+ "loss": 4.5154266357421875,
+ "step": 1318
+ },
+ {
+ "epoch": 18.321397379912664,
+ "grad_norm": 0.018646420910954475,
+ "learning_rate": 0.0006,
+ "loss": 4.5224714279174805,
+ "step": 1319
+ },
+ {
+ "epoch": 18.335371179039303,
+ "grad_norm": 0.01907474733889103,
+ "learning_rate": 0.0006,
+ "loss": 4.5055437088012695,
+ "step": 1320
+ },
+ {
+ "epoch": 18.349344978165938,
+ "grad_norm": 0.018179846927523613,
+ "learning_rate": 0.0006,
+ "loss": 4.462161064147949,
+ "step": 1321
+ },
+ {
+ "epoch": 18.363318777292577,
+ "grad_norm": 0.016697878018021584,
+ "learning_rate": 0.0006,
+ "loss": 4.473718643188477,
+ "step": 1322
+ },
+ {
+ "epoch": 18.377292576419215,
+ "grad_norm": 0.017504019662737846,
+ "learning_rate": 0.0006,
+ "loss": 4.474348068237305,
+ "step": 1323
+ },
+ {
+ "epoch": 18.39126637554585,
+ "grad_norm": 0.018889913335442543,
+ "learning_rate": 0.0006,
+ "loss": 4.4531636238098145,
+ "step": 1324
+ },
+ {
+ "epoch": 18.40524017467249,
+ "grad_norm": 0.019167251884937286,
+ "learning_rate": 0.0006,
+ "loss": 4.4278364181518555,
+ "step": 1325
+ },
+ {
+ "epoch": 18.419213973799128,
+ "grad_norm": 0.015895741060376167,
+ "learning_rate": 0.0006,
+ "loss": 4.474592208862305,
+ "step": 1326
+ },
+ {
+ "epoch": 18.433187772925763,
+ "grad_norm": 0.01531882956624031,
+ "learning_rate": 0.0006,
+ "loss": 4.403757572174072,
+ "step": 1327
+ },
+ {
+ "epoch": 18.4471615720524,
+ "grad_norm": 0.01565849967300892,
+ "learning_rate": 0.0006,
+ "loss": 4.507664680480957,
+ "step": 1328
+ },
+ {
+ "epoch": 18.46113537117904,
+ "grad_norm": 0.01602315716445446,
+ "learning_rate": 0.0006,
+ "loss": 4.5648908615112305,
+ "step": 1329
+ },
+ {
+ "epoch": 18.475109170305675,
+ "grad_norm": 0.016467366367578506,
+ "learning_rate": 0.0006,
+ "loss": 4.428762912750244,
+ "step": 1330
+ },
+ {
+ "epoch": 18.489082969432314,
+ "grad_norm": 0.01604127697646618,
+ "learning_rate": 0.0006,
+ "loss": 4.527771472930908,
+ "step": 1331
+ },
+ {
+ "epoch": 18.503056768558952,
+ "grad_norm": 0.018825586885213852,
+ "learning_rate": 0.0006,
+ "loss": 4.485386848449707,
+ "step": 1332
+ },
+ {
+ "epoch": 18.51703056768559,
+ "grad_norm": 0.019258007407188416,
+ "learning_rate": 0.0006,
+ "loss": 4.414972305297852,
+ "step": 1333
+ },
+ {
+ "epoch": 18.531004366812226,
+ "grad_norm": 0.018192021176218987,
+ "learning_rate": 0.0006,
+ "loss": 4.412166595458984,
+ "step": 1334
+ },
+ {
+ "epoch": 18.544978165938865,
+ "grad_norm": 0.016120698302984238,
+ "learning_rate": 0.0006,
+ "loss": 4.484317779541016,
+ "step": 1335
+ },
+ {
+ "epoch": 18.558951965065503,
+ "grad_norm": 0.016496622934937477,
+ "learning_rate": 0.0006,
+ "loss": 4.514986515045166,
+ "step": 1336
+ },
+ {
+ "epoch": 18.57292576419214,
+ "grad_norm": 0.01795116998255253,
+ "learning_rate": 0.0006,
+ "loss": 4.41473388671875,
+ "step": 1337
+ },
+ {
+ "epoch": 18.586899563318777,
+ "grad_norm": 0.019062276929616928,
+ "learning_rate": 0.0006,
+ "loss": 4.601996898651123,
+ "step": 1338
+ },
+ {
+ "epoch": 18.600873362445416,
+ "grad_norm": 0.017174631357192993,
+ "learning_rate": 0.0006,
+ "loss": 4.571218490600586,
+ "step": 1339
+ },
+ {
+ "epoch": 18.61484716157205,
+ "grad_norm": 0.014015606604516506,
+ "learning_rate": 0.0006,
+ "loss": 4.526606559753418,
+ "step": 1340
+ },
+ {
+ "epoch": 18.62882096069869,
+ "grad_norm": 0.015358510427176952,
+ "learning_rate": 0.0006,
+ "loss": 4.604286193847656,
+ "step": 1341
+ },
+ {
+ "epoch": 18.64279475982533,
+ "grad_norm": 0.01619391329586506,
+ "learning_rate": 0.0006,
+ "loss": 4.560883522033691,
+ "step": 1342
+ },
+ {
+ "epoch": 18.656768558951963,
+ "grad_norm": 0.015972964465618134,
+ "learning_rate": 0.0006,
+ "loss": 4.479755401611328,
+ "step": 1343
+ },
+ {
+ "epoch": 18.670742358078602,
+ "grad_norm": 0.016861025243997574,
+ "learning_rate": 0.0006,
+ "loss": 4.474471092224121,
+ "step": 1344
+ },
+ {
+ "epoch": 18.68471615720524,
+ "grad_norm": 0.01822078227996826,
+ "learning_rate": 0.0006,
+ "loss": 4.509489059448242,
+ "step": 1345
+ },
+ {
+ "epoch": 18.69868995633188,
+ "grad_norm": 0.019415007904171944,
+ "learning_rate": 0.0006,
+ "loss": 4.508941173553467,
+ "step": 1346
+ },
+ {
+ "epoch": 18.712663755458514,
+ "grad_norm": 0.019856126978993416,
+ "learning_rate": 0.0006,
+ "loss": 4.446670055389404,
+ "step": 1347
+ },
+ {
+ "epoch": 18.726637554585153,
+ "grad_norm": 0.01842520199716091,
+ "learning_rate": 0.0006,
+ "loss": 4.524422645568848,
+ "step": 1348
+ },
+ {
+ "epoch": 18.74061135371179,
+ "grad_norm": 0.017196480184793472,
+ "learning_rate": 0.0006,
+ "loss": 4.498466968536377,
+ "step": 1349
+ },
+ {
+ "epoch": 18.754585152838427,
+ "grad_norm": 0.01823010854423046,
+ "learning_rate": 0.0006,
+ "loss": 4.534684181213379,
+ "step": 1350
+ },
+ {
+ "epoch": 18.768558951965066,
+ "grad_norm": 0.020398065447807312,
+ "learning_rate": 0.0006,
+ "loss": 4.456494331359863,
+ "step": 1351
+ },
+ {
+ "epoch": 18.782532751091704,
+ "grad_norm": 0.022992262616753578,
+ "learning_rate": 0.0006,
+ "loss": 4.476374626159668,
+ "step": 1352
+ },
+ {
+ "epoch": 18.79650655021834,
+ "grad_norm": 0.02235420234501362,
+ "learning_rate": 0.0006,
+ "loss": 4.587776184082031,
+ "step": 1353
+ },
+ {
+ "epoch": 18.810480349344978,
+ "grad_norm": 0.02093626745045185,
+ "learning_rate": 0.0006,
+ "loss": 4.424172401428223,
+ "step": 1354
+ },
+ {
+ "epoch": 18.824454148471617,
+ "grad_norm": 0.01997302658855915,
+ "learning_rate": 0.0006,
+ "loss": 4.448066711425781,
+ "step": 1355
+ },
+ {
+ "epoch": 18.83842794759825,
+ "grad_norm": 0.018265459686517715,
+ "learning_rate": 0.0006,
+ "loss": 4.436610698699951,
+ "step": 1356
+ },
+ {
+ "epoch": 18.85240174672489,
+ "grad_norm": 0.01673833839595318,
+ "learning_rate": 0.0006,
+ "loss": 4.5502214431762695,
+ "step": 1357
+ },
+ {
+ "epoch": 18.86637554585153,
+ "grad_norm": 0.017908833920955658,
+ "learning_rate": 0.0006,
+ "loss": 4.550530433654785,
+ "step": 1358
+ },
+ {
+ "epoch": 18.880349344978168,
+ "grad_norm": 0.01796654611825943,
+ "learning_rate": 0.0006,
+ "loss": 4.531463146209717,
+ "step": 1359
+ },
+ {
+ "epoch": 18.894323144104803,
+ "grad_norm": 0.01646183803677559,
+ "learning_rate": 0.0006,
+ "loss": 4.498582363128662,
+ "step": 1360
+ },
+ {
+ "epoch": 18.90829694323144,
+ "grad_norm": 0.015430327504873276,
+ "learning_rate": 0.0006,
+ "loss": 4.52571439743042,
+ "step": 1361
+ },
+ {
+ "epoch": 18.92227074235808,
+ "grad_norm": 0.016527952626347542,
+ "learning_rate": 0.0006,
+ "loss": 4.547013759613037,
+ "step": 1362
+ },
+ {
+ "epoch": 18.936244541484715,
+ "grad_norm": 0.016263660043478012,
+ "learning_rate": 0.0006,
+ "loss": 4.468754291534424,
+ "step": 1363
+ },
+ {
+ "epoch": 18.950218340611354,
+ "grad_norm": 0.016347210854291916,
+ "learning_rate": 0.0006,
+ "loss": 4.450741767883301,
+ "step": 1364
+ },
+ {
+ "epoch": 18.964192139737992,
+ "grad_norm": 0.01701398566365242,
+ "learning_rate": 0.0006,
+ "loss": 4.534899711608887,
+ "step": 1365
+ },
+ {
+ "epoch": 18.978165938864628,
+ "grad_norm": 0.017693081870675087,
+ "learning_rate": 0.0006,
+ "loss": 4.460553169250488,
+ "step": 1366
+ },
+ {
+ "epoch": 18.992139737991266,
+ "grad_norm": 0.018746308982372284,
+ "learning_rate": 0.0006,
+ "loss": 4.538792610168457,
+ "step": 1367
+ },
+ {
+ "epoch": 19.0,
+ "grad_norm": 0.020099416375160217,
+ "learning_rate": 0.0006,
+ "loss": 4.449617385864258,
+ "step": 1368
+ },
+ {
+ "epoch": 19.0,
+ "eval_loss": 4.7756571769714355,
+ "eval_runtime": 57.1804,
+ "eval_samples_per_second": 42.707,
+ "eval_steps_per_second": 1.347,
+ "step": 1368
+ },
+ {
+ "epoch": 19.01397379912664,
+ "grad_norm": 0.018726933747529984,
+ "learning_rate": 0.0006,
+ "loss": 4.446059703826904,
+ "step": 1369
+ },
+ {
+ "epoch": 19.027947598253274,
+ "grad_norm": 0.01956087350845337,
+ "learning_rate": 0.0006,
+ "loss": 4.4520039558410645,
+ "step": 1370
+ },
+ {
+ "epoch": 19.041921397379912,
+ "grad_norm": 0.020092811435461044,
+ "learning_rate": 0.0006,
+ "loss": 4.543343544006348,
+ "step": 1371
+ },
+ {
+ "epoch": 19.05589519650655,
+ "grad_norm": 0.01843622885644436,
+ "learning_rate": 0.0006,
+ "loss": 4.535579204559326,
+ "step": 1372
+ },
+ {
+ "epoch": 19.069868995633186,
+ "grad_norm": 0.016989488154649734,
+ "learning_rate": 0.0006,
+ "loss": 4.52316951751709,
+ "step": 1373
+ },
+ {
+ "epoch": 19.083842794759825,
+ "grad_norm": 0.01805422455072403,
+ "learning_rate": 0.0006,
+ "loss": 4.5393266677856445,
+ "step": 1374
+ },
+ {
+ "epoch": 19.097816593886463,
+ "grad_norm": 0.01874629408121109,
+ "learning_rate": 0.0006,
+ "loss": 4.422059059143066,
+ "step": 1375
+ },
+ {
+ "epoch": 19.111790393013102,
+ "grad_norm": 0.01804221048951149,
+ "learning_rate": 0.0006,
+ "loss": 4.50508451461792,
+ "step": 1376
+ },
+ {
+ "epoch": 19.125764192139737,
+ "grad_norm": 0.01785474829375744,
+ "learning_rate": 0.0006,
+ "loss": 4.412599563598633,
+ "step": 1377
+ },
+ {
+ "epoch": 19.139737991266376,
+ "grad_norm": 0.01971791312098503,
+ "learning_rate": 0.0006,
+ "loss": 4.518318176269531,
+ "step": 1378
+ },
+ {
+ "epoch": 19.153711790393015,
+ "grad_norm": 0.02061633951961994,
+ "learning_rate": 0.0006,
+ "loss": 4.332393169403076,
+ "step": 1379
+ },
+ {
+ "epoch": 19.16768558951965,
+ "grad_norm": 0.020563429221510887,
+ "learning_rate": 0.0006,
+ "loss": 4.445863723754883,
+ "step": 1380
+ },
+ {
+ "epoch": 19.18165938864629,
+ "grad_norm": 0.019995713606476784,
+ "learning_rate": 0.0006,
+ "loss": 4.497312068939209,
+ "step": 1381
+ },
+ {
+ "epoch": 19.195633187772927,
+ "grad_norm": 0.019823070615530014,
+ "learning_rate": 0.0006,
+ "loss": 4.475085258483887,
+ "step": 1382
+ },
+ {
+ "epoch": 19.209606986899562,
+ "grad_norm": 0.019039355218410492,
+ "learning_rate": 0.0006,
+ "loss": 4.50432014465332,
+ "step": 1383
+ },
+ {
+ "epoch": 19.2235807860262,
+ "grad_norm": 0.017876973375678062,
+ "learning_rate": 0.0006,
+ "loss": 4.384788513183594,
+ "step": 1384
+ },
+ {
+ "epoch": 19.23755458515284,
+ "grad_norm": 0.018697699531912804,
+ "learning_rate": 0.0006,
+ "loss": 4.410778045654297,
+ "step": 1385
+ },
+ {
+ "epoch": 19.251528384279474,
+ "grad_norm": 0.018826806917786598,
+ "learning_rate": 0.0006,
+ "loss": 4.524034023284912,
+ "step": 1386
+ },
+ {
+ "epoch": 19.265502183406113,
+ "grad_norm": 0.01824226602911949,
+ "learning_rate": 0.0006,
+ "loss": 4.526393890380859,
+ "step": 1387
+ },
+ {
+ "epoch": 19.27947598253275,
+ "grad_norm": 0.01753208599984646,
+ "learning_rate": 0.0006,
+ "loss": 4.493703365325928,
+ "step": 1388
+ },
+ {
+ "epoch": 19.29344978165939,
+ "grad_norm": 0.01742432825267315,
+ "learning_rate": 0.0006,
+ "loss": 4.392831802368164,
+ "step": 1389
+ },
+ {
+ "epoch": 19.307423580786025,
+ "grad_norm": 0.017099356278777122,
+ "learning_rate": 0.0006,
+ "loss": 4.578668594360352,
+ "step": 1390
+ },
+ {
+ "epoch": 19.321397379912664,
+ "grad_norm": 0.017946023494005203,
+ "learning_rate": 0.0006,
+ "loss": 4.456718444824219,
+ "step": 1391
+ },
+ {
+ "epoch": 19.335371179039303,
+ "grad_norm": 0.017705943435430527,
+ "learning_rate": 0.0006,
+ "loss": 4.553682327270508,
+ "step": 1392
+ },
+ {
+ "epoch": 19.349344978165938,
+ "grad_norm": 0.016647523269057274,
+ "learning_rate": 0.0006,
+ "loss": 4.525016784667969,
+ "step": 1393
+ },
+ {
+ "epoch": 19.363318777292577,
+ "grad_norm": 0.015407416969537735,
+ "learning_rate": 0.0006,
+ "loss": 4.500231742858887,
+ "step": 1394
+ },
+ {
+ "epoch": 19.377292576419215,
+ "grad_norm": 0.017264019697904587,
+ "learning_rate": 0.0006,
+ "loss": 4.4615349769592285,
+ "step": 1395
+ },
+ {
+ "epoch": 19.39126637554585,
+ "grad_norm": 0.01871555484831333,
+ "learning_rate": 0.0006,
+ "loss": 4.281026840209961,
+ "step": 1396
+ },
+ {
+ "epoch": 19.40524017467249,
+ "grad_norm": 0.017361650243401527,
+ "learning_rate": 0.0006,
+ "loss": 4.479382514953613,
+ "step": 1397
+ },
+ {
+ "epoch": 19.419213973799128,
+ "grad_norm": 0.019781362265348434,
+ "learning_rate": 0.0006,
+ "loss": 4.402320861816406,
+ "step": 1398
+ },
+ {
+ "epoch": 19.433187772925763,
+ "grad_norm": 0.019043298438191414,
+ "learning_rate": 0.0006,
+ "loss": 4.476983070373535,
+ "step": 1399
+ },
+ {
+ "epoch": 19.4471615720524,
+ "grad_norm": 0.017137985676527023,
+ "learning_rate": 0.0006,
+ "loss": 4.418689250946045,
+ "step": 1400
+ },
+ {
+ "epoch": 19.46113537117904,
+ "grad_norm": 0.016481833532452583,
+ "learning_rate": 0.0006,
+ "loss": 4.579145431518555,
+ "step": 1401
+ },
+ {
+ "epoch": 19.475109170305675,
+ "grad_norm": 0.016507714986801147,
+ "learning_rate": 0.0006,
+ "loss": 4.519158363342285,
+ "step": 1402
+ },
+ {
+ "epoch": 19.489082969432314,
+ "grad_norm": 0.018092872574925423,
+ "learning_rate": 0.0006,
+ "loss": 4.513748645782471,
+ "step": 1403
+ },
+ {
+ "epoch": 19.503056768558952,
+ "grad_norm": 0.018153440207242966,
+ "learning_rate": 0.0006,
+ "loss": 4.645384311676025,
+ "step": 1404
+ },
+ {
+ "epoch": 19.51703056768559,
+ "grad_norm": 0.01657119207084179,
+ "learning_rate": 0.0006,
+ "loss": 4.399757385253906,
+ "step": 1405
+ },
+ {
+ "epoch": 19.531004366812226,
+ "grad_norm": 0.0156280305236578,
+ "learning_rate": 0.0006,
+ "loss": 4.467836380004883,
+ "step": 1406
+ },
+ {
+ "epoch": 19.544978165938865,
+ "grad_norm": 0.016838403418660164,
+ "learning_rate": 0.0006,
+ "loss": 4.529176235198975,
+ "step": 1407
+ },
+ {
+ "epoch": 19.558951965065503,
+ "grad_norm": 0.01743236929178238,
+ "learning_rate": 0.0006,
+ "loss": 4.440032482147217,
+ "step": 1408
+ },
+ {
+ "epoch": 19.57292576419214,
+ "grad_norm": 0.01553610060364008,
+ "learning_rate": 0.0006,
+ "loss": 4.425588607788086,
+ "step": 1409
+ },
+ {
+ "epoch": 19.586899563318777,
+ "grad_norm": 0.017259271815419197,
+ "learning_rate": 0.0006,
+ "loss": 4.405452728271484,
+ "step": 1410
+ },
+ {
+ "epoch": 19.600873362445416,
+ "grad_norm": 0.015568481758236885,
+ "learning_rate": 0.0006,
+ "loss": 4.444279193878174,
+ "step": 1411
+ },
+ {
+ "epoch": 19.61484716157205,
+ "grad_norm": 0.015331248752772808,
+ "learning_rate": 0.0006,
+ "loss": 4.451846122741699,
+ "step": 1412
+ },
+ {
+ "epoch": 19.62882096069869,
+ "grad_norm": 0.015667378902435303,
+ "learning_rate": 0.0006,
+ "loss": 4.533931255340576,
+ "step": 1413
+ },
+ {
+ "epoch": 19.64279475982533,
+ "grad_norm": 0.016177602112293243,
+ "learning_rate": 0.0006,
+ "loss": 4.467313289642334,
+ "step": 1414
+ },
+ {
+ "epoch": 19.656768558951963,
+ "grad_norm": 0.01733129471540451,
+ "learning_rate": 0.0006,
+ "loss": 4.496893882751465,
+ "step": 1415
+ },
+ {
+ "epoch": 19.670742358078602,
+ "grad_norm": 0.016539594158530235,
+ "learning_rate": 0.0006,
+ "loss": 4.401961326599121,
+ "step": 1416
+ },
+ {
+ "epoch": 19.68471615720524,
+ "grad_norm": 0.016058551147580147,
+ "learning_rate": 0.0006,
+ "loss": 4.503375053405762,
+ "step": 1417
+ },
+ {
+ "epoch": 19.69868995633188,
+ "grad_norm": 0.01715777814388275,
+ "learning_rate": 0.0006,
+ "loss": 4.3808369636535645,
+ "step": 1418
+ },
+ {
+ "epoch": 19.712663755458514,
+ "grad_norm": 0.01681119203567505,
+ "learning_rate": 0.0006,
+ "loss": 4.5469560623168945,
+ "step": 1419
+ },
+ {
+ "epoch": 19.726637554585153,
+ "grad_norm": 0.01634550653398037,
+ "learning_rate": 0.0006,
+ "loss": 4.328197956085205,
+ "step": 1420
+ },
+ {
+ "epoch": 19.74061135371179,
+ "grad_norm": 0.01549347210675478,
+ "learning_rate": 0.0006,
+ "loss": 4.534621238708496,
+ "step": 1421
+ },
+ {
+ "epoch": 19.754585152838427,
+ "grad_norm": 0.01654665172100067,
+ "learning_rate": 0.0006,
+ "loss": 4.4724225997924805,
+ "step": 1422
+ },
+ {
+ "epoch": 19.768558951965066,
+ "grad_norm": 0.017119025811553,
+ "learning_rate": 0.0006,
+ "loss": 4.3897247314453125,
+ "step": 1423
+ },
+ {
+ "epoch": 19.782532751091704,
+ "grad_norm": 0.01603223942220211,
+ "learning_rate": 0.0006,
+ "loss": 4.412731170654297,
+ "step": 1424
+ },
+ {
+ "epoch": 19.79650655021834,
+ "grad_norm": 0.016774067655205727,
+ "learning_rate": 0.0006,
+ "loss": 4.54707145690918,
+ "step": 1425
+ },
+ {
+ "epoch": 19.810480349344978,
+ "grad_norm": 0.016453025862574577,
+ "learning_rate": 0.0006,
+ "loss": 4.462501049041748,
+ "step": 1426
+ },
+ {
+ "epoch": 19.824454148471617,
+ "grad_norm": 0.01597629487514496,
+ "learning_rate": 0.0006,
+ "loss": 4.229758262634277,
+ "step": 1427
+ },
+ {
+ "epoch": 19.83842794759825,
+ "grad_norm": 0.017999034374952316,
+ "learning_rate": 0.0006,
+ "loss": 4.389263153076172,
+ "step": 1428
+ },
+ {
+ "epoch": 19.85240174672489,
+ "grad_norm": 0.02165898121893406,
+ "learning_rate": 0.0006,
+ "loss": 4.342130661010742,
+ "step": 1429
+ },
+ {
+ "epoch": 19.86637554585153,
+ "grad_norm": 0.02222822792828083,
+ "learning_rate": 0.0006,
+ "loss": 4.492661476135254,
+ "step": 1430
+ },
+ {
+ "epoch": 19.880349344978168,
+ "grad_norm": 0.019378066062927246,
+ "learning_rate": 0.0006,
+ "loss": 4.4614105224609375,
+ "step": 1431
+ },
+ {
+ "epoch": 19.894323144104803,
+ "grad_norm": 0.019908083602786064,
+ "learning_rate": 0.0006,
+ "loss": 4.370929718017578,
+ "step": 1432
+ },
+ {
+ "epoch": 19.90829694323144,
+ "grad_norm": 0.022296471521258354,
+ "learning_rate": 0.0006,
+ "loss": 4.438320159912109,
+ "step": 1433
+ },
+ {
+ "epoch": 19.92227074235808,
+ "grad_norm": 0.021810343489050865,
+ "learning_rate": 0.0006,
+ "loss": 4.462067127227783,
+ "step": 1434
+ },
+ {
+ "epoch": 19.936244541484715,
+ "grad_norm": 0.02129209227859974,
+ "learning_rate": 0.0006,
+ "loss": 4.541261196136475,
+ "step": 1435
+ },
+ {
+ "epoch": 19.950218340611354,
+ "grad_norm": 0.02079629711806774,
+ "learning_rate": 0.0006,
+ "loss": 4.36712646484375,
+ "step": 1436
+ },
+ {
+ "epoch": 19.964192139737992,
+ "grad_norm": 0.022501740604639053,
+ "learning_rate": 0.0006,
+ "loss": 4.5279083251953125,
+ "step": 1437
+ },
+ {
+ "epoch": 19.978165938864628,
+ "grad_norm": 0.019121676683425903,
+ "learning_rate": 0.0006,
+ "loss": 4.477789878845215,
+ "step": 1438
+ },
+ {
+ "epoch": 19.992139737991266,
+ "grad_norm": 0.017429588362574577,
+ "learning_rate": 0.0006,
+ "loss": 4.4023847579956055,
+ "step": 1439
+ },
+ {
+ "epoch": 20.0,
+ "grad_norm": 0.018857665359973907,
+ "learning_rate": 0.0006,
+ "loss": 4.494743347167969,
+ "step": 1440
+ },
+ {
+ "epoch": 20.0,
+ "eval_loss": 4.694555759429932,
+ "eval_runtime": 56.4146,
+ "eval_samples_per_second": 43.287,
+ "eval_steps_per_second": 1.365,
+ "step": 1440
+ },
+ {
+ "epoch": 20.01397379912664,
+ "grad_norm": 0.016505861654877663,
+ "learning_rate": 0.0006,
+ "loss": 4.506229877471924,
+ "step": 1441
+ },
+ {
+ "epoch": 20.027947598253274,
+ "grad_norm": 0.017571842297911644,
+ "learning_rate": 0.0006,
+ "loss": 4.384073257446289,
+ "step": 1442
+ },
+ {
+ "epoch": 20.041921397379912,
+ "grad_norm": 0.019398365169763565,
+ "learning_rate": 0.0006,
+ "loss": 4.482177734375,
+ "step": 1443
+ },
+ {
+ "epoch": 20.05589519650655,
+ "grad_norm": 0.02011062018573284,
+ "learning_rate": 0.0006,
+ "loss": 4.440773963928223,
+ "step": 1444
+ },
+ {
+ "epoch": 20.069868995633186,
+ "grad_norm": 0.019026953727006912,
+ "learning_rate": 0.0006,
+ "loss": 4.507168769836426,
+ "step": 1445
+ },
+ {
+ "epoch": 20.083842794759825,
+ "grad_norm": 0.020506031811237335,
+ "learning_rate": 0.0006,
+ "loss": 4.469079971313477,
+ "step": 1446
+ },
+ {
+ "epoch": 20.097816593886463,
+ "grad_norm": 0.019608162343502045,
+ "learning_rate": 0.0006,
+ "loss": 4.395927429199219,
+ "step": 1447
+ },
+ {
+ "epoch": 20.111790393013102,
+ "grad_norm": 0.019419124349951744,
+ "learning_rate": 0.0006,
+ "loss": 4.484235763549805,
+ "step": 1448
+ },
+ {
+ "epoch": 20.125764192139737,
+ "grad_norm": 0.021775128319859505,
+ "learning_rate": 0.0006,
+ "loss": 4.333173751831055,
+ "step": 1449
+ },
+ {
+ "epoch": 20.139737991266376,
+ "grad_norm": 0.025512272492051125,
+ "learning_rate": 0.0006,
+ "loss": 4.460362434387207,
+ "step": 1450
+ },
+ {
+ "epoch": 20.153711790393015,
+ "grad_norm": 0.025157401338219643,
+ "learning_rate": 0.0006,
+ "loss": 4.4361395835876465,
+ "step": 1451
+ },
+ {
+ "epoch": 20.16768558951965,
+ "grad_norm": 0.019862616434693336,
+ "learning_rate": 0.0006,
+ "loss": 4.426807403564453,
+ "step": 1452
+ },
+ {
+ "epoch": 20.18165938864629,
+ "grad_norm": 0.01932152360677719,
+ "learning_rate": 0.0006,
+ "loss": 4.375677108764648,
+ "step": 1453
+ },
+ {
+ "epoch": 20.195633187772927,
+ "grad_norm": 0.01760999858379364,
+ "learning_rate": 0.0006,
+ "loss": 4.42259407043457,
+ "step": 1454
+ },
+ {
+ "epoch": 20.209606986899562,
+ "grad_norm": 0.016998112201690674,
+ "learning_rate": 0.0006,
+ "loss": 4.481280326843262,
+ "step": 1455
+ },
+ {
+ "epoch": 20.2235807860262,
+ "grad_norm": 0.017199190333485603,
+ "learning_rate": 0.0006,
+ "loss": 4.496712684631348,
+ "step": 1456
+ },
+ {
+ "epoch": 20.23755458515284,
+ "grad_norm": 0.016833679750561714,
+ "learning_rate": 0.0006,
+ "loss": 4.40736722946167,
+ "step": 1457
+ },
+ {
+ "epoch": 20.251528384279474,
+ "grad_norm": 0.017546730116009712,
+ "learning_rate": 0.0006,
+ "loss": 4.4915266036987305,
+ "step": 1458
+ },
+ {
+ "epoch": 20.265502183406113,
+ "grad_norm": 0.016595078632235527,
+ "learning_rate": 0.0006,
+ "loss": 4.460815906524658,
+ "step": 1459
+ },
+ {
+ "epoch": 20.27947598253275,
+ "grad_norm": 0.017728568986058235,
+ "learning_rate": 0.0006,
+ "loss": 4.346048831939697,
+ "step": 1460
+ },
+ {
+ "epoch": 20.29344978165939,
+ "grad_norm": 0.01849197782576084,
+ "learning_rate": 0.0006,
+ "loss": 4.399681091308594,
+ "step": 1461
+ },
+ {
+ "epoch": 20.307423580786025,
+ "grad_norm": 0.018336588516831398,
+ "learning_rate": 0.0006,
+ "loss": 4.414183616638184,
+ "step": 1462
+ },
+ {
+ "epoch": 20.321397379912664,
+ "grad_norm": 0.018086519092321396,
+ "learning_rate": 0.0006,
+ "loss": 4.346586227416992,
+ "step": 1463
+ },
+ {
+ "epoch": 20.335371179039303,
+ "grad_norm": 0.019697299227118492,
+ "learning_rate": 0.0006,
+ "loss": 4.4420952796936035,
+ "step": 1464
+ },
+ {
+ "epoch": 20.349344978165938,
+ "grad_norm": 0.02198074571788311,
+ "learning_rate": 0.0006,
+ "loss": 4.588729381561279,
+ "step": 1465
+ },
+ {
+ "epoch": 20.363318777292577,
+ "grad_norm": 0.022813329473137856,
+ "learning_rate": 0.0006,
+ "loss": 4.49729585647583,
+ "step": 1466
+ },
+ {
+ "epoch": 20.377292576419215,
+ "grad_norm": 0.022006649523973465,
+ "learning_rate": 0.0006,
+ "loss": 4.475641250610352,
+ "step": 1467
+ },
+ {
+ "epoch": 20.39126637554585,
+ "grad_norm": 0.023188438266515732,
+ "learning_rate": 0.0006,
+ "loss": 4.404322624206543,
+ "step": 1468
+ },
+ {
+ "epoch": 20.40524017467249,
+ "grad_norm": 0.022226877510547638,
+ "learning_rate": 0.0006,
+ "loss": 4.50904655456543,
+ "step": 1469
+ },
+ {
+ "epoch": 20.419213973799128,
+ "grad_norm": 0.021082637831568718,
+ "learning_rate": 0.0006,
+ "loss": 4.339011192321777,
+ "step": 1470
+ },
+ {
+ "epoch": 20.433187772925763,
+ "grad_norm": 0.02013542130589485,
+ "learning_rate": 0.0006,
+ "loss": 4.516942024230957,
+ "step": 1471
+ },
+ {
+ "epoch": 20.4471615720524,
+ "grad_norm": 0.01898968778550625,
+ "learning_rate": 0.0006,
+ "loss": 4.315804481506348,
+ "step": 1472
+ },
+ {
+ "epoch": 20.46113537117904,
+ "grad_norm": 0.019450880587100983,
+ "learning_rate": 0.0006,
+ "loss": 4.4794921875,
+ "step": 1473
+ },
+ {
+ "epoch": 20.475109170305675,
+ "grad_norm": 0.018577802926301956,
+ "learning_rate": 0.0006,
+ "loss": 4.422323703765869,
+ "step": 1474
+ },
+ {
+ "epoch": 20.489082969432314,
+ "grad_norm": 0.01778412237763405,
+ "learning_rate": 0.0006,
+ "loss": 4.429789066314697,
+ "step": 1475
+ },
+ {
+ "epoch": 20.503056768558952,
+ "grad_norm": 0.017134476453065872,
+ "learning_rate": 0.0006,
+ "loss": 4.416459083557129,
+ "step": 1476
+ },
+ {
+ "epoch": 20.51703056768559,
+ "grad_norm": 0.017153557389974594,
+ "learning_rate": 0.0006,
+ "loss": 4.446170806884766,
+ "step": 1477
+ },
+ {
+ "epoch": 20.531004366812226,
+ "grad_norm": 0.014998532831668854,
+ "learning_rate": 0.0006,
+ "loss": 4.419447898864746,
+ "step": 1478
+ },
+ {
+ "epoch": 20.544978165938865,
+ "grad_norm": 0.015877850353717804,
+ "learning_rate": 0.0006,
+ "loss": 4.312338829040527,
+ "step": 1479
+ },
+ {
+ "epoch": 20.558951965065503,
+ "grad_norm": 0.01569489948451519,
+ "learning_rate": 0.0006,
+ "loss": 4.572905540466309,
+ "step": 1480
+ },
+ {
+ "epoch": 20.57292576419214,
+ "grad_norm": 0.01591474376618862,
+ "learning_rate": 0.0006,
+ "loss": 4.414907455444336,
+ "step": 1481
+ },
+ {
+ "epoch": 20.586899563318777,
+ "grad_norm": 0.015395903028547764,
+ "learning_rate": 0.0006,
+ "loss": 4.549131393432617,
+ "step": 1482
+ },
+ {
+ "epoch": 20.600873362445416,
+ "grad_norm": 0.015715358778834343,
+ "learning_rate": 0.0006,
+ "loss": 4.530194282531738,
+ "step": 1483
+ },
+ {
+ "epoch": 20.61484716157205,
+ "grad_norm": 0.016042975708842278,
+ "learning_rate": 0.0006,
+ "loss": 4.322149276733398,
+ "step": 1484
+ },
+ {
+ "epoch": 20.62882096069869,
+ "grad_norm": 0.01664186641573906,
+ "learning_rate": 0.0006,
+ "loss": 4.424866676330566,
+ "step": 1485
+ },
+ {
+ "epoch": 20.64279475982533,
+ "grad_norm": 0.01631534844636917,
+ "learning_rate": 0.0006,
+ "loss": 4.441752910614014,
+ "step": 1486
+ },
+ {
+ "epoch": 20.656768558951963,
+ "grad_norm": 0.014057058840990067,
+ "learning_rate": 0.0006,
+ "loss": 4.399982452392578,
+ "step": 1487
+ },
+ {
+ "epoch": 20.670742358078602,
+ "grad_norm": 0.015076273120939732,
+ "learning_rate": 0.0006,
+ "loss": 4.441324234008789,
+ "step": 1488
+ },
+ {
+ "epoch": 20.68471615720524,
+ "grad_norm": 0.014138277620077133,
+ "learning_rate": 0.0006,
+ "loss": 4.4312543869018555,
+ "step": 1489
+ },
+ {
+ "epoch": 20.69868995633188,
+ "grad_norm": 0.013762440532445908,
+ "learning_rate": 0.0006,
+ "loss": 4.387807846069336,
+ "step": 1490
+ },
+ {
+ "epoch": 20.712663755458514,
+ "grad_norm": 0.015058411285281181,
+ "learning_rate": 0.0006,
+ "loss": 4.594747543334961,
+ "step": 1491
+ },
+ {
+ "epoch": 20.726637554585153,
+ "grad_norm": 0.013827123679220676,
+ "learning_rate": 0.0006,
+ "loss": 4.434604167938232,
+ "step": 1492
+ },
+ {
+ "epoch": 20.74061135371179,
+ "grad_norm": 0.015315295197069645,
+ "learning_rate": 0.0006,
+ "loss": 4.588066101074219,
+ "step": 1493
+ },
+ {
+ "epoch": 20.754585152838427,
+ "grad_norm": 0.018678732216358185,
+ "learning_rate": 0.0006,
+ "loss": 4.3514204025268555,
+ "step": 1494
+ },
+ {
+ "epoch": 20.768558951965066,
+ "grad_norm": 0.019521350041031837,
+ "learning_rate": 0.0006,
+ "loss": 4.437658786773682,
+ "step": 1495
+ },
+ {
+ "epoch": 20.782532751091704,
+ "grad_norm": 0.018363403156399727,
+ "learning_rate": 0.0006,
+ "loss": 4.495002269744873,
+ "step": 1496
+ },
+ {
+ "epoch": 20.79650655021834,
+ "grad_norm": 0.016830725595355034,
+ "learning_rate": 0.0006,
+ "loss": 4.402471542358398,
+ "step": 1497
+ },
+ {
+ "epoch": 20.810480349344978,
+ "grad_norm": 0.01722072623670101,
+ "learning_rate": 0.0006,
+ "loss": 4.437653541564941,
+ "step": 1498
+ },
+ {
+ "epoch": 20.824454148471617,
+ "grad_norm": 0.017459599301218987,
+ "learning_rate": 0.0006,
+ "loss": 4.366603851318359,
+ "step": 1499
+ },
+ {
+ "epoch": 20.83842794759825,
+ "grad_norm": 0.018902219831943512,
+ "learning_rate": 0.0006,
+ "loss": 4.501561164855957,
+ "step": 1500
+ },
+ {
+ "epoch": 20.85240174672489,
+ "grad_norm": 0.020814096555113792,
+ "learning_rate": 0.0006,
+ "loss": 4.535295486450195,
+ "step": 1501
+ },
+ {
+ "epoch": 20.86637554585153,
+ "grad_norm": 0.018990976735949516,
+ "learning_rate": 0.0006,
+ "loss": 4.333580017089844,
+ "step": 1502
+ },
+ {
+ "epoch": 20.880349344978168,
+ "grad_norm": 0.01837880350649357,
+ "learning_rate": 0.0006,
+ "loss": 4.313388824462891,
+ "step": 1503
+ },
+ {
+ "epoch": 20.894323144104803,
+ "grad_norm": 0.019820360466837883,
+ "learning_rate": 0.0006,
+ "loss": 4.388157844543457,
+ "step": 1504
+ },
+ {
+ "epoch": 20.90829694323144,
+ "grad_norm": 0.020181143656373024,
+ "learning_rate": 0.0006,
+ "loss": 4.572224140167236,
+ "step": 1505
+ },
+ {
+ "epoch": 20.92227074235808,
+ "grad_norm": 0.021206077188253403,
+ "learning_rate": 0.0006,
+ "loss": 4.3127970695495605,
+ "step": 1506
+ },
+ {
+ "epoch": 20.936244541484715,
+ "grad_norm": 0.017962060868740082,
+ "learning_rate": 0.0006,
+ "loss": 4.384734630584717,
+ "step": 1507
+ },
+ {
+ "epoch": 20.950218340611354,
+ "grad_norm": 0.01737220585346222,
+ "learning_rate": 0.0006,
+ "loss": 4.4075751304626465,
+ "step": 1508
+ },
+ {
+ "epoch": 20.964192139737992,
+ "grad_norm": 0.01839268207550049,
+ "learning_rate": 0.0006,
+ "loss": 4.553779125213623,
+ "step": 1509
+ },
+ {
+ "epoch": 20.978165938864628,
+ "grad_norm": 0.020102640613913536,
+ "learning_rate": 0.0006,
+ "loss": 4.399398326873779,
+ "step": 1510
+ },
+ {
+ "epoch": 20.992139737991266,
+ "grad_norm": 0.018831463530659676,
+ "learning_rate": 0.0006,
+ "loss": 4.358499526977539,
+ "step": 1511
+ },
+ {
+ "epoch": 21.0,
+ "grad_norm": 0.018254734575748444,
+ "learning_rate": 0.0006,
+ "loss": 4.384119510650635,
+ "step": 1512
+ },
+ {
+ "epoch": 21.0,
+ "eval_loss": 4.7524237632751465,
+ "eval_runtime": 57.8812,
+ "eval_samples_per_second": 42.19,
+ "eval_steps_per_second": 1.33,
+ "step": 1512
+ },
+ {
+ "epoch": 21.01397379912664,
+ "grad_norm": 0.019728587940335274,
+ "learning_rate": 0.0006,
+ "loss": 4.32363224029541,
+ "step": 1513
+ },
+ {
+ "epoch": 21.027947598253274,
+ "grad_norm": 0.02216625027358532,
+ "learning_rate": 0.0006,
+ "loss": 4.436988353729248,
+ "step": 1514
+ },
+ {
+ "epoch": 21.041921397379912,
+ "grad_norm": 0.02227606810629368,
+ "learning_rate": 0.0006,
+ "loss": 4.406063556671143,
+ "step": 1515
+ },
+ {
+ "epoch": 21.05589519650655,
+ "grad_norm": 0.020187009125947952,
+ "learning_rate": 0.0006,
+ "loss": 4.324402332305908,
+ "step": 1516
+ },
+ {
+ "epoch": 21.069868995633186,
+ "grad_norm": 0.02095157466828823,
+ "learning_rate": 0.0006,
+ "loss": 4.456544876098633,
+ "step": 1517
+ },
+ {
+ "epoch": 21.083842794759825,
+ "grad_norm": 0.02146013453602791,
+ "learning_rate": 0.0006,
+ "loss": 4.368387222290039,
+ "step": 1518
+ },
+ {
+ "epoch": 21.097816593886463,
+ "grad_norm": 0.02162035182118416,
+ "learning_rate": 0.0006,
+ "loss": 4.48838996887207,
+ "step": 1519
+ },
+ {
+ "epoch": 21.111790393013102,
+ "grad_norm": 0.021745676174759865,
+ "learning_rate": 0.0006,
+ "loss": 4.30778694152832,
+ "step": 1520
+ },
+ {
+ "epoch": 21.125764192139737,
+ "grad_norm": 0.02173588052392006,
+ "learning_rate": 0.0006,
+ "loss": 4.433149814605713,
+ "step": 1521
+ },
+ {
+ "epoch": 21.139737991266376,
+ "grad_norm": 0.020950648933649063,
+ "learning_rate": 0.0006,
+ "loss": 4.520089149475098,
+ "step": 1522
+ },
+ {
+ "epoch": 21.153711790393015,
+ "grad_norm": 0.020684752613306046,
+ "learning_rate": 0.0006,
+ "loss": 4.372437477111816,
+ "step": 1523
+ },
+ {
+ "epoch": 21.16768558951965,
+ "grad_norm": 0.019472869113087654,
+ "learning_rate": 0.0006,
+ "loss": 4.437015533447266,
+ "step": 1524
+ },
+ {
+ "epoch": 21.18165938864629,
+ "grad_norm": 0.01846962980926037,
+ "learning_rate": 0.0006,
+ "loss": 4.449531555175781,
+ "step": 1525
+ },
+ {
+ "epoch": 21.195633187772927,
+ "grad_norm": 0.017377803102135658,
+ "learning_rate": 0.0006,
+ "loss": 4.458279132843018,
+ "step": 1526
+ },
+ {
+ "epoch": 21.209606986899562,
+ "grad_norm": 0.016002200543880463,
+ "learning_rate": 0.0006,
+ "loss": 4.266297817230225,
+ "step": 1527
+ },
+ {
+ "epoch": 21.2235807860262,
+ "grad_norm": 0.015575903467833996,
+ "learning_rate": 0.0006,
+ "loss": 4.342632293701172,
+ "step": 1528
+ },
+ {
+ "epoch": 21.23755458515284,
+ "grad_norm": 0.015634698793292046,
+ "learning_rate": 0.0006,
+ "loss": 4.338623523712158,
+ "step": 1529
+ },
+ {
+ "epoch": 21.251528384279474,
+ "grad_norm": 0.01449581328779459,
+ "learning_rate": 0.0006,
+ "loss": 4.354330062866211,
+ "step": 1530
+ },
+ {
+ "epoch": 21.265502183406113,
+ "grad_norm": 0.01487616915255785,
+ "learning_rate": 0.0006,
+ "loss": 4.486835479736328,
+ "step": 1531
+ },
+ {
+ "epoch": 21.27947598253275,
+ "grad_norm": 0.016346555203199387,
+ "learning_rate": 0.0006,
+ "loss": 4.407098770141602,
+ "step": 1532
+ },
+ {
+ "epoch": 21.29344978165939,
+ "grad_norm": 0.01682182401418686,
+ "learning_rate": 0.0006,
+ "loss": 4.436809539794922,
+ "step": 1533
+ },
+ {
+ "epoch": 21.307423580786025,
+ "grad_norm": 0.017704004421830177,
+ "learning_rate": 0.0006,
+ "loss": 4.3854780197143555,
+ "step": 1534
+ },
+ {
+ "epoch": 21.321397379912664,
+ "grad_norm": 0.018437420949339867,
+ "learning_rate": 0.0006,
+ "loss": 4.413283824920654,
+ "step": 1535
+ },
+ {
+ "epoch": 21.335371179039303,
+ "grad_norm": 0.016911081969738007,
+ "learning_rate": 0.0006,
+ "loss": 4.357306003570557,
+ "step": 1536
+ },
+ {
+ "epoch": 21.349344978165938,
+ "grad_norm": 0.01642223633825779,
+ "learning_rate": 0.0006,
+ "loss": 4.422322750091553,
+ "step": 1537
+ },
+ {
+ "epoch": 21.363318777292577,
+ "grad_norm": 0.01780877821147442,
+ "learning_rate": 0.0006,
+ "loss": 4.387620449066162,
+ "step": 1538
+ },
+ {
+ "epoch": 21.377292576419215,
+ "grad_norm": 0.01653975248336792,
+ "learning_rate": 0.0006,
+ "loss": 4.327267646789551,
+ "step": 1539
+ },
+ {
+ "epoch": 21.39126637554585,
+ "grad_norm": 0.01572156324982643,
+ "learning_rate": 0.0006,
+ "loss": 4.475979804992676,
+ "step": 1540
+ },
+ {
+ "epoch": 21.40524017467249,
+ "grad_norm": 0.017156066372990608,
+ "learning_rate": 0.0006,
+ "loss": 4.4011125564575195,
+ "step": 1541
+ },
+ {
+ "epoch": 21.419213973799128,
+ "grad_norm": 0.01788829080760479,
+ "learning_rate": 0.0006,
+ "loss": 4.4174089431762695,
+ "step": 1542
+ },
+ {
+ "epoch": 21.433187772925763,
+ "grad_norm": 0.01728985086083412,
+ "learning_rate": 0.0006,
+ "loss": 4.429244041442871,
+ "step": 1543
+ },
+ {
+ "epoch": 21.4471615720524,
+ "grad_norm": 0.016534466296434402,
+ "learning_rate": 0.0006,
+ "loss": 4.326068878173828,
+ "step": 1544
+ },
+ {
+ "epoch": 21.46113537117904,
+ "grad_norm": 0.016582584008574486,
+ "learning_rate": 0.0006,
+ "loss": 4.2975969314575195,
+ "step": 1545
+ },
+ {
+ "epoch": 21.475109170305675,
+ "grad_norm": 0.018034178763628006,
+ "learning_rate": 0.0006,
+ "loss": 4.501319408416748,
+ "step": 1546
+ },
+ {
+ "epoch": 21.489082969432314,
+ "grad_norm": 0.017827702686190605,
+ "learning_rate": 0.0006,
+ "loss": 4.502284526824951,
+ "step": 1547
+ },
+ {
+ "epoch": 21.503056768558952,
+ "grad_norm": 0.017182299867272377,
+ "learning_rate": 0.0006,
+ "loss": 4.380828857421875,
+ "step": 1548
+ },
+ {
+ "epoch": 21.51703056768559,
+ "grad_norm": 0.019101588055491447,
+ "learning_rate": 0.0006,
+ "loss": 4.31564998626709,
+ "step": 1549
+ },
+ {
+ "epoch": 21.531004366812226,
+ "grad_norm": 0.0204677302390337,
+ "learning_rate": 0.0006,
+ "loss": 4.3840131759643555,
+ "step": 1550
+ },
+ {
+ "epoch": 21.544978165938865,
+ "grad_norm": 0.0188993439078331,
+ "learning_rate": 0.0006,
+ "loss": 4.4257731437683105,
+ "step": 1551
+ },
+ {
+ "epoch": 21.558951965065503,
+ "grad_norm": 0.01730230636894703,
+ "learning_rate": 0.0006,
+ "loss": 4.445132255554199,
+ "step": 1552
+ },
+ {
+ "epoch": 21.57292576419214,
+ "grad_norm": 0.01801574043929577,
+ "learning_rate": 0.0006,
+ "loss": 4.427064895629883,
+ "step": 1553
+ },
+ {
+ "epoch": 21.586899563318777,
+ "grad_norm": 0.018417565152049065,
+ "learning_rate": 0.0006,
+ "loss": 4.319595813751221,
+ "step": 1554
+ },
+ {
+ "epoch": 21.600873362445416,
+ "grad_norm": 0.01683160290122032,
+ "learning_rate": 0.0006,
+ "loss": 4.349218368530273,
+ "step": 1555
+ },
+ {
+ "epoch": 21.61484716157205,
+ "grad_norm": 0.016973106190562248,
+ "learning_rate": 0.0006,
+ "loss": 4.375990390777588,
+ "step": 1556
+ },
+ {
+ "epoch": 21.62882096069869,
+ "grad_norm": 0.015688583254814148,
+ "learning_rate": 0.0006,
+ "loss": 4.408170223236084,
+ "step": 1557
+ },
+ {
+ "epoch": 21.64279475982533,
+ "grad_norm": 0.01573043316602707,
+ "learning_rate": 0.0006,
+ "loss": 4.416255950927734,
+ "step": 1558
+ },
+ {
+ "epoch": 21.656768558951963,
+ "grad_norm": 0.01744293048977852,
+ "learning_rate": 0.0006,
+ "loss": 4.477141380310059,
+ "step": 1559
+ },
+ {
+ "epoch": 21.670742358078602,
+ "grad_norm": 0.017503513023257256,
+ "learning_rate": 0.0006,
+ "loss": 4.3367486000061035,
+ "step": 1560
+ },
+ {
+ "epoch": 21.68471615720524,
+ "grad_norm": 0.016592150554060936,
+ "learning_rate": 0.0006,
+ "loss": 4.314785480499268,
+ "step": 1561
+ },
+ {
+ "epoch": 21.69868995633188,
+ "grad_norm": 0.015774084255099297,
+ "learning_rate": 0.0006,
+ "loss": 4.448209285736084,
+ "step": 1562
+ },
+ {
+ "epoch": 21.712663755458514,
+ "grad_norm": 0.015779150649905205,
+ "learning_rate": 0.0006,
+ "loss": 4.392690658569336,
+ "step": 1563
+ },
+ {
+ "epoch": 21.726637554585153,
+ "grad_norm": 0.015392648056149483,
+ "learning_rate": 0.0006,
+ "loss": 4.437741756439209,
+ "step": 1564
+ },
+ {
+ "epoch": 21.74061135371179,
+ "grad_norm": 0.01530836895108223,
+ "learning_rate": 0.0006,
+ "loss": 4.424595832824707,
+ "step": 1565
+ },
+ {
+ "epoch": 21.754585152838427,
+ "grad_norm": 0.015600779093801975,
+ "learning_rate": 0.0006,
+ "loss": 4.335417747497559,
+ "step": 1566
+ },
+ {
+ "epoch": 21.768558951965066,
+ "grad_norm": 0.014646518044173717,
+ "learning_rate": 0.0006,
+ "loss": 4.279081344604492,
+ "step": 1567
+ },
+ {
+ "epoch": 21.782532751091704,
+ "grad_norm": 0.014132864773273468,
+ "learning_rate": 0.0006,
+ "loss": 4.345950126647949,
+ "step": 1568
+ },
+ {
+ "epoch": 21.79650655021834,
+ "grad_norm": 0.01425724383443594,
+ "learning_rate": 0.0006,
+ "loss": 4.277588844299316,
+ "step": 1569
+ },
+ {
+ "epoch": 21.810480349344978,
+ "grad_norm": 0.015396572649478912,
+ "learning_rate": 0.0006,
+ "loss": 4.326196670532227,
+ "step": 1570
+ },
+ {
+ "epoch": 21.824454148471617,
+ "grad_norm": 0.01683669537305832,
+ "learning_rate": 0.0006,
+ "loss": 4.3294172286987305,
+ "step": 1571
+ },
+ {
+ "epoch": 21.83842794759825,
+ "grad_norm": 0.016365647315979004,
+ "learning_rate": 0.0006,
+ "loss": 4.523321628570557,
+ "step": 1572
+ },
+ {
+ "epoch": 21.85240174672489,
+ "grad_norm": 0.017142774537205696,
+ "learning_rate": 0.0006,
+ "loss": 4.475842475891113,
+ "step": 1573
+ },
+ {
+ "epoch": 21.86637554585153,
+ "grad_norm": 0.015950839966535568,
+ "learning_rate": 0.0006,
+ "loss": 4.397187232971191,
+ "step": 1574
+ },
+ {
+ "epoch": 21.880349344978168,
+ "grad_norm": 0.016180362552404404,
+ "learning_rate": 0.0006,
+ "loss": 4.414778709411621,
+ "step": 1575
+ },
+ {
+ "epoch": 21.894323144104803,
+ "grad_norm": 0.017615556716918945,
+ "learning_rate": 0.0006,
+ "loss": 4.348567962646484,
+ "step": 1576
+ },
+ {
+ "epoch": 21.90829694323144,
+ "grad_norm": 0.01896277628839016,
+ "learning_rate": 0.0006,
+ "loss": 4.431762218475342,
+ "step": 1577
+ },
+ {
+ "epoch": 21.92227074235808,
+ "grad_norm": 0.01990971714258194,
+ "learning_rate": 0.0006,
+ "loss": 4.431285858154297,
+ "step": 1578
+ },
+ {
+ "epoch": 21.936244541484715,
+ "grad_norm": 0.0201258547604084,
+ "learning_rate": 0.0006,
+ "loss": 4.309872627258301,
+ "step": 1579
+ },
+ {
+ "epoch": 21.950218340611354,
+ "grad_norm": 0.019620198756456375,
+ "learning_rate": 0.0006,
+ "loss": 4.484044551849365,
+ "step": 1580
+ },
+ {
+ "epoch": 21.964192139737992,
+ "grad_norm": 0.01771123707294464,
+ "learning_rate": 0.0006,
+ "loss": 4.368093013763428,
+ "step": 1581
+ },
+ {
+ "epoch": 21.978165938864628,
+ "grad_norm": 0.017501654103398323,
+ "learning_rate": 0.0006,
+ "loss": 4.372651100158691,
+ "step": 1582
+ },
+ {
+ "epoch": 21.992139737991266,
+ "grad_norm": 0.01693608984351158,
+ "learning_rate": 0.0006,
+ "loss": 4.379167556762695,
+ "step": 1583
+ },
+ {
+ "epoch": 22.0,
+ "grad_norm": 0.02002849616110325,
+ "learning_rate": 0.0006,
+ "loss": 4.329627990722656,
+ "step": 1584
+ },
+ {
+ "epoch": 22.0,
+ "eval_loss": 4.723696231842041,
+ "eval_runtime": 56.1758,
+ "eval_samples_per_second": 43.471,
+ "eval_steps_per_second": 1.371,
+ "step": 1584
+ },
+ {
+ "epoch": 22.01397379912664,
+ "grad_norm": 0.02030305378139019,
+ "learning_rate": 0.0006,
+ "loss": 4.355566024780273,
+ "step": 1585
+ },
+ {
+ "epoch": 22.027947598253274,
+ "grad_norm": 0.018638934940099716,
+ "learning_rate": 0.0006,
+ "loss": 4.314009189605713,
+ "step": 1586
+ },
+ {
+ "epoch": 22.041921397379912,
+ "grad_norm": 0.019917072728276253,
+ "learning_rate": 0.0006,
+ "loss": 4.367525577545166,
+ "step": 1587
+ },
+ {
+ "epoch": 22.05589519650655,
+ "grad_norm": 0.020785167813301086,
+ "learning_rate": 0.0006,
+ "loss": 4.375561714172363,
+ "step": 1588
+ },
+ {
+ "epoch": 22.069868995633186,
+ "grad_norm": 0.021214453503489494,
+ "learning_rate": 0.0006,
+ "loss": 4.418305397033691,
+ "step": 1589
+ },
+ {
+ "epoch": 22.083842794759825,
+ "grad_norm": 0.019051678478717804,
+ "learning_rate": 0.0006,
+ "loss": 4.215126991271973,
+ "step": 1590
+ },
+ {
+ "epoch": 22.097816593886463,
+ "grad_norm": 0.02323366142809391,
+ "learning_rate": 0.0006,
+ "loss": 4.306807994842529,
+ "step": 1591
+ },
+ {
+ "epoch": 22.111790393013102,
+ "grad_norm": 0.026186738163232803,
+ "learning_rate": 0.0006,
+ "loss": 4.408508777618408,
+ "step": 1592
+ },
+ {
+ "epoch": 22.125764192139737,
+ "grad_norm": 0.027361227199435234,
+ "learning_rate": 0.0006,
+ "loss": 4.329322814941406,
+ "step": 1593
+ },
+ {
+ "epoch": 22.139737991266376,
+ "grad_norm": 0.026564646512269974,
+ "learning_rate": 0.0006,
+ "loss": 4.346642017364502,
+ "step": 1594
+ },
+ {
+ "epoch": 22.153711790393015,
+ "grad_norm": 0.025073667988181114,
+ "learning_rate": 0.0006,
+ "loss": 4.3984198570251465,
+ "step": 1595
+ },
+ {
+ "epoch": 22.16768558951965,
+ "grad_norm": 0.02347794733941555,
+ "learning_rate": 0.0006,
+ "loss": 4.352538108825684,
+ "step": 1596
+ },
+ {
+ "epoch": 22.18165938864629,
+ "grad_norm": 0.02203468047082424,
+ "learning_rate": 0.0006,
+ "loss": 4.298148155212402,
+ "step": 1597
+ },
+ {
+ "epoch": 22.195633187772927,
+ "grad_norm": 0.020621957257390022,
+ "learning_rate": 0.0006,
+ "loss": 4.480309963226318,
+ "step": 1598
+ },
+ {
+ "epoch": 22.209606986899562,
+ "grad_norm": 0.023710055276751518,
+ "learning_rate": 0.0006,
+ "loss": 4.339317321777344,
+ "step": 1599
+ },
+ {
+ "epoch": 22.2235807860262,
+ "grad_norm": 0.0233482476323843,
+ "learning_rate": 0.0006,
+ "loss": 4.436896324157715,
+ "step": 1600
+ },
+ {
+ "epoch": 22.23755458515284,
+ "grad_norm": 0.020122403278946877,
+ "learning_rate": 0.0006,
+ "loss": 4.236002445220947,
+ "step": 1601
+ },
+ {
+ "epoch": 22.251528384279474,
+ "grad_norm": 0.01957266591489315,
+ "learning_rate": 0.0006,
+ "loss": 4.497328281402588,
+ "step": 1602
+ },
+ {
+ "epoch": 22.265502183406113,
+ "grad_norm": 0.018964895978569984,
+ "learning_rate": 0.0006,
+ "loss": 4.36360502243042,
+ "step": 1603
+ },
+ {
+ "epoch": 22.27947598253275,
+ "grad_norm": 0.018902769312262535,
+ "learning_rate": 0.0006,
+ "loss": 4.427093505859375,
+ "step": 1604
+ },
+ {
+ "epoch": 22.29344978165939,
+ "grad_norm": 0.018918119370937347,
+ "learning_rate": 0.0006,
+ "loss": 4.294950485229492,
+ "step": 1605
+ },
+ {
+ "epoch": 22.307423580786025,
+ "grad_norm": 0.019315775483846664,
+ "learning_rate": 0.0006,
+ "loss": 4.427952289581299,
+ "step": 1606
+ },
+ {
+ "epoch": 22.321397379912664,
+ "grad_norm": 0.018350522965192795,
+ "learning_rate": 0.0006,
+ "loss": 4.272945404052734,
+ "step": 1607
+ },
+ {
+ "epoch": 22.335371179039303,
+ "grad_norm": 0.01759534701704979,
+ "learning_rate": 0.0006,
+ "loss": 4.348023891448975,
+ "step": 1608
+ },
+ {
+ "epoch": 22.349344978165938,
+ "grad_norm": 0.01646057888865471,
+ "learning_rate": 0.0006,
+ "loss": 4.278518199920654,
+ "step": 1609
+ },
+ {
+ "epoch": 22.363318777292577,
+ "grad_norm": 0.015875060111284256,
+ "learning_rate": 0.0006,
+ "loss": 4.423527240753174,
+ "step": 1610
+ },
+ {
+ "epoch": 22.377292576419215,
+ "grad_norm": 0.01638529635965824,
+ "learning_rate": 0.0006,
+ "loss": 4.321723937988281,
+ "step": 1611
+ },
+ {
+ "epoch": 22.39126637554585,
+ "grad_norm": 0.016942061483860016,
+ "learning_rate": 0.0006,
+ "loss": 4.346164703369141,
+ "step": 1612
+ },
+ {
+ "epoch": 22.40524017467249,
+ "grad_norm": 0.015209621749818325,
+ "learning_rate": 0.0006,
+ "loss": 4.434340000152588,
+ "step": 1613
+ },
+ {
+ "epoch": 22.419213973799128,
+ "grad_norm": 0.014423009008169174,
+ "learning_rate": 0.0006,
+ "loss": 4.342560768127441,
+ "step": 1614
+ },
+ {
+ "epoch": 22.433187772925763,
+ "grad_norm": 0.014369525946676731,
+ "learning_rate": 0.0006,
+ "loss": 4.296586990356445,
+ "step": 1615
+ },
+ {
+ "epoch": 22.4471615720524,
+ "grad_norm": 0.01392540242522955,
+ "learning_rate": 0.0006,
+ "loss": 4.433176040649414,
+ "step": 1616
+ },
+ {
+ "epoch": 22.46113537117904,
+ "grad_norm": 0.013206899166107178,
+ "learning_rate": 0.0006,
+ "loss": 4.32962703704834,
+ "step": 1617
+ },
+ {
+ "epoch": 22.475109170305675,
+ "grad_norm": 0.013427077792584896,
+ "learning_rate": 0.0006,
+ "loss": 4.420437335968018,
+ "step": 1618
+ },
+ {
+ "epoch": 22.489082969432314,
+ "grad_norm": 0.014194196090102196,
+ "learning_rate": 0.0006,
+ "loss": 4.410478591918945,
+ "step": 1619
+ },
+ {
+ "epoch": 22.503056768558952,
+ "grad_norm": 0.01362364087253809,
+ "learning_rate": 0.0006,
+ "loss": 4.338932037353516,
+ "step": 1620
+ },
+ {
+ "epoch": 22.51703056768559,
+ "grad_norm": 0.014299589209258556,
+ "learning_rate": 0.0006,
+ "loss": 4.34089994430542,
+ "step": 1621
+ },
+ {
+ "epoch": 22.531004366812226,
+ "grad_norm": 0.014061295427381992,
+ "learning_rate": 0.0006,
+ "loss": 4.29201602935791,
+ "step": 1622
+ },
+ {
+ "epoch": 22.544978165938865,
+ "grad_norm": 0.013945229351520538,
+ "learning_rate": 0.0006,
+ "loss": 4.326052188873291,
+ "step": 1623
+ },
+ {
+ "epoch": 22.558951965065503,
+ "grad_norm": 0.01387068536132574,
+ "learning_rate": 0.0006,
+ "loss": 4.393521308898926,
+ "step": 1624
+ },
+ {
+ "epoch": 22.57292576419214,
+ "grad_norm": 0.013931898400187492,
+ "learning_rate": 0.0006,
+ "loss": 4.3739752769470215,
+ "step": 1625
+ },
+ {
+ "epoch": 22.586899563318777,
+ "grad_norm": 0.01497740764170885,
+ "learning_rate": 0.0006,
+ "loss": 4.3991618156433105,
+ "step": 1626
+ },
+ {
+ "epoch": 22.600873362445416,
+ "grad_norm": 0.015842905268073082,
+ "learning_rate": 0.0006,
+ "loss": 4.405340671539307,
+ "step": 1627
+ },
+ {
+ "epoch": 22.61484716157205,
+ "grad_norm": 0.01580633595585823,
+ "learning_rate": 0.0006,
+ "loss": 4.280102729797363,
+ "step": 1628
+ },
+ {
+ "epoch": 22.62882096069869,
+ "grad_norm": 0.016567382961511612,
+ "learning_rate": 0.0006,
+ "loss": 4.407657146453857,
+ "step": 1629
+ },
+ {
+ "epoch": 22.64279475982533,
+ "grad_norm": 0.01648622378706932,
+ "learning_rate": 0.0006,
+ "loss": 4.392590522766113,
+ "step": 1630
+ },
+ {
+ "epoch": 22.656768558951963,
+ "grad_norm": 0.017066307365894318,
+ "learning_rate": 0.0006,
+ "loss": 4.324862003326416,
+ "step": 1631
+ },
+ {
+ "epoch": 22.670742358078602,
+ "grad_norm": 0.01637062057852745,
+ "learning_rate": 0.0006,
+ "loss": 4.34598445892334,
+ "step": 1632
+ },
+ {
+ "epoch": 22.68471615720524,
+ "grad_norm": 0.018038960173726082,
+ "learning_rate": 0.0006,
+ "loss": 4.320711612701416,
+ "step": 1633
+ },
+ {
+ "epoch": 22.69868995633188,
+ "grad_norm": 0.018096428364515305,
+ "learning_rate": 0.0006,
+ "loss": 4.536527156829834,
+ "step": 1634
+ },
+ {
+ "epoch": 22.712663755458514,
+ "grad_norm": 0.017889291048049927,
+ "learning_rate": 0.0006,
+ "loss": 4.388131141662598,
+ "step": 1635
+ },
+ {
+ "epoch": 22.726637554585153,
+ "grad_norm": 0.018771521747112274,
+ "learning_rate": 0.0006,
+ "loss": 4.357668876647949,
+ "step": 1636
+ },
+ {
+ "epoch": 22.74061135371179,
+ "grad_norm": 0.02100873365998268,
+ "learning_rate": 0.0006,
+ "loss": 4.293972015380859,
+ "step": 1637
+ },
+ {
+ "epoch": 22.754585152838427,
+ "grad_norm": 0.017715785652399063,
+ "learning_rate": 0.0006,
+ "loss": 4.349849224090576,
+ "step": 1638
+ },
+ {
+ "epoch": 22.768558951965066,
+ "grad_norm": 0.015546333976089954,
+ "learning_rate": 0.0006,
+ "loss": 4.311728000640869,
+ "step": 1639
+ },
+ {
+ "epoch": 22.782532751091704,
+ "grad_norm": 0.01605990156531334,
+ "learning_rate": 0.0006,
+ "loss": 4.499123573303223,
+ "step": 1640
+ },
+ {
+ "epoch": 22.79650655021834,
+ "grad_norm": 0.016642142087221146,
+ "learning_rate": 0.0006,
+ "loss": 4.4319257736206055,
+ "step": 1641
+ },
+ {
+ "epoch": 22.810480349344978,
+ "grad_norm": 0.016120387241244316,
+ "learning_rate": 0.0006,
+ "loss": 4.4373650550842285,
+ "step": 1642
+ },
+ {
+ "epoch": 22.824454148471617,
+ "grad_norm": 0.015174890868365765,
+ "learning_rate": 0.0006,
+ "loss": 4.314082145690918,
+ "step": 1643
+ },
+ {
+ "epoch": 22.83842794759825,
+ "grad_norm": 0.01551714539527893,
+ "learning_rate": 0.0006,
+ "loss": 4.360363960266113,
+ "step": 1644
+ },
+ {
+ "epoch": 22.85240174672489,
+ "grad_norm": 0.014966742135584354,
+ "learning_rate": 0.0006,
+ "loss": 4.346253395080566,
+ "step": 1645
+ },
+ {
+ "epoch": 22.86637554585153,
+ "grad_norm": 0.014288028702139854,
+ "learning_rate": 0.0006,
+ "loss": 4.464366436004639,
+ "step": 1646
+ },
+ {
+ "epoch": 22.880349344978168,
+ "grad_norm": 0.015291682444512844,
+ "learning_rate": 0.0006,
+ "loss": 4.504101753234863,
+ "step": 1647
+ },
+ {
+ "epoch": 22.894323144104803,
+ "grad_norm": 0.014534426853060722,
+ "learning_rate": 0.0006,
+ "loss": 4.462172031402588,
+ "step": 1648
+ },
+ {
+ "epoch": 22.90829694323144,
+ "grad_norm": 0.014129819348454475,
+ "learning_rate": 0.0006,
+ "loss": 4.350126266479492,
+ "step": 1649
+ },
+ {
+ "epoch": 22.92227074235808,
+ "grad_norm": 0.014932668767869473,
+ "learning_rate": 0.0006,
+ "loss": 4.374392032623291,
+ "step": 1650
+ },
+ {
+ "epoch": 22.936244541484715,
+ "grad_norm": 0.014647853560745716,
+ "learning_rate": 0.0006,
+ "loss": 4.252264499664307,
+ "step": 1651
+ },
+ {
+ "epoch": 22.950218340611354,
+ "grad_norm": 0.013553360477089882,
+ "learning_rate": 0.0006,
+ "loss": 4.379729270935059,
+ "step": 1652
+ },
+ {
+ "epoch": 22.964192139737992,
+ "grad_norm": 0.014606194570660591,
+ "learning_rate": 0.0006,
+ "loss": 4.363372802734375,
+ "step": 1653
+ },
+ {
+ "epoch": 22.978165938864628,
+ "grad_norm": 0.015000631101429462,
+ "learning_rate": 0.0006,
+ "loss": 4.408048629760742,
+ "step": 1654
+ },
+ {
+ "epoch": 22.992139737991266,
+ "grad_norm": 0.015370816923677921,
+ "learning_rate": 0.0006,
+ "loss": 4.297451019287109,
+ "step": 1655
+ },
+ {
+ "epoch": 23.0,
+ "grad_norm": 0.01709144562482834,
+ "learning_rate": 0.0006,
+ "loss": 4.294776916503906,
+ "step": 1656
+ },
+ {
+ "epoch": 23.0,
+ "eval_loss": 4.678618907928467,
+ "eval_runtime": 56.1668,
+ "eval_samples_per_second": 43.478,
+ "eval_steps_per_second": 1.371,
+ "step": 1656
+ },
+ {
+ "epoch": 23.01397379912664,
+ "grad_norm": 0.01727290451526642,
+ "learning_rate": 0.0006,
+ "loss": 4.311701774597168,
+ "step": 1657
+ },
+ {
+ "epoch": 23.027947598253274,
+ "grad_norm": 0.020645759999752045,
+ "learning_rate": 0.0006,
+ "loss": 4.429897308349609,
+ "step": 1658
+ },
+ {
+ "epoch": 23.041921397379912,
+ "grad_norm": 0.02204226516187191,
+ "learning_rate": 0.0006,
+ "loss": 4.365151882171631,
+ "step": 1659
+ },
+ {
+ "epoch": 23.05589519650655,
+ "grad_norm": 0.023069269955158234,
+ "learning_rate": 0.0006,
+ "loss": 4.3710832595825195,
+ "step": 1660
+ },
+ {
+ "epoch": 23.069868995633186,
+ "grad_norm": 0.024151179939508438,
+ "learning_rate": 0.0006,
+ "loss": 4.403108596801758,
+ "step": 1661
+ },
+ {
+ "epoch": 23.083842794759825,
+ "grad_norm": 0.02511233650147915,
+ "learning_rate": 0.0006,
+ "loss": 4.414128303527832,
+ "step": 1662
+ },
+ {
+ "epoch": 23.097816593886463,
+ "grad_norm": 0.025802113115787506,
+ "learning_rate": 0.0006,
+ "loss": 4.388265609741211,
+ "step": 1663
+ },
+ {
+ "epoch": 23.111790393013102,
+ "grad_norm": 0.02377348765730858,
+ "learning_rate": 0.0006,
+ "loss": 4.357505798339844,
+ "step": 1664
+ },
+ {
+ "epoch": 23.125764192139737,
+ "grad_norm": 0.020876651629805565,
+ "learning_rate": 0.0006,
+ "loss": 4.23043966293335,
+ "step": 1665
+ },
+ {
+ "epoch": 23.139737991266376,
+ "grad_norm": 0.01959727331995964,
+ "learning_rate": 0.0006,
+ "loss": 4.280970096588135,
+ "step": 1666
+ },
+ {
+ "epoch": 23.153711790393015,
+ "grad_norm": 0.018678616732358932,
+ "learning_rate": 0.0006,
+ "loss": 4.346716403961182,
+ "step": 1667
+ },
+ {
+ "epoch": 23.16768558951965,
+ "grad_norm": 0.02108944021165371,
+ "learning_rate": 0.0006,
+ "loss": 4.277904510498047,
+ "step": 1668
+ },
+ {
+ "epoch": 23.18165938864629,
+ "grad_norm": 0.020221566781401634,
+ "learning_rate": 0.0006,
+ "loss": 4.367693901062012,
+ "step": 1669
+ },
+ {
+ "epoch": 23.195633187772927,
+ "grad_norm": 0.019599711522459984,
+ "learning_rate": 0.0006,
+ "loss": 4.257092475891113,
+ "step": 1670
+ },
+ {
+ "epoch": 23.209606986899562,
+ "grad_norm": 0.019833361729979515,
+ "learning_rate": 0.0006,
+ "loss": 4.540737628936768,
+ "step": 1671
+ },
+ {
+ "epoch": 23.2235807860262,
+ "grad_norm": 0.020352095365524292,
+ "learning_rate": 0.0006,
+ "loss": 4.360923767089844,
+ "step": 1672
+ },
+ {
+ "epoch": 23.23755458515284,
+ "grad_norm": 0.01939748041331768,
+ "learning_rate": 0.0006,
+ "loss": 4.393110275268555,
+ "step": 1673
+ },
+ {
+ "epoch": 23.251528384279474,
+ "grad_norm": 0.01884354278445244,
+ "learning_rate": 0.0006,
+ "loss": 4.383938789367676,
+ "step": 1674
+ },
+ {
+ "epoch": 23.265502183406113,
+ "grad_norm": 0.02008470520377159,
+ "learning_rate": 0.0006,
+ "loss": 4.346344947814941,
+ "step": 1675
+ },
+ {
+ "epoch": 23.27947598253275,
+ "grad_norm": 0.019065195694565773,
+ "learning_rate": 0.0006,
+ "loss": 4.36611795425415,
+ "step": 1676
+ },
+ {
+ "epoch": 23.29344978165939,
+ "grad_norm": 0.01938541606068611,
+ "learning_rate": 0.0006,
+ "loss": 4.4009599685668945,
+ "step": 1677
+ },
+ {
+ "epoch": 23.307423580786025,
+ "grad_norm": 0.019044723361730576,
+ "learning_rate": 0.0006,
+ "loss": 4.340169906616211,
+ "step": 1678
+ },
+ {
+ "epoch": 23.321397379912664,
+ "grad_norm": 0.018850378692150116,
+ "learning_rate": 0.0006,
+ "loss": 4.37674617767334,
+ "step": 1679
+ },
+ {
+ "epoch": 23.335371179039303,
+ "grad_norm": 0.019076095893979073,
+ "learning_rate": 0.0006,
+ "loss": 4.226874351501465,
+ "step": 1680
+ },
+ {
+ "epoch": 23.349344978165938,
+ "grad_norm": 0.017543038353323936,
+ "learning_rate": 0.0006,
+ "loss": 4.337751388549805,
+ "step": 1681
+ },
+ {
+ "epoch": 23.363318777292577,
+ "grad_norm": 0.017170218750834465,
+ "learning_rate": 0.0006,
+ "loss": 4.282512664794922,
+ "step": 1682
+ },
+ {
+ "epoch": 23.377292576419215,
+ "grad_norm": 0.01574796624481678,
+ "learning_rate": 0.0006,
+ "loss": 4.30638313293457,
+ "step": 1683
+ },
+ {
+ "epoch": 23.39126637554585,
+ "grad_norm": 0.015160846523940563,
+ "learning_rate": 0.0006,
+ "loss": 4.336716175079346,
+ "step": 1684
+ },
+ {
+ "epoch": 23.40524017467249,
+ "grad_norm": 0.01534526702016592,
+ "learning_rate": 0.0006,
+ "loss": 4.2174787521362305,
+ "step": 1685
+ },
+ {
+ "epoch": 23.419213973799128,
+ "grad_norm": 0.015289999544620514,
+ "learning_rate": 0.0006,
+ "loss": 4.272546291351318,
+ "step": 1686
+ },
+ {
+ "epoch": 23.433187772925763,
+ "grad_norm": 0.014894185587763786,
+ "learning_rate": 0.0006,
+ "loss": 4.296045303344727,
+ "step": 1687
+ },
+ {
+ "epoch": 23.4471615720524,
+ "grad_norm": 0.014869502745568752,
+ "learning_rate": 0.0006,
+ "loss": 4.425461769104004,
+ "step": 1688
+ },
+ {
+ "epoch": 23.46113537117904,
+ "grad_norm": 0.01474483497440815,
+ "learning_rate": 0.0006,
+ "loss": 4.361867427825928,
+ "step": 1689
+ },
+ {
+ "epoch": 23.475109170305675,
+ "grad_norm": 0.015299245715141296,
+ "learning_rate": 0.0006,
+ "loss": 4.364523410797119,
+ "step": 1690
+ },
+ {
+ "epoch": 23.489082969432314,
+ "grad_norm": 0.015201340429484844,
+ "learning_rate": 0.0006,
+ "loss": 4.285182952880859,
+ "step": 1691
+ },
+ {
+ "epoch": 23.503056768558952,
+ "grad_norm": 0.016057293862104416,
+ "learning_rate": 0.0006,
+ "loss": 4.284873962402344,
+ "step": 1692
+ },
+ {
+ "epoch": 23.51703056768559,
+ "grad_norm": 0.015871062874794006,
+ "learning_rate": 0.0006,
+ "loss": 4.326470851898193,
+ "step": 1693
+ },
+ {
+ "epoch": 23.531004366812226,
+ "grad_norm": 0.01586288772523403,
+ "learning_rate": 0.0006,
+ "loss": 4.324883460998535,
+ "step": 1694
+ },
+ {
+ "epoch": 23.544978165938865,
+ "grad_norm": 0.016092827543616295,
+ "learning_rate": 0.0006,
+ "loss": 4.3258867263793945,
+ "step": 1695
+ },
+ {
+ "epoch": 23.558951965065503,
+ "grad_norm": 0.01582016795873642,
+ "learning_rate": 0.0006,
+ "loss": 4.358892440795898,
+ "step": 1696
+ },
+ {
+ "epoch": 23.57292576419214,
+ "grad_norm": 0.01688452623784542,
+ "learning_rate": 0.0006,
+ "loss": 4.280842304229736,
+ "step": 1697
+ },
+ {
+ "epoch": 23.586899563318777,
+ "grad_norm": 0.015865886583924294,
+ "learning_rate": 0.0006,
+ "loss": 4.245567321777344,
+ "step": 1698
+ },
+ {
+ "epoch": 23.600873362445416,
+ "grad_norm": 0.015303662046790123,
+ "learning_rate": 0.0006,
+ "loss": 4.383112907409668,
+ "step": 1699
+ },
+ {
+ "epoch": 23.61484716157205,
+ "grad_norm": 0.015649516135454178,
+ "learning_rate": 0.0006,
+ "loss": 4.519082069396973,
+ "step": 1700
+ },
+ {
+ "epoch": 23.62882096069869,
+ "grad_norm": 0.0167120099067688,
+ "learning_rate": 0.0006,
+ "loss": 4.447847366333008,
+ "step": 1701
+ },
+ {
+ "epoch": 23.64279475982533,
+ "grad_norm": 0.016368716955184937,
+ "learning_rate": 0.0006,
+ "loss": 4.434186935424805,
+ "step": 1702
+ },
+ {
+ "epoch": 23.656768558951963,
+ "grad_norm": 0.017379503697156906,
+ "learning_rate": 0.0006,
+ "loss": 4.317594528198242,
+ "step": 1703
+ },
+ {
+ "epoch": 23.670742358078602,
+ "grad_norm": 0.018573811277747154,
+ "learning_rate": 0.0006,
+ "loss": 4.205793857574463,
+ "step": 1704
+ },
+ {
+ "epoch": 23.68471615720524,
+ "grad_norm": 0.01749090477824211,
+ "learning_rate": 0.0006,
+ "loss": 4.3985700607299805,
+ "step": 1705
+ },
+ {
+ "epoch": 23.69868995633188,
+ "grad_norm": 0.0165668074041605,
+ "learning_rate": 0.0006,
+ "loss": 4.329226970672607,
+ "step": 1706
+ },
+ {
+ "epoch": 23.712663755458514,
+ "grad_norm": 0.016913846135139465,
+ "learning_rate": 0.0006,
+ "loss": 4.388485908508301,
+ "step": 1707
+ },
+ {
+ "epoch": 23.726637554585153,
+ "grad_norm": 0.016813859343528748,
+ "learning_rate": 0.0006,
+ "loss": 4.3395915031433105,
+ "step": 1708
+ },
+ {
+ "epoch": 23.74061135371179,
+ "grad_norm": 0.01676975190639496,
+ "learning_rate": 0.0006,
+ "loss": 4.254745960235596,
+ "step": 1709
+ },
+ {
+ "epoch": 23.754585152838427,
+ "grad_norm": 0.016370350494980812,
+ "learning_rate": 0.0006,
+ "loss": 4.2612223625183105,
+ "step": 1710
+ },
+ {
+ "epoch": 23.768558951965066,
+ "grad_norm": 0.01696198247373104,
+ "learning_rate": 0.0006,
+ "loss": 4.327524662017822,
+ "step": 1711
+ },
+ {
+ "epoch": 23.782532751091704,
+ "grad_norm": 0.0189108457416296,
+ "learning_rate": 0.0006,
+ "loss": 4.325839519500732,
+ "step": 1712
+ },
+ {
+ "epoch": 23.79650655021834,
+ "grad_norm": 0.01862351782619953,
+ "learning_rate": 0.0006,
+ "loss": 4.333662509918213,
+ "step": 1713
+ },
+ {
+ "epoch": 23.810480349344978,
+ "grad_norm": 0.01655014418065548,
+ "learning_rate": 0.0006,
+ "loss": 4.39305305480957,
+ "step": 1714
+ },
+ {
+ "epoch": 23.824454148471617,
+ "grad_norm": 0.015899376943707466,
+ "learning_rate": 0.0006,
+ "loss": 4.390053749084473,
+ "step": 1715
+ },
+ {
+ "epoch": 23.83842794759825,
+ "grad_norm": 0.016519056633114815,
+ "learning_rate": 0.0006,
+ "loss": 4.255903244018555,
+ "step": 1716
+ },
+ {
+ "epoch": 23.85240174672489,
+ "grad_norm": 0.0167539119720459,
+ "learning_rate": 0.0006,
+ "loss": 4.255727767944336,
+ "step": 1717
+ },
+ {
+ "epoch": 23.86637554585153,
+ "grad_norm": 0.0171353816986084,
+ "learning_rate": 0.0006,
+ "loss": 4.376740455627441,
+ "step": 1718
+ },
+ {
+ "epoch": 23.880349344978168,
+ "grad_norm": 0.016719117760658264,
+ "learning_rate": 0.0006,
+ "loss": 4.3376688957214355,
+ "step": 1719
+ },
+ {
+ "epoch": 23.894323144104803,
+ "grad_norm": 0.01561494916677475,
+ "learning_rate": 0.0006,
+ "loss": 4.3345746994018555,
+ "step": 1720
+ },
+ {
+ "epoch": 23.90829694323144,
+ "grad_norm": 0.016303174197673798,
+ "learning_rate": 0.0006,
+ "loss": 4.332371234893799,
+ "step": 1721
+ },
+ {
+ "epoch": 23.92227074235808,
+ "grad_norm": 0.016722865402698517,
+ "learning_rate": 0.0006,
+ "loss": 4.267125129699707,
+ "step": 1722
+ },
+ {
+ "epoch": 23.936244541484715,
+ "grad_norm": 0.017072124406695366,
+ "learning_rate": 0.0006,
+ "loss": 4.322751998901367,
+ "step": 1723
+ },
+ {
+ "epoch": 23.950218340611354,
+ "grad_norm": 0.01699932850897312,
+ "learning_rate": 0.0006,
+ "loss": 4.378422737121582,
+ "step": 1724
+ },
+ {
+ "epoch": 23.964192139737992,
+ "grad_norm": 0.01742720976471901,
+ "learning_rate": 0.0006,
+ "loss": 4.345083236694336,
+ "step": 1725
+ },
+ {
+ "epoch": 23.978165938864628,
+ "grad_norm": 0.018797501921653748,
+ "learning_rate": 0.0006,
+ "loss": 4.290443420410156,
+ "step": 1726
+ },
+ {
+ "epoch": 23.992139737991266,
+ "grad_norm": 0.018090683966875076,
+ "learning_rate": 0.0006,
+ "loss": 4.377580642700195,
+ "step": 1727
+ },
+ {
+ "epoch": 24.0,
+ "grad_norm": 0.017305459827184677,
+ "learning_rate": 0.0006,
+ "loss": 4.364426612854004,
+ "step": 1728
+ },
+ {
+ "epoch": 24.0,
+ "eval_loss": 4.647762298583984,
+ "eval_runtime": 56.5402,
+ "eval_samples_per_second": 43.191,
+ "eval_steps_per_second": 1.362,
+ "step": 1728
+ },
+ {
+ "epoch": 24.01397379912664,
+ "grad_norm": 0.016493607312440872,
+ "learning_rate": 0.0006,
+ "loss": 4.252255916595459,
+ "step": 1729
+ },
+ {
+ "epoch": 24.027947598253274,
+ "grad_norm": 0.016529450193047523,
+ "learning_rate": 0.0006,
+ "loss": 4.352746963500977,
+ "step": 1730
+ },
+ {
+ "epoch": 24.041921397379912,
+ "grad_norm": 0.016940191388130188,
+ "learning_rate": 0.0006,
+ "loss": 4.301870822906494,
+ "step": 1731
+ },
+ {
+ "epoch": 24.05589519650655,
+ "grad_norm": 0.01752558909356594,
+ "learning_rate": 0.0006,
+ "loss": 4.2762451171875,
+ "step": 1732
+ },
+ {
+ "epoch": 24.069868995633186,
+ "grad_norm": 0.017112884670495987,
+ "learning_rate": 0.0006,
+ "loss": 4.2659149169921875,
+ "step": 1733
+ },
+ {
+ "epoch": 24.083842794759825,
+ "grad_norm": 0.01900586113333702,
+ "learning_rate": 0.0006,
+ "loss": 4.306058883666992,
+ "step": 1734
+ },
+ {
+ "epoch": 24.097816593886463,
+ "grad_norm": 0.019447680562734604,
+ "learning_rate": 0.0006,
+ "loss": 4.450128555297852,
+ "step": 1735
+ },
+ {
+ "epoch": 24.111790393013102,
+ "grad_norm": 0.01948186755180359,
+ "learning_rate": 0.0006,
+ "loss": 4.202869415283203,
+ "step": 1736
+ },
+ {
+ "epoch": 24.125764192139737,
+ "grad_norm": 0.01716383546590805,
+ "learning_rate": 0.0006,
+ "loss": 4.318876266479492,
+ "step": 1737
+ },
+ {
+ "epoch": 24.139737991266376,
+ "grad_norm": 0.017049988731741905,
+ "learning_rate": 0.0006,
+ "loss": 4.237238883972168,
+ "step": 1738
+ },
+ {
+ "epoch": 24.153711790393015,
+ "grad_norm": 0.015169923193752766,
+ "learning_rate": 0.0006,
+ "loss": 4.289426803588867,
+ "step": 1739
+ },
+ {
+ "epoch": 24.16768558951965,
+ "grad_norm": 0.014760667458176613,
+ "learning_rate": 0.0006,
+ "loss": 4.356924057006836,
+ "step": 1740
+ },
+ {
+ "epoch": 24.18165938864629,
+ "grad_norm": 0.015127750113606453,
+ "learning_rate": 0.0006,
+ "loss": 4.142916679382324,
+ "step": 1741
+ },
+ {
+ "epoch": 24.195633187772927,
+ "grad_norm": 0.01549555640667677,
+ "learning_rate": 0.0006,
+ "loss": 4.345563888549805,
+ "step": 1742
+ },
+ {
+ "epoch": 24.209606986899562,
+ "grad_norm": 0.016393091529607773,
+ "learning_rate": 0.0006,
+ "loss": 4.417495250701904,
+ "step": 1743
+ },
+ {
+ "epoch": 24.2235807860262,
+ "grad_norm": 0.01785266026854515,
+ "learning_rate": 0.0006,
+ "loss": 4.3553547859191895,
+ "step": 1744
+ },
+ {
+ "epoch": 24.23755458515284,
+ "grad_norm": 0.017574617639183998,
+ "learning_rate": 0.0006,
+ "loss": 4.241024494171143,
+ "step": 1745
+ },
+ {
+ "epoch": 24.251528384279474,
+ "grad_norm": 0.019060306251049042,
+ "learning_rate": 0.0006,
+ "loss": 4.3685150146484375,
+ "step": 1746
+ },
+ {
+ "epoch": 24.265502183406113,
+ "grad_norm": 0.0214633010327816,
+ "learning_rate": 0.0006,
+ "loss": 4.323866844177246,
+ "step": 1747
+ },
+ {
+ "epoch": 24.27947598253275,
+ "grad_norm": 0.021055003628134727,
+ "learning_rate": 0.0006,
+ "loss": 4.278757095336914,
+ "step": 1748
+ },
+ {
+ "epoch": 24.29344978165939,
+ "grad_norm": 0.018618572503328323,
+ "learning_rate": 0.0006,
+ "loss": 4.28849983215332,
+ "step": 1749
+ },
+ {
+ "epoch": 24.307423580786025,
+ "grad_norm": 0.018666435033082962,
+ "learning_rate": 0.0006,
+ "loss": 4.368590354919434,
+ "step": 1750
+ },
+ {
+ "epoch": 24.321397379912664,
+ "grad_norm": 0.020626146346330643,
+ "learning_rate": 0.0006,
+ "loss": 4.2951836585998535,
+ "step": 1751
+ },
+ {
+ "epoch": 24.335371179039303,
+ "grad_norm": 0.01929759792983532,
+ "learning_rate": 0.0006,
+ "loss": 4.3664703369140625,
+ "step": 1752
+ },
+ {
+ "epoch": 24.349344978165938,
+ "grad_norm": 0.020956002175807953,
+ "learning_rate": 0.0006,
+ "loss": 4.345308303833008,
+ "step": 1753
+ },
+ {
+ "epoch": 24.363318777292577,
+ "grad_norm": 0.019227957352995872,
+ "learning_rate": 0.0006,
+ "loss": 4.349849224090576,
+ "step": 1754
+ },
+ {
+ "epoch": 24.377292576419215,
+ "grad_norm": 0.015970097854733467,
+ "learning_rate": 0.0006,
+ "loss": 4.223012447357178,
+ "step": 1755
+ },
+ {
+ "epoch": 24.39126637554585,
+ "grad_norm": 0.015951646491885185,
+ "learning_rate": 0.0006,
+ "loss": 4.236420631408691,
+ "step": 1756
+ },
+ {
+ "epoch": 24.40524017467249,
+ "grad_norm": 0.016032515093684196,
+ "learning_rate": 0.0006,
+ "loss": 4.226899147033691,
+ "step": 1757
+ },
+ {
+ "epoch": 24.419213973799128,
+ "grad_norm": 0.01573132909834385,
+ "learning_rate": 0.0006,
+ "loss": 4.315916061401367,
+ "step": 1758
+ },
+ {
+ "epoch": 24.433187772925763,
+ "grad_norm": 0.015367005951702595,
+ "learning_rate": 0.0006,
+ "loss": 4.358939170837402,
+ "step": 1759
+ },
+ {
+ "epoch": 24.4471615720524,
+ "grad_norm": 0.01669907011091709,
+ "learning_rate": 0.0006,
+ "loss": 4.232147216796875,
+ "step": 1760
+ },
+ {
+ "epoch": 24.46113537117904,
+ "grad_norm": 0.018708152696490288,
+ "learning_rate": 0.0006,
+ "loss": 4.329852104187012,
+ "step": 1761
+ },
+ {
+ "epoch": 24.475109170305675,
+ "grad_norm": 0.019035987555980682,
+ "learning_rate": 0.0006,
+ "loss": 4.260340213775635,
+ "step": 1762
+ },
+ {
+ "epoch": 24.489082969432314,
+ "grad_norm": 0.01832144521176815,
+ "learning_rate": 0.0006,
+ "loss": 4.359543800354004,
+ "step": 1763
+ },
+ {
+ "epoch": 24.503056768558952,
+ "grad_norm": 0.015313294716179371,
+ "learning_rate": 0.0006,
+ "loss": 4.279088020324707,
+ "step": 1764
+ },
+ {
+ "epoch": 24.51703056768559,
+ "grad_norm": 0.015803923830389977,
+ "learning_rate": 0.0006,
+ "loss": 4.230682849884033,
+ "step": 1765
+ },
+ {
+ "epoch": 24.531004366812226,
+ "grad_norm": 0.016748374328017235,
+ "learning_rate": 0.0006,
+ "loss": 4.349661827087402,
+ "step": 1766
+ },
+ {
+ "epoch": 24.544978165938865,
+ "grad_norm": 0.015732428058981895,
+ "learning_rate": 0.0006,
+ "loss": 4.372441291809082,
+ "step": 1767
+ },
+ {
+ "epoch": 24.558951965065503,
+ "grad_norm": 0.016564829275012016,
+ "learning_rate": 0.0006,
+ "loss": 4.323090553283691,
+ "step": 1768
+ },
+ {
+ "epoch": 24.57292576419214,
+ "grad_norm": 0.01603010483086109,
+ "learning_rate": 0.0006,
+ "loss": 4.347898006439209,
+ "step": 1769
+ },
+ {
+ "epoch": 24.586899563318777,
+ "grad_norm": 0.01674809865653515,
+ "learning_rate": 0.0006,
+ "loss": 4.269165515899658,
+ "step": 1770
+ },
+ {
+ "epoch": 24.600873362445416,
+ "grad_norm": 0.017333930358290672,
+ "learning_rate": 0.0006,
+ "loss": 4.353298187255859,
+ "step": 1771
+ },
+ {
+ "epoch": 24.61484716157205,
+ "grad_norm": 0.017519580200314522,
+ "learning_rate": 0.0006,
+ "loss": 4.265199661254883,
+ "step": 1772
+ },
+ {
+ "epoch": 24.62882096069869,
+ "grad_norm": 0.017675306648015976,
+ "learning_rate": 0.0006,
+ "loss": 4.322596549987793,
+ "step": 1773
+ },
+ {
+ "epoch": 24.64279475982533,
+ "grad_norm": 0.017166534438729286,
+ "learning_rate": 0.0006,
+ "loss": 4.413900375366211,
+ "step": 1774
+ },
+ {
+ "epoch": 24.656768558951963,
+ "grad_norm": 0.016362346708774567,
+ "learning_rate": 0.0006,
+ "loss": 4.307236671447754,
+ "step": 1775
+ },
+ {
+ "epoch": 24.670742358078602,
+ "grad_norm": 0.016086872667074203,
+ "learning_rate": 0.0006,
+ "loss": 4.235316276550293,
+ "step": 1776
+ },
+ {
+ "epoch": 24.68471615720524,
+ "grad_norm": 0.01689537614583969,
+ "learning_rate": 0.0006,
+ "loss": 4.390591621398926,
+ "step": 1777
+ },
+ {
+ "epoch": 24.69868995633188,
+ "grad_norm": 0.016836725175380707,
+ "learning_rate": 0.0006,
+ "loss": 4.2990031242370605,
+ "step": 1778
+ },
+ {
+ "epoch": 24.712663755458514,
+ "grad_norm": 0.015627335757017136,
+ "learning_rate": 0.0006,
+ "loss": 4.166745185852051,
+ "step": 1779
+ },
+ {
+ "epoch": 24.726637554585153,
+ "grad_norm": 0.015007016249001026,
+ "learning_rate": 0.0006,
+ "loss": 4.295273303985596,
+ "step": 1780
+ },
+ {
+ "epoch": 24.74061135371179,
+ "grad_norm": 0.016084209084510803,
+ "learning_rate": 0.0006,
+ "loss": 4.342424392700195,
+ "step": 1781
+ },
+ {
+ "epoch": 24.754585152838427,
+ "grad_norm": 0.01595073565840721,
+ "learning_rate": 0.0006,
+ "loss": 4.3060102462768555,
+ "step": 1782
+ },
+ {
+ "epoch": 24.768558951965066,
+ "grad_norm": 0.016113461926579475,
+ "learning_rate": 0.0006,
+ "loss": 4.267908096313477,
+ "step": 1783
+ },
+ {
+ "epoch": 24.782532751091704,
+ "grad_norm": 0.01564556173980236,
+ "learning_rate": 0.0006,
+ "loss": 4.324642658233643,
+ "step": 1784
+ },
+ {
+ "epoch": 24.79650655021834,
+ "grad_norm": 0.01575268618762493,
+ "learning_rate": 0.0006,
+ "loss": 4.328181743621826,
+ "step": 1785
+ },
+ {
+ "epoch": 24.810480349344978,
+ "grad_norm": 0.016830824315547943,
+ "learning_rate": 0.0006,
+ "loss": 4.247906684875488,
+ "step": 1786
+ },
+ {
+ "epoch": 24.824454148471617,
+ "grad_norm": 0.017989547923207283,
+ "learning_rate": 0.0006,
+ "loss": 4.228281021118164,
+ "step": 1787
+ },
+ {
+ "epoch": 24.83842794759825,
+ "grad_norm": 0.016783086583018303,
+ "learning_rate": 0.0006,
+ "loss": 4.2880964279174805,
+ "step": 1788
+ },
+ {
+ "epoch": 24.85240174672489,
+ "grad_norm": 0.016467561945319176,
+ "learning_rate": 0.0006,
+ "loss": 4.36118745803833,
+ "step": 1789
+ },
+ {
+ "epoch": 24.86637554585153,
+ "grad_norm": 0.01780523732304573,
+ "learning_rate": 0.0006,
+ "loss": 4.172321319580078,
+ "step": 1790
+ },
+ {
+ "epoch": 24.880349344978168,
+ "grad_norm": 0.017102031037211418,
+ "learning_rate": 0.0006,
+ "loss": 4.352826118469238,
+ "step": 1791
+ },
+ {
+ "epoch": 24.894323144104803,
+ "grad_norm": 0.016409218311309814,
+ "learning_rate": 0.0006,
+ "loss": 4.212162017822266,
+ "step": 1792
+ },
+ {
+ "epoch": 24.90829694323144,
+ "grad_norm": 0.018660584464669228,
+ "learning_rate": 0.0006,
+ "loss": 4.346684455871582,
+ "step": 1793
+ },
+ {
+ "epoch": 24.92227074235808,
+ "grad_norm": 0.017389440909028053,
+ "learning_rate": 0.0006,
+ "loss": 4.197913646697998,
+ "step": 1794
+ },
+ {
+ "epoch": 24.936244541484715,
+ "grad_norm": 0.01596485823392868,
+ "learning_rate": 0.0006,
+ "loss": 4.370617866516113,
+ "step": 1795
+ },
+ {
+ "epoch": 24.950218340611354,
+ "grad_norm": 0.01709776371717453,
+ "learning_rate": 0.0006,
+ "loss": 4.313847064971924,
+ "step": 1796
+ },
+ {
+ "epoch": 24.964192139737992,
+ "grad_norm": 0.016571134328842163,
+ "learning_rate": 0.0006,
+ "loss": 4.1895856857299805,
+ "step": 1797
+ },
+ {
+ "epoch": 24.978165938864628,
+ "grad_norm": 0.016851048916578293,
+ "learning_rate": 0.0006,
+ "loss": 4.350772857666016,
+ "step": 1798
+ },
+ {
+ "epoch": 24.992139737991266,
+ "grad_norm": 0.017007440328598022,
+ "learning_rate": 0.0006,
+ "loss": 4.399008750915527,
+ "step": 1799
+ },
+ {
+ "epoch": 25.0,
+ "grad_norm": 0.01791212148964405,
+ "learning_rate": 0.0006,
+ "loss": 4.306121826171875,
+ "step": 1800
+ },
+ {
+ "epoch": 25.0,
+ "eval_loss": 4.636417865753174,
+ "eval_runtime": 56.8296,
+ "eval_samples_per_second": 42.971,
+ "eval_steps_per_second": 1.355,
+ "step": 1800
+ },
+ {
+ "epoch": 25.01397379912664,
+ "grad_norm": 0.01729617640376091,
+ "learning_rate": 0.0006,
+ "loss": 4.257018089294434,
+ "step": 1801
+ },
+ {
+ "epoch": 25.027947598253274,
+ "grad_norm": 0.017466643825173378,
+ "learning_rate": 0.0006,
+ "loss": 4.307605266571045,
+ "step": 1802
+ },
+ {
+ "epoch": 25.041921397379912,
+ "grad_norm": 0.018614279106259346,
+ "learning_rate": 0.0006,
+ "loss": 4.360141277313232,
+ "step": 1803
+ },
+ {
+ "epoch": 25.05589519650655,
+ "grad_norm": 0.02010771445930004,
+ "learning_rate": 0.0006,
+ "loss": 4.196428298950195,
+ "step": 1804
+ },
+ {
+ "epoch": 25.069868995633186,
+ "grad_norm": 0.02090228535234928,
+ "learning_rate": 0.0006,
+ "loss": 4.288947105407715,
+ "step": 1805
+ },
+ {
+ "epoch": 25.083842794759825,
+ "grad_norm": 0.02050255797803402,
+ "learning_rate": 0.0006,
+ "loss": 4.259582042694092,
+ "step": 1806
+ },
+ {
+ "epoch": 25.097816593886463,
+ "grad_norm": 0.021844089031219482,
+ "learning_rate": 0.0006,
+ "loss": 4.405138969421387,
+ "step": 1807
+ },
+ {
+ "epoch": 25.111790393013102,
+ "grad_norm": 0.022287018597126007,
+ "learning_rate": 0.0006,
+ "loss": 4.333793640136719,
+ "step": 1808
+ },
+ {
+ "epoch": 25.125764192139737,
+ "grad_norm": 0.02288329415023327,
+ "learning_rate": 0.0006,
+ "loss": 4.3910980224609375,
+ "step": 1809
+ },
+ {
+ "epoch": 25.139737991266376,
+ "grad_norm": 0.02373199723660946,
+ "learning_rate": 0.0006,
+ "loss": 4.32921028137207,
+ "step": 1810
+ },
+ {
+ "epoch": 25.153711790393015,
+ "grad_norm": 0.024960234761238098,
+ "learning_rate": 0.0006,
+ "loss": 4.295609474182129,
+ "step": 1811
+ },
+ {
+ "epoch": 25.16768558951965,
+ "grad_norm": 0.026743892580270767,
+ "learning_rate": 0.0006,
+ "loss": 4.315308094024658,
+ "step": 1812
+ },
+ {
+ "epoch": 25.18165938864629,
+ "grad_norm": 0.022138185799121857,
+ "learning_rate": 0.0006,
+ "loss": 4.360208511352539,
+ "step": 1813
+ },
+ {
+ "epoch": 25.195633187772927,
+ "grad_norm": 0.022120574489235878,
+ "learning_rate": 0.0006,
+ "loss": 4.1973772048950195,
+ "step": 1814
+ },
+ {
+ "epoch": 25.209606986899562,
+ "grad_norm": 0.02230502851307392,
+ "learning_rate": 0.0006,
+ "loss": 4.285558700561523,
+ "step": 1815
+ },
+ {
+ "epoch": 25.2235807860262,
+ "grad_norm": 0.022301986813545227,
+ "learning_rate": 0.0006,
+ "loss": 4.238343238830566,
+ "step": 1816
+ },
+ {
+ "epoch": 25.23755458515284,
+ "grad_norm": 0.020050447434186935,
+ "learning_rate": 0.0006,
+ "loss": 4.298235893249512,
+ "step": 1817
+ },
+ {
+ "epoch": 25.251528384279474,
+ "grad_norm": 0.01987724006175995,
+ "learning_rate": 0.0006,
+ "loss": 4.301384449005127,
+ "step": 1818
+ },
+ {
+ "epoch": 25.265502183406113,
+ "grad_norm": 0.020440855994820595,
+ "learning_rate": 0.0006,
+ "loss": 4.269218444824219,
+ "step": 1819
+ },
+ {
+ "epoch": 25.27947598253275,
+ "grad_norm": 0.01891893707215786,
+ "learning_rate": 0.0006,
+ "loss": 4.2401957511901855,
+ "step": 1820
+ },
+ {
+ "epoch": 25.29344978165939,
+ "grad_norm": 0.01903688907623291,
+ "learning_rate": 0.0006,
+ "loss": 4.139670372009277,
+ "step": 1821
+ },
+ {
+ "epoch": 25.307423580786025,
+ "grad_norm": 0.018856938928365707,
+ "learning_rate": 0.0006,
+ "loss": 4.233214378356934,
+ "step": 1822
+ },
+ {
+ "epoch": 25.321397379912664,
+ "grad_norm": 0.017791323363780975,
+ "learning_rate": 0.0006,
+ "loss": 4.2342023849487305,
+ "step": 1823
+ },
+ {
+ "epoch": 25.335371179039303,
+ "grad_norm": 0.017304565757513046,
+ "learning_rate": 0.0006,
+ "loss": 4.355198860168457,
+ "step": 1824
+ },
+ {
+ "epoch": 25.349344978165938,
+ "grad_norm": 0.01576206088066101,
+ "learning_rate": 0.0006,
+ "loss": 4.114147186279297,
+ "step": 1825
+ },
+ {
+ "epoch": 25.363318777292577,
+ "grad_norm": 0.015105154365301132,
+ "learning_rate": 0.0006,
+ "loss": 4.313795566558838,
+ "step": 1826
+ },
+ {
+ "epoch": 25.377292576419215,
+ "grad_norm": 0.01563587784767151,
+ "learning_rate": 0.0006,
+ "loss": 4.308282852172852,
+ "step": 1827
+ },
+ {
+ "epoch": 25.39126637554585,
+ "grad_norm": 0.015980906784534454,
+ "learning_rate": 0.0006,
+ "loss": 4.345149040222168,
+ "step": 1828
+ },
+ {
+ "epoch": 25.40524017467249,
+ "grad_norm": 0.01669284701347351,
+ "learning_rate": 0.0006,
+ "loss": 4.237611293792725,
+ "step": 1829
+ },
+ {
+ "epoch": 25.419213973799128,
+ "grad_norm": 0.016202064231038094,
+ "learning_rate": 0.0006,
+ "loss": 4.3072309494018555,
+ "step": 1830
+ },
+ {
+ "epoch": 25.433187772925763,
+ "grad_norm": 0.01478662807494402,
+ "learning_rate": 0.0006,
+ "loss": 4.276215076446533,
+ "step": 1831
+ },
+ {
+ "epoch": 25.4471615720524,
+ "grad_norm": 0.015168731100857258,
+ "learning_rate": 0.0006,
+ "loss": 4.37509822845459,
+ "step": 1832
+ },
+ {
+ "epoch": 25.46113537117904,
+ "grad_norm": 0.01648980937898159,
+ "learning_rate": 0.0006,
+ "loss": 4.349167346954346,
+ "step": 1833
+ },
+ {
+ "epoch": 25.475109170305675,
+ "grad_norm": 0.01708398200571537,
+ "learning_rate": 0.0006,
+ "loss": 4.346580505371094,
+ "step": 1834
+ },
+ {
+ "epoch": 25.489082969432314,
+ "grad_norm": 0.017370784655213356,
+ "learning_rate": 0.0006,
+ "loss": 4.299466133117676,
+ "step": 1835
+ },
+ {
+ "epoch": 25.503056768558952,
+ "grad_norm": 0.016248228028416634,
+ "learning_rate": 0.0006,
+ "loss": 4.184610366821289,
+ "step": 1836
+ },
+ {
+ "epoch": 25.51703056768559,
+ "grad_norm": 0.017106913030147552,
+ "learning_rate": 0.0006,
+ "loss": 4.366856098175049,
+ "step": 1837
+ },
+ {
+ "epoch": 25.531004366812226,
+ "grad_norm": 0.016071965917944908,
+ "learning_rate": 0.0006,
+ "loss": 4.297691345214844,
+ "step": 1838
+ },
+ {
+ "epoch": 25.544978165938865,
+ "grad_norm": 0.016451986506581306,
+ "learning_rate": 0.0006,
+ "loss": 4.229983329772949,
+ "step": 1839
+ },
+ {
+ "epoch": 25.558951965065503,
+ "grad_norm": 0.016195081174373627,
+ "learning_rate": 0.0006,
+ "loss": 4.259893417358398,
+ "step": 1840
+ },
+ {
+ "epoch": 25.57292576419214,
+ "grad_norm": 0.0165175162255764,
+ "learning_rate": 0.0006,
+ "loss": 4.3348236083984375,
+ "step": 1841
+ },
+ {
+ "epoch": 25.586899563318777,
+ "grad_norm": 0.01651517115533352,
+ "learning_rate": 0.0006,
+ "loss": 4.319541931152344,
+ "step": 1842
+ },
+ {
+ "epoch": 25.600873362445416,
+ "grad_norm": 0.016143690794706345,
+ "learning_rate": 0.0006,
+ "loss": 4.345652103424072,
+ "step": 1843
+ },
+ {
+ "epoch": 25.61484716157205,
+ "grad_norm": 0.015140696428716183,
+ "learning_rate": 0.0006,
+ "loss": 4.212047576904297,
+ "step": 1844
+ },
+ {
+ "epoch": 25.62882096069869,
+ "grad_norm": 0.014637443237006664,
+ "learning_rate": 0.0006,
+ "loss": 4.187453269958496,
+ "step": 1845
+ },
+ {
+ "epoch": 25.64279475982533,
+ "grad_norm": 0.014585614204406738,
+ "learning_rate": 0.0006,
+ "loss": 4.253936767578125,
+ "step": 1846
+ },
+ {
+ "epoch": 25.656768558951963,
+ "grad_norm": 0.015158289112150669,
+ "learning_rate": 0.0006,
+ "loss": 4.34612512588501,
+ "step": 1847
+ },
+ {
+ "epoch": 25.670742358078602,
+ "grad_norm": 0.015202849172055721,
+ "learning_rate": 0.0006,
+ "loss": 4.270634651184082,
+ "step": 1848
+ },
+ {
+ "epoch": 25.68471615720524,
+ "grad_norm": 0.015363575890660286,
+ "learning_rate": 0.0006,
+ "loss": 4.271080493927002,
+ "step": 1849
+ },
+ {
+ "epoch": 25.69868995633188,
+ "grad_norm": 0.014360226690769196,
+ "learning_rate": 0.0006,
+ "loss": 4.375033855438232,
+ "step": 1850
+ },
+ {
+ "epoch": 25.712663755458514,
+ "grad_norm": 0.014263181947171688,
+ "learning_rate": 0.0006,
+ "loss": 4.424138069152832,
+ "step": 1851
+ },
+ {
+ "epoch": 25.726637554585153,
+ "grad_norm": 0.014398688450455666,
+ "learning_rate": 0.0006,
+ "loss": 4.410154819488525,
+ "step": 1852
+ },
+ {
+ "epoch": 25.74061135371179,
+ "grad_norm": 0.016131488606333733,
+ "learning_rate": 0.0006,
+ "loss": 4.324014663696289,
+ "step": 1853
+ },
+ {
+ "epoch": 25.754585152838427,
+ "grad_norm": 0.017568735405802727,
+ "learning_rate": 0.0006,
+ "loss": 4.367861747741699,
+ "step": 1854
+ },
+ {
+ "epoch": 25.768558951965066,
+ "grad_norm": 0.017907511442899704,
+ "learning_rate": 0.0006,
+ "loss": 4.352965354919434,
+ "step": 1855
+ },
+ {
+ "epoch": 25.782532751091704,
+ "grad_norm": 0.016918016597628593,
+ "learning_rate": 0.0006,
+ "loss": 4.252397537231445,
+ "step": 1856
+ },
+ {
+ "epoch": 25.79650655021834,
+ "grad_norm": 0.017274610698223114,
+ "learning_rate": 0.0006,
+ "loss": 4.342309951782227,
+ "step": 1857
+ },
+ {
+ "epoch": 25.810480349344978,
+ "grad_norm": 0.017879825085401535,
+ "learning_rate": 0.0006,
+ "loss": 4.330987453460693,
+ "step": 1858
+ },
+ {
+ "epoch": 25.824454148471617,
+ "grad_norm": 0.01736099272966385,
+ "learning_rate": 0.0006,
+ "loss": 4.307255744934082,
+ "step": 1859
+ },
+ {
+ "epoch": 25.83842794759825,
+ "grad_norm": 0.016575824469327927,
+ "learning_rate": 0.0006,
+ "loss": 4.262016773223877,
+ "step": 1860
+ },
+ {
+ "epoch": 25.85240174672489,
+ "grad_norm": 0.015181140042841434,
+ "learning_rate": 0.0006,
+ "loss": 4.291863441467285,
+ "step": 1861
+ },
+ {
+ "epoch": 25.86637554585153,
+ "grad_norm": 0.015965687111020088,
+ "learning_rate": 0.0006,
+ "loss": 4.424836158752441,
+ "step": 1862
+ },
+ {
+ "epoch": 25.880349344978168,
+ "grad_norm": 0.01687219925224781,
+ "learning_rate": 0.0006,
+ "loss": 4.343531131744385,
+ "step": 1863
+ },
+ {
+ "epoch": 25.894323144104803,
+ "grad_norm": 0.015115504153072834,
+ "learning_rate": 0.0006,
+ "loss": 4.357808589935303,
+ "step": 1864
+ },
+ {
+ "epoch": 25.90829694323144,
+ "grad_norm": 0.014150998555123806,
+ "learning_rate": 0.0006,
+ "loss": 4.324550151824951,
+ "step": 1865
+ },
+ {
+ "epoch": 25.92227074235808,
+ "grad_norm": 0.014989510178565979,
+ "learning_rate": 0.0006,
+ "loss": 4.380428314208984,
+ "step": 1866
+ },
+ {
+ "epoch": 25.936244541484715,
+ "grad_norm": 0.01527960691601038,
+ "learning_rate": 0.0006,
+ "loss": 4.399374961853027,
+ "step": 1867
+ },
+ {
+ "epoch": 25.950218340611354,
+ "grad_norm": 0.015434633940458298,
+ "learning_rate": 0.0006,
+ "loss": 4.260984420776367,
+ "step": 1868
+ },
+ {
+ "epoch": 25.964192139737992,
+ "grad_norm": 0.015370228327810764,
+ "learning_rate": 0.0006,
+ "loss": 4.3340253829956055,
+ "step": 1869
+ },
+ {
+ "epoch": 25.978165938864628,
+ "grad_norm": 0.015348542481660843,
+ "learning_rate": 0.0006,
+ "loss": 4.369132995605469,
+ "step": 1870
+ },
+ {
+ "epoch": 25.992139737991266,
+ "grad_norm": 0.015354936942458153,
+ "learning_rate": 0.0006,
+ "loss": 4.381937026977539,
+ "step": 1871
+ },
+ {
+ "epoch": 26.0,
+ "grad_norm": 0.01796272210776806,
+ "learning_rate": 0.0006,
+ "loss": 4.286768436431885,
+ "step": 1872
+ },
+ {
+ "epoch": 26.0,
+ "eval_loss": 4.576984405517578,
+ "eval_runtime": 57.2091,
+ "eval_samples_per_second": 42.685,
+ "eval_steps_per_second": 1.346,
+ "step": 1872
+ },
+ {
+ "epoch": 26.01397379912664,
+ "grad_norm": 0.01748845726251602,
+ "learning_rate": 0.0006,
+ "loss": 4.337504863739014,
+ "step": 1873
+ },
+ {
+ "epoch": 26.027947598253274,
+ "grad_norm": 0.019244296476244926,
+ "learning_rate": 0.0006,
+ "loss": 4.264280796051025,
+ "step": 1874
+ },
+ {
+ "epoch": 26.041921397379912,
+ "grad_norm": 0.0215692650526762,
+ "learning_rate": 0.0006,
+ "loss": 4.350830078125,
+ "step": 1875
+ },
+ {
+ "epoch": 26.05589519650655,
+ "grad_norm": 0.022189252078533173,
+ "learning_rate": 0.0006,
+ "loss": 4.299098968505859,
+ "step": 1876
+ },
+ {
+ "epoch": 26.069868995633186,
+ "grad_norm": 0.022384043782949448,
+ "learning_rate": 0.0006,
+ "loss": 4.1185173988342285,
+ "step": 1877
+ },
+ {
+ "epoch": 26.083842794759825,
+ "grad_norm": 0.02076014317572117,
+ "learning_rate": 0.0006,
+ "loss": 4.297612190246582,
+ "step": 1878
+ },
+ {
+ "epoch": 26.097816593886463,
+ "grad_norm": 0.020843051373958588,
+ "learning_rate": 0.0006,
+ "loss": 4.294098854064941,
+ "step": 1879
+ },
+ {
+ "epoch": 26.111790393013102,
+ "grad_norm": 0.020654456689953804,
+ "learning_rate": 0.0006,
+ "loss": 4.267856597900391,
+ "step": 1880
+ },
+ {
+ "epoch": 26.125764192139737,
+ "grad_norm": 0.01860472559928894,
+ "learning_rate": 0.0006,
+ "loss": 4.284232139587402,
+ "step": 1881
+ },
+ {
+ "epoch": 26.139737991266376,
+ "grad_norm": 0.018719421699643135,
+ "learning_rate": 0.0006,
+ "loss": 4.277214527130127,
+ "step": 1882
+ },
+ {
+ "epoch": 26.153711790393015,
+ "grad_norm": 0.020903829485177994,
+ "learning_rate": 0.0006,
+ "loss": 4.254947662353516,
+ "step": 1883
+ },
+ {
+ "epoch": 26.16768558951965,
+ "grad_norm": 0.02020149491727352,
+ "learning_rate": 0.0006,
+ "loss": 4.2526936531066895,
+ "step": 1884
+ },
+ {
+ "epoch": 26.18165938864629,
+ "grad_norm": 0.0180392786860466,
+ "learning_rate": 0.0006,
+ "loss": 4.309922695159912,
+ "step": 1885
+ },
+ {
+ "epoch": 26.195633187772927,
+ "grad_norm": 0.018163125962018967,
+ "learning_rate": 0.0006,
+ "loss": 4.13860559463501,
+ "step": 1886
+ },
+ {
+ "epoch": 26.209606986899562,
+ "grad_norm": 0.01853291131556034,
+ "learning_rate": 0.0006,
+ "loss": 4.190926551818848,
+ "step": 1887
+ },
+ {
+ "epoch": 26.2235807860262,
+ "grad_norm": 0.01757677271962166,
+ "learning_rate": 0.0006,
+ "loss": 4.163999557495117,
+ "step": 1888
+ },
+ {
+ "epoch": 26.23755458515284,
+ "grad_norm": 0.01767992228269577,
+ "learning_rate": 0.0006,
+ "loss": 4.184430122375488,
+ "step": 1889
+ },
+ {
+ "epoch": 26.251528384279474,
+ "grad_norm": 0.018139973282814026,
+ "learning_rate": 0.0006,
+ "loss": 4.300766944885254,
+ "step": 1890
+ },
+ {
+ "epoch": 26.265502183406113,
+ "grad_norm": 0.020184461027383804,
+ "learning_rate": 0.0006,
+ "loss": 4.232537269592285,
+ "step": 1891
+ },
+ {
+ "epoch": 26.27947598253275,
+ "grad_norm": 0.02028856985270977,
+ "learning_rate": 0.0006,
+ "loss": 4.324883460998535,
+ "step": 1892
+ },
+ {
+ "epoch": 26.29344978165939,
+ "grad_norm": 0.017120616510510445,
+ "learning_rate": 0.0006,
+ "loss": 4.23493766784668,
+ "step": 1893
+ },
+ {
+ "epoch": 26.307423580786025,
+ "grad_norm": 0.016202978789806366,
+ "learning_rate": 0.0006,
+ "loss": 4.197602272033691,
+ "step": 1894
+ },
+ {
+ "epoch": 26.321397379912664,
+ "grad_norm": 0.017231421545147896,
+ "learning_rate": 0.0006,
+ "loss": 4.302001953125,
+ "step": 1895
+ },
+ {
+ "epoch": 26.335371179039303,
+ "grad_norm": 0.01725655607879162,
+ "learning_rate": 0.0006,
+ "loss": 4.356800079345703,
+ "step": 1896
+ },
+ {
+ "epoch": 26.349344978165938,
+ "grad_norm": 0.016657443717122078,
+ "learning_rate": 0.0006,
+ "loss": 4.266307830810547,
+ "step": 1897
+ },
+ {
+ "epoch": 26.363318777292577,
+ "grad_norm": 0.01616556942462921,
+ "learning_rate": 0.0006,
+ "loss": 4.307868003845215,
+ "step": 1898
+ },
+ {
+ "epoch": 26.377292576419215,
+ "grad_norm": 0.016621306538581848,
+ "learning_rate": 0.0006,
+ "loss": 4.172905921936035,
+ "step": 1899
+ },
+ {
+ "epoch": 26.39126637554585,
+ "grad_norm": 0.016243206337094307,
+ "learning_rate": 0.0006,
+ "loss": 4.162896156311035,
+ "step": 1900
+ },
+ {
+ "epoch": 26.40524017467249,
+ "grad_norm": 0.016663808375597,
+ "learning_rate": 0.0006,
+ "loss": 4.258232116699219,
+ "step": 1901
+ },
+ {
+ "epoch": 26.419213973799128,
+ "grad_norm": 0.01677478663623333,
+ "learning_rate": 0.0006,
+ "loss": 4.336090087890625,
+ "step": 1902
+ },
+ {
+ "epoch": 26.433187772925763,
+ "grad_norm": 0.01641303487122059,
+ "learning_rate": 0.0006,
+ "loss": 4.191476821899414,
+ "step": 1903
+ },
+ {
+ "epoch": 26.4471615720524,
+ "grad_norm": 0.016700323671102524,
+ "learning_rate": 0.0006,
+ "loss": 4.299427509307861,
+ "step": 1904
+ },
+ {
+ "epoch": 26.46113537117904,
+ "grad_norm": 0.017595035955309868,
+ "learning_rate": 0.0006,
+ "loss": 4.312828063964844,
+ "step": 1905
+ },
+ {
+ "epoch": 26.475109170305675,
+ "grad_norm": 0.020300284028053284,
+ "learning_rate": 0.0006,
+ "loss": 4.28331184387207,
+ "step": 1906
+ },
+ {
+ "epoch": 26.489082969432314,
+ "grad_norm": 0.02198321744799614,
+ "learning_rate": 0.0006,
+ "loss": 4.3280839920043945,
+ "step": 1907
+ },
+ {
+ "epoch": 26.503056768558952,
+ "grad_norm": 0.02003200724720955,
+ "learning_rate": 0.0006,
+ "loss": 4.325623989105225,
+ "step": 1908
+ },
+ {
+ "epoch": 26.51703056768559,
+ "grad_norm": 0.01870287023484707,
+ "learning_rate": 0.0006,
+ "loss": 4.251203536987305,
+ "step": 1909
+ },
+ {
+ "epoch": 26.531004366812226,
+ "grad_norm": 0.01835448481142521,
+ "learning_rate": 0.0006,
+ "loss": 4.335144996643066,
+ "step": 1910
+ },
+ {
+ "epoch": 26.544978165938865,
+ "grad_norm": 0.02046327292919159,
+ "learning_rate": 0.0006,
+ "loss": 4.238023281097412,
+ "step": 1911
+ },
+ {
+ "epoch": 26.558951965065503,
+ "grad_norm": 0.019771311432123184,
+ "learning_rate": 0.0006,
+ "loss": 4.353594779968262,
+ "step": 1912
+ },
+ {
+ "epoch": 26.57292576419214,
+ "grad_norm": 0.018090544268488884,
+ "learning_rate": 0.0006,
+ "loss": 4.304343223571777,
+ "step": 1913
+ },
+ {
+ "epoch": 26.586899563318777,
+ "grad_norm": 0.017489777877926826,
+ "learning_rate": 0.0006,
+ "loss": 4.204375267028809,
+ "step": 1914
+ },
+ {
+ "epoch": 26.600873362445416,
+ "grad_norm": 0.016785891726613045,
+ "learning_rate": 0.0006,
+ "loss": 4.286669731140137,
+ "step": 1915
+ },
+ {
+ "epoch": 26.61484716157205,
+ "grad_norm": 0.016382789239287376,
+ "learning_rate": 0.0006,
+ "loss": 4.347662448883057,
+ "step": 1916
+ },
+ {
+ "epoch": 26.62882096069869,
+ "grad_norm": 0.016581876203417778,
+ "learning_rate": 0.0006,
+ "loss": 4.2716193199157715,
+ "step": 1917
+ },
+ {
+ "epoch": 26.64279475982533,
+ "grad_norm": 0.015627184882760048,
+ "learning_rate": 0.0006,
+ "loss": 4.311489105224609,
+ "step": 1918
+ },
+ {
+ "epoch": 26.656768558951963,
+ "grad_norm": 0.015908483415842056,
+ "learning_rate": 0.0006,
+ "loss": 4.262099742889404,
+ "step": 1919
+ },
+ {
+ "epoch": 26.670742358078602,
+ "grad_norm": 0.015576236881315708,
+ "learning_rate": 0.0006,
+ "loss": 4.275871276855469,
+ "step": 1920
+ },
+ {
+ "epoch": 26.68471615720524,
+ "grad_norm": 0.014621064998209476,
+ "learning_rate": 0.0006,
+ "loss": 4.184260845184326,
+ "step": 1921
+ },
+ {
+ "epoch": 26.69868995633188,
+ "grad_norm": 0.014797811396420002,
+ "learning_rate": 0.0006,
+ "loss": 4.3412628173828125,
+ "step": 1922
+ },
+ {
+ "epoch": 26.712663755458514,
+ "grad_norm": 0.015610570088028908,
+ "learning_rate": 0.0006,
+ "loss": 4.181094646453857,
+ "step": 1923
+ },
+ {
+ "epoch": 26.726637554585153,
+ "grad_norm": 0.016048265621066093,
+ "learning_rate": 0.0006,
+ "loss": 4.365290641784668,
+ "step": 1924
+ },
+ {
+ "epoch": 26.74061135371179,
+ "grad_norm": 0.016604071483016014,
+ "learning_rate": 0.0006,
+ "loss": 4.2628912925720215,
+ "step": 1925
+ },
+ {
+ "epoch": 26.754585152838427,
+ "grad_norm": 0.017044221982359886,
+ "learning_rate": 0.0006,
+ "loss": 4.378512382507324,
+ "step": 1926
+ },
+ {
+ "epoch": 26.768558951965066,
+ "grad_norm": 0.01746384985744953,
+ "learning_rate": 0.0006,
+ "loss": 4.3888773918151855,
+ "step": 1927
+ },
+ {
+ "epoch": 26.782532751091704,
+ "grad_norm": 0.01572711206972599,
+ "learning_rate": 0.0006,
+ "loss": 4.189935684204102,
+ "step": 1928
+ },
+ {
+ "epoch": 26.79650655021834,
+ "grad_norm": 0.014976629987359047,
+ "learning_rate": 0.0006,
+ "loss": 4.280525207519531,
+ "step": 1929
+ },
+ {
+ "epoch": 26.810480349344978,
+ "grad_norm": 0.014804664999246597,
+ "learning_rate": 0.0006,
+ "loss": 4.270630836486816,
+ "step": 1930
+ },
+ {
+ "epoch": 26.824454148471617,
+ "grad_norm": 0.014641920104622841,
+ "learning_rate": 0.0006,
+ "loss": 4.319505214691162,
+ "step": 1931
+ },
+ {
+ "epoch": 26.83842794759825,
+ "grad_norm": 0.014928505755960941,
+ "learning_rate": 0.0006,
+ "loss": 4.180147647857666,
+ "step": 1932
+ },
+ {
+ "epoch": 26.85240174672489,
+ "grad_norm": 0.016130365431308746,
+ "learning_rate": 0.0006,
+ "loss": 4.260660648345947,
+ "step": 1933
+ },
+ {
+ "epoch": 26.86637554585153,
+ "grad_norm": 0.015642493963241577,
+ "learning_rate": 0.0006,
+ "loss": 4.304064750671387,
+ "step": 1934
+ },
+ {
+ "epoch": 26.880349344978168,
+ "grad_norm": 0.015988217666745186,
+ "learning_rate": 0.0006,
+ "loss": 4.2548017501831055,
+ "step": 1935
+ },
+ {
+ "epoch": 26.894323144104803,
+ "grad_norm": 0.016501398757100105,
+ "learning_rate": 0.0006,
+ "loss": 4.260058879852295,
+ "step": 1936
+ },
+ {
+ "epoch": 26.90829694323144,
+ "grad_norm": 0.01645040698349476,
+ "learning_rate": 0.0006,
+ "loss": 4.483081340789795,
+ "step": 1937
+ },
+ {
+ "epoch": 26.92227074235808,
+ "grad_norm": 0.016140863299369812,
+ "learning_rate": 0.0006,
+ "loss": 4.1776838302612305,
+ "step": 1938
+ },
+ {
+ "epoch": 26.936244541484715,
+ "grad_norm": 0.016681019216775894,
+ "learning_rate": 0.0006,
+ "loss": 4.224747657775879,
+ "step": 1939
+ },
+ {
+ "epoch": 26.950218340611354,
+ "grad_norm": 0.01650378853082657,
+ "learning_rate": 0.0006,
+ "loss": 4.305703163146973,
+ "step": 1940
+ },
+ {
+ "epoch": 26.964192139737992,
+ "grad_norm": 0.017051683738827705,
+ "learning_rate": 0.0006,
+ "loss": 4.285835266113281,
+ "step": 1941
+ },
+ {
+ "epoch": 26.978165938864628,
+ "grad_norm": 0.016894327476620674,
+ "learning_rate": 0.0006,
+ "loss": 4.318901538848877,
+ "step": 1942
+ },
+ {
+ "epoch": 26.992139737991266,
+ "grad_norm": 0.017313100397586823,
+ "learning_rate": 0.0006,
+ "loss": 4.281434535980225,
+ "step": 1943
+ },
+ {
+ "epoch": 27.0,
+ "grad_norm": 0.018450884148478508,
+ "learning_rate": 0.0006,
+ "loss": 4.335104942321777,
+ "step": 1944
+ },
+ {
+ "epoch": 27.0,
+ "eval_loss": 4.640687465667725,
+ "eval_runtime": 56.9624,
+ "eval_samples_per_second": 42.87,
+ "eval_steps_per_second": 1.352,
+ "step": 1944
+ },
+ {
+ "epoch": 27.01397379912664,
+ "grad_norm": 0.017755698412656784,
+ "learning_rate": 0.0006,
+ "loss": 4.278824329376221,
+ "step": 1945
+ },
+ {
+ "epoch": 27.027947598253274,
+ "grad_norm": 0.01724310778081417,
+ "learning_rate": 0.0006,
+ "loss": 4.311049461364746,
+ "step": 1946
+ },
+ {
+ "epoch": 27.041921397379912,
+ "grad_norm": 0.01690123789012432,
+ "learning_rate": 0.0006,
+ "loss": 4.261994361877441,
+ "step": 1947
+ },
+ {
+ "epoch": 27.05589519650655,
+ "grad_norm": 0.018406696617603302,
+ "learning_rate": 0.0006,
+ "loss": 4.1925153732299805,
+ "step": 1948
+ },
+ {
+ "epoch": 27.069868995633186,
+ "grad_norm": 0.019945867359638214,
+ "learning_rate": 0.0006,
+ "loss": 4.108537197113037,
+ "step": 1949
+ },
+ {
+ "epoch": 27.083842794759825,
+ "grad_norm": 0.020316628739237785,
+ "learning_rate": 0.0006,
+ "loss": 4.209043025970459,
+ "step": 1950
+ },
+ {
+ "epoch": 27.097816593886463,
+ "grad_norm": 0.019690370187163353,
+ "learning_rate": 0.0006,
+ "loss": 4.195789337158203,
+ "step": 1951
+ },
+ {
+ "epoch": 27.111790393013102,
+ "grad_norm": 0.01785232499241829,
+ "learning_rate": 0.0006,
+ "loss": 4.21466064453125,
+ "step": 1952
+ },
+ {
+ "epoch": 27.125764192139737,
+ "grad_norm": 0.0164767038077116,
+ "learning_rate": 0.0006,
+ "loss": 4.270247459411621,
+ "step": 1953
+ },
+ {
+ "epoch": 27.139737991266376,
+ "grad_norm": 0.018008455634117126,
+ "learning_rate": 0.0006,
+ "loss": 4.201042175292969,
+ "step": 1954
+ },
+ {
+ "epoch": 27.153711790393015,
+ "grad_norm": 0.01807340793311596,
+ "learning_rate": 0.0006,
+ "loss": 4.25289249420166,
+ "step": 1955
+ },
+ {
+ "epoch": 27.16768558951965,
+ "grad_norm": 0.016246909275650978,
+ "learning_rate": 0.0006,
+ "loss": 4.116283416748047,
+ "step": 1956
+ },
+ {
+ "epoch": 27.18165938864629,
+ "grad_norm": 0.017545776441693306,
+ "learning_rate": 0.0006,
+ "loss": 4.316999435424805,
+ "step": 1957
+ },
+ {
+ "epoch": 27.195633187772927,
+ "grad_norm": 0.017915375530719757,
+ "learning_rate": 0.0006,
+ "loss": 4.253859519958496,
+ "step": 1958
+ },
+ {
+ "epoch": 27.209606986899562,
+ "grad_norm": 0.018112223595380783,
+ "learning_rate": 0.0006,
+ "loss": 4.2541022300720215,
+ "step": 1959
+ },
+ {
+ "epoch": 27.2235807860262,
+ "grad_norm": 0.01846139505505562,
+ "learning_rate": 0.0006,
+ "loss": 4.20438289642334,
+ "step": 1960
+ },
+ {
+ "epoch": 27.23755458515284,
+ "grad_norm": 0.01943897269666195,
+ "learning_rate": 0.0006,
+ "loss": 4.2799577713012695,
+ "step": 1961
+ },
+ {
+ "epoch": 27.251528384279474,
+ "grad_norm": 0.018253348767757416,
+ "learning_rate": 0.0006,
+ "loss": 4.1720499992370605,
+ "step": 1962
+ },
+ {
+ "epoch": 27.265502183406113,
+ "grad_norm": 0.018292246386408806,
+ "learning_rate": 0.0006,
+ "loss": 4.28231143951416,
+ "step": 1963
+ },
+ {
+ "epoch": 27.27947598253275,
+ "grad_norm": 0.01943155936896801,
+ "learning_rate": 0.0006,
+ "loss": 4.314976692199707,
+ "step": 1964
+ },
+ {
+ "epoch": 27.29344978165939,
+ "grad_norm": 0.019064994528889656,
+ "learning_rate": 0.0006,
+ "loss": 4.25328254699707,
+ "step": 1965
+ },
+ {
+ "epoch": 27.307423580786025,
+ "grad_norm": 0.017423273995518684,
+ "learning_rate": 0.0006,
+ "loss": 4.195992469787598,
+ "step": 1966
+ },
+ {
+ "epoch": 27.321397379912664,
+ "grad_norm": 0.01677924580872059,
+ "learning_rate": 0.0006,
+ "loss": 4.260043144226074,
+ "step": 1967
+ },
+ {
+ "epoch": 27.335371179039303,
+ "grad_norm": 0.01720673404633999,
+ "learning_rate": 0.0006,
+ "loss": 4.222780704498291,
+ "step": 1968
+ },
+ {
+ "epoch": 27.349344978165938,
+ "grad_norm": 0.017300743609666824,
+ "learning_rate": 0.0006,
+ "loss": 4.109524250030518,
+ "step": 1969
+ },
+ {
+ "epoch": 27.363318777292577,
+ "grad_norm": 0.016065871343016624,
+ "learning_rate": 0.0006,
+ "loss": 4.36108922958374,
+ "step": 1970
+ },
+ {
+ "epoch": 27.377292576419215,
+ "grad_norm": 0.016626616939902306,
+ "learning_rate": 0.0006,
+ "loss": 4.220630645751953,
+ "step": 1971
+ },
+ {
+ "epoch": 27.39126637554585,
+ "grad_norm": 0.016482602804899216,
+ "learning_rate": 0.0006,
+ "loss": 4.344797134399414,
+ "step": 1972
+ },
+ {
+ "epoch": 27.40524017467249,
+ "grad_norm": 0.01727953553199768,
+ "learning_rate": 0.0006,
+ "loss": 4.296645164489746,
+ "step": 1973
+ },
+ {
+ "epoch": 27.419213973799128,
+ "grad_norm": 0.01522731315344572,
+ "learning_rate": 0.0006,
+ "loss": 4.223751068115234,
+ "step": 1974
+ },
+ {
+ "epoch": 27.433187772925763,
+ "grad_norm": 0.01581776700913906,
+ "learning_rate": 0.0006,
+ "loss": 4.351379871368408,
+ "step": 1975
+ },
+ {
+ "epoch": 27.4471615720524,
+ "grad_norm": 0.016394605860114098,
+ "learning_rate": 0.0006,
+ "loss": 4.2412919998168945,
+ "step": 1976
+ },
+ {
+ "epoch": 27.46113537117904,
+ "grad_norm": 0.01751169003546238,
+ "learning_rate": 0.0006,
+ "loss": 4.257606506347656,
+ "step": 1977
+ },
+ {
+ "epoch": 27.475109170305675,
+ "grad_norm": 0.017452310770750046,
+ "learning_rate": 0.0006,
+ "loss": 4.380134582519531,
+ "step": 1978
+ },
+ {
+ "epoch": 27.489082969432314,
+ "grad_norm": 0.017741898074746132,
+ "learning_rate": 0.0006,
+ "loss": 4.279632568359375,
+ "step": 1979
+ },
+ {
+ "epoch": 27.503056768558952,
+ "grad_norm": 0.017680590972304344,
+ "learning_rate": 0.0006,
+ "loss": 4.273406028747559,
+ "step": 1980
+ },
+ {
+ "epoch": 27.51703056768559,
+ "grad_norm": 0.01864718459546566,
+ "learning_rate": 0.0006,
+ "loss": 4.278583526611328,
+ "step": 1981
+ },
+ {
+ "epoch": 27.531004366812226,
+ "grad_norm": 0.019558541476726532,
+ "learning_rate": 0.0006,
+ "loss": 4.211919784545898,
+ "step": 1982
+ },
+ {
+ "epoch": 27.544978165938865,
+ "grad_norm": 0.019467400386929512,
+ "learning_rate": 0.0006,
+ "loss": 4.276822566986084,
+ "step": 1983
+ },
+ {
+ "epoch": 27.558951965065503,
+ "grad_norm": 0.01756172813475132,
+ "learning_rate": 0.0006,
+ "loss": 4.165379047393799,
+ "step": 1984
+ },
+ {
+ "epoch": 27.57292576419214,
+ "grad_norm": 0.017458263784646988,
+ "learning_rate": 0.0006,
+ "loss": 4.2992401123046875,
+ "step": 1985
+ },
+ {
+ "epoch": 27.586899563318777,
+ "grad_norm": 0.018089665099978447,
+ "learning_rate": 0.0006,
+ "loss": 4.243378639221191,
+ "step": 1986
+ },
+ {
+ "epoch": 27.600873362445416,
+ "grad_norm": 0.019535524770617485,
+ "learning_rate": 0.0006,
+ "loss": 4.321477890014648,
+ "step": 1987
+ },
+ {
+ "epoch": 27.61484716157205,
+ "grad_norm": 0.01997970975935459,
+ "learning_rate": 0.0006,
+ "loss": 4.123082637786865,
+ "step": 1988
+ },
+ {
+ "epoch": 27.62882096069869,
+ "grad_norm": 0.01913067139685154,
+ "learning_rate": 0.0006,
+ "loss": 4.289896488189697,
+ "step": 1989
+ },
+ {
+ "epoch": 27.64279475982533,
+ "grad_norm": 0.0169806070625782,
+ "learning_rate": 0.0006,
+ "loss": 4.247004985809326,
+ "step": 1990
+ },
+ {
+ "epoch": 27.656768558951963,
+ "grad_norm": 0.017315007746219635,
+ "learning_rate": 0.0006,
+ "loss": 4.2095441818237305,
+ "step": 1991
+ },
+ {
+ "epoch": 27.670742358078602,
+ "grad_norm": 0.015098759904503822,
+ "learning_rate": 0.0006,
+ "loss": 4.145560264587402,
+ "step": 1992
+ },
+ {
+ "epoch": 27.68471615720524,
+ "grad_norm": 0.016240514814853668,
+ "learning_rate": 0.0006,
+ "loss": 4.254927635192871,
+ "step": 1993
+ },
+ {
+ "epoch": 27.69868995633188,
+ "grad_norm": 0.01621108129620552,
+ "learning_rate": 0.0006,
+ "loss": 4.219725608825684,
+ "step": 1994
+ },
+ {
+ "epoch": 27.712663755458514,
+ "grad_norm": 0.015191009268164635,
+ "learning_rate": 0.0006,
+ "loss": 4.194345474243164,
+ "step": 1995
+ },
+ {
+ "epoch": 27.726637554585153,
+ "grad_norm": 0.014387983828783035,
+ "learning_rate": 0.0006,
+ "loss": 4.211467742919922,
+ "step": 1996
+ },
+ {
+ "epoch": 27.74061135371179,
+ "grad_norm": 0.015460561029613018,
+ "learning_rate": 0.0006,
+ "loss": 4.358700752258301,
+ "step": 1997
+ },
+ {
+ "epoch": 27.754585152838427,
+ "grad_norm": 0.015100076794624329,
+ "learning_rate": 0.0006,
+ "loss": 4.171223163604736,
+ "step": 1998
+ },
+ {
+ "epoch": 27.768558951965066,
+ "grad_norm": 0.015607891604304314,
+ "learning_rate": 0.0006,
+ "loss": 4.217874526977539,
+ "step": 1999
+ },
+ {
+ "epoch": 27.782532751091704,
+ "grad_norm": 0.015650689601898193,
+ "learning_rate": 0.0006,
+ "loss": 4.280780792236328,
+ "step": 2000
+ },
+ {
+ "epoch": 27.79650655021834,
+ "grad_norm": 0.01580994389951229,
+ "learning_rate": 0.0006,
+ "loss": 4.159212112426758,
+ "step": 2001
+ },
+ {
+ "epoch": 27.810480349344978,
+ "grad_norm": 0.016607413068413734,
+ "learning_rate": 0.0006,
+ "loss": 4.238853454589844,
+ "step": 2002
+ },
+ {
+ "epoch": 27.824454148471617,
+ "grad_norm": 0.017383860424160957,
+ "learning_rate": 0.0006,
+ "loss": 4.233944892883301,
+ "step": 2003
+ },
+ {
+ "epoch": 27.83842794759825,
+ "grad_norm": 0.014966659247875214,
+ "learning_rate": 0.0006,
+ "loss": 4.167392730712891,
+ "step": 2004
+ },
+ {
+ "epoch": 27.85240174672489,
+ "grad_norm": 0.014589878730475903,
+ "learning_rate": 0.0006,
+ "loss": 4.2390899658203125,
+ "step": 2005
+ },
+ {
+ "epoch": 27.86637554585153,
+ "grad_norm": 0.017179450020194054,
+ "learning_rate": 0.0006,
+ "loss": 4.365785598754883,
+ "step": 2006
+ },
+ {
+ "epoch": 27.880349344978168,
+ "grad_norm": 0.01893901452422142,
+ "learning_rate": 0.0006,
+ "loss": 4.289430141448975,
+ "step": 2007
+ },
+ {
+ "epoch": 27.894323144104803,
+ "grad_norm": 0.018619712442159653,
+ "learning_rate": 0.0006,
+ "loss": 4.310665130615234,
+ "step": 2008
+ },
+ {
+ "epoch": 27.90829694323144,
+ "grad_norm": 0.019180385395884514,
+ "learning_rate": 0.0006,
+ "loss": 4.27454137802124,
+ "step": 2009
+ },
+ {
+ "epoch": 27.92227074235808,
+ "grad_norm": 0.017811523750424385,
+ "learning_rate": 0.0006,
+ "loss": 4.312562465667725,
+ "step": 2010
+ },
+ {
+ "epoch": 27.936244541484715,
+ "grad_norm": 0.016583573073148727,
+ "learning_rate": 0.0006,
+ "loss": 4.202253341674805,
+ "step": 2011
+ },
+ {
+ "epoch": 27.950218340611354,
+ "grad_norm": 0.01658615842461586,
+ "learning_rate": 0.0006,
+ "loss": 4.183433532714844,
+ "step": 2012
+ },
+ {
+ "epoch": 27.964192139737992,
+ "grad_norm": 0.015327401459217072,
+ "learning_rate": 0.0006,
+ "loss": 4.2125139236450195,
+ "step": 2013
+ },
+ {
+ "epoch": 27.978165938864628,
+ "grad_norm": 0.015004601329565048,
+ "learning_rate": 0.0006,
+ "loss": 4.257392883300781,
+ "step": 2014
+ },
+ {
+ "epoch": 27.992139737991266,
+ "grad_norm": 0.015259161591529846,
+ "learning_rate": 0.0006,
+ "loss": 4.241818428039551,
+ "step": 2015
+ },
+ {
+ "epoch": 28.0,
+ "grad_norm": 0.016808776184916496,
+ "learning_rate": 0.0006,
+ "loss": 4.146241188049316,
+ "step": 2016
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 28800,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 400,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": true,
+ "should_log": false,
+ "should_save": false,
+ "should_training_stop": false
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 8.594917411730227e+17,
+ "train_batch_size": 8,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2016/training_args.bin b/runs/l2r50-i2-fulle-lm/checkpoint-2016/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..cad232831c222cc7f8b938b8bbe42f61e7d89c3a
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-2016/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:f7c40e3bfa2f5ac1450ca1243e3f2a895d44a06ed5995e6aefd97f78d229d8c8
+size 4792
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2304/chat_template.jinja b/runs/l2r50-i2-fulle-lm/checkpoint-2304/chat_template.jinja
new file mode 100644
index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-2304/chat_template.jinja
@@ -0,0 +1,85 @@
+{%- if tools %}
+ {{- '<|im_start|>system\n' }}
+ {%- if messages[0].role == 'system' %}
+ {{- messages[0].content + '\n\n' }}
+ {%- endif %}
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }}
+ {%- for tool in tools %}
+ {{- "\n" }}
+ {{- tool | tojson }}
+ {%- endfor %}
+ {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }}
+{%- else %}
+ {%- if messages[0].role == 'system' %}
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
+ {%- endif %}
+{%- endif %}
+{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
+{%- for message in messages[::-1] %}
+ {%- set index = (messages|length - 1) - loop.index0 %}
+ {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %}
+ {%- set ns.multi_step_tool = false %}
+ {%- set ns.last_query_index = index %}
+ {%- endif %}
+{%- endfor %}
+{%- for message in messages %}
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
+ {%- elif message.role == "assistant" %}
+ {%- set content = message.content %}
+ {%- set reasoning_content = '' %}
+ {%- if message.reasoning_content is defined and message.reasoning_content is not none %}
+ {%- set reasoning_content = message.reasoning_content %}
+ {%- else %}
+ {%- if '' in message.content %}
+ {%- set content = message.content.split('')[-1].lstrip('\n') %}
+ {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %}
+ {%- endif %}
+ {%- endif %}
+ {%- if loop.index0 > ns.last_query_index %}
+ {%- if loop.last or (not loop.last and reasoning_content) %}
+ {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- if message.tool_calls %}
+ {%- for tool_call in message.tool_calls %}
+ {%- if (loop.first and content) or (not loop.first) %}
+ {{- '\n' }}
+ {%- endif %}
+ {%- if tool_call.function %}
+ {%- set tool_call = tool_call.function %}
+ {%- endif %}
+ {{- '\n{"name": "' }}
+ {{- tool_call.name }}
+ {{- '", "arguments": ' }}
+ {%- if tool_call.arguments is string %}
+ {{- tool_call.arguments }}
+ {%- else %}
+ {{- tool_call.arguments | tojson }}
+ {%- endif %}
+ {{- '}\n' }}
+ {%- endfor %}
+ {%- endif %}
+ {{- '<|im_end|>\n' }}
+ {%- elif message.role == "tool" %}
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
+ {{- '<|im_start|>user' }}
+ {%- endif %}
+ {{- '\n\n' }}
+ {{- message.content }}
+ {{- '\n' }}
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
+ {{- '<|im_end|>\n' }}
+ {%- endif %}
+ {%- endif %}
+{%- endfor %}
+{%- if add_generation_prompt %}
+ {{- '<|im_start|>assistant\n' }}
+ {%- if enable_thinking is defined and enable_thinking is false %}
+ {{- '\n\n\n\n' }}
+ {%- endif %}
+{%- endif %}
\ No newline at end of file
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2304/config.json b/runs/l2r50-i2-fulle-lm/checkpoint-2304/config.json
new file mode 100644
index 0000000000000000000000000000000000000000..22f250953c6ce8fc91ee4411856dd73041a415e2
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-2304/config.json
@@ -0,0 +1,32 @@
+{
+ "architectures": [
+ "LlamaForCausalLM"
+ ],
+ "attention_bias": false,
+ "attention_dropout": 0.0,
+ "bos_token_id": null,
+ "dtype": "float32",
+ "eos_token_id": 151645,
+ "head_dim": 128,
+ "hidden_act": "silu",
+ "hidden_size": 512,
+ "initializer_range": 0.02,
+ "intermediate_size": 1536,
+ "max_position_embeddings": 2048,
+ "mlp_bias": false,
+ "model_type": "llama",
+ "num_attention_heads": 4,
+ "num_hidden_layers": 20,
+ "num_key_value_heads": 4,
+ "pad_token_id": 151645,
+ "pretraining_tp": 1,
+ "rms_norm_eps": 1e-06,
+ "rope_parameters": {
+ "rope_theta": 10000.0,
+ "rope_type": "default"
+ },
+ "tie_word_embeddings": true,
+ "transformers_version": "5.5.0",
+ "use_cache": false,
+ "vocab_size": 151673
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2304/generation_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-2304/generation_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-2304/generation_config.json
@@ -0,0 +1,11 @@
+{
+ "_from_model_config": true,
+ "eos_token_id": [
+ 151645
+ ],
+ "output_attentions": false,
+ "output_hidden_states": false,
+ "pad_token_id": 151645,
+ "transformers_version": "5.5.0",
+ "use_cache": true
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2304/model.safetensors b/runs/l2r50-i2-fulle-lm/checkpoint-2304/model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..4337a74bfb72bc6067e870fcd49a307876c60026
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-2304/model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:612be7549a62726fad1ba1b692e96ff4def8b87fe5608f4921acfb3519bca075
+size 583360328
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2304/optimizer.pt b/runs/l2r50-i2-fulle-lm/checkpoint-2304/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..67837922fac0e574d863ddd1e699f0129b800285
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-2304/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:c82e59d4958422cd14ffb582877b08f0d0a432632dd755c3bea74ededea0619d
+size 1166833530
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2304/rng_state_0.pth b/runs/l2r50-i2-fulle-lm/checkpoint-2304/rng_state_0.pth
new file mode 100644
index 0000000000000000000000000000000000000000..92fdc90c18815d9c22b41c21650399ae0343759f
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-2304/rng_state_0.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:d426ace1ca5421fc259c01bfd79180bbe5cdfc7bd4b796e00828010457442e33
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2304/rng_state_1.pth b/runs/l2r50-i2-fulle-lm/checkpoint-2304/rng_state_1.pth
new file mode 100644
index 0000000000000000000000000000000000000000..94d6c5a589dd3c39bd39deb11b0e5e4f96dd736d
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-2304/rng_state_1.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:c83a37072b1808e469343f57559d31ce3929503dc3811a85a5a42679611c3b7c
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2304/rng_state_2.pth b/runs/l2r50-i2-fulle-lm/checkpoint-2304/rng_state_2.pth
new file mode 100644
index 0000000000000000000000000000000000000000..b68205c575bd5490955dcfa1cd063fbb48a82438
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-2304/rng_state_2.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:bb27032a62c16f6a117bd2e85c89f4d1b272c3ac5b94a1352c5a6816c78ddc8f
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2304/rng_state_3.pth b/runs/l2r50-i2-fulle-lm/checkpoint-2304/rng_state_3.pth
new file mode 100644
index 0000000000000000000000000000000000000000..262450393b4af9c8e06fe55890185cdfea10912f
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-2304/rng_state_3.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:1b69ca38c0bb75456b1a48beda9c6d53ff29e394c82869433936bd6cc5e9193f
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2304/scheduler.pt b/runs/l2r50-i2-fulle-lm/checkpoint-2304/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..f90b8c27b6972bbb4a1ca0bc0dffdc47237e677d
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-2304/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:b057fe862d340f767a25cc1e8c8a0bb24146431f818f82ef2cc9b3d15cf7bfad
+size 1064
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2304/tokenizer.json b/runs/l2r50-i2-fulle-lm/checkpoint-2304/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..53786016f070e11184813f3c83101dc5e93e392e
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-2304/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:c3c63fda1646a555448e8f03662542e04b11d103721319e4d9b267d7da03ceb8
+size 11424463
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2304/tokenizer_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-2304/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..98e2fb9b2cc54785dc9d871b8e503f8b8f2defe7
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-2304/tokenizer_config.json
@@ -0,0 +1,20 @@
+{
+ "add_prefix_space": false,
+ "backend": "tokenizers",
+ "bos_token": null,
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "<|im_end|>",
+ "errors": "replace",
+ "extra_special_tokens": [
+ "<|l2r_pred|>",
+ "<|r2l_pred|>",
+ "<|next_1_pred|>",
+ "<|next_2_pred|>"
+ ],
+ "is_local": false,
+ "model_max_length": 131072,
+ "pad_token": "<|im_end|>",
+ "split_special_tokens": false,
+ "tokenizer_class": "Qwen2Tokenizer",
+ "unk_token": null
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2304/trainer_state.json b/runs/l2r50-i2-fulle-lm/checkpoint-2304/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..3a478f1679d0f2987a542859944e53bb753ddd9a
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-2304/trainer_state.json
@@ -0,0 +1,16410 @@
+{
+ "best_global_step": null,
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 32.0,
+ "eval_steps": 500,
+ "global_step": 2304,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.013973799126637555,
+ "grad_norm": 0.1739506870508194,
+ "learning_rate": 0.0,
+ "loss": 11.99299430847168,
+ "step": 1
+ },
+ {
+ "epoch": 0.02794759825327511,
+ "grad_norm": 0.17216597497463226,
+ "learning_rate": 5.999999999999999e-06,
+ "loss": 11.988704681396484,
+ "step": 2
+ },
+ {
+ "epoch": 0.04192139737991266,
+ "grad_norm": 0.17459715902805328,
+ "learning_rate": 1.1999999999999999e-05,
+ "loss": 11.945581436157227,
+ "step": 3
+ },
+ {
+ "epoch": 0.05589519650655022,
+ "grad_norm": 0.17577414214611053,
+ "learning_rate": 1.7999999999999997e-05,
+ "loss": 11.859772682189941,
+ "step": 4
+ },
+ {
+ "epoch": 0.06986899563318777,
+ "grad_norm": 0.16163651645183563,
+ "learning_rate": 2.3999999999999997e-05,
+ "loss": 11.744415283203125,
+ "step": 5
+ },
+ {
+ "epoch": 0.08384279475982533,
+ "grad_norm": 0.14409932494163513,
+ "learning_rate": 2.9999999999999997e-05,
+ "loss": 11.618099212646484,
+ "step": 6
+ },
+ {
+ "epoch": 0.09781659388646288,
+ "grad_norm": 0.13079848885536194,
+ "learning_rate": 3.5999999999999994e-05,
+ "loss": 11.505935668945312,
+ "step": 7
+ },
+ {
+ "epoch": 0.11179039301310044,
+ "grad_norm": 0.11964613944292068,
+ "learning_rate": 4.2e-05,
+ "loss": 11.416532516479492,
+ "step": 8
+ },
+ {
+ "epoch": 0.125764192139738,
+ "grad_norm": 0.11345142871141434,
+ "learning_rate": 4.7999999999999994e-05,
+ "loss": 11.33206558227539,
+ "step": 9
+ },
+ {
+ "epoch": 0.13973799126637554,
+ "grad_norm": 0.11033952981233597,
+ "learning_rate": 5.399999999999999e-05,
+ "loss": 11.259052276611328,
+ "step": 10
+ },
+ {
+ "epoch": 0.1537117903930131,
+ "grad_norm": 0.10776489973068237,
+ "learning_rate": 5.9999999999999995e-05,
+ "loss": 11.203384399414062,
+ "step": 11
+ },
+ {
+ "epoch": 0.16768558951965065,
+ "grad_norm": 0.10721946507692337,
+ "learning_rate": 6.599999999999999e-05,
+ "loss": 11.153295516967773,
+ "step": 12
+ },
+ {
+ "epoch": 0.18165938864628822,
+ "grad_norm": 0.10673430562019348,
+ "learning_rate": 7.199999999999999e-05,
+ "loss": 11.110200881958008,
+ "step": 13
+ },
+ {
+ "epoch": 0.19563318777292577,
+ "grad_norm": 0.10630254447460175,
+ "learning_rate": 7.8e-05,
+ "loss": 11.067477226257324,
+ "step": 14
+ },
+ {
+ "epoch": 0.2096069868995633,
+ "grad_norm": 0.10631615668535233,
+ "learning_rate": 8.4e-05,
+ "loss": 11.022941589355469,
+ "step": 15
+ },
+ {
+ "epoch": 0.22358078602620088,
+ "grad_norm": 0.10609443485736847,
+ "learning_rate": 8.999999999999999e-05,
+ "loss": 10.970746040344238,
+ "step": 16
+ },
+ {
+ "epoch": 0.23755458515283842,
+ "grad_norm": 0.10648232698440552,
+ "learning_rate": 9.599999999999999e-05,
+ "loss": 10.91267204284668,
+ "step": 17
+ },
+ {
+ "epoch": 0.251528384279476,
+ "grad_norm": 0.10584589838981628,
+ "learning_rate": 0.000102,
+ "loss": 10.854547500610352,
+ "step": 18
+ },
+ {
+ "epoch": 0.26550218340611353,
+ "grad_norm": 0.10591720044612885,
+ "learning_rate": 0.00010799999999999998,
+ "loss": 10.788917541503906,
+ "step": 19
+ },
+ {
+ "epoch": 0.2794759825327511,
+ "grad_norm": 0.10611672699451447,
+ "learning_rate": 0.00011399999999999999,
+ "loss": 10.717556953430176,
+ "step": 20
+ },
+ {
+ "epoch": 0.2934497816593886,
+ "grad_norm": 0.10455359518527985,
+ "learning_rate": 0.00011999999999999999,
+ "loss": 10.650186538696289,
+ "step": 21
+ },
+ {
+ "epoch": 0.3074235807860262,
+ "grad_norm": 0.1047525629401207,
+ "learning_rate": 0.00012599999999999997,
+ "loss": 10.56808090209961,
+ "step": 22
+ },
+ {
+ "epoch": 0.32139737991266376,
+ "grad_norm": 0.10511605441570282,
+ "learning_rate": 0.00013199999999999998,
+ "loss": 10.484028816223145,
+ "step": 23
+ },
+ {
+ "epoch": 0.3353711790393013,
+ "grad_norm": 0.10467177629470825,
+ "learning_rate": 0.000138,
+ "loss": 10.399947166442871,
+ "step": 24
+ },
+ {
+ "epoch": 0.34934497816593885,
+ "grad_norm": 0.10432136058807373,
+ "learning_rate": 0.00014399999999999998,
+ "loss": 10.312714576721191,
+ "step": 25
+ },
+ {
+ "epoch": 0.36331877729257644,
+ "grad_norm": 0.10527262091636658,
+ "learning_rate": 0.00015,
+ "loss": 10.212261199951172,
+ "step": 26
+ },
+ {
+ "epoch": 0.377292576419214,
+ "grad_norm": 0.103610560297966,
+ "learning_rate": 0.000156,
+ "loss": 10.132034301757812,
+ "step": 27
+ },
+ {
+ "epoch": 0.39126637554585153,
+ "grad_norm": 0.10553059726953506,
+ "learning_rate": 0.000162,
+ "loss": 10.020434379577637,
+ "step": 28
+ },
+ {
+ "epoch": 0.4052401746724891,
+ "grad_norm": 0.1037299782037735,
+ "learning_rate": 0.000168,
+ "loss": 9.934466361999512,
+ "step": 29
+ },
+ {
+ "epoch": 0.4192139737991266,
+ "grad_norm": 0.1042739674448967,
+ "learning_rate": 0.00017399999999999997,
+ "loss": 9.822539329528809,
+ "step": 30
+ },
+ {
+ "epoch": 0.4331877729257642,
+ "grad_norm": 0.10300865024328232,
+ "learning_rate": 0.00017999999999999998,
+ "loss": 9.734994888305664,
+ "step": 31
+ },
+ {
+ "epoch": 0.44716157205240176,
+ "grad_norm": 0.10291654616594315,
+ "learning_rate": 0.000186,
+ "loss": 9.624505996704102,
+ "step": 32
+ },
+ {
+ "epoch": 0.4611353711790393,
+ "grad_norm": 0.10210851579904556,
+ "learning_rate": 0.00019199999999999998,
+ "loss": 9.518205642700195,
+ "step": 33
+ },
+ {
+ "epoch": 0.47510917030567684,
+ "grad_norm": 0.10222529619932175,
+ "learning_rate": 0.000198,
+ "loss": 9.412222862243652,
+ "step": 34
+ },
+ {
+ "epoch": 0.4890829694323144,
+ "grad_norm": 0.09992019087076187,
+ "learning_rate": 0.000204,
+ "loss": 9.326566696166992,
+ "step": 35
+ },
+ {
+ "epoch": 0.503056768558952,
+ "grad_norm": 0.09931084513664246,
+ "learning_rate": 0.00020999999999999998,
+ "loss": 9.210111618041992,
+ "step": 36
+ },
+ {
+ "epoch": 0.5170305676855895,
+ "grad_norm": 0.09928256273269653,
+ "learning_rate": 0.00021599999999999996,
+ "loss": 9.100493431091309,
+ "step": 37
+ },
+ {
+ "epoch": 0.5310043668122271,
+ "grad_norm": 0.09843816608190536,
+ "learning_rate": 0.00022199999999999998,
+ "loss": 8.996543884277344,
+ "step": 38
+ },
+ {
+ "epoch": 0.5449781659388646,
+ "grad_norm": 0.09825374186038971,
+ "learning_rate": 0.00022799999999999999,
+ "loss": 8.888991355895996,
+ "step": 39
+ },
+ {
+ "epoch": 0.5589519650655022,
+ "grad_norm": 0.09602385759353638,
+ "learning_rate": 0.000234,
+ "loss": 8.802907943725586,
+ "step": 40
+ },
+ {
+ "epoch": 0.5729257641921397,
+ "grad_norm": 0.09414064139127731,
+ "learning_rate": 0.00023999999999999998,
+ "loss": 8.715423583984375,
+ "step": 41
+ },
+ {
+ "epoch": 0.5868995633187772,
+ "grad_norm": 0.09273893386125565,
+ "learning_rate": 0.00024599999999999996,
+ "loss": 8.61374282836914,
+ "step": 42
+ },
+ {
+ "epoch": 0.6008733624454149,
+ "grad_norm": 0.09079599380493164,
+ "learning_rate": 0.00025199999999999995,
+ "loss": 8.502904891967773,
+ "step": 43
+ },
+ {
+ "epoch": 0.6148471615720524,
+ "grad_norm": 0.08605825901031494,
+ "learning_rate": 0.000258,
+ "loss": 8.453008651733398,
+ "step": 44
+ },
+ {
+ "epoch": 0.62882096069869,
+ "grad_norm": 0.08338996022939682,
+ "learning_rate": 0.00026399999999999997,
+ "loss": 8.365163803100586,
+ "step": 45
+ },
+ {
+ "epoch": 0.6427947598253275,
+ "grad_norm": 0.08150867372751236,
+ "learning_rate": 0.00027,
+ "loss": 8.285372734069824,
+ "step": 46
+ },
+ {
+ "epoch": 0.6567685589519651,
+ "grad_norm": 0.07794807851314545,
+ "learning_rate": 0.000276,
+ "loss": 8.185811996459961,
+ "step": 47
+ },
+ {
+ "epoch": 0.6707423580786026,
+ "grad_norm": 0.07592720538377762,
+ "learning_rate": 0.00028199999999999997,
+ "loss": 8.098482131958008,
+ "step": 48
+ },
+ {
+ "epoch": 0.6847161572052402,
+ "grad_norm": 0.0709897130727768,
+ "learning_rate": 0.00028799999999999995,
+ "loss": 8.042091369628906,
+ "step": 49
+ },
+ {
+ "epoch": 0.6986899563318777,
+ "grad_norm": 0.06582305580377579,
+ "learning_rate": 0.000294,
+ "loss": 7.972844123840332,
+ "step": 50
+ },
+ {
+ "epoch": 0.7126637554585152,
+ "grad_norm": 0.059452351182699203,
+ "learning_rate": 0.0003,
+ "loss": 7.924579620361328,
+ "step": 51
+ },
+ {
+ "epoch": 0.7266375545851529,
+ "grad_norm": 0.052426718175411224,
+ "learning_rate": 0.00030599999999999996,
+ "loss": 7.876336097717285,
+ "step": 52
+ },
+ {
+ "epoch": 0.7406113537117904,
+ "grad_norm": 0.04768150672316551,
+ "learning_rate": 0.000312,
+ "loss": 7.8102707862854,
+ "step": 53
+ },
+ {
+ "epoch": 0.754585152838428,
+ "grad_norm": 0.03970075398683548,
+ "learning_rate": 0.000318,
+ "loss": 7.791913986206055,
+ "step": 54
+ },
+ {
+ "epoch": 0.7685589519650655,
+ "grad_norm": 0.03559752553701401,
+ "learning_rate": 0.000324,
+ "loss": 7.756960868835449,
+ "step": 55
+ },
+ {
+ "epoch": 0.7825327510917031,
+ "grad_norm": 0.029321353882551193,
+ "learning_rate": 0.00033,
+ "loss": 7.724878787994385,
+ "step": 56
+ },
+ {
+ "epoch": 0.7965065502183406,
+ "grad_norm": 0.023485427722334862,
+ "learning_rate": 0.000336,
+ "loss": 7.72553014755249,
+ "step": 57
+ },
+ {
+ "epoch": 0.8104803493449781,
+ "grad_norm": 0.019319474697113037,
+ "learning_rate": 0.00034199999999999996,
+ "loss": 7.709493160247803,
+ "step": 58
+ },
+ {
+ "epoch": 0.8244541484716157,
+ "grad_norm": 0.016705691814422607,
+ "learning_rate": 0.00034799999999999995,
+ "loss": 7.6873297691345215,
+ "step": 59
+ },
+ {
+ "epoch": 0.8384279475982532,
+ "grad_norm": 0.017667734995484352,
+ "learning_rate": 0.00035399999999999993,
+ "loss": 7.677059173583984,
+ "step": 60
+ },
+ {
+ "epoch": 0.8524017467248908,
+ "grad_norm": 0.01671992614865303,
+ "learning_rate": 0.00035999999999999997,
+ "loss": 7.643512725830078,
+ "step": 61
+ },
+ {
+ "epoch": 0.8663755458515284,
+ "grad_norm": 0.012288172729313374,
+ "learning_rate": 0.00036599999999999995,
+ "loss": 7.658343315124512,
+ "step": 62
+ },
+ {
+ "epoch": 0.880349344978166,
+ "grad_norm": 0.012004776857793331,
+ "learning_rate": 0.000372,
+ "loss": 7.649759292602539,
+ "step": 63
+ },
+ {
+ "epoch": 0.8943231441048035,
+ "grad_norm": 0.012487445957958698,
+ "learning_rate": 0.00037799999999999997,
+ "loss": 7.631503105163574,
+ "step": 64
+ },
+ {
+ "epoch": 0.9082969432314411,
+ "grad_norm": 0.01489241048693657,
+ "learning_rate": 0.00038399999999999996,
+ "loss": 7.647110462188721,
+ "step": 65
+ },
+ {
+ "epoch": 0.9222707423580786,
+ "grad_norm": 0.01428980939090252,
+ "learning_rate": 0.00039,
+ "loss": 7.636983871459961,
+ "step": 66
+ },
+ {
+ "epoch": 0.9362445414847161,
+ "grad_norm": 0.010711144655942917,
+ "learning_rate": 0.000396,
+ "loss": 7.621130466461182,
+ "step": 67
+ },
+ {
+ "epoch": 0.9502183406113537,
+ "grad_norm": 0.010178886353969574,
+ "learning_rate": 0.000402,
+ "loss": 7.593924522399902,
+ "step": 68
+ },
+ {
+ "epoch": 0.9641921397379912,
+ "grad_norm": 0.010604874230921268,
+ "learning_rate": 0.000408,
+ "loss": 7.6015496253967285,
+ "step": 69
+ },
+ {
+ "epoch": 0.9781659388646288,
+ "grad_norm": 0.009751426056027412,
+ "learning_rate": 0.0004139999999999999,
+ "loss": 7.571964263916016,
+ "step": 70
+ },
+ {
+ "epoch": 0.9921397379912664,
+ "grad_norm": 0.009816481731832027,
+ "learning_rate": 0.00041999999999999996,
+ "loss": 7.57344913482666,
+ "step": 71
+ },
+ {
+ "epoch": 1.0,
+ "grad_norm": 0.008665637113153934,
+ "learning_rate": 0.00042599999999999995,
+ "loss": 7.579050064086914,
+ "step": 72
+ },
+ {
+ "epoch": 1.0,
+ "eval_loss": 7.581204891204834,
+ "eval_runtime": 59.0091,
+ "eval_samples_per_second": 41.383,
+ "eval_steps_per_second": 1.305,
+ "step": 72
+ },
+ {
+ "epoch": 1.0139737991266375,
+ "grad_norm": 0.009684201329946518,
+ "learning_rate": 0.00043199999999999993,
+ "loss": 7.568594932556152,
+ "step": 73
+ },
+ {
+ "epoch": 1.027947598253275,
+ "grad_norm": 0.025783710181713104,
+ "learning_rate": 0.00043799999999999997,
+ "loss": 7.578948497772217,
+ "step": 74
+ },
+ {
+ "epoch": 1.0419213973799126,
+ "grad_norm": 0.0327819399535656,
+ "learning_rate": 0.00044399999999999995,
+ "loss": 7.548173904418945,
+ "step": 75
+ },
+ {
+ "epoch": 1.0558951965065502,
+ "grad_norm": 0.021604498848319054,
+ "learning_rate": 0.00045,
+ "loss": 7.552953720092773,
+ "step": 76
+ },
+ {
+ "epoch": 1.0698689956331877,
+ "grad_norm": 0.010230105370283127,
+ "learning_rate": 0.00045599999999999997,
+ "loss": 7.5213212966918945,
+ "step": 77
+ },
+ {
+ "epoch": 1.0838427947598253,
+ "grad_norm": 0.023315589874982834,
+ "learning_rate": 0.00046199999999999995,
+ "loss": 7.501640319824219,
+ "step": 78
+ },
+ {
+ "epoch": 1.0978165938864628,
+ "grad_norm": 0.01109944935888052,
+ "learning_rate": 0.000468,
+ "loss": 7.539519786834717,
+ "step": 79
+ },
+ {
+ "epoch": 1.1117903930131003,
+ "grad_norm": 0.026093758642673492,
+ "learning_rate": 0.000474,
+ "loss": 7.525928974151611,
+ "step": 80
+ },
+ {
+ "epoch": 1.125764192139738,
+ "grad_norm": 0.03274895250797272,
+ "learning_rate": 0.00047999999999999996,
+ "loss": 7.535602569580078,
+ "step": 81
+ },
+ {
+ "epoch": 1.1397379912663754,
+ "grad_norm": 0.01543602254241705,
+ "learning_rate": 0.000486,
+ "loss": 7.513683795928955,
+ "step": 82
+ },
+ {
+ "epoch": 1.1537117903930132,
+ "grad_norm": 0.028377249836921692,
+ "learning_rate": 0.0004919999999999999,
+ "loss": 7.49384880065918,
+ "step": 83
+ },
+ {
+ "epoch": 1.1676855895196507,
+ "grad_norm": 0.04121481627225876,
+ "learning_rate": 0.000498,
+ "loss": 7.488447189331055,
+ "step": 84
+ },
+ {
+ "epoch": 1.1816593886462883,
+ "grad_norm": 0.013282302767038345,
+ "learning_rate": 0.0005039999999999999,
+ "loss": 7.463010787963867,
+ "step": 85
+ },
+ {
+ "epoch": 1.1956331877729258,
+ "grad_norm": 0.03200364485383034,
+ "learning_rate": 0.0005099999999999999,
+ "loss": 7.473515033721924,
+ "step": 86
+ },
+ {
+ "epoch": 1.2096069868995634,
+ "grad_norm": 0.020768029615283012,
+ "learning_rate": 0.000516,
+ "loss": 7.464688777923584,
+ "step": 87
+ },
+ {
+ "epoch": 1.223580786026201,
+ "grad_norm": 0.021715788170695305,
+ "learning_rate": 0.000522,
+ "loss": 7.4835028648376465,
+ "step": 88
+ },
+ {
+ "epoch": 1.2375545851528384,
+ "grad_norm": 0.026255754753947258,
+ "learning_rate": 0.0005279999999999999,
+ "loss": 7.443392753601074,
+ "step": 89
+ },
+ {
+ "epoch": 1.251528384279476,
+ "grad_norm": 0.022325951606035233,
+ "learning_rate": 0.000534,
+ "loss": 7.457373142242432,
+ "step": 90
+ },
+ {
+ "epoch": 1.2655021834061135,
+ "grad_norm": 0.009270387701690197,
+ "learning_rate": 0.00054,
+ "loss": 7.455986976623535,
+ "step": 91
+ },
+ {
+ "epoch": 1.279475982532751,
+ "grad_norm": 0.02369544468820095,
+ "learning_rate": 0.0005459999999999999,
+ "loss": 7.442216873168945,
+ "step": 92
+ },
+ {
+ "epoch": 1.2934497816593886,
+ "grad_norm": 0.008474506437778473,
+ "learning_rate": 0.000552,
+ "loss": 7.4294915199279785,
+ "step": 93
+ },
+ {
+ "epoch": 1.3074235807860262,
+ "grad_norm": 0.017043249681591988,
+ "learning_rate": 0.000558,
+ "loss": 7.406932830810547,
+ "step": 94
+ },
+ {
+ "epoch": 1.3213973799126637,
+ "grad_norm": 0.011813412420451641,
+ "learning_rate": 0.0005639999999999999,
+ "loss": 7.415900230407715,
+ "step": 95
+ },
+ {
+ "epoch": 1.3353711790393012,
+ "grad_norm": 0.01086883433163166,
+ "learning_rate": 0.00057,
+ "loss": 7.452998638153076,
+ "step": 96
+ },
+ {
+ "epoch": 1.3493449781659388,
+ "grad_norm": 0.014700951054692268,
+ "learning_rate": 0.0005759999999999999,
+ "loss": 7.410249710083008,
+ "step": 97
+ },
+ {
+ "epoch": 1.3633187772925766,
+ "grad_norm": 0.014512976631522179,
+ "learning_rate": 0.0005819999999999999,
+ "loss": 7.429132461547852,
+ "step": 98
+ },
+ {
+ "epoch": 1.3772925764192139,
+ "grad_norm": 0.012519920244812965,
+ "learning_rate": 0.000588,
+ "loss": 7.430817604064941,
+ "step": 99
+ },
+ {
+ "epoch": 1.3912663755458516,
+ "grad_norm": 0.007132664322853088,
+ "learning_rate": 0.0005939999999999999,
+ "loss": 7.405234336853027,
+ "step": 100
+ },
+ {
+ "epoch": 1.405240174672489,
+ "grad_norm": 0.019489388912916183,
+ "learning_rate": 0.0006,
+ "loss": 7.427791595458984,
+ "step": 101
+ },
+ {
+ "epoch": 1.4192139737991267,
+ "grad_norm": 0.027252664789557457,
+ "learning_rate": 0.0006,
+ "loss": 7.41761589050293,
+ "step": 102
+ },
+ {
+ "epoch": 1.4331877729257643,
+ "grad_norm": 0.02501637488603592,
+ "learning_rate": 0.0006,
+ "loss": 7.416735649108887,
+ "step": 103
+ },
+ {
+ "epoch": 1.4471615720524018,
+ "grad_norm": 0.00989772193133831,
+ "learning_rate": 0.0006,
+ "loss": 7.3756632804870605,
+ "step": 104
+ },
+ {
+ "epoch": 1.4611353711790394,
+ "grad_norm": 0.01622949168086052,
+ "learning_rate": 0.0006,
+ "loss": 7.400811672210693,
+ "step": 105
+ },
+ {
+ "epoch": 1.475109170305677,
+ "grad_norm": 0.012407763861119747,
+ "learning_rate": 0.0006,
+ "loss": 7.37675666809082,
+ "step": 106
+ },
+ {
+ "epoch": 1.4890829694323144,
+ "grad_norm": 0.010353565216064453,
+ "learning_rate": 0.0006,
+ "loss": 7.4013142585754395,
+ "step": 107
+ },
+ {
+ "epoch": 1.503056768558952,
+ "grad_norm": 0.0120553532615304,
+ "learning_rate": 0.0006,
+ "loss": 7.379873275756836,
+ "step": 108
+ },
+ {
+ "epoch": 1.5170305676855895,
+ "grad_norm": 0.01610225811600685,
+ "learning_rate": 0.0006,
+ "loss": 7.371631145477295,
+ "step": 109
+ },
+ {
+ "epoch": 1.531004366812227,
+ "grad_norm": 0.01159879844635725,
+ "learning_rate": 0.0006,
+ "loss": 7.389007568359375,
+ "step": 110
+ },
+ {
+ "epoch": 1.5449781659388646,
+ "grad_norm": 0.008013816550374031,
+ "learning_rate": 0.0006,
+ "loss": 7.395920753479004,
+ "step": 111
+ },
+ {
+ "epoch": 1.5589519650655022,
+ "grad_norm": 0.016334809362888336,
+ "learning_rate": 0.0006,
+ "loss": 7.393121719360352,
+ "step": 112
+ },
+ {
+ "epoch": 1.5729257641921397,
+ "grad_norm": 0.018821721896529198,
+ "learning_rate": 0.0006,
+ "loss": 7.367543697357178,
+ "step": 113
+ },
+ {
+ "epoch": 1.5868995633187772,
+ "grad_norm": 0.014088994823396206,
+ "learning_rate": 0.0006,
+ "loss": 7.385537624359131,
+ "step": 114
+ },
+ {
+ "epoch": 1.600873362445415,
+ "grad_norm": 0.005864050704985857,
+ "learning_rate": 0.0006,
+ "loss": 7.365346908569336,
+ "step": 115
+ },
+ {
+ "epoch": 1.6148471615720523,
+ "grad_norm": 0.011745430529117584,
+ "learning_rate": 0.0006,
+ "loss": 7.390227317810059,
+ "step": 116
+ },
+ {
+ "epoch": 1.62882096069869,
+ "grad_norm": 0.017205573618412018,
+ "learning_rate": 0.0006,
+ "loss": 7.3558549880981445,
+ "step": 117
+ },
+ {
+ "epoch": 1.6427947598253274,
+ "grad_norm": 0.021256346255540848,
+ "learning_rate": 0.0006,
+ "loss": 7.377346992492676,
+ "step": 118
+ },
+ {
+ "epoch": 1.6567685589519652,
+ "grad_norm": 0.019335204735398293,
+ "learning_rate": 0.0006,
+ "loss": 7.346194267272949,
+ "step": 119
+ },
+ {
+ "epoch": 1.6707423580786025,
+ "grad_norm": 0.008134003728628159,
+ "learning_rate": 0.0006,
+ "loss": 7.3063578605651855,
+ "step": 120
+ },
+ {
+ "epoch": 1.6847161572052403,
+ "grad_norm": 0.017613831907510757,
+ "learning_rate": 0.0006,
+ "loss": 7.338171005249023,
+ "step": 121
+ },
+ {
+ "epoch": 1.6986899563318776,
+ "grad_norm": 0.00890734139829874,
+ "learning_rate": 0.0006,
+ "loss": 7.355384826660156,
+ "step": 122
+ },
+ {
+ "epoch": 1.7126637554585153,
+ "grad_norm": 0.013448155485093594,
+ "learning_rate": 0.0006,
+ "loss": 7.344296455383301,
+ "step": 123
+ },
+ {
+ "epoch": 1.726637554585153,
+ "grad_norm": 0.009241633117198944,
+ "learning_rate": 0.0006,
+ "loss": 7.3454060554504395,
+ "step": 124
+ },
+ {
+ "epoch": 1.7406113537117904,
+ "grad_norm": 0.018561311066150665,
+ "learning_rate": 0.0006,
+ "loss": 7.32348108291626,
+ "step": 125
+ },
+ {
+ "epoch": 1.754585152838428,
+ "grad_norm": 0.025667704641819,
+ "learning_rate": 0.0006,
+ "loss": 7.356142520904541,
+ "step": 126
+ },
+ {
+ "epoch": 1.7685589519650655,
+ "grad_norm": 0.036232151091098785,
+ "learning_rate": 0.0006,
+ "loss": 7.363539695739746,
+ "step": 127
+ },
+ {
+ "epoch": 1.782532751091703,
+ "grad_norm": 0.059483714401721954,
+ "learning_rate": 0.0006,
+ "loss": 7.402331352233887,
+ "step": 128
+ },
+ {
+ "epoch": 1.7965065502183406,
+ "grad_norm": 0.05365009233355522,
+ "learning_rate": 0.0006,
+ "loss": 7.40457820892334,
+ "step": 129
+ },
+ {
+ "epoch": 1.8104803493449781,
+ "grad_norm": 0.02738415263593197,
+ "learning_rate": 0.0006,
+ "loss": 7.3756208419799805,
+ "step": 130
+ },
+ {
+ "epoch": 1.8244541484716157,
+ "grad_norm": 0.035017091780900955,
+ "learning_rate": 0.0006,
+ "loss": 7.379263877868652,
+ "step": 131
+ },
+ {
+ "epoch": 1.8384279475982532,
+ "grad_norm": 0.02819863148033619,
+ "learning_rate": 0.0006,
+ "loss": 7.327673435211182,
+ "step": 132
+ },
+ {
+ "epoch": 1.8524017467248908,
+ "grad_norm": 0.017754238098859787,
+ "learning_rate": 0.0006,
+ "loss": 7.333420276641846,
+ "step": 133
+ },
+ {
+ "epoch": 1.8663755458515285,
+ "grad_norm": 0.024685733020305634,
+ "learning_rate": 0.0006,
+ "loss": 7.34493350982666,
+ "step": 134
+ },
+ {
+ "epoch": 1.8803493449781659,
+ "grad_norm": 0.019426532089710236,
+ "learning_rate": 0.0006,
+ "loss": 7.328448295593262,
+ "step": 135
+ },
+ {
+ "epoch": 1.8943231441048036,
+ "grad_norm": 0.02106391079723835,
+ "learning_rate": 0.0006,
+ "loss": 7.327917098999023,
+ "step": 136
+ },
+ {
+ "epoch": 1.908296943231441,
+ "grad_norm": 0.02056879736483097,
+ "learning_rate": 0.0006,
+ "loss": 7.3311991691589355,
+ "step": 137
+ },
+ {
+ "epoch": 1.9222707423580787,
+ "grad_norm": 0.016865141689777374,
+ "learning_rate": 0.0006,
+ "loss": 7.312766075134277,
+ "step": 138
+ },
+ {
+ "epoch": 1.936244541484716,
+ "grad_norm": 0.017549855634570122,
+ "learning_rate": 0.0006,
+ "loss": 7.31191349029541,
+ "step": 139
+ },
+ {
+ "epoch": 1.9502183406113538,
+ "grad_norm": 0.015930157154798508,
+ "learning_rate": 0.0006,
+ "loss": 7.330901622772217,
+ "step": 140
+ },
+ {
+ "epoch": 1.9641921397379911,
+ "grad_norm": 0.013283558189868927,
+ "learning_rate": 0.0006,
+ "loss": 7.298318386077881,
+ "step": 141
+ },
+ {
+ "epoch": 1.9781659388646289,
+ "grad_norm": 0.012023469433188438,
+ "learning_rate": 0.0006,
+ "loss": 7.282337188720703,
+ "step": 142
+ },
+ {
+ "epoch": 1.9921397379912664,
+ "grad_norm": 0.01459511648863554,
+ "learning_rate": 0.0006,
+ "loss": 7.282924175262451,
+ "step": 143
+ },
+ {
+ "epoch": 2.0,
+ "grad_norm": 0.017474235966801643,
+ "learning_rate": 0.0006,
+ "loss": 7.254333972930908,
+ "step": 144
+ },
+ {
+ "epoch": 2.0,
+ "eval_loss": 7.318708896636963,
+ "eval_runtime": 59.6555,
+ "eval_samples_per_second": 40.935,
+ "eval_steps_per_second": 1.291,
+ "step": 144
+ },
+ {
+ "epoch": 2.0139737991266378,
+ "grad_norm": 0.019757507368922234,
+ "learning_rate": 0.0006,
+ "loss": 7.30288553237915,
+ "step": 145
+ },
+ {
+ "epoch": 2.027947598253275,
+ "grad_norm": 0.014571248553693295,
+ "learning_rate": 0.0006,
+ "loss": 7.274681568145752,
+ "step": 146
+ },
+ {
+ "epoch": 2.041921397379913,
+ "grad_norm": 0.015744198113679886,
+ "learning_rate": 0.0006,
+ "loss": 7.278632640838623,
+ "step": 147
+ },
+ {
+ "epoch": 2.05589519650655,
+ "grad_norm": 0.028048858046531677,
+ "learning_rate": 0.0006,
+ "loss": 7.26558780670166,
+ "step": 148
+ },
+ {
+ "epoch": 2.069868995633188,
+ "grad_norm": 0.06547307223081589,
+ "learning_rate": 0.0006,
+ "loss": 7.32567834854126,
+ "step": 149
+ },
+ {
+ "epoch": 2.0838427947598253,
+ "grad_norm": 0.05969763547182083,
+ "learning_rate": 0.0006,
+ "loss": 7.395836353302002,
+ "step": 150
+ },
+ {
+ "epoch": 2.097816593886463,
+ "grad_norm": 0.032574281096458435,
+ "learning_rate": 0.0006,
+ "loss": 7.318150997161865,
+ "step": 151
+ },
+ {
+ "epoch": 2.1117903930131003,
+ "grad_norm": 0.029452962800860405,
+ "learning_rate": 0.0006,
+ "loss": 7.308066368103027,
+ "step": 152
+ },
+ {
+ "epoch": 2.125764192139738,
+ "grad_norm": 0.03161991387605667,
+ "learning_rate": 0.0006,
+ "loss": 7.306290626525879,
+ "step": 153
+ },
+ {
+ "epoch": 2.1397379912663754,
+ "grad_norm": 0.027096295729279518,
+ "learning_rate": 0.0006,
+ "loss": 7.297796249389648,
+ "step": 154
+ },
+ {
+ "epoch": 2.153711790393013,
+ "grad_norm": 0.028657056391239166,
+ "learning_rate": 0.0006,
+ "loss": 7.317424774169922,
+ "step": 155
+ },
+ {
+ "epoch": 2.1676855895196505,
+ "grad_norm": 0.02046995982527733,
+ "learning_rate": 0.0006,
+ "loss": 7.297986030578613,
+ "step": 156
+ },
+ {
+ "epoch": 2.1816593886462883,
+ "grad_norm": 0.02220369316637516,
+ "learning_rate": 0.0006,
+ "loss": 7.268237590789795,
+ "step": 157
+ },
+ {
+ "epoch": 2.1956331877729256,
+ "grad_norm": 0.018381357192993164,
+ "learning_rate": 0.0006,
+ "loss": 7.259415626525879,
+ "step": 158
+ },
+ {
+ "epoch": 2.2096069868995634,
+ "grad_norm": 0.0204413291066885,
+ "learning_rate": 0.0006,
+ "loss": 7.258173942565918,
+ "step": 159
+ },
+ {
+ "epoch": 2.2235807860262007,
+ "grad_norm": 0.015874288976192474,
+ "learning_rate": 0.0006,
+ "loss": 7.281834602355957,
+ "step": 160
+ },
+ {
+ "epoch": 2.2375545851528384,
+ "grad_norm": 0.01634068600833416,
+ "learning_rate": 0.0006,
+ "loss": 7.2469868659973145,
+ "step": 161
+ },
+ {
+ "epoch": 2.251528384279476,
+ "grad_norm": 0.01728747971355915,
+ "learning_rate": 0.0006,
+ "loss": 7.280995845794678,
+ "step": 162
+ },
+ {
+ "epoch": 2.2655021834061135,
+ "grad_norm": 0.016341188922524452,
+ "learning_rate": 0.0006,
+ "loss": 7.273357391357422,
+ "step": 163
+ },
+ {
+ "epoch": 2.279475982532751,
+ "grad_norm": 0.016125807538628578,
+ "learning_rate": 0.0006,
+ "loss": 7.273510932922363,
+ "step": 164
+ },
+ {
+ "epoch": 2.2934497816593886,
+ "grad_norm": 0.01629287749528885,
+ "learning_rate": 0.0006,
+ "loss": 7.267029762268066,
+ "step": 165
+ },
+ {
+ "epoch": 2.3074235807860264,
+ "grad_norm": 0.015296131372451782,
+ "learning_rate": 0.0006,
+ "loss": 7.211989402770996,
+ "step": 166
+ },
+ {
+ "epoch": 2.3213973799126637,
+ "grad_norm": 0.01709570363163948,
+ "learning_rate": 0.0006,
+ "loss": 7.242851734161377,
+ "step": 167
+ },
+ {
+ "epoch": 2.3353711790393015,
+ "grad_norm": 0.012552388943731785,
+ "learning_rate": 0.0006,
+ "loss": 7.241696834564209,
+ "step": 168
+ },
+ {
+ "epoch": 2.349344978165939,
+ "grad_norm": 0.014571291394531727,
+ "learning_rate": 0.0006,
+ "loss": 7.270681381225586,
+ "step": 169
+ },
+ {
+ "epoch": 2.3633187772925766,
+ "grad_norm": 0.013389437459409237,
+ "learning_rate": 0.0006,
+ "loss": 7.191600799560547,
+ "step": 170
+ },
+ {
+ "epoch": 2.377292576419214,
+ "grad_norm": 0.011895314790308475,
+ "learning_rate": 0.0006,
+ "loss": 7.213610649108887,
+ "step": 171
+ },
+ {
+ "epoch": 2.3912663755458516,
+ "grad_norm": 0.01345642190426588,
+ "learning_rate": 0.0006,
+ "loss": 7.193787574768066,
+ "step": 172
+ },
+ {
+ "epoch": 2.405240174672489,
+ "grad_norm": 0.0091794328764081,
+ "learning_rate": 0.0006,
+ "loss": 7.233364105224609,
+ "step": 173
+ },
+ {
+ "epoch": 2.4192139737991267,
+ "grad_norm": 0.013767086900770664,
+ "learning_rate": 0.0006,
+ "loss": 7.217272758483887,
+ "step": 174
+ },
+ {
+ "epoch": 2.433187772925764,
+ "grad_norm": 0.015134960412979126,
+ "learning_rate": 0.0006,
+ "loss": 7.192791938781738,
+ "step": 175
+ },
+ {
+ "epoch": 2.447161572052402,
+ "grad_norm": 0.01790648326277733,
+ "learning_rate": 0.0006,
+ "loss": 7.229467391967773,
+ "step": 176
+ },
+ {
+ "epoch": 2.461135371179039,
+ "grad_norm": 0.015362377278506756,
+ "learning_rate": 0.0006,
+ "loss": 7.180346488952637,
+ "step": 177
+ },
+ {
+ "epoch": 2.475109170305677,
+ "grad_norm": 0.010505498386919498,
+ "learning_rate": 0.0006,
+ "loss": 7.188833236694336,
+ "step": 178
+ },
+ {
+ "epoch": 2.489082969432314,
+ "grad_norm": 0.009129747748374939,
+ "learning_rate": 0.0006,
+ "loss": 7.203153133392334,
+ "step": 179
+ },
+ {
+ "epoch": 2.503056768558952,
+ "grad_norm": 0.01147883478552103,
+ "learning_rate": 0.0006,
+ "loss": 7.163824558258057,
+ "step": 180
+ },
+ {
+ "epoch": 2.5170305676855893,
+ "grad_norm": 0.011338942684233189,
+ "learning_rate": 0.0006,
+ "loss": 7.16287899017334,
+ "step": 181
+ },
+ {
+ "epoch": 2.531004366812227,
+ "grad_norm": 0.013277675025165081,
+ "learning_rate": 0.0006,
+ "loss": 7.155905723571777,
+ "step": 182
+ },
+ {
+ "epoch": 2.544978165938865,
+ "grad_norm": 0.015269347466528416,
+ "learning_rate": 0.0006,
+ "loss": 7.189371109008789,
+ "step": 183
+ },
+ {
+ "epoch": 2.558951965065502,
+ "grad_norm": 0.017905596643686295,
+ "learning_rate": 0.0006,
+ "loss": 7.177547931671143,
+ "step": 184
+ },
+ {
+ "epoch": 2.5729257641921395,
+ "grad_norm": 0.024431610479950905,
+ "learning_rate": 0.0006,
+ "loss": 7.192169189453125,
+ "step": 185
+ },
+ {
+ "epoch": 2.5868995633187772,
+ "grad_norm": 0.03656961768865585,
+ "learning_rate": 0.0006,
+ "loss": 7.124387741088867,
+ "step": 186
+ },
+ {
+ "epoch": 2.600873362445415,
+ "grad_norm": 0.04337048903107643,
+ "learning_rate": 0.0006,
+ "loss": 7.140231609344482,
+ "step": 187
+ },
+ {
+ "epoch": 2.6148471615720523,
+ "grad_norm": 0.04047228768467903,
+ "learning_rate": 0.0006,
+ "loss": 7.171927452087402,
+ "step": 188
+ },
+ {
+ "epoch": 2.62882096069869,
+ "grad_norm": 0.07825770974159241,
+ "learning_rate": 0.0006,
+ "loss": 7.207631587982178,
+ "step": 189
+ },
+ {
+ "epoch": 2.6427947598253274,
+ "grad_norm": 0.07118063420057297,
+ "learning_rate": 0.0006,
+ "loss": 7.209488391876221,
+ "step": 190
+ },
+ {
+ "epoch": 2.656768558951965,
+ "grad_norm": 0.05153293535113335,
+ "learning_rate": 0.0006,
+ "loss": 7.14793062210083,
+ "step": 191
+ },
+ {
+ "epoch": 2.6707423580786025,
+ "grad_norm": 0.03318839520215988,
+ "learning_rate": 0.0006,
+ "loss": 7.163010597229004,
+ "step": 192
+ },
+ {
+ "epoch": 2.6847161572052403,
+ "grad_norm": 0.029333539307117462,
+ "learning_rate": 0.0006,
+ "loss": 7.159955978393555,
+ "step": 193
+ },
+ {
+ "epoch": 2.6986899563318776,
+ "grad_norm": 0.029405880719423294,
+ "learning_rate": 0.0006,
+ "loss": 7.136178493499756,
+ "step": 194
+ },
+ {
+ "epoch": 2.7126637554585153,
+ "grad_norm": 0.037815988063812256,
+ "learning_rate": 0.0006,
+ "loss": 7.183186054229736,
+ "step": 195
+ },
+ {
+ "epoch": 2.726637554585153,
+ "grad_norm": 0.020307250320911407,
+ "learning_rate": 0.0006,
+ "loss": 7.133164882659912,
+ "step": 196
+ },
+ {
+ "epoch": 2.7406113537117904,
+ "grad_norm": 0.026048069819808006,
+ "learning_rate": 0.0006,
+ "loss": 7.124977111816406,
+ "step": 197
+ },
+ {
+ "epoch": 2.7545851528384278,
+ "grad_norm": 0.027316724881529808,
+ "learning_rate": 0.0006,
+ "loss": 7.07265567779541,
+ "step": 198
+ },
+ {
+ "epoch": 2.7685589519650655,
+ "grad_norm": 0.021629009395837784,
+ "learning_rate": 0.0006,
+ "loss": 7.150020599365234,
+ "step": 199
+ },
+ {
+ "epoch": 2.7825327510917033,
+ "grad_norm": 0.01605929434299469,
+ "learning_rate": 0.0006,
+ "loss": 7.10919189453125,
+ "step": 200
+ },
+ {
+ "epoch": 2.7965065502183406,
+ "grad_norm": 0.020800089463591576,
+ "learning_rate": 0.0006,
+ "loss": 7.132045269012451,
+ "step": 201
+ },
+ {
+ "epoch": 2.810480349344978,
+ "grad_norm": 0.019225774332880974,
+ "learning_rate": 0.0006,
+ "loss": 7.128364562988281,
+ "step": 202
+ },
+ {
+ "epoch": 2.8244541484716157,
+ "grad_norm": 0.01561372634023428,
+ "learning_rate": 0.0006,
+ "loss": 7.083669662475586,
+ "step": 203
+ },
+ {
+ "epoch": 2.8384279475982535,
+ "grad_norm": 0.021969519555568695,
+ "learning_rate": 0.0006,
+ "loss": 7.050841331481934,
+ "step": 204
+ },
+ {
+ "epoch": 2.8524017467248908,
+ "grad_norm": 0.021453695371747017,
+ "learning_rate": 0.0006,
+ "loss": 7.097085475921631,
+ "step": 205
+ },
+ {
+ "epoch": 2.8663755458515285,
+ "grad_norm": 0.022836022078990936,
+ "learning_rate": 0.0006,
+ "loss": 7.10405969619751,
+ "step": 206
+ },
+ {
+ "epoch": 2.880349344978166,
+ "grad_norm": 0.02430851384997368,
+ "learning_rate": 0.0006,
+ "loss": 7.045181751251221,
+ "step": 207
+ },
+ {
+ "epoch": 2.8943231441048036,
+ "grad_norm": 0.020166993141174316,
+ "learning_rate": 0.0006,
+ "loss": 7.068183898925781,
+ "step": 208
+ },
+ {
+ "epoch": 2.908296943231441,
+ "grad_norm": 0.0118505684658885,
+ "learning_rate": 0.0006,
+ "loss": 7.07132625579834,
+ "step": 209
+ },
+ {
+ "epoch": 2.9222707423580787,
+ "grad_norm": 0.023772427812218666,
+ "learning_rate": 0.0006,
+ "loss": 7.055238723754883,
+ "step": 210
+ },
+ {
+ "epoch": 2.936244541484716,
+ "grad_norm": 0.02350712940096855,
+ "learning_rate": 0.0006,
+ "loss": 7.055898666381836,
+ "step": 211
+ },
+ {
+ "epoch": 2.950218340611354,
+ "grad_norm": 0.017244907096028328,
+ "learning_rate": 0.0006,
+ "loss": 7.0099711418151855,
+ "step": 212
+ },
+ {
+ "epoch": 2.964192139737991,
+ "grad_norm": 0.021565575152635574,
+ "learning_rate": 0.0006,
+ "loss": 6.991751670837402,
+ "step": 213
+ },
+ {
+ "epoch": 2.978165938864629,
+ "grad_norm": 0.03241860866546631,
+ "learning_rate": 0.0006,
+ "loss": 7.005980491638184,
+ "step": 214
+ },
+ {
+ "epoch": 2.992139737991266,
+ "grad_norm": 0.04613726586103439,
+ "learning_rate": 0.0006,
+ "loss": 7.027859687805176,
+ "step": 215
+ },
+ {
+ "epoch": 3.0,
+ "grad_norm": 0.06404894590377808,
+ "learning_rate": 0.0006,
+ "loss": 7.123725891113281,
+ "step": 216
+ },
+ {
+ "epoch": 3.0,
+ "eval_loss": 7.070647716522217,
+ "eval_runtime": 58.5801,
+ "eval_samples_per_second": 41.687,
+ "eval_steps_per_second": 1.314,
+ "step": 216
+ },
+ {
+ "epoch": 3.0139737991266378,
+ "grad_norm": 0.05542432889342308,
+ "learning_rate": 0.0006,
+ "loss": 7.078299522399902,
+ "step": 217
+ },
+ {
+ "epoch": 3.027947598253275,
+ "grad_norm": 0.026484526693820953,
+ "learning_rate": 0.0006,
+ "loss": 6.961159706115723,
+ "step": 218
+ },
+ {
+ "epoch": 3.041921397379913,
+ "grad_norm": 0.028276391327381134,
+ "learning_rate": 0.0006,
+ "loss": 7.019961833953857,
+ "step": 219
+ },
+ {
+ "epoch": 3.05589519650655,
+ "grad_norm": 0.024486595764756203,
+ "learning_rate": 0.0006,
+ "loss": 7.039677619934082,
+ "step": 220
+ },
+ {
+ "epoch": 3.069868995633188,
+ "grad_norm": 0.017620893195271492,
+ "learning_rate": 0.0006,
+ "loss": 6.9270782470703125,
+ "step": 221
+ },
+ {
+ "epoch": 3.0838427947598253,
+ "grad_norm": 0.02547302283346653,
+ "learning_rate": 0.0006,
+ "loss": 6.924233436584473,
+ "step": 222
+ },
+ {
+ "epoch": 3.097816593886463,
+ "grad_norm": 0.02539009042084217,
+ "learning_rate": 0.0006,
+ "loss": 6.973979949951172,
+ "step": 223
+ },
+ {
+ "epoch": 3.1117903930131003,
+ "grad_norm": 0.03256227448582649,
+ "learning_rate": 0.0006,
+ "loss": 6.933725357055664,
+ "step": 224
+ },
+ {
+ "epoch": 3.125764192139738,
+ "grad_norm": 0.04651800915598869,
+ "learning_rate": 0.0006,
+ "loss": 6.951044082641602,
+ "step": 225
+ },
+ {
+ "epoch": 3.1397379912663754,
+ "grad_norm": 0.04822975769639015,
+ "learning_rate": 0.0006,
+ "loss": 6.9614410400390625,
+ "step": 226
+ },
+ {
+ "epoch": 3.153711790393013,
+ "grad_norm": 0.031165020540356636,
+ "learning_rate": 0.0006,
+ "loss": 6.91193151473999,
+ "step": 227
+ },
+ {
+ "epoch": 3.1676855895196505,
+ "grad_norm": 0.029754292219877243,
+ "learning_rate": 0.0006,
+ "loss": 6.91802978515625,
+ "step": 228
+ },
+ {
+ "epoch": 3.1816593886462883,
+ "grad_norm": 0.02193152718245983,
+ "learning_rate": 0.0006,
+ "loss": 6.928090572357178,
+ "step": 229
+ },
+ {
+ "epoch": 3.1956331877729256,
+ "grad_norm": 0.02428470179438591,
+ "learning_rate": 0.0006,
+ "loss": 6.90396785736084,
+ "step": 230
+ },
+ {
+ "epoch": 3.2096069868995634,
+ "grad_norm": 0.023374345153570175,
+ "learning_rate": 0.0006,
+ "loss": 6.885858535766602,
+ "step": 231
+ },
+ {
+ "epoch": 3.2235807860262007,
+ "grad_norm": 0.023405000567436218,
+ "learning_rate": 0.0006,
+ "loss": 6.8855767250061035,
+ "step": 232
+ },
+ {
+ "epoch": 3.2375545851528384,
+ "grad_norm": 0.017277134582400322,
+ "learning_rate": 0.0006,
+ "loss": 6.901449203491211,
+ "step": 233
+ },
+ {
+ "epoch": 3.251528384279476,
+ "grad_norm": 0.019496750086545944,
+ "learning_rate": 0.0006,
+ "loss": 6.885687828063965,
+ "step": 234
+ },
+ {
+ "epoch": 3.2655021834061135,
+ "grad_norm": 0.016893276944756508,
+ "learning_rate": 0.0006,
+ "loss": 6.872369766235352,
+ "step": 235
+ },
+ {
+ "epoch": 3.279475982532751,
+ "grad_norm": 0.02058715932071209,
+ "learning_rate": 0.0006,
+ "loss": 6.823280334472656,
+ "step": 236
+ },
+ {
+ "epoch": 3.2934497816593886,
+ "grad_norm": 0.015530755743384361,
+ "learning_rate": 0.0006,
+ "loss": 6.860499858856201,
+ "step": 237
+ },
+ {
+ "epoch": 3.3074235807860264,
+ "grad_norm": 0.01934926211833954,
+ "learning_rate": 0.0006,
+ "loss": 6.817448616027832,
+ "step": 238
+ },
+ {
+ "epoch": 3.3213973799126637,
+ "grad_norm": 0.017413552850484848,
+ "learning_rate": 0.0006,
+ "loss": 6.881955146789551,
+ "step": 239
+ },
+ {
+ "epoch": 3.3353711790393015,
+ "grad_norm": 0.026478875428438187,
+ "learning_rate": 0.0006,
+ "loss": 6.827404499053955,
+ "step": 240
+ },
+ {
+ "epoch": 3.349344978165939,
+ "grad_norm": 0.032611701637506485,
+ "learning_rate": 0.0006,
+ "loss": 6.78285551071167,
+ "step": 241
+ },
+ {
+ "epoch": 3.3633187772925766,
+ "grad_norm": 0.041900306940078735,
+ "learning_rate": 0.0006,
+ "loss": 6.783053874969482,
+ "step": 242
+ },
+ {
+ "epoch": 3.377292576419214,
+ "grad_norm": 0.04149497672915459,
+ "learning_rate": 0.0006,
+ "loss": 6.745220184326172,
+ "step": 243
+ },
+ {
+ "epoch": 3.3912663755458516,
+ "grad_norm": 0.023949826136231422,
+ "learning_rate": 0.0006,
+ "loss": 6.841533184051514,
+ "step": 244
+ },
+ {
+ "epoch": 3.405240174672489,
+ "grad_norm": 0.028601842001080513,
+ "learning_rate": 0.0006,
+ "loss": 6.852852821350098,
+ "step": 245
+ },
+ {
+ "epoch": 3.4192139737991267,
+ "grad_norm": 0.027365155518054962,
+ "learning_rate": 0.0006,
+ "loss": 6.74376106262207,
+ "step": 246
+ },
+ {
+ "epoch": 3.433187772925764,
+ "grad_norm": 0.02856568619608879,
+ "learning_rate": 0.0006,
+ "loss": 6.74906063079834,
+ "step": 247
+ },
+ {
+ "epoch": 3.447161572052402,
+ "grad_norm": 0.019172416999936104,
+ "learning_rate": 0.0006,
+ "loss": 6.809206008911133,
+ "step": 248
+ },
+ {
+ "epoch": 3.461135371179039,
+ "grad_norm": 0.02161746844649315,
+ "learning_rate": 0.0006,
+ "loss": 6.789888381958008,
+ "step": 249
+ },
+ {
+ "epoch": 3.475109170305677,
+ "grad_norm": 0.020273666828870773,
+ "learning_rate": 0.0006,
+ "loss": 6.73512077331543,
+ "step": 250
+ },
+ {
+ "epoch": 3.489082969432314,
+ "grad_norm": 0.022329283878207207,
+ "learning_rate": 0.0006,
+ "loss": 6.826148509979248,
+ "step": 251
+ },
+ {
+ "epoch": 3.503056768558952,
+ "grad_norm": 0.02284284122288227,
+ "learning_rate": 0.0006,
+ "loss": 6.748521327972412,
+ "step": 252
+ },
+ {
+ "epoch": 3.5170305676855893,
+ "grad_norm": 0.026368247345089912,
+ "learning_rate": 0.0006,
+ "loss": 6.645879745483398,
+ "step": 253
+ },
+ {
+ "epoch": 3.531004366812227,
+ "grad_norm": 0.03627244383096695,
+ "learning_rate": 0.0006,
+ "loss": 6.666064739227295,
+ "step": 254
+ },
+ {
+ "epoch": 3.544978165938865,
+ "grad_norm": 0.04140935093164444,
+ "learning_rate": 0.0006,
+ "loss": 6.712275981903076,
+ "step": 255
+ },
+ {
+ "epoch": 3.558951965065502,
+ "grad_norm": 0.0390462800860405,
+ "learning_rate": 0.0006,
+ "loss": 6.665900230407715,
+ "step": 256
+ },
+ {
+ "epoch": 3.5729257641921395,
+ "grad_norm": 0.05178583785891533,
+ "learning_rate": 0.0006,
+ "loss": 6.753298759460449,
+ "step": 257
+ },
+ {
+ "epoch": 3.5868995633187772,
+ "grad_norm": 0.054328061640262604,
+ "learning_rate": 0.0006,
+ "loss": 6.732633113861084,
+ "step": 258
+ },
+ {
+ "epoch": 3.600873362445415,
+ "grad_norm": 0.04590460658073425,
+ "learning_rate": 0.0006,
+ "loss": 6.65250301361084,
+ "step": 259
+ },
+ {
+ "epoch": 3.6148471615720523,
+ "grad_norm": 0.044331666082143784,
+ "learning_rate": 0.0006,
+ "loss": 6.68654203414917,
+ "step": 260
+ },
+ {
+ "epoch": 3.62882096069869,
+ "grad_norm": 0.05308730527758598,
+ "learning_rate": 0.0006,
+ "loss": 6.642482757568359,
+ "step": 261
+ },
+ {
+ "epoch": 3.6427947598253274,
+ "grad_norm": 0.03999016433954239,
+ "learning_rate": 0.0006,
+ "loss": 6.625227451324463,
+ "step": 262
+ },
+ {
+ "epoch": 3.656768558951965,
+ "grad_norm": 0.04088086634874344,
+ "learning_rate": 0.0006,
+ "loss": 6.67623233795166,
+ "step": 263
+ },
+ {
+ "epoch": 3.6707423580786025,
+ "grad_norm": 0.03615636005997658,
+ "learning_rate": 0.0006,
+ "loss": 6.6869001388549805,
+ "step": 264
+ },
+ {
+ "epoch": 3.6847161572052403,
+ "grad_norm": 0.02817360684275627,
+ "learning_rate": 0.0006,
+ "loss": 6.674601078033447,
+ "step": 265
+ },
+ {
+ "epoch": 3.6986899563318776,
+ "grad_norm": 0.032586123794317245,
+ "learning_rate": 0.0006,
+ "loss": 6.612994194030762,
+ "step": 266
+ },
+ {
+ "epoch": 3.7126637554585153,
+ "grad_norm": 0.031117867678403854,
+ "learning_rate": 0.0006,
+ "loss": 6.6244354248046875,
+ "step": 267
+ },
+ {
+ "epoch": 3.726637554585153,
+ "grad_norm": 0.028330687433481216,
+ "learning_rate": 0.0006,
+ "loss": 6.669557094573975,
+ "step": 268
+ },
+ {
+ "epoch": 3.7406113537117904,
+ "grad_norm": 0.021489202976226807,
+ "learning_rate": 0.0006,
+ "loss": 6.632035732269287,
+ "step": 269
+ },
+ {
+ "epoch": 3.7545851528384278,
+ "grad_norm": 0.024269424378871918,
+ "learning_rate": 0.0006,
+ "loss": 6.660867691040039,
+ "step": 270
+ },
+ {
+ "epoch": 3.7685589519650655,
+ "grad_norm": 0.02237873338162899,
+ "learning_rate": 0.0006,
+ "loss": 6.602587699890137,
+ "step": 271
+ },
+ {
+ "epoch": 3.7825327510917033,
+ "grad_norm": 0.026503929868340492,
+ "learning_rate": 0.0006,
+ "loss": 6.599200248718262,
+ "step": 272
+ },
+ {
+ "epoch": 3.7965065502183406,
+ "grad_norm": 0.03687124326825142,
+ "learning_rate": 0.0006,
+ "loss": 6.567296504974365,
+ "step": 273
+ },
+ {
+ "epoch": 3.810480349344978,
+ "grad_norm": 0.04708952456712723,
+ "learning_rate": 0.0006,
+ "loss": 6.621265411376953,
+ "step": 274
+ },
+ {
+ "epoch": 3.8244541484716157,
+ "grad_norm": 0.041846614331007004,
+ "learning_rate": 0.0006,
+ "loss": 6.550536155700684,
+ "step": 275
+ },
+ {
+ "epoch": 3.8384279475982535,
+ "grad_norm": 0.031249675899744034,
+ "learning_rate": 0.0006,
+ "loss": 6.54278564453125,
+ "step": 276
+ },
+ {
+ "epoch": 3.8524017467248908,
+ "grad_norm": 0.023082256317138672,
+ "learning_rate": 0.0006,
+ "loss": 6.544787406921387,
+ "step": 277
+ },
+ {
+ "epoch": 3.8663755458515285,
+ "grad_norm": 0.020995570346713066,
+ "learning_rate": 0.0006,
+ "loss": 6.518143653869629,
+ "step": 278
+ },
+ {
+ "epoch": 3.880349344978166,
+ "grad_norm": 0.020839324221014977,
+ "learning_rate": 0.0006,
+ "loss": 6.546667098999023,
+ "step": 279
+ },
+ {
+ "epoch": 3.8943231441048036,
+ "grad_norm": 0.018377432599663734,
+ "learning_rate": 0.0006,
+ "loss": 6.478307723999023,
+ "step": 280
+ },
+ {
+ "epoch": 3.908296943231441,
+ "grad_norm": 0.021713724359869957,
+ "learning_rate": 0.0006,
+ "loss": 6.485574245452881,
+ "step": 281
+ },
+ {
+ "epoch": 3.9222707423580787,
+ "grad_norm": 0.01359301246702671,
+ "learning_rate": 0.0006,
+ "loss": 6.570268630981445,
+ "step": 282
+ },
+ {
+ "epoch": 3.936244541484716,
+ "grad_norm": 0.016233332455158234,
+ "learning_rate": 0.0006,
+ "loss": 6.57150936126709,
+ "step": 283
+ },
+ {
+ "epoch": 3.950218340611354,
+ "grad_norm": 0.020687608048319817,
+ "learning_rate": 0.0006,
+ "loss": 6.527956008911133,
+ "step": 284
+ },
+ {
+ "epoch": 3.964192139737991,
+ "grad_norm": 0.02032964862883091,
+ "learning_rate": 0.0006,
+ "loss": 6.468169689178467,
+ "step": 285
+ },
+ {
+ "epoch": 3.978165938864629,
+ "grad_norm": 0.016676288098096848,
+ "learning_rate": 0.0006,
+ "loss": 6.448418617248535,
+ "step": 286
+ },
+ {
+ "epoch": 3.992139737991266,
+ "grad_norm": 0.0168539360165596,
+ "learning_rate": 0.0006,
+ "loss": 6.461180686950684,
+ "step": 287
+ },
+ {
+ "epoch": 4.0,
+ "grad_norm": 0.02666233666241169,
+ "learning_rate": 0.0006,
+ "loss": 6.344893932342529,
+ "step": 288
+ },
+ {
+ "epoch": 4.0,
+ "eval_loss": 6.5384979248046875,
+ "eval_runtime": 58.814,
+ "eval_samples_per_second": 41.521,
+ "eval_steps_per_second": 1.309,
+ "step": 288
+ },
+ {
+ "epoch": 4.013973799126638,
+ "grad_norm": 0.04154708981513977,
+ "learning_rate": 0.0006,
+ "loss": 6.41855525970459,
+ "step": 289
+ },
+ {
+ "epoch": 4.0279475982532755,
+ "grad_norm": 0.04895668104290962,
+ "learning_rate": 0.0006,
+ "loss": 6.499302864074707,
+ "step": 290
+ },
+ {
+ "epoch": 4.041921397379912,
+ "grad_norm": 0.034464482218027115,
+ "learning_rate": 0.0006,
+ "loss": 6.438611030578613,
+ "step": 291
+ },
+ {
+ "epoch": 4.05589519650655,
+ "grad_norm": 0.06845773756504059,
+ "learning_rate": 0.0006,
+ "loss": 6.538361072540283,
+ "step": 292
+ },
+ {
+ "epoch": 4.069868995633188,
+ "grad_norm": 0.053768858313560486,
+ "learning_rate": 0.0006,
+ "loss": 6.527524948120117,
+ "step": 293
+ },
+ {
+ "epoch": 4.083842794759826,
+ "grad_norm": 0.039037931710481644,
+ "learning_rate": 0.0006,
+ "loss": 6.503847122192383,
+ "step": 294
+ },
+ {
+ "epoch": 4.097816593886463,
+ "grad_norm": 0.03962196782231331,
+ "learning_rate": 0.0006,
+ "loss": 6.475178241729736,
+ "step": 295
+ },
+ {
+ "epoch": 4.1117903930131,
+ "grad_norm": 0.049101535230875015,
+ "learning_rate": 0.0006,
+ "loss": 6.474587917327881,
+ "step": 296
+ },
+ {
+ "epoch": 4.125764192139738,
+ "grad_norm": 0.02830282226204872,
+ "learning_rate": 0.0006,
+ "loss": 6.45065975189209,
+ "step": 297
+ },
+ {
+ "epoch": 4.139737991266376,
+ "grad_norm": 0.023153482005000114,
+ "learning_rate": 0.0006,
+ "loss": 6.5218706130981445,
+ "step": 298
+ },
+ {
+ "epoch": 4.153711790393013,
+ "grad_norm": 0.025064002722501755,
+ "learning_rate": 0.0006,
+ "loss": 6.456475257873535,
+ "step": 299
+ },
+ {
+ "epoch": 4.1676855895196505,
+ "grad_norm": 0.02357092685997486,
+ "learning_rate": 0.0006,
+ "loss": 6.485952377319336,
+ "step": 300
+ },
+ {
+ "epoch": 4.181659388646288,
+ "grad_norm": 0.021832305938005447,
+ "learning_rate": 0.0006,
+ "loss": 6.425506591796875,
+ "step": 301
+ },
+ {
+ "epoch": 4.195633187772926,
+ "grad_norm": 0.0227900929749012,
+ "learning_rate": 0.0006,
+ "loss": 6.439155578613281,
+ "step": 302
+ },
+ {
+ "epoch": 4.209606986899563,
+ "grad_norm": 0.015961607918143272,
+ "learning_rate": 0.0006,
+ "loss": 6.460862636566162,
+ "step": 303
+ },
+ {
+ "epoch": 4.223580786026201,
+ "grad_norm": 0.017876390367746353,
+ "learning_rate": 0.0006,
+ "loss": 6.44564151763916,
+ "step": 304
+ },
+ {
+ "epoch": 4.2375545851528384,
+ "grad_norm": 0.013703204691410065,
+ "learning_rate": 0.0006,
+ "loss": 6.418414115905762,
+ "step": 305
+ },
+ {
+ "epoch": 4.251528384279476,
+ "grad_norm": 0.015240306034684181,
+ "learning_rate": 0.0006,
+ "loss": 6.411238670349121,
+ "step": 306
+ },
+ {
+ "epoch": 4.265502183406113,
+ "grad_norm": 0.01587662845849991,
+ "learning_rate": 0.0006,
+ "loss": 6.395642280578613,
+ "step": 307
+ },
+ {
+ "epoch": 4.279475982532751,
+ "grad_norm": 0.014650700613856316,
+ "learning_rate": 0.0006,
+ "loss": 6.302793025970459,
+ "step": 308
+ },
+ {
+ "epoch": 4.293449781659389,
+ "grad_norm": 0.016533225774765015,
+ "learning_rate": 0.0006,
+ "loss": 6.380588054656982,
+ "step": 309
+ },
+ {
+ "epoch": 4.307423580786026,
+ "grad_norm": 0.019963741302490234,
+ "learning_rate": 0.0006,
+ "loss": 6.345336437225342,
+ "step": 310
+ },
+ {
+ "epoch": 4.321397379912664,
+ "grad_norm": 0.02237936295568943,
+ "learning_rate": 0.0006,
+ "loss": 6.387770652770996,
+ "step": 311
+ },
+ {
+ "epoch": 4.335371179039301,
+ "grad_norm": 0.026658328250050545,
+ "learning_rate": 0.0006,
+ "loss": 6.323662757873535,
+ "step": 312
+ },
+ {
+ "epoch": 4.349344978165939,
+ "grad_norm": 0.02852284163236618,
+ "learning_rate": 0.0006,
+ "loss": 6.382204055786133,
+ "step": 313
+ },
+ {
+ "epoch": 4.3633187772925766,
+ "grad_norm": 0.028815090656280518,
+ "learning_rate": 0.0006,
+ "loss": 6.402390480041504,
+ "step": 314
+ },
+ {
+ "epoch": 4.377292576419214,
+ "grad_norm": 0.029393529519438744,
+ "learning_rate": 0.0006,
+ "loss": 6.280278205871582,
+ "step": 315
+ },
+ {
+ "epoch": 4.391266375545851,
+ "grad_norm": 0.02186041697859764,
+ "learning_rate": 0.0006,
+ "loss": 6.302707672119141,
+ "step": 316
+ },
+ {
+ "epoch": 4.405240174672489,
+ "grad_norm": 0.02071218006312847,
+ "learning_rate": 0.0006,
+ "loss": 6.3365020751953125,
+ "step": 317
+ },
+ {
+ "epoch": 4.419213973799127,
+ "grad_norm": 0.02779117226600647,
+ "learning_rate": 0.0006,
+ "loss": 6.252157211303711,
+ "step": 318
+ },
+ {
+ "epoch": 4.4331877729257645,
+ "grad_norm": 0.03470654785633087,
+ "learning_rate": 0.0006,
+ "loss": 6.319070339202881,
+ "step": 319
+ },
+ {
+ "epoch": 4.447161572052401,
+ "grad_norm": 0.037067804485559464,
+ "learning_rate": 0.0006,
+ "loss": 6.314116477966309,
+ "step": 320
+ },
+ {
+ "epoch": 4.461135371179039,
+ "grad_norm": 0.032182756811380386,
+ "learning_rate": 0.0006,
+ "loss": 6.267298698425293,
+ "step": 321
+ },
+ {
+ "epoch": 4.475109170305677,
+ "grad_norm": 0.026305217295885086,
+ "learning_rate": 0.0006,
+ "loss": 6.331688404083252,
+ "step": 322
+ },
+ {
+ "epoch": 4.489082969432315,
+ "grad_norm": 0.027228357270359993,
+ "learning_rate": 0.0006,
+ "loss": 6.31139612197876,
+ "step": 323
+ },
+ {
+ "epoch": 4.503056768558952,
+ "grad_norm": 0.025781169533729553,
+ "learning_rate": 0.0006,
+ "loss": 6.194684982299805,
+ "step": 324
+ },
+ {
+ "epoch": 4.517030567685589,
+ "grad_norm": 0.029186977073550224,
+ "learning_rate": 0.0006,
+ "loss": 6.31306266784668,
+ "step": 325
+ },
+ {
+ "epoch": 4.531004366812227,
+ "grad_norm": 0.023608777672052383,
+ "learning_rate": 0.0006,
+ "loss": 6.285243034362793,
+ "step": 326
+ },
+ {
+ "epoch": 4.544978165938865,
+ "grad_norm": 0.02267594076693058,
+ "learning_rate": 0.0006,
+ "loss": 6.30342960357666,
+ "step": 327
+ },
+ {
+ "epoch": 4.558951965065502,
+ "grad_norm": 0.01926310732960701,
+ "learning_rate": 0.0006,
+ "loss": 6.280606269836426,
+ "step": 328
+ },
+ {
+ "epoch": 4.5729257641921395,
+ "grad_norm": 0.026484837755560875,
+ "learning_rate": 0.0006,
+ "loss": 6.237980365753174,
+ "step": 329
+ },
+ {
+ "epoch": 4.586899563318777,
+ "grad_norm": 0.026228854432702065,
+ "learning_rate": 0.0006,
+ "loss": 6.3018293380737305,
+ "step": 330
+ },
+ {
+ "epoch": 4.600873362445415,
+ "grad_norm": 0.022096967324614525,
+ "learning_rate": 0.0006,
+ "loss": 6.253863334655762,
+ "step": 331
+ },
+ {
+ "epoch": 4.614847161572053,
+ "grad_norm": 0.027223890647292137,
+ "learning_rate": 0.0006,
+ "loss": 6.166882038116455,
+ "step": 332
+ },
+ {
+ "epoch": 4.62882096069869,
+ "grad_norm": 0.03160124272108078,
+ "learning_rate": 0.0006,
+ "loss": 6.230309963226318,
+ "step": 333
+ },
+ {
+ "epoch": 4.642794759825327,
+ "grad_norm": 0.04110539332032204,
+ "learning_rate": 0.0006,
+ "loss": 6.2334818840026855,
+ "step": 334
+ },
+ {
+ "epoch": 4.656768558951965,
+ "grad_norm": 0.04355933889746666,
+ "learning_rate": 0.0006,
+ "loss": 6.160956382751465,
+ "step": 335
+ },
+ {
+ "epoch": 4.670742358078603,
+ "grad_norm": 0.03611086308956146,
+ "learning_rate": 0.0006,
+ "loss": 6.23399543762207,
+ "step": 336
+ },
+ {
+ "epoch": 4.68471615720524,
+ "grad_norm": 0.051811400800943375,
+ "learning_rate": 0.0006,
+ "loss": 6.241855144500732,
+ "step": 337
+ },
+ {
+ "epoch": 4.698689956331878,
+ "grad_norm": 0.057434193789958954,
+ "learning_rate": 0.0006,
+ "loss": 6.220993995666504,
+ "step": 338
+ },
+ {
+ "epoch": 4.712663755458515,
+ "grad_norm": 0.04420953616499901,
+ "learning_rate": 0.0006,
+ "loss": 6.300461292266846,
+ "step": 339
+ },
+ {
+ "epoch": 4.726637554585153,
+ "grad_norm": 0.03515457361936569,
+ "learning_rate": 0.0006,
+ "loss": 6.269349098205566,
+ "step": 340
+ },
+ {
+ "epoch": 4.74061135371179,
+ "grad_norm": 0.03222600743174553,
+ "learning_rate": 0.0006,
+ "loss": 6.248613357543945,
+ "step": 341
+ },
+ {
+ "epoch": 4.754585152838428,
+ "grad_norm": 0.0384046845138073,
+ "learning_rate": 0.0006,
+ "loss": 6.266979217529297,
+ "step": 342
+ },
+ {
+ "epoch": 4.7685589519650655,
+ "grad_norm": 0.0514867827296257,
+ "learning_rate": 0.0006,
+ "loss": 6.234103202819824,
+ "step": 343
+ },
+ {
+ "epoch": 4.782532751091703,
+ "grad_norm": 0.050747793167829514,
+ "learning_rate": 0.0006,
+ "loss": 6.322582244873047,
+ "step": 344
+ },
+ {
+ "epoch": 4.796506550218341,
+ "grad_norm": 0.040878988802433014,
+ "learning_rate": 0.0006,
+ "loss": 6.212509632110596,
+ "step": 345
+ },
+ {
+ "epoch": 4.810480349344978,
+ "grad_norm": 0.034655820578336716,
+ "learning_rate": 0.0006,
+ "loss": 6.265879154205322,
+ "step": 346
+ },
+ {
+ "epoch": 4.824454148471616,
+ "grad_norm": 0.04202224686741829,
+ "learning_rate": 0.0006,
+ "loss": 6.2896409034729,
+ "step": 347
+ },
+ {
+ "epoch": 4.8384279475982535,
+ "grad_norm": 0.03632277995347977,
+ "learning_rate": 0.0006,
+ "loss": 6.253917217254639,
+ "step": 348
+ },
+ {
+ "epoch": 4.85240174672489,
+ "grad_norm": 0.028645765036344528,
+ "learning_rate": 0.0006,
+ "loss": 6.207965850830078,
+ "step": 349
+ },
+ {
+ "epoch": 4.866375545851528,
+ "grad_norm": 0.025305170565843582,
+ "learning_rate": 0.0006,
+ "loss": 6.168068885803223,
+ "step": 350
+ },
+ {
+ "epoch": 4.880349344978166,
+ "grad_norm": 0.022137803956866264,
+ "learning_rate": 0.0006,
+ "loss": 6.235283374786377,
+ "step": 351
+ },
+ {
+ "epoch": 4.894323144104804,
+ "grad_norm": 0.01814538985490799,
+ "learning_rate": 0.0006,
+ "loss": 6.259641647338867,
+ "step": 352
+ },
+ {
+ "epoch": 4.908296943231441,
+ "grad_norm": 0.018465086817741394,
+ "learning_rate": 0.0006,
+ "loss": 6.111738681793213,
+ "step": 353
+ },
+ {
+ "epoch": 4.922270742358078,
+ "grad_norm": 0.018638933077454567,
+ "learning_rate": 0.0006,
+ "loss": 6.225642204284668,
+ "step": 354
+ },
+ {
+ "epoch": 4.936244541484716,
+ "grad_norm": 0.016461260616779327,
+ "learning_rate": 0.0006,
+ "loss": 6.140647888183594,
+ "step": 355
+ },
+ {
+ "epoch": 4.950218340611354,
+ "grad_norm": 0.015458385460078716,
+ "learning_rate": 0.0006,
+ "loss": 6.155224800109863,
+ "step": 356
+ },
+ {
+ "epoch": 4.964192139737992,
+ "grad_norm": 0.014934048056602478,
+ "learning_rate": 0.0006,
+ "loss": 6.101555824279785,
+ "step": 357
+ },
+ {
+ "epoch": 4.978165938864628,
+ "grad_norm": 0.013979545794427395,
+ "learning_rate": 0.0006,
+ "loss": 6.161718845367432,
+ "step": 358
+ },
+ {
+ "epoch": 4.992139737991266,
+ "grad_norm": 0.01640394888818264,
+ "learning_rate": 0.0006,
+ "loss": 6.055768966674805,
+ "step": 359
+ },
+ {
+ "epoch": 5.0,
+ "grad_norm": 0.015135680325329304,
+ "learning_rate": 0.0006,
+ "loss": 6.169695854187012,
+ "step": 360
+ },
+ {
+ "epoch": 5.0,
+ "eval_loss": 6.16408109664917,
+ "eval_runtime": 59.3649,
+ "eval_samples_per_second": 41.135,
+ "eval_steps_per_second": 1.297,
+ "step": 360
+ },
+ {
+ "epoch": 5.013973799126638,
+ "grad_norm": 0.012849048711359501,
+ "learning_rate": 0.0006,
+ "loss": 6.055830001831055,
+ "step": 361
+ },
+ {
+ "epoch": 5.0279475982532755,
+ "grad_norm": 0.018169576302170753,
+ "learning_rate": 0.0006,
+ "loss": 6.082655429840088,
+ "step": 362
+ },
+ {
+ "epoch": 5.041921397379912,
+ "grad_norm": 0.024293430149555206,
+ "learning_rate": 0.0006,
+ "loss": 6.109032154083252,
+ "step": 363
+ },
+ {
+ "epoch": 5.05589519650655,
+ "grad_norm": 0.03266787901520729,
+ "learning_rate": 0.0006,
+ "loss": 6.030613899230957,
+ "step": 364
+ },
+ {
+ "epoch": 5.069868995633188,
+ "grad_norm": 0.03572266176342964,
+ "learning_rate": 0.0006,
+ "loss": 6.050390720367432,
+ "step": 365
+ },
+ {
+ "epoch": 5.083842794759826,
+ "grad_norm": 0.032625213265419006,
+ "learning_rate": 0.0006,
+ "loss": 6.169313430786133,
+ "step": 366
+ },
+ {
+ "epoch": 5.097816593886463,
+ "grad_norm": 0.04051872715353966,
+ "learning_rate": 0.0006,
+ "loss": 6.10398006439209,
+ "step": 367
+ },
+ {
+ "epoch": 5.1117903930131,
+ "grad_norm": 0.0351213738322258,
+ "learning_rate": 0.0006,
+ "loss": 6.121973514556885,
+ "step": 368
+ },
+ {
+ "epoch": 5.125764192139738,
+ "grad_norm": 0.03523759916424751,
+ "learning_rate": 0.0006,
+ "loss": 6.0626654624938965,
+ "step": 369
+ },
+ {
+ "epoch": 5.139737991266376,
+ "grad_norm": 0.02706182189285755,
+ "learning_rate": 0.0006,
+ "loss": 6.049354076385498,
+ "step": 370
+ },
+ {
+ "epoch": 5.153711790393013,
+ "grad_norm": 0.024212589487433434,
+ "learning_rate": 0.0006,
+ "loss": 6.034826278686523,
+ "step": 371
+ },
+ {
+ "epoch": 5.1676855895196505,
+ "grad_norm": 0.024798082187771797,
+ "learning_rate": 0.0006,
+ "loss": 6.004058837890625,
+ "step": 372
+ },
+ {
+ "epoch": 5.181659388646288,
+ "grad_norm": 0.02325567416846752,
+ "learning_rate": 0.0006,
+ "loss": 5.986461639404297,
+ "step": 373
+ },
+ {
+ "epoch": 5.195633187772926,
+ "grad_norm": 0.019060570746660233,
+ "learning_rate": 0.0006,
+ "loss": 6.059736251831055,
+ "step": 374
+ },
+ {
+ "epoch": 5.209606986899563,
+ "grad_norm": 0.016822976991534233,
+ "learning_rate": 0.0006,
+ "loss": 6.072957515716553,
+ "step": 375
+ },
+ {
+ "epoch": 5.223580786026201,
+ "grad_norm": 0.017218658700585365,
+ "learning_rate": 0.0006,
+ "loss": 6.031739234924316,
+ "step": 376
+ },
+ {
+ "epoch": 5.2375545851528384,
+ "grad_norm": 0.014889383688569069,
+ "learning_rate": 0.0006,
+ "loss": 6.025674343109131,
+ "step": 377
+ },
+ {
+ "epoch": 5.251528384279476,
+ "grad_norm": 0.01708107627928257,
+ "learning_rate": 0.0006,
+ "loss": 6.090451240539551,
+ "step": 378
+ },
+ {
+ "epoch": 5.265502183406113,
+ "grad_norm": 0.020260317251086235,
+ "learning_rate": 0.0006,
+ "loss": 6.094257354736328,
+ "step": 379
+ },
+ {
+ "epoch": 5.279475982532751,
+ "grad_norm": 0.020110400393605232,
+ "learning_rate": 0.0006,
+ "loss": 6.021188735961914,
+ "step": 380
+ },
+ {
+ "epoch": 5.293449781659389,
+ "grad_norm": 0.02027830481529236,
+ "learning_rate": 0.0006,
+ "loss": 6.022156238555908,
+ "step": 381
+ },
+ {
+ "epoch": 5.307423580786026,
+ "grad_norm": 0.022747062146663666,
+ "learning_rate": 0.0006,
+ "loss": 6.011836051940918,
+ "step": 382
+ },
+ {
+ "epoch": 5.321397379912664,
+ "grad_norm": 0.01990230567753315,
+ "learning_rate": 0.0006,
+ "loss": 6.003054618835449,
+ "step": 383
+ },
+ {
+ "epoch": 5.335371179039301,
+ "grad_norm": 0.01596238650381565,
+ "learning_rate": 0.0006,
+ "loss": 5.991410732269287,
+ "step": 384
+ },
+ {
+ "epoch": 5.349344978165939,
+ "grad_norm": 0.015847105532884598,
+ "learning_rate": 0.0006,
+ "loss": 6.100622177124023,
+ "step": 385
+ },
+ {
+ "epoch": 5.3633187772925766,
+ "grad_norm": 0.016179397702217102,
+ "learning_rate": 0.0006,
+ "loss": 6.032659530639648,
+ "step": 386
+ },
+ {
+ "epoch": 5.377292576419214,
+ "grad_norm": 0.018256602808833122,
+ "learning_rate": 0.0006,
+ "loss": 5.961983680725098,
+ "step": 387
+ },
+ {
+ "epoch": 5.391266375545851,
+ "grad_norm": 0.02805912122130394,
+ "learning_rate": 0.0006,
+ "loss": 5.983541965484619,
+ "step": 388
+ },
+ {
+ "epoch": 5.405240174672489,
+ "grad_norm": 0.039082255214452744,
+ "learning_rate": 0.0006,
+ "loss": 6.021474361419678,
+ "step": 389
+ },
+ {
+ "epoch": 5.419213973799127,
+ "grad_norm": 0.036757659167051315,
+ "learning_rate": 0.0006,
+ "loss": 6.005046367645264,
+ "step": 390
+ },
+ {
+ "epoch": 5.4331877729257645,
+ "grad_norm": 0.035277366638183594,
+ "learning_rate": 0.0006,
+ "loss": 6.036296844482422,
+ "step": 391
+ },
+ {
+ "epoch": 5.447161572052401,
+ "grad_norm": 0.034404635429382324,
+ "learning_rate": 0.0006,
+ "loss": 5.994539260864258,
+ "step": 392
+ },
+ {
+ "epoch": 5.461135371179039,
+ "grad_norm": 0.0377206988632679,
+ "learning_rate": 0.0006,
+ "loss": 5.941534519195557,
+ "step": 393
+ },
+ {
+ "epoch": 5.475109170305677,
+ "grad_norm": 0.0389922633767128,
+ "learning_rate": 0.0006,
+ "loss": 5.983644485473633,
+ "step": 394
+ },
+ {
+ "epoch": 5.489082969432315,
+ "grad_norm": 0.04176778718829155,
+ "learning_rate": 0.0006,
+ "loss": 6.030922889709473,
+ "step": 395
+ },
+ {
+ "epoch": 5.503056768558952,
+ "grad_norm": 0.0337153784930706,
+ "learning_rate": 0.0006,
+ "loss": 5.961367130279541,
+ "step": 396
+ },
+ {
+ "epoch": 5.517030567685589,
+ "grad_norm": 0.036116816103458405,
+ "learning_rate": 0.0006,
+ "loss": 5.819280624389648,
+ "step": 397
+ },
+ {
+ "epoch": 5.531004366812227,
+ "grad_norm": 0.030724041163921356,
+ "learning_rate": 0.0006,
+ "loss": 5.906380653381348,
+ "step": 398
+ },
+ {
+ "epoch": 5.544978165938865,
+ "grad_norm": 0.030264202505350113,
+ "learning_rate": 0.0006,
+ "loss": 5.975833415985107,
+ "step": 399
+ },
+ {
+ "epoch": 5.558951965065502,
+ "grad_norm": 0.032096318900585175,
+ "learning_rate": 0.0006,
+ "loss": 6.001348972320557,
+ "step": 400
+ },
+ {
+ "epoch": 5.5729257641921395,
+ "grad_norm": 0.030579356476664543,
+ "learning_rate": 0.0006,
+ "loss": 5.928768634796143,
+ "step": 401
+ },
+ {
+ "epoch": 5.586899563318777,
+ "grad_norm": 0.028241973370313644,
+ "learning_rate": 0.0006,
+ "loss": 5.92582893371582,
+ "step": 402
+ },
+ {
+ "epoch": 5.600873362445415,
+ "grad_norm": 0.023086953908205032,
+ "learning_rate": 0.0006,
+ "loss": 5.892926216125488,
+ "step": 403
+ },
+ {
+ "epoch": 5.614847161572053,
+ "grad_norm": 0.02547992393374443,
+ "learning_rate": 0.0006,
+ "loss": 5.903195381164551,
+ "step": 404
+ },
+ {
+ "epoch": 5.62882096069869,
+ "grad_norm": 0.023089831694960594,
+ "learning_rate": 0.0006,
+ "loss": 5.912033557891846,
+ "step": 405
+ },
+ {
+ "epoch": 5.642794759825327,
+ "grad_norm": 0.025446368381381035,
+ "learning_rate": 0.0006,
+ "loss": 5.926138401031494,
+ "step": 406
+ },
+ {
+ "epoch": 5.656768558951965,
+ "grad_norm": 0.03748747706413269,
+ "learning_rate": 0.0006,
+ "loss": 5.913451194763184,
+ "step": 407
+ },
+ {
+ "epoch": 5.670742358078603,
+ "grad_norm": 0.03493810072541237,
+ "learning_rate": 0.0006,
+ "loss": 5.97633695602417,
+ "step": 408
+ },
+ {
+ "epoch": 5.68471615720524,
+ "grad_norm": 0.023131275549530983,
+ "learning_rate": 0.0006,
+ "loss": 5.899571418762207,
+ "step": 409
+ },
+ {
+ "epoch": 5.698689956331878,
+ "grad_norm": 0.02986014075577259,
+ "learning_rate": 0.0006,
+ "loss": 5.894903182983398,
+ "step": 410
+ },
+ {
+ "epoch": 5.712663755458515,
+ "grad_norm": 0.030171144753694534,
+ "learning_rate": 0.0006,
+ "loss": 5.876595497131348,
+ "step": 411
+ },
+ {
+ "epoch": 5.726637554585153,
+ "grad_norm": 0.029377546161413193,
+ "learning_rate": 0.0006,
+ "loss": 5.911637306213379,
+ "step": 412
+ },
+ {
+ "epoch": 5.74061135371179,
+ "grad_norm": 0.029829490929841995,
+ "learning_rate": 0.0006,
+ "loss": 5.819629669189453,
+ "step": 413
+ },
+ {
+ "epoch": 5.754585152838428,
+ "grad_norm": 0.024139299988746643,
+ "learning_rate": 0.0006,
+ "loss": 5.81728458404541,
+ "step": 414
+ },
+ {
+ "epoch": 5.7685589519650655,
+ "grad_norm": 0.021453432738780975,
+ "learning_rate": 0.0006,
+ "loss": 5.987326145172119,
+ "step": 415
+ },
+ {
+ "epoch": 5.782532751091703,
+ "grad_norm": 0.017942246049642563,
+ "learning_rate": 0.0006,
+ "loss": 5.893008232116699,
+ "step": 416
+ },
+ {
+ "epoch": 5.796506550218341,
+ "grad_norm": 0.019724637269973755,
+ "learning_rate": 0.0006,
+ "loss": 5.912441253662109,
+ "step": 417
+ },
+ {
+ "epoch": 5.810480349344978,
+ "grad_norm": 0.018597912043333054,
+ "learning_rate": 0.0006,
+ "loss": 5.860394477844238,
+ "step": 418
+ },
+ {
+ "epoch": 5.824454148471616,
+ "grad_norm": 0.016232412308454514,
+ "learning_rate": 0.0006,
+ "loss": 5.94991397857666,
+ "step": 419
+ },
+ {
+ "epoch": 5.8384279475982535,
+ "grad_norm": 0.01667204685509205,
+ "learning_rate": 0.0006,
+ "loss": 5.8506364822387695,
+ "step": 420
+ },
+ {
+ "epoch": 5.85240174672489,
+ "grad_norm": 0.014232151210308075,
+ "learning_rate": 0.0006,
+ "loss": 5.928424835205078,
+ "step": 421
+ },
+ {
+ "epoch": 5.866375545851528,
+ "grad_norm": 0.01570476032793522,
+ "learning_rate": 0.0006,
+ "loss": 5.953692436218262,
+ "step": 422
+ },
+ {
+ "epoch": 5.880349344978166,
+ "grad_norm": 0.01586179807782173,
+ "learning_rate": 0.0006,
+ "loss": 5.820014476776123,
+ "step": 423
+ },
+ {
+ "epoch": 5.894323144104804,
+ "grad_norm": 0.014729145914316177,
+ "learning_rate": 0.0006,
+ "loss": 5.860154628753662,
+ "step": 424
+ },
+ {
+ "epoch": 5.908296943231441,
+ "grad_norm": 0.013111459091305733,
+ "learning_rate": 0.0006,
+ "loss": 5.85988187789917,
+ "step": 425
+ },
+ {
+ "epoch": 5.922270742358078,
+ "grad_norm": 0.012937922962009907,
+ "learning_rate": 0.0006,
+ "loss": 5.756265640258789,
+ "step": 426
+ },
+ {
+ "epoch": 5.936244541484716,
+ "grad_norm": 0.01377453189343214,
+ "learning_rate": 0.0006,
+ "loss": 5.8333048820495605,
+ "step": 427
+ },
+ {
+ "epoch": 5.950218340611354,
+ "grad_norm": 0.014054795727133751,
+ "learning_rate": 0.0006,
+ "loss": 5.803333759307861,
+ "step": 428
+ },
+ {
+ "epoch": 5.964192139737992,
+ "grad_norm": 0.01674959622323513,
+ "learning_rate": 0.0006,
+ "loss": 5.8718485832214355,
+ "step": 429
+ },
+ {
+ "epoch": 5.978165938864628,
+ "grad_norm": 0.018733017146587372,
+ "learning_rate": 0.0006,
+ "loss": 5.875979900360107,
+ "step": 430
+ },
+ {
+ "epoch": 5.992139737991266,
+ "grad_norm": 0.02088983915746212,
+ "learning_rate": 0.0006,
+ "loss": 5.763634204864502,
+ "step": 431
+ },
+ {
+ "epoch": 6.0,
+ "grad_norm": 0.026705985888838768,
+ "learning_rate": 0.0006,
+ "loss": 5.6922478675842285,
+ "step": 432
+ },
+ {
+ "epoch": 6.0,
+ "eval_loss": 5.847958087921143,
+ "eval_runtime": 58.5451,
+ "eval_samples_per_second": 41.711,
+ "eval_steps_per_second": 1.315,
+ "step": 432
+ },
+ {
+ "epoch": 6.013973799126638,
+ "grad_norm": 0.03181544691324234,
+ "learning_rate": 0.0006,
+ "loss": 5.786368370056152,
+ "step": 433
+ },
+ {
+ "epoch": 6.0279475982532755,
+ "grad_norm": 0.03238112851977348,
+ "learning_rate": 0.0006,
+ "loss": 5.810310363769531,
+ "step": 434
+ },
+ {
+ "epoch": 6.041921397379912,
+ "grad_norm": 0.03916756808757782,
+ "learning_rate": 0.0006,
+ "loss": 5.7761993408203125,
+ "step": 435
+ },
+ {
+ "epoch": 6.05589519650655,
+ "grad_norm": 0.044009730219841,
+ "learning_rate": 0.0006,
+ "loss": 5.898112773895264,
+ "step": 436
+ },
+ {
+ "epoch": 6.069868995633188,
+ "grad_norm": 0.045235246419906616,
+ "learning_rate": 0.0006,
+ "loss": 5.876371383666992,
+ "step": 437
+ },
+ {
+ "epoch": 6.083842794759826,
+ "grad_norm": 0.050218384712934494,
+ "learning_rate": 0.0006,
+ "loss": 5.805103302001953,
+ "step": 438
+ },
+ {
+ "epoch": 6.097816593886463,
+ "grad_norm": 0.05444490164518356,
+ "learning_rate": 0.0006,
+ "loss": 5.826424598693848,
+ "step": 439
+ },
+ {
+ "epoch": 6.1117903930131,
+ "grad_norm": 0.04890982061624527,
+ "learning_rate": 0.0006,
+ "loss": 5.7735395431518555,
+ "step": 440
+ },
+ {
+ "epoch": 6.125764192139738,
+ "grad_norm": 0.05478639155626297,
+ "learning_rate": 0.0006,
+ "loss": 5.823677062988281,
+ "step": 441
+ },
+ {
+ "epoch": 6.139737991266376,
+ "grad_norm": 0.057562313973903656,
+ "learning_rate": 0.0006,
+ "loss": 5.889334678649902,
+ "step": 442
+ },
+ {
+ "epoch": 6.153711790393013,
+ "grad_norm": 0.06447703391313553,
+ "learning_rate": 0.0006,
+ "loss": 5.84621524810791,
+ "step": 443
+ },
+ {
+ "epoch": 6.1676855895196505,
+ "grad_norm": 0.048861872404813766,
+ "learning_rate": 0.0006,
+ "loss": 5.931595802307129,
+ "step": 444
+ },
+ {
+ "epoch": 6.181659388646288,
+ "grad_norm": 0.05521339178085327,
+ "learning_rate": 0.0006,
+ "loss": 5.8229780197143555,
+ "step": 445
+ },
+ {
+ "epoch": 6.195633187772926,
+ "grad_norm": 0.053666990250349045,
+ "learning_rate": 0.0006,
+ "loss": 5.879191875457764,
+ "step": 446
+ },
+ {
+ "epoch": 6.209606986899563,
+ "grad_norm": 0.0451025515794754,
+ "learning_rate": 0.0006,
+ "loss": 5.801628112792969,
+ "step": 447
+ },
+ {
+ "epoch": 6.223580786026201,
+ "grad_norm": 0.04291655868291855,
+ "learning_rate": 0.0006,
+ "loss": 5.914680480957031,
+ "step": 448
+ },
+ {
+ "epoch": 6.2375545851528384,
+ "grad_norm": 0.035102490335702896,
+ "learning_rate": 0.0006,
+ "loss": 5.843682765960693,
+ "step": 449
+ },
+ {
+ "epoch": 6.251528384279476,
+ "grad_norm": 0.03403995931148529,
+ "learning_rate": 0.0006,
+ "loss": 5.827154159545898,
+ "step": 450
+ },
+ {
+ "epoch": 6.265502183406113,
+ "grad_norm": 0.03444375470280647,
+ "learning_rate": 0.0006,
+ "loss": 5.800136566162109,
+ "step": 451
+ },
+ {
+ "epoch": 6.279475982532751,
+ "grad_norm": 0.03165159001946449,
+ "learning_rate": 0.0006,
+ "loss": 5.906252384185791,
+ "step": 452
+ },
+ {
+ "epoch": 6.293449781659389,
+ "grad_norm": 0.026153093203902245,
+ "learning_rate": 0.0006,
+ "loss": 5.7423200607299805,
+ "step": 453
+ },
+ {
+ "epoch": 6.307423580786026,
+ "grad_norm": 0.024357657879590988,
+ "learning_rate": 0.0006,
+ "loss": 5.701364517211914,
+ "step": 454
+ },
+ {
+ "epoch": 6.321397379912664,
+ "grad_norm": 0.021508049219846725,
+ "learning_rate": 0.0006,
+ "loss": 5.8328094482421875,
+ "step": 455
+ },
+ {
+ "epoch": 6.335371179039301,
+ "grad_norm": 0.017313921824097633,
+ "learning_rate": 0.0006,
+ "loss": 5.792145252227783,
+ "step": 456
+ },
+ {
+ "epoch": 6.349344978165939,
+ "grad_norm": 0.018563397228717804,
+ "learning_rate": 0.0006,
+ "loss": 5.732672214508057,
+ "step": 457
+ },
+ {
+ "epoch": 6.3633187772925766,
+ "grad_norm": 0.016568679362535477,
+ "learning_rate": 0.0006,
+ "loss": 5.81948184967041,
+ "step": 458
+ },
+ {
+ "epoch": 6.377292576419214,
+ "grad_norm": 0.014133770950138569,
+ "learning_rate": 0.0006,
+ "loss": 5.734982490539551,
+ "step": 459
+ },
+ {
+ "epoch": 6.391266375545851,
+ "grad_norm": 0.014114447869360447,
+ "learning_rate": 0.0006,
+ "loss": 5.70073127746582,
+ "step": 460
+ },
+ {
+ "epoch": 6.405240174672489,
+ "grad_norm": 0.012907393276691437,
+ "learning_rate": 0.0006,
+ "loss": 5.758626937866211,
+ "step": 461
+ },
+ {
+ "epoch": 6.419213973799127,
+ "grad_norm": 0.012673444114625454,
+ "learning_rate": 0.0006,
+ "loss": 5.775138854980469,
+ "step": 462
+ },
+ {
+ "epoch": 6.4331877729257645,
+ "grad_norm": 0.01311410591006279,
+ "learning_rate": 0.0006,
+ "loss": 5.70920991897583,
+ "step": 463
+ },
+ {
+ "epoch": 6.447161572052401,
+ "grad_norm": 0.012935544364154339,
+ "learning_rate": 0.0006,
+ "loss": 5.781479835510254,
+ "step": 464
+ },
+ {
+ "epoch": 6.461135371179039,
+ "grad_norm": 0.01439738366752863,
+ "learning_rate": 0.0006,
+ "loss": 5.691315650939941,
+ "step": 465
+ },
+ {
+ "epoch": 6.475109170305677,
+ "grad_norm": 0.01205186452716589,
+ "learning_rate": 0.0006,
+ "loss": 5.662499904632568,
+ "step": 466
+ },
+ {
+ "epoch": 6.489082969432315,
+ "grad_norm": 0.011575652286410332,
+ "learning_rate": 0.0006,
+ "loss": 5.6271772384643555,
+ "step": 467
+ },
+ {
+ "epoch": 6.503056768558952,
+ "grad_norm": 0.011672502383589745,
+ "learning_rate": 0.0006,
+ "loss": 5.761662006378174,
+ "step": 468
+ },
+ {
+ "epoch": 6.517030567685589,
+ "grad_norm": 0.011541751213371754,
+ "learning_rate": 0.0006,
+ "loss": 5.762078285217285,
+ "step": 469
+ },
+ {
+ "epoch": 6.531004366812227,
+ "grad_norm": 0.011396365240216255,
+ "learning_rate": 0.0006,
+ "loss": 5.67873477935791,
+ "step": 470
+ },
+ {
+ "epoch": 6.544978165938865,
+ "grad_norm": 0.010688499547541142,
+ "learning_rate": 0.0006,
+ "loss": 5.717051982879639,
+ "step": 471
+ },
+ {
+ "epoch": 6.558951965065502,
+ "grad_norm": 0.012224317528307438,
+ "learning_rate": 0.0006,
+ "loss": 5.707948684692383,
+ "step": 472
+ },
+ {
+ "epoch": 6.5729257641921395,
+ "grad_norm": 0.011856825090944767,
+ "learning_rate": 0.0006,
+ "loss": 5.70878791809082,
+ "step": 473
+ },
+ {
+ "epoch": 6.586899563318777,
+ "grad_norm": 0.01199064590036869,
+ "learning_rate": 0.0006,
+ "loss": 5.708697319030762,
+ "step": 474
+ },
+ {
+ "epoch": 6.600873362445415,
+ "grad_norm": 0.01219446212053299,
+ "learning_rate": 0.0006,
+ "loss": 5.6061577796936035,
+ "step": 475
+ },
+ {
+ "epoch": 6.614847161572053,
+ "grad_norm": 0.013397484086453915,
+ "learning_rate": 0.0006,
+ "loss": 5.624789714813232,
+ "step": 476
+ },
+ {
+ "epoch": 6.62882096069869,
+ "grad_norm": 0.01483136136084795,
+ "learning_rate": 0.0006,
+ "loss": 5.721141338348389,
+ "step": 477
+ },
+ {
+ "epoch": 6.642794759825327,
+ "grad_norm": 0.0194488987326622,
+ "learning_rate": 0.0006,
+ "loss": 5.604279518127441,
+ "step": 478
+ },
+ {
+ "epoch": 6.656768558951965,
+ "grad_norm": 0.02079896256327629,
+ "learning_rate": 0.0006,
+ "loss": 5.586322784423828,
+ "step": 479
+ },
+ {
+ "epoch": 6.670742358078603,
+ "grad_norm": 0.018996229395270348,
+ "learning_rate": 0.0006,
+ "loss": 5.5987772941589355,
+ "step": 480
+ },
+ {
+ "epoch": 6.68471615720524,
+ "grad_norm": 0.015779603272676468,
+ "learning_rate": 0.0006,
+ "loss": 5.588602066040039,
+ "step": 481
+ },
+ {
+ "epoch": 6.698689956331878,
+ "grad_norm": 0.015322973020374775,
+ "learning_rate": 0.0006,
+ "loss": 5.686467170715332,
+ "step": 482
+ },
+ {
+ "epoch": 6.712663755458515,
+ "grad_norm": 0.015282213687896729,
+ "learning_rate": 0.0006,
+ "loss": 5.664676666259766,
+ "step": 483
+ },
+ {
+ "epoch": 6.726637554585153,
+ "grad_norm": 0.016589272767305374,
+ "learning_rate": 0.0006,
+ "loss": 5.666739463806152,
+ "step": 484
+ },
+ {
+ "epoch": 6.74061135371179,
+ "grad_norm": 0.01811421848833561,
+ "learning_rate": 0.0006,
+ "loss": 5.647593021392822,
+ "step": 485
+ },
+ {
+ "epoch": 6.754585152838428,
+ "grad_norm": 0.019652029499411583,
+ "learning_rate": 0.0006,
+ "loss": 5.628726959228516,
+ "step": 486
+ },
+ {
+ "epoch": 6.7685589519650655,
+ "grad_norm": 0.02118665538728237,
+ "learning_rate": 0.0006,
+ "loss": 5.6580705642700195,
+ "step": 487
+ },
+ {
+ "epoch": 6.782532751091703,
+ "grad_norm": 0.02237832546234131,
+ "learning_rate": 0.0006,
+ "loss": 5.635931968688965,
+ "step": 488
+ },
+ {
+ "epoch": 6.796506550218341,
+ "grad_norm": 0.023897631093859673,
+ "learning_rate": 0.0006,
+ "loss": 5.517949104309082,
+ "step": 489
+ },
+ {
+ "epoch": 6.810480349344978,
+ "grad_norm": 0.02442227490246296,
+ "learning_rate": 0.0006,
+ "loss": 5.600021839141846,
+ "step": 490
+ },
+ {
+ "epoch": 6.824454148471616,
+ "grad_norm": 0.024675287306308746,
+ "learning_rate": 0.0006,
+ "loss": 5.6639909744262695,
+ "step": 491
+ },
+ {
+ "epoch": 6.8384279475982535,
+ "grad_norm": 0.030372392386198044,
+ "learning_rate": 0.0006,
+ "loss": 5.582888603210449,
+ "step": 492
+ },
+ {
+ "epoch": 6.85240174672489,
+ "grad_norm": 0.03237690031528473,
+ "learning_rate": 0.0006,
+ "loss": 5.573896884918213,
+ "step": 493
+ },
+ {
+ "epoch": 6.866375545851528,
+ "grad_norm": 0.03511589393019676,
+ "learning_rate": 0.0006,
+ "loss": 5.565878868103027,
+ "step": 494
+ },
+ {
+ "epoch": 6.880349344978166,
+ "grad_norm": 0.03883938118815422,
+ "learning_rate": 0.0006,
+ "loss": 5.58540678024292,
+ "step": 495
+ },
+ {
+ "epoch": 6.894323144104804,
+ "grad_norm": 0.04175502806901932,
+ "learning_rate": 0.0006,
+ "loss": 5.6095194816589355,
+ "step": 496
+ },
+ {
+ "epoch": 6.908296943231441,
+ "grad_norm": 0.04013441503047943,
+ "learning_rate": 0.0006,
+ "loss": 5.654401779174805,
+ "step": 497
+ },
+ {
+ "epoch": 6.922270742358078,
+ "grad_norm": 0.04618770629167557,
+ "learning_rate": 0.0006,
+ "loss": 5.6483473777771,
+ "step": 498
+ },
+ {
+ "epoch": 6.936244541484716,
+ "grad_norm": 0.04507308453321457,
+ "learning_rate": 0.0006,
+ "loss": 5.5856218338012695,
+ "step": 499
+ },
+ {
+ "epoch": 6.950218340611354,
+ "grad_norm": 0.03151383996009827,
+ "learning_rate": 0.0006,
+ "loss": 5.594086647033691,
+ "step": 500
+ },
+ {
+ "epoch": 6.964192139737992,
+ "grad_norm": 0.026883797720074654,
+ "learning_rate": 0.0006,
+ "loss": 5.607676029205322,
+ "step": 501
+ },
+ {
+ "epoch": 6.978165938864628,
+ "grad_norm": 0.02981836162507534,
+ "learning_rate": 0.0006,
+ "loss": 5.563666820526123,
+ "step": 502
+ },
+ {
+ "epoch": 6.992139737991266,
+ "grad_norm": 0.03233477845788002,
+ "learning_rate": 0.0006,
+ "loss": 5.667543411254883,
+ "step": 503
+ },
+ {
+ "epoch": 7.0,
+ "grad_norm": 0.029927456751465797,
+ "learning_rate": 0.0006,
+ "loss": 5.733482837677002,
+ "step": 504
+ },
+ {
+ "epoch": 7.0,
+ "eval_loss": 5.679076671600342,
+ "eval_runtime": 59.5965,
+ "eval_samples_per_second": 40.976,
+ "eval_steps_per_second": 1.292,
+ "step": 504
+ },
+ {
+ "epoch": 7.013973799126638,
+ "grad_norm": 0.027081597596406937,
+ "learning_rate": 0.0006,
+ "loss": 5.634487152099609,
+ "step": 505
+ },
+ {
+ "epoch": 7.0279475982532755,
+ "grad_norm": 0.031959421932697296,
+ "learning_rate": 0.0006,
+ "loss": 5.643294811248779,
+ "step": 506
+ },
+ {
+ "epoch": 7.041921397379912,
+ "grad_norm": 0.02802063524723053,
+ "learning_rate": 0.0006,
+ "loss": 5.573239326477051,
+ "step": 507
+ },
+ {
+ "epoch": 7.05589519650655,
+ "grad_norm": 0.032579705119132996,
+ "learning_rate": 0.0006,
+ "loss": 5.627150058746338,
+ "step": 508
+ },
+ {
+ "epoch": 7.069868995633188,
+ "grad_norm": 0.02503928542137146,
+ "learning_rate": 0.0006,
+ "loss": 5.6045308113098145,
+ "step": 509
+ },
+ {
+ "epoch": 7.083842794759826,
+ "grad_norm": 0.02421317622065544,
+ "learning_rate": 0.0006,
+ "loss": 5.522153377532959,
+ "step": 510
+ },
+ {
+ "epoch": 7.097816593886463,
+ "grad_norm": 0.021983426064252853,
+ "learning_rate": 0.0006,
+ "loss": 5.632939338684082,
+ "step": 511
+ },
+ {
+ "epoch": 7.1117903930131,
+ "grad_norm": 0.021933183073997498,
+ "learning_rate": 0.0006,
+ "loss": 5.592185974121094,
+ "step": 512
+ },
+ {
+ "epoch": 7.125764192139738,
+ "grad_norm": 0.02287031151354313,
+ "learning_rate": 0.0006,
+ "loss": 5.613700866699219,
+ "step": 513
+ },
+ {
+ "epoch": 7.139737991266376,
+ "grad_norm": 0.02281602844595909,
+ "learning_rate": 0.0006,
+ "loss": 5.551383018493652,
+ "step": 514
+ },
+ {
+ "epoch": 7.153711790393013,
+ "grad_norm": 0.021050360053777695,
+ "learning_rate": 0.0006,
+ "loss": 5.55275821685791,
+ "step": 515
+ },
+ {
+ "epoch": 7.1676855895196505,
+ "grad_norm": 0.018299926072359085,
+ "learning_rate": 0.0006,
+ "loss": 5.528225421905518,
+ "step": 516
+ },
+ {
+ "epoch": 7.181659388646288,
+ "grad_norm": 0.02024853602051735,
+ "learning_rate": 0.0006,
+ "loss": 5.522453784942627,
+ "step": 517
+ },
+ {
+ "epoch": 7.195633187772926,
+ "grad_norm": 0.02135239914059639,
+ "learning_rate": 0.0006,
+ "loss": 5.527522563934326,
+ "step": 518
+ },
+ {
+ "epoch": 7.209606986899563,
+ "grad_norm": 0.023394184187054634,
+ "learning_rate": 0.0006,
+ "loss": 5.4535017013549805,
+ "step": 519
+ },
+ {
+ "epoch": 7.223580786026201,
+ "grad_norm": 0.025952080264687538,
+ "learning_rate": 0.0006,
+ "loss": 5.5374674797058105,
+ "step": 520
+ },
+ {
+ "epoch": 7.2375545851528384,
+ "grad_norm": 0.022287718951702118,
+ "learning_rate": 0.0006,
+ "loss": 5.493753433227539,
+ "step": 521
+ },
+ {
+ "epoch": 7.251528384279476,
+ "grad_norm": 0.02018447406589985,
+ "learning_rate": 0.0006,
+ "loss": 5.510575294494629,
+ "step": 522
+ },
+ {
+ "epoch": 7.265502183406113,
+ "grad_norm": 0.02050507813692093,
+ "learning_rate": 0.0006,
+ "loss": 5.466026782989502,
+ "step": 523
+ },
+ {
+ "epoch": 7.279475982532751,
+ "grad_norm": 0.023688802495598793,
+ "learning_rate": 0.0006,
+ "loss": 5.485815525054932,
+ "step": 524
+ },
+ {
+ "epoch": 7.293449781659389,
+ "grad_norm": 0.021078843623399734,
+ "learning_rate": 0.0006,
+ "loss": 5.415581703186035,
+ "step": 525
+ },
+ {
+ "epoch": 7.307423580786026,
+ "grad_norm": 0.018266011029481888,
+ "learning_rate": 0.0006,
+ "loss": 5.402815818786621,
+ "step": 526
+ },
+ {
+ "epoch": 7.321397379912664,
+ "grad_norm": 0.019693493843078613,
+ "learning_rate": 0.0006,
+ "loss": 5.504674911499023,
+ "step": 527
+ },
+ {
+ "epoch": 7.335371179039301,
+ "grad_norm": 0.02031373605132103,
+ "learning_rate": 0.0006,
+ "loss": 5.503837585449219,
+ "step": 528
+ },
+ {
+ "epoch": 7.349344978165939,
+ "grad_norm": 0.017636626958847046,
+ "learning_rate": 0.0006,
+ "loss": 5.530580997467041,
+ "step": 529
+ },
+ {
+ "epoch": 7.3633187772925766,
+ "grad_norm": 0.01787244901061058,
+ "learning_rate": 0.0006,
+ "loss": 5.450218677520752,
+ "step": 530
+ },
+ {
+ "epoch": 7.377292576419214,
+ "grad_norm": 0.0170669574290514,
+ "learning_rate": 0.0006,
+ "loss": 5.481570243835449,
+ "step": 531
+ },
+ {
+ "epoch": 7.391266375545851,
+ "grad_norm": 0.01802165061235428,
+ "learning_rate": 0.0006,
+ "loss": 5.386394500732422,
+ "step": 532
+ },
+ {
+ "epoch": 7.405240174672489,
+ "grad_norm": 0.022949472069740295,
+ "learning_rate": 0.0006,
+ "loss": 5.424929618835449,
+ "step": 533
+ },
+ {
+ "epoch": 7.419213973799127,
+ "grad_norm": 0.034213390201330185,
+ "learning_rate": 0.0006,
+ "loss": 5.346663475036621,
+ "step": 534
+ },
+ {
+ "epoch": 7.4331877729257645,
+ "grad_norm": 0.03847593814134598,
+ "learning_rate": 0.0006,
+ "loss": 5.418482780456543,
+ "step": 535
+ },
+ {
+ "epoch": 7.447161572052401,
+ "grad_norm": 0.029393676668405533,
+ "learning_rate": 0.0006,
+ "loss": 5.456090927124023,
+ "step": 536
+ },
+ {
+ "epoch": 7.461135371179039,
+ "grad_norm": 0.03388667106628418,
+ "learning_rate": 0.0006,
+ "loss": 5.572983741760254,
+ "step": 537
+ },
+ {
+ "epoch": 7.475109170305677,
+ "grad_norm": 0.039690613746643066,
+ "learning_rate": 0.0006,
+ "loss": 5.414067268371582,
+ "step": 538
+ },
+ {
+ "epoch": 7.489082969432315,
+ "grad_norm": 0.03630776330828667,
+ "learning_rate": 0.0006,
+ "loss": 5.523739814758301,
+ "step": 539
+ },
+ {
+ "epoch": 7.503056768558952,
+ "grad_norm": 0.03356870263814926,
+ "learning_rate": 0.0006,
+ "loss": 5.444767951965332,
+ "step": 540
+ },
+ {
+ "epoch": 7.517030567685589,
+ "grad_norm": 0.030812304466962814,
+ "learning_rate": 0.0006,
+ "loss": 5.461244106292725,
+ "step": 541
+ },
+ {
+ "epoch": 7.531004366812227,
+ "grad_norm": 0.03319217637181282,
+ "learning_rate": 0.0006,
+ "loss": 5.502161026000977,
+ "step": 542
+ },
+ {
+ "epoch": 7.544978165938865,
+ "grad_norm": 0.032997481524944305,
+ "learning_rate": 0.0006,
+ "loss": 5.4162278175354,
+ "step": 543
+ },
+ {
+ "epoch": 7.558951965065502,
+ "grad_norm": 0.03364962339401245,
+ "learning_rate": 0.0006,
+ "loss": 5.415736198425293,
+ "step": 544
+ },
+ {
+ "epoch": 7.5729257641921395,
+ "grad_norm": 0.0344221405684948,
+ "learning_rate": 0.0006,
+ "loss": 5.541967868804932,
+ "step": 545
+ },
+ {
+ "epoch": 7.586899563318777,
+ "grad_norm": 0.029609503224492073,
+ "learning_rate": 0.0006,
+ "loss": 5.440614223480225,
+ "step": 546
+ },
+ {
+ "epoch": 7.600873362445415,
+ "grad_norm": 0.029055926948785782,
+ "learning_rate": 0.0006,
+ "loss": 5.462865352630615,
+ "step": 547
+ },
+ {
+ "epoch": 7.614847161572053,
+ "grad_norm": 0.02658848837018013,
+ "learning_rate": 0.0006,
+ "loss": 5.444467544555664,
+ "step": 548
+ },
+ {
+ "epoch": 7.62882096069869,
+ "grad_norm": 0.026276404038071632,
+ "learning_rate": 0.0006,
+ "loss": 5.462278842926025,
+ "step": 549
+ },
+ {
+ "epoch": 7.642794759825327,
+ "grad_norm": 0.0286889486014843,
+ "learning_rate": 0.0006,
+ "loss": 5.4849534034729,
+ "step": 550
+ },
+ {
+ "epoch": 7.656768558951965,
+ "grad_norm": 0.026849817484617233,
+ "learning_rate": 0.0006,
+ "loss": 5.431473731994629,
+ "step": 551
+ },
+ {
+ "epoch": 7.670742358078603,
+ "grad_norm": 0.02312396466732025,
+ "learning_rate": 0.0006,
+ "loss": 5.45462703704834,
+ "step": 552
+ },
+ {
+ "epoch": 7.68471615720524,
+ "grad_norm": 0.026197485625743866,
+ "learning_rate": 0.0006,
+ "loss": 5.452552318572998,
+ "step": 553
+ },
+ {
+ "epoch": 7.698689956331878,
+ "grad_norm": 0.02747255191206932,
+ "learning_rate": 0.0006,
+ "loss": 5.447169780731201,
+ "step": 554
+ },
+ {
+ "epoch": 7.712663755458515,
+ "grad_norm": 0.028123432770371437,
+ "learning_rate": 0.0006,
+ "loss": 5.394400596618652,
+ "step": 555
+ },
+ {
+ "epoch": 7.726637554585153,
+ "grad_norm": 0.02599870041012764,
+ "learning_rate": 0.0006,
+ "loss": 5.398341178894043,
+ "step": 556
+ },
+ {
+ "epoch": 7.74061135371179,
+ "grad_norm": 0.022171657532453537,
+ "learning_rate": 0.0006,
+ "loss": 5.368820667266846,
+ "step": 557
+ },
+ {
+ "epoch": 7.754585152838428,
+ "grad_norm": 0.022309480234980583,
+ "learning_rate": 0.0006,
+ "loss": 5.416772365570068,
+ "step": 558
+ },
+ {
+ "epoch": 7.7685589519650655,
+ "grad_norm": 0.024074165150523186,
+ "learning_rate": 0.0006,
+ "loss": 5.359002113342285,
+ "step": 559
+ },
+ {
+ "epoch": 7.782532751091703,
+ "grad_norm": 0.02636653557419777,
+ "learning_rate": 0.0006,
+ "loss": 5.484368324279785,
+ "step": 560
+ },
+ {
+ "epoch": 7.796506550218341,
+ "grad_norm": 0.02203752100467682,
+ "learning_rate": 0.0006,
+ "loss": 5.375498294830322,
+ "step": 561
+ },
+ {
+ "epoch": 7.810480349344978,
+ "grad_norm": 0.019924819469451904,
+ "learning_rate": 0.0006,
+ "loss": 5.334723472595215,
+ "step": 562
+ },
+ {
+ "epoch": 7.824454148471616,
+ "grad_norm": 0.018755966797471046,
+ "learning_rate": 0.0006,
+ "loss": 5.348781108856201,
+ "step": 563
+ },
+ {
+ "epoch": 7.8384279475982535,
+ "grad_norm": 0.016747845336794853,
+ "learning_rate": 0.0006,
+ "loss": 5.433079719543457,
+ "step": 564
+ },
+ {
+ "epoch": 7.85240174672489,
+ "grad_norm": 0.0170395877212286,
+ "learning_rate": 0.0006,
+ "loss": 5.346794128417969,
+ "step": 565
+ },
+ {
+ "epoch": 7.866375545851528,
+ "grad_norm": 0.016074176877737045,
+ "learning_rate": 0.0006,
+ "loss": 5.3277740478515625,
+ "step": 566
+ },
+ {
+ "epoch": 7.880349344978166,
+ "grad_norm": 0.014119806699454784,
+ "learning_rate": 0.0006,
+ "loss": 5.483551979064941,
+ "step": 567
+ },
+ {
+ "epoch": 7.894323144104804,
+ "grad_norm": 0.015270394273102283,
+ "learning_rate": 0.0006,
+ "loss": 5.378519535064697,
+ "step": 568
+ },
+ {
+ "epoch": 7.908296943231441,
+ "grad_norm": 0.015677539631724358,
+ "learning_rate": 0.0006,
+ "loss": 5.290210247039795,
+ "step": 569
+ },
+ {
+ "epoch": 7.922270742358078,
+ "grad_norm": 0.015930423513054848,
+ "learning_rate": 0.0006,
+ "loss": 5.427360534667969,
+ "step": 570
+ },
+ {
+ "epoch": 7.936244541484716,
+ "grad_norm": 0.016641564667224884,
+ "learning_rate": 0.0006,
+ "loss": 5.301599979400635,
+ "step": 571
+ },
+ {
+ "epoch": 7.950218340611354,
+ "grad_norm": 0.018293552100658417,
+ "learning_rate": 0.0006,
+ "loss": 5.277889251708984,
+ "step": 572
+ },
+ {
+ "epoch": 7.964192139737992,
+ "grad_norm": 0.017618799582123756,
+ "learning_rate": 0.0006,
+ "loss": 5.297194480895996,
+ "step": 573
+ },
+ {
+ "epoch": 7.978165938864628,
+ "grad_norm": 0.015034242533147335,
+ "learning_rate": 0.0006,
+ "loss": 5.42428731918335,
+ "step": 574
+ },
+ {
+ "epoch": 7.992139737991266,
+ "grad_norm": 0.01617511734366417,
+ "learning_rate": 0.0006,
+ "loss": 5.420044898986816,
+ "step": 575
+ },
+ {
+ "epoch": 8.0,
+ "grad_norm": 0.01837257295846939,
+ "learning_rate": 0.0006,
+ "loss": 5.240387916564941,
+ "step": 576
+ },
+ {
+ "epoch": 8.0,
+ "eval_loss": 5.423932075500488,
+ "eval_runtime": 58.0841,
+ "eval_samples_per_second": 42.042,
+ "eval_steps_per_second": 1.326,
+ "step": 576
+ },
+ {
+ "epoch": 8.013973799126637,
+ "grad_norm": 0.020953809842467308,
+ "learning_rate": 0.0006,
+ "loss": 5.4320783615112305,
+ "step": 577
+ },
+ {
+ "epoch": 8.027947598253276,
+ "grad_norm": 0.0227745920419693,
+ "learning_rate": 0.0006,
+ "loss": 5.310197830200195,
+ "step": 578
+ },
+ {
+ "epoch": 8.041921397379912,
+ "grad_norm": 0.021372603252530098,
+ "learning_rate": 0.0006,
+ "loss": 5.339468002319336,
+ "step": 579
+ },
+ {
+ "epoch": 8.055895196506551,
+ "grad_norm": 0.022401731461286545,
+ "learning_rate": 0.0006,
+ "loss": 5.279829978942871,
+ "step": 580
+ },
+ {
+ "epoch": 8.069868995633188,
+ "grad_norm": 0.023942379280924797,
+ "learning_rate": 0.0006,
+ "loss": 5.372323989868164,
+ "step": 581
+ },
+ {
+ "epoch": 8.083842794759825,
+ "grad_norm": 0.024380534887313843,
+ "learning_rate": 0.0006,
+ "loss": 5.253486633300781,
+ "step": 582
+ },
+ {
+ "epoch": 8.097816593886463,
+ "grad_norm": 0.03246625140309334,
+ "learning_rate": 0.0006,
+ "loss": 5.305774211883545,
+ "step": 583
+ },
+ {
+ "epoch": 8.1117903930131,
+ "grad_norm": 0.030399736016988754,
+ "learning_rate": 0.0006,
+ "loss": 5.370976448059082,
+ "step": 584
+ },
+ {
+ "epoch": 8.125764192139737,
+ "grad_norm": 0.027148913592100143,
+ "learning_rate": 0.0006,
+ "loss": 5.368010520935059,
+ "step": 585
+ },
+ {
+ "epoch": 8.139737991266376,
+ "grad_norm": 0.02925540879368782,
+ "learning_rate": 0.0006,
+ "loss": 5.303013801574707,
+ "step": 586
+ },
+ {
+ "epoch": 8.153711790393013,
+ "grad_norm": 0.02915453538298607,
+ "learning_rate": 0.0006,
+ "loss": 5.329667091369629,
+ "step": 587
+ },
+ {
+ "epoch": 8.167685589519651,
+ "grad_norm": 0.031732626259326935,
+ "learning_rate": 0.0006,
+ "loss": 5.3634934425354,
+ "step": 588
+ },
+ {
+ "epoch": 8.181659388646288,
+ "grad_norm": 0.036166731268167496,
+ "learning_rate": 0.0006,
+ "loss": 5.223980903625488,
+ "step": 589
+ },
+ {
+ "epoch": 8.195633187772925,
+ "grad_norm": 0.035654108971357346,
+ "learning_rate": 0.0006,
+ "loss": 5.361059188842773,
+ "step": 590
+ },
+ {
+ "epoch": 8.209606986899564,
+ "grad_norm": 0.0365324430167675,
+ "learning_rate": 0.0006,
+ "loss": 5.336151599884033,
+ "step": 591
+ },
+ {
+ "epoch": 8.2235807860262,
+ "grad_norm": 0.03578880429267883,
+ "learning_rate": 0.0006,
+ "loss": 5.434076309204102,
+ "step": 592
+ },
+ {
+ "epoch": 8.237554585152838,
+ "grad_norm": 0.035734012722969055,
+ "learning_rate": 0.0006,
+ "loss": 5.330893516540527,
+ "step": 593
+ },
+ {
+ "epoch": 8.251528384279476,
+ "grad_norm": 0.03196857124567032,
+ "learning_rate": 0.0006,
+ "loss": 5.306009769439697,
+ "step": 594
+ },
+ {
+ "epoch": 8.265502183406113,
+ "grad_norm": 0.032042086124420166,
+ "learning_rate": 0.0006,
+ "loss": 5.343267917633057,
+ "step": 595
+ },
+ {
+ "epoch": 8.279475982532752,
+ "grad_norm": 0.03160746395587921,
+ "learning_rate": 0.0006,
+ "loss": 5.2353715896606445,
+ "step": 596
+ },
+ {
+ "epoch": 8.293449781659389,
+ "grad_norm": 0.026689818128943443,
+ "learning_rate": 0.0006,
+ "loss": 5.254042625427246,
+ "step": 597
+ },
+ {
+ "epoch": 8.307423580786025,
+ "grad_norm": 0.02880188450217247,
+ "learning_rate": 0.0006,
+ "loss": 5.326833248138428,
+ "step": 598
+ },
+ {
+ "epoch": 8.321397379912664,
+ "grad_norm": 0.027516381815075874,
+ "learning_rate": 0.0006,
+ "loss": 5.342336177825928,
+ "step": 599
+ },
+ {
+ "epoch": 8.335371179039301,
+ "grad_norm": 0.027875030413269997,
+ "learning_rate": 0.0006,
+ "loss": 5.200719833374023,
+ "step": 600
+ },
+ {
+ "epoch": 8.34934497816594,
+ "grad_norm": 0.0268265288323164,
+ "learning_rate": 0.0006,
+ "loss": 5.249449729919434,
+ "step": 601
+ },
+ {
+ "epoch": 8.363318777292577,
+ "grad_norm": 0.024824608117341995,
+ "learning_rate": 0.0006,
+ "loss": 5.318952560424805,
+ "step": 602
+ },
+ {
+ "epoch": 8.377292576419213,
+ "grad_norm": 0.022990627214312553,
+ "learning_rate": 0.0006,
+ "loss": 5.244993209838867,
+ "step": 603
+ },
+ {
+ "epoch": 8.391266375545852,
+ "grad_norm": 0.022804604843258858,
+ "learning_rate": 0.0006,
+ "loss": 5.3236846923828125,
+ "step": 604
+ },
+ {
+ "epoch": 8.405240174672489,
+ "grad_norm": 0.02419872209429741,
+ "learning_rate": 0.0006,
+ "loss": 5.255486488342285,
+ "step": 605
+ },
+ {
+ "epoch": 8.419213973799126,
+ "grad_norm": 0.021952059119939804,
+ "learning_rate": 0.0006,
+ "loss": 5.3511552810668945,
+ "step": 606
+ },
+ {
+ "epoch": 8.433187772925764,
+ "grad_norm": 0.022375497967004776,
+ "learning_rate": 0.0006,
+ "loss": 5.294790267944336,
+ "step": 607
+ },
+ {
+ "epoch": 8.447161572052401,
+ "grad_norm": 0.019242815673351288,
+ "learning_rate": 0.0006,
+ "loss": 5.252618789672852,
+ "step": 608
+ },
+ {
+ "epoch": 8.46113537117904,
+ "grad_norm": 0.0173486340790987,
+ "learning_rate": 0.0006,
+ "loss": 5.30147647857666,
+ "step": 609
+ },
+ {
+ "epoch": 8.475109170305677,
+ "grad_norm": 0.01612604409456253,
+ "learning_rate": 0.0006,
+ "loss": 5.23960542678833,
+ "step": 610
+ },
+ {
+ "epoch": 8.489082969432314,
+ "grad_norm": 0.014293976128101349,
+ "learning_rate": 0.0006,
+ "loss": 5.119932651519775,
+ "step": 611
+ },
+ {
+ "epoch": 8.503056768558952,
+ "grad_norm": 0.014453536830842495,
+ "learning_rate": 0.0006,
+ "loss": 5.166794776916504,
+ "step": 612
+ },
+ {
+ "epoch": 8.51703056768559,
+ "grad_norm": 0.014685769565403461,
+ "learning_rate": 0.0006,
+ "loss": 5.264822959899902,
+ "step": 613
+ },
+ {
+ "epoch": 8.531004366812226,
+ "grad_norm": 0.01427740603685379,
+ "learning_rate": 0.0006,
+ "loss": 5.278433799743652,
+ "step": 614
+ },
+ {
+ "epoch": 8.544978165938865,
+ "grad_norm": 0.014864951372146606,
+ "learning_rate": 0.0006,
+ "loss": 5.214512348175049,
+ "step": 615
+ },
+ {
+ "epoch": 8.558951965065502,
+ "grad_norm": 0.01565164513885975,
+ "learning_rate": 0.0006,
+ "loss": 5.183066368103027,
+ "step": 616
+ },
+ {
+ "epoch": 8.57292576419214,
+ "grad_norm": 0.01779816672205925,
+ "learning_rate": 0.0006,
+ "loss": 5.163185119628906,
+ "step": 617
+ },
+ {
+ "epoch": 8.586899563318777,
+ "grad_norm": 0.017254827544093132,
+ "learning_rate": 0.0006,
+ "loss": 5.243555068969727,
+ "step": 618
+ },
+ {
+ "epoch": 8.600873362445414,
+ "grad_norm": 0.01650584116578102,
+ "learning_rate": 0.0006,
+ "loss": 5.234884738922119,
+ "step": 619
+ },
+ {
+ "epoch": 8.614847161572053,
+ "grad_norm": 0.017021115869283676,
+ "learning_rate": 0.0006,
+ "loss": 5.12271785736084,
+ "step": 620
+ },
+ {
+ "epoch": 8.62882096069869,
+ "grad_norm": 0.01773759163916111,
+ "learning_rate": 0.0006,
+ "loss": 5.242153167724609,
+ "step": 621
+ },
+ {
+ "epoch": 8.642794759825328,
+ "grad_norm": 0.015679042786359787,
+ "learning_rate": 0.0006,
+ "loss": 5.170779228210449,
+ "step": 622
+ },
+ {
+ "epoch": 8.656768558951965,
+ "grad_norm": 0.013760549947619438,
+ "learning_rate": 0.0006,
+ "loss": 5.238644599914551,
+ "step": 623
+ },
+ {
+ "epoch": 8.670742358078602,
+ "grad_norm": 0.014571522362530231,
+ "learning_rate": 0.0006,
+ "loss": 5.177056312561035,
+ "step": 624
+ },
+ {
+ "epoch": 8.68471615720524,
+ "grad_norm": 0.016209015622735023,
+ "learning_rate": 0.0006,
+ "loss": 5.163750648498535,
+ "step": 625
+ },
+ {
+ "epoch": 8.698689956331878,
+ "grad_norm": 0.016813941299915314,
+ "learning_rate": 0.0006,
+ "loss": 5.183428764343262,
+ "step": 626
+ },
+ {
+ "epoch": 8.712663755458514,
+ "grad_norm": 0.020985357463359833,
+ "learning_rate": 0.0006,
+ "loss": 5.229465007781982,
+ "step": 627
+ },
+ {
+ "epoch": 8.726637554585153,
+ "grad_norm": 0.02679632417857647,
+ "learning_rate": 0.0006,
+ "loss": 5.257368087768555,
+ "step": 628
+ },
+ {
+ "epoch": 8.74061135371179,
+ "grad_norm": 0.025756070390343666,
+ "learning_rate": 0.0006,
+ "loss": 5.253736972808838,
+ "step": 629
+ },
+ {
+ "epoch": 8.754585152838429,
+ "grad_norm": 0.02643490768969059,
+ "learning_rate": 0.0006,
+ "loss": 5.228633403778076,
+ "step": 630
+ },
+ {
+ "epoch": 8.768558951965066,
+ "grad_norm": 0.03303210437297821,
+ "learning_rate": 0.0006,
+ "loss": 5.163540363311768,
+ "step": 631
+ },
+ {
+ "epoch": 8.782532751091702,
+ "grad_norm": 0.030432431027293205,
+ "learning_rate": 0.0006,
+ "loss": 5.243169784545898,
+ "step": 632
+ },
+ {
+ "epoch": 8.796506550218341,
+ "grad_norm": 0.028631288558244705,
+ "learning_rate": 0.0006,
+ "loss": 5.166398048400879,
+ "step": 633
+ },
+ {
+ "epoch": 8.810480349344978,
+ "grad_norm": 0.026188310235738754,
+ "learning_rate": 0.0006,
+ "loss": 5.128122329711914,
+ "step": 634
+ },
+ {
+ "epoch": 8.824454148471617,
+ "grad_norm": 0.025308528915047646,
+ "learning_rate": 0.0006,
+ "loss": 5.211284637451172,
+ "step": 635
+ },
+ {
+ "epoch": 8.838427947598253,
+ "grad_norm": 0.020657729357481003,
+ "learning_rate": 0.0006,
+ "loss": 5.161575794219971,
+ "step": 636
+ },
+ {
+ "epoch": 8.85240174672489,
+ "grad_norm": 0.021901234984397888,
+ "learning_rate": 0.0006,
+ "loss": 5.20050048828125,
+ "step": 637
+ },
+ {
+ "epoch": 8.866375545851529,
+ "grad_norm": 0.021219318732619286,
+ "learning_rate": 0.0006,
+ "loss": 5.211699485778809,
+ "step": 638
+ },
+ {
+ "epoch": 8.880349344978166,
+ "grad_norm": 0.0200974028557539,
+ "learning_rate": 0.0006,
+ "loss": 5.092503547668457,
+ "step": 639
+ },
+ {
+ "epoch": 8.894323144104803,
+ "grad_norm": 0.023804882541298866,
+ "learning_rate": 0.0006,
+ "loss": 5.216068267822266,
+ "step": 640
+ },
+ {
+ "epoch": 8.908296943231441,
+ "grad_norm": 0.026088010519742966,
+ "learning_rate": 0.0006,
+ "loss": 5.155592918395996,
+ "step": 641
+ },
+ {
+ "epoch": 8.922270742358078,
+ "grad_norm": 0.02601276896893978,
+ "learning_rate": 0.0006,
+ "loss": 5.197238922119141,
+ "step": 642
+ },
+ {
+ "epoch": 8.936244541484717,
+ "grad_norm": 0.020387211814522743,
+ "learning_rate": 0.0006,
+ "loss": 5.1955437660217285,
+ "step": 643
+ },
+ {
+ "epoch": 8.950218340611354,
+ "grad_norm": 0.019214622676372528,
+ "learning_rate": 0.0006,
+ "loss": 5.206346035003662,
+ "step": 644
+ },
+ {
+ "epoch": 8.96419213973799,
+ "grad_norm": 0.019674314185976982,
+ "learning_rate": 0.0006,
+ "loss": 5.219857215881348,
+ "step": 645
+ },
+ {
+ "epoch": 8.97816593886463,
+ "grad_norm": 0.020318234339356422,
+ "learning_rate": 0.0006,
+ "loss": 5.191132545471191,
+ "step": 646
+ },
+ {
+ "epoch": 8.992139737991266,
+ "grad_norm": 0.021428676322102547,
+ "learning_rate": 0.0006,
+ "loss": 5.214695930480957,
+ "step": 647
+ },
+ {
+ "epoch": 9.0,
+ "grad_norm": 0.023235054686665535,
+ "learning_rate": 0.0006,
+ "loss": 5.229091644287109,
+ "step": 648
+ },
+ {
+ "epoch": 9.0,
+ "eval_loss": 5.289593696594238,
+ "eval_runtime": 56.9256,
+ "eval_samples_per_second": 42.898,
+ "eval_steps_per_second": 1.353,
+ "step": 648
+ },
+ {
+ "epoch": 9.013973799126637,
+ "grad_norm": 0.02320289984345436,
+ "learning_rate": 0.0006,
+ "loss": 4.983669281005859,
+ "step": 649
+ },
+ {
+ "epoch": 9.027947598253276,
+ "grad_norm": 0.018521282821893692,
+ "learning_rate": 0.0006,
+ "loss": 5.200575828552246,
+ "step": 650
+ },
+ {
+ "epoch": 9.041921397379912,
+ "grad_norm": 0.0180169939994812,
+ "learning_rate": 0.0006,
+ "loss": 5.148033618927002,
+ "step": 651
+ },
+ {
+ "epoch": 9.055895196506551,
+ "grad_norm": 0.019969860091805458,
+ "learning_rate": 0.0006,
+ "loss": 5.099125385284424,
+ "step": 652
+ },
+ {
+ "epoch": 9.069868995633188,
+ "grad_norm": 0.01643305830657482,
+ "learning_rate": 0.0006,
+ "loss": 5.092347145080566,
+ "step": 653
+ },
+ {
+ "epoch": 9.083842794759825,
+ "grad_norm": 0.016022363677620888,
+ "learning_rate": 0.0006,
+ "loss": 5.129279613494873,
+ "step": 654
+ },
+ {
+ "epoch": 9.097816593886463,
+ "grad_norm": 0.01657041162252426,
+ "learning_rate": 0.0006,
+ "loss": 5.159562587738037,
+ "step": 655
+ },
+ {
+ "epoch": 9.1117903930131,
+ "grad_norm": 0.017924228683114052,
+ "learning_rate": 0.0006,
+ "loss": 5.165497779846191,
+ "step": 656
+ },
+ {
+ "epoch": 9.125764192139737,
+ "grad_norm": 0.021592361852526665,
+ "learning_rate": 0.0006,
+ "loss": 5.204775810241699,
+ "step": 657
+ },
+ {
+ "epoch": 9.139737991266376,
+ "grad_norm": 0.026924120262265205,
+ "learning_rate": 0.0006,
+ "loss": 5.128296852111816,
+ "step": 658
+ },
+ {
+ "epoch": 9.153711790393013,
+ "grad_norm": 0.023156899958848953,
+ "learning_rate": 0.0006,
+ "loss": 5.1255340576171875,
+ "step": 659
+ },
+ {
+ "epoch": 9.167685589519651,
+ "grad_norm": 0.020670367404818535,
+ "learning_rate": 0.0006,
+ "loss": 5.128479957580566,
+ "step": 660
+ },
+ {
+ "epoch": 9.181659388646288,
+ "grad_norm": 0.025447173044085503,
+ "learning_rate": 0.0006,
+ "loss": 5.154821872711182,
+ "step": 661
+ },
+ {
+ "epoch": 9.195633187772925,
+ "grad_norm": 0.027940068393945694,
+ "learning_rate": 0.0006,
+ "loss": 5.143466472625732,
+ "step": 662
+ },
+ {
+ "epoch": 9.209606986899564,
+ "grad_norm": 0.026847844943404198,
+ "learning_rate": 0.0006,
+ "loss": 5.045665740966797,
+ "step": 663
+ },
+ {
+ "epoch": 9.2235807860262,
+ "grad_norm": 0.02340601570904255,
+ "learning_rate": 0.0006,
+ "loss": 5.21933126449585,
+ "step": 664
+ },
+ {
+ "epoch": 9.237554585152838,
+ "grad_norm": 0.02340371161699295,
+ "learning_rate": 0.0006,
+ "loss": 4.942949295043945,
+ "step": 665
+ },
+ {
+ "epoch": 9.251528384279476,
+ "grad_norm": 0.02221992425620556,
+ "learning_rate": 0.0006,
+ "loss": 5.197099208831787,
+ "step": 666
+ },
+ {
+ "epoch": 9.265502183406113,
+ "grad_norm": 0.023508677259087563,
+ "learning_rate": 0.0006,
+ "loss": 5.022897720336914,
+ "step": 667
+ },
+ {
+ "epoch": 9.279475982532752,
+ "grad_norm": 0.026120394468307495,
+ "learning_rate": 0.0006,
+ "loss": 5.087725639343262,
+ "step": 668
+ },
+ {
+ "epoch": 9.293449781659389,
+ "grad_norm": 0.027273228392004967,
+ "learning_rate": 0.0006,
+ "loss": 5.090963840484619,
+ "step": 669
+ },
+ {
+ "epoch": 9.307423580786025,
+ "grad_norm": 0.03241586685180664,
+ "learning_rate": 0.0006,
+ "loss": 5.047842979431152,
+ "step": 670
+ },
+ {
+ "epoch": 9.321397379912664,
+ "grad_norm": 0.030245667323470116,
+ "learning_rate": 0.0006,
+ "loss": 5.104803085327148,
+ "step": 671
+ },
+ {
+ "epoch": 9.335371179039301,
+ "grad_norm": 0.027698364108800888,
+ "learning_rate": 0.0006,
+ "loss": 5.0644097328186035,
+ "step": 672
+ },
+ {
+ "epoch": 9.34934497816594,
+ "grad_norm": 0.029692554846405983,
+ "learning_rate": 0.0006,
+ "loss": 5.111942291259766,
+ "step": 673
+ },
+ {
+ "epoch": 9.363318777292577,
+ "grad_norm": 0.03328656405210495,
+ "learning_rate": 0.0006,
+ "loss": 5.1574554443359375,
+ "step": 674
+ },
+ {
+ "epoch": 9.377292576419213,
+ "grad_norm": 0.031938180327415466,
+ "learning_rate": 0.0006,
+ "loss": 5.190371990203857,
+ "step": 675
+ },
+ {
+ "epoch": 9.391266375545852,
+ "grad_norm": 0.024858945980668068,
+ "learning_rate": 0.0006,
+ "loss": 5.121420860290527,
+ "step": 676
+ },
+ {
+ "epoch": 9.405240174672489,
+ "grad_norm": 0.023033203557133675,
+ "learning_rate": 0.0006,
+ "loss": 5.0974345207214355,
+ "step": 677
+ },
+ {
+ "epoch": 9.419213973799126,
+ "grad_norm": 0.022083545103669167,
+ "learning_rate": 0.0006,
+ "loss": 5.046319961547852,
+ "step": 678
+ },
+ {
+ "epoch": 9.433187772925764,
+ "grad_norm": 0.0228437427431345,
+ "learning_rate": 0.0006,
+ "loss": 5.119963645935059,
+ "step": 679
+ },
+ {
+ "epoch": 9.447161572052401,
+ "grad_norm": 0.0242351982742548,
+ "learning_rate": 0.0006,
+ "loss": 5.0444560050964355,
+ "step": 680
+ },
+ {
+ "epoch": 9.46113537117904,
+ "grad_norm": 0.02401185780763626,
+ "learning_rate": 0.0006,
+ "loss": 5.063238620758057,
+ "step": 681
+ },
+ {
+ "epoch": 9.475109170305677,
+ "grad_norm": 0.02560959942638874,
+ "learning_rate": 0.0006,
+ "loss": 5.035224914550781,
+ "step": 682
+ },
+ {
+ "epoch": 9.489082969432314,
+ "grad_norm": 0.024985626339912415,
+ "learning_rate": 0.0006,
+ "loss": 5.049284934997559,
+ "step": 683
+ },
+ {
+ "epoch": 9.503056768558952,
+ "grad_norm": 0.023178115487098694,
+ "learning_rate": 0.0006,
+ "loss": 5.2266621589660645,
+ "step": 684
+ },
+ {
+ "epoch": 9.51703056768559,
+ "grad_norm": 0.0215512253344059,
+ "learning_rate": 0.0006,
+ "loss": 5.1875715255737305,
+ "step": 685
+ },
+ {
+ "epoch": 9.531004366812226,
+ "grad_norm": 0.01662837713956833,
+ "learning_rate": 0.0006,
+ "loss": 5.137012481689453,
+ "step": 686
+ },
+ {
+ "epoch": 9.544978165938865,
+ "grad_norm": 0.01952764391899109,
+ "learning_rate": 0.0006,
+ "loss": 5.072500228881836,
+ "step": 687
+ },
+ {
+ "epoch": 9.558951965065502,
+ "grad_norm": 0.017521077767014503,
+ "learning_rate": 0.0006,
+ "loss": 5.045711994171143,
+ "step": 688
+ },
+ {
+ "epoch": 9.57292576419214,
+ "grad_norm": 0.017355265095829964,
+ "learning_rate": 0.0006,
+ "loss": 5.02881383895874,
+ "step": 689
+ },
+ {
+ "epoch": 9.586899563318777,
+ "grad_norm": 0.015587719157338142,
+ "learning_rate": 0.0006,
+ "loss": 5.132787704467773,
+ "step": 690
+ },
+ {
+ "epoch": 9.600873362445414,
+ "grad_norm": 0.017469845712184906,
+ "learning_rate": 0.0006,
+ "loss": 5.105447292327881,
+ "step": 691
+ },
+ {
+ "epoch": 9.614847161572053,
+ "grad_norm": 0.019630100578069687,
+ "learning_rate": 0.0006,
+ "loss": 5.032464981079102,
+ "step": 692
+ },
+ {
+ "epoch": 9.62882096069869,
+ "grad_norm": 0.01784994639456272,
+ "learning_rate": 0.0006,
+ "loss": 5.081345558166504,
+ "step": 693
+ },
+ {
+ "epoch": 9.642794759825328,
+ "grad_norm": 0.01872754842042923,
+ "learning_rate": 0.0006,
+ "loss": 4.9772114753723145,
+ "step": 694
+ },
+ {
+ "epoch": 9.656768558951965,
+ "grad_norm": 0.01999032311141491,
+ "learning_rate": 0.0006,
+ "loss": 5.089740753173828,
+ "step": 695
+ },
+ {
+ "epoch": 9.670742358078602,
+ "grad_norm": 0.01710767112672329,
+ "learning_rate": 0.0006,
+ "loss": 4.9869890213012695,
+ "step": 696
+ },
+ {
+ "epoch": 9.68471615720524,
+ "grad_norm": 0.015144161880016327,
+ "learning_rate": 0.0006,
+ "loss": 5.022564888000488,
+ "step": 697
+ },
+ {
+ "epoch": 9.698689956331878,
+ "grad_norm": 0.01618543453514576,
+ "learning_rate": 0.0006,
+ "loss": 5.035921573638916,
+ "step": 698
+ },
+ {
+ "epoch": 9.712663755458514,
+ "grad_norm": 0.014653387479484081,
+ "learning_rate": 0.0006,
+ "loss": 4.914456367492676,
+ "step": 699
+ },
+ {
+ "epoch": 9.726637554585153,
+ "grad_norm": 0.013434977270662785,
+ "learning_rate": 0.0006,
+ "loss": 5.063764572143555,
+ "step": 700
+ },
+ {
+ "epoch": 9.74061135371179,
+ "grad_norm": 0.015128864906728268,
+ "learning_rate": 0.0006,
+ "loss": 5.052367687225342,
+ "step": 701
+ },
+ {
+ "epoch": 9.754585152838429,
+ "grad_norm": 0.014210768043994904,
+ "learning_rate": 0.0006,
+ "loss": 5.016489028930664,
+ "step": 702
+ },
+ {
+ "epoch": 9.768558951965066,
+ "grad_norm": 0.015191680751740932,
+ "learning_rate": 0.0006,
+ "loss": 4.990176200866699,
+ "step": 703
+ },
+ {
+ "epoch": 9.782532751091702,
+ "grad_norm": 0.018285127356648445,
+ "learning_rate": 0.0006,
+ "loss": 5.015164375305176,
+ "step": 704
+ },
+ {
+ "epoch": 9.796506550218341,
+ "grad_norm": 0.0204079058021307,
+ "learning_rate": 0.0006,
+ "loss": 5.091142654418945,
+ "step": 705
+ },
+ {
+ "epoch": 9.810480349344978,
+ "grad_norm": 0.01980586163699627,
+ "learning_rate": 0.0006,
+ "loss": 5.03350830078125,
+ "step": 706
+ },
+ {
+ "epoch": 9.824454148471617,
+ "grad_norm": 0.02048729732632637,
+ "learning_rate": 0.0006,
+ "loss": 5.03611421585083,
+ "step": 707
+ },
+ {
+ "epoch": 9.838427947598253,
+ "grad_norm": 0.02058163844048977,
+ "learning_rate": 0.0006,
+ "loss": 4.9613447189331055,
+ "step": 708
+ },
+ {
+ "epoch": 9.85240174672489,
+ "grad_norm": 0.018004924058914185,
+ "learning_rate": 0.0006,
+ "loss": 4.966541767120361,
+ "step": 709
+ },
+ {
+ "epoch": 9.866375545851529,
+ "grad_norm": 0.0198152307420969,
+ "learning_rate": 0.0006,
+ "loss": 5.01750373840332,
+ "step": 710
+ },
+ {
+ "epoch": 9.880349344978166,
+ "grad_norm": 0.018532969057559967,
+ "learning_rate": 0.0006,
+ "loss": 5.097424507141113,
+ "step": 711
+ },
+ {
+ "epoch": 9.894323144104803,
+ "grad_norm": 0.018939971923828125,
+ "learning_rate": 0.0006,
+ "loss": 5.177223205566406,
+ "step": 712
+ },
+ {
+ "epoch": 9.908296943231441,
+ "grad_norm": 0.017969170585274696,
+ "learning_rate": 0.0006,
+ "loss": 5.118013858795166,
+ "step": 713
+ },
+ {
+ "epoch": 9.922270742358078,
+ "grad_norm": 0.018021270632743835,
+ "learning_rate": 0.0006,
+ "loss": 5.034787178039551,
+ "step": 714
+ },
+ {
+ "epoch": 9.936244541484717,
+ "grad_norm": 0.017009710893034935,
+ "learning_rate": 0.0006,
+ "loss": 5.003920555114746,
+ "step": 715
+ },
+ {
+ "epoch": 9.950218340611354,
+ "grad_norm": 0.017722049728035927,
+ "learning_rate": 0.0006,
+ "loss": 5.031939506530762,
+ "step": 716
+ },
+ {
+ "epoch": 9.96419213973799,
+ "grad_norm": 0.017134087160229683,
+ "learning_rate": 0.0006,
+ "loss": 5.093474388122559,
+ "step": 717
+ },
+ {
+ "epoch": 9.97816593886463,
+ "grad_norm": 0.018511613830924034,
+ "learning_rate": 0.0006,
+ "loss": 5.078604698181152,
+ "step": 718
+ },
+ {
+ "epoch": 9.992139737991266,
+ "grad_norm": 0.021206334233283997,
+ "learning_rate": 0.0006,
+ "loss": 5.052572727203369,
+ "step": 719
+ },
+ {
+ "epoch": 10.0,
+ "grad_norm": 0.02239062264561653,
+ "learning_rate": 0.0006,
+ "loss": 5.063547134399414,
+ "step": 720
+ },
+ {
+ "epoch": 10.0,
+ "eval_loss": 5.173044681549072,
+ "eval_runtime": 56.7287,
+ "eval_samples_per_second": 43.047,
+ "eval_steps_per_second": 1.357,
+ "step": 720
+ },
+ {
+ "epoch": 10.013973799126637,
+ "grad_norm": 0.020757969468832016,
+ "learning_rate": 0.0006,
+ "loss": 5.045648574829102,
+ "step": 721
+ },
+ {
+ "epoch": 10.027947598253276,
+ "grad_norm": 0.022697702050209045,
+ "learning_rate": 0.0006,
+ "loss": 4.993833065032959,
+ "step": 722
+ },
+ {
+ "epoch": 10.041921397379912,
+ "grad_norm": 0.02434331737458706,
+ "learning_rate": 0.0006,
+ "loss": 5.073219299316406,
+ "step": 723
+ },
+ {
+ "epoch": 10.055895196506551,
+ "grad_norm": 0.025499660521745682,
+ "learning_rate": 0.0006,
+ "loss": 5.019072532653809,
+ "step": 724
+ },
+ {
+ "epoch": 10.069868995633188,
+ "grad_norm": 0.029281053692102432,
+ "learning_rate": 0.0006,
+ "loss": 4.959851264953613,
+ "step": 725
+ },
+ {
+ "epoch": 10.083842794759825,
+ "grad_norm": 0.033878110349178314,
+ "learning_rate": 0.0006,
+ "loss": 5.132717609405518,
+ "step": 726
+ },
+ {
+ "epoch": 10.097816593886463,
+ "grad_norm": 0.03385764732956886,
+ "learning_rate": 0.0006,
+ "loss": 5.0482892990112305,
+ "step": 727
+ },
+ {
+ "epoch": 10.1117903930131,
+ "grad_norm": 0.03466648980975151,
+ "learning_rate": 0.0006,
+ "loss": 4.946855545043945,
+ "step": 728
+ },
+ {
+ "epoch": 10.125764192139737,
+ "grad_norm": 0.028546305373311043,
+ "learning_rate": 0.0006,
+ "loss": 5.027774333953857,
+ "step": 729
+ },
+ {
+ "epoch": 10.139737991266376,
+ "grad_norm": 0.030575869604945183,
+ "learning_rate": 0.0006,
+ "loss": 4.9768476486206055,
+ "step": 730
+ },
+ {
+ "epoch": 10.153711790393013,
+ "grad_norm": 0.03614303469657898,
+ "learning_rate": 0.0006,
+ "loss": 5.078113555908203,
+ "step": 731
+ },
+ {
+ "epoch": 10.167685589519651,
+ "grad_norm": 0.040991660207509995,
+ "learning_rate": 0.0006,
+ "loss": 5.033185958862305,
+ "step": 732
+ },
+ {
+ "epoch": 10.181659388646288,
+ "grad_norm": 0.04551699757575989,
+ "learning_rate": 0.0006,
+ "loss": 5.0820746421813965,
+ "step": 733
+ },
+ {
+ "epoch": 10.195633187772925,
+ "grad_norm": 0.040193989872932434,
+ "learning_rate": 0.0006,
+ "loss": 5.059700965881348,
+ "step": 734
+ },
+ {
+ "epoch": 10.209606986899564,
+ "grad_norm": 0.035851605236530304,
+ "learning_rate": 0.0006,
+ "loss": 5.142467498779297,
+ "step": 735
+ },
+ {
+ "epoch": 10.2235807860262,
+ "grad_norm": 0.034200169146060944,
+ "learning_rate": 0.0006,
+ "loss": 5.072312355041504,
+ "step": 736
+ },
+ {
+ "epoch": 10.237554585152838,
+ "grad_norm": 0.033621110022068024,
+ "learning_rate": 0.0006,
+ "loss": 5.08037805557251,
+ "step": 737
+ },
+ {
+ "epoch": 10.251528384279476,
+ "grad_norm": 0.037372369319200516,
+ "learning_rate": 0.0006,
+ "loss": 4.998425483703613,
+ "step": 738
+ },
+ {
+ "epoch": 10.265502183406113,
+ "grad_norm": 0.037044707685709,
+ "learning_rate": 0.0006,
+ "loss": 4.923993110656738,
+ "step": 739
+ },
+ {
+ "epoch": 10.279475982532752,
+ "grad_norm": 0.030454808846116066,
+ "learning_rate": 0.0006,
+ "loss": 5.0889997482299805,
+ "step": 740
+ },
+ {
+ "epoch": 10.293449781659389,
+ "grad_norm": 0.030969638377428055,
+ "learning_rate": 0.0006,
+ "loss": 4.984112739562988,
+ "step": 741
+ },
+ {
+ "epoch": 10.307423580786025,
+ "grad_norm": 0.0276983380317688,
+ "learning_rate": 0.0006,
+ "loss": 4.955412864685059,
+ "step": 742
+ },
+ {
+ "epoch": 10.321397379912664,
+ "grad_norm": 0.02716052532196045,
+ "learning_rate": 0.0006,
+ "loss": 5.0098876953125,
+ "step": 743
+ },
+ {
+ "epoch": 10.335371179039301,
+ "grad_norm": 0.02403552085161209,
+ "learning_rate": 0.0006,
+ "loss": 5.165160655975342,
+ "step": 744
+ },
+ {
+ "epoch": 10.34934497816594,
+ "grad_norm": 0.0257862601429224,
+ "learning_rate": 0.0006,
+ "loss": 5.001547813415527,
+ "step": 745
+ },
+ {
+ "epoch": 10.363318777292577,
+ "grad_norm": 0.02300378680229187,
+ "learning_rate": 0.0006,
+ "loss": 5.038888931274414,
+ "step": 746
+ },
+ {
+ "epoch": 10.377292576419213,
+ "grad_norm": 0.019766854122281075,
+ "learning_rate": 0.0006,
+ "loss": 4.924291610717773,
+ "step": 747
+ },
+ {
+ "epoch": 10.391266375545852,
+ "grad_norm": 0.01888395845890045,
+ "learning_rate": 0.0006,
+ "loss": 5.059971332550049,
+ "step": 748
+ },
+ {
+ "epoch": 10.405240174672489,
+ "grad_norm": 0.015577499754726887,
+ "learning_rate": 0.0006,
+ "loss": 4.987686634063721,
+ "step": 749
+ },
+ {
+ "epoch": 10.419213973799126,
+ "grad_norm": 0.016012346372008324,
+ "learning_rate": 0.0006,
+ "loss": 5.015157699584961,
+ "step": 750
+ },
+ {
+ "epoch": 10.433187772925764,
+ "grad_norm": 0.016277998685836792,
+ "learning_rate": 0.0006,
+ "loss": 5.004931449890137,
+ "step": 751
+ },
+ {
+ "epoch": 10.447161572052401,
+ "grad_norm": 0.017504561692476273,
+ "learning_rate": 0.0006,
+ "loss": 4.902827739715576,
+ "step": 752
+ },
+ {
+ "epoch": 10.46113537117904,
+ "grad_norm": 0.01569022797048092,
+ "learning_rate": 0.0006,
+ "loss": 4.954863548278809,
+ "step": 753
+ },
+ {
+ "epoch": 10.475109170305677,
+ "grad_norm": 0.014114422723650932,
+ "learning_rate": 0.0006,
+ "loss": 5.00656795501709,
+ "step": 754
+ },
+ {
+ "epoch": 10.489082969432314,
+ "grad_norm": 0.01573832333087921,
+ "learning_rate": 0.0006,
+ "loss": 4.978353023529053,
+ "step": 755
+ },
+ {
+ "epoch": 10.503056768558952,
+ "grad_norm": 0.016480036079883575,
+ "learning_rate": 0.0006,
+ "loss": 4.953327655792236,
+ "step": 756
+ },
+ {
+ "epoch": 10.51703056768559,
+ "grad_norm": 0.014008025638759136,
+ "learning_rate": 0.0006,
+ "loss": 5.050840377807617,
+ "step": 757
+ },
+ {
+ "epoch": 10.531004366812226,
+ "grad_norm": 0.013147998601198196,
+ "learning_rate": 0.0006,
+ "loss": 4.974964141845703,
+ "step": 758
+ },
+ {
+ "epoch": 10.544978165938865,
+ "grad_norm": 0.013807603158056736,
+ "learning_rate": 0.0006,
+ "loss": 4.927907943725586,
+ "step": 759
+ },
+ {
+ "epoch": 10.558951965065502,
+ "grad_norm": 0.013555224984884262,
+ "learning_rate": 0.0006,
+ "loss": 4.980690956115723,
+ "step": 760
+ },
+ {
+ "epoch": 10.57292576419214,
+ "grad_norm": 0.013045408762991428,
+ "learning_rate": 0.0006,
+ "loss": 4.9436116218566895,
+ "step": 761
+ },
+ {
+ "epoch": 10.586899563318777,
+ "grad_norm": 0.012858862057328224,
+ "learning_rate": 0.0006,
+ "loss": 4.9595770835876465,
+ "step": 762
+ },
+ {
+ "epoch": 10.600873362445414,
+ "grad_norm": 0.01468253880739212,
+ "learning_rate": 0.0006,
+ "loss": 4.90725040435791,
+ "step": 763
+ },
+ {
+ "epoch": 10.614847161572053,
+ "grad_norm": 0.013224679045379162,
+ "learning_rate": 0.0006,
+ "loss": 4.877615928649902,
+ "step": 764
+ },
+ {
+ "epoch": 10.62882096069869,
+ "grad_norm": 0.013555348850786686,
+ "learning_rate": 0.0006,
+ "loss": 4.9456610679626465,
+ "step": 765
+ },
+ {
+ "epoch": 10.642794759825328,
+ "grad_norm": 0.014387160539627075,
+ "learning_rate": 0.0006,
+ "loss": 4.9233717918396,
+ "step": 766
+ },
+ {
+ "epoch": 10.656768558951965,
+ "grad_norm": 0.014354088343679905,
+ "learning_rate": 0.0006,
+ "loss": 4.983511924743652,
+ "step": 767
+ },
+ {
+ "epoch": 10.670742358078602,
+ "grad_norm": 0.01332700252532959,
+ "learning_rate": 0.0006,
+ "loss": 4.934416770935059,
+ "step": 768
+ },
+ {
+ "epoch": 10.68471615720524,
+ "grad_norm": 0.011543313041329384,
+ "learning_rate": 0.0006,
+ "loss": 4.96388053894043,
+ "step": 769
+ },
+ {
+ "epoch": 10.698689956331878,
+ "grad_norm": 0.012330878525972366,
+ "learning_rate": 0.0006,
+ "loss": 4.946396350860596,
+ "step": 770
+ },
+ {
+ "epoch": 10.712663755458514,
+ "grad_norm": 0.01379779726266861,
+ "learning_rate": 0.0006,
+ "loss": 4.981019020080566,
+ "step": 771
+ },
+ {
+ "epoch": 10.726637554585153,
+ "grad_norm": 0.014612431637942791,
+ "learning_rate": 0.0006,
+ "loss": 4.894567966461182,
+ "step": 772
+ },
+ {
+ "epoch": 10.74061135371179,
+ "grad_norm": 0.013062899932265282,
+ "learning_rate": 0.0006,
+ "loss": 5.033038139343262,
+ "step": 773
+ },
+ {
+ "epoch": 10.754585152838429,
+ "grad_norm": 0.013898049481213093,
+ "learning_rate": 0.0006,
+ "loss": 4.823636054992676,
+ "step": 774
+ },
+ {
+ "epoch": 10.768558951965066,
+ "grad_norm": 0.01502019353210926,
+ "learning_rate": 0.0006,
+ "loss": 5.020156383514404,
+ "step": 775
+ },
+ {
+ "epoch": 10.782532751091702,
+ "grad_norm": 0.013810204342007637,
+ "learning_rate": 0.0006,
+ "loss": 4.935647010803223,
+ "step": 776
+ },
+ {
+ "epoch": 10.796506550218341,
+ "grad_norm": 0.013616513460874557,
+ "learning_rate": 0.0006,
+ "loss": 4.901486396789551,
+ "step": 777
+ },
+ {
+ "epoch": 10.810480349344978,
+ "grad_norm": 0.013601024635136127,
+ "learning_rate": 0.0006,
+ "loss": 4.931873321533203,
+ "step": 778
+ },
+ {
+ "epoch": 10.824454148471617,
+ "grad_norm": 0.013668350875377655,
+ "learning_rate": 0.0006,
+ "loss": 4.937256813049316,
+ "step": 779
+ },
+ {
+ "epoch": 10.838427947598253,
+ "grad_norm": 0.01581823080778122,
+ "learning_rate": 0.0006,
+ "loss": 4.9611406326293945,
+ "step": 780
+ },
+ {
+ "epoch": 10.85240174672489,
+ "grad_norm": 0.019478484988212585,
+ "learning_rate": 0.0006,
+ "loss": 4.9212236404418945,
+ "step": 781
+ },
+ {
+ "epoch": 10.866375545851529,
+ "grad_norm": 0.026921333745121956,
+ "learning_rate": 0.0006,
+ "loss": 4.88501501083374,
+ "step": 782
+ },
+ {
+ "epoch": 10.880349344978166,
+ "grad_norm": 0.03155914694070816,
+ "learning_rate": 0.0006,
+ "loss": 4.989967346191406,
+ "step": 783
+ },
+ {
+ "epoch": 10.894323144104803,
+ "grad_norm": 0.029689429327845573,
+ "learning_rate": 0.0006,
+ "loss": 4.936000823974609,
+ "step": 784
+ },
+ {
+ "epoch": 10.908296943231441,
+ "grad_norm": 0.026626063510775566,
+ "learning_rate": 0.0006,
+ "loss": 4.984785079956055,
+ "step": 785
+ },
+ {
+ "epoch": 10.922270742358078,
+ "grad_norm": 0.026168212294578552,
+ "learning_rate": 0.0006,
+ "loss": 4.908282279968262,
+ "step": 786
+ },
+ {
+ "epoch": 10.936244541484717,
+ "grad_norm": 0.02828158251941204,
+ "learning_rate": 0.0006,
+ "loss": 4.926390647888184,
+ "step": 787
+ },
+ {
+ "epoch": 10.950218340611354,
+ "grad_norm": 0.02649078331887722,
+ "learning_rate": 0.0006,
+ "loss": 5.00400972366333,
+ "step": 788
+ },
+ {
+ "epoch": 10.96419213973799,
+ "grad_norm": 0.0256856270134449,
+ "learning_rate": 0.0006,
+ "loss": 5.0317487716674805,
+ "step": 789
+ },
+ {
+ "epoch": 10.97816593886463,
+ "grad_norm": 0.024728883057832718,
+ "learning_rate": 0.0006,
+ "loss": 4.9299635887146,
+ "step": 790
+ },
+ {
+ "epoch": 10.992139737991266,
+ "grad_norm": 0.026241116225719452,
+ "learning_rate": 0.0006,
+ "loss": 4.947319984436035,
+ "step": 791
+ },
+ {
+ "epoch": 11.0,
+ "grad_norm": 0.023526743054389954,
+ "learning_rate": 0.0006,
+ "loss": 4.94429349899292,
+ "step": 792
+ },
+ {
+ "epoch": 11.0,
+ "eval_loss": 5.104345798492432,
+ "eval_runtime": 57.0955,
+ "eval_samples_per_second": 42.77,
+ "eval_steps_per_second": 1.349,
+ "step": 792
+ },
+ {
+ "epoch": 11.013973799126637,
+ "grad_norm": 0.0216788612306118,
+ "learning_rate": 0.0006,
+ "loss": 4.881950855255127,
+ "step": 793
+ },
+ {
+ "epoch": 11.027947598253276,
+ "grad_norm": 0.01830463856458664,
+ "learning_rate": 0.0006,
+ "loss": 4.971016883850098,
+ "step": 794
+ },
+ {
+ "epoch": 11.041921397379912,
+ "grad_norm": 0.02204521745443344,
+ "learning_rate": 0.0006,
+ "loss": 4.894623756408691,
+ "step": 795
+ },
+ {
+ "epoch": 11.055895196506551,
+ "grad_norm": 0.0210255216807127,
+ "learning_rate": 0.0006,
+ "loss": 4.834009170532227,
+ "step": 796
+ },
+ {
+ "epoch": 11.069868995633188,
+ "grad_norm": 0.019797448068857193,
+ "learning_rate": 0.0006,
+ "loss": 5.086714744567871,
+ "step": 797
+ },
+ {
+ "epoch": 11.083842794759825,
+ "grad_norm": 0.02036641724407673,
+ "learning_rate": 0.0006,
+ "loss": 4.920056343078613,
+ "step": 798
+ },
+ {
+ "epoch": 11.097816593886463,
+ "grad_norm": 0.01966066285967827,
+ "learning_rate": 0.0006,
+ "loss": 4.879859447479248,
+ "step": 799
+ },
+ {
+ "epoch": 11.1117903930131,
+ "grad_norm": 0.0202149897813797,
+ "learning_rate": 0.0006,
+ "loss": 4.933640480041504,
+ "step": 800
+ },
+ {
+ "epoch": 11.125764192139737,
+ "grad_norm": 0.01893039606511593,
+ "learning_rate": 0.0006,
+ "loss": 4.901970863342285,
+ "step": 801
+ },
+ {
+ "epoch": 11.139737991266376,
+ "grad_norm": 0.016040155664086342,
+ "learning_rate": 0.0006,
+ "loss": 4.905491828918457,
+ "step": 802
+ },
+ {
+ "epoch": 11.153711790393013,
+ "grad_norm": 0.01720350608229637,
+ "learning_rate": 0.0006,
+ "loss": 4.798816680908203,
+ "step": 803
+ },
+ {
+ "epoch": 11.167685589519651,
+ "grad_norm": 0.017002852633595467,
+ "learning_rate": 0.0006,
+ "loss": 4.851859092712402,
+ "step": 804
+ },
+ {
+ "epoch": 11.181659388646288,
+ "grad_norm": 0.015594680793583393,
+ "learning_rate": 0.0006,
+ "loss": 4.90328311920166,
+ "step": 805
+ },
+ {
+ "epoch": 11.195633187772925,
+ "grad_norm": 0.014904462732374668,
+ "learning_rate": 0.0006,
+ "loss": 4.904215335845947,
+ "step": 806
+ },
+ {
+ "epoch": 11.209606986899564,
+ "grad_norm": 0.01366228237748146,
+ "learning_rate": 0.0006,
+ "loss": 4.9813055992126465,
+ "step": 807
+ },
+ {
+ "epoch": 11.2235807860262,
+ "grad_norm": 0.014719638973474503,
+ "learning_rate": 0.0006,
+ "loss": 4.927306175231934,
+ "step": 808
+ },
+ {
+ "epoch": 11.237554585152838,
+ "grad_norm": 0.015016036108136177,
+ "learning_rate": 0.0006,
+ "loss": 4.874215126037598,
+ "step": 809
+ },
+ {
+ "epoch": 11.251528384279476,
+ "grad_norm": 0.014925424009561539,
+ "learning_rate": 0.0006,
+ "loss": 4.838257789611816,
+ "step": 810
+ },
+ {
+ "epoch": 11.265502183406113,
+ "grad_norm": 0.014335823245346546,
+ "learning_rate": 0.0006,
+ "loss": 4.940608024597168,
+ "step": 811
+ },
+ {
+ "epoch": 11.279475982532752,
+ "grad_norm": 0.014987512491643429,
+ "learning_rate": 0.0006,
+ "loss": 4.831531524658203,
+ "step": 812
+ },
+ {
+ "epoch": 11.293449781659389,
+ "grad_norm": 0.016087457537651062,
+ "learning_rate": 0.0006,
+ "loss": 4.863227367401123,
+ "step": 813
+ },
+ {
+ "epoch": 11.307423580786025,
+ "grad_norm": 0.01671300269663334,
+ "learning_rate": 0.0006,
+ "loss": 4.8535661697387695,
+ "step": 814
+ },
+ {
+ "epoch": 11.321397379912664,
+ "grad_norm": 0.015229142270982265,
+ "learning_rate": 0.0006,
+ "loss": 4.948635101318359,
+ "step": 815
+ },
+ {
+ "epoch": 11.335371179039301,
+ "grad_norm": 0.01532587967813015,
+ "learning_rate": 0.0006,
+ "loss": 4.9035797119140625,
+ "step": 816
+ },
+ {
+ "epoch": 11.34934497816594,
+ "grad_norm": 0.015715476125478745,
+ "learning_rate": 0.0006,
+ "loss": 4.894709587097168,
+ "step": 817
+ },
+ {
+ "epoch": 11.363318777292577,
+ "grad_norm": 0.014156977646052837,
+ "learning_rate": 0.0006,
+ "loss": 4.960206985473633,
+ "step": 818
+ },
+ {
+ "epoch": 11.377292576419213,
+ "grad_norm": 0.017076566815376282,
+ "learning_rate": 0.0006,
+ "loss": 5.005850791931152,
+ "step": 819
+ },
+ {
+ "epoch": 11.391266375545852,
+ "grad_norm": 0.0172084029763937,
+ "learning_rate": 0.0006,
+ "loss": 4.903355598449707,
+ "step": 820
+ },
+ {
+ "epoch": 11.405240174672489,
+ "grad_norm": 0.0180258359760046,
+ "learning_rate": 0.0006,
+ "loss": 4.842667579650879,
+ "step": 821
+ },
+ {
+ "epoch": 11.419213973799126,
+ "grad_norm": 0.022130120545625687,
+ "learning_rate": 0.0006,
+ "loss": 4.799970626831055,
+ "step": 822
+ },
+ {
+ "epoch": 11.433187772925764,
+ "grad_norm": 0.02384897693991661,
+ "learning_rate": 0.0006,
+ "loss": 4.860746383666992,
+ "step": 823
+ },
+ {
+ "epoch": 11.447161572052401,
+ "grad_norm": 0.021881572902202606,
+ "learning_rate": 0.0006,
+ "loss": 4.918107032775879,
+ "step": 824
+ },
+ {
+ "epoch": 11.46113537117904,
+ "grad_norm": 0.023085080087184906,
+ "learning_rate": 0.0006,
+ "loss": 4.878946304321289,
+ "step": 825
+ },
+ {
+ "epoch": 11.475109170305677,
+ "grad_norm": 0.02291012555360794,
+ "learning_rate": 0.0006,
+ "loss": 4.958887100219727,
+ "step": 826
+ },
+ {
+ "epoch": 11.489082969432314,
+ "grad_norm": 0.0199968870729208,
+ "learning_rate": 0.0006,
+ "loss": 4.80955696105957,
+ "step": 827
+ },
+ {
+ "epoch": 11.503056768558952,
+ "grad_norm": 0.018702229484915733,
+ "learning_rate": 0.0006,
+ "loss": 5.0276408195495605,
+ "step": 828
+ },
+ {
+ "epoch": 11.51703056768559,
+ "grad_norm": 0.017787311226129532,
+ "learning_rate": 0.0006,
+ "loss": 4.96018123626709,
+ "step": 829
+ },
+ {
+ "epoch": 11.531004366812226,
+ "grad_norm": 0.01704501546919346,
+ "learning_rate": 0.0006,
+ "loss": 4.914553165435791,
+ "step": 830
+ },
+ {
+ "epoch": 11.544978165938865,
+ "grad_norm": 0.017519677057862282,
+ "learning_rate": 0.0006,
+ "loss": 4.798098087310791,
+ "step": 831
+ },
+ {
+ "epoch": 11.558951965065502,
+ "grad_norm": 0.018511991947889328,
+ "learning_rate": 0.0006,
+ "loss": 4.817785263061523,
+ "step": 832
+ },
+ {
+ "epoch": 11.57292576419214,
+ "grad_norm": 0.01876644790172577,
+ "learning_rate": 0.0006,
+ "loss": 4.818995475769043,
+ "step": 833
+ },
+ {
+ "epoch": 11.586899563318777,
+ "grad_norm": 0.01753568835556507,
+ "learning_rate": 0.0006,
+ "loss": 4.811938285827637,
+ "step": 834
+ },
+ {
+ "epoch": 11.600873362445414,
+ "grad_norm": 0.017046531662344933,
+ "learning_rate": 0.0006,
+ "loss": 4.914674282073975,
+ "step": 835
+ },
+ {
+ "epoch": 11.614847161572053,
+ "grad_norm": 0.01802094094455242,
+ "learning_rate": 0.0006,
+ "loss": 4.993284225463867,
+ "step": 836
+ },
+ {
+ "epoch": 11.62882096069869,
+ "grad_norm": 0.019493764266371727,
+ "learning_rate": 0.0006,
+ "loss": 4.895628452301025,
+ "step": 837
+ },
+ {
+ "epoch": 11.642794759825328,
+ "grad_norm": 0.019461210817098618,
+ "learning_rate": 0.0006,
+ "loss": 4.869269371032715,
+ "step": 838
+ },
+ {
+ "epoch": 11.656768558951965,
+ "grad_norm": 0.01981567218899727,
+ "learning_rate": 0.0006,
+ "loss": 4.863052845001221,
+ "step": 839
+ },
+ {
+ "epoch": 11.670742358078602,
+ "grad_norm": 0.019894491881132126,
+ "learning_rate": 0.0006,
+ "loss": 4.8775787353515625,
+ "step": 840
+ },
+ {
+ "epoch": 11.68471615720524,
+ "grad_norm": 0.01875786855816841,
+ "learning_rate": 0.0006,
+ "loss": 4.817763328552246,
+ "step": 841
+ },
+ {
+ "epoch": 11.698689956331878,
+ "grad_norm": 0.020926695317029953,
+ "learning_rate": 0.0006,
+ "loss": 4.7470526695251465,
+ "step": 842
+ },
+ {
+ "epoch": 11.712663755458514,
+ "grad_norm": 0.021477103233337402,
+ "learning_rate": 0.0006,
+ "loss": 4.762682914733887,
+ "step": 843
+ },
+ {
+ "epoch": 11.726637554585153,
+ "grad_norm": 0.020628711208701134,
+ "learning_rate": 0.0006,
+ "loss": 4.807985782623291,
+ "step": 844
+ },
+ {
+ "epoch": 11.74061135371179,
+ "grad_norm": 0.021489612758159637,
+ "learning_rate": 0.0006,
+ "loss": 4.843091011047363,
+ "step": 845
+ },
+ {
+ "epoch": 11.754585152838429,
+ "grad_norm": 0.021194491535425186,
+ "learning_rate": 0.0006,
+ "loss": 4.8300251960754395,
+ "step": 846
+ },
+ {
+ "epoch": 11.768558951965066,
+ "grad_norm": 0.01833350397646427,
+ "learning_rate": 0.0006,
+ "loss": 5.000202178955078,
+ "step": 847
+ },
+ {
+ "epoch": 11.782532751091702,
+ "grad_norm": 0.018096931278705597,
+ "learning_rate": 0.0006,
+ "loss": 4.880648136138916,
+ "step": 848
+ },
+ {
+ "epoch": 11.796506550218341,
+ "grad_norm": 0.0197161752730608,
+ "learning_rate": 0.0006,
+ "loss": 4.788316249847412,
+ "step": 849
+ },
+ {
+ "epoch": 11.810480349344978,
+ "grad_norm": 0.0203181654214859,
+ "learning_rate": 0.0006,
+ "loss": 4.870944976806641,
+ "step": 850
+ },
+ {
+ "epoch": 11.824454148471617,
+ "grad_norm": 0.01908455230295658,
+ "learning_rate": 0.0006,
+ "loss": 4.890144348144531,
+ "step": 851
+ },
+ {
+ "epoch": 11.838427947598253,
+ "grad_norm": 0.021281801164150238,
+ "learning_rate": 0.0006,
+ "loss": 4.89578914642334,
+ "step": 852
+ },
+ {
+ "epoch": 11.85240174672489,
+ "grad_norm": 0.02124273031949997,
+ "learning_rate": 0.0006,
+ "loss": 4.883749008178711,
+ "step": 853
+ },
+ {
+ "epoch": 11.866375545851529,
+ "grad_norm": 0.020969398319721222,
+ "learning_rate": 0.0006,
+ "loss": 4.809142589569092,
+ "step": 854
+ },
+ {
+ "epoch": 11.880349344978166,
+ "grad_norm": 0.01967657171189785,
+ "learning_rate": 0.0006,
+ "loss": 4.821864604949951,
+ "step": 855
+ },
+ {
+ "epoch": 11.894323144104803,
+ "grad_norm": 0.020887333899736404,
+ "learning_rate": 0.0006,
+ "loss": 4.917985916137695,
+ "step": 856
+ },
+ {
+ "epoch": 11.908296943231441,
+ "grad_norm": 0.019203083589673042,
+ "learning_rate": 0.0006,
+ "loss": 4.824960708618164,
+ "step": 857
+ },
+ {
+ "epoch": 11.922270742358078,
+ "grad_norm": 0.020162401720881462,
+ "learning_rate": 0.0006,
+ "loss": 4.772617816925049,
+ "step": 858
+ },
+ {
+ "epoch": 11.936244541484717,
+ "grad_norm": 0.02251487225294113,
+ "learning_rate": 0.0006,
+ "loss": 4.735418319702148,
+ "step": 859
+ },
+ {
+ "epoch": 11.950218340611354,
+ "grad_norm": 0.020313527435064316,
+ "learning_rate": 0.0006,
+ "loss": 4.939233779907227,
+ "step": 860
+ },
+ {
+ "epoch": 11.96419213973799,
+ "grad_norm": 0.021635599434375763,
+ "learning_rate": 0.0006,
+ "loss": 4.861807346343994,
+ "step": 861
+ },
+ {
+ "epoch": 11.97816593886463,
+ "grad_norm": 0.02285311557352543,
+ "learning_rate": 0.0006,
+ "loss": 4.932732582092285,
+ "step": 862
+ },
+ {
+ "epoch": 11.992139737991266,
+ "grad_norm": 0.02072535641491413,
+ "learning_rate": 0.0006,
+ "loss": 4.796767711639404,
+ "step": 863
+ },
+ {
+ "epoch": 12.0,
+ "grad_norm": 0.020588304847478867,
+ "learning_rate": 0.0006,
+ "loss": 4.997537612915039,
+ "step": 864
+ },
+ {
+ "epoch": 12.0,
+ "eval_loss": 5.044651985168457,
+ "eval_runtime": 56.9145,
+ "eval_samples_per_second": 42.906,
+ "eval_steps_per_second": 1.353,
+ "step": 864
+ },
+ {
+ "epoch": 12.013973799126637,
+ "grad_norm": 0.017597166821360588,
+ "learning_rate": 0.0006,
+ "loss": 4.830524444580078,
+ "step": 865
+ },
+ {
+ "epoch": 12.027947598253276,
+ "grad_norm": 0.01716788113117218,
+ "learning_rate": 0.0006,
+ "loss": 4.894325256347656,
+ "step": 866
+ },
+ {
+ "epoch": 12.041921397379912,
+ "grad_norm": 0.017712457105517387,
+ "learning_rate": 0.0006,
+ "loss": 4.868297576904297,
+ "step": 867
+ },
+ {
+ "epoch": 12.055895196506551,
+ "grad_norm": 0.021486390382051468,
+ "learning_rate": 0.0006,
+ "loss": 4.922049522399902,
+ "step": 868
+ },
+ {
+ "epoch": 12.069868995633188,
+ "grad_norm": 0.024560654535889626,
+ "learning_rate": 0.0006,
+ "loss": 4.734241008758545,
+ "step": 869
+ },
+ {
+ "epoch": 12.083842794759825,
+ "grad_norm": 0.025870252400636673,
+ "learning_rate": 0.0006,
+ "loss": 4.868566989898682,
+ "step": 870
+ },
+ {
+ "epoch": 12.097816593886463,
+ "grad_norm": 0.0277389045804739,
+ "learning_rate": 0.0006,
+ "loss": 4.755006790161133,
+ "step": 871
+ },
+ {
+ "epoch": 12.1117903930131,
+ "grad_norm": 0.029100844636559486,
+ "learning_rate": 0.0006,
+ "loss": 4.909667015075684,
+ "step": 872
+ },
+ {
+ "epoch": 12.125764192139737,
+ "grad_norm": 0.031833067536354065,
+ "learning_rate": 0.0006,
+ "loss": 4.858674049377441,
+ "step": 873
+ },
+ {
+ "epoch": 12.139737991266376,
+ "grad_norm": 0.032709237188100815,
+ "learning_rate": 0.0006,
+ "loss": 4.823696136474609,
+ "step": 874
+ },
+ {
+ "epoch": 12.153711790393013,
+ "grad_norm": 0.02995380386710167,
+ "learning_rate": 0.0006,
+ "loss": 4.850235462188721,
+ "step": 875
+ },
+ {
+ "epoch": 12.167685589519651,
+ "grad_norm": 0.034098610281944275,
+ "learning_rate": 0.0006,
+ "loss": 4.8169989585876465,
+ "step": 876
+ },
+ {
+ "epoch": 12.181659388646288,
+ "grad_norm": 0.03377080336213112,
+ "learning_rate": 0.0006,
+ "loss": 4.90738582611084,
+ "step": 877
+ },
+ {
+ "epoch": 12.195633187772925,
+ "grad_norm": 0.029780825600028038,
+ "learning_rate": 0.0006,
+ "loss": 4.811163902282715,
+ "step": 878
+ },
+ {
+ "epoch": 12.209606986899564,
+ "grad_norm": 0.03118024580180645,
+ "learning_rate": 0.0006,
+ "loss": 4.861286640167236,
+ "step": 879
+ },
+ {
+ "epoch": 12.2235807860262,
+ "grad_norm": 0.033052004873752594,
+ "learning_rate": 0.0006,
+ "loss": 4.865021228790283,
+ "step": 880
+ },
+ {
+ "epoch": 12.237554585152838,
+ "grad_norm": 0.029070032760500908,
+ "learning_rate": 0.0006,
+ "loss": 4.947303771972656,
+ "step": 881
+ },
+ {
+ "epoch": 12.251528384279476,
+ "grad_norm": 0.028355715796351433,
+ "learning_rate": 0.0006,
+ "loss": 4.804648399353027,
+ "step": 882
+ },
+ {
+ "epoch": 12.265502183406113,
+ "grad_norm": 0.031257227063179016,
+ "learning_rate": 0.0006,
+ "loss": 4.757718086242676,
+ "step": 883
+ },
+ {
+ "epoch": 12.279475982532752,
+ "grad_norm": 0.031438373029232025,
+ "learning_rate": 0.0006,
+ "loss": 4.851268768310547,
+ "step": 884
+ },
+ {
+ "epoch": 12.293449781659389,
+ "grad_norm": 0.030492138117551804,
+ "learning_rate": 0.0006,
+ "loss": 4.7891035079956055,
+ "step": 885
+ },
+ {
+ "epoch": 12.307423580786025,
+ "grad_norm": 0.026115991175174713,
+ "learning_rate": 0.0006,
+ "loss": 4.86474609375,
+ "step": 886
+ },
+ {
+ "epoch": 12.321397379912664,
+ "grad_norm": 0.024184465408325195,
+ "learning_rate": 0.0006,
+ "loss": 4.860456943511963,
+ "step": 887
+ },
+ {
+ "epoch": 12.335371179039301,
+ "grad_norm": 0.025437982752919197,
+ "learning_rate": 0.0006,
+ "loss": 4.806251049041748,
+ "step": 888
+ },
+ {
+ "epoch": 12.34934497816594,
+ "grad_norm": 0.022390231490135193,
+ "learning_rate": 0.0006,
+ "loss": 4.807086944580078,
+ "step": 889
+ },
+ {
+ "epoch": 12.363318777292577,
+ "grad_norm": 0.020174294710159302,
+ "learning_rate": 0.0006,
+ "loss": 4.78292179107666,
+ "step": 890
+ },
+ {
+ "epoch": 12.377292576419213,
+ "grad_norm": 0.019734324887394905,
+ "learning_rate": 0.0006,
+ "loss": 4.784509658813477,
+ "step": 891
+ },
+ {
+ "epoch": 12.391266375545852,
+ "grad_norm": 0.016733523458242416,
+ "learning_rate": 0.0006,
+ "loss": 4.811755180358887,
+ "step": 892
+ },
+ {
+ "epoch": 12.405240174672489,
+ "grad_norm": 0.01570146717131138,
+ "learning_rate": 0.0006,
+ "loss": 4.720088005065918,
+ "step": 893
+ },
+ {
+ "epoch": 12.419213973799126,
+ "grad_norm": 0.01601393334567547,
+ "learning_rate": 0.0006,
+ "loss": 4.783543109893799,
+ "step": 894
+ },
+ {
+ "epoch": 12.433187772925764,
+ "grad_norm": 0.014035487547516823,
+ "learning_rate": 0.0006,
+ "loss": 4.848340034484863,
+ "step": 895
+ },
+ {
+ "epoch": 12.447161572052401,
+ "grad_norm": 0.015248487703502178,
+ "learning_rate": 0.0006,
+ "loss": 4.773287773132324,
+ "step": 896
+ },
+ {
+ "epoch": 12.46113537117904,
+ "grad_norm": 0.01532268151640892,
+ "learning_rate": 0.0006,
+ "loss": 4.799646854400635,
+ "step": 897
+ },
+ {
+ "epoch": 12.475109170305677,
+ "grad_norm": 0.013982534408569336,
+ "learning_rate": 0.0006,
+ "loss": 4.8090500831604,
+ "step": 898
+ },
+ {
+ "epoch": 12.489082969432314,
+ "grad_norm": 0.014081164263188839,
+ "learning_rate": 0.0006,
+ "loss": 4.751888751983643,
+ "step": 899
+ },
+ {
+ "epoch": 12.503056768558952,
+ "grad_norm": 0.01380133256316185,
+ "learning_rate": 0.0006,
+ "loss": 4.78144645690918,
+ "step": 900
+ },
+ {
+ "epoch": 12.51703056768559,
+ "grad_norm": 0.013555269688367844,
+ "learning_rate": 0.0006,
+ "loss": 4.779257774353027,
+ "step": 901
+ },
+ {
+ "epoch": 12.531004366812226,
+ "grad_norm": 0.013075701892375946,
+ "learning_rate": 0.0006,
+ "loss": 4.820705413818359,
+ "step": 902
+ },
+ {
+ "epoch": 12.544978165938865,
+ "grad_norm": 0.01276316400617361,
+ "learning_rate": 0.0006,
+ "loss": 4.836733818054199,
+ "step": 903
+ },
+ {
+ "epoch": 12.558951965065502,
+ "grad_norm": 0.013033892959356308,
+ "learning_rate": 0.0006,
+ "loss": 4.764178276062012,
+ "step": 904
+ },
+ {
+ "epoch": 12.57292576419214,
+ "grad_norm": 0.013996980153024197,
+ "learning_rate": 0.0006,
+ "loss": 4.767983436584473,
+ "step": 905
+ },
+ {
+ "epoch": 12.586899563318777,
+ "grad_norm": 0.015222184360027313,
+ "learning_rate": 0.0006,
+ "loss": 4.759721755981445,
+ "step": 906
+ },
+ {
+ "epoch": 12.600873362445414,
+ "grad_norm": 0.01623941957950592,
+ "learning_rate": 0.0006,
+ "loss": 4.849811553955078,
+ "step": 907
+ },
+ {
+ "epoch": 12.614847161572053,
+ "grad_norm": 0.015711182728409767,
+ "learning_rate": 0.0006,
+ "loss": 4.7829084396362305,
+ "step": 908
+ },
+ {
+ "epoch": 12.62882096069869,
+ "grad_norm": 0.014137467369437218,
+ "learning_rate": 0.0006,
+ "loss": 4.73321533203125,
+ "step": 909
+ },
+ {
+ "epoch": 12.642794759825328,
+ "grad_norm": 0.012738462537527084,
+ "learning_rate": 0.0006,
+ "loss": 4.870052337646484,
+ "step": 910
+ },
+ {
+ "epoch": 12.656768558951965,
+ "grad_norm": 0.013683130033314228,
+ "learning_rate": 0.0006,
+ "loss": 4.82658052444458,
+ "step": 911
+ },
+ {
+ "epoch": 12.670742358078602,
+ "grad_norm": 0.014409830793738365,
+ "learning_rate": 0.0006,
+ "loss": 4.707494735717773,
+ "step": 912
+ },
+ {
+ "epoch": 12.68471615720524,
+ "grad_norm": 0.013806809671223164,
+ "learning_rate": 0.0006,
+ "loss": 4.778928279876709,
+ "step": 913
+ },
+ {
+ "epoch": 12.698689956331878,
+ "grad_norm": 0.012857912108302116,
+ "learning_rate": 0.0006,
+ "loss": 4.890717506408691,
+ "step": 914
+ },
+ {
+ "epoch": 12.712663755458514,
+ "grad_norm": 0.013828758150339127,
+ "learning_rate": 0.0006,
+ "loss": 4.8357439041137695,
+ "step": 915
+ },
+ {
+ "epoch": 12.726637554585153,
+ "grad_norm": 0.01478351280093193,
+ "learning_rate": 0.0006,
+ "loss": 4.869986534118652,
+ "step": 916
+ },
+ {
+ "epoch": 12.74061135371179,
+ "grad_norm": 0.01326421182602644,
+ "learning_rate": 0.0006,
+ "loss": 4.830646514892578,
+ "step": 917
+ },
+ {
+ "epoch": 12.754585152838429,
+ "grad_norm": 0.013335433788597584,
+ "learning_rate": 0.0006,
+ "loss": 4.7786149978637695,
+ "step": 918
+ },
+ {
+ "epoch": 12.768558951965066,
+ "grad_norm": 0.014261147007346153,
+ "learning_rate": 0.0006,
+ "loss": 4.788693428039551,
+ "step": 919
+ },
+ {
+ "epoch": 12.782532751091702,
+ "grad_norm": 0.014274783432483673,
+ "learning_rate": 0.0006,
+ "loss": 4.77295446395874,
+ "step": 920
+ },
+ {
+ "epoch": 12.796506550218341,
+ "grad_norm": 0.015074674971401691,
+ "learning_rate": 0.0006,
+ "loss": 4.830754280090332,
+ "step": 921
+ },
+ {
+ "epoch": 12.810480349344978,
+ "grad_norm": 0.015599515289068222,
+ "learning_rate": 0.0006,
+ "loss": 4.837759017944336,
+ "step": 922
+ },
+ {
+ "epoch": 12.824454148471617,
+ "grad_norm": 0.014595242217183113,
+ "learning_rate": 0.0006,
+ "loss": 4.882965564727783,
+ "step": 923
+ },
+ {
+ "epoch": 12.838427947598253,
+ "grad_norm": 0.014951467514038086,
+ "learning_rate": 0.0006,
+ "loss": 4.709748268127441,
+ "step": 924
+ },
+ {
+ "epoch": 12.85240174672489,
+ "grad_norm": 0.01539295818656683,
+ "learning_rate": 0.0006,
+ "loss": 4.791160583496094,
+ "step": 925
+ },
+ {
+ "epoch": 12.866375545851529,
+ "grad_norm": 0.015987906605005264,
+ "learning_rate": 0.0006,
+ "loss": 4.661876678466797,
+ "step": 926
+ },
+ {
+ "epoch": 12.880349344978166,
+ "grad_norm": 0.016527341678738594,
+ "learning_rate": 0.0006,
+ "loss": 4.721952438354492,
+ "step": 927
+ },
+ {
+ "epoch": 12.894323144104803,
+ "grad_norm": 0.01637161523103714,
+ "learning_rate": 0.0006,
+ "loss": 4.704446792602539,
+ "step": 928
+ },
+ {
+ "epoch": 12.908296943231441,
+ "grad_norm": 0.015221050009131432,
+ "learning_rate": 0.0006,
+ "loss": 4.771907806396484,
+ "step": 929
+ },
+ {
+ "epoch": 12.922270742358078,
+ "grad_norm": 0.016242166981101036,
+ "learning_rate": 0.0006,
+ "loss": 4.806267738342285,
+ "step": 930
+ },
+ {
+ "epoch": 12.936244541484717,
+ "grad_norm": 0.01628357358276844,
+ "learning_rate": 0.0006,
+ "loss": 4.7252326011657715,
+ "step": 931
+ },
+ {
+ "epoch": 12.950218340611354,
+ "grad_norm": 0.017428133636713028,
+ "learning_rate": 0.0006,
+ "loss": 4.765689849853516,
+ "step": 932
+ },
+ {
+ "epoch": 12.96419213973799,
+ "grad_norm": 0.01870061084628105,
+ "learning_rate": 0.0006,
+ "loss": 4.814488887786865,
+ "step": 933
+ },
+ {
+ "epoch": 12.97816593886463,
+ "grad_norm": 0.019796060398221016,
+ "learning_rate": 0.0006,
+ "loss": 4.739185810089111,
+ "step": 934
+ },
+ {
+ "epoch": 12.992139737991266,
+ "grad_norm": 0.01985820196568966,
+ "learning_rate": 0.0006,
+ "loss": 4.810608863830566,
+ "step": 935
+ },
+ {
+ "epoch": 13.0,
+ "grad_norm": 0.020834477618336678,
+ "learning_rate": 0.0006,
+ "loss": 4.727858066558838,
+ "step": 936
+ },
+ {
+ "epoch": 13.0,
+ "eval_loss": 4.979116916656494,
+ "eval_runtime": 57.2877,
+ "eval_samples_per_second": 42.627,
+ "eval_steps_per_second": 1.344,
+ "step": 936
+ },
+ {
+ "epoch": 13.013973799126637,
+ "grad_norm": 0.02025519125163555,
+ "learning_rate": 0.0006,
+ "loss": 4.628875732421875,
+ "step": 937
+ },
+ {
+ "epoch": 13.027947598253276,
+ "grad_norm": 0.019596368074417114,
+ "learning_rate": 0.0006,
+ "loss": 4.694699287414551,
+ "step": 938
+ },
+ {
+ "epoch": 13.041921397379912,
+ "grad_norm": 0.020494934171438217,
+ "learning_rate": 0.0006,
+ "loss": 4.691821098327637,
+ "step": 939
+ },
+ {
+ "epoch": 13.055895196506551,
+ "grad_norm": 0.020121760666370392,
+ "learning_rate": 0.0006,
+ "loss": 4.71478271484375,
+ "step": 940
+ },
+ {
+ "epoch": 13.069868995633188,
+ "grad_norm": 0.021091405302286148,
+ "learning_rate": 0.0006,
+ "loss": 4.695915222167969,
+ "step": 941
+ },
+ {
+ "epoch": 13.083842794759825,
+ "grad_norm": 0.019787423312664032,
+ "learning_rate": 0.0006,
+ "loss": 4.8652448654174805,
+ "step": 942
+ },
+ {
+ "epoch": 13.097816593886463,
+ "grad_norm": 0.0185833927243948,
+ "learning_rate": 0.0006,
+ "loss": 4.7065887451171875,
+ "step": 943
+ },
+ {
+ "epoch": 13.1117903930131,
+ "grad_norm": 0.021220475435256958,
+ "learning_rate": 0.0006,
+ "loss": 4.741232395172119,
+ "step": 944
+ },
+ {
+ "epoch": 13.125764192139737,
+ "grad_norm": 0.02136622555553913,
+ "learning_rate": 0.0006,
+ "loss": 4.705967903137207,
+ "step": 945
+ },
+ {
+ "epoch": 13.139737991266376,
+ "grad_norm": 0.0191587433218956,
+ "learning_rate": 0.0006,
+ "loss": 4.758667945861816,
+ "step": 946
+ },
+ {
+ "epoch": 13.153711790393013,
+ "grad_norm": 0.018290022388100624,
+ "learning_rate": 0.0006,
+ "loss": 4.760666847229004,
+ "step": 947
+ },
+ {
+ "epoch": 13.167685589519651,
+ "grad_norm": 0.021704409271478653,
+ "learning_rate": 0.0006,
+ "loss": 4.750555992126465,
+ "step": 948
+ },
+ {
+ "epoch": 13.181659388646288,
+ "grad_norm": 0.021668918430805206,
+ "learning_rate": 0.0006,
+ "loss": 4.840903282165527,
+ "step": 949
+ },
+ {
+ "epoch": 13.195633187772925,
+ "grad_norm": 0.019497141242027283,
+ "learning_rate": 0.0006,
+ "loss": 4.738969326019287,
+ "step": 950
+ },
+ {
+ "epoch": 13.209606986899564,
+ "grad_norm": 0.018500596284866333,
+ "learning_rate": 0.0006,
+ "loss": 4.75154972076416,
+ "step": 951
+ },
+ {
+ "epoch": 13.2235807860262,
+ "grad_norm": 0.01872239261865616,
+ "learning_rate": 0.0006,
+ "loss": 4.811738014221191,
+ "step": 952
+ },
+ {
+ "epoch": 13.237554585152838,
+ "grad_norm": 0.020321547985076904,
+ "learning_rate": 0.0006,
+ "loss": 4.815384387969971,
+ "step": 953
+ },
+ {
+ "epoch": 13.251528384279476,
+ "grad_norm": 0.02283575013279915,
+ "learning_rate": 0.0006,
+ "loss": 4.74582052230835,
+ "step": 954
+ },
+ {
+ "epoch": 13.265502183406113,
+ "grad_norm": 0.02474398910999298,
+ "learning_rate": 0.0006,
+ "loss": 4.738375186920166,
+ "step": 955
+ },
+ {
+ "epoch": 13.279475982532752,
+ "grad_norm": 0.02065764181315899,
+ "learning_rate": 0.0006,
+ "loss": 4.746875286102295,
+ "step": 956
+ },
+ {
+ "epoch": 13.293449781659389,
+ "grad_norm": 0.020534520968794823,
+ "learning_rate": 0.0006,
+ "loss": 4.656914710998535,
+ "step": 957
+ },
+ {
+ "epoch": 13.307423580786025,
+ "grad_norm": 0.0206378772854805,
+ "learning_rate": 0.0006,
+ "loss": 4.842615604400635,
+ "step": 958
+ },
+ {
+ "epoch": 13.321397379912664,
+ "grad_norm": 0.017935654148459435,
+ "learning_rate": 0.0006,
+ "loss": 4.717827796936035,
+ "step": 959
+ },
+ {
+ "epoch": 13.335371179039301,
+ "grad_norm": 0.016958363354206085,
+ "learning_rate": 0.0006,
+ "loss": 4.681070327758789,
+ "step": 960
+ },
+ {
+ "epoch": 13.34934497816594,
+ "grad_norm": 0.015735412016510963,
+ "learning_rate": 0.0006,
+ "loss": 4.7957048416137695,
+ "step": 961
+ },
+ {
+ "epoch": 13.363318777292577,
+ "grad_norm": 0.01563699170947075,
+ "learning_rate": 0.0006,
+ "loss": 4.701192378997803,
+ "step": 962
+ },
+ {
+ "epoch": 13.377292576419213,
+ "grad_norm": 0.015207760035991669,
+ "learning_rate": 0.0006,
+ "loss": 4.73256254196167,
+ "step": 963
+ },
+ {
+ "epoch": 13.391266375545852,
+ "grad_norm": 0.015396883711218834,
+ "learning_rate": 0.0006,
+ "loss": 4.76412296295166,
+ "step": 964
+ },
+ {
+ "epoch": 13.405240174672489,
+ "grad_norm": 0.015690583735704422,
+ "learning_rate": 0.0006,
+ "loss": 4.791157245635986,
+ "step": 965
+ },
+ {
+ "epoch": 13.419213973799126,
+ "grad_norm": 0.01755724661052227,
+ "learning_rate": 0.0006,
+ "loss": 4.845666885375977,
+ "step": 966
+ },
+ {
+ "epoch": 13.433187772925764,
+ "grad_norm": 0.019443422555923462,
+ "learning_rate": 0.0006,
+ "loss": 4.5902299880981445,
+ "step": 967
+ },
+ {
+ "epoch": 13.447161572052401,
+ "grad_norm": 0.02118338644504547,
+ "learning_rate": 0.0006,
+ "loss": 4.740627288818359,
+ "step": 968
+ },
+ {
+ "epoch": 13.46113537117904,
+ "grad_norm": 0.02174839936196804,
+ "learning_rate": 0.0006,
+ "loss": 4.814560890197754,
+ "step": 969
+ },
+ {
+ "epoch": 13.475109170305677,
+ "grad_norm": 0.022860374301671982,
+ "learning_rate": 0.0006,
+ "loss": 4.66440486907959,
+ "step": 970
+ },
+ {
+ "epoch": 13.489082969432314,
+ "grad_norm": 0.024742677807807922,
+ "learning_rate": 0.0006,
+ "loss": 4.7451653480529785,
+ "step": 971
+ },
+ {
+ "epoch": 13.503056768558952,
+ "grad_norm": 0.027068907395005226,
+ "learning_rate": 0.0006,
+ "loss": 4.802584171295166,
+ "step": 972
+ },
+ {
+ "epoch": 13.51703056768559,
+ "grad_norm": 0.026316625997424126,
+ "learning_rate": 0.0006,
+ "loss": 4.842397689819336,
+ "step": 973
+ },
+ {
+ "epoch": 13.531004366812226,
+ "grad_norm": 0.024610962718725204,
+ "learning_rate": 0.0006,
+ "loss": 4.650721073150635,
+ "step": 974
+ },
+ {
+ "epoch": 13.544978165938865,
+ "grad_norm": 0.024122603237628937,
+ "learning_rate": 0.0006,
+ "loss": 4.6984357833862305,
+ "step": 975
+ },
+ {
+ "epoch": 13.558951965065502,
+ "grad_norm": 0.022699708119034767,
+ "learning_rate": 0.0006,
+ "loss": 4.793007850646973,
+ "step": 976
+ },
+ {
+ "epoch": 13.57292576419214,
+ "grad_norm": 0.021706534549593925,
+ "learning_rate": 0.0006,
+ "loss": 4.691362380981445,
+ "step": 977
+ },
+ {
+ "epoch": 13.586899563318777,
+ "grad_norm": 0.021363291889429092,
+ "learning_rate": 0.0006,
+ "loss": 4.76762580871582,
+ "step": 978
+ },
+ {
+ "epoch": 13.600873362445414,
+ "grad_norm": 0.023476677015423775,
+ "learning_rate": 0.0006,
+ "loss": 4.777202606201172,
+ "step": 979
+ },
+ {
+ "epoch": 13.614847161572053,
+ "grad_norm": 0.02457100711762905,
+ "learning_rate": 0.0006,
+ "loss": 4.590703010559082,
+ "step": 980
+ },
+ {
+ "epoch": 13.62882096069869,
+ "grad_norm": 0.02469099499285221,
+ "learning_rate": 0.0006,
+ "loss": 4.640185356140137,
+ "step": 981
+ },
+ {
+ "epoch": 13.642794759825328,
+ "grad_norm": 0.024914603680372238,
+ "learning_rate": 0.0006,
+ "loss": 4.772201061248779,
+ "step": 982
+ },
+ {
+ "epoch": 13.656768558951965,
+ "grad_norm": 0.025133173912763596,
+ "learning_rate": 0.0006,
+ "loss": 4.743337154388428,
+ "step": 983
+ },
+ {
+ "epoch": 13.670742358078602,
+ "grad_norm": 0.02034137211740017,
+ "learning_rate": 0.0006,
+ "loss": 4.752127647399902,
+ "step": 984
+ },
+ {
+ "epoch": 13.68471615720524,
+ "grad_norm": 0.019645661115646362,
+ "learning_rate": 0.0006,
+ "loss": 4.838453769683838,
+ "step": 985
+ },
+ {
+ "epoch": 13.698689956331878,
+ "grad_norm": 0.01900297962129116,
+ "learning_rate": 0.0006,
+ "loss": 4.804868221282959,
+ "step": 986
+ },
+ {
+ "epoch": 13.712663755458514,
+ "grad_norm": 0.020292341709136963,
+ "learning_rate": 0.0006,
+ "loss": 4.609743118286133,
+ "step": 987
+ },
+ {
+ "epoch": 13.726637554585153,
+ "grad_norm": 0.017014775425195694,
+ "learning_rate": 0.0006,
+ "loss": 4.740612983703613,
+ "step": 988
+ },
+ {
+ "epoch": 13.74061135371179,
+ "grad_norm": 0.015199844725430012,
+ "learning_rate": 0.0006,
+ "loss": 4.719701766967773,
+ "step": 989
+ },
+ {
+ "epoch": 13.754585152838429,
+ "grad_norm": 0.015125414356589317,
+ "learning_rate": 0.0006,
+ "loss": 4.686114311218262,
+ "step": 990
+ },
+ {
+ "epoch": 13.768558951965066,
+ "grad_norm": 0.014779540710151196,
+ "learning_rate": 0.0006,
+ "loss": 4.674347877502441,
+ "step": 991
+ },
+ {
+ "epoch": 13.782532751091702,
+ "grad_norm": 0.015255163423717022,
+ "learning_rate": 0.0006,
+ "loss": 4.6309309005737305,
+ "step": 992
+ },
+ {
+ "epoch": 13.796506550218341,
+ "grad_norm": 0.013692197389900684,
+ "learning_rate": 0.0006,
+ "loss": 4.698497295379639,
+ "step": 993
+ },
+ {
+ "epoch": 13.810480349344978,
+ "grad_norm": 0.014529623091220856,
+ "learning_rate": 0.0006,
+ "loss": 4.628443717956543,
+ "step": 994
+ },
+ {
+ "epoch": 13.824454148471617,
+ "grad_norm": 0.01576659083366394,
+ "learning_rate": 0.0006,
+ "loss": 4.788015842437744,
+ "step": 995
+ },
+ {
+ "epoch": 13.838427947598253,
+ "grad_norm": 0.015834277495741844,
+ "learning_rate": 0.0006,
+ "loss": 4.672703742980957,
+ "step": 996
+ },
+ {
+ "epoch": 13.85240174672489,
+ "grad_norm": 0.015539503656327724,
+ "learning_rate": 0.0006,
+ "loss": 4.731503486633301,
+ "step": 997
+ },
+ {
+ "epoch": 13.866375545851529,
+ "grad_norm": 0.013463746756315231,
+ "learning_rate": 0.0006,
+ "loss": 4.7006330490112305,
+ "step": 998
+ },
+ {
+ "epoch": 13.880349344978166,
+ "grad_norm": 0.016377726569771767,
+ "learning_rate": 0.0006,
+ "loss": 4.636074066162109,
+ "step": 999
+ },
+ {
+ "epoch": 13.894323144104803,
+ "grad_norm": 0.016660034656524658,
+ "learning_rate": 0.0006,
+ "loss": 4.742693901062012,
+ "step": 1000
+ },
+ {
+ "epoch": 13.908296943231441,
+ "grad_norm": 0.014902938157320023,
+ "learning_rate": 0.0006,
+ "loss": 4.769811630249023,
+ "step": 1001
+ },
+ {
+ "epoch": 13.922270742358078,
+ "grad_norm": 0.014883476309478283,
+ "learning_rate": 0.0006,
+ "loss": 4.731159210205078,
+ "step": 1002
+ },
+ {
+ "epoch": 13.936244541484717,
+ "grad_norm": 0.014494217932224274,
+ "learning_rate": 0.0006,
+ "loss": 4.659485816955566,
+ "step": 1003
+ },
+ {
+ "epoch": 13.950218340611354,
+ "grad_norm": 0.015426991507411003,
+ "learning_rate": 0.0006,
+ "loss": 4.755229949951172,
+ "step": 1004
+ },
+ {
+ "epoch": 13.96419213973799,
+ "grad_norm": 0.016648586839437485,
+ "learning_rate": 0.0006,
+ "loss": 4.74111270904541,
+ "step": 1005
+ },
+ {
+ "epoch": 13.97816593886463,
+ "grad_norm": 0.016980541869997978,
+ "learning_rate": 0.0006,
+ "loss": 4.74072265625,
+ "step": 1006
+ },
+ {
+ "epoch": 13.992139737991266,
+ "grad_norm": 0.018070152029395103,
+ "learning_rate": 0.0006,
+ "loss": 4.732975006103516,
+ "step": 1007
+ },
+ {
+ "epoch": 14.0,
+ "grad_norm": 0.019910665228962898,
+ "learning_rate": 0.0006,
+ "loss": 4.634697914123535,
+ "step": 1008
+ },
+ {
+ "epoch": 14.0,
+ "eval_loss": 4.850640296936035,
+ "eval_runtime": 57.3039,
+ "eval_samples_per_second": 42.615,
+ "eval_steps_per_second": 1.344,
+ "step": 1008
+ },
+ {
+ "epoch": 14.013973799126637,
+ "grad_norm": 0.01980053074657917,
+ "learning_rate": 0.0006,
+ "loss": 4.628744602203369,
+ "step": 1009
+ },
+ {
+ "epoch": 14.027947598253276,
+ "grad_norm": 0.02497623674571514,
+ "learning_rate": 0.0006,
+ "loss": 4.75930118560791,
+ "step": 1010
+ },
+ {
+ "epoch": 14.041921397379912,
+ "grad_norm": 0.02794097363948822,
+ "learning_rate": 0.0006,
+ "loss": 4.691340446472168,
+ "step": 1011
+ },
+ {
+ "epoch": 14.055895196506551,
+ "grad_norm": 0.023725535720586777,
+ "learning_rate": 0.0006,
+ "loss": 4.733664512634277,
+ "step": 1012
+ },
+ {
+ "epoch": 14.069868995633188,
+ "grad_norm": 0.022966474294662476,
+ "learning_rate": 0.0006,
+ "loss": 4.664808750152588,
+ "step": 1013
+ },
+ {
+ "epoch": 14.083842794759825,
+ "grad_norm": 0.02530563250184059,
+ "learning_rate": 0.0006,
+ "loss": 4.8047285079956055,
+ "step": 1014
+ },
+ {
+ "epoch": 14.097816593886463,
+ "grad_norm": 0.02491082064807415,
+ "learning_rate": 0.0006,
+ "loss": 4.667956352233887,
+ "step": 1015
+ },
+ {
+ "epoch": 14.1117903930131,
+ "grad_norm": 0.022110294550657272,
+ "learning_rate": 0.0006,
+ "loss": 4.666140556335449,
+ "step": 1016
+ },
+ {
+ "epoch": 14.125764192139737,
+ "grad_norm": 0.02065849117934704,
+ "learning_rate": 0.0006,
+ "loss": 4.581960678100586,
+ "step": 1017
+ },
+ {
+ "epoch": 14.139737991266376,
+ "grad_norm": 0.01941298507153988,
+ "learning_rate": 0.0006,
+ "loss": 4.614226341247559,
+ "step": 1018
+ },
+ {
+ "epoch": 14.153711790393013,
+ "grad_norm": 0.022441856563091278,
+ "learning_rate": 0.0006,
+ "loss": 4.638352394104004,
+ "step": 1019
+ },
+ {
+ "epoch": 14.167685589519651,
+ "grad_norm": 0.021519694477319717,
+ "learning_rate": 0.0006,
+ "loss": 4.700160980224609,
+ "step": 1020
+ },
+ {
+ "epoch": 14.181659388646288,
+ "grad_norm": 0.018956486135721207,
+ "learning_rate": 0.0006,
+ "loss": 4.649593353271484,
+ "step": 1021
+ },
+ {
+ "epoch": 14.195633187772925,
+ "grad_norm": 0.016681140288710594,
+ "learning_rate": 0.0006,
+ "loss": 4.603952407836914,
+ "step": 1022
+ },
+ {
+ "epoch": 14.209606986899564,
+ "grad_norm": 0.01788809522986412,
+ "learning_rate": 0.0006,
+ "loss": 4.669160842895508,
+ "step": 1023
+ },
+ {
+ "epoch": 14.2235807860262,
+ "grad_norm": 0.016047228127717972,
+ "learning_rate": 0.0006,
+ "loss": 4.6367950439453125,
+ "step": 1024
+ },
+ {
+ "epoch": 14.237554585152838,
+ "grad_norm": 0.01614239625632763,
+ "learning_rate": 0.0006,
+ "loss": 4.664529323577881,
+ "step": 1025
+ },
+ {
+ "epoch": 14.251528384279476,
+ "grad_norm": 0.016362549737095833,
+ "learning_rate": 0.0006,
+ "loss": 4.702373504638672,
+ "step": 1026
+ },
+ {
+ "epoch": 14.265502183406113,
+ "grad_norm": 0.015800409018993378,
+ "learning_rate": 0.0006,
+ "loss": 4.7594428062438965,
+ "step": 1027
+ },
+ {
+ "epoch": 14.279475982532752,
+ "grad_norm": 0.016730139032006264,
+ "learning_rate": 0.0006,
+ "loss": 4.677244186401367,
+ "step": 1028
+ },
+ {
+ "epoch": 14.293449781659389,
+ "grad_norm": 0.01661759801208973,
+ "learning_rate": 0.0006,
+ "loss": 4.633679389953613,
+ "step": 1029
+ },
+ {
+ "epoch": 14.307423580786025,
+ "grad_norm": 0.017673347145318985,
+ "learning_rate": 0.0006,
+ "loss": 4.649107456207275,
+ "step": 1030
+ },
+ {
+ "epoch": 14.321397379912664,
+ "grad_norm": 0.019243352115154266,
+ "learning_rate": 0.0006,
+ "loss": 4.611257553100586,
+ "step": 1031
+ },
+ {
+ "epoch": 14.335371179039301,
+ "grad_norm": 0.019817164167761803,
+ "learning_rate": 0.0006,
+ "loss": 4.588086128234863,
+ "step": 1032
+ },
+ {
+ "epoch": 14.34934497816594,
+ "grad_norm": 0.022292211651802063,
+ "learning_rate": 0.0006,
+ "loss": 4.783066749572754,
+ "step": 1033
+ },
+ {
+ "epoch": 14.363318777292577,
+ "grad_norm": 0.02140096202492714,
+ "learning_rate": 0.0006,
+ "loss": 4.604718208312988,
+ "step": 1034
+ },
+ {
+ "epoch": 14.377292576419213,
+ "grad_norm": 0.0187817495316267,
+ "learning_rate": 0.0006,
+ "loss": 4.721773147583008,
+ "step": 1035
+ },
+ {
+ "epoch": 14.391266375545852,
+ "grad_norm": 0.019879210740327835,
+ "learning_rate": 0.0006,
+ "loss": 4.789286136627197,
+ "step": 1036
+ },
+ {
+ "epoch": 14.405240174672489,
+ "grad_norm": 0.01885785162448883,
+ "learning_rate": 0.0006,
+ "loss": 4.669743537902832,
+ "step": 1037
+ },
+ {
+ "epoch": 14.419213973799126,
+ "grad_norm": 0.018324270844459534,
+ "learning_rate": 0.0006,
+ "loss": 4.65901517868042,
+ "step": 1038
+ },
+ {
+ "epoch": 14.433187772925764,
+ "grad_norm": 0.017664821818470955,
+ "learning_rate": 0.0006,
+ "loss": 4.667839527130127,
+ "step": 1039
+ },
+ {
+ "epoch": 14.447161572052401,
+ "grad_norm": 0.015702905133366585,
+ "learning_rate": 0.0006,
+ "loss": 4.64700984954834,
+ "step": 1040
+ },
+ {
+ "epoch": 14.46113537117904,
+ "grad_norm": 0.016674194484949112,
+ "learning_rate": 0.0006,
+ "loss": 4.694093227386475,
+ "step": 1041
+ },
+ {
+ "epoch": 14.475109170305677,
+ "grad_norm": 0.016869477927684784,
+ "learning_rate": 0.0006,
+ "loss": 4.619544982910156,
+ "step": 1042
+ },
+ {
+ "epoch": 14.489082969432314,
+ "grad_norm": 0.016557540744543076,
+ "learning_rate": 0.0006,
+ "loss": 4.568667411804199,
+ "step": 1043
+ },
+ {
+ "epoch": 14.503056768558952,
+ "grad_norm": 0.014500945806503296,
+ "learning_rate": 0.0006,
+ "loss": 4.561741828918457,
+ "step": 1044
+ },
+ {
+ "epoch": 14.51703056768559,
+ "grad_norm": 0.014228739775717258,
+ "learning_rate": 0.0006,
+ "loss": 4.654277324676514,
+ "step": 1045
+ },
+ {
+ "epoch": 14.531004366812226,
+ "grad_norm": 0.013894238509237766,
+ "learning_rate": 0.0006,
+ "loss": 4.5867719650268555,
+ "step": 1046
+ },
+ {
+ "epoch": 14.544978165938865,
+ "grad_norm": 0.01550068985670805,
+ "learning_rate": 0.0006,
+ "loss": 4.61168098449707,
+ "step": 1047
+ },
+ {
+ "epoch": 14.558951965065502,
+ "grad_norm": 0.015940772369503975,
+ "learning_rate": 0.0006,
+ "loss": 4.691247940063477,
+ "step": 1048
+ },
+ {
+ "epoch": 14.57292576419214,
+ "grad_norm": 0.01714535616338253,
+ "learning_rate": 0.0006,
+ "loss": 4.61183500289917,
+ "step": 1049
+ },
+ {
+ "epoch": 14.586899563318777,
+ "grad_norm": 0.017925038933753967,
+ "learning_rate": 0.0006,
+ "loss": 4.719022750854492,
+ "step": 1050
+ },
+ {
+ "epoch": 14.600873362445414,
+ "grad_norm": 0.018141552805900574,
+ "learning_rate": 0.0006,
+ "loss": 4.760281562805176,
+ "step": 1051
+ },
+ {
+ "epoch": 14.614847161572053,
+ "grad_norm": 0.017131328582763672,
+ "learning_rate": 0.0006,
+ "loss": 4.709680080413818,
+ "step": 1052
+ },
+ {
+ "epoch": 14.62882096069869,
+ "grad_norm": 0.018730048090219498,
+ "learning_rate": 0.0006,
+ "loss": 4.5994720458984375,
+ "step": 1053
+ },
+ {
+ "epoch": 14.642794759825328,
+ "grad_norm": 0.018894298002123833,
+ "learning_rate": 0.0006,
+ "loss": 4.669070720672607,
+ "step": 1054
+ },
+ {
+ "epoch": 14.656768558951965,
+ "grad_norm": 0.017643198370933533,
+ "learning_rate": 0.0006,
+ "loss": 4.641790390014648,
+ "step": 1055
+ },
+ {
+ "epoch": 14.670742358078602,
+ "grad_norm": 0.017647244036197662,
+ "learning_rate": 0.0006,
+ "loss": 4.653899192810059,
+ "step": 1056
+ },
+ {
+ "epoch": 14.68471615720524,
+ "grad_norm": 0.017304198816418648,
+ "learning_rate": 0.0006,
+ "loss": 4.7487473487854,
+ "step": 1057
+ },
+ {
+ "epoch": 14.698689956331878,
+ "grad_norm": 0.01638447679579258,
+ "learning_rate": 0.0006,
+ "loss": 4.71945858001709,
+ "step": 1058
+ },
+ {
+ "epoch": 14.712663755458514,
+ "grad_norm": 0.015892071649432182,
+ "learning_rate": 0.0006,
+ "loss": 4.744903564453125,
+ "step": 1059
+ },
+ {
+ "epoch": 14.726637554585153,
+ "grad_norm": 0.017420614138245583,
+ "learning_rate": 0.0006,
+ "loss": 4.781563758850098,
+ "step": 1060
+ },
+ {
+ "epoch": 14.74061135371179,
+ "grad_norm": 0.016498520970344543,
+ "learning_rate": 0.0006,
+ "loss": 4.622779846191406,
+ "step": 1061
+ },
+ {
+ "epoch": 14.754585152838429,
+ "grad_norm": 0.017482127994298935,
+ "learning_rate": 0.0006,
+ "loss": 4.62825870513916,
+ "step": 1062
+ },
+ {
+ "epoch": 14.768558951965066,
+ "grad_norm": 0.019018379971385002,
+ "learning_rate": 0.0006,
+ "loss": 4.693194389343262,
+ "step": 1063
+ },
+ {
+ "epoch": 14.782532751091702,
+ "grad_norm": 0.01993914321064949,
+ "learning_rate": 0.0006,
+ "loss": 4.742903709411621,
+ "step": 1064
+ },
+ {
+ "epoch": 14.796506550218341,
+ "grad_norm": 0.02301335707306862,
+ "learning_rate": 0.0006,
+ "loss": 4.613251686096191,
+ "step": 1065
+ },
+ {
+ "epoch": 14.810480349344978,
+ "grad_norm": 0.024088222533464432,
+ "learning_rate": 0.0006,
+ "loss": 4.594305992126465,
+ "step": 1066
+ },
+ {
+ "epoch": 14.824454148471617,
+ "grad_norm": 0.02493320405483246,
+ "learning_rate": 0.0006,
+ "loss": 4.700588226318359,
+ "step": 1067
+ },
+ {
+ "epoch": 14.838427947598253,
+ "grad_norm": 0.02501937560737133,
+ "learning_rate": 0.0006,
+ "loss": 4.601742744445801,
+ "step": 1068
+ },
+ {
+ "epoch": 14.85240174672489,
+ "grad_norm": 0.023490751162171364,
+ "learning_rate": 0.0006,
+ "loss": 4.589211463928223,
+ "step": 1069
+ },
+ {
+ "epoch": 14.866375545851529,
+ "grad_norm": 0.019695665687322617,
+ "learning_rate": 0.0006,
+ "loss": 4.635081768035889,
+ "step": 1070
+ },
+ {
+ "epoch": 14.880349344978166,
+ "grad_norm": 0.019276980310678482,
+ "learning_rate": 0.0006,
+ "loss": 4.737002372741699,
+ "step": 1071
+ },
+ {
+ "epoch": 14.894323144104803,
+ "grad_norm": 0.022430511191487312,
+ "learning_rate": 0.0006,
+ "loss": 4.808295726776123,
+ "step": 1072
+ },
+ {
+ "epoch": 14.908296943231441,
+ "grad_norm": 0.021543916314840317,
+ "learning_rate": 0.0006,
+ "loss": 4.559475421905518,
+ "step": 1073
+ },
+ {
+ "epoch": 14.922270742358078,
+ "grad_norm": 0.019885241985321045,
+ "learning_rate": 0.0006,
+ "loss": 4.647948741912842,
+ "step": 1074
+ },
+ {
+ "epoch": 14.936244541484717,
+ "grad_norm": 0.02115175500512123,
+ "learning_rate": 0.0006,
+ "loss": 4.680505752563477,
+ "step": 1075
+ },
+ {
+ "epoch": 14.950218340611354,
+ "grad_norm": 0.02156151458621025,
+ "learning_rate": 0.0006,
+ "loss": 4.673477649688721,
+ "step": 1076
+ },
+ {
+ "epoch": 14.96419213973799,
+ "grad_norm": 0.02184601128101349,
+ "learning_rate": 0.0006,
+ "loss": 4.644402027130127,
+ "step": 1077
+ },
+ {
+ "epoch": 14.97816593886463,
+ "grad_norm": 0.0211983360350132,
+ "learning_rate": 0.0006,
+ "loss": 4.821558475494385,
+ "step": 1078
+ },
+ {
+ "epoch": 14.992139737991266,
+ "grad_norm": 0.019636567682027817,
+ "learning_rate": 0.0006,
+ "loss": 4.578868865966797,
+ "step": 1079
+ },
+ {
+ "epoch": 15.0,
+ "grad_norm": 0.019997350871562958,
+ "learning_rate": 0.0006,
+ "loss": 4.691307544708252,
+ "step": 1080
+ },
+ {
+ "epoch": 15.0,
+ "eval_loss": 4.872478008270264,
+ "eval_runtime": 56.8513,
+ "eval_samples_per_second": 42.954,
+ "eval_steps_per_second": 1.354,
+ "step": 1080
+ },
+ {
+ "epoch": 15.013973799126637,
+ "grad_norm": 0.018830662593245506,
+ "learning_rate": 0.0006,
+ "loss": 4.761779308319092,
+ "step": 1081
+ },
+ {
+ "epoch": 15.027947598253276,
+ "grad_norm": 0.016876481473445892,
+ "learning_rate": 0.0006,
+ "loss": 4.61148738861084,
+ "step": 1082
+ },
+ {
+ "epoch": 15.041921397379912,
+ "grad_norm": 0.015550863929092884,
+ "learning_rate": 0.0006,
+ "loss": 4.676124095916748,
+ "step": 1083
+ },
+ {
+ "epoch": 15.055895196506551,
+ "grad_norm": 0.017178896814584732,
+ "learning_rate": 0.0006,
+ "loss": 4.588935852050781,
+ "step": 1084
+ },
+ {
+ "epoch": 15.069868995633188,
+ "grad_norm": 0.015465447679162025,
+ "learning_rate": 0.0006,
+ "loss": 4.605499267578125,
+ "step": 1085
+ },
+ {
+ "epoch": 15.083842794759825,
+ "grad_norm": 0.01606195978820324,
+ "learning_rate": 0.0006,
+ "loss": 4.723373889923096,
+ "step": 1086
+ },
+ {
+ "epoch": 15.097816593886463,
+ "grad_norm": 0.016521936282515526,
+ "learning_rate": 0.0006,
+ "loss": 4.646608352661133,
+ "step": 1087
+ },
+ {
+ "epoch": 15.1117903930131,
+ "grad_norm": 0.01589205674827099,
+ "learning_rate": 0.0006,
+ "loss": 4.498138427734375,
+ "step": 1088
+ },
+ {
+ "epoch": 15.125764192139737,
+ "grad_norm": 0.01524856686592102,
+ "learning_rate": 0.0006,
+ "loss": 4.594562530517578,
+ "step": 1089
+ },
+ {
+ "epoch": 15.139737991266376,
+ "grad_norm": 0.014714641496539116,
+ "learning_rate": 0.0006,
+ "loss": 4.630744934082031,
+ "step": 1090
+ },
+ {
+ "epoch": 15.153711790393013,
+ "grad_norm": 0.015333331190049648,
+ "learning_rate": 0.0006,
+ "loss": 4.682388782501221,
+ "step": 1091
+ },
+ {
+ "epoch": 15.167685589519651,
+ "grad_norm": 0.01550073828548193,
+ "learning_rate": 0.0006,
+ "loss": 4.692652702331543,
+ "step": 1092
+ },
+ {
+ "epoch": 15.181659388646288,
+ "grad_norm": 0.014783318154513836,
+ "learning_rate": 0.0006,
+ "loss": 4.6223835945129395,
+ "step": 1093
+ },
+ {
+ "epoch": 15.195633187772925,
+ "grad_norm": 0.014372751116752625,
+ "learning_rate": 0.0006,
+ "loss": 4.52687406539917,
+ "step": 1094
+ },
+ {
+ "epoch": 15.209606986899564,
+ "grad_norm": 0.014779130928218365,
+ "learning_rate": 0.0006,
+ "loss": 4.637009620666504,
+ "step": 1095
+ },
+ {
+ "epoch": 15.2235807860262,
+ "grad_norm": 0.015697073191404343,
+ "learning_rate": 0.0006,
+ "loss": 4.682981014251709,
+ "step": 1096
+ },
+ {
+ "epoch": 15.237554585152838,
+ "grad_norm": 0.016523664817214012,
+ "learning_rate": 0.0006,
+ "loss": 4.705722808837891,
+ "step": 1097
+ },
+ {
+ "epoch": 15.251528384279476,
+ "grad_norm": 0.017633996903896332,
+ "learning_rate": 0.0006,
+ "loss": 4.574686050415039,
+ "step": 1098
+ },
+ {
+ "epoch": 15.265502183406113,
+ "grad_norm": 0.019955076277256012,
+ "learning_rate": 0.0006,
+ "loss": 4.634450912475586,
+ "step": 1099
+ },
+ {
+ "epoch": 15.279475982532752,
+ "grad_norm": 0.02132618986070156,
+ "learning_rate": 0.0006,
+ "loss": 4.56110143661499,
+ "step": 1100
+ },
+ {
+ "epoch": 15.293449781659389,
+ "grad_norm": 0.0207593385130167,
+ "learning_rate": 0.0006,
+ "loss": 4.629396438598633,
+ "step": 1101
+ },
+ {
+ "epoch": 15.307423580786025,
+ "grad_norm": 0.017882540822029114,
+ "learning_rate": 0.0006,
+ "loss": 4.730963230133057,
+ "step": 1102
+ },
+ {
+ "epoch": 15.321397379912664,
+ "grad_norm": 0.01917876861989498,
+ "learning_rate": 0.0006,
+ "loss": 4.64585542678833,
+ "step": 1103
+ },
+ {
+ "epoch": 15.335371179039301,
+ "grad_norm": 0.01943974196910858,
+ "learning_rate": 0.0006,
+ "loss": 4.656181335449219,
+ "step": 1104
+ },
+ {
+ "epoch": 15.34934497816594,
+ "grad_norm": 0.019904715940356255,
+ "learning_rate": 0.0006,
+ "loss": 4.619329452514648,
+ "step": 1105
+ },
+ {
+ "epoch": 15.363318777292577,
+ "grad_norm": 0.019174346700310707,
+ "learning_rate": 0.0006,
+ "loss": 4.647892475128174,
+ "step": 1106
+ },
+ {
+ "epoch": 15.377292576419213,
+ "grad_norm": 0.018244758248329163,
+ "learning_rate": 0.0006,
+ "loss": 4.56166934967041,
+ "step": 1107
+ },
+ {
+ "epoch": 15.391266375545852,
+ "grad_norm": 0.018914205953478813,
+ "learning_rate": 0.0006,
+ "loss": 4.648063659667969,
+ "step": 1108
+ },
+ {
+ "epoch": 15.405240174672489,
+ "grad_norm": 0.01961071416735649,
+ "learning_rate": 0.0006,
+ "loss": 4.665525436401367,
+ "step": 1109
+ },
+ {
+ "epoch": 15.419213973799126,
+ "grad_norm": 0.01841311901807785,
+ "learning_rate": 0.0006,
+ "loss": 4.611342430114746,
+ "step": 1110
+ },
+ {
+ "epoch": 15.433187772925764,
+ "grad_norm": 0.01863427273929119,
+ "learning_rate": 0.0006,
+ "loss": 4.664256572723389,
+ "step": 1111
+ },
+ {
+ "epoch": 15.447161572052401,
+ "grad_norm": 0.020506877452135086,
+ "learning_rate": 0.0006,
+ "loss": 4.631237506866455,
+ "step": 1112
+ },
+ {
+ "epoch": 15.46113537117904,
+ "grad_norm": 0.02223833277821541,
+ "learning_rate": 0.0006,
+ "loss": 4.668654441833496,
+ "step": 1113
+ },
+ {
+ "epoch": 15.475109170305677,
+ "grad_norm": 0.023336702957749367,
+ "learning_rate": 0.0006,
+ "loss": 4.562833309173584,
+ "step": 1114
+ },
+ {
+ "epoch": 15.489082969432314,
+ "grad_norm": 0.023052413016557693,
+ "learning_rate": 0.0006,
+ "loss": 4.621532917022705,
+ "step": 1115
+ },
+ {
+ "epoch": 15.503056768558952,
+ "grad_norm": 0.023420924320816994,
+ "learning_rate": 0.0006,
+ "loss": 4.665465354919434,
+ "step": 1116
+ },
+ {
+ "epoch": 15.51703056768559,
+ "grad_norm": 0.024071281775832176,
+ "learning_rate": 0.0006,
+ "loss": 4.595980644226074,
+ "step": 1117
+ },
+ {
+ "epoch": 15.531004366812226,
+ "grad_norm": 0.023406682536005974,
+ "learning_rate": 0.0006,
+ "loss": 4.642635345458984,
+ "step": 1118
+ },
+ {
+ "epoch": 15.544978165938865,
+ "grad_norm": 0.02138647809624672,
+ "learning_rate": 0.0006,
+ "loss": 4.704513072967529,
+ "step": 1119
+ },
+ {
+ "epoch": 15.558951965065502,
+ "grad_norm": 0.0221384447067976,
+ "learning_rate": 0.0006,
+ "loss": 4.567503452301025,
+ "step": 1120
+ },
+ {
+ "epoch": 15.57292576419214,
+ "grad_norm": 0.020879851654171944,
+ "learning_rate": 0.0006,
+ "loss": 4.664057731628418,
+ "step": 1121
+ },
+ {
+ "epoch": 15.586899563318777,
+ "grad_norm": 0.021478967741131783,
+ "learning_rate": 0.0006,
+ "loss": 4.539627552032471,
+ "step": 1122
+ },
+ {
+ "epoch": 15.600873362445414,
+ "grad_norm": 0.022748656570911407,
+ "learning_rate": 0.0006,
+ "loss": 4.5654497146606445,
+ "step": 1123
+ },
+ {
+ "epoch": 15.614847161572053,
+ "grad_norm": 0.021870248019695282,
+ "learning_rate": 0.0006,
+ "loss": 4.56899356842041,
+ "step": 1124
+ },
+ {
+ "epoch": 15.62882096069869,
+ "grad_norm": 0.021711355075240135,
+ "learning_rate": 0.0006,
+ "loss": 4.657808303833008,
+ "step": 1125
+ },
+ {
+ "epoch": 15.642794759825328,
+ "grad_norm": 0.022810909897089005,
+ "learning_rate": 0.0006,
+ "loss": 4.6526689529418945,
+ "step": 1126
+ },
+ {
+ "epoch": 15.656768558951965,
+ "grad_norm": 0.022561343386769295,
+ "learning_rate": 0.0006,
+ "loss": 4.534887790679932,
+ "step": 1127
+ },
+ {
+ "epoch": 15.670742358078602,
+ "grad_norm": 0.019909802824258804,
+ "learning_rate": 0.0006,
+ "loss": 4.600790023803711,
+ "step": 1128
+ },
+ {
+ "epoch": 15.68471615720524,
+ "grad_norm": 0.020092325285077095,
+ "learning_rate": 0.0006,
+ "loss": 4.773830413818359,
+ "step": 1129
+ },
+ {
+ "epoch": 15.698689956331878,
+ "grad_norm": 0.018834445625543594,
+ "learning_rate": 0.0006,
+ "loss": 4.628868103027344,
+ "step": 1130
+ },
+ {
+ "epoch": 15.712663755458514,
+ "grad_norm": 0.017552798613905907,
+ "learning_rate": 0.0006,
+ "loss": 4.647500038146973,
+ "step": 1131
+ },
+ {
+ "epoch": 15.726637554585153,
+ "grad_norm": 0.01751251146197319,
+ "learning_rate": 0.0006,
+ "loss": 4.590545654296875,
+ "step": 1132
+ },
+ {
+ "epoch": 15.74061135371179,
+ "grad_norm": 0.01699099875986576,
+ "learning_rate": 0.0006,
+ "loss": 4.509699821472168,
+ "step": 1133
+ },
+ {
+ "epoch": 15.754585152838429,
+ "grad_norm": 0.015808649361133575,
+ "learning_rate": 0.0006,
+ "loss": 4.606992244720459,
+ "step": 1134
+ },
+ {
+ "epoch": 15.768558951965066,
+ "grad_norm": 0.01655443385243416,
+ "learning_rate": 0.0006,
+ "loss": 4.585219860076904,
+ "step": 1135
+ },
+ {
+ "epoch": 15.782532751091702,
+ "grad_norm": 0.01708163693547249,
+ "learning_rate": 0.0006,
+ "loss": 4.668147087097168,
+ "step": 1136
+ },
+ {
+ "epoch": 15.796506550218341,
+ "grad_norm": 0.017161913216114044,
+ "learning_rate": 0.0006,
+ "loss": 4.7170634269714355,
+ "step": 1137
+ },
+ {
+ "epoch": 15.810480349344978,
+ "grad_norm": 0.017262674868106842,
+ "learning_rate": 0.0006,
+ "loss": 4.572132110595703,
+ "step": 1138
+ },
+ {
+ "epoch": 15.824454148471617,
+ "grad_norm": 0.017713576555252075,
+ "learning_rate": 0.0006,
+ "loss": 4.702538967132568,
+ "step": 1139
+ },
+ {
+ "epoch": 15.838427947598253,
+ "grad_norm": 0.018730733543634415,
+ "learning_rate": 0.0006,
+ "loss": 4.629721641540527,
+ "step": 1140
+ },
+ {
+ "epoch": 15.85240174672489,
+ "grad_norm": 0.017960187047719955,
+ "learning_rate": 0.0006,
+ "loss": 4.661039352416992,
+ "step": 1141
+ },
+ {
+ "epoch": 15.866375545851529,
+ "grad_norm": 0.01750079356133938,
+ "learning_rate": 0.0006,
+ "loss": 4.503710746765137,
+ "step": 1142
+ },
+ {
+ "epoch": 15.880349344978166,
+ "grad_norm": 0.016630422323942184,
+ "learning_rate": 0.0006,
+ "loss": 4.6582136154174805,
+ "step": 1143
+ },
+ {
+ "epoch": 15.894323144104803,
+ "grad_norm": 0.015896832570433617,
+ "learning_rate": 0.0006,
+ "loss": 4.596693992614746,
+ "step": 1144
+ },
+ {
+ "epoch": 15.908296943231441,
+ "grad_norm": 0.01566407084465027,
+ "learning_rate": 0.0006,
+ "loss": 4.6337199211120605,
+ "step": 1145
+ },
+ {
+ "epoch": 15.922270742358078,
+ "grad_norm": 0.016823330894112587,
+ "learning_rate": 0.0006,
+ "loss": 4.686994552612305,
+ "step": 1146
+ },
+ {
+ "epoch": 15.936244541484717,
+ "grad_norm": 0.016208255663514137,
+ "learning_rate": 0.0006,
+ "loss": 4.6205549240112305,
+ "step": 1147
+ },
+ {
+ "epoch": 15.950218340611354,
+ "grad_norm": 0.0171707421541214,
+ "learning_rate": 0.0006,
+ "loss": 4.582968711853027,
+ "step": 1148
+ },
+ {
+ "epoch": 15.96419213973799,
+ "grad_norm": 0.017009446397423744,
+ "learning_rate": 0.0006,
+ "loss": 4.510274887084961,
+ "step": 1149
+ },
+ {
+ "epoch": 15.97816593886463,
+ "grad_norm": 0.01565811224281788,
+ "learning_rate": 0.0006,
+ "loss": 4.55834436416626,
+ "step": 1150
+ },
+ {
+ "epoch": 15.992139737991266,
+ "grad_norm": 0.015276037156581879,
+ "learning_rate": 0.0006,
+ "loss": 4.537870407104492,
+ "step": 1151
+ },
+ {
+ "epoch": 16.0,
+ "grad_norm": 0.017305882647633553,
+ "learning_rate": 0.0006,
+ "loss": 4.611305236816406,
+ "step": 1152
+ },
+ {
+ "epoch": 16.0,
+ "eval_loss": 4.775513172149658,
+ "eval_runtime": 56.7341,
+ "eval_samples_per_second": 43.043,
+ "eval_steps_per_second": 1.357,
+ "step": 1152
+ },
+ {
+ "epoch": 16.01397379912664,
+ "grad_norm": 0.016752436757087708,
+ "learning_rate": 0.0006,
+ "loss": 4.541312217712402,
+ "step": 1153
+ },
+ {
+ "epoch": 16.027947598253274,
+ "grad_norm": 0.01643134094774723,
+ "learning_rate": 0.0006,
+ "loss": 4.585402965545654,
+ "step": 1154
+ },
+ {
+ "epoch": 16.041921397379912,
+ "grad_norm": 0.01888720504939556,
+ "learning_rate": 0.0006,
+ "loss": 4.597825050354004,
+ "step": 1155
+ },
+ {
+ "epoch": 16.05589519650655,
+ "grad_norm": 0.022044632583856583,
+ "learning_rate": 0.0006,
+ "loss": 4.532772064208984,
+ "step": 1156
+ },
+ {
+ "epoch": 16.069868995633186,
+ "grad_norm": 0.022986987605690956,
+ "learning_rate": 0.0006,
+ "loss": 4.551351547241211,
+ "step": 1157
+ },
+ {
+ "epoch": 16.083842794759825,
+ "grad_norm": 0.020116791129112244,
+ "learning_rate": 0.0006,
+ "loss": 4.484149932861328,
+ "step": 1158
+ },
+ {
+ "epoch": 16.097816593886463,
+ "grad_norm": 0.021442247554659843,
+ "learning_rate": 0.0006,
+ "loss": 4.515316486358643,
+ "step": 1159
+ },
+ {
+ "epoch": 16.111790393013102,
+ "grad_norm": 0.020974772050976753,
+ "learning_rate": 0.0006,
+ "loss": 4.51606559753418,
+ "step": 1160
+ },
+ {
+ "epoch": 16.125764192139737,
+ "grad_norm": 0.021274864673614502,
+ "learning_rate": 0.0006,
+ "loss": 4.587329864501953,
+ "step": 1161
+ },
+ {
+ "epoch": 16.139737991266376,
+ "grad_norm": 0.022799784317612648,
+ "learning_rate": 0.0006,
+ "loss": 4.568149089813232,
+ "step": 1162
+ },
+ {
+ "epoch": 16.153711790393015,
+ "grad_norm": 0.021784933283925056,
+ "learning_rate": 0.0006,
+ "loss": 4.515336990356445,
+ "step": 1163
+ },
+ {
+ "epoch": 16.16768558951965,
+ "grad_norm": 0.024261271581053734,
+ "learning_rate": 0.0006,
+ "loss": 4.5127081871032715,
+ "step": 1164
+ },
+ {
+ "epoch": 16.18165938864629,
+ "grad_norm": 0.02248375490307808,
+ "learning_rate": 0.0006,
+ "loss": 4.754251003265381,
+ "step": 1165
+ },
+ {
+ "epoch": 16.195633187772927,
+ "grad_norm": 0.02136492356657982,
+ "learning_rate": 0.0006,
+ "loss": 4.58193826675415,
+ "step": 1166
+ },
+ {
+ "epoch": 16.209606986899562,
+ "grad_norm": 0.020727066323161125,
+ "learning_rate": 0.0006,
+ "loss": 4.564209938049316,
+ "step": 1167
+ },
+ {
+ "epoch": 16.2235807860262,
+ "grad_norm": 0.017789119854569435,
+ "learning_rate": 0.0006,
+ "loss": 4.698680877685547,
+ "step": 1168
+ },
+ {
+ "epoch": 16.23755458515284,
+ "grad_norm": 0.017603034153580666,
+ "learning_rate": 0.0006,
+ "loss": 4.501171588897705,
+ "step": 1169
+ },
+ {
+ "epoch": 16.251528384279474,
+ "grad_norm": 0.017345862463116646,
+ "learning_rate": 0.0006,
+ "loss": 4.593590259552002,
+ "step": 1170
+ },
+ {
+ "epoch": 16.265502183406113,
+ "grad_norm": 0.018047425895929337,
+ "learning_rate": 0.0006,
+ "loss": 4.597467422485352,
+ "step": 1171
+ },
+ {
+ "epoch": 16.27947598253275,
+ "grad_norm": 0.01947222836315632,
+ "learning_rate": 0.0006,
+ "loss": 4.5072221755981445,
+ "step": 1172
+ },
+ {
+ "epoch": 16.29344978165939,
+ "grad_norm": 0.019598500803112984,
+ "learning_rate": 0.0006,
+ "loss": 4.601351737976074,
+ "step": 1173
+ },
+ {
+ "epoch": 16.307423580786025,
+ "grad_norm": 0.020002514123916626,
+ "learning_rate": 0.0006,
+ "loss": 4.580431938171387,
+ "step": 1174
+ },
+ {
+ "epoch": 16.321397379912664,
+ "grad_norm": 0.0216530729085207,
+ "learning_rate": 0.0006,
+ "loss": 4.424681663513184,
+ "step": 1175
+ },
+ {
+ "epoch": 16.335371179039303,
+ "grad_norm": 0.01949235238134861,
+ "learning_rate": 0.0006,
+ "loss": 4.751883029937744,
+ "step": 1176
+ },
+ {
+ "epoch": 16.349344978165938,
+ "grad_norm": 0.017779016867280006,
+ "learning_rate": 0.0006,
+ "loss": 4.541490077972412,
+ "step": 1177
+ },
+ {
+ "epoch": 16.363318777292577,
+ "grad_norm": 0.01909121684730053,
+ "learning_rate": 0.0006,
+ "loss": 4.594141006469727,
+ "step": 1178
+ },
+ {
+ "epoch": 16.377292576419215,
+ "grad_norm": 0.019894849509000778,
+ "learning_rate": 0.0006,
+ "loss": 4.555070400238037,
+ "step": 1179
+ },
+ {
+ "epoch": 16.39126637554585,
+ "grad_norm": 0.01993032731115818,
+ "learning_rate": 0.0006,
+ "loss": 4.521401405334473,
+ "step": 1180
+ },
+ {
+ "epoch": 16.40524017467249,
+ "grad_norm": 0.0210788045078516,
+ "learning_rate": 0.0006,
+ "loss": 4.672842979431152,
+ "step": 1181
+ },
+ {
+ "epoch": 16.419213973799128,
+ "grad_norm": 0.02274598926305771,
+ "learning_rate": 0.0006,
+ "loss": 4.629504203796387,
+ "step": 1182
+ },
+ {
+ "epoch": 16.433187772925763,
+ "grad_norm": 0.020448120310902596,
+ "learning_rate": 0.0006,
+ "loss": 4.548209190368652,
+ "step": 1183
+ },
+ {
+ "epoch": 16.4471615720524,
+ "grad_norm": 0.017565643414855003,
+ "learning_rate": 0.0006,
+ "loss": 4.529054641723633,
+ "step": 1184
+ },
+ {
+ "epoch": 16.46113537117904,
+ "grad_norm": 0.019504237920045853,
+ "learning_rate": 0.0006,
+ "loss": 4.48182487487793,
+ "step": 1185
+ },
+ {
+ "epoch": 16.475109170305675,
+ "grad_norm": 0.016823742538690567,
+ "learning_rate": 0.0006,
+ "loss": 4.608696937561035,
+ "step": 1186
+ },
+ {
+ "epoch": 16.489082969432314,
+ "grad_norm": 0.01602894812822342,
+ "learning_rate": 0.0006,
+ "loss": 4.620062351226807,
+ "step": 1187
+ },
+ {
+ "epoch": 16.503056768558952,
+ "grad_norm": 0.016615642234683037,
+ "learning_rate": 0.0006,
+ "loss": 4.69378662109375,
+ "step": 1188
+ },
+ {
+ "epoch": 16.51703056768559,
+ "grad_norm": 0.018576730042696,
+ "learning_rate": 0.0006,
+ "loss": 4.537041664123535,
+ "step": 1189
+ },
+ {
+ "epoch": 16.531004366812226,
+ "grad_norm": 0.018611151725053787,
+ "learning_rate": 0.0006,
+ "loss": 4.496169090270996,
+ "step": 1190
+ },
+ {
+ "epoch": 16.544978165938865,
+ "grad_norm": 0.018622413277626038,
+ "learning_rate": 0.0006,
+ "loss": 4.557358264923096,
+ "step": 1191
+ },
+ {
+ "epoch": 16.558951965065503,
+ "grad_norm": 0.01823544315993786,
+ "learning_rate": 0.0006,
+ "loss": 4.730792999267578,
+ "step": 1192
+ },
+ {
+ "epoch": 16.57292576419214,
+ "grad_norm": 0.018796846270561218,
+ "learning_rate": 0.0006,
+ "loss": 4.781791687011719,
+ "step": 1193
+ },
+ {
+ "epoch": 16.586899563318777,
+ "grad_norm": 0.01728292554616928,
+ "learning_rate": 0.0006,
+ "loss": 4.515257835388184,
+ "step": 1194
+ },
+ {
+ "epoch": 16.600873362445416,
+ "grad_norm": 0.01699589006602764,
+ "learning_rate": 0.0006,
+ "loss": 4.5996904373168945,
+ "step": 1195
+ },
+ {
+ "epoch": 16.61484716157205,
+ "grad_norm": 0.015905970707535744,
+ "learning_rate": 0.0006,
+ "loss": 4.601301193237305,
+ "step": 1196
+ },
+ {
+ "epoch": 16.62882096069869,
+ "grad_norm": 0.01338571310043335,
+ "learning_rate": 0.0006,
+ "loss": 4.557708740234375,
+ "step": 1197
+ },
+ {
+ "epoch": 16.64279475982533,
+ "grad_norm": 0.015298855490982533,
+ "learning_rate": 0.0006,
+ "loss": 4.542287349700928,
+ "step": 1198
+ },
+ {
+ "epoch": 16.656768558951963,
+ "grad_norm": 0.014753367751836777,
+ "learning_rate": 0.0006,
+ "loss": 4.595787048339844,
+ "step": 1199
+ },
+ {
+ "epoch": 16.670742358078602,
+ "grad_norm": 0.015527483075857162,
+ "learning_rate": 0.0006,
+ "loss": 4.607529163360596,
+ "step": 1200
+ },
+ {
+ "epoch": 16.68471615720524,
+ "grad_norm": 0.015042162500321865,
+ "learning_rate": 0.0006,
+ "loss": 4.686626434326172,
+ "step": 1201
+ },
+ {
+ "epoch": 16.69868995633188,
+ "grad_norm": 0.0159393772482872,
+ "learning_rate": 0.0006,
+ "loss": 4.47656774520874,
+ "step": 1202
+ },
+ {
+ "epoch": 16.712663755458514,
+ "grad_norm": 0.01532787922769785,
+ "learning_rate": 0.0006,
+ "loss": 4.55678653717041,
+ "step": 1203
+ },
+ {
+ "epoch": 16.726637554585153,
+ "grad_norm": 0.016332164406776428,
+ "learning_rate": 0.0006,
+ "loss": 4.51441764831543,
+ "step": 1204
+ },
+ {
+ "epoch": 16.74061135371179,
+ "grad_norm": 0.018901726230978966,
+ "learning_rate": 0.0006,
+ "loss": 4.57227897644043,
+ "step": 1205
+ },
+ {
+ "epoch": 16.754585152838427,
+ "grad_norm": 0.018589433282613754,
+ "learning_rate": 0.0006,
+ "loss": 4.550070762634277,
+ "step": 1206
+ },
+ {
+ "epoch": 16.768558951965066,
+ "grad_norm": 0.01634034886956215,
+ "learning_rate": 0.0006,
+ "loss": 4.586910724639893,
+ "step": 1207
+ },
+ {
+ "epoch": 16.782532751091704,
+ "grad_norm": 0.017599433660507202,
+ "learning_rate": 0.0006,
+ "loss": 4.5891923904418945,
+ "step": 1208
+ },
+ {
+ "epoch": 16.79650655021834,
+ "grad_norm": 0.01878446340560913,
+ "learning_rate": 0.0006,
+ "loss": 4.57227087020874,
+ "step": 1209
+ },
+ {
+ "epoch": 16.810480349344978,
+ "grad_norm": 0.01803469844162464,
+ "learning_rate": 0.0006,
+ "loss": 4.583845138549805,
+ "step": 1210
+ },
+ {
+ "epoch": 16.824454148471617,
+ "grad_norm": 0.018505144864320755,
+ "learning_rate": 0.0006,
+ "loss": 4.561151504516602,
+ "step": 1211
+ },
+ {
+ "epoch": 16.83842794759825,
+ "grad_norm": 0.01920081488788128,
+ "learning_rate": 0.0006,
+ "loss": 4.506834030151367,
+ "step": 1212
+ },
+ {
+ "epoch": 16.85240174672489,
+ "grad_norm": 0.0177712831646204,
+ "learning_rate": 0.0006,
+ "loss": 4.557195663452148,
+ "step": 1213
+ },
+ {
+ "epoch": 16.86637554585153,
+ "grad_norm": 0.019113918766379356,
+ "learning_rate": 0.0006,
+ "loss": 4.651227951049805,
+ "step": 1214
+ },
+ {
+ "epoch": 16.880349344978168,
+ "grad_norm": 0.02012539468705654,
+ "learning_rate": 0.0006,
+ "loss": 4.4952239990234375,
+ "step": 1215
+ },
+ {
+ "epoch": 16.894323144104803,
+ "grad_norm": 0.019993646070361137,
+ "learning_rate": 0.0006,
+ "loss": 4.517433166503906,
+ "step": 1216
+ },
+ {
+ "epoch": 16.90829694323144,
+ "grad_norm": 0.020041441544890404,
+ "learning_rate": 0.0006,
+ "loss": 4.489407539367676,
+ "step": 1217
+ },
+ {
+ "epoch": 16.92227074235808,
+ "grad_norm": 0.0183611661195755,
+ "learning_rate": 0.0006,
+ "loss": 4.5435638427734375,
+ "step": 1218
+ },
+ {
+ "epoch": 16.936244541484715,
+ "grad_norm": 0.019138429313898087,
+ "learning_rate": 0.0006,
+ "loss": 4.545949459075928,
+ "step": 1219
+ },
+ {
+ "epoch": 16.950218340611354,
+ "grad_norm": 0.018554266542196274,
+ "learning_rate": 0.0006,
+ "loss": 4.579620361328125,
+ "step": 1220
+ },
+ {
+ "epoch": 16.964192139737992,
+ "grad_norm": 0.0193545650690794,
+ "learning_rate": 0.0006,
+ "loss": 4.588255405426025,
+ "step": 1221
+ },
+ {
+ "epoch": 16.978165938864628,
+ "grad_norm": 0.019389165565371513,
+ "learning_rate": 0.0006,
+ "loss": 4.520698547363281,
+ "step": 1222
+ },
+ {
+ "epoch": 16.992139737991266,
+ "grad_norm": 0.016601379960775375,
+ "learning_rate": 0.0006,
+ "loss": 4.4809064865112305,
+ "step": 1223
+ },
+ {
+ "epoch": 17.0,
+ "grad_norm": 0.01771303080022335,
+ "learning_rate": 0.0006,
+ "loss": 4.721122741699219,
+ "step": 1224
+ },
+ {
+ "epoch": 17.0,
+ "eval_loss": 4.7903666496276855,
+ "eval_runtime": 56.9826,
+ "eval_samples_per_second": 42.855,
+ "eval_steps_per_second": 1.351,
+ "step": 1224
+ },
+ {
+ "epoch": 17.01397379912664,
+ "grad_norm": 0.017720289528369904,
+ "learning_rate": 0.0006,
+ "loss": 4.523834705352783,
+ "step": 1225
+ },
+ {
+ "epoch": 17.027947598253274,
+ "grad_norm": 0.020137161016464233,
+ "learning_rate": 0.0006,
+ "loss": 4.574061870574951,
+ "step": 1226
+ },
+ {
+ "epoch": 17.041921397379912,
+ "grad_norm": 0.02002314105629921,
+ "learning_rate": 0.0006,
+ "loss": 4.512407302856445,
+ "step": 1227
+ },
+ {
+ "epoch": 17.05589519650655,
+ "grad_norm": 0.019628183916211128,
+ "learning_rate": 0.0006,
+ "loss": 4.502533435821533,
+ "step": 1228
+ },
+ {
+ "epoch": 17.069868995633186,
+ "grad_norm": 0.019808458164334297,
+ "learning_rate": 0.0006,
+ "loss": 4.463156700134277,
+ "step": 1229
+ },
+ {
+ "epoch": 17.083842794759825,
+ "grad_norm": 0.021250786259770393,
+ "learning_rate": 0.0006,
+ "loss": 4.622682094573975,
+ "step": 1230
+ },
+ {
+ "epoch": 17.097816593886463,
+ "grad_norm": 0.022031694650650024,
+ "learning_rate": 0.0006,
+ "loss": 4.451882839202881,
+ "step": 1231
+ },
+ {
+ "epoch": 17.111790393013102,
+ "grad_norm": 0.01898987777531147,
+ "learning_rate": 0.0006,
+ "loss": 4.529331684112549,
+ "step": 1232
+ },
+ {
+ "epoch": 17.125764192139737,
+ "grad_norm": 0.018749834969639778,
+ "learning_rate": 0.0006,
+ "loss": 4.435815334320068,
+ "step": 1233
+ },
+ {
+ "epoch": 17.139737991266376,
+ "grad_norm": 0.01896674931049347,
+ "learning_rate": 0.0006,
+ "loss": 4.484029293060303,
+ "step": 1234
+ },
+ {
+ "epoch": 17.153711790393015,
+ "grad_norm": 0.018286997452378273,
+ "learning_rate": 0.0006,
+ "loss": 4.523510932922363,
+ "step": 1235
+ },
+ {
+ "epoch": 17.16768558951965,
+ "grad_norm": 0.017273059114813805,
+ "learning_rate": 0.0006,
+ "loss": 4.451767444610596,
+ "step": 1236
+ },
+ {
+ "epoch": 17.18165938864629,
+ "grad_norm": 0.016499150544404984,
+ "learning_rate": 0.0006,
+ "loss": 4.5568647384643555,
+ "step": 1237
+ },
+ {
+ "epoch": 17.195633187772927,
+ "grad_norm": 0.015934469178318977,
+ "learning_rate": 0.0006,
+ "loss": 4.4639997482299805,
+ "step": 1238
+ },
+ {
+ "epoch": 17.209606986899562,
+ "grad_norm": 0.01714209094643593,
+ "learning_rate": 0.0006,
+ "loss": 4.547614097595215,
+ "step": 1239
+ },
+ {
+ "epoch": 17.2235807860262,
+ "grad_norm": 0.019167495891451836,
+ "learning_rate": 0.0006,
+ "loss": 4.571097373962402,
+ "step": 1240
+ },
+ {
+ "epoch": 17.23755458515284,
+ "grad_norm": 0.018276996910572052,
+ "learning_rate": 0.0006,
+ "loss": 4.489764213562012,
+ "step": 1241
+ },
+ {
+ "epoch": 17.251528384279474,
+ "grad_norm": 0.021296866238117218,
+ "learning_rate": 0.0006,
+ "loss": 4.526913642883301,
+ "step": 1242
+ },
+ {
+ "epoch": 17.265502183406113,
+ "grad_norm": 0.022606629878282547,
+ "learning_rate": 0.0006,
+ "loss": 4.521624565124512,
+ "step": 1243
+ },
+ {
+ "epoch": 17.27947598253275,
+ "grad_norm": 0.021204711869359016,
+ "learning_rate": 0.0006,
+ "loss": 4.500864028930664,
+ "step": 1244
+ },
+ {
+ "epoch": 17.29344978165939,
+ "grad_norm": 0.01988849975168705,
+ "learning_rate": 0.0006,
+ "loss": 4.45162296295166,
+ "step": 1245
+ },
+ {
+ "epoch": 17.307423580786025,
+ "grad_norm": 0.020268384367227554,
+ "learning_rate": 0.0006,
+ "loss": 4.536198616027832,
+ "step": 1246
+ },
+ {
+ "epoch": 17.321397379912664,
+ "grad_norm": 0.020185677334666252,
+ "learning_rate": 0.0006,
+ "loss": 4.440262794494629,
+ "step": 1247
+ },
+ {
+ "epoch": 17.335371179039303,
+ "grad_norm": 0.020106853917241096,
+ "learning_rate": 0.0006,
+ "loss": 4.475907325744629,
+ "step": 1248
+ },
+ {
+ "epoch": 17.349344978165938,
+ "grad_norm": 0.018883297219872475,
+ "learning_rate": 0.0006,
+ "loss": 4.640661716461182,
+ "step": 1249
+ },
+ {
+ "epoch": 17.363318777292577,
+ "grad_norm": 0.01852620206773281,
+ "learning_rate": 0.0006,
+ "loss": 4.578243255615234,
+ "step": 1250
+ },
+ {
+ "epoch": 17.377292576419215,
+ "grad_norm": 0.01886451430618763,
+ "learning_rate": 0.0006,
+ "loss": 4.546026229858398,
+ "step": 1251
+ },
+ {
+ "epoch": 17.39126637554585,
+ "grad_norm": 0.01777452416718006,
+ "learning_rate": 0.0006,
+ "loss": 4.521542549133301,
+ "step": 1252
+ },
+ {
+ "epoch": 17.40524017467249,
+ "grad_norm": 0.016626447439193726,
+ "learning_rate": 0.0006,
+ "loss": 4.52629280090332,
+ "step": 1253
+ },
+ {
+ "epoch": 17.419213973799128,
+ "grad_norm": 0.01599477231502533,
+ "learning_rate": 0.0006,
+ "loss": 4.4796037673950195,
+ "step": 1254
+ },
+ {
+ "epoch": 17.433187772925763,
+ "grad_norm": 0.017021458595991135,
+ "learning_rate": 0.0006,
+ "loss": 4.468813419342041,
+ "step": 1255
+ },
+ {
+ "epoch": 17.4471615720524,
+ "grad_norm": 0.016925333067774773,
+ "learning_rate": 0.0006,
+ "loss": 4.584898948669434,
+ "step": 1256
+ },
+ {
+ "epoch": 17.46113537117904,
+ "grad_norm": 0.01714998669922352,
+ "learning_rate": 0.0006,
+ "loss": 4.523544788360596,
+ "step": 1257
+ },
+ {
+ "epoch": 17.475109170305675,
+ "grad_norm": 0.01802288182079792,
+ "learning_rate": 0.0006,
+ "loss": 4.595581531524658,
+ "step": 1258
+ },
+ {
+ "epoch": 17.489082969432314,
+ "grad_norm": 0.0180289875715971,
+ "learning_rate": 0.0006,
+ "loss": 4.394440650939941,
+ "step": 1259
+ },
+ {
+ "epoch": 17.503056768558952,
+ "grad_norm": 0.018504245206713676,
+ "learning_rate": 0.0006,
+ "loss": 4.4794416427612305,
+ "step": 1260
+ },
+ {
+ "epoch": 17.51703056768559,
+ "grad_norm": 0.019973233342170715,
+ "learning_rate": 0.0006,
+ "loss": 4.4241437911987305,
+ "step": 1261
+ },
+ {
+ "epoch": 17.531004366812226,
+ "grad_norm": 0.020063083618879318,
+ "learning_rate": 0.0006,
+ "loss": 4.48068380355835,
+ "step": 1262
+ },
+ {
+ "epoch": 17.544978165938865,
+ "grad_norm": 0.02066640742123127,
+ "learning_rate": 0.0006,
+ "loss": 4.628993034362793,
+ "step": 1263
+ },
+ {
+ "epoch": 17.558951965065503,
+ "grad_norm": 0.02206416428089142,
+ "learning_rate": 0.0006,
+ "loss": 4.564790725708008,
+ "step": 1264
+ },
+ {
+ "epoch": 17.57292576419214,
+ "grad_norm": 0.020640335977077484,
+ "learning_rate": 0.0006,
+ "loss": 4.578705787658691,
+ "step": 1265
+ },
+ {
+ "epoch": 17.586899563318777,
+ "grad_norm": 0.019093405455350876,
+ "learning_rate": 0.0006,
+ "loss": 4.5023603439331055,
+ "step": 1266
+ },
+ {
+ "epoch": 17.600873362445416,
+ "grad_norm": 0.019459193572402,
+ "learning_rate": 0.0006,
+ "loss": 4.594531059265137,
+ "step": 1267
+ },
+ {
+ "epoch": 17.61484716157205,
+ "grad_norm": 0.02021171525120735,
+ "learning_rate": 0.0006,
+ "loss": 4.770299911499023,
+ "step": 1268
+ },
+ {
+ "epoch": 17.62882096069869,
+ "grad_norm": 0.021161191165447235,
+ "learning_rate": 0.0006,
+ "loss": 4.5380401611328125,
+ "step": 1269
+ },
+ {
+ "epoch": 17.64279475982533,
+ "grad_norm": 0.021013403311371803,
+ "learning_rate": 0.0006,
+ "loss": 4.455524921417236,
+ "step": 1270
+ },
+ {
+ "epoch": 17.656768558951963,
+ "grad_norm": 0.020336592569947243,
+ "learning_rate": 0.0006,
+ "loss": 4.533672332763672,
+ "step": 1271
+ },
+ {
+ "epoch": 17.670742358078602,
+ "grad_norm": 0.020159434527158737,
+ "learning_rate": 0.0006,
+ "loss": 4.596151351928711,
+ "step": 1272
+ },
+ {
+ "epoch": 17.68471615720524,
+ "grad_norm": 0.020432960242033005,
+ "learning_rate": 0.0006,
+ "loss": 4.55531644821167,
+ "step": 1273
+ },
+ {
+ "epoch": 17.69868995633188,
+ "grad_norm": 0.025466229766607285,
+ "learning_rate": 0.0006,
+ "loss": 4.566912651062012,
+ "step": 1274
+ },
+ {
+ "epoch": 17.712663755458514,
+ "grad_norm": 0.026537558063864708,
+ "learning_rate": 0.0006,
+ "loss": 4.526003837585449,
+ "step": 1275
+ },
+ {
+ "epoch": 17.726637554585153,
+ "grad_norm": 0.024397607892751694,
+ "learning_rate": 0.0006,
+ "loss": 4.541755676269531,
+ "step": 1276
+ },
+ {
+ "epoch": 17.74061135371179,
+ "grad_norm": 0.022062312811613083,
+ "learning_rate": 0.0006,
+ "loss": 4.6532793045043945,
+ "step": 1277
+ },
+ {
+ "epoch": 17.754585152838427,
+ "grad_norm": 0.02204282023012638,
+ "learning_rate": 0.0006,
+ "loss": 4.582825660705566,
+ "step": 1278
+ },
+ {
+ "epoch": 17.768558951965066,
+ "grad_norm": 0.020844416692852974,
+ "learning_rate": 0.0006,
+ "loss": 4.522336483001709,
+ "step": 1279
+ },
+ {
+ "epoch": 17.782532751091704,
+ "grad_norm": 0.019815703853964806,
+ "learning_rate": 0.0006,
+ "loss": 4.607398509979248,
+ "step": 1280
+ },
+ {
+ "epoch": 17.79650655021834,
+ "grad_norm": 0.016676954925060272,
+ "learning_rate": 0.0006,
+ "loss": 4.446206092834473,
+ "step": 1281
+ },
+ {
+ "epoch": 17.810480349344978,
+ "grad_norm": 0.0166452769190073,
+ "learning_rate": 0.0006,
+ "loss": 4.505931854248047,
+ "step": 1282
+ },
+ {
+ "epoch": 17.824454148471617,
+ "grad_norm": 0.015928907319903374,
+ "learning_rate": 0.0006,
+ "loss": 4.455267429351807,
+ "step": 1283
+ },
+ {
+ "epoch": 17.83842794759825,
+ "grad_norm": 0.015219560824334621,
+ "learning_rate": 0.0006,
+ "loss": 4.5623674392700195,
+ "step": 1284
+ },
+ {
+ "epoch": 17.85240174672489,
+ "grad_norm": 0.015125355683267117,
+ "learning_rate": 0.0006,
+ "loss": 4.554104804992676,
+ "step": 1285
+ },
+ {
+ "epoch": 17.86637554585153,
+ "grad_norm": 0.015612718649208546,
+ "learning_rate": 0.0006,
+ "loss": 4.484847068786621,
+ "step": 1286
+ },
+ {
+ "epoch": 17.880349344978168,
+ "grad_norm": 0.015215662308037281,
+ "learning_rate": 0.0006,
+ "loss": 4.4802093505859375,
+ "step": 1287
+ },
+ {
+ "epoch": 17.894323144104803,
+ "grad_norm": 0.01511684525758028,
+ "learning_rate": 0.0006,
+ "loss": 4.537298679351807,
+ "step": 1288
+ },
+ {
+ "epoch": 17.90829694323144,
+ "grad_norm": 0.013841322623193264,
+ "learning_rate": 0.0006,
+ "loss": 4.427488327026367,
+ "step": 1289
+ },
+ {
+ "epoch": 17.92227074235808,
+ "grad_norm": 0.015115964226424694,
+ "learning_rate": 0.0006,
+ "loss": 4.4547119140625,
+ "step": 1290
+ },
+ {
+ "epoch": 17.936244541484715,
+ "grad_norm": 0.016411006450653076,
+ "learning_rate": 0.0006,
+ "loss": 4.52994966506958,
+ "step": 1291
+ },
+ {
+ "epoch": 17.950218340611354,
+ "grad_norm": 0.017313152551651,
+ "learning_rate": 0.0006,
+ "loss": 4.5337653160095215,
+ "step": 1292
+ },
+ {
+ "epoch": 17.964192139737992,
+ "grad_norm": 0.016559796407818794,
+ "learning_rate": 0.0006,
+ "loss": 4.5277910232543945,
+ "step": 1293
+ },
+ {
+ "epoch": 17.978165938864628,
+ "grad_norm": 0.015854761004447937,
+ "learning_rate": 0.0006,
+ "loss": 4.6334309577941895,
+ "step": 1294
+ },
+ {
+ "epoch": 17.992139737991266,
+ "grad_norm": 0.015658238902688026,
+ "learning_rate": 0.0006,
+ "loss": 4.613372802734375,
+ "step": 1295
+ },
+ {
+ "epoch": 18.0,
+ "grad_norm": 0.0169155802577734,
+ "learning_rate": 0.0006,
+ "loss": 4.527356147766113,
+ "step": 1296
+ },
+ {
+ "epoch": 18.0,
+ "eval_loss": 4.764543056488037,
+ "eval_runtime": 56.5998,
+ "eval_samples_per_second": 43.145,
+ "eval_steps_per_second": 1.36,
+ "step": 1296
+ },
+ {
+ "epoch": 18.01397379912664,
+ "grad_norm": 0.015348963439464569,
+ "learning_rate": 0.0006,
+ "loss": 4.466423988342285,
+ "step": 1297
+ },
+ {
+ "epoch": 18.027947598253274,
+ "grad_norm": 0.01768588088452816,
+ "learning_rate": 0.0006,
+ "loss": 4.545933246612549,
+ "step": 1298
+ },
+ {
+ "epoch": 18.041921397379912,
+ "grad_norm": 0.017381226643919945,
+ "learning_rate": 0.0006,
+ "loss": 4.494060039520264,
+ "step": 1299
+ },
+ {
+ "epoch": 18.05589519650655,
+ "grad_norm": 0.019002335146069527,
+ "learning_rate": 0.0006,
+ "loss": 4.395200729370117,
+ "step": 1300
+ },
+ {
+ "epoch": 18.069868995633186,
+ "grad_norm": 0.01827932707965374,
+ "learning_rate": 0.0006,
+ "loss": 4.424848556518555,
+ "step": 1301
+ },
+ {
+ "epoch": 18.083842794759825,
+ "grad_norm": 0.016383031383156776,
+ "learning_rate": 0.0006,
+ "loss": 4.494722366333008,
+ "step": 1302
+ },
+ {
+ "epoch": 18.097816593886463,
+ "grad_norm": 0.016265157610177994,
+ "learning_rate": 0.0006,
+ "loss": 4.416522979736328,
+ "step": 1303
+ },
+ {
+ "epoch": 18.111790393013102,
+ "grad_norm": 0.01702386513352394,
+ "learning_rate": 0.0006,
+ "loss": 4.600103378295898,
+ "step": 1304
+ },
+ {
+ "epoch": 18.125764192139737,
+ "grad_norm": 0.018495498225092888,
+ "learning_rate": 0.0006,
+ "loss": 4.588401794433594,
+ "step": 1305
+ },
+ {
+ "epoch": 18.139737991266376,
+ "grad_norm": 0.017921684309840202,
+ "learning_rate": 0.0006,
+ "loss": 4.478075981140137,
+ "step": 1306
+ },
+ {
+ "epoch": 18.153711790393015,
+ "grad_norm": 0.01801278069615364,
+ "learning_rate": 0.0006,
+ "loss": 4.413186550140381,
+ "step": 1307
+ },
+ {
+ "epoch": 18.16768558951965,
+ "grad_norm": 0.017211275175213814,
+ "learning_rate": 0.0006,
+ "loss": 4.461341857910156,
+ "step": 1308
+ },
+ {
+ "epoch": 18.18165938864629,
+ "grad_norm": 0.017589328810572624,
+ "learning_rate": 0.0006,
+ "loss": 4.546969413757324,
+ "step": 1309
+ },
+ {
+ "epoch": 18.195633187772927,
+ "grad_norm": 0.01797698810696602,
+ "learning_rate": 0.0006,
+ "loss": 4.530797004699707,
+ "step": 1310
+ },
+ {
+ "epoch": 18.209606986899562,
+ "grad_norm": 0.020408619195222855,
+ "learning_rate": 0.0006,
+ "loss": 4.62999153137207,
+ "step": 1311
+ },
+ {
+ "epoch": 18.2235807860262,
+ "grad_norm": 0.01994038000702858,
+ "learning_rate": 0.0006,
+ "loss": 4.586811065673828,
+ "step": 1312
+ },
+ {
+ "epoch": 18.23755458515284,
+ "grad_norm": 0.019381340593099594,
+ "learning_rate": 0.0006,
+ "loss": 4.4769368171691895,
+ "step": 1313
+ },
+ {
+ "epoch": 18.251528384279474,
+ "grad_norm": 0.020484555512666702,
+ "learning_rate": 0.0006,
+ "loss": 4.524302959442139,
+ "step": 1314
+ },
+ {
+ "epoch": 18.265502183406113,
+ "grad_norm": 0.02426416426897049,
+ "learning_rate": 0.0006,
+ "loss": 4.561705589294434,
+ "step": 1315
+ },
+ {
+ "epoch": 18.27947598253275,
+ "grad_norm": 0.024748101830482483,
+ "learning_rate": 0.0006,
+ "loss": 4.4654316902160645,
+ "step": 1316
+ },
+ {
+ "epoch": 18.29344978165939,
+ "grad_norm": 0.020703328773379326,
+ "learning_rate": 0.0006,
+ "loss": 4.53033447265625,
+ "step": 1317
+ },
+ {
+ "epoch": 18.307423580786025,
+ "grad_norm": 0.01765458658337593,
+ "learning_rate": 0.0006,
+ "loss": 4.5154266357421875,
+ "step": 1318
+ },
+ {
+ "epoch": 18.321397379912664,
+ "grad_norm": 0.018646420910954475,
+ "learning_rate": 0.0006,
+ "loss": 4.5224714279174805,
+ "step": 1319
+ },
+ {
+ "epoch": 18.335371179039303,
+ "grad_norm": 0.01907474733889103,
+ "learning_rate": 0.0006,
+ "loss": 4.5055437088012695,
+ "step": 1320
+ },
+ {
+ "epoch": 18.349344978165938,
+ "grad_norm": 0.018179846927523613,
+ "learning_rate": 0.0006,
+ "loss": 4.462161064147949,
+ "step": 1321
+ },
+ {
+ "epoch": 18.363318777292577,
+ "grad_norm": 0.016697878018021584,
+ "learning_rate": 0.0006,
+ "loss": 4.473718643188477,
+ "step": 1322
+ },
+ {
+ "epoch": 18.377292576419215,
+ "grad_norm": 0.017504019662737846,
+ "learning_rate": 0.0006,
+ "loss": 4.474348068237305,
+ "step": 1323
+ },
+ {
+ "epoch": 18.39126637554585,
+ "grad_norm": 0.018889913335442543,
+ "learning_rate": 0.0006,
+ "loss": 4.4531636238098145,
+ "step": 1324
+ },
+ {
+ "epoch": 18.40524017467249,
+ "grad_norm": 0.019167251884937286,
+ "learning_rate": 0.0006,
+ "loss": 4.4278364181518555,
+ "step": 1325
+ },
+ {
+ "epoch": 18.419213973799128,
+ "grad_norm": 0.015895741060376167,
+ "learning_rate": 0.0006,
+ "loss": 4.474592208862305,
+ "step": 1326
+ },
+ {
+ "epoch": 18.433187772925763,
+ "grad_norm": 0.01531882956624031,
+ "learning_rate": 0.0006,
+ "loss": 4.403757572174072,
+ "step": 1327
+ },
+ {
+ "epoch": 18.4471615720524,
+ "grad_norm": 0.01565849967300892,
+ "learning_rate": 0.0006,
+ "loss": 4.507664680480957,
+ "step": 1328
+ },
+ {
+ "epoch": 18.46113537117904,
+ "grad_norm": 0.01602315716445446,
+ "learning_rate": 0.0006,
+ "loss": 4.5648908615112305,
+ "step": 1329
+ },
+ {
+ "epoch": 18.475109170305675,
+ "grad_norm": 0.016467366367578506,
+ "learning_rate": 0.0006,
+ "loss": 4.428762912750244,
+ "step": 1330
+ },
+ {
+ "epoch": 18.489082969432314,
+ "grad_norm": 0.01604127697646618,
+ "learning_rate": 0.0006,
+ "loss": 4.527771472930908,
+ "step": 1331
+ },
+ {
+ "epoch": 18.503056768558952,
+ "grad_norm": 0.018825586885213852,
+ "learning_rate": 0.0006,
+ "loss": 4.485386848449707,
+ "step": 1332
+ },
+ {
+ "epoch": 18.51703056768559,
+ "grad_norm": 0.019258007407188416,
+ "learning_rate": 0.0006,
+ "loss": 4.414972305297852,
+ "step": 1333
+ },
+ {
+ "epoch": 18.531004366812226,
+ "grad_norm": 0.018192021176218987,
+ "learning_rate": 0.0006,
+ "loss": 4.412166595458984,
+ "step": 1334
+ },
+ {
+ "epoch": 18.544978165938865,
+ "grad_norm": 0.016120698302984238,
+ "learning_rate": 0.0006,
+ "loss": 4.484317779541016,
+ "step": 1335
+ },
+ {
+ "epoch": 18.558951965065503,
+ "grad_norm": 0.016496622934937477,
+ "learning_rate": 0.0006,
+ "loss": 4.514986515045166,
+ "step": 1336
+ },
+ {
+ "epoch": 18.57292576419214,
+ "grad_norm": 0.01795116998255253,
+ "learning_rate": 0.0006,
+ "loss": 4.41473388671875,
+ "step": 1337
+ },
+ {
+ "epoch": 18.586899563318777,
+ "grad_norm": 0.019062276929616928,
+ "learning_rate": 0.0006,
+ "loss": 4.601996898651123,
+ "step": 1338
+ },
+ {
+ "epoch": 18.600873362445416,
+ "grad_norm": 0.017174631357192993,
+ "learning_rate": 0.0006,
+ "loss": 4.571218490600586,
+ "step": 1339
+ },
+ {
+ "epoch": 18.61484716157205,
+ "grad_norm": 0.014015606604516506,
+ "learning_rate": 0.0006,
+ "loss": 4.526606559753418,
+ "step": 1340
+ },
+ {
+ "epoch": 18.62882096069869,
+ "grad_norm": 0.015358510427176952,
+ "learning_rate": 0.0006,
+ "loss": 4.604286193847656,
+ "step": 1341
+ },
+ {
+ "epoch": 18.64279475982533,
+ "grad_norm": 0.01619391329586506,
+ "learning_rate": 0.0006,
+ "loss": 4.560883522033691,
+ "step": 1342
+ },
+ {
+ "epoch": 18.656768558951963,
+ "grad_norm": 0.015972964465618134,
+ "learning_rate": 0.0006,
+ "loss": 4.479755401611328,
+ "step": 1343
+ },
+ {
+ "epoch": 18.670742358078602,
+ "grad_norm": 0.016861025243997574,
+ "learning_rate": 0.0006,
+ "loss": 4.474471092224121,
+ "step": 1344
+ },
+ {
+ "epoch": 18.68471615720524,
+ "grad_norm": 0.01822078227996826,
+ "learning_rate": 0.0006,
+ "loss": 4.509489059448242,
+ "step": 1345
+ },
+ {
+ "epoch": 18.69868995633188,
+ "grad_norm": 0.019415007904171944,
+ "learning_rate": 0.0006,
+ "loss": 4.508941173553467,
+ "step": 1346
+ },
+ {
+ "epoch": 18.712663755458514,
+ "grad_norm": 0.019856126978993416,
+ "learning_rate": 0.0006,
+ "loss": 4.446670055389404,
+ "step": 1347
+ },
+ {
+ "epoch": 18.726637554585153,
+ "grad_norm": 0.01842520199716091,
+ "learning_rate": 0.0006,
+ "loss": 4.524422645568848,
+ "step": 1348
+ },
+ {
+ "epoch": 18.74061135371179,
+ "grad_norm": 0.017196480184793472,
+ "learning_rate": 0.0006,
+ "loss": 4.498466968536377,
+ "step": 1349
+ },
+ {
+ "epoch": 18.754585152838427,
+ "grad_norm": 0.01823010854423046,
+ "learning_rate": 0.0006,
+ "loss": 4.534684181213379,
+ "step": 1350
+ },
+ {
+ "epoch": 18.768558951965066,
+ "grad_norm": 0.020398065447807312,
+ "learning_rate": 0.0006,
+ "loss": 4.456494331359863,
+ "step": 1351
+ },
+ {
+ "epoch": 18.782532751091704,
+ "grad_norm": 0.022992262616753578,
+ "learning_rate": 0.0006,
+ "loss": 4.476374626159668,
+ "step": 1352
+ },
+ {
+ "epoch": 18.79650655021834,
+ "grad_norm": 0.02235420234501362,
+ "learning_rate": 0.0006,
+ "loss": 4.587776184082031,
+ "step": 1353
+ },
+ {
+ "epoch": 18.810480349344978,
+ "grad_norm": 0.02093626745045185,
+ "learning_rate": 0.0006,
+ "loss": 4.424172401428223,
+ "step": 1354
+ },
+ {
+ "epoch": 18.824454148471617,
+ "grad_norm": 0.01997302658855915,
+ "learning_rate": 0.0006,
+ "loss": 4.448066711425781,
+ "step": 1355
+ },
+ {
+ "epoch": 18.83842794759825,
+ "grad_norm": 0.018265459686517715,
+ "learning_rate": 0.0006,
+ "loss": 4.436610698699951,
+ "step": 1356
+ },
+ {
+ "epoch": 18.85240174672489,
+ "grad_norm": 0.01673833839595318,
+ "learning_rate": 0.0006,
+ "loss": 4.5502214431762695,
+ "step": 1357
+ },
+ {
+ "epoch": 18.86637554585153,
+ "grad_norm": 0.017908833920955658,
+ "learning_rate": 0.0006,
+ "loss": 4.550530433654785,
+ "step": 1358
+ },
+ {
+ "epoch": 18.880349344978168,
+ "grad_norm": 0.01796654611825943,
+ "learning_rate": 0.0006,
+ "loss": 4.531463146209717,
+ "step": 1359
+ },
+ {
+ "epoch": 18.894323144104803,
+ "grad_norm": 0.01646183803677559,
+ "learning_rate": 0.0006,
+ "loss": 4.498582363128662,
+ "step": 1360
+ },
+ {
+ "epoch": 18.90829694323144,
+ "grad_norm": 0.015430327504873276,
+ "learning_rate": 0.0006,
+ "loss": 4.52571439743042,
+ "step": 1361
+ },
+ {
+ "epoch": 18.92227074235808,
+ "grad_norm": 0.016527952626347542,
+ "learning_rate": 0.0006,
+ "loss": 4.547013759613037,
+ "step": 1362
+ },
+ {
+ "epoch": 18.936244541484715,
+ "grad_norm": 0.016263660043478012,
+ "learning_rate": 0.0006,
+ "loss": 4.468754291534424,
+ "step": 1363
+ },
+ {
+ "epoch": 18.950218340611354,
+ "grad_norm": 0.016347210854291916,
+ "learning_rate": 0.0006,
+ "loss": 4.450741767883301,
+ "step": 1364
+ },
+ {
+ "epoch": 18.964192139737992,
+ "grad_norm": 0.01701398566365242,
+ "learning_rate": 0.0006,
+ "loss": 4.534899711608887,
+ "step": 1365
+ },
+ {
+ "epoch": 18.978165938864628,
+ "grad_norm": 0.017693081870675087,
+ "learning_rate": 0.0006,
+ "loss": 4.460553169250488,
+ "step": 1366
+ },
+ {
+ "epoch": 18.992139737991266,
+ "grad_norm": 0.018746308982372284,
+ "learning_rate": 0.0006,
+ "loss": 4.538792610168457,
+ "step": 1367
+ },
+ {
+ "epoch": 19.0,
+ "grad_norm": 0.020099416375160217,
+ "learning_rate": 0.0006,
+ "loss": 4.449617385864258,
+ "step": 1368
+ },
+ {
+ "epoch": 19.0,
+ "eval_loss": 4.7756571769714355,
+ "eval_runtime": 57.1804,
+ "eval_samples_per_second": 42.707,
+ "eval_steps_per_second": 1.347,
+ "step": 1368
+ },
+ {
+ "epoch": 19.01397379912664,
+ "grad_norm": 0.018726933747529984,
+ "learning_rate": 0.0006,
+ "loss": 4.446059703826904,
+ "step": 1369
+ },
+ {
+ "epoch": 19.027947598253274,
+ "grad_norm": 0.01956087350845337,
+ "learning_rate": 0.0006,
+ "loss": 4.4520039558410645,
+ "step": 1370
+ },
+ {
+ "epoch": 19.041921397379912,
+ "grad_norm": 0.020092811435461044,
+ "learning_rate": 0.0006,
+ "loss": 4.543343544006348,
+ "step": 1371
+ },
+ {
+ "epoch": 19.05589519650655,
+ "grad_norm": 0.01843622885644436,
+ "learning_rate": 0.0006,
+ "loss": 4.535579204559326,
+ "step": 1372
+ },
+ {
+ "epoch": 19.069868995633186,
+ "grad_norm": 0.016989488154649734,
+ "learning_rate": 0.0006,
+ "loss": 4.52316951751709,
+ "step": 1373
+ },
+ {
+ "epoch": 19.083842794759825,
+ "grad_norm": 0.01805422455072403,
+ "learning_rate": 0.0006,
+ "loss": 4.5393266677856445,
+ "step": 1374
+ },
+ {
+ "epoch": 19.097816593886463,
+ "grad_norm": 0.01874629408121109,
+ "learning_rate": 0.0006,
+ "loss": 4.422059059143066,
+ "step": 1375
+ },
+ {
+ "epoch": 19.111790393013102,
+ "grad_norm": 0.01804221048951149,
+ "learning_rate": 0.0006,
+ "loss": 4.50508451461792,
+ "step": 1376
+ },
+ {
+ "epoch": 19.125764192139737,
+ "grad_norm": 0.01785474829375744,
+ "learning_rate": 0.0006,
+ "loss": 4.412599563598633,
+ "step": 1377
+ },
+ {
+ "epoch": 19.139737991266376,
+ "grad_norm": 0.01971791312098503,
+ "learning_rate": 0.0006,
+ "loss": 4.518318176269531,
+ "step": 1378
+ },
+ {
+ "epoch": 19.153711790393015,
+ "grad_norm": 0.02061633951961994,
+ "learning_rate": 0.0006,
+ "loss": 4.332393169403076,
+ "step": 1379
+ },
+ {
+ "epoch": 19.16768558951965,
+ "grad_norm": 0.020563429221510887,
+ "learning_rate": 0.0006,
+ "loss": 4.445863723754883,
+ "step": 1380
+ },
+ {
+ "epoch": 19.18165938864629,
+ "grad_norm": 0.019995713606476784,
+ "learning_rate": 0.0006,
+ "loss": 4.497312068939209,
+ "step": 1381
+ },
+ {
+ "epoch": 19.195633187772927,
+ "grad_norm": 0.019823070615530014,
+ "learning_rate": 0.0006,
+ "loss": 4.475085258483887,
+ "step": 1382
+ },
+ {
+ "epoch": 19.209606986899562,
+ "grad_norm": 0.019039355218410492,
+ "learning_rate": 0.0006,
+ "loss": 4.50432014465332,
+ "step": 1383
+ },
+ {
+ "epoch": 19.2235807860262,
+ "grad_norm": 0.017876973375678062,
+ "learning_rate": 0.0006,
+ "loss": 4.384788513183594,
+ "step": 1384
+ },
+ {
+ "epoch": 19.23755458515284,
+ "grad_norm": 0.018697699531912804,
+ "learning_rate": 0.0006,
+ "loss": 4.410778045654297,
+ "step": 1385
+ },
+ {
+ "epoch": 19.251528384279474,
+ "grad_norm": 0.018826806917786598,
+ "learning_rate": 0.0006,
+ "loss": 4.524034023284912,
+ "step": 1386
+ },
+ {
+ "epoch": 19.265502183406113,
+ "grad_norm": 0.01824226602911949,
+ "learning_rate": 0.0006,
+ "loss": 4.526393890380859,
+ "step": 1387
+ },
+ {
+ "epoch": 19.27947598253275,
+ "grad_norm": 0.01753208599984646,
+ "learning_rate": 0.0006,
+ "loss": 4.493703365325928,
+ "step": 1388
+ },
+ {
+ "epoch": 19.29344978165939,
+ "grad_norm": 0.01742432825267315,
+ "learning_rate": 0.0006,
+ "loss": 4.392831802368164,
+ "step": 1389
+ },
+ {
+ "epoch": 19.307423580786025,
+ "grad_norm": 0.017099356278777122,
+ "learning_rate": 0.0006,
+ "loss": 4.578668594360352,
+ "step": 1390
+ },
+ {
+ "epoch": 19.321397379912664,
+ "grad_norm": 0.017946023494005203,
+ "learning_rate": 0.0006,
+ "loss": 4.456718444824219,
+ "step": 1391
+ },
+ {
+ "epoch": 19.335371179039303,
+ "grad_norm": 0.017705943435430527,
+ "learning_rate": 0.0006,
+ "loss": 4.553682327270508,
+ "step": 1392
+ },
+ {
+ "epoch": 19.349344978165938,
+ "grad_norm": 0.016647523269057274,
+ "learning_rate": 0.0006,
+ "loss": 4.525016784667969,
+ "step": 1393
+ },
+ {
+ "epoch": 19.363318777292577,
+ "grad_norm": 0.015407416969537735,
+ "learning_rate": 0.0006,
+ "loss": 4.500231742858887,
+ "step": 1394
+ },
+ {
+ "epoch": 19.377292576419215,
+ "grad_norm": 0.017264019697904587,
+ "learning_rate": 0.0006,
+ "loss": 4.4615349769592285,
+ "step": 1395
+ },
+ {
+ "epoch": 19.39126637554585,
+ "grad_norm": 0.01871555484831333,
+ "learning_rate": 0.0006,
+ "loss": 4.281026840209961,
+ "step": 1396
+ },
+ {
+ "epoch": 19.40524017467249,
+ "grad_norm": 0.017361650243401527,
+ "learning_rate": 0.0006,
+ "loss": 4.479382514953613,
+ "step": 1397
+ },
+ {
+ "epoch": 19.419213973799128,
+ "grad_norm": 0.019781362265348434,
+ "learning_rate": 0.0006,
+ "loss": 4.402320861816406,
+ "step": 1398
+ },
+ {
+ "epoch": 19.433187772925763,
+ "grad_norm": 0.019043298438191414,
+ "learning_rate": 0.0006,
+ "loss": 4.476983070373535,
+ "step": 1399
+ },
+ {
+ "epoch": 19.4471615720524,
+ "grad_norm": 0.017137985676527023,
+ "learning_rate": 0.0006,
+ "loss": 4.418689250946045,
+ "step": 1400
+ },
+ {
+ "epoch": 19.46113537117904,
+ "grad_norm": 0.016481833532452583,
+ "learning_rate": 0.0006,
+ "loss": 4.579145431518555,
+ "step": 1401
+ },
+ {
+ "epoch": 19.475109170305675,
+ "grad_norm": 0.016507714986801147,
+ "learning_rate": 0.0006,
+ "loss": 4.519158363342285,
+ "step": 1402
+ },
+ {
+ "epoch": 19.489082969432314,
+ "grad_norm": 0.018092872574925423,
+ "learning_rate": 0.0006,
+ "loss": 4.513748645782471,
+ "step": 1403
+ },
+ {
+ "epoch": 19.503056768558952,
+ "grad_norm": 0.018153440207242966,
+ "learning_rate": 0.0006,
+ "loss": 4.645384311676025,
+ "step": 1404
+ },
+ {
+ "epoch": 19.51703056768559,
+ "grad_norm": 0.01657119207084179,
+ "learning_rate": 0.0006,
+ "loss": 4.399757385253906,
+ "step": 1405
+ },
+ {
+ "epoch": 19.531004366812226,
+ "grad_norm": 0.0156280305236578,
+ "learning_rate": 0.0006,
+ "loss": 4.467836380004883,
+ "step": 1406
+ },
+ {
+ "epoch": 19.544978165938865,
+ "grad_norm": 0.016838403418660164,
+ "learning_rate": 0.0006,
+ "loss": 4.529176235198975,
+ "step": 1407
+ },
+ {
+ "epoch": 19.558951965065503,
+ "grad_norm": 0.01743236929178238,
+ "learning_rate": 0.0006,
+ "loss": 4.440032482147217,
+ "step": 1408
+ },
+ {
+ "epoch": 19.57292576419214,
+ "grad_norm": 0.01553610060364008,
+ "learning_rate": 0.0006,
+ "loss": 4.425588607788086,
+ "step": 1409
+ },
+ {
+ "epoch": 19.586899563318777,
+ "grad_norm": 0.017259271815419197,
+ "learning_rate": 0.0006,
+ "loss": 4.405452728271484,
+ "step": 1410
+ },
+ {
+ "epoch": 19.600873362445416,
+ "grad_norm": 0.015568481758236885,
+ "learning_rate": 0.0006,
+ "loss": 4.444279193878174,
+ "step": 1411
+ },
+ {
+ "epoch": 19.61484716157205,
+ "grad_norm": 0.015331248752772808,
+ "learning_rate": 0.0006,
+ "loss": 4.451846122741699,
+ "step": 1412
+ },
+ {
+ "epoch": 19.62882096069869,
+ "grad_norm": 0.015667378902435303,
+ "learning_rate": 0.0006,
+ "loss": 4.533931255340576,
+ "step": 1413
+ },
+ {
+ "epoch": 19.64279475982533,
+ "grad_norm": 0.016177602112293243,
+ "learning_rate": 0.0006,
+ "loss": 4.467313289642334,
+ "step": 1414
+ },
+ {
+ "epoch": 19.656768558951963,
+ "grad_norm": 0.01733129471540451,
+ "learning_rate": 0.0006,
+ "loss": 4.496893882751465,
+ "step": 1415
+ },
+ {
+ "epoch": 19.670742358078602,
+ "grad_norm": 0.016539594158530235,
+ "learning_rate": 0.0006,
+ "loss": 4.401961326599121,
+ "step": 1416
+ },
+ {
+ "epoch": 19.68471615720524,
+ "grad_norm": 0.016058551147580147,
+ "learning_rate": 0.0006,
+ "loss": 4.503375053405762,
+ "step": 1417
+ },
+ {
+ "epoch": 19.69868995633188,
+ "grad_norm": 0.01715777814388275,
+ "learning_rate": 0.0006,
+ "loss": 4.3808369636535645,
+ "step": 1418
+ },
+ {
+ "epoch": 19.712663755458514,
+ "grad_norm": 0.01681119203567505,
+ "learning_rate": 0.0006,
+ "loss": 4.5469560623168945,
+ "step": 1419
+ },
+ {
+ "epoch": 19.726637554585153,
+ "grad_norm": 0.01634550653398037,
+ "learning_rate": 0.0006,
+ "loss": 4.328197956085205,
+ "step": 1420
+ },
+ {
+ "epoch": 19.74061135371179,
+ "grad_norm": 0.01549347210675478,
+ "learning_rate": 0.0006,
+ "loss": 4.534621238708496,
+ "step": 1421
+ },
+ {
+ "epoch": 19.754585152838427,
+ "grad_norm": 0.01654665172100067,
+ "learning_rate": 0.0006,
+ "loss": 4.4724225997924805,
+ "step": 1422
+ },
+ {
+ "epoch": 19.768558951965066,
+ "grad_norm": 0.017119025811553,
+ "learning_rate": 0.0006,
+ "loss": 4.3897247314453125,
+ "step": 1423
+ },
+ {
+ "epoch": 19.782532751091704,
+ "grad_norm": 0.01603223942220211,
+ "learning_rate": 0.0006,
+ "loss": 4.412731170654297,
+ "step": 1424
+ },
+ {
+ "epoch": 19.79650655021834,
+ "grad_norm": 0.016774067655205727,
+ "learning_rate": 0.0006,
+ "loss": 4.54707145690918,
+ "step": 1425
+ },
+ {
+ "epoch": 19.810480349344978,
+ "grad_norm": 0.016453025862574577,
+ "learning_rate": 0.0006,
+ "loss": 4.462501049041748,
+ "step": 1426
+ },
+ {
+ "epoch": 19.824454148471617,
+ "grad_norm": 0.01597629487514496,
+ "learning_rate": 0.0006,
+ "loss": 4.229758262634277,
+ "step": 1427
+ },
+ {
+ "epoch": 19.83842794759825,
+ "grad_norm": 0.017999034374952316,
+ "learning_rate": 0.0006,
+ "loss": 4.389263153076172,
+ "step": 1428
+ },
+ {
+ "epoch": 19.85240174672489,
+ "grad_norm": 0.02165898121893406,
+ "learning_rate": 0.0006,
+ "loss": 4.342130661010742,
+ "step": 1429
+ },
+ {
+ "epoch": 19.86637554585153,
+ "grad_norm": 0.02222822792828083,
+ "learning_rate": 0.0006,
+ "loss": 4.492661476135254,
+ "step": 1430
+ },
+ {
+ "epoch": 19.880349344978168,
+ "grad_norm": 0.019378066062927246,
+ "learning_rate": 0.0006,
+ "loss": 4.4614105224609375,
+ "step": 1431
+ },
+ {
+ "epoch": 19.894323144104803,
+ "grad_norm": 0.019908083602786064,
+ "learning_rate": 0.0006,
+ "loss": 4.370929718017578,
+ "step": 1432
+ },
+ {
+ "epoch": 19.90829694323144,
+ "grad_norm": 0.022296471521258354,
+ "learning_rate": 0.0006,
+ "loss": 4.438320159912109,
+ "step": 1433
+ },
+ {
+ "epoch": 19.92227074235808,
+ "grad_norm": 0.021810343489050865,
+ "learning_rate": 0.0006,
+ "loss": 4.462067127227783,
+ "step": 1434
+ },
+ {
+ "epoch": 19.936244541484715,
+ "grad_norm": 0.02129209227859974,
+ "learning_rate": 0.0006,
+ "loss": 4.541261196136475,
+ "step": 1435
+ },
+ {
+ "epoch": 19.950218340611354,
+ "grad_norm": 0.02079629711806774,
+ "learning_rate": 0.0006,
+ "loss": 4.36712646484375,
+ "step": 1436
+ },
+ {
+ "epoch": 19.964192139737992,
+ "grad_norm": 0.022501740604639053,
+ "learning_rate": 0.0006,
+ "loss": 4.5279083251953125,
+ "step": 1437
+ },
+ {
+ "epoch": 19.978165938864628,
+ "grad_norm": 0.019121676683425903,
+ "learning_rate": 0.0006,
+ "loss": 4.477789878845215,
+ "step": 1438
+ },
+ {
+ "epoch": 19.992139737991266,
+ "grad_norm": 0.017429588362574577,
+ "learning_rate": 0.0006,
+ "loss": 4.4023847579956055,
+ "step": 1439
+ },
+ {
+ "epoch": 20.0,
+ "grad_norm": 0.018857665359973907,
+ "learning_rate": 0.0006,
+ "loss": 4.494743347167969,
+ "step": 1440
+ },
+ {
+ "epoch": 20.0,
+ "eval_loss": 4.694555759429932,
+ "eval_runtime": 56.4146,
+ "eval_samples_per_second": 43.287,
+ "eval_steps_per_second": 1.365,
+ "step": 1440
+ },
+ {
+ "epoch": 20.01397379912664,
+ "grad_norm": 0.016505861654877663,
+ "learning_rate": 0.0006,
+ "loss": 4.506229877471924,
+ "step": 1441
+ },
+ {
+ "epoch": 20.027947598253274,
+ "grad_norm": 0.017571842297911644,
+ "learning_rate": 0.0006,
+ "loss": 4.384073257446289,
+ "step": 1442
+ },
+ {
+ "epoch": 20.041921397379912,
+ "grad_norm": 0.019398365169763565,
+ "learning_rate": 0.0006,
+ "loss": 4.482177734375,
+ "step": 1443
+ },
+ {
+ "epoch": 20.05589519650655,
+ "grad_norm": 0.02011062018573284,
+ "learning_rate": 0.0006,
+ "loss": 4.440773963928223,
+ "step": 1444
+ },
+ {
+ "epoch": 20.069868995633186,
+ "grad_norm": 0.019026953727006912,
+ "learning_rate": 0.0006,
+ "loss": 4.507168769836426,
+ "step": 1445
+ },
+ {
+ "epoch": 20.083842794759825,
+ "grad_norm": 0.020506031811237335,
+ "learning_rate": 0.0006,
+ "loss": 4.469079971313477,
+ "step": 1446
+ },
+ {
+ "epoch": 20.097816593886463,
+ "grad_norm": 0.019608162343502045,
+ "learning_rate": 0.0006,
+ "loss": 4.395927429199219,
+ "step": 1447
+ },
+ {
+ "epoch": 20.111790393013102,
+ "grad_norm": 0.019419124349951744,
+ "learning_rate": 0.0006,
+ "loss": 4.484235763549805,
+ "step": 1448
+ },
+ {
+ "epoch": 20.125764192139737,
+ "grad_norm": 0.021775128319859505,
+ "learning_rate": 0.0006,
+ "loss": 4.333173751831055,
+ "step": 1449
+ },
+ {
+ "epoch": 20.139737991266376,
+ "grad_norm": 0.025512272492051125,
+ "learning_rate": 0.0006,
+ "loss": 4.460362434387207,
+ "step": 1450
+ },
+ {
+ "epoch": 20.153711790393015,
+ "grad_norm": 0.025157401338219643,
+ "learning_rate": 0.0006,
+ "loss": 4.4361395835876465,
+ "step": 1451
+ },
+ {
+ "epoch": 20.16768558951965,
+ "grad_norm": 0.019862616434693336,
+ "learning_rate": 0.0006,
+ "loss": 4.426807403564453,
+ "step": 1452
+ },
+ {
+ "epoch": 20.18165938864629,
+ "grad_norm": 0.01932152360677719,
+ "learning_rate": 0.0006,
+ "loss": 4.375677108764648,
+ "step": 1453
+ },
+ {
+ "epoch": 20.195633187772927,
+ "grad_norm": 0.01760999858379364,
+ "learning_rate": 0.0006,
+ "loss": 4.42259407043457,
+ "step": 1454
+ },
+ {
+ "epoch": 20.209606986899562,
+ "grad_norm": 0.016998112201690674,
+ "learning_rate": 0.0006,
+ "loss": 4.481280326843262,
+ "step": 1455
+ },
+ {
+ "epoch": 20.2235807860262,
+ "grad_norm": 0.017199190333485603,
+ "learning_rate": 0.0006,
+ "loss": 4.496712684631348,
+ "step": 1456
+ },
+ {
+ "epoch": 20.23755458515284,
+ "grad_norm": 0.016833679750561714,
+ "learning_rate": 0.0006,
+ "loss": 4.40736722946167,
+ "step": 1457
+ },
+ {
+ "epoch": 20.251528384279474,
+ "grad_norm": 0.017546730116009712,
+ "learning_rate": 0.0006,
+ "loss": 4.4915266036987305,
+ "step": 1458
+ },
+ {
+ "epoch": 20.265502183406113,
+ "grad_norm": 0.016595078632235527,
+ "learning_rate": 0.0006,
+ "loss": 4.460815906524658,
+ "step": 1459
+ },
+ {
+ "epoch": 20.27947598253275,
+ "grad_norm": 0.017728568986058235,
+ "learning_rate": 0.0006,
+ "loss": 4.346048831939697,
+ "step": 1460
+ },
+ {
+ "epoch": 20.29344978165939,
+ "grad_norm": 0.01849197782576084,
+ "learning_rate": 0.0006,
+ "loss": 4.399681091308594,
+ "step": 1461
+ },
+ {
+ "epoch": 20.307423580786025,
+ "grad_norm": 0.018336588516831398,
+ "learning_rate": 0.0006,
+ "loss": 4.414183616638184,
+ "step": 1462
+ },
+ {
+ "epoch": 20.321397379912664,
+ "grad_norm": 0.018086519092321396,
+ "learning_rate": 0.0006,
+ "loss": 4.346586227416992,
+ "step": 1463
+ },
+ {
+ "epoch": 20.335371179039303,
+ "grad_norm": 0.019697299227118492,
+ "learning_rate": 0.0006,
+ "loss": 4.4420952796936035,
+ "step": 1464
+ },
+ {
+ "epoch": 20.349344978165938,
+ "grad_norm": 0.02198074571788311,
+ "learning_rate": 0.0006,
+ "loss": 4.588729381561279,
+ "step": 1465
+ },
+ {
+ "epoch": 20.363318777292577,
+ "grad_norm": 0.022813329473137856,
+ "learning_rate": 0.0006,
+ "loss": 4.49729585647583,
+ "step": 1466
+ },
+ {
+ "epoch": 20.377292576419215,
+ "grad_norm": 0.022006649523973465,
+ "learning_rate": 0.0006,
+ "loss": 4.475641250610352,
+ "step": 1467
+ },
+ {
+ "epoch": 20.39126637554585,
+ "grad_norm": 0.023188438266515732,
+ "learning_rate": 0.0006,
+ "loss": 4.404322624206543,
+ "step": 1468
+ },
+ {
+ "epoch": 20.40524017467249,
+ "grad_norm": 0.022226877510547638,
+ "learning_rate": 0.0006,
+ "loss": 4.50904655456543,
+ "step": 1469
+ },
+ {
+ "epoch": 20.419213973799128,
+ "grad_norm": 0.021082637831568718,
+ "learning_rate": 0.0006,
+ "loss": 4.339011192321777,
+ "step": 1470
+ },
+ {
+ "epoch": 20.433187772925763,
+ "grad_norm": 0.02013542130589485,
+ "learning_rate": 0.0006,
+ "loss": 4.516942024230957,
+ "step": 1471
+ },
+ {
+ "epoch": 20.4471615720524,
+ "grad_norm": 0.01898968778550625,
+ "learning_rate": 0.0006,
+ "loss": 4.315804481506348,
+ "step": 1472
+ },
+ {
+ "epoch": 20.46113537117904,
+ "grad_norm": 0.019450880587100983,
+ "learning_rate": 0.0006,
+ "loss": 4.4794921875,
+ "step": 1473
+ },
+ {
+ "epoch": 20.475109170305675,
+ "grad_norm": 0.018577802926301956,
+ "learning_rate": 0.0006,
+ "loss": 4.422323703765869,
+ "step": 1474
+ },
+ {
+ "epoch": 20.489082969432314,
+ "grad_norm": 0.01778412237763405,
+ "learning_rate": 0.0006,
+ "loss": 4.429789066314697,
+ "step": 1475
+ },
+ {
+ "epoch": 20.503056768558952,
+ "grad_norm": 0.017134476453065872,
+ "learning_rate": 0.0006,
+ "loss": 4.416459083557129,
+ "step": 1476
+ },
+ {
+ "epoch": 20.51703056768559,
+ "grad_norm": 0.017153557389974594,
+ "learning_rate": 0.0006,
+ "loss": 4.446170806884766,
+ "step": 1477
+ },
+ {
+ "epoch": 20.531004366812226,
+ "grad_norm": 0.014998532831668854,
+ "learning_rate": 0.0006,
+ "loss": 4.419447898864746,
+ "step": 1478
+ },
+ {
+ "epoch": 20.544978165938865,
+ "grad_norm": 0.015877850353717804,
+ "learning_rate": 0.0006,
+ "loss": 4.312338829040527,
+ "step": 1479
+ },
+ {
+ "epoch": 20.558951965065503,
+ "grad_norm": 0.01569489948451519,
+ "learning_rate": 0.0006,
+ "loss": 4.572905540466309,
+ "step": 1480
+ },
+ {
+ "epoch": 20.57292576419214,
+ "grad_norm": 0.01591474376618862,
+ "learning_rate": 0.0006,
+ "loss": 4.414907455444336,
+ "step": 1481
+ },
+ {
+ "epoch": 20.586899563318777,
+ "grad_norm": 0.015395903028547764,
+ "learning_rate": 0.0006,
+ "loss": 4.549131393432617,
+ "step": 1482
+ },
+ {
+ "epoch": 20.600873362445416,
+ "grad_norm": 0.015715358778834343,
+ "learning_rate": 0.0006,
+ "loss": 4.530194282531738,
+ "step": 1483
+ },
+ {
+ "epoch": 20.61484716157205,
+ "grad_norm": 0.016042975708842278,
+ "learning_rate": 0.0006,
+ "loss": 4.322149276733398,
+ "step": 1484
+ },
+ {
+ "epoch": 20.62882096069869,
+ "grad_norm": 0.01664186641573906,
+ "learning_rate": 0.0006,
+ "loss": 4.424866676330566,
+ "step": 1485
+ },
+ {
+ "epoch": 20.64279475982533,
+ "grad_norm": 0.01631534844636917,
+ "learning_rate": 0.0006,
+ "loss": 4.441752910614014,
+ "step": 1486
+ },
+ {
+ "epoch": 20.656768558951963,
+ "grad_norm": 0.014057058840990067,
+ "learning_rate": 0.0006,
+ "loss": 4.399982452392578,
+ "step": 1487
+ },
+ {
+ "epoch": 20.670742358078602,
+ "grad_norm": 0.015076273120939732,
+ "learning_rate": 0.0006,
+ "loss": 4.441324234008789,
+ "step": 1488
+ },
+ {
+ "epoch": 20.68471615720524,
+ "grad_norm": 0.014138277620077133,
+ "learning_rate": 0.0006,
+ "loss": 4.4312543869018555,
+ "step": 1489
+ },
+ {
+ "epoch": 20.69868995633188,
+ "grad_norm": 0.013762440532445908,
+ "learning_rate": 0.0006,
+ "loss": 4.387807846069336,
+ "step": 1490
+ },
+ {
+ "epoch": 20.712663755458514,
+ "grad_norm": 0.015058411285281181,
+ "learning_rate": 0.0006,
+ "loss": 4.594747543334961,
+ "step": 1491
+ },
+ {
+ "epoch": 20.726637554585153,
+ "grad_norm": 0.013827123679220676,
+ "learning_rate": 0.0006,
+ "loss": 4.434604167938232,
+ "step": 1492
+ },
+ {
+ "epoch": 20.74061135371179,
+ "grad_norm": 0.015315295197069645,
+ "learning_rate": 0.0006,
+ "loss": 4.588066101074219,
+ "step": 1493
+ },
+ {
+ "epoch": 20.754585152838427,
+ "grad_norm": 0.018678732216358185,
+ "learning_rate": 0.0006,
+ "loss": 4.3514204025268555,
+ "step": 1494
+ },
+ {
+ "epoch": 20.768558951965066,
+ "grad_norm": 0.019521350041031837,
+ "learning_rate": 0.0006,
+ "loss": 4.437658786773682,
+ "step": 1495
+ },
+ {
+ "epoch": 20.782532751091704,
+ "grad_norm": 0.018363403156399727,
+ "learning_rate": 0.0006,
+ "loss": 4.495002269744873,
+ "step": 1496
+ },
+ {
+ "epoch": 20.79650655021834,
+ "grad_norm": 0.016830725595355034,
+ "learning_rate": 0.0006,
+ "loss": 4.402471542358398,
+ "step": 1497
+ },
+ {
+ "epoch": 20.810480349344978,
+ "grad_norm": 0.01722072623670101,
+ "learning_rate": 0.0006,
+ "loss": 4.437653541564941,
+ "step": 1498
+ },
+ {
+ "epoch": 20.824454148471617,
+ "grad_norm": 0.017459599301218987,
+ "learning_rate": 0.0006,
+ "loss": 4.366603851318359,
+ "step": 1499
+ },
+ {
+ "epoch": 20.83842794759825,
+ "grad_norm": 0.018902219831943512,
+ "learning_rate": 0.0006,
+ "loss": 4.501561164855957,
+ "step": 1500
+ },
+ {
+ "epoch": 20.85240174672489,
+ "grad_norm": 0.020814096555113792,
+ "learning_rate": 0.0006,
+ "loss": 4.535295486450195,
+ "step": 1501
+ },
+ {
+ "epoch": 20.86637554585153,
+ "grad_norm": 0.018990976735949516,
+ "learning_rate": 0.0006,
+ "loss": 4.333580017089844,
+ "step": 1502
+ },
+ {
+ "epoch": 20.880349344978168,
+ "grad_norm": 0.01837880350649357,
+ "learning_rate": 0.0006,
+ "loss": 4.313388824462891,
+ "step": 1503
+ },
+ {
+ "epoch": 20.894323144104803,
+ "grad_norm": 0.019820360466837883,
+ "learning_rate": 0.0006,
+ "loss": 4.388157844543457,
+ "step": 1504
+ },
+ {
+ "epoch": 20.90829694323144,
+ "grad_norm": 0.020181143656373024,
+ "learning_rate": 0.0006,
+ "loss": 4.572224140167236,
+ "step": 1505
+ },
+ {
+ "epoch": 20.92227074235808,
+ "grad_norm": 0.021206077188253403,
+ "learning_rate": 0.0006,
+ "loss": 4.3127970695495605,
+ "step": 1506
+ },
+ {
+ "epoch": 20.936244541484715,
+ "grad_norm": 0.017962060868740082,
+ "learning_rate": 0.0006,
+ "loss": 4.384734630584717,
+ "step": 1507
+ },
+ {
+ "epoch": 20.950218340611354,
+ "grad_norm": 0.01737220585346222,
+ "learning_rate": 0.0006,
+ "loss": 4.4075751304626465,
+ "step": 1508
+ },
+ {
+ "epoch": 20.964192139737992,
+ "grad_norm": 0.01839268207550049,
+ "learning_rate": 0.0006,
+ "loss": 4.553779125213623,
+ "step": 1509
+ },
+ {
+ "epoch": 20.978165938864628,
+ "grad_norm": 0.020102640613913536,
+ "learning_rate": 0.0006,
+ "loss": 4.399398326873779,
+ "step": 1510
+ },
+ {
+ "epoch": 20.992139737991266,
+ "grad_norm": 0.018831463530659676,
+ "learning_rate": 0.0006,
+ "loss": 4.358499526977539,
+ "step": 1511
+ },
+ {
+ "epoch": 21.0,
+ "grad_norm": 0.018254734575748444,
+ "learning_rate": 0.0006,
+ "loss": 4.384119510650635,
+ "step": 1512
+ },
+ {
+ "epoch": 21.0,
+ "eval_loss": 4.7524237632751465,
+ "eval_runtime": 57.8812,
+ "eval_samples_per_second": 42.19,
+ "eval_steps_per_second": 1.33,
+ "step": 1512
+ },
+ {
+ "epoch": 21.01397379912664,
+ "grad_norm": 0.019728587940335274,
+ "learning_rate": 0.0006,
+ "loss": 4.32363224029541,
+ "step": 1513
+ },
+ {
+ "epoch": 21.027947598253274,
+ "grad_norm": 0.02216625027358532,
+ "learning_rate": 0.0006,
+ "loss": 4.436988353729248,
+ "step": 1514
+ },
+ {
+ "epoch": 21.041921397379912,
+ "grad_norm": 0.02227606810629368,
+ "learning_rate": 0.0006,
+ "loss": 4.406063556671143,
+ "step": 1515
+ },
+ {
+ "epoch": 21.05589519650655,
+ "grad_norm": 0.020187009125947952,
+ "learning_rate": 0.0006,
+ "loss": 4.324402332305908,
+ "step": 1516
+ },
+ {
+ "epoch": 21.069868995633186,
+ "grad_norm": 0.02095157466828823,
+ "learning_rate": 0.0006,
+ "loss": 4.456544876098633,
+ "step": 1517
+ },
+ {
+ "epoch": 21.083842794759825,
+ "grad_norm": 0.02146013453602791,
+ "learning_rate": 0.0006,
+ "loss": 4.368387222290039,
+ "step": 1518
+ },
+ {
+ "epoch": 21.097816593886463,
+ "grad_norm": 0.02162035182118416,
+ "learning_rate": 0.0006,
+ "loss": 4.48838996887207,
+ "step": 1519
+ },
+ {
+ "epoch": 21.111790393013102,
+ "grad_norm": 0.021745676174759865,
+ "learning_rate": 0.0006,
+ "loss": 4.30778694152832,
+ "step": 1520
+ },
+ {
+ "epoch": 21.125764192139737,
+ "grad_norm": 0.02173588052392006,
+ "learning_rate": 0.0006,
+ "loss": 4.433149814605713,
+ "step": 1521
+ },
+ {
+ "epoch": 21.139737991266376,
+ "grad_norm": 0.020950648933649063,
+ "learning_rate": 0.0006,
+ "loss": 4.520089149475098,
+ "step": 1522
+ },
+ {
+ "epoch": 21.153711790393015,
+ "grad_norm": 0.020684752613306046,
+ "learning_rate": 0.0006,
+ "loss": 4.372437477111816,
+ "step": 1523
+ },
+ {
+ "epoch": 21.16768558951965,
+ "grad_norm": 0.019472869113087654,
+ "learning_rate": 0.0006,
+ "loss": 4.437015533447266,
+ "step": 1524
+ },
+ {
+ "epoch": 21.18165938864629,
+ "grad_norm": 0.01846962980926037,
+ "learning_rate": 0.0006,
+ "loss": 4.449531555175781,
+ "step": 1525
+ },
+ {
+ "epoch": 21.195633187772927,
+ "grad_norm": 0.017377803102135658,
+ "learning_rate": 0.0006,
+ "loss": 4.458279132843018,
+ "step": 1526
+ },
+ {
+ "epoch": 21.209606986899562,
+ "grad_norm": 0.016002200543880463,
+ "learning_rate": 0.0006,
+ "loss": 4.266297817230225,
+ "step": 1527
+ },
+ {
+ "epoch": 21.2235807860262,
+ "grad_norm": 0.015575903467833996,
+ "learning_rate": 0.0006,
+ "loss": 4.342632293701172,
+ "step": 1528
+ },
+ {
+ "epoch": 21.23755458515284,
+ "grad_norm": 0.015634698793292046,
+ "learning_rate": 0.0006,
+ "loss": 4.338623523712158,
+ "step": 1529
+ },
+ {
+ "epoch": 21.251528384279474,
+ "grad_norm": 0.01449581328779459,
+ "learning_rate": 0.0006,
+ "loss": 4.354330062866211,
+ "step": 1530
+ },
+ {
+ "epoch": 21.265502183406113,
+ "grad_norm": 0.01487616915255785,
+ "learning_rate": 0.0006,
+ "loss": 4.486835479736328,
+ "step": 1531
+ },
+ {
+ "epoch": 21.27947598253275,
+ "grad_norm": 0.016346555203199387,
+ "learning_rate": 0.0006,
+ "loss": 4.407098770141602,
+ "step": 1532
+ },
+ {
+ "epoch": 21.29344978165939,
+ "grad_norm": 0.01682182401418686,
+ "learning_rate": 0.0006,
+ "loss": 4.436809539794922,
+ "step": 1533
+ },
+ {
+ "epoch": 21.307423580786025,
+ "grad_norm": 0.017704004421830177,
+ "learning_rate": 0.0006,
+ "loss": 4.3854780197143555,
+ "step": 1534
+ },
+ {
+ "epoch": 21.321397379912664,
+ "grad_norm": 0.018437420949339867,
+ "learning_rate": 0.0006,
+ "loss": 4.413283824920654,
+ "step": 1535
+ },
+ {
+ "epoch": 21.335371179039303,
+ "grad_norm": 0.016911081969738007,
+ "learning_rate": 0.0006,
+ "loss": 4.357306003570557,
+ "step": 1536
+ },
+ {
+ "epoch": 21.349344978165938,
+ "grad_norm": 0.01642223633825779,
+ "learning_rate": 0.0006,
+ "loss": 4.422322750091553,
+ "step": 1537
+ },
+ {
+ "epoch": 21.363318777292577,
+ "grad_norm": 0.01780877821147442,
+ "learning_rate": 0.0006,
+ "loss": 4.387620449066162,
+ "step": 1538
+ },
+ {
+ "epoch": 21.377292576419215,
+ "grad_norm": 0.01653975248336792,
+ "learning_rate": 0.0006,
+ "loss": 4.327267646789551,
+ "step": 1539
+ },
+ {
+ "epoch": 21.39126637554585,
+ "grad_norm": 0.01572156324982643,
+ "learning_rate": 0.0006,
+ "loss": 4.475979804992676,
+ "step": 1540
+ },
+ {
+ "epoch": 21.40524017467249,
+ "grad_norm": 0.017156066372990608,
+ "learning_rate": 0.0006,
+ "loss": 4.4011125564575195,
+ "step": 1541
+ },
+ {
+ "epoch": 21.419213973799128,
+ "grad_norm": 0.01788829080760479,
+ "learning_rate": 0.0006,
+ "loss": 4.4174089431762695,
+ "step": 1542
+ },
+ {
+ "epoch": 21.433187772925763,
+ "grad_norm": 0.01728985086083412,
+ "learning_rate": 0.0006,
+ "loss": 4.429244041442871,
+ "step": 1543
+ },
+ {
+ "epoch": 21.4471615720524,
+ "grad_norm": 0.016534466296434402,
+ "learning_rate": 0.0006,
+ "loss": 4.326068878173828,
+ "step": 1544
+ },
+ {
+ "epoch": 21.46113537117904,
+ "grad_norm": 0.016582584008574486,
+ "learning_rate": 0.0006,
+ "loss": 4.2975969314575195,
+ "step": 1545
+ },
+ {
+ "epoch": 21.475109170305675,
+ "grad_norm": 0.018034178763628006,
+ "learning_rate": 0.0006,
+ "loss": 4.501319408416748,
+ "step": 1546
+ },
+ {
+ "epoch": 21.489082969432314,
+ "grad_norm": 0.017827702686190605,
+ "learning_rate": 0.0006,
+ "loss": 4.502284526824951,
+ "step": 1547
+ },
+ {
+ "epoch": 21.503056768558952,
+ "grad_norm": 0.017182299867272377,
+ "learning_rate": 0.0006,
+ "loss": 4.380828857421875,
+ "step": 1548
+ },
+ {
+ "epoch": 21.51703056768559,
+ "grad_norm": 0.019101588055491447,
+ "learning_rate": 0.0006,
+ "loss": 4.31564998626709,
+ "step": 1549
+ },
+ {
+ "epoch": 21.531004366812226,
+ "grad_norm": 0.0204677302390337,
+ "learning_rate": 0.0006,
+ "loss": 4.3840131759643555,
+ "step": 1550
+ },
+ {
+ "epoch": 21.544978165938865,
+ "grad_norm": 0.0188993439078331,
+ "learning_rate": 0.0006,
+ "loss": 4.4257731437683105,
+ "step": 1551
+ },
+ {
+ "epoch": 21.558951965065503,
+ "grad_norm": 0.01730230636894703,
+ "learning_rate": 0.0006,
+ "loss": 4.445132255554199,
+ "step": 1552
+ },
+ {
+ "epoch": 21.57292576419214,
+ "grad_norm": 0.01801574043929577,
+ "learning_rate": 0.0006,
+ "loss": 4.427064895629883,
+ "step": 1553
+ },
+ {
+ "epoch": 21.586899563318777,
+ "grad_norm": 0.018417565152049065,
+ "learning_rate": 0.0006,
+ "loss": 4.319595813751221,
+ "step": 1554
+ },
+ {
+ "epoch": 21.600873362445416,
+ "grad_norm": 0.01683160290122032,
+ "learning_rate": 0.0006,
+ "loss": 4.349218368530273,
+ "step": 1555
+ },
+ {
+ "epoch": 21.61484716157205,
+ "grad_norm": 0.016973106190562248,
+ "learning_rate": 0.0006,
+ "loss": 4.375990390777588,
+ "step": 1556
+ },
+ {
+ "epoch": 21.62882096069869,
+ "grad_norm": 0.015688583254814148,
+ "learning_rate": 0.0006,
+ "loss": 4.408170223236084,
+ "step": 1557
+ },
+ {
+ "epoch": 21.64279475982533,
+ "grad_norm": 0.01573043316602707,
+ "learning_rate": 0.0006,
+ "loss": 4.416255950927734,
+ "step": 1558
+ },
+ {
+ "epoch": 21.656768558951963,
+ "grad_norm": 0.01744293048977852,
+ "learning_rate": 0.0006,
+ "loss": 4.477141380310059,
+ "step": 1559
+ },
+ {
+ "epoch": 21.670742358078602,
+ "grad_norm": 0.017503513023257256,
+ "learning_rate": 0.0006,
+ "loss": 4.3367486000061035,
+ "step": 1560
+ },
+ {
+ "epoch": 21.68471615720524,
+ "grad_norm": 0.016592150554060936,
+ "learning_rate": 0.0006,
+ "loss": 4.314785480499268,
+ "step": 1561
+ },
+ {
+ "epoch": 21.69868995633188,
+ "grad_norm": 0.015774084255099297,
+ "learning_rate": 0.0006,
+ "loss": 4.448209285736084,
+ "step": 1562
+ },
+ {
+ "epoch": 21.712663755458514,
+ "grad_norm": 0.015779150649905205,
+ "learning_rate": 0.0006,
+ "loss": 4.392690658569336,
+ "step": 1563
+ },
+ {
+ "epoch": 21.726637554585153,
+ "grad_norm": 0.015392648056149483,
+ "learning_rate": 0.0006,
+ "loss": 4.437741756439209,
+ "step": 1564
+ },
+ {
+ "epoch": 21.74061135371179,
+ "grad_norm": 0.01530836895108223,
+ "learning_rate": 0.0006,
+ "loss": 4.424595832824707,
+ "step": 1565
+ },
+ {
+ "epoch": 21.754585152838427,
+ "grad_norm": 0.015600779093801975,
+ "learning_rate": 0.0006,
+ "loss": 4.335417747497559,
+ "step": 1566
+ },
+ {
+ "epoch": 21.768558951965066,
+ "grad_norm": 0.014646518044173717,
+ "learning_rate": 0.0006,
+ "loss": 4.279081344604492,
+ "step": 1567
+ },
+ {
+ "epoch": 21.782532751091704,
+ "grad_norm": 0.014132864773273468,
+ "learning_rate": 0.0006,
+ "loss": 4.345950126647949,
+ "step": 1568
+ },
+ {
+ "epoch": 21.79650655021834,
+ "grad_norm": 0.01425724383443594,
+ "learning_rate": 0.0006,
+ "loss": 4.277588844299316,
+ "step": 1569
+ },
+ {
+ "epoch": 21.810480349344978,
+ "grad_norm": 0.015396572649478912,
+ "learning_rate": 0.0006,
+ "loss": 4.326196670532227,
+ "step": 1570
+ },
+ {
+ "epoch": 21.824454148471617,
+ "grad_norm": 0.01683669537305832,
+ "learning_rate": 0.0006,
+ "loss": 4.3294172286987305,
+ "step": 1571
+ },
+ {
+ "epoch": 21.83842794759825,
+ "grad_norm": 0.016365647315979004,
+ "learning_rate": 0.0006,
+ "loss": 4.523321628570557,
+ "step": 1572
+ },
+ {
+ "epoch": 21.85240174672489,
+ "grad_norm": 0.017142774537205696,
+ "learning_rate": 0.0006,
+ "loss": 4.475842475891113,
+ "step": 1573
+ },
+ {
+ "epoch": 21.86637554585153,
+ "grad_norm": 0.015950839966535568,
+ "learning_rate": 0.0006,
+ "loss": 4.397187232971191,
+ "step": 1574
+ },
+ {
+ "epoch": 21.880349344978168,
+ "grad_norm": 0.016180362552404404,
+ "learning_rate": 0.0006,
+ "loss": 4.414778709411621,
+ "step": 1575
+ },
+ {
+ "epoch": 21.894323144104803,
+ "grad_norm": 0.017615556716918945,
+ "learning_rate": 0.0006,
+ "loss": 4.348567962646484,
+ "step": 1576
+ },
+ {
+ "epoch": 21.90829694323144,
+ "grad_norm": 0.01896277628839016,
+ "learning_rate": 0.0006,
+ "loss": 4.431762218475342,
+ "step": 1577
+ },
+ {
+ "epoch": 21.92227074235808,
+ "grad_norm": 0.01990971714258194,
+ "learning_rate": 0.0006,
+ "loss": 4.431285858154297,
+ "step": 1578
+ },
+ {
+ "epoch": 21.936244541484715,
+ "grad_norm": 0.0201258547604084,
+ "learning_rate": 0.0006,
+ "loss": 4.309872627258301,
+ "step": 1579
+ },
+ {
+ "epoch": 21.950218340611354,
+ "grad_norm": 0.019620198756456375,
+ "learning_rate": 0.0006,
+ "loss": 4.484044551849365,
+ "step": 1580
+ },
+ {
+ "epoch": 21.964192139737992,
+ "grad_norm": 0.01771123707294464,
+ "learning_rate": 0.0006,
+ "loss": 4.368093013763428,
+ "step": 1581
+ },
+ {
+ "epoch": 21.978165938864628,
+ "grad_norm": 0.017501654103398323,
+ "learning_rate": 0.0006,
+ "loss": 4.372651100158691,
+ "step": 1582
+ },
+ {
+ "epoch": 21.992139737991266,
+ "grad_norm": 0.01693608984351158,
+ "learning_rate": 0.0006,
+ "loss": 4.379167556762695,
+ "step": 1583
+ },
+ {
+ "epoch": 22.0,
+ "grad_norm": 0.02002849616110325,
+ "learning_rate": 0.0006,
+ "loss": 4.329627990722656,
+ "step": 1584
+ },
+ {
+ "epoch": 22.0,
+ "eval_loss": 4.723696231842041,
+ "eval_runtime": 56.1758,
+ "eval_samples_per_second": 43.471,
+ "eval_steps_per_second": 1.371,
+ "step": 1584
+ },
+ {
+ "epoch": 22.01397379912664,
+ "grad_norm": 0.02030305378139019,
+ "learning_rate": 0.0006,
+ "loss": 4.355566024780273,
+ "step": 1585
+ },
+ {
+ "epoch": 22.027947598253274,
+ "grad_norm": 0.018638934940099716,
+ "learning_rate": 0.0006,
+ "loss": 4.314009189605713,
+ "step": 1586
+ },
+ {
+ "epoch": 22.041921397379912,
+ "grad_norm": 0.019917072728276253,
+ "learning_rate": 0.0006,
+ "loss": 4.367525577545166,
+ "step": 1587
+ },
+ {
+ "epoch": 22.05589519650655,
+ "grad_norm": 0.020785167813301086,
+ "learning_rate": 0.0006,
+ "loss": 4.375561714172363,
+ "step": 1588
+ },
+ {
+ "epoch": 22.069868995633186,
+ "grad_norm": 0.021214453503489494,
+ "learning_rate": 0.0006,
+ "loss": 4.418305397033691,
+ "step": 1589
+ },
+ {
+ "epoch": 22.083842794759825,
+ "grad_norm": 0.019051678478717804,
+ "learning_rate": 0.0006,
+ "loss": 4.215126991271973,
+ "step": 1590
+ },
+ {
+ "epoch": 22.097816593886463,
+ "grad_norm": 0.02323366142809391,
+ "learning_rate": 0.0006,
+ "loss": 4.306807994842529,
+ "step": 1591
+ },
+ {
+ "epoch": 22.111790393013102,
+ "grad_norm": 0.026186738163232803,
+ "learning_rate": 0.0006,
+ "loss": 4.408508777618408,
+ "step": 1592
+ },
+ {
+ "epoch": 22.125764192139737,
+ "grad_norm": 0.027361227199435234,
+ "learning_rate": 0.0006,
+ "loss": 4.329322814941406,
+ "step": 1593
+ },
+ {
+ "epoch": 22.139737991266376,
+ "grad_norm": 0.026564646512269974,
+ "learning_rate": 0.0006,
+ "loss": 4.346642017364502,
+ "step": 1594
+ },
+ {
+ "epoch": 22.153711790393015,
+ "grad_norm": 0.025073667988181114,
+ "learning_rate": 0.0006,
+ "loss": 4.3984198570251465,
+ "step": 1595
+ },
+ {
+ "epoch": 22.16768558951965,
+ "grad_norm": 0.02347794733941555,
+ "learning_rate": 0.0006,
+ "loss": 4.352538108825684,
+ "step": 1596
+ },
+ {
+ "epoch": 22.18165938864629,
+ "grad_norm": 0.02203468047082424,
+ "learning_rate": 0.0006,
+ "loss": 4.298148155212402,
+ "step": 1597
+ },
+ {
+ "epoch": 22.195633187772927,
+ "grad_norm": 0.020621957257390022,
+ "learning_rate": 0.0006,
+ "loss": 4.480309963226318,
+ "step": 1598
+ },
+ {
+ "epoch": 22.209606986899562,
+ "grad_norm": 0.023710055276751518,
+ "learning_rate": 0.0006,
+ "loss": 4.339317321777344,
+ "step": 1599
+ },
+ {
+ "epoch": 22.2235807860262,
+ "grad_norm": 0.0233482476323843,
+ "learning_rate": 0.0006,
+ "loss": 4.436896324157715,
+ "step": 1600
+ },
+ {
+ "epoch": 22.23755458515284,
+ "grad_norm": 0.020122403278946877,
+ "learning_rate": 0.0006,
+ "loss": 4.236002445220947,
+ "step": 1601
+ },
+ {
+ "epoch": 22.251528384279474,
+ "grad_norm": 0.01957266591489315,
+ "learning_rate": 0.0006,
+ "loss": 4.497328281402588,
+ "step": 1602
+ },
+ {
+ "epoch": 22.265502183406113,
+ "grad_norm": 0.018964895978569984,
+ "learning_rate": 0.0006,
+ "loss": 4.36360502243042,
+ "step": 1603
+ },
+ {
+ "epoch": 22.27947598253275,
+ "grad_norm": 0.018902769312262535,
+ "learning_rate": 0.0006,
+ "loss": 4.427093505859375,
+ "step": 1604
+ },
+ {
+ "epoch": 22.29344978165939,
+ "grad_norm": 0.018918119370937347,
+ "learning_rate": 0.0006,
+ "loss": 4.294950485229492,
+ "step": 1605
+ },
+ {
+ "epoch": 22.307423580786025,
+ "grad_norm": 0.019315775483846664,
+ "learning_rate": 0.0006,
+ "loss": 4.427952289581299,
+ "step": 1606
+ },
+ {
+ "epoch": 22.321397379912664,
+ "grad_norm": 0.018350522965192795,
+ "learning_rate": 0.0006,
+ "loss": 4.272945404052734,
+ "step": 1607
+ },
+ {
+ "epoch": 22.335371179039303,
+ "grad_norm": 0.01759534701704979,
+ "learning_rate": 0.0006,
+ "loss": 4.348023891448975,
+ "step": 1608
+ },
+ {
+ "epoch": 22.349344978165938,
+ "grad_norm": 0.01646057888865471,
+ "learning_rate": 0.0006,
+ "loss": 4.278518199920654,
+ "step": 1609
+ },
+ {
+ "epoch": 22.363318777292577,
+ "grad_norm": 0.015875060111284256,
+ "learning_rate": 0.0006,
+ "loss": 4.423527240753174,
+ "step": 1610
+ },
+ {
+ "epoch": 22.377292576419215,
+ "grad_norm": 0.01638529635965824,
+ "learning_rate": 0.0006,
+ "loss": 4.321723937988281,
+ "step": 1611
+ },
+ {
+ "epoch": 22.39126637554585,
+ "grad_norm": 0.016942061483860016,
+ "learning_rate": 0.0006,
+ "loss": 4.346164703369141,
+ "step": 1612
+ },
+ {
+ "epoch": 22.40524017467249,
+ "grad_norm": 0.015209621749818325,
+ "learning_rate": 0.0006,
+ "loss": 4.434340000152588,
+ "step": 1613
+ },
+ {
+ "epoch": 22.419213973799128,
+ "grad_norm": 0.014423009008169174,
+ "learning_rate": 0.0006,
+ "loss": 4.342560768127441,
+ "step": 1614
+ },
+ {
+ "epoch": 22.433187772925763,
+ "grad_norm": 0.014369525946676731,
+ "learning_rate": 0.0006,
+ "loss": 4.296586990356445,
+ "step": 1615
+ },
+ {
+ "epoch": 22.4471615720524,
+ "grad_norm": 0.01392540242522955,
+ "learning_rate": 0.0006,
+ "loss": 4.433176040649414,
+ "step": 1616
+ },
+ {
+ "epoch": 22.46113537117904,
+ "grad_norm": 0.013206899166107178,
+ "learning_rate": 0.0006,
+ "loss": 4.32962703704834,
+ "step": 1617
+ },
+ {
+ "epoch": 22.475109170305675,
+ "grad_norm": 0.013427077792584896,
+ "learning_rate": 0.0006,
+ "loss": 4.420437335968018,
+ "step": 1618
+ },
+ {
+ "epoch": 22.489082969432314,
+ "grad_norm": 0.014194196090102196,
+ "learning_rate": 0.0006,
+ "loss": 4.410478591918945,
+ "step": 1619
+ },
+ {
+ "epoch": 22.503056768558952,
+ "grad_norm": 0.01362364087253809,
+ "learning_rate": 0.0006,
+ "loss": 4.338932037353516,
+ "step": 1620
+ },
+ {
+ "epoch": 22.51703056768559,
+ "grad_norm": 0.014299589209258556,
+ "learning_rate": 0.0006,
+ "loss": 4.34089994430542,
+ "step": 1621
+ },
+ {
+ "epoch": 22.531004366812226,
+ "grad_norm": 0.014061295427381992,
+ "learning_rate": 0.0006,
+ "loss": 4.29201602935791,
+ "step": 1622
+ },
+ {
+ "epoch": 22.544978165938865,
+ "grad_norm": 0.013945229351520538,
+ "learning_rate": 0.0006,
+ "loss": 4.326052188873291,
+ "step": 1623
+ },
+ {
+ "epoch": 22.558951965065503,
+ "grad_norm": 0.01387068536132574,
+ "learning_rate": 0.0006,
+ "loss": 4.393521308898926,
+ "step": 1624
+ },
+ {
+ "epoch": 22.57292576419214,
+ "grad_norm": 0.013931898400187492,
+ "learning_rate": 0.0006,
+ "loss": 4.3739752769470215,
+ "step": 1625
+ },
+ {
+ "epoch": 22.586899563318777,
+ "grad_norm": 0.01497740764170885,
+ "learning_rate": 0.0006,
+ "loss": 4.3991618156433105,
+ "step": 1626
+ },
+ {
+ "epoch": 22.600873362445416,
+ "grad_norm": 0.015842905268073082,
+ "learning_rate": 0.0006,
+ "loss": 4.405340671539307,
+ "step": 1627
+ },
+ {
+ "epoch": 22.61484716157205,
+ "grad_norm": 0.01580633595585823,
+ "learning_rate": 0.0006,
+ "loss": 4.280102729797363,
+ "step": 1628
+ },
+ {
+ "epoch": 22.62882096069869,
+ "grad_norm": 0.016567382961511612,
+ "learning_rate": 0.0006,
+ "loss": 4.407657146453857,
+ "step": 1629
+ },
+ {
+ "epoch": 22.64279475982533,
+ "grad_norm": 0.01648622378706932,
+ "learning_rate": 0.0006,
+ "loss": 4.392590522766113,
+ "step": 1630
+ },
+ {
+ "epoch": 22.656768558951963,
+ "grad_norm": 0.017066307365894318,
+ "learning_rate": 0.0006,
+ "loss": 4.324862003326416,
+ "step": 1631
+ },
+ {
+ "epoch": 22.670742358078602,
+ "grad_norm": 0.01637062057852745,
+ "learning_rate": 0.0006,
+ "loss": 4.34598445892334,
+ "step": 1632
+ },
+ {
+ "epoch": 22.68471615720524,
+ "grad_norm": 0.018038960173726082,
+ "learning_rate": 0.0006,
+ "loss": 4.320711612701416,
+ "step": 1633
+ },
+ {
+ "epoch": 22.69868995633188,
+ "grad_norm": 0.018096428364515305,
+ "learning_rate": 0.0006,
+ "loss": 4.536527156829834,
+ "step": 1634
+ },
+ {
+ "epoch": 22.712663755458514,
+ "grad_norm": 0.017889291048049927,
+ "learning_rate": 0.0006,
+ "loss": 4.388131141662598,
+ "step": 1635
+ },
+ {
+ "epoch": 22.726637554585153,
+ "grad_norm": 0.018771521747112274,
+ "learning_rate": 0.0006,
+ "loss": 4.357668876647949,
+ "step": 1636
+ },
+ {
+ "epoch": 22.74061135371179,
+ "grad_norm": 0.02100873365998268,
+ "learning_rate": 0.0006,
+ "loss": 4.293972015380859,
+ "step": 1637
+ },
+ {
+ "epoch": 22.754585152838427,
+ "grad_norm": 0.017715785652399063,
+ "learning_rate": 0.0006,
+ "loss": 4.349849224090576,
+ "step": 1638
+ },
+ {
+ "epoch": 22.768558951965066,
+ "grad_norm": 0.015546333976089954,
+ "learning_rate": 0.0006,
+ "loss": 4.311728000640869,
+ "step": 1639
+ },
+ {
+ "epoch": 22.782532751091704,
+ "grad_norm": 0.01605990156531334,
+ "learning_rate": 0.0006,
+ "loss": 4.499123573303223,
+ "step": 1640
+ },
+ {
+ "epoch": 22.79650655021834,
+ "grad_norm": 0.016642142087221146,
+ "learning_rate": 0.0006,
+ "loss": 4.4319257736206055,
+ "step": 1641
+ },
+ {
+ "epoch": 22.810480349344978,
+ "grad_norm": 0.016120387241244316,
+ "learning_rate": 0.0006,
+ "loss": 4.4373650550842285,
+ "step": 1642
+ },
+ {
+ "epoch": 22.824454148471617,
+ "grad_norm": 0.015174890868365765,
+ "learning_rate": 0.0006,
+ "loss": 4.314082145690918,
+ "step": 1643
+ },
+ {
+ "epoch": 22.83842794759825,
+ "grad_norm": 0.01551714539527893,
+ "learning_rate": 0.0006,
+ "loss": 4.360363960266113,
+ "step": 1644
+ },
+ {
+ "epoch": 22.85240174672489,
+ "grad_norm": 0.014966742135584354,
+ "learning_rate": 0.0006,
+ "loss": 4.346253395080566,
+ "step": 1645
+ },
+ {
+ "epoch": 22.86637554585153,
+ "grad_norm": 0.014288028702139854,
+ "learning_rate": 0.0006,
+ "loss": 4.464366436004639,
+ "step": 1646
+ },
+ {
+ "epoch": 22.880349344978168,
+ "grad_norm": 0.015291682444512844,
+ "learning_rate": 0.0006,
+ "loss": 4.504101753234863,
+ "step": 1647
+ },
+ {
+ "epoch": 22.894323144104803,
+ "grad_norm": 0.014534426853060722,
+ "learning_rate": 0.0006,
+ "loss": 4.462172031402588,
+ "step": 1648
+ },
+ {
+ "epoch": 22.90829694323144,
+ "grad_norm": 0.014129819348454475,
+ "learning_rate": 0.0006,
+ "loss": 4.350126266479492,
+ "step": 1649
+ },
+ {
+ "epoch": 22.92227074235808,
+ "grad_norm": 0.014932668767869473,
+ "learning_rate": 0.0006,
+ "loss": 4.374392032623291,
+ "step": 1650
+ },
+ {
+ "epoch": 22.936244541484715,
+ "grad_norm": 0.014647853560745716,
+ "learning_rate": 0.0006,
+ "loss": 4.252264499664307,
+ "step": 1651
+ },
+ {
+ "epoch": 22.950218340611354,
+ "grad_norm": 0.013553360477089882,
+ "learning_rate": 0.0006,
+ "loss": 4.379729270935059,
+ "step": 1652
+ },
+ {
+ "epoch": 22.964192139737992,
+ "grad_norm": 0.014606194570660591,
+ "learning_rate": 0.0006,
+ "loss": 4.363372802734375,
+ "step": 1653
+ },
+ {
+ "epoch": 22.978165938864628,
+ "grad_norm": 0.015000631101429462,
+ "learning_rate": 0.0006,
+ "loss": 4.408048629760742,
+ "step": 1654
+ },
+ {
+ "epoch": 22.992139737991266,
+ "grad_norm": 0.015370816923677921,
+ "learning_rate": 0.0006,
+ "loss": 4.297451019287109,
+ "step": 1655
+ },
+ {
+ "epoch": 23.0,
+ "grad_norm": 0.01709144562482834,
+ "learning_rate": 0.0006,
+ "loss": 4.294776916503906,
+ "step": 1656
+ },
+ {
+ "epoch": 23.0,
+ "eval_loss": 4.678618907928467,
+ "eval_runtime": 56.1668,
+ "eval_samples_per_second": 43.478,
+ "eval_steps_per_second": 1.371,
+ "step": 1656
+ },
+ {
+ "epoch": 23.01397379912664,
+ "grad_norm": 0.01727290451526642,
+ "learning_rate": 0.0006,
+ "loss": 4.311701774597168,
+ "step": 1657
+ },
+ {
+ "epoch": 23.027947598253274,
+ "grad_norm": 0.020645759999752045,
+ "learning_rate": 0.0006,
+ "loss": 4.429897308349609,
+ "step": 1658
+ },
+ {
+ "epoch": 23.041921397379912,
+ "grad_norm": 0.02204226516187191,
+ "learning_rate": 0.0006,
+ "loss": 4.365151882171631,
+ "step": 1659
+ },
+ {
+ "epoch": 23.05589519650655,
+ "grad_norm": 0.023069269955158234,
+ "learning_rate": 0.0006,
+ "loss": 4.3710832595825195,
+ "step": 1660
+ },
+ {
+ "epoch": 23.069868995633186,
+ "grad_norm": 0.024151179939508438,
+ "learning_rate": 0.0006,
+ "loss": 4.403108596801758,
+ "step": 1661
+ },
+ {
+ "epoch": 23.083842794759825,
+ "grad_norm": 0.02511233650147915,
+ "learning_rate": 0.0006,
+ "loss": 4.414128303527832,
+ "step": 1662
+ },
+ {
+ "epoch": 23.097816593886463,
+ "grad_norm": 0.025802113115787506,
+ "learning_rate": 0.0006,
+ "loss": 4.388265609741211,
+ "step": 1663
+ },
+ {
+ "epoch": 23.111790393013102,
+ "grad_norm": 0.02377348765730858,
+ "learning_rate": 0.0006,
+ "loss": 4.357505798339844,
+ "step": 1664
+ },
+ {
+ "epoch": 23.125764192139737,
+ "grad_norm": 0.020876651629805565,
+ "learning_rate": 0.0006,
+ "loss": 4.23043966293335,
+ "step": 1665
+ },
+ {
+ "epoch": 23.139737991266376,
+ "grad_norm": 0.01959727331995964,
+ "learning_rate": 0.0006,
+ "loss": 4.280970096588135,
+ "step": 1666
+ },
+ {
+ "epoch": 23.153711790393015,
+ "grad_norm": 0.018678616732358932,
+ "learning_rate": 0.0006,
+ "loss": 4.346716403961182,
+ "step": 1667
+ },
+ {
+ "epoch": 23.16768558951965,
+ "grad_norm": 0.02108944021165371,
+ "learning_rate": 0.0006,
+ "loss": 4.277904510498047,
+ "step": 1668
+ },
+ {
+ "epoch": 23.18165938864629,
+ "grad_norm": 0.020221566781401634,
+ "learning_rate": 0.0006,
+ "loss": 4.367693901062012,
+ "step": 1669
+ },
+ {
+ "epoch": 23.195633187772927,
+ "grad_norm": 0.019599711522459984,
+ "learning_rate": 0.0006,
+ "loss": 4.257092475891113,
+ "step": 1670
+ },
+ {
+ "epoch": 23.209606986899562,
+ "grad_norm": 0.019833361729979515,
+ "learning_rate": 0.0006,
+ "loss": 4.540737628936768,
+ "step": 1671
+ },
+ {
+ "epoch": 23.2235807860262,
+ "grad_norm": 0.020352095365524292,
+ "learning_rate": 0.0006,
+ "loss": 4.360923767089844,
+ "step": 1672
+ },
+ {
+ "epoch": 23.23755458515284,
+ "grad_norm": 0.01939748041331768,
+ "learning_rate": 0.0006,
+ "loss": 4.393110275268555,
+ "step": 1673
+ },
+ {
+ "epoch": 23.251528384279474,
+ "grad_norm": 0.01884354278445244,
+ "learning_rate": 0.0006,
+ "loss": 4.383938789367676,
+ "step": 1674
+ },
+ {
+ "epoch": 23.265502183406113,
+ "grad_norm": 0.02008470520377159,
+ "learning_rate": 0.0006,
+ "loss": 4.346344947814941,
+ "step": 1675
+ },
+ {
+ "epoch": 23.27947598253275,
+ "grad_norm": 0.019065195694565773,
+ "learning_rate": 0.0006,
+ "loss": 4.36611795425415,
+ "step": 1676
+ },
+ {
+ "epoch": 23.29344978165939,
+ "grad_norm": 0.01938541606068611,
+ "learning_rate": 0.0006,
+ "loss": 4.4009599685668945,
+ "step": 1677
+ },
+ {
+ "epoch": 23.307423580786025,
+ "grad_norm": 0.019044723361730576,
+ "learning_rate": 0.0006,
+ "loss": 4.340169906616211,
+ "step": 1678
+ },
+ {
+ "epoch": 23.321397379912664,
+ "grad_norm": 0.018850378692150116,
+ "learning_rate": 0.0006,
+ "loss": 4.37674617767334,
+ "step": 1679
+ },
+ {
+ "epoch": 23.335371179039303,
+ "grad_norm": 0.019076095893979073,
+ "learning_rate": 0.0006,
+ "loss": 4.226874351501465,
+ "step": 1680
+ },
+ {
+ "epoch": 23.349344978165938,
+ "grad_norm": 0.017543038353323936,
+ "learning_rate": 0.0006,
+ "loss": 4.337751388549805,
+ "step": 1681
+ },
+ {
+ "epoch": 23.363318777292577,
+ "grad_norm": 0.017170218750834465,
+ "learning_rate": 0.0006,
+ "loss": 4.282512664794922,
+ "step": 1682
+ },
+ {
+ "epoch": 23.377292576419215,
+ "grad_norm": 0.01574796624481678,
+ "learning_rate": 0.0006,
+ "loss": 4.30638313293457,
+ "step": 1683
+ },
+ {
+ "epoch": 23.39126637554585,
+ "grad_norm": 0.015160846523940563,
+ "learning_rate": 0.0006,
+ "loss": 4.336716175079346,
+ "step": 1684
+ },
+ {
+ "epoch": 23.40524017467249,
+ "grad_norm": 0.01534526702016592,
+ "learning_rate": 0.0006,
+ "loss": 4.2174787521362305,
+ "step": 1685
+ },
+ {
+ "epoch": 23.419213973799128,
+ "grad_norm": 0.015289999544620514,
+ "learning_rate": 0.0006,
+ "loss": 4.272546291351318,
+ "step": 1686
+ },
+ {
+ "epoch": 23.433187772925763,
+ "grad_norm": 0.014894185587763786,
+ "learning_rate": 0.0006,
+ "loss": 4.296045303344727,
+ "step": 1687
+ },
+ {
+ "epoch": 23.4471615720524,
+ "grad_norm": 0.014869502745568752,
+ "learning_rate": 0.0006,
+ "loss": 4.425461769104004,
+ "step": 1688
+ },
+ {
+ "epoch": 23.46113537117904,
+ "grad_norm": 0.01474483497440815,
+ "learning_rate": 0.0006,
+ "loss": 4.361867427825928,
+ "step": 1689
+ },
+ {
+ "epoch": 23.475109170305675,
+ "grad_norm": 0.015299245715141296,
+ "learning_rate": 0.0006,
+ "loss": 4.364523410797119,
+ "step": 1690
+ },
+ {
+ "epoch": 23.489082969432314,
+ "grad_norm": 0.015201340429484844,
+ "learning_rate": 0.0006,
+ "loss": 4.285182952880859,
+ "step": 1691
+ },
+ {
+ "epoch": 23.503056768558952,
+ "grad_norm": 0.016057293862104416,
+ "learning_rate": 0.0006,
+ "loss": 4.284873962402344,
+ "step": 1692
+ },
+ {
+ "epoch": 23.51703056768559,
+ "grad_norm": 0.015871062874794006,
+ "learning_rate": 0.0006,
+ "loss": 4.326470851898193,
+ "step": 1693
+ },
+ {
+ "epoch": 23.531004366812226,
+ "grad_norm": 0.01586288772523403,
+ "learning_rate": 0.0006,
+ "loss": 4.324883460998535,
+ "step": 1694
+ },
+ {
+ "epoch": 23.544978165938865,
+ "grad_norm": 0.016092827543616295,
+ "learning_rate": 0.0006,
+ "loss": 4.3258867263793945,
+ "step": 1695
+ },
+ {
+ "epoch": 23.558951965065503,
+ "grad_norm": 0.01582016795873642,
+ "learning_rate": 0.0006,
+ "loss": 4.358892440795898,
+ "step": 1696
+ },
+ {
+ "epoch": 23.57292576419214,
+ "grad_norm": 0.01688452623784542,
+ "learning_rate": 0.0006,
+ "loss": 4.280842304229736,
+ "step": 1697
+ },
+ {
+ "epoch": 23.586899563318777,
+ "grad_norm": 0.015865886583924294,
+ "learning_rate": 0.0006,
+ "loss": 4.245567321777344,
+ "step": 1698
+ },
+ {
+ "epoch": 23.600873362445416,
+ "grad_norm": 0.015303662046790123,
+ "learning_rate": 0.0006,
+ "loss": 4.383112907409668,
+ "step": 1699
+ },
+ {
+ "epoch": 23.61484716157205,
+ "grad_norm": 0.015649516135454178,
+ "learning_rate": 0.0006,
+ "loss": 4.519082069396973,
+ "step": 1700
+ },
+ {
+ "epoch": 23.62882096069869,
+ "grad_norm": 0.0167120099067688,
+ "learning_rate": 0.0006,
+ "loss": 4.447847366333008,
+ "step": 1701
+ },
+ {
+ "epoch": 23.64279475982533,
+ "grad_norm": 0.016368716955184937,
+ "learning_rate": 0.0006,
+ "loss": 4.434186935424805,
+ "step": 1702
+ },
+ {
+ "epoch": 23.656768558951963,
+ "grad_norm": 0.017379503697156906,
+ "learning_rate": 0.0006,
+ "loss": 4.317594528198242,
+ "step": 1703
+ },
+ {
+ "epoch": 23.670742358078602,
+ "grad_norm": 0.018573811277747154,
+ "learning_rate": 0.0006,
+ "loss": 4.205793857574463,
+ "step": 1704
+ },
+ {
+ "epoch": 23.68471615720524,
+ "grad_norm": 0.01749090477824211,
+ "learning_rate": 0.0006,
+ "loss": 4.3985700607299805,
+ "step": 1705
+ },
+ {
+ "epoch": 23.69868995633188,
+ "grad_norm": 0.0165668074041605,
+ "learning_rate": 0.0006,
+ "loss": 4.329226970672607,
+ "step": 1706
+ },
+ {
+ "epoch": 23.712663755458514,
+ "grad_norm": 0.016913846135139465,
+ "learning_rate": 0.0006,
+ "loss": 4.388485908508301,
+ "step": 1707
+ },
+ {
+ "epoch": 23.726637554585153,
+ "grad_norm": 0.016813859343528748,
+ "learning_rate": 0.0006,
+ "loss": 4.3395915031433105,
+ "step": 1708
+ },
+ {
+ "epoch": 23.74061135371179,
+ "grad_norm": 0.01676975190639496,
+ "learning_rate": 0.0006,
+ "loss": 4.254745960235596,
+ "step": 1709
+ },
+ {
+ "epoch": 23.754585152838427,
+ "grad_norm": 0.016370350494980812,
+ "learning_rate": 0.0006,
+ "loss": 4.2612223625183105,
+ "step": 1710
+ },
+ {
+ "epoch": 23.768558951965066,
+ "grad_norm": 0.01696198247373104,
+ "learning_rate": 0.0006,
+ "loss": 4.327524662017822,
+ "step": 1711
+ },
+ {
+ "epoch": 23.782532751091704,
+ "grad_norm": 0.0189108457416296,
+ "learning_rate": 0.0006,
+ "loss": 4.325839519500732,
+ "step": 1712
+ },
+ {
+ "epoch": 23.79650655021834,
+ "grad_norm": 0.01862351782619953,
+ "learning_rate": 0.0006,
+ "loss": 4.333662509918213,
+ "step": 1713
+ },
+ {
+ "epoch": 23.810480349344978,
+ "grad_norm": 0.01655014418065548,
+ "learning_rate": 0.0006,
+ "loss": 4.39305305480957,
+ "step": 1714
+ },
+ {
+ "epoch": 23.824454148471617,
+ "grad_norm": 0.015899376943707466,
+ "learning_rate": 0.0006,
+ "loss": 4.390053749084473,
+ "step": 1715
+ },
+ {
+ "epoch": 23.83842794759825,
+ "grad_norm": 0.016519056633114815,
+ "learning_rate": 0.0006,
+ "loss": 4.255903244018555,
+ "step": 1716
+ },
+ {
+ "epoch": 23.85240174672489,
+ "grad_norm": 0.0167539119720459,
+ "learning_rate": 0.0006,
+ "loss": 4.255727767944336,
+ "step": 1717
+ },
+ {
+ "epoch": 23.86637554585153,
+ "grad_norm": 0.0171353816986084,
+ "learning_rate": 0.0006,
+ "loss": 4.376740455627441,
+ "step": 1718
+ },
+ {
+ "epoch": 23.880349344978168,
+ "grad_norm": 0.016719117760658264,
+ "learning_rate": 0.0006,
+ "loss": 4.3376688957214355,
+ "step": 1719
+ },
+ {
+ "epoch": 23.894323144104803,
+ "grad_norm": 0.01561494916677475,
+ "learning_rate": 0.0006,
+ "loss": 4.3345746994018555,
+ "step": 1720
+ },
+ {
+ "epoch": 23.90829694323144,
+ "grad_norm": 0.016303174197673798,
+ "learning_rate": 0.0006,
+ "loss": 4.332371234893799,
+ "step": 1721
+ },
+ {
+ "epoch": 23.92227074235808,
+ "grad_norm": 0.016722865402698517,
+ "learning_rate": 0.0006,
+ "loss": 4.267125129699707,
+ "step": 1722
+ },
+ {
+ "epoch": 23.936244541484715,
+ "grad_norm": 0.017072124406695366,
+ "learning_rate": 0.0006,
+ "loss": 4.322751998901367,
+ "step": 1723
+ },
+ {
+ "epoch": 23.950218340611354,
+ "grad_norm": 0.01699932850897312,
+ "learning_rate": 0.0006,
+ "loss": 4.378422737121582,
+ "step": 1724
+ },
+ {
+ "epoch": 23.964192139737992,
+ "grad_norm": 0.01742720976471901,
+ "learning_rate": 0.0006,
+ "loss": 4.345083236694336,
+ "step": 1725
+ },
+ {
+ "epoch": 23.978165938864628,
+ "grad_norm": 0.018797501921653748,
+ "learning_rate": 0.0006,
+ "loss": 4.290443420410156,
+ "step": 1726
+ },
+ {
+ "epoch": 23.992139737991266,
+ "grad_norm": 0.018090683966875076,
+ "learning_rate": 0.0006,
+ "loss": 4.377580642700195,
+ "step": 1727
+ },
+ {
+ "epoch": 24.0,
+ "grad_norm": 0.017305459827184677,
+ "learning_rate": 0.0006,
+ "loss": 4.364426612854004,
+ "step": 1728
+ },
+ {
+ "epoch": 24.0,
+ "eval_loss": 4.647762298583984,
+ "eval_runtime": 56.5402,
+ "eval_samples_per_second": 43.191,
+ "eval_steps_per_second": 1.362,
+ "step": 1728
+ },
+ {
+ "epoch": 24.01397379912664,
+ "grad_norm": 0.016493607312440872,
+ "learning_rate": 0.0006,
+ "loss": 4.252255916595459,
+ "step": 1729
+ },
+ {
+ "epoch": 24.027947598253274,
+ "grad_norm": 0.016529450193047523,
+ "learning_rate": 0.0006,
+ "loss": 4.352746963500977,
+ "step": 1730
+ },
+ {
+ "epoch": 24.041921397379912,
+ "grad_norm": 0.016940191388130188,
+ "learning_rate": 0.0006,
+ "loss": 4.301870822906494,
+ "step": 1731
+ },
+ {
+ "epoch": 24.05589519650655,
+ "grad_norm": 0.01752558909356594,
+ "learning_rate": 0.0006,
+ "loss": 4.2762451171875,
+ "step": 1732
+ },
+ {
+ "epoch": 24.069868995633186,
+ "grad_norm": 0.017112884670495987,
+ "learning_rate": 0.0006,
+ "loss": 4.2659149169921875,
+ "step": 1733
+ },
+ {
+ "epoch": 24.083842794759825,
+ "grad_norm": 0.01900586113333702,
+ "learning_rate": 0.0006,
+ "loss": 4.306058883666992,
+ "step": 1734
+ },
+ {
+ "epoch": 24.097816593886463,
+ "grad_norm": 0.019447680562734604,
+ "learning_rate": 0.0006,
+ "loss": 4.450128555297852,
+ "step": 1735
+ },
+ {
+ "epoch": 24.111790393013102,
+ "grad_norm": 0.01948186755180359,
+ "learning_rate": 0.0006,
+ "loss": 4.202869415283203,
+ "step": 1736
+ },
+ {
+ "epoch": 24.125764192139737,
+ "grad_norm": 0.01716383546590805,
+ "learning_rate": 0.0006,
+ "loss": 4.318876266479492,
+ "step": 1737
+ },
+ {
+ "epoch": 24.139737991266376,
+ "grad_norm": 0.017049988731741905,
+ "learning_rate": 0.0006,
+ "loss": 4.237238883972168,
+ "step": 1738
+ },
+ {
+ "epoch": 24.153711790393015,
+ "grad_norm": 0.015169923193752766,
+ "learning_rate": 0.0006,
+ "loss": 4.289426803588867,
+ "step": 1739
+ },
+ {
+ "epoch": 24.16768558951965,
+ "grad_norm": 0.014760667458176613,
+ "learning_rate": 0.0006,
+ "loss": 4.356924057006836,
+ "step": 1740
+ },
+ {
+ "epoch": 24.18165938864629,
+ "grad_norm": 0.015127750113606453,
+ "learning_rate": 0.0006,
+ "loss": 4.142916679382324,
+ "step": 1741
+ },
+ {
+ "epoch": 24.195633187772927,
+ "grad_norm": 0.01549555640667677,
+ "learning_rate": 0.0006,
+ "loss": 4.345563888549805,
+ "step": 1742
+ },
+ {
+ "epoch": 24.209606986899562,
+ "grad_norm": 0.016393091529607773,
+ "learning_rate": 0.0006,
+ "loss": 4.417495250701904,
+ "step": 1743
+ },
+ {
+ "epoch": 24.2235807860262,
+ "grad_norm": 0.01785266026854515,
+ "learning_rate": 0.0006,
+ "loss": 4.3553547859191895,
+ "step": 1744
+ },
+ {
+ "epoch": 24.23755458515284,
+ "grad_norm": 0.017574617639183998,
+ "learning_rate": 0.0006,
+ "loss": 4.241024494171143,
+ "step": 1745
+ },
+ {
+ "epoch": 24.251528384279474,
+ "grad_norm": 0.019060306251049042,
+ "learning_rate": 0.0006,
+ "loss": 4.3685150146484375,
+ "step": 1746
+ },
+ {
+ "epoch": 24.265502183406113,
+ "grad_norm": 0.0214633010327816,
+ "learning_rate": 0.0006,
+ "loss": 4.323866844177246,
+ "step": 1747
+ },
+ {
+ "epoch": 24.27947598253275,
+ "grad_norm": 0.021055003628134727,
+ "learning_rate": 0.0006,
+ "loss": 4.278757095336914,
+ "step": 1748
+ },
+ {
+ "epoch": 24.29344978165939,
+ "grad_norm": 0.018618572503328323,
+ "learning_rate": 0.0006,
+ "loss": 4.28849983215332,
+ "step": 1749
+ },
+ {
+ "epoch": 24.307423580786025,
+ "grad_norm": 0.018666435033082962,
+ "learning_rate": 0.0006,
+ "loss": 4.368590354919434,
+ "step": 1750
+ },
+ {
+ "epoch": 24.321397379912664,
+ "grad_norm": 0.020626146346330643,
+ "learning_rate": 0.0006,
+ "loss": 4.2951836585998535,
+ "step": 1751
+ },
+ {
+ "epoch": 24.335371179039303,
+ "grad_norm": 0.01929759792983532,
+ "learning_rate": 0.0006,
+ "loss": 4.3664703369140625,
+ "step": 1752
+ },
+ {
+ "epoch": 24.349344978165938,
+ "grad_norm": 0.020956002175807953,
+ "learning_rate": 0.0006,
+ "loss": 4.345308303833008,
+ "step": 1753
+ },
+ {
+ "epoch": 24.363318777292577,
+ "grad_norm": 0.019227957352995872,
+ "learning_rate": 0.0006,
+ "loss": 4.349849224090576,
+ "step": 1754
+ },
+ {
+ "epoch": 24.377292576419215,
+ "grad_norm": 0.015970097854733467,
+ "learning_rate": 0.0006,
+ "loss": 4.223012447357178,
+ "step": 1755
+ },
+ {
+ "epoch": 24.39126637554585,
+ "grad_norm": 0.015951646491885185,
+ "learning_rate": 0.0006,
+ "loss": 4.236420631408691,
+ "step": 1756
+ },
+ {
+ "epoch": 24.40524017467249,
+ "grad_norm": 0.016032515093684196,
+ "learning_rate": 0.0006,
+ "loss": 4.226899147033691,
+ "step": 1757
+ },
+ {
+ "epoch": 24.419213973799128,
+ "grad_norm": 0.01573132909834385,
+ "learning_rate": 0.0006,
+ "loss": 4.315916061401367,
+ "step": 1758
+ },
+ {
+ "epoch": 24.433187772925763,
+ "grad_norm": 0.015367005951702595,
+ "learning_rate": 0.0006,
+ "loss": 4.358939170837402,
+ "step": 1759
+ },
+ {
+ "epoch": 24.4471615720524,
+ "grad_norm": 0.01669907011091709,
+ "learning_rate": 0.0006,
+ "loss": 4.232147216796875,
+ "step": 1760
+ },
+ {
+ "epoch": 24.46113537117904,
+ "grad_norm": 0.018708152696490288,
+ "learning_rate": 0.0006,
+ "loss": 4.329852104187012,
+ "step": 1761
+ },
+ {
+ "epoch": 24.475109170305675,
+ "grad_norm": 0.019035987555980682,
+ "learning_rate": 0.0006,
+ "loss": 4.260340213775635,
+ "step": 1762
+ },
+ {
+ "epoch": 24.489082969432314,
+ "grad_norm": 0.01832144521176815,
+ "learning_rate": 0.0006,
+ "loss": 4.359543800354004,
+ "step": 1763
+ },
+ {
+ "epoch": 24.503056768558952,
+ "grad_norm": 0.015313294716179371,
+ "learning_rate": 0.0006,
+ "loss": 4.279088020324707,
+ "step": 1764
+ },
+ {
+ "epoch": 24.51703056768559,
+ "grad_norm": 0.015803923830389977,
+ "learning_rate": 0.0006,
+ "loss": 4.230682849884033,
+ "step": 1765
+ },
+ {
+ "epoch": 24.531004366812226,
+ "grad_norm": 0.016748374328017235,
+ "learning_rate": 0.0006,
+ "loss": 4.349661827087402,
+ "step": 1766
+ },
+ {
+ "epoch": 24.544978165938865,
+ "grad_norm": 0.015732428058981895,
+ "learning_rate": 0.0006,
+ "loss": 4.372441291809082,
+ "step": 1767
+ },
+ {
+ "epoch": 24.558951965065503,
+ "grad_norm": 0.016564829275012016,
+ "learning_rate": 0.0006,
+ "loss": 4.323090553283691,
+ "step": 1768
+ },
+ {
+ "epoch": 24.57292576419214,
+ "grad_norm": 0.01603010483086109,
+ "learning_rate": 0.0006,
+ "loss": 4.347898006439209,
+ "step": 1769
+ },
+ {
+ "epoch": 24.586899563318777,
+ "grad_norm": 0.01674809865653515,
+ "learning_rate": 0.0006,
+ "loss": 4.269165515899658,
+ "step": 1770
+ },
+ {
+ "epoch": 24.600873362445416,
+ "grad_norm": 0.017333930358290672,
+ "learning_rate": 0.0006,
+ "loss": 4.353298187255859,
+ "step": 1771
+ },
+ {
+ "epoch": 24.61484716157205,
+ "grad_norm": 0.017519580200314522,
+ "learning_rate": 0.0006,
+ "loss": 4.265199661254883,
+ "step": 1772
+ },
+ {
+ "epoch": 24.62882096069869,
+ "grad_norm": 0.017675306648015976,
+ "learning_rate": 0.0006,
+ "loss": 4.322596549987793,
+ "step": 1773
+ },
+ {
+ "epoch": 24.64279475982533,
+ "grad_norm": 0.017166534438729286,
+ "learning_rate": 0.0006,
+ "loss": 4.413900375366211,
+ "step": 1774
+ },
+ {
+ "epoch": 24.656768558951963,
+ "grad_norm": 0.016362346708774567,
+ "learning_rate": 0.0006,
+ "loss": 4.307236671447754,
+ "step": 1775
+ },
+ {
+ "epoch": 24.670742358078602,
+ "grad_norm": 0.016086872667074203,
+ "learning_rate": 0.0006,
+ "loss": 4.235316276550293,
+ "step": 1776
+ },
+ {
+ "epoch": 24.68471615720524,
+ "grad_norm": 0.01689537614583969,
+ "learning_rate": 0.0006,
+ "loss": 4.390591621398926,
+ "step": 1777
+ },
+ {
+ "epoch": 24.69868995633188,
+ "grad_norm": 0.016836725175380707,
+ "learning_rate": 0.0006,
+ "loss": 4.2990031242370605,
+ "step": 1778
+ },
+ {
+ "epoch": 24.712663755458514,
+ "grad_norm": 0.015627335757017136,
+ "learning_rate": 0.0006,
+ "loss": 4.166745185852051,
+ "step": 1779
+ },
+ {
+ "epoch": 24.726637554585153,
+ "grad_norm": 0.015007016249001026,
+ "learning_rate": 0.0006,
+ "loss": 4.295273303985596,
+ "step": 1780
+ },
+ {
+ "epoch": 24.74061135371179,
+ "grad_norm": 0.016084209084510803,
+ "learning_rate": 0.0006,
+ "loss": 4.342424392700195,
+ "step": 1781
+ },
+ {
+ "epoch": 24.754585152838427,
+ "grad_norm": 0.01595073565840721,
+ "learning_rate": 0.0006,
+ "loss": 4.3060102462768555,
+ "step": 1782
+ },
+ {
+ "epoch": 24.768558951965066,
+ "grad_norm": 0.016113461926579475,
+ "learning_rate": 0.0006,
+ "loss": 4.267908096313477,
+ "step": 1783
+ },
+ {
+ "epoch": 24.782532751091704,
+ "grad_norm": 0.01564556173980236,
+ "learning_rate": 0.0006,
+ "loss": 4.324642658233643,
+ "step": 1784
+ },
+ {
+ "epoch": 24.79650655021834,
+ "grad_norm": 0.01575268618762493,
+ "learning_rate": 0.0006,
+ "loss": 4.328181743621826,
+ "step": 1785
+ },
+ {
+ "epoch": 24.810480349344978,
+ "grad_norm": 0.016830824315547943,
+ "learning_rate": 0.0006,
+ "loss": 4.247906684875488,
+ "step": 1786
+ },
+ {
+ "epoch": 24.824454148471617,
+ "grad_norm": 0.017989547923207283,
+ "learning_rate": 0.0006,
+ "loss": 4.228281021118164,
+ "step": 1787
+ },
+ {
+ "epoch": 24.83842794759825,
+ "grad_norm": 0.016783086583018303,
+ "learning_rate": 0.0006,
+ "loss": 4.2880964279174805,
+ "step": 1788
+ },
+ {
+ "epoch": 24.85240174672489,
+ "grad_norm": 0.016467561945319176,
+ "learning_rate": 0.0006,
+ "loss": 4.36118745803833,
+ "step": 1789
+ },
+ {
+ "epoch": 24.86637554585153,
+ "grad_norm": 0.01780523732304573,
+ "learning_rate": 0.0006,
+ "loss": 4.172321319580078,
+ "step": 1790
+ },
+ {
+ "epoch": 24.880349344978168,
+ "grad_norm": 0.017102031037211418,
+ "learning_rate": 0.0006,
+ "loss": 4.352826118469238,
+ "step": 1791
+ },
+ {
+ "epoch": 24.894323144104803,
+ "grad_norm": 0.016409218311309814,
+ "learning_rate": 0.0006,
+ "loss": 4.212162017822266,
+ "step": 1792
+ },
+ {
+ "epoch": 24.90829694323144,
+ "grad_norm": 0.018660584464669228,
+ "learning_rate": 0.0006,
+ "loss": 4.346684455871582,
+ "step": 1793
+ },
+ {
+ "epoch": 24.92227074235808,
+ "grad_norm": 0.017389440909028053,
+ "learning_rate": 0.0006,
+ "loss": 4.197913646697998,
+ "step": 1794
+ },
+ {
+ "epoch": 24.936244541484715,
+ "grad_norm": 0.01596485823392868,
+ "learning_rate": 0.0006,
+ "loss": 4.370617866516113,
+ "step": 1795
+ },
+ {
+ "epoch": 24.950218340611354,
+ "grad_norm": 0.01709776371717453,
+ "learning_rate": 0.0006,
+ "loss": 4.313847064971924,
+ "step": 1796
+ },
+ {
+ "epoch": 24.964192139737992,
+ "grad_norm": 0.016571134328842163,
+ "learning_rate": 0.0006,
+ "loss": 4.1895856857299805,
+ "step": 1797
+ },
+ {
+ "epoch": 24.978165938864628,
+ "grad_norm": 0.016851048916578293,
+ "learning_rate": 0.0006,
+ "loss": 4.350772857666016,
+ "step": 1798
+ },
+ {
+ "epoch": 24.992139737991266,
+ "grad_norm": 0.017007440328598022,
+ "learning_rate": 0.0006,
+ "loss": 4.399008750915527,
+ "step": 1799
+ },
+ {
+ "epoch": 25.0,
+ "grad_norm": 0.01791212148964405,
+ "learning_rate": 0.0006,
+ "loss": 4.306121826171875,
+ "step": 1800
+ },
+ {
+ "epoch": 25.0,
+ "eval_loss": 4.636417865753174,
+ "eval_runtime": 56.8296,
+ "eval_samples_per_second": 42.971,
+ "eval_steps_per_second": 1.355,
+ "step": 1800
+ },
+ {
+ "epoch": 25.01397379912664,
+ "grad_norm": 0.01729617640376091,
+ "learning_rate": 0.0006,
+ "loss": 4.257018089294434,
+ "step": 1801
+ },
+ {
+ "epoch": 25.027947598253274,
+ "grad_norm": 0.017466643825173378,
+ "learning_rate": 0.0006,
+ "loss": 4.307605266571045,
+ "step": 1802
+ },
+ {
+ "epoch": 25.041921397379912,
+ "grad_norm": 0.018614279106259346,
+ "learning_rate": 0.0006,
+ "loss": 4.360141277313232,
+ "step": 1803
+ },
+ {
+ "epoch": 25.05589519650655,
+ "grad_norm": 0.02010771445930004,
+ "learning_rate": 0.0006,
+ "loss": 4.196428298950195,
+ "step": 1804
+ },
+ {
+ "epoch": 25.069868995633186,
+ "grad_norm": 0.02090228535234928,
+ "learning_rate": 0.0006,
+ "loss": 4.288947105407715,
+ "step": 1805
+ },
+ {
+ "epoch": 25.083842794759825,
+ "grad_norm": 0.02050255797803402,
+ "learning_rate": 0.0006,
+ "loss": 4.259582042694092,
+ "step": 1806
+ },
+ {
+ "epoch": 25.097816593886463,
+ "grad_norm": 0.021844089031219482,
+ "learning_rate": 0.0006,
+ "loss": 4.405138969421387,
+ "step": 1807
+ },
+ {
+ "epoch": 25.111790393013102,
+ "grad_norm": 0.022287018597126007,
+ "learning_rate": 0.0006,
+ "loss": 4.333793640136719,
+ "step": 1808
+ },
+ {
+ "epoch": 25.125764192139737,
+ "grad_norm": 0.02288329415023327,
+ "learning_rate": 0.0006,
+ "loss": 4.3910980224609375,
+ "step": 1809
+ },
+ {
+ "epoch": 25.139737991266376,
+ "grad_norm": 0.02373199723660946,
+ "learning_rate": 0.0006,
+ "loss": 4.32921028137207,
+ "step": 1810
+ },
+ {
+ "epoch": 25.153711790393015,
+ "grad_norm": 0.024960234761238098,
+ "learning_rate": 0.0006,
+ "loss": 4.295609474182129,
+ "step": 1811
+ },
+ {
+ "epoch": 25.16768558951965,
+ "grad_norm": 0.026743892580270767,
+ "learning_rate": 0.0006,
+ "loss": 4.315308094024658,
+ "step": 1812
+ },
+ {
+ "epoch": 25.18165938864629,
+ "grad_norm": 0.022138185799121857,
+ "learning_rate": 0.0006,
+ "loss": 4.360208511352539,
+ "step": 1813
+ },
+ {
+ "epoch": 25.195633187772927,
+ "grad_norm": 0.022120574489235878,
+ "learning_rate": 0.0006,
+ "loss": 4.1973772048950195,
+ "step": 1814
+ },
+ {
+ "epoch": 25.209606986899562,
+ "grad_norm": 0.02230502851307392,
+ "learning_rate": 0.0006,
+ "loss": 4.285558700561523,
+ "step": 1815
+ },
+ {
+ "epoch": 25.2235807860262,
+ "grad_norm": 0.022301986813545227,
+ "learning_rate": 0.0006,
+ "loss": 4.238343238830566,
+ "step": 1816
+ },
+ {
+ "epoch": 25.23755458515284,
+ "grad_norm": 0.020050447434186935,
+ "learning_rate": 0.0006,
+ "loss": 4.298235893249512,
+ "step": 1817
+ },
+ {
+ "epoch": 25.251528384279474,
+ "grad_norm": 0.01987724006175995,
+ "learning_rate": 0.0006,
+ "loss": 4.301384449005127,
+ "step": 1818
+ },
+ {
+ "epoch": 25.265502183406113,
+ "grad_norm": 0.020440855994820595,
+ "learning_rate": 0.0006,
+ "loss": 4.269218444824219,
+ "step": 1819
+ },
+ {
+ "epoch": 25.27947598253275,
+ "grad_norm": 0.01891893707215786,
+ "learning_rate": 0.0006,
+ "loss": 4.2401957511901855,
+ "step": 1820
+ },
+ {
+ "epoch": 25.29344978165939,
+ "grad_norm": 0.01903688907623291,
+ "learning_rate": 0.0006,
+ "loss": 4.139670372009277,
+ "step": 1821
+ },
+ {
+ "epoch": 25.307423580786025,
+ "grad_norm": 0.018856938928365707,
+ "learning_rate": 0.0006,
+ "loss": 4.233214378356934,
+ "step": 1822
+ },
+ {
+ "epoch": 25.321397379912664,
+ "grad_norm": 0.017791323363780975,
+ "learning_rate": 0.0006,
+ "loss": 4.2342023849487305,
+ "step": 1823
+ },
+ {
+ "epoch": 25.335371179039303,
+ "grad_norm": 0.017304565757513046,
+ "learning_rate": 0.0006,
+ "loss": 4.355198860168457,
+ "step": 1824
+ },
+ {
+ "epoch": 25.349344978165938,
+ "grad_norm": 0.01576206088066101,
+ "learning_rate": 0.0006,
+ "loss": 4.114147186279297,
+ "step": 1825
+ },
+ {
+ "epoch": 25.363318777292577,
+ "grad_norm": 0.015105154365301132,
+ "learning_rate": 0.0006,
+ "loss": 4.313795566558838,
+ "step": 1826
+ },
+ {
+ "epoch": 25.377292576419215,
+ "grad_norm": 0.01563587784767151,
+ "learning_rate": 0.0006,
+ "loss": 4.308282852172852,
+ "step": 1827
+ },
+ {
+ "epoch": 25.39126637554585,
+ "grad_norm": 0.015980906784534454,
+ "learning_rate": 0.0006,
+ "loss": 4.345149040222168,
+ "step": 1828
+ },
+ {
+ "epoch": 25.40524017467249,
+ "grad_norm": 0.01669284701347351,
+ "learning_rate": 0.0006,
+ "loss": 4.237611293792725,
+ "step": 1829
+ },
+ {
+ "epoch": 25.419213973799128,
+ "grad_norm": 0.016202064231038094,
+ "learning_rate": 0.0006,
+ "loss": 4.3072309494018555,
+ "step": 1830
+ },
+ {
+ "epoch": 25.433187772925763,
+ "grad_norm": 0.01478662807494402,
+ "learning_rate": 0.0006,
+ "loss": 4.276215076446533,
+ "step": 1831
+ },
+ {
+ "epoch": 25.4471615720524,
+ "grad_norm": 0.015168731100857258,
+ "learning_rate": 0.0006,
+ "loss": 4.37509822845459,
+ "step": 1832
+ },
+ {
+ "epoch": 25.46113537117904,
+ "grad_norm": 0.01648980937898159,
+ "learning_rate": 0.0006,
+ "loss": 4.349167346954346,
+ "step": 1833
+ },
+ {
+ "epoch": 25.475109170305675,
+ "grad_norm": 0.01708398200571537,
+ "learning_rate": 0.0006,
+ "loss": 4.346580505371094,
+ "step": 1834
+ },
+ {
+ "epoch": 25.489082969432314,
+ "grad_norm": 0.017370784655213356,
+ "learning_rate": 0.0006,
+ "loss": 4.299466133117676,
+ "step": 1835
+ },
+ {
+ "epoch": 25.503056768558952,
+ "grad_norm": 0.016248228028416634,
+ "learning_rate": 0.0006,
+ "loss": 4.184610366821289,
+ "step": 1836
+ },
+ {
+ "epoch": 25.51703056768559,
+ "grad_norm": 0.017106913030147552,
+ "learning_rate": 0.0006,
+ "loss": 4.366856098175049,
+ "step": 1837
+ },
+ {
+ "epoch": 25.531004366812226,
+ "grad_norm": 0.016071965917944908,
+ "learning_rate": 0.0006,
+ "loss": 4.297691345214844,
+ "step": 1838
+ },
+ {
+ "epoch": 25.544978165938865,
+ "grad_norm": 0.016451986506581306,
+ "learning_rate": 0.0006,
+ "loss": 4.229983329772949,
+ "step": 1839
+ },
+ {
+ "epoch": 25.558951965065503,
+ "grad_norm": 0.016195081174373627,
+ "learning_rate": 0.0006,
+ "loss": 4.259893417358398,
+ "step": 1840
+ },
+ {
+ "epoch": 25.57292576419214,
+ "grad_norm": 0.0165175162255764,
+ "learning_rate": 0.0006,
+ "loss": 4.3348236083984375,
+ "step": 1841
+ },
+ {
+ "epoch": 25.586899563318777,
+ "grad_norm": 0.01651517115533352,
+ "learning_rate": 0.0006,
+ "loss": 4.319541931152344,
+ "step": 1842
+ },
+ {
+ "epoch": 25.600873362445416,
+ "grad_norm": 0.016143690794706345,
+ "learning_rate": 0.0006,
+ "loss": 4.345652103424072,
+ "step": 1843
+ },
+ {
+ "epoch": 25.61484716157205,
+ "grad_norm": 0.015140696428716183,
+ "learning_rate": 0.0006,
+ "loss": 4.212047576904297,
+ "step": 1844
+ },
+ {
+ "epoch": 25.62882096069869,
+ "grad_norm": 0.014637443237006664,
+ "learning_rate": 0.0006,
+ "loss": 4.187453269958496,
+ "step": 1845
+ },
+ {
+ "epoch": 25.64279475982533,
+ "grad_norm": 0.014585614204406738,
+ "learning_rate": 0.0006,
+ "loss": 4.253936767578125,
+ "step": 1846
+ },
+ {
+ "epoch": 25.656768558951963,
+ "grad_norm": 0.015158289112150669,
+ "learning_rate": 0.0006,
+ "loss": 4.34612512588501,
+ "step": 1847
+ },
+ {
+ "epoch": 25.670742358078602,
+ "grad_norm": 0.015202849172055721,
+ "learning_rate": 0.0006,
+ "loss": 4.270634651184082,
+ "step": 1848
+ },
+ {
+ "epoch": 25.68471615720524,
+ "grad_norm": 0.015363575890660286,
+ "learning_rate": 0.0006,
+ "loss": 4.271080493927002,
+ "step": 1849
+ },
+ {
+ "epoch": 25.69868995633188,
+ "grad_norm": 0.014360226690769196,
+ "learning_rate": 0.0006,
+ "loss": 4.375033855438232,
+ "step": 1850
+ },
+ {
+ "epoch": 25.712663755458514,
+ "grad_norm": 0.014263181947171688,
+ "learning_rate": 0.0006,
+ "loss": 4.424138069152832,
+ "step": 1851
+ },
+ {
+ "epoch": 25.726637554585153,
+ "grad_norm": 0.014398688450455666,
+ "learning_rate": 0.0006,
+ "loss": 4.410154819488525,
+ "step": 1852
+ },
+ {
+ "epoch": 25.74061135371179,
+ "grad_norm": 0.016131488606333733,
+ "learning_rate": 0.0006,
+ "loss": 4.324014663696289,
+ "step": 1853
+ },
+ {
+ "epoch": 25.754585152838427,
+ "grad_norm": 0.017568735405802727,
+ "learning_rate": 0.0006,
+ "loss": 4.367861747741699,
+ "step": 1854
+ },
+ {
+ "epoch": 25.768558951965066,
+ "grad_norm": 0.017907511442899704,
+ "learning_rate": 0.0006,
+ "loss": 4.352965354919434,
+ "step": 1855
+ },
+ {
+ "epoch": 25.782532751091704,
+ "grad_norm": 0.016918016597628593,
+ "learning_rate": 0.0006,
+ "loss": 4.252397537231445,
+ "step": 1856
+ },
+ {
+ "epoch": 25.79650655021834,
+ "grad_norm": 0.017274610698223114,
+ "learning_rate": 0.0006,
+ "loss": 4.342309951782227,
+ "step": 1857
+ },
+ {
+ "epoch": 25.810480349344978,
+ "grad_norm": 0.017879825085401535,
+ "learning_rate": 0.0006,
+ "loss": 4.330987453460693,
+ "step": 1858
+ },
+ {
+ "epoch": 25.824454148471617,
+ "grad_norm": 0.01736099272966385,
+ "learning_rate": 0.0006,
+ "loss": 4.307255744934082,
+ "step": 1859
+ },
+ {
+ "epoch": 25.83842794759825,
+ "grad_norm": 0.016575824469327927,
+ "learning_rate": 0.0006,
+ "loss": 4.262016773223877,
+ "step": 1860
+ },
+ {
+ "epoch": 25.85240174672489,
+ "grad_norm": 0.015181140042841434,
+ "learning_rate": 0.0006,
+ "loss": 4.291863441467285,
+ "step": 1861
+ },
+ {
+ "epoch": 25.86637554585153,
+ "grad_norm": 0.015965687111020088,
+ "learning_rate": 0.0006,
+ "loss": 4.424836158752441,
+ "step": 1862
+ },
+ {
+ "epoch": 25.880349344978168,
+ "grad_norm": 0.01687219925224781,
+ "learning_rate": 0.0006,
+ "loss": 4.343531131744385,
+ "step": 1863
+ },
+ {
+ "epoch": 25.894323144104803,
+ "grad_norm": 0.015115504153072834,
+ "learning_rate": 0.0006,
+ "loss": 4.357808589935303,
+ "step": 1864
+ },
+ {
+ "epoch": 25.90829694323144,
+ "grad_norm": 0.014150998555123806,
+ "learning_rate": 0.0006,
+ "loss": 4.324550151824951,
+ "step": 1865
+ },
+ {
+ "epoch": 25.92227074235808,
+ "grad_norm": 0.014989510178565979,
+ "learning_rate": 0.0006,
+ "loss": 4.380428314208984,
+ "step": 1866
+ },
+ {
+ "epoch": 25.936244541484715,
+ "grad_norm": 0.01527960691601038,
+ "learning_rate": 0.0006,
+ "loss": 4.399374961853027,
+ "step": 1867
+ },
+ {
+ "epoch": 25.950218340611354,
+ "grad_norm": 0.015434633940458298,
+ "learning_rate": 0.0006,
+ "loss": 4.260984420776367,
+ "step": 1868
+ },
+ {
+ "epoch": 25.964192139737992,
+ "grad_norm": 0.015370228327810764,
+ "learning_rate": 0.0006,
+ "loss": 4.3340253829956055,
+ "step": 1869
+ },
+ {
+ "epoch": 25.978165938864628,
+ "grad_norm": 0.015348542481660843,
+ "learning_rate": 0.0006,
+ "loss": 4.369132995605469,
+ "step": 1870
+ },
+ {
+ "epoch": 25.992139737991266,
+ "grad_norm": 0.015354936942458153,
+ "learning_rate": 0.0006,
+ "loss": 4.381937026977539,
+ "step": 1871
+ },
+ {
+ "epoch": 26.0,
+ "grad_norm": 0.01796272210776806,
+ "learning_rate": 0.0006,
+ "loss": 4.286768436431885,
+ "step": 1872
+ },
+ {
+ "epoch": 26.0,
+ "eval_loss": 4.576984405517578,
+ "eval_runtime": 57.2091,
+ "eval_samples_per_second": 42.685,
+ "eval_steps_per_second": 1.346,
+ "step": 1872
+ },
+ {
+ "epoch": 26.01397379912664,
+ "grad_norm": 0.01748845726251602,
+ "learning_rate": 0.0006,
+ "loss": 4.337504863739014,
+ "step": 1873
+ },
+ {
+ "epoch": 26.027947598253274,
+ "grad_norm": 0.019244296476244926,
+ "learning_rate": 0.0006,
+ "loss": 4.264280796051025,
+ "step": 1874
+ },
+ {
+ "epoch": 26.041921397379912,
+ "grad_norm": 0.0215692650526762,
+ "learning_rate": 0.0006,
+ "loss": 4.350830078125,
+ "step": 1875
+ },
+ {
+ "epoch": 26.05589519650655,
+ "grad_norm": 0.022189252078533173,
+ "learning_rate": 0.0006,
+ "loss": 4.299098968505859,
+ "step": 1876
+ },
+ {
+ "epoch": 26.069868995633186,
+ "grad_norm": 0.022384043782949448,
+ "learning_rate": 0.0006,
+ "loss": 4.1185173988342285,
+ "step": 1877
+ },
+ {
+ "epoch": 26.083842794759825,
+ "grad_norm": 0.02076014317572117,
+ "learning_rate": 0.0006,
+ "loss": 4.297612190246582,
+ "step": 1878
+ },
+ {
+ "epoch": 26.097816593886463,
+ "grad_norm": 0.020843051373958588,
+ "learning_rate": 0.0006,
+ "loss": 4.294098854064941,
+ "step": 1879
+ },
+ {
+ "epoch": 26.111790393013102,
+ "grad_norm": 0.020654456689953804,
+ "learning_rate": 0.0006,
+ "loss": 4.267856597900391,
+ "step": 1880
+ },
+ {
+ "epoch": 26.125764192139737,
+ "grad_norm": 0.01860472559928894,
+ "learning_rate": 0.0006,
+ "loss": 4.284232139587402,
+ "step": 1881
+ },
+ {
+ "epoch": 26.139737991266376,
+ "grad_norm": 0.018719421699643135,
+ "learning_rate": 0.0006,
+ "loss": 4.277214527130127,
+ "step": 1882
+ },
+ {
+ "epoch": 26.153711790393015,
+ "grad_norm": 0.020903829485177994,
+ "learning_rate": 0.0006,
+ "loss": 4.254947662353516,
+ "step": 1883
+ },
+ {
+ "epoch": 26.16768558951965,
+ "grad_norm": 0.02020149491727352,
+ "learning_rate": 0.0006,
+ "loss": 4.2526936531066895,
+ "step": 1884
+ },
+ {
+ "epoch": 26.18165938864629,
+ "grad_norm": 0.0180392786860466,
+ "learning_rate": 0.0006,
+ "loss": 4.309922695159912,
+ "step": 1885
+ },
+ {
+ "epoch": 26.195633187772927,
+ "grad_norm": 0.018163125962018967,
+ "learning_rate": 0.0006,
+ "loss": 4.13860559463501,
+ "step": 1886
+ },
+ {
+ "epoch": 26.209606986899562,
+ "grad_norm": 0.01853291131556034,
+ "learning_rate": 0.0006,
+ "loss": 4.190926551818848,
+ "step": 1887
+ },
+ {
+ "epoch": 26.2235807860262,
+ "grad_norm": 0.01757677271962166,
+ "learning_rate": 0.0006,
+ "loss": 4.163999557495117,
+ "step": 1888
+ },
+ {
+ "epoch": 26.23755458515284,
+ "grad_norm": 0.01767992228269577,
+ "learning_rate": 0.0006,
+ "loss": 4.184430122375488,
+ "step": 1889
+ },
+ {
+ "epoch": 26.251528384279474,
+ "grad_norm": 0.018139973282814026,
+ "learning_rate": 0.0006,
+ "loss": 4.300766944885254,
+ "step": 1890
+ },
+ {
+ "epoch": 26.265502183406113,
+ "grad_norm": 0.020184461027383804,
+ "learning_rate": 0.0006,
+ "loss": 4.232537269592285,
+ "step": 1891
+ },
+ {
+ "epoch": 26.27947598253275,
+ "grad_norm": 0.02028856985270977,
+ "learning_rate": 0.0006,
+ "loss": 4.324883460998535,
+ "step": 1892
+ },
+ {
+ "epoch": 26.29344978165939,
+ "grad_norm": 0.017120616510510445,
+ "learning_rate": 0.0006,
+ "loss": 4.23493766784668,
+ "step": 1893
+ },
+ {
+ "epoch": 26.307423580786025,
+ "grad_norm": 0.016202978789806366,
+ "learning_rate": 0.0006,
+ "loss": 4.197602272033691,
+ "step": 1894
+ },
+ {
+ "epoch": 26.321397379912664,
+ "grad_norm": 0.017231421545147896,
+ "learning_rate": 0.0006,
+ "loss": 4.302001953125,
+ "step": 1895
+ },
+ {
+ "epoch": 26.335371179039303,
+ "grad_norm": 0.01725655607879162,
+ "learning_rate": 0.0006,
+ "loss": 4.356800079345703,
+ "step": 1896
+ },
+ {
+ "epoch": 26.349344978165938,
+ "grad_norm": 0.016657443717122078,
+ "learning_rate": 0.0006,
+ "loss": 4.266307830810547,
+ "step": 1897
+ },
+ {
+ "epoch": 26.363318777292577,
+ "grad_norm": 0.01616556942462921,
+ "learning_rate": 0.0006,
+ "loss": 4.307868003845215,
+ "step": 1898
+ },
+ {
+ "epoch": 26.377292576419215,
+ "grad_norm": 0.016621306538581848,
+ "learning_rate": 0.0006,
+ "loss": 4.172905921936035,
+ "step": 1899
+ },
+ {
+ "epoch": 26.39126637554585,
+ "grad_norm": 0.016243206337094307,
+ "learning_rate": 0.0006,
+ "loss": 4.162896156311035,
+ "step": 1900
+ },
+ {
+ "epoch": 26.40524017467249,
+ "grad_norm": 0.016663808375597,
+ "learning_rate": 0.0006,
+ "loss": 4.258232116699219,
+ "step": 1901
+ },
+ {
+ "epoch": 26.419213973799128,
+ "grad_norm": 0.01677478663623333,
+ "learning_rate": 0.0006,
+ "loss": 4.336090087890625,
+ "step": 1902
+ },
+ {
+ "epoch": 26.433187772925763,
+ "grad_norm": 0.01641303487122059,
+ "learning_rate": 0.0006,
+ "loss": 4.191476821899414,
+ "step": 1903
+ },
+ {
+ "epoch": 26.4471615720524,
+ "grad_norm": 0.016700323671102524,
+ "learning_rate": 0.0006,
+ "loss": 4.299427509307861,
+ "step": 1904
+ },
+ {
+ "epoch": 26.46113537117904,
+ "grad_norm": 0.017595035955309868,
+ "learning_rate": 0.0006,
+ "loss": 4.312828063964844,
+ "step": 1905
+ },
+ {
+ "epoch": 26.475109170305675,
+ "grad_norm": 0.020300284028053284,
+ "learning_rate": 0.0006,
+ "loss": 4.28331184387207,
+ "step": 1906
+ },
+ {
+ "epoch": 26.489082969432314,
+ "grad_norm": 0.02198321744799614,
+ "learning_rate": 0.0006,
+ "loss": 4.3280839920043945,
+ "step": 1907
+ },
+ {
+ "epoch": 26.503056768558952,
+ "grad_norm": 0.02003200724720955,
+ "learning_rate": 0.0006,
+ "loss": 4.325623989105225,
+ "step": 1908
+ },
+ {
+ "epoch": 26.51703056768559,
+ "grad_norm": 0.01870287023484707,
+ "learning_rate": 0.0006,
+ "loss": 4.251203536987305,
+ "step": 1909
+ },
+ {
+ "epoch": 26.531004366812226,
+ "grad_norm": 0.01835448481142521,
+ "learning_rate": 0.0006,
+ "loss": 4.335144996643066,
+ "step": 1910
+ },
+ {
+ "epoch": 26.544978165938865,
+ "grad_norm": 0.02046327292919159,
+ "learning_rate": 0.0006,
+ "loss": 4.238023281097412,
+ "step": 1911
+ },
+ {
+ "epoch": 26.558951965065503,
+ "grad_norm": 0.019771311432123184,
+ "learning_rate": 0.0006,
+ "loss": 4.353594779968262,
+ "step": 1912
+ },
+ {
+ "epoch": 26.57292576419214,
+ "grad_norm": 0.018090544268488884,
+ "learning_rate": 0.0006,
+ "loss": 4.304343223571777,
+ "step": 1913
+ },
+ {
+ "epoch": 26.586899563318777,
+ "grad_norm": 0.017489777877926826,
+ "learning_rate": 0.0006,
+ "loss": 4.204375267028809,
+ "step": 1914
+ },
+ {
+ "epoch": 26.600873362445416,
+ "grad_norm": 0.016785891726613045,
+ "learning_rate": 0.0006,
+ "loss": 4.286669731140137,
+ "step": 1915
+ },
+ {
+ "epoch": 26.61484716157205,
+ "grad_norm": 0.016382789239287376,
+ "learning_rate": 0.0006,
+ "loss": 4.347662448883057,
+ "step": 1916
+ },
+ {
+ "epoch": 26.62882096069869,
+ "grad_norm": 0.016581876203417778,
+ "learning_rate": 0.0006,
+ "loss": 4.2716193199157715,
+ "step": 1917
+ },
+ {
+ "epoch": 26.64279475982533,
+ "grad_norm": 0.015627184882760048,
+ "learning_rate": 0.0006,
+ "loss": 4.311489105224609,
+ "step": 1918
+ },
+ {
+ "epoch": 26.656768558951963,
+ "grad_norm": 0.015908483415842056,
+ "learning_rate": 0.0006,
+ "loss": 4.262099742889404,
+ "step": 1919
+ },
+ {
+ "epoch": 26.670742358078602,
+ "grad_norm": 0.015576236881315708,
+ "learning_rate": 0.0006,
+ "loss": 4.275871276855469,
+ "step": 1920
+ },
+ {
+ "epoch": 26.68471615720524,
+ "grad_norm": 0.014621064998209476,
+ "learning_rate": 0.0006,
+ "loss": 4.184260845184326,
+ "step": 1921
+ },
+ {
+ "epoch": 26.69868995633188,
+ "grad_norm": 0.014797811396420002,
+ "learning_rate": 0.0006,
+ "loss": 4.3412628173828125,
+ "step": 1922
+ },
+ {
+ "epoch": 26.712663755458514,
+ "grad_norm": 0.015610570088028908,
+ "learning_rate": 0.0006,
+ "loss": 4.181094646453857,
+ "step": 1923
+ },
+ {
+ "epoch": 26.726637554585153,
+ "grad_norm": 0.016048265621066093,
+ "learning_rate": 0.0006,
+ "loss": 4.365290641784668,
+ "step": 1924
+ },
+ {
+ "epoch": 26.74061135371179,
+ "grad_norm": 0.016604071483016014,
+ "learning_rate": 0.0006,
+ "loss": 4.2628912925720215,
+ "step": 1925
+ },
+ {
+ "epoch": 26.754585152838427,
+ "grad_norm": 0.017044221982359886,
+ "learning_rate": 0.0006,
+ "loss": 4.378512382507324,
+ "step": 1926
+ },
+ {
+ "epoch": 26.768558951965066,
+ "grad_norm": 0.01746384985744953,
+ "learning_rate": 0.0006,
+ "loss": 4.3888773918151855,
+ "step": 1927
+ },
+ {
+ "epoch": 26.782532751091704,
+ "grad_norm": 0.01572711206972599,
+ "learning_rate": 0.0006,
+ "loss": 4.189935684204102,
+ "step": 1928
+ },
+ {
+ "epoch": 26.79650655021834,
+ "grad_norm": 0.014976629987359047,
+ "learning_rate": 0.0006,
+ "loss": 4.280525207519531,
+ "step": 1929
+ },
+ {
+ "epoch": 26.810480349344978,
+ "grad_norm": 0.014804664999246597,
+ "learning_rate": 0.0006,
+ "loss": 4.270630836486816,
+ "step": 1930
+ },
+ {
+ "epoch": 26.824454148471617,
+ "grad_norm": 0.014641920104622841,
+ "learning_rate": 0.0006,
+ "loss": 4.319505214691162,
+ "step": 1931
+ },
+ {
+ "epoch": 26.83842794759825,
+ "grad_norm": 0.014928505755960941,
+ "learning_rate": 0.0006,
+ "loss": 4.180147647857666,
+ "step": 1932
+ },
+ {
+ "epoch": 26.85240174672489,
+ "grad_norm": 0.016130365431308746,
+ "learning_rate": 0.0006,
+ "loss": 4.260660648345947,
+ "step": 1933
+ },
+ {
+ "epoch": 26.86637554585153,
+ "grad_norm": 0.015642493963241577,
+ "learning_rate": 0.0006,
+ "loss": 4.304064750671387,
+ "step": 1934
+ },
+ {
+ "epoch": 26.880349344978168,
+ "grad_norm": 0.015988217666745186,
+ "learning_rate": 0.0006,
+ "loss": 4.2548017501831055,
+ "step": 1935
+ },
+ {
+ "epoch": 26.894323144104803,
+ "grad_norm": 0.016501398757100105,
+ "learning_rate": 0.0006,
+ "loss": 4.260058879852295,
+ "step": 1936
+ },
+ {
+ "epoch": 26.90829694323144,
+ "grad_norm": 0.01645040698349476,
+ "learning_rate": 0.0006,
+ "loss": 4.483081340789795,
+ "step": 1937
+ },
+ {
+ "epoch": 26.92227074235808,
+ "grad_norm": 0.016140863299369812,
+ "learning_rate": 0.0006,
+ "loss": 4.1776838302612305,
+ "step": 1938
+ },
+ {
+ "epoch": 26.936244541484715,
+ "grad_norm": 0.016681019216775894,
+ "learning_rate": 0.0006,
+ "loss": 4.224747657775879,
+ "step": 1939
+ },
+ {
+ "epoch": 26.950218340611354,
+ "grad_norm": 0.01650378853082657,
+ "learning_rate": 0.0006,
+ "loss": 4.305703163146973,
+ "step": 1940
+ },
+ {
+ "epoch": 26.964192139737992,
+ "grad_norm": 0.017051683738827705,
+ "learning_rate": 0.0006,
+ "loss": 4.285835266113281,
+ "step": 1941
+ },
+ {
+ "epoch": 26.978165938864628,
+ "grad_norm": 0.016894327476620674,
+ "learning_rate": 0.0006,
+ "loss": 4.318901538848877,
+ "step": 1942
+ },
+ {
+ "epoch": 26.992139737991266,
+ "grad_norm": 0.017313100397586823,
+ "learning_rate": 0.0006,
+ "loss": 4.281434535980225,
+ "step": 1943
+ },
+ {
+ "epoch": 27.0,
+ "grad_norm": 0.018450884148478508,
+ "learning_rate": 0.0006,
+ "loss": 4.335104942321777,
+ "step": 1944
+ },
+ {
+ "epoch": 27.0,
+ "eval_loss": 4.640687465667725,
+ "eval_runtime": 56.9624,
+ "eval_samples_per_second": 42.87,
+ "eval_steps_per_second": 1.352,
+ "step": 1944
+ },
+ {
+ "epoch": 27.01397379912664,
+ "grad_norm": 0.017755698412656784,
+ "learning_rate": 0.0006,
+ "loss": 4.278824329376221,
+ "step": 1945
+ },
+ {
+ "epoch": 27.027947598253274,
+ "grad_norm": 0.01724310778081417,
+ "learning_rate": 0.0006,
+ "loss": 4.311049461364746,
+ "step": 1946
+ },
+ {
+ "epoch": 27.041921397379912,
+ "grad_norm": 0.01690123789012432,
+ "learning_rate": 0.0006,
+ "loss": 4.261994361877441,
+ "step": 1947
+ },
+ {
+ "epoch": 27.05589519650655,
+ "grad_norm": 0.018406696617603302,
+ "learning_rate": 0.0006,
+ "loss": 4.1925153732299805,
+ "step": 1948
+ },
+ {
+ "epoch": 27.069868995633186,
+ "grad_norm": 0.019945867359638214,
+ "learning_rate": 0.0006,
+ "loss": 4.108537197113037,
+ "step": 1949
+ },
+ {
+ "epoch": 27.083842794759825,
+ "grad_norm": 0.020316628739237785,
+ "learning_rate": 0.0006,
+ "loss": 4.209043025970459,
+ "step": 1950
+ },
+ {
+ "epoch": 27.097816593886463,
+ "grad_norm": 0.019690370187163353,
+ "learning_rate": 0.0006,
+ "loss": 4.195789337158203,
+ "step": 1951
+ },
+ {
+ "epoch": 27.111790393013102,
+ "grad_norm": 0.01785232499241829,
+ "learning_rate": 0.0006,
+ "loss": 4.21466064453125,
+ "step": 1952
+ },
+ {
+ "epoch": 27.125764192139737,
+ "grad_norm": 0.0164767038077116,
+ "learning_rate": 0.0006,
+ "loss": 4.270247459411621,
+ "step": 1953
+ },
+ {
+ "epoch": 27.139737991266376,
+ "grad_norm": 0.018008455634117126,
+ "learning_rate": 0.0006,
+ "loss": 4.201042175292969,
+ "step": 1954
+ },
+ {
+ "epoch": 27.153711790393015,
+ "grad_norm": 0.01807340793311596,
+ "learning_rate": 0.0006,
+ "loss": 4.25289249420166,
+ "step": 1955
+ },
+ {
+ "epoch": 27.16768558951965,
+ "grad_norm": 0.016246909275650978,
+ "learning_rate": 0.0006,
+ "loss": 4.116283416748047,
+ "step": 1956
+ },
+ {
+ "epoch": 27.18165938864629,
+ "grad_norm": 0.017545776441693306,
+ "learning_rate": 0.0006,
+ "loss": 4.316999435424805,
+ "step": 1957
+ },
+ {
+ "epoch": 27.195633187772927,
+ "grad_norm": 0.017915375530719757,
+ "learning_rate": 0.0006,
+ "loss": 4.253859519958496,
+ "step": 1958
+ },
+ {
+ "epoch": 27.209606986899562,
+ "grad_norm": 0.018112223595380783,
+ "learning_rate": 0.0006,
+ "loss": 4.2541022300720215,
+ "step": 1959
+ },
+ {
+ "epoch": 27.2235807860262,
+ "grad_norm": 0.01846139505505562,
+ "learning_rate": 0.0006,
+ "loss": 4.20438289642334,
+ "step": 1960
+ },
+ {
+ "epoch": 27.23755458515284,
+ "grad_norm": 0.01943897269666195,
+ "learning_rate": 0.0006,
+ "loss": 4.2799577713012695,
+ "step": 1961
+ },
+ {
+ "epoch": 27.251528384279474,
+ "grad_norm": 0.018253348767757416,
+ "learning_rate": 0.0006,
+ "loss": 4.1720499992370605,
+ "step": 1962
+ },
+ {
+ "epoch": 27.265502183406113,
+ "grad_norm": 0.018292246386408806,
+ "learning_rate": 0.0006,
+ "loss": 4.28231143951416,
+ "step": 1963
+ },
+ {
+ "epoch": 27.27947598253275,
+ "grad_norm": 0.01943155936896801,
+ "learning_rate": 0.0006,
+ "loss": 4.314976692199707,
+ "step": 1964
+ },
+ {
+ "epoch": 27.29344978165939,
+ "grad_norm": 0.019064994528889656,
+ "learning_rate": 0.0006,
+ "loss": 4.25328254699707,
+ "step": 1965
+ },
+ {
+ "epoch": 27.307423580786025,
+ "grad_norm": 0.017423273995518684,
+ "learning_rate": 0.0006,
+ "loss": 4.195992469787598,
+ "step": 1966
+ },
+ {
+ "epoch": 27.321397379912664,
+ "grad_norm": 0.01677924580872059,
+ "learning_rate": 0.0006,
+ "loss": 4.260043144226074,
+ "step": 1967
+ },
+ {
+ "epoch": 27.335371179039303,
+ "grad_norm": 0.01720673404633999,
+ "learning_rate": 0.0006,
+ "loss": 4.222780704498291,
+ "step": 1968
+ },
+ {
+ "epoch": 27.349344978165938,
+ "grad_norm": 0.017300743609666824,
+ "learning_rate": 0.0006,
+ "loss": 4.109524250030518,
+ "step": 1969
+ },
+ {
+ "epoch": 27.363318777292577,
+ "grad_norm": 0.016065871343016624,
+ "learning_rate": 0.0006,
+ "loss": 4.36108922958374,
+ "step": 1970
+ },
+ {
+ "epoch": 27.377292576419215,
+ "grad_norm": 0.016626616939902306,
+ "learning_rate": 0.0006,
+ "loss": 4.220630645751953,
+ "step": 1971
+ },
+ {
+ "epoch": 27.39126637554585,
+ "grad_norm": 0.016482602804899216,
+ "learning_rate": 0.0006,
+ "loss": 4.344797134399414,
+ "step": 1972
+ },
+ {
+ "epoch": 27.40524017467249,
+ "grad_norm": 0.01727953553199768,
+ "learning_rate": 0.0006,
+ "loss": 4.296645164489746,
+ "step": 1973
+ },
+ {
+ "epoch": 27.419213973799128,
+ "grad_norm": 0.01522731315344572,
+ "learning_rate": 0.0006,
+ "loss": 4.223751068115234,
+ "step": 1974
+ },
+ {
+ "epoch": 27.433187772925763,
+ "grad_norm": 0.01581776700913906,
+ "learning_rate": 0.0006,
+ "loss": 4.351379871368408,
+ "step": 1975
+ },
+ {
+ "epoch": 27.4471615720524,
+ "grad_norm": 0.016394605860114098,
+ "learning_rate": 0.0006,
+ "loss": 4.2412919998168945,
+ "step": 1976
+ },
+ {
+ "epoch": 27.46113537117904,
+ "grad_norm": 0.01751169003546238,
+ "learning_rate": 0.0006,
+ "loss": 4.257606506347656,
+ "step": 1977
+ },
+ {
+ "epoch": 27.475109170305675,
+ "grad_norm": 0.017452310770750046,
+ "learning_rate": 0.0006,
+ "loss": 4.380134582519531,
+ "step": 1978
+ },
+ {
+ "epoch": 27.489082969432314,
+ "grad_norm": 0.017741898074746132,
+ "learning_rate": 0.0006,
+ "loss": 4.279632568359375,
+ "step": 1979
+ },
+ {
+ "epoch": 27.503056768558952,
+ "grad_norm": 0.017680590972304344,
+ "learning_rate": 0.0006,
+ "loss": 4.273406028747559,
+ "step": 1980
+ },
+ {
+ "epoch": 27.51703056768559,
+ "grad_norm": 0.01864718459546566,
+ "learning_rate": 0.0006,
+ "loss": 4.278583526611328,
+ "step": 1981
+ },
+ {
+ "epoch": 27.531004366812226,
+ "grad_norm": 0.019558541476726532,
+ "learning_rate": 0.0006,
+ "loss": 4.211919784545898,
+ "step": 1982
+ },
+ {
+ "epoch": 27.544978165938865,
+ "grad_norm": 0.019467400386929512,
+ "learning_rate": 0.0006,
+ "loss": 4.276822566986084,
+ "step": 1983
+ },
+ {
+ "epoch": 27.558951965065503,
+ "grad_norm": 0.01756172813475132,
+ "learning_rate": 0.0006,
+ "loss": 4.165379047393799,
+ "step": 1984
+ },
+ {
+ "epoch": 27.57292576419214,
+ "grad_norm": 0.017458263784646988,
+ "learning_rate": 0.0006,
+ "loss": 4.2992401123046875,
+ "step": 1985
+ },
+ {
+ "epoch": 27.586899563318777,
+ "grad_norm": 0.018089665099978447,
+ "learning_rate": 0.0006,
+ "loss": 4.243378639221191,
+ "step": 1986
+ },
+ {
+ "epoch": 27.600873362445416,
+ "grad_norm": 0.019535524770617485,
+ "learning_rate": 0.0006,
+ "loss": 4.321477890014648,
+ "step": 1987
+ },
+ {
+ "epoch": 27.61484716157205,
+ "grad_norm": 0.01997970975935459,
+ "learning_rate": 0.0006,
+ "loss": 4.123082637786865,
+ "step": 1988
+ },
+ {
+ "epoch": 27.62882096069869,
+ "grad_norm": 0.01913067139685154,
+ "learning_rate": 0.0006,
+ "loss": 4.289896488189697,
+ "step": 1989
+ },
+ {
+ "epoch": 27.64279475982533,
+ "grad_norm": 0.0169806070625782,
+ "learning_rate": 0.0006,
+ "loss": 4.247004985809326,
+ "step": 1990
+ },
+ {
+ "epoch": 27.656768558951963,
+ "grad_norm": 0.017315007746219635,
+ "learning_rate": 0.0006,
+ "loss": 4.2095441818237305,
+ "step": 1991
+ },
+ {
+ "epoch": 27.670742358078602,
+ "grad_norm": 0.015098759904503822,
+ "learning_rate": 0.0006,
+ "loss": 4.145560264587402,
+ "step": 1992
+ },
+ {
+ "epoch": 27.68471615720524,
+ "grad_norm": 0.016240514814853668,
+ "learning_rate": 0.0006,
+ "loss": 4.254927635192871,
+ "step": 1993
+ },
+ {
+ "epoch": 27.69868995633188,
+ "grad_norm": 0.01621108129620552,
+ "learning_rate": 0.0006,
+ "loss": 4.219725608825684,
+ "step": 1994
+ },
+ {
+ "epoch": 27.712663755458514,
+ "grad_norm": 0.015191009268164635,
+ "learning_rate": 0.0006,
+ "loss": 4.194345474243164,
+ "step": 1995
+ },
+ {
+ "epoch": 27.726637554585153,
+ "grad_norm": 0.014387983828783035,
+ "learning_rate": 0.0006,
+ "loss": 4.211467742919922,
+ "step": 1996
+ },
+ {
+ "epoch": 27.74061135371179,
+ "grad_norm": 0.015460561029613018,
+ "learning_rate": 0.0006,
+ "loss": 4.358700752258301,
+ "step": 1997
+ },
+ {
+ "epoch": 27.754585152838427,
+ "grad_norm": 0.015100076794624329,
+ "learning_rate": 0.0006,
+ "loss": 4.171223163604736,
+ "step": 1998
+ },
+ {
+ "epoch": 27.768558951965066,
+ "grad_norm": 0.015607891604304314,
+ "learning_rate": 0.0006,
+ "loss": 4.217874526977539,
+ "step": 1999
+ },
+ {
+ "epoch": 27.782532751091704,
+ "grad_norm": 0.015650689601898193,
+ "learning_rate": 0.0006,
+ "loss": 4.280780792236328,
+ "step": 2000
+ },
+ {
+ "epoch": 27.79650655021834,
+ "grad_norm": 0.01580994389951229,
+ "learning_rate": 0.0006,
+ "loss": 4.159212112426758,
+ "step": 2001
+ },
+ {
+ "epoch": 27.810480349344978,
+ "grad_norm": 0.016607413068413734,
+ "learning_rate": 0.0006,
+ "loss": 4.238853454589844,
+ "step": 2002
+ },
+ {
+ "epoch": 27.824454148471617,
+ "grad_norm": 0.017383860424160957,
+ "learning_rate": 0.0006,
+ "loss": 4.233944892883301,
+ "step": 2003
+ },
+ {
+ "epoch": 27.83842794759825,
+ "grad_norm": 0.014966659247875214,
+ "learning_rate": 0.0006,
+ "loss": 4.167392730712891,
+ "step": 2004
+ },
+ {
+ "epoch": 27.85240174672489,
+ "grad_norm": 0.014589878730475903,
+ "learning_rate": 0.0006,
+ "loss": 4.2390899658203125,
+ "step": 2005
+ },
+ {
+ "epoch": 27.86637554585153,
+ "grad_norm": 0.017179450020194054,
+ "learning_rate": 0.0006,
+ "loss": 4.365785598754883,
+ "step": 2006
+ },
+ {
+ "epoch": 27.880349344978168,
+ "grad_norm": 0.01893901452422142,
+ "learning_rate": 0.0006,
+ "loss": 4.289430141448975,
+ "step": 2007
+ },
+ {
+ "epoch": 27.894323144104803,
+ "grad_norm": 0.018619712442159653,
+ "learning_rate": 0.0006,
+ "loss": 4.310665130615234,
+ "step": 2008
+ },
+ {
+ "epoch": 27.90829694323144,
+ "grad_norm": 0.019180385395884514,
+ "learning_rate": 0.0006,
+ "loss": 4.27454137802124,
+ "step": 2009
+ },
+ {
+ "epoch": 27.92227074235808,
+ "grad_norm": 0.017811523750424385,
+ "learning_rate": 0.0006,
+ "loss": 4.312562465667725,
+ "step": 2010
+ },
+ {
+ "epoch": 27.936244541484715,
+ "grad_norm": 0.016583573073148727,
+ "learning_rate": 0.0006,
+ "loss": 4.202253341674805,
+ "step": 2011
+ },
+ {
+ "epoch": 27.950218340611354,
+ "grad_norm": 0.01658615842461586,
+ "learning_rate": 0.0006,
+ "loss": 4.183433532714844,
+ "step": 2012
+ },
+ {
+ "epoch": 27.964192139737992,
+ "grad_norm": 0.015327401459217072,
+ "learning_rate": 0.0006,
+ "loss": 4.2125139236450195,
+ "step": 2013
+ },
+ {
+ "epoch": 27.978165938864628,
+ "grad_norm": 0.015004601329565048,
+ "learning_rate": 0.0006,
+ "loss": 4.257392883300781,
+ "step": 2014
+ },
+ {
+ "epoch": 27.992139737991266,
+ "grad_norm": 0.015259161591529846,
+ "learning_rate": 0.0006,
+ "loss": 4.241818428039551,
+ "step": 2015
+ },
+ {
+ "epoch": 28.0,
+ "grad_norm": 0.016808776184916496,
+ "learning_rate": 0.0006,
+ "loss": 4.146241188049316,
+ "step": 2016
+ },
+ {
+ "epoch": 28.0,
+ "eval_loss": 4.663090705871582,
+ "eval_runtime": 56.4559,
+ "eval_samples_per_second": 43.255,
+ "eval_steps_per_second": 1.364,
+ "step": 2016
+ },
+ {
+ "epoch": 28.01397379912664,
+ "grad_norm": 0.015784382820129395,
+ "learning_rate": 0.0006,
+ "loss": 4.044191360473633,
+ "step": 2017
+ },
+ {
+ "epoch": 28.027947598253274,
+ "grad_norm": 0.015636710450053215,
+ "learning_rate": 0.0006,
+ "loss": 4.128754615783691,
+ "step": 2018
+ },
+ {
+ "epoch": 28.041921397379912,
+ "grad_norm": 0.015364975668489933,
+ "learning_rate": 0.0006,
+ "loss": 4.191835403442383,
+ "step": 2019
+ },
+ {
+ "epoch": 28.05589519650655,
+ "grad_norm": 0.015015700832009315,
+ "learning_rate": 0.0006,
+ "loss": 4.120022773742676,
+ "step": 2020
+ },
+ {
+ "epoch": 28.069868995633186,
+ "grad_norm": 0.016666799783706665,
+ "learning_rate": 0.0006,
+ "loss": 4.33906364440918,
+ "step": 2021
+ },
+ {
+ "epoch": 28.083842794759825,
+ "grad_norm": 0.01780104823410511,
+ "learning_rate": 0.0006,
+ "loss": 4.279458999633789,
+ "step": 2022
+ },
+ {
+ "epoch": 28.097816593886463,
+ "grad_norm": 0.01870882883667946,
+ "learning_rate": 0.0006,
+ "loss": 4.34728479385376,
+ "step": 2023
+ },
+ {
+ "epoch": 28.111790393013102,
+ "grad_norm": 0.018406381830573082,
+ "learning_rate": 0.0006,
+ "loss": 4.211311340332031,
+ "step": 2024
+ },
+ {
+ "epoch": 28.125764192139737,
+ "grad_norm": 0.017833448946475983,
+ "learning_rate": 0.0006,
+ "loss": 4.223254203796387,
+ "step": 2025
+ },
+ {
+ "epoch": 28.139737991266376,
+ "grad_norm": 0.017552688717842102,
+ "learning_rate": 0.0006,
+ "loss": 4.208823204040527,
+ "step": 2026
+ },
+ {
+ "epoch": 28.153711790393015,
+ "grad_norm": 0.018500229343771935,
+ "learning_rate": 0.0006,
+ "loss": 4.198025226593018,
+ "step": 2027
+ },
+ {
+ "epoch": 28.16768558951965,
+ "grad_norm": 0.01782156340777874,
+ "learning_rate": 0.0006,
+ "loss": 4.179978370666504,
+ "step": 2028
+ },
+ {
+ "epoch": 28.18165938864629,
+ "grad_norm": 0.019580967724323273,
+ "learning_rate": 0.0006,
+ "loss": 4.291114807128906,
+ "step": 2029
+ },
+ {
+ "epoch": 28.195633187772927,
+ "grad_norm": 0.020804036408662796,
+ "learning_rate": 0.0006,
+ "loss": 4.220607280731201,
+ "step": 2030
+ },
+ {
+ "epoch": 28.209606986899562,
+ "grad_norm": 0.019595693796873093,
+ "learning_rate": 0.0006,
+ "loss": 4.2905731201171875,
+ "step": 2031
+ },
+ {
+ "epoch": 28.2235807860262,
+ "grad_norm": 0.019247086718678474,
+ "learning_rate": 0.0006,
+ "loss": 4.2467360496521,
+ "step": 2032
+ },
+ {
+ "epoch": 28.23755458515284,
+ "grad_norm": 0.01958036608994007,
+ "learning_rate": 0.0006,
+ "loss": 4.286137580871582,
+ "step": 2033
+ },
+ {
+ "epoch": 28.251528384279474,
+ "grad_norm": 0.021354753524065018,
+ "learning_rate": 0.0006,
+ "loss": 4.234755039215088,
+ "step": 2034
+ },
+ {
+ "epoch": 28.265502183406113,
+ "grad_norm": 0.02206375077366829,
+ "learning_rate": 0.0006,
+ "loss": 4.149641036987305,
+ "step": 2035
+ },
+ {
+ "epoch": 28.27947598253275,
+ "grad_norm": 0.02201310358941555,
+ "learning_rate": 0.0006,
+ "loss": 4.1928629875183105,
+ "step": 2036
+ },
+ {
+ "epoch": 28.29344978165939,
+ "grad_norm": 0.01961950585246086,
+ "learning_rate": 0.0006,
+ "loss": 4.211404323577881,
+ "step": 2037
+ },
+ {
+ "epoch": 28.307423580786025,
+ "grad_norm": 0.019211795181035995,
+ "learning_rate": 0.0006,
+ "loss": 4.200976848602295,
+ "step": 2038
+ },
+ {
+ "epoch": 28.321397379912664,
+ "grad_norm": 0.01779370754957199,
+ "learning_rate": 0.0006,
+ "loss": 4.249148845672607,
+ "step": 2039
+ },
+ {
+ "epoch": 28.335371179039303,
+ "grad_norm": 0.018635908141732216,
+ "learning_rate": 0.0006,
+ "loss": 4.11806058883667,
+ "step": 2040
+ },
+ {
+ "epoch": 28.349344978165938,
+ "grad_norm": 0.019792694598436356,
+ "learning_rate": 0.0006,
+ "loss": 4.335224151611328,
+ "step": 2041
+ },
+ {
+ "epoch": 28.363318777292577,
+ "grad_norm": 0.021731717512011528,
+ "learning_rate": 0.0006,
+ "loss": 4.302778244018555,
+ "step": 2042
+ },
+ {
+ "epoch": 28.377292576419215,
+ "grad_norm": 0.020823447033762932,
+ "learning_rate": 0.0006,
+ "loss": 4.260610580444336,
+ "step": 2043
+ },
+ {
+ "epoch": 28.39126637554585,
+ "grad_norm": 0.019385412335395813,
+ "learning_rate": 0.0006,
+ "loss": 4.2247724533081055,
+ "step": 2044
+ },
+ {
+ "epoch": 28.40524017467249,
+ "grad_norm": 0.019539108499884605,
+ "learning_rate": 0.0006,
+ "loss": 4.236655235290527,
+ "step": 2045
+ },
+ {
+ "epoch": 28.419213973799128,
+ "grad_norm": 0.01942690648138523,
+ "learning_rate": 0.0006,
+ "loss": 4.251850605010986,
+ "step": 2046
+ },
+ {
+ "epoch": 28.433187772925763,
+ "grad_norm": 0.019667886197566986,
+ "learning_rate": 0.0006,
+ "loss": 4.222312927246094,
+ "step": 2047
+ },
+ {
+ "epoch": 28.4471615720524,
+ "grad_norm": 0.02029012329876423,
+ "learning_rate": 0.0006,
+ "loss": 4.246252059936523,
+ "step": 2048
+ },
+ {
+ "epoch": 28.46113537117904,
+ "grad_norm": 0.01784469373524189,
+ "learning_rate": 0.0006,
+ "loss": 4.192128658294678,
+ "step": 2049
+ },
+ {
+ "epoch": 28.475109170305675,
+ "grad_norm": 0.0160287544131279,
+ "learning_rate": 0.0006,
+ "loss": 4.208607196807861,
+ "step": 2050
+ },
+ {
+ "epoch": 28.489082969432314,
+ "grad_norm": 0.015348346903920174,
+ "learning_rate": 0.0006,
+ "loss": 4.236598014831543,
+ "step": 2051
+ },
+ {
+ "epoch": 28.503056768558952,
+ "grad_norm": 0.015198206529021263,
+ "learning_rate": 0.0006,
+ "loss": 4.235552787780762,
+ "step": 2052
+ },
+ {
+ "epoch": 28.51703056768559,
+ "grad_norm": 0.01523754745721817,
+ "learning_rate": 0.0006,
+ "loss": 4.1941728591918945,
+ "step": 2053
+ },
+ {
+ "epoch": 28.531004366812226,
+ "grad_norm": 0.015559614636003971,
+ "learning_rate": 0.0006,
+ "loss": 4.294757843017578,
+ "step": 2054
+ },
+ {
+ "epoch": 28.544978165938865,
+ "grad_norm": 0.01568358950316906,
+ "learning_rate": 0.0006,
+ "loss": 4.213065147399902,
+ "step": 2055
+ },
+ {
+ "epoch": 28.558951965065503,
+ "grad_norm": 0.015336516313254833,
+ "learning_rate": 0.0006,
+ "loss": 4.240293979644775,
+ "step": 2056
+ },
+ {
+ "epoch": 28.57292576419214,
+ "grad_norm": 0.015041496604681015,
+ "learning_rate": 0.0006,
+ "loss": 4.273397445678711,
+ "step": 2057
+ },
+ {
+ "epoch": 28.586899563318777,
+ "grad_norm": 0.01610128954052925,
+ "learning_rate": 0.0006,
+ "loss": 4.125369071960449,
+ "step": 2058
+ },
+ {
+ "epoch": 28.600873362445416,
+ "grad_norm": 0.015974976122379303,
+ "learning_rate": 0.0006,
+ "loss": 4.133459091186523,
+ "step": 2059
+ },
+ {
+ "epoch": 28.61484716157205,
+ "grad_norm": 0.015936799347400665,
+ "learning_rate": 0.0006,
+ "loss": 4.306667327880859,
+ "step": 2060
+ },
+ {
+ "epoch": 28.62882096069869,
+ "grad_norm": 0.01795247197151184,
+ "learning_rate": 0.0006,
+ "loss": 4.205156326293945,
+ "step": 2061
+ },
+ {
+ "epoch": 28.64279475982533,
+ "grad_norm": 0.019639814272522926,
+ "learning_rate": 0.0006,
+ "loss": 4.277614593505859,
+ "step": 2062
+ },
+ {
+ "epoch": 28.656768558951963,
+ "grad_norm": 0.01722300611436367,
+ "learning_rate": 0.0006,
+ "loss": 4.191032886505127,
+ "step": 2063
+ },
+ {
+ "epoch": 28.670742358078602,
+ "grad_norm": 0.01612633652985096,
+ "learning_rate": 0.0006,
+ "loss": 4.20717191696167,
+ "step": 2064
+ },
+ {
+ "epoch": 28.68471615720524,
+ "grad_norm": 0.01603279449045658,
+ "learning_rate": 0.0006,
+ "loss": 4.228884220123291,
+ "step": 2065
+ },
+ {
+ "epoch": 28.69868995633188,
+ "grad_norm": 0.016405927017331123,
+ "learning_rate": 0.0006,
+ "loss": 4.201292991638184,
+ "step": 2066
+ },
+ {
+ "epoch": 28.712663755458514,
+ "grad_norm": 0.016644183546304703,
+ "learning_rate": 0.0006,
+ "loss": 4.239282131195068,
+ "step": 2067
+ },
+ {
+ "epoch": 28.726637554585153,
+ "grad_norm": 0.0175587497651577,
+ "learning_rate": 0.0006,
+ "loss": 4.289038181304932,
+ "step": 2068
+ },
+ {
+ "epoch": 28.74061135371179,
+ "grad_norm": 0.018223397433757782,
+ "learning_rate": 0.0006,
+ "loss": 4.2266998291015625,
+ "step": 2069
+ },
+ {
+ "epoch": 28.754585152838427,
+ "grad_norm": 0.018680868670344353,
+ "learning_rate": 0.0006,
+ "loss": 4.348849296569824,
+ "step": 2070
+ },
+ {
+ "epoch": 28.768558951965066,
+ "grad_norm": 0.016726728528738022,
+ "learning_rate": 0.0006,
+ "loss": 4.292080879211426,
+ "step": 2071
+ },
+ {
+ "epoch": 28.782532751091704,
+ "grad_norm": 0.017386795952916145,
+ "learning_rate": 0.0006,
+ "loss": 4.208870887756348,
+ "step": 2072
+ },
+ {
+ "epoch": 28.79650655021834,
+ "grad_norm": 0.01865178905427456,
+ "learning_rate": 0.0006,
+ "loss": 4.054150581359863,
+ "step": 2073
+ },
+ {
+ "epoch": 28.810480349344978,
+ "grad_norm": 0.01776815392076969,
+ "learning_rate": 0.0006,
+ "loss": 4.202404499053955,
+ "step": 2074
+ },
+ {
+ "epoch": 28.824454148471617,
+ "grad_norm": 0.01643497310578823,
+ "learning_rate": 0.0006,
+ "loss": 4.131855010986328,
+ "step": 2075
+ },
+ {
+ "epoch": 28.83842794759825,
+ "grad_norm": 0.01574273779988289,
+ "learning_rate": 0.0006,
+ "loss": 4.296788692474365,
+ "step": 2076
+ },
+ {
+ "epoch": 28.85240174672489,
+ "grad_norm": 0.015613908879458904,
+ "learning_rate": 0.0006,
+ "loss": 4.210949897766113,
+ "step": 2077
+ },
+ {
+ "epoch": 28.86637554585153,
+ "grad_norm": 0.015543212182819843,
+ "learning_rate": 0.0006,
+ "loss": 4.218136310577393,
+ "step": 2078
+ },
+ {
+ "epoch": 28.880349344978168,
+ "grad_norm": 0.014681251719594002,
+ "learning_rate": 0.0006,
+ "loss": 4.19401741027832,
+ "step": 2079
+ },
+ {
+ "epoch": 28.894323144104803,
+ "grad_norm": 0.014393828809261322,
+ "learning_rate": 0.0006,
+ "loss": 4.198468208312988,
+ "step": 2080
+ },
+ {
+ "epoch": 28.90829694323144,
+ "grad_norm": 0.01580170728266239,
+ "learning_rate": 0.0006,
+ "loss": 4.224969863891602,
+ "step": 2081
+ },
+ {
+ "epoch": 28.92227074235808,
+ "grad_norm": 0.01484165620058775,
+ "learning_rate": 0.0006,
+ "loss": 4.12236213684082,
+ "step": 2082
+ },
+ {
+ "epoch": 28.936244541484715,
+ "grad_norm": 0.01497623696923256,
+ "learning_rate": 0.0006,
+ "loss": 4.24141788482666,
+ "step": 2083
+ },
+ {
+ "epoch": 28.950218340611354,
+ "grad_norm": 0.015080046840012074,
+ "learning_rate": 0.0006,
+ "loss": 4.284151077270508,
+ "step": 2084
+ },
+ {
+ "epoch": 28.964192139737992,
+ "grad_norm": 0.016207212582230568,
+ "learning_rate": 0.0006,
+ "loss": 4.298693656921387,
+ "step": 2085
+ },
+ {
+ "epoch": 28.978165938864628,
+ "grad_norm": 0.015474417246878147,
+ "learning_rate": 0.0006,
+ "loss": 4.2412261962890625,
+ "step": 2086
+ },
+ {
+ "epoch": 28.992139737991266,
+ "grad_norm": 0.01515351701527834,
+ "learning_rate": 0.0006,
+ "loss": 4.3054423332214355,
+ "step": 2087
+ },
+ {
+ "epoch": 29.0,
+ "grad_norm": 0.016411345452070236,
+ "learning_rate": 0.0006,
+ "loss": 4.104830741882324,
+ "step": 2088
+ },
+ {
+ "epoch": 29.0,
+ "eval_loss": 4.661567687988281,
+ "eval_runtime": 57.1244,
+ "eval_samples_per_second": 42.749,
+ "eval_steps_per_second": 1.348,
+ "step": 2088
+ },
+ {
+ "epoch": 29.01397379912664,
+ "grad_norm": 0.01665342226624489,
+ "learning_rate": 0.0006,
+ "loss": 4.311391830444336,
+ "step": 2089
+ },
+ {
+ "epoch": 29.027947598253274,
+ "grad_norm": 0.018053214997053146,
+ "learning_rate": 0.0006,
+ "loss": 4.135597229003906,
+ "step": 2090
+ },
+ {
+ "epoch": 29.041921397379912,
+ "grad_norm": 0.018213748931884766,
+ "learning_rate": 0.0006,
+ "loss": 4.303299903869629,
+ "step": 2091
+ },
+ {
+ "epoch": 29.05589519650655,
+ "grad_norm": 0.018460242077708244,
+ "learning_rate": 0.0006,
+ "loss": 4.20850944519043,
+ "step": 2092
+ },
+ {
+ "epoch": 29.069868995633186,
+ "grad_norm": 0.018365809693932533,
+ "learning_rate": 0.0006,
+ "loss": 4.275847434997559,
+ "step": 2093
+ },
+ {
+ "epoch": 29.083842794759825,
+ "grad_norm": 0.018346186727285385,
+ "learning_rate": 0.0006,
+ "loss": 4.202091217041016,
+ "step": 2094
+ },
+ {
+ "epoch": 29.097816593886463,
+ "grad_norm": 0.017920587211847305,
+ "learning_rate": 0.0006,
+ "loss": 4.182092666625977,
+ "step": 2095
+ },
+ {
+ "epoch": 29.111790393013102,
+ "grad_norm": 0.01812003180384636,
+ "learning_rate": 0.0006,
+ "loss": 4.259721755981445,
+ "step": 2096
+ },
+ {
+ "epoch": 29.125764192139737,
+ "grad_norm": 0.01700986735522747,
+ "learning_rate": 0.0006,
+ "loss": 4.067573070526123,
+ "step": 2097
+ },
+ {
+ "epoch": 29.139737991266376,
+ "grad_norm": 0.018069760873913765,
+ "learning_rate": 0.0006,
+ "loss": 4.234400749206543,
+ "step": 2098
+ },
+ {
+ "epoch": 29.153711790393015,
+ "grad_norm": 0.01905428245663643,
+ "learning_rate": 0.0006,
+ "loss": 4.16609001159668,
+ "step": 2099
+ },
+ {
+ "epoch": 29.16768558951965,
+ "grad_norm": 0.018711242824792862,
+ "learning_rate": 0.0006,
+ "loss": 4.249088764190674,
+ "step": 2100
+ },
+ {
+ "epoch": 29.18165938864629,
+ "grad_norm": 0.018721306696534157,
+ "learning_rate": 0.0006,
+ "loss": 4.14540958404541,
+ "step": 2101
+ },
+ {
+ "epoch": 29.195633187772927,
+ "grad_norm": 0.020139874890446663,
+ "learning_rate": 0.0006,
+ "loss": 4.049248695373535,
+ "step": 2102
+ },
+ {
+ "epoch": 29.209606986899562,
+ "grad_norm": 0.020849574357271194,
+ "learning_rate": 0.0006,
+ "loss": 4.235224723815918,
+ "step": 2103
+ },
+ {
+ "epoch": 29.2235807860262,
+ "grad_norm": 0.018693549558520317,
+ "learning_rate": 0.0006,
+ "loss": 4.232961177825928,
+ "step": 2104
+ },
+ {
+ "epoch": 29.23755458515284,
+ "grad_norm": 0.017889857292175293,
+ "learning_rate": 0.0006,
+ "loss": 4.171422004699707,
+ "step": 2105
+ },
+ {
+ "epoch": 29.251528384279474,
+ "grad_norm": 0.016690224409103394,
+ "learning_rate": 0.0006,
+ "loss": 4.244717597961426,
+ "step": 2106
+ },
+ {
+ "epoch": 29.265502183406113,
+ "grad_norm": 0.01708616502583027,
+ "learning_rate": 0.0006,
+ "loss": 4.125068664550781,
+ "step": 2107
+ },
+ {
+ "epoch": 29.27947598253275,
+ "grad_norm": 0.017928237095475197,
+ "learning_rate": 0.0006,
+ "loss": 4.208292007446289,
+ "step": 2108
+ },
+ {
+ "epoch": 29.29344978165939,
+ "grad_norm": 0.018250394612550735,
+ "learning_rate": 0.0006,
+ "loss": 4.118630886077881,
+ "step": 2109
+ },
+ {
+ "epoch": 29.307423580786025,
+ "grad_norm": 0.020160246640443802,
+ "learning_rate": 0.0006,
+ "loss": 4.206025123596191,
+ "step": 2110
+ },
+ {
+ "epoch": 29.321397379912664,
+ "grad_norm": 0.020045453682541847,
+ "learning_rate": 0.0006,
+ "loss": 4.2758684158325195,
+ "step": 2111
+ },
+ {
+ "epoch": 29.335371179039303,
+ "grad_norm": 0.019556907936930656,
+ "learning_rate": 0.0006,
+ "loss": 4.188272953033447,
+ "step": 2112
+ },
+ {
+ "epoch": 29.349344978165938,
+ "grad_norm": 0.019185200333595276,
+ "learning_rate": 0.0006,
+ "loss": 4.17451286315918,
+ "step": 2113
+ },
+ {
+ "epoch": 29.363318777292577,
+ "grad_norm": 0.01816629432141781,
+ "learning_rate": 0.0006,
+ "loss": 4.165246963500977,
+ "step": 2114
+ },
+ {
+ "epoch": 29.377292576419215,
+ "grad_norm": 0.01865146867930889,
+ "learning_rate": 0.0006,
+ "loss": 4.148184776306152,
+ "step": 2115
+ },
+ {
+ "epoch": 29.39126637554585,
+ "grad_norm": 0.018338464200496674,
+ "learning_rate": 0.0006,
+ "loss": 4.288936614990234,
+ "step": 2116
+ },
+ {
+ "epoch": 29.40524017467249,
+ "grad_norm": 0.018986187875270844,
+ "learning_rate": 0.0006,
+ "loss": 4.183379650115967,
+ "step": 2117
+ },
+ {
+ "epoch": 29.419213973799128,
+ "grad_norm": 0.018409637734293938,
+ "learning_rate": 0.0006,
+ "loss": 4.057574272155762,
+ "step": 2118
+ },
+ {
+ "epoch": 29.433187772925763,
+ "grad_norm": 0.016664542257785797,
+ "learning_rate": 0.0006,
+ "loss": 4.18517541885376,
+ "step": 2119
+ },
+ {
+ "epoch": 29.4471615720524,
+ "grad_norm": 0.016486089676618576,
+ "learning_rate": 0.0006,
+ "loss": 4.179367542266846,
+ "step": 2120
+ },
+ {
+ "epoch": 29.46113537117904,
+ "grad_norm": 0.01787523552775383,
+ "learning_rate": 0.0006,
+ "loss": 4.184126377105713,
+ "step": 2121
+ },
+ {
+ "epoch": 29.475109170305675,
+ "grad_norm": 0.016574613749980927,
+ "learning_rate": 0.0006,
+ "loss": 4.290170192718506,
+ "step": 2122
+ },
+ {
+ "epoch": 29.489082969432314,
+ "grad_norm": 0.01643376611173153,
+ "learning_rate": 0.0006,
+ "loss": 4.052913665771484,
+ "step": 2123
+ },
+ {
+ "epoch": 29.503056768558952,
+ "grad_norm": 0.016503969207406044,
+ "learning_rate": 0.0006,
+ "loss": 4.335184097290039,
+ "step": 2124
+ },
+ {
+ "epoch": 29.51703056768559,
+ "grad_norm": 0.017596479505300522,
+ "learning_rate": 0.0006,
+ "loss": 4.165539741516113,
+ "step": 2125
+ },
+ {
+ "epoch": 29.531004366812226,
+ "grad_norm": 0.017468582838773727,
+ "learning_rate": 0.0006,
+ "loss": 4.315964698791504,
+ "step": 2126
+ },
+ {
+ "epoch": 29.544978165938865,
+ "grad_norm": 0.01724618673324585,
+ "learning_rate": 0.0006,
+ "loss": 4.243681907653809,
+ "step": 2127
+ },
+ {
+ "epoch": 29.558951965065503,
+ "grad_norm": 0.018384616822004318,
+ "learning_rate": 0.0006,
+ "loss": 4.112447261810303,
+ "step": 2128
+ },
+ {
+ "epoch": 29.57292576419214,
+ "grad_norm": 0.01824074238538742,
+ "learning_rate": 0.0006,
+ "loss": 4.236065864562988,
+ "step": 2129
+ },
+ {
+ "epoch": 29.586899563318777,
+ "grad_norm": 0.01625971868634224,
+ "learning_rate": 0.0006,
+ "loss": 4.18143892288208,
+ "step": 2130
+ },
+ {
+ "epoch": 29.600873362445416,
+ "grad_norm": 0.015537494793534279,
+ "learning_rate": 0.0006,
+ "loss": 4.225147247314453,
+ "step": 2131
+ },
+ {
+ "epoch": 29.61484716157205,
+ "grad_norm": 0.015818974003195763,
+ "learning_rate": 0.0006,
+ "loss": 4.3100175857543945,
+ "step": 2132
+ },
+ {
+ "epoch": 29.62882096069869,
+ "grad_norm": 0.015909671783447266,
+ "learning_rate": 0.0006,
+ "loss": 4.169775009155273,
+ "step": 2133
+ },
+ {
+ "epoch": 29.64279475982533,
+ "grad_norm": 0.01769658550620079,
+ "learning_rate": 0.0006,
+ "loss": 4.278904914855957,
+ "step": 2134
+ },
+ {
+ "epoch": 29.656768558951963,
+ "grad_norm": 0.015641704201698303,
+ "learning_rate": 0.0006,
+ "loss": 4.2426228523254395,
+ "step": 2135
+ },
+ {
+ "epoch": 29.670742358078602,
+ "grad_norm": 0.01565658301115036,
+ "learning_rate": 0.0006,
+ "loss": 4.232687950134277,
+ "step": 2136
+ },
+ {
+ "epoch": 29.68471615720524,
+ "grad_norm": 0.016485940665006638,
+ "learning_rate": 0.0006,
+ "loss": 4.209839820861816,
+ "step": 2137
+ },
+ {
+ "epoch": 29.69868995633188,
+ "grad_norm": 0.017622729763388634,
+ "learning_rate": 0.0006,
+ "loss": 4.196017265319824,
+ "step": 2138
+ },
+ {
+ "epoch": 29.712663755458514,
+ "grad_norm": 0.018320564180612564,
+ "learning_rate": 0.0006,
+ "loss": 4.225683689117432,
+ "step": 2139
+ },
+ {
+ "epoch": 29.726637554585153,
+ "grad_norm": 0.018140794709324837,
+ "learning_rate": 0.0006,
+ "loss": 4.311944961547852,
+ "step": 2140
+ },
+ {
+ "epoch": 29.74061135371179,
+ "grad_norm": 0.016227100044488907,
+ "learning_rate": 0.0006,
+ "loss": 4.230257511138916,
+ "step": 2141
+ },
+ {
+ "epoch": 29.754585152838427,
+ "grad_norm": 0.015759488567709923,
+ "learning_rate": 0.0006,
+ "loss": 4.119174003601074,
+ "step": 2142
+ },
+ {
+ "epoch": 29.768558951965066,
+ "grad_norm": 0.01757766678929329,
+ "learning_rate": 0.0006,
+ "loss": 4.165826797485352,
+ "step": 2143
+ },
+ {
+ "epoch": 29.782532751091704,
+ "grad_norm": 0.019661923870444298,
+ "learning_rate": 0.0006,
+ "loss": 4.392203330993652,
+ "step": 2144
+ },
+ {
+ "epoch": 29.79650655021834,
+ "grad_norm": 0.019478755071759224,
+ "learning_rate": 0.0006,
+ "loss": 4.348371982574463,
+ "step": 2145
+ },
+ {
+ "epoch": 29.810480349344978,
+ "grad_norm": 0.019149569794535637,
+ "learning_rate": 0.0006,
+ "loss": 4.34104061126709,
+ "step": 2146
+ },
+ {
+ "epoch": 29.824454148471617,
+ "grad_norm": 0.017707521095871925,
+ "learning_rate": 0.0006,
+ "loss": 4.228226661682129,
+ "step": 2147
+ },
+ {
+ "epoch": 29.83842794759825,
+ "grad_norm": 0.015615378506481647,
+ "learning_rate": 0.0006,
+ "loss": 4.175541877746582,
+ "step": 2148
+ },
+ {
+ "epoch": 29.85240174672489,
+ "grad_norm": 0.016111129894852638,
+ "learning_rate": 0.0006,
+ "loss": 4.267936706542969,
+ "step": 2149
+ },
+ {
+ "epoch": 29.86637554585153,
+ "grad_norm": 0.01528779324144125,
+ "learning_rate": 0.0006,
+ "loss": 4.186014175415039,
+ "step": 2150
+ },
+ {
+ "epoch": 29.880349344978168,
+ "grad_norm": 0.014925646595656872,
+ "learning_rate": 0.0006,
+ "loss": 4.174860000610352,
+ "step": 2151
+ },
+ {
+ "epoch": 29.894323144104803,
+ "grad_norm": 0.016917673870921135,
+ "learning_rate": 0.0006,
+ "loss": 4.33908748626709,
+ "step": 2152
+ },
+ {
+ "epoch": 29.90829694323144,
+ "grad_norm": 0.016007075086236,
+ "learning_rate": 0.0006,
+ "loss": 4.266141414642334,
+ "step": 2153
+ },
+ {
+ "epoch": 29.92227074235808,
+ "grad_norm": 0.016178956255316734,
+ "learning_rate": 0.0006,
+ "loss": 4.271215915679932,
+ "step": 2154
+ },
+ {
+ "epoch": 29.936244541484715,
+ "grad_norm": 0.016180871054530144,
+ "learning_rate": 0.0006,
+ "loss": 4.278801441192627,
+ "step": 2155
+ },
+ {
+ "epoch": 29.950218340611354,
+ "grad_norm": 0.015335503034293652,
+ "learning_rate": 0.0006,
+ "loss": 4.194978713989258,
+ "step": 2156
+ },
+ {
+ "epoch": 29.964192139737992,
+ "grad_norm": 0.016580764204263687,
+ "learning_rate": 0.0006,
+ "loss": 4.290809631347656,
+ "step": 2157
+ },
+ {
+ "epoch": 29.978165938864628,
+ "grad_norm": 0.01623128168284893,
+ "learning_rate": 0.0006,
+ "loss": 4.205451488494873,
+ "step": 2158
+ },
+ {
+ "epoch": 29.992139737991266,
+ "grad_norm": 0.014904248528182507,
+ "learning_rate": 0.0006,
+ "loss": 4.24924373626709,
+ "step": 2159
+ },
+ {
+ "epoch": 30.0,
+ "grad_norm": 0.017089299857616425,
+ "learning_rate": 0.0006,
+ "loss": 4.222072601318359,
+ "step": 2160
+ },
+ {
+ "epoch": 30.0,
+ "eval_loss": 4.585411071777344,
+ "eval_runtime": 57.1164,
+ "eval_samples_per_second": 42.755,
+ "eval_steps_per_second": 1.348,
+ "step": 2160
+ },
+ {
+ "epoch": 30.01397379912664,
+ "grad_norm": 0.017226383090019226,
+ "learning_rate": 0.0006,
+ "loss": 4.19467830657959,
+ "step": 2161
+ },
+ {
+ "epoch": 30.027947598253274,
+ "grad_norm": 0.01635066606104374,
+ "learning_rate": 0.0006,
+ "loss": 4.176875591278076,
+ "step": 2162
+ },
+ {
+ "epoch": 30.041921397379912,
+ "grad_norm": 0.01735362410545349,
+ "learning_rate": 0.0006,
+ "loss": 4.13667106628418,
+ "step": 2163
+ },
+ {
+ "epoch": 30.05589519650655,
+ "grad_norm": 0.017047109082341194,
+ "learning_rate": 0.0006,
+ "loss": 4.136329650878906,
+ "step": 2164
+ },
+ {
+ "epoch": 30.069868995633186,
+ "grad_norm": 0.018116053193807602,
+ "learning_rate": 0.0006,
+ "loss": 4.228706359863281,
+ "step": 2165
+ },
+ {
+ "epoch": 30.083842794759825,
+ "grad_norm": 0.01791389286518097,
+ "learning_rate": 0.0006,
+ "loss": 4.1597747802734375,
+ "step": 2166
+ },
+ {
+ "epoch": 30.097816593886463,
+ "grad_norm": 0.017793502658605576,
+ "learning_rate": 0.0006,
+ "loss": 4.0807976722717285,
+ "step": 2167
+ },
+ {
+ "epoch": 30.111790393013102,
+ "grad_norm": 0.01827503927052021,
+ "learning_rate": 0.0006,
+ "loss": 4.170646667480469,
+ "step": 2168
+ },
+ {
+ "epoch": 30.125764192139737,
+ "grad_norm": 0.017228564247488976,
+ "learning_rate": 0.0006,
+ "loss": 4.146934509277344,
+ "step": 2169
+ },
+ {
+ "epoch": 30.139737991266376,
+ "grad_norm": 0.01757437363266945,
+ "learning_rate": 0.0006,
+ "loss": 4.1859130859375,
+ "step": 2170
+ },
+ {
+ "epoch": 30.153711790393015,
+ "grad_norm": 0.01701093092560768,
+ "learning_rate": 0.0006,
+ "loss": 4.126981258392334,
+ "step": 2171
+ },
+ {
+ "epoch": 30.16768558951965,
+ "grad_norm": 0.015220489352941513,
+ "learning_rate": 0.0006,
+ "loss": 4.164018630981445,
+ "step": 2172
+ },
+ {
+ "epoch": 30.18165938864629,
+ "grad_norm": 0.016254238784313202,
+ "learning_rate": 0.0006,
+ "loss": 4.258275508880615,
+ "step": 2173
+ },
+ {
+ "epoch": 30.195633187772927,
+ "grad_norm": 0.016609614714980125,
+ "learning_rate": 0.0006,
+ "loss": 4.222233772277832,
+ "step": 2174
+ },
+ {
+ "epoch": 30.209606986899562,
+ "grad_norm": 0.016531318426132202,
+ "learning_rate": 0.0006,
+ "loss": 4.144489288330078,
+ "step": 2175
+ },
+ {
+ "epoch": 30.2235807860262,
+ "grad_norm": 0.01567668654024601,
+ "learning_rate": 0.0006,
+ "loss": 4.181916236877441,
+ "step": 2176
+ },
+ {
+ "epoch": 30.23755458515284,
+ "grad_norm": 0.017195681110024452,
+ "learning_rate": 0.0006,
+ "loss": 4.189693927764893,
+ "step": 2177
+ },
+ {
+ "epoch": 30.251528384279474,
+ "grad_norm": 0.016882948577404022,
+ "learning_rate": 0.0006,
+ "loss": 4.21604061126709,
+ "step": 2178
+ },
+ {
+ "epoch": 30.265502183406113,
+ "grad_norm": 0.016261622309684753,
+ "learning_rate": 0.0006,
+ "loss": 4.227639198303223,
+ "step": 2179
+ },
+ {
+ "epoch": 30.27947598253275,
+ "grad_norm": 0.016411812976002693,
+ "learning_rate": 0.0006,
+ "loss": 4.130362510681152,
+ "step": 2180
+ },
+ {
+ "epoch": 30.29344978165939,
+ "grad_norm": 0.01732019893825054,
+ "learning_rate": 0.0006,
+ "loss": 4.230443000793457,
+ "step": 2181
+ },
+ {
+ "epoch": 30.307423580786025,
+ "grad_norm": 0.01825905591249466,
+ "learning_rate": 0.0006,
+ "loss": 4.175817489624023,
+ "step": 2182
+ },
+ {
+ "epoch": 30.321397379912664,
+ "grad_norm": 0.01887713558971882,
+ "learning_rate": 0.0006,
+ "loss": 4.169020652770996,
+ "step": 2183
+ },
+ {
+ "epoch": 30.335371179039303,
+ "grad_norm": 0.019391661509871483,
+ "learning_rate": 0.0006,
+ "loss": 4.176318645477295,
+ "step": 2184
+ },
+ {
+ "epoch": 30.349344978165938,
+ "grad_norm": 0.020840002223849297,
+ "learning_rate": 0.0006,
+ "loss": 4.174228668212891,
+ "step": 2185
+ },
+ {
+ "epoch": 30.363318777292577,
+ "grad_norm": 0.020193178206682205,
+ "learning_rate": 0.0006,
+ "loss": 4.127919673919678,
+ "step": 2186
+ },
+ {
+ "epoch": 30.377292576419215,
+ "grad_norm": 0.018449561670422554,
+ "learning_rate": 0.0006,
+ "loss": 4.229487895965576,
+ "step": 2187
+ },
+ {
+ "epoch": 30.39126637554585,
+ "grad_norm": 0.01711983233690262,
+ "learning_rate": 0.0006,
+ "loss": 4.201651573181152,
+ "step": 2188
+ },
+ {
+ "epoch": 30.40524017467249,
+ "grad_norm": 0.017293395474553108,
+ "learning_rate": 0.0006,
+ "loss": 4.168704032897949,
+ "step": 2189
+ },
+ {
+ "epoch": 30.419213973799128,
+ "grad_norm": 0.018526358529925346,
+ "learning_rate": 0.0006,
+ "loss": 4.171864032745361,
+ "step": 2190
+ },
+ {
+ "epoch": 30.433187772925763,
+ "grad_norm": 0.019287291914224625,
+ "learning_rate": 0.0006,
+ "loss": 4.039534568786621,
+ "step": 2191
+ },
+ {
+ "epoch": 30.4471615720524,
+ "grad_norm": 0.016902444884181023,
+ "learning_rate": 0.0006,
+ "loss": 4.282787799835205,
+ "step": 2192
+ },
+ {
+ "epoch": 30.46113537117904,
+ "grad_norm": 0.01656750962138176,
+ "learning_rate": 0.0006,
+ "loss": 4.213098526000977,
+ "step": 2193
+ },
+ {
+ "epoch": 30.475109170305675,
+ "grad_norm": 0.016805065795779228,
+ "learning_rate": 0.0006,
+ "loss": 4.1583476066589355,
+ "step": 2194
+ },
+ {
+ "epoch": 30.489082969432314,
+ "grad_norm": 0.017166638746857643,
+ "learning_rate": 0.0006,
+ "loss": 4.188634395599365,
+ "step": 2195
+ },
+ {
+ "epoch": 30.503056768558952,
+ "grad_norm": 0.01592106744647026,
+ "learning_rate": 0.0006,
+ "loss": 4.234043121337891,
+ "step": 2196
+ },
+ {
+ "epoch": 30.51703056768559,
+ "grad_norm": 0.015270989388227463,
+ "learning_rate": 0.0006,
+ "loss": 4.301420211791992,
+ "step": 2197
+ },
+ {
+ "epoch": 30.531004366812226,
+ "grad_norm": 0.015194724313914776,
+ "learning_rate": 0.0006,
+ "loss": 4.18703031539917,
+ "step": 2198
+ },
+ {
+ "epoch": 30.544978165938865,
+ "grad_norm": 0.015587746165692806,
+ "learning_rate": 0.0006,
+ "loss": 4.138928413391113,
+ "step": 2199
+ },
+ {
+ "epoch": 30.558951965065503,
+ "grad_norm": 0.016029570251703262,
+ "learning_rate": 0.0006,
+ "loss": 4.290863990783691,
+ "step": 2200
+ },
+ {
+ "epoch": 30.57292576419214,
+ "grad_norm": 0.015651309862732887,
+ "learning_rate": 0.0006,
+ "loss": 4.272062301635742,
+ "step": 2201
+ },
+ {
+ "epoch": 30.586899563318777,
+ "grad_norm": 0.015992306172847748,
+ "learning_rate": 0.0006,
+ "loss": 4.2487406730651855,
+ "step": 2202
+ },
+ {
+ "epoch": 30.600873362445416,
+ "grad_norm": 0.014969157986342907,
+ "learning_rate": 0.0006,
+ "loss": 4.115347862243652,
+ "step": 2203
+ },
+ {
+ "epoch": 30.61484716157205,
+ "grad_norm": 0.014639027416706085,
+ "learning_rate": 0.0006,
+ "loss": 4.2207489013671875,
+ "step": 2204
+ },
+ {
+ "epoch": 30.62882096069869,
+ "grad_norm": 0.015059413388371468,
+ "learning_rate": 0.0006,
+ "loss": 4.274044990539551,
+ "step": 2205
+ },
+ {
+ "epoch": 30.64279475982533,
+ "grad_norm": 0.0166641678661108,
+ "learning_rate": 0.0006,
+ "loss": 4.191373348236084,
+ "step": 2206
+ },
+ {
+ "epoch": 30.656768558951963,
+ "grad_norm": 0.017806345596909523,
+ "learning_rate": 0.0006,
+ "loss": 4.139800548553467,
+ "step": 2207
+ },
+ {
+ "epoch": 30.670742358078602,
+ "grad_norm": 0.018712421879172325,
+ "learning_rate": 0.0006,
+ "loss": 4.035965442657471,
+ "step": 2208
+ },
+ {
+ "epoch": 30.68471615720524,
+ "grad_norm": 0.018457243219017982,
+ "learning_rate": 0.0006,
+ "loss": 4.2579240798950195,
+ "step": 2209
+ },
+ {
+ "epoch": 30.69868995633188,
+ "grad_norm": 0.01920999586582184,
+ "learning_rate": 0.0006,
+ "loss": 4.131112098693848,
+ "step": 2210
+ },
+ {
+ "epoch": 30.712663755458514,
+ "grad_norm": 0.01729944534599781,
+ "learning_rate": 0.0006,
+ "loss": 4.254059314727783,
+ "step": 2211
+ },
+ {
+ "epoch": 30.726637554585153,
+ "grad_norm": 0.016914231702685356,
+ "learning_rate": 0.0006,
+ "loss": 4.233428955078125,
+ "step": 2212
+ },
+ {
+ "epoch": 30.74061135371179,
+ "grad_norm": 0.01913098618388176,
+ "learning_rate": 0.0006,
+ "loss": 4.23525333404541,
+ "step": 2213
+ },
+ {
+ "epoch": 30.754585152838427,
+ "grad_norm": 0.019150281324982643,
+ "learning_rate": 0.0006,
+ "loss": 4.213695049285889,
+ "step": 2214
+ },
+ {
+ "epoch": 30.768558951965066,
+ "grad_norm": 0.018947452306747437,
+ "learning_rate": 0.0006,
+ "loss": 4.151697158813477,
+ "step": 2215
+ },
+ {
+ "epoch": 30.782532751091704,
+ "grad_norm": 0.016058508306741714,
+ "learning_rate": 0.0006,
+ "loss": 4.220280170440674,
+ "step": 2216
+ },
+ {
+ "epoch": 30.79650655021834,
+ "grad_norm": 0.017488619312644005,
+ "learning_rate": 0.0006,
+ "loss": 4.154428482055664,
+ "step": 2217
+ },
+ {
+ "epoch": 30.810480349344978,
+ "grad_norm": 0.017355090007185936,
+ "learning_rate": 0.0006,
+ "loss": 4.256086349487305,
+ "step": 2218
+ },
+ {
+ "epoch": 30.824454148471617,
+ "grad_norm": 0.016636032611131668,
+ "learning_rate": 0.0006,
+ "loss": 4.227564334869385,
+ "step": 2219
+ },
+ {
+ "epoch": 30.83842794759825,
+ "grad_norm": 0.016273891553282738,
+ "learning_rate": 0.0006,
+ "loss": 4.200466156005859,
+ "step": 2220
+ },
+ {
+ "epoch": 30.85240174672489,
+ "grad_norm": 0.016939911991357803,
+ "learning_rate": 0.0006,
+ "loss": 4.1571221351623535,
+ "step": 2221
+ },
+ {
+ "epoch": 30.86637554585153,
+ "grad_norm": 0.018332522362470627,
+ "learning_rate": 0.0006,
+ "loss": 4.228398323059082,
+ "step": 2222
+ },
+ {
+ "epoch": 30.880349344978168,
+ "grad_norm": 0.019508114084601402,
+ "learning_rate": 0.0006,
+ "loss": 4.222960472106934,
+ "step": 2223
+ },
+ {
+ "epoch": 30.894323144104803,
+ "grad_norm": 0.018440278246998787,
+ "learning_rate": 0.0006,
+ "loss": 4.261051177978516,
+ "step": 2224
+ },
+ {
+ "epoch": 30.90829694323144,
+ "grad_norm": 0.018113229423761368,
+ "learning_rate": 0.0006,
+ "loss": 4.109235763549805,
+ "step": 2225
+ },
+ {
+ "epoch": 30.92227074235808,
+ "grad_norm": 0.015392083674669266,
+ "learning_rate": 0.0006,
+ "loss": 4.21950101852417,
+ "step": 2226
+ },
+ {
+ "epoch": 30.936244541484715,
+ "grad_norm": 0.01633290946483612,
+ "learning_rate": 0.0006,
+ "loss": 4.176053047180176,
+ "step": 2227
+ },
+ {
+ "epoch": 30.950218340611354,
+ "grad_norm": 0.016236383467912674,
+ "learning_rate": 0.0006,
+ "loss": 4.314309120178223,
+ "step": 2228
+ },
+ {
+ "epoch": 30.964192139737992,
+ "grad_norm": 0.015284373424947262,
+ "learning_rate": 0.0006,
+ "loss": 4.303133487701416,
+ "step": 2229
+ },
+ {
+ "epoch": 30.978165938864628,
+ "grad_norm": 0.015982897952198982,
+ "learning_rate": 0.0006,
+ "loss": 4.25739860534668,
+ "step": 2230
+ },
+ {
+ "epoch": 30.992139737991266,
+ "grad_norm": 0.016682198271155357,
+ "learning_rate": 0.0006,
+ "loss": 4.292210578918457,
+ "step": 2231
+ },
+ {
+ "epoch": 31.0,
+ "grad_norm": 0.017521467059850693,
+ "learning_rate": 0.0006,
+ "loss": 4.170637130737305,
+ "step": 2232
+ },
+ {
+ "epoch": 31.0,
+ "eval_loss": 4.642786979675293,
+ "eval_runtime": 57.029,
+ "eval_samples_per_second": 42.82,
+ "eval_steps_per_second": 1.35,
+ "step": 2232
+ },
+ {
+ "epoch": 31.01397379912664,
+ "grad_norm": 0.016607852652668953,
+ "learning_rate": 0.0006,
+ "loss": 4.106224536895752,
+ "step": 2233
+ },
+ {
+ "epoch": 31.027947598253274,
+ "grad_norm": 0.018448904156684875,
+ "learning_rate": 0.0006,
+ "loss": 4.224459648132324,
+ "step": 2234
+ },
+ {
+ "epoch": 31.041921397379912,
+ "grad_norm": 0.020481228828430176,
+ "learning_rate": 0.0006,
+ "loss": 4.305548667907715,
+ "step": 2235
+ },
+ {
+ "epoch": 31.05589519650655,
+ "grad_norm": 0.01999707892537117,
+ "learning_rate": 0.0006,
+ "loss": 4.253663539886475,
+ "step": 2236
+ },
+ {
+ "epoch": 31.069868995633186,
+ "grad_norm": 0.020428303629159927,
+ "learning_rate": 0.0006,
+ "loss": 4.166954040527344,
+ "step": 2237
+ },
+ {
+ "epoch": 31.083842794759825,
+ "grad_norm": 0.022608213126659393,
+ "learning_rate": 0.0006,
+ "loss": 4.172819137573242,
+ "step": 2238
+ },
+ {
+ "epoch": 31.097816593886463,
+ "grad_norm": 0.024669520556926727,
+ "learning_rate": 0.0006,
+ "loss": 4.106586456298828,
+ "step": 2239
+ },
+ {
+ "epoch": 31.111790393013102,
+ "grad_norm": 0.02247351035475731,
+ "learning_rate": 0.0006,
+ "loss": 4.160271644592285,
+ "step": 2240
+ },
+ {
+ "epoch": 31.125764192139737,
+ "grad_norm": 0.0207810141146183,
+ "learning_rate": 0.0006,
+ "loss": 4.272491931915283,
+ "step": 2241
+ },
+ {
+ "epoch": 31.139737991266376,
+ "grad_norm": 0.021576499566435814,
+ "learning_rate": 0.0006,
+ "loss": 4.171474456787109,
+ "step": 2242
+ },
+ {
+ "epoch": 31.153711790393015,
+ "grad_norm": 0.02337106689810753,
+ "learning_rate": 0.0006,
+ "loss": 4.094712734222412,
+ "step": 2243
+ },
+ {
+ "epoch": 31.16768558951965,
+ "grad_norm": 0.02219875156879425,
+ "learning_rate": 0.0006,
+ "loss": 4.1800994873046875,
+ "step": 2244
+ },
+ {
+ "epoch": 31.18165938864629,
+ "grad_norm": 0.022680504247546196,
+ "learning_rate": 0.0006,
+ "loss": 4.2054362297058105,
+ "step": 2245
+ },
+ {
+ "epoch": 31.195633187772927,
+ "grad_norm": 0.0208734143525362,
+ "learning_rate": 0.0006,
+ "loss": 4.06068229675293,
+ "step": 2246
+ },
+ {
+ "epoch": 31.209606986899562,
+ "grad_norm": 0.01811066083610058,
+ "learning_rate": 0.0006,
+ "loss": 4.2465128898620605,
+ "step": 2247
+ },
+ {
+ "epoch": 31.2235807860262,
+ "grad_norm": 0.018571637570858,
+ "learning_rate": 0.0006,
+ "loss": 4.191955089569092,
+ "step": 2248
+ },
+ {
+ "epoch": 31.23755458515284,
+ "grad_norm": 0.018598807975649834,
+ "learning_rate": 0.0006,
+ "loss": 4.200872421264648,
+ "step": 2249
+ },
+ {
+ "epoch": 31.251528384279474,
+ "grad_norm": 0.017577625811100006,
+ "learning_rate": 0.0006,
+ "loss": 4.148700714111328,
+ "step": 2250
+ },
+ {
+ "epoch": 31.265502183406113,
+ "grad_norm": 0.01817980408668518,
+ "learning_rate": 0.0006,
+ "loss": 4.256978988647461,
+ "step": 2251
+ },
+ {
+ "epoch": 31.27947598253275,
+ "grad_norm": 0.017983486875891685,
+ "learning_rate": 0.0006,
+ "loss": 4.175005912780762,
+ "step": 2252
+ },
+ {
+ "epoch": 31.29344978165939,
+ "grad_norm": 0.016751961782574654,
+ "learning_rate": 0.0006,
+ "loss": 4.237518310546875,
+ "step": 2253
+ },
+ {
+ "epoch": 31.307423580786025,
+ "grad_norm": 0.016407202929258347,
+ "learning_rate": 0.0006,
+ "loss": 4.197269439697266,
+ "step": 2254
+ },
+ {
+ "epoch": 31.321397379912664,
+ "grad_norm": 0.01578645594418049,
+ "learning_rate": 0.0006,
+ "loss": 4.179751396179199,
+ "step": 2255
+ },
+ {
+ "epoch": 31.335371179039303,
+ "grad_norm": 0.016001226380467415,
+ "learning_rate": 0.0006,
+ "loss": 4.08821964263916,
+ "step": 2256
+ },
+ {
+ "epoch": 31.349344978165938,
+ "grad_norm": 0.015122811309993267,
+ "learning_rate": 0.0006,
+ "loss": 4.047828674316406,
+ "step": 2257
+ },
+ {
+ "epoch": 31.363318777292577,
+ "grad_norm": 0.015399965457618237,
+ "learning_rate": 0.0006,
+ "loss": 4.1506524085998535,
+ "step": 2258
+ },
+ {
+ "epoch": 31.377292576419215,
+ "grad_norm": 0.01534605398774147,
+ "learning_rate": 0.0006,
+ "loss": 4.27952766418457,
+ "step": 2259
+ },
+ {
+ "epoch": 31.39126637554585,
+ "grad_norm": 0.015461381524801254,
+ "learning_rate": 0.0006,
+ "loss": 4.27102518081665,
+ "step": 2260
+ },
+ {
+ "epoch": 31.40524017467249,
+ "grad_norm": 0.016769153997302055,
+ "learning_rate": 0.0006,
+ "loss": 4.191592216491699,
+ "step": 2261
+ },
+ {
+ "epoch": 31.419213973799128,
+ "grad_norm": 0.016557959839701653,
+ "learning_rate": 0.0006,
+ "loss": 4.159705638885498,
+ "step": 2262
+ },
+ {
+ "epoch": 31.433187772925763,
+ "grad_norm": 0.01659569889307022,
+ "learning_rate": 0.0006,
+ "loss": 4.2136664390563965,
+ "step": 2263
+ },
+ {
+ "epoch": 31.4471615720524,
+ "grad_norm": 0.017101937904953957,
+ "learning_rate": 0.0006,
+ "loss": 4.2716498374938965,
+ "step": 2264
+ },
+ {
+ "epoch": 31.46113537117904,
+ "grad_norm": 0.01722985878586769,
+ "learning_rate": 0.0006,
+ "loss": 4.300534248352051,
+ "step": 2265
+ },
+ {
+ "epoch": 31.475109170305675,
+ "grad_norm": 0.016800185665488243,
+ "learning_rate": 0.0006,
+ "loss": 4.113240718841553,
+ "step": 2266
+ },
+ {
+ "epoch": 31.489082969432314,
+ "grad_norm": 0.01603875495493412,
+ "learning_rate": 0.0006,
+ "loss": 4.157105445861816,
+ "step": 2267
+ },
+ {
+ "epoch": 31.503056768558952,
+ "grad_norm": 0.015903932973742485,
+ "learning_rate": 0.0006,
+ "loss": 4.233669757843018,
+ "step": 2268
+ },
+ {
+ "epoch": 31.51703056768559,
+ "grad_norm": 0.0160983856767416,
+ "learning_rate": 0.0006,
+ "loss": 4.147704124450684,
+ "step": 2269
+ },
+ {
+ "epoch": 31.531004366812226,
+ "grad_norm": 0.01604730263352394,
+ "learning_rate": 0.0006,
+ "loss": 4.1465911865234375,
+ "step": 2270
+ },
+ {
+ "epoch": 31.544978165938865,
+ "grad_norm": 0.016205785796046257,
+ "learning_rate": 0.0006,
+ "loss": 4.106704235076904,
+ "step": 2271
+ },
+ {
+ "epoch": 31.558951965065503,
+ "grad_norm": 0.017008036375045776,
+ "learning_rate": 0.0006,
+ "loss": 4.191000938415527,
+ "step": 2272
+ },
+ {
+ "epoch": 31.57292576419214,
+ "grad_norm": 0.01699635200202465,
+ "learning_rate": 0.0006,
+ "loss": 4.187042236328125,
+ "step": 2273
+ },
+ {
+ "epoch": 31.586899563318777,
+ "grad_norm": 0.01692967116832733,
+ "learning_rate": 0.0006,
+ "loss": 4.208805561065674,
+ "step": 2274
+ },
+ {
+ "epoch": 31.600873362445416,
+ "grad_norm": 0.017608756199479103,
+ "learning_rate": 0.0006,
+ "loss": 4.200512886047363,
+ "step": 2275
+ },
+ {
+ "epoch": 31.61484716157205,
+ "grad_norm": 0.016590990126132965,
+ "learning_rate": 0.0006,
+ "loss": 4.194705009460449,
+ "step": 2276
+ },
+ {
+ "epoch": 31.62882096069869,
+ "grad_norm": 0.0170235987752676,
+ "learning_rate": 0.0006,
+ "loss": 4.210326194763184,
+ "step": 2277
+ },
+ {
+ "epoch": 31.64279475982533,
+ "grad_norm": 0.01604801043868065,
+ "learning_rate": 0.0006,
+ "loss": 4.196024417877197,
+ "step": 2278
+ },
+ {
+ "epoch": 31.656768558951963,
+ "grad_norm": 0.015694202855229378,
+ "learning_rate": 0.0006,
+ "loss": 4.178236484527588,
+ "step": 2279
+ },
+ {
+ "epoch": 31.670742358078602,
+ "grad_norm": 0.015473959036171436,
+ "learning_rate": 0.0006,
+ "loss": 4.259339809417725,
+ "step": 2280
+ },
+ {
+ "epoch": 31.68471615720524,
+ "grad_norm": 0.014689362607896328,
+ "learning_rate": 0.0006,
+ "loss": 3.9702324867248535,
+ "step": 2281
+ },
+ {
+ "epoch": 31.69868995633188,
+ "grad_norm": 0.015245123766362667,
+ "learning_rate": 0.0006,
+ "loss": 4.197025775909424,
+ "step": 2282
+ },
+ {
+ "epoch": 31.712663755458514,
+ "grad_norm": 0.016148347407579422,
+ "learning_rate": 0.0006,
+ "loss": 4.231856346130371,
+ "step": 2283
+ },
+ {
+ "epoch": 31.726637554585153,
+ "grad_norm": 0.015169818885624409,
+ "learning_rate": 0.0006,
+ "loss": 4.230169296264648,
+ "step": 2284
+ },
+ {
+ "epoch": 31.74061135371179,
+ "grad_norm": 0.014843013137578964,
+ "learning_rate": 0.0006,
+ "loss": 4.195633411407471,
+ "step": 2285
+ },
+ {
+ "epoch": 31.754585152838427,
+ "grad_norm": 0.015229340642690659,
+ "learning_rate": 0.0006,
+ "loss": 4.231280326843262,
+ "step": 2286
+ },
+ {
+ "epoch": 31.768558951965066,
+ "grad_norm": 0.01551708485931158,
+ "learning_rate": 0.0006,
+ "loss": 4.17064094543457,
+ "step": 2287
+ },
+ {
+ "epoch": 31.782532751091704,
+ "grad_norm": 0.015863575041294098,
+ "learning_rate": 0.0006,
+ "loss": 4.1579484939575195,
+ "step": 2288
+ },
+ {
+ "epoch": 31.79650655021834,
+ "grad_norm": 0.015118198469281197,
+ "learning_rate": 0.0006,
+ "loss": 4.1626787185668945,
+ "step": 2289
+ },
+ {
+ "epoch": 31.810480349344978,
+ "grad_norm": 0.015220776200294495,
+ "learning_rate": 0.0006,
+ "loss": 4.135746955871582,
+ "step": 2290
+ },
+ {
+ "epoch": 31.824454148471617,
+ "grad_norm": 0.014529784210026264,
+ "learning_rate": 0.0006,
+ "loss": 4.164931297302246,
+ "step": 2291
+ },
+ {
+ "epoch": 31.83842794759825,
+ "grad_norm": 0.015533296391367912,
+ "learning_rate": 0.0006,
+ "loss": 4.112464904785156,
+ "step": 2292
+ },
+ {
+ "epoch": 31.85240174672489,
+ "grad_norm": 0.016562335193157196,
+ "learning_rate": 0.0006,
+ "loss": 4.164898872375488,
+ "step": 2293
+ },
+ {
+ "epoch": 31.86637554585153,
+ "grad_norm": 0.016258113086223602,
+ "learning_rate": 0.0006,
+ "loss": 4.1979265213012695,
+ "step": 2294
+ },
+ {
+ "epoch": 31.880349344978168,
+ "grad_norm": 0.01558110024780035,
+ "learning_rate": 0.0006,
+ "loss": 4.174675941467285,
+ "step": 2295
+ },
+ {
+ "epoch": 31.894323144104803,
+ "grad_norm": 0.016412515193223953,
+ "learning_rate": 0.0006,
+ "loss": 4.217909336090088,
+ "step": 2296
+ },
+ {
+ "epoch": 31.90829694323144,
+ "grad_norm": 0.015118442475795746,
+ "learning_rate": 0.0006,
+ "loss": 4.137143611907959,
+ "step": 2297
+ },
+ {
+ "epoch": 31.92227074235808,
+ "grad_norm": 0.01506161317229271,
+ "learning_rate": 0.0006,
+ "loss": 4.196122169494629,
+ "step": 2298
+ },
+ {
+ "epoch": 31.936244541484715,
+ "grad_norm": 0.017165271565318108,
+ "learning_rate": 0.0006,
+ "loss": 4.14834451675415,
+ "step": 2299
+ },
+ {
+ "epoch": 31.950218340611354,
+ "grad_norm": 0.01696440577507019,
+ "learning_rate": 0.0006,
+ "loss": 4.218868732452393,
+ "step": 2300
+ },
+ {
+ "epoch": 31.964192139737992,
+ "grad_norm": 0.017349394038319588,
+ "learning_rate": 0.0006,
+ "loss": 4.2754011154174805,
+ "step": 2301
+ },
+ {
+ "epoch": 31.978165938864628,
+ "grad_norm": 0.018387921154499054,
+ "learning_rate": 0.0006,
+ "loss": 4.282623291015625,
+ "step": 2302
+ },
+ {
+ "epoch": 31.992139737991266,
+ "grad_norm": 0.018837926909327507,
+ "learning_rate": 0.0006,
+ "loss": 4.185425281524658,
+ "step": 2303
+ },
+ {
+ "epoch": 32.0,
+ "grad_norm": 0.018998464569449425,
+ "learning_rate": 0.0006,
+ "loss": 4.151291847229004,
+ "step": 2304
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 28800,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 400,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": true,
+ "should_log": false,
+ "should_save": false,
+ "should_training_stop": false
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 9.822762756263117e+17,
+ "train_batch_size": 8,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2304/training_args.bin b/runs/l2r50-i2-fulle-lm/checkpoint-2304/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..cad232831c222cc7f8b938b8bbe42f61e7d89c3a
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-2304/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:f7c40e3bfa2f5ac1450ca1243e3f2a895d44a06ed5995e6aefd97f78d229d8c8
+size 4792
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2592/chat_template.jinja b/runs/l2r50-i2-fulle-lm/checkpoint-2592/chat_template.jinja
new file mode 100644
index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-2592/chat_template.jinja
@@ -0,0 +1,85 @@
+{%- if tools %}
+ {{- '<|im_start|>system\n' }}
+ {%- if messages[0].role == 'system' %}
+ {{- messages[0].content + '\n\n' }}
+ {%- endif %}
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }}
+ {%- for tool in tools %}
+ {{- "\n" }}
+ {{- tool | tojson }}
+ {%- endfor %}
+ {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }}
+{%- else %}
+ {%- if messages[0].role == 'system' %}
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
+ {%- endif %}
+{%- endif %}
+{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
+{%- for message in messages[::-1] %}
+ {%- set index = (messages|length - 1) - loop.index0 %}
+ {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %}
+ {%- set ns.multi_step_tool = false %}
+ {%- set ns.last_query_index = index %}
+ {%- endif %}
+{%- endfor %}
+{%- for message in messages %}
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
+ {%- elif message.role == "assistant" %}
+ {%- set content = message.content %}
+ {%- set reasoning_content = '' %}
+ {%- if message.reasoning_content is defined and message.reasoning_content is not none %}
+ {%- set reasoning_content = message.reasoning_content %}
+ {%- else %}
+ {%- if '' in message.content %}
+ {%- set content = message.content.split('')[-1].lstrip('\n') %}
+ {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %}
+ {%- endif %}
+ {%- endif %}
+ {%- if loop.index0 > ns.last_query_index %}
+ {%- if loop.last or (not loop.last and reasoning_content) %}
+ {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- if message.tool_calls %}
+ {%- for tool_call in message.tool_calls %}
+ {%- if (loop.first and content) or (not loop.first) %}
+ {{- '\n' }}
+ {%- endif %}
+ {%- if tool_call.function %}
+ {%- set tool_call = tool_call.function %}
+ {%- endif %}
+ {{- '\n{"name": "' }}
+ {{- tool_call.name }}
+ {{- '", "arguments": ' }}
+ {%- if tool_call.arguments is string %}
+ {{- tool_call.arguments }}
+ {%- else %}
+ {{- tool_call.arguments | tojson }}
+ {%- endif %}
+ {{- '}\n' }}
+ {%- endfor %}
+ {%- endif %}
+ {{- '<|im_end|>\n' }}
+ {%- elif message.role == "tool" %}
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
+ {{- '<|im_start|>user' }}
+ {%- endif %}
+ {{- '\n\n' }}
+ {{- message.content }}
+ {{- '\n' }}
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
+ {{- '<|im_end|>\n' }}
+ {%- endif %}
+ {%- endif %}
+{%- endfor %}
+{%- if add_generation_prompt %}
+ {{- '<|im_start|>assistant\n' }}
+ {%- if enable_thinking is defined and enable_thinking is false %}
+ {{- '\n\n\n\n' }}
+ {%- endif %}
+{%- endif %}
\ No newline at end of file
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2592/config.json b/runs/l2r50-i2-fulle-lm/checkpoint-2592/config.json
new file mode 100644
index 0000000000000000000000000000000000000000..22f250953c6ce8fc91ee4411856dd73041a415e2
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-2592/config.json
@@ -0,0 +1,32 @@
+{
+ "architectures": [
+ "LlamaForCausalLM"
+ ],
+ "attention_bias": false,
+ "attention_dropout": 0.0,
+ "bos_token_id": null,
+ "dtype": "float32",
+ "eos_token_id": 151645,
+ "head_dim": 128,
+ "hidden_act": "silu",
+ "hidden_size": 512,
+ "initializer_range": 0.02,
+ "intermediate_size": 1536,
+ "max_position_embeddings": 2048,
+ "mlp_bias": false,
+ "model_type": "llama",
+ "num_attention_heads": 4,
+ "num_hidden_layers": 20,
+ "num_key_value_heads": 4,
+ "pad_token_id": 151645,
+ "pretraining_tp": 1,
+ "rms_norm_eps": 1e-06,
+ "rope_parameters": {
+ "rope_theta": 10000.0,
+ "rope_type": "default"
+ },
+ "tie_word_embeddings": true,
+ "transformers_version": "5.5.0",
+ "use_cache": false,
+ "vocab_size": 151673
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2592/generation_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-2592/generation_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-2592/generation_config.json
@@ -0,0 +1,11 @@
+{
+ "_from_model_config": true,
+ "eos_token_id": [
+ 151645
+ ],
+ "output_attentions": false,
+ "output_hidden_states": false,
+ "pad_token_id": 151645,
+ "transformers_version": "5.5.0",
+ "use_cache": true
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2592/model.safetensors b/runs/l2r50-i2-fulle-lm/checkpoint-2592/model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..b8932b2095bc39e6c64d178cd9d435f89a28ee9b
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-2592/model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:159af0671f14991d93a0cf3de43bbb61f98764ddd20b8c4d113e6d53b75ea79e
+size 583360328
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2592/optimizer.pt b/runs/l2r50-i2-fulle-lm/checkpoint-2592/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..7b3be601378ad7adc03184c81d0c367d0fedfecf
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-2592/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:ee5657c939446cb739c72fd10af9dfc17bf998cb81acec1a67f046885c0e8798
+size 1166833530
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2592/rng_state_0.pth b/runs/l2r50-i2-fulle-lm/checkpoint-2592/rng_state_0.pth
new file mode 100644
index 0000000000000000000000000000000000000000..a0fdb39149ff58d0081c949f5c4ac75fad91bdea
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-2592/rng_state_0.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:b9fb29c856023c989256c2258f7638a00483c73b0e057199d87f49bde9adc001
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2592/rng_state_1.pth b/runs/l2r50-i2-fulle-lm/checkpoint-2592/rng_state_1.pth
new file mode 100644
index 0000000000000000000000000000000000000000..a4a189614e11e6247011118d86c4300b64bec057
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-2592/rng_state_1.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:8dcfc3024cdc624110f90d7ee29b0d17ef0cc4e4a51c6bb3887cfb1aa7d94b2b
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2592/rng_state_2.pth b/runs/l2r50-i2-fulle-lm/checkpoint-2592/rng_state_2.pth
new file mode 100644
index 0000000000000000000000000000000000000000..3ca337805739ff7a42b8bda27c5bc07f46adbdf8
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-2592/rng_state_2.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:f72e1db756ecb49e4b074f22485247cd548215c7452157ef8ff95dbe70728a87
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2592/rng_state_3.pth b/runs/l2r50-i2-fulle-lm/checkpoint-2592/rng_state_3.pth
new file mode 100644
index 0000000000000000000000000000000000000000..387d4ca2cead233b29667cf1b24c44a90092d396
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-2592/rng_state_3.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:6e2fea0eba78e390537679e1702fa655f12d93c6e64ab30b2ec5f73f5825e46b
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2592/scheduler.pt b/runs/l2r50-i2-fulle-lm/checkpoint-2592/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..f5637f631cc7ed2167f366c33ad859c0b7d97027
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-2592/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:0eee245b547fe91bdb4d6b4cbe52724d52118d751812b2781fcb358a63c34082
+size 1064
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2592/tokenizer.json b/runs/l2r50-i2-fulle-lm/checkpoint-2592/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..53786016f070e11184813f3c83101dc5e93e392e
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-2592/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:c3c63fda1646a555448e8f03662542e04b11d103721319e4d9b267d7da03ceb8
+size 11424463
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2592/tokenizer_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-2592/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..98e2fb9b2cc54785dc9d871b8e503f8b8f2defe7
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-2592/tokenizer_config.json
@@ -0,0 +1,20 @@
+{
+ "add_prefix_space": false,
+ "backend": "tokenizers",
+ "bos_token": null,
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "<|im_end|>",
+ "errors": "replace",
+ "extra_special_tokens": [
+ "<|l2r_pred|>",
+ "<|r2l_pred|>",
+ "<|next_1_pred|>",
+ "<|next_2_pred|>"
+ ],
+ "is_local": false,
+ "model_max_length": 131072,
+ "pad_token": "<|im_end|>",
+ "split_special_tokens": false,
+ "tokenizer_class": "Qwen2Tokenizer",
+ "unk_token": null
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2592/trainer_state.json b/runs/l2r50-i2-fulle-lm/checkpoint-2592/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..31e125ba45f1e965b57c37b69148d069d61b3643
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-2592/trainer_state.json
@@ -0,0 +1,18458 @@
+{
+ "best_global_step": null,
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 36.0,
+ "eval_steps": 500,
+ "global_step": 2592,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.013973799126637555,
+ "grad_norm": 0.1739506870508194,
+ "learning_rate": 0.0,
+ "loss": 11.99299430847168,
+ "step": 1
+ },
+ {
+ "epoch": 0.02794759825327511,
+ "grad_norm": 0.17216597497463226,
+ "learning_rate": 5.999999999999999e-06,
+ "loss": 11.988704681396484,
+ "step": 2
+ },
+ {
+ "epoch": 0.04192139737991266,
+ "grad_norm": 0.17459715902805328,
+ "learning_rate": 1.1999999999999999e-05,
+ "loss": 11.945581436157227,
+ "step": 3
+ },
+ {
+ "epoch": 0.05589519650655022,
+ "grad_norm": 0.17577414214611053,
+ "learning_rate": 1.7999999999999997e-05,
+ "loss": 11.859772682189941,
+ "step": 4
+ },
+ {
+ "epoch": 0.06986899563318777,
+ "grad_norm": 0.16163651645183563,
+ "learning_rate": 2.3999999999999997e-05,
+ "loss": 11.744415283203125,
+ "step": 5
+ },
+ {
+ "epoch": 0.08384279475982533,
+ "grad_norm": 0.14409932494163513,
+ "learning_rate": 2.9999999999999997e-05,
+ "loss": 11.618099212646484,
+ "step": 6
+ },
+ {
+ "epoch": 0.09781659388646288,
+ "grad_norm": 0.13079848885536194,
+ "learning_rate": 3.5999999999999994e-05,
+ "loss": 11.505935668945312,
+ "step": 7
+ },
+ {
+ "epoch": 0.11179039301310044,
+ "grad_norm": 0.11964613944292068,
+ "learning_rate": 4.2e-05,
+ "loss": 11.416532516479492,
+ "step": 8
+ },
+ {
+ "epoch": 0.125764192139738,
+ "grad_norm": 0.11345142871141434,
+ "learning_rate": 4.7999999999999994e-05,
+ "loss": 11.33206558227539,
+ "step": 9
+ },
+ {
+ "epoch": 0.13973799126637554,
+ "grad_norm": 0.11033952981233597,
+ "learning_rate": 5.399999999999999e-05,
+ "loss": 11.259052276611328,
+ "step": 10
+ },
+ {
+ "epoch": 0.1537117903930131,
+ "grad_norm": 0.10776489973068237,
+ "learning_rate": 5.9999999999999995e-05,
+ "loss": 11.203384399414062,
+ "step": 11
+ },
+ {
+ "epoch": 0.16768558951965065,
+ "grad_norm": 0.10721946507692337,
+ "learning_rate": 6.599999999999999e-05,
+ "loss": 11.153295516967773,
+ "step": 12
+ },
+ {
+ "epoch": 0.18165938864628822,
+ "grad_norm": 0.10673430562019348,
+ "learning_rate": 7.199999999999999e-05,
+ "loss": 11.110200881958008,
+ "step": 13
+ },
+ {
+ "epoch": 0.19563318777292577,
+ "grad_norm": 0.10630254447460175,
+ "learning_rate": 7.8e-05,
+ "loss": 11.067477226257324,
+ "step": 14
+ },
+ {
+ "epoch": 0.2096069868995633,
+ "grad_norm": 0.10631615668535233,
+ "learning_rate": 8.4e-05,
+ "loss": 11.022941589355469,
+ "step": 15
+ },
+ {
+ "epoch": 0.22358078602620088,
+ "grad_norm": 0.10609443485736847,
+ "learning_rate": 8.999999999999999e-05,
+ "loss": 10.970746040344238,
+ "step": 16
+ },
+ {
+ "epoch": 0.23755458515283842,
+ "grad_norm": 0.10648232698440552,
+ "learning_rate": 9.599999999999999e-05,
+ "loss": 10.91267204284668,
+ "step": 17
+ },
+ {
+ "epoch": 0.251528384279476,
+ "grad_norm": 0.10584589838981628,
+ "learning_rate": 0.000102,
+ "loss": 10.854547500610352,
+ "step": 18
+ },
+ {
+ "epoch": 0.26550218340611353,
+ "grad_norm": 0.10591720044612885,
+ "learning_rate": 0.00010799999999999998,
+ "loss": 10.788917541503906,
+ "step": 19
+ },
+ {
+ "epoch": 0.2794759825327511,
+ "grad_norm": 0.10611672699451447,
+ "learning_rate": 0.00011399999999999999,
+ "loss": 10.717556953430176,
+ "step": 20
+ },
+ {
+ "epoch": 0.2934497816593886,
+ "grad_norm": 0.10455359518527985,
+ "learning_rate": 0.00011999999999999999,
+ "loss": 10.650186538696289,
+ "step": 21
+ },
+ {
+ "epoch": 0.3074235807860262,
+ "grad_norm": 0.1047525629401207,
+ "learning_rate": 0.00012599999999999997,
+ "loss": 10.56808090209961,
+ "step": 22
+ },
+ {
+ "epoch": 0.32139737991266376,
+ "grad_norm": 0.10511605441570282,
+ "learning_rate": 0.00013199999999999998,
+ "loss": 10.484028816223145,
+ "step": 23
+ },
+ {
+ "epoch": 0.3353711790393013,
+ "grad_norm": 0.10467177629470825,
+ "learning_rate": 0.000138,
+ "loss": 10.399947166442871,
+ "step": 24
+ },
+ {
+ "epoch": 0.34934497816593885,
+ "grad_norm": 0.10432136058807373,
+ "learning_rate": 0.00014399999999999998,
+ "loss": 10.312714576721191,
+ "step": 25
+ },
+ {
+ "epoch": 0.36331877729257644,
+ "grad_norm": 0.10527262091636658,
+ "learning_rate": 0.00015,
+ "loss": 10.212261199951172,
+ "step": 26
+ },
+ {
+ "epoch": 0.377292576419214,
+ "grad_norm": 0.103610560297966,
+ "learning_rate": 0.000156,
+ "loss": 10.132034301757812,
+ "step": 27
+ },
+ {
+ "epoch": 0.39126637554585153,
+ "grad_norm": 0.10553059726953506,
+ "learning_rate": 0.000162,
+ "loss": 10.020434379577637,
+ "step": 28
+ },
+ {
+ "epoch": 0.4052401746724891,
+ "grad_norm": 0.1037299782037735,
+ "learning_rate": 0.000168,
+ "loss": 9.934466361999512,
+ "step": 29
+ },
+ {
+ "epoch": 0.4192139737991266,
+ "grad_norm": 0.1042739674448967,
+ "learning_rate": 0.00017399999999999997,
+ "loss": 9.822539329528809,
+ "step": 30
+ },
+ {
+ "epoch": 0.4331877729257642,
+ "grad_norm": 0.10300865024328232,
+ "learning_rate": 0.00017999999999999998,
+ "loss": 9.734994888305664,
+ "step": 31
+ },
+ {
+ "epoch": 0.44716157205240176,
+ "grad_norm": 0.10291654616594315,
+ "learning_rate": 0.000186,
+ "loss": 9.624505996704102,
+ "step": 32
+ },
+ {
+ "epoch": 0.4611353711790393,
+ "grad_norm": 0.10210851579904556,
+ "learning_rate": 0.00019199999999999998,
+ "loss": 9.518205642700195,
+ "step": 33
+ },
+ {
+ "epoch": 0.47510917030567684,
+ "grad_norm": 0.10222529619932175,
+ "learning_rate": 0.000198,
+ "loss": 9.412222862243652,
+ "step": 34
+ },
+ {
+ "epoch": 0.4890829694323144,
+ "grad_norm": 0.09992019087076187,
+ "learning_rate": 0.000204,
+ "loss": 9.326566696166992,
+ "step": 35
+ },
+ {
+ "epoch": 0.503056768558952,
+ "grad_norm": 0.09931084513664246,
+ "learning_rate": 0.00020999999999999998,
+ "loss": 9.210111618041992,
+ "step": 36
+ },
+ {
+ "epoch": 0.5170305676855895,
+ "grad_norm": 0.09928256273269653,
+ "learning_rate": 0.00021599999999999996,
+ "loss": 9.100493431091309,
+ "step": 37
+ },
+ {
+ "epoch": 0.5310043668122271,
+ "grad_norm": 0.09843816608190536,
+ "learning_rate": 0.00022199999999999998,
+ "loss": 8.996543884277344,
+ "step": 38
+ },
+ {
+ "epoch": 0.5449781659388646,
+ "grad_norm": 0.09825374186038971,
+ "learning_rate": 0.00022799999999999999,
+ "loss": 8.888991355895996,
+ "step": 39
+ },
+ {
+ "epoch": 0.5589519650655022,
+ "grad_norm": 0.09602385759353638,
+ "learning_rate": 0.000234,
+ "loss": 8.802907943725586,
+ "step": 40
+ },
+ {
+ "epoch": 0.5729257641921397,
+ "grad_norm": 0.09414064139127731,
+ "learning_rate": 0.00023999999999999998,
+ "loss": 8.715423583984375,
+ "step": 41
+ },
+ {
+ "epoch": 0.5868995633187772,
+ "grad_norm": 0.09273893386125565,
+ "learning_rate": 0.00024599999999999996,
+ "loss": 8.61374282836914,
+ "step": 42
+ },
+ {
+ "epoch": 0.6008733624454149,
+ "grad_norm": 0.09079599380493164,
+ "learning_rate": 0.00025199999999999995,
+ "loss": 8.502904891967773,
+ "step": 43
+ },
+ {
+ "epoch": 0.6148471615720524,
+ "grad_norm": 0.08605825901031494,
+ "learning_rate": 0.000258,
+ "loss": 8.453008651733398,
+ "step": 44
+ },
+ {
+ "epoch": 0.62882096069869,
+ "grad_norm": 0.08338996022939682,
+ "learning_rate": 0.00026399999999999997,
+ "loss": 8.365163803100586,
+ "step": 45
+ },
+ {
+ "epoch": 0.6427947598253275,
+ "grad_norm": 0.08150867372751236,
+ "learning_rate": 0.00027,
+ "loss": 8.285372734069824,
+ "step": 46
+ },
+ {
+ "epoch": 0.6567685589519651,
+ "grad_norm": 0.07794807851314545,
+ "learning_rate": 0.000276,
+ "loss": 8.185811996459961,
+ "step": 47
+ },
+ {
+ "epoch": 0.6707423580786026,
+ "grad_norm": 0.07592720538377762,
+ "learning_rate": 0.00028199999999999997,
+ "loss": 8.098482131958008,
+ "step": 48
+ },
+ {
+ "epoch": 0.6847161572052402,
+ "grad_norm": 0.0709897130727768,
+ "learning_rate": 0.00028799999999999995,
+ "loss": 8.042091369628906,
+ "step": 49
+ },
+ {
+ "epoch": 0.6986899563318777,
+ "grad_norm": 0.06582305580377579,
+ "learning_rate": 0.000294,
+ "loss": 7.972844123840332,
+ "step": 50
+ },
+ {
+ "epoch": 0.7126637554585152,
+ "grad_norm": 0.059452351182699203,
+ "learning_rate": 0.0003,
+ "loss": 7.924579620361328,
+ "step": 51
+ },
+ {
+ "epoch": 0.7266375545851529,
+ "grad_norm": 0.052426718175411224,
+ "learning_rate": 0.00030599999999999996,
+ "loss": 7.876336097717285,
+ "step": 52
+ },
+ {
+ "epoch": 0.7406113537117904,
+ "grad_norm": 0.04768150672316551,
+ "learning_rate": 0.000312,
+ "loss": 7.8102707862854,
+ "step": 53
+ },
+ {
+ "epoch": 0.754585152838428,
+ "grad_norm": 0.03970075398683548,
+ "learning_rate": 0.000318,
+ "loss": 7.791913986206055,
+ "step": 54
+ },
+ {
+ "epoch": 0.7685589519650655,
+ "grad_norm": 0.03559752553701401,
+ "learning_rate": 0.000324,
+ "loss": 7.756960868835449,
+ "step": 55
+ },
+ {
+ "epoch": 0.7825327510917031,
+ "grad_norm": 0.029321353882551193,
+ "learning_rate": 0.00033,
+ "loss": 7.724878787994385,
+ "step": 56
+ },
+ {
+ "epoch": 0.7965065502183406,
+ "grad_norm": 0.023485427722334862,
+ "learning_rate": 0.000336,
+ "loss": 7.72553014755249,
+ "step": 57
+ },
+ {
+ "epoch": 0.8104803493449781,
+ "grad_norm": 0.019319474697113037,
+ "learning_rate": 0.00034199999999999996,
+ "loss": 7.709493160247803,
+ "step": 58
+ },
+ {
+ "epoch": 0.8244541484716157,
+ "grad_norm": 0.016705691814422607,
+ "learning_rate": 0.00034799999999999995,
+ "loss": 7.6873297691345215,
+ "step": 59
+ },
+ {
+ "epoch": 0.8384279475982532,
+ "grad_norm": 0.017667734995484352,
+ "learning_rate": 0.00035399999999999993,
+ "loss": 7.677059173583984,
+ "step": 60
+ },
+ {
+ "epoch": 0.8524017467248908,
+ "grad_norm": 0.01671992614865303,
+ "learning_rate": 0.00035999999999999997,
+ "loss": 7.643512725830078,
+ "step": 61
+ },
+ {
+ "epoch": 0.8663755458515284,
+ "grad_norm": 0.012288172729313374,
+ "learning_rate": 0.00036599999999999995,
+ "loss": 7.658343315124512,
+ "step": 62
+ },
+ {
+ "epoch": 0.880349344978166,
+ "grad_norm": 0.012004776857793331,
+ "learning_rate": 0.000372,
+ "loss": 7.649759292602539,
+ "step": 63
+ },
+ {
+ "epoch": 0.8943231441048035,
+ "grad_norm": 0.012487445957958698,
+ "learning_rate": 0.00037799999999999997,
+ "loss": 7.631503105163574,
+ "step": 64
+ },
+ {
+ "epoch": 0.9082969432314411,
+ "grad_norm": 0.01489241048693657,
+ "learning_rate": 0.00038399999999999996,
+ "loss": 7.647110462188721,
+ "step": 65
+ },
+ {
+ "epoch": 0.9222707423580786,
+ "grad_norm": 0.01428980939090252,
+ "learning_rate": 0.00039,
+ "loss": 7.636983871459961,
+ "step": 66
+ },
+ {
+ "epoch": 0.9362445414847161,
+ "grad_norm": 0.010711144655942917,
+ "learning_rate": 0.000396,
+ "loss": 7.621130466461182,
+ "step": 67
+ },
+ {
+ "epoch": 0.9502183406113537,
+ "grad_norm": 0.010178886353969574,
+ "learning_rate": 0.000402,
+ "loss": 7.593924522399902,
+ "step": 68
+ },
+ {
+ "epoch": 0.9641921397379912,
+ "grad_norm": 0.010604874230921268,
+ "learning_rate": 0.000408,
+ "loss": 7.6015496253967285,
+ "step": 69
+ },
+ {
+ "epoch": 0.9781659388646288,
+ "grad_norm": 0.009751426056027412,
+ "learning_rate": 0.0004139999999999999,
+ "loss": 7.571964263916016,
+ "step": 70
+ },
+ {
+ "epoch": 0.9921397379912664,
+ "grad_norm": 0.009816481731832027,
+ "learning_rate": 0.00041999999999999996,
+ "loss": 7.57344913482666,
+ "step": 71
+ },
+ {
+ "epoch": 1.0,
+ "grad_norm": 0.008665637113153934,
+ "learning_rate": 0.00042599999999999995,
+ "loss": 7.579050064086914,
+ "step": 72
+ },
+ {
+ "epoch": 1.0,
+ "eval_loss": 7.581204891204834,
+ "eval_runtime": 59.0091,
+ "eval_samples_per_second": 41.383,
+ "eval_steps_per_second": 1.305,
+ "step": 72
+ },
+ {
+ "epoch": 1.0139737991266375,
+ "grad_norm": 0.009684201329946518,
+ "learning_rate": 0.00043199999999999993,
+ "loss": 7.568594932556152,
+ "step": 73
+ },
+ {
+ "epoch": 1.027947598253275,
+ "grad_norm": 0.025783710181713104,
+ "learning_rate": 0.00043799999999999997,
+ "loss": 7.578948497772217,
+ "step": 74
+ },
+ {
+ "epoch": 1.0419213973799126,
+ "grad_norm": 0.0327819399535656,
+ "learning_rate": 0.00044399999999999995,
+ "loss": 7.548173904418945,
+ "step": 75
+ },
+ {
+ "epoch": 1.0558951965065502,
+ "grad_norm": 0.021604498848319054,
+ "learning_rate": 0.00045,
+ "loss": 7.552953720092773,
+ "step": 76
+ },
+ {
+ "epoch": 1.0698689956331877,
+ "grad_norm": 0.010230105370283127,
+ "learning_rate": 0.00045599999999999997,
+ "loss": 7.5213212966918945,
+ "step": 77
+ },
+ {
+ "epoch": 1.0838427947598253,
+ "grad_norm": 0.023315589874982834,
+ "learning_rate": 0.00046199999999999995,
+ "loss": 7.501640319824219,
+ "step": 78
+ },
+ {
+ "epoch": 1.0978165938864628,
+ "grad_norm": 0.01109944935888052,
+ "learning_rate": 0.000468,
+ "loss": 7.539519786834717,
+ "step": 79
+ },
+ {
+ "epoch": 1.1117903930131003,
+ "grad_norm": 0.026093758642673492,
+ "learning_rate": 0.000474,
+ "loss": 7.525928974151611,
+ "step": 80
+ },
+ {
+ "epoch": 1.125764192139738,
+ "grad_norm": 0.03274895250797272,
+ "learning_rate": 0.00047999999999999996,
+ "loss": 7.535602569580078,
+ "step": 81
+ },
+ {
+ "epoch": 1.1397379912663754,
+ "grad_norm": 0.01543602254241705,
+ "learning_rate": 0.000486,
+ "loss": 7.513683795928955,
+ "step": 82
+ },
+ {
+ "epoch": 1.1537117903930132,
+ "grad_norm": 0.028377249836921692,
+ "learning_rate": 0.0004919999999999999,
+ "loss": 7.49384880065918,
+ "step": 83
+ },
+ {
+ "epoch": 1.1676855895196507,
+ "grad_norm": 0.04121481627225876,
+ "learning_rate": 0.000498,
+ "loss": 7.488447189331055,
+ "step": 84
+ },
+ {
+ "epoch": 1.1816593886462883,
+ "grad_norm": 0.013282302767038345,
+ "learning_rate": 0.0005039999999999999,
+ "loss": 7.463010787963867,
+ "step": 85
+ },
+ {
+ "epoch": 1.1956331877729258,
+ "grad_norm": 0.03200364485383034,
+ "learning_rate": 0.0005099999999999999,
+ "loss": 7.473515033721924,
+ "step": 86
+ },
+ {
+ "epoch": 1.2096069868995634,
+ "grad_norm": 0.020768029615283012,
+ "learning_rate": 0.000516,
+ "loss": 7.464688777923584,
+ "step": 87
+ },
+ {
+ "epoch": 1.223580786026201,
+ "grad_norm": 0.021715788170695305,
+ "learning_rate": 0.000522,
+ "loss": 7.4835028648376465,
+ "step": 88
+ },
+ {
+ "epoch": 1.2375545851528384,
+ "grad_norm": 0.026255754753947258,
+ "learning_rate": 0.0005279999999999999,
+ "loss": 7.443392753601074,
+ "step": 89
+ },
+ {
+ "epoch": 1.251528384279476,
+ "grad_norm": 0.022325951606035233,
+ "learning_rate": 0.000534,
+ "loss": 7.457373142242432,
+ "step": 90
+ },
+ {
+ "epoch": 1.2655021834061135,
+ "grad_norm": 0.009270387701690197,
+ "learning_rate": 0.00054,
+ "loss": 7.455986976623535,
+ "step": 91
+ },
+ {
+ "epoch": 1.279475982532751,
+ "grad_norm": 0.02369544468820095,
+ "learning_rate": 0.0005459999999999999,
+ "loss": 7.442216873168945,
+ "step": 92
+ },
+ {
+ "epoch": 1.2934497816593886,
+ "grad_norm": 0.008474506437778473,
+ "learning_rate": 0.000552,
+ "loss": 7.4294915199279785,
+ "step": 93
+ },
+ {
+ "epoch": 1.3074235807860262,
+ "grad_norm": 0.017043249681591988,
+ "learning_rate": 0.000558,
+ "loss": 7.406932830810547,
+ "step": 94
+ },
+ {
+ "epoch": 1.3213973799126637,
+ "grad_norm": 0.011813412420451641,
+ "learning_rate": 0.0005639999999999999,
+ "loss": 7.415900230407715,
+ "step": 95
+ },
+ {
+ "epoch": 1.3353711790393012,
+ "grad_norm": 0.01086883433163166,
+ "learning_rate": 0.00057,
+ "loss": 7.452998638153076,
+ "step": 96
+ },
+ {
+ "epoch": 1.3493449781659388,
+ "grad_norm": 0.014700951054692268,
+ "learning_rate": 0.0005759999999999999,
+ "loss": 7.410249710083008,
+ "step": 97
+ },
+ {
+ "epoch": 1.3633187772925766,
+ "grad_norm": 0.014512976631522179,
+ "learning_rate": 0.0005819999999999999,
+ "loss": 7.429132461547852,
+ "step": 98
+ },
+ {
+ "epoch": 1.3772925764192139,
+ "grad_norm": 0.012519920244812965,
+ "learning_rate": 0.000588,
+ "loss": 7.430817604064941,
+ "step": 99
+ },
+ {
+ "epoch": 1.3912663755458516,
+ "grad_norm": 0.007132664322853088,
+ "learning_rate": 0.0005939999999999999,
+ "loss": 7.405234336853027,
+ "step": 100
+ },
+ {
+ "epoch": 1.405240174672489,
+ "grad_norm": 0.019489388912916183,
+ "learning_rate": 0.0006,
+ "loss": 7.427791595458984,
+ "step": 101
+ },
+ {
+ "epoch": 1.4192139737991267,
+ "grad_norm": 0.027252664789557457,
+ "learning_rate": 0.0006,
+ "loss": 7.41761589050293,
+ "step": 102
+ },
+ {
+ "epoch": 1.4331877729257643,
+ "grad_norm": 0.02501637488603592,
+ "learning_rate": 0.0006,
+ "loss": 7.416735649108887,
+ "step": 103
+ },
+ {
+ "epoch": 1.4471615720524018,
+ "grad_norm": 0.00989772193133831,
+ "learning_rate": 0.0006,
+ "loss": 7.3756632804870605,
+ "step": 104
+ },
+ {
+ "epoch": 1.4611353711790394,
+ "grad_norm": 0.01622949168086052,
+ "learning_rate": 0.0006,
+ "loss": 7.400811672210693,
+ "step": 105
+ },
+ {
+ "epoch": 1.475109170305677,
+ "grad_norm": 0.012407763861119747,
+ "learning_rate": 0.0006,
+ "loss": 7.37675666809082,
+ "step": 106
+ },
+ {
+ "epoch": 1.4890829694323144,
+ "grad_norm": 0.010353565216064453,
+ "learning_rate": 0.0006,
+ "loss": 7.4013142585754395,
+ "step": 107
+ },
+ {
+ "epoch": 1.503056768558952,
+ "grad_norm": 0.0120553532615304,
+ "learning_rate": 0.0006,
+ "loss": 7.379873275756836,
+ "step": 108
+ },
+ {
+ "epoch": 1.5170305676855895,
+ "grad_norm": 0.01610225811600685,
+ "learning_rate": 0.0006,
+ "loss": 7.371631145477295,
+ "step": 109
+ },
+ {
+ "epoch": 1.531004366812227,
+ "grad_norm": 0.01159879844635725,
+ "learning_rate": 0.0006,
+ "loss": 7.389007568359375,
+ "step": 110
+ },
+ {
+ "epoch": 1.5449781659388646,
+ "grad_norm": 0.008013816550374031,
+ "learning_rate": 0.0006,
+ "loss": 7.395920753479004,
+ "step": 111
+ },
+ {
+ "epoch": 1.5589519650655022,
+ "grad_norm": 0.016334809362888336,
+ "learning_rate": 0.0006,
+ "loss": 7.393121719360352,
+ "step": 112
+ },
+ {
+ "epoch": 1.5729257641921397,
+ "grad_norm": 0.018821721896529198,
+ "learning_rate": 0.0006,
+ "loss": 7.367543697357178,
+ "step": 113
+ },
+ {
+ "epoch": 1.5868995633187772,
+ "grad_norm": 0.014088994823396206,
+ "learning_rate": 0.0006,
+ "loss": 7.385537624359131,
+ "step": 114
+ },
+ {
+ "epoch": 1.600873362445415,
+ "grad_norm": 0.005864050704985857,
+ "learning_rate": 0.0006,
+ "loss": 7.365346908569336,
+ "step": 115
+ },
+ {
+ "epoch": 1.6148471615720523,
+ "grad_norm": 0.011745430529117584,
+ "learning_rate": 0.0006,
+ "loss": 7.390227317810059,
+ "step": 116
+ },
+ {
+ "epoch": 1.62882096069869,
+ "grad_norm": 0.017205573618412018,
+ "learning_rate": 0.0006,
+ "loss": 7.3558549880981445,
+ "step": 117
+ },
+ {
+ "epoch": 1.6427947598253274,
+ "grad_norm": 0.021256346255540848,
+ "learning_rate": 0.0006,
+ "loss": 7.377346992492676,
+ "step": 118
+ },
+ {
+ "epoch": 1.6567685589519652,
+ "grad_norm": 0.019335204735398293,
+ "learning_rate": 0.0006,
+ "loss": 7.346194267272949,
+ "step": 119
+ },
+ {
+ "epoch": 1.6707423580786025,
+ "grad_norm": 0.008134003728628159,
+ "learning_rate": 0.0006,
+ "loss": 7.3063578605651855,
+ "step": 120
+ },
+ {
+ "epoch": 1.6847161572052403,
+ "grad_norm": 0.017613831907510757,
+ "learning_rate": 0.0006,
+ "loss": 7.338171005249023,
+ "step": 121
+ },
+ {
+ "epoch": 1.6986899563318776,
+ "grad_norm": 0.00890734139829874,
+ "learning_rate": 0.0006,
+ "loss": 7.355384826660156,
+ "step": 122
+ },
+ {
+ "epoch": 1.7126637554585153,
+ "grad_norm": 0.013448155485093594,
+ "learning_rate": 0.0006,
+ "loss": 7.344296455383301,
+ "step": 123
+ },
+ {
+ "epoch": 1.726637554585153,
+ "grad_norm": 0.009241633117198944,
+ "learning_rate": 0.0006,
+ "loss": 7.3454060554504395,
+ "step": 124
+ },
+ {
+ "epoch": 1.7406113537117904,
+ "grad_norm": 0.018561311066150665,
+ "learning_rate": 0.0006,
+ "loss": 7.32348108291626,
+ "step": 125
+ },
+ {
+ "epoch": 1.754585152838428,
+ "grad_norm": 0.025667704641819,
+ "learning_rate": 0.0006,
+ "loss": 7.356142520904541,
+ "step": 126
+ },
+ {
+ "epoch": 1.7685589519650655,
+ "grad_norm": 0.036232151091098785,
+ "learning_rate": 0.0006,
+ "loss": 7.363539695739746,
+ "step": 127
+ },
+ {
+ "epoch": 1.782532751091703,
+ "grad_norm": 0.059483714401721954,
+ "learning_rate": 0.0006,
+ "loss": 7.402331352233887,
+ "step": 128
+ },
+ {
+ "epoch": 1.7965065502183406,
+ "grad_norm": 0.05365009233355522,
+ "learning_rate": 0.0006,
+ "loss": 7.40457820892334,
+ "step": 129
+ },
+ {
+ "epoch": 1.8104803493449781,
+ "grad_norm": 0.02738415263593197,
+ "learning_rate": 0.0006,
+ "loss": 7.3756208419799805,
+ "step": 130
+ },
+ {
+ "epoch": 1.8244541484716157,
+ "grad_norm": 0.035017091780900955,
+ "learning_rate": 0.0006,
+ "loss": 7.379263877868652,
+ "step": 131
+ },
+ {
+ "epoch": 1.8384279475982532,
+ "grad_norm": 0.02819863148033619,
+ "learning_rate": 0.0006,
+ "loss": 7.327673435211182,
+ "step": 132
+ },
+ {
+ "epoch": 1.8524017467248908,
+ "grad_norm": 0.017754238098859787,
+ "learning_rate": 0.0006,
+ "loss": 7.333420276641846,
+ "step": 133
+ },
+ {
+ "epoch": 1.8663755458515285,
+ "grad_norm": 0.024685733020305634,
+ "learning_rate": 0.0006,
+ "loss": 7.34493350982666,
+ "step": 134
+ },
+ {
+ "epoch": 1.8803493449781659,
+ "grad_norm": 0.019426532089710236,
+ "learning_rate": 0.0006,
+ "loss": 7.328448295593262,
+ "step": 135
+ },
+ {
+ "epoch": 1.8943231441048036,
+ "grad_norm": 0.02106391079723835,
+ "learning_rate": 0.0006,
+ "loss": 7.327917098999023,
+ "step": 136
+ },
+ {
+ "epoch": 1.908296943231441,
+ "grad_norm": 0.02056879736483097,
+ "learning_rate": 0.0006,
+ "loss": 7.3311991691589355,
+ "step": 137
+ },
+ {
+ "epoch": 1.9222707423580787,
+ "grad_norm": 0.016865141689777374,
+ "learning_rate": 0.0006,
+ "loss": 7.312766075134277,
+ "step": 138
+ },
+ {
+ "epoch": 1.936244541484716,
+ "grad_norm": 0.017549855634570122,
+ "learning_rate": 0.0006,
+ "loss": 7.31191349029541,
+ "step": 139
+ },
+ {
+ "epoch": 1.9502183406113538,
+ "grad_norm": 0.015930157154798508,
+ "learning_rate": 0.0006,
+ "loss": 7.330901622772217,
+ "step": 140
+ },
+ {
+ "epoch": 1.9641921397379911,
+ "grad_norm": 0.013283558189868927,
+ "learning_rate": 0.0006,
+ "loss": 7.298318386077881,
+ "step": 141
+ },
+ {
+ "epoch": 1.9781659388646289,
+ "grad_norm": 0.012023469433188438,
+ "learning_rate": 0.0006,
+ "loss": 7.282337188720703,
+ "step": 142
+ },
+ {
+ "epoch": 1.9921397379912664,
+ "grad_norm": 0.01459511648863554,
+ "learning_rate": 0.0006,
+ "loss": 7.282924175262451,
+ "step": 143
+ },
+ {
+ "epoch": 2.0,
+ "grad_norm": 0.017474235966801643,
+ "learning_rate": 0.0006,
+ "loss": 7.254333972930908,
+ "step": 144
+ },
+ {
+ "epoch": 2.0,
+ "eval_loss": 7.318708896636963,
+ "eval_runtime": 59.6555,
+ "eval_samples_per_second": 40.935,
+ "eval_steps_per_second": 1.291,
+ "step": 144
+ },
+ {
+ "epoch": 2.0139737991266378,
+ "grad_norm": 0.019757507368922234,
+ "learning_rate": 0.0006,
+ "loss": 7.30288553237915,
+ "step": 145
+ },
+ {
+ "epoch": 2.027947598253275,
+ "grad_norm": 0.014571248553693295,
+ "learning_rate": 0.0006,
+ "loss": 7.274681568145752,
+ "step": 146
+ },
+ {
+ "epoch": 2.041921397379913,
+ "grad_norm": 0.015744198113679886,
+ "learning_rate": 0.0006,
+ "loss": 7.278632640838623,
+ "step": 147
+ },
+ {
+ "epoch": 2.05589519650655,
+ "grad_norm": 0.028048858046531677,
+ "learning_rate": 0.0006,
+ "loss": 7.26558780670166,
+ "step": 148
+ },
+ {
+ "epoch": 2.069868995633188,
+ "grad_norm": 0.06547307223081589,
+ "learning_rate": 0.0006,
+ "loss": 7.32567834854126,
+ "step": 149
+ },
+ {
+ "epoch": 2.0838427947598253,
+ "grad_norm": 0.05969763547182083,
+ "learning_rate": 0.0006,
+ "loss": 7.395836353302002,
+ "step": 150
+ },
+ {
+ "epoch": 2.097816593886463,
+ "grad_norm": 0.032574281096458435,
+ "learning_rate": 0.0006,
+ "loss": 7.318150997161865,
+ "step": 151
+ },
+ {
+ "epoch": 2.1117903930131003,
+ "grad_norm": 0.029452962800860405,
+ "learning_rate": 0.0006,
+ "loss": 7.308066368103027,
+ "step": 152
+ },
+ {
+ "epoch": 2.125764192139738,
+ "grad_norm": 0.03161991387605667,
+ "learning_rate": 0.0006,
+ "loss": 7.306290626525879,
+ "step": 153
+ },
+ {
+ "epoch": 2.1397379912663754,
+ "grad_norm": 0.027096295729279518,
+ "learning_rate": 0.0006,
+ "loss": 7.297796249389648,
+ "step": 154
+ },
+ {
+ "epoch": 2.153711790393013,
+ "grad_norm": 0.028657056391239166,
+ "learning_rate": 0.0006,
+ "loss": 7.317424774169922,
+ "step": 155
+ },
+ {
+ "epoch": 2.1676855895196505,
+ "grad_norm": 0.02046995982527733,
+ "learning_rate": 0.0006,
+ "loss": 7.297986030578613,
+ "step": 156
+ },
+ {
+ "epoch": 2.1816593886462883,
+ "grad_norm": 0.02220369316637516,
+ "learning_rate": 0.0006,
+ "loss": 7.268237590789795,
+ "step": 157
+ },
+ {
+ "epoch": 2.1956331877729256,
+ "grad_norm": 0.018381357192993164,
+ "learning_rate": 0.0006,
+ "loss": 7.259415626525879,
+ "step": 158
+ },
+ {
+ "epoch": 2.2096069868995634,
+ "grad_norm": 0.0204413291066885,
+ "learning_rate": 0.0006,
+ "loss": 7.258173942565918,
+ "step": 159
+ },
+ {
+ "epoch": 2.2235807860262007,
+ "grad_norm": 0.015874288976192474,
+ "learning_rate": 0.0006,
+ "loss": 7.281834602355957,
+ "step": 160
+ },
+ {
+ "epoch": 2.2375545851528384,
+ "grad_norm": 0.01634068600833416,
+ "learning_rate": 0.0006,
+ "loss": 7.2469868659973145,
+ "step": 161
+ },
+ {
+ "epoch": 2.251528384279476,
+ "grad_norm": 0.01728747971355915,
+ "learning_rate": 0.0006,
+ "loss": 7.280995845794678,
+ "step": 162
+ },
+ {
+ "epoch": 2.2655021834061135,
+ "grad_norm": 0.016341188922524452,
+ "learning_rate": 0.0006,
+ "loss": 7.273357391357422,
+ "step": 163
+ },
+ {
+ "epoch": 2.279475982532751,
+ "grad_norm": 0.016125807538628578,
+ "learning_rate": 0.0006,
+ "loss": 7.273510932922363,
+ "step": 164
+ },
+ {
+ "epoch": 2.2934497816593886,
+ "grad_norm": 0.01629287749528885,
+ "learning_rate": 0.0006,
+ "loss": 7.267029762268066,
+ "step": 165
+ },
+ {
+ "epoch": 2.3074235807860264,
+ "grad_norm": 0.015296131372451782,
+ "learning_rate": 0.0006,
+ "loss": 7.211989402770996,
+ "step": 166
+ },
+ {
+ "epoch": 2.3213973799126637,
+ "grad_norm": 0.01709570363163948,
+ "learning_rate": 0.0006,
+ "loss": 7.242851734161377,
+ "step": 167
+ },
+ {
+ "epoch": 2.3353711790393015,
+ "grad_norm": 0.012552388943731785,
+ "learning_rate": 0.0006,
+ "loss": 7.241696834564209,
+ "step": 168
+ },
+ {
+ "epoch": 2.349344978165939,
+ "grad_norm": 0.014571291394531727,
+ "learning_rate": 0.0006,
+ "loss": 7.270681381225586,
+ "step": 169
+ },
+ {
+ "epoch": 2.3633187772925766,
+ "grad_norm": 0.013389437459409237,
+ "learning_rate": 0.0006,
+ "loss": 7.191600799560547,
+ "step": 170
+ },
+ {
+ "epoch": 2.377292576419214,
+ "grad_norm": 0.011895314790308475,
+ "learning_rate": 0.0006,
+ "loss": 7.213610649108887,
+ "step": 171
+ },
+ {
+ "epoch": 2.3912663755458516,
+ "grad_norm": 0.01345642190426588,
+ "learning_rate": 0.0006,
+ "loss": 7.193787574768066,
+ "step": 172
+ },
+ {
+ "epoch": 2.405240174672489,
+ "grad_norm": 0.0091794328764081,
+ "learning_rate": 0.0006,
+ "loss": 7.233364105224609,
+ "step": 173
+ },
+ {
+ "epoch": 2.4192139737991267,
+ "grad_norm": 0.013767086900770664,
+ "learning_rate": 0.0006,
+ "loss": 7.217272758483887,
+ "step": 174
+ },
+ {
+ "epoch": 2.433187772925764,
+ "grad_norm": 0.015134960412979126,
+ "learning_rate": 0.0006,
+ "loss": 7.192791938781738,
+ "step": 175
+ },
+ {
+ "epoch": 2.447161572052402,
+ "grad_norm": 0.01790648326277733,
+ "learning_rate": 0.0006,
+ "loss": 7.229467391967773,
+ "step": 176
+ },
+ {
+ "epoch": 2.461135371179039,
+ "grad_norm": 0.015362377278506756,
+ "learning_rate": 0.0006,
+ "loss": 7.180346488952637,
+ "step": 177
+ },
+ {
+ "epoch": 2.475109170305677,
+ "grad_norm": 0.010505498386919498,
+ "learning_rate": 0.0006,
+ "loss": 7.188833236694336,
+ "step": 178
+ },
+ {
+ "epoch": 2.489082969432314,
+ "grad_norm": 0.009129747748374939,
+ "learning_rate": 0.0006,
+ "loss": 7.203153133392334,
+ "step": 179
+ },
+ {
+ "epoch": 2.503056768558952,
+ "grad_norm": 0.01147883478552103,
+ "learning_rate": 0.0006,
+ "loss": 7.163824558258057,
+ "step": 180
+ },
+ {
+ "epoch": 2.5170305676855893,
+ "grad_norm": 0.011338942684233189,
+ "learning_rate": 0.0006,
+ "loss": 7.16287899017334,
+ "step": 181
+ },
+ {
+ "epoch": 2.531004366812227,
+ "grad_norm": 0.013277675025165081,
+ "learning_rate": 0.0006,
+ "loss": 7.155905723571777,
+ "step": 182
+ },
+ {
+ "epoch": 2.544978165938865,
+ "grad_norm": 0.015269347466528416,
+ "learning_rate": 0.0006,
+ "loss": 7.189371109008789,
+ "step": 183
+ },
+ {
+ "epoch": 2.558951965065502,
+ "grad_norm": 0.017905596643686295,
+ "learning_rate": 0.0006,
+ "loss": 7.177547931671143,
+ "step": 184
+ },
+ {
+ "epoch": 2.5729257641921395,
+ "grad_norm": 0.024431610479950905,
+ "learning_rate": 0.0006,
+ "loss": 7.192169189453125,
+ "step": 185
+ },
+ {
+ "epoch": 2.5868995633187772,
+ "grad_norm": 0.03656961768865585,
+ "learning_rate": 0.0006,
+ "loss": 7.124387741088867,
+ "step": 186
+ },
+ {
+ "epoch": 2.600873362445415,
+ "grad_norm": 0.04337048903107643,
+ "learning_rate": 0.0006,
+ "loss": 7.140231609344482,
+ "step": 187
+ },
+ {
+ "epoch": 2.6148471615720523,
+ "grad_norm": 0.04047228768467903,
+ "learning_rate": 0.0006,
+ "loss": 7.171927452087402,
+ "step": 188
+ },
+ {
+ "epoch": 2.62882096069869,
+ "grad_norm": 0.07825770974159241,
+ "learning_rate": 0.0006,
+ "loss": 7.207631587982178,
+ "step": 189
+ },
+ {
+ "epoch": 2.6427947598253274,
+ "grad_norm": 0.07118063420057297,
+ "learning_rate": 0.0006,
+ "loss": 7.209488391876221,
+ "step": 190
+ },
+ {
+ "epoch": 2.656768558951965,
+ "grad_norm": 0.05153293535113335,
+ "learning_rate": 0.0006,
+ "loss": 7.14793062210083,
+ "step": 191
+ },
+ {
+ "epoch": 2.6707423580786025,
+ "grad_norm": 0.03318839520215988,
+ "learning_rate": 0.0006,
+ "loss": 7.163010597229004,
+ "step": 192
+ },
+ {
+ "epoch": 2.6847161572052403,
+ "grad_norm": 0.029333539307117462,
+ "learning_rate": 0.0006,
+ "loss": 7.159955978393555,
+ "step": 193
+ },
+ {
+ "epoch": 2.6986899563318776,
+ "grad_norm": 0.029405880719423294,
+ "learning_rate": 0.0006,
+ "loss": 7.136178493499756,
+ "step": 194
+ },
+ {
+ "epoch": 2.7126637554585153,
+ "grad_norm": 0.037815988063812256,
+ "learning_rate": 0.0006,
+ "loss": 7.183186054229736,
+ "step": 195
+ },
+ {
+ "epoch": 2.726637554585153,
+ "grad_norm": 0.020307250320911407,
+ "learning_rate": 0.0006,
+ "loss": 7.133164882659912,
+ "step": 196
+ },
+ {
+ "epoch": 2.7406113537117904,
+ "grad_norm": 0.026048069819808006,
+ "learning_rate": 0.0006,
+ "loss": 7.124977111816406,
+ "step": 197
+ },
+ {
+ "epoch": 2.7545851528384278,
+ "grad_norm": 0.027316724881529808,
+ "learning_rate": 0.0006,
+ "loss": 7.07265567779541,
+ "step": 198
+ },
+ {
+ "epoch": 2.7685589519650655,
+ "grad_norm": 0.021629009395837784,
+ "learning_rate": 0.0006,
+ "loss": 7.150020599365234,
+ "step": 199
+ },
+ {
+ "epoch": 2.7825327510917033,
+ "grad_norm": 0.01605929434299469,
+ "learning_rate": 0.0006,
+ "loss": 7.10919189453125,
+ "step": 200
+ },
+ {
+ "epoch": 2.7965065502183406,
+ "grad_norm": 0.020800089463591576,
+ "learning_rate": 0.0006,
+ "loss": 7.132045269012451,
+ "step": 201
+ },
+ {
+ "epoch": 2.810480349344978,
+ "grad_norm": 0.019225774332880974,
+ "learning_rate": 0.0006,
+ "loss": 7.128364562988281,
+ "step": 202
+ },
+ {
+ "epoch": 2.8244541484716157,
+ "grad_norm": 0.01561372634023428,
+ "learning_rate": 0.0006,
+ "loss": 7.083669662475586,
+ "step": 203
+ },
+ {
+ "epoch": 2.8384279475982535,
+ "grad_norm": 0.021969519555568695,
+ "learning_rate": 0.0006,
+ "loss": 7.050841331481934,
+ "step": 204
+ },
+ {
+ "epoch": 2.8524017467248908,
+ "grad_norm": 0.021453695371747017,
+ "learning_rate": 0.0006,
+ "loss": 7.097085475921631,
+ "step": 205
+ },
+ {
+ "epoch": 2.8663755458515285,
+ "grad_norm": 0.022836022078990936,
+ "learning_rate": 0.0006,
+ "loss": 7.10405969619751,
+ "step": 206
+ },
+ {
+ "epoch": 2.880349344978166,
+ "grad_norm": 0.02430851384997368,
+ "learning_rate": 0.0006,
+ "loss": 7.045181751251221,
+ "step": 207
+ },
+ {
+ "epoch": 2.8943231441048036,
+ "grad_norm": 0.020166993141174316,
+ "learning_rate": 0.0006,
+ "loss": 7.068183898925781,
+ "step": 208
+ },
+ {
+ "epoch": 2.908296943231441,
+ "grad_norm": 0.0118505684658885,
+ "learning_rate": 0.0006,
+ "loss": 7.07132625579834,
+ "step": 209
+ },
+ {
+ "epoch": 2.9222707423580787,
+ "grad_norm": 0.023772427812218666,
+ "learning_rate": 0.0006,
+ "loss": 7.055238723754883,
+ "step": 210
+ },
+ {
+ "epoch": 2.936244541484716,
+ "grad_norm": 0.02350712940096855,
+ "learning_rate": 0.0006,
+ "loss": 7.055898666381836,
+ "step": 211
+ },
+ {
+ "epoch": 2.950218340611354,
+ "grad_norm": 0.017244907096028328,
+ "learning_rate": 0.0006,
+ "loss": 7.0099711418151855,
+ "step": 212
+ },
+ {
+ "epoch": 2.964192139737991,
+ "grad_norm": 0.021565575152635574,
+ "learning_rate": 0.0006,
+ "loss": 6.991751670837402,
+ "step": 213
+ },
+ {
+ "epoch": 2.978165938864629,
+ "grad_norm": 0.03241860866546631,
+ "learning_rate": 0.0006,
+ "loss": 7.005980491638184,
+ "step": 214
+ },
+ {
+ "epoch": 2.992139737991266,
+ "grad_norm": 0.04613726586103439,
+ "learning_rate": 0.0006,
+ "loss": 7.027859687805176,
+ "step": 215
+ },
+ {
+ "epoch": 3.0,
+ "grad_norm": 0.06404894590377808,
+ "learning_rate": 0.0006,
+ "loss": 7.123725891113281,
+ "step": 216
+ },
+ {
+ "epoch": 3.0,
+ "eval_loss": 7.070647716522217,
+ "eval_runtime": 58.5801,
+ "eval_samples_per_second": 41.687,
+ "eval_steps_per_second": 1.314,
+ "step": 216
+ },
+ {
+ "epoch": 3.0139737991266378,
+ "grad_norm": 0.05542432889342308,
+ "learning_rate": 0.0006,
+ "loss": 7.078299522399902,
+ "step": 217
+ },
+ {
+ "epoch": 3.027947598253275,
+ "grad_norm": 0.026484526693820953,
+ "learning_rate": 0.0006,
+ "loss": 6.961159706115723,
+ "step": 218
+ },
+ {
+ "epoch": 3.041921397379913,
+ "grad_norm": 0.028276391327381134,
+ "learning_rate": 0.0006,
+ "loss": 7.019961833953857,
+ "step": 219
+ },
+ {
+ "epoch": 3.05589519650655,
+ "grad_norm": 0.024486595764756203,
+ "learning_rate": 0.0006,
+ "loss": 7.039677619934082,
+ "step": 220
+ },
+ {
+ "epoch": 3.069868995633188,
+ "grad_norm": 0.017620893195271492,
+ "learning_rate": 0.0006,
+ "loss": 6.9270782470703125,
+ "step": 221
+ },
+ {
+ "epoch": 3.0838427947598253,
+ "grad_norm": 0.02547302283346653,
+ "learning_rate": 0.0006,
+ "loss": 6.924233436584473,
+ "step": 222
+ },
+ {
+ "epoch": 3.097816593886463,
+ "grad_norm": 0.02539009042084217,
+ "learning_rate": 0.0006,
+ "loss": 6.973979949951172,
+ "step": 223
+ },
+ {
+ "epoch": 3.1117903930131003,
+ "grad_norm": 0.03256227448582649,
+ "learning_rate": 0.0006,
+ "loss": 6.933725357055664,
+ "step": 224
+ },
+ {
+ "epoch": 3.125764192139738,
+ "grad_norm": 0.04651800915598869,
+ "learning_rate": 0.0006,
+ "loss": 6.951044082641602,
+ "step": 225
+ },
+ {
+ "epoch": 3.1397379912663754,
+ "grad_norm": 0.04822975769639015,
+ "learning_rate": 0.0006,
+ "loss": 6.9614410400390625,
+ "step": 226
+ },
+ {
+ "epoch": 3.153711790393013,
+ "grad_norm": 0.031165020540356636,
+ "learning_rate": 0.0006,
+ "loss": 6.91193151473999,
+ "step": 227
+ },
+ {
+ "epoch": 3.1676855895196505,
+ "grad_norm": 0.029754292219877243,
+ "learning_rate": 0.0006,
+ "loss": 6.91802978515625,
+ "step": 228
+ },
+ {
+ "epoch": 3.1816593886462883,
+ "grad_norm": 0.02193152718245983,
+ "learning_rate": 0.0006,
+ "loss": 6.928090572357178,
+ "step": 229
+ },
+ {
+ "epoch": 3.1956331877729256,
+ "grad_norm": 0.02428470179438591,
+ "learning_rate": 0.0006,
+ "loss": 6.90396785736084,
+ "step": 230
+ },
+ {
+ "epoch": 3.2096069868995634,
+ "grad_norm": 0.023374345153570175,
+ "learning_rate": 0.0006,
+ "loss": 6.885858535766602,
+ "step": 231
+ },
+ {
+ "epoch": 3.2235807860262007,
+ "grad_norm": 0.023405000567436218,
+ "learning_rate": 0.0006,
+ "loss": 6.8855767250061035,
+ "step": 232
+ },
+ {
+ "epoch": 3.2375545851528384,
+ "grad_norm": 0.017277134582400322,
+ "learning_rate": 0.0006,
+ "loss": 6.901449203491211,
+ "step": 233
+ },
+ {
+ "epoch": 3.251528384279476,
+ "grad_norm": 0.019496750086545944,
+ "learning_rate": 0.0006,
+ "loss": 6.885687828063965,
+ "step": 234
+ },
+ {
+ "epoch": 3.2655021834061135,
+ "grad_norm": 0.016893276944756508,
+ "learning_rate": 0.0006,
+ "loss": 6.872369766235352,
+ "step": 235
+ },
+ {
+ "epoch": 3.279475982532751,
+ "grad_norm": 0.02058715932071209,
+ "learning_rate": 0.0006,
+ "loss": 6.823280334472656,
+ "step": 236
+ },
+ {
+ "epoch": 3.2934497816593886,
+ "grad_norm": 0.015530755743384361,
+ "learning_rate": 0.0006,
+ "loss": 6.860499858856201,
+ "step": 237
+ },
+ {
+ "epoch": 3.3074235807860264,
+ "grad_norm": 0.01934926211833954,
+ "learning_rate": 0.0006,
+ "loss": 6.817448616027832,
+ "step": 238
+ },
+ {
+ "epoch": 3.3213973799126637,
+ "grad_norm": 0.017413552850484848,
+ "learning_rate": 0.0006,
+ "loss": 6.881955146789551,
+ "step": 239
+ },
+ {
+ "epoch": 3.3353711790393015,
+ "grad_norm": 0.026478875428438187,
+ "learning_rate": 0.0006,
+ "loss": 6.827404499053955,
+ "step": 240
+ },
+ {
+ "epoch": 3.349344978165939,
+ "grad_norm": 0.032611701637506485,
+ "learning_rate": 0.0006,
+ "loss": 6.78285551071167,
+ "step": 241
+ },
+ {
+ "epoch": 3.3633187772925766,
+ "grad_norm": 0.041900306940078735,
+ "learning_rate": 0.0006,
+ "loss": 6.783053874969482,
+ "step": 242
+ },
+ {
+ "epoch": 3.377292576419214,
+ "grad_norm": 0.04149497672915459,
+ "learning_rate": 0.0006,
+ "loss": 6.745220184326172,
+ "step": 243
+ },
+ {
+ "epoch": 3.3912663755458516,
+ "grad_norm": 0.023949826136231422,
+ "learning_rate": 0.0006,
+ "loss": 6.841533184051514,
+ "step": 244
+ },
+ {
+ "epoch": 3.405240174672489,
+ "grad_norm": 0.028601842001080513,
+ "learning_rate": 0.0006,
+ "loss": 6.852852821350098,
+ "step": 245
+ },
+ {
+ "epoch": 3.4192139737991267,
+ "grad_norm": 0.027365155518054962,
+ "learning_rate": 0.0006,
+ "loss": 6.74376106262207,
+ "step": 246
+ },
+ {
+ "epoch": 3.433187772925764,
+ "grad_norm": 0.02856568619608879,
+ "learning_rate": 0.0006,
+ "loss": 6.74906063079834,
+ "step": 247
+ },
+ {
+ "epoch": 3.447161572052402,
+ "grad_norm": 0.019172416999936104,
+ "learning_rate": 0.0006,
+ "loss": 6.809206008911133,
+ "step": 248
+ },
+ {
+ "epoch": 3.461135371179039,
+ "grad_norm": 0.02161746844649315,
+ "learning_rate": 0.0006,
+ "loss": 6.789888381958008,
+ "step": 249
+ },
+ {
+ "epoch": 3.475109170305677,
+ "grad_norm": 0.020273666828870773,
+ "learning_rate": 0.0006,
+ "loss": 6.73512077331543,
+ "step": 250
+ },
+ {
+ "epoch": 3.489082969432314,
+ "grad_norm": 0.022329283878207207,
+ "learning_rate": 0.0006,
+ "loss": 6.826148509979248,
+ "step": 251
+ },
+ {
+ "epoch": 3.503056768558952,
+ "grad_norm": 0.02284284122288227,
+ "learning_rate": 0.0006,
+ "loss": 6.748521327972412,
+ "step": 252
+ },
+ {
+ "epoch": 3.5170305676855893,
+ "grad_norm": 0.026368247345089912,
+ "learning_rate": 0.0006,
+ "loss": 6.645879745483398,
+ "step": 253
+ },
+ {
+ "epoch": 3.531004366812227,
+ "grad_norm": 0.03627244383096695,
+ "learning_rate": 0.0006,
+ "loss": 6.666064739227295,
+ "step": 254
+ },
+ {
+ "epoch": 3.544978165938865,
+ "grad_norm": 0.04140935093164444,
+ "learning_rate": 0.0006,
+ "loss": 6.712275981903076,
+ "step": 255
+ },
+ {
+ "epoch": 3.558951965065502,
+ "grad_norm": 0.0390462800860405,
+ "learning_rate": 0.0006,
+ "loss": 6.665900230407715,
+ "step": 256
+ },
+ {
+ "epoch": 3.5729257641921395,
+ "grad_norm": 0.05178583785891533,
+ "learning_rate": 0.0006,
+ "loss": 6.753298759460449,
+ "step": 257
+ },
+ {
+ "epoch": 3.5868995633187772,
+ "grad_norm": 0.054328061640262604,
+ "learning_rate": 0.0006,
+ "loss": 6.732633113861084,
+ "step": 258
+ },
+ {
+ "epoch": 3.600873362445415,
+ "grad_norm": 0.04590460658073425,
+ "learning_rate": 0.0006,
+ "loss": 6.65250301361084,
+ "step": 259
+ },
+ {
+ "epoch": 3.6148471615720523,
+ "grad_norm": 0.044331666082143784,
+ "learning_rate": 0.0006,
+ "loss": 6.68654203414917,
+ "step": 260
+ },
+ {
+ "epoch": 3.62882096069869,
+ "grad_norm": 0.05308730527758598,
+ "learning_rate": 0.0006,
+ "loss": 6.642482757568359,
+ "step": 261
+ },
+ {
+ "epoch": 3.6427947598253274,
+ "grad_norm": 0.03999016433954239,
+ "learning_rate": 0.0006,
+ "loss": 6.625227451324463,
+ "step": 262
+ },
+ {
+ "epoch": 3.656768558951965,
+ "grad_norm": 0.04088086634874344,
+ "learning_rate": 0.0006,
+ "loss": 6.67623233795166,
+ "step": 263
+ },
+ {
+ "epoch": 3.6707423580786025,
+ "grad_norm": 0.03615636005997658,
+ "learning_rate": 0.0006,
+ "loss": 6.6869001388549805,
+ "step": 264
+ },
+ {
+ "epoch": 3.6847161572052403,
+ "grad_norm": 0.02817360684275627,
+ "learning_rate": 0.0006,
+ "loss": 6.674601078033447,
+ "step": 265
+ },
+ {
+ "epoch": 3.6986899563318776,
+ "grad_norm": 0.032586123794317245,
+ "learning_rate": 0.0006,
+ "loss": 6.612994194030762,
+ "step": 266
+ },
+ {
+ "epoch": 3.7126637554585153,
+ "grad_norm": 0.031117867678403854,
+ "learning_rate": 0.0006,
+ "loss": 6.6244354248046875,
+ "step": 267
+ },
+ {
+ "epoch": 3.726637554585153,
+ "grad_norm": 0.028330687433481216,
+ "learning_rate": 0.0006,
+ "loss": 6.669557094573975,
+ "step": 268
+ },
+ {
+ "epoch": 3.7406113537117904,
+ "grad_norm": 0.021489202976226807,
+ "learning_rate": 0.0006,
+ "loss": 6.632035732269287,
+ "step": 269
+ },
+ {
+ "epoch": 3.7545851528384278,
+ "grad_norm": 0.024269424378871918,
+ "learning_rate": 0.0006,
+ "loss": 6.660867691040039,
+ "step": 270
+ },
+ {
+ "epoch": 3.7685589519650655,
+ "grad_norm": 0.02237873338162899,
+ "learning_rate": 0.0006,
+ "loss": 6.602587699890137,
+ "step": 271
+ },
+ {
+ "epoch": 3.7825327510917033,
+ "grad_norm": 0.026503929868340492,
+ "learning_rate": 0.0006,
+ "loss": 6.599200248718262,
+ "step": 272
+ },
+ {
+ "epoch": 3.7965065502183406,
+ "grad_norm": 0.03687124326825142,
+ "learning_rate": 0.0006,
+ "loss": 6.567296504974365,
+ "step": 273
+ },
+ {
+ "epoch": 3.810480349344978,
+ "grad_norm": 0.04708952456712723,
+ "learning_rate": 0.0006,
+ "loss": 6.621265411376953,
+ "step": 274
+ },
+ {
+ "epoch": 3.8244541484716157,
+ "grad_norm": 0.041846614331007004,
+ "learning_rate": 0.0006,
+ "loss": 6.550536155700684,
+ "step": 275
+ },
+ {
+ "epoch": 3.8384279475982535,
+ "grad_norm": 0.031249675899744034,
+ "learning_rate": 0.0006,
+ "loss": 6.54278564453125,
+ "step": 276
+ },
+ {
+ "epoch": 3.8524017467248908,
+ "grad_norm": 0.023082256317138672,
+ "learning_rate": 0.0006,
+ "loss": 6.544787406921387,
+ "step": 277
+ },
+ {
+ "epoch": 3.8663755458515285,
+ "grad_norm": 0.020995570346713066,
+ "learning_rate": 0.0006,
+ "loss": 6.518143653869629,
+ "step": 278
+ },
+ {
+ "epoch": 3.880349344978166,
+ "grad_norm": 0.020839324221014977,
+ "learning_rate": 0.0006,
+ "loss": 6.546667098999023,
+ "step": 279
+ },
+ {
+ "epoch": 3.8943231441048036,
+ "grad_norm": 0.018377432599663734,
+ "learning_rate": 0.0006,
+ "loss": 6.478307723999023,
+ "step": 280
+ },
+ {
+ "epoch": 3.908296943231441,
+ "grad_norm": 0.021713724359869957,
+ "learning_rate": 0.0006,
+ "loss": 6.485574245452881,
+ "step": 281
+ },
+ {
+ "epoch": 3.9222707423580787,
+ "grad_norm": 0.01359301246702671,
+ "learning_rate": 0.0006,
+ "loss": 6.570268630981445,
+ "step": 282
+ },
+ {
+ "epoch": 3.936244541484716,
+ "grad_norm": 0.016233332455158234,
+ "learning_rate": 0.0006,
+ "loss": 6.57150936126709,
+ "step": 283
+ },
+ {
+ "epoch": 3.950218340611354,
+ "grad_norm": 0.020687608048319817,
+ "learning_rate": 0.0006,
+ "loss": 6.527956008911133,
+ "step": 284
+ },
+ {
+ "epoch": 3.964192139737991,
+ "grad_norm": 0.02032964862883091,
+ "learning_rate": 0.0006,
+ "loss": 6.468169689178467,
+ "step": 285
+ },
+ {
+ "epoch": 3.978165938864629,
+ "grad_norm": 0.016676288098096848,
+ "learning_rate": 0.0006,
+ "loss": 6.448418617248535,
+ "step": 286
+ },
+ {
+ "epoch": 3.992139737991266,
+ "grad_norm": 0.0168539360165596,
+ "learning_rate": 0.0006,
+ "loss": 6.461180686950684,
+ "step": 287
+ },
+ {
+ "epoch": 4.0,
+ "grad_norm": 0.02666233666241169,
+ "learning_rate": 0.0006,
+ "loss": 6.344893932342529,
+ "step": 288
+ },
+ {
+ "epoch": 4.0,
+ "eval_loss": 6.5384979248046875,
+ "eval_runtime": 58.814,
+ "eval_samples_per_second": 41.521,
+ "eval_steps_per_second": 1.309,
+ "step": 288
+ },
+ {
+ "epoch": 4.013973799126638,
+ "grad_norm": 0.04154708981513977,
+ "learning_rate": 0.0006,
+ "loss": 6.41855525970459,
+ "step": 289
+ },
+ {
+ "epoch": 4.0279475982532755,
+ "grad_norm": 0.04895668104290962,
+ "learning_rate": 0.0006,
+ "loss": 6.499302864074707,
+ "step": 290
+ },
+ {
+ "epoch": 4.041921397379912,
+ "grad_norm": 0.034464482218027115,
+ "learning_rate": 0.0006,
+ "loss": 6.438611030578613,
+ "step": 291
+ },
+ {
+ "epoch": 4.05589519650655,
+ "grad_norm": 0.06845773756504059,
+ "learning_rate": 0.0006,
+ "loss": 6.538361072540283,
+ "step": 292
+ },
+ {
+ "epoch": 4.069868995633188,
+ "grad_norm": 0.053768858313560486,
+ "learning_rate": 0.0006,
+ "loss": 6.527524948120117,
+ "step": 293
+ },
+ {
+ "epoch": 4.083842794759826,
+ "grad_norm": 0.039037931710481644,
+ "learning_rate": 0.0006,
+ "loss": 6.503847122192383,
+ "step": 294
+ },
+ {
+ "epoch": 4.097816593886463,
+ "grad_norm": 0.03962196782231331,
+ "learning_rate": 0.0006,
+ "loss": 6.475178241729736,
+ "step": 295
+ },
+ {
+ "epoch": 4.1117903930131,
+ "grad_norm": 0.049101535230875015,
+ "learning_rate": 0.0006,
+ "loss": 6.474587917327881,
+ "step": 296
+ },
+ {
+ "epoch": 4.125764192139738,
+ "grad_norm": 0.02830282226204872,
+ "learning_rate": 0.0006,
+ "loss": 6.45065975189209,
+ "step": 297
+ },
+ {
+ "epoch": 4.139737991266376,
+ "grad_norm": 0.023153482005000114,
+ "learning_rate": 0.0006,
+ "loss": 6.5218706130981445,
+ "step": 298
+ },
+ {
+ "epoch": 4.153711790393013,
+ "grad_norm": 0.025064002722501755,
+ "learning_rate": 0.0006,
+ "loss": 6.456475257873535,
+ "step": 299
+ },
+ {
+ "epoch": 4.1676855895196505,
+ "grad_norm": 0.02357092685997486,
+ "learning_rate": 0.0006,
+ "loss": 6.485952377319336,
+ "step": 300
+ },
+ {
+ "epoch": 4.181659388646288,
+ "grad_norm": 0.021832305938005447,
+ "learning_rate": 0.0006,
+ "loss": 6.425506591796875,
+ "step": 301
+ },
+ {
+ "epoch": 4.195633187772926,
+ "grad_norm": 0.0227900929749012,
+ "learning_rate": 0.0006,
+ "loss": 6.439155578613281,
+ "step": 302
+ },
+ {
+ "epoch": 4.209606986899563,
+ "grad_norm": 0.015961607918143272,
+ "learning_rate": 0.0006,
+ "loss": 6.460862636566162,
+ "step": 303
+ },
+ {
+ "epoch": 4.223580786026201,
+ "grad_norm": 0.017876390367746353,
+ "learning_rate": 0.0006,
+ "loss": 6.44564151763916,
+ "step": 304
+ },
+ {
+ "epoch": 4.2375545851528384,
+ "grad_norm": 0.013703204691410065,
+ "learning_rate": 0.0006,
+ "loss": 6.418414115905762,
+ "step": 305
+ },
+ {
+ "epoch": 4.251528384279476,
+ "grad_norm": 0.015240306034684181,
+ "learning_rate": 0.0006,
+ "loss": 6.411238670349121,
+ "step": 306
+ },
+ {
+ "epoch": 4.265502183406113,
+ "grad_norm": 0.01587662845849991,
+ "learning_rate": 0.0006,
+ "loss": 6.395642280578613,
+ "step": 307
+ },
+ {
+ "epoch": 4.279475982532751,
+ "grad_norm": 0.014650700613856316,
+ "learning_rate": 0.0006,
+ "loss": 6.302793025970459,
+ "step": 308
+ },
+ {
+ "epoch": 4.293449781659389,
+ "grad_norm": 0.016533225774765015,
+ "learning_rate": 0.0006,
+ "loss": 6.380588054656982,
+ "step": 309
+ },
+ {
+ "epoch": 4.307423580786026,
+ "grad_norm": 0.019963741302490234,
+ "learning_rate": 0.0006,
+ "loss": 6.345336437225342,
+ "step": 310
+ },
+ {
+ "epoch": 4.321397379912664,
+ "grad_norm": 0.02237936295568943,
+ "learning_rate": 0.0006,
+ "loss": 6.387770652770996,
+ "step": 311
+ },
+ {
+ "epoch": 4.335371179039301,
+ "grad_norm": 0.026658328250050545,
+ "learning_rate": 0.0006,
+ "loss": 6.323662757873535,
+ "step": 312
+ },
+ {
+ "epoch": 4.349344978165939,
+ "grad_norm": 0.02852284163236618,
+ "learning_rate": 0.0006,
+ "loss": 6.382204055786133,
+ "step": 313
+ },
+ {
+ "epoch": 4.3633187772925766,
+ "grad_norm": 0.028815090656280518,
+ "learning_rate": 0.0006,
+ "loss": 6.402390480041504,
+ "step": 314
+ },
+ {
+ "epoch": 4.377292576419214,
+ "grad_norm": 0.029393529519438744,
+ "learning_rate": 0.0006,
+ "loss": 6.280278205871582,
+ "step": 315
+ },
+ {
+ "epoch": 4.391266375545851,
+ "grad_norm": 0.02186041697859764,
+ "learning_rate": 0.0006,
+ "loss": 6.302707672119141,
+ "step": 316
+ },
+ {
+ "epoch": 4.405240174672489,
+ "grad_norm": 0.02071218006312847,
+ "learning_rate": 0.0006,
+ "loss": 6.3365020751953125,
+ "step": 317
+ },
+ {
+ "epoch": 4.419213973799127,
+ "grad_norm": 0.02779117226600647,
+ "learning_rate": 0.0006,
+ "loss": 6.252157211303711,
+ "step": 318
+ },
+ {
+ "epoch": 4.4331877729257645,
+ "grad_norm": 0.03470654785633087,
+ "learning_rate": 0.0006,
+ "loss": 6.319070339202881,
+ "step": 319
+ },
+ {
+ "epoch": 4.447161572052401,
+ "grad_norm": 0.037067804485559464,
+ "learning_rate": 0.0006,
+ "loss": 6.314116477966309,
+ "step": 320
+ },
+ {
+ "epoch": 4.461135371179039,
+ "grad_norm": 0.032182756811380386,
+ "learning_rate": 0.0006,
+ "loss": 6.267298698425293,
+ "step": 321
+ },
+ {
+ "epoch": 4.475109170305677,
+ "grad_norm": 0.026305217295885086,
+ "learning_rate": 0.0006,
+ "loss": 6.331688404083252,
+ "step": 322
+ },
+ {
+ "epoch": 4.489082969432315,
+ "grad_norm": 0.027228357270359993,
+ "learning_rate": 0.0006,
+ "loss": 6.31139612197876,
+ "step": 323
+ },
+ {
+ "epoch": 4.503056768558952,
+ "grad_norm": 0.025781169533729553,
+ "learning_rate": 0.0006,
+ "loss": 6.194684982299805,
+ "step": 324
+ },
+ {
+ "epoch": 4.517030567685589,
+ "grad_norm": 0.029186977073550224,
+ "learning_rate": 0.0006,
+ "loss": 6.31306266784668,
+ "step": 325
+ },
+ {
+ "epoch": 4.531004366812227,
+ "grad_norm": 0.023608777672052383,
+ "learning_rate": 0.0006,
+ "loss": 6.285243034362793,
+ "step": 326
+ },
+ {
+ "epoch": 4.544978165938865,
+ "grad_norm": 0.02267594076693058,
+ "learning_rate": 0.0006,
+ "loss": 6.30342960357666,
+ "step": 327
+ },
+ {
+ "epoch": 4.558951965065502,
+ "grad_norm": 0.01926310732960701,
+ "learning_rate": 0.0006,
+ "loss": 6.280606269836426,
+ "step": 328
+ },
+ {
+ "epoch": 4.5729257641921395,
+ "grad_norm": 0.026484837755560875,
+ "learning_rate": 0.0006,
+ "loss": 6.237980365753174,
+ "step": 329
+ },
+ {
+ "epoch": 4.586899563318777,
+ "grad_norm": 0.026228854432702065,
+ "learning_rate": 0.0006,
+ "loss": 6.3018293380737305,
+ "step": 330
+ },
+ {
+ "epoch": 4.600873362445415,
+ "grad_norm": 0.022096967324614525,
+ "learning_rate": 0.0006,
+ "loss": 6.253863334655762,
+ "step": 331
+ },
+ {
+ "epoch": 4.614847161572053,
+ "grad_norm": 0.027223890647292137,
+ "learning_rate": 0.0006,
+ "loss": 6.166882038116455,
+ "step": 332
+ },
+ {
+ "epoch": 4.62882096069869,
+ "grad_norm": 0.03160124272108078,
+ "learning_rate": 0.0006,
+ "loss": 6.230309963226318,
+ "step": 333
+ },
+ {
+ "epoch": 4.642794759825327,
+ "grad_norm": 0.04110539332032204,
+ "learning_rate": 0.0006,
+ "loss": 6.2334818840026855,
+ "step": 334
+ },
+ {
+ "epoch": 4.656768558951965,
+ "grad_norm": 0.04355933889746666,
+ "learning_rate": 0.0006,
+ "loss": 6.160956382751465,
+ "step": 335
+ },
+ {
+ "epoch": 4.670742358078603,
+ "grad_norm": 0.03611086308956146,
+ "learning_rate": 0.0006,
+ "loss": 6.23399543762207,
+ "step": 336
+ },
+ {
+ "epoch": 4.68471615720524,
+ "grad_norm": 0.051811400800943375,
+ "learning_rate": 0.0006,
+ "loss": 6.241855144500732,
+ "step": 337
+ },
+ {
+ "epoch": 4.698689956331878,
+ "grad_norm": 0.057434193789958954,
+ "learning_rate": 0.0006,
+ "loss": 6.220993995666504,
+ "step": 338
+ },
+ {
+ "epoch": 4.712663755458515,
+ "grad_norm": 0.04420953616499901,
+ "learning_rate": 0.0006,
+ "loss": 6.300461292266846,
+ "step": 339
+ },
+ {
+ "epoch": 4.726637554585153,
+ "grad_norm": 0.03515457361936569,
+ "learning_rate": 0.0006,
+ "loss": 6.269349098205566,
+ "step": 340
+ },
+ {
+ "epoch": 4.74061135371179,
+ "grad_norm": 0.03222600743174553,
+ "learning_rate": 0.0006,
+ "loss": 6.248613357543945,
+ "step": 341
+ },
+ {
+ "epoch": 4.754585152838428,
+ "grad_norm": 0.0384046845138073,
+ "learning_rate": 0.0006,
+ "loss": 6.266979217529297,
+ "step": 342
+ },
+ {
+ "epoch": 4.7685589519650655,
+ "grad_norm": 0.0514867827296257,
+ "learning_rate": 0.0006,
+ "loss": 6.234103202819824,
+ "step": 343
+ },
+ {
+ "epoch": 4.782532751091703,
+ "grad_norm": 0.050747793167829514,
+ "learning_rate": 0.0006,
+ "loss": 6.322582244873047,
+ "step": 344
+ },
+ {
+ "epoch": 4.796506550218341,
+ "grad_norm": 0.040878988802433014,
+ "learning_rate": 0.0006,
+ "loss": 6.212509632110596,
+ "step": 345
+ },
+ {
+ "epoch": 4.810480349344978,
+ "grad_norm": 0.034655820578336716,
+ "learning_rate": 0.0006,
+ "loss": 6.265879154205322,
+ "step": 346
+ },
+ {
+ "epoch": 4.824454148471616,
+ "grad_norm": 0.04202224686741829,
+ "learning_rate": 0.0006,
+ "loss": 6.2896409034729,
+ "step": 347
+ },
+ {
+ "epoch": 4.8384279475982535,
+ "grad_norm": 0.03632277995347977,
+ "learning_rate": 0.0006,
+ "loss": 6.253917217254639,
+ "step": 348
+ },
+ {
+ "epoch": 4.85240174672489,
+ "grad_norm": 0.028645765036344528,
+ "learning_rate": 0.0006,
+ "loss": 6.207965850830078,
+ "step": 349
+ },
+ {
+ "epoch": 4.866375545851528,
+ "grad_norm": 0.025305170565843582,
+ "learning_rate": 0.0006,
+ "loss": 6.168068885803223,
+ "step": 350
+ },
+ {
+ "epoch": 4.880349344978166,
+ "grad_norm": 0.022137803956866264,
+ "learning_rate": 0.0006,
+ "loss": 6.235283374786377,
+ "step": 351
+ },
+ {
+ "epoch": 4.894323144104804,
+ "grad_norm": 0.01814538985490799,
+ "learning_rate": 0.0006,
+ "loss": 6.259641647338867,
+ "step": 352
+ },
+ {
+ "epoch": 4.908296943231441,
+ "grad_norm": 0.018465086817741394,
+ "learning_rate": 0.0006,
+ "loss": 6.111738681793213,
+ "step": 353
+ },
+ {
+ "epoch": 4.922270742358078,
+ "grad_norm": 0.018638933077454567,
+ "learning_rate": 0.0006,
+ "loss": 6.225642204284668,
+ "step": 354
+ },
+ {
+ "epoch": 4.936244541484716,
+ "grad_norm": 0.016461260616779327,
+ "learning_rate": 0.0006,
+ "loss": 6.140647888183594,
+ "step": 355
+ },
+ {
+ "epoch": 4.950218340611354,
+ "grad_norm": 0.015458385460078716,
+ "learning_rate": 0.0006,
+ "loss": 6.155224800109863,
+ "step": 356
+ },
+ {
+ "epoch": 4.964192139737992,
+ "grad_norm": 0.014934048056602478,
+ "learning_rate": 0.0006,
+ "loss": 6.101555824279785,
+ "step": 357
+ },
+ {
+ "epoch": 4.978165938864628,
+ "grad_norm": 0.013979545794427395,
+ "learning_rate": 0.0006,
+ "loss": 6.161718845367432,
+ "step": 358
+ },
+ {
+ "epoch": 4.992139737991266,
+ "grad_norm": 0.01640394888818264,
+ "learning_rate": 0.0006,
+ "loss": 6.055768966674805,
+ "step": 359
+ },
+ {
+ "epoch": 5.0,
+ "grad_norm": 0.015135680325329304,
+ "learning_rate": 0.0006,
+ "loss": 6.169695854187012,
+ "step": 360
+ },
+ {
+ "epoch": 5.0,
+ "eval_loss": 6.16408109664917,
+ "eval_runtime": 59.3649,
+ "eval_samples_per_second": 41.135,
+ "eval_steps_per_second": 1.297,
+ "step": 360
+ },
+ {
+ "epoch": 5.013973799126638,
+ "grad_norm": 0.012849048711359501,
+ "learning_rate": 0.0006,
+ "loss": 6.055830001831055,
+ "step": 361
+ },
+ {
+ "epoch": 5.0279475982532755,
+ "grad_norm": 0.018169576302170753,
+ "learning_rate": 0.0006,
+ "loss": 6.082655429840088,
+ "step": 362
+ },
+ {
+ "epoch": 5.041921397379912,
+ "grad_norm": 0.024293430149555206,
+ "learning_rate": 0.0006,
+ "loss": 6.109032154083252,
+ "step": 363
+ },
+ {
+ "epoch": 5.05589519650655,
+ "grad_norm": 0.03266787901520729,
+ "learning_rate": 0.0006,
+ "loss": 6.030613899230957,
+ "step": 364
+ },
+ {
+ "epoch": 5.069868995633188,
+ "grad_norm": 0.03572266176342964,
+ "learning_rate": 0.0006,
+ "loss": 6.050390720367432,
+ "step": 365
+ },
+ {
+ "epoch": 5.083842794759826,
+ "grad_norm": 0.032625213265419006,
+ "learning_rate": 0.0006,
+ "loss": 6.169313430786133,
+ "step": 366
+ },
+ {
+ "epoch": 5.097816593886463,
+ "grad_norm": 0.04051872715353966,
+ "learning_rate": 0.0006,
+ "loss": 6.10398006439209,
+ "step": 367
+ },
+ {
+ "epoch": 5.1117903930131,
+ "grad_norm": 0.0351213738322258,
+ "learning_rate": 0.0006,
+ "loss": 6.121973514556885,
+ "step": 368
+ },
+ {
+ "epoch": 5.125764192139738,
+ "grad_norm": 0.03523759916424751,
+ "learning_rate": 0.0006,
+ "loss": 6.0626654624938965,
+ "step": 369
+ },
+ {
+ "epoch": 5.139737991266376,
+ "grad_norm": 0.02706182189285755,
+ "learning_rate": 0.0006,
+ "loss": 6.049354076385498,
+ "step": 370
+ },
+ {
+ "epoch": 5.153711790393013,
+ "grad_norm": 0.024212589487433434,
+ "learning_rate": 0.0006,
+ "loss": 6.034826278686523,
+ "step": 371
+ },
+ {
+ "epoch": 5.1676855895196505,
+ "grad_norm": 0.024798082187771797,
+ "learning_rate": 0.0006,
+ "loss": 6.004058837890625,
+ "step": 372
+ },
+ {
+ "epoch": 5.181659388646288,
+ "grad_norm": 0.02325567416846752,
+ "learning_rate": 0.0006,
+ "loss": 5.986461639404297,
+ "step": 373
+ },
+ {
+ "epoch": 5.195633187772926,
+ "grad_norm": 0.019060570746660233,
+ "learning_rate": 0.0006,
+ "loss": 6.059736251831055,
+ "step": 374
+ },
+ {
+ "epoch": 5.209606986899563,
+ "grad_norm": 0.016822976991534233,
+ "learning_rate": 0.0006,
+ "loss": 6.072957515716553,
+ "step": 375
+ },
+ {
+ "epoch": 5.223580786026201,
+ "grad_norm": 0.017218658700585365,
+ "learning_rate": 0.0006,
+ "loss": 6.031739234924316,
+ "step": 376
+ },
+ {
+ "epoch": 5.2375545851528384,
+ "grad_norm": 0.014889383688569069,
+ "learning_rate": 0.0006,
+ "loss": 6.025674343109131,
+ "step": 377
+ },
+ {
+ "epoch": 5.251528384279476,
+ "grad_norm": 0.01708107627928257,
+ "learning_rate": 0.0006,
+ "loss": 6.090451240539551,
+ "step": 378
+ },
+ {
+ "epoch": 5.265502183406113,
+ "grad_norm": 0.020260317251086235,
+ "learning_rate": 0.0006,
+ "loss": 6.094257354736328,
+ "step": 379
+ },
+ {
+ "epoch": 5.279475982532751,
+ "grad_norm": 0.020110400393605232,
+ "learning_rate": 0.0006,
+ "loss": 6.021188735961914,
+ "step": 380
+ },
+ {
+ "epoch": 5.293449781659389,
+ "grad_norm": 0.02027830481529236,
+ "learning_rate": 0.0006,
+ "loss": 6.022156238555908,
+ "step": 381
+ },
+ {
+ "epoch": 5.307423580786026,
+ "grad_norm": 0.022747062146663666,
+ "learning_rate": 0.0006,
+ "loss": 6.011836051940918,
+ "step": 382
+ },
+ {
+ "epoch": 5.321397379912664,
+ "grad_norm": 0.01990230567753315,
+ "learning_rate": 0.0006,
+ "loss": 6.003054618835449,
+ "step": 383
+ },
+ {
+ "epoch": 5.335371179039301,
+ "grad_norm": 0.01596238650381565,
+ "learning_rate": 0.0006,
+ "loss": 5.991410732269287,
+ "step": 384
+ },
+ {
+ "epoch": 5.349344978165939,
+ "grad_norm": 0.015847105532884598,
+ "learning_rate": 0.0006,
+ "loss": 6.100622177124023,
+ "step": 385
+ },
+ {
+ "epoch": 5.3633187772925766,
+ "grad_norm": 0.016179397702217102,
+ "learning_rate": 0.0006,
+ "loss": 6.032659530639648,
+ "step": 386
+ },
+ {
+ "epoch": 5.377292576419214,
+ "grad_norm": 0.018256602808833122,
+ "learning_rate": 0.0006,
+ "loss": 5.961983680725098,
+ "step": 387
+ },
+ {
+ "epoch": 5.391266375545851,
+ "grad_norm": 0.02805912122130394,
+ "learning_rate": 0.0006,
+ "loss": 5.983541965484619,
+ "step": 388
+ },
+ {
+ "epoch": 5.405240174672489,
+ "grad_norm": 0.039082255214452744,
+ "learning_rate": 0.0006,
+ "loss": 6.021474361419678,
+ "step": 389
+ },
+ {
+ "epoch": 5.419213973799127,
+ "grad_norm": 0.036757659167051315,
+ "learning_rate": 0.0006,
+ "loss": 6.005046367645264,
+ "step": 390
+ },
+ {
+ "epoch": 5.4331877729257645,
+ "grad_norm": 0.035277366638183594,
+ "learning_rate": 0.0006,
+ "loss": 6.036296844482422,
+ "step": 391
+ },
+ {
+ "epoch": 5.447161572052401,
+ "grad_norm": 0.034404635429382324,
+ "learning_rate": 0.0006,
+ "loss": 5.994539260864258,
+ "step": 392
+ },
+ {
+ "epoch": 5.461135371179039,
+ "grad_norm": 0.0377206988632679,
+ "learning_rate": 0.0006,
+ "loss": 5.941534519195557,
+ "step": 393
+ },
+ {
+ "epoch": 5.475109170305677,
+ "grad_norm": 0.0389922633767128,
+ "learning_rate": 0.0006,
+ "loss": 5.983644485473633,
+ "step": 394
+ },
+ {
+ "epoch": 5.489082969432315,
+ "grad_norm": 0.04176778718829155,
+ "learning_rate": 0.0006,
+ "loss": 6.030922889709473,
+ "step": 395
+ },
+ {
+ "epoch": 5.503056768558952,
+ "grad_norm": 0.0337153784930706,
+ "learning_rate": 0.0006,
+ "loss": 5.961367130279541,
+ "step": 396
+ },
+ {
+ "epoch": 5.517030567685589,
+ "grad_norm": 0.036116816103458405,
+ "learning_rate": 0.0006,
+ "loss": 5.819280624389648,
+ "step": 397
+ },
+ {
+ "epoch": 5.531004366812227,
+ "grad_norm": 0.030724041163921356,
+ "learning_rate": 0.0006,
+ "loss": 5.906380653381348,
+ "step": 398
+ },
+ {
+ "epoch": 5.544978165938865,
+ "grad_norm": 0.030264202505350113,
+ "learning_rate": 0.0006,
+ "loss": 5.975833415985107,
+ "step": 399
+ },
+ {
+ "epoch": 5.558951965065502,
+ "grad_norm": 0.032096318900585175,
+ "learning_rate": 0.0006,
+ "loss": 6.001348972320557,
+ "step": 400
+ },
+ {
+ "epoch": 5.5729257641921395,
+ "grad_norm": 0.030579356476664543,
+ "learning_rate": 0.0006,
+ "loss": 5.928768634796143,
+ "step": 401
+ },
+ {
+ "epoch": 5.586899563318777,
+ "grad_norm": 0.028241973370313644,
+ "learning_rate": 0.0006,
+ "loss": 5.92582893371582,
+ "step": 402
+ },
+ {
+ "epoch": 5.600873362445415,
+ "grad_norm": 0.023086953908205032,
+ "learning_rate": 0.0006,
+ "loss": 5.892926216125488,
+ "step": 403
+ },
+ {
+ "epoch": 5.614847161572053,
+ "grad_norm": 0.02547992393374443,
+ "learning_rate": 0.0006,
+ "loss": 5.903195381164551,
+ "step": 404
+ },
+ {
+ "epoch": 5.62882096069869,
+ "grad_norm": 0.023089831694960594,
+ "learning_rate": 0.0006,
+ "loss": 5.912033557891846,
+ "step": 405
+ },
+ {
+ "epoch": 5.642794759825327,
+ "grad_norm": 0.025446368381381035,
+ "learning_rate": 0.0006,
+ "loss": 5.926138401031494,
+ "step": 406
+ },
+ {
+ "epoch": 5.656768558951965,
+ "grad_norm": 0.03748747706413269,
+ "learning_rate": 0.0006,
+ "loss": 5.913451194763184,
+ "step": 407
+ },
+ {
+ "epoch": 5.670742358078603,
+ "grad_norm": 0.03493810072541237,
+ "learning_rate": 0.0006,
+ "loss": 5.97633695602417,
+ "step": 408
+ },
+ {
+ "epoch": 5.68471615720524,
+ "grad_norm": 0.023131275549530983,
+ "learning_rate": 0.0006,
+ "loss": 5.899571418762207,
+ "step": 409
+ },
+ {
+ "epoch": 5.698689956331878,
+ "grad_norm": 0.02986014075577259,
+ "learning_rate": 0.0006,
+ "loss": 5.894903182983398,
+ "step": 410
+ },
+ {
+ "epoch": 5.712663755458515,
+ "grad_norm": 0.030171144753694534,
+ "learning_rate": 0.0006,
+ "loss": 5.876595497131348,
+ "step": 411
+ },
+ {
+ "epoch": 5.726637554585153,
+ "grad_norm": 0.029377546161413193,
+ "learning_rate": 0.0006,
+ "loss": 5.911637306213379,
+ "step": 412
+ },
+ {
+ "epoch": 5.74061135371179,
+ "grad_norm": 0.029829490929841995,
+ "learning_rate": 0.0006,
+ "loss": 5.819629669189453,
+ "step": 413
+ },
+ {
+ "epoch": 5.754585152838428,
+ "grad_norm": 0.024139299988746643,
+ "learning_rate": 0.0006,
+ "loss": 5.81728458404541,
+ "step": 414
+ },
+ {
+ "epoch": 5.7685589519650655,
+ "grad_norm": 0.021453432738780975,
+ "learning_rate": 0.0006,
+ "loss": 5.987326145172119,
+ "step": 415
+ },
+ {
+ "epoch": 5.782532751091703,
+ "grad_norm": 0.017942246049642563,
+ "learning_rate": 0.0006,
+ "loss": 5.893008232116699,
+ "step": 416
+ },
+ {
+ "epoch": 5.796506550218341,
+ "grad_norm": 0.019724637269973755,
+ "learning_rate": 0.0006,
+ "loss": 5.912441253662109,
+ "step": 417
+ },
+ {
+ "epoch": 5.810480349344978,
+ "grad_norm": 0.018597912043333054,
+ "learning_rate": 0.0006,
+ "loss": 5.860394477844238,
+ "step": 418
+ },
+ {
+ "epoch": 5.824454148471616,
+ "grad_norm": 0.016232412308454514,
+ "learning_rate": 0.0006,
+ "loss": 5.94991397857666,
+ "step": 419
+ },
+ {
+ "epoch": 5.8384279475982535,
+ "grad_norm": 0.01667204685509205,
+ "learning_rate": 0.0006,
+ "loss": 5.8506364822387695,
+ "step": 420
+ },
+ {
+ "epoch": 5.85240174672489,
+ "grad_norm": 0.014232151210308075,
+ "learning_rate": 0.0006,
+ "loss": 5.928424835205078,
+ "step": 421
+ },
+ {
+ "epoch": 5.866375545851528,
+ "grad_norm": 0.01570476032793522,
+ "learning_rate": 0.0006,
+ "loss": 5.953692436218262,
+ "step": 422
+ },
+ {
+ "epoch": 5.880349344978166,
+ "grad_norm": 0.01586179807782173,
+ "learning_rate": 0.0006,
+ "loss": 5.820014476776123,
+ "step": 423
+ },
+ {
+ "epoch": 5.894323144104804,
+ "grad_norm": 0.014729145914316177,
+ "learning_rate": 0.0006,
+ "loss": 5.860154628753662,
+ "step": 424
+ },
+ {
+ "epoch": 5.908296943231441,
+ "grad_norm": 0.013111459091305733,
+ "learning_rate": 0.0006,
+ "loss": 5.85988187789917,
+ "step": 425
+ },
+ {
+ "epoch": 5.922270742358078,
+ "grad_norm": 0.012937922962009907,
+ "learning_rate": 0.0006,
+ "loss": 5.756265640258789,
+ "step": 426
+ },
+ {
+ "epoch": 5.936244541484716,
+ "grad_norm": 0.01377453189343214,
+ "learning_rate": 0.0006,
+ "loss": 5.8333048820495605,
+ "step": 427
+ },
+ {
+ "epoch": 5.950218340611354,
+ "grad_norm": 0.014054795727133751,
+ "learning_rate": 0.0006,
+ "loss": 5.803333759307861,
+ "step": 428
+ },
+ {
+ "epoch": 5.964192139737992,
+ "grad_norm": 0.01674959622323513,
+ "learning_rate": 0.0006,
+ "loss": 5.8718485832214355,
+ "step": 429
+ },
+ {
+ "epoch": 5.978165938864628,
+ "grad_norm": 0.018733017146587372,
+ "learning_rate": 0.0006,
+ "loss": 5.875979900360107,
+ "step": 430
+ },
+ {
+ "epoch": 5.992139737991266,
+ "grad_norm": 0.02088983915746212,
+ "learning_rate": 0.0006,
+ "loss": 5.763634204864502,
+ "step": 431
+ },
+ {
+ "epoch": 6.0,
+ "grad_norm": 0.026705985888838768,
+ "learning_rate": 0.0006,
+ "loss": 5.6922478675842285,
+ "step": 432
+ },
+ {
+ "epoch": 6.0,
+ "eval_loss": 5.847958087921143,
+ "eval_runtime": 58.5451,
+ "eval_samples_per_second": 41.711,
+ "eval_steps_per_second": 1.315,
+ "step": 432
+ },
+ {
+ "epoch": 6.013973799126638,
+ "grad_norm": 0.03181544691324234,
+ "learning_rate": 0.0006,
+ "loss": 5.786368370056152,
+ "step": 433
+ },
+ {
+ "epoch": 6.0279475982532755,
+ "grad_norm": 0.03238112851977348,
+ "learning_rate": 0.0006,
+ "loss": 5.810310363769531,
+ "step": 434
+ },
+ {
+ "epoch": 6.041921397379912,
+ "grad_norm": 0.03916756808757782,
+ "learning_rate": 0.0006,
+ "loss": 5.7761993408203125,
+ "step": 435
+ },
+ {
+ "epoch": 6.05589519650655,
+ "grad_norm": 0.044009730219841,
+ "learning_rate": 0.0006,
+ "loss": 5.898112773895264,
+ "step": 436
+ },
+ {
+ "epoch": 6.069868995633188,
+ "grad_norm": 0.045235246419906616,
+ "learning_rate": 0.0006,
+ "loss": 5.876371383666992,
+ "step": 437
+ },
+ {
+ "epoch": 6.083842794759826,
+ "grad_norm": 0.050218384712934494,
+ "learning_rate": 0.0006,
+ "loss": 5.805103302001953,
+ "step": 438
+ },
+ {
+ "epoch": 6.097816593886463,
+ "grad_norm": 0.05444490164518356,
+ "learning_rate": 0.0006,
+ "loss": 5.826424598693848,
+ "step": 439
+ },
+ {
+ "epoch": 6.1117903930131,
+ "grad_norm": 0.04890982061624527,
+ "learning_rate": 0.0006,
+ "loss": 5.7735395431518555,
+ "step": 440
+ },
+ {
+ "epoch": 6.125764192139738,
+ "grad_norm": 0.05478639155626297,
+ "learning_rate": 0.0006,
+ "loss": 5.823677062988281,
+ "step": 441
+ },
+ {
+ "epoch": 6.139737991266376,
+ "grad_norm": 0.057562313973903656,
+ "learning_rate": 0.0006,
+ "loss": 5.889334678649902,
+ "step": 442
+ },
+ {
+ "epoch": 6.153711790393013,
+ "grad_norm": 0.06447703391313553,
+ "learning_rate": 0.0006,
+ "loss": 5.84621524810791,
+ "step": 443
+ },
+ {
+ "epoch": 6.1676855895196505,
+ "grad_norm": 0.048861872404813766,
+ "learning_rate": 0.0006,
+ "loss": 5.931595802307129,
+ "step": 444
+ },
+ {
+ "epoch": 6.181659388646288,
+ "grad_norm": 0.05521339178085327,
+ "learning_rate": 0.0006,
+ "loss": 5.8229780197143555,
+ "step": 445
+ },
+ {
+ "epoch": 6.195633187772926,
+ "grad_norm": 0.053666990250349045,
+ "learning_rate": 0.0006,
+ "loss": 5.879191875457764,
+ "step": 446
+ },
+ {
+ "epoch": 6.209606986899563,
+ "grad_norm": 0.0451025515794754,
+ "learning_rate": 0.0006,
+ "loss": 5.801628112792969,
+ "step": 447
+ },
+ {
+ "epoch": 6.223580786026201,
+ "grad_norm": 0.04291655868291855,
+ "learning_rate": 0.0006,
+ "loss": 5.914680480957031,
+ "step": 448
+ },
+ {
+ "epoch": 6.2375545851528384,
+ "grad_norm": 0.035102490335702896,
+ "learning_rate": 0.0006,
+ "loss": 5.843682765960693,
+ "step": 449
+ },
+ {
+ "epoch": 6.251528384279476,
+ "grad_norm": 0.03403995931148529,
+ "learning_rate": 0.0006,
+ "loss": 5.827154159545898,
+ "step": 450
+ },
+ {
+ "epoch": 6.265502183406113,
+ "grad_norm": 0.03444375470280647,
+ "learning_rate": 0.0006,
+ "loss": 5.800136566162109,
+ "step": 451
+ },
+ {
+ "epoch": 6.279475982532751,
+ "grad_norm": 0.03165159001946449,
+ "learning_rate": 0.0006,
+ "loss": 5.906252384185791,
+ "step": 452
+ },
+ {
+ "epoch": 6.293449781659389,
+ "grad_norm": 0.026153093203902245,
+ "learning_rate": 0.0006,
+ "loss": 5.7423200607299805,
+ "step": 453
+ },
+ {
+ "epoch": 6.307423580786026,
+ "grad_norm": 0.024357657879590988,
+ "learning_rate": 0.0006,
+ "loss": 5.701364517211914,
+ "step": 454
+ },
+ {
+ "epoch": 6.321397379912664,
+ "grad_norm": 0.021508049219846725,
+ "learning_rate": 0.0006,
+ "loss": 5.8328094482421875,
+ "step": 455
+ },
+ {
+ "epoch": 6.335371179039301,
+ "grad_norm": 0.017313921824097633,
+ "learning_rate": 0.0006,
+ "loss": 5.792145252227783,
+ "step": 456
+ },
+ {
+ "epoch": 6.349344978165939,
+ "grad_norm": 0.018563397228717804,
+ "learning_rate": 0.0006,
+ "loss": 5.732672214508057,
+ "step": 457
+ },
+ {
+ "epoch": 6.3633187772925766,
+ "grad_norm": 0.016568679362535477,
+ "learning_rate": 0.0006,
+ "loss": 5.81948184967041,
+ "step": 458
+ },
+ {
+ "epoch": 6.377292576419214,
+ "grad_norm": 0.014133770950138569,
+ "learning_rate": 0.0006,
+ "loss": 5.734982490539551,
+ "step": 459
+ },
+ {
+ "epoch": 6.391266375545851,
+ "grad_norm": 0.014114447869360447,
+ "learning_rate": 0.0006,
+ "loss": 5.70073127746582,
+ "step": 460
+ },
+ {
+ "epoch": 6.405240174672489,
+ "grad_norm": 0.012907393276691437,
+ "learning_rate": 0.0006,
+ "loss": 5.758626937866211,
+ "step": 461
+ },
+ {
+ "epoch": 6.419213973799127,
+ "grad_norm": 0.012673444114625454,
+ "learning_rate": 0.0006,
+ "loss": 5.775138854980469,
+ "step": 462
+ },
+ {
+ "epoch": 6.4331877729257645,
+ "grad_norm": 0.01311410591006279,
+ "learning_rate": 0.0006,
+ "loss": 5.70920991897583,
+ "step": 463
+ },
+ {
+ "epoch": 6.447161572052401,
+ "grad_norm": 0.012935544364154339,
+ "learning_rate": 0.0006,
+ "loss": 5.781479835510254,
+ "step": 464
+ },
+ {
+ "epoch": 6.461135371179039,
+ "grad_norm": 0.01439738366752863,
+ "learning_rate": 0.0006,
+ "loss": 5.691315650939941,
+ "step": 465
+ },
+ {
+ "epoch": 6.475109170305677,
+ "grad_norm": 0.01205186452716589,
+ "learning_rate": 0.0006,
+ "loss": 5.662499904632568,
+ "step": 466
+ },
+ {
+ "epoch": 6.489082969432315,
+ "grad_norm": 0.011575652286410332,
+ "learning_rate": 0.0006,
+ "loss": 5.6271772384643555,
+ "step": 467
+ },
+ {
+ "epoch": 6.503056768558952,
+ "grad_norm": 0.011672502383589745,
+ "learning_rate": 0.0006,
+ "loss": 5.761662006378174,
+ "step": 468
+ },
+ {
+ "epoch": 6.517030567685589,
+ "grad_norm": 0.011541751213371754,
+ "learning_rate": 0.0006,
+ "loss": 5.762078285217285,
+ "step": 469
+ },
+ {
+ "epoch": 6.531004366812227,
+ "grad_norm": 0.011396365240216255,
+ "learning_rate": 0.0006,
+ "loss": 5.67873477935791,
+ "step": 470
+ },
+ {
+ "epoch": 6.544978165938865,
+ "grad_norm": 0.010688499547541142,
+ "learning_rate": 0.0006,
+ "loss": 5.717051982879639,
+ "step": 471
+ },
+ {
+ "epoch": 6.558951965065502,
+ "grad_norm": 0.012224317528307438,
+ "learning_rate": 0.0006,
+ "loss": 5.707948684692383,
+ "step": 472
+ },
+ {
+ "epoch": 6.5729257641921395,
+ "grad_norm": 0.011856825090944767,
+ "learning_rate": 0.0006,
+ "loss": 5.70878791809082,
+ "step": 473
+ },
+ {
+ "epoch": 6.586899563318777,
+ "grad_norm": 0.01199064590036869,
+ "learning_rate": 0.0006,
+ "loss": 5.708697319030762,
+ "step": 474
+ },
+ {
+ "epoch": 6.600873362445415,
+ "grad_norm": 0.01219446212053299,
+ "learning_rate": 0.0006,
+ "loss": 5.6061577796936035,
+ "step": 475
+ },
+ {
+ "epoch": 6.614847161572053,
+ "grad_norm": 0.013397484086453915,
+ "learning_rate": 0.0006,
+ "loss": 5.624789714813232,
+ "step": 476
+ },
+ {
+ "epoch": 6.62882096069869,
+ "grad_norm": 0.01483136136084795,
+ "learning_rate": 0.0006,
+ "loss": 5.721141338348389,
+ "step": 477
+ },
+ {
+ "epoch": 6.642794759825327,
+ "grad_norm": 0.0194488987326622,
+ "learning_rate": 0.0006,
+ "loss": 5.604279518127441,
+ "step": 478
+ },
+ {
+ "epoch": 6.656768558951965,
+ "grad_norm": 0.02079896256327629,
+ "learning_rate": 0.0006,
+ "loss": 5.586322784423828,
+ "step": 479
+ },
+ {
+ "epoch": 6.670742358078603,
+ "grad_norm": 0.018996229395270348,
+ "learning_rate": 0.0006,
+ "loss": 5.5987772941589355,
+ "step": 480
+ },
+ {
+ "epoch": 6.68471615720524,
+ "grad_norm": 0.015779603272676468,
+ "learning_rate": 0.0006,
+ "loss": 5.588602066040039,
+ "step": 481
+ },
+ {
+ "epoch": 6.698689956331878,
+ "grad_norm": 0.015322973020374775,
+ "learning_rate": 0.0006,
+ "loss": 5.686467170715332,
+ "step": 482
+ },
+ {
+ "epoch": 6.712663755458515,
+ "grad_norm": 0.015282213687896729,
+ "learning_rate": 0.0006,
+ "loss": 5.664676666259766,
+ "step": 483
+ },
+ {
+ "epoch": 6.726637554585153,
+ "grad_norm": 0.016589272767305374,
+ "learning_rate": 0.0006,
+ "loss": 5.666739463806152,
+ "step": 484
+ },
+ {
+ "epoch": 6.74061135371179,
+ "grad_norm": 0.01811421848833561,
+ "learning_rate": 0.0006,
+ "loss": 5.647593021392822,
+ "step": 485
+ },
+ {
+ "epoch": 6.754585152838428,
+ "grad_norm": 0.019652029499411583,
+ "learning_rate": 0.0006,
+ "loss": 5.628726959228516,
+ "step": 486
+ },
+ {
+ "epoch": 6.7685589519650655,
+ "grad_norm": 0.02118665538728237,
+ "learning_rate": 0.0006,
+ "loss": 5.6580705642700195,
+ "step": 487
+ },
+ {
+ "epoch": 6.782532751091703,
+ "grad_norm": 0.02237832546234131,
+ "learning_rate": 0.0006,
+ "loss": 5.635931968688965,
+ "step": 488
+ },
+ {
+ "epoch": 6.796506550218341,
+ "grad_norm": 0.023897631093859673,
+ "learning_rate": 0.0006,
+ "loss": 5.517949104309082,
+ "step": 489
+ },
+ {
+ "epoch": 6.810480349344978,
+ "grad_norm": 0.02442227490246296,
+ "learning_rate": 0.0006,
+ "loss": 5.600021839141846,
+ "step": 490
+ },
+ {
+ "epoch": 6.824454148471616,
+ "grad_norm": 0.024675287306308746,
+ "learning_rate": 0.0006,
+ "loss": 5.6639909744262695,
+ "step": 491
+ },
+ {
+ "epoch": 6.8384279475982535,
+ "grad_norm": 0.030372392386198044,
+ "learning_rate": 0.0006,
+ "loss": 5.582888603210449,
+ "step": 492
+ },
+ {
+ "epoch": 6.85240174672489,
+ "grad_norm": 0.03237690031528473,
+ "learning_rate": 0.0006,
+ "loss": 5.573896884918213,
+ "step": 493
+ },
+ {
+ "epoch": 6.866375545851528,
+ "grad_norm": 0.03511589393019676,
+ "learning_rate": 0.0006,
+ "loss": 5.565878868103027,
+ "step": 494
+ },
+ {
+ "epoch": 6.880349344978166,
+ "grad_norm": 0.03883938118815422,
+ "learning_rate": 0.0006,
+ "loss": 5.58540678024292,
+ "step": 495
+ },
+ {
+ "epoch": 6.894323144104804,
+ "grad_norm": 0.04175502806901932,
+ "learning_rate": 0.0006,
+ "loss": 5.6095194816589355,
+ "step": 496
+ },
+ {
+ "epoch": 6.908296943231441,
+ "grad_norm": 0.04013441503047943,
+ "learning_rate": 0.0006,
+ "loss": 5.654401779174805,
+ "step": 497
+ },
+ {
+ "epoch": 6.922270742358078,
+ "grad_norm": 0.04618770629167557,
+ "learning_rate": 0.0006,
+ "loss": 5.6483473777771,
+ "step": 498
+ },
+ {
+ "epoch": 6.936244541484716,
+ "grad_norm": 0.04507308453321457,
+ "learning_rate": 0.0006,
+ "loss": 5.5856218338012695,
+ "step": 499
+ },
+ {
+ "epoch": 6.950218340611354,
+ "grad_norm": 0.03151383996009827,
+ "learning_rate": 0.0006,
+ "loss": 5.594086647033691,
+ "step": 500
+ },
+ {
+ "epoch": 6.964192139737992,
+ "grad_norm": 0.026883797720074654,
+ "learning_rate": 0.0006,
+ "loss": 5.607676029205322,
+ "step": 501
+ },
+ {
+ "epoch": 6.978165938864628,
+ "grad_norm": 0.02981836162507534,
+ "learning_rate": 0.0006,
+ "loss": 5.563666820526123,
+ "step": 502
+ },
+ {
+ "epoch": 6.992139737991266,
+ "grad_norm": 0.03233477845788002,
+ "learning_rate": 0.0006,
+ "loss": 5.667543411254883,
+ "step": 503
+ },
+ {
+ "epoch": 7.0,
+ "grad_norm": 0.029927456751465797,
+ "learning_rate": 0.0006,
+ "loss": 5.733482837677002,
+ "step": 504
+ },
+ {
+ "epoch": 7.0,
+ "eval_loss": 5.679076671600342,
+ "eval_runtime": 59.5965,
+ "eval_samples_per_second": 40.976,
+ "eval_steps_per_second": 1.292,
+ "step": 504
+ },
+ {
+ "epoch": 7.013973799126638,
+ "grad_norm": 0.027081597596406937,
+ "learning_rate": 0.0006,
+ "loss": 5.634487152099609,
+ "step": 505
+ },
+ {
+ "epoch": 7.0279475982532755,
+ "grad_norm": 0.031959421932697296,
+ "learning_rate": 0.0006,
+ "loss": 5.643294811248779,
+ "step": 506
+ },
+ {
+ "epoch": 7.041921397379912,
+ "grad_norm": 0.02802063524723053,
+ "learning_rate": 0.0006,
+ "loss": 5.573239326477051,
+ "step": 507
+ },
+ {
+ "epoch": 7.05589519650655,
+ "grad_norm": 0.032579705119132996,
+ "learning_rate": 0.0006,
+ "loss": 5.627150058746338,
+ "step": 508
+ },
+ {
+ "epoch": 7.069868995633188,
+ "grad_norm": 0.02503928542137146,
+ "learning_rate": 0.0006,
+ "loss": 5.6045308113098145,
+ "step": 509
+ },
+ {
+ "epoch": 7.083842794759826,
+ "grad_norm": 0.02421317622065544,
+ "learning_rate": 0.0006,
+ "loss": 5.522153377532959,
+ "step": 510
+ },
+ {
+ "epoch": 7.097816593886463,
+ "grad_norm": 0.021983426064252853,
+ "learning_rate": 0.0006,
+ "loss": 5.632939338684082,
+ "step": 511
+ },
+ {
+ "epoch": 7.1117903930131,
+ "grad_norm": 0.021933183073997498,
+ "learning_rate": 0.0006,
+ "loss": 5.592185974121094,
+ "step": 512
+ },
+ {
+ "epoch": 7.125764192139738,
+ "grad_norm": 0.02287031151354313,
+ "learning_rate": 0.0006,
+ "loss": 5.613700866699219,
+ "step": 513
+ },
+ {
+ "epoch": 7.139737991266376,
+ "grad_norm": 0.02281602844595909,
+ "learning_rate": 0.0006,
+ "loss": 5.551383018493652,
+ "step": 514
+ },
+ {
+ "epoch": 7.153711790393013,
+ "grad_norm": 0.021050360053777695,
+ "learning_rate": 0.0006,
+ "loss": 5.55275821685791,
+ "step": 515
+ },
+ {
+ "epoch": 7.1676855895196505,
+ "grad_norm": 0.018299926072359085,
+ "learning_rate": 0.0006,
+ "loss": 5.528225421905518,
+ "step": 516
+ },
+ {
+ "epoch": 7.181659388646288,
+ "grad_norm": 0.02024853602051735,
+ "learning_rate": 0.0006,
+ "loss": 5.522453784942627,
+ "step": 517
+ },
+ {
+ "epoch": 7.195633187772926,
+ "grad_norm": 0.02135239914059639,
+ "learning_rate": 0.0006,
+ "loss": 5.527522563934326,
+ "step": 518
+ },
+ {
+ "epoch": 7.209606986899563,
+ "grad_norm": 0.023394184187054634,
+ "learning_rate": 0.0006,
+ "loss": 5.4535017013549805,
+ "step": 519
+ },
+ {
+ "epoch": 7.223580786026201,
+ "grad_norm": 0.025952080264687538,
+ "learning_rate": 0.0006,
+ "loss": 5.5374674797058105,
+ "step": 520
+ },
+ {
+ "epoch": 7.2375545851528384,
+ "grad_norm": 0.022287718951702118,
+ "learning_rate": 0.0006,
+ "loss": 5.493753433227539,
+ "step": 521
+ },
+ {
+ "epoch": 7.251528384279476,
+ "grad_norm": 0.02018447406589985,
+ "learning_rate": 0.0006,
+ "loss": 5.510575294494629,
+ "step": 522
+ },
+ {
+ "epoch": 7.265502183406113,
+ "grad_norm": 0.02050507813692093,
+ "learning_rate": 0.0006,
+ "loss": 5.466026782989502,
+ "step": 523
+ },
+ {
+ "epoch": 7.279475982532751,
+ "grad_norm": 0.023688802495598793,
+ "learning_rate": 0.0006,
+ "loss": 5.485815525054932,
+ "step": 524
+ },
+ {
+ "epoch": 7.293449781659389,
+ "grad_norm": 0.021078843623399734,
+ "learning_rate": 0.0006,
+ "loss": 5.415581703186035,
+ "step": 525
+ },
+ {
+ "epoch": 7.307423580786026,
+ "grad_norm": 0.018266011029481888,
+ "learning_rate": 0.0006,
+ "loss": 5.402815818786621,
+ "step": 526
+ },
+ {
+ "epoch": 7.321397379912664,
+ "grad_norm": 0.019693493843078613,
+ "learning_rate": 0.0006,
+ "loss": 5.504674911499023,
+ "step": 527
+ },
+ {
+ "epoch": 7.335371179039301,
+ "grad_norm": 0.02031373605132103,
+ "learning_rate": 0.0006,
+ "loss": 5.503837585449219,
+ "step": 528
+ },
+ {
+ "epoch": 7.349344978165939,
+ "grad_norm": 0.017636626958847046,
+ "learning_rate": 0.0006,
+ "loss": 5.530580997467041,
+ "step": 529
+ },
+ {
+ "epoch": 7.3633187772925766,
+ "grad_norm": 0.01787244901061058,
+ "learning_rate": 0.0006,
+ "loss": 5.450218677520752,
+ "step": 530
+ },
+ {
+ "epoch": 7.377292576419214,
+ "grad_norm": 0.0170669574290514,
+ "learning_rate": 0.0006,
+ "loss": 5.481570243835449,
+ "step": 531
+ },
+ {
+ "epoch": 7.391266375545851,
+ "grad_norm": 0.01802165061235428,
+ "learning_rate": 0.0006,
+ "loss": 5.386394500732422,
+ "step": 532
+ },
+ {
+ "epoch": 7.405240174672489,
+ "grad_norm": 0.022949472069740295,
+ "learning_rate": 0.0006,
+ "loss": 5.424929618835449,
+ "step": 533
+ },
+ {
+ "epoch": 7.419213973799127,
+ "grad_norm": 0.034213390201330185,
+ "learning_rate": 0.0006,
+ "loss": 5.346663475036621,
+ "step": 534
+ },
+ {
+ "epoch": 7.4331877729257645,
+ "grad_norm": 0.03847593814134598,
+ "learning_rate": 0.0006,
+ "loss": 5.418482780456543,
+ "step": 535
+ },
+ {
+ "epoch": 7.447161572052401,
+ "grad_norm": 0.029393676668405533,
+ "learning_rate": 0.0006,
+ "loss": 5.456090927124023,
+ "step": 536
+ },
+ {
+ "epoch": 7.461135371179039,
+ "grad_norm": 0.03388667106628418,
+ "learning_rate": 0.0006,
+ "loss": 5.572983741760254,
+ "step": 537
+ },
+ {
+ "epoch": 7.475109170305677,
+ "grad_norm": 0.039690613746643066,
+ "learning_rate": 0.0006,
+ "loss": 5.414067268371582,
+ "step": 538
+ },
+ {
+ "epoch": 7.489082969432315,
+ "grad_norm": 0.03630776330828667,
+ "learning_rate": 0.0006,
+ "loss": 5.523739814758301,
+ "step": 539
+ },
+ {
+ "epoch": 7.503056768558952,
+ "grad_norm": 0.03356870263814926,
+ "learning_rate": 0.0006,
+ "loss": 5.444767951965332,
+ "step": 540
+ },
+ {
+ "epoch": 7.517030567685589,
+ "grad_norm": 0.030812304466962814,
+ "learning_rate": 0.0006,
+ "loss": 5.461244106292725,
+ "step": 541
+ },
+ {
+ "epoch": 7.531004366812227,
+ "grad_norm": 0.03319217637181282,
+ "learning_rate": 0.0006,
+ "loss": 5.502161026000977,
+ "step": 542
+ },
+ {
+ "epoch": 7.544978165938865,
+ "grad_norm": 0.032997481524944305,
+ "learning_rate": 0.0006,
+ "loss": 5.4162278175354,
+ "step": 543
+ },
+ {
+ "epoch": 7.558951965065502,
+ "grad_norm": 0.03364962339401245,
+ "learning_rate": 0.0006,
+ "loss": 5.415736198425293,
+ "step": 544
+ },
+ {
+ "epoch": 7.5729257641921395,
+ "grad_norm": 0.0344221405684948,
+ "learning_rate": 0.0006,
+ "loss": 5.541967868804932,
+ "step": 545
+ },
+ {
+ "epoch": 7.586899563318777,
+ "grad_norm": 0.029609503224492073,
+ "learning_rate": 0.0006,
+ "loss": 5.440614223480225,
+ "step": 546
+ },
+ {
+ "epoch": 7.600873362445415,
+ "grad_norm": 0.029055926948785782,
+ "learning_rate": 0.0006,
+ "loss": 5.462865352630615,
+ "step": 547
+ },
+ {
+ "epoch": 7.614847161572053,
+ "grad_norm": 0.02658848837018013,
+ "learning_rate": 0.0006,
+ "loss": 5.444467544555664,
+ "step": 548
+ },
+ {
+ "epoch": 7.62882096069869,
+ "grad_norm": 0.026276404038071632,
+ "learning_rate": 0.0006,
+ "loss": 5.462278842926025,
+ "step": 549
+ },
+ {
+ "epoch": 7.642794759825327,
+ "grad_norm": 0.0286889486014843,
+ "learning_rate": 0.0006,
+ "loss": 5.4849534034729,
+ "step": 550
+ },
+ {
+ "epoch": 7.656768558951965,
+ "grad_norm": 0.026849817484617233,
+ "learning_rate": 0.0006,
+ "loss": 5.431473731994629,
+ "step": 551
+ },
+ {
+ "epoch": 7.670742358078603,
+ "grad_norm": 0.02312396466732025,
+ "learning_rate": 0.0006,
+ "loss": 5.45462703704834,
+ "step": 552
+ },
+ {
+ "epoch": 7.68471615720524,
+ "grad_norm": 0.026197485625743866,
+ "learning_rate": 0.0006,
+ "loss": 5.452552318572998,
+ "step": 553
+ },
+ {
+ "epoch": 7.698689956331878,
+ "grad_norm": 0.02747255191206932,
+ "learning_rate": 0.0006,
+ "loss": 5.447169780731201,
+ "step": 554
+ },
+ {
+ "epoch": 7.712663755458515,
+ "grad_norm": 0.028123432770371437,
+ "learning_rate": 0.0006,
+ "loss": 5.394400596618652,
+ "step": 555
+ },
+ {
+ "epoch": 7.726637554585153,
+ "grad_norm": 0.02599870041012764,
+ "learning_rate": 0.0006,
+ "loss": 5.398341178894043,
+ "step": 556
+ },
+ {
+ "epoch": 7.74061135371179,
+ "grad_norm": 0.022171657532453537,
+ "learning_rate": 0.0006,
+ "loss": 5.368820667266846,
+ "step": 557
+ },
+ {
+ "epoch": 7.754585152838428,
+ "grad_norm": 0.022309480234980583,
+ "learning_rate": 0.0006,
+ "loss": 5.416772365570068,
+ "step": 558
+ },
+ {
+ "epoch": 7.7685589519650655,
+ "grad_norm": 0.024074165150523186,
+ "learning_rate": 0.0006,
+ "loss": 5.359002113342285,
+ "step": 559
+ },
+ {
+ "epoch": 7.782532751091703,
+ "grad_norm": 0.02636653557419777,
+ "learning_rate": 0.0006,
+ "loss": 5.484368324279785,
+ "step": 560
+ },
+ {
+ "epoch": 7.796506550218341,
+ "grad_norm": 0.02203752100467682,
+ "learning_rate": 0.0006,
+ "loss": 5.375498294830322,
+ "step": 561
+ },
+ {
+ "epoch": 7.810480349344978,
+ "grad_norm": 0.019924819469451904,
+ "learning_rate": 0.0006,
+ "loss": 5.334723472595215,
+ "step": 562
+ },
+ {
+ "epoch": 7.824454148471616,
+ "grad_norm": 0.018755966797471046,
+ "learning_rate": 0.0006,
+ "loss": 5.348781108856201,
+ "step": 563
+ },
+ {
+ "epoch": 7.8384279475982535,
+ "grad_norm": 0.016747845336794853,
+ "learning_rate": 0.0006,
+ "loss": 5.433079719543457,
+ "step": 564
+ },
+ {
+ "epoch": 7.85240174672489,
+ "grad_norm": 0.0170395877212286,
+ "learning_rate": 0.0006,
+ "loss": 5.346794128417969,
+ "step": 565
+ },
+ {
+ "epoch": 7.866375545851528,
+ "grad_norm": 0.016074176877737045,
+ "learning_rate": 0.0006,
+ "loss": 5.3277740478515625,
+ "step": 566
+ },
+ {
+ "epoch": 7.880349344978166,
+ "grad_norm": 0.014119806699454784,
+ "learning_rate": 0.0006,
+ "loss": 5.483551979064941,
+ "step": 567
+ },
+ {
+ "epoch": 7.894323144104804,
+ "grad_norm": 0.015270394273102283,
+ "learning_rate": 0.0006,
+ "loss": 5.378519535064697,
+ "step": 568
+ },
+ {
+ "epoch": 7.908296943231441,
+ "grad_norm": 0.015677539631724358,
+ "learning_rate": 0.0006,
+ "loss": 5.290210247039795,
+ "step": 569
+ },
+ {
+ "epoch": 7.922270742358078,
+ "grad_norm": 0.015930423513054848,
+ "learning_rate": 0.0006,
+ "loss": 5.427360534667969,
+ "step": 570
+ },
+ {
+ "epoch": 7.936244541484716,
+ "grad_norm": 0.016641564667224884,
+ "learning_rate": 0.0006,
+ "loss": 5.301599979400635,
+ "step": 571
+ },
+ {
+ "epoch": 7.950218340611354,
+ "grad_norm": 0.018293552100658417,
+ "learning_rate": 0.0006,
+ "loss": 5.277889251708984,
+ "step": 572
+ },
+ {
+ "epoch": 7.964192139737992,
+ "grad_norm": 0.017618799582123756,
+ "learning_rate": 0.0006,
+ "loss": 5.297194480895996,
+ "step": 573
+ },
+ {
+ "epoch": 7.978165938864628,
+ "grad_norm": 0.015034242533147335,
+ "learning_rate": 0.0006,
+ "loss": 5.42428731918335,
+ "step": 574
+ },
+ {
+ "epoch": 7.992139737991266,
+ "grad_norm": 0.01617511734366417,
+ "learning_rate": 0.0006,
+ "loss": 5.420044898986816,
+ "step": 575
+ },
+ {
+ "epoch": 8.0,
+ "grad_norm": 0.01837257295846939,
+ "learning_rate": 0.0006,
+ "loss": 5.240387916564941,
+ "step": 576
+ },
+ {
+ "epoch": 8.0,
+ "eval_loss": 5.423932075500488,
+ "eval_runtime": 58.0841,
+ "eval_samples_per_second": 42.042,
+ "eval_steps_per_second": 1.326,
+ "step": 576
+ },
+ {
+ "epoch": 8.013973799126637,
+ "grad_norm": 0.020953809842467308,
+ "learning_rate": 0.0006,
+ "loss": 5.4320783615112305,
+ "step": 577
+ },
+ {
+ "epoch": 8.027947598253276,
+ "grad_norm": 0.0227745920419693,
+ "learning_rate": 0.0006,
+ "loss": 5.310197830200195,
+ "step": 578
+ },
+ {
+ "epoch": 8.041921397379912,
+ "grad_norm": 0.021372603252530098,
+ "learning_rate": 0.0006,
+ "loss": 5.339468002319336,
+ "step": 579
+ },
+ {
+ "epoch": 8.055895196506551,
+ "grad_norm": 0.022401731461286545,
+ "learning_rate": 0.0006,
+ "loss": 5.279829978942871,
+ "step": 580
+ },
+ {
+ "epoch": 8.069868995633188,
+ "grad_norm": 0.023942379280924797,
+ "learning_rate": 0.0006,
+ "loss": 5.372323989868164,
+ "step": 581
+ },
+ {
+ "epoch": 8.083842794759825,
+ "grad_norm": 0.024380534887313843,
+ "learning_rate": 0.0006,
+ "loss": 5.253486633300781,
+ "step": 582
+ },
+ {
+ "epoch": 8.097816593886463,
+ "grad_norm": 0.03246625140309334,
+ "learning_rate": 0.0006,
+ "loss": 5.305774211883545,
+ "step": 583
+ },
+ {
+ "epoch": 8.1117903930131,
+ "grad_norm": 0.030399736016988754,
+ "learning_rate": 0.0006,
+ "loss": 5.370976448059082,
+ "step": 584
+ },
+ {
+ "epoch": 8.125764192139737,
+ "grad_norm": 0.027148913592100143,
+ "learning_rate": 0.0006,
+ "loss": 5.368010520935059,
+ "step": 585
+ },
+ {
+ "epoch": 8.139737991266376,
+ "grad_norm": 0.02925540879368782,
+ "learning_rate": 0.0006,
+ "loss": 5.303013801574707,
+ "step": 586
+ },
+ {
+ "epoch": 8.153711790393013,
+ "grad_norm": 0.02915453538298607,
+ "learning_rate": 0.0006,
+ "loss": 5.329667091369629,
+ "step": 587
+ },
+ {
+ "epoch": 8.167685589519651,
+ "grad_norm": 0.031732626259326935,
+ "learning_rate": 0.0006,
+ "loss": 5.3634934425354,
+ "step": 588
+ },
+ {
+ "epoch": 8.181659388646288,
+ "grad_norm": 0.036166731268167496,
+ "learning_rate": 0.0006,
+ "loss": 5.223980903625488,
+ "step": 589
+ },
+ {
+ "epoch": 8.195633187772925,
+ "grad_norm": 0.035654108971357346,
+ "learning_rate": 0.0006,
+ "loss": 5.361059188842773,
+ "step": 590
+ },
+ {
+ "epoch": 8.209606986899564,
+ "grad_norm": 0.0365324430167675,
+ "learning_rate": 0.0006,
+ "loss": 5.336151599884033,
+ "step": 591
+ },
+ {
+ "epoch": 8.2235807860262,
+ "grad_norm": 0.03578880429267883,
+ "learning_rate": 0.0006,
+ "loss": 5.434076309204102,
+ "step": 592
+ },
+ {
+ "epoch": 8.237554585152838,
+ "grad_norm": 0.035734012722969055,
+ "learning_rate": 0.0006,
+ "loss": 5.330893516540527,
+ "step": 593
+ },
+ {
+ "epoch": 8.251528384279476,
+ "grad_norm": 0.03196857124567032,
+ "learning_rate": 0.0006,
+ "loss": 5.306009769439697,
+ "step": 594
+ },
+ {
+ "epoch": 8.265502183406113,
+ "grad_norm": 0.032042086124420166,
+ "learning_rate": 0.0006,
+ "loss": 5.343267917633057,
+ "step": 595
+ },
+ {
+ "epoch": 8.279475982532752,
+ "grad_norm": 0.03160746395587921,
+ "learning_rate": 0.0006,
+ "loss": 5.2353715896606445,
+ "step": 596
+ },
+ {
+ "epoch": 8.293449781659389,
+ "grad_norm": 0.026689818128943443,
+ "learning_rate": 0.0006,
+ "loss": 5.254042625427246,
+ "step": 597
+ },
+ {
+ "epoch": 8.307423580786025,
+ "grad_norm": 0.02880188450217247,
+ "learning_rate": 0.0006,
+ "loss": 5.326833248138428,
+ "step": 598
+ },
+ {
+ "epoch": 8.321397379912664,
+ "grad_norm": 0.027516381815075874,
+ "learning_rate": 0.0006,
+ "loss": 5.342336177825928,
+ "step": 599
+ },
+ {
+ "epoch": 8.335371179039301,
+ "grad_norm": 0.027875030413269997,
+ "learning_rate": 0.0006,
+ "loss": 5.200719833374023,
+ "step": 600
+ },
+ {
+ "epoch": 8.34934497816594,
+ "grad_norm": 0.0268265288323164,
+ "learning_rate": 0.0006,
+ "loss": 5.249449729919434,
+ "step": 601
+ },
+ {
+ "epoch": 8.363318777292577,
+ "grad_norm": 0.024824608117341995,
+ "learning_rate": 0.0006,
+ "loss": 5.318952560424805,
+ "step": 602
+ },
+ {
+ "epoch": 8.377292576419213,
+ "grad_norm": 0.022990627214312553,
+ "learning_rate": 0.0006,
+ "loss": 5.244993209838867,
+ "step": 603
+ },
+ {
+ "epoch": 8.391266375545852,
+ "grad_norm": 0.022804604843258858,
+ "learning_rate": 0.0006,
+ "loss": 5.3236846923828125,
+ "step": 604
+ },
+ {
+ "epoch": 8.405240174672489,
+ "grad_norm": 0.02419872209429741,
+ "learning_rate": 0.0006,
+ "loss": 5.255486488342285,
+ "step": 605
+ },
+ {
+ "epoch": 8.419213973799126,
+ "grad_norm": 0.021952059119939804,
+ "learning_rate": 0.0006,
+ "loss": 5.3511552810668945,
+ "step": 606
+ },
+ {
+ "epoch": 8.433187772925764,
+ "grad_norm": 0.022375497967004776,
+ "learning_rate": 0.0006,
+ "loss": 5.294790267944336,
+ "step": 607
+ },
+ {
+ "epoch": 8.447161572052401,
+ "grad_norm": 0.019242815673351288,
+ "learning_rate": 0.0006,
+ "loss": 5.252618789672852,
+ "step": 608
+ },
+ {
+ "epoch": 8.46113537117904,
+ "grad_norm": 0.0173486340790987,
+ "learning_rate": 0.0006,
+ "loss": 5.30147647857666,
+ "step": 609
+ },
+ {
+ "epoch": 8.475109170305677,
+ "grad_norm": 0.01612604409456253,
+ "learning_rate": 0.0006,
+ "loss": 5.23960542678833,
+ "step": 610
+ },
+ {
+ "epoch": 8.489082969432314,
+ "grad_norm": 0.014293976128101349,
+ "learning_rate": 0.0006,
+ "loss": 5.119932651519775,
+ "step": 611
+ },
+ {
+ "epoch": 8.503056768558952,
+ "grad_norm": 0.014453536830842495,
+ "learning_rate": 0.0006,
+ "loss": 5.166794776916504,
+ "step": 612
+ },
+ {
+ "epoch": 8.51703056768559,
+ "grad_norm": 0.014685769565403461,
+ "learning_rate": 0.0006,
+ "loss": 5.264822959899902,
+ "step": 613
+ },
+ {
+ "epoch": 8.531004366812226,
+ "grad_norm": 0.01427740603685379,
+ "learning_rate": 0.0006,
+ "loss": 5.278433799743652,
+ "step": 614
+ },
+ {
+ "epoch": 8.544978165938865,
+ "grad_norm": 0.014864951372146606,
+ "learning_rate": 0.0006,
+ "loss": 5.214512348175049,
+ "step": 615
+ },
+ {
+ "epoch": 8.558951965065502,
+ "grad_norm": 0.01565164513885975,
+ "learning_rate": 0.0006,
+ "loss": 5.183066368103027,
+ "step": 616
+ },
+ {
+ "epoch": 8.57292576419214,
+ "grad_norm": 0.01779816672205925,
+ "learning_rate": 0.0006,
+ "loss": 5.163185119628906,
+ "step": 617
+ },
+ {
+ "epoch": 8.586899563318777,
+ "grad_norm": 0.017254827544093132,
+ "learning_rate": 0.0006,
+ "loss": 5.243555068969727,
+ "step": 618
+ },
+ {
+ "epoch": 8.600873362445414,
+ "grad_norm": 0.01650584116578102,
+ "learning_rate": 0.0006,
+ "loss": 5.234884738922119,
+ "step": 619
+ },
+ {
+ "epoch": 8.614847161572053,
+ "grad_norm": 0.017021115869283676,
+ "learning_rate": 0.0006,
+ "loss": 5.12271785736084,
+ "step": 620
+ },
+ {
+ "epoch": 8.62882096069869,
+ "grad_norm": 0.01773759163916111,
+ "learning_rate": 0.0006,
+ "loss": 5.242153167724609,
+ "step": 621
+ },
+ {
+ "epoch": 8.642794759825328,
+ "grad_norm": 0.015679042786359787,
+ "learning_rate": 0.0006,
+ "loss": 5.170779228210449,
+ "step": 622
+ },
+ {
+ "epoch": 8.656768558951965,
+ "grad_norm": 0.013760549947619438,
+ "learning_rate": 0.0006,
+ "loss": 5.238644599914551,
+ "step": 623
+ },
+ {
+ "epoch": 8.670742358078602,
+ "grad_norm": 0.014571522362530231,
+ "learning_rate": 0.0006,
+ "loss": 5.177056312561035,
+ "step": 624
+ },
+ {
+ "epoch": 8.68471615720524,
+ "grad_norm": 0.016209015622735023,
+ "learning_rate": 0.0006,
+ "loss": 5.163750648498535,
+ "step": 625
+ },
+ {
+ "epoch": 8.698689956331878,
+ "grad_norm": 0.016813941299915314,
+ "learning_rate": 0.0006,
+ "loss": 5.183428764343262,
+ "step": 626
+ },
+ {
+ "epoch": 8.712663755458514,
+ "grad_norm": 0.020985357463359833,
+ "learning_rate": 0.0006,
+ "loss": 5.229465007781982,
+ "step": 627
+ },
+ {
+ "epoch": 8.726637554585153,
+ "grad_norm": 0.02679632417857647,
+ "learning_rate": 0.0006,
+ "loss": 5.257368087768555,
+ "step": 628
+ },
+ {
+ "epoch": 8.74061135371179,
+ "grad_norm": 0.025756070390343666,
+ "learning_rate": 0.0006,
+ "loss": 5.253736972808838,
+ "step": 629
+ },
+ {
+ "epoch": 8.754585152838429,
+ "grad_norm": 0.02643490768969059,
+ "learning_rate": 0.0006,
+ "loss": 5.228633403778076,
+ "step": 630
+ },
+ {
+ "epoch": 8.768558951965066,
+ "grad_norm": 0.03303210437297821,
+ "learning_rate": 0.0006,
+ "loss": 5.163540363311768,
+ "step": 631
+ },
+ {
+ "epoch": 8.782532751091702,
+ "grad_norm": 0.030432431027293205,
+ "learning_rate": 0.0006,
+ "loss": 5.243169784545898,
+ "step": 632
+ },
+ {
+ "epoch": 8.796506550218341,
+ "grad_norm": 0.028631288558244705,
+ "learning_rate": 0.0006,
+ "loss": 5.166398048400879,
+ "step": 633
+ },
+ {
+ "epoch": 8.810480349344978,
+ "grad_norm": 0.026188310235738754,
+ "learning_rate": 0.0006,
+ "loss": 5.128122329711914,
+ "step": 634
+ },
+ {
+ "epoch": 8.824454148471617,
+ "grad_norm": 0.025308528915047646,
+ "learning_rate": 0.0006,
+ "loss": 5.211284637451172,
+ "step": 635
+ },
+ {
+ "epoch": 8.838427947598253,
+ "grad_norm": 0.020657729357481003,
+ "learning_rate": 0.0006,
+ "loss": 5.161575794219971,
+ "step": 636
+ },
+ {
+ "epoch": 8.85240174672489,
+ "grad_norm": 0.021901234984397888,
+ "learning_rate": 0.0006,
+ "loss": 5.20050048828125,
+ "step": 637
+ },
+ {
+ "epoch": 8.866375545851529,
+ "grad_norm": 0.021219318732619286,
+ "learning_rate": 0.0006,
+ "loss": 5.211699485778809,
+ "step": 638
+ },
+ {
+ "epoch": 8.880349344978166,
+ "grad_norm": 0.0200974028557539,
+ "learning_rate": 0.0006,
+ "loss": 5.092503547668457,
+ "step": 639
+ },
+ {
+ "epoch": 8.894323144104803,
+ "grad_norm": 0.023804882541298866,
+ "learning_rate": 0.0006,
+ "loss": 5.216068267822266,
+ "step": 640
+ },
+ {
+ "epoch": 8.908296943231441,
+ "grad_norm": 0.026088010519742966,
+ "learning_rate": 0.0006,
+ "loss": 5.155592918395996,
+ "step": 641
+ },
+ {
+ "epoch": 8.922270742358078,
+ "grad_norm": 0.02601276896893978,
+ "learning_rate": 0.0006,
+ "loss": 5.197238922119141,
+ "step": 642
+ },
+ {
+ "epoch": 8.936244541484717,
+ "grad_norm": 0.020387211814522743,
+ "learning_rate": 0.0006,
+ "loss": 5.1955437660217285,
+ "step": 643
+ },
+ {
+ "epoch": 8.950218340611354,
+ "grad_norm": 0.019214622676372528,
+ "learning_rate": 0.0006,
+ "loss": 5.206346035003662,
+ "step": 644
+ },
+ {
+ "epoch": 8.96419213973799,
+ "grad_norm": 0.019674314185976982,
+ "learning_rate": 0.0006,
+ "loss": 5.219857215881348,
+ "step": 645
+ },
+ {
+ "epoch": 8.97816593886463,
+ "grad_norm": 0.020318234339356422,
+ "learning_rate": 0.0006,
+ "loss": 5.191132545471191,
+ "step": 646
+ },
+ {
+ "epoch": 8.992139737991266,
+ "grad_norm": 0.021428676322102547,
+ "learning_rate": 0.0006,
+ "loss": 5.214695930480957,
+ "step": 647
+ },
+ {
+ "epoch": 9.0,
+ "grad_norm": 0.023235054686665535,
+ "learning_rate": 0.0006,
+ "loss": 5.229091644287109,
+ "step": 648
+ },
+ {
+ "epoch": 9.0,
+ "eval_loss": 5.289593696594238,
+ "eval_runtime": 56.9256,
+ "eval_samples_per_second": 42.898,
+ "eval_steps_per_second": 1.353,
+ "step": 648
+ },
+ {
+ "epoch": 9.013973799126637,
+ "grad_norm": 0.02320289984345436,
+ "learning_rate": 0.0006,
+ "loss": 4.983669281005859,
+ "step": 649
+ },
+ {
+ "epoch": 9.027947598253276,
+ "grad_norm": 0.018521282821893692,
+ "learning_rate": 0.0006,
+ "loss": 5.200575828552246,
+ "step": 650
+ },
+ {
+ "epoch": 9.041921397379912,
+ "grad_norm": 0.0180169939994812,
+ "learning_rate": 0.0006,
+ "loss": 5.148033618927002,
+ "step": 651
+ },
+ {
+ "epoch": 9.055895196506551,
+ "grad_norm": 0.019969860091805458,
+ "learning_rate": 0.0006,
+ "loss": 5.099125385284424,
+ "step": 652
+ },
+ {
+ "epoch": 9.069868995633188,
+ "grad_norm": 0.01643305830657482,
+ "learning_rate": 0.0006,
+ "loss": 5.092347145080566,
+ "step": 653
+ },
+ {
+ "epoch": 9.083842794759825,
+ "grad_norm": 0.016022363677620888,
+ "learning_rate": 0.0006,
+ "loss": 5.129279613494873,
+ "step": 654
+ },
+ {
+ "epoch": 9.097816593886463,
+ "grad_norm": 0.01657041162252426,
+ "learning_rate": 0.0006,
+ "loss": 5.159562587738037,
+ "step": 655
+ },
+ {
+ "epoch": 9.1117903930131,
+ "grad_norm": 0.017924228683114052,
+ "learning_rate": 0.0006,
+ "loss": 5.165497779846191,
+ "step": 656
+ },
+ {
+ "epoch": 9.125764192139737,
+ "grad_norm": 0.021592361852526665,
+ "learning_rate": 0.0006,
+ "loss": 5.204775810241699,
+ "step": 657
+ },
+ {
+ "epoch": 9.139737991266376,
+ "grad_norm": 0.026924120262265205,
+ "learning_rate": 0.0006,
+ "loss": 5.128296852111816,
+ "step": 658
+ },
+ {
+ "epoch": 9.153711790393013,
+ "grad_norm": 0.023156899958848953,
+ "learning_rate": 0.0006,
+ "loss": 5.1255340576171875,
+ "step": 659
+ },
+ {
+ "epoch": 9.167685589519651,
+ "grad_norm": 0.020670367404818535,
+ "learning_rate": 0.0006,
+ "loss": 5.128479957580566,
+ "step": 660
+ },
+ {
+ "epoch": 9.181659388646288,
+ "grad_norm": 0.025447173044085503,
+ "learning_rate": 0.0006,
+ "loss": 5.154821872711182,
+ "step": 661
+ },
+ {
+ "epoch": 9.195633187772925,
+ "grad_norm": 0.027940068393945694,
+ "learning_rate": 0.0006,
+ "loss": 5.143466472625732,
+ "step": 662
+ },
+ {
+ "epoch": 9.209606986899564,
+ "grad_norm": 0.026847844943404198,
+ "learning_rate": 0.0006,
+ "loss": 5.045665740966797,
+ "step": 663
+ },
+ {
+ "epoch": 9.2235807860262,
+ "grad_norm": 0.02340601570904255,
+ "learning_rate": 0.0006,
+ "loss": 5.21933126449585,
+ "step": 664
+ },
+ {
+ "epoch": 9.237554585152838,
+ "grad_norm": 0.02340371161699295,
+ "learning_rate": 0.0006,
+ "loss": 4.942949295043945,
+ "step": 665
+ },
+ {
+ "epoch": 9.251528384279476,
+ "grad_norm": 0.02221992425620556,
+ "learning_rate": 0.0006,
+ "loss": 5.197099208831787,
+ "step": 666
+ },
+ {
+ "epoch": 9.265502183406113,
+ "grad_norm": 0.023508677259087563,
+ "learning_rate": 0.0006,
+ "loss": 5.022897720336914,
+ "step": 667
+ },
+ {
+ "epoch": 9.279475982532752,
+ "grad_norm": 0.026120394468307495,
+ "learning_rate": 0.0006,
+ "loss": 5.087725639343262,
+ "step": 668
+ },
+ {
+ "epoch": 9.293449781659389,
+ "grad_norm": 0.027273228392004967,
+ "learning_rate": 0.0006,
+ "loss": 5.090963840484619,
+ "step": 669
+ },
+ {
+ "epoch": 9.307423580786025,
+ "grad_norm": 0.03241586685180664,
+ "learning_rate": 0.0006,
+ "loss": 5.047842979431152,
+ "step": 670
+ },
+ {
+ "epoch": 9.321397379912664,
+ "grad_norm": 0.030245667323470116,
+ "learning_rate": 0.0006,
+ "loss": 5.104803085327148,
+ "step": 671
+ },
+ {
+ "epoch": 9.335371179039301,
+ "grad_norm": 0.027698364108800888,
+ "learning_rate": 0.0006,
+ "loss": 5.0644097328186035,
+ "step": 672
+ },
+ {
+ "epoch": 9.34934497816594,
+ "grad_norm": 0.029692554846405983,
+ "learning_rate": 0.0006,
+ "loss": 5.111942291259766,
+ "step": 673
+ },
+ {
+ "epoch": 9.363318777292577,
+ "grad_norm": 0.03328656405210495,
+ "learning_rate": 0.0006,
+ "loss": 5.1574554443359375,
+ "step": 674
+ },
+ {
+ "epoch": 9.377292576419213,
+ "grad_norm": 0.031938180327415466,
+ "learning_rate": 0.0006,
+ "loss": 5.190371990203857,
+ "step": 675
+ },
+ {
+ "epoch": 9.391266375545852,
+ "grad_norm": 0.024858945980668068,
+ "learning_rate": 0.0006,
+ "loss": 5.121420860290527,
+ "step": 676
+ },
+ {
+ "epoch": 9.405240174672489,
+ "grad_norm": 0.023033203557133675,
+ "learning_rate": 0.0006,
+ "loss": 5.0974345207214355,
+ "step": 677
+ },
+ {
+ "epoch": 9.419213973799126,
+ "grad_norm": 0.022083545103669167,
+ "learning_rate": 0.0006,
+ "loss": 5.046319961547852,
+ "step": 678
+ },
+ {
+ "epoch": 9.433187772925764,
+ "grad_norm": 0.0228437427431345,
+ "learning_rate": 0.0006,
+ "loss": 5.119963645935059,
+ "step": 679
+ },
+ {
+ "epoch": 9.447161572052401,
+ "grad_norm": 0.0242351982742548,
+ "learning_rate": 0.0006,
+ "loss": 5.0444560050964355,
+ "step": 680
+ },
+ {
+ "epoch": 9.46113537117904,
+ "grad_norm": 0.02401185780763626,
+ "learning_rate": 0.0006,
+ "loss": 5.063238620758057,
+ "step": 681
+ },
+ {
+ "epoch": 9.475109170305677,
+ "grad_norm": 0.02560959942638874,
+ "learning_rate": 0.0006,
+ "loss": 5.035224914550781,
+ "step": 682
+ },
+ {
+ "epoch": 9.489082969432314,
+ "grad_norm": 0.024985626339912415,
+ "learning_rate": 0.0006,
+ "loss": 5.049284934997559,
+ "step": 683
+ },
+ {
+ "epoch": 9.503056768558952,
+ "grad_norm": 0.023178115487098694,
+ "learning_rate": 0.0006,
+ "loss": 5.2266621589660645,
+ "step": 684
+ },
+ {
+ "epoch": 9.51703056768559,
+ "grad_norm": 0.0215512253344059,
+ "learning_rate": 0.0006,
+ "loss": 5.1875715255737305,
+ "step": 685
+ },
+ {
+ "epoch": 9.531004366812226,
+ "grad_norm": 0.01662837713956833,
+ "learning_rate": 0.0006,
+ "loss": 5.137012481689453,
+ "step": 686
+ },
+ {
+ "epoch": 9.544978165938865,
+ "grad_norm": 0.01952764391899109,
+ "learning_rate": 0.0006,
+ "loss": 5.072500228881836,
+ "step": 687
+ },
+ {
+ "epoch": 9.558951965065502,
+ "grad_norm": 0.017521077767014503,
+ "learning_rate": 0.0006,
+ "loss": 5.045711994171143,
+ "step": 688
+ },
+ {
+ "epoch": 9.57292576419214,
+ "grad_norm": 0.017355265095829964,
+ "learning_rate": 0.0006,
+ "loss": 5.02881383895874,
+ "step": 689
+ },
+ {
+ "epoch": 9.586899563318777,
+ "grad_norm": 0.015587719157338142,
+ "learning_rate": 0.0006,
+ "loss": 5.132787704467773,
+ "step": 690
+ },
+ {
+ "epoch": 9.600873362445414,
+ "grad_norm": 0.017469845712184906,
+ "learning_rate": 0.0006,
+ "loss": 5.105447292327881,
+ "step": 691
+ },
+ {
+ "epoch": 9.614847161572053,
+ "grad_norm": 0.019630100578069687,
+ "learning_rate": 0.0006,
+ "loss": 5.032464981079102,
+ "step": 692
+ },
+ {
+ "epoch": 9.62882096069869,
+ "grad_norm": 0.01784994639456272,
+ "learning_rate": 0.0006,
+ "loss": 5.081345558166504,
+ "step": 693
+ },
+ {
+ "epoch": 9.642794759825328,
+ "grad_norm": 0.01872754842042923,
+ "learning_rate": 0.0006,
+ "loss": 4.9772114753723145,
+ "step": 694
+ },
+ {
+ "epoch": 9.656768558951965,
+ "grad_norm": 0.01999032311141491,
+ "learning_rate": 0.0006,
+ "loss": 5.089740753173828,
+ "step": 695
+ },
+ {
+ "epoch": 9.670742358078602,
+ "grad_norm": 0.01710767112672329,
+ "learning_rate": 0.0006,
+ "loss": 4.9869890213012695,
+ "step": 696
+ },
+ {
+ "epoch": 9.68471615720524,
+ "grad_norm": 0.015144161880016327,
+ "learning_rate": 0.0006,
+ "loss": 5.022564888000488,
+ "step": 697
+ },
+ {
+ "epoch": 9.698689956331878,
+ "grad_norm": 0.01618543453514576,
+ "learning_rate": 0.0006,
+ "loss": 5.035921573638916,
+ "step": 698
+ },
+ {
+ "epoch": 9.712663755458514,
+ "grad_norm": 0.014653387479484081,
+ "learning_rate": 0.0006,
+ "loss": 4.914456367492676,
+ "step": 699
+ },
+ {
+ "epoch": 9.726637554585153,
+ "grad_norm": 0.013434977270662785,
+ "learning_rate": 0.0006,
+ "loss": 5.063764572143555,
+ "step": 700
+ },
+ {
+ "epoch": 9.74061135371179,
+ "grad_norm": 0.015128864906728268,
+ "learning_rate": 0.0006,
+ "loss": 5.052367687225342,
+ "step": 701
+ },
+ {
+ "epoch": 9.754585152838429,
+ "grad_norm": 0.014210768043994904,
+ "learning_rate": 0.0006,
+ "loss": 5.016489028930664,
+ "step": 702
+ },
+ {
+ "epoch": 9.768558951965066,
+ "grad_norm": 0.015191680751740932,
+ "learning_rate": 0.0006,
+ "loss": 4.990176200866699,
+ "step": 703
+ },
+ {
+ "epoch": 9.782532751091702,
+ "grad_norm": 0.018285127356648445,
+ "learning_rate": 0.0006,
+ "loss": 5.015164375305176,
+ "step": 704
+ },
+ {
+ "epoch": 9.796506550218341,
+ "grad_norm": 0.0204079058021307,
+ "learning_rate": 0.0006,
+ "loss": 5.091142654418945,
+ "step": 705
+ },
+ {
+ "epoch": 9.810480349344978,
+ "grad_norm": 0.01980586163699627,
+ "learning_rate": 0.0006,
+ "loss": 5.03350830078125,
+ "step": 706
+ },
+ {
+ "epoch": 9.824454148471617,
+ "grad_norm": 0.02048729732632637,
+ "learning_rate": 0.0006,
+ "loss": 5.03611421585083,
+ "step": 707
+ },
+ {
+ "epoch": 9.838427947598253,
+ "grad_norm": 0.02058163844048977,
+ "learning_rate": 0.0006,
+ "loss": 4.9613447189331055,
+ "step": 708
+ },
+ {
+ "epoch": 9.85240174672489,
+ "grad_norm": 0.018004924058914185,
+ "learning_rate": 0.0006,
+ "loss": 4.966541767120361,
+ "step": 709
+ },
+ {
+ "epoch": 9.866375545851529,
+ "grad_norm": 0.0198152307420969,
+ "learning_rate": 0.0006,
+ "loss": 5.01750373840332,
+ "step": 710
+ },
+ {
+ "epoch": 9.880349344978166,
+ "grad_norm": 0.018532969057559967,
+ "learning_rate": 0.0006,
+ "loss": 5.097424507141113,
+ "step": 711
+ },
+ {
+ "epoch": 9.894323144104803,
+ "grad_norm": 0.018939971923828125,
+ "learning_rate": 0.0006,
+ "loss": 5.177223205566406,
+ "step": 712
+ },
+ {
+ "epoch": 9.908296943231441,
+ "grad_norm": 0.017969170585274696,
+ "learning_rate": 0.0006,
+ "loss": 5.118013858795166,
+ "step": 713
+ },
+ {
+ "epoch": 9.922270742358078,
+ "grad_norm": 0.018021270632743835,
+ "learning_rate": 0.0006,
+ "loss": 5.034787178039551,
+ "step": 714
+ },
+ {
+ "epoch": 9.936244541484717,
+ "grad_norm": 0.017009710893034935,
+ "learning_rate": 0.0006,
+ "loss": 5.003920555114746,
+ "step": 715
+ },
+ {
+ "epoch": 9.950218340611354,
+ "grad_norm": 0.017722049728035927,
+ "learning_rate": 0.0006,
+ "loss": 5.031939506530762,
+ "step": 716
+ },
+ {
+ "epoch": 9.96419213973799,
+ "grad_norm": 0.017134087160229683,
+ "learning_rate": 0.0006,
+ "loss": 5.093474388122559,
+ "step": 717
+ },
+ {
+ "epoch": 9.97816593886463,
+ "grad_norm": 0.018511613830924034,
+ "learning_rate": 0.0006,
+ "loss": 5.078604698181152,
+ "step": 718
+ },
+ {
+ "epoch": 9.992139737991266,
+ "grad_norm": 0.021206334233283997,
+ "learning_rate": 0.0006,
+ "loss": 5.052572727203369,
+ "step": 719
+ },
+ {
+ "epoch": 10.0,
+ "grad_norm": 0.02239062264561653,
+ "learning_rate": 0.0006,
+ "loss": 5.063547134399414,
+ "step": 720
+ },
+ {
+ "epoch": 10.0,
+ "eval_loss": 5.173044681549072,
+ "eval_runtime": 56.7287,
+ "eval_samples_per_second": 43.047,
+ "eval_steps_per_second": 1.357,
+ "step": 720
+ },
+ {
+ "epoch": 10.013973799126637,
+ "grad_norm": 0.020757969468832016,
+ "learning_rate": 0.0006,
+ "loss": 5.045648574829102,
+ "step": 721
+ },
+ {
+ "epoch": 10.027947598253276,
+ "grad_norm": 0.022697702050209045,
+ "learning_rate": 0.0006,
+ "loss": 4.993833065032959,
+ "step": 722
+ },
+ {
+ "epoch": 10.041921397379912,
+ "grad_norm": 0.02434331737458706,
+ "learning_rate": 0.0006,
+ "loss": 5.073219299316406,
+ "step": 723
+ },
+ {
+ "epoch": 10.055895196506551,
+ "grad_norm": 0.025499660521745682,
+ "learning_rate": 0.0006,
+ "loss": 5.019072532653809,
+ "step": 724
+ },
+ {
+ "epoch": 10.069868995633188,
+ "grad_norm": 0.029281053692102432,
+ "learning_rate": 0.0006,
+ "loss": 4.959851264953613,
+ "step": 725
+ },
+ {
+ "epoch": 10.083842794759825,
+ "grad_norm": 0.033878110349178314,
+ "learning_rate": 0.0006,
+ "loss": 5.132717609405518,
+ "step": 726
+ },
+ {
+ "epoch": 10.097816593886463,
+ "grad_norm": 0.03385764732956886,
+ "learning_rate": 0.0006,
+ "loss": 5.0482892990112305,
+ "step": 727
+ },
+ {
+ "epoch": 10.1117903930131,
+ "grad_norm": 0.03466648980975151,
+ "learning_rate": 0.0006,
+ "loss": 4.946855545043945,
+ "step": 728
+ },
+ {
+ "epoch": 10.125764192139737,
+ "grad_norm": 0.028546305373311043,
+ "learning_rate": 0.0006,
+ "loss": 5.027774333953857,
+ "step": 729
+ },
+ {
+ "epoch": 10.139737991266376,
+ "grad_norm": 0.030575869604945183,
+ "learning_rate": 0.0006,
+ "loss": 4.9768476486206055,
+ "step": 730
+ },
+ {
+ "epoch": 10.153711790393013,
+ "grad_norm": 0.03614303469657898,
+ "learning_rate": 0.0006,
+ "loss": 5.078113555908203,
+ "step": 731
+ },
+ {
+ "epoch": 10.167685589519651,
+ "grad_norm": 0.040991660207509995,
+ "learning_rate": 0.0006,
+ "loss": 5.033185958862305,
+ "step": 732
+ },
+ {
+ "epoch": 10.181659388646288,
+ "grad_norm": 0.04551699757575989,
+ "learning_rate": 0.0006,
+ "loss": 5.0820746421813965,
+ "step": 733
+ },
+ {
+ "epoch": 10.195633187772925,
+ "grad_norm": 0.040193989872932434,
+ "learning_rate": 0.0006,
+ "loss": 5.059700965881348,
+ "step": 734
+ },
+ {
+ "epoch": 10.209606986899564,
+ "grad_norm": 0.035851605236530304,
+ "learning_rate": 0.0006,
+ "loss": 5.142467498779297,
+ "step": 735
+ },
+ {
+ "epoch": 10.2235807860262,
+ "grad_norm": 0.034200169146060944,
+ "learning_rate": 0.0006,
+ "loss": 5.072312355041504,
+ "step": 736
+ },
+ {
+ "epoch": 10.237554585152838,
+ "grad_norm": 0.033621110022068024,
+ "learning_rate": 0.0006,
+ "loss": 5.08037805557251,
+ "step": 737
+ },
+ {
+ "epoch": 10.251528384279476,
+ "grad_norm": 0.037372369319200516,
+ "learning_rate": 0.0006,
+ "loss": 4.998425483703613,
+ "step": 738
+ },
+ {
+ "epoch": 10.265502183406113,
+ "grad_norm": 0.037044707685709,
+ "learning_rate": 0.0006,
+ "loss": 4.923993110656738,
+ "step": 739
+ },
+ {
+ "epoch": 10.279475982532752,
+ "grad_norm": 0.030454808846116066,
+ "learning_rate": 0.0006,
+ "loss": 5.0889997482299805,
+ "step": 740
+ },
+ {
+ "epoch": 10.293449781659389,
+ "grad_norm": 0.030969638377428055,
+ "learning_rate": 0.0006,
+ "loss": 4.984112739562988,
+ "step": 741
+ },
+ {
+ "epoch": 10.307423580786025,
+ "grad_norm": 0.0276983380317688,
+ "learning_rate": 0.0006,
+ "loss": 4.955412864685059,
+ "step": 742
+ },
+ {
+ "epoch": 10.321397379912664,
+ "grad_norm": 0.02716052532196045,
+ "learning_rate": 0.0006,
+ "loss": 5.0098876953125,
+ "step": 743
+ },
+ {
+ "epoch": 10.335371179039301,
+ "grad_norm": 0.02403552085161209,
+ "learning_rate": 0.0006,
+ "loss": 5.165160655975342,
+ "step": 744
+ },
+ {
+ "epoch": 10.34934497816594,
+ "grad_norm": 0.0257862601429224,
+ "learning_rate": 0.0006,
+ "loss": 5.001547813415527,
+ "step": 745
+ },
+ {
+ "epoch": 10.363318777292577,
+ "grad_norm": 0.02300378680229187,
+ "learning_rate": 0.0006,
+ "loss": 5.038888931274414,
+ "step": 746
+ },
+ {
+ "epoch": 10.377292576419213,
+ "grad_norm": 0.019766854122281075,
+ "learning_rate": 0.0006,
+ "loss": 4.924291610717773,
+ "step": 747
+ },
+ {
+ "epoch": 10.391266375545852,
+ "grad_norm": 0.01888395845890045,
+ "learning_rate": 0.0006,
+ "loss": 5.059971332550049,
+ "step": 748
+ },
+ {
+ "epoch": 10.405240174672489,
+ "grad_norm": 0.015577499754726887,
+ "learning_rate": 0.0006,
+ "loss": 4.987686634063721,
+ "step": 749
+ },
+ {
+ "epoch": 10.419213973799126,
+ "grad_norm": 0.016012346372008324,
+ "learning_rate": 0.0006,
+ "loss": 5.015157699584961,
+ "step": 750
+ },
+ {
+ "epoch": 10.433187772925764,
+ "grad_norm": 0.016277998685836792,
+ "learning_rate": 0.0006,
+ "loss": 5.004931449890137,
+ "step": 751
+ },
+ {
+ "epoch": 10.447161572052401,
+ "grad_norm": 0.017504561692476273,
+ "learning_rate": 0.0006,
+ "loss": 4.902827739715576,
+ "step": 752
+ },
+ {
+ "epoch": 10.46113537117904,
+ "grad_norm": 0.01569022797048092,
+ "learning_rate": 0.0006,
+ "loss": 4.954863548278809,
+ "step": 753
+ },
+ {
+ "epoch": 10.475109170305677,
+ "grad_norm": 0.014114422723650932,
+ "learning_rate": 0.0006,
+ "loss": 5.00656795501709,
+ "step": 754
+ },
+ {
+ "epoch": 10.489082969432314,
+ "grad_norm": 0.01573832333087921,
+ "learning_rate": 0.0006,
+ "loss": 4.978353023529053,
+ "step": 755
+ },
+ {
+ "epoch": 10.503056768558952,
+ "grad_norm": 0.016480036079883575,
+ "learning_rate": 0.0006,
+ "loss": 4.953327655792236,
+ "step": 756
+ },
+ {
+ "epoch": 10.51703056768559,
+ "grad_norm": 0.014008025638759136,
+ "learning_rate": 0.0006,
+ "loss": 5.050840377807617,
+ "step": 757
+ },
+ {
+ "epoch": 10.531004366812226,
+ "grad_norm": 0.013147998601198196,
+ "learning_rate": 0.0006,
+ "loss": 4.974964141845703,
+ "step": 758
+ },
+ {
+ "epoch": 10.544978165938865,
+ "grad_norm": 0.013807603158056736,
+ "learning_rate": 0.0006,
+ "loss": 4.927907943725586,
+ "step": 759
+ },
+ {
+ "epoch": 10.558951965065502,
+ "grad_norm": 0.013555224984884262,
+ "learning_rate": 0.0006,
+ "loss": 4.980690956115723,
+ "step": 760
+ },
+ {
+ "epoch": 10.57292576419214,
+ "grad_norm": 0.013045408762991428,
+ "learning_rate": 0.0006,
+ "loss": 4.9436116218566895,
+ "step": 761
+ },
+ {
+ "epoch": 10.586899563318777,
+ "grad_norm": 0.012858862057328224,
+ "learning_rate": 0.0006,
+ "loss": 4.9595770835876465,
+ "step": 762
+ },
+ {
+ "epoch": 10.600873362445414,
+ "grad_norm": 0.01468253880739212,
+ "learning_rate": 0.0006,
+ "loss": 4.90725040435791,
+ "step": 763
+ },
+ {
+ "epoch": 10.614847161572053,
+ "grad_norm": 0.013224679045379162,
+ "learning_rate": 0.0006,
+ "loss": 4.877615928649902,
+ "step": 764
+ },
+ {
+ "epoch": 10.62882096069869,
+ "grad_norm": 0.013555348850786686,
+ "learning_rate": 0.0006,
+ "loss": 4.9456610679626465,
+ "step": 765
+ },
+ {
+ "epoch": 10.642794759825328,
+ "grad_norm": 0.014387160539627075,
+ "learning_rate": 0.0006,
+ "loss": 4.9233717918396,
+ "step": 766
+ },
+ {
+ "epoch": 10.656768558951965,
+ "grad_norm": 0.014354088343679905,
+ "learning_rate": 0.0006,
+ "loss": 4.983511924743652,
+ "step": 767
+ },
+ {
+ "epoch": 10.670742358078602,
+ "grad_norm": 0.01332700252532959,
+ "learning_rate": 0.0006,
+ "loss": 4.934416770935059,
+ "step": 768
+ },
+ {
+ "epoch": 10.68471615720524,
+ "grad_norm": 0.011543313041329384,
+ "learning_rate": 0.0006,
+ "loss": 4.96388053894043,
+ "step": 769
+ },
+ {
+ "epoch": 10.698689956331878,
+ "grad_norm": 0.012330878525972366,
+ "learning_rate": 0.0006,
+ "loss": 4.946396350860596,
+ "step": 770
+ },
+ {
+ "epoch": 10.712663755458514,
+ "grad_norm": 0.01379779726266861,
+ "learning_rate": 0.0006,
+ "loss": 4.981019020080566,
+ "step": 771
+ },
+ {
+ "epoch": 10.726637554585153,
+ "grad_norm": 0.014612431637942791,
+ "learning_rate": 0.0006,
+ "loss": 4.894567966461182,
+ "step": 772
+ },
+ {
+ "epoch": 10.74061135371179,
+ "grad_norm": 0.013062899932265282,
+ "learning_rate": 0.0006,
+ "loss": 5.033038139343262,
+ "step": 773
+ },
+ {
+ "epoch": 10.754585152838429,
+ "grad_norm": 0.013898049481213093,
+ "learning_rate": 0.0006,
+ "loss": 4.823636054992676,
+ "step": 774
+ },
+ {
+ "epoch": 10.768558951965066,
+ "grad_norm": 0.01502019353210926,
+ "learning_rate": 0.0006,
+ "loss": 5.020156383514404,
+ "step": 775
+ },
+ {
+ "epoch": 10.782532751091702,
+ "grad_norm": 0.013810204342007637,
+ "learning_rate": 0.0006,
+ "loss": 4.935647010803223,
+ "step": 776
+ },
+ {
+ "epoch": 10.796506550218341,
+ "grad_norm": 0.013616513460874557,
+ "learning_rate": 0.0006,
+ "loss": 4.901486396789551,
+ "step": 777
+ },
+ {
+ "epoch": 10.810480349344978,
+ "grad_norm": 0.013601024635136127,
+ "learning_rate": 0.0006,
+ "loss": 4.931873321533203,
+ "step": 778
+ },
+ {
+ "epoch": 10.824454148471617,
+ "grad_norm": 0.013668350875377655,
+ "learning_rate": 0.0006,
+ "loss": 4.937256813049316,
+ "step": 779
+ },
+ {
+ "epoch": 10.838427947598253,
+ "grad_norm": 0.01581823080778122,
+ "learning_rate": 0.0006,
+ "loss": 4.9611406326293945,
+ "step": 780
+ },
+ {
+ "epoch": 10.85240174672489,
+ "grad_norm": 0.019478484988212585,
+ "learning_rate": 0.0006,
+ "loss": 4.9212236404418945,
+ "step": 781
+ },
+ {
+ "epoch": 10.866375545851529,
+ "grad_norm": 0.026921333745121956,
+ "learning_rate": 0.0006,
+ "loss": 4.88501501083374,
+ "step": 782
+ },
+ {
+ "epoch": 10.880349344978166,
+ "grad_norm": 0.03155914694070816,
+ "learning_rate": 0.0006,
+ "loss": 4.989967346191406,
+ "step": 783
+ },
+ {
+ "epoch": 10.894323144104803,
+ "grad_norm": 0.029689429327845573,
+ "learning_rate": 0.0006,
+ "loss": 4.936000823974609,
+ "step": 784
+ },
+ {
+ "epoch": 10.908296943231441,
+ "grad_norm": 0.026626063510775566,
+ "learning_rate": 0.0006,
+ "loss": 4.984785079956055,
+ "step": 785
+ },
+ {
+ "epoch": 10.922270742358078,
+ "grad_norm": 0.026168212294578552,
+ "learning_rate": 0.0006,
+ "loss": 4.908282279968262,
+ "step": 786
+ },
+ {
+ "epoch": 10.936244541484717,
+ "grad_norm": 0.02828158251941204,
+ "learning_rate": 0.0006,
+ "loss": 4.926390647888184,
+ "step": 787
+ },
+ {
+ "epoch": 10.950218340611354,
+ "grad_norm": 0.02649078331887722,
+ "learning_rate": 0.0006,
+ "loss": 5.00400972366333,
+ "step": 788
+ },
+ {
+ "epoch": 10.96419213973799,
+ "grad_norm": 0.0256856270134449,
+ "learning_rate": 0.0006,
+ "loss": 5.0317487716674805,
+ "step": 789
+ },
+ {
+ "epoch": 10.97816593886463,
+ "grad_norm": 0.024728883057832718,
+ "learning_rate": 0.0006,
+ "loss": 4.9299635887146,
+ "step": 790
+ },
+ {
+ "epoch": 10.992139737991266,
+ "grad_norm": 0.026241116225719452,
+ "learning_rate": 0.0006,
+ "loss": 4.947319984436035,
+ "step": 791
+ },
+ {
+ "epoch": 11.0,
+ "grad_norm": 0.023526743054389954,
+ "learning_rate": 0.0006,
+ "loss": 4.94429349899292,
+ "step": 792
+ },
+ {
+ "epoch": 11.0,
+ "eval_loss": 5.104345798492432,
+ "eval_runtime": 57.0955,
+ "eval_samples_per_second": 42.77,
+ "eval_steps_per_second": 1.349,
+ "step": 792
+ },
+ {
+ "epoch": 11.013973799126637,
+ "grad_norm": 0.0216788612306118,
+ "learning_rate": 0.0006,
+ "loss": 4.881950855255127,
+ "step": 793
+ },
+ {
+ "epoch": 11.027947598253276,
+ "grad_norm": 0.01830463856458664,
+ "learning_rate": 0.0006,
+ "loss": 4.971016883850098,
+ "step": 794
+ },
+ {
+ "epoch": 11.041921397379912,
+ "grad_norm": 0.02204521745443344,
+ "learning_rate": 0.0006,
+ "loss": 4.894623756408691,
+ "step": 795
+ },
+ {
+ "epoch": 11.055895196506551,
+ "grad_norm": 0.0210255216807127,
+ "learning_rate": 0.0006,
+ "loss": 4.834009170532227,
+ "step": 796
+ },
+ {
+ "epoch": 11.069868995633188,
+ "grad_norm": 0.019797448068857193,
+ "learning_rate": 0.0006,
+ "loss": 5.086714744567871,
+ "step": 797
+ },
+ {
+ "epoch": 11.083842794759825,
+ "grad_norm": 0.02036641724407673,
+ "learning_rate": 0.0006,
+ "loss": 4.920056343078613,
+ "step": 798
+ },
+ {
+ "epoch": 11.097816593886463,
+ "grad_norm": 0.01966066285967827,
+ "learning_rate": 0.0006,
+ "loss": 4.879859447479248,
+ "step": 799
+ },
+ {
+ "epoch": 11.1117903930131,
+ "grad_norm": 0.0202149897813797,
+ "learning_rate": 0.0006,
+ "loss": 4.933640480041504,
+ "step": 800
+ },
+ {
+ "epoch": 11.125764192139737,
+ "grad_norm": 0.01893039606511593,
+ "learning_rate": 0.0006,
+ "loss": 4.901970863342285,
+ "step": 801
+ },
+ {
+ "epoch": 11.139737991266376,
+ "grad_norm": 0.016040155664086342,
+ "learning_rate": 0.0006,
+ "loss": 4.905491828918457,
+ "step": 802
+ },
+ {
+ "epoch": 11.153711790393013,
+ "grad_norm": 0.01720350608229637,
+ "learning_rate": 0.0006,
+ "loss": 4.798816680908203,
+ "step": 803
+ },
+ {
+ "epoch": 11.167685589519651,
+ "grad_norm": 0.017002852633595467,
+ "learning_rate": 0.0006,
+ "loss": 4.851859092712402,
+ "step": 804
+ },
+ {
+ "epoch": 11.181659388646288,
+ "grad_norm": 0.015594680793583393,
+ "learning_rate": 0.0006,
+ "loss": 4.90328311920166,
+ "step": 805
+ },
+ {
+ "epoch": 11.195633187772925,
+ "grad_norm": 0.014904462732374668,
+ "learning_rate": 0.0006,
+ "loss": 4.904215335845947,
+ "step": 806
+ },
+ {
+ "epoch": 11.209606986899564,
+ "grad_norm": 0.01366228237748146,
+ "learning_rate": 0.0006,
+ "loss": 4.9813055992126465,
+ "step": 807
+ },
+ {
+ "epoch": 11.2235807860262,
+ "grad_norm": 0.014719638973474503,
+ "learning_rate": 0.0006,
+ "loss": 4.927306175231934,
+ "step": 808
+ },
+ {
+ "epoch": 11.237554585152838,
+ "grad_norm": 0.015016036108136177,
+ "learning_rate": 0.0006,
+ "loss": 4.874215126037598,
+ "step": 809
+ },
+ {
+ "epoch": 11.251528384279476,
+ "grad_norm": 0.014925424009561539,
+ "learning_rate": 0.0006,
+ "loss": 4.838257789611816,
+ "step": 810
+ },
+ {
+ "epoch": 11.265502183406113,
+ "grad_norm": 0.014335823245346546,
+ "learning_rate": 0.0006,
+ "loss": 4.940608024597168,
+ "step": 811
+ },
+ {
+ "epoch": 11.279475982532752,
+ "grad_norm": 0.014987512491643429,
+ "learning_rate": 0.0006,
+ "loss": 4.831531524658203,
+ "step": 812
+ },
+ {
+ "epoch": 11.293449781659389,
+ "grad_norm": 0.016087457537651062,
+ "learning_rate": 0.0006,
+ "loss": 4.863227367401123,
+ "step": 813
+ },
+ {
+ "epoch": 11.307423580786025,
+ "grad_norm": 0.01671300269663334,
+ "learning_rate": 0.0006,
+ "loss": 4.8535661697387695,
+ "step": 814
+ },
+ {
+ "epoch": 11.321397379912664,
+ "grad_norm": 0.015229142270982265,
+ "learning_rate": 0.0006,
+ "loss": 4.948635101318359,
+ "step": 815
+ },
+ {
+ "epoch": 11.335371179039301,
+ "grad_norm": 0.01532587967813015,
+ "learning_rate": 0.0006,
+ "loss": 4.9035797119140625,
+ "step": 816
+ },
+ {
+ "epoch": 11.34934497816594,
+ "grad_norm": 0.015715476125478745,
+ "learning_rate": 0.0006,
+ "loss": 4.894709587097168,
+ "step": 817
+ },
+ {
+ "epoch": 11.363318777292577,
+ "grad_norm": 0.014156977646052837,
+ "learning_rate": 0.0006,
+ "loss": 4.960206985473633,
+ "step": 818
+ },
+ {
+ "epoch": 11.377292576419213,
+ "grad_norm": 0.017076566815376282,
+ "learning_rate": 0.0006,
+ "loss": 5.005850791931152,
+ "step": 819
+ },
+ {
+ "epoch": 11.391266375545852,
+ "grad_norm": 0.0172084029763937,
+ "learning_rate": 0.0006,
+ "loss": 4.903355598449707,
+ "step": 820
+ },
+ {
+ "epoch": 11.405240174672489,
+ "grad_norm": 0.0180258359760046,
+ "learning_rate": 0.0006,
+ "loss": 4.842667579650879,
+ "step": 821
+ },
+ {
+ "epoch": 11.419213973799126,
+ "grad_norm": 0.022130120545625687,
+ "learning_rate": 0.0006,
+ "loss": 4.799970626831055,
+ "step": 822
+ },
+ {
+ "epoch": 11.433187772925764,
+ "grad_norm": 0.02384897693991661,
+ "learning_rate": 0.0006,
+ "loss": 4.860746383666992,
+ "step": 823
+ },
+ {
+ "epoch": 11.447161572052401,
+ "grad_norm": 0.021881572902202606,
+ "learning_rate": 0.0006,
+ "loss": 4.918107032775879,
+ "step": 824
+ },
+ {
+ "epoch": 11.46113537117904,
+ "grad_norm": 0.023085080087184906,
+ "learning_rate": 0.0006,
+ "loss": 4.878946304321289,
+ "step": 825
+ },
+ {
+ "epoch": 11.475109170305677,
+ "grad_norm": 0.02291012555360794,
+ "learning_rate": 0.0006,
+ "loss": 4.958887100219727,
+ "step": 826
+ },
+ {
+ "epoch": 11.489082969432314,
+ "grad_norm": 0.0199968870729208,
+ "learning_rate": 0.0006,
+ "loss": 4.80955696105957,
+ "step": 827
+ },
+ {
+ "epoch": 11.503056768558952,
+ "grad_norm": 0.018702229484915733,
+ "learning_rate": 0.0006,
+ "loss": 5.0276408195495605,
+ "step": 828
+ },
+ {
+ "epoch": 11.51703056768559,
+ "grad_norm": 0.017787311226129532,
+ "learning_rate": 0.0006,
+ "loss": 4.96018123626709,
+ "step": 829
+ },
+ {
+ "epoch": 11.531004366812226,
+ "grad_norm": 0.01704501546919346,
+ "learning_rate": 0.0006,
+ "loss": 4.914553165435791,
+ "step": 830
+ },
+ {
+ "epoch": 11.544978165938865,
+ "grad_norm": 0.017519677057862282,
+ "learning_rate": 0.0006,
+ "loss": 4.798098087310791,
+ "step": 831
+ },
+ {
+ "epoch": 11.558951965065502,
+ "grad_norm": 0.018511991947889328,
+ "learning_rate": 0.0006,
+ "loss": 4.817785263061523,
+ "step": 832
+ },
+ {
+ "epoch": 11.57292576419214,
+ "grad_norm": 0.01876644790172577,
+ "learning_rate": 0.0006,
+ "loss": 4.818995475769043,
+ "step": 833
+ },
+ {
+ "epoch": 11.586899563318777,
+ "grad_norm": 0.01753568835556507,
+ "learning_rate": 0.0006,
+ "loss": 4.811938285827637,
+ "step": 834
+ },
+ {
+ "epoch": 11.600873362445414,
+ "grad_norm": 0.017046531662344933,
+ "learning_rate": 0.0006,
+ "loss": 4.914674282073975,
+ "step": 835
+ },
+ {
+ "epoch": 11.614847161572053,
+ "grad_norm": 0.01802094094455242,
+ "learning_rate": 0.0006,
+ "loss": 4.993284225463867,
+ "step": 836
+ },
+ {
+ "epoch": 11.62882096069869,
+ "grad_norm": 0.019493764266371727,
+ "learning_rate": 0.0006,
+ "loss": 4.895628452301025,
+ "step": 837
+ },
+ {
+ "epoch": 11.642794759825328,
+ "grad_norm": 0.019461210817098618,
+ "learning_rate": 0.0006,
+ "loss": 4.869269371032715,
+ "step": 838
+ },
+ {
+ "epoch": 11.656768558951965,
+ "grad_norm": 0.01981567218899727,
+ "learning_rate": 0.0006,
+ "loss": 4.863052845001221,
+ "step": 839
+ },
+ {
+ "epoch": 11.670742358078602,
+ "grad_norm": 0.019894491881132126,
+ "learning_rate": 0.0006,
+ "loss": 4.8775787353515625,
+ "step": 840
+ },
+ {
+ "epoch": 11.68471615720524,
+ "grad_norm": 0.01875786855816841,
+ "learning_rate": 0.0006,
+ "loss": 4.817763328552246,
+ "step": 841
+ },
+ {
+ "epoch": 11.698689956331878,
+ "grad_norm": 0.020926695317029953,
+ "learning_rate": 0.0006,
+ "loss": 4.7470526695251465,
+ "step": 842
+ },
+ {
+ "epoch": 11.712663755458514,
+ "grad_norm": 0.021477103233337402,
+ "learning_rate": 0.0006,
+ "loss": 4.762682914733887,
+ "step": 843
+ },
+ {
+ "epoch": 11.726637554585153,
+ "grad_norm": 0.020628711208701134,
+ "learning_rate": 0.0006,
+ "loss": 4.807985782623291,
+ "step": 844
+ },
+ {
+ "epoch": 11.74061135371179,
+ "grad_norm": 0.021489612758159637,
+ "learning_rate": 0.0006,
+ "loss": 4.843091011047363,
+ "step": 845
+ },
+ {
+ "epoch": 11.754585152838429,
+ "grad_norm": 0.021194491535425186,
+ "learning_rate": 0.0006,
+ "loss": 4.8300251960754395,
+ "step": 846
+ },
+ {
+ "epoch": 11.768558951965066,
+ "grad_norm": 0.01833350397646427,
+ "learning_rate": 0.0006,
+ "loss": 5.000202178955078,
+ "step": 847
+ },
+ {
+ "epoch": 11.782532751091702,
+ "grad_norm": 0.018096931278705597,
+ "learning_rate": 0.0006,
+ "loss": 4.880648136138916,
+ "step": 848
+ },
+ {
+ "epoch": 11.796506550218341,
+ "grad_norm": 0.0197161752730608,
+ "learning_rate": 0.0006,
+ "loss": 4.788316249847412,
+ "step": 849
+ },
+ {
+ "epoch": 11.810480349344978,
+ "grad_norm": 0.0203181654214859,
+ "learning_rate": 0.0006,
+ "loss": 4.870944976806641,
+ "step": 850
+ },
+ {
+ "epoch": 11.824454148471617,
+ "grad_norm": 0.01908455230295658,
+ "learning_rate": 0.0006,
+ "loss": 4.890144348144531,
+ "step": 851
+ },
+ {
+ "epoch": 11.838427947598253,
+ "grad_norm": 0.021281801164150238,
+ "learning_rate": 0.0006,
+ "loss": 4.89578914642334,
+ "step": 852
+ },
+ {
+ "epoch": 11.85240174672489,
+ "grad_norm": 0.02124273031949997,
+ "learning_rate": 0.0006,
+ "loss": 4.883749008178711,
+ "step": 853
+ },
+ {
+ "epoch": 11.866375545851529,
+ "grad_norm": 0.020969398319721222,
+ "learning_rate": 0.0006,
+ "loss": 4.809142589569092,
+ "step": 854
+ },
+ {
+ "epoch": 11.880349344978166,
+ "grad_norm": 0.01967657171189785,
+ "learning_rate": 0.0006,
+ "loss": 4.821864604949951,
+ "step": 855
+ },
+ {
+ "epoch": 11.894323144104803,
+ "grad_norm": 0.020887333899736404,
+ "learning_rate": 0.0006,
+ "loss": 4.917985916137695,
+ "step": 856
+ },
+ {
+ "epoch": 11.908296943231441,
+ "grad_norm": 0.019203083589673042,
+ "learning_rate": 0.0006,
+ "loss": 4.824960708618164,
+ "step": 857
+ },
+ {
+ "epoch": 11.922270742358078,
+ "grad_norm": 0.020162401720881462,
+ "learning_rate": 0.0006,
+ "loss": 4.772617816925049,
+ "step": 858
+ },
+ {
+ "epoch": 11.936244541484717,
+ "grad_norm": 0.02251487225294113,
+ "learning_rate": 0.0006,
+ "loss": 4.735418319702148,
+ "step": 859
+ },
+ {
+ "epoch": 11.950218340611354,
+ "grad_norm": 0.020313527435064316,
+ "learning_rate": 0.0006,
+ "loss": 4.939233779907227,
+ "step": 860
+ },
+ {
+ "epoch": 11.96419213973799,
+ "grad_norm": 0.021635599434375763,
+ "learning_rate": 0.0006,
+ "loss": 4.861807346343994,
+ "step": 861
+ },
+ {
+ "epoch": 11.97816593886463,
+ "grad_norm": 0.02285311557352543,
+ "learning_rate": 0.0006,
+ "loss": 4.932732582092285,
+ "step": 862
+ },
+ {
+ "epoch": 11.992139737991266,
+ "grad_norm": 0.02072535641491413,
+ "learning_rate": 0.0006,
+ "loss": 4.796767711639404,
+ "step": 863
+ },
+ {
+ "epoch": 12.0,
+ "grad_norm": 0.020588304847478867,
+ "learning_rate": 0.0006,
+ "loss": 4.997537612915039,
+ "step": 864
+ },
+ {
+ "epoch": 12.0,
+ "eval_loss": 5.044651985168457,
+ "eval_runtime": 56.9145,
+ "eval_samples_per_second": 42.906,
+ "eval_steps_per_second": 1.353,
+ "step": 864
+ },
+ {
+ "epoch": 12.013973799126637,
+ "grad_norm": 0.017597166821360588,
+ "learning_rate": 0.0006,
+ "loss": 4.830524444580078,
+ "step": 865
+ },
+ {
+ "epoch": 12.027947598253276,
+ "grad_norm": 0.01716788113117218,
+ "learning_rate": 0.0006,
+ "loss": 4.894325256347656,
+ "step": 866
+ },
+ {
+ "epoch": 12.041921397379912,
+ "grad_norm": 0.017712457105517387,
+ "learning_rate": 0.0006,
+ "loss": 4.868297576904297,
+ "step": 867
+ },
+ {
+ "epoch": 12.055895196506551,
+ "grad_norm": 0.021486390382051468,
+ "learning_rate": 0.0006,
+ "loss": 4.922049522399902,
+ "step": 868
+ },
+ {
+ "epoch": 12.069868995633188,
+ "grad_norm": 0.024560654535889626,
+ "learning_rate": 0.0006,
+ "loss": 4.734241008758545,
+ "step": 869
+ },
+ {
+ "epoch": 12.083842794759825,
+ "grad_norm": 0.025870252400636673,
+ "learning_rate": 0.0006,
+ "loss": 4.868566989898682,
+ "step": 870
+ },
+ {
+ "epoch": 12.097816593886463,
+ "grad_norm": 0.0277389045804739,
+ "learning_rate": 0.0006,
+ "loss": 4.755006790161133,
+ "step": 871
+ },
+ {
+ "epoch": 12.1117903930131,
+ "grad_norm": 0.029100844636559486,
+ "learning_rate": 0.0006,
+ "loss": 4.909667015075684,
+ "step": 872
+ },
+ {
+ "epoch": 12.125764192139737,
+ "grad_norm": 0.031833067536354065,
+ "learning_rate": 0.0006,
+ "loss": 4.858674049377441,
+ "step": 873
+ },
+ {
+ "epoch": 12.139737991266376,
+ "grad_norm": 0.032709237188100815,
+ "learning_rate": 0.0006,
+ "loss": 4.823696136474609,
+ "step": 874
+ },
+ {
+ "epoch": 12.153711790393013,
+ "grad_norm": 0.02995380386710167,
+ "learning_rate": 0.0006,
+ "loss": 4.850235462188721,
+ "step": 875
+ },
+ {
+ "epoch": 12.167685589519651,
+ "grad_norm": 0.034098610281944275,
+ "learning_rate": 0.0006,
+ "loss": 4.8169989585876465,
+ "step": 876
+ },
+ {
+ "epoch": 12.181659388646288,
+ "grad_norm": 0.03377080336213112,
+ "learning_rate": 0.0006,
+ "loss": 4.90738582611084,
+ "step": 877
+ },
+ {
+ "epoch": 12.195633187772925,
+ "grad_norm": 0.029780825600028038,
+ "learning_rate": 0.0006,
+ "loss": 4.811163902282715,
+ "step": 878
+ },
+ {
+ "epoch": 12.209606986899564,
+ "grad_norm": 0.03118024580180645,
+ "learning_rate": 0.0006,
+ "loss": 4.861286640167236,
+ "step": 879
+ },
+ {
+ "epoch": 12.2235807860262,
+ "grad_norm": 0.033052004873752594,
+ "learning_rate": 0.0006,
+ "loss": 4.865021228790283,
+ "step": 880
+ },
+ {
+ "epoch": 12.237554585152838,
+ "grad_norm": 0.029070032760500908,
+ "learning_rate": 0.0006,
+ "loss": 4.947303771972656,
+ "step": 881
+ },
+ {
+ "epoch": 12.251528384279476,
+ "grad_norm": 0.028355715796351433,
+ "learning_rate": 0.0006,
+ "loss": 4.804648399353027,
+ "step": 882
+ },
+ {
+ "epoch": 12.265502183406113,
+ "grad_norm": 0.031257227063179016,
+ "learning_rate": 0.0006,
+ "loss": 4.757718086242676,
+ "step": 883
+ },
+ {
+ "epoch": 12.279475982532752,
+ "grad_norm": 0.031438373029232025,
+ "learning_rate": 0.0006,
+ "loss": 4.851268768310547,
+ "step": 884
+ },
+ {
+ "epoch": 12.293449781659389,
+ "grad_norm": 0.030492138117551804,
+ "learning_rate": 0.0006,
+ "loss": 4.7891035079956055,
+ "step": 885
+ },
+ {
+ "epoch": 12.307423580786025,
+ "grad_norm": 0.026115991175174713,
+ "learning_rate": 0.0006,
+ "loss": 4.86474609375,
+ "step": 886
+ },
+ {
+ "epoch": 12.321397379912664,
+ "grad_norm": 0.024184465408325195,
+ "learning_rate": 0.0006,
+ "loss": 4.860456943511963,
+ "step": 887
+ },
+ {
+ "epoch": 12.335371179039301,
+ "grad_norm": 0.025437982752919197,
+ "learning_rate": 0.0006,
+ "loss": 4.806251049041748,
+ "step": 888
+ },
+ {
+ "epoch": 12.34934497816594,
+ "grad_norm": 0.022390231490135193,
+ "learning_rate": 0.0006,
+ "loss": 4.807086944580078,
+ "step": 889
+ },
+ {
+ "epoch": 12.363318777292577,
+ "grad_norm": 0.020174294710159302,
+ "learning_rate": 0.0006,
+ "loss": 4.78292179107666,
+ "step": 890
+ },
+ {
+ "epoch": 12.377292576419213,
+ "grad_norm": 0.019734324887394905,
+ "learning_rate": 0.0006,
+ "loss": 4.784509658813477,
+ "step": 891
+ },
+ {
+ "epoch": 12.391266375545852,
+ "grad_norm": 0.016733523458242416,
+ "learning_rate": 0.0006,
+ "loss": 4.811755180358887,
+ "step": 892
+ },
+ {
+ "epoch": 12.405240174672489,
+ "grad_norm": 0.01570146717131138,
+ "learning_rate": 0.0006,
+ "loss": 4.720088005065918,
+ "step": 893
+ },
+ {
+ "epoch": 12.419213973799126,
+ "grad_norm": 0.01601393334567547,
+ "learning_rate": 0.0006,
+ "loss": 4.783543109893799,
+ "step": 894
+ },
+ {
+ "epoch": 12.433187772925764,
+ "grad_norm": 0.014035487547516823,
+ "learning_rate": 0.0006,
+ "loss": 4.848340034484863,
+ "step": 895
+ },
+ {
+ "epoch": 12.447161572052401,
+ "grad_norm": 0.015248487703502178,
+ "learning_rate": 0.0006,
+ "loss": 4.773287773132324,
+ "step": 896
+ },
+ {
+ "epoch": 12.46113537117904,
+ "grad_norm": 0.01532268151640892,
+ "learning_rate": 0.0006,
+ "loss": 4.799646854400635,
+ "step": 897
+ },
+ {
+ "epoch": 12.475109170305677,
+ "grad_norm": 0.013982534408569336,
+ "learning_rate": 0.0006,
+ "loss": 4.8090500831604,
+ "step": 898
+ },
+ {
+ "epoch": 12.489082969432314,
+ "grad_norm": 0.014081164263188839,
+ "learning_rate": 0.0006,
+ "loss": 4.751888751983643,
+ "step": 899
+ },
+ {
+ "epoch": 12.503056768558952,
+ "grad_norm": 0.01380133256316185,
+ "learning_rate": 0.0006,
+ "loss": 4.78144645690918,
+ "step": 900
+ },
+ {
+ "epoch": 12.51703056768559,
+ "grad_norm": 0.013555269688367844,
+ "learning_rate": 0.0006,
+ "loss": 4.779257774353027,
+ "step": 901
+ },
+ {
+ "epoch": 12.531004366812226,
+ "grad_norm": 0.013075701892375946,
+ "learning_rate": 0.0006,
+ "loss": 4.820705413818359,
+ "step": 902
+ },
+ {
+ "epoch": 12.544978165938865,
+ "grad_norm": 0.01276316400617361,
+ "learning_rate": 0.0006,
+ "loss": 4.836733818054199,
+ "step": 903
+ },
+ {
+ "epoch": 12.558951965065502,
+ "grad_norm": 0.013033892959356308,
+ "learning_rate": 0.0006,
+ "loss": 4.764178276062012,
+ "step": 904
+ },
+ {
+ "epoch": 12.57292576419214,
+ "grad_norm": 0.013996980153024197,
+ "learning_rate": 0.0006,
+ "loss": 4.767983436584473,
+ "step": 905
+ },
+ {
+ "epoch": 12.586899563318777,
+ "grad_norm": 0.015222184360027313,
+ "learning_rate": 0.0006,
+ "loss": 4.759721755981445,
+ "step": 906
+ },
+ {
+ "epoch": 12.600873362445414,
+ "grad_norm": 0.01623941957950592,
+ "learning_rate": 0.0006,
+ "loss": 4.849811553955078,
+ "step": 907
+ },
+ {
+ "epoch": 12.614847161572053,
+ "grad_norm": 0.015711182728409767,
+ "learning_rate": 0.0006,
+ "loss": 4.7829084396362305,
+ "step": 908
+ },
+ {
+ "epoch": 12.62882096069869,
+ "grad_norm": 0.014137467369437218,
+ "learning_rate": 0.0006,
+ "loss": 4.73321533203125,
+ "step": 909
+ },
+ {
+ "epoch": 12.642794759825328,
+ "grad_norm": 0.012738462537527084,
+ "learning_rate": 0.0006,
+ "loss": 4.870052337646484,
+ "step": 910
+ },
+ {
+ "epoch": 12.656768558951965,
+ "grad_norm": 0.013683130033314228,
+ "learning_rate": 0.0006,
+ "loss": 4.82658052444458,
+ "step": 911
+ },
+ {
+ "epoch": 12.670742358078602,
+ "grad_norm": 0.014409830793738365,
+ "learning_rate": 0.0006,
+ "loss": 4.707494735717773,
+ "step": 912
+ },
+ {
+ "epoch": 12.68471615720524,
+ "grad_norm": 0.013806809671223164,
+ "learning_rate": 0.0006,
+ "loss": 4.778928279876709,
+ "step": 913
+ },
+ {
+ "epoch": 12.698689956331878,
+ "grad_norm": 0.012857912108302116,
+ "learning_rate": 0.0006,
+ "loss": 4.890717506408691,
+ "step": 914
+ },
+ {
+ "epoch": 12.712663755458514,
+ "grad_norm": 0.013828758150339127,
+ "learning_rate": 0.0006,
+ "loss": 4.8357439041137695,
+ "step": 915
+ },
+ {
+ "epoch": 12.726637554585153,
+ "grad_norm": 0.01478351280093193,
+ "learning_rate": 0.0006,
+ "loss": 4.869986534118652,
+ "step": 916
+ },
+ {
+ "epoch": 12.74061135371179,
+ "grad_norm": 0.01326421182602644,
+ "learning_rate": 0.0006,
+ "loss": 4.830646514892578,
+ "step": 917
+ },
+ {
+ "epoch": 12.754585152838429,
+ "grad_norm": 0.013335433788597584,
+ "learning_rate": 0.0006,
+ "loss": 4.7786149978637695,
+ "step": 918
+ },
+ {
+ "epoch": 12.768558951965066,
+ "grad_norm": 0.014261147007346153,
+ "learning_rate": 0.0006,
+ "loss": 4.788693428039551,
+ "step": 919
+ },
+ {
+ "epoch": 12.782532751091702,
+ "grad_norm": 0.014274783432483673,
+ "learning_rate": 0.0006,
+ "loss": 4.77295446395874,
+ "step": 920
+ },
+ {
+ "epoch": 12.796506550218341,
+ "grad_norm": 0.015074674971401691,
+ "learning_rate": 0.0006,
+ "loss": 4.830754280090332,
+ "step": 921
+ },
+ {
+ "epoch": 12.810480349344978,
+ "grad_norm": 0.015599515289068222,
+ "learning_rate": 0.0006,
+ "loss": 4.837759017944336,
+ "step": 922
+ },
+ {
+ "epoch": 12.824454148471617,
+ "grad_norm": 0.014595242217183113,
+ "learning_rate": 0.0006,
+ "loss": 4.882965564727783,
+ "step": 923
+ },
+ {
+ "epoch": 12.838427947598253,
+ "grad_norm": 0.014951467514038086,
+ "learning_rate": 0.0006,
+ "loss": 4.709748268127441,
+ "step": 924
+ },
+ {
+ "epoch": 12.85240174672489,
+ "grad_norm": 0.01539295818656683,
+ "learning_rate": 0.0006,
+ "loss": 4.791160583496094,
+ "step": 925
+ },
+ {
+ "epoch": 12.866375545851529,
+ "grad_norm": 0.015987906605005264,
+ "learning_rate": 0.0006,
+ "loss": 4.661876678466797,
+ "step": 926
+ },
+ {
+ "epoch": 12.880349344978166,
+ "grad_norm": 0.016527341678738594,
+ "learning_rate": 0.0006,
+ "loss": 4.721952438354492,
+ "step": 927
+ },
+ {
+ "epoch": 12.894323144104803,
+ "grad_norm": 0.01637161523103714,
+ "learning_rate": 0.0006,
+ "loss": 4.704446792602539,
+ "step": 928
+ },
+ {
+ "epoch": 12.908296943231441,
+ "grad_norm": 0.015221050009131432,
+ "learning_rate": 0.0006,
+ "loss": 4.771907806396484,
+ "step": 929
+ },
+ {
+ "epoch": 12.922270742358078,
+ "grad_norm": 0.016242166981101036,
+ "learning_rate": 0.0006,
+ "loss": 4.806267738342285,
+ "step": 930
+ },
+ {
+ "epoch": 12.936244541484717,
+ "grad_norm": 0.01628357358276844,
+ "learning_rate": 0.0006,
+ "loss": 4.7252326011657715,
+ "step": 931
+ },
+ {
+ "epoch": 12.950218340611354,
+ "grad_norm": 0.017428133636713028,
+ "learning_rate": 0.0006,
+ "loss": 4.765689849853516,
+ "step": 932
+ },
+ {
+ "epoch": 12.96419213973799,
+ "grad_norm": 0.01870061084628105,
+ "learning_rate": 0.0006,
+ "loss": 4.814488887786865,
+ "step": 933
+ },
+ {
+ "epoch": 12.97816593886463,
+ "grad_norm": 0.019796060398221016,
+ "learning_rate": 0.0006,
+ "loss": 4.739185810089111,
+ "step": 934
+ },
+ {
+ "epoch": 12.992139737991266,
+ "grad_norm": 0.01985820196568966,
+ "learning_rate": 0.0006,
+ "loss": 4.810608863830566,
+ "step": 935
+ },
+ {
+ "epoch": 13.0,
+ "grad_norm": 0.020834477618336678,
+ "learning_rate": 0.0006,
+ "loss": 4.727858066558838,
+ "step": 936
+ },
+ {
+ "epoch": 13.0,
+ "eval_loss": 4.979116916656494,
+ "eval_runtime": 57.2877,
+ "eval_samples_per_second": 42.627,
+ "eval_steps_per_second": 1.344,
+ "step": 936
+ },
+ {
+ "epoch": 13.013973799126637,
+ "grad_norm": 0.02025519125163555,
+ "learning_rate": 0.0006,
+ "loss": 4.628875732421875,
+ "step": 937
+ },
+ {
+ "epoch": 13.027947598253276,
+ "grad_norm": 0.019596368074417114,
+ "learning_rate": 0.0006,
+ "loss": 4.694699287414551,
+ "step": 938
+ },
+ {
+ "epoch": 13.041921397379912,
+ "grad_norm": 0.020494934171438217,
+ "learning_rate": 0.0006,
+ "loss": 4.691821098327637,
+ "step": 939
+ },
+ {
+ "epoch": 13.055895196506551,
+ "grad_norm": 0.020121760666370392,
+ "learning_rate": 0.0006,
+ "loss": 4.71478271484375,
+ "step": 940
+ },
+ {
+ "epoch": 13.069868995633188,
+ "grad_norm": 0.021091405302286148,
+ "learning_rate": 0.0006,
+ "loss": 4.695915222167969,
+ "step": 941
+ },
+ {
+ "epoch": 13.083842794759825,
+ "grad_norm": 0.019787423312664032,
+ "learning_rate": 0.0006,
+ "loss": 4.8652448654174805,
+ "step": 942
+ },
+ {
+ "epoch": 13.097816593886463,
+ "grad_norm": 0.0185833927243948,
+ "learning_rate": 0.0006,
+ "loss": 4.7065887451171875,
+ "step": 943
+ },
+ {
+ "epoch": 13.1117903930131,
+ "grad_norm": 0.021220475435256958,
+ "learning_rate": 0.0006,
+ "loss": 4.741232395172119,
+ "step": 944
+ },
+ {
+ "epoch": 13.125764192139737,
+ "grad_norm": 0.02136622555553913,
+ "learning_rate": 0.0006,
+ "loss": 4.705967903137207,
+ "step": 945
+ },
+ {
+ "epoch": 13.139737991266376,
+ "grad_norm": 0.0191587433218956,
+ "learning_rate": 0.0006,
+ "loss": 4.758667945861816,
+ "step": 946
+ },
+ {
+ "epoch": 13.153711790393013,
+ "grad_norm": 0.018290022388100624,
+ "learning_rate": 0.0006,
+ "loss": 4.760666847229004,
+ "step": 947
+ },
+ {
+ "epoch": 13.167685589519651,
+ "grad_norm": 0.021704409271478653,
+ "learning_rate": 0.0006,
+ "loss": 4.750555992126465,
+ "step": 948
+ },
+ {
+ "epoch": 13.181659388646288,
+ "grad_norm": 0.021668918430805206,
+ "learning_rate": 0.0006,
+ "loss": 4.840903282165527,
+ "step": 949
+ },
+ {
+ "epoch": 13.195633187772925,
+ "grad_norm": 0.019497141242027283,
+ "learning_rate": 0.0006,
+ "loss": 4.738969326019287,
+ "step": 950
+ },
+ {
+ "epoch": 13.209606986899564,
+ "grad_norm": 0.018500596284866333,
+ "learning_rate": 0.0006,
+ "loss": 4.75154972076416,
+ "step": 951
+ },
+ {
+ "epoch": 13.2235807860262,
+ "grad_norm": 0.01872239261865616,
+ "learning_rate": 0.0006,
+ "loss": 4.811738014221191,
+ "step": 952
+ },
+ {
+ "epoch": 13.237554585152838,
+ "grad_norm": 0.020321547985076904,
+ "learning_rate": 0.0006,
+ "loss": 4.815384387969971,
+ "step": 953
+ },
+ {
+ "epoch": 13.251528384279476,
+ "grad_norm": 0.02283575013279915,
+ "learning_rate": 0.0006,
+ "loss": 4.74582052230835,
+ "step": 954
+ },
+ {
+ "epoch": 13.265502183406113,
+ "grad_norm": 0.02474398910999298,
+ "learning_rate": 0.0006,
+ "loss": 4.738375186920166,
+ "step": 955
+ },
+ {
+ "epoch": 13.279475982532752,
+ "grad_norm": 0.02065764181315899,
+ "learning_rate": 0.0006,
+ "loss": 4.746875286102295,
+ "step": 956
+ },
+ {
+ "epoch": 13.293449781659389,
+ "grad_norm": 0.020534520968794823,
+ "learning_rate": 0.0006,
+ "loss": 4.656914710998535,
+ "step": 957
+ },
+ {
+ "epoch": 13.307423580786025,
+ "grad_norm": 0.0206378772854805,
+ "learning_rate": 0.0006,
+ "loss": 4.842615604400635,
+ "step": 958
+ },
+ {
+ "epoch": 13.321397379912664,
+ "grad_norm": 0.017935654148459435,
+ "learning_rate": 0.0006,
+ "loss": 4.717827796936035,
+ "step": 959
+ },
+ {
+ "epoch": 13.335371179039301,
+ "grad_norm": 0.016958363354206085,
+ "learning_rate": 0.0006,
+ "loss": 4.681070327758789,
+ "step": 960
+ },
+ {
+ "epoch": 13.34934497816594,
+ "grad_norm": 0.015735412016510963,
+ "learning_rate": 0.0006,
+ "loss": 4.7957048416137695,
+ "step": 961
+ },
+ {
+ "epoch": 13.363318777292577,
+ "grad_norm": 0.01563699170947075,
+ "learning_rate": 0.0006,
+ "loss": 4.701192378997803,
+ "step": 962
+ },
+ {
+ "epoch": 13.377292576419213,
+ "grad_norm": 0.015207760035991669,
+ "learning_rate": 0.0006,
+ "loss": 4.73256254196167,
+ "step": 963
+ },
+ {
+ "epoch": 13.391266375545852,
+ "grad_norm": 0.015396883711218834,
+ "learning_rate": 0.0006,
+ "loss": 4.76412296295166,
+ "step": 964
+ },
+ {
+ "epoch": 13.405240174672489,
+ "grad_norm": 0.015690583735704422,
+ "learning_rate": 0.0006,
+ "loss": 4.791157245635986,
+ "step": 965
+ },
+ {
+ "epoch": 13.419213973799126,
+ "grad_norm": 0.01755724661052227,
+ "learning_rate": 0.0006,
+ "loss": 4.845666885375977,
+ "step": 966
+ },
+ {
+ "epoch": 13.433187772925764,
+ "grad_norm": 0.019443422555923462,
+ "learning_rate": 0.0006,
+ "loss": 4.5902299880981445,
+ "step": 967
+ },
+ {
+ "epoch": 13.447161572052401,
+ "grad_norm": 0.02118338644504547,
+ "learning_rate": 0.0006,
+ "loss": 4.740627288818359,
+ "step": 968
+ },
+ {
+ "epoch": 13.46113537117904,
+ "grad_norm": 0.02174839936196804,
+ "learning_rate": 0.0006,
+ "loss": 4.814560890197754,
+ "step": 969
+ },
+ {
+ "epoch": 13.475109170305677,
+ "grad_norm": 0.022860374301671982,
+ "learning_rate": 0.0006,
+ "loss": 4.66440486907959,
+ "step": 970
+ },
+ {
+ "epoch": 13.489082969432314,
+ "grad_norm": 0.024742677807807922,
+ "learning_rate": 0.0006,
+ "loss": 4.7451653480529785,
+ "step": 971
+ },
+ {
+ "epoch": 13.503056768558952,
+ "grad_norm": 0.027068907395005226,
+ "learning_rate": 0.0006,
+ "loss": 4.802584171295166,
+ "step": 972
+ },
+ {
+ "epoch": 13.51703056768559,
+ "grad_norm": 0.026316625997424126,
+ "learning_rate": 0.0006,
+ "loss": 4.842397689819336,
+ "step": 973
+ },
+ {
+ "epoch": 13.531004366812226,
+ "grad_norm": 0.024610962718725204,
+ "learning_rate": 0.0006,
+ "loss": 4.650721073150635,
+ "step": 974
+ },
+ {
+ "epoch": 13.544978165938865,
+ "grad_norm": 0.024122603237628937,
+ "learning_rate": 0.0006,
+ "loss": 4.6984357833862305,
+ "step": 975
+ },
+ {
+ "epoch": 13.558951965065502,
+ "grad_norm": 0.022699708119034767,
+ "learning_rate": 0.0006,
+ "loss": 4.793007850646973,
+ "step": 976
+ },
+ {
+ "epoch": 13.57292576419214,
+ "grad_norm": 0.021706534549593925,
+ "learning_rate": 0.0006,
+ "loss": 4.691362380981445,
+ "step": 977
+ },
+ {
+ "epoch": 13.586899563318777,
+ "grad_norm": 0.021363291889429092,
+ "learning_rate": 0.0006,
+ "loss": 4.76762580871582,
+ "step": 978
+ },
+ {
+ "epoch": 13.600873362445414,
+ "grad_norm": 0.023476677015423775,
+ "learning_rate": 0.0006,
+ "loss": 4.777202606201172,
+ "step": 979
+ },
+ {
+ "epoch": 13.614847161572053,
+ "grad_norm": 0.02457100711762905,
+ "learning_rate": 0.0006,
+ "loss": 4.590703010559082,
+ "step": 980
+ },
+ {
+ "epoch": 13.62882096069869,
+ "grad_norm": 0.02469099499285221,
+ "learning_rate": 0.0006,
+ "loss": 4.640185356140137,
+ "step": 981
+ },
+ {
+ "epoch": 13.642794759825328,
+ "grad_norm": 0.024914603680372238,
+ "learning_rate": 0.0006,
+ "loss": 4.772201061248779,
+ "step": 982
+ },
+ {
+ "epoch": 13.656768558951965,
+ "grad_norm": 0.025133173912763596,
+ "learning_rate": 0.0006,
+ "loss": 4.743337154388428,
+ "step": 983
+ },
+ {
+ "epoch": 13.670742358078602,
+ "grad_norm": 0.02034137211740017,
+ "learning_rate": 0.0006,
+ "loss": 4.752127647399902,
+ "step": 984
+ },
+ {
+ "epoch": 13.68471615720524,
+ "grad_norm": 0.019645661115646362,
+ "learning_rate": 0.0006,
+ "loss": 4.838453769683838,
+ "step": 985
+ },
+ {
+ "epoch": 13.698689956331878,
+ "grad_norm": 0.01900297962129116,
+ "learning_rate": 0.0006,
+ "loss": 4.804868221282959,
+ "step": 986
+ },
+ {
+ "epoch": 13.712663755458514,
+ "grad_norm": 0.020292341709136963,
+ "learning_rate": 0.0006,
+ "loss": 4.609743118286133,
+ "step": 987
+ },
+ {
+ "epoch": 13.726637554585153,
+ "grad_norm": 0.017014775425195694,
+ "learning_rate": 0.0006,
+ "loss": 4.740612983703613,
+ "step": 988
+ },
+ {
+ "epoch": 13.74061135371179,
+ "grad_norm": 0.015199844725430012,
+ "learning_rate": 0.0006,
+ "loss": 4.719701766967773,
+ "step": 989
+ },
+ {
+ "epoch": 13.754585152838429,
+ "grad_norm": 0.015125414356589317,
+ "learning_rate": 0.0006,
+ "loss": 4.686114311218262,
+ "step": 990
+ },
+ {
+ "epoch": 13.768558951965066,
+ "grad_norm": 0.014779540710151196,
+ "learning_rate": 0.0006,
+ "loss": 4.674347877502441,
+ "step": 991
+ },
+ {
+ "epoch": 13.782532751091702,
+ "grad_norm": 0.015255163423717022,
+ "learning_rate": 0.0006,
+ "loss": 4.6309309005737305,
+ "step": 992
+ },
+ {
+ "epoch": 13.796506550218341,
+ "grad_norm": 0.013692197389900684,
+ "learning_rate": 0.0006,
+ "loss": 4.698497295379639,
+ "step": 993
+ },
+ {
+ "epoch": 13.810480349344978,
+ "grad_norm": 0.014529623091220856,
+ "learning_rate": 0.0006,
+ "loss": 4.628443717956543,
+ "step": 994
+ },
+ {
+ "epoch": 13.824454148471617,
+ "grad_norm": 0.01576659083366394,
+ "learning_rate": 0.0006,
+ "loss": 4.788015842437744,
+ "step": 995
+ },
+ {
+ "epoch": 13.838427947598253,
+ "grad_norm": 0.015834277495741844,
+ "learning_rate": 0.0006,
+ "loss": 4.672703742980957,
+ "step": 996
+ },
+ {
+ "epoch": 13.85240174672489,
+ "grad_norm": 0.015539503656327724,
+ "learning_rate": 0.0006,
+ "loss": 4.731503486633301,
+ "step": 997
+ },
+ {
+ "epoch": 13.866375545851529,
+ "grad_norm": 0.013463746756315231,
+ "learning_rate": 0.0006,
+ "loss": 4.7006330490112305,
+ "step": 998
+ },
+ {
+ "epoch": 13.880349344978166,
+ "grad_norm": 0.016377726569771767,
+ "learning_rate": 0.0006,
+ "loss": 4.636074066162109,
+ "step": 999
+ },
+ {
+ "epoch": 13.894323144104803,
+ "grad_norm": 0.016660034656524658,
+ "learning_rate": 0.0006,
+ "loss": 4.742693901062012,
+ "step": 1000
+ },
+ {
+ "epoch": 13.908296943231441,
+ "grad_norm": 0.014902938157320023,
+ "learning_rate": 0.0006,
+ "loss": 4.769811630249023,
+ "step": 1001
+ },
+ {
+ "epoch": 13.922270742358078,
+ "grad_norm": 0.014883476309478283,
+ "learning_rate": 0.0006,
+ "loss": 4.731159210205078,
+ "step": 1002
+ },
+ {
+ "epoch": 13.936244541484717,
+ "grad_norm": 0.014494217932224274,
+ "learning_rate": 0.0006,
+ "loss": 4.659485816955566,
+ "step": 1003
+ },
+ {
+ "epoch": 13.950218340611354,
+ "grad_norm": 0.015426991507411003,
+ "learning_rate": 0.0006,
+ "loss": 4.755229949951172,
+ "step": 1004
+ },
+ {
+ "epoch": 13.96419213973799,
+ "grad_norm": 0.016648586839437485,
+ "learning_rate": 0.0006,
+ "loss": 4.74111270904541,
+ "step": 1005
+ },
+ {
+ "epoch": 13.97816593886463,
+ "grad_norm": 0.016980541869997978,
+ "learning_rate": 0.0006,
+ "loss": 4.74072265625,
+ "step": 1006
+ },
+ {
+ "epoch": 13.992139737991266,
+ "grad_norm": 0.018070152029395103,
+ "learning_rate": 0.0006,
+ "loss": 4.732975006103516,
+ "step": 1007
+ },
+ {
+ "epoch": 14.0,
+ "grad_norm": 0.019910665228962898,
+ "learning_rate": 0.0006,
+ "loss": 4.634697914123535,
+ "step": 1008
+ },
+ {
+ "epoch": 14.0,
+ "eval_loss": 4.850640296936035,
+ "eval_runtime": 57.3039,
+ "eval_samples_per_second": 42.615,
+ "eval_steps_per_second": 1.344,
+ "step": 1008
+ },
+ {
+ "epoch": 14.013973799126637,
+ "grad_norm": 0.01980053074657917,
+ "learning_rate": 0.0006,
+ "loss": 4.628744602203369,
+ "step": 1009
+ },
+ {
+ "epoch": 14.027947598253276,
+ "grad_norm": 0.02497623674571514,
+ "learning_rate": 0.0006,
+ "loss": 4.75930118560791,
+ "step": 1010
+ },
+ {
+ "epoch": 14.041921397379912,
+ "grad_norm": 0.02794097363948822,
+ "learning_rate": 0.0006,
+ "loss": 4.691340446472168,
+ "step": 1011
+ },
+ {
+ "epoch": 14.055895196506551,
+ "grad_norm": 0.023725535720586777,
+ "learning_rate": 0.0006,
+ "loss": 4.733664512634277,
+ "step": 1012
+ },
+ {
+ "epoch": 14.069868995633188,
+ "grad_norm": 0.022966474294662476,
+ "learning_rate": 0.0006,
+ "loss": 4.664808750152588,
+ "step": 1013
+ },
+ {
+ "epoch": 14.083842794759825,
+ "grad_norm": 0.02530563250184059,
+ "learning_rate": 0.0006,
+ "loss": 4.8047285079956055,
+ "step": 1014
+ },
+ {
+ "epoch": 14.097816593886463,
+ "grad_norm": 0.02491082064807415,
+ "learning_rate": 0.0006,
+ "loss": 4.667956352233887,
+ "step": 1015
+ },
+ {
+ "epoch": 14.1117903930131,
+ "grad_norm": 0.022110294550657272,
+ "learning_rate": 0.0006,
+ "loss": 4.666140556335449,
+ "step": 1016
+ },
+ {
+ "epoch": 14.125764192139737,
+ "grad_norm": 0.02065849117934704,
+ "learning_rate": 0.0006,
+ "loss": 4.581960678100586,
+ "step": 1017
+ },
+ {
+ "epoch": 14.139737991266376,
+ "grad_norm": 0.01941298507153988,
+ "learning_rate": 0.0006,
+ "loss": 4.614226341247559,
+ "step": 1018
+ },
+ {
+ "epoch": 14.153711790393013,
+ "grad_norm": 0.022441856563091278,
+ "learning_rate": 0.0006,
+ "loss": 4.638352394104004,
+ "step": 1019
+ },
+ {
+ "epoch": 14.167685589519651,
+ "grad_norm": 0.021519694477319717,
+ "learning_rate": 0.0006,
+ "loss": 4.700160980224609,
+ "step": 1020
+ },
+ {
+ "epoch": 14.181659388646288,
+ "grad_norm": 0.018956486135721207,
+ "learning_rate": 0.0006,
+ "loss": 4.649593353271484,
+ "step": 1021
+ },
+ {
+ "epoch": 14.195633187772925,
+ "grad_norm": 0.016681140288710594,
+ "learning_rate": 0.0006,
+ "loss": 4.603952407836914,
+ "step": 1022
+ },
+ {
+ "epoch": 14.209606986899564,
+ "grad_norm": 0.01788809522986412,
+ "learning_rate": 0.0006,
+ "loss": 4.669160842895508,
+ "step": 1023
+ },
+ {
+ "epoch": 14.2235807860262,
+ "grad_norm": 0.016047228127717972,
+ "learning_rate": 0.0006,
+ "loss": 4.6367950439453125,
+ "step": 1024
+ },
+ {
+ "epoch": 14.237554585152838,
+ "grad_norm": 0.01614239625632763,
+ "learning_rate": 0.0006,
+ "loss": 4.664529323577881,
+ "step": 1025
+ },
+ {
+ "epoch": 14.251528384279476,
+ "grad_norm": 0.016362549737095833,
+ "learning_rate": 0.0006,
+ "loss": 4.702373504638672,
+ "step": 1026
+ },
+ {
+ "epoch": 14.265502183406113,
+ "grad_norm": 0.015800409018993378,
+ "learning_rate": 0.0006,
+ "loss": 4.7594428062438965,
+ "step": 1027
+ },
+ {
+ "epoch": 14.279475982532752,
+ "grad_norm": 0.016730139032006264,
+ "learning_rate": 0.0006,
+ "loss": 4.677244186401367,
+ "step": 1028
+ },
+ {
+ "epoch": 14.293449781659389,
+ "grad_norm": 0.01661759801208973,
+ "learning_rate": 0.0006,
+ "loss": 4.633679389953613,
+ "step": 1029
+ },
+ {
+ "epoch": 14.307423580786025,
+ "grad_norm": 0.017673347145318985,
+ "learning_rate": 0.0006,
+ "loss": 4.649107456207275,
+ "step": 1030
+ },
+ {
+ "epoch": 14.321397379912664,
+ "grad_norm": 0.019243352115154266,
+ "learning_rate": 0.0006,
+ "loss": 4.611257553100586,
+ "step": 1031
+ },
+ {
+ "epoch": 14.335371179039301,
+ "grad_norm": 0.019817164167761803,
+ "learning_rate": 0.0006,
+ "loss": 4.588086128234863,
+ "step": 1032
+ },
+ {
+ "epoch": 14.34934497816594,
+ "grad_norm": 0.022292211651802063,
+ "learning_rate": 0.0006,
+ "loss": 4.783066749572754,
+ "step": 1033
+ },
+ {
+ "epoch": 14.363318777292577,
+ "grad_norm": 0.02140096202492714,
+ "learning_rate": 0.0006,
+ "loss": 4.604718208312988,
+ "step": 1034
+ },
+ {
+ "epoch": 14.377292576419213,
+ "grad_norm": 0.0187817495316267,
+ "learning_rate": 0.0006,
+ "loss": 4.721773147583008,
+ "step": 1035
+ },
+ {
+ "epoch": 14.391266375545852,
+ "grad_norm": 0.019879210740327835,
+ "learning_rate": 0.0006,
+ "loss": 4.789286136627197,
+ "step": 1036
+ },
+ {
+ "epoch": 14.405240174672489,
+ "grad_norm": 0.01885785162448883,
+ "learning_rate": 0.0006,
+ "loss": 4.669743537902832,
+ "step": 1037
+ },
+ {
+ "epoch": 14.419213973799126,
+ "grad_norm": 0.018324270844459534,
+ "learning_rate": 0.0006,
+ "loss": 4.65901517868042,
+ "step": 1038
+ },
+ {
+ "epoch": 14.433187772925764,
+ "grad_norm": 0.017664821818470955,
+ "learning_rate": 0.0006,
+ "loss": 4.667839527130127,
+ "step": 1039
+ },
+ {
+ "epoch": 14.447161572052401,
+ "grad_norm": 0.015702905133366585,
+ "learning_rate": 0.0006,
+ "loss": 4.64700984954834,
+ "step": 1040
+ },
+ {
+ "epoch": 14.46113537117904,
+ "grad_norm": 0.016674194484949112,
+ "learning_rate": 0.0006,
+ "loss": 4.694093227386475,
+ "step": 1041
+ },
+ {
+ "epoch": 14.475109170305677,
+ "grad_norm": 0.016869477927684784,
+ "learning_rate": 0.0006,
+ "loss": 4.619544982910156,
+ "step": 1042
+ },
+ {
+ "epoch": 14.489082969432314,
+ "grad_norm": 0.016557540744543076,
+ "learning_rate": 0.0006,
+ "loss": 4.568667411804199,
+ "step": 1043
+ },
+ {
+ "epoch": 14.503056768558952,
+ "grad_norm": 0.014500945806503296,
+ "learning_rate": 0.0006,
+ "loss": 4.561741828918457,
+ "step": 1044
+ },
+ {
+ "epoch": 14.51703056768559,
+ "grad_norm": 0.014228739775717258,
+ "learning_rate": 0.0006,
+ "loss": 4.654277324676514,
+ "step": 1045
+ },
+ {
+ "epoch": 14.531004366812226,
+ "grad_norm": 0.013894238509237766,
+ "learning_rate": 0.0006,
+ "loss": 4.5867719650268555,
+ "step": 1046
+ },
+ {
+ "epoch": 14.544978165938865,
+ "grad_norm": 0.01550068985670805,
+ "learning_rate": 0.0006,
+ "loss": 4.61168098449707,
+ "step": 1047
+ },
+ {
+ "epoch": 14.558951965065502,
+ "grad_norm": 0.015940772369503975,
+ "learning_rate": 0.0006,
+ "loss": 4.691247940063477,
+ "step": 1048
+ },
+ {
+ "epoch": 14.57292576419214,
+ "grad_norm": 0.01714535616338253,
+ "learning_rate": 0.0006,
+ "loss": 4.61183500289917,
+ "step": 1049
+ },
+ {
+ "epoch": 14.586899563318777,
+ "grad_norm": 0.017925038933753967,
+ "learning_rate": 0.0006,
+ "loss": 4.719022750854492,
+ "step": 1050
+ },
+ {
+ "epoch": 14.600873362445414,
+ "grad_norm": 0.018141552805900574,
+ "learning_rate": 0.0006,
+ "loss": 4.760281562805176,
+ "step": 1051
+ },
+ {
+ "epoch": 14.614847161572053,
+ "grad_norm": 0.017131328582763672,
+ "learning_rate": 0.0006,
+ "loss": 4.709680080413818,
+ "step": 1052
+ },
+ {
+ "epoch": 14.62882096069869,
+ "grad_norm": 0.018730048090219498,
+ "learning_rate": 0.0006,
+ "loss": 4.5994720458984375,
+ "step": 1053
+ },
+ {
+ "epoch": 14.642794759825328,
+ "grad_norm": 0.018894298002123833,
+ "learning_rate": 0.0006,
+ "loss": 4.669070720672607,
+ "step": 1054
+ },
+ {
+ "epoch": 14.656768558951965,
+ "grad_norm": 0.017643198370933533,
+ "learning_rate": 0.0006,
+ "loss": 4.641790390014648,
+ "step": 1055
+ },
+ {
+ "epoch": 14.670742358078602,
+ "grad_norm": 0.017647244036197662,
+ "learning_rate": 0.0006,
+ "loss": 4.653899192810059,
+ "step": 1056
+ },
+ {
+ "epoch": 14.68471615720524,
+ "grad_norm": 0.017304198816418648,
+ "learning_rate": 0.0006,
+ "loss": 4.7487473487854,
+ "step": 1057
+ },
+ {
+ "epoch": 14.698689956331878,
+ "grad_norm": 0.01638447679579258,
+ "learning_rate": 0.0006,
+ "loss": 4.71945858001709,
+ "step": 1058
+ },
+ {
+ "epoch": 14.712663755458514,
+ "grad_norm": 0.015892071649432182,
+ "learning_rate": 0.0006,
+ "loss": 4.744903564453125,
+ "step": 1059
+ },
+ {
+ "epoch": 14.726637554585153,
+ "grad_norm": 0.017420614138245583,
+ "learning_rate": 0.0006,
+ "loss": 4.781563758850098,
+ "step": 1060
+ },
+ {
+ "epoch": 14.74061135371179,
+ "grad_norm": 0.016498520970344543,
+ "learning_rate": 0.0006,
+ "loss": 4.622779846191406,
+ "step": 1061
+ },
+ {
+ "epoch": 14.754585152838429,
+ "grad_norm": 0.017482127994298935,
+ "learning_rate": 0.0006,
+ "loss": 4.62825870513916,
+ "step": 1062
+ },
+ {
+ "epoch": 14.768558951965066,
+ "grad_norm": 0.019018379971385002,
+ "learning_rate": 0.0006,
+ "loss": 4.693194389343262,
+ "step": 1063
+ },
+ {
+ "epoch": 14.782532751091702,
+ "grad_norm": 0.01993914321064949,
+ "learning_rate": 0.0006,
+ "loss": 4.742903709411621,
+ "step": 1064
+ },
+ {
+ "epoch": 14.796506550218341,
+ "grad_norm": 0.02301335707306862,
+ "learning_rate": 0.0006,
+ "loss": 4.613251686096191,
+ "step": 1065
+ },
+ {
+ "epoch": 14.810480349344978,
+ "grad_norm": 0.024088222533464432,
+ "learning_rate": 0.0006,
+ "loss": 4.594305992126465,
+ "step": 1066
+ },
+ {
+ "epoch": 14.824454148471617,
+ "grad_norm": 0.02493320405483246,
+ "learning_rate": 0.0006,
+ "loss": 4.700588226318359,
+ "step": 1067
+ },
+ {
+ "epoch": 14.838427947598253,
+ "grad_norm": 0.02501937560737133,
+ "learning_rate": 0.0006,
+ "loss": 4.601742744445801,
+ "step": 1068
+ },
+ {
+ "epoch": 14.85240174672489,
+ "grad_norm": 0.023490751162171364,
+ "learning_rate": 0.0006,
+ "loss": 4.589211463928223,
+ "step": 1069
+ },
+ {
+ "epoch": 14.866375545851529,
+ "grad_norm": 0.019695665687322617,
+ "learning_rate": 0.0006,
+ "loss": 4.635081768035889,
+ "step": 1070
+ },
+ {
+ "epoch": 14.880349344978166,
+ "grad_norm": 0.019276980310678482,
+ "learning_rate": 0.0006,
+ "loss": 4.737002372741699,
+ "step": 1071
+ },
+ {
+ "epoch": 14.894323144104803,
+ "grad_norm": 0.022430511191487312,
+ "learning_rate": 0.0006,
+ "loss": 4.808295726776123,
+ "step": 1072
+ },
+ {
+ "epoch": 14.908296943231441,
+ "grad_norm": 0.021543916314840317,
+ "learning_rate": 0.0006,
+ "loss": 4.559475421905518,
+ "step": 1073
+ },
+ {
+ "epoch": 14.922270742358078,
+ "grad_norm": 0.019885241985321045,
+ "learning_rate": 0.0006,
+ "loss": 4.647948741912842,
+ "step": 1074
+ },
+ {
+ "epoch": 14.936244541484717,
+ "grad_norm": 0.02115175500512123,
+ "learning_rate": 0.0006,
+ "loss": 4.680505752563477,
+ "step": 1075
+ },
+ {
+ "epoch": 14.950218340611354,
+ "grad_norm": 0.02156151458621025,
+ "learning_rate": 0.0006,
+ "loss": 4.673477649688721,
+ "step": 1076
+ },
+ {
+ "epoch": 14.96419213973799,
+ "grad_norm": 0.02184601128101349,
+ "learning_rate": 0.0006,
+ "loss": 4.644402027130127,
+ "step": 1077
+ },
+ {
+ "epoch": 14.97816593886463,
+ "grad_norm": 0.0211983360350132,
+ "learning_rate": 0.0006,
+ "loss": 4.821558475494385,
+ "step": 1078
+ },
+ {
+ "epoch": 14.992139737991266,
+ "grad_norm": 0.019636567682027817,
+ "learning_rate": 0.0006,
+ "loss": 4.578868865966797,
+ "step": 1079
+ },
+ {
+ "epoch": 15.0,
+ "grad_norm": 0.019997350871562958,
+ "learning_rate": 0.0006,
+ "loss": 4.691307544708252,
+ "step": 1080
+ },
+ {
+ "epoch": 15.0,
+ "eval_loss": 4.872478008270264,
+ "eval_runtime": 56.8513,
+ "eval_samples_per_second": 42.954,
+ "eval_steps_per_second": 1.354,
+ "step": 1080
+ },
+ {
+ "epoch": 15.013973799126637,
+ "grad_norm": 0.018830662593245506,
+ "learning_rate": 0.0006,
+ "loss": 4.761779308319092,
+ "step": 1081
+ },
+ {
+ "epoch": 15.027947598253276,
+ "grad_norm": 0.016876481473445892,
+ "learning_rate": 0.0006,
+ "loss": 4.61148738861084,
+ "step": 1082
+ },
+ {
+ "epoch": 15.041921397379912,
+ "grad_norm": 0.015550863929092884,
+ "learning_rate": 0.0006,
+ "loss": 4.676124095916748,
+ "step": 1083
+ },
+ {
+ "epoch": 15.055895196506551,
+ "grad_norm": 0.017178896814584732,
+ "learning_rate": 0.0006,
+ "loss": 4.588935852050781,
+ "step": 1084
+ },
+ {
+ "epoch": 15.069868995633188,
+ "grad_norm": 0.015465447679162025,
+ "learning_rate": 0.0006,
+ "loss": 4.605499267578125,
+ "step": 1085
+ },
+ {
+ "epoch": 15.083842794759825,
+ "grad_norm": 0.01606195978820324,
+ "learning_rate": 0.0006,
+ "loss": 4.723373889923096,
+ "step": 1086
+ },
+ {
+ "epoch": 15.097816593886463,
+ "grad_norm": 0.016521936282515526,
+ "learning_rate": 0.0006,
+ "loss": 4.646608352661133,
+ "step": 1087
+ },
+ {
+ "epoch": 15.1117903930131,
+ "grad_norm": 0.01589205674827099,
+ "learning_rate": 0.0006,
+ "loss": 4.498138427734375,
+ "step": 1088
+ },
+ {
+ "epoch": 15.125764192139737,
+ "grad_norm": 0.01524856686592102,
+ "learning_rate": 0.0006,
+ "loss": 4.594562530517578,
+ "step": 1089
+ },
+ {
+ "epoch": 15.139737991266376,
+ "grad_norm": 0.014714641496539116,
+ "learning_rate": 0.0006,
+ "loss": 4.630744934082031,
+ "step": 1090
+ },
+ {
+ "epoch": 15.153711790393013,
+ "grad_norm": 0.015333331190049648,
+ "learning_rate": 0.0006,
+ "loss": 4.682388782501221,
+ "step": 1091
+ },
+ {
+ "epoch": 15.167685589519651,
+ "grad_norm": 0.01550073828548193,
+ "learning_rate": 0.0006,
+ "loss": 4.692652702331543,
+ "step": 1092
+ },
+ {
+ "epoch": 15.181659388646288,
+ "grad_norm": 0.014783318154513836,
+ "learning_rate": 0.0006,
+ "loss": 4.6223835945129395,
+ "step": 1093
+ },
+ {
+ "epoch": 15.195633187772925,
+ "grad_norm": 0.014372751116752625,
+ "learning_rate": 0.0006,
+ "loss": 4.52687406539917,
+ "step": 1094
+ },
+ {
+ "epoch": 15.209606986899564,
+ "grad_norm": 0.014779130928218365,
+ "learning_rate": 0.0006,
+ "loss": 4.637009620666504,
+ "step": 1095
+ },
+ {
+ "epoch": 15.2235807860262,
+ "grad_norm": 0.015697073191404343,
+ "learning_rate": 0.0006,
+ "loss": 4.682981014251709,
+ "step": 1096
+ },
+ {
+ "epoch": 15.237554585152838,
+ "grad_norm": 0.016523664817214012,
+ "learning_rate": 0.0006,
+ "loss": 4.705722808837891,
+ "step": 1097
+ },
+ {
+ "epoch": 15.251528384279476,
+ "grad_norm": 0.017633996903896332,
+ "learning_rate": 0.0006,
+ "loss": 4.574686050415039,
+ "step": 1098
+ },
+ {
+ "epoch": 15.265502183406113,
+ "grad_norm": 0.019955076277256012,
+ "learning_rate": 0.0006,
+ "loss": 4.634450912475586,
+ "step": 1099
+ },
+ {
+ "epoch": 15.279475982532752,
+ "grad_norm": 0.02132618986070156,
+ "learning_rate": 0.0006,
+ "loss": 4.56110143661499,
+ "step": 1100
+ },
+ {
+ "epoch": 15.293449781659389,
+ "grad_norm": 0.0207593385130167,
+ "learning_rate": 0.0006,
+ "loss": 4.629396438598633,
+ "step": 1101
+ },
+ {
+ "epoch": 15.307423580786025,
+ "grad_norm": 0.017882540822029114,
+ "learning_rate": 0.0006,
+ "loss": 4.730963230133057,
+ "step": 1102
+ },
+ {
+ "epoch": 15.321397379912664,
+ "grad_norm": 0.01917876861989498,
+ "learning_rate": 0.0006,
+ "loss": 4.64585542678833,
+ "step": 1103
+ },
+ {
+ "epoch": 15.335371179039301,
+ "grad_norm": 0.01943974196910858,
+ "learning_rate": 0.0006,
+ "loss": 4.656181335449219,
+ "step": 1104
+ },
+ {
+ "epoch": 15.34934497816594,
+ "grad_norm": 0.019904715940356255,
+ "learning_rate": 0.0006,
+ "loss": 4.619329452514648,
+ "step": 1105
+ },
+ {
+ "epoch": 15.363318777292577,
+ "grad_norm": 0.019174346700310707,
+ "learning_rate": 0.0006,
+ "loss": 4.647892475128174,
+ "step": 1106
+ },
+ {
+ "epoch": 15.377292576419213,
+ "grad_norm": 0.018244758248329163,
+ "learning_rate": 0.0006,
+ "loss": 4.56166934967041,
+ "step": 1107
+ },
+ {
+ "epoch": 15.391266375545852,
+ "grad_norm": 0.018914205953478813,
+ "learning_rate": 0.0006,
+ "loss": 4.648063659667969,
+ "step": 1108
+ },
+ {
+ "epoch": 15.405240174672489,
+ "grad_norm": 0.01961071416735649,
+ "learning_rate": 0.0006,
+ "loss": 4.665525436401367,
+ "step": 1109
+ },
+ {
+ "epoch": 15.419213973799126,
+ "grad_norm": 0.01841311901807785,
+ "learning_rate": 0.0006,
+ "loss": 4.611342430114746,
+ "step": 1110
+ },
+ {
+ "epoch": 15.433187772925764,
+ "grad_norm": 0.01863427273929119,
+ "learning_rate": 0.0006,
+ "loss": 4.664256572723389,
+ "step": 1111
+ },
+ {
+ "epoch": 15.447161572052401,
+ "grad_norm": 0.020506877452135086,
+ "learning_rate": 0.0006,
+ "loss": 4.631237506866455,
+ "step": 1112
+ },
+ {
+ "epoch": 15.46113537117904,
+ "grad_norm": 0.02223833277821541,
+ "learning_rate": 0.0006,
+ "loss": 4.668654441833496,
+ "step": 1113
+ },
+ {
+ "epoch": 15.475109170305677,
+ "grad_norm": 0.023336702957749367,
+ "learning_rate": 0.0006,
+ "loss": 4.562833309173584,
+ "step": 1114
+ },
+ {
+ "epoch": 15.489082969432314,
+ "grad_norm": 0.023052413016557693,
+ "learning_rate": 0.0006,
+ "loss": 4.621532917022705,
+ "step": 1115
+ },
+ {
+ "epoch": 15.503056768558952,
+ "grad_norm": 0.023420924320816994,
+ "learning_rate": 0.0006,
+ "loss": 4.665465354919434,
+ "step": 1116
+ },
+ {
+ "epoch": 15.51703056768559,
+ "grad_norm": 0.024071281775832176,
+ "learning_rate": 0.0006,
+ "loss": 4.595980644226074,
+ "step": 1117
+ },
+ {
+ "epoch": 15.531004366812226,
+ "grad_norm": 0.023406682536005974,
+ "learning_rate": 0.0006,
+ "loss": 4.642635345458984,
+ "step": 1118
+ },
+ {
+ "epoch": 15.544978165938865,
+ "grad_norm": 0.02138647809624672,
+ "learning_rate": 0.0006,
+ "loss": 4.704513072967529,
+ "step": 1119
+ },
+ {
+ "epoch": 15.558951965065502,
+ "grad_norm": 0.0221384447067976,
+ "learning_rate": 0.0006,
+ "loss": 4.567503452301025,
+ "step": 1120
+ },
+ {
+ "epoch": 15.57292576419214,
+ "grad_norm": 0.020879851654171944,
+ "learning_rate": 0.0006,
+ "loss": 4.664057731628418,
+ "step": 1121
+ },
+ {
+ "epoch": 15.586899563318777,
+ "grad_norm": 0.021478967741131783,
+ "learning_rate": 0.0006,
+ "loss": 4.539627552032471,
+ "step": 1122
+ },
+ {
+ "epoch": 15.600873362445414,
+ "grad_norm": 0.022748656570911407,
+ "learning_rate": 0.0006,
+ "loss": 4.5654497146606445,
+ "step": 1123
+ },
+ {
+ "epoch": 15.614847161572053,
+ "grad_norm": 0.021870248019695282,
+ "learning_rate": 0.0006,
+ "loss": 4.56899356842041,
+ "step": 1124
+ },
+ {
+ "epoch": 15.62882096069869,
+ "grad_norm": 0.021711355075240135,
+ "learning_rate": 0.0006,
+ "loss": 4.657808303833008,
+ "step": 1125
+ },
+ {
+ "epoch": 15.642794759825328,
+ "grad_norm": 0.022810909897089005,
+ "learning_rate": 0.0006,
+ "loss": 4.6526689529418945,
+ "step": 1126
+ },
+ {
+ "epoch": 15.656768558951965,
+ "grad_norm": 0.022561343386769295,
+ "learning_rate": 0.0006,
+ "loss": 4.534887790679932,
+ "step": 1127
+ },
+ {
+ "epoch": 15.670742358078602,
+ "grad_norm": 0.019909802824258804,
+ "learning_rate": 0.0006,
+ "loss": 4.600790023803711,
+ "step": 1128
+ },
+ {
+ "epoch": 15.68471615720524,
+ "grad_norm": 0.020092325285077095,
+ "learning_rate": 0.0006,
+ "loss": 4.773830413818359,
+ "step": 1129
+ },
+ {
+ "epoch": 15.698689956331878,
+ "grad_norm": 0.018834445625543594,
+ "learning_rate": 0.0006,
+ "loss": 4.628868103027344,
+ "step": 1130
+ },
+ {
+ "epoch": 15.712663755458514,
+ "grad_norm": 0.017552798613905907,
+ "learning_rate": 0.0006,
+ "loss": 4.647500038146973,
+ "step": 1131
+ },
+ {
+ "epoch": 15.726637554585153,
+ "grad_norm": 0.01751251146197319,
+ "learning_rate": 0.0006,
+ "loss": 4.590545654296875,
+ "step": 1132
+ },
+ {
+ "epoch": 15.74061135371179,
+ "grad_norm": 0.01699099875986576,
+ "learning_rate": 0.0006,
+ "loss": 4.509699821472168,
+ "step": 1133
+ },
+ {
+ "epoch": 15.754585152838429,
+ "grad_norm": 0.015808649361133575,
+ "learning_rate": 0.0006,
+ "loss": 4.606992244720459,
+ "step": 1134
+ },
+ {
+ "epoch": 15.768558951965066,
+ "grad_norm": 0.01655443385243416,
+ "learning_rate": 0.0006,
+ "loss": 4.585219860076904,
+ "step": 1135
+ },
+ {
+ "epoch": 15.782532751091702,
+ "grad_norm": 0.01708163693547249,
+ "learning_rate": 0.0006,
+ "loss": 4.668147087097168,
+ "step": 1136
+ },
+ {
+ "epoch": 15.796506550218341,
+ "grad_norm": 0.017161913216114044,
+ "learning_rate": 0.0006,
+ "loss": 4.7170634269714355,
+ "step": 1137
+ },
+ {
+ "epoch": 15.810480349344978,
+ "grad_norm": 0.017262674868106842,
+ "learning_rate": 0.0006,
+ "loss": 4.572132110595703,
+ "step": 1138
+ },
+ {
+ "epoch": 15.824454148471617,
+ "grad_norm": 0.017713576555252075,
+ "learning_rate": 0.0006,
+ "loss": 4.702538967132568,
+ "step": 1139
+ },
+ {
+ "epoch": 15.838427947598253,
+ "grad_norm": 0.018730733543634415,
+ "learning_rate": 0.0006,
+ "loss": 4.629721641540527,
+ "step": 1140
+ },
+ {
+ "epoch": 15.85240174672489,
+ "grad_norm": 0.017960187047719955,
+ "learning_rate": 0.0006,
+ "loss": 4.661039352416992,
+ "step": 1141
+ },
+ {
+ "epoch": 15.866375545851529,
+ "grad_norm": 0.01750079356133938,
+ "learning_rate": 0.0006,
+ "loss": 4.503710746765137,
+ "step": 1142
+ },
+ {
+ "epoch": 15.880349344978166,
+ "grad_norm": 0.016630422323942184,
+ "learning_rate": 0.0006,
+ "loss": 4.6582136154174805,
+ "step": 1143
+ },
+ {
+ "epoch": 15.894323144104803,
+ "grad_norm": 0.015896832570433617,
+ "learning_rate": 0.0006,
+ "loss": 4.596693992614746,
+ "step": 1144
+ },
+ {
+ "epoch": 15.908296943231441,
+ "grad_norm": 0.01566407084465027,
+ "learning_rate": 0.0006,
+ "loss": 4.6337199211120605,
+ "step": 1145
+ },
+ {
+ "epoch": 15.922270742358078,
+ "grad_norm": 0.016823330894112587,
+ "learning_rate": 0.0006,
+ "loss": 4.686994552612305,
+ "step": 1146
+ },
+ {
+ "epoch": 15.936244541484717,
+ "grad_norm": 0.016208255663514137,
+ "learning_rate": 0.0006,
+ "loss": 4.6205549240112305,
+ "step": 1147
+ },
+ {
+ "epoch": 15.950218340611354,
+ "grad_norm": 0.0171707421541214,
+ "learning_rate": 0.0006,
+ "loss": 4.582968711853027,
+ "step": 1148
+ },
+ {
+ "epoch": 15.96419213973799,
+ "grad_norm": 0.017009446397423744,
+ "learning_rate": 0.0006,
+ "loss": 4.510274887084961,
+ "step": 1149
+ },
+ {
+ "epoch": 15.97816593886463,
+ "grad_norm": 0.01565811224281788,
+ "learning_rate": 0.0006,
+ "loss": 4.55834436416626,
+ "step": 1150
+ },
+ {
+ "epoch": 15.992139737991266,
+ "grad_norm": 0.015276037156581879,
+ "learning_rate": 0.0006,
+ "loss": 4.537870407104492,
+ "step": 1151
+ },
+ {
+ "epoch": 16.0,
+ "grad_norm": 0.017305882647633553,
+ "learning_rate": 0.0006,
+ "loss": 4.611305236816406,
+ "step": 1152
+ },
+ {
+ "epoch": 16.0,
+ "eval_loss": 4.775513172149658,
+ "eval_runtime": 56.7341,
+ "eval_samples_per_second": 43.043,
+ "eval_steps_per_second": 1.357,
+ "step": 1152
+ },
+ {
+ "epoch": 16.01397379912664,
+ "grad_norm": 0.016752436757087708,
+ "learning_rate": 0.0006,
+ "loss": 4.541312217712402,
+ "step": 1153
+ },
+ {
+ "epoch": 16.027947598253274,
+ "grad_norm": 0.01643134094774723,
+ "learning_rate": 0.0006,
+ "loss": 4.585402965545654,
+ "step": 1154
+ },
+ {
+ "epoch": 16.041921397379912,
+ "grad_norm": 0.01888720504939556,
+ "learning_rate": 0.0006,
+ "loss": 4.597825050354004,
+ "step": 1155
+ },
+ {
+ "epoch": 16.05589519650655,
+ "grad_norm": 0.022044632583856583,
+ "learning_rate": 0.0006,
+ "loss": 4.532772064208984,
+ "step": 1156
+ },
+ {
+ "epoch": 16.069868995633186,
+ "grad_norm": 0.022986987605690956,
+ "learning_rate": 0.0006,
+ "loss": 4.551351547241211,
+ "step": 1157
+ },
+ {
+ "epoch": 16.083842794759825,
+ "grad_norm": 0.020116791129112244,
+ "learning_rate": 0.0006,
+ "loss": 4.484149932861328,
+ "step": 1158
+ },
+ {
+ "epoch": 16.097816593886463,
+ "grad_norm": 0.021442247554659843,
+ "learning_rate": 0.0006,
+ "loss": 4.515316486358643,
+ "step": 1159
+ },
+ {
+ "epoch": 16.111790393013102,
+ "grad_norm": 0.020974772050976753,
+ "learning_rate": 0.0006,
+ "loss": 4.51606559753418,
+ "step": 1160
+ },
+ {
+ "epoch": 16.125764192139737,
+ "grad_norm": 0.021274864673614502,
+ "learning_rate": 0.0006,
+ "loss": 4.587329864501953,
+ "step": 1161
+ },
+ {
+ "epoch": 16.139737991266376,
+ "grad_norm": 0.022799784317612648,
+ "learning_rate": 0.0006,
+ "loss": 4.568149089813232,
+ "step": 1162
+ },
+ {
+ "epoch": 16.153711790393015,
+ "grad_norm": 0.021784933283925056,
+ "learning_rate": 0.0006,
+ "loss": 4.515336990356445,
+ "step": 1163
+ },
+ {
+ "epoch": 16.16768558951965,
+ "grad_norm": 0.024261271581053734,
+ "learning_rate": 0.0006,
+ "loss": 4.5127081871032715,
+ "step": 1164
+ },
+ {
+ "epoch": 16.18165938864629,
+ "grad_norm": 0.02248375490307808,
+ "learning_rate": 0.0006,
+ "loss": 4.754251003265381,
+ "step": 1165
+ },
+ {
+ "epoch": 16.195633187772927,
+ "grad_norm": 0.02136492356657982,
+ "learning_rate": 0.0006,
+ "loss": 4.58193826675415,
+ "step": 1166
+ },
+ {
+ "epoch": 16.209606986899562,
+ "grad_norm": 0.020727066323161125,
+ "learning_rate": 0.0006,
+ "loss": 4.564209938049316,
+ "step": 1167
+ },
+ {
+ "epoch": 16.2235807860262,
+ "grad_norm": 0.017789119854569435,
+ "learning_rate": 0.0006,
+ "loss": 4.698680877685547,
+ "step": 1168
+ },
+ {
+ "epoch": 16.23755458515284,
+ "grad_norm": 0.017603034153580666,
+ "learning_rate": 0.0006,
+ "loss": 4.501171588897705,
+ "step": 1169
+ },
+ {
+ "epoch": 16.251528384279474,
+ "grad_norm": 0.017345862463116646,
+ "learning_rate": 0.0006,
+ "loss": 4.593590259552002,
+ "step": 1170
+ },
+ {
+ "epoch": 16.265502183406113,
+ "grad_norm": 0.018047425895929337,
+ "learning_rate": 0.0006,
+ "loss": 4.597467422485352,
+ "step": 1171
+ },
+ {
+ "epoch": 16.27947598253275,
+ "grad_norm": 0.01947222836315632,
+ "learning_rate": 0.0006,
+ "loss": 4.5072221755981445,
+ "step": 1172
+ },
+ {
+ "epoch": 16.29344978165939,
+ "grad_norm": 0.019598500803112984,
+ "learning_rate": 0.0006,
+ "loss": 4.601351737976074,
+ "step": 1173
+ },
+ {
+ "epoch": 16.307423580786025,
+ "grad_norm": 0.020002514123916626,
+ "learning_rate": 0.0006,
+ "loss": 4.580431938171387,
+ "step": 1174
+ },
+ {
+ "epoch": 16.321397379912664,
+ "grad_norm": 0.0216530729085207,
+ "learning_rate": 0.0006,
+ "loss": 4.424681663513184,
+ "step": 1175
+ },
+ {
+ "epoch": 16.335371179039303,
+ "grad_norm": 0.01949235238134861,
+ "learning_rate": 0.0006,
+ "loss": 4.751883029937744,
+ "step": 1176
+ },
+ {
+ "epoch": 16.349344978165938,
+ "grad_norm": 0.017779016867280006,
+ "learning_rate": 0.0006,
+ "loss": 4.541490077972412,
+ "step": 1177
+ },
+ {
+ "epoch": 16.363318777292577,
+ "grad_norm": 0.01909121684730053,
+ "learning_rate": 0.0006,
+ "loss": 4.594141006469727,
+ "step": 1178
+ },
+ {
+ "epoch": 16.377292576419215,
+ "grad_norm": 0.019894849509000778,
+ "learning_rate": 0.0006,
+ "loss": 4.555070400238037,
+ "step": 1179
+ },
+ {
+ "epoch": 16.39126637554585,
+ "grad_norm": 0.01993032731115818,
+ "learning_rate": 0.0006,
+ "loss": 4.521401405334473,
+ "step": 1180
+ },
+ {
+ "epoch": 16.40524017467249,
+ "grad_norm": 0.0210788045078516,
+ "learning_rate": 0.0006,
+ "loss": 4.672842979431152,
+ "step": 1181
+ },
+ {
+ "epoch": 16.419213973799128,
+ "grad_norm": 0.02274598926305771,
+ "learning_rate": 0.0006,
+ "loss": 4.629504203796387,
+ "step": 1182
+ },
+ {
+ "epoch": 16.433187772925763,
+ "grad_norm": 0.020448120310902596,
+ "learning_rate": 0.0006,
+ "loss": 4.548209190368652,
+ "step": 1183
+ },
+ {
+ "epoch": 16.4471615720524,
+ "grad_norm": 0.017565643414855003,
+ "learning_rate": 0.0006,
+ "loss": 4.529054641723633,
+ "step": 1184
+ },
+ {
+ "epoch": 16.46113537117904,
+ "grad_norm": 0.019504237920045853,
+ "learning_rate": 0.0006,
+ "loss": 4.48182487487793,
+ "step": 1185
+ },
+ {
+ "epoch": 16.475109170305675,
+ "grad_norm": 0.016823742538690567,
+ "learning_rate": 0.0006,
+ "loss": 4.608696937561035,
+ "step": 1186
+ },
+ {
+ "epoch": 16.489082969432314,
+ "grad_norm": 0.01602894812822342,
+ "learning_rate": 0.0006,
+ "loss": 4.620062351226807,
+ "step": 1187
+ },
+ {
+ "epoch": 16.503056768558952,
+ "grad_norm": 0.016615642234683037,
+ "learning_rate": 0.0006,
+ "loss": 4.69378662109375,
+ "step": 1188
+ },
+ {
+ "epoch": 16.51703056768559,
+ "grad_norm": 0.018576730042696,
+ "learning_rate": 0.0006,
+ "loss": 4.537041664123535,
+ "step": 1189
+ },
+ {
+ "epoch": 16.531004366812226,
+ "grad_norm": 0.018611151725053787,
+ "learning_rate": 0.0006,
+ "loss": 4.496169090270996,
+ "step": 1190
+ },
+ {
+ "epoch": 16.544978165938865,
+ "grad_norm": 0.018622413277626038,
+ "learning_rate": 0.0006,
+ "loss": 4.557358264923096,
+ "step": 1191
+ },
+ {
+ "epoch": 16.558951965065503,
+ "grad_norm": 0.01823544315993786,
+ "learning_rate": 0.0006,
+ "loss": 4.730792999267578,
+ "step": 1192
+ },
+ {
+ "epoch": 16.57292576419214,
+ "grad_norm": 0.018796846270561218,
+ "learning_rate": 0.0006,
+ "loss": 4.781791687011719,
+ "step": 1193
+ },
+ {
+ "epoch": 16.586899563318777,
+ "grad_norm": 0.01728292554616928,
+ "learning_rate": 0.0006,
+ "loss": 4.515257835388184,
+ "step": 1194
+ },
+ {
+ "epoch": 16.600873362445416,
+ "grad_norm": 0.01699589006602764,
+ "learning_rate": 0.0006,
+ "loss": 4.5996904373168945,
+ "step": 1195
+ },
+ {
+ "epoch": 16.61484716157205,
+ "grad_norm": 0.015905970707535744,
+ "learning_rate": 0.0006,
+ "loss": 4.601301193237305,
+ "step": 1196
+ },
+ {
+ "epoch": 16.62882096069869,
+ "grad_norm": 0.01338571310043335,
+ "learning_rate": 0.0006,
+ "loss": 4.557708740234375,
+ "step": 1197
+ },
+ {
+ "epoch": 16.64279475982533,
+ "grad_norm": 0.015298855490982533,
+ "learning_rate": 0.0006,
+ "loss": 4.542287349700928,
+ "step": 1198
+ },
+ {
+ "epoch": 16.656768558951963,
+ "grad_norm": 0.014753367751836777,
+ "learning_rate": 0.0006,
+ "loss": 4.595787048339844,
+ "step": 1199
+ },
+ {
+ "epoch": 16.670742358078602,
+ "grad_norm": 0.015527483075857162,
+ "learning_rate": 0.0006,
+ "loss": 4.607529163360596,
+ "step": 1200
+ },
+ {
+ "epoch": 16.68471615720524,
+ "grad_norm": 0.015042162500321865,
+ "learning_rate": 0.0006,
+ "loss": 4.686626434326172,
+ "step": 1201
+ },
+ {
+ "epoch": 16.69868995633188,
+ "grad_norm": 0.0159393772482872,
+ "learning_rate": 0.0006,
+ "loss": 4.47656774520874,
+ "step": 1202
+ },
+ {
+ "epoch": 16.712663755458514,
+ "grad_norm": 0.01532787922769785,
+ "learning_rate": 0.0006,
+ "loss": 4.55678653717041,
+ "step": 1203
+ },
+ {
+ "epoch": 16.726637554585153,
+ "grad_norm": 0.016332164406776428,
+ "learning_rate": 0.0006,
+ "loss": 4.51441764831543,
+ "step": 1204
+ },
+ {
+ "epoch": 16.74061135371179,
+ "grad_norm": 0.018901726230978966,
+ "learning_rate": 0.0006,
+ "loss": 4.57227897644043,
+ "step": 1205
+ },
+ {
+ "epoch": 16.754585152838427,
+ "grad_norm": 0.018589433282613754,
+ "learning_rate": 0.0006,
+ "loss": 4.550070762634277,
+ "step": 1206
+ },
+ {
+ "epoch": 16.768558951965066,
+ "grad_norm": 0.01634034886956215,
+ "learning_rate": 0.0006,
+ "loss": 4.586910724639893,
+ "step": 1207
+ },
+ {
+ "epoch": 16.782532751091704,
+ "grad_norm": 0.017599433660507202,
+ "learning_rate": 0.0006,
+ "loss": 4.5891923904418945,
+ "step": 1208
+ },
+ {
+ "epoch": 16.79650655021834,
+ "grad_norm": 0.01878446340560913,
+ "learning_rate": 0.0006,
+ "loss": 4.57227087020874,
+ "step": 1209
+ },
+ {
+ "epoch": 16.810480349344978,
+ "grad_norm": 0.01803469844162464,
+ "learning_rate": 0.0006,
+ "loss": 4.583845138549805,
+ "step": 1210
+ },
+ {
+ "epoch": 16.824454148471617,
+ "grad_norm": 0.018505144864320755,
+ "learning_rate": 0.0006,
+ "loss": 4.561151504516602,
+ "step": 1211
+ },
+ {
+ "epoch": 16.83842794759825,
+ "grad_norm": 0.01920081488788128,
+ "learning_rate": 0.0006,
+ "loss": 4.506834030151367,
+ "step": 1212
+ },
+ {
+ "epoch": 16.85240174672489,
+ "grad_norm": 0.0177712831646204,
+ "learning_rate": 0.0006,
+ "loss": 4.557195663452148,
+ "step": 1213
+ },
+ {
+ "epoch": 16.86637554585153,
+ "grad_norm": 0.019113918766379356,
+ "learning_rate": 0.0006,
+ "loss": 4.651227951049805,
+ "step": 1214
+ },
+ {
+ "epoch": 16.880349344978168,
+ "grad_norm": 0.02012539468705654,
+ "learning_rate": 0.0006,
+ "loss": 4.4952239990234375,
+ "step": 1215
+ },
+ {
+ "epoch": 16.894323144104803,
+ "grad_norm": 0.019993646070361137,
+ "learning_rate": 0.0006,
+ "loss": 4.517433166503906,
+ "step": 1216
+ },
+ {
+ "epoch": 16.90829694323144,
+ "grad_norm": 0.020041441544890404,
+ "learning_rate": 0.0006,
+ "loss": 4.489407539367676,
+ "step": 1217
+ },
+ {
+ "epoch": 16.92227074235808,
+ "grad_norm": 0.0183611661195755,
+ "learning_rate": 0.0006,
+ "loss": 4.5435638427734375,
+ "step": 1218
+ },
+ {
+ "epoch": 16.936244541484715,
+ "grad_norm": 0.019138429313898087,
+ "learning_rate": 0.0006,
+ "loss": 4.545949459075928,
+ "step": 1219
+ },
+ {
+ "epoch": 16.950218340611354,
+ "grad_norm": 0.018554266542196274,
+ "learning_rate": 0.0006,
+ "loss": 4.579620361328125,
+ "step": 1220
+ },
+ {
+ "epoch": 16.964192139737992,
+ "grad_norm": 0.0193545650690794,
+ "learning_rate": 0.0006,
+ "loss": 4.588255405426025,
+ "step": 1221
+ },
+ {
+ "epoch": 16.978165938864628,
+ "grad_norm": 0.019389165565371513,
+ "learning_rate": 0.0006,
+ "loss": 4.520698547363281,
+ "step": 1222
+ },
+ {
+ "epoch": 16.992139737991266,
+ "grad_norm": 0.016601379960775375,
+ "learning_rate": 0.0006,
+ "loss": 4.4809064865112305,
+ "step": 1223
+ },
+ {
+ "epoch": 17.0,
+ "grad_norm": 0.01771303080022335,
+ "learning_rate": 0.0006,
+ "loss": 4.721122741699219,
+ "step": 1224
+ },
+ {
+ "epoch": 17.0,
+ "eval_loss": 4.7903666496276855,
+ "eval_runtime": 56.9826,
+ "eval_samples_per_second": 42.855,
+ "eval_steps_per_second": 1.351,
+ "step": 1224
+ },
+ {
+ "epoch": 17.01397379912664,
+ "grad_norm": 0.017720289528369904,
+ "learning_rate": 0.0006,
+ "loss": 4.523834705352783,
+ "step": 1225
+ },
+ {
+ "epoch": 17.027947598253274,
+ "grad_norm": 0.020137161016464233,
+ "learning_rate": 0.0006,
+ "loss": 4.574061870574951,
+ "step": 1226
+ },
+ {
+ "epoch": 17.041921397379912,
+ "grad_norm": 0.02002314105629921,
+ "learning_rate": 0.0006,
+ "loss": 4.512407302856445,
+ "step": 1227
+ },
+ {
+ "epoch": 17.05589519650655,
+ "grad_norm": 0.019628183916211128,
+ "learning_rate": 0.0006,
+ "loss": 4.502533435821533,
+ "step": 1228
+ },
+ {
+ "epoch": 17.069868995633186,
+ "grad_norm": 0.019808458164334297,
+ "learning_rate": 0.0006,
+ "loss": 4.463156700134277,
+ "step": 1229
+ },
+ {
+ "epoch": 17.083842794759825,
+ "grad_norm": 0.021250786259770393,
+ "learning_rate": 0.0006,
+ "loss": 4.622682094573975,
+ "step": 1230
+ },
+ {
+ "epoch": 17.097816593886463,
+ "grad_norm": 0.022031694650650024,
+ "learning_rate": 0.0006,
+ "loss": 4.451882839202881,
+ "step": 1231
+ },
+ {
+ "epoch": 17.111790393013102,
+ "grad_norm": 0.01898987777531147,
+ "learning_rate": 0.0006,
+ "loss": 4.529331684112549,
+ "step": 1232
+ },
+ {
+ "epoch": 17.125764192139737,
+ "grad_norm": 0.018749834969639778,
+ "learning_rate": 0.0006,
+ "loss": 4.435815334320068,
+ "step": 1233
+ },
+ {
+ "epoch": 17.139737991266376,
+ "grad_norm": 0.01896674931049347,
+ "learning_rate": 0.0006,
+ "loss": 4.484029293060303,
+ "step": 1234
+ },
+ {
+ "epoch": 17.153711790393015,
+ "grad_norm": 0.018286997452378273,
+ "learning_rate": 0.0006,
+ "loss": 4.523510932922363,
+ "step": 1235
+ },
+ {
+ "epoch": 17.16768558951965,
+ "grad_norm": 0.017273059114813805,
+ "learning_rate": 0.0006,
+ "loss": 4.451767444610596,
+ "step": 1236
+ },
+ {
+ "epoch": 17.18165938864629,
+ "grad_norm": 0.016499150544404984,
+ "learning_rate": 0.0006,
+ "loss": 4.5568647384643555,
+ "step": 1237
+ },
+ {
+ "epoch": 17.195633187772927,
+ "grad_norm": 0.015934469178318977,
+ "learning_rate": 0.0006,
+ "loss": 4.4639997482299805,
+ "step": 1238
+ },
+ {
+ "epoch": 17.209606986899562,
+ "grad_norm": 0.01714209094643593,
+ "learning_rate": 0.0006,
+ "loss": 4.547614097595215,
+ "step": 1239
+ },
+ {
+ "epoch": 17.2235807860262,
+ "grad_norm": 0.019167495891451836,
+ "learning_rate": 0.0006,
+ "loss": 4.571097373962402,
+ "step": 1240
+ },
+ {
+ "epoch": 17.23755458515284,
+ "grad_norm": 0.018276996910572052,
+ "learning_rate": 0.0006,
+ "loss": 4.489764213562012,
+ "step": 1241
+ },
+ {
+ "epoch": 17.251528384279474,
+ "grad_norm": 0.021296866238117218,
+ "learning_rate": 0.0006,
+ "loss": 4.526913642883301,
+ "step": 1242
+ },
+ {
+ "epoch": 17.265502183406113,
+ "grad_norm": 0.022606629878282547,
+ "learning_rate": 0.0006,
+ "loss": 4.521624565124512,
+ "step": 1243
+ },
+ {
+ "epoch": 17.27947598253275,
+ "grad_norm": 0.021204711869359016,
+ "learning_rate": 0.0006,
+ "loss": 4.500864028930664,
+ "step": 1244
+ },
+ {
+ "epoch": 17.29344978165939,
+ "grad_norm": 0.01988849975168705,
+ "learning_rate": 0.0006,
+ "loss": 4.45162296295166,
+ "step": 1245
+ },
+ {
+ "epoch": 17.307423580786025,
+ "grad_norm": 0.020268384367227554,
+ "learning_rate": 0.0006,
+ "loss": 4.536198616027832,
+ "step": 1246
+ },
+ {
+ "epoch": 17.321397379912664,
+ "grad_norm": 0.020185677334666252,
+ "learning_rate": 0.0006,
+ "loss": 4.440262794494629,
+ "step": 1247
+ },
+ {
+ "epoch": 17.335371179039303,
+ "grad_norm": 0.020106853917241096,
+ "learning_rate": 0.0006,
+ "loss": 4.475907325744629,
+ "step": 1248
+ },
+ {
+ "epoch": 17.349344978165938,
+ "grad_norm": 0.018883297219872475,
+ "learning_rate": 0.0006,
+ "loss": 4.640661716461182,
+ "step": 1249
+ },
+ {
+ "epoch": 17.363318777292577,
+ "grad_norm": 0.01852620206773281,
+ "learning_rate": 0.0006,
+ "loss": 4.578243255615234,
+ "step": 1250
+ },
+ {
+ "epoch": 17.377292576419215,
+ "grad_norm": 0.01886451430618763,
+ "learning_rate": 0.0006,
+ "loss": 4.546026229858398,
+ "step": 1251
+ },
+ {
+ "epoch": 17.39126637554585,
+ "grad_norm": 0.01777452416718006,
+ "learning_rate": 0.0006,
+ "loss": 4.521542549133301,
+ "step": 1252
+ },
+ {
+ "epoch": 17.40524017467249,
+ "grad_norm": 0.016626447439193726,
+ "learning_rate": 0.0006,
+ "loss": 4.52629280090332,
+ "step": 1253
+ },
+ {
+ "epoch": 17.419213973799128,
+ "grad_norm": 0.01599477231502533,
+ "learning_rate": 0.0006,
+ "loss": 4.4796037673950195,
+ "step": 1254
+ },
+ {
+ "epoch": 17.433187772925763,
+ "grad_norm": 0.017021458595991135,
+ "learning_rate": 0.0006,
+ "loss": 4.468813419342041,
+ "step": 1255
+ },
+ {
+ "epoch": 17.4471615720524,
+ "grad_norm": 0.016925333067774773,
+ "learning_rate": 0.0006,
+ "loss": 4.584898948669434,
+ "step": 1256
+ },
+ {
+ "epoch": 17.46113537117904,
+ "grad_norm": 0.01714998669922352,
+ "learning_rate": 0.0006,
+ "loss": 4.523544788360596,
+ "step": 1257
+ },
+ {
+ "epoch": 17.475109170305675,
+ "grad_norm": 0.01802288182079792,
+ "learning_rate": 0.0006,
+ "loss": 4.595581531524658,
+ "step": 1258
+ },
+ {
+ "epoch": 17.489082969432314,
+ "grad_norm": 0.0180289875715971,
+ "learning_rate": 0.0006,
+ "loss": 4.394440650939941,
+ "step": 1259
+ },
+ {
+ "epoch": 17.503056768558952,
+ "grad_norm": 0.018504245206713676,
+ "learning_rate": 0.0006,
+ "loss": 4.4794416427612305,
+ "step": 1260
+ },
+ {
+ "epoch": 17.51703056768559,
+ "grad_norm": 0.019973233342170715,
+ "learning_rate": 0.0006,
+ "loss": 4.4241437911987305,
+ "step": 1261
+ },
+ {
+ "epoch": 17.531004366812226,
+ "grad_norm": 0.020063083618879318,
+ "learning_rate": 0.0006,
+ "loss": 4.48068380355835,
+ "step": 1262
+ },
+ {
+ "epoch": 17.544978165938865,
+ "grad_norm": 0.02066640742123127,
+ "learning_rate": 0.0006,
+ "loss": 4.628993034362793,
+ "step": 1263
+ },
+ {
+ "epoch": 17.558951965065503,
+ "grad_norm": 0.02206416428089142,
+ "learning_rate": 0.0006,
+ "loss": 4.564790725708008,
+ "step": 1264
+ },
+ {
+ "epoch": 17.57292576419214,
+ "grad_norm": 0.020640335977077484,
+ "learning_rate": 0.0006,
+ "loss": 4.578705787658691,
+ "step": 1265
+ },
+ {
+ "epoch": 17.586899563318777,
+ "grad_norm": 0.019093405455350876,
+ "learning_rate": 0.0006,
+ "loss": 4.5023603439331055,
+ "step": 1266
+ },
+ {
+ "epoch": 17.600873362445416,
+ "grad_norm": 0.019459193572402,
+ "learning_rate": 0.0006,
+ "loss": 4.594531059265137,
+ "step": 1267
+ },
+ {
+ "epoch": 17.61484716157205,
+ "grad_norm": 0.02021171525120735,
+ "learning_rate": 0.0006,
+ "loss": 4.770299911499023,
+ "step": 1268
+ },
+ {
+ "epoch": 17.62882096069869,
+ "grad_norm": 0.021161191165447235,
+ "learning_rate": 0.0006,
+ "loss": 4.5380401611328125,
+ "step": 1269
+ },
+ {
+ "epoch": 17.64279475982533,
+ "grad_norm": 0.021013403311371803,
+ "learning_rate": 0.0006,
+ "loss": 4.455524921417236,
+ "step": 1270
+ },
+ {
+ "epoch": 17.656768558951963,
+ "grad_norm": 0.020336592569947243,
+ "learning_rate": 0.0006,
+ "loss": 4.533672332763672,
+ "step": 1271
+ },
+ {
+ "epoch": 17.670742358078602,
+ "grad_norm": 0.020159434527158737,
+ "learning_rate": 0.0006,
+ "loss": 4.596151351928711,
+ "step": 1272
+ },
+ {
+ "epoch": 17.68471615720524,
+ "grad_norm": 0.020432960242033005,
+ "learning_rate": 0.0006,
+ "loss": 4.55531644821167,
+ "step": 1273
+ },
+ {
+ "epoch": 17.69868995633188,
+ "grad_norm": 0.025466229766607285,
+ "learning_rate": 0.0006,
+ "loss": 4.566912651062012,
+ "step": 1274
+ },
+ {
+ "epoch": 17.712663755458514,
+ "grad_norm": 0.026537558063864708,
+ "learning_rate": 0.0006,
+ "loss": 4.526003837585449,
+ "step": 1275
+ },
+ {
+ "epoch": 17.726637554585153,
+ "grad_norm": 0.024397607892751694,
+ "learning_rate": 0.0006,
+ "loss": 4.541755676269531,
+ "step": 1276
+ },
+ {
+ "epoch": 17.74061135371179,
+ "grad_norm": 0.022062312811613083,
+ "learning_rate": 0.0006,
+ "loss": 4.6532793045043945,
+ "step": 1277
+ },
+ {
+ "epoch": 17.754585152838427,
+ "grad_norm": 0.02204282023012638,
+ "learning_rate": 0.0006,
+ "loss": 4.582825660705566,
+ "step": 1278
+ },
+ {
+ "epoch": 17.768558951965066,
+ "grad_norm": 0.020844416692852974,
+ "learning_rate": 0.0006,
+ "loss": 4.522336483001709,
+ "step": 1279
+ },
+ {
+ "epoch": 17.782532751091704,
+ "grad_norm": 0.019815703853964806,
+ "learning_rate": 0.0006,
+ "loss": 4.607398509979248,
+ "step": 1280
+ },
+ {
+ "epoch": 17.79650655021834,
+ "grad_norm": 0.016676954925060272,
+ "learning_rate": 0.0006,
+ "loss": 4.446206092834473,
+ "step": 1281
+ },
+ {
+ "epoch": 17.810480349344978,
+ "grad_norm": 0.0166452769190073,
+ "learning_rate": 0.0006,
+ "loss": 4.505931854248047,
+ "step": 1282
+ },
+ {
+ "epoch": 17.824454148471617,
+ "grad_norm": 0.015928907319903374,
+ "learning_rate": 0.0006,
+ "loss": 4.455267429351807,
+ "step": 1283
+ },
+ {
+ "epoch": 17.83842794759825,
+ "grad_norm": 0.015219560824334621,
+ "learning_rate": 0.0006,
+ "loss": 4.5623674392700195,
+ "step": 1284
+ },
+ {
+ "epoch": 17.85240174672489,
+ "grad_norm": 0.015125355683267117,
+ "learning_rate": 0.0006,
+ "loss": 4.554104804992676,
+ "step": 1285
+ },
+ {
+ "epoch": 17.86637554585153,
+ "grad_norm": 0.015612718649208546,
+ "learning_rate": 0.0006,
+ "loss": 4.484847068786621,
+ "step": 1286
+ },
+ {
+ "epoch": 17.880349344978168,
+ "grad_norm": 0.015215662308037281,
+ "learning_rate": 0.0006,
+ "loss": 4.4802093505859375,
+ "step": 1287
+ },
+ {
+ "epoch": 17.894323144104803,
+ "grad_norm": 0.01511684525758028,
+ "learning_rate": 0.0006,
+ "loss": 4.537298679351807,
+ "step": 1288
+ },
+ {
+ "epoch": 17.90829694323144,
+ "grad_norm": 0.013841322623193264,
+ "learning_rate": 0.0006,
+ "loss": 4.427488327026367,
+ "step": 1289
+ },
+ {
+ "epoch": 17.92227074235808,
+ "grad_norm": 0.015115964226424694,
+ "learning_rate": 0.0006,
+ "loss": 4.4547119140625,
+ "step": 1290
+ },
+ {
+ "epoch": 17.936244541484715,
+ "grad_norm": 0.016411006450653076,
+ "learning_rate": 0.0006,
+ "loss": 4.52994966506958,
+ "step": 1291
+ },
+ {
+ "epoch": 17.950218340611354,
+ "grad_norm": 0.017313152551651,
+ "learning_rate": 0.0006,
+ "loss": 4.5337653160095215,
+ "step": 1292
+ },
+ {
+ "epoch": 17.964192139737992,
+ "grad_norm": 0.016559796407818794,
+ "learning_rate": 0.0006,
+ "loss": 4.5277910232543945,
+ "step": 1293
+ },
+ {
+ "epoch": 17.978165938864628,
+ "grad_norm": 0.015854761004447937,
+ "learning_rate": 0.0006,
+ "loss": 4.6334309577941895,
+ "step": 1294
+ },
+ {
+ "epoch": 17.992139737991266,
+ "grad_norm": 0.015658238902688026,
+ "learning_rate": 0.0006,
+ "loss": 4.613372802734375,
+ "step": 1295
+ },
+ {
+ "epoch": 18.0,
+ "grad_norm": 0.0169155802577734,
+ "learning_rate": 0.0006,
+ "loss": 4.527356147766113,
+ "step": 1296
+ },
+ {
+ "epoch": 18.0,
+ "eval_loss": 4.764543056488037,
+ "eval_runtime": 56.5998,
+ "eval_samples_per_second": 43.145,
+ "eval_steps_per_second": 1.36,
+ "step": 1296
+ },
+ {
+ "epoch": 18.01397379912664,
+ "grad_norm": 0.015348963439464569,
+ "learning_rate": 0.0006,
+ "loss": 4.466423988342285,
+ "step": 1297
+ },
+ {
+ "epoch": 18.027947598253274,
+ "grad_norm": 0.01768588088452816,
+ "learning_rate": 0.0006,
+ "loss": 4.545933246612549,
+ "step": 1298
+ },
+ {
+ "epoch": 18.041921397379912,
+ "grad_norm": 0.017381226643919945,
+ "learning_rate": 0.0006,
+ "loss": 4.494060039520264,
+ "step": 1299
+ },
+ {
+ "epoch": 18.05589519650655,
+ "grad_norm": 0.019002335146069527,
+ "learning_rate": 0.0006,
+ "loss": 4.395200729370117,
+ "step": 1300
+ },
+ {
+ "epoch": 18.069868995633186,
+ "grad_norm": 0.01827932707965374,
+ "learning_rate": 0.0006,
+ "loss": 4.424848556518555,
+ "step": 1301
+ },
+ {
+ "epoch": 18.083842794759825,
+ "grad_norm": 0.016383031383156776,
+ "learning_rate": 0.0006,
+ "loss": 4.494722366333008,
+ "step": 1302
+ },
+ {
+ "epoch": 18.097816593886463,
+ "grad_norm": 0.016265157610177994,
+ "learning_rate": 0.0006,
+ "loss": 4.416522979736328,
+ "step": 1303
+ },
+ {
+ "epoch": 18.111790393013102,
+ "grad_norm": 0.01702386513352394,
+ "learning_rate": 0.0006,
+ "loss": 4.600103378295898,
+ "step": 1304
+ },
+ {
+ "epoch": 18.125764192139737,
+ "grad_norm": 0.018495498225092888,
+ "learning_rate": 0.0006,
+ "loss": 4.588401794433594,
+ "step": 1305
+ },
+ {
+ "epoch": 18.139737991266376,
+ "grad_norm": 0.017921684309840202,
+ "learning_rate": 0.0006,
+ "loss": 4.478075981140137,
+ "step": 1306
+ },
+ {
+ "epoch": 18.153711790393015,
+ "grad_norm": 0.01801278069615364,
+ "learning_rate": 0.0006,
+ "loss": 4.413186550140381,
+ "step": 1307
+ },
+ {
+ "epoch": 18.16768558951965,
+ "grad_norm": 0.017211275175213814,
+ "learning_rate": 0.0006,
+ "loss": 4.461341857910156,
+ "step": 1308
+ },
+ {
+ "epoch": 18.18165938864629,
+ "grad_norm": 0.017589328810572624,
+ "learning_rate": 0.0006,
+ "loss": 4.546969413757324,
+ "step": 1309
+ },
+ {
+ "epoch": 18.195633187772927,
+ "grad_norm": 0.01797698810696602,
+ "learning_rate": 0.0006,
+ "loss": 4.530797004699707,
+ "step": 1310
+ },
+ {
+ "epoch": 18.209606986899562,
+ "grad_norm": 0.020408619195222855,
+ "learning_rate": 0.0006,
+ "loss": 4.62999153137207,
+ "step": 1311
+ },
+ {
+ "epoch": 18.2235807860262,
+ "grad_norm": 0.01994038000702858,
+ "learning_rate": 0.0006,
+ "loss": 4.586811065673828,
+ "step": 1312
+ },
+ {
+ "epoch": 18.23755458515284,
+ "grad_norm": 0.019381340593099594,
+ "learning_rate": 0.0006,
+ "loss": 4.4769368171691895,
+ "step": 1313
+ },
+ {
+ "epoch": 18.251528384279474,
+ "grad_norm": 0.020484555512666702,
+ "learning_rate": 0.0006,
+ "loss": 4.524302959442139,
+ "step": 1314
+ },
+ {
+ "epoch": 18.265502183406113,
+ "grad_norm": 0.02426416426897049,
+ "learning_rate": 0.0006,
+ "loss": 4.561705589294434,
+ "step": 1315
+ },
+ {
+ "epoch": 18.27947598253275,
+ "grad_norm": 0.024748101830482483,
+ "learning_rate": 0.0006,
+ "loss": 4.4654316902160645,
+ "step": 1316
+ },
+ {
+ "epoch": 18.29344978165939,
+ "grad_norm": 0.020703328773379326,
+ "learning_rate": 0.0006,
+ "loss": 4.53033447265625,
+ "step": 1317
+ },
+ {
+ "epoch": 18.307423580786025,
+ "grad_norm": 0.01765458658337593,
+ "learning_rate": 0.0006,
+ "loss": 4.5154266357421875,
+ "step": 1318
+ },
+ {
+ "epoch": 18.321397379912664,
+ "grad_norm": 0.018646420910954475,
+ "learning_rate": 0.0006,
+ "loss": 4.5224714279174805,
+ "step": 1319
+ },
+ {
+ "epoch": 18.335371179039303,
+ "grad_norm": 0.01907474733889103,
+ "learning_rate": 0.0006,
+ "loss": 4.5055437088012695,
+ "step": 1320
+ },
+ {
+ "epoch": 18.349344978165938,
+ "grad_norm": 0.018179846927523613,
+ "learning_rate": 0.0006,
+ "loss": 4.462161064147949,
+ "step": 1321
+ },
+ {
+ "epoch": 18.363318777292577,
+ "grad_norm": 0.016697878018021584,
+ "learning_rate": 0.0006,
+ "loss": 4.473718643188477,
+ "step": 1322
+ },
+ {
+ "epoch": 18.377292576419215,
+ "grad_norm": 0.017504019662737846,
+ "learning_rate": 0.0006,
+ "loss": 4.474348068237305,
+ "step": 1323
+ },
+ {
+ "epoch": 18.39126637554585,
+ "grad_norm": 0.018889913335442543,
+ "learning_rate": 0.0006,
+ "loss": 4.4531636238098145,
+ "step": 1324
+ },
+ {
+ "epoch": 18.40524017467249,
+ "grad_norm": 0.019167251884937286,
+ "learning_rate": 0.0006,
+ "loss": 4.4278364181518555,
+ "step": 1325
+ },
+ {
+ "epoch": 18.419213973799128,
+ "grad_norm": 0.015895741060376167,
+ "learning_rate": 0.0006,
+ "loss": 4.474592208862305,
+ "step": 1326
+ },
+ {
+ "epoch": 18.433187772925763,
+ "grad_norm": 0.01531882956624031,
+ "learning_rate": 0.0006,
+ "loss": 4.403757572174072,
+ "step": 1327
+ },
+ {
+ "epoch": 18.4471615720524,
+ "grad_norm": 0.01565849967300892,
+ "learning_rate": 0.0006,
+ "loss": 4.507664680480957,
+ "step": 1328
+ },
+ {
+ "epoch": 18.46113537117904,
+ "grad_norm": 0.01602315716445446,
+ "learning_rate": 0.0006,
+ "loss": 4.5648908615112305,
+ "step": 1329
+ },
+ {
+ "epoch": 18.475109170305675,
+ "grad_norm": 0.016467366367578506,
+ "learning_rate": 0.0006,
+ "loss": 4.428762912750244,
+ "step": 1330
+ },
+ {
+ "epoch": 18.489082969432314,
+ "grad_norm": 0.01604127697646618,
+ "learning_rate": 0.0006,
+ "loss": 4.527771472930908,
+ "step": 1331
+ },
+ {
+ "epoch": 18.503056768558952,
+ "grad_norm": 0.018825586885213852,
+ "learning_rate": 0.0006,
+ "loss": 4.485386848449707,
+ "step": 1332
+ },
+ {
+ "epoch": 18.51703056768559,
+ "grad_norm": 0.019258007407188416,
+ "learning_rate": 0.0006,
+ "loss": 4.414972305297852,
+ "step": 1333
+ },
+ {
+ "epoch": 18.531004366812226,
+ "grad_norm": 0.018192021176218987,
+ "learning_rate": 0.0006,
+ "loss": 4.412166595458984,
+ "step": 1334
+ },
+ {
+ "epoch": 18.544978165938865,
+ "grad_norm": 0.016120698302984238,
+ "learning_rate": 0.0006,
+ "loss": 4.484317779541016,
+ "step": 1335
+ },
+ {
+ "epoch": 18.558951965065503,
+ "grad_norm": 0.016496622934937477,
+ "learning_rate": 0.0006,
+ "loss": 4.514986515045166,
+ "step": 1336
+ },
+ {
+ "epoch": 18.57292576419214,
+ "grad_norm": 0.01795116998255253,
+ "learning_rate": 0.0006,
+ "loss": 4.41473388671875,
+ "step": 1337
+ },
+ {
+ "epoch": 18.586899563318777,
+ "grad_norm": 0.019062276929616928,
+ "learning_rate": 0.0006,
+ "loss": 4.601996898651123,
+ "step": 1338
+ },
+ {
+ "epoch": 18.600873362445416,
+ "grad_norm": 0.017174631357192993,
+ "learning_rate": 0.0006,
+ "loss": 4.571218490600586,
+ "step": 1339
+ },
+ {
+ "epoch": 18.61484716157205,
+ "grad_norm": 0.014015606604516506,
+ "learning_rate": 0.0006,
+ "loss": 4.526606559753418,
+ "step": 1340
+ },
+ {
+ "epoch": 18.62882096069869,
+ "grad_norm": 0.015358510427176952,
+ "learning_rate": 0.0006,
+ "loss": 4.604286193847656,
+ "step": 1341
+ },
+ {
+ "epoch": 18.64279475982533,
+ "grad_norm": 0.01619391329586506,
+ "learning_rate": 0.0006,
+ "loss": 4.560883522033691,
+ "step": 1342
+ },
+ {
+ "epoch": 18.656768558951963,
+ "grad_norm": 0.015972964465618134,
+ "learning_rate": 0.0006,
+ "loss": 4.479755401611328,
+ "step": 1343
+ },
+ {
+ "epoch": 18.670742358078602,
+ "grad_norm": 0.016861025243997574,
+ "learning_rate": 0.0006,
+ "loss": 4.474471092224121,
+ "step": 1344
+ },
+ {
+ "epoch": 18.68471615720524,
+ "grad_norm": 0.01822078227996826,
+ "learning_rate": 0.0006,
+ "loss": 4.509489059448242,
+ "step": 1345
+ },
+ {
+ "epoch": 18.69868995633188,
+ "grad_norm": 0.019415007904171944,
+ "learning_rate": 0.0006,
+ "loss": 4.508941173553467,
+ "step": 1346
+ },
+ {
+ "epoch": 18.712663755458514,
+ "grad_norm": 0.019856126978993416,
+ "learning_rate": 0.0006,
+ "loss": 4.446670055389404,
+ "step": 1347
+ },
+ {
+ "epoch": 18.726637554585153,
+ "grad_norm": 0.01842520199716091,
+ "learning_rate": 0.0006,
+ "loss": 4.524422645568848,
+ "step": 1348
+ },
+ {
+ "epoch": 18.74061135371179,
+ "grad_norm": 0.017196480184793472,
+ "learning_rate": 0.0006,
+ "loss": 4.498466968536377,
+ "step": 1349
+ },
+ {
+ "epoch": 18.754585152838427,
+ "grad_norm": 0.01823010854423046,
+ "learning_rate": 0.0006,
+ "loss": 4.534684181213379,
+ "step": 1350
+ },
+ {
+ "epoch": 18.768558951965066,
+ "grad_norm": 0.020398065447807312,
+ "learning_rate": 0.0006,
+ "loss": 4.456494331359863,
+ "step": 1351
+ },
+ {
+ "epoch": 18.782532751091704,
+ "grad_norm": 0.022992262616753578,
+ "learning_rate": 0.0006,
+ "loss": 4.476374626159668,
+ "step": 1352
+ },
+ {
+ "epoch": 18.79650655021834,
+ "grad_norm": 0.02235420234501362,
+ "learning_rate": 0.0006,
+ "loss": 4.587776184082031,
+ "step": 1353
+ },
+ {
+ "epoch": 18.810480349344978,
+ "grad_norm": 0.02093626745045185,
+ "learning_rate": 0.0006,
+ "loss": 4.424172401428223,
+ "step": 1354
+ },
+ {
+ "epoch": 18.824454148471617,
+ "grad_norm": 0.01997302658855915,
+ "learning_rate": 0.0006,
+ "loss": 4.448066711425781,
+ "step": 1355
+ },
+ {
+ "epoch": 18.83842794759825,
+ "grad_norm": 0.018265459686517715,
+ "learning_rate": 0.0006,
+ "loss": 4.436610698699951,
+ "step": 1356
+ },
+ {
+ "epoch": 18.85240174672489,
+ "grad_norm": 0.01673833839595318,
+ "learning_rate": 0.0006,
+ "loss": 4.5502214431762695,
+ "step": 1357
+ },
+ {
+ "epoch": 18.86637554585153,
+ "grad_norm": 0.017908833920955658,
+ "learning_rate": 0.0006,
+ "loss": 4.550530433654785,
+ "step": 1358
+ },
+ {
+ "epoch": 18.880349344978168,
+ "grad_norm": 0.01796654611825943,
+ "learning_rate": 0.0006,
+ "loss": 4.531463146209717,
+ "step": 1359
+ },
+ {
+ "epoch": 18.894323144104803,
+ "grad_norm": 0.01646183803677559,
+ "learning_rate": 0.0006,
+ "loss": 4.498582363128662,
+ "step": 1360
+ },
+ {
+ "epoch": 18.90829694323144,
+ "grad_norm": 0.015430327504873276,
+ "learning_rate": 0.0006,
+ "loss": 4.52571439743042,
+ "step": 1361
+ },
+ {
+ "epoch": 18.92227074235808,
+ "grad_norm": 0.016527952626347542,
+ "learning_rate": 0.0006,
+ "loss": 4.547013759613037,
+ "step": 1362
+ },
+ {
+ "epoch": 18.936244541484715,
+ "grad_norm": 0.016263660043478012,
+ "learning_rate": 0.0006,
+ "loss": 4.468754291534424,
+ "step": 1363
+ },
+ {
+ "epoch": 18.950218340611354,
+ "grad_norm": 0.016347210854291916,
+ "learning_rate": 0.0006,
+ "loss": 4.450741767883301,
+ "step": 1364
+ },
+ {
+ "epoch": 18.964192139737992,
+ "grad_norm": 0.01701398566365242,
+ "learning_rate": 0.0006,
+ "loss": 4.534899711608887,
+ "step": 1365
+ },
+ {
+ "epoch": 18.978165938864628,
+ "grad_norm": 0.017693081870675087,
+ "learning_rate": 0.0006,
+ "loss": 4.460553169250488,
+ "step": 1366
+ },
+ {
+ "epoch": 18.992139737991266,
+ "grad_norm": 0.018746308982372284,
+ "learning_rate": 0.0006,
+ "loss": 4.538792610168457,
+ "step": 1367
+ },
+ {
+ "epoch": 19.0,
+ "grad_norm": 0.020099416375160217,
+ "learning_rate": 0.0006,
+ "loss": 4.449617385864258,
+ "step": 1368
+ },
+ {
+ "epoch": 19.0,
+ "eval_loss": 4.7756571769714355,
+ "eval_runtime": 57.1804,
+ "eval_samples_per_second": 42.707,
+ "eval_steps_per_second": 1.347,
+ "step": 1368
+ },
+ {
+ "epoch": 19.01397379912664,
+ "grad_norm": 0.018726933747529984,
+ "learning_rate": 0.0006,
+ "loss": 4.446059703826904,
+ "step": 1369
+ },
+ {
+ "epoch": 19.027947598253274,
+ "grad_norm": 0.01956087350845337,
+ "learning_rate": 0.0006,
+ "loss": 4.4520039558410645,
+ "step": 1370
+ },
+ {
+ "epoch": 19.041921397379912,
+ "grad_norm": 0.020092811435461044,
+ "learning_rate": 0.0006,
+ "loss": 4.543343544006348,
+ "step": 1371
+ },
+ {
+ "epoch": 19.05589519650655,
+ "grad_norm": 0.01843622885644436,
+ "learning_rate": 0.0006,
+ "loss": 4.535579204559326,
+ "step": 1372
+ },
+ {
+ "epoch": 19.069868995633186,
+ "grad_norm": 0.016989488154649734,
+ "learning_rate": 0.0006,
+ "loss": 4.52316951751709,
+ "step": 1373
+ },
+ {
+ "epoch": 19.083842794759825,
+ "grad_norm": 0.01805422455072403,
+ "learning_rate": 0.0006,
+ "loss": 4.5393266677856445,
+ "step": 1374
+ },
+ {
+ "epoch": 19.097816593886463,
+ "grad_norm": 0.01874629408121109,
+ "learning_rate": 0.0006,
+ "loss": 4.422059059143066,
+ "step": 1375
+ },
+ {
+ "epoch": 19.111790393013102,
+ "grad_norm": 0.01804221048951149,
+ "learning_rate": 0.0006,
+ "loss": 4.50508451461792,
+ "step": 1376
+ },
+ {
+ "epoch": 19.125764192139737,
+ "grad_norm": 0.01785474829375744,
+ "learning_rate": 0.0006,
+ "loss": 4.412599563598633,
+ "step": 1377
+ },
+ {
+ "epoch": 19.139737991266376,
+ "grad_norm": 0.01971791312098503,
+ "learning_rate": 0.0006,
+ "loss": 4.518318176269531,
+ "step": 1378
+ },
+ {
+ "epoch": 19.153711790393015,
+ "grad_norm": 0.02061633951961994,
+ "learning_rate": 0.0006,
+ "loss": 4.332393169403076,
+ "step": 1379
+ },
+ {
+ "epoch": 19.16768558951965,
+ "grad_norm": 0.020563429221510887,
+ "learning_rate": 0.0006,
+ "loss": 4.445863723754883,
+ "step": 1380
+ },
+ {
+ "epoch": 19.18165938864629,
+ "grad_norm": 0.019995713606476784,
+ "learning_rate": 0.0006,
+ "loss": 4.497312068939209,
+ "step": 1381
+ },
+ {
+ "epoch": 19.195633187772927,
+ "grad_norm": 0.019823070615530014,
+ "learning_rate": 0.0006,
+ "loss": 4.475085258483887,
+ "step": 1382
+ },
+ {
+ "epoch": 19.209606986899562,
+ "grad_norm": 0.019039355218410492,
+ "learning_rate": 0.0006,
+ "loss": 4.50432014465332,
+ "step": 1383
+ },
+ {
+ "epoch": 19.2235807860262,
+ "grad_norm": 0.017876973375678062,
+ "learning_rate": 0.0006,
+ "loss": 4.384788513183594,
+ "step": 1384
+ },
+ {
+ "epoch": 19.23755458515284,
+ "grad_norm": 0.018697699531912804,
+ "learning_rate": 0.0006,
+ "loss": 4.410778045654297,
+ "step": 1385
+ },
+ {
+ "epoch": 19.251528384279474,
+ "grad_norm": 0.018826806917786598,
+ "learning_rate": 0.0006,
+ "loss": 4.524034023284912,
+ "step": 1386
+ },
+ {
+ "epoch": 19.265502183406113,
+ "grad_norm": 0.01824226602911949,
+ "learning_rate": 0.0006,
+ "loss": 4.526393890380859,
+ "step": 1387
+ },
+ {
+ "epoch": 19.27947598253275,
+ "grad_norm": 0.01753208599984646,
+ "learning_rate": 0.0006,
+ "loss": 4.493703365325928,
+ "step": 1388
+ },
+ {
+ "epoch": 19.29344978165939,
+ "grad_norm": 0.01742432825267315,
+ "learning_rate": 0.0006,
+ "loss": 4.392831802368164,
+ "step": 1389
+ },
+ {
+ "epoch": 19.307423580786025,
+ "grad_norm": 0.017099356278777122,
+ "learning_rate": 0.0006,
+ "loss": 4.578668594360352,
+ "step": 1390
+ },
+ {
+ "epoch": 19.321397379912664,
+ "grad_norm": 0.017946023494005203,
+ "learning_rate": 0.0006,
+ "loss": 4.456718444824219,
+ "step": 1391
+ },
+ {
+ "epoch": 19.335371179039303,
+ "grad_norm": 0.017705943435430527,
+ "learning_rate": 0.0006,
+ "loss": 4.553682327270508,
+ "step": 1392
+ },
+ {
+ "epoch": 19.349344978165938,
+ "grad_norm": 0.016647523269057274,
+ "learning_rate": 0.0006,
+ "loss": 4.525016784667969,
+ "step": 1393
+ },
+ {
+ "epoch": 19.363318777292577,
+ "grad_norm": 0.015407416969537735,
+ "learning_rate": 0.0006,
+ "loss": 4.500231742858887,
+ "step": 1394
+ },
+ {
+ "epoch": 19.377292576419215,
+ "grad_norm": 0.017264019697904587,
+ "learning_rate": 0.0006,
+ "loss": 4.4615349769592285,
+ "step": 1395
+ },
+ {
+ "epoch": 19.39126637554585,
+ "grad_norm": 0.01871555484831333,
+ "learning_rate": 0.0006,
+ "loss": 4.281026840209961,
+ "step": 1396
+ },
+ {
+ "epoch": 19.40524017467249,
+ "grad_norm": 0.017361650243401527,
+ "learning_rate": 0.0006,
+ "loss": 4.479382514953613,
+ "step": 1397
+ },
+ {
+ "epoch": 19.419213973799128,
+ "grad_norm": 0.019781362265348434,
+ "learning_rate": 0.0006,
+ "loss": 4.402320861816406,
+ "step": 1398
+ },
+ {
+ "epoch": 19.433187772925763,
+ "grad_norm": 0.019043298438191414,
+ "learning_rate": 0.0006,
+ "loss": 4.476983070373535,
+ "step": 1399
+ },
+ {
+ "epoch": 19.4471615720524,
+ "grad_norm": 0.017137985676527023,
+ "learning_rate": 0.0006,
+ "loss": 4.418689250946045,
+ "step": 1400
+ },
+ {
+ "epoch": 19.46113537117904,
+ "grad_norm": 0.016481833532452583,
+ "learning_rate": 0.0006,
+ "loss": 4.579145431518555,
+ "step": 1401
+ },
+ {
+ "epoch": 19.475109170305675,
+ "grad_norm": 0.016507714986801147,
+ "learning_rate": 0.0006,
+ "loss": 4.519158363342285,
+ "step": 1402
+ },
+ {
+ "epoch": 19.489082969432314,
+ "grad_norm": 0.018092872574925423,
+ "learning_rate": 0.0006,
+ "loss": 4.513748645782471,
+ "step": 1403
+ },
+ {
+ "epoch": 19.503056768558952,
+ "grad_norm": 0.018153440207242966,
+ "learning_rate": 0.0006,
+ "loss": 4.645384311676025,
+ "step": 1404
+ },
+ {
+ "epoch": 19.51703056768559,
+ "grad_norm": 0.01657119207084179,
+ "learning_rate": 0.0006,
+ "loss": 4.399757385253906,
+ "step": 1405
+ },
+ {
+ "epoch": 19.531004366812226,
+ "grad_norm": 0.0156280305236578,
+ "learning_rate": 0.0006,
+ "loss": 4.467836380004883,
+ "step": 1406
+ },
+ {
+ "epoch": 19.544978165938865,
+ "grad_norm": 0.016838403418660164,
+ "learning_rate": 0.0006,
+ "loss": 4.529176235198975,
+ "step": 1407
+ },
+ {
+ "epoch": 19.558951965065503,
+ "grad_norm": 0.01743236929178238,
+ "learning_rate": 0.0006,
+ "loss": 4.440032482147217,
+ "step": 1408
+ },
+ {
+ "epoch": 19.57292576419214,
+ "grad_norm": 0.01553610060364008,
+ "learning_rate": 0.0006,
+ "loss": 4.425588607788086,
+ "step": 1409
+ },
+ {
+ "epoch": 19.586899563318777,
+ "grad_norm": 0.017259271815419197,
+ "learning_rate": 0.0006,
+ "loss": 4.405452728271484,
+ "step": 1410
+ },
+ {
+ "epoch": 19.600873362445416,
+ "grad_norm": 0.015568481758236885,
+ "learning_rate": 0.0006,
+ "loss": 4.444279193878174,
+ "step": 1411
+ },
+ {
+ "epoch": 19.61484716157205,
+ "grad_norm": 0.015331248752772808,
+ "learning_rate": 0.0006,
+ "loss": 4.451846122741699,
+ "step": 1412
+ },
+ {
+ "epoch": 19.62882096069869,
+ "grad_norm": 0.015667378902435303,
+ "learning_rate": 0.0006,
+ "loss": 4.533931255340576,
+ "step": 1413
+ },
+ {
+ "epoch": 19.64279475982533,
+ "grad_norm": 0.016177602112293243,
+ "learning_rate": 0.0006,
+ "loss": 4.467313289642334,
+ "step": 1414
+ },
+ {
+ "epoch": 19.656768558951963,
+ "grad_norm": 0.01733129471540451,
+ "learning_rate": 0.0006,
+ "loss": 4.496893882751465,
+ "step": 1415
+ },
+ {
+ "epoch": 19.670742358078602,
+ "grad_norm": 0.016539594158530235,
+ "learning_rate": 0.0006,
+ "loss": 4.401961326599121,
+ "step": 1416
+ },
+ {
+ "epoch": 19.68471615720524,
+ "grad_norm": 0.016058551147580147,
+ "learning_rate": 0.0006,
+ "loss": 4.503375053405762,
+ "step": 1417
+ },
+ {
+ "epoch": 19.69868995633188,
+ "grad_norm": 0.01715777814388275,
+ "learning_rate": 0.0006,
+ "loss": 4.3808369636535645,
+ "step": 1418
+ },
+ {
+ "epoch": 19.712663755458514,
+ "grad_norm": 0.01681119203567505,
+ "learning_rate": 0.0006,
+ "loss": 4.5469560623168945,
+ "step": 1419
+ },
+ {
+ "epoch": 19.726637554585153,
+ "grad_norm": 0.01634550653398037,
+ "learning_rate": 0.0006,
+ "loss": 4.328197956085205,
+ "step": 1420
+ },
+ {
+ "epoch": 19.74061135371179,
+ "grad_norm": 0.01549347210675478,
+ "learning_rate": 0.0006,
+ "loss": 4.534621238708496,
+ "step": 1421
+ },
+ {
+ "epoch": 19.754585152838427,
+ "grad_norm": 0.01654665172100067,
+ "learning_rate": 0.0006,
+ "loss": 4.4724225997924805,
+ "step": 1422
+ },
+ {
+ "epoch": 19.768558951965066,
+ "grad_norm": 0.017119025811553,
+ "learning_rate": 0.0006,
+ "loss": 4.3897247314453125,
+ "step": 1423
+ },
+ {
+ "epoch": 19.782532751091704,
+ "grad_norm": 0.01603223942220211,
+ "learning_rate": 0.0006,
+ "loss": 4.412731170654297,
+ "step": 1424
+ },
+ {
+ "epoch": 19.79650655021834,
+ "grad_norm": 0.016774067655205727,
+ "learning_rate": 0.0006,
+ "loss": 4.54707145690918,
+ "step": 1425
+ },
+ {
+ "epoch": 19.810480349344978,
+ "grad_norm": 0.016453025862574577,
+ "learning_rate": 0.0006,
+ "loss": 4.462501049041748,
+ "step": 1426
+ },
+ {
+ "epoch": 19.824454148471617,
+ "grad_norm": 0.01597629487514496,
+ "learning_rate": 0.0006,
+ "loss": 4.229758262634277,
+ "step": 1427
+ },
+ {
+ "epoch": 19.83842794759825,
+ "grad_norm": 0.017999034374952316,
+ "learning_rate": 0.0006,
+ "loss": 4.389263153076172,
+ "step": 1428
+ },
+ {
+ "epoch": 19.85240174672489,
+ "grad_norm": 0.02165898121893406,
+ "learning_rate": 0.0006,
+ "loss": 4.342130661010742,
+ "step": 1429
+ },
+ {
+ "epoch": 19.86637554585153,
+ "grad_norm": 0.02222822792828083,
+ "learning_rate": 0.0006,
+ "loss": 4.492661476135254,
+ "step": 1430
+ },
+ {
+ "epoch": 19.880349344978168,
+ "grad_norm": 0.019378066062927246,
+ "learning_rate": 0.0006,
+ "loss": 4.4614105224609375,
+ "step": 1431
+ },
+ {
+ "epoch": 19.894323144104803,
+ "grad_norm": 0.019908083602786064,
+ "learning_rate": 0.0006,
+ "loss": 4.370929718017578,
+ "step": 1432
+ },
+ {
+ "epoch": 19.90829694323144,
+ "grad_norm": 0.022296471521258354,
+ "learning_rate": 0.0006,
+ "loss": 4.438320159912109,
+ "step": 1433
+ },
+ {
+ "epoch": 19.92227074235808,
+ "grad_norm": 0.021810343489050865,
+ "learning_rate": 0.0006,
+ "loss": 4.462067127227783,
+ "step": 1434
+ },
+ {
+ "epoch": 19.936244541484715,
+ "grad_norm": 0.02129209227859974,
+ "learning_rate": 0.0006,
+ "loss": 4.541261196136475,
+ "step": 1435
+ },
+ {
+ "epoch": 19.950218340611354,
+ "grad_norm": 0.02079629711806774,
+ "learning_rate": 0.0006,
+ "loss": 4.36712646484375,
+ "step": 1436
+ },
+ {
+ "epoch": 19.964192139737992,
+ "grad_norm": 0.022501740604639053,
+ "learning_rate": 0.0006,
+ "loss": 4.5279083251953125,
+ "step": 1437
+ },
+ {
+ "epoch": 19.978165938864628,
+ "grad_norm": 0.019121676683425903,
+ "learning_rate": 0.0006,
+ "loss": 4.477789878845215,
+ "step": 1438
+ },
+ {
+ "epoch": 19.992139737991266,
+ "grad_norm": 0.017429588362574577,
+ "learning_rate": 0.0006,
+ "loss": 4.4023847579956055,
+ "step": 1439
+ },
+ {
+ "epoch": 20.0,
+ "grad_norm": 0.018857665359973907,
+ "learning_rate": 0.0006,
+ "loss": 4.494743347167969,
+ "step": 1440
+ },
+ {
+ "epoch": 20.0,
+ "eval_loss": 4.694555759429932,
+ "eval_runtime": 56.4146,
+ "eval_samples_per_second": 43.287,
+ "eval_steps_per_second": 1.365,
+ "step": 1440
+ },
+ {
+ "epoch": 20.01397379912664,
+ "grad_norm": 0.016505861654877663,
+ "learning_rate": 0.0006,
+ "loss": 4.506229877471924,
+ "step": 1441
+ },
+ {
+ "epoch": 20.027947598253274,
+ "grad_norm": 0.017571842297911644,
+ "learning_rate": 0.0006,
+ "loss": 4.384073257446289,
+ "step": 1442
+ },
+ {
+ "epoch": 20.041921397379912,
+ "grad_norm": 0.019398365169763565,
+ "learning_rate": 0.0006,
+ "loss": 4.482177734375,
+ "step": 1443
+ },
+ {
+ "epoch": 20.05589519650655,
+ "grad_norm": 0.02011062018573284,
+ "learning_rate": 0.0006,
+ "loss": 4.440773963928223,
+ "step": 1444
+ },
+ {
+ "epoch": 20.069868995633186,
+ "grad_norm": 0.019026953727006912,
+ "learning_rate": 0.0006,
+ "loss": 4.507168769836426,
+ "step": 1445
+ },
+ {
+ "epoch": 20.083842794759825,
+ "grad_norm": 0.020506031811237335,
+ "learning_rate": 0.0006,
+ "loss": 4.469079971313477,
+ "step": 1446
+ },
+ {
+ "epoch": 20.097816593886463,
+ "grad_norm": 0.019608162343502045,
+ "learning_rate": 0.0006,
+ "loss": 4.395927429199219,
+ "step": 1447
+ },
+ {
+ "epoch": 20.111790393013102,
+ "grad_norm": 0.019419124349951744,
+ "learning_rate": 0.0006,
+ "loss": 4.484235763549805,
+ "step": 1448
+ },
+ {
+ "epoch": 20.125764192139737,
+ "grad_norm": 0.021775128319859505,
+ "learning_rate": 0.0006,
+ "loss": 4.333173751831055,
+ "step": 1449
+ },
+ {
+ "epoch": 20.139737991266376,
+ "grad_norm": 0.025512272492051125,
+ "learning_rate": 0.0006,
+ "loss": 4.460362434387207,
+ "step": 1450
+ },
+ {
+ "epoch": 20.153711790393015,
+ "grad_norm": 0.025157401338219643,
+ "learning_rate": 0.0006,
+ "loss": 4.4361395835876465,
+ "step": 1451
+ },
+ {
+ "epoch": 20.16768558951965,
+ "grad_norm": 0.019862616434693336,
+ "learning_rate": 0.0006,
+ "loss": 4.426807403564453,
+ "step": 1452
+ },
+ {
+ "epoch": 20.18165938864629,
+ "grad_norm": 0.01932152360677719,
+ "learning_rate": 0.0006,
+ "loss": 4.375677108764648,
+ "step": 1453
+ },
+ {
+ "epoch": 20.195633187772927,
+ "grad_norm": 0.01760999858379364,
+ "learning_rate": 0.0006,
+ "loss": 4.42259407043457,
+ "step": 1454
+ },
+ {
+ "epoch": 20.209606986899562,
+ "grad_norm": 0.016998112201690674,
+ "learning_rate": 0.0006,
+ "loss": 4.481280326843262,
+ "step": 1455
+ },
+ {
+ "epoch": 20.2235807860262,
+ "grad_norm": 0.017199190333485603,
+ "learning_rate": 0.0006,
+ "loss": 4.496712684631348,
+ "step": 1456
+ },
+ {
+ "epoch": 20.23755458515284,
+ "grad_norm": 0.016833679750561714,
+ "learning_rate": 0.0006,
+ "loss": 4.40736722946167,
+ "step": 1457
+ },
+ {
+ "epoch": 20.251528384279474,
+ "grad_norm": 0.017546730116009712,
+ "learning_rate": 0.0006,
+ "loss": 4.4915266036987305,
+ "step": 1458
+ },
+ {
+ "epoch": 20.265502183406113,
+ "grad_norm": 0.016595078632235527,
+ "learning_rate": 0.0006,
+ "loss": 4.460815906524658,
+ "step": 1459
+ },
+ {
+ "epoch": 20.27947598253275,
+ "grad_norm": 0.017728568986058235,
+ "learning_rate": 0.0006,
+ "loss": 4.346048831939697,
+ "step": 1460
+ },
+ {
+ "epoch": 20.29344978165939,
+ "grad_norm": 0.01849197782576084,
+ "learning_rate": 0.0006,
+ "loss": 4.399681091308594,
+ "step": 1461
+ },
+ {
+ "epoch": 20.307423580786025,
+ "grad_norm": 0.018336588516831398,
+ "learning_rate": 0.0006,
+ "loss": 4.414183616638184,
+ "step": 1462
+ },
+ {
+ "epoch": 20.321397379912664,
+ "grad_norm": 0.018086519092321396,
+ "learning_rate": 0.0006,
+ "loss": 4.346586227416992,
+ "step": 1463
+ },
+ {
+ "epoch": 20.335371179039303,
+ "grad_norm": 0.019697299227118492,
+ "learning_rate": 0.0006,
+ "loss": 4.4420952796936035,
+ "step": 1464
+ },
+ {
+ "epoch": 20.349344978165938,
+ "grad_norm": 0.02198074571788311,
+ "learning_rate": 0.0006,
+ "loss": 4.588729381561279,
+ "step": 1465
+ },
+ {
+ "epoch": 20.363318777292577,
+ "grad_norm": 0.022813329473137856,
+ "learning_rate": 0.0006,
+ "loss": 4.49729585647583,
+ "step": 1466
+ },
+ {
+ "epoch": 20.377292576419215,
+ "grad_norm": 0.022006649523973465,
+ "learning_rate": 0.0006,
+ "loss": 4.475641250610352,
+ "step": 1467
+ },
+ {
+ "epoch": 20.39126637554585,
+ "grad_norm": 0.023188438266515732,
+ "learning_rate": 0.0006,
+ "loss": 4.404322624206543,
+ "step": 1468
+ },
+ {
+ "epoch": 20.40524017467249,
+ "grad_norm": 0.022226877510547638,
+ "learning_rate": 0.0006,
+ "loss": 4.50904655456543,
+ "step": 1469
+ },
+ {
+ "epoch": 20.419213973799128,
+ "grad_norm": 0.021082637831568718,
+ "learning_rate": 0.0006,
+ "loss": 4.339011192321777,
+ "step": 1470
+ },
+ {
+ "epoch": 20.433187772925763,
+ "grad_norm": 0.02013542130589485,
+ "learning_rate": 0.0006,
+ "loss": 4.516942024230957,
+ "step": 1471
+ },
+ {
+ "epoch": 20.4471615720524,
+ "grad_norm": 0.01898968778550625,
+ "learning_rate": 0.0006,
+ "loss": 4.315804481506348,
+ "step": 1472
+ },
+ {
+ "epoch": 20.46113537117904,
+ "grad_norm": 0.019450880587100983,
+ "learning_rate": 0.0006,
+ "loss": 4.4794921875,
+ "step": 1473
+ },
+ {
+ "epoch": 20.475109170305675,
+ "grad_norm": 0.018577802926301956,
+ "learning_rate": 0.0006,
+ "loss": 4.422323703765869,
+ "step": 1474
+ },
+ {
+ "epoch": 20.489082969432314,
+ "grad_norm": 0.01778412237763405,
+ "learning_rate": 0.0006,
+ "loss": 4.429789066314697,
+ "step": 1475
+ },
+ {
+ "epoch": 20.503056768558952,
+ "grad_norm": 0.017134476453065872,
+ "learning_rate": 0.0006,
+ "loss": 4.416459083557129,
+ "step": 1476
+ },
+ {
+ "epoch": 20.51703056768559,
+ "grad_norm": 0.017153557389974594,
+ "learning_rate": 0.0006,
+ "loss": 4.446170806884766,
+ "step": 1477
+ },
+ {
+ "epoch": 20.531004366812226,
+ "grad_norm": 0.014998532831668854,
+ "learning_rate": 0.0006,
+ "loss": 4.419447898864746,
+ "step": 1478
+ },
+ {
+ "epoch": 20.544978165938865,
+ "grad_norm": 0.015877850353717804,
+ "learning_rate": 0.0006,
+ "loss": 4.312338829040527,
+ "step": 1479
+ },
+ {
+ "epoch": 20.558951965065503,
+ "grad_norm": 0.01569489948451519,
+ "learning_rate": 0.0006,
+ "loss": 4.572905540466309,
+ "step": 1480
+ },
+ {
+ "epoch": 20.57292576419214,
+ "grad_norm": 0.01591474376618862,
+ "learning_rate": 0.0006,
+ "loss": 4.414907455444336,
+ "step": 1481
+ },
+ {
+ "epoch": 20.586899563318777,
+ "grad_norm": 0.015395903028547764,
+ "learning_rate": 0.0006,
+ "loss": 4.549131393432617,
+ "step": 1482
+ },
+ {
+ "epoch": 20.600873362445416,
+ "grad_norm": 0.015715358778834343,
+ "learning_rate": 0.0006,
+ "loss": 4.530194282531738,
+ "step": 1483
+ },
+ {
+ "epoch": 20.61484716157205,
+ "grad_norm": 0.016042975708842278,
+ "learning_rate": 0.0006,
+ "loss": 4.322149276733398,
+ "step": 1484
+ },
+ {
+ "epoch": 20.62882096069869,
+ "grad_norm": 0.01664186641573906,
+ "learning_rate": 0.0006,
+ "loss": 4.424866676330566,
+ "step": 1485
+ },
+ {
+ "epoch": 20.64279475982533,
+ "grad_norm": 0.01631534844636917,
+ "learning_rate": 0.0006,
+ "loss": 4.441752910614014,
+ "step": 1486
+ },
+ {
+ "epoch": 20.656768558951963,
+ "grad_norm": 0.014057058840990067,
+ "learning_rate": 0.0006,
+ "loss": 4.399982452392578,
+ "step": 1487
+ },
+ {
+ "epoch": 20.670742358078602,
+ "grad_norm": 0.015076273120939732,
+ "learning_rate": 0.0006,
+ "loss": 4.441324234008789,
+ "step": 1488
+ },
+ {
+ "epoch": 20.68471615720524,
+ "grad_norm": 0.014138277620077133,
+ "learning_rate": 0.0006,
+ "loss": 4.4312543869018555,
+ "step": 1489
+ },
+ {
+ "epoch": 20.69868995633188,
+ "grad_norm": 0.013762440532445908,
+ "learning_rate": 0.0006,
+ "loss": 4.387807846069336,
+ "step": 1490
+ },
+ {
+ "epoch": 20.712663755458514,
+ "grad_norm": 0.015058411285281181,
+ "learning_rate": 0.0006,
+ "loss": 4.594747543334961,
+ "step": 1491
+ },
+ {
+ "epoch": 20.726637554585153,
+ "grad_norm": 0.013827123679220676,
+ "learning_rate": 0.0006,
+ "loss": 4.434604167938232,
+ "step": 1492
+ },
+ {
+ "epoch": 20.74061135371179,
+ "grad_norm": 0.015315295197069645,
+ "learning_rate": 0.0006,
+ "loss": 4.588066101074219,
+ "step": 1493
+ },
+ {
+ "epoch": 20.754585152838427,
+ "grad_norm": 0.018678732216358185,
+ "learning_rate": 0.0006,
+ "loss": 4.3514204025268555,
+ "step": 1494
+ },
+ {
+ "epoch": 20.768558951965066,
+ "grad_norm": 0.019521350041031837,
+ "learning_rate": 0.0006,
+ "loss": 4.437658786773682,
+ "step": 1495
+ },
+ {
+ "epoch": 20.782532751091704,
+ "grad_norm": 0.018363403156399727,
+ "learning_rate": 0.0006,
+ "loss": 4.495002269744873,
+ "step": 1496
+ },
+ {
+ "epoch": 20.79650655021834,
+ "grad_norm": 0.016830725595355034,
+ "learning_rate": 0.0006,
+ "loss": 4.402471542358398,
+ "step": 1497
+ },
+ {
+ "epoch": 20.810480349344978,
+ "grad_norm": 0.01722072623670101,
+ "learning_rate": 0.0006,
+ "loss": 4.437653541564941,
+ "step": 1498
+ },
+ {
+ "epoch": 20.824454148471617,
+ "grad_norm": 0.017459599301218987,
+ "learning_rate": 0.0006,
+ "loss": 4.366603851318359,
+ "step": 1499
+ },
+ {
+ "epoch": 20.83842794759825,
+ "grad_norm": 0.018902219831943512,
+ "learning_rate": 0.0006,
+ "loss": 4.501561164855957,
+ "step": 1500
+ },
+ {
+ "epoch": 20.85240174672489,
+ "grad_norm": 0.020814096555113792,
+ "learning_rate": 0.0006,
+ "loss": 4.535295486450195,
+ "step": 1501
+ },
+ {
+ "epoch": 20.86637554585153,
+ "grad_norm": 0.018990976735949516,
+ "learning_rate": 0.0006,
+ "loss": 4.333580017089844,
+ "step": 1502
+ },
+ {
+ "epoch": 20.880349344978168,
+ "grad_norm": 0.01837880350649357,
+ "learning_rate": 0.0006,
+ "loss": 4.313388824462891,
+ "step": 1503
+ },
+ {
+ "epoch": 20.894323144104803,
+ "grad_norm": 0.019820360466837883,
+ "learning_rate": 0.0006,
+ "loss": 4.388157844543457,
+ "step": 1504
+ },
+ {
+ "epoch": 20.90829694323144,
+ "grad_norm": 0.020181143656373024,
+ "learning_rate": 0.0006,
+ "loss": 4.572224140167236,
+ "step": 1505
+ },
+ {
+ "epoch": 20.92227074235808,
+ "grad_norm": 0.021206077188253403,
+ "learning_rate": 0.0006,
+ "loss": 4.3127970695495605,
+ "step": 1506
+ },
+ {
+ "epoch": 20.936244541484715,
+ "grad_norm": 0.017962060868740082,
+ "learning_rate": 0.0006,
+ "loss": 4.384734630584717,
+ "step": 1507
+ },
+ {
+ "epoch": 20.950218340611354,
+ "grad_norm": 0.01737220585346222,
+ "learning_rate": 0.0006,
+ "loss": 4.4075751304626465,
+ "step": 1508
+ },
+ {
+ "epoch": 20.964192139737992,
+ "grad_norm": 0.01839268207550049,
+ "learning_rate": 0.0006,
+ "loss": 4.553779125213623,
+ "step": 1509
+ },
+ {
+ "epoch": 20.978165938864628,
+ "grad_norm": 0.020102640613913536,
+ "learning_rate": 0.0006,
+ "loss": 4.399398326873779,
+ "step": 1510
+ },
+ {
+ "epoch": 20.992139737991266,
+ "grad_norm": 0.018831463530659676,
+ "learning_rate": 0.0006,
+ "loss": 4.358499526977539,
+ "step": 1511
+ },
+ {
+ "epoch": 21.0,
+ "grad_norm": 0.018254734575748444,
+ "learning_rate": 0.0006,
+ "loss": 4.384119510650635,
+ "step": 1512
+ },
+ {
+ "epoch": 21.0,
+ "eval_loss": 4.7524237632751465,
+ "eval_runtime": 57.8812,
+ "eval_samples_per_second": 42.19,
+ "eval_steps_per_second": 1.33,
+ "step": 1512
+ },
+ {
+ "epoch": 21.01397379912664,
+ "grad_norm": 0.019728587940335274,
+ "learning_rate": 0.0006,
+ "loss": 4.32363224029541,
+ "step": 1513
+ },
+ {
+ "epoch": 21.027947598253274,
+ "grad_norm": 0.02216625027358532,
+ "learning_rate": 0.0006,
+ "loss": 4.436988353729248,
+ "step": 1514
+ },
+ {
+ "epoch": 21.041921397379912,
+ "grad_norm": 0.02227606810629368,
+ "learning_rate": 0.0006,
+ "loss": 4.406063556671143,
+ "step": 1515
+ },
+ {
+ "epoch": 21.05589519650655,
+ "grad_norm": 0.020187009125947952,
+ "learning_rate": 0.0006,
+ "loss": 4.324402332305908,
+ "step": 1516
+ },
+ {
+ "epoch": 21.069868995633186,
+ "grad_norm": 0.02095157466828823,
+ "learning_rate": 0.0006,
+ "loss": 4.456544876098633,
+ "step": 1517
+ },
+ {
+ "epoch": 21.083842794759825,
+ "grad_norm": 0.02146013453602791,
+ "learning_rate": 0.0006,
+ "loss": 4.368387222290039,
+ "step": 1518
+ },
+ {
+ "epoch": 21.097816593886463,
+ "grad_norm": 0.02162035182118416,
+ "learning_rate": 0.0006,
+ "loss": 4.48838996887207,
+ "step": 1519
+ },
+ {
+ "epoch": 21.111790393013102,
+ "grad_norm": 0.021745676174759865,
+ "learning_rate": 0.0006,
+ "loss": 4.30778694152832,
+ "step": 1520
+ },
+ {
+ "epoch": 21.125764192139737,
+ "grad_norm": 0.02173588052392006,
+ "learning_rate": 0.0006,
+ "loss": 4.433149814605713,
+ "step": 1521
+ },
+ {
+ "epoch": 21.139737991266376,
+ "grad_norm": 0.020950648933649063,
+ "learning_rate": 0.0006,
+ "loss": 4.520089149475098,
+ "step": 1522
+ },
+ {
+ "epoch": 21.153711790393015,
+ "grad_norm": 0.020684752613306046,
+ "learning_rate": 0.0006,
+ "loss": 4.372437477111816,
+ "step": 1523
+ },
+ {
+ "epoch": 21.16768558951965,
+ "grad_norm": 0.019472869113087654,
+ "learning_rate": 0.0006,
+ "loss": 4.437015533447266,
+ "step": 1524
+ },
+ {
+ "epoch": 21.18165938864629,
+ "grad_norm": 0.01846962980926037,
+ "learning_rate": 0.0006,
+ "loss": 4.449531555175781,
+ "step": 1525
+ },
+ {
+ "epoch": 21.195633187772927,
+ "grad_norm": 0.017377803102135658,
+ "learning_rate": 0.0006,
+ "loss": 4.458279132843018,
+ "step": 1526
+ },
+ {
+ "epoch": 21.209606986899562,
+ "grad_norm": 0.016002200543880463,
+ "learning_rate": 0.0006,
+ "loss": 4.266297817230225,
+ "step": 1527
+ },
+ {
+ "epoch": 21.2235807860262,
+ "grad_norm": 0.015575903467833996,
+ "learning_rate": 0.0006,
+ "loss": 4.342632293701172,
+ "step": 1528
+ },
+ {
+ "epoch": 21.23755458515284,
+ "grad_norm": 0.015634698793292046,
+ "learning_rate": 0.0006,
+ "loss": 4.338623523712158,
+ "step": 1529
+ },
+ {
+ "epoch": 21.251528384279474,
+ "grad_norm": 0.01449581328779459,
+ "learning_rate": 0.0006,
+ "loss": 4.354330062866211,
+ "step": 1530
+ },
+ {
+ "epoch": 21.265502183406113,
+ "grad_norm": 0.01487616915255785,
+ "learning_rate": 0.0006,
+ "loss": 4.486835479736328,
+ "step": 1531
+ },
+ {
+ "epoch": 21.27947598253275,
+ "grad_norm": 0.016346555203199387,
+ "learning_rate": 0.0006,
+ "loss": 4.407098770141602,
+ "step": 1532
+ },
+ {
+ "epoch": 21.29344978165939,
+ "grad_norm": 0.01682182401418686,
+ "learning_rate": 0.0006,
+ "loss": 4.436809539794922,
+ "step": 1533
+ },
+ {
+ "epoch": 21.307423580786025,
+ "grad_norm": 0.017704004421830177,
+ "learning_rate": 0.0006,
+ "loss": 4.3854780197143555,
+ "step": 1534
+ },
+ {
+ "epoch": 21.321397379912664,
+ "grad_norm": 0.018437420949339867,
+ "learning_rate": 0.0006,
+ "loss": 4.413283824920654,
+ "step": 1535
+ },
+ {
+ "epoch": 21.335371179039303,
+ "grad_norm": 0.016911081969738007,
+ "learning_rate": 0.0006,
+ "loss": 4.357306003570557,
+ "step": 1536
+ },
+ {
+ "epoch": 21.349344978165938,
+ "grad_norm": 0.01642223633825779,
+ "learning_rate": 0.0006,
+ "loss": 4.422322750091553,
+ "step": 1537
+ },
+ {
+ "epoch": 21.363318777292577,
+ "grad_norm": 0.01780877821147442,
+ "learning_rate": 0.0006,
+ "loss": 4.387620449066162,
+ "step": 1538
+ },
+ {
+ "epoch": 21.377292576419215,
+ "grad_norm": 0.01653975248336792,
+ "learning_rate": 0.0006,
+ "loss": 4.327267646789551,
+ "step": 1539
+ },
+ {
+ "epoch": 21.39126637554585,
+ "grad_norm": 0.01572156324982643,
+ "learning_rate": 0.0006,
+ "loss": 4.475979804992676,
+ "step": 1540
+ },
+ {
+ "epoch": 21.40524017467249,
+ "grad_norm": 0.017156066372990608,
+ "learning_rate": 0.0006,
+ "loss": 4.4011125564575195,
+ "step": 1541
+ },
+ {
+ "epoch": 21.419213973799128,
+ "grad_norm": 0.01788829080760479,
+ "learning_rate": 0.0006,
+ "loss": 4.4174089431762695,
+ "step": 1542
+ },
+ {
+ "epoch": 21.433187772925763,
+ "grad_norm": 0.01728985086083412,
+ "learning_rate": 0.0006,
+ "loss": 4.429244041442871,
+ "step": 1543
+ },
+ {
+ "epoch": 21.4471615720524,
+ "grad_norm": 0.016534466296434402,
+ "learning_rate": 0.0006,
+ "loss": 4.326068878173828,
+ "step": 1544
+ },
+ {
+ "epoch": 21.46113537117904,
+ "grad_norm": 0.016582584008574486,
+ "learning_rate": 0.0006,
+ "loss": 4.2975969314575195,
+ "step": 1545
+ },
+ {
+ "epoch": 21.475109170305675,
+ "grad_norm": 0.018034178763628006,
+ "learning_rate": 0.0006,
+ "loss": 4.501319408416748,
+ "step": 1546
+ },
+ {
+ "epoch": 21.489082969432314,
+ "grad_norm": 0.017827702686190605,
+ "learning_rate": 0.0006,
+ "loss": 4.502284526824951,
+ "step": 1547
+ },
+ {
+ "epoch": 21.503056768558952,
+ "grad_norm": 0.017182299867272377,
+ "learning_rate": 0.0006,
+ "loss": 4.380828857421875,
+ "step": 1548
+ },
+ {
+ "epoch": 21.51703056768559,
+ "grad_norm": 0.019101588055491447,
+ "learning_rate": 0.0006,
+ "loss": 4.31564998626709,
+ "step": 1549
+ },
+ {
+ "epoch": 21.531004366812226,
+ "grad_norm": 0.0204677302390337,
+ "learning_rate": 0.0006,
+ "loss": 4.3840131759643555,
+ "step": 1550
+ },
+ {
+ "epoch": 21.544978165938865,
+ "grad_norm": 0.0188993439078331,
+ "learning_rate": 0.0006,
+ "loss": 4.4257731437683105,
+ "step": 1551
+ },
+ {
+ "epoch": 21.558951965065503,
+ "grad_norm": 0.01730230636894703,
+ "learning_rate": 0.0006,
+ "loss": 4.445132255554199,
+ "step": 1552
+ },
+ {
+ "epoch": 21.57292576419214,
+ "grad_norm": 0.01801574043929577,
+ "learning_rate": 0.0006,
+ "loss": 4.427064895629883,
+ "step": 1553
+ },
+ {
+ "epoch": 21.586899563318777,
+ "grad_norm": 0.018417565152049065,
+ "learning_rate": 0.0006,
+ "loss": 4.319595813751221,
+ "step": 1554
+ },
+ {
+ "epoch": 21.600873362445416,
+ "grad_norm": 0.01683160290122032,
+ "learning_rate": 0.0006,
+ "loss": 4.349218368530273,
+ "step": 1555
+ },
+ {
+ "epoch": 21.61484716157205,
+ "grad_norm": 0.016973106190562248,
+ "learning_rate": 0.0006,
+ "loss": 4.375990390777588,
+ "step": 1556
+ },
+ {
+ "epoch": 21.62882096069869,
+ "grad_norm": 0.015688583254814148,
+ "learning_rate": 0.0006,
+ "loss": 4.408170223236084,
+ "step": 1557
+ },
+ {
+ "epoch": 21.64279475982533,
+ "grad_norm": 0.01573043316602707,
+ "learning_rate": 0.0006,
+ "loss": 4.416255950927734,
+ "step": 1558
+ },
+ {
+ "epoch": 21.656768558951963,
+ "grad_norm": 0.01744293048977852,
+ "learning_rate": 0.0006,
+ "loss": 4.477141380310059,
+ "step": 1559
+ },
+ {
+ "epoch": 21.670742358078602,
+ "grad_norm": 0.017503513023257256,
+ "learning_rate": 0.0006,
+ "loss": 4.3367486000061035,
+ "step": 1560
+ },
+ {
+ "epoch": 21.68471615720524,
+ "grad_norm": 0.016592150554060936,
+ "learning_rate": 0.0006,
+ "loss": 4.314785480499268,
+ "step": 1561
+ },
+ {
+ "epoch": 21.69868995633188,
+ "grad_norm": 0.015774084255099297,
+ "learning_rate": 0.0006,
+ "loss": 4.448209285736084,
+ "step": 1562
+ },
+ {
+ "epoch": 21.712663755458514,
+ "grad_norm": 0.015779150649905205,
+ "learning_rate": 0.0006,
+ "loss": 4.392690658569336,
+ "step": 1563
+ },
+ {
+ "epoch": 21.726637554585153,
+ "grad_norm": 0.015392648056149483,
+ "learning_rate": 0.0006,
+ "loss": 4.437741756439209,
+ "step": 1564
+ },
+ {
+ "epoch": 21.74061135371179,
+ "grad_norm": 0.01530836895108223,
+ "learning_rate": 0.0006,
+ "loss": 4.424595832824707,
+ "step": 1565
+ },
+ {
+ "epoch": 21.754585152838427,
+ "grad_norm": 0.015600779093801975,
+ "learning_rate": 0.0006,
+ "loss": 4.335417747497559,
+ "step": 1566
+ },
+ {
+ "epoch": 21.768558951965066,
+ "grad_norm": 0.014646518044173717,
+ "learning_rate": 0.0006,
+ "loss": 4.279081344604492,
+ "step": 1567
+ },
+ {
+ "epoch": 21.782532751091704,
+ "grad_norm": 0.014132864773273468,
+ "learning_rate": 0.0006,
+ "loss": 4.345950126647949,
+ "step": 1568
+ },
+ {
+ "epoch": 21.79650655021834,
+ "grad_norm": 0.01425724383443594,
+ "learning_rate": 0.0006,
+ "loss": 4.277588844299316,
+ "step": 1569
+ },
+ {
+ "epoch": 21.810480349344978,
+ "grad_norm": 0.015396572649478912,
+ "learning_rate": 0.0006,
+ "loss": 4.326196670532227,
+ "step": 1570
+ },
+ {
+ "epoch": 21.824454148471617,
+ "grad_norm": 0.01683669537305832,
+ "learning_rate": 0.0006,
+ "loss": 4.3294172286987305,
+ "step": 1571
+ },
+ {
+ "epoch": 21.83842794759825,
+ "grad_norm": 0.016365647315979004,
+ "learning_rate": 0.0006,
+ "loss": 4.523321628570557,
+ "step": 1572
+ },
+ {
+ "epoch": 21.85240174672489,
+ "grad_norm": 0.017142774537205696,
+ "learning_rate": 0.0006,
+ "loss": 4.475842475891113,
+ "step": 1573
+ },
+ {
+ "epoch": 21.86637554585153,
+ "grad_norm": 0.015950839966535568,
+ "learning_rate": 0.0006,
+ "loss": 4.397187232971191,
+ "step": 1574
+ },
+ {
+ "epoch": 21.880349344978168,
+ "grad_norm": 0.016180362552404404,
+ "learning_rate": 0.0006,
+ "loss": 4.414778709411621,
+ "step": 1575
+ },
+ {
+ "epoch": 21.894323144104803,
+ "grad_norm": 0.017615556716918945,
+ "learning_rate": 0.0006,
+ "loss": 4.348567962646484,
+ "step": 1576
+ },
+ {
+ "epoch": 21.90829694323144,
+ "grad_norm": 0.01896277628839016,
+ "learning_rate": 0.0006,
+ "loss": 4.431762218475342,
+ "step": 1577
+ },
+ {
+ "epoch": 21.92227074235808,
+ "grad_norm": 0.01990971714258194,
+ "learning_rate": 0.0006,
+ "loss": 4.431285858154297,
+ "step": 1578
+ },
+ {
+ "epoch": 21.936244541484715,
+ "grad_norm": 0.0201258547604084,
+ "learning_rate": 0.0006,
+ "loss": 4.309872627258301,
+ "step": 1579
+ },
+ {
+ "epoch": 21.950218340611354,
+ "grad_norm": 0.019620198756456375,
+ "learning_rate": 0.0006,
+ "loss": 4.484044551849365,
+ "step": 1580
+ },
+ {
+ "epoch": 21.964192139737992,
+ "grad_norm": 0.01771123707294464,
+ "learning_rate": 0.0006,
+ "loss": 4.368093013763428,
+ "step": 1581
+ },
+ {
+ "epoch": 21.978165938864628,
+ "grad_norm": 0.017501654103398323,
+ "learning_rate": 0.0006,
+ "loss": 4.372651100158691,
+ "step": 1582
+ },
+ {
+ "epoch": 21.992139737991266,
+ "grad_norm": 0.01693608984351158,
+ "learning_rate": 0.0006,
+ "loss": 4.379167556762695,
+ "step": 1583
+ },
+ {
+ "epoch": 22.0,
+ "grad_norm": 0.02002849616110325,
+ "learning_rate": 0.0006,
+ "loss": 4.329627990722656,
+ "step": 1584
+ },
+ {
+ "epoch": 22.0,
+ "eval_loss": 4.723696231842041,
+ "eval_runtime": 56.1758,
+ "eval_samples_per_second": 43.471,
+ "eval_steps_per_second": 1.371,
+ "step": 1584
+ },
+ {
+ "epoch": 22.01397379912664,
+ "grad_norm": 0.02030305378139019,
+ "learning_rate": 0.0006,
+ "loss": 4.355566024780273,
+ "step": 1585
+ },
+ {
+ "epoch": 22.027947598253274,
+ "grad_norm": 0.018638934940099716,
+ "learning_rate": 0.0006,
+ "loss": 4.314009189605713,
+ "step": 1586
+ },
+ {
+ "epoch": 22.041921397379912,
+ "grad_norm": 0.019917072728276253,
+ "learning_rate": 0.0006,
+ "loss": 4.367525577545166,
+ "step": 1587
+ },
+ {
+ "epoch": 22.05589519650655,
+ "grad_norm": 0.020785167813301086,
+ "learning_rate": 0.0006,
+ "loss": 4.375561714172363,
+ "step": 1588
+ },
+ {
+ "epoch": 22.069868995633186,
+ "grad_norm": 0.021214453503489494,
+ "learning_rate": 0.0006,
+ "loss": 4.418305397033691,
+ "step": 1589
+ },
+ {
+ "epoch": 22.083842794759825,
+ "grad_norm": 0.019051678478717804,
+ "learning_rate": 0.0006,
+ "loss": 4.215126991271973,
+ "step": 1590
+ },
+ {
+ "epoch": 22.097816593886463,
+ "grad_norm": 0.02323366142809391,
+ "learning_rate": 0.0006,
+ "loss": 4.306807994842529,
+ "step": 1591
+ },
+ {
+ "epoch": 22.111790393013102,
+ "grad_norm": 0.026186738163232803,
+ "learning_rate": 0.0006,
+ "loss": 4.408508777618408,
+ "step": 1592
+ },
+ {
+ "epoch": 22.125764192139737,
+ "grad_norm": 0.027361227199435234,
+ "learning_rate": 0.0006,
+ "loss": 4.329322814941406,
+ "step": 1593
+ },
+ {
+ "epoch": 22.139737991266376,
+ "grad_norm": 0.026564646512269974,
+ "learning_rate": 0.0006,
+ "loss": 4.346642017364502,
+ "step": 1594
+ },
+ {
+ "epoch": 22.153711790393015,
+ "grad_norm": 0.025073667988181114,
+ "learning_rate": 0.0006,
+ "loss": 4.3984198570251465,
+ "step": 1595
+ },
+ {
+ "epoch": 22.16768558951965,
+ "grad_norm": 0.02347794733941555,
+ "learning_rate": 0.0006,
+ "loss": 4.352538108825684,
+ "step": 1596
+ },
+ {
+ "epoch": 22.18165938864629,
+ "grad_norm": 0.02203468047082424,
+ "learning_rate": 0.0006,
+ "loss": 4.298148155212402,
+ "step": 1597
+ },
+ {
+ "epoch": 22.195633187772927,
+ "grad_norm": 0.020621957257390022,
+ "learning_rate": 0.0006,
+ "loss": 4.480309963226318,
+ "step": 1598
+ },
+ {
+ "epoch": 22.209606986899562,
+ "grad_norm": 0.023710055276751518,
+ "learning_rate": 0.0006,
+ "loss": 4.339317321777344,
+ "step": 1599
+ },
+ {
+ "epoch": 22.2235807860262,
+ "grad_norm": 0.0233482476323843,
+ "learning_rate": 0.0006,
+ "loss": 4.436896324157715,
+ "step": 1600
+ },
+ {
+ "epoch": 22.23755458515284,
+ "grad_norm": 0.020122403278946877,
+ "learning_rate": 0.0006,
+ "loss": 4.236002445220947,
+ "step": 1601
+ },
+ {
+ "epoch": 22.251528384279474,
+ "grad_norm": 0.01957266591489315,
+ "learning_rate": 0.0006,
+ "loss": 4.497328281402588,
+ "step": 1602
+ },
+ {
+ "epoch": 22.265502183406113,
+ "grad_norm": 0.018964895978569984,
+ "learning_rate": 0.0006,
+ "loss": 4.36360502243042,
+ "step": 1603
+ },
+ {
+ "epoch": 22.27947598253275,
+ "grad_norm": 0.018902769312262535,
+ "learning_rate": 0.0006,
+ "loss": 4.427093505859375,
+ "step": 1604
+ },
+ {
+ "epoch": 22.29344978165939,
+ "grad_norm": 0.018918119370937347,
+ "learning_rate": 0.0006,
+ "loss": 4.294950485229492,
+ "step": 1605
+ },
+ {
+ "epoch": 22.307423580786025,
+ "grad_norm": 0.019315775483846664,
+ "learning_rate": 0.0006,
+ "loss": 4.427952289581299,
+ "step": 1606
+ },
+ {
+ "epoch": 22.321397379912664,
+ "grad_norm": 0.018350522965192795,
+ "learning_rate": 0.0006,
+ "loss": 4.272945404052734,
+ "step": 1607
+ },
+ {
+ "epoch": 22.335371179039303,
+ "grad_norm": 0.01759534701704979,
+ "learning_rate": 0.0006,
+ "loss": 4.348023891448975,
+ "step": 1608
+ },
+ {
+ "epoch": 22.349344978165938,
+ "grad_norm": 0.01646057888865471,
+ "learning_rate": 0.0006,
+ "loss": 4.278518199920654,
+ "step": 1609
+ },
+ {
+ "epoch": 22.363318777292577,
+ "grad_norm": 0.015875060111284256,
+ "learning_rate": 0.0006,
+ "loss": 4.423527240753174,
+ "step": 1610
+ },
+ {
+ "epoch": 22.377292576419215,
+ "grad_norm": 0.01638529635965824,
+ "learning_rate": 0.0006,
+ "loss": 4.321723937988281,
+ "step": 1611
+ },
+ {
+ "epoch": 22.39126637554585,
+ "grad_norm": 0.016942061483860016,
+ "learning_rate": 0.0006,
+ "loss": 4.346164703369141,
+ "step": 1612
+ },
+ {
+ "epoch": 22.40524017467249,
+ "grad_norm": 0.015209621749818325,
+ "learning_rate": 0.0006,
+ "loss": 4.434340000152588,
+ "step": 1613
+ },
+ {
+ "epoch": 22.419213973799128,
+ "grad_norm": 0.014423009008169174,
+ "learning_rate": 0.0006,
+ "loss": 4.342560768127441,
+ "step": 1614
+ },
+ {
+ "epoch": 22.433187772925763,
+ "grad_norm": 0.014369525946676731,
+ "learning_rate": 0.0006,
+ "loss": 4.296586990356445,
+ "step": 1615
+ },
+ {
+ "epoch": 22.4471615720524,
+ "grad_norm": 0.01392540242522955,
+ "learning_rate": 0.0006,
+ "loss": 4.433176040649414,
+ "step": 1616
+ },
+ {
+ "epoch": 22.46113537117904,
+ "grad_norm": 0.013206899166107178,
+ "learning_rate": 0.0006,
+ "loss": 4.32962703704834,
+ "step": 1617
+ },
+ {
+ "epoch": 22.475109170305675,
+ "grad_norm": 0.013427077792584896,
+ "learning_rate": 0.0006,
+ "loss": 4.420437335968018,
+ "step": 1618
+ },
+ {
+ "epoch": 22.489082969432314,
+ "grad_norm": 0.014194196090102196,
+ "learning_rate": 0.0006,
+ "loss": 4.410478591918945,
+ "step": 1619
+ },
+ {
+ "epoch": 22.503056768558952,
+ "grad_norm": 0.01362364087253809,
+ "learning_rate": 0.0006,
+ "loss": 4.338932037353516,
+ "step": 1620
+ },
+ {
+ "epoch": 22.51703056768559,
+ "grad_norm": 0.014299589209258556,
+ "learning_rate": 0.0006,
+ "loss": 4.34089994430542,
+ "step": 1621
+ },
+ {
+ "epoch": 22.531004366812226,
+ "grad_norm": 0.014061295427381992,
+ "learning_rate": 0.0006,
+ "loss": 4.29201602935791,
+ "step": 1622
+ },
+ {
+ "epoch": 22.544978165938865,
+ "grad_norm": 0.013945229351520538,
+ "learning_rate": 0.0006,
+ "loss": 4.326052188873291,
+ "step": 1623
+ },
+ {
+ "epoch": 22.558951965065503,
+ "grad_norm": 0.01387068536132574,
+ "learning_rate": 0.0006,
+ "loss": 4.393521308898926,
+ "step": 1624
+ },
+ {
+ "epoch": 22.57292576419214,
+ "grad_norm": 0.013931898400187492,
+ "learning_rate": 0.0006,
+ "loss": 4.3739752769470215,
+ "step": 1625
+ },
+ {
+ "epoch": 22.586899563318777,
+ "grad_norm": 0.01497740764170885,
+ "learning_rate": 0.0006,
+ "loss": 4.3991618156433105,
+ "step": 1626
+ },
+ {
+ "epoch": 22.600873362445416,
+ "grad_norm": 0.015842905268073082,
+ "learning_rate": 0.0006,
+ "loss": 4.405340671539307,
+ "step": 1627
+ },
+ {
+ "epoch": 22.61484716157205,
+ "grad_norm": 0.01580633595585823,
+ "learning_rate": 0.0006,
+ "loss": 4.280102729797363,
+ "step": 1628
+ },
+ {
+ "epoch": 22.62882096069869,
+ "grad_norm": 0.016567382961511612,
+ "learning_rate": 0.0006,
+ "loss": 4.407657146453857,
+ "step": 1629
+ },
+ {
+ "epoch": 22.64279475982533,
+ "grad_norm": 0.01648622378706932,
+ "learning_rate": 0.0006,
+ "loss": 4.392590522766113,
+ "step": 1630
+ },
+ {
+ "epoch": 22.656768558951963,
+ "grad_norm": 0.017066307365894318,
+ "learning_rate": 0.0006,
+ "loss": 4.324862003326416,
+ "step": 1631
+ },
+ {
+ "epoch": 22.670742358078602,
+ "grad_norm": 0.01637062057852745,
+ "learning_rate": 0.0006,
+ "loss": 4.34598445892334,
+ "step": 1632
+ },
+ {
+ "epoch": 22.68471615720524,
+ "grad_norm": 0.018038960173726082,
+ "learning_rate": 0.0006,
+ "loss": 4.320711612701416,
+ "step": 1633
+ },
+ {
+ "epoch": 22.69868995633188,
+ "grad_norm": 0.018096428364515305,
+ "learning_rate": 0.0006,
+ "loss": 4.536527156829834,
+ "step": 1634
+ },
+ {
+ "epoch": 22.712663755458514,
+ "grad_norm": 0.017889291048049927,
+ "learning_rate": 0.0006,
+ "loss": 4.388131141662598,
+ "step": 1635
+ },
+ {
+ "epoch": 22.726637554585153,
+ "grad_norm": 0.018771521747112274,
+ "learning_rate": 0.0006,
+ "loss": 4.357668876647949,
+ "step": 1636
+ },
+ {
+ "epoch": 22.74061135371179,
+ "grad_norm": 0.02100873365998268,
+ "learning_rate": 0.0006,
+ "loss": 4.293972015380859,
+ "step": 1637
+ },
+ {
+ "epoch": 22.754585152838427,
+ "grad_norm": 0.017715785652399063,
+ "learning_rate": 0.0006,
+ "loss": 4.349849224090576,
+ "step": 1638
+ },
+ {
+ "epoch": 22.768558951965066,
+ "grad_norm": 0.015546333976089954,
+ "learning_rate": 0.0006,
+ "loss": 4.311728000640869,
+ "step": 1639
+ },
+ {
+ "epoch": 22.782532751091704,
+ "grad_norm": 0.01605990156531334,
+ "learning_rate": 0.0006,
+ "loss": 4.499123573303223,
+ "step": 1640
+ },
+ {
+ "epoch": 22.79650655021834,
+ "grad_norm": 0.016642142087221146,
+ "learning_rate": 0.0006,
+ "loss": 4.4319257736206055,
+ "step": 1641
+ },
+ {
+ "epoch": 22.810480349344978,
+ "grad_norm": 0.016120387241244316,
+ "learning_rate": 0.0006,
+ "loss": 4.4373650550842285,
+ "step": 1642
+ },
+ {
+ "epoch": 22.824454148471617,
+ "grad_norm": 0.015174890868365765,
+ "learning_rate": 0.0006,
+ "loss": 4.314082145690918,
+ "step": 1643
+ },
+ {
+ "epoch": 22.83842794759825,
+ "grad_norm": 0.01551714539527893,
+ "learning_rate": 0.0006,
+ "loss": 4.360363960266113,
+ "step": 1644
+ },
+ {
+ "epoch": 22.85240174672489,
+ "grad_norm": 0.014966742135584354,
+ "learning_rate": 0.0006,
+ "loss": 4.346253395080566,
+ "step": 1645
+ },
+ {
+ "epoch": 22.86637554585153,
+ "grad_norm": 0.014288028702139854,
+ "learning_rate": 0.0006,
+ "loss": 4.464366436004639,
+ "step": 1646
+ },
+ {
+ "epoch": 22.880349344978168,
+ "grad_norm": 0.015291682444512844,
+ "learning_rate": 0.0006,
+ "loss": 4.504101753234863,
+ "step": 1647
+ },
+ {
+ "epoch": 22.894323144104803,
+ "grad_norm": 0.014534426853060722,
+ "learning_rate": 0.0006,
+ "loss": 4.462172031402588,
+ "step": 1648
+ },
+ {
+ "epoch": 22.90829694323144,
+ "grad_norm": 0.014129819348454475,
+ "learning_rate": 0.0006,
+ "loss": 4.350126266479492,
+ "step": 1649
+ },
+ {
+ "epoch": 22.92227074235808,
+ "grad_norm": 0.014932668767869473,
+ "learning_rate": 0.0006,
+ "loss": 4.374392032623291,
+ "step": 1650
+ },
+ {
+ "epoch": 22.936244541484715,
+ "grad_norm": 0.014647853560745716,
+ "learning_rate": 0.0006,
+ "loss": 4.252264499664307,
+ "step": 1651
+ },
+ {
+ "epoch": 22.950218340611354,
+ "grad_norm": 0.013553360477089882,
+ "learning_rate": 0.0006,
+ "loss": 4.379729270935059,
+ "step": 1652
+ },
+ {
+ "epoch": 22.964192139737992,
+ "grad_norm": 0.014606194570660591,
+ "learning_rate": 0.0006,
+ "loss": 4.363372802734375,
+ "step": 1653
+ },
+ {
+ "epoch": 22.978165938864628,
+ "grad_norm": 0.015000631101429462,
+ "learning_rate": 0.0006,
+ "loss": 4.408048629760742,
+ "step": 1654
+ },
+ {
+ "epoch": 22.992139737991266,
+ "grad_norm": 0.015370816923677921,
+ "learning_rate": 0.0006,
+ "loss": 4.297451019287109,
+ "step": 1655
+ },
+ {
+ "epoch": 23.0,
+ "grad_norm": 0.01709144562482834,
+ "learning_rate": 0.0006,
+ "loss": 4.294776916503906,
+ "step": 1656
+ },
+ {
+ "epoch": 23.0,
+ "eval_loss": 4.678618907928467,
+ "eval_runtime": 56.1668,
+ "eval_samples_per_second": 43.478,
+ "eval_steps_per_second": 1.371,
+ "step": 1656
+ },
+ {
+ "epoch": 23.01397379912664,
+ "grad_norm": 0.01727290451526642,
+ "learning_rate": 0.0006,
+ "loss": 4.311701774597168,
+ "step": 1657
+ },
+ {
+ "epoch": 23.027947598253274,
+ "grad_norm": 0.020645759999752045,
+ "learning_rate": 0.0006,
+ "loss": 4.429897308349609,
+ "step": 1658
+ },
+ {
+ "epoch": 23.041921397379912,
+ "grad_norm": 0.02204226516187191,
+ "learning_rate": 0.0006,
+ "loss": 4.365151882171631,
+ "step": 1659
+ },
+ {
+ "epoch": 23.05589519650655,
+ "grad_norm": 0.023069269955158234,
+ "learning_rate": 0.0006,
+ "loss": 4.3710832595825195,
+ "step": 1660
+ },
+ {
+ "epoch": 23.069868995633186,
+ "grad_norm": 0.024151179939508438,
+ "learning_rate": 0.0006,
+ "loss": 4.403108596801758,
+ "step": 1661
+ },
+ {
+ "epoch": 23.083842794759825,
+ "grad_norm": 0.02511233650147915,
+ "learning_rate": 0.0006,
+ "loss": 4.414128303527832,
+ "step": 1662
+ },
+ {
+ "epoch": 23.097816593886463,
+ "grad_norm": 0.025802113115787506,
+ "learning_rate": 0.0006,
+ "loss": 4.388265609741211,
+ "step": 1663
+ },
+ {
+ "epoch": 23.111790393013102,
+ "grad_norm": 0.02377348765730858,
+ "learning_rate": 0.0006,
+ "loss": 4.357505798339844,
+ "step": 1664
+ },
+ {
+ "epoch": 23.125764192139737,
+ "grad_norm": 0.020876651629805565,
+ "learning_rate": 0.0006,
+ "loss": 4.23043966293335,
+ "step": 1665
+ },
+ {
+ "epoch": 23.139737991266376,
+ "grad_norm": 0.01959727331995964,
+ "learning_rate": 0.0006,
+ "loss": 4.280970096588135,
+ "step": 1666
+ },
+ {
+ "epoch": 23.153711790393015,
+ "grad_norm": 0.018678616732358932,
+ "learning_rate": 0.0006,
+ "loss": 4.346716403961182,
+ "step": 1667
+ },
+ {
+ "epoch": 23.16768558951965,
+ "grad_norm": 0.02108944021165371,
+ "learning_rate": 0.0006,
+ "loss": 4.277904510498047,
+ "step": 1668
+ },
+ {
+ "epoch": 23.18165938864629,
+ "grad_norm": 0.020221566781401634,
+ "learning_rate": 0.0006,
+ "loss": 4.367693901062012,
+ "step": 1669
+ },
+ {
+ "epoch": 23.195633187772927,
+ "grad_norm": 0.019599711522459984,
+ "learning_rate": 0.0006,
+ "loss": 4.257092475891113,
+ "step": 1670
+ },
+ {
+ "epoch": 23.209606986899562,
+ "grad_norm": 0.019833361729979515,
+ "learning_rate": 0.0006,
+ "loss": 4.540737628936768,
+ "step": 1671
+ },
+ {
+ "epoch": 23.2235807860262,
+ "grad_norm": 0.020352095365524292,
+ "learning_rate": 0.0006,
+ "loss": 4.360923767089844,
+ "step": 1672
+ },
+ {
+ "epoch": 23.23755458515284,
+ "grad_norm": 0.01939748041331768,
+ "learning_rate": 0.0006,
+ "loss": 4.393110275268555,
+ "step": 1673
+ },
+ {
+ "epoch": 23.251528384279474,
+ "grad_norm": 0.01884354278445244,
+ "learning_rate": 0.0006,
+ "loss": 4.383938789367676,
+ "step": 1674
+ },
+ {
+ "epoch": 23.265502183406113,
+ "grad_norm": 0.02008470520377159,
+ "learning_rate": 0.0006,
+ "loss": 4.346344947814941,
+ "step": 1675
+ },
+ {
+ "epoch": 23.27947598253275,
+ "grad_norm": 0.019065195694565773,
+ "learning_rate": 0.0006,
+ "loss": 4.36611795425415,
+ "step": 1676
+ },
+ {
+ "epoch": 23.29344978165939,
+ "grad_norm": 0.01938541606068611,
+ "learning_rate": 0.0006,
+ "loss": 4.4009599685668945,
+ "step": 1677
+ },
+ {
+ "epoch": 23.307423580786025,
+ "grad_norm": 0.019044723361730576,
+ "learning_rate": 0.0006,
+ "loss": 4.340169906616211,
+ "step": 1678
+ },
+ {
+ "epoch": 23.321397379912664,
+ "grad_norm": 0.018850378692150116,
+ "learning_rate": 0.0006,
+ "loss": 4.37674617767334,
+ "step": 1679
+ },
+ {
+ "epoch": 23.335371179039303,
+ "grad_norm": 0.019076095893979073,
+ "learning_rate": 0.0006,
+ "loss": 4.226874351501465,
+ "step": 1680
+ },
+ {
+ "epoch": 23.349344978165938,
+ "grad_norm": 0.017543038353323936,
+ "learning_rate": 0.0006,
+ "loss": 4.337751388549805,
+ "step": 1681
+ },
+ {
+ "epoch": 23.363318777292577,
+ "grad_norm": 0.017170218750834465,
+ "learning_rate": 0.0006,
+ "loss": 4.282512664794922,
+ "step": 1682
+ },
+ {
+ "epoch": 23.377292576419215,
+ "grad_norm": 0.01574796624481678,
+ "learning_rate": 0.0006,
+ "loss": 4.30638313293457,
+ "step": 1683
+ },
+ {
+ "epoch": 23.39126637554585,
+ "grad_norm": 0.015160846523940563,
+ "learning_rate": 0.0006,
+ "loss": 4.336716175079346,
+ "step": 1684
+ },
+ {
+ "epoch": 23.40524017467249,
+ "grad_norm": 0.01534526702016592,
+ "learning_rate": 0.0006,
+ "loss": 4.2174787521362305,
+ "step": 1685
+ },
+ {
+ "epoch": 23.419213973799128,
+ "grad_norm": 0.015289999544620514,
+ "learning_rate": 0.0006,
+ "loss": 4.272546291351318,
+ "step": 1686
+ },
+ {
+ "epoch": 23.433187772925763,
+ "grad_norm": 0.014894185587763786,
+ "learning_rate": 0.0006,
+ "loss": 4.296045303344727,
+ "step": 1687
+ },
+ {
+ "epoch": 23.4471615720524,
+ "grad_norm": 0.014869502745568752,
+ "learning_rate": 0.0006,
+ "loss": 4.425461769104004,
+ "step": 1688
+ },
+ {
+ "epoch": 23.46113537117904,
+ "grad_norm": 0.01474483497440815,
+ "learning_rate": 0.0006,
+ "loss": 4.361867427825928,
+ "step": 1689
+ },
+ {
+ "epoch": 23.475109170305675,
+ "grad_norm": 0.015299245715141296,
+ "learning_rate": 0.0006,
+ "loss": 4.364523410797119,
+ "step": 1690
+ },
+ {
+ "epoch": 23.489082969432314,
+ "grad_norm": 0.015201340429484844,
+ "learning_rate": 0.0006,
+ "loss": 4.285182952880859,
+ "step": 1691
+ },
+ {
+ "epoch": 23.503056768558952,
+ "grad_norm": 0.016057293862104416,
+ "learning_rate": 0.0006,
+ "loss": 4.284873962402344,
+ "step": 1692
+ },
+ {
+ "epoch": 23.51703056768559,
+ "grad_norm": 0.015871062874794006,
+ "learning_rate": 0.0006,
+ "loss": 4.326470851898193,
+ "step": 1693
+ },
+ {
+ "epoch": 23.531004366812226,
+ "grad_norm": 0.01586288772523403,
+ "learning_rate": 0.0006,
+ "loss": 4.324883460998535,
+ "step": 1694
+ },
+ {
+ "epoch": 23.544978165938865,
+ "grad_norm": 0.016092827543616295,
+ "learning_rate": 0.0006,
+ "loss": 4.3258867263793945,
+ "step": 1695
+ },
+ {
+ "epoch": 23.558951965065503,
+ "grad_norm": 0.01582016795873642,
+ "learning_rate": 0.0006,
+ "loss": 4.358892440795898,
+ "step": 1696
+ },
+ {
+ "epoch": 23.57292576419214,
+ "grad_norm": 0.01688452623784542,
+ "learning_rate": 0.0006,
+ "loss": 4.280842304229736,
+ "step": 1697
+ },
+ {
+ "epoch": 23.586899563318777,
+ "grad_norm": 0.015865886583924294,
+ "learning_rate": 0.0006,
+ "loss": 4.245567321777344,
+ "step": 1698
+ },
+ {
+ "epoch": 23.600873362445416,
+ "grad_norm": 0.015303662046790123,
+ "learning_rate": 0.0006,
+ "loss": 4.383112907409668,
+ "step": 1699
+ },
+ {
+ "epoch": 23.61484716157205,
+ "grad_norm": 0.015649516135454178,
+ "learning_rate": 0.0006,
+ "loss": 4.519082069396973,
+ "step": 1700
+ },
+ {
+ "epoch": 23.62882096069869,
+ "grad_norm": 0.0167120099067688,
+ "learning_rate": 0.0006,
+ "loss": 4.447847366333008,
+ "step": 1701
+ },
+ {
+ "epoch": 23.64279475982533,
+ "grad_norm": 0.016368716955184937,
+ "learning_rate": 0.0006,
+ "loss": 4.434186935424805,
+ "step": 1702
+ },
+ {
+ "epoch": 23.656768558951963,
+ "grad_norm": 0.017379503697156906,
+ "learning_rate": 0.0006,
+ "loss": 4.317594528198242,
+ "step": 1703
+ },
+ {
+ "epoch": 23.670742358078602,
+ "grad_norm": 0.018573811277747154,
+ "learning_rate": 0.0006,
+ "loss": 4.205793857574463,
+ "step": 1704
+ },
+ {
+ "epoch": 23.68471615720524,
+ "grad_norm": 0.01749090477824211,
+ "learning_rate": 0.0006,
+ "loss": 4.3985700607299805,
+ "step": 1705
+ },
+ {
+ "epoch": 23.69868995633188,
+ "grad_norm": 0.0165668074041605,
+ "learning_rate": 0.0006,
+ "loss": 4.329226970672607,
+ "step": 1706
+ },
+ {
+ "epoch": 23.712663755458514,
+ "grad_norm": 0.016913846135139465,
+ "learning_rate": 0.0006,
+ "loss": 4.388485908508301,
+ "step": 1707
+ },
+ {
+ "epoch": 23.726637554585153,
+ "grad_norm": 0.016813859343528748,
+ "learning_rate": 0.0006,
+ "loss": 4.3395915031433105,
+ "step": 1708
+ },
+ {
+ "epoch": 23.74061135371179,
+ "grad_norm": 0.01676975190639496,
+ "learning_rate": 0.0006,
+ "loss": 4.254745960235596,
+ "step": 1709
+ },
+ {
+ "epoch": 23.754585152838427,
+ "grad_norm": 0.016370350494980812,
+ "learning_rate": 0.0006,
+ "loss": 4.2612223625183105,
+ "step": 1710
+ },
+ {
+ "epoch": 23.768558951965066,
+ "grad_norm": 0.01696198247373104,
+ "learning_rate": 0.0006,
+ "loss": 4.327524662017822,
+ "step": 1711
+ },
+ {
+ "epoch": 23.782532751091704,
+ "grad_norm": 0.0189108457416296,
+ "learning_rate": 0.0006,
+ "loss": 4.325839519500732,
+ "step": 1712
+ },
+ {
+ "epoch": 23.79650655021834,
+ "grad_norm": 0.01862351782619953,
+ "learning_rate": 0.0006,
+ "loss": 4.333662509918213,
+ "step": 1713
+ },
+ {
+ "epoch": 23.810480349344978,
+ "grad_norm": 0.01655014418065548,
+ "learning_rate": 0.0006,
+ "loss": 4.39305305480957,
+ "step": 1714
+ },
+ {
+ "epoch": 23.824454148471617,
+ "grad_norm": 0.015899376943707466,
+ "learning_rate": 0.0006,
+ "loss": 4.390053749084473,
+ "step": 1715
+ },
+ {
+ "epoch": 23.83842794759825,
+ "grad_norm": 0.016519056633114815,
+ "learning_rate": 0.0006,
+ "loss": 4.255903244018555,
+ "step": 1716
+ },
+ {
+ "epoch": 23.85240174672489,
+ "grad_norm": 0.0167539119720459,
+ "learning_rate": 0.0006,
+ "loss": 4.255727767944336,
+ "step": 1717
+ },
+ {
+ "epoch": 23.86637554585153,
+ "grad_norm": 0.0171353816986084,
+ "learning_rate": 0.0006,
+ "loss": 4.376740455627441,
+ "step": 1718
+ },
+ {
+ "epoch": 23.880349344978168,
+ "grad_norm": 0.016719117760658264,
+ "learning_rate": 0.0006,
+ "loss": 4.3376688957214355,
+ "step": 1719
+ },
+ {
+ "epoch": 23.894323144104803,
+ "grad_norm": 0.01561494916677475,
+ "learning_rate": 0.0006,
+ "loss": 4.3345746994018555,
+ "step": 1720
+ },
+ {
+ "epoch": 23.90829694323144,
+ "grad_norm": 0.016303174197673798,
+ "learning_rate": 0.0006,
+ "loss": 4.332371234893799,
+ "step": 1721
+ },
+ {
+ "epoch": 23.92227074235808,
+ "grad_norm": 0.016722865402698517,
+ "learning_rate": 0.0006,
+ "loss": 4.267125129699707,
+ "step": 1722
+ },
+ {
+ "epoch": 23.936244541484715,
+ "grad_norm": 0.017072124406695366,
+ "learning_rate": 0.0006,
+ "loss": 4.322751998901367,
+ "step": 1723
+ },
+ {
+ "epoch": 23.950218340611354,
+ "grad_norm": 0.01699932850897312,
+ "learning_rate": 0.0006,
+ "loss": 4.378422737121582,
+ "step": 1724
+ },
+ {
+ "epoch": 23.964192139737992,
+ "grad_norm": 0.01742720976471901,
+ "learning_rate": 0.0006,
+ "loss": 4.345083236694336,
+ "step": 1725
+ },
+ {
+ "epoch": 23.978165938864628,
+ "grad_norm": 0.018797501921653748,
+ "learning_rate": 0.0006,
+ "loss": 4.290443420410156,
+ "step": 1726
+ },
+ {
+ "epoch": 23.992139737991266,
+ "grad_norm": 0.018090683966875076,
+ "learning_rate": 0.0006,
+ "loss": 4.377580642700195,
+ "step": 1727
+ },
+ {
+ "epoch": 24.0,
+ "grad_norm": 0.017305459827184677,
+ "learning_rate": 0.0006,
+ "loss": 4.364426612854004,
+ "step": 1728
+ },
+ {
+ "epoch": 24.0,
+ "eval_loss": 4.647762298583984,
+ "eval_runtime": 56.5402,
+ "eval_samples_per_second": 43.191,
+ "eval_steps_per_second": 1.362,
+ "step": 1728
+ },
+ {
+ "epoch": 24.01397379912664,
+ "grad_norm": 0.016493607312440872,
+ "learning_rate": 0.0006,
+ "loss": 4.252255916595459,
+ "step": 1729
+ },
+ {
+ "epoch": 24.027947598253274,
+ "grad_norm": 0.016529450193047523,
+ "learning_rate": 0.0006,
+ "loss": 4.352746963500977,
+ "step": 1730
+ },
+ {
+ "epoch": 24.041921397379912,
+ "grad_norm": 0.016940191388130188,
+ "learning_rate": 0.0006,
+ "loss": 4.301870822906494,
+ "step": 1731
+ },
+ {
+ "epoch": 24.05589519650655,
+ "grad_norm": 0.01752558909356594,
+ "learning_rate": 0.0006,
+ "loss": 4.2762451171875,
+ "step": 1732
+ },
+ {
+ "epoch": 24.069868995633186,
+ "grad_norm": 0.017112884670495987,
+ "learning_rate": 0.0006,
+ "loss": 4.2659149169921875,
+ "step": 1733
+ },
+ {
+ "epoch": 24.083842794759825,
+ "grad_norm": 0.01900586113333702,
+ "learning_rate": 0.0006,
+ "loss": 4.306058883666992,
+ "step": 1734
+ },
+ {
+ "epoch": 24.097816593886463,
+ "grad_norm": 0.019447680562734604,
+ "learning_rate": 0.0006,
+ "loss": 4.450128555297852,
+ "step": 1735
+ },
+ {
+ "epoch": 24.111790393013102,
+ "grad_norm": 0.01948186755180359,
+ "learning_rate": 0.0006,
+ "loss": 4.202869415283203,
+ "step": 1736
+ },
+ {
+ "epoch": 24.125764192139737,
+ "grad_norm": 0.01716383546590805,
+ "learning_rate": 0.0006,
+ "loss": 4.318876266479492,
+ "step": 1737
+ },
+ {
+ "epoch": 24.139737991266376,
+ "grad_norm": 0.017049988731741905,
+ "learning_rate": 0.0006,
+ "loss": 4.237238883972168,
+ "step": 1738
+ },
+ {
+ "epoch": 24.153711790393015,
+ "grad_norm": 0.015169923193752766,
+ "learning_rate": 0.0006,
+ "loss": 4.289426803588867,
+ "step": 1739
+ },
+ {
+ "epoch": 24.16768558951965,
+ "grad_norm": 0.014760667458176613,
+ "learning_rate": 0.0006,
+ "loss": 4.356924057006836,
+ "step": 1740
+ },
+ {
+ "epoch": 24.18165938864629,
+ "grad_norm": 0.015127750113606453,
+ "learning_rate": 0.0006,
+ "loss": 4.142916679382324,
+ "step": 1741
+ },
+ {
+ "epoch": 24.195633187772927,
+ "grad_norm": 0.01549555640667677,
+ "learning_rate": 0.0006,
+ "loss": 4.345563888549805,
+ "step": 1742
+ },
+ {
+ "epoch": 24.209606986899562,
+ "grad_norm": 0.016393091529607773,
+ "learning_rate": 0.0006,
+ "loss": 4.417495250701904,
+ "step": 1743
+ },
+ {
+ "epoch": 24.2235807860262,
+ "grad_norm": 0.01785266026854515,
+ "learning_rate": 0.0006,
+ "loss": 4.3553547859191895,
+ "step": 1744
+ },
+ {
+ "epoch": 24.23755458515284,
+ "grad_norm": 0.017574617639183998,
+ "learning_rate": 0.0006,
+ "loss": 4.241024494171143,
+ "step": 1745
+ },
+ {
+ "epoch": 24.251528384279474,
+ "grad_norm": 0.019060306251049042,
+ "learning_rate": 0.0006,
+ "loss": 4.3685150146484375,
+ "step": 1746
+ },
+ {
+ "epoch": 24.265502183406113,
+ "grad_norm": 0.0214633010327816,
+ "learning_rate": 0.0006,
+ "loss": 4.323866844177246,
+ "step": 1747
+ },
+ {
+ "epoch": 24.27947598253275,
+ "grad_norm": 0.021055003628134727,
+ "learning_rate": 0.0006,
+ "loss": 4.278757095336914,
+ "step": 1748
+ },
+ {
+ "epoch": 24.29344978165939,
+ "grad_norm": 0.018618572503328323,
+ "learning_rate": 0.0006,
+ "loss": 4.28849983215332,
+ "step": 1749
+ },
+ {
+ "epoch": 24.307423580786025,
+ "grad_norm": 0.018666435033082962,
+ "learning_rate": 0.0006,
+ "loss": 4.368590354919434,
+ "step": 1750
+ },
+ {
+ "epoch": 24.321397379912664,
+ "grad_norm": 0.020626146346330643,
+ "learning_rate": 0.0006,
+ "loss": 4.2951836585998535,
+ "step": 1751
+ },
+ {
+ "epoch": 24.335371179039303,
+ "grad_norm": 0.01929759792983532,
+ "learning_rate": 0.0006,
+ "loss": 4.3664703369140625,
+ "step": 1752
+ },
+ {
+ "epoch": 24.349344978165938,
+ "grad_norm": 0.020956002175807953,
+ "learning_rate": 0.0006,
+ "loss": 4.345308303833008,
+ "step": 1753
+ },
+ {
+ "epoch": 24.363318777292577,
+ "grad_norm": 0.019227957352995872,
+ "learning_rate": 0.0006,
+ "loss": 4.349849224090576,
+ "step": 1754
+ },
+ {
+ "epoch": 24.377292576419215,
+ "grad_norm": 0.015970097854733467,
+ "learning_rate": 0.0006,
+ "loss": 4.223012447357178,
+ "step": 1755
+ },
+ {
+ "epoch": 24.39126637554585,
+ "grad_norm": 0.015951646491885185,
+ "learning_rate": 0.0006,
+ "loss": 4.236420631408691,
+ "step": 1756
+ },
+ {
+ "epoch": 24.40524017467249,
+ "grad_norm": 0.016032515093684196,
+ "learning_rate": 0.0006,
+ "loss": 4.226899147033691,
+ "step": 1757
+ },
+ {
+ "epoch": 24.419213973799128,
+ "grad_norm": 0.01573132909834385,
+ "learning_rate": 0.0006,
+ "loss": 4.315916061401367,
+ "step": 1758
+ },
+ {
+ "epoch": 24.433187772925763,
+ "grad_norm": 0.015367005951702595,
+ "learning_rate": 0.0006,
+ "loss": 4.358939170837402,
+ "step": 1759
+ },
+ {
+ "epoch": 24.4471615720524,
+ "grad_norm": 0.01669907011091709,
+ "learning_rate": 0.0006,
+ "loss": 4.232147216796875,
+ "step": 1760
+ },
+ {
+ "epoch": 24.46113537117904,
+ "grad_norm": 0.018708152696490288,
+ "learning_rate": 0.0006,
+ "loss": 4.329852104187012,
+ "step": 1761
+ },
+ {
+ "epoch": 24.475109170305675,
+ "grad_norm": 0.019035987555980682,
+ "learning_rate": 0.0006,
+ "loss": 4.260340213775635,
+ "step": 1762
+ },
+ {
+ "epoch": 24.489082969432314,
+ "grad_norm": 0.01832144521176815,
+ "learning_rate": 0.0006,
+ "loss": 4.359543800354004,
+ "step": 1763
+ },
+ {
+ "epoch": 24.503056768558952,
+ "grad_norm": 0.015313294716179371,
+ "learning_rate": 0.0006,
+ "loss": 4.279088020324707,
+ "step": 1764
+ },
+ {
+ "epoch": 24.51703056768559,
+ "grad_norm": 0.015803923830389977,
+ "learning_rate": 0.0006,
+ "loss": 4.230682849884033,
+ "step": 1765
+ },
+ {
+ "epoch": 24.531004366812226,
+ "grad_norm": 0.016748374328017235,
+ "learning_rate": 0.0006,
+ "loss": 4.349661827087402,
+ "step": 1766
+ },
+ {
+ "epoch": 24.544978165938865,
+ "grad_norm": 0.015732428058981895,
+ "learning_rate": 0.0006,
+ "loss": 4.372441291809082,
+ "step": 1767
+ },
+ {
+ "epoch": 24.558951965065503,
+ "grad_norm": 0.016564829275012016,
+ "learning_rate": 0.0006,
+ "loss": 4.323090553283691,
+ "step": 1768
+ },
+ {
+ "epoch": 24.57292576419214,
+ "grad_norm": 0.01603010483086109,
+ "learning_rate": 0.0006,
+ "loss": 4.347898006439209,
+ "step": 1769
+ },
+ {
+ "epoch": 24.586899563318777,
+ "grad_norm": 0.01674809865653515,
+ "learning_rate": 0.0006,
+ "loss": 4.269165515899658,
+ "step": 1770
+ },
+ {
+ "epoch": 24.600873362445416,
+ "grad_norm": 0.017333930358290672,
+ "learning_rate": 0.0006,
+ "loss": 4.353298187255859,
+ "step": 1771
+ },
+ {
+ "epoch": 24.61484716157205,
+ "grad_norm": 0.017519580200314522,
+ "learning_rate": 0.0006,
+ "loss": 4.265199661254883,
+ "step": 1772
+ },
+ {
+ "epoch": 24.62882096069869,
+ "grad_norm": 0.017675306648015976,
+ "learning_rate": 0.0006,
+ "loss": 4.322596549987793,
+ "step": 1773
+ },
+ {
+ "epoch": 24.64279475982533,
+ "grad_norm": 0.017166534438729286,
+ "learning_rate": 0.0006,
+ "loss": 4.413900375366211,
+ "step": 1774
+ },
+ {
+ "epoch": 24.656768558951963,
+ "grad_norm": 0.016362346708774567,
+ "learning_rate": 0.0006,
+ "loss": 4.307236671447754,
+ "step": 1775
+ },
+ {
+ "epoch": 24.670742358078602,
+ "grad_norm": 0.016086872667074203,
+ "learning_rate": 0.0006,
+ "loss": 4.235316276550293,
+ "step": 1776
+ },
+ {
+ "epoch": 24.68471615720524,
+ "grad_norm": 0.01689537614583969,
+ "learning_rate": 0.0006,
+ "loss": 4.390591621398926,
+ "step": 1777
+ },
+ {
+ "epoch": 24.69868995633188,
+ "grad_norm": 0.016836725175380707,
+ "learning_rate": 0.0006,
+ "loss": 4.2990031242370605,
+ "step": 1778
+ },
+ {
+ "epoch": 24.712663755458514,
+ "grad_norm": 0.015627335757017136,
+ "learning_rate": 0.0006,
+ "loss": 4.166745185852051,
+ "step": 1779
+ },
+ {
+ "epoch": 24.726637554585153,
+ "grad_norm": 0.015007016249001026,
+ "learning_rate": 0.0006,
+ "loss": 4.295273303985596,
+ "step": 1780
+ },
+ {
+ "epoch": 24.74061135371179,
+ "grad_norm": 0.016084209084510803,
+ "learning_rate": 0.0006,
+ "loss": 4.342424392700195,
+ "step": 1781
+ },
+ {
+ "epoch": 24.754585152838427,
+ "grad_norm": 0.01595073565840721,
+ "learning_rate": 0.0006,
+ "loss": 4.3060102462768555,
+ "step": 1782
+ },
+ {
+ "epoch": 24.768558951965066,
+ "grad_norm": 0.016113461926579475,
+ "learning_rate": 0.0006,
+ "loss": 4.267908096313477,
+ "step": 1783
+ },
+ {
+ "epoch": 24.782532751091704,
+ "grad_norm": 0.01564556173980236,
+ "learning_rate": 0.0006,
+ "loss": 4.324642658233643,
+ "step": 1784
+ },
+ {
+ "epoch": 24.79650655021834,
+ "grad_norm": 0.01575268618762493,
+ "learning_rate": 0.0006,
+ "loss": 4.328181743621826,
+ "step": 1785
+ },
+ {
+ "epoch": 24.810480349344978,
+ "grad_norm": 0.016830824315547943,
+ "learning_rate": 0.0006,
+ "loss": 4.247906684875488,
+ "step": 1786
+ },
+ {
+ "epoch": 24.824454148471617,
+ "grad_norm": 0.017989547923207283,
+ "learning_rate": 0.0006,
+ "loss": 4.228281021118164,
+ "step": 1787
+ },
+ {
+ "epoch": 24.83842794759825,
+ "grad_norm": 0.016783086583018303,
+ "learning_rate": 0.0006,
+ "loss": 4.2880964279174805,
+ "step": 1788
+ },
+ {
+ "epoch": 24.85240174672489,
+ "grad_norm": 0.016467561945319176,
+ "learning_rate": 0.0006,
+ "loss": 4.36118745803833,
+ "step": 1789
+ },
+ {
+ "epoch": 24.86637554585153,
+ "grad_norm": 0.01780523732304573,
+ "learning_rate": 0.0006,
+ "loss": 4.172321319580078,
+ "step": 1790
+ },
+ {
+ "epoch": 24.880349344978168,
+ "grad_norm": 0.017102031037211418,
+ "learning_rate": 0.0006,
+ "loss": 4.352826118469238,
+ "step": 1791
+ },
+ {
+ "epoch": 24.894323144104803,
+ "grad_norm": 0.016409218311309814,
+ "learning_rate": 0.0006,
+ "loss": 4.212162017822266,
+ "step": 1792
+ },
+ {
+ "epoch": 24.90829694323144,
+ "grad_norm": 0.018660584464669228,
+ "learning_rate": 0.0006,
+ "loss": 4.346684455871582,
+ "step": 1793
+ },
+ {
+ "epoch": 24.92227074235808,
+ "grad_norm": 0.017389440909028053,
+ "learning_rate": 0.0006,
+ "loss": 4.197913646697998,
+ "step": 1794
+ },
+ {
+ "epoch": 24.936244541484715,
+ "grad_norm": 0.01596485823392868,
+ "learning_rate": 0.0006,
+ "loss": 4.370617866516113,
+ "step": 1795
+ },
+ {
+ "epoch": 24.950218340611354,
+ "grad_norm": 0.01709776371717453,
+ "learning_rate": 0.0006,
+ "loss": 4.313847064971924,
+ "step": 1796
+ },
+ {
+ "epoch": 24.964192139737992,
+ "grad_norm": 0.016571134328842163,
+ "learning_rate": 0.0006,
+ "loss": 4.1895856857299805,
+ "step": 1797
+ },
+ {
+ "epoch": 24.978165938864628,
+ "grad_norm": 0.016851048916578293,
+ "learning_rate": 0.0006,
+ "loss": 4.350772857666016,
+ "step": 1798
+ },
+ {
+ "epoch": 24.992139737991266,
+ "grad_norm": 0.017007440328598022,
+ "learning_rate": 0.0006,
+ "loss": 4.399008750915527,
+ "step": 1799
+ },
+ {
+ "epoch": 25.0,
+ "grad_norm": 0.01791212148964405,
+ "learning_rate": 0.0006,
+ "loss": 4.306121826171875,
+ "step": 1800
+ },
+ {
+ "epoch": 25.0,
+ "eval_loss": 4.636417865753174,
+ "eval_runtime": 56.8296,
+ "eval_samples_per_second": 42.971,
+ "eval_steps_per_second": 1.355,
+ "step": 1800
+ },
+ {
+ "epoch": 25.01397379912664,
+ "grad_norm": 0.01729617640376091,
+ "learning_rate": 0.0006,
+ "loss": 4.257018089294434,
+ "step": 1801
+ },
+ {
+ "epoch": 25.027947598253274,
+ "grad_norm": 0.017466643825173378,
+ "learning_rate": 0.0006,
+ "loss": 4.307605266571045,
+ "step": 1802
+ },
+ {
+ "epoch": 25.041921397379912,
+ "grad_norm": 0.018614279106259346,
+ "learning_rate": 0.0006,
+ "loss": 4.360141277313232,
+ "step": 1803
+ },
+ {
+ "epoch": 25.05589519650655,
+ "grad_norm": 0.02010771445930004,
+ "learning_rate": 0.0006,
+ "loss": 4.196428298950195,
+ "step": 1804
+ },
+ {
+ "epoch": 25.069868995633186,
+ "grad_norm": 0.02090228535234928,
+ "learning_rate": 0.0006,
+ "loss": 4.288947105407715,
+ "step": 1805
+ },
+ {
+ "epoch": 25.083842794759825,
+ "grad_norm": 0.02050255797803402,
+ "learning_rate": 0.0006,
+ "loss": 4.259582042694092,
+ "step": 1806
+ },
+ {
+ "epoch": 25.097816593886463,
+ "grad_norm": 0.021844089031219482,
+ "learning_rate": 0.0006,
+ "loss": 4.405138969421387,
+ "step": 1807
+ },
+ {
+ "epoch": 25.111790393013102,
+ "grad_norm": 0.022287018597126007,
+ "learning_rate": 0.0006,
+ "loss": 4.333793640136719,
+ "step": 1808
+ },
+ {
+ "epoch": 25.125764192139737,
+ "grad_norm": 0.02288329415023327,
+ "learning_rate": 0.0006,
+ "loss": 4.3910980224609375,
+ "step": 1809
+ },
+ {
+ "epoch": 25.139737991266376,
+ "grad_norm": 0.02373199723660946,
+ "learning_rate": 0.0006,
+ "loss": 4.32921028137207,
+ "step": 1810
+ },
+ {
+ "epoch": 25.153711790393015,
+ "grad_norm": 0.024960234761238098,
+ "learning_rate": 0.0006,
+ "loss": 4.295609474182129,
+ "step": 1811
+ },
+ {
+ "epoch": 25.16768558951965,
+ "grad_norm": 0.026743892580270767,
+ "learning_rate": 0.0006,
+ "loss": 4.315308094024658,
+ "step": 1812
+ },
+ {
+ "epoch": 25.18165938864629,
+ "grad_norm": 0.022138185799121857,
+ "learning_rate": 0.0006,
+ "loss": 4.360208511352539,
+ "step": 1813
+ },
+ {
+ "epoch": 25.195633187772927,
+ "grad_norm": 0.022120574489235878,
+ "learning_rate": 0.0006,
+ "loss": 4.1973772048950195,
+ "step": 1814
+ },
+ {
+ "epoch": 25.209606986899562,
+ "grad_norm": 0.02230502851307392,
+ "learning_rate": 0.0006,
+ "loss": 4.285558700561523,
+ "step": 1815
+ },
+ {
+ "epoch": 25.2235807860262,
+ "grad_norm": 0.022301986813545227,
+ "learning_rate": 0.0006,
+ "loss": 4.238343238830566,
+ "step": 1816
+ },
+ {
+ "epoch": 25.23755458515284,
+ "grad_norm": 0.020050447434186935,
+ "learning_rate": 0.0006,
+ "loss": 4.298235893249512,
+ "step": 1817
+ },
+ {
+ "epoch": 25.251528384279474,
+ "grad_norm": 0.01987724006175995,
+ "learning_rate": 0.0006,
+ "loss": 4.301384449005127,
+ "step": 1818
+ },
+ {
+ "epoch": 25.265502183406113,
+ "grad_norm": 0.020440855994820595,
+ "learning_rate": 0.0006,
+ "loss": 4.269218444824219,
+ "step": 1819
+ },
+ {
+ "epoch": 25.27947598253275,
+ "grad_norm": 0.01891893707215786,
+ "learning_rate": 0.0006,
+ "loss": 4.2401957511901855,
+ "step": 1820
+ },
+ {
+ "epoch": 25.29344978165939,
+ "grad_norm": 0.01903688907623291,
+ "learning_rate": 0.0006,
+ "loss": 4.139670372009277,
+ "step": 1821
+ },
+ {
+ "epoch": 25.307423580786025,
+ "grad_norm": 0.018856938928365707,
+ "learning_rate": 0.0006,
+ "loss": 4.233214378356934,
+ "step": 1822
+ },
+ {
+ "epoch": 25.321397379912664,
+ "grad_norm": 0.017791323363780975,
+ "learning_rate": 0.0006,
+ "loss": 4.2342023849487305,
+ "step": 1823
+ },
+ {
+ "epoch": 25.335371179039303,
+ "grad_norm": 0.017304565757513046,
+ "learning_rate": 0.0006,
+ "loss": 4.355198860168457,
+ "step": 1824
+ },
+ {
+ "epoch": 25.349344978165938,
+ "grad_norm": 0.01576206088066101,
+ "learning_rate": 0.0006,
+ "loss": 4.114147186279297,
+ "step": 1825
+ },
+ {
+ "epoch": 25.363318777292577,
+ "grad_norm": 0.015105154365301132,
+ "learning_rate": 0.0006,
+ "loss": 4.313795566558838,
+ "step": 1826
+ },
+ {
+ "epoch": 25.377292576419215,
+ "grad_norm": 0.01563587784767151,
+ "learning_rate": 0.0006,
+ "loss": 4.308282852172852,
+ "step": 1827
+ },
+ {
+ "epoch": 25.39126637554585,
+ "grad_norm": 0.015980906784534454,
+ "learning_rate": 0.0006,
+ "loss": 4.345149040222168,
+ "step": 1828
+ },
+ {
+ "epoch": 25.40524017467249,
+ "grad_norm": 0.01669284701347351,
+ "learning_rate": 0.0006,
+ "loss": 4.237611293792725,
+ "step": 1829
+ },
+ {
+ "epoch": 25.419213973799128,
+ "grad_norm": 0.016202064231038094,
+ "learning_rate": 0.0006,
+ "loss": 4.3072309494018555,
+ "step": 1830
+ },
+ {
+ "epoch": 25.433187772925763,
+ "grad_norm": 0.01478662807494402,
+ "learning_rate": 0.0006,
+ "loss": 4.276215076446533,
+ "step": 1831
+ },
+ {
+ "epoch": 25.4471615720524,
+ "grad_norm": 0.015168731100857258,
+ "learning_rate": 0.0006,
+ "loss": 4.37509822845459,
+ "step": 1832
+ },
+ {
+ "epoch": 25.46113537117904,
+ "grad_norm": 0.01648980937898159,
+ "learning_rate": 0.0006,
+ "loss": 4.349167346954346,
+ "step": 1833
+ },
+ {
+ "epoch": 25.475109170305675,
+ "grad_norm": 0.01708398200571537,
+ "learning_rate": 0.0006,
+ "loss": 4.346580505371094,
+ "step": 1834
+ },
+ {
+ "epoch": 25.489082969432314,
+ "grad_norm": 0.017370784655213356,
+ "learning_rate": 0.0006,
+ "loss": 4.299466133117676,
+ "step": 1835
+ },
+ {
+ "epoch": 25.503056768558952,
+ "grad_norm": 0.016248228028416634,
+ "learning_rate": 0.0006,
+ "loss": 4.184610366821289,
+ "step": 1836
+ },
+ {
+ "epoch": 25.51703056768559,
+ "grad_norm": 0.017106913030147552,
+ "learning_rate": 0.0006,
+ "loss": 4.366856098175049,
+ "step": 1837
+ },
+ {
+ "epoch": 25.531004366812226,
+ "grad_norm": 0.016071965917944908,
+ "learning_rate": 0.0006,
+ "loss": 4.297691345214844,
+ "step": 1838
+ },
+ {
+ "epoch": 25.544978165938865,
+ "grad_norm": 0.016451986506581306,
+ "learning_rate": 0.0006,
+ "loss": 4.229983329772949,
+ "step": 1839
+ },
+ {
+ "epoch": 25.558951965065503,
+ "grad_norm": 0.016195081174373627,
+ "learning_rate": 0.0006,
+ "loss": 4.259893417358398,
+ "step": 1840
+ },
+ {
+ "epoch": 25.57292576419214,
+ "grad_norm": 0.0165175162255764,
+ "learning_rate": 0.0006,
+ "loss": 4.3348236083984375,
+ "step": 1841
+ },
+ {
+ "epoch": 25.586899563318777,
+ "grad_norm": 0.01651517115533352,
+ "learning_rate": 0.0006,
+ "loss": 4.319541931152344,
+ "step": 1842
+ },
+ {
+ "epoch": 25.600873362445416,
+ "grad_norm": 0.016143690794706345,
+ "learning_rate": 0.0006,
+ "loss": 4.345652103424072,
+ "step": 1843
+ },
+ {
+ "epoch": 25.61484716157205,
+ "grad_norm": 0.015140696428716183,
+ "learning_rate": 0.0006,
+ "loss": 4.212047576904297,
+ "step": 1844
+ },
+ {
+ "epoch": 25.62882096069869,
+ "grad_norm": 0.014637443237006664,
+ "learning_rate": 0.0006,
+ "loss": 4.187453269958496,
+ "step": 1845
+ },
+ {
+ "epoch": 25.64279475982533,
+ "grad_norm": 0.014585614204406738,
+ "learning_rate": 0.0006,
+ "loss": 4.253936767578125,
+ "step": 1846
+ },
+ {
+ "epoch": 25.656768558951963,
+ "grad_norm": 0.015158289112150669,
+ "learning_rate": 0.0006,
+ "loss": 4.34612512588501,
+ "step": 1847
+ },
+ {
+ "epoch": 25.670742358078602,
+ "grad_norm": 0.015202849172055721,
+ "learning_rate": 0.0006,
+ "loss": 4.270634651184082,
+ "step": 1848
+ },
+ {
+ "epoch": 25.68471615720524,
+ "grad_norm": 0.015363575890660286,
+ "learning_rate": 0.0006,
+ "loss": 4.271080493927002,
+ "step": 1849
+ },
+ {
+ "epoch": 25.69868995633188,
+ "grad_norm": 0.014360226690769196,
+ "learning_rate": 0.0006,
+ "loss": 4.375033855438232,
+ "step": 1850
+ },
+ {
+ "epoch": 25.712663755458514,
+ "grad_norm": 0.014263181947171688,
+ "learning_rate": 0.0006,
+ "loss": 4.424138069152832,
+ "step": 1851
+ },
+ {
+ "epoch": 25.726637554585153,
+ "grad_norm": 0.014398688450455666,
+ "learning_rate": 0.0006,
+ "loss": 4.410154819488525,
+ "step": 1852
+ },
+ {
+ "epoch": 25.74061135371179,
+ "grad_norm": 0.016131488606333733,
+ "learning_rate": 0.0006,
+ "loss": 4.324014663696289,
+ "step": 1853
+ },
+ {
+ "epoch": 25.754585152838427,
+ "grad_norm": 0.017568735405802727,
+ "learning_rate": 0.0006,
+ "loss": 4.367861747741699,
+ "step": 1854
+ },
+ {
+ "epoch": 25.768558951965066,
+ "grad_norm": 0.017907511442899704,
+ "learning_rate": 0.0006,
+ "loss": 4.352965354919434,
+ "step": 1855
+ },
+ {
+ "epoch": 25.782532751091704,
+ "grad_norm": 0.016918016597628593,
+ "learning_rate": 0.0006,
+ "loss": 4.252397537231445,
+ "step": 1856
+ },
+ {
+ "epoch": 25.79650655021834,
+ "grad_norm": 0.017274610698223114,
+ "learning_rate": 0.0006,
+ "loss": 4.342309951782227,
+ "step": 1857
+ },
+ {
+ "epoch": 25.810480349344978,
+ "grad_norm": 0.017879825085401535,
+ "learning_rate": 0.0006,
+ "loss": 4.330987453460693,
+ "step": 1858
+ },
+ {
+ "epoch": 25.824454148471617,
+ "grad_norm": 0.01736099272966385,
+ "learning_rate": 0.0006,
+ "loss": 4.307255744934082,
+ "step": 1859
+ },
+ {
+ "epoch": 25.83842794759825,
+ "grad_norm": 0.016575824469327927,
+ "learning_rate": 0.0006,
+ "loss": 4.262016773223877,
+ "step": 1860
+ },
+ {
+ "epoch": 25.85240174672489,
+ "grad_norm": 0.015181140042841434,
+ "learning_rate": 0.0006,
+ "loss": 4.291863441467285,
+ "step": 1861
+ },
+ {
+ "epoch": 25.86637554585153,
+ "grad_norm": 0.015965687111020088,
+ "learning_rate": 0.0006,
+ "loss": 4.424836158752441,
+ "step": 1862
+ },
+ {
+ "epoch": 25.880349344978168,
+ "grad_norm": 0.01687219925224781,
+ "learning_rate": 0.0006,
+ "loss": 4.343531131744385,
+ "step": 1863
+ },
+ {
+ "epoch": 25.894323144104803,
+ "grad_norm": 0.015115504153072834,
+ "learning_rate": 0.0006,
+ "loss": 4.357808589935303,
+ "step": 1864
+ },
+ {
+ "epoch": 25.90829694323144,
+ "grad_norm": 0.014150998555123806,
+ "learning_rate": 0.0006,
+ "loss": 4.324550151824951,
+ "step": 1865
+ },
+ {
+ "epoch": 25.92227074235808,
+ "grad_norm": 0.014989510178565979,
+ "learning_rate": 0.0006,
+ "loss": 4.380428314208984,
+ "step": 1866
+ },
+ {
+ "epoch": 25.936244541484715,
+ "grad_norm": 0.01527960691601038,
+ "learning_rate": 0.0006,
+ "loss": 4.399374961853027,
+ "step": 1867
+ },
+ {
+ "epoch": 25.950218340611354,
+ "grad_norm": 0.015434633940458298,
+ "learning_rate": 0.0006,
+ "loss": 4.260984420776367,
+ "step": 1868
+ },
+ {
+ "epoch": 25.964192139737992,
+ "grad_norm": 0.015370228327810764,
+ "learning_rate": 0.0006,
+ "loss": 4.3340253829956055,
+ "step": 1869
+ },
+ {
+ "epoch": 25.978165938864628,
+ "grad_norm": 0.015348542481660843,
+ "learning_rate": 0.0006,
+ "loss": 4.369132995605469,
+ "step": 1870
+ },
+ {
+ "epoch": 25.992139737991266,
+ "grad_norm": 0.015354936942458153,
+ "learning_rate": 0.0006,
+ "loss": 4.381937026977539,
+ "step": 1871
+ },
+ {
+ "epoch": 26.0,
+ "grad_norm": 0.01796272210776806,
+ "learning_rate": 0.0006,
+ "loss": 4.286768436431885,
+ "step": 1872
+ },
+ {
+ "epoch": 26.0,
+ "eval_loss": 4.576984405517578,
+ "eval_runtime": 57.2091,
+ "eval_samples_per_second": 42.685,
+ "eval_steps_per_second": 1.346,
+ "step": 1872
+ },
+ {
+ "epoch": 26.01397379912664,
+ "grad_norm": 0.01748845726251602,
+ "learning_rate": 0.0006,
+ "loss": 4.337504863739014,
+ "step": 1873
+ },
+ {
+ "epoch": 26.027947598253274,
+ "grad_norm": 0.019244296476244926,
+ "learning_rate": 0.0006,
+ "loss": 4.264280796051025,
+ "step": 1874
+ },
+ {
+ "epoch": 26.041921397379912,
+ "grad_norm": 0.0215692650526762,
+ "learning_rate": 0.0006,
+ "loss": 4.350830078125,
+ "step": 1875
+ },
+ {
+ "epoch": 26.05589519650655,
+ "grad_norm": 0.022189252078533173,
+ "learning_rate": 0.0006,
+ "loss": 4.299098968505859,
+ "step": 1876
+ },
+ {
+ "epoch": 26.069868995633186,
+ "grad_norm": 0.022384043782949448,
+ "learning_rate": 0.0006,
+ "loss": 4.1185173988342285,
+ "step": 1877
+ },
+ {
+ "epoch": 26.083842794759825,
+ "grad_norm": 0.02076014317572117,
+ "learning_rate": 0.0006,
+ "loss": 4.297612190246582,
+ "step": 1878
+ },
+ {
+ "epoch": 26.097816593886463,
+ "grad_norm": 0.020843051373958588,
+ "learning_rate": 0.0006,
+ "loss": 4.294098854064941,
+ "step": 1879
+ },
+ {
+ "epoch": 26.111790393013102,
+ "grad_norm": 0.020654456689953804,
+ "learning_rate": 0.0006,
+ "loss": 4.267856597900391,
+ "step": 1880
+ },
+ {
+ "epoch": 26.125764192139737,
+ "grad_norm": 0.01860472559928894,
+ "learning_rate": 0.0006,
+ "loss": 4.284232139587402,
+ "step": 1881
+ },
+ {
+ "epoch": 26.139737991266376,
+ "grad_norm": 0.018719421699643135,
+ "learning_rate": 0.0006,
+ "loss": 4.277214527130127,
+ "step": 1882
+ },
+ {
+ "epoch": 26.153711790393015,
+ "grad_norm": 0.020903829485177994,
+ "learning_rate": 0.0006,
+ "loss": 4.254947662353516,
+ "step": 1883
+ },
+ {
+ "epoch": 26.16768558951965,
+ "grad_norm": 0.02020149491727352,
+ "learning_rate": 0.0006,
+ "loss": 4.2526936531066895,
+ "step": 1884
+ },
+ {
+ "epoch": 26.18165938864629,
+ "grad_norm": 0.0180392786860466,
+ "learning_rate": 0.0006,
+ "loss": 4.309922695159912,
+ "step": 1885
+ },
+ {
+ "epoch": 26.195633187772927,
+ "grad_norm": 0.018163125962018967,
+ "learning_rate": 0.0006,
+ "loss": 4.13860559463501,
+ "step": 1886
+ },
+ {
+ "epoch": 26.209606986899562,
+ "grad_norm": 0.01853291131556034,
+ "learning_rate": 0.0006,
+ "loss": 4.190926551818848,
+ "step": 1887
+ },
+ {
+ "epoch": 26.2235807860262,
+ "grad_norm": 0.01757677271962166,
+ "learning_rate": 0.0006,
+ "loss": 4.163999557495117,
+ "step": 1888
+ },
+ {
+ "epoch": 26.23755458515284,
+ "grad_norm": 0.01767992228269577,
+ "learning_rate": 0.0006,
+ "loss": 4.184430122375488,
+ "step": 1889
+ },
+ {
+ "epoch": 26.251528384279474,
+ "grad_norm": 0.018139973282814026,
+ "learning_rate": 0.0006,
+ "loss": 4.300766944885254,
+ "step": 1890
+ },
+ {
+ "epoch": 26.265502183406113,
+ "grad_norm": 0.020184461027383804,
+ "learning_rate": 0.0006,
+ "loss": 4.232537269592285,
+ "step": 1891
+ },
+ {
+ "epoch": 26.27947598253275,
+ "grad_norm": 0.02028856985270977,
+ "learning_rate": 0.0006,
+ "loss": 4.324883460998535,
+ "step": 1892
+ },
+ {
+ "epoch": 26.29344978165939,
+ "grad_norm": 0.017120616510510445,
+ "learning_rate": 0.0006,
+ "loss": 4.23493766784668,
+ "step": 1893
+ },
+ {
+ "epoch": 26.307423580786025,
+ "grad_norm": 0.016202978789806366,
+ "learning_rate": 0.0006,
+ "loss": 4.197602272033691,
+ "step": 1894
+ },
+ {
+ "epoch": 26.321397379912664,
+ "grad_norm": 0.017231421545147896,
+ "learning_rate": 0.0006,
+ "loss": 4.302001953125,
+ "step": 1895
+ },
+ {
+ "epoch": 26.335371179039303,
+ "grad_norm": 0.01725655607879162,
+ "learning_rate": 0.0006,
+ "loss": 4.356800079345703,
+ "step": 1896
+ },
+ {
+ "epoch": 26.349344978165938,
+ "grad_norm": 0.016657443717122078,
+ "learning_rate": 0.0006,
+ "loss": 4.266307830810547,
+ "step": 1897
+ },
+ {
+ "epoch": 26.363318777292577,
+ "grad_norm": 0.01616556942462921,
+ "learning_rate": 0.0006,
+ "loss": 4.307868003845215,
+ "step": 1898
+ },
+ {
+ "epoch": 26.377292576419215,
+ "grad_norm": 0.016621306538581848,
+ "learning_rate": 0.0006,
+ "loss": 4.172905921936035,
+ "step": 1899
+ },
+ {
+ "epoch": 26.39126637554585,
+ "grad_norm": 0.016243206337094307,
+ "learning_rate": 0.0006,
+ "loss": 4.162896156311035,
+ "step": 1900
+ },
+ {
+ "epoch": 26.40524017467249,
+ "grad_norm": 0.016663808375597,
+ "learning_rate": 0.0006,
+ "loss": 4.258232116699219,
+ "step": 1901
+ },
+ {
+ "epoch": 26.419213973799128,
+ "grad_norm": 0.01677478663623333,
+ "learning_rate": 0.0006,
+ "loss": 4.336090087890625,
+ "step": 1902
+ },
+ {
+ "epoch": 26.433187772925763,
+ "grad_norm": 0.01641303487122059,
+ "learning_rate": 0.0006,
+ "loss": 4.191476821899414,
+ "step": 1903
+ },
+ {
+ "epoch": 26.4471615720524,
+ "grad_norm": 0.016700323671102524,
+ "learning_rate": 0.0006,
+ "loss": 4.299427509307861,
+ "step": 1904
+ },
+ {
+ "epoch": 26.46113537117904,
+ "grad_norm": 0.017595035955309868,
+ "learning_rate": 0.0006,
+ "loss": 4.312828063964844,
+ "step": 1905
+ },
+ {
+ "epoch": 26.475109170305675,
+ "grad_norm": 0.020300284028053284,
+ "learning_rate": 0.0006,
+ "loss": 4.28331184387207,
+ "step": 1906
+ },
+ {
+ "epoch": 26.489082969432314,
+ "grad_norm": 0.02198321744799614,
+ "learning_rate": 0.0006,
+ "loss": 4.3280839920043945,
+ "step": 1907
+ },
+ {
+ "epoch": 26.503056768558952,
+ "grad_norm": 0.02003200724720955,
+ "learning_rate": 0.0006,
+ "loss": 4.325623989105225,
+ "step": 1908
+ },
+ {
+ "epoch": 26.51703056768559,
+ "grad_norm": 0.01870287023484707,
+ "learning_rate": 0.0006,
+ "loss": 4.251203536987305,
+ "step": 1909
+ },
+ {
+ "epoch": 26.531004366812226,
+ "grad_norm": 0.01835448481142521,
+ "learning_rate": 0.0006,
+ "loss": 4.335144996643066,
+ "step": 1910
+ },
+ {
+ "epoch": 26.544978165938865,
+ "grad_norm": 0.02046327292919159,
+ "learning_rate": 0.0006,
+ "loss": 4.238023281097412,
+ "step": 1911
+ },
+ {
+ "epoch": 26.558951965065503,
+ "grad_norm": 0.019771311432123184,
+ "learning_rate": 0.0006,
+ "loss": 4.353594779968262,
+ "step": 1912
+ },
+ {
+ "epoch": 26.57292576419214,
+ "grad_norm": 0.018090544268488884,
+ "learning_rate": 0.0006,
+ "loss": 4.304343223571777,
+ "step": 1913
+ },
+ {
+ "epoch": 26.586899563318777,
+ "grad_norm": 0.017489777877926826,
+ "learning_rate": 0.0006,
+ "loss": 4.204375267028809,
+ "step": 1914
+ },
+ {
+ "epoch": 26.600873362445416,
+ "grad_norm": 0.016785891726613045,
+ "learning_rate": 0.0006,
+ "loss": 4.286669731140137,
+ "step": 1915
+ },
+ {
+ "epoch": 26.61484716157205,
+ "grad_norm": 0.016382789239287376,
+ "learning_rate": 0.0006,
+ "loss": 4.347662448883057,
+ "step": 1916
+ },
+ {
+ "epoch": 26.62882096069869,
+ "grad_norm": 0.016581876203417778,
+ "learning_rate": 0.0006,
+ "loss": 4.2716193199157715,
+ "step": 1917
+ },
+ {
+ "epoch": 26.64279475982533,
+ "grad_norm": 0.015627184882760048,
+ "learning_rate": 0.0006,
+ "loss": 4.311489105224609,
+ "step": 1918
+ },
+ {
+ "epoch": 26.656768558951963,
+ "grad_norm": 0.015908483415842056,
+ "learning_rate": 0.0006,
+ "loss": 4.262099742889404,
+ "step": 1919
+ },
+ {
+ "epoch": 26.670742358078602,
+ "grad_norm": 0.015576236881315708,
+ "learning_rate": 0.0006,
+ "loss": 4.275871276855469,
+ "step": 1920
+ },
+ {
+ "epoch": 26.68471615720524,
+ "grad_norm": 0.014621064998209476,
+ "learning_rate": 0.0006,
+ "loss": 4.184260845184326,
+ "step": 1921
+ },
+ {
+ "epoch": 26.69868995633188,
+ "grad_norm": 0.014797811396420002,
+ "learning_rate": 0.0006,
+ "loss": 4.3412628173828125,
+ "step": 1922
+ },
+ {
+ "epoch": 26.712663755458514,
+ "grad_norm": 0.015610570088028908,
+ "learning_rate": 0.0006,
+ "loss": 4.181094646453857,
+ "step": 1923
+ },
+ {
+ "epoch": 26.726637554585153,
+ "grad_norm": 0.016048265621066093,
+ "learning_rate": 0.0006,
+ "loss": 4.365290641784668,
+ "step": 1924
+ },
+ {
+ "epoch": 26.74061135371179,
+ "grad_norm": 0.016604071483016014,
+ "learning_rate": 0.0006,
+ "loss": 4.2628912925720215,
+ "step": 1925
+ },
+ {
+ "epoch": 26.754585152838427,
+ "grad_norm": 0.017044221982359886,
+ "learning_rate": 0.0006,
+ "loss": 4.378512382507324,
+ "step": 1926
+ },
+ {
+ "epoch": 26.768558951965066,
+ "grad_norm": 0.01746384985744953,
+ "learning_rate": 0.0006,
+ "loss": 4.3888773918151855,
+ "step": 1927
+ },
+ {
+ "epoch": 26.782532751091704,
+ "grad_norm": 0.01572711206972599,
+ "learning_rate": 0.0006,
+ "loss": 4.189935684204102,
+ "step": 1928
+ },
+ {
+ "epoch": 26.79650655021834,
+ "grad_norm": 0.014976629987359047,
+ "learning_rate": 0.0006,
+ "loss": 4.280525207519531,
+ "step": 1929
+ },
+ {
+ "epoch": 26.810480349344978,
+ "grad_norm": 0.014804664999246597,
+ "learning_rate": 0.0006,
+ "loss": 4.270630836486816,
+ "step": 1930
+ },
+ {
+ "epoch": 26.824454148471617,
+ "grad_norm": 0.014641920104622841,
+ "learning_rate": 0.0006,
+ "loss": 4.319505214691162,
+ "step": 1931
+ },
+ {
+ "epoch": 26.83842794759825,
+ "grad_norm": 0.014928505755960941,
+ "learning_rate": 0.0006,
+ "loss": 4.180147647857666,
+ "step": 1932
+ },
+ {
+ "epoch": 26.85240174672489,
+ "grad_norm": 0.016130365431308746,
+ "learning_rate": 0.0006,
+ "loss": 4.260660648345947,
+ "step": 1933
+ },
+ {
+ "epoch": 26.86637554585153,
+ "grad_norm": 0.015642493963241577,
+ "learning_rate": 0.0006,
+ "loss": 4.304064750671387,
+ "step": 1934
+ },
+ {
+ "epoch": 26.880349344978168,
+ "grad_norm": 0.015988217666745186,
+ "learning_rate": 0.0006,
+ "loss": 4.2548017501831055,
+ "step": 1935
+ },
+ {
+ "epoch": 26.894323144104803,
+ "grad_norm": 0.016501398757100105,
+ "learning_rate": 0.0006,
+ "loss": 4.260058879852295,
+ "step": 1936
+ },
+ {
+ "epoch": 26.90829694323144,
+ "grad_norm": 0.01645040698349476,
+ "learning_rate": 0.0006,
+ "loss": 4.483081340789795,
+ "step": 1937
+ },
+ {
+ "epoch": 26.92227074235808,
+ "grad_norm": 0.016140863299369812,
+ "learning_rate": 0.0006,
+ "loss": 4.1776838302612305,
+ "step": 1938
+ },
+ {
+ "epoch": 26.936244541484715,
+ "grad_norm": 0.016681019216775894,
+ "learning_rate": 0.0006,
+ "loss": 4.224747657775879,
+ "step": 1939
+ },
+ {
+ "epoch": 26.950218340611354,
+ "grad_norm": 0.01650378853082657,
+ "learning_rate": 0.0006,
+ "loss": 4.305703163146973,
+ "step": 1940
+ },
+ {
+ "epoch": 26.964192139737992,
+ "grad_norm": 0.017051683738827705,
+ "learning_rate": 0.0006,
+ "loss": 4.285835266113281,
+ "step": 1941
+ },
+ {
+ "epoch": 26.978165938864628,
+ "grad_norm": 0.016894327476620674,
+ "learning_rate": 0.0006,
+ "loss": 4.318901538848877,
+ "step": 1942
+ },
+ {
+ "epoch": 26.992139737991266,
+ "grad_norm": 0.017313100397586823,
+ "learning_rate": 0.0006,
+ "loss": 4.281434535980225,
+ "step": 1943
+ },
+ {
+ "epoch": 27.0,
+ "grad_norm": 0.018450884148478508,
+ "learning_rate": 0.0006,
+ "loss": 4.335104942321777,
+ "step": 1944
+ },
+ {
+ "epoch": 27.0,
+ "eval_loss": 4.640687465667725,
+ "eval_runtime": 56.9624,
+ "eval_samples_per_second": 42.87,
+ "eval_steps_per_second": 1.352,
+ "step": 1944
+ },
+ {
+ "epoch": 27.01397379912664,
+ "grad_norm": 0.017755698412656784,
+ "learning_rate": 0.0006,
+ "loss": 4.278824329376221,
+ "step": 1945
+ },
+ {
+ "epoch": 27.027947598253274,
+ "grad_norm": 0.01724310778081417,
+ "learning_rate": 0.0006,
+ "loss": 4.311049461364746,
+ "step": 1946
+ },
+ {
+ "epoch": 27.041921397379912,
+ "grad_norm": 0.01690123789012432,
+ "learning_rate": 0.0006,
+ "loss": 4.261994361877441,
+ "step": 1947
+ },
+ {
+ "epoch": 27.05589519650655,
+ "grad_norm": 0.018406696617603302,
+ "learning_rate": 0.0006,
+ "loss": 4.1925153732299805,
+ "step": 1948
+ },
+ {
+ "epoch": 27.069868995633186,
+ "grad_norm": 0.019945867359638214,
+ "learning_rate": 0.0006,
+ "loss": 4.108537197113037,
+ "step": 1949
+ },
+ {
+ "epoch": 27.083842794759825,
+ "grad_norm": 0.020316628739237785,
+ "learning_rate": 0.0006,
+ "loss": 4.209043025970459,
+ "step": 1950
+ },
+ {
+ "epoch": 27.097816593886463,
+ "grad_norm": 0.019690370187163353,
+ "learning_rate": 0.0006,
+ "loss": 4.195789337158203,
+ "step": 1951
+ },
+ {
+ "epoch": 27.111790393013102,
+ "grad_norm": 0.01785232499241829,
+ "learning_rate": 0.0006,
+ "loss": 4.21466064453125,
+ "step": 1952
+ },
+ {
+ "epoch": 27.125764192139737,
+ "grad_norm": 0.0164767038077116,
+ "learning_rate": 0.0006,
+ "loss": 4.270247459411621,
+ "step": 1953
+ },
+ {
+ "epoch": 27.139737991266376,
+ "grad_norm": 0.018008455634117126,
+ "learning_rate": 0.0006,
+ "loss": 4.201042175292969,
+ "step": 1954
+ },
+ {
+ "epoch": 27.153711790393015,
+ "grad_norm": 0.01807340793311596,
+ "learning_rate": 0.0006,
+ "loss": 4.25289249420166,
+ "step": 1955
+ },
+ {
+ "epoch": 27.16768558951965,
+ "grad_norm": 0.016246909275650978,
+ "learning_rate": 0.0006,
+ "loss": 4.116283416748047,
+ "step": 1956
+ },
+ {
+ "epoch": 27.18165938864629,
+ "grad_norm": 0.017545776441693306,
+ "learning_rate": 0.0006,
+ "loss": 4.316999435424805,
+ "step": 1957
+ },
+ {
+ "epoch": 27.195633187772927,
+ "grad_norm": 0.017915375530719757,
+ "learning_rate": 0.0006,
+ "loss": 4.253859519958496,
+ "step": 1958
+ },
+ {
+ "epoch": 27.209606986899562,
+ "grad_norm": 0.018112223595380783,
+ "learning_rate": 0.0006,
+ "loss": 4.2541022300720215,
+ "step": 1959
+ },
+ {
+ "epoch": 27.2235807860262,
+ "grad_norm": 0.01846139505505562,
+ "learning_rate": 0.0006,
+ "loss": 4.20438289642334,
+ "step": 1960
+ },
+ {
+ "epoch": 27.23755458515284,
+ "grad_norm": 0.01943897269666195,
+ "learning_rate": 0.0006,
+ "loss": 4.2799577713012695,
+ "step": 1961
+ },
+ {
+ "epoch": 27.251528384279474,
+ "grad_norm": 0.018253348767757416,
+ "learning_rate": 0.0006,
+ "loss": 4.1720499992370605,
+ "step": 1962
+ },
+ {
+ "epoch": 27.265502183406113,
+ "grad_norm": 0.018292246386408806,
+ "learning_rate": 0.0006,
+ "loss": 4.28231143951416,
+ "step": 1963
+ },
+ {
+ "epoch": 27.27947598253275,
+ "grad_norm": 0.01943155936896801,
+ "learning_rate": 0.0006,
+ "loss": 4.314976692199707,
+ "step": 1964
+ },
+ {
+ "epoch": 27.29344978165939,
+ "grad_norm": 0.019064994528889656,
+ "learning_rate": 0.0006,
+ "loss": 4.25328254699707,
+ "step": 1965
+ },
+ {
+ "epoch": 27.307423580786025,
+ "grad_norm": 0.017423273995518684,
+ "learning_rate": 0.0006,
+ "loss": 4.195992469787598,
+ "step": 1966
+ },
+ {
+ "epoch": 27.321397379912664,
+ "grad_norm": 0.01677924580872059,
+ "learning_rate": 0.0006,
+ "loss": 4.260043144226074,
+ "step": 1967
+ },
+ {
+ "epoch": 27.335371179039303,
+ "grad_norm": 0.01720673404633999,
+ "learning_rate": 0.0006,
+ "loss": 4.222780704498291,
+ "step": 1968
+ },
+ {
+ "epoch": 27.349344978165938,
+ "grad_norm": 0.017300743609666824,
+ "learning_rate": 0.0006,
+ "loss": 4.109524250030518,
+ "step": 1969
+ },
+ {
+ "epoch": 27.363318777292577,
+ "grad_norm": 0.016065871343016624,
+ "learning_rate": 0.0006,
+ "loss": 4.36108922958374,
+ "step": 1970
+ },
+ {
+ "epoch": 27.377292576419215,
+ "grad_norm": 0.016626616939902306,
+ "learning_rate": 0.0006,
+ "loss": 4.220630645751953,
+ "step": 1971
+ },
+ {
+ "epoch": 27.39126637554585,
+ "grad_norm": 0.016482602804899216,
+ "learning_rate": 0.0006,
+ "loss": 4.344797134399414,
+ "step": 1972
+ },
+ {
+ "epoch": 27.40524017467249,
+ "grad_norm": 0.01727953553199768,
+ "learning_rate": 0.0006,
+ "loss": 4.296645164489746,
+ "step": 1973
+ },
+ {
+ "epoch": 27.419213973799128,
+ "grad_norm": 0.01522731315344572,
+ "learning_rate": 0.0006,
+ "loss": 4.223751068115234,
+ "step": 1974
+ },
+ {
+ "epoch": 27.433187772925763,
+ "grad_norm": 0.01581776700913906,
+ "learning_rate": 0.0006,
+ "loss": 4.351379871368408,
+ "step": 1975
+ },
+ {
+ "epoch": 27.4471615720524,
+ "grad_norm": 0.016394605860114098,
+ "learning_rate": 0.0006,
+ "loss": 4.2412919998168945,
+ "step": 1976
+ },
+ {
+ "epoch": 27.46113537117904,
+ "grad_norm": 0.01751169003546238,
+ "learning_rate": 0.0006,
+ "loss": 4.257606506347656,
+ "step": 1977
+ },
+ {
+ "epoch": 27.475109170305675,
+ "grad_norm": 0.017452310770750046,
+ "learning_rate": 0.0006,
+ "loss": 4.380134582519531,
+ "step": 1978
+ },
+ {
+ "epoch": 27.489082969432314,
+ "grad_norm": 0.017741898074746132,
+ "learning_rate": 0.0006,
+ "loss": 4.279632568359375,
+ "step": 1979
+ },
+ {
+ "epoch": 27.503056768558952,
+ "grad_norm": 0.017680590972304344,
+ "learning_rate": 0.0006,
+ "loss": 4.273406028747559,
+ "step": 1980
+ },
+ {
+ "epoch": 27.51703056768559,
+ "grad_norm": 0.01864718459546566,
+ "learning_rate": 0.0006,
+ "loss": 4.278583526611328,
+ "step": 1981
+ },
+ {
+ "epoch": 27.531004366812226,
+ "grad_norm": 0.019558541476726532,
+ "learning_rate": 0.0006,
+ "loss": 4.211919784545898,
+ "step": 1982
+ },
+ {
+ "epoch": 27.544978165938865,
+ "grad_norm": 0.019467400386929512,
+ "learning_rate": 0.0006,
+ "loss": 4.276822566986084,
+ "step": 1983
+ },
+ {
+ "epoch": 27.558951965065503,
+ "grad_norm": 0.01756172813475132,
+ "learning_rate": 0.0006,
+ "loss": 4.165379047393799,
+ "step": 1984
+ },
+ {
+ "epoch": 27.57292576419214,
+ "grad_norm": 0.017458263784646988,
+ "learning_rate": 0.0006,
+ "loss": 4.2992401123046875,
+ "step": 1985
+ },
+ {
+ "epoch": 27.586899563318777,
+ "grad_norm": 0.018089665099978447,
+ "learning_rate": 0.0006,
+ "loss": 4.243378639221191,
+ "step": 1986
+ },
+ {
+ "epoch": 27.600873362445416,
+ "grad_norm": 0.019535524770617485,
+ "learning_rate": 0.0006,
+ "loss": 4.321477890014648,
+ "step": 1987
+ },
+ {
+ "epoch": 27.61484716157205,
+ "grad_norm": 0.01997970975935459,
+ "learning_rate": 0.0006,
+ "loss": 4.123082637786865,
+ "step": 1988
+ },
+ {
+ "epoch": 27.62882096069869,
+ "grad_norm": 0.01913067139685154,
+ "learning_rate": 0.0006,
+ "loss": 4.289896488189697,
+ "step": 1989
+ },
+ {
+ "epoch": 27.64279475982533,
+ "grad_norm": 0.0169806070625782,
+ "learning_rate": 0.0006,
+ "loss": 4.247004985809326,
+ "step": 1990
+ },
+ {
+ "epoch": 27.656768558951963,
+ "grad_norm": 0.017315007746219635,
+ "learning_rate": 0.0006,
+ "loss": 4.2095441818237305,
+ "step": 1991
+ },
+ {
+ "epoch": 27.670742358078602,
+ "grad_norm": 0.015098759904503822,
+ "learning_rate": 0.0006,
+ "loss": 4.145560264587402,
+ "step": 1992
+ },
+ {
+ "epoch": 27.68471615720524,
+ "grad_norm": 0.016240514814853668,
+ "learning_rate": 0.0006,
+ "loss": 4.254927635192871,
+ "step": 1993
+ },
+ {
+ "epoch": 27.69868995633188,
+ "grad_norm": 0.01621108129620552,
+ "learning_rate": 0.0006,
+ "loss": 4.219725608825684,
+ "step": 1994
+ },
+ {
+ "epoch": 27.712663755458514,
+ "grad_norm": 0.015191009268164635,
+ "learning_rate": 0.0006,
+ "loss": 4.194345474243164,
+ "step": 1995
+ },
+ {
+ "epoch": 27.726637554585153,
+ "grad_norm": 0.014387983828783035,
+ "learning_rate": 0.0006,
+ "loss": 4.211467742919922,
+ "step": 1996
+ },
+ {
+ "epoch": 27.74061135371179,
+ "grad_norm": 0.015460561029613018,
+ "learning_rate": 0.0006,
+ "loss": 4.358700752258301,
+ "step": 1997
+ },
+ {
+ "epoch": 27.754585152838427,
+ "grad_norm": 0.015100076794624329,
+ "learning_rate": 0.0006,
+ "loss": 4.171223163604736,
+ "step": 1998
+ },
+ {
+ "epoch": 27.768558951965066,
+ "grad_norm": 0.015607891604304314,
+ "learning_rate": 0.0006,
+ "loss": 4.217874526977539,
+ "step": 1999
+ },
+ {
+ "epoch": 27.782532751091704,
+ "grad_norm": 0.015650689601898193,
+ "learning_rate": 0.0006,
+ "loss": 4.280780792236328,
+ "step": 2000
+ },
+ {
+ "epoch": 27.79650655021834,
+ "grad_norm": 0.01580994389951229,
+ "learning_rate": 0.0006,
+ "loss": 4.159212112426758,
+ "step": 2001
+ },
+ {
+ "epoch": 27.810480349344978,
+ "grad_norm": 0.016607413068413734,
+ "learning_rate": 0.0006,
+ "loss": 4.238853454589844,
+ "step": 2002
+ },
+ {
+ "epoch": 27.824454148471617,
+ "grad_norm": 0.017383860424160957,
+ "learning_rate": 0.0006,
+ "loss": 4.233944892883301,
+ "step": 2003
+ },
+ {
+ "epoch": 27.83842794759825,
+ "grad_norm": 0.014966659247875214,
+ "learning_rate": 0.0006,
+ "loss": 4.167392730712891,
+ "step": 2004
+ },
+ {
+ "epoch": 27.85240174672489,
+ "grad_norm": 0.014589878730475903,
+ "learning_rate": 0.0006,
+ "loss": 4.2390899658203125,
+ "step": 2005
+ },
+ {
+ "epoch": 27.86637554585153,
+ "grad_norm": 0.017179450020194054,
+ "learning_rate": 0.0006,
+ "loss": 4.365785598754883,
+ "step": 2006
+ },
+ {
+ "epoch": 27.880349344978168,
+ "grad_norm": 0.01893901452422142,
+ "learning_rate": 0.0006,
+ "loss": 4.289430141448975,
+ "step": 2007
+ },
+ {
+ "epoch": 27.894323144104803,
+ "grad_norm": 0.018619712442159653,
+ "learning_rate": 0.0006,
+ "loss": 4.310665130615234,
+ "step": 2008
+ },
+ {
+ "epoch": 27.90829694323144,
+ "grad_norm": 0.019180385395884514,
+ "learning_rate": 0.0006,
+ "loss": 4.27454137802124,
+ "step": 2009
+ },
+ {
+ "epoch": 27.92227074235808,
+ "grad_norm": 0.017811523750424385,
+ "learning_rate": 0.0006,
+ "loss": 4.312562465667725,
+ "step": 2010
+ },
+ {
+ "epoch": 27.936244541484715,
+ "grad_norm": 0.016583573073148727,
+ "learning_rate": 0.0006,
+ "loss": 4.202253341674805,
+ "step": 2011
+ },
+ {
+ "epoch": 27.950218340611354,
+ "grad_norm": 0.01658615842461586,
+ "learning_rate": 0.0006,
+ "loss": 4.183433532714844,
+ "step": 2012
+ },
+ {
+ "epoch": 27.964192139737992,
+ "grad_norm": 0.015327401459217072,
+ "learning_rate": 0.0006,
+ "loss": 4.2125139236450195,
+ "step": 2013
+ },
+ {
+ "epoch": 27.978165938864628,
+ "grad_norm": 0.015004601329565048,
+ "learning_rate": 0.0006,
+ "loss": 4.257392883300781,
+ "step": 2014
+ },
+ {
+ "epoch": 27.992139737991266,
+ "grad_norm": 0.015259161591529846,
+ "learning_rate": 0.0006,
+ "loss": 4.241818428039551,
+ "step": 2015
+ },
+ {
+ "epoch": 28.0,
+ "grad_norm": 0.016808776184916496,
+ "learning_rate": 0.0006,
+ "loss": 4.146241188049316,
+ "step": 2016
+ },
+ {
+ "epoch": 28.0,
+ "eval_loss": 4.663090705871582,
+ "eval_runtime": 56.4559,
+ "eval_samples_per_second": 43.255,
+ "eval_steps_per_second": 1.364,
+ "step": 2016
+ },
+ {
+ "epoch": 28.01397379912664,
+ "grad_norm": 0.015784382820129395,
+ "learning_rate": 0.0006,
+ "loss": 4.044191360473633,
+ "step": 2017
+ },
+ {
+ "epoch": 28.027947598253274,
+ "grad_norm": 0.015636710450053215,
+ "learning_rate": 0.0006,
+ "loss": 4.128754615783691,
+ "step": 2018
+ },
+ {
+ "epoch": 28.041921397379912,
+ "grad_norm": 0.015364975668489933,
+ "learning_rate": 0.0006,
+ "loss": 4.191835403442383,
+ "step": 2019
+ },
+ {
+ "epoch": 28.05589519650655,
+ "grad_norm": 0.015015700832009315,
+ "learning_rate": 0.0006,
+ "loss": 4.120022773742676,
+ "step": 2020
+ },
+ {
+ "epoch": 28.069868995633186,
+ "grad_norm": 0.016666799783706665,
+ "learning_rate": 0.0006,
+ "loss": 4.33906364440918,
+ "step": 2021
+ },
+ {
+ "epoch": 28.083842794759825,
+ "grad_norm": 0.01780104823410511,
+ "learning_rate": 0.0006,
+ "loss": 4.279458999633789,
+ "step": 2022
+ },
+ {
+ "epoch": 28.097816593886463,
+ "grad_norm": 0.01870882883667946,
+ "learning_rate": 0.0006,
+ "loss": 4.34728479385376,
+ "step": 2023
+ },
+ {
+ "epoch": 28.111790393013102,
+ "grad_norm": 0.018406381830573082,
+ "learning_rate": 0.0006,
+ "loss": 4.211311340332031,
+ "step": 2024
+ },
+ {
+ "epoch": 28.125764192139737,
+ "grad_norm": 0.017833448946475983,
+ "learning_rate": 0.0006,
+ "loss": 4.223254203796387,
+ "step": 2025
+ },
+ {
+ "epoch": 28.139737991266376,
+ "grad_norm": 0.017552688717842102,
+ "learning_rate": 0.0006,
+ "loss": 4.208823204040527,
+ "step": 2026
+ },
+ {
+ "epoch": 28.153711790393015,
+ "grad_norm": 0.018500229343771935,
+ "learning_rate": 0.0006,
+ "loss": 4.198025226593018,
+ "step": 2027
+ },
+ {
+ "epoch": 28.16768558951965,
+ "grad_norm": 0.01782156340777874,
+ "learning_rate": 0.0006,
+ "loss": 4.179978370666504,
+ "step": 2028
+ },
+ {
+ "epoch": 28.18165938864629,
+ "grad_norm": 0.019580967724323273,
+ "learning_rate": 0.0006,
+ "loss": 4.291114807128906,
+ "step": 2029
+ },
+ {
+ "epoch": 28.195633187772927,
+ "grad_norm": 0.020804036408662796,
+ "learning_rate": 0.0006,
+ "loss": 4.220607280731201,
+ "step": 2030
+ },
+ {
+ "epoch": 28.209606986899562,
+ "grad_norm": 0.019595693796873093,
+ "learning_rate": 0.0006,
+ "loss": 4.2905731201171875,
+ "step": 2031
+ },
+ {
+ "epoch": 28.2235807860262,
+ "grad_norm": 0.019247086718678474,
+ "learning_rate": 0.0006,
+ "loss": 4.2467360496521,
+ "step": 2032
+ },
+ {
+ "epoch": 28.23755458515284,
+ "grad_norm": 0.01958036608994007,
+ "learning_rate": 0.0006,
+ "loss": 4.286137580871582,
+ "step": 2033
+ },
+ {
+ "epoch": 28.251528384279474,
+ "grad_norm": 0.021354753524065018,
+ "learning_rate": 0.0006,
+ "loss": 4.234755039215088,
+ "step": 2034
+ },
+ {
+ "epoch": 28.265502183406113,
+ "grad_norm": 0.02206375077366829,
+ "learning_rate": 0.0006,
+ "loss": 4.149641036987305,
+ "step": 2035
+ },
+ {
+ "epoch": 28.27947598253275,
+ "grad_norm": 0.02201310358941555,
+ "learning_rate": 0.0006,
+ "loss": 4.1928629875183105,
+ "step": 2036
+ },
+ {
+ "epoch": 28.29344978165939,
+ "grad_norm": 0.01961950585246086,
+ "learning_rate": 0.0006,
+ "loss": 4.211404323577881,
+ "step": 2037
+ },
+ {
+ "epoch": 28.307423580786025,
+ "grad_norm": 0.019211795181035995,
+ "learning_rate": 0.0006,
+ "loss": 4.200976848602295,
+ "step": 2038
+ },
+ {
+ "epoch": 28.321397379912664,
+ "grad_norm": 0.01779370754957199,
+ "learning_rate": 0.0006,
+ "loss": 4.249148845672607,
+ "step": 2039
+ },
+ {
+ "epoch": 28.335371179039303,
+ "grad_norm": 0.018635908141732216,
+ "learning_rate": 0.0006,
+ "loss": 4.11806058883667,
+ "step": 2040
+ },
+ {
+ "epoch": 28.349344978165938,
+ "grad_norm": 0.019792694598436356,
+ "learning_rate": 0.0006,
+ "loss": 4.335224151611328,
+ "step": 2041
+ },
+ {
+ "epoch": 28.363318777292577,
+ "grad_norm": 0.021731717512011528,
+ "learning_rate": 0.0006,
+ "loss": 4.302778244018555,
+ "step": 2042
+ },
+ {
+ "epoch": 28.377292576419215,
+ "grad_norm": 0.020823447033762932,
+ "learning_rate": 0.0006,
+ "loss": 4.260610580444336,
+ "step": 2043
+ },
+ {
+ "epoch": 28.39126637554585,
+ "grad_norm": 0.019385412335395813,
+ "learning_rate": 0.0006,
+ "loss": 4.2247724533081055,
+ "step": 2044
+ },
+ {
+ "epoch": 28.40524017467249,
+ "grad_norm": 0.019539108499884605,
+ "learning_rate": 0.0006,
+ "loss": 4.236655235290527,
+ "step": 2045
+ },
+ {
+ "epoch": 28.419213973799128,
+ "grad_norm": 0.01942690648138523,
+ "learning_rate": 0.0006,
+ "loss": 4.251850605010986,
+ "step": 2046
+ },
+ {
+ "epoch": 28.433187772925763,
+ "grad_norm": 0.019667886197566986,
+ "learning_rate": 0.0006,
+ "loss": 4.222312927246094,
+ "step": 2047
+ },
+ {
+ "epoch": 28.4471615720524,
+ "grad_norm": 0.02029012329876423,
+ "learning_rate": 0.0006,
+ "loss": 4.246252059936523,
+ "step": 2048
+ },
+ {
+ "epoch": 28.46113537117904,
+ "grad_norm": 0.01784469373524189,
+ "learning_rate": 0.0006,
+ "loss": 4.192128658294678,
+ "step": 2049
+ },
+ {
+ "epoch": 28.475109170305675,
+ "grad_norm": 0.0160287544131279,
+ "learning_rate": 0.0006,
+ "loss": 4.208607196807861,
+ "step": 2050
+ },
+ {
+ "epoch": 28.489082969432314,
+ "grad_norm": 0.015348346903920174,
+ "learning_rate": 0.0006,
+ "loss": 4.236598014831543,
+ "step": 2051
+ },
+ {
+ "epoch": 28.503056768558952,
+ "grad_norm": 0.015198206529021263,
+ "learning_rate": 0.0006,
+ "loss": 4.235552787780762,
+ "step": 2052
+ },
+ {
+ "epoch": 28.51703056768559,
+ "grad_norm": 0.01523754745721817,
+ "learning_rate": 0.0006,
+ "loss": 4.1941728591918945,
+ "step": 2053
+ },
+ {
+ "epoch": 28.531004366812226,
+ "grad_norm": 0.015559614636003971,
+ "learning_rate": 0.0006,
+ "loss": 4.294757843017578,
+ "step": 2054
+ },
+ {
+ "epoch": 28.544978165938865,
+ "grad_norm": 0.01568358950316906,
+ "learning_rate": 0.0006,
+ "loss": 4.213065147399902,
+ "step": 2055
+ },
+ {
+ "epoch": 28.558951965065503,
+ "grad_norm": 0.015336516313254833,
+ "learning_rate": 0.0006,
+ "loss": 4.240293979644775,
+ "step": 2056
+ },
+ {
+ "epoch": 28.57292576419214,
+ "grad_norm": 0.015041496604681015,
+ "learning_rate": 0.0006,
+ "loss": 4.273397445678711,
+ "step": 2057
+ },
+ {
+ "epoch": 28.586899563318777,
+ "grad_norm": 0.01610128954052925,
+ "learning_rate": 0.0006,
+ "loss": 4.125369071960449,
+ "step": 2058
+ },
+ {
+ "epoch": 28.600873362445416,
+ "grad_norm": 0.015974976122379303,
+ "learning_rate": 0.0006,
+ "loss": 4.133459091186523,
+ "step": 2059
+ },
+ {
+ "epoch": 28.61484716157205,
+ "grad_norm": 0.015936799347400665,
+ "learning_rate": 0.0006,
+ "loss": 4.306667327880859,
+ "step": 2060
+ },
+ {
+ "epoch": 28.62882096069869,
+ "grad_norm": 0.01795247197151184,
+ "learning_rate": 0.0006,
+ "loss": 4.205156326293945,
+ "step": 2061
+ },
+ {
+ "epoch": 28.64279475982533,
+ "grad_norm": 0.019639814272522926,
+ "learning_rate": 0.0006,
+ "loss": 4.277614593505859,
+ "step": 2062
+ },
+ {
+ "epoch": 28.656768558951963,
+ "grad_norm": 0.01722300611436367,
+ "learning_rate": 0.0006,
+ "loss": 4.191032886505127,
+ "step": 2063
+ },
+ {
+ "epoch": 28.670742358078602,
+ "grad_norm": 0.01612633652985096,
+ "learning_rate": 0.0006,
+ "loss": 4.20717191696167,
+ "step": 2064
+ },
+ {
+ "epoch": 28.68471615720524,
+ "grad_norm": 0.01603279449045658,
+ "learning_rate": 0.0006,
+ "loss": 4.228884220123291,
+ "step": 2065
+ },
+ {
+ "epoch": 28.69868995633188,
+ "grad_norm": 0.016405927017331123,
+ "learning_rate": 0.0006,
+ "loss": 4.201292991638184,
+ "step": 2066
+ },
+ {
+ "epoch": 28.712663755458514,
+ "grad_norm": 0.016644183546304703,
+ "learning_rate": 0.0006,
+ "loss": 4.239282131195068,
+ "step": 2067
+ },
+ {
+ "epoch": 28.726637554585153,
+ "grad_norm": 0.0175587497651577,
+ "learning_rate": 0.0006,
+ "loss": 4.289038181304932,
+ "step": 2068
+ },
+ {
+ "epoch": 28.74061135371179,
+ "grad_norm": 0.018223397433757782,
+ "learning_rate": 0.0006,
+ "loss": 4.2266998291015625,
+ "step": 2069
+ },
+ {
+ "epoch": 28.754585152838427,
+ "grad_norm": 0.018680868670344353,
+ "learning_rate": 0.0006,
+ "loss": 4.348849296569824,
+ "step": 2070
+ },
+ {
+ "epoch": 28.768558951965066,
+ "grad_norm": 0.016726728528738022,
+ "learning_rate": 0.0006,
+ "loss": 4.292080879211426,
+ "step": 2071
+ },
+ {
+ "epoch": 28.782532751091704,
+ "grad_norm": 0.017386795952916145,
+ "learning_rate": 0.0006,
+ "loss": 4.208870887756348,
+ "step": 2072
+ },
+ {
+ "epoch": 28.79650655021834,
+ "grad_norm": 0.01865178905427456,
+ "learning_rate": 0.0006,
+ "loss": 4.054150581359863,
+ "step": 2073
+ },
+ {
+ "epoch": 28.810480349344978,
+ "grad_norm": 0.01776815392076969,
+ "learning_rate": 0.0006,
+ "loss": 4.202404499053955,
+ "step": 2074
+ },
+ {
+ "epoch": 28.824454148471617,
+ "grad_norm": 0.01643497310578823,
+ "learning_rate": 0.0006,
+ "loss": 4.131855010986328,
+ "step": 2075
+ },
+ {
+ "epoch": 28.83842794759825,
+ "grad_norm": 0.01574273779988289,
+ "learning_rate": 0.0006,
+ "loss": 4.296788692474365,
+ "step": 2076
+ },
+ {
+ "epoch": 28.85240174672489,
+ "grad_norm": 0.015613908879458904,
+ "learning_rate": 0.0006,
+ "loss": 4.210949897766113,
+ "step": 2077
+ },
+ {
+ "epoch": 28.86637554585153,
+ "grad_norm": 0.015543212182819843,
+ "learning_rate": 0.0006,
+ "loss": 4.218136310577393,
+ "step": 2078
+ },
+ {
+ "epoch": 28.880349344978168,
+ "grad_norm": 0.014681251719594002,
+ "learning_rate": 0.0006,
+ "loss": 4.19401741027832,
+ "step": 2079
+ },
+ {
+ "epoch": 28.894323144104803,
+ "grad_norm": 0.014393828809261322,
+ "learning_rate": 0.0006,
+ "loss": 4.198468208312988,
+ "step": 2080
+ },
+ {
+ "epoch": 28.90829694323144,
+ "grad_norm": 0.01580170728266239,
+ "learning_rate": 0.0006,
+ "loss": 4.224969863891602,
+ "step": 2081
+ },
+ {
+ "epoch": 28.92227074235808,
+ "grad_norm": 0.01484165620058775,
+ "learning_rate": 0.0006,
+ "loss": 4.12236213684082,
+ "step": 2082
+ },
+ {
+ "epoch": 28.936244541484715,
+ "grad_norm": 0.01497623696923256,
+ "learning_rate": 0.0006,
+ "loss": 4.24141788482666,
+ "step": 2083
+ },
+ {
+ "epoch": 28.950218340611354,
+ "grad_norm": 0.015080046840012074,
+ "learning_rate": 0.0006,
+ "loss": 4.284151077270508,
+ "step": 2084
+ },
+ {
+ "epoch": 28.964192139737992,
+ "grad_norm": 0.016207212582230568,
+ "learning_rate": 0.0006,
+ "loss": 4.298693656921387,
+ "step": 2085
+ },
+ {
+ "epoch": 28.978165938864628,
+ "grad_norm": 0.015474417246878147,
+ "learning_rate": 0.0006,
+ "loss": 4.2412261962890625,
+ "step": 2086
+ },
+ {
+ "epoch": 28.992139737991266,
+ "grad_norm": 0.01515351701527834,
+ "learning_rate": 0.0006,
+ "loss": 4.3054423332214355,
+ "step": 2087
+ },
+ {
+ "epoch": 29.0,
+ "grad_norm": 0.016411345452070236,
+ "learning_rate": 0.0006,
+ "loss": 4.104830741882324,
+ "step": 2088
+ },
+ {
+ "epoch": 29.0,
+ "eval_loss": 4.661567687988281,
+ "eval_runtime": 57.1244,
+ "eval_samples_per_second": 42.749,
+ "eval_steps_per_second": 1.348,
+ "step": 2088
+ },
+ {
+ "epoch": 29.01397379912664,
+ "grad_norm": 0.01665342226624489,
+ "learning_rate": 0.0006,
+ "loss": 4.311391830444336,
+ "step": 2089
+ },
+ {
+ "epoch": 29.027947598253274,
+ "grad_norm": 0.018053214997053146,
+ "learning_rate": 0.0006,
+ "loss": 4.135597229003906,
+ "step": 2090
+ },
+ {
+ "epoch": 29.041921397379912,
+ "grad_norm": 0.018213748931884766,
+ "learning_rate": 0.0006,
+ "loss": 4.303299903869629,
+ "step": 2091
+ },
+ {
+ "epoch": 29.05589519650655,
+ "grad_norm": 0.018460242077708244,
+ "learning_rate": 0.0006,
+ "loss": 4.20850944519043,
+ "step": 2092
+ },
+ {
+ "epoch": 29.069868995633186,
+ "grad_norm": 0.018365809693932533,
+ "learning_rate": 0.0006,
+ "loss": 4.275847434997559,
+ "step": 2093
+ },
+ {
+ "epoch": 29.083842794759825,
+ "grad_norm": 0.018346186727285385,
+ "learning_rate": 0.0006,
+ "loss": 4.202091217041016,
+ "step": 2094
+ },
+ {
+ "epoch": 29.097816593886463,
+ "grad_norm": 0.017920587211847305,
+ "learning_rate": 0.0006,
+ "loss": 4.182092666625977,
+ "step": 2095
+ },
+ {
+ "epoch": 29.111790393013102,
+ "grad_norm": 0.01812003180384636,
+ "learning_rate": 0.0006,
+ "loss": 4.259721755981445,
+ "step": 2096
+ },
+ {
+ "epoch": 29.125764192139737,
+ "grad_norm": 0.01700986735522747,
+ "learning_rate": 0.0006,
+ "loss": 4.067573070526123,
+ "step": 2097
+ },
+ {
+ "epoch": 29.139737991266376,
+ "grad_norm": 0.018069760873913765,
+ "learning_rate": 0.0006,
+ "loss": 4.234400749206543,
+ "step": 2098
+ },
+ {
+ "epoch": 29.153711790393015,
+ "grad_norm": 0.01905428245663643,
+ "learning_rate": 0.0006,
+ "loss": 4.16609001159668,
+ "step": 2099
+ },
+ {
+ "epoch": 29.16768558951965,
+ "grad_norm": 0.018711242824792862,
+ "learning_rate": 0.0006,
+ "loss": 4.249088764190674,
+ "step": 2100
+ },
+ {
+ "epoch": 29.18165938864629,
+ "grad_norm": 0.018721306696534157,
+ "learning_rate": 0.0006,
+ "loss": 4.14540958404541,
+ "step": 2101
+ },
+ {
+ "epoch": 29.195633187772927,
+ "grad_norm": 0.020139874890446663,
+ "learning_rate": 0.0006,
+ "loss": 4.049248695373535,
+ "step": 2102
+ },
+ {
+ "epoch": 29.209606986899562,
+ "grad_norm": 0.020849574357271194,
+ "learning_rate": 0.0006,
+ "loss": 4.235224723815918,
+ "step": 2103
+ },
+ {
+ "epoch": 29.2235807860262,
+ "grad_norm": 0.018693549558520317,
+ "learning_rate": 0.0006,
+ "loss": 4.232961177825928,
+ "step": 2104
+ },
+ {
+ "epoch": 29.23755458515284,
+ "grad_norm": 0.017889857292175293,
+ "learning_rate": 0.0006,
+ "loss": 4.171422004699707,
+ "step": 2105
+ },
+ {
+ "epoch": 29.251528384279474,
+ "grad_norm": 0.016690224409103394,
+ "learning_rate": 0.0006,
+ "loss": 4.244717597961426,
+ "step": 2106
+ },
+ {
+ "epoch": 29.265502183406113,
+ "grad_norm": 0.01708616502583027,
+ "learning_rate": 0.0006,
+ "loss": 4.125068664550781,
+ "step": 2107
+ },
+ {
+ "epoch": 29.27947598253275,
+ "grad_norm": 0.017928237095475197,
+ "learning_rate": 0.0006,
+ "loss": 4.208292007446289,
+ "step": 2108
+ },
+ {
+ "epoch": 29.29344978165939,
+ "grad_norm": 0.018250394612550735,
+ "learning_rate": 0.0006,
+ "loss": 4.118630886077881,
+ "step": 2109
+ },
+ {
+ "epoch": 29.307423580786025,
+ "grad_norm": 0.020160246640443802,
+ "learning_rate": 0.0006,
+ "loss": 4.206025123596191,
+ "step": 2110
+ },
+ {
+ "epoch": 29.321397379912664,
+ "grad_norm": 0.020045453682541847,
+ "learning_rate": 0.0006,
+ "loss": 4.2758684158325195,
+ "step": 2111
+ },
+ {
+ "epoch": 29.335371179039303,
+ "grad_norm": 0.019556907936930656,
+ "learning_rate": 0.0006,
+ "loss": 4.188272953033447,
+ "step": 2112
+ },
+ {
+ "epoch": 29.349344978165938,
+ "grad_norm": 0.019185200333595276,
+ "learning_rate": 0.0006,
+ "loss": 4.17451286315918,
+ "step": 2113
+ },
+ {
+ "epoch": 29.363318777292577,
+ "grad_norm": 0.01816629432141781,
+ "learning_rate": 0.0006,
+ "loss": 4.165246963500977,
+ "step": 2114
+ },
+ {
+ "epoch": 29.377292576419215,
+ "grad_norm": 0.01865146867930889,
+ "learning_rate": 0.0006,
+ "loss": 4.148184776306152,
+ "step": 2115
+ },
+ {
+ "epoch": 29.39126637554585,
+ "grad_norm": 0.018338464200496674,
+ "learning_rate": 0.0006,
+ "loss": 4.288936614990234,
+ "step": 2116
+ },
+ {
+ "epoch": 29.40524017467249,
+ "grad_norm": 0.018986187875270844,
+ "learning_rate": 0.0006,
+ "loss": 4.183379650115967,
+ "step": 2117
+ },
+ {
+ "epoch": 29.419213973799128,
+ "grad_norm": 0.018409637734293938,
+ "learning_rate": 0.0006,
+ "loss": 4.057574272155762,
+ "step": 2118
+ },
+ {
+ "epoch": 29.433187772925763,
+ "grad_norm": 0.016664542257785797,
+ "learning_rate": 0.0006,
+ "loss": 4.18517541885376,
+ "step": 2119
+ },
+ {
+ "epoch": 29.4471615720524,
+ "grad_norm": 0.016486089676618576,
+ "learning_rate": 0.0006,
+ "loss": 4.179367542266846,
+ "step": 2120
+ },
+ {
+ "epoch": 29.46113537117904,
+ "grad_norm": 0.01787523552775383,
+ "learning_rate": 0.0006,
+ "loss": 4.184126377105713,
+ "step": 2121
+ },
+ {
+ "epoch": 29.475109170305675,
+ "grad_norm": 0.016574613749980927,
+ "learning_rate": 0.0006,
+ "loss": 4.290170192718506,
+ "step": 2122
+ },
+ {
+ "epoch": 29.489082969432314,
+ "grad_norm": 0.01643376611173153,
+ "learning_rate": 0.0006,
+ "loss": 4.052913665771484,
+ "step": 2123
+ },
+ {
+ "epoch": 29.503056768558952,
+ "grad_norm": 0.016503969207406044,
+ "learning_rate": 0.0006,
+ "loss": 4.335184097290039,
+ "step": 2124
+ },
+ {
+ "epoch": 29.51703056768559,
+ "grad_norm": 0.017596479505300522,
+ "learning_rate": 0.0006,
+ "loss": 4.165539741516113,
+ "step": 2125
+ },
+ {
+ "epoch": 29.531004366812226,
+ "grad_norm": 0.017468582838773727,
+ "learning_rate": 0.0006,
+ "loss": 4.315964698791504,
+ "step": 2126
+ },
+ {
+ "epoch": 29.544978165938865,
+ "grad_norm": 0.01724618673324585,
+ "learning_rate": 0.0006,
+ "loss": 4.243681907653809,
+ "step": 2127
+ },
+ {
+ "epoch": 29.558951965065503,
+ "grad_norm": 0.018384616822004318,
+ "learning_rate": 0.0006,
+ "loss": 4.112447261810303,
+ "step": 2128
+ },
+ {
+ "epoch": 29.57292576419214,
+ "grad_norm": 0.01824074238538742,
+ "learning_rate": 0.0006,
+ "loss": 4.236065864562988,
+ "step": 2129
+ },
+ {
+ "epoch": 29.586899563318777,
+ "grad_norm": 0.01625971868634224,
+ "learning_rate": 0.0006,
+ "loss": 4.18143892288208,
+ "step": 2130
+ },
+ {
+ "epoch": 29.600873362445416,
+ "grad_norm": 0.015537494793534279,
+ "learning_rate": 0.0006,
+ "loss": 4.225147247314453,
+ "step": 2131
+ },
+ {
+ "epoch": 29.61484716157205,
+ "grad_norm": 0.015818974003195763,
+ "learning_rate": 0.0006,
+ "loss": 4.3100175857543945,
+ "step": 2132
+ },
+ {
+ "epoch": 29.62882096069869,
+ "grad_norm": 0.015909671783447266,
+ "learning_rate": 0.0006,
+ "loss": 4.169775009155273,
+ "step": 2133
+ },
+ {
+ "epoch": 29.64279475982533,
+ "grad_norm": 0.01769658550620079,
+ "learning_rate": 0.0006,
+ "loss": 4.278904914855957,
+ "step": 2134
+ },
+ {
+ "epoch": 29.656768558951963,
+ "grad_norm": 0.015641704201698303,
+ "learning_rate": 0.0006,
+ "loss": 4.2426228523254395,
+ "step": 2135
+ },
+ {
+ "epoch": 29.670742358078602,
+ "grad_norm": 0.01565658301115036,
+ "learning_rate": 0.0006,
+ "loss": 4.232687950134277,
+ "step": 2136
+ },
+ {
+ "epoch": 29.68471615720524,
+ "grad_norm": 0.016485940665006638,
+ "learning_rate": 0.0006,
+ "loss": 4.209839820861816,
+ "step": 2137
+ },
+ {
+ "epoch": 29.69868995633188,
+ "grad_norm": 0.017622729763388634,
+ "learning_rate": 0.0006,
+ "loss": 4.196017265319824,
+ "step": 2138
+ },
+ {
+ "epoch": 29.712663755458514,
+ "grad_norm": 0.018320564180612564,
+ "learning_rate": 0.0006,
+ "loss": 4.225683689117432,
+ "step": 2139
+ },
+ {
+ "epoch": 29.726637554585153,
+ "grad_norm": 0.018140794709324837,
+ "learning_rate": 0.0006,
+ "loss": 4.311944961547852,
+ "step": 2140
+ },
+ {
+ "epoch": 29.74061135371179,
+ "grad_norm": 0.016227100044488907,
+ "learning_rate": 0.0006,
+ "loss": 4.230257511138916,
+ "step": 2141
+ },
+ {
+ "epoch": 29.754585152838427,
+ "grad_norm": 0.015759488567709923,
+ "learning_rate": 0.0006,
+ "loss": 4.119174003601074,
+ "step": 2142
+ },
+ {
+ "epoch": 29.768558951965066,
+ "grad_norm": 0.01757766678929329,
+ "learning_rate": 0.0006,
+ "loss": 4.165826797485352,
+ "step": 2143
+ },
+ {
+ "epoch": 29.782532751091704,
+ "grad_norm": 0.019661923870444298,
+ "learning_rate": 0.0006,
+ "loss": 4.392203330993652,
+ "step": 2144
+ },
+ {
+ "epoch": 29.79650655021834,
+ "grad_norm": 0.019478755071759224,
+ "learning_rate": 0.0006,
+ "loss": 4.348371982574463,
+ "step": 2145
+ },
+ {
+ "epoch": 29.810480349344978,
+ "grad_norm": 0.019149569794535637,
+ "learning_rate": 0.0006,
+ "loss": 4.34104061126709,
+ "step": 2146
+ },
+ {
+ "epoch": 29.824454148471617,
+ "grad_norm": 0.017707521095871925,
+ "learning_rate": 0.0006,
+ "loss": 4.228226661682129,
+ "step": 2147
+ },
+ {
+ "epoch": 29.83842794759825,
+ "grad_norm": 0.015615378506481647,
+ "learning_rate": 0.0006,
+ "loss": 4.175541877746582,
+ "step": 2148
+ },
+ {
+ "epoch": 29.85240174672489,
+ "grad_norm": 0.016111129894852638,
+ "learning_rate": 0.0006,
+ "loss": 4.267936706542969,
+ "step": 2149
+ },
+ {
+ "epoch": 29.86637554585153,
+ "grad_norm": 0.01528779324144125,
+ "learning_rate": 0.0006,
+ "loss": 4.186014175415039,
+ "step": 2150
+ },
+ {
+ "epoch": 29.880349344978168,
+ "grad_norm": 0.014925646595656872,
+ "learning_rate": 0.0006,
+ "loss": 4.174860000610352,
+ "step": 2151
+ },
+ {
+ "epoch": 29.894323144104803,
+ "grad_norm": 0.016917673870921135,
+ "learning_rate": 0.0006,
+ "loss": 4.33908748626709,
+ "step": 2152
+ },
+ {
+ "epoch": 29.90829694323144,
+ "grad_norm": 0.016007075086236,
+ "learning_rate": 0.0006,
+ "loss": 4.266141414642334,
+ "step": 2153
+ },
+ {
+ "epoch": 29.92227074235808,
+ "grad_norm": 0.016178956255316734,
+ "learning_rate": 0.0006,
+ "loss": 4.271215915679932,
+ "step": 2154
+ },
+ {
+ "epoch": 29.936244541484715,
+ "grad_norm": 0.016180871054530144,
+ "learning_rate": 0.0006,
+ "loss": 4.278801441192627,
+ "step": 2155
+ },
+ {
+ "epoch": 29.950218340611354,
+ "grad_norm": 0.015335503034293652,
+ "learning_rate": 0.0006,
+ "loss": 4.194978713989258,
+ "step": 2156
+ },
+ {
+ "epoch": 29.964192139737992,
+ "grad_norm": 0.016580764204263687,
+ "learning_rate": 0.0006,
+ "loss": 4.290809631347656,
+ "step": 2157
+ },
+ {
+ "epoch": 29.978165938864628,
+ "grad_norm": 0.01623128168284893,
+ "learning_rate": 0.0006,
+ "loss": 4.205451488494873,
+ "step": 2158
+ },
+ {
+ "epoch": 29.992139737991266,
+ "grad_norm": 0.014904248528182507,
+ "learning_rate": 0.0006,
+ "loss": 4.24924373626709,
+ "step": 2159
+ },
+ {
+ "epoch": 30.0,
+ "grad_norm": 0.017089299857616425,
+ "learning_rate": 0.0006,
+ "loss": 4.222072601318359,
+ "step": 2160
+ },
+ {
+ "epoch": 30.0,
+ "eval_loss": 4.585411071777344,
+ "eval_runtime": 57.1164,
+ "eval_samples_per_second": 42.755,
+ "eval_steps_per_second": 1.348,
+ "step": 2160
+ },
+ {
+ "epoch": 30.01397379912664,
+ "grad_norm": 0.017226383090019226,
+ "learning_rate": 0.0006,
+ "loss": 4.19467830657959,
+ "step": 2161
+ },
+ {
+ "epoch": 30.027947598253274,
+ "grad_norm": 0.01635066606104374,
+ "learning_rate": 0.0006,
+ "loss": 4.176875591278076,
+ "step": 2162
+ },
+ {
+ "epoch": 30.041921397379912,
+ "grad_norm": 0.01735362410545349,
+ "learning_rate": 0.0006,
+ "loss": 4.13667106628418,
+ "step": 2163
+ },
+ {
+ "epoch": 30.05589519650655,
+ "grad_norm": 0.017047109082341194,
+ "learning_rate": 0.0006,
+ "loss": 4.136329650878906,
+ "step": 2164
+ },
+ {
+ "epoch": 30.069868995633186,
+ "grad_norm": 0.018116053193807602,
+ "learning_rate": 0.0006,
+ "loss": 4.228706359863281,
+ "step": 2165
+ },
+ {
+ "epoch": 30.083842794759825,
+ "grad_norm": 0.01791389286518097,
+ "learning_rate": 0.0006,
+ "loss": 4.1597747802734375,
+ "step": 2166
+ },
+ {
+ "epoch": 30.097816593886463,
+ "grad_norm": 0.017793502658605576,
+ "learning_rate": 0.0006,
+ "loss": 4.0807976722717285,
+ "step": 2167
+ },
+ {
+ "epoch": 30.111790393013102,
+ "grad_norm": 0.01827503927052021,
+ "learning_rate": 0.0006,
+ "loss": 4.170646667480469,
+ "step": 2168
+ },
+ {
+ "epoch": 30.125764192139737,
+ "grad_norm": 0.017228564247488976,
+ "learning_rate": 0.0006,
+ "loss": 4.146934509277344,
+ "step": 2169
+ },
+ {
+ "epoch": 30.139737991266376,
+ "grad_norm": 0.01757437363266945,
+ "learning_rate": 0.0006,
+ "loss": 4.1859130859375,
+ "step": 2170
+ },
+ {
+ "epoch": 30.153711790393015,
+ "grad_norm": 0.01701093092560768,
+ "learning_rate": 0.0006,
+ "loss": 4.126981258392334,
+ "step": 2171
+ },
+ {
+ "epoch": 30.16768558951965,
+ "grad_norm": 0.015220489352941513,
+ "learning_rate": 0.0006,
+ "loss": 4.164018630981445,
+ "step": 2172
+ },
+ {
+ "epoch": 30.18165938864629,
+ "grad_norm": 0.016254238784313202,
+ "learning_rate": 0.0006,
+ "loss": 4.258275508880615,
+ "step": 2173
+ },
+ {
+ "epoch": 30.195633187772927,
+ "grad_norm": 0.016609614714980125,
+ "learning_rate": 0.0006,
+ "loss": 4.222233772277832,
+ "step": 2174
+ },
+ {
+ "epoch": 30.209606986899562,
+ "grad_norm": 0.016531318426132202,
+ "learning_rate": 0.0006,
+ "loss": 4.144489288330078,
+ "step": 2175
+ },
+ {
+ "epoch": 30.2235807860262,
+ "grad_norm": 0.01567668654024601,
+ "learning_rate": 0.0006,
+ "loss": 4.181916236877441,
+ "step": 2176
+ },
+ {
+ "epoch": 30.23755458515284,
+ "grad_norm": 0.017195681110024452,
+ "learning_rate": 0.0006,
+ "loss": 4.189693927764893,
+ "step": 2177
+ },
+ {
+ "epoch": 30.251528384279474,
+ "grad_norm": 0.016882948577404022,
+ "learning_rate": 0.0006,
+ "loss": 4.21604061126709,
+ "step": 2178
+ },
+ {
+ "epoch": 30.265502183406113,
+ "grad_norm": 0.016261622309684753,
+ "learning_rate": 0.0006,
+ "loss": 4.227639198303223,
+ "step": 2179
+ },
+ {
+ "epoch": 30.27947598253275,
+ "grad_norm": 0.016411812976002693,
+ "learning_rate": 0.0006,
+ "loss": 4.130362510681152,
+ "step": 2180
+ },
+ {
+ "epoch": 30.29344978165939,
+ "grad_norm": 0.01732019893825054,
+ "learning_rate": 0.0006,
+ "loss": 4.230443000793457,
+ "step": 2181
+ },
+ {
+ "epoch": 30.307423580786025,
+ "grad_norm": 0.01825905591249466,
+ "learning_rate": 0.0006,
+ "loss": 4.175817489624023,
+ "step": 2182
+ },
+ {
+ "epoch": 30.321397379912664,
+ "grad_norm": 0.01887713558971882,
+ "learning_rate": 0.0006,
+ "loss": 4.169020652770996,
+ "step": 2183
+ },
+ {
+ "epoch": 30.335371179039303,
+ "grad_norm": 0.019391661509871483,
+ "learning_rate": 0.0006,
+ "loss": 4.176318645477295,
+ "step": 2184
+ },
+ {
+ "epoch": 30.349344978165938,
+ "grad_norm": 0.020840002223849297,
+ "learning_rate": 0.0006,
+ "loss": 4.174228668212891,
+ "step": 2185
+ },
+ {
+ "epoch": 30.363318777292577,
+ "grad_norm": 0.020193178206682205,
+ "learning_rate": 0.0006,
+ "loss": 4.127919673919678,
+ "step": 2186
+ },
+ {
+ "epoch": 30.377292576419215,
+ "grad_norm": 0.018449561670422554,
+ "learning_rate": 0.0006,
+ "loss": 4.229487895965576,
+ "step": 2187
+ },
+ {
+ "epoch": 30.39126637554585,
+ "grad_norm": 0.01711983233690262,
+ "learning_rate": 0.0006,
+ "loss": 4.201651573181152,
+ "step": 2188
+ },
+ {
+ "epoch": 30.40524017467249,
+ "grad_norm": 0.017293395474553108,
+ "learning_rate": 0.0006,
+ "loss": 4.168704032897949,
+ "step": 2189
+ },
+ {
+ "epoch": 30.419213973799128,
+ "grad_norm": 0.018526358529925346,
+ "learning_rate": 0.0006,
+ "loss": 4.171864032745361,
+ "step": 2190
+ },
+ {
+ "epoch": 30.433187772925763,
+ "grad_norm": 0.019287291914224625,
+ "learning_rate": 0.0006,
+ "loss": 4.039534568786621,
+ "step": 2191
+ },
+ {
+ "epoch": 30.4471615720524,
+ "grad_norm": 0.016902444884181023,
+ "learning_rate": 0.0006,
+ "loss": 4.282787799835205,
+ "step": 2192
+ },
+ {
+ "epoch": 30.46113537117904,
+ "grad_norm": 0.01656750962138176,
+ "learning_rate": 0.0006,
+ "loss": 4.213098526000977,
+ "step": 2193
+ },
+ {
+ "epoch": 30.475109170305675,
+ "grad_norm": 0.016805065795779228,
+ "learning_rate": 0.0006,
+ "loss": 4.1583476066589355,
+ "step": 2194
+ },
+ {
+ "epoch": 30.489082969432314,
+ "grad_norm": 0.017166638746857643,
+ "learning_rate": 0.0006,
+ "loss": 4.188634395599365,
+ "step": 2195
+ },
+ {
+ "epoch": 30.503056768558952,
+ "grad_norm": 0.01592106744647026,
+ "learning_rate": 0.0006,
+ "loss": 4.234043121337891,
+ "step": 2196
+ },
+ {
+ "epoch": 30.51703056768559,
+ "grad_norm": 0.015270989388227463,
+ "learning_rate": 0.0006,
+ "loss": 4.301420211791992,
+ "step": 2197
+ },
+ {
+ "epoch": 30.531004366812226,
+ "grad_norm": 0.015194724313914776,
+ "learning_rate": 0.0006,
+ "loss": 4.18703031539917,
+ "step": 2198
+ },
+ {
+ "epoch": 30.544978165938865,
+ "grad_norm": 0.015587746165692806,
+ "learning_rate": 0.0006,
+ "loss": 4.138928413391113,
+ "step": 2199
+ },
+ {
+ "epoch": 30.558951965065503,
+ "grad_norm": 0.016029570251703262,
+ "learning_rate": 0.0006,
+ "loss": 4.290863990783691,
+ "step": 2200
+ },
+ {
+ "epoch": 30.57292576419214,
+ "grad_norm": 0.015651309862732887,
+ "learning_rate": 0.0006,
+ "loss": 4.272062301635742,
+ "step": 2201
+ },
+ {
+ "epoch": 30.586899563318777,
+ "grad_norm": 0.015992306172847748,
+ "learning_rate": 0.0006,
+ "loss": 4.2487406730651855,
+ "step": 2202
+ },
+ {
+ "epoch": 30.600873362445416,
+ "grad_norm": 0.014969157986342907,
+ "learning_rate": 0.0006,
+ "loss": 4.115347862243652,
+ "step": 2203
+ },
+ {
+ "epoch": 30.61484716157205,
+ "grad_norm": 0.014639027416706085,
+ "learning_rate": 0.0006,
+ "loss": 4.2207489013671875,
+ "step": 2204
+ },
+ {
+ "epoch": 30.62882096069869,
+ "grad_norm": 0.015059413388371468,
+ "learning_rate": 0.0006,
+ "loss": 4.274044990539551,
+ "step": 2205
+ },
+ {
+ "epoch": 30.64279475982533,
+ "grad_norm": 0.0166641678661108,
+ "learning_rate": 0.0006,
+ "loss": 4.191373348236084,
+ "step": 2206
+ },
+ {
+ "epoch": 30.656768558951963,
+ "grad_norm": 0.017806345596909523,
+ "learning_rate": 0.0006,
+ "loss": 4.139800548553467,
+ "step": 2207
+ },
+ {
+ "epoch": 30.670742358078602,
+ "grad_norm": 0.018712421879172325,
+ "learning_rate": 0.0006,
+ "loss": 4.035965442657471,
+ "step": 2208
+ },
+ {
+ "epoch": 30.68471615720524,
+ "grad_norm": 0.018457243219017982,
+ "learning_rate": 0.0006,
+ "loss": 4.2579240798950195,
+ "step": 2209
+ },
+ {
+ "epoch": 30.69868995633188,
+ "grad_norm": 0.01920999586582184,
+ "learning_rate": 0.0006,
+ "loss": 4.131112098693848,
+ "step": 2210
+ },
+ {
+ "epoch": 30.712663755458514,
+ "grad_norm": 0.01729944534599781,
+ "learning_rate": 0.0006,
+ "loss": 4.254059314727783,
+ "step": 2211
+ },
+ {
+ "epoch": 30.726637554585153,
+ "grad_norm": 0.016914231702685356,
+ "learning_rate": 0.0006,
+ "loss": 4.233428955078125,
+ "step": 2212
+ },
+ {
+ "epoch": 30.74061135371179,
+ "grad_norm": 0.01913098618388176,
+ "learning_rate": 0.0006,
+ "loss": 4.23525333404541,
+ "step": 2213
+ },
+ {
+ "epoch": 30.754585152838427,
+ "grad_norm": 0.019150281324982643,
+ "learning_rate": 0.0006,
+ "loss": 4.213695049285889,
+ "step": 2214
+ },
+ {
+ "epoch": 30.768558951965066,
+ "grad_norm": 0.018947452306747437,
+ "learning_rate": 0.0006,
+ "loss": 4.151697158813477,
+ "step": 2215
+ },
+ {
+ "epoch": 30.782532751091704,
+ "grad_norm": 0.016058508306741714,
+ "learning_rate": 0.0006,
+ "loss": 4.220280170440674,
+ "step": 2216
+ },
+ {
+ "epoch": 30.79650655021834,
+ "grad_norm": 0.017488619312644005,
+ "learning_rate": 0.0006,
+ "loss": 4.154428482055664,
+ "step": 2217
+ },
+ {
+ "epoch": 30.810480349344978,
+ "grad_norm": 0.017355090007185936,
+ "learning_rate": 0.0006,
+ "loss": 4.256086349487305,
+ "step": 2218
+ },
+ {
+ "epoch": 30.824454148471617,
+ "grad_norm": 0.016636032611131668,
+ "learning_rate": 0.0006,
+ "loss": 4.227564334869385,
+ "step": 2219
+ },
+ {
+ "epoch": 30.83842794759825,
+ "grad_norm": 0.016273891553282738,
+ "learning_rate": 0.0006,
+ "loss": 4.200466156005859,
+ "step": 2220
+ },
+ {
+ "epoch": 30.85240174672489,
+ "grad_norm": 0.016939911991357803,
+ "learning_rate": 0.0006,
+ "loss": 4.1571221351623535,
+ "step": 2221
+ },
+ {
+ "epoch": 30.86637554585153,
+ "grad_norm": 0.018332522362470627,
+ "learning_rate": 0.0006,
+ "loss": 4.228398323059082,
+ "step": 2222
+ },
+ {
+ "epoch": 30.880349344978168,
+ "grad_norm": 0.019508114084601402,
+ "learning_rate": 0.0006,
+ "loss": 4.222960472106934,
+ "step": 2223
+ },
+ {
+ "epoch": 30.894323144104803,
+ "grad_norm": 0.018440278246998787,
+ "learning_rate": 0.0006,
+ "loss": 4.261051177978516,
+ "step": 2224
+ },
+ {
+ "epoch": 30.90829694323144,
+ "grad_norm": 0.018113229423761368,
+ "learning_rate": 0.0006,
+ "loss": 4.109235763549805,
+ "step": 2225
+ },
+ {
+ "epoch": 30.92227074235808,
+ "grad_norm": 0.015392083674669266,
+ "learning_rate": 0.0006,
+ "loss": 4.21950101852417,
+ "step": 2226
+ },
+ {
+ "epoch": 30.936244541484715,
+ "grad_norm": 0.01633290946483612,
+ "learning_rate": 0.0006,
+ "loss": 4.176053047180176,
+ "step": 2227
+ },
+ {
+ "epoch": 30.950218340611354,
+ "grad_norm": 0.016236383467912674,
+ "learning_rate": 0.0006,
+ "loss": 4.314309120178223,
+ "step": 2228
+ },
+ {
+ "epoch": 30.964192139737992,
+ "grad_norm": 0.015284373424947262,
+ "learning_rate": 0.0006,
+ "loss": 4.303133487701416,
+ "step": 2229
+ },
+ {
+ "epoch": 30.978165938864628,
+ "grad_norm": 0.015982897952198982,
+ "learning_rate": 0.0006,
+ "loss": 4.25739860534668,
+ "step": 2230
+ },
+ {
+ "epoch": 30.992139737991266,
+ "grad_norm": 0.016682198271155357,
+ "learning_rate": 0.0006,
+ "loss": 4.292210578918457,
+ "step": 2231
+ },
+ {
+ "epoch": 31.0,
+ "grad_norm": 0.017521467059850693,
+ "learning_rate": 0.0006,
+ "loss": 4.170637130737305,
+ "step": 2232
+ },
+ {
+ "epoch": 31.0,
+ "eval_loss": 4.642786979675293,
+ "eval_runtime": 57.029,
+ "eval_samples_per_second": 42.82,
+ "eval_steps_per_second": 1.35,
+ "step": 2232
+ },
+ {
+ "epoch": 31.01397379912664,
+ "grad_norm": 0.016607852652668953,
+ "learning_rate": 0.0006,
+ "loss": 4.106224536895752,
+ "step": 2233
+ },
+ {
+ "epoch": 31.027947598253274,
+ "grad_norm": 0.018448904156684875,
+ "learning_rate": 0.0006,
+ "loss": 4.224459648132324,
+ "step": 2234
+ },
+ {
+ "epoch": 31.041921397379912,
+ "grad_norm": 0.020481228828430176,
+ "learning_rate": 0.0006,
+ "loss": 4.305548667907715,
+ "step": 2235
+ },
+ {
+ "epoch": 31.05589519650655,
+ "grad_norm": 0.01999707892537117,
+ "learning_rate": 0.0006,
+ "loss": 4.253663539886475,
+ "step": 2236
+ },
+ {
+ "epoch": 31.069868995633186,
+ "grad_norm": 0.020428303629159927,
+ "learning_rate": 0.0006,
+ "loss": 4.166954040527344,
+ "step": 2237
+ },
+ {
+ "epoch": 31.083842794759825,
+ "grad_norm": 0.022608213126659393,
+ "learning_rate": 0.0006,
+ "loss": 4.172819137573242,
+ "step": 2238
+ },
+ {
+ "epoch": 31.097816593886463,
+ "grad_norm": 0.024669520556926727,
+ "learning_rate": 0.0006,
+ "loss": 4.106586456298828,
+ "step": 2239
+ },
+ {
+ "epoch": 31.111790393013102,
+ "grad_norm": 0.02247351035475731,
+ "learning_rate": 0.0006,
+ "loss": 4.160271644592285,
+ "step": 2240
+ },
+ {
+ "epoch": 31.125764192139737,
+ "grad_norm": 0.0207810141146183,
+ "learning_rate": 0.0006,
+ "loss": 4.272491931915283,
+ "step": 2241
+ },
+ {
+ "epoch": 31.139737991266376,
+ "grad_norm": 0.021576499566435814,
+ "learning_rate": 0.0006,
+ "loss": 4.171474456787109,
+ "step": 2242
+ },
+ {
+ "epoch": 31.153711790393015,
+ "grad_norm": 0.02337106689810753,
+ "learning_rate": 0.0006,
+ "loss": 4.094712734222412,
+ "step": 2243
+ },
+ {
+ "epoch": 31.16768558951965,
+ "grad_norm": 0.02219875156879425,
+ "learning_rate": 0.0006,
+ "loss": 4.1800994873046875,
+ "step": 2244
+ },
+ {
+ "epoch": 31.18165938864629,
+ "grad_norm": 0.022680504247546196,
+ "learning_rate": 0.0006,
+ "loss": 4.2054362297058105,
+ "step": 2245
+ },
+ {
+ "epoch": 31.195633187772927,
+ "grad_norm": 0.0208734143525362,
+ "learning_rate": 0.0006,
+ "loss": 4.06068229675293,
+ "step": 2246
+ },
+ {
+ "epoch": 31.209606986899562,
+ "grad_norm": 0.01811066083610058,
+ "learning_rate": 0.0006,
+ "loss": 4.2465128898620605,
+ "step": 2247
+ },
+ {
+ "epoch": 31.2235807860262,
+ "grad_norm": 0.018571637570858,
+ "learning_rate": 0.0006,
+ "loss": 4.191955089569092,
+ "step": 2248
+ },
+ {
+ "epoch": 31.23755458515284,
+ "grad_norm": 0.018598807975649834,
+ "learning_rate": 0.0006,
+ "loss": 4.200872421264648,
+ "step": 2249
+ },
+ {
+ "epoch": 31.251528384279474,
+ "grad_norm": 0.017577625811100006,
+ "learning_rate": 0.0006,
+ "loss": 4.148700714111328,
+ "step": 2250
+ },
+ {
+ "epoch": 31.265502183406113,
+ "grad_norm": 0.01817980408668518,
+ "learning_rate": 0.0006,
+ "loss": 4.256978988647461,
+ "step": 2251
+ },
+ {
+ "epoch": 31.27947598253275,
+ "grad_norm": 0.017983486875891685,
+ "learning_rate": 0.0006,
+ "loss": 4.175005912780762,
+ "step": 2252
+ },
+ {
+ "epoch": 31.29344978165939,
+ "grad_norm": 0.016751961782574654,
+ "learning_rate": 0.0006,
+ "loss": 4.237518310546875,
+ "step": 2253
+ },
+ {
+ "epoch": 31.307423580786025,
+ "grad_norm": 0.016407202929258347,
+ "learning_rate": 0.0006,
+ "loss": 4.197269439697266,
+ "step": 2254
+ },
+ {
+ "epoch": 31.321397379912664,
+ "grad_norm": 0.01578645594418049,
+ "learning_rate": 0.0006,
+ "loss": 4.179751396179199,
+ "step": 2255
+ },
+ {
+ "epoch": 31.335371179039303,
+ "grad_norm": 0.016001226380467415,
+ "learning_rate": 0.0006,
+ "loss": 4.08821964263916,
+ "step": 2256
+ },
+ {
+ "epoch": 31.349344978165938,
+ "grad_norm": 0.015122811309993267,
+ "learning_rate": 0.0006,
+ "loss": 4.047828674316406,
+ "step": 2257
+ },
+ {
+ "epoch": 31.363318777292577,
+ "grad_norm": 0.015399965457618237,
+ "learning_rate": 0.0006,
+ "loss": 4.1506524085998535,
+ "step": 2258
+ },
+ {
+ "epoch": 31.377292576419215,
+ "grad_norm": 0.01534605398774147,
+ "learning_rate": 0.0006,
+ "loss": 4.27952766418457,
+ "step": 2259
+ },
+ {
+ "epoch": 31.39126637554585,
+ "grad_norm": 0.015461381524801254,
+ "learning_rate": 0.0006,
+ "loss": 4.27102518081665,
+ "step": 2260
+ },
+ {
+ "epoch": 31.40524017467249,
+ "grad_norm": 0.016769153997302055,
+ "learning_rate": 0.0006,
+ "loss": 4.191592216491699,
+ "step": 2261
+ },
+ {
+ "epoch": 31.419213973799128,
+ "grad_norm": 0.016557959839701653,
+ "learning_rate": 0.0006,
+ "loss": 4.159705638885498,
+ "step": 2262
+ },
+ {
+ "epoch": 31.433187772925763,
+ "grad_norm": 0.01659569889307022,
+ "learning_rate": 0.0006,
+ "loss": 4.2136664390563965,
+ "step": 2263
+ },
+ {
+ "epoch": 31.4471615720524,
+ "grad_norm": 0.017101937904953957,
+ "learning_rate": 0.0006,
+ "loss": 4.2716498374938965,
+ "step": 2264
+ },
+ {
+ "epoch": 31.46113537117904,
+ "grad_norm": 0.01722985878586769,
+ "learning_rate": 0.0006,
+ "loss": 4.300534248352051,
+ "step": 2265
+ },
+ {
+ "epoch": 31.475109170305675,
+ "grad_norm": 0.016800185665488243,
+ "learning_rate": 0.0006,
+ "loss": 4.113240718841553,
+ "step": 2266
+ },
+ {
+ "epoch": 31.489082969432314,
+ "grad_norm": 0.01603875495493412,
+ "learning_rate": 0.0006,
+ "loss": 4.157105445861816,
+ "step": 2267
+ },
+ {
+ "epoch": 31.503056768558952,
+ "grad_norm": 0.015903932973742485,
+ "learning_rate": 0.0006,
+ "loss": 4.233669757843018,
+ "step": 2268
+ },
+ {
+ "epoch": 31.51703056768559,
+ "grad_norm": 0.0160983856767416,
+ "learning_rate": 0.0006,
+ "loss": 4.147704124450684,
+ "step": 2269
+ },
+ {
+ "epoch": 31.531004366812226,
+ "grad_norm": 0.01604730263352394,
+ "learning_rate": 0.0006,
+ "loss": 4.1465911865234375,
+ "step": 2270
+ },
+ {
+ "epoch": 31.544978165938865,
+ "grad_norm": 0.016205785796046257,
+ "learning_rate": 0.0006,
+ "loss": 4.106704235076904,
+ "step": 2271
+ },
+ {
+ "epoch": 31.558951965065503,
+ "grad_norm": 0.017008036375045776,
+ "learning_rate": 0.0006,
+ "loss": 4.191000938415527,
+ "step": 2272
+ },
+ {
+ "epoch": 31.57292576419214,
+ "grad_norm": 0.01699635200202465,
+ "learning_rate": 0.0006,
+ "loss": 4.187042236328125,
+ "step": 2273
+ },
+ {
+ "epoch": 31.586899563318777,
+ "grad_norm": 0.01692967116832733,
+ "learning_rate": 0.0006,
+ "loss": 4.208805561065674,
+ "step": 2274
+ },
+ {
+ "epoch": 31.600873362445416,
+ "grad_norm": 0.017608756199479103,
+ "learning_rate": 0.0006,
+ "loss": 4.200512886047363,
+ "step": 2275
+ },
+ {
+ "epoch": 31.61484716157205,
+ "grad_norm": 0.016590990126132965,
+ "learning_rate": 0.0006,
+ "loss": 4.194705009460449,
+ "step": 2276
+ },
+ {
+ "epoch": 31.62882096069869,
+ "grad_norm": 0.0170235987752676,
+ "learning_rate": 0.0006,
+ "loss": 4.210326194763184,
+ "step": 2277
+ },
+ {
+ "epoch": 31.64279475982533,
+ "grad_norm": 0.01604801043868065,
+ "learning_rate": 0.0006,
+ "loss": 4.196024417877197,
+ "step": 2278
+ },
+ {
+ "epoch": 31.656768558951963,
+ "grad_norm": 0.015694202855229378,
+ "learning_rate": 0.0006,
+ "loss": 4.178236484527588,
+ "step": 2279
+ },
+ {
+ "epoch": 31.670742358078602,
+ "grad_norm": 0.015473959036171436,
+ "learning_rate": 0.0006,
+ "loss": 4.259339809417725,
+ "step": 2280
+ },
+ {
+ "epoch": 31.68471615720524,
+ "grad_norm": 0.014689362607896328,
+ "learning_rate": 0.0006,
+ "loss": 3.9702324867248535,
+ "step": 2281
+ },
+ {
+ "epoch": 31.69868995633188,
+ "grad_norm": 0.015245123766362667,
+ "learning_rate": 0.0006,
+ "loss": 4.197025775909424,
+ "step": 2282
+ },
+ {
+ "epoch": 31.712663755458514,
+ "grad_norm": 0.016148347407579422,
+ "learning_rate": 0.0006,
+ "loss": 4.231856346130371,
+ "step": 2283
+ },
+ {
+ "epoch": 31.726637554585153,
+ "grad_norm": 0.015169818885624409,
+ "learning_rate": 0.0006,
+ "loss": 4.230169296264648,
+ "step": 2284
+ },
+ {
+ "epoch": 31.74061135371179,
+ "grad_norm": 0.014843013137578964,
+ "learning_rate": 0.0006,
+ "loss": 4.195633411407471,
+ "step": 2285
+ },
+ {
+ "epoch": 31.754585152838427,
+ "grad_norm": 0.015229340642690659,
+ "learning_rate": 0.0006,
+ "loss": 4.231280326843262,
+ "step": 2286
+ },
+ {
+ "epoch": 31.768558951965066,
+ "grad_norm": 0.01551708485931158,
+ "learning_rate": 0.0006,
+ "loss": 4.17064094543457,
+ "step": 2287
+ },
+ {
+ "epoch": 31.782532751091704,
+ "grad_norm": 0.015863575041294098,
+ "learning_rate": 0.0006,
+ "loss": 4.1579484939575195,
+ "step": 2288
+ },
+ {
+ "epoch": 31.79650655021834,
+ "grad_norm": 0.015118198469281197,
+ "learning_rate": 0.0006,
+ "loss": 4.1626787185668945,
+ "step": 2289
+ },
+ {
+ "epoch": 31.810480349344978,
+ "grad_norm": 0.015220776200294495,
+ "learning_rate": 0.0006,
+ "loss": 4.135746955871582,
+ "step": 2290
+ },
+ {
+ "epoch": 31.824454148471617,
+ "grad_norm": 0.014529784210026264,
+ "learning_rate": 0.0006,
+ "loss": 4.164931297302246,
+ "step": 2291
+ },
+ {
+ "epoch": 31.83842794759825,
+ "grad_norm": 0.015533296391367912,
+ "learning_rate": 0.0006,
+ "loss": 4.112464904785156,
+ "step": 2292
+ },
+ {
+ "epoch": 31.85240174672489,
+ "grad_norm": 0.016562335193157196,
+ "learning_rate": 0.0006,
+ "loss": 4.164898872375488,
+ "step": 2293
+ },
+ {
+ "epoch": 31.86637554585153,
+ "grad_norm": 0.016258113086223602,
+ "learning_rate": 0.0006,
+ "loss": 4.1979265213012695,
+ "step": 2294
+ },
+ {
+ "epoch": 31.880349344978168,
+ "grad_norm": 0.01558110024780035,
+ "learning_rate": 0.0006,
+ "loss": 4.174675941467285,
+ "step": 2295
+ },
+ {
+ "epoch": 31.894323144104803,
+ "grad_norm": 0.016412515193223953,
+ "learning_rate": 0.0006,
+ "loss": 4.217909336090088,
+ "step": 2296
+ },
+ {
+ "epoch": 31.90829694323144,
+ "grad_norm": 0.015118442475795746,
+ "learning_rate": 0.0006,
+ "loss": 4.137143611907959,
+ "step": 2297
+ },
+ {
+ "epoch": 31.92227074235808,
+ "grad_norm": 0.01506161317229271,
+ "learning_rate": 0.0006,
+ "loss": 4.196122169494629,
+ "step": 2298
+ },
+ {
+ "epoch": 31.936244541484715,
+ "grad_norm": 0.017165271565318108,
+ "learning_rate": 0.0006,
+ "loss": 4.14834451675415,
+ "step": 2299
+ },
+ {
+ "epoch": 31.950218340611354,
+ "grad_norm": 0.01696440577507019,
+ "learning_rate": 0.0006,
+ "loss": 4.218868732452393,
+ "step": 2300
+ },
+ {
+ "epoch": 31.964192139737992,
+ "grad_norm": 0.017349394038319588,
+ "learning_rate": 0.0006,
+ "loss": 4.2754011154174805,
+ "step": 2301
+ },
+ {
+ "epoch": 31.978165938864628,
+ "grad_norm": 0.018387921154499054,
+ "learning_rate": 0.0006,
+ "loss": 4.282623291015625,
+ "step": 2302
+ },
+ {
+ "epoch": 31.992139737991266,
+ "grad_norm": 0.018837926909327507,
+ "learning_rate": 0.0006,
+ "loss": 4.185425281524658,
+ "step": 2303
+ },
+ {
+ "epoch": 32.0,
+ "grad_norm": 0.018998464569449425,
+ "learning_rate": 0.0006,
+ "loss": 4.151291847229004,
+ "step": 2304
+ },
+ {
+ "epoch": 32.0,
+ "eval_loss": 4.522614002227783,
+ "eval_runtime": 57.0106,
+ "eval_samples_per_second": 42.834,
+ "eval_steps_per_second": 1.351,
+ "step": 2304
+ },
+ {
+ "epoch": 32.01397379912664,
+ "grad_norm": 0.019036833196878433,
+ "learning_rate": 0.0006,
+ "loss": 4.183202266693115,
+ "step": 2305
+ },
+ {
+ "epoch": 32.02794759825328,
+ "grad_norm": 0.02207397110760212,
+ "learning_rate": 0.0006,
+ "loss": 4.17875862121582,
+ "step": 2306
+ },
+ {
+ "epoch": 32.041921397379916,
+ "grad_norm": 0.022008279338479042,
+ "learning_rate": 0.0006,
+ "loss": 4.08042049407959,
+ "step": 2307
+ },
+ {
+ "epoch": 32.05589519650655,
+ "grad_norm": 0.022407017648220062,
+ "learning_rate": 0.0006,
+ "loss": 4.112189769744873,
+ "step": 2308
+ },
+ {
+ "epoch": 32.069868995633186,
+ "grad_norm": 0.021599747240543365,
+ "learning_rate": 0.0006,
+ "loss": 4.126455307006836,
+ "step": 2309
+ },
+ {
+ "epoch": 32.083842794759825,
+ "grad_norm": 0.020340966060757637,
+ "learning_rate": 0.0006,
+ "loss": 4.080538272857666,
+ "step": 2310
+ },
+ {
+ "epoch": 32.09781659388646,
+ "grad_norm": 0.020956892520189285,
+ "learning_rate": 0.0006,
+ "loss": 4.171943664550781,
+ "step": 2311
+ },
+ {
+ "epoch": 32.1117903930131,
+ "grad_norm": 0.019966667518019676,
+ "learning_rate": 0.0006,
+ "loss": 4.144007682800293,
+ "step": 2312
+ },
+ {
+ "epoch": 32.12576419213974,
+ "grad_norm": 0.018256952986121178,
+ "learning_rate": 0.0006,
+ "loss": 4.208107948303223,
+ "step": 2313
+ },
+ {
+ "epoch": 32.13973799126637,
+ "grad_norm": 0.017289170995354652,
+ "learning_rate": 0.0006,
+ "loss": 4.182024955749512,
+ "step": 2314
+ },
+ {
+ "epoch": 32.15371179039301,
+ "grad_norm": 0.017656700685620308,
+ "learning_rate": 0.0006,
+ "loss": 4.068373680114746,
+ "step": 2315
+ },
+ {
+ "epoch": 32.16768558951965,
+ "grad_norm": 0.017903361469507217,
+ "learning_rate": 0.0006,
+ "loss": 4.2345428466796875,
+ "step": 2316
+ },
+ {
+ "epoch": 32.18165938864629,
+ "grad_norm": 0.01785109005868435,
+ "learning_rate": 0.0006,
+ "loss": 4.219776153564453,
+ "step": 2317
+ },
+ {
+ "epoch": 32.19563318777293,
+ "grad_norm": 0.01887257769703865,
+ "learning_rate": 0.0006,
+ "loss": 4.139206409454346,
+ "step": 2318
+ },
+ {
+ "epoch": 32.209606986899566,
+ "grad_norm": 0.01924685761332512,
+ "learning_rate": 0.0006,
+ "loss": 4.121052265167236,
+ "step": 2319
+ },
+ {
+ "epoch": 32.223580786026204,
+ "grad_norm": 0.020574072375893593,
+ "learning_rate": 0.0006,
+ "loss": 4.13262414932251,
+ "step": 2320
+ },
+ {
+ "epoch": 32.237554585152836,
+ "grad_norm": 0.023849064484238625,
+ "learning_rate": 0.0006,
+ "loss": 4.239119529724121,
+ "step": 2321
+ },
+ {
+ "epoch": 32.251528384279474,
+ "grad_norm": 0.026239361613988876,
+ "learning_rate": 0.0006,
+ "loss": 4.248993873596191,
+ "step": 2322
+ },
+ {
+ "epoch": 32.26550218340611,
+ "grad_norm": 0.023980243131518364,
+ "learning_rate": 0.0006,
+ "loss": 4.159852504730225,
+ "step": 2323
+ },
+ {
+ "epoch": 32.27947598253275,
+ "grad_norm": 0.020442405715584755,
+ "learning_rate": 0.0006,
+ "loss": 4.133546829223633,
+ "step": 2324
+ },
+ {
+ "epoch": 32.29344978165939,
+ "grad_norm": 0.020661186426877975,
+ "learning_rate": 0.0006,
+ "loss": 4.078243732452393,
+ "step": 2325
+ },
+ {
+ "epoch": 32.30742358078603,
+ "grad_norm": 0.02046387456357479,
+ "learning_rate": 0.0006,
+ "loss": 4.172012805938721,
+ "step": 2326
+ },
+ {
+ "epoch": 32.32139737991266,
+ "grad_norm": 0.02089597098529339,
+ "learning_rate": 0.0006,
+ "loss": 4.0604705810546875,
+ "step": 2327
+ },
+ {
+ "epoch": 32.3353711790393,
+ "grad_norm": 0.01875591278076172,
+ "learning_rate": 0.0006,
+ "loss": 4.125911235809326,
+ "step": 2328
+ },
+ {
+ "epoch": 32.34934497816594,
+ "grad_norm": 0.018606796860694885,
+ "learning_rate": 0.0006,
+ "loss": 4.097362518310547,
+ "step": 2329
+ },
+ {
+ "epoch": 32.36331877729258,
+ "grad_norm": 0.01766626164317131,
+ "learning_rate": 0.0006,
+ "loss": 4.118616580963135,
+ "step": 2330
+ },
+ {
+ "epoch": 32.377292576419215,
+ "grad_norm": 0.017692960798740387,
+ "learning_rate": 0.0006,
+ "loss": 4.1457672119140625,
+ "step": 2331
+ },
+ {
+ "epoch": 32.391266375545854,
+ "grad_norm": 0.017619704827666283,
+ "learning_rate": 0.0006,
+ "loss": 4.175961494445801,
+ "step": 2332
+ },
+ {
+ "epoch": 32.40524017467249,
+ "grad_norm": 0.01784413494169712,
+ "learning_rate": 0.0006,
+ "loss": 4.051291465759277,
+ "step": 2333
+ },
+ {
+ "epoch": 32.419213973799124,
+ "grad_norm": 0.018508171662688255,
+ "learning_rate": 0.0006,
+ "loss": 4.0638556480407715,
+ "step": 2334
+ },
+ {
+ "epoch": 32.43318777292576,
+ "grad_norm": 0.01731436885893345,
+ "learning_rate": 0.0006,
+ "loss": 4.1805419921875,
+ "step": 2335
+ },
+ {
+ "epoch": 32.4471615720524,
+ "grad_norm": 0.015886498615145683,
+ "learning_rate": 0.0006,
+ "loss": 4.151026725769043,
+ "step": 2336
+ },
+ {
+ "epoch": 32.46113537117904,
+ "grad_norm": 0.016583820804953575,
+ "learning_rate": 0.0006,
+ "loss": 4.203604698181152,
+ "step": 2337
+ },
+ {
+ "epoch": 32.47510917030568,
+ "grad_norm": 0.0176097322255373,
+ "learning_rate": 0.0006,
+ "loss": 4.135159015655518,
+ "step": 2338
+ },
+ {
+ "epoch": 32.48908296943232,
+ "grad_norm": 0.016657903790473938,
+ "learning_rate": 0.0006,
+ "loss": 4.126681804656982,
+ "step": 2339
+ },
+ {
+ "epoch": 32.50305676855895,
+ "grad_norm": 0.016866130754351616,
+ "learning_rate": 0.0006,
+ "loss": 4.162341117858887,
+ "step": 2340
+ },
+ {
+ "epoch": 32.51703056768559,
+ "grad_norm": 0.016286680474877357,
+ "learning_rate": 0.0006,
+ "loss": 4.261801242828369,
+ "step": 2341
+ },
+ {
+ "epoch": 32.531004366812226,
+ "grad_norm": 0.01481365505605936,
+ "learning_rate": 0.0006,
+ "loss": 4.110016822814941,
+ "step": 2342
+ },
+ {
+ "epoch": 32.544978165938865,
+ "grad_norm": 0.016605013981461525,
+ "learning_rate": 0.0006,
+ "loss": 4.230024814605713,
+ "step": 2343
+ },
+ {
+ "epoch": 32.5589519650655,
+ "grad_norm": 0.01645175740122795,
+ "learning_rate": 0.0006,
+ "loss": 4.174612045288086,
+ "step": 2344
+ },
+ {
+ "epoch": 32.57292576419214,
+ "grad_norm": 0.014725026674568653,
+ "learning_rate": 0.0006,
+ "loss": 4.066371440887451,
+ "step": 2345
+ },
+ {
+ "epoch": 32.58689956331878,
+ "grad_norm": 0.015003837645053864,
+ "learning_rate": 0.0006,
+ "loss": 4.193343639373779,
+ "step": 2346
+ },
+ {
+ "epoch": 32.60087336244541,
+ "grad_norm": 0.015346302650868893,
+ "learning_rate": 0.0006,
+ "loss": 4.210659980773926,
+ "step": 2347
+ },
+ {
+ "epoch": 32.61484716157205,
+ "grad_norm": 0.015135962516069412,
+ "learning_rate": 0.0006,
+ "loss": 4.204618453979492,
+ "step": 2348
+ },
+ {
+ "epoch": 32.62882096069869,
+ "grad_norm": 0.01601956970989704,
+ "learning_rate": 0.0006,
+ "loss": 4.145564556121826,
+ "step": 2349
+ },
+ {
+ "epoch": 32.64279475982533,
+ "grad_norm": 0.01517851185053587,
+ "learning_rate": 0.0006,
+ "loss": 4.051673889160156,
+ "step": 2350
+ },
+ {
+ "epoch": 32.65676855895197,
+ "grad_norm": 0.014894292689859867,
+ "learning_rate": 0.0006,
+ "loss": 4.184818267822266,
+ "step": 2351
+ },
+ {
+ "epoch": 32.670742358078606,
+ "grad_norm": 0.015720834955573082,
+ "learning_rate": 0.0006,
+ "loss": 4.220036506652832,
+ "step": 2352
+ },
+ {
+ "epoch": 32.68471615720524,
+ "grad_norm": 0.01573433168232441,
+ "learning_rate": 0.0006,
+ "loss": 4.076773166656494,
+ "step": 2353
+ },
+ {
+ "epoch": 32.698689956331876,
+ "grad_norm": 0.015828052535653114,
+ "learning_rate": 0.0006,
+ "loss": 4.116329669952393,
+ "step": 2354
+ },
+ {
+ "epoch": 32.712663755458514,
+ "grad_norm": 0.015462259761989117,
+ "learning_rate": 0.0006,
+ "loss": 4.1155242919921875,
+ "step": 2355
+ },
+ {
+ "epoch": 32.72663755458515,
+ "grad_norm": 0.015763849020004272,
+ "learning_rate": 0.0006,
+ "loss": 4.169687747955322,
+ "step": 2356
+ },
+ {
+ "epoch": 32.74061135371179,
+ "grad_norm": 0.014935145154595375,
+ "learning_rate": 0.0006,
+ "loss": 4.2090911865234375,
+ "step": 2357
+ },
+ {
+ "epoch": 32.75458515283843,
+ "grad_norm": 0.015119184739887714,
+ "learning_rate": 0.0006,
+ "loss": 4.131838798522949,
+ "step": 2358
+ },
+ {
+ "epoch": 32.76855895196506,
+ "grad_norm": 0.014833525754511356,
+ "learning_rate": 0.0006,
+ "loss": 4.17618465423584,
+ "step": 2359
+ },
+ {
+ "epoch": 32.7825327510917,
+ "grad_norm": 0.01588359661400318,
+ "learning_rate": 0.0006,
+ "loss": 4.163723945617676,
+ "step": 2360
+ },
+ {
+ "epoch": 32.79650655021834,
+ "grad_norm": 0.0168188214302063,
+ "learning_rate": 0.0006,
+ "loss": 4.178189277648926,
+ "step": 2361
+ },
+ {
+ "epoch": 32.81048034934498,
+ "grad_norm": 0.01726095750927925,
+ "learning_rate": 0.0006,
+ "loss": 4.144223690032959,
+ "step": 2362
+ },
+ {
+ "epoch": 32.82445414847162,
+ "grad_norm": 0.01701410301029682,
+ "learning_rate": 0.0006,
+ "loss": 4.209896087646484,
+ "step": 2363
+ },
+ {
+ "epoch": 32.838427947598255,
+ "grad_norm": 0.01682264916598797,
+ "learning_rate": 0.0006,
+ "loss": 4.26223087310791,
+ "step": 2364
+ },
+ {
+ "epoch": 32.852401746724894,
+ "grad_norm": 0.015597456134855747,
+ "learning_rate": 0.0006,
+ "loss": 4.292664051055908,
+ "step": 2365
+ },
+ {
+ "epoch": 32.866375545851525,
+ "grad_norm": 0.015258579514920712,
+ "learning_rate": 0.0006,
+ "loss": 4.165006637573242,
+ "step": 2366
+ },
+ {
+ "epoch": 32.880349344978164,
+ "grad_norm": 0.01662200503051281,
+ "learning_rate": 0.0006,
+ "loss": 4.309786319732666,
+ "step": 2367
+ },
+ {
+ "epoch": 32.8943231441048,
+ "grad_norm": 0.015691161155700684,
+ "learning_rate": 0.0006,
+ "loss": 4.167202949523926,
+ "step": 2368
+ },
+ {
+ "epoch": 32.90829694323144,
+ "grad_norm": 0.014870637096464634,
+ "learning_rate": 0.0006,
+ "loss": 4.142415523529053,
+ "step": 2369
+ },
+ {
+ "epoch": 32.92227074235808,
+ "grad_norm": 0.01567252166569233,
+ "learning_rate": 0.0006,
+ "loss": 4.2312092781066895,
+ "step": 2370
+ },
+ {
+ "epoch": 32.93624454148472,
+ "grad_norm": 0.014948616735637188,
+ "learning_rate": 0.0006,
+ "loss": 4.1514129638671875,
+ "step": 2371
+ },
+ {
+ "epoch": 32.95021834061135,
+ "grad_norm": 0.014654500409960747,
+ "learning_rate": 0.0006,
+ "loss": 4.205010890960693,
+ "step": 2372
+ },
+ {
+ "epoch": 32.96419213973799,
+ "grad_norm": 0.015059148892760277,
+ "learning_rate": 0.0006,
+ "loss": 4.148405075073242,
+ "step": 2373
+ },
+ {
+ "epoch": 32.97816593886463,
+ "grad_norm": 0.015026576817035675,
+ "learning_rate": 0.0006,
+ "loss": 4.182535171508789,
+ "step": 2374
+ },
+ {
+ "epoch": 32.992139737991266,
+ "grad_norm": 0.016394320875406265,
+ "learning_rate": 0.0006,
+ "loss": 4.129051208496094,
+ "step": 2375
+ },
+ {
+ "epoch": 33.0,
+ "grad_norm": 0.01867043413221836,
+ "learning_rate": 0.0006,
+ "loss": 4.185556411743164,
+ "step": 2376
+ },
+ {
+ "epoch": 33.0,
+ "eval_loss": 4.640405654907227,
+ "eval_runtime": 56.5861,
+ "eval_samples_per_second": 43.155,
+ "eval_steps_per_second": 1.361,
+ "step": 2376
+ },
+ {
+ "epoch": 33.01397379912664,
+ "grad_norm": 0.018626727163791656,
+ "learning_rate": 0.0006,
+ "loss": 4.143031597137451,
+ "step": 2377
+ },
+ {
+ "epoch": 33.02794759825328,
+ "grad_norm": 0.019283363595604897,
+ "learning_rate": 0.0006,
+ "loss": 4.117845058441162,
+ "step": 2378
+ },
+ {
+ "epoch": 33.041921397379916,
+ "grad_norm": 0.018185822293162346,
+ "learning_rate": 0.0006,
+ "loss": 4.076862812042236,
+ "step": 2379
+ },
+ {
+ "epoch": 33.05589519650655,
+ "grad_norm": 0.019411196932196617,
+ "learning_rate": 0.0006,
+ "loss": 4.124156475067139,
+ "step": 2380
+ },
+ {
+ "epoch": 33.069868995633186,
+ "grad_norm": 0.019248466938734055,
+ "learning_rate": 0.0006,
+ "loss": 4.201735019683838,
+ "step": 2381
+ },
+ {
+ "epoch": 33.083842794759825,
+ "grad_norm": 0.018856720998883247,
+ "learning_rate": 0.0006,
+ "loss": 4.00296688079834,
+ "step": 2382
+ },
+ {
+ "epoch": 33.09781659388646,
+ "grad_norm": 0.018554294481873512,
+ "learning_rate": 0.0006,
+ "loss": 4.270577430725098,
+ "step": 2383
+ },
+ {
+ "epoch": 33.1117903930131,
+ "grad_norm": 0.019946224987506866,
+ "learning_rate": 0.0006,
+ "loss": 4.119589328765869,
+ "step": 2384
+ },
+ {
+ "epoch": 33.12576419213974,
+ "grad_norm": 0.021395670250058174,
+ "learning_rate": 0.0006,
+ "loss": 4.128847122192383,
+ "step": 2385
+ },
+ {
+ "epoch": 33.13973799126637,
+ "grad_norm": 0.02009674906730652,
+ "learning_rate": 0.0006,
+ "loss": 4.183228969573975,
+ "step": 2386
+ },
+ {
+ "epoch": 33.15371179039301,
+ "grad_norm": 0.01820223033428192,
+ "learning_rate": 0.0006,
+ "loss": 4.131601810455322,
+ "step": 2387
+ },
+ {
+ "epoch": 33.16768558951965,
+ "grad_norm": 0.018653156235814095,
+ "learning_rate": 0.0006,
+ "loss": 4.116430282592773,
+ "step": 2388
+ },
+ {
+ "epoch": 33.18165938864629,
+ "grad_norm": 0.016828063875436783,
+ "learning_rate": 0.0006,
+ "loss": 4.054663181304932,
+ "step": 2389
+ },
+ {
+ "epoch": 33.19563318777293,
+ "grad_norm": 0.017025409266352654,
+ "learning_rate": 0.0006,
+ "loss": 4.110044956207275,
+ "step": 2390
+ },
+ {
+ "epoch": 33.209606986899566,
+ "grad_norm": 0.016608452424407005,
+ "learning_rate": 0.0006,
+ "loss": 4.102668285369873,
+ "step": 2391
+ },
+ {
+ "epoch": 33.223580786026204,
+ "grad_norm": 0.01743577979505062,
+ "learning_rate": 0.0006,
+ "loss": 4.166791915893555,
+ "step": 2392
+ },
+ {
+ "epoch": 33.237554585152836,
+ "grad_norm": 0.017253734171390533,
+ "learning_rate": 0.0006,
+ "loss": 4.217832088470459,
+ "step": 2393
+ },
+ {
+ "epoch": 33.251528384279474,
+ "grad_norm": 0.017937595024704933,
+ "learning_rate": 0.0006,
+ "loss": 4.2173967361450195,
+ "step": 2394
+ },
+ {
+ "epoch": 33.26550218340611,
+ "grad_norm": 0.01985993981361389,
+ "learning_rate": 0.0006,
+ "loss": 4.17849588394165,
+ "step": 2395
+ },
+ {
+ "epoch": 33.27947598253275,
+ "grad_norm": 0.019243355840444565,
+ "learning_rate": 0.0006,
+ "loss": 4.086523056030273,
+ "step": 2396
+ },
+ {
+ "epoch": 33.29344978165939,
+ "grad_norm": 0.01924416422843933,
+ "learning_rate": 0.0006,
+ "loss": 4.128574371337891,
+ "step": 2397
+ },
+ {
+ "epoch": 33.30742358078603,
+ "grad_norm": 0.01983758620917797,
+ "learning_rate": 0.0006,
+ "loss": 4.061581611633301,
+ "step": 2398
+ },
+ {
+ "epoch": 33.32139737991266,
+ "grad_norm": 0.01930462196469307,
+ "learning_rate": 0.0006,
+ "loss": 4.091731071472168,
+ "step": 2399
+ },
+ {
+ "epoch": 33.3353711790393,
+ "grad_norm": 0.017988424748182297,
+ "learning_rate": 0.0006,
+ "loss": 4.155599117279053,
+ "step": 2400
+ },
+ {
+ "epoch": 33.34934497816594,
+ "grad_norm": 0.015194767154753208,
+ "learning_rate": 0.0006,
+ "loss": 4.144134044647217,
+ "step": 2401
+ },
+ {
+ "epoch": 33.36331877729258,
+ "grad_norm": 0.01536081824451685,
+ "learning_rate": 0.0006,
+ "loss": 4.127272129058838,
+ "step": 2402
+ },
+ {
+ "epoch": 33.377292576419215,
+ "grad_norm": 0.015431297942996025,
+ "learning_rate": 0.0006,
+ "loss": 4.3213019371032715,
+ "step": 2403
+ },
+ {
+ "epoch": 33.391266375545854,
+ "grad_norm": 0.015740280970931053,
+ "learning_rate": 0.0006,
+ "loss": 4.121467590332031,
+ "step": 2404
+ },
+ {
+ "epoch": 33.40524017467249,
+ "grad_norm": 0.015379384160041809,
+ "learning_rate": 0.0006,
+ "loss": 4.095216751098633,
+ "step": 2405
+ },
+ {
+ "epoch": 33.419213973799124,
+ "grad_norm": 0.01528073474764824,
+ "learning_rate": 0.0006,
+ "loss": 4.13047981262207,
+ "step": 2406
+ },
+ {
+ "epoch": 33.43318777292576,
+ "grad_norm": 0.015739979222416878,
+ "learning_rate": 0.0006,
+ "loss": 4.2533135414123535,
+ "step": 2407
+ },
+ {
+ "epoch": 33.4471615720524,
+ "grad_norm": 0.01519764307886362,
+ "learning_rate": 0.0006,
+ "loss": 4.131782054901123,
+ "step": 2408
+ },
+ {
+ "epoch": 33.46113537117904,
+ "grad_norm": 0.015425996854901314,
+ "learning_rate": 0.0006,
+ "loss": 4.02811336517334,
+ "step": 2409
+ },
+ {
+ "epoch": 33.47510917030568,
+ "grad_norm": 0.015378830023109913,
+ "learning_rate": 0.0006,
+ "loss": 4.08652925491333,
+ "step": 2410
+ },
+ {
+ "epoch": 33.48908296943232,
+ "grad_norm": 0.01473075058311224,
+ "learning_rate": 0.0006,
+ "loss": 4.214364528656006,
+ "step": 2411
+ },
+ {
+ "epoch": 33.50305676855895,
+ "grad_norm": 0.015141637064516544,
+ "learning_rate": 0.0006,
+ "loss": 4.155712127685547,
+ "step": 2412
+ },
+ {
+ "epoch": 33.51703056768559,
+ "grad_norm": 0.01641593500971794,
+ "learning_rate": 0.0006,
+ "loss": 4.097907066345215,
+ "step": 2413
+ },
+ {
+ "epoch": 33.531004366812226,
+ "grad_norm": 0.015182029455900192,
+ "learning_rate": 0.0006,
+ "loss": 4.238499641418457,
+ "step": 2414
+ },
+ {
+ "epoch": 33.544978165938865,
+ "grad_norm": 0.01643732376396656,
+ "learning_rate": 0.0006,
+ "loss": 4.0661492347717285,
+ "step": 2415
+ },
+ {
+ "epoch": 33.5589519650655,
+ "grad_norm": 0.015614228323101997,
+ "learning_rate": 0.0006,
+ "loss": 4.120795249938965,
+ "step": 2416
+ },
+ {
+ "epoch": 33.57292576419214,
+ "grad_norm": 0.016066187992691994,
+ "learning_rate": 0.0006,
+ "loss": 4.096465110778809,
+ "step": 2417
+ },
+ {
+ "epoch": 33.58689956331878,
+ "grad_norm": 0.017214342951774597,
+ "learning_rate": 0.0006,
+ "loss": 4.242628574371338,
+ "step": 2418
+ },
+ {
+ "epoch": 33.60087336244541,
+ "grad_norm": 0.01787586510181427,
+ "learning_rate": 0.0006,
+ "loss": 4.149589538574219,
+ "step": 2419
+ },
+ {
+ "epoch": 33.61484716157205,
+ "grad_norm": 0.019455241039395332,
+ "learning_rate": 0.0006,
+ "loss": 4.267796039581299,
+ "step": 2420
+ },
+ {
+ "epoch": 33.62882096069869,
+ "grad_norm": 0.01984674669802189,
+ "learning_rate": 0.0006,
+ "loss": 4.175220966339111,
+ "step": 2421
+ },
+ {
+ "epoch": 33.64279475982533,
+ "grad_norm": 0.021933717653155327,
+ "learning_rate": 0.0006,
+ "loss": 4.106647491455078,
+ "step": 2422
+ },
+ {
+ "epoch": 33.65676855895197,
+ "grad_norm": 0.02123589813709259,
+ "learning_rate": 0.0006,
+ "loss": 4.204996109008789,
+ "step": 2423
+ },
+ {
+ "epoch": 33.670742358078606,
+ "grad_norm": 0.019814589992165565,
+ "learning_rate": 0.0006,
+ "loss": 4.186171531677246,
+ "step": 2424
+ },
+ {
+ "epoch": 33.68471615720524,
+ "grad_norm": 0.02059728465974331,
+ "learning_rate": 0.0006,
+ "loss": 4.170509338378906,
+ "step": 2425
+ },
+ {
+ "epoch": 33.698689956331876,
+ "grad_norm": 0.018483608961105347,
+ "learning_rate": 0.0006,
+ "loss": 4.085585594177246,
+ "step": 2426
+ },
+ {
+ "epoch": 33.712663755458514,
+ "grad_norm": 0.018403852358460426,
+ "learning_rate": 0.0006,
+ "loss": 4.2551774978637695,
+ "step": 2427
+ },
+ {
+ "epoch": 33.72663755458515,
+ "grad_norm": 0.018653830513358116,
+ "learning_rate": 0.0006,
+ "loss": 4.091766357421875,
+ "step": 2428
+ },
+ {
+ "epoch": 33.74061135371179,
+ "grad_norm": 0.017692334949970245,
+ "learning_rate": 0.0006,
+ "loss": 4.100838661193848,
+ "step": 2429
+ },
+ {
+ "epoch": 33.75458515283843,
+ "grad_norm": 0.016029126942157745,
+ "learning_rate": 0.0006,
+ "loss": 4.1530680656433105,
+ "step": 2430
+ },
+ {
+ "epoch": 33.76855895196506,
+ "grad_norm": 0.016095684841275215,
+ "learning_rate": 0.0006,
+ "loss": 4.20544958114624,
+ "step": 2431
+ },
+ {
+ "epoch": 33.7825327510917,
+ "grad_norm": 0.016375798732042313,
+ "learning_rate": 0.0006,
+ "loss": 4.199177265167236,
+ "step": 2432
+ },
+ {
+ "epoch": 33.79650655021834,
+ "grad_norm": 0.01590786688029766,
+ "learning_rate": 0.0006,
+ "loss": 4.200799465179443,
+ "step": 2433
+ },
+ {
+ "epoch": 33.81048034934498,
+ "grad_norm": 0.014896097593009472,
+ "learning_rate": 0.0006,
+ "loss": 4.185850143432617,
+ "step": 2434
+ },
+ {
+ "epoch": 33.82445414847162,
+ "grad_norm": 0.01438513956964016,
+ "learning_rate": 0.0006,
+ "loss": 4.214962005615234,
+ "step": 2435
+ },
+ {
+ "epoch": 33.838427947598255,
+ "grad_norm": 0.014459339901804924,
+ "learning_rate": 0.0006,
+ "loss": 4.196972370147705,
+ "step": 2436
+ },
+ {
+ "epoch": 33.852401746724894,
+ "grad_norm": 0.014782169833779335,
+ "learning_rate": 0.0006,
+ "loss": 4.0885910987854,
+ "step": 2437
+ },
+ {
+ "epoch": 33.866375545851525,
+ "grad_norm": 0.01520370040088892,
+ "learning_rate": 0.0006,
+ "loss": 4.0323262214660645,
+ "step": 2438
+ },
+ {
+ "epoch": 33.880349344978164,
+ "grad_norm": 0.014056282117962837,
+ "learning_rate": 0.0006,
+ "loss": 4.1386518478393555,
+ "step": 2439
+ },
+ {
+ "epoch": 33.8943231441048,
+ "grad_norm": 0.013567990623414516,
+ "learning_rate": 0.0006,
+ "loss": 4.095754623413086,
+ "step": 2440
+ },
+ {
+ "epoch": 33.90829694323144,
+ "grad_norm": 0.013306716457009315,
+ "learning_rate": 0.0006,
+ "loss": 4.1623077392578125,
+ "step": 2441
+ },
+ {
+ "epoch": 33.92227074235808,
+ "grad_norm": 0.014047150500118732,
+ "learning_rate": 0.0006,
+ "loss": 4.142541885375977,
+ "step": 2442
+ },
+ {
+ "epoch": 33.93624454148472,
+ "grad_norm": 0.01377617847174406,
+ "learning_rate": 0.0006,
+ "loss": 4.268551826477051,
+ "step": 2443
+ },
+ {
+ "epoch": 33.95021834061135,
+ "grad_norm": 0.014671621844172478,
+ "learning_rate": 0.0006,
+ "loss": 4.198196887969971,
+ "step": 2444
+ },
+ {
+ "epoch": 33.96419213973799,
+ "grad_norm": 0.014806744642555714,
+ "learning_rate": 0.0006,
+ "loss": 4.057899475097656,
+ "step": 2445
+ },
+ {
+ "epoch": 33.97816593886463,
+ "grad_norm": 0.015734272077679634,
+ "learning_rate": 0.0006,
+ "loss": 4.163505554199219,
+ "step": 2446
+ },
+ {
+ "epoch": 33.992139737991266,
+ "grad_norm": 0.015250146389007568,
+ "learning_rate": 0.0006,
+ "loss": 4.076294898986816,
+ "step": 2447
+ },
+ {
+ "epoch": 34.0,
+ "grad_norm": 0.016733702272176743,
+ "learning_rate": 0.0006,
+ "loss": 4.138420104980469,
+ "step": 2448
+ },
+ {
+ "epoch": 34.0,
+ "eval_loss": 4.542503356933594,
+ "eval_runtime": 56.983,
+ "eval_samples_per_second": 42.855,
+ "eval_steps_per_second": 1.351,
+ "step": 2448
+ },
+ {
+ "epoch": 34.01397379912664,
+ "grad_norm": 0.015866613015532494,
+ "learning_rate": 0.0006,
+ "loss": 4.032081127166748,
+ "step": 2449
+ },
+ {
+ "epoch": 34.02794759825328,
+ "grad_norm": 0.01946570910513401,
+ "learning_rate": 0.0006,
+ "loss": 4.122320175170898,
+ "step": 2450
+ },
+ {
+ "epoch": 34.041921397379916,
+ "grad_norm": 0.020265938714146614,
+ "learning_rate": 0.0006,
+ "loss": 4.06460428237915,
+ "step": 2451
+ },
+ {
+ "epoch": 34.05589519650655,
+ "grad_norm": 0.018887171521782875,
+ "learning_rate": 0.0006,
+ "loss": 4.096906661987305,
+ "step": 2452
+ },
+ {
+ "epoch": 34.069868995633186,
+ "grad_norm": 0.017304692417383194,
+ "learning_rate": 0.0006,
+ "loss": 4.137261390686035,
+ "step": 2453
+ },
+ {
+ "epoch": 34.083842794759825,
+ "grad_norm": 0.017029082402586937,
+ "learning_rate": 0.0006,
+ "loss": 4.081112861633301,
+ "step": 2454
+ },
+ {
+ "epoch": 34.09781659388646,
+ "grad_norm": 0.016414813697338104,
+ "learning_rate": 0.0006,
+ "loss": 4.140783309936523,
+ "step": 2455
+ },
+ {
+ "epoch": 34.1117903930131,
+ "grad_norm": 0.018238281831145287,
+ "learning_rate": 0.0006,
+ "loss": 4.161116600036621,
+ "step": 2456
+ },
+ {
+ "epoch": 34.12576419213974,
+ "grad_norm": 0.019929232075810432,
+ "learning_rate": 0.0006,
+ "loss": 4.243722438812256,
+ "step": 2457
+ },
+ {
+ "epoch": 34.13973799126637,
+ "grad_norm": 0.019375786185264587,
+ "learning_rate": 0.0006,
+ "loss": 4.134873390197754,
+ "step": 2458
+ },
+ {
+ "epoch": 34.15371179039301,
+ "grad_norm": 0.018151137977838516,
+ "learning_rate": 0.0006,
+ "loss": 4.098745346069336,
+ "step": 2459
+ },
+ {
+ "epoch": 34.16768558951965,
+ "grad_norm": 0.017054934054613113,
+ "learning_rate": 0.0006,
+ "loss": 4.194662570953369,
+ "step": 2460
+ },
+ {
+ "epoch": 34.18165938864629,
+ "grad_norm": 0.018220555037260056,
+ "learning_rate": 0.0006,
+ "loss": 4.128866195678711,
+ "step": 2461
+ },
+ {
+ "epoch": 34.19563318777293,
+ "grad_norm": 0.01692879945039749,
+ "learning_rate": 0.0006,
+ "loss": 4.062477111816406,
+ "step": 2462
+ },
+ {
+ "epoch": 34.209606986899566,
+ "grad_norm": 0.01674199476838112,
+ "learning_rate": 0.0006,
+ "loss": 4.186149597167969,
+ "step": 2463
+ },
+ {
+ "epoch": 34.223580786026204,
+ "grad_norm": 0.016527455300092697,
+ "learning_rate": 0.0006,
+ "loss": 4.0734028816223145,
+ "step": 2464
+ },
+ {
+ "epoch": 34.237554585152836,
+ "grad_norm": 0.016448386013507843,
+ "learning_rate": 0.0006,
+ "loss": 4.098896026611328,
+ "step": 2465
+ },
+ {
+ "epoch": 34.251528384279474,
+ "grad_norm": 0.014644953422248363,
+ "learning_rate": 0.0006,
+ "loss": 4.1031904220581055,
+ "step": 2466
+ },
+ {
+ "epoch": 34.26550218340611,
+ "grad_norm": 0.016649341210722923,
+ "learning_rate": 0.0006,
+ "loss": 4.148791313171387,
+ "step": 2467
+ },
+ {
+ "epoch": 34.27947598253275,
+ "grad_norm": 0.017442714422941208,
+ "learning_rate": 0.0006,
+ "loss": 4.070799827575684,
+ "step": 2468
+ },
+ {
+ "epoch": 34.29344978165939,
+ "grad_norm": 0.017852013930678368,
+ "learning_rate": 0.0006,
+ "loss": 4.031924247741699,
+ "step": 2469
+ },
+ {
+ "epoch": 34.30742358078603,
+ "grad_norm": 0.016371270641684532,
+ "learning_rate": 0.0006,
+ "loss": 4.149420738220215,
+ "step": 2470
+ },
+ {
+ "epoch": 34.32139737991266,
+ "grad_norm": 0.015846814960241318,
+ "learning_rate": 0.0006,
+ "loss": 4.15427303314209,
+ "step": 2471
+ },
+ {
+ "epoch": 34.3353711790393,
+ "grad_norm": 0.017411163076758385,
+ "learning_rate": 0.0006,
+ "loss": 4.1304192543029785,
+ "step": 2472
+ },
+ {
+ "epoch": 34.34934497816594,
+ "grad_norm": 0.01780046708881855,
+ "learning_rate": 0.0006,
+ "loss": 4.191732406616211,
+ "step": 2473
+ },
+ {
+ "epoch": 34.36331877729258,
+ "grad_norm": 0.01958383060991764,
+ "learning_rate": 0.0006,
+ "loss": 4.068355083465576,
+ "step": 2474
+ },
+ {
+ "epoch": 34.377292576419215,
+ "grad_norm": 0.022173317149281502,
+ "learning_rate": 0.0006,
+ "loss": 4.182929039001465,
+ "step": 2475
+ },
+ {
+ "epoch": 34.391266375545854,
+ "grad_norm": 0.022389156743884087,
+ "learning_rate": 0.0006,
+ "loss": 4.110218048095703,
+ "step": 2476
+ },
+ {
+ "epoch": 34.40524017467249,
+ "grad_norm": 0.022478511556982994,
+ "learning_rate": 0.0006,
+ "loss": 4.095479488372803,
+ "step": 2477
+ },
+ {
+ "epoch": 34.419213973799124,
+ "grad_norm": 0.019504647701978683,
+ "learning_rate": 0.0006,
+ "loss": 4.101964950561523,
+ "step": 2478
+ },
+ {
+ "epoch": 34.43318777292576,
+ "grad_norm": 0.019378576427698135,
+ "learning_rate": 0.0006,
+ "loss": 4.14620304107666,
+ "step": 2479
+ },
+ {
+ "epoch": 34.4471615720524,
+ "grad_norm": 0.017993204295635223,
+ "learning_rate": 0.0006,
+ "loss": 4.183492660522461,
+ "step": 2480
+ },
+ {
+ "epoch": 34.46113537117904,
+ "grad_norm": 0.01814805157482624,
+ "learning_rate": 0.0006,
+ "loss": 4.151247978210449,
+ "step": 2481
+ },
+ {
+ "epoch": 34.47510917030568,
+ "grad_norm": 0.018622448667883873,
+ "learning_rate": 0.0006,
+ "loss": 4.020938396453857,
+ "step": 2482
+ },
+ {
+ "epoch": 34.48908296943232,
+ "grad_norm": 0.018360132351517677,
+ "learning_rate": 0.0006,
+ "loss": 4.064734935760498,
+ "step": 2483
+ },
+ {
+ "epoch": 34.50305676855895,
+ "grad_norm": 0.017715711146593094,
+ "learning_rate": 0.0006,
+ "loss": 4.185687065124512,
+ "step": 2484
+ },
+ {
+ "epoch": 34.51703056768559,
+ "grad_norm": 0.01714295893907547,
+ "learning_rate": 0.0006,
+ "loss": 4.1044721603393555,
+ "step": 2485
+ },
+ {
+ "epoch": 34.531004366812226,
+ "grad_norm": 0.016943803057074547,
+ "learning_rate": 0.0006,
+ "loss": 4.106837272644043,
+ "step": 2486
+ },
+ {
+ "epoch": 34.544978165938865,
+ "grad_norm": 0.0182423684746027,
+ "learning_rate": 0.0006,
+ "loss": 4.1041340827941895,
+ "step": 2487
+ },
+ {
+ "epoch": 34.5589519650655,
+ "grad_norm": 0.01636587455868721,
+ "learning_rate": 0.0006,
+ "loss": 3.9735546112060547,
+ "step": 2488
+ },
+ {
+ "epoch": 34.57292576419214,
+ "grad_norm": 0.014593546278774738,
+ "learning_rate": 0.0006,
+ "loss": 4.041980266571045,
+ "step": 2489
+ },
+ {
+ "epoch": 34.58689956331878,
+ "grad_norm": 0.015436965972185135,
+ "learning_rate": 0.0006,
+ "loss": 4.301148414611816,
+ "step": 2490
+ },
+ {
+ "epoch": 34.60087336244541,
+ "grad_norm": 0.015075269155204296,
+ "learning_rate": 0.0006,
+ "loss": 4.118527412414551,
+ "step": 2491
+ },
+ {
+ "epoch": 34.61484716157205,
+ "grad_norm": 0.01701514609158039,
+ "learning_rate": 0.0006,
+ "loss": 4.245968818664551,
+ "step": 2492
+ },
+ {
+ "epoch": 34.62882096069869,
+ "grad_norm": 0.018047887831926346,
+ "learning_rate": 0.0006,
+ "loss": 4.227261066436768,
+ "step": 2493
+ },
+ {
+ "epoch": 34.64279475982533,
+ "grad_norm": 0.018499085679650307,
+ "learning_rate": 0.0006,
+ "loss": 4.121360778808594,
+ "step": 2494
+ },
+ {
+ "epoch": 34.65676855895197,
+ "grad_norm": 0.018987711519002914,
+ "learning_rate": 0.0006,
+ "loss": 4.131564617156982,
+ "step": 2495
+ },
+ {
+ "epoch": 34.670742358078606,
+ "grad_norm": 0.01834476925432682,
+ "learning_rate": 0.0006,
+ "loss": 4.224941253662109,
+ "step": 2496
+ },
+ {
+ "epoch": 34.68471615720524,
+ "grad_norm": 0.017507942393422127,
+ "learning_rate": 0.0006,
+ "loss": 4.1394548416137695,
+ "step": 2497
+ },
+ {
+ "epoch": 34.698689956331876,
+ "grad_norm": 0.01680580899119377,
+ "learning_rate": 0.0006,
+ "loss": 4.1073431968688965,
+ "step": 2498
+ },
+ {
+ "epoch": 34.712663755458514,
+ "grad_norm": 0.015233514830470085,
+ "learning_rate": 0.0006,
+ "loss": 4.136994361877441,
+ "step": 2499
+ },
+ {
+ "epoch": 34.72663755458515,
+ "grad_norm": 0.015482917428016663,
+ "learning_rate": 0.0006,
+ "loss": 4.110719680786133,
+ "step": 2500
+ },
+ {
+ "epoch": 34.74061135371179,
+ "grad_norm": 0.015940461307764053,
+ "learning_rate": 0.0006,
+ "loss": 4.207851886749268,
+ "step": 2501
+ },
+ {
+ "epoch": 34.75458515283843,
+ "grad_norm": 0.015355597250163555,
+ "learning_rate": 0.0006,
+ "loss": 4.169469833374023,
+ "step": 2502
+ },
+ {
+ "epoch": 34.76855895196506,
+ "grad_norm": 0.015974000096321106,
+ "learning_rate": 0.0006,
+ "loss": 4.110061168670654,
+ "step": 2503
+ },
+ {
+ "epoch": 34.7825327510917,
+ "grad_norm": 0.016088927164673805,
+ "learning_rate": 0.0006,
+ "loss": 4.155452728271484,
+ "step": 2504
+ },
+ {
+ "epoch": 34.79650655021834,
+ "grad_norm": 0.016156120225787163,
+ "learning_rate": 0.0006,
+ "loss": 4.178092956542969,
+ "step": 2505
+ },
+ {
+ "epoch": 34.81048034934498,
+ "grad_norm": 0.014928647316992283,
+ "learning_rate": 0.0006,
+ "loss": 4.05457878112793,
+ "step": 2506
+ },
+ {
+ "epoch": 34.82445414847162,
+ "grad_norm": 0.015289328061044216,
+ "learning_rate": 0.0006,
+ "loss": 4.18869686126709,
+ "step": 2507
+ },
+ {
+ "epoch": 34.838427947598255,
+ "grad_norm": 0.015799039974808693,
+ "learning_rate": 0.0006,
+ "loss": 4.106073379516602,
+ "step": 2508
+ },
+ {
+ "epoch": 34.852401746724894,
+ "grad_norm": 0.015824418514966965,
+ "learning_rate": 0.0006,
+ "loss": 4.0634589195251465,
+ "step": 2509
+ },
+ {
+ "epoch": 34.866375545851525,
+ "grad_norm": 0.015321656130254269,
+ "learning_rate": 0.0006,
+ "loss": 4.093090057373047,
+ "step": 2510
+ },
+ {
+ "epoch": 34.880349344978164,
+ "grad_norm": 0.014380166307091713,
+ "learning_rate": 0.0006,
+ "loss": 4.1750640869140625,
+ "step": 2511
+ },
+ {
+ "epoch": 34.8943231441048,
+ "grad_norm": 0.014962567947804928,
+ "learning_rate": 0.0006,
+ "loss": 4.185277938842773,
+ "step": 2512
+ },
+ {
+ "epoch": 34.90829694323144,
+ "grad_norm": 0.014467543922364712,
+ "learning_rate": 0.0006,
+ "loss": 4.1600799560546875,
+ "step": 2513
+ },
+ {
+ "epoch": 34.92227074235808,
+ "grad_norm": 0.01545181218534708,
+ "learning_rate": 0.0006,
+ "loss": 4.096299648284912,
+ "step": 2514
+ },
+ {
+ "epoch": 34.93624454148472,
+ "grad_norm": 0.01605340465903282,
+ "learning_rate": 0.0006,
+ "loss": 4.203323841094971,
+ "step": 2515
+ },
+ {
+ "epoch": 34.95021834061135,
+ "grad_norm": 0.015203815884888172,
+ "learning_rate": 0.0006,
+ "loss": 4.201385021209717,
+ "step": 2516
+ },
+ {
+ "epoch": 34.96419213973799,
+ "grad_norm": 0.015149409882724285,
+ "learning_rate": 0.0006,
+ "loss": 4.088164329528809,
+ "step": 2517
+ },
+ {
+ "epoch": 34.97816593886463,
+ "grad_norm": 0.014885558746755123,
+ "learning_rate": 0.0006,
+ "loss": 4.073911666870117,
+ "step": 2518
+ },
+ {
+ "epoch": 34.992139737991266,
+ "grad_norm": 0.015003419481217861,
+ "learning_rate": 0.0006,
+ "loss": 4.157113552093506,
+ "step": 2519
+ },
+ {
+ "epoch": 35.0,
+ "grad_norm": 0.01641916297376156,
+ "learning_rate": 0.0006,
+ "loss": 3.970052719116211,
+ "step": 2520
+ },
+ {
+ "epoch": 35.0,
+ "eval_loss": 4.641692161560059,
+ "eval_runtime": 56.8836,
+ "eval_samples_per_second": 42.93,
+ "eval_steps_per_second": 1.354,
+ "step": 2520
+ },
+ {
+ "epoch": 35.01397379912664,
+ "grad_norm": 0.015894673764705658,
+ "learning_rate": 0.0006,
+ "loss": 4.104437351226807,
+ "step": 2521
+ },
+ {
+ "epoch": 35.02794759825328,
+ "grad_norm": 0.0173867866396904,
+ "learning_rate": 0.0006,
+ "loss": 4.091726303100586,
+ "step": 2522
+ },
+ {
+ "epoch": 35.041921397379916,
+ "grad_norm": 0.019272832199931145,
+ "learning_rate": 0.0006,
+ "loss": 4.217905044555664,
+ "step": 2523
+ },
+ {
+ "epoch": 35.05589519650655,
+ "grad_norm": 0.021820371970534325,
+ "learning_rate": 0.0006,
+ "loss": 4.111428260803223,
+ "step": 2524
+ },
+ {
+ "epoch": 35.069868995633186,
+ "grad_norm": 0.02340741828083992,
+ "learning_rate": 0.0006,
+ "loss": 4.115771293640137,
+ "step": 2525
+ },
+ {
+ "epoch": 35.083842794759825,
+ "grad_norm": 0.023760691285133362,
+ "learning_rate": 0.0006,
+ "loss": 4.132383346557617,
+ "step": 2526
+ },
+ {
+ "epoch": 35.09781659388646,
+ "grad_norm": 0.02073468267917633,
+ "learning_rate": 0.0006,
+ "loss": 4.272691249847412,
+ "step": 2527
+ },
+ {
+ "epoch": 35.1117903930131,
+ "grad_norm": 0.02197718247771263,
+ "learning_rate": 0.0006,
+ "loss": 4.105586528778076,
+ "step": 2528
+ },
+ {
+ "epoch": 35.12576419213974,
+ "grad_norm": 0.0200254675000906,
+ "learning_rate": 0.0006,
+ "loss": 3.968820095062256,
+ "step": 2529
+ },
+ {
+ "epoch": 35.13973799126637,
+ "grad_norm": 0.01929587684571743,
+ "learning_rate": 0.0006,
+ "loss": 4.045468807220459,
+ "step": 2530
+ },
+ {
+ "epoch": 35.15371179039301,
+ "grad_norm": 0.01886162906885147,
+ "learning_rate": 0.0006,
+ "loss": 4.185571670532227,
+ "step": 2531
+ },
+ {
+ "epoch": 35.16768558951965,
+ "grad_norm": 0.019789839163422585,
+ "learning_rate": 0.0006,
+ "loss": 3.9734551906585693,
+ "step": 2532
+ },
+ {
+ "epoch": 35.18165938864629,
+ "grad_norm": 0.0198142658919096,
+ "learning_rate": 0.0006,
+ "loss": 4.2662506103515625,
+ "step": 2533
+ },
+ {
+ "epoch": 35.19563318777293,
+ "grad_norm": 0.02087012305855751,
+ "learning_rate": 0.0006,
+ "loss": 4.175931930541992,
+ "step": 2534
+ },
+ {
+ "epoch": 35.209606986899566,
+ "grad_norm": 0.020294535905122757,
+ "learning_rate": 0.0006,
+ "loss": 4.060324668884277,
+ "step": 2535
+ },
+ {
+ "epoch": 35.223580786026204,
+ "grad_norm": 0.01908983662724495,
+ "learning_rate": 0.0006,
+ "loss": 4.217939853668213,
+ "step": 2536
+ },
+ {
+ "epoch": 35.237554585152836,
+ "grad_norm": 0.0178892333060503,
+ "learning_rate": 0.0006,
+ "loss": 4.190216541290283,
+ "step": 2537
+ },
+ {
+ "epoch": 35.251528384279474,
+ "grad_norm": 0.017892315983772278,
+ "learning_rate": 0.0006,
+ "loss": 4.067419052124023,
+ "step": 2538
+ },
+ {
+ "epoch": 35.26550218340611,
+ "grad_norm": 0.018556250259280205,
+ "learning_rate": 0.0006,
+ "loss": 4.105958461761475,
+ "step": 2539
+ },
+ {
+ "epoch": 35.27947598253275,
+ "grad_norm": 0.01944422349333763,
+ "learning_rate": 0.0006,
+ "loss": 4.155775547027588,
+ "step": 2540
+ },
+ {
+ "epoch": 35.29344978165939,
+ "grad_norm": 0.017333146184682846,
+ "learning_rate": 0.0006,
+ "loss": 4.244513511657715,
+ "step": 2541
+ },
+ {
+ "epoch": 35.30742358078603,
+ "grad_norm": 0.01639244332909584,
+ "learning_rate": 0.0006,
+ "loss": 4.167815685272217,
+ "step": 2542
+ },
+ {
+ "epoch": 35.32139737991266,
+ "grad_norm": 0.01610143668949604,
+ "learning_rate": 0.0006,
+ "loss": 4.121486663818359,
+ "step": 2543
+ },
+ {
+ "epoch": 35.3353711790393,
+ "grad_norm": 0.01656411588191986,
+ "learning_rate": 0.0006,
+ "loss": 4.1755499839782715,
+ "step": 2544
+ },
+ {
+ "epoch": 35.34934497816594,
+ "grad_norm": 0.015075297094881535,
+ "learning_rate": 0.0006,
+ "loss": 4.102992534637451,
+ "step": 2545
+ },
+ {
+ "epoch": 35.36331877729258,
+ "grad_norm": 0.01538170501589775,
+ "learning_rate": 0.0006,
+ "loss": 4.161603927612305,
+ "step": 2546
+ },
+ {
+ "epoch": 35.377292576419215,
+ "grad_norm": 0.01483312901109457,
+ "learning_rate": 0.0006,
+ "loss": 4.085355758666992,
+ "step": 2547
+ },
+ {
+ "epoch": 35.391266375545854,
+ "grad_norm": 0.014894738793373108,
+ "learning_rate": 0.0006,
+ "loss": 4.083585262298584,
+ "step": 2548
+ },
+ {
+ "epoch": 35.40524017467249,
+ "grad_norm": 0.01475780550390482,
+ "learning_rate": 0.0006,
+ "loss": 4.012645721435547,
+ "step": 2549
+ },
+ {
+ "epoch": 35.419213973799124,
+ "grad_norm": 0.014531653374433517,
+ "learning_rate": 0.0006,
+ "loss": 4.197113990783691,
+ "step": 2550
+ },
+ {
+ "epoch": 35.43318777292576,
+ "grad_norm": 0.015316493809223175,
+ "learning_rate": 0.0006,
+ "loss": 4.227970123291016,
+ "step": 2551
+ },
+ {
+ "epoch": 35.4471615720524,
+ "grad_norm": 0.014685571193695068,
+ "learning_rate": 0.0006,
+ "loss": 4.089415550231934,
+ "step": 2552
+ },
+ {
+ "epoch": 35.46113537117904,
+ "grad_norm": 0.015337035991251469,
+ "learning_rate": 0.0006,
+ "loss": 4.080583572387695,
+ "step": 2553
+ },
+ {
+ "epoch": 35.47510917030568,
+ "grad_norm": 0.016754208132624626,
+ "learning_rate": 0.0006,
+ "loss": 4.163935661315918,
+ "step": 2554
+ },
+ {
+ "epoch": 35.48908296943232,
+ "grad_norm": 0.017535798251628876,
+ "learning_rate": 0.0006,
+ "loss": 4.093027591705322,
+ "step": 2555
+ },
+ {
+ "epoch": 35.50305676855895,
+ "grad_norm": 0.017336975783109665,
+ "learning_rate": 0.0006,
+ "loss": 4.127961158752441,
+ "step": 2556
+ },
+ {
+ "epoch": 35.51703056768559,
+ "grad_norm": 0.016242152079939842,
+ "learning_rate": 0.0006,
+ "loss": 4.155593395233154,
+ "step": 2557
+ },
+ {
+ "epoch": 35.531004366812226,
+ "grad_norm": 0.015121322125196457,
+ "learning_rate": 0.0006,
+ "loss": 3.984004020690918,
+ "step": 2558
+ },
+ {
+ "epoch": 35.544978165938865,
+ "grad_norm": 0.015178884379565716,
+ "learning_rate": 0.0006,
+ "loss": 4.111564636230469,
+ "step": 2559
+ },
+ {
+ "epoch": 35.5589519650655,
+ "grad_norm": 0.014777131378650665,
+ "learning_rate": 0.0006,
+ "loss": 4.041402816772461,
+ "step": 2560
+ },
+ {
+ "epoch": 35.57292576419214,
+ "grad_norm": 0.015197676606476307,
+ "learning_rate": 0.0006,
+ "loss": 4.133706092834473,
+ "step": 2561
+ },
+ {
+ "epoch": 35.58689956331878,
+ "grad_norm": 0.015905972570180893,
+ "learning_rate": 0.0006,
+ "loss": 4.1069135665893555,
+ "step": 2562
+ },
+ {
+ "epoch": 35.60087336244541,
+ "grad_norm": 0.015942877158522606,
+ "learning_rate": 0.0006,
+ "loss": 4.0720624923706055,
+ "step": 2563
+ },
+ {
+ "epoch": 35.61484716157205,
+ "grad_norm": 0.016697583720088005,
+ "learning_rate": 0.0006,
+ "loss": 4.167448043823242,
+ "step": 2564
+ },
+ {
+ "epoch": 35.62882096069869,
+ "grad_norm": 0.015603577718138695,
+ "learning_rate": 0.0006,
+ "loss": 4.127327919006348,
+ "step": 2565
+ },
+ {
+ "epoch": 35.64279475982533,
+ "grad_norm": 0.017153460532426834,
+ "learning_rate": 0.0006,
+ "loss": 4.056543350219727,
+ "step": 2566
+ },
+ {
+ "epoch": 35.65676855895197,
+ "grad_norm": 0.016731958836317062,
+ "learning_rate": 0.0006,
+ "loss": 4.0681843757629395,
+ "step": 2567
+ },
+ {
+ "epoch": 35.670742358078606,
+ "grad_norm": 0.016167359426617622,
+ "learning_rate": 0.0006,
+ "loss": 4.174992084503174,
+ "step": 2568
+ },
+ {
+ "epoch": 35.68471615720524,
+ "grad_norm": 0.01615108922123909,
+ "learning_rate": 0.0006,
+ "loss": 4.1383161544799805,
+ "step": 2569
+ },
+ {
+ "epoch": 35.698689956331876,
+ "grad_norm": 0.015750816091895103,
+ "learning_rate": 0.0006,
+ "loss": 4.031595706939697,
+ "step": 2570
+ },
+ {
+ "epoch": 35.712663755458514,
+ "grad_norm": 0.015416189096868038,
+ "learning_rate": 0.0006,
+ "loss": 4.067498207092285,
+ "step": 2571
+ },
+ {
+ "epoch": 35.72663755458515,
+ "grad_norm": 0.01570792682468891,
+ "learning_rate": 0.0006,
+ "loss": 4.1338300704956055,
+ "step": 2572
+ },
+ {
+ "epoch": 35.74061135371179,
+ "grad_norm": 0.016367752104997635,
+ "learning_rate": 0.0006,
+ "loss": 4.152523040771484,
+ "step": 2573
+ },
+ {
+ "epoch": 35.75458515283843,
+ "grad_norm": 0.016349803656339645,
+ "learning_rate": 0.0006,
+ "loss": 4.101985454559326,
+ "step": 2574
+ },
+ {
+ "epoch": 35.76855895196506,
+ "grad_norm": 0.01663660630583763,
+ "learning_rate": 0.0006,
+ "loss": 4.190130710601807,
+ "step": 2575
+ },
+ {
+ "epoch": 35.7825327510917,
+ "grad_norm": 0.015599170699715614,
+ "learning_rate": 0.0006,
+ "loss": 4.0792436599731445,
+ "step": 2576
+ },
+ {
+ "epoch": 35.79650655021834,
+ "grad_norm": 0.015472427010536194,
+ "learning_rate": 0.0006,
+ "loss": 4.186446666717529,
+ "step": 2577
+ },
+ {
+ "epoch": 35.81048034934498,
+ "grad_norm": 0.015520439483225346,
+ "learning_rate": 0.0006,
+ "loss": 4.246635437011719,
+ "step": 2578
+ },
+ {
+ "epoch": 35.82445414847162,
+ "grad_norm": 0.016998281702399254,
+ "learning_rate": 0.0006,
+ "loss": 4.13435173034668,
+ "step": 2579
+ },
+ {
+ "epoch": 35.838427947598255,
+ "grad_norm": 0.016290361061692238,
+ "learning_rate": 0.0006,
+ "loss": 4.107428073883057,
+ "step": 2580
+ },
+ {
+ "epoch": 35.852401746724894,
+ "grad_norm": 0.0165516659617424,
+ "learning_rate": 0.0006,
+ "loss": 4.1612138748168945,
+ "step": 2581
+ },
+ {
+ "epoch": 35.866375545851525,
+ "grad_norm": 0.01736091636121273,
+ "learning_rate": 0.0006,
+ "loss": 4.1072869300842285,
+ "step": 2582
+ },
+ {
+ "epoch": 35.880349344978164,
+ "grad_norm": 0.01787031814455986,
+ "learning_rate": 0.0006,
+ "loss": 4.1076812744140625,
+ "step": 2583
+ },
+ {
+ "epoch": 35.8943231441048,
+ "grad_norm": 0.015437985770404339,
+ "learning_rate": 0.0006,
+ "loss": 4.097488880157471,
+ "step": 2584
+ },
+ {
+ "epoch": 35.90829694323144,
+ "grad_norm": 0.016306741163134575,
+ "learning_rate": 0.0006,
+ "loss": 4.105017185211182,
+ "step": 2585
+ },
+ {
+ "epoch": 35.92227074235808,
+ "grad_norm": 0.015723245218396187,
+ "learning_rate": 0.0006,
+ "loss": 4.205283164978027,
+ "step": 2586
+ },
+ {
+ "epoch": 35.93624454148472,
+ "grad_norm": 0.014626456424593925,
+ "learning_rate": 0.0006,
+ "loss": 4.086552619934082,
+ "step": 2587
+ },
+ {
+ "epoch": 35.95021834061135,
+ "grad_norm": 0.014348454773426056,
+ "learning_rate": 0.0006,
+ "loss": 4.171329975128174,
+ "step": 2588
+ },
+ {
+ "epoch": 35.96419213973799,
+ "grad_norm": 0.015598480589687824,
+ "learning_rate": 0.0006,
+ "loss": 4.140815258026123,
+ "step": 2589
+ },
+ {
+ "epoch": 35.97816593886463,
+ "grad_norm": 0.017126062884926796,
+ "learning_rate": 0.0006,
+ "loss": 4.157176971435547,
+ "step": 2590
+ },
+ {
+ "epoch": 35.992139737991266,
+ "grad_norm": 0.01860199309885502,
+ "learning_rate": 0.0006,
+ "loss": 4.252880096435547,
+ "step": 2591
+ },
+ {
+ "epoch": 36.0,
+ "grad_norm": 0.021780220791697502,
+ "learning_rate": 0.0006,
+ "loss": 4.249861717224121,
+ "step": 2592
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 28800,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 400,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": true,
+ "should_log": false,
+ "should_save": false,
+ "should_training_stop": false
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 1.1050608100796006e+18,
+ "train_batch_size": 8,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2592/training_args.bin b/runs/l2r50-i2-fulle-lm/checkpoint-2592/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..cad232831c222cc7f8b938b8bbe42f61e7d89c3a
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-2592/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:f7c40e3bfa2f5ac1450ca1243e3f2a895d44a06ed5995e6aefd97f78d229d8c8
+size 4792
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-288/chat_template.jinja b/runs/l2r50-i2-fulle-lm/checkpoint-288/chat_template.jinja
new file mode 100644
index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-288/chat_template.jinja
@@ -0,0 +1,85 @@
+{%- if tools %}
+ {{- '<|im_start|>system\n' }}
+ {%- if messages[0].role == 'system' %}
+ {{- messages[0].content + '\n\n' }}
+ {%- endif %}
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }}
+ {%- for tool in tools %}
+ {{- "\n" }}
+ {{- tool | tojson }}
+ {%- endfor %}
+ {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }}
+{%- else %}
+ {%- if messages[0].role == 'system' %}
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
+ {%- endif %}
+{%- endif %}
+{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
+{%- for message in messages[::-1] %}
+ {%- set index = (messages|length - 1) - loop.index0 %}
+ {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %}
+ {%- set ns.multi_step_tool = false %}
+ {%- set ns.last_query_index = index %}
+ {%- endif %}
+{%- endfor %}
+{%- for message in messages %}
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
+ {%- elif message.role == "assistant" %}
+ {%- set content = message.content %}
+ {%- set reasoning_content = '' %}
+ {%- if message.reasoning_content is defined and message.reasoning_content is not none %}
+ {%- set reasoning_content = message.reasoning_content %}
+ {%- else %}
+ {%- if '' in message.content %}
+ {%- set content = message.content.split('')[-1].lstrip('\n') %}
+ {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %}
+ {%- endif %}
+ {%- endif %}
+ {%- if loop.index0 > ns.last_query_index %}
+ {%- if loop.last or (not loop.last and reasoning_content) %}
+ {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- if message.tool_calls %}
+ {%- for tool_call in message.tool_calls %}
+ {%- if (loop.first and content) or (not loop.first) %}
+ {{- '\n' }}
+ {%- endif %}
+ {%- if tool_call.function %}
+ {%- set tool_call = tool_call.function %}
+ {%- endif %}
+ {{- '\n{"name": "' }}
+ {{- tool_call.name }}
+ {{- '", "arguments": ' }}
+ {%- if tool_call.arguments is string %}
+ {{- tool_call.arguments }}
+ {%- else %}
+ {{- tool_call.arguments | tojson }}
+ {%- endif %}
+ {{- '}\n' }}
+ {%- endfor %}
+ {%- endif %}
+ {{- '<|im_end|>\n' }}
+ {%- elif message.role == "tool" %}
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
+ {{- '<|im_start|>user' }}
+ {%- endif %}
+ {{- '\n\n' }}
+ {{- message.content }}
+ {{- '\n' }}
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
+ {{- '<|im_end|>\n' }}
+ {%- endif %}
+ {%- endif %}
+{%- endfor %}
+{%- if add_generation_prompt %}
+ {{- '<|im_start|>assistant\n' }}
+ {%- if enable_thinking is defined and enable_thinking is false %}
+ {{- '\n\n\n\n' }}
+ {%- endif %}
+{%- endif %}
\ No newline at end of file
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-288/config.json b/runs/l2r50-i2-fulle-lm/checkpoint-288/config.json
new file mode 100644
index 0000000000000000000000000000000000000000..22f250953c6ce8fc91ee4411856dd73041a415e2
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-288/config.json
@@ -0,0 +1,32 @@
+{
+ "architectures": [
+ "LlamaForCausalLM"
+ ],
+ "attention_bias": false,
+ "attention_dropout": 0.0,
+ "bos_token_id": null,
+ "dtype": "float32",
+ "eos_token_id": 151645,
+ "head_dim": 128,
+ "hidden_act": "silu",
+ "hidden_size": 512,
+ "initializer_range": 0.02,
+ "intermediate_size": 1536,
+ "max_position_embeddings": 2048,
+ "mlp_bias": false,
+ "model_type": "llama",
+ "num_attention_heads": 4,
+ "num_hidden_layers": 20,
+ "num_key_value_heads": 4,
+ "pad_token_id": 151645,
+ "pretraining_tp": 1,
+ "rms_norm_eps": 1e-06,
+ "rope_parameters": {
+ "rope_theta": 10000.0,
+ "rope_type": "default"
+ },
+ "tie_word_embeddings": true,
+ "transformers_version": "5.5.0",
+ "use_cache": false,
+ "vocab_size": 151673
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-288/generation_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-288/generation_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-288/generation_config.json
@@ -0,0 +1,11 @@
+{
+ "_from_model_config": true,
+ "eos_token_id": [
+ 151645
+ ],
+ "output_attentions": false,
+ "output_hidden_states": false,
+ "pad_token_id": 151645,
+ "transformers_version": "5.5.0",
+ "use_cache": true
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-288/model.safetensors b/runs/l2r50-i2-fulle-lm/checkpoint-288/model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..932e2392f1f7ca6b990b2f73c2370b3470207aff
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-288/model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:0b73f8b5606f1126045d5c65fba98a80f8bf4d48c7ee9d040ff33f8e6df39d35
+size 583360328
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-288/optimizer.pt b/runs/l2r50-i2-fulle-lm/checkpoint-288/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..142e0688e153e7072df672f1970d3ffd4e3d39c9
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-288/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:9bf29c4366db961bc92ec5cbb00d9ecc6660ab160276331129e62e7dafe8bfce
+size 1166833530
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-288/rng_state_0.pth b/runs/l2r50-i2-fulle-lm/checkpoint-288/rng_state_0.pth
new file mode 100644
index 0000000000000000000000000000000000000000..1f77348e5e2510858cdf15d1d52ccae69d3c15ec
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-288/rng_state_0.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:4be5e023fa0cbf8e69b3989a76be3c51c1003595cef9a1cd6a981e96df33dc81
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-288/rng_state_1.pth b/runs/l2r50-i2-fulle-lm/checkpoint-288/rng_state_1.pth
new file mode 100644
index 0000000000000000000000000000000000000000..1be981878581dc4cc5e85e632fb1f24089818b49
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-288/rng_state_1.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:f07d0e39393ed810f745b19bf8ca8fa3347107f534dbdfb8793f199d1dba009c
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-288/rng_state_2.pth b/runs/l2r50-i2-fulle-lm/checkpoint-288/rng_state_2.pth
new file mode 100644
index 0000000000000000000000000000000000000000..6fbcd414b36d096739a953ed547b1dff430947f8
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-288/rng_state_2.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:48b06f9ecad190f56f234653cd52c124e3cf5e35fea91d2ea5b309bab6a46c3c
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-288/rng_state_3.pth b/runs/l2r50-i2-fulle-lm/checkpoint-288/rng_state_3.pth
new file mode 100644
index 0000000000000000000000000000000000000000..36c6f976b01de6004b9d1389ad29e9283efba9b3
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-288/rng_state_3.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:ed7c98726399543e24ee5ac4b4f3ea212ab29fcfec2a586bb2dc00d6f002e8de
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-288/scheduler.pt b/runs/l2r50-i2-fulle-lm/checkpoint-288/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..e7f2e4bf5abc860ae45f2d5dbda8115542dd3611
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-288/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:4d1eab1d6f570939dc1e4b9d10383b7354ef244100fb2b34ff0ac8b58642b63c
+size 1064
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-288/tokenizer.json b/runs/l2r50-i2-fulle-lm/checkpoint-288/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..53786016f070e11184813f3c83101dc5e93e392e
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-288/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:c3c63fda1646a555448e8f03662542e04b11d103721319e4d9b267d7da03ceb8
+size 11424463
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-288/tokenizer_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-288/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..98e2fb9b2cc54785dc9d871b8e503f8b8f2defe7
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-288/tokenizer_config.json
@@ -0,0 +1,20 @@
+{
+ "add_prefix_space": false,
+ "backend": "tokenizers",
+ "bos_token": null,
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "<|im_end|>",
+ "errors": "replace",
+ "extra_special_tokens": [
+ "<|l2r_pred|>",
+ "<|r2l_pred|>",
+ "<|next_1_pred|>",
+ "<|next_2_pred|>"
+ ],
+ "is_local": false,
+ "model_max_length": 131072,
+ "pad_token": "<|im_end|>",
+ "split_special_tokens": false,
+ "tokenizer_class": "Qwen2Tokenizer",
+ "unk_token": null
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-288/trainer_state.json b/runs/l2r50-i2-fulle-lm/checkpoint-288/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..534ab212ab85e7abb844693a88e2f7df98a1a732
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-288/trainer_state.json
@@ -0,0 +1,2074 @@
+{
+ "best_global_step": null,
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 4.0,
+ "eval_steps": 500,
+ "global_step": 288,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.013973799126637555,
+ "grad_norm": 0.1739506870508194,
+ "learning_rate": 0.0,
+ "loss": 11.99299430847168,
+ "step": 1
+ },
+ {
+ "epoch": 0.02794759825327511,
+ "grad_norm": 0.17216597497463226,
+ "learning_rate": 5.999999999999999e-06,
+ "loss": 11.988704681396484,
+ "step": 2
+ },
+ {
+ "epoch": 0.04192139737991266,
+ "grad_norm": 0.17459715902805328,
+ "learning_rate": 1.1999999999999999e-05,
+ "loss": 11.945581436157227,
+ "step": 3
+ },
+ {
+ "epoch": 0.05589519650655022,
+ "grad_norm": 0.17577414214611053,
+ "learning_rate": 1.7999999999999997e-05,
+ "loss": 11.859772682189941,
+ "step": 4
+ },
+ {
+ "epoch": 0.06986899563318777,
+ "grad_norm": 0.16163651645183563,
+ "learning_rate": 2.3999999999999997e-05,
+ "loss": 11.744415283203125,
+ "step": 5
+ },
+ {
+ "epoch": 0.08384279475982533,
+ "grad_norm": 0.14409932494163513,
+ "learning_rate": 2.9999999999999997e-05,
+ "loss": 11.618099212646484,
+ "step": 6
+ },
+ {
+ "epoch": 0.09781659388646288,
+ "grad_norm": 0.13079848885536194,
+ "learning_rate": 3.5999999999999994e-05,
+ "loss": 11.505935668945312,
+ "step": 7
+ },
+ {
+ "epoch": 0.11179039301310044,
+ "grad_norm": 0.11964613944292068,
+ "learning_rate": 4.2e-05,
+ "loss": 11.416532516479492,
+ "step": 8
+ },
+ {
+ "epoch": 0.125764192139738,
+ "grad_norm": 0.11345142871141434,
+ "learning_rate": 4.7999999999999994e-05,
+ "loss": 11.33206558227539,
+ "step": 9
+ },
+ {
+ "epoch": 0.13973799126637554,
+ "grad_norm": 0.11033952981233597,
+ "learning_rate": 5.399999999999999e-05,
+ "loss": 11.259052276611328,
+ "step": 10
+ },
+ {
+ "epoch": 0.1537117903930131,
+ "grad_norm": 0.10776489973068237,
+ "learning_rate": 5.9999999999999995e-05,
+ "loss": 11.203384399414062,
+ "step": 11
+ },
+ {
+ "epoch": 0.16768558951965065,
+ "grad_norm": 0.10721946507692337,
+ "learning_rate": 6.599999999999999e-05,
+ "loss": 11.153295516967773,
+ "step": 12
+ },
+ {
+ "epoch": 0.18165938864628822,
+ "grad_norm": 0.10673430562019348,
+ "learning_rate": 7.199999999999999e-05,
+ "loss": 11.110200881958008,
+ "step": 13
+ },
+ {
+ "epoch": 0.19563318777292577,
+ "grad_norm": 0.10630254447460175,
+ "learning_rate": 7.8e-05,
+ "loss": 11.067477226257324,
+ "step": 14
+ },
+ {
+ "epoch": 0.2096069868995633,
+ "grad_norm": 0.10631615668535233,
+ "learning_rate": 8.4e-05,
+ "loss": 11.022941589355469,
+ "step": 15
+ },
+ {
+ "epoch": 0.22358078602620088,
+ "grad_norm": 0.10609443485736847,
+ "learning_rate": 8.999999999999999e-05,
+ "loss": 10.970746040344238,
+ "step": 16
+ },
+ {
+ "epoch": 0.23755458515283842,
+ "grad_norm": 0.10648232698440552,
+ "learning_rate": 9.599999999999999e-05,
+ "loss": 10.91267204284668,
+ "step": 17
+ },
+ {
+ "epoch": 0.251528384279476,
+ "grad_norm": 0.10584589838981628,
+ "learning_rate": 0.000102,
+ "loss": 10.854547500610352,
+ "step": 18
+ },
+ {
+ "epoch": 0.26550218340611353,
+ "grad_norm": 0.10591720044612885,
+ "learning_rate": 0.00010799999999999998,
+ "loss": 10.788917541503906,
+ "step": 19
+ },
+ {
+ "epoch": 0.2794759825327511,
+ "grad_norm": 0.10611672699451447,
+ "learning_rate": 0.00011399999999999999,
+ "loss": 10.717556953430176,
+ "step": 20
+ },
+ {
+ "epoch": 0.2934497816593886,
+ "grad_norm": 0.10455359518527985,
+ "learning_rate": 0.00011999999999999999,
+ "loss": 10.650186538696289,
+ "step": 21
+ },
+ {
+ "epoch": 0.3074235807860262,
+ "grad_norm": 0.1047525629401207,
+ "learning_rate": 0.00012599999999999997,
+ "loss": 10.56808090209961,
+ "step": 22
+ },
+ {
+ "epoch": 0.32139737991266376,
+ "grad_norm": 0.10511605441570282,
+ "learning_rate": 0.00013199999999999998,
+ "loss": 10.484028816223145,
+ "step": 23
+ },
+ {
+ "epoch": 0.3353711790393013,
+ "grad_norm": 0.10467177629470825,
+ "learning_rate": 0.000138,
+ "loss": 10.399947166442871,
+ "step": 24
+ },
+ {
+ "epoch": 0.34934497816593885,
+ "grad_norm": 0.10432136058807373,
+ "learning_rate": 0.00014399999999999998,
+ "loss": 10.312714576721191,
+ "step": 25
+ },
+ {
+ "epoch": 0.36331877729257644,
+ "grad_norm": 0.10527262091636658,
+ "learning_rate": 0.00015,
+ "loss": 10.212261199951172,
+ "step": 26
+ },
+ {
+ "epoch": 0.377292576419214,
+ "grad_norm": 0.103610560297966,
+ "learning_rate": 0.000156,
+ "loss": 10.132034301757812,
+ "step": 27
+ },
+ {
+ "epoch": 0.39126637554585153,
+ "grad_norm": 0.10553059726953506,
+ "learning_rate": 0.000162,
+ "loss": 10.020434379577637,
+ "step": 28
+ },
+ {
+ "epoch": 0.4052401746724891,
+ "grad_norm": 0.1037299782037735,
+ "learning_rate": 0.000168,
+ "loss": 9.934466361999512,
+ "step": 29
+ },
+ {
+ "epoch": 0.4192139737991266,
+ "grad_norm": 0.1042739674448967,
+ "learning_rate": 0.00017399999999999997,
+ "loss": 9.822539329528809,
+ "step": 30
+ },
+ {
+ "epoch": 0.4331877729257642,
+ "grad_norm": 0.10300865024328232,
+ "learning_rate": 0.00017999999999999998,
+ "loss": 9.734994888305664,
+ "step": 31
+ },
+ {
+ "epoch": 0.44716157205240176,
+ "grad_norm": 0.10291654616594315,
+ "learning_rate": 0.000186,
+ "loss": 9.624505996704102,
+ "step": 32
+ },
+ {
+ "epoch": 0.4611353711790393,
+ "grad_norm": 0.10210851579904556,
+ "learning_rate": 0.00019199999999999998,
+ "loss": 9.518205642700195,
+ "step": 33
+ },
+ {
+ "epoch": 0.47510917030567684,
+ "grad_norm": 0.10222529619932175,
+ "learning_rate": 0.000198,
+ "loss": 9.412222862243652,
+ "step": 34
+ },
+ {
+ "epoch": 0.4890829694323144,
+ "grad_norm": 0.09992019087076187,
+ "learning_rate": 0.000204,
+ "loss": 9.326566696166992,
+ "step": 35
+ },
+ {
+ "epoch": 0.503056768558952,
+ "grad_norm": 0.09931084513664246,
+ "learning_rate": 0.00020999999999999998,
+ "loss": 9.210111618041992,
+ "step": 36
+ },
+ {
+ "epoch": 0.5170305676855895,
+ "grad_norm": 0.09928256273269653,
+ "learning_rate": 0.00021599999999999996,
+ "loss": 9.100493431091309,
+ "step": 37
+ },
+ {
+ "epoch": 0.5310043668122271,
+ "grad_norm": 0.09843816608190536,
+ "learning_rate": 0.00022199999999999998,
+ "loss": 8.996543884277344,
+ "step": 38
+ },
+ {
+ "epoch": 0.5449781659388646,
+ "grad_norm": 0.09825374186038971,
+ "learning_rate": 0.00022799999999999999,
+ "loss": 8.888991355895996,
+ "step": 39
+ },
+ {
+ "epoch": 0.5589519650655022,
+ "grad_norm": 0.09602385759353638,
+ "learning_rate": 0.000234,
+ "loss": 8.802907943725586,
+ "step": 40
+ },
+ {
+ "epoch": 0.5729257641921397,
+ "grad_norm": 0.09414064139127731,
+ "learning_rate": 0.00023999999999999998,
+ "loss": 8.715423583984375,
+ "step": 41
+ },
+ {
+ "epoch": 0.5868995633187772,
+ "grad_norm": 0.09273893386125565,
+ "learning_rate": 0.00024599999999999996,
+ "loss": 8.61374282836914,
+ "step": 42
+ },
+ {
+ "epoch": 0.6008733624454149,
+ "grad_norm": 0.09079599380493164,
+ "learning_rate": 0.00025199999999999995,
+ "loss": 8.502904891967773,
+ "step": 43
+ },
+ {
+ "epoch": 0.6148471615720524,
+ "grad_norm": 0.08605825901031494,
+ "learning_rate": 0.000258,
+ "loss": 8.453008651733398,
+ "step": 44
+ },
+ {
+ "epoch": 0.62882096069869,
+ "grad_norm": 0.08338996022939682,
+ "learning_rate": 0.00026399999999999997,
+ "loss": 8.365163803100586,
+ "step": 45
+ },
+ {
+ "epoch": 0.6427947598253275,
+ "grad_norm": 0.08150867372751236,
+ "learning_rate": 0.00027,
+ "loss": 8.285372734069824,
+ "step": 46
+ },
+ {
+ "epoch": 0.6567685589519651,
+ "grad_norm": 0.07794807851314545,
+ "learning_rate": 0.000276,
+ "loss": 8.185811996459961,
+ "step": 47
+ },
+ {
+ "epoch": 0.6707423580786026,
+ "grad_norm": 0.07592720538377762,
+ "learning_rate": 0.00028199999999999997,
+ "loss": 8.098482131958008,
+ "step": 48
+ },
+ {
+ "epoch": 0.6847161572052402,
+ "grad_norm": 0.0709897130727768,
+ "learning_rate": 0.00028799999999999995,
+ "loss": 8.042091369628906,
+ "step": 49
+ },
+ {
+ "epoch": 0.6986899563318777,
+ "grad_norm": 0.06582305580377579,
+ "learning_rate": 0.000294,
+ "loss": 7.972844123840332,
+ "step": 50
+ },
+ {
+ "epoch": 0.7126637554585152,
+ "grad_norm": 0.059452351182699203,
+ "learning_rate": 0.0003,
+ "loss": 7.924579620361328,
+ "step": 51
+ },
+ {
+ "epoch": 0.7266375545851529,
+ "grad_norm": 0.052426718175411224,
+ "learning_rate": 0.00030599999999999996,
+ "loss": 7.876336097717285,
+ "step": 52
+ },
+ {
+ "epoch": 0.7406113537117904,
+ "grad_norm": 0.04768150672316551,
+ "learning_rate": 0.000312,
+ "loss": 7.8102707862854,
+ "step": 53
+ },
+ {
+ "epoch": 0.754585152838428,
+ "grad_norm": 0.03970075398683548,
+ "learning_rate": 0.000318,
+ "loss": 7.791913986206055,
+ "step": 54
+ },
+ {
+ "epoch": 0.7685589519650655,
+ "grad_norm": 0.03559752553701401,
+ "learning_rate": 0.000324,
+ "loss": 7.756960868835449,
+ "step": 55
+ },
+ {
+ "epoch": 0.7825327510917031,
+ "grad_norm": 0.029321353882551193,
+ "learning_rate": 0.00033,
+ "loss": 7.724878787994385,
+ "step": 56
+ },
+ {
+ "epoch": 0.7965065502183406,
+ "grad_norm": 0.023485427722334862,
+ "learning_rate": 0.000336,
+ "loss": 7.72553014755249,
+ "step": 57
+ },
+ {
+ "epoch": 0.8104803493449781,
+ "grad_norm": 0.019319474697113037,
+ "learning_rate": 0.00034199999999999996,
+ "loss": 7.709493160247803,
+ "step": 58
+ },
+ {
+ "epoch": 0.8244541484716157,
+ "grad_norm": 0.016705691814422607,
+ "learning_rate": 0.00034799999999999995,
+ "loss": 7.6873297691345215,
+ "step": 59
+ },
+ {
+ "epoch": 0.8384279475982532,
+ "grad_norm": 0.017667734995484352,
+ "learning_rate": 0.00035399999999999993,
+ "loss": 7.677059173583984,
+ "step": 60
+ },
+ {
+ "epoch": 0.8524017467248908,
+ "grad_norm": 0.01671992614865303,
+ "learning_rate": 0.00035999999999999997,
+ "loss": 7.643512725830078,
+ "step": 61
+ },
+ {
+ "epoch": 0.8663755458515284,
+ "grad_norm": 0.012288172729313374,
+ "learning_rate": 0.00036599999999999995,
+ "loss": 7.658343315124512,
+ "step": 62
+ },
+ {
+ "epoch": 0.880349344978166,
+ "grad_norm": 0.012004776857793331,
+ "learning_rate": 0.000372,
+ "loss": 7.649759292602539,
+ "step": 63
+ },
+ {
+ "epoch": 0.8943231441048035,
+ "grad_norm": 0.012487445957958698,
+ "learning_rate": 0.00037799999999999997,
+ "loss": 7.631503105163574,
+ "step": 64
+ },
+ {
+ "epoch": 0.9082969432314411,
+ "grad_norm": 0.01489241048693657,
+ "learning_rate": 0.00038399999999999996,
+ "loss": 7.647110462188721,
+ "step": 65
+ },
+ {
+ "epoch": 0.9222707423580786,
+ "grad_norm": 0.01428980939090252,
+ "learning_rate": 0.00039,
+ "loss": 7.636983871459961,
+ "step": 66
+ },
+ {
+ "epoch": 0.9362445414847161,
+ "grad_norm": 0.010711144655942917,
+ "learning_rate": 0.000396,
+ "loss": 7.621130466461182,
+ "step": 67
+ },
+ {
+ "epoch": 0.9502183406113537,
+ "grad_norm": 0.010178886353969574,
+ "learning_rate": 0.000402,
+ "loss": 7.593924522399902,
+ "step": 68
+ },
+ {
+ "epoch": 0.9641921397379912,
+ "grad_norm": 0.010604874230921268,
+ "learning_rate": 0.000408,
+ "loss": 7.6015496253967285,
+ "step": 69
+ },
+ {
+ "epoch": 0.9781659388646288,
+ "grad_norm": 0.009751426056027412,
+ "learning_rate": 0.0004139999999999999,
+ "loss": 7.571964263916016,
+ "step": 70
+ },
+ {
+ "epoch": 0.9921397379912664,
+ "grad_norm": 0.009816481731832027,
+ "learning_rate": 0.00041999999999999996,
+ "loss": 7.57344913482666,
+ "step": 71
+ },
+ {
+ "epoch": 1.0,
+ "grad_norm": 0.008665637113153934,
+ "learning_rate": 0.00042599999999999995,
+ "loss": 7.579050064086914,
+ "step": 72
+ },
+ {
+ "epoch": 1.0,
+ "eval_loss": 7.581204891204834,
+ "eval_runtime": 59.0091,
+ "eval_samples_per_second": 41.383,
+ "eval_steps_per_second": 1.305,
+ "step": 72
+ },
+ {
+ "epoch": 1.0139737991266375,
+ "grad_norm": 0.009684201329946518,
+ "learning_rate": 0.00043199999999999993,
+ "loss": 7.568594932556152,
+ "step": 73
+ },
+ {
+ "epoch": 1.027947598253275,
+ "grad_norm": 0.025783710181713104,
+ "learning_rate": 0.00043799999999999997,
+ "loss": 7.578948497772217,
+ "step": 74
+ },
+ {
+ "epoch": 1.0419213973799126,
+ "grad_norm": 0.0327819399535656,
+ "learning_rate": 0.00044399999999999995,
+ "loss": 7.548173904418945,
+ "step": 75
+ },
+ {
+ "epoch": 1.0558951965065502,
+ "grad_norm": 0.021604498848319054,
+ "learning_rate": 0.00045,
+ "loss": 7.552953720092773,
+ "step": 76
+ },
+ {
+ "epoch": 1.0698689956331877,
+ "grad_norm": 0.010230105370283127,
+ "learning_rate": 0.00045599999999999997,
+ "loss": 7.5213212966918945,
+ "step": 77
+ },
+ {
+ "epoch": 1.0838427947598253,
+ "grad_norm": 0.023315589874982834,
+ "learning_rate": 0.00046199999999999995,
+ "loss": 7.501640319824219,
+ "step": 78
+ },
+ {
+ "epoch": 1.0978165938864628,
+ "grad_norm": 0.01109944935888052,
+ "learning_rate": 0.000468,
+ "loss": 7.539519786834717,
+ "step": 79
+ },
+ {
+ "epoch": 1.1117903930131003,
+ "grad_norm": 0.026093758642673492,
+ "learning_rate": 0.000474,
+ "loss": 7.525928974151611,
+ "step": 80
+ },
+ {
+ "epoch": 1.125764192139738,
+ "grad_norm": 0.03274895250797272,
+ "learning_rate": 0.00047999999999999996,
+ "loss": 7.535602569580078,
+ "step": 81
+ },
+ {
+ "epoch": 1.1397379912663754,
+ "grad_norm": 0.01543602254241705,
+ "learning_rate": 0.000486,
+ "loss": 7.513683795928955,
+ "step": 82
+ },
+ {
+ "epoch": 1.1537117903930132,
+ "grad_norm": 0.028377249836921692,
+ "learning_rate": 0.0004919999999999999,
+ "loss": 7.49384880065918,
+ "step": 83
+ },
+ {
+ "epoch": 1.1676855895196507,
+ "grad_norm": 0.04121481627225876,
+ "learning_rate": 0.000498,
+ "loss": 7.488447189331055,
+ "step": 84
+ },
+ {
+ "epoch": 1.1816593886462883,
+ "grad_norm": 0.013282302767038345,
+ "learning_rate": 0.0005039999999999999,
+ "loss": 7.463010787963867,
+ "step": 85
+ },
+ {
+ "epoch": 1.1956331877729258,
+ "grad_norm": 0.03200364485383034,
+ "learning_rate": 0.0005099999999999999,
+ "loss": 7.473515033721924,
+ "step": 86
+ },
+ {
+ "epoch": 1.2096069868995634,
+ "grad_norm": 0.020768029615283012,
+ "learning_rate": 0.000516,
+ "loss": 7.464688777923584,
+ "step": 87
+ },
+ {
+ "epoch": 1.223580786026201,
+ "grad_norm": 0.021715788170695305,
+ "learning_rate": 0.000522,
+ "loss": 7.4835028648376465,
+ "step": 88
+ },
+ {
+ "epoch": 1.2375545851528384,
+ "grad_norm": 0.026255754753947258,
+ "learning_rate": 0.0005279999999999999,
+ "loss": 7.443392753601074,
+ "step": 89
+ },
+ {
+ "epoch": 1.251528384279476,
+ "grad_norm": 0.022325951606035233,
+ "learning_rate": 0.000534,
+ "loss": 7.457373142242432,
+ "step": 90
+ },
+ {
+ "epoch": 1.2655021834061135,
+ "grad_norm": 0.009270387701690197,
+ "learning_rate": 0.00054,
+ "loss": 7.455986976623535,
+ "step": 91
+ },
+ {
+ "epoch": 1.279475982532751,
+ "grad_norm": 0.02369544468820095,
+ "learning_rate": 0.0005459999999999999,
+ "loss": 7.442216873168945,
+ "step": 92
+ },
+ {
+ "epoch": 1.2934497816593886,
+ "grad_norm": 0.008474506437778473,
+ "learning_rate": 0.000552,
+ "loss": 7.4294915199279785,
+ "step": 93
+ },
+ {
+ "epoch": 1.3074235807860262,
+ "grad_norm": 0.017043249681591988,
+ "learning_rate": 0.000558,
+ "loss": 7.406932830810547,
+ "step": 94
+ },
+ {
+ "epoch": 1.3213973799126637,
+ "grad_norm": 0.011813412420451641,
+ "learning_rate": 0.0005639999999999999,
+ "loss": 7.415900230407715,
+ "step": 95
+ },
+ {
+ "epoch": 1.3353711790393012,
+ "grad_norm": 0.01086883433163166,
+ "learning_rate": 0.00057,
+ "loss": 7.452998638153076,
+ "step": 96
+ },
+ {
+ "epoch": 1.3493449781659388,
+ "grad_norm": 0.014700951054692268,
+ "learning_rate": 0.0005759999999999999,
+ "loss": 7.410249710083008,
+ "step": 97
+ },
+ {
+ "epoch": 1.3633187772925766,
+ "grad_norm": 0.014512976631522179,
+ "learning_rate": 0.0005819999999999999,
+ "loss": 7.429132461547852,
+ "step": 98
+ },
+ {
+ "epoch": 1.3772925764192139,
+ "grad_norm": 0.012519920244812965,
+ "learning_rate": 0.000588,
+ "loss": 7.430817604064941,
+ "step": 99
+ },
+ {
+ "epoch": 1.3912663755458516,
+ "grad_norm": 0.007132664322853088,
+ "learning_rate": 0.0005939999999999999,
+ "loss": 7.405234336853027,
+ "step": 100
+ },
+ {
+ "epoch": 1.405240174672489,
+ "grad_norm": 0.019489388912916183,
+ "learning_rate": 0.0006,
+ "loss": 7.427791595458984,
+ "step": 101
+ },
+ {
+ "epoch": 1.4192139737991267,
+ "grad_norm": 0.027252664789557457,
+ "learning_rate": 0.0006,
+ "loss": 7.41761589050293,
+ "step": 102
+ },
+ {
+ "epoch": 1.4331877729257643,
+ "grad_norm": 0.02501637488603592,
+ "learning_rate": 0.0006,
+ "loss": 7.416735649108887,
+ "step": 103
+ },
+ {
+ "epoch": 1.4471615720524018,
+ "grad_norm": 0.00989772193133831,
+ "learning_rate": 0.0006,
+ "loss": 7.3756632804870605,
+ "step": 104
+ },
+ {
+ "epoch": 1.4611353711790394,
+ "grad_norm": 0.01622949168086052,
+ "learning_rate": 0.0006,
+ "loss": 7.400811672210693,
+ "step": 105
+ },
+ {
+ "epoch": 1.475109170305677,
+ "grad_norm": 0.012407763861119747,
+ "learning_rate": 0.0006,
+ "loss": 7.37675666809082,
+ "step": 106
+ },
+ {
+ "epoch": 1.4890829694323144,
+ "grad_norm": 0.010353565216064453,
+ "learning_rate": 0.0006,
+ "loss": 7.4013142585754395,
+ "step": 107
+ },
+ {
+ "epoch": 1.503056768558952,
+ "grad_norm": 0.0120553532615304,
+ "learning_rate": 0.0006,
+ "loss": 7.379873275756836,
+ "step": 108
+ },
+ {
+ "epoch": 1.5170305676855895,
+ "grad_norm": 0.01610225811600685,
+ "learning_rate": 0.0006,
+ "loss": 7.371631145477295,
+ "step": 109
+ },
+ {
+ "epoch": 1.531004366812227,
+ "grad_norm": 0.01159879844635725,
+ "learning_rate": 0.0006,
+ "loss": 7.389007568359375,
+ "step": 110
+ },
+ {
+ "epoch": 1.5449781659388646,
+ "grad_norm": 0.008013816550374031,
+ "learning_rate": 0.0006,
+ "loss": 7.395920753479004,
+ "step": 111
+ },
+ {
+ "epoch": 1.5589519650655022,
+ "grad_norm": 0.016334809362888336,
+ "learning_rate": 0.0006,
+ "loss": 7.393121719360352,
+ "step": 112
+ },
+ {
+ "epoch": 1.5729257641921397,
+ "grad_norm": 0.018821721896529198,
+ "learning_rate": 0.0006,
+ "loss": 7.367543697357178,
+ "step": 113
+ },
+ {
+ "epoch": 1.5868995633187772,
+ "grad_norm": 0.014088994823396206,
+ "learning_rate": 0.0006,
+ "loss": 7.385537624359131,
+ "step": 114
+ },
+ {
+ "epoch": 1.600873362445415,
+ "grad_norm": 0.005864050704985857,
+ "learning_rate": 0.0006,
+ "loss": 7.365346908569336,
+ "step": 115
+ },
+ {
+ "epoch": 1.6148471615720523,
+ "grad_norm": 0.011745430529117584,
+ "learning_rate": 0.0006,
+ "loss": 7.390227317810059,
+ "step": 116
+ },
+ {
+ "epoch": 1.62882096069869,
+ "grad_norm": 0.017205573618412018,
+ "learning_rate": 0.0006,
+ "loss": 7.3558549880981445,
+ "step": 117
+ },
+ {
+ "epoch": 1.6427947598253274,
+ "grad_norm": 0.021256346255540848,
+ "learning_rate": 0.0006,
+ "loss": 7.377346992492676,
+ "step": 118
+ },
+ {
+ "epoch": 1.6567685589519652,
+ "grad_norm": 0.019335204735398293,
+ "learning_rate": 0.0006,
+ "loss": 7.346194267272949,
+ "step": 119
+ },
+ {
+ "epoch": 1.6707423580786025,
+ "grad_norm": 0.008134003728628159,
+ "learning_rate": 0.0006,
+ "loss": 7.3063578605651855,
+ "step": 120
+ },
+ {
+ "epoch": 1.6847161572052403,
+ "grad_norm": 0.017613831907510757,
+ "learning_rate": 0.0006,
+ "loss": 7.338171005249023,
+ "step": 121
+ },
+ {
+ "epoch": 1.6986899563318776,
+ "grad_norm": 0.00890734139829874,
+ "learning_rate": 0.0006,
+ "loss": 7.355384826660156,
+ "step": 122
+ },
+ {
+ "epoch": 1.7126637554585153,
+ "grad_norm": 0.013448155485093594,
+ "learning_rate": 0.0006,
+ "loss": 7.344296455383301,
+ "step": 123
+ },
+ {
+ "epoch": 1.726637554585153,
+ "grad_norm": 0.009241633117198944,
+ "learning_rate": 0.0006,
+ "loss": 7.3454060554504395,
+ "step": 124
+ },
+ {
+ "epoch": 1.7406113537117904,
+ "grad_norm": 0.018561311066150665,
+ "learning_rate": 0.0006,
+ "loss": 7.32348108291626,
+ "step": 125
+ },
+ {
+ "epoch": 1.754585152838428,
+ "grad_norm": 0.025667704641819,
+ "learning_rate": 0.0006,
+ "loss": 7.356142520904541,
+ "step": 126
+ },
+ {
+ "epoch": 1.7685589519650655,
+ "grad_norm": 0.036232151091098785,
+ "learning_rate": 0.0006,
+ "loss": 7.363539695739746,
+ "step": 127
+ },
+ {
+ "epoch": 1.782532751091703,
+ "grad_norm": 0.059483714401721954,
+ "learning_rate": 0.0006,
+ "loss": 7.402331352233887,
+ "step": 128
+ },
+ {
+ "epoch": 1.7965065502183406,
+ "grad_norm": 0.05365009233355522,
+ "learning_rate": 0.0006,
+ "loss": 7.40457820892334,
+ "step": 129
+ },
+ {
+ "epoch": 1.8104803493449781,
+ "grad_norm": 0.02738415263593197,
+ "learning_rate": 0.0006,
+ "loss": 7.3756208419799805,
+ "step": 130
+ },
+ {
+ "epoch": 1.8244541484716157,
+ "grad_norm": 0.035017091780900955,
+ "learning_rate": 0.0006,
+ "loss": 7.379263877868652,
+ "step": 131
+ },
+ {
+ "epoch": 1.8384279475982532,
+ "grad_norm": 0.02819863148033619,
+ "learning_rate": 0.0006,
+ "loss": 7.327673435211182,
+ "step": 132
+ },
+ {
+ "epoch": 1.8524017467248908,
+ "grad_norm": 0.017754238098859787,
+ "learning_rate": 0.0006,
+ "loss": 7.333420276641846,
+ "step": 133
+ },
+ {
+ "epoch": 1.8663755458515285,
+ "grad_norm": 0.024685733020305634,
+ "learning_rate": 0.0006,
+ "loss": 7.34493350982666,
+ "step": 134
+ },
+ {
+ "epoch": 1.8803493449781659,
+ "grad_norm": 0.019426532089710236,
+ "learning_rate": 0.0006,
+ "loss": 7.328448295593262,
+ "step": 135
+ },
+ {
+ "epoch": 1.8943231441048036,
+ "grad_norm": 0.02106391079723835,
+ "learning_rate": 0.0006,
+ "loss": 7.327917098999023,
+ "step": 136
+ },
+ {
+ "epoch": 1.908296943231441,
+ "grad_norm": 0.02056879736483097,
+ "learning_rate": 0.0006,
+ "loss": 7.3311991691589355,
+ "step": 137
+ },
+ {
+ "epoch": 1.9222707423580787,
+ "grad_norm": 0.016865141689777374,
+ "learning_rate": 0.0006,
+ "loss": 7.312766075134277,
+ "step": 138
+ },
+ {
+ "epoch": 1.936244541484716,
+ "grad_norm": 0.017549855634570122,
+ "learning_rate": 0.0006,
+ "loss": 7.31191349029541,
+ "step": 139
+ },
+ {
+ "epoch": 1.9502183406113538,
+ "grad_norm": 0.015930157154798508,
+ "learning_rate": 0.0006,
+ "loss": 7.330901622772217,
+ "step": 140
+ },
+ {
+ "epoch": 1.9641921397379911,
+ "grad_norm": 0.013283558189868927,
+ "learning_rate": 0.0006,
+ "loss": 7.298318386077881,
+ "step": 141
+ },
+ {
+ "epoch": 1.9781659388646289,
+ "grad_norm": 0.012023469433188438,
+ "learning_rate": 0.0006,
+ "loss": 7.282337188720703,
+ "step": 142
+ },
+ {
+ "epoch": 1.9921397379912664,
+ "grad_norm": 0.01459511648863554,
+ "learning_rate": 0.0006,
+ "loss": 7.282924175262451,
+ "step": 143
+ },
+ {
+ "epoch": 2.0,
+ "grad_norm": 0.017474235966801643,
+ "learning_rate": 0.0006,
+ "loss": 7.254333972930908,
+ "step": 144
+ },
+ {
+ "epoch": 2.0,
+ "eval_loss": 7.318708896636963,
+ "eval_runtime": 59.6555,
+ "eval_samples_per_second": 40.935,
+ "eval_steps_per_second": 1.291,
+ "step": 144
+ },
+ {
+ "epoch": 2.0139737991266378,
+ "grad_norm": 0.019757507368922234,
+ "learning_rate": 0.0006,
+ "loss": 7.30288553237915,
+ "step": 145
+ },
+ {
+ "epoch": 2.027947598253275,
+ "grad_norm": 0.014571248553693295,
+ "learning_rate": 0.0006,
+ "loss": 7.274681568145752,
+ "step": 146
+ },
+ {
+ "epoch": 2.041921397379913,
+ "grad_norm": 0.015744198113679886,
+ "learning_rate": 0.0006,
+ "loss": 7.278632640838623,
+ "step": 147
+ },
+ {
+ "epoch": 2.05589519650655,
+ "grad_norm": 0.028048858046531677,
+ "learning_rate": 0.0006,
+ "loss": 7.26558780670166,
+ "step": 148
+ },
+ {
+ "epoch": 2.069868995633188,
+ "grad_norm": 0.06547307223081589,
+ "learning_rate": 0.0006,
+ "loss": 7.32567834854126,
+ "step": 149
+ },
+ {
+ "epoch": 2.0838427947598253,
+ "grad_norm": 0.05969763547182083,
+ "learning_rate": 0.0006,
+ "loss": 7.395836353302002,
+ "step": 150
+ },
+ {
+ "epoch": 2.097816593886463,
+ "grad_norm": 0.032574281096458435,
+ "learning_rate": 0.0006,
+ "loss": 7.318150997161865,
+ "step": 151
+ },
+ {
+ "epoch": 2.1117903930131003,
+ "grad_norm": 0.029452962800860405,
+ "learning_rate": 0.0006,
+ "loss": 7.308066368103027,
+ "step": 152
+ },
+ {
+ "epoch": 2.125764192139738,
+ "grad_norm": 0.03161991387605667,
+ "learning_rate": 0.0006,
+ "loss": 7.306290626525879,
+ "step": 153
+ },
+ {
+ "epoch": 2.1397379912663754,
+ "grad_norm": 0.027096295729279518,
+ "learning_rate": 0.0006,
+ "loss": 7.297796249389648,
+ "step": 154
+ },
+ {
+ "epoch": 2.153711790393013,
+ "grad_norm": 0.028657056391239166,
+ "learning_rate": 0.0006,
+ "loss": 7.317424774169922,
+ "step": 155
+ },
+ {
+ "epoch": 2.1676855895196505,
+ "grad_norm": 0.02046995982527733,
+ "learning_rate": 0.0006,
+ "loss": 7.297986030578613,
+ "step": 156
+ },
+ {
+ "epoch": 2.1816593886462883,
+ "grad_norm": 0.02220369316637516,
+ "learning_rate": 0.0006,
+ "loss": 7.268237590789795,
+ "step": 157
+ },
+ {
+ "epoch": 2.1956331877729256,
+ "grad_norm": 0.018381357192993164,
+ "learning_rate": 0.0006,
+ "loss": 7.259415626525879,
+ "step": 158
+ },
+ {
+ "epoch": 2.2096069868995634,
+ "grad_norm": 0.0204413291066885,
+ "learning_rate": 0.0006,
+ "loss": 7.258173942565918,
+ "step": 159
+ },
+ {
+ "epoch": 2.2235807860262007,
+ "grad_norm": 0.015874288976192474,
+ "learning_rate": 0.0006,
+ "loss": 7.281834602355957,
+ "step": 160
+ },
+ {
+ "epoch": 2.2375545851528384,
+ "grad_norm": 0.01634068600833416,
+ "learning_rate": 0.0006,
+ "loss": 7.2469868659973145,
+ "step": 161
+ },
+ {
+ "epoch": 2.251528384279476,
+ "grad_norm": 0.01728747971355915,
+ "learning_rate": 0.0006,
+ "loss": 7.280995845794678,
+ "step": 162
+ },
+ {
+ "epoch": 2.2655021834061135,
+ "grad_norm": 0.016341188922524452,
+ "learning_rate": 0.0006,
+ "loss": 7.273357391357422,
+ "step": 163
+ },
+ {
+ "epoch": 2.279475982532751,
+ "grad_norm": 0.016125807538628578,
+ "learning_rate": 0.0006,
+ "loss": 7.273510932922363,
+ "step": 164
+ },
+ {
+ "epoch": 2.2934497816593886,
+ "grad_norm": 0.01629287749528885,
+ "learning_rate": 0.0006,
+ "loss": 7.267029762268066,
+ "step": 165
+ },
+ {
+ "epoch": 2.3074235807860264,
+ "grad_norm": 0.015296131372451782,
+ "learning_rate": 0.0006,
+ "loss": 7.211989402770996,
+ "step": 166
+ },
+ {
+ "epoch": 2.3213973799126637,
+ "grad_norm": 0.01709570363163948,
+ "learning_rate": 0.0006,
+ "loss": 7.242851734161377,
+ "step": 167
+ },
+ {
+ "epoch": 2.3353711790393015,
+ "grad_norm": 0.012552388943731785,
+ "learning_rate": 0.0006,
+ "loss": 7.241696834564209,
+ "step": 168
+ },
+ {
+ "epoch": 2.349344978165939,
+ "grad_norm": 0.014571291394531727,
+ "learning_rate": 0.0006,
+ "loss": 7.270681381225586,
+ "step": 169
+ },
+ {
+ "epoch": 2.3633187772925766,
+ "grad_norm": 0.013389437459409237,
+ "learning_rate": 0.0006,
+ "loss": 7.191600799560547,
+ "step": 170
+ },
+ {
+ "epoch": 2.377292576419214,
+ "grad_norm": 0.011895314790308475,
+ "learning_rate": 0.0006,
+ "loss": 7.213610649108887,
+ "step": 171
+ },
+ {
+ "epoch": 2.3912663755458516,
+ "grad_norm": 0.01345642190426588,
+ "learning_rate": 0.0006,
+ "loss": 7.193787574768066,
+ "step": 172
+ },
+ {
+ "epoch": 2.405240174672489,
+ "grad_norm": 0.0091794328764081,
+ "learning_rate": 0.0006,
+ "loss": 7.233364105224609,
+ "step": 173
+ },
+ {
+ "epoch": 2.4192139737991267,
+ "grad_norm": 0.013767086900770664,
+ "learning_rate": 0.0006,
+ "loss": 7.217272758483887,
+ "step": 174
+ },
+ {
+ "epoch": 2.433187772925764,
+ "grad_norm": 0.015134960412979126,
+ "learning_rate": 0.0006,
+ "loss": 7.192791938781738,
+ "step": 175
+ },
+ {
+ "epoch": 2.447161572052402,
+ "grad_norm": 0.01790648326277733,
+ "learning_rate": 0.0006,
+ "loss": 7.229467391967773,
+ "step": 176
+ },
+ {
+ "epoch": 2.461135371179039,
+ "grad_norm": 0.015362377278506756,
+ "learning_rate": 0.0006,
+ "loss": 7.180346488952637,
+ "step": 177
+ },
+ {
+ "epoch": 2.475109170305677,
+ "grad_norm": 0.010505498386919498,
+ "learning_rate": 0.0006,
+ "loss": 7.188833236694336,
+ "step": 178
+ },
+ {
+ "epoch": 2.489082969432314,
+ "grad_norm": 0.009129747748374939,
+ "learning_rate": 0.0006,
+ "loss": 7.203153133392334,
+ "step": 179
+ },
+ {
+ "epoch": 2.503056768558952,
+ "grad_norm": 0.01147883478552103,
+ "learning_rate": 0.0006,
+ "loss": 7.163824558258057,
+ "step": 180
+ },
+ {
+ "epoch": 2.5170305676855893,
+ "grad_norm": 0.011338942684233189,
+ "learning_rate": 0.0006,
+ "loss": 7.16287899017334,
+ "step": 181
+ },
+ {
+ "epoch": 2.531004366812227,
+ "grad_norm": 0.013277675025165081,
+ "learning_rate": 0.0006,
+ "loss": 7.155905723571777,
+ "step": 182
+ },
+ {
+ "epoch": 2.544978165938865,
+ "grad_norm": 0.015269347466528416,
+ "learning_rate": 0.0006,
+ "loss": 7.189371109008789,
+ "step": 183
+ },
+ {
+ "epoch": 2.558951965065502,
+ "grad_norm": 0.017905596643686295,
+ "learning_rate": 0.0006,
+ "loss": 7.177547931671143,
+ "step": 184
+ },
+ {
+ "epoch": 2.5729257641921395,
+ "grad_norm": 0.024431610479950905,
+ "learning_rate": 0.0006,
+ "loss": 7.192169189453125,
+ "step": 185
+ },
+ {
+ "epoch": 2.5868995633187772,
+ "grad_norm": 0.03656961768865585,
+ "learning_rate": 0.0006,
+ "loss": 7.124387741088867,
+ "step": 186
+ },
+ {
+ "epoch": 2.600873362445415,
+ "grad_norm": 0.04337048903107643,
+ "learning_rate": 0.0006,
+ "loss": 7.140231609344482,
+ "step": 187
+ },
+ {
+ "epoch": 2.6148471615720523,
+ "grad_norm": 0.04047228768467903,
+ "learning_rate": 0.0006,
+ "loss": 7.171927452087402,
+ "step": 188
+ },
+ {
+ "epoch": 2.62882096069869,
+ "grad_norm": 0.07825770974159241,
+ "learning_rate": 0.0006,
+ "loss": 7.207631587982178,
+ "step": 189
+ },
+ {
+ "epoch": 2.6427947598253274,
+ "grad_norm": 0.07118063420057297,
+ "learning_rate": 0.0006,
+ "loss": 7.209488391876221,
+ "step": 190
+ },
+ {
+ "epoch": 2.656768558951965,
+ "grad_norm": 0.05153293535113335,
+ "learning_rate": 0.0006,
+ "loss": 7.14793062210083,
+ "step": 191
+ },
+ {
+ "epoch": 2.6707423580786025,
+ "grad_norm": 0.03318839520215988,
+ "learning_rate": 0.0006,
+ "loss": 7.163010597229004,
+ "step": 192
+ },
+ {
+ "epoch": 2.6847161572052403,
+ "grad_norm": 0.029333539307117462,
+ "learning_rate": 0.0006,
+ "loss": 7.159955978393555,
+ "step": 193
+ },
+ {
+ "epoch": 2.6986899563318776,
+ "grad_norm": 0.029405880719423294,
+ "learning_rate": 0.0006,
+ "loss": 7.136178493499756,
+ "step": 194
+ },
+ {
+ "epoch": 2.7126637554585153,
+ "grad_norm": 0.037815988063812256,
+ "learning_rate": 0.0006,
+ "loss": 7.183186054229736,
+ "step": 195
+ },
+ {
+ "epoch": 2.726637554585153,
+ "grad_norm": 0.020307250320911407,
+ "learning_rate": 0.0006,
+ "loss": 7.133164882659912,
+ "step": 196
+ },
+ {
+ "epoch": 2.7406113537117904,
+ "grad_norm": 0.026048069819808006,
+ "learning_rate": 0.0006,
+ "loss": 7.124977111816406,
+ "step": 197
+ },
+ {
+ "epoch": 2.7545851528384278,
+ "grad_norm": 0.027316724881529808,
+ "learning_rate": 0.0006,
+ "loss": 7.07265567779541,
+ "step": 198
+ },
+ {
+ "epoch": 2.7685589519650655,
+ "grad_norm": 0.021629009395837784,
+ "learning_rate": 0.0006,
+ "loss": 7.150020599365234,
+ "step": 199
+ },
+ {
+ "epoch": 2.7825327510917033,
+ "grad_norm": 0.01605929434299469,
+ "learning_rate": 0.0006,
+ "loss": 7.10919189453125,
+ "step": 200
+ },
+ {
+ "epoch": 2.7965065502183406,
+ "grad_norm": 0.020800089463591576,
+ "learning_rate": 0.0006,
+ "loss": 7.132045269012451,
+ "step": 201
+ },
+ {
+ "epoch": 2.810480349344978,
+ "grad_norm": 0.019225774332880974,
+ "learning_rate": 0.0006,
+ "loss": 7.128364562988281,
+ "step": 202
+ },
+ {
+ "epoch": 2.8244541484716157,
+ "grad_norm": 0.01561372634023428,
+ "learning_rate": 0.0006,
+ "loss": 7.083669662475586,
+ "step": 203
+ },
+ {
+ "epoch": 2.8384279475982535,
+ "grad_norm": 0.021969519555568695,
+ "learning_rate": 0.0006,
+ "loss": 7.050841331481934,
+ "step": 204
+ },
+ {
+ "epoch": 2.8524017467248908,
+ "grad_norm": 0.021453695371747017,
+ "learning_rate": 0.0006,
+ "loss": 7.097085475921631,
+ "step": 205
+ },
+ {
+ "epoch": 2.8663755458515285,
+ "grad_norm": 0.022836022078990936,
+ "learning_rate": 0.0006,
+ "loss": 7.10405969619751,
+ "step": 206
+ },
+ {
+ "epoch": 2.880349344978166,
+ "grad_norm": 0.02430851384997368,
+ "learning_rate": 0.0006,
+ "loss": 7.045181751251221,
+ "step": 207
+ },
+ {
+ "epoch": 2.8943231441048036,
+ "grad_norm": 0.020166993141174316,
+ "learning_rate": 0.0006,
+ "loss": 7.068183898925781,
+ "step": 208
+ },
+ {
+ "epoch": 2.908296943231441,
+ "grad_norm": 0.0118505684658885,
+ "learning_rate": 0.0006,
+ "loss": 7.07132625579834,
+ "step": 209
+ },
+ {
+ "epoch": 2.9222707423580787,
+ "grad_norm": 0.023772427812218666,
+ "learning_rate": 0.0006,
+ "loss": 7.055238723754883,
+ "step": 210
+ },
+ {
+ "epoch": 2.936244541484716,
+ "grad_norm": 0.02350712940096855,
+ "learning_rate": 0.0006,
+ "loss": 7.055898666381836,
+ "step": 211
+ },
+ {
+ "epoch": 2.950218340611354,
+ "grad_norm": 0.017244907096028328,
+ "learning_rate": 0.0006,
+ "loss": 7.0099711418151855,
+ "step": 212
+ },
+ {
+ "epoch": 2.964192139737991,
+ "grad_norm": 0.021565575152635574,
+ "learning_rate": 0.0006,
+ "loss": 6.991751670837402,
+ "step": 213
+ },
+ {
+ "epoch": 2.978165938864629,
+ "grad_norm": 0.03241860866546631,
+ "learning_rate": 0.0006,
+ "loss": 7.005980491638184,
+ "step": 214
+ },
+ {
+ "epoch": 2.992139737991266,
+ "grad_norm": 0.04613726586103439,
+ "learning_rate": 0.0006,
+ "loss": 7.027859687805176,
+ "step": 215
+ },
+ {
+ "epoch": 3.0,
+ "grad_norm": 0.06404894590377808,
+ "learning_rate": 0.0006,
+ "loss": 7.123725891113281,
+ "step": 216
+ },
+ {
+ "epoch": 3.0,
+ "eval_loss": 7.070647716522217,
+ "eval_runtime": 58.5801,
+ "eval_samples_per_second": 41.687,
+ "eval_steps_per_second": 1.314,
+ "step": 216
+ },
+ {
+ "epoch": 3.0139737991266378,
+ "grad_norm": 0.05542432889342308,
+ "learning_rate": 0.0006,
+ "loss": 7.078299522399902,
+ "step": 217
+ },
+ {
+ "epoch": 3.027947598253275,
+ "grad_norm": 0.026484526693820953,
+ "learning_rate": 0.0006,
+ "loss": 6.961159706115723,
+ "step": 218
+ },
+ {
+ "epoch": 3.041921397379913,
+ "grad_norm": 0.028276391327381134,
+ "learning_rate": 0.0006,
+ "loss": 7.019961833953857,
+ "step": 219
+ },
+ {
+ "epoch": 3.05589519650655,
+ "grad_norm": 0.024486595764756203,
+ "learning_rate": 0.0006,
+ "loss": 7.039677619934082,
+ "step": 220
+ },
+ {
+ "epoch": 3.069868995633188,
+ "grad_norm": 0.017620893195271492,
+ "learning_rate": 0.0006,
+ "loss": 6.9270782470703125,
+ "step": 221
+ },
+ {
+ "epoch": 3.0838427947598253,
+ "grad_norm": 0.02547302283346653,
+ "learning_rate": 0.0006,
+ "loss": 6.924233436584473,
+ "step": 222
+ },
+ {
+ "epoch": 3.097816593886463,
+ "grad_norm": 0.02539009042084217,
+ "learning_rate": 0.0006,
+ "loss": 6.973979949951172,
+ "step": 223
+ },
+ {
+ "epoch": 3.1117903930131003,
+ "grad_norm": 0.03256227448582649,
+ "learning_rate": 0.0006,
+ "loss": 6.933725357055664,
+ "step": 224
+ },
+ {
+ "epoch": 3.125764192139738,
+ "grad_norm": 0.04651800915598869,
+ "learning_rate": 0.0006,
+ "loss": 6.951044082641602,
+ "step": 225
+ },
+ {
+ "epoch": 3.1397379912663754,
+ "grad_norm": 0.04822975769639015,
+ "learning_rate": 0.0006,
+ "loss": 6.9614410400390625,
+ "step": 226
+ },
+ {
+ "epoch": 3.153711790393013,
+ "grad_norm": 0.031165020540356636,
+ "learning_rate": 0.0006,
+ "loss": 6.91193151473999,
+ "step": 227
+ },
+ {
+ "epoch": 3.1676855895196505,
+ "grad_norm": 0.029754292219877243,
+ "learning_rate": 0.0006,
+ "loss": 6.91802978515625,
+ "step": 228
+ },
+ {
+ "epoch": 3.1816593886462883,
+ "grad_norm": 0.02193152718245983,
+ "learning_rate": 0.0006,
+ "loss": 6.928090572357178,
+ "step": 229
+ },
+ {
+ "epoch": 3.1956331877729256,
+ "grad_norm": 0.02428470179438591,
+ "learning_rate": 0.0006,
+ "loss": 6.90396785736084,
+ "step": 230
+ },
+ {
+ "epoch": 3.2096069868995634,
+ "grad_norm": 0.023374345153570175,
+ "learning_rate": 0.0006,
+ "loss": 6.885858535766602,
+ "step": 231
+ },
+ {
+ "epoch": 3.2235807860262007,
+ "grad_norm": 0.023405000567436218,
+ "learning_rate": 0.0006,
+ "loss": 6.8855767250061035,
+ "step": 232
+ },
+ {
+ "epoch": 3.2375545851528384,
+ "grad_norm": 0.017277134582400322,
+ "learning_rate": 0.0006,
+ "loss": 6.901449203491211,
+ "step": 233
+ },
+ {
+ "epoch": 3.251528384279476,
+ "grad_norm": 0.019496750086545944,
+ "learning_rate": 0.0006,
+ "loss": 6.885687828063965,
+ "step": 234
+ },
+ {
+ "epoch": 3.2655021834061135,
+ "grad_norm": 0.016893276944756508,
+ "learning_rate": 0.0006,
+ "loss": 6.872369766235352,
+ "step": 235
+ },
+ {
+ "epoch": 3.279475982532751,
+ "grad_norm": 0.02058715932071209,
+ "learning_rate": 0.0006,
+ "loss": 6.823280334472656,
+ "step": 236
+ },
+ {
+ "epoch": 3.2934497816593886,
+ "grad_norm": 0.015530755743384361,
+ "learning_rate": 0.0006,
+ "loss": 6.860499858856201,
+ "step": 237
+ },
+ {
+ "epoch": 3.3074235807860264,
+ "grad_norm": 0.01934926211833954,
+ "learning_rate": 0.0006,
+ "loss": 6.817448616027832,
+ "step": 238
+ },
+ {
+ "epoch": 3.3213973799126637,
+ "grad_norm": 0.017413552850484848,
+ "learning_rate": 0.0006,
+ "loss": 6.881955146789551,
+ "step": 239
+ },
+ {
+ "epoch": 3.3353711790393015,
+ "grad_norm": 0.026478875428438187,
+ "learning_rate": 0.0006,
+ "loss": 6.827404499053955,
+ "step": 240
+ },
+ {
+ "epoch": 3.349344978165939,
+ "grad_norm": 0.032611701637506485,
+ "learning_rate": 0.0006,
+ "loss": 6.78285551071167,
+ "step": 241
+ },
+ {
+ "epoch": 3.3633187772925766,
+ "grad_norm": 0.041900306940078735,
+ "learning_rate": 0.0006,
+ "loss": 6.783053874969482,
+ "step": 242
+ },
+ {
+ "epoch": 3.377292576419214,
+ "grad_norm": 0.04149497672915459,
+ "learning_rate": 0.0006,
+ "loss": 6.745220184326172,
+ "step": 243
+ },
+ {
+ "epoch": 3.3912663755458516,
+ "grad_norm": 0.023949826136231422,
+ "learning_rate": 0.0006,
+ "loss": 6.841533184051514,
+ "step": 244
+ },
+ {
+ "epoch": 3.405240174672489,
+ "grad_norm": 0.028601842001080513,
+ "learning_rate": 0.0006,
+ "loss": 6.852852821350098,
+ "step": 245
+ },
+ {
+ "epoch": 3.4192139737991267,
+ "grad_norm": 0.027365155518054962,
+ "learning_rate": 0.0006,
+ "loss": 6.74376106262207,
+ "step": 246
+ },
+ {
+ "epoch": 3.433187772925764,
+ "grad_norm": 0.02856568619608879,
+ "learning_rate": 0.0006,
+ "loss": 6.74906063079834,
+ "step": 247
+ },
+ {
+ "epoch": 3.447161572052402,
+ "grad_norm": 0.019172416999936104,
+ "learning_rate": 0.0006,
+ "loss": 6.809206008911133,
+ "step": 248
+ },
+ {
+ "epoch": 3.461135371179039,
+ "grad_norm": 0.02161746844649315,
+ "learning_rate": 0.0006,
+ "loss": 6.789888381958008,
+ "step": 249
+ },
+ {
+ "epoch": 3.475109170305677,
+ "grad_norm": 0.020273666828870773,
+ "learning_rate": 0.0006,
+ "loss": 6.73512077331543,
+ "step": 250
+ },
+ {
+ "epoch": 3.489082969432314,
+ "grad_norm": 0.022329283878207207,
+ "learning_rate": 0.0006,
+ "loss": 6.826148509979248,
+ "step": 251
+ },
+ {
+ "epoch": 3.503056768558952,
+ "grad_norm": 0.02284284122288227,
+ "learning_rate": 0.0006,
+ "loss": 6.748521327972412,
+ "step": 252
+ },
+ {
+ "epoch": 3.5170305676855893,
+ "grad_norm": 0.026368247345089912,
+ "learning_rate": 0.0006,
+ "loss": 6.645879745483398,
+ "step": 253
+ },
+ {
+ "epoch": 3.531004366812227,
+ "grad_norm": 0.03627244383096695,
+ "learning_rate": 0.0006,
+ "loss": 6.666064739227295,
+ "step": 254
+ },
+ {
+ "epoch": 3.544978165938865,
+ "grad_norm": 0.04140935093164444,
+ "learning_rate": 0.0006,
+ "loss": 6.712275981903076,
+ "step": 255
+ },
+ {
+ "epoch": 3.558951965065502,
+ "grad_norm": 0.0390462800860405,
+ "learning_rate": 0.0006,
+ "loss": 6.665900230407715,
+ "step": 256
+ },
+ {
+ "epoch": 3.5729257641921395,
+ "grad_norm": 0.05178583785891533,
+ "learning_rate": 0.0006,
+ "loss": 6.753298759460449,
+ "step": 257
+ },
+ {
+ "epoch": 3.5868995633187772,
+ "grad_norm": 0.054328061640262604,
+ "learning_rate": 0.0006,
+ "loss": 6.732633113861084,
+ "step": 258
+ },
+ {
+ "epoch": 3.600873362445415,
+ "grad_norm": 0.04590460658073425,
+ "learning_rate": 0.0006,
+ "loss": 6.65250301361084,
+ "step": 259
+ },
+ {
+ "epoch": 3.6148471615720523,
+ "grad_norm": 0.044331666082143784,
+ "learning_rate": 0.0006,
+ "loss": 6.68654203414917,
+ "step": 260
+ },
+ {
+ "epoch": 3.62882096069869,
+ "grad_norm": 0.05308730527758598,
+ "learning_rate": 0.0006,
+ "loss": 6.642482757568359,
+ "step": 261
+ },
+ {
+ "epoch": 3.6427947598253274,
+ "grad_norm": 0.03999016433954239,
+ "learning_rate": 0.0006,
+ "loss": 6.625227451324463,
+ "step": 262
+ },
+ {
+ "epoch": 3.656768558951965,
+ "grad_norm": 0.04088086634874344,
+ "learning_rate": 0.0006,
+ "loss": 6.67623233795166,
+ "step": 263
+ },
+ {
+ "epoch": 3.6707423580786025,
+ "grad_norm": 0.03615636005997658,
+ "learning_rate": 0.0006,
+ "loss": 6.6869001388549805,
+ "step": 264
+ },
+ {
+ "epoch": 3.6847161572052403,
+ "grad_norm": 0.02817360684275627,
+ "learning_rate": 0.0006,
+ "loss": 6.674601078033447,
+ "step": 265
+ },
+ {
+ "epoch": 3.6986899563318776,
+ "grad_norm": 0.032586123794317245,
+ "learning_rate": 0.0006,
+ "loss": 6.612994194030762,
+ "step": 266
+ },
+ {
+ "epoch": 3.7126637554585153,
+ "grad_norm": 0.031117867678403854,
+ "learning_rate": 0.0006,
+ "loss": 6.6244354248046875,
+ "step": 267
+ },
+ {
+ "epoch": 3.726637554585153,
+ "grad_norm": 0.028330687433481216,
+ "learning_rate": 0.0006,
+ "loss": 6.669557094573975,
+ "step": 268
+ },
+ {
+ "epoch": 3.7406113537117904,
+ "grad_norm": 0.021489202976226807,
+ "learning_rate": 0.0006,
+ "loss": 6.632035732269287,
+ "step": 269
+ },
+ {
+ "epoch": 3.7545851528384278,
+ "grad_norm": 0.024269424378871918,
+ "learning_rate": 0.0006,
+ "loss": 6.660867691040039,
+ "step": 270
+ },
+ {
+ "epoch": 3.7685589519650655,
+ "grad_norm": 0.02237873338162899,
+ "learning_rate": 0.0006,
+ "loss": 6.602587699890137,
+ "step": 271
+ },
+ {
+ "epoch": 3.7825327510917033,
+ "grad_norm": 0.026503929868340492,
+ "learning_rate": 0.0006,
+ "loss": 6.599200248718262,
+ "step": 272
+ },
+ {
+ "epoch": 3.7965065502183406,
+ "grad_norm": 0.03687124326825142,
+ "learning_rate": 0.0006,
+ "loss": 6.567296504974365,
+ "step": 273
+ },
+ {
+ "epoch": 3.810480349344978,
+ "grad_norm": 0.04708952456712723,
+ "learning_rate": 0.0006,
+ "loss": 6.621265411376953,
+ "step": 274
+ },
+ {
+ "epoch": 3.8244541484716157,
+ "grad_norm": 0.041846614331007004,
+ "learning_rate": 0.0006,
+ "loss": 6.550536155700684,
+ "step": 275
+ },
+ {
+ "epoch": 3.8384279475982535,
+ "grad_norm": 0.031249675899744034,
+ "learning_rate": 0.0006,
+ "loss": 6.54278564453125,
+ "step": 276
+ },
+ {
+ "epoch": 3.8524017467248908,
+ "grad_norm": 0.023082256317138672,
+ "learning_rate": 0.0006,
+ "loss": 6.544787406921387,
+ "step": 277
+ },
+ {
+ "epoch": 3.8663755458515285,
+ "grad_norm": 0.020995570346713066,
+ "learning_rate": 0.0006,
+ "loss": 6.518143653869629,
+ "step": 278
+ },
+ {
+ "epoch": 3.880349344978166,
+ "grad_norm": 0.020839324221014977,
+ "learning_rate": 0.0006,
+ "loss": 6.546667098999023,
+ "step": 279
+ },
+ {
+ "epoch": 3.8943231441048036,
+ "grad_norm": 0.018377432599663734,
+ "learning_rate": 0.0006,
+ "loss": 6.478307723999023,
+ "step": 280
+ },
+ {
+ "epoch": 3.908296943231441,
+ "grad_norm": 0.021713724359869957,
+ "learning_rate": 0.0006,
+ "loss": 6.485574245452881,
+ "step": 281
+ },
+ {
+ "epoch": 3.9222707423580787,
+ "grad_norm": 0.01359301246702671,
+ "learning_rate": 0.0006,
+ "loss": 6.570268630981445,
+ "step": 282
+ },
+ {
+ "epoch": 3.936244541484716,
+ "grad_norm": 0.016233332455158234,
+ "learning_rate": 0.0006,
+ "loss": 6.57150936126709,
+ "step": 283
+ },
+ {
+ "epoch": 3.950218340611354,
+ "grad_norm": 0.020687608048319817,
+ "learning_rate": 0.0006,
+ "loss": 6.527956008911133,
+ "step": 284
+ },
+ {
+ "epoch": 3.964192139737991,
+ "grad_norm": 0.02032964862883091,
+ "learning_rate": 0.0006,
+ "loss": 6.468169689178467,
+ "step": 285
+ },
+ {
+ "epoch": 3.978165938864629,
+ "grad_norm": 0.016676288098096848,
+ "learning_rate": 0.0006,
+ "loss": 6.448418617248535,
+ "step": 286
+ },
+ {
+ "epoch": 3.992139737991266,
+ "grad_norm": 0.0168539360165596,
+ "learning_rate": 0.0006,
+ "loss": 6.461180686950684,
+ "step": 287
+ },
+ {
+ "epoch": 4.0,
+ "grad_norm": 0.02666233666241169,
+ "learning_rate": 0.0006,
+ "loss": 6.344893932342529,
+ "step": 288
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 28800,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 400,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": true,
+ "should_log": false,
+ "should_save": false,
+ "should_training_stop": false
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 1.2278453445328896e+17,
+ "train_batch_size": 8,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-288/training_args.bin b/runs/l2r50-i2-fulle-lm/checkpoint-288/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..cad232831c222cc7f8b938b8bbe42f61e7d89c3a
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-288/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:f7c40e3bfa2f5ac1450ca1243e3f2a895d44a06ed5995e6aefd97f78d229d8c8
+size 4792
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2880/chat_template.jinja b/runs/l2r50-i2-fulle-lm/checkpoint-2880/chat_template.jinja
new file mode 100644
index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-2880/chat_template.jinja
@@ -0,0 +1,85 @@
+{%- if tools %}
+ {{- '<|im_start|>system\n' }}
+ {%- if messages[0].role == 'system' %}
+ {{- messages[0].content + '\n\n' }}
+ {%- endif %}
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }}
+ {%- for tool in tools %}
+ {{- "\n" }}
+ {{- tool | tojson }}
+ {%- endfor %}
+ {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }}
+{%- else %}
+ {%- if messages[0].role == 'system' %}
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
+ {%- endif %}
+{%- endif %}
+{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
+{%- for message in messages[::-1] %}
+ {%- set index = (messages|length - 1) - loop.index0 %}
+ {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %}
+ {%- set ns.multi_step_tool = false %}
+ {%- set ns.last_query_index = index %}
+ {%- endif %}
+{%- endfor %}
+{%- for message in messages %}
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
+ {%- elif message.role == "assistant" %}
+ {%- set content = message.content %}
+ {%- set reasoning_content = '' %}
+ {%- if message.reasoning_content is defined and message.reasoning_content is not none %}
+ {%- set reasoning_content = message.reasoning_content %}
+ {%- else %}
+ {%- if '' in message.content %}
+ {%- set content = message.content.split('')[-1].lstrip('\n') %}
+ {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %}
+ {%- endif %}
+ {%- endif %}
+ {%- if loop.index0 > ns.last_query_index %}
+ {%- if loop.last or (not loop.last and reasoning_content) %}
+ {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- if message.tool_calls %}
+ {%- for tool_call in message.tool_calls %}
+ {%- if (loop.first and content) or (not loop.first) %}
+ {{- '\n' }}
+ {%- endif %}
+ {%- if tool_call.function %}
+ {%- set tool_call = tool_call.function %}
+ {%- endif %}
+ {{- '\n{"name": "' }}
+ {{- tool_call.name }}
+ {{- '", "arguments": ' }}
+ {%- if tool_call.arguments is string %}
+ {{- tool_call.arguments }}
+ {%- else %}
+ {{- tool_call.arguments | tojson }}
+ {%- endif %}
+ {{- '}\n' }}
+ {%- endfor %}
+ {%- endif %}
+ {{- '<|im_end|>\n' }}
+ {%- elif message.role == "tool" %}
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
+ {{- '<|im_start|>user' }}
+ {%- endif %}
+ {{- '\n\n' }}
+ {{- message.content }}
+ {{- '\n' }}
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
+ {{- '<|im_end|>\n' }}
+ {%- endif %}
+ {%- endif %}
+{%- endfor %}
+{%- if add_generation_prompt %}
+ {{- '<|im_start|>assistant\n' }}
+ {%- if enable_thinking is defined and enable_thinking is false %}
+ {{- '\n\n\n\n' }}
+ {%- endif %}
+{%- endif %}
\ No newline at end of file
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2880/config.json b/runs/l2r50-i2-fulle-lm/checkpoint-2880/config.json
new file mode 100644
index 0000000000000000000000000000000000000000..22f250953c6ce8fc91ee4411856dd73041a415e2
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-2880/config.json
@@ -0,0 +1,32 @@
+{
+ "architectures": [
+ "LlamaForCausalLM"
+ ],
+ "attention_bias": false,
+ "attention_dropout": 0.0,
+ "bos_token_id": null,
+ "dtype": "float32",
+ "eos_token_id": 151645,
+ "head_dim": 128,
+ "hidden_act": "silu",
+ "hidden_size": 512,
+ "initializer_range": 0.02,
+ "intermediate_size": 1536,
+ "max_position_embeddings": 2048,
+ "mlp_bias": false,
+ "model_type": "llama",
+ "num_attention_heads": 4,
+ "num_hidden_layers": 20,
+ "num_key_value_heads": 4,
+ "pad_token_id": 151645,
+ "pretraining_tp": 1,
+ "rms_norm_eps": 1e-06,
+ "rope_parameters": {
+ "rope_theta": 10000.0,
+ "rope_type": "default"
+ },
+ "tie_word_embeddings": true,
+ "transformers_version": "5.5.0",
+ "use_cache": false,
+ "vocab_size": 151673
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2880/generation_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-2880/generation_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-2880/generation_config.json
@@ -0,0 +1,11 @@
+{
+ "_from_model_config": true,
+ "eos_token_id": [
+ 151645
+ ],
+ "output_attentions": false,
+ "output_hidden_states": false,
+ "pad_token_id": 151645,
+ "transformers_version": "5.5.0",
+ "use_cache": true
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2880/model.safetensors b/runs/l2r50-i2-fulle-lm/checkpoint-2880/model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..7a51c5c1ccac4c8d7d276fc3783b893948605be4
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-2880/model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:1bb6f8b646626a86e79f911e3e834b2a10e372457b71dee43ebcd5f20db74bcf
+size 583360328
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2880/optimizer.pt b/runs/l2r50-i2-fulle-lm/checkpoint-2880/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..c8cfb763c02c5f0dcc46e80ccc2e2afcaaa2c2ff
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-2880/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:e2c25bdb61fa498e65d81733fbe7a6bc7da13085b041bd3dd615273bdbe1b41e
+size 1166833530
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2880/rng_state_0.pth b/runs/l2r50-i2-fulle-lm/checkpoint-2880/rng_state_0.pth
new file mode 100644
index 0000000000000000000000000000000000000000..46aa5d638d90b38e6e23e319e89d3ad4c2f9f4d2
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-2880/rng_state_0.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:5b9c0da4d03bbbcc4b70f3bfcb5cb2f9f49da5ef15eeea1bf8411056223551c1
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2880/rng_state_1.pth b/runs/l2r50-i2-fulle-lm/checkpoint-2880/rng_state_1.pth
new file mode 100644
index 0000000000000000000000000000000000000000..c94fd8f927e40f13f0dddd0206934032e8478afc
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-2880/rng_state_1.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:2a2e93731721a79a7062c5c650729198de80400972629e91fe965623d5298d09
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2880/rng_state_2.pth b/runs/l2r50-i2-fulle-lm/checkpoint-2880/rng_state_2.pth
new file mode 100644
index 0000000000000000000000000000000000000000..caabbc0e5bea10106f5ec02ac960e5854486d067
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-2880/rng_state_2.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:8ca72e3ce18e7d072c336c12a9ac59d6061211259635b7cea434b569953bcd7e
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2880/rng_state_3.pth b/runs/l2r50-i2-fulle-lm/checkpoint-2880/rng_state_3.pth
new file mode 100644
index 0000000000000000000000000000000000000000..c24519070b860089a9d5b1e9e5f6ef121eb04590
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-2880/rng_state_3.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:62b58282799b63651225abb40495daf3718399e4ae4f5d57549b5ba4ac2f8e58
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2880/scheduler.pt b/runs/l2r50-i2-fulle-lm/checkpoint-2880/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..0a5f79a713b34dd2d8d63f191629551c649dd599
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-2880/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:dc3ec3427ea842499c98ae42d325c555883f86219e2410d0c1a47e90ade53b1c
+size 1064
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2880/tokenizer.json b/runs/l2r50-i2-fulle-lm/checkpoint-2880/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..53786016f070e11184813f3c83101dc5e93e392e
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-2880/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:c3c63fda1646a555448e8f03662542e04b11d103721319e4d9b267d7da03ceb8
+size 11424463
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2880/tokenizer_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-2880/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..98e2fb9b2cc54785dc9d871b8e503f8b8f2defe7
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-2880/tokenizer_config.json
@@ -0,0 +1,20 @@
+{
+ "add_prefix_space": false,
+ "backend": "tokenizers",
+ "bos_token": null,
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "<|im_end|>",
+ "errors": "replace",
+ "extra_special_tokens": [
+ "<|l2r_pred|>",
+ "<|r2l_pred|>",
+ "<|next_1_pred|>",
+ "<|next_2_pred|>"
+ ],
+ "is_local": false,
+ "model_max_length": 131072,
+ "pad_token": "<|im_end|>",
+ "split_special_tokens": false,
+ "tokenizer_class": "Qwen2Tokenizer",
+ "unk_token": null
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2880/trainer_state.json b/runs/l2r50-i2-fulle-lm/checkpoint-2880/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..aebf5254a61e0691e8d39c066bb1d6807395e8df
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-2880/trainer_state.json
@@ -0,0 +1,20506 @@
+{
+ "best_global_step": null,
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 40.0,
+ "eval_steps": 500,
+ "global_step": 2880,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.013973799126637555,
+ "grad_norm": 0.1739506870508194,
+ "learning_rate": 0.0,
+ "loss": 11.99299430847168,
+ "step": 1
+ },
+ {
+ "epoch": 0.02794759825327511,
+ "grad_norm": 0.17216597497463226,
+ "learning_rate": 5.999999999999999e-06,
+ "loss": 11.988704681396484,
+ "step": 2
+ },
+ {
+ "epoch": 0.04192139737991266,
+ "grad_norm": 0.17459715902805328,
+ "learning_rate": 1.1999999999999999e-05,
+ "loss": 11.945581436157227,
+ "step": 3
+ },
+ {
+ "epoch": 0.05589519650655022,
+ "grad_norm": 0.17577414214611053,
+ "learning_rate": 1.7999999999999997e-05,
+ "loss": 11.859772682189941,
+ "step": 4
+ },
+ {
+ "epoch": 0.06986899563318777,
+ "grad_norm": 0.16163651645183563,
+ "learning_rate": 2.3999999999999997e-05,
+ "loss": 11.744415283203125,
+ "step": 5
+ },
+ {
+ "epoch": 0.08384279475982533,
+ "grad_norm": 0.14409932494163513,
+ "learning_rate": 2.9999999999999997e-05,
+ "loss": 11.618099212646484,
+ "step": 6
+ },
+ {
+ "epoch": 0.09781659388646288,
+ "grad_norm": 0.13079848885536194,
+ "learning_rate": 3.5999999999999994e-05,
+ "loss": 11.505935668945312,
+ "step": 7
+ },
+ {
+ "epoch": 0.11179039301310044,
+ "grad_norm": 0.11964613944292068,
+ "learning_rate": 4.2e-05,
+ "loss": 11.416532516479492,
+ "step": 8
+ },
+ {
+ "epoch": 0.125764192139738,
+ "grad_norm": 0.11345142871141434,
+ "learning_rate": 4.7999999999999994e-05,
+ "loss": 11.33206558227539,
+ "step": 9
+ },
+ {
+ "epoch": 0.13973799126637554,
+ "grad_norm": 0.11033952981233597,
+ "learning_rate": 5.399999999999999e-05,
+ "loss": 11.259052276611328,
+ "step": 10
+ },
+ {
+ "epoch": 0.1537117903930131,
+ "grad_norm": 0.10776489973068237,
+ "learning_rate": 5.9999999999999995e-05,
+ "loss": 11.203384399414062,
+ "step": 11
+ },
+ {
+ "epoch": 0.16768558951965065,
+ "grad_norm": 0.10721946507692337,
+ "learning_rate": 6.599999999999999e-05,
+ "loss": 11.153295516967773,
+ "step": 12
+ },
+ {
+ "epoch": 0.18165938864628822,
+ "grad_norm": 0.10673430562019348,
+ "learning_rate": 7.199999999999999e-05,
+ "loss": 11.110200881958008,
+ "step": 13
+ },
+ {
+ "epoch": 0.19563318777292577,
+ "grad_norm": 0.10630254447460175,
+ "learning_rate": 7.8e-05,
+ "loss": 11.067477226257324,
+ "step": 14
+ },
+ {
+ "epoch": 0.2096069868995633,
+ "grad_norm": 0.10631615668535233,
+ "learning_rate": 8.4e-05,
+ "loss": 11.022941589355469,
+ "step": 15
+ },
+ {
+ "epoch": 0.22358078602620088,
+ "grad_norm": 0.10609443485736847,
+ "learning_rate": 8.999999999999999e-05,
+ "loss": 10.970746040344238,
+ "step": 16
+ },
+ {
+ "epoch": 0.23755458515283842,
+ "grad_norm": 0.10648232698440552,
+ "learning_rate": 9.599999999999999e-05,
+ "loss": 10.91267204284668,
+ "step": 17
+ },
+ {
+ "epoch": 0.251528384279476,
+ "grad_norm": 0.10584589838981628,
+ "learning_rate": 0.000102,
+ "loss": 10.854547500610352,
+ "step": 18
+ },
+ {
+ "epoch": 0.26550218340611353,
+ "grad_norm": 0.10591720044612885,
+ "learning_rate": 0.00010799999999999998,
+ "loss": 10.788917541503906,
+ "step": 19
+ },
+ {
+ "epoch": 0.2794759825327511,
+ "grad_norm": 0.10611672699451447,
+ "learning_rate": 0.00011399999999999999,
+ "loss": 10.717556953430176,
+ "step": 20
+ },
+ {
+ "epoch": 0.2934497816593886,
+ "grad_norm": 0.10455359518527985,
+ "learning_rate": 0.00011999999999999999,
+ "loss": 10.650186538696289,
+ "step": 21
+ },
+ {
+ "epoch": 0.3074235807860262,
+ "grad_norm": 0.1047525629401207,
+ "learning_rate": 0.00012599999999999997,
+ "loss": 10.56808090209961,
+ "step": 22
+ },
+ {
+ "epoch": 0.32139737991266376,
+ "grad_norm": 0.10511605441570282,
+ "learning_rate": 0.00013199999999999998,
+ "loss": 10.484028816223145,
+ "step": 23
+ },
+ {
+ "epoch": 0.3353711790393013,
+ "grad_norm": 0.10467177629470825,
+ "learning_rate": 0.000138,
+ "loss": 10.399947166442871,
+ "step": 24
+ },
+ {
+ "epoch": 0.34934497816593885,
+ "grad_norm": 0.10432136058807373,
+ "learning_rate": 0.00014399999999999998,
+ "loss": 10.312714576721191,
+ "step": 25
+ },
+ {
+ "epoch": 0.36331877729257644,
+ "grad_norm": 0.10527262091636658,
+ "learning_rate": 0.00015,
+ "loss": 10.212261199951172,
+ "step": 26
+ },
+ {
+ "epoch": 0.377292576419214,
+ "grad_norm": 0.103610560297966,
+ "learning_rate": 0.000156,
+ "loss": 10.132034301757812,
+ "step": 27
+ },
+ {
+ "epoch": 0.39126637554585153,
+ "grad_norm": 0.10553059726953506,
+ "learning_rate": 0.000162,
+ "loss": 10.020434379577637,
+ "step": 28
+ },
+ {
+ "epoch": 0.4052401746724891,
+ "grad_norm": 0.1037299782037735,
+ "learning_rate": 0.000168,
+ "loss": 9.934466361999512,
+ "step": 29
+ },
+ {
+ "epoch": 0.4192139737991266,
+ "grad_norm": 0.1042739674448967,
+ "learning_rate": 0.00017399999999999997,
+ "loss": 9.822539329528809,
+ "step": 30
+ },
+ {
+ "epoch": 0.4331877729257642,
+ "grad_norm": 0.10300865024328232,
+ "learning_rate": 0.00017999999999999998,
+ "loss": 9.734994888305664,
+ "step": 31
+ },
+ {
+ "epoch": 0.44716157205240176,
+ "grad_norm": 0.10291654616594315,
+ "learning_rate": 0.000186,
+ "loss": 9.624505996704102,
+ "step": 32
+ },
+ {
+ "epoch": 0.4611353711790393,
+ "grad_norm": 0.10210851579904556,
+ "learning_rate": 0.00019199999999999998,
+ "loss": 9.518205642700195,
+ "step": 33
+ },
+ {
+ "epoch": 0.47510917030567684,
+ "grad_norm": 0.10222529619932175,
+ "learning_rate": 0.000198,
+ "loss": 9.412222862243652,
+ "step": 34
+ },
+ {
+ "epoch": 0.4890829694323144,
+ "grad_norm": 0.09992019087076187,
+ "learning_rate": 0.000204,
+ "loss": 9.326566696166992,
+ "step": 35
+ },
+ {
+ "epoch": 0.503056768558952,
+ "grad_norm": 0.09931084513664246,
+ "learning_rate": 0.00020999999999999998,
+ "loss": 9.210111618041992,
+ "step": 36
+ },
+ {
+ "epoch": 0.5170305676855895,
+ "grad_norm": 0.09928256273269653,
+ "learning_rate": 0.00021599999999999996,
+ "loss": 9.100493431091309,
+ "step": 37
+ },
+ {
+ "epoch": 0.5310043668122271,
+ "grad_norm": 0.09843816608190536,
+ "learning_rate": 0.00022199999999999998,
+ "loss": 8.996543884277344,
+ "step": 38
+ },
+ {
+ "epoch": 0.5449781659388646,
+ "grad_norm": 0.09825374186038971,
+ "learning_rate": 0.00022799999999999999,
+ "loss": 8.888991355895996,
+ "step": 39
+ },
+ {
+ "epoch": 0.5589519650655022,
+ "grad_norm": 0.09602385759353638,
+ "learning_rate": 0.000234,
+ "loss": 8.802907943725586,
+ "step": 40
+ },
+ {
+ "epoch": 0.5729257641921397,
+ "grad_norm": 0.09414064139127731,
+ "learning_rate": 0.00023999999999999998,
+ "loss": 8.715423583984375,
+ "step": 41
+ },
+ {
+ "epoch": 0.5868995633187772,
+ "grad_norm": 0.09273893386125565,
+ "learning_rate": 0.00024599999999999996,
+ "loss": 8.61374282836914,
+ "step": 42
+ },
+ {
+ "epoch": 0.6008733624454149,
+ "grad_norm": 0.09079599380493164,
+ "learning_rate": 0.00025199999999999995,
+ "loss": 8.502904891967773,
+ "step": 43
+ },
+ {
+ "epoch": 0.6148471615720524,
+ "grad_norm": 0.08605825901031494,
+ "learning_rate": 0.000258,
+ "loss": 8.453008651733398,
+ "step": 44
+ },
+ {
+ "epoch": 0.62882096069869,
+ "grad_norm": 0.08338996022939682,
+ "learning_rate": 0.00026399999999999997,
+ "loss": 8.365163803100586,
+ "step": 45
+ },
+ {
+ "epoch": 0.6427947598253275,
+ "grad_norm": 0.08150867372751236,
+ "learning_rate": 0.00027,
+ "loss": 8.285372734069824,
+ "step": 46
+ },
+ {
+ "epoch": 0.6567685589519651,
+ "grad_norm": 0.07794807851314545,
+ "learning_rate": 0.000276,
+ "loss": 8.185811996459961,
+ "step": 47
+ },
+ {
+ "epoch": 0.6707423580786026,
+ "grad_norm": 0.07592720538377762,
+ "learning_rate": 0.00028199999999999997,
+ "loss": 8.098482131958008,
+ "step": 48
+ },
+ {
+ "epoch": 0.6847161572052402,
+ "grad_norm": 0.0709897130727768,
+ "learning_rate": 0.00028799999999999995,
+ "loss": 8.042091369628906,
+ "step": 49
+ },
+ {
+ "epoch": 0.6986899563318777,
+ "grad_norm": 0.06582305580377579,
+ "learning_rate": 0.000294,
+ "loss": 7.972844123840332,
+ "step": 50
+ },
+ {
+ "epoch": 0.7126637554585152,
+ "grad_norm": 0.059452351182699203,
+ "learning_rate": 0.0003,
+ "loss": 7.924579620361328,
+ "step": 51
+ },
+ {
+ "epoch": 0.7266375545851529,
+ "grad_norm": 0.052426718175411224,
+ "learning_rate": 0.00030599999999999996,
+ "loss": 7.876336097717285,
+ "step": 52
+ },
+ {
+ "epoch": 0.7406113537117904,
+ "grad_norm": 0.04768150672316551,
+ "learning_rate": 0.000312,
+ "loss": 7.8102707862854,
+ "step": 53
+ },
+ {
+ "epoch": 0.754585152838428,
+ "grad_norm": 0.03970075398683548,
+ "learning_rate": 0.000318,
+ "loss": 7.791913986206055,
+ "step": 54
+ },
+ {
+ "epoch": 0.7685589519650655,
+ "grad_norm": 0.03559752553701401,
+ "learning_rate": 0.000324,
+ "loss": 7.756960868835449,
+ "step": 55
+ },
+ {
+ "epoch": 0.7825327510917031,
+ "grad_norm": 0.029321353882551193,
+ "learning_rate": 0.00033,
+ "loss": 7.724878787994385,
+ "step": 56
+ },
+ {
+ "epoch": 0.7965065502183406,
+ "grad_norm": 0.023485427722334862,
+ "learning_rate": 0.000336,
+ "loss": 7.72553014755249,
+ "step": 57
+ },
+ {
+ "epoch": 0.8104803493449781,
+ "grad_norm": 0.019319474697113037,
+ "learning_rate": 0.00034199999999999996,
+ "loss": 7.709493160247803,
+ "step": 58
+ },
+ {
+ "epoch": 0.8244541484716157,
+ "grad_norm": 0.016705691814422607,
+ "learning_rate": 0.00034799999999999995,
+ "loss": 7.6873297691345215,
+ "step": 59
+ },
+ {
+ "epoch": 0.8384279475982532,
+ "grad_norm": 0.017667734995484352,
+ "learning_rate": 0.00035399999999999993,
+ "loss": 7.677059173583984,
+ "step": 60
+ },
+ {
+ "epoch": 0.8524017467248908,
+ "grad_norm": 0.01671992614865303,
+ "learning_rate": 0.00035999999999999997,
+ "loss": 7.643512725830078,
+ "step": 61
+ },
+ {
+ "epoch": 0.8663755458515284,
+ "grad_norm": 0.012288172729313374,
+ "learning_rate": 0.00036599999999999995,
+ "loss": 7.658343315124512,
+ "step": 62
+ },
+ {
+ "epoch": 0.880349344978166,
+ "grad_norm": 0.012004776857793331,
+ "learning_rate": 0.000372,
+ "loss": 7.649759292602539,
+ "step": 63
+ },
+ {
+ "epoch": 0.8943231441048035,
+ "grad_norm": 0.012487445957958698,
+ "learning_rate": 0.00037799999999999997,
+ "loss": 7.631503105163574,
+ "step": 64
+ },
+ {
+ "epoch": 0.9082969432314411,
+ "grad_norm": 0.01489241048693657,
+ "learning_rate": 0.00038399999999999996,
+ "loss": 7.647110462188721,
+ "step": 65
+ },
+ {
+ "epoch": 0.9222707423580786,
+ "grad_norm": 0.01428980939090252,
+ "learning_rate": 0.00039,
+ "loss": 7.636983871459961,
+ "step": 66
+ },
+ {
+ "epoch": 0.9362445414847161,
+ "grad_norm": 0.010711144655942917,
+ "learning_rate": 0.000396,
+ "loss": 7.621130466461182,
+ "step": 67
+ },
+ {
+ "epoch": 0.9502183406113537,
+ "grad_norm": 0.010178886353969574,
+ "learning_rate": 0.000402,
+ "loss": 7.593924522399902,
+ "step": 68
+ },
+ {
+ "epoch": 0.9641921397379912,
+ "grad_norm": 0.010604874230921268,
+ "learning_rate": 0.000408,
+ "loss": 7.6015496253967285,
+ "step": 69
+ },
+ {
+ "epoch": 0.9781659388646288,
+ "grad_norm": 0.009751426056027412,
+ "learning_rate": 0.0004139999999999999,
+ "loss": 7.571964263916016,
+ "step": 70
+ },
+ {
+ "epoch": 0.9921397379912664,
+ "grad_norm": 0.009816481731832027,
+ "learning_rate": 0.00041999999999999996,
+ "loss": 7.57344913482666,
+ "step": 71
+ },
+ {
+ "epoch": 1.0,
+ "grad_norm": 0.008665637113153934,
+ "learning_rate": 0.00042599999999999995,
+ "loss": 7.579050064086914,
+ "step": 72
+ },
+ {
+ "epoch": 1.0,
+ "eval_loss": 7.581204891204834,
+ "eval_runtime": 59.0091,
+ "eval_samples_per_second": 41.383,
+ "eval_steps_per_second": 1.305,
+ "step": 72
+ },
+ {
+ "epoch": 1.0139737991266375,
+ "grad_norm": 0.009684201329946518,
+ "learning_rate": 0.00043199999999999993,
+ "loss": 7.568594932556152,
+ "step": 73
+ },
+ {
+ "epoch": 1.027947598253275,
+ "grad_norm": 0.025783710181713104,
+ "learning_rate": 0.00043799999999999997,
+ "loss": 7.578948497772217,
+ "step": 74
+ },
+ {
+ "epoch": 1.0419213973799126,
+ "grad_norm": 0.0327819399535656,
+ "learning_rate": 0.00044399999999999995,
+ "loss": 7.548173904418945,
+ "step": 75
+ },
+ {
+ "epoch": 1.0558951965065502,
+ "grad_norm": 0.021604498848319054,
+ "learning_rate": 0.00045,
+ "loss": 7.552953720092773,
+ "step": 76
+ },
+ {
+ "epoch": 1.0698689956331877,
+ "grad_norm": 0.010230105370283127,
+ "learning_rate": 0.00045599999999999997,
+ "loss": 7.5213212966918945,
+ "step": 77
+ },
+ {
+ "epoch": 1.0838427947598253,
+ "grad_norm": 0.023315589874982834,
+ "learning_rate": 0.00046199999999999995,
+ "loss": 7.501640319824219,
+ "step": 78
+ },
+ {
+ "epoch": 1.0978165938864628,
+ "grad_norm": 0.01109944935888052,
+ "learning_rate": 0.000468,
+ "loss": 7.539519786834717,
+ "step": 79
+ },
+ {
+ "epoch": 1.1117903930131003,
+ "grad_norm": 0.026093758642673492,
+ "learning_rate": 0.000474,
+ "loss": 7.525928974151611,
+ "step": 80
+ },
+ {
+ "epoch": 1.125764192139738,
+ "grad_norm": 0.03274895250797272,
+ "learning_rate": 0.00047999999999999996,
+ "loss": 7.535602569580078,
+ "step": 81
+ },
+ {
+ "epoch": 1.1397379912663754,
+ "grad_norm": 0.01543602254241705,
+ "learning_rate": 0.000486,
+ "loss": 7.513683795928955,
+ "step": 82
+ },
+ {
+ "epoch": 1.1537117903930132,
+ "grad_norm": 0.028377249836921692,
+ "learning_rate": 0.0004919999999999999,
+ "loss": 7.49384880065918,
+ "step": 83
+ },
+ {
+ "epoch": 1.1676855895196507,
+ "grad_norm": 0.04121481627225876,
+ "learning_rate": 0.000498,
+ "loss": 7.488447189331055,
+ "step": 84
+ },
+ {
+ "epoch": 1.1816593886462883,
+ "grad_norm": 0.013282302767038345,
+ "learning_rate": 0.0005039999999999999,
+ "loss": 7.463010787963867,
+ "step": 85
+ },
+ {
+ "epoch": 1.1956331877729258,
+ "grad_norm": 0.03200364485383034,
+ "learning_rate": 0.0005099999999999999,
+ "loss": 7.473515033721924,
+ "step": 86
+ },
+ {
+ "epoch": 1.2096069868995634,
+ "grad_norm": 0.020768029615283012,
+ "learning_rate": 0.000516,
+ "loss": 7.464688777923584,
+ "step": 87
+ },
+ {
+ "epoch": 1.223580786026201,
+ "grad_norm": 0.021715788170695305,
+ "learning_rate": 0.000522,
+ "loss": 7.4835028648376465,
+ "step": 88
+ },
+ {
+ "epoch": 1.2375545851528384,
+ "grad_norm": 0.026255754753947258,
+ "learning_rate": 0.0005279999999999999,
+ "loss": 7.443392753601074,
+ "step": 89
+ },
+ {
+ "epoch": 1.251528384279476,
+ "grad_norm": 0.022325951606035233,
+ "learning_rate": 0.000534,
+ "loss": 7.457373142242432,
+ "step": 90
+ },
+ {
+ "epoch": 1.2655021834061135,
+ "grad_norm": 0.009270387701690197,
+ "learning_rate": 0.00054,
+ "loss": 7.455986976623535,
+ "step": 91
+ },
+ {
+ "epoch": 1.279475982532751,
+ "grad_norm": 0.02369544468820095,
+ "learning_rate": 0.0005459999999999999,
+ "loss": 7.442216873168945,
+ "step": 92
+ },
+ {
+ "epoch": 1.2934497816593886,
+ "grad_norm": 0.008474506437778473,
+ "learning_rate": 0.000552,
+ "loss": 7.4294915199279785,
+ "step": 93
+ },
+ {
+ "epoch": 1.3074235807860262,
+ "grad_norm": 0.017043249681591988,
+ "learning_rate": 0.000558,
+ "loss": 7.406932830810547,
+ "step": 94
+ },
+ {
+ "epoch": 1.3213973799126637,
+ "grad_norm": 0.011813412420451641,
+ "learning_rate": 0.0005639999999999999,
+ "loss": 7.415900230407715,
+ "step": 95
+ },
+ {
+ "epoch": 1.3353711790393012,
+ "grad_norm": 0.01086883433163166,
+ "learning_rate": 0.00057,
+ "loss": 7.452998638153076,
+ "step": 96
+ },
+ {
+ "epoch": 1.3493449781659388,
+ "grad_norm": 0.014700951054692268,
+ "learning_rate": 0.0005759999999999999,
+ "loss": 7.410249710083008,
+ "step": 97
+ },
+ {
+ "epoch": 1.3633187772925766,
+ "grad_norm": 0.014512976631522179,
+ "learning_rate": 0.0005819999999999999,
+ "loss": 7.429132461547852,
+ "step": 98
+ },
+ {
+ "epoch": 1.3772925764192139,
+ "grad_norm": 0.012519920244812965,
+ "learning_rate": 0.000588,
+ "loss": 7.430817604064941,
+ "step": 99
+ },
+ {
+ "epoch": 1.3912663755458516,
+ "grad_norm": 0.007132664322853088,
+ "learning_rate": 0.0005939999999999999,
+ "loss": 7.405234336853027,
+ "step": 100
+ },
+ {
+ "epoch": 1.405240174672489,
+ "grad_norm": 0.019489388912916183,
+ "learning_rate": 0.0006,
+ "loss": 7.427791595458984,
+ "step": 101
+ },
+ {
+ "epoch": 1.4192139737991267,
+ "grad_norm": 0.027252664789557457,
+ "learning_rate": 0.0006,
+ "loss": 7.41761589050293,
+ "step": 102
+ },
+ {
+ "epoch": 1.4331877729257643,
+ "grad_norm": 0.02501637488603592,
+ "learning_rate": 0.0006,
+ "loss": 7.416735649108887,
+ "step": 103
+ },
+ {
+ "epoch": 1.4471615720524018,
+ "grad_norm": 0.00989772193133831,
+ "learning_rate": 0.0006,
+ "loss": 7.3756632804870605,
+ "step": 104
+ },
+ {
+ "epoch": 1.4611353711790394,
+ "grad_norm": 0.01622949168086052,
+ "learning_rate": 0.0006,
+ "loss": 7.400811672210693,
+ "step": 105
+ },
+ {
+ "epoch": 1.475109170305677,
+ "grad_norm": 0.012407763861119747,
+ "learning_rate": 0.0006,
+ "loss": 7.37675666809082,
+ "step": 106
+ },
+ {
+ "epoch": 1.4890829694323144,
+ "grad_norm": 0.010353565216064453,
+ "learning_rate": 0.0006,
+ "loss": 7.4013142585754395,
+ "step": 107
+ },
+ {
+ "epoch": 1.503056768558952,
+ "grad_norm": 0.0120553532615304,
+ "learning_rate": 0.0006,
+ "loss": 7.379873275756836,
+ "step": 108
+ },
+ {
+ "epoch": 1.5170305676855895,
+ "grad_norm": 0.01610225811600685,
+ "learning_rate": 0.0006,
+ "loss": 7.371631145477295,
+ "step": 109
+ },
+ {
+ "epoch": 1.531004366812227,
+ "grad_norm": 0.01159879844635725,
+ "learning_rate": 0.0006,
+ "loss": 7.389007568359375,
+ "step": 110
+ },
+ {
+ "epoch": 1.5449781659388646,
+ "grad_norm": 0.008013816550374031,
+ "learning_rate": 0.0006,
+ "loss": 7.395920753479004,
+ "step": 111
+ },
+ {
+ "epoch": 1.5589519650655022,
+ "grad_norm": 0.016334809362888336,
+ "learning_rate": 0.0006,
+ "loss": 7.393121719360352,
+ "step": 112
+ },
+ {
+ "epoch": 1.5729257641921397,
+ "grad_norm": 0.018821721896529198,
+ "learning_rate": 0.0006,
+ "loss": 7.367543697357178,
+ "step": 113
+ },
+ {
+ "epoch": 1.5868995633187772,
+ "grad_norm": 0.014088994823396206,
+ "learning_rate": 0.0006,
+ "loss": 7.385537624359131,
+ "step": 114
+ },
+ {
+ "epoch": 1.600873362445415,
+ "grad_norm": 0.005864050704985857,
+ "learning_rate": 0.0006,
+ "loss": 7.365346908569336,
+ "step": 115
+ },
+ {
+ "epoch": 1.6148471615720523,
+ "grad_norm": 0.011745430529117584,
+ "learning_rate": 0.0006,
+ "loss": 7.390227317810059,
+ "step": 116
+ },
+ {
+ "epoch": 1.62882096069869,
+ "grad_norm": 0.017205573618412018,
+ "learning_rate": 0.0006,
+ "loss": 7.3558549880981445,
+ "step": 117
+ },
+ {
+ "epoch": 1.6427947598253274,
+ "grad_norm": 0.021256346255540848,
+ "learning_rate": 0.0006,
+ "loss": 7.377346992492676,
+ "step": 118
+ },
+ {
+ "epoch": 1.6567685589519652,
+ "grad_norm": 0.019335204735398293,
+ "learning_rate": 0.0006,
+ "loss": 7.346194267272949,
+ "step": 119
+ },
+ {
+ "epoch": 1.6707423580786025,
+ "grad_norm": 0.008134003728628159,
+ "learning_rate": 0.0006,
+ "loss": 7.3063578605651855,
+ "step": 120
+ },
+ {
+ "epoch": 1.6847161572052403,
+ "grad_norm": 0.017613831907510757,
+ "learning_rate": 0.0006,
+ "loss": 7.338171005249023,
+ "step": 121
+ },
+ {
+ "epoch": 1.6986899563318776,
+ "grad_norm": 0.00890734139829874,
+ "learning_rate": 0.0006,
+ "loss": 7.355384826660156,
+ "step": 122
+ },
+ {
+ "epoch": 1.7126637554585153,
+ "grad_norm": 0.013448155485093594,
+ "learning_rate": 0.0006,
+ "loss": 7.344296455383301,
+ "step": 123
+ },
+ {
+ "epoch": 1.726637554585153,
+ "grad_norm": 0.009241633117198944,
+ "learning_rate": 0.0006,
+ "loss": 7.3454060554504395,
+ "step": 124
+ },
+ {
+ "epoch": 1.7406113537117904,
+ "grad_norm": 0.018561311066150665,
+ "learning_rate": 0.0006,
+ "loss": 7.32348108291626,
+ "step": 125
+ },
+ {
+ "epoch": 1.754585152838428,
+ "grad_norm": 0.025667704641819,
+ "learning_rate": 0.0006,
+ "loss": 7.356142520904541,
+ "step": 126
+ },
+ {
+ "epoch": 1.7685589519650655,
+ "grad_norm": 0.036232151091098785,
+ "learning_rate": 0.0006,
+ "loss": 7.363539695739746,
+ "step": 127
+ },
+ {
+ "epoch": 1.782532751091703,
+ "grad_norm": 0.059483714401721954,
+ "learning_rate": 0.0006,
+ "loss": 7.402331352233887,
+ "step": 128
+ },
+ {
+ "epoch": 1.7965065502183406,
+ "grad_norm": 0.05365009233355522,
+ "learning_rate": 0.0006,
+ "loss": 7.40457820892334,
+ "step": 129
+ },
+ {
+ "epoch": 1.8104803493449781,
+ "grad_norm": 0.02738415263593197,
+ "learning_rate": 0.0006,
+ "loss": 7.3756208419799805,
+ "step": 130
+ },
+ {
+ "epoch": 1.8244541484716157,
+ "grad_norm": 0.035017091780900955,
+ "learning_rate": 0.0006,
+ "loss": 7.379263877868652,
+ "step": 131
+ },
+ {
+ "epoch": 1.8384279475982532,
+ "grad_norm": 0.02819863148033619,
+ "learning_rate": 0.0006,
+ "loss": 7.327673435211182,
+ "step": 132
+ },
+ {
+ "epoch": 1.8524017467248908,
+ "grad_norm": 0.017754238098859787,
+ "learning_rate": 0.0006,
+ "loss": 7.333420276641846,
+ "step": 133
+ },
+ {
+ "epoch": 1.8663755458515285,
+ "grad_norm": 0.024685733020305634,
+ "learning_rate": 0.0006,
+ "loss": 7.34493350982666,
+ "step": 134
+ },
+ {
+ "epoch": 1.8803493449781659,
+ "grad_norm": 0.019426532089710236,
+ "learning_rate": 0.0006,
+ "loss": 7.328448295593262,
+ "step": 135
+ },
+ {
+ "epoch": 1.8943231441048036,
+ "grad_norm": 0.02106391079723835,
+ "learning_rate": 0.0006,
+ "loss": 7.327917098999023,
+ "step": 136
+ },
+ {
+ "epoch": 1.908296943231441,
+ "grad_norm": 0.02056879736483097,
+ "learning_rate": 0.0006,
+ "loss": 7.3311991691589355,
+ "step": 137
+ },
+ {
+ "epoch": 1.9222707423580787,
+ "grad_norm": 0.016865141689777374,
+ "learning_rate": 0.0006,
+ "loss": 7.312766075134277,
+ "step": 138
+ },
+ {
+ "epoch": 1.936244541484716,
+ "grad_norm": 0.017549855634570122,
+ "learning_rate": 0.0006,
+ "loss": 7.31191349029541,
+ "step": 139
+ },
+ {
+ "epoch": 1.9502183406113538,
+ "grad_norm": 0.015930157154798508,
+ "learning_rate": 0.0006,
+ "loss": 7.330901622772217,
+ "step": 140
+ },
+ {
+ "epoch": 1.9641921397379911,
+ "grad_norm": 0.013283558189868927,
+ "learning_rate": 0.0006,
+ "loss": 7.298318386077881,
+ "step": 141
+ },
+ {
+ "epoch": 1.9781659388646289,
+ "grad_norm": 0.012023469433188438,
+ "learning_rate": 0.0006,
+ "loss": 7.282337188720703,
+ "step": 142
+ },
+ {
+ "epoch": 1.9921397379912664,
+ "grad_norm": 0.01459511648863554,
+ "learning_rate": 0.0006,
+ "loss": 7.282924175262451,
+ "step": 143
+ },
+ {
+ "epoch": 2.0,
+ "grad_norm": 0.017474235966801643,
+ "learning_rate": 0.0006,
+ "loss": 7.254333972930908,
+ "step": 144
+ },
+ {
+ "epoch": 2.0,
+ "eval_loss": 7.318708896636963,
+ "eval_runtime": 59.6555,
+ "eval_samples_per_second": 40.935,
+ "eval_steps_per_second": 1.291,
+ "step": 144
+ },
+ {
+ "epoch": 2.0139737991266378,
+ "grad_norm": 0.019757507368922234,
+ "learning_rate": 0.0006,
+ "loss": 7.30288553237915,
+ "step": 145
+ },
+ {
+ "epoch": 2.027947598253275,
+ "grad_norm": 0.014571248553693295,
+ "learning_rate": 0.0006,
+ "loss": 7.274681568145752,
+ "step": 146
+ },
+ {
+ "epoch": 2.041921397379913,
+ "grad_norm": 0.015744198113679886,
+ "learning_rate": 0.0006,
+ "loss": 7.278632640838623,
+ "step": 147
+ },
+ {
+ "epoch": 2.05589519650655,
+ "grad_norm": 0.028048858046531677,
+ "learning_rate": 0.0006,
+ "loss": 7.26558780670166,
+ "step": 148
+ },
+ {
+ "epoch": 2.069868995633188,
+ "grad_norm": 0.06547307223081589,
+ "learning_rate": 0.0006,
+ "loss": 7.32567834854126,
+ "step": 149
+ },
+ {
+ "epoch": 2.0838427947598253,
+ "grad_norm": 0.05969763547182083,
+ "learning_rate": 0.0006,
+ "loss": 7.395836353302002,
+ "step": 150
+ },
+ {
+ "epoch": 2.097816593886463,
+ "grad_norm": 0.032574281096458435,
+ "learning_rate": 0.0006,
+ "loss": 7.318150997161865,
+ "step": 151
+ },
+ {
+ "epoch": 2.1117903930131003,
+ "grad_norm": 0.029452962800860405,
+ "learning_rate": 0.0006,
+ "loss": 7.308066368103027,
+ "step": 152
+ },
+ {
+ "epoch": 2.125764192139738,
+ "grad_norm": 0.03161991387605667,
+ "learning_rate": 0.0006,
+ "loss": 7.306290626525879,
+ "step": 153
+ },
+ {
+ "epoch": 2.1397379912663754,
+ "grad_norm": 0.027096295729279518,
+ "learning_rate": 0.0006,
+ "loss": 7.297796249389648,
+ "step": 154
+ },
+ {
+ "epoch": 2.153711790393013,
+ "grad_norm": 0.028657056391239166,
+ "learning_rate": 0.0006,
+ "loss": 7.317424774169922,
+ "step": 155
+ },
+ {
+ "epoch": 2.1676855895196505,
+ "grad_norm": 0.02046995982527733,
+ "learning_rate": 0.0006,
+ "loss": 7.297986030578613,
+ "step": 156
+ },
+ {
+ "epoch": 2.1816593886462883,
+ "grad_norm": 0.02220369316637516,
+ "learning_rate": 0.0006,
+ "loss": 7.268237590789795,
+ "step": 157
+ },
+ {
+ "epoch": 2.1956331877729256,
+ "grad_norm": 0.018381357192993164,
+ "learning_rate": 0.0006,
+ "loss": 7.259415626525879,
+ "step": 158
+ },
+ {
+ "epoch": 2.2096069868995634,
+ "grad_norm": 0.0204413291066885,
+ "learning_rate": 0.0006,
+ "loss": 7.258173942565918,
+ "step": 159
+ },
+ {
+ "epoch": 2.2235807860262007,
+ "grad_norm": 0.015874288976192474,
+ "learning_rate": 0.0006,
+ "loss": 7.281834602355957,
+ "step": 160
+ },
+ {
+ "epoch": 2.2375545851528384,
+ "grad_norm": 0.01634068600833416,
+ "learning_rate": 0.0006,
+ "loss": 7.2469868659973145,
+ "step": 161
+ },
+ {
+ "epoch": 2.251528384279476,
+ "grad_norm": 0.01728747971355915,
+ "learning_rate": 0.0006,
+ "loss": 7.280995845794678,
+ "step": 162
+ },
+ {
+ "epoch": 2.2655021834061135,
+ "grad_norm": 0.016341188922524452,
+ "learning_rate": 0.0006,
+ "loss": 7.273357391357422,
+ "step": 163
+ },
+ {
+ "epoch": 2.279475982532751,
+ "grad_norm": 0.016125807538628578,
+ "learning_rate": 0.0006,
+ "loss": 7.273510932922363,
+ "step": 164
+ },
+ {
+ "epoch": 2.2934497816593886,
+ "grad_norm": 0.01629287749528885,
+ "learning_rate": 0.0006,
+ "loss": 7.267029762268066,
+ "step": 165
+ },
+ {
+ "epoch": 2.3074235807860264,
+ "grad_norm": 0.015296131372451782,
+ "learning_rate": 0.0006,
+ "loss": 7.211989402770996,
+ "step": 166
+ },
+ {
+ "epoch": 2.3213973799126637,
+ "grad_norm": 0.01709570363163948,
+ "learning_rate": 0.0006,
+ "loss": 7.242851734161377,
+ "step": 167
+ },
+ {
+ "epoch": 2.3353711790393015,
+ "grad_norm": 0.012552388943731785,
+ "learning_rate": 0.0006,
+ "loss": 7.241696834564209,
+ "step": 168
+ },
+ {
+ "epoch": 2.349344978165939,
+ "grad_norm": 0.014571291394531727,
+ "learning_rate": 0.0006,
+ "loss": 7.270681381225586,
+ "step": 169
+ },
+ {
+ "epoch": 2.3633187772925766,
+ "grad_norm": 0.013389437459409237,
+ "learning_rate": 0.0006,
+ "loss": 7.191600799560547,
+ "step": 170
+ },
+ {
+ "epoch": 2.377292576419214,
+ "grad_norm": 0.011895314790308475,
+ "learning_rate": 0.0006,
+ "loss": 7.213610649108887,
+ "step": 171
+ },
+ {
+ "epoch": 2.3912663755458516,
+ "grad_norm": 0.01345642190426588,
+ "learning_rate": 0.0006,
+ "loss": 7.193787574768066,
+ "step": 172
+ },
+ {
+ "epoch": 2.405240174672489,
+ "grad_norm": 0.0091794328764081,
+ "learning_rate": 0.0006,
+ "loss": 7.233364105224609,
+ "step": 173
+ },
+ {
+ "epoch": 2.4192139737991267,
+ "grad_norm": 0.013767086900770664,
+ "learning_rate": 0.0006,
+ "loss": 7.217272758483887,
+ "step": 174
+ },
+ {
+ "epoch": 2.433187772925764,
+ "grad_norm": 0.015134960412979126,
+ "learning_rate": 0.0006,
+ "loss": 7.192791938781738,
+ "step": 175
+ },
+ {
+ "epoch": 2.447161572052402,
+ "grad_norm": 0.01790648326277733,
+ "learning_rate": 0.0006,
+ "loss": 7.229467391967773,
+ "step": 176
+ },
+ {
+ "epoch": 2.461135371179039,
+ "grad_norm": 0.015362377278506756,
+ "learning_rate": 0.0006,
+ "loss": 7.180346488952637,
+ "step": 177
+ },
+ {
+ "epoch": 2.475109170305677,
+ "grad_norm": 0.010505498386919498,
+ "learning_rate": 0.0006,
+ "loss": 7.188833236694336,
+ "step": 178
+ },
+ {
+ "epoch": 2.489082969432314,
+ "grad_norm": 0.009129747748374939,
+ "learning_rate": 0.0006,
+ "loss": 7.203153133392334,
+ "step": 179
+ },
+ {
+ "epoch": 2.503056768558952,
+ "grad_norm": 0.01147883478552103,
+ "learning_rate": 0.0006,
+ "loss": 7.163824558258057,
+ "step": 180
+ },
+ {
+ "epoch": 2.5170305676855893,
+ "grad_norm": 0.011338942684233189,
+ "learning_rate": 0.0006,
+ "loss": 7.16287899017334,
+ "step": 181
+ },
+ {
+ "epoch": 2.531004366812227,
+ "grad_norm": 0.013277675025165081,
+ "learning_rate": 0.0006,
+ "loss": 7.155905723571777,
+ "step": 182
+ },
+ {
+ "epoch": 2.544978165938865,
+ "grad_norm": 0.015269347466528416,
+ "learning_rate": 0.0006,
+ "loss": 7.189371109008789,
+ "step": 183
+ },
+ {
+ "epoch": 2.558951965065502,
+ "grad_norm": 0.017905596643686295,
+ "learning_rate": 0.0006,
+ "loss": 7.177547931671143,
+ "step": 184
+ },
+ {
+ "epoch": 2.5729257641921395,
+ "grad_norm": 0.024431610479950905,
+ "learning_rate": 0.0006,
+ "loss": 7.192169189453125,
+ "step": 185
+ },
+ {
+ "epoch": 2.5868995633187772,
+ "grad_norm": 0.03656961768865585,
+ "learning_rate": 0.0006,
+ "loss": 7.124387741088867,
+ "step": 186
+ },
+ {
+ "epoch": 2.600873362445415,
+ "grad_norm": 0.04337048903107643,
+ "learning_rate": 0.0006,
+ "loss": 7.140231609344482,
+ "step": 187
+ },
+ {
+ "epoch": 2.6148471615720523,
+ "grad_norm": 0.04047228768467903,
+ "learning_rate": 0.0006,
+ "loss": 7.171927452087402,
+ "step": 188
+ },
+ {
+ "epoch": 2.62882096069869,
+ "grad_norm": 0.07825770974159241,
+ "learning_rate": 0.0006,
+ "loss": 7.207631587982178,
+ "step": 189
+ },
+ {
+ "epoch": 2.6427947598253274,
+ "grad_norm": 0.07118063420057297,
+ "learning_rate": 0.0006,
+ "loss": 7.209488391876221,
+ "step": 190
+ },
+ {
+ "epoch": 2.656768558951965,
+ "grad_norm": 0.05153293535113335,
+ "learning_rate": 0.0006,
+ "loss": 7.14793062210083,
+ "step": 191
+ },
+ {
+ "epoch": 2.6707423580786025,
+ "grad_norm": 0.03318839520215988,
+ "learning_rate": 0.0006,
+ "loss": 7.163010597229004,
+ "step": 192
+ },
+ {
+ "epoch": 2.6847161572052403,
+ "grad_norm": 0.029333539307117462,
+ "learning_rate": 0.0006,
+ "loss": 7.159955978393555,
+ "step": 193
+ },
+ {
+ "epoch": 2.6986899563318776,
+ "grad_norm": 0.029405880719423294,
+ "learning_rate": 0.0006,
+ "loss": 7.136178493499756,
+ "step": 194
+ },
+ {
+ "epoch": 2.7126637554585153,
+ "grad_norm": 0.037815988063812256,
+ "learning_rate": 0.0006,
+ "loss": 7.183186054229736,
+ "step": 195
+ },
+ {
+ "epoch": 2.726637554585153,
+ "grad_norm": 0.020307250320911407,
+ "learning_rate": 0.0006,
+ "loss": 7.133164882659912,
+ "step": 196
+ },
+ {
+ "epoch": 2.7406113537117904,
+ "grad_norm": 0.026048069819808006,
+ "learning_rate": 0.0006,
+ "loss": 7.124977111816406,
+ "step": 197
+ },
+ {
+ "epoch": 2.7545851528384278,
+ "grad_norm": 0.027316724881529808,
+ "learning_rate": 0.0006,
+ "loss": 7.07265567779541,
+ "step": 198
+ },
+ {
+ "epoch": 2.7685589519650655,
+ "grad_norm": 0.021629009395837784,
+ "learning_rate": 0.0006,
+ "loss": 7.150020599365234,
+ "step": 199
+ },
+ {
+ "epoch": 2.7825327510917033,
+ "grad_norm": 0.01605929434299469,
+ "learning_rate": 0.0006,
+ "loss": 7.10919189453125,
+ "step": 200
+ },
+ {
+ "epoch": 2.7965065502183406,
+ "grad_norm": 0.020800089463591576,
+ "learning_rate": 0.0006,
+ "loss": 7.132045269012451,
+ "step": 201
+ },
+ {
+ "epoch": 2.810480349344978,
+ "grad_norm": 0.019225774332880974,
+ "learning_rate": 0.0006,
+ "loss": 7.128364562988281,
+ "step": 202
+ },
+ {
+ "epoch": 2.8244541484716157,
+ "grad_norm": 0.01561372634023428,
+ "learning_rate": 0.0006,
+ "loss": 7.083669662475586,
+ "step": 203
+ },
+ {
+ "epoch": 2.8384279475982535,
+ "grad_norm": 0.021969519555568695,
+ "learning_rate": 0.0006,
+ "loss": 7.050841331481934,
+ "step": 204
+ },
+ {
+ "epoch": 2.8524017467248908,
+ "grad_norm": 0.021453695371747017,
+ "learning_rate": 0.0006,
+ "loss": 7.097085475921631,
+ "step": 205
+ },
+ {
+ "epoch": 2.8663755458515285,
+ "grad_norm": 0.022836022078990936,
+ "learning_rate": 0.0006,
+ "loss": 7.10405969619751,
+ "step": 206
+ },
+ {
+ "epoch": 2.880349344978166,
+ "grad_norm": 0.02430851384997368,
+ "learning_rate": 0.0006,
+ "loss": 7.045181751251221,
+ "step": 207
+ },
+ {
+ "epoch": 2.8943231441048036,
+ "grad_norm": 0.020166993141174316,
+ "learning_rate": 0.0006,
+ "loss": 7.068183898925781,
+ "step": 208
+ },
+ {
+ "epoch": 2.908296943231441,
+ "grad_norm": 0.0118505684658885,
+ "learning_rate": 0.0006,
+ "loss": 7.07132625579834,
+ "step": 209
+ },
+ {
+ "epoch": 2.9222707423580787,
+ "grad_norm": 0.023772427812218666,
+ "learning_rate": 0.0006,
+ "loss": 7.055238723754883,
+ "step": 210
+ },
+ {
+ "epoch": 2.936244541484716,
+ "grad_norm": 0.02350712940096855,
+ "learning_rate": 0.0006,
+ "loss": 7.055898666381836,
+ "step": 211
+ },
+ {
+ "epoch": 2.950218340611354,
+ "grad_norm": 0.017244907096028328,
+ "learning_rate": 0.0006,
+ "loss": 7.0099711418151855,
+ "step": 212
+ },
+ {
+ "epoch": 2.964192139737991,
+ "grad_norm": 0.021565575152635574,
+ "learning_rate": 0.0006,
+ "loss": 6.991751670837402,
+ "step": 213
+ },
+ {
+ "epoch": 2.978165938864629,
+ "grad_norm": 0.03241860866546631,
+ "learning_rate": 0.0006,
+ "loss": 7.005980491638184,
+ "step": 214
+ },
+ {
+ "epoch": 2.992139737991266,
+ "grad_norm": 0.04613726586103439,
+ "learning_rate": 0.0006,
+ "loss": 7.027859687805176,
+ "step": 215
+ },
+ {
+ "epoch": 3.0,
+ "grad_norm": 0.06404894590377808,
+ "learning_rate": 0.0006,
+ "loss": 7.123725891113281,
+ "step": 216
+ },
+ {
+ "epoch": 3.0,
+ "eval_loss": 7.070647716522217,
+ "eval_runtime": 58.5801,
+ "eval_samples_per_second": 41.687,
+ "eval_steps_per_second": 1.314,
+ "step": 216
+ },
+ {
+ "epoch": 3.0139737991266378,
+ "grad_norm": 0.05542432889342308,
+ "learning_rate": 0.0006,
+ "loss": 7.078299522399902,
+ "step": 217
+ },
+ {
+ "epoch": 3.027947598253275,
+ "grad_norm": 0.026484526693820953,
+ "learning_rate": 0.0006,
+ "loss": 6.961159706115723,
+ "step": 218
+ },
+ {
+ "epoch": 3.041921397379913,
+ "grad_norm": 0.028276391327381134,
+ "learning_rate": 0.0006,
+ "loss": 7.019961833953857,
+ "step": 219
+ },
+ {
+ "epoch": 3.05589519650655,
+ "grad_norm": 0.024486595764756203,
+ "learning_rate": 0.0006,
+ "loss": 7.039677619934082,
+ "step": 220
+ },
+ {
+ "epoch": 3.069868995633188,
+ "grad_norm": 0.017620893195271492,
+ "learning_rate": 0.0006,
+ "loss": 6.9270782470703125,
+ "step": 221
+ },
+ {
+ "epoch": 3.0838427947598253,
+ "grad_norm": 0.02547302283346653,
+ "learning_rate": 0.0006,
+ "loss": 6.924233436584473,
+ "step": 222
+ },
+ {
+ "epoch": 3.097816593886463,
+ "grad_norm": 0.02539009042084217,
+ "learning_rate": 0.0006,
+ "loss": 6.973979949951172,
+ "step": 223
+ },
+ {
+ "epoch": 3.1117903930131003,
+ "grad_norm": 0.03256227448582649,
+ "learning_rate": 0.0006,
+ "loss": 6.933725357055664,
+ "step": 224
+ },
+ {
+ "epoch": 3.125764192139738,
+ "grad_norm": 0.04651800915598869,
+ "learning_rate": 0.0006,
+ "loss": 6.951044082641602,
+ "step": 225
+ },
+ {
+ "epoch": 3.1397379912663754,
+ "grad_norm": 0.04822975769639015,
+ "learning_rate": 0.0006,
+ "loss": 6.9614410400390625,
+ "step": 226
+ },
+ {
+ "epoch": 3.153711790393013,
+ "grad_norm": 0.031165020540356636,
+ "learning_rate": 0.0006,
+ "loss": 6.91193151473999,
+ "step": 227
+ },
+ {
+ "epoch": 3.1676855895196505,
+ "grad_norm": 0.029754292219877243,
+ "learning_rate": 0.0006,
+ "loss": 6.91802978515625,
+ "step": 228
+ },
+ {
+ "epoch": 3.1816593886462883,
+ "grad_norm": 0.02193152718245983,
+ "learning_rate": 0.0006,
+ "loss": 6.928090572357178,
+ "step": 229
+ },
+ {
+ "epoch": 3.1956331877729256,
+ "grad_norm": 0.02428470179438591,
+ "learning_rate": 0.0006,
+ "loss": 6.90396785736084,
+ "step": 230
+ },
+ {
+ "epoch": 3.2096069868995634,
+ "grad_norm": 0.023374345153570175,
+ "learning_rate": 0.0006,
+ "loss": 6.885858535766602,
+ "step": 231
+ },
+ {
+ "epoch": 3.2235807860262007,
+ "grad_norm": 0.023405000567436218,
+ "learning_rate": 0.0006,
+ "loss": 6.8855767250061035,
+ "step": 232
+ },
+ {
+ "epoch": 3.2375545851528384,
+ "grad_norm": 0.017277134582400322,
+ "learning_rate": 0.0006,
+ "loss": 6.901449203491211,
+ "step": 233
+ },
+ {
+ "epoch": 3.251528384279476,
+ "grad_norm": 0.019496750086545944,
+ "learning_rate": 0.0006,
+ "loss": 6.885687828063965,
+ "step": 234
+ },
+ {
+ "epoch": 3.2655021834061135,
+ "grad_norm": 0.016893276944756508,
+ "learning_rate": 0.0006,
+ "loss": 6.872369766235352,
+ "step": 235
+ },
+ {
+ "epoch": 3.279475982532751,
+ "grad_norm": 0.02058715932071209,
+ "learning_rate": 0.0006,
+ "loss": 6.823280334472656,
+ "step": 236
+ },
+ {
+ "epoch": 3.2934497816593886,
+ "grad_norm": 0.015530755743384361,
+ "learning_rate": 0.0006,
+ "loss": 6.860499858856201,
+ "step": 237
+ },
+ {
+ "epoch": 3.3074235807860264,
+ "grad_norm": 0.01934926211833954,
+ "learning_rate": 0.0006,
+ "loss": 6.817448616027832,
+ "step": 238
+ },
+ {
+ "epoch": 3.3213973799126637,
+ "grad_norm": 0.017413552850484848,
+ "learning_rate": 0.0006,
+ "loss": 6.881955146789551,
+ "step": 239
+ },
+ {
+ "epoch": 3.3353711790393015,
+ "grad_norm": 0.026478875428438187,
+ "learning_rate": 0.0006,
+ "loss": 6.827404499053955,
+ "step": 240
+ },
+ {
+ "epoch": 3.349344978165939,
+ "grad_norm": 0.032611701637506485,
+ "learning_rate": 0.0006,
+ "loss": 6.78285551071167,
+ "step": 241
+ },
+ {
+ "epoch": 3.3633187772925766,
+ "grad_norm": 0.041900306940078735,
+ "learning_rate": 0.0006,
+ "loss": 6.783053874969482,
+ "step": 242
+ },
+ {
+ "epoch": 3.377292576419214,
+ "grad_norm": 0.04149497672915459,
+ "learning_rate": 0.0006,
+ "loss": 6.745220184326172,
+ "step": 243
+ },
+ {
+ "epoch": 3.3912663755458516,
+ "grad_norm": 0.023949826136231422,
+ "learning_rate": 0.0006,
+ "loss": 6.841533184051514,
+ "step": 244
+ },
+ {
+ "epoch": 3.405240174672489,
+ "grad_norm": 0.028601842001080513,
+ "learning_rate": 0.0006,
+ "loss": 6.852852821350098,
+ "step": 245
+ },
+ {
+ "epoch": 3.4192139737991267,
+ "grad_norm": 0.027365155518054962,
+ "learning_rate": 0.0006,
+ "loss": 6.74376106262207,
+ "step": 246
+ },
+ {
+ "epoch": 3.433187772925764,
+ "grad_norm": 0.02856568619608879,
+ "learning_rate": 0.0006,
+ "loss": 6.74906063079834,
+ "step": 247
+ },
+ {
+ "epoch": 3.447161572052402,
+ "grad_norm": 0.019172416999936104,
+ "learning_rate": 0.0006,
+ "loss": 6.809206008911133,
+ "step": 248
+ },
+ {
+ "epoch": 3.461135371179039,
+ "grad_norm": 0.02161746844649315,
+ "learning_rate": 0.0006,
+ "loss": 6.789888381958008,
+ "step": 249
+ },
+ {
+ "epoch": 3.475109170305677,
+ "grad_norm": 0.020273666828870773,
+ "learning_rate": 0.0006,
+ "loss": 6.73512077331543,
+ "step": 250
+ },
+ {
+ "epoch": 3.489082969432314,
+ "grad_norm": 0.022329283878207207,
+ "learning_rate": 0.0006,
+ "loss": 6.826148509979248,
+ "step": 251
+ },
+ {
+ "epoch": 3.503056768558952,
+ "grad_norm": 0.02284284122288227,
+ "learning_rate": 0.0006,
+ "loss": 6.748521327972412,
+ "step": 252
+ },
+ {
+ "epoch": 3.5170305676855893,
+ "grad_norm": 0.026368247345089912,
+ "learning_rate": 0.0006,
+ "loss": 6.645879745483398,
+ "step": 253
+ },
+ {
+ "epoch": 3.531004366812227,
+ "grad_norm": 0.03627244383096695,
+ "learning_rate": 0.0006,
+ "loss": 6.666064739227295,
+ "step": 254
+ },
+ {
+ "epoch": 3.544978165938865,
+ "grad_norm": 0.04140935093164444,
+ "learning_rate": 0.0006,
+ "loss": 6.712275981903076,
+ "step": 255
+ },
+ {
+ "epoch": 3.558951965065502,
+ "grad_norm": 0.0390462800860405,
+ "learning_rate": 0.0006,
+ "loss": 6.665900230407715,
+ "step": 256
+ },
+ {
+ "epoch": 3.5729257641921395,
+ "grad_norm": 0.05178583785891533,
+ "learning_rate": 0.0006,
+ "loss": 6.753298759460449,
+ "step": 257
+ },
+ {
+ "epoch": 3.5868995633187772,
+ "grad_norm": 0.054328061640262604,
+ "learning_rate": 0.0006,
+ "loss": 6.732633113861084,
+ "step": 258
+ },
+ {
+ "epoch": 3.600873362445415,
+ "grad_norm": 0.04590460658073425,
+ "learning_rate": 0.0006,
+ "loss": 6.65250301361084,
+ "step": 259
+ },
+ {
+ "epoch": 3.6148471615720523,
+ "grad_norm": 0.044331666082143784,
+ "learning_rate": 0.0006,
+ "loss": 6.68654203414917,
+ "step": 260
+ },
+ {
+ "epoch": 3.62882096069869,
+ "grad_norm": 0.05308730527758598,
+ "learning_rate": 0.0006,
+ "loss": 6.642482757568359,
+ "step": 261
+ },
+ {
+ "epoch": 3.6427947598253274,
+ "grad_norm": 0.03999016433954239,
+ "learning_rate": 0.0006,
+ "loss": 6.625227451324463,
+ "step": 262
+ },
+ {
+ "epoch": 3.656768558951965,
+ "grad_norm": 0.04088086634874344,
+ "learning_rate": 0.0006,
+ "loss": 6.67623233795166,
+ "step": 263
+ },
+ {
+ "epoch": 3.6707423580786025,
+ "grad_norm": 0.03615636005997658,
+ "learning_rate": 0.0006,
+ "loss": 6.6869001388549805,
+ "step": 264
+ },
+ {
+ "epoch": 3.6847161572052403,
+ "grad_norm": 0.02817360684275627,
+ "learning_rate": 0.0006,
+ "loss": 6.674601078033447,
+ "step": 265
+ },
+ {
+ "epoch": 3.6986899563318776,
+ "grad_norm": 0.032586123794317245,
+ "learning_rate": 0.0006,
+ "loss": 6.612994194030762,
+ "step": 266
+ },
+ {
+ "epoch": 3.7126637554585153,
+ "grad_norm": 0.031117867678403854,
+ "learning_rate": 0.0006,
+ "loss": 6.6244354248046875,
+ "step": 267
+ },
+ {
+ "epoch": 3.726637554585153,
+ "grad_norm": 0.028330687433481216,
+ "learning_rate": 0.0006,
+ "loss": 6.669557094573975,
+ "step": 268
+ },
+ {
+ "epoch": 3.7406113537117904,
+ "grad_norm": 0.021489202976226807,
+ "learning_rate": 0.0006,
+ "loss": 6.632035732269287,
+ "step": 269
+ },
+ {
+ "epoch": 3.7545851528384278,
+ "grad_norm": 0.024269424378871918,
+ "learning_rate": 0.0006,
+ "loss": 6.660867691040039,
+ "step": 270
+ },
+ {
+ "epoch": 3.7685589519650655,
+ "grad_norm": 0.02237873338162899,
+ "learning_rate": 0.0006,
+ "loss": 6.602587699890137,
+ "step": 271
+ },
+ {
+ "epoch": 3.7825327510917033,
+ "grad_norm": 0.026503929868340492,
+ "learning_rate": 0.0006,
+ "loss": 6.599200248718262,
+ "step": 272
+ },
+ {
+ "epoch": 3.7965065502183406,
+ "grad_norm": 0.03687124326825142,
+ "learning_rate": 0.0006,
+ "loss": 6.567296504974365,
+ "step": 273
+ },
+ {
+ "epoch": 3.810480349344978,
+ "grad_norm": 0.04708952456712723,
+ "learning_rate": 0.0006,
+ "loss": 6.621265411376953,
+ "step": 274
+ },
+ {
+ "epoch": 3.8244541484716157,
+ "grad_norm": 0.041846614331007004,
+ "learning_rate": 0.0006,
+ "loss": 6.550536155700684,
+ "step": 275
+ },
+ {
+ "epoch": 3.8384279475982535,
+ "grad_norm": 0.031249675899744034,
+ "learning_rate": 0.0006,
+ "loss": 6.54278564453125,
+ "step": 276
+ },
+ {
+ "epoch": 3.8524017467248908,
+ "grad_norm": 0.023082256317138672,
+ "learning_rate": 0.0006,
+ "loss": 6.544787406921387,
+ "step": 277
+ },
+ {
+ "epoch": 3.8663755458515285,
+ "grad_norm": 0.020995570346713066,
+ "learning_rate": 0.0006,
+ "loss": 6.518143653869629,
+ "step": 278
+ },
+ {
+ "epoch": 3.880349344978166,
+ "grad_norm": 0.020839324221014977,
+ "learning_rate": 0.0006,
+ "loss": 6.546667098999023,
+ "step": 279
+ },
+ {
+ "epoch": 3.8943231441048036,
+ "grad_norm": 0.018377432599663734,
+ "learning_rate": 0.0006,
+ "loss": 6.478307723999023,
+ "step": 280
+ },
+ {
+ "epoch": 3.908296943231441,
+ "grad_norm": 0.021713724359869957,
+ "learning_rate": 0.0006,
+ "loss": 6.485574245452881,
+ "step": 281
+ },
+ {
+ "epoch": 3.9222707423580787,
+ "grad_norm": 0.01359301246702671,
+ "learning_rate": 0.0006,
+ "loss": 6.570268630981445,
+ "step": 282
+ },
+ {
+ "epoch": 3.936244541484716,
+ "grad_norm": 0.016233332455158234,
+ "learning_rate": 0.0006,
+ "loss": 6.57150936126709,
+ "step": 283
+ },
+ {
+ "epoch": 3.950218340611354,
+ "grad_norm": 0.020687608048319817,
+ "learning_rate": 0.0006,
+ "loss": 6.527956008911133,
+ "step": 284
+ },
+ {
+ "epoch": 3.964192139737991,
+ "grad_norm": 0.02032964862883091,
+ "learning_rate": 0.0006,
+ "loss": 6.468169689178467,
+ "step": 285
+ },
+ {
+ "epoch": 3.978165938864629,
+ "grad_norm": 0.016676288098096848,
+ "learning_rate": 0.0006,
+ "loss": 6.448418617248535,
+ "step": 286
+ },
+ {
+ "epoch": 3.992139737991266,
+ "grad_norm": 0.0168539360165596,
+ "learning_rate": 0.0006,
+ "loss": 6.461180686950684,
+ "step": 287
+ },
+ {
+ "epoch": 4.0,
+ "grad_norm": 0.02666233666241169,
+ "learning_rate": 0.0006,
+ "loss": 6.344893932342529,
+ "step": 288
+ },
+ {
+ "epoch": 4.0,
+ "eval_loss": 6.5384979248046875,
+ "eval_runtime": 58.814,
+ "eval_samples_per_second": 41.521,
+ "eval_steps_per_second": 1.309,
+ "step": 288
+ },
+ {
+ "epoch": 4.013973799126638,
+ "grad_norm": 0.04154708981513977,
+ "learning_rate": 0.0006,
+ "loss": 6.41855525970459,
+ "step": 289
+ },
+ {
+ "epoch": 4.0279475982532755,
+ "grad_norm": 0.04895668104290962,
+ "learning_rate": 0.0006,
+ "loss": 6.499302864074707,
+ "step": 290
+ },
+ {
+ "epoch": 4.041921397379912,
+ "grad_norm": 0.034464482218027115,
+ "learning_rate": 0.0006,
+ "loss": 6.438611030578613,
+ "step": 291
+ },
+ {
+ "epoch": 4.05589519650655,
+ "grad_norm": 0.06845773756504059,
+ "learning_rate": 0.0006,
+ "loss": 6.538361072540283,
+ "step": 292
+ },
+ {
+ "epoch": 4.069868995633188,
+ "grad_norm": 0.053768858313560486,
+ "learning_rate": 0.0006,
+ "loss": 6.527524948120117,
+ "step": 293
+ },
+ {
+ "epoch": 4.083842794759826,
+ "grad_norm": 0.039037931710481644,
+ "learning_rate": 0.0006,
+ "loss": 6.503847122192383,
+ "step": 294
+ },
+ {
+ "epoch": 4.097816593886463,
+ "grad_norm": 0.03962196782231331,
+ "learning_rate": 0.0006,
+ "loss": 6.475178241729736,
+ "step": 295
+ },
+ {
+ "epoch": 4.1117903930131,
+ "grad_norm": 0.049101535230875015,
+ "learning_rate": 0.0006,
+ "loss": 6.474587917327881,
+ "step": 296
+ },
+ {
+ "epoch": 4.125764192139738,
+ "grad_norm": 0.02830282226204872,
+ "learning_rate": 0.0006,
+ "loss": 6.45065975189209,
+ "step": 297
+ },
+ {
+ "epoch": 4.139737991266376,
+ "grad_norm": 0.023153482005000114,
+ "learning_rate": 0.0006,
+ "loss": 6.5218706130981445,
+ "step": 298
+ },
+ {
+ "epoch": 4.153711790393013,
+ "grad_norm": 0.025064002722501755,
+ "learning_rate": 0.0006,
+ "loss": 6.456475257873535,
+ "step": 299
+ },
+ {
+ "epoch": 4.1676855895196505,
+ "grad_norm": 0.02357092685997486,
+ "learning_rate": 0.0006,
+ "loss": 6.485952377319336,
+ "step": 300
+ },
+ {
+ "epoch": 4.181659388646288,
+ "grad_norm": 0.021832305938005447,
+ "learning_rate": 0.0006,
+ "loss": 6.425506591796875,
+ "step": 301
+ },
+ {
+ "epoch": 4.195633187772926,
+ "grad_norm": 0.0227900929749012,
+ "learning_rate": 0.0006,
+ "loss": 6.439155578613281,
+ "step": 302
+ },
+ {
+ "epoch": 4.209606986899563,
+ "grad_norm": 0.015961607918143272,
+ "learning_rate": 0.0006,
+ "loss": 6.460862636566162,
+ "step": 303
+ },
+ {
+ "epoch": 4.223580786026201,
+ "grad_norm": 0.017876390367746353,
+ "learning_rate": 0.0006,
+ "loss": 6.44564151763916,
+ "step": 304
+ },
+ {
+ "epoch": 4.2375545851528384,
+ "grad_norm": 0.013703204691410065,
+ "learning_rate": 0.0006,
+ "loss": 6.418414115905762,
+ "step": 305
+ },
+ {
+ "epoch": 4.251528384279476,
+ "grad_norm": 0.015240306034684181,
+ "learning_rate": 0.0006,
+ "loss": 6.411238670349121,
+ "step": 306
+ },
+ {
+ "epoch": 4.265502183406113,
+ "grad_norm": 0.01587662845849991,
+ "learning_rate": 0.0006,
+ "loss": 6.395642280578613,
+ "step": 307
+ },
+ {
+ "epoch": 4.279475982532751,
+ "grad_norm": 0.014650700613856316,
+ "learning_rate": 0.0006,
+ "loss": 6.302793025970459,
+ "step": 308
+ },
+ {
+ "epoch": 4.293449781659389,
+ "grad_norm": 0.016533225774765015,
+ "learning_rate": 0.0006,
+ "loss": 6.380588054656982,
+ "step": 309
+ },
+ {
+ "epoch": 4.307423580786026,
+ "grad_norm": 0.019963741302490234,
+ "learning_rate": 0.0006,
+ "loss": 6.345336437225342,
+ "step": 310
+ },
+ {
+ "epoch": 4.321397379912664,
+ "grad_norm": 0.02237936295568943,
+ "learning_rate": 0.0006,
+ "loss": 6.387770652770996,
+ "step": 311
+ },
+ {
+ "epoch": 4.335371179039301,
+ "grad_norm": 0.026658328250050545,
+ "learning_rate": 0.0006,
+ "loss": 6.323662757873535,
+ "step": 312
+ },
+ {
+ "epoch": 4.349344978165939,
+ "grad_norm": 0.02852284163236618,
+ "learning_rate": 0.0006,
+ "loss": 6.382204055786133,
+ "step": 313
+ },
+ {
+ "epoch": 4.3633187772925766,
+ "grad_norm": 0.028815090656280518,
+ "learning_rate": 0.0006,
+ "loss": 6.402390480041504,
+ "step": 314
+ },
+ {
+ "epoch": 4.377292576419214,
+ "grad_norm": 0.029393529519438744,
+ "learning_rate": 0.0006,
+ "loss": 6.280278205871582,
+ "step": 315
+ },
+ {
+ "epoch": 4.391266375545851,
+ "grad_norm": 0.02186041697859764,
+ "learning_rate": 0.0006,
+ "loss": 6.302707672119141,
+ "step": 316
+ },
+ {
+ "epoch": 4.405240174672489,
+ "grad_norm": 0.02071218006312847,
+ "learning_rate": 0.0006,
+ "loss": 6.3365020751953125,
+ "step": 317
+ },
+ {
+ "epoch": 4.419213973799127,
+ "grad_norm": 0.02779117226600647,
+ "learning_rate": 0.0006,
+ "loss": 6.252157211303711,
+ "step": 318
+ },
+ {
+ "epoch": 4.4331877729257645,
+ "grad_norm": 0.03470654785633087,
+ "learning_rate": 0.0006,
+ "loss": 6.319070339202881,
+ "step": 319
+ },
+ {
+ "epoch": 4.447161572052401,
+ "grad_norm": 0.037067804485559464,
+ "learning_rate": 0.0006,
+ "loss": 6.314116477966309,
+ "step": 320
+ },
+ {
+ "epoch": 4.461135371179039,
+ "grad_norm": 0.032182756811380386,
+ "learning_rate": 0.0006,
+ "loss": 6.267298698425293,
+ "step": 321
+ },
+ {
+ "epoch": 4.475109170305677,
+ "grad_norm": 0.026305217295885086,
+ "learning_rate": 0.0006,
+ "loss": 6.331688404083252,
+ "step": 322
+ },
+ {
+ "epoch": 4.489082969432315,
+ "grad_norm": 0.027228357270359993,
+ "learning_rate": 0.0006,
+ "loss": 6.31139612197876,
+ "step": 323
+ },
+ {
+ "epoch": 4.503056768558952,
+ "grad_norm": 0.025781169533729553,
+ "learning_rate": 0.0006,
+ "loss": 6.194684982299805,
+ "step": 324
+ },
+ {
+ "epoch": 4.517030567685589,
+ "grad_norm": 0.029186977073550224,
+ "learning_rate": 0.0006,
+ "loss": 6.31306266784668,
+ "step": 325
+ },
+ {
+ "epoch": 4.531004366812227,
+ "grad_norm": 0.023608777672052383,
+ "learning_rate": 0.0006,
+ "loss": 6.285243034362793,
+ "step": 326
+ },
+ {
+ "epoch": 4.544978165938865,
+ "grad_norm": 0.02267594076693058,
+ "learning_rate": 0.0006,
+ "loss": 6.30342960357666,
+ "step": 327
+ },
+ {
+ "epoch": 4.558951965065502,
+ "grad_norm": 0.01926310732960701,
+ "learning_rate": 0.0006,
+ "loss": 6.280606269836426,
+ "step": 328
+ },
+ {
+ "epoch": 4.5729257641921395,
+ "grad_norm": 0.026484837755560875,
+ "learning_rate": 0.0006,
+ "loss": 6.237980365753174,
+ "step": 329
+ },
+ {
+ "epoch": 4.586899563318777,
+ "grad_norm": 0.026228854432702065,
+ "learning_rate": 0.0006,
+ "loss": 6.3018293380737305,
+ "step": 330
+ },
+ {
+ "epoch": 4.600873362445415,
+ "grad_norm": 0.022096967324614525,
+ "learning_rate": 0.0006,
+ "loss": 6.253863334655762,
+ "step": 331
+ },
+ {
+ "epoch": 4.614847161572053,
+ "grad_norm": 0.027223890647292137,
+ "learning_rate": 0.0006,
+ "loss": 6.166882038116455,
+ "step": 332
+ },
+ {
+ "epoch": 4.62882096069869,
+ "grad_norm": 0.03160124272108078,
+ "learning_rate": 0.0006,
+ "loss": 6.230309963226318,
+ "step": 333
+ },
+ {
+ "epoch": 4.642794759825327,
+ "grad_norm": 0.04110539332032204,
+ "learning_rate": 0.0006,
+ "loss": 6.2334818840026855,
+ "step": 334
+ },
+ {
+ "epoch": 4.656768558951965,
+ "grad_norm": 0.04355933889746666,
+ "learning_rate": 0.0006,
+ "loss": 6.160956382751465,
+ "step": 335
+ },
+ {
+ "epoch": 4.670742358078603,
+ "grad_norm": 0.03611086308956146,
+ "learning_rate": 0.0006,
+ "loss": 6.23399543762207,
+ "step": 336
+ },
+ {
+ "epoch": 4.68471615720524,
+ "grad_norm": 0.051811400800943375,
+ "learning_rate": 0.0006,
+ "loss": 6.241855144500732,
+ "step": 337
+ },
+ {
+ "epoch": 4.698689956331878,
+ "grad_norm": 0.057434193789958954,
+ "learning_rate": 0.0006,
+ "loss": 6.220993995666504,
+ "step": 338
+ },
+ {
+ "epoch": 4.712663755458515,
+ "grad_norm": 0.04420953616499901,
+ "learning_rate": 0.0006,
+ "loss": 6.300461292266846,
+ "step": 339
+ },
+ {
+ "epoch": 4.726637554585153,
+ "grad_norm": 0.03515457361936569,
+ "learning_rate": 0.0006,
+ "loss": 6.269349098205566,
+ "step": 340
+ },
+ {
+ "epoch": 4.74061135371179,
+ "grad_norm": 0.03222600743174553,
+ "learning_rate": 0.0006,
+ "loss": 6.248613357543945,
+ "step": 341
+ },
+ {
+ "epoch": 4.754585152838428,
+ "grad_norm": 0.0384046845138073,
+ "learning_rate": 0.0006,
+ "loss": 6.266979217529297,
+ "step": 342
+ },
+ {
+ "epoch": 4.7685589519650655,
+ "grad_norm": 0.0514867827296257,
+ "learning_rate": 0.0006,
+ "loss": 6.234103202819824,
+ "step": 343
+ },
+ {
+ "epoch": 4.782532751091703,
+ "grad_norm": 0.050747793167829514,
+ "learning_rate": 0.0006,
+ "loss": 6.322582244873047,
+ "step": 344
+ },
+ {
+ "epoch": 4.796506550218341,
+ "grad_norm": 0.040878988802433014,
+ "learning_rate": 0.0006,
+ "loss": 6.212509632110596,
+ "step": 345
+ },
+ {
+ "epoch": 4.810480349344978,
+ "grad_norm": 0.034655820578336716,
+ "learning_rate": 0.0006,
+ "loss": 6.265879154205322,
+ "step": 346
+ },
+ {
+ "epoch": 4.824454148471616,
+ "grad_norm": 0.04202224686741829,
+ "learning_rate": 0.0006,
+ "loss": 6.2896409034729,
+ "step": 347
+ },
+ {
+ "epoch": 4.8384279475982535,
+ "grad_norm": 0.03632277995347977,
+ "learning_rate": 0.0006,
+ "loss": 6.253917217254639,
+ "step": 348
+ },
+ {
+ "epoch": 4.85240174672489,
+ "grad_norm": 0.028645765036344528,
+ "learning_rate": 0.0006,
+ "loss": 6.207965850830078,
+ "step": 349
+ },
+ {
+ "epoch": 4.866375545851528,
+ "grad_norm": 0.025305170565843582,
+ "learning_rate": 0.0006,
+ "loss": 6.168068885803223,
+ "step": 350
+ },
+ {
+ "epoch": 4.880349344978166,
+ "grad_norm": 0.022137803956866264,
+ "learning_rate": 0.0006,
+ "loss": 6.235283374786377,
+ "step": 351
+ },
+ {
+ "epoch": 4.894323144104804,
+ "grad_norm": 0.01814538985490799,
+ "learning_rate": 0.0006,
+ "loss": 6.259641647338867,
+ "step": 352
+ },
+ {
+ "epoch": 4.908296943231441,
+ "grad_norm": 0.018465086817741394,
+ "learning_rate": 0.0006,
+ "loss": 6.111738681793213,
+ "step": 353
+ },
+ {
+ "epoch": 4.922270742358078,
+ "grad_norm": 0.018638933077454567,
+ "learning_rate": 0.0006,
+ "loss": 6.225642204284668,
+ "step": 354
+ },
+ {
+ "epoch": 4.936244541484716,
+ "grad_norm": 0.016461260616779327,
+ "learning_rate": 0.0006,
+ "loss": 6.140647888183594,
+ "step": 355
+ },
+ {
+ "epoch": 4.950218340611354,
+ "grad_norm": 0.015458385460078716,
+ "learning_rate": 0.0006,
+ "loss": 6.155224800109863,
+ "step": 356
+ },
+ {
+ "epoch": 4.964192139737992,
+ "grad_norm": 0.014934048056602478,
+ "learning_rate": 0.0006,
+ "loss": 6.101555824279785,
+ "step": 357
+ },
+ {
+ "epoch": 4.978165938864628,
+ "grad_norm": 0.013979545794427395,
+ "learning_rate": 0.0006,
+ "loss": 6.161718845367432,
+ "step": 358
+ },
+ {
+ "epoch": 4.992139737991266,
+ "grad_norm": 0.01640394888818264,
+ "learning_rate": 0.0006,
+ "loss": 6.055768966674805,
+ "step": 359
+ },
+ {
+ "epoch": 5.0,
+ "grad_norm": 0.015135680325329304,
+ "learning_rate": 0.0006,
+ "loss": 6.169695854187012,
+ "step": 360
+ },
+ {
+ "epoch": 5.0,
+ "eval_loss": 6.16408109664917,
+ "eval_runtime": 59.3649,
+ "eval_samples_per_second": 41.135,
+ "eval_steps_per_second": 1.297,
+ "step": 360
+ },
+ {
+ "epoch": 5.013973799126638,
+ "grad_norm": 0.012849048711359501,
+ "learning_rate": 0.0006,
+ "loss": 6.055830001831055,
+ "step": 361
+ },
+ {
+ "epoch": 5.0279475982532755,
+ "grad_norm": 0.018169576302170753,
+ "learning_rate": 0.0006,
+ "loss": 6.082655429840088,
+ "step": 362
+ },
+ {
+ "epoch": 5.041921397379912,
+ "grad_norm": 0.024293430149555206,
+ "learning_rate": 0.0006,
+ "loss": 6.109032154083252,
+ "step": 363
+ },
+ {
+ "epoch": 5.05589519650655,
+ "grad_norm": 0.03266787901520729,
+ "learning_rate": 0.0006,
+ "loss": 6.030613899230957,
+ "step": 364
+ },
+ {
+ "epoch": 5.069868995633188,
+ "grad_norm": 0.03572266176342964,
+ "learning_rate": 0.0006,
+ "loss": 6.050390720367432,
+ "step": 365
+ },
+ {
+ "epoch": 5.083842794759826,
+ "grad_norm": 0.032625213265419006,
+ "learning_rate": 0.0006,
+ "loss": 6.169313430786133,
+ "step": 366
+ },
+ {
+ "epoch": 5.097816593886463,
+ "grad_norm": 0.04051872715353966,
+ "learning_rate": 0.0006,
+ "loss": 6.10398006439209,
+ "step": 367
+ },
+ {
+ "epoch": 5.1117903930131,
+ "grad_norm": 0.0351213738322258,
+ "learning_rate": 0.0006,
+ "loss": 6.121973514556885,
+ "step": 368
+ },
+ {
+ "epoch": 5.125764192139738,
+ "grad_norm": 0.03523759916424751,
+ "learning_rate": 0.0006,
+ "loss": 6.0626654624938965,
+ "step": 369
+ },
+ {
+ "epoch": 5.139737991266376,
+ "grad_norm": 0.02706182189285755,
+ "learning_rate": 0.0006,
+ "loss": 6.049354076385498,
+ "step": 370
+ },
+ {
+ "epoch": 5.153711790393013,
+ "grad_norm": 0.024212589487433434,
+ "learning_rate": 0.0006,
+ "loss": 6.034826278686523,
+ "step": 371
+ },
+ {
+ "epoch": 5.1676855895196505,
+ "grad_norm": 0.024798082187771797,
+ "learning_rate": 0.0006,
+ "loss": 6.004058837890625,
+ "step": 372
+ },
+ {
+ "epoch": 5.181659388646288,
+ "grad_norm": 0.02325567416846752,
+ "learning_rate": 0.0006,
+ "loss": 5.986461639404297,
+ "step": 373
+ },
+ {
+ "epoch": 5.195633187772926,
+ "grad_norm": 0.019060570746660233,
+ "learning_rate": 0.0006,
+ "loss": 6.059736251831055,
+ "step": 374
+ },
+ {
+ "epoch": 5.209606986899563,
+ "grad_norm": 0.016822976991534233,
+ "learning_rate": 0.0006,
+ "loss": 6.072957515716553,
+ "step": 375
+ },
+ {
+ "epoch": 5.223580786026201,
+ "grad_norm": 0.017218658700585365,
+ "learning_rate": 0.0006,
+ "loss": 6.031739234924316,
+ "step": 376
+ },
+ {
+ "epoch": 5.2375545851528384,
+ "grad_norm": 0.014889383688569069,
+ "learning_rate": 0.0006,
+ "loss": 6.025674343109131,
+ "step": 377
+ },
+ {
+ "epoch": 5.251528384279476,
+ "grad_norm": 0.01708107627928257,
+ "learning_rate": 0.0006,
+ "loss": 6.090451240539551,
+ "step": 378
+ },
+ {
+ "epoch": 5.265502183406113,
+ "grad_norm": 0.020260317251086235,
+ "learning_rate": 0.0006,
+ "loss": 6.094257354736328,
+ "step": 379
+ },
+ {
+ "epoch": 5.279475982532751,
+ "grad_norm": 0.020110400393605232,
+ "learning_rate": 0.0006,
+ "loss": 6.021188735961914,
+ "step": 380
+ },
+ {
+ "epoch": 5.293449781659389,
+ "grad_norm": 0.02027830481529236,
+ "learning_rate": 0.0006,
+ "loss": 6.022156238555908,
+ "step": 381
+ },
+ {
+ "epoch": 5.307423580786026,
+ "grad_norm": 0.022747062146663666,
+ "learning_rate": 0.0006,
+ "loss": 6.011836051940918,
+ "step": 382
+ },
+ {
+ "epoch": 5.321397379912664,
+ "grad_norm": 0.01990230567753315,
+ "learning_rate": 0.0006,
+ "loss": 6.003054618835449,
+ "step": 383
+ },
+ {
+ "epoch": 5.335371179039301,
+ "grad_norm": 0.01596238650381565,
+ "learning_rate": 0.0006,
+ "loss": 5.991410732269287,
+ "step": 384
+ },
+ {
+ "epoch": 5.349344978165939,
+ "grad_norm": 0.015847105532884598,
+ "learning_rate": 0.0006,
+ "loss": 6.100622177124023,
+ "step": 385
+ },
+ {
+ "epoch": 5.3633187772925766,
+ "grad_norm": 0.016179397702217102,
+ "learning_rate": 0.0006,
+ "loss": 6.032659530639648,
+ "step": 386
+ },
+ {
+ "epoch": 5.377292576419214,
+ "grad_norm": 0.018256602808833122,
+ "learning_rate": 0.0006,
+ "loss": 5.961983680725098,
+ "step": 387
+ },
+ {
+ "epoch": 5.391266375545851,
+ "grad_norm": 0.02805912122130394,
+ "learning_rate": 0.0006,
+ "loss": 5.983541965484619,
+ "step": 388
+ },
+ {
+ "epoch": 5.405240174672489,
+ "grad_norm": 0.039082255214452744,
+ "learning_rate": 0.0006,
+ "loss": 6.021474361419678,
+ "step": 389
+ },
+ {
+ "epoch": 5.419213973799127,
+ "grad_norm": 0.036757659167051315,
+ "learning_rate": 0.0006,
+ "loss": 6.005046367645264,
+ "step": 390
+ },
+ {
+ "epoch": 5.4331877729257645,
+ "grad_norm": 0.035277366638183594,
+ "learning_rate": 0.0006,
+ "loss": 6.036296844482422,
+ "step": 391
+ },
+ {
+ "epoch": 5.447161572052401,
+ "grad_norm": 0.034404635429382324,
+ "learning_rate": 0.0006,
+ "loss": 5.994539260864258,
+ "step": 392
+ },
+ {
+ "epoch": 5.461135371179039,
+ "grad_norm": 0.0377206988632679,
+ "learning_rate": 0.0006,
+ "loss": 5.941534519195557,
+ "step": 393
+ },
+ {
+ "epoch": 5.475109170305677,
+ "grad_norm": 0.0389922633767128,
+ "learning_rate": 0.0006,
+ "loss": 5.983644485473633,
+ "step": 394
+ },
+ {
+ "epoch": 5.489082969432315,
+ "grad_norm": 0.04176778718829155,
+ "learning_rate": 0.0006,
+ "loss": 6.030922889709473,
+ "step": 395
+ },
+ {
+ "epoch": 5.503056768558952,
+ "grad_norm": 0.0337153784930706,
+ "learning_rate": 0.0006,
+ "loss": 5.961367130279541,
+ "step": 396
+ },
+ {
+ "epoch": 5.517030567685589,
+ "grad_norm": 0.036116816103458405,
+ "learning_rate": 0.0006,
+ "loss": 5.819280624389648,
+ "step": 397
+ },
+ {
+ "epoch": 5.531004366812227,
+ "grad_norm": 0.030724041163921356,
+ "learning_rate": 0.0006,
+ "loss": 5.906380653381348,
+ "step": 398
+ },
+ {
+ "epoch": 5.544978165938865,
+ "grad_norm": 0.030264202505350113,
+ "learning_rate": 0.0006,
+ "loss": 5.975833415985107,
+ "step": 399
+ },
+ {
+ "epoch": 5.558951965065502,
+ "grad_norm": 0.032096318900585175,
+ "learning_rate": 0.0006,
+ "loss": 6.001348972320557,
+ "step": 400
+ },
+ {
+ "epoch": 5.5729257641921395,
+ "grad_norm": 0.030579356476664543,
+ "learning_rate": 0.0006,
+ "loss": 5.928768634796143,
+ "step": 401
+ },
+ {
+ "epoch": 5.586899563318777,
+ "grad_norm": 0.028241973370313644,
+ "learning_rate": 0.0006,
+ "loss": 5.92582893371582,
+ "step": 402
+ },
+ {
+ "epoch": 5.600873362445415,
+ "grad_norm": 0.023086953908205032,
+ "learning_rate": 0.0006,
+ "loss": 5.892926216125488,
+ "step": 403
+ },
+ {
+ "epoch": 5.614847161572053,
+ "grad_norm": 0.02547992393374443,
+ "learning_rate": 0.0006,
+ "loss": 5.903195381164551,
+ "step": 404
+ },
+ {
+ "epoch": 5.62882096069869,
+ "grad_norm": 0.023089831694960594,
+ "learning_rate": 0.0006,
+ "loss": 5.912033557891846,
+ "step": 405
+ },
+ {
+ "epoch": 5.642794759825327,
+ "grad_norm": 0.025446368381381035,
+ "learning_rate": 0.0006,
+ "loss": 5.926138401031494,
+ "step": 406
+ },
+ {
+ "epoch": 5.656768558951965,
+ "grad_norm": 0.03748747706413269,
+ "learning_rate": 0.0006,
+ "loss": 5.913451194763184,
+ "step": 407
+ },
+ {
+ "epoch": 5.670742358078603,
+ "grad_norm": 0.03493810072541237,
+ "learning_rate": 0.0006,
+ "loss": 5.97633695602417,
+ "step": 408
+ },
+ {
+ "epoch": 5.68471615720524,
+ "grad_norm": 0.023131275549530983,
+ "learning_rate": 0.0006,
+ "loss": 5.899571418762207,
+ "step": 409
+ },
+ {
+ "epoch": 5.698689956331878,
+ "grad_norm": 0.02986014075577259,
+ "learning_rate": 0.0006,
+ "loss": 5.894903182983398,
+ "step": 410
+ },
+ {
+ "epoch": 5.712663755458515,
+ "grad_norm": 0.030171144753694534,
+ "learning_rate": 0.0006,
+ "loss": 5.876595497131348,
+ "step": 411
+ },
+ {
+ "epoch": 5.726637554585153,
+ "grad_norm": 0.029377546161413193,
+ "learning_rate": 0.0006,
+ "loss": 5.911637306213379,
+ "step": 412
+ },
+ {
+ "epoch": 5.74061135371179,
+ "grad_norm": 0.029829490929841995,
+ "learning_rate": 0.0006,
+ "loss": 5.819629669189453,
+ "step": 413
+ },
+ {
+ "epoch": 5.754585152838428,
+ "grad_norm": 0.024139299988746643,
+ "learning_rate": 0.0006,
+ "loss": 5.81728458404541,
+ "step": 414
+ },
+ {
+ "epoch": 5.7685589519650655,
+ "grad_norm": 0.021453432738780975,
+ "learning_rate": 0.0006,
+ "loss": 5.987326145172119,
+ "step": 415
+ },
+ {
+ "epoch": 5.782532751091703,
+ "grad_norm": 0.017942246049642563,
+ "learning_rate": 0.0006,
+ "loss": 5.893008232116699,
+ "step": 416
+ },
+ {
+ "epoch": 5.796506550218341,
+ "grad_norm": 0.019724637269973755,
+ "learning_rate": 0.0006,
+ "loss": 5.912441253662109,
+ "step": 417
+ },
+ {
+ "epoch": 5.810480349344978,
+ "grad_norm": 0.018597912043333054,
+ "learning_rate": 0.0006,
+ "loss": 5.860394477844238,
+ "step": 418
+ },
+ {
+ "epoch": 5.824454148471616,
+ "grad_norm": 0.016232412308454514,
+ "learning_rate": 0.0006,
+ "loss": 5.94991397857666,
+ "step": 419
+ },
+ {
+ "epoch": 5.8384279475982535,
+ "grad_norm": 0.01667204685509205,
+ "learning_rate": 0.0006,
+ "loss": 5.8506364822387695,
+ "step": 420
+ },
+ {
+ "epoch": 5.85240174672489,
+ "grad_norm": 0.014232151210308075,
+ "learning_rate": 0.0006,
+ "loss": 5.928424835205078,
+ "step": 421
+ },
+ {
+ "epoch": 5.866375545851528,
+ "grad_norm": 0.01570476032793522,
+ "learning_rate": 0.0006,
+ "loss": 5.953692436218262,
+ "step": 422
+ },
+ {
+ "epoch": 5.880349344978166,
+ "grad_norm": 0.01586179807782173,
+ "learning_rate": 0.0006,
+ "loss": 5.820014476776123,
+ "step": 423
+ },
+ {
+ "epoch": 5.894323144104804,
+ "grad_norm": 0.014729145914316177,
+ "learning_rate": 0.0006,
+ "loss": 5.860154628753662,
+ "step": 424
+ },
+ {
+ "epoch": 5.908296943231441,
+ "grad_norm": 0.013111459091305733,
+ "learning_rate": 0.0006,
+ "loss": 5.85988187789917,
+ "step": 425
+ },
+ {
+ "epoch": 5.922270742358078,
+ "grad_norm": 0.012937922962009907,
+ "learning_rate": 0.0006,
+ "loss": 5.756265640258789,
+ "step": 426
+ },
+ {
+ "epoch": 5.936244541484716,
+ "grad_norm": 0.01377453189343214,
+ "learning_rate": 0.0006,
+ "loss": 5.8333048820495605,
+ "step": 427
+ },
+ {
+ "epoch": 5.950218340611354,
+ "grad_norm": 0.014054795727133751,
+ "learning_rate": 0.0006,
+ "loss": 5.803333759307861,
+ "step": 428
+ },
+ {
+ "epoch": 5.964192139737992,
+ "grad_norm": 0.01674959622323513,
+ "learning_rate": 0.0006,
+ "loss": 5.8718485832214355,
+ "step": 429
+ },
+ {
+ "epoch": 5.978165938864628,
+ "grad_norm": 0.018733017146587372,
+ "learning_rate": 0.0006,
+ "loss": 5.875979900360107,
+ "step": 430
+ },
+ {
+ "epoch": 5.992139737991266,
+ "grad_norm": 0.02088983915746212,
+ "learning_rate": 0.0006,
+ "loss": 5.763634204864502,
+ "step": 431
+ },
+ {
+ "epoch": 6.0,
+ "grad_norm": 0.026705985888838768,
+ "learning_rate": 0.0006,
+ "loss": 5.6922478675842285,
+ "step": 432
+ },
+ {
+ "epoch": 6.0,
+ "eval_loss": 5.847958087921143,
+ "eval_runtime": 58.5451,
+ "eval_samples_per_second": 41.711,
+ "eval_steps_per_second": 1.315,
+ "step": 432
+ },
+ {
+ "epoch": 6.013973799126638,
+ "grad_norm": 0.03181544691324234,
+ "learning_rate": 0.0006,
+ "loss": 5.786368370056152,
+ "step": 433
+ },
+ {
+ "epoch": 6.0279475982532755,
+ "grad_norm": 0.03238112851977348,
+ "learning_rate": 0.0006,
+ "loss": 5.810310363769531,
+ "step": 434
+ },
+ {
+ "epoch": 6.041921397379912,
+ "grad_norm": 0.03916756808757782,
+ "learning_rate": 0.0006,
+ "loss": 5.7761993408203125,
+ "step": 435
+ },
+ {
+ "epoch": 6.05589519650655,
+ "grad_norm": 0.044009730219841,
+ "learning_rate": 0.0006,
+ "loss": 5.898112773895264,
+ "step": 436
+ },
+ {
+ "epoch": 6.069868995633188,
+ "grad_norm": 0.045235246419906616,
+ "learning_rate": 0.0006,
+ "loss": 5.876371383666992,
+ "step": 437
+ },
+ {
+ "epoch": 6.083842794759826,
+ "grad_norm": 0.050218384712934494,
+ "learning_rate": 0.0006,
+ "loss": 5.805103302001953,
+ "step": 438
+ },
+ {
+ "epoch": 6.097816593886463,
+ "grad_norm": 0.05444490164518356,
+ "learning_rate": 0.0006,
+ "loss": 5.826424598693848,
+ "step": 439
+ },
+ {
+ "epoch": 6.1117903930131,
+ "grad_norm": 0.04890982061624527,
+ "learning_rate": 0.0006,
+ "loss": 5.7735395431518555,
+ "step": 440
+ },
+ {
+ "epoch": 6.125764192139738,
+ "grad_norm": 0.05478639155626297,
+ "learning_rate": 0.0006,
+ "loss": 5.823677062988281,
+ "step": 441
+ },
+ {
+ "epoch": 6.139737991266376,
+ "grad_norm": 0.057562313973903656,
+ "learning_rate": 0.0006,
+ "loss": 5.889334678649902,
+ "step": 442
+ },
+ {
+ "epoch": 6.153711790393013,
+ "grad_norm": 0.06447703391313553,
+ "learning_rate": 0.0006,
+ "loss": 5.84621524810791,
+ "step": 443
+ },
+ {
+ "epoch": 6.1676855895196505,
+ "grad_norm": 0.048861872404813766,
+ "learning_rate": 0.0006,
+ "loss": 5.931595802307129,
+ "step": 444
+ },
+ {
+ "epoch": 6.181659388646288,
+ "grad_norm": 0.05521339178085327,
+ "learning_rate": 0.0006,
+ "loss": 5.8229780197143555,
+ "step": 445
+ },
+ {
+ "epoch": 6.195633187772926,
+ "grad_norm": 0.053666990250349045,
+ "learning_rate": 0.0006,
+ "loss": 5.879191875457764,
+ "step": 446
+ },
+ {
+ "epoch": 6.209606986899563,
+ "grad_norm": 0.0451025515794754,
+ "learning_rate": 0.0006,
+ "loss": 5.801628112792969,
+ "step": 447
+ },
+ {
+ "epoch": 6.223580786026201,
+ "grad_norm": 0.04291655868291855,
+ "learning_rate": 0.0006,
+ "loss": 5.914680480957031,
+ "step": 448
+ },
+ {
+ "epoch": 6.2375545851528384,
+ "grad_norm": 0.035102490335702896,
+ "learning_rate": 0.0006,
+ "loss": 5.843682765960693,
+ "step": 449
+ },
+ {
+ "epoch": 6.251528384279476,
+ "grad_norm": 0.03403995931148529,
+ "learning_rate": 0.0006,
+ "loss": 5.827154159545898,
+ "step": 450
+ },
+ {
+ "epoch": 6.265502183406113,
+ "grad_norm": 0.03444375470280647,
+ "learning_rate": 0.0006,
+ "loss": 5.800136566162109,
+ "step": 451
+ },
+ {
+ "epoch": 6.279475982532751,
+ "grad_norm": 0.03165159001946449,
+ "learning_rate": 0.0006,
+ "loss": 5.906252384185791,
+ "step": 452
+ },
+ {
+ "epoch": 6.293449781659389,
+ "grad_norm": 0.026153093203902245,
+ "learning_rate": 0.0006,
+ "loss": 5.7423200607299805,
+ "step": 453
+ },
+ {
+ "epoch": 6.307423580786026,
+ "grad_norm": 0.024357657879590988,
+ "learning_rate": 0.0006,
+ "loss": 5.701364517211914,
+ "step": 454
+ },
+ {
+ "epoch": 6.321397379912664,
+ "grad_norm": 0.021508049219846725,
+ "learning_rate": 0.0006,
+ "loss": 5.8328094482421875,
+ "step": 455
+ },
+ {
+ "epoch": 6.335371179039301,
+ "grad_norm": 0.017313921824097633,
+ "learning_rate": 0.0006,
+ "loss": 5.792145252227783,
+ "step": 456
+ },
+ {
+ "epoch": 6.349344978165939,
+ "grad_norm": 0.018563397228717804,
+ "learning_rate": 0.0006,
+ "loss": 5.732672214508057,
+ "step": 457
+ },
+ {
+ "epoch": 6.3633187772925766,
+ "grad_norm": 0.016568679362535477,
+ "learning_rate": 0.0006,
+ "loss": 5.81948184967041,
+ "step": 458
+ },
+ {
+ "epoch": 6.377292576419214,
+ "grad_norm": 0.014133770950138569,
+ "learning_rate": 0.0006,
+ "loss": 5.734982490539551,
+ "step": 459
+ },
+ {
+ "epoch": 6.391266375545851,
+ "grad_norm": 0.014114447869360447,
+ "learning_rate": 0.0006,
+ "loss": 5.70073127746582,
+ "step": 460
+ },
+ {
+ "epoch": 6.405240174672489,
+ "grad_norm": 0.012907393276691437,
+ "learning_rate": 0.0006,
+ "loss": 5.758626937866211,
+ "step": 461
+ },
+ {
+ "epoch": 6.419213973799127,
+ "grad_norm": 0.012673444114625454,
+ "learning_rate": 0.0006,
+ "loss": 5.775138854980469,
+ "step": 462
+ },
+ {
+ "epoch": 6.4331877729257645,
+ "grad_norm": 0.01311410591006279,
+ "learning_rate": 0.0006,
+ "loss": 5.70920991897583,
+ "step": 463
+ },
+ {
+ "epoch": 6.447161572052401,
+ "grad_norm": 0.012935544364154339,
+ "learning_rate": 0.0006,
+ "loss": 5.781479835510254,
+ "step": 464
+ },
+ {
+ "epoch": 6.461135371179039,
+ "grad_norm": 0.01439738366752863,
+ "learning_rate": 0.0006,
+ "loss": 5.691315650939941,
+ "step": 465
+ },
+ {
+ "epoch": 6.475109170305677,
+ "grad_norm": 0.01205186452716589,
+ "learning_rate": 0.0006,
+ "loss": 5.662499904632568,
+ "step": 466
+ },
+ {
+ "epoch": 6.489082969432315,
+ "grad_norm": 0.011575652286410332,
+ "learning_rate": 0.0006,
+ "loss": 5.6271772384643555,
+ "step": 467
+ },
+ {
+ "epoch": 6.503056768558952,
+ "grad_norm": 0.011672502383589745,
+ "learning_rate": 0.0006,
+ "loss": 5.761662006378174,
+ "step": 468
+ },
+ {
+ "epoch": 6.517030567685589,
+ "grad_norm": 0.011541751213371754,
+ "learning_rate": 0.0006,
+ "loss": 5.762078285217285,
+ "step": 469
+ },
+ {
+ "epoch": 6.531004366812227,
+ "grad_norm": 0.011396365240216255,
+ "learning_rate": 0.0006,
+ "loss": 5.67873477935791,
+ "step": 470
+ },
+ {
+ "epoch": 6.544978165938865,
+ "grad_norm": 0.010688499547541142,
+ "learning_rate": 0.0006,
+ "loss": 5.717051982879639,
+ "step": 471
+ },
+ {
+ "epoch": 6.558951965065502,
+ "grad_norm": 0.012224317528307438,
+ "learning_rate": 0.0006,
+ "loss": 5.707948684692383,
+ "step": 472
+ },
+ {
+ "epoch": 6.5729257641921395,
+ "grad_norm": 0.011856825090944767,
+ "learning_rate": 0.0006,
+ "loss": 5.70878791809082,
+ "step": 473
+ },
+ {
+ "epoch": 6.586899563318777,
+ "grad_norm": 0.01199064590036869,
+ "learning_rate": 0.0006,
+ "loss": 5.708697319030762,
+ "step": 474
+ },
+ {
+ "epoch": 6.600873362445415,
+ "grad_norm": 0.01219446212053299,
+ "learning_rate": 0.0006,
+ "loss": 5.6061577796936035,
+ "step": 475
+ },
+ {
+ "epoch": 6.614847161572053,
+ "grad_norm": 0.013397484086453915,
+ "learning_rate": 0.0006,
+ "loss": 5.624789714813232,
+ "step": 476
+ },
+ {
+ "epoch": 6.62882096069869,
+ "grad_norm": 0.01483136136084795,
+ "learning_rate": 0.0006,
+ "loss": 5.721141338348389,
+ "step": 477
+ },
+ {
+ "epoch": 6.642794759825327,
+ "grad_norm": 0.0194488987326622,
+ "learning_rate": 0.0006,
+ "loss": 5.604279518127441,
+ "step": 478
+ },
+ {
+ "epoch": 6.656768558951965,
+ "grad_norm": 0.02079896256327629,
+ "learning_rate": 0.0006,
+ "loss": 5.586322784423828,
+ "step": 479
+ },
+ {
+ "epoch": 6.670742358078603,
+ "grad_norm": 0.018996229395270348,
+ "learning_rate": 0.0006,
+ "loss": 5.5987772941589355,
+ "step": 480
+ },
+ {
+ "epoch": 6.68471615720524,
+ "grad_norm": 0.015779603272676468,
+ "learning_rate": 0.0006,
+ "loss": 5.588602066040039,
+ "step": 481
+ },
+ {
+ "epoch": 6.698689956331878,
+ "grad_norm": 0.015322973020374775,
+ "learning_rate": 0.0006,
+ "loss": 5.686467170715332,
+ "step": 482
+ },
+ {
+ "epoch": 6.712663755458515,
+ "grad_norm": 0.015282213687896729,
+ "learning_rate": 0.0006,
+ "loss": 5.664676666259766,
+ "step": 483
+ },
+ {
+ "epoch": 6.726637554585153,
+ "grad_norm": 0.016589272767305374,
+ "learning_rate": 0.0006,
+ "loss": 5.666739463806152,
+ "step": 484
+ },
+ {
+ "epoch": 6.74061135371179,
+ "grad_norm": 0.01811421848833561,
+ "learning_rate": 0.0006,
+ "loss": 5.647593021392822,
+ "step": 485
+ },
+ {
+ "epoch": 6.754585152838428,
+ "grad_norm": 0.019652029499411583,
+ "learning_rate": 0.0006,
+ "loss": 5.628726959228516,
+ "step": 486
+ },
+ {
+ "epoch": 6.7685589519650655,
+ "grad_norm": 0.02118665538728237,
+ "learning_rate": 0.0006,
+ "loss": 5.6580705642700195,
+ "step": 487
+ },
+ {
+ "epoch": 6.782532751091703,
+ "grad_norm": 0.02237832546234131,
+ "learning_rate": 0.0006,
+ "loss": 5.635931968688965,
+ "step": 488
+ },
+ {
+ "epoch": 6.796506550218341,
+ "grad_norm": 0.023897631093859673,
+ "learning_rate": 0.0006,
+ "loss": 5.517949104309082,
+ "step": 489
+ },
+ {
+ "epoch": 6.810480349344978,
+ "grad_norm": 0.02442227490246296,
+ "learning_rate": 0.0006,
+ "loss": 5.600021839141846,
+ "step": 490
+ },
+ {
+ "epoch": 6.824454148471616,
+ "grad_norm": 0.024675287306308746,
+ "learning_rate": 0.0006,
+ "loss": 5.6639909744262695,
+ "step": 491
+ },
+ {
+ "epoch": 6.8384279475982535,
+ "grad_norm": 0.030372392386198044,
+ "learning_rate": 0.0006,
+ "loss": 5.582888603210449,
+ "step": 492
+ },
+ {
+ "epoch": 6.85240174672489,
+ "grad_norm": 0.03237690031528473,
+ "learning_rate": 0.0006,
+ "loss": 5.573896884918213,
+ "step": 493
+ },
+ {
+ "epoch": 6.866375545851528,
+ "grad_norm": 0.03511589393019676,
+ "learning_rate": 0.0006,
+ "loss": 5.565878868103027,
+ "step": 494
+ },
+ {
+ "epoch": 6.880349344978166,
+ "grad_norm": 0.03883938118815422,
+ "learning_rate": 0.0006,
+ "loss": 5.58540678024292,
+ "step": 495
+ },
+ {
+ "epoch": 6.894323144104804,
+ "grad_norm": 0.04175502806901932,
+ "learning_rate": 0.0006,
+ "loss": 5.6095194816589355,
+ "step": 496
+ },
+ {
+ "epoch": 6.908296943231441,
+ "grad_norm": 0.04013441503047943,
+ "learning_rate": 0.0006,
+ "loss": 5.654401779174805,
+ "step": 497
+ },
+ {
+ "epoch": 6.922270742358078,
+ "grad_norm": 0.04618770629167557,
+ "learning_rate": 0.0006,
+ "loss": 5.6483473777771,
+ "step": 498
+ },
+ {
+ "epoch": 6.936244541484716,
+ "grad_norm": 0.04507308453321457,
+ "learning_rate": 0.0006,
+ "loss": 5.5856218338012695,
+ "step": 499
+ },
+ {
+ "epoch": 6.950218340611354,
+ "grad_norm": 0.03151383996009827,
+ "learning_rate": 0.0006,
+ "loss": 5.594086647033691,
+ "step": 500
+ },
+ {
+ "epoch": 6.964192139737992,
+ "grad_norm": 0.026883797720074654,
+ "learning_rate": 0.0006,
+ "loss": 5.607676029205322,
+ "step": 501
+ },
+ {
+ "epoch": 6.978165938864628,
+ "grad_norm": 0.02981836162507534,
+ "learning_rate": 0.0006,
+ "loss": 5.563666820526123,
+ "step": 502
+ },
+ {
+ "epoch": 6.992139737991266,
+ "grad_norm": 0.03233477845788002,
+ "learning_rate": 0.0006,
+ "loss": 5.667543411254883,
+ "step": 503
+ },
+ {
+ "epoch": 7.0,
+ "grad_norm": 0.029927456751465797,
+ "learning_rate": 0.0006,
+ "loss": 5.733482837677002,
+ "step": 504
+ },
+ {
+ "epoch": 7.0,
+ "eval_loss": 5.679076671600342,
+ "eval_runtime": 59.5965,
+ "eval_samples_per_second": 40.976,
+ "eval_steps_per_second": 1.292,
+ "step": 504
+ },
+ {
+ "epoch": 7.013973799126638,
+ "grad_norm": 0.027081597596406937,
+ "learning_rate": 0.0006,
+ "loss": 5.634487152099609,
+ "step": 505
+ },
+ {
+ "epoch": 7.0279475982532755,
+ "grad_norm": 0.031959421932697296,
+ "learning_rate": 0.0006,
+ "loss": 5.643294811248779,
+ "step": 506
+ },
+ {
+ "epoch": 7.041921397379912,
+ "grad_norm": 0.02802063524723053,
+ "learning_rate": 0.0006,
+ "loss": 5.573239326477051,
+ "step": 507
+ },
+ {
+ "epoch": 7.05589519650655,
+ "grad_norm": 0.032579705119132996,
+ "learning_rate": 0.0006,
+ "loss": 5.627150058746338,
+ "step": 508
+ },
+ {
+ "epoch": 7.069868995633188,
+ "grad_norm": 0.02503928542137146,
+ "learning_rate": 0.0006,
+ "loss": 5.6045308113098145,
+ "step": 509
+ },
+ {
+ "epoch": 7.083842794759826,
+ "grad_norm": 0.02421317622065544,
+ "learning_rate": 0.0006,
+ "loss": 5.522153377532959,
+ "step": 510
+ },
+ {
+ "epoch": 7.097816593886463,
+ "grad_norm": 0.021983426064252853,
+ "learning_rate": 0.0006,
+ "loss": 5.632939338684082,
+ "step": 511
+ },
+ {
+ "epoch": 7.1117903930131,
+ "grad_norm": 0.021933183073997498,
+ "learning_rate": 0.0006,
+ "loss": 5.592185974121094,
+ "step": 512
+ },
+ {
+ "epoch": 7.125764192139738,
+ "grad_norm": 0.02287031151354313,
+ "learning_rate": 0.0006,
+ "loss": 5.613700866699219,
+ "step": 513
+ },
+ {
+ "epoch": 7.139737991266376,
+ "grad_norm": 0.02281602844595909,
+ "learning_rate": 0.0006,
+ "loss": 5.551383018493652,
+ "step": 514
+ },
+ {
+ "epoch": 7.153711790393013,
+ "grad_norm": 0.021050360053777695,
+ "learning_rate": 0.0006,
+ "loss": 5.55275821685791,
+ "step": 515
+ },
+ {
+ "epoch": 7.1676855895196505,
+ "grad_norm": 0.018299926072359085,
+ "learning_rate": 0.0006,
+ "loss": 5.528225421905518,
+ "step": 516
+ },
+ {
+ "epoch": 7.181659388646288,
+ "grad_norm": 0.02024853602051735,
+ "learning_rate": 0.0006,
+ "loss": 5.522453784942627,
+ "step": 517
+ },
+ {
+ "epoch": 7.195633187772926,
+ "grad_norm": 0.02135239914059639,
+ "learning_rate": 0.0006,
+ "loss": 5.527522563934326,
+ "step": 518
+ },
+ {
+ "epoch": 7.209606986899563,
+ "grad_norm": 0.023394184187054634,
+ "learning_rate": 0.0006,
+ "loss": 5.4535017013549805,
+ "step": 519
+ },
+ {
+ "epoch": 7.223580786026201,
+ "grad_norm": 0.025952080264687538,
+ "learning_rate": 0.0006,
+ "loss": 5.5374674797058105,
+ "step": 520
+ },
+ {
+ "epoch": 7.2375545851528384,
+ "grad_norm": 0.022287718951702118,
+ "learning_rate": 0.0006,
+ "loss": 5.493753433227539,
+ "step": 521
+ },
+ {
+ "epoch": 7.251528384279476,
+ "grad_norm": 0.02018447406589985,
+ "learning_rate": 0.0006,
+ "loss": 5.510575294494629,
+ "step": 522
+ },
+ {
+ "epoch": 7.265502183406113,
+ "grad_norm": 0.02050507813692093,
+ "learning_rate": 0.0006,
+ "loss": 5.466026782989502,
+ "step": 523
+ },
+ {
+ "epoch": 7.279475982532751,
+ "grad_norm": 0.023688802495598793,
+ "learning_rate": 0.0006,
+ "loss": 5.485815525054932,
+ "step": 524
+ },
+ {
+ "epoch": 7.293449781659389,
+ "grad_norm": 0.021078843623399734,
+ "learning_rate": 0.0006,
+ "loss": 5.415581703186035,
+ "step": 525
+ },
+ {
+ "epoch": 7.307423580786026,
+ "grad_norm": 0.018266011029481888,
+ "learning_rate": 0.0006,
+ "loss": 5.402815818786621,
+ "step": 526
+ },
+ {
+ "epoch": 7.321397379912664,
+ "grad_norm": 0.019693493843078613,
+ "learning_rate": 0.0006,
+ "loss": 5.504674911499023,
+ "step": 527
+ },
+ {
+ "epoch": 7.335371179039301,
+ "grad_norm": 0.02031373605132103,
+ "learning_rate": 0.0006,
+ "loss": 5.503837585449219,
+ "step": 528
+ },
+ {
+ "epoch": 7.349344978165939,
+ "grad_norm": 0.017636626958847046,
+ "learning_rate": 0.0006,
+ "loss": 5.530580997467041,
+ "step": 529
+ },
+ {
+ "epoch": 7.3633187772925766,
+ "grad_norm": 0.01787244901061058,
+ "learning_rate": 0.0006,
+ "loss": 5.450218677520752,
+ "step": 530
+ },
+ {
+ "epoch": 7.377292576419214,
+ "grad_norm": 0.0170669574290514,
+ "learning_rate": 0.0006,
+ "loss": 5.481570243835449,
+ "step": 531
+ },
+ {
+ "epoch": 7.391266375545851,
+ "grad_norm": 0.01802165061235428,
+ "learning_rate": 0.0006,
+ "loss": 5.386394500732422,
+ "step": 532
+ },
+ {
+ "epoch": 7.405240174672489,
+ "grad_norm": 0.022949472069740295,
+ "learning_rate": 0.0006,
+ "loss": 5.424929618835449,
+ "step": 533
+ },
+ {
+ "epoch": 7.419213973799127,
+ "grad_norm": 0.034213390201330185,
+ "learning_rate": 0.0006,
+ "loss": 5.346663475036621,
+ "step": 534
+ },
+ {
+ "epoch": 7.4331877729257645,
+ "grad_norm": 0.03847593814134598,
+ "learning_rate": 0.0006,
+ "loss": 5.418482780456543,
+ "step": 535
+ },
+ {
+ "epoch": 7.447161572052401,
+ "grad_norm": 0.029393676668405533,
+ "learning_rate": 0.0006,
+ "loss": 5.456090927124023,
+ "step": 536
+ },
+ {
+ "epoch": 7.461135371179039,
+ "grad_norm": 0.03388667106628418,
+ "learning_rate": 0.0006,
+ "loss": 5.572983741760254,
+ "step": 537
+ },
+ {
+ "epoch": 7.475109170305677,
+ "grad_norm": 0.039690613746643066,
+ "learning_rate": 0.0006,
+ "loss": 5.414067268371582,
+ "step": 538
+ },
+ {
+ "epoch": 7.489082969432315,
+ "grad_norm": 0.03630776330828667,
+ "learning_rate": 0.0006,
+ "loss": 5.523739814758301,
+ "step": 539
+ },
+ {
+ "epoch": 7.503056768558952,
+ "grad_norm": 0.03356870263814926,
+ "learning_rate": 0.0006,
+ "loss": 5.444767951965332,
+ "step": 540
+ },
+ {
+ "epoch": 7.517030567685589,
+ "grad_norm": 0.030812304466962814,
+ "learning_rate": 0.0006,
+ "loss": 5.461244106292725,
+ "step": 541
+ },
+ {
+ "epoch": 7.531004366812227,
+ "grad_norm": 0.03319217637181282,
+ "learning_rate": 0.0006,
+ "loss": 5.502161026000977,
+ "step": 542
+ },
+ {
+ "epoch": 7.544978165938865,
+ "grad_norm": 0.032997481524944305,
+ "learning_rate": 0.0006,
+ "loss": 5.4162278175354,
+ "step": 543
+ },
+ {
+ "epoch": 7.558951965065502,
+ "grad_norm": 0.03364962339401245,
+ "learning_rate": 0.0006,
+ "loss": 5.415736198425293,
+ "step": 544
+ },
+ {
+ "epoch": 7.5729257641921395,
+ "grad_norm": 0.0344221405684948,
+ "learning_rate": 0.0006,
+ "loss": 5.541967868804932,
+ "step": 545
+ },
+ {
+ "epoch": 7.586899563318777,
+ "grad_norm": 0.029609503224492073,
+ "learning_rate": 0.0006,
+ "loss": 5.440614223480225,
+ "step": 546
+ },
+ {
+ "epoch": 7.600873362445415,
+ "grad_norm": 0.029055926948785782,
+ "learning_rate": 0.0006,
+ "loss": 5.462865352630615,
+ "step": 547
+ },
+ {
+ "epoch": 7.614847161572053,
+ "grad_norm": 0.02658848837018013,
+ "learning_rate": 0.0006,
+ "loss": 5.444467544555664,
+ "step": 548
+ },
+ {
+ "epoch": 7.62882096069869,
+ "grad_norm": 0.026276404038071632,
+ "learning_rate": 0.0006,
+ "loss": 5.462278842926025,
+ "step": 549
+ },
+ {
+ "epoch": 7.642794759825327,
+ "grad_norm": 0.0286889486014843,
+ "learning_rate": 0.0006,
+ "loss": 5.4849534034729,
+ "step": 550
+ },
+ {
+ "epoch": 7.656768558951965,
+ "grad_norm": 0.026849817484617233,
+ "learning_rate": 0.0006,
+ "loss": 5.431473731994629,
+ "step": 551
+ },
+ {
+ "epoch": 7.670742358078603,
+ "grad_norm": 0.02312396466732025,
+ "learning_rate": 0.0006,
+ "loss": 5.45462703704834,
+ "step": 552
+ },
+ {
+ "epoch": 7.68471615720524,
+ "grad_norm": 0.026197485625743866,
+ "learning_rate": 0.0006,
+ "loss": 5.452552318572998,
+ "step": 553
+ },
+ {
+ "epoch": 7.698689956331878,
+ "grad_norm": 0.02747255191206932,
+ "learning_rate": 0.0006,
+ "loss": 5.447169780731201,
+ "step": 554
+ },
+ {
+ "epoch": 7.712663755458515,
+ "grad_norm": 0.028123432770371437,
+ "learning_rate": 0.0006,
+ "loss": 5.394400596618652,
+ "step": 555
+ },
+ {
+ "epoch": 7.726637554585153,
+ "grad_norm": 0.02599870041012764,
+ "learning_rate": 0.0006,
+ "loss": 5.398341178894043,
+ "step": 556
+ },
+ {
+ "epoch": 7.74061135371179,
+ "grad_norm": 0.022171657532453537,
+ "learning_rate": 0.0006,
+ "loss": 5.368820667266846,
+ "step": 557
+ },
+ {
+ "epoch": 7.754585152838428,
+ "grad_norm": 0.022309480234980583,
+ "learning_rate": 0.0006,
+ "loss": 5.416772365570068,
+ "step": 558
+ },
+ {
+ "epoch": 7.7685589519650655,
+ "grad_norm": 0.024074165150523186,
+ "learning_rate": 0.0006,
+ "loss": 5.359002113342285,
+ "step": 559
+ },
+ {
+ "epoch": 7.782532751091703,
+ "grad_norm": 0.02636653557419777,
+ "learning_rate": 0.0006,
+ "loss": 5.484368324279785,
+ "step": 560
+ },
+ {
+ "epoch": 7.796506550218341,
+ "grad_norm": 0.02203752100467682,
+ "learning_rate": 0.0006,
+ "loss": 5.375498294830322,
+ "step": 561
+ },
+ {
+ "epoch": 7.810480349344978,
+ "grad_norm": 0.019924819469451904,
+ "learning_rate": 0.0006,
+ "loss": 5.334723472595215,
+ "step": 562
+ },
+ {
+ "epoch": 7.824454148471616,
+ "grad_norm": 0.018755966797471046,
+ "learning_rate": 0.0006,
+ "loss": 5.348781108856201,
+ "step": 563
+ },
+ {
+ "epoch": 7.8384279475982535,
+ "grad_norm": 0.016747845336794853,
+ "learning_rate": 0.0006,
+ "loss": 5.433079719543457,
+ "step": 564
+ },
+ {
+ "epoch": 7.85240174672489,
+ "grad_norm": 0.0170395877212286,
+ "learning_rate": 0.0006,
+ "loss": 5.346794128417969,
+ "step": 565
+ },
+ {
+ "epoch": 7.866375545851528,
+ "grad_norm": 0.016074176877737045,
+ "learning_rate": 0.0006,
+ "loss": 5.3277740478515625,
+ "step": 566
+ },
+ {
+ "epoch": 7.880349344978166,
+ "grad_norm": 0.014119806699454784,
+ "learning_rate": 0.0006,
+ "loss": 5.483551979064941,
+ "step": 567
+ },
+ {
+ "epoch": 7.894323144104804,
+ "grad_norm": 0.015270394273102283,
+ "learning_rate": 0.0006,
+ "loss": 5.378519535064697,
+ "step": 568
+ },
+ {
+ "epoch": 7.908296943231441,
+ "grad_norm": 0.015677539631724358,
+ "learning_rate": 0.0006,
+ "loss": 5.290210247039795,
+ "step": 569
+ },
+ {
+ "epoch": 7.922270742358078,
+ "grad_norm": 0.015930423513054848,
+ "learning_rate": 0.0006,
+ "loss": 5.427360534667969,
+ "step": 570
+ },
+ {
+ "epoch": 7.936244541484716,
+ "grad_norm": 0.016641564667224884,
+ "learning_rate": 0.0006,
+ "loss": 5.301599979400635,
+ "step": 571
+ },
+ {
+ "epoch": 7.950218340611354,
+ "grad_norm": 0.018293552100658417,
+ "learning_rate": 0.0006,
+ "loss": 5.277889251708984,
+ "step": 572
+ },
+ {
+ "epoch": 7.964192139737992,
+ "grad_norm": 0.017618799582123756,
+ "learning_rate": 0.0006,
+ "loss": 5.297194480895996,
+ "step": 573
+ },
+ {
+ "epoch": 7.978165938864628,
+ "grad_norm": 0.015034242533147335,
+ "learning_rate": 0.0006,
+ "loss": 5.42428731918335,
+ "step": 574
+ },
+ {
+ "epoch": 7.992139737991266,
+ "grad_norm": 0.01617511734366417,
+ "learning_rate": 0.0006,
+ "loss": 5.420044898986816,
+ "step": 575
+ },
+ {
+ "epoch": 8.0,
+ "grad_norm": 0.01837257295846939,
+ "learning_rate": 0.0006,
+ "loss": 5.240387916564941,
+ "step": 576
+ },
+ {
+ "epoch": 8.0,
+ "eval_loss": 5.423932075500488,
+ "eval_runtime": 58.0841,
+ "eval_samples_per_second": 42.042,
+ "eval_steps_per_second": 1.326,
+ "step": 576
+ },
+ {
+ "epoch": 8.013973799126637,
+ "grad_norm": 0.020953809842467308,
+ "learning_rate": 0.0006,
+ "loss": 5.4320783615112305,
+ "step": 577
+ },
+ {
+ "epoch": 8.027947598253276,
+ "grad_norm": 0.0227745920419693,
+ "learning_rate": 0.0006,
+ "loss": 5.310197830200195,
+ "step": 578
+ },
+ {
+ "epoch": 8.041921397379912,
+ "grad_norm": 0.021372603252530098,
+ "learning_rate": 0.0006,
+ "loss": 5.339468002319336,
+ "step": 579
+ },
+ {
+ "epoch": 8.055895196506551,
+ "grad_norm": 0.022401731461286545,
+ "learning_rate": 0.0006,
+ "loss": 5.279829978942871,
+ "step": 580
+ },
+ {
+ "epoch": 8.069868995633188,
+ "grad_norm": 0.023942379280924797,
+ "learning_rate": 0.0006,
+ "loss": 5.372323989868164,
+ "step": 581
+ },
+ {
+ "epoch": 8.083842794759825,
+ "grad_norm": 0.024380534887313843,
+ "learning_rate": 0.0006,
+ "loss": 5.253486633300781,
+ "step": 582
+ },
+ {
+ "epoch": 8.097816593886463,
+ "grad_norm": 0.03246625140309334,
+ "learning_rate": 0.0006,
+ "loss": 5.305774211883545,
+ "step": 583
+ },
+ {
+ "epoch": 8.1117903930131,
+ "grad_norm": 0.030399736016988754,
+ "learning_rate": 0.0006,
+ "loss": 5.370976448059082,
+ "step": 584
+ },
+ {
+ "epoch": 8.125764192139737,
+ "grad_norm": 0.027148913592100143,
+ "learning_rate": 0.0006,
+ "loss": 5.368010520935059,
+ "step": 585
+ },
+ {
+ "epoch": 8.139737991266376,
+ "grad_norm": 0.02925540879368782,
+ "learning_rate": 0.0006,
+ "loss": 5.303013801574707,
+ "step": 586
+ },
+ {
+ "epoch": 8.153711790393013,
+ "grad_norm": 0.02915453538298607,
+ "learning_rate": 0.0006,
+ "loss": 5.329667091369629,
+ "step": 587
+ },
+ {
+ "epoch": 8.167685589519651,
+ "grad_norm": 0.031732626259326935,
+ "learning_rate": 0.0006,
+ "loss": 5.3634934425354,
+ "step": 588
+ },
+ {
+ "epoch": 8.181659388646288,
+ "grad_norm": 0.036166731268167496,
+ "learning_rate": 0.0006,
+ "loss": 5.223980903625488,
+ "step": 589
+ },
+ {
+ "epoch": 8.195633187772925,
+ "grad_norm": 0.035654108971357346,
+ "learning_rate": 0.0006,
+ "loss": 5.361059188842773,
+ "step": 590
+ },
+ {
+ "epoch": 8.209606986899564,
+ "grad_norm": 0.0365324430167675,
+ "learning_rate": 0.0006,
+ "loss": 5.336151599884033,
+ "step": 591
+ },
+ {
+ "epoch": 8.2235807860262,
+ "grad_norm": 0.03578880429267883,
+ "learning_rate": 0.0006,
+ "loss": 5.434076309204102,
+ "step": 592
+ },
+ {
+ "epoch": 8.237554585152838,
+ "grad_norm": 0.035734012722969055,
+ "learning_rate": 0.0006,
+ "loss": 5.330893516540527,
+ "step": 593
+ },
+ {
+ "epoch": 8.251528384279476,
+ "grad_norm": 0.03196857124567032,
+ "learning_rate": 0.0006,
+ "loss": 5.306009769439697,
+ "step": 594
+ },
+ {
+ "epoch": 8.265502183406113,
+ "grad_norm": 0.032042086124420166,
+ "learning_rate": 0.0006,
+ "loss": 5.343267917633057,
+ "step": 595
+ },
+ {
+ "epoch": 8.279475982532752,
+ "grad_norm": 0.03160746395587921,
+ "learning_rate": 0.0006,
+ "loss": 5.2353715896606445,
+ "step": 596
+ },
+ {
+ "epoch": 8.293449781659389,
+ "grad_norm": 0.026689818128943443,
+ "learning_rate": 0.0006,
+ "loss": 5.254042625427246,
+ "step": 597
+ },
+ {
+ "epoch": 8.307423580786025,
+ "grad_norm": 0.02880188450217247,
+ "learning_rate": 0.0006,
+ "loss": 5.326833248138428,
+ "step": 598
+ },
+ {
+ "epoch": 8.321397379912664,
+ "grad_norm": 0.027516381815075874,
+ "learning_rate": 0.0006,
+ "loss": 5.342336177825928,
+ "step": 599
+ },
+ {
+ "epoch": 8.335371179039301,
+ "grad_norm": 0.027875030413269997,
+ "learning_rate": 0.0006,
+ "loss": 5.200719833374023,
+ "step": 600
+ },
+ {
+ "epoch": 8.34934497816594,
+ "grad_norm": 0.0268265288323164,
+ "learning_rate": 0.0006,
+ "loss": 5.249449729919434,
+ "step": 601
+ },
+ {
+ "epoch": 8.363318777292577,
+ "grad_norm": 0.024824608117341995,
+ "learning_rate": 0.0006,
+ "loss": 5.318952560424805,
+ "step": 602
+ },
+ {
+ "epoch": 8.377292576419213,
+ "grad_norm": 0.022990627214312553,
+ "learning_rate": 0.0006,
+ "loss": 5.244993209838867,
+ "step": 603
+ },
+ {
+ "epoch": 8.391266375545852,
+ "grad_norm": 0.022804604843258858,
+ "learning_rate": 0.0006,
+ "loss": 5.3236846923828125,
+ "step": 604
+ },
+ {
+ "epoch": 8.405240174672489,
+ "grad_norm": 0.02419872209429741,
+ "learning_rate": 0.0006,
+ "loss": 5.255486488342285,
+ "step": 605
+ },
+ {
+ "epoch": 8.419213973799126,
+ "grad_norm": 0.021952059119939804,
+ "learning_rate": 0.0006,
+ "loss": 5.3511552810668945,
+ "step": 606
+ },
+ {
+ "epoch": 8.433187772925764,
+ "grad_norm": 0.022375497967004776,
+ "learning_rate": 0.0006,
+ "loss": 5.294790267944336,
+ "step": 607
+ },
+ {
+ "epoch": 8.447161572052401,
+ "grad_norm": 0.019242815673351288,
+ "learning_rate": 0.0006,
+ "loss": 5.252618789672852,
+ "step": 608
+ },
+ {
+ "epoch": 8.46113537117904,
+ "grad_norm": 0.0173486340790987,
+ "learning_rate": 0.0006,
+ "loss": 5.30147647857666,
+ "step": 609
+ },
+ {
+ "epoch": 8.475109170305677,
+ "grad_norm": 0.01612604409456253,
+ "learning_rate": 0.0006,
+ "loss": 5.23960542678833,
+ "step": 610
+ },
+ {
+ "epoch": 8.489082969432314,
+ "grad_norm": 0.014293976128101349,
+ "learning_rate": 0.0006,
+ "loss": 5.119932651519775,
+ "step": 611
+ },
+ {
+ "epoch": 8.503056768558952,
+ "grad_norm": 0.014453536830842495,
+ "learning_rate": 0.0006,
+ "loss": 5.166794776916504,
+ "step": 612
+ },
+ {
+ "epoch": 8.51703056768559,
+ "grad_norm": 0.014685769565403461,
+ "learning_rate": 0.0006,
+ "loss": 5.264822959899902,
+ "step": 613
+ },
+ {
+ "epoch": 8.531004366812226,
+ "grad_norm": 0.01427740603685379,
+ "learning_rate": 0.0006,
+ "loss": 5.278433799743652,
+ "step": 614
+ },
+ {
+ "epoch": 8.544978165938865,
+ "grad_norm": 0.014864951372146606,
+ "learning_rate": 0.0006,
+ "loss": 5.214512348175049,
+ "step": 615
+ },
+ {
+ "epoch": 8.558951965065502,
+ "grad_norm": 0.01565164513885975,
+ "learning_rate": 0.0006,
+ "loss": 5.183066368103027,
+ "step": 616
+ },
+ {
+ "epoch": 8.57292576419214,
+ "grad_norm": 0.01779816672205925,
+ "learning_rate": 0.0006,
+ "loss": 5.163185119628906,
+ "step": 617
+ },
+ {
+ "epoch": 8.586899563318777,
+ "grad_norm": 0.017254827544093132,
+ "learning_rate": 0.0006,
+ "loss": 5.243555068969727,
+ "step": 618
+ },
+ {
+ "epoch": 8.600873362445414,
+ "grad_norm": 0.01650584116578102,
+ "learning_rate": 0.0006,
+ "loss": 5.234884738922119,
+ "step": 619
+ },
+ {
+ "epoch": 8.614847161572053,
+ "grad_norm": 0.017021115869283676,
+ "learning_rate": 0.0006,
+ "loss": 5.12271785736084,
+ "step": 620
+ },
+ {
+ "epoch": 8.62882096069869,
+ "grad_norm": 0.01773759163916111,
+ "learning_rate": 0.0006,
+ "loss": 5.242153167724609,
+ "step": 621
+ },
+ {
+ "epoch": 8.642794759825328,
+ "grad_norm": 0.015679042786359787,
+ "learning_rate": 0.0006,
+ "loss": 5.170779228210449,
+ "step": 622
+ },
+ {
+ "epoch": 8.656768558951965,
+ "grad_norm": 0.013760549947619438,
+ "learning_rate": 0.0006,
+ "loss": 5.238644599914551,
+ "step": 623
+ },
+ {
+ "epoch": 8.670742358078602,
+ "grad_norm": 0.014571522362530231,
+ "learning_rate": 0.0006,
+ "loss": 5.177056312561035,
+ "step": 624
+ },
+ {
+ "epoch": 8.68471615720524,
+ "grad_norm": 0.016209015622735023,
+ "learning_rate": 0.0006,
+ "loss": 5.163750648498535,
+ "step": 625
+ },
+ {
+ "epoch": 8.698689956331878,
+ "grad_norm": 0.016813941299915314,
+ "learning_rate": 0.0006,
+ "loss": 5.183428764343262,
+ "step": 626
+ },
+ {
+ "epoch": 8.712663755458514,
+ "grad_norm": 0.020985357463359833,
+ "learning_rate": 0.0006,
+ "loss": 5.229465007781982,
+ "step": 627
+ },
+ {
+ "epoch": 8.726637554585153,
+ "grad_norm": 0.02679632417857647,
+ "learning_rate": 0.0006,
+ "loss": 5.257368087768555,
+ "step": 628
+ },
+ {
+ "epoch": 8.74061135371179,
+ "grad_norm": 0.025756070390343666,
+ "learning_rate": 0.0006,
+ "loss": 5.253736972808838,
+ "step": 629
+ },
+ {
+ "epoch": 8.754585152838429,
+ "grad_norm": 0.02643490768969059,
+ "learning_rate": 0.0006,
+ "loss": 5.228633403778076,
+ "step": 630
+ },
+ {
+ "epoch": 8.768558951965066,
+ "grad_norm": 0.03303210437297821,
+ "learning_rate": 0.0006,
+ "loss": 5.163540363311768,
+ "step": 631
+ },
+ {
+ "epoch": 8.782532751091702,
+ "grad_norm": 0.030432431027293205,
+ "learning_rate": 0.0006,
+ "loss": 5.243169784545898,
+ "step": 632
+ },
+ {
+ "epoch": 8.796506550218341,
+ "grad_norm": 0.028631288558244705,
+ "learning_rate": 0.0006,
+ "loss": 5.166398048400879,
+ "step": 633
+ },
+ {
+ "epoch": 8.810480349344978,
+ "grad_norm": 0.026188310235738754,
+ "learning_rate": 0.0006,
+ "loss": 5.128122329711914,
+ "step": 634
+ },
+ {
+ "epoch": 8.824454148471617,
+ "grad_norm": 0.025308528915047646,
+ "learning_rate": 0.0006,
+ "loss": 5.211284637451172,
+ "step": 635
+ },
+ {
+ "epoch": 8.838427947598253,
+ "grad_norm": 0.020657729357481003,
+ "learning_rate": 0.0006,
+ "loss": 5.161575794219971,
+ "step": 636
+ },
+ {
+ "epoch": 8.85240174672489,
+ "grad_norm": 0.021901234984397888,
+ "learning_rate": 0.0006,
+ "loss": 5.20050048828125,
+ "step": 637
+ },
+ {
+ "epoch": 8.866375545851529,
+ "grad_norm": 0.021219318732619286,
+ "learning_rate": 0.0006,
+ "loss": 5.211699485778809,
+ "step": 638
+ },
+ {
+ "epoch": 8.880349344978166,
+ "grad_norm": 0.0200974028557539,
+ "learning_rate": 0.0006,
+ "loss": 5.092503547668457,
+ "step": 639
+ },
+ {
+ "epoch": 8.894323144104803,
+ "grad_norm": 0.023804882541298866,
+ "learning_rate": 0.0006,
+ "loss": 5.216068267822266,
+ "step": 640
+ },
+ {
+ "epoch": 8.908296943231441,
+ "grad_norm": 0.026088010519742966,
+ "learning_rate": 0.0006,
+ "loss": 5.155592918395996,
+ "step": 641
+ },
+ {
+ "epoch": 8.922270742358078,
+ "grad_norm": 0.02601276896893978,
+ "learning_rate": 0.0006,
+ "loss": 5.197238922119141,
+ "step": 642
+ },
+ {
+ "epoch": 8.936244541484717,
+ "grad_norm": 0.020387211814522743,
+ "learning_rate": 0.0006,
+ "loss": 5.1955437660217285,
+ "step": 643
+ },
+ {
+ "epoch": 8.950218340611354,
+ "grad_norm": 0.019214622676372528,
+ "learning_rate": 0.0006,
+ "loss": 5.206346035003662,
+ "step": 644
+ },
+ {
+ "epoch": 8.96419213973799,
+ "grad_norm": 0.019674314185976982,
+ "learning_rate": 0.0006,
+ "loss": 5.219857215881348,
+ "step": 645
+ },
+ {
+ "epoch": 8.97816593886463,
+ "grad_norm": 0.020318234339356422,
+ "learning_rate": 0.0006,
+ "loss": 5.191132545471191,
+ "step": 646
+ },
+ {
+ "epoch": 8.992139737991266,
+ "grad_norm": 0.021428676322102547,
+ "learning_rate": 0.0006,
+ "loss": 5.214695930480957,
+ "step": 647
+ },
+ {
+ "epoch": 9.0,
+ "grad_norm": 0.023235054686665535,
+ "learning_rate": 0.0006,
+ "loss": 5.229091644287109,
+ "step": 648
+ },
+ {
+ "epoch": 9.0,
+ "eval_loss": 5.289593696594238,
+ "eval_runtime": 56.9256,
+ "eval_samples_per_second": 42.898,
+ "eval_steps_per_second": 1.353,
+ "step": 648
+ },
+ {
+ "epoch": 9.013973799126637,
+ "grad_norm": 0.02320289984345436,
+ "learning_rate": 0.0006,
+ "loss": 4.983669281005859,
+ "step": 649
+ },
+ {
+ "epoch": 9.027947598253276,
+ "grad_norm": 0.018521282821893692,
+ "learning_rate": 0.0006,
+ "loss": 5.200575828552246,
+ "step": 650
+ },
+ {
+ "epoch": 9.041921397379912,
+ "grad_norm": 0.0180169939994812,
+ "learning_rate": 0.0006,
+ "loss": 5.148033618927002,
+ "step": 651
+ },
+ {
+ "epoch": 9.055895196506551,
+ "grad_norm": 0.019969860091805458,
+ "learning_rate": 0.0006,
+ "loss": 5.099125385284424,
+ "step": 652
+ },
+ {
+ "epoch": 9.069868995633188,
+ "grad_norm": 0.01643305830657482,
+ "learning_rate": 0.0006,
+ "loss": 5.092347145080566,
+ "step": 653
+ },
+ {
+ "epoch": 9.083842794759825,
+ "grad_norm": 0.016022363677620888,
+ "learning_rate": 0.0006,
+ "loss": 5.129279613494873,
+ "step": 654
+ },
+ {
+ "epoch": 9.097816593886463,
+ "grad_norm": 0.01657041162252426,
+ "learning_rate": 0.0006,
+ "loss": 5.159562587738037,
+ "step": 655
+ },
+ {
+ "epoch": 9.1117903930131,
+ "grad_norm": 0.017924228683114052,
+ "learning_rate": 0.0006,
+ "loss": 5.165497779846191,
+ "step": 656
+ },
+ {
+ "epoch": 9.125764192139737,
+ "grad_norm": 0.021592361852526665,
+ "learning_rate": 0.0006,
+ "loss": 5.204775810241699,
+ "step": 657
+ },
+ {
+ "epoch": 9.139737991266376,
+ "grad_norm": 0.026924120262265205,
+ "learning_rate": 0.0006,
+ "loss": 5.128296852111816,
+ "step": 658
+ },
+ {
+ "epoch": 9.153711790393013,
+ "grad_norm": 0.023156899958848953,
+ "learning_rate": 0.0006,
+ "loss": 5.1255340576171875,
+ "step": 659
+ },
+ {
+ "epoch": 9.167685589519651,
+ "grad_norm": 0.020670367404818535,
+ "learning_rate": 0.0006,
+ "loss": 5.128479957580566,
+ "step": 660
+ },
+ {
+ "epoch": 9.181659388646288,
+ "grad_norm": 0.025447173044085503,
+ "learning_rate": 0.0006,
+ "loss": 5.154821872711182,
+ "step": 661
+ },
+ {
+ "epoch": 9.195633187772925,
+ "grad_norm": 0.027940068393945694,
+ "learning_rate": 0.0006,
+ "loss": 5.143466472625732,
+ "step": 662
+ },
+ {
+ "epoch": 9.209606986899564,
+ "grad_norm": 0.026847844943404198,
+ "learning_rate": 0.0006,
+ "loss": 5.045665740966797,
+ "step": 663
+ },
+ {
+ "epoch": 9.2235807860262,
+ "grad_norm": 0.02340601570904255,
+ "learning_rate": 0.0006,
+ "loss": 5.21933126449585,
+ "step": 664
+ },
+ {
+ "epoch": 9.237554585152838,
+ "grad_norm": 0.02340371161699295,
+ "learning_rate": 0.0006,
+ "loss": 4.942949295043945,
+ "step": 665
+ },
+ {
+ "epoch": 9.251528384279476,
+ "grad_norm": 0.02221992425620556,
+ "learning_rate": 0.0006,
+ "loss": 5.197099208831787,
+ "step": 666
+ },
+ {
+ "epoch": 9.265502183406113,
+ "grad_norm": 0.023508677259087563,
+ "learning_rate": 0.0006,
+ "loss": 5.022897720336914,
+ "step": 667
+ },
+ {
+ "epoch": 9.279475982532752,
+ "grad_norm": 0.026120394468307495,
+ "learning_rate": 0.0006,
+ "loss": 5.087725639343262,
+ "step": 668
+ },
+ {
+ "epoch": 9.293449781659389,
+ "grad_norm": 0.027273228392004967,
+ "learning_rate": 0.0006,
+ "loss": 5.090963840484619,
+ "step": 669
+ },
+ {
+ "epoch": 9.307423580786025,
+ "grad_norm": 0.03241586685180664,
+ "learning_rate": 0.0006,
+ "loss": 5.047842979431152,
+ "step": 670
+ },
+ {
+ "epoch": 9.321397379912664,
+ "grad_norm": 0.030245667323470116,
+ "learning_rate": 0.0006,
+ "loss": 5.104803085327148,
+ "step": 671
+ },
+ {
+ "epoch": 9.335371179039301,
+ "grad_norm": 0.027698364108800888,
+ "learning_rate": 0.0006,
+ "loss": 5.0644097328186035,
+ "step": 672
+ },
+ {
+ "epoch": 9.34934497816594,
+ "grad_norm": 0.029692554846405983,
+ "learning_rate": 0.0006,
+ "loss": 5.111942291259766,
+ "step": 673
+ },
+ {
+ "epoch": 9.363318777292577,
+ "grad_norm": 0.03328656405210495,
+ "learning_rate": 0.0006,
+ "loss": 5.1574554443359375,
+ "step": 674
+ },
+ {
+ "epoch": 9.377292576419213,
+ "grad_norm": 0.031938180327415466,
+ "learning_rate": 0.0006,
+ "loss": 5.190371990203857,
+ "step": 675
+ },
+ {
+ "epoch": 9.391266375545852,
+ "grad_norm": 0.024858945980668068,
+ "learning_rate": 0.0006,
+ "loss": 5.121420860290527,
+ "step": 676
+ },
+ {
+ "epoch": 9.405240174672489,
+ "grad_norm": 0.023033203557133675,
+ "learning_rate": 0.0006,
+ "loss": 5.0974345207214355,
+ "step": 677
+ },
+ {
+ "epoch": 9.419213973799126,
+ "grad_norm": 0.022083545103669167,
+ "learning_rate": 0.0006,
+ "loss": 5.046319961547852,
+ "step": 678
+ },
+ {
+ "epoch": 9.433187772925764,
+ "grad_norm": 0.0228437427431345,
+ "learning_rate": 0.0006,
+ "loss": 5.119963645935059,
+ "step": 679
+ },
+ {
+ "epoch": 9.447161572052401,
+ "grad_norm": 0.0242351982742548,
+ "learning_rate": 0.0006,
+ "loss": 5.0444560050964355,
+ "step": 680
+ },
+ {
+ "epoch": 9.46113537117904,
+ "grad_norm": 0.02401185780763626,
+ "learning_rate": 0.0006,
+ "loss": 5.063238620758057,
+ "step": 681
+ },
+ {
+ "epoch": 9.475109170305677,
+ "grad_norm": 0.02560959942638874,
+ "learning_rate": 0.0006,
+ "loss": 5.035224914550781,
+ "step": 682
+ },
+ {
+ "epoch": 9.489082969432314,
+ "grad_norm": 0.024985626339912415,
+ "learning_rate": 0.0006,
+ "loss": 5.049284934997559,
+ "step": 683
+ },
+ {
+ "epoch": 9.503056768558952,
+ "grad_norm": 0.023178115487098694,
+ "learning_rate": 0.0006,
+ "loss": 5.2266621589660645,
+ "step": 684
+ },
+ {
+ "epoch": 9.51703056768559,
+ "grad_norm": 0.0215512253344059,
+ "learning_rate": 0.0006,
+ "loss": 5.1875715255737305,
+ "step": 685
+ },
+ {
+ "epoch": 9.531004366812226,
+ "grad_norm": 0.01662837713956833,
+ "learning_rate": 0.0006,
+ "loss": 5.137012481689453,
+ "step": 686
+ },
+ {
+ "epoch": 9.544978165938865,
+ "grad_norm": 0.01952764391899109,
+ "learning_rate": 0.0006,
+ "loss": 5.072500228881836,
+ "step": 687
+ },
+ {
+ "epoch": 9.558951965065502,
+ "grad_norm": 0.017521077767014503,
+ "learning_rate": 0.0006,
+ "loss": 5.045711994171143,
+ "step": 688
+ },
+ {
+ "epoch": 9.57292576419214,
+ "grad_norm": 0.017355265095829964,
+ "learning_rate": 0.0006,
+ "loss": 5.02881383895874,
+ "step": 689
+ },
+ {
+ "epoch": 9.586899563318777,
+ "grad_norm": 0.015587719157338142,
+ "learning_rate": 0.0006,
+ "loss": 5.132787704467773,
+ "step": 690
+ },
+ {
+ "epoch": 9.600873362445414,
+ "grad_norm": 0.017469845712184906,
+ "learning_rate": 0.0006,
+ "loss": 5.105447292327881,
+ "step": 691
+ },
+ {
+ "epoch": 9.614847161572053,
+ "grad_norm": 0.019630100578069687,
+ "learning_rate": 0.0006,
+ "loss": 5.032464981079102,
+ "step": 692
+ },
+ {
+ "epoch": 9.62882096069869,
+ "grad_norm": 0.01784994639456272,
+ "learning_rate": 0.0006,
+ "loss": 5.081345558166504,
+ "step": 693
+ },
+ {
+ "epoch": 9.642794759825328,
+ "grad_norm": 0.01872754842042923,
+ "learning_rate": 0.0006,
+ "loss": 4.9772114753723145,
+ "step": 694
+ },
+ {
+ "epoch": 9.656768558951965,
+ "grad_norm": 0.01999032311141491,
+ "learning_rate": 0.0006,
+ "loss": 5.089740753173828,
+ "step": 695
+ },
+ {
+ "epoch": 9.670742358078602,
+ "grad_norm": 0.01710767112672329,
+ "learning_rate": 0.0006,
+ "loss": 4.9869890213012695,
+ "step": 696
+ },
+ {
+ "epoch": 9.68471615720524,
+ "grad_norm": 0.015144161880016327,
+ "learning_rate": 0.0006,
+ "loss": 5.022564888000488,
+ "step": 697
+ },
+ {
+ "epoch": 9.698689956331878,
+ "grad_norm": 0.01618543453514576,
+ "learning_rate": 0.0006,
+ "loss": 5.035921573638916,
+ "step": 698
+ },
+ {
+ "epoch": 9.712663755458514,
+ "grad_norm": 0.014653387479484081,
+ "learning_rate": 0.0006,
+ "loss": 4.914456367492676,
+ "step": 699
+ },
+ {
+ "epoch": 9.726637554585153,
+ "grad_norm": 0.013434977270662785,
+ "learning_rate": 0.0006,
+ "loss": 5.063764572143555,
+ "step": 700
+ },
+ {
+ "epoch": 9.74061135371179,
+ "grad_norm": 0.015128864906728268,
+ "learning_rate": 0.0006,
+ "loss": 5.052367687225342,
+ "step": 701
+ },
+ {
+ "epoch": 9.754585152838429,
+ "grad_norm": 0.014210768043994904,
+ "learning_rate": 0.0006,
+ "loss": 5.016489028930664,
+ "step": 702
+ },
+ {
+ "epoch": 9.768558951965066,
+ "grad_norm": 0.015191680751740932,
+ "learning_rate": 0.0006,
+ "loss": 4.990176200866699,
+ "step": 703
+ },
+ {
+ "epoch": 9.782532751091702,
+ "grad_norm": 0.018285127356648445,
+ "learning_rate": 0.0006,
+ "loss": 5.015164375305176,
+ "step": 704
+ },
+ {
+ "epoch": 9.796506550218341,
+ "grad_norm": 0.0204079058021307,
+ "learning_rate": 0.0006,
+ "loss": 5.091142654418945,
+ "step": 705
+ },
+ {
+ "epoch": 9.810480349344978,
+ "grad_norm": 0.01980586163699627,
+ "learning_rate": 0.0006,
+ "loss": 5.03350830078125,
+ "step": 706
+ },
+ {
+ "epoch": 9.824454148471617,
+ "grad_norm": 0.02048729732632637,
+ "learning_rate": 0.0006,
+ "loss": 5.03611421585083,
+ "step": 707
+ },
+ {
+ "epoch": 9.838427947598253,
+ "grad_norm": 0.02058163844048977,
+ "learning_rate": 0.0006,
+ "loss": 4.9613447189331055,
+ "step": 708
+ },
+ {
+ "epoch": 9.85240174672489,
+ "grad_norm": 0.018004924058914185,
+ "learning_rate": 0.0006,
+ "loss": 4.966541767120361,
+ "step": 709
+ },
+ {
+ "epoch": 9.866375545851529,
+ "grad_norm": 0.0198152307420969,
+ "learning_rate": 0.0006,
+ "loss": 5.01750373840332,
+ "step": 710
+ },
+ {
+ "epoch": 9.880349344978166,
+ "grad_norm": 0.018532969057559967,
+ "learning_rate": 0.0006,
+ "loss": 5.097424507141113,
+ "step": 711
+ },
+ {
+ "epoch": 9.894323144104803,
+ "grad_norm": 0.018939971923828125,
+ "learning_rate": 0.0006,
+ "loss": 5.177223205566406,
+ "step": 712
+ },
+ {
+ "epoch": 9.908296943231441,
+ "grad_norm": 0.017969170585274696,
+ "learning_rate": 0.0006,
+ "loss": 5.118013858795166,
+ "step": 713
+ },
+ {
+ "epoch": 9.922270742358078,
+ "grad_norm": 0.018021270632743835,
+ "learning_rate": 0.0006,
+ "loss": 5.034787178039551,
+ "step": 714
+ },
+ {
+ "epoch": 9.936244541484717,
+ "grad_norm": 0.017009710893034935,
+ "learning_rate": 0.0006,
+ "loss": 5.003920555114746,
+ "step": 715
+ },
+ {
+ "epoch": 9.950218340611354,
+ "grad_norm": 0.017722049728035927,
+ "learning_rate": 0.0006,
+ "loss": 5.031939506530762,
+ "step": 716
+ },
+ {
+ "epoch": 9.96419213973799,
+ "grad_norm": 0.017134087160229683,
+ "learning_rate": 0.0006,
+ "loss": 5.093474388122559,
+ "step": 717
+ },
+ {
+ "epoch": 9.97816593886463,
+ "grad_norm": 0.018511613830924034,
+ "learning_rate": 0.0006,
+ "loss": 5.078604698181152,
+ "step": 718
+ },
+ {
+ "epoch": 9.992139737991266,
+ "grad_norm": 0.021206334233283997,
+ "learning_rate": 0.0006,
+ "loss": 5.052572727203369,
+ "step": 719
+ },
+ {
+ "epoch": 10.0,
+ "grad_norm": 0.02239062264561653,
+ "learning_rate": 0.0006,
+ "loss": 5.063547134399414,
+ "step": 720
+ },
+ {
+ "epoch": 10.0,
+ "eval_loss": 5.173044681549072,
+ "eval_runtime": 56.7287,
+ "eval_samples_per_second": 43.047,
+ "eval_steps_per_second": 1.357,
+ "step": 720
+ },
+ {
+ "epoch": 10.013973799126637,
+ "grad_norm": 0.020757969468832016,
+ "learning_rate": 0.0006,
+ "loss": 5.045648574829102,
+ "step": 721
+ },
+ {
+ "epoch": 10.027947598253276,
+ "grad_norm": 0.022697702050209045,
+ "learning_rate": 0.0006,
+ "loss": 4.993833065032959,
+ "step": 722
+ },
+ {
+ "epoch": 10.041921397379912,
+ "grad_norm": 0.02434331737458706,
+ "learning_rate": 0.0006,
+ "loss": 5.073219299316406,
+ "step": 723
+ },
+ {
+ "epoch": 10.055895196506551,
+ "grad_norm": 0.025499660521745682,
+ "learning_rate": 0.0006,
+ "loss": 5.019072532653809,
+ "step": 724
+ },
+ {
+ "epoch": 10.069868995633188,
+ "grad_norm": 0.029281053692102432,
+ "learning_rate": 0.0006,
+ "loss": 4.959851264953613,
+ "step": 725
+ },
+ {
+ "epoch": 10.083842794759825,
+ "grad_norm": 0.033878110349178314,
+ "learning_rate": 0.0006,
+ "loss": 5.132717609405518,
+ "step": 726
+ },
+ {
+ "epoch": 10.097816593886463,
+ "grad_norm": 0.03385764732956886,
+ "learning_rate": 0.0006,
+ "loss": 5.0482892990112305,
+ "step": 727
+ },
+ {
+ "epoch": 10.1117903930131,
+ "grad_norm": 0.03466648980975151,
+ "learning_rate": 0.0006,
+ "loss": 4.946855545043945,
+ "step": 728
+ },
+ {
+ "epoch": 10.125764192139737,
+ "grad_norm": 0.028546305373311043,
+ "learning_rate": 0.0006,
+ "loss": 5.027774333953857,
+ "step": 729
+ },
+ {
+ "epoch": 10.139737991266376,
+ "grad_norm": 0.030575869604945183,
+ "learning_rate": 0.0006,
+ "loss": 4.9768476486206055,
+ "step": 730
+ },
+ {
+ "epoch": 10.153711790393013,
+ "grad_norm": 0.03614303469657898,
+ "learning_rate": 0.0006,
+ "loss": 5.078113555908203,
+ "step": 731
+ },
+ {
+ "epoch": 10.167685589519651,
+ "grad_norm": 0.040991660207509995,
+ "learning_rate": 0.0006,
+ "loss": 5.033185958862305,
+ "step": 732
+ },
+ {
+ "epoch": 10.181659388646288,
+ "grad_norm": 0.04551699757575989,
+ "learning_rate": 0.0006,
+ "loss": 5.0820746421813965,
+ "step": 733
+ },
+ {
+ "epoch": 10.195633187772925,
+ "grad_norm": 0.040193989872932434,
+ "learning_rate": 0.0006,
+ "loss": 5.059700965881348,
+ "step": 734
+ },
+ {
+ "epoch": 10.209606986899564,
+ "grad_norm": 0.035851605236530304,
+ "learning_rate": 0.0006,
+ "loss": 5.142467498779297,
+ "step": 735
+ },
+ {
+ "epoch": 10.2235807860262,
+ "grad_norm": 0.034200169146060944,
+ "learning_rate": 0.0006,
+ "loss": 5.072312355041504,
+ "step": 736
+ },
+ {
+ "epoch": 10.237554585152838,
+ "grad_norm": 0.033621110022068024,
+ "learning_rate": 0.0006,
+ "loss": 5.08037805557251,
+ "step": 737
+ },
+ {
+ "epoch": 10.251528384279476,
+ "grad_norm": 0.037372369319200516,
+ "learning_rate": 0.0006,
+ "loss": 4.998425483703613,
+ "step": 738
+ },
+ {
+ "epoch": 10.265502183406113,
+ "grad_norm": 0.037044707685709,
+ "learning_rate": 0.0006,
+ "loss": 4.923993110656738,
+ "step": 739
+ },
+ {
+ "epoch": 10.279475982532752,
+ "grad_norm": 0.030454808846116066,
+ "learning_rate": 0.0006,
+ "loss": 5.0889997482299805,
+ "step": 740
+ },
+ {
+ "epoch": 10.293449781659389,
+ "grad_norm": 0.030969638377428055,
+ "learning_rate": 0.0006,
+ "loss": 4.984112739562988,
+ "step": 741
+ },
+ {
+ "epoch": 10.307423580786025,
+ "grad_norm": 0.0276983380317688,
+ "learning_rate": 0.0006,
+ "loss": 4.955412864685059,
+ "step": 742
+ },
+ {
+ "epoch": 10.321397379912664,
+ "grad_norm": 0.02716052532196045,
+ "learning_rate": 0.0006,
+ "loss": 5.0098876953125,
+ "step": 743
+ },
+ {
+ "epoch": 10.335371179039301,
+ "grad_norm": 0.02403552085161209,
+ "learning_rate": 0.0006,
+ "loss": 5.165160655975342,
+ "step": 744
+ },
+ {
+ "epoch": 10.34934497816594,
+ "grad_norm": 0.0257862601429224,
+ "learning_rate": 0.0006,
+ "loss": 5.001547813415527,
+ "step": 745
+ },
+ {
+ "epoch": 10.363318777292577,
+ "grad_norm": 0.02300378680229187,
+ "learning_rate": 0.0006,
+ "loss": 5.038888931274414,
+ "step": 746
+ },
+ {
+ "epoch": 10.377292576419213,
+ "grad_norm": 0.019766854122281075,
+ "learning_rate": 0.0006,
+ "loss": 4.924291610717773,
+ "step": 747
+ },
+ {
+ "epoch": 10.391266375545852,
+ "grad_norm": 0.01888395845890045,
+ "learning_rate": 0.0006,
+ "loss": 5.059971332550049,
+ "step": 748
+ },
+ {
+ "epoch": 10.405240174672489,
+ "grad_norm": 0.015577499754726887,
+ "learning_rate": 0.0006,
+ "loss": 4.987686634063721,
+ "step": 749
+ },
+ {
+ "epoch": 10.419213973799126,
+ "grad_norm": 0.016012346372008324,
+ "learning_rate": 0.0006,
+ "loss": 5.015157699584961,
+ "step": 750
+ },
+ {
+ "epoch": 10.433187772925764,
+ "grad_norm": 0.016277998685836792,
+ "learning_rate": 0.0006,
+ "loss": 5.004931449890137,
+ "step": 751
+ },
+ {
+ "epoch": 10.447161572052401,
+ "grad_norm": 0.017504561692476273,
+ "learning_rate": 0.0006,
+ "loss": 4.902827739715576,
+ "step": 752
+ },
+ {
+ "epoch": 10.46113537117904,
+ "grad_norm": 0.01569022797048092,
+ "learning_rate": 0.0006,
+ "loss": 4.954863548278809,
+ "step": 753
+ },
+ {
+ "epoch": 10.475109170305677,
+ "grad_norm": 0.014114422723650932,
+ "learning_rate": 0.0006,
+ "loss": 5.00656795501709,
+ "step": 754
+ },
+ {
+ "epoch": 10.489082969432314,
+ "grad_norm": 0.01573832333087921,
+ "learning_rate": 0.0006,
+ "loss": 4.978353023529053,
+ "step": 755
+ },
+ {
+ "epoch": 10.503056768558952,
+ "grad_norm": 0.016480036079883575,
+ "learning_rate": 0.0006,
+ "loss": 4.953327655792236,
+ "step": 756
+ },
+ {
+ "epoch": 10.51703056768559,
+ "grad_norm": 0.014008025638759136,
+ "learning_rate": 0.0006,
+ "loss": 5.050840377807617,
+ "step": 757
+ },
+ {
+ "epoch": 10.531004366812226,
+ "grad_norm": 0.013147998601198196,
+ "learning_rate": 0.0006,
+ "loss": 4.974964141845703,
+ "step": 758
+ },
+ {
+ "epoch": 10.544978165938865,
+ "grad_norm": 0.013807603158056736,
+ "learning_rate": 0.0006,
+ "loss": 4.927907943725586,
+ "step": 759
+ },
+ {
+ "epoch": 10.558951965065502,
+ "grad_norm": 0.013555224984884262,
+ "learning_rate": 0.0006,
+ "loss": 4.980690956115723,
+ "step": 760
+ },
+ {
+ "epoch": 10.57292576419214,
+ "grad_norm": 0.013045408762991428,
+ "learning_rate": 0.0006,
+ "loss": 4.9436116218566895,
+ "step": 761
+ },
+ {
+ "epoch": 10.586899563318777,
+ "grad_norm": 0.012858862057328224,
+ "learning_rate": 0.0006,
+ "loss": 4.9595770835876465,
+ "step": 762
+ },
+ {
+ "epoch": 10.600873362445414,
+ "grad_norm": 0.01468253880739212,
+ "learning_rate": 0.0006,
+ "loss": 4.90725040435791,
+ "step": 763
+ },
+ {
+ "epoch": 10.614847161572053,
+ "grad_norm": 0.013224679045379162,
+ "learning_rate": 0.0006,
+ "loss": 4.877615928649902,
+ "step": 764
+ },
+ {
+ "epoch": 10.62882096069869,
+ "grad_norm": 0.013555348850786686,
+ "learning_rate": 0.0006,
+ "loss": 4.9456610679626465,
+ "step": 765
+ },
+ {
+ "epoch": 10.642794759825328,
+ "grad_norm": 0.014387160539627075,
+ "learning_rate": 0.0006,
+ "loss": 4.9233717918396,
+ "step": 766
+ },
+ {
+ "epoch": 10.656768558951965,
+ "grad_norm": 0.014354088343679905,
+ "learning_rate": 0.0006,
+ "loss": 4.983511924743652,
+ "step": 767
+ },
+ {
+ "epoch": 10.670742358078602,
+ "grad_norm": 0.01332700252532959,
+ "learning_rate": 0.0006,
+ "loss": 4.934416770935059,
+ "step": 768
+ },
+ {
+ "epoch": 10.68471615720524,
+ "grad_norm": 0.011543313041329384,
+ "learning_rate": 0.0006,
+ "loss": 4.96388053894043,
+ "step": 769
+ },
+ {
+ "epoch": 10.698689956331878,
+ "grad_norm": 0.012330878525972366,
+ "learning_rate": 0.0006,
+ "loss": 4.946396350860596,
+ "step": 770
+ },
+ {
+ "epoch": 10.712663755458514,
+ "grad_norm": 0.01379779726266861,
+ "learning_rate": 0.0006,
+ "loss": 4.981019020080566,
+ "step": 771
+ },
+ {
+ "epoch": 10.726637554585153,
+ "grad_norm": 0.014612431637942791,
+ "learning_rate": 0.0006,
+ "loss": 4.894567966461182,
+ "step": 772
+ },
+ {
+ "epoch": 10.74061135371179,
+ "grad_norm": 0.013062899932265282,
+ "learning_rate": 0.0006,
+ "loss": 5.033038139343262,
+ "step": 773
+ },
+ {
+ "epoch": 10.754585152838429,
+ "grad_norm": 0.013898049481213093,
+ "learning_rate": 0.0006,
+ "loss": 4.823636054992676,
+ "step": 774
+ },
+ {
+ "epoch": 10.768558951965066,
+ "grad_norm": 0.01502019353210926,
+ "learning_rate": 0.0006,
+ "loss": 5.020156383514404,
+ "step": 775
+ },
+ {
+ "epoch": 10.782532751091702,
+ "grad_norm": 0.013810204342007637,
+ "learning_rate": 0.0006,
+ "loss": 4.935647010803223,
+ "step": 776
+ },
+ {
+ "epoch": 10.796506550218341,
+ "grad_norm": 0.013616513460874557,
+ "learning_rate": 0.0006,
+ "loss": 4.901486396789551,
+ "step": 777
+ },
+ {
+ "epoch": 10.810480349344978,
+ "grad_norm": 0.013601024635136127,
+ "learning_rate": 0.0006,
+ "loss": 4.931873321533203,
+ "step": 778
+ },
+ {
+ "epoch": 10.824454148471617,
+ "grad_norm": 0.013668350875377655,
+ "learning_rate": 0.0006,
+ "loss": 4.937256813049316,
+ "step": 779
+ },
+ {
+ "epoch": 10.838427947598253,
+ "grad_norm": 0.01581823080778122,
+ "learning_rate": 0.0006,
+ "loss": 4.9611406326293945,
+ "step": 780
+ },
+ {
+ "epoch": 10.85240174672489,
+ "grad_norm": 0.019478484988212585,
+ "learning_rate": 0.0006,
+ "loss": 4.9212236404418945,
+ "step": 781
+ },
+ {
+ "epoch": 10.866375545851529,
+ "grad_norm": 0.026921333745121956,
+ "learning_rate": 0.0006,
+ "loss": 4.88501501083374,
+ "step": 782
+ },
+ {
+ "epoch": 10.880349344978166,
+ "grad_norm": 0.03155914694070816,
+ "learning_rate": 0.0006,
+ "loss": 4.989967346191406,
+ "step": 783
+ },
+ {
+ "epoch": 10.894323144104803,
+ "grad_norm": 0.029689429327845573,
+ "learning_rate": 0.0006,
+ "loss": 4.936000823974609,
+ "step": 784
+ },
+ {
+ "epoch": 10.908296943231441,
+ "grad_norm": 0.026626063510775566,
+ "learning_rate": 0.0006,
+ "loss": 4.984785079956055,
+ "step": 785
+ },
+ {
+ "epoch": 10.922270742358078,
+ "grad_norm": 0.026168212294578552,
+ "learning_rate": 0.0006,
+ "loss": 4.908282279968262,
+ "step": 786
+ },
+ {
+ "epoch": 10.936244541484717,
+ "grad_norm": 0.02828158251941204,
+ "learning_rate": 0.0006,
+ "loss": 4.926390647888184,
+ "step": 787
+ },
+ {
+ "epoch": 10.950218340611354,
+ "grad_norm": 0.02649078331887722,
+ "learning_rate": 0.0006,
+ "loss": 5.00400972366333,
+ "step": 788
+ },
+ {
+ "epoch": 10.96419213973799,
+ "grad_norm": 0.0256856270134449,
+ "learning_rate": 0.0006,
+ "loss": 5.0317487716674805,
+ "step": 789
+ },
+ {
+ "epoch": 10.97816593886463,
+ "grad_norm": 0.024728883057832718,
+ "learning_rate": 0.0006,
+ "loss": 4.9299635887146,
+ "step": 790
+ },
+ {
+ "epoch": 10.992139737991266,
+ "grad_norm": 0.026241116225719452,
+ "learning_rate": 0.0006,
+ "loss": 4.947319984436035,
+ "step": 791
+ },
+ {
+ "epoch": 11.0,
+ "grad_norm": 0.023526743054389954,
+ "learning_rate": 0.0006,
+ "loss": 4.94429349899292,
+ "step": 792
+ },
+ {
+ "epoch": 11.0,
+ "eval_loss": 5.104345798492432,
+ "eval_runtime": 57.0955,
+ "eval_samples_per_second": 42.77,
+ "eval_steps_per_second": 1.349,
+ "step": 792
+ },
+ {
+ "epoch": 11.013973799126637,
+ "grad_norm": 0.0216788612306118,
+ "learning_rate": 0.0006,
+ "loss": 4.881950855255127,
+ "step": 793
+ },
+ {
+ "epoch": 11.027947598253276,
+ "grad_norm": 0.01830463856458664,
+ "learning_rate": 0.0006,
+ "loss": 4.971016883850098,
+ "step": 794
+ },
+ {
+ "epoch": 11.041921397379912,
+ "grad_norm": 0.02204521745443344,
+ "learning_rate": 0.0006,
+ "loss": 4.894623756408691,
+ "step": 795
+ },
+ {
+ "epoch": 11.055895196506551,
+ "grad_norm": 0.0210255216807127,
+ "learning_rate": 0.0006,
+ "loss": 4.834009170532227,
+ "step": 796
+ },
+ {
+ "epoch": 11.069868995633188,
+ "grad_norm": 0.019797448068857193,
+ "learning_rate": 0.0006,
+ "loss": 5.086714744567871,
+ "step": 797
+ },
+ {
+ "epoch": 11.083842794759825,
+ "grad_norm": 0.02036641724407673,
+ "learning_rate": 0.0006,
+ "loss": 4.920056343078613,
+ "step": 798
+ },
+ {
+ "epoch": 11.097816593886463,
+ "grad_norm": 0.01966066285967827,
+ "learning_rate": 0.0006,
+ "loss": 4.879859447479248,
+ "step": 799
+ },
+ {
+ "epoch": 11.1117903930131,
+ "grad_norm": 0.0202149897813797,
+ "learning_rate": 0.0006,
+ "loss": 4.933640480041504,
+ "step": 800
+ },
+ {
+ "epoch": 11.125764192139737,
+ "grad_norm": 0.01893039606511593,
+ "learning_rate": 0.0006,
+ "loss": 4.901970863342285,
+ "step": 801
+ },
+ {
+ "epoch": 11.139737991266376,
+ "grad_norm": 0.016040155664086342,
+ "learning_rate": 0.0006,
+ "loss": 4.905491828918457,
+ "step": 802
+ },
+ {
+ "epoch": 11.153711790393013,
+ "grad_norm": 0.01720350608229637,
+ "learning_rate": 0.0006,
+ "loss": 4.798816680908203,
+ "step": 803
+ },
+ {
+ "epoch": 11.167685589519651,
+ "grad_norm": 0.017002852633595467,
+ "learning_rate": 0.0006,
+ "loss": 4.851859092712402,
+ "step": 804
+ },
+ {
+ "epoch": 11.181659388646288,
+ "grad_norm": 0.015594680793583393,
+ "learning_rate": 0.0006,
+ "loss": 4.90328311920166,
+ "step": 805
+ },
+ {
+ "epoch": 11.195633187772925,
+ "grad_norm": 0.014904462732374668,
+ "learning_rate": 0.0006,
+ "loss": 4.904215335845947,
+ "step": 806
+ },
+ {
+ "epoch": 11.209606986899564,
+ "grad_norm": 0.01366228237748146,
+ "learning_rate": 0.0006,
+ "loss": 4.9813055992126465,
+ "step": 807
+ },
+ {
+ "epoch": 11.2235807860262,
+ "grad_norm": 0.014719638973474503,
+ "learning_rate": 0.0006,
+ "loss": 4.927306175231934,
+ "step": 808
+ },
+ {
+ "epoch": 11.237554585152838,
+ "grad_norm": 0.015016036108136177,
+ "learning_rate": 0.0006,
+ "loss": 4.874215126037598,
+ "step": 809
+ },
+ {
+ "epoch": 11.251528384279476,
+ "grad_norm": 0.014925424009561539,
+ "learning_rate": 0.0006,
+ "loss": 4.838257789611816,
+ "step": 810
+ },
+ {
+ "epoch": 11.265502183406113,
+ "grad_norm": 0.014335823245346546,
+ "learning_rate": 0.0006,
+ "loss": 4.940608024597168,
+ "step": 811
+ },
+ {
+ "epoch": 11.279475982532752,
+ "grad_norm": 0.014987512491643429,
+ "learning_rate": 0.0006,
+ "loss": 4.831531524658203,
+ "step": 812
+ },
+ {
+ "epoch": 11.293449781659389,
+ "grad_norm": 0.016087457537651062,
+ "learning_rate": 0.0006,
+ "loss": 4.863227367401123,
+ "step": 813
+ },
+ {
+ "epoch": 11.307423580786025,
+ "grad_norm": 0.01671300269663334,
+ "learning_rate": 0.0006,
+ "loss": 4.8535661697387695,
+ "step": 814
+ },
+ {
+ "epoch": 11.321397379912664,
+ "grad_norm": 0.015229142270982265,
+ "learning_rate": 0.0006,
+ "loss": 4.948635101318359,
+ "step": 815
+ },
+ {
+ "epoch": 11.335371179039301,
+ "grad_norm": 0.01532587967813015,
+ "learning_rate": 0.0006,
+ "loss": 4.9035797119140625,
+ "step": 816
+ },
+ {
+ "epoch": 11.34934497816594,
+ "grad_norm": 0.015715476125478745,
+ "learning_rate": 0.0006,
+ "loss": 4.894709587097168,
+ "step": 817
+ },
+ {
+ "epoch": 11.363318777292577,
+ "grad_norm": 0.014156977646052837,
+ "learning_rate": 0.0006,
+ "loss": 4.960206985473633,
+ "step": 818
+ },
+ {
+ "epoch": 11.377292576419213,
+ "grad_norm": 0.017076566815376282,
+ "learning_rate": 0.0006,
+ "loss": 5.005850791931152,
+ "step": 819
+ },
+ {
+ "epoch": 11.391266375545852,
+ "grad_norm": 0.0172084029763937,
+ "learning_rate": 0.0006,
+ "loss": 4.903355598449707,
+ "step": 820
+ },
+ {
+ "epoch": 11.405240174672489,
+ "grad_norm": 0.0180258359760046,
+ "learning_rate": 0.0006,
+ "loss": 4.842667579650879,
+ "step": 821
+ },
+ {
+ "epoch": 11.419213973799126,
+ "grad_norm": 0.022130120545625687,
+ "learning_rate": 0.0006,
+ "loss": 4.799970626831055,
+ "step": 822
+ },
+ {
+ "epoch": 11.433187772925764,
+ "grad_norm": 0.02384897693991661,
+ "learning_rate": 0.0006,
+ "loss": 4.860746383666992,
+ "step": 823
+ },
+ {
+ "epoch": 11.447161572052401,
+ "grad_norm": 0.021881572902202606,
+ "learning_rate": 0.0006,
+ "loss": 4.918107032775879,
+ "step": 824
+ },
+ {
+ "epoch": 11.46113537117904,
+ "grad_norm": 0.023085080087184906,
+ "learning_rate": 0.0006,
+ "loss": 4.878946304321289,
+ "step": 825
+ },
+ {
+ "epoch": 11.475109170305677,
+ "grad_norm": 0.02291012555360794,
+ "learning_rate": 0.0006,
+ "loss": 4.958887100219727,
+ "step": 826
+ },
+ {
+ "epoch": 11.489082969432314,
+ "grad_norm": 0.0199968870729208,
+ "learning_rate": 0.0006,
+ "loss": 4.80955696105957,
+ "step": 827
+ },
+ {
+ "epoch": 11.503056768558952,
+ "grad_norm": 0.018702229484915733,
+ "learning_rate": 0.0006,
+ "loss": 5.0276408195495605,
+ "step": 828
+ },
+ {
+ "epoch": 11.51703056768559,
+ "grad_norm": 0.017787311226129532,
+ "learning_rate": 0.0006,
+ "loss": 4.96018123626709,
+ "step": 829
+ },
+ {
+ "epoch": 11.531004366812226,
+ "grad_norm": 0.01704501546919346,
+ "learning_rate": 0.0006,
+ "loss": 4.914553165435791,
+ "step": 830
+ },
+ {
+ "epoch": 11.544978165938865,
+ "grad_norm": 0.017519677057862282,
+ "learning_rate": 0.0006,
+ "loss": 4.798098087310791,
+ "step": 831
+ },
+ {
+ "epoch": 11.558951965065502,
+ "grad_norm": 0.018511991947889328,
+ "learning_rate": 0.0006,
+ "loss": 4.817785263061523,
+ "step": 832
+ },
+ {
+ "epoch": 11.57292576419214,
+ "grad_norm": 0.01876644790172577,
+ "learning_rate": 0.0006,
+ "loss": 4.818995475769043,
+ "step": 833
+ },
+ {
+ "epoch": 11.586899563318777,
+ "grad_norm": 0.01753568835556507,
+ "learning_rate": 0.0006,
+ "loss": 4.811938285827637,
+ "step": 834
+ },
+ {
+ "epoch": 11.600873362445414,
+ "grad_norm": 0.017046531662344933,
+ "learning_rate": 0.0006,
+ "loss": 4.914674282073975,
+ "step": 835
+ },
+ {
+ "epoch": 11.614847161572053,
+ "grad_norm": 0.01802094094455242,
+ "learning_rate": 0.0006,
+ "loss": 4.993284225463867,
+ "step": 836
+ },
+ {
+ "epoch": 11.62882096069869,
+ "grad_norm": 0.019493764266371727,
+ "learning_rate": 0.0006,
+ "loss": 4.895628452301025,
+ "step": 837
+ },
+ {
+ "epoch": 11.642794759825328,
+ "grad_norm": 0.019461210817098618,
+ "learning_rate": 0.0006,
+ "loss": 4.869269371032715,
+ "step": 838
+ },
+ {
+ "epoch": 11.656768558951965,
+ "grad_norm": 0.01981567218899727,
+ "learning_rate": 0.0006,
+ "loss": 4.863052845001221,
+ "step": 839
+ },
+ {
+ "epoch": 11.670742358078602,
+ "grad_norm": 0.019894491881132126,
+ "learning_rate": 0.0006,
+ "loss": 4.8775787353515625,
+ "step": 840
+ },
+ {
+ "epoch": 11.68471615720524,
+ "grad_norm": 0.01875786855816841,
+ "learning_rate": 0.0006,
+ "loss": 4.817763328552246,
+ "step": 841
+ },
+ {
+ "epoch": 11.698689956331878,
+ "grad_norm": 0.020926695317029953,
+ "learning_rate": 0.0006,
+ "loss": 4.7470526695251465,
+ "step": 842
+ },
+ {
+ "epoch": 11.712663755458514,
+ "grad_norm": 0.021477103233337402,
+ "learning_rate": 0.0006,
+ "loss": 4.762682914733887,
+ "step": 843
+ },
+ {
+ "epoch": 11.726637554585153,
+ "grad_norm": 0.020628711208701134,
+ "learning_rate": 0.0006,
+ "loss": 4.807985782623291,
+ "step": 844
+ },
+ {
+ "epoch": 11.74061135371179,
+ "grad_norm": 0.021489612758159637,
+ "learning_rate": 0.0006,
+ "loss": 4.843091011047363,
+ "step": 845
+ },
+ {
+ "epoch": 11.754585152838429,
+ "grad_norm": 0.021194491535425186,
+ "learning_rate": 0.0006,
+ "loss": 4.8300251960754395,
+ "step": 846
+ },
+ {
+ "epoch": 11.768558951965066,
+ "grad_norm": 0.01833350397646427,
+ "learning_rate": 0.0006,
+ "loss": 5.000202178955078,
+ "step": 847
+ },
+ {
+ "epoch": 11.782532751091702,
+ "grad_norm": 0.018096931278705597,
+ "learning_rate": 0.0006,
+ "loss": 4.880648136138916,
+ "step": 848
+ },
+ {
+ "epoch": 11.796506550218341,
+ "grad_norm": 0.0197161752730608,
+ "learning_rate": 0.0006,
+ "loss": 4.788316249847412,
+ "step": 849
+ },
+ {
+ "epoch": 11.810480349344978,
+ "grad_norm": 0.0203181654214859,
+ "learning_rate": 0.0006,
+ "loss": 4.870944976806641,
+ "step": 850
+ },
+ {
+ "epoch": 11.824454148471617,
+ "grad_norm": 0.01908455230295658,
+ "learning_rate": 0.0006,
+ "loss": 4.890144348144531,
+ "step": 851
+ },
+ {
+ "epoch": 11.838427947598253,
+ "grad_norm": 0.021281801164150238,
+ "learning_rate": 0.0006,
+ "loss": 4.89578914642334,
+ "step": 852
+ },
+ {
+ "epoch": 11.85240174672489,
+ "grad_norm": 0.02124273031949997,
+ "learning_rate": 0.0006,
+ "loss": 4.883749008178711,
+ "step": 853
+ },
+ {
+ "epoch": 11.866375545851529,
+ "grad_norm": 0.020969398319721222,
+ "learning_rate": 0.0006,
+ "loss": 4.809142589569092,
+ "step": 854
+ },
+ {
+ "epoch": 11.880349344978166,
+ "grad_norm": 0.01967657171189785,
+ "learning_rate": 0.0006,
+ "loss": 4.821864604949951,
+ "step": 855
+ },
+ {
+ "epoch": 11.894323144104803,
+ "grad_norm": 0.020887333899736404,
+ "learning_rate": 0.0006,
+ "loss": 4.917985916137695,
+ "step": 856
+ },
+ {
+ "epoch": 11.908296943231441,
+ "grad_norm": 0.019203083589673042,
+ "learning_rate": 0.0006,
+ "loss": 4.824960708618164,
+ "step": 857
+ },
+ {
+ "epoch": 11.922270742358078,
+ "grad_norm": 0.020162401720881462,
+ "learning_rate": 0.0006,
+ "loss": 4.772617816925049,
+ "step": 858
+ },
+ {
+ "epoch": 11.936244541484717,
+ "grad_norm": 0.02251487225294113,
+ "learning_rate": 0.0006,
+ "loss": 4.735418319702148,
+ "step": 859
+ },
+ {
+ "epoch": 11.950218340611354,
+ "grad_norm": 0.020313527435064316,
+ "learning_rate": 0.0006,
+ "loss": 4.939233779907227,
+ "step": 860
+ },
+ {
+ "epoch": 11.96419213973799,
+ "grad_norm": 0.021635599434375763,
+ "learning_rate": 0.0006,
+ "loss": 4.861807346343994,
+ "step": 861
+ },
+ {
+ "epoch": 11.97816593886463,
+ "grad_norm": 0.02285311557352543,
+ "learning_rate": 0.0006,
+ "loss": 4.932732582092285,
+ "step": 862
+ },
+ {
+ "epoch": 11.992139737991266,
+ "grad_norm": 0.02072535641491413,
+ "learning_rate": 0.0006,
+ "loss": 4.796767711639404,
+ "step": 863
+ },
+ {
+ "epoch": 12.0,
+ "grad_norm": 0.020588304847478867,
+ "learning_rate": 0.0006,
+ "loss": 4.997537612915039,
+ "step": 864
+ },
+ {
+ "epoch": 12.0,
+ "eval_loss": 5.044651985168457,
+ "eval_runtime": 56.9145,
+ "eval_samples_per_second": 42.906,
+ "eval_steps_per_second": 1.353,
+ "step": 864
+ },
+ {
+ "epoch": 12.013973799126637,
+ "grad_norm": 0.017597166821360588,
+ "learning_rate": 0.0006,
+ "loss": 4.830524444580078,
+ "step": 865
+ },
+ {
+ "epoch": 12.027947598253276,
+ "grad_norm": 0.01716788113117218,
+ "learning_rate": 0.0006,
+ "loss": 4.894325256347656,
+ "step": 866
+ },
+ {
+ "epoch": 12.041921397379912,
+ "grad_norm": 0.017712457105517387,
+ "learning_rate": 0.0006,
+ "loss": 4.868297576904297,
+ "step": 867
+ },
+ {
+ "epoch": 12.055895196506551,
+ "grad_norm": 0.021486390382051468,
+ "learning_rate": 0.0006,
+ "loss": 4.922049522399902,
+ "step": 868
+ },
+ {
+ "epoch": 12.069868995633188,
+ "grad_norm": 0.024560654535889626,
+ "learning_rate": 0.0006,
+ "loss": 4.734241008758545,
+ "step": 869
+ },
+ {
+ "epoch": 12.083842794759825,
+ "grad_norm": 0.025870252400636673,
+ "learning_rate": 0.0006,
+ "loss": 4.868566989898682,
+ "step": 870
+ },
+ {
+ "epoch": 12.097816593886463,
+ "grad_norm": 0.0277389045804739,
+ "learning_rate": 0.0006,
+ "loss": 4.755006790161133,
+ "step": 871
+ },
+ {
+ "epoch": 12.1117903930131,
+ "grad_norm": 0.029100844636559486,
+ "learning_rate": 0.0006,
+ "loss": 4.909667015075684,
+ "step": 872
+ },
+ {
+ "epoch": 12.125764192139737,
+ "grad_norm": 0.031833067536354065,
+ "learning_rate": 0.0006,
+ "loss": 4.858674049377441,
+ "step": 873
+ },
+ {
+ "epoch": 12.139737991266376,
+ "grad_norm": 0.032709237188100815,
+ "learning_rate": 0.0006,
+ "loss": 4.823696136474609,
+ "step": 874
+ },
+ {
+ "epoch": 12.153711790393013,
+ "grad_norm": 0.02995380386710167,
+ "learning_rate": 0.0006,
+ "loss": 4.850235462188721,
+ "step": 875
+ },
+ {
+ "epoch": 12.167685589519651,
+ "grad_norm": 0.034098610281944275,
+ "learning_rate": 0.0006,
+ "loss": 4.8169989585876465,
+ "step": 876
+ },
+ {
+ "epoch": 12.181659388646288,
+ "grad_norm": 0.03377080336213112,
+ "learning_rate": 0.0006,
+ "loss": 4.90738582611084,
+ "step": 877
+ },
+ {
+ "epoch": 12.195633187772925,
+ "grad_norm": 0.029780825600028038,
+ "learning_rate": 0.0006,
+ "loss": 4.811163902282715,
+ "step": 878
+ },
+ {
+ "epoch": 12.209606986899564,
+ "grad_norm": 0.03118024580180645,
+ "learning_rate": 0.0006,
+ "loss": 4.861286640167236,
+ "step": 879
+ },
+ {
+ "epoch": 12.2235807860262,
+ "grad_norm": 0.033052004873752594,
+ "learning_rate": 0.0006,
+ "loss": 4.865021228790283,
+ "step": 880
+ },
+ {
+ "epoch": 12.237554585152838,
+ "grad_norm": 0.029070032760500908,
+ "learning_rate": 0.0006,
+ "loss": 4.947303771972656,
+ "step": 881
+ },
+ {
+ "epoch": 12.251528384279476,
+ "grad_norm": 0.028355715796351433,
+ "learning_rate": 0.0006,
+ "loss": 4.804648399353027,
+ "step": 882
+ },
+ {
+ "epoch": 12.265502183406113,
+ "grad_norm": 0.031257227063179016,
+ "learning_rate": 0.0006,
+ "loss": 4.757718086242676,
+ "step": 883
+ },
+ {
+ "epoch": 12.279475982532752,
+ "grad_norm": 0.031438373029232025,
+ "learning_rate": 0.0006,
+ "loss": 4.851268768310547,
+ "step": 884
+ },
+ {
+ "epoch": 12.293449781659389,
+ "grad_norm": 0.030492138117551804,
+ "learning_rate": 0.0006,
+ "loss": 4.7891035079956055,
+ "step": 885
+ },
+ {
+ "epoch": 12.307423580786025,
+ "grad_norm": 0.026115991175174713,
+ "learning_rate": 0.0006,
+ "loss": 4.86474609375,
+ "step": 886
+ },
+ {
+ "epoch": 12.321397379912664,
+ "grad_norm": 0.024184465408325195,
+ "learning_rate": 0.0006,
+ "loss": 4.860456943511963,
+ "step": 887
+ },
+ {
+ "epoch": 12.335371179039301,
+ "grad_norm": 0.025437982752919197,
+ "learning_rate": 0.0006,
+ "loss": 4.806251049041748,
+ "step": 888
+ },
+ {
+ "epoch": 12.34934497816594,
+ "grad_norm": 0.022390231490135193,
+ "learning_rate": 0.0006,
+ "loss": 4.807086944580078,
+ "step": 889
+ },
+ {
+ "epoch": 12.363318777292577,
+ "grad_norm": 0.020174294710159302,
+ "learning_rate": 0.0006,
+ "loss": 4.78292179107666,
+ "step": 890
+ },
+ {
+ "epoch": 12.377292576419213,
+ "grad_norm": 0.019734324887394905,
+ "learning_rate": 0.0006,
+ "loss": 4.784509658813477,
+ "step": 891
+ },
+ {
+ "epoch": 12.391266375545852,
+ "grad_norm": 0.016733523458242416,
+ "learning_rate": 0.0006,
+ "loss": 4.811755180358887,
+ "step": 892
+ },
+ {
+ "epoch": 12.405240174672489,
+ "grad_norm": 0.01570146717131138,
+ "learning_rate": 0.0006,
+ "loss": 4.720088005065918,
+ "step": 893
+ },
+ {
+ "epoch": 12.419213973799126,
+ "grad_norm": 0.01601393334567547,
+ "learning_rate": 0.0006,
+ "loss": 4.783543109893799,
+ "step": 894
+ },
+ {
+ "epoch": 12.433187772925764,
+ "grad_norm": 0.014035487547516823,
+ "learning_rate": 0.0006,
+ "loss": 4.848340034484863,
+ "step": 895
+ },
+ {
+ "epoch": 12.447161572052401,
+ "grad_norm": 0.015248487703502178,
+ "learning_rate": 0.0006,
+ "loss": 4.773287773132324,
+ "step": 896
+ },
+ {
+ "epoch": 12.46113537117904,
+ "grad_norm": 0.01532268151640892,
+ "learning_rate": 0.0006,
+ "loss": 4.799646854400635,
+ "step": 897
+ },
+ {
+ "epoch": 12.475109170305677,
+ "grad_norm": 0.013982534408569336,
+ "learning_rate": 0.0006,
+ "loss": 4.8090500831604,
+ "step": 898
+ },
+ {
+ "epoch": 12.489082969432314,
+ "grad_norm": 0.014081164263188839,
+ "learning_rate": 0.0006,
+ "loss": 4.751888751983643,
+ "step": 899
+ },
+ {
+ "epoch": 12.503056768558952,
+ "grad_norm": 0.01380133256316185,
+ "learning_rate": 0.0006,
+ "loss": 4.78144645690918,
+ "step": 900
+ },
+ {
+ "epoch": 12.51703056768559,
+ "grad_norm": 0.013555269688367844,
+ "learning_rate": 0.0006,
+ "loss": 4.779257774353027,
+ "step": 901
+ },
+ {
+ "epoch": 12.531004366812226,
+ "grad_norm": 0.013075701892375946,
+ "learning_rate": 0.0006,
+ "loss": 4.820705413818359,
+ "step": 902
+ },
+ {
+ "epoch": 12.544978165938865,
+ "grad_norm": 0.01276316400617361,
+ "learning_rate": 0.0006,
+ "loss": 4.836733818054199,
+ "step": 903
+ },
+ {
+ "epoch": 12.558951965065502,
+ "grad_norm": 0.013033892959356308,
+ "learning_rate": 0.0006,
+ "loss": 4.764178276062012,
+ "step": 904
+ },
+ {
+ "epoch": 12.57292576419214,
+ "grad_norm": 0.013996980153024197,
+ "learning_rate": 0.0006,
+ "loss": 4.767983436584473,
+ "step": 905
+ },
+ {
+ "epoch": 12.586899563318777,
+ "grad_norm": 0.015222184360027313,
+ "learning_rate": 0.0006,
+ "loss": 4.759721755981445,
+ "step": 906
+ },
+ {
+ "epoch": 12.600873362445414,
+ "grad_norm": 0.01623941957950592,
+ "learning_rate": 0.0006,
+ "loss": 4.849811553955078,
+ "step": 907
+ },
+ {
+ "epoch": 12.614847161572053,
+ "grad_norm": 0.015711182728409767,
+ "learning_rate": 0.0006,
+ "loss": 4.7829084396362305,
+ "step": 908
+ },
+ {
+ "epoch": 12.62882096069869,
+ "grad_norm": 0.014137467369437218,
+ "learning_rate": 0.0006,
+ "loss": 4.73321533203125,
+ "step": 909
+ },
+ {
+ "epoch": 12.642794759825328,
+ "grad_norm": 0.012738462537527084,
+ "learning_rate": 0.0006,
+ "loss": 4.870052337646484,
+ "step": 910
+ },
+ {
+ "epoch": 12.656768558951965,
+ "grad_norm": 0.013683130033314228,
+ "learning_rate": 0.0006,
+ "loss": 4.82658052444458,
+ "step": 911
+ },
+ {
+ "epoch": 12.670742358078602,
+ "grad_norm": 0.014409830793738365,
+ "learning_rate": 0.0006,
+ "loss": 4.707494735717773,
+ "step": 912
+ },
+ {
+ "epoch": 12.68471615720524,
+ "grad_norm": 0.013806809671223164,
+ "learning_rate": 0.0006,
+ "loss": 4.778928279876709,
+ "step": 913
+ },
+ {
+ "epoch": 12.698689956331878,
+ "grad_norm": 0.012857912108302116,
+ "learning_rate": 0.0006,
+ "loss": 4.890717506408691,
+ "step": 914
+ },
+ {
+ "epoch": 12.712663755458514,
+ "grad_norm": 0.013828758150339127,
+ "learning_rate": 0.0006,
+ "loss": 4.8357439041137695,
+ "step": 915
+ },
+ {
+ "epoch": 12.726637554585153,
+ "grad_norm": 0.01478351280093193,
+ "learning_rate": 0.0006,
+ "loss": 4.869986534118652,
+ "step": 916
+ },
+ {
+ "epoch": 12.74061135371179,
+ "grad_norm": 0.01326421182602644,
+ "learning_rate": 0.0006,
+ "loss": 4.830646514892578,
+ "step": 917
+ },
+ {
+ "epoch": 12.754585152838429,
+ "grad_norm": 0.013335433788597584,
+ "learning_rate": 0.0006,
+ "loss": 4.7786149978637695,
+ "step": 918
+ },
+ {
+ "epoch": 12.768558951965066,
+ "grad_norm": 0.014261147007346153,
+ "learning_rate": 0.0006,
+ "loss": 4.788693428039551,
+ "step": 919
+ },
+ {
+ "epoch": 12.782532751091702,
+ "grad_norm": 0.014274783432483673,
+ "learning_rate": 0.0006,
+ "loss": 4.77295446395874,
+ "step": 920
+ },
+ {
+ "epoch": 12.796506550218341,
+ "grad_norm": 0.015074674971401691,
+ "learning_rate": 0.0006,
+ "loss": 4.830754280090332,
+ "step": 921
+ },
+ {
+ "epoch": 12.810480349344978,
+ "grad_norm": 0.015599515289068222,
+ "learning_rate": 0.0006,
+ "loss": 4.837759017944336,
+ "step": 922
+ },
+ {
+ "epoch": 12.824454148471617,
+ "grad_norm": 0.014595242217183113,
+ "learning_rate": 0.0006,
+ "loss": 4.882965564727783,
+ "step": 923
+ },
+ {
+ "epoch": 12.838427947598253,
+ "grad_norm": 0.014951467514038086,
+ "learning_rate": 0.0006,
+ "loss": 4.709748268127441,
+ "step": 924
+ },
+ {
+ "epoch": 12.85240174672489,
+ "grad_norm": 0.01539295818656683,
+ "learning_rate": 0.0006,
+ "loss": 4.791160583496094,
+ "step": 925
+ },
+ {
+ "epoch": 12.866375545851529,
+ "grad_norm": 0.015987906605005264,
+ "learning_rate": 0.0006,
+ "loss": 4.661876678466797,
+ "step": 926
+ },
+ {
+ "epoch": 12.880349344978166,
+ "grad_norm": 0.016527341678738594,
+ "learning_rate": 0.0006,
+ "loss": 4.721952438354492,
+ "step": 927
+ },
+ {
+ "epoch": 12.894323144104803,
+ "grad_norm": 0.01637161523103714,
+ "learning_rate": 0.0006,
+ "loss": 4.704446792602539,
+ "step": 928
+ },
+ {
+ "epoch": 12.908296943231441,
+ "grad_norm": 0.015221050009131432,
+ "learning_rate": 0.0006,
+ "loss": 4.771907806396484,
+ "step": 929
+ },
+ {
+ "epoch": 12.922270742358078,
+ "grad_norm": 0.016242166981101036,
+ "learning_rate": 0.0006,
+ "loss": 4.806267738342285,
+ "step": 930
+ },
+ {
+ "epoch": 12.936244541484717,
+ "grad_norm": 0.01628357358276844,
+ "learning_rate": 0.0006,
+ "loss": 4.7252326011657715,
+ "step": 931
+ },
+ {
+ "epoch": 12.950218340611354,
+ "grad_norm": 0.017428133636713028,
+ "learning_rate": 0.0006,
+ "loss": 4.765689849853516,
+ "step": 932
+ },
+ {
+ "epoch": 12.96419213973799,
+ "grad_norm": 0.01870061084628105,
+ "learning_rate": 0.0006,
+ "loss": 4.814488887786865,
+ "step": 933
+ },
+ {
+ "epoch": 12.97816593886463,
+ "grad_norm": 0.019796060398221016,
+ "learning_rate": 0.0006,
+ "loss": 4.739185810089111,
+ "step": 934
+ },
+ {
+ "epoch": 12.992139737991266,
+ "grad_norm": 0.01985820196568966,
+ "learning_rate": 0.0006,
+ "loss": 4.810608863830566,
+ "step": 935
+ },
+ {
+ "epoch": 13.0,
+ "grad_norm": 0.020834477618336678,
+ "learning_rate": 0.0006,
+ "loss": 4.727858066558838,
+ "step": 936
+ },
+ {
+ "epoch": 13.0,
+ "eval_loss": 4.979116916656494,
+ "eval_runtime": 57.2877,
+ "eval_samples_per_second": 42.627,
+ "eval_steps_per_second": 1.344,
+ "step": 936
+ },
+ {
+ "epoch": 13.013973799126637,
+ "grad_norm": 0.02025519125163555,
+ "learning_rate": 0.0006,
+ "loss": 4.628875732421875,
+ "step": 937
+ },
+ {
+ "epoch": 13.027947598253276,
+ "grad_norm": 0.019596368074417114,
+ "learning_rate": 0.0006,
+ "loss": 4.694699287414551,
+ "step": 938
+ },
+ {
+ "epoch": 13.041921397379912,
+ "grad_norm": 0.020494934171438217,
+ "learning_rate": 0.0006,
+ "loss": 4.691821098327637,
+ "step": 939
+ },
+ {
+ "epoch": 13.055895196506551,
+ "grad_norm": 0.020121760666370392,
+ "learning_rate": 0.0006,
+ "loss": 4.71478271484375,
+ "step": 940
+ },
+ {
+ "epoch": 13.069868995633188,
+ "grad_norm": 0.021091405302286148,
+ "learning_rate": 0.0006,
+ "loss": 4.695915222167969,
+ "step": 941
+ },
+ {
+ "epoch": 13.083842794759825,
+ "grad_norm": 0.019787423312664032,
+ "learning_rate": 0.0006,
+ "loss": 4.8652448654174805,
+ "step": 942
+ },
+ {
+ "epoch": 13.097816593886463,
+ "grad_norm": 0.0185833927243948,
+ "learning_rate": 0.0006,
+ "loss": 4.7065887451171875,
+ "step": 943
+ },
+ {
+ "epoch": 13.1117903930131,
+ "grad_norm": 0.021220475435256958,
+ "learning_rate": 0.0006,
+ "loss": 4.741232395172119,
+ "step": 944
+ },
+ {
+ "epoch": 13.125764192139737,
+ "grad_norm": 0.02136622555553913,
+ "learning_rate": 0.0006,
+ "loss": 4.705967903137207,
+ "step": 945
+ },
+ {
+ "epoch": 13.139737991266376,
+ "grad_norm": 0.0191587433218956,
+ "learning_rate": 0.0006,
+ "loss": 4.758667945861816,
+ "step": 946
+ },
+ {
+ "epoch": 13.153711790393013,
+ "grad_norm": 0.018290022388100624,
+ "learning_rate": 0.0006,
+ "loss": 4.760666847229004,
+ "step": 947
+ },
+ {
+ "epoch": 13.167685589519651,
+ "grad_norm": 0.021704409271478653,
+ "learning_rate": 0.0006,
+ "loss": 4.750555992126465,
+ "step": 948
+ },
+ {
+ "epoch": 13.181659388646288,
+ "grad_norm": 0.021668918430805206,
+ "learning_rate": 0.0006,
+ "loss": 4.840903282165527,
+ "step": 949
+ },
+ {
+ "epoch": 13.195633187772925,
+ "grad_norm": 0.019497141242027283,
+ "learning_rate": 0.0006,
+ "loss": 4.738969326019287,
+ "step": 950
+ },
+ {
+ "epoch": 13.209606986899564,
+ "grad_norm": 0.018500596284866333,
+ "learning_rate": 0.0006,
+ "loss": 4.75154972076416,
+ "step": 951
+ },
+ {
+ "epoch": 13.2235807860262,
+ "grad_norm": 0.01872239261865616,
+ "learning_rate": 0.0006,
+ "loss": 4.811738014221191,
+ "step": 952
+ },
+ {
+ "epoch": 13.237554585152838,
+ "grad_norm": 0.020321547985076904,
+ "learning_rate": 0.0006,
+ "loss": 4.815384387969971,
+ "step": 953
+ },
+ {
+ "epoch": 13.251528384279476,
+ "grad_norm": 0.02283575013279915,
+ "learning_rate": 0.0006,
+ "loss": 4.74582052230835,
+ "step": 954
+ },
+ {
+ "epoch": 13.265502183406113,
+ "grad_norm": 0.02474398910999298,
+ "learning_rate": 0.0006,
+ "loss": 4.738375186920166,
+ "step": 955
+ },
+ {
+ "epoch": 13.279475982532752,
+ "grad_norm": 0.02065764181315899,
+ "learning_rate": 0.0006,
+ "loss": 4.746875286102295,
+ "step": 956
+ },
+ {
+ "epoch": 13.293449781659389,
+ "grad_norm": 0.020534520968794823,
+ "learning_rate": 0.0006,
+ "loss": 4.656914710998535,
+ "step": 957
+ },
+ {
+ "epoch": 13.307423580786025,
+ "grad_norm": 0.0206378772854805,
+ "learning_rate": 0.0006,
+ "loss": 4.842615604400635,
+ "step": 958
+ },
+ {
+ "epoch": 13.321397379912664,
+ "grad_norm": 0.017935654148459435,
+ "learning_rate": 0.0006,
+ "loss": 4.717827796936035,
+ "step": 959
+ },
+ {
+ "epoch": 13.335371179039301,
+ "grad_norm": 0.016958363354206085,
+ "learning_rate": 0.0006,
+ "loss": 4.681070327758789,
+ "step": 960
+ },
+ {
+ "epoch": 13.34934497816594,
+ "grad_norm": 0.015735412016510963,
+ "learning_rate": 0.0006,
+ "loss": 4.7957048416137695,
+ "step": 961
+ },
+ {
+ "epoch": 13.363318777292577,
+ "grad_norm": 0.01563699170947075,
+ "learning_rate": 0.0006,
+ "loss": 4.701192378997803,
+ "step": 962
+ },
+ {
+ "epoch": 13.377292576419213,
+ "grad_norm": 0.015207760035991669,
+ "learning_rate": 0.0006,
+ "loss": 4.73256254196167,
+ "step": 963
+ },
+ {
+ "epoch": 13.391266375545852,
+ "grad_norm": 0.015396883711218834,
+ "learning_rate": 0.0006,
+ "loss": 4.76412296295166,
+ "step": 964
+ },
+ {
+ "epoch": 13.405240174672489,
+ "grad_norm": 0.015690583735704422,
+ "learning_rate": 0.0006,
+ "loss": 4.791157245635986,
+ "step": 965
+ },
+ {
+ "epoch": 13.419213973799126,
+ "grad_norm": 0.01755724661052227,
+ "learning_rate": 0.0006,
+ "loss": 4.845666885375977,
+ "step": 966
+ },
+ {
+ "epoch": 13.433187772925764,
+ "grad_norm": 0.019443422555923462,
+ "learning_rate": 0.0006,
+ "loss": 4.5902299880981445,
+ "step": 967
+ },
+ {
+ "epoch": 13.447161572052401,
+ "grad_norm": 0.02118338644504547,
+ "learning_rate": 0.0006,
+ "loss": 4.740627288818359,
+ "step": 968
+ },
+ {
+ "epoch": 13.46113537117904,
+ "grad_norm": 0.02174839936196804,
+ "learning_rate": 0.0006,
+ "loss": 4.814560890197754,
+ "step": 969
+ },
+ {
+ "epoch": 13.475109170305677,
+ "grad_norm": 0.022860374301671982,
+ "learning_rate": 0.0006,
+ "loss": 4.66440486907959,
+ "step": 970
+ },
+ {
+ "epoch": 13.489082969432314,
+ "grad_norm": 0.024742677807807922,
+ "learning_rate": 0.0006,
+ "loss": 4.7451653480529785,
+ "step": 971
+ },
+ {
+ "epoch": 13.503056768558952,
+ "grad_norm": 0.027068907395005226,
+ "learning_rate": 0.0006,
+ "loss": 4.802584171295166,
+ "step": 972
+ },
+ {
+ "epoch": 13.51703056768559,
+ "grad_norm": 0.026316625997424126,
+ "learning_rate": 0.0006,
+ "loss": 4.842397689819336,
+ "step": 973
+ },
+ {
+ "epoch": 13.531004366812226,
+ "grad_norm": 0.024610962718725204,
+ "learning_rate": 0.0006,
+ "loss": 4.650721073150635,
+ "step": 974
+ },
+ {
+ "epoch": 13.544978165938865,
+ "grad_norm": 0.024122603237628937,
+ "learning_rate": 0.0006,
+ "loss": 4.6984357833862305,
+ "step": 975
+ },
+ {
+ "epoch": 13.558951965065502,
+ "grad_norm": 0.022699708119034767,
+ "learning_rate": 0.0006,
+ "loss": 4.793007850646973,
+ "step": 976
+ },
+ {
+ "epoch": 13.57292576419214,
+ "grad_norm": 0.021706534549593925,
+ "learning_rate": 0.0006,
+ "loss": 4.691362380981445,
+ "step": 977
+ },
+ {
+ "epoch": 13.586899563318777,
+ "grad_norm": 0.021363291889429092,
+ "learning_rate": 0.0006,
+ "loss": 4.76762580871582,
+ "step": 978
+ },
+ {
+ "epoch": 13.600873362445414,
+ "grad_norm": 0.023476677015423775,
+ "learning_rate": 0.0006,
+ "loss": 4.777202606201172,
+ "step": 979
+ },
+ {
+ "epoch": 13.614847161572053,
+ "grad_norm": 0.02457100711762905,
+ "learning_rate": 0.0006,
+ "loss": 4.590703010559082,
+ "step": 980
+ },
+ {
+ "epoch": 13.62882096069869,
+ "grad_norm": 0.02469099499285221,
+ "learning_rate": 0.0006,
+ "loss": 4.640185356140137,
+ "step": 981
+ },
+ {
+ "epoch": 13.642794759825328,
+ "grad_norm": 0.024914603680372238,
+ "learning_rate": 0.0006,
+ "loss": 4.772201061248779,
+ "step": 982
+ },
+ {
+ "epoch": 13.656768558951965,
+ "grad_norm": 0.025133173912763596,
+ "learning_rate": 0.0006,
+ "loss": 4.743337154388428,
+ "step": 983
+ },
+ {
+ "epoch": 13.670742358078602,
+ "grad_norm": 0.02034137211740017,
+ "learning_rate": 0.0006,
+ "loss": 4.752127647399902,
+ "step": 984
+ },
+ {
+ "epoch": 13.68471615720524,
+ "grad_norm": 0.019645661115646362,
+ "learning_rate": 0.0006,
+ "loss": 4.838453769683838,
+ "step": 985
+ },
+ {
+ "epoch": 13.698689956331878,
+ "grad_norm": 0.01900297962129116,
+ "learning_rate": 0.0006,
+ "loss": 4.804868221282959,
+ "step": 986
+ },
+ {
+ "epoch": 13.712663755458514,
+ "grad_norm": 0.020292341709136963,
+ "learning_rate": 0.0006,
+ "loss": 4.609743118286133,
+ "step": 987
+ },
+ {
+ "epoch": 13.726637554585153,
+ "grad_norm": 0.017014775425195694,
+ "learning_rate": 0.0006,
+ "loss": 4.740612983703613,
+ "step": 988
+ },
+ {
+ "epoch": 13.74061135371179,
+ "grad_norm": 0.015199844725430012,
+ "learning_rate": 0.0006,
+ "loss": 4.719701766967773,
+ "step": 989
+ },
+ {
+ "epoch": 13.754585152838429,
+ "grad_norm": 0.015125414356589317,
+ "learning_rate": 0.0006,
+ "loss": 4.686114311218262,
+ "step": 990
+ },
+ {
+ "epoch": 13.768558951965066,
+ "grad_norm": 0.014779540710151196,
+ "learning_rate": 0.0006,
+ "loss": 4.674347877502441,
+ "step": 991
+ },
+ {
+ "epoch": 13.782532751091702,
+ "grad_norm": 0.015255163423717022,
+ "learning_rate": 0.0006,
+ "loss": 4.6309309005737305,
+ "step": 992
+ },
+ {
+ "epoch": 13.796506550218341,
+ "grad_norm": 0.013692197389900684,
+ "learning_rate": 0.0006,
+ "loss": 4.698497295379639,
+ "step": 993
+ },
+ {
+ "epoch": 13.810480349344978,
+ "grad_norm": 0.014529623091220856,
+ "learning_rate": 0.0006,
+ "loss": 4.628443717956543,
+ "step": 994
+ },
+ {
+ "epoch": 13.824454148471617,
+ "grad_norm": 0.01576659083366394,
+ "learning_rate": 0.0006,
+ "loss": 4.788015842437744,
+ "step": 995
+ },
+ {
+ "epoch": 13.838427947598253,
+ "grad_norm": 0.015834277495741844,
+ "learning_rate": 0.0006,
+ "loss": 4.672703742980957,
+ "step": 996
+ },
+ {
+ "epoch": 13.85240174672489,
+ "grad_norm": 0.015539503656327724,
+ "learning_rate": 0.0006,
+ "loss": 4.731503486633301,
+ "step": 997
+ },
+ {
+ "epoch": 13.866375545851529,
+ "grad_norm": 0.013463746756315231,
+ "learning_rate": 0.0006,
+ "loss": 4.7006330490112305,
+ "step": 998
+ },
+ {
+ "epoch": 13.880349344978166,
+ "grad_norm": 0.016377726569771767,
+ "learning_rate": 0.0006,
+ "loss": 4.636074066162109,
+ "step": 999
+ },
+ {
+ "epoch": 13.894323144104803,
+ "grad_norm": 0.016660034656524658,
+ "learning_rate": 0.0006,
+ "loss": 4.742693901062012,
+ "step": 1000
+ },
+ {
+ "epoch": 13.908296943231441,
+ "grad_norm": 0.014902938157320023,
+ "learning_rate": 0.0006,
+ "loss": 4.769811630249023,
+ "step": 1001
+ },
+ {
+ "epoch": 13.922270742358078,
+ "grad_norm": 0.014883476309478283,
+ "learning_rate": 0.0006,
+ "loss": 4.731159210205078,
+ "step": 1002
+ },
+ {
+ "epoch": 13.936244541484717,
+ "grad_norm": 0.014494217932224274,
+ "learning_rate": 0.0006,
+ "loss": 4.659485816955566,
+ "step": 1003
+ },
+ {
+ "epoch": 13.950218340611354,
+ "grad_norm": 0.015426991507411003,
+ "learning_rate": 0.0006,
+ "loss": 4.755229949951172,
+ "step": 1004
+ },
+ {
+ "epoch": 13.96419213973799,
+ "grad_norm": 0.016648586839437485,
+ "learning_rate": 0.0006,
+ "loss": 4.74111270904541,
+ "step": 1005
+ },
+ {
+ "epoch": 13.97816593886463,
+ "grad_norm": 0.016980541869997978,
+ "learning_rate": 0.0006,
+ "loss": 4.74072265625,
+ "step": 1006
+ },
+ {
+ "epoch": 13.992139737991266,
+ "grad_norm": 0.018070152029395103,
+ "learning_rate": 0.0006,
+ "loss": 4.732975006103516,
+ "step": 1007
+ },
+ {
+ "epoch": 14.0,
+ "grad_norm": 0.019910665228962898,
+ "learning_rate": 0.0006,
+ "loss": 4.634697914123535,
+ "step": 1008
+ },
+ {
+ "epoch": 14.0,
+ "eval_loss": 4.850640296936035,
+ "eval_runtime": 57.3039,
+ "eval_samples_per_second": 42.615,
+ "eval_steps_per_second": 1.344,
+ "step": 1008
+ },
+ {
+ "epoch": 14.013973799126637,
+ "grad_norm": 0.01980053074657917,
+ "learning_rate": 0.0006,
+ "loss": 4.628744602203369,
+ "step": 1009
+ },
+ {
+ "epoch": 14.027947598253276,
+ "grad_norm": 0.02497623674571514,
+ "learning_rate": 0.0006,
+ "loss": 4.75930118560791,
+ "step": 1010
+ },
+ {
+ "epoch": 14.041921397379912,
+ "grad_norm": 0.02794097363948822,
+ "learning_rate": 0.0006,
+ "loss": 4.691340446472168,
+ "step": 1011
+ },
+ {
+ "epoch": 14.055895196506551,
+ "grad_norm": 0.023725535720586777,
+ "learning_rate": 0.0006,
+ "loss": 4.733664512634277,
+ "step": 1012
+ },
+ {
+ "epoch": 14.069868995633188,
+ "grad_norm": 0.022966474294662476,
+ "learning_rate": 0.0006,
+ "loss": 4.664808750152588,
+ "step": 1013
+ },
+ {
+ "epoch": 14.083842794759825,
+ "grad_norm": 0.02530563250184059,
+ "learning_rate": 0.0006,
+ "loss": 4.8047285079956055,
+ "step": 1014
+ },
+ {
+ "epoch": 14.097816593886463,
+ "grad_norm": 0.02491082064807415,
+ "learning_rate": 0.0006,
+ "loss": 4.667956352233887,
+ "step": 1015
+ },
+ {
+ "epoch": 14.1117903930131,
+ "grad_norm": 0.022110294550657272,
+ "learning_rate": 0.0006,
+ "loss": 4.666140556335449,
+ "step": 1016
+ },
+ {
+ "epoch": 14.125764192139737,
+ "grad_norm": 0.02065849117934704,
+ "learning_rate": 0.0006,
+ "loss": 4.581960678100586,
+ "step": 1017
+ },
+ {
+ "epoch": 14.139737991266376,
+ "grad_norm": 0.01941298507153988,
+ "learning_rate": 0.0006,
+ "loss": 4.614226341247559,
+ "step": 1018
+ },
+ {
+ "epoch": 14.153711790393013,
+ "grad_norm": 0.022441856563091278,
+ "learning_rate": 0.0006,
+ "loss": 4.638352394104004,
+ "step": 1019
+ },
+ {
+ "epoch": 14.167685589519651,
+ "grad_norm": 0.021519694477319717,
+ "learning_rate": 0.0006,
+ "loss": 4.700160980224609,
+ "step": 1020
+ },
+ {
+ "epoch": 14.181659388646288,
+ "grad_norm": 0.018956486135721207,
+ "learning_rate": 0.0006,
+ "loss": 4.649593353271484,
+ "step": 1021
+ },
+ {
+ "epoch": 14.195633187772925,
+ "grad_norm": 0.016681140288710594,
+ "learning_rate": 0.0006,
+ "loss": 4.603952407836914,
+ "step": 1022
+ },
+ {
+ "epoch": 14.209606986899564,
+ "grad_norm": 0.01788809522986412,
+ "learning_rate": 0.0006,
+ "loss": 4.669160842895508,
+ "step": 1023
+ },
+ {
+ "epoch": 14.2235807860262,
+ "grad_norm": 0.016047228127717972,
+ "learning_rate": 0.0006,
+ "loss": 4.6367950439453125,
+ "step": 1024
+ },
+ {
+ "epoch": 14.237554585152838,
+ "grad_norm": 0.01614239625632763,
+ "learning_rate": 0.0006,
+ "loss": 4.664529323577881,
+ "step": 1025
+ },
+ {
+ "epoch": 14.251528384279476,
+ "grad_norm": 0.016362549737095833,
+ "learning_rate": 0.0006,
+ "loss": 4.702373504638672,
+ "step": 1026
+ },
+ {
+ "epoch": 14.265502183406113,
+ "grad_norm": 0.015800409018993378,
+ "learning_rate": 0.0006,
+ "loss": 4.7594428062438965,
+ "step": 1027
+ },
+ {
+ "epoch": 14.279475982532752,
+ "grad_norm": 0.016730139032006264,
+ "learning_rate": 0.0006,
+ "loss": 4.677244186401367,
+ "step": 1028
+ },
+ {
+ "epoch": 14.293449781659389,
+ "grad_norm": 0.01661759801208973,
+ "learning_rate": 0.0006,
+ "loss": 4.633679389953613,
+ "step": 1029
+ },
+ {
+ "epoch": 14.307423580786025,
+ "grad_norm": 0.017673347145318985,
+ "learning_rate": 0.0006,
+ "loss": 4.649107456207275,
+ "step": 1030
+ },
+ {
+ "epoch": 14.321397379912664,
+ "grad_norm": 0.019243352115154266,
+ "learning_rate": 0.0006,
+ "loss": 4.611257553100586,
+ "step": 1031
+ },
+ {
+ "epoch": 14.335371179039301,
+ "grad_norm": 0.019817164167761803,
+ "learning_rate": 0.0006,
+ "loss": 4.588086128234863,
+ "step": 1032
+ },
+ {
+ "epoch": 14.34934497816594,
+ "grad_norm": 0.022292211651802063,
+ "learning_rate": 0.0006,
+ "loss": 4.783066749572754,
+ "step": 1033
+ },
+ {
+ "epoch": 14.363318777292577,
+ "grad_norm": 0.02140096202492714,
+ "learning_rate": 0.0006,
+ "loss": 4.604718208312988,
+ "step": 1034
+ },
+ {
+ "epoch": 14.377292576419213,
+ "grad_norm": 0.0187817495316267,
+ "learning_rate": 0.0006,
+ "loss": 4.721773147583008,
+ "step": 1035
+ },
+ {
+ "epoch": 14.391266375545852,
+ "grad_norm": 0.019879210740327835,
+ "learning_rate": 0.0006,
+ "loss": 4.789286136627197,
+ "step": 1036
+ },
+ {
+ "epoch": 14.405240174672489,
+ "grad_norm": 0.01885785162448883,
+ "learning_rate": 0.0006,
+ "loss": 4.669743537902832,
+ "step": 1037
+ },
+ {
+ "epoch": 14.419213973799126,
+ "grad_norm": 0.018324270844459534,
+ "learning_rate": 0.0006,
+ "loss": 4.65901517868042,
+ "step": 1038
+ },
+ {
+ "epoch": 14.433187772925764,
+ "grad_norm": 0.017664821818470955,
+ "learning_rate": 0.0006,
+ "loss": 4.667839527130127,
+ "step": 1039
+ },
+ {
+ "epoch": 14.447161572052401,
+ "grad_norm": 0.015702905133366585,
+ "learning_rate": 0.0006,
+ "loss": 4.64700984954834,
+ "step": 1040
+ },
+ {
+ "epoch": 14.46113537117904,
+ "grad_norm": 0.016674194484949112,
+ "learning_rate": 0.0006,
+ "loss": 4.694093227386475,
+ "step": 1041
+ },
+ {
+ "epoch": 14.475109170305677,
+ "grad_norm": 0.016869477927684784,
+ "learning_rate": 0.0006,
+ "loss": 4.619544982910156,
+ "step": 1042
+ },
+ {
+ "epoch": 14.489082969432314,
+ "grad_norm": 0.016557540744543076,
+ "learning_rate": 0.0006,
+ "loss": 4.568667411804199,
+ "step": 1043
+ },
+ {
+ "epoch": 14.503056768558952,
+ "grad_norm": 0.014500945806503296,
+ "learning_rate": 0.0006,
+ "loss": 4.561741828918457,
+ "step": 1044
+ },
+ {
+ "epoch": 14.51703056768559,
+ "grad_norm": 0.014228739775717258,
+ "learning_rate": 0.0006,
+ "loss": 4.654277324676514,
+ "step": 1045
+ },
+ {
+ "epoch": 14.531004366812226,
+ "grad_norm": 0.013894238509237766,
+ "learning_rate": 0.0006,
+ "loss": 4.5867719650268555,
+ "step": 1046
+ },
+ {
+ "epoch": 14.544978165938865,
+ "grad_norm": 0.01550068985670805,
+ "learning_rate": 0.0006,
+ "loss": 4.61168098449707,
+ "step": 1047
+ },
+ {
+ "epoch": 14.558951965065502,
+ "grad_norm": 0.015940772369503975,
+ "learning_rate": 0.0006,
+ "loss": 4.691247940063477,
+ "step": 1048
+ },
+ {
+ "epoch": 14.57292576419214,
+ "grad_norm": 0.01714535616338253,
+ "learning_rate": 0.0006,
+ "loss": 4.61183500289917,
+ "step": 1049
+ },
+ {
+ "epoch": 14.586899563318777,
+ "grad_norm": 0.017925038933753967,
+ "learning_rate": 0.0006,
+ "loss": 4.719022750854492,
+ "step": 1050
+ },
+ {
+ "epoch": 14.600873362445414,
+ "grad_norm": 0.018141552805900574,
+ "learning_rate": 0.0006,
+ "loss": 4.760281562805176,
+ "step": 1051
+ },
+ {
+ "epoch": 14.614847161572053,
+ "grad_norm": 0.017131328582763672,
+ "learning_rate": 0.0006,
+ "loss": 4.709680080413818,
+ "step": 1052
+ },
+ {
+ "epoch": 14.62882096069869,
+ "grad_norm": 0.018730048090219498,
+ "learning_rate": 0.0006,
+ "loss": 4.5994720458984375,
+ "step": 1053
+ },
+ {
+ "epoch": 14.642794759825328,
+ "grad_norm": 0.018894298002123833,
+ "learning_rate": 0.0006,
+ "loss": 4.669070720672607,
+ "step": 1054
+ },
+ {
+ "epoch": 14.656768558951965,
+ "grad_norm": 0.017643198370933533,
+ "learning_rate": 0.0006,
+ "loss": 4.641790390014648,
+ "step": 1055
+ },
+ {
+ "epoch": 14.670742358078602,
+ "grad_norm": 0.017647244036197662,
+ "learning_rate": 0.0006,
+ "loss": 4.653899192810059,
+ "step": 1056
+ },
+ {
+ "epoch": 14.68471615720524,
+ "grad_norm": 0.017304198816418648,
+ "learning_rate": 0.0006,
+ "loss": 4.7487473487854,
+ "step": 1057
+ },
+ {
+ "epoch": 14.698689956331878,
+ "grad_norm": 0.01638447679579258,
+ "learning_rate": 0.0006,
+ "loss": 4.71945858001709,
+ "step": 1058
+ },
+ {
+ "epoch": 14.712663755458514,
+ "grad_norm": 0.015892071649432182,
+ "learning_rate": 0.0006,
+ "loss": 4.744903564453125,
+ "step": 1059
+ },
+ {
+ "epoch": 14.726637554585153,
+ "grad_norm": 0.017420614138245583,
+ "learning_rate": 0.0006,
+ "loss": 4.781563758850098,
+ "step": 1060
+ },
+ {
+ "epoch": 14.74061135371179,
+ "grad_norm": 0.016498520970344543,
+ "learning_rate": 0.0006,
+ "loss": 4.622779846191406,
+ "step": 1061
+ },
+ {
+ "epoch": 14.754585152838429,
+ "grad_norm": 0.017482127994298935,
+ "learning_rate": 0.0006,
+ "loss": 4.62825870513916,
+ "step": 1062
+ },
+ {
+ "epoch": 14.768558951965066,
+ "grad_norm": 0.019018379971385002,
+ "learning_rate": 0.0006,
+ "loss": 4.693194389343262,
+ "step": 1063
+ },
+ {
+ "epoch": 14.782532751091702,
+ "grad_norm": 0.01993914321064949,
+ "learning_rate": 0.0006,
+ "loss": 4.742903709411621,
+ "step": 1064
+ },
+ {
+ "epoch": 14.796506550218341,
+ "grad_norm": 0.02301335707306862,
+ "learning_rate": 0.0006,
+ "loss": 4.613251686096191,
+ "step": 1065
+ },
+ {
+ "epoch": 14.810480349344978,
+ "grad_norm": 0.024088222533464432,
+ "learning_rate": 0.0006,
+ "loss": 4.594305992126465,
+ "step": 1066
+ },
+ {
+ "epoch": 14.824454148471617,
+ "grad_norm": 0.02493320405483246,
+ "learning_rate": 0.0006,
+ "loss": 4.700588226318359,
+ "step": 1067
+ },
+ {
+ "epoch": 14.838427947598253,
+ "grad_norm": 0.02501937560737133,
+ "learning_rate": 0.0006,
+ "loss": 4.601742744445801,
+ "step": 1068
+ },
+ {
+ "epoch": 14.85240174672489,
+ "grad_norm": 0.023490751162171364,
+ "learning_rate": 0.0006,
+ "loss": 4.589211463928223,
+ "step": 1069
+ },
+ {
+ "epoch": 14.866375545851529,
+ "grad_norm": 0.019695665687322617,
+ "learning_rate": 0.0006,
+ "loss": 4.635081768035889,
+ "step": 1070
+ },
+ {
+ "epoch": 14.880349344978166,
+ "grad_norm": 0.019276980310678482,
+ "learning_rate": 0.0006,
+ "loss": 4.737002372741699,
+ "step": 1071
+ },
+ {
+ "epoch": 14.894323144104803,
+ "grad_norm": 0.022430511191487312,
+ "learning_rate": 0.0006,
+ "loss": 4.808295726776123,
+ "step": 1072
+ },
+ {
+ "epoch": 14.908296943231441,
+ "grad_norm": 0.021543916314840317,
+ "learning_rate": 0.0006,
+ "loss": 4.559475421905518,
+ "step": 1073
+ },
+ {
+ "epoch": 14.922270742358078,
+ "grad_norm": 0.019885241985321045,
+ "learning_rate": 0.0006,
+ "loss": 4.647948741912842,
+ "step": 1074
+ },
+ {
+ "epoch": 14.936244541484717,
+ "grad_norm": 0.02115175500512123,
+ "learning_rate": 0.0006,
+ "loss": 4.680505752563477,
+ "step": 1075
+ },
+ {
+ "epoch": 14.950218340611354,
+ "grad_norm": 0.02156151458621025,
+ "learning_rate": 0.0006,
+ "loss": 4.673477649688721,
+ "step": 1076
+ },
+ {
+ "epoch": 14.96419213973799,
+ "grad_norm": 0.02184601128101349,
+ "learning_rate": 0.0006,
+ "loss": 4.644402027130127,
+ "step": 1077
+ },
+ {
+ "epoch": 14.97816593886463,
+ "grad_norm": 0.0211983360350132,
+ "learning_rate": 0.0006,
+ "loss": 4.821558475494385,
+ "step": 1078
+ },
+ {
+ "epoch": 14.992139737991266,
+ "grad_norm": 0.019636567682027817,
+ "learning_rate": 0.0006,
+ "loss": 4.578868865966797,
+ "step": 1079
+ },
+ {
+ "epoch": 15.0,
+ "grad_norm": 0.019997350871562958,
+ "learning_rate": 0.0006,
+ "loss": 4.691307544708252,
+ "step": 1080
+ },
+ {
+ "epoch": 15.0,
+ "eval_loss": 4.872478008270264,
+ "eval_runtime": 56.8513,
+ "eval_samples_per_second": 42.954,
+ "eval_steps_per_second": 1.354,
+ "step": 1080
+ },
+ {
+ "epoch": 15.013973799126637,
+ "grad_norm": 0.018830662593245506,
+ "learning_rate": 0.0006,
+ "loss": 4.761779308319092,
+ "step": 1081
+ },
+ {
+ "epoch": 15.027947598253276,
+ "grad_norm": 0.016876481473445892,
+ "learning_rate": 0.0006,
+ "loss": 4.61148738861084,
+ "step": 1082
+ },
+ {
+ "epoch": 15.041921397379912,
+ "grad_norm": 0.015550863929092884,
+ "learning_rate": 0.0006,
+ "loss": 4.676124095916748,
+ "step": 1083
+ },
+ {
+ "epoch": 15.055895196506551,
+ "grad_norm": 0.017178896814584732,
+ "learning_rate": 0.0006,
+ "loss": 4.588935852050781,
+ "step": 1084
+ },
+ {
+ "epoch": 15.069868995633188,
+ "grad_norm": 0.015465447679162025,
+ "learning_rate": 0.0006,
+ "loss": 4.605499267578125,
+ "step": 1085
+ },
+ {
+ "epoch": 15.083842794759825,
+ "grad_norm": 0.01606195978820324,
+ "learning_rate": 0.0006,
+ "loss": 4.723373889923096,
+ "step": 1086
+ },
+ {
+ "epoch": 15.097816593886463,
+ "grad_norm": 0.016521936282515526,
+ "learning_rate": 0.0006,
+ "loss": 4.646608352661133,
+ "step": 1087
+ },
+ {
+ "epoch": 15.1117903930131,
+ "grad_norm": 0.01589205674827099,
+ "learning_rate": 0.0006,
+ "loss": 4.498138427734375,
+ "step": 1088
+ },
+ {
+ "epoch": 15.125764192139737,
+ "grad_norm": 0.01524856686592102,
+ "learning_rate": 0.0006,
+ "loss": 4.594562530517578,
+ "step": 1089
+ },
+ {
+ "epoch": 15.139737991266376,
+ "grad_norm": 0.014714641496539116,
+ "learning_rate": 0.0006,
+ "loss": 4.630744934082031,
+ "step": 1090
+ },
+ {
+ "epoch": 15.153711790393013,
+ "grad_norm": 0.015333331190049648,
+ "learning_rate": 0.0006,
+ "loss": 4.682388782501221,
+ "step": 1091
+ },
+ {
+ "epoch": 15.167685589519651,
+ "grad_norm": 0.01550073828548193,
+ "learning_rate": 0.0006,
+ "loss": 4.692652702331543,
+ "step": 1092
+ },
+ {
+ "epoch": 15.181659388646288,
+ "grad_norm": 0.014783318154513836,
+ "learning_rate": 0.0006,
+ "loss": 4.6223835945129395,
+ "step": 1093
+ },
+ {
+ "epoch": 15.195633187772925,
+ "grad_norm": 0.014372751116752625,
+ "learning_rate": 0.0006,
+ "loss": 4.52687406539917,
+ "step": 1094
+ },
+ {
+ "epoch": 15.209606986899564,
+ "grad_norm": 0.014779130928218365,
+ "learning_rate": 0.0006,
+ "loss": 4.637009620666504,
+ "step": 1095
+ },
+ {
+ "epoch": 15.2235807860262,
+ "grad_norm": 0.015697073191404343,
+ "learning_rate": 0.0006,
+ "loss": 4.682981014251709,
+ "step": 1096
+ },
+ {
+ "epoch": 15.237554585152838,
+ "grad_norm": 0.016523664817214012,
+ "learning_rate": 0.0006,
+ "loss": 4.705722808837891,
+ "step": 1097
+ },
+ {
+ "epoch": 15.251528384279476,
+ "grad_norm": 0.017633996903896332,
+ "learning_rate": 0.0006,
+ "loss": 4.574686050415039,
+ "step": 1098
+ },
+ {
+ "epoch": 15.265502183406113,
+ "grad_norm": 0.019955076277256012,
+ "learning_rate": 0.0006,
+ "loss": 4.634450912475586,
+ "step": 1099
+ },
+ {
+ "epoch": 15.279475982532752,
+ "grad_norm": 0.02132618986070156,
+ "learning_rate": 0.0006,
+ "loss": 4.56110143661499,
+ "step": 1100
+ },
+ {
+ "epoch": 15.293449781659389,
+ "grad_norm": 0.0207593385130167,
+ "learning_rate": 0.0006,
+ "loss": 4.629396438598633,
+ "step": 1101
+ },
+ {
+ "epoch": 15.307423580786025,
+ "grad_norm": 0.017882540822029114,
+ "learning_rate": 0.0006,
+ "loss": 4.730963230133057,
+ "step": 1102
+ },
+ {
+ "epoch": 15.321397379912664,
+ "grad_norm": 0.01917876861989498,
+ "learning_rate": 0.0006,
+ "loss": 4.64585542678833,
+ "step": 1103
+ },
+ {
+ "epoch": 15.335371179039301,
+ "grad_norm": 0.01943974196910858,
+ "learning_rate": 0.0006,
+ "loss": 4.656181335449219,
+ "step": 1104
+ },
+ {
+ "epoch": 15.34934497816594,
+ "grad_norm": 0.019904715940356255,
+ "learning_rate": 0.0006,
+ "loss": 4.619329452514648,
+ "step": 1105
+ },
+ {
+ "epoch": 15.363318777292577,
+ "grad_norm": 0.019174346700310707,
+ "learning_rate": 0.0006,
+ "loss": 4.647892475128174,
+ "step": 1106
+ },
+ {
+ "epoch": 15.377292576419213,
+ "grad_norm": 0.018244758248329163,
+ "learning_rate": 0.0006,
+ "loss": 4.56166934967041,
+ "step": 1107
+ },
+ {
+ "epoch": 15.391266375545852,
+ "grad_norm": 0.018914205953478813,
+ "learning_rate": 0.0006,
+ "loss": 4.648063659667969,
+ "step": 1108
+ },
+ {
+ "epoch": 15.405240174672489,
+ "grad_norm": 0.01961071416735649,
+ "learning_rate": 0.0006,
+ "loss": 4.665525436401367,
+ "step": 1109
+ },
+ {
+ "epoch": 15.419213973799126,
+ "grad_norm": 0.01841311901807785,
+ "learning_rate": 0.0006,
+ "loss": 4.611342430114746,
+ "step": 1110
+ },
+ {
+ "epoch": 15.433187772925764,
+ "grad_norm": 0.01863427273929119,
+ "learning_rate": 0.0006,
+ "loss": 4.664256572723389,
+ "step": 1111
+ },
+ {
+ "epoch": 15.447161572052401,
+ "grad_norm": 0.020506877452135086,
+ "learning_rate": 0.0006,
+ "loss": 4.631237506866455,
+ "step": 1112
+ },
+ {
+ "epoch": 15.46113537117904,
+ "grad_norm": 0.02223833277821541,
+ "learning_rate": 0.0006,
+ "loss": 4.668654441833496,
+ "step": 1113
+ },
+ {
+ "epoch": 15.475109170305677,
+ "grad_norm": 0.023336702957749367,
+ "learning_rate": 0.0006,
+ "loss": 4.562833309173584,
+ "step": 1114
+ },
+ {
+ "epoch": 15.489082969432314,
+ "grad_norm": 0.023052413016557693,
+ "learning_rate": 0.0006,
+ "loss": 4.621532917022705,
+ "step": 1115
+ },
+ {
+ "epoch": 15.503056768558952,
+ "grad_norm": 0.023420924320816994,
+ "learning_rate": 0.0006,
+ "loss": 4.665465354919434,
+ "step": 1116
+ },
+ {
+ "epoch": 15.51703056768559,
+ "grad_norm": 0.024071281775832176,
+ "learning_rate": 0.0006,
+ "loss": 4.595980644226074,
+ "step": 1117
+ },
+ {
+ "epoch": 15.531004366812226,
+ "grad_norm": 0.023406682536005974,
+ "learning_rate": 0.0006,
+ "loss": 4.642635345458984,
+ "step": 1118
+ },
+ {
+ "epoch": 15.544978165938865,
+ "grad_norm": 0.02138647809624672,
+ "learning_rate": 0.0006,
+ "loss": 4.704513072967529,
+ "step": 1119
+ },
+ {
+ "epoch": 15.558951965065502,
+ "grad_norm": 0.0221384447067976,
+ "learning_rate": 0.0006,
+ "loss": 4.567503452301025,
+ "step": 1120
+ },
+ {
+ "epoch": 15.57292576419214,
+ "grad_norm": 0.020879851654171944,
+ "learning_rate": 0.0006,
+ "loss": 4.664057731628418,
+ "step": 1121
+ },
+ {
+ "epoch": 15.586899563318777,
+ "grad_norm": 0.021478967741131783,
+ "learning_rate": 0.0006,
+ "loss": 4.539627552032471,
+ "step": 1122
+ },
+ {
+ "epoch": 15.600873362445414,
+ "grad_norm": 0.022748656570911407,
+ "learning_rate": 0.0006,
+ "loss": 4.5654497146606445,
+ "step": 1123
+ },
+ {
+ "epoch": 15.614847161572053,
+ "grad_norm": 0.021870248019695282,
+ "learning_rate": 0.0006,
+ "loss": 4.56899356842041,
+ "step": 1124
+ },
+ {
+ "epoch": 15.62882096069869,
+ "grad_norm": 0.021711355075240135,
+ "learning_rate": 0.0006,
+ "loss": 4.657808303833008,
+ "step": 1125
+ },
+ {
+ "epoch": 15.642794759825328,
+ "grad_norm": 0.022810909897089005,
+ "learning_rate": 0.0006,
+ "loss": 4.6526689529418945,
+ "step": 1126
+ },
+ {
+ "epoch": 15.656768558951965,
+ "grad_norm": 0.022561343386769295,
+ "learning_rate": 0.0006,
+ "loss": 4.534887790679932,
+ "step": 1127
+ },
+ {
+ "epoch": 15.670742358078602,
+ "grad_norm": 0.019909802824258804,
+ "learning_rate": 0.0006,
+ "loss": 4.600790023803711,
+ "step": 1128
+ },
+ {
+ "epoch": 15.68471615720524,
+ "grad_norm": 0.020092325285077095,
+ "learning_rate": 0.0006,
+ "loss": 4.773830413818359,
+ "step": 1129
+ },
+ {
+ "epoch": 15.698689956331878,
+ "grad_norm": 0.018834445625543594,
+ "learning_rate": 0.0006,
+ "loss": 4.628868103027344,
+ "step": 1130
+ },
+ {
+ "epoch": 15.712663755458514,
+ "grad_norm": 0.017552798613905907,
+ "learning_rate": 0.0006,
+ "loss": 4.647500038146973,
+ "step": 1131
+ },
+ {
+ "epoch": 15.726637554585153,
+ "grad_norm": 0.01751251146197319,
+ "learning_rate": 0.0006,
+ "loss": 4.590545654296875,
+ "step": 1132
+ },
+ {
+ "epoch": 15.74061135371179,
+ "grad_norm": 0.01699099875986576,
+ "learning_rate": 0.0006,
+ "loss": 4.509699821472168,
+ "step": 1133
+ },
+ {
+ "epoch": 15.754585152838429,
+ "grad_norm": 0.015808649361133575,
+ "learning_rate": 0.0006,
+ "loss": 4.606992244720459,
+ "step": 1134
+ },
+ {
+ "epoch": 15.768558951965066,
+ "grad_norm": 0.01655443385243416,
+ "learning_rate": 0.0006,
+ "loss": 4.585219860076904,
+ "step": 1135
+ },
+ {
+ "epoch": 15.782532751091702,
+ "grad_norm": 0.01708163693547249,
+ "learning_rate": 0.0006,
+ "loss": 4.668147087097168,
+ "step": 1136
+ },
+ {
+ "epoch": 15.796506550218341,
+ "grad_norm": 0.017161913216114044,
+ "learning_rate": 0.0006,
+ "loss": 4.7170634269714355,
+ "step": 1137
+ },
+ {
+ "epoch": 15.810480349344978,
+ "grad_norm": 0.017262674868106842,
+ "learning_rate": 0.0006,
+ "loss": 4.572132110595703,
+ "step": 1138
+ },
+ {
+ "epoch": 15.824454148471617,
+ "grad_norm": 0.017713576555252075,
+ "learning_rate": 0.0006,
+ "loss": 4.702538967132568,
+ "step": 1139
+ },
+ {
+ "epoch": 15.838427947598253,
+ "grad_norm": 0.018730733543634415,
+ "learning_rate": 0.0006,
+ "loss": 4.629721641540527,
+ "step": 1140
+ },
+ {
+ "epoch": 15.85240174672489,
+ "grad_norm": 0.017960187047719955,
+ "learning_rate": 0.0006,
+ "loss": 4.661039352416992,
+ "step": 1141
+ },
+ {
+ "epoch": 15.866375545851529,
+ "grad_norm": 0.01750079356133938,
+ "learning_rate": 0.0006,
+ "loss": 4.503710746765137,
+ "step": 1142
+ },
+ {
+ "epoch": 15.880349344978166,
+ "grad_norm": 0.016630422323942184,
+ "learning_rate": 0.0006,
+ "loss": 4.6582136154174805,
+ "step": 1143
+ },
+ {
+ "epoch": 15.894323144104803,
+ "grad_norm": 0.015896832570433617,
+ "learning_rate": 0.0006,
+ "loss": 4.596693992614746,
+ "step": 1144
+ },
+ {
+ "epoch": 15.908296943231441,
+ "grad_norm": 0.01566407084465027,
+ "learning_rate": 0.0006,
+ "loss": 4.6337199211120605,
+ "step": 1145
+ },
+ {
+ "epoch": 15.922270742358078,
+ "grad_norm": 0.016823330894112587,
+ "learning_rate": 0.0006,
+ "loss": 4.686994552612305,
+ "step": 1146
+ },
+ {
+ "epoch": 15.936244541484717,
+ "grad_norm": 0.016208255663514137,
+ "learning_rate": 0.0006,
+ "loss": 4.6205549240112305,
+ "step": 1147
+ },
+ {
+ "epoch": 15.950218340611354,
+ "grad_norm": 0.0171707421541214,
+ "learning_rate": 0.0006,
+ "loss": 4.582968711853027,
+ "step": 1148
+ },
+ {
+ "epoch": 15.96419213973799,
+ "grad_norm": 0.017009446397423744,
+ "learning_rate": 0.0006,
+ "loss": 4.510274887084961,
+ "step": 1149
+ },
+ {
+ "epoch": 15.97816593886463,
+ "grad_norm": 0.01565811224281788,
+ "learning_rate": 0.0006,
+ "loss": 4.55834436416626,
+ "step": 1150
+ },
+ {
+ "epoch": 15.992139737991266,
+ "grad_norm": 0.015276037156581879,
+ "learning_rate": 0.0006,
+ "loss": 4.537870407104492,
+ "step": 1151
+ },
+ {
+ "epoch": 16.0,
+ "grad_norm": 0.017305882647633553,
+ "learning_rate": 0.0006,
+ "loss": 4.611305236816406,
+ "step": 1152
+ },
+ {
+ "epoch": 16.0,
+ "eval_loss": 4.775513172149658,
+ "eval_runtime": 56.7341,
+ "eval_samples_per_second": 43.043,
+ "eval_steps_per_second": 1.357,
+ "step": 1152
+ },
+ {
+ "epoch": 16.01397379912664,
+ "grad_norm": 0.016752436757087708,
+ "learning_rate": 0.0006,
+ "loss": 4.541312217712402,
+ "step": 1153
+ },
+ {
+ "epoch": 16.027947598253274,
+ "grad_norm": 0.01643134094774723,
+ "learning_rate": 0.0006,
+ "loss": 4.585402965545654,
+ "step": 1154
+ },
+ {
+ "epoch": 16.041921397379912,
+ "grad_norm": 0.01888720504939556,
+ "learning_rate": 0.0006,
+ "loss": 4.597825050354004,
+ "step": 1155
+ },
+ {
+ "epoch": 16.05589519650655,
+ "grad_norm": 0.022044632583856583,
+ "learning_rate": 0.0006,
+ "loss": 4.532772064208984,
+ "step": 1156
+ },
+ {
+ "epoch": 16.069868995633186,
+ "grad_norm": 0.022986987605690956,
+ "learning_rate": 0.0006,
+ "loss": 4.551351547241211,
+ "step": 1157
+ },
+ {
+ "epoch": 16.083842794759825,
+ "grad_norm": 0.020116791129112244,
+ "learning_rate": 0.0006,
+ "loss": 4.484149932861328,
+ "step": 1158
+ },
+ {
+ "epoch": 16.097816593886463,
+ "grad_norm": 0.021442247554659843,
+ "learning_rate": 0.0006,
+ "loss": 4.515316486358643,
+ "step": 1159
+ },
+ {
+ "epoch": 16.111790393013102,
+ "grad_norm": 0.020974772050976753,
+ "learning_rate": 0.0006,
+ "loss": 4.51606559753418,
+ "step": 1160
+ },
+ {
+ "epoch": 16.125764192139737,
+ "grad_norm": 0.021274864673614502,
+ "learning_rate": 0.0006,
+ "loss": 4.587329864501953,
+ "step": 1161
+ },
+ {
+ "epoch": 16.139737991266376,
+ "grad_norm": 0.022799784317612648,
+ "learning_rate": 0.0006,
+ "loss": 4.568149089813232,
+ "step": 1162
+ },
+ {
+ "epoch": 16.153711790393015,
+ "grad_norm": 0.021784933283925056,
+ "learning_rate": 0.0006,
+ "loss": 4.515336990356445,
+ "step": 1163
+ },
+ {
+ "epoch": 16.16768558951965,
+ "grad_norm": 0.024261271581053734,
+ "learning_rate": 0.0006,
+ "loss": 4.5127081871032715,
+ "step": 1164
+ },
+ {
+ "epoch": 16.18165938864629,
+ "grad_norm": 0.02248375490307808,
+ "learning_rate": 0.0006,
+ "loss": 4.754251003265381,
+ "step": 1165
+ },
+ {
+ "epoch": 16.195633187772927,
+ "grad_norm": 0.02136492356657982,
+ "learning_rate": 0.0006,
+ "loss": 4.58193826675415,
+ "step": 1166
+ },
+ {
+ "epoch": 16.209606986899562,
+ "grad_norm": 0.020727066323161125,
+ "learning_rate": 0.0006,
+ "loss": 4.564209938049316,
+ "step": 1167
+ },
+ {
+ "epoch": 16.2235807860262,
+ "grad_norm": 0.017789119854569435,
+ "learning_rate": 0.0006,
+ "loss": 4.698680877685547,
+ "step": 1168
+ },
+ {
+ "epoch": 16.23755458515284,
+ "grad_norm": 0.017603034153580666,
+ "learning_rate": 0.0006,
+ "loss": 4.501171588897705,
+ "step": 1169
+ },
+ {
+ "epoch": 16.251528384279474,
+ "grad_norm": 0.017345862463116646,
+ "learning_rate": 0.0006,
+ "loss": 4.593590259552002,
+ "step": 1170
+ },
+ {
+ "epoch": 16.265502183406113,
+ "grad_norm": 0.018047425895929337,
+ "learning_rate": 0.0006,
+ "loss": 4.597467422485352,
+ "step": 1171
+ },
+ {
+ "epoch": 16.27947598253275,
+ "grad_norm": 0.01947222836315632,
+ "learning_rate": 0.0006,
+ "loss": 4.5072221755981445,
+ "step": 1172
+ },
+ {
+ "epoch": 16.29344978165939,
+ "grad_norm": 0.019598500803112984,
+ "learning_rate": 0.0006,
+ "loss": 4.601351737976074,
+ "step": 1173
+ },
+ {
+ "epoch": 16.307423580786025,
+ "grad_norm": 0.020002514123916626,
+ "learning_rate": 0.0006,
+ "loss": 4.580431938171387,
+ "step": 1174
+ },
+ {
+ "epoch": 16.321397379912664,
+ "grad_norm": 0.0216530729085207,
+ "learning_rate": 0.0006,
+ "loss": 4.424681663513184,
+ "step": 1175
+ },
+ {
+ "epoch": 16.335371179039303,
+ "grad_norm": 0.01949235238134861,
+ "learning_rate": 0.0006,
+ "loss": 4.751883029937744,
+ "step": 1176
+ },
+ {
+ "epoch": 16.349344978165938,
+ "grad_norm": 0.017779016867280006,
+ "learning_rate": 0.0006,
+ "loss": 4.541490077972412,
+ "step": 1177
+ },
+ {
+ "epoch": 16.363318777292577,
+ "grad_norm": 0.01909121684730053,
+ "learning_rate": 0.0006,
+ "loss": 4.594141006469727,
+ "step": 1178
+ },
+ {
+ "epoch": 16.377292576419215,
+ "grad_norm": 0.019894849509000778,
+ "learning_rate": 0.0006,
+ "loss": 4.555070400238037,
+ "step": 1179
+ },
+ {
+ "epoch": 16.39126637554585,
+ "grad_norm": 0.01993032731115818,
+ "learning_rate": 0.0006,
+ "loss": 4.521401405334473,
+ "step": 1180
+ },
+ {
+ "epoch": 16.40524017467249,
+ "grad_norm": 0.0210788045078516,
+ "learning_rate": 0.0006,
+ "loss": 4.672842979431152,
+ "step": 1181
+ },
+ {
+ "epoch": 16.419213973799128,
+ "grad_norm": 0.02274598926305771,
+ "learning_rate": 0.0006,
+ "loss": 4.629504203796387,
+ "step": 1182
+ },
+ {
+ "epoch": 16.433187772925763,
+ "grad_norm": 0.020448120310902596,
+ "learning_rate": 0.0006,
+ "loss": 4.548209190368652,
+ "step": 1183
+ },
+ {
+ "epoch": 16.4471615720524,
+ "grad_norm": 0.017565643414855003,
+ "learning_rate": 0.0006,
+ "loss": 4.529054641723633,
+ "step": 1184
+ },
+ {
+ "epoch": 16.46113537117904,
+ "grad_norm": 0.019504237920045853,
+ "learning_rate": 0.0006,
+ "loss": 4.48182487487793,
+ "step": 1185
+ },
+ {
+ "epoch": 16.475109170305675,
+ "grad_norm": 0.016823742538690567,
+ "learning_rate": 0.0006,
+ "loss": 4.608696937561035,
+ "step": 1186
+ },
+ {
+ "epoch": 16.489082969432314,
+ "grad_norm": 0.01602894812822342,
+ "learning_rate": 0.0006,
+ "loss": 4.620062351226807,
+ "step": 1187
+ },
+ {
+ "epoch": 16.503056768558952,
+ "grad_norm": 0.016615642234683037,
+ "learning_rate": 0.0006,
+ "loss": 4.69378662109375,
+ "step": 1188
+ },
+ {
+ "epoch": 16.51703056768559,
+ "grad_norm": 0.018576730042696,
+ "learning_rate": 0.0006,
+ "loss": 4.537041664123535,
+ "step": 1189
+ },
+ {
+ "epoch": 16.531004366812226,
+ "grad_norm": 0.018611151725053787,
+ "learning_rate": 0.0006,
+ "loss": 4.496169090270996,
+ "step": 1190
+ },
+ {
+ "epoch": 16.544978165938865,
+ "grad_norm": 0.018622413277626038,
+ "learning_rate": 0.0006,
+ "loss": 4.557358264923096,
+ "step": 1191
+ },
+ {
+ "epoch": 16.558951965065503,
+ "grad_norm": 0.01823544315993786,
+ "learning_rate": 0.0006,
+ "loss": 4.730792999267578,
+ "step": 1192
+ },
+ {
+ "epoch": 16.57292576419214,
+ "grad_norm": 0.018796846270561218,
+ "learning_rate": 0.0006,
+ "loss": 4.781791687011719,
+ "step": 1193
+ },
+ {
+ "epoch": 16.586899563318777,
+ "grad_norm": 0.01728292554616928,
+ "learning_rate": 0.0006,
+ "loss": 4.515257835388184,
+ "step": 1194
+ },
+ {
+ "epoch": 16.600873362445416,
+ "grad_norm": 0.01699589006602764,
+ "learning_rate": 0.0006,
+ "loss": 4.5996904373168945,
+ "step": 1195
+ },
+ {
+ "epoch": 16.61484716157205,
+ "grad_norm": 0.015905970707535744,
+ "learning_rate": 0.0006,
+ "loss": 4.601301193237305,
+ "step": 1196
+ },
+ {
+ "epoch": 16.62882096069869,
+ "grad_norm": 0.01338571310043335,
+ "learning_rate": 0.0006,
+ "loss": 4.557708740234375,
+ "step": 1197
+ },
+ {
+ "epoch": 16.64279475982533,
+ "grad_norm": 0.015298855490982533,
+ "learning_rate": 0.0006,
+ "loss": 4.542287349700928,
+ "step": 1198
+ },
+ {
+ "epoch": 16.656768558951963,
+ "grad_norm": 0.014753367751836777,
+ "learning_rate": 0.0006,
+ "loss": 4.595787048339844,
+ "step": 1199
+ },
+ {
+ "epoch": 16.670742358078602,
+ "grad_norm": 0.015527483075857162,
+ "learning_rate": 0.0006,
+ "loss": 4.607529163360596,
+ "step": 1200
+ },
+ {
+ "epoch": 16.68471615720524,
+ "grad_norm": 0.015042162500321865,
+ "learning_rate": 0.0006,
+ "loss": 4.686626434326172,
+ "step": 1201
+ },
+ {
+ "epoch": 16.69868995633188,
+ "grad_norm": 0.0159393772482872,
+ "learning_rate": 0.0006,
+ "loss": 4.47656774520874,
+ "step": 1202
+ },
+ {
+ "epoch": 16.712663755458514,
+ "grad_norm": 0.01532787922769785,
+ "learning_rate": 0.0006,
+ "loss": 4.55678653717041,
+ "step": 1203
+ },
+ {
+ "epoch": 16.726637554585153,
+ "grad_norm": 0.016332164406776428,
+ "learning_rate": 0.0006,
+ "loss": 4.51441764831543,
+ "step": 1204
+ },
+ {
+ "epoch": 16.74061135371179,
+ "grad_norm": 0.018901726230978966,
+ "learning_rate": 0.0006,
+ "loss": 4.57227897644043,
+ "step": 1205
+ },
+ {
+ "epoch": 16.754585152838427,
+ "grad_norm": 0.018589433282613754,
+ "learning_rate": 0.0006,
+ "loss": 4.550070762634277,
+ "step": 1206
+ },
+ {
+ "epoch": 16.768558951965066,
+ "grad_norm": 0.01634034886956215,
+ "learning_rate": 0.0006,
+ "loss": 4.586910724639893,
+ "step": 1207
+ },
+ {
+ "epoch": 16.782532751091704,
+ "grad_norm": 0.017599433660507202,
+ "learning_rate": 0.0006,
+ "loss": 4.5891923904418945,
+ "step": 1208
+ },
+ {
+ "epoch": 16.79650655021834,
+ "grad_norm": 0.01878446340560913,
+ "learning_rate": 0.0006,
+ "loss": 4.57227087020874,
+ "step": 1209
+ },
+ {
+ "epoch": 16.810480349344978,
+ "grad_norm": 0.01803469844162464,
+ "learning_rate": 0.0006,
+ "loss": 4.583845138549805,
+ "step": 1210
+ },
+ {
+ "epoch": 16.824454148471617,
+ "grad_norm": 0.018505144864320755,
+ "learning_rate": 0.0006,
+ "loss": 4.561151504516602,
+ "step": 1211
+ },
+ {
+ "epoch": 16.83842794759825,
+ "grad_norm": 0.01920081488788128,
+ "learning_rate": 0.0006,
+ "loss": 4.506834030151367,
+ "step": 1212
+ },
+ {
+ "epoch": 16.85240174672489,
+ "grad_norm": 0.0177712831646204,
+ "learning_rate": 0.0006,
+ "loss": 4.557195663452148,
+ "step": 1213
+ },
+ {
+ "epoch": 16.86637554585153,
+ "grad_norm": 0.019113918766379356,
+ "learning_rate": 0.0006,
+ "loss": 4.651227951049805,
+ "step": 1214
+ },
+ {
+ "epoch": 16.880349344978168,
+ "grad_norm": 0.02012539468705654,
+ "learning_rate": 0.0006,
+ "loss": 4.4952239990234375,
+ "step": 1215
+ },
+ {
+ "epoch": 16.894323144104803,
+ "grad_norm": 0.019993646070361137,
+ "learning_rate": 0.0006,
+ "loss": 4.517433166503906,
+ "step": 1216
+ },
+ {
+ "epoch": 16.90829694323144,
+ "grad_norm": 0.020041441544890404,
+ "learning_rate": 0.0006,
+ "loss": 4.489407539367676,
+ "step": 1217
+ },
+ {
+ "epoch": 16.92227074235808,
+ "grad_norm": 0.0183611661195755,
+ "learning_rate": 0.0006,
+ "loss": 4.5435638427734375,
+ "step": 1218
+ },
+ {
+ "epoch": 16.936244541484715,
+ "grad_norm": 0.019138429313898087,
+ "learning_rate": 0.0006,
+ "loss": 4.545949459075928,
+ "step": 1219
+ },
+ {
+ "epoch": 16.950218340611354,
+ "grad_norm": 0.018554266542196274,
+ "learning_rate": 0.0006,
+ "loss": 4.579620361328125,
+ "step": 1220
+ },
+ {
+ "epoch": 16.964192139737992,
+ "grad_norm": 0.0193545650690794,
+ "learning_rate": 0.0006,
+ "loss": 4.588255405426025,
+ "step": 1221
+ },
+ {
+ "epoch": 16.978165938864628,
+ "grad_norm": 0.019389165565371513,
+ "learning_rate": 0.0006,
+ "loss": 4.520698547363281,
+ "step": 1222
+ },
+ {
+ "epoch": 16.992139737991266,
+ "grad_norm": 0.016601379960775375,
+ "learning_rate": 0.0006,
+ "loss": 4.4809064865112305,
+ "step": 1223
+ },
+ {
+ "epoch": 17.0,
+ "grad_norm": 0.01771303080022335,
+ "learning_rate": 0.0006,
+ "loss": 4.721122741699219,
+ "step": 1224
+ },
+ {
+ "epoch": 17.0,
+ "eval_loss": 4.7903666496276855,
+ "eval_runtime": 56.9826,
+ "eval_samples_per_second": 42.855,
+ "eval_steps_per_second": 1.351,
+ "step": 1224
+ },
+ {
+ "epoch": 17.01397379912664,
+ "grad_norm": 0.017720289528369904,
+ "learning_rate": 0.0006,
+ "loss": 4.523834705352783,
+ "step": 1225
+ },
+ {
+ "epoch": 17.027947598253274,
+ "grad_norm": 0.020137161016464233,
+ "learning_rate": 0.0006,
+ "loss": 4.574061870574951,
+ "step": 1226
+ },
+ {
+ "epoch": 17.041921397379912,
+ "grad_norm": 0.02002314105629921,
+ "learning_rate": 0.0006,
+ "loss": 4.512407302856445,
+ "step": 1227
+ },
+ {
+ "epoch": 17.05589519650655,
+ "grad_norm": 0.019628183916211128,
+ "learning_rate": 0.0006,
+ "loss": 4.502533435821533,
+ "step": 1228
+ },
+ {
+ "epoch": 17.069868995633186,
+ "grad_norm": 0.019808458164334297,
+ "learning_rate": 0.0006,
+ "loss": 4.463156700134277,
+ "step": 1229
+ },
+ {
+ "epoch": 17.083842794759825,
+ "grad_norm": 0.021250786259770393,
+ "learning_rate": 0.0006,
+ "loss": 4.622682094573975,
+ "step": 1230
+ },
+ {
+ "epoch": 17.097816593886463,
+ "grad_norm": 0.022031694650650024,
+ "learning_rate": 0.0006,
+ "loss": 4.451882839202881,
+ "step": 1231
+ },
+ {
+ "epoch": 17.111790393013102,
+ "grad_norm": 0.01898987777531147,
+ "learning_rate": 0.0006,
+ "loss": 4.529331684112549,
+ "step": 1232
+ },
+ {
+ "epoch": 17.125764192139737,
+ "grad_norm": 0.018749834969639778,
+ "learning_rate": 0.0006,
+ "loss": 4.435815334320068,
+ "step": 1233
+ },
+ {
+ "epoch": 17.139737991266376,
+ "grad_norm": 0.01896674931049347,
+ "learning_rate": 0.0006,
+ "loss": 4.484029293060303,
+ "step": 1234
+ },
+ {
+ "epoch": 17.153711790393015,
+ "grad_norm": 0.018286997452378273,
+ "learning_rate": 0.0006,
+ "loss": 4.523510932922363,
+ "step": 1235
+ },
+ {
+ "epoch": 17.16768558951965,
+ "grad_norm": 0.017273059114813805,
+ "learning_rate": 0.0006,
+ "loss": 4.451767444610596,
+ "step": 1236
+ },
+ {
+ "epoch": 17.18165938864629,
+ "grad_norm": 0.016499150544404984,
+ "learning_rate": 0.0006,
+ "loss": 4.5568647384643555,
+ "step": 1237
+ },
+ {
+ "epoch": 17.195633187772927,
+ "grad_norm": 0.015934469178318977,
+ "learning_rate": 0.0006,
+ "loss": 4.4639997482299805,
+ "step": 1238
+ },
+ {
+ "epoch": 17.209606986899562,
+ "grad_norm": 0.01714209094643593,
+ "learning_rate": 0.0006,
+ "loss": 4.547614097595215,
+ "step": 1239
+ },
+ {
+ "epoch": 17.2235807860262,
+ "grad_norm": 0.019167495891451836,
+ "learning_rate": 0.0006,
+ "loss": 4.571097373962402,
+ "step": 1240
+ },
+ {
+ "epoch": 17.23755458515284,
+ "grad_norm": 0.018276996910572052,
+ "learning_rate": 0.0006,
+ "loss": 4.489764213562012,
+ "step": 1241
+ },
+ {
+ "epoch": 17.251528384279474,
+ "grad_norm": 0.021296866238117218,
+ "learning_rate": 0.0006,
+ "loss": 4.526913642883301,
+ "step": 1242
+ },
+ {
+ "epoch": 17.265502183406113,
+ "grad_norm": 0.022606629878282547,
+ "learning_rate": 0.0006,
+ "loss": 4.521624565124512,
+ "step": 1243
+ },
+ {
+ "epoch": 17.27947598253275,
+ "grad_norm": 0.021204711869359016,
+ "learning_rate": 0.0006,
+ "loss": 4.500864028930664,
+ "step": 1244
+ },
+ {
+ "epoch": 17.29344978165939,
+ "grad_norm": 0.01988849975168705,
+ "learning_rate": 0.0006,
+ "loss": 4.45162296295166,
+ "step": 1245
+ },
+ {
+ "epoch": 17.307423580786025,
+ "grad_norm": 0.020268384367227554,
+ "learning_rate": 0.0006,
+ "loss": 4.536198616027832,
+ "step": 1246
+ },
+ {
+ "epoch": 17.321397379912664,
+ "grad_norm": 0.020185677334666252,
+ "learning_rate": 0.0006,
+ "loss": 4.440262794494629,
+ "step": 1247
+ },
+ {
+ "epoch": 17.335371179039303,
+ "grad_norm": 0.020106853917241096,
+ "learning_rate": 0.0006,
+ "loss": 4.475907325744629,
+ "step": 1248
+ },
+ {
+ "epoch": 17.349344978165938,
+ "grad_norm": 0.018883297219872475,
+ "learning_rate": 0.0006,
+ "loss": 4.640661716461182,
+ "step": 1249
+ },
+ {
+ "epoch": 17.363318777292577,
+ "grad_norm": 0.01852620206773281,
+ "learning_rate": 0.0006,
+ "loss": 4.578243255615234,
+ "step": 1250
+ },
+ {
+ "epoch": 17.377292576419215,
+ "grad_norm": 0.01886451430618763,
+ "learning_rate": 0.0006,
+ "loss": 4.546026229858398,
+ "step": 1251
+ },
+ {
+ "epoch": 17.39126637554585,
+ "grad_norm": 0.01777452416718006,
+ "learning_rate": 0.0006,
+ "loss": 4.521542549133301,
+ "step": 1252
+ },
+ {
+ "epoch": 17.40524017467249,
+ "grad_norm": 0.016626447439193726,
+ "learning_rate": 0.0006,
+ "loss": 4.52629280090332,
+ "step": 1253
+ },
+ {
+ "epoch": 17.419213973799128,
+ "grad_norm": 0.01599477231502533,
+ "learning_rate": 0.0006,
+ "loss": 4.4796037673950195,
+ "step": 1254
+ },
+ {
+ "epoch": 17.433187772925763,
+ "grad_norm": 0.017021458595991135,
+ "learning_rate": 0.0006,
+ "loss": 4.468813419342041,
+ "step": 1255
+ },
+ {
+ "epoch": 17.4471615720524,
+ "grad_norm": 0.016925333067774773,
+ "learning_rate": 0.0006,
+ "loss": 4.584898948669434,
+ "step": 1256
+ },
+ {
+ "epoch": 17.46113537117904,
+ "grad_norm": 0.01714998669922352,
+ "learning_rate": 0.0006,
+ "loss": 4.523544788360596,
+ "step": 1257
+ },
+ {
+ "epoch": 17.475109170305675,
+ "grad_norm": 0.01802288182079792,
+ "learning_rate": 0.0006,
+ "loss": 4.595581531524658,
+ "step": 1258
+ },
+ {
+ "epoch": 17.489082969432314,
+ "grad_norm": 0.0180289875715971,
+ "learning_rate": 0.0006,
+ "loss": 4.394440650939941,
+ "step": 1259
+ },
+ {
+ "epoch": 17.503056768558952,
+ "grad_norm": 0.018504245206713676,
+ "learning_rate": 0.0006,
+ "loss": 4.4794416427612305,
+ "step": 1260
+ },
+ {
+ "epoch": 17.51703056768559,
+ "grad_norm": 0.019973233342170715,
+ "learning_rate": 0.0006,
+ "loss": 4.4241437911987305,
+ "step": 1261
+ },
+ {
+ "epoch": 17.531004366812226,
+ "grad_norm": 0.020063083618879318,
+ "learning_rate": 0.0006,
+ "loss": 4.48068380355835,
+ "step": 1262
+ },
+ {
+ "epoch": 17.544978165938865,
+ "grad_norm": 0.02066640742123127,
+ "learning_rate": 0.0006,
+ "loss": 4.628993034362793,
+ "step": 1263
+ },
+ {
+ "epoch": 17.558951965065503,
+ "grad_norm": 0.02206416428089142,
+ "learning_rate": 0.0006,
+ "loss": 4.564790725708008,
+ "step": 1264
+ },
+ {
+ "epoch": 17.57292576419214,
+ "grad_norm": 0.020640335977077484,
+ "learning_rate": 0.0006,
+ "loss": 4.578705787658691,
+ "step": 1265
+ },
+ {
+ "epoch": 17.586899563318777,
+ "grad_norm": 0.019093405455350876,
+ "learning_rate": 0.0006,
+ "loss": 4.5023603439331055,
+ "step": 1266
+ },
+ {
+ "epoch": 17.600873362445416,
+ "grad_norm": 0.019459193572402,
+ "learning_rate": 0.0006,
+ "loss": 4.594531059265137,
+ "step": 1267
+ },
+ {
+ "epoch": 17.61484716157205,
+ "grad_norm": 0.02021171525120735,
+ "learning_rate": 0.0006,
+ "loss": 4.770299911499023,
+ "step": 1268
+ },
+ {
+ "epoch": 17.62882096069869,
+ "grad_norm": 0.021161191165447235,
+ "learning_rate": 0.0006,
+ "loss": 4.5380401611328125,
+ "step": 1269
+ },
+ {
+ "epoch": 17.64279475982533,
+ "grad_norm": 0.021013403311371803,
+ "learning_rate": 0.0006,
+ "loss": 4.455524921417236,
+ "step": 1270
+ },
+ {
+ "epoch": 17.656768558951963,
+ "grad_norm": 0.020336592569947243,
+ "learning_rate": 0.0006,
+ "loss": 4.533672332763672,
+ "step": 1271
+ },
+ {
+ "epoch": 17.670742358078602,
+ "grad_norm": 0.020159434527158737,
+ "learning_rate": 0.0006,
+ "loss": 4.596151351928711,
+ "step": 1272
+ },
+ {
+ "epoch": 17.68471615720524,
+ "grad_norm": 0.020432960242033005,
+ "learning_rate": 0.0006,
+ "loss": 4.55531644821167,
+ "step": 1273
+ },
+ {
+ "epoch": 17.69868995633188,
+ "grad_norm": 0.025466229766607285,
+ "learning_rate": 0.0006,
+ "loss": 4.566912651062012,
+ "step": 1274
+ },
+ {
+ "epoch": 17.712663755458514,
+ "grad_norm": 0.026537558063864708,
+ "learning_rate": 0.0006,
+ "loss": 4.526003837585449,
+ "step": 1275
+ },
+ {
+ "epoch": 17.726637554585153,
+ "grad_norm": 0.024397607892751694,
+ "learning_rate": 0.0006,
+ "loss": 4.541755676269531,
+ "step": 1276
+ },
+ {
+ "epoch": 17.74061135371179,
+ "grad_norm": 0.022062312811613083,
+ "learning_rate": 0.0006,
+ "loss": 4.6532793045043945,
+ "step": 1277
+ },
+ {
+ "epoch": 17.754585152838427,
+ "grad_norm": 0.02204282023012638,
+ "learning_rate": 0.0006,
+ "loss": 4.582825660705566,
+ "step": 1278
+ },
+ {
+ "epoch": 17.768558951965066,
+ "grad_norm": 0.020844416692852974,
+ "learning_rate": 0.0006,
+ "loss": 4.522336483001709,
+ "step": 1279
+ },
+ {
+ "epoch": 17.782532751091704,
+ "grad_norm": 0.019815703853964806,
+ "learning_rate": 0.0006,
+ "loss": 4.607398509979248,
+ "step": 1280
+ },
+ {
+ "epoch": 17.79650655021834,
+ "grad_norm": 0.016676954925060272,
+ "learning_rate": 0.0006,
+ "loss": 4.446206092834473,
+ "step": 1281
+ },
+ {
+ "epoch": 17.810480349344978,
+ "grad_norm": 0.0166452769190073,
+ "learning_rate": 0.0006,
+ "loss": 4.505931854248047,
+ "step": 1282
+ },
+ {
+ "epoch": 17.824454148471617,
+ "grad_norm": 0.015928907319903374,
+ "learning_rate": 0.0006,
+ "loss": 4.455267429351807,
+ "step": 1283
+ },
+ {
+ "epoch": 17.83842794759825,
+ "grad_norm": 0.015219560824334621,
+ "learning_rate": 0.0006,
+ "loss": 4.5623674392700195,
+ "step": 1284
+ },
+ {
+ "epoch": 17.85240174672489,
+ "grad_norm": 0.015125355683267117,
+ "learning_rate": 0.0006,
+ "loss": 4.554104804992676,
+ "step": 1285
+ },
+ {
+ "epoch": 17.86637554585153,
+ "grad_norm": 0.015612718649208546,
+ "learning_rate": 0.0006,
+ "loss": 4.484847068786621,
+ "step": 1286
+ },
+ {
+ "epoch": 17.880349344978168,
+ "grad_norm": 0.015215662308037281,
+ "learning_rate": 0.0006,
+ "loss": 4.4802093505859375,
+ "step": 1287
+ },
+ {
+ "epoch": 17.894323144104803,
+ "grad_norm": 0.01511684525758028,
+ "learning_rate": 0.0006,
+ "loss": 4.537298679351807,
+ "step": 1288
+ },
+ {
+ "epoch": 17.90829694323144,
+ "grad_norm": 0.013841322623193264,
+ "learning_rate": 0.0006,
+ "loss": 4.427488327026367,
+ "step": 1289
+ },
+ {
+ "epoch": 17.92227074235808,
+ "grad_norm": 0.015115964226424694,
+ "learning_rate": 0.0006,
+ "loss": 4.4547119140625,
+ "step": 1290
+ },
+ {
+ "epoch": 17.936244541484715,
+ "grad_norm": 0.016411006450653076,
+ "learning_rate": 0.0006,
+ "loss": 4.52994966506958,
+ "step": 1291
+ },
+ {
+ "epoch": 17.950218340611354,
+ "grad_norm": 0.017313152551651,
+ "learning_rate": 0.0006,
+ "loss": 4.5337653160095215,
+ "step": 1292
+ },
+ {
+ "epoch": 17.964192139737992,
+ "grad_norm": 0.016559796407818794,
+ "learning_rate": 0.0006,
+ "loss": 4.5277910232543945,
+ "step": 1293
+ },
+ {
+ "epoch": 17.978165938864628,
+ "grad_norm": 0.015854761004447937,
+ "learning_rate": 0.0006,
+ "loss": 4.6334309577941895,
+ "step": 1294
+ },
+ {
+ "epoch": 17.992139737991266,
+ "grad_norm": 0.015658238902688026,
+ "learning_rate": 0.0006,
+ "loss": 4.613372802734375,
+ "step": 1295
+ },
+ {
+ "epoch": 18.0,
+ "grad_norm": 0.0169155802577734,
+ "learning_rate": 0.0006,
+ "loss": 4.527356147766113,
+ "step": 1296
+ },
+ {
+ "epoch": 18.0,
+ "eval_loss": 4.764543056488037,
+ "eval_runtime": 56.5998,
+ "eval_samples_per_second": 43.145,
+ "eval_steps_per_second": 1.36,
+ "step": 1296
+ },
+ {
+ "epoch": 18.01397379912664,
+ "grad_norm": 0.015348963439464569,
+ "learning_rate": 0.0006,
+ "loss": 4.466423988342285,
+ "step": 1297
+ },
+ {
+ "epoch": 18.027947598253274,
+ "grad_norm": 0.01768588088452816,
+ "learning_rate": 0.0006,
+ "loss": 4.545933246612549,
+ "step": 1298
+ },
+ {
+ "epoch": 18.041921397379912,
+ "grad_norm": 0.017381226643919945,
+ "learning_rate": 0.0006,
+ "loss": 4.494060039520264,
+ "step": 1299
+ },
+ {
+ "epoch": 18.05589519650655,
+ "grad_norm": 0.019002335146069527,
+ "learning_rate": 0.0006,
+ "loss": 4.395200729370117,
+ "step": 1300
+ },
+ {
+ "epoch": 18.069868995633186,
+ "grad_norm": 0.01827932707965374,
+ "learning_rate": 0.0006,
+ "loss": 4.424848556518555,
+ "step": 1301
+ },
+ {
+ "epoch": 18.083842794759825,
+ "grad_norm": 0.016383031383156776,
+ "learning_rate": 0.0006,
+ "loss": 4.494722366333008,
+ "step": 1302
+ },
+ {
+ "epoch": 18.097816593886463,
+ "grad_norm": 0.016265157610177994,
+ "learning_rate": 0.0006,
+ "loss": 4.416522979736328,
+ "step": 1303
+ },
+ {
+ "epoch": 18.111790393013102,
+ "grad_norm": 0.01702386513352394,
+ "learning_rate": 0.0006,
+ "loss": 4.600103378295898,
+ "step": 1304
+ },
+ {
+ "epoch": 18.125764192139737,
+ "grad_norm": 0.018495498225092888,
+ "learning_rate": 0.0006,
+ "loss": 4.588401794433594,
+ "step": 1305
+ },
+ {
+ "epoch": 18.139737991266376,
+ "grad_norm": 0.017921684309840202,
+ "learning_rate": 0.0006,
+ "loss": 4.478075981140137,
+ "step": 1306
+ },
+ {
+ "epoch": 18.153711790393015,
+ "grad_norm": 0.01801278069615364,
+ "learning_rate": 0.0006,
+ "loss": 4.413186550140381,
+ "step": 1307
+ },
+ {
+ "epoch": 18.16768558951965,
+ "grad_norm": 0.017211275175213814,
+ "learning_rate": 0.0006,
+ "loss": 4.461341857910156,
+ "step": 1308
+ },
+ {
+ "epoch": 18.18165938864629,
+ "grad_norm": 0.017589328810572624,
+ "learning_rate": 0.0006,
+ "loss": 4.546969413757324,
+ "step": 1309
+ },
+ {
+ "epoch": 18.195633187772927,
+ "grad_norm": 0.01797698810696602,
+ "learning_rate": 0.0006,
+ "loss": 4.530797004699707,
+ "step": 1310
+ },
+ {
+ "epoch": 18.209606986899562,
+ "grad_norm": 0.020408619195222855,
+ "learning_rate": 0.0006,
+ "loss": 4.62999153137207,
+ "step": 1311
+ },
+ {
+ "epoch": 18.2235807860262,
+ "grad_norm": 0.01994038000702858,
+ "learning_rate": 0.0006,
+ "loss": 4.586811065673828,
+ "step": 1312
+ },
+ {
+ "epoch": 18.23755458515284,
+ "grad_norm": 0.019381340593099594,
+ "learning_rate": 0.0006,
+ "loss": 4.4769368171691895,
+ "step": 1313
+ },
+ {
+ "epoch": 18.251528384279474,
+ "grad_norm": 0.020484555512666702,
+ "learning_rate": 0.0006,
+ "loss": 4.524302959442139,
+ "step": 1314
+ },
+ {
+ "epoch": 18.265502183406113,
+ "grad_norm": 0.02426416426897049,
+ "learning_rate": 0.0006,
+ "loss": 4.561705589294434,
+ "step": 1315
+ },
+ {
+ "epoch": 18.27947598253275,
+ "grad_norm": 0.024748101830482483,
+ "learning_rate": 0.0006,
+ "loss": 4.4654316902160645,
+ "step": 1316
+ },
+ {
+ "epoch": 18.29344978165939,
+ "grad_norm": 0.020703328773379326,
+ "learning_rate": 0.0006,
+ "loss": 4.53033447265625,
+ "step": 1317
+ },
+ {
+ "epoch": 18.307423580786025,
+ "grad_norm": 0.01765458658337593,
+ "learning_rate": 0.0006,
+ "loss": 4.5154266357421875,
+ "step": 1318
+ },
+ {
+ "epoch": 18.321397379912664,
+ "grad_norm": 0.018646420910954475,
+ "learning_rate": 0.0006,
+ "loss": 4.5224714279174805,
+ "step": 1319
+ },
+ {
+ "epoch": 18.335371179039303,
+ "grad_norm": 0.01907474733889103,
+ "learning_rate": 0.0006,
+ "loss": 4.5055437088012695,
+ "step": 1320
+ },
+ {
+ "epoch": 18.349344978165938,
+ "grad_norm": 0.018179846927523613,
+ "learning_rate": 0.0006,
+ "loss": 4.462161064147949,
+ "step": 1321
+ },
+ {
+ "epoch": 18.363318777292577,
+ "grad_norm": 0.016697878018021584,
+ "learning_rate": 0.0006,
+ "loss": 4.473718643188477,
+ "step": 1322
+ },
+ {
+ "epoch": 18.377292576419215,
+ "grad_norm": 0.017504019662737846,
+ "learning_rate": 0.0006,
+ "loss": 4.474348068237305,
+ "step": 1323
+ },
+ {
+ "epoch": 18.39126637554585,
+ "grad_norm": 0.018889913335442543,
+ "learning_rate": 0.0006,
+ "loss": 4.4531636238098145,
+ "step": 1324
+ },
+ {
+ "epoch": 18.40524017467249,
+ "grad_norm": 0.019167251884937286,
+ "learning_rate": 0.0006,
+ "loss": 4.4278364181518555,
+ "step": 1325
+ },
+ {
+ "epoch": 18.419213973799128,
+ "grad_norm": 0.015895741060376167,
+ "learning_rate": 0.0006,
+ "loss": 4.474592208862305,
+ "step": 1326
+ },
+ {
+ "epoch": 18.433187772925763,
+ "grad_norm": 0.01531882956624031,
+ "learning_rate": 0.0006,
+ "loss": 4.403757572174072,
+ "step": 1327
+ },
+ {
+ "epoch": 18.4471615720524,
+ "grad_norm": 0.01565849967300892,
+ "learning_rate": 0.0006,
+ "loss": 4.507664680480957,
+ "step": 1328
+ },
+ {
+ "epoch": 18.46113537117904,
+ "grad_norm": 0.01602315716445446,
+ "learning_rate": 0.0006,
+ "loss": 4.5648908615112305,
+ "step": 1329
+ },
+ {
+ "epoch": 18.475109170305675,
+ "grad_norm": 0.016467366367578506,
+ "learning_rate": 0.0006,
+ "loss": 4.428762912750244,
+ "step": 1330
+ },
+ {
+ "epoch": 18.489082969432314,
+ "grad_norm": 0.01604127697646618,
+ "learning_rate": 0.0006,
+ "loss": 4.527771472930908,
+ "step": 1331
+ },
+ {
+ "epoch": 18.503056768558952,
+ "grad_norm": 0.018825586885213852,
+ "learning_rate": 0.0006,
+ "loss": 4.485386848449707,
+ "step": 1332
+ },
+ {
+ "epoch": 18.51703056768559,
+ "grad_norm": 0.019258007407188416,
+ "learning_rate": 0.0006,
+ "loss": 4.414972305297852,
+ "step": 1333
+ },
+ {
+ "epoch": 18.531004366812226,
+ "grad_norm": 0.018192021176218987,
+ "learning_rate": 0.0006,
+ "loss": 4.412166595458984,
+ "step": 1334
+ },
+ {
+ "epoch": 18.544978165938865,
+ "grad_norm": 0.016120698302984238,
+ "learning_rate": 0.0006,
+ "loss": 4.484317779541016,
+ "step": 1335
+ },
+ {
+ "epoch": 18.558951965065503,
+ "grad_norm": 0.016496622934937477,
+ "learning_rate": 0.0006,
+ "loss": 4.514986515045166,
+ "step": 1336
+ },
+ {
+ "epoch": 18.57292576419214,
+ "grad_norm": 0.01795116998255253,
+ "learning_rate": 0.0006,
+ "loss": 4.41473388671875,
+ "step": 1337
+ },
+ {
+ "epoch": 18.586899563318777,
+ "grad_norm": 0.019062276929616928,
+ "learning_rate": 0.0006,
+ "loss": 4.601996898651123,
+ "step": 1338
+ },
+ {
+ "epoch": 18.600873362445416,
+ "grad_norm": 0.017174631357192993,
+ "learning_rate": 0.0006,
+ "loss": 4.571218490600586,
+ "step": 1339
+ },
+ {
+ "epoch": 18.61484716157205,
+ "grad_norm": 0.014015606604516506,
+ "learning_rate": 0.0006,
+ "loss": 4.526606559753418,
+ "step": 1340
+ },
+ {
+ "epoch": 18.62882096069869,
+ "grad_norm": 0.015358510427176952,
+ "learning_rate": 0.0006,
+ "loss": 4.604286193847656,
+ "step": 1341
+ },
+ {
+ "epoch": 18.64279475982533,
+ "grad_norm": 0.01619391329586506,
+ "learning_rate": 0.0006,
+ "loss": 4.560883522033691,
+ "step": 1342
+ },
+ {
+ "epoch": 18.656768558951963,
+ "grad_norm": 0.015972964465618134,
+ "learning_rate": 0.0006,
+ "loss": 4.479755401611328,
+ "step": 1343
+ },
+ {
+ "epoch": 18.670742358078602,
+ "grad_norm": 0.016861025243997574,
+ "learning_rate": 0.0006,
+ "loss": 4.474471092224121,
+ "step": 1344
+ },
+ {
+ "epoch": 18.68471615720524,
+ "grad_norm": 0.01822078227996826,
+ "learning_rate": 0.0006,
+ "loss": 4.509489059448242,
+ "step": 1345
+ },
+ {
+ "epoch": 18.69868995633188,
+ "grad_norm": 0.019415007904171944,
+ "learning_rate": 0.0006,
+ "loss": 4.508941173553467,
+ "step": 1346
+ },
+ {
+ "epoch": 18.712663755458514,
+ "grad_norm": 0.019856126978993416,
+ "learning_rate": 0.0006,
+ "loss": 4.446670055389404,
+ "step": 1347
+ },
+ {
+ "epoch": 18.726637554585153,
+ "grad_norm": 0.01842520199716091,
+ "learning_rate": 0.0006,
+ "loss": 4.524422645568848,
+ "step": 1348
+ },
+ {
+ "epoch": 18.74061135371179,
+ "grad_norm": 0.017196480184793472,
+ "learning_rate": 0.0006,
+ "loss": 4.498466968536377,
+ "step": 1349
+ },
+ {
+ "epoch": 18.754585152838427,
+ "grad_norm": 0.01823010854423046,
+ "learning_rate": 0.0006,
+ "loss": 4.534684181213379,
+ "step": 1350
+ },
+ {
+ "epoch": 18.768558951965066,
+ "grad_norm": 0.020398065447807312,
+ "learning_rate": 0.0006,
+ "loss": 4.456494331359863,
+ "step": 1351
+ },
+ {
+ "epoch": 18.782532751091704,
+ "grad_norm": 0.022992262616753578,
+ "learning_rate": 0.0006,
+ "loss": 4.476374626159668,
+ "step": 1352
+ },
+ {
+ "epoch": 18.79650655021834,
+ "grad_norm": 0.02235420234501362,
+ "learning_rate": 0.0006,
+ "loss": 4.587776184082031,
+ "step": 1353
+ },
+ {
+ "epoch": 18.810480349344978,
+ "grad_norm": 0.02093626745045185,
+ "learning_rate": 0.0006,
+ "loss": 4.424172401428223,
+ "step": 1354
+ },
+ {
+ "epoch": 18.824454148471617,
+ "grad_norm": 0.01997302658855915,
+ "learning_rate": 0.0006,
+ "loss": 4.448066711425781,
+ "step": 1355
+ },
+ {
+ "epoch": 18.83842794759825,
+ "grad_norm": 0.018265459686517715,
+ "learning_rate": 0.0006,
+ "loss": 4.436610698699951,
+ "step": 1356
+ },
+ {
+ "epoch": 18.85240174672489,
+ "grad_norm": 0.01673833839595318,
+ "learning_rate": 0.0006,
+ "loss": 4.5502214431762695,
+ "step": 1357
+ },
+ {
+ "epoch": 18.86637554585153,
+ "grad_norm": 0.017908833920955658,
+ "learning_rate": 0.0006,
+ "loss": 4.550530433654785,
+ "step": 1358
+ },
+ {
+ "epoch": 18.880349344978168,
+ "grad_norm": 0.01796654611825943,
+ "learning_rate": 0.0006,
+ "loss": 4.531463146209717,
+ "step": 1359
+ },
+ {
+ "epoch": 18.894323144104803,
+ "grad_norm": 0.01646183803677559,
+ "learning_rate": 0.0006,
+ "loss": 4.498582363128662,
+ "step": 1360
+ },
+ {
+ "epoch": 18.90829694323144,
+ "grad_norm": 0.015430327504873276,
+ "learning_rate": 0.0006,
+ "loss": 4.52571439743042,
+ "step": 1361
+ },
+ {
+ "epoch": 18.92227074235808,
+ "grad_norm": 0.016527952626347542,
+ "learning_rate": 0.0006,
+ "loss": 4.547013759613037,
+ "step": 1362
+ },
+ {
+ "epoch": 18.936244541484715,
+ "grad_norm": 0.016263660043478012,
+ "learning_rate": 0.0006,
+ "loss": 4.468754291534424,
+ "step": 1363
+ },
+ {
+ "epoch": 18.950218340611354,
+ "grad_norm": 0.016347210854291916,
+ "learning_rate": 0.0006,
+ "loss": 4.450741767883301,
+ "step": 1364
+ },
+ {
+ "epoch": 18.964192139737992,
+ "grad_norm": 0.01701398566365242,
+ "learning_rate": 0.0006,
+ "loss": 4.534899711608887,
+ "step": 1365
+ },
+ {
+ "epoch": 18.978165938864628,
+ "grad_norm": 0.017693081870675087,
+ "learning_rate": 0.0006,
+ "loss": 4.460553169250488,
+ "step": 1366
+ },
+ {
+ "epoch": 18.992139737991266,
+ "grad_norm": 0.018746308982372284,
+ "learning_rate": 0.0006,
+ "loss": 4.538792610168457,
+ "step": 1367
+ },
+ {
+ "epoch": 19.0,
+ "grad_norm": 0.020099416375160217,
+ "learning_rate": 0.0006,
+ "loss": 4.449617385864258,
+ "step": 1368
+ },
+ {
+ "epoch": 19.0,
+ "eval_loss": 4.7756571769714355,
+ "eval_runtime": 57.1804,
+ "eval_samples_per_second": 42.707,
+ "eval_steps_per_second": 1.347,
+ "step": 1368
+ },
+ {
+ "epoch": 19.01397379912664,
+ "grad_norm": 0.018726933747529984,
+ "learning_rate": 0.0006,
+ "loss": 4.446059703826904,
+ "step": 1369
+ },
+ {
+ "epoch": 19.027947598253274,
+ "grad_norm": 0.01956087350845337,
+ "learning_rate": 0.0006,
+ "loss": 4.4520039558410645,
+ "step": 1370
+ },
+ {
+ "epoch": 19.041921397379912,
+ "grad_norm": 0.020092811435461044,
+ "learning_rate": 0.0006,
+ "loss": 4.543343544006348,
+ "step": 1371
+ },
+ {
+ "epoch": 19.05589519650655,
+ "grad_norm": 0.01843622885644436,
+ "learning_rate": 0.0006,
+ "loss": 4.535579204559326,
+ "step": 1372
+ },
+ {
+ "epoch": 19.069868995633186,
+ "grad_norm": 0.016989488154649734,
+ "learning_rate": 0.0006,
+ "loss": 4.52316951751709,
+ "step": 1373
+ },
+ {
+ "epoch": 19.083842794759825,
+ "grad_norm": 0.01805422455072403,
+ "learning_rate": 0.0006,
+ "loss": 4.5393266677856445,
+ "step": 1374
+ },
+ {
+ "epoch": 19.097816593886463,
+ "grad_norm": 0.01874629408121109,
+ "learning_rate": 0.0006,
+ "loss": 4.422059059143066,
+ "step": 1375
+ },
+ {
+ "epoch": 19.111790393013102,
+ "grad_norm": 0.01804221048951149,
+ "learning_rate": 0.0006,
+ "loss": 4.50508451461792,
+ "step": 1376
+ },
+ {
+ "epoch": 19.125764192139737,
+ "grad_norm": 0.01785474829375744,
+ "learning_rate": 0.0006,
+ "loss": 4.412599563598633,
+ "step": 1377
+ },
+ {
+ "epoch": 19.139737991266376,
+ "grad_norm": 0.01971791312098503,
+ "learning_rate": 0.0006,
+ "loss": 4.518318176269531,
+ "step": 1378
+ },
+ {
+ "epoch": 19.153711790393015,
+ "grad_norm": 0.02061633951961994,
+ "learning_rate": 0.0006,
+ "loss": 4.332393169403076,
+ "step": 1379
+ },
+ {
+ "epoch": 19.16768558951965,
+ "grad_norm": 0.020563429221510887,
+ "learning_rate": 0.0006,
+ "loss": 4.445863723754883,
+ "step": 1380
+ },
+ {
+ "epoch": 19.18165938864629,
+ "grad_norm": 0.019995713606476784,
+ "learning_rate": 0.0006,
+ "loss": 4.497312068939209,
+ "step": 1381
+ },
+ {
+ "epoch": 19.195633187772927,
+ "grad_norm": 0.019823070615530014,
+ "learning_rate": 0.0006,
+ "loss": 4.475085258483887,
+ "step": 1382
+ },
+ {
+ "epoch": 19.209606986899562,
+ "grad_norm": 0.019039355218410492,
+ "learning_rate": 0.0006,
+ "loss": 4.50432014465332,
+ "step": 1383
+ },
+ {
+ "epoch": 19.2235807860262,
+ "grad_norm": 0.017876973375678062,
+ "learning_rate": 0.0006,
+ "loss": 4.384788513183594,
+ "step": 1384
+ },
+ {
+ "epoch": 19.23755458515284,
+ "grad_norm": 0.018697699531912804,
+ "learning_rate": 0.0006,
+ "loss": 4.410778045654297,
+ "step": 1385
+ },
+ {
+ "epoch": 19.251528384279474,
+ "grad_norm": 0.018826806917786598,
+ "learning_rate": 0.0006,
+ "loss": 4.524034023284912,
+ "step": 1386
+ },
+ {
+ "epoch": 19.265502183406113,
+ "grad_norm": 0.01824226602911949,
+ "learning_rate": 0.0006,
+ "loss": 4.526393890380859,
+ "step": 1387
+ },
+ {
+ "epoch": 19.27947598253275,
+ "grad_norm": 0.01753208599984646,
+ "learning_rate": 0.0006,
+ "loss": 4.493703365325928,
+ "step": 1388
+ },
+ {
+ "epoch": 19.29344978165939,
+ "grad_norm": 0.01742432825267315,
+ "learning_rate": 0.0006,
+ "loss": 4.392831802368164,
+ "step": 1389
+ },
+ {
+ "epoch": 19.307423580786025,
+ "grad_norm": 0.017099356278777122,
+ "learning_rate": 0.0006,
+ "loss": 4.578668594360352,
+ "step": 1390
+ },
+ {
+ "epoch": 19.321397379912664,
+ "grad_norm": 0.017946023494005203,
+ "learning_rate": 0.0006,
+ "loss": 4.456718444824219,
+ "step": 1391
+ },
+ {
+ "epoch": 19.335371179039303,
+ "grad_norm": 0.017705943435430527,
+ "learning_rate": 0.0006,
+ "loss": 4.553682327270508,
+ "step": 1392
+ },
+ {
+ "epoch": 19.349344978165938,
+ "grad_norm": 0.016647523269057274,
+ "learning_rate": 0.0006,
+ "loss": 4.525016784667969,
+ "step": 1393
+ },
+ {
+ "epoch": 19.363318777292577,
+ "grad_norm": 0.015407416969537735,
+ "learning_rate": 0.0006,
+ "loss": 4.500231742858887,
+ "step": 1394
+ },
+ {
+ "epoch": 19.377292576419215,
+ "grad_norm": 0.017264019697904587,
+ "learning_rate": 0.0006,
+ "loss": 4.4615349769592285,
+ "step": 1395
+ },
+ {
+ "epoch": 19.39126637554585,
+ "grad_norm": 0.01871555484831333,
+ "learning_rate": 0.0006,
+ "loss": 4.281026840209961,
+ "step": 1396
+ },
+ {
+ "epoch": 19.40524017467249,
+ "grad_norm": 0.017361650243401527,
+ "learning_rate": 0.0006,
+ "loss": 4.479382514953613,
+ "step": 1397
+ },
+ {
+ "epoch": 19.419213973799128,
+ "grad_norm": 0.019781362265348434,
+ "learning_rate": 0.0006,
+ "loss": 4.402320861816406,
+ "step": 1398
+ },
+ {
+ "epoch": 19.433187772925763,
+ "grad_norm": 0.019043298438191414,
+ "learning_rate": 0.0006,
+ "loss": 4.476983070373535,
+ "step": 1399
+ },
+ {
+ "epoch": 19.4471615720524,
+ "grad_norm": 0.017137985676527023,
+ "learning_rate": 0.0006,
+ "loss": 4.418689250946045,
+ "step": 1400
+ },
+ {
+ "epoch": 19.46113537117904,
+ "grad_norm": 0.016481833532452583,
+ "learning_rate": 0.0006,
+ "loss": 4.579145431518555,
+ "step": 1401
+ },
+ {
+ "epoch": 19.475109170305675,
+ "grad_norm": 0.016507714986801147,
+ "learning_rate": 0.0006,
+ "loss": 4.519158363342285,
+ "step": 1402
+ },
+ {
+ "epoch": 19.489082969432314,
+ "grad_norm": 0.018092872574925423,
+ "learning_rate": 0.0006,
+ "loss": 4.513748645782471,
+ "step": 1403
+ },
+ {
+ "epoch": 19.503056768558952,
+ "grad_norm": 0.018153440207242966,
+ "learning_rate": 0.0006,
+ "loss": 4.645384311676025,
+ "step": 1404
+ },
+ {
+ "epoch": 19.51703056768559,
+ "grad_norm": 0.01657119207084179,
+ "learning_rate": 0.0006,
+ "loss": 4.399757385253906,
+ "step": 1405
+ },
+ {
+ "epoch": 19.531004366812226,
+ "grad_norm": 0.0156280305236578,
+ "learning_rate": 0.0006,
+ "loss": 4.467836380004883,
+ "step": 1406
+ },
+ {
+ "epoch": 19.544978165938865,
+ "grad_norm": 0.016838403418660164,
+ "learning_rate": 0.0006,
+ "loss": 4.529176235198975,
+ "step": 1407
+ },
+ {
+ "epoch": 19.558951965065503,
+ "grad_norm": 0.01743236929178238,
+ "learning_rate": 0.0006,
+ "loss": 4.440032482147217,
+ "step": 1408
+ },
+ {
+ "epoch": 19.57292576419214,
+ "grad_norm": 0.01553610060364008,
+ "learning_rate": 0.0006,
+ "loss": 4.425588607788086,
+ "step": 1409
+ },
+ {
+ "epoch": 19.586899563318777,
+ "grad_norm": 0.017259271815419197,
+ "learning_rate": 0.0006,
+ "loss": 4.405452728271484,
+ "step": 1410
+ },
+ {
+ "epoch": 19.600873362445416,
+ "grad_norm": 0.015568481758236885,
+ "learning_rate": 0.0006,
+ "loss": 4.444279193878174,
+ "step": 1411
+ },
+ {
+ "epoch": 19.61484716157205,
+ "grad_norm": 0.015331248752772808,
+ "learning_rate": 0.0006,
+ "loss": 4.451846122741699,
+ "step": 1412
+ },
+ {
+ "epoch": 19.62882096069869,
+ "grad_norm": 0.015667378902435303,
+ "learning_rate": 0.0006,
+ "loss": 4.533931255340576,
+ "step": 1413
+ },
+ {
+ "epoch": 19.64279475982533,
+ "grad_norm": 0.016177602112293243,
+ "learning_rate": 0.0006,
+ "loss": 4.467313289642334,
+ "step": 1414
+ },
+ {
+ "epoch": 19.656768558951963,
+ "grad_norm": 0.01733129471540451,
+ "learning_rate": 0.0006,
+ "loss": 4.496893882751465,
+ "step": 1415
+ },
+ {
+ "epoch": 19.670742358078602,
+ "grad_norm": 0.016539594158530235,
+ "learning_rate": 0.0006,
+ "loss": 4.401961326599121,
+ "step": 1416
+ },
+ {
+ "epoch": 19.68471615720524,
+ "grad_norm": 0.016058551147580147,
+ "learning_rate": 0.0006,
+ "loss": 4.503375053405762,
+ "step": 1417
+ },
+ {
+ "epoch": 19.69868995633188,
+ "grad_norm": 0.01715777814388275,
+ "learning_rate": 0.0006,
+ "loss": 4.3808369636535645,
+ "step": 1418
+ },
+ {
+ "epoch": 19.712663755458514,
+ "grad_norm": 0.01681119203567505,
+ "learning_rate": 0.0006,
+ "loss": 4.5469560623168945,
+ "step": 1419
+ },
+ {
+ "epoch": 19.726637554585153,
+ "grad_norm": 0.01634550653398037,
+ "learning_rate": 0.0006,
+ "loss": 4.328197956085205,
+ "step": 1420
+ },
+ {
+ "epoch": 19.74061135371179,
+ "grad_norm": 0.01549347210675478,
+ "learning_rate": 0.0006,
+ "loss": 4.534621238708496,
+ "step": 1421
+ },
+ {
+ "epoch": 19.754585152838427,
+ "grad_norm": 0.01654665172100067,
+ "learning_rate": 0.0006,
+ "loss": 4.4724225997924805,
+ "step": 1422
+ },
+ {
+ "epoch": 19.768558951965066,
+ "grad_norm": 0.017119025811553,
+ "learning_rate": 0.0006,
+ "loss": 4.3897247314453125,
+ "step": 1423
+ },
+ {
+ "epoch": 19.782532751091704,
+ "grad_norm": 0.01603223942220211,
+ "learning_rate": 0.0006,
+ "loss": 4.412731170654297,
+ "step": 1424
+ },
+ {
+ "epoch": 19.79650655021834,
+ "grad_norm": 0.016774067655205727,
+ "learning_rate": 0.0006,
+ "loss": 4.54707145690918,
+ "step": 1425
+ },
+ {
+ "epoch": 19.810480349344978,
+ "grad_norm": 0.016453025862574577,
+ "learning_rate": 0.0006,
+ "loss": 4.462501049041748,
+ "step": 1426
+ },
+ {
+ "epoch": 19.824454148471617,
+ "grad_norm": 0.01597629487514496,
+ "learning_rate": 0.0006,
+ "loss": 4.229758262634277,
+ "step": 1427
+ },
+ {
+ "epoch": 19.83842794759825,
+ "grad_norm": 0.017999034374952316,
+ "learning_rate": 0.0006,
+ "loss": 4.389263153076172,
+ "step": 1428
+ },
+ {
+ "epoch": 19.85240174672489,
+ "grad_norm": 0.02165898121893406,
+ "learning_rate": 0.0006,
+ "loss": 4.342130661010742,
+ "step": 1429
+ },
+ {
+ "epoch": 19.86637554585153,
+ "grad_norm": 0.02222822792828083,
+ "learning_rate": 0.0006,
+ "loss": 4.492661476135254,
+ "step": 1430
+ },
+ {
+ "epoch": 19.880349344978168,
+ "grad_norm": 0.019378066062927246,
+ "learning_rate": 0.0006,
+ "loss": 4.4614105224609375,
+ "step": 1431
+ },
+ {
+ "epoch": 19.894323144104803,
+ "grad_norm": 0.019908083602786064,
+ "learning_rate": 0.0006,
+ "loss": 4.370929718017578,
+ "step": 1432
+ },
+ {
+ "epoch": 19.90829694323144,
+ "grad_norm": 0.022296471521258354,
+ "learning_rate": 0.0006,
+ "loss": 4.438320159912109,
+ "step": 1433
+ },
+ {
+ "epoch": 19.92227074235808,
+ "grad_norm": 0.021810343489050865,
+ "learning_rate": 0.0006,
+ "loss": 4.462067127227783,
+ "step": 1434
+ },
+ {
+ "epoch": 19.936244541484715,
+ "grad_norm": 0.02129209227859974,
+ "learning_rate": 0.0006,
+ "loss": 4.541261196136475,
+ "step": 1435
+ },
+ {
+ "epoch": 19.950218340611354,
+ "grad_norm": 0.02079629711806774,
+ "learning_rate": 0.0006,
+ "loss": 4.36712646484375,
+ "step": 1436
+ },
+ {
+ "epoch": 19.964192139737992,
+ "grad_norm": 0.022501740604639053,
+ "learning_rate": 0.0006,
+ "loss": 4.5279083251953125,
+ "step": 1437
+ },
+ {
+ "epoch": 19.978165938864628,
+ "grad_norm": 0.019121676683425903,
+ "learning_rate": 0.0006,
+ "loss": 4.477789878845215,
+ "step": 1438
+ },
+ {
+ "epoch": 19.992139737991266,
+ "grad_norm": 0.017429588362574577,
+ "learning_rate": 0.0006,
+ "loss": 4.4023847579956055,
+ "step": 1439
+ },
+ {
+ "epoch": 20.0,
+ "grad_norm": 0.018857665359973907,
+ "learning_rate": 0.0006,
+ "loss": 4.494743347167969,
+ "step": 1440
+ },
+ {
+ "epoch": 20.0,
+ "eval_loss": 4.694555759429932,
+ "eval_runtime": 56.4146,
+ "eval_samples_per_second": 43.287,
+ "eval_steps_per_second": 1.365,
+ "step": 1440
+ },
+ {
+ "epoch": 20.01397379912664,
+ "grad_norm": 0.016505861654877663,
+ "learning_rate": 0.0006,
+ "loss": 4.506229877471924,
+ "step": 1441
+ },
+ {
+ "epoch": 20.027947598253274,
+ "grad_norm": 0.017571842297911644,
+ "learning_rate": 0.0006,
+ "loss": 4.384073257446289,
+ "step": 1442
+ },
+ {
+ "epoch": 20.041921397379912,
+ "grad_norm": 0.019398365169763565,
+ "learning_rate": 0.0006,
+ "loss": 4.482177734375,
+ "step": 1443
+ },
+ {
+ "epoch": 20.05589519650655,
+ "grad_norm": 0.02011062018573284,
+ "learning_rate": 0.0006,
+ "loss": 4.440773963928223,
+ "step": 1444
+ },
+ {
+ "epoch": 20.069868995633186,
+ "grad_norm": 0.019026953727006912,
+ "learning_rate": 0.0006,
+ "loss": 4.507168769836426,
+ "step": 1445
+ },
+ {
+ "epoch": 20.083842794759825,
+ "grad_norm": 0.020506031811237335,
+ "learning_rate": 0.0006,
+ "loss": 4.469079971313477,
+ "step": 1446
+ },
+ {
+ "epoch": 20.097816593886463,
+ "grad_norm": 0.019608162343502045,
+ "learning_rate": 0.0006,
+ "loss": 4.395927429199219,
+ "step": 1447
+ },
+ {
+ "epoch": 20.111790393013102,
+ "grad_norm": 0.019419124349951744,
+ "learning_rate": 0.0006,
+ "loss": 4.484235763549805,
+ "step": 1448
+ },
+ {
+ "epoch": 20.125764192139737,
+ "grad_norm": 0.021775128319859505,
+ "learning_rate": 0.0006,
+ "loss": 4.333173751831055,
+ "step": 1449
+ },
+ {
+ "epoch": 20.139737991266376,
+ "grad_norm": 0.025512272492051125,
+ "learning_rate": 0.0006,
+ "loss": 4.460362434387207,
+ "step": 1450
+ },
+ {
+ "epoch": 20.153711790393015,
+ "grad_norm": 0.025157401338219643,
+ "learning_rate": 0.0006,
+ "loss": 4.4361395835876465,
+ "step": 1451
+ },
+ {
+ "epoch": 20.16768558951965,
+ "grad_norm": 0.019862616434693336,
+ "learning_rate": 0.0006,
+ "loss": 4.426807403564453,
+ "step": 1452
+ },
+ {
+ "epoch": 20.18165938864629,
+ "grad_norm": 0.01932152360677719,
+ "learning_rate": 0.0006,
+ "loss": 4.375677108764648,
+ "step": 1453
+ },
+ {
+ "epoch": 20.195633187772927,
+ "grad_norm": 0.01760999858379364,
+ "learning_rate": 0.0006,
+ "loss": 4.42259407043457,
+ "step": 1454
+ },
+ {
+ "epoch": 20.209606986899562,
+ "grad_norm": 0.016998112201690674,
+ "learning_rate": 0.0006,
+ "loss": 4.481280326843262,
+ "step": 1455
+ },
+ {
+ "epoch": 20.2235807860262,
+ "grad_norm": 0.017199190333485603,
+ "learning_rate": 0.0006,
+ "loss": 4.496712684631348,
+ "step": 1456
+ },
+ {
+ "epoch": 20.23755458515284,
+ "grad_norm": 0.016833679750561714,
+ "learning_rate": 0.0006,
+ "loss": 4.40736722946167,
+ "step": 1457
+ },
+ {
+ "epoch": 20.251528384279474,
+ "grad_norm": 0.017546730116009712,
+ "learning_rate": 0.0006,
+ "loss": 4.4915266036987305,
+ "step": 1458
+ },
+ {
+ "epoch": 20.265502183406113,
+ "grad_norm": 0.016595078632235527,
+ "learning_rate": 0.0006,
+ "loss": 4.460815906524658,
+ "step": 1459
+ },
+ {
+ "epoch": 20.27947598253275,
+ "grad_norm": 0.017728568986058235,
+ "learning_rate": 0.0006,
+ "loss": 4.346048831939697,
+ "step": 1460
+ },
+ {
+ "epoch": 20.29344978165939,
+ "grad_norm": 0.01849197782576084,
+ "learning_rate": 0.0006,
+ "loss": 4.399681091308594,
+ "step": 1461
+ },
+ {
+ "epoch": 20.307423580786025,
+ "grad_norm": 0.018336588516831398,
+ "learning_rate": 0.0006,
+ "loss": 4.414183616638184,
+ "step": 1462
+ },
+ {
+ "epoch": 20.321397379912664,
+ "grad_norm": 0.018086519092321396,
+ "learning_rate": 0.0006,
+ "loss": 4.346586227416992,
+ "step": 1463
+ },
+ {
+ "epoch": 20.335371179039303,
+ "grad_norm": 0.019697299227118492,
+ "learning_rate": 0.0006,
+ "loss": 4.4420952796936035,
+ "step": 1464
+ },
+ {
+ "epoch": 20.349344978165938,
+ "grad_norm": 0.02198074571788311,
+ "learning_rate": 0.0006,
+ "loss": 4.588729381561279,
+ "step": 1465
+ },
+ {
+ "epoch": 20.363318777292577,
+ "grad_norm": 0.022813329473137856,
+ "learning_rate": 0.0006,
+ "loss": 4.49729585647583,
+ "step": 1466
+ },
+ {
+ "epoch": 20.377292576419215,
+ "grad_norm": 0.022006649523973465,
+ "learning_rate": 0.0006,
+ "loss": 4.475641250610352,
+ "step": 1467
+ },
+ {
+ "epoch": 20.39126637554585,
+ "grad_norm": 0.023188438266515732,
+ "learning_rate": 0.0006,
+ "loss": 4.404322624206543,
+ "step": 1468
+ },
+ {
+ "epoch": 20.40524017467249,
+ "grad_norm": 0.022226877510547638,
+ "learning_rate": 0.0006,
+ "loss": 4.50904655456543,
+ "step": 1469
+ },
+ {
+ "epoch": 20.419213973799128,
+ "grad_norm": 0.021082637831568718,
+ "learning_rate": 0.0006,
+ "loss": 4.339011192321777,
+ "step": 1470
+ },
+ {
+ "epoch": 20.433187772925763,
+ "grad_norm": 0.02013542130589485,
+ "learning_rate": 0.0006,
+ "loss": 4.516942024230957,
+ "step": 1471
+ },
+ {
+ "epoch": 20.4471615720524,
+ "grad_norm": 0.01898968778550625,
+ "learning_rate": 0.0006,
+ "loss": 4.315804481506348,
+ "step": 1472
+ },
+ {
+ "epoch": 20.46113537117904,
+ "grad_norm": 0.019450880587100983,
+ "learning_rate": 0.0006,
+ "loss": 4.4794921875,
+ "step": 1473
+ },
+ {
+ "epoch": 20.475109170305675,
+ "grad_norm": 0.018577802926301956,
+ "learning_rate": 0.0006,
+ "loss": 4.422323703765869,
+ "step": 1474
+ },
+ {
+ "epoch": 20.489082969432314,
+ "grad_norm": 0.01778412237763405,
+ "learning_rate": 0.0006,
+ "loss": 4.429789066314697,
+ "step": 1475
+ },
+ {
+ "epoch": 20.503056768558952,
+ "grad_norm": 0.017134476453065872,
+ "learning_rate": 0.0006,
+ "loss": 4.416459083557129,
+ "step": 1476
+ },
+ {
+ "epoch": 20.51703056768559,
+ "grad_norm": 0.017153557389974594,
+ "learning_rate": 0.0006,
+ "loss": 4.446170806884766,
+ "step": 1477
+ },
+ {
+ "epoch": 20.531004366812226,
+ "grad_norm": 0.014998532831668854,
+ "learning_rate": 0.0006,
+ "loss": 4.419447898864746,
+ "step": 1478
+ },
+ {
+ "epoch": 20.544978165938865,
+ "grad_norm": 0.015877850353717804,
+ "learning_rate": 0.0006,
+ "loss": 4.312338829040527,
+ "step": 1479
+ },
+ {
+ "epoch": 20.558951965065503,
+ "grad_norm": 0.01569489948451519,
+ "learning_rate": 0.0006,
+ "loss": 4.572905540466309,
+ "step": 1480
+ },
+ {
+ "epoch": 20.57292576419214,
+ "grad_norm": 0.01591474376618862,
+ "learning_rate": 0.0006,
+ "loss": 4.414907455444336,
+ "step": 1481
+ },
+ {
+ "epoch": 20.586899563318777,
+ "grad_norm": 0.015395903028547764,
+ "learning_rate": 0.0006,
+ "loss": 4.549131393432617,
+ "step": 1482
+ },
+ {
+ "epoch": 20.600873362445416,
+ "grad_norm": 0.015715358778834343,
+ "learning_rate": 0.0006,
+ "loss": 4.530194282531738,
+ "step": 1483
+ },
+ {
+ "epoch": 20.61484716157205,
+ "grad_norm": 0.016042975708842278,
+ "learning_rate": 0.0006,
+ "loss": 4.322149276733398,
+ "step": 1484
+ },
+ {
+ "epoch": 20.62882096069869,
+ "grad_norm": 0.01664186641573906,
+ "learning_rate": 0.0006,
+ "loss": 4.424866676330566,
+ "step": 1485
+ },
+ {
+ "epoch": 20.64279475982533,
+ "grad_norm": 0.01631534844636917,
+ "learning_rate": 0.0006,
+ "loss": 4.441752910614014,
+ "step": 1486
+ },
+ {
+ "epoch": 20.656768558951963,
+ "grad_norm": 0.014057058840990067,
+ "learning_rate": 0.0006,
+ "loss": 4.399982452392578,
+ "step": 1487
+ },
+ {
+ "epoch": 20.670742358078602,
+ "grad_norm": 0.015076273120939732,
+ "learning_rate": 0.0006,
+ "loss": 4.441324234008789,
+ "step": 1488
+ },
+ {
+ "epoch": 20.68471615720524,
+ "grad_norm": 0.014138277620077133,
+ "learning_rate": 0.0006,
+ "loss": 4.4312543869018555,
+ "step": 1489
+ },
+ {
+ "epoch": 20.69868995633188,
+ "grad_norm": 0.013762440532445908,
+ "learning_rate": 0.0006,
+ "loss": 4.387807846069336,
+ "step": 1490
+ },
+ {
+ "epoch": 20.712663755458514,
+ "grad_norm": 0.015058411285281181,
+ "learning_rate": 0.0006,
+ "loss": 4.594747543334961,
+ "step": 1491
+ },
+ {
+ "epoch": 20.726637554585153,
+ "grad_norm": 0.013827123679220676,
+ "learning_rate": 0.0006,
+ "loss": 4.434604167938232,
+ "step": 1492
+ },
+ {
+ "epoch": 20.74061135371179,
+ "grad_norm": 0.015315295197069645,
+ "learning_rate": 0.0006,
+ "loss": 4.588066101074219,
+ "step": 1493
+ },
+ {
+ "epoch": 20.754585152838427,
+ "grad_norm": 0.018678732216358185,
+ "learning_rate": 0.0006,
+ "loss": 4.3514204025268555,
+ "step": 1494
+ },
+ {
+ "epoch": 20.768558951965066,
+ "grad_norm": 0.019521350041031837,
+ "learning_rate": 0.0006,
+ "loss": 4.437658786773682,
+ "step": 1495
+ },
+ {
+ "epoch": 20.782532751091704,
+ "grad_norm": 0.018363403156399727,
+ "learning_rate": 0.0006,
+ "loss": 4.495002269744873,
+ "step": 1496
+ },
+ {
+ "epoch": 20.79650655021834,
+ "grad_norm": 0.016830725595355034,
+ "learning_rate": 0.0006,
+ "loss": 4.402471542358398,
+ "step": 1497
+ },
+ {
+ "epoch": 20.810480349344978,
+ "grad_norm": 0.01722072623670101,
+ "learning_rate": 0.0006,
+ "loss": 4.437653541564941,
+ "step": 1498
+ },
+ {
+ "epoch": 20.824454148471617,
+ "grad_norm": 0.017459599301218987,
+ "learning_rate": 0.0006,
+ "loss": 4.366603851318359,
+ "step": 1499
+ },
+ {
+ "epoch": 20.83842794759825,
+ "grad_norm": 0.018902219831943512,
+ "learning_rate": 0.0006,
+ "loss": 4.501561164855957,
+ "step": 1500
+ },
+ {
+ "epoch": 20.85240174672489,
+ "grad_norm": 0.020814096555113792,
+ "learning_rate": 0.0006,
+ "loss": 4.535295486450195,
+ "step": 1501
+ },
+ {
+ "epoch": 20.86637554585153,
+ "grad_norm": 0.018990976735949516,
+ "learning_rate": 0.0006,
+ "loss": 4.333580017089844,
+ "step": 1502
+ },
+ {
+ "epoch": 20.880349344978168,
+ "grad_norm": 0.01837880350649357,
+ "learning_rate": 0.0006,
+ "loss": 4.313388824462891,
+ "step": 1503
+ },
+ {
+ "epoch": 20.894323144104803,
+ "grad_norm": 0.019820360466837883,
+ "learning_rate": 0.0006,
+ "loss": 4.388157844543457,
+ "step": 1504
+ },
+ {
+ "epoch": 20.90829694323144,
+ "grad_norm": 0.020181143656373024,
+ "learning_rate": 0.0006,
+ "loss": 4.572224140167236,
+ "step": 1505
+ },
+ {
+ "epoch": 20.92227074235808,
+ "grad_norm": 0.021206077188253403,
+ "learning_rate": 0.0006,
+ "loss": 4.3127970695495605,
+ "step": 1506
+ },
+ {
+ "epoch": 20.936244541484715,
+ "grad_norm": 0.017962060868740082,
+ "learning_rate": 0.0006,
+ "loss": 4.384734630584717,
+ "step": 1507
+ },
+ {
+ "epoch": 20.950218340611354,
+ "grad_norm": 0.01737220585346222,
+ "learning_rate": 0.0006,
+ "loss": 4.4075751304626465,
+ "step": 1508
+ },
+ {
+ "epoch": 20.964192139737992,
+ "grad_norm": 0.01839268207550049,
+ "learning_rate": 0.0006,
+ "loss": 4.553779125213623,
+ "step": 1509
+ },
+ {
+ "epoch": 20.978165938864628,
+ "grad_norm": 0.020102640613913536,
+ "learning_rate": 0.0006,
+ "loss": 4.399398326873779,
+ "step": 1510
+ },
+ {
+ "epoch": 20.992139737991266,
+ "grad_norm": 0.018831463530659676,
+ "learning_rate": 0.0006,
+ "loss": 4.358499526977539,
+ "step": 1511
+ },
+ {
+ "epoch": 21.0,
+ "grad_norm": 0.018254734575748444,
+ "learning_rate": 0.0006,
+ "loss": 4.384119510650635,
+ "step": 1512
+ },
+ {
+ "epoch": 21.0,
+ "eval_loss": 4.7524237632751465,
+ "eval_runtime": 57.8812,
+ "eval_samples_per_second": 42.19,
+ "eval_steps_per_second": 1.33,
+ "step": 1512
+ },
+ {
+ "epoch": 21.01397379912664,
+ "grad_norm": 0.019728587940335274,
+ "learning_rate": 0.0006,
+ "loss": 4.32363224029541,
+ "step": 1513
+ },
+ {
+ "epoch": 21.027947598253274,
+ "grad_norm": 0.02216625027358532,
+ "learning_rate": 0.0006,
+ "loss": 4.436988353729248,
+ "step": 1514
+ },
+ {
+ "epoch": 21.041921397379912,
+ "grad_norm": 0.02227606810629368,
+ "learning_rate": 0.0006,
+ "loss": 4.406063556671143,
+ "step": 1515
+ },
+ {
+ "epoch": 21.05589519650655,
+ "grad_norm": 0.020187009125947952,
+ "learning_rate": 0.0006,
+ "loss": 4.324402332305908,
+ "step": 1516
+ },
+ {
+ "epoch": 21.069868995633186,
+ "grad_norm": 0.02095157466828823,
+ "learning_rate": 0.0006,
+ "loss": 4.456544876098633,
+ "step": 1517
+ },
+ {
+ "epoch": 21.083842794759825,
+ "grad_norm": 0.02146013453602791,
+ "learning_rate": 0.0006,
+ "loss": 4.368387222290039,
+ "step": 1518
+ },
+ {
+ "epoch": 21.097816593886463,
+ "grad_norm": 0.02162035182118416,
+ "learning_rate": 0.0006,
+ "loss": 4.48838996887207,
+ "step": 1519
+ },
+ {
+ "epoch": 21.111790393013102,
+ "grad_norm": 0.021745676174759865,
+ "learning_rate": 0.0006,
+ "loss": 4.30778694152832,
+ "step": 1520
+ },
+ {
+ "epoch": 21.125764192139737,
+ "grad_norm": 0.02173588052392006,
+ "learning_rate": 0.0006,
+ "loss": 4.433149814605713,
+ "step": 1521
+ },
+ {
+ "epoch": 21.139737991266376,
+ "grad_norm": 0.020950648933649063,
+ "learning_rate": 0.0006,
+ "loss": 4.520089149475098,
+ "step": 1522
+ },
+ {
+ "epoch": 21.153711790393015,
+ "grad_norm": 0.020684752613306046,
+ "learning_rate": 0.0006,
+ "loss": 4.372437477111816,
+ "step": 1523
+ },
+ {
+ "epoch": 21.16768558951965,
+ "grad_norm": 0.019472869113087654,
+ "learning_rate": 0.0006,
+ "loss": 4.437015533447266,
+ "step": 1524
+ },
+ {
+ "epoch": 21.18165938864629,
+ "grad_norm": 0.01846962980926037,
+ "learning_rate": 0.0006,
+ "loss": 4.449531555175781,
+ "step": 1525
+ },
+ {
+ "epoch": 21.195633187772927,
+ "grad_norm": 0.017377803102135658,
+ "learning_rate": 0.0006,
+ "loss": 4.458279132843018,
+ "step": 1526
+ },
+ {
+ "epoch": 21.209606986899562,
+ "grad_norm": 0.016002200543880463,
+ "learning_rate": 0.0006,
+ "loss": 4.266297817230225,
+ "step": 1527
+ },
+ {
+ "epoch": 21.2235807860262,
+ "grad_norm": 0.015575903467833996,
+ "learning_rate": 0.0006,
+ "loss": 4.342632293701172,
+ "step": 1528
+ },
+ {
+ "epoch": 21.23755458515284,
+ "grad_norm": 0.015634698793292046,
+ "learning_rate": 0.0006,
+ "loss": 4.338623523712158,
+ "step": 1529
+ },
+ {
+ "epoch": 21.251528384279474,
+ "grad_norm": 0.01449581328779459,
+ "learning_rate": 0.0006,
+ "loss": 4.354330062866211,
+ "step": 1530
+ },
+ {
+ "epoch": 21.265502183406113,
+ "grad_norm": 0.01487616915255785,
+ "learning_rate": 0.0006,
+ "loss": 4.486835479736328,
+ "step": 1531
+ },
+ {
+ "epoch": 21.27947598253275,
+ "grad_norm": 0.016346555203199387,
+ "learning_rate": 0.0006,
+ "loss": 4.407098770141602,
+ "step": 1532
+ },
+ {
+ "epoch": 21.29344978165939,
+ "grad_norm": 0.01682182401418686,
+ "learning_rate": 0.0006,
+ "loss": 4.436809539794922,
+ "step": 1533
+ },
+ {
+ "epoch": 21.307423580786025,
+ "grad_norm": 0.017704004421830177,
+ "learning_rate": 0.0006,
+ "loss": 4.3854780197143555,
+ "step": 1534
+ },
+ {
+ "epoch": 21.321397379912664,
+ "grad_norm": 0.018437420949339867,
+ "learning_rate": 0.0006,
+ "loss": 4.413283824920654,
+ "step": 1535
+ },
+ {
+ "epoch": 21.335371179039303,
+ "grad_norm": 0.016911081969738007,
+ "learning_rate": 0.0006,
+ "loss": 4.357306003570557,
+ "step": 1536
+ },
+ {
+ "epoch": 21.349344978165938,
+ "grad_norm": 0.01642223633825779,
+ "learning_rate": 0.0006,
+ "loss": 4.422322750091553,
+ "step": 1537
+ },
+ {
+ "epoch": 21.363318777292577,
+ "grad_norm": 0.01780877821147442,
+ "learning_rate": 0.0006,
+ "loss": 4.387620449066162,
+ "step": 1538
+ },
+ {
+ "epoch": 21.377292576419215,
+ "grad_norm": 0.01653975248336792,
+ "learning_rate": 0.0006,
+ "loss": 4.327267646789551,
+ "step": 1539
+ },
+ {
+ "epoch": 21.39126637554585,
+ "grad_norm": 0.01572156324982643,
+ "learning_rate": 0.0006,
+ "loss": 4.475979804992676,
+ "step": 1540
+ },
+ {
+ "epoch": 21.40524017467249,
+ "grad_norm": 0.017156066372990608,
+ "learning_rate": 0.0006,
+ "loss": 4.4011125564575195,
+ "step": 1541
+ },
+ {
+ "epoch": 21.419213973799128,
+ "grad_norm": 0.01788829080760479,
+ "learning_rate": 0.0006,
+ "loss": 4.4174089431762695,
+ "step": 1542
+ },
+ {
+ "epoch": 21.433187772925763,
+ "grad_norm": 0.01728985086083412,
+ "learning_rate": 0.0006,
+ "loss": 4.429244041442871,
+ "step": 1543
+ },
+ {
+ "epoch": 21.4471615720524,
+ "grad_norm": 0.016534466296434402,
+ "learning_rate": 0.0006,
+ "loss": 4.326068878173828,
+ "step": 1544
+ },
+ {
+ "epoch": 21.46113537117904,
+ "grad_norm": 0.016582584008574486,
+ "learning_rate": 0.0006,
+ "loss": 4.2975969314575195,
+ "step": 1545
+ },
+ {
+ "epoch": 21.475109170305675,
+ "grad_norm": 0.018034178763628006,
+ "learning_rate": 0.0006,
+ "loss": 4.501319408416748,
+ "step": 1546
+ },
+ {
+ "epoch": 21.489082969432314,
+ "grad_norm": 0.017827702686190605,
+ "learning_rate": 0.0006,
+ "loss": 4.502284526824951,
+ "step": 1547
+ },
+ {
+ "epoch": 21.503056768558952,
+ "grad_norm": 0.017182299867272377,
+ "learning_rate": 0.0006,
+ "loss": 4.380828857421875,
+ "step": 1548
+ },
+ {
+ "epoch": 21.51703056768559,
+ "grad_norm": 0.019101588055491447,
+ "learning_rate": 0.0006,
+ "loss": 4.31564998626709,
+ "step": 1549
+ },
+ {
+ "epoch": 21.531004366812226,
+ "grad_norm": 0.0204677302390337,
+ "learning_rate": 0.0006,
+ "loss": 4.3840131759643555,
+ "step": 1550
+ },
+ {
+ "epoch": 21.544978165938865,
+ "grad_norm": 0.0188993439078331,
+ "learning_rate": 0.0006,
+ "loss": 4.4257731437683105,
+ "step": 1551
+ },
+ {
+ "epoch": 21.558951965065503,
+ "grad_norm": 0.01730230636894703,
+ "learning_rate": 0.0006,
+ "loss": 4.445132255554199,
+ "step": 1552
+ },
+ {
+ "epoch": 21.57292576419214,
+ "grad_norm": 0.01801574043929577,
+ "learning_rate": 0.0006,
+ "loss": 4.427064895629883,
+ "step": 1553
+ },
+ {
+ "epoch": 21.586899563318777,
+ "grad_norm": 0.018417565152049065,
+ "learning_rate": 0.0006,
+ "loss": 4.319595813751221,
+ "step": 1554
+ },
+ {
+ "epoch": 21.600873362445416,
+ "grad_norm": 0.01683160290122032,
+ "learning_rate": 0.0006,
+ "loss": 4.349218368530273,
+ "step": 1555
+ },
+ {
+ "epoch": 21.61484716157205,
+ "grad_norm": 0.016973106190562248,
+ "learning_rate": 0.0006,
+ "loss": 4.375990390777588,
+ "step": 1556
+ },
+ {
+ "epoch": 21.62882096069869,
+ "grad_norm": 0.015688583254814148,
+ "learning_rate": 0.0006,
+ "loss": 4.408170223236084,
+ "step": 1557
+ },
+ {
+ "epoch": 21.64279475982533,
+ "grad_norm": 0.01573043316602707,
+ "learning_rate": 0.0006,
+ "loss": 4.416255950927734,
+ "step": 1558
+ },
+ {
+ "epoch": 21.656768558951963,
+ "grad_norm": 0.01744293048977852,
+ "learning_rate": 0.0006,
+ "loss": 4.477141380310059,
+ "step": 1559
+ },
+ {
+ "epoch": 21.670742358078602,
+ "grad_norm": 0.017503513023257256,
+ "learning_rate": 0.0006,
+ "loss": 4.3367486000061035,
+ "step": 1560
+ },
+ {
+ "epoch": 21.68471615720524,
+ "grad_norm": 0.016592150554060936,
+ "learning_rate": 0.0006,
+ "loss": 4.314785480499268,
+ "step": 1561
+ },
+ {
+ "epoch": 21.69868995633188,
+ "grad_norm": 0.015774084255099297,
+ "learning_rate": 0.0006,
+ "loss": 4.448209285736084,
+ "step": 1562
+ },
+ {
+ "epoch": 21.712663755458514,
+ "grad_norm": 0.015779150649905205,
+ "learning_rate": 0.0006,
+ "loss": 4.392690658569336,
+ "step": 1563
+ },
+ {
+ "epoch": 21.726637554585153,
+ "grad_norm": 0.015392648056149483,
+ "learning_rate": 0.0006,
+ "loss": 4.437741756439209,
+ "step": 1564
+ },
+ {
+ "epoch": 21.74061135371179,
+ "grad_norm": 0.01530836895108223,
+ "learning_rate": 0.0006,
+ "loss": 4.424595832824707,
+ "step": 1565
+ },
+ {
+ "epoch": 21.754585152838427,
+ "grad_norm": 0.015600779093801975,
+ "learning_rate": 0.0006,
+ "loss": 4.335417747497559,
+ "step": 1566
+ },
+ {
+ "epoch": 21.768558951965066,
+ "grad_norm": 0.014646518044173717,
+ "learning_rate": 0.0006,
+ "loss": 4.279081344604492,
+ "step": 1567
+ },
+ {
+ "epoch": 21.782532751091704,
+ "grad_norm": 0.014132864773273468,
+ "learning_rate": 0.0006,
+ "loss": 4.345950126647949,
+ "step": 1568
+ },
+ {
+ "epoch": 21.79650655021834,
+ "grad_norm": 0.01425724383443594,
+ "learning_rate": 0.0006,
+ "loss": 4.277588844299316,
+ "step": 1569
+ },
+ {
+ "epoch": 21.810480349344978,
+ "grad_norm": 0.015396572649478912,
+ "learning_rate": 0.0006,
+ "loss": 4.326196670532227,
+ "step": 1570
+ },
+ {
+ "epoch": 21.824454148471617,
+ "grad_norm": 0.01683669537305832,
+ "learning_rate": 0.0006,
+ "loss": 4.3294172286987305,
+ "step": 1571
+ },
+ {
+ "epoch": 21.83842794759825,
+ "grad_norm": 0.016365647315979004,
+ "learning_rate": 0.0006,
+ "loss": 4.523321628570557,
+ "step": 1572
+ },
+ {
+ "epoch": 21.85240174672489,
+ "grad_norm": 0.017142774537205696,
+ "learning_rate": 0.0006,
+ "loss": 4.475842475891113,
+ "step": 1573
+ },
+ {
+ "epoch": 21.86637554585153,
+ "grad_norm": 0.015950839966535568,
+ "learning_rate": 0.0006,
+ "loss": 4.397187232971191,
+ "step": 1574
+ },
+ {
+ "epoch": 21.880349344978168,
+ "grad_norm": 0.016180362552404404,
+ "learning_rate": 0.0006,
+ "loss": 4.414778709411621,
+ "step": 1575
+ },
+ {
+ "epoch": 21.894323144104803,
+ "grad_norm": 0.017615556716918945,
+ "learning_rate": 0.0006,
+ "loss": 4.348567962646484,
+ "step": 1576
+ },
+ {
+ "epoch": 21.90829694323144,
+ "grad_norm": 0.01896277628839016,
+ "learning_rate": 0.0006,
+ "loss": 4.431762218475342,
+ "step": 1577
+ },
+ {
+ "epoch": 21.92227074235808,
+ "grad_norm": 0.01990971714258194,
+ "learning_rate": 0.0006,
+ "loss": 4.431285858154297,
+ "step": 1578
+ },
+ {
+ "epoch": 21.936244541484715,
+ "grad_norm": 0.0201258547604084,
+ "learning_rate": 0.0006,
+ "loss": 4.309872627258301,
+ "step": 1579
+ },
+ {
+ "epoch": 21.950218340611354,
+ "grad_norm": 0.019620198756456375,
+ "learning_rate": 0.0006,
+ "loss": 4.484044551849365,
+ "step": 1580
+ },
+ {
+ "epoch": 21.964192139737992,
+ "grad_norm": 0.01771123707294464,
+ "learning_rate": 0.0006,
+ "loss": 4.368093013763428,
+ "step": 1581
+ },
+ {
+ "epoch": 21.978165938864628,
+ "grad_norm": 0.017501654103398323,
+ "learning_rate": 0.0006,
+ "loss": 4.372651100158691,
+ "step": 1582
+ },
+ {
+ "epoch": 21.992139737991266,
+ "grad_norm": 0.01693608984351158,
+ "learning_rate": 0.0006,
+ "loss": 4.379167556762695,
+ "step": 1583
+ },
+ {
+ "epoch": 22.0,
+ "grad_norm": 0.02002849616110325,
+ "learning_rate": 0.0006,
+ "loss": 4.329627990722656,
+ "step": 1584
+ },
+ {
+ "epoch": 22.0,
+ "eval_loss": 4.723696231842041,
+ "eval_runtime": 56.1758,
+ "eval_samples_per_second": 43.471,
+ "eval_steps_per_second": 1.371,
+ "step": 1584
+ },
+ {
+ "epoch": 22.01397379912664,
+ "grad_norm": 0.02030305378139019,
+ "learning_rate": 0.0006,
+ "loss": 4.355566024780273,
+ "step": 1585
+ },
+ {
+ "epoch": 22.027947598253274,
+ "grad_norm": 0.018638934940099716,
+ "learning_rate": 0.0006,
+ "loss": 4.314009189605713,
+ "step": 1586
+ },
+ {
+ "epoch": 22.041921397379912,
+ "grad_norm": 0.019917072728276253,
+ "learning_rate": 0.0006,
+ "loss": 4.367525577545166,
+ "step": 1587
+ },
+ {
+ "epoch": 22.05589519650655,
+ "grad_norm": 0.020785167813301086,
+ "learning_rate": 0.0006,
+ "loss": 4.375561714172363,
+ "step": 1588
+ },
+ {
+ "epoch": 22.069868995633186,
+ "grad_norm": 0.021214453503489494,
+ "learning_rate": 0.0006,
+ "loss": 4.418305397033691,
+ "step": 1589
+ },
+ {
+ "epoch": 22.083842794759825,
+ "grad_norm": 0.019051678478717804,
+ "learning_rate": 0.0006,
+ "loss": 4.215126991271973,
+ "step": 1590
+ },
+ {
+ "epoch": 22.097816593886463,
+ "grad_norm": 0.02323366142809391,
+ "learning_rate": 0.0006,
+ "loss": 4.306807994842529,
+ "step": 1591
+ },
+ {
+ "epoch": 22.111790393013102,
+ "grad_norm": 0.026186738163232803,
+ "learning_rate": 0.0006,
+ "loss": 4.408508777618408,
+ "step": 1592
+ },
+ {
+ "epoch": 22.125764192139737,
+ "grad_norm": 0.027361227199435234,
+ "learning_rate": 0.0006,
+ "loss": 4.329322814941406,
+ "step": 1593
+ },
+ {
+ "epoch": 22.139737991266376,
+ "grad_norm": 0.026564646512269974,
+ "learning_rate": 0.0006,
+ "loss": 4.346642017364502,
+ "step": 1594
+ },
+ {
+ "epoch": 22.153711790393015,
+ "grad_norm": 0.025073667988181114,
+ "learning_rate": 0.0006,
+ "loss": 4.3984198570251465,
+ "step": 1595
+ },
+ {
+ "epoch": 22.16768558951965,
+ "grad_norm": 0.02347794733941555,
+ "learning_rate": 0.0006,
+ "loss": 4.352538108825684,
+ "step": 1596
+ },
+ {
+ "epoch": 22.18165938864629,
+ "grad_norm": 0.02203468047082424,
+ "learning_rate": 0.0006,
+ "loss": 4.298148155212402,
+ "step": 1597
+ },
+ {
+ "epoch": 22.195633187772927,
+ "grad_norm": 0.020621957257390022,
+ "learning_rate": 0.0006,
+ "loss": 4.480309963226318,
+ "step": 1598
+ },
+ {
+ "epoch": 22.209606986899562,
+ "grad_norm": 0.023710055276751518,
+ "learning_rate": 0.0006,
+ "loss": 4.339317321777344,
+ "step": 1599
+ },
+ {
+ "epoch": 22.2235807860262,
+ "grad_norm": 0.0233482476323843,
+ "learning_rate": 0.0006,
+ "loss": 4.436896324157715,
+ "step": 1600
+ },
+ {
+ "epoch": 22.23755458515284,
+ "grad_norm": 0.020122403278946877,
+ "learning_rate": 0.0006,
+ "loss": 4.236002445220947,
+ "step": 1601
+ },
+ {
+ "epoch": 22.251528384279474,
+ "grad_norm": 0.01957266591489315,
+ "learning_rate": 0.0006,
+ "loss": 4.497328281402588,
+ "step": 1602
+ },
+ {
+ "epoch": 22.265502183406113,
+ "grad_norm": 0.018964895978569984,
+ "learning_rate": 0.0006,
+ "loss": 4.36360502243042,
+ "step": 1603
+ },
+ {
+ "epoch": 22.27947598253275,
+ "grad_norm": 0.018902769312262535,
+ "learning_rate": 0.0006,
+ "loss": 4.427093505859375,
+ "step": 1604
+ },
+ {
+ "epoch": 22.29344978165939,
+ "grad_norm": 0.018918119370937347,
+ "learning_rate": 0.0006,
+ "loss": 4.294950485229492,
+ "step": 1605
+ },
+ {
+ "epoch": 22.307423580786025,
+ "grad_norm": 0.019315775483846664,
+ "learning_rate": 0.0006,
+ "loss": 4.427952289581299,
+ "step": 1606
+ },
+ {
+ "epoch": 22.321397379912664,
+ "grad_norm": 0.018350522965192795,
+ "learning_rate": 0.0006,
+ "loss": 4.272945404052734,
+ "step": 1607
+ },
+ {
+ "epoch": 22.335371179039303,
+ "grad_norm": 0.01759534701704979,
+ "learning_rate": 0.0006,
+ "loss": 4.348023891448975,
+ "step": 1608
+ },
+ {
+ "epoch": 22.349344978165938,
+ "grad_norm": 0.01646057888865471,
+ "learning_rate": 0.0006,
+ "loss": 4.278518199920654,
+ "step": 1609
+ },
+ {
+ "epoch": 22.363318777292577,
+ "grad_norm": 0.015875060111284256,
+ "learning_rate": 0.0006,
+ "loss": 4.423527240753174,
+ "step": 1610
+ },
+ {
+ "epoch": 22.377292576419215,
+ "grad_norm": 0.01638529635965824,
+ "learning_rate": 0.0006,
+ "loss": 4.321723937988281,
+ "step": 1611
+ },
+ {
+ "epoch": 22.39126637554585,
+ "grad_norm": 0.016942061483860016,
+ "learning_rate": 0.0006,
+ "loss": 4.346164703369141,
+ "step": 1612
+ },
+ {
+ "epoch": 22.40524017467249,
+ "grad_norm": 0.015209621749818325,
+ "learning_rate": 0.0006,
+ "loss": 4.434340000152588,
+ "step": 1613
+ },
+ {
+ "epoch": 22.419213973799128,
+ "grad_norm": 0.014423009008169174,
+ "learning_rate": 0.0006,
+ "loss": 4.342560768127441,
+ "step": 1614
+ },
+ {
+ "epoch": 22.433187772925763,
+ "grad_norm": 0.014369525946676731,
+ "learning_rate": 0.0006,
+ "loss": 4.296586990356445,
+ "step": 1615
+ },
+ {
+ "epoch": 22.4471615720524,
+ "grad_norm": 0.01392540242522955,
+ "learning_rate": 0.0006,
+ "loss": 4.433176040649414,
+ "step": 1616
+ },
+ {
+ "epoch": 22.46113537117904,
+ "grad_norm": 0.013206899166107178,
+ "learning_rate": 0.0006,
+ "loss": 4.32962703704834,
+ "step": 1617
+ },
+ {
+ "epoch": 22.475109170305675,
+ "grad_norm": 0.013427077792584896,
+ "learning_rate": 0.0006,
+ "loss": 4.420437335968018,
+ "step": 1618
+ },
+ {
+ "epoch": 22.489082969432314,
+ "grad_norm": 0.014194196090102196,
+ "learning_rate": 0.0006,
+ "loss": 4.410478591918945,
+ "step": 1619
+ },
+ {
+ "epoch": 22.503056768558952,
+ "grad_norm": 0.01362364087253809,
+ "learning_rate": 0.0006,
+ "loss": 4.338932037353516,
+ "step": 1620
+ },
+ {
+ "epoch": 22.51703056768559,
+ "grad_norm": 0.014299589209258556,
+ "learning_rate": 0.0006,
+ "loss": 4.34089994430542,
+ "step": 1621
+ },
+ {
+ "epoch": 22.531004366812226,
+ "grad_norm": 0.014061295427381992,
+ "learning_rate": 0.0006,
+ "loss": 4.29201602935791,
+ "step": 1622
+ },
+ {
+ "epoch": 22.544978165938865,
+ "grad_norm": 0.013945229351520538,
+ "learning_rate": 0.0006,
+ "loss": 4.326052188873291,
+ "step": 1623
+ },
+ {
+ "epoch": 22.558951965065503,
+ "grad_norm": 0.01387068536132574,
+ "learning_rate": 0.0006,
+ "loss": 4.393521308898926,
+ "step": 1624
+ },
+ {
+ "epoch": 22.57292576419214,
+ "grad_norm": 0.013931898400187492,
+ "learning_rate": 0.0006,
+ "loss": 4.3739752769470215,
+ "step": 1625
+ },
+ {
+ "epoch": 22.586899563318777,
+ "grad_norm": 0.01497740764170885,
+ "learning_rate": 0.0006,
+ "loss": 4.3991618156433105,
+ "step": 1626
+ },
+ {
+ "epoch": 22.600873362445416,
+ "grad_norm": 0.015842905268073082,
+ "learning_rate": 0.0006,
+ "loss": 4.405340671539307,
+ "step": 1627
+ },
+ {
+ "epoch": 22.61484716157205,
+ "grad_norm": 0.01580633595585823,
+ "learning_rate": 0.0006,
+ "loss": 4.280102729797363,
+ "step": 1628
+ },
+ {
+ "epoch": 22.62882096069869,
+ "grad_norm": 0.016567382961511612,
+ "learning_rate": 0.0006,
+ "loss": 4.407657146453857,
+ "step": 1629
+ },
+ {
+ "epoch": 22.64279475982533,
+ "grad_norm": 0.01648622378706932,
+ "learning_rate": 0.0006,
+ "loss": 4.392590522766113,
+ "step": 1630
+ },
+ {
+ "epoch": 22.656768558951963,
+ "grad_norm": 0.017066307365894318,
+ "learning_rate": 0.0006,
+ "loss": 4.324862003326416,
+ "step": 1631
+ },
+ {
+ "epoch": 22.670742358078602,
+ "grad_norm": 0.01637062057852745,
+ "learning_rate": 0.0006,
+ "loss": 4.34598445892334,
+ "step": 1632
+ },
+ {
+ "epoch": 22.68471615720524,
+ "grad_norm": 0.018038960173726082,
+ "learning_rate": 0.0006,
+ "loss": 4.320711612701416,
+ "step": 1633
+ },
+ {
+ "epoch": 22.69868995633188,
+ "grad_norm": 0.018096428364515305,
+ "learning_rate": 0.0006,
+ "loss": 4.536527156829834,
+ "step": 1634
+ },
+ {
+ "epoch": 22.712663755458514,
+ "grad_norm": 0.017889291048049927,
+ "learning_rate": 0.0006,
+ "loss": 4.388131141662598,
+ "step": 1635
+ },
+ {
+ "epoch": 22.726637554585153,
+ "grad_norm": 0.018771521747112274,
+ "learning_rate": 0.0006,
+ "loss": 4.357668876647949,
+ "step": 1636
+ },
+ {
+ "epoch": 22.74061135371179,
+ "grad_norm": 0.02100873365998268,
+ "learning_rate": 0.0006,
+ "loss": 4.293972015380859,
+ "step": 1637
+ },
+ {
+ "epoch": 22.754585152838427,
+ "grad_norm": 0.017715785652399063,
+ "learning_rate": 0.0006,
+ "loss": 4.349849224090576,
+ "step": 1638
+ },
+ {
+ "epoch": 22.768558951965066,
+ "grad_norm": 0.015546333976089954,
+ "learning_rate": 0.0006,
+ "loss": 4.311728000640869,
+ "step": 1639
+ },
+ {
+ "epoch": 22.782532751091704,
+ "grad_norm": 0.01605990156531334,
+ "learning_rate": 0.0006,
+ "loss": 4.499123573303223,
+ "step": 1640
+ },
+ {
+ "epoch": 22.79650655021834,
+ "grad_norm": 0.016642142087221146,
+ "learning_rate": 0.0006,
+ "loss": 4.4319257736206055,
+ "step": 1641
+ },
+ {
+ "epoch": 22.810480349344978,
+ "grad_norm": 0.016120387241244316,
+ "learning_rate": 0.0006,
+ "loss": 4.4373650550842285,
+ "step": 1642
+ },
+ {
+ "epoch": 22.824454148471617,
+ "grad_norm": 0.015174890868365765,
+ "learning_rate": 0.0006,
+ "loss": 4.314082145690918,
+ "step": 1643
+ },
+ {
+ "epoch": 22.83842794759825,
+ "grad_norm": 0.01551714539527893,
+ "learning_rate": 0.0006,
+ "loss": 4.360363960266113,
+ "step": 1644
+ },
+ {
+ "epoch": 22.85240174672489,
+ "grad_norm": 0.014966742135584354,
+ "learning_rate": 0.0006,
+ "loss": 4.346253395080566,
+ "step": 1645
+ },
+ {
+ "epoch": 22.86637554585153,
+ "grad_norm": 0.014288028702139854,
+ "learning_rate": 0.0006,
+ "loss": 4.464366436004639,
+ "step": 1646
+ },
+ {
+ "epoch": 22.880349344978168,
+ "grad_norm": 0.015291682444512844,
+ "learning_rate": 0.0006,
+ "loss": 4.504101753234863,
+ "step": 1647
+ },
+ {
+ "epoch": 22.894323144104803,
+ "grad_norm": 0.014534426853060722,
+ "learning_rate": 0.0006,
+ "loss": 4.462172031402588,
+ "step": 1648
+ },
+ {
+ "epoch": 22.90829694323144,
+ "grad_norm": 0.014129819348454475,
+ "learning_rate": 0.0006,
+ "loss": 4.350126266479492,
+ "step": 1649
+ },
+ {
+ "epoch": 22.92227074235808,
+ "grad_norm": 0.014932668767869473,
+ "learning_rate": 0.0006,
+ "loss": 4.374392032623291,
+ "step": 1650
+ },
+ {
+ "epoch": 22.936244541484715,
+ "grad_norm": 0.014647853560745716,
+ "learning_rate": 0.0006,
+ "loss": 4.252264499664307,
+ "step": 1651
+ },
+ {
+ "epoch": 22.950218340611354,
+ "grad_norm": 0.013553360477089882,
+ "learning_rate": 0.0006,
+ "loss": 4.379729270935059,
+ "step": 1652
+ },
+ {
+ "epoch": 22.964192139737992,
+ "grad_norm": 0.014606194570660591,
+ "learning_rate": 0.0006,
+ "loss": 4.363372802734375,
+ "step": 1653
+ },
+ {
+ "epoch": 22.978165938864628,
+ "grad_norm": 0.015000631101429462,
+ "learning_rate": 0.0006,
+ "loss": 4.408048629760742,
+ "step": 1654
+ },
+ {
+ "epoch": 22.992139737991266,
+ "grad_norm": 0.015370816923677921,
+ "learning_rate": 0.0006,
+ "loss": 4.297451019287109,
+ "step": 1655
+ },
+ {
+ "epoch": 23.0,
+ "grad_norm": 0.01709144562482834,
+ "learning_rate": 0.0006,
+ "loss": 4.294776916503906,
+ "step": 1656
+ },
+ {
+ "epoch": 23.0,
+ "eval_loss": 4.678618907928467,
+ "eval_runtime": 56.1668,
+ "eval_samples_per_second": 43.478,
+ "eval_steps_per_second": 1.371,
+ "step": 1656
+ },
+ {
+ "epoch": 23.01397379912664,
+ "grad_norm": 0.01727290451526642,
+ "learning_rate": 0.0006,
+ "loss": 4.311701774597168,
+ "step": 1657
+ },
+ {
+ "epoch": 23.027947598253274,
+ "grad_norm": 0.020645759999752045,
+ "learning_rate": 0.0006,
+ "loss": 4.429897308349609,
+ "step": 1658
+ },
+ {
+ "epoch": 23.041921397379912,
+ "grad_norm": 0.02204226516187191,
+ "learning_rate": 0.0006,
+ "loss": 4.365151882171631,
+ "step": 1659
+ },
+ {
+ "epoch": 23.05589519650655,
+ "grad_norm": 0.023069269955158234,
+ "learning_rate": 0.0006,
+ "loss": 4.3710832595825195,
+ "step": 1660
+ },
+ {
+ "epoch": 23.069868995633186,
+ "grad_norm": 0.024151179939508438,
+ "learning_rate": 0.0006,
+ "loss": 4.403108596801758,
+ "step": 1661
+ },
+ {
+ "epoch": 23.083842794759825,
+ "grad_norm": 0.02511233650147915,
+ "learning_rate": 0.0006,
+ "loss": 4.414128303527832,
+ "step": 1662
+ },
+ {
+ "epoch": 23.097816593886463,
+ "grad_norm": 0.025802113115787506,
+ "learning_rate": 0.0006,
+ "loss": 4.388265609741211,
+ "step": 1663
+ },
+ {
+ "epoch": 23.111790393013102,
+ "grad_norm": 0.02377348765730858,
+ "learning_rate": 0.0006,
+ "loss": 4.357505798339844,
+ "step": 1664
+ },
+ {
+ "epoch": 23.125764192139737,
+ "grad_norm": 0.020876651629805565,
+ "learning_rate": 0.0006,
+ "loss": 4.23043966293335,
+ "step": 1665
+ },
+ {
+ "epoch": 23.139737991266376,
+ "grad_norm": 0.01959727331995964,
+ "learning_rate": 0.0006,
+ "loss": 4.280970096588135,
+ "step": 1666
+ },
+ {
+ "epoch": 23.153711790393015,
+ "grad_norm": 0.018678616732358932,
+ "learning_rate": 0.0006,
+ "loss": 4.346716403961182,
+ "step": 1667
+ },
+ {
+ "epoch": 23.16768558951965,
+ "grad_norm": 0.02108944021165371,
+ "learning_rate": 0.0006,
+ "loss": 4.277904510498047,
+ "step": 1668
+ },
+ {
+ "epoch": 23.18165938864629,
+ "grad_norm": 0.020221566781401634,
+ "learning_rate": 0.0006,
+ "loss": 4.367693901062012,
+ "step": 1669
+ },
+ {
+ "epoch": 23.195633187772927,
+ "grad_norm": 0.019599711522459984,
+ "learning_rate": 0.0006,
+ "loss": 4.257092475891113,
+ "step": 1670
+ },
+ {
+ "epoch": 23.209606986899562,
+ "grad_norm": 0.019833361729979515,
+ "learning_rate": 0.0006,
+ "loss": 4.540737628936768,
+ "step": 1671
+ },
+ {
+ "epoch": 23.2235807860262,
+ "grad_norm": 0.020352095365524292,
+ "learning_rate": 0.0006,
+ "loss": 4.360923767089844,
+ "step": 1672
+ },
+ {
+ "epoch": 23.23755458515284,
+ "grad_norm": 0.01939748041331768,
+ "learning_rate": 0.0006,
+ "loss": 4.393110275268555,
+ "step": 1673
+ },
+ {
+ "epoch": 23.251528384279474,
+ "grad_norm": 0.01884354278445244,
+ "learning_rate": 0.0006,
+ "loss": 4.383938789367676,
+ "step": 1674
+ },
+ {
+ "epoch": 23.265502183406113,
+ "grad_norm": 0.02008470520377159,
+ "learning_rate": 0.0006,
+ "loss": 4.346344947814941,
+ "step": 1675
+ },
+ {
+ "epoch": 23.27947598253275,
+ "grad_norm": 0.019065195694565773,
+ "learning_rate": 0.0006,
+ "loss": 4.36611795425415,
+ "step": 1676
+ },
+ {
+ "epoch": 23.29344978165939,
+ "grad_norm": 0.01938541606068611,
+ "learning_rate": 0.0006,
+ "loss": 4.4009599685668945,
+ "step": 1677
+ },
+ {
+ "epoch": 23.307423580786025,
+ "grad_norm": 0.019044723361730576,
+ "learning_rate": 0.0006,
+ "loss": 4.340169906616211,
+ "step": 1678
+ },
+ {
+ "epoch": 23.321397379912664,
+ "grad_norm": 0.018850378692150116,
+ "learning_rate": 0.0006,
+ "loss": 4.37674617767334,
+ "step": 1679
+ },
+ {
+ "epoch": 23.335371179039303,
+ "grad_norm": 0.019076095893979073,
+ "learning_rate": 0.0006,
+ "loss": 4.226874351501465,
+ "step": 1680
+ },
+ {
+ "epoch": 23.349344978165938,
+ "grad_norm": 0.017543038353323936,
+ "learning_rate": 0.0006,
+ "loss": 4.337751388549805,
+ "step": 1681
+ },
+ {
+ "epoch": 23.363318777292577,
+ "grad_norm": 0.017170218750834465,
+ "learning_rate": 0.0006,
+ "loss": 4.282512664794922,
+ "step": 1682
+ },
+ {
+ "epoch": 23.377292576419215,
+ "grad_norm": 0.01574796624481678,
+ "learning_rate": 0.0006,
+ "loss": 4.30638313293457,
+ "step": 1683
+ },
+ {
+ "epoch": 23.39126637554585,
+ "grad_norm": 0.015160846523940563,
+ "learning_rate": 0.0006,
+ "loss": 4.336716175079346,
+ "step": 1684
+ },
+ {
+ "epoch": 23.40524017467249,
+ "grad_norm": 0.01534526702016592,
+ "learning_rate": 0.0006,
+ "loss": 4.2174787521362305,
+ "step": 1685
+ },
+ {
+ "epoch": 23.419213973799128,
+ "grad_norm": 0.015289999544620514,
+ "learning_rate": 0.0006,
+ "loss": 4.272546291351318,
+ "step": 1686
+ },
+ {
+ "epoch": 23.433187772925763,
+ "grad_norm": 0.014894185587763786,
+ "learning_rate": 0.0006,
+ "loss": 4.296045303344727,
+ "step": 1687
+ },
+ {
+ "epoch": 23.4471615720524,
+ "grad_norm": 0.014869502745568752,
+ "learning_rate": 0.0006,
+ "loss": 4.425461769104004,
+ "step": 1688
+ },
+ {
+ "epoch": 23.46113537117904,
+ "grad_norm": 0.01474483497440815,
+ "learning_rate": 0.0006,
+ "loss": 4.361867427825928,
+ "step": 1689
+ },
+ {
+ "epoch": 23.475109170305675,
+ "grad_norm": 0.015299245715141296,
+ "learning_rate": 0.0006,
+ "loss": 4.364523410797119,
+ "step": 1690
+ },
+ {
+ "epoch": 23.489082969432314,
+ "grad_norm": 0.015201340429484844,
+ "learning_rate": 0.0006,
+ "loss": 4.285182952880859,
+ "step": 1691
+ },
+ {
+ "epoch": 23.503056768558952,
+ "grad_norm": 0.016057293862104416,
+ "learning_rate": 0.0006,
+ "loss": 4.284873962402344,
+ "step": 1692
+ },
+ {
+ "epoch": 23.51703056768559,
+ "grad_norm": 0.015871062874794006,
+ "learning_rate": 0.0006,
+ "loss": 4.326470851898193,
+ "step": 1693
+ },
+ {
+ "epoch": 23.531004366812226,
+ "grad_norm": 0.01586288772523403,
+ "learning_rate": 0.0006,
+ "loss": 4.324883460998535,
+ "step": 1694
+ },
+ {
+ "epoch": 23.544978165938865,
+ "grad_norm": 0.016092827543616295,
+ "learning_rate": 0.0006,
+ "loss": 4.3258867263793945,
+ "step": 1695
+ },
+ {
+ "epoch": 23.558951965065503,
+ "grad_norm": 0.01582016795873642,
+ "learning_rate": 0.0006,
+ "loss": 4.358892440795898,
+ "step": 1696
+ },
+ {
+ "epoch": 23.57292576419214,
+ "grad_norm": 0.01688452623784542,
+ "learning_rate": 0.0006,
+ "loss": 4.280842304229736,
+ "step": 1697
+ },
+ {
+ "epoch": 23.586899563318777,
+ "grad_norm": 0.015865886583924294,
+ "learning_rate": 0.0006,
+ "loss": 4.245567321777344,
+ "step": 1698
+ },
+ {
+ "epoch": 23.600873362445416,
+ "grad_norm": 0.015303662046790123,
+ "learning_rate": 0.0006,
+ "loss": 4.383112907409668,
+ "step": 1699
+ },
+ {
+ "epoch": 23.61484716157205,
+ "grad_norm": 0.015649516135454178,
+ "learning_rate": 0.0006,
+ "loss": 4.519082069396973,
+ "step": 1700
+ },
+ {
+ "epoch": 23.62882096069869,
+ "grad_norm": 0.0167120099067688,
+ "learning_rate": 0.0006,
+ "loss": 4.447847366333008,
+ "step": 1701
+ },
+ {
+ "epoch": 23.64279475982533,
+ "grad_norm": 0.016368716955184937,
+ "learning_rate": 0.0006,
+ "loss": 4.434186935424805,
+ "step": 1702
+ },
+ {
+ "epoch": 23.656768558951963,
+ "grad_norm": 0.017379503697156906,
+ "learning_rate": 0.0006,
+ "loss": 4.317594528198242,
+ "step": 1703
+ },
+ {
+ "epoch": 23.670742358078602,
+ "grad_norm": 0.018573811277747154,
+ "learning_rate": 0.0006,
+ "loss": 4.205793857574463,
+ "step": 1704
+ },
+ {
+ "epoch": 23.68471615720524,
+ "grad_norm": 0.01749090477824211,
+ "learning_rate": 0.0006,
+ "loss": 4.3985700607299805,
+ "step": 1705
+ },
+ {
+ "epoch": 23.69868995633188,
+ "grad_norm": 0.0165668074041605,
+ "learning_rate": 0.0006,
+ "loss": 4.329226970672607,
+ "step": 1706
+ },
+ {
+ "epoch": 23.712663755458514,
+ "grad_norm": 0.016913846135139465,
+ "learning_rate": 0.0006,
+ "loss": 4.388485908508301,
+ "step": 1707
+ },
+ {
+ "epoch": 23.726637554585153,
+ "grad_norm": 0.016813859343528748,
+ "learning_rate": 0.0006,
+ "loss": 4.3395915031433105,
+ "step": 1708
+ },
+ {
+ "epoch": 23.74061135371179,
+ "grad_norm": 0.01676975190639496,
+ "learning_rate": 0.0006,
+ "loss": 4.254745960235596,
+ "step": 1709
+ },
+ {
+ "epoch": 23.754585152838427,
+ "grad_norm": 0.016370350494980812,
+ "learning_rate": 0.0006,
+ "loss": 4.2612223625183105,
+ "step": 1710
+ },
+ {
+ "epoch": 23.768558951965066,
+ "grad_norm": 0.01696198247373104,
+ "learning_rate": 0.0006,
+ "loss": 4.327524662017822,
+ "step": 1711
+ },
+ {
+ "epoch": 23.782532751091704,
+ "grad_norm": 0.0189108457416296,
+ "learning_rate": 0.0006,
+ "loss": 4.325839519500732,
+ "step": 1712
+ },
+ {
+ "epoch": 23.79650655021834,
+ "grad_norm": 0.01862351782619953,
+ "learning_rate": 0.0006,
+ "loss": 4.333662509918213,
+ "step": 1713
+ },
+ {
+ "epoch": 23.810480349344978,
+ "grad_norm": 0.01655014418065548,
+ "learning_rate": 0.0006,
+ "loss": 4.39305305480957,
+ "step": 1714
+ },
+ {
+ "epoch": 23.824454148471617,
+ "grad_norm": 0.015899376943707466,
+ "learning_rate": 0.0006,
+ "loss": 4.390053749084473,
+ "step": 1715
+ },
+ {
+ "epoch": 23.83842794759825,
+ "grad_norm": 0.016519056633114815,
+ "learning_rate": 0.0006,
+ "loss": 4.255903244018555,
+ "step": 1716
+ },
+ {
+ "epoch": 23.85240174672489,
+ "grad_norm": 0.0167539119720459,
+ "learning_rate": 0.0006,
+ "loss": 4.255727767944336,
+ "step": 1717
+ },
+ {
+ "epoch": 23.86637554585153,
+ "grad_norm": 0.0171353816986084,
+ "learning_rate": 0.0006,
+ "loss": 4.376740455627441,
+ "step": 1718
+ },
+ {
+ "epoch": 23.880349344978168,
+ "grad_norm": 0.016719117760658264,
+ "learning_rate": 0.0006,
+ "loss": 4.3376688957214355,
+ "step": 1719
+ },
+ {
+ "epoch": 23.894323144104803,
+ "grad_norm": 0.01561494916677475,
+ "learning_rate": 0.0006,
+ "loss": 4.3345746994018555,
+ "step": 1720
+ },
+ {
+ "epoch": 23.90829694323144,
+ "grad_norm": 0.016303174197673798,
+ "learning_rate": 0.0006,
+ "loss": 4.332371234893799,
+ "step": 1721
+ },
+ {
+ "epoch": 23.92227074235808,
+ "grad_norm": 0.016722865402698517,
+ "learning_rate": 0.0006,
+ "loss": 4.267125129699707,
+ "step": 1722
+ },
+ {
+ "epoch": 23.936244541484715,
+ "grad_norm": 0.017072124406695366,
+ "learning_rate": 0.0006,
+ "loss": 4.322751998901367,
+ "step": 1723
+ },
+ {
+ "epoch": 23.950218340611354,
+ "grad_norm": 0.01699932850897312,
+ "learning_rate": 0.0006,
+ "loss": 4.378422737121582,
+ "step": 1724
+ },
+ {
+ "epoch": 23.964192139737992,
+ "grad_norm": 0.01742720976471901,
+ "learning_rate": 0.0006,
+ "loss": 4.345083236694336,
+ "step": 1725
+ },
+ {
+ "epoch": 23.978165938864628,
+ "grad_norm": 0.018797501921653748,
+ "learning_rate": 0.0006,
+ "loss": 4.290443420410156,
+ "step": 1726
+ },
+ {
+ "epoch": 23.992139737991266,
+ "grad_norm": 0.018090683966875076,
+ "learning_rate": 0.0006,
+ "loss": 4.377580642700195,
+ "step": 1727
+ },
+ {
+ "epoch": 24.0,
+ "grad_norm": 0.017305459827184677,
+ "learning_rate": 0.0006,
+ "loss": 4.364426612854004,
+ "step": 1728
+ },
+ {
+ "epoch": 24.0,
+ "eval_loss": 4.647762298583984,
+ "eval_runtime": 56.5402,
+ "eval_samples_per_second": 43.191,
+ "eval_steps_per_second": 1.362,
+ "step": 1728
+ },
+ {
+ "epoch": 24.01397379912664,
+ "grad_norm": 0.016493607312440872,
+ "learning_rate": 0.0006,
+ "loss": 4.252255916595459,
+ "step": 1729
+ },
+ {
+ "epoch": 24.027947598253274,
+ "grad_norm": 0.016529450193047523,
+ "learning_rate": 0.0006,
+ "loss": 4.352746963500977,
+ "step": 1730
+ },
+ {
+ "epoch": 24.041921397379912,
+ "grad_norm": 0.016940191388130188,
+ "learning_rate": 0.0006,
+ "loss": 4.301870822906494,
+ "step": 1731
+ },
+ {
+ "epoch": 24.05589519650655,
+ "grad_norm": 0.01752558909356594,
+ "learning_rate": 0.0006,
+ "loss": 4.2762451171875,
+ "step": 1732
+ },
+ {
+ "epoch": 24.069868995633186,
+ "grad_norm": 0.017112884670495987,
+ "learning_rate": 0.0006,
+ "loss": 4.2659149169921875,
+ "step": 1733
+ },
+ {
+ "epoch": 24.083842794759825,
+ "grad_norm": 0.01900586113333702,
+ "learning_rate": 0.0006,
+ "loss": 4.306058883666992,
+ "step": 1734
+ },
+ {
+ "epoch": 24.097816593886463,
+ "grad_norm": 0.019447680562734604,
+ "learning_rate": 0.0006,
+ "loss": 4.450128555297852,
+ "step": 1735
+ },
+ {
+ "epoch": 24.111790393013102,
+ "grad_norm": 0.01948186755180359,
+ "learning_rate": 0.0006,
+ "loss": 4.202869415283203,
+ "step": 1736
+ },
+ {
+ "epoch": 24.125764192139737,
+ "grad_norm": 0.01716383546590805,
+ "learning_rate": 0.0006,
+ "loss": 4.318876266479492,
+ "step": 1737
+ },
+ {
+ "epoch": 24.139737991266376,
+ "grad_norm": 0.017049988731741905,
+ "learning_rate": 0.0006,
+ "loss": 4.237238883972168,
+ "step": 1738
+ },
+ {
+ "epoch": 24.153711790393015,
+ "grad_norm": 0.015169923193752766,
+ "learning_rate": 0.0006,
+ "loss": 4.289426803588867,
+ "step": 1739
+ },
+ {
+ "epoch": 24.16768558951965,
+ "grad_norm": 0.014760667458176613,
+ "learning_rate": 0.0006,
+ "loss": 4.356924057006836,
+ "step": 1740
+ },
+ {
+ "epoch": 24.18165938864629,
+ "grad_norm": 0.015127750113606453,
+ "learning_rate": 0.0006,
+ "loss": 4.142916679382324,
+ "step": 1741
+ },
+ {
+ "epoch": 24.195633187772927,
+ "grad_norm": 0.01549555640667677,
+ "learning_rate": 0.0006,
+ "loss": 4.345563888549805,
+ "step": 1742
+ },
+ {
+ "epoch": 24.209606986899562,
+ "grad_norm": 0.016393091529607773,
+ "learning_rate": 0.0006,
+ "loss": 4.417495250701904,
+ "step": 1743
+ },
+ {
+ "epoch": 24.2235807860262,
+ "grad_norm": 0.01785266026854515,
+ "learning_rate": 0.0006,
+ "loss": 4.3553547859191895,
+ "step": 1744
+ },
+ {
+ "epoch": 24.23755458515284,
+ "grad_norm": 0.017574617639183998,
+ "learning_rate": 0.0006,
+ "loss": 4.241024494171143,
+ "step": 1745
+ },
+ {
+ "epoch": 24.251528384279474,
+ "grad_norm": 0.019060306251049042,
+ "learning_rate": 0.0006,
+ "loss": 4.3685150146484375,
+ "step": 1746
+ },
+ {
+ "epoch": 24.265502183406113,
+ "grad_norm": 0.0214633010327816,
+ "learning_rate": 0.0006,
+ "loss": 4.323866844177246,
+ "step": 1747
+ },
+ {
+ "epoch": 24.27947598253275,
+ "grad_norm": 0.021055003628134727,
+ "learning_rate": 0.0006,
+ "loss": 4.278757095336914,
+ "step": 1748
+ },
+ {
+ "epoch": 24.29344978165939,
+ "grad_norm": 0.018618572503328323,
+ "learning_rate": 0.0006,
+ "loss": 4.28849983215332,
+ "step": 1749
+ },
+ {
+ "epoch": 24.307423580786025,
+ "grad_norm": 0.018666435033082962,
+ "learning_rate": 0.0006,
+ "loss": 4.368590354919434,
+ "step": 1750
+ },
+ {
+ "epoch": 24.321397379912664,
+ "grad_norm": 0.020626146346330643,
+ "learning_rate": 0.0006,
+ "loss": 4.2951836585998535,
+ "step": 1751
+ },
+ {
+ "epoch": 24.335371179039303,
+ "grad_norm": 0.01929759792983532,
+ "learning_rate": 0.0006,
+ "loss": 4.3664703369140625,
+ "step": 1752
+ },
+ {
+ "epoch": 24.349344978165938,
+ "grad_norm": 0.020956002175807953,
+ "learning_rate": 0.0006,
+ "loss": 4.345308303833008,
+ "step": 1753
+ },
+ {
+ "epoch": 24.363318777292577,
+ "grad_norm": 0.019227957352995872,
+ "learning_rate": 0.0006,
+ "loss": 4.349849224090576,
+ "step": 1754
+ },
+ {
+ "epoch": 24.377292576419215,
+ "grad_norm": 0.015970097854733467,
+ "learning_rate": 0.0006,
+ "loss": 4.223012447357178,
+ "step": 1755
+ },
+ {
+ "epoch": 24.39126637554585,
+ "grad_norm": 0.015951646491885185,
+ "learning_rate": 0.0006,
+ "loss": 4.236420631408691,
+ "step": 1756
+ },
+ {
+ "epoch": 24.40524017467249,
+ "grad_norm": 0.016032515093684196,
+ "learning_rate": 0.0006,
+ "loss": 4.226899147033691,
+ "step": 1757
+ },
+ {
+ "epoch": 24.419213973799128,
+ "grad_norm": 0.01573132909834385,
+ "learning_rate": 0.0006,
+ "loss": 4.315916061401367,
+ "step": 1758
+ },
+ {
+ "epoch": 24.433187772925763,
+ "grad_norm": 0.015367005951702595,
+ "learning_rate": 0.0006,
+ "loss": 4.358939170837402,
+ "step": 1759
+ },
+ {
+ "epoch": 24.4471615720524,
+ "grad_norm": 0.01669907011091709,
+ "learning_rate": 0.0006,
+ "loss": 4.232147216796875,
+ "step": 1760
+ },
+ {
+ "epoch": 24.46113537117904,
+ "grad_norm": 0.018708152696490288,
+ "learning_rate": 0.0006,
+ "loss": 4.329852104187012,
+ "step": 1761
+ },
+ {
+ "epoch": 24.475109170305675,
+ "grad_norm": 0.019035987555980682,
+ "learning_rate": 0.0006,
+ "loss": 4.260340213775635,
+ "step": 1762
+ },
+ {
+ "epoch": 24.489082969432314,
+ "grad_norm": 0.01832144521176815,
+ "learning_rate": 0.0006,
+ "loss": 4.359543800354004,
+ "step": 1763
+ },
+ {
+ "epoch": 24.503056768558952,
+ "grad_norm": 0.015313294716179371,
+ "learning_rate": 0.0006,
+ "loss": 4.279088020324707,
+ "step": 1764
+ },
+ {
+ "epoch": 24.51703056768559,
+ "grad_norm": 0.015803923830389977,
+ "learning_rate": 0.0006,
+ "loss": 4.230682849884033,
+ "step": 1765
+ },
+ {
+ "epoch": 24.531004366812226,
+ "grad_norm": 0.016748374328017235,
+ "learning_rate": 0.0006,
+ "loss": 4.349661827087402,
+ "step": 1766
+ },
+ {
+ "epoch": 24.544978165938865,
+ "grad_norm": 0.015732428058981895,
+ "learning_rate": 0.0006,
+ "loss": 4.372441291809082,
+ "step": 1767
+ },
+ {
+ "epoch": 24.558951965065503,
+ "grad_norm": 0.016564829275012016,
+ "learning_rate": 0.0006,
+ "loss": 4.323090553283691,
+ "step": 1768
+ },
+ {
+ "epoch": 24.57292576419214,
+ "grad_norm": 0.01603010483086109,
+ "learning_rate": 0.0006,
+ "loss": 4.347898006439209,
+ "step": 1769
+ },
+ {
+ "epoch": 24.586899563318777,
+ "grad_norm": 0.01674809865653515,
+ "learning_rate": 0.0006,
+ "loss": 4.269165515899658,
+ "step": 1770
+ },
+ {
+ "epoch": 24.600873362445416,
+ "grad_norm": 0.017333930358290672,
+ "learning_rate": 0.0006,
+ "loss": 4.353298187255859,
+ "step": 1771
+ },
+ {
+ "epoch": 24.61484716157205,
+ "grad_norm": 0.017519580200314522,
+ "learning_rate": 0.0006,
+ "loss": 4.265199661254883,
+ "step": 1772
+ },
+ {
+ "epoch": 24.62882096069869,
+ "grad_norm": 0.017675306648015976,
+ "learning_rate": 0.0006,
+ "loss": 4.322596549987793,
+ "step": 1773
+ },
+ {
+ "epoch": 24.64279475982533,
+ "grad_norm": 0.017166534438729286,
+ "learning_rate": 0.0006,
+ "loss": 4.413900375366211,
+ "step": 1774
+ },
+ {
+ "epoch": 24.656768558951963,
+ "grad_norm": 0.016362346708774567,
+ "learning_rate": 0.0006,
+ "loss": 4.307236671447754,
+ "step": 1775
+ },
+ {
+ "epoch": 24.670742358078602,
+ "grad_norm": 0.016086872667074203,
+ "learning_rate": 0.0006,
+ "loss": 4.235316276550293,
+ "step": 1776
+ },
+ {
+ "epoch": 24.68471615720524,
+ "grad_norm": 0.01689537614583969,
+ "learning_rate": 0.0006,
+ "loss": 4.390591621398926,
+ "step": 1777
+ },
+ {
+ "epoch": 24.69868995633188,
+ "grad_norm": 0.016836725175380707,
+ "learning_rate": 0.0006,
+ "loss": 4.2990031242370605,
+ "step": 1778
+ },
+ {
+ "epoch": 24.712663755458514,
+ "grad_norm": 0.015627335757017136,
+ "learning_rate": 0.0006,
+ "loss": 4.166745185852051,
+ "step": 1779
+ },
+ {
+ "epoch": 24.726637554585153,
+ "grad_norm": 0.015007016249001026,
+ "learning_rate": 0.0006,
+ "loss": 4.295273303985596,
+ "step": 1780
+ },
+ {
+ "epoch": 24.74061135371179,
+ "grad_norm": 0.016084209084510803,
+ "learning_rate": 0.0006,
+ "loss": 4.342424392700195,
+ "step": 1781
+ },
+ {
+ "epoch": 24.754585152838427,
+ "grad_norm": 0.01595073565840721,
+ "learning_rate": 0.0006,
+ "loss": 4.3060102462768555,
+ "step": 1782
+ },
+ {
+ "epoch": 24.768558951965066,
+ "grad_norm": 0.016113461926579475,
+ "learning_rate": 0.0006,
+ "loss": 4.267908096313477,
+ "step": 1783
+ },
+ {
+ "epoch": 24.782532751091704,
+ "grad_norm": 0.01564556173980236,
+ "learning_rate": 0.0006,
+ "loss": 4.324642658233643,
+ "step": 1784
+ },
+ {
+ "epoch": 24.79650655021834,
+ "grad_norm": 0.01575268618762493,
+ "learning_rate": 0.0006,
+ "loss": 4.328181743621826,
+ "step": 1785
+ },
+ {
+ "epoch": 24.810480349344978,
+ "grad_norm": 0.016830824315547943,
+ "learning_rate": 0.0006,
+ "loss": 4.247906684875488,
+ "step": 1786
+ },
+ {
+ "epoch": 24.824454148471617,
+ "grad_norm": 0.017989547923207283,
+ "learning_rate": 0.0006,
+ "loss": 4.228281021118164,
+ "step": 1787
+ },
+ {
+ "epoch": 24.83842794759825,
+ "grad_norm": 0.016783086583018303,
+ "learning_rate": 0.0006,
+ "loss": 4.2880964279174805,
+ "step": 1788
+ },
+ {
+ "epoch": 24.85240174672489,
+ "grad_norm": 0.016467561945319176,
+ "learning_rate": 0.0006,
+ "loss": 4.36118745803833,
+ "step": 1789
+ },
+ {
+ "epoch": 24.86637554585153,
+ "grad_norm": 0.01780523732304573,
+ "learning_rate": 0.0006,
+ "loss": 4.172321319580078,
+ "step": 1790
+ },
+ {
+ "epoch": 24.880349344978168,
+ "grad_norm": 0.017102031037211418,
+ "learning_rate": 0.0006,
+ "loss": 4.352826118469238,
+ "step": 1791
+ },
+ {
+ "epoch": 24.894323144104803,
+ "grad_norm": 0.016409218311309814,
+ "learning_rate": 0.0006,
+ "loss": 4.212162017822266,
+ "step": 1792
+ },
+ {
+ "epoch": 24.90829694323144,
+ "grad_norm": 0.018660584464669228,
+ "learning_rate": 0.0006,
+ "loss": 4.346684455871582,
+ "step": 1793
+ },
+ {
+ "epoch": 24.92227074235808,
+ "grad_norm": 0.017389440909028053,
+ "learning_rate": 0.0006,
+ "loss": 4.197913646697998,
+ "step": 1794
+ },
+ {
+ "epoch": 24.936244541484715,
+ "grad_norm": 0.01596485823392868,
+ "learning_rate": 0.0006,
+ "loss": 4.370617866516113,
+ "step": 1795
+ },
+ {
+ "epoch": 24.950218340611354,
+ "grad_norm": 0.01709776371717453,
+ "learning_rate": 0.0006,
+ "loss": 4.313847064971924,
+ "step": 1796
+ },
+ {
+ "epoch": 24.964192139737992,
+ "grad_norm": 0.016571134328842163,
+ "learning_rate": 0.0006,
+ "loss": 4.1895856857299805,
+ "step": 1797
+ },
+ {
+ "epoch": 24.978165938864628,
+ "grad_norm": 0.016851048916578293,
+ "learning_rate": 0.0006,
+ "loss": 4.350772857666016,
+ "step": 1798
+ },
+ {
+ "epoch": 24.992139737991266,
+ "grad_norm": 0.017007440328598022,
+ "learning_rate": 0.0006,
+ "loss": 4.399008750915527,
+ "step": 1799
+ },
+ {
+ "epoch": 25.0,
+ "grad_norm": 0.01791212148964405,
+ "learning_rate": 0.0006,
+ "loss": 4.306121826171875,
+ "step": 1800
+ },
+ {
+ "epoch": 25.0,
+ "eval_loss": 4.636417865753174,
+ "eval_runtime": 56.8296,
+ "eval_samples_per_second": 42.971,
+ "eval_steps_per_second": 1.355,
+ "step": 1800
+ },
+ {
+ "epoch": 25.01397379912664,
+ "grad_norm": 0.01729617640376091,
+ "learning_rate": 0.0006,
+ "loss": 4.257018089294434,
+ "step": 1801
+ },
+ {
+ "epoch": 25.027947598253274,
+ "grad_norm": 0.017466643825173378,
+ "learning_rate": 0.0006,
+ "loss": 4.307605266571045,
+ "step": 1802
+ },
+ {
+ "epoch": 25.041921397379912,
+ "grad_norm": 0.018614279106259346,
+ "learning_rate": 0.0006,
+ "loss": 4.360141277313232,
+ "step": 1803
+ },
+ {
+ "epoch": 25.05589519650655,
+ "grad_norm": 0.02010771445930004,
+ "learning_rate": 0.0006,
+ "loss": 4.196428298950195,
+ "step": 1804
+ },
+ {
+ "epoch": 25.069868995633186,
+ "grad_norm": 0.02090228535234928,
+ "learning_rate": 0.0006,
+ "loss": 4.288947105407715,
+ "step": 1805
+ },
+ {
+ "epoch": 25.083842794759825,
+ "grad_norm": 0.02050255797803402,
+ "learning_rate": 0.0006,
+ "loss": 4.259582042694092,
+ "step": 1806
+ },
+ {
+ "epoch": 25.097816593886463,
+ "grad_norm": 0.021844089031219482,
+ "learning_rate": 0.0006,
+ "loss": 4.405138969421387,
+ "step": 1807
+ },
+ {
+ "epoch": 25.111790393013102,
+ "grad_norm": 0.022287018597126007,
+ "learning_rate": 0.0006,
+ "loss": 4.333793640136719,
+ "step": 1808
+ },
+ {
+ "epoch": 25.125764192139737,
+ "grad_norm": 0.02288329415023327,
+ "learning_rate": 0.0006,
+ "loss": 4.3910980224609375,
+ "step": 1809
+ },
+ {
+ "epoch": 25.139737991266376,
+ "grad_norm": 0.02373199723660946,
+ "learning_rate": 0.0006,
+ "loss": 4.32921028137207,
+ "step": 1810
+ },
+ {
+ "epoch": 25.153711790393015,
+ "grad_norm": 0.024960234761238098,
+ "learning_rate": 0.0006,
+ "loss": 4.295609474182129,
+ "step": 1811
+ },
+ {
+ "epoch": 25.16768558951965,
+ "grad_norm": 0.026743892580270767,
+ "learning_rate": 0.0006,
+ "loss": 4.315308094024658,
+ "step": 1812
+ },
+ {
+ "epoch": 25.18165938864629,
+ "grad_norm": 0.022138185799121857,
+ "learning_rate": 0.0006,
+ "loss": 4.360208511352539,
+ "step": 1813
+ },
+ {
+ "epoch": 25.195633187772927,
+ "grad_norm": 0.022120574489235878,
+ "learning_rate": 0.0006,
+ "loss": 4.1973772048950195,
+ "step": 1814
+ },
+ {
+ "epoch": 25.209606986899562,
+ "grad_norm": 0.02230502851307392,
+ "learning_rate": 0.0006,
+ "loss": 4.285558700561523,
+ "step": 1815
+ },
+ {
+ "epoch": 25.2235807860262,
+ "grad_norm": 0.022301986813545227,
+ "learning_rate": 0.0006,
+ "loss": 4.238343238830566,
+ "step": 1816
+ },
+ {
+ "epoch": 25.23755458515284,
+ "grad_norm": 0.020050447434186935,
+ "learning_rate": 0.0006,
+ "loss": 4.298235893249512,
+ "step": 1817
+ },
+ {
+ "epoch": 25.251528384279474,
+ "grad_norm": 0.01987724006175995,
+ "learning_rate": 0.0006,
+ "loss": 4.301384449005127,
+ "step": 1818
+ },
+ {
+ "epoch": 25.265502183406113,
+ "grad_norm": 0.020440855994820595,
+ "learning_rate": 0.0006,
+ "loss": 4.269218444824219,
+ "step": 1819
+ },
+ {
+ "epoch": 25.27947598253275,
+ "grad_norm": 0.01891893707215786,
+ "learning_rate": 0.0006,
+ "loss": 4.2401957511901855,
+ "step": 1820
+ },
+ {
+ "epoch": 25.29344978165939,
+ "grad_norm": 0.01903688907623291,
+ "learning_rate": 0.0006,
+ "loss": 4.139670372009277,
+ "step": 1821
+ },
+ {
+ "epoch": 25.307423580786025,
+ "grad_norm": 0.018856938928365707,
+ "learning_rate": 0.0006,
+ "loss": 4.233214378356934,
+ "step": 1822
+ },
+ {
+ "epoch": 25.321397379912664,
+ "grad_norm": 0.017791323363780975,
+ "learning_rate": 0.0006,
+ "loss": 4.2342023849487305,
+ "step": 1823
+ },
+ {
+ "epoch": 25.335371179039303,
+ "grad_norm": 0.017304565757513046,
+ "learning_rate": 0.0006,
+ "loss": 4.355198860168457,
+ "step": 1824
+ },
+ {
+ "epoch": 25.349344978165938,
+ "grad_norm": 0.01576206088066101,
+ "learning_rate": 0.0006,
+ "loss": 4.114147186279297,
+ "step": 1825
+ },
+ {
+ "epoch": 25.363318777292577,
+ "grad_norm": 0.015105154365301132,
+ "learning_rate": 0.0006,
+ "loss": 4.313795566558838,
+ "step": 1826
+ },
+ {
+ "epoch": 25.377292576419215,
+ "grad_norm": 0.01563587784767151,
+ "learning_rate": 0.0006,
+ "loss": 4.308282852172852,
+ "step": 1827
+ },
+ {
+ "epoch": 25.39126637554585,
+ "grad_norm": 0.015980906784534454,
+ "learning_rate": 0.0006,
+ "loss": 4.345149040222168,
+ "step": 1828
+ },
+ {
+ "epoch": 25.40524017467249,
+ "grad_norm": 0.01669284701347351,
+ "learning_rate": 0.0006,
+ "loss": 4.237611293792725,
+ "step": 1829
+ },
+ {
+ "epoch": 25.419213973799128,
+ "grad_norm": 0.016202064231038094,
+ "learning_rate": 0.0006,
+ "loss": 4.3072309494018555,
+ "step": 1830
+ },
+ {
+ "epoch": 25.433187772925763,
+ "grad_norm": 0.01478662807494402,
+ "learning_rate": 0.0006,
+ "loss": 4.276215076446533,
+ "step": 1831
+ },
+ {
+ "epoch": 25.4471615720524,
+ "grad_norm": 0.015168731100857258,
+ "learning_rate": 0.0006,
+ "loss": 4.37509822845459,
+ "step": 1832
+ },
+ {
+ "epoch": 25.46113537117904,
+ "grad_norm": 0.01648980937898159,
+ "learning_rate": 0.0006,
+ "loss": 4.349167346954346,
+ "step": 1833
+ },
+ {
+ "epoch": 25.475109170305675,
+ "grad_norm": 0.01708398200571537,
+ "learning_rate": 0.0006,
+ "loss": 4.346580505371094,
+ "step": 1834
+ },
+ {
+ "epoch": 25.489082969432314,
+ "grad_norm": 0.017370784655213356,
+ "learning_rate": 0.0006,
+ "loss": 4.299466133117676,
+ "step": 1835
+ },
+ {
+ "epoch": 25.503056768558952,
+ "grad_norm": 0.016248228028416634,
+ "learning_rate": 0.0006,
+ "loss": 4.184610366821289,
+ "step": 1836
+ },
+ {
+ "epoch": 25.51703056768559,
+ "grad_norm": 0.017106913030147552,
+ "learning_rate": 0.0006,
+ "loss": 4.366856098175049,
+ "step": 1837
+ },
+ {
+ "epoch": 25.531004366812226,
+ "grad_norm": 0.016071965917944908,
+ "learning_rate": 0.0006,
+ "loss": 4.297691345214844,
+ "step": 1838
+ },
+ {
+ "epoch": 25.544978165938865,
+ "grad_norm": 0.016451986506581306,
+ "learning_rate": 0.0006,
+ "loss": 4.229983329772949,
+ "step": 1839
+ },
+ {
+ "epoch": 25.558951965065503,
+ "grad_norm": 0.016195081174373627,
+ "learning_rate": 0.0006,
+ "loss": 4.259893417358398,
+ "step": 1840
+ },
+ {
+ "epoch": 25.57292576419214,
+ "grad_norm": 0.0165175162255764,
+ "learning_rate": 0.0006,
+ "loss": 4.3348236083984375,
+ "step": 1841
+ },
+ {
+ "epoch": 25.586899563318777,
+ "grad_norm": 0.01651517115533352,
+ "learning_rate": 0.0006,
+ "loss": 4.319541931152344,
+ "step": 1842
+ },
+ {
+ "epoch": 25.600873362445416,
+ "grad_norm": 0.016143690794706345,
+ "learning_rate": 0.0006,
+ "loss": 4.345652103424072,
+ "step": 1843
+ },
+ {
+ "epoch": 25.61484716157205,
+ "grad_norm": 0.015140696428716183,
+ "learning_rate": 0.0006,
+ "loss": 4.212047576904297,
+ "step": 1844
+ },
+ {
+ "epoch": 25.62882096069869,
+ "grad_norm": 0.014637443237006664,
+ "learning_rate": 0.0006,
+ "loss": 4.187453269958496,
+ "step": 1845
+ },
+ {
+ "epoch": 25.64279475982533,
+ "grad_norm": 0.014585614204406738,
+ "learning_rate": 0.0006,
+ "loss": 4.253936767578125,
+ "step": 1846
+ },
+ {
+ "epoch": 25.656768558951963,
+ "grad_norm": 0.015158289112150669,
+ "learning_rate": 0.0006,
+ "loss": 4.34612512588501,
+ "step": 1847
+ },
+ {
+ "epoch": 25.670742358078602,
+ "grad_norm": 0.015202849172055721,
+ "learning_rate": 0.0006,
+ "loss": 4.270634651184082,
+ "step": 1848
+ },
+ {
+ "epoch": 25.68471615720524,
+ "grad_norm": 0.015363575890660286,
+ "learning_rate": 0.0006,
+ "loss": 4.271080493927002,
+ "step": 1849
+ },
+ {
+ "epoch": 25.69868995633188,
+ "grad_norm": 0.014360226690769196,
+ "learning_rate": 0.0006,
+ "loss": 4.375033855438232,
+ "step": 1850
+ },
+ {
+ "epoch": 25.712663755458514,
+ "grad_norm": 0.014263181947171688,
+ "learning_rate": 0.0006,
+ "loss": 4.424138069152832,
+ "step": 1851
+ },
+ {
+ "epoch": 25.726637554585153,
+ "grad_norm": 0.014398688450455666,
+ "learning_rate": 0.0006,
+ "loss": 4.410154819488525,
+ "step": 1852
+ },
+ {
+ "epoch": 25.74061135371179,
+ "grad_norm": 0.016131488606333733,
+ "learning_rate": 0.0006,
+ "loss": 4.324014663696289,
+ "step": 1853
+ },
+ {
+ "epoch": 25.754585152838427,
+ "grad_norm": 0.017568735405802727,
+ "learning_rate": 0.0006,
+ "loss": 4.367861747741699,
+ "step": 1854
+ },
+ {
+ "epoch": 25.768558951965066,
+ "grad_norm": 0.017907511442899704,
+ "learning_rate": 0.0006,
+ "loss": 4.352965354919434,
+ "step": 1855
+ },
+ {
+ "epoch": 25.782532751091704,
+ "grad_norm": 0.016918016597628593,
+ "learning_rate": 0.0006,
+ "loss": 4.252397537231445,
+ "step": 1856
+ },
+ {
+ "epoch": 25.79650655021834,
+ "grad_norm": 0.017274610698223114,
+ "learning_rate": 0.0006,
+ "loss": 4.342309951782227,
+ "step": 1857
+ },
+ {
+ "epoch": 25.810480349344978,
+ "grad_norm": 0.017879825085401535,
+ "learning_rate": 0.0006,
+ "loss": 4.330987453460693,
+ "step": 1858
+ },
+ {
+ "epoch": 25.824454148471617,
+ "grad_norm": 0.01736099272966385,
+ "learning_rate": 0.0006,
+ "loss": 4.307255744934082,
+ "step": 1859
+ },
+ {
+ "epoch": 25.83842794759825,
+ "grad_norm": 0.016575824469327927,
+ "learning_rate": 0.0006,
+ "loss": 4.262016773223877,
+ "step": 1860
+ },
+ {
+ "epoch": 25.85240174672489,
+ "grad_norm": 0.015181140042841434,
+ "learning_rate": 0.0006,
+ "loss": 4.291863441467285,
+ "step": 1861
+ },
+ {
+ "epoch": 25.86637554585153,
+ "grad_norm": 0.015965687111020088,
+ "learning_rate": 0.0006,
+ "loss": 4.424836158752441,
+ "step": 1862
+ },
+ {
+ "epoch": 25.880349344978168,
+ "grad_norm": 0.01687219925224781,
+ "learning_rate": 0.0006,
+ "loss": 4.343531131744385,
+ "step": 1863
+ },
+ {
+ "epoch": 25.894323144104803,
+ "grad_norm": 0.015115504153072834,
+ "learning_rate": 0.0006,
+ "loss": 4.357808589935303,
+ "step": 1864
+ },
+ {
+ "epoch": 25.90829694323144,
+ "grad_norm": 0.014150998555123806,
+ "learning_rate": 0.0006,
+ "loss": 4.324550151824951,
+ "step": 1865
+ },
+ {
+ "epoch": 25.92227074235808,
+ "grad_norm": 0.014989510178565979,
+ "learning_rate": 0.0006,
+ "loss": 4.380428314208984,
+ "step": 1866
+ },
+ {
+ "epoch": 25.936244541484715,
+ "grad_norm": 0.01527960691601038,
+ "learning_rate": 0.0006,
+ "loss": 4.399374961853027,
+ "step": 1867
+ },
+ {
+ "epoch": 25.950218340611354,
+ "grad_norm": 0.015434633940458298,
+ "learning_rate": 0.0006,
+ "loss": 4.260984420776367,
+ "step": 1868
+ },
+ {
+ "epoch": 25.964192139737992,
+ "grad_norm": 0.015370228327810764,
+ "learning_rate": 0.0006,
+ "loss": 4.3340253829956055,
+ "step": 1869
+ },
+ {
+ "epoch": 25.978165938864628,
+ "grad_norm": 0.015348542481660843,
+ "learning_rate": 0.0006,
+ "loss": 4.369132995605469,
+ "step": 1870
+ },
+ {
+ "epoch": 25.992139737991266,
+ "grad_norm": 0.015354936942458153,
+ "learning_rate": 0.0006,
+ "loss": 4.381937026977539,
+ "step": 1871
+ },
+ {
+ "epoch": 26.0,
+ "grad_norm": 0.01796272210776806,
+ "learning_rate": 0.0006,
+ "loss": 4.286768436431885,
+ "step": 1872
+ },
+ {
+ "epoch": 26.0,
+ "eval_loss": 4.576984405517578,
+ "eval_runtime": 57.2091,
+ "eval_samples_per_second": 42.685,
+ "eval_steps_per_second": 1.346,
+ "step": 1872
+ },
+ {
+ "epoch": 26.01397379912664,
+ "grad_norm": 0.01748845726251602,
+ "learning_rate": 0.0006,
+ "loss": 4.337504863739014,
+ "step": 1873
+ },
+ {
+ "epoch": 26.027947598253274,
+ "grad_norm": 0.019244296476244926,
+ "learning_rate": 0.0006,
+ "loss": 4.264280796051025,
+ "step": 1874
+ },
+ {
+ "epoch": 26.041921397379912,
+ "grad_norm": 0.0215692650526762,
+ "learning_rate": 0.0006,
+ "loss": 4.350830078125,
+ "step": 1875
+ },
+ {
+ "epoch": 26.05589519650655,
+ "grad_norm": 0.022189252078533173,
+ "learning_rate": 0.0006,
+ "loss": 4.299098968505859,
+ "step": 1876
+ },
+ {
+ "epoch": 26.069868995633186,
+ "grad_norm": 0.022384043782949448,
+ "learning_rate": 0.0006,
+ "loss": 4.1185173988342285,
+ "step": 1877
+ },
+ {
+ "epoch": 26.083842794759825,
+ "grad_norm": 0.02076014317572117,
+ "learning_rate": 0.0006,
+ "loss": 4.297612190246582,
+ "step": 1878
+ },
+ {
+ "epoch": 26.097816593886463,
+ "grad_norm": 0.020843051373958588,
+ "learning_rate": 0.0006,
+ "loss": 4.294098854064941,
+ "step": 1879
+ },
+ {
+ "epoch": 26.111790393013102,
+ "grad_norm": 0.020654456689953804,
+ "learning_rate": 0.0006,
+ "loss": 4.267856597900391,
+ "step": 1880
+ },
+ {
+ "epoch": 26.125764192139737,
+ "grad_norm": 0.01860472559928894,
+ "learning_rate": 0.0006,
+ "loss": 4.284232139587402,
+ "step": 1881
+ },
+ {
+ "epoch": 26.139737991266376,
+ "grad_norm": 0.018719421699643135,
+ "learning_rate": 0.0006,
+ "loss": 4.277214527130127,
+ "step": 1882
+ },
+ {
+ "epoch": 26.153711790393015,
+ "grad_norm": 0.020903829485177994,
+ "learning_rate": 0.0006,
+ "loss": 4.254947662353516,
+ "step": 1883
+ },
+ {
+ "epoch": 26.16768558951965,
+ "grad_norm": 0.02020149491727352,
+ "learning_rate": 0.0006,
+ "loss": 4.2526936531066895,
+ "step": 1884
+ },
+ {
+ "epoch": 26.18165938864629,
+ "grad_norm": 0.0180392786860466,
+ "learning_rate": 0.0006,
+ "loss": 4.309922695159912,
+ "step": 1885
+ },
+ {
+ "epoch": 26.195633187772927,
+ "grad_norm": 0.018163125962018967,
+ "learning_rate": 0.0006,
+ "loss": 4.13860559463501,
+ "step": 1886
+ },
+ {
+ "epoch": 26.209606986899562,
+ "grad_norm": 0.01853291131556034,
+ "learning_rate": 0.0006,
+ "loss": 4.190926551818848,
+ "step": 1887
+ },
+ {
+ "epoch": 26.2235807860262,
+ "grad_norm": 0.01757677271962166,
+ "learning_rate": 0.0006,
+ "loss": 4.163999557495117,
+ "step": 1888
+ },
+ {
+ "epoch": 26.23755458515284,
+ "grad_norm": 0.01767992228269577,
+ "learning_rate": 0.0006,
+ "loss": 4.184430122375488,
+ "step": 1889
+ },
+ {
+ "epoch": 26.251528384279474,
+ "grad_norm": 0.018139973282814026,
+ "learning_rate": 0.0006,
+ "loss": 4.300766944885254,
+ "step": 1890
+ },
+ {
+ "epoch": 26.265502183406113,
+ "grad_norm": 0.020184461027383804,
+ "learning_rate": 0.0006,
+ "loss": 4.232537269592285,
+ "step": 1891
+ },
+ {
+ "epoch": 26.27947598253275,
+ "grad_norm": 0.02028856985270977,
+ "learning_rate": 0.0006,
+ "loss": 4.324883460998535,
+ "step": 1892
+ },
+ {
+ "epoch": 26.29344978165939,
+ "grad_norm": 0.017120616510510445,
+ "learning_rate": 0.0006,
+ "loss": 4.23493766784668,
+ "step": 1893
+ },
+ {
+ "epoch": 26.307423580786025,
+ "grad_norm": 0.016202978789806366,
+ "learning_rate": 0.0006,
+ "loss": 4.197602272033691,
+ "step": 1894
+ },
+ {
+ "epoch": 26.321397379912664,
+ "grad_norm": 0.017231421545147896,
+ "learning_rate": 0.0006,
+ "loss": 4.302001953125,
+ "step": 1895
+ },
+ {
+ "epoch": 26.335371179039303,
+ "grad_norm": 0.01725655607879162,
+ "learning_rate": 0.0006,
+ "loss": 4.356800079345703,
+ "step": 1896
+ },
+ {
+ "epoch": 26.349344978165938,
+ "grad_norm": 0.016657443717122078,
+ "learning_rate": 0.0006,
+ "loss": 4.266307830810547,
+ "step": 1897
+ },
+ {
+ "epoch": 26.363318777292577,
+ "grad_norm": 0.01616556942462921,
+ "learning_rate": 0.0006,
+ "loss": 4.307868003845215,
+ "step": 1898
+ },
+ {
+ "epoch": 26.377292576419215,
+ "grad_norm": 0.016621306538581848,
+ "learning_rate": 0.0006,
+ "loss": 4.172905921936035,
+ "step": 1899
+ },
+ {
+ "epoch": 26.39126637554585,
+ "grad_norm": 0.016243206337094307,
+ "learning_rate": 0.0006,
+ "loss": 4.162896156311035,
+ "step": 1900
+ },
+ {
+ "epoch": 26.40524017467249,
+ "grad_norm": 0.016663808375597,
+ "learning_rate": 0.0006,
+ "loss": 4.258232116699219,
+ "step": 1901
+ },
+ {
+ "epoch": 26.419213973799128,
+ "grad_norm": 0.01677478663623333,
+ "learning_rate": 0.0006,
+ "loss": 4.336090087890625,
+ "step": 1902
+ },
+ {
+ "epoch": 26.433187772925763,
+ "grad_norm": 0.01641303487122059,
+ "learning_rate": 0.0006,
+ "loss": 4.191476821899414,
+ "step": 1903
+ },
+ {
+ "epoch": 26.4471615720524,
+ "grad_norm": 0.016700323671102524,
+ "learning_rate": 0.0006,
+ "loss": 4.299427509307861,
+ "step": 1904
+ },
+ {
+ "epoch": 26.46113537117904,
+ "grad_norm": 0.017595035955309868,
+ "learning_rate": 0.0006,
+ "loss": 4.312828063964844,
+ "step": 1905
+ },
+ {
+ "epoch": 26.475109170305675,
+ "grad_norm": 0.020300284028053284,
+ "learning_rate": 0.0006,
+ "loss": 4.28331184387207,
+ "step": 1906
+ },
+ {
+ "epoch": 26.489082969432314,
+ "grad_norm": 0.02198321744799614,
+ "learning_rate": 0.0006,
+ "loss": 4.3280839920043945,
+ "step": 1907
+ },
+ {
+ "epoch": 26.503056768558952,
+ "grad_norm": 0.02003200724720955,
+ "learning_rate": 0.0006,
+ "loss": 4.325623989105225,
+ "step": 1908
+ },
+ {
+ "epoch": 26.51703056768559,
+ "grad_norm": 0.01870287023484707,
+ "learning_rate": 0.0006,
+ "loss": 4.251203536987305,
+ "step": 1909
+ },
+ {
+ "epoch": 26.531004366812226,
+ "grad_norm": 0.01835448481142521,
+ "learning_rate": 0.0006,
+ "loss": 4.335144996643066,
+ "step": 1910
+ },
+ {
+ "epoch": 26.544978165938865,
+ "grad_norm": 0.02046327292919159,
+ "learning_rate": 0.0006,
+ "loss": 4.238023281097412,
+ "step": 1911
+ },
+ {
+ "epoch": 26.558951965065503,
+ "grad_norm": 0.019771311432123184,
+ "learning_rate": 0.0006,
+ "loss": 4.353594779968262,
+ "step": 1912
+ },
+ {
+ "epoch": 26.57292576419214,
+ "grad_norm": 0.018090544268488884,
+ "learning_rate": 0.0006,
+ "loss": 4.304343223571777,
+ "step": 1913
+ },
+ {
+ "epoch": 26.586899563318777,
+ "grad_norm": 0.017489777877926826,
+ "learning_rate": 0.0006,
+ "loss": 4.204375267028809,
+ "step": 1914
+ },
+ {
+ "epoch": 26.600873362445416,
+ "grad_norm": 0.016785891726613045,
+ "learning_rate": 0.0006,
+ "loss": 4.286669731140137,
+ "step": 1915
+ },
+ {
+ "epoch": 26.61484716157205,
+ "grad_norm": 0.016382789239287376,
+ "learning_rate": 0.0006,
+ "loss": 4.347662448883057,
+ "step": 1916
+ },
+ {
+ "epoch": 26.62882096069869,
+ "grad_norm": 0.016581876203417778,
+ "learning_rate": 0.0006,
+ "loss": 4.2716193199157715,
+ "step": 1917
+ },
+ {
+ "epoch": 26.64279475982533,
+ "grad_norm": 0.015627184882760048,
+ "learning_rate": 0.0006,
+ "loss": 4.311489105224609,
+ "step": 1918
+ },
+ {
+ "epoch": 26.656768558951963,
+ "grad_norm": 0.015908483415842056,
+ "learning_rate": 0.0006,
+ "loss": 4.262099742889404,
+ "step": 1919
+ },
+ {
+ "epoch": 26.670742358078602,
+ "grad_norm": 0.015576236881315708,
+ "learning_rate": 0.0006,
+ "loss": 4.275871276855469,
+ "step": 1920
+ },
+ {
+ "epoch": 26.68471615720524,
+ "grad_norm": 0.014621064998209476,
+ "learning_rate": 0.0006,
+ "loss": 4.184260845184326,
+ "step": 1921
+ },
+ {
+ "epoch": 26.69868995633188,
+ "grad_norm": 0.014797811396420002,
+ "learning_rate": 0.0006,
+ "loss": 4.3412628173828125,
+ "step": 1922
+ },
+ {
+ "epoch": 26.712663755458514,
+ "grad_norm": 0.015610570088028908,
+ "learning_rate": 0.0006,
+ "loss": 4.181094646453857,
+ "step": 1923
+ },
+ {
+ "epoch": 26.726637554585153,
+ "grad_norm": 0.016048265621066093,
+ "learning_rate": 0.0006,
+ "loss": 4.365290641784668,
+ "step": 1924
+ },
+ {
+ "epoch": 26.74061135371179,
+ "grad_norm": 0.016604071483016014,
+ "learning_rate": 0.0006,
+ "loss": 4.2628912925720215,
+ "step": 1925
+ },
+ {
+ "epoch": 26.754585152838427,
+ "grad_norm": 0.017044221982359886,
+ "learning_rate": 0.0006,
+ "loss": 4.378512382507324,
+ "step": 1926
+ },
+ {
+ "epoch": 26.768558951965066,
+ "grad_norm": 0.01746384985744953,
+ "learning_rate": 0.0006,
+ "loss": 4.3888773918151855,
+ "step": 1927
+ },
+ {
+ "epoch": 26.782532751091704,
+ "grad_norm": 0.01572711206972599,
+ "learning_rate": 0.0006,
+ "loss": 4.189935684204102,
+ "step": 1928
+ },
+ {
+ "epoch": 26.79650655021834,
+ "grad_norm": 0.014976629987359047,
+ "learning_rate": 0.0006,
+ "loss": 4.280525207519531,
+ "step": 1929
+ },
+ {
+ "epoch": 26.810480349344978,
+ "grad_norm": 0.014804664999246597,
+ "learning_rate": 0.0006,
+ "loss": 4.270630836486816,
+ "step": 1930
+ },
+ {
+ "epoch": 26.824454148471617,
+ "grad_norm": 0.014641920104622841,
+ "learning_rate": 0.0006,
+ "loss": 4.319505214691162,
+ "step": 1931
+ },
+ {
+ "epoch": 26.83842794759825,
+ "grad_norm": 0.014928505755960941,
+ "learning_rate": 0.0006,
+ "loss": 4.180147647857666,
+ "step": 1932
+ },
+ {
+ "epoch": 26.85240174672489,
+ "grad_norm": 0.016130365431308746,
+ "learning_rate": 0.0006,
+ "loss": 4.260660648345947,
+ "step": 1933
+ },
+ {
+ "epoch": 26.86637554585153,
+ "grad_norm": 0.015642493963241577,
+ "learning_rate": 0.0006,
+ "loss": 4.304064750671387,
+ "step": 1934
+ },
+ {
+ "epoch": 26.880349344978168,
+ "grad_norm": 0.015988217666745186,
+ "learning_rate": 0.0006,
+ "loss": 4.2548017501831055,
+ "step": 1935
+ },
+ {
+ "epoch": 26.894323144104803,
+ "grad_norm": 0.016501398757100105,
+ "learning_rate": 0.0006,
+ "loss": 4.260058879852295,
+ "step": 1936
+ },
+ {
+ "epoch": 26.90829694323144,
+ "grad_norm": 0.01645040698349476,
+ "learning_rate": 0.0006,
+ "loss": 4.483081340789795,
+ "step": 1937
+ },
+ {
+ "epoch": 26.92227074235808,
+ "grad_norm": 0.016140863299369812,
+ "learning_rate": 0.0006,
+ "loss": 4.1776838302612305,
+ "step": 1938
+ },
+ {
+ "epoch": 26.936244541484715,
+ "grad_norm": 0.016681019216775894,
+ "learning_rate": 0.0006,
+ "loss": 4.224747657775879,
+ "step": 1939
+ },
+ {
+ "epoch": 26.950218340611354,
+ "grad_norm": 0.01650378853082657,
+ "learning_rate": 0.0006,
+ "loss": 4.305703163146973,
+ "step": 1940
+ },
+ {
+ "epoch": 26.964192139737992,
+ "grad_norm": 0.017051683738827705,
+ "learning_rate": 0.0006,
+ "loss": 4.285835266113281,
+ "step": 1941
+ },
+ {
+ "epoch": 26.978165938864628,
+ "grad_norm": 0.016894327476620674,
+ "learning_rate": 0.0006,
+ "loss": 4.318901538848877,
+ "step": 1942
+ },
+ {
+ "epoch": 26.992139737991266,
+ "grad_norm": 0.017313100397586823,
+ "learning_rate": 0.0006,
+ "loss": 4.281434535980225,
+ "step": 1943
+ },
+ {
+ "epoch": 27.0,
+ "grad_norm": 0.018450884148478508,
+ "learning_rate": 0.0006,
+ "loss": 4.335104942321777,
+ "step": 1944
+ },
+ {
+ "epoch": 27.0,
+ "eval_loss": 4.640687465667725,
+ "eval_runtime": 56.9624,
+ "eval_samples_per_second": 42.87,
+ "eval_steps_per_second": 1.352,
+ "step": 1944
+ },
+ {
+ "epoch": 27.01397379912664,
+ "grad_norm": 0.017755698412656784,
+ "learning_rate": 0.0006,
+ "loss": 4.278824329376221,
+ "step": 1945
+ },
+ {
+ "epoch": 27.027947598253274,
+ "grad_norm": 0.01724310778081417,
+ "learning_rate": 0.0006,
+ "loss": 4.311049461364746,
+ "step": 1946
+ },
+ {
+ "epoch": 27.041921397379912,
+ "grad_norm": 0.01690123789012432,
+ "learning_rate": 0.0006,
+ "loss": 4.261994361877441,
+ "step": 1947
+ },
+ {
+ "epoch": 27.05589519650655,
+ "grad_norm": 0.018406696617603302,
+ "learning_rate": 0.0006,
+ "loss": 4.1925153732299805,
+ "step": 1948
+ },
+ {
+ "epoch": 27.069868995633186,
+ "grad_norm": 0.019945867359638214,
+ "learning_rate": 0.0006,
+ "loss": 4.108537197113037,
+ "step": 1949
+ },
+ {
+ "epoch": 27.083842794759825,
+ "grad_norm": 0.020316628739237785,
+ "learning_rate": 0.0006,
+ "loss": 4.209043025970459,
+ "step": 1950
+ },
+ {
+ "epoch": 27.097816593886463,
+ "grad_norm": 0.019690370187163353,
+ "learning_rate": 0.0006,
+ "loss": 4.195789337158203,
+ "step": 1951
+ },
+ {
+ "epoch": 27.111790393013102,
+ "grad_norm": 0.01785232499241829,
+ "learning_rate": 0.0006,
+ "loss": 4.21466064453125,
+ "step": 1952
+ },
+ {
+ "epoch": 27.125764192139737,
+ "grad_norm": 0.0164767038077116,
+ "learning_rate": 0.0006,
+ "loss": 4.270247459411621,
+ "step": 1953
+ },
+ {
+ "epoch": 27.139737991266376,
+ "grad_norm": 0.018008455634117126,
+ "learning_rate": 0.0006,
+ "loss": 4.201042175292969,
+ "step": 1954
+ },
+ {
+ "epoch": 27.153711790393015,
+ "grad_norm": 0.01807340793311596,
+ "learning_rate": 0.0006,
+ "loss": 4.25289249420166,
+ "step": 1955
+ },
+ {
+ "epoch": 27.16768558951965,
+ "grad_norm": 0.016246909275650978,
+ "learning_rate": 0.0006,
+ "loss": 4.116283416748047,
+ "step": 1956
+ },
+ {
+ "epoch": 27.18165938864629,
+ "grad_norm": 0.017545776441693306,
+ "learning_rate": 0.0006,
+ "loss": 4.316999435424805,
+ "step": 1957
+ },
+ {
+ "epoch": 27.195633187772927,
+ "grad_norm": 0.017915375530719757,
+ "learning_rate": 0.0006,
+ "loss": 4.253859519958496,
+ "step": 1958
+ },
+ {
+ "epoch": 27.209606986899562,
+ "grad_norm": 0.018112223595380783,
+ "learning_rate": 0.0006,
+ "loss": 4.2541022300720215,
+ "step": 1959
+ },
+ {
+ "epoch": 27.2235807860262,
+ "grad_norm": 0.01846139505505562,
+ "learning_rate": 0.0006,
+ "loss": 4.20438289642334,
+ "step": 1960
+ },
+ {
+ "epoch": 27.23755458515284,
+ "grad_norm": 0.01943897269666195,
+ "learning_rate": 0.0006,
+ "loss": 4.2799577713012695,
+ "step": 1961
+ },
+ {
+ "epoch": 27.251528384279474,
+ "grad_norm": 0.018253348767757416,
+ "learning_rate": 0.0006,
+ "loss": 4.1720499992370605,
+ "step": 1962
+ },
+ {
+ "epoch": 27.265502183406113,
+ "grad_norm": 0.018292246386408806,
+ "learning_rate": 0.0006,
+ "loss": 4.28231143951416,
+ "step": 1963
+ },
+ {
+ "epoch": 27.27947598253275,
+ "grad_norm": 0.01943155936896801,
+ "learning_rate": 0.0006,
+ "loss": 4.314976692199707,
+ "step": 1964
+ },
+ {
+ "epoch": 27.29344978165939,
+ "grad_norm": 0.019064994528889656,
+ "learning_rate": 0.0006,
+ "loss": 4.25328254699707,
+ "step": 1965
+ },
+ {
+ "epoch": 27.307423580786025,
+ "grad_norm": 0.017423273995518684,
+ "learning_rate": 0.0006,
+ "loss": 4.195992469787598,
+ "step": 1966
+ },
+ {
+ "epoch": 27.321397379912664,
+ "grad_norm": 0.01677924580872059,
+ "learning_rate": 0.0006,
+ "loss": 4.260043144226074,
+ "step": 1967
+ },
+ {
+ "epoch": 27.335371179039303,
+ "grad_norm": 0.01720673404633999,
+ "learning_rate": 0.0006,
+ "loss": 4.222780704498291,
+ "step": 1968
+ },
+ {
+ "epoch": 27.349344978165938,
+ "grad_norm": 0.017300743609666824,
+ "learning_rate": 0.0006,
+ "loss": 4.109524250030518,
+ "step": 1969
+ },
+ {
+ "epoch": 27.363318777292577,
+ "grad_norm": 0.016065871343016624,
+ "learning_rate": 0.0006,
+ "loss": 4.36108922958374,
+ "step": 1970
+ },
+ {
+ "epoch": 27.377292576419215,
+ "grad_norm": 0.016626616939902306,
+ "learning_rate": 0.0006,
+ "loss": 4.220630645751953,
+ "step": 1971
+ },
+ {
+ "epoch": 27.39126637554585,
+ "grad_norm": 0.016482602804899216,
+ "learning_rate": 0.0006,
+ "loss": 4.344797134399414,
+ "step": 1972
+ },
+ {
+ "epoch": 27.40524017467249,
+ "grad_norm": 0.01727953553199768,
+ "learning_rate": 0.0006,
+ "loss": 4.296645164489746,
+ "step": 1973
+ },
+ {
+ "epoch": 27.419213973799128,
+ "grad_norm": 0.01522731315344572,
+ "learning_rate": 0.0006,
+ "loss": 4.223751068115234,
+ "step": 1974
+ },
+ {
+ "epoch": 27.433187772925763,
+ "grad_norm": 0.01581776700913906,
+ "learning_rate": 0.0006,
+ "loss": 4.351379871368408,
+ "step": 1975
+ },
+ {
+ "epoch": 27.4471615720524,
+ "grad_norm": 0.016394605860114098,
+ "learning_rate": 0.0006,
+ "loss": 4.2412919998168945,
+ "step": 1976
+ },
+ {
+ "epoch": 27.46113537117904,
+ "grad_norm": 0.01751169003546238,
+ "learning_rate": 0.0006,
+ "loss": 4.257606506347656,
+ "step": 1977
+ },
+ {
+ "epoch": 27.475109170305675,
+ "grad_norm": 0.017452310770750046,
+ "learning_rate": 0.0006,
+ "loss": 4.380134582519531,
+ "step": 1978
+ },
+ {
+ "epoch": 27.489082969432314,
+ "grad_norm": 0.017741898074746132,
+ "learning_rate": 0.0006,
+ "loss": 4.279632568359375,
+ "step": 1979
+ },
+ {
+ "epoch": 27.503056768558952,
+ "grad_norm": 0.017680590972304344,
+ "learning_rate": 0.0006,
+ "loss": 4.273406028747559,
+ "step": 1980
+ },
+ {
+ "epoch": 27.51703056768559,
+ "grad_norm": 0.01864718459546566,
+ "learning_rate": 0.0006,
+ "loss": 4.278583526611328,
+ "step": 1981
+ },
+ {
+ "epoch": 27.531004366812226,
+ "grad_norm": 0.019558541476726532,
+ "learning_rate": 0.0006,
+ "loss": 4.211919784545898,
+ "step": 1982
+ },
+ {
+ "epoch": 27.544978165938865,
+ "grad_norm": 0.019467400386929512,
+ "learning_rate": 0.0006,
+ "loss": 4.276822566986084,
+ "step": 1983
+ },
+ {
+ "epoch": 27.558951965065503,
+ "grad_norm": 0.01756172813475132,
+ "learning_rate": 0.0006,
+ "loss": 4.165379047393799,
+ "step": 1984
+ },
+ {
+ "epoch": 27.57292576419214,
+ "grad_norm": 0.017458263784646988,
+ "learning_rate": 0.0006,
+ "loss": 4.2992401123046875,
+ "step": 1985
+ },
+ {
+ "epoch": 27.586899563318777,
+ "grad_norm": 0.018089665099978447,
+ "learning_rate": 0.0006,
+ "loss": 4.243378639221191,
+ "step": 1986
+ },
+ {
+ "epoch": 27.600873362445416,
+ "grad_norm": 0.019535524770617485,
+ "learning_rate": 0.0006,
+ "loss": 4.321477890014648,
+ "step": 1987
+ },
+ {
+ "epoch": 27.61484716157205,
+ "grad_norm": 0.01997970975935459,
+ "learning_rate": 0.0006,
+ "loss": 4.123082637786865,
+ "step": 1988
+ },
+ {
+ "epoch": 27.62882096069869,
+ "grad_norm": 0.01913067139685154,
+ "learning_rate": 0.0006,
+ "loss": 4.289896488189697,
+ "step": 1989
+ },
+ {
+ "epoch": 27.64279475982533,
+ "grad_norm": 0.0169806070625782,
+ "learning_rate": 0.0006,
+ "loss": 4.247004985809326,
+ "step": 1990
+ },
+ {
+ "epoch": 27.656768558951963,
+ "grad_norm": 0.017315007746219635,
+ "learning_rate": 0.0006,
+ "loss": 4.2095441818237305,
+ "step": 1991
+ },
+ {
+ "epoch": 27.670742358078602,
+ "grad_norm": 0.015098759904503822,
+ "learning_rate": 0.0006,
+ "loss": 4.145560264587402,
+ "step": 1992
+ },
+ {
+ "epoch": 27.68471615720524,
+ "grad_norm": 0.016240514814853668,
+ "learning_rate": 0.0006,
+ "loss": 4.254927635192871,
+ "step": 1993
+ },
+ {
+ "epoch": 27.69868995633188,
+ "grad_norm": 0.01621108129620552,
+ "learning_rate": 0.0006,
+ "loss": 4.219725608825684,
+ "step": 1994
+ },
+ {
+ "epoch": 27.712663755458514,
+ "grad_norm": 0.015191009268164635,
+ "learning_rate": 0.0006,
+ "loss": 4.194345474243164,
+ "step": 1995
+ },
+ {
+ "epoch": 27.726637554585153,
+ "grad_norm": 0.014387983828783035,
+ "learning_rate": 0.0006,
+ "loss": 4.211467742919922,
+ "step": 1996
+ },
+ {
+ "epoch": 27.74061135371179,
+ "grad_norm": 0.015460561029613018,
+ "learning_rate": 0.0006,
+ "loss": 4.358700752258301,
+ "step": 1997
+ },
+ {
+ "epoch": 27.754585152838427,
+ "grad_norm": 0.015100076794624329,
+ "learning_rate": 0.0006,
+ "loss": 4.171223163604736,
+ "step": 1998
+ },
+ {
+ "epoch": 27.768558951965066,
+ "grad_norm": 0.015607891604304314,
+ "learning_rate": 0.0006,
+ "loss": 4.217874526977539,
+ "step": 1999
+ },
+ {
+ "epoch": 27.782532751091704,
+ "grad_norm": 0.015650689601898193,
+ "learning_rate": 0.0006,
+ "loss": 4.280780792236328,
+ "step": 2000
+ },
+ {
+ "epoch": 27.79650655021834,
+ "grad_norm": 0.01580994389951229,
+ "learning_rate": 0.0006,
+ "loss": 4.159212112426758,
+ "step": 2001
+ },
+ {
+ "epoch": 27.810480349344978,
+ "grad_norm": 0.016607413068413734,
+ "learning_rate": 0.0006,
+ "loss": 4.238853454589844,
+ "step": 2002
+ },
+ {
+ "epoch": 27.824454148471617,
+ "grad_norm": 0.017383860424160957,
+ "learning_rate": 0.0006,
+ "loss": 4.233944892883301,
+ "step": 2003
+ },
+ {
+ "epoch": 27.83842794759825,
+ "grad_norm": 0.014966659247875214,
+ "learning_rate": 0.0006,
+ "loss": 4.167392730712891,
+ "step": 2004
+ },
+ {
+ "epoch": 27.85240174672489,
+ "grad_norm": 0.014589878730475903,
+ "learning_rate": 0.0006,
+ "loss": 4.2390899658203125,
+ "step": 2005
+ },
+ {
+ "epoch": 27.86637554585153,
+ "grad_norm": 0.017179450020194054,
+ "learning_rate": 0.0006,
+ "loss": 4.365785598754883,
+ "step": 2006
+ },
+ {
+ "epoch": 27.880349344978168,
+ "grad_norm": 0.01893901452422142,
+ "learning_rate": 0.0006,
+ "loss": 4.289430141448975,
+ "step": 2007
+ },
+ {
+ "epoch": 27.894323144104803,
+ "grad_norm": 0.018619712442159653,
+ "learning_rate": 0.0006,
+ "loss": 4.310665130615234,
+ "step": 2008
+ },
+ {
+ "epoch": 27.90829694323144,
+ "grad_norm": 0.019180385395884514,
+ "learning_rate": 0.0006,
+ "loss": 4.27454137802124,
+ "step": 2009
+ },
+ {
+ "epoch": 27.92227074235808,
+ "grad_norm": 0.017811523750424385,
+ "learning_rate": 0.0006,
+ "loss": 4.312562465667725,
+ "step": 2010
+ },
+ {
+ "epoch": 27.936244541484715,
+ "grad_norm": 0.016583573073148727,
+ "learning_rate": 0.0006,
+ "loss": 4.202253341674805,
+ "step": 2011
+ },
+ {
+ "epoch": 27.950218340611354,
+ "grad_norm": 0.01658615842461586,
+ "learning_rate": 0.0006,
+ "loss": 4.183433532714844,
+ "step": 2012
+ },
+ {
+ "epoch": 27.964192139737992,
+ "grad_norm": 0.015327401459217072,
+ "learning_rate": 0.0006,
+ "loss": 4.2125139236450195,
+ "step": 2013
+ },
+ {
+ "epoch": 27.978165938864628,
+ "grad_norm": 0.015004601329565048,
+ "learning_rate": 0.0006,
+ "loss": 4.257392883300781,
+ "step": 2014
+ },
+ {
+ "epoch": 27.992139737991266,
+ "grad_norm": 0.015259161591529846,
+ "learning_rate": 0.0006,
+ "loss": 4.241818428039551,
+ "step": 2015
+ },
+ {
+ "epoch": 28.0,
+ "grad_norm": 0.016808776184916496,
+ "learning_rate": 0.0006,
+ "loss": 4.146241188049316,
+ "step": 2016
+ },
+ {
+ "epoch": 28.0,
+ "eval_loss": 4.663090705871582,
+ "eval_runtime": 56.4559,
+ "eval_samples_per_second": 43.255,
+ "eval_steps_per_second": 1.364,
+ "step": 2016
+ },
+ {
+ "epoch": 28.01397379912664,
+ "grad_norm": 0.015784382820129395,
+ "learning_rate": 0.0006,
+ "loss": 4.044191360473633,
+ "step": 2017
+ },
+ {
+ "epoch": 28.027947598253274,
+ "grad_norm": 0.015636710450053215,
+ "learning_rate": 0.0006,
+ "loss": 4.128754615783691,
+ "step": 2018
+ },
+ {
+ "epoch": 28.041921397379912,
+ "grad_norm": 0.015364975668489933,
+ "learning_rate": 0.0006,
+ "loss": 4.191835403442383,
+ "step": 2019
+ },
+ {
+ "epoch": 28.05589519650655,
+ "grad_norm": 0.015015700832009315,
+ "learning_rate": 0.0006,
+ "loss": 4.120022773742676,
+ "step": 2020
+ },
+ {
+ "epoch": 28.069868995633186,
+ "grad_norm": 0.016666799783706665,
+ "learning_rate": 0.0006,
+ "loss": 4.33906364440918,
+ "step": 2021
+ },
+ {
+ "epoch": 28.083842794759825,
+ "grad_norm": 0.01780104823410511,
+ "learning_rate": 0.0006,
+ "loss": 4.279458999633789,
+ "step": 2022
+ },
+ {
+ "epoch": 28.097816593886463,
+ "grad_norm": 0.01870882883667946,
+ "learning_rate": 0.0006,
+ "loss": 4.34728479385376,
+ "step": 2023
+ },
+ {
+ "epoch": 28.111790393013102,
+ "grad_norm": 0.018406381830573082,
+ "learning_rate": 0.0006,
+ "loss": 4.211311340332031,
+ "step": 2024
+ },
+ {
+ "epoch": 28.125764192139737,
+ "grad_norm": 0.017833448946475983,
+ "learning_rate": 0.0006,
+ "loss": 4.223254203796387,
+ "step": 2025
+ },
+ {
+ "epoch": 28.139737991266376,
+ "grad_norm": 0.017552688717842102,
+ "learning_rate": 0.0006,
+ "loss": 4.208823204040527,
+ "step": 2026
+ },
+ {
+ "epoch": 28.153711790393015,
+ "grad_norm": 0.018500229343771935,
+ "learning_rate": 0.0006,
+ "loss": 4.198025226593018,
+ "step": 2027
+ },
+ {
+ "epoch": 28.16768558951965,
+ "grad_norm": 0.01782156340777874,
+ "learning_rate": 0.0006,
+ "loss": 4.179978370666504,
+ "step": 2028
+ },
+ {
+ "epoch": 28.18165938864629,
+ "grad_norm": 0.019580967724323273,
+ "learning_rate": 0.0006,
+ "loss": 4.291114807128906,
+ "step": 2029
+ },
+ {
+ "epoch": 28.195633187772927,
+ "grad_norm": 0.020804036408662796,
+ "learning_rate": 0.0006,
+ "loss": 4.220607280731201,
+ "step": 2030
+ },
+ {
+ "epoch": 28.209606986899562,
+ "grad_norm": 0.019595693796873093,
+ "learning_rate": 0.0006,
+ "loss": 4.2905731201171875,
+ "step": 2031
+ },
+ {
+ "epoch": 28.2235807860262,
+ "grad_norm": 0.019247086718678474,
+ "learning_rate": 0.0006,
+ "loss": 4.2467360496521,
+ "step": 2032
+ },
+ {
+ "epoch": 28.23755458515284,
+ "grad_norm": 0.01958036608994007,
+ "learning_rate": 0.0006,
+ "loss": 4.286137580871582,
+ "step": 2033
+ },
+ {
+ "epoch": 28.251528384279474,
+ "grad_norm": 0.021354753524065018,
+ "learning_rate": 0.0006,
+ "loss": 4.234755039215088,
+ "step": 2034
+ },
+ {
+ "epoch": 28.265502183406113,
+ "grad_norm": 0.02206375077366829,
+ "learning_rate": 0.0006,
+ "loss": 4.149641036987305,
+ "step": 2035
+ },
+ {
+ "epoch": 28.27947598253275,
+ "grad_norm": 0.02201310358941555,
+ "learning_rate": 0.0006,
+ "loss": 4.1928629875183105,
+ "step": 2036
+ },
+ {
+ "epoch": 28.29344978165939,
+ "grad_norm": 0.01961950585246086,
+ "learning_rate": 0.0006,
+ "loss": 4.211404323577881,
+ "step": 2037
+ },
+ {
+ "epoch": 28.307423580786025,
+ "grad_norm": 0.019211795181035995,
+ "learning_rate": 0.0006,
+ "loss": 4.200976848602295,
+ "step": 2038
+ },
+ {
+ "epoch": 28.321397379912664,
+ "grad_norm": 0.01779370754957199,
+ "learning_rate": 0.0006,
+ "loss": 4.249148845672607,
+ "step": 2039
+ },
+ {
+ "epoch": 28.335371179039303,
+ "grad_norm": 0.018635908141732216,
+ "learning_rate": 0.0006,
+ "loss": 4.11806058883667,
+ "step": 2040
+ },
+ {
+ "epoch": 28.349344978165938,
+ "grad_norm": 0.019792694598436356,
+ "learning_rate": 0.0006,
+ "loss": 4.335224151611328,
+ "step": 2041
+ },
+ {
+ "epoch": 28.363318777292577,
+ "grad_norm": 0.021731717512011528,
+ "learning_rate": 0.0006,
+ "loss": 4.302778244018555,
+ "step": 2042
+ },
+ {
+ "epoch": 28.377292576419215,
+ "grad_norm": 0.020823447033762932,
+ "learning_rate": 0.0006,
+ "loss": 4.260610580444336,
+ "step": 2043
+ },
+ {
+ "epoch": 28.39126637554585,
+ "grad_norm": 0.019385412335395813,
+ "learning_rate": 0.0006,
+ "loss": 4.2247724533081055,
+ "step": 2044
+ },
+ {
+ "epoch": 28.40524017467249,
+ "grad_norm": 0.019539108499884605,
+ "learning_rate": 0.0006,
+ "loss": 4.236655235290527,
+ "step": 2045
+ },
+ {
+ "epoch": 28.419213973799128,
+ "grad_norm": 0.01942690648138523,
+ "learning_rate": 0.0006,
+ "loss": 4.251850605010986,
+ "step": 2046
+ },
+ {
+ "epoch": 28.433187772925763,
+ "grad_norm": 0.019667886197566986,
+ "learning_rate": 0.0006,
+ "loss": 4.222312927246094,
+ "step": 2047
+ },
+ {
+ "epoch": 28.4471615720524,
+ "grad_norm": 0.02029012329876423,
+ "learning_rate": 0.0006,
+ "loss": 4.246252059936523,
+ "step": 2048
+ },
+ {
+ "epoch": 28.46113537117904,
+ "grad_norm": 0.01784469373524189,
+ "learning_rate": 0.0006,
+ "loss": 4.192128658294678,
+ "step": 2049
+ },
+ {
+ "epoch": 28.475109170305675,
+ "grad_norm": 0.0160287544131279,
+ "learning_rate": 0.0006,
+ "loss": 4.208607196807861,
+ "step": 2050
+ },
+ {
+ "epoch": 28.489082969432314,
+ "grad_norm": 0.015348346903920174,
+ "learning_rate": 0.0006,
+ "loss": 4.236598014831543,
+ "step": 2051
+ },
+ {
+ "epoch": 28.503056768558952,
+ "grad_norm": 0.015198206529021263,
+ "learning_rate": 0.0006,
+ "loss": 4.235552787780762,
+ "step": 2052
+ },
+ {
+ "epoch": 28.51703056768559,
+ "grad_norm": 0.01523754745721817,
+ "learning_rate": 0.0006,
+ "loss": 4.1941728591918945,
+ "step": 2053
+ },
+ {
+ "epoch": 28.531004366812226,
+ "grad_norm": 0.015559614636003971,
+ "learning_rate": 0.0006,
+ "loss": 4.294757843017578,
+ "step": 2054
+ },
+ {
+ "epoch": 28.544978165938865,
+ "grad_norm": 0.01568358950316906,
+ "learning_rate": 0.0006,
+ "loss": 4.213065147399902,
+ "step": 2055
+ },
+ {
+ "epoch": 28.558951965065503,
+ "grad_norm": 0.015336516313254833,
+ "learning_rate": 0.0006,
+ "loss": 4.240293979644775,
+ "step": 2056
+ },
+ {
+ "epoch": 28.57292576419214,
+ "grad_norm": 0.015041496604681015,
+ "learning_rate": 0.0006,
+ "loss": 4.273397445678711,
+ "step": 2057
+ },
+ {
+ "epoch": 28.586899563318777,
+ "grad_norm": 0.01610128954052925,
+ "learning_rate": 0.0006,
+ "loss": 4.125369071960449,
+ "step": 2058
+ },
+ {
+ "epoch": 28.600873362445416,
+ "grad_norm": 0.015974976122379303,
+ "learning_rate": 0.0006,
+ "loss": 4.133459091186523,
+ "step": 2059
+ },
+ {
+ "epoch": 28.61484716157205,
+ "grad_norm": 0.015936799347400665,
+ "learning_rate": 0.0006,
+ "loss": 4.306667327880859,
+ "step": 2060
+ },
+ {
+ "epoch": 28.62882096069869,
+ "grad_norm": 0.01795247197151184,
+ "learning_rate": 0.0006,
+ "loss": 4.205156326293945,
+ "step": 2061
+ },
+ {
+ "epoch": 28.64279475982533,
+ "grad_norm": 0.019639814272522926,
+ "learning_rate": 0.0006,
+ "loss": 4.277614593505859,
+ "step": 2062
+ },
+ {
+ "epoch": 28.656768558951963,
+ "grad_norm": 0.01722300611436367,
+ "learning_rate": 0.0006,
+ "loss": 4.191032886505127,
+ "step": 2063
+ },
+ {
+ "epoch": 28.670742358078602,
+ "grad_norm": 0.01612633652985096,
+ "learning_rate": 0.0006,
+ "loss": 4.20717191696167,
+ "step": 2064
+ },
+ {
+ "epoch": 28.68471615720524,
+ "grad_norm": 0.01603279449045658,
+ "learning_rate": 0.0006,
+ "loss": 4.228884220123291,
+ "step": 2065
+ },
+ {
+ "epoch": 28.69868995633188,
+ "grad_norm": 0.016405927017331123,
+ "learning_rate": 0.0006,
+ "loss": 4.201292991638184,
+ "step": 2066
+ },
+ {
+ "epoch": 28.712663755458514,
+ "grad_norm": 0.016644183546304703,
+ "learning_rate": 0.0006,
+ "loss": 4.239282131195068,
+ "step": 2067
+ },
+ {
+ "epoch": 28.726637554585153,
+ "grad_norm": 0.0175587497651577,
+ "learning_rate": 0.0006,
+ "loss": 4.289038181304932,
+ "step": 2068
+ },
+ {
+ "epoch": 28.74061135371179,
+ "grad_norm": 0.018223397433757782,
+ "learning_rate": 0.0006,
+ "loss": 4.2266998291015625,
+ "step": 2069
+ },
+ {
+ "epoch": 28.754585152838427,
+ "grad_norm": 0.018680868670344353,
+ "learning_rate": 0.0006,
+ "loss": 4.348849296569824,
+ "step": 2070
+ },
+ {
+ "epoch": 28.768558951965066,
+ "grad_norm": 0.016726728528738022,
+ "learning_rate": 0.0006,
+ "loss": 4.292080879211426,
+ "step": 2071
+ },
+ {
+ "epoch": 28.782532751091704,
+ "grad_norm": 0.017386795952916145,
+ "learning_rate": 0.0006,
+ "loss": 4.208870887756348,
+ "step": 2072
+ },
+ {
+ "epoch": 28.79650655021834,
+ "grad_norm": 0.01865178905427456,
+ "learning_rate": 0.0006,
+ "loss": 4.054150581359863,
+ "step": 2073
+ },
+ {
+ "epoch": 28.810480349344978,
+ "grad_norm": 0.01776815392076969,
+ "learning_rate": 0.0006,
+ "loss": 4.202404499053955,
+ "step": 2074
+ },
+ {
+ "epoch": 28.824454148471617,
+ "grad_norm": 0.01643497310578823,
+ "learning_rate": 0.0006,
+ "loss": 4.131855010986328,
+ "step": 2075
+ },
+ {
+ "epoch": 28.83842794759825,
+ "grad_norm": 0.01574273779988289,
+ "learning_rate": 0.0006,
+ "loss": 4.296788692474365,
+ "step": 2076
+ },
+ {
+ "epoch": 28.85240174672489,
+ "grad_norm": 0.015613908879458904,
+ "learning_rate": 0.0006,
+ "loss": 4.210949897766113,
+ "step": 2077
+ },
+ {
+ "epoch": 28.86637554585153,
+ "grad_norm": 0.015543212182819843,
+ "learning_rate": 0.0006,
+ "loss": 4.218136310577393,
+ "step": 2078
+ },
+ {
+ "epoch": 28.880349344978168,
+ "grad_norm": 0.014681251719594002,
+ "learning_rate": 0.0006,
+ "loss": 4.19401741027832,
+ "step": 2079
+ },
+ {
+ "epoch": 28.894323144104803,
+ "grad_norm": 0.014393828809261322,
+ "learning_rate": 0.0006,
+ "loss": 4.198468208312988,
+ "step": 2080
+ },
+ {
+ "epoch": 28.90829694323144,
+ "grad_norm": 0.01580170728266239,
+ "learning_rate": 0.0006,
+ "loss": 4.224969863891602,
+ "step": 2081
+ },
+ {
+ "epoch": 28.92227074235808,
+ "grad_norm": 0.01484165620058775,
+ "learning_rate": 0.0006,
+ "loss": 4.12236213684082,
+ "step": 2082
+ },
+ {
+ "epoch": 28.936244541484715,
+ "grad_norm": 0.01497623696923256,
+ "learning_rate": 0.0006,
+ "loss": 4.24141788482666,
+ "step": 2083
+ },
+ {
+ "epoch": 28.950218340611354,
+ "grad_norm": 0.015080046840012074,
+ "learning_rate": 0.0006,
+ "loss": 4.284151077270508,
+ "step": 2084
+ },
+ {
+ "epoch": 28.964192139737992,
+ "grad_norm": 0.016207212582230568,
+ "learning_rate": 0.0006,
+ "loss": 4.298693656921387,
+ "step": 2085
+ },
+ {
+ "epoch": 28.978165938864628,
+ "grad_norm": 0.015474417246878147,
+ "learning_rate": 0.0006,
+ "loss": 4.2412261962890625,
+ "step": 2086
+ },
+ {
+ "epoch": 28.992139737991266,
+ "grad_norm": 0.01515351701527834,
+ "learning_rate": 0.0006,
+ "loss": 4.3054423332214355,
+ "step": 2087
+ },
+ {
+ "epoch": 29.0,
+ "grad_norm": 0.016411345452070236,
+ "learning_rate": 0.0006,
+ "loss": 4.104830741882324,
+ "step": 2088
+ },
+ {
+ "epoch": 29.0,
+ "eval_loss": 4.661567687988281,
+ "eval_runtime": 57.1244,
+ "eval_samples_per_second": 42.749,
+ "eval_steps_per_second": 1.348,
+ "step": 2088
+ },
+ {
+ "epoch": 29.01397379912664,
+ "grad_norm": 0.01665342226624489,
+ "learning_rate": 0.0006,
+ "loss": 4.311391830444336,
+ "step": 2089
+ },
+ {
+ "epoch": 29.027947598253274,
+ "grad_norm": 0.018053214997053146,
+ "learning_rate": 0.0006,
+ "loss": 4.135597229003906,
+ "step": 2090
+ },
+ {
+ "epoch": 29.041921397379912,
+ "grad_norm": 0.018213748931884766,
+ "learning_rate": 0.0006,
+ "loss": 4.303299903869629,
+ "step": 2091
+ },
+ {
+ "epoch": 29.05589519650655,
+ "grad_norm": 0.018460242077708244,
+ "learning_rate": 0.0006,
+ "loss": 4.20850944519043,
+ "step": 2092
+ },
+ {
+ "epoch": 29.069868995633186,
+ "grad_norm": 0.018365809693932533,
+ "learning_rate": 0.0006,
+ "loss": 4.275847434997559,
+ "step": 2093
+ },
+ {
+ "epoch": 29.083842794759825,
+ "grad_norm": 0.018346186727285385,
+ "learning_rate": 0.0006,
+ "loss": 4.202091217041016,
+ "step": 2094
+ },
+ {
+ "epoch": 29.097816593886463,
+ "grad_norm": 0.017920587211847305,
+ "learning_rate": 0.0006,
+ "loss": 4.182092666625977,
+ "step": 2095
+ },
+ {
+ "epoch": 29.111790393013102,
+ "grad_norm": 0.01812003180384636,
+ "learning_rate": 0.0006,
+ "loss": 4.259721755981445,
+ "step": 2096
+ },
+ {
+ "epoch": 29.125764192139737,
+ "grad_norm": 0.01700986735522747,
+ "learning_rate": 0.0006,
+ "loss": 4.067573070526123,
+ "step": 2097
+ },
+ {
+ "epoch": 29.139737991266376,
+ "grad_norm": 0.018069760873913765,
+ "learning_rate": 0.0006,
+ "loss": 4.234400749206543,
+ "step": 2098
+ },
+ {
+ "epoch": 29.153711790393015,
+ "grad_norm": 0.01905428245663643,
+ "learning_rate": 0.0006,
+ "loss": 4.16609001159668,
+ "step": 2099
+ },
+ {
+ "epoch": 29.16768558951965,
+ "grad_norm": 0.018711242824792862,
+ "learning_rate": 0.0006,
+ "loss": 4.249088764190674,
+ "step": 2100
+ },
+ {
+ "epoch": 29.18165938864629,
+ "grad_norm": 0.018721306696534157,
+ "learning_rate": 0.0006,
+ "loss": 4.14540958404541,
+ "step": 2101
+ },
+ {
+ "epoch": 29.195633187772927,
+ "grad_norm": 0.020139874890446663,
+ "learning_rate": 0.0006,
+ "loss": 4.049248695373535,
+ "step": 2102
+ },
+ {
+ "epoch": 29.209606986899562,
+ "grad_norm": 0.020849574357271194,
+ "learning_rate": 0.0006,
+ "loss": 4.235224723815918,
+ "step": 2103
+ },
+ {
+ "epoch": 29.2235807860262,
+ "grad_norm": 0.018693549558520317,
+ "learning_rate": 0.0006,
+ "loss": 4.232961177825928,
+ "step": 2104
+ },
+ {
+ "epoch": 29.23755458515284,
+ "grad_norm": 0.017889857292175293,
+ "learning_rate": 0.0006,
+ "loss": 4.171422004699707,
+ "step": 2105
+ },
+ {
+ "epoch": 29.251528384279474,
+ "grad_norm": 0.016690224409103394,
+ "learning_rate": 0.0006,
+ "loss": 4.244717597961426,
+ "step": 2106
+ },
+ {
+ "epoch": 29.265502183406113,
+ "grad_norm": 0.01708616502583027,
+ "learning_rate": 0.0006,
+ "loss": 4.125068664550781,
+ "step": 2107
+ },
+ {
+ "epoch": 29.27947598253275,
+ "grad_norm": 0.017928237095475197,
+ "learning_rate": 0.0006,
+ "loss": 4.208292007446289,
+ "step": 2108
+ },
+ {
+ "epoch": 29.29344978165939,
+ "grad_norm": 0.018250394612550735,
+ "learning_rate": 0.0006,
+ "loss": 4.118630886077881,
+ "step": 2109
+ },
+ {
+ "epoch": 29.307423580786025,
+ "grad_norm": 0.020160246640443802,
+ "learning_rate": 0.0006,
+ "loss": 4.206025123596191,
+ "step": 2110
+ },
+ {
+ "epoch": 29.321397379912664,
+ "grad_norm": 0.020045453682541847,
+ "learning_rate": 0.0006,
+ "loss": 4.2758684158325195,
+ "step": 2111
+ },
+ {
+ "epoch": 29.335371179039303,
+ "grad_norm": 0.019556907936930656,
+ "learning_rate": 0.0006,
+ "loss": 4.188272953033447,
+ "step": 2112
+ },
+ {
+ "epoch": 29.349344978165938,
+ "grad_norm": 0.019185200333595276,
+ "learning_rate": 0.0006,
+ "loss": 4.17451286315918,
+ "step": 2113
+ },
+ {
+ "epoch": 29.363318777292577,
+ "grad_norm": 0.01816629432141781,
+ "learning_rate": 0.0006,
+ "loss": 4.165246963500977,
+ "step": 2114
+ },
+ {
+ "epoch": 29.377292576419215,
+ "grad_norm": 0.01865146867930889,
+ "learning_rate": 0.0006,
+ "loss": 4.148184776306152,
+ "step": 2115
+ },
+ {
+ "epoch": 29.39126637554585,
+ "grad_norm": 0.018338464200496674,
+ "learning_rate": 0.0006,
+ "loss": 4.288936614990234,
+ "step": 2116
+ },
+ {
+ "epoch": 29.40524017467249,
+ "grad_norm": 0.018986187875270844,
+ "learning_rate": 0.0006,
+ "loss": 4.183379650115967,
+ "step": 2117
+ },
+ {
+ "epoch": 29.419213973799128,
+ "grad_norm": 0.018409637734293938,
+ "learning_rate": 0.0006,
+ "loss": 4.057574272155762,
+ "step": 2118
+ },
+ {
+ "epoch": 29.433187772925763,
+ "grad_norm": 0.016664542257785797,
+ "learning_rate": 0.0006,
+ "loss": 4.18517541885376,
+ "step": 2119
+ },
+ {
+ "epoch": 29.4471615720524,
+ "grad_norm": 0.016486089676618576,
+ "learning_rate": 0.0006,
+ "loss": 4.179367542266846,
+ "step": 2120
+ },
+ {
+ "epoch": 29.46113537117904,
+ "grad_norm": 0.01787523552775383,
+ "learning_rate": 0.0006,
+ "loss": 4.184126377105713,
+ "step": 2121
+ },
+ {
+ "epoch": 29.475109170305675,
+ "grad_norm": 0.016574613749980927,
+ "learning_rate": 0.0006,
+ "loss": 4.290170192718506,
+ "step": 2122
+ },
+ {
+ "epoch": 29.489082969432314,
+ "grad_norm": 0.01643376611173153,
+ "learning_rate": 0.0006,
+ "loss": 4.052913665771484,
+ "step": 2123
+ },
+ {
+ "epoch": 29.503056768558952,
+ "grad_norm": 0.016503969207406044,
+ "learning_rate": 0.0006,
+ "loss": 4.335184097290039,
+ "step": 2124
+ },
+ {
+ "epoch": 29.51703056768559,
+ "grad_norm": 0.017596479505300522,
+ "learning_rate": 0.0006,
+ "loss": 4.165539741516113,
+ "step": 2125
+ },
+ {
+ "epoch": 29.531004366812226,
+ "grad_norm": 0.017468582838773727,
+ "learning_rate": 0.0006,
+ "loss": 4.315964698791504,
+ "step": 2126
+ },
+ {
+ "epoch": 29.544978165938865,
+ "grad_norm": 0.01724618673324585,
+ "learning_rate": 0.0006,
+ "loss": 4.243681907653809,
+ "step": 2127
+ },
+ {
+ "epoch": 29.558951965065503,
+ "grad_norm": 0.018384616822004318,
+ "learning_rate": 0.0006,
+ "loss": 4.112447261810303,
+ "step": 2128
+ },
+ {
+ "epoch": 29.57292576419214,
+ "grad_norm": 0.01824074238538742,
+ "learning_rate": 0.0006,
+ "loss": 4.236065864562988,
+ "step": 2129
+ },
+ {
+ "epoch": 29.586899563318777,
+ "grad_norm": 0.01625971868634224,
+ "learning_rate": 0.0006,
+ "loss": 4.18143892288208,
+ "step": 2130
+ },
+ {
+ "epoch": 29.600873362445416,
+ "grad_norm": 0.015537494793534279,
+ "learning_rate": 0.0006,
+ "loss": 4.225147247314453,
+ "step": 2131
+ },
+ {
+ "epoch": 29.61484716157205,
+ "grad_norm": 0.015818974003195763,
+ "learning_rate": 0.0006,
+ "loss": 4.3100175857543945,
+ "step": 2132
+ },
+ {
+ "epoch": 29.62882096069869,
+ "grad_norm": 0.015909671783447266,
+ "learning_rate": 0.0006,
+ "loss": 4.169775009155273,
+ "step": 2133
+ },
+ {
+ "epoch": 29.64279475982533,
+ "grad_norm": 0.01769658550620079,
+ "learning_rate": 0.0006,
+ "loss": 4.278904914855957,
+ "step": 2134
+ },
+ {
+ "epoch": 29.656768558951963,
+ "grad_norm": 0.015641704201698303,
+ "learning_rate": 0.0006,
+ "loss": 4.2426228523254395,
+ "step": 2135
+ },
+ {
+ "epoch": 29.670742358078602,
+ "grad_norm": 0.01565658301115036,
+ "learning_rate": 0.0006,
+ "loss": 4.232687950134277,
+ "step": 2136
+ },
+ {
+ "epoch": 29.68471615720524,
+ "grad_norm": 0.016485940665006638,
+ "learning_rate": 0.0006,
+ "loss": 4.209839820861816,
+ "step": 2137
+ },
+ {
+ "epoch": 29.69868995633188,
+ "grad_norm": 0.017622729763388634,
+ "learning_rate": 0.0006,
+ "loss": 4.196017265319824,
+ "step": 2138
+ },
+ {
+ "epoch": 29.712663755458514,
+ "grad_norm": 0.018320564180612564,
+ "learning_rate": 0.0006,
+ "loss": 4.225683689117432,
+ "step": 2139
+ },
+ {
+ "epoch": 29.726637554585153,
+ "grad_norm": 0.018140794709324837,
+ "learning_rate": 0.0006,
+ "loss": 4.311944961547852,
+ "step": 2140
+ },
+ {
+ "epoch": 29.74061135371179,
+ "grad_norm": 0.016227100044488907,
+ "learning_rate": 0.0006,
+ "loss": 4.230257511138916,
+ "step": 2141
+ },
+ {
+ "epoch": 29.754585152838427,
+ "grad_norm": 0.015759488567709923,
+ "learning_rate": 0.0006,
+ "loss": 4.119174003601074,
+ "step": 2142
+ },
+ {
+ "epoch": 29.768558951965066,
+ "grad_norm": 0.01757766678929329,
+ "learning_rate": 0.0006,
+ "loss": 4.165826797485352,
+ "step": 2143
+ },
+ {
+ "epoch": 29.782532751091704,
+ "grad_norm": 0.019661923870444298,
+ "learning_rate": 0.0006,
+ "loss": 4.392203330993652,
+ "step": 2144
+ },
+ {
+ "epoch": 29.79650655021834,
+ "grad_norm": 0.019478755071759224,
+ "learning_rate": 0.0006,
+ "loss": 4.348371982574463,
+ "step": 2145
+ },
+ {
+ "epoch": 29.810480349344978,
+ "grad_norm": 0.019149569794535637,
+ "learning_rate": 0.0006,
+ "loss": 4.34104061126709,
+ "step": 2146
+ },
+ {
+ "epoch": 29.824454148471617,
+ "grad_norm": 0.017707521095871925,
+ "learning_rate": 0.0006,
+ "loss": 4.228226661682129,
+ "step": 2147
+ },
+ {
+ "epoch": 29.83842794759825,
+ "grad_norm": 0.015615378506481647,
+ "learning_rate": 0.0006,
+ "loss": 4.175541877746582,
+ "step": 2148
+ },
+ {
+ "epoch": 29.85240174672489,
+ "grad_norm": 0.016111129894852638,
+ "learning_rate": 0.0006,
+ "loss": 4.267936706542969,
+ "step": 2149
+ },
+ {
+ "epoch": 29.86637554585153,
+ "grad_norm": 0.01528779324144125,
+ "learning_rate": 0.0006,
+ "loss": 4.186014175415039,
+ "step": 2150
+ },
+ {
+ "epoch": 29.880349344978168,
+ "grad_norm": 0.014925646595656872,
+ "learning_rate": 0.0006,
+ "loss": 4.174860000610352,
+ "step": 2151
+ },
+ {
+ "epoch": 29.894323144104803,
+ "grad_norm": 0.016917673870921135,
+ "learning_rate": 0.0006,
+ "loss": 4.33908748626709,
+ "step": 2152
+ },
+ {
+ "epoch": 29.90829694323144,
+ "grad_norm": 0.016007075086236,
+ "learning_rate": 0.0006,
+ "loss": 4.266141414642334,
+ "step": 2153
+ },
+ {
+ "epoch": 29.92227074235808,
+ "grad_norm": 0.016178956255316734,
+ "learning_rate": 0.0006,
+ "loss": 4.271215915679932,
+ "step": 2154
+ },
+ {
+ "epoch": 29.936244541484715,
+ "grad_norm": 0.016180871054530144,
+ "learning_rate": 0.0006,
+ "loss": 4.278801441192627,
+ "step": 2155
+ },
+ {
+ "epoch": 29.950218340611354,
+ "grad_norm": 0.015335503034293652,
+ "learning_rate": 0.0006,
+ "loss": 4.194978713989258,
+ "step": 2156
+ },
+ {
+ "epoch": 29.964192139737992,
+ "grad_norm": 0.016580764204263687,
+ "learning_rate": 0.0006,
+ "loss": 4.290809631347656,
+ "step": 2157
+ },
+ {
+ "epoch": 29.978165938864628,
+ "grad_norm": 0.01623128168284893,
+ "learning_rate": 0.0006,
+ "loss": 4.205451488494873,
+ "step": 2158
+ },
+ {
+ "epoch": 29.992139737991266,
+ "grad_norm": 0.014904248528182507,
+ "learning_rate": 0.0006,
+ "loss": 4.24924373626709,
+ "step": 2159
+ },
+ {
+ "epoch": 30.0,
+ "grad_norm": 0.017089299857616425,
+ "learning_rate": 0.0006,
+ "loss": 4.222072601318359,
+ "step": 2160
+ },
+ {
+ "epoch": 30.0,
+ "eval_loss": 4.585411071777344,
+ "eval_runtime": 57.1164,
+ "eval_samples_per_second": 42.755,
+ "eval_steps_per_second": 1.348,
+ "step": 2160
+ },
+ {
+ "epoch": 30.01397379912664,
+ "grad_norm": 0.017226383090019226,
+ "learning_rate": 0.0006,
+ "loss": 4.19467830657959,
+ "step": 2161
+ },
+ {
+ "epoch": 30.027947598253274,
+ "grad_norm": 0.01635066606104374,
+ "learning_rate": 0.0006,
+ "loss": 4.176875591278076,
+ "step": 2162
+ },
+ {
+ "epoch": 30.041921397379912,
+ "grad_norm": 0.01735362410545349,
+ "learning_rate": 0.0006,
+ "loss": 4.13667106628418,
+ "step": 2163
+ },
+ {
+ "epoch": 30.05589519650655,
+ "grad_norm": 0.017047109082341194,
+ "learning_rate": 0.0006,
+ "loss": 4.136329650878906,
+ "step": 2164
+ },
+ {
+ "epoch": 30.069868995633186,
+ "grad_norm": 0.018116053193807602,
+ "learning_rate": 0.0006,
+ "loss": 4.228706359863281,
+ "step": 2165
+ },
+ {
+ "epoch": 30.083842794759825,
+ "grad_norm": 0.01791389286518097,
+ "learning_rate": 0.0006,
+ "loss": 4.1597747802734375,
+ "step": 2166
+ },
+ {
+ "epoch": 30.097816593886463,
+ "grad_norm": 0.017793502658605576,
+ "learning_rate": 0.0006,
+ "loss": 4.0807976722717285,
+ "step": 2167
+ },
+ {
+ "epoch": 30.111790393013102,
+ "grad_norm": 0.01827503927052021,
+ "learning_rate": 0.0006,
+ "loss": 4.170646667480469,
+ "step": 2168
+ },
+ {
+ "epoch": 30.125764192139737,
+ "grad_norm": 0.017228564247488976,
+ "learning_rate": 0.0006,
+ "loss": 4.146934509277344,
+ "step": 2169
+ },
+ {
+ "epoch": 30.139737991266376,
+ "grad_norm": 0.01757437363266945,
+ "learning_rate": 0.0006,
+ "loss": 4.1859130859375,
+ "step": 2170
+ },
+ {
+ "epoch": 30.153711790393015,
+ "grad_norm": 0.01701093092560768,
+ "learning_rate": 0.0006,
+ "loss": 4.126981258392334,
+ "step": 2171
+ },
+ {
+ "epoch": 30.16768558951965,
+ "grad_norm": 0.015220489352941513,
+ "learning_rate": 0.0006,
+ "loss": 4.164018630981445,
+ "step": 2172
+ },
+ {
+ "epoch": 30.18165938864629,
+ "grad_norm": 0.016254238784313202,
+ "learning_rate": 0.0006,
+ "loss": 4.258275508880615,
+ "step": 2173
+ },
+ {
+ "epoch": 30.195633187772927,
+ "grad_norm": 0.016609614714980125,
+ "learning_rate": 0.0006,
+ "loss": 4.222233772277832,
+ "step": 2174
+ },
+ {
+ "epoch": 30.209606986899562,
+ "grad_norm": 0.016531318426132202,
+ "learning_rate": 0.0006,
+ "loss": 4.144489288330078,
+ "step": 2175
+ },
+ {
+ "epoch": 30.2235807860262,
+ "grad_norm": 0.01567668654024601,
+ "learning_rate": 0.0006,
+ "loss": 4.181916236877441,
+ "step": 2176
+ },
+ {
+ "epoch": 30.23755458515284,
+ "grad_norm": 0.017195681110024452,
+ "learning_rate": 0.0006,
+ "loss": 4.189693927764893,
+ "step": 2177
+ },
+ {
+ "epoch": 30.251528384279474,
+ "grad_norm": 0.016882948577404022,
+ "learning_rate": 0.0006,
+ "loss": 4.21604061126709,
+ "step": 2178
+ },
+ {
+ "epoch": 30.265502183406113,
+ "grad_norm": 0.016261622309684753,
+ "learning_rate": 0.0006,
+ "loss": 4.227639198303223,
+ "step": 2179
+ },
+ {
+ "epoch": 30.27947598253275,
+ "grad_norm": 0.016411812976002693,
+ "learning_rate": 0.0006,
+ "loss": 4.130362510681152,
+ "step": 2180
+ },
+ {
+ "epoch": 30.29344978165939,
+ "grad_norm": 0.01732019893825054,
+ "learning_rate": 0.0006,
+ "loss": 4.230443000793457,
+ "step": 2181
+ },
+ {
+ "epoch": 30.307423580786025,
+ "grad_norm": 0.01825905591249466,
+ "learning_rate": 0.0006,
+ "loss": 4.175817489624023,
+ "step": 2182
+ },
+ {
+ "epoch": 30.321397379912664,
+ "grad_norm": 0.01887713558971882,
+ "learning_rate": 0.0006,
+ "loss": 4.169020652770996,
+ "step": 2183
+ },
+ {
+ "epoch": 30.335371179039303,
+ "grad_norm": 0.019391661509871483,
+ "learning_rate": 0.0006,
+ "loss": 4.176318645477295,
+ "step": 2184
+ },
+ {
+ "epoch": 30.349344978165938,
+ "grad_norm": 0.020840002223849297,
+ "learning_rate": 0.0006,
+ "loss": 4.174228668212891,
+ "step": 2185
+ },
+ {
+ "epoch": 30.363318777292577,
+ "grad_norm": 0.020193178206682205,
+ "learning_rate": 0.0006,
+ "loss": 4.127919673919678,
+ "step": 2186
+ },
+ {
+ "epoch": 30.377292576419215,
+ "grad_norm": 0.018449561670422554,
+ "learning_rate": 0.0006,
+ "loss": 4.229487895965576,
+ "step": 2187
+ },
+ {
+ "epoch": 30.39126637554585,
+ "grad_norm": 0.01711983233690262,
+ "learning_rate": 0.0006,
+ "loss": 4.201651573181152,
+ "step": 2188
+ },
+ {
+ "epoch": 30.40524017467249,
+ "grad_norm": 0.017293395474553108,
+ "learning_rate": 0.0006,
+ "loss": 4.168704032897949,
+ "step": 2189
+ },
+ {
+ "epoch": 30.419213973799128,
+ "grad_norm": 0.018526358529925346,
+ "learning_rate": 0.0006,
+ "loss": 4.171864032745361,
+ "step": 2190
+ },
+ {
+ "epoch": 30.433187772925763,
+ "grad_norm": 0.019287291914224625,
+ "learning_rate": 0.0006,
+ "loss": 4.039534568786621,
+ "step": 2191
+ },
+ {
+ "epoch": 30.4471615720524,
+ "grad_norm": 0.016902444884181023,
+ "learning_rate": 0.0006,
+ "loss": 4.282787799835205,
+ "step": 2192
+ },
+ {
+ "epoch": 30.46113537117904,
+ "grad_norm": 0.01656750962138176,
+ "learning_rate": 0.0006,
+ "loss": 4.213098526000977,
+ "step": 2193
+ },
+ {
+ "epoch": 30.475109170305675,
+ "grad_norm": 0.016805065795779228,
+ "learning_rate": 0.0006,
+ "loss": 4.1583476066589355,
+ "step": 2194
+ },
+ {
+ "epoch": 30.489082969432314,
+ "grad_norm": 0.017166638746857643,
+ "learning_rate": 0.0006,
+ "loss": 4.188634395599365,
+ "step": 2195
+ },
+ {
+ "epoch": 30.503056768558952,
+ "grad_norm": 0.01592106744647026,
+ "learning_rate": 0.0006,
+ "loss": 4.234043121337891,
+ "step": 2196
+ },
+ {
+ "epoch": 30.51703056768559,
+ "grad_norm": 0.015270989388227463,
+ "learning_rate": 0.0006,
+ "loss": 4.301420211791992,
+ "step": 2197
+ },
+ {
+ "epoch": 30.531004366812226,
+ "grad_norm": 0.015194724313914776,
+ "learning_rate": 0.0006,
+ "loss": 4.18703031539917,
+ "step": 2198
+ },
+ {
+ "epoch": 30.544978165938865,
+ "grad_norm": 0.015587746165692806,
+ "learning_rate": 0.0006,
+ "loss": 4.138928413391113,
+ "step": 2199
+ },
+ {
+ "epoch": 30.558951965065503,
+ "grad_norm": 0.016029570251703262,
+ "learning_rate": 0.0006,
+ "loss": 4.290863990783691,
+ "step": 2200
+ },
+ {
+ "epoch": 30.57292576419214,
+ "grad_norm": 0.015651309862732887,
+ "learning_rate": 0.0006,
+ "loss": 4.272062301635742,
+ "step": 2201
+ },
+ {
+ "epoch": 30.586899563318777,
+ "grad_norm": 0.015992306172847748,
+ "learning_rate": 0.0006,
+ "loss": 4.2487406730651855,
+ "step": 2202
+ },
+ {
+ "epoch": 30.600873362445416,
+ "grad_norm": 0.014969157986342907,
+ "learning_rate": 0.0006,
+ "loss": 4.115347862243652,
+ "step": 2203
+ },
+ {
+ "epoch": 30.61484716157205,
+ "grad_norm": 0.014639027416706085,
+ "learning_rate": 0.0006,
+ "loss": 4.2207489013671875,
+ "step": 2204
+ },
+ {
+ "epoch": 30.62882096069869,
+ "grad_norm": 0.015059413388371468,
+ "learning_rate": 0.0006,
+ "loss": 4.274044990539551,
+ "step": 2205
+ },
+ {
+ "epoch": 30.64279475982533,
+ "grad_norm": 0.0166641678661108,
+ "learning_rate": 0.0006,
+ "loss": 4.191373348236084,
+ "step": 2206
+ },
+ {
+ "epoch": 30.656768558951963,
+ "grad_norm": 0.017806345596909523,
+ "learning_rate": 0.0006,
+ "loss": 4.139800548553467,
+ "step": 2207
+ },
+ {
+ "epoch": 30.670742358078602,
+ "grad_norm": 0.018712421879172325,
+ "learning_rate": 0.0006,
+ "loss": 4.035965442657471,
+ "step": 2208
+ },
+ {
+ "epoch": 30.68471615720524,
+ "grad_norm": 0.018457243219017982,
+ "learning_rate": 0.0006,
+ "loss": 4.2579240798950195,
+ "step": 2209
+ },
+ {
+ "epoch": 30.69868995633188,
+ "grad_norm": 0.01920999586582184,
+ "learning_rate": 0.0006,
+ "loss": 4.131112098693848,
+ "step": 2210
+ },
+ {
+ "epoch": 30.712663755458514,
+ "grad_norm": 0.01729944534599781,
+ "learning_rate": 0.0006,
+ "loss": 4.254059314727783,
+ "step": 2211
+ },
+ {
+ "epoch": 30.726637554585153,
+ "grad_norm": 0.016914231702685356,
+ "learning_rate": 0.0006,
+ "loss": 4.233428955078125,
+ "step": 2212
+ },
+ {
+ "epoch": 30.74061135371179,
+ "grad_norm": 0.01913098618388176,
+ "learning_rate": 0.0006,
+ "loss": 4.23525333404541,
+ "step": 2213
+ },
+ {
+ "epoch": 30.754585152838427,
+ "grad_norm": 0.019150281324982643,
+ "learning_rate": 0.0006,
+ "loss": 4.213695049285889,
+ "step": 2214
+ },
+ {
+ "epoch": 30.768558951965066,
+ "grad_norm": 0.018947452306747437,
+ "learning_rate": 0.0006,
+ "loss": 4.151697158813477,
+ "step": 2215
+ },
+ {
+ "epoch": 30.782532751091704,
+ "grad_norm": 0.016058508306741714,
+ "learning_rate": 0.0006,
+ "loss": 4.220280170440674,
+ "step": 2216
+ },
+ {
+ "epoch": 30.79650655021834,
+ "grad_norm": 0.017488619312644005,
+ "learning_rate": 0.0006,
+ "loss": 4.154428482055664,
+ "step": 2217
+ },
+ {
+ "epoch": 30.810480349344978,
+ "grad_norm": 0.017355090007185936,
+ "learning_rate": 0.0006,
+ "loss": 4.256086349487305,
+ "step": 2218
+ },
+ {
+ "epoch": 30.824454148471617,
+ "grad_norm": 0.016636032611131668,
+ "learning_rate": 0.0006,
+ "loss": 4.227564334869385,
+ "step": 2219
+ },
+ {
+ "epoch": 30.83842794759825,
+ "grad_norm": 0.016273891553282738,
+ "learning_rate": 0.0006,
+ "loss": 4.200466156005859,
+ "step": 2220
+ },
+ {
+ "epoch": 30.85240174672489,
+ "grad_norm": 0.016939911991357803,
+ "learning_rate": 0.0006,
+ "loss": 4.1571221351623535,
+ "step": 2221
+ },
+ {
+ "epoch": 30.86637554585153,
+ "grad_norm": 0.018332522362470627,
+ "learning_rate": 0.0006,
+ "loss": 4.228398323059082,
+ "step": 2222
+ },
+ {
+ "epoch": 30.880349344978168,
+ "grad_norm": 0.019508114084601402,
+ "learning_rate": 0.0006,
+ "loss": 4.222960472106934,
+ "step": 2223
+ },
+ {
+ "epoch": 30.894323144104803,
+ "grad_norm": 0.018440278246998787,
+ "learning_rate": 0.0006,
+ "loss": 4.261051177978516,
+ "step": 2224
+ },
+ {
+ "epoch": 30.90829694323144,
+ "grad_norm": 0.018113229423761368,
+ "learning_rate": 0.0006,
+ "loss": 4.109235763549805,
+ "step": 2225
+ },
+ {
+ "epoch": 30.92227074235808,
+ "grad_norm": 0.015392083674669266,
+ "learning_rate": 0.0006,
+ "loss": 4.21950101852417,
+ "step": 2226
+ },
+ {
+ "epoch": 30.936244541484715,
+ "grad_norm": 0.01633290946483612,
+ "learning_rate": 0.0006,
+ "loss": 4.176053047180176,
+ "step": 2227
+ },
+ {
+ "epoch": 30.950218340611354,
+ "grad_norm": 0.016236383467912674,
+ "learning_rate": 0.0006,
+ "loss": 4.314309120178223,
+ "step": 2228
+ },
+ {
+ "epoch": 30.964192139737992,
+ "grad_norm": 0.015284373424947262,
+ "learning_rate": 0.0006,
+ "loss": 4.303133487701416,
+ "step": 2229
+ },
+ {
+ "epoch": 30.978165938864628,
+ "grad_norm": 0.015982897952198982,
+ "learning_rate": 0.0006,
+ "loss": 4.25739860534668,
+ "step": 2230
+ },
+ {
+ "epoch": 30.992139737991266,
+ "grad_norm": 0.016682198271155357,
+ "learning_rate": 0.0006,
+ "loss": 4.292210578918457,
+ "step": 2231
+ },
+ {
+ "epoch": 31.0,
+ "grad_norm": 0.017521467059850693,
+ "learning_rate": 0.0006,
+ "loss": 4.170637130737305,
+ "step": 2232
+ },
+ {
+ "epoch": 31.0,
+ "eval_loss": 4.642786979675293,
+ "eval_runtime": 57.029,
+ "eval_samples_per_second": 42.82,
+ "eval_steps_per_second": 1.35,
+ "step": 2232
+ },
+ {
+ "epoch": 31.01397379912664,
+ "grad_norm": 0.016607852652668953,
+ "learning_rate": 0.0006,
+ "loss": 4.106224536895752,
+ "step": 2233
+ },
+ {
+ "epoch": 31.027947598253274,
+ "grad_norm": 0.018448904156684875,
+ "learning_rate": 0.0006,
+ "loss": 4.224459648132324,
+ "step": 2234
+ },
+ {
+ "epoch": 31.041921397379912,
+ "grad_norm": 0.020481228828430176,
+ "learning_rate": 0.0006,
+ "loss": 4.305548667907715,
+ "step": 2235
+ },
+ {
+ "epoch": 31.05589519650655,
+ "grad_norm": 0.01999707892537117,
+ "learning_rate": 0.0006,
+ "loss": 4.253663539886475,
+ "step": 2236
+ },
+ {
+ "epoch": 31.069868995633186,
+ "grad_norm": 0.020428303629159927,
+ "learning_rate": 0.0006,
+ "loss": 4.166954040527344,
+ "step": 2237
+ },
+ {
+ "epoch": 31.083842794759825,
+ "grad_norm": 0.022608213126659393,
+ "learning_rate": 0.0006,
+ "loss": 4.172819137573242,
+ "step": 2238
+ },
+ {
+ "epoch": 31.097816593886463,
+ "grad_norm": 0.024669520556926727,
+ "learning_rate": 0.0006,
+ "loss": 4.106586456298828,
+ "step": 2239
+ },
+ {
+ "epoch": 31.111790393013102,
+ "grad_norm": 0.02247351035475731,
+ "learning_rate": 0.0006,
+ "loss": 4.160271644592285,
+ "step": 2240
+ },
+ {
+ "epoch": 31.125764192139737,
+ "grad_norm": 0.0207810141146183,
+ "learning_rate": 0.0006,
+ "loss": 4.272491931915283,
+ "step": 2241
+ },
+ {
+ "epoch": 31.139737991266376,
+ "grad_norm": 0.021576499566435814,
+ "learning_rate": 0.0006,
+ "loss": 4.171474456787109,
+ "step": 2242
+ },
+ {
+ "epoch": 31.153711790393015,
+ "grad_norm": 0.02337106689810753,
+ "learning_rate": 0.0006,
+ "loss": 4.094712734222412,
+ "step": 2243
+ },
+ {
+ "epoch": 31.16768558951965,
+ "grad_norm": 0.02219875156879425,
+ "learning_rate": 0.0006,
+ "loss": 4.1800994873046875,
+ "step": 2244
+ },
+ {
+ "epoch": 31.18165938864629,
+ "grad_norm": 0.022680504247546196,
+ "learning_rate": 0.0006,
+ "loss": 4.2054362297058105,
+ "step": 2245
+ },
+ {
+ "epoch": 31.195633187772927,
+ "grad_norm": 0.0208734143525362,
+ "learning_rate": 0.0006,
+ "loss": 4.06068229675293,
+ "step": 2246
+ },
+ {
+ "epoch": 31.209606986899562,
+ "grad_norm": 0.01811066083610058,
+ "learning_rate": 0.0006,
+ "loss": 4.2465128898620605,
+ "step": 2247
+ },
+ {
+ "epoch": 31.2235807860262,
+ "grad_norm": 0.018571637570858,
+ "learning_rate": 0.0006,
+ "loss": 4.191955089569092,
+ "step": 2248
+ },
+ {
+ "epoch": 31.23755458515284,
+ "grad_norm": 0.018598807975649834,
+ "learning_rate": 0.0006,
+ "loss": 4.200872421264648,
+ "step": 2249
+ },
+ {
+ "epoch": 31.251528384279474,
+ "grad_norm": 0.017577625811100006,
+ "learning_rate": 0.0006,
+ "loss": 4.148700714111328,
+ "step": 2250
+ },
+ {
+ "epoch": 31.265502183406113,
+ "grad_norm": 0.01817980408668518,
+ "learning_rate": 0.0006,
+ "loss": 4.256978988647461,
+ "step": 2251
+ },
+ {
+ "epoch": 31.27947598253275,
+ "grad_norm": 0.017983486875891685,
+ "learning_rate": 0.0006,
+ "loss": 4.175005912780762,
+ "step": 2252
+ },
+ {
+ "epoch": 31.29344978165939,
+ "grad_norm": 0.016751961782574654,
+ "learning_rate": 0.0006,
+ "loss": 4.237518310546875,
+ "step": 2253
+ },
+ {
+ "epoch": 31.307423580786025,
+ "grad_norm": 0.016407202929258347,
+ "learning_rate": 0.0006,
+ "loss": 4.197269439697266,
+ "step": 2254
+ },
+ {
+ "epoch": 31.321397379912664,
+ "grad_norm": 0.01578645594418049,
+ "learning_rate": 0.0006,
+ "loss": 4.179751396179199,
+ "step": 2255
+ },
+ {
+ "epoch": 31.335371179039303,
+ "grad_norm": 0.016001226380467415,
+ "learning_rate": 0.0006,
+ "loss": 4.08821964263916,
+ "step": 2256
+ },
+ {
+ "epoch": 31.349344978165938,
+ "grad_norm": 0.015122811309993267,
+ "learning_rate": 0.0006,
+ "loss": 4.047828674316406,
+ "step": 2257
+ },
+ {
+ "epoch": 31.363318777292577,
+ "grad_norm": 0.015399965457618237,
+ "learning_rate": 0.0006,
+ "loss": 4.1506524085998535,
+ "step": 2258
+ },
+ {
+ "epoch": 31.377292576419215,
+ "grad_norm": 0.01534605398774147,
+ "learning_rate": 0.0006,
+ "loss": 4.27952766418457,
+ "step": 2259
+ },
+ {
+ "epoch": 31.39126637554585,
+ "grad_norm": 0.015461381524801254,
+ "learning_rate": 0.0006,
+ "loss": 4.27102518081665,
+ "step": 2260
+ },
+ {
+ "epoch": 31.40524017467249,
+ "grad_norm": 0.016769153997302055,
+ "learning_rate": 0.0006,
+ "loss": 4.191592216491699,
+ "step": 2261
+ },
+ {
+ "epoch": 31.419213973799128,
+ "grad_norm": 0.016557959839701653,
+ "learning_rate": 0.0006,
+ "loss": 4.159705638885498,
+ "step": 2262
+ },
+ {
+ "epoch": 31.433187772925763,
+ "grad_norm": 0.01659569889307022,
+ "learning_rate": 0.0006,
+ "loss": 4.2136664390563965,
+ "step": 2263
+ },
+ {
+ "epoch": 31.4471615720524,
+ "grad_norm": 0.017101937904953957,
+ "learning_rate": 0.0006,
+ "loss": 4.2716498374938965,
+ "step": 2264
+ },
+ {
+ "epoch": 31.46113537117904,
+ "grad_norm": 0.01722985878586769,
+ "learning_rate": 0.0006,
+ "loss": 4.300534248352051,
+ "step": 2265
+ },
+ {
+ "epoch": 31.475109170305675,
+ "grad_norm": 0.016800185665488243,
+ "learning_rate": 0.0006,
+ "loss": 4.113240718841553,
+ "step": 2266
+ },
+ {
+ "epoch": 31.489082969432314,
+ "grad_norm": 0.01603875495493412,
+ "learning_rate": 0.0006,
+ "loss": 4.157105445861816,
+ "step": 2267
+ },
+ {
+ "epoch": 31.503056768558952,
+ "grad_norm": 0.015903932973742485,
+ "learning_rate": 0.0006,
+ "loss": 4.233669757843018,
+ "step": 2268
+ },
+ {
+ "epoch": 31.51703056768559,
+ "grad_norm": 0.0160983856767416,
+ "learning_rate": 0.0006,
+ "loss": 4.147704124450684,
+ "step": 2269
+ },
+ {
+ "epoch": 31.531004366812226,
+ "grad_norm": 0.01604730263352394,
+ "learning_rate": 0.0006,
+ "loss": 4.1465911865234375,
+ "step": 2270
+ },
+ {
+ "epoch": 31.544978165938865,
+ "grad_norm": 0.016205785796046257,
+ "learning_rate": 0.0006,
+ "loss": 4.106704235076904,
+ "step": 2271
+ },
+ {
+ "epoch": 31.558951965065503,
+ "grad_norm": 0.017008036375045776,
+ "learning_rate": 0.0006,
+ "loss": 4.191000938415527,
+ "step": 2272
+ },
+ {
+ "epoch": 31.57292576419214,
+ "grad_norm": 0.01699635200202465,
+ "learning_rate": 0.0006,
+ "loss": 4.187042236328125,
+ "step": 2273
+ },
+ {
+ "epoch": 31.586899563318777,
+ "grad_norm": 0.01692967116832733,
+ "learning_rate": 0.0006,
+ "loss": 4.208805561065674,
+ "step": 2274
+ },
+ {
+ "epoch": 31.600873362445416,
+ "grad_norm": 0.017608756199479103,
+ "learning_rate": 0.0006,
+ "loss": 4.200512886047363,
+ "step": 2275
+ },
+ {
+ "epoch": 31.61484716157205,
+ "grad_norm": 0.016590990126132965,
+ "learning_rate": 0.0006,
+ "loss": 4.194705009460449,
+ "step": 2276
+ },
+ {
+ "epoch": 31.62882096069869,
+ "grad_norm": 0.0170235987752676,
+ "learning_rate": 0.0006,
+ "loss": 4.210326194763184,
+ "step": 2277
+ },
+ {
+ "epoch": 31.64279475982533,
+ "grad_norm": 0.01604801043868065,
+ "learning_rate": 0.0006,
+ "loss": 4.196024417877197,
+ "step": 2278
+ },
+ {
+ "epoch": 31.656768558951963,
+ "grad_norm": 0.015694202855229378,
+ "learning_rate": 0.0006,
+ "loss": 4.178236484527588,
+ "step": 2279
+ },
+ {
+ "epoch": 31.670742358078602,
+ "grad_norm": 0.015473959036171436,
+ "learning_rate": 0.0006,
+ "loss": 4.259339809417725,
+ "step": 2280
+ },
+ {
+ "epoch": 31.68471615720524,
+ "grad_norm": 0.014689362607896328,
+ "learning_rate": 0.0006,
+ "loss": 3.9702324867248535,
+ "step": 2281
+ },
+ {
+ "epoch": 31.69868995633188,
+ "grad_norm": 0.015245123766362667,
+ "learning_rate": 0.0006,
+ "loss": 4.197025775909424,
+ "step": 2282
+ },
+ {
+ "epoch": 31.712663755458514,
+ "grad_norm": 0.016148347407579422,
+ "learning_rate": 0.0006,
+ "loss": 4.231856346130371,
+ "step": 2283
+ },
+ {
+ "epoch": 31.726637554585153,
+ "grad_norm": 0.015169818885624409,
+ "learning_rate": 0.0006,
+ "loss": 4.230169296264648,
+ "step": 2284
+ },
+ {
+ "epoch": 31.74061135371179,
+ "grad_norm": 0.014843013137578964,
+ "learning_rate": 0.0006,
+ "loss": 4.195633411407471,
+ "step": 2285
+ },
+ {
+ "epoch": 31.754585152838427,
+ "grad_norm": 0.015229340642690659,
+ "learning_rate": 0.0006,
+ "loss": 4.231280326843262,
+ "step": 2286
+ },
+ {
+ "epoch": 31.768558951965066,
+ "grad_norm": 0.01551708485931158,
+ "learning_rate": 0.0006,
+ "loss": 4.17064094543457,
+ "step": 2287
+ },
+ {
+ "epoch": 31.782532751091704,
+ "grad_norm": 0.015863575041294098,
+ "learning_rate": 0.0006,
+ "loss": 4.1579484939575195,
+ "step": 2288
+ },
+ {
+ "epoch": 31.79650655021834,
+ "grad_norm": 0.015118198469281197,
+ "learning_rate": 0.0006,
+ "loss": 4.1626787185668945,
+ "step": 2289
+ },
+ {
+ "epoch": 31.810480349344978,
+ "grad_norm": 0.015220776200294495,
+ "learning_rate": 0.0006,
+ "loss": 4.135746955871582,
+ "step": 2290
+ },
+ {
+ "epoch": 31.824454148471617,
+ "grad_norm": 0.014529784210026264,
+ "learning_rate": 0.0006,
+ "loss": 4.164931297302246,
+ "step": 2291
+ },
+ {
+ "epoch": 31.83842794759825,
+ "grad_norm": 0.015533296391367912,
+ "learning_rate": 0.0006,
+ "loss": 4.112464904785156,
+ "step": 2292
+ },
+ {
+ "epoch": 31.85240174672489,
+ "grad_norm": 0.016562335193157196,
+ "learning_rate": 0.0006,
+ "loss": 4.164898872375488,
+ "step": 2293
+ },
+ {
+ "epoch": 31.86637554585153,
+ "grad_norm": 0.016258113086223602,
+ "learning_rate": 0.0006,
+ "loss": 4.1979265213012695,
+ "step": 2294
+ },
+ {
+ "epoch": 31.880349344978168,
+ "grad_norm": 0.01558110024780035,
+ "learning_rate": 0.0006,
+ "loss": 4.174675941467285,
+ "step": 2295
+ },
+ {
+ "epoch": 31.894323144104803,
+ "grad_norm": 0.016412515193223953,
+ "learning_rate": 0.0006,
+ "loss": 4.217909336090088,
+ "step": 2296
+ },
+ {
+ "epoch": 31.90829694323144,
+ "grad_norm": 0.015118442475795746,
+ "learning_rate": 0.0006,
+ "loss": 4.137143611907959,
+ "step": 2297
+ },
+ {
+ "epoch": 31.92227074235808,
+ "grad_norm": 0.01506161317229271,
+ "learning_rate": 0.0006,
+ "loss": 4.196122169494629,
+ "step": 2298
+ },
+ {
+ "epoch": 31.936244541484715,
+ "grad_norm": 0.017165271565318108,
+ "learning_rate": 0.0006,
+ "loss": 4.14834451675415,
+ "step": 2299
+ },
+ {
+ "epoch": 31.950218340611354,
+ "grad_norm": 0.01696440577507019,
+ "learning_rate": 0.0006,
+ "loss": 4.218868732452393,
+ "step": 2300
+ },
+ {
+ "epoch": 31.964192139737992,
+ "grad_norm": 0.017349394038319588,
+ "learning_rate": 0.0006,
+ "loss": 4.2754011154174805,
+ "step": 2301
+ },
+ {
+ "epoch": 31.978165938864628,
+ "grad_norm": 0.018387921154499054,
+ "learning_rate": 0.0006,
+ "loss": 4.282623291015625,
+ "step": 2302
+ },
+ {
+ "epoch": 31.992139737991266,
+ "grad_norm": 0.018837926909327507,
+ "learning_rate": 0.0006,
+ "loss": 4.185425281524658,
+ "step": 2303
+ },
+ {
+ "epoch": 32.0,
+ "grad_norm": 0.018998464569449425,
+ "learning_rate": 0.0006,
+ "loss": 4.151291847229004,
+ "step": 2304
+ },
+ {
+ "epoch": 32.0,
+ "eval_loss": 4.522614002227783,
+ "eval_runtime": 57.0106,
+ "eval_samples_per_second": 42.834,
+ "eval_steps_per_second": 1.351,
+ "step": 2304
+ },
+ {
+ "epoch": 32.01397379912664,
+ "grad_norm": 0.019036833196878433,
+ "learning_rate": 0.0006,
+ "loss": 4.183202266693115,
+ "step": 2305
+ },
+ {
+ "epoch": 32.02794759825328,
+ "grad_norm": 0.02207397110760212,
+ "learning_rate": 0.0006,
+ "loss": 4.17875862121582,
+ "step": 2306
+ },
+ {
+ "epoch": 32.041921397379916,
+ "grad_norm": 0.022008279338479042,
+ "learning_rate": 0.0006,
+ "loss": 4.08042049407959,
+ "step": 2307
+ },
+ {
+ "epoch": 32.05589519650655,
+ "grad_norm": 0.022407017648220062,
+ "learning_rate": 0.0006,
+ "loss": 4.112189769744873,
+ "step": 2308
+ },
+ {
+ "epoch": 32.069868995633186,
+ "grad_norm": 0.021599747240543365,
+ "learning_rate": 0.0006,
+ "loss": 4.126455307006836,
+ "step": 2309
+ },
+ {
+ "epoch": 32.083842794759825,
+ "grad_norm": 0.020340966060757637,
+ "learning_rate": 0.0006,
+ "loss": 4.080538272857666,
+ "step": 2310
+ },
+ {
+ "epoch": 32.09781659388646,
+ "grad_norm": 0.020956892520189285,
+ "learning_rate": 0.0006,
+ "loss": 4.171943664550781,
+ "step": 2311
+ },
+ {
+ "epoch": 32.1117903930131,
+ "grad_norm": 0.019966667518019676,
+ "learning_rate": 0.0006,
+ "loss": 4.144007682800293,
+ "step": 2312
+ },
+ {
+ "epoch": 32.12576419213974,
+ "grad_norm": 0.018256952986121178,
+ "learning_rate": 0.0006,
+ "loss": 4.208107948303223,
+ "step": 2313
+ },
+ {
+ "epoch": 32.13973799126637,
+ "grad_norm": 0.017289170995354652,
+ "learning_rate": 0.0006,
+ "loss": 4.182024955749512,
+ "step": 2314
+ },
+ {
+ "epoch": 32.15371179039301,
+ "grad_norm": 0.017656700685620308,
+ "learning_rate": 0.0006,
+ "loss": 4.068373680114746,
+ "step": 2315
+ },
+ {
+ "epoch": 32.16768558951965,
+ "grad_norm": 0.017903361469507217,
+ "learning_rate": 0.0006,
+ "loss": 4.2345428466796875,
+ "step": 2316
+ },
+ {
+ "epoch": 32.18165938864629,
+ "grad_norm": 0.01785109005868435,
+ "learning_rate": 0.0006,
+ "loss": 4.219776153564453,
+ "step": 2317
+ },
+ {
+ "epoch": 32.19563318777293,
+ "grad_norm": 0.01887257769703865,
+ "learning_rate": 0.0006,
+ "loss": 4.139206409454346,
+ "step": 2318
+ },
+ {
+ "epoch": 32.209606986899566,
+ "grad_norm": 0.01924685761332512,
+ "learning_rate": 0.0006,
+ "loss": 4.121052265167236,
+ "step": 2319
+ },
+ {
+ "epoch": 32.223580786026204,
+ "grad_norm": 0.020574072375893593,
+ "learning_rate": 0.0006,
+ "loss": 4.13262414932251,
+ "step": 2320
+ },
+ {
+ "epoch": 32.237554585152836,
+ "grad_norm": 0.023849064484238625,
+ "learning_rate": 0.0006,
+ "loss": 4.239119529724121,
+ "step": 2321
+ },
+ {
+ "epoch": 32.251528384279474,
+ "grad_norm": 0.026239361613988876,
+ "learning_rate": 0.0006,
+ "loss": 4.248993873596191,
+ "step": 2322
+ },
+ {
+ "epoch": 32.26550218340611,
+ "grad_norm": 0.023980243131518364,
+ "learning_rate": 0.0006,
+ "loss": 4.159852504730225,
+ "step": 2323
+ },
+ {
+ "epoch": 32.27947598253275,
+ "grad_norm": 0.020442405715584755,
+ "learning_rate": 0.0006,
+ "loss": 4.133546829223633,
+ "step": 2324
+ },
+ {
+ "epoch": 32.29344978165939,
+ "grad_norm": 0.020661186426877975,
+ "learning_rate": 0.0006,
+ "loss": 4.078243732452393,
+ "step": 2325
+ },
+ {
+ "epoch": 32.30742358078603,
+ "grad_norm": 0.02046387456357479,
+ "learning_rate": 0.0006,
+ "loss": 4.172012805938721,
+ "step": 2326
+ },
+ {
+ "epoch": 32.32139737991266,
+ "grad_norm": 0.02089597098529339,
+ "learning_rate": 0.0006,
+ "loss": 4.0604705810546875,
+ "step": 2327
+ },
+ {
+ "epoch": 32.3353711790393,
+ "grad_norm": 0.01875591278076172,
+ "learning_rate": 0.0006,
+ "loss": 4.125911235809326,
+ "step": 2328
+ },
+ {
+ "epoch": 32.34934497816594,
+ "grad_norm": 0.018606796860694885,
+ "learning_rate": 0.0006,
+ "loss": 4.097362518310547,
+ "step": 2329
+ },
+ {
+ "epoch": 32.36331877729258,
+ "grad_norm": 0.01766626164317131,
+ "learning_rate": 0.0006,
+ "loss": 4.118616580963135,
+ "step": 2330
+ },
+ {
+ "epoch": 32.377292576419215,
+ "grad_norm": 0.017692960798740387,
+ "learning_rate": 0.0006,
+ "loss": 4.1457672119140625,
+ "step": 2331
+ },
+ {
+ "epoch": 32.391266375545854,
+ "grad_norm": 0.017619704827666283,
+ "learning_rate": 0.0006,
+ "loss": 4.175961494445801,
+ "step": 2332
+ },
+ {
+ "epoch": 32.40524017467249,
+ "grad_norm": 0.01784413494169712,
+ "learning_rate": 0.0006,
+ "loss": 4.051291465759277,
+ "step": 2333
+ },
+ {
+ "epoch": 32.419213973799124,
+ "grad_norm": 0.018508171662688255,
+ "learning_rate": 0.0006,
+ "loss": 4.0638556480407715,
+ "step": 2334
+ },
+ {
+ "epoch": 32.43318777292576,
+ "grad_norm": 0.01731436885893345,
+ "learning_rate": 0.0006,
+ "loss": 4.1805419921875,
+ "step": 2335
+ },
+ {
+ "epoch": 32.4471615720524,
+ "grad_norm": 0.015886498615145683,
+ "learning_rate": 0.0006,
+ "loss": 4.151026725769043,
+ "step": 2336
+ },
+ {
+ "epoch": 32.46113537117904,
+ "grad_norm": 0.016583820804953575,
+ "learning_rate": 0.0006,
+ "loss": 4.203604698181152,
+ "step": 2337
+ },
+ {
+ "epoch": 32.47510917030568,
+ "grad_norm": 0.0176097322255373,
+ "learning_rate": 0.0006,
+ "loss": 4.135159015655518,
+ "step": 2338
+ },
+ {
+ "epoch": 32.48908296943232,
+ "grad_norm": 0.016657903790473938,
+ "learning_rate": 0.0006,
+ "loss": 4.126681804656982,
+ "step": 2339
+ },
+ {
+ "epoch": 32.50305676855895,
+ "grad_norm": 0.016866130754351616,
+ "learning_rate": 0.0006,
+ "loss": 4.162341117858887,
+ "step": 2340
+ },
+ {
+ "epoch": 32.51703056768559,
+ "grad_norm": 0.016286680474877357,
+ "learning_rate": 0.0006,
+ "loss": 4.261801242828369,
+ "step": 2341
+ },
+ {
+ "epoch": 32.531004366812226,
+ "grad_norm": 0.01481365505605936,
+ "learning_rate": 0.0006,
+ "loss": 4.110016822814941,
+ "step": 2342
+ },
+ {
+ "epoch": 32.544978165938865,
+ "grad_norm": 0.016605013981461525,
+ "learning_rate": 0.0006,
+ "loss": 4.230024814605713,
+ "step": 2343
+ },
+ {
+ "epoch": 32.5589519650655,
+ "grad_norm": 0.01645175740122795,
+ "learning_rate": 0.0006,
+ "loss": 4.174612045288086,
+ "step": 2344
+ },
+ {
+ "epoch": 32.57292576419214,
+ "grad_norm": 0.014725026674568653,
+ "learning_rate": 0.0006,
+ "loss": 4.066371440887451,
+ "step": 2345
+ },
+ {
+ "epoch": 32.58689956331878,
+ "grad_norm": 0.015003837645053864,
+ "learning_rate": 0.0006,
+ "loss": 4.193343639373779,
+ "step": 2346
+ },
+ {
+ "epoch": 32.60087336244541,
+ "grad_norm": 0.015346302650868893,
+ "learning_rate": 0.0006,
+ "loss": 4.210659980773926,
+ "step": 2347
+ },
+ {
+ "epoch": 32.61484716157205,
+ "grad_norm": 0.015135962516069412,
+ "learning_rate": 0.0006,
+ "loss": 4.204618453979492,
+ "step": 2348
+ },
+ {
+ "epoch": 32.62882096069869,
+ "grad_norm": 0.01601956970989704,
+ "learning_rate": 0.0006,
+ "loss": 4.145564556121826,
+ "step": 2349
+ },
+ {
+ "epoch": 32.64279475982533,
+ "grad_norm": 0.01517851185053587,
+ "learning_rate": 0.0006,
+ "loss": 4.051673889160156,
+ "step": 2350
+ },
+ {
+ "epoch": 32.65676855895197,
+ "grad_norm": 0.014894292689859867,
+ "learning_rate": 0.0006,
+ "loss": 4.184818267822266,
+ "step": 2351
+ },
+ {
+ "epoch": 32.670742358078606,
+ "grad_norm": 0.015720834955573082,
+ "learning_rate": 0.0006,
+ "loss": 4.220036506652832,
+ "step": 2352
+ },
+ {
+ "epoch": 32.68471615720524,
+ "grad_norm": 0.01573433168232441,
+ "learning_rate": 0.0006,
+ "loss": 4.076773166656494,
+ "step": 2353
+ },
+ {
+ "epoch": 32.698689956331876,
+ "grad_norm": 0.015828052535653114,
+ "learning_rate": 0.0006,
+ "loss": 4.116329669952393,
+ "step": 2354
+ },
+ {
+ "epoch": 32.712663755458514,
+ "grad_norm": 0.015462259761989117,
+ "learning_rate": 0.0006,
+ "loss": 4.1155242919921875,
+ "step": 2355
+ },
+ {
+ "epoch": 32.72663755458515,
+ "grad_norm": 0.015763849020004272,
+ "learning_rate": 0.0006,
+ "loss": 4.169687747955322,
+ "step": 2356
+ },
+ {
+ "epoch": 32.74061135371179,
+ "grad_norm": 0.014935145154595375,
+ "learning_rate": 0.0006,
+ "loss": 4.2090911865234375,
+ "step": 2357
+ },
+ {
+ "epoch": 32.75458515283843,
+ "grad_norm": 0.015119184739887714,
+ "learning_rate": 0.0006,
+ "loss": 4.131838798522949,
+ "step": 2358
+ },
+ {
+ "epoch": 32.76855895196506,
+ "grad_norm": 0.014833525754511356,
+ "learning_rate": 0.0006,
+ "loss": 4.17618465423584,
+ "step": 2359
+ },
+ {
+ "epoch": 32.7825327510917,
+ "grad_norm": 0.01588359661400318,
+ "learning_rate": 0.0006,
+ "loss": 4.163723945617676,
+ "step": 2360
+ },
+ {
+ "epoch": 32.79650655021834,
+ "grad_norm": 0.0168188214302063,
+ "learning_rate": 0.0006,
+ "loss": 4.178189277648926,
+ "step": 2361
+ },
+ {
+ "epoch": 32.81048034934498,
+ "grad_norm": 0.01726095750927925,
+ "learning_rate": 0.0006,
+ "loss": 4.144223690032959,
+ "step": 2362
+ },
+ {
+ "epoch": 32.82445414847162,
+ "grad_norm": 0.01701410301029682,
+ "learning_rate": 0.0006,
+ "loss": 4.209896087646484,
+ "step": 2363
+ },
+ {
+ "epoch": 32.838427947598255,
+ "grad_norm": 0.01682264916598797,
+ "learning_rate": 0.0006,
+ "loss": 4.26223087310791,
+ "step": 2364
+ },
+ {
+ "epoch": 32.852401746724894,
+ "grad_norm": 0.015597456134855747,
+ "learning_rate": 0.0006,
+ "loss": 4.292664051055908,
+ "step": 2365
+ },
+ {
+ "epoch": 32.866375545851525,
+ "grad_norm": 0.015258579514920712,
+ "learning_rate": 0.0006,
+ "loss": 4.165006637573242,
+ "step": 2366
+ },
+ {
+ "epoch": 32.880349344978164,
+ "grad_norm": 0.01662200503051281,
+ "learning_rate": 0.0006,
+ "loss": 4.309786319732666,
+ "step": 2367
+ },
+ {
+ "epoch": 32.8943231441048,
+ "grad_norm": 0.015691161155700684,
+ "learning_rate": 0.0006,
+ "loss": 4.167202949523926,
+ "step": 2368
+ },
+ {
+ "epoch": 32.90829694323144,
+ "grad_norm": 0.014870637096464634,
+ "learning_rate": 0.0006,
+ "loss": 4.142415523529053,
+ "step": 2369
+ },
+ {
+ "epoch": 32.92227074235808,
+ "grad_norm": 0.01567252166569233,
+ "learning_rate": 0.0006,
+ "loss": 4.2312092781066895,
+ "step": 2370
+ },
+ {
+ "epoch": 32.93624454148472,
+ "grad_norm": 0.014948616735637188,
+ "learning_rate": 0.0006,
+ "loss": 4.1514129638671875,
+ "step": 2371
+ },
+ {
+ "epoch": 32.95021834061135,
+ "grad_norm": 0.014654500409960747,
+ "learning_rate": 0.0006,
+ "loss": 4.205010890960693,
+ "step": 2372
+ },
+ {
+ "epoch": 32.96419213973799,
+ "grad_norm": 0.015059148892760277,
+ "learning_rate": 0.0006,
+ "loss": 4.148405075073242,
+ "step": 2373
+ },
+ {
+ "epoch": 32.97816593886463,
+ "grad_norm": 0.015026576817035675,
+ "learning_rate": 0.0006,
+ "loss": 4.182535171508789,
+ "step": 2374
+ },
+ {
+ "epoch": 32.992139737991266,
+ "grad_norm": 0.016394320875406265,
+ "learning_rate": 0.0006,
+ "loss": 4.129051208496094,
+ "step": 2375
+ },
+ {
+ "epoch": 33.0,
+ "grad_norm": 0.01867043413221836,
+ "learning_rate": 0.0006,
+ "loss": 4.185556411743164,
+ "step": 2376
+ },
+ {
+ "epoch": 33.0,
+ "eval_loss": 4.640405654907227,
+ "eval_runtime": 56.5861,
+ "eval_samples_per_second": 43.155,
+ "eval_steps_per_second": 1.361,
+ "step": 2376
+ },
+ {
+ "epoch": 33.01397379912664,
+ "grad_norm": 0.018626727163791656,
+ "learning_rate": 0.0006,
+ "loss": 4.143031597137451,
+ "step": 2377
+ },
+ {
+ "epoch": 33.02794759825328,
+ "grad_norm": 0.019283363595604897,
+ "learning_rate": 0.0006,
+ "loss": 4.117845058441162,
+ "step": 2378
+ },
+ {
+ "epoch": 33.041921397379916,
+ "grad_norm": 0.018185822293162346,
+ "learning_rate": 0.0006,
+ "loss": 4.076862812042236,
+ "step": 2379
+ },
+ {
+ "epoch": 33.05589519650655,
+ "grad_norm": 0.019411196932196617,
+ "learning_rate": 0.0006,
+ "loss": 4.124156475067139,
+ "step": 2380
+ },
+ {
+ "epoch": 33.069868995633186,
+ "grad_norm": 0.019248466938734055,
+ "learning_rate": 0.0006,
+ "loss": 4.201735019683838,
+ "step": 2381
+ },
+ {
+ "epoch": 33.083842794759825,
+ "grad_norm": 0.018856720998883247,
+ "learning_rate": 0.0006,
+ "loss": 4.00296688079834,
+ "step": 2382
+ },
+ {
+ "epoch": 33.09781659388646,
+ "grad_norm": 0.018554294481873512,
+ "learning_rate": 0.0006,
+ "loss": 4.270577430725098,
+ "step": 2383
+ },
+ {
+ "epoch": 33.1117903930131,
+ "grad_norm": 0.019946224987506866,
+ "learning_rate": 0.0006,
+ "loss": 4.119589328765869,
+ "step": 2384
+ },
+ {
+ "epoch": 33.12576419213974,
+ "grad_norm": 0.021395670250058174,
+ "learning_rate": 0.0006,
+ "loss": 4.128847122192383,
+ "step": 2385
+ },
+ {
+ "epoch": 33.13973799126637,
+ "grad_norm": 0.02009674906730652,
+ "learning_rate": 0.0006,
+ "loss": 4.183228969573975,
+ "step": 2386
+ },
+ {
+ "epoch": 33.15371179039301,
+ "grad_norm": 0.01820223033428192,
+ "learning_rate": 0.0006,
+ "loss": 4.131601810455322,
+ "step": 2387
+ },
+ {
+ "epoch": 33.16768558951965,
+ "grad_norm": 0.018653156235814095,
+ "learning_rate": 0.0006,
+ "loss": 4.116430282592773,
+ "step": 2388
+ },
+ {
+ "epoch": 33.18165938864629,
+ "grad_norm": 0.016828063875436783,
+ "learning_rate": 0.0006,
+ "loss": 4.054663181304932,
+ "step": 2389
+ },
+ {
+ "epoch": 33.19563318777293,
+ "grad_norm": 0.017025409266352654,
+ "learning_rate": 0.0006,
+ "loss": 4.110044956207275,
+ "step": 2390
+ },
+ {
+ "epoch": 33.209606986899566,
+ "grad_norm": 0.016608452424407005,
+ "learning_rate": 0.0006,
+ "loss": 4.102668285369873,
+ "step": 2391
+ },
+ {
+ "epoch": 33.223580786026204,
+ "grad_norm": 0.01743577979505062,
+ "learning_rate": 0.0006,
+ "loss": 4.166791915893555,
+ "step": 2392
+ },
+ {
+ "epoch": 33.237554585152836,
+ "grad_norm": 0.017253734171390533,
+ "learning_rate": 0.0006,
+ "loss": 4.217832088470459,
+ "step": 2393
+ },
+ {
+ "epoch": 33.251528384279474,
+ "grad_norm": 0.017937595024704933,
+ "learning_rate": 0.0006,
+ "loss": 4.2173967361450195,
+ "step": 2394
+ },
+ {
+ "epoch": 33.26550218340611,
+ "grad_norm": 0.01985993981361389,
+ "learning_rate": 0.0006,
+ "loss": 4.17849588394165,
+ "step": 2395
+ },
+ {
+ "epoch": 33.27947598253275,
+ "grad_norm": 0.019243355840444565,
+ "learning_rate": 0.0006,
+ "loss": 4.086523056030273,
+ "step": 2396
+ },
+ {
+ "epoch": 33.29344978165939,
+ "grad_norm": 0.01924416422843933,
+ "learning_rate": 0.0006,
+ "loss": 4.128574371337891,
+ "step": 2397
+ },
+ {
+ "epoch": 33.30742358078603,
+ "grad_norm": 0.01983758620917797,
+ "learning_rate": 0.0006,
+ "loss": 4.061581611633301,
+ "step": 2398
+ },
+ {
+ "epoch": 33.32139737991266,
+ "grad_norm": 0.01930462196469307,
+ "learning_rate": 0.0006,
+ "loss": 4.091731071472168,
+ "step": 2399
+ },
+ {
+ "epoch": 33.3353711790393,
+ "grad_norm": 0.017988424748182297,
+ "learning_rate": 0.0006,
+ "loss": 4.155599117279053,
+ "step": 2400
+ },
+ {
+ "epoch": 33.34934497816594,
+ "grad_norm": 0.015194767154753208,
+ "learning_rate": 0.0006,
+ "loss": 4.144134044647217,
+ "step": 2401
+ },
+ {
+ "epoch": 33.36331877729258,
+ "grad_norm": 0.01536081824451685,
+ "learning_rate": 0.0006,
+ "loss": 4.127272129058838,
+ "step": 2402
+ },
+ {
+ "epoch": 33.377292576419215,
+ "grad_norm": 0.015431297942996025,
+ "learning_rate": 0.0006,
+ "loss": 4.3213019371032715,
+ "step": 2403
+ },
+ {
+ "epoch": 33.391266375545854,
+ "grad_norm": 0.015740280970931053,
+ "learning_rate": 0.0006,
+ "loss": 4.121467590332031,
+ "step": 2404
+ },
+ {
+ "epoch": 33.40524017467249,
+ "grad_norm": 0.015379384160041809,
+ "learning_rate": 0.0006,
+ "loss": 4.095216751098633,
+ "step": 2405
+ },
+ {
+ "epoch": 33.419213973799124,
+ "grad_norm": 0.01528073474764824,
+ "learning_rate": 0.0006,
+ "loss": 4.13047981262207,
+ "step": 2406
+ },
+ {
+ "epoch": 33.43318777292576,
+ "grad_norm": 0.015739979222416878,
+ "learning_rate": 0.0006,
+ "loss": 4.2533135414123535,
+ "step": 2407
+ },
+ {
+ "epoch": 33.4471615720524,
+ "grad_norm": 0.01519764307886362,
+ "learning_rate": 0.0006,
+ "loss": 4.131782054901123,
+ "step": 2408
+ },
+ {
+ "epoch": 33.46113537117904,
+ "grad_norm": 0.015425996854901314,
+ "learning_rate": 0.0006,
+ "loss": 4.02811336517334,
+ "step": 2409
+ },
+ {
+ "epoch": 33.47510917030568,
+ "grad_norm": 0.015378830023109913,
+ "learning_rate": 0.0006,
+ "loss": 4.08652925491333,
+ "step": 2410
+ },
+ {
+ "epoch": 33.48908296943232,
+ "grad_norm": 0.01473075058311224,
+ "learning_rate": 0.0006,
+ "loss": 4.214364528656006,
+ "step": 2411
+ },
+ {
+ "epoch": 33.50305676855895,
+ "grad_norm": 0.015141637064516544,
+ "learning_rate": 0.0006,
+ "loss": 4.155712127685547,
+ "step": 2412
+ },
+ {
+ "epoch": 33.51703056768559,
+ "grad_norm": 0.01641593500971794,
+ "learning_rate": 0.0006,
+ "loss": 4.097907066345215,
+ "step": 2413
+ },
+ {
+ "epoch": 33.531004366812226,
+ "grad_norm": 0.015182029455900192,
+ "learning_rate": 0.0006,
+ "loss": 4.238499641418457,
+ "step": 2414
+ },
+ {
+ "epoch": 33.544978165938865,
+ "grad_norm": 0.01643732376396656,
+ "learning_rate": 0.0006,
+ "loss": 4.0661492347717285,
+ "step": 2415
+ },
+ {
+ "epoch": 33.5589519650655,
+ "grad_norm": 0.015614228323101997,
+ "learning_rate": 0.0006,
+ "loss": 4.120795249938965,
+ "step": 2416
+ },
+ {
+ "epoch": 33.57292576419214,
+ "grad_norm": 0.016066187992691994,
+ "learning_rate": 0.0006,
+ "loss": 4.096465110778809,
+ "step": 2417
+ },
+ {
+ "epoch": 33.58689956331878,
+ "grad_norm": 0.017214342951774597,
+ "learning_rate": 0.0006,
+ "loss": 4.242628574371338,
+ "step": 2418
+ },
+ {
+ "epoch": 33.60087336244541,
+ "grad_norm": 0.01787586510181427,
+ "learning_rate": 0.0006,
+ "loss": 4.149589538574219,
+ "step": 2419
+ },
+ {
+ "epoch": 33.61484716157205,
+ "grad_norm": 0.019455241039395332,
+ "learning_rate": 0.0006,
+ "loss": 4.267796039581299,
+ "step": 2420
+ },
+ {
+ "epoch": 33.62882096069869,
+ "grad_norm": 0.01984674669802189,
+ "learning_rate": 0.0006,
+ "loss": 4.175220966339111,
+ "step": 2421
+ },
+ {
+ "epoch": 33.64279475982533,
+ "grad_norm": 0.021933717653155327,
+ "learning_rate": 0.0006,
+ "loss": 4.106647491455078,
+ "step": 2422
+ },
+ {
+ "epoch": 33.65676855895197,
+ "grad_norm": 0.02123589813709259,
+ "learning_rate": 0.0006,
+ "loss": 4.204996109008789,
+ "step": 2423
+ },
+ {
+ "epoch": 33.670742358078606,
+ "grad_norm": 0.019814589992165565,
+ "learning_rate": 0.0006,
+ "loss": 4.186171531677246,
+ "step": 2424
+ },
+ {
+ "epoch": 33.68471615720524,
+ "grad_norm": 0.02059728465974331,
+ "learning_rate": 0.0006,
+ "loss": 4.170509338378906,
+ "step": 2425
+ },
+ {
+ "epoch": 33.698689956331876,
+ "grad_norm": 0.018483608961105347,
+ "learning_rate": 0.0006,
+ "loss": 4.085585594177246,
+ "step": 2426
+ },
+ {
+ "epoch": 33.712663755458514,
+ "grad_norm": 0.018403852358460426,
+ "learning_rate": 0.0006,
+ "loss": 4.2551774978637695,
+ "step": 2427
+ },
+ {
+ "epoch": 33.72663755458515,
+ "grad_norm": 0.018653830513358116,
+ "learning_rate": 0.0006,
+ "loss": 4.091766357421875,
+ "step": 2428
+ },
+ {
+ "epoch": 33.74061135371179,
+ "grad_norm": 0.017692334949970245,
+ "learning_rate": 0.0006,
+ "loss": 4.100838661193848,
+ "step": 2429
+ },
+ {
+ "epoch": 33.75458515283843,
+ "grad_norm": 0.016029126942157745,
+ "learning_rate": 0.0006,
+ "loss": 4.1530680656433105,
+ "step": 2430
+ },
+ {
+ "epoch": 33.76855895196506,
+ "grad_norm": 0.016095684841275215,
+ "learning_rate": 0.0006,
+ "loss": 4.20544958114624,
+ "step": 2431
+ },
+ {
+ "epoch": 33.7825327510917,
+ "grad_norm": 0.016375798732042313,
+ "learning_rate": 0.0006,
+ "loss": 4.199177265167236,
+ "step": 2432
+ },
+ {
+ "epoch": 33.79650655021834,
+ "grad_norm": 0.01590786688029766,
+ "learning_rate": 0.0006,
+ "loss": 4.200799465179443,
+ "step": 2433
+ },
+ {
+ "epoch": 33.81048034934498,
+ "grad_norm": 0.014896097593009472,
+ "learning_rate": 0.0006,
+ "loss": 4.185850143432617,
+ "step": 2434
+ },
+ {
+ "epoch": 33.82445414847162,
+ "grad_norm": 0.01438513956964016,
+ "learning_rate": 0.0006,
+ "loss": 4.214962005615234,
+ "step": 2435
+ },
+ {
+ "epoch": 33.838427947598255,
+ "grad_norm": 0.014459339901804924,
+ "learning_rate": 0.0006,
+ "loss": 4.196972370147705,
+ "step": 2436
+ },
+ {
+ "epoch": 33.852401746724894,
+ "grad_norm": 0.014782169833779335,
+ "learning_rate": 0.0006,
+ "loss": 4.0885910987854,
+ "step": 2437
+ },
+ {
+ "epoch": 33.866375545851525,
+ "grad_norm": 0.01520370040088892,
+ "learning_rate": 0.0006,
+ "loss": 4.0323262214660645,
+ "step": 2438
+ },
+ {
+ "epoch": 33.880349344978164,
+ "grad_norm": 0.014056282117962837,
+ "learning_rate": 0.0006,
+ "loss": 4.1386518478393555,
+ "step": 2439
+ },
+ {
+ "epoch": 33.8943231441048,
+ "grad_norm": 0.013567990623414516,
+ "learning_rate": 0.0006,
+ "loss": 4.095754623413086,
+ "step": 2440
+ },
+ {
+ "epoch": 33.90829694323144,
+ "grad_norm": 0.013306716457009315,
+ "learning_rate": 0.0006,
+ "loss": 4.1623077392578125,
+ "step": 2441
+ },
+ {
+ "epoch": 33.92227074235808,
+ "grad_norm": 0.014047150500118732,
+ "learning_rate": 0.0006,
+ "loss": 4.142541885375977,
+ "step": 2442
+ },
+ {
+ "epoch": 33.93624454148472,
+ "grad_norm": 0.01377617847174406,
+ "learning_rate": 0.0006,
+ "loss": 4.268551826477051,
+ "step": 2443
+ },
+ {
+ "epoch": 33.95021834061135,
+ "grad_norm": 0.014671621844172478,
+ "learning_rate": 0.0006,
+ "loss": 4.198196887969971,
+ "step": 2444
+ },
+ {
+ "epoch": 33.96419213973799,
+ "grad_norm": 0.014806744642555714,
+ "learning_rate": 0.0006,
+ "loss": 4.057899475097656,
+ "step": 2445
+ },
+ {
+ "epoch": 33.97816593886463,
+ "grad_norm": 0.015734272077679634,
+ "learning_rate": 0.0006,
+ "loss": 4.163505554199219,
+ "step": 2446
+ },
+ {
+ "epoch": 33.992139737991266,
+ "grad_norm": 0.015250146389007568,
+ "learning_rate": 0.0006,
+ "loss": 4.076294898986816,
+ "step": 2447
+ },
+ {
+ "epoch": 34.0,
+ "grad_norm": 0.016733702272176743,
+ "learning_rate": 0.0006,
+ "loss": 4.138420104980469,
+ "step": 2448
+ },
+ {
+ "epoch": 34.0,
+ "eval_loss": 4.542503356933594,
+ "eval_runtime": 56.983,
+ "eval_samples_per_second": 42.855,
+ "eval_steps_per_second": 1.351,
+ "step": 2448
+ },
+ {
+ "epoch": 34.01397379912664,
+ "grad_norm": 0.015866613015532494,
+ "learning_rate": 0.0006,
+ "loss": 4.032081127166748,
+ "step": 2449
+ },
+ {
+ "epoch": 34.02794759825328,
+ "grad_norm": 0.01946570910513401,
+ "learning_rate": 0.0006,
+ "loss": 4.122320175170898,
+ "step": 2450
+ },
+ {
+ "epoch": 34.041921397379916,
+ "grad_norm": 0.020265938714146614,
+ "learning_rate": 0.0006,
+ "loss": 4.06460428237915,
+ "step": 2451
+ },
+ {
+ "epoch": 34.05589519650655,
+ "grad_norm": 0.018887171521782875,
+ "learning_rate": 0.0006,
+ "loss": 4.096906661987305,
+ "step": 2452
+ },
+ {
+ "epoch": 34.069868995633186,
+ "grad_norm": 0.017304692417383194,
+ "learning_rate": 0.0006,
+ "loss": 4.137261390686035,
+ "step": 2453
+ },
+ {
+ "epoch": 34.083842794759825,
+ "grad_norm": 0.017029082402586937,
+ "learning_rate": 0.0006,
+ "loss": 4.081112861633301,
+ "step": 2454
+ },
+ {
+ "epoch": 34.09781659388646,
+ "grad_norm": 0.016414813697338104,
+ "learning_rate": 0.0006,
+ "loss": 4.140783309936523,
+ "step": 2455
+ },
+ {
+ "epoch": 34.1117903930131,
+ "grad_norm": 0.018238281831145287,
+ "learning_rate": 0.0006,
+ "loss": 4.161116600036621,
+ "step": 2456
+ },
+ {
+ "epoch": 34.12576419213974,
+ "grad_norm": 0.019929232075810432,
+ "learning_rate": 0.0006,
+ "loss": 4.243722438812256,
+ "step": 2457
+ },
+ {
+ "epoch": 34.13973799126637,
+ "grad_norm": 0.019375786185264587,
+ "learning_rate": 0.0006,
+ "loss": 4.134873390197754,
+ "step": 2458
+ },
+ {
+ "epoch": 34.15371179039301,
+ "grad_norm": 0.018151137977838516,
+ "learning_rate": 0.0006,
+ "loss": 4.098745346069336,
+ "step": 2459
+ },
+ {
+ "epoch": 34.16768558951965,
+ "grad_norm": 0.017054934054613113,
+ "learning_rate": 0.0006,
+ "loss": 4.194662570953369,
+ "step": 2460
+ },
+ {
+ "epoch": 34.18165938864629,
+ "grad_norm": 0.018220555037260056,
+ "learning_rate": 0.0006,
+ "loss": 4.128866195678711,
+ "step": 2461
+ },
+ {
+ "epoch": 34.19563318777293,
+ "grad_norm": 0.01692879945039749,
+ "learning_rate": 0.0006,
+ "loss": 4.062477111816406,
+ "step": 2462
+ },
+ {
+ "epoch": 34.209606986899566,
+ "grad_norm": 0.01674199476838112,
+ "learning_rate": 0.0006,
+ "loss": 4.186149597167969,
+ "step": 2463
+ },
+ {
+ "epoch": 34.223580786026204,
+ "grad_norm": 0.016527455300092697,
+ "learning_rate": 0.0006,
+ "loss": 4.0734028816223145,
+ "step": 2464
+ },
+ {
+ "epoch": 34.237554585152836,
+ "grad_norm": 0.016448386013507843,
+ "learning_rate": 0.0006,
+ "loss": 4.098896026611328,
+ "step": 2465
+ },
+ {
+ "epoch": 34.251528384279474,
+ "grad_norm": 0.014644953422248363,
+ "learning_rate": 0.0006,
+ "loss": 4.1031904220581055,
+ "step": 2466
+ },
+ {
+ "epoch": 34.26550218340611,
+ "grad_norm": 0.016649341210722923,
+ "learning_rate": 0.0006,
+ "loss": 4.148791313171387,
+ "step": 2467
+ },
+ {
+ "epoch": 34.27947598253275,
+ "grad_norm": 0.017442714422941208,
+ "learning_rate": 0.0006,
+ "loss": 4.070799827575684,
+ "step": 2468
+ },
+ {
+ "epoch": 34.29344978165939,
+ "grad_norm": 0.017852013930678368,
+ "learning_rate": 0.0006,
+ "loss": 4.031924247741699,
+ "step": 2469
+ },
+ {
+ "epoch": 34.30742358078603,
+ "grad_norm": 0.016371270641684532,
+ "learning_rate": 0.0006,
+ "loss": 4.149420738220215,
+ "step": 2470
+ },
+ {
+ "epoch": 34.32139737991266,
+ "grad_norm": 0.015846814960241318,
+ "learning_rate": 0.0006,
+ "loss": 4.15427303314209,
+ "step": 2471
+ },
+ {
+ "epoch": 34.3353711790393,
+ "grad_norm": 0.017411163076758385,
+ "learning_rate": 0.0006,
+ "loss": 4.1304192543029785,
+ "step": 2472
+ },
+ {
+ "epoch": 34.34934497816594,
+ "grad_norm": 0.01780046708881855,
+ "learning_rate": 0.0006,
+ "loss": 4.191732406616211,
+ "step": 2473
+ },
+ {
+ "epoch": 34.36331877729258,
+ "grad_norm": 0.01958383060991764,
+ "learning_rate": 0.0006,
+ "loss": 4.068355083465576,
+ "step": 2474
+ },
+ {
+ "epoch": 34.377292576419215,
+ "grad_norm": 0.022173317149281502,
+ "learning_rate": 0.0006,
+ "loss": 4.182929039001465,
+ "step": 2475
+ },
+ {
+ "epoch": 34.391266375545854,
+ "grad_norm": 0.022389156743884087,
+ "learning_rate": 0.0006,
+ "loss": 4.110218048095703,
+ "step": 2476
+ },
+ {
+ "epoch": 34.40524017467249,
+ "grad_norm": 0.022478511556982994,
+ "learning_rate": 0.0006,
+ "loss": 4.095479488372803,
+ "step": 2477
+ },
+ {
+ "epoch": 34.419213973799124,
+ "grad_norm": 0.019504647701978683,
+ "learning_rate": 0.0006,
+ "loss": 4.101964950561523,
+ "step": 2478
+ },
+ {
+ "epoch": 34.43318777292576,
+ "grad_norm": 0.019378576427698135,
+ "learning_rate": 0.0006,
+ "loss": 4.14620304107666,
+ "step": 2479
+ },
+ {
+ "epoch": 34.4471615720524,
+ "grad_norm": 0.017993204295635223,
+ "learning_rate": 0.0006,
+ "loss": 4.183492660522461,
+ "step": 2480
+ },
+ {
+ "epoch": 34.46113537117904,
+ "grad_norm": 0.01814805157482624,
+ "learning_rate": 0.0006,
+ "loss": 4.151247978210449,
+ "step": 2481
+ },
+ {
+ "epoch": 34.47510917030568,
+ "grad_norm": 0.018622448667883873,
+ "learning_rate": 0.0006,
+ "loss": 4.020938396453857,
+ "step": 2482
+ },
+ {
+ "epoch": 34.48908296943232,
+ "grad_norm": 0.018360132351517677,
+ "learning_rate": 0.0006,
+ "loss": 4.064734935760498,
+ "step": 2483
+ },
+ {
+ "epoch": 34.50305676855895,
+ "grad_norm": 0.017715711146593094,
+ "learning_rate": 0.0006,
+ "loss": 4.185687065124512,
+ "step": 2484
+ },
+ {
+ "epoch": 34.51703056768559,
+ "grad_norm": 0.01714295893907547,
+ "learning_rate": 0.0006,
+ "loss": 4.1044721603393555,
+ "step": 2485
+ },
+ {
+ "epoch": 34.531004366812226,
+ "grad_norm": 0.016943803057074547,
+ "learning_rate": 0.0006,
+ "loss": 4.106837272644043,
+ "step": 2486
+ },
+ {
+ "epoch": 34.544978165938865,
+ "grad_norm": 0.0182423684746027,
+ "learning_rate": 0.0006,
+ "loss": 4.1041340827941895,
+ "step": 2487
+ },
+ {
+ "epoch": 34.5589519650655,
+ "grad_norm": 0.01636587455868721,
+ "learning_rate": 0.0006,
+ "loss": 3.9735546112060547,
+ "step": 2488
+ },
+ {
+ "epoch": 34.57292576419214,
+ "grad_norm": 0.014593546278774738,
+ "learning_rate": 0.0006,
+ "loss": 4.041980266571045,
+ "step": 2489
+ },
+ {
+ "epoch": 34.58689956331878,
+ "grad_norm": 0.015436965972185135,
+ "learning_rate": 0.0006,
+ "loss": 4.301148414611816,
+ "step": 2490
+ },
+ {
+ "epoch": 34.60087336244541,
+ "grad_norm": 0.015075269155204296,
+ "learning_rate": 0.0006,
+ "loss": 4.118527412414551,
+ "step": 2491
+ },
+ {
+ "epoch": 34.61484716157205,
+ "grad_norm": 0.01701514609158039,
+ "learning_rate": 0.0006,
+ "loss": 4.245968818664551,
+ "step": 2492
+ },
+ {
+ "epoch": 34.62882096069869,
+ "grad_norm": 0.018047887831926346,
+ "learning_rate": 0.0006,
+ "loss": 4.227261066436768,
+ "step": 2493
+ },
+ {
+ "epoch": 34.64279475982533,
+ "grad_norm": 0.018499085679650307,
+ "learning_rate": 0.0006,
+ "loss": 4.121360778808594,
+ "step": 2494
+ },
+ {
+ "epoch": 34.65676855895197,
+ "grad_norm": 0.018987711519002914,
+ "learning_rate": 0.0006,
+ "loss": 4.131564617156982,
+ "step": 2495
+ },
+ {
+ "epoch": 34.670742358078606,
+ "grad_norm": 0.01834476925432682,
+ "learning_rate": 0.0006,
+ "loss": 4.224941253662109,
+ "step": 2496
+ },
+ {
+ "epoch": 34.68471615720524,
+ "grad_norm": 0.017507942393422127,
+ "learning_rate": 0.0006,
+ "loss": 4.1394548416137695,
+ "step": 2497
+ },
+ {
+ "epoch": 34.698689956331876,
+ "grad_norm": 0.01680580899119377,
+ "learning_rate": 0.0006,
+ "loss": 4.1073431968688965,
+ "step": 2498
+ },
+ {
+ "epoch": 34.712663755458514,
+ "grad_norm": 0.015233514830470085,
+ "learning_rate": 0.0006,
+ "loss": 4.136994361877441,
+ "step": 2499
+ },
+ {
+ "epoch": 34.72663755458515,
+ "grad_norm": 0.015482917428016663,
+ "learning_rate": 0.0006,
+ "loss": 4.110719680786133,
+ "step": 2500
+ },
+ {
+ "epoch": 34.74061135371179,
+ "grad_norm": 0.015940461307764053,
+ "learning_rate": 0.0006,
+ "loss": 4.207851886749268,
+ "step": 2501
+ },
+ {
+ "epoch": 34.75458515283843,
+ "grad_norm": 0.015355597250163555,
+ "learning_rate": 0.0006,
+ "loss": 4.169469833374023,
+ "step": 2502
+ },
+ {
+ "epoch": 34.76855895196506,
+ "grad_norm": 0.015974000096321106,
+ "learning_rate": 0.0006,
+ "loss": 4.110061168670654,
+ "step": 2503
+ },
+ {
+ "epoch": 34.7825327510917,
+ "grad_norm": 0.016088927164673805,
+ "learning_rate": 0.0006,
+ "loss": 4.155452728271484,
+ "step": 2504
+ },
+ {
+ "epoch": 34.79650655021834,
+ "grad_norm": 0.016156120225787163,
+ "learning_rate": 0.0006,
+ "loss": 4.178092956542969,
+ "step": 2505
+ },
+ {
+ "epoch": 34.81048034934498,
+ "grad_norm": 0.014928647316992283,
+ "learning_rate": 0.0006,
+ "loss": 4.05457878112793,
+ "step": 2506
+ },
+ {
+ "epoch": 34.82445414847162,
+ "grad_norm": 0.015289328061044216,
+ "learning_rate": 0.0006,
+ "loss": 4.18869686126709,
+ "step": 2507
+ },
+ {
+ "epoch": 34.838427947598255,
+ "grad_norm": 0.015799039974808693,
+ "learning_rate": 0.0006,
+ "loss": 4.106073379516602,
+ "step": 2508
+ },
+ {
+ "epoch": 34.852401746724894,
+ "grad_norm": 0.015824418514966965,
+ "learning_rate": 0.0006,
+ "loss": 4.0634589195251465,
+ "step": 2509
+ },
+ {
+ "epoch": 34.866375545851525,
+ "grad_norm": 0.015321656130254269,
+ "learning_rate": 0.0006,
+ "loss": 4.093090057373047,
+ "step": 2510
+ },
+ {
+ "epoch": 34.880349344978164,
+ "grad_norm": 0.014380166307091713,
+ "learning_rate": 0.0006,
+ "loss": 4.1750640869140625,
+ "step": 2511
+ },
+ {
+ "epoch": 34.8943231441048,
+ "grad_norm": 0.014962567947804928,
+ "learning_rate": 0.0006,
+ "loss": 4.185277938842773,
+ "step": 2512
+ },
+ {
+ "epoch": 34.90829694323144,
+ "grad_norm": 0.014467543922364712,
+ "learning_rate": 0.0006,
+ "loss": 4.1600799560546875,
+ "step": 2513
+ },
+ {
+ "epoch": 34.92227074235808,
+ "grad_norm": 0.01545181218534708,
+ "learning_rate": 0.0006,
+ "loss": 4.096299648284912,
+ "step": 2514
+ },
+ {
+ "epoch": 34.93624454148472,
+ "grad_norm": 0.01605340465903282,
+ "learning_rate": 0.0006,
+ "loss": 4.203323841094971,
+ "step": 2515
+ },
+ {
+ "epoch": 34.95021834061135,
+ "grad_norm": 0.015203815884888172,
+ "learning_rate": 0.0006,
+ "loss": 4.201385021209717,
+ "step": 2516
+ },
+ {
+ "epoch": 34.96419213973799,
+ "grad_norm": 0.015149409882724285,
+ "learning_rate": 0.0006,
+ "loss": 4.088164329528809,
+ "step": 2517
+ },
+ {
+ "epoch": 34.97816593886463,
+ "grad_norm": 0.014885558746755123,
+ "learning_rate": 0.0006,
+ "loss": 4.073911666870117,
+ "step": 2518
+ },
+ {
+ "epoch": 34.992139737991266,
+ "grad_norm": 0.015003419481217861,
+ "learning_rate": 0.0006,
+ "loss": 4.157113552093506,
+ "step": 2519
+ },
+ {
+ "epoch": 35.0,
+ "grad_norm": 0.01641916297376156,
+ "learning_rate": 0.0006,
+ "loss": 3.970052719116211,
+ "step": 2520
+ },
+ {
+ "epoch": 35.0,
+ "eval_loss": 4.641692161560059,
+ "eval_runtime": 56.8836,
+ "eval_samples_per_second": 42.93,
+ "eval_steps_per_second": 1.354,
+ "step": 2520
+ },
+ {
+ "epoch": 35.01397379912664,
+ "grad_norm": 0.015894673764705658,
+ "learning_rate": 0.0006,
+ "loss": 4.104437351226807,
+ "step": 2521
+ },
+ {
+ "epoch": 35.02794759825328,
+ "grad_norm": 0.0173867866396904,
+ "learning_rate": 0.0006,
+ "loss": 4.091726303100586,
+ "step": 2522
+ },
+ {
+ "epoch": 35.041921397379916,
+ "grad_norm": 0.019272832199931145,
+ "learning_rate": 0.0006,
+ "loss": 4.217905044555664,
+ "step": 2523
+ },
+ {
+ "epoch": 35.05589519650655,
+ "grad_norm": 0.021820371970534325,
+ "learning_rate": 0.0006,
+ "loss": 4.111428260803223,
+ "step": 2524
+ },
+ {
+ "epoch": 35.069868995633186,
+ "grad_norm": 0.02340741828083992,
+ "learning_rate": 0.0006,
+ "loss": 4.115771293640137,
+ "step": 2525
+ },
+ {
+ "epoch": 35.083842794759825,
+ "grad_norm": 0.023760691285133362,
+ "learning_rate": 0.0006,
+ "loss": 4.132383346557617,
+ "step": 2526
+ },
+ {
+ "epoch": 35.09781659388646,
+ "grad_norm": 0.02073468267917633,
+ "learning_rate": 0.0006,
+ "loss": 4.272691249847412,
+ "step": 2527
+ },
+ {
+ "epoch": 35.1117903930131,
+ "grad_norm": 0.02197718247771263,
+ "learning_rate": 0.0006,
+ "loss": 4.105586528778076,
+ "step": 2528
+ },
+ {
+ "epoch": 35.12576419213974,
+ "grad_norm": 0.0200254675000906,
+ "learning_rate": 0.0006,
+ "loss": 3.968820095062256,
+ "step": 2529
+ },
+ {
+ "epoch": 35.13973799126637,
+ "grad_norm": 0.01929587684571743,
+ "learning_rate": 0.0006,
+ "loss": 4.045468807220459,
+ "step": 2530
+ },
+ {
+ "epoch": 35.15371179039301,
+ "grad_norm": 0.01886162906885147,
+ "learning_rate": 0.0006,
+ "loss": 4.185571670532227,
+ "step": 2531
+ },
+ {
+ "epoch": 35.16768558951965,
+ "grad_norm": 0.019789839163422585,
+ "learning_rate": 0.0006,
+ "loss": 3.9734551906585693,
+ "step": 2532
+ },
+ {
+ "epoch": 35.18165938864629,
+ "grad_norm": 0.0198142658919096,
+ "learning_rate": 0.0006,
+ "loss": 4.2662506103515625,
+ "step": 2533
+ },
+ {
+ "epoch": 35.19563318777293,
+ "grad_norm": 0.02087012305855751,
+ "learning_rate": 0.0006,
+ "loss": 4.175931930541992,
+ "step": 2534
+ },
+ {
+ "epoch": 35.209606986899566,
+ "grad_norm": 0.020294535905122757,
+ "learning_rate": 0.0006,
+ "loss": 4.060324668884277,
+ "step": 2535
+ },
+ {
+ "epoch": 35.223580786026204,
+ "grad_norm": 0.01908983662724495,
+ "learning_rate": 0.0006,
+ "loss": 4.217939853668213,
+ "step": 2536
+ },
+ {
+ "epoch": 35.237554585152836,
+ "grad_norm": 0.0178892333060503,
+ "learning_rate": 0.0006,
+ "loss": 4.190216541290283,
+ "step": 2537
+ },
+ {
+ "epoch": 35.251528384279474,
+ "grad_norm": 0.017892315983772278,
+ "learning_rate": 0.0006,
+ "loss": 4.067419052124023,
+ "step": 2538
+ },
+ {
+ "epoch": 35.26550218340611,
+ "grad_norm": 0.018556250259280205,
+ "learning_rate": 0.0006,
+ "loss": 4.105958461761475,
+ "step": 2539
+ },
+ {
+ "epoch": 35.27947598253275,
+ "grad_norm": 0.01944422349333763,
+ "learning_rate": 0.0006,
+ "loss": 4.155775547027588,
+ "step": 2540
+ },
+ {
+ "epoch": 35.29344978165939,
+ "grad_norm": 0.017333146184682846,
+ "learning_rate": 0.0006,
+ "loss": 4.244513511657715,
+ "step": 2541
+ },
+ {
+ "epoch": 35.30742358078603,
+ "grad_norm": 0.01639244332909584,
+ "learning_rate": 0.0006,
+ "loss": 4.167815685272217,
+ "step": 2542
+ },
+ {
+ "epoch": 35.32139737991266,
+ "grad_norm": 0.01610143668949604,
+ "learning_rate": 0.0006,
+ "loss": 4.121486663818359,
+ "step": 2543
+ },
+ {
+ "epoch": 35.3353711790393,
+ "grad_norm": 0.01656411588191986,
+ "learning_rate": 0.0006,
+ "loss": 4.1755499839782715,
+ "step": 2544
+ },
+ {
+ "epoch": 35.34934497816594,
+ "grad_norm": 0.015075297094881535,
+ "learning_rate": 0.0006,
+ "loss": 4.102992534637451,
+ "step": 2545
+ },
+ {
+ "epoch": 35.36331877729258,
+ "grad_norm": 0.01538170501589775,
+ "learning_rate": 0.0006,
+ "loss": 4.161603927612305,
+ "step": 2546
+ },
+ {
+ "epoch": 35.377292576419215,
+ "grad_norm": 0.01483312901109457,
+ "learning_rate": 0.0006,
+ "loss": 4.085355758666992,
+ "step": 2547
+ },
+ {
+ "epoch": 35.391266375545854,
+ "grad_norm": 0.014894738793373108,
+ "learning_rate": 0.0006,
+ "loss": 4.083585262298584,
+ "step": 2548
+ },
+ {
+ "epoch": 35.40524017467249,
+ "grad_norm": 0.01475780550390482,
+ "learning_rate": 0.0006,
+ "loss": 4.012645721435547,
+ "step": 2549
+ },
+ {
+ "epoch": 35.419213973799124,
+ "grad_norm": 0.014531653374433517,
+ "learning_rate": 0.0006,
+ "loss": 4.197113990783691,
+ "step": 2550
+ },
+ {
+ "epoch": 35.43318777292576,
+ "grad_norm": 0.015316493809223175,
+ "learning_rate": 0.0006,
+ "loss": 4.227970123291016,
+ "step": 2551
+ },
+ {
+ "epoch": 35.4471615720524,
+ "grad_norm": 0.014685571193695068,
+ "learning_rate": 0.0006,
+ "loss": 4.089415550231934,
+ "step": 2552
+ },
+ {
+ "epoch": 35.46113537117904,
+ "grad_norm": 0.015337035991251469,
+ "learning_rate": 0.0006,
+ "loss": 4.080583572387695,
+ "step": 2553
+ },
+ {
+ "epoch": 35.47510917030568,
+ "grad_norm": 0.016754208132624626,
+ "learning_rate": 0.0006,
+ "loss": 4.163935661315918,
+ "step": 2554
+ },
+ {
+ "epoch": 35.48908296943232,
+ "grad_norm": 0.017535798251628876,
+ "learning_rate": 0.0006,
+ "loss": 4.093027591705322,
+ "step": 2555
+ },
+ {
+ "epoch": 35.50305676855895,
+ "grad_norm": 0.017336975783109665,
+ "learning_rate": 0.0006,
+ "loss": 4.127961158752441,
+ "step": 2556
+ },
+ {
+ "epoch": 35.51703056768559,
+ "grad_norm": 0.016242152079939842,
+ "learning_rate": 0.0006,
+ "loss": 4.155593395233154,
+ "step": 2557
+ },
+ {
+ "epoch": 35.531004366812226,
+ "grad_norm": 0.015121322125196457,
+ "learning_rate": 0.0006,
+ "loss": 3.984004020690918,
+ "step": 2558
+ },
+ {
+ "epoch": 35.544978165938865,
+ "grad_norm": 0.015178884379565716,
+ "learning_rate": 0.0006,
+ "loss": 4.111564636230469,
+ "step": 2559
+ },
+ {
+ "epoch": 35.5589519650655,
+ "grad_norm": 0.014777131378650665,
+ "learning_rate": 0.0006,
+ "loss": 4.041402816772461,
+ "step": 2560
+ },
+ {
+ "epoch": 35.57292576419214,
+ "grad_norm": 0.015197676606476307,
+ "learning_rate": 0.0006,
+ "loss": 4.133706092834473,
+ "step": 2561
+ },
+ {
+ "epoch": 35.58689956331878,
+ "grad_norm": 0.015905972570180893,
+ "learning_rate": 0.0006,
+ "loss": 4.1069135665893555,
+ "step": 2562
+ },
+ {
+ "epoch": 35.60087336244541,
+ "grad_norm": 0.015942877158522606,
+ "learning_rate": 0.0006,
+ "loss": 4.0720624923706055,
+ "step": 2563
+ },
+ {
+ "epoch": 35.61484716157205,
+ "grad_norm": 0.016697583720088005,
+ "learning_rate": 0.0006,
+ "loss": 4.167448043823242,
+ "step": 2564
+ },
+ {
+ "epoch": 35.62882096069869,
+ "grad_norm": 0.015603577718138695,
+ "learning_rate": 0.0006,
+ "loss": 4.127327919006348,
+ "step": 2565
+ },
+ {
+ "epoch": 35.64279475982533,
+ "grad_norm": 0.017153460532426834,
+ "learning_rate": 0.0006,
+ "loss": 4.056543350219727,
+ "step": 2566
+ },
+ {
+ "epoch": 35.65676855895197,
+ "grad_norm": 0.016731958836317062,
+ "learning_rate": 0.0006,
+ "loss": 4.0681843757629395,
+ "step": 2567
+ },
+ {
+ "epoch": 35.670742358078606,
+ "grad_norm": 0.016167359426617622,
+ "learning_rate": 0.0006,
+ "loss": 4.174992084503174,
+ "step": 2568
+ },
+ {
+ "epoch": 35.68471615720524,
+ "grad_norm": 0.01615108922123909,
+ "learning_rate": 0.0006,
+ "loss": 4.1383161544799805,
+ "step": 2569
+ },
+ {
+ "epoch": 35.698689956331876,
+ "grad_norm": 0.015750816091895103,
+ "learning_rate": 0.0006,
+ "loss": 4.031595706939697,
+ "step": 2570
+ },
+ {
+ "epoch": 35.712663755458514,
+ "grad_norm": 0.015416189096868038,
+ "learning_rate": 0.0006,
+ "loss": 4.067498207092285,
+ "step": 2571
+ },
+ {
+ "epoch": 35.72663755458515,
+ "grad_norm": 0.01570792682468891,
+ "learning_rate": 0.0006,
+ "loss": 4.1338300704956055,
+ "step": 2572
+ },
+ {
+ "epoch": 35.74061135371179,
+ "grad_norm": 0.016367752104997635,
+ "learning_rate": 0.0006,
+ "loss": 4.152523040771484,
+ "step": 2573
+ },
+ {
+ "epoch": 35.75458515283843,
+ "grad_norm": 0.016349803656339645,
+ "learning_rate": 0.0006,
+ "loss": 4.101985454559326,
+ "step": 2574
+ },
+ {
+ "epoch": 35.76855895196506,
+ "grad_norm": 0.01663660630583763,
+ "learning_rate": 0.0006,
+ "loss": 4.190130710601807,
+ "step": 2575
+ },
+ {
+ "epoch": 35.7825327510917,
+ "grad_norm": 0.015599170699715614,
+ "learning_rate": 0.0006,
+ "loss": 4.0792436599731445,
+ "step": 2576
+ },
+ {
+ "epoch": 35.79650655021834,
+ "grad_norm": 0.015472427010536194,
+ "learning_rate": 0.0006,
+ "loss": 4.186446666717529,
+ "step": 2577
+ },
+ {
+ "epoch": 35.81048034934498,
+ "grad_norm": 0.015520439483225346,
+ "learning_rate": 0.0006,
+ "loss": 4.246635437011719,
+ "step": 2578
+ },
+ {
+ "epoch": 35.82445414847162,
+ "grad_norm": 0.016998281702399254,
+ "learning_rate": 0.0006,
+ "loss": 4.13435173034668,
+ "step": 2579
+ },
+ {
+ "epoch": 35.838427947598255,
+ "grad_norm": 0.016290361061692238,
+ "learning_rate": 0.0006,
+ "loss": 4.107428073883057,
+ "step": 2580
+ },
+ {
+ "epoch": 35.852401746724894,
+ "grad_norm": 0.0165516659617424,
+ "learning_rate": 0.0006,
+ "loss": 4.1612138748168945,
+ "step": 2581
+ },
+ {
+ "epoch": 35.866375545851525,
+ "grad_norm": 0.01736091636121273,
+ "learning_rate": 0.0006,
+ "loss": 4.1072869300842285,
+ "step": 2582
+ },
+ {
+ "epoch": 35.880349344978164,
+ "grad_norm": 0.01787031814455986,
+ "learning_rate": 0.0006,
+ "loss": 4.1076812744140625,
+ "step": 2583
+ },
+ {
+ "epoch": 35.8943231441048,
+ "grad_norm": 0.015437985770404339,
+ "learning_rate": 0.0006,
+ "loss": 4.097488880157471,
+ "step": 2584
+ },
+ {
+ "epoch": 35.90829694323144,
+ "grad_norm": 0.016306741163134575,
+ "learning_rate": 0.0006,
+ "loss": 4.105017185211182,
+ "step": 2585
+ },
+ {
+ "epoch": 35.92227074235808,
+ "grad_norm": 0.015723245218396187,
+ "learning_rate": 0.0006,
+ "loss": 4.205283164978027,
+ "step": 2586
+ },
+ {
+ "epoch": 35.93624454148472,
+ "grad_norm": 0.014626456424593925,
+ "learning_rate": 0.0006,
+ "loss": 4.086552619934082,
+ "step": 2587
+ },
+ {
+ "epoch": 35.95021834061135,
+ "grad_norm": 0.014348454773426056,
+ "learning_rate": 0.0006,
+ "loss": 4.171329975128174,
+ "step": 2588
+ },
+ {
+ "epoch": 35.96419213973799,
+ "grad_norm": 0.015598480589687824,
+ "learning_rate": 0.0006,
+ "loss": 4.140815258026123,
+ "step": 2589
+ },
+ {
+ "epoch": 35.97816593886463,
+ "grad_norm": 0.017126062884926796,
+ "learning_rate": 0.0006,
+ "loss": 4.157176971435547,
+ "step": 2590
+ },
+ {
+ "epoch": 35.992139737991266,
+ "grad_norm": 0.01860199309885502,
+ "learning_rate": 0.0006,
+ "loss": 4.252880096435547,
+ "step": 2591
+ },
+ {
+ "epoch": 36.0,
+ "grad_norm": 0.021780220791697502,
+ "learning_rate": 0.0006,
+ "loss": 4.249861717224121,
+ "step": 2592
+ },
+ {
+ "epoch": 36.0,
+ "eval_loss": 4.606466293334961,
+ "eval_runtime": 56.6217,
+ "eval_samples_per_second": 43.128,
+ "eval_steps_per_second": 1.36,
+ "step": 2592
+ },
+ {
+ "epoch": 36.01397379912664,
+ "grad_norm": 0.020892828702926636,
+ "learning_rate": 0.0006,
+ "loss": 4.102055549621582,
+ "step": 2593
+ },
+ {
+ "epoch": 36.02794759825328,
+ "grad_norm": 0.020306255668401718,
+ "learning_rate": 0.0006,
+ "loss": 4.184308052062988,
+ "step": 2594
+ },
+ {
+ "epoch": 36.041921397379916,
+ "grad_norm": 0.02025097794830799,
+ "learning_rate": 0.0006,
+ "loss": 3.9691858291625977,
+ "step": 2595
+ },
+ {
+ "epoch": 36.05589519650655,
+ "grad_norm": 0.020254382863640785,
+ "learning_rate": 0.0006,
+ "loss": 4.02957820892334,
+ "step": 2596
+ },
+ {
+ "epoch": 36.069868995633186,
+ "grad_norm": 0.019307196140289307,
+ "learning_rate": 0.0006,
+ "loss": 4.033181667327881,
+ "step": 2597
+ },
+ {
+ "epoch": 36.083842794759825,
+ "grad_norm": 0.0198476854711771,
+ "learning_rate": 0.0006,
+ "loss": 4.0816874504089355,
+ "step": 2598
+ },
+ {
+ "epoch": 36.09781659388646,
+ "grad_norm": 0.019916515797376633,
+ "learning_rate": 0.0006,
+ "loss": 4.164889812469482,
+ "step": 2599
+ },
+ {
+ "epoch": 36.1117903930131,
+ "grad_norm": 0.0189144778996706,
+ "learning_rate": 0.0006,
+ "loss": 4.038427352905273,
+ "step": 2600
+ },
+ {
+ "epoch": 36.12576419213974,
+ "grad_norm": 0.01754768192768097,
+ "learning_rate": 0.0006,
+ "loss": 4.010427474975586,
+ "step": 2601
+ },
+ {
+ "epoch": 36.13973799126637,
+ "grad_norm": 0.016862990334630013,
+ "learning_rate": 0.0006,
+ "loss": 4.035970211029053,
+ "step": 2602
+ },
+ {
+ "epoch": 36.15371179039301,
+ "grad_norm": 0.019616752862930298,
+ "learning_rate": 0.0006,
+ "loss": 4.079898834228516,
+ "step": 2603
+ },
+ {
+ "epoch": 36.16768558951965,
+ "grad_norm": 0.020131465047597885,
+ "learning_rate": 0.0006,
+ "loss": 4.144684791564941,
+ "step": 2604
+ },
+ {
+ "epoch": 36.18165938864629,
+ "grad_norm": 0.02072521299123764,
+ "learning_rate": 0.0006,
+ "loss": 4.013660430908203,
+ "step": 2605
+ },
+ {
+ "epoch": 36.19563318777293,
+ "grad_norm": 0.01924615167081356,
+ "learning_rate": 0.0006,
+ "loss": 4.071398735046387,
+ "step": 2606
+ },
+ {
+ "epoch": 36.209606986899566,
+ "grad_norm": 0.02013954520225525,
+ "learning_rate": 0.0006,
+ "loss": 4.127094268798828,
+ "step": 2607
+ },
+ {
+ "epoch": 36.223580786026204,
+ "grad_norm": 0.02042008936405182,
+ "learning_rate": 0.0006,
+ "loss": 4.154261589050293,
+ "step": 2608
+ },
+ {
+ "epoch": 36.237554585152836,
+ "grad_norm": 0.021998876705765724,
+ "learning_rate": 0.0006,
+ "loss": 4.204150199890137,
+ "step": 2609
+ },
+ {
+ "epoch": 36.251528384279474,
+ "grad_norm": 0.022946683689951897,
+ "learning_rate": 0.0006,
+ "loss": 4.013632774353027,
+ "step": 2610
+ },
+ {
+ "epoch": 36.26550218340611,
+ "grad_norm": 0.02061830274760723,
+ "learning_rate": 0.0006,
+ "loss": 4.171084403991699,
+ "step": 2611
+ },
+ {
+ "epoch": 36.27947598253275,
+ "grad_norm": 0.01883152686059475,
+ "learning_rate": 0.0006,
+ "loss": 4.149848461151123,
+ "step": 2612
+ },
+ {
+ "epoch": 36.29344978165939,
+ "grad_norm": 0.019945282489061356,
+ "learning_rate": 0.0006,
+ "loss": 4.174656867980957,
+ "step": 2613
+ },
+ {
+ "epoch": 36.30742358078603,
+ "grad_norm": 0.020751051604747772,
+ "learning_rate": 0.0006,
+ "loss": 4.080747127532959,
+ "step": 2614
+ },
+ {
+ "epoch": 36.32139737991266,
+ "grad_norm": 0.01879691146314144,
+ "learning_rate": 0.0006,
+ "loss": 4.057356357574463,
+ "step": 2615
+ },
+ {
+ "epoch": 36.3353711790393,
+ "grad_norm": 0.019037071615457535,
+ "learning_rate": 0.0006,
+ "loss": 4.204607009887695,
+ "step": 2616
+ },
+ {
+ "epoch": 36.34934497816594,
+ "grad_norm": 0.01872078701853752,
+ "learning_rate": 0.0006,
+ "loss": 4.050586700439453,
+ "step": 2617
+ },
+ {
+ "epoch": 36.36331877729258,
+ "grad_norm": 0.018092811107635498,
+ "learning_rate": 0.0006,
+ "loss": 4.102832794189453,
+ "step": 2618
+ },
+ {
+ "epoch": 36.377292576419215,
+ "grad_norm": 0.017010487616062164,
+ "learning_rate": 0.0006,
+ "loss": 4.191788673400879,
+ "step": 2619
+ },
+ {
+ "epoch": 36.391266375545854,
+ "grad_norm": 0.01691143773496151,
+ "learning_rate": 0.0006,
+ "loss": 4.160456657409668,
+ "step": 2620
+ },
+ {
+ "epoch": 36.40524017467249,
+ "grad_norm": 0.017046816647052765,
+ "learning_rate": 0.0006,
+ "loss": 4.073828220367432,
+ "step": 2621
+ },
+ {
+ "epoch": 36.419213973799124,
+ "grad_norm": 0.01707981713116169,
+ "learning_rate": 0.0006,
+ "loss": 4.0980095863342285,
+ "step": 2622
+ },
+ {
+ "epoch": 36.43318777292576,
+ "grad_norm": 0.01630452275276184,
+ "learning_rate": 0.0006,
+ "loss": 4.101321220397949,
+ "step": 2623
+ },
+ {
+ "epoch": 36.4471615720524,
+ "grad_norm": 0.01576915942132473,
+ "learning_rate": 0.0006,
+ "loss": 4.158249855041504,
+ "step": 2624
+ },
+ {
+ "epoch": 36.46113537117904,
+ "grad_norm": 0.016179712489247322,
+ "learning_rate": 0.0006,
+ "loss": 4.077852249145508,
+ "step": 2625
+ },
+ {
+ "epoch": 36.47510917030568,
+ "grad_norm": 0.01609169878065586,
+ "learning_rate": 0.0006,
+ "loss": 4.187317848205566,
+ "step": 2626
+ },
+ {
+ "epoch": 36.48908296943232,
+ "grad_norm": 0.017148254439234734,
+ "learning_rate": 0.0006,
+ "loss": 4.109071731567383,
+ "step": 2627
+ },
+ {
+ "epoch": 36.50305676855895,
+ "grad_norm": 0.015842624008655548,
+ "learning_rate": 0.0006,
+ "loss": 4.145096302032471,
+ "step": 2628
+ },
+ {
+ "epoch": 36.51703056768559,
+ "grad_norm": 0.014906194061040878,
+ "learning_rate": 0.0006,
+ "loss": 4.028440475463867,
+ "step": 2629
+ },
+ {
+ "epoch": 36.531004366812226,
+ "grad_norm": 0.01485038734972477,
+ "learning_rate": 0.0006,
+ "loss": 4.175201416015625,
+ "step": 2630
+ },
+ {
+ "epoch": 36.544978165938865,
+ "grad_norm": 0.015834150835871696,
+ "learning_rate": 0.0006,
+ "loss": 4.076434135437012,
+ "step": 2631
+ },
+ {
+ "epoch": 36.5589519650655,
+ "grad_norm": 0.015411358326673508,
+ "learning_rate": 0.0006,
+ "loss": 4.03912878036499,
+ "step": 2632
+ },
+ {
+ "epoch": 36.57292576419214,
+ "grad_norm": 0.014762846753001213,
+ "learning_rate": 0.0006,
+ "loss": 4.019827365875244,
+ "step": 2633
+ },
+ {
+ "epoch": 36.58689956331878,
+ "grad_norm": 0.015269497409462929,
+ "learning_rate": 0.0006,
+ "loss": 4.116957187652588,
+ "step": 2634
+ },
+ {
+ "epoch": 36.60087336244541,
+ "grad_norm": 0.015171617269515991,
+ "learning_rate": 0.0006,
+ "loss": 4.2069926261901855,
+ "step": 2635
+ },
+ {
+ "epoch": 36.61484716157205,
+ "grad_norm": 0.014770902693271637,
+ "learning_rate": 0.0006,
+ "loss": 4.092021942138672,
+ "step": 2636
+ },
+ {
+ "epoch": 36.62882096069869,
+ "grad_norm": 0.014957531355321407,
+ "learning_rate": 0.0006,
+ "loss": 4.16819953918457,
+ "step": 2637
+ },
+ {
+ "epoch": 36.64279475982533,
+ "grad_norm": 0.014774451963603497,
+ "learning_rate": 0.0006,
+ "loss": 4.046073913574219,
+ "step": 2638
+ },
+ {
+ "epoch": 36.65676855895197,
+ "grad_norm": 0.014320445246994495,
+ "learning_rate": 0.0006,
+ "loss": 3.977792501449585,
+ "step": 2639
+ },
+ {
+ "epoch": 36.670742358078606,
+ "grad_norm": 0.01413130946457386,
+ "learning_rate": 0.0006,
+ "loss": 4.130649566650391,
+ "step": 2640
+ },
+ {
+ "epoch": 36.68471615720524,
+ "grad_norm": 0.014336307533085346,
+ "learning_rate": 0.0006,
+ "loss": 4.161357402801514,
+ "step": 2641
+ },
+ {
+ "epoch": 36.698689956331876,
+ "grad_norm": 0.014516931027173996,
+ "learning_rate": 0.0006,
+ "loss": 4.14248514175415,
+ "step": 2642
+ },
+ {
+ "epoch": 36.712663755458514,
+ "grad_norm": 0.014377126470208168,
+ "learning_rate": 0.0006,
+ "loss": 4.062711715698242,
+ "step": 2643
+ },
+ {
+ "epoch": 36.72663755458515,
+ "grad_norm": 0.014332668855786324,
+ "learning_rate": 0.0006,
+ "loss": 4.06423807144165,
+ "step": 2644
+ },
+ {
+ "epoch": 36.74061135371179,
+ "grad_norm": 0.014030571095645428,
+ "learning_rate": 0.0006,
+ "loss": 4.178613662719727,
+ "step": 2645
+ },
+ {
+ "epoch": 36.75458515283843,
+ "grad_norm": 0.014376776292920113,
+ "learning_rate": 0.0006,
+ "loss": 4.073130130767822,
+ "step": 2646
+ },
+ {
+ "epoch": 36.76855895196506,
+ "grad_norm": 0.013827716931700706,
+ "learning_rate": 0.0006,
+ "loss": 4.043869972229004,
+ "step": 2647
+ },
+ {
+ "epoch": 36.7825327510917,
+ "grad_norm": 0.015548762865364552,
+ "learning_rate": 0.0006,
+ "loss": 4.134049415588379,
+ "step": 2648
+ },
+ {
+ "epoch": 36.79650655021834,
+ "grad_norm": 0.017041970044374466,
+ "learning_rate": 0.0006,
+ "loss": 4.205024242401123,
+ "step": 2649
+ },
+ {
+ "epoch": 36.81048034934498,
+ "grad_norm": 0.017638275399804115,
+ "learning_rate": 0.0006,
+ "loss": 4.199924468994141,
+ "step": 2650
+ },
+ {
+ "epoch": 36.82445414847162,
+ "grad_norm": 0.017771627753973007,
+ "learning_rate": 0.0006,
+ "loss": 4.15706729888916,
+ "step": 2651
+ },
+ {
+ "epoch": 36.838427947598255,
+ "grad_norm": 0.015933562070131302,
+ "learning_rate": 0.0006,
+ "loss": 4.220742225646973,
+ "step": 2652
+ },
+ {
+ "epoch": 36.852401746724894,
+ "grad_norm": 0.01668144389986992,
+ "learning_rate": 0.0006,
+ "loss": 4.161426544189453,
+ "step": 2653
+ },
+ {
+ "epoch": 36.866375545851525,
+ "grad_norm": 0.018014680594205856,
+ "learning_rate": 0.0006,
+ "loss": 4.183046817779541,
+ "step": 2654
+ },
+ {
+ "epoch": 36.880349344978164,
+ "grad_norm": 0.018808670341968536,
+ "learning_rate": 0.0006,
+ "loss": 4.105293273925781,
+ "step": 2655
+ },
+ {
+ "epoch": 36.8943231441048,
+ "grad_norm": 0.018003476783633232,
+ "learning_rate": 0.0006,
+ "loss": 4.103171348571777,
+ "step": 2656
+ },
+ {
+ "epoch": 36.90829694323144,
+ "grad_norm": 0.016865914687514305,
+ "learning_rate": 0.0006,
+ "loss": 4.125634670257568,
+ "step": 2657
+ },
+ {
+ "epoch": 36.92227074235808,
+ "grad_norm": 0.018371153622865677,
+ "learning_rate": 0.0006,
+ "loss": 4.192461013793945,
+ "step": 2658
+ },
+ {
+ "epoch": 36.93624454148472,
+ "grad_norm": 0.019045377150177956,
+ "learning_rate": 0.0006,
+ "loss": 4.198105812072754,
+ "step": 2659
+ },
+ {
+ "epoch": 36.95021834061135,
+ "grad_norm": 0.01594168320298195,
+ "learning_rate": 0.0006,
+ "loss": 4.169788360595703,
+ "step": 2660
+ },
+ {
+ "epoch": 36.96419213973799,
+ "grad_norm": 0.015645820647478104,
+ "learning_rate": 0.0006,
+ "loss": 4.047277450561523,
+ "step": 2661
+ },
+ {
+ "epoch": 36.97816593886463,
+ "grad_norm": 0.016313156113028526,
+ "learning_rate": 0.0006,
+ "loss": 4.087392807006836,
+ "step": 2662
+ },
+ {
+ "epoch": 36.992139737991266,
+ "grad_norm": 0.017050625756382942,
+ "learning_rate": 0.0006,
+ "loss": 4.0908002853393555,
+ "step": 2663
+ },
+ {
+ "epoch": 37.0,
+ "grad_norm": 0.017730968073010445,
+ "learning_rate": 0.0006,
+ "loss": 4.110957622528076,
+ "step": 2664
+ },
+ {
+ "epoch": 37.0,
+ "eval_loss": 4.563793182373047,
+ "eval_runtime": 56.5922,
+ "eval_samples_per_second": 43.151,
+ "eval_steps_per_second": 1.361,
+ "step": 2664
+ },
+ {
+ "epoch": 37.01397379912664,
+ "grad_norm": 0.015687910839915276,
+ "learning_rate": 0.0006,
+ "loss": 4.143589019775391,
+ "step": 2665
+ },
+ {
+ "epoch": 37.02794759825328,
+ "grad_norm": 0.017451366409659386,
+ "learning_rate": 0.0006,
+ "loss": 4.0326666831970215,
+ "step": 2666
+ },
+ {
+ "epoch": 37.041921397379916,
+ "grad_norm": 0.01909538544714451,
+ "learning_rate": 0.0006,
+ "loss": 3.9706249237060547,
+ "step": 2667
+ },
+ {
+ "epoch": 37.05589519650655,
+ "grad_norm": 0.01988379657268524,
+ "learning_rate": 0.0006,
+ "loss": 3.9531009197235107,
+ "step": 2668
+ },
+ {
+ "epoch": 37.069868995633186,
+ "grad_norm": 0.02035115472972393,
+ "learning_rate": 0.0006,
+ "loss": 3.962214469909668,
+ "step": 2669
+ },
+ {
+ "epoch": 37.083842794759825,
+ "grad_norm": 0.01952001079916954,
+ "learning_rate": 0.0006,
+ "loss": 4.182262897491455,
+ "step": 2670
+ },
+ {
+ "epoch": 37.09781659388646,
+ "grad_norm": 0.01854325644671917,
+ "learning_rate": 0.0006,
+ "loss": 4.127663612365723,
+ "step": 2671
+ },
+ {
+ "epoch": 37.1117903930131,
+ "grad_norm": 0.017757616937160492,
+ "learning_rate": 0.0006,
+ "loss": 4.100804805755615,
+ "step": 2672
+ },
+ {
+ "epoch": 37.12576419213974,
+ "grad_norm": 0.016712479293346405,
+ "learning_rate": 0.0006,
+ "loss": 4.026284217834473,
+ "step": 2673
+ },
+ {
+ "epoch": 37.13973799126637,
+ "grad_norm": 0.0168139711022377,
+ "learning_rate": 0.0006,
+ "loss": 4.183882713317871,
+ "step": 2674
+ },
+ {
+ "epoch": 37.15371179039301,
+ "grad_norm": 0.017060423269867897,
+ "learning_rate": 0.0006,
+ "loss": 4.131832599639893,
+ "step": 2675
+ },
+ {
+ "epoch": 37.16768558951965,
+ "grad_norm": 0.01726631447672844,
+ "learning_rate": 0.0006,
+ "loss": 3.9156057834625244,
+ "step": 2676
+ },
+ {
+ "epoch": 37.18165938864629,
+ "grad_norm": 0.016348622739315033,
+ "learning_rate": 0.0006,
+ "loss": 4.1731367111206055,
+ "step": 2677
+ },
+ {
+ "epoch": 37.19563318777293,
+ "grad_norm": 0.016640059649944305,
+ "learning_rate": 0.0006,
+ "loss": 4.10997200012207,
+ "step": 2678
+ },
+ {
+ "epoch": 37.209606986899566,
+ "grad_norm": 0.01601342298090458,
+ "learning_rate": 0.0006,
+ "loss": 4.081145286560059,
+ "step": 2679
+ },
+ {
+ "epoch": 37.223580786026204,
+ "grad_norm": 0.01735631190240383,
+ "learning_rate": 0.0006,
+ "loss": 4.202108860015869,
+ "step": 2680
+ },
+ {
+ "epoch": 37.237554585152836,
+ "grad_norm": 0.019164685159921646,
+ "learning_rate": 0.0006,
+ "loss": 4.104109764099121,
+ "step": 2681
+ },
+ {
+ "epoch": 37.251528384279474,
+ "grad_norm": 0.01881759613752365,
+ "learning_rate": 0.0006,
+ "loss": 4.145532131195068,
+ "step": 2682
+ },
+ {
+ "epoch": 37.26550218340611,
+ "grad_norm": 0.017656555399298668,
+ "learning_rate": 0.0006,
+ "loss": 4.012782096862793,
+ "step": 2683
+ },
+ {
+ "epoch": 37.27947598253275,
+ "grad_norm": 0.01624823734164238,
+ "learning_rate": 0.0006,
+ "loss": 4.001727104187012,
+ "step": 2684
+ },
+ {
+ "epoch": 37.29344978165939,
+ "grad_norm": 0.016973139718174934,
+ "learning_rate": 0.0006,
+ "loss": 4.095399856567383,
+ "step": 2685
+ },
+ {
+ "epoch": 37.30742358078603,
+ "grad_norm": 0.015503671951591969,
+ "learning_rate": 0.0006,
+ "loss": 4.034431457519531,
+ "step": 2686
+ },
+ {
+ "epoch": 37.32139737991266,
+ "grad_norm": 0.015370250679552555,
+ "learning_rate": 0.0006,
+ "loss": 4.075836181640625,
+ "step": 2687
+ },
+ {
+ "epoch": 37.3353711790393,
+ "grad_norm": 0.01564021222293377,
+ "learning_rate": 0.0006,
+ "loss": 4.125202178955078,
+ "step": 2688
+ },
+ {
+ "epoch": 37.34934497816594,
+ "grad_norm": 0.016717975959181786,
+ "learning_rate": 0.0006,
+ "loss": 4.153017044067383,
+ "step": 2689
+ },
+ {
+ "epoch": 37.36331877729258,
+ "grad_norm": 0.015263617038726807,
+ "learning_rate": 0.0006,
+ "loss": 4.021814823150635,
+ "step": 2690
+ },
+ {
+ "epoch": 37.377292576419215,
+ "grad_norm": 0.016497142612934113,
+ "learning_rate": 0.0006,
+ "loss": 4.069095611572266,
+ "step": 2691
+ },
+ {
+ "epoch": 37.391266375545854,
+ "grad_norm": 0.01669355481863022,
+ "learning_rate": 0.0006,
+ "loss": 4.174587726593018,
+ "step": 2692
+ },
+ {
+ "epoch": 37.40524017467249,
+ "grad_norm": 0.019695773720741272,
+ "learning_rate": 0.0006,
+ "loss": 4.0073041915893555,
+ "step": 2693
+ },
+ {
+ "epoch": 37.419213973799124,
+ "grad_norm": 0.023720644414424896,
+ "learning_rate": 0.0006,
+ "loss": 4.111948013305664,
+ "step": 2694
+ },
+ {
+ "epoch": 37.43318777292576,
+ "grad_norm": 0.02399350143969059,
+ "learning_rate": 0.0006,
+ "loss": 4.081164360046387,
+ "step": 2695
+ },
+ {
+ "epoch": 37.4471615720524,
+ "grad_norm": 0.019630271941423416,
+ "learning_rate": 0.0006,
+ "loss": 4.065731048583984,
+ "step": 2696
+ },
+ {
+ "epoch": 37.46113537117904,
+ "grad_norm": 0.019679008051753044,
+ "learning_rate": 0.0006,
+ "loss": 3.993852376937866,
+ "step": 2697
+ },
+ {
+ "epoch": 37.47510917030568,
+ "grad_norm": 0.020597299560904503,
+ "learning_rate": 0.0006,
+ "loss": 4.275555610656738,
+ "step": 2698
+ },
+ {
+ "epoch": 37.48908296943232,
+ "grad_norm": 0.01983889564871788,
+ "learning_rate": 0.0006,
+ "loss": 4.254708290100098,
+ "step": 2699
+ },
+ {
+ "epoch": 37.50305676855895,
+ "grad_norm": 0.019150668755173683,
+ "learning_rate": 0.0006,
+ "loss": 4.097885608673096,
+ "step": 2700
+ },
+ {
+ "epoch": 37.51703056768559,
+ "grad_norm": 0.018847715109586716,
+ "learning_rate": 0.0006,
+ "loss": 4.054081439971924,
+ "step": 2701
+ },
+ {
+ "epoch": 37.531004366812226,
+ "grad_norm": 0.019607288762927055,
+ "learning_rate": 0.0006,
+ "loss": 3.9891440868377686,
+ "step": 2702
+ },
+ {
+ "epoch": 37.544978165938865,
+ "grad_norm": 0.018863150849938393,
+ "learning_rate": 0.0006,
+ "loss": 4.084877014160156,
+ "step": 2703
+ },
+ {
+ "epoch": 37.5589519650655,
+ "grad_norm": 0.017340239137411118,
+ "learning_rate": 0.0006,
+ "loss": 4.113290786743164,
+ "step": 2704
+ },
+ {
+ "epoch": 37.57292576419214,
+ "grad_norm": 0.015628693625330925,
+ "learning_rate": 0.0006,
+ "loss": 4.1483001708984375,
+ "step": 2705
+ },
+ {
+ "epoch": 37.58689956331878,
+ "grad_norm": 0.016493283212184906,
+ "learning_rate": 0.0006,
+ "loss": 4.1040143966674805,
+ "step": 2706
+ },
+ {
+ "epoch": 37.60087336244541,
+ "grad_norm": 0.016732443124055862,
+ "learning_rate": 0.0006,
+ "loss": 4.07205057144165,
+ "step": 2707
+ },
+ {
+ "epoch": 37.61484716157205,
+ "grad_norm": 0.017176156863570213,
+ "learning_rate": 0.0006,
+ "loss": 4.1279754638671875,
+ "step": 2708
+ },
+ {
+ "epoch": 37.62882096069869,
+ "grad_norm": 0.015730608254671097,
+ "learning_rate": 0.0006,
+ "loss": 4.086658000946045,
+ "step": 2709
+ },
+ {
+ "epoch": 37.64279475982533,
+ "grad_norm": 0.016111090779304504,
+ "learning_rate": 0.0006,
+ "loss": 4.064335823059082,
+ "step": 2710
+ },
+ {
+ "epoch": 37.65676855895197,
+ "grad_norm": 0.01769379712641239,
+ "learning_rate": 0.0006,
+ "loss": 4.116047382354736,
+ "step": 2711
+ },
+ {
+ "epoch": 37.670742358078606,
+ "grad_norm": 0.01683102548122406,
+ "learning_rate": 0.0006,
+ "loss": 4.095986843109131,
+ "step": 2712
+ },
+ {
+ "epoch": 37.68471615720524,
+ "grad_norm": 0.015726450830698013,
+ "learning_rate": 0.0006,
+ "loss": 4.10775089263916,
+ "step": 2713
+ },
+ {
+ "epoch": 37.698689956331876,
+ "grad_norm": 0.014760235324501991,
+ "learning_rate": 0.0006,
+ "loss": 4.1234893798828125,
+ "step": 2714
+ },
+ {
+ "epoch": 37.712663755458514,
+ "grad_norm": 0.014933438040316105,
+ "learning_rate": 0.0006,
+ "loss": 4.079200744628906,
+ "step": 2715
+ },
+ {
+ "epoch": 37.72663755458515,
+ "grad_norm": 0.015723997727036476,
+ "learning_rate": 0.0006,
+ "loss": 4.07865047454834,
+ "step": 2716
+ },
+ {
+ "epoch": 37.74061135371179,
+ "grad_norm": 0.014539001509547234,
+ "learning_rate": 0.0006,
+ "loss": 4.036242485046387,
+ "step": 2717
+ },
+ {
+ "epoch": 37.75458515283843,
+ "grad_norm": 0.014844979159533978,
+ "learning_rate": 0.0006,
+ "loss": 4.1051483154296875,
+ "step": 2718
+ },
+ {
+ "epoch": 37.76855895196506,
+ "grad_norm": 0.015939844772219658,
+ "learning_rate": 0.0006,
+ "loss": 4.017253875732422,
+ "step": 2719
+ },
+ {
+ "epoch": 37.7825327510917,
+ "grad_norm": 0.014771555550396442,
+ "learning_rate": 0.0006,
+ "loss": 4.0694661140441895,
+ "step": 2720
+ },
+ {
+ "epoch": 37.79650655021834,
+ "grad_norm": 0.015777921304106712,
+ "learning_rate": 0.0006,
+ "loss": 3.947587490081787,
+ "step": 2721
+ },
+ {
+ "epoch": 37.81048034934498,
+ "grad_norm": 0.017451291903853416,
+ "learning_rate": 0.0006,
+ "loss": 4.093625068664551,
+ "step": 2722
+ },
+ {
+ "epoch": 37.82445414847162,
+ "grad_norm": 0.016829246655106544,
+ "learning_rate": 0.0006,
+ "loss": 4.09030818939209,
+ "step": 2723
+ },
+ {
+ "epoch": 37.838427947598255,
+ "grad_norm": 0.016984397545456886,
+ "learning_rate": 0.0006,
+ "loss": 4.140626907348633,
+ "step": 2724
+ },
+ {
+ "epoch": 37.852401746724894,
+ "grad_norm": 0.01732603646814823,
+ "learning_rate": 0.0006,
+ "loss": 4.128276824951172,
+ "step": 2725
+ },
+ {
+ "epoch": 37.866375545851525,
+ "grad_norm": 0.01744619384407997,
+ "learning_rate": 0.0006,
+ "loss": 4.103130340576172,
+ "step": 2726
+ },
+ {
+ "epoch": 37.880349344978164,
+ "grad_norm": 0.016331320628523827,
+ "learning_rate": 0.0006,
+ "loss": 4.037951469421387,
+ "step": 2727
+ },
+ {
+ "epoch": 37.8943231441048,
+ "grad_norm": 0.015919430181384087,
+ "learning_rate": 0.0006,
+ "loss": 4.121513366699219,
+ "step": 2728
+ },
+ {
+ "epoch": 37.90829694323144,
+ "grad_norm": 0.016153695061802864,
+ "learning_rate": 0.0006,
+ "loss": 4.099384307861328,
+ "step": 2729
+ },
+ {
+ "epoch": 37.92227074235808,
+ "grad_norm": 0.01628800667822361,
+ "learning_rate": 0.0006,
+ "loss": 4.092325687408447,
+ "step": 2730
+ },
+ {
+ "epoch": 37.93624454148472,
+ "grad_norm": 0.017285684123635292,
+ "learning_rate": 0.0006,
+ "loss": 3.9828906059265137,
+ "step": 2731
+ },
+ {
+ "epoch": 37.95021834061135,
+ "grad_norm": 0.016047554090619087,
+ "learning_rate": 0.0006,
+ "loss": 4.100419044494629,
+ "step": 2732
+ },
+ {
+ "epoch": 37.96419213973799,
+ "grad_norm": 0.01562986709177494,
+ "learning_rate": 0.0006,
+ "loss": 4.065832614898682,
+ "step": 2733
+ },
+ {
+ "epoch": 37.97816593886463,
+ "grad_norm": 0.01677163504064083,
+ "learning_rate": 0.0006,
+ "loss": 4.077518463134766,
+ "step": 2734
+ },
+ {
+ "epoch": 37.992139737991266,
+ "grad_norm": 0.01755492016673088,
+ "learning_rate": 0.0006,
+ "loss": 4.1076154708862305,
+ "step": 2735
+ },
+ {
+ "epoch": 38.0,
+ "grad_norm": 0.01929762214422226,
+ "learning_rate": 0.0006,
+ "loss": 4.09603214263916,
+ "step": 2736
+ },
+ {
+ "epoch": 38.0,
+ "eval_loss": 4.582442283630371,
+ "eval_runtime": 62.8649,
+ "eval_samples_per_second": 38.845,
+ "eval_steps_per_second": 1.225,
+ "step": 2736
+ },
+ {
+ "epoch": 38.01397379912664,
+ "grad_norm": 0.019129136577248573,
+ "learning_rate": 0.0006,
+ "loss": 4.095005989074707,
+ "step": 2737
+ },
+ {
+ "epoch": 38.02794759825328,
+ "grad_norm": 0.019167408347129822,
+ "learning_rate": 0.0006,
+ "loss": 4.054426670074463,
+ "step": 2738
+ },
+ {
+ "epoch": 38.041921397379916,
+ "grad_norm": 0.01789460889995098,
+ "learning_rate": 0.0006,
+ "loss": 4.0434417724609375,
+ "step": 2739
+ },
+ {
+ "epoch": 38.05589519650655,
+ "grad_norm": 0.015947027131915092,
+ "learning_rate": 0.0006,
+ "loss": 4.074516296386719,
+ "step": 2740
+ },
+ {
+ "epoch": 38.069868995633186,
+ "grad_norm": 0.016667252406477928,
+ "learning_rate": 0.0006,
+ "loss": 4.108545303344727,
+ "step": 2741
+ },
+ {
+ "epoch": 38.083842794759825,
+ "grad_norm": 0.016014399006962776,
+ "learning_rate": 0.0006,
+ "loss": 4.026925086975098,
+ "step": 2742
+ },
+ {
+ "epoch": 38.09781659388646,
+ "grad_norm": 0.015397695824503899,
+ "learning_rate": 0.0006,
+ "loss": 4.163970470428467,
+ "step": 2743
+ },
+ {
+ "epoch": 38.1117903930131,
+ "grad_norm": 0.01563289947807789,
+ "learning_rate": 0.0006,
+ "loss": 4.116578578948975,
+ "step": 2744
+ },
+ {
+ "epoch": 38.12576419213974,
+ "grad_norm": 0.015512318350374699,
+ "learning_rate": 0.0006,
+ "loss": 4.096311569213867,
+ "step": 2745
+ },
+ {
+ "epoch": 38.13973799126637,
+ "grad_norm": 0.015426642261445522,
+ "learning_rate": 0.0006,
+ "loss": 4.017065048217773,
+ "step": 2746
+ },
+ {
+ "epoch": 38.15371179039301,
+ "grad_norm": 0.015115435235202312,
+ "learning_rate": 0.0006,
+ "loss": 4.005434989929199,
+ "step": 2747
+ },
+ {
+ "epoch": 38.16768558951965,
+ "grad_norm": 0.015235894359648228,
+ "learning_rate": 0.0006,
+ "loss": 3.982743263244629,
+ "step": 2748
+ },
+ {
+ "epoch": 38.18165938864629,
+ "grad_norm": 0.016095101833343506,
+ "learning_rate": 0.0006,
+ "loss": 4.0849385261535645,
+ "step": 2749
+ },
+ {
+ "epoch": 38.19563318777293,
+ "grad_norm": 0.015201197937130928,
+ "learning_rate": 0.0006,
+ "loss": 4.103863716125488,
+ "step": 2750
+ },
+ {
+ "epoch": 38.209606986899566,
+ "grad_norm": 0.015687478706240654,
+ "learning_rate": 0.0006,
+ "loss": 4.12074089050293,
+ "step": 2751
+ },
+ {
+ "epoch": 38.223580786026204,
+ "grad_norm": 0.015201589092612267,
+ "learning_rate": 0.0006,
+ "loss": 4.056872367858887,
+ "step": 2752
+ },
+ {
+ "epoch": 38.237554585152836,
+ "grad_norm": 0.015154894441366196,
+ "learning_rate": 0.0006,
+ "loss": 4.089670181274414,
+ "step": 2753
+ },
+ {
+ "epoch": 38.251528384279474,
+ "grad_norm": 0.01586608588695526,
+ "learning_rate": 0.0006,
+ "loss": 4.155317306518555,
+ "step": 2754
+ },
+ {
+ "epoch": 38.26550218340611,
+ "grad_norm": 0.01609654724597931,
+ "learning_rate": 0.0006,
+ "loss": 3.995800495147705,
+ "step": 2755
+ },
+ {
+ "epoch": 38.27947598253275,
+ "grad_norm": 0.015495185740292072,
+ "learning_rate": 0.0006,
+ "loss": 3.995769739151001,
+ "step": 2756
+ },
+ {
+ "epoch": 38.29344978165939,
+ "grad_norm": 0.017042381688952446,
+ "learning_rate": 0.0006,
+ "loss": 4.132832050323486,
+ "step": 2757
+ },
+ {
+ "epoch": 38.30742358078603,
+ "grad_norm": 0.018706614151597023,
+ "learning_rate": 0.0006,
+ "loss": 4.038820743560791,
+ "step": 2758
+ },
+ {
+ "epoch": 38.32139737991266,
+ "grad_norm": 0.018316026777029037,
+ "learning_rate": 0.0006,
+ "loss": 3.972052574157715,
+ "step": 2759
+ },
+ {
+ "epoch": 38.3353711790393,
+ "grad_norm": 0.018375389277935028,
+ "learning_rate": 0.0006,
+ "loss": 4.152097702026367,
+ "step": 2760
+ },
+ {
+ "epoch": 38.34934497816594,
+ "grad_norm": 0.016879886388778687,
+ "learning_rate": 0.0006,
+ "loss": 4.020869731903076,
+ "step": 2761
+ },
+ {
+ "epoch": 38.36331877729258,
+ "grad_norm": 0.015728183090686798,
+ "learning_rate": 0.0006,
+ "loss": 4.008016586303711,
+ "step": 2762
+ },
+ {
+ "epoch": 38.377292576419215,
+ "grad_norm": 0.016147758811712265,
+ "learning_rate": 0.0006,
+ "loss": 4.035854339599609,
+ "step": 2763
+ },
+ {
+ "epoch": 38.391266375545854,
+ "grad_norm": 0.017774797976017,
+ "learning_rate": 0.0006,
+ "loss": 4.122506141662598,
+ "step": 2764
+ },
+ {
+ "epoch": 38.40524017467249,
+ "grad_norm": 0.016590069979429245,
+ "learning_rate": 0.0006,
+ "loss": 4.082847595214844,
+ "step": 2765
+ },
+ {
+ "epoch": 38.419213973799124,
+ "grad_norm": 0.016544528305530548,
+ "learning_rate": 0.0006,
+ "loss": 4.184270858764648,
+ "step": 2766
+ },
+ {
+ "epoch": 38.43318777292576,
+ "grad_norm": 0.017972925677895546,
+ "learning_rate": 0.0006,
+ "loss": 3.987558603286743,
+ "step": 2767
+ },
+ {
+ "epoch": 38.4471615720524,
+ "grad_norm": 0.018562227487564087,
+ "learning_rate": 0.0006,
+ "loss": 4.1193461418151855,
+ "step": 2768
+ },
+ {
+ "epoch": 38.46113537117904,
+ "grad_norm": 0.018896685913205147,
+ "learning_rate": 0.0006,
+ "loss": 4.057808876037598,
+ "step": 2769
+ },
+ {
+ "epoch": 38.47510917030568,
+ "grad_norm": 0.016793809831142426,
+ "learning_rate": 0.0006,
+ "loss": 4.069523334503174,
+ "step": 2770
+ },
+ {
+ "epoch": 38.48908296943232,
+ "grad_norm": 0.01698676310479641,
+ "learning_rate": 0.0006,
+ "loss": 3.997624158859253,
+ "step": 2771
+ },
+ {
+ "epoch": 38.50305676855895,
+ "grad_norm": 0.016784202307462692,
+ "learning_rate": 0.0006,
+ "loss": 4.174795627593994,
+ "step": 2772
+ },
+ {
+ "epoch": 38.51703056768559,
+ "grad_norm": 0.016785014420747757,
+ "learning_rate": 0.0006,
+ "loss": 3.999321222305298,
+ "step": 2773
+ },
+ {
+ "epoch": 38.531004366812226,
+ "grad_norm": 0.018615184351801872,
+ "learning_rate": 0.0006,
+ "loss": 4.191237926483154,
+ "step": 2774
+ },
+ {
+ "epoch": 38.544978165938865,
+ "grad_norm": 0.019090790301561356,
+ "learning_rate": 0.0006,
+ "loss": 4.106966018676758,
+ "step": 2775
+ },
+ {
+ "epoch": 38.5589519650655,
+ "grad_norm": 0.018298320472240448,
+ "learning_rate": 0.0006,
+ "loss": 4.063363552093506,
+ "step": 2776
+ },
+ {
+ "epoch": 38.57292576419214,
+ "grad_norm": 0.018541090190410614,
+ "learning_rate": 0.0006,
+ "loss": 4.101193428039551,
+ "step": 2777
+ },
+ {
+ "epoch": 38.58689956331878,
+ "grad_norm": 0.01849970407783985,
+ "learning_rate": 0.0006,
+ "loss": 4.071335792541504,
+ "step": 2778
+ },
+ {
+ "epoch": 38.60087336244541,
+ "grad_norm": 0.018309077247977257,
+ "learning_rate": 0.0006,
+ "loss": 4.245305061340332,
+ "step": 2779
+ },
+ {
+ "epoch": 38.61484716157205,
+ "grad_norm": 0.017681840807199478,
+ "learning_rate": 0.0006,
+ "loss": 4.081173419952393,
+ "step": 2780
+ },
+ {
+ "epoch": 38.62882096069869,
+ "grad_norm": 0.016286224126815796,
+ "learning_rate": 0.0006,
+ "loss": 4.024351596832275,
+ "step": 2781
+ },
+ {
+ "epoch": 38.64279475982533,
+ "grad_norm": 0.015837235376238823,
+ "learning_rate": 0.0006,
+ "loss": 4.028205394744873,
+ "step": 2782
+ },
+ {
+ "epoch": 38.65676855895197,
+ "grad_norm": 0.016062596812844276,
+ "learning_rate": 0.0006,
+ "loss": 4.131122589111328,
+ "step": 2783
+ },
+ {
+ "epoch": 38.670742358078606,
+ "grad_norm": 0.01711239665746689,
+ "learning_rate": 0.0006,
+ "loss": 4.124416351318359,
+ "step": 2784
+ },
+ {
+ "epoch": 38.68471615720524,
+ "grad_norm": 0.015304913744330406,
+ "learning_rate": 0.0006,
+ "loss": 4.056122779846191,
+ "step": 2785
+ },
+ {
+ "epoch": 38.698689956331876,
+ "grad_norm": 0.015971887856721878,
+ "learning_rate": 0.0006,
+ "loss": 4.132555961608887,
+ "step": 2786
+ },
+ {
+ "epoch": 38.712663755458514,
+ "grad_norm": 0.015787221491336823,
+ "learning_rate": 0.0006,
+ "loss": 4.051596164703369,
+ "step": 2787
+ },
+ {
+ "epoch": 38.72663755458515,
+ "grad_norm": 0.01605444774031639,
+ "learning_rate": 0.0006,
+ "loss": 4.191808700561523,
+ "step": 2788
+ },
+ {
+ "epoch": 38.74061135371179,
+ "grad_norm": 0.016383009031414986,
+ "learning_rate": 0.0006,
+ "loss": 4.0648417472839355,
+ "step": 2789
+ },
+ {
+ "epoch": 38.75458515283843,
+ "grad_norm": 0.016135554760694504,
+ "learning_rate": 0.0006,
+ "loss": 3.988563060760498,
+ "step": 2790
+ },
+ {
+ "epoch": 38.76855895196506,
+ "grad_norm": 0.016634158790111542,
+ "learning_rate": 0.0006,
+ "loss": 4.133493423461914,
+ "step": 2791
+ },
+ {
+ "epoch": 38.7825327510917,
+ "grad_norm": 0.01688210666179657,
+ "learning_rate": 0.0006,
+ "loss": 4.091277122497559,
+ "step": 2792
+ },
+ {
+ "epoch": 38.79650655021834,
+ "grad_norm": 0.016102122142910957,
+ "learning_rate": 0.0006,
+ "loss": 4.051219940185547,
+ "step": 2793
+ },
+ {
+ "epoch": 38.81048034934498,
+ "grad_norm": 0.016585038974881172,
+ "learning_rate": 0.0006,
+ "loss": 4.058587074279785,
+ "step": 2794
+ },
+ {
+ "epoch": 38.82445414847162,
+ "grad_norm": 0.017700951546430588,
+ "learning_rate": 0.0006,
+ "loss": 4.016030311584473,
+ "step": 2795
+ },
+ {
+ "epoch": 38.838427947598255,
+ "grad_norm": 0.017922203987836838,
+ "learning_rate": 0.0006,
+ "loss": 4.079977989196777,
+ "step": 2796
+ },
+ {
+ "epoch": 38.852401746724894,
+ "grad_norm": 0.017299890518188477,
+ "learning_rate": 0.0006,
+ "loss": 4.062984943389893,
+ "step": 2797
+ },
+ {
+ "epoch": 38.866375545851525,
+ "grad_norm": 0.018648581579327583,
+ "learning_rate": 0.0006,
+ "loss": 4.065717697143555,
+ "step": 2798
+ },
+ {
+ "epoch": 38.880349344978164,
+ "grad_norm": 0.019057920202612877,
+ "learning_rate": 0.0006,
+ "loss": 4.227892875671387,
+ "step": 2799
+ },
+ {
+ "epoch": 38.8943231441048,
+ "grad_norm": 0.018824422731995583,
+ "learning_rate": 0.0006,
+ "loss": 4.116058349609375,
+ "step": 2800
+ },
+ {
+ "epoch": 38.90829694323144,
+ "grad_norm": 0.019084978848695755,
+ "learning_rate": 0.0006,
+ "loss": 4.061677932739258,
+ "step": 2801
+ },
+ {
+ "epoch": 38.92227074235808,
+ "grad_norm": 0.017842981964349747,
+ "learning_rate": 0.0006,
+ "loss": 4.083813667297363,
+ "step": 2802
+ },
+ {
+ "epoch": 38.93624454148472,
+ "grad_norm": 0.01880858838558197,
+ "learning_rate": 0.0006,
+ "loss": 4.152561187744141,
+ "step": 2803
+ },
+ {
+ "epoch": 38.95021834061135,
+ "grad_norm": 0.017400087788701057,
+ "learning_rate": 0.0006,
+ "loss": 4.154635429382324,
+ "step": 2804
+ },
+ {
+ "epoch": 38.96419213973799,
+ "grad_norm": 0.017682507634162903,
+ "learning_rate": 0.0006,
+ "loss": 4.151128768920898,
+ "step": 2805
+ },
+ {
+ "epoch": 38.97816593886463,
+ "grad_norm": 0.015803052112460136,
+ "learning_rate": 0.0006,
+ "loss": 4.10498571395874,
+ "step": 2806
+ },
+ {
+ "epoch": 38.992139737991266,
+ "grad_norm": 0.01582511141896248,
+ "learning_rate": 0.0006,
+ "loss": 4.140159606933594,
+ "step": 2807
+ },
+ {
+ "epoch": 39.0,
+ "grad_norm": 0.01925666816532612,
+ "learning_rate": 0.0006,
+ "loss": 4.126394271850586,
+ "step": 2808
+ },
+ {
+ "epoch": 39.0,
+ "eval_loss": 4.615617275238037,
+ "eval_runtime": 56.8895,
+ "eval_samples_per_second": 42.925,
+ "eval_steps_per_second": 1.354,
+ "step": 2808
+ },
+ {
+ "epoch": 39.01397379912664,
+ "grad_norm": 0.01677853614091873,
+ "learning_rate": 0.0006,
+ "loss": 4.102395057678223,
+ "step": 2809
+ },
+ {
+ "epoch": 39.02794759825328,
+ "grad_norm": 0.017901137471199036,
+ "learning_rate": 0.0006,
+ "loss": 3.976609230041504,
+ "step": 2810
+ },
+ {
+ "epoch": 39.041921397379916,
+ "grad_norm": 0.017978668212890625,
+ "learning_rate": 0.0006,
+ "loss": 3.944416046142578,
+ "step": 2811
+ },
+ {
+ "epoch": 39.05589519650655,
+ "grad_norm": 0.019967492669820786,
+ "learning_rate": 0.0006,
+ "loss": 3.992234468460083,
+ "step": 2812
+ },
+ {
+ "epoch": 39.069868995633186,
+ "grad_norm": 0.018937544897198677,
+ "learning_rate": 0.0006,
+ "loss": 4.028573036193848,
+ "step": 2813
+ },
+ {
+ "epoch": 39.083842794759825,
+ "grad_norm": 0.016852952539920807,
+ "learning_rate": 0.0006,
+ "loss": 4.062442779541016,
+ "step": 2814
+ },
+ {
+ "epoch": 39.09781659388646,
+ "grad_norm": 0.016643647104501724,
+ "learning_rate": 0.0006,
+ "loss": 4.08871603012085,
+ "step": 2815
+ },
+ {
+ "epoch": 39.1117903930131,
+ "grad_norm": 0.016872866079211235,
+ "learning_rate": 0.0006,
+ "loss": 3.9514245986938477,
+ "step": 2816
+ },
+ {
+ "epoch": 39.12576419213974,
+ "grad_norm": 0.018045669421553612,
+ "learning_rate": 0.0006,
+ "loss": 4.038842678070068,
+ "step": 2817
+ },
+ {
+ "epoch": 39.13973799126637,
+ "grad_norm": 0.017853403463959694,
+ "learning_rate": 0.0006,
+ "loss": 3.93709135055542,
+ "step": 2818
+ },
+ {
+ "epoch": 39.15371179039301,
+ "grad_norm": 0.017507443204522133,
+ "learning_rate": 0.0006,
+ "loss": 3.98665714263916,
+ "step": 2819
+ },
+ {
+ "epoch": 39.16768558951965,
+ "grad_norm": 0.01766853965818882,
+ "learning_rate": 0.0006,
+ "loss": 3.937126636505127,
+ "step": 2820
+ },
+ {
+ "epoch": 39.18165938864629,
+ "grad_norm": 0.016660021618008614,
+ "learning_rate": 0.0006,
+ "loss": 3.8397014141082764,
+ "step": 2821
+ },
+ {
+ "epoch": 39.19563318777293,
+ "grad_norm": 0.017719309777021408,
+ "learning_rate": 0.0006,
+ "loss": 4.003579616546631,
+ "step": 2822
+ },
+ {
+ "epoch": 39.209606986899566,
+ "grad_norm": 0.018234414979815483,
+ "learning_rate": 0.0006,
+ "loss": 4.053859710693359,
+ "step": 2823
+ },
+ {
+ "epoch": 39.223580786026204,
+ "grad_norm": 0.018777024000883102,
+ "learning_rate": 0.0006,
+ "loss": 4.056182861328125,
+ "step": 2824
+ },
+ {
+ "epoch": 39.237554585152836,
+ "grad_norm": 0.017443502321839333,
+ "learning_rate": 0.0006,
+ "loss": 4.025068283081055,
+ "step": 2825
+ },
+ {
+ "epoch": 39.251528384279474,
+ "grad_norm": 0.016500331461429596,
+ "learning_rate": 0.0006,
+ "loss": 4.101631164550781,
+ "step": 2826
+ },
+ {
+ "epoch": 39.26550218340611,
+ "grad_norm": 0.016101544722914696,
+ "learning_rate": 0.0006,
+ "loss": 4.027082443237305,
+ "step": 2827
+ },
+ {
+ "epoch": 39.27947598253275,
+ "grad_norm": 0.015471171587705612,
+ "learning_rate": 0.0006,
+ "loss": 4.022909641265869,
+ "step": 2828
+ },
+ {
+ "epoch": 39.29344978165939,
+ "grad_norm": 0.015779748558998108,
+ "learning_rate": 0.0006,
+ "loss": 4.18796968460083,
+ "step": 2829
+ },
+ {
+ "epoch": 39.30742358078603,
+ "grad_norm": 0.015522679314017296,
+ "learning_rate": 0.0006,
+ "loss": 4.005777359008789,
+ "step": 2830
+ },
+ {
+ "epoch": 39.32139737991266,
+ "grad_norm": 0.015226504765450954,
+ "learning_rate": 0.0006,
+ "loss": 4.108416557312012,
+ "step": 2831
+ },
+ {
+ "epoch": 39.3353711790393,
+ "grad_norm": 0.015095778740942478,
+ "learning_rate": 0.0006,
+ "loss": 4.08029842376709,
+ "step": 2832
+ },
+ {
+ "epoch": 39.34934497816594,
+ "grad_norm": 0.015601031482219696,
+ "learning_rate": 0.0006,
+ "loss": 3.989492416381836,
+ "step": 2833
+ },
+ {
+ "epoch": 39.36331877729258,
+ "grad_norm": 0.014825052581727505,
+ "learning_rate": 0.0006,
+ "loss": 3.998520612716675,
+ "step": 2834
+ },
+ {
+ "epoch": 39.377292576419215,
+ "grad_norm": 0.014821100980043411,
+ "learning_rate": 0.0006,
+ "loss": 4.001777172088623,
+ "step": 2835
+ },
+ {
+ "epoch": 39.391266375545854,
+ "grad_norm": 0.015333754941821098,
+ "learning_rate": 0.0006,
+ "loss": 4.11985969543457,
+ "step": 2836
+ },
+ {
+ "epoch": 39.40524017467249,
+ "grad_norm": 0.015940699726343155,
+ "learning_rate": 0.0006,
+ "loss": 4.025373935699463,
+ "step": 2837
+ },
+ {
+ "epoch": 39.419213973799124,
+ "grad_norm": 0.015096287243068218,
+ "learning_rate": 0.0006,
+ "loss": 4.068551063537598,
+ "step": 2838
+ },
+ {
+ "epoch": 39.43318777292576,
+ "grad_norm": 0.01533548254519701,
+ "learning_rate": 0.0006,
+ "loss": 3.9800801277160645,
+ "step": 2839
+ },
+ {
+ "epoch": 39.4471615720524,
+ "grad_norm": 0.014648732729256153,
+ "learning_rate": 0.0006,
+ "loss": 3.9670190811157227,
+ "step": 2840
+ },
+ {
+ "epoch": 39.46113537117904,
+ "grad_norm": 0.015759671106934547,
+ "learning_rate": 0.0006,
+ "loss": 4.050443649291992,
+ "step": 2841
+ },
+ {
+ "epoch": 39.47510917030568,
+ "grad_norm": 0.016054073348641396,
+ "learning_rate": 0.0006,
+ "loss": 4.22734260559082,
+ "step": 2842
+ },
+ {
+ "epoch": 39.48908296943232,
+ "grad_norm": 0.0160264540463686,
+ "learning_rate": 0.0006,
+ "loss": 4.116174221038818,
+ "step": 2843
+ },
+ {
+ "epoch": 39.50305676855895,
+ "grad_norm": 0.015426610596477985,
+ "learning_rate": 0.0006,
+ "loss": 4.000178337097168,
+ "step": 2844
+ },
+ {
+ "epoch": 39.51703056768559,
+ "grad_norm": 0.01520642638206482,
+ "learning_rate": 0.0006,
+ "loss": 4.218410968780518,
+ "step": 2845
+ },
+ {
+ "epoch": 39.531004366812226,
+ "grad_norm": 0.015305949375033379,
+ "learning_rate": 0.0006,
+ "loss": 4.025891304016113,
+ "step": 2846
+ },
+ {
+ "epoch": 39.544978165938865,
+ "grad_norm": 0.015461564064025879,
+ "learning_rate": 0.0006,
+ "loss": 4.100950717926025,
+ "step": 2847
+ },
+ {
+ "epoch": 39.5589519650655,
+ "grad_norm": 0.01712625101208687,
+ "learning_rate": 0.0006,
+ "loss": 4.0620856285095215,
+ "step": 2848
+ },
+ {
+ "epoch": 39.57292576419214,
+ "grad_norm": 0.017133301123976707,
+ "learning_rate": 0.0006,
+ "loss": 4.097138404846191,
+ "step": 2849
+ },
+ {
+ "epoch": 39.58689956331878,
+ "grad_norm": 0.01683625392615795,
+ "learning_rate": 0.0006,
+ "loss": 4.084364891052246,
+ "step": 2850
+ },
+ {
+ "epoch": 39.60087336244541,
+ "grad_norm": 0.0165710486471653,
+ "learning_rate": 0.0006,
+ "loss": 4.042361259460449,
+ "step": 2851
+ },
+ {
+ "epoch": 39.61484716157205,
+ "grad_norm": 0.01660393364727497,
+ "learning_rate": 0.0006,
+ "loss": 4.1395487785339355,
+ "step": 2852
+ },
+ {
+ "epoch": 39.62882096069869,
+ "grad_norm": 0.017025845125317574,
+ "learning_rate": 0.0006,
+ "loss": 4.1207075119018555,
+ "step": 2853
+ },
+ {
+ "epoch": 39.64279475982533,
+ "grad_norm": 0.01660497486591339,
+ "learning_rate": 0.0006,
+ "loss": 4.103846073150635,
+ "step": 2854
+ },
+ {
+ "epoch": 39.65676855895197,
+ "grad_norm": 0.01607946865260601,
+ "learning_rate": 0.0006,
+ "loss": 4.189016342163086,
+ "step": 2855
+ },
+ {
+ "epoch": 39.670742358078606,
+ "grad_norm": 0.01614505983889103,
+ "learning_rate": 0.0006,
+ "loss": 4.079860687255859,
+ "step": 2856
+ },
+ {
+ "epoch": 39.68471615720524,
+ "grad_norm": 0.017085086554288864,
+ "learning_rate": 0.0006,
+ "loss": 3.9774293899536133,
+ "step": 2857
+ },
+ {
+ "epoch": 39.698689956331876,
+ "grad_norm": 0.01816520094871521,
+ "learning_rate": 0.0006,
+ "loss": 4.113666534423828,
+ "step": 2858
+ },
+ {
+ "epoch": 39.712663755458514,
+ "grad_norm": 0.016597526147961617,
+ "learning_rate": 0.0006,
+ "loss": 4.068242073059082,
+ "step": 2859
+ },
+ {
+ "epoch": 39.72663755458515,
+ "grad_norm": 0.01672235131263733,
+ "learning_rate": 0.0006,
+ "loss": 4.024444580078125,
+ "step": 2860
+ },
+ {
+ "epoch": 39.74061135371179,
+ "grad_norm": 0.020093556493520737,
+ "learning_rate": 0.0006,
+ "loss": 4.064124584197998,
+ "step": 2861
+ },
+ {
+ "epoch": 39.75458515283843,
+ "grad_norm": 0.023141304031014442,
+ "learning_rate": 0.0006,
+ "loss": 4.130439758300781,
+ "step": 2862
+ },
+ {
+ "epoch": 39.76855895196506,
+ "grad_norm": 0.02289372682571411,
+ "learning_rate": 0.0006,
+ "loss": 4.14794921875,
+ "step": 2863
+ },
+ {
+ "epoch": 39.7825327510917,
+ "grad_norm": 0.01919020712375641,
+ "learning_rate": 0.0006,
+ "loss": 4.123961925506592,
+ "step": 2864
+ },
+ {
+ "epoch": 39.79650655021834,
+ "grad_norm": 0.016932329162955284,
+ "learning_rate": 0.0006,
+ "loss": 4.0944132804870605,
+ "step": 2865
+ },
+ {
+ "epoch": 39.81048034934498,
+ "grad_norm": 0.016644535586237907,
+ "learning_rate": 0.0006,
+ "loss": 4.124269962310791,
+ "step": 2866
+ },
+ {
+ "epoch": 39.82445414847162,
+ "grad_norm": 0.016302216798067093,
+ "learning_rate": 0.0006,
+ "loss": 4.0267205238342285,
+ "step": 2867
+ },
+ {
+ "epoch": 39.838427947598255,
+ "grad_norm": 0.01673300378024578,
+ "learning_rate": 0.0006,
+ "loss": 4.10650634765625,
+ "step": 2868
+ },
+ {
+ "epoch": 39.852401746724894,
+ "grad_norm": 0.016581356525421143,
+ "learning_rate": 0.0006,
+ "loss": 4.1547956466674805,
+ "step": 2869
+ },
+ {
+ "epoch": 39.866375545851525,
+ "grad_norm": 0.01691797934472561,
+ "learning_rate": 0.0006,
+ "loss": 4.056936740875244,
+ "step": 2870
+ },
+ {
+ "epoch": 39.880349344978164,
+ "grad_norm": 0.01596726104617119,
+ "learning_rate": 0.0006,
+ "loss": 3.9891414642333984,
+ "step": 2871
+ },
+ {
+ "epoch": 39.8943231441048,
+ "grad_norm": 0.015615091659128666,
+ "learning_rate": 0.0006,
+ "loss": 4.137948036193848,
+ "step": 2872
+ },
+ {
+ "epoch": 39.90829694323144,
+ "grad_norm": 0.015703413635492325,
+ "learning_rate": 0.0006,
+ "loss": 3.9844930171966553,
+ "step": 2873
+ },
+ {
+ "epoch": 39.92227074235808,
+ "grad_norm": 0.015453094616532326,
+ "learning_rate": 0.0006,
+ "loss": 4.098443508148193,
+ "step": 2874
+ },
+ {
+ "epoch": 39.93624454148472,
+ "grad_norm": 0.01449158601462841,
+ "learning_rate": 0.0006,
+ "loss": 4.090852737426758,
+ "step": 2875
+ },
+ {
+ "epoch": 39.95021834061135,
+ "grad_norm": 0.01537515688687563,
+ "learning_rate": 0.0006,
+ "loss": 4.239612579345703,
+ "step": 2876
+ },
+ {
+ "epoch": 39.96419213973799,
+ "grad_norm": 0.015989921987056732,
+ "learning_rate": 0.0006,
+ "loss": 4.142073154449463,
+ "step": 2877
+ },
+ {
+ "epoch": 39.97816593886463,
+ "grad_norm": 0.015985611826181412,
+ "learning_rate": 0.0006,
+ "loss": 4.100076675415039,
+ "step": 2878
+ },
+ {
+ "epoch": 39.992139737991266,
+ "grad_norm": 0.015260833315551281,
+ "learning_rate": 0.0006,
+ "loss": 4.068576335906982,
+ "step": 2879
+ },
+ {
+ "epoch": 40.0,
+ "grad_norm": 0.018047941848635674,
+ "learning_rate": 0.0006,
+ "loss": 4.005878925323486,
+ "step": 2880
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 28800,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 400,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": true,
+ "should_log": false,
+ "should_save": false,
+ "should_training_stop": false
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 1.2278453445328896e+18,
+ "train_batch_size": 8,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2880/training_args.bin b/runs/l2r50-i2-fulle-lm/checkpoint-2880/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..cad232831c222cc7f8b938b8bbe42f61e7d89c3a
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-2880/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:f7c40e3bfa2f5ac1450ca1243e3f2a895d44a06ed5995e6aefd97f78d229d8c8
+size 4792
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-3168/chat_template.jinja b/runs/l2r50-i2-fulle-lm/checkpoint-3168/chat_template.jinja
new file mode 100644
index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-3168/chat_template.jinja
@@ -0,0 +1,85 @@
+{%- if tools %}
+ {{- '<|im_start|>system\n' }}
+ {%- if messages[0].role == 'system' %}
+ {{- messages[0].content + '\n\n' }}
+ {%- endif %}
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }}
+ {%- for tool in tools %}
+ {{- "\n" }}
+ {{- tool | tojson }}
+ {%- endfor %}
+ {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }}
+{%- else %}
+ {%- if messages[0].role == 'system' %}
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
+ {%- endif %}
+{%- endif %}
+{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
+{%- for message in messages[::-1] %}
+ {%- set index = (messages|length - 1) - loop.index0 %}
+ {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %}
+ {%- set ns.multi_step_tool = false %}
+ {%- set ns.last_query_index = index %}
+ {%- endif %}
+{%- endfor %}
+{%- for message in messages %}
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
+ {%- elif message.role == "assistant" %}
+ {%- set content = message.content %}
+ {%- set reasoning_content = '' %}
+ {%- if message.reasoning_content is defined and message.reasoning_content is not none %}
+ {%- set reasoning_content = message.reasoning_content %}
+ {%- else %}
+ {%- if '' in message.content %}
+ {%- set content = message.content.split('')[-1].lstrip('\n') %}
+ {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %}
+ {%- endif %}
+ {%- endif %}
+ {%- if loop.index0 > ns.last_query_index %}
+ {%- if loop.last or (not loop.last and reasoning_content) %}
+ {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- if message.tool_calls %}
+ {%- for tool_call in message.tool_calls %}
+ {%- if (loop.first and content) or (not loop.first) %}
+ {{- '\n' }}
+ {%- endif %}
+ {%- if tool_call.function %}
+ {%- set tool_call = tool_call.function %}
+ {%- endif %}
+ {{- '\n{"name": "' }}
+ {{- tool_call.name }}
+ {{- '", "arguments": ' }}
+ {%- if tool_call.arguments is string %}
+ {{- tool_call.arguments }}
+ {%- else %}
+ {{- tool_call.arguments | tojson }}
+ {%- endif %}
+ {{- '}\n' }}
+ {%- endfor %}
+ {%- endif %}
+ {{- '<|im_end|>\n' }}
+ {%- elif message.role == "tool" %}
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
+ {{- '<|im_start|>user' }}
+ {%- endif %}
+ {{- '\n\n' }}
+ {{- message.content }}
+ {{- '\n' }}
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
+ {{- '<|im_end|>\n' }}
+ {%- endif %}
+ {%- endif %}
+{%- endfor %}
+{%- if add_generation_prompt %}
+ {{- '<|im_start|>assistant\n' }}
+ {%- if enable_thinking is defined and enable_thinking is false %}
+ {{- '\n\n\n\n' }}
+ {%- endif %}
+{%- endif %}
\ No newline at end of file
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-3168/config.json b/runs/l2r50-i2-fulle-lm/checkpoint-3168/config.json
new file mode 100644
index 0000000000000000000000000000000000000000..22f250953c6ce8fc91ee4411856dd73041a415e2
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-3168/config.json
@@ -0,0 +1,32 @@
+{
+ "architectures": [
+ "LlamaForCausalLM"
+ ],
+ "attention_bias": false,
+ "attention_dropout": 0.0,
+ "bos_token_id": null,
+ "dtype": "float32",
+ "eos_token_id": 151645,
+ "head_dim": 128,
+ "hidden_act": "silu",
+ "hidden_size": 512,
+ "initializer_range": 0.02,
+ "intermediate_size": 1536,
+ "max_position_embeddings": 2048,
+ "mlp_bias": false,
+ "model_type": "llama",
+ "num_attention_heads": 4,
+ "num_hidden_layers": 20,
+ "num_key_value_heads": 4,
+ "pad_token_id": 151645,
+ "pretraining_tp": 1,
+ "rms_norm_eps": 1e-06,
+ "rope_parameters": {
+ "rope_theta": 10000.0,
+ "rope_type": "default"
+ },
+ "tie_word_embeddings": true,
+ "transformers_version": "5.5.0",
+ "use_cache": false,
+ "vocab_size": 151673
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-3168/generation_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-3168/generation_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-3168/generation_config.json
@@ -0,0 +1,11 @@
+{
+ "_from_model_config": true,
+ "eos_token_id": [
+ 151645
+ ],
+ "output_attentions": false,
+ "output_hidden_states": false,
+ "pad_token_id": 151645,
+ "transformers_version": "5.5.0",
+ "use_cache": true
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-3168/model.safetensors b/runs/l2r50-i2-fulle-lm/checkpoint-3168/model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..0dcfa3a30b5fa6752cd1e6f79057359c369bbfc6
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-3168/model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:e6e814bb9e83b141a3d571b77177fe21003a41a1673ed9f2cf38a18db57f9b1b
+size 583360328
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-3168/optimizer.pt b/runs/l2r50-i2-fulle-lm/checkpoint-3168/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..5b2fa2f340cfdfa488848d3eaa140c5a4a1d51ee
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-3168/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:23dcb35da4d7c749c248d6d94637c0c19abadf9e3a4fd88122c9935ae5798d5d
+size 1166833530
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-3168/rng_state_0.pth b/runs/l2r50-i2-fulle-lm/checkpoint-3168/rng_state_0.pth
new file mode 100644
index 0000000000000000000000000000000000000000..5cead0762b939d4cfee78cbe3a60d54daa25efb0
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-3168/rng_state_0.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:3e593743986927a9cd1648c0e4bd6c7f8a21b611ad69fcae341973d2d7515960
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-3168/rng_state_1.pth b/runs/l2r50-i2-fulle-lm/checkpoint-3168/rng_state_1.pth
new file mode 100644
index 0000000000000000000000000000000000000000..77b5bcfd10229dee27d6ac33a93e7b8364000f70
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-3168/rng_state_1.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:6b03a37a26d716a393d503dc387470854bc62410e93ec96de951f0aff2c7d4cc
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-3168/rng_state_2.pth b/runs/l2r50-i2-fulle-lm/checkpoint-3168/rng_state_2.pth
new file mode 100644
index 0000000000000000000000000000000000000000..71f30120f7b4f413225d384ebd49b9888ae690a0
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-3168/rng_state_2.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:3dbf5cf5bc8065e48c4cdfbb7805e10a738b6f2e0271c2419220921c94cc9122
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-3168/rng_state_3.pth b/runs/l2r50-i2-fulle-lm/checkpoint-3168/rng_state_3.pth
new file mode 100644
index 0000000000000000000000000000000000000000..8f5ccdd68668ff317221c6f2f95b7735455ecd05
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-3168/rng_state_3.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:7e782ad710797b96d3abdb28995d2a41bd12fdcc553cfa86a98bfd43032b0b4b
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-3168/scheduler.pt b/runs/l2r50-i2-fulle-lm/checkpoint-3168/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..197051978d6c74fdad6b7032be258e0e7820d474
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-3168/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:8e78dacd009cacb02d7e6aed95e856956ae6e4311674c02b7eda2409a3c3c392
+size 1064
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-3168/tokenizer.json b/runs/l2r50-i2-fulle-lm/checkpoint-3168/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..53786016f070e11184813f3c83101dc5e93e392e
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-3168/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:c3c63fda1646a555448e8f03662542e04b11d103721319e4d9b267d7da03ceb8
+size 11424463
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-3168/tokenizer_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-3168/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..98e2fb9b2cc54785dc9d871b8e503f8b8f2defe7
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-3168/tokenizer_config.json
@@ -0,0 +1,20 @@
+{
+ "add_prefix_space": false,
+ "backend": "tokenizers",
+ "bos_token": null,
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "<|im_end|>",
+ "errors": "replace",
+ "extra_special_tokens": [
+ "<|l2r_pred|>",
+ "<|r2l_pred|>",
+ "<|next_1_pred|>",
+ "<|next_2_pred|>"
+ ],
+ "is_local": false,
+ "model_max_length": 131072,
+ "pad_token": "<|im_end|>",
+ "split_special_tokens": false,
+ "tokenizer_class": "Qwen2Tokenizer",
+ "unk_token": null
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-3168/trainer_state.json b/runs/l2r50-i2-fulle-lm/checkpoint-3168/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..23ae523aaa95147509a158856f3bec083db76ac5
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-3168/trainer_state.json
@@ -0,0 +1,22554 @@
+{
+ "best_global_step": null,
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 44.0,
+ "eval_steps": 500,
+ "global_step": 3168,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.013973799126637555,
+ "grad_norm": 0.1739506870508194,
+ "learning_rate": 0.0,
+ "loss": 11.99299430847168,
+ "step": 1
+ },
+ {
+ "epoch": 0.02794759825327511,
+ "grad_norm": 0.17216597497463226,
+ "learning_rate": 5.999999999999999e-06,
+ "loss": 11.988704681396484,
+ "step": 2
+ },
+ {
+ "epoch": 0.04192139737991266,
+ "grad_norm": 0.17459715902805328,
+ "learning_rate": 1.1999999999999999e-05,
+ "loss": 11.945581436157227,
+ "step": 3
+ },
+ {
+ "epoch": 0.05589519650655022,
+ "grad_norm": 0.17577414214611053,
+ "learning_rate": 1.7999999999999997e-05,
+ "loss": 11.859772682189941,
+ "step": 4
+ },
+ {
+ "epoch": 0.06986899563318777,
+ "grad_norm": 0.16163651645183563,
+ "learning_rate": 2.3999999999999997e-05,
+ "loss": 11.744415283203125,
+ "step": 5
+ },
+ {
+ "epoch": 0.08384279475982533,
+ "grad_norm": 0.14409932494163513,
+ "learning_rate": 2.9999999999999997e-05,
+ "loss": 11.618099212646484,
+ "step": 6
+ },
+ {
+ "epoch": 0.09781659388646288,
+ "grad_norm": 0.13079848885536194,
+ "learning_rate": 3.5999999999999994e-05,
+ "loss": 11.505935668945312,
+ "step": 7
+ },
+ {
+ "epoch": 0.11179039301310044,
+ "grad_norm": 0.11964613944292068,
+ "learning_rate": 4.2e-05,
+ "loss": 11.416532516479492,
+ "step": 8
+ },
+ {
+ "epoch": 0.125764192139738,
+ "grad_norm": 0.11345142871141434,
+ "learning_rate": 4.7999999999999994e-05,
+ "loss": 11.33206558227539,
+ "step": 9
+ },
+ {
+ "epoch": 0.13973799126637554,
+ "grad_norm": 0.11033952981233597,
+ "learning_rate": 5.399999999999999e-05,
+ "loss": 11.259052276611328,
+ "step": 10
+ },
+ {
+ "epoch": 0.1537117903930131,
+ "grad_norm": 0.10776489973068237,
+ "learning_rate": 5.9999999999999995e-05,
+ "loss": 11.203384399414062,
+ "step": 11
+ },
+ {
+ "epoch": 0.16768558951965065,
+ "grad_norm": 0.10721946507692337,
+ "learning_rate": 6.599999999999999e-05,
+ "loss": 11.153295516967773,
+ "step": 12
+ },
+ {
+ "epoch": 0.18165938864628822,
+ "grad_norm": 0.10673430562019348,
+ "learning_rate": 7.199999999999999e-05,
+ "loss": 11.110200881958008,
+ "step": 13
+ },
+ {
+ "epoch": 0.19563318777292577,
+ "grad_norm": 0.10630254447460175,
+ "learning_rate": 7.8e-05,
+ "loss": 11.067477226257324,
+ "step": 14
+ },
+ {
+ "epoch": 0.2096069868995633,
+ "grad_norm": 0.10631615668535233,
+ "learning_rate": 8.4e-05,
+ "loss": 11.022941589355469,
+ "step": 15
+ },
+ {
+ "epoch": 0.22358078602620088,
+ "grad_norm": 0.10609443485736847,
+ "learning_rate": 8.999999999999999e-05,
+ "loss": 10.970746040344238,
+ "step": 16
+ },
+ {
+ "epoch": 0.23755458515283842,
+ "grad_norm": 0.10648232698440552,
+ "learning_rate": 9.599999999999999e-05,
+ "loss": 10.91267204284668,
+ "step": 17
+ },
+ {
+ "epoch": 0.251528384279476,
+ "grad_norm": 0.10584589838981628,
+ "learning_rate": 0.000102,
+ "loss": 10.854547500610352,
+ "step": 18
+ },
+ {
+ "epoch": 0.26550218340611353,
+ "grad_norm": 0.10591720044612885,
+ "learning_rate": 0.00010799999999999998,
+ "loss": 10.788917541503906,
+ "step": 19
+ },
+ {
+ "epoch": 0.2794759825327511,
+ "grad_norm": 0.10611672699451447,
+ "learning_rate": 0.00011399999999999999,
+ "loss": 10.717556953430176,
+ "step": 20
+ },
+ {
+ "epoch": 0.2934497816593886,
+ "grad_norm": 0.10455359518527985,
+ "learning_rate": 0.00011999999999999999,
+ "loss": 10.650186538696289,
+ "step": 21
+ },
+ {
+ "epoch": 0.3074235807860262,
+ "grad_norm": 0.1047525629401207,
+ "learning_rate": 0.00012599999999999997,
+ "loss": 10.56808090209961,
+ "step": 22
+ },
+ {
+ "epoch": 0.32139737991266376,
+ "grad_norm": 0.10511605441570282,
+ "learning_rate": 0.00013199999999999998,
+ "loss": 10.484028816223145,
+ "step": 23
+ },
+ {
+ "epoch": 0.3353711790393013,
+ "grad_norm": 0.10467177629470825,
+ "learning_rate": 0.000138,
+ "loss": 10.399947166442871,
+ "step": 24
+ },
+ {
+ "epoch": 0.34934497816593885,
+ "grad_norm": 0.10432136058807373,
+ "learning_rate": 0.00014399999999999998,
+ "loss": 10.312714576721191,
+ "step": 25
+ },
+ {
+ "epoch": 0.36331877729257644,
+ "grad_norm": 0.10527262091636658,
+ "learning_rate": 0.00015,
+ "loss": 10.212261199951172,
+ "step": 26
+ },
+ {
+ "epoch": 0.377292576419214,
+ "grad_norm": 0.103610560297966,
+ "learning_rate": 0.000156,
+ "loss": 10.132034301757812,
+ "step": 27
+ },
+ {
+ "epoch": 0.39126637554585153,
+ "grad_norm": 0.10553059726953506,
+ "learning_rate": 0.000162,
+ "loss": 10.020434379577637,
+ "step": 28
+ },
+ {
+ "epoch": 0.4052401746724891,
+ "grad_norm": 0.1037299782037735,
+ "learning_rate": 0.000168,
+ "loss": 9.934466361999512,
+ "step": 29
+ },
+ {
+ "epoch": 0.4192139737991266,
+ "grad_norm": 0.1042739674448967,
+ "learning_rate": 0.00017399999999999997,
+ "loss": 9.822539329528809,
+ "step": 30
+ },
+ {
+ "epoch": 0.4331877729257642,
+ "grad_norm": 0.10300865024328232,
+ "learning_rate": 0.00017999999999999998,
+ "loss": 9.734994888305664,
+ "step": 31
+ },
+ {
+ "epoch": 0.44716157205240176,
+ "grad_norm": 0.10291654616594315,
+ "learning_rate": 0.000186,
+ "loss": 9.624505996704102,
+ "step": 32
+ },
+ {
+ "epoch": 0.4611353711790393,
+ "grad_norm": 0.10210851579904556,
+ "learning_rate": 0.00019199999999999998,
+ "loss": 9.518205642700195,
+ "step": 33
+ },
+ {
+ "epoch": 0.47510917030567684,
+ "grad_norm": 0.10222529619932175,
+ "learning_rate": 0.000198,
+ "loss": 9.412222862243652,
+ "step": 34
+ },
+ {
+ "epoch": 0.4890829694323144,
+ "grad_norm": 0.09992019087076187,
+ "learning_rate": 0.000204,
+ "loss": 9.326566696166992,
+ "step": 35
+ },
+ {
+ "epoch": 0.503056768558952,
+ "grad_norm": 0.09931084513664246,
+ "learning_rate": 0.00020999999999999998,
+ "loss": 9.210111618041992,
+ "step": 36
+ },
+ {
+ "epoch": 0.5170305676855895,
+ "grad_norm": 0.09928256273269653,
+ "learning_rate": 0.00021599999999999996,
+ "loss": 9.100493431091309,
+ "step": 37
+ },
+ {
+ "epoch": 0.5310043668122271,
+ "grad_norm": 0.09843816608190536,
+ "learning_rate": 0.00022199999999999998,
+ "loss": 8.996543884277344,
+ "step": 38
+ },
+ {
+ "epoch": 0.5449781659388646,
+ "grad_norm": 0.09825374186038971,
+ "learning_rate": 0.00022799999999999999,
+ "loss": 8.888991355895996,
+ "step": 39
+ },
+ {
+ "epoch": 0.5589519650655022,
+ "grad_norm": 0.09602385759353638,
+ "learning_rate": 0.000234,
+ "loss": 8.802907943725586,
+ "step": 40
+ },
+ {
+ "epoch": 0.5729257641921397,
+ "grad_norm": 0.09414064139127731,
+ "learning_rate": 0.00023999999999999998,
+ "loss": 8.715423583984375,
+ "step": 41
+ },
+ {
+ "epoch": 0.5868995633187772,
+ "grad_norm": 0.09273893386125565,
+ "learning_rate": 0.00024599999999999996,
+ "loss": 8.61374282836914,
+ "step": 42
+ },
+ {
+ "epoch": 0.6008733624454149,
+ "grad_norm": 0.09079599380493164,
+ "learning_rate": 0.00025199999999999995,
+ "loss": 8.502904891967773,
+ "step": 43
+ },
+ {
+ "epoch": 0.6148471615720524,
+ "grad_norm": 0.08605825901031494,
+ "learning_rate": 0.000258,
+ "loss": 8.453008651733398,
+ "step": 44
+ },
+ {
+ "epoch": 0.62882096069869,
+ "grad_norm": 0.08338996022939682,
+ "learning_rate": 0.00026399999999999997,
+ "loss": 8.365163803100586,
+ "step": 45
+ },
+ {
+ "epoch": 0.6427947598253275,
+ "grad_norm": 0.08150867372751236,
+ "learning_rate": 0.00027,
+ "loss": 8.285372734069824,
+ "step": 46
+ },
+ {
+ "epoch": 0.6567685589519651,
+ "grad_norm": 0.07794807851314545,
+ "learning_rate": 0.000276,
+ "loss": 8.185811996459961,
+ "step": 47
+ },
+ {
+ "epoch": 0.6707423580786026,
+ "grad_norm": 0.07592720538377762,
+ "learning_rate": 0.00028199999999999997,
+ "loss": 8.098482131958008,
+ "step": 48
+ },
+ {
+ "epoch": 0.6847161572052402,
+ "grad_norm": 0.0709897130727768,
+ "learning_rate": 0.00028799999999999995,
+ "loss": 8.042091369628906,
+ "step": 49
+ },
+ {
+ "epoch": 0.6986899563318777,
+ "grad_norm": 0.06582305580377579,
+ "learning_rate": 0.000294,
+ "loss": 7.972844123840332,
+ "step": 50
+ },
+ {
+ "epoch": 0.7126637554585152,
+ "grad_norm": 0.059452351182699203,
+ "learning_rate": 0.0003,
+ "loss": 7.924579620361328,
+ "step": 51
+ },
+ {
+ "epoch": 0.7266375545851529,
+ "grad_norm": 0.052426718175411224,
+ "learning_rate": 0.00030599999999999996,
+ "loss": 7.876336097717285,
+ "step": 52
+ },
+ {
+ "epoch": 0.7406113537117904,
+ "grad_norm": 0.04768150672316551,
+ "learning_rate": 0.000312,
+ "loss": 7.8102707862854,
+ "step": 53
+ },
+ {
+ "epoch": 0.754585152838428,
+ "grad_norm": 0.03970075398683548,
+ "learning_rate": 0.000318,
+ "loss": 7.791913986206055,
+ "step": 54
+ },
+ {
+ "epoch": 0.7685589519650655,
+ "grad_norm": 0.03559752553701401,
+ "learning_rate": 0.000324,
+ "loss": 7.756960868835449,
+ "step": 55
+ },
+ {
+ "epoch": 0.7825327510917031,
+ "grad_norm": 0.029321353882551193,
+ "learning_rate": 0.00033,
+ "loss": 7.724878787994385,
+ "step": 56
+ },
+ {
+ "epoch": 0.7965065502183406,
+ "grad_norm": 0.023485427722334862,
+ "learning_rate": 0.000336,
+ "loss": 7.72553014755249,
+ "step": 57
+ },
+ {
+ "epoch": 0.8104803493449781,
+ "grad_norm": 0.019319474697113037,
+ "learning_rate": 0.00034199999999999996,
+ "loss": 7.709493160247803,
+ "step": 58
+ },
+ {
+ "epoch": 0.8244541484716157,
+ "grad_norm": 0.016705691814422607,
+ "learning_rate": 0.00034799999999999995,
+ "loss": 7.6873297691345215,
+ "step": 59
+ },
+ {
+ "epoch": 0.8384279475982532,
+ "grad_norm": 0.017667734995484352,
+ "learning_rate": 0.00035399999999999993,
+ "loss": 7.677059173583984,
+ "step": 60
+ },
+ {
+ "epoch": 0.8524017467248908,
+ "grad_norm": 0.01671992614865303,
+ "learning_rate": 0.00035999999999999997,
+ "loss": 7.643512725830078,
+ "step": 61
+ },
+ {
+ "epoch": 0.8663755458515284,
+ "grad_norm": 0.012288172729313374,
+ "learning_rate": 0.00036599999999999995,
+ "loss": 7.658343315124512,
+ "step": 62
+ },
+ {
+ "epoch": 0.880349344978166,
+ "grad_norm": 0.012004776857793331,
+ "learning_rate": 0.000372,
+ "loss": 7.649759292602539,
+ "step": 63
+ },
+ {
+ "epoch": 0.8943231441048035,
+ "grad_norm": 0.012487445957958698,
+ "learning_rate": 0.00037799999999999997,
+ "loss": 7.631503105163574,
+ "step": 64
+ },
+ {
+ "epoch": 0.9082969432314411,
+ "grad_norm": 0.01489241048693657,
+ "learning_rate": 0.00038399999999999996,
+ "loss": 7.647110462188721,
+ "step": 65
+ },
+ {
+ "epoch": 0.9222707423580786,
+ "grad_norm": 0.01428980939090252,
+ "learning_rate": 0.00039,
+ "loss": 7.636983871459961,
+ "step": 66
+ },
+ {
+ "epoch": 0.9362445414847161,
+ "grad_norm": 0.010711144655942917,
+ "learning_rate": 0.000396,
+ "loss": 7.621130466461182,
+ "step": 67
+ },
+ {
+ "epoch": 0.9502183406113537,
+ "grad_norm": 0.010178886353969574,
+ "learning_rate": 0.000402,
+ "loss": 7.593924522399902,
+ "step": 68
+ },
+ {
+ "epoch": 0.9641921397379912,
+ "grad_norm": 0.010604874230921268,
+ "learning_rate": 0.000408,
+ "loss": 7.6015496253967285,
+ "step": 69
+ },
+ {
+ "epoch": 0.9781659388646288,
+ "grad_norm": 0.009751426056027412,
+ "learning_rate": 0.0004139999999999999,
+ "loss": 7.571964263916016,
+ "step": 70
+ },
+ {
+ "epoch": 0.9921397379912664,
+ "grad_norm": 0.009816481731832027,
+ "learning_rate": 0.00041999999999999996,
+ "loss": 7.57344913482666,
+ "step": 71
+ },
+ {
+ "epoch": 1.0,
+ "grad_norm": 0.008665637113153934,
+ "learning_rate": 0.00042599999999999995,
+ "loss": 7.579050064086914,
+ "step": 72
+ },
+ {
+ "epoch": 1.0,
+ "eval_loss": 7.581204891204834,
+ "eval_runtime": 59.0091,
+ "eval_samples_per_second": 41.383,
+ "eval_steps_per_second": 1.305,
+ "step": 72
+ },
+ {
+ "epoch": 1.0139737991266375,
+ "grad_norm": 0.009684201329946518,
+ "learning_rate": 0.00043199999999999993,
+ "loss": 7.568594932556152,
+ "step": 73
+ },
+ {
+ "epoch": 1.027947598253275,
+ "grad_norm": 0.025783710181713104,
+ "learning_rate": 0.00043799999999999997,
+ "loss": 7.578948497772217,
+ "step": 74
+ },
+ {
+ "epoch": 1.0419213973799126,
+ "grad_norm": 0.0327819399535656,
+ "learning_rate": 0.00044399999999999995,
+ "loss": 7.548173904418945,
+ "step": 75
+ },
+ {
+ "epoch": 1.0558951965065502,
+ "grad_norm": 0.021604498848319054,
+ "learning_rate": 0.00045,
+ "loss": 7.552953720092773,
+ "step": 76
+ },
+ {
+ "epoch": 1.0698689956331877,
+ "grad_norm": 0.010230105370283127,
+ "learning_rate": 0.00045599999999999997,
+ "loss": 7.5213212966918945,
+ "step": 77
+ },
+ {
+ "epoch": 1.0838427947598253,
+ "grad_norm": 0.023315589874982834,
+ "learning_rate": 0.00046199999999999995,
+ "loss": 7.501640319824219,
+ "step": 78
+ },
+ {
+ "epoch": 1.0978165938864628,
+ "grad_norm": 0.01109944935888052,
+ "learning_rate": 0.000468,
+ "loss": 7.539519786834717,
+ "step": 79
+ },
+ {
+ "epoch": 1.1117903930131003,
+ "grad_norm": 0.026093758642673492,
+ "learning_rate": 0.000474,
+ "loss": 7.525928974151611,
+ "step": 80
+ },
+ {
+ "epoch": 1.125764192139738,
+ "grad_norm": 0.03274895250797272,
+ "learning_rate": 0.00047999999999999996,
+ "loss": 7.535602569580078,
+ "step": 81
+ },
+ {
+ "epoch": 1.1397379912663754,
+ "grad_norm": 0.01543602254241705,
+ "learning_rate": 0.000486,
+ "loss": 7.513683795928955,
+ "step": 82
+ },
+ {
+ "epoch": 1.1537117903930132,
+ "grad_norm": 0.028377249836921692,
+ "learning_rate": 0.0004919999999999999,
+ "loss": 7.49384880065918,
+ "step": 83
+ },
+ {
+ "epoch": 1.1676855895196507,
+ "grad_norm": 0.04121481627225876,
+ "learning_rate": 0.000498,
+ "loss": 7.488447189331055,
+ "step": 84
+ },
+ {
+ "epoch": 1.1816593886462883,
+ "grad_norm": 0.013282302767038345,
+ "learning_rate": 0.0005039999999999999,
+ "loss": 7.463010787963867,
+ "step": 85
+ },
+ {
+ "epoch": 1.1956331877729258,
+ "grad_norm": 0.03200364485383034,
+ "learning_rate": 0.0005099999999999999,
+ "loss": 7.473515033721924,
+ "step": 86
+ },
+ {
+ "epoch": 1.2096069868995634,
+ "grad_norm": 0.020768029615283012,
+ "learning_rate": 0.000516,
+ "loss": 7.464688777923584,
+ "step": 87
+ },
+ {
+ "epoch": 1.223580786026201,
+ "grad_norm": 0.021715788170695305,
+ "learning_rate": 0.000522,
+ "loss": 7.4835028648376465,
+ "step": 88
+ },
+ {
+ "epoch": 1.2375545851528384,
+ "grad_norm": 0.026255754753947258,
+ "learning_rate": 0.0005279999999999999,
+ "loss": 7.443392753601074,
+ "step": 89
+ },
+ {
+ "epoch": 1.251528384279476,
+ "grad_norm": 0.022325951606035233,
+ "learning_rate": 0.000534,
+ "loss": 7.457373142242432,
+ "step": 90
+ },
+ {
+ "epoch": 1.2655021834061135,
+ "grad_norm": 0.009270387701690197,
+ "learning_rate": 0.00054,
+ "loss": 7.455986976623535,
+ "step": 91
+ },
+ {
+ "epoch": 1.279475982532751,
+ "grad_norm": 0.02369544468820095,
+ "learning_rate": 0.0005459999999999999,
+ "loss": 7.442216873168945,
+ "step": 92
+ },
+ {
+ "epoch": 1.2934497816593886,
+ "grad_norm": 0.008474506437778473,
+ "learning_rate": 0.000552,
+ "loss": 7.4294915199279785,
+ "step": 93
+ },
+ {
+ "epoch": 1.3074235807860262,
+ "grad_norm": 0.017043249681591988,
+ "learning_rate": 0.000558,
+ "loss": 7.406932830810547,
+ "step": 94
+ },
+ {
+ "epoch": 1.3213973799126637,
+ "grad_norm": 0.011813412420451641,
+ "learning_rate": 0.0005639999999999999,
+ "loss": 7.415900230407715,
+ "step": 95
+ },
+ {
+ "epoch": 1.3353711790393012,
+ "grad_norm": 0.01086883433163166,
+ "learning_rate": 0.00057,
+ "loss": 7.452998638153076,
+ "step": 96
+ },
+ {
+ "epoch": 1.3493449781659388,
+ "grad_norm": 0.014700951054692268,
+ "learning_rate": 0.0005759999999999999,
+ "loss": 7.410249710083008,
+ "step": 97
+ },
+ {
+ "epoch": 1.3633187772925766,
+ "grad_norm": 0.014512976631522179,
+ "learning_rate": 0.0005819999999999999,
+ "loss": 7.429132461547852,
+ "step": 98
+ },
+ {
+ "epoch": 1.3772925764192139,
+ "grad_norm": 0.012519920244812965,
+ "learning_rate": 0.000588,
+ "loss": 7.430817604064941,
+ "step": 99
+ },
+ {
+ "epoch": 1.3912663755458516,
+ "grad_norm": 0.007132664322853088,
+ "learning_rate": 0.0005939999999999999,
+ "loss": 7.405234336853027,
+ "step": 100
+ },
+ {
+ "epoch": 1.405240174672489,
+ "grad_norm": 0.019489388912916183,
+ "learning_rate": 0.0006,
+ "loss": 7.427791595458984,
+ "step": 101
+ },
+ {
+ "epoch": 1.4192139737991267,
+ "grad_norm": 0.027252664789557457,
+ "learning_rate": 0.0006,
+ "loss": 7.41761589050293,
+ "step": 102
+ },
+ {
+ "epoch": 1.4331877729257643,
+ "grad_norm": 0.02501637488603592,
+ "learning_rate": 0.0006,
+ "loss": 7.416735649108887,
+ "step": 103
+ },
+ {
+ "epoch": 1.4471615720524018,
+ "grad_norm": 0.00989772193133831,
+ "learning_rate": 0.0006,
+ "loss": 7.3756632804870605,
+ "step": 104
+ },
+ {
+ "epoch": 1.4611353711790394,
+ "grad_norm": 0.01622949168086052,
+ "learning_rate": 0.0006,
+ "loss": 7.400811672210693,
+ "step": 105
+ },
+ {
+ "epoch": 1.475109170305677,
+ "grad_norm": 0.012407763861119747,
+ "learning_rate": 0.0006,
+ "loss": 7.37675666809082,
+ "step": 106
+ },
+ {
+ "epoch": 1.4890829694323144,
+ "grad_norm": 0.010353565216064453,
+ "learning_rate": 0.0006,
+ "loss": 7.4013142585754395,
+ "step": 107
+ },
+ {
+ "epoch": 1.503056768558952,
+ "grad_norm": 0.0120553532615304,
+ "learning_rate": 0.0006,
+ "loss": 7.379873275756836,
+ "step": 108
+ },
+ {
+ "epoch": 1.5170305676855895,
+ "grad_norm": 0.01610225811600685,
+ "learning_rate": 0.0006,
+ "loss": 7.371631145477295,
+ "step": 109
+ },
+ {
+ "epoch": 1.531004366812227,
+ "grad_norm": 0.01159879844635725,
+ "learning_rate": 0.0006,
+ "loss": 7.389007568359375,
+ "step": 110
+ },
+ {
+ "epoch": 1.5449781659388646,
+ "grad_norm": 0.008013816550374031,
+ "learning_rate": 0.0006,
+ "loss": 7.395920753479004,
+ "step": 111
+ },
+ {
+ "epoch": 1.5589519650655022,
+ "grad_norm": 0.016334809362888336,
+ "learning_rate": 0.0006,
+ "loss": 7.393121719360352,
+ "step": 112
+ },
+ {
+ "epoch": 1.5729257641921397,
+ "grad_norm": 0.018821721896529198,
+ "learning_rate": 0.0006,
+ "loss": 7.367543697357178,
+ "step": 113
+ },
+ {
+ "epoch": 1.5868995633187772,
+ "grad_norm": 0.014088994823396206,
+ "learning_rate": 0.0006,
+ "loss": 7.385537624359131,
+ "step": 114
+ },
+ {
+ "epoch": 1.600873362445415,
+ "grad_norm": 0.005864050704985857,
+ "learning_rate": 0.0006,
+ "loss": 7.365346908569336,
+ "step": 115
+ },
+ {
+ "epoch": 1.6148471615720523,
+ "grad_norm": 0.011745430529117584,
+ "learning_rate": 0.0006,
+ "loss": 7.390227317810059,
+ "step": 116
+ },
+ {
+ "epoch": 1.62882096069869,
+ "grad_norm": 0.017205573618412018,
+ "learning_rate": 0.0006,
+ "loss": 7.3558549880981445,
+ "step": 117
+ },
+ {
+ "epoch": 1.6427947598253274,
+ "grad_norm": 0.021256346255540848,
+ "learning_rate": 0.0006,
+ "loss": 7.377346992492676,
+ "step": 118
+ },
+ {
+ "epoch": 1.6567685589519652,
+ "grad_norm": 0.019335204735398293,
+ "learning_rate": 0.0006,
+ "loss": 7.346194267272949,
+ "step": 119
+ },
+ {
+ "epoch": 1.6707423580786025,
+ "grad_norm": 0.008134003728628159,
+ "learning_rate": 0.0006,
+ "loss": 7.3063578605651855,
+ "step": 120
+ },
+ {
+ "epoch": 1.6847161572052403,
+ "grad_norm": 0.017613831907510757,
+ "learning_rate": 0.0006,
+ "loss": 7.338171005249023,
+ "step": 121
+ },
+ {
+ "epoch": 1.6986899563318776,
+ "grad_norm": 0.00890734139829874,
+ "learning_rate": 0.0006,
+ "loss": 7.355384826660156,
+ "step": 122
+ },
+ {
+ "epoch": 1.7126637554585153,
+ "grad_norm": 0.013448155485093594,
+ "learning_rate": 0.0006,
+ "loss": 7.344296455383301,
+ "step": 123
+ },
+ {
+ "epoch": 1.726637554585153,
+ "grad_norm": 0.009241633117198944,
+ "learning_rate": 0.0006,
+ "loss": 7.3454060554504395,
+ "step": 124
+ },
+ {
+ "epoch": 1.7406113537117904,
+ "grad_norm": 0.018561311066150665,
+ "learning_rate": 0.0006,
+ "loss": 7.32348108291626,
+ "step": 125
+ },
+ {
+ "epoch": 1.754585152838428,
+ "grad_norm": 0.025667704641819,
+ "learning_rate": 0.0006,
+ "loss": 7.356142520904541,
+ "step": 126
+ },
+ {
+ "epoch": 1.7685589519650655,
+ "grad_norm": 0.036232151091098785,
+ "learning_rate": 0.0006,
+ "loss": 7.363539695739746,
+ "step": 127
+ },
+ {
+ "epoch": 1.782532751091703,
+ "grad_norm": 0.059483714401721954,
+ "learning_rate": 0.0006,
+ "loss": 7.402331352233887,
+ "step": 128
+ },
+ {
+ "epoch": 1.7965065502183406,
+ "grad_norm": 0.05365009233355522,
+ "learning_rate": 0.0006,
+ "loss": 7.40457820892334,
+ "step": 129
+ },
+ {
+ "epoch": 1.8104803493449781,
+ "grad_norm": 0.02738415263593197,
+ "learning_rate": 0.0006,
+ "loss": 7.3756208419799805,
+ "step": 130
+ },
+ {
+ "epoch": 1.8244541484716157,
+ "grad_norm": 0.035017091780900955,
+ "learning_rate": 0.0006,
+ "loss": 7.379263877868652,
+ "step": 131
+ },
+ {
+ "epoch": 1.8384279475982532,
+ "grad_norm": 0.02819863148033619,
+ "learning_rate": 0.0006,
+ "loss": 7.327673435211182,
+ "step": 132
+ },
+ {
+ "epoch": 1.8524017467248908,
+ "grad_norm": 0.017754238098859787,
+ "learning_rate": 0.0006,
+ "loss": 7.333420276641846,
+ "step": 133
+ },
+ {
+ "epoch": 1.8663755458515285,
+ "grad_norm": 0.024685733020305634,
+ "learning_rate": 0.0006,
+ "loss": 7.34493350982666,
+ "step": 134
+ },
+ {
+ "epoch": 1.8803493449781659,
+ "grad_norm": 0.019426532089710236,
+ "learning_rate": 0.0006,
+ "loss": 7.328448295593262,
+ "step": 135
+ },
+ {
+ "epoch": 1.8943231441048036,
+ "grad_norm": 0.02106391079723835,
+ "learning_rate": 0.0006,
+ "loss": 7.327917098999023,
+ "step": 136
+ },
+ {
+ "epoch": 1.908296943231441,
+ "grad_norm": 0.02056879736483097,
+ "learning_rate": 0.0006,
+ "loss": 7.3311991691589355,
+ "step": 137
+ },
+ {
+ "epoch": 1.9222707423580787,
+ "grad_norm": 0.016865141689777374,
+ "learning_rate": 0.0006,
+ "loss": 7.312766075134277,
+ "step": 138
+ },
+ {
+ "epoch": 1.936244541484716,
+ "grad_norm": 0.017549855634570122,
+ "learning_rate": 0.0006,
+ "loss": 7.31191349029541,
+ "step": 139
+ },
+ {
+ "epoch": 1.9502183406113538,
+ "grad_norm": 0.015930157154798508,
+ "learning_rate": 0.0006,
+ "loss": 7.330901622772217,
+ "step": 140
+ },
+ {
+ "epoch": 1.9641921397379911,
+ "grad_norm": 0.013283558189868927,
+ "learning_rate": 0.0006,
+ "loss": 7.298318386077881,
+ "step": 141
+ },
+ {
+ "epoch": 1.9781659388646289,
+ "grad_norm": 0.012023469433188438,
+ "learning_rate": 0.0006,
+ "loss": 7.282337188720703,
+ "step": 142
+ },
+ {
+ "epoch": 1.9921397379912664,
+ "grad_norm": 0.01459511648863554,
+ "learning_rate": 0.0006,
+ "loss": 7.282924175262451,
+ "step": 143
+ },
+ {
+ "epoch": 2.0,
+ "grad_norm": 0.017474235966801643,
+ "learning_rate": 0.0006,
+ "loss": 7.254333972930908,
+ "step": 144
+ },
+ {
+ "epoch": 2.0,
+ "eval_loss": 7.318708896636963,
+ "eval_runtime": 59.6555,
+ "eval_samples_per_second": 40.935,
+ "eval_steps_per_second": 1.291,
+ "step": 144
+ },
+ {
+ "epoch": 2.0139737991266378,
+ "grad_norm": 0.019757507368922234,
+ "learning_rate": 0.0006,
+ "loss": 7.30288553237915,
+ "step": 145
+ },
+ {
+ "epoch": 2.027947598253275,
+ "grad_norm": 0.014571248553693295,
+ "learning_rate": 0.0006,
+ "loss": 7.274681568145752,
+ "step": 146
+ },
+ {
+ "epoch": 2.041921397379913,
+ "grad_norm": 0.015744198113679886,
+ "learning_rate": 0.0006,
+ "loss": 7.278632640838623,
+ "step": 147
+ },
+ {
+ "epoch": 2.05589519650655,
+ "grad_norm": 0.028048858046531677,
+ "learning_rate": 0.0006,
+ "loss": 7.26558780670166,
+ "step": 148
+ },
+ {
+ "epoch": 2.069868995633188,
+ "grad_norm": 0.06547307223081589,
+ "learning_rate": 0.0006,
+ "loss": 7.32567834854126,
+ "step": 149
+ },
+ {
+ "epoch": 2.0838427947598253,
+ "grad_norm": 0.05969763547182083,
+ "learning_rate": 0.0006,
+ "loss": 7.395836353302002,
+ "step": 150
+ },
+ {
+ "epoch": 2.097816593886463,
+ "grad_norm": 0.032574281096458435,
+ "learning_rate": 0.0006,
+ "loss": 7.318150997161865,
+ "step": 151
+ },
+ {
+ "epoch": 2.1117903930131003,
+ "grad_norm": 0.029452962800860405,
+ "learning_rate": 0.0006,
+ "loss": 7.308066368103027,
+ "step": 152
+ },
+ {
+ "epoch": 2.125764192139738,
+ "grad_norm": 0.03161991387605667,
+ "learning_rate": 0.0006,
+ "loss": 7.306290626525879,
+ "step": 153
+ },
+ {
+ "epoch": 2.1397379912663754,
+ "grad_norm": 0.027096295729279518,
+ "learning_rate": 0.0006,
+ "loss": 7.297796249389648,
+ "step": 154
+ },
+ {
+ "epoch": 2.153711790393013,
+ "grad_norm": 0.028657056391239166,
+ "learning_rate": 0.0006,
+ "loss": 7.317424774169922,
+ "step": 155
+ },
+ {
+ "epoch": 2.1676855895196505,
+ "grad_norm": 0.02046995982527733,
+ "learning_rate": 0.0006,
+ "loss": 7.297986030578613,
+ "step": 156
+ },
+ {
+ "epoch": 2.1816593886462883,
+ "grad_norm": 0.02220369316637516,
+ "learning_rate": 0.0006,
+ "loss": 7.268237590789795,
+ "step": 157
+ },
+ {
+ "epoch": 2.1956331877729256,
+ "grad_norm": 0.018381357192993164,
+ "learning_rate": 0.0006,
+ "loss": 7.259415626525879,
+ "step": 158
+ },
+ {
+ "epoch": 2.2096069868995634,
+ "grad_norm": 0.0204413291066885,
+ "learning_rate": 0.0006,
+ "loss": 7.258173942565918,
+ "step": 159
+ },
+ {
+ "epoch": 2.2235807860262007,
+ "grad_norm": 0.015874288976192474,
+ "learning_rate": 0.0006,
+ "loss": 7.281834602355957,
+ "step": 160
+ },
+ {
+ "epoch": 2.2375545851528384,
+ "grad_norm": 0.01634068600833416,
+ "learning_rate": 0.0006,
+ "loss": 7.2469868659973145,
+ "step": 161
+ },
+ {
+ "epoch": 2.251528384279476,
+ "grad_norm": 0.01728747971355915,
+ "learning_rate": 0.0006,
+ "loss": 7.280995845794678,
+ "step": 162
+ },
+ {
+ "epoch": 2.2655021834061135,
+ "grad_norm": 0.016341188922524452,
+ "learning_rate": 0.0006,
+ "loss": 7.273357391357422,
+ "step": 163
+ },
+ {
+ "epoch": 2.279475982532751,
+ "grad_norm": 0.016125807538628578,
+ "learning_rate": 0.0006,
+ "loss": 7.273510932922363,
+ "step": 164
+ },
+ {
+ "epoch": 2.2934497816593886,
+ "grad_norm": 0.01629287749528885,
+ "learning_rate": 0.0006,
+ "loss": 7.267029762268066,
+ "step": 165
+ },
+ {
+ "epoch": 2.3074235807860264,
+ "grad_norm": 0.015296131372451782,
+ "learning_rate": 0.0006,
+ "loss": 7.211989402770996,
+ "step": 166
+ },
+ {
+ "epoch": 2.3213973799126637,
+ "grad_norm": 0.01709570363163948,
+ "learning_rate": 0.0006,
+ "loss": 7.242851734161377,
+ "step": 167
+ },
+ {
+ "epoch": 2.3353711790393015,
+ "grad_norm": 0.012552388943731785,
+ "learning_rate": 0.0006,
+ "loss": 7.241696834564209,
+ "step": 168
+ },
+ {
+ "epoch": 2.349344978165939,
+ "grad_norm": 0.014571291394531727,
+ "learning_rate": 0.0006,
+ "loss": 7.270681381225586,
+ "step": 169
+ },
+ {
+ "epoch": 2.3633187772925766,
+ "grad_norm": 0.013389437459409237,
+ "learning_rate": 0.0006,
+ "loss": 7.191600799560547,
+ "step": 170
+ },
+ {
+ "epoch": 2.377292576419214,
+ "grad_norm": 0.011895314790308475,
+ "learning_rate": 0.0006,
+ "loss": 7.213610649108887,
+ "step": 171
+ },
+ {
+ "epoch": 2.3912663755458516,
+ "grad_norm": 0.01345642190426588,
+ "learning_rate": 0.0006,
+ "loss": 7.193787574768066,
+ "step": 172
+ },
+ {
+ "epoch": 2.405240174672489,
+ "grad_norm": 0.0091794328764081,
+ "learning_rate": 0.0006,
+ "loss": 7.233364105224609,
+ "step": 173
+ },
+ {
+ "epoch": 2.4192139737991267,
+ "grad_norm": 0.013767086900770664,
+ "learning_rate": 0.0006,
+ "loss": 7.217272758483887,
+ "step": 174
+ },
+ {
+ "epoch": 2.433187772925764,
+ "grad_norm": 0.015134960412979126,
+ "learning_rate": 0.0006,
+ "loss": 7.192791938781738,
+ "step": 175
+ },
+ {
+ "epoch": 2.447161572052402,
+ "grad_norm": 0.01790648326277733,
+ "learning_rate": 0.0006,
+ "loss": 7.229467391967773,
+ "step": 176
+ },
+ {
+ "epoch": 2.461135371179039,
+ "grad_norm": 0.015362377278506756,
+ "learning_rate": 0.0006,
+ "loss": 7.180346488952637,
+ "step": 177
+ },
+ {
+ "epoch": 2.475109170305677,
+ "grad_norm": 0.010505498386919498,
+ "learning_rate": 0.0006,
+ "loss": 7.188833236694336,
+ "step": 178
+ },
+ {
+ "epoch": 2.489082969432314,
+ "grad_norm": 0.009129747748374939,
+ "learning_rate": 0.0006,
+ "loss": 7.203153133392334,
+ "step": 179
+ },
+ {
+ "epoch": 2.503056768558952,
+ "grad_norm": 0.01147883478552103,
+ "learning_rate": 0.0006,
+ "loss": 7.163824558258057,
+ "step": 180
+ },
+ {
+ "epoch": 2.5170305676855893,
+ "grad_norm": 0.011338942684233189,
+ "learning_rate": 0.0006,
+ "loss": 7.16287899017334,
+ "step": 181
+ },
+ {
+ "epoch": 2.531004366812227,
+ "grad_norm": 0.013277675025165081,
+ "learning_rate": 0.0006,
+ "loss": 7.155905723571777,
+ "step": 182
+ },
+ {
+ "epoch": 2.544978165938865,
+ "grad_norm": 0.015269347466528416,
+ "learning_rate": 0.0006,
+ "loss": 7.189371109008789,
+ "step": 183
+ },
+ {
+ "epoch": 2.558951965065502,
+ "grad_norm": 0.017905596643686295,
+ "learning_rate": 0.0006,
+ "loss": 7.177547931671143,
+ "step": 184
+ },
+ {
+ "epoch": 2.5729257641921395,
+ "grad_norm": 0.024431610479950905,
+ "learning_rate": 0.0006,
+ "loss": 7.192169189453125,
+ "step": 185
+ },
+ {
+ "epoch": 2.5868995633187772,
+ "grad_norm": 0.03656961768865585,
+ "learning_rate": 0.0006,
+ "loss": 7.124387741088867,
+ "step": 186
+ },
+ {
+ "epoch": 2.600873362445415,
+ "grad_norm": 0.04337048903107643,
+ "learning_rate": 0.0006,
+ "loss": 7.140231609344482,
+ "step": 187
+ },
+ {
+ "epoch": 2.6148471615720523,
+ "grad_norm": 0.04047228768467903,
+ "learning_rate": 0.0006,
+ "loss": 7.171927452087402,
+ "step": 188
+ },
+ {
+ "epoch": 2.62882096069869,
+ "grad_norm": 0.07825770974159241,
+ "learning_rate": 0.0006,
+ "loss": 7.207631587982178,
+ "step": 189
+ },
+ {
+ "epoch": 2.6427947598253274,
+ "grad_norm": 0.07118063420057297,
+ "learning_rate": 0.0006,
+ "loss": 7.209488391876221,
+ "step": 190
+ },
+ {
+ "epoch": 2.656768558951965,
+ "grad_norm": 0.05153293535113335,
+ "learning_rate": 0.0006,
+ "loss": 7.14793062210083,
+ "step": 191
+ },
+ {
+ "epoch": 2.6707423580786025,
+ "grad_norm": 0.03318839520215988,
+ "learning_rate": 0.0006,
+ "loss": 7.163010597229004,
+ "step": 192
+ },
+ {
+ "epoch": 2.6847161572052403,
+ "grad_norm": 0.029333539307117462,
+ "learning_rate": 0.0006,
+ "loss": 7.159955978393555,
+ "step": 193
+ },
+ {
+ "epoch": 2.6986899563318776,
+ "grad_norm": 0.029405880719423294,
+ "learning_rate": 0.0006,
+ "loss": 7.136178493499756,
+ "step": 194
+ },
+ {
+ "epoch": 2.7126637554585153,
+ "grad_norm": 0.037815988063812256,
+ "learning_rate": 0.0006,
+ "loss": 7.183186054229736,
+ "step": 195
+ },
+ {
+ "epoch": 2.726637554585153,
+ "grad_norm": 0.020307250320911407,
+ "learning_rate": 0.0006,
+ "loss": 7.133164882659912,
+ "step": 196
+ },
+ {
+ "epoch": 2.7406113537117904,
+ "grad_norm": 0.026048069819808006,
+ "learning_rate": 0.0006,
+ "loss": 7.124977111816406,
+ "step": 197
+ },
+ {
+ "epoch": 2.7545851528384278,
+ "grad_norm": 0.027316724881529808,
+ "learning_rate": 0.0006,
+ "loss": 7.07265567779541,
+ "step": 198
+ },
+ {
+ "epoch": 2.7685589519650655,
+ "grad_norm": 0.021629009395837784,
+ "learning_rate": 0.0006,
+ "loss": 7.150020599365234,
+ "step": 199
+ },
+ {
+ "epoch": 2.7825327510917033,
+ "grad_norm": 0.01605929434299469,
+ "learning_rate": 0.0006,
+ "loss": 7.10919189453125,
+ "step": 200
+ },
+ {
+ "epoch": 2.7965065502183406,
+ "grad_norm": 0.020800089463591576,
+ "learning_rate": 0.0006,
+ "loss": 7.132045269012451,
+ "step": 201
+ },
+ {
+ "epoch": 2.810480349344978,
+ "grad_norm": 0.019225774332880974,
+ "learning_rate": 0.0006,
+ "loss": 7.128364562988281,
+ "step": 202
+ },
+ {
+ "epoch": 2.8244541484716157,
+ "grad_norm": 0.01561372634023428,
+ "learning_rate": 0.0006,
+ "loss": 7.083669662475586,
+ "step": 203
+ },
+ {
+ "epoch": 2.8384279475982535,
+ "grad_norm": 0.021969519555568695,
+ "learning_rate": 0.0006,
+ "loss": 7.050841331481934,
+ "step": 204
+ },
+ {
+ "epoch": 2.8524017467248908,
+ "grad_norm": 0.021453695371747017,
+ "learning_rate": 0.0006,
+ "loss": 7.097085475921631,
+ "step": 205
+ },
+ {
+ "epoch": 2.8663755458515285,
+ "grad_norm": 0.022836022078990936,
+ "learning_rate": 0.0006,
+ "loss": 7.10405969619751,
+ "step": 206
+ },
+ {
+ "epoch": 2.880349344978166,
+ "grad_norm": 0.02430851384997368,
+ "learning_rate": 0.0006,
+ "loss": 7.045181751251221,
+ "step": 207
+ },
+ {
+ "epoch": 2.8943231441048036,
+ "grad_norm": 0.020166993141174316,
+ "learning_rate": 0.0006,
+ "loss": 7.068183898925781,
+ "step": 208
+ },
+ {
+ "epoch": 2.908296943231441,
+ "grad_norm": 0.0118505684658885,
+ "learning_rate": 0.0006,
+ "loss": 7.07132625579834,
+ "step": 209
+ },
+ {
+ "epoch": 2.9222707423580787,
+ "grad_norm": 0.023772427812218666,
+ "learning_rate": 0.0006,
+ "loss": 7.055238723754883,
+ "step": 210
+ },
+ {
+ "epoch": 2.936244541484716,
+ "grad_norm": 0.02350712940096855,
+ "learning_rate": 0.0006,
+ "loss": 7.055898666381836,
+ "step": 211
+ },
+ {
+ "epoch": 2.950218340611354,
+ "grad_norm": 0.017244907096028328,
+ "learning_rate": 0.0006,
+ "loss": 7.0099711418151855,
+ "step": 212
+ },
+ {
+ "epoch": 2.964192139737991,
+ "grad_norm": 0.021565575152635574,
+ "learning_rate": 0.0006,
+ "loss": 6.991751670837402,
+ "step": 213
+ },
+ {
+ "epoch": 2.978165938864629,
+ "grad_norm": 0.03241860866546631,
+ "learning_rate": 0.0006,
+ "loss": 7.005980491638184,
+ "step": 214
+ },
+ {
+ "epoch": 2.992139737991266,
+ "grad_norm": 0.04613726586103439,
+ "learning_rate": 0.0006,
+ "loss": 7.027859687805176,
+ "step": 215
+ },
+ {
+ "epoch": 3.0,
+ "grad_norm": 0.06404894590377808,
+ "learning_rate": 0.0006,
+ "loss": 7.123725891113281,
+ "step": 216
+ },
+ {
+ "epoch": 3.0,
+ "eval_loss": 7.070647716522217,
+ "eval_runtime": 58.5801,
+ "eval_samples_per_second": 41.687,
+ "eval_steps_per_second": 1.314,
+ "step": 216
+ },
+ {
+ "epoch": 3.0139737991266378,
+ "grad_norm": 0.05542432889342308,
+ "learning_rate": 0.0006,
+ "loss": 7.078299522399902,
+ "step": 217
+ },
+ {
+ "epoch": 3.027947598253275,
+ "grad_norm": 0.026484526693820953,
+ "learning_rate": 0.0006,
+ "loss": 6.961159706115723,
+ "step": 218
+ },
+ {
+ "epoch": 3.041921397379913,
+ "grad_norm": 0.028276391327381134,
+ "learning_rate": 0.0006,
+ "loss": 7.019961833953857,
+ "step": 219
+ },
+ {
+ "epoch": 3.05589519650655,
+ "grad_norm": 0.024486595764756203,
+ "learning_rate": 0.0006,
+ "loss": 7.039677619934082,
+ "step": 220
+ },
+ {
+ "epoch": 3.069868995633188,
+ "grad_norm": 0.017620893195271492,
+ "learning_rate": 0.0006,
+ "loss": 6.9270782470703125,
+ "step": 221
+ },
+ {
+ "epoch": 3.0838427947598253,
+ "grad_norm": 0.02547302283346653,
+ "learning_rate": 0.0006,
+ "loss": 6.924233436584473,
+ "step": 222
+ },
+ {
+ "epoch": 3.097816593886463,
+ "grad_norm": 0.02539009042084217,
+ "learning_rate": 0.0006,
+ "loss": 6.973979949951172,
+ "step": 223
+ },
+ {
+ "epoch": 3.1117903930131003,
+ "grad_norm": 0.03256227448582649,
+ "learning_rate": 0.0006,
+ "loss": 6.933725357055664,
+ "step": 224
+ },
+ {
+ "epoch": 3.125764192139738,
+ "grad_norm": 0.04651800915598869,
+ "learning_rate": 0.0006,
+ "loss": 6.951044082641602,
+ "step": 225
+ },
+ {
+ "epoch": 3.1397379912663754,
+ "grad_norm": 0.04822975769639015,
+ "learning_rate": 0.0006,
+ "loss": 6.9614410400390625,
+ "step": 226
+ },
+ {
+ "epoch": 3.153711790393013,
+ "grad_norm": 0.031165020540356636,
+ "learning_rate": 0.0006,
+ "loss": 6.91193151473999,
+ "step": 227
+ },
+ {
+ "epoch": 3.1676855895196505,
+ "grad_norm": 0.029754292219877243,
+ "learning_rate": 0.0006,
+ "loss": 6.91802978515625,
+ "step": 228
+ },
+ {
+ "epoch": 3.1816593886462883,
+ "grad_norm": 0.02193152718245983,
+ "learning_rate": 0.0006,
+ "loss": 6.928090572357178,
+ "step": 229
+ },
+ {
+ "epoch": 3.1956331877729256,
+ "grad_norm": 0.02428470179438591,
+ "learning_rate": 0.0006,
+ "loss": 6.90396785736084,
+ "step": 230
+ },
+ {
+ "epoch": 3.2096069868995634,
+ "grad_norm": 0.023374345153570175,
+ "learning_rate": 0.0006,
+ "loss": 6.885858535766602,
+ "step": 231
+ },
+ {
+ "epoch": 3.2235807860262007,
+ "grad_norm": 0.023405000567436218,
+ "learning_rate": 0.0006,
+ "loss": 6.8855767250061035,
+ "step": 232
+ },
+ {
+ "epoch": 3.2375545851528384,
+ "grad_norm": 0.017277134582400322,
+ "learning_rate": 0.0006,
+ "loss": 6.901449203491211,
+ "step": 233
+ },
+ {
+ "epoch": 3.251528384279476,
+ "grad_norm": 0.019496750086545944,
+ "learning_rate": 0.0006,
+ "loss": 6.885687828063965,
+ "step": 234
+ },
+ {
+ "epoch": 3.2655021834061135,
+ "grad_norm": 0.016893276944756508,
+ "learning_rate": 0.0006,
+ "loss": 6.872369766235352,
+ "step": 235
+ },
+ {
+ "epoch": 3.279475982532751,
+ "grad_norm": 0.02058715932071209,
+ "learning_rate": 0.0006,
+ "loss": 6.823280334472656,
+ "step": 236
+ },
+ {
+ "epoch": 3.2934497816593886,
+ "grad_norm": 0.015530755743384361,
+ "learning_rate": 0.0006,
+ "loss": 6.860499858856201,
+ "step": 237
+ },
+ {
+ "epoch": 3.3074235807860264,
+ "grad_norm": 0.01934926211833954,
+ "learning_rate": 0.0006,
+ "loss": 6.817448616027832,
+ "step": 238
+ },
+ {
+ "epoch": 3.3213973799126637,
+ "grad_norm": 0.017413552850484848,
+ "learning_rate": 0.0006,
+ "loss": 6.881955146789551,
+ "step": 239
+ },
+ {
+ "epoch": 3.3353711790393015,
+ "grad_norm": 0.026478875428438187,
+ "learning_rate": 0.0006,
+ "loss": 6.827404499053955,
+ "step": 240
+ },
+ {
+ "epoch": 3.349344978165939,
+ "grad_norm": 0.032611701637506485,
+ "learning_rate": 0.0006,
+ "loss": 6.78285551071167,
+ "step": 241
+ },
+ {
+ "epoch": 3.3633187772925766,
+ "grad_norm": 0.041900306940078735,
+ "learning_rate": 0.0006,
+ "loss": 6.783053874969482,
+ "step": 242
+ },
+ {
+ "epoch": 3.377292576419214,
+ "grad_norm": 0.04149497672915459,
+ "learning_rate": 0.0006,
+ "loss": 6.745220184326172,
+ "step": 243
+ },
+ {
+ "epoch": 3.3912663755458516,
+ "grad_norm": 0.023949826136231422,
+ "learning_rate": 0.0006,
+ "loss": 6.841533184051514,
+ "step": 244
+ },
+ {
+ "epoch": 3.405240174672489,
+ "grad_norm": 0.028601842001080513,
+ "learning_rate": 0.0006,
+ "loss": 6.852852821350098,
+ "step": 245
+ },
+ {
+ "epoch": 3.4192139737991267,
+ "grad_norm": 0.027365155518054962,
+ "learning_rate": 0.0006,
+ "loss": 6.74376106262207,
+ "step": 246
+ },
+ {
+ "epoch": 3.433187772925764,
+ "grad_norm": 0.02856568619608879,
+ "learning_rate": 0.0006,
+ "loss": 6.74906063079834,
+ "step": 247
+ },
+ {
+ "epoch": 3.447161572052402,
+ "grad_norm": 0.019172416999936104,
+ "learning_rate": 0.0006,
+ "loss": 6.809206008911133,
+ "step": 248
+ },
+ {
+ "epoch": 3.461135371179039,
+ "grad_norm": 0.02161746844649315,
+ "learning_rate": 0.0006,
+ "loss": 6.789888381958008,
+ "step": 249
+ },
+ {
+ "epoch": 3.475109170305677,
+ "grad_norm": 0.020273666828870773,
+ "learning_rate": 0.0006,
+ "loss": 6.73512077331543,
+ "step": 250
+ },
+ {
+ "epoch": 3.489082969432314,
+ "grad_norm": 0.022329283878207207,
+ "learning_rate": 0.0006,
+ "loss": 6.826148509979248,
+ "step": 251
+ },
+ {
+ "epoch": 3.503056768558952,
+ "grad_norm": 0.02284284122288227,
+ "learning_rate": 0.0006,
+ "loss": 6.748521327972412,
+ "step": 252
+ },
+ {
+ "epoch": 3.5170305676855893,
+ "grad_norm": 0.026368247345089912,
+ "learning_rate": 0.0006,
+ "loss": 6.645879745483398,
+ "step": 253
+ },
+ {
+ "epoch": 3.531004366812227,
+ "grad_norm": 0.03627244383096695,
+ "learning_rate": 0.0006,
+ "loss": 6.666064739227295,
+ "step": 254
+ },
+ {
+ "epoch": 3.544978165938865,
+ "grad_norm": 0.04140935093164444,
+ "learning_rate": 0.0006,
+ "loss": 6.712275981903076,
+ "step": 255
+ },
+ {
+ "epoch": 3.558951965065502,
+ "grad_norm": 0.0390462800860405,
+ "learning_rate": 0.0006,
+ "loss": 6.665900230407715,
+ "step": 256
+ },
+ {
+ "epoch": 3.5729257641921395,
+ "grad_norm": 0.05178583785891533,
+ "learning_rate": 0.0006,
+ "loss": 6.753298759460449,
+ "step": 257
+ },
+ {
+ "epoch": 3.5868995633187772,
+ "grad_norm": 0.054328061640262604,
+ "learning_rate": 0.0006,
+ "loss": 6.732633113861084,
+ "step": 258
+ },
+ {
+ "epoch": 3.600873362445415,
+ "grad_norm": 0.04590460658073425,
+ "learning_rate": 0.0006,
+ "loss": 6.65250301361084,
+ "step": 259
+ },
+ {
+ "epoch": 3.6148471615720523,
+ "grad_norm": 0.044331666082143784,
+ "learning_rate": 0.0006,
+ "loss": 6.68654203414917,
+ "step": 260
+ },
+ {
+ "epoch": 3.62882096069869,
+ "grad_norm": 0.05308730527758598,
+ "learning_rate": 0.0006,
+ "loss": 6.642482757568359,
+ "step": 261
+ },
+ {
+ "epoch": 3.6427947598253274,
+ "grad_norm": 0.03999016433954239,
+ "learning_rate": 0.0006,
+ "loss": 6.625227451324463,
+ "step": 262
+ },
+ {
+ "epoch": 3.656768558951965,
+ "grad_norm": 0.04088086634874344,
+ "learning_rate": 0.0006,
+ "loss": 6.67623233795166,
+ "step": 263
+ },
+ {
+ "epoch": 3.6707423580786025,
+ "grad_norm": 0.03615636005997658,
+ "learning_rate": 0.0006,
+ "loss": 6.6869001388549805,
+ "step": 264
+ },
+ {
+ "epoch": 3.6847161572052403,
+ "grad_norm": 0.02817360684275627,
+ "learning_rate": 0.0006,
+ "loss": 6.674601078033447,
+ "step": 265
+ },
+ {
+ "epoch": 3.6986899563318776,
+ "grad_norm": 0.032586123794317245,
+ "learning_rate": 0.0006,
+ "loss": 6.612994194030762,
+ "step": 266
+ },
+ {
+ "epoch": 3.7126637554585153,
+ "grad_norm": 0.031117867678403854,
+ "learning_rate": 0.0006,
+ "loss": 6.6244354248046875,
+ "step": 267
+ },
+ {
+ "epoch": 3.726637554585153,
+ "grad_norm": 0.028330687433481216,
+ "learning_rate": 0.0006,
+ "loss": 6.669557094573975,
+ "step": 268
+ },
+ {
+ "epoch": 3.7406113537117904,
+ "grad_norm": 0.021489202976226807,
+ "learning_rate": 0.0006,
+ "loss": 6.632035732269287,
+ "step": 269
+ },
+ {
+ "epoch": 3.7545851528384278,
+ "grad_norm": 0.024269424378871918,
+ "learning_rate": 0.0006,
+ "loss": 6.660867691040039,
+ "step": 270
+ },
+ {
+ "epoch": 3.7685589519650655,
+ "grad_norm": 0.02237873338162899,
+ "learning_rate": 0.0006,
+ "loss": 6.602587699890137,
+ "step": 271
+ },
+ {
+ "epoch": 3.7825327510917033,
+ "grad_norm": 0.026503929868340492,
+ "learning_rate": 0.0006,
+ "loss": 6.599200248718262,
+ "step": 272
+ },
+ {
+ "epoch": 3.7965065502183406,
+ "grad_norm": 0.03687124326825142,
+ "learning_rate": 0.0006,
+ "loss": 6.567296504974365,
+ "step": 273
+ },
+ {
+ "epoch": 3.810480349344978,
+ "grad_norm": 0.04708952456712723,
+ "learning_rate": 0.0006,
+ "loss": 6.621265411376953,
+ "step": 274
+ },
+ {
+ "epoch": 3.8244541484716157,
+ "grad_norm": 0.041846614331007004,
+ "learning_rate": 0.0006,
+ "loss": 6.550536155700684,
+ "step": 275
+ },
+ {
+ "epoch": 3.8384279475982535,
+ "grad_norm": 0.031249675899744034,
+ "learning_rate": 0.0006,
+ "loss": 6.54278564453125,
+ "step": 276
+ },
+ {
+ "epoch": 3.8524017467248908,
+ "grad_norm": 0.023082256317138672,
+ "learning_rate": 0.0006,
+ "loss": 6.544787406921387,
+ "step": 277
+ },
+ {
+ "epoch": 3.8663755458515285,
+ "grad_norm": 0.020995570346713066,
+ "learning_rate": 0.0006,
+ "loss": 6.518143653869629,
+ "step": 278
+ },
+ {
+ "epoch": 3.880349344978166,
+ "grad_norm": 0.020839324221014977,
+ "learning_rate": 0.0006,
+ "loss": 6.546667098999023,
+ "step": 279
+ },
+ {
+ "epoch": 3.8943231441048036,
+ "grad_norm": 0.018377432599663734,
+ "learning_rate": 0.0006,
+ "loss": 6.478307723999023,
+ "step": 280
+ },
+ {
+ "epoch": 3.908296943231441,
+ "grad_norm": 0.021713724359869957,
+ "learning_rate": 0.0006,
+ "loss": 6.485574245452881,
+ "step": 281
+ },
+ {
+ "epoch": 3.9222707423580787,
+ "grad_norm": 0.01359301246702671,
+ "learning_rate": 0.0006,
+ "loss": 6.570268630981445,
+ "step": 282
+ },
+ {
+ "epoch": 3.936244541484716,
+ "grad_norm": 0.016233332455158234,
+ "learning_rate": 0.0006,
+ "loss": 6.57150936126709,
+ "step": 283
+ },
+ {
+ "epoch": 3.950218340611354,
+ "grad_norm": 0.020687608048319817,
+ "learning_rate": 0.0006,
+ "loss": 6.527956008911133,
+ "step": 284
+ },
+ {
+ "epoch": 3.964192139737991,
+ "grad_norm": 0.02032964862883091,
+ "learning_rate": 0.0006,
+ "loss": 6.468169689178467,
+ "step": 285
+ },
+ {
+ "epoch": 3.978165938864629,
+ "grad_norm": 0.016676288098096848,
+ "learning_rate": 0.0006,
+ "loss": 6.448418617248535,
+ "step": 286
+ },
+ {
+ "epoch": 3.992139737991266,
+ "grad_norm": 0.0168539360165596,
+ "learning_rate": 0.0006,
+ "loss": 6.461180686950684,
+ "step": 287
+ },
+ {
+ "epoch": 4.0,
+ "grad_norm": 0.02666233666241169,
+ "learning_rate": 0.0006,
+ "loss": 6.344893932342529,
+ "step": 288
+ },
+ {
+ "epoch": 4.0,
+ "eval_loss": 6.5384979248046875,
+ "eval_runtime": 58.814,
+ "eval_samples_per_second": 41.521,
+ "eval_steps_per_second": 1.309,
+ "step": 288
+ },
+ {
+ "epoch": 4.013973799126638,
+ "grad_norm": 0.04154708981513977,
+ "learning_rate": 0.0006,
+ "loss": 6.41855525970459,
+ "step": 289
+ },
+ {
+ "epoch": 4.0279475982532755,
+ "grad_norm": 0.04895668104290962,
+ "learning_rate": 0.0006,
+ "loss": 6.499302864074707,
+ "step": 290
+ },
+ {
+ "epoch": 4.041921397379912,
+ "grad_norm": 0.034464482218027115,
+ "learning_rate": 0.0006,
+ "loss": 6.438611030578613,
+ "step": 291
+ },
+ {
+ "epoch": 4.05589519650655,
+ "grad_norm": 0.06845773756504059,
+ "learning_rate": 0.0006,
+ "loss": 6.538361072540283,
+ "step": 292
+ },
+ {
+ "epoch": 4.069868995633188,
+ "grad_norm": 0.053768858313560486,
+ "learning_rate": 0.0006,
+ "loss": 6.527524948120117,
+ "step": 293
+ },
+ {
+ "epoch": 4.083842794759826,
+ "grad_norm": 0.039037931710481644,
+ "learning_rate": 0.0006,
+ "loss": 6.503847122192383,
+ "step": 294
+ },
+ {
+ "epoch": 4.097816593886463,
+ "grad_norm": 0.03962196782231331,
+ "learning_rate": 0.0006,
+ "loss": 6.475178241729736,
+ "step": 295
+ },
+ {
+ "epoch": 4.1117903930131,
+ "grad_norm": 0.049101535230875015,
+ "learning_rate": 0.0006,
+ "loss": 6.474587917327881,
+ "step": 296
+ },
+ {
+ "epoch": 4.125764192139738,
+ "grad_norm": 0.02830282226204872,
+ "learning_rate": 0.0006,
+ "loss": 6.45065975189209,
+ "step": 297
+ },
+ {
+ "epoch": 4.139737991266376,
+ "grad_norm": 0.023153482005000114,
+ "learning_rate": 0.0006,
+ "loss": 6.5218706130981445,
+ "step": 298
+ },
+ {
+ "epoch": 4.153711790393013,
+ "grad_norm": 0.025064002722501755,
+ "learning_rate": 0.0006,
+ "loss": 6.456475257873535,
+ "step": 299
+ },
+ {
+ "epoch": 4.1676855895196505,
+ "grad_norm": 0.02357092685997486,
+ "learning_rate": 0.0006,
+ "loss": 6.485952377319336,
+ "step": 300
+ },
+ {
+ "epoch": 4.181659388646288,
+ "grad_norm": 0.021832305938005447,
+ "learning_rate": 0.0006,
+ "loss": 6.425506591796875,
+ "step": 301
+ },
+ {
+ "epoch": 4.195633187772926,
+ "grad_norm": 0.0227900929749012,
+ "learning_rate": 0.0006,
+ "loss": 6.439155578613281,
+ "step": 302
+ },
+ {
+ "epoch": 4.209606986899563,
+ "grad_norm": 0.015961607918143272,
+ "learning_rate": 0.0006,
+ "loss": 6.460862636566162,
+ "step": 303
+ },
+ {
+ "epoch": 4.223580786026201,
+ "grad_norm": 0.017876390367746353,
+ "learning_rate": 0.0006,
+ "loss": 6.44564151763916,
+ "step": 304
+ },
+ {
+ "epoch": 4.2375545851528384,
+ "grad_norm": 0.013703204691410065,
+ "learning_rate": 0.0006,
+ "loss": 6.418414115905762,
+ "step": 305
+ },
+ {
+ "epoch": 4.251528384279476,
+ "grad_norm": 0.015240306034684181,
+ "learning_rate": 0.0006,
+ "loss": 6.411238670349121,
+ "step": 306
+ },
+ {
+ "epoch": 4.265502183406113,
+ "grad_norm": 0.01587662845849991,
+ "learning_rate": 0.0006,
+ "loss": 6.395642280578613,
+ "step": 307
+ },
+ {
+ "epoch": 4.279475982532751,
+ "grad_norm": 0.014650700613856316,
+ "learning_rate": 0.0006,
+ "loss": 6.302793025970459,
+ "step": 308
+ },
+ {
+ "epoch": 4.293449781659389,
+ "grad_norm": 0.016533225774765015,
+ "learning_rate": 0.0006,
+ "loss": 6.380588054656982,
+ "step": 309
+ },
+ {
+ "epoch": 4.307423580786026,
+ "grad_norm": 0.019963741302490234,
+ "learning_rate": 0.0006,
+ "loss": 6.345336437225342,
+ "step": 310
+ },
+ {
+ "epoch": 4.321397379912664,
+ "grad_norm": 0.02237936295568943,
+ "learning_rate": 0.0006,
+ "loss": 6.387770652770996,
+ "step": 311
+ },
+ {
+ "epoch": 4.335371179039301,
+ "grad_norm": 0.026658328250050545,
+ "learning_rate": 0.0006,
+ "loss": 6.323662757873535,
+ "step": 312
+ },
+ {
+ "epoch": 4.349344978165939,
+ "grad_norm": 0.02852284163236618,
+ "learning_rate": 0.0006,
+ "loss": 6.382204055786133,
+ "step": 313
+ },
+ {
+ "epoch": 4.3633187772925766,
+ "grad_norm": 0.028815090656280518,
+ "learning_rate": 0.0006,
+ "loss": 6.402390480041504,
+ "step": 314
+ },
+ {
+ "epoch": 4.377292576419214,
+ "grad_norm": 0.029393529519438744,
+ "learning_rate": 0.0006,
+ "loss": 6.280278205871582,
+ "step": 315
+ },
+ {
+ "epoch": 4.391266375545851,
+ "grad_norm": 0.02186041697859764,
+ "learning_rate": 0.0006,
+ "loss": 6.302707672119141,
+ "step": 316
+ },
+ {
+ "epoch": 4.405240174672489,
+ "grad_norm": 0.02071218006312847,
+ "learning_rate": 0.0006,
+ "loss": 6.3365020751953125,
+ "step": 317
+ },
+ {
+ "epoch": 4.419213973799127,
+ "grad_norm": 0.02779117226600647,
+ "learning_rate": 0.0006,
+ "loss": 6.252157211303711,
+ "step": 318
+ },
+ {
+ "epoch": 4.4331877729257645,
+ "grad_norm": 0.03470654785633087,
+ "learning_rate": 0.0006,
+ "loss": 6.319070339202881,
+ "step": 319
+ },
+ {
+ "epoch": 4.447161572052401,
+ "grad_norm": 0.037067804485559464,
+ "learning_rate": 0.0006,
+ "loss": 6.314116477966309,
+ "step": 320
+ },
+ {
+ "epoch": 4.461135371179039,
+ "grad_norm": 0.032182756811380386,
+ "learning_rate": 0.0006,
+ "loss": 6.267298698425293,
+ "step": 321
+ },
+ {
+ "epoch": 4.475109170305677,
+ "grad_norm": 0.026305217295885086,
+ "learning_rate": 0.0006,
+ "loss": 6.331688404083252,
+ "step": 322
+ },
+ {
+ "epoch": 4.489082969432315,
+ "grad_norm": 0.027228357270359993,
+ "learning_rate": 0.0006,
+ "loss": 6.31139612197876,
+ "step": 323
+ },
+ {
+ "epoch": 4.503056768558952,
+ "grad_norm": 0.025781169533729553,
+ "learning_rate": 0.0006,
+ "loss": 6.194684982299805,
+ "step": 324
+ },
+ {
+ "epoch": 4.517030567685589,
+ "grad_norm": 0.029186977073550224,
+ "learning_rate": 0.0006,
+ "loss": 6.31306266784668,
+ "step": 325
+ },
+ {
+ "epoch": 4.531004366812227,
+ "grad_norm": 0.023608777672052383,
+ "learning_rate": 0.0006,
+ "loss": 6.285243034362793,
+ "step": 326
+ },
+ {
+ "epoch": 4.544978165938865,
+ "grad_norm": 0.02267594076693058,
+ "learning_rate": 0.0006,
+ "loss": 6.30342960357666,
+ "step": 327
+ },
+ {
+ "epoch": 4.558951965065502,
+ "grad_norm": 0.01926310732960701,
+ "learning_rate": 0.0006,
+ "loss": 6.280606269836426,
+ "step": 328
+ },
+ {
+ "epoch": 4.5729257641921395,
+ "grad_norm": 0.026484837755560875,
+ "learning_rate": 0.0006,
+ "loss": 6.237980365753174,
+ "step": 329
+ },
+ {
+ "epoch": 4.586899563318777,
+ "grad_norm": 0.026228854432702065,
+ "learning_rate": 0.0006,
+ "loss": 6.3018293380737305,
+ "step": 330
+ },
+ {
+ "epoch": 4.600873362445415,
+ "grad_norm": 0.022096967324614525,
+ "learning_rate": 0.0006,
+ "loss": 6.253863334655762,
+ "step": 331
+ },
+ {
+ "epoch": 4.614847161572053,
+ "grad_norm": 0.027223890647292137,
+ "learning_rate": 0.0006,
+ "loss": 6.166882038116455,
+ "step": 332
+ },
+ {
+ "epoch": 4.62882096069869,
+ "grad_norm": 0.03160124272108078,
+ "learning_rate": 0.0006,
+ "loss": 6.230309963226318,
+ "step": 333
+ },
+ {
+ "epoch": 4.642794759825327,
+ "grad_norm": 0.04110539332032204,
+ "learning_rate": 0.0006,
+ "loss": 6.2334818840026855,
+ "step": 334
+ },
+ {
+ "epoch": 4.656768558951965,
+ "grad_norm": 0.04355933889746666,
+ "learning_rate": 0.0006,
+ "loss": 6.160956382751465,
+ "step": 335
+ },
+ {
+ "epoch": 4.670742358078603,
+ "grad_norm": 0.03611086308956146,
+ "learning_rate": 0.0006,
+ "loss": 6.23399543762207,
+ "step": 336
+ },
+ {
+ "epoch": 4.68471615720524,
+ "grad_norm": 0.051811400800943375,
+ "learning_rate": 0.0006,
+ "loss": 6.241855144500732,
+ "step": 337
+ },
+ {
+ "epoch": 4.698689956331878,
+ "grad_norm": 0.057434193789958954,
+ "learning_rate": 0.0006,
+ "loss": 6.220993995666504,
+ "step": 338
+ },
+ {
+ "epoch": 4.712663755458515,
+ "grad_norm": 0.04420953616499901,
+ "learning_rate": 0.0006,
+ "loss": 6.300461292266846,
+ "step": 339
+ },
+ {
+ "epoch": 4.726637554585153,
+ "grad_norm": 0.03515457361936569,
+ "learning_rate": 0.0006,
+ "loss": 6.269349098205566,
+ "step": 340
+ },
+ {
+ "epoch": 4.74061135371179,
+ "grad_norm": 0.03222600743174553,
+ "learning_rate": 0.0006,
+ "loss": 6.248613357543945,
+ "step": 341
+ },
+ {
+ "epoch": 4.754585152838428,
+ "grad_norm": 0.0384046845138073,
+ "learning_rate": 0.0006,
+ "loss": 6.266979217529297,
+ "step": 342
+ },
+ {
+ "epoch": 4.7685589519650655,
+ "grad_norm": 0.0514867827296257,
+ "learning_rate": 0.0006,
+ "loss": 6.234103202819824,
+ "step": 343
+ },
+ {
+ "epoch": 4.782532751091703,
+ "grad_norm": 0.050747793167829514,
+ "learning_rate": 0.0006,
+ "loss": 6.322582244873047,
+ "step": 344
+ },
+ {
+ "epoch": 4.796506550218341,
+ "grad_norm": 0.040878988802433014,
+ "learning_rate": 0.0006,
+ "loss": 6.212509632110596,
+ "step": 345
+ },
+ {
+ "epoch": 4.810480349344978,
+ "grad_norm": 0.034655820578336716,
+ "learning_rate": 0.0006,
+ "loss": 6.265879154205322,
+ "step": 346
+ },
+ {
+ "epoch": 4.824454148471616,
+ "grad_norm": 0.04202224686741829,
+ "learning_rate": 0.0006,
+ "loss": 6.2896409034729,
+ "step": 347
+ },
+ {
+ "epoch": 4.8384279475982535,
+ "grad_norm": 0.03632277995347977,
+ "learning_rate": 0.0006,
+ "loss": 6.253917217254639,
+ "step": 348
+ },
+ {
+ "epoch": 4.85240174672489,
+ "grad_norm": 0.028645765036344528,
+ "learning_rate": 0.0006,
+ "loss": 6.207965850830078,
+ "step": 349
+ },
+ {
+ "epoch": 4.866375545851528,
+ "grad_norm": 0.025305170565843582,
+ "learning_rate": 0.0006,
+ "loss": 6.168068885803223,
+ "step": 350
+ },
+ {
+ "epoch": 4.880349344978166,
+ "grad_norm": 0.022137803956866264,
+ "learning_rate": 0.0006,
+ "loss": 6.235283374786377,
+ "step": 351
+ },
+ {
+ "epoch": 4.894323144104804,
+ "grad_norm": 0.01814538985490799,
+ "learning_rate": 0.0006,
+ "loss": 6.259641647338867,
+ "step": 352
+ },
+ {
+ "epoch": 4.908296943231441,
+ "grad_norm": 0.018465086817741394,
+ "learning_rate": 0.0006,
+ "loss": 6.111738681793213,
+ "step": 353
+ },
+ {
+ "epoch": 4.922270742358078,
+ "grad_norm": 0.018638933077454567,
+ "learning_rate": 0.0006,
+ "loss": 6.225642204284668,
+ "step": 354
+ },
+ {
+ "epoch": 4.936244541484716,
+ "grad_norm": 0.016461260616779327,
+ "learning_rate": 0.0006,
+ "loss": 6.140647888183594,
+ "step": 355
+ },
+ {
+ "epoch": 4.950218340611354,
+ "grad_norm": 0.015458385460078716,
+ "learning_rate": 0.0006,
+ "loss": 6.155224800109863,
+ "step": 356
+ },
+ {
+ "epoch": 4.964192139737992,
+ "grad_norm": 0.014934048056602478,
+ "learning_rate": 0.0006,
+ "loss": 6.101555824279785,
+ "step": 357
+ },
+ {
+ "epoch": 4.978165938864628,
+ "grad_norm": 0.013979545794427395,
+ "learning_rate": 0.0006,
+ "loss": 6.161718845367432,
+ "step": 358
+ },
+ {
+ "epoch": 4.992139737991266,
+ "grad_norm": 0.01640394888818264,
+ "learning_rate": 0.0006,
+ "loss": 6.055768966674805,
+ "step": 359
+ },
+ {
+ "epoch": 5.0,
+ "grad_norm": 0.015135680325329304,
+ "learning_rate": 0.0006,
+ "loss": 6.169695854187012,
+ "step": 360
+ },
+ {
+ "epoch": 5.0,
+ "eval_loss": 6.16408109664917,
+ "eval_runtime": 59.3649,
+ "eval_samples_per_second": 41.135,
+ "eval_steps_per_second": 1.297,
+ "step": 360
+ },
+ {
+ "epoch": 5.013973799126638,
+ "grad_norm": 0.012849048711359501,
+ "learning_rate": 0.0006,
+ "loss": 6.055830001831055,
+ "step": 361
+ },
+ {
+ "epoch": 5.0279475982532755,
+ "grad_norm": 0.018169576302170753,
+ "learning_rate": 0.0006,
+ "loss": 6.082655429840088,
+ "step": 362
+ },
+ {
+ "epoch": 5.041921397379912,
+ "grad_norm": 0.024293430149555206,
+ "learning_rate": 0.0006,
+ "loss": 6.109032154083252,
+ "step": 363
+ },
+ {
+ "epoch": 5.05589519650655,
+ "grad_norm": 0.03266787901520729,
+ "learning_rate": 0.0006,
+ "loss": 6.030613899230957,
+ "step": 364
+ },
+ {
+ "epoch": 5.069868995633188,
+ "grad_norm": 0.03572266176342964,
+ "learning_rate": 0.0006,
+ "loss": 6.050390720367432,
+ "step": 365
+ },
+ {
+ "epoch": 5.083842794759826,
+ "grad_norm": 0.032625213265419006,
+ "learning_rate": 0.0006,
+ "loss": 6.169313430786133,
+ "step": 366
+ },
+ {
+ "epoch": 5.097816593886463,
+ "grad_norm": 0.04051872715353966,
+ "learning_rate": 0.0006,
+ "loss": 6.10398006439209,
+ "step": 367
+ },
+ {
+ "epoch": 5.1117903930131,
+ "grad_norm": 0.0351213738322258,
+ "learning_rate": 0.0006,
+ "loss": 6.121973514556885,
+ "step": 368
+ },
+ {
+ "epoch": 5.125764192139738,
+ "grad_norm": 0.03523759916424751,
+ "learning_rate": 0.0006,
+ "loss": 6.0626654624938965,
+ "step": 369
+ },
+ {
+ "epoch": 5.139737991266376,
+ "grad_norm": 0.02706182189285755,
+ "learning_rate": 0.0006,
+ "loss": 6.049354076385498,
+ "step": 370
+ },
+ {
+ "epoch": 5.153711790393013,
+ "grad_norm": 0.024212589487433434,
+ "learning_rate": 0.0006,
+ "loss": 6.034826278686523,
+ "step": 371
+ },
+ {
+ "epoch": 5.1676855895196505,
+ "grad_norm": 0.024798082187771797,
+ "learning_rate": 0.0006,
+ "loss": 6.004058837890625,
+ "step": 372
+ },
+ {
+ "epoch": 5.181659388646288,
+ "grad_norm": 0.02325567416846752,
+ "learning_rate": 0.0006,
+ "loss": 5.986461639404297,
+ "step": 373
+ },
+ {
+ "epoch": 5.195633187772926,
+ "grad_norm": 0.019060570746660233,
+ "learning_rate": 0.0006,
+ "loss": 6.059736251831055,
+ "step": 374
+ },
+ {
+ "epoch": 5.209606986899563,
+ "grad_norm": 0.016822976991534233,
+ "learning_rate": 0.0006,
+ "loss": 6.072957515716553,
+ "step": 375
+ },
+ {
+ "epoch": 5.223580786026201,
+ "grad_norm": 0.017218658700585365,
+ "learning_rate": 0.0006,
+ "loss": 6.031739234924316,
+ "step": 376
+ },
+ {
+ "epoch": 5.2375545851528384,
+ "grad_norm": 0.014889383688569069,
+ "learning_rate": 0.0006,
+ "loss": 6.025674343109131,
+ "step": 377
+ },
+ {
+ "epoch": 5.251528384279476,
+ "grad_norm": 0.01708107627928257,
+ "learning_rate": 0.0006,
+ "loss": 6.090451240539551,
+ "step": 378
+ },
+ {
+ "epoch": 5.265502183406113,
+ "grad_norm": 0.020260317251086235,
+ "learning_rate": 0.0006,
+ "loss": 6.094257354736328,
+ "step": 379
+ },
+ {
+ "epoch": 5.279475982532751,
+ "grad_norm": 0.020110400393605232,
+ "learning_rate": 0.0006,
+ "loss": 6.021188735961914,
+ "step": 380
+ },
+ {
+ "epoch": 5.293449781659389,
+ "grad_norm": 0.02027830481529236,
+ "learning_rate": 0.0006,
+ "loss": 6.022156238555908,
+ "step": 381
+ },
+ {
+ "epoch": 5.307423580786026,
+ "grad_norm": 0.022747062146663666,
+ "learning_rate": 0.0006,
+ "loss": 6.011836051940918,
+ "step": 382
+ },
+ {
+ "epoch": 5.321397379912664,
+ "grad_norm": 0.01990230567753315,
+ "learning_rate": 0.0006,
+ "loss": 6.003054618835449,
+ "step": 383
+ },
+ {
+ "epoch": 5.335371179039301,
+ "grad_norm": 0.01596238650381565,
+ "learning_rate": 0.0006,
+ "loss": 5.991410732269287,
+ "step": 384
+ },
+ {
+ "epoch": 5.349344978165939,
+ "grad_norm": 0.015847105532884598,
+ "learning_rate": 0.0006,
+ "loss": 6.100622177124023,
+ "step": 385
+ },
+ {
+ "epoch": 5.3633187772925766,
+ "grad_norm": 0.016179397702217102,
+ "learning_rate": 0.0006,
+ "loss": 6.032659530639648,
+ "step": 386
+ },
+ {
+ "epoch": 5.377292576419214,
+ "grad_norm": 0.018256602808833122,
+ "learning_rate": 0.0006,
+ "loss": 5.961983680725098,
+ "step": 387
+ },
+ {
+ "epoch": 5.391266375545851,
+ "grad_norm": 0.02805912122130394,
+ "learning_rate": 0.0006,
+ "loss": 5.983541965484619,
+ "step": 388
+ },
+ {
+ "epoch": 5.405240174672489,
+ "grad_norm": 0.039082255214452744,
+ "learning_rate": 0.0006,
+ "loss": 6.021474361419678,
+ "step": 389
+ },
+ {
+ "epoch": 5.419213973799127,
+ "grad_norm": 0.036757659167051315,
+ "learning_rate": 0.0006,
+ "loss": 6.005046367645264,
+ "step": 390
+ },
+ {
+ "epoch": 5.4331877729257645,
+ "grad_norm": 0.035277366638183594,
+ "learning_rate": 0.0006,
+ "loss": 6.036296844482422,
+ "step": 391
+ },
+ {
+ "epoch": 5.447161572052401,
+ "grad_norm": 0.034404635429382324,
+ "learning_rate": 0.0006,
+ "loss": 5.994539260864258,
+ "step": 392
+ },
+ {
+ "epoch": 5.461135371179039,
+ "grad_norm": 0.0377206988632679,
+ "learning_rate": 0.0006,
+ "loss": 5.941534519195557,
+ "step": 393
+ },
+ {
+ "epoch": 5.475109170305677,
+ "grad_norm": 0.0389922633767128,
+ "learning_rate": 0.0006,
+ "loss": 5.983644485473633,
+ "step": 394
+ },
+ {
+ "epoch": 5.489082969432315,
+ "grad_norm": 0.04176778718829155,
+ "learning_rate": 0.0006,
+ "loss": 6.030922889709473,
+ "step": 395
+ },
+ {
+ "epoch": 5.503056768558952,
+ "grad_norm": 0.0337153784930706,
+ "learning_rate": 0.0006,
+ "loss": 5.961367130279541,
+ "step": 396
+ },
+ {
+ "epoch": 5.517030567685589,
+ "grad_norm": 0.036116816103458405,
+ "learning_rate": 0.0006,
+ "loss": 5.819280624389648,
+ "step": 397
+ },
+ {
+ "epoch": 5.531004366812227,
+ "grad_norm": 0.030724041163921356,
+ "learning_rate": 0.0006,
+ "loss": 5.906380653381348,
+ "step": 398
+ },
+ {
+ "epoch": 5.544978165938865,
+ "grad_norm": 0.030264202505350113,
+ "learning_rate": 0.0006,
+ "loss": 5.975833415985107,
+ "step": 399
+ },
+ {
+ "epoch": 5.558951965065502,
+ "grad_norm": 0.032096318900585175,
+ "learning_rate": 0.0006,
+ "loss": 6.001348972320557,
+ "step": 400
+ },
+ {
+ "epoch": 5.5729257641921395,
+ "grad_norm": 0.030579356476664543,
+ "learning_rate": 0.0006,
+ "loss": 5.928768634796143,
+ "step": 401
+ },
+ {
+ "epoch": 5.586899563318777,
+ "grad_norm": 0.028241973370313644,
+ "learning_rate": 0.0006,
+ "loss": 5.92582893371582,
+ "step": 402
+ },
+ {
+ "epoch": 5.600873362445415,
+ "grad_norm": 0.023086953908205032,
+ "learning_rate": 0.0006,
+ "loss": 5.892926216125488,
+ "step": 403
+ },
+ {
+ "epoch": 5.614847161572053,
+ "grad_norm": 0.02547992393374443,
+ "learning_rate": 0.0006,
+ "loss": 5.903195381164551,
+ "step": 404
+ },
+ {
+ "epoch": 5.62882096069869,
+ "grad_norm": 0.023089831694960594,
+ "learning_rate": 0.0006,
+ "loss": 5.912033557891846,
+ "step": 405
+ },
+ {
+ "epoch": 5.642794759825327,
+ "grad_norm": 0.025446368381381035,
+ "learning_rate": 0.0006,
+ "loss": 5.926138401031494,
+ "step": 406
+ },
+ {
+ "epoch": 5.656768558951965,
+ "grad_norm": 0.03748747706413269,
+ "learning_rate": 0.0006,
+ "loss": 5.913451194763184,
+ "step": 407
+ },
+ {
+ "epoch": 5.670742358078603,
+ "grad_norm": 0.03493810072541237,
+ "learning_rate": 0.0006,
+ "loss": 5.97633695602417,
+ "step": 408
+ },
+ {
+ "epoch": 5.68471615720524,
+ "grad_norm": 0.023131275549530983,
+ "learning_rate": 0.0006,
+ "loss": 5.899571418762207,
+ "step": 409
+ },
+ {
+ "epoch": 5.698689956331878,
+ "grad_norm": 0.02986014075577259,
+ "learning_rate": 0.0006,
+ "loss": 5.894903182983398,
+ "step": 410
+ },
+ {
+ "epoch": 5.712663755458515,
+ "grad_norm": 0.030171144753694534,
+ "learning_rate": 0.0006,
+ "loss": 5.876595497131348,
+ "step": 411
+ },
+ {
+ "epoch": 5.726637554585153,
+ "grad_norm": 0.029377546161413193,
+ "learning_rate": 0.0006,
+ "loss": 5.911637306213379,
+ "step": 412
+ },
+ {
+ "epoch": 5.74061135371179,
+ "grad_norm": 0.029829490929841995,
+ "learning_rate": 0.0006,
+ "loss": 5.819629669189453,
+ "step": 413
+ },
+ {
+ "epoch": 5.754585152838428,
+ "grad_norm": 0.024139299988746643,
+ "learning_rate": 0.0006,
+ "loss": 5.81728458404541,
+ "step": 414
+ },
+ {
+ "epoch": 5.7685589519650655,
+ "grad_norm": 0.021453432738780975,
+ "learning_rate": 0.0006,
+ "loss": 5.987326145172119,
+ "step": 415
+ },
+ {
+ "epoch": 5.782532751091703,
+ "grad_norm": 0.017942246049642563,
+ "learning_rate": 0.0006,
+ "loss": 5.893008232116699,
+ "step": 416
+ },
+ {
+ "epoch": 5.796506550218341,
+ "grad_norm": 0.019724637269973755,
+ "learning_rate": 0.0006,
+ "loss": 5.912441253662109,
+ "step": 417
+ },
+ {
+ "epoch": 5.810480349344978,
+ "grad_norm": 0.018597912043333054,
+ "learning_rate": 0.0006,
+ "loss": 5.860394477844238,
+ "step": 418
+ },
+ {
+ "epoch": 5.824454148471616,
+ "grad_norm": 0.016232412308454514,
+ "learning_rate": 0.0006,
+ "loss": 5.94991397857666,
+ "step": 419
+ },
+ {
+ "epoch": 5.8384279475982535,
+ "grad_norm": 0.01667204685509205,
+ "learning_rate": 0.0006,
+ "loss": 5.8506364822387695,
+ "step": 420
+ },
+ {
+ "epoch": 5.85240174672489,
+ "grad_norm": 0.014232151210308075,
+ "learning_rate": 0.0006,
+ "loss": 5.928424835205078,
+ "step": 421
+ },
+ {
+ "epoch": 5.866375545851528,
+ "grad_norm": 0.01570476032793522,
+ "learning_rate": 0.0006,
+ "loss": 5.953692436218262,
+ "step": 422
+ },
+ {
+ "epoch": 5.880349344978166,
+ "grad_norm": 0.01586179807782173,
+ "learning_rate": 0.0006,
+ "loss": 5.820014476776123,
+ "step": 423
+ },
+ {
+ "epoch": 5.894323144104804,
+ "grad_norm": 0.014729145914316177,
+ "learning_rate": 0.0006,
+ "loss": 5.860154628753662,
+ "step": 424
+ },
+ {
+ "epoch": 5.908296943231441,
+ "grad_norm": 0.013111459091305733,
+ "learning_rate": 0.0006,
+ "loss": 5.85988187789917,
+ "step": 425
+ },
+ {
+ "epoch": 5.922270742358078,
+ "grad_norm": 0.012937922962009907,
+ "learning_rate": 0.0006,
+ "loss": 5.756265640258789,
+ "step": 426
+ },
+ {
+ "epoch": 5.936244541484716,
+ "grad_norm": 0.01377453189343214,
+ "learning_rate": 0.0006,
+ "loss": 5.8333048820495605,
+ "step": 427
+ },
+ {
+ "epoch": 5.950218340611354,
+ "grad_norm": 0.014054795727133751,
+ "learning_rate": 0.0006,
+ "loss": 5.803333759307861,
+ "step": 428
+ },
+ {
+ "epoch": 5.964192139737992,
+ "grad_norm": 0.01674959622323513,
+ "learning_rate": 0.0006,
+ "loss": 5.8718485832214355,
+ "step": 429
+ },
+ {
+ "epoch": 5.978165938864628,
+ "grad_norm": 0.018733017146587372,
+ "learning_rate": 0.0006,
+ "loss": 5.875979900360107,
+ "step": 430
+ },
+ {
+ "epoch": 5.992139737991266,
+ "grad_norm": 0.02088983915746212,
+ "learning_rate": 0.0006,
+ "loss": 5.763634204864502,
+ "step": 431
+ },
+ {
+ "epoch": 6.0,
+ "grad_norm": 0.026705985888838768,
+ "learning_rate": 0.0006,
+ "loss": 5.6922478675842285,
+ "step": 432
+ },
+ {
+ "epoch": 6.0,
+ "eval_loss": 5.847958087921143,
+ "eval_runtime": 58.5451,
+ "eval_samples_per_second": 41.711,
+ "eval_steps_per_second": 1.315,
+ "step": 432
+ },
+ {
+ "epoch": 6.013973799126638,
+ "grad_norm": 0.03181544691324234,
+ "learning_rate": 0.0006,
+ "loss": 5.786368370056152,
+ "step": 433
+ },
+ {
+ "epoch": 6.0279475982532755,
+ "grad_norm": 0.03238112851977348,
+ "learning_rate": 0.0006,
+ "loss": 5.810310363769531,
+ "step": 434
+ },
+ {
+ "epoch": 6.041921397379912,
+ "grad_norm": 0.03916756808757782,
+ "learning_rate": 0.0006,
+ "loss": 5.7761993408203125,
+ "step": 435
+ },
+ {
+ "epoch": 6.05589519650655,
+ "grad_norm": 0.044009730219841,
+ "learning_rate": 0.0006,
+ "loss": 5.898112773895264,
+ "step": 436
+ },
+ {
+ "epoch": 6.069868995633188,
+ "grad_norm": 0.045235246419906616,
+ "learning_rate": 0.0006,
+ "loss": 5.876371383666992,
+ "step": 437
+ },
+ {
+ "epoch": 6.083842794759826,
+ "grad_norm": 0.050218384712934494,
+ "learning_rate": 0.0006,
+ "loss": 5.805103302001953,
+ "step": 438
+ },
+ {
+ "epoch": 6.097816593886463,
+ "grad_norm": 0.05444490164518356,
+ "learning_rate": 0.0006,
+ "loss": 5.826424598693848,
+ "step": 439
+ },
+ {
+ "epoch": 6.1117903930131,
+ "grad_norm": 0.04890982061624527,
+ "learning_rate": 0.0006,
+ "loss": 5.7735395431518555,
+ "step": 440
+ },
+ {
+ "epoch": 6.125764192139738,
+ "grad_norm": 0.05478639155626297,
+ "learning_rate": 0.0006,
+ "loss": 5.823677062988281,
+ "step": 441
+ },
+ {
+ "epoch": 6.139737991266376,
+ "grad_norm": 0.057562313973903656,
+ "learning_rate": 0.0006,
+ "loss": 5.889334678649902,
+ "step": 442
+ },
+ {
+ "epoch": 6.153711790393013,
+ "grad_norm": 0.06447703391313553,
+ "learning_rate": 0.0006,
+ "loss": 5.84621524810791,
+ "step": 443
+ },
+ {
+ "epoch": 6.1676855895196505,
+ "grad_norm": 0.048861872404813766,
+ "learning_rate": 0.0006,
+ "loss": 5.931595802307129,
+ "step": 444
+ },
+ {
+ "epoch": 6.181659388646288,
+ "grad_norm": 0.05521339178085327,
+ "learning_rate": 0.0006,
+ "loss": 5.8229780197143555,
+ "step": 445
+ },
+ {
+ "epoch": 6.195633187772926,
+ "grad_norm": 0.053666990250349045,
+ "learning_rate": 0.0006,
+ "loss": 5.879191875457764,
+ "step": 446
+ },
+ {
+ "epoch": 6.209606986899563,
+ "grad_norm": 0.0451025515794754,
+ "learning_rate": 0.0006,
+ "loss": 5.801628112792969,
+ "step": 447
+ },
+ {
+ "epoch": 6.223580786026201,
+ "grad_norm": 0.04291655868291855,
+ "learning_rate": 0.0006,
+ "loss": 5.914680480957031,
+ "step": 448
+ },
+ {
+ "epoch": 6.2375545851528384,
+ "grad_norm": 0.035102490335702896,
+ "learning_rate": 0.0006,
+ "loss": 5.843682765960693,
+ "step": 449
+ },
+ {
+ "epoch": 6.251528384279476,
+ "grad_norm": 0.03403995931148529,
+ "learning_rate": 0.0006,
+ "loss": 5.827154159545898,
+ "step": 450
+ },
+ {
+ "epoch": 6.265502183406113,
+ "grad_norm": 0.03444375470280647,
+ "learning_rate": 0.0006,
+ "loss": 5.800136566162109,
+ "step": 451
+ },
+ {
+ "epoch": 6.279475982532751,
+ "grad_norm": 0.03165159001946449,
+ "learning_rate": 0.0006,
+ "loss": 5.906252384185791,
+ "step": 452
+ },
+ {
+ "epoch": 6.293449781659389,
+ "grad_norm": 0.026153093203902245,
+ "learning_rate": 0.0006,
+ "loss": 5.7423200607299805,
+ "step": 453
+ },
+ {
+ "epoch": 6.307423580786026,
+ "grad_norm": 0.024357657879590988,
+ "learning_rate": 0.0006,
+ "loss": 5.701364517211914,
+ "step": 454
+ },
+ {
+ "epoch": 6.321397379912664,
+ "grad_norm": 0.021508049219846725,
+ "learning_rate": 0.0006,
+ "loss": 5.8328094482421875,
+ "step": 455
+ },
+ {
+ "epoch": 6.335371179039301,
+ "grad_norm": 0.017313921824097633,
+ "learning_rate": 0.0006,
+ "loss": 5.792145252227783,
+ "step": 456
+ },
+ {
+ "epoch": 6.349344978165939,
+ "grad_norm": 0.018563397228717804,
+ "learning_rate": 0.0006,
+ "loss": 5.732672214508057,
+ "step": 457
+ },
+ {
+ "epoch": 6.3633187772925766,
+ "grad_norm": 0.016568679362535477,
+ "learning_rate": 0.0006,
+ "loss": 5.81948184967041,
+ "step": 458
+ },
+ {
+ "epoch": 6.377292576419214,
+ "grad_norm": 0.014133770950138569,
+ "learning_rate": 0.0006,
+ "loss": 5.734982490539551,
+ "step": 459
+ },
+ {
+ "epoch": 6.391266375545851,
+ "grad_norm": 0.014114447869360447,
+ "learning_rate": 0.0006,
+ "loss": 5.70073127746582,
+ "step": 460
+ },
+ {
+ "epoch": 6.405240174672489,
+ "grad_norm": 0.012907393276691437,
+ "learning_rate": 0.0006,
+ "loss": 5.758626937866211,
+ "step": 461
+ },
+ {
+ "epoch": 6.419213973799127,
+ "grad_norm": 0.012673444114625454,
+ "learning_rate": 0.0006,
+ "loss": 5.775138854980469,
+ "step": 462
+ },
+ {
+ "epoch": 6.4331877729257645,
+ "grad_norm": 0.01311410591006279,
+ "learning_rate": 0.0006,
+ "loss": 5.70920991897583,
+ "step": 463
+ },
+ {
+ "epoch": 6.447161572052401,
+ "grad_norm": 0.012935544364154339,
+ "learning_rate": 0.0006,
+ "loss": 5.781479835510254,
+ "step": 464
+ },
+ {
+ "epoch": 6.461135371179039,
+ "grad_norm": 0.01439738366752863,
+ "learning_rate": 0.0006,
+ "loss": 5.691315650939941,
+ "step": 465
+ },
+ {
+ "epoch": 6.475109170305677,
+ "grad_norm": 0.01205186452716589,
+ "learning_rate": 0.0006,
+ "loss": 5.662499904632568,
+ "step": 466
+ },
+ {
+ "epoch": 6.489082969432315,
+ "grad_norm": 0.011575652286410332,
+ "learning_rate": 0.0006,
+ "loss": 5.6271772384643555,
+ "step": 467
+ },
+ {
+ "epoch": 6.503056768558952,
+ "grad_norm": 0.011672502383589745,
+ "learning_rate": 0.0006,
+ "loss": 5.761662006378174,
+ "step": 468
+ },
+ {
+ "epoch": 6.517030567685589,
+ "grad_norm": 0.011541751213371754,
+ "learning_rate": 0.0006,
+ "loss": 5.762078285217285,
+ "step": 469
+ },
+ {
+ "epoch": 6.531004366812227,
+ "grad_norm": 0.011396365240216255,
+ "learning_rate": 0.0006,
+ "loss": 5.67873477935791,
+ "step": 470
+ },
+ {
+ "epoch": 6.544978165938865,
+ "grad_norm": 0.010688499547541142,
+ "learning_rate": 0.0006,
+ "loss": 5.717051982879639,
+ "step": 471
+ },
+ {
+ "epoch": 6.558951965065502,
+ "grad_norm": 0.012224317528307438,
+ "learning_rate": 0.0006,
+ "loss": 5.707948684692383,
+ "step": 472
+ },
+ {
+ "epoch": 6.5729257641921395,
+ "grad_norm": 0.011856825090944767,
+ "learning_rate": 0.0006,
+ "loss": 5.70878791809082,
+ "step": 473
+ },
+ {
+ "epoch": 6.586899563318777,
+ "grad_norm": 0.01199064590036869,
+ "learning_rate": 0.0006,
+ "loss": 5.708697319030762,
+ "step": 474
+ },
+ {
+ "epoch": 6.600873362445415,
+ "grad_norm": 0.01219446212053299,
+ "learning_rate": 0.0006,
+ "loss": 5.6061577796936035,
+ "step": 475
+ },
+ {
+ "epoch": 6.614847161572053,
+ "grad_norm": 0.013397484086453915,
+ "learning_rate": 0.0006,
+ "loss": 5.624789714813232,
+ "step": 476
+ },
+ {
+ "epoch": 6.62882096069869,
+ "grad_norm": 0.01483136136084795,
+ "learning_rate": 0.0006,
+ "loss": 5.721141338348389,
+ "step": 477
+ },
+ {
+ "epoch": 6.642794759825327,
+ "grad_norm": 0.0194488987326622,
+ "learning_rate": 0.0006,
+ "loss": 5.604279518127441,
+ "step": 478
+ },
+ {
+ "epoch": 6.656768558951965,
+ "grad_norm": 0.02079896256327629,
+ "learning_rate": 0.0006,
+ "loss": 5.586322784423828,
+ "step": 479
+ },
+ {
+ "epoch": 6.670742358078603,
+ "grad_norm": 0.018996229395270348,
+ "learning_rate": 0.0006,
+ "loss": 5.5987772941589355,
+ "step": 480
+ },
+ {
+ "epoch": 6.68471615720524,
+ "grad_norm": 0.015779603272676468,
+ "learning_rate": 0.0006,
+ "loss": 5.588602066040039,
+ "step": 481
+ },
+ {
+ "epoch": 6.698689956331878,
+ "grad_norm": 0.015322973020374775,
+ "learning_rate": 0.0006,
+ "loss": 5.686467170715332,
+ "step": 482
+ },
+ {
+ "epoch": 6.712663755458515,
+ "grad_norm": 0.015282213687896729,
+ "learning_rate": 0.0006,
+ "loss": 5.664676666259766,
+ "step": 483
+ },
+ {
+ "epoch": 6.726637554585153,
+ "grad_norm": 0.016589272767305374,
+ "learning_rate": 0.0006,
+ "loss": 5.666739463806152,
+ "step": 484
+ },
+ {
+ "epoch": 6.74061135371179,
+ "grad_norm": 0.01811421848833561,
+ "learning_rate": 0.0006,
+ "loss": 5.647593021392822,
+ "step": 485
+ },
+ {
+ "epoch": 6.754585152838428,
+ "grad_norm": 0.019652029499411583,
+ "learning_rate": 0.0006,
+ "loss": 5.628726959228516,
+ "step": 486
+ },
+ {
+ "epoch": 6.7685589519650655,
+ "grad_norm": 0.02118665538728237,
+ "learning_rate": 0.0006,
+ "loss": 5.6580705642700195,
+ "step": 487
+ },
+ {
+ "epoch": 6.782532751091703,
+ "grad_norm": 0.02237832546234131,
+ "learning_rate": 0.0006,
+ "loss": 5.635931968688965,
+ "step": 488
+ },
+ {
+ "epoch": 6.796506550218341,
+ "grad_norm": 0.023897631093859673,
+ "learning_rate": 0.0006,
+ "loss": 5.517949104309082,
+ "step": 489
+ },
+ {
+ "epoch": 6.810480349344978,
+ "grad_norm": 0.02442227490246296,
+ "learning_rate": 0.0006,
+ "loss": 5.600021839141846,
+ "step": 490
+ },
+ {
+ "epoch": 6.824454148471616,
+ "grad_norm": 0.024675287306308746,
+ "learning_rate": 0.0006,
+ "loss": 5.6639909744262695,
+ "step": 491
+ },
+ {
+ "epoch": 6.8384279475982535,
+ "grad_norm": 0.030372392386198044,
+ "learning_rate": 0.0006,
+ "loss": 5.582888603210449,
+ "step": 492
+ },
+ {
+ "epoch": 6.85240174672489,
+ "grad_norm": 0.03237690031528473,
+ "learning_rate": 0.0006,
+ "loss": 5.573896884918213,
+ "step": 493
+ },
+ {
+ "epoch": 6.866375545851528,
+ "grad_norm": 0.03511589393019676,
+ "learning_rate": 0.0006,
+ "loss": 5.565878868103027,
+ "step": 494
+ },
+ {
+ "epoch": 6.880349344978166,
+ "grad_norm": 0.03883938118815422,
+ "learning_rate": 0.0006,
+ "loss": 5.58540678024292,
+ "step": 495
+ },
+ {
+ "epoch": 6.894323144104804,
+ "grad_norm": 0.04175502806901932,
+ "learning_rate": 0.0006,
+ "loss": 5.6095194816589355,
+ "step": 496
+ },
+ {
+ "epoch": 6.908296943231441,
+ "grad_norm": 0.04013441503047943,
+ "learning_rate": 0.0006,
+ "loss": 5.654401779174805,
+ "step": 497
+ },
+ {
+ "epoch": 6.922270742358078,
+ "grad_norm": 0.04618770629167557,
+ "learning_rate": 0.0006,
+ "loss": 5.6483473777771,
+ "step": 498
+ },
+ {
+ "epoch": 6.936244541484716,
+ "grad_norm": 0.04507308453321457,
+ "learning_rate": 0.0006,
+ "loss": 5.5856218338012695,
+ "step": 499
+ },
+ {
+ "epoch": 6.950218340611354,
+ "grad_norm": 0.03151383996009827,
+ "learning_rate": 0.0006,
+ "loss": 5.594086647033691,
+ "step": 500
+ },
+ {
+ "epoch": 6.964192139737992,
+ "grad_norm": 0.026883797720074654,
+ "learning_rate": 0.0006,
+ "loss": 5.607676029205322,
+ "step": 501
+ },
+ {
+ "epoch": 6.978165938864628,
+ "grad_norm": 0.02981836162507534,
+ "learning_rate": 0.0006,
+ "loss": 5.563666820526123,
+ "step": 502
+ },
+ {
+ "epoch": 6.992139737991266,
+ "grad_norm": 0.03233477845788002,
+ "learning_rate": 0.0006,
+ "loss": 5.667543411254883,
+ "step": 503
+ },
+ {
+ "epoch": 7.0,
+ "grad_norm": 0.029927456751465797,
+ "learning_rate": 0.0006,
+ "loss": 5.733482837677002,
+ "step": 504
+ },
+ {
+ "epoch": 7.0,
+ "eval_loss": 5.679076671600342,
+ "eval_runtime": 59.5965,
+ "eval_samples_per_second": 40.976,
+ "eval_steps_per_second": 1.292,
+ "step": 504
+ },
+ {
+ "epoch": 7.013973799126638,
+ "grad_norm": 0.027081597596406937,
+ "learning_rate": 0.0006,
+ "loss": 5.634487152099609,
+ "step": 505
+ },
+ {
+ "epoch": 7.0279475982532755,
+ "grad_norm": 0.031959421932697296,
+ "learning_rate": 0.0006,
+ "loss": 5.643294811248779,
+ "step": 506
+ },
+ {
+ "epoch": 7.041921397379912,
+ "grad_norm": 0.02802063524723053,
+ "learning_rate": 0.0006,
+ "loss": 5.573239326477051,
+ "step": 507
+ },
+ {
+ "epoch": 7.05589519650655,
+ "grad_norm": 0.032579705119132996,
+ "learning_rate": 0.0006,
+ "loss": 5.627150058746338,
+ "step": 508
+ },
+ {
+ "epoch": 7.069868995633188,
+ "grad_norm": 0.02503928542137146,
+ "learning_rate": 0.0006,
+ "loss": 5.6045308113098145,
+ "step": 509
+ },
+ {
+ "epoch": 7.083842794759826,
+ "grad_norm": 0.02421317622065544,
+ "learning_rate": 0.0006,
+ "loss": 5.522153377532959,
+ "step": 510
+ },
+ {
+ "epoch": 7.097816593886463,
+ "grad_norm": 0.021983426064252853,
+ "learning_rate": 0.0006,
+ "loss": 5.632939338684082,
+ "step": 511
+ },
+ {
+ "epoch": 7.1117903930131,
+ "grad_norm": 0.021933183073997498,
+ "learning_rate": 0.0006,
+ "loss": 5.592185974121094,
+ "step": 512
+ },
+ {
+ "epoch": 7.125764192139738,
+ "grad_norm": 0.02287031151354313,
+ "learning_rate": 0.0006,
+ "loss": 5.613700866699219,
+ "step": 513
+ },
+ {
+ "epoch": 7.139737991266376,
+ "grad_norm": 0.02281602844595909,
+ "learning_rate": 0.0006,
+ "loss": 5.551383018493652,
+ "step": 514
+ },
+ {
+ "epoch": 7.153711790393013,
+ "grad_norm": 0.021050360053777695,
+ "learning_rate": 0.0006,
+ "loss": 5.55275821685791,
+ "step": 515
+ },
+ {
+ "epoch": 7.1676855895196505,
+ "grad_norm": 0.018299926072359085,
+ "learning_rate": 0.0006,
+ "loss": 5.528225421905518,
+ "step": 516
+ },
+ {
+ "epoch": 7.181659388646288,
+ "grad_norm": 0.02024853602051735,
+ "learning_rate": 0.0006,
+ "loss": 5.522453784942627,
+ "step": 517
+ },
+ {
+ "epoch": 7.195633187772926,
+ "grad_norm": 0.02135239914059639,
+ "learning_rate": 0.0006,
+ "loss": 5.527522563934326,
+ "step": 518
+ },
+ {
+ "epoch": 7.209606986899563,
+ "grad_norm": 0.023394184187054634,
+ "learning_rate": 0.0006,
+ "loss": 5.4535017013549805,
+ "step": 519
+ },
+ {
+ "epoch": 7.223580786026201,
+ "grad_norm": 0.025952080264687538,
+ "learning_rate": 0.0006,
+ "loss": 5.5374674797058105,
+ "step": 520
+ },
+ {
+ "epoch": 7.2375545851528384,
+ "grad_norm": 0.022287718951702118,
+ "learning_rate": 0.0006,
+ "loss": 5.493753433227539,
+ "step": 521
+ },
+ {
+ "epoch": 7.251528384279476,
+ "grad_norm": 0.02018447406589985,
+ "learning_rate": 0.0006,
+ "loss": 5.510575294494629,
+ "step": 522
+ },
+ {
+ "epoch": 7.265502183406113,
+ "grad_norm": 0.02050507813692093,
+ "learning_rate": 0.0006,
+ "loss": 5.466026782989502,
+ "step": 523
+ },
+ {
+ "epoch": 7.279475982532751,
+ "grad_norm": 0.023688802495598793,
+ "learning_rate": 0.0006,
+ "loss": 5.485815525054932,
+ "step": 524
+ },
+ {
+ "epoch": 7.293449781659389,
+ "grad_norm": 0.021078843623399734,
+ "learning_rate": 0.0006,
+ "loss": 5.415581703186035,
+ "step": 525
+ },
+ {
+ "epoch": 7.307423580786026,
+ "grad_norm": 0.018266011029481888,
+ "learning_rate": 0.0006,
+ "loss": 5.402815818786621,
+ "step": 526
+ },
+ {
+ "epoch": 7.321397379912664,
+ "grad_norm": 0.019693493843078613,
+ "learning_rate": 0.0006,
+ "loss": 5.504674911499023,
+ "step": 527
+ },
+ {
+ "epoch": 7.335371179039301,
+ "grad_norm": 0.02031373605132103,
+ "learning_rate": 0.0006,
+ "loss": 5.503837585449219,
+ "step": 528
+ },
+ {
+ "epoch": 7.349344978165939,
+ "grad_norm": 0.017636626958847046,
+ "learning_rate": 0.0006,
+ "loss": 5.530580997467041,
+ "step": 529
+ },
+ {
+ "epoch": 7.3633187772925766,
+ "grad_norm": 0.01787244901061058,
+ "learning_rate": 0.0006,
+ "loss": 5.450218677520752,
+ "step": 530
+ },
+ {
+ "epoch": 7.377292576419214,
+ "grad_norm": 0.0170669574290514,
+ "learning_rate": 0.0006,
+ "loss": 5.481570243835449,
+ "step": 531
+ },
+ {
+ "epoch": 7.391266375545851,
+ "grad_norm": 0.01802165061235428,
+ "learning_rate": 0.0006,
+ "loss": 5.386394500732422,
+ "step": 532
+ },
+ {
+ "epoch": 7.405240174672489,
+ "grad_norm": 0.022949472069740295,
+ "learning_rate": 0.0006,
+ "loss": 5.424929618835449,
+ "step": 533
+ },
+ {
+ "epoch": 7.419213973799127,
+ "grad_norm": 0.034213390201330185,
+ "learning_rate": 0.0006,
+ "loss": 5.346663475036621,
+ "step": 534
+ },
+ {
+ "epoch": 7.4331877729257645,
+ "grad_norm": 0.03847593814134598,
+ "learning_rate": 0.0006,
+ "loss": 5.418482780456543,
+ "step": 535
+ },
+ {
+ "epoch": 7.447161572052401,
+ "grad_norm": 0.029393676668405533,
+ "learning_rate": 0.0006,
+ "loss": 5.456090927124023,
+ "step": 536
+ },
+ {
+ "epoch": 7.461135371179039,
+ "grad_norm": 0.03388667106628418,
+ "learning_rate": 0.0006,
+ "loss": 5.572983741760254,
+ "step": 537
+ },
+ {
+ "epoch": 7.475109170305677,
+ "grad_norm": 0.039690613746643066,
+ "learning_rate": 0.0006,
+ "loss": 5.414067268371582,
+ "step": 538
+ },
+ {
+ "epoch": 7.489082969432315,
+ "grad_norm": 0.03630776330828667,
+ "learning_rate": 0.0006,
+ "loss": 5.523739814758301,
+ "step": 539
+ },
+ {
+ "epoch": 7.503056768558952,
+ "grad_norm": 0.03356870263814926,
+ "learning_rate": 0.0006,
+ "loss": 5.444767951965332,
+ "step": 540
+ },
+ {
+ "epoch": 7.517030567685589,
+ "grad_norm": 0.030812304466962814,
+ "learning_rate": 0.0006,
+ "loss": 5.461244106292725,
+ "step": 541
+ },
+ {
+ "epoch": 7.531004366812227,
+ "grad_norm": 0.03319217637181282,
+ "learning_rate": 0.0006,
+ "loss": 5.502161026000977,
+ "step": 542
+ },
+ {
+ "epoch": 7.544978165938865,
+ "grad_norm": 0.032997481524944305,
+ "learning_rate": 0.0006,
+ "loss": 5.4162278175354,
+ "step": 543
+ },
+ {
+ "epoch": 7.558951965065502,
+ "grad_norm": 0.03364962339401245,
+ "learning_rate": 0.0006,
+ "loss": 5.415736198425293,
+ "step": 544
+ },
+ {
+ "epoch": 7.5729257641921395,
+ "grad_norm": 0.0344221405684948,
+ "learning_rate": 0.0006,
+ "loss": 5.541967868804932,
+ "step": 545
+ },
+ {
+ "epoch": 7.586899563318777,
+ "grad_norm": 0.029609503224492073,
+ "learning_rate": 0.0006,
+ "loss": 5.440614223480225,
+ "step": 546
+ },
+ {
+ "epoch": 7.600873362445415,
+ "grad_norm": 0.029055926948785782,
+ "learning_rate": 0.0006,
+ "loss": 5.462865352630615,
+ "step": 547
+ },
+ {
+ "epoch": 7.614847161572053,
+ "grad_norm": 0.02658848837018013,
+ "learning_rate": 0.0006,
+ "loss": 5.444467544555664,
+ "step": 548
+ },
+ {
+ "epoch": 7.62882096069869,
+ "grad_norm": 0.026276404038071632,
+ "learning_rate": 0.0006,
+ "loss": 5.462278842926025,
+ "step": 549
+ },
+ {
+ "epoch": 7.642794759825327,
+ "grad_norm": 0.0286889486014843,
+ "learning_rate": 0.0006,
+ "loss": 5.4849534034729,
+ "step": 550
+ },
+ {
+ "epoch": 7.656768558951965,
+ "grad_norm": 0.026849817484617233,
+ "learning_rate": 0.0006,
+ "loss": 5.431473731994629,
+ "step": 551
+ },
+ {
+ "epoch": 7.670742358078603,
+ "grad_norm": 0.02312396466732025,
+ "learning_rate": 0.0006,
+ "loss": 5.45462703704834,
+ "step": 552
+ },
+ {
+ "epoch": 7.68471615720524,
+ "grad_norm": 0.026197485625743866,
+ "learning_rate": 0.0006,
+ "loss": 5.452552318572998,
+ "step": 553
+ },
+ {
+ "epoch": 7.698689956331878,
+ "grad_norm": 0.02747255191206932,
+ "learning_rate": 0.0006,
+ "loss": 5.447169780731201,
+ "step": 554
+ },
+ {
+ "epoch": 7.712663755458515,
+ "grad_norm": 0.028123432770371437,
+ "learning_rate": 0.0006,
+ "loss": 5.394400596618652,
+ "step": 555
+ },
+ {
+ "epoch": 7.726637554585153,
+ "grad_norm": 0.02599870041012764,
+ "learning_rate": 0.0006,
+ "loss": 5.398341178894043,
+ "step": 556
+ },
+ {
+ "epoch": 7.74061135371179,
+ "grad_norm": 0.022171657532453537,
+ "learning_rate": 0.0006,
+ "loss": 5.368820667266846,
+ "step": 557
+ },
+ {
+ "epoch": 7.754585152838428,
+ "grad_norm": 0.022309480234980583,
+ "learning_rate": 0.0006,
+ "loss": 5.416772365570068,
+ "step": 558
+ },
+ {
+ "epoch": 7.7685589519650655,
+ "grad_norm": 0.024074165150523186,
+ "learning_rate": 0.0006,
+ "loss": 5.359002113342285,
+ "step": 559
+ },
+ {
+ "epoch": 7.782532751091703,
+ "grad_norm": 0.02636653557419777,
+ "learning_rate": 0.0006,
+ "loss": 5.484368324279785,
+ "step": 560
+ },
+ {
+ "epoch": 7.796506550218341,
+ "grad_norm": 0.02203752100467682,
+ "learning_rate": 0.0006,
+ "loss": 5.375498294830322,
+ "step": 561
+ },
+ {
+ "epoch": 7.810480349344978,
+ "grad_norm": 0.019924819469451904,
+ "learning_rate": 0.0006,
+ "loss": 5.334723472595215,
+ "step": 562
+ },
+ {
+ "epoch": 7.824454148471616,
+ "grad_norm": 0.018755966797471046,
+ "learning_rate": 0.0006,
+ "loss": 5.348781108856201,
+ "step": 563
+ },
+ {
+ "epoch": 7.8384279475982535,
+ "grad_norm": 0.016747845336794853,
+ "learning_rate": 0.0006,
+ "loss": 5.433079719543457,
+ "step": 564
+ },
+ {
+ "epoch": 7.85240174672489,
+ "grad_norm": 0.0170395877212286,
+ "learning_rate": 0.0006,
+ "loss": 5.346794128417969,
+ "step": 565
+ },
+ {
+ "epoch": 7.866375545851528,
+ "grad_norm": 0.016074176877737045,
+ "learning_rate": 0.0006,
+ "loss": 5.3277740478515625,
+ "step": 566
+ },
+ {
+ "epoch": 7.880349344978166,
+ "grad_norm": 0.014119806699454784,
+ "learning_rate": 0.0006,
+ "loss": 5.483551979064941,
+ "step": 567
+ },
+ {
+ "epoch": 7.894323144104804,
+ "grad_norm": 0.015270394273102283,
+ "learning_rate": 0.0006,
+ "loss": 5.378519535064697,
+ "step": 568
+ },
+ {
+ "epoch": 7.908296943231441,
+ "grad_norm": 0.015677539631724358,
+ "learning_rate": 0.0006,
+ "loss": 5.290210247039795,
+ "step": 569
+ },
+ {
+ "epoch": 7.922270742358078,
+ "grad_norm": 0.015930423513054848,
+ "learning_rate": 0.0006,
+ "loss": 5.427360534667969,
+ "step": 570
+ },
+ {
+ "epoch": 7.936244541484716,
+ "grad_norm": 0.016641564667224884,
+ "learning_rate": 0.0006,
+ "loss": 5.301599979400635,
+ "step": 571
+ },
+ {
+ "epoch": 7.950218340611354,
+ "grad_norm": 0.018293552100658417,
+ "learning_rate": 0.0006,
+ "loss": 5.277889251708984,
+ "step": 572
+ },
+ {
+ "epoch": 7.964192139737992,
+ "grad_norm": 0.017618799582123756,
+ "learning_rate": 0.0006,
+ "loss": 5.297194480895996,
+ "step": 573
+ },
+ {
+ "epoch": 7.978165938864628,
+ "grad_norm": 0.015034242533147335,
+ "learning_rate": 0.0006,
+ "loss": 5.42428731918335,
+ "step": 574
+ },
+ {
+ "epoch": 7.992139737991266,
+ "grad_norm": 0.01617511734366417,
+ "learning_rate": 0.0006,
+ "loss": 5.420044898986816,
+ "step": 575
+ },
+ {
+ "epoch": 8.0,
+ "grad_norm": 0.01837257295846939,
+ "learning_rate": 0.0006,
+ "loss": 5.240387916564941,
+ "step": 576
+ },
+ {
+ "epoch": 8.0,
+ "eval_loss": 5.423932075500488,
+ "eval_runtime": 58.0841,
+ "eval_samples_per_second": 42.042,
+ "eval_steps_per_second": 1.326,
+ "step": 576
+ },
+ {
+ "epoch": 8.013973799126637,
+ "grad_norm": 0.020953809842467308,
+ "learning_rate": 0.0006,
+ "loss": 5.4320783615112305,
+ "step": 577
+ },
+ {
+ "epoch": 8.027947598253276,
+ "grad_norm": 0.0227745920419693,
+ "learning_rate": 0.0006,
+ "loss": 5.310197830200195,
+ "step": 578
+ },
+ {
+ "epoch": 8.041921397379912,
+ "grad_norm": 0.021372603252530098,
+ "learning_rate": 0.0006,
+ "loss": 5.339468002319336,
+ "step": 579
+ },
+ {
+ "epoch": 8.055895196506551,
+ "grad_norm": 0.022401731461286545,
+ "learning_rate": 0.0006,
+ "loss": 5.279829978942871,
+ "step": 580
+ },
+ {
+ "epoch": 8.069868995633188,
+ "grad_norm": 0.023942379280924797,
+ "learning_rate": 0.0006,
+ "loss": 5.372323989868164,
+ "step": 581
+ },
+ {
+ "epoch": 8.083842794759825,
+ "grad_norm": 0.024380534887313843,
+ "learning_rate": 0.0006,
+ "loss": 5.253486633300781,
+ "step": 582
+ },
+ {
+ "epoch": 8.097816593886463,
+ "grad_norm": 0.03246625140309334,
+ "learning_rate": 0.0006,
+ "loss": 5.305774211883545,
+ "step": 583
+ },
+ {
+ "epoch": 8.1117903930131,
+ "grad_norm": 0.030399736016988754,
+ "learning_rate": 0.0006,
+ "loss": 5.370976448059082,
+ "step": 584
+ },
+ {
+ "epoch": 8.125764192139737,
+ "grad_norm": 0.027148913592100143,
+ "learning_rate": 0.0006,
+ "loss": 5.368010520935059,
+ "step": 585
+ },
+ {
+ "epoch": 8.139737991266376,
+ "grad_norm": 0.02925540879368782,
+ "learning_rate": 0.0006,
+ "loss": 5.303013801574707,
+ "step": 586
+ },
+ {
+ "epoch": 8.153711790393013,
+ "grad_norm": 0.02915453538298607,
+ "learning_rate": 0.0006,
+ "loss": 5.329667091369629,
+ "step": 587
+ },
+ {
+ "epoch": 8.167685589519651,
+ "grad_norm": 0.031732626259326935,
+ "learning_rate": 0.0006,
+ "loss": 5.3634934425354,
+ "step": 588
+ },
+ {
+ "epoch": 8.181659388646288,
+ "grad_norm": 0.036166731268167496,
+ "learning_rate": 0.0006,
+ "loss": 5.223980903625488,
+ "step": 589
+ },
+ {
+ "epoch": 8.195633187772925,
+ "grad_norm": 0.035654108971357346,
+ "learning_rate": 0.0006,
+ "loss": 5.361059188842773,
+ "step": 590
+ },
+ {
+ "epoch": 8.209606986899564,
+ "grad_norm": 0.0365324430167675,
+ "learning_rate": 0.0006,
+ "loss": 5.336151599884033,
+ "step": 591
+ },
+ {
+ "epoch": 8.2235807860262,
+ "grad_norm": 0.03578880429267883,
+ "learning_rate": 0.0006,
+ "loss": 5.434076309204102,
+ "step": 592
+ },
+ {
+ "epoch": 8.237554585152838,
+ "grad_norm": 0.035734012722969055,
+ "learning_rate": 0.0006,
+ "loss": 5.330893516540527,
+ "step": 593
+ },
+ {
+ "epoch": 8.251528384279476,
+ "grad_norm": 0.03196857124567032,
+ "learning_rate": 0.0006,
+ "loss": 5.306009769439697,
+ "step": 594
+ },
+ {
+ "epoch": 8.265502183406113,
+ "grad_norm": 0.032042086124420166,
+ "learning_rate": 0.0006,
+ "loss": 5.343267917633057,
+ "step": 595
+ },
+ {
+ "epoch": 8.279475982532752,
+ "grad_norm": 0.03160746395587921,
+ "learning_rate": 0.0006,
+ "loss": 5.2353715896606445,
+ "step": 596
+ },
+ {
+ "epoch": 8.293449781659389,
+ "grad_norm": 0.026689818128943443,
+ "learning_rate": 0.0006,
+ "loss": 5.254042625427246,
+ "step": 597
+ },
+ {
+ "epoch": 8.307423580786025,
+ "grad_norm": 0.02880188450217247,
+ "learning_rate": 0.0006,
+ "loss": 5.326833248138428,
+ "step": 598
+ },
+ {
+ "epoch": 8.321397379912664,
+ "grad_norm": 0.027516381815075874,
+ "learning_rate": 0.0006,
+ "loss": 5.342336177825928,
+ "step": 599
+ },
+ {
+ "epoch": 8.335371179039301,
+ "grad_norm": 0.027875030413269997,
+ "learning_rate": 0.0006,
+ "loss": 5.200719833374023,
+ "step": 600
+ },
+ {
+ "epoch": 8.34934497816594,
+ "grad_norm": 0.0268265288323164,
+ "learning_rate": 0.0006,
+ "loss": 5.249449729919434,
+ "step": 601
+ },
+ {
+ "epoch": 8.363318777292577,
+ "grad_norm": 0.024824608117341995,
+ "learning_rate": 0.0006,
+ "loss": 5.318952560424805,
+ "step": 602
+ },
+ {
+ "epoch": 8.377292576419213,
+ "grad_norm": 0.022990627214312553,
+ "learning_rate": 0.0006,
+ "loss": 5.244993209838867,
+ "step": 603
+ },
+ {
+ "epoch": 8.391266375545852,
+ "grad_norm": 0.022804604843258858,
+ "learning_rate": 0.0006,
+ "loss": 5.3236846923828125,
+ "step": 604
+ },
+ {
+ "epoch": 8.405240174672489,
+ "grad_norm": 0.02419872209429741,
+ "learning_rate": 0.0006,
+ "loss": 5.255486488342285,
+ "step": 605
+ },
+ {
+ "epoch": 8.419213973799126,
+ "grad_norm": 0.021952059119939804,
+ "learning_rate": 0.0006,
+ "loss": 5.3511552810668945,
+ "step": 606
+ },
+ {
+ "epoch": 8.433187772925764,
+ "grad_norm": 0.022375497967004776,
+ "learning_rate": 0.0006,
+ "loss": 5.294790267944336,
+ "step": 607
+ },
+ {
+ "epoch": 8.447161572052401,
+ "grad_norm": 0.019242815673351288,
+ "learning_rate": 0.0006,
+ "loss": 5.252618789672852,
+ "step": 608
+ },
+ {
+ "epoch": 8.46113537117904,
+ "grad_norm": 0.0173486340790987,
+ "learning_rate": 0.0006,
+ "loss": 5.30147647857666,
+ "step": 609
+ },
+ {
+ "epoch": 8.475109170305677,
+ "grad_norm": 0.01612604409456253,
+ "learning_rate": 0.0006,
+ "loss": 5.23960542678833,
+ "step": 610
+ },
+ {
+ "epoch": 8.489082969432314,
+ "grad_norm": 0.014293976128101349,
+ "learning_rate": 0.0006,
+ "loss": 5.119932651519775,
+ "step": 611
+ },
+ {
+ "epoch": 8.503056768558952,
+ "grad_norm": 0.014453536830842495,
+ "learning_rate": 0.0006,
+ "loss": 5.166794776916504,
+ "step": 612
+ },
+ {
+ "epoch": 8.51703056768559,
+ "grad_norm": 0.014685769565403461,
+ "learning_rate": 0.0006,
+ "loss": 5.264822959899902,
+ "step": 613
+ },
+ {
+ "epoch": 8.531004366812226,
+ "grad_norm": 0.01427740603685379,
+ "learning_rate": 0.0006,
+ "loss": 5.278433799743652,
+ "step": 614
+ },
+ {
+ "epoch": 8.544978165938865,
+ "grad_norm": 0.014864951372146606,
+ "learning_rate": 0.0006,
+ "loss": 5.214512348175049,
+ "step": 615
+ },
+ {
+ "epoch": 8.558951965065502,
+ "grad_norm": 0.01565164513885975,
+ "learning_rate": 0.0006,
+ "loss": 5.183066368103027,
+ "step": 616
+ },
+ {
+ "epoch": 8.57292576419214,
+ "grad_norm": 0.01779816672205925,
+ "learning_rate": 0.0006,
+ "loss": 5.163185119628906,
+ "step": 617
+ },
+ {
+ "epoch": 8.586899563318777,
+ "grad_norm": 0.017254827544093132,
+ "learning_rate": 0.0006,
+ "loss": 5.243555068969727,
+ "step": 618
+ },
+ {
+ "epoch": 8.600873362445414,
+ "grad_norm": 0.01650584116578102,
+ "learning_rate": 0.0006,
+ "loss": 5.234884738922119,
+ "step": 619
+ },
+ {
+ "epoch": 8.614847161572053,
+ "grad_norm": 0.017021115869283676,
+ "learning_rate": 0.0006,
+ "loss": 5.12271785736084,
+ "step": 620
+ },
+ {
+ "epoch": 8.62882096069869,
+ "grad_norm": 0.01773759163916111,
+ "learning_rate": 0.0006,
+ "loss": 5.242153167724609,
+ "step": 621
+ },
+ {
+ "epoch": 8.642794759825328,
+ "grad_norm": 0.015679042786359787,
+ "learning_rate": 0.0006,
+ "loss": 5.170779228210449,
+ "step": 622
+ },
+ {
+ "epoch": 8.656768558951965,
+ "grad_norm": 0.013760549947619438,
+ "learning_rate": 0.0006,
+ "loss": 5.238644599914551,
+ "step": 623
+ },
+ {
+ "epoch": 8.670742358078602,
+ "grad_norm": 0.014571522362530231,
+ "learning_rate": 0.0006,
+ "loss": 5.177056312561035,
+ "step": 624
+ },
+ {
+ "epoch": 8.68471615720524,
+ "grad_norm": 0.016209015622735023,
+ "learning_rate": 0.0006,
+ "loss": 5.163750648498535,
+ "step": 625
+ },
+ {
+ "epoch": 8.698689956331878,
+ "grad_norm": 0.016813941299915314,
+ "learning_rate": 0.0006,
+ "loss": 5.183428764343262,
+ "step": 626
+ },
+ {
+ "epoch": 8.712663755458514,
+ "grad_norm": 0.020985357463359833,
+ "learning_rate": 0.0006,
+ "loss": 5.229465007781982,
+ "step": 627
+ },
+ {
+ "epoch": 8.726637554585153,
+ "grad_norm": 0.02679632417857647,
+ "learning_rate": 0.0006,
+ "loss": 5.257368087768555,
+ "step": 628
+ },
+ {
+ "epoch": 8.74061135371179,
+ "grad_norm": 0.025756070390343666,
+ "learning_rate": 0.0006,
+ "loss": 5.253736972808838,
+ "step": 629
+ },
+ {
+ "epoch": 8.754585152838429,
+ "grad_norm": 0.02643490768969059,
+ "learning_rate": 0.0006,
+ "loss": 5.228633403778076,
+ "step": 630
+ },
+ {
+ "epoch": 8.768558951965066,
+ "grad_norm": 0.03303210437297821,
+ "learning_rate": 0.0006,
+ "loss": 5.163540363311768,
+ "step": 631
+ },
+ {
+ "epoch": 8.782532751091702,
+ "grad_norm": 0.030432431027293205,
+ "learning_rate": 0.0006,
+ "loss": 5.243169784545898,
+ "step": 632
+ },
+ {
+ "epoch": 8.796506550218341,
+ "grad_norm": 0.028631288558244705,
+ "learning_rate": 0.0006,
+ "loss": 5.166398048400879,
+ "step": 633
+ },
+ {
+ "epoch": 8.810480349344978,
+ "grad_norm": 0.026188310235738754,
+ "learning_rate": 0.0006,
+ "loss": 5.128122329711914,
+ "step": 634
+ },
+ {
+ "epoch": 8.824454148471617,
+ "grad_norm": 0.025308528915047646,
+ "learning_rate": 0.0006,
+ "loss": 5.211284637451172,
+ "step": 635
+ },
+ {
+ "epoch": 8.838427947598253,
+ "grad_norm": 0.020657729357481003,
+ "learning_rate": 0.0006,
+ "loss": 5.161575794219971,
+ "step": 636
+ },
+ {
+ "epoch": 8.85240174672489,
+ "grad_norm": 0.021901234984397888,
+ "learning_rate": 0.0006,
+ "loss": 5.20050048828125,
+ "step": 637
+ },
+ {
+ "epoch": 8.866375545851529,
+ "grad_norm": 0.021219318732619286,
+ "learning_rate": 0.0006,
+ "loss": 5.211699485778809,
+ "step": 638
+ },
+ {
+ "epoch": 8.880349344978166,
+ "grad_norm": 0.0200974028557539,
+ "learning_rate": 0.0006,
+ "loss": 5.092503547668457,
+ "step": 639
+ },
+ {
+ "epoch": 8.894323144104803,
+ "grad_norm": 0.023804882541298866,
+ "learning_rate": 0.0006,
+ "loss": 5.216068267822266,
+ "step": 640
+ },
+ {
+ "epoch": 8.908296943231441,
+ "grad_norm": 0.026088010519742966,
+ "learning_rate": 0.0006,
+ "loss": 5.155592918395996,
+ "step": 641
+ },
+ {
+ "epoch": 8.922270742358078,
+ "grad_norm": 0.02601276896893978,
+ "learning_rate": 0.0006,
+ "loss": 5.197238922119141,
+ "step": 642
+ },
+ {
+ "epoch": 8.936244541484717,
+ "grad_norm": 0.020387211814522743,
+ "learning_rate": 0.0006,
+ "loss": 5.1955437660217285,
+ "step": 643
+ },
+ {
+ "epoch": 8.950218340611354,
+ "grad_norm": 0.019214622676372528,
+ "learning_rate": 0.0006,
+ "loss": 5.206346035003662,
+ "step": 644
+ },
+ {
+ "epoch": 8.96419213973799,
+ "grad_norm": 0.019674314185976982,
+ "learning_rate": 0.0006,
+ "loss": 5.219857215881348,
+ "step": 645
+ },
+ {
+ "epoch": 8.97816593886463,
+ "grad_norm": 0.020318234339356422,
+ "learning_rate": 0.0006,
+ "loss": 5.191132545471191,
+ "step": 646
+ },
+ {
+ "epoch": 8.992139737991266,
+ "grad_norm": 0.021428676322102547,
+ "learning_rate": 0.0006,
+ "loss": 5.214695930480957,
+ "step": 647
+ },
+ {
+ "epoch": 9.0,
+ "grad_norm": 0.023235054686665535,
+ "learning_rate": 0.0006,
+ "loss": 5.229091644287109,
+ "step": 648
+ },
+ {
+ "epoch": 9.0,
+ "eval_loss": 5.289593696594238,
+ "eval_runtime": 56.9256,
+ "eval_samples_per_second": 42.898,
+ "eval_steps_per_second": 1.353,
+ "step": 648
+ },
+ {
+ "epoch": 9.013973799126637,
+ "grad_norm": 0.02320289984345436,
+ "learning_rate": 0.0006,
+ "loss": 4.983669281005859,
+ "step": 649
+ },
+ {
+ "epoch": 9.027947598253276,
+ "grad_norm": 0.018521282821893692,
+ "learning_rate": 0.0006,
+ "loss": 5.200575828552246,
+ "step": 650
+ },
+ {
+ "epoch": 9.041921397379912,
+ "grad_norm": 0.0180169939994812,
+ "learning_rate": 0.0006,
+ "loss": 5.148033618927002,
+ "step": 651
+ },
+ {
+ "epoch": 9.055895196506551,
+ "grad_norm": 0.019969860091805458,
+ "learning_rate": 0.0006,
+ "loss": 5.099125385284424,
+ "step": 652
+ },
+ {
+ "epoch": 9.069868995633188,
+ "grad_norm": 0.01643305830657482,
+ "learning_rate": 0.0006,
+ "loss": 5.092347145080566,
+ "step": 653
+ },
+ {
+ "epoch": 9.083842794759825,
+ "grad_norm": 0.016022363677620888,
+ "learning_rate": 0.0006,
+ "loss": 5.129279613494873,
+ "step": 654
+ },
+ {
+ "epoch": 9.097816593886463,
+ "grad_norm": 0.01657041162252426,
+ "learning_rate": 0.0006,
+ "loss": 5.159562587738037,
+ "step": 655
+ },
+ {
+ "epoch": 9.1117903930131,
+ "grad_norm": 0.017924228683114052,
+ "learning_rate": 0.0006,
+ "loss": 5.165497779846191,
+ "step": 656
+ },
+ {
+ "epoch": 9.125764192139737,
+ "grad_norm": 0.021592361852526665,
+ "learning_rate": 0.0006,
+ "loss": 5.204775810241699,
+ "step": 657
+ },
+ {
+ "epoch": 9.139737991266376,
+ "grad_norm": 0.026924120262265205,
+ "learning_rate": 0.0006,
+ "loss": 5.128296852111816,
+ "step": 658
+ },
+ {
+ "epoch": 9.153711790393013,
+ "grad_norm": 0.023156899958848953,
+ "learning_rate": 0.0006,
+ "loss": 5.1255340576171875,
+ "step": 659
+ },
+ {
+ "epoch": 9.167685589519651,
+ "grad_norm": 0.020670367404818535,
+ "learning_rate": 0.0006,
+ "loss": 5.128479957580566,
+ "step": 660
+ },
+ {
+ "epoch": 9.181659388646288,
+ "grad_norm": 0.025447173044085503,
+ "learning_rate": 0.0006,
+ "loss": 5.154821872711182,
+ "step": 661
+ },
+ {
+ "epoch": 9.195633187772925,
+ "grad_norm": 0.027940068393945694,
+ "learning_rate": 0.0006,
+ "loss": 5.143466472625732,
+ "step": 662
+ },
+ {
+ "epoch": 9.209606986899564,
+ "grad_norm": 0.026847844943404198,
+ "learning_rate": 0.0006,
+ "loss": 5.045665740966797,
+ "step": 663
+ },
+ {
+ "epoch": 9.2235807860262,
+ "grad_norm": 0.02340601570904255,
+ "learning_rate": 0.0006,
+ "loss": 5.21933126449585,
+ "step": 664
+ },
+ {
+ "epoch": 9.237554585152838,
+ "grad_norm": 0.02340371161699295,
+ "learning_rate": 0.0006,
+ "loss": 4.942949295043945,
+ "step": 665
+ },
+ {
+ "epoch": 9.251528384279476,
+ "grad_norm": 0.02221992425620556,
+ "learning_rate": 0.0006,
+ "loss": 5.197099208831787,
+ "step": 666
+ },
+ {
+ "epoch": 9.265502183406113,
+ "grad_norm": 0.023508677259087563,
+ "learning_rate": 0.0006,
+ "loss": 5.022897720336914,
+ "step": 667
+ },
+ {
+ "epoch": 9.279475982532752,
+ "grad_norm": 0.026120394468307495,
+ "learning_rate": 0.0006,
+ "loss": 5.087725639343262,
+ "step": 668
+ },
+ {
+ "epoch": 9.293449781659389,
+ "grad_norm": 0.027273228392004967,
+ "learning_rate": 0.0006,
+ "loss": 5.090963840484619,
+ "step": 669
+ },
+ {
+ "epoch": 9.307423580786025,
+ "grad_norm": 0.03241586685180664,
+ "learning_rate": 0.0006,
+ "loss": 5.047842979431152,
+ "step": 670
+ },
+ {
+ "epoch": 9.321397379912664,
+ "grad_norm": 0.030245667323470116,
+ "learning_rate": 0.0006,
+ "loss": 5.104803085327148,
+ "step": 671
+ },
+ {
+ "epoch": 9.335371179039301,
+ "grad_norm": 0.027698364108800888,
+ "learning_rate": 0.0006,
+ "loss": 5.0644097328186035,
+ "step": 672
+ },
+ {
+ "epoch": 9.34934497816594,
+ "grad_norm": 0.029692554846405983,
+ "learning_rate": 0.0006,
+ "loss": 5.111942291259766,
+ "step": 673
+ },
+ {
+ "epoch": 9.363318777292577,
+ "grad_norm": 0.03328656405210495,
+ "learning_rate": 0.0006,
+ "loss": 5.1574554443359375,
+ "step": 674
+ },
+ {
+ "epoch": 9.377292576419213,
+ "grad_norm": 0.031938180327415466,
+ "learning_rate": 0.0006,
+ "loss": 5.190371990203857,
+ "step": 675
+ },
+ {
+ "epoch": 9.391266375545852,
+ "grad_norm": 0.024858945980668068,
+ "learning_rate": 0.0006,
+ "loss": 5.121420860290527,
+ "step": 676
+ },
+ {
+ "epoch": 9.405240174672489,
+ "grad_norm": 0.023033203557133675,
+ "learning_rate": 0.0006,
+ "loss": 5.0974345207214355,
+ "step": 677
+ },
+ {
+ "epoch": 9.419213973799126,
+ "grad_norm": 0.022083545103669167,
+ "learning_rate": 0.0006,
+ "loss": 5.046319961547852,
+ "step": 678
+ },
+ {
+ "epoch": 9.433187772925764,
+ "grad_norm": 0.0228437427431345,
+ "learning_rate": 0.0006,
+ "loss": 5.119963645935059,
+ "step": 679
+ },
+ {
+ "epoch": 9.447161572052401,
+ "grad_norm": 0.0242351982742548,
+ "learning_rate": 0.0006,
+ "loss": 5.0444560050964355,
+ "step": 680
+ },
+ {
+ "epoch": 9.46113537117904,
+ "grad_norm": 0.02401185780763626,
+ "learning_rate": 0.0006,
+ "loss": 5.063238620758057,
+ "step": 681
+ },
+ {
+ "epoch": 9.475109170305677,
+ "grad_norm": 0.02560959942638874,
+ "learning_rate": 0.0006,
+ "loss": 5.035224914550781,
+ "step": 682
+ },
+ {
+ "epoch": 9.489082969432314,
+ "grad_norm": 0.024985626339912415,
+ "learning_rate": 0.0006,
+ "loss": 5.049284934997559,
+ "step": 683
+ },
+ {
+ "epoch": 9.503056768558952,
+ "grad_norm": 0.023178115487098694,
+ "learning_rate": 0.0006,
+ "loss": 5.2266621589660645,
+ "step": 684
+ },
+ {
+ "epoch": 9.51703056768559,
+ "grad_norm": 0.0215512253344059,
+ "learning_rate": 0.0006,
+ "loss": 5.1875715255737305,
+ "step": 685
+ },
+ {
+ "epoch": 9.531004366812226,
+ "grad_norm": 0.01662837713956833,
+ "learning_rate": 0.0006,
+ "loss": 5.137012481689453,
+ "step": 686
+ },
+ {
+ "epoch": 9.544978165938865,
+ "grad_norm": 0.01952764391899109,
+ "learning_rate": 0.0006,
+ "loss": 5.072500228881836,
+ "step": 687
+ },
+ {
+ "epoch": 9.558951965065502,
+ "grad_norm": 0.017521077767014503,
+ "learning_rate": 0.0006,
+ "loss": 5.045711994171143,
+ "step": 688
+ },
+ {
+ "epoch": 9.57292576419214,
+ "grad_norm": 0.017355265095829964,
+ "learning_rate": 0.0006,
+ "loss": 5.02881383895874,
+ "step": 689
+ },
+ {
+ "epoch": 9.586899563318777,
+ "grad_norm": 0.015587719157338142,
+ "learning_rate": 0.0006,
+ "loss": 5.132787704467773,
+ "step": 690
+ },
+ {
+ "epoch": 9.600873362445414,
+ "grad_norm": 0.017469845712184906,
+ "learning_rate": 0.0006,
+ "loss": 5.105447292327881,
+ "step": 691
+ },
+ {
+ "epoch": 9.614847161572053,
+ "grad_norm": 0.019630100578069687,
+ "learning_rate": 0.0006,
+ "loss": 5.032464981079102,
+ "step": 692
+ },
+ {
+ "epoch": 9.62882096069869,
+ "grad_norm": 0.01784994639456272,
+ "learning_rate": 0.0006,
+ "loss": 5.081345558166504,
+ "step": 693
+ },
+ {
+ "epoch": 9.642794759825328,
+ "grad_norm": 0.01872754842042923,
+ "learning_rate": 0.0006,
+ "loss": 4.9772114753723145,
+ "step": 694
+ },
+ {
+ "epoch": 9.656768558951965,
+ "grad_norm": 0.01999032311141491,
+ "learning_rate": 0.0006,
+ "loss": 5.089740753173828,
+ "step": 695
+ },
+ {
+ "epoch": 9.670742358078602,
+ "grad_norm": 0.01710767112672329,
+ "learning_rate": 0.0006,
+ "loss": 4.9869890213012695,
+ "step": 696
+ },
+ {
+ "epoch": 9.68471615720524,
+ "grad_norm": 0.015144161880016327,
+ "learning_rate": 0.0006,
+ "loss": 5.022564888000488,
+ "step": 697
+ },
+ {
+ "epoch": 9.698689956331878,
+ "grad_norm": 0.01618543453514576,
+ "learning_rate": 0.0006,
+ "loss": 5.035921573638916,
+ "step": 698
+ },
+ {
+ "epoch": 9.712663755458514,
+ "grad_norm": 0.014653387479484081,
+ "learning_rate": 0.0006,
+ "loss": 4.914456367492676,
+ "step": 699
+ },
+ {
+ "epoch": 9.726637554585153,
+ "grad_norm": 0.013434977270662785,
+ "learning_rate": 0.0006,
+ "loss": 5.063764572143555,
+ "step": 700
+ },
+ {
+ "epoch": 9.74061135371179,
+ "grad_norm": 0.015128864906728268,
+ "learning_rate": 0.0006,
+ "loss": 5.052367687225342,
+ "step": 701
+ },
+ {
+ "epoch": 9.754585152838429,
+ "grad_norm": 0.014210768043994904,
+ "learning_rate": 0.0006,
+ "loss": 5.016489028930664,
+ "step": 702
+ },
+ {
+ "epoch": 9.768558951965066,
+ "grad_norm": 0.015191680751740932,
+ "learning_rate": 0.0006,
+ "loss": 4.990176200866699,
+ "step": 703
+ },
+ {
+ "epoch": 9.782532751091702,
+ "grad_norm": 0.018285127356648445,
+ "learning_rate": 0.0006,
+ "loss": 5.015164375305176,
+ "step": 704
+ },
+ {
+ "epoch": 9.796506550218341,
+ "grad_norm": 0.0204079058021307,
+ "learning_rate": 0.0006,
+ "loss": 5.091142654418945,
+ "step": 705
+ },
+ {
+ "epoch": 9.810480349344978,
+ "grad_norm": 0.01980586163699627,
+ "learning_rate": 0.0006,
+ "loss": 5.03350830078125,
+ "step": 706
+ },
+ {
+ "epoch": 9.824454148471617,
+ "grad_norm": 0.02048729732632637,
+ "learning_rate": 0.0006,
+ "loss": 5.03611421585083,
+ "step": 707
+ },
+ {
+ "epoch": 9.838427947598253,
+ "grad_norm": 0.02058163844048977,
+ "learning_rate": 0.0006,
+ "loss": 4.9613447189331055,
+ "step": 708
+ },
+ {
+ "epoch": 9.85240174672489,
+ "grad_norm": 0.018004924058914185,
+ "learning_rate": 0.0006,
+ "loss": 4.966541767120361,
+ "step": 709
+ },
+ {
+ "epoch": 9.866375545851529,
+ "grad_norm": 0.0198152307420969,
+ "learning_rate": 0.0006,
+ "loss": 5.01750373840332,
+ "step": 710
+ },
+ {
+ "epoch": 9.880349344978166,
+ "grad_norm": 0.018532969057559967,
+ "learning_rate": 0.0006,
+ "loss": 5.097424507141113,
+ "step": 711
+ },
+ {
+ "epoch": 9.894323144104803,
+ "grad_norm": 0.018939971923828125,
+ "learning_rate": 0.0006,
+ "loss": 5.177223205566406,
+ "step": 712
+ },
+ {
+ "epoch": 9.908296943231441,
+ "grad_norm": 0.017969170585274696,
+ "learning_rate": 0.0006,
+ "loss": 5.118013858795166,
+ "step": 713
+ },
+ {
+ "epoch": 9.922270742358078,
+ "grad_norm": 0.018021270632743835,
+ "learning_rate": 0.0006,
+ "loss": 5.034787178039551,
+ "step": 714
+ },
+ {
+ "epoch": 9.936244541484717,
+ "grad_norm": 0.017009710893034935,
+ "learning_rate": 0.0006,
+ "loss": 5.003920555114746,
+ "step": 715
+ },
+ {
+ "epoch": 9.950218340611354,
+ "grad_norm": 0.017722049728035927,
+ "learning_rate": 0.0006,
+ "loss": 5.031939506530762,
+ "step": 716
+ },
+ {
+ "epoch": 9.96419213973799,
+ "grad_norm": 0.017134087160229683,
+ "learning_rate": 0.0006,
+ "loss": 5.093474388122559,
+ "step": 717
+ },
+ {
+ "epoch": 9.97816593886463,
+ "grad_norm": 0.018511613830924034,
+ "learning_rate": 0.0006,
+ "loss": 5.078604698181152,
+ "step": 718
+ },
+ {
+ "epoch": 9.992139737991266,
+ "grad_norm": 0.021206334233283997,
+ "learning_rate": 0.0006,
+ "loss": 5.052572727203369,
+ "step": 719
+ },
+ {
+ "epoch": 10.0,
+ "grad_norm": 0.02239062264561653,
+ "learning_rate": 0.0006,
+ "loss": 5.063547134399414,
+ "step": 720
+ },
+ {
+ "epoch": 10.0,
+ "eval_loss": 5.173044681549072,
+ "eval_runtime": 56.7287,
+ "eval_samples_per_second": 43.047,
+ "eval_steps_per_second": 1.357,
+ "step": 720
+ },
+ {
+ "epoch": 10.013973799126637,
+ "grad_norm": 0.020757969468832016,
+ "learning_rate": 0.0006,
+ "loss": 5.045648574829102,
+ "step": 721
+ },
+ {
+ "epoch": 10.027947598253276,
+ "grad_norm": 0.022697702050209045,
+ "learning_rate": 0.0006,
+ "loss": 4.993833065032959,
+ "step": 722
+ },
+ {
+ "epoch": 10.041921397379912,
+ "grad_norm": 0.02434331737458706,
+ "learning_rate": 0.0006,
+ "loss": 5.073219299316406,
+ "step": 723
+ },
+ {
+ "epoch": 10.055895196506551,
+ "grad_norm": 0.025499660521745682,
+ "learning_rate": 0.0006,
+ "loss": 5.019072532653809,
+ "step": 724
+ },
+ {
+ "epoch": 10.069868995633188,
+ "grad_norm": 0.029281053692102432,
+ "learning_rate": 0.0006,
+ "loss": 4.959851264953613,
+ "step": 725
+ },
+ {
+ "epoch": 10.083842794759825,
+ "grad_norm": 0.033878110349178314,
+ "learning_rate": 0.0006,
+ "loss": 5.132717609405518,
+ "step": 726
+ },
+ {
+ "epoch": 10.097816593886463,
+ "grad_norm": 0.03385764732956886,
+ "learning_rate": 0.0006,
+ "loss": 5.0482892990112305,
+ "step": 727
+ },
+ {
+ "epoch": 10.1117903930131,
+ "grad_norm": 0.03466648980975151,
+ "learning_rate": 0.0006,
+ "loss": 4.946855545043945,
+ "step": 728
+ },
+ {
+ "epoch": 10.125764192139737,
+ "grad_norm": 0.028546305373311043,
+ "learning_rate": 0.0006,
+ "loss": 5.027774333953857,
+ "step": 729
+ },
+ {
+ "epoch": 10.139737991266376,
+ "grad_norm": 0.030575869604945183,
+ "learning_rate": 0.0006,
+ "loss": 4.9768476486206055,
+ "step": 730
+ },
+ {
+ "epoch": 10.153711790393013,
+ "grad_norm": 0.03614303469657898,
+ "learning_rate": 0.0006,
+ "loss": 5.078113555908203,
+ "step": 731
+ },
+ {
+ "epoch": 10.167685589519651,
+ "grad_norm": 0.040991660207509995,
+ "learning_rate": 0.0006,
+ "loss": 5.033185958862305,
+ "step": 732
+ },
+ {
+ "epoch": 10.181659388646288,
+ "grad_norm": 0.04551699757575989,
+ "learning_rate": 0.0006,
+ "loss": 5.0820746421813965,
+ "step": 733
+ },
+ {
+ "epoch": 10.195633187772925,
+ "grad_norm": 0.040193989872932434,
+ "learning_rate": 0.0006,
+ "loss": 5.059700965881348,
+ "step": 734
+ },
+ {
+ "epoch": 10.209606986899564,
+ "grad_norm": 0.035851605236530304,
+ "learning_rate": 0.0006,
+ "loss": 5.142467498779297,
+ "step": 735
+ },
+ {
+ "epoch": 10.2235807860262,
+ "grad_norm": 0.034200169146060944,
+ "learning_rate": 0.0006,
+ "loss": 5.072312355041504,
+ "step": 736
+ },
+ {
+ "epoch": 10.237554585152838,
+ "grad_norm": 0.033621110022068024,
+ "learning_rate": 0.0006,
+ "loss": 5.08037805557251,
+ "step": 737
+ },
+ {
+ "epoch": 10.251528384279476,
+ "grad_norm": 0.037372369319200516,
+ "learning_rate": 0.0006,
+ "loss": 4.998425483703613,
+ "step": 738
+ },
+ {
+ "epoch": 10.265502183406113,
+ "grad_norm": 0.037044707685709,
+ "learning_rate": 0.0006,
+ "loss": 4.923993110656738,
+ "step": 739
+ },
+ {
+ "epoch": 10.279475982532752,
+ "grad_norm": 0.030454808846116066,
+ "learning_rate": 0.0006,
+ "loss": 5.0889997482299805,
+ "step": 740
+ },
+ {
+ "epoch": 10.293449781659389,
+ "grad_norm": 0.030969638377428055,
+ "learning_rate": 0.0006,
+ "loss": 4.984112739562988,
+ "step": 741
+ },
+ {
+ "epoch": 10.307423580786025,
+ "grad_norm": 0.0276983380317688,
+ "learning_rate": 0.0006,
+ "loss": 4.955412864685059,
+ "step": 742
+ },
+ {
+ "epoch": 10.321397379912664,
+ "grad_norm": 0.02716052532196045,
+ "learning_rate": 0.0006,
+ "loss": 5.0098876953125,
+ "step": 743
+ },
+ {
+ "epoch": 10.335371179039301,
+ "grad_norm": 0.02403552085161209,
+ "learning_rate": 0.0006,
+ "loss": 5.165160655975342,
+ "step": 744
+ },
+ {
+ "epoch": 10.34934497816594,
+ "grad_norm": 0.0257862601429224,
+ "learning_rate": 0.0006,
+ "loss": 5.001547813415527,
+ "step": 745
+ },
+ {
+ "epoch": 10.363318777292577,
+ "grad_norm": 0.02300378680229187,
+ "learning_rate": 0.0006,
+ "loss": 5.038888931274414,
+ "step": 746
+ },
+ {
+ "epoch": 10.377292576419213,
+ "grad_norm": 0.019766854122281075,
+ "learning_rate": 0.0006,
+ "loss": 4.924291610717773,
+ "step": 747
+ },
+ {
+ "epoch": 10.391266375545852,
+ "grad_norm": 0.01888395845890045,
+ "learning_rate": 0.0006,
+ "loss": 5.059971332550049,
+ "step": 748
+ },
+ {
+ "epoch": 10.405240174672489,
+ "grad_norm": 0.015577499754726887,
+ "learning_rate": 0.0006,
+ "loss": 4.987686634063721,
+ "step": 749
+ },
+ {
+ "epoch": 10.419213973799126,
+ "grad_norm": 0.016012346372008324,
+ "learning_rate": 0.0006,
+ "loss": 5.015157699584961,
+ "step": 750
+ },
+ {
+ "epoch": 10.433187772925764,
+ "grad_norm": 0.016277998685836792,
+ "learning_rate": 0.0006,
+ "loss": 5.004931449890137,
+ "step": 751
+ },
+ {
+ "epoch": 10.447161572052401,
+ "grad_norm": 0.017504561692476273,
+ "learning_rate": 0.0006,
+ "loss": 4.902827739715576,
+ "step": 752
+ },
+ {
+ "epoch": 10.46113537117904,
+ "grad_norm": 0.01569022797048092,
+ "learning_rate": 0.0006,
+ "loss": 4.954863548278809,
+ "step": 753
+ },
+ {
+ "epoch": 10.475109170305677,
+ "grad_norm": 0.014114422723650932,
+ "learning_rate": 0.0006,
+ "loss": 5.00656795501709,
+ "step": 754
+ },
+ {
+ "epoch": 10.489082969432314,
+ "grad_norm": 0.01573832333087921,
+ "learning_rate": 0.0006,
+ "loss": 4.978353023529053,
+ "step": 755
+ },
+ {
+ "epoch": 10.503056768558952,
+ "grad_norm": 0.016480036079883575,
+ "learning_rate": 0.0006,
+ "loss": 4.953327655792236,
+ "step": 756
+ },
+ {
+ "epoch": 10.51703056768559,
+ "grad_norm": 0.014008025638759136,
+ "learning_rate": 0.0006,
+ "loss": 5.050840377807617,
+ "step": 757
+ },
+ {
+ "epoch": 10.531004366812226,
+ "grad_norm": 0.013147998601198196,
+ "learning_rate": 0.0006,
+ "loss": 4.974964141845703,
+ "step": 758
+ },
+ {
+ "epoch": 10.544978165938865,
+ "grad_norm": 0.013807603158056736,
+ "learning_rate": 0.0006,
+ "loss": 4.927907943725586,
+ "step": 759
+ },
+ {
+ "epoch": 10.558951965065502,
+ "grad_norm": 0.013555224984884262,
+ "learning_rate": 0.0006,
+ "loss": 4.980690956115723,
+ "step": 760
+ },
+ {
+ "epoch": 10.57292576419214,
+ "grad_norm": 0.013045408762991428,
+ "learning_rate": 0.0006,
+ "loss": 4.9436116218566895,
+ "step": 761
+ },
+ {
+ "epoch": 10.586899563318777,
+ "grad_norm": 0.012858862057328224,
+ "learning_rate": 0.0006,
+ "loss": 4.9595770835876465,
+ "step": 762
+ },
+ {
+ "epoch": 10.600873362445414,
+ "grad_norm": 0.01468253880739212,
+ "learning_rate": 0.0006,
+ "loss": 4.90725040435791,
+ "step": 763
+ },
+ {
+ "epoch": 10.614847161572053,
+ "grad_norm": 0.013224679045379162,
+ "learning_rate": 0.0006,
+ "loss": 4.877615928649902,
+ "step": 764
+ },
+ {
+ "epoch": 10.62882096069869,
+ "grad_norm": 0.013555348850786686,
+ "learning_rate": 0.0006,
+ "loss": 4.9456610679626465,
+ "step": 765
+ },
+ {
+ "epoch": 10.642794759825328,
+ "grad_norm": 0.014387160539627075,
+ "learning_rate": 0.0006,
+ "loss": 4.9233717918396,
+ "step": 766
+ },
+ {
+ "epoch": 10.656768558951965,
+ "grad_norm": 0.014354088343679905,
+ "learning_rate": 0.0006,
+ "loss": 4.983511924743652,
+ "step": 767
+ },
+ {
+ "epoch": 10.670742358078602,
+ "grad_norm": 0.01332700252532959,
+ "learning_rate": 0.0006,
+ "loss": 4.934416770935059,
+ "step": 768
+ },
+ {
+ "epoch": 10.68471615720524,
+ "grad_norm": 0.011543313041329384,
+ "learning_rate": 0.0006,
+ "loss": 4.96388053894043,
+ "step": 769
+ },
+ {
+ "epoch": 10.698689956331878,
+ "grad_norm": 0.012330878525972366,
+ "learning_rate": 0.0006,
+ "loss": 4.946396350860596,
+ "step": 770
+ },
+ {
+ "epoch": 10.712663755458514,
+ "grad_norm": 0.01379779726266861,
+ "learning_rate": 0.0006,
+ "loss": 4.981019020080566,
+ "step": 771
+ },
+ {
+ "epoch": 10.726637554585153,
+ "grad_norm": 0.014612431637942791,
+ "learning_rate": 0.0006,
+ "loss": 4.894567966461182,
+ "step": 772
+ },
+ {
+ "epoch": 10.74061135371179,
+ "grad_norm": 0.013062899932265282,
+ "learning_rate": 0.0006,
+ "loss": 5.033038139343262,
+ "step": 773
+ },
+ {
+ "epoch": 10.754585152838429,
+ "grad_norm": 0.013898049481213093,
+ "learning_rate": 0.0006,
+ "loss": 4.823636054992676,
+ "step": 774
+ },
+ {
+ "epoch": 10.768558951965066,
+ "grad_norm": 0.01502019353210926,
+ "learning_rate": 0.0006,
+ "loss": 5.020156383514404,
+ "step": 775
+ },
+ {
+ "epoch": 10.782532751091702,
+ "grad_norm": 0.013810204342007637,
+ "learning_rate": 0.0006,
+ "loss": 4.935647010803223,
+ "step": 776
+ },
+ {
+ "epoch": 10.796506550218341,
+ "grad_norm": 0.013616513460874557,
+ "learning_rate": 0.0006,
+ "loss": 4.901486396789551,
+ "step": 777
+ },
+ {
+ "epoch": 10.810480349344978,
+ "grad_norm": 0.013601024635136127,
+ "learning_rate": 0.0006,
+ "loss": 4.931873321533203,
+ "step": 778
+ },
+ {
+ "epoch": 10.824454148471617,
+ "grad_norm": 0.013668350875377655,
+ "learning_rate": 0.0006,
+ "loss": 4.937256813049316,
+ "step": 779
+ },
+ {
+ "epoch": 10.838427947598253,
+ "grad_norm": 0.01581823080778122,
+ "learning_rate": 0.0006,
+ "loss": 4.9611406326293945,
+ "step": 780
+ },
+ {
+ "epoch": 10.85240174672489,
+ "grad_norm": 0.019478484988212585,
+ "learning_rate": 0.0006,
+ "loss": 4.9212236404418945,
+ "step": 781
+ },
+ {
+ "epoch": 10.866375545851529,
+ "grad_norm": 0.026921333745121956,
+ "learning_rate": 0.0006,
+ "loss": 4.88501501083374,
+ "step": 782
+ },
+ {
+ "epoch": 10.880349344978166,
+ "grad_norm": 0.03155914694070816,
+ "learning_rate": 0.0006,
+ "loss": 4.989967346191406,
+ "step": 783
+ },
+ {
+ "epoch": 10.894323144104803,
+ "grad_norm": 0.029689429327845573,
+ "learning_rate": 0.0006,
+ "loss": 4.936000823974609,
+ "step": 784
+ },
+ {
+ "epoch": 10.908296943231441,
+ "grad_norm": 0.026626063510775566,
+ "learning_rate": 0.0006,
+ "loss": 4.984785079956055,
+ "step": 785
+ },
+ {
+ "epoch": 10.922270742358078,
+ "grad_norm": 0.026168212294578552,
+ "learning_rate": 0.0006,
+ "loss": 4.908282279968262,
+ "step": 786
+ },
+ {
+ "epoch": 10.936244541484717,
+ "grad_norm": 0.02828158251941204,
+ "learning_rate": 0.0006,
+ "loss": 4.926390647888184,
+ "step": 787
+ },
+ {
+ "epoch": 10.950218340611354,
+ "grad_norm": 0.02649078331887722,
+ "learning_rate": 0.0006,
+ "loss": 5.00400972366333,
+ "step": 788
+ },
+ {
+ "epoch": 10.96419213973799,
+ "grad_norm": 0.0256856270134449,
+ "learning_rate": 0.0006,
+ "loss": 5.0317487716674805,
+ "step": 789
+ },
+ {
+ "epoch": 10.97816593886463,
+ "grad_norm": 0.024728883057832718,
+ "learning_rate": 0.0006,
+ "loss": 4.9299635887146,
+ "step": 790
+ },
+ {
+ "epoch": 10.992139737991266,
+ "grad_norm": 0.026241116225719452,
+ "learning_rate": 0.0006,
+ "loss": 4.947319984436035,
+ "step": 791
+ },
+ {
+ "epoch": 11.0,
+ "grad_norm": 0.023526743054389954,
+ "learning_rate": 0.0006,
+ "loss": 4.94429349899292,
+ "step": 792
+ },
+ {
+ "epoch": 11.0,
+ "eval_loss": 5.104345798492432,
+ "eval_runtime": 57.0955,
+ "eval_samples_per_second": 42.77,
+ "eval_steps_per_second": 1.349,
+ "step": 792
+ },
+ {
+ "epoch": 11.013973799126637,
+ "grad_norm": 0.0216788612306118,
+ "learning_rate": 0.0006,
+ "loss": 4.881950855255127,
+ "step": 793
+ },
+ {
+ "epoch": 11.027947598253276,
+ "grad_norm": 0.01830463856458664,
+ "learning_rate": 0.0006,
+ "loss": 4.971016883850098,
+ "step": 794
+ },
+ {
+ "epoch": 11.041921397379912,
+ "grad_norm": 0.02204521745443344,
+ "learning_rate": 0.0006,
+ "loss": 4.894623756408691,
+ "step": 795
+ },
+ {
+ "epoch": 11.055895196506551,
+ "grad_norm": 0.0210255216807127,
+ "learning_rate": 0.0006,
+ "loss": 4.834009170532227,
+ "step": 796
+ },
+ {
+ "epoch": 11.069868995633188,
+ "grad_norm": 0.019797448068857193,
+ "learning_rate": 0.0006,
+ "loss": 5.086714744567871,
+ "step": 797
+ },
+ {
+ "epoch": 11.083842794759825,
+ "grad_norm": 0.02036641724407673,
+ "learning_rate": 0.0006,
+ "loss": 4.920056343078613,
+ "step": 798
+ },
+ {
+ "epoch": 11.097816593886463,
+ "grad_norm": 0.01966066285967827,
+ "learning_rate": 0.0006,
+ "loss": 4.879859447479248,
+ "step": 799
+ },
+ {
+ "epoch": 11.1117903930131,
+ "grad_norm": 0.0202149897813797,
+ "learning_rate": 0.0006,
+ "loss": 4.933640480041504,
+ "step": 800
+ },
+ {
+ "epoch": 11.125764192139737,
+ "grad_norm": 0.01893039606511593,
+ "learning_rate": 0.0006,
+ "loss": 4.901970863342285,
+ "step": 801
+ },
+ {
+ "epoch": 11.139737991266376,
+ "grad_norm": 0.016040155664086342,
+ "learning_rate": 0.0006,
+ "loss": 4.905491828918457,
+ "step": 802
+ },
+ {
+ "epoch": 11.153711790393013,
+ "grad_norm": 0.01720350608229637,
+ "learning_rate": 0.0006,
+ "loss": 4.798816680908203,
+ "step": 803
+ },
+ {
+ "epoch": 11.167685589519651,
+ "grad_norm": 0.017002852633595467,
+ "learning_rate": 0.0006,
+ "loss": 4.851859092712402,
+ "step": 804
+ },
+ {
+ "epoch": 11.181659388646288,
+ "grad_norm": 0.015594680793583393,
+ "learning_rate": 0.0006,
+ "loss": 4.90328311920166,
+ "step": 805
+ },
+ {
+ "epoch": 11.195633187772925,
+ "grad_norm": 0.014904462732374668,
+ "learning_rate": 0.0006,
+ "loss": 4.904215335845947,
+ "step": 806
+ },
+ {
+ "epoch": 11.209606986899564,
+ "grad_norm": 0.01366228237748146,
+ "learning_rate": 0.0006,
+ "loss": 4.9813055992126465,
+ "step": 807
+ },
+ {
+ "epoch": 11.2235807860262,
+ "grad_norm": 0.014719638973474503,
+ "learning_rate": 0.0006,
+ "loss": 4.927306175231934,
+ "step": 808
+ },
+ {
+ "epoch": 11.237554585152838,
+ "grad_norm": 0.015016036108136177,
+ "learning_rate": 0.0006,
+ "loss": 4.874215126037598,
+ "step": 809
+ },
+ {
+ "epoch": 11.251528384279476,
+ "grad_norm": 0.014925424009561539,
+ "learning_rate": 0.0006,
+ "loss": 4.838257789611816,
+ "step": 810
+ },
+ {
+ "epoch": 11.265502183406113,
+ "grad_norm": 0.014335823245346546,
+ "learning_rate": 0.0006,
+ "loss": 4.940608024597168,
+ "step": 811
+ },
+ {
+ "epoch": 11.279475982532752,
+ "grad_norm": 0.014987512491643429,
+ "learning_rate": 0.0006,
+ "loss": 4.831531524658203,
+ "step": 812
+ },
+ {
+ "epoch": 11.293449781659389,
+ "grad_norm": 0.016087457537651062,
+ "learning_rate": 0.0006,
+ "loss": 4.863227367401123,
+ "step": 813
+ },
+ {
+ "epoch": 11.307423580786025,
+ "grad_norm": 0.01671300269663334,
+ "learning_rate": 0.0006,
+ "loss": 4.8535661697387695,
+ "step": 814
+ },
+ {
+ "epoch": 11.321397379912664,
+ "grad_norm": 0.015229142270982265,
+ "learning_rate": 0.0006,
+ "loss": 4.948635101318359,
+ "step": 815
+ },
+ {
+ "epoch": 11.335371179039301,
+ "grad_norm": 0.01532587967813015,
+ "learning_rate": 0.0006,
+ "loss": 4.9035797119140625,
+ "step": 816
+ },
+ {
+ "epoch": 11.34934497816594,
+ "grad_norm": 0.015715476125478745,
+ "learning_rate": 0.0006,
+ "loss": 4.894709587097168,
+ "step": 817
+ },
+ {
+ "epoch": 11.363318777292577,
+ "grad_norm": 0.014156977646052837,
+ "learning_rate": 0.0006,
+ "loss": 4.960206985473633,
+ "step": 818
+ },
+ {
+ "epoch": 11.377292576419213,
+ "grad_norm": 0.017076566815376282,
+ "learning_rate": 0.0006,
+ "loss": 5.005850791931152,
+ "step": 819
+ },
+ {
+ "epoch": 11.391266375545852,
+ "grad_norm": 0.0172084029763937,
+ "learning_rate": 0.0006,
+ "loss": 4.903355598449707,
+ "step": 820
+ },
+ {
+ "epoch": 11.405240174672489,
+ "grad_norm": 0.0180258359760046,
+ "learning_rate": 0.0006,
+ "loss": 4.842667579650879,
+ "step": 821
+ },
+ {
+ "epoch": 11.419213973799126,
+ "grad_norm": 0.022130120545625687,
+ "learning_rate": 0.0006,
+ "loss": 4.799970626831055,
+ "step": 822
+ },
+ {
+ "epoch": 11.433187772925764,
+ "grad_norm": 0.02384897693991661,
+ "learning_rate": 0.0006,
+ "loss": 4.860746383666992,
+ "step": 823
+ },
+ {
+ "epoch": 11.447161572052401,
+ "grad_norm": 0.021881572902202606,
+ "learning_rate": 0.0006,
+ "loss": 4.918107032775879,
+ "step": 824
+ },
+ {
+ "epoch": 11.46113537117904,
+ "grad_norm": 0.023085080087184906,
+ "learning_rate": 0.0006,
+ "loss": 4.878946304321289,
+ "step": 825
+ },
+ {
+ "epoch": 11.475109170305677,
+ "grad_norm": 0.02291012555360794,
+ "learning_rate": 0.0006,
+ "loss": 4.958887100219727,
+ "step": 826
+ },
+ {
+ "epoch": 11.489082969432314,
+ "grad_norm": 0.0199968870729208,
+ "learning_rate": 0.0006,
+ "loss": 4.80955696105957,
+ "step": 827
+ },
+ {
+ "epoch": 11.503056768558952,
+ "grad_norm": 0.018702229484915733,
+ "learning_rate": 0.0006,
+ "loss": 5.0276408195495605,
+ "step": 828
+ },
+ {
+ "epoch": 11.51703056768559,
+ "grad_norm": 0.017787311226129532,
+ "learning_rate": 0.0006,
+ "loss": 4.96018123626709,
+ "step": 829
+ },
+ {
+ "epoch": 11.531004366812226,
+ "grad_norm": 0.01704501546919346,
+ "learning_rate": 0.0006,
+ "loss": 4.914553165435791,
+ "step": 830
+ },
+ {
+ "epoch": 11.544978165938865,
+ "grad_norm": 0.017519677057862282,
+ "learning_rate": 0.0006,
+ "loss": 4.798098087310791,
+ "step": 831
+ },
+ {
+ "epoch": 11.558951965065502,
+ "grad_norm": 0.018511991947889328,
+ "learning_rate": 0.0006,
+ "loss": 4.817785263061523,
+ "step": 832
+ },
+ {
+ "epoch": 11.57292576419214,
+ "grad_norm": 0.01876644790172577,
+ "learning_rate": 0.0006,
+ "loss": 4.818995475769043,
+ "step": 833
+ },
+ {
+ "epoch": 11.586899563318777,
+ "grad_norm": 0.01753568835556507,
+ "learning_rate": 0.0006,
+ "loss": 4.811938285827637,
+ "step": 834
+ },
+ {
+ "epoch": 11.600873362445414,
+ "grad_norm": 0.017046531662344933,
+ "learning_rate": 0.0006,
+ "loss": 4.914674282073975,
+ "step": 835
+ },
+ {
+ "epoch": 11.614847161572053,
+ "grad_norm": 0.01802094094455242,
+ "learning_rate": 0.0006,
+ "loss": 4.993284225463867,
+ "step": 836
+ },
+ {
+ "epoch": 11.62882096069869,
+ "grad_norm": 0.019493764266371727,
+ "learning_rate": 0.0006,
+ "loss": 4.895628452301025,
+ "step": 837
+ },
+ {
+ "epoch": 11.642794759825328,
+ "grad_norm": 0.019461210817098618,
+ "learning_rate": 0.0006,
+ "loss": 4.869269371032715,
+ "step": 838
+ },
+ {
+ "epoch": 11.656768558951965,
+ "grad_norm": 0.01981567218899727,
+ "learning_rate": 0.0006,
+ "loss": 4.863052845001221,
+ "step": 839
+ },
+ {
+ "epoch": 11.670742358078602,
+ "grad_norm": 0.019894491881132126,
+ "learning_rate": 0.0006,
+ "loss": 4.8775787353515625,
+ "step": 840
+ },
+ {
+ "epoch": 11.68471615720524,
+ "grad_norm": 0.01875786855816841,
+ "learning_rate": 0.0006,
+ "loss": 4.817763328552246,
+ "step": 841
+ },
+ {
+ "epoch": 11.698689956331878,
+ "grad_norm": 0.020926695317029953,
+ "learning_rate": 0.0006,
+ "loss": 4.7470526695251465,
+ "step": 842
+ },
+ {
+ "epoch": 11.712663755458514,
+ "grad_norm": 0.021477103233337402,
+ "learning_rate": 0.0006,
+ "loss": 4.762682914733887,
+ "step": 843
+ },
+ {
+ "epoch": 11.726637554585153,
+ "grad_norm": 0.020628711208701134,
+ "learning_rate": 0.0006,
+ "loss": 4.807985782623291,
+ "step": 844
+ },
+ {
+ "epoch": 11.74061135371179,
+ "grad_norm": 0.021489612758159637,
+ "learning_rate": 0.0006,
+ "loss": 4.843091011047363,
+ "step": 845
+ },
+ {
+ "epoch": 11.754585152838429,
+ "grad_norm": 0.021194491535425186,
+ "learning_rate": 0.0006,
+ "loss": 4.8300251960754395,
+ "step": 846
+ },
+ {
+ "epoch": 11.768558951965066,
+ "grad_norm": 0.01833350397646427,
+ "learning_rate": 0.0006,
+ "loss": 5.000202178955078,
+ "step": 847
+ },
+ {
+ "epoch": 11.782532751091702,
+ "grad_norm": 0.018096931278705597,
+ "learning_rate": 0.0006,
+ "loss": 4.880648136138916,
+ "step": 848
+ },
+ {
+ "epoch": 11.796506550218341,
+ "grad_norm": 0.0197161752730608,
+ "learning_rate": 0.0006,
+ "loss": 4.788316249847412,
+ "step": 849
+ },
+ {
+ "epoch": 11.810480349344978,
+ "grad_norm": 0.0203181654214859,
+ "learning_rate": 0.0006,
+ "loss": 4.870944976806641,
+ "step": 850
+ },
+ {
+ "epoch": 11.824454148471617,
+ "grad_norm": 0.01908455230295658,
+ "learning_rate": 0.0006,
+ "loss": 4.890144348144531,
+ "step": 851
+ },
+ {
+ "epoch": 11.838427947598253,
+ "grad_norm": 0.021281801164150238,
+ "learning_rate": 0.0006,
+ "loss": 4.89578914642334,
+ "step": 852
+ },
+ {
+ "epoch": 11.85240174672489,
+ "grad_norm": 0.02124273031949997,
+ "learning_rate": 0.0006,
+ "loss": 4.883749008178711,
+ "step": 853
+ },
+ {
+ "epoch": 11.866375545851529,
+ "grad_norm": 0.020969398319721222,
+ "learning_rate": 0.0006,
+ "loss": 4.809142589569092,
+ "step": 854
+ },
+ {
+ "epoch": 11.880349344978166,
+ "grad_norm": 0.01967657171189785,
+ "learning_rate": 0.0006,
+ "loss": 4.821864604949951,
+ "step": 855
+ },
+ {
+ "epoch": 11.894323144104803,
+ "grad_norm": 0.020887333899736404,
+ "learning_rate": 0.0006,
+ "loss": 4.917985916137695,
+ "step": 856
+ },
+ {
+ "epoch": 11.908296943231441,
+ "grad_norm": 0.019203083589673042,
+ "learning_rate": 0.0006,
+ "loss": 4.824960708618164,
+ "step": 857
+ },
+ {
+ "epoch": 11.922270742358078,
+ "grad_norm": 0.020162401720881462,
+ "learning_rate": 0.0006,
+ "loss": 4.772617816925049,
+ "step": 858
+ },
+ {
+ "epoch": 11.936244541484717,
+ "grad_norm": 0.02251487225294113,
+ "learning_rate": 0.0006,
+ "loss": 4.735418319702148,
+ "step": 859
+ },
+ {
+ "epoch": 11.950218340611354,
+ "grad_norm": 0.020313527435064316,
+ "learning_rate": 0.0006,
+ "loss": 4.939233779907227,
+ "step": 860
+ },
+ {
+ "epoch": 11.96419213973799,
+ "grad_norm": 0.021635599434375763,
+ "learning_rate": 0.0006,
+ "loss": 4.861807346343994,
+ "step": 861
+ },
+ {
+ "epoch": 11.97816593886463,
+ "grad_norm": 0.02285311557352543,
+ "learning_rate": 0.0006,
+ "loss": 4.932732582092285,
+ "step": 862
+ },
+ {
+ "epoch": 11.992139737991266,
+ "grad_norm": 0.02072535641491413,
+ "learning_rate": 0.0006,
+ "loss": 4.796767711639404,
+ "step": 863
+ },
+ {
+ "epoch": 12.0,
+ "grad_norm": 0.020588304847478867,
+ "learning_rate": 0.0006,
+ "loss": 4.997537612915039,
+ "step": 864
+ },
+ {
+ "epoch": 12.0,
+ "eval_loss": 5.044651985168457,
+ "eval_runtime": 56.9145,
+ "eval_samples_per_second": 42.906,
+ "eval_steps_per_second": 1.353,
+ "step": 864
+ },
+ {
+ "epoch": 12.013973799126637,
+ "grad_norm": 0.017597166821360588,
+ "learning_rate": 0.0006,
+ "loss": 4.830524444580078,
+ "step": 865
+ },
+ {
+ "epoch": 12.027947598253276,
+ "grad_norm": 0.01716788113117218,
+ "learning_rate": 0.0006,
+ "loss": 4.894325256347656,
+ "step": 866
+ },
+ {
+ "epoch": 12.041921397379912,
+ "grad_norm": 0.017712457105517387,
+ "learning_rate": 0.0006,
+ "loss": 4.868297576904297,
+ "step": 867
+ },
+ {
+ "epoch": 12.055895196506551,
+ "grad_norm": 0.021486390382051468,
+ "learning_rate": 0.0006,
+ "loss": 4.922049522399902,
+ "step": 868
+ },
+ {
+ "epoch": 12.069868995633188,
+ "grad_norm": 0.024560654535889626,
+ "learning_rate": 0.0006,
+ "loss": 4.734241008758545,
+ "step": 869
+ },
+ {
+ "epoch": 12.083842794759825,
+ "grad_norm": 0.025870252400636673,
+ "learning_rate": 0.0006,
+ "loss": 4.868566989898682,
+ "step": 870
+ },
+ {
+ "epoch": 12.097816593886463,
+ "grad_norm": 0.0277389045804739,
+ "learning_rate": 0.0006,
+ "loss": 4.755006790161133,
+ "step": 871
+ },
+ {
+ "epoch": 12.1117903930131,
+ "grad_norm": 0.029100844636559486,
+ "learning_rate": 0.0006,
+ "loss": 4.909667015075684,
+ "step": 872
+ },
+ {
+ "epoch": 12.125764192139737,
+ "grad_norm": 0.031833067536354065,
+ "learning_rate": 0.0006,
+ "loss": 4.858674049377441,
+ "step": 873
+ },
+ {
+ "epoch": 12.139737991266376,
+ "grad_norm": 0.032709237188100815,
+ "learning_rate": 0.0006,
+ "loss": 4.823696136474609,
+ "step": 874
+ },
+ {
+ "epoch": 12.153711790393013,
+ "grad_norm": 0.02995380386710167,
+ "learning_rate": 0.0006,
+ "loss": 4.850235462188721,
+ "step": 875
+ },
+ {
+ "epoch": 12.167685589519651,
+ "grad_norm": 0.034098610281944275,
+ "learning_rate": 0.0006,
+ "loss": 4.8169989585876465,
+ "step": 876
+ },
+ {
+ "epoch": 12.181659388646288,
+ "grad_norm": 0.03377080336213112,
+ "learning_rate": 0.0006,
+ "loss": 4.90738582611084,
+ "step": 877
+ },
+ {
+ "epoch": 12.195633187772925,
+ "grad_norm": 0.029780825600028038,
+ "learning_rate": 0.0006,
+ "loss": 4.811163902282715,
+ "step": 878
+ },
+ {
+ "epoch": 12.209606986899564,
+ "grad_norm": 0.03118024580180645,
+ "learning_rate": 0.0006,
+ "loss": 4.861286640167236,
+ "step": 879
+ },
+ {
+ "epoch": 12.2235807860262,
+ "grad_norm": 0.033052004873752594,
+ "learning_rate": 0.0006,
+ "loss": 4.865021228790283,
+ "step": 880
+ },
+ {
+ "epoch": 12.237554585152838,
+ "grad_norm": 0.029070032760500908,
+ "learning_rate": 0.0006,
+ "loss": 4.947303771972656,
+ "step": 881
+ },
+ {
+ "epoch": 12.251528384279476,
+ "grad_norm": 0.028355715796351433,
+ "learning_rate": 0.0006,
+ "loss": 4.804648399353027,
+ "step": 882
+ },
+ {
+ "epoch": 12.265502183406113,
+ "grad_norm": 0.031257227063179016,
+ "learning_rate": 0.0006,
+ "loss": 4.757718086242676,
+ "step": 883
+ },
+ {
+ "epoch": 12.279475982532752,
+ "grad_norm": 0.031438373029232025,
+ "learning_rate": 0.0006,
+ "loss": 4.851268768310547,
+ "step": 884
+ },
+ {
+ "epoch": 12.293449781659389,
+ "grad_norm": 0.030492138117551804,
+ "learning_rate": 0.0006,
+ "loss": 4.7891035079956055,
+ "step": 885
+ },
+ {
+ "epoch": 12.307423580786025,
+ "grad_norm": 0.026115991175174713,
+ "learning_rate": 0.0006,
+ "loss": 4.86474609375,
+ "step": 886
+ },
+ {
+ "epoch": 12.321397379912664,
+ "grad_norm": 0.024184465408325195,
+ "learning_rate": 0.0006,
+ "loss": 4.860456943511963,
+ "step": 887
+ },
+ {
+ "epoch": 12.335371179039301,
+ "grad_norm": 0.025437982752919197,
+ "learning_rate": 0.0006,
+ "loss": 4.806251049041748,
+ "step": 888
+ },
+ {
+ "epoch": 12.34934497816594,
+ "grad_norm": 0.022390231490135193,
+ "learning_rate": 0.0006,
+ "loss": 4.807086944580078,
+ "step": 889
+ },
+ {
+ "epoch": 12.363318777292577,
+ "grad_norm": 0.020174294710159302,
+ "learning_rate": 0.0006,
+ "loss": 4.78292179107666,
+ "step": 890
+ },
+ {
+ "epoch": 12.377292576419213,
+ "grad_norm": 0.019734324887394905,
+ "learning_rate": 0.0006,
+ "loss": 4.784509658813477,
+ "step": 891
+ },
+ {
+ "epoch": 12.391266375545852,
+ "grad_norm": 0.016733523458242416,
+ "learning_rate": 0.0006,
+ "loss": 4.811755180358887,
+ "step": 892
+ },
+ {
+ "epoch": 12.405240174672489,
+ "grad_norm": 0.01570146717131138,
+ "learning_rate": 0.0006,
+ "loss": 4.720088005065918,
+ "step": 893
+ },
+ {
+ "epoch": 12.419213973799126,
+ "grad_norm": 0.01601393334567547,
+ "learning_rate": 0.0006,
+ "loss": 4.783543109893799,
+ "step": 894
+ },
+ {
+ "epoch": 12.433187772925764,
+ "grad_norm": 0.014035487547516823,
+ "learning_rate": 0.0006,
+ "loss": 4.848340034484863,
+ "step": 895
+ },
+ {
+ "epoch": 12.447161572052401,
+ "grad_norm": 0.015248487703502178,
+ "learning_rate": 0.0006,
+ "loss": 4.773287773132324,
+ "step": 896
+ },
+ {
+ "epoch": 12.46113537117904,
+ "grad_norm": 0.01532268151640892,
+ "learning_rate": 0.0006,
+ "loss": 4.799646854400635,
+ "step": 897
+ },
+ {
+ "epoch": 12.475109170305677,
+ "grad_norm": 0.013982534408569336,
+ "learning_rate": 0.0006,
+ "loss": 4.8090500831604,
+ "step": 898
+ },
+ {
+ "epoch": 12.489082969432314,
+ "grad_norm": 0.014081164263188839,
+ "learning_rate": 0.0006,
+ "loss": 4.751888751983643,
+ "step": 899
+ },
+ {
+ "epoch": 12.503056768558952,
+ "grad_norm": 0.01380133256316185,
+ "learning_rate": 0.0006,
+ "loss": 4.78144645690918,
+ "step": 900
+ },
+ {
+ "epoch": 12.51703056768559,
+ "grad_norm": 0.013555269688367844,
+ "learning_rate": 0.0006,
+ "loss": 4.779257774353027,
+ "step": 901
+ },
+ {
+ "epoch": 12.531004366812226,
+ "grad_norm": 0.013075701892375946,
+ "learning_rate": 0.0006,
+ "loss": 4.820705413818359,
+ "step": 902
+ },
+ {
+ "epoch": 12.544978165938865,
+ "grad_norm": 0.01276316400617361,
+ "learning_rate": 0.0006,
+ "loss": 4.836733818054199,
+ "step": 903
+ },
+ {
+ "epoch": 12.558951965065502,
+ "grad_norm": 0.013033892959356308,
+ "learning_rate": 0.0006,
+ "loss": 4.764178276062012,
+ "step": 904
+ },
+ {
+ "epoch": 12.57292576419214,
+ "grad_norm": 0.013996980153024197,
+ "learning_rate": 0.0006,
+ "loss": 4.767983436584473,
+ "step": 905
+ },
+ {
+ "epoch": 12.586899563318777,
+ "grad_norm": 0.015222184360027313,
+ "learning_rate": 0.0006,
+ "loss": 4.759721755981445,
+ "step": 906
+ },
+ {
+ "epoch": 12.600873362445414,
+ "grad_norm": 0.01623941957950592,
+ "learning_rate": 0.0006,
+ "loss": 4.849811553955078,
+ "step": 907
+ },
+ {
+ "epoch": 12.614847161572053,
+ "grad_norm": 0.015711182728409767,
+ "learning_rate": 0.0006,
+ "loss": 4.7829084396362305,
+ "step": 908
+ },
+ {
+ "epoch": 12.62882096069869,
+ "grad_norm": 0.014137467369437218,
+ "learning_rate": 0.0006,
+ "loss": 4.73321533203125,
+ "step": 909
+ },
+ {
+ "epoch": 12.642794759825328,
+ "grad_norm": 0.012738462537527084,
+ "learning_rate": 0.0006,
+ "loss": 4.870052337646484,
+ "step": 910
+ },
+ {
+ "epoch": 12.656768558951965,
+ "grad_norm": 0.013683130033314228,
+ "learning_rate": 0.0006,
+ "loss": 4.82658052444458,
+ "step": 911
+ },
+ {
+ "epoch": 12.670742358078602,
+ "grad_norm": 0.014409830793738365,
+ "learning_rate": 0.0006,
+ "loss": 4.707494735717773,
+ "step": 912
+ },
+ {
+ "epoch": 12.68471615720524,
+ "grad_norm": 0.013806809671223164,
+ "learning_rate": 0.0006,
+ "loss": 4.778928279876709,
+ "step": 913
+ },
+ {
+ "epoch": 12.698689956331878,
+ "grad_norm": 0.012857912108302116,
+ "learning_rate": 0.0006,
+ "loss": 4.890717506408691,
+ "step": 914
+ },
+ {
+ "epoch": 12.712663755458514,
+ "grad_norm": 0.013828758150339127,
+ "learning_rate": 0.0006,
+ "loss": 4.8357439041137695,
+ "step": 915
+ },
+ {
+ "epoch": 12.726637554585153,
+ "grad_norm": 0.01478351280093193,
+ "learning_rate": 0.0006,
+ "loss": 4.869986534118652,
+ "step": 916
+ },
+ {
+ "epoch": 12.74061135371179,
+ "grad_norm": 0.01326421182602644,
+ "learning_rate": 0.0006,
+ "loss": 4.830646514892578,
+ "step": 917
+ },
+ {
+ "epoch": 12.754585152838429,
+ "grad_norm": 0.013335433788597584,
+ "learning_rate": 0.0006,
+ "loss": 4.7786149978637695,
+ "step": 918
+ },
+ {
+ "epoch": 12.768558951965066,
+ "grad_norm": 0.014261147007346153,
+ "learning_rate": 0.0006,
+ "loss": 4.788693428039551,
+ "step": 919
+ },
+ {
+ "epoch": 12.782532751091702,
+ "grad_norm": 0.014274783432483673,
+ "learning_rate": 0.0006,
+ "loss": 4.77295446395874,
+ "step": 920
+ },
+ {
+ "epoch": 12.796506550218341,
+ "grad_norm": 0.015074674971401691,
+ "learning_rate": 0.0006,
+ "loss": 4.830754280090332,
+ "step": 921
+ },
+ {
+ "epoch": 12.810480349344978,
+ "grad_norm": 0.015599515289068222,
+ "learning_rate": 0.0006,
+ "loss": 4.837759017944336,
+ "step": 922
+ },
+ {
+ "epoch": 12.824454148471617,
+ "grad_norm": 0.014595242217183113,
+ "learning_rate": 0.0006,
+ "loss": 4.882965564727783,
+ "step": 923
+ },
+ {
+ "epoch": 12.838427947598253,
+ "grad_norm": 0.014951467514038086,
+ "learning_rate": 0.0006,
+ "loss": 4.709748268127441,
+ "step": 924
+ },
+ {
+ "epoch": 12.85240174672489,
+ "grad_norm": 0.01539295818656683,
+ "learning_rate": 0.0006,
+ "loss": 4.791160583496094,
+ "step": 925
+ },
+ {
+ "epoch": 12.866375545851529,
+ "grad_norm": 0.015987906605005264,
+ "learning_rate": 0.0006,
+ "loss": 4.661876678466797,
+ "step": 926
+ },
+ {
+ "epoch": 12.880349344978166,
+ "grad_norm": 0.016527341678738594,
+ "learning_rate": 0.0006,
+ "loss": 4.721952438354492,
+ "step": 927
+ },
+ {
+ "epoch": 12.894323144104803,
+ "grad_norm": 0.01637161523103714,
+ "learning_rate": 0.0006,
+ "loss": 4.704446792602539,
+ "step": 928
+ },
+ {
+ "epoch": 12.908296943231441,
+ "grad_norm": 0.015221050009131432,
+ "learning_rate": 0.0006,
+ "loss": 4.771907806396484,
+ "step": 929
+ },
+ {
+ "epoch": 12.922270742358078,
+ "grad_norm": 0.016242166981101036,
+ "learning_rate": 0.0006,
+ "loss": 4.806267738342285,
+ "step": 930
+ },
+ {
+ "epoch": 12.936244541484717,
+ "grad_norm": 0.01628357358276844,
+ "learning_rate": 0.0006,
+ "loss": 4.7252326011657715,
+ "step": 931
+ },
+ {
+ "epoch": 12.950218340611354,
+ "grad_norm": 0.017428133636713028,
+ "learning_rate": 0.0006,
+ "loss": 4.765689849853516,
+ "step": 932
+ },
+ {
+ "epoch": 12.96419213973799,
+ "grad_norm": 0.01870061084628105,
+ "learning_rate": 0.0006,
+ "loss": 4.814488887786865,
+ "step": 933
+ },
+ {
+ "epoch": 12.97816593886463,
+ "grad_norm": 0.019796060398221016,
+ "learning_rate": 0.0006,
+ "loss": 4.739185810089111,
+ "step": 934
+ },
+ {
+ "epoch": 12.992139737991266,
+ "grad_norm": 0.01985820196568966,
+ "learning_rate": 0.0006,
+ "loss": 4.810608863830566,
+ "step": 935
+ },
+ {
+ "epoch": 13.0,
+ "grad_norm": 0.020834477618336678,
+ "learning_rate": 0.0006,
+ "loss": 4.727858066558838,
+ "step": 936
+ },
+ {
+ "epoch": 13.0,
+ "eval_loss": 4.979116916656494,
+ "eval_runtime": 57.2877,
+ "eval_samples_per_second": 42.627,
+ "eval_steps_per_second": 1.344,
+ "step": 936
+ },
+ {
+ "epoch": 13.013973799126637,
+ "grad_norm": 0.02025519125163555,
+ "learning_rate": 0.0006,
+ "loss": 4.628875732421875,
+ "step": 937
+ },
+ {
+ "epoch": 13.027947598253276,
+ "grad_norm": 0.019596368074417114,
+ "learning_rate": 0.0006,
+ "loss": 4.694699287414551,
+ "step": 938
+ },
+ {
+ "epoch": 13.041921397379912,
+ "grad_norm": 0.020494934171438217,
+ "learning_rate": 0.0006,
+ "loss": 4.691821098327637,
+ "step": 939
+ },
+ {
+ "epoch": 13.055895196506551,
+ "grad_norm": 0.020121760666370392,
+ "learning_rate": 0.0006,
+ "loss": 4.71478271484375,
+ "step": 940
+ },
+ {
+ "epoch": 13.069868995633188,
+ "grad_norm": 0.021091405302286148,
+ "learning_rate": 0.0006,
+ "loss": 4.695915222167969,
+ "step": 941
+ },
+ {
+ "epoch": 13.083842794759825,
+ "grad_norm": 0.019787423312664032,
+ "learning_rate": 0.0006,
+ "loss": 4.8652448654174805,
+ "step": 942
+ },
+ {
+ "epoch": 13.097816593886463,
+ "grad_norm": 0.0185833927243948,
+ "learning_rate": 0.0006,
+ "loss": 4.7065887451171875,
+ "step": 943
+ },
+ {
+ "epoch": 13.1117903930131,
+ "grad_norm": 0.021220475435256958,
+ "learning_rate": 0.0006,
+ "loss": 4.741232395172119,
+ "step": 944
+ },
+ {
+ "epoch": 13.125764192139737,
+ "grad_norm": 0.02136622555553913,
+ "learning_rate": 0.0006,
+ "loss": 4.705967903137207,
+ "step": 945
+ },
+ {
+ "epoch": 13.139737991266376,
+ "grad_norm": 0.0191587433218956,
+ "learning_rate": 0.0006,
+ "loss": 4.758667945861816,
+ "step": 946
+ },
+ {
+ "epoch": 13.153711790393013,
+ "grad_norm": 0.018290022388100624,
+ "learning_rate": 0.0006,
+ "loss": 4.760666847229004,
+ "step": 947
+ },
+ {
+ "epoch": 13.167685589519651,
+ "grad_norm": 0.021704409271478653,
+ "learning_rate": 0.0006,
+ "loss": 4.750555992126465,
+ "step": 948
+ },
+ {
+ "epoch": 13.181659388646288,
+ "grad_norm": 0.021668918430805206,
+ "learning_rate": 0.0006,
+ "loss": 4.840903282165527,
+ "step": 949
+ },
+ {
+ "epoch": 13.195633187772925,
+ "grad_norm": 0.019497141242027283,
+ "learning_rate": 0.0006,
+ "loss": 4.738969326019287,
+ "step": 950
+ },
+ {
+ "epoch": 13.209606986899564,
+ "grad_norm": 0.018500596284866333,
+ "learning_rate": 0.0006,
+ "loss": 4.75154972076416,
+ "step": 951
+ },
+ {
+ "epoch": 13.2235807860262,
+ "grad_norm": 0.01872239261865616,
+ "learning_rate": 0.0006,
+ "loss": 4.811738014221191,
+ "step": 952
+ },
+ {
+ "epoch": 13.237554585152838,
+ "grad_norm": 0.020321547985076904,
+ "learning_rate": 0.0006,
+ "loss": 4.815384387969971,
+ "step": 953
+ },
+ {
+ "epoch": 13.251528384279476,
+ "grad_norm": 0.02283575013279915,
+ "learning_rate": 0.0006,
+ "loss": 4.74582052230835,
+ "step": 954
+ },
+ {
+ "epoch": 13.265502183406113,
+ "grad_norm": 0.02474398910999298,
+ "learning_rate": 0.0006,
+ "loss": 4.738375186920166,
+ "step": 955
+ },
+ {
+ "epoch": 13.279475982532752,
+ "grad_norm": 0.02065764181315899,
+ "learning_rate": 0.0006,
+ "loss": 4.746875286102295,
+ "step": 956
+ },
+ {
+ "epoch": 13.293449781659389,
+ "grad_norm": 0.020534520968794823,
+ "learning_rate": 0.0006,
+ "loss": 4.656914710998535,
+ "step": 957
+ },
+ {
+ "epoch": 13.307423580786025,
+ "grad_norm": 0.0206378772854805,
+ "learning_rate": 0.0006,
+ "loss": 4.842615604400635,
+ "step": 958
+ },
+ {
+ "epoch": 13.321397379912664,
+ "grad_norm": 0.017935654148459435,
+ "learning_rate": 0.0006,
+ "loss": 4.717827796936035,
+ "step": 959
+ },
+ {
+ "epoch": 13.335371179039301,
+ "grad_norm": 0.016958363354206085,
+ "learning_rate": 0.0006,
+ "loss": 4.681070327758789,
+ "step": 960
+ },
+ {
+ "epoch": 13.34934497816594,
+ "grad_norm": 0.015735412016510963,
+ "learning_rate": 0.0006,
+ "loss": 4.7957048416137695,
+ "step": 961
+ },
+ {
+ "epoch": 13.363318777292577,
+ "grad_norm": 0.01563699170947075,
+ "learning_rate": 0.0006,
+ "loss": 4.701192378997803,
+ "step": 962
+ },
+ {
+ "epoch": 13.377292576419213,
+ "grad_norm": 0.015207760035991669,
+ "learning_rate": 0.0006,
+ "loss": 4.73256254196167,
+ "step": 963
+ },
+ {
+ "epoch": 13.391266375545852,
+ "grad_norm": 0.015396883711218834,
+ "learning_rate": 0.0006,
+ "loss": 4.76412296295166,
+ "step": 964
+ },
+ {
+ "epoch": 13.405240174672489,
+ "grad_norm": 0.015690583735704422,
+ "learning_rate": 0.0006,
+ "loss": 4.791157245635986,
+ "step": 965
+ },
+ {
+ "epoch": 13.419213973799126,
+ "grad_norm": 0.01755724661052227,
+ "learning_rate": 0.0006,
+ "loss": 4.845666885375977,
+ "step": 966
+ },
+ {
+ "epoch": 13.433187772925764,
+ "grad_norm": 0.019443422555923462,
+ "learning_rate": 0.0006,
+ "loss": 4.5902299880981445,
+ "step": 967
+ },
+ {
+ "epoch": 13.447161572052401,
+ "grad_norm": 0.02118338644504547,
+ "learning_rate": 0.0006,
+ "loss": 4.740627288818359,
+ "step": 968
+ },
+ {
+ "epoch": 13.46113537117904,
+ "grad_norm": 0.02174839936196804,
+ "learning_rate": 0.0006,
+ "loss": 4.814560890197754,
+ "step": 969
+ },
+ {
+ "epoch": 13.475109170305677,
+ "grad_norm": 0.022860374301671982,
+ "learning_rate": 0.0006,
+ "loss": 4.66440486907959,
+ "step": 970
+ },
+ {
+ "epoch": 13.489082969432314,
+ "grad_norm": 0.024742677807807922,
+ "learning_rate": 0.0006,
+ "loss": 4.7451653480529785,
+ "step": 971
+ },
+ {
+ "epoch": 13.503056768558952,
+ "grad_norm": 0.027068907395005226,
+ "learning_rate": 0.0006,
+ "loss": 4.802584171295166,
+ "step": 972
+ },
+ {
+ "epoch": 13.51703056768559,
+ "grad_norm": 0.026316625997424126,
+ "learning_rate": 0.0006,
+ "loss": 4.842397689819336,
+ "step": 973
+ },
+ {
+ "epoch": 13.531004366812226,
+ "grad_norm": 0.024610962718725204,
+ "learning_rate": 0.0006,
+ "loss": 4.650721073150635,
+ "step": 974
+ },
+ {
+ "epoch": 13.544978165938865,
+ "grad_norm": 0.024122603237628937,
+ "learning_rate": 0.0006,
+ "loss": 4.6984357833862305,
+ "step": 975
+ },
+ {
+ "epoch": 13.558951965065502,
+ "grad_norm": 0.022699708119034767,
+ "learning_rate": 0.0006,
+ "loss": 4.793007850646973,
+ "step": 976
+ },
+ {
+ "epoch": 13.57292576419214,
+ "grad_norm": 0.021706534549593925,
+ "learning_rate": 0.0006,
+ "loss": 4.691362380981445,
+ "step": 977
+ },
+ {
+ "epoch": 13.586899563318777,
+ "grad_norm": 0.021363291889429092,
+ "learning_rate": 0.0006,
+ "loss": 4.76762580871582,
+ "step": 978
+ },
+ {
+ "epoch": 13.600873362445414,
+ "grad_norm": 0.023476677015423775,
+ "learning_rate": 0.0006,
+ "loss": 4.777202606201172,
+ "step": 979
+ },
+ {
+ "epoch": 13.614847161572053,
+ "grad_norm": 0.02457100711762905,
+ "learning_rate": 0.0006,
+ "loss": 4.590703010559082,
+ "step": 980
+ },
+ {
+ "epoch": 13.62882096069869,
+ "grad_norm": 0.02469099499285221,
+ "learning_rate": 0.0006,
+ "loss": 4.640185356140137,
+ "step": 981
+ },
+ {
+ "epoch": 13.642794759825328,
+ "grad_norm": 0.024914603680372238,
+ "learning_rate": 0.0006,
+ "loss": 4.772201061248779,
+ "step": 982
+ },
+ {
+ "epoch": 13.656768558951965,
+ "grad_norm": 0.025133173912763596,
+ "learning_rate": 0.0006,
+ "loss": 4.743337154388428,
+ "step": 983
+ },
+ {
+ "epoch": 13.670742358078602,
+ "grad_norm": 0.02034137211740017,
+ "learning_rate": 0.0006,
+ "loss": 4.752127647399902,
+ "step": 984
+ },
+ {
+ "epoch": 13.68471615720524,
+ "grad_norm": 0.019645661115646362,
+ "learning_rate": 0.0006,
+ "loss": 4.838453769683838,
+ "step": 985
+ },
+ {
+ "epoch": 13.698689956331878,
+ "grad_norm": 0.01900297962129116,
+ "learning_rate": 0.0006,
+ "loss": 4.804868221282959,
+ "step": 986
+ },
+ {
+ "epoch": 13.712663755458514,
+ "grad_norm": 0.020292341709136963,
+ "learning_rate": 0.0006,
+ "loss": 4.609743118286133,
+ "step": 987
+ },
+ {
+ "epoch": 13.726637554585153,
+ "grad_norm": 0.017014775425195694,
+ "learning_rate": 0.0006,
+ "loss": 4.740612983703613,
+ "step": 988
+ },
+ {
+ "epoch": 13.74061135371179,
+ "grad_norm": 0.015199844725430012,
+ "learning_rate": 0.0006,
+ "loss": 4.719701766967773,
+ "step": 989
+ },
+ {
+ "epoch": 13.754585152838429,
+ "grad_norm": 0.015125414356589317,
+ "learning_rate": 0.0006,
+ "loss": 4.686114311218262,
+ "step": 990
+ },
+ {
+ "epoch": 13.768558951965066,
+ "grad_norm": 0.014779540710151196,
+ "learning_rate": 0.0006,
+ "loss": 4.674347877502441,
+ "step": 991
+ },
+ {
+ "epoch": 13.782532751091702,
+ "grad_norm": 0.015255163423717022,
+ "learning_rate": 0.0006,
+ "loss": 4.6309309005737305,
+ "step": 992
+ },
+ {
+ "epoch": 13.796506550218341,
+ "grad_norm": 0.013692197389900684,
+ "learning_rate": 0.0006,
+ "loss": 4.698497295379639,
+ "step": 993
+ },
+ {
+ "epoch": 13.810480349344978,
+ "grad_norm": 0.014529623091220856,
+ "learning_rate": 0.0006,
+ "loss": 4.628443717956543,
+ "step": 994
+ },
+ {
+ "epoch": 13.824454148471617,
+ "grad_norm": 0.01576659083366394,
+ "learning_rate": 0.0006,
+ "loss": 4.788015842437744,
+ "step": 995
+ },
+ {
+ "epoch": 13.838427947598253,
+ "grad_norm": 0.015834277495741844,
+ "learning_rate": 0.0006,
+ "loss": 4.672703742980957,
+ "step": 996
+ },
+ {
+ "epoch": 13.85240174672489,
+ "grad_norm": 0.015539503656327724,
+ "learning_rate": 0.0006,
+ "loss": 4.731503486633301,
+ "step": 997
+ },
+ {
+ "epoch": 13.866375545851529,
+ "grad_norm": 0.013463746756315231,
+ "learning_rate": 0.0006,
+ "loss": 4.7006330490112305,
+ "step": 998
+ },
+ {
+ "epoch": 13.880349344978166,
+ "grad_norm": 0.016377726569771767,
+ "learning_rate": 0.0006,
+ "loss": 4.636074066162109,
+ "step": 999
+ },
+ {
+ "epoch": 13.894323144104803,
+ "grad_norm": 0.016660034656524658,
+ "learning_rate": 0.0006,
+ "loss": 4.742693901062012,
+ "step": 1000
+ },
+ {
+ "epoch": 13.908296943231441,
+ "grad_norm": 0.014902938157320023,
+ "learning_rate": 0.0006,
+ "loss": 4.769811630249023,
+ "step": 1001
+ },
+ {
+ "epoch": 13.922270742358078,
+ "grad_norm": 0.014883476309478283,
+ "learning_rate": 0.0006,
+ "loss": 4.731159210205078,
+ "step": 1002
+ },
+ {
+ "epoch": 13.936244541484717,
+ "grad_norm": 0.014494217932224274,
+ "learning_rate": 0.0006,
+ "loss": 4.659485816955566,
+ "step": 1003
+ },
+ {
+ "epoch": 13.950218340611354,
+ "grad_norm": 0.015426991507411003,
+ "learning_rate": 0.0006,
+ "loss": 4.755229949951172,
+ "step": 1004
+ },
+ {
+ "epoch": 13.96419213973799,
+ "grad_norm": 0.016648586839437485,
+ "learning_rate": 0.0006,
+ "loss": 4.74111270904541,
+ "step": 1005
+ },
+ {
+ "epoch": 13.97816593886463,
+ "grad_norm": 0.016980541869997978,
+ "learning_rate": 0.0006,
+ "loss": 4.74072265625,
+ "step": 1006
+ },
+ {
+ "epoch": 13.992139737991266,
+ "grad_norm": 0.018070152029395103,
+ "learning_rate": 0.0006,
+ "loss": 4.732975006103516,
+ "step": 1007
+ },
+ {
+ "epoch": 14.0,
+ "grad_norm": 0.019910665228962898,
+ "learning_rate": 0.0006,
+ "loss": 4.634697914123535,
+ "step": 1008
+ },
+ {
+ "epoch": 14.0,
+ "eval_loss": 4.850640296936035,
+ "eval_runtime": 57.3039,
+ "eval_samples_per_second": 42.615,
+ "eval_steps_per_second": 1.344,
+ "step": 1008
+ },
+ {
+ "epoch": 14.013973799126637,
+ "grad_norm": 0.01980053074657917,
+ "learning_rate": 0.0006,
+ "loss": 4.628744602203369,
+ "step": 1009
+ },
+ {
+ "epoch": 14.027947598253276,
+ "grad_norm": 0.02497623674571514,
+ "learning_rate": 0.0006,
+ "loss": 4.75930118560791,
+ "step": 1010
+ },
+ {
+ "epoch": 14.041921397379912,
+ "grad_norm": 0.02794097363948822,
+ "learning_rate": 0.0006,
+ "loss": 4.691340446472168,
+ "step": 1011
+ },
+ {
+ "epoch": 14.055895196506551,
+ "grad_norm": 0.023725535720586777,
+ "learning_rate": 0.0006,
+ "loss": 4.733664512634277,
+ "step": 1012
+ },
+ {
+ "epoch": 14.069868995633188,
+ "grad_norm": 0.022966474294662476,
+ "learning_rate": 0.0006,
+ "loss": 4.664808750152588,
+ "step": 1013
+ },
+ {
+ "epoch": 14.083842794759825,
+ "grad_norm": 0.02530563250184059,
+ "learning_rate": 0.0006,
+ "loss": 4.8047285079956055,
+ "step": 1014
+ },
+ {
+ "epoch": 14.097816593886463,
+ "grad_norm": 0.02491082064807415,
+ "learning_rate": 0.0006,
+ "loss": 4.667956352233887,
+ "step": 1015
+ },
+ {
+ "epoch": 14.1117903930131,
+ "grad_norm": 0.022110294550657272,
+ "learning_rate": 0.0006,
+ "loss": 4.666140556335449,
+ "step": 1016
+ },
+ {
+ "epoch": 14.125764192139737,
+ "grad_norm": 0.02065849117934704,
+ "learning_rate": 0.0006,
+ "loss": 4.581960678100586,
+ "step": 1017
+ },
+ {
+ "epoch": 14.139737991266376,
+ "grad_norm": 0.01941298507153988,
+ "learning_rate": 0.0006,
+ "loss": 4.614226341247559,
+ "step": 1018
+ },
+ {
+ "epoch": 14.153711790393013,
+ "grad_norm": 0.022441856563091278,
+ "learning_rate": 0.0006,
+ "loss": 4.638352394104004,
+ "step": 1019
+ },
+ {
+ "epoch": 14.167685589519651,
+ "grad_norm": 0.021519694477319717,
+ "learning_rate": 0.0006,
+ "loss": 4.700160980224609,
+ "step": 1020
+ },
+ {
+ "epoch": 14.181659388646288,
+ "grad_norm": 0.018956486135721207,
+ "learning_rate": 0.0006,
+ "loss": 4.649593353271484,
+ "step": 1021
+ },
+ {
+ "epoch": 14.195633187772925,
+ "grad_norm": 0.016681140288710594,
+ "learning_rate": 0.0006,
+ "loss": 4.603952407836914,
+ "step": 1022
+ },
+ {
+ "epoch": 14.209606986899564,
+ "grad_norm": 0.01788809522986412,
+ "learning_rate": 0.0006,
+ "loss": 4.669160842895508,
+ "step": 1023
+ },
+ {
+ "epoch": 14.2235807860262,
+ "grad_norm": 0.016047228127717972,
+ "learning_rate": 0.0006,
+ "loss": 4.6367950439453125,
+ "step": 1024
+ },
+ {
+ "epoch": 14.237554585152838,
+ "grad_norm": 0.01614239625632763,
+ "learning_rate": 0.0006,
+ "loss": 4.664529323577881,
+ "step": 1025
+ },
+ {
+ "epoch": 14.251528384279476,
+ "grad_norm": 0.016362549737095833,
+ "learning_rate": 0.0006,
+ "loss": 4.702373504638672,
+ "step": 1026
+ },
+ {
+ "epoch": 14.265502183406113,
+ "grad_norm": 0.015800409018993378,
+ "learning_rate": 0.0006,
+ "loss": 4.7594428062438965,
+ "step": 1027
+ },
+ {
+ "epoch": 14.279475982532752,
+ "grad_norm": 0.016730139032006264,
+ "learning_rate": 0.0006,
+ "loss": 4.677244186401367,
+ "step": 1028
+ },
+ {
+ "epoch": 14.293449781659389,
+ "grad_norm": 0.01661759801208973,
+ "learning_rate": 0.0006,
+ "loss": 4.633679389953613,
+ "step": 1029
+ },
+ {
+ "epoch": 14.307423580786025,
+ "grad_norm": 0.017673347145318985,
+ "learning_rate": 0.0006,
+ "loss": 4.649107456207275,
+ "step": 1030
+ },
+ {
+ "epoch": 14.321397379912664,
+ "grad_norm": 0.019243352115154266,
+ "learning_rate": 0.0006,
+ "loss": 4.611257553100586,
+ "step": 1031
+ },
+ {
+ "epoch": 14.335371179039301,
+ "grad_norm": 0.019817164167761803,
+ "learning_rate": 0.0006,
+ "loss": 4.588086128234863,
+ "step": 1032
+ },
+ {
+ "epoch": 14.34934497816594,
+ "grad_norm": 0.022292211651802063,
+ "learning_rate": 0.0006,
+ "loss": 4.783066749572754,
+ "step": 1033
+ },
+ {
+ "epoch": 14.363318777292577,
+ "grad_norm": 0.02140096202492714,
+ "learning_rate": 0.0006,
+ "loss": 4.604718208312988,
+ "step": 1034
+ },
+ {
+ "epoch": 14.377292576419213,
+ "grad_norm": 0.0187817495316267,
+ "learning_rate": 0.0006,
+ "loss": 4.721773147583008,
+ "step": 1035
+ },
+ {
+ "epoch": 14.391266375545852,
+ "grad_norm": 0.019879210740327835,
+ "learning_rate": 0.0006,
+ "loss": 4.789286136627197,
+ "step": 1036
+ },
+ {
+ "epoch": 14.405240174672489,
+ "grad_norm": 0.01885785162448883,
+ "learning_rate": 0.0006,
+ "loss": 4.669743537902832,
+ "step": 1037
+ },
+ {
+ "epoch": 14.419213973799126,
+ "grad_norm": 0.018324270844459534,
+ "learning_rate": 0.0006,
+ "loss": 4.65901517868042,
+ "step": 1038
+ },
+ {
+ "epoch": 14.433187772925764,
+ "grad_norm": 0.017664821818470955,
+ "learning_rate": 0.0006,
+ "loss": 4.667839527130127,
+ "step": 1039
+ },
+ {
+ "epoch": 14.447161572052401,
+ "grad_norm": 0.015702905133366585,
+ "learning_rate": 0.0006,
+ "loss": 4.64700984954834,
+ "step": 1040
+ },
+ {
+ "epoch": 14.46113537117904,
+ "grad_norm": 0.016674194484949112,
+ "learning_rate": 0.0006,
+ "loss": 4.694093227386475,
+ "step": 1041
+ },
+ {
+ "epoch": 14.475109170305677,
+ "grad_norm": 0.016869477927684784,
+ "learning_rate": 0.0006,
+ "loss": 4.619544982910156,
+ "step": 1042
+ },
+ {
+ "epoch": 14.489082969432314,
+ "grad_norm": 0.016557540744543076,
+ "learning_rate": 0.0006,
+ "loss": 4.568667411804199,
+ "step": 1043
+ },
+ {
+ "epoch": 14.503056768558952,
+ "grad_norm": 0.014500945806503296,
+ "learning_rate": 0.0006,
+ "loss": 4.561741828918457,
+ "step": 1044
+ },
+ {
+ "epoch": 14.51703056768559,
+ "grad_norm": 0.014228739775717258,
+ "learning_rate": 0.0006,
+ "loss": 4.654277324676514,
+ "step": 1045
+ },
+ {
+ "epoch": 14.531004366812226,
+ "grad_norm": 0.013894238509237766,
+ "learning_rate": 0.0006,
+ "loss": 4.5867719650268555,
+ "step": 1046
+ },
+ {
+ "epoch": 14.544978165938865,
+ "grad_norm": 0.01550068985670805,
+ "learning_rate": 0.0006,
+ "loss": 4.61168098449707,
+ "step": 1047
+ },
+ {
+ "epoch": 14.558951965065502,
+ "grad_norm": 0.015940772369503975,
+ "learning_rate": 0.0006,
+ "loss": 4.691247940063477,
+ "step": 1048
+ },
+ {
+ "epoch": 14.57292576419214,
+ "grad_norm": 0.01714535616338253,
+ "learning_rate": 0.0006,
+ "loss": 4.61183500289917,
+ "step": 1049
+ },
+ {
+ "epoch": 14.586899563318777,
+ "grad_norm": 0.017925038933753967,
+ "learning_rate": 0.0006,
+ "loss": 4.719022750854492,
+ "step": 1050
+ },
+ {
+ "epoch": 14.600873362445414,
+ "grad_norm": 0.018141552805900574,
+ "learning_rate": 0.0006,
+ "loss": 4.760281562805176,
+ "step": 1051
+ },
+ {
+ "epoch": 14.614847161572053,
+ "grad_norm": 0.017131328582763672,
+ "learning_rate": 0.0006,
+ "loss": 4.709680080413818,
+ "step": 1052
+ },
+ {
+ "epoch": 14.62882096069869,
+ "grad_norm": 0.018730048090219498,
+ "learning_rate": 0.0006,
+ "loss": 4.5994720458984375,
+ "step": 1053
+ },
+ {
+ "epoch": 14.642794759825328,
+ "grad_norm": 0.018894298002123833,
+ "learning_rate": 0.0006,
+ "loss": 4.669070720672607,
+ "step": 1054
+ },
+ {
+ "epoch": 14.656768558951965,
+ "grad_norm": 0.017643198370933533,
+ "learning_rate": 0.0006,
+ "loss": 4.641790390014648,
+ "step": 1055
+ },
+ {
+ "epoch": 14.670742358078602,
+ "grad_norm": 0.017647244036197662,
+ "learning_rate": 0.0006,
+ "loss": 4.653899192810059,
+ "step": 1056
+ },
+ {
+ "epoch": 14.68471615720524,
+ "grad_norm": 0.017304198816418648,
+ "learning_rate": 0.0006,
+ "loss": 4.7487473487854,
+ "step": 1057
+ },
+ {
+ "epoch": 14.698689956331878,
+ "grad_norm": 0.01638447679579258,
+ "learning_rate": 0.0006,
+ "loss": 4.71945858001709,
+ "step": 1058
+ },
+ {
+ "epoch": 14.712663755458514,
+ "grad_norm": 0.015892071649432182,
+ "learning_rate": 0.0006,
+ "loss": 4.744903564453125,
+ "step": 1059
+ },
+ {
+ "epoch": 14.726637554585153,
+ "grad_norm": 0.017420614138245583,
+ "learning_rate": 0.0006,
+ "loss": 4.781563758850098,
+ "step": 1060
+ },
+ {
+ "epoch": 14.74061135371179,
+ "grad_norm": 0.016498520970344543,
+ "learning_rate": 0.0006,
+ "loss": 4.622779846191406,
+ "step": 1061
+ },
+ {
+ "epoch": 14.754585152838429,
+ "grad_norm": 0.017482127994298935,
+ "learning_rate": 0.0006,
+ "loss": 4.62825870513916,
+ "step": 1062
+ },
+ {
+ "epoch": 14.768558951965066,
+ "grad_norm": 0.019018379971385002,
+ "learning_rate": 0.0006,
+ "loss": 4.693194389343262,
+ "step": 1063
+ },
+ {
+ "epoch": 14.782532751091702,
+ "grad_norm": 0.01993914321064949,
+ "learning_rate": 0.0006,
+ "loss": 4.742903709411621,
+ "step": 1064
+ },
+ {
+ "epoch": 14.796506550218341,
+ "grad_norm": 0.02301335707306862,
+ "learning_rate": 0.0006,
+ "loss": 4.613251686096191,
+ "step": 1065
+ },
+ {
+ "epoch": 14.810480349344978,
+ "grad_norm": 0.024088222533464432,
+ "learning_rate": 0.0006,
+ "loss": 4.594305992126465,
+ "step": 1066
+ },
+ {
+ "epoch": 14.824454148471617,
+ "grad_norm": 0.02493320405483246,
+ "learning_rate": 0.0006,
+ "loss": 4.700588226318359,
+ "step": 1067
+ },
+ {
+ "epoch": 14.838427947598253,
+ "grad_norm": 0.02501937560737133,
+ "learning_rate": 0.0006,
+ "loss": 4.601742744445801,
+ "step": 1068
+ },
+ {
+ "epoch": 14.85240174672489,
+ "grad_norm": 0.023490751162171364,
+ "learning_rate": 0.0006,
+ "loss": 4.589211463928223,
+ "step": 1069
+ },
+ {
+ "epoch": 14.866375545851529,
+ "grad_norm": 0.019695665687322617,
+ "learning_rate": 0.0006,
+ "loss": 4.635081768035889,
+ "step": 1070
+ },
+ {
+ "epoch": 14.880349344978166,
+ "grad_norm": 0.019276980310678482,
+ "learning_rate": 0.0006,
+ "loss": 4.737002372741699,
+ "step": 1071
+ },
+ {
+ "epoch": 14.894323144104803,
+ "grad_norm": 0.022430511191487312,
+ "learning_rate": 0.0006,
+ "loss": 4.808295726776123,
+ "step": 1072
+ },
+ {
+ "epoch": 14.908296943231441,
+ "grad_norm": 0.021543916314840317,
+ "learning_rate": 0.0006,
+ "loss": 4.559475421905518,
+ "step": 1073
+ },
+ {
+ "epoch": 14.922270742358078,
+ "grad_norm": 0.019885241985321045,
+ "learning_rate": 0.0006,
+ "loss": 4.647948741912842,
+ "step": 1074
+ },
+ {
+ "epoch": 14.936244541484717,
+ "grad_norm": 0.02115175500512123,
+ "learning_rate": 0.0006,
+ "loss": 4.680505752563477,
+ "step": 1075
+ },
+ {
+ "epoch": 14.950218340611354,
+ "grad_norm": 0.02156151458621025,
+ "learning_rate": 0.0006,
+ "loss": 4.673477649688721,
+ "step": 1076
+ },
+ {
+ "epoch": 14.96419213973799,
+ "grad_norm": 0.02184601128101349,
+ "learning_rate": 0.0006,
+ "loss": 4.644402027130127,
+ "step": 1077
+ },
+ {
+ "epoch": 14.97816593886463,
+ "grad_norm": 0.0211983360350132,
+ "learning_rate": 0.0006,
+ "loss": 4.821558475494385,
+ "step": 1078
+ },
+ {
+ "epoch": 14.992139737991266,
+ "grad_norm": 0.019636567682027817,
+ "learning_rate": 0.0006,
+ "loss": 4.578868865966797,
+ "step": 1079
+ },
+ {
+ "epoch": 15.0,
+ "grad_norm": 0.019997350871562958,
+ "learning_rate": 0.0006,
+ "loss": 4.691307544708252,
+ "step": 1080
+ },
+ {
+ "epoch": 15.0,
+ "eval_loss": 4.872478008270264,
+ "eval_runtime": 56.8513,
+ "eval_samples_per_second": 42.954,
+ "eval_steps_per_second": 1.354,
+ "step": 1080
+ },
+ {
+ "epoch": 15.013973799126637,
+ "grad_norm": 0.018830662593245506,
+ "learning_rate": 0.0006,
+ "loss": 4.761779308319092,
+ "step": 1081
+ },
+ {
+ "epoch": 15.027947598253276,
+ "grad_norm": 0.016876481473445892,
+ "learning_rate": 0.0006,
+ "loss": 4.61148738861084,
+ "step": 1082
+ },
+ {
+ "epoch": 15.041921397379912,
+ "grad_norm": 0.015550863929092884,
+ "learning_rate": 0.0006,
+ "loss": 4.676124095916748,
+ "step": 1083
+ },
+ {
+ "epoch": 15.055895196506551,
+ "grad_norm": 0.017178896814584732,
+ "learning_rate": 0.0006,
+ "loss": 4.588935852050781,
+ "step": 1084
+ },
+ {
+ "epoch": 15.069868995633188,
+ "grad_norm": 0.015465447679162025,
+ "learning_rate": 0.0006,
+ "loss": 4.605499267578125,
+ "step": 1085
+ },
+ {
+ "epoch": 15.083842794759825,
+ "grad_norm": 0.01606195978820324,
+ "learning_rate": 0.0006,
+ "loss": 4.723373889923096,
+ "step": 1086
+ },
+ {
+ "epoch": 15.097816593886463,
+ "grad_norm": 0.016521936282515526,
+ "learning_rate": 0.0006,
+ "loss": 4.646608352661133,
+ "step": 1087
+ },
+ {
+ "epoch": 15.1117903930131,
+ "grad_norm": 0.01589205674827099,
+ "learning_rate": 0.0006,
+ "loss": 4.498138427734375,
+ "step": 1088
+ },
+ {
+ "epoch": 15.125764192139737,
+ "grad_norm": 0.01524856686592102,
+ "learning_rate": 0.0006,
+ "loss": 4.594562530517578,
+ "step": 1089
+ },
+ {
+ "epoch": 15.139737991266376,
+ "grad_norm": 0.014714641496539116,
+ "learning_rate": 0.0006,
+ "loss": 4.630744934082031,
+ "step": 1090
+ },
+ {
+ "epoch": 15.153711790393013,
+ "grad_norm": 0.015333331190049648,
+ "learning_rate": 0.0006,
+ "loss": 4.682388782501221,
+ "step": 1091
+ },
+ {
+ "epoch": 15.167685589519651,
+ "grad_norm": 0.01550073828548193,
+ "learning_rate": 0.0006,
+ "loss": 4.692652702331543,
+ "step": 1092
+ },
+ {
+ "epoch": 15.181659388646288,
+ "grad_norm": 0.014783318154513836,
+ "learning_rate": 0.0006,
+ "loss": 4.6223835945129395,
+ "step": 1093
+ },
+ {
+ "epoch": 15.195633187772925,
+ "grad_norm": 0.014372751116752625,
+ "learning_rate": 0.0006,
+ "loss": 4.52687406539917,
+ "step": 1094
+ },
+ {
+ "epoch": 15.209606986899564,
+ "grad_norm": 0.014779130928218365,
+ "learning_rate": 0.0006,
+ "loss": 4.637009620666504,
+ "step": 1095
+ },
+ {
+ "epoch": 15.2235807860262,
+ "grad_norm": 0.015697073191404343,
+ "learning_rate": 0.0006,
+ "loss": 4.682981014251709,
+ "step": 1096
+ },
+ {
+ "epoch": 15.237554585152838,
+ "grad_norm": 0.016523664817214012,
+ "learning_rate": 0.0006,
+ "loss": 4.705722808837891,
+ "step": 1097
+ },
+ {
+ "epoch": 15.251528384279476,
+ "grad_norm": 0.017633996903896332,
+ "learning_rate": 0.0006,
+ "loss": 4.574686050415039,
+ "step": 1098
+ },
+ {
+ "epoch": 15.265502183406113,
+ "grad_norm": 0.019955076277256012,
+ "learning_rate": 0.0006,
+ "loss": 4.634450912475586,
+ "step": 1099
+ },
+ {
+ "epoch": 15.279475982532752,
+ "grad_norm": 0.02132618986070156,
+ "learning_rate": 0.0006,
+ "loss": 4.56110143661499,
+ "step": 1100
+ },
+ {
+ "epoch": 15.293449781659389,
+ "grad_norm": 0.0207593385130167,
+ "learning_rate": 0.0006,
+ "loss": 4.629396438598633,
+ "step": 1101
+ },
+ {
+ "epoch": 15.307423580786025,
+ "grad_norm": 0.017882540822029114,
+ "learning_rate": 0.0006,
+ "loss": 4.730963230133057,
+ "step": 1102
+ },
+ {
+ "epoch": 15.321397379912664,
+ "grad_norm": 0.01917876861989498,
+ "learning_rate": 0.0006,
+ "loss": 4.64585542678833,
+ "step": 1103
+ },
+ {
+ "epoch": 15.335371179039301,
+ "grad_norm": 0.01943974196910858,
+ "learning_rate": 0.0006,
+ "loss": 4.656181335449219,
+ "step": 1104
+ },
+ {
+ "epoch": 15.34934497816594,
+ "grad_norm": 0.019904715940356255,
+ "learning_rate": 0.0006,
+ "loss": 4.619329452514648,
+ "step": 1105
+ },
+ {
+ "epoch": 15.363318777292577,
+ "grad_norm": 0.019174346700310707,
+ "learning_rate": 0.0006,
+ "loss": 4.647892475128174,
+ "step": 1106
+ },
+ {
+ "epoch": 15.377292576419213,
+ "grad_norm": 0.018244758248329163,
+ "learning_rate": 0.0006,
+ "loss": 4.56166934967041,
+ "step": 1107
+ },
+ {
+ "epoch": 15.391266375545852,
+ "grad_norm": 0.018914205953478813,
+ "learning_rate": 0.0006,
+ "loss": 4.648063659667969,
+ "step": 1108
+ },
+ {
+ "epoch": 15.405240174672489,
+ "grad_norm": 0.01961071416735649,
+ "learning_rate": 0.0006,
+ "loss": 4.665525436401367,
+ "step": 1109
+ },
+ {
+ "epoch": 15.419213973799126,
+ "grad_norm": 0.01841311901807785,
+ "learning_rate": 0.0006,
+ "loss": 4.611342430114746,
+ "step": 1110
+ },
+ {
+ "epoch": 15.433187772925764,
+ "grad_norm": 0.01863427273929119,
+ "learning_rate": 0.0006,
+ "loss": 4.664256572723389,
+ "step": 1111
+ },
+ {
+ "epoch": 15.447161572052401,
+ "grad_norm": 0.020506877452135086,
+ "learning_rate": 0.0006,
+ "loss": 4.631237506866455,
+ "step": 1112
+ },
+ {
+ "epoch": 15.46113537117904,
+ "grad_norm": 0.02223833277821541,
+ "learning_rate": 0.0006,
+ "loss": 4.668654441833496,
+ "step": 1113
+ },
+ {
+ "epoch": 15.475109170305677,
+ "grad_norm": 0.023336702957749367,
+ "learning_rate": 0.0006,
+ "loss": 4.562833309173584,
+ "step": 1114
+ },
+ {
+ "epoch": 15.489082969432314,
+ "grad_norm": 0.023052413016557693,
+ "learning_rate": 0.0006,
+ "loss": 4.621532917022705,
+ "step": 1115
+ },
+ {
+ "epoch": 15.503056768558952,
+ "grad_norm": 0.023420924320816994,
+ "learning_rate": 0.0006,
+ "loss": 4.665465354919434,
+ "step": 1116
+ },
+ {
+ "epoch": 15.51703056768559,
+ "grad_norm": 0.024071281775832176,
+ "learning_rate": 0.0006,
+ "loss": 4.595980644226074,
+ "step": 1117
+ },
+ {
+ "epoch": 15.531004366812226,
+ "grad_norm": 0.023406682536005974,
+ "learning_rate": 0.0006,
+ "loss": 4.642635345458984,
+ "step": 1118
+ },
+ {
+ "epoch": 15.544978165938865,
+ "grad_norm": 0.02138647809624672,
+ "learning_rate": 0.0006,
+ "loss": 4.704513072967529,
+ "step": 1119
+ },
+ {
+ "epoch": 15.558951965065502,
+ "grad_norm": 0.0221384447067976,
+ "learning_rate": 0.0006,
+ "loss": 4.567503452301025,
+ "step": 1120
+ },
+ {
+ "epoch": 15.57292576419214,
+ "grad_norm": 0.020879851654171944,
+ "learning_rate": 0.0006,
+ "loss": 4.664057731628418,
+ "step": 1121
+ },
+ {
+ "epoch": 15.586899563318777,
+ "grad_norm": 0.021478967741131783,
+ "learning_rate": 0.0006,
+ "loss": 4.539627552032471,
+ "step": 1122
+ },
+ {
+ "epoch": 15.600873362445414,
+ "grad_norm": 0.022748656570911407,
+ "learning_rate": 0.0006,
+ "loss": 4.5654497146606445,
+ "step": 1123
+ },
+ {
+ "epoch": 15.614847161572053,
+ "grad_norm": 0.021870248019695282,
+ "learning_rate": 0.0006,
+ "loss": 4.56899356842041,
+ "step": 1124
+ },
+ {
+ "epoch": 15.62882096069869,
+ "grad_norm": 0.021711355075240135,
+ "learning_rate": 0.0006,
+ "loss": 4.657808303833008,
+ "step": 1125
+ },
+ {
+ "epoch": 15.642794759825328,
+ "grad_norm": 0.022810909897089005,
+ "learning_rate": 0.0006,
+ "loss": 4.6526689529418945,
+ "step": 1126
+ },
+ {
+ "epoch": 15.656768558951965,
+ "grad_norm": 0.022561343386769295,
+ "learning_rate": 0.0006,
+ "loss": 4.534887790679932,
+ "step": 1127
+ },
+ {
+ "epoch": 15.670742358078602,
+ "grad_norm": 0.019909802824258804,
+ "learning_rate": 0.0006,
+ "loss": 4.600790023803711,
+ "step": 1128
+ },
+ {
+ "epoch": 15.68471615720524,
+ "grad_norm": 0.020092325285077095,
+ "learning_rate": 0.0006,
+ "loss": 4.773830413818359,
+ "step": 1129
+ },
+ {
+ "epoch": 15.698689956331878,
+ "grad_norm": 0.018834445625543594,
+ "learning_rate": 0.0006,
+ "loss": 4.628868103027344,
+ "step": 1130
+ },
+ {
+ "epoch": 15.712663755458514,
+ "grad_norm": 0.017552798613905907,
+ "learning_rate": 0.0006,
+ "loss": 4.647500038146973,
+ "step": 1131
+ },
+ {
+ "epoch": 15.726637554585153,
+ "grad_norm": 0.01751251146197319,
+ "learning_rate": 0.0006,
+ "loss": 4.590545654296875,
+ "step": 1132
+ },
+ {
+ "epoch": 15.74061135371179,
+ "grad_norm": 0.01699099875986576,
+ "learning_rate": 0.0006,
+ "loss": 4.509699821472168,
+ "step": 1133
+ },
+ {
+ "epoch": 15.754585152838429,
+ "grad_norm": 0.015808649361133575,
+ "learning_rate": 0.0006,
+ "loss": 4.606992244720459,
+ "step": 1134
+ },
+ {
+ "epoch": 15.768558951965066,
+ "grad_norm": 0.01655443385243416,
+ "learning_rate": 0.0006,
+ "loss": 4.585219860076904,
+ "step": 1135
+ },
+ {
+ "epoch": 15.782532751091702,
+ "grad_norm": 0.01708163693547249,
+ "learning_rate": 0.0006,
+ "loss": 4.668147087097168,
+ "step": 1136
+ },
+ {
+ "epoch": 15.796506550218341,
+ "grad_norm": 0.017161913216114044,
+ "learning_rate": 0.0006,
+ "loss": 4.7170634269714355,
+ "step": 1137
+ },
+ {
+ "epoch": 15.810480349344978,
+ "grad_norm": 0.017262674868106842,
+ "learning_rate": 0.0006,
+ "loss": 4.572132110595703,
+ "step": 1138
+ },
+ {
+ "epoch": 15.824454148471617,
+ "grad_norm": 0.017713576555252075,
+ "learning_rate": 0.0006,
+ "loss": 4.702538967132568,
+ "step": 1139
+ },
+ {
+ "epoch": 15.838427947598253,
+ "grad_norm": 0.018730733543634415,
+ "learning_rate": 0.0006,
+ "loss": 4.629721641540527,
+ "step": 1140
+ },
+ {
+ "epoch": 15.85240174672489,
+ "grad_norm": 0.017960187047719955,
+ "learning_rate": 0.0006,
+ "loss": 4.661039352416992,
+ "step": 1141
+ },
+ {
+ "epoch": 15.866375545851529,
+ "grad_norm": 0.01750079356133938,
+ "learning_rate": 0.0006,
+ "loss": 4.503710746765137,
+ "step": 1142
+ },
+ {
+ "epoch": 15.880349344978166,
+ "grad_norm": 0.016630422323942184,
+ "learning_rate": 0.0006,
+ "loss": 4.6582136154174805,
+ "step": 1143
+ },
+ {
+ "epoch": 15.894323144104803,
+ "grad_norm": 0.015896832570433617,
+ "learning_rate": 0.0006,
+ "loss": 4.596693992614746,
+ "step": 1144
+ },
+ {
+ "epoch": 15.908296943231441,
+ "grad_norm": 0.01566407084465027,
+ "learning_rate": 0.0006,
+ "loss": 4.6337199211120605,
+ "step": 1145
+ },
+ {
+ "epoch": 15.922270742358078,
+ "grad_norm": 0.016823330894112587,
+ "learning_rate": 0.0006,
+ "loss": 4.686994552612305,
+ "step": 1146
+ },
+ {
+ "epoch": 15.936244541484717,
+ "grad_norm": 0.016208255663514137,
+ "learning_rate": 0.0006,
+ "loss": 4.6205549240112305,
+ "step": 1147
+ },
+ {
+ "epoch": 15.950218340611354,
+ "grad_norm": 0.0171707421541214,
+ "learning_rate": 0.0006,
+ "loss": 4.582968711853027,
+ "step": 1148
+ },
+ {
+ "epoch": 15.96419213973799,
+ "grad_norm": 0.017009446397423744,
+ "learning_rate": 0.0006,
+ "loss": 4.510274887084961,
+ "step": 1149
+ },
+ {
+ "epoch": 15.97816593886463,
+ "grad_norm": 0.01565811224281788,
+ "learning_rate": 0.0006,
+ "loss": 4.55834436416626,
+ "step": 1150
+ },
+ {
+ "epoch": 15.992139737991266,
+ "grad_norm": 0.015276037156581879,
+ "learning_rate": 0.0006,
+ "loss": 4.537870407104492,
+ "step": 1151
+ },
+ {
+ "epoch": 16.0,
+ "grad_norm": 0.017305882647633553,
+ "learning_rate": 0.0006,
+ "loss": 4.611305236816406,
+ "step": 1152
+ },
+ {
+ "epoch": 16.0,
+ "eval_loss": 4.775513172149658,
+ "eval_runtime": 56.7341,
+ "eval_samples_per_second": 43.043,
+ "eval_steps_per_second": 1.357,
+ "step": 1152
+ },
+ {
+ "epoch": 16.01397379912664,
+ "grad_norm": 0.016752436757087708,
+ "learning_rate": 0.0006,
+ "loss": 4.541312217712402,
+ "step": 1153
+ },
+ {
+ "epoch": 16.027947598253274,
+ "grad_norm": 0.01643134094774723,
+ "learning_rate": 0.0006,
+ "loss": 4.585402965545654,
+ "step": 1154
+ },
+ {
+ "epoch": 16.041921397379912,
+ "grad_norm": 0.01888720504939556,
+ "learning_rate": 0.0006,
+ "loss": 4.597825050354004,
+ "step": 1155
+ },
+ {
+ "epoch": 16.05589519650655,
+ "grad_norm": 0.022044632583856583,
+ "learning_rate": 0.0006,
+ "loss": 4.532772064208984,
+ "step": 1156
+ },
+ {
+ "epoch": 16.069868995633186,
+ "grad_norm": 0.022986987605690956,
+ "learning_rate": 0.0006,
+ "loss": 4.551351547241211,
+ "step": 1157
+ },
+ {
+ "epoch": 16.083842794759825,
+ "grad_norm": 0.020116791129112244,
+ "learning_rate": 0.0006,
+ "loss": 4.484149932861328,
+ "step": 1158
+ },
+ {
+ "epoch": 16.097816593886463,
+ "grad_norm": 0.021442247554659843,
+ "learning_rate": 0.0006,
+ "loss": 4.515316486358643,
+ "step": 1159
+ },
+ {
+ "epoch": 16.111790393013102,
+ "grad_norm": 0.020974772050976753,
+ "learning_rate": 0.0006,
+ "loss": 4.51606559753418,
+ "step": 1160
+ },
+ {
+ "epoch": 16.125764192139737,
+ "grad_norm": 0.021274864673614502,
+ "learning_rate": 0.0006,
+ "loss": 4.587329864501953,
+ "step": 1161
+ },
+ {
+ "epoch": 16.139737991266376,
+ "grad_norm": 0.022799784317612648,
+ "learning_rate": 0.0006,
+ "loss": 4.568149089813232,
+ "step": 1162
+ },
+ {
+ "epoch": 16.153711790393015,
+ "grad_norm": 0.021784933283925056,
+ "learning_rate": 0.0006,
+ "loss": 4.515336990356445,
+ "step": 1163
+ },
+ {
+ "epoch": 16.16768558951965,
+ "grad_norm": 0.024261271581053734,
+ "learning_rate": 0.0006,
+ "loss": 4.5127081871032715,
+ "step": 1164
+ },
+ {
+ "epoch": 16.18165938864629,
+ "grad_norm": 0.02248375490307808,
+ "learning_rate": 0.0006,
+ "loss": 4.754251003265381,
+ "step": 1165
+ },
+ {
+ "epoch": 16.195633187772927,
+ "grad_norm": 0.02136492356657982,
+ "learning_rate": 0.0006,
+ "loss": 4.58193826675415,
+ "step": 1166
+ },
+ {
+ "epoch": 16.209606986899562,
+ "grad_norm": 0.020727066323161125,
+ "learning_rate": 0.0006,
+ "loss": 4.564209938049316,
+ "step": 1167
+ },
+ {
+ "epoch": 16.2235807860262,
+ "grad_norm": 0.017789119854569435,
+ "learning_rate": 0.0006,
+ "loss": 4.698680877685547,
+ "step": 1168
+ },
+ {
+ "epoch": 16.23755458515284,
+ "grad_norm": 0.017603034153580666,
+ "learning_rate": 0.0006,
+ "loss": 4.501171588897705,
+ "step": 1169
+ },
+ {
+ "epoch": 16.251528384279474,
+ "grad_norm": 0.017345862463116646,
+ "learning_rate": 0.0006,
+ "loss": 4.593590259552002,
+ "step": 1170
+ },
+ {
+ "epoch": 16.265502183406113,
+ "grad_norm": 0.018047425895929337,
+ "learning_rate": 0.0006,
+ "loss": 4.597467422485352,
+ "step": 1171
+ },
+ {
+ "epoch": 16.27947598253275,
+ "grad_norm": 0.01947222836315632,
+ "learning_rate": 0.0006,
+ "loss": 4.5072221755981445,
+ "step": 1172
+ },
+ {
+ "epoch": 16.29344978165939,
+ "grad_norm": 0.019598500803112984,
+ "learning_rate": 0.0006,
+ "loss": 4.601351737976074,
+ "step": 1173
+ },
+ {
+ "epoch": 16.307423580786025,
+ "grad_norm": 0.020002514123916626,
+ "learning_rate": 0.0006,
+ "loss": 4.580431938171387,
+ "step": 1174
+ },
+ {
+ "epoch": 16.321397379912664,
+ "grad_norm": 0.0216530729085207,
+ "learning_rate": 0.0006,
+ "loss": 4.424681663513184,
+ "step": 1175
+ },
+ {
+ "epoch": 16.335371179039303,
+ "grad_norm": 0.01949235238134861,
+ "learning_rate": 0.0006,
+ "loss": 4.751883029937744,
+ "step": 1176
+ },
+ {
+ "epoch": 16.349344978165938,
+ "grad_norm": 0.017779016867280006,
+ "learning_rate": 0.0006,
+ "loss": 4.541490077972412,
+ "step": 1177
+ },
+ {
+ "epoch": 16.363318777292577,
+ "grad_norm": 0.01909121684730053,
+ "learning_rate": 0.0006,
+ "loss": 4.594141006469727,
+ "step": 1178
+ },
+ {
+ "epoch": 16.377292576419215,
+ "grad_norm": 0.019894849509000778,
+ "learning_rate": 0.0006,
+ "loss": 4.555070400238037,
+ "step": 1179
+ },
+ {
+ "epoch": 16.39126637554585,
+ "grad_norm": 0.01993032731115818,
+ "learning_rate": 0.0006,
+ "loss": 4.521401405334473,
+ "step": 1180
+ },
+ {
+ "epoch": 16.40524017467249,
+ "grad_norm": 0.0210788045078516,
+ "learning_rate": 0.0006,
+ "loss": 4.672842979431152,
+ "step": 1181
+ },
+ {
+ "epoch": 16.419213973799128,
+ "grad_norm": 0.02274598926305771,
+ "learning_rate": 0.0006,
+ "loss": 4.629504203796387,
+ "step": 1182
+ },
+ {
+ "epoch": 16.433187772925763,
+ "grad_norm": 0.020448120310902596,
+ "learning_rate": 0.0006,
+ "loss": 4.548209190368652,
+ "step": 1183
+ },
+ {
+ "epoch": 16.4471615720524,
+ "grad_norm": 0.017565643414855003,
+ "learning_rate": 0.0006,
+ "loss": 4.529054641723633,
+ "step": 1184
+ },
+ {
+ "epoch": 16.46113537117904,
+ "grad_norm": 0.019504237920045853,
+ "learning_rate": 0.0006,
+ "loss": 4.48182487487793,
+ "step": 1185
+ },
+ {
+ "epoch": 16.475109170305675,
+ "grad_norm": 0.016823742538690567,
+ "learning_rate": 0.0006,
+ "loss": 4.608696937561035,
+ "step": 1186
+ },
+ {
+ "epoch": 16.489082969432314,
+ "grad_norm": 0.01602894812822342,
+ "learning_rate": 0.0006,
+ "loss": 4.620062351226807,
+ "step": 1187
+ },
+ {
+ "epoch": 16.503056768558952,
+ "grad_norm": 0.016615642234683037,
+ "learning_rate": 0.0006,
+ "loss": 4.69378662109375,
+ "step": 1188
+ },
+ {
+ "epoch": 16.51703056768559,
+ "grad_norm": 0.018576730042696,
+ "learning_rate": 0.0006,
+ "loss": 4.537041664123535,
+ "step": 1189
+ },
+ {
+ "epoch": 16.531004366812226,
+ "grad_norm": 0.018611151725053787,
+ "learning_rate": 0.0006,
+ "loss": 4.496169090270996,
+ "step": 1190
+ },
+ {
+ "epoch": 16.544978165938865,
+ "grad_norm": 0.018622413277626038,
+ "learning_rate": 0.0006,
+ "loss": 4.557358264923096,
+ "step": 1191
+ },
+ {
+ "epoch": 16.558951965065503,
+ "grad_norm": 0.01823544315993786,
+ "learning_rate": 0.0006,
+ "loss": 4.730792999267578,
+ "step": 1192
+ },
+ {
+ "epoch": 16.57292576419214,
+ "grad_norm": 0.018796846270561218,
+ "learning_rate": 0.0006,
+ "loss": 4.781791687011719,
+ "step": 1193
+ },
+ {
+ "epoch": 16.586899563318777,
+ "grad_norm": 0.01728292554616928,
+ "learning_rate": 0.0006,
+ "loss": 4.515257835388184,
+ "step": 1194
+ },
+ {
+ "epoch": 16.600873362445416,
+ "grad_norm": 0.01699589006602764,
+ "learning_rate": 0.0006,
+ "loss": 4.5996904373168945,
+ "step": 1195
+ },
+ {
+ "epoch": 16.61484716157205,
+ "grad_norm": 0.015905970707535744,
+ "learning_rate": 0.0006,
+ "loss": 4.601301193237305,
+ "step": 1196
+ },
+ {
+ "epoch": 16.62882096069869,
+ "grad_norm": 0.01338571310043335,
+ "learning_rate": 0.0006,
+ "loss": 4.557708740234375,
+ "step": 1197
+ },
+ {
+ "epoch": 16.64279475982533,
+ "grad_norm": 0.015298855490982533,
+ "learning_rate": 0.0006,
+ "loss": 4.542287349700928,
+ "step": 1198
+ },
+ {
+ "epoch": 16.656768558951963,
+ "grad_norm": 0.014753367751836777,
+ "learning_rate": 0.0006,
+ "loss": 4.595787048339844,
+ "step": 1199
+ },
+ {
+ "epoch": 16.670742358078602,
+ "grad_norm": 0.015527483075857162,
+ "learning_rate": 0.0006,
+ "loss": 4.607529163360596,
+ "step": 1200
+ },
+ {
+ "epoch": 16.68471615720524,
+ "grad_norm": 0.015042162500321865,
+ "learning_rate": 0.0006,
+ "loss": 4.686626434326172,
+ "step": 1201
+ },
+ {
+ "epoch": 16.69868995633188,
+ "grad_norm": 0.0159393772482872,
+ "learning_rate": 0.0006,
+ "loss": 4.47656774520874,
+ "step": 1202
+ },
+ {
+ "epoch": 16.712663755458514,
+ "grad_norm": 0.01532787922769785,
+ "learning_rate": 0.0006,
+ "loss": 4.55678653717041,
+ "step": 1203
+ },
+ {
+ "epoch": 16.726637554585153,
+ "grad_norm": 0.016332164406776428,
+ "learning_rate": 0.0006,
+ "loss": 4.51441764831543,
+ "step": 1204
+ },
+ {
+ "epoch": 16.74061135371179,
+ "grad_norm": 0.018901726230978966,
+ "learning_rate": 0.0006,
+ "loss": 4.57227897644043,
+ "step": 1205
+ },
+ {
+ "epoch": 16.754585152838427,
+ "grad_norm": 0.018589433282613754,
+ "learning_rate": 0.0006,
+ "loss": 4.550070762634277,
+ "step": 1206
+ },
+ {
+ "epoch": 16.768558951965066,
+ "grad_norm": 0.01634034886956215,
+ "learning_rate": 0.0006,
+ "loss": 4.586910724639893,
+ "step": 1207
+ },
+ {
+ "epoch": 16.782532751091704,
+ "grad_norm": 0.017599433660507202,
+ "learning_rate": 0.0006,
+ "loss": 4.5891923904418945,
+ "step": 1208
+ },
+ {
+ "epoch": 16.79650655021834,
+ "grad_norm": 0.01878446340560913,
+ "learning_rate": 0.0006,
+ "loss": 4.57227087020874,
+ "step": 1209
+ },
+ {
+ "epoch": 16.810480349344978,
+ "grad_norm": 0.01803469844162464,
+ "learning_rate": 0.0006,
+ "loss": 4.583845138549805,
+ "step": 1210
+ },
+ {
+ "epoch": 16.824454148471617,
+ "grad_norm": 0.018505144864320755,
+ "learning_rate": 0.0006,
+ "loss": 4.561151504516602,
+ "step": 1211
+ },
+ {
+ "epoch": 16.83842794759825,
+ "grad_norm": 0.01920081488788128,
+ "learning_rate": 0.0006,
+ "loss": 4.506834030151367,
+ "step": 1212
+ },
+ {
+ "epoch": 16.85240174672489,
+ "grad_norm": 0.0177712831646204,
+ "learning_rate": 0.0006,
+ "loss": 4.557195663452148,
+ "step": 1213
+ },
+ {
+ "epoch": 16.86637554585153,
+ "grad_norm": 0.019113918766379356,
+ "learning_rate": 0.0006,
+ "loss": 4.651227951049805,
+ "step": 1214
+ },
+ {
+ "epoch": 16.880349344978168,
+ "grad_norm": 0.02012539468705654,
+ "learning_rate": 0.0006,
+ "loss": 4.4952239990234375,
+ "step": 1215
+ },
+ {
+ "epoch": 16.894323144104803,
+ "grad_norm": 0.019993646070361137,
+ "learning_rate": 0.0006,
+ "loss": 4.517433166503906,
+ "step": 1216
+ },
+ {
+ "epoch": 16.90829694323144,
+ "grad_norm": 0.020041441544890404,
+ "learning_rate": 0.0006,
+ "loss": 4.489407539367676,
+ "step": 1217
+ },
+ {
+ "epoch": 16.92227074235808,
+ "grad_norm": 0.0183611661195755,
+ "learning_rate": 0.0006,
+ "loss": 4.5435638427734375,
+ "step": 1218
+ },
+ {
+ "epoch": 16.936244541484715,
+ "grad_norm": 0.019138429313898087,
+ "learning_rate": 0.0006,
+ "loss": 4.545949459075928,
+ "step": 1219
+ },
+ {
+ "epoch": 16.950218340611354,
+ "grad_norm": 0.018554266542196274,
+ "learning_rate": 0.0006,
+ "loss": 4.579620361328125,
+ "step": 1220
+ },
+ {
+ "epoch": 16.964192139737992,
+ "grad_norm": 0.0193545650690794,
+ "learning_rate": 0.0006,
+ "loss": 4.588255405426025,
+ "step": 1221
+ },
+ {
+ "epoch": 16.978165938864628,
+ "grad_norm": 0.019389165565371513,
+ "learning_rate": 0.0006,
+ "loss": 4.520698547363281,
+ "step": 1222
+ },
+ {
+ "epoch": 16.992139737991266,
+ "grad_norm": 0.016601379960775375,
+ "learning_rate": 0.0006,
+ "loss": 4.4809064865112305,
+ "step": 1223
+ },
+ {
+ "epoch": 17.0,
+ "grad_norm": 0.01771303080022335,
+ "learning_rate": 0.0006,
+ "loss": 4.721122741699219,
+ "step": 1224
+ },
+ {
+ "epoch": 17.0,
+ "eval_loss": 4.7903666496276855,
+ "eval_runtime": 56.9826,
+ "eval_samples_per_second": 42.855,
+ "eval_steps_per_second": 1.351,
+ "step": 1224
+ },
+ {
+ "epoch": 17.01397379912664,
+ "grad_norm": 0.017720289528369904,
+ "learning_rate": 0.0006,
+ "loss": 4.523834705352783,
+ "step": 1225
+ },
+ {
+ "epoch": 17.027947598253274,
+ "grad_norm": 0.020137161016464233,
+ "learning_rate": 0.0006,
+ "loss": 4.574061870574951,
+ "step": 1226
+ },
+ {
+ "epoch": 17.041921397379912,
+ "grad_norm": 0.02002314105629921,
+ "learning_rate": 0.0006,
+ "loss": 4.512407302856445,
+ "step": 1227
+ },
+ {
+ "epoch": 17.05589519650655,
+ "grad_norm": 0.019628183916211128,
+ "learning_rate": 0.0006,
+ "loss": 4.502533435821533,
+ "step": 1228
+ },
+ {
+ "epoch": 17.069868995633186,
+ "grad_norm": 0.019808458164334297,
+ "learning_rate": 0.0006,
+ "loss": 4.463156700134277,
+ "step": 1229
+ },
+ {
+ "epoch": 17.083842794759825,
+ "grad_norm": 0.021250786259770393,
+ "learning_rate": 0.0006,
+ "loss": 4.622682094573975,
+ "step": 1230
+ },
+ {
+ "epoch": 17.097816593886463,
+ "grad_norm": 0.022031694650650024,
+ "learning_rate": 0.0006,
+ "loss": 4.451882839202881,
+ "step": 1231
+ },
+ {
+ "epoch": 17.111790393013102,
+ "grad_norm": 0.01898987777531147,
+ "learning_rate": 0.0006,
+ "loss": 4.529331684112549,
+ "step": 1232
+ },
+ {
+ "epoch": 17.125764192139737,
+ "grad_norm": 0.018749834969639778,
+ "learning_rate": 0.0006,
+ "loss": 4.435815334320068,
+ "step": 1233
+ },
+ {
+ "epoch": 17.139737991266376,
+ "grad_norm": 0.01896674931049347,
+ "learning_rate": 0.0006,
+ "loss": 4.484029293060303,
+ "step": 1234
+ },
+ {
+ "epoch": 17.153711790393015,
+ "grad_norm": 0.018286997452378273,
+ "learning_rate": 0.0006,
+ "loss": 4.523510932922363,
+ "step": 1235
+ },
+ {
+ "epoch": 17.16768558951965,
+ "grad_norm": 0.017273059114813805,
+ "learning_rate": 0.0006,
+ "loss": 4.451767444610596,
+ "step": 1236
+ },
+ {
+ "epoch": 17.18165938864629,
+ "grad_norm": 0.016499150544404984,
+ "learning_rate": 0.0006,
+ "loss": 4.5568647384643555,
+ "step": 1237
+ },
+ {
+ "epoch": 17.195633187772927,
+ "grad_norm": 0.015934469178318977,
+ "learning_rate": 0.0006,
+ "loss": 4.4639997482299805,
+ "step": 1238
+ },
+ {
+ "epoch": 17.209606986899562,
+ "grad_norm": 0.01714209094643593,
+ "learning_rate": 0.0006,
+ "loss": 4.547614097595215,
+ "step": 1239
+ },
+ {
+ "epoch": 17.2235807860262,
+ "grad_norm": 0.019167495891451836,
+ "learning_rate": 0.0006,
+ "loss": 4.571097373962402,
+ "step": 1240
+ },
+ {
+ "epoch": 17.23755458515284,
+ "grad_norm": 0.018276996910572052,
+ "learning_rate": 0.0006,
+ "loss": 4.489764213562012,
+ "step": 1241
+ },
+ {
+ "epoch": 17.251528384279474,
+ "grad_norm": 0.021296866238117218,
+ "learning_rate": 0.0006,
+ "loss": 4.526913642883301,
+ "step": 1242
+ },
+ {
+ "epoch": 17.265502183406113,
+ "grad_norm": 0.022606629878282547,
+ "learning_rate": 0.0006,
+ "loss": 4.521624565124512,
+ "step": 1243
+ },
+ {
+ "epoch": 17.27947598253275,
+ "grad_norm": 0.021204711869359016,
+ "learning_rate": 0.0006,
+ "loss": 4.500864028930664,
+ "step": 1244
+ },
+ {
+ "epoch": 17.29344978165939,
+ "grad_norm": 0.01988849975168705,
+ "learning_rate": 0.0006,
+ "loss": 4.45162296295166,
+ "step": 1245
+ },
+ {
+ "epoch": 17.307423580786025,
+ "grad_norm": 0.020268384367227554,
+ "learning_rate": 0.0006,
+ "loss": 4.536198616027832,
+ "step": 1246
+ },
+ {
+ "epoch": 17.321397379912664,
+ "grad_norm": 0.020185677334666252,
+ "learning_rate": 0.0006,
+ "loss": 4.440262794494629,
+ "step": 1247
+ },
+ {
+ "epoch": 17.335371179039303,
+ "grad_norm": 0.020106853917241096,
+ "learning_rate": 0.0006,
+ "loss": 4.475907325744629,
+ "step": 1248
+ },
+ {
+ "epoch": 17.349344978165938,
+ "grad_norm": 0.018883297219872475,
+ "learning_rate": 0.0006,
+ "loss": 4.640661716461182,
+ "step": 1249
+ },
+ {
+ "epoch": 17.363318777292577,
+ "grad_norm": 0.01852620206773281,
+ "learning_rate": 0.0006,
+ "loss": 4.578243255615234,
+ "step": 1250
+ },
+ {
+ "epoch": 17.377292576419215,
+ "grad_norm": 0.01886451430618763,
+ "learning_rate": 0.0006,
+ "loss": 4.546026229858398,
+ "step": 1251
+ },
+ {
+ "epoch": 17.39126637554585,
+ "grad_norm": 0.01777452416718006,
+ "learning_rate": 0.0006,
+ "loss": 4.521542549133301,
+ "step": 1252
+ },
+ {
+ "epoch": 17.40524017467249,
+ "grad_norm": 0.016626447439193726,
+ "learning_rate": 0.0006,
+ "loss": 4.52629280090332,
+ "step": 1253
+ },
+ {
+ "epoch": 17.419213973799128,
+ "grad_norm": 0.01599477231502533,
+ "learning_rate": 0.0006,
+ "loss": 4.4796037673950195,
+ "step": 1254
+ },
+ {
+ "epoch": 17.433187772925763,
+ "grad_norm": 0.017021458595991135,
+ "learning_rate": 0.0006,
+ "loss": 4.468813419342041,
+ "step": 1255
+ },
+ {
+ "epoch": 17.4471615720524,
+ "grad_norm": 0.016925333067774773,
+ "learning_rate": 0.0006,
+ "loss": 4.584898948669434,
+ "step": 1256
+ },
+ {
+ "epoch": 17.46113537117904,
+ "grad_norm": 0.01714998669922352,
+ "learning_rate": 0.0006,
+ "loss": 4.523544788360596,
+ "step": 1257
+ },
+ {
+ "epoch": 17.475109170305675,
+ "grad_norm": 0.01802288182079792,
+ "learning_rate": 0.0006,
+ "loss": 4.595581531524658,
+ "step": 1258
+ },
+ {
+ "epoch": 17.489082969432314,
+ "grad_norm": 0.0180289875715971,
+ "learning_rate": 0.0006,
+ "loss": 4.394440650939941,
+ "step": 1259
+ },
+ {
+ "epoch": 17.503056768558952,
+ "grad_norm": 0.018504245206713676,
+ "learning_rate": 0.0006,
+ "loss": 4.4794416427612305,
+ "step": 1260
+ },
+ {
+ "epoch": 17.51703056768559,
+ "grad_norm": 0.019973233342170715,
+ "learning_rate": 0.0006,
+ "loss": 4.4241437911987305,
+ "step": 1261
+ },
+ {
+ "epoch": 17.531004366812226,
+ "grad_norm": 0.020063083618879318,
+ "learning_rate": 0.0006,
+ "loss": 4.48068380355835,
+ "step": 1262
+ },
+ {
+ "epoch": 17.544978165938865,
+ "grad_norm": 0.02066640742123127,
+ "learning_rate": 0.0006,
+ "loss": 4.628993034362793,
+ "step": 1263
+ },
+ {
+ "epoch": 17.558951965065503,
+ "grad_norm": 0.02206416428089142,
+ "learning_rate": 0.0006,
+ "loss": 4.564790725708008,
+ "step": 1264
+ },
+ {
+ "epoch": 17.57292576419214,
+ "grad_norm": 0.020640335977077484,
+ "learning_rate": 0.0006,
+ "loss": 4.578705787658691,
+ "step": 1265
+ },
+ {
+ "epoch": 17.586899563318777,
+ "grad_norm": 0.019093405455350876,
+ "learning_rate": 0.0006,
+ "loss": 4.5023603439331055,
+ "step": 1266
+ },
+ {
+ "epoch": 17.600873362445416,
+ "grad_norm": 0.019459193572402,
+ "learning_rate": 0.0006,
+ "loss": 4.594531059265137,
+ "step": 1267
+ },
+ {
+ "epoch": 17.61484716157205,
+ "grad_norm": 0.02021171525120735,
+ "learning_rate": 0.0006,
+ "loss": 4.770299911499023,
+ "step": 1268
+ },
+ {
+ "epoch": 17.62882096069869,
+ "grad_norm": 0.021161191165447235,
+ "learning_rate": 0.0006,
+ "loss": 4.5380401611328125,
+ "step": 1269
+ },
+ {
+ "epoch": 17.64279475982533,
+ "grad_norm": 0.021013403311371803,
+ "learning_rate": 0.0006,
+ "loss": 4.455524921417236,
+ "step": 1270
+ },
+ {
+ "epoch": 17.656768558951963,
+ "grad_norm": 0.020336592569947243,
+ "learning_rate": 0.0006,
+ "loss": 4.533672332763672,
+ "step": 1271
+ },
+ {
+ "epoch": 17.670742358078602,
+ "grad_norm": 0.020159434527158737,
+ "learning_rate": 0.0006,
+ "loss": 4.596151351928711,
+ "step": 1272
+ },
+ {
+ "epoch": 17.68471615720524,
+ "grad_norm": 0.020432960242033005,
+ "learning_rate": 0.0006,
+ "loss": 4.55531644821167,
+ "step": 1273
+ },
+ {
+ "epoch": 17.69868995633188,
+ "grad_norm": 0.025466229766607285,
+ "learning_rate": 0.0006,
+ "loss": 4.566912651062012,
+ "step": 1274
+ },
+ {
+ "epoch": 17.712663755458514,
+ "grad_norm": 0.026537558063864708,
+ "learning_rate": 0.0006,
+ "loss": 4.526003837585449,
+ "step": 1275
+ },
+ {
+ "epoch": 17.726637554585153,
+ "grad_norm": 0.024397607892751694,
+ "learning_rate": 0.0006,
+ "loss": 4.541755676269531,
+ "step": 1276
+ },
+ {
+ "epoch": 17.74061135371179,
+ "grad_norm": 0.022062312811613083,
+ "learning_rate": 0.0006,
+ "loss": 4.6532793045043945,
+ "step": 1277
+ },
+ {
+ "epoch": 17.754585152838427,
+ "grad_norm": 0.02204282023012638,
+ "learning_rate": 0.0006,
+ "loss": 4.582825660705566,
+ "step": 1278
+ },
+ {
+ "epoch": 17.768558951965066,
+ "grad_norm": 0.020844416692852974,
+ "learning_rate": 0.0006,
+ "loss": 4.522336483001709,
+ "step": 1279
+ },
+ {
+ "epoch": 17.782532751091704,
+ "grad_norm": 0.019815703853964806,
+ "learning_rate": 0.0006,
+ "loss": 4.607398509979248,
+ "step": 1280
+ },
+ {
+ "epoch": 17.79650655021834,
+ "grad_norm": 0.016676954925060272,
+ "learning_rate": 0.0006,
+ "loss": 4.446206092834473,
+ "step": 1281
+ },
+ {
+ "epoch": 17.810480349344978,
+ "grad_norm": 0.0166452769190073,
+ "learning_rate": 0.0006,
+ "loss": 4.505931854248047,
+ "step": 1282
+ },
+ {
+ "epoch": 17.824454148471617,
+ "grad_norm": 0.015928907319903374,
+ "learning_rate": 0.0006,
+ "loss": 4.455267429351807,
+ "step": 1283
+ },
+ {
+ "epoch": 17.83842794759825,
+ "grad_norm": 0.015219560824334621,
+ "learning_rate": 0.0006,
+ "loss": 4.5623674392700195,
+ "step": 1284
+ },
+ {
+ "epoch": 17.85240174672489,
+ "grad_norm": 0.015125355683267117,
+ "learning_rate": 0.0006,
+ "loss": 4.554104804992676,
+ "step": 1285
+ },
+ {
+ "epoch": 17.86637554585153,
+ "grad_norm": 0.015612718649208546,
+ "learning_rate": 0.0006,
+ "loss": 4.484847068786621,
+ "step": 1286
+ },
+ {
+ "epoch": 17.880349344978168,
+ "grad_norm": 0.015215662308037281,
+ "learning_rate": 0.0006,
+ "loss": 4.4802093505859375,
+ "step": 1287
+ },
+ {
+ "epoch": 17.894323144104803,
+ "grad_norm": 0.01511684525758028,
+ "learning_rate": 0.0006,
+ "loss": 4.537298679351807,
+ "step": 1288
+ },
+ {
+ "epoch": 17.90829694323144,
+ "grad_norm": 0.013841322623193264,
+ "learning_rate": 0.0006,
+ "loss": 4.427488327026367,
+ "step": 1289
+ },
+ {
+ "epoch": 17.92227074235808,
+ "grad_norm": 0.015115964226424694,
+ "learning_rate": 0.0006,
+ "loss": 4.4547119140625,
+ "step": 1290
+ },
+ {
+ "epoch": 17.936244541484715,
+ "grad_norm": 0.016411006450653076,
+ "learning_rate": 0.0006,
+ "loss": 4.52994966506958,
+ "step": 1291
+ },
+ {
+ "epoch": 17.950218340611354,
+ "grad_norm": 0.017313152551651,
+ "learning_rate": 0.0006,
+ "loss": 4.5337653160095215,
+ "step": 1292
+ },
+ {
+ "epoch": 17.964192139737992,
+ "grad_norm": 0.016559796407818794,
+ "learning_rate": 0.0006,
+ "loss": 4.5277910232543945,
+ "step": 1293
+ },
+ {
+ "epoch": 17.978165938864628,
+ "grad_norm": 0.015854761004447937,
+ "learning_rate": 0.0006,
+ "loss": 4.6334309577941895,
+ "step": 1294
+ },
+ {
+ "epoch": 17.992139737991266,
+ "grad_norm": 0.015658238902688026,
+ "learning_rate": 0.0006,
+ "loss": 4.613372802734375,
+ "step": 1295
+ },
+ {
+ "epoch": 18.0,
+ "grad_norm": 0.0169155802577734,
+ "learning_rate": 0.0006,
+ "loss": 4.527356147766113,
+ "step": 1296
+ },
+ {
+ "epoch": 18.0,
+ "eval_loss": 4.764543056488037,
+ "eval_runtime": 56.5998,
+ "eval_samples_per_second": 43.145,
+ "eval_steps_per_second": 1.36,
+ "step": 1296
+ },
+ {
+ "epoch": 18.01397379912664,
+ "grad_norm": 0.015348963439464569,
+ "learning_rate": 0.0006,
+ "loss": 4.466423988342285,
+ "step": 1297
+ },
+ {
+ "epoch": 18.027947598253274,
+ "grad_norm": 0.01768588088452816,
+ "learning_rate": 0.0006,
+ "loss": 4.545933246612549,
+ "step": 1298
+ },
+ {
+ "epoch": 18.041921397379912,
+ "grad_norm": 0.017381226643919945,
+ "learning_rate": 0.0006,
+ "loss": 4.494060039520264,
+ "step": 1299
+ },
+ {
+ "epoch": 18.05589519650655,
+ "grad_norm": 0.019002335146069527,
+ "learning_rate": 0.0006,
+ "loss": 4.395200729370117,
+ "step": 1300
+ },
+ {
+ "epoch": 18.069868995633186,
+ "grad_norm": 0.01827932707965374,
+ "learning_rate": 0.0006,
+ "loss": 4.424848556518555,
+ "step": 1301
+ },
+ {
+ "epoch": 18.083842794759825,
+ "grad_norm": 0.016383031383156776,
+ "learning_rate": 0.0006,
+ "loss": 4.494722366333008,
+ "step": 1302
+ },
+ {
+ "epoch": 18.097816593886463,
+ "grad_norm": 0.016265157610177994,
+ "learning_rate": 0.0006,
+ "loss": 4.416522979736328,
+ "step": 1303
+ },
+ {
+ "epoch": 18.111790393013102,
+ "grad_norm": 0.01702386513352394,
+ "learning_rate": 0.0006,
+ "loss": 4.600103378295898,
+ "step": 1304
+ },
+ {
+ "epoch": 18.125764192139737,
+ "grad_norm": 0.018495498225092888,
+ "learning_rate": 0.0006,
+ "loss": 4.588401794433594,
+ "step": 1305
+ },
+ {
+ "epoch": 18.139737991266376,
+ "grad_norm": 0.017921684309840202,
+ "learning_rate": 0.0006,
+ "loss": 4.478075981140137,
+ "step": 1306
+ },
+ {
+ "epoch": 18.153711790393015,
+ "grad_norm": 0.01801278069615364,
+ "learning_rate": 0.0006,
+ "loss": 4.413186550140381,
+ "step": 1307
+ },
+ {
+ "epoch": 18.16768558951965,
+ "grad_norm": 0.017211275175213814,
+ "learning_rate": 0.0006,
+ "loss": 4.461341857910156,
+ "step": 1308
+ },
+ {
+ "epoch": 18.18165938864629,
+ "grad_norm": 0.017589328810572624,
+ "learning_rate": 0.0006,
+ "loss": 4.546969413757324,
+ "step": 1309
+ },
+ {
+ "epoch": 18.195633187772927,
+ "grad_norm": 0.01797698810696602,
+ "learning_rate": 0.0006,
+ "loss": 4.530797004699707,
+ "step": 1310
+ },
+ {
+ "epoch": 18.209606986899562,
+ "grad_norm": 0.020408619195222855,
+ "learning_rate": 0.0006,
+ "loss": 4.62999153137207,
+ "step": 1311
+ },
+ {
+ "epoch": 18.2235807860262,
+ "grad_norm": 0.01994038000702858,
+ "learning_rate": 0.0006,
+ "loss": 4.586811065673828,
+ "step": 1312
+ },
+ {
+ "epoch": 18.23755458515284,
+ "grad_norm": 0.019381340593099594,
+ "learning_rate": 0.0006,
+ "loss": 4.4769368171691895,
+ "step": 1313
+ },
+ {
+ "epoch": 18.251528384279474,
+ "grad_norm": 0.020484555512666702,
+ "learning_rate": 0.0006,
+ "loss": 4.524302959442139,
+ "step": 1314
+ },
+ {
+ "epoch": 18.265502183406113,
+ "grad_norm": 0.02426416426897049,
+ "learning_rate": 0.0006,
+ "loss": 4.561705589294434,
+ "step": 1315
+ },
+ {
+ "epoch": 18.27947598253275,
+ "grad_norm": 0.024748101830482483,
+ "learning_rate": 0.0006,
+ "loss": 4.4654316902160645,
+ "step": 1316
+ },
+ {
+ "epoch": 18.29344978165939,
+ "grad_norm": 0.020703328773379326,
+ "learning_rate": 0.0006,
+ "loss": 4.53033447265625,
+ "step": 1317
+ },
+ {
+ "epoch": 18.307423580786025,
+ "grad_norm": 0.01765458658337593,
+ "learning_rate": 0.0006,
+ "loss": 4.5154266357421875,
+ "step": 1318
+ },
+ {
+ "epoch": 18.321397379912664,
+ "grad_norm": 0.018646420910954475,
+ "learning_rate": 0.0006,
+ "loss": 4.5224714279174805,
+ "step": 1319
+ },
+ {
+ "epoch": 18.335371179039303,
+ "grad_norm": 0.01907474733889103,
+ "learning_rate": 0.0006,
+ "loss": 4.5055437088012695,
+ "step": 1320
+ },
+ {
+ "epoch": 18.349344978165938,
+ "grad_norm": 0.018179846927523613,
+ "learning_rate": 0.0006,
+ "loss": 4.462161064147949,
+ "step": 1321
+ },
+ {
+ "epoch": 18.363318777292577,
+ "grad_norm": 0.016697878018021584,
+ "learning_rate": 0.0006,
+ "loss": 4.473718643188477,
+ "step": 1322
+ },
+ {
+ "epoch": 18.377292576419215,
+ "grad_norm": 0.017504019662737846,
+ "learning_rate": 0.0006,
+ "loss": 4.474348068237305,
+ "step": 1323
+ },
+ {
+ "epoch": 18.39126637554585,
+ "grad_norm": 0.018889913335442543,
+ "learning_rate": 0.0006,
+ "loss": 4.4531636238098145,
+ "step": 1324
+ },
+ {
+ "epoch": 18.40524017467249,
+ "grad_norm": 0.019167251884937286,
+ "learning_rate": 0.0006,
+ "loss": 4.4278364181518555,
+ "step": 1325
+ },
+ {
+ "epoch": 18.419213973799128,
+ "grad_norm": 0.015895741060376167,
+ "learning_rate": 0.0006,
+ "loss": 4.474592208862305,
+ "step": 1326
+ },
+ {
+ "epoch": 18.433187772925763,
+ "grad_norm": 0.01531882956624031,
+ "learning_rate": 0.0006,
+ "loss": 4.403757572174072,
+ "step": 1327
+ },
+ {
+ "epoch": 18.4471615720524,
+ "grad_norm": 0.01565849967300892,
+ "learning_rate": 0.0006,
+ "loss": 4.507664680480957,
+ "step": 1328
+ },
+ {
+ "epoch": 18.46113537117904,
+ "grad_norm": 0.01602315716445446,
+ "learning_rate": 0.0006,
+ "loss": 4.5648908615112305,
+ "step": 1329
+ },
+ {
+ "epoch": 18.475109170305675,
+ "grad_norm": 0.016467366367578506,
+ "learning_rate": 0.0006,
+ "loss": 4.428762912750244,
+ "step": 1330
+ },
+ {
+ "epoch": 18.489082969432314,
+ "grad_norm": 0.01604127697646618,
+ "learning_rate": 0.0006,
+ "loss": 4.527771472930908,
+ "step": 1331
+ },
+ {
+ "epoch": 18.503056768558952,
+ "grad_norm": 0.018825586885213852,
+ "learning_rate": 0.0006,
+ "loss": 4.485386848449707,
+ "step": 1332
+ },
+ {
+ "epoch": 18.51703056768559,
+ "grad_norm": 0.019258007407188416,
+ "learning_rate": 0.0006,
+ "loss": 4.414972305297852,
+ "step": 1333
+ },
+ {
+ "epoch": 18.531004366812226,
+ "grad_norm": 0.018192021176218987,
+ "learning_rate": 0.0006,
+ "loss": 4.412166595458984,
+ "step": 1334
+ },
+ {
+ "epoch": 18.544978165938865,
+ "grad_norm": 0.016120698302984238,
+ "learning_rate": 0.0006,
+ "loss": 4.484317779541016,
+ "step": 1335
+ },
+ {
+ "epoch": 18.558951965065503,
+ "grad_norm": 0.016496622934937477,
+ "learning_rate": 0.0006,
+ "loss": 4.514986515045166,
+ "step": 1336
+ },
+ {
+ "epoch": 18.57292576419214,
+ "grad_norm": 0.01795116998255253,
+ "learning_rate": 0.0006,
+ "loss": 4.41473388671875,
+ "step": 1337
+ },
+ {
+ "epoch": 18.586899563318777,
+ "grad_norm": 0.019062276929616928,
+ "learning_rate": 0.0006,
+ "loss": 4.601996898651123,
+ "step": 1338
+ },
+ {
+ "epoch": 18.600873362445416,
+ "grad_norm": 0.017174631357192993,
+ "learning_rate": 0.0006,
+ "loss": 4.571218490600586,
+ "step": 1339
+ },
+ {
+ "epoch": 18.61484716157205,
+ "grad_norm": 0.014015606604516506,
+ "learning_rate": 0.0006,
+ "loss": 4.526606559753418,
+ "step": 1340
+ },
+ {
+ "epoch": 18.62882096069869,
+ "grad_norm": 0.015358510427176952,
+ "learning_rate": 0.0006,
+ "loss": 4.604286193847656,
+ "step": 1341
+ },
+ {
+ "epoch": 18.64279475982533,
+ "grad_norm": 0.01619391329586506,
+ "learning_rate": 0.0006,
+ "loss": 4.560883522033691,
+ "step": 1342
+ },
+ {
+ "epoch": 18.656768558951963,
+ "grad_norm": 0.015972964465618134,
+ "learning_rate": 0.0006,
+ "loss": 4.479755401611328,
+ "step": 1343
+ },
+ {
+ "epoch": 18.670742358078602,
+ "grad_norm": 0.016861025243997574,
+ "learning_rate": 0.0006,
+ "loss": 4.474471092224121,
+ "step": 1344
+ },
+ {
+ "epoch": 18.68471615720524,
+ "grad_norm": 0.01822078227996826,
+ "learning_rate": 0.0006,
+ "loss": 4.509489059448242,
+ "step": 1345
+ },
+ {
+ "epoch": 18.69868995633188,
+ "grad_norm": 0.019415007904171944,
+ "learning_rate": 0.0006,
+ "loss": 4.508941173553467,
+ "step": 1346
+ },
+ {
+ "epoch": 18.712663755458514,
+ "grad_norm": 0.019856126978993416,
+ "learning_rate": 0.0006,
+ "loss": 4.446670055389404,
+ "step": 1347
+ },
+ {
+ "epoch": 18.726637554585153,
+ "grad_norm": 0.01842520199716091,
+ "learning_rate": 0.0006,
+ "loss": 4.524422645568848,
+ "step": 1348
+ },
+ {
+ "epoch": 18.74061135371179,
+ "grad_norm": 0.017196480184793472,
+ "learning_rate": 0.0006,
+ "loss": 4.498466968536377,
+ "step": 1349
+ },
+ {
+ "epoch": 18.754585152838427,
+ "grad_norm": 0.01823010854423046,
+ "learning_rate": 0.0006,
+ "loss": 4.534684181213379,
+ "step": 1350
+ },
+ {
+ "epoch": 18.768558951965066,
+ "grad_norm": 0.020398065447807312,
+ "learning_rate": 0.0006,
+ "loss": 4.456494331359863,
+ "step": 1351
+ },
+ {
+ "epoch": 18.782532751091704,
+ "grad_norm": 0.022992262616753578,
+ "learning_rate": 0.0006,
+ "loss": 4.476374626159668,
+ "step": 1352
+ },
+ {
+ "epoch": 18.79650655021834,
+ "grad_norm": 0.02235420234501362,
+ "learning_rate": 0.0006,
+ "loss": 4.587776184082031,
+ "step": 1353
+ },
+ {
+ "epoch": 18.810480349344978,
+ "grad_norm": 0.02093626745045185,
+ "learning_rate": 0.0006,
+ "loss": 4.424172401428223,
+ "step": 1354
+ },
+ {
+ "epoch": 18.824454148471617,
+ "grad_norm": 0.01997302658855915,
+ "learning_rate": 0.0006,
+ "loss": 4.448066711425781,
+ "step": 1355
+ },
+ {
+ "epoch": 18.83842794759825,
+ "grad_norm": 0.018265459686517715,
+ "learning_rate": 0.0006,
+ "loss": 4.436610698699951,
+ "step": 1356
+ },
+ {
+ "epoch": 18.85240174672489,
+ "grad_norm": 0.01673833839595318,
+ "learning_rate": 0.0006,
+ "loss": 4.5502214431762695,
+ "step": 1357
+ },
+ {
+ "epoch": 18.86637554585153,
+ "grad_norm": 0.017908833920955658,
+ "learning_rate": 0.0006,
+ "loss": 4.550530433654785,
+ "step": 1358
+ },
+ {
+ "epoch": 18.880349344978168,
+ "grad_norm": 0.01796654611825943,
+ "learning_rate": 0.0006,
+ "loss": 4.531463146209717,
+ "step": 1359
+ },
+ {
+ "epoch": 18.894323144104803,
+ "grad_norm": 0.01646183803677559,
+ "learning_rate": 0.0006,
+ "loss": 4.498582363128662,
+ "step": 1360
+ },
+ {
+ "epoch": 18.90829694323144,
+ "grad_norm": 0.015430327504873276,
+ "learning_rate": 0.0006,
+ "loss": 4.52571439743042,
+ "step": 1361
+ },
+ {
+ "epoch": 18.92227074235808,
+ "grad_norm": 0.016527952626347542,
+ "learning_rate": 0.0006,
+ "loss": 4.547013759613037,
+ "step": 1362
+ },
+ {
+ "epoch": 18.936244541484715,
+ "grad_norm": 0.016263660043478012,
+ "learning_rate": 0.0006,
+ "loss": 4.468754291534424,
+ "step": 1363
+ },
+ {
+ "epoch": 18.950218340611354,
+ "grad_norm": 0.016347210854291916,
+ "learning_rate": 0.0006,
+ "loss": 4.450741767883301,
+ "step": 1364
+ },
+ {
+ "epoch": 18.964192139737992,
+ "grad_norm": 0.01701398566365242,
+ "learning_rate": 0.0006,
+ "loss": 4.534899711608887,
+ "step": 1365
+ },
+ {
+ "epoch": 18.978165938864628,
+ "grad_norm": 0.017693081870675087,
+ "learning_rate": 0.0006,
+ "loss": 4.460553169250488,
+ "step": 1366
+ },
+ {
+ "epoch": 18.992139737991266,
+ "grad_norm": 0.018746308982372284,
+ "learning_rate": 0.0006,
+ "loss": 4.538792610168457,
+ "step": 1367
+ },
+ {
+ "epoch": 19.0,
+ "grad_norm": 0.020099416375160217,
+ "learning_rate": 0.0006,
+ "loss": 4.449617385864258,
+ "step": 1368
+ },
+ {
+ "epoch": 19.0,
+ "eval_loss": 4.7756571769714355,
+ "eval_runtime": 57.1804,
+ "eval_samples_per_second": 42.707,
+ "eval_steps_per_second": 1.347,
+ "step": 1368
+ },
+ {
+ "epoch": 19.01397379912664,
+ "grad_norm": 0.018726933747529984,
+ "learning_rate": 0.0006,
+ "loss": 4.446059703826904,
+ "step": 1369
+ },
+ {
+ "epoch": 19.027947598253274,
+ "grad_norm": 0.01956087350845337,
+ "learning_rate": 0.0006,
+ "loss": 4.4520039558410645,
+ "step": 1370
+ },
+ {
+ "epoch": 19.041921397379912,
+ "grad_norm": 0.020092811435461044,
+ "learning_rate": 0.0006,
+ "loss": 4.543343544006348,
+ "step": 1371
+ },
+ {
+ "epoch": 19.05589519650655,
+ "grad_norm": 0.01843622885644436,
+ "learning_rate": 0.0006,
+ "loss": 4.535579204559326,
+ "step": 1372
+ },
+ {
+ "epoch": 19.069868995633186,
+ "grad_norm": 0.016989488154649734,
+ "learning_rate": 0.0006,
+ "loss": 4.52316951751709,
+ "step": 1373
+ },
+ {
+ "epoch": 19.083842794759825,
+ "grad_norm": 0.01805422455072403,
+ "learning_rate": 0.0006,
+ "loss": 4.5393266677856445,
+ "step": 1374
+ },
+ {
+ "epoch": 19.097816593886463,
+ "grad_norm": 0.01874629408121109,
+ "learning_rate": 0.0006,
+ "loss": 4.422059059143066,
+ "step": 1375
+ },
+ {
+ "epoch": 19.111790393013102,
+ "grad_norm": 0.01804221048951149,
+ "learning_rate": 0.0006,
+ "loss": 4.50508451461792,
+ "step": 1376
+ },
+ {
+ "epoch": 19.125764192139737,
+ "grad_norm": 0.01785474829375744,
+ "learning_rate": 0.0006,
+ "loss": 4.412599563598633,
+ "step": 1377
+ },
+ {
+ "epoch": 19.139737991266376,
+ "grad_norm": 0.01971791312098503,
+ "learning_rate": 0.0006,
+ "loss": 4.518318176269531,
+ "step": 1378
+ },
+ {
+ "epoch": 19.153711790393015,
+ "grad_norm": 0.02061633951961994,
+ "learning_rate": 0.0006,
+ "loss": 4.332393169403076,
+ "step": 1379
+ },
+ {
+ "epoch": 19.16768558951965,
+ "grad_norm": 0.020563429221510887,
+ "learning_rate": 0.0006,
+ "loss": 4.445863723754883,
+ "step": 1380
+ },
+ {
+ "epoch": 19.18165938864629,
+ "grad_norm": 0.019995713606476784,
+ "learning_rate": 0.0006,
+ "loss": 4.497312068939209,
+ "step": 1381
+ },
+ {
+ "epoch": 19.195633187772927,
+ "grad_norm": 0.019823070615530014,
+ "learning_rate": 0.0006,
+ "loss": 4.475085258483887,
+ "step": 1382
+ },
+ {
+ "epoch": 19.209606986899562,
+ "grad_norm": 0.019039355218410492,
+ "learning_rate": 0.0006,
+ "loss": 4.50432014465332,
+ "step": 1383
+ },
+ {
+ "epoch": 19.2235807860262,
+ "grad_norm": 0.017876973375678062,
+ "learning_rate": 0.0006,
+ "loss": 4.384788513183594,
+ "step": 1384
+ },
+ {
+ "epoch": 19.23755458515284,
+ "grad_norm": 0.018697699531912804,
+ "learning_rate": 0.0006,
+ "loss": 4.410778045654297,
+ "step": 1385
+ },
+ {
+ "epoch": 19.251528384279474,
+ "grad_norm": 0.018826806917786598,
+ "learning_rate": 0.0006,
+ "loss": 4.524034023284912,
+ "step": 1386
+ },
+ {
+ "epoch": 19.265502183406113,
+ "grad_norm": 0.01824226602911949,
+ "learning_rate": 0.0006,
+ "loss": 4.526393890380859,
+ "step": 1387
+ },
+ {
+ "epoch": 19.27947598253275,
+ "grad_norm": 0.01753208599984646,
+ "learning_rate": 0.0006,
+ "loss": 4.493703365325928,
+ "step": 1388
+ },
+ {
+ "epoch": 19.29344978165939,
+ "grad_norm": 0.01742432825267315,
+ "learning_rate": 0.0006,
+ "loss": 4.392831802368164,
+ "step": 1389
+ },
+ {
+ "epoch": 19.307423580786025,
+ "grad_norm": 0.017099356278777122,
+ "learning_rate": 0.0006,
+ "loss": 4.578668594360352,
+ "step": 1390
+ },
+ {
+ "epoch": 19.321397379912664,
+ "grad_norm": 0.017946023494005203,
+ "learning_rate": 0.0006,
+ "loss": 4.456718444824219,
+ "step": 1391
+ },
+ {
+ "epoch": 19.335371179039303,
+ "grad_norm": 0.017705943435430527,
+ "learning_rate": 0.0006,
+ "loss": 4.553682327270508,
+ "step": 1392
+ },
+ {
+ "epoch": 19.349344978165938,
+ "grad_norm": 0.016647523269057274,
+ "learning_rate": 0.0006,
+ "loss": 4.525016784667969,
+ "step": 1393
+ },
+ {
+ "epoch": 19.363318777292577,
+ "grad_norm": 0.015407416969537735,
+ "learning_rate": 0.0006,
+ "loss": 4.500231742858887,
+ "step": 1394
+ },
+ {
+ "epoch": 19.377292576419215,
+ "grad_norm": 0.017264019697904587,
+ "learning_rate": 0.0006,
+ "loss": 4.4615349769592285,
+ "step": 1395
+ },
+ {
+ "epoch": 19.39126637554585,
+ "grad_norm": 0.01871555484831333,
+ "learning_rate": 0.0006,
+ "loss": 4.281026840209961,
+ "step": 1396
+ },
+ {
+ "epoch": 19.40524017467249,
+ "grad_norm": 0.017361650243401527,
+ "learning_rate": 0.0006,
+ "loss": 4.479382514953613,
+ "step": 1397
+ },
+ {
+ "epoch": 19.419213973799128,
+ "grad_norm": 0.019781362265348434,
+ "learning_rate": 0.0006,
+ "loss": 4.402320861816406,
+ "step": 1398
+ },
+ {
+ "epoch": 19.433187772925763,
+ "grad_norm": 0.019043298438191414,
+ "learning_rate": 0.0006,
+ "loss": 4.476983070373535,
+ "step": 1399
+ },
+ {
+ "epoch": 19.4471615720524,
+ "grad_norm": 0.017137985676527023,
+ "learning_rate": 0.0006,
+ "loss": 4.418689250946045,
+ "step": 1400
+ },
+ {
+ "epoch": 19.46113537117904,
+ "grad_norm": 0.016481833532452583,
+ "learning_rate": 0.0006,
+ "loss": 4.579145431518555,
+ "step": 1401
+ },
+ {
+ "epoch": 19.475109170305675,
+ "grad_norm": 0.016507714986801147,
+ "learning_rate": 0.0006,
+ "loss": 4.519158363342285,
+ "step": 1402
+ },
+ {
+ "epoch": 19.489082969432314,
+ "grad_norm": 0.018092872574925423,
+ "learning_rate": 0.0006,
+ "loss": 4.513748645782471,
+ "step": 1403
+ },
+ {
+ "epoch": 19.503056768558952,
+ "grad_norm": 0.018153440207242966,
+ "learning_rate": 0.0006,
+ "loss": 4.645384311676025,
+ "step": 1404
+ },
+ {
+ "epoch": 19.51703056768559,
+ "grad_norm": 0.01657119207084179,
+ "learning_rate": 0.0006,
+ "loss": 4.399757385253906,
+ "step": 1405
+ },
+ {
+ "epoch": 19.531004366812226,
+ "grad_norm": 0.0156280305236578,
+ "learning_rate": 0.0006,
+ "loss": 4.467836380004883,
+ "step": 1406
+ },
+ {
+ "epoch": 19.544978165938865,
+ "grad_norm": 0.016838403418660164,
+ "learning_rate": 0.0006,
+ "loss": 4.529176235198975,
+ "step": 1407
+ },
+ {
+ "epoch": 19.558951965065503,
+ "grad_norm": 0.01743236929178238,
+ "learning_rate": 0.0006,
+ "loss": 4.440032482147217,
+ "step": 1408
+ },
+ {
+ "epoch": 19.57292576419214,
+ "grad_norm": 0.01553610060364008,
+ "learning_rate": 0.0006,
+ "loss": 4.425588607788086,
+ "step": 1409
+ },
+ {
+ "epoch": 19.586899563318777,
+ "grad_norm": 0.017259271815419197,
+ "learning_rate": 0.0006,
+ "loss": 4.405452728271484,
+ "step": 1410
+ },
+ {
+ "epoch": 19.600873362445416,
+ "grad_norm": 0.015568481758236885,
+ "learning_rate": 0.0006,
+ "loss": 4.444279193878174,
+ "step": 1411
+ },
+ {
+ "epoch": 19.61484716157205,
+ "grad_norm": 0.015331248752772808,
+ "learning_rate": 0.0006,
+ "loss": 4.451846122741699,
+ "step": 1412
+ },
+ {
+ "epoch": 19.62882096069869,
+ "grad_norm": 0.015667378902435303,
+ "learning_rate": 0.0006,
+ "loss": 4.533931255340576,
+ "step": 1413
+ },
+ {
+ "epoch": 19.64279475982533,
+ "grad_norm": 0.016177602112293243,
+ "learning_rate": 0.0006,
+ "loss": 4.467313289642334,
+ "step": 1414
+ },
+ {
+ "epoch": 19.656768558951963,
+ "grad_norm": 0.01733129471540451,
+ "learning_rate": 0.0006,
+ "loss": 4.496893882751465,
+ "step": 1415
+ },
+ {
+ "epoch": 19.670742358078602,
+ "grad_norm": 0.016539594158530235,
+ "learning_rate": 0.0006,
+ "loss": 4.401961326599121,
+ "step": 1416
+ },
+ {
+ "epoch": 19.68471615720524,
+ "grad_norm": 0.016058551147580147,
+ "learning_rate": 0.0006,
+ "loss": 4.503375053405762,
+ "step": 1417
+ },
+ {
+ "epoch": 19.69868995633188,
+ "grad_norm": 0.01715777814388275,
+ "learning_rate": 0.0006,
+ "loss": 4.3808369636535645,
+ "step": 1418
+ },
+ {
+ "epoch": 19.712663755458514,
+ "grad_norm": 0.01681119203567505,
+ "learning_rate": 0.0006,
+ "loss": 4.5469560623168945,
+ "step": 1419
+ },
+ {
+ "epoch": 19.726637554585153,
+ "grad_norm": 0.01634550653398037,
+ "learning_rate": 0.0006,
+ "loss": 4.328197956085205,
+ "step": 1420
+ },
+ {
+ "epoch": 19.74061135371179,
+ "grad_norm": 0.01549347210675478,
+ "learning_rate": 0.0006,
+ "loss": 4.534621238708496,
+ "step": 1421
+ },
+ {
+ "epoch": 19.754585152838427,
+ "grad_norm": 0.01654665172100067,
+ "learning_rate": 0.0006,
+ "loss": 4.4724225997924805,
+ "step": 1422
+ },
+ {
+ "epoch": 19.768558951965066,
+ "grad_norm": 0.017119025811553,
+ "learning_rate": 0.0006,
+ "loss": 4.3897247314453125,
+ "step": 1423
+ },
+ {
+ "epoch": 19.782532751091704,
+ "grad_norm": 0.01603223942220211,
+ "learning_rate": 0.0006,
+ "loss": 4.412731170654297,
+ "step": 1424
+ },
+ {
+ "epoch": 19.79650655021834,
+ "grad_norm": 0.016774067655205727,
+ "learning_rate": 0.0006,
+ "loss": 4.54707145690918,
+ "step": 1425
+ },
+ {
+ "epoch": 19.810480349344978,
+ "grad_norm": 0.016453025862574577,
+ "learning_rate": 0.0006,
+ "loss": 4.462501049041748,
+ "step": 1426
+ },
+ {
+ "epoch": 19.824454148471617,
+ "grad_norm": 0.01597629487514496,
+ "learning_rate": 0.0006,
+ "loss": 4.229758262634277,
+ "step": 1427
+ },
+ {
+ "epoch": 19.83842794759825,
+ "grad_norm": 0.017999034374952316,
+ "learning_rate": 0.0006,
+ "loss": 4.389263153076172,
+ "step": 1428
+ },
+ {
+ "epoch": 19.85240174672489,
+ "grad_norm": 0.02165898121893406,
+ "learning_rate": 0.0006,
+ "loss": 4.342130661010742,
+ "step": 1429
+ },
+ {
+ "epoch": 19.86637554585153,
+ "grad_norm": 0.02222822792828083,
+ "learning_rate": 0.0006,
+ "loss": 4.492661476135254,
+ "step": 1430
+ },
+ {
+ "epoch": 19.880349344978168,
+ "grad_norm": 0.019378066062927246,
+ "learning_rate": 0.0006,
+ "loss": 4.4614105224609375,
+ "step": 1431
+ },
+ {
+ "epoch": 19.894323144104803,
+ "grad_norm": 0.019908083602786064,
+ "learning_rate": 0.0006,
+ "loss": 4.370929718017578,
+ "step": 1432
+ },
+ {
+ "epoch": 19.90829694323144,
+ "grad_norm": 0.022296471521258354,
+ "learning_rate": 0.0006,
+ "loss": 4.438320159912109,
+ "step": 1433
+ },
+ {
+ "epoch": 19.92227074235808,
+ "grad_norm": 0.021810343489050865,
+ "learning_rate": 0.0006,
+ "loss": 4.462067127227783,
+ "step": 1434
+ },
+ {
+ "epoch": 19.936244541484715,
+ "grad_norm": 0.02129209227859974,
+ "learning_rate": 0.0006,
+ "loss": 4.541261196136475,
+ "step": 1435
+ },
+ {
+ "epoch": 19.950218340611354,
+ "grad_norm": 0.02079629711806774,
+ "learning_rate": 0.0006,
+ "loss": 4.36712646484375,
+ "step": 1436
+ },
+ {
+ "epoch": 19.964192139737992,
+ "grad_norm": 0.022501740604639053,
+ "learning_rate": 0.0006,
+ "loss": 4.5279083251953125,
+ "step": 1437
+ },
+ {
+ "epoch": 19.978165938864628,
+ "grad_norm": 0.019121676683425903,
+ "learning_rate": 0.0006,
+ "loss": 4.477789878845215,
+ "step": 1438
+ },
+ {
+ "epoch": 19.992139737991266,
+ "grad_norm": 0.017429588362574577,
+ "learning_rate": 0.0006,
+ "loss": 4.4023847579956055,
+ "step": 1439
+ },
+ {
+ "epoch": 20.0,
+ "grad_norm": 0.018857665359973907,
+ "learning_rate": 0.0006,
+ "loss": 4.494743347167969,
+ "step": 1440
+ },
+ {
+ "epoch": 20.0,
+ "eval_loss": 4.694555759429932,
+ "eval_runtime": 56.4146,
+ "eval_samples_per_second": 43.287,
+ "eval_steps_per_second": 1.365,
+ "step": 1440
+ },
+ {
+ "epoch": 20.01397379912664,
+ "grad_norm": 0.016505861654877663,
+ "learning_rate": 0.0006,
+ "loss": 4.506229877471924,
+ "step": 1441
+ },
+ {
+ "epoch": 20.027947598253274,
+ "grad_norm": 0.017571842297911644,
+ "learning_rate": 0.0006,
+ "loss": 4.384073257446289,
+ "step": 1442
+ },
+ {
+ "epoch": 20.041921397379912,
+ "grad_norm": 0.019398365169763565,
+ "learning_rate": 0.0006,
+ "loss": 4.482177734375,
+ "step": 1443
+ },
+ {
+ "epoch": 20.05589519650655,
+ "grad_norm": 0.02011062018573284,
+ "learning_rate": 0.0006,
+ "loss": 4.440773963928223,
+ "step": 1444
+ },
+ {
+ "epoch": 20.069868995633186,
+ "grad_norm": 0.019026953727006912,
+ "learning_rate": 0.0006,
+ "loss": 4.507168769836426,
+ "step": 1445
+ },
+ {
+ "epoch": 20.083842794759825,
+ "grad_norm": 0.020506031811237335,
+ "learning_rate": 0.0006,
+ "loss": 4.469079971313477,
+ "step": 1446
+ },
+ {
+ "epoch": 20.097816593886463,
+ "grad_norm": 0.019608162343502045,
+ "learning_rate": 0.0006,
+ "loss": 4.395927429199219,
+ "step": 1447
+ },
+ {
+ "epoch": 20.111790393013102,
+ "grad_norm": 0.019419124349951744,
+ "learning_rate": 0.0006,
+ "loss": 4.484235763549805,
+ "step": 1448
+ },
+ {
+ "epoch": 20.125764192139737,
+ "grad_norm": 0.021775128319859505,
+ "learning_rate": 0.0006,
+ "loss": 4.333173751831055,
+ "step": 1449
+ },
+ {
+ "epoch": 20.139737991266376,
+ "grad_norm": 0.025512272492051125,
+ "learning_rate": 0.0006,
+ "loss": 4.460362434387207,
+ "step": 1450
+ },
+ {
+ "epoch": 20.153711790393015,
+ "grad_norm": 0.025157401338219643,
+ "learning_rate": 0.0006,
+ "loss": 4.4361395835876465,
+ "step": 1451
+ },
+ {
+ "epoch": 20.16768558951965,
+ "grad_norm": 0.019862616434693336,
+ "learning_rate": 0.0006,
+ "loss": 4.426807403564453,
+ "step": 1452
+ },
+ {
+ "epoch": 20.18165938864629,
+ "grad_norm": 0.01932152360677719,
+ "learning_rate": 0.0006,
+ "loss": 4.375677108764648,
+ "step": 1453
+ },
+ {
+ "epoch": 20.195633187772927,
+ "grad_norm": 0.01760999858379364,
+ "learning_rate": 0.0006,
+ "loss": 4.42259407043457,
+ "step": 1454
+ },
+ {
+ "epoch": 20.209606986899562,
+ "grad_norm": 0.016998112201690674,
+ "learning_rate": 0.0006,
+ "loss": 4.481280326843262,
+ "step": 1455
+ },
+ {
+ "epoch": 20.2235807860262,
+ "grad_norm": 0.017199190333485603,
+ "learning_rate": 0.0006,
+ "loss": 4.496712684631348,
+ "step": 1456
+ },
+ {
+ "epoch": 20.23755458515284,
+ "grad_norm": 0.016833679750561714,
+ "learning_rate": 0.0006,
+ "loss": 4.40736722946167,
+ "step": 1457
+ },
+ {
+ "epoch": 20.251528384279474,
+ "grad_norm": 0.017546730116009712,
+ "learning_rate": 0.0006,
+ "loss": 4.4915266036987305,
+ "step": 1458
+ },
+ {
+ "epoch": 20.265502183406113,
+ "grad_norm": 0.016595078632235527,
+ "learning_rate": 0.0006,
+ "loss": 4.460815906524658,
+ "step": 1459
+ },
+ {
+ "epoch": 20.27947598253275,
+ "grad_norm": 0.017728568986058235,
+ "learning_rate": 0.0006,
+ "loss": 4.346048831939697,
+ "step": 1460
+ },
+ {
+ "epoch": 20.29344978165939,
+ "grad_norm": 0.01849197782576084,
+ "learning_rate": 0.0006,
+ "loss": 4.399681091308594,
+ "step": 1461
+ },
+ {
+ "epoch": 20.307423580786025,
+ "grad_norm": 0.018336588516831398,
+ "learning_rate": 0.0006,
+ "loss": 4.414183616638184,
+ "step": 1462
+ },
+ {
+ "epoch": 20.321397379912664,
+ "grad_norm": 0.018086519092321396,
+ "learning_rate": 0.0006,
+ "loss": 4.346586227416992,
+ "step": 1463
+ },
+ {
+ "epoch": 20.335371179039303,
+ "grad_norm": 0.019697299227118492,
+ "learning_rate": 0.0006,
+ "loss": 4.4420952796936035,
+ "step": 1464
+ },
+ {
+ "epoch": 20.349344978165938,
+ "grad_norm": 0.02198074571788311,
+ "learning_rate": 0.0006,
+ "loss": 4.588729381561279,
+ "step": 1465
+ },
+ {
+ "epoch": 20.363318777292577,
+ "grad_norm": 0.022813329473137856,
+ "learning_rate": 0.0006,
+ "loss": 4.49729585647583,
+ "step": 1466
+ },
+ {
+ "epoch": 20.377292576419215,
+ "grad_norm": 0.022006649523973465,
+ "learning_rate": 0.0006,
+ "loss": 4.475641250610352,
+ "step": 1467
+ },
+ {
+ "epoch": 20.39126637554585,
+ "grad_norm": 0.023188438266515732,
+ "learning_rate": 0.0006,
+ "loss": 4.404322624206543,
+ "step": 1468
+ },
+ {
+ "epoch": 20.40524017467249,
+ "grad_norm": 0.022226877510547638,
+ "learning_rate": 0.0006,
+ "loss": 4.50904655456543,
+ "step": 1469
+ },
+ {
+ "epoch": 20.419213973799128,
+ "grad_norm": 0.021082637831568718,
+ "learning_rate": 0.0006,
+ "loss": 4.339011192321777,
+ "step": 1470
+ },
+ {
+ "epoch": 20.433187772925763,
+ "grad_norm": 0.02013542130589485,
+ "learning_rate": 0.0006,
+ "loss": 4.516942024230957,
+ "step": 1471
+ },
+ {
+ "epoch": 20.4471615720524,
+ "grad_norm": 0.01898968778550625,
+ "learning_rate": 0.0006,
+ "loss": 4.315804481506348,
+ "step": 1472
+ },
+ {
+ "epoch": 20.46113537117904,
+ "grad_norm": 0.019450880587100983,
+ "learning_rate": 0.0006,
+ "loss": 4.4794921875,
+ "step": 1473
+ },
+ {
+ "epoch": 20.475109170305675,
+ "grad_norm": 0.018577802926301956,
+ "learning_rate": 0.0006,
+ "loss": 4.422323703765869,
+ "step": 1474
+ },
+ {
+ "epoch": 20.489082969432314,
+ "grad_norm": 0.01778412237763405,
+ "learning_rate": 0.0006,
+ "loss": 4.429789066314697,
+ "step": 1475
+ },
+ {
+ "epoch": 20.503056768558952,
+ "grad_norm": 0.017134476453065872,
+ "learning_rate": 0.0006,
+ "loss": 4.416459083557129,
+ "step": 1476
+ },
+ {
+ "epoch": 20.51703056768559,
+ "grad_norm": 0.017153557389974594,
+ "learning_rate": 0.0006,
+ "loss": 4.446170806884766,
+ "step": 1477
+ },
+ {
+ "epoch": 20.531004366812226,
+ "grad_norm": 0.014998532831668854,
+ "learning_rate": 0.0006,
+ "loss": 4.419447898864746,
+ "step": 1478
+ },
+ {
+ "epoch": 20.544978165938865,
+ "grad_norm": 0.015877850353717804,
+ "learning_rate": 0.0006,
+ "loss": 4.312338829040527,
+ "step": 1479
+ },
+ {
+ "epoch": 20.558951965065503,
+ "grad_norm": 0.01569489948451519,
+ "learning_rate": 0.0006,
+ "loss": 4.572905540466309,
+ "step": 1480
+ },
+ {
+ "epoch": 20.57292576419214,
+ "grad_norm": 0.01591474376618862,
+ "learning_rate": 0.0006,
+ "loss": 4.414907455444336,
+ "step": 1481
+ },
+ {
+ "epoch": 20.586899563318777,
+ "grad_norm": 0.015395903028547764,
+ "learning_rate": 0.0006,
+ "loss": 4.549131393432617,
+ "step": 1482
+ },
+ {
+ "epoch": 20.600873362445416,
+ "grad_norm": 0.015715358778834343,
+ "learning_rate": 0.0006,
+ "loss": 4.530194282531738,
+ "step": 1483
+ },
+ {
+ "epoch": 20.61484716157205,
+ "grad_norm": 0.016042975708842278,
+ "learning_rate": 0.0006,
+ "loss": 4.322149276733398,
+ "step": 1484
+ },
+ {
+ "epoch": 20.62882096069869,
+ "grad_norm": 0.01664186641573906,
+ "learning_rate": 0.0006,
+ "loss": 4.424866676330566,
+ "step": 1485
+ },
+ {
+ "epoch": 20.64279475982533,
+ "grad_norm": 0.01631534844636917,
+ "learning_rate": 0.0006,
+ "loss": 4.441752910614014,
+ "step": 1486
+ },
+ {
+ "epoch": 20.656768558951963,
+ "grad_norm": 0.014057058840990067,
+ "learning_rate": 0.0006,
+ "loss": 4.399982452392578,
+ "step": 1487
+ },
+ {
+ "epoch": 20.670742358078602,
+ "grad_norm": 0.015076273120939732,
+ "learning_rate": 0.0006,
+ "loss": 4.441324234008789,
+ "step": 1488
+ },
+ {
+ "epoch": 20.68471615720524,
+ "grad_norm": 0.014138277620077133,
+ "learning_rate": 0.0006,
+ "loss": 4.4312543869018555,
+ "step": 1489
+ },
+ {
+ "epoch": 20.69868995633188,
+ "grad_norm": 0.013762440532445908,
+ "learning_rate": 0.0006,
+ "loss": 4.387807846069336,
+ "step": 1490
+ },
+ {
+ "epoch": 20.712663755458514,
+ "grad_norm": 0.015058411285281181,
+ "learning_rate": 0.0006,
+ "loss": 4.594747543334961,
+ "step": 1491
+ },
+ {
+ "epoch": 20.726637554585153,
+ "grad_norm": 0.013827123679220676,
+ "learning_rate": 0.0006,
+ "loss": 4.434604167938232,
+ "step": 1492
+ },
+ {
+ "epoch": 20.74061135371179,
+ "grad_norm": 0.015315295197069645,
+ "learning_rate": 0.0006,
+ "loss": 4.588066101074219,
+ "step": 1493
+ },
+ {
+ "epoch": 20.754585152838427,
+ "grad_norm": 0.018678732216358185,
+ "learning_rate": 0.0006,
+ "loss": 4.3514204025268555,
+ "step": 1494
+ },
+ {
+ "epoch": 20.768558951965066,
+ "grad_norm": 0.019521350041031837,
+ "learning_rate": 0.0006,
+ "loss": 4.437658786773682,
+ "step": 1495
+ },
+ {
+ "epoch": 20.782532751091704,
+ "grad_norm": 0.018363403156399727,
+ "learning_rate": 0.0006,
+ "loss": 4.495002269744873,
+ "step": 1496
+ },
+ {
+ "epoch": 20.79650655021834,
+ "grad_norm": 0.016830725595355034,
+ "learning_rate": 0.0006,
+ "loss": 4.402471542358398,
+ "step": 1497
+ },
+ {
+ "epoch": 20.810480349344978,
+ "grad_norm": 0.01722072623670101,
+ "learning_rate": 0.0006,
+ "loss": 4.437653541564941,
+ "step": 1498
+ },
+ {
+ "epoch": 20.824454148471617,
+ "grad_norm": 0.017459599301218987,
+ "learning_rate": 0.0006,
+ "loss": 4.366603851318359,
+ "step": 1499
+ },
+ {
+ "epoch": 20.83842794759825,
+ "grad_norm": 0.018902219831943512,
+ "learning_rate": 0.0006,
+ "loss": 4.501561164855957,
+ "step": 1500
+ },
+ {
+ "epoch": 20.85240174672489,
+ "grad_norm": 0.020814096555113792,
+ "learning_rate": 0.0006,
+ "loss": 4.535295486450195,
+ "step": 1501
+ },
+ {
+ "epoch": 20.86637554585153,
+ "grad_norm": 0.018990976735949516,
+ "learning_rate": 0.0006,
+ "loss": 4.333580017089844,
+ "step": 1502
+ },
+ {
+ "epoch": 20.880349344978168,
+ "grad_norm": 0.01837880350649357,
+ "learning_rate": 0.0006,
+ "loss": 4.313388824462891,
+ "step": 1503
+ },
+ {
+ "epoch": 20.894323144104803,
+ "grad_norm": 0.019820360466837883,
+ "learning_rate": 0.0006,
+ "loss": 4.388157844543457,
+ "step": 1504
+ },
+ {
+ "epoch": 20.90829694323144,
+ "grad_norm": 0.020181143656373024,
+ "learning_rate": 0.0006,
+ "loss": 4.572224140167236,
+ "step": 1505
+ },
+ {
+ "epoch": 20.92227074235808,
+ "grad_norm": 0.021206077188253403,
+ "learning_rate": 0.0006,
+ "loss": 4.3127970695495605,
+ "step": 1506
+ },
+ {
+ "epoch": 20.936244541484715,
+ "grad_norm": 0.017962060868740082,
+ "learning_rate": 0.0006,
+ "loss": 4.384734630584717,
+ "step": 1507
+ },
+ {
+ "epoch": 20.950218340611354,
+ "grad_norm": 0.01737220585346222,
+ "learning_rate": 0.0006,
+ "loss": 4.4075751304626465,
+ "step": 1508
+ },
+ {
+ "epoch": 20.964192139737992,
+ "grad_norm": 0.01839268207550049,
+ "learning_rate": 0.0006,
+ "loss": 4.553779125213623,
+ "step": 1509
+ },
+ {
+ "epoch": 20.978165938864628,
+ "grad_norm": 0.020102640613913536,
+ "learning_rate": 0.0006,
+ "loss": 4.399398326873779,
+ "step": 1510
+ },
+ {
+ "epoch": 20.992139737991266,
+ "grad_norm": 0.018831463530659676,
+ "learning_rate": 0.0006,
+ "loss": 4.358499526977539,
+ "step": 1511
+ },
+ {
+ "epoch": 21.0,
+ "grad_norm": 0.018254734575748444,
+ "learning_rate": 0.0006,
+ "loss": 4.384119510650635,
+ "step": 1512
+ },
+ {
+ "epoch": 21.0,
+ "eval_loss": 4.7524237632751465,
+ "eval_runtime": 57.8812,
+ "eval_samples_per_second": 42.19,
+ "eval_steps_per_second": 1.33,
+ "step": 1512
+ },
+ {
+ "epoch": 21.01397379912664,
+ "grad_norm": 0.019728587940335274,
+ "learning_rate": 0.0006,
+ "loss": 4.32363224029541,
+ "step": 1513
+ },
+ {
+ "epoch": 21.027947598253274,
+ "grad_norm": 0.02216625027358532,
+ "learning_rate": 0.0006,
+ "loss": 4.436988353729248,
+ "step": 1514
+ },
+ {
+ "epoch": 21.041921397379912,
+ "grad_norm": 0.02227606810629368,
+ "learning_rate": 0.0006,
+ "loss": 4.406063556671143,
+ "step": 1515
+ },
+ {
+ "epoch": 21.05589519650655,
+ "grad_norm": 0.020187009125947952,
+ "learning_rate": 0.0006,
+ "loss": 4.324402332305908,
+ "step": 1516
+ },
+ {
+ "epoch": 21.069868995633186,
+ "grad_norm": 0.02095157466828823,
+ "learning_rate": 0.0006,
+ "loss": 4.456544876098633,
+ "step": 1517
+ },
+ {
+ "epoch": 21.083842794759825,
+ "grad_norm": 0.02146013453602791,
+ "learning_rate": 0.0006,
+ "loss": 4.368387222290039,
+ "step": 1518
+ },
+ {
+ "epoch": 21.097816593886463,
+ "grad_norm": 0.02162035182118416,
+ "learning_rate": 0.0006,
+ "loss": 4.48838996887207,
+ "step": 1519
+ },
+ {
+ "epoch": 21.111790393013102,
+ "grad_norm": 0.021745676174759865,
+ "learning_rate": 0.0006,
+ "loss": 4.30778694152832,
+ "step": 1520
+ },
+ {
+ "epoch": 21.125764192139737,
+ "grad_norm": 0.02173588052392006,
+ "learning_rate": 0.0006,
+ "loss": 4.433149814605713,
+ "step": 1521
+ },
+ {
+ "epoch": 21.139737991266376,
+ "grad_norm": 0.020950648933649063,
+ "learning_rate": 0.0006,
+ "loss": 4.520089149475098,
+ "step": 1522
+ },
+ {
+ "epoch": 21.153711790393015,
+ "grad_norm": 0.020684752613306046,
+ "learning_rate": 0.0006,
+ "loss": 4.372437477111816,
+ "step": 1523
+ },
+ {
+ "epoch": 21.16768558951965,
+ "grad_norm": 0.019472869113087654,
+ "learning_rate": 0.0006,
+ "loss": 4.437015533447266,
+ "step": 1524
+ },
+ {
+ "epoch": 21.18165938864629,
+ "grad_norm": 0.01846962980926037,
+ "learning_rate": 0.0006,
+ "loss": 4.449531555175781,
+ "step": 1525
+ },
+ {
+ "epoch": 21.195633187772927,
+ "grad_norm": 0.017377803102135658,
+ "learning_rate": 0.0006,
+ "loss": 4.458279132843018,
+ "step": 1526
+ },
+ {
+ "epoch": 21.209606986899562,
+ "grad_norm": 0.016002200543880463,
+ "learning_rate": 0.0006,
+ "loss": 4.266297817230225,
+ "step": 1527
+ },
+ {
+ "epoch": 21.2235807860262,
+ "grad_norm": 0.015575903467833996,
+ "learning_rate": 0.0006,
+ "loss": 4.342632293701172,
+ "step": 1528
+ },
+ {
+ "epoch": 21.23755458515284,
+ "grad_norm": 0.015634698793292046,
+ "learning_rate": 0.0006,
+ "loss": 4.338623523712158,
+ "step": 1529
+ },
+ {
+ "epoch": 21.251528384279474,
+ "grad_norm": 0.01449581328779459,
+ "learning_rate": 0.0006,
+ "loss": 4.354330062866211,
+ "step": 1530
+ },
+ {
+ "epoch": 21.265502183406113,
+ "grad_norm": 0.01487616915255785,
+ "learning_rate": 0.0006,
+ "loss": 4.486835479736328,
+ "step": 1531
+ },
+ {
+ "epoch": 21.27947598253275,
+ "grad_norm": 0.016346555203199387,
+ "learning_rate": 0.0006,
+ "loss": 4.407098770141602,
+ "step": 1532
+ },
+ {
+ "epoch": 21.29344978165939,
+ "grad_norm": 0.01682182401418686,
+ "learning_rate": 0.0006,
+ "loss": 4.436809539794922,
+ "step": 1533
+ },
+ {
+ "epoch": 21.307423580786025,
+ "grad_norm": 0.017704004421830177,
+ "learning_rate": 0.0006,
+ "loss": 4.3854780197143555,
+ "step": 1534
+ },
+ {
+ "epoch": 21.321397379912664,
+ "grad_norm": 0.018437420949339867,
+ "learning_rate": 0.0006,
+ "loss": 4.413283824920654,
+ "step": 1535
+ },
+ {
+ "epoch": 21.335371179039303,
+ "grad_norm": 0.016911081969738007,
+ "learning_rate": 0.0006,
+ "loss": 4.357306003570557,
+ "step": 1536
+ },
+ {
+ "epoch": 21.349344978165938,
+ "grad_norm": 0.01642223633825779,
+ "learning_rate": 0.0006,
+ "loss": 4.422322750091553,
+ "step": 1537
+ },
+ {
+ "epoch": 21.363318777292577,
+ "grad_norm": 0.01780877821147442,
+ "learning_rate": 0.0006,
+ "loss": 4.387620449066162,
+ "step": 1538
+ },
+ {
+ "epoch": 21.377292576419215,
+ "grad_norm": 0.01653975248336792,
+ "learning_rate": 0.0006,
+ "loss": 4.327267646789551,
+ "step": 1539
+ },
+ {
+ "epoch": 21.39126637554585,
+ "grad_norm": 0.01572156324982643,
+ "learning_rate": 0.0006,
+ "loss": 4.475979804992676,
+ "step": 1540
+ },
+ {
+ "epoch": 21.40524017467249,
+ "grad_norm": 0.017156066372990608,
+ "learning_rate": 0.0006,
+ "loss": 4.4011125564575195,
+ "step": 1541
+ },
+ {
+ "epoch": 21.419213973799128,
+ "grad_norm": 0.01788829080760479,
+ "learning_rate": 0.0006,
+ "loss": 4.4174089431762695,
+ "step": 1542
+ },
+ {
+ "epoch": 21.433187772925763,
+ "grad_norm": 0.01728985086083412,
+ "learning_rate": 0.0006,
+ "loss": 4.429244041442871,
+ "step": 1543
+ },
+ {
+ "epoch": 21.4471615720524,
+ "grad_norm": 0.016534466296434402,
+ "learning_rate": 0.0006,
+ "loss": 4.326068878173828,
+ "step": 1544
+ },
+ {
+ "epoch": 21.46113537117904,
+ "grad_norm": 0.016582584008574486,
+ "learning_rate": 0.0006,
+ "loss": 4.2975969314575195,
+ "step": 1545
+ },
+ {
+ "epoch": 21.475109170305675,
+ "grad_norm": 0.018034178763628006,
+ "learning_rate": 0.0006,
+ "loss": 4.501319408416748,
+ "step": 1546
+ },
+ {
+ "epoch": 21.489082969432314,
+ "grad_norm": 0.017827702686190605,
+ "learning_rate": 0.0006,
+ "loss": 4.502284526824951,
+ "step": 1547
+ },
+ {
+ "epoch": 21.503056768558952,
+ "grad_norm": 0.017182299867272377,
+ "learning_rate": 0.0006,
+ "loss": 4.380828857421875,
+ "step": 1548
+ },
+ {
+ "epoch": 21.51703056768559,
+ "grad_norm": 0.019101588055491447,
+ "learning_rate": 0.0006,
+ "loss": 4.31564998626709,
+ "step": 1549
+ },
+ {
+ "epoch": 21.531004366812226,
+ "grad_norm": 0.0204677302390337,
+ "learning_rate": 0.0006,
+ "loss": 4.3840131759643555,
+ "step": 1550
+ },
+ {
+ "epoch": 21.544978165938865,
+ "grad_norm": 0.0188993439078331,
+ "learning_rate": 0.0006,
+ "loss": 4.4257731437683105,
+ "step": 1551
+ },
+ {
+ "epoch": 21.558951965065503,
+ "grad_norm": 0.01730230636894703,
+ "learning_rate": 0.0006,
+ "loss": 4.445132255554199,
+ "step": 1552
+ },
+ {
+ "epoch": 21.57292576419214,
+ "grad_norm": 0.01801574043929577,
+ "learning_rate": 0.0006,
+ "loss": 4.427064895629883,
+ "step": 1553
+ },
+ {
+ "epoch": 21.586899563318777,
+ "grad_norm": 0.018417565152049065,
+ "learning_rate": 0.0006,
+ "loss": 4.319595813751221,
+ "step": 1554
+ },
+ {
+ "epoch": 21.600873362445416,
+ "grad_norm": 0.01683160290122032,
+ "learning_rate": 0.0006,
+ "loss": 4.349218368530273,
+ "step": 1555
+ },
+ {
+ "epoch": 21.61484716157205,
+ "grad_norm": 0.016973106190562248,
+ "learning_rate": 0.0006,
+ "loss": 4.375990390777588,
+ "step": 1556
+ },
+ {
+ "epoch": 21.62882096069869,
+ "grad_norm": 0.015688583254814148,
+ "learning_rate": 0.0006,
+ "loss": 4.408170223236084,
+ "step": 1557
+ },
+ {
+ "epoch": 21.64279475982533,
+ "grad_norm": 0.01573043316602707,
+ "learning_rate": 0.0006,
+ "loss": 4.416255950927734,
+ "step": 1558
+ },
+ {
+ "epoch": 21.656768558951963,
+ "grad_norm": 0.01744293048977852,
+ "learning_rate": 0.0006,
+ "loss": 4.477141380310059,
+ "step": 1559
+ },
+ {
+ "epoch": 21.670742358078602,
+ "grad_norm": 0.017503513023257256,
+ "learning_rate": 0.0006,
+ "loss": 4.3367486000061035,
+ "step": 1560
+ },
+ {
+ "epoch": 21.68471615720524,
+ "grad_norm": 0.016592150554060936,
+ "learning_rate": 0.0006,
+ "loss": 4.314785480499268,
+ "step": 1561
+ },
+ {
+ "epoch": 21.69868995633188,
+ "grad_norm": 0.015774084255099297,
+ "learning_rate": 0.0006,
+ "loss": 4.448209285736084,
+ "step": 1562
+ },
+ {
+ "epoch": 21.712663755458514,
+ "grad_norm": 0.015779150649905205,
+ "learning_rate": 0.0006,
+ "loss": 4.392690658569336,
+ "step": 1563
+ },
+ {
+ "epoch": 21.726637554585153,
+ "grad_norm": 0.015392648056149483,
+ "learning_rate": 0.0006,
+ "loss": 4.437741756439209,
+ "step": 1564
+ },
+ {
+ "epoch": 21.74061135371179,
+ "grad_norm": 0.01530836895108223,
+ "learning_rate": 0.0006,
+ "loss": 4.424595832824707,
+ "step": 1565
+ },
+ {
+ "epoch": 21.754585152838427,
+ "grad_norm": 0.015600779093801975,
+ "learning_rate": 0.0006,
+ "loss": 4.335417747497559,
+ "step": 1566
+ },
+ {
+ "epoch": 21.768558951965066,
+ "grad_norm": 0.014646518044173717,
+ "learning_rate": 0.0006,
+ "loss": 4.279081344604492,
+ "step": 1567
+ },
+ {
+ "epoch": 21.782532751091704,
+ "grad_norm": 0.014132864773273468,
+ "learning_rate": 0.0006,
+ "loss": 4.345950126647949,
+ "step": 1568
+ },
+ {
+ "epoch": 21.79650655021834,
+ "grad_norm": 0.01425724383443594,
+ "learning_rate": 0.0006,
+ "loss": 4.277588844299316,
+ "step": 1569
+ },
+ {
+ "epoch": 21.810480349344978,
+ "grad_norm": 0.015396572649478912,
+ "learning_rate": 0.0006,
+ "loss": 4.326196670532227,
+ "step": 1570
+ },
+ {
+ "epoch": 21.824454148471617,
+ "grad_norm": 0.01683669537305832,
+ "learning_rate": 0.0006,
+ "loss": 4.3294172286987305,
+ "step": 1571
+ },
+ {
+ "epoch": 21.83842794759825,
+ "grad_norm": 0.016365647315979004,
+ "learning_rate": 0.0006,
+ "loss": 4.523321628570557,
+ "step": 1572
+ },
+ {
+ "epoch": 21.85240174672489,
+ "grad_norm": 0.017142774537205696,
+ "learning_rate": 0.0006,
+ "loss": 4.475842475891113,
+ "step": 1573
+ },
+ {
+ "epoch": 21.86637554585153,
+ "grad_norm": 0.015950839966535568,
+ "learning_rate": 0.0006,
+ "loss": 4.397187232971191,
+ "step": 1574
+ },
+ {
+ "epoch": 21.880349344978168,
+ "grad_norm": 0.016180362552404404,
+ "learning_rate": 0.0006,
+ "loss": 4.414778709411621,
+ "step": 1575
+ },
+ {
+ "epoch": 21.894323144104803,
+ "grad_norm": 0.017615556716918945,
+ "learning_rate": 0.0006,
+ "loss": 4.348567962646484,
+ "step": 1576
+ },
+ {
+ "epoch": 21.90829694323144,
+ "grad_norm": 0.01896277628839016,
+ "learning_rate": 0.0006,
+ "loss": 4.431762218475342,
+ "step": 1577
+ },
+ {
+ "epoch": 21.92227074235808,
+ "grad_norm": 0.01990971714258194,
+ "learning_rate": 0.0006,
+ "loss": 4.431285858154297,
+ "step": 1578
+ },
+ {
+ "epoch": 21.936244541484715,
+ "grad_norm": 0.0201258547604084,
+ "learning_rate": 0.0006,
+ "loss": 4.309872627258301,
+ "step": 1579
+ },
+ {
+ "epoch": 21.950218340611354,
+ "grad_norm": 0.019620198756456375,
+ "learning_rate": 0.0006,
+ "loss": 4.484044551849365,
+ "step": 1580
+ },
+ {
+ "epoch": 21.964192139737992,
+ "grad_norm": 0.01771123707294464,
+ "learning_rate": 0.0006,
+ "loss": 4.368093013763428,
+ "step": 1581
+ },
+ {
+ "epoch": 21.978165938864628,
+ "grad_norm": 0.017501654103398323,
+ "learning_rate": 0.0006,
+ "loss": 4.372651100158691,
+ "step": 1582
+ },
+ {
+ "epoch": 21.992139737991266,
+ "grad_norm": 0.01693608984351158,
+ "learning_rate": 0.0006,
+ "loss": 4.379167556762695,
+ "step": 1583
+ },
+ {
+ "epoch": 22.0,
+ "grad_norm": 0.02002849616110325,
+ "learning_rate": 0.0006,
+ "loss": 4.329627990722656,
+ "step": 1584
+ },
+ {
+ "epoch": 22.0,
+ "eval_loss": 4.723696231842041,
+ "eval_runtime": 56.1758,
+ "eval_samples_per_second": 43.471,
+ "eval_steps_per_second": 1.371,
+ "step": 1584
+ },
+ {
+ "epoch": 22.01397379912664,
+ "grad_norm": 0.02030305378139019,
+ "learning_rate": 0.0006,
+ "loss": 4.355566024780273,
+ "step": 1585
+ },
+ {
+ "epoch": 22.027947598253274,
+ "grad_norm": 0.018638934940099716,
+ "learning_rate": 0.0006,
+ "loss": 4.314009189605713,
+ "step": 1586
+ },
+ {
+ "epoch": 22.041921397379912,
+ "grad_norm": 0.019917072728276253,
+ "learning_rate": 0.0006,
+ "loss": 4.367525577545166,
+ "step": 1587
+ },
+ {
+ "epoch": 22.05589519650655,
+ "grad_norm": 0.020785167813301086,
+ "learning_rate": 0.0006,
+ "loss": 4.375561714172363,
+ "step": 1588
+ },
+ {
+ "epoch": 22.069868995633186,
+ "grad_norm": 0.021214453503489494,
+ "learning_rate": 0.0006,
+ "loss": 4.418305397033691,
+ "step": 1589
+ },
+ {
+ "epoch": 22.083842794759825,
+ "grad_norm": 0.019051678478717804,
+ "learning_rate": 0.0006,
+ "loss": 4.215126991271973,
+ "step": 1590
+ },
+ {
+ "epoch": 22.097816593886463,
+ "grad_norm": 0.02323366142809391,
+ "learning_rate": 0.0006,
+ "loss": 4.306807994842529,
+ "step": 1591
+ },
+ {
+ "epoch": 22.111790393013102,
+ "grad_norm": 0.026186738163232803,
+ "learning_rate": 0.0006,
+ "loss": 4.408508777618408,
+ "step": 1592
+ },
+ {
+ "epoch": 22.125764192139737,
+ "grad_norm": 0.027361227199435234,
+ "learning_rate": 0.0006,
+ "loss": 4.329322814941406,
+ "step": 1593
+ },
+ {
+ "epoch": 22.139737991266376,
+ "grad_norm": 0.026564646512269974,
+ "learning_rate": 0.0006,
+ "loss": 4.346642017364502,
+ "step": 1594
+ },
+ {
+ "epoch": 22.153711790393015,
+ "grad_norm": 0.025073667988181114,
+ "learning_rate": 0.0006,
+ "loss": 4.3984198570251465,
+ "step": 1595
+ },
+ {
+ "epoch": 22.16768558951965,
+ "grad_norm": 0.02347794733941555,
+ "learning_rate": 0.0006,
+ "loss": 4.352538108825684,
+ "step": 1596
+ },
+ {
+ "epoch": 22.18165938864629,
+ "grad_norm": 0.02203468047082424,
+ "learning_rate": 0.0006,
+ "loss": 4.298148155212402,
+ "step": 1597
+ },
+ {
+ "epoch": 22.195633187772927,
+ "grad_norm": 0.020621957257390022,
+ "learning_rate": 0.0006,
+ "loss": 4.480309963226318,
+ "step": 1598
+ },
+ {
+ "epoch": 22.209606986899562,
+ "grad_norm": 0.023710055276751518,
+ "learning_rate": 0.0006,
+ "loss": 4.339317321777344,
+ "step": 1599
+ },
+ {
+ "epoch": 22.2235807860262,
+ "grad_norm": 0.0233482476323843,
+ "learning_rate": 0.0006,
+ "loss": 4.436896324157715,
+ "step": 1600
+ },
+ {
+ "epoch": 22.23755458515284,
+ "grad_norm": 0.020122403278946877,
+ "learning_rate": 0.0006,
+ "loss": 4.236002445220947,
+ "step": 1601
+ },
+ {
+ "epoch": 22.251528384279474,
+ "grad_norm": 0.01957266591489315,
+ "learning_rate": 0.0006,
+ "loss": 4.497328281402588,
+ "step": 1602
+ },
+ {
+ "epoch": 22.265502183406113,
+ "grad_norm": 0.018964895978569984,
+ "learning_rate": 0.0006,
+ "loss": 4.36360502243042,
+ "step": 1603
+ },
+ {
+ "epoch": 22.27947598253275,
+ "grad_norm": 0.018902769312262535,
+ "learning_rate": 0.0006,
+ "loss": 4.427093505859375,
+ "step": 1604
+ },
+ {
+ "epoch": 22.29344978165939,
+ "grad_norm": 0.018918119370937347,
+ "learning_rate": 0.0006,
+ "loss": 4.294950485229492,
+ "step": 1605
+ },
+ {
+ "epoch": 22.307423580786025,
+ "grad_norm": 0.019315775483846664,
+ "learning_rate": 0.0006,
+ "loss": 4.427952289581299,
+ "step": 1606
+ },
+ {
+ "epoch": 22.321397379912664,
+ "grad_norm": 0.018350522965192795,
+ "learning_rate": 0.0006,
+ "loss": 4.272945404052734,
+ "step": 1607
+ },
+ {
+ "epoch": 22.335371179039303,
+ "grad_norm": 0.01759534701704979,
+ "learning_rate": 0.0006,
+ "loss": 4.348023891448975,
+ "step": 1608
+ },
+ {
+ "epoch": 22.349344978165938,
+ "grad_norm": 0.01646057888865471,
+ "learning_rate": 0.0006,
+ "loss": 4.278518199920654,
+ "step": 1609
+ },
+ {
+ "epoch": 22.363318777292577,
+ "grad_norm": 0.015875060111284256,
+ "learning_rate": 0.0006,
+ "loss": 4.423527240753174,
+ "step": 1610
+ },
+ {
+ "epoch": 22.377292576419215,
+ "grad_norm": 0.01638529635965824,
+ "learning_rate": 0.0006,
+ "loss": 4.321723937988281,
+ "step": 1611
+ },
+ {
+ "epoch": 22.39126637554585,
+ "grad_norm": 0.016942061483860016,
+ "learning_rate": 0.0006,
+ "loss": 4.346164703369141,
+ "step": 1612
+ },
+ {
+ "epoch": 22.40524017467249,
+ "grad_norm": 0.015209621749818325,
+ "learning_rate": 0.0006,
+ "loss": 4.434340000152588,
+ "step": 1613
+ },
+ {
+ "epoch": 22.419213973799128,
+ "grad_norm": 0.014423009008169174,
+ "learning_rate": 0.0006,
+ "loss": 4.342560768127441,
+ "step": 1614
+ },
+ {
+ "epoch": 22.433187772925763,
+ "grad_norm": 0.014369525946676731,
+ "learning_rate": 0.0006,
+ "loss": 4.296586990356445,
+ "step": 1615
+ },
+ {
+ "epoch": 22.4471615720524,
+ "grad_norm": 0.01392540242522955,
+ "learning_rate": 0.0006,
+ "loss": 4.433176040649414,
+ "step": 1616
+ },
+ {
+ "epoch": 22.46113537117904,
+ "grad_norm": 0.013206899166107178,
+ "learning_rate": 0.0006,
+ "loss": 4.32962703704834,
+ "step": 1617
+ },
+ {
+ "epoch": 22.475109170305675,
+ "grad_norm": 0.013427077792584896,
+ "learning_rate": 0.0006,
+ "loss": 4.420437335968018,
+ "step": 1618
+ },
+ {
+ "epoch": 22.489082969432314,
+ "grad_norm": 0.014194196090102196,
+ "learning_rate": 0.0006,
+ "loss": 4.410478591918945,
+ "step": 1619
+ },
+ {
+ "epoch": 22.503056768558952,
+ "grad_norm": 0.01362364087253809,
+ "learning_rate": 0.0006,
+ "loss": 4.338932037353516,
+ "step": 1620
+ },
+ {
+ "epoch": 22.51703056768559,
+ "grad_norm": 0.014299589209258556,
+ "learning_rate": 0.0006,
+ "loss": 4.34089994430542,
+ "step": 1621
+ },
+ {
+ "epoch": 22.531004366812226,
+ "grad_norm": 0.014061295427381992,
+ "learning_rate": 0.0006,
+ "loss": 4.29201602935791,
+ "step": 1622
+ },
+ {
+ "epoch": 22.544978165938865,
+ "grad_norm": 0.013945229351520538,
+ "learning_rate": 0.0006,
+ "loss": 4.326052188873291,
+ "step": 1623
+ },
+ {
+ "epoch": 22.558951965065503,
+ "grad_norm": 0.01387068536132574,
+ "learning_rate": 0.0006,
+ "loss": 4.393521308898926,
+ "step": 1624
+ },
+ {
+ "epoch": 22.57292576419214,
+ "grad_norm": 0.013931898400187492,
+ "learning_rate": 0.0006,
+ "loss": 4.3739752769470215,
+ "step": 1625
+ },
+ {
+ "epoch": 22.586899563318777,
+ "grad_norm": 0.01497740764170885,
+ "learning_rate": 0.0006,
+ "loss": 4.3991618156433105,
+ "step": 1626
+ },
+ {
+ "epoch": 22.600873362445416,
+ "grad_norm": 0.015842905268073082,
+ "learning_rate": 0.0006,
+ "loss": 4.405340671539307,
+ "step": 1627
+ },
+ {
+ "epoch": 22.61484716157205,
+ "grad_norm": 0.01580633595585823,
+ "learning_rate": 0.0006,
+ "loss": 4.280102729797363,
+ "step": 1628
+ },
+ {
+ "epoch": 22.62882096069869,
+ "grad_norm": 0.016567382961511612,
+ "learning_rate": 0.0006,
+ "loss": 4.407657146453857,
+ "step": 1629
+ },
+ {
+ "epoch": 22.64279475982533,
+ "grad_norm": 0.01648622378706932,
+ "learning_rate": 0.0006,
+ "loss": 4.392590522766113,
+ "step": 1630
+ },
+ {
+ "epoch": 22.656768558951963,
+ "grad_norm": 0.017066307365894318,
+ "learning_rate": 0.0006,
+ "loss": 4.324862003326416,
+ "step": 1631
+ },
+ {
+ "epoch": 22.670742358078602,
+ "grad_norm": 0.01637062057852745,
+ "learning_rate": 0.0006,
+ "loss": 4.34598445892334,
+ "step": 1632
+ },
+ {
+ "epoch": 22.68471615720524,
+ "grad_norm": 0.018038960173726082,
+ "learning_rate": 0.0006,
+ "loss": 4.320711612701416,
+ "step": 1633
+ },
+ {
+ "epoch": 22.69868995633188,
+ "grad_norm": 0.018096428364515305,
+ "learning_rate": 0.0006,
+ "loss": 4.536527156829834,
+ "step": 1634
+ },
+ {
+ "epoch": 22.712663755458514,
+ "grad_norm": 0.017889291048049927,
+ "learning_rate": 0.0006,
+ "loss": 4.388131141662598,
+ "step": 1635
+ },
+ {
+ "epoch": 22.726637554585153,
+ "grad_norm": 0.018771521747112274,
+ "learning_rate": 0.0006,
+ "loss": 4.357668876647949,
+ "step": 1636
+ },
+ {
+ "epoch": 22.74061135371179,
+ "grad_norm": 0.02100873365998268,
+ "learning_rate": 0.0006,
+ "loss": 4.293972015380859,
+ "step": 1637
+ },
+ {
+ "epoch": 22.754585152838427,
+ "grad_norm": 0.017715785652399063,
+ "learning_rate": 0.0006,
+ "loss": 4.349849224090576,
+ "step": 1638
+ },
+ {
+ "epoch": 22.768558951965066,
+ "grad_norm": 0.015546333976089954,
+ "learning_rate": 0.0006,
+ "loss": 4.311728000640869,
+ "step": 1639
+ },
+ {
+ "epoch": 22.782532751091704,
+ "grad_norm": 0.01605990156531334,
+ "learning_rate": 0.0006,
+ "loss": 4.499123573303223,
+ "step": 1640
+ },
+ {
+ "epoch": 22.79650655021834,
+ "grad_norm": 0.016642142087221146,
+ "learning_rate": 0.0006,
+ "loss": 4.4319257736206055,
+ "step": 1641
+ },
+ {
+ "epoch": 22.810480349344978,
+ "grad_norm": 0.016120387241244316,
+ "learning_rate": 0.0006,
+ "loss": 4.4373650550842285,
+ "step": 1642
+ },
+ {
+ "epoch": 22.824454148471617,
+ "grad_norm": 0.015174890868365765,
+ "learning_rate": 0.0006,
+ "loss": 4.314082145690918,
+ "step": 1643
+ },
+ {
+ "epoch": 22.83842794759825,
+ "grad_norm": 0.01551714539527893,
+ "learning_rate": 0.0006,
+ "loss": 4.360363960266113,
+ "step": 1644
+ },
+ {
+ "epoch": 22.85240174672489,
+ "grad_norm": 0.014966742135584354,
+ "learning_rate": 0.0006,
+ "loss": 4.346253395080566,
+ "step": 1645
+ },
+ {
+ "epoch": 22.86637554585153,
+ "grad_norm": 0.014288028702139854,
+ "learning_rate": 0.0006,
+ "loss": 4.464366436004639,
+ "step": 1646
+ },
+ {
+ "epoch": 22.880349344978168,
+ "grad_norm": 0.015291682444512844,
+ "learning_rate": 0.0006,
+ "loss": 4.504101753234863,
+ "step": 1647
+ },
+ {
+ "epoch": 22.894323144104803,
+ "grad_norm": 0.014534426853060722,
+ "learning_rate": 0.0006,
+ "loss": 4.462172031402588,
+ "step": 1648
+ },
+ {
+ "epoch": 22.90829694323144,
+ "grad_norm": 0.014129819348454475,
+ "learning_rate": 0.0006,
+ "loss": 4.350126266479492,
+ "step": 1649
+ },
+ {
+ "epoch": 22.92227074235808,
+ "grad_norm": 0.014932668767869473,
+ "learning_rate": 0.0006,
+ "loss": 4.374392032623291,
+ "step": 1650
+ },
+ {
+ "epoch": 22.936244541484715,
+ "grad_norm": 0.014647853560745716,
+ "learning_rate": 0.0006,
+ "loss": 4.252264499664307,
+ "step": 1651
+ },
+ {
+ "epoch": 22.950218340611354,
+ "grad_norm": 0.013553360477089882,
+ "learning_rate": 0.0006,
+ "loss": 4.379729270935059,
+ "step": 1652
+ },
+ {
+ "epoch": 22.964192139737992,
+ "grad_norm": 0.014606194570660591,
+ "learning_rate": 0.0006,
+ "loss": 4.363372802734375,
+ "step": 1653
+ },
+ {
+ "epoch": 22.978165938864628,
+ "grad_norm": 0.015000631101429462,
+ "learning_rate": 0.0006,
+ "loss": 4.408048629760742,
+ "step": 1654
+ },
+ {
+ "epoch": 22.992139737991266,
+ "grad_norm": 0.015370816923677921,
+ "learning_rate": 0.0006,
+ "loss": 4.297451019287109,
+ "step": 1655
+ },
+ {
+ "epoch": 23.0,
+ "grad_norm": 0.01709144562482834,
+ "learning_rate": 0.0006,
+ "loss": 4.294776916503906,
+ "step": 1656
+ },
+ {
+ "epoch": 23.0,
+ "eval_loss": 4.678618907928467,
+ "eval_runtime": 56.1668,
+ "eval_samples_per_second": 43.478,
+ "eval_steps_per_second": 1.371,
+ "step": 1656
+ },
+ {
+ "epoch": 23.01397379912664,
+ "grad_norm": 0.01727290451526642,
+ "learning_rate": 0.0006,
+ "loss": 4.311701774597168,
+ "step": 1657
+ },
+ {
+ "epoch": 23.027947598253274,
+ "grad_norm": 0.020645759999752045,
+ "learning_rate": 0.0006,
+ "loss": 4.429897308349609,
+ "step": 1658
+ },
+ {
+ "epoch": 23.041921397379912,
+ "grad_norm": 0.02204226516187191,
+ "learning_rate": 0.0006,
+ "loss": 4.365151882171631,
+ "step": 1659
+ },
+ {
+ "epoch": 23.05589519650655,
+ "grad_norm": 0.023069269955158234,
+ "learning_rate": 0.0006,
+ "loss": 4.3710832595825195,
+ "step": 1660
+ },
+ {
+ "epoch": 23.069868995633186,
+ "grad_norm": 0.024151179939508438,
+ "learning_rate": 0.0006,
+ "loss": 4.403108596801758,
+ "step": 1661
+ },
+ {
+ "epoch": 23.083842794759825,
+ "grad_norm": 0.02511233650147915,
+ "learning_rate": 0.0006,
+ "loss": 4.414128303527832,
+ "step": 1662
+ },
+ {
+ "epoch": 23.097816593886463,
+ "grad_norm": 0.025802113115787506,
+ "learning_rate": 0.0006,
+ "loss": 4.388265609741211,
+ "step": 1663
+ },
+ {
+ "epoch": 23.111790393013102,
+ "grad_norm": 0.02377348765730858,
+ "learning_rate": 0.0006,
+ "loss": 4.357505798339844,
+ "step": 1664
+ },
+ {
+ "epoch": 23.125764192139737,
+ "grad_norm": 0.020876651629805565,
+ "learning_rate": 0.0006,
+ "loss": 4.23043966293335,
+ "step": 1665
+ },
+ {
+ "epoch": 23.139737991266376,
+ "grad_norm": 0.01959727331995964,
+ "learning_rate": 0.0006,
+ "loss": 4.280970096588135,
+ "step": 1666
+ },
+ {
+ "epoch": 23.153711790393015,
+ "grad_norm": 0.018678616732358932,
+ "learning_rate": 0.0006,
+ "loss": 4.346716403961182,
+ "step": 1667
+ },
+ {
+ "epoch": 23.16768558951965,
+ "grad_norm": 0.02108944021165371,
+ "learning_rate": 0.0006,
+ "loss": 4.277904510498047,
+ "step": 1668
+ },
+ {
+ "epoch": 23.18165938864629,
+ "grad_norm": 0.020221566781401634,
+ "learning_rate": 0.0006,
+ "loss": 4.367693901062012,
+ "step": 1669
+ },
+ {
+ "epoch": 23.195633187772927,
+ "grad_norm": 0.019599711522459984,
+ "learning_rate": 0.0006,
+ "loss": 4.257092475891113,
+ "step": 1670
+ },
+ {
+ "epoch": 23.209606986899562,
+ "grad_norm": 0.019833361729979515,
+ "learning_rate": 0.0006,
+ "loss": 4.540737628936768,
+ "step": 1671
+ },
+ {
+ "epoch": 23.2235807860262,
+ "grad_norm": 0.020352095365524292,
+ "learning_rate": 0.0006,
+ "loss": 4.360923767089844,
+ "step": 1672
+ },
+ {
+ "epoch": 23.23755458515284,
+ "grad_norm": 0.01939748041331768,
+ "learning_rate": 0.0006,
+ "loss": 4.393110275268555,
+ "step": 1673
+ },
+ {
+ "epoch": 23.251528384279474,
+ "grad_norm": 0.01884354278445244,
+ "learning_rate": 0.0006,
+ "loss": 4.383938789367676,
+ "step": 1674
+ },
+ {
+ "epoch": 23.265502183406113,
+ "grad_norm": 0.02008470520377159,
+ "learning_rate": 0.0006,
+ "loss": 4.346344947814941,
+ "step": 1675
+ },
+ {
+ "epoch": 23.27947598253275,
+ "grad_norm": 0.019065195694565773,
+ "learning_rate": 0.0006,
+ "loss": 4.36611795425415,
+ "step": 1676
+ },
+ {
+ "epoch": 23.29344978165939,
+ "grad_norm": 0.01938541606068611,
+ "learning_rate": 0.0006,
+ "loss": 4.4009599685668945,
+ "step": 1677
+ },
+ {
+ "epoch": 23.307423580786025,
+ "grad_norm": 0.019044723361730576,
+ "learning_rate": 0.0006,
+ "loss": 4.340169906616211,
+ "step": 1678
+ },
+ {
+ "epoch": 23.321397379912664,
+ "grad_norm": 0.018850378692150116,
+ "learning_rate": 0.0006,
+ "loss": 4.37674617767334,
+ "step": 1679
+ },
+ {
+ "epoch": 23.335371179039303,
+ "grad_norm": 0.019076095893979073,
+ "learning_rate": 0.0006,
+ "loss": 4.226874351501465,
+ "step": 1680
+ },
+ {
+ "epoch": 23.349344978165938,
+ "grad_norm": 0.017543038353323936,
+ "learning_rate": 0.0006,
+ "loss": 4.337751388549805,
+ "step": 1681
+ },
+ {
+ "epoch": 23.363318777292577,
+ "grad_norm": 0.017170218750834465,
+ "learning_rate": 0.0006,
+ "loss": 4.282512664794922,
+ "step": 1682
+ },
+ {
+ "epoch": 23.377292576419215,
+ "grad_norm": 0.01574796624481678,
+ "learning_rate": 0.0006,
+ "loss": 4.30638313293457,
+ "step": 1683
+ },
+ {
+ "epoch": 23.39126637554585,
+ "grad_norm": 0.015160846523940563,
+ "learning_rate": 0.0006,
+ "loss": 4.336716175079346,
+ "step": 1684
+ },
+ {
+ "epoch": 23.40524017467249,
+ "grad_norm": 0.01534526702016592,
+ "learning_rate": 0.0006,
+ "loss": 4.2174787521362305,
+ "step": 1685
+ },
+ {
+ "epoch": 23.419213973799128,
+ "grad_norm": 0.015289999544620514,
+ "learning_rate": 0.0006,
+ "loss": 4.272546291351318,
+ "step": 1686
+ },
+ {
+ "epoch": 23.433187772925763,
+ "grad_norm": 0.014894185587763786,
+ "learning_rate": 0.0006,
+ "loss": 4.296045303344727,
+ "step": 1687
+ },
+ {
+ "epoch": 23.4471615720524,
+ "grad_norm": 0.014869502745568752,
+ "learning_rate": 0.0006,
+ "loss": 4.425461769104004,
+ "step": 1688
+ },
+ {
+ "epoch": 23.46113537117904,
+ "grad_norm": 0.01474483497440815,
+ "learning_rate": 0.0006,
+ "loss": 4.361867427825928,
+ "step": 1689
+ },
+ {
+ "epoch": 23.475109170305675,
+ "grad_norm": 0.015299245715141296,
+ "learning_rate": 0.0006,
+ "loss": 4.364523410797119,
+ "step": 1690
+ },
+ {
+ "epoch": 23.489082969432314,
+ "grad_norm": 0.015201340429484844,
+ "learning_rate": 0.0006,
+ "loss": 4.285182952880859,
+ "step": 1691
+ },
+ {
+ "epoch": 23.503056768558952,
+ "grad_norm": 0.016057293862104416,
+ "learning_rate": 0.0006,
+ "loss": 4.284873962402344,
+ "step": 1692
+ },
+ {
+ "epoch": 23.51703056768559,
+ "grad_norm": 0.015871062874794006,
+ "learning_rate": 0.0006,
+ "loss": 4.326470851898193,
+ "step": 1693
+ },
+ {
+ "epoch": 23.531004366812226,
+ "grad_norm": 0.01586288772523403,
+ "learning_rate": 0.0006,
+ "loss": 4.324883460998535,
+ "step": 1694
+ },
+ {
+ "epoch": 23.544978165938865,
+ "grad_norm": 0.016092827543616295,
+ "learning_rate": 0.0006,
+ "loss": 4.3258867263793945,
+ "step": 1695
+ },
+ {
+ "epoch": 23.558951965065503,
+ "grad_norm": 0.01582016795873642,
+ "learning_rate": 0.0006,
+ "loss": 4.358892440795898,
+ "step": 1696
+ },
+ {
+ "epoch": 23.57292576419214,
+ "grad_norm": 0.01688452623784542,
+ "learning_rate": 0.0006,
+ "loss": 4.280842304229736,
+ "step": 1697
+ },
+ {
+ "epoch": 23.586899563318777,
+ "grad_norm": 0.015865886583924294,
+ "learning_rate": 0.0006,
+ "loss": 4.245567321777344,
+ "step": 1698
+ },
+ {
+ "epoch": 23.600873362445416,
+ "grad_norm": 0.015303662046790123,
+ "learning_rate": 0.0006,
+ "loss": 4.383112907409668,
+ "step": 1699
+ },
+ {
+ "epoch": 23.61484716157205,
+ "grad_norm": 0.015649516135454178,
+ "learning_rate": 0.0006,
+ "loss": 4.519082069396973,
+ "step": 1700
+ },
+ {
+ "epoch": 23.62882096069869,
+ "grad_norm": 0.0167120099067688,
+ "learning_rate": 0.0006,
+ "loss": 4.447847366333008,
+ "step": 1701
+ },
+ {
+ "epoch": 23.64279475982533,
+ "grad_norm": 0.016368716955184937,
+ "learning_rate": 0.0006,
+ "loss": 4.434186935424805,
+ "step": 1702
+ },
+ {
+ "epoch": 23.656768558951963,
+ "grad_norm": 0.017379503697156906,
+ "learning_rate": 0.0006,
+ "loss": 4.317594528198242,
+ "step": 1703
+ },
+ {
+ "epoch": 23.670742358078602,
+ "grad_norm": 0.018573811277747154,
+ "learning_rate": 0.0006,
+ "loss": 4.205793857574463,
+ "step": 1704
+ },
+ {
+ "epoch": 23.68471615720524,
+ "grad_norm": 0.01749090477824211,
+ "learning_rate": 0.0006,
+ "loss": 4.3985700607299805,
+ "step": 1705
+ },
+ {
+ "epoch": 23.69868995633188,
+ "grad_norm": 0.0165668074041605,
+ "learning_rate": 0.0006,
+ "loss": 4.329226970672607,
+ "step": 1706
+ },
+ {
+ "epoch": 23.712663755458514,
+ "grad_norm": 0.016913846135139465,
+ "learning_rate": 0.0006,
+ "loss": 4.388485908508301,
+ "step": 1707
+ },
+ {
+ "epoch": 23.726637554585153,
+ "grad_norm": 0.016813859343528748,
+ "learning_rate": 0.0006,
+ "loss": 4.3395915031433105,
+ "step": 1708
+ },
+ {
+ "epoch": 23.74061135371179,
+ "grad_norm": 0.01676975190639496,
+ "learning_rate": 0.0006,
+ "loss": 4.254745960235596,
+ "step": 1709
+ },
+ {
+ "epoch": 23.754585152838427,
+ "grad_norm": 0.016370350494980812,
+ "learning_rate": 0.0006,
+ "loss": 4.2612223625183105,
+ "step": 1710
+ },
+ {
+ "epoch": 23.768558951965066,
+ "grad_norm": 0.01696198247373104,
+ "learning_rate": 0.0006,
+ "loss": 4.327524662017822,
+ "step": 1711
+ },
+ {
+ "epoch": 23.782532751091704,
+ "grad_norm": 0.0189108457416296,
+ "learning_rate": 0.0006,
+ "loss": 4.325839519500732,
+ "step": 1712
+ },
+ {
+ "epoch": 23.79650655021834,
+ "grad_norm": 0.01862351782619953,
+ "learning_rate": 0.0006,
+ "loss": 4.333662509918213,
+ "step": 1713
+ },
+ {
+ "epoch": 23.810480349344978,
+ "grad_norm": 0.01655014418065548,
+ "learning_rate": 0.0006,
+ "loss": 4.39305305480957,
+ "step": 1714
+ },
+ {
+ "epoch": 23.824454148471617,
+ "grad_norm": 0.015899376943707466,
+ "learning_rate": 0.0006,
+ "loss": 4.390053749084473,
+ "step": 1715
+ },
+ {
+ "epoch": 23.83842794759825,
+ "grad_norm": 0.016519056633114815,
+ "learning_rate": 0.0006,
+ "loss": 4.255903244018555,
+ "step": 1716
+ },
+ {
+ "epoch": 23.85240174672489,
+ "grad_norm": 0.0167539119720459,
+ "learning_rate": 0.0006,
+ "loss": 4.255727767944336,
+ "step": 1717
+ },
+ {
+ "epoch": 23.86637554585153,
+ "grad_norm": 0.0171353816986084,
+ "learning_rate": 0.0006,
+ "loss": 4.376740455627441,
+ "step": 1718
+ },
+ {
+ "epoch": 23.880349344978168,
+ "grad_norm": 0.016719117760658264,
+ "learning_rate": 0.0006,
+ "loss": 4.3376688957214355,
+ "step": 1719
+ },
+ {
+ "epoch": 23.894323144104803,
+ "grad_norm": 0.01561494916677475,
+ "learning_rate": 0.0006,
+ "loss": 4.3345746994018555,
+ "step": 1720
+ },
+ {
+ "epoch": 23.90829694323144,
+ "grad_norm": 0.016303174197673798,
+ "learning_rate": 0.0006,
+ "loss": 4.332371234893799,
+ "step": 1721
+ },
+ {
+ "epoch": 23.92227074235808,
+ "grad_norm": 0.016722865402698517,
+ "learning_rate": 0.0006,
+ "loss": 4.267125129699707,
+ "step": 1722
+ },
+ {
+ "epoch": 23.936244541484715,
+ "grad_norm": 0.017072124406695366,
+ "learning_rate": 0.0006,
+ "loss": 4.322751998901367,
+ "step": 1723
+ },
+ {
+ "epoch": 23.950218340611354,
+ "grad_norm": 0.01699932850897312,
+ "learning_rate": 0.0006,
+ "loss": 4.378422737121582,
+ "step": 1724
+ },
+ {
+ "epoch": 23.964192139737992,
+ "grad_norm": 0.01742720976471901,
+ "learning_rate": 0.0006,
+ "loss": 4.345083236694336,
+ "step": 1725
+ },
+ {
+ "epoch": 23.978165938864628,
+ "grad_norm": 0.018797501921653748,
+ "learning_rate": 0.0006,
+ "loss": 4.290443420410156,
+ "step": 1726
+ },
+ {
+ "epoch": 23.992139737991266,
+ "grad_norm": 0.018090683966875076,
+ "learning_rate": 0.0006,
+ "loss": 4.377580642700195,
+ "step": 1727
+ },
+ {
+ "epoch": 24.0,
+ "grad_norm": 0.017305459827184677,
+ "learning_rate": 0.0006,
+ "loss": 4.364426612854004,
+ "step": 1728
+ },
+ {
+ "epoch": 24.0,
+ "eval_loss": 4.647762298583984,
+ "eval_runtime": 56.5402,
+ "eval_samples_per_second": 43.191,
+ "eval_steps_per_second": 1.362,
+ "step": 1728
+ },
+ {
+ "epoch": 24.01397379912664,
+ "grad_norm": 0.016493607312440872,
+ "learning_rate": 0.0006,
+ "loss": 4.252255916595459,
+ "step": 1729
+ },
+ {
+ "epoch": 24.027947598253274,
+ "grad_norm": 0.016529450193047523,
+ "learning_rate": 0.0006,
+ "loss": 4.352746963500977,
+ "step": 1730
+ },
+ {
+ "epoch": 24.041921397379912,
+ "grad_norm": 0.016940191388130188,
+ "learning_rate": 0.0006,
+ "loss": 4.301870822906494,
+ "step": 1731
+ },
+ {
+ "epoch": 24.05589519650655,
+ "grad_norm": 0.01752558909356594,
+ "learning_rate": 0.0006,
+ "loss": 4.2762451171875,
+ "step": 1732
+ },
+ {
+ "epoch": 24.069868995633186,
+ "grad_norm": 0.017112884670495987,
+ "learning_rate": 0.0006,
+ "loss": 4.2659149169921875,
+ "step": 1733
+ },
+ {
+ "epoch": 24.083842794759825,
+ "grad_norm": 0.01900586113333702,
+ "learning_rate": 0.0006,
+ "loss": 4.306058883666992,
+ "step": 1734
+ },
+ {
+ "epoch": 24.097816593886463,
+ "grad_norm": 0.019447680562734604,
+ "learning_rate": 0.0006,
+ "loss": 4.450128555297852,
+ "step": 1735
+ },
+ {
+ "epoch": 24.111790393013102,
+ "grad_norm": 0.01948186755180359,
+ "learning_rate": 0.0006,
+ "loss": 4.202869415283203,
+ "step": 1736
+ },
+ {
+ "epoch": 24.125764192139737,
+ "grad_norm": 0.01716383546590805,
+ "learning_rate": 0.0006,
+ "loss": 4.318876266479492,
+ "step": 1737
+ },
+ {
+ "epoch": 24.139737991266376,
+ "grad_norm": 0.017049988731741905,
+ "learning_rate": 0.0006,
+ "loss": 4.237238883972168,
+ "step": 1738
+ },
+ {
+ "epoch": 24.153711790393015,
+ "grad_norm": 0.015169923193752766,
+ "learning_rate": 0.0006,
+ "loss": 4.289426803588867,
+ "step": 1739
+ },
+ {
+ "epoch": 24.16768558951965,
+ "grad_norm": 0.014760667458176613,
+ "learning_rate": 0.0006,
+ "loss": 4.356924057006836,
+ "step": 1740
+ },
+ {
+ "epoch": 24.18165938864629,
+ "grad_norm": 0.015127750113606453,
+ "learning_rate": 0.0006,
+ "loss": 4.142916679382324,
+ "step": 1741
+ },
+ {
+ "epoch": 24.195633187772927,
+ "grad_norm": 0.01549555640667677,
+ "learning_rate": 0.0006,
+ "loss": 4.345563888549805,
+ "step": 1742
+ },
+ {
+ "epoch": 24.209606986899562,
+ "grad_norm": 0.016393091529607773,
+ "learning_rate": 0.0006,
+ "loss": 4.417495250701904,
+ "step": 1743
+ },
+ {
+ "epoch": 24.2235807860262,
+ "grad_norm": 0.01785266026854515,
+ "learning_rate": 0.0006,
+ "loss": 4.3553547859191895,
+ "step": 1744
+ },
+ {
+ "epoch": 24.23755458515284,
+ "grad_norm": 0.017574617639183998,
+ "learning_rate": 0.0006,
+ "loss": 4.241024494171143,
+ "step": 1745
+ },
+ {
+ "epoch": 24.251528384279474,
+ "grad_norm": 0.019060306251049042,
+ "learning_rate": 0.0006,
+ "loss": 4.3685150146484375,
+ "step": 1746
+ },
+ {
+ "epoch": 24.265502183406113,
+ "grad_norm": 0.0214633010327816,
+ "learning_rate": 0.0006,
+ "loss": 4.323866844177246,
+ "step": 1747
+ },
+ {
+ "epoch": 24.27947598253275,
+ "grad_norm": 0.021055003628134727,
+ "learning_rate": 0.0006,
+ "loss": 4.278757095336914,
+ "step": 1748
+ },
+ {
+ "epoch": 24.29344978165939,
+ "grad_norm": 0.018618572503328323,
+ "learning_rate": 0.0006,
+ "loss": 4.28849983215332,
+ "step": 1749
+ },
+ {
+ "epoch": 24.307423580786025,
+ "grad_norm": 0.018666435033082962,
+ "learning_rate": 0.0006,
+ "loss": 4.368590354919434,
+ "step": 1750
+ },
+ {
+ "epoch": 24.321397379912664,
+ "grad_norm": 0.020626146346330643,
+ "learning_rate": 0.0006,
+ "loss": 4.2951836585998535,
+ "step": 1751
+ },
+ {
+ "epoch": 24.335371179039303,
+ "grad_norm": 0.01929759792983532,
+ "learning_rate": 0.0006,
+ "loss": 4.3664703369140625,
+ "step": 1752
+ },
+ {
+ "epoch": 24.349344978165938,
+ "grad_norm": 0.020956002175807953,
+ "learning_rate": 0.0006,
+ "loss": 4.345308303833008,
+ "step": 1753
+ },
+ {
+ "epoch": 24.363318777292577,
+ "grad_norm": 0.019227957352995872,
+ "learning_rate": 0.0006,
+ "loss": 4.349849224090576,
+ "step": 1754
+ },
+ {
+ "epoch": 24.377292576419215,
+ "grad_norm": 0.015970097854733467,
+ "learning_rate": 0.0006,
+ "loss": 4.223012447357178,
+ "step": 1755
+ },
+ {
+ "epoch": 24.39126637554585,
+ "grad_norm": 0.015951646491885185,
+ "learning_rate": 0.0006,
+ "loss": 4.236420631408691,
+ "step": 1756
+ },
+ {
+ "epoch": 24.40524017467249,
+ "grad_norm": 0.016032515093684196,
+ "learning_rate": 0.0006,
+ "loss": 4.226899147033691,
+ "step": 1757
+ },
+ {
+ "epoch": 24.419213973799128,
+ "grad_norm": 0.01573132909834385,
+ "learning_rate": 0.0006,
+ "loss": 4.315916061401367,
+ "step": 1758
+ },
+ {
+ "epoch": 24.433187772925763,
+ "grad_norm": 0.015367005951702595,
+ "learning_rate": 0.0006,
+ "loss": 4.358939170837402,
+ "step": 1759
+ },
+ {
+ "epoch": 24.4471615720524,
+ "grad_norm": 0.01669907011091709,
+ "learning_rate": 0.0006,
+ "loss": 4.232147216796875,
+ "step": 1760
+ },
+ {
+ "epoch": 24.46113537117904,
+ "grad_norm": 0.018708152696490288,
+ "learning_rate": 0.0006,
+ "loss": 4.329852104187012,
+ "step": 1761
+ },
+ {
+ "epoch": 24.475109170305675,
+ "grad_norm": 0.019035987555980682,
+ "learning_rate": 0.0006,
+ "loss": 4.260340213775635,
+ "step": 1762
+ },
+ {
+ "epoch": 24.489082969432314,
+ "grad_norm": 0.01832144521176815,
+ "learning_rate": 0.0006,
+ "loss": 4.359543800354004,
+ "step": 1763
+ },
+ {
+ "epoch": 24.503056768558952,
+ "grad_norm": 0.015313294716179371,
+ "learning_rate": 0.0006,
+ "loss": 4.279088020324707,
+ "step": 1764
+ },
+ {
+ "epoch": 24.51703056768559,
+ "grad_norm": 0.015803923830389977,
+ "learning_rate": 0.0006,
+ "loss": 4.230682849884033,
+ "step": 1765
+ },
+ {
+ "epoch": 24.531004366812226,
+ "grad_norm": 0.016748374328017235,
+ "learning_rate": 0.0006,
+ "loss": 4.349661827087402,
+ "step": 1766
+ },
+ {
+ "epoch": 24.544978165938865,
+ "grad_norm": 0.015732428058981895,
+ "learning_rate": 0.0006,
+ "loss": 4.372441291809082,
+ "step": 1767
+ },
+ {
+ "epoch": 24.558951965065503,
+ "grad_norm": 0.016564829275012016,
+ "learning_rate": 0.0006,
+ "loss": 4.323090553283691,
+ "step": 1768
+ },
+ {
+ "epoch": 24.57292576419214,
+ "grad_norm": 0.01603010483086109,
+ "learning_rate": 0.0006,
+ "loss": 4.347898006439209,
+ "step": 1769
+ },
+ {
+ "epoch": 24.586899563318777,
+ "grad_norm": 0.01674809865653515,
+ "learning_rate": 0.0006,
+ "loss": 4.269165515899658,
+ "step": 1770
+ },
+ {
+ "epoch": 24.600873362445416,
+ "grad_norm": 0.017333930358290672,
+ "learning_rate": 0.0006,
+ "loss": 4.353298187255859,
+ "step": 1771
+ },
+ {
+ "epoch": 24.61484716157205,
+ "grad_norm": 0.017519580200314522,
+ "learning_rate": 0.0006,
+ "loss": 4.265199661254883,
+ "step": 1772
+ },
+ {
+ "epoch": 24.62882096069869,
+ "grad_norm": 0.017675306648015976,
+ "learning_rate": 0.0006,
+ "loss": 4.322596549987793,
+ "step": 1773
+ },
+ {
+ "epoch": 24.64279475982533,
+ "grad_norm": 0.017166534438729286,
+ "learning_rate": 0.0006,
+ "loss": 4.413900375366211,
+ "step": 1774
+ },
+ {
+ "epoch": 24.656768558951963,
+ "grad_norm": 0.016362346708774567,
+ "learning_rate": 0.0006,
+ "loss": 4.307236671447754,
+ "step": 1775
+ },
+ {
+ "epoch": 24.670742358078602,
+ "grad_norm": 0.016086872667074203,
+ "learning_rate": 0.0006,
+ "loss": 4.235316276550293,
+ "step": 1776
+ },
+ {
+ "epoch": 24.68471615720524,
+ "grad_norm": 0.01689537614583969,
+ "learning_rate": 0.0006,
+ "loss": 4.390591621398926,
+ "step": 1777
+ },
+ {
+ "epoch": 24.69868995633188,
+ "grad_norm": 0.016836725175380707,
+ "learning_rate": 0.0006,
+ "loss": 4.2990031242370605,
+ "step": 1778
+ },
+ {
+ "epoch": 24.712663755458514,
+ "grad_norm": 0.015627335757017136,
+ "learning_rate": 0.0006,
+ "loss": 4.166745185852051,
+ "step": 1779
+ },
+ {
+ "epoch": 24.726637554585153,
+ "grad_norm": 0.015007016249001026,
+ "learning_rate": 0.0006,
+ "loss": 4.295273303985596,
+ "step": 1780
+ },
+ {
+ "epoch": 24.74061135371179,
+ "grad_norm": 0.016084209084510803,
+ "learning_rate": 0.0006,
+ "loss": 4.342424392700195,
+ "step": 1781
+ },
+ {
+ "epoch": 24.754585152838427,
+ "grad_norm": 0.01595073565840721,
+ "learning_rate": 0.0006,
+ "loss": 4.3060102462768555,
+ "step": 1782
+ },
+ {
+ "epoch": 24.768558951965066,
+ "grad_norm": 0.016113461926579475,
+ "learning_rate": 0.0006,
+ "loss": 4.267908096313477,
+ "step": 1783
+ },
+ {
+ "epoch": 24.782532751091704,
+ "grad_norm": 0.01564556173980236,
+ "learning_rate": 0.0006,
+ "loss": 4.324642658233643,
+ "step": 1784
+ },
+ {
+ "epoch": 24.79650655021834,
+ "grad_norm": 0.01575268618762493,
+ "learning_rate": 0.0006,
+ "loss": 4.328181743621826,
+ "step": 1785
+ },
+ {
+ "epoch": 24.810480349344978,
+ "grad_norm": 0.016830824315547943,
+ "learning_rate": 0.0006,
+ "loss": 4.247906684875488,
+ "step": 1786
+ },
+ {
+ "epoch": 24.824454148471617,
+ "grad_norm": 0.017989547923207283,
+ "learning_rate": 0.0006,
+ "loss": 4.228281021118164,
+ "step": 1787
+ },
+ {
+ "epoch": 24.83842794759825,
+ "grad_norm": 0.016783086583018303,
+ "learning_rate": 0.0006,
+ "loss": 4.2880964279174805,
+ "step": 1788
+ },
+ {
+ "epoch": 24.85240174672489,
+ "grad_norm": 0.016467561945319176,
+ "learning_rate": 0.0006,
+ "loss": 4.36118745803833,
+ "step": 1789
+ },
+ {
+ "epoch": 24.86637554585153,
+ "grad_norm": 0.01780523732304573,
+ "learning_rate": 0.0006,
+ "loss": 4.172321319580078,
+ "step": 1790
+ },
+ {
+ "epoch": 24.880349344978168,
+ "grad_norm": 0.017102031037211418,
+ "learning_rate": 0.0006,
+ "loss": 4.352826118469238,
+ "step": 1791
+ },
+ {
+ "epoch": 24.894323144104803,
+ "grad_norm": 0.016409218311309814,
+ "learning_rate": 0.0006,
+ "loss": 4.212162017822266,
+ "step": 1792
+ },
+ {
+ "epoch": 24.90829694323144,
+ "grad_norm": 0.018660584464669228,
+ "learning_rate": 0.0006,
+ "loss": 4.346684455871582,
+ "step": 1793
+ },
+ {
+ "epoch": 24.92227074235808,
+ "grad_norm": 0.017389440909028053,
+ "learning_rate": 0.0006,
+ "loss": 4.197913646697998,
+ "step": 1794
+ },
+ {
+ "epoch": 24.936244541484715,
+ "grad_norm": 0.01596485823392868,
+ "learning_rate": 0.0006,
+ "loss": 4.370617866516113,
+ "step": 1795
+ },
+ {
+ "epoch": 24.950218340611354,
+ "grad_norm": 0.01709776371717453,
+ "learning_rate": 0.0006,
+ "loss": 4.313847064971924,
+ "step": 1796
+ },
+ {
+ "epoch": 24.964192139737992,
+ "grad_norm": 0.016571134328842163,
+ "learning_rate": 0.0006,
+ "loss": 4.1895856857299805,
+ "step": 1797
+ },
+ {
+ "epoch": 24.978165938864628,
+ "grad_norm": 0.016851048916578293,
+ "learning_rate": 0.0006,
+ "loss": 4.350772857666016,
+ "step": 1798
+ },
+ {
+ "epoch": 24.992139737991266,
+ "grad_norm": 0.017007440328598022,
+ "learning_rate": 0.0006,
+ "loss": 4.399008750915527,
+ "step": 1799
+ },
+ {
+ "epoch": 25.0,
+ "grad_norm": 0.01791212148964405,
+ "learning_rate": 0.0006,
+ "loss": 4.306121826171875,
+ "step": 1800
+ },
+ {
+ "epoch": 25.0,
+ "eval_loss": 4.636417865753174,
+ "eval_runtime": 56.8296,
+ "eval_samples_per_second": 42.971,
+ "eval_steps_per_second": 1.355,
+ "step": 1800
+ },
+ {
+ "epoch": 25.01397379912664,
+ "grad_norm": 0.01729617640376091,
+ "learning_rate": 0.0006,
+ "loss": 4.257018089294434,
+ "step": 1801
+ },
+ {
+ "epoch": 25.027947598253274,
+ "grad_norm": 0.017466643825173378,
+ "learning_rate": 0.0006,
+ "loss": 4.307605266571045,
+ "step": 1802
+ },
+ {
+ "epoch": 25.041921397379912,
+ "grad_norm": 0.018614279106259346,
+ "learning_rate": 0.0006,
+ "loss": 4.360141277313232,
+ "step": 1803
+ },
+ {
+ "epoch": 25.05589519650655,
+ "grad_norm": 0.02010771445930004,
+ "learning_rate": 0.0006,
+ "loss": 4.196428298950195,
+ "step": 1804
+ },
+ {
+ "epoch": 25.069868995633186,
+ "grad_norm": 0.02090228535234928,
+ "learning_rate": 0.0006,
+ "loss": 4.288947105407715,
+ "step": 1805
+ },
+ {
+ "epoch": 25.083842794759825,
+ "grad_norm": 0.02050255797803402,
+ "learning_rate": 0.0006,
+ "loss": 4.259582042694092,
+ "step": 1806
+ },
+ {
+ "epoch": 25.097816593886463,
+ "grad_norm": 0.021844089031219482,
+ "learning_rate": 0.0006,
+ "loss": 4.405138969421387,
+ "step": 1807
+ },
+ {
+ "epoch": 25.111790393013102,
+ "grad_norm": 0.022287018597126007,
+ "learning_rate": 0.0006,
+ "loss": 4.333793640136719,
+ "step": 1808
+ },
+ {
+ "epoch": 25.125764192139737,
+ "grad_norm": 0.02288329415023327,
+ "learning_rate": 0.0006,
+ "loss": 4.3910980224609375,
+ "step": 1809
+ },
+ {
+ "epoch": 25.139737991266376,
+ "grad_norm": 0.02373199723660946,
+ "learning_rate": 0.0006,
+ "loss": 4.32921028137207,
+ "step": 1810
+ },
+ {
+ "epoch": 25.153711790393015,
+ "grad_norm": 0.024960234761238098,
+ "learning_rate": 0.0006,
+ "loss": 4.295609474182129,
+ "step": 1811
+ },
+ {
+ "epoch": 25.16768558951965,
+ "grad_norm": 0.026743892580270767,
+ "learning_rate": 0.0006,
+ "loss": 4.315308094024658,
+ "step": 1812
+ },
+ {
+ "epoch": 25.18165938864629,
+ "grad_norm": 0.022138185799121857,
+ "learning_rate": 0.0006,
+ "loss": 4.360208511352539,
+ "step": 1813
+ },
+ {
+ "epoch": 25.195633187772927,
+ "grad_norm": 0.022120574489235878,
+ "learning_rate": 0.0006,
+ "loss": 4.1973772048950195,
+ "step": 1814
+ },
+ {
+ "epoch": 25.209606986899562,
+ "grad_norm": 0.02230502851307392,
+ "learning_rate": 0.0006,
+ "loss": 4.285558700561523,
+ "step": 1815
+ },
+ {
+ "epoch": 25.2235807860262,
+ "grad_norm": 0.022301986813545227,
+ "learning_rate": 0.0006,
+ "loss": 4.238343238830566,
+ "step": 1816
+ },
+ {
+ "epoch": 25.23755458515284,
+ "grad_norm": 0.020050447434186935,
+ "learning_rate": 0.0006,
+ "loss": 4.298235893249512,
+ "step": 1817
+ },
+ {
+ "epoch": 25.251528384279474,
+ "grad_norm": 0.01987724006175995,
+ "learning_rate": 0.0006,
+ "loss": 4.301384449005127,
+ "step": 1818
+ },
+ {
+ "epoch": 25.265502183406113,
+ "grad_norm": 0.020440855994820595,
+ "learning_rate": 0.0006,
+ "loss": 4.269218444824219,
+ "step": 1819
+ },
+ {
+ "epoch": 25.27947598253275,
+ "grad_norm": 0.01891893707215786,
+ "learning_rate": 0.0006,
+ "loss": 4.2401957511901855,
+ "step": 1820
+ },
+ {
+ "epoch": 25.29344978165939,
+ "grad_norm": 0.01903688907623291,
+ "learning_rate": 0.0006,
+ "loss": 4.139670372009277,
+ "step": 1821
+ },
+ {
+ "epoch": 25.307423580786025,
+ "grad_norm": 0.018856938928365707,
+ "learning_rate": 0.0006,
+ "loss": 4.233214378356934,
+ "step": 1822
+ },
+ {
+ "epoch": 25.321397379912664,
+ "grad_norm": 0.017791323363780975,
+ "learning_rate": 0.0006,
+ "loss": 4.2342023849487305,
+ "step": 1823
+ },
+ {
+ "epoch": 25.335371179039303,
+ "grad_norm": 0.017304565757513046,
+ "learning_rate": 0.0006,
+ "loss": 4.355198860168457,
+ "step": 1824
+ },
+ {
+ "epoch": 25.349344978165938,
+ "grad_norm": 0.01576206088066101,
+ "learning_rate": 0.0006,
+ "loss": 4.114147186279297,
+ "step": 1825
+ },
+ {
+ "epoch": 25.363318777292577,
+ "grad_norm": 0.015105154365301132,
+ "learning_rate": 0.0006,
+ "loss": 4.313795566558838,
+ "step": 1826
+ },
+ {
+ "epoch": 25.377292576419215,
+ "grad_norm": 0.01563587784767151,
+ "learning_rate": 0.0006,
+ "loss": 4.308282852172852,
+ "step": 1827
+ },
+ {
+ "epoch": 25.39126637554585,
+ "grad_norm": 0.015980906784534454,
+ "learning_rate": 0.0006,
+ "loss": 4.345149040222168,
+ "step": 1828
+ },
+ {
+ "epoch": 25.40524017467249,
+ "grad_norm": 0.01669284701347351,
+ "learning_rate": 0.0006,
+ "loss": 4.237611293792725,
+ "step": 1829
+ },
+ {
+ "epoch": 25.419213973799128,
+ "grad_norm": 0.016202064231038094,
+ "learning_rate": 0.0006,
+ "loss": 4.3072309494018555,
+ "step": 1830
+ },
+ {
+ "epoch": 25.433187772925763,
+ "grad_norm": 0.01478662807494402,
+ "learning_rate": 0.0006,
+ "loss": 4.276215076446533,
+ "step": 1831
+ },
+ {
+ "epoch": 25.4471615720524,
+ "grad_norm": 0.015168731100857258,
+ "learning_rate": 0.0006,
+ "loss": 4.37509822845459,
+ "step": 1832
+ },
+ {
+ "epoch": 25.46113537117904,
+ "grad_norm": 0.01648980937898159,
+ "learning_rate": 0.0006,
+ "loss": 4.349167346954346,
+ "step": 1833
+ },
+ {
+ "epoch": 25.475109170305675,
+ "grad_norm": 0.01708398200571537,
+ "learning_rate": 0.0006,
+ "loss": 4.346580505371094,
+ "step": 1834
+ },
+ {
+ "epoch": 25.489082969432314,
+ "grad_norm": 0.017370784655213356,
+ "learning_rate": 0.0006,
+ "loss": 4.299466133117676,
+ "step": 1835
+ },
+ {
+ "epoch": 25.503056768558952,
+ "grad_norm": 0.016248228028416634,
+ "learning_rate": 0.0006,
+ "loss": 4.184610366821289,
+ "step": 1836
+ },
+ {
+ "epoch": 25.51703056768559,
+ "grad_norm": 0.017106913030147552,
+ "learning_rate": 0.0006,
+ "loss": 4.366856098175049,
+ "step": 1837
+ },
+ {
+ "epoch": 25.531004366812226,
+ "grad_norm": 0.016071965917944908,
+ "learning_rate": 0.0006,
+ "loss": 4.297691345214844,
+ "step": 1838
+ },
+ {
+ "epoch": 25.544978165938865,
+ "grad_norm": 0.016451986506581306,
+ "learning_rate": 0.0006,
+ "loss": 4.229983329772949,
+ "step": 1839
+ },
+ {
+ "epoch": 25.558951965065503,
+ "grad_norm": 0.016195081174373627,
+ "learning_rate": 0.0006,
+ "loss": 4.259893417358398,
+ "step": 1840
+ },
+ {
+ "epoch": 25.57292576419214,
+ "grad_norm": 0.0165175162255764,
+ "learning_rate": 0.0006,
+ "loss": 4.3348236083984375,
+ "step": 1841
+ },
+ {
+ "epoch": 25.586899563318777,
+ "grad_norm": 0.01651517115533352,
+ "learning_rate": 0.0006,
+ "loss": 4.319541931152344,
+ "step": 1842
+ },
+ {
+ "epoch": 25.600873362445416,
+ "grad_norm": 0.016143690794706345,
+ "learning_rate": 0.0006,
+ "loss": 4.345652103424072,
+ "step": 1843
+ },
+ {
+ "epoch": 25.61484716157205,
+ "grad_norm": 0.015140696428716183,
+ "learning_rate": 0.0006,
+ "loss": 4.212047576904297,
+ "step": 1844
+ },
+ {
+ "epoch": 25.62882096069869,
+ "grad_norm": 0.014637443237006664,
+ "learning_rate": 0.0006,
+ "loss": 4.187453269958496,
+ "step": 1845
+ },
+ {
+ "epoch": 25.64279475982533,
+ "grad_norm": 0.014585614204406738,
+ "learning_rate": 0.0006,
+ "loss": 4.253936767578125,
+ "step": 1846
+ },
+ {
+ "epoch": 25.656768558951963,
+ "grad_norm": 0.015158289112150669,
+ "learning_rate": 0.0006,
+ "loss": 4.34612512588501,
+ "step": 1847
+ },
+ {
+ "epoch": 25.670742358078602,
+ "grad_norm": 0.015202849172055721,
+ "learning_rate": 0.0006,
+ "loss": 4.270634651184082,
+ "step": 1848
+ },
+ {
+ "epoch": 25.68471615720524,
+ "grad_norm": 0.015363575890660286,
+ "learning_rate": 0.0006,
+ "loss": 4.271080493927002,
+ "step": 1849
+ },
+ {
+ "epoch": 25.69868995633188,
+ "grad_norm": 0.014360226690769196,
+ "learning_rate": 0.0006,
+ "loss": 4.375033855438232,
+ "step": 1850
+ },
+ {
+ "epoch": 25.712663755458514,
+ "grad_norm": 0.014263181947171688,
+ "learning_rate": 0.0006,
+ "loss": 4.424138069152832,
+ "step": 1851
+ },
+ {
+ "epoch": 25.726637554585153,
+ "grad_norm": 0.014398688450455666,
+ "learning_rate": 0.0006,
+ "loss": 4.410154819488525,
+ "step": 1852
+ },
+ {
+ "epoch": 25.74061135371179,
+ "grad_norm": 0.016131488606333733,
+ "learning_rate": 0.0006,
+ "loss": 4.324014663696289,
+ "step": 1853
+ },
+ {
+ "epoch": 25.754585152838427,
+ "grad_norm": 0.017568735405802727,
+ "learning_rate": 0.0006,
+ "loss": 4.367861747741699,
+ "step": 1854
+ },
+ {
+ "epoch": 25.768558951965066,
+ "grad_norm": 0.017907511442899704,
+ "learning_rate": 0.0006,
+ "loss": 4.352965354919434,
+ "step": 1855
+ },
+ {
+ "epoch": 25.782532751091704,
+ "grad_norm": 0.016918016597628593,
+ "learning_rate": 0.0006,
+ "loss": 4.252397537231445,
+ "step": 1856
+ },
+ {
+ "epoch": 25.79650655021834,
+ "grad_norm": 0.017274610698223114,
+ "learning_rate": 0.0006,
+ "loss": 4.342309951782227,
+ "step": 1857
+ },
+ {
+ "epoch": 25.810480349344978,
+ "grad_norm": 0.017879825085401535,
+ "learning_rate": 0.0006,
+ "loss": 4.330987453460693,
+ "step": 1858
+ },
+ {
+ "epoch": 25.824454148471617,
+ "grad_norm": 0.01736099272966385,
+ "learning_rate": 0.0006,
+ "loss": 4.307255744934082,
+ "step": 1859
+ },
+ {
+ "epoch": 25.83842794759825,
+ "grad_norm": 0.016575824469327927,
+ "learning_rate": 0.0006,
+ "loss": 4.262016773223877,
+ "step": 1860
+ },
+ {
+ "epoch": 25.85240174672489,
+ "grad_norm": 0.015181140042841434,
+ "learning_rate": 0.0006,
+ "loss": 4.291863441467285,
+ "step": 1861
+ },
+ {
+ "epoch": 25.86637554585153,
+ "grad_norm": 0.015965687111020088,
+ "learning_rate": 0.0006,
+ "loss": 4.424836158752441,
+ "step": 1862
+ },
+ {
+ "epoch": 25.880349344978168,
+ "grad_norm": 0.01687219925224781,
+ "learning_rate": 0.0006,
+ "loss": 4.343531131744385,
+ "step": 1863
+ },
+ {
+ "epoch": 25.894323144104803,
+ "grad_norm": 0.015115504153072834,
+ "learning_rate": 0.0006,
+ "loss": 4.357808589935303,
+ "step": 1864
+ },
+ {
+ "epoch": 25.90829694323144,
+ "grad_norm": 0.014150998555123806,
+ "learning_rate": 0.0006,
+ "loss": 4.324550151824951,
+ "step": 1865
+ },
+ {
+ "epoch": 25.92227074235808,
+ "grad_norm": 0.014989510178565979,
+ "learning_rate": 0.0006,
+ "loss": 4.380428314208984,
+ "step": 1866
+ },
+ {
+ "epoch": 25.936244541484715,
+ "grad_norm": 0.01527960691601038,
+ "learning_rate": 0.0006,
+ "loss": 4.399374961853027,
+ "step": 1867
+ },
+ {
+ "epoch": 25.950218340611354,
+ "grad_norm": 0.015434633940458298,
+ "learning_rate": 0.0006,
+ "loss": 4.260984420776367,
+ "step": 1868
+ },
+ {
+ "epoch": 25.964192139737992,
+ "grad_norm": 0.015370228327810764,
+ "learning_rate": 0.0006,
+ "loss": 4.3340253829956055,
+ "step": 1869
+ },
+ {
+ "epoch": 25.978165938864628,
+ "grad_norm": 0.015348542481660843,
+ "learning_rate": 0.0006,
+ "loss": 4.369132995605469,
+ "step": 1870
+ },
+ {
+ "epoch": 25.992139737991266,
+ "grad_norm": 0.015354936942458153,
+ "learning_rate": 0.0006,
+ "loss": 4.381937026977539,
+ "step": 1871
+ },
+ {
+ "epoch": 26.0,
+ "grad_norm": 0.01796272210776806,
+ "learning_rate": 0.0006,
+ "loss": 4.286768436431885,
+ "step": 1872
+ },
+ {
+ "epoch": 26.0,
+ "eval_loss": 4.576984405517578,
+ "eval_runtime": 57.2091,
+ "eval_samples_per_second": 42.685,
+ "eval_steps_per_second": 1.346,
+ "step": 1872
+ },
+ {
+ "epoch": 26.01397379912664,
+ "grad_norm": 0.01748845726251602,
+ "learning_rate": 0.0006,
+ "loss": 4.337504863739014,
+ "step": 1873
+ },
+ {
+ "epoch": 26.027947598253274,
+ "grad_norm": 0.019244296476244926,
+ "learning_rate": 0.0006,
+ "loss": 4.264280796051025,
+ "step": 1874
+ },
+ {
+ "epoch": 26.041921397379912,
+ "grad_norm": 0.0215692650526762,
+ "learning_rate": 0.0006,
+ "loss": 4.350830078125,
+ "step": 1875
+ },
+ {
+ "epoch": 26.05589519650655,
+ "grad_norm": 0.022189252078533173,
+ "learning_rate": 0.0006,
+ "loss": 4.299098968505859,
+ "step": 1876
+ },
+ {
+ "epoch": 26.069868995633186,
+ "grad_norm": 0.022384043782949448,
+ "learning_rate": 0.0006,
+ "loss": 4.1185173988342285,
+ "step": 1877
+ },
+ {
+ "epoch": 26.083842794759825,
+ "grad_norm": 0.02076014317572117,
+ "learning_rate": 0.0006,
+ "loss": 4.297612190246582,
+ "step": 1878
+ },
+ {
+ "epoch": 26.097816593886463,
+ "grad_norm": 0.020843051373958588,
+ "learning_rate": 0.0006,
+ "loss": 4.294098854064941,
+ "step": 1879
+ },
+ {
+ "epoch": 26.111790393013102,
+ "grad_norm": 0.020654456689953804,
+ "learning_rate": 0.0006,
+ "loss": 4.267856597900391,
+ "step": 1880
+ },
+ {
+ "epoch": 26.125764192139737,
+ "grad_norm": 0.01860472559928894,
+ "learning_rate": 0.0006,
+ "loss": 4.284232139587402,
+ "step": 1881
+ },
+ {
+ "epoch": 26.139737991266376,
+ "grad_norm": 0.018719421699643135,
+ "learning_rate": 0.0006,
+ "loss": 4.277214527130127,
+ "step": 1882
+ },
+ {
+ "epoch": 26.153711790393015,
+ "grad_norm": 0.020903829485177994,
+ "learning_rate": 0.0006,
+ "loss": 4.254947662353516,
+ "step": 1883
+ },
+ {
+ "epoch": 26.16768558951965,
+ "grad_norm": 0.02020149491727352,
+ "learning_rate": 0.0006,
+ "loss": 4.2526936531066895,
+ "step": 1884
+ },
+ {
+ "epoch": 26.18165938864629,
+ "grad_norm": 0.0180392786860466,
+ "learning_rate": 0.0006,
+ "loss": 4.309922695159912,
+ "step": 1885
+ },
+ {
+ "epoch": 26.195633187772927,
+ "grad_norm": 0.018163125962018967,
+ "learning_rate": 0.0006,
+ "loss": 4.13860559463501,
+ "step": 1886
+ },
+ {
+ "epoch": 26.209606986899562,
+ "grad_norm": 0.01853291131556034,
+ "learning_rate": 0.0006,
+ "loss": 4.190926551818848,
+ "step": 1887
+ },
+ {
+ "epoch": 26.2235807860262,
+ "grad_norm": 0.01757677271962166,
+ "learning_rate": 0.0006,
+ "loss": 4.163999557495117,
+ "step": 1888
+ },
+ {
+ "epoch": 26.23755458515284,
+ "grad_norm": 0.01767992228269577,
+ "learning_rate": 0.0006,
+ "loss": 4.184430122375488,
+ "step": 1889
+ },
+ {
+ "epoch": 26.251528384279474,
+ "grad_norm": 0.018139973282814026,
+ "learning_rate": 0.0006,
+ "loss": 4.300766944885254,
+ "step": 1890
+ },
+ {
+ "epoch": 26.265502183406113,
+ "grad_norm": 0.020184461027383804,
+ "learning_rate": 0.0006,
+ "loss": 4.232537269592285,
+ "step": 1891
+ },
+ {
+ "epoch": 26.27947598253275,
+ "grad_norm": 0.02028856985270977,
+ "learning_rate": 0.0006,
+ "loss": 4.324883460998535,
+ "step": 1892
+ },
+ {
+ "epoch": 26.29344978165939,
+ "grad_norm": 0.017120616510510445,
+ "learning_rate": 0.0006,
+ "loss": 4.23493766784668,
+ "step": 1893
+ },
+ {
+ "epoch": 26.307423580786025,
+ "grad_norm": 0.016202978789806366,
+ "learning_rate": 0.0006,
+ "loss": 4.197602272033691,
+ "step": 1894
+ },
+ {
+ "epoch": 26.321397379912664,
+ "grad_norm": 0.017231421545147896,
+ "learning_rate": 0.0006,
+ "loss": 4.302001953125,
+ "step": 1895
+ },
+ {
+ "epoch": 26.335371179039303,
+ "grad_norm": 0.01725655607879162,
+ "learning_rate": 0.0006,
+ "loss": 4.356800079345703,
+ "step": 1896
+ },
+ {
+ "epoch": 26.349344978165938,
+ "grad_norm": 0.016657443717122078,
+ "learning_rate": 0.0006,
+ "loss": 4.266307830810547,
+ "step": 1897
+ },
+ {
+ "epoch": 26.363318777292577,
+ "grad_norm": 0.01616556942462921,
+ "learning_rate": 0.0006,
+ "loss": 4.307868003845215,
+ "step": 1898
+ },
+ {
+ "epoch": 26.377292576419215,
+ "grad_norm": 0.016621306538581848,
+ "learning_rate": 0.0006,
+ "loss": 4.172905921936035,
+ "step": 1899
+ },
+ {
+ "epoch": 26.39126637554585,
+ "grad_norm": 0.016243206337094307,
+ "learning_rate": 0.0006,
+ "loss": 4.162896156311035,
+ "step": 1900
+ },
+ {
+ "epoch": 26.40524017467249,
+ "grad_norm": 0.016663808375597,
+ "learning_rate": 0.0006,
+ "loss": 4.258232116699219,
+ "step": 1901
+ },
+ {
+ "epoch": 26.419213973799128,
+ "grad_norm": 0.01677478663623333,
+ "learning_rate": 0.0006,
+ "loss": 4.336090087890625,
+ "step": 1902
+ },
+ {
+ "epoch": 26.433187772925763,
+ "grad_norm": 0.01641303487122059,
+ "learning_rate": 0.0006,
+ "loss": 4.191476821899414,
+ "step": 1903
+ },
+ {
+ "epoch": 26.4471615720524,
+ "grad_norm": 0.016700323671102524,
+ "learning_rate": 0.0006,
+ "loss": 4.299427509307861,
+ "step": 1904
+ },
+ {
+ "epoch": 26.46113537117904,
+ "grad_norm": 0.017595035955309868,
+ "learning_rate": 0.0006,
+ "loss": 4.312828063964844,
+ "step": 1905
+ },
+ {
+ "epoch": 26.475109170305675,
+ "grad_norm": 0.020300284028053284,
+ "learning_rate": 0.0006,
+ "loss": 4.28331184387207,
+ "step": 1906
+ },
+ {
+ "epoch": 26.489082969432314,
+ "grad_norm": 0.02198321744799614,
+ "learning_rate": 0.0006,
+ "loss": 4.3280839920043945,
+ "step": 1907
+ },
+ {
+ "epoch": 26.503056768558952,
+ "grad_norm": 0.02003200724720955,
+ "learning_rate": 0.0006,
+ "loss": 4.325623989105225,
+ "step": 1908
+ },
+ {
+ "epoch": 26.51703056768559,
+ "grad_norm": 0.01870287023484707,
+ "learning_rate": 0.0006,
+ "loss": 4.251203536987305,
+ "step": 1909
+ },
+ {
+ "epoch": 26.531004366812226,
+ "grad_norm": 0.01835448481142521,
+ "learning_rate": 0.0006,
+ "loss": 4.335144996643066,
+ "step": 1910
+ },
+ {
+ "epoch": 26.544978165938865,
+ "grad_norm": 0.02046327292919159,
+ "learning_rate": 0.0006,
+ "loss": 4.238023281097412,
+ "step": 1911
+ },
+ {
+ "epoch": 26.558951965065503,
+ "grad_norm": 0.019771311432123184,
+ "learning_rate": 0.0006,
+ "loss": 4.353594779968262,
+ "step": 1912
+ },
+ {
+ "epoch": 26.57292576419214,
+ "grad_norm": 0.018090544268488884,
+ "learning_rate": 0.0006,
+ "loss": 4.304343223571777,
+ "step": 1913
+ },
+ {
+ "epoch": 26.586899563318777,
+ "grad_norm": 0.017489777877926826,
+ "learning_rate": 0.0006,
+ "loss": 4.204375267028809,
+ "step": 1914
+ },
+ {
+ "epoch": 26.600873362445416,
+ "grad_norm": 0.016785891726613045,
+ "learning_rate": 0.0006,
+ "loss": 4.286669731140137,
+ "step": 1915
+ },
+ {
+ "epoch": 26.61484716157205,
+ "grad_norm": 0.016382789239287376,
+ "learning_rate": 0.0006,
+ "loss": 4.347662448883057,
+ "step": 1916
+ },
+ {
+ "epoch": 26.62882096069869,
+ "grad_norm": 0.016581876203417778,
+ "learning_rate": 0.0006,
+ "loss": 4.2716193199157715,
+ "step": 1917
+ },
+ {
+ "epoch": 26.64279475982533,
+ "grad_norm": 0.015627184882760048,
+ "learning_rate": 0.0006,
+ "loss": 4.311489105224609,
+ "step": 1918
+ },
+ {
+ "epoch": 26.656768558951963,
+ "grad_norm": 0.015908483415842056,
+ "learning_rate": 0.0006,
+ "loss": 4.262099742889404,
+ "step": 1919
+ },
+ {
+ "epoch": 26.670742358078602,
+ "grad_norm": 0.015576236881315708,
+ "learning_rate": 0.0006,
+ "loss": 4.275871276855469,
+ "step": 1920
+ },
+ {
+ "epoch": 26.68471615720524,
+ "grad_norm": 0.014621064998209476,
+ "learning_rate": 0.0006,
+ "loss": 4.184260845184326,
+ "step": 1921
+ },
+ {
+ "epoch": 26.69868995633188,
+ "grad_norm": 0.014797811396420002,
+ "learning_rate": 0.0006,
+ "loss": 4.3412628173828125,
+ "step": 1922
+ },
+ {
+ "epoch": 26.712663755458514,
+ "grad_norm": 0.015610570088028908,
+ "learning_rate": 0.0006,
+ "loss": 4.181094646453857,
+ "step": 1923
+ },
+ {
+ "epoch": 26.726637554585153,
+ "grad_norm": 0.016048265621066093,
+ "learning_rate": 0.0006,
+ "loss": 4.365290641784668,
+ "step": 1924
+ },
+ {
+ "epoch": 26.74061135371179,
+ "grad_norm": 0.016604071483016014,
+ "learning_rate": 0.0006,
+ "loss": 4.2628912925720215,
+ "step": 1925
+ },
+ {
+ "epoch": 26.754585152838427,
+ "grad_norm": 0.017044221982359886,
+ "learning_rate": 0.0006,
+ "loss": 4.378512382507324,
+ "step": 1926
+ },
+ {
+ "epoch": 26.768558951965066,
+ "grad_norm": 0.01746384985744953,
+ "learning_rate": 0.0006,
+ "loss": 4.3888773918151855,
+ "step": 1927
+ },
+ {
+ "epoch": 26.782532751091704,
+ "grad_norm": 0.01572711206972599,
+ "learning_rate": 0.0006,
+ "loss": 4.189935684204102,
+ "step": 1928
+ },
+ {
+ "epoch": 26.79650655021834,
+ "grad_norm": 0.014976629987359047,
+ "learning_rate": 0.0006,
+ "loss": 4.280525207519531,
+ "step": 1929
+ },
+ {
+ "epoch": 26.810480349344978,
+ "grad_norm": 0.014804664999246597,
+ "learning_rate": 0.0006,
+ "loss": 4.270630836486816,
+ "step": 1930
+ },
+ {
+ "epoch": 26.824454148471617,
+ "grad_norm": 0.014641920104622841,
+ "learning_rate": 0.0006,
+ "loss": 4.319505214691162,
+ "step": 1931
+ },
+ {
+ "epoch": 26.83842794759825,
+ "grad_norm": 0.014928505755960941,
+ "learning_rate": 0.0006,
+ "loss": 4.180147647857666,
+ "step": 1932
+ },
+ {
+ "epoch": 26.85240174672489,
+ "grad_norm": 0.016130365431308746,
+ "learning_rate": 0.0006,
+ "loss": 4.260660648345947,
+ "step": 1933
+ },
+ {
+ "epoch": 26.86637554585153,
+ "grad_norm": 0.015642493963241577,
+ "learning_rate": 0.0006,
+ "loss": 4.304064750671387,
+ "step": 1934
+ },
+ {
+ "epoch": 26.880349344978168,
+ "grad_norm": 0.015988217666745186,
+ "learning_rate": 0.0006,
+ "loss": 4.2548017501831055,
+ "step": 1935
+ },
+ {
+ "epoch": 26.894323144104803,
+ "grad_norm": 0.016501398757100105,
+ "learning_rate": 0.0006,
+ "loss": 4.260058879852295,
+ "step": 1936
+ },
+ {
+ "epoch": 26.90829694323144,
+ "grad_norm": 0.01645040698349476,
+ "learning_rate": 0.0006,
+ "loss": 4.483081340789795,
+ "step": 1937
+ },
+ {
+ "epoch": 26.92227074235808,
+ "grad_norm": 0.016140863299369812,
+ "learning_rate": 0.0006,
+ "loss": 4.1776838302612305,
+ "step": 1938
+ },
+ {
+ "epoch": 26.936244541484715,
+ "grad_norm": 0.016681019216775894,
+ "learning_rate": 0.0006,
+ "loss": 4.224747657775879,
+ "step": 1939
+ },
+ {
+ "epoch": 26.950218340611354,
+ "grad_norm": 0.01650378853082657,
+ "learning_rate": 0.0006,
+ "loss": 4.305703163146973,
+ "step": 1940
+ },
+ {
+ "epoch": 26.964192139737992,
+ "grad_norm": 0.017051683738827705,
+ "learning_rate": 0.0006,
+ "loss": 4.285835266113281,
+ "step": 1941
+ },
+ {
+ "epoch": 26.978165938864628,
+ "grad_norm": 0.016894327476620674,
+ "learning_rate": 0.0006,
+ "loss": 4.318901538848877,
+ "step": 1942
+ },
+ {
+ "epoch": 26.992139737991266,
+ "grad_norm": 0.017313100397586823,
+ "learning_rate": 0.0006,
+ "loss": 4.281434535980225,
+ "step": 1943
+ },
+ {
+ "epoch": 27.0,
+ "grad_norm": 0.018450884148478508,
+ "learning_rate": 0.0006,
+ "loss": 4.335104942321777,
+ "step": 1944
+ },
+ {
+ "epoch": 27.0,
+ "eval_loss": 4.640687465667725,
+ "eval_runtime": 56.9624,
+ "eval_samples_per_second": 42.87,
+ "eval_steps_per_second": 1.352,
+ "step": 1944
+ },
+ {
+ "epoch": 27.01397379912664,
+ "grad_norm": 0.017755698412656784,
+ "learning_rate": 0.0006,
+ "loss": 4.278824329376221,
+ "step": 1945
+ },
+ {
+ "epoch": 27.027947598253274,
+ "grad_norm": 0.01724310778081417,
+ "learning_rate": 0.0006,
+ "loss": 4.311049461364746,
+ "step": 1946
+ },
+ {
+ "epoch": 27.041921397379912,
+ "grad_norm": 0.01690123789012432,
+ "learning_rate": 0.0006,
+ "loss": 4.261994361877441,
+ "step": 1947
+ },
+ {
+ "epoch": 27.05589519650655,
+ "grad_norm": 0.018406696617603302,
+ "learning_rate": 0.0006,
+ "loss": 4.1925153732299805,
+ "step": 1948
+ },
+ {
+ "epoch": 27.069868995633186,
+ "grad_norm": 0.019945867359638214,
+ "learning_rate": 0.0006,
+ "loss": 4.108537197113037,
+ "step": 1949
+ },
+ {
+ "epoch": 27.083842794759825,
+ "grad_norm": 0.020316628739237785,
+ "learning_rate": 0.0006,
+ "loss": 4.209043025970459,
+ "step": 1950
+ },
+ {
+ "epoch": 27.097816593886463,
+ "grad_norm": 0.019690370187163353,
+ "learning_rate": 0.0006,
+ "loss": 4.195789337158203,
+ "step": 1951
+ },
+ {
+ "epoch": 27.111790393013102,
+ "grad_norm": 0.01785232499241829,
+ "learning_rate": 0.0006,
+ "loss": 4.21466064453125,
+ "step": 1952
+ },
+ {
+ "epoch": 27.125764192139737,
+ "grad_norm": 0.0164767038077116,
+ "learning_rate": 0.0006,
+ "loss": 4.270247459411621,
+ "step": 1953
+ },
+ {
+ "epoch": 27.139737991266376,
+ "grad_norm": 0.018008455634117126,
+ "learning_rate": 0.0006,
+ "loss": 4.201042175292969,
+ "step": 1954
+ },
+ {
+ "epoch": 27.153711790393015,
+ "grad_norm": 0.01807340793311596,
+ "learning_rate": 0.0006,
+ "loss": 4.25289249420166,
+ "step": 1955
+ },
+ {
+ "epoch": 27.16768558951965,
+ "grad_norm": 0.016246909275650978,
+ "learning_rate": 0.0006,
+ "loss": 4.116283416748047,
+ "step": 1956
+ },
+ {
+ "epoch": 27.18165938864629,
+ "grad_norm": 0.017545776441693306,
+ "learning_rate": 0.0006,
+ "loss": 4.316999435424805,
+ "step": 1957
+ },
+ {
+ "epoch": 27.195633187772927,
+ "grad_norm": 0.017915375530719757,
+ "learning_rate": 0.0006,
+ "loss": 4.253859519958496,
+ "step": 1958
+ },
+ {
+ "epoch": 27.209606986899562,
+ "grad_norm": 0.018112223595380783,
+ "learning_rate": 0.0006,
+ "loss": 4.2541022300720215,
+ "step": 1959
+ },
+ {
+ "epoch": 27.2235807860262,
+ "grad_norm": 0.01846139505505562,
+ "learning_rate": 0.0006,
+ "loss": 4.20438289642334,
+ "step": 1960
+ },
+ {
+ "epoch": 27.23755458515284,
+ "grad_norm": 0.01943897269666195,
+ "learning_rate": 0.0006,
+ "loss": 4.2799577713012695,
+ "step": 1961
+ },
+ {
+ "epoch": 27.251528384279474,
+ "grad_norm": 0.018253348767757416,
+ "learning_rate": 0.0006,
+ "loss": 4.1720499992370605,
+ "step": 1962
+ },
+ {
+ "epoch": 27.265502183406113,
+ "grad_norm": 0.018292246386408806,
+ "learning_rate": 0.0006,
+ "loss": 4.28231143951416,
+ "step": 1963
+ },
+ {
+ "epoch": 27.27947598253275,
+ "grad_norm": 0.01943155936896801,
+ "learning_rate": 0.0006,
+ "loss": 4.314976692199707,
+ "step": 1964
+ },
+ {
+ "epoch": 27.29344978165939,
+ "grad_norm": 0.019064994528889656,
+ "learning_rate": 0.0006,
+ "loss": 4.25328254699707,
+ "step": 1965
+ },
+ {
+ "epoch": 27.307423580786025,
+ "grad_norm": 0.017423273995518684,
+ "learning_rate": 0.0006,
+ "loss": 4.195992469787598,
+ "step": 1966
+ },
+ {
+ "epoch": 27.321397379912664,
+ "grad_norm": 0.01677924580872059,
+ "learning_rate": 0.0006,
+ "loss": 4.260043144226074,
+ "step": 1967
+ },
+ {
+ "epoch": 27.335371179039303,
+ "grad_norm": 0.01720673404633999,
+ "learning_rate": 0.0006,
+ "loss": 4.222780704498291,
+ "step": 1968
+ },
+ {
+ "epoch": 27.349344978165938,
+ "grad_norm": 0.017300743609666824,
+ "learning_rate": 0.0006,
+ "loss": 4.109524250030518,
+ "step": 1969
+ },
+ {
+ "epoch": 27.363318777292577,
+ "grad_norm": 0.016065871343016624,
+ "learning_rate": 0.0006,
+ "loss": 4.36108922958374,
+ "step": 1970
+ },
+ {
+ "epoch": 27.377292576419215,
+ "grad_norm": 0.016626616939902306,
+ "learning_rate": 0.0006,
+ "loss": 4.220630645751953,
+ "step": 1971
+ },
+ {
+ "epoch": 27.39126637554585,
+ "grad_norm": 0.016482602804899216,
+ "learning_rate": 0.0006,
+ "loss": 4.344797134399414,
+ "step": 1972
+ },
+ {
+ "epoch": 27.40524017467249,
+ "grad_norm": 0.01727953553199768,
+ "learning_rate": 0.0006,
+ "loss": 4.296645164489746,
+ "step": 1973
+ },
+ {
+ "epoch": 27.419213973799128,
+ "grad_norm": 0.01522731315344572,
+ "learning_rate": 0.0006,
+ "loss": 4.223751068115234,
+ "step": 1974
+ },
+ {
+ "epoch": 27.433187772925763,
+ "grad_norm": 0.01581776700913906,
+ "learning_rate": 0.0006,
+ "loss": 4.351379871368408,
+ "step": 1975
+ },
+ {
+ "epoch": 27.4471615720524,
+ "grad_norm": 0.016394605860114098,
+ "learning_rate": 0.0006,
+ "loss": 4.2412919998168945,
+ "step": 1976
+ },
+ {
+ "epoch": 27.46113537117904,
+ "grad_norm": 0.01751169003546238,
+ "learning_rate": 0.0006,
+ "loss": 4.257606506347656,
+ "step": 1977
+ },
+ {
+ "epoch": 27.475109170305675,
+ "grad_norm": 0.017452310770750046,
+ "learning_rate": 0.0006,
+ "loss": 4.380134582519531,
+ "step": 1978
+ },
+ {
+ "epoch": 27.489082969432314,
+ "grad_norm": 0.017741898074746132,
+ "learning_rate": 0.0006,
+ "loss": 4.279632568359375,
+ "step": 1979
+ },
+ {
+ "epoch": 27.503056768558952,
+ "grad_norm": 0.017680590972304344,
+ "learning_rate": 0.0006,
+ "loss": 4.273406028747559,
+ "step": 1980
+ },
+ {
+ "epoch": 27.51703056768559,
+ "grad_norm": 0.01864718459546566,
+ "learning_rate": 0.0006,
+ "loss": 4.278583526611328,
+ "step": 1981
+ },
+ {
+ "epoch": 27.531004366812226,
+ "grad_norm": 0.019558541476726532,
+ "learning_rate": 0.0006,
+ "loss": 4.211919784545898,
+ "step": 1982
+ },
+ {
+ "epoch": 27.544978165938865,
+ "grad_norm": 0.019467400386929512,
+ "learning_rate": 0.0006,
+ "loss": 4.276822566986084,
+ "step": 1983
+ },
+ {
+ "epoch": 27.558951965065503,
+ "grad_norm": 0.01756172813475132,
+ "learning_rate": 0.0006,
+ "loss": 4.165379047393799,
+ "step": 1984
+ },
+ {
+ "epoch": 27.57292576419214,
+ "grad_norm": 0.017458263784646988,
+ "learning_rate": 0.0006,
+ "loss": 4.2992401123046875,
+ "step": 1985
+ },
+ {
+ "epoch": 27.586899563318777,
+ "grad_norm": 0.018089665099978447,
+ "learning_rate": 0.0006,
+ "loss": 4.243378639221191,
+ "step": 1986
+ },
+ {
+ "epoch": 27.600873362445416,
+ "grad_norm": 0.019535524770617485,
+ "learning_rate": 0.0006,
+ "loss": 4.321477890014648,
+ "step": 1987
+ },
+ {
+ "epoch": 27.61484716157205,
+ "grad_norm": 0.01997970975935459,
+ "learning_rate": 0.0006,
+ "loss": 4.123082637786865,
+ "step": 1988
+ },
+ {
+ "epoch": 27.62882096069869,
+ "grad_norm": 0.01913067139685154,
+ "learning_rate": 0.0006,
+ "loss": 4.289896488189697,
+ "step": 1989
+ },
+ {
+ "epoch": 27.64279475982533,
+ "grad_norm": 0.0169806070625782,
+ "learning_rate": 0.0006,
+ "loss": 4.247004985809326,
+ "step": 1990
+ },
+ {
+ "epoch": 27.656768558951963,
+ "grad_norm": 0.017315007746219635,
+ "learning_rate": 0.0006,
+ "loss": 4.2095441818237305,
+ "step": 1991
+ },
+ {
+ "epoch": 27.670742358078602,
+ "grad_norm": 0.015098759904503822,
+ "learning_rate": 0.0006,
+ "loss": 4.145560264587402,
+ "step": 1992
+ },
+ {
+ "epoch": 27.68471615720524,
+ "grad_norm": 0.016240514814853668,
+ "learning_rate": 0.0006,
+ "loss": 4.254927635192871,
+ "step": 1993
+ },
+ {
+ "epoch": 27.69868995633188,
+ "grad_norm": 0.01621108129620552,
+ "learning_rate": 0.0006,
+ "loss": 4.219725608825684,
+ "step": 1994
+ },
+ {
+ "epoch": 27.712663755458514,
+ "grad_norm": 0.015191009268164635,
+ "learning_rate": 0.0006,
+ "loss": 4.194345474243164,
+ "step": 1995
+ },
+ {
+ "epoch": 27.726637554585153,
+ "grad_norm": 0.014387983828783035,
+ "learning_rate": 0.0006,
+ "loss": 4.211467742919922,
+ "step": 1996
+ },
+ {
+ "epoch": 27.74061135371179,
+ "grad_norm": 0.015460561029613018,
+ "learning_rate": 0.0006,
+ "loss": 4.358700752258301,
+ "step": 1997
+ },
+ {
+ "epoch": 27.754585152838427,
+ "grad_norm": 0.015100076794624329,
+ "learning_rate": 0.0006,
+ "loss": 4.171223163604736,
+ "step": 1998
+ },
+ {
+ "epoch": 27.768558951965066,
+ "grad_norm": 0.015607891604304314,
+ "learning_rate": 0.0006,
+ "loss": 4.217874526977539,
+ "step": 1999
+ },
+ {
+ "epoch": 27.782532751091704,
+ "grad_norm": 0.015650689601898193,
+ "learning_rate": 0.0006,
+ "loss": 4.280780792236328,
+ "step": 2000
+ },
+ {
+ "epoch": 27.79650655021834,
+ "grad_norm": 0.01580994389951229,
+ "learning_rate": 0.0006,
+ "loss": 4.159212112426758,
+ "step": 2001
+ },
+ {
+ "epoch": 27.810480349344978,
+ "grad_norm": 0.016607413068413734,
+ "learning_rate": 0.0006,
+ "loss": 4.238853454589844,
+ "step": 2002
+ },
+ {
+ "epoch": 27.824454148471617,
+ "grad_norm": 0.017383860424160957,
+ "learning_rate": 0.0006,
+ "loss": 4.233944892883301,
+ "step": 2003
+ },
+ {
+ "epoch": 27.83842794759825,
+ "grad_norm": 0.014966659247875214,
+ "learning_rate": 0.0006,
+ "loss": 4.167392730712891,
+ "step": 2004
+ },
+ {
+ "epoch": 27.85240174672489,
+ "grad_norm": 0.014589878730475903,
+ "learning_rate": 0.0006,
+ "loss": 4.2390899658203125,
+ "step": 2005
+ },
+ {
+ "epoch": 27.86637554585153,
+ "grad_norm": 0.017179450020194054,
+ "learning_rate": 0.0006,
+ "loss": 4.365785598754883,
+ "step": 2006
+ },
+ {
+ "epoch": 27.880349344978168,
+ "grad_norm": 0.01893901452422142,
+ "learning_rate": 0.0006,
+ "loss": 4.289430141448975,
+ "step": 2007
+ },
+ {
+ "epoch": 27.894323144104803,
+ "grad_norm": 0.018619712442159653,
+ "learning_rate": 0.0006,
+ "loss": 4.310665130615234,
+ "step": 2008
+ },
+ {
+ "epoch": 27.90829694323144,
+ "grad_norm": 0.019180385395884514,
+ "learning_rate": 0.0006,
+ "loss": 4.27454137802124,
+ "step": 2009
+ },
+ {
+ "epoch": 27.92227074235808,
+ "grad_norm": 0.017811523750424385,
+ "learning_rate": 0.0006,
+ "loss": 4.312562465667725,
+ "step": 2010
+ },
+ {
+ "epoch": 27.936244541484715,
+ "grad_norm": 0.016583573073148727,
+ "learning_rate": 0.0006,
+ "loss": 4.202253341674805,
+ "step": 2011
+ },
+ {
+ "epoch": 27.950218340611354,
+ "grad_norm": 0.01658615842461586,
+ "learning_rate": 0.0006,
+ "loss": 4.183433532714844,
+ "step": 2012
+ },
+ {
+ "epoch": 27.964192139737992,
+ "grad_norm": 0.015327401459217072,
+ "learning_rate": 0.0006,
+ "loss": 4.2125139236450195,
+ "step": 2013
+ },
+ {
+ "epoch": 27.978165938864628,
+ "grad_norm": 0.015004601329565048,
+ "learning_rate": 0.0006,
+ "loss": 4.257392883300781,
+ "step": 2014
+ },
+ {
+ "epoch": 27.992139737991266,
+ "grad_norm": 0.015259161591529846,
+ "learning_rate": 0.0006,
+ "loss": 4.241818428039551,
+ "step": 2015
+ },
+ {
+ "epoch": 28.0,
+ "grad_norm": 0.016808776184916496,
+ "learning_rate": 0.0006,
+ "loss": 4.146241188049316,
+ "step": 2016
+ },
+ {
+ "epoch": 28.0,
+ "eval_loss": 4.663090705871582,
+ "eval_runtime": 56.4559,
+ "eval_samples_per_second": 43.255,
+ "eval_steps_per_second": 1.364,
+ "step": 2016
+ },
+ {
+ "epoch": 28.01397379912664,
+ "grad_norm": 0.015784382820129395,
+ "learning_rate": 0.0006,
+ "loss": 4.044191360473633,
+ "step": 2017
+ },
+ {
+ "epoch": 28.027947598253274,
+ "grad_norm": 0.015636710450053215,
+ "learning_rate": 0.0006,
+ "loss": 4.128754615783691,
+ "step": 2018
+ },
+ {
+ "epoch": 28.041921397379912,
+ "grad_norm": 0.015364975668489933,
+ "learning_rate": 0.0006,
+ "loss": 4.191835403442383,
+ "step": 2019
+ },
+ {
+ "epoch": 28.05589519650655,
+ "grad_norm": 0.015015700832009315,
+ "learning_rate": 0.0006,
+ "loss": 4.120022773742676,
+ "step": 2020
+ },
+ {
+ "epoch": 28.069868995633186,
+ "grad_norm": 0.016666799783706665,
+ "learning_rate": 0.0006,
+ "loss": 4.33906364440918,
+ "step": 2021
+ },
+ {
+ "epoch": 28.083842794759825,
+ "grad_norm": 0.01780104823410511,
+ "learning_rate": 0.0006,
+ "loss": 4.279458999633789,
+ "step": 2022
+ },
+ {
+ "epoch": 28.097816593886463,
+ "grad_norm": 0.01870882883667946,
+ "learning_rate": 0.0006,
+ "loss": 4.34728479385376,
+ "step": 2023
+ },
+ {
+ "epoch": 28.111790393013102,
+ "grad_norm": 0.018406381830573082,
+ "learning_rate": 0.0006,
+ "loss": 4.211311340332031,
+ "step": 2024
+ },
+ {
+ "epoch": 28.125764192139737,
+ "grad_norm": 0.017833448946475983,
+ "learning_rate": 0.0006,
+ "loss": 4.223254203796387,
+ "step": 2025
+ },
+ {
+ "epoch": 28.139737991266376,
+ "grad_norm": 0.017552688717842102,
+ "learning_rate": 0.0006,
+ "loss": 4.208823204040527,
+ "step": 2026
+ },
+ {
+ "epoch": 28.153711790393015,
+ "grad_norm": 0.018500229343771935,
+ "learning_rate": 0.0006,
+ "loss": 4.198025226593018,
+ "step": 2027
+ },
+ {
+ "epoch": 28.16768558951965,
+ "grad_norm": 0.01782156340777874,
+ "learning_rate": 0.0006,
+ "loss": 4.179978370666504,
+ "step": 2028
+ },
+ {
+ "epoch": 28.18165938864629,
+ "grad_norm": 0.019580967724323273,
+ "learning_rate": 0.0006,
+ "loss": 4.291114807128906,
+ "step": 2029
+ },
+ {
+ "epoch": 28.195633187772927,
+ "grad_norm": 0.020804036408662796,
+ "learning_rate": 0.0006,
+ "loss": 4.220607280731201,
+ "step": 2030
+ },
+ {
+ "epoch": 28.209606986899562,
+ "grad_norm": 0.019595693796873093,
+ "learning_rate": 0.0006,
+ "loss": 4.2905731201171875,
+ "step": 2031
+ },
+ {
+ "epoch": 28.2235807860262,
+ "grad_norm": 0.019247086718678474,
+ "learning_rate": 0.0006,
+ "loss": 4.2467360496521,
+ "step": 2032
+ },
+ {
+ "epoch": 28.23755458515284,
+ "grad_norm": 0.01958036608994007,
+ "learning_rate": 0.0006,
+ "loss": 4.286137580871582,
+ "step": 2033
+ },
+ {
+ "epoch": 28.251528384279474,
+ "grad_norm": 0.021354753524065018,
+ "learning_rate": 0.0006,
+ "loss": 4.234755039215088,
+ "step": 2034
+ },
+ {
+ "epoch": 28.265502183406113,
+ "grad_norm": 0.02206375077366829,
+ "learning_rate": 0.0006,
+ "loss": 4.149641036987305,
+ "step": 2035
+ },
+ {
+ "epoch": 28.27947598253275,
+ "grad_norm": 0.02201310358941555,
+ "learning_rate": 0.0006,
+ "loss": 4.1928629875183105,
+ "step": 2036
+ },
+ {
+ "epoch": 28.29344978165939,
+ "grad_norm": 0.01961950585246086,
+ "learning_rate": 0.0006,
+ "loss": 4.211404323577881,
+ "step": 2037
+ },
+ {
+ "epoch": 28.307423580786025,
+ "grad_norm": 0.019211795181035995,
+ "learning_rate": 0.0006,
+ "loss": 4.200976848602295,
+ "step": 2038
+ },
+ {
+ "epoch": 28.321397379912664,
+ "grad_norm": 0.01779370754957199,
+ "learning_rate": 0.0006,
+ "loss": 4.249148845672607,
+ "step": 2039
+ },
+ {
+ "epoch": 28.335371179039303,
+ "grad_norm": 0.018635908141732216,
+ "learning_rate": 0.0006,
+ "loss": 4.11806058883667,
+ "step": 2040
+ },
+ {
+ "epoch": 28.349344978165938,
+ "grad_norm": 0.019792694598436356,
+ "learning_rate": 0.0006,
+ "loss": 4.335224151611328,
+ "step": 2041
+ },
+ {
+ "epoch": 28.363318777292577,
+ "grad_norm": 0.021731717512011528,
+ "learning_rate": 0.0006,
+ "loss": 4.302778244018555,
+ "step": 2042
+ },
+ {
+ "epoch": 28.377292576419215,
+ "grad_norm": 0.020823447033762932,
+ "learning_rate": 0.0006,
+ "loss": 4.260610580444336,
+ "step": 2043
+ },
+ {
+ "epoch": 28.39126637554585,
+ "grad_norm": 0.019385412335395813,
+ "learning_rate": 0.0006,
+ "loss": 4.2247724533081055,
+ "step": 2044
+ },
+ {
+ "epoch": 28.40524017467249,
+ "grad_norm": 0.019539108499884605,
+ "learning_rate": 0.0006,
+ "loss": 4.236655235290527,
+ "step": 2045
+ },
+ {
+ "epoch": 28.419213973799128,
+ "grad_norm": 0.01942690648138523,
+ "learning_rate": 0.0006,
+ "loss": 4.251850605010986,
+ "step": 2046
+ },
+ {
+ "epoch": 28.433187772925763,
+ "grad_norm": 0.019667886197566986,
+ "learning_rate": 0.0006,
+ "loss": 4.222312927246094,
+ "step": 2047
+ },
+ {
+ "epoch": 28.4471615720524,
+ "grad_norm": 0.02029012329876423,
+ "learning_rate": 0.0006,
+ "loss": 4.246252059936523,
+ "step": 2048
+ },
+ {
+ "epoch": 28.46113537117904,
+ "grad_norm": 0.01784469373524189,
+ "learning_rate": 0.0006,
+ "loss": 4.192128658294678,
+ "step": 2049
+ },
+ {
+ "epoch": 28.475109170305675,
+ "grad_norm": 0.0160287544131279,
+ "learning_rate": 0.0006,
+ "loss": 4.208607196807861,
+ "step": 2050
+ },
+ {
+ "epoch": 28.489082969432314,
+ "grad_norm": 0.015348346903920174,
+ "learning_rate": 0.0006,
+ "loss": 4.236598014831543,
+ "step": 2051
+ },
+ {
+ "epoch": 28.503056768558952,
+ "grad_norm": 0.015198206529021263,
+ "learning_rate": 0.0006,
+ "loss": 4.235552787780762,
+ "step": 2052
+ },
+ {
+ "epoch": 28.51703056768559,
+ "grad_norm": 0.01523754745721817,
+ "learning_rate": 0.0006,
+ "loss": 4.1941728591918945,
+ "step": 2053
+ },
+ {
+ "epoch": 28.531004366812226,
+ "grad_norm": 0.015559614636003971,
+ "learning_rate": 0.0006,
+ "loss": 4.294757843017578,
+ "step": 2054
+ },
+ {
+ "epoch": 28.544978165938865,
+ "grad_norm": 0.01568358950316906,
+ "learning_rate": 0.0006,
+ "loss": 4.213065147399902,
+ "step": 2055
+ },
+ {
+ "epoch": 28.558951965065503,
+ "grad_norm": 0.015336516313254833,
+ "learning_rate": 0.0006,
+ "loss": 4.240293979644775,
+ "step": 2056
+ },
+ {
+ "epoch": 28.57292576419214,
+ "grad_norm": 0.015041496604681015,
+ "learning_rate": 0.0006,
+ "loss": 4.273397445678711,
+ "step": 2057
+ },
+ {
+ "epoch": 28.586899563318777,
+ "grad_norm": 0.01610128954052925,
+ "learning_rate": 0.0006,
+ "loss": 4.125369071960449,
+ "step": 2058
+ },
+ {
+ "epoch": 28.600873362445416,
+ "grad_norm": 0.015974976122379303,
+ "learning_rate": 0.0006,
+ "loss": 4.133459091186523,
+ "step": 2059
+ },
+ {
+ "epoch": 28.61484716157205,
+ "grad_norm": 0.015936799347400665,
+ "learning_rate": 0.0006,
+ "loss": 4.306667327880859,
+ "step": 2060
+ },
+ {
+ "epoch": 28.62882096069869,
+ "grad_norm": 0.01795247197151184,
+ "learning_rate": 0.0006,
+ "loss": 4.205156326293945,
+ "step": 2061
+ },
+ {
+ "epoch": 28.64279475982533,
+ "grad_norm": 0.019639814272522926,
+ "learning_rate": 0.0006,
+ "loss": 4.277614593505859,
+ "step": 2062
+ },
+ {
+ "epoch": 28.656768558951963,
+ "grad_norm": 0.01722300611436367,
+ "learning_rate": 0.0006,
+ "loss": 4.191032886505127,
+ "step": 2063
+ },
+ {
+ "epoch": 28.670742358078602,
+ "grad_norm": 0.01612633652985096,
+ "learning_rate": 0.0006,
+ "loss": 4.20717191696167,
+ "step": 2064
+ },
+ {
+ "epoch": 28.68471615720524,
+ "grad_norm": 0.01603279449045658,
+ "learning_rate": 0.0006,
+ "loss": 4.228884220123291,
+ "step": 2065
+ },
+ {
+ "epoch": 28.69868995633188,
+ "grad_norm": 0.016405927017331123,
+ "learning_rate": 0.0006,
+ "loss": 4.201292991638184,
+ "step": 2066
+ },
+ {
+ "epoch": 28.712663755458514,
+ "grad_norm": 0.016644183546304703,
+ "learning_rate": 0.0006,
+ "loss": 4.239282131195068,
+ "step": 2067
+ },
+ {
+ "epoch": 28.726637554585153,
+ "grad_norm": 0.0175587497651577,
+ "learning_rate": 0.0006,
+ "loss": 4.289038181304932,
+ "step": 2068
+ },
+ {
+ "epoch": 28.74061135371179,
+ "grad_norm": 0.018223397433757782,
+ "learning_rate": 0.0006,
+ "loss": 4.2266998291015625,
+ "step": 2069
+ },
+ {
+ "epoch": 28.754585152838427,
+ "grad_norm": 0.018680868670344353,
+ "learning_rate": 0.0006,
+ "loss": 4.348849296569824,
+ "step": 2070
+ },
+ {
+ "epoch": 28.768558951965066,
+ "grad_norm": 0.016726728528738022,
+ "learning_rate": 0.0006,
+ "loss": 4.292080879211426,
+ "step": 2071
+ },
+ {
+ "epoch": 28.782532751091704,
+ "grad_norm": 0.017386795952916145,
+ "learning_rate": 0.0006,
+ "loss": 4.208870887756348,
+ "step": 2072
+ },
+ {
+ "epoch": 28.79650655021834,
+ "grad_norm": 0.01865178905427456,
+ "learning_rate": 0.0006,
+ "loss": 4.054150581359863,
+ "step": 2073
+ },
+ {
+ "epoch": 28.810480349344978,
+ "grad_norm": 0.01776815392076969,
+ "learning_rate": 0.0006,
+ "loss": 4.202404499053955,
+ "step": 2074
+ },
+ {
+ "epoch": 28.824454148471617,
+ "grad_norm": 0.01643497310578823,
+ "learning_rate": 0.0006,
+ "loss": 4.131855010986328,
+ "step": 2075
+ },
+ {
+ "epoch": 28.83842794759825,
+ "grad_norm": 0.01574273779988289,
+ "learning_rate": 0.0006,
+ "loss": 4.296788692474365,
+ "step": 2076
+ },
+ {
+ "epoch": 28.85240174672489,
+ "grad_norm": 0.015613908879458904,
+ "learning_rate": 0.0006,
+ "loss": 4.210949897766113,
+ "step": 2077
+ },
+ {
+ "epoch": 28.86637554585153,
+ "grad_norm": 0.015543212182819843,
+ "learning_rate": 0.0006,
+ "loss": 4.218136310577393,
+ "step": 2078
+ },
+ {
+ "epoch": 28.880349344978168,
+ "grad_norm": 0.014681251719594002,
+ "learning_rate": 0.0006,
+ "loss": 4.19401741027832,
+ "step": 2079
+ },
+ {
+ "epoch": 28.894323144104803,
+ "grad_norm": 0.014393828809261322,
+ "learning_rate": 0.0006,
+ "loss": 4.198468208312988,
+ "step": 2080
+ },
+ {
+ "epoch": 28.90829694323144,
+ "grad_norm": 0.01580170728266239,
+ "learning_rate": 0.0006,
+ "loss": 4.224969863891602,
+ "step": 2081
+ },
+ {
+ "epoch": 28.92227074235808,
+ "grad_norm": 0.01484165620058775,
+ "learning_rate": 0.0006,
+ "loss": 4.12236213684082,
+ "step": 2082
+ },
+ {
+ "epoch": 28.936244541484715,
+ "grad_norm": 0.01497623696923256,
+ "learning_rate": 0.0006,
+ "loss": 4.24141788482666,
+ "step": 2083
+ },
+ {
+ "epoch": 28.950218340611354,
+ "grad_norm": 0.015080046840012074,
+ "learning_rate": 0.0006,
+ "loss": 4.284151077270508,
+ "step": 2084
+ },
+ {
+ "epoch": 28.964192139737992,
+ "grad_norm": 0.016207212582230568,
+ "learning_rate": 0.0006,
+ "loss": 4.298693656921387,
+ "step": 2085
+ },
+ {
+ "epoch": 28.978165938864628,
+ "grad_norm": 0.015474417246878147,
+ "learning_rate": 0.0006,
+ "loss": 4.2412261962890625,
+ "step": 2086
+ },
+ {
+ "epoch": 28.992139737991266,
+ "grad_norm": 0.01515351701527834,
+ "learning_rate": 0.0006,
+ "loss": 4.3054423332214355,
+ "step": 2087
+ },
+ {
+ "epoch": 29.0,
+ "grad_norm": 0.016411345452070236,
+ "learning_rate": 0.0006,
+ "loss": 4.104830741882324,
+ "step": 2088
+ },
+ {
+ "epoch": 29.0,
+ "eval_loss": 4.661567687988281,
+ "eval_runtime": 57.1244,
+ "eval_samples_per_second": 42.749,
+ "eval_steps_per_second": 1.348,
+ "step": 2088
+ },
+ {
+ "epoch": 29.01397379912664,
+ "grad_norm": 0.01665342226624489,
+ "learning_rate": 0.0006,
+ "loss": 4.311391830444336,
+ "step": 2089
+ },
+ {
+ "epoch": 29.027947598253274,
+ "grad_norm": 0.018053214997053146,
+ "learning_rate": 0.0006,
+ "loss": 4.135597229003906,
+ "step": 2090
+ },
+ {
+ "epoch": 29.041921397379912,
+ "grad_norm": 0.018213748931884766,
+ "learning_rate": 0.0006,
+ "loss": 4.303299903869629,
+ "step": 2091
+ },
+ {
+ "epoch": 29.05589519650655,
+ "grad_norm": 0.018460242077708244,
+ "learning_rate": 0.0006,
+ "loss": 4.20850944519043,
+ "step": 2092
+ },
+ {
+ "epoch": 29.069868995633186,
+ "grad_norm": 0.018365809693932533,
+ "learning_rate": 0.0006,
+ "loss": 4.275847434997559,
+ "step": 2093
+ },
+ {
+ "epoch": 29.083842794759825,
+ "grad_norm": 0.018346186727285385,
+ "learning_rate": 0.0006,
+ "loss": 4.202091217041016,
+ "step": 2094
+ },
+ {
+ "epoch": 29.097816593886463,
+ "grad_norm": 0.017920587211847305,
+ "learning_rate": 0.0006,
+ "loss": 4.182092666625977,
+ "step": 2095
+ },
+ {
+ "epoch": 29.111790393013102,
+ "grad_norm": 0.01812003180384636,
+ "learning_rate": 0.0006,
+ "loss": 4.259721755981445,
+ "step": 2096
+ },
+ {
+ "epoch": 29.125764192139737,
+ "grad_norm": 0.01700986735522747,
+ "learning_rate": 0.0006,
+ "loss": 4.067573070526123,
+ "step": 2097
+ },
+ {
+ "epoch": 29.139737991266376,
+ "grad_norm": 0.018069760873913765,
+ "learning_rate": 0.0006,
+ "loss": 4.234400749206543,
+ "step": 2098
+ },
+ {
+ "epoch": 29.153711790393015,
+ "grad_norm": 0.01905428245663643,
+ "learning_rate": 0.0006,
+ "loss": 4.16609001159668,
+ "step": 2099
+ },
+ {
+ "epoch": 29.16768558951965,
+ "grad_norm": 0.018711242824792862,
+ "learning_rate": 0.0006,
+ "loss": 4.249088764190674,
+ "step": 2100
+ },
+ {
+ "epoch": 29.18165938864629,
+ "grad_norm": 0.018721306696534157,
+ "learning_rate": 0.0006,
+ "loss": 4.14540958404541,
+ "step": 2101
+ },
+ {
+ "epoch": 29.195633187772927,
+ "grad_norm": 0.020139874890446663,
+ "learning_rate": 0.0006,
+ "loss": 4.049248695373535,
+ "step": 2102
+ },
+ {
+ "epoch": 29.209606986899562,
+ "grad_norm": 0.020849574357271194,
+ "learning_rate": 0.0006,
+ "loss": 4.235224723815918,
+ "step": 2103
+ },
+ {
+ "epoch": 29.2235807860262,
+ "grad_norm": 0.018693549558520317,
+ "learning_rate": 0.0006,
+ "loss": 4.232961177825928,
+ "step": 2104
+ },
+ {
+ "epoch": 29.23755458515284,
+ "grad_norm": 0.017889857292175293,
+ "learning_rate": 0.0006,
+ "loss": 4.171422004699707,
+ "step": 2105
+ },
+ {
+ "epoch": 29.251528384279474,
+ "grad_norm": 0.016690224409103394,
+ "learning_rate": 0.0006,
+ "loss": 4.244717597961426,
+ "step": 2106
+ },
+ {
+ "epoch": 29.265502183406113,
+ "grad_norm": 0.01708616502583027,
+ "learning_rate": 0.0006,
+ "loss": 4.125068664550781,
+ "step": 2107
+ },
+ {
+ "epoch": 29.27947598253275,
+ "grad_norm": 0.017928237095475197,
+ "learning_rate": 0.0006,
+ "loss": 4.208292007446289,
+ "step": 2108
+ },
+ {
+ "epoch": 29.29344978165939,
+ "grad_norm": 0.018250394612550735,
+ "learning_rate": 0.0006,
+ "loss": 4.118630886077881,
+ "step": 2109
+ },
+ {
+ "epoch": 29.307423580786025,
+ "grad_norm": 0.020160246640443802,
+ "learning_rate": 0.0006,
+ "loss": 4.206025123596191,
+ "step": 2110
+ },
+ {
+ "epoch": 29.321397379912664,
+ "grad_norm": 0.020045453682541847,
+ "learning_rate": 0.0006,
+ "loss": 4.2758684158325195,
+ "step": 2111
+ },
+ {
+ "epoch": 29.335371179039303,
+ "grad_norm": 0.019556907936930656,
+ "learning_rate": 0.0006,
+ "loss": 4.188272953033447,
+ "step": 2112
+ },
+ {
+ "epoch": 29.349344978165938,
+ "grad_norm": 0.019185200333595276,
+ "learning_rate": 0.0006,
+ "loss": 4.17451286315918,
+ "step": 2113
+ },
+ {
+ "epoch": 29.363318777292577,
+ "grad_norm": 0.01816629432141781,
+ "learning_rate": 0.0006,
+ "loss": 4.165246963500977,
+ "step": 2114
+ },
+ {
+ "epoch": 29.377292576419215,
+ "grad_norm": 0.01865146867930889,
+ "learning_rate": 0.0006,
+ "loss": 4.148184776306152,
+ "step": 2115
+ },
+ {
+ "epoch": 29.39126637554585,
+ "grad_norm": 0.018338464200496674,
+ "learning_rate": 0.0006,
+ "loss": 4.288936614990234,
+ "step": 2116
+ },
+ {
+ "epoch": 29.40524017467249,
+ "grad_norm": 0.018986187875270844,
+ "learning_rate": 0.0006,
+ "loss": 4.183379650115967,
+ "step": 2117
+ },
+ {
+ "epoch": 29.419213973799128,
+ "grad_norm": 0.018409637734293938,
+ "learning_rate": 0.0006,
+ "loss": 4.057574272155762,
+ "step": 2118
+ },
+ {
+ "epoch": 29.433187772925763,
+ "grad_norm": 0.016664542257785797,
+ "learning_rate": 0.0006,
+ "loss": 4.18517541885376,
+ "step": 2119
+ },
+ {
+ "epoch": 29.4471615720524,
+ "grad_norm": 0.016486089676618576,
+ "learning_rate": 0.0006,
+ "loss": 4.179367542266846,
+ "step": 2120
+ },
+ {
+ "epoch": 29.46113537117904,
+ "grad_norm": 0.01787523552775383,
+ "learning_rate": 0.0006,
+ "loss": 4.184126377105713,
+ "step": 2121
+ },
+ {
+ "epoch": 29.475109170305675,
+ "grad_norm": 0.016574613749980927,
+ "learning_rate": 0.0006,
+ "loss": 4.290170192718506,
+ "step": 2122
+ },
+ {
+ "epoch": 29.489082969432314,
+ "grad_norm": 0.01643376611173153,
+ "learning_rate": 0.0006,
+ "loss": 4.052913665771484,
+ "step": 2123
+ },
+ {
+ "epoch": 29.503056768558952,
+ "grad_norm": 0.016503969207406044,
+ "learning_rate": 0.0006,
+ "loss": 4.335184097290039,
+ "step": 2124
+ },
+ {
+ "epoch": 29.51703056768559,
+ "grad_norm": 0.017596479505300522,
+ "learning_rate": 0.0006,
+ "loss": 4.165539741516113,
+ "step": 2125
+ },
+ {
+ "epoch": 29.531004366812226,
+ "grad_norm": 0.017468582838773727,
+ "learning_rate": 0.0006,
+ "loss": 4.315964698791504,
+ "step": 2126
+ },
+ {
+ "epoch": 29.544978165938865,
+ "grad_norm": 0.01724618673324585,
+ "learning_rate": 0.0006,
+ "loss": 4.243681907653809,
+ "step": 2127
+ },
+ {
+ "epoch": 29.558951965065503,
+ "grad_norm": 0.018384616822004318,
+ "learning_rate": 0.0006,
+ "loss": 4.112447261810303,
+ "step": 2128
+ },
+ {
+ "epoch": 29.57292576419214,
+ "grad_norm": 0.01824074238538742,
+ "learning_rate": 0.0006,
+ "loss": 4.236065864562988,
+ "step": 2129
+ },
+ {
+ "epoch": 29.586899563318777,
+ "grad_norm": 0.01625971868634224,
+ "learning_rate": 0.0006,
+ "loss": 4.18143892288208,
+ "step": 2130
+ },
+ {
+ "epoch": 29.600873362445416,
+ "grad_norm": 0.015537494793534279,
+ "learning_rate": 0.0006,
+ "loss": 4.225147247314453,
+ "step": 2131
+ },
+ {
+ "epoch": 29.61484716157205,
+ "grad_norm": 0.015818974003195763,
+ "learning_rate": 0.0006,
+ "loss": 4.3100175857543945,
+ "step": 2132
+ },
+ {
+ "epoch": 29.62882096069869,
+ "grad_norm": 0.015909671783447266,
+ "learning_rate": 0.0006,
+ "loss": 4.169775009155273,
+ "step": 2133
+ },
+ {
+ "epoch": 29.64279475982533,
+ "grad_norm": 0.01769658550620079,
+ "learning_rate": 0.0006,
+ "loss": 4.278904914855957,
+ "step": 2134
+ },
+ {
+ "epoch": 29.656768558951963,
+ "grad_norm": 0.015641704201698303,
+ "learning_rate": 0.0006,
+ "loss": 4.2426228523254395,
+ "step": 2135
+ },
+ {
+ "epoch": 29.670742358078602,
+ "grad_norm": 0.01565658301115036,
+ "learning_rate": 0.0006,
+ "loss": 4.232687950134277,
+ "step": 2136
+ },
+ {
+ "epoch": 29.68471615720524,
+ "grad_norm": 0.016485940665006638,
+ "learning_rate": 0.0006,
+ "loss": 4.209839820861816,
+ "step": 2137
+ },
+ {
+ "epoch": 29.69868995633188,
+ "grad_norm": 0.017622729763388634,
+ "learning_rate": 0.0006,
+ "loss": 4.196017265319824,
+ "step": 2138
+ },
+ {
+ "epoch": 29.712663755458514,
+ "grad_norm": 0.018320564180612564,
+ "learning_rate": 0.0006,
+ "loss": 4.225683689117432,
+ "step": 2139
+ },
+ {
+ "epoch": 29.726637554585153,
+ "grad_norm": 0.018140794709324837,
+ "learning_rate": 0.0006,
+ "loss": 4.311944961547852,
+ "step": 2140
+ },
+ {
+ "epoch": 29.74061135371179,
+ "grad_norm": 0.016227100044488907,
+ "learning_rate": 0.0006,
+ "loss": 4.230257511138916,
+ "step": 2141
+ },
+ {
+ "epoch": 29.754585152838427,
+ "grad_norm": 0.015759488567709923,
+ "learning_rate": 0.0006,
+ "loss": 4.119174003601074,
+ "step": 2142
+ },
+ {
+ "epoch": 29.768558951965066,
+ "grad_norm": 0.01757766678929329,
+ "learning_rate": 0.0006,
+ "loss": 4.165826797485352,
+ "step": 2143
+ },
+ {
+ "epoch": 29.782532751091704,
+ "grad_norm": 0.019661923870444298,
+ "learning_rate": 0.0006,
+ "loss": 4.392203330993652,
+ "step": 2144
+ },
+ {
+ "epoch": 29.79650655021834,
+ "grad_norm": 0.019478755071759224,
+ "learning_rate": 0.0006,
+ "loss": 4.348371982574463,
+ "step": 2145
+ },
+ {
+ "epoch": 29.810480349344978,
+ "grad_norm": 0.019149569794535637,
+ "learning_rate": 0.0006,
+ "loss": 4.34104061126709,
+ "step": 2146
+ },
+ {
+ "epoch": 29.824454148471617,
+ "grad_norm": 0.017707521095871925,
+ "learning_rate": 0.0006,
+ "loss": 4.228226661682129,
+ "step": 2147
+ },
+ {
+ "epoch": 29.83842794759825,
+ "grad_norm": 0.015615378506481647,
+ "learning_rate": 0.0006,
+ "loss": 4.175541877746582,
+ "step": 2148
+ },
+ {
+ "epoch": 29.85240174672489,
+ "grad_norm": 0.016111129894852638,
+ "learning_rate": 0.0006,
+ "loss": 4.267936706542969,
+ "step": 2149
+ },
+ {
+ "epoch": 29.86637554585153,
+ "grad_norm": 0.01528779324144125,
+ "learning_rate": 0.0006,
+ "loss": 4.186014175415039,
+ "step": 2150
+ },
+ {
+ "epoch": 29.880349344978168,
+ "grad_norm": 0.014925646595656872,
+ "learning_rate": 0.0006,
+ "loss": 4.174860000610352,
+ "step": 2151
+ },
+ {
+ "epoch": 29.894323144104803,
+ "grad_norm": 0.016917673870921135,
+ "learning_rate": 0.0006,
+ "loss": 4.33908748626709,
+ "step": 2152
+ },
+ {
+ "epoch": 29.90829694323144,
+ "grad_norm": 0.016007075086236,
+ "learning_rate": 0.0006,
+ "loss": 4.266141414642334,
+ "step": 2153
+ },
+ {
+ "epoch": 29.92227074235808,
+ "grad_norm": 0.016178956255316734,
+ "learning_rate": 0.0006,
+ "loss": 4.271215915679932,
+ "step": 2154
+ },
+ {
+ "epoch": 29.936244541484715,
+ "grad_norm": 0.016180871054530144,
+ "learning_rate": 0.0006,
+ "loss": 4.278801441192627,
+ "step": 2155
+ },
+ {
+ "epoch": 29.950218340611354,
+ "grad_norm": 0.015335503034293652,
+ "learning_rate": 0.0006,
+ "loss": 4.194978713989258,
+ "step": 2156
+ },
+ {
+ "epoch": 29.964192139737992,
+ "grad_norm": 0.016580764204263687,
+ "learning_rate": 0.0006,
+ "loss": 4.290809631347656,
+ "step": 2157
+ },
+ {
+ "epoch": 29.978165938864628,
+ "grad_norm": 0.01623128168284893,
+ "learning_rate": 0.0006,
+ "loss": 4.205451488494873,
+ "step": 2158
+ },
+ {
+ "epoch": 29.992139737991266,
+ "grad_norm": 0.014904248528182507,
+ "learning_rate": 0.0006,
+ "loss": 4.24924373626709,
+ "step": 2159
+ },
+ {
+ "epoch": 30.0,
+ "grad_norm": 0.017089299857616425,
+ "learning_rate": 0.0006,
+ "loss": 4.222072601318359,
+ "step": 2160
+ },
+ {
+ "epoch": 30.0,
+ "eval_loss": 4.585411071777344,
+ "eval_runtime": 57.1164,
+ "eval_samples_per_second": 42.755,
+ "eval_steps_per_second": 1.348,
+ "step": 2160
+ },
+ {
+ "epoch": 30.01397379912664,
+ "grad_norm": 0.017226383090019226,
+ "learning_rate": 0.0006,
+ "loss": 4.19467830657959,
+ "step": 2161
+ },
+ {
+ "epoch": 30.027947598253274,
+ "grad_norm": 0.01635066606104374,
+ "learning_rate": 0.0006,
+ "loss": 4.176875591278076,
+ "step": 2162
+ },
+ {
+ "epoch": 30.041921397379912,
+ "grad_norm": 0.01735362410545349,
+ "learning_rate": 0.0006,
+ "loss": 4.13667106628418,
+ "step": 2163
+ },
+ {
+ "epoch": 30.05589519650655,
+ "grad_norm": 0.017047109082341194,
+ "learning_rate": 0.0006,
+ "loss": 4.136329650878906,
+ "step": 2164
+ },
+ {
+ "epoch": 30.069868995633186,
+ "grad_norm": 0.018116053193807602,
+ "learning_rate": 0.0006,
+ "loss": 4.228706359863281,
+ "step": 2165
+ },
+ {
+ "epoch": 30.083842794759825,
+ "grad_norm": 0.01791389286518097,
+ "learning_rate": 0.0006,
+ "loss": 4.1597747802734375,
+ "step": 2166
+ },
+ {
+ "epoch": 30.097816593886463,
+ "grad_norm": 0.017793502658605576,
+ "learning_rate": 0.0006,
+ "loss": 4.0807976722717285,
+ "step": 2167
+ },
+ {
+ "epoch": 30.111790393013102,
+ "grad_norm": 0.01827503927052021,
+ "learning_rate": 0.0006,
+ "loss": 4.170646667480469,
+ "step": 2168
+ },
+ {
+ "epoch": 30.125764192139737,
+ "grad_norm": 0.017228564247488976,
+ "learning_rate": 0.0006,
+ "loss": 4.146934509277344,
+ "step": 2169
+ },
+ {
+ "epoch": 30.139737991266376,
+ "grad_norm": 0.01757437363266945,
+ "learning_rate": 0.0006,
+ "loss": 4.1859130859375,
+ "step": 2170
+ },
+ {
+ "epoch": 30.153711790393015,
+ "grad_norm": 0.01701093092560768,
+ "learning_rate": 0.0006,
+ "loss": 4.126981258392334,
+ "step": 2171
+ },
+ {
+ "epoch": 30.16768558951965,
+ "grad_norm": 0.015220489352941513,
+ "learning_rate": 0.0006,
+ "loss": 4.164018630981445,
+ "step": 2172
+ },
+ {
+ "epoch": 30.18165938864629,
+ "grad_norm": 0.016254238784313202,
+ "learning_rate": 0.0006,
+ "loss": 4.258275508880615,
+ "step": 2173
+ },
+ {
+ "epoch": 30.195633187772927,
+ "grad_norm": 0.016609614714980125,
+ "learning_rate": 0.0006,
+ "loss": 4.222233772277832,
+ "step": 2174
+ },
+ {
+ "epoch": 30.209606986899562,
+ "grad_norm": 0.016531318426132202,
+ "learning_rate": 0.0006,
+ "loss": 4.144489288330078,
+ "step": 2175
+ },
+ {
+ "epoch": 30.2235807860262,
+ "grad_norm": 0.01567668654024601,
+ "learning_rate": 0.0006,
+ "loss": 4.181916236877441,
+ "step": 2176
+ },
+ {
+ "epoch": 30.23755458515284,
+ "grad_norm": 0.017195681110024452,
+ "learning_rate": 0.0006,
+ "loss": 4.189693927764893,
+ "step": 2177
+ },
+ {
+ "epoch": 30.251528384279474,
+ "grad_norm": 0.016882948577404022,
+ "learning_rate": 0.0006,
+ "loss": 4.21604061126709,
+ "step": 2178
+ },
+ {
+ "epoch": 30.265502183406113,
+ "grad_norm": 0.016261622309684753,
+ "learning_rate": 0.0006,
+ "loss": 4.227639198303223,
+ "step": 2179
+ },
+ {
+ "epoch": 30.27947598253275,
+ "grad_norm": 0.016411812976002693,
+ "learning_rate": 0.0006,
+ "loss": 4.130362510681152,
+ "step": 2180
+ },
+ {
+ "epoch": 30.29344978165939,
+ "grad_norm": 0.01732019893825054,
+ "learning_rate": 0.0006,
+ "loss": 4.230443000793457,
+ "step": 2181
+ },
+ {
+ "epoch": 30.307423580786025,
+ "grad_norm": 0.01825905591249466,
+ "learning_rate": 0.0006,
+ "loss": 4.175817489624023,
+ "step": 2182
+ },
+ {
+ "epoch": 30.321397379912664,
+ "grad_norm": 0.01887713558971882,
+ "learning_rate": 0.0006,
+ "loss": 4.169020652770996,
+ "step": 2183
+ },
+ {
+ "epoch": 30.335371179039303,
+ "grad_norm": 0.019391661509871483,
+ "learning_rate": 0.0006,
+ "loss": 4.176318645477295,
+ "step": 2184
+ },
+ {
+ "epoch": 30.349344978165938,
+ "grad_norm": 0.020840002223849297,
+ "learning_rate": 0.0006,
+ "loss": 4.174228668212891,
+ "step": 2185
+ },
+ {
+ "epoch": 30.363318777292577,
+ "grad_norm": 0.020193178206682205,
+ "learning_rate": 0.0006,
+ "loss": 4.127919673919678,
+ "step": 2186
+ },
+ {
+ "epoch": 30.377292576419215,
+ "grad_norm": 0.018449561670422554,
+ "learning_rate": 0.0006,
+ "loss": 4.229487895965576,
+ "step": 2187
+ },
+ {
+ "epoch": 30.39126637554585,
+ "grad_norm": 0.01711983233690262,
+ "learning_rate": 0.0006,
+ "loss": 4.201651573181152,
+ "step": 2188
+ },
+ {
+ "epoch": 30.40524017467249,
+ "grad_norm": 0.017293395474553108,
+ "learning_rate": 0.0006,
+ "loss": 4.168704032897949,
+ "step": 2189
+ },
+ {
+ "epoch": 30.419213973799128,
+ "grad_norm": 0.018526358529925346,
+ "learning_rate": 0.0006,
+ "loss": 4.171864032745361,
+ "step": 2190
+ },
+ {
+ "epoch": 30.433187772925763,
+ "grad_norm": 0.019287291914224625,
+ "learning_rate": 0.0006,
+ "loss": 4.039534568786621,
+ "step": 2191
+ },
+ {
+ "epoch": 30.4471615720524,
+ "grad_norm": 0.016902444884181023,
+ "learning_rate": 0.0006,
+ "loss": 4.282787799835205,
+ "step": 2192
+ },
+ {
+ "epoch": 30.46113537117904,
+ "grad_norm": 0.01656750962138176,
+ "learning_rate": 0.0006,
+ "loss": 4.213098526000977,
+ "step": 2193
+ },
+ {
+ "epoch": 30.475109170305675,
+ "grad_norm": 0.016805065795779228,
+ "learning_rate": 0.0006,
+ "loss": 4.1583476066589355,
+ "step": 2194
+ },
+ {
+ "epoch": 30.489082969432314,
+ "grad_norm": 0.017166638746857643,
+ "learning_rate": 0.0006,
+ "loss": 4.188634395599365,
+ "step": 2195
+ },
+ {
+ "epoch": 30.503056768558952,
+ "grad_norm": 0.01592106744647026,
+ "learning_rate": 0.0006,
+ "loss": 4.234043121337891,
+ "step": 2196
+ },
+ {
+ "epoch": 30.51703056768559,
+ "grad_norm": 0.015270989388227463,
+ "learning_rate": 0.0006,
+ "loss": 4.301420211791992,
+ "step": 2197
+ },
+ {
+ "epoch": 30.531004366812226,
+ "grad_norm": 0.015194724313914776,
+ "learning_rate": 0.0006,
+ "loss": 4.18703031539917,
+ "step": 2198
+ },
+ {
+ "epoch": 30.544978165938865,
+ "grad_norm": 0.015587746165692806,
+ "learning_rate": 0.0006,
+ "loss": 4.138928413391113,
+ "step": 2199
+ },
+ {
+ "epoch": 30.558951965065503,
+ "grad_norm": 0.016029570251703262,
+ "learning_rate": 0.0006,
+ "loss": 4.290863990783691,
+ "step": 2200
+ },
+ {
+ "epoch": 30.57292576419214,
+ "grad_norm": 0.015651309862732887,
+ "learning_rate": 0.0006,
+ "loss": 4.272062301635742,
+ "step": 2201
+ },
+ {
+ "epoch": 30.586899563318777,
+ "grad_norm": 0.015992306172847748,
+ "learning_rate": 0.0006,
+ "loss": 4.2487406730651855,
+ "step": 2202
+ },
+ {
+ "epoch": 30.600873362445416,
+ "grad_norm": 0.014969157986342907,
+ "learning_rate": 0.0006,
+ "loss": 4.115347862243652,
+ "step": 2203
+ },
+ {
+ "epoch": 30.61484716157205,
+ "grad_norm": 0.014639027416706085,
+ "learning_rate": 0.0006,
+ "loss": 4.2207489013671875,
+ "step": 2204
+ },
+ {
+ "epoch": 30.62882096069869,
+ "grad_norm": 0.015059413388371468,
+ "learning_rate": 0.0006,
+ "loss": 4.274044990539551,
+ "step": 2205
+ },
+ {
+ "epoch": 30.64279475982533,
+ "grad_norm": 0.0166641678661108,
+ "learning_rate": 0.0006,
+ "loss": 4.191373348236084,
+ "step": 2206
+ },
+ {
+ "epoch": 30.656768558951963,
+ "grad_norm": 0.017806345596909523,
+ "learning_rate": 0.0006,
+ "loss": 4.139800548553467,
+ "step": 2207
+ },
+ {
+ "epoch": 30.670742358078602,
+ "grad_norm": 0.018712421879172325,
+ "learning_rate": 0.0006,
+ "loss": 4.035965442657471,
+ "step": 2208
+ },
+ {
+ "epoch": 30.68471615720524,
+ "grad_norm": 0.018457243219017982,
+ "learning_rate": 0.0006,
+ "loss": 4.2579240798950195,
+ "step": 2209
+ },
+ {
+ "epoch": 30.69868995633188,
+ "grad_norm": 0.01920999586582184,
+ "learning_rate": 0.0006,
+ "loss": 4.131112098693848,
+ "step": 2210
+ },
+ {
+ "epoch": 30.712663755458514,
+ "grad_norm": 0.01729944534599781,
+ "learning_rate": 0.0006,
+ "loss": 4.254059314727783,
+ "step": 2211
+ },
+ {
+ "epoch": 30.726637554585153,
+ "grad_norm": 0.016914231702685356,
+ "learning_rate": 0.0006,
+ "loss": 4.233428955078125,
+ "step": 2212
+ },
+ {
+ "epoch": 30.74061135371179,
+ "grad_norm": 0.01913098618388176,
+ "learning_rate": 0.0006,
+ "loss": 4.23525333404541,
+ "step": 2213
+ },
+ {
+ "epoch": 30.754585152838427,
+ "grad_norm": 0.019150281324982643,
+ "learning_rate": 0.0006,
+ "loss": 4.213695049285889,
+ "step": 2214
+ },
+ {
+ "epoch": 30.768558951965066,
+ "grad_norm": 0.018947452306747437,
+ "learning_rate": 0.0006,
+ "loss": 4.151697158813477,
+ "step": 2215
+ },
+ {
+ "epoch": 30.782532751091704,
+ "grad_norm": 0.016058508306741714,
+ "learning_rate": 0.0006,
+ "loss": 4.220280170440674,
+ "step": 2216
+ },
+ {
+ "epoch": 30.79650655021834,
+ "grad_norm": 0.017488619312644005,
+ "learning_rate": 0.0006,
+ "loss": 4.154428482055664,
+ "step": 2217
+ },
+ {
+ "epoch": 30.810480349344978,
+ "grad_norm": 0.017355090007185936,
+ "learning_rate": 0.0006,
+ "loss": 4.256086349487305,
+ "step": 2218
+ },
+ {
+ "epoch": 30.824454148471617,
+ "grad_norm": 0.016636032611131668,
+ "learning_rate": 0.0006,
+ "loss": 4.227564334869385,
+ "step": 2219
+ },
+ {
+ "epoch": 30.83842794759825,
+ "grad_norm": 0.016273891553282738,
+ "learning_rate": 0.0006,
+ "loss": 4.200466156005859,
+ "step": 2220
+ },
+ {
+ "epoch": 30.85240174672489,
+ "grad_norm": 0.016939911991357803,
+ "learning_rate": 0.0006,
+ "loss": 4.1571221351623535,
+ "step": 2221
+ },
+ {
+ "epoch": 30.86637554585153,
+ "grad_norm": 0.018332522362470627,
+ "learning_rate": 0.0006,
+ "loss": 4.228398323059082,
+ "step": 2222
+ },
+ {
+ "epoch": 30.880349344978168,
+ "grad_norm": 0.019508114084601402,
+ "learning_rate": 0.0006,
+ "loss": 4.222960472106934,
+ "step": 2223
+ },
+ {
+ "epoch": 30.894323144104803,
+ "grad_norm": 0.018440278246998787,
+ "learning_rate": 0.0006,
+ "loss": 4.261051177978516,
+ "step": 2224
+ },
+ {
+ "epoch": 30.90829694323144,
+ "grad_norm": 0.018113229423761368,
+ "learning_rate": 0.0006,
+ "loss": 4.109235763549805,
+ "step": 2225
+ },
+ {
+ "epoch": 30.92227074235808,
+ "grad_norm": 0.015392083674669266,
+ "learning_rate": 0.0006,
+ "loss": 4.21950101852417,
+ "step": 2226
+ },
+ {
+ "epoch": 30.936244541484715,
+ "grad_norm": 0.01633290946483612,
+ "learning_rate": 0.0006,
+ "loss": 4.176053047180176,
+ "step": 2227
+ },
+ {
+ "epoch": 30.950218340611354,
+ "grad_norm": 0.016236383467912674,
+ "learning_rate": 0.0006,
+ "loss": 4.314309120178223,
+ "step": 2228
+ },
+ {
+ "epoch": 30.964192139737992,
+ "grad_norm": 0.015284373424947262,
+ "learning_rate": 0.0006,
+ "loss": 4.303133487701416,
+ "step": 2229
+ },
+ {
+ "epoch": 30.978165938864628,
+ "grad_norm": 0.015982897952198982,
+ "learning_rate": 0.0006,
+ "loss": 4.25739860534668,
+ "step": 2230
+ },
+ {
+ "epoch": 30.992139737991266,
+ "grad_norm": 0.016682198271155357,
+ "learning_rate": 0.0006,
+ "loss": 4.292210578918457,
+ "step": 2231
+ },
+ {
+ "epoch": 31.0,
+ "grad_norm": 0.017521467059850693,
+ "learning_rate": 0.0006,
+ "loss": 4.170637130737305,
+ "step": 2232
+ },
+ {
+ "epoch": 31.0,
+ "eval_loss": 4.642786979675293,
+ "eval_runtime": 57.029,
+ "eval_samples_per_second": 42.82,
+ "eval_steps_per_second": 1.35,
+ "step": 2232
+ },
+ {
+ "epoch": 31.01397379912664,
+ "grad_norm": 0.016607852652668953,
+ "learning_rate": 0.0006,
+ "loss": 4.106224536895752,
+ "step": 2233
+ },
+ {
+ "epoch": 31.027947598253274,
+ "grad_norm": 0.018448904156684875,
+ "learning_rate": 0.0006,
+ "loss": 4.224459648132324,
+ "step": 2234
+ },
+ {
+ "epoch": 31.041921397379912,
+ "grad_norm": 0.020481228828430176,
+ "learning_rate": 0.0006,
+ "loss": 4.305548667907715,
+ "step": 2235
+ },
+ {
+ "epoch": 31.05589519650655,
+ "grad_norm": 0.01999707892537117,
+ "learning_rate": 0.0006,
+ "loss": 4.253663539886475,
+ "step": 2236
+ },
+ {
+ "epoch": 31.069868995633186,
+ "grad_norm": 0.020428303629159927,
+ "learning_rate": 0.0006,
+ "loss": 4.166954040527344,
+ "step": 2237
+ },
+ {
+ "epoch": 31.083842794759825,
+ "grad_norm": 0.022608213126659393,
+ "learning_rate": 0.0006,
+ "loss": 4.172819137573242,
+ "step": 2238
+ },
+ {
+ "epoch": 31.097816593886463,
+ "grad_norm": 0.024669520556926727,
+ "learning_rate": 0.0006,
+ "loss": 4.106586456298828,
+ "step": 2239
+ },
+ {
+ "epoch": 31.111790393013102,
+ "grad_norm": 0.02247351035475731,
+ "learning_rate": 0.0006,
+ "loss": 4.160271644592285,
+ "step": 2240
+ },
+ {
+ "epoch": 31.125764192139737,
+ "grad_norm": 0.0207810141146183,
+ "learning_rate": 0.0006,
+ "loss": 4.272491931915283,
+ "step": 2241
+ },
+ {
+ "epoch": 31.139737991266376,
+ "grad_norm": 0.021576499566435814,
+ "learning_rate": 0.0006,
+ "loss": 4.171474456787109,
+ "step": 2242
+ },
+ {
+ "epoch": 31.153711790393015,
+ "grad_norm": 0.02337106689810753,
+ "learning_rate": 0.0006,
+ "loss": 4.094712734222412,
+ "step": 2243
+ },
+ {
+ "epoch": 31.16768558951965,
+ "grad_norm": 0.02219875156879425,
+ "learning_rate": 0.0006,
+ "loss": 4.1800994873046875,
+ "step": 2244
+ },
+ {
+ "epoch": 31.18165938864629,
+ "grad_norm": 0.022680504247546196,
+ "learning_rate": 0.0006,
+ "loss": 4.2054362297058105,
+ "step": 2245
+ },
+ {
+ "epoch": 31.195633187772927,
+ "grad_norm": 0.0208734143525362,
+ "learning_rate": 0.0006,
+ "loss": 4.06068229675293,
+ "step": 2246
+ },
+ {
+ "epoch": 31.209606986899562,
+ "grad_norm": 0.01811066083610058,
+ "learning_rate": 0.0006,
+ "loss": 4.2465128898620605,
+ "step": 2247
+ },
+ {
+ "epoch": 31.2235807860262,
+ "grad_norm": 0.018571637570858,
+ "learning_rate": 0.0006,
+ "loss": 4.191955089569092,
+ "step": 2248
+ },
+ {
+ "epoch": 31.23755458515284,
+ "grad_norm": 0.018598807975649834,
+ "learning_rate": 0.0006,
+ "loss": 4.200872421264648,
+ "step": 2249
+ },
+ {
+ "epoch": 31.251528384279474,
+ "grad_norm": 0.017577625811100006,
+ "learning_rate": 0.0006,
+ "loss": 4.148700714111328,
+ "step": 2250
+ },
+ {
+ "epoch": 31.265502183406113,
+ "grad_norm": 0.01817980408668518,
+ "learning_rate": 0.0006,
+ "loss": 4.256978988647461,
+ "step": 2251
+ },
+ {
+ "epoch": 31.27947598253275,
+ "grad_norm": 0.017983486875891685,
+ "learning_rate": 0.0006,
+ "loss": 4.175005912780762,
+ "step": 2252
+ },
+ {
+ "epoch": 31.29344978165939,
+ "grad_norm": 0.016751961782574654,
+ "learning_rate": 0.0006,
+ "loss": 4.237518310546875,
+ "step": 2253
+ },
+ {
+ "epoch": 31.307423580786025,
+ "grad_norm": 0.016407202929258347,
+ "learning_rate": 0.0006,
+ "loss": 4.197269439697266,
+ "step": 2254
+ },
+ {
+ "epoch": 31.321397379912664,
+ "grad_norm": 0.01578645594418049,
+ "learning_rate": 0.0006,
+ "loss": 4.179751396179199,
+ "step": 2255
+ },
+ {
+ "epoch": 31.335371179039303,
+ "grad_norm": 0.016001226380467415,
+ "learning_rate": 0.0006,
+ "loss": 4.08821964263916,
+ "step": 2256
+ },
+ {
+ "epoch": 31.349344978165938,
+ "grad_norm": 0.015122811309993267,
+ "learning_rate": 0.0006,
+ "loss": 4.047828674316406,
+ "step": 2257
+ },
+ {
+ "epoch": 31.363318777292577,
+ "grad_norm": 0.015399965457618237,
+ "learning_rate": 0.0006,
+ "loss": 4.1506524085998535,
+ "step": 2258
+ },
+ {
+ "epoch": 31.377292576419215,
+ "grad_norm": 0.01534605398774147,
+ "learning_rate": 0.0006,
+ "loss": 4.27952766418457,
+ "step": 2259
+ },
+ {
+ "epoch": 31.39126637554585,
+ "grad_norm": 0.015461381524801254,
+ "learning_rate": 0.0006,
+ "loss": 4.27102518081665,
+ "step": 2260
+ },
+ {
+ "epoch": 31.40524017467249,
+ "grad_norm": 0.016769153997302055,
+ "learning_rate": 0.0006,
+ "loss": 4.191592216491699,
+ "step": 2261
+ },
+ {
+ "epoch": 31.419213973799128,
+ "grad_norm": 0.016557959839701653,
+ "learning_rate": 0.0006,
+ "loss": 4.159705638885498,
+ "step": 2262
+ },
+ {
+ "epoch": 31.433187772925763,
+ "grad_norm": 0.01659569889307022,
+ "learning_rate": 0.0006,
+ "loss": 4.2136664390563965,
+ "step": 2263
+ },
+ {
+ "epoch": 31.4471615720524,
+ "grad_norm": 0.017101937904953957,
+ "learning_rate": 0.0006,
+ "loss": 4.2716498374938965,
+ "step": 2264
+ },
+ {
+ "epoch": 31.46113537117904,
+ "grad_norm": 0.01722985878586769,
+ "learning_rate": 0.0006,
+ "loss": 4.300534248352051,
+ "step": 2265
+ },
+ {
+ "epoch": 31.475109170305675,
+ "grad_norm": 0.016800185665488243,
+ "learning_rate": 0.0006,
+ "loss": 4.113240718841553,
+ "step": 2266
+ },
+ {
+ "epoch": 31.489082969432314,
+ "grad_norm": 0.01603875495493412,
+ "learning_rate": 0.0006,
+ "loss": 4.157105445861816,
+ "step": 2267
+ },
+ {
+ "epoch": 31.503056768558952,
+ "grad_norm": 0.015903932973742485,
+ "learning_rate": 0.0006,
+ "loss": 4.233669757843018,
+ "step": 2268
+ },
+ {
+ "epoch": 31.51703056768559,
+ "grad_norm": 0.0160983856767416,
+ "learning_rate": 0.0006,
+ "loss": 4.147704124450684,
+ "step": 2269
+ },
+ {
+ "epoch": 31.531004366812226,
+ "grad_norm": 0.01604730263352394,
+ "learning_rate": 0.0006,
+ "loss": 4.1465911865234375,
+ "step": 2270
+ },
+ {
+ "epoch": 31.544978165938865,
+ "grad_norm": 0.016205785796046257,
+ "learning_rate": 0.0006,
+ "loss": 4.106704235076904,
+ "step": 2271
+ },
+ {
+ "epoch": 31.558951965065503,
+ "grad_norm": 0.017008036375045776,
+ "learning_rate": 0.0006,
+ "loss": 4.191000938415527,
+ "step": 2272
+ },
+ {
+ "epoch": 31.57292576419214,
+ "grad_norm": 0.01699635200202465,
+ "learning_rate": 0.0006,
+ "loss": 4.187042236328125,
+ "step": 2273
+ },
+ {
+ "epoch": 31.586899563318777,
+ "grad_norm": 0.01692967116832733,
+ "learning_rate": 0.0006,
+ "loss": 4.208805561065674,
+ "step": 2274
+ },
+ {
+ "epoch": 31.600873362445416,
+ "grad_norm": 0.017608756199479103,
+ "learning_rate": 0.0006,
+ "loss": 4.200512886047363,
+ "step": 2275
+ },
+ {
+ "epoch": 31.61484716157205,
+ "grad_norm": 0.016590990126132965,
+ "learning_rate": 0.0006,
+ "loss": 4.194705009460449,
+ "step": 2276
+ },
+ {
+ "epoch": 31.62882096069869,
+ "grad_norm": 0.0170235987752676,
+ "learning_rate": 0.0006,
+ "loss": 4.210326194763184,
+ "step": 2277
+ },
+ {
+ "epoch": 31.64279475982533,
+ "grad_norm": 0.01604801043868065,
+ "learning_rate": 0.0006,
+ "loss": 4.196024417877197,
+ "step": 2278
+ },
+ {
+ "epoch": 31.656768558951963,
+ "grad_norm": 0.015694202855229378,
+ "learning_rate": 0.0006,
+ "loss": 4.178236484527588,
+ "step": 2279
+ },
+ {
+ "epoch": 31.670742358078602,
+ "grad_norm": 0.015473959036171436,
+ "learning_rate": 0.0006,
+ "loss": 4.259339809417725,
+ "step": 2280
+ },
+ {
+ "epoch": 31.68471615720524,
+ "grad_norm": 0.014689362607896328,
+ "learning_rate": 0.0006,
+ "loss": 3.9702324867248535,
+ "step": 2281
+ },
+ {
+ "epoch": 31.69868995633188,
+ "grad_norm": 0.015245123766362667,
+ "learning_rate": 0.0006,
+ "loss": 4.197025775909424,
+ "step": 2282
+ },
+ {
+ "epoch": 31.712663755458514,
+ "grad_norm": 0.016148347407579422,
+ "learning_rate": 0.0006,
+ "loss": 4.231856346130371,
+ "step": 2283
+ },
+ {
+ "epoch": 31.726637554585153,
+ "grad_norm": 0.015169818885624409,
+ "learning_rate": 0.0006,
+ "loss": 4.230169296264648,
+ "step": 2284
+ },
+ {
+ "epoch": 31.74061135371179,
+ "grad_norm": 0.014843013137578964,
+ "learning_rate": 0.0006,
+ "loss": 4.195633411407471,
+ "step": 2285
+ },
+ {
+ "epoch": 31.754585152838427,
+ "grad_norm": 0.015229340642690659,
+ "learning_rate": 0.0006,
+ "loss": 4.231280326843262,
+ "step": 2286
+ },
+ {
+ "epoch": 31.768558951965066,
+ "grad_norm": 0.01551708485931158,
+ "learning_rate": 0.0006,
+ "loss": 4.17064094543457,
+ "step": 2287
+ },
+ {
+ "epoch": 31.782532751091704,
+ "grad_norm": 0.015863575041294098,
+ "learning_rate": 0.0006,
+ "loss": 4.1579484939575195,
+ "step": 2288
+ },
+ {
+ "epoch": 31.79650655021834,
+ "grad_norm": 0.015118198469281197,
+ "learning_rate": 0.0006,
+ "loss": 4.1626787185668945,
+ "step": 2289
+ },
+ {
+ "epoch": 31.810480349344978,
+ "grad_norm": 0.015220776200294495,
+ "learning_rate": 0.0006,
+ "loss": 4.135746955871582,
+ "step": 2290
+ },
+ {
+ "epoch": 31.824454148471617,
+ "grad_norm": 0.014529784210026264,
+ "learning_rate": 0.0006,
+ "loss": 4.164931297302246,
+ "step": 2291
+ },
+ {
+ "epoch": 31.83842794759825,
+ "grad_norm": 0.015533296391367912,
+ "learning_rate": 0.0006,
+ "loss": 4.112464904785156,
+ "step": 2292
+ },
+ {
+ "epoch": 31.85240174672489,
+ "grad_norm": 0.016562335193157196,
+ "learning_rate": 0.0006,
+ "loss": 4.164898872375488,
+ "step": 2293
+ },
+ {
+ "epoch": 31.86637554585153,
+ "grad_norm": 0.016258113086223602,
+ "learning_rate": 0.0006,
+ "loss": 4.1979265213012695,
+ "step": 2294
+ },
+ {
+ "epoch": 31.880349344978168,
+ "grad_norm": 0.01558110024780035,
+ "learning_rate": 0.0006,
+ "loss": 4.174675941467285,
+ "step": 2295
+ },
+ {
+ "epoch": 31.894323144104803,
+ "grad_norm": 0.016412515193223953,
+ "learning_rate": 0.0006,
+ "loss": 4.217909336090088,
+ "step": 2296
+ },
+ {
+ "epoch": 31.90829694323144,
+ "grad_norm": 0.015118442475795746,
+ "learning_rate": 0.0006,
+ "loss": 4.137143611907959,
+ "step": 2297
+ },
+ {
+ "epoch": 31.92227074235808,
+ "grad_norm": 0.01506161317229271,
+ "learning_rate": 0.0006,
+ "loss": 4.196122169494629,
+ "step": 2298
+ },
+ {
+ "epoch": 31.936244541484715,
+ "grad_norm": 0.017165271565318108,
+ "learning_rate": 0.0006,
+ "loss": 4.14834451675415,
+ "step": 2299
+ },
+ {
+ "epoch": 31.950218340611354,
+ "grad_norm": 0.01696440577507019,
+ "learning_rate": 0.0006,
+ "loss": 4.218868732452393,
+ "step": 2300
+ },
+ {
+ "epoch": 31.964192139737992,
+ "grad_norm": 0.017349394038319588,
+ "learning_rate": 0.0006,
+ "loss": 4.2754011154174805,
+ "step": 2301
+ },
+ {
+ "epoch": 31.978165938864628,
+ "grad_norm": 0.018387921154499054,
+ "learning_rate": 0.0006,
+ "loss": 4.282623291015625,
+ "step": 2302
+ },
+ {
+ "epoch": 31.992139737991266,
+ "grad_norm": 0.018837926909327507,
+ "learning_rate": 0.0006,
+ "loss": 4.185425281524658,
+ "step": 2303
+ },
+ {
+ "epoch": 32.0,
+ "grad_norm": 0.018998464569449425,
+ "learning_rate": 0.0006,
+ "loss": 4.151291847229004,
+ "step": 2304
+ },
+ {
+ "epoch": 32.0,
+ "eval_loss": 4.522614002227783,
+ "eval_runtime": 57.0106,
+ "eval_samples_per_second": 42.834,
+ "eval_steps_per_second": 1.351,
+ "step": 2304
+ },
+ {
+ "epoch": 32.01397379912664,
+ "grad_norm": 0.019036833196878433,
+ "learning_rate": 0.0006,
+ "loss": 4.183202266693115,
+ "step": 2305
+ },
+ {
+ "epoch": 32.02794759825328,
+ "grad_norm": 0.02207397110760212,
+ "learning_rate": 0.0006,
+ "loss": 4.17875862121582,
+ "step": 2306
+ },
+ {
+ "epoch": 32.041921397379916,
+ "grad_norm": 0.022008279338479042,
+ "learning_rate": 0.0006,
+ "loss": 4.08042049407959,
+ "step": 2307
+ },
+ {
+ "epoch": 32.05589519650655,
+ "grad_norm": 0.022407017648220062,
+ "learning_rate": 0.0006,
+ "loss": 4.112189769744873,
+ "step": 2308
+ },
+ {
+ "epoch": 32.069868995633186,
+ "grad_norm": 0.021599747240543365,
+ "learning_rate": 0.0006,
+ "loss": 4.126455307006836,
+ "step": 2309
+ },
+ {
+ "epoch": 32.083842794759825,
+ "grad_norm": 0.020340966060757637,
+ "learning_rate": 0.0006,
+ "loss": 4.080538272857666,
+ "step": 2310
+ },
+ {
+ "epoch": 32.09781659388646,
+ "grad_norm": 0.020956892520189285,
+ "learning_rate": 0.0006,
+ "loss": 4.171943664550781,
+ "step": 2311
+ },
+ {
+ "epoch": 32.1117903930131,
+ "grad_norm": 0.019966667518019676,
+ "learning_rate": 0.0006,
+ "loss": 4.144007682800293,
+ "step": 2312
+ },
+ {
+ "epoch": 32.12576419213974,
+ "grad_norm": 0.018256952986121178,
+ "learning_rate": 0.0006,
+ "loss": 4.208107948303223,
+ "step": 2313
+ },
+ {
+ "epoch": 32.13973799126637,
+ "grad_norm": 0.017289170995354652,
+ "learning_rate": 0.0006,
+ "loss": 4.182024955749512,
+ "step": 2314
+ },
+ {
+ "epoch": 32.15371179039301,
+ "grad_norm": 0.017656700685620308,
+ "learning_rate": 0.0006,
+ "loss": 4.068373680114746,
+ "step": 2315
+ },
+ {
+ "epoch": 32.16768558951965,
+ "grad_norm": 0.017903361469507217,
+ "learning_rate": 0.0006,
+ "loss": 4.2345428466796875,
+ "step": 2316
+ },
+ {
+ "epoch": 32.18165938864629,
+ "grad_norm": 0.01785109005868435,
+ "learning_rate": 0.0006,
+ "loss": 4.219776153564453,
+ "step": 2317
+ },
+ {
+ "epoch": 32.19563318777293,
+ "grad_norm": 0.01887257769703865,
+ "learning_rate": 0.0006,
+ "loss": 4.139206409454346,
+ "step": 2318
+ },
+ {
+ "epoch": 32.209606986899566,
+ "grad_norm": 0.01924685761332512,
+ "learning_rate": 0.0006,
+ "loss": 4.121052265167236,
+ "step": 2319
+ },
+ {
+ "epoch": 32.223580786026204,
+ "grad_norm": 0.020574072375893593,
+ "learning_rate": 0.0006,
+ "loss": 4.13262414932251,
+ "step": 2320
+ },
+ {
+ "epoch": 32.237554585152836,
+ "grad_norm": 0.023849064484238625,
+ "learning_rate": 0.0006,
+ "loss": 4.239119529724121,
+ "step": 2321
+ },
+ {
+ "epoch": 32.251528384279474,
+ "grad_norm": 0.026239361613988876,
+ "learning_rate": 0.0006,
+ "loss": 4.248993873596191,
+ "step": 2322
+ },
+ {
+ "epoch": 32.26550218340611,
+ "grad_norm": 0.023980243131518364,
+ "learning_rate": 0.0006,
+ "loss": 4.159852504730225,
+ "step": 2323
+ },
+ {
+ "epoch": 32.27947598253275,
+ "grad_norm": 0.020442405715584755,
+ "learning_rate": 0.0006,
+ "loss": 4.133546829223633,
+ "step": 2324
+ },
+ {
+ "epoch": 32.29344978165939,
+ "grad_norm": 0.020661186426877975,
+ "learning_rate": 0.0006,
+ "loss": 4.078243732452393,
+ "step": 2325
+ },
+ {
+ "epoch": 32.30742358078603,
+ "grad_norm": 0.02046387456357479,
+ "learning_rate": 0.0006,
+ "loss": 4.172012805938721,
+ "step": 2326
+ },
+ {
+ "epoch": 32.32139737991266,
+ "grad_norm": 0.02089597098529339,
+ "learning_rate": 0.0006,
+ "loss": 4.0604705810546875,
+ "step": 2327
+ },
+ {
+ "epoch": 32.3353711790393,
+ "grad_norm": 0.01875591278076172,
+ "learning_rate": 0.0006,
+ "loss": 4.125911235809326,
+ "step": 2328
+ },
+ {
+ "epoch": 32.34934497816594,
+ "grad_norm": 0.018606796860694885,
+ "learning_rate": 0.0006,
+ "loss": 4.097362518310547,
+ "step": 2329
+ },
+ {
+ "epoch": 32.36331877729258,
+ "grad_norm": 0.01766626164317131,
+ "learning_rate": 0.0006,
+ "loss": 4.118616580963135,
+ "step": 2330
+ },
+ {
+ "epoch": 32.377292576419215,
+ "grad_norm": 0.017692960798740387,
+ "learning_rate": 0.0006,
+ "loss": 4.1457672119140625,
+ "step": 2331
+ },
+ {
+ "epoch": 32.391266375545854,
+ "grad_norm": 0.017619704827666283,
+ "learning_rate": 0.0006,
+ "loss": 4.175961494445801,
+ "step": 2332
+ },
+ {
+ "epoch": 32.40524017467249,
+ "grad_norm": 0.01784413494169712,
+ "learning_rate": 0.0006,
+ "loss": 4.051291465759277,
+ "step": 2333
+ },
+ {
+ "epoch": 32.419213973799124,
+ "grad_norm": 0.018508171662688255,
+ "learning_rate": 0.0006,
+ "loss": 4.0638556480407715,
+ "step": 2334
+ },
+ {
+ "epoch": 32.43318777292576,
+ "grad_norm": 0.01731436885893345,
+ "learning_rate": 0.0006,
+ "loss": 4.1805419921875,
+ "step": 2335
+ },
+ {
+ "epoch": 32.4471615720524,
+ "grad_norm": 0.015886498615145683,
+ "learning_rate": 0.0006,
+ "loss": 4.151026725769043,
+ "step": 2336
+ },
+ {
+ "epoch": 32.46113537117904,
+ "grad_norm": 0.016583820804953575,
+ "learning_rate": 0.0006,
+ "loss": 4.203604698181152,
+ "step": 2337
+ },
+ {
+ "epoch": 32.47510917030568,
+ "grad_norm": 0.0176097322255373,
+ "learning_rate": 0.0006,
+ "loss": 4.135159015655518,
+ "step": 2338
+ },
+ {
+ "epoch": 32.48908296943232,
+ "grad_norm": 0.016657903790473938,
+ "learning_rate": 0.0006,
+ "loss": 4.126681804656982,
+ "step": 2339
+ },
+ {
+ "epoch": 32.50305676855895,
+ "grad_norm": 0.016866130754351616,
+ "learning_rate": 0.0006,
+ "loss": 4.162341117858887,
+ "step": 2340
+ },
+ {
+ "epoch": 32.51703056768559,
+ "grad_norm": 0.016286680474877357,
+ "learning_rate": 0.0006,
+ "loss": 4.261801242828369,
+ "step": 2341
+ },
+ {
+ "epoch": 32.531004366812226,
+ "grad_norm": 0.01481365505605936,
+ "learning_rate": 0.0006,
+ "loss": 4.110016822814941,
+ "step": 2342
+ },
+ {
+ "epoch": 32.544978165938865,
+ "grad_norm": 0.016605013981461525,
+ "learning_rate": 0.0006,
+ "loss": 4.230024814605713,
+ "step": 2343
+ },
+ {
+ "epoch": 32.5589519650655,
+ "grad_norm": 0.01645175740122795,
+ "learning_rate": 0.0006,
+ "loss": 4.174612045288086,
+ "step": 2344
+ },
+ {
+ "epoch": 32.57292576419214,
+ "grad_norm": 0.014725026674568653,
+ "learning_rate": 0.0006,
+ "loss": 4.066371440887451,
+ "step": 2345
+ },
+ {
+ "epoch": 32.58689956331878,
+ "grad_norm": 0.015003837645053864,
+ "learning_rate": 0.0006,
+ "loss": 4.193343639373779,
+ "step": 2346
+ },
+ {
+ "epoch": 32.60087336244541,
+ "grad_norm": 0.015346302650868893,
+ "learning_rate": 0.0006,
+ "loss": 4.210659980773926,
+ "step": 2347
+ },
+ {
+ "epoch": 32.61484716157205,
+ "grad_norm": 0.015135962516069412,
+ "learning_rate": 0.0006,
+ "loss": 4.204618453979492,
+ "step": 2348
+ },
+ {
+ "epoch": 32.62882096069869,
+ "grad_norm": 0.01601956970989704,
+ "learning_rate": 0.0006,
+ "loss": 4.145564556121826,
+ "step": 2349
+ },
+ {
+ "epoch": 32.64279475982533,
+ "grad_norm": 0.01517851185053587,
+ "learning_rate": 0.0006,
+ "loss": 4.051673889160156,
+ "step": 2350
+ },
+ {
+ "epoch": 32.65676855895197,
+ "grad_norm": 0.014894292689859867,
+ "learning_rate": 0.0006,
+ "loss": 4.184818267822266,
+ "step": 2351
+ },
+ {
+ "epoch": 32.670742358078606,
+ "grad_norm": 0.015720834955573082,
+ "learning_rate": 0.0006,
+ "loss": 4.220036506652832,
+ "step": 2352
+ },
+ {
+ "epoch": 32.68471615720524,
+ "grad_norm": 0.01573433168232441,
+ "learning_rate": 0.0006,
+ "loss": 4.076773166656494,
+ "step": 2353
+ },
+ {
+ "epoch": 32.698689956331876,
+ "grad_norm": 0.015828052535653114,
+ "learning_rate": 0.0006,
+ "loss": 4.116329669952393,
+ "step": 2354
+ },
+ {
+ "epoch": 32.712663755458514,
+ "grad_norm": 0.015462259761989117,
+ "learning_rate": 0.0006,
+ "loss": 4.1155242919921875,
+ "step": 2355
+ },
+ {
+ "epoch": 32.72663755458515,
+ "grad_norm": 0.015763849020004272,
+ "learning_rate": 0.0006,
+ "loss": 4.169687747955322,
+ "step": 2356
+ },
+ {
+ "epoch": 32.74061135371179,
+ "grad_norm": 0.014935145154595375,
+ "learning_rate": 0.0006,
+ "loss": 4.2090911865234375,
+ "step": 2357
+ },
+ {
+ "epoch": 32.75458515283843,
+ "grad_norm": 0.015119184739887714,
+ "learning_rate": 0.0006,
+ "loss": 4.131838798522949,
+ "step": 2358
+ },
+ {
+ "epoch": 32.76855895196506,
+ "grad_norm": 0.014833525754511356,
+ "learning_rate": 0.0006,
+ "loss": 4.17618465423584,
+ "step": 2359
+ },
+ {
+ "epoch": 32.7825327510917,
+ "grad_norm": 0.01588359661400318,
+ "learning_rate": 0.0006,
+ "loss": 4.163723945617676,
+ "step": 2360
+ },
+ {
+ "epoch": 32.79650655021834,
+ "grad_norm": 0.0168188214302063,
+ "learning_rate": 0.0006,
+ "loss": 4.178189277648926,
+ "step": 2361
+ },
+ {
+ "epoch": 32.81048034934498,
+ "grad_norm": 0.01726095750927925,
+ "learning_rate": 0.0006,
+ "loss": 4.144223690032959,
+ "step": 2362
+ },
+ {
+ "epoch": 32.82445414847162,
+ "grad_norm": 0.01701410301029682,
+ "learning_rate": 0.0006,
+ "loss": 4.209896087646484,
+ "step": 2363
+ },
+ {
+ "epoch": 32.838427947598255,
+ "grad_norm": 0.01682264916598797,
+ "learning_rate": 0.0006,
+ "loss": 4.26223087310791,
+ "step": 2364
+ },
+ {
+ "epoch": 32.852401746724894,
+ "grad_norm": 0.015597456134855747,
+ "learning_rate": 0.0006,
+ "loss": 4.292664051055908,
+ "step": 2365
+ },
+ {
+ "epoch": 32.866375545851525,
+ "grad_norm": 0.015258579514920712,
+ "learning_rate": 0.0006,
+ "loss": 4.165006637573242,
+ "step": 2366
+ },
+ {
+ "epoch": 32.880349344978164,
+ "grad_norm": 0.01662200503051281,
+ "learning_rate": 0.0006,
+ "loss": 4.309786319732666,
+ "step": 2367
+ },
+ {
+ "epoch": 32.8943231441048,
+ "grad_norm": 0.015691161155700684,
+ "learning_rate": 0.0006,
+ "loss": 4.167202949523926,
+ "step": 2368
+ },
+ {
+ "epoch": 32.90829694323144,
+ "grad_norm": 0.014870637096464634,
+ "learning_rate": 0.0006,
+ "loss": 4.142415523529053,
+ "step": 2369
+ },
+ {
+ "epoch": 32.92227074235808,
+ "grad_norm": 0.01567252166569233,
+ "learning_rate": 0.0006,
+ "loss": 4.2312092781066895,
+ "step": 2370
+ },
+ {
+ "epoch": 32.93624454148472,
+ "grad_norm": 0.014948616735637188,
+ "learning_rate": 0.0006,
+ "loss": 4.1514129638671875,
+ "step": 2371
+ },
+ {
+ "epoch": 32.95021834061135,
+ "grad_norm": 0.014654500409960747,
+ "learning_rate": 0.0006,
+ "loss": 4.205010890960693,
+ "step": 2372
+ },
+ {
+ "epoch": 32.96419213973799,
+ "grad_norm": 0.015059148892760277,
+ "learning_rate": 0.0006,
+ "loss": 4.148405075073242,
+ "step": 2373
+ },
+ {
+ "epoch": 32.97816593886463,
+ "grad_norm": 0.015026576817035675,
+ "learning_rate": 0.0006,
+ "loss": 4.182535171508789,
+ "step": 2374
+ },
+ {
+ "epoch": 32.992139737991266,
+ "grad_norm": 0.016394320875406265,
+ "learning_rate": 0.0006,
+ "loss": 4.129051208496094,
+ "step": 2375
+ },
+ {
+ "epoch": 33.0,
+ "grad_norm": 0.01867043413221836,
+ "learning_rate": 0.0006,
+ "loss": 4.185556411743164,
+ "step": 2376
+ },
+ {
+ "epoch": 33.0,
+ "eval_loss": 4.640405654907227,
+ "eval_runtime": 56.5861,
+ "eval_samples_per_second": 43.155,
+ "eval_steps_per_second": 1.361,
+ "step": 2376
+ },
+ {
+ "epoch": 33.01397379912664,
+ "grad_norm": 0.018626727163791656,
+ "learning_rate": 0.0006,
+ "loss": 4.143031597137451,
+ "step": 2377
+ },
+ {
+ "epoch": 33.02794759825328,
+ "grad_norm": 0.019283363595604897,
+ "learning_rate": 0.0006,
+ "loss": 4.117845058441162,
+ "step": 2378
+ },
+ {
+ "epoch": 33.041921397379916,
+ "grad_norm": 0.018185822293162346,
+ "learning_rate": 0.0006,
+ "loss": 4.076862812042236,
+ "step": 2379
+ },
+ {
+ "epoch": 33.05589519650655,
+ "grad_norm": 0.019411196932196617,
+ "learning_rate": 0.0006,
+ "loss": 4.124156475067139,
+ "step": 2380
+ },
+ {
+ "epoch": 33.069868995633186,
+ "grad_norm": 0.019248466938734055,
+ "learning_rate": 0.0006,
+ "loss": 4.201735019683838,
+ "step": 2381
+ },
+ {
+ "epoch": 33.083842794759825,
+ "grad_norm": 0.018856720998883247,
+ "learning_rate": 0.0006,
+ "loss": 4.00296688079834,
+ "step": 2382
+ },
+ {
+ "epoch": 33.09781659388646,
+ "grad_norm": 0.018554294481873512,
+ "learning_rate": 0.0006,
+ "loss": 4.270577430725098,
+ "step": 2383
+ },
+ {
+ "epoch": 33.1117903930131,
+ "grad_norm": 0.019946224987506866,
+ "learning_rate": 0.0006,
+ "loss": 4.119589328765869,
+ "step": 2384
+ },
+ {
+ "epoch": 33.12576419213974,
+ "grad_norm": 0.021395670250058174,
+ "learning_rate": 0.0006,
+ "loss": 4.128847122192383,
+ "step": 2385
+ },
+ {
+ "epoch": 33.13973799126637,
+ "grad_norm": 0.02009674906730652,
+ "learning_rate": 0.0006,
+ "loss": 4.183228969573975,
+ "step": 2386
+ },
+ {
+ "epoch": 33.15371179039301,
+ "grad_norm": 0.01820223033428192,
+ "learning_rate": 0.0006,
+ "loss": 4.131601810455322,
+ "step": 2387
+ },
+ {
+ "epoch": 33.16768558951965,
+ "grad_norm": 0.018653156235814095,
+ "learning_rate": 0.0006,
+ "loss": 4.116430282592773,
+ "step": 2388
+ },
+ {
+ "epoch": 33.18165938864629,
+ "grad_norm": 0.016828063875436783,
+ "learning_rate": 0.0006,
+ "loss": 4.054663181304932,
+ "step": 2389
+ },
+ {
+ "epoch": 33.19563318777293,
+ "grad_norm": 0.017025409266352654,
+ "learning_rate": 0.0006,
+ "loss": 4.110044956207275,
+ "step": 2390
+ },
+ {
+ "epoch": 33.209606986899566,
+ "grad_norm": 0.016608452424407005,
+ "learning_rate": 0.0006,
+ "loss": 4.102668285369873,
+ "step": 2391
+ },
+ {
+ "epoch": 33.223580786026204,
+ "grad_norm": 0.01743577979505062,
+ "learning_rate": 0.0006,
+ "loss": 4.166791915893555,
+ "step": 2392
+ },
+ {
+ "epoch": 33.237554585152836,
+ "grad_norm": 0.017253734171390533,
+ "learning_rate": 0.0006,
+ "loss": 4.217832088470459,
+ "step": 2393
+ },
+ {
+ "epoch": 33.251528384279474,
+ "grad_norm": 0.017937595024704933,
+ "learning_rate": 0.0006,
+ "loss": 4.2173967361450195,
+ "step": 2394
+ },
+ {
+ "epoch": 33.26550218340611,
+ "grad_norm": 0.01985993981361389,
+ "learning_rate": 0.0006,
+ "loss": 4.17849588394165,
+ "step": 2395
+ },
+ {
+ "epoch": 33.27947598253275,
+ "grad_norm": 0.019243355840444565,
+ "learning_rate": 0.0006,
+ "loss": 4.086523056030273,
+ "step": 2396
+ },
+ {
+ "epoch": 33.29344978165939,
+ "grad_norm": 0.01924416422843933,
+ "learning_rate": 0.0006,
+ "loss": 4.128574371337891,
+ "step": 2397
+ },
+ {
+ "epoch": 33.30742358078603,
+ "grad_norm": 0.01983758620917797,
+ "learning_rate": 0.0006,
+ "loss": 4.061581611633301,
+ "step": 2398
+ },
+ {
+ "epoch": 33.32139737991266,
+ "grad_norm": 0.01930462196469307,
+ "learning_rate": 0.0006,
+ "loss": 4.091731071472168,
+ "step": 2399
+ },
+ {
+ "epoch": 33.3353711790393,
+ "grad_norm": 0.017988424748182297,
+ "learning_rate": 0.0006,
+ "loss": 4.155599117279053,
+ "step": 2400
+ },
+ {
+ "epoch": 33.34934497816594,
+ "grad_norm": 0.015194767154753208,
+ "learning_rate": 0.0006,
+ "loss": 4.144134044647217,
+ "step": 2401
+ },
+ {
+ "epoch": 33.36331877729258,
+ "grad_norm": 0.01536081824451685,
+ "learning_rate": 0.0006,
+ "loss": 4.127272129058838,
+ "step": 2402
+ },
+ {
+ "epoch": 33.377292576419215,
+ "grad_norm": 0.015431297942996025,
+ "learning_rate": 0.0006,
+ "loss": 4.3213019371032715,
+ "step": 2403
+ },
+ {
+ "epoch": 33.391266375545854,
+ "grad_norm": 0.015740280970931053,
+ "learning_rate": 0.0006,
+ "loss": 4.121467590332031,
+ "step": 2404
+ },
+ {
+ "epoch": 33.40524017467249,
+ "grad_norm": 0.015379384160041809,
+ "learning_rate": 0.0006,
+ "loss": 4.095216751098633,
+ "step": 2405
+ },
+ {
+ "epoch": 33.419213973799124,
+ "grad_norm": 0.01528073474764824,
+ "learning_rate": 0.0006,
+ "loss": 4.13047981262207,
+ "step": 2406
+ },
+ {
+ "epoch": 33.43318777292576,
+ "grad_norm": 0.015739979222416878,
+ "learning_rate": 0.0006,
+ "loss": 4.2533135414123535,
+ "step": 2407
+ },
+ {
+ "epoch": 33.4471615720524,
+ "grad_norm": 0.01519764307886362,
+ "learning_rate": 0.0006,
+ "loss": 4.131782054901123,
+ "step": 2408
+ },
+ {
+ "epoch": 33.46113537117904,
+ "grad_norm": 0.015425996854901314,
+ "learning_rate": 0.0006,
+ "loss": 4.02811336517334,
+ "step": 2409
+ },
+ {
+ "epoch": 33.47510917030568,
+ "grad_norm": 0.015378830023109913,
+ "learning_rate": 0.0006,
+ "loss": 4.08652925491333,
+ "step": 2410
+ },
+ {
+ "epoch": 33.48908296943232,
+ "grad_norm": 0.01473075058311224,
+ "learning_rate": 0.0006,
+ "loss": 4.214364528656006,
+ "step": 2411
+ },
+ {
+ "epoch": 33.50305676855895,
+ "grad_norm": 0.015141637064516544,
+ "learning_rate": 0.0006,
+ "loss": 4.155712127685547,
+ "step": 2412
+ },
+ {
+ "epoch": 33.51703056768559,
+ "grad_norm": 0.01641593500971794,
+ "learning_rate": 0.0006,
+ "loss": 4.097907066345215,
+ "step": 2413
+ },
+ {
+ "epoch": 33.531004366812226,
+ "grad_norm": 0.015182029455900192,
+ "learning_rate": 0.0006,
+ "loss": 4.238499641418457,
+ "step": 2414
+ },
+ {
+ "epoch": 33.544978165938865,
+ "grad_norm": 0.01643732376396656,
+ "learning_rate": 0.0006,
+ "loss": 4.0661492347717285,
+ "step": 2415
+ },
+ {
+ "epoch": 33.5589519650655,
+ "grad_norm": 0.015614228323101997,
+ "learning_rate": 0.0006,
+ "loss": 4.120795249938965,
+ "step": 2416
+ },
+ {
+ "epoch": 33.57292576419214,
+ "grad_norm": 0.016066187992691994,
+ "learning_rate": 0.0006,
+ "loss": 4.096465110778809,
+ "step": 2417
+ },
+ {
+ "epoch": 33.58689956331878,
+ "grad_norm": 0.017214342951774597,
+ "learning_rate": 0.0006,
+ "loss": 4.242628574371338,
+ "step": 2418
+ },
+ {
+ "epoch": 33.60087336244541,
+ "grad_norm": 0.01787586510181427,
+ "learning_rate": 0.0006,
+ "loss": 4.149589538574219,
+ "step": 2419
+ },
+ {
+ "epoch": 33.61484716157205,
+ "grad_norm": 0.019455241039395332,
+ "learning_rate": 0.0006,
+ "loss": 4.267796039581299,
+ "step": 2420
+ },
+ {
+ "epoch": 33.62882096069869,
+ "grad_norm": 0.01984674669802189,
+ "learning_rate": 0.0006,
+ "loss": 4.175220966339111,
+ "step": 2421
+ },
+ {
+ "epoch": 33.64279475982533,
+ "grad_norm": 0.021933717653155327,
+ "learning_rate": 0.0006,
+ "loss": 4.106647491455078,
+ "step": 2422
+ },
+ {
+ "epoch": 33.65676855895197,
+ "grad_norm": 0.02123589813709259,
+ "learning_rate": 0.0006,
+ "loss": 4.204996109008789,
+ "step": 2423
+ },
+ {
+ "epoch": 33.670742358078606,
+ "grad_norm": 0.019814589992165565,
+ "learning_rate": 0.0006,
+ "loss": 4.186171531677246,
+ "step": 2424
+ },
+ {
+ "epoch": 33.68471615720524,
+ "grad_norm": 0.02059728465974331,
+ "learning_rate": 0.0006,
+ "loss": 4.170509338378906,
+ "step": 2425
+ },
+ {
+ "epoch": 33.698689956331876,
+ "grad_norm": 0.018483608961105347,
+ "learning_rate": 0.0006,
+ "loss": 4.085585594177246,
+ "step": 2426
+ },
+ {
+ "epoch": 33.712663755458514,
+ "grad_norm": 0.018403852358460426,
+ "learning_rate": 0.0006,
+ "loss": 4.2551774978637695,
+ "step": 2427
+ },
+ {
+ "epoch": 33.72663755458515,
+ "grad_norm": 0.018653830513358116,
+ "learning_rate": 0.0006,
+ "loss": 4.091766357421875,
+ "step": 2428
+ },
+ {
+ "epoch": 33.74061135371179,
+ "grad_norm": 0.017692334949970245,
+ "learning_rate": 0.0006,
+ "loss": 4.100838661193848,
+ "step": 2429
+ },
+ {
+ "epoch": 33.75458515283843,
+ "grad_norm": 0.016029126942157745,
+ "learning_rate": 0.0006,
+ "loss": 4.1530680656433105,
+ "step": 2430
+ },
+ {
+ "epoch": 33.76855895196506,
+ "grad_norm": 0.016095684841275215,
+ "learning_rate": 0.0006,
+ "loss": 4.20544958114624,
+ "step": 2431
+ },
+ {
+ "epoch": 33.7825327510917,
+ "grad_norm": 0.016375798732042313,
+ "learning_rate": 0.0006,
+ "loss": 4.199177265167236,
+ "step": 2432
+ },
+ {
+ "epoch": 33.79650655021834,
+ "grad_norm": 0.01590786688029766,
+ "learning_rate": 0.0006,
+ "loss": 4.200799465179443,
+ "step": 2433
+ },
+ {
+ "epoch": 33.81048034934498,
+ "grad_norm": 0.014896097593009472,
+ "learning_rate": 0.0006,
+ "loss": 4.185850143432617,
+ "step": 2434
+ },
+ {
+ "epoch": 33.82445414847162,
+ "grad_norm": 0.01438513956964016,
+ "learning_rate": 0.0006,
+ "loss": 4.214962005615234,
+ "step": 2435
+ },
+ {
+ "epoch": 33.838427947598255,
+ "grad_norm": 0.014459339901804924,
+ "learning_rate": 0.0006,
+ "loss": 4.196972370147705,
+ "step": 2436
+ },
+ {
+ "epoch": 33.852401746724894,
+ "grad_norm": 0.014782169833779335,
+ "learning_rate": 0.0006,
+ "loss": 4.0885910987854,
+ "step": 2437
+ },
+ {
+ "epoch": 33.866375545851525,
+ "grad_norm": 0.01520370040088892,
+ "learning_rate": 0.0006,
+ "loss": 4.0323262214660645,
+ "step": 2438
+ },
+ {
+ "epoch": 33.880349344978164,
+ "grad_norm": 0.014056282117962837,
+ "learning_rate": 0.0006,
+ "loss": 4.1386518478393555,
+ "step": 2439
+ },
+ {
+ "epoch": 33.8943231441048,
+ "grad_norm": 0.013567990623414516,
+ "learning_rate": 0.0006,
+ "loss": 4.095754623413086,
+ "step": 2440
+ },
+ {
+ "epoch": 33.90829694323144,
+ "grad_norm": 0.013306716457009315,
+ "learning_rate": 0.0006,
+ "loss": 4.1623077392578125,
+ "step": 2441
+ },
+ {
+ "epoch": 33.92227074235808,
+ "grad_norm": 0.014047150500118732,
+ "learning_rate": 0.0006,
+ "loss": 4.142541885375977,
+ "step": 2442
+ },
+ {
+ "epoch": 33.93624454148472,
+ "grad_norm": 0.01377617847174406,
+ "learning_rate": 0.0006,
+ "loss": 4.268551826477051,
+ "step": 2443
+ },
+ {
+ "epoch": 33.95021834061135,
+ "grad_norm": 0.014671621844172478,
+ "learning_rate": 0.0006,
+ "loss": 4.198196887969971,
+ "step": 2444
+ },
+ {
+ "epoch": 33.96419213973799,
+ "grad_norm": 0.014806744642555714,
+ "learning_rate": 0.0006,
+ "loss": 4.057899475097656,
+ "step": 2445
+ },
+ {
+ "epoch": 33.97816593886463,
+ "grad_norm": 0.015734272077679634,
+ "learning_rate": 0.0006,
+ "loss": 4.163505554199219,
+ "step": 2446
+ },
+ {
+ "epoch": 33.992139737991266,
+ "grad_norm": 0.015250146389007568,
+ "learning_rate": 0.0006,
+ "loss": 4.076294898986816,
+ "step": 2447
+ },
+ {
+ "epoch": 34.0,
+ "grad_norm": 0.016733702272176743,
+ "learning_rate": 0.0006,
+ "loss": 4.138420104980469,
+ "step": 2448
+ },
+ {
+ "epoch": 34.0,
+ "eval_loss": 4.542503356933594,
+ "eval_runtime": 56.983,
+ "eval_samples_per_second": 42.855,
+ "eval_steps_per_second": 1.351,
+ "step": 2448
+ },
+ {
+ "epoch": 34.01397379912664,
+ "grad_norm": 0.015866613015532494,
+ "learning_rate": 0.0006,
+ "loss": 4.032081127166748,
+ "step": 2449
+ },
+ {
+ "epoch": 34.02794759825328,
+ "grad_norm": 0.01946570910513401,
+ "learning_rate": 0.0006,
+ "loss": 4.122320175170898,
+ "step": 2450
+ },
+ {
+ "epoch": 34.041921397379916,
+ "grad_norm": 0.020265938714146614,
+ "learning_rate": 0.0006,
+ "loss": 4.06460428237915,
+ "step": 2451
+ },
+ {
+ "epoch": 34.05589519650655,
+ "grad_norm": 0.018887171521782875,
+ "learning_rate": 0.0006,
+ "loss": 4.096906661987305,
+ "step": 2452
+ },
+ {
+ "epoch": 34.069868995633186,
+ "grad_norm": 0.017304692417383194,
+ "learning_rate": 0.0006,
+ "loss": 4.137261390686035,
+ "step": 2453
+ },
+ {
+ "epoch": 34.083842794759825,
+ "grad_norm": 0.017029082402586937,
+ "learning_rate": 0.0006,
+ "loss": 4.081112861633301,
+ "step": 2454
+ },
+ {
+ "epoch": 34.09781659388646,
+ "grad_norm": 0.016414813697338104,
+ "learning_rate": 0.0006,
+ "loss": 4.140783309936523,
+ "step": 2455
+ },
+ {
+ "epoch": 34.1117903930131,
+ "grad_norm": 0.018238281831145287,
+ "learning_rate": 0.0006,
+ "loss": 4.161116600036621,
+ "step": 2456
+ },
+ {
+ "epoch": 34.12576419213974,
+ "grad_norm": 0.019929232075810432,
+ "learning_rate": 0.0006,
+ "loss": 4.243722438812256,
+ "step": 2457
+ },
+ {
+ "epoch": 34.13973799126637,
+ "grad_norm": 0.019375786185264587,
+ "learning_rate": 0.0006,
+ "loss": 4.134873390197754,
+ "step": 2458
+ },
+ {
+ "epoch": 34.15371179039301,
+ "grad_norm": 0.018151137977838516,
+ "learning_rate": 0.0006,
+ "loss": 4.098745346069336,
+ "step": 2459
+ },
+ {
+ "epoch": 34.16768558951965,
+ "grad_norm": 0.017054934054613113,
+ "learning_rate": 0.0006,
+ "loss": 4.194662570953369,
+ "step": 2460
+ },
+ {
+ "epoch": 34.18165938864629,
+ "grad_norm": 0.018220555037260056,
+ "learning_rate": 0.0006,
+ "loss": 4.128866195678711,
+ "step": 2461
+ },
+ {
+ "epoch": 34.19563318777293,
+ "grad_norm": 0.01692879945039749,
+ "learning_rate": 0.0006,
+ "loss": 4.062477111816406,
+ "step": 2462
+ },
+ {
+ "epoch": 34.209606986899566,
+ "grad_norm": 0.01674199476838112,
+ "learning_rate": 0.0006,
+ "loss": 4.186149597167969,
+ "step": 2463
+ },
+ {
+ "epoch": 34.223580786026204,
+ "grad_norm": 0.016527455300092697,
+ "learning_rate": 0.0006,
+ "loss": 4.0734028816223145,
+ "step": 2464
+ },
+ {
+ "epoch": 34.237554585152836,
+ "grad_norm": 0.016448386013507843,
+ "learning_rate": 0.0006,
+ "loss": 4.098896026611328,
+ "step": 2465
+ },
+ {
+ "epoch": 34.251528384279474,
+ "grad_norm": 0.014644953422248363,
+ "learning_rate": 0.0006,
+ "loss": 4.1031904220581055,
+ "step": 2466
+ },
+ {
+ "epoch": 34.26550218340611,
+ "grad_norm": 0.016649341210722923,
+ "learning_rate": 0.0006,
+ "loss": 4.148791313171387,
+ "step": 2467
+ },
+ {
+ "epoch": 34.27947598253275,
+ "grad_norm": 0.017442714422941208,
+ "learning_rate": 0.0006,
+ "loss": 4.070799827575684,
+ "step": 2468
+ },
+ {
+ "epoch": 34.29344978165939,
+ "grad_norm": 0.017852013930678368,
+ "learning_rate": 0.0006,
+ "loss": 4.031924247741699,
+ "step": 2469
+ },
+ {
+ "epoch": 34.30742358078603,
+ "grad_norm": 0.016371270641684532,
+ "learning_rate": 0.0006,
+ "loss": 4.149420738220215,
+ "step": 2470
+ },
+ {
+ "epoch": 34.32139737991266,
+ "grad_norm": 0.015846814960241318,
+ "learning_rate": 0.0006,
+ "loss": 4.15427303314209,
+ "step": 2471
+ },
+ {
+ "epoch": 34.3353711790393,
+ "grad_norm": 0.017411163076758385,
+ "learning_rate": 0.0006,
+ "loss": 4.1304192543029785,
+ "step": 2472
+ },
+ {
+ "epoch": 34.34934497816594,
+ "grad_norm": 0.01780046708881855,
+ "learning_rate": 0.0006,
+ "loss": 4.191732406616211,
+ "step": 2473
+ },
+ {
+ "epoch": 34.36331877729258,
+ "grad_norm": 0.01958383060991764,
+ "learning_rate": 0.0006,
+ "loss": 4.068355083465576,
+ "step": 2474
+ },
+ {
+ "epoch": 34.377292576419215,
+ "grad_norm": 0.022173317149281502,
+ "learning_rate": 0.0006,
+ "loss": 4.182929039001465,
+ "step": 2475
+ },
+ {
+ "epoch": 34.391266375545854,
+ "grad_norm": 0.022389156743884087,
+ "learning_rate": 0.0006,
+ "loss": 4.110218048095703,
+ "step": 2476
+ },
+ {
+ "epoch": 34.40524017467249,
+ "grad_norm": 0.022478511556982994,
+ "learning_rate": 0.0006,
+ "loss": 4.095479488372803,
+ "step": 2477
+ },
+ {
+ "epoch": 34.419213973799124,
+ "grad_norm": 0.019504647701978683,
+ "learning_rate": 0.0006,
+ "loss": 4.101964950561523,
+ "step": 2478
+ },
+ {
+ "epoch": 34.43318777292576,
+ "grad_norm": 0.019378576427698135,
+ "learning_rate": 0.0006,
+ "loss": 4.14620304107666,
+ "step": 2479
+ },
+ {
+ "epoch": 34.4471615720524,
+ "grad_norm": 0.017993204295635223,
+ "learning_rate": 0.0006,
+ "loss": 4.183492660522461,
+ "step": 2480
+ },
+ {
+ "epoch": 34.46113537117904,
+ "grad_norm": 0.01814805157482624,
+ "learning_rate": 0.0006,
+ "loss": 4.151247978210449,
+ "step": 2481
+ },
+ {
+ "epoch": 34.47510917030568,
+ "grad_norm": 0.018622448667883873,
+ "learning_rate": 0.0006,
+ "loss": 4.020938396453857,
+ "step": 2482
+ },
+ {
+ "epoch": 34.48908296943232,
+ "grad_norm": 0.018360132351517677,
+ "learning_rate": 0.0006,
+ "loss": 4.064734935760498,
+ "step": 2483
+ },
+ {
+ "epoch": 34.50305676855895,
+ "grad_norm": 0.017715711146593094,
+ "learning_rate": 0.0006,
+ "loss": 4.185687065124512,
+ "step": 2484
+ },
+ {
+ "epoch": 34.51703056768559,
+ "grad_norm": 0.01714295893907547,
+ "learning_rate": 0.0006,
+ "loss": 4.1044721603393555,
+ "step": 2485
+ },
+ {
+ "epoch": 34.531004366812226,
+ "grad_norm": 0.016943803057074547,
+ "learning_rate": 0.0006,
+ "loss": 4.106837272644043,
+ "step": 2486
+ },
+ {
+ "epoch": 34.544978165938865,
+ "grad_norm": 0.0182423684746027,
+ "learning_rate": 0.0006,
+ "loss": 4.1041340827941895,
+ "step": 2487
+ },
+ {
+ "epoch": 34.5589519650655,
+ "grad_norm": 0.01636587455868721,
+ "learning_rate": 0.0006,
+ "loss": 3.9735546112060547,
+ "step": 2488
+ },
+ {
+ "epoch": 34.57292576419214,
+ "grad_norm": 0.014593546278774738,
+ "learning_rate": 0.0006,
+ "loss": 4.041980266571045,
+ "step": 2489
+ },
+ {
+ "epoch": 34.58689956331878,
+ "grad_norm": 0.015436965972185135,
+ "learning_rate": 0.0006,
+ "loss": 4.301148414611816,
+ "step": 2490
+ },
+ {
+ "epoch": 34.60087336244541,
+ "grad_norm": 0.015075269155204296,
+ "learning_rate": 0.0006,
+ "loss": 4.118527412414551,
+ "step": 2491
+ },
+ {
+ "epoch": 34.61484716157205,
+ "grad_norm": 0.01701514609158039,
+ "learning_rate": 0.0006,
+ "loss": 4.245968818664551,
+ "step": 2492
+ },
+ {
+ "epoch": 34.62882096069869,
+ "grad_norm": 0.018047887831926346,
+ "learning_rate": 0.0006,
+ "loss": 4.227261066436768,
+ "step": 2493
+ },
+ {
+ "epoch": 34.64279475982533,
+ "grad_norm": 0.018499085679650307,
+ "learning_rate": 0.0006,
+ "loss": 4.121360778808594,
+ "step": 2494
+ },
+ {
+ "epoch": 34.65676855895197,
+ "grad_norm": 0.018987711519002914,
+ "learning_rate": 0.0006,
+ "loss": 4.131564617156982,
+ "step": 2495
+ },
+ {
+ "epoch": 34.670742358078606,
+ "grad_norm": 0.01834476925432682,
+ "learning_rate": 0.0006,
+ "loss": 4.224941253662109,
+ "step": 2496
+ },
+ {
+ "epoch": 34.68471615720524,
+ "grad_norm": 0.017507942393422127,
+ "learning_rate": 0.0006,
+ "loss": 4.1394548416137695,
+ "step": 2497
+ },
+ {
+ "epoch": 34.698689956331876,
+ "grad_norm": 0.01680580899119377,
+ "learning_rate": 0.0006,
+ "loss": 4.1073431968688965,
+ "step": 2498
+ },
+ {
+ "epoch": 34.712663755458514,
+ "grad_norm": 0.015233514830470085,
+ "learning_rate": 0.0006,
+ "loss": 4.136994361877441,
+ "step": 2499
+ },
+ {
+ "epoch": 34.72663755458515,
+ "grad_norm": 0.015482917428016663,
+ "learning_rate": 0.0006,
+ "loss": 4.110719680786133,
+ "step": 2500
+ },
+ {
+ "epoch": 34.74061135371179,
+ "grad_norm": 0.015940461307764053,
+ "learning_rate": 0.0006,
+ "loss": 4.207851886749268,
+ "step": 2501
+ },
+ {
+ "epoch": 34.75458515283843,
+ "grad_norm": 0.015355597250163555,
+ "learning_rate": 0.0006,
+ "loss": 4.169469833374023,
+ "step": 2502
+ },
+ {
+ "epoch": 34.76855895196506,
+ "grad_norm": 0.015974000096321106,
+ "learning_rate": 0.0006,
+ "loss": 4.110061168670654,
+ "step": 2503
+ },
+ {
+ "epoch": 34.7825327510917,
+ "grad_norm": 0.016088927164673805,
+ "learning_rate": 0.0006,
+ "loss": 4.155452728271484,
+ "step": 2504
+ },
+ {
+ "epoch": 34.79650655021834,
+ "grad_norm": 0.016156120225787163,
+ "learning_rate": 0.0006,
+ "loss": 4.178092956542969,
+ "step": 2505
+ },
+ {
+ "epoch": 34.81048034934498,
+ "grad_norm": 0.014928647316992283,
+ "learning_rate": 0.0006,
+ "loss": 4.05457878112793,
+ "step": 2506
+ },
+ {
+ "epoch": 34.82445414847162,
+ "grad_norm": 0.015289328061044216,
+ "learning_rate": 0.0006,
+ "loss": 4.18869686126709,
+ "step": 2507
+ },
+ {
+ "epoch": 34.838427947598255,
+ "grad_norm": 0.015799039974808693,
+ "learning_rate": 0.0006,
+ "loss": 4.106073379516602,
+ "step": 2508
+ },
+ {
+ "epoch": 34.852401746724894,
+ "grad_norm": 0.015824418514966965,
+ "learning_rate": 0.0006,
+ "loss": 4.0634589195251465,
+ "step": 2509
+ },
+ {
+ "epoch": 34.866375545851525,
+ "grad_norm": 0.015321656130254269,
+ "learning_rate": 0.0006,
+ "loss": 4.093090057373047,
+ "step": 2510
+ },
+ {
+ "epoch": 34.880349344978164,
+ "grad_norm": 0.014380166307091713,
+ "learning_rate": 0.0006,
+ "loss": 4.1750640869140625,
+ "step": 2511
+ },
+ {
+ "epoch": 34.8943231441048,
+ "grad_norm": 0.014962567947804928,
+ "learning_rate": 0.0006,
+ "loss": 4.185277938842773,
+ "step": 2512
+ },
+ {
+ "epoch": 34.90829694323144,
+ "grad_norm": 0.014467543922364712,
+ "learning_rate": 0.0006,
+ "loss": 4.1600799560546875,
+ "step": 2513
+ },
+ {
+ "epoch": 34.92227074235808,
+ "grad_norm": 0.01545181218534708,
+ "learning_rate": 0.0006,
+ "loss": 4.096299648284912,
+ "step": 2514
+ },
+ {
+ "epoch": 34.93624454148472,
+ "grad_norm": 0.01605340465903282,
+ "learning_rate": 0.0006,
+ "loss": 4.203323841094971,
+ "step": 2515
+ },
+ {
+ "epoch": 34.95021834061135,
+ "grad_norm": 0.015203815884888172,
+ "learning_rate": 0.0006,
+ "loss": 4.201385021209717,
+ "step": 2516
+ },
+ {
+ "epoch": 34.96419213973799,
+ "grad_norm": 0.015149409882724285,
+ "learning_rate": 0.0006,
+ "loss": 4.088164329528809,
+ "step": 2517
+ },
+ {
+ "epoch": 34.97816593886463,
+ "grad_norm": 0.014885558746755123,
+ "learning_rate": 0.0006,
+ "loss": 4.073911666870117,
+ "step": 2518
+ },
+ {
+ "epoch": 34.992139737991266,
+ "grad_norm": 0.015003419481217861,
+ "learning_rate": 0.0006,
+ "loss": 4.157113552093506,
+ "step": 2519
+ },
+ {
+ "epoch": 35.0,
+ "grad_norm": 0.01641916297376156,
+ "learning_rate": 0.0006,
+ "loss": 3.970052719116211,
+ "step": 2520
+ },
+ {
+ "epoch": 35.0,
+ "eval_loss": 4.641692161560059,
+ "eval_runtime": 56.8836,
+ "eval_samples_per_second": 42.93,
+ "eval_steps_per_second": 1.354,
+ "step": 2520
+ },
+ {
+ "epoch": 35.01397379912664,
+ "grad_norm": 0.015894673764705658,
+ "learning_rate": 0.0006,
+ "loss": 4.104437351226807,
+ "step": 2521
+ },
+ {
+ "epoch": 35.02794759825328,
+ "grad_norm": 0.0173867866396904,
+ "learning_rate": 0.0006,
+ "loss": 4.091726303100586,
+ "step": 2522
+ },
+ {
+ "epoch": 35.041921397379916,
+ "grad_norm": 0.019272832199931145,
+ "learning_rate": 0.0006,
+ "loss": 4.217905044555664,
+ "step": 2523
+ },
+ {
+ "epoch": 35.05589519650655,
+ "grad_norm": 0.021820371970534325,
+ "learning_rate": 0.0006,
+ "loss": 4.111428260803223,
+ "step": 2524
+ },
+ {
+ "epoch": 35.069868995633186,
+ "grad_norm": 0.02340741828083992,
+ "learning_rate": 0.0006,
+ "loss": 4.115771293640137,
+ "step": 2525
+ },
+ {
+ "epoch": 35.083842794759825,
+ "grad_norm": 0.023760691285133362,
+ "learning_rate": 0.0006,
+ "loss": 4.132383346557617,
+ "step": 2526
+ },
+ {
+ "epoch": 35.09781659388646,
+ "grad_norm": 0.02073468267917633,
+ "learning_rate": 0.0006,
+ "loss": 4.272691249847412,
+ "step": 2527
+ },
+ {
+ "epoch": 35.1117903930131,
+ "grad_norm": 0.02197718247771263,
+ "learning_rate": 0.0006,
+ "loss": 4.105586528778076,
+ "step": 2528
+ },
+ {
+ "epoch": 35.12576419213974,
+ "grad_norm": 0.0200254675000906,
+ "learning_rate": 0.0006,
+ "loss": 3.968820095062256,
+ "step": 2529
+ },
+ {
+ "epoch": 35.13973799126637,
+ "grad_norm": 0.01929587684571743,
+ "learning_rate": 0.0006,
+ "loss": 4.045468807220459,
+ "step": 2530
+ },
+ {
+ "epoch": 35.15371179039301,
+ "grad_norm": 0.01886162906885147,
+ "learning_rate": 0.0006,
+ "loss": 4.185571670532227,
+ "step": 2531
+ },
+ {
+ "epoch": 35.16768558951965,
+ "grad_norm": 0.019789839163422585,
+ "learning_rate": 0.0006,
+ "loss": 3.9734551906585693,
+ "step": 2532
+ },
+ {
+ "epoch": 35.18165938864629,
+ "grad_norm": 0.0198142658919096,
+ "learning_rate": 0.0006,
+ "loss": 4.2662506103515625,
+ "step": 2533
+ },
+ {
+ "epoch": 35.19563318777293,
+ "grad_norm": 0.02087012305855751,
+ "learning_rate": 0.0006,
+ "loss": 4.175931930541992,
+ "step": 2534
+ },
+ {
+ "epoch": 35.209606986899566,
+ "grad_norm": 0.020294535905122757,
+ "learning_rate": 0.0006,
+ "loss": 4.060324668884277,
+ "step": 2535
+ },
+ {
+ "epoch": 35.223580786026204,
+ "grad_norm": 0.01908983662724495,
+ "learning_rate": 0.0006,
+ "loss": 4.217939853668213,
+ "step": 2536
+ },
+ {
+ "epoch": 35.237554585152836,
+ "grad_norm": 0.0178892333060503,
+ "learning_rate": 0.0006,
+ "loss": 4.190216541290283,
+ "step": 2537
+ },
+ {
+ "epoch": 35.251528384279474,
+ "grad_norm": 0.017892315983772278,
+ "learning_rate": 0.0006,
+ "loss": 4.067419052124023,
+ "step": 2538
+ },
+ {
+ "epoch": 35.26550218340611,
+ "grad_norm": 0.018556250259280205,
+ "learning_rate": 0.0006,
+ "loss": 4.105958461761475,
+ "step": 2539
+ },
+ {
+ "epoch": 35.27947598253275,
+ "grad_norm": 0.01944422349333763,
+ "learning_rate": 0.0006,
+ "loss": 4.155775547027588,
+ "step": 2540
+ },
+ {
+ "epoch": 35.29344978165939,
+ "grad_norm": 0.017333146184682846,
+ "learning_rate": 0.0006,
+ "loss": 4.244513511657715,
+ "step": 2541
+ },
+ {
+ "epoch": 35.30742358078603,
+ "grad_norm": 0.01639244332909584,
+ "learning_rate": 0.0006,
+ "loss": 4.167815685272217,
+ "step": 2542
+ },
+ {
+ "epoch": 35.32139737991266,
+ "grad_norm": 0.01610143668949604,
+ "learning_rate": 0.0006,
+ "loss": 4.121486663818359,
+ "step": 2543
+ },
+ {
+ "epoch": 35.3353711790393,
+ "grad_norm": 0.01656411588191986,
+ "learning_rate": 0.0006,
+ "loss": 4.1755499839782715,
+ "step": 2544
+ },
+ {
+ "epoch": 35.34934497816594,
+ "grad_norm": 0.015075297094881535,
+ "learning_rate": 0.0006,
+ "loss": 4.102992534637451,
+ "step": 2545
+ },
+ {
+ "epoch": 35.36331877729258,
+ "grad_norm": 0.01538170501589775,
+ "learning_rate": 0.0006,
+ "loss": 4.161603927612305,
+ "step": 2546
+ },
+ {
+ "epoch": 35.377292576419215,
+ "grad_norm": 0.01483312901109457,
+ "learning_rate": 0.0006,
+ "loss": 4.085355758666992,
+ "step": 2547
+ },
+ {
+ "epoch": 35.391266375545854,
+ "grad_norm": 0.014894738793373108,
+ "learning_rate": 0.0006,
+ "loss": 4.083585262298584,
+ "step": 2548
+ },
+ {
+ "epoch": 35.40524017467249,
+ "grad_norm": 0.01475780550390482,
+ "learning_rate": 0.0006,
+ "loss": 4.012645721435547,
+ "step": 2549
+ },
+ {
+ "epoch": 35.419213973799124,
+ "grad_norm": 0.014531653374433517,
+ "learning_rate": 0.0006,
+ "loss": 4.197113990783691,
+ "step": 2550
+ },
+ {
+ "epoch": 35.43318777292576,
+ "grad_norm": 0.015316493809223175,
+ "learning_rate": 0.0006,
+ "loss": 4.227970123291016,
+ "step": 2551
+ },
+ {
+ "epoch": 35.4471615720524,
+ "grad_norm": 0.014685571193695068,
+ "learning_rate": 0.0006,
+ "loss": 4.089415550231934,
+ "step": 2552
+ },
+ {
+ "epoch": 35.46113537117904,
+ "grad_norm": 0.015337035991251469,
+ "learning_rate": 0.0006,
+ "loss": 4.080583572387695,
+ "step": 2553
+ },
+ {
+ "epoch": 35.47510917030568,
+ "grad_norm": 0.016754208132624626,
+ "learning_rate": 0.0006,
+ "loss": 4.163935661315918,
+ "step": 2554
+ },
+ {
+ "epoch": 35.48908296943232,
+ "grad_norm": 0.017535798251628876,
+ "learning_rate": 0.0006,
+ "loss": 4.093027591705322,
+ "step": 2555
+ },
+ {
+ "epoch": 35.50305676855895,
+ "grad_norm": 0.017336975783109665,
+ "learning_rate": 0.0006,
+ "loss": 4.127961158752441,
+ "step": 2556
+ },
+ {
+ "epoch": 35.51703056768559,
+ "grad_norm": 0.016242152079939842,
+ "learning_rate": 0.0006,
+ "loss": 4.155593395233154,
+ "step": 2557
+ },
+ {
+ "epoch": 35.531004366812226,
+ "grad_norm": 0.015121322125196457,
+ "learning_rate": 0.0006,
+ "loss": 3.984004020690918,
+ "step": 2558
+ },
+ {
+ "epoch": 35.544978165938865,
+ "grad_norm": 0.015178884379565716,
+ "learning_rate": 0.0006,
+ "loss": 4.111564636230469,
+ "step": 2559
+ },
+ {
+ "epoch": 35.5589519650655,
+ "grad_norm": 0.014777131378650665,
+ "learning_rate": 0.0006,
+ "loss": 4.041402816772461,
+ "step": 2560
+ },
+ {
+ "epoch": 35.57292576419214,
+ "grad_norm": 0.015197676606476307,
+ "learning_rate": 0.0006,
+ "loss": 4.133706092834473,
+ "step": 2561
+ },
+ {
+ "epoch": 35.58689956331878,
+ "grad_norm": 0.015905972570180893,
+ "learning_rate": 0.0006,
+ "loss": 4.1069135665893555,
+ "step": 2562
+ },
+ {
+ "epoch": 35.60087336244541,
+ "grad_norm": 0.015942877158522606,
+ "learning_rate": 0.0006,
+ "loss": 4.0720624923706055,
+ "step": 2563
+ },
+ {
+ "epoch": 35.61484716157205,
+ "grad_norm": 0.016697583720088005,
+ "learning_rate": 0.0006,
+ "loss": 4.167448043823242,
+ "step": 2564
+ },
+ {
+ "epoch": 35.62882096069869,
+ "grad_norm": 0.015603577718138695,
+ "learning_rate": 0.0006,
+ "loss": 4.127327919006348,
+ "step": 2565
+ },
+ {
+ "epoch": 35.64279475982533,
+ "grad_norm": 0.017153460532426834,
+ "learning_rate": 0.0006,
+ "loss": 4.056543350219727,
+ "step": 2566
+ },
+ {
+ "epoch": 35.65676855895197,
+ "grad_norm": 0.016731958836317062,
+ "learning_rate": 0.0006,
+ "loss": 4.0681843757629395,
+ "step": 2567
+ },
+ {
+ "epoch": 35.670742358078606,
+ "grad_norm": 0.016167359426617622,
+ "learning_rate": 0.0006,
+ "loss": 4.174992084503174,
+ "step": 2568
+ },
+ {
+ "epoch": 35.68471615720524,
+ "grad_norm": 0.01615108922123909,
+ "learning_rate": 0.0006,
+ "loss": 4.1383161544799805,
+ "step": 2569
+ },
+ {
+ "epoch": 35.698689956331876,
+ "grad_norm": 0.015750816091895103,
+ "learning_rate": 0.0006,
+ "loss": 4.031595706939697,
+ "step": 2570
+ },
+ {
+ "epoch": 35.712663755458514,
+ "grad_norm": 0.015416189096868038,
+ "learning_rate": 0.0006,
+ "loss": 4.067498207092285,
+ "step": 2571
+ },
+ {
+ "epoch": 35.72663755458515,
+ "grad_norm": 0.01570792682468891,
+ "learning_rate": 0.0006,
+ "loss": 4.1338300704956055,
+ "step": 2572
+ },
+ {
+ "epoch": 35.74061135371179,
+ "grad_norm": 0.016367752104997635,
+ "learning_rate": 0.0006,
+ "loss": 4.152523040771484,
+ "step": 2573
+ },
+ {
+ "epoch": 35.75458515283843,
+ "grad_norm": 0.016349803656339645,
+ "learning_rate": 0.0006,
+ "loss": 4.101985454559326,
+ "step": 2574
+ },
+ {
+ "epoch": 35.76855895196506,
+ "grad_norm": 0.01663660630583763,
+ "learning_rate": 0.0006,
+ "loss": 4.190130710601807,
+ "step": 2575
+ },
+ {
+ "epoch": 35.7825327510917,
+ "grad_norm": 0.015599170699715614,
+ "learning_rate": 0.0006,
+ "loss": 4.0792436599731445,
+ "step": 2576
+ },
+ {
+ "epoch": 35.79650655021834,
+ "grad_norm": 0.015472427010536194,
+ "learning_rate": 0.0006,
+ "loss": 4.186446666717529,
+ "step": 2577
+ },
+ {
+ "epoch": 35.81048034934498,
+ "grad_norm": 0.015520439483225346,
+ "learning_rate": 0.0006,
+ "loss": 4.246635437011719,
+ "step": 2578
+ },
+ {
+ "epoch": 35.82445414847162,
+ "grad_norm": 0.016998281702399254,
+ "learning_rate": 0.0006,
+ "loss": 4.13435173034668,
+ "step": 2579
+ },
+ {
+ "epoch": 35.838427947598255,
+ "grad_norm": 0.016290361061692238,
+ "learning_rate": 0.0006,
+ "loss": 4.107428073883057,
+ "step": 2580
+ },
+ {
+ "epoch": 35.852401746724894,
+ "grad_norm": 0.0165516659617424,
+ "learning_rate": 0.0006,
+ "loss": 4.1612138748168945,
+ "step": 2581
+ },
+ {
+ "epoch": 35.866375545851525,
+ "grad_norm": 0.01736091636121273,
+ "learning_rate": 0.0006,
+ "loss": 4.1072869300842285,
+ "step": 2582
+ },
+ {
+ "epoch": 35.880349344978164,
+ "grad_norm": 0.01787031814455986,
+ "learning_rate": 0.0006,
+ "loss": 4.1076812744140625,
+ "step": 2583
+ },
+ {
+ "epoch": 35.8943231441048,
+ "grad_norm": 0.015437985770404339,
+ "learning_rate": 0.0006,
+ "loss": 4.097488880157471,
+ "step": 2584
+ },
+ {
+ "epoch": 35.90829694323144,
+ "grad_norm": 0.016306741163134575,
+ "learning_rate": 0.0006,
+ "loss": 4.105017185211182,
+ "step": 2585
+ },
+ {
+ "epoch": 35.92227074235808,
+ "grad_norm": 0.015723245218396187,
+ "learning_rate": 0.0006,
+ "loss": 4.205283164978027,
+ "step": 2586
+ },
+ {
+ "epoch": 35.93624454148472,
+ "grad_norm": 0.014626456424593925,
+ "learning_rate": 0.0006,
+ "loss": 4.086552619934082,
+ "step": 2587
+ },
+ {
+ "epoch": 35.95021834061135,
+ "grad_norm": 0.014348454773426056,
+ "learning_rate": 0.0006,
+ "loss": 4.171329975128174,
+ "step": 2588
+ },
+ {
+ "epoch": 35.96419213973799,
+ "grad_norm": 0.015598480589687824,
+ "learning_rate": 0.0006,
+ "loss": 4.140815258026123,
+ "step": 2589
+ },
+ {
+ "epoch": 35.97816593886463,
+ "grad_norm": 0.017126062884926796,
+ "learning_rate": 0.0006,
+ "loss": 4.157176971435547,
+ "step": 2590
+ },
+ {
+ "epoch": 35.992139737991266,
+ "grad_norm": 0.01860199309885502,
+ "learning_rate": 0.0006,
+ "loss": 4.252880096435547,
+ "step": 2591
+ },
+ {
+ "epoch": 36.0,
+ "grad_norm": 0.021780220791697502,
+ "learning_rate": 0.0006,
+ "loss": 4.249861717224121,
+ "step": 2592
+ },
+ {
+ "epoch": 36.0,
+ "eval_loss": 4.606466293334961,
+ "eval_runtime": 56.6217,
+ "eval_samples_per_second": 43.128,
+ "eval_steps_per_second": 1.36,
+ "step": 2592
+ },
+ {
+ "epoch": 36.01397379912664,
+ "grad_norm": 0.020892828702926636,
+ "learning_rate": 0.0006,
+ "loss": 4.102055549621582,
+ "step": 2593
+ },
+ {
+ "epoch": 36.02794759825328,
+ "grad_norm": 0.020306255668401718,
+ "learning_rate": 0.0006,
+ "loss": 4.184308052062988,
+ "step": 2594
+ },
+ {
+ "epoch": 36.041921397379916,
+ "grad_norm": 0.02025097794830799,
+ "learning_rate": 0.0006,
+ "loss": 3.9691858291625977,
+ "step": 2595
+ },
+ {
+ "epoch": 36.05589519650655,
+ "grad_norm": 0.020254382863640785,
+ "learning_rate": 0.0006,
+ "loss": 4.02957820892334,
+ "step": 2596
+ },
+ {
+ "epoch": 36.069868995633186,
+ "grad_norm": 0.019307196140289307,
+ "learning_rate": 0.0006,
+ "loss": 4.033181667327881,
+ "step": 2597
+ },
+ {
+ "epoch": 36.083842794759825,
+ "grad_norm": 0.0198476854711771,
+ "learning_rate": 0.0006,
+ "loss": 4.0816874504089355,
+ "step": 2598
+ },
+ {
+ "epoch": 36.09781659388646,
+ "grad_norm": 0.019916515797376633,
+ "learning_rate": 0.0006,
+ "loss": 4.164889812469482,
+ "step": 2599
+ },
+ {
+ "epoch": 36.1117903930131,
+ "grad_norm": 0.0189144778996706,
+ "learning_rate": 0.0006,
+ "loss": 4.038427352905273,
+ "step": 2600
+ },
+ {
+ "epoch": 36.12576419213974,
+ "grad_norm": 0.01754768192768097,
+ "learning_rate": 0.0006,
+ "loss": 4.010427474975586,
+ "step": 2601
+ },
+ {
+ "epoch": 36.13973799126637,
+ "grad_norm": 0.016862990334630013,
+ "learning_rate": 0.0006,
+ "loss": 4.035970211029053,
+ "step": 2602
+ },
+ {
+ "epoch": 36.15371179039301,
+ "grad_norm": 0.019616752862930298,
+ "learning_rate": 0.0006,
+ "loss": 4.079898834228516,
+ "step": 2603
+ },
+ {
+ "epoch": 36.16768558951965,
+ "grad_norm": 0.020131465047597885,
+ "learning_rate": 0.0006,
+ "loss": 4.144684791564941,
+ "step": 2604
+ },
+ {
+ "epoch": 36.18165938864629,
+ "grad_norm": 0.02072521299123764,
+ "learning_rate": 0.0006,
+ "loss": 4.013660430908203,
+ "step": 2605
+ },
+ {
+ "epoch": 36.19563318777293,
+ "grad_norm": 0.01924615167081356,
+ "learning_rate": 0.0006,
+ "loss": 4.071398735046387,
+ "step": 2606
+ },
+ {
+ "epoch": 36.209606986899566,
+ "grad_norm": 0.02013954520225525,
+ "learning_rate": 0.0006,
+ "loss": 4.127094268798828,
+ "step": 2607
+ },
+ {
+ "epoch": 36.223580786026204,
+ "grad_norm": 0.02042008936405182,
+ "learning_rate": 0.0006,
+ "loss": 4.154261589050293,
+ "step": 2608
+ },
+ {
+ "epoch": 36.237554585152836,
+ "grad_norm": 0.021998876705765724,
+ "learning_rate": 0.0006,
+ "loss": 4.204150199890137,
+ "step": 2609
+ },
+ {
+ "epoch": 36.251528384279474,
+ "grad_norm": 0.022946683689951897,
+ "learning_rate": 0.0006,
+ "loss": 4.013632774353027,
+ "step": 2610
+ },
+ {
+ "epoch": 36.26550218340611,
+ "grad_norm": 0.02061830274760723,
+ "learning_rate": 0.0006,
+ "loss": 4.171084403991699,
+ "step": 2611
+ },
+ {
+ "epoch": 36.27947598253275,
+ "grad_norm": 0.01883152686059475,
+ "learning_rate": 0.0006,
+ "loss": 4.149848461151123,
+ "step": 2612
+ },
+ {
+ "epoch": 36.29344978165939,
+ "grad_norm": 0.019945282489061356,
+ "learning_rate": 0.0006,
+ "loss": 4.174656867980957,
+ "step": 2613
+ },
+ {
+ "epoch": 36.30742358078603,
+ "grad_norm": 0.020751051604747772,
+ "learning_rate": 0.0006,
+ "loss": 4.080747127532959,
+ "step": 2614
+ },
+ {
+ "epoch": 36.32139737991266,
+ "grad_norm": 0.01879691146314144,
+ "learning_rate": 0.0006,
+ "loss": 4.057356357574463,
+ "step": 2615
+ },
+ {
+ "epoch": 36.3353711790393,
+ "grad_norm": 0.019037071615457535,
+ "learning_rate": 0.0006,
+ "loss": 4.204607009887695,
+ "step": 2616
+ },
+ {
+ "epoch": 36.34934497816594,
+ "grad_norm": 0.01872078701853752,
+ "learning_rate": 0.0006,
+ "loss": 4.050586700439453,
+ "step": 2617
+ },
+ {
+ "epoch": 36.36331877729258,
+ "grad_norm": 0.018092811107635498,
+ "learning_rate": 0.0006,
+ "loss": 4.102832794189453,
+ "step": 2618
+ },
+ {
+ "epoch": 36.377292576419215,
+ "grad_norm": 0.017010487616062164,
+ "learning_rate": 0.0006,
+ "loss": 4.191788673400879,
+ "step": 2619
+ },
+ {
+ "epoch": 36.391266375545854,
+ "grad_norm": 0.01691143773496151,
+ "learning_rate": 0.0006,
+ "loss": 4.160456657409668,
+ "step": 2620
+ },
+ {
+ "epoch": 36.40524017467249,
+ "grad_norm": 0.017046816647052765,
+ "learning_rate": 0.0006,
+ "loss": 4.073828220367432,
+ "step": 2621
+ },
+ {
+ "epoch": 36.419213973799124,
+ "grad_norm": 0.01707981713116169,
+ "learning_rate": 0.0006,
+ "loss": 4.0980095863342285,
+ "step": 2622
+ },
+ {
+ "epoch": 36.43318777292576,
+ "grad_norm": 0.01630452275276184,
+ "learning_rate": 0.0006,
+ "loss": 4.101321220397949,
+ "step": 2623
+ },
+ {
+ "epoch": 36.4471615720524,
+ "grad_norm": 0.01576915942132473,
+ "learning_rate": 0.0006,
+ "loss": 4.158249855041504,
+ "step": 2624
+ },
+ {
+ "epoch": 36.46113537117904,
+ "grad_norm": 0.016179712489247322,
+ "learning_rate": 0.0006,
+ "loss": 4.077852249145508,
+ "step": 2625
+ },
+ {
+ "epoch": 36.47510917030568,
+ "grad_norm": 0.01609169878065586,
+ "learning_rate": 0.0006,
+ "loss": 4.187317848205566,
+ "step": 2626
+ },
+ {
+ "epoch": 36.48908296943232,
+ "grad_norm": 0.017148254439234734,
+ "learning_rate": 0.0006,
+ "loss": 4.109071731567383,
+ "step": 2627
+ },
+ {
+ "epoch": 36.50305676855895,
+ "grad_norm": 0.015842624008655548,
+ "learning_rate": 0.0006,
+ "loss": 4.145096302032471,
+ "step": 2628
+ },
+ {
+ "epoch": 36.51703056768559,
+ "grad_norm": 0.014906194061040878,
+ "learning_rate": 0.0006,
+ "loss": 4.028440475463867,
+ "step": 2629
+ },
+ {
+ "epoch": 36.531004366812226,
+ "grad_norm": 0.01485038734972477,
+ "learning_rate": 0.0006,
+ "loss": 4.175201416015625,
+ "step": 2630
+ },
+ {
+ "epoch": 36.544978165938865,
+ "grad_norm": 0.015834150835871696,
+ "learning_rate": 0.0006,
+ "loss": 4.076434135437012,
+ "step": 2631
+ },
+ {
+ "epoch": 36.5589519650655,
+ "grad_norm": 0.015411358326673508,
+ "learning_rate": 0.0006,
+ "loss": 4.03912878036499,
+ "step": 2632
+ },
+ {
+ "epoch": 36.57292576419214,
+ "grad_norm": 0.014762846753001213,
+ "learning_rate": 0.0006,
+ "loss": 4.019827365875244,
+ "step": 2633
+ },
+ {
+ "epoch": 36.58689956331878,
+ "grad_norm": 0.015269497409462929,
+ "learning_rate": 0.0006,
+ "loss": 4.116957187652588,
+ "step": 2634
+ },
+ {
+ "epoch": 36.60087336244541,
+ "grad_norm": 0.015171617269515991,
+ "learning_rate": 0.0006,
+ "loss": 4.2069926261901855,
+ "step": 2635
+ },
+ {
+ "epoch": 36.61484716157205,
+ "grad_norm": 0.014770902693271637,
+ "learning_rate": 0.0006,
+ "loss": 4.092021942138672,
+ "step": 2636
+ },
+ {
+ "epoch": 36.62882096069869,
+ "grad_norm": 0.014957531355321407,
+ "learning_rate": 0.0006,
+ "loss": 4.16819953918457,
+ "step": 2637
+ },
+ {
+ "epoch": 36.64279475982533,
+ "grad_norm": 0.014774451963603497,
+ "learning_rate": 0.0006,
+ "loss": 4.046073913574219,
+ "step": 2638
+ },
+ {
+ "epoch": 36.65676855895197,
+ "grad_norm": 0.014320445246994495,
+ "learning_rate": 0.0006,
+ "loss": 3.977792501449585,
+ "step": 2639
+ },
+ {
+ "epoch": 36.670742358078606,
+ "grad_norm": 0.01413130946457386,
+ "learning_rate": 0.0006,
+ "loss": 4.130649566650391,
+ "step": 2640
+ },
+ {
+ "epoch": 36.68471615720524,
+ "grad_norm": 0.014336307533085346,
+ "learning_rate": 0.0006,
+ "loss": 4.161357402801514,
+ "step": 2641
+ },
+ {
+ "epoch": 36.698689956331876,
+ "grad_norm": 0.014516931027173996,
+ "learning_rate": 0.0006,
+ "loss": 4.14248514175415,
+ "step": 2642
+ },
+ {
+ "epoch": 36.712663755458514,
+ "grad_norm": 0.014377126470208168,
+ "learning_rate": 0.0006,
+ "loss": 4.062711715698242,
+ "step": 2643
+ },
+ {
+ "epoch": 36.72663755458515,
+ "grad_norm": 0.014332668855786324,
+ "learning_rate": 0.0006,
+ "loss": 4.06423807144165,
+ "step": 2644
+ },
+ {
+ "epoch": 36.74061135371179,
+ "grad_norm": 0.014030571095645428,
+ "learning_rate": 0.0006,
+ "loss": 4.178613662719727,
+ "step": 2645
+ },
+ {
+ "epoch": 36.75458515283843,
+ "grad_norm": 0.014376776292920113,
+ "learning_rate": 0.0006,
+ "loss": 4.073130130767822,
+ "step": 2646
+ },
+ {
+ "epoch": 36.76855895196506,
+ "grad_norm": 0.013827716931700706,
+ "learning_rate": 0.0006,
+ "loss": 4.043869972229004,
+ "step": 2647
+ },
+ {
+ "epoch": 36.7825327510917,
+ "grad_norm": 0.015548762865364552,
+ "learning_rate": 0.0006,
+ "loss": 4.134049415588379,
+ "step": 2648
+ },
+ {
+ "epoch": 36.79650655021834,
+ "grad_norm": 0.017041970044374466,
+ "learning_rate": 0.0006,
+ "loss": 4.205024242401123,
+ "step": 2649
+ },
+ {
+ "epoch": 36.81048034934498,
+ "grad_norm": 0.017638275399804115,
+ "learning_rate": 0.0006,
+ "loss": 4.199924468994141,
+ "step": 2650
+ },
+ {
+ "epoch": 36.82445414847162,
+ "grad_norm": 0.017771627753973007,
+ "learning_rate": 0.0006,
+ "loss": 4.15706729888916,
+ "step": 2651
+ },
+ {
+ "epoch": 36.838427947598255,
+ "grad_norm": 0.015933562070131302,
+ "learning_rate": 0.0006,
+ "loss": 4.220742225646973,
+ "step": 2652
+ },
+ {
+ "epoch": 36.852401746724894,
+ "grad_norm": 0.01668144389986992,
+ "learning_rate": 0.0006,
+ "loss": 4.161426544189453,
+ "step": 2653
+ },
+ {
+ "epoch": 36.866375545851525,
+ "grad_norm": 0.018014680594205856,
+ "learning_rate": 0.0006,
+ "loss": 4.183046817779541,
+ "step": 2654
+ },
+ {
+ "epoch": 36.880349344978164,
+ "grad_norm": 0.018808670341968536,
+ "learning_rate": 0.0006,
+ "loss": 4.105293273925781,
+ "step": 2655
+ },
+ {
+ "epoch": 36.8943231441048,
+ "grad_norm": 0.018003476783633232,
+ "learning_rate": 0.0006,
+ "loss": 4.103171348571777,
+ "step": 2656
+ },
+ {
+ "epoch": 36.90829694323144,
+ "grad_norm": 0.016865914687514305,
+ "learning_rate": 0.0006,
+ "loss": 4.125634670257568,
+ "step": 2657
+ },
+ {
+ "epoch": 36.92227074235808,
+ "grad_norm": 0.018371153622865677,
+ "learning_rate": 0.0006,
+ "loss": 4.192461013793945,
+ "step": 2658
+ },
+ {
+ "epoch": 36.93624454148472,
+ "grad_norm": 0.019045377150177956,
+ "learning_rate": 0.0006,
+ "loss": 4.198105812072754,
+ "step": 2659
+ },
+ {
+ "epoch": 36.95021834061135,
+ "grad_norm": 0.01594168320298195,
+ "learning_rate": 0.0006,
+ "loss": 4.169788360595703,
+ "step": 2660
+ },
+ {
+ "epoch": 36.96419213973799,
+ "grad_norm": 0.015645820647478104,
+ "learning_rate": 0.0006,
+ "loss": 4.047277450561523,
+ "step": 2661
+ },
+ {
+ "epoch": 36.97816593886463,
+ "grad_norm": 0.016313156113028526,
+ "learning_rate": 0.0006,
+ "loss": 4.087392807006836,
+ "step": 2662
+ },
+ {
+ "epoch": 36.992139737991266,
+ "grad_norm": 0.017050625756382942,
+ "learning_rate": 0.0006,
+ "loss": 4.0908002853393555,
+ "step": 2663
+ },
+ {
+ "epoch": 37.0,
+ "grad_norm": 0.017730968073010445,
+ "learning_rate": 0.0006,
+ "loss": 4.110957622528076,
+ "step": 2664
+ },
+ {
+ "epoch": 37.0,
+ "eval_loss": 4.563793182373047,
+ "eval_runtime": 56.5922,
+ "eval_samples_per_second": 43.151,
+ "eval_steps_per_second": 1.361,
+ "step": 2664
+ },
+ {
+ "epoch": 37.01397379912664,
+ "grad_norm": 0.015687910839915276,
+ "learning_rate": 0.0006,
+ "loss": 4.143589019775391,
+ "step": 2665
+ },
+ {
+ "epoch": 37.02794759825328,
+ "grad_norm": 0.017451366409659386,
+ "learning_rate": 0.0006,
+ "loss": 4.0326666831970215,
+ "step": 2666
+ },
+ {
+ "epoch": 37.041921397379916,
+ "grad_norm": 0.01909538544714451,
+ "learning_rate": 0.0006,
+ "loss": 3.9706249237060547,
+ "step": 2667
+ },
+ {
+ "epoch": 37.05589519650655,
+ "grad_norm": 0.01988379657268524,
+ "learning_rate": 0.0006,
+ "loss": 3.9531009197235107,
+ "step": 2668
+ },
+ {
+ "epoch": 37.069868995633186,
+ "grad_norm": 0.02035115472972393,
+ "learning_rate": 0.0006,
+ "loss": 3.962214469909668,
+ "step": 2669
+ },
+ {
+ "epoch": 37.083842794759825,
+ "grad_norm": 0.01952001079916954,
+ "learning_rate": 0.0006,
+ "loss": 4.182262897491455,
+ "step": 2670
+ },
+ {
+ "epoch": 37.09781659388646,
+ "grad_norm": 0.01854325644671917,
+ "learning_rate": 0.0006,
+ "loss": 4.127663612365723,
+ "step": 2671
+ },
+ {
+ "epoch": 37.1117903930131,
+ "grad_norm": 0.017757616937160492,
+ "learning_rate": 0.0006,
+ "loss": 4.100804805755615,
+ "step": 2672
+ },
+ {
+ "epoch": 37.12576419213974,
+ "grad_norm": 0.016712479293346405,
+ "learning_rate": 0.0006,
+ "loss": 4.026284217834473,
+ "step": 2673
+ },
+ {
+ "epoch": 37.13973799126637,
+ "grad_norm": 0.0168139711022377,
+ "learning_rate": 0.0006,
+ "loss": 4.183882713317871,
+ "step": 2674
+ },
+ {
+ "epoch": 37.15371179039301,
+ "grad_norm": 0.017060423269867897,
+ "learning_rate": 0.0006,
+ "loss": 4.131832599639893,
+ "step": 2675
+ },
+ {
+ "epoch": 37.16768558951965,
+ "grad_norm": 0.01726631447672844,
+ "learning_rate": 0.0006,
+ "loss": 3.9156057834625244,
+ "step": 2676
+ },
+ {
+ "epoch": 37.18165938864629,
+ "grad_norm": 0.016348622739315033,
+ "learning_rate": 0.0006,
+ "loss": 4.1731367111206055,
+ "step": 2677
+ },
+ {
+ "epoch": 37.19563318777293,
+ "grad_norm": 0.016640059649944305,
+ "learning_rate": 0.0006,
+ "loss": 4.10997200012207,
+ "step": 2678
+ },
+ {
+ "epoch": 37.209606986899566,
+ "grad_norm": 0.01601342298090458,
+ "learning_rate": 0.0006,
+ "loss": 4.081145286560059,
+ "step": 2679
+ },
+ {
+ "epoch": 37.223580786026204,
+ "grad_norm": 0.01735631190240383,
+ "learning_rate": 0.0006,
+ "loss": 4.202108860015869,
+ "step": 2680
+ },
+ {
+ "epoch": 37.237554585152836,
+ "grad_norm": 0.019164685159921646,
+ "learning_rate": 0.0006,
+ "loss": 4.104109764099121,
+ "step": 2681
+ },
+ {
+ "epoch": 37.251528384279474,
+ "grad_norm": 0.01881759613752365,
+ "learning_rate": 0.0006,
+ "loss": 4.145532131195068,
+ "step": 2682
+ },
+ {
+ "epoch": 37.26550218340611,
+ "grad_norm": 0.017656555399298668,
+ "learning_rate": 0.0006,
+ "loss": 4.012782096862793,
+ "step": 2683
+ },
+ {
+ "epoch": 37.27947598253275,
+ "grad_norm": 0.01624823734164238,
+ "learning_rate": 0.0006,
+ "loss": 4.001727104187012,
+ "step": 2684
+ },
+ {
+ "epoch": 37.29344978165939,
+ "grad_norm": 0.016973139718174934,
+ "learning_rate": 0.0006,
+ "loss": 4.095399856567383,
+ "step": 2685
+ },
+ {
+ "epoch": 37.30742358078603,
+ "grad_norm": 0.015503671951591969,
+ "learning_rate": 0.0006,
+ "loss": 4.034431457519531,
+ "step": 2686
+ },
+ {
+ "epoch": 37.32139737991266,
+ "grad_norm": 0.015370250679552555,
+ "learning_rate": 0.0006,
+ "loss": 4.075836181640625,
+ "step": 2687
+ },
+ {
+ "epoch": 37.3353711790393,
+ "grad_norm": 0.01564021222293377,
+ "learning_rate": 0.0006,
+ "loss": 4.125202178955078,
+ "step": 2688
+ },
+ {
+ "epoch": 37.34934497816594,
+ "grad_norm": 0.016717975959181786,
+ "learning_rate": 0.0006,
+ "loss": 4.153017044067383,
+ "step": 2689
+ },
+ {
+ "epoch": 37.36331877729258,
+ "grad_norm": 0.015263617038726807,
+ "learning_rate": 0.0006,
+ "loss": 4.021814823150635,
+ "step": 2690
+ },
+ {
+ "epoch": 37.377292576419215,
+ "grad_norm": 0.016497142612934113,
+ "learning_rate": 0.0006,
+ "loss": 4.069095611572266,
+ "step": 2691
+ },
+ {
+ "epoch": 37.391266375545854,
+ "grad_norm": 0.01669355481863022,
+ "learning_rate": 0.0006,
+ "loss": 4.174587726593018,
+ "step": 2692
+ },
+ {
+ "epoch": 37.40524017467249,
+ "grad_norm": 0.019695773720741272,
+ "learning_rate": 0.0006,
+ "loss": 4.0073041915893555,
+ "step": 2693
+ },
+ {
+ "epoch": 37.419213973799124,
+ "grad_norm": 0.023720644414424896,
+ "learning_rate": 0.0006,
+ "loss": 4.111948013305664,
+ "step": 2694
+ },
+ {
+ "epoch": 37.43318777292576,
+ "grad_norm": 0.02399350143969059,
+ "learning_rate": 0.0006,
+ "loss": 4.081164360046387,
+ "step": 2695
+ },
+ {
+ "epoch": 37.4471615720524,
+ "grad_norm": 0.019630271941423416,
+ "learning_rate": 0.0006,
+ "loss": 4.065731048583984,
+ "step": 2696
+ },
+ {
+ "epoch": 37.46113537117904,
+ "grad_norm": 0.019679008051753044,
+ "learning_rate": 0.0006,
+ "loss": 3.993852376937866,
+ "step": 2697
+ },
+ {
+ "epoch": 37.47510917030568,
+ "grad_norm": 0.020597299560904503,
+ "learning_rate": 0.0006,
+ "loss": 4.275555610656738,
+ "step": 2698
+ },
+ {
+ "epoch": 37.48908296943232,
+ "grad_norm": 0.01983889564871788,
+ "learning_rate": 0.0006,
+ "loss": 4.254708290100098,
+ "step": 2699
+ },
+ {
+ "epoch": 37.50305676855895,
+ "grad_norm": 0.019150668755173683,
+ "learning_rate": 0.0006,
+ "loss": 4.097885608673096,
+ "step": 2700
+ },
+ {
+ "epoch": 37.51703056768559,
+ "grad_norm": 0.018847715109586716,
+ "learning_rate": 0.0006,
+ "loss": 4.054081439971924,
+ "step": 2701
+ },
+ {
+ "epoch": 37.531004366812226,
+ "grad_norm": 0.019607288762927055,
+ "learning_rate": 0.0006,
+ "loss": 3.9891440868377686,
+ "step": 2702
+ },
+ {
+ "epoch": 37.544978165938865,
+ "grad_norm": 0.018863150849938393,
+ "learning_rate": 0.0006,
+ "loss": 4.084877014160156,
+ "step": 2703
+ },
+ {
+ "epoch": 37.5589519650655,
+ "grad_norm": 0.017340239137411118,
+ "learning_rate": 0.0006,
+ "loss": 4.113290786743164,
+ "step": 2704
+ },
+ {
+ "epoch": 37.57292576419214,
+ "grad_norm": 0.015628693625330925,
+ "learning_rate": 0.0006,
+ "loss": 4.1483001708984375,
+ "step": 2705
+ },
+ {
+ "epoch": 37.58689956331878,
+ "grad_norm": 0.016493283212184906,
+ "learning_rate": 0.0006,
+ "loss": 4.1040143966674805,
+ "step": 2706
+ },
+ {
+ "epoch": 37.60087336244541,
+ "grad_norm": 0.016732443124055862,
+ "learning_rate": 0.0006,
+ "loss": 4.07205057144165,
+ "step": 2707
+ },
+ {
+ "epoch": 37.61484716157205,
+ "grad_norm": 0.017176156863570213,
+ "learning_rate": 0.0006,
+ "loss": 4.1279754638671875,
+ "step": 2708
+ },
+ {
+ "epoch": 37.62882096069869,
+ "grad_norm": 0.015730608254671097,
+ "learning_rate": 0.0006,
+ "loss": 4.086658000946045,
+ "step": 2709
+ },
+ {
+ "epoch": 37.64279475982533,
+ "grad_norm": 0.016111090779304504,
+ "learning_rate": 0.0006,
+ "loss": 4.064335823059082,
+ "step": 2710
+ },
+ {
+ "epoch": 37.65676855895197,
+ "grad_norm": 0.01769379712641239,
+ "learning_rate": 0.0006,
+ "loss": 4.116047382354736,
+ "step": 2711
+ },
+ {
+ "epoch": 37.670742358078606,
+ "grad_norm": 0.01683102548122406,
+ "learning_rate": 0.0006,
+ "loss": 4.095986843109131,
+ "step": 2712
+ },
+ {
+ "epoch": 37.68471615720524,
+ "grad_norm": 0.015726450830698013,
+ "learning_rate": 0.0006,
+ "loss": 4.10775089263916,
+ "step": 2713
+ },
+ {
+ "epoch": 37.698689956331876,
+ "grad_norm": 0.014760235324501991,
+ "learning_rate": 0.0006,
+ "loss": 4.1234893798828125,
+ "step": 2714
+ },
+ {
+ "epoch": 37.712663755458514,
+ "grad_norm": 0.014933438040316105,
+ "learning_rate": 0.0006,
+ "loss": 4.079200744628906,
+ "step": 2715
+ },
+ {
+ "epoch": 37.72663755458515,
+ "grad_norm": 0.015723997727036476,
+ "learning_rate": 0.0006,
+ "loss": 4.07865047454834,
+ "step": 2716
+ },
+ {
+ "epoch": 37.74061135371179,
+ "grad_norm": 0.014539001509547234,
+ "learning_rate": 0.0006,
+ "loss": 4.036242485046387,
+ "step": 2717
+ },
+ {
+ "epoch": 37.75458515283843,
+ "grad_norm": 0.014844979159533978,
+ "learning_rate": 0.0006,
+ "loss": 4.1051483154296875,
+ "step": 2718
+ },
+ {
+ "epoch": 37.76855895196506,
+ "grad_norm": 0.015939844772219658,
+ "learning_rate": 0.0006,
+ "loss": 4.017253875732422,
+ "step": 2719
+ },
+ {
+ "epoch": 37.7825327510917,
+ "grad_norm": 0.014771555550396442,
+ "learning_rate": 0.0006,
+ "loss": 4.0694661140441895,
+ "step": 2720
+ },
+ {
+ "epoch": 37.79650655021834,
+ "grad_norm": 0.015777921304106712,
+ "learning_rate": 0.0006,
+ "loss": 3.947587490081787,
+ "step": 2721
+ },
+ {
+ "epoch": 37.81048034934498,
+ "grad_norm": 0.017451291903853416,
+ "learning_rate": 0.0006,
+ "loss": 4.093625068664551,
+ "step": 2722
+ },
+ {
+ "epoch": 37.82445414847162,
+ "grad_norm": 0.016829246655106544,
+ "learning_rate": 0.0006,
+ "loss": 4.09030818939209,
+ "step": 2723
+ },
+ {
+ "epoch": 37.838427947598255,
+ "grad_norm": 0.016984397545456886,
+ "learning_rate": 0.0006,
+ "loss": 4.140626907348633,
+ "step": 2724
+ },
+ {
+ "epoch": 37.852401746724894,
+ "grad_norm": 0.01732603646814823,
+ "learning_rate": 0.0006,
+ "loss": 4.128276824951172,
+ "step": 2725
+ },
+ {
+ "epoch": 37.866375545851525,
+ "grad_norm": 0.01744619384407997,
+ "learning_rate": 0.0006,
+ "loss": 4.103130340576172,
+ "step": 2726
+ },
+ {
+ "epoch": 37.880349344978164,
+ "grad_norm": 0.016331320628523827,
+ "learning_rate": 0.0006,
+ "loss": 4.037951469421387,
+ "step": 2727
+ },
+ {
+ "epoch": 37.8943231441048,
+ "grad_norm": 0.015919430181384087,
+ "learning_rate": 0.0006,
+ "loss": 4.121513366699219,
+ "step": 2728
+ },
+ {
+ "epoch": 37.90829694323144,
+ "grad_norm": 0.016153695061802864,
+ "learning_rate": 0.0006,
+ "loss": 4.099384307861328,
+ "step": 2729
+ },
+ {
+ "epoch": 37.92227074235808,
+ "grad_norm": 0.01628800667822361,
+ "learning_rate": 0.0006,
+ "loss": 4.092325687408447,
+ "step": 2730
+ },
+ {
+ "epoch": 37.93624454148472,
+ "grad_norm": 0.017285684123635292,
+ "learning_rate": 0.0006,
+ "loss": 3.9828906059265137,
+ "step": 2731
+ },
+ {
+ "epoch": 37.95021834061135,
+ "grad_norm": 0.016047554090619087,
+ "learning_rate": 0.0006,
+ "loss": 4.100419044494629,
+ "step": 2732
+ },
+ {
+ "epoch": 37.96419213973799,
+ "grad_norm": 0.01562986709177494,
+ "learning_rate": 0.0006,
+ "loss": 4.065832614898682,
+ "step": 2733
+ },
+ {
+ "epoch": 37.97816593886463,
+ "grad_norm": 0.01677163504064083,
+ "learning_rate": 0.0006,
+ "loss": 4.077518463134766,
+ "step": 2734
+ },
+ {
+ "epoch": 37.992139737991266,
+ "grad_norm": 0.01755492016673088,
+ "learning_rate": 0.0006,
+ "loss": 4.1076154708862305,
+ "step": 2735
+ },
+ {
+ "epoch": 38.0,
+ "grad_norm": 0.01929762214422226,
+ "learning_rate": 0.0006,
+ "loss": 4.09603214263916,
+ "step": 2736
+ },
+ {
+ "epoch": 38.0,
+ "eval_loss": 4.582442283630371,
+ "eval_runtime": 62.8649,
+ "eval_samples_per_second": 38.845,
+ "eval_steps_per_second": 1.225,
+ "step": 2736
+ },
+ {
+ "epoch": 38.01397379912664,
+ "grad_norm": 0.019129136577248573,
+ "learning_rate": 0.0006,
+ "loss": 4.095005989074707,
+ "step": 2737
+ },
+ {
+ "epoch": 38.02794759825328,
+ "grad_norm": 0.019167408347129822,
+ "learning_rate": 0.0006,
+ "loss": 4.054426670074463,
+ "step": 2738
+ },
+ {
+ "epoch": 38.041921397379916,
+ "grad_norm": 0.01789460889995098,
+ "learning_rate": 0.0006,
+ "loss": 4.0434417724609375,
+ "step": 2739
+ },
+ {
+ "epoch": 38.05589519650655,
+ "grad_norm": 0.015947027131915092,
+ "learning_rate": 0.0006,
+ "loss": 4.074516296386719,
+ "step": 2740
+ },
+ {
+ "epoch": 38.069868995633186,
+ "grad_norm": 0.016667252406477928,
+ "learning_rate": 0.0006,
+ "loss": 4.108545303344727,
+ "step": 2741
+ },
+ {
+ "epoch": 38.083842794759825,
+ "grad_norm": 0.016014399006962776,
+ "learning_rate": 0.0006,
+ "loss": 4.026925086975098,
+ "step": 2742
+ },
+ {
+ "epoch": 38.09781659388646,
+ "grad_norm": 0.015397695824503899,
+ "learning_rate": 0.0006,
+ "loss": 4.163970470428467,
+ "step": 2743
+ },
+ {
+ "epoch": 38.1117903930131,
+ "grad_norm": 0.01563289947807789,
+ "learning_rate": 0.0006,
+ "loss": 4.116578578948975,
+ "step": 2744
+ },
+ {
+ "epoch": 38.12576419213974,
+ "grad_norm": 0.015512318350374699,
+ "learning_rate": 0.0006,
+ "loss": 4.096311569213867,
+ "step": 2745
+ },
+ {
+ "epoch": 38.13973799126637,
+ "grad_norm": 0.015426642261445522,
+ "learning_rate": 0.0006,
+ "loss": 4.017065048217773,
+ "step": 2746
+ },
+ {
+ "epoch": 38.15371179039301,
+ "grad_norm": 0.015115435235202312,
+ "learning_rate": 0.0006,
+ "loss": 4.005434989929199,
+ "step": 2747
+ },
+ {
+ "epoch": 38.16768558951965,
+ "grad_norm": 0.015235894359648228,
+ "learning_rate": 0.0006,
+ "loss": 3.982743263244629,
+ "step": 2748
+ },
+ {
+ "epoch": 38.18165938864629,
+ "grad_norm": 0.016095101833343506,
+ "learning_rate": 0.0006,
+ "loss": 4.0849385261535645,
+ "step": 2749
+ },
+ {
+ "epoch": 38.19563318777293,
+ "grad_norm": 0.015201197937130928,
+ "learning_rate": 0.0006,
+ "loss": 4.103863716125488,
+ "step": 2750
+ },
+ {
+ "epoch": 38.209606986899566,
+ "grad_norm": 0.015687478706240654,
+ "learning_rate": 0.0006,
+ "loss": 4.12074089050293,
+ "step": 2751
+ },
+ {
+ "epoch": 38.223580786026204,
+ "grad_norm": 0.015201589092612267,
+ "learning_rate": 0.0006,
+ "loss": 4.056872367858887,
+ "step": 2752
+ },
+ {
+ "epoch": 38.237554585152836,
+ "grad_norm": 0.015154894441366196,
+ "learning_rate": 0.0006,
+ "loss": 4.089670181274414,
+ "step": 2753
+ },
+ {
+ "epoch": 38.251528384279474,
+ "grad_norm": 0.01586608588695526,
+ "learning_rate": 0.0006,
+ "loss": 4.155317306518555,
+ "step": 2754
+ },
+ {
+ "epoch": 38.26550218340611,
+ "grad_norm": 0.01609654724597931,
+ "learning_rate": 0.0006,
+ "loss": 3.995800495147705,
+ "step": 2755
+ },
+ {
+ "epoch": 38.27947598253275,
+ "grad_norm": 0.015495185740292072,
+ "learning_rate": 0.0006,
+ "loss": 3.995769739151001,
+ "step": 2756
+ },
+ {
+ "epoch": 38.29344978165939,
+ "grad_norm": 0.017042381688952446,
+ "learning_rate": 0.0006,
+ "loss": 4.132832050323486,
+ "step": 2757
+ },
+ {
+ "epoch": 38.30742358078603,
+ "grad_norm": 0.018706614151597023,
+ "learning_rate": 0.0006,
+ "loss": 4.038820743560791,
+ "step": 2758
+ },
+ {
+ "epoch": 38.32139737991266,
+ "grad_norm": 0.018316026777029037,
+ "learning_rate": 0.0006,
+ "loss": 3.972052574157715,
+ "step": 2759
+ },
+ {
+ "epoch": 38.3353711790393,
+ "grad_norm": 0.018375389277935028,
+ "learning_rate": 0.0006,
+ "loss": 4.152097702026367,
+ "step": 2760
+ },
+ {
+ "epoch": 38.34934497816594,
+ "grad_norm": 0.016879886388778687,
+ "learning_rate": 0.0006,
+ "loss": 4.020869731903076,
+ "step": 2761
+ },
+ {
+ "epoch": 38.36331877729258,
+ "grad_norm": 0.015728183090686798,
+ "learning_rate": 0.0006,
+ "loss": 4.008016586303711,
+ "step": 2762
+ },
+ {
+ "epoch": 38.377292576419215,
+ "grad_norm": 0.016147758811712265,
+ "learning_rate": 0.0006,
+ "loss": 4.035854339599609,
+ "step": 2763
+ },
+ {
+ "epoch": 38.391266375545854,
+ "grad_norm": 0.017774797976017,
+ "learning_rate": 0.0006,
+ "loss": 4.122506141662598,
+ "step": 2764
+ },
+ {
+ "epoch": 38.40524017467249,
+ "grad_norm": 0.016590069979429245,
+ "learning_rate": 0.0006,
+ "loss": 4.082847595214844,
+ "step": 2765
+ },
+ {
+ "epoch": 38.419213973799124,
+ "grad_norm": 0.016544528305530548,
+ "learning_rate": 0.0006,
+ "loss": 4.184270858764648,
+ "step": 2766
+ },
+ {
+ "epoch": 38.43318777292576,
+ "grad_norm": 0.017972925677895546,
+ "learning_rate": 0.0006,
+ "loss": 3.987558603286743,
+ "step": 2767
+ },
+ {
+ "epoch": 38.4471615720524,
+ "grad_norm": 0.018562227487564087,
+ "learning_rate": 0.0006,
+ "loss": 4.1193461418151855,
+ "step": 2768
+ },
+ {
+ "epoch": 38.46113537117904,
+ "grad_norm": 0.018896685913205147,
+ "learning_rate": 0.0006,
+ "loss": 4.057808876037598,
+ "step": 2769
+ },
+ {
+ "epoch": 38.47510917030568,
+ "grad_norm": 0.016793809831142426,
+ "learning_rate": 0.0006,
+ "loss": 4.069523334503174,
+ "step": 2770
+ },
+ {
+ "epoch": 38.48908296943232,
+ "grad_norm": 0.01698676310479641,
+ "learning_rate": 0.0006,
+ "loss": 3.997624158859253,
+ "step": 2771
+ },
+ {
+ "epoch": 38.50305676855895,
+ "grad_norm": 0.016784202307462692,
+ "learning_rate": 0.0006,
+ "loss": 4.174795627593994,
+ "step": 2772
+ },
+ {
+ "epoch": 38.51703056768559,
+ "grad_norm": 0.016785014420747757,
+ "learning_rate": 0.0006,
+ "loss": 3.999321222305298,
+ "step": 2773
+ },
+ {
+ "epoch": 38.531004366812226,
+ "grad_norm": 0.018615184351801872,
+ "learning_rate": 0.0006,
+ "loss": 4.191237926483154,
+ "step": 2774
+ },
+ {
+ "epoch": 38.544978165938865,
+ "grad_norm": 0.019090790301561356,
+ "learning_rate": 0.0006,
+ "loss": 4.106966018676758,
+ "step": 2775
+ },
+ {
+ "epoch": 38.5589519650655,
+ "grad_norm": 0.018298320472240448,
+ "learning_rate": 0.0006,
+ "loss": 4.063363552093506,
+ "step": 2776
+ },
+ {
+ "epoch": 38.57292576419214,
+ "grad_norm": 0.018541090190410614,
+ "learning_rate": 0.0006,
+ "loss": 4.101193428039551,
+ "step": 2777
+ },
+ {
+ "epoch": 38.58689956331878,
+ "grad_norm": 0.01849970407783985,
+ "learning_rate": 0.0006,
+ "loss": 4.071335792541504,
+ "step": 2778
+ },
+ {
+ "epoch": 38.60087336244541,
+ "grad_norm": 0.018309077247977257,
+ "learning_rate": 0.0006,
+ "loss": 4.245305061340332,
+ "step": 2779
+ },
+ {
+ "epoch": 38.61484716157205,
+ "grad_norm": 0.017681840807199478,
+ "learning_rate": 0.0006,
+ "loss": 4.081173419952393,
+ "step": 2780
+ },
+ {
+ "epoch": 38.62882096069869,
+ "grad_norm": 0.016286224126815796,
+ "learning_rate": 0.0006,
+ "loss": 4.024351596832275,
+ "step": 2781
+ },
+ {
+ "epoch": 38.64279475982533,
+ "grad_norm": 0.015837235376238823,
+ "learning_rate": 0.0006,
+ "loss": 4.028205394744873,
+ "step": 2782
+ },
+ {
+ "epoch": 38.65676855895197,
+ "grad_norm": 0.016062596812844276,
+ "learning_rate": 0.0006,
+ "loss": 4.131122589111328,
+ "step": 2783
+ },
+ {
+ "epoch": 38.670742358078606,
+ "grad_norm": 0.01711239665746689,
+ "learning_rate": 0.0006,
+ "loss": 4.124416351318359,
+ "step": 2784
+ },
+ {
+ "epoch": 38.68471615720524,
+ "grad_norm": 0.015304913744330406,
+ "learning_rate": 0.0006,
+ "loss": 4.056122779846191,
+ "step": 2785
+ },
+ {
+ "epoch": 38.698689956331876,
+ "grad_norm": 0.015971887856721878,
+ "learning_rate": 0.0006,
+ "loss": 4.132555961608887,
+ "step": 2786
+ },
+ {
+ "epoch": 38.712663755458514,
+ "grad_norm": 0.015787221491336823,
+ "learning_rate": 0.0006,
+ "loss": 4.051596164703369,
+ "step": 2787
+ },
+ {
+ "epoch": 38.72663755458515,
+ "grad_norm": 0.01605444774031639,
+ "learning_rate": 0.0006,
+ "loss": 4.191808700561523,
+ "step": 2788
+ },
+ {
+ "epoch": 38.74061135371179,
+ "grad_norm": 0.016383009031414986,
+ "learning_rate": 0.0006,
+ "loss": 4.0648417472839355,
+ "step": 2789
+ },
+ {
+ "epoch": 38.75458515283843,
+ "grad_norm": 0.016135554760694504,
+ "learning_rate": 0.0006,
+ "loss": 3.988563060760498,
+ "step": 2790
+ },
+ {
+ "epoch": 38.76855895196506,
+ "grad_norm": 0.016634158790111542,
+ "learning_rate": 0.0006,
+ "loss": 4.133493423461914,
+ "step": 2791
+ },
+ {
+ "epoch": 38.7825327510917,
+ "grad_norm": 0.01688210666179657,
+ "learning_rate": 0.0006,
+ "loss": 4.091277122497559,
+ "step": 2792
+ },
+ {
+ "epoch": 38.79650655021834,
+ "grad_norm": 0.016102122142910957,
+ "learning_rate": 0.0006,
+ "loss": 4.051219940185547,
+ "step": 2793
+ },
+ {
+ "epoch": 38.81048034934498,
+ "grad_norm": 0.016585038974881172,
+ "learning_rate": 0.0006,
+ "loss": 4.058587074279785,
+ "step": 2794
+ },
+ {
+ "epoch": 38.82445414847162,
+ "grad_norm": 0.017700951546430588,
+ "learning_rate": 0.0006,
+ "loss": 4.016030311584473,
+ "step": 2795
+ },
+ {
+ "epoch": 38.838427947598255,
+ "grad_norm": 0.017922203987836838,
+ "learning_rate": 0.0006,
+ "loss": 4.079977989196777,
+ "step": 2796
+ },
+ {
+ "epoch": 38.852401746724894,
+ "grad_norm": 0.017299890518188477,
+ "learning_rate": 0.0006,
+ "loss": 4.062984943389893,
+ "step": 2797
+ },
+ {
+ "epoch": 38.866375545851525,
+ "grad_norm": 0.018648581579327583,
+ "learning_rate": 0.0006,
+ "loss": 4.065717697143555,
+ "step": 2798
+ },
+ {
+ "epoch": 38.880349344978164,
+ "grad_norm": 0.019057920202612877,
+ "learning_rate": 0.0006,
+ "loss": 4.227892875671387,
+ "step": 2799
+ },
+ {
+ "epoch": 38.8943231441048,
+ "grad_norm": 0.018824422731995583,
+ "learning_rate": 0.0006,
+ "loss": 4.116058349609375,
+ "step": 2800
+ },
+ {
+ "epoch": 38.90829694323144,
+ "grad_norm": 0.019084978848695755,
+ "learning_rate": 0.0006,
+ "loss": 4.061677932739258,
+ "step": 2801
+ },
+ {
+ "epoch": 38.92227074235808,
+ "grad_norm": 0.017842981964349747,
+ "learning_rate": 0.0006,
+ "loss": 4.083813667297363,
+ "step": 2802
+ },
+ {
+ "epoch": 38.93624454148472,
+ "grad_norm": 0.01880858838558197,
+ "learning_rate": 0.0006,
+ "loss": 4.152561187744141,
+ "step": 2803
+ },
+ {
+ "epoch": 38.95021834061135,
+ "grad_norm": 0.017400087788701057,
+ "learning_rate": 0.0006,
+ "loss": 4.154635429382324,
+ "step": 2804
+ },
+ {
+ "epoch": 38.96419213973799,
+ "grad_norm": 0.017682507634162903,
+ "learning_rate": 0.0006,
+ "loss": 4.151128768920898,
+ "step": 2805
+ },
+ {
+ "epoch": 38.97816593886463,
+ "grad_norm": 0.015803052112460136,
+ "learning_rate": 0.0006,
+ "loss": 4.10498571395874,
+ "step": 2806
+ },
+ {
+ "epoch": 38.992139737991266,
+ "grad_norm": 0.01582511141896248,
+ "learning_rate": 0.0006,
+ "loss": 4.140159606933594,
+ "step": 2807
+ },
+ {
+ "epoch": 39.0,
+ "grad_norm": 0.01925666816532612,
+ "learning_rate": 0.0006,
+ "loss": 4.126394271850586,
+ "step": 2808
+ },
+ {
+ "epoch": 39.0,
+ "eval_loss": 4.615617275238037,
+ "eval_runtime": 56.8895,
+ "eval_samples_per_second": 42.925,
+ "eval_steps_per_second": 1.354,
+ "step": 2808
+ },
+ {
+ "epoch": 39.01397379912664,
+ "grad_norm": 0.01677853614091873,
+ "learning_rate": 0.0006,
+ "loss": 4.102395057678223,
+ "step": 2809
+ },
+ {
+ "epoch": 39.02794759825328,
+ "grad_norm": 0.017901137471199036,
+ "learning_rate": 0.0006,
+ "loss": 3.976609230041504,
+ "step": 2810
+ },
+ {
+ "epoch": 39.041921397379916,
+ "grad_norm": 0.017978668212890625,
+ "learning_rate": 0.0006,
+ "loss": 3.944416046142578,
+ "step": 2811
+ },
+ {
+ "epoch": 39.05589519650655,
+ "grad_norm": 0.019967492669820786,
+ "learning_rate": 0.0006,
+ "loss": 3.992234468460083,
+ "step": 2812
+ },
+ {
+ "epoch": 39.069868995633186,
+ "grad_norm": 0.018937544897198677,
+ "learning_rate": 0.0006,
+ "loss": 4.028573036193848,
+ "step": 2813
+ },
+ {
+ "epoch": 39.083842794759825,
+ "grad_norm": 0.016852952539920807,
+ "learning_rate": 0.0006,
+ "loss": 4.062442779541016,
+ "step": 2814
+ },
+ {
+ "epoch": 39.09781659388646,
+ "grad_norm": 0.016643647104501724,
+ "learning_rate": 0.0006,
+ "loss": 4.08871603012085,
+ "step": 2815
+ },
+ {
+ "epoch": 39.1117903930131,
+ "grad_norm": 0.016872866079211235,
+ "learning_rate": 0.0006,
+ "loss": 3.9514245986938477,
+ "step": 2816
+ },
+ {
+ "epoch": 39.12576419213974,
+ "grad_norm": 0.018045669421553612,
+ "learning_rate": 0.0006,
+ "loss": 4.038842678070068,
+ "step": 2817
+ },
+ {
+ "epoch": 39.13973799126637,
+ "grad_norm": 0.017853403463959694,
+ "learning_rate": 0.0006,
+ "loss": 3.93709135055542,
+ "step": 2818
+ },
+ {
+ "epoch": 39.15371179039301,
+ "grad_norm": 0.017507443204522133,
+ "learning_rate": 0.0006,
+ "loss": 3.98665714263916,
+ "step": 2819
+ },
+ {
+ "epoch": 39.16768558951965,
+ "grad_norm": 0.01766853965818882,
+ "learning_rate": 0.0006,
+ "loss": 3.937126636505127,
+ "step": 2820
+ },
+ {
+ "epoch": 39.18165938864629,
+ "grad_norm": 0.016660021618008614,
+ "learning_rate": 0.0006,
+ "loss": 3.8397014141082764,
+ "step": 2821
+ },
+ {
+ "epoch": 39.19563318777293,
+ "grad_norm": 0.017719309777021408,
+ "learning_rate": 0.0006,
+ "loss": 4.003579616546631,
+ "step": 2822
+ },
+ {
+ "epoch": 39.209606986899566,
+ "grad_norm": 0.018234414979815483,
+ "learning_rate": 0.0006,
+ "loss": 4.053859710693359,
+ "step": 2823
+ },
+ {
+ "epoch": 39.223580786026204,
+ "grad_norm": 0.018777024000883102,
+ "learning_rate": 0.0006,
+ "loss": 4.056182861328125,
+ "step": 2824
+ },
+ {
+ "epoch": 39.237554585152836,
+ "grad_norm": 0.017443502321839333,
+ "learning_rate": 0.0006,
+ "loss": 4.025068283081055,
+ "step": 2825
+ },
+ {
+ "epoch": 39.251528384279474,
+ "grad_norm": 0.016500331461429596,
+ "learning_rate": 0.0006,
+ "loss": 4.101631164550781,
+ "step": 2826
+ },
+ {
+ "epoch": 39.26550218340611,
+ "grad_norm": 0.016101544722914696,
+ "learning_rate": 0.0006,
+ "loss": 4.027082443237305,
+ "step": 2827
+ },
+ {
+ "epoch": 39.27947598253275,
+ "grad_norm": 0.015471171587705612,
+ "learning_rate": 0.0006,
+ "loss": 4.022909641265869,
+ "step": 2828
+ },
+ {
+ "epoch": 39.29344978165939,
+ "grad_norm": 0.015779748558998108,
+ "learning_rate": 0.0006,
+ "loss": 4.18796968460083,
+ "step": 2829
+ },
+ {
+ "epoch": 39.30742358078603,
+ "grad_norm": 0.015522679314017296,
+ "learning_rate": 0.0006,
+ "loss": 4.005777359008789,
+ "step": 2830
+ },
+ {
+ "epoch": 39.32139737991266,
+ "grad_norm": 0.015226504765450954,
+ "learning_rate": 0.0006,
+ "loss": 4.108416557312012,
+ "step": 2831
+ },
+ {
+ "epoch": 39.3353711790393,
+ "grad_norm": 0.015095778740942478,
+ "learning_rate": 0.0006,
+ "loss": 4.08029842376709,
+ "step": 2832
+ },
+ {
+ "epoch": 39.34934497816594,
+ "grad_norm": 0.015601031482219696,
+ "learning_rate": 0.0006,
+ "loss": 3.989492416381836,
+ "step": 2833
+ },
+ {
+ "epoch": 39.36331877729258,
+ "grad_norm": 0.014825052581727505,
+ "learning_rate": 0.0006,
+ "loss": 3.998520612716675,
+ "step": 2834
+ },
+ {
+ "epoch": 39.377292576419215,
+ "grad_norm": 0.014821100980043411,
+ "learning_rate": 0.0006,
+ "loss": 4.001777172088623,
+ "step": 2835
+ },
+ {
+ "epoch": 39.391266375545854,
+ "grad_norm": 0.015333754941821098,
+ "learning_rate": 0.0006,
+ "loss": 4.11985969543457,
+ "step": 2836
+ },
+ {
+ "epoch": 39.40524017467249,
+ "grad_norm": 0.015940699726343155,
+ "learning_rate": 0.0006,
+ "loss": 4.025373935699463,
+ "step": 2837
+ },
+ {
+ "epoch": 39.419213973799124,
+ "grad_norm": 0.015096287243068218,
+ "learning_rate": 0.0006,
+ "loss": 4.068551063537598,
+ "step": 2838
+ },
+ {
+ "epoch": 39.43318777292576,
+ "grad_norm": 0.01533548254519701,
+ "learning_rate": 0.0006,
+ "loss": 3.9800801277160645,
+ "step": 2839
+ },
+ {
+ "epoch": 39.4471615720524,
+ "grad_norm": 0.014648732729256153,
+ "learning_rate": 0.0006,
+ "loss": 3.9670190811157227,
+ "step": 2840
+ },
+ {
+ "epoch": 39.46113537117904,
+ "grad_norm": 0.015759671106934547,
+ "learning_rate": 0.0006,
+ "loss": 4.050443649291992,
+ "step": 2841
+ },
+ {
+ "epoch": 39.47510917030568,
+ "grad_norm": 0.016054073348641396,
+ "learning_rate": 0.0006,
+ "loss": 4.22734260559082,
+ "step": 2842
+ },
+ {
+ "epoch": 39.48908296943232,
+ "grad_norm": 0.0160264540463686,
+ "learning_rate": 0.0006,
+ "loss": 4.116174221038818,
+ "step": 2843
+ },
+ {
+ "epoch": 39.50305676855895,
+ "grad_norm": 0.015426610596477985,
+ "learning_rate": 0.0006,
+ "loss": 4.000178337097168,
+ "step": 2844
+ },
+ {
+ "epoch": 39.51703056768559,
+ "grad_norm": 0.01520642638206482,
+ "learning_rate": 0.0006,
+ "loss": 4.218410968780518,
+ "step": 2845
+ },
+ {
+ "epoch": 39.531004366812226,
+ "grad_norm": 0.015305949375033379,
+ "learning_rate": 0.0006,
+ "loss": 4.025891304016113,
+ "step": 2846
+ },
+ {
+ "epoch": 39.544978165938865,
+ "grad_norm": 0.015461564064025879,
+ "learning_rate": 0.0006,
+ "loss": 4.100950717926025,
+ "step": 2847
+ },
+ {
+ "epoch": 39.5589519650655,
+ "grad_norm": 0.01712625101208687,
+ "learning_rate": 0.0006,
+ "loss": 4.0620856285095215,
+ "step": 2848
+ },
+ {
+ "epoch": 39.57292576419214,
+ "grad_norm": 0.017133301123976707,
+ "learning_rate": 0.0006,
+ "loss": 4.097138404846191,
+ "step": 2849
+ },
+ {
+ "epoch": 39.58689956331878,
+ "grad_norm": 0.01683625392615795,
+ "learning_rate": 0.0006,
+ "loss": 4.084364891052246,
+ "step": 2850
+ },
+ {
+ "epoch": 39.60087336244541,
+ "grad_norm": 0.0165710486471653,
+ "learning_rate": 0.0006,
+ "loss": 4.042361259460449,
+ "step": 2851
+ },
+ {
+ "epoch": 39.61484716157205,
+ "grad_norm": 0.01660393364727497,
+ "learning_rate": 0.0006,
+ "loss": 4.1395487785339355,
+ "step": 2852
+ },
+ {
+ "epoch": 39.62882096069869,
+ "grad_norm": 0.017025845125317574,
+ "learning_rate": 0.0006,
+ "loss": 4.1207075119018555,
+ "step": 2853
+ },
+ {
+ "epoch": 39.64279475982533,
+ "grad_norm": 0.01660497486591339,
+ "learning_rate": 0.0006,
+ "loss": 4.103846073150635,
+ "step": 2854
+ },
+ {
+ "epoch": 39.65676855895197,
+ "grad_norm": 0.01607946865260601,
+ "learning_rate": 0.0006,
+ "loss": 4.189016342163086,
+ "step": 2855
+ },
+ {
+ "epoch": 39.670742358078606,
+ "grad_norm": 0.01614505983889103,
+ "learning_rate": 0.0006,
+ "loss": 4.079860687255859,
+ "step": 2856
+ },
+ {
+ "epoch": 39.68471615720524,
+ "grad_norm": 0.017085086554288864,
+ "learning_rate": 0.0006,
+ "loss": 3.9774293899536133,
+ "step": 2857
+ },
+ {
+ "epoch": 39.698689956331876,
+ "grad_norm": 0.01816520094871521,
+ "learning_rate": 0.0006,
+ "loss": 4.113666534423828,
+ "step": 2858
+ },
+ {
+ "epoch": 39.712663755458514,
+ "grad_norm": 0.016597526147961617,
+ "learning_rate": 0.0006,
+ "loss": 4.068242073059082,
+ "step": 2859
+ },
+ {
+ "epoch": 39.72663755458515,
+ "grad_norm": 0.01672235131263733,
+ "learning_rate": 0.0006,
+ "loss": 4.024444580078125,
+ "step": 2860
+ },
+ {
+ "epoch": 39.74061135371179,
+ "grad_norm": 0.020093556493520737,
+ "learning_rate": 0.0006,
+ "loss": 4.064124584197998,
+ "step": 2861
+ },
+ {
+ "epoch": 39.75458515283843,
+ "grad_norm": 0.023141304031014442,
+ "learning_rate": 0.0006,
+ "loss": 4.130439758300781,
+ "step": 2862
+ },
+ {
+ "epoch": 39.76855895196506,
+ "grad_norm": 0.02289372682571411,
+ "learning_rate": 0.0006,
+ "loss": 4.14794921875,
+ "step": 2863
+ },
+ {
+ "epoch": 39.7825327510917,
+ "grad_norm": 0.01919020712375641,
+ "learning_rate": 0.0006,
+ "loss": 4.123961925506592,
+ "step": 2864
+ },
+ {
+ "epoch": 39.79650655021834,
+ "grad_norm": 0.016932329162955284,
+ "learning_rate": 0.0006,
+ "loss": 4.0944132804870605,
+ "step": 2865
+ },
+ {
+ "epoch": 39.81048034934498,
+ "grad_norm": 0.016644535586237907,
+ "learning_rate": 0.0006,
+ "loss": 4.124269962310791,
+ "step": 2866
+ },
+ {
+ "epoch": 39.82445414847162,
+ "grad_norm": 0.016302216798067093,
+ "learning_rate": 0.0006,
+ "loss": 4.0267205238342285,
+ "step": 2867
+ },
+ {
+ "epoch": 39.838427947598255,
+ "grad_norm": 0.01673300378024578,
+ "learning_rate": 0.0006,
+ "loss": 4.10650634765625,
+ "step": 2868
+ },
+ {
+ "epoch": 39.852401746724894,
+ "grad_norm": 0.016581356525421143,
+ "learning_rate": 0.0006,
+ "loss": 4.1547956466674805,
+ "step": 2869
+ },
+ {
+ "epoch": 39.866375545851525,
+ "grad_norm": 0.01691797934472561,
+ "learning_rate": 0.0006,
+ "loss": 4.056936740875244,
+ "step": 2870
+ },
+ {
+ "epoch": 39.880349344978164,
+ "grad_norm": 0.01596726104617119,
+ "learning_rate": 0.0006,
+ "loss": 3.9891414642333984,
+ "step": 2871
+ },
+ {
+ "epoch": 39.8943231441048,
+ "grad_norm": 0.015615091659128666,
+ "learning_rate": 0.0006,
+ "loss": 4.137948036193848,
+ "step": 2872
+ },
+ {
+ "epoch": 39.90829694323144,
+ "grad_norm": 0.015703413635492325,
+ "learning_rate": 0.0006,
+ "loss": 3.9844930171966553,
+ "step": 2873
+ },
+ {
+ "epoch": 39.92227074235808,
+ "grad_norm": 0.015453094616532326,
+ "learning_rate": 0.0006,
+ "loss": 4.098443508148193,
+ "step": 2874
+ },
+ {
+ "epoch": 39.93624454148472,
+ "grad_norm": 0.01449158601462841,
+ "learning_rate": 0.0006,
+ "loss": 4.090852737426758,
+ "step": 2875
+ },
+ {
+ "epoch": 39.95021834061135,
+ "grad_norm": 0.01537515688687563,
+ "learning_rate": 0.0006,
+ "loss": 4.239612579345703,
+ "step": 2876
+ },
+ {
+ "epoch": 39.96419213973799,
+ "grad_norm": 0.015989921987056732,
+ "learning_rate": 0.0006,
+ "loss": 4.142073154449463,
+ "step": 2877
+ },
+ {
+ "epoch": 39.97816593886463,
+ "grad_norm": 0.015985611826181412,
+ "learning_rate": 0.0006,
+ "loss": 4.100076675415039,
+ "step": 2878
+ },
+ {
+ "epoch": 39.992139737991266,
+ "grad_norm": 0.015260833315551281,
+ "learning_rate": 0.0006,
+ "loss": 4.068576335906982,
+ "step": 2879
+ },
+ {
+ "epoch": 40.0,
+ "grad_norm": 0.018047941848635674,
+ "learning_rate": 0.0006,
+ "loss": 4.005878925323486,
+ "step": 2880
+ },
+ {
+ "epoch": 40.0,
+ "eval_loss": 4.622848987579346,
+ "eval_runtime": 56.9618,
+ "eval_samples_per_second": 42.871,
+ "eval_steps_per_second": 1.352,
+ "step": 2880
+ },
+ {
+ "epoch": 40.01397379912664,
+ "grad_norm": 0.015976838767528534,
+ "learning_rate": 0.0006,
+ "loss": 4.035689353942871,
+ "step": 2881
+ },
+ {
+ "epoch": 40.02794759825328,
+ "grad_norm": 0.015176220797002316,
+ "learning_rate": 0.0006,
+ "loss": 3.932197332382202,
+ "step": 2882
+ },
+ {
+ "epoch": 40.041921397379916,
+ "grad_norm": 0.016676228493452072,
+ "learning_rate": 0.0006,
+ "loss": 4.014171123504639,
+ "step": 2883
+ },
+ {
+ "epoch": 40.05589519650655,
+ "grad_norm": 0.01698281429708004,
+ "learning_rate": 0.0006,
+ "loss": 4.117893218994141,
+ "step": 2884
+ },
+ {
+ "epoch": 40.069868995633186,
+ "grad_norm": 0.017872486263513565,
+ "learning_rate": 0.0006,
+ "loss": 4.132119178771973,
+ "step": 2885
+ },
+ {
+ "epoch": 40.083842794759825,
+ "grad_norm": 0.017615554854273796,
+ "learning_rate": 0.0006,
+ "loss": 4.072000980377197,
+ "step": 2886
+ },
+ {
+ "epoch": 40.09781659388646,
+ "grad_norm": 0.017469394952058792,
+ "learning_rate": 0.0006,
+ "loss": 4.017726421356201,
+ "step": 2887
+ },
+ {
+ "epoch": 40.1117903930131,
+ "grad_norm": 0.01954762637615204,
+ "learning_rate": 0.0006,
+ "loss": 4.014443874359131,
+ "step": 2888
+ },
+ {
+ "epoch": 40.12576419213974,
+ "grad_norm": 0.018935853615403175,
+ "learning_rate": 0.0006,
+ "loss": 4.0042948722839355,
+ "step": 2889
+ },
+ {
+ "epoch": 40.13973799126637,
+ "grad_norm": 0.019382692873477936,
+ "learning_rate": 0.0006,
+ "loss": 4.06160831451416,
+ "step": 2890
+ },
+ {
+ "epoch": 40.15371179039301,
+ "grad_norm": 0.019677115604281425,
+ "learning_rate": 0.0006,
+ "loss": 4.13466215133667,
+ "step": 2891
+ },
+ {
+ "epoch": 40.16768558951965,
+ "grad_norm": 0.020365122705698013,
+ "learning_rate": 0.0006,
+ "loss": 4.095000267028809,
+ "step": 2892
+ },
+ {
+ "epoch": 40.18165938864629,
+ "grad_norm": 0.021842138841748238,
+ "learning_rate": 0.0006,
+ "loss": 3.9877853393554688,
+ "step": 2893
+ },
+ {
+ "epoch": 40.19563318777293,
+ "grad_norm": 0.02354881353676319,
+ "learning_rate": 0.0006,
+ "loss": 3.9430112838745117,
+ "step": 2894
+ },
+ {
+ "epoch": 40.209606986899566,
+ "grad_norm": 0.026602735742926598,
+ "learning_rate": 0.0006,
+ "loss": 4.012076377868652,
+ "step": 2895
+ },
+ {
+ "epoch": 40.223580786026204,
+ "grad_norm": 0.023976484313607216,
+ "learning_rate": 0.0006,
+ "loss": 3.9768595695495605,
+ "step": 2896
+ },
+ {
+ "epoch": 40.237554585152836,
+ "grad_norm": 0.02336616814136505,
+ "learning_rate": 0.0006,
+ "loss": 4.1096038818359375,
+ "step": 2897
+ },
+ {
+ "epoch": 40.251528384279474,
+ "grad_norm": 0.02426023595035076,
+ "learning_rate": 0.0006,
+ "loss": 4.157189846038818,
+ "step": 2898
+ },
+ {
+ "epoch": 40.26550218340611,
+ "grad_norm": 0.021572696045041084,
+ "learning_rate": 0.0006,
+ "loss": 3.979175090789795,
+ "step": 2899
+ },
+ {
+ "epoch": 40.27947598253275,
+ "grad_norm": 0.08300432562828064,
+ "learning_rate": 0.0006,
+ "loss": 4.092336654663086,
+ "step": 2900
+ },
+ {
+ "epoch": 40.29344978165939,
+ "grad_norm": 0.7323967814445496,
+ "learning_rate": 0.0006,
+ "loss": 5.337990760803223,
+ "step": 2901
+ },
+ {
+ "epoch": 40.30742358078603,
+ "grad_norm": 0.3363337814807892,
+ "learning_rate": 0.0006,
+ "loss": 6.40043830871582,
+ "step": 2902
+ },
+ {
+ "epoch": 40.32139737991266,
+ "grad_norm": 0.598686933517456,
+ "learning_rate": 0.0006,
+ "loss": 7.004765510559082,
+ "step": 2903
+ },
+ {
+ "epoch": 40.3353711790393,
+ "grad_norm": 0.5998364686965942,
+ "learning_rate": 0.0006,
+ "loss": 8.184747695922852,
+ "step": 2904
+ },
+ {
+ "epoch": 40.34934497816594,
+ "grad_norm": 1.408208966255188,
+ "learning_rate": 0.0006,
+ "loss": 8.88978385925293,
+ "step": 2905
+ },
+ {
+ "epoch": 40.36331877729258,
+ "grad_norm": 0.7829785943031311,
+ "learning_rate": 0.0006,
+ "loss": 11.13601303100586,
+ "step": 2906
+ },
+ {
+ "epoch": 40.377292576419215,
+ "grad_norm": 0.7769215106964111,
+ "learning_rate": 0.0006,
+ "loss": 9.90416145324707,
+ "step": 2907
+ },
+ {
+ "epoch": 40.391266375545854,
+ "grad_norm": 0.46508079767227173,
+ "learning_rate": 0.0006,
+ "loss": 9.647209167480469,
+ "step": 2908
+ },
+ {
+ "epoch": 40.40524017467249,
+ "grad_norm": 1.0228968858718872,
+ "learning_rate": 0.0006,
+ "loss": 11.685434341430664,
+ "step": 2909
+ },
+ {
+ "epoch": 40.419213973799124,
+ "grad_norm": 0.20553778111934662,
+ "learning_rate": 0.0006,
+ "loss": 8.621708869934082,
+ "step": 2910
+ },
+ {
+ "epoch": 40.43318777292576,
+ "grad_norm": 0.18466424942016602,
+ "learning_rate": 0.0006,
+ "loss": 8.590787887573242,
+ "step": 2911
+ },
+ {
+ "epoch": 40.4471615720524,
+ "grad_norm": 0.10458594560623169,
+ "learning_rate": 0.0006,
+ "loss": 8.324458122253418,
+ "step": 2912
+ },
+ {
+ "epoch": 40.46113537117904,
+ "grad_norm": 0.1096842959523201,
+ "learning_rate": 0.0006,
+ "loss": 8.300851821899414,
+ "step": 2913
+ },
+ {
+ "epoch": 40.47510917030568,
+ "grad_norm": 0.10501226782798767,
+ "learning_rate": 0.0006,
+ "loss": 8.21009635925293,
+ "step": 2914
+ },
+ {
+ "epoch": 40.48908296943232,
+ "grad_norm": 0.13704165816307068,
+ "learning_rate": 0.0006,
+ "loss": 8.141355514526367,
+ "step": 2915
+ },
+ {
+ "epoch": 40.50305676855895,
+ "grad_norm": 0.12129108607769012,
+ "learning_rate": 0.0006,
+ "loss": 8.074612617492676,
+ "step": 2916
+ },
+ {
+ "epoch": 40.51703056768559,
+ "grad_norm": 0.10028897225856781,
+ "learning_rate": 0.0006,
+ "loss": 8.00384521484375,
+ "step": 2917
+ },
+ {
+ "epoch": 40.531004366812226,
+ "grad_norm": 0.11764246970415115,
+ "learning_rate": 0.0006,
+ "loss": 8.078598022460938,
+ "step": 2918
+ },
+ {
+ "epoch": 40.544978165938865,
+ "grad_norm": 0.10081959515810013,
+ "learning_rate": 0.0006,
+ "loss": 8.008938789367676,
+ "step": 2919
+ },
+ {
+ "epoch": 40.5589519650655,
+ "grad_norm": 0.1003170907497406,
+ "learning_rate": 0.0006,
+ "loss": 7.962940216064453,
+ "step": 2920
+ },
+ {
+ "epoch": 40.57292576419214,
+ "grad_norm": 0.08763492107391357,
+ "learning_rate": 0.0006,
+ "loss": 7.909419536590576,
+ "step": 2921
+ },
+ {
+ "epoch": 40.58689956331878,
+ "grad_norm": 0.0636807233095169,
+ "learning_rate": 0.0006,
+ "loss": 7.878905773162842,
+ "step": 2922
+ },
+ {
+ "epoch": 40.60087336244541,
+ "grad_norm": 0.06005936488509178,
+ "learning_rate": 0.0006,
+ "loss": 7.850980758666992,
+ "step": 2923
+ },
+ {
+ "epoch": 40.61484716157205,
+ "grad_norm": 0.06415296345949173,
+ "learning_rate": 0.0006,
+ "loss": 7.830353260040283,
+ "step": 2924
+ },
+ {
+ "epoch": 40.62882096069869,
+ "grad_norm": 0.0732007846236229,
+ "learning_rate": 0.0006,
+ "loss": 7.830746650695801,
+ "step": 2925
+ },
+ {
+ "epoch": 40.64279475982533,
+ "grad_norm": 0.06628110259771347,
+ "learning_rate": 0.0006,
+ "loss": 7.82661771774292,
+ "step": 2926
+ },
+ {
+ "epoch": 40.65676855895197,
+ "grad_norm": 0.05544218048453331,
+ "learning_rate": 0.0006,
+ "loss": 7.814305782318115,
+ "step": 2927
+ },
+ {
+ "epoch": 40.670742358078606,
+ "grad_norm": 0.050327103585004807,
+ "learning_rate": 0.0006,
+ "loss": 7.764055252075195,
+ "step": 2928
+ },
+ {
+ "epoch": 40.68471615720524,
+ "grad_norm": 0.04918425530195236,
+ "learning_rate": 0.0006,
+ "loss": 7.744625091552734,
+ "step": 2929
+ },
+ {
+ "epoch": 40.698689956331876,
+ "grad_norm": 0.042862266302108765,
+ "learning_rate": 0.0006,
+ "loss": 7.764426231384277,
+ "step": 2930
+ },
+ {
+ "epoch": 40.712663755458514,
+ "grad_norm": 0.04338089004158974,
+ "learning_rate": 0.0006,
+ "loss": 7.712530136108398,
+ "step": 2931
+ },
+ {
+ "epoch": 40.72663755458515,
+ "grad_norm": 0.04127737879753113,
+ "learning_rate": 0.0006,
+ "loss": 7.742180347442627,
+ "step": 2932
+ },
+ {
+ "epoch": 40.74061135371179,
+ "grad_norm": 0.03358631581068039,
+ "learning_rate": 0.0006,
+ "loss": 7.714834213256836,
+ "step": 2933
+ },
+ {
+ "epoch": 40.75458515283843,
+ "grad_norm": 0.02845790982246399,
+ "learning_rate": 0.0006,
+ "loss": 7.7021026611328125,
+ "step": 2934
+ },
+ {
+ "epoch": 40.76855895196506,
+ "grad_norm": 0.03012419492006302,
+ "learning_rate": 0.0006,
+ "loss": 7.712377548217773,
+ "step": 2935
+ },
+ {
+ "epoch": 40.7825327510917,
+ "grad_norm": 0.032981179654598236,
+ "learning_rate": 0.0006,
+ "loss": 7.666385650634766,
+ "step": 2936
+ },
+ {
+ "epoch": 40.79650655021834,
+ "grad_norm": 0.03357074037194252,
+ "learning_rate": 0.0006,
+ "loss": 7.649097919464111,
+ "step": 2937
+ },
+ {
+ "epoch": 40.81048034934498,
+ "grad_norm": 0.030914708971977234,
+ "learning_rate": 0.0006,
+ "loss": 7.649196624755859,
+ "step": 2938
+ },
+ {
+ "epoch": 40.82445414847162,
+ "grad_norm": 0.026296131312847137,
+ "learning_rate": 0.0006,
+ "loss": 7.629964828491211,
+ "step": 2939
+ },
+ {
+ "epoch": 40.838427947598255,
+ "grad_norm": 0.026677245274186134,
+ "learning_rate": 0.0006,
+ "loss": 7.599973678588867,
+ "step": 2940
+ },
+ {
+ "epoch": 40.852401746724894,
+ "grad_norm": 0.023208219558000565,
+ "learning_rate": 0.0006,
+ "loss": 7.601624488830566,
+ "step": 2941
+ },
+ {
+ "epoch": 40.866375545851525,
+ "grad_norm": 0.020741475746035576,
+ "learning_rate": 0.0006,
+ "loss": 7.603902339935303,
+ "step": 2942
+ },
+ {
+ "epoch": 40.880349344978164,
+ "grad_norm": 0.019221751019358635,
+ "learning_rate": 0.0006,
+ "loss": 7.5535888671875,
+ "step": 2943
+ },
+ {
+ "epoch": 40.8943231441048,
+ "grad_norm": 0.019491951912641525,
+ "learning_rate": 0.0006,
+ "loss": 7.54597282409668,
+ "step": 2944
+ },
+ {
+ "epoch": 40.90829694323144,
+ "grad_norm": 0.019702572375535965,
+ "learning_rate": 0.0006,
+ "loss": 7.537567138671875,
+ "step": 2945
+ },
+ {
+ "epoch": 40.92227074235808,
+ "grad_norm": 0.01768800988793373,
+ "learning_rate": 0.0006,
+ "loss": 7.512800216674805,
+ "step": 2946
+ },
+ {
+ "epoch": 40.93624454148472,
+ "grad_norm": 0.01926138624548912,
+ "learning_rate": 0.0006,
+ "loss": 7.509932518005371,
+ "step": 2947
+ },
+ {
+ "epoch": 40.95021834061135,
+ "grad_norm": 0.019030006602406502,
+ "learning_rate": 0.0006,
+ "loss": 7.497983932495117,
+ "step": 2948
+ },
+ {
+ "epoch": 40.96419213973799,
+ "grad_norm": 0.017313841730356216,
+ "learning_rate": 0.0006,
+ "loss": 7.463518142700195,
+ "step": 2949
+ },
+ {
+ "epoch": 40.97816593886463,
+ "grad_norm": 0.01310762483626604,
+ "learning_rate": 0.0006,
+ "loss": 7.474668502807617,
+ "step": 2950
+ },
+ {
+ "epoch": 40.992139737991266,
+ "grad_norm": 0.013008282519876957,
+ "learning_rate": 0.0006,
+ "loss": 7.45982551574707,
+ "step": 2951
+ },
+ {
+ "epoch": 41.0,
+ "grad_norm": 0.01441363524645567,
+ "learning_rate": 0.0006,
+ "loss": 7.446529388427734,
+ "step": 2952
+ },
+ {
+ "epoch": 41.0,
+ "eval_loss": 7.486120223999023,
+ "eval_runtime": 70.4986,
+ "eval_samples_per_second": 34.639,
+ "eval_steps_per_second": 1.092,
+ "step": 2952
+ },
+ {
+ "epoch": 41.01397379912664,
+ "grad_norm": 0.014061855152249336,
+ "learning_rate": 0.0006,
+ "loss": 7.4692182540893555,
+ "step": 2953
+ },
+ {
+ "epoch": 41.02794759825328,
+ "grad_norm": 0.01587500050663948,
+ "learning_rate": 0.0006,
+ "loss": 7.4182024002075195,
+ "step": 2954
+ },
+ {
+ "epoch": 41.041921397379916,
+ "grad_norm": 0.014783984050154686,
+ "learning_rate": 0.0006,
+ "loss": 7.466778755187988,
+ "step": 2955
+ },
+ {
+ "epoch": 41.05589519650655,
+ "grad_norm": 0.011560854502022266,
+ "learning_rate": 0.0006,
+ "loss": 7.438002586364746,
+ "step": 2956
+ },
+ {
+ "epoch": 41.069868995633186,
+ "grad_norm": 0.00970957800745964,
+ "learning_rate": 0.0006,
+ "loss": 7.442694187164307,
+ "step": 2957
+ },
+ {
+ "epoch": 41.083842794759825,
+ "grad_norm": 0.009633459150791168,
+ "learning_rate": 0.0006,
+ "loss": 7.458884239196777,
+ "step": 2958
+ },
+ {
+ "epoch": 41.09781659388646,
+ "grad_norm": 0.010664350353181362,
+ "learning_rate": 0.0006,
+ "loss": 7.48378324508667,
+ "step": 2959
+ },
+ {
+ "epoch": 41.1117903930131,
+ "grad_norm": 0.011440591886639595,
+ "learning_rate": 0.0006,
+ "loss": 7.474484920501709,
+ "step": 2960
+ },
+ {
+ "epoch": 41.12576419213974,
+ "grad_norm": 0.01150908600538969,
+ "learning_rate": 0.0006,
+ "loss": 7.417902946472168,
+ "step": 2961
+ },
+ {
+ "epoch": 41.13973799126637,
+ "grad_norm": 0.010375697165727615,
+ "learning_rate": 0.0006,
+ "loss": 7.403279781341553,
+ "step": 2962
+ },
+ {
+ "epoch": 41.15371179039301,
+ "grad_norm": 0.009075703099370003,
+ "learning_rate": 0.0006,
+ "loss": 7.447347164154053,
+ "step": 2963
+ },
+ {
+ "epoch": 41.16768558951965,
+ "grad_norm": 0.008892177604138851,
+ "learning_rate": 0.0006,
+ "loss": 7.402681350708008,
+ "step": 2964
+ },
+ {
+ "epoch": 41.18165938864629,
+ "grad_norm": 0.009492084383964539,
+ "learning_rate": 0.0006,
+ "loss": 7.3871355056762695,
+ "step": 2965
+ },
+ {
+ "epoch": 41.19563318777293,
+ "grad_norm": 0.009139460511505604,
+ "learning_rate": 0.0006,
+ "loss": 7.419617176055908,
+ "step": 2966
+ },
+ {
+ "epoch": 41.209606986899566,
+ "grad_norm": 0.007959138602018356,
+ "learning_rate": 0.0006,
+ "loss": 7.374949932098389,
+ "step": 2967
+ },
+ {
+ "epoch": 41.223580786026204,
+ "grad_norm": 0.007571548223495483,
+ "learning_rate": 0.0006,
+ "loss": 7.4205241203308105,
+ "step": 2968
+ },
+ {
+ "epoch": 41.237554585152836,
+ "grad_norm": 0.007885400205850601,
+ "learning_rate": 0.0006,
+ "loss": 7.416543006896973,
+ "step": 2969
+ },
+ {
+ "epoch": 41.251528384279474,
+ "grad_norm": 0.008650158531963825,
+ "learning_rate": 0.0006,
+ "loss": 7.357326984405518,
+ "step": 2970
+ },
+ {
+ "epoch": 41.26550218340611,
+ "grad_norm": 0.007483262103050947,
+ "learning_rate": 0.0006,
+ "loss": 7.410891056060791,
+ "step": 2971
+ },
+ {
+ "epoch": 41.27947598253275,
+ "grad_norm": 0.00869277399033308,
+ "learning_rate": 0.0006,
+ "loss": 7.392127990722656,
+ "step": 2972
+ },
+ {
+ "epoch": 41.29344978165939,
+ "grad_norm": 0.00885078776627779,
+ "learning_rate": 0.0006,
+ "loss": 7.405449867248535,
+ "step": 2973
+ },
+ {
+ "epoch": 41.30742358078603,
+ "grad_norm": 0.00958226341754198,
+ "learning_rate": 0.0006,
+ "loss": 7.399868488311768,
+ "step": 2974
+ },
+ {
+ "epoch": 41.32139737991266,
+ "grad_norm": 0.006991599686443806,
+ "learning_rate": 0.0006,
+ "loss": 7.35369873046875,
+ "step": 2975
+ },
+ {
+ "epoch": 41.3353711790393,
+ "grad_norm": 0.007415142375975847,
+ "learning_rate": 0.0006,
+ "loss": 7.358353137969971,
+ "step": 2976
+ },
+ {
+ "epoch": 41.34934497816594,
+ "grad_norm": 0.007024371065199375,
+ "learning_rate": 0.0006,
+ "loss": 7.4070892333984375,
+ "step": 2977
+ },
+ {
+ "epoch": 41.36331877729258,
+ "grad_norm": 0.007234332151710987,
+ "learning_rate": 0.0006,
+ "loss": 7.379120349884033,
+ "step": 2978
+ },
+ {
+ "epoch": 41.377292576419215,
+ "grad_norm": 0.00798892229795456,
+ "learning_rate": 0.0006,
+ "loss": 7.365455627441406,
+ "step": 2979
+ },
+ {
+ "epoch": 41.391266375545854,
+ "grad_norm": 0.007303398102521896,
+ "learning_rate": 0.0006,
+ "loss": 7.341798782348633,
+ "step": 2980
+ },
+ {
+ "epoch": 41.40524017467249,
+ "grad_norm": 0.007119669578969479,
+ "learning_rate": 0.0006,
+ "loss": 7.363280296325684,
+ "step": 2981
+ },
+ {
+ "epoch": 41.419213973799124,
+ "grad_norm": 0.007457619532942772,
+ "learning_rate": 0.0006,
+ "loss": 7.39388370513916,
+ "step": 2982
+ },
+ {
+ "epoch": 41.43318777292576,
+ "grad_norm": 0.007905688136816025,
+ "learning_rate": 0.0006,
+ "loss": 7.380652904510498,
+ "step": 2983
+ },
+ {
+ "epoch": 41.4471615720524,
+ "grad_norm": 0.006212855689227581,
+ "learning_rate": 0.0006,
+ "loss": 7.398769378662109,
+ "step": 2984
+ },
+ {
+ "epoch": 41.46113537117904,
+ "grad_norm": 0.0070641543716192245,
+ "learning_rate": 0.0006,
+ "loss": 7.3503546714782715,
+ "step": 2985
+ },
+ {
+ "epoch": 41.47510917030568,
+ "grad_norm": 0.00787805300205946,
+ "learning_rate": 0.0006,
+ "loss": 7.347678184509277,
+ "step": 2986
+ },
+ {
+ "epoch": 41.48908296943232,
+ "grad_norm": 0.006760426796972752,
+ "learning_rate": 0.0006,
+ "loss": 7.380006790161133,
+ "step": 2987
+ },
+ {
+ "epoch": 41.50305676855895,
+ "grad_norm": 0.00772385997697711,
+ "learning_rate": 0.0006,
+ "loss": 7.363680362701416,
+ "step": 2988
+ },
+ {
+ "epoch": 41.51703056768559,
+ "grad_norm": 0.007373564876616001,
+ "learning_rate": 0.0006,
+ "loss": 7.333014488220215,
+ "step": 2989
+ },
+ {
+ "epoch": 41.531004366812226,
+ "grad_norm": 0.006717653013765812,
+ "learning_rate": 0.0006,
+ "loss": 7.360076427459717,
+ "step": 2990
+ },
+ {
+ "epoch": 41.544978165938865,
+ "grad_norm": 0.007191413082182407,
+ "learning_rate": 0.0006,
+ "loss": 7.397030830383301,
+ "step": 2991
+ },
+ {
+ "epoch": 41.5589519650655,
+ "grad_norm": 0.007831960916519165,
+ "learning_rate": 0.0006,
+ "loss": 7.369959831237793,
+ "step": 2992
+ },
+ {
+ "epoch": 41.57292576419214,
+ "grad_norm": 0.006947481073439121,
+ "learning_rate": 0.0006,
+ "loss": 7.333254337310791,
+ "step": 2993
+ },
+ {
+ "epoch": 41.58689956331878,
+ "grad_norm": 0.006243606563657522,
+ "learning_rate": 0.0006,
+ "loss": 7.327378273010254,
+ "step": 2994
+ },
+ {
+ "epoch": 41.60087336244541,
+ "grad_norm": 0.006298339460045099,
+ "learning_rate": 0.0006,
+ "loss": 7.323482036590576,
+ "step": 2995
+ },
+ {
+ "epoch": 41.61484716157205,
+ "grad_norm": 0.007587412837892771,
+ "learning_rate": 0.0006,
+ "loss": 7.313789367675781,
+ "step": 2996
+ },
+ {
+ "epoch": 41.62882096069869,
+ "grad_norm": 0.007140746805816889,
+ "learning_rate": 0.0006,
+ "loss": 7.356067657470703,
+ "step": 2997
+ },
+ {
+ "epoch": 41.64279475982533,
+ "grad_norm": 0.008315625600516796,
+ "learning_rate": 0.0006,
+ "loss": 7.3339643478393555,
+ "step": 2998
+ },
+ {
+ "epoch": 41.65676855895197,
+ "grad_norm": 0.006547102238982916,
+ "learning_rate": 0.0006,
+ "loss": 7.342191696166992,
+ "step": 2999
+ },
+ {
+ "epoch": 41.670742358078606,
+ "grad_norm": 0.006068122107535601,
+ "learning_rate": 0.0006,
+ "loss": 7.321015357971191,
+ "step": 3000
+ },
+ {
+ "epoch": 41.68471615720524,
+ "grad_norm": 0.007316991221159697,
+ "learning_rate": 0.0006,
+ "loss": 7.346490859985352,
+ "step": 3001
+ },
+ {
+ "epoch": 41.698689956331876,
+ "grad_norm": 0.0067388564348220825,
+ "learning_rate": 0.0006,
+ "loss": 7.340723991394043,
+ "step": 3002
+ },
+ {
+ "epoch": 41.712663755458514,
+ "grad_norm": 0.006839963141828775,
+ "learning_rate": 0.0006,
+ "loss": 7.329365253448486,
+ "step": 3003
+ },
+ {
+ "epoch": 41.72663755458515,
+ "grad_norm": 0.00805041566491127,
+ "learning_rate": 0.0006,
+ "loss": 7.298934459686279,
+ "step": 3004
+ },
+ {
+ "epoch": 41.74061135371179,
+ "grad_norm": 0.007197503466159105,
+ "learning_rate": 0.0006,
+ "loss": 7.331496238708496,
+ "step": 3005
+ },
+ {
+ "epoch": 41.75458515283843,
+ "grad_norm": 0.007220805156975985,
+ "learning_rate": 0.0006,
+ "loss": 7.317695140838623,
+ "step": 3006
+ },
+ {
+ "epoch": 41.76855895196506,
+ "grad_norm": 0.00874276738613844,
+ "learning_rate": 0.0006,
+ "loss": 7.272614479064941,
+ "step": 3007
+ },
+ {
+ "epoch": 41.7825327510917,
+ "grad_norm": 0.008386565372347832,
+ "learning_rate": 0.0006,
+ "loss": 7.295244216918945,
+ "step": 3008
+ },
+ {
+ "epoch": 41.79650655021834,
+ "grad_norm": 0.0077730282209813595,
+ "learning_rate": 0.0006,
+ "loss": 7.3395490646362305,
+ "step": 3009
+ },
+ {
+ "epoch": 41.81048034934498,
+ "grad_norm": 0.008441480807960033,
+ "learning_rate": 0.0006,
+ "loss": 7.319921493530273,
+ "step": 3010
+ },
+ {
+ "epoch": 41.82445414847162,
+ "grad_norm": 0.007398269604891539,
+ "learning_rate": 0.0006,
+ "loss": 7.3029465675354,
+ "step": 3011
+ },
+ {
+ "epoch": 41.838427947598255,
+ "grad_norm": 0.006878677289932966,
+ "learning_rate": 0.0006,
+ "loss": 7.318467140197754,
+ "step": 3012
+ },
+ {
+ "epoch": 41.852401746724894,
+ "grad_norm": 0.007616598624736071,
+ "learning_rate": 0.0006,
+ "loss": 7.3112077713012695,
+ "step": 3013
+ },
+ {
+ "epoch": 41.866375545851525,
+ "grad_norm": 0.011629699729382992,
+ "learning_rate": 0.0006,
+ "loss": 7.32523250579834,
+ "step": 3014
+ },
+ {
+ "epoch": 41.880349344978164,
+ "grad_norm": 0.011011249385774136,
+ "learning_rate": 0.0006,
+ "loss": 7.287247657775879,
+ "step": 3015
+ },
+ {
+ "epoch": 41.8943231441048,
+ "grad_norm": 0.010208295658230782,
+ "learning_rate": 0.0006,
+ "loss": 7.292117118835449,
+ "step": 3016
+ },
+ {
+ "epoch": 41.90829694323144,
+ "grad_norm": 0.008651095442473888,
+ "learning_rate": 0.0006,
+ "loss": 7.302201271057129,
+ "step": 3017
+ },
+ {
+ "epoch": 41.92227074235808,
+ "grad_norm": 0.006692920345813036,
+ "learning_rate": 0.0006,
+ "loss": 7.319568634033203,
+ "step": 3018
+ },
+ {
+ "epoch": 41.93624454148472,
+ "grad_norm": 0.010033348575234413,
+ "learning_rate": 0.0006,
+ "loss": 7.2738566398620605,
+ "step": 3019
+ },
+ {
+ "epoch": 41.95021834061135,
+ "grad_norm": 0.011790832504630089,
+ "learning_rate": 0.0006,
+ "loss": 7.298540115356445,
+ "step": 3020
+ },
+ {
+ "epoch": 41.96419213973799,
+ "grad_norm": 0.018622662872076035,
+ "learning_rate": 0.0006,
+ "loss": 7.267182350158691,
+ "step": 3021
+ },
+ {
+ "epoch": 41.97816593886463,
+ "grad_norm": 0.02520257607102394,
+ "learning_rate": 0.0006,
+ "loss": 7.261566162109375,
+ "step": 3022
+ },
+ {
+ "epoch": 41.992139737991266,
+ "grad_norm": 0.036332644522190094,
+ "learning_rate": 0.0006,
+ "loss": 7.270719051361084,
+ "step": 3023
+ },
+ {
+ "epoch": 42.0,
+ "grad_norm": 0.04717305675148964,
+ "learning_rate": 0.0006,
+ "loss": 7.293332099914551,
+ "step": 3024
+ },
+ {
+ "epoch": 42.0,
+ "eval_loss": 7.354280471801758,
+ "eval_runtime": 56.4353,
+ "eval_samples_per_second": 43.271,
+ "eval_steps_per_second": 1.364,
+ "step": 3024
+ },
+ {
+ "epoch": 42.01397379912664,
+ "grad_norm": 0.06661626696586609,
+ "learning_rate": 0.0006,
+ "loss": 7.338189601898193,
+ "step": 3025
+ },
+ {
+ "epoch": 42.02794759825328,
+ "grad_norm": 0.05452054366469383,
+ "learning_rate": 0.0006,
+ "loss": 7.316764831542969,
+ "step": 3026
+ },
+ {
+ "epoch": 42.041921397379916,
+ "grad_norm": 0.025638775900006294,
+ "learning_rate": 0.0006,
+ "loss": 7.34204626083374,
+ "step": 3027
+ },
+ {
+ "epoch": 42.05589519650655,
+ "grad_norm": 0.022398799657821655,
+ "learning_rate": 0.0006,
+ "loss": 7.262149333953857,
+ "step": 3028
+ },
+ {
+ "epoch": 42.069868995633186,
+ "grad_norm": 0.03713499382138252,
+ "learning_rate": 0.0006,
+ "loss": 7.27695369720459,
+ "step": 3029
+ },
+ {
+ "epoch": 42.083842794759825,
+ "grad_norm": 0.014949247241020203,
+ "learning_rate": 0.0006,
+ "loss": 7.296508312225342,
+ "step": 3030
+ },
+ {
+ "epoch": 42.09781659388646,
+ "grad_norm": 0.02703995071351528,
+ "learning_rate": 0.0006,
+ "loss": 7.258018493652344,
+ "step": 3031
+ },
+ {
+ "epoch": 42.1117903930131,
+ "grad_norm": 0.020352816209197044,
+ "learning_rate": 0.0006,
+ "loss": 7.241450309753418,
+ "step": 3032
+ },
+ {
+ "epoch": 42.12576419213974,
+ "grad_norm": 0.014578479342162609,
+ "learning_rate": 0.0006,
+ "loss": 7.257323265075684,
+ "step": 3033
+ },
+ {
+ "epoch": 42.13973799126637,
+ "grad_norm": 0.026053044945001602,
+ "learning_rate": 0.0006,
+ "loss": 7.254680633544922,
+ "step": 3034
+ },
+ {
+ "epoch": 42.15371179039301,
+ "grad_norm": 0.015120275318622589,
+ "learning_rate": 0.0006,
+ "loss": 7.261074066162109,
+ "step": 3035
+ },
+ {
+ "epoch": 42.16768558951965,
+ "grad_norm": 0.016822177916765213,
+ "learning_rate": 0.0006,
+ "loss": 7.279064178466797,
+ "step": 3036
+ },
+ {
+ "epoch": 42.18165938864629,
+ "grad_norm": 0.016367588192224503,
+ "learning_rate": 0.0006,
+ "loss": 7.251429557800293,
+ "step": 3037
+ },
+ {
+ "epoch": 42.19563318777293,
+ "grad_norm": 0.014350418001413345,
+ "learning_rate": 0.0006,
+ "loss": 7.259264945983887,
+ "step": 3038
+ },
+ {
+ "epoch": 42.209606986899566,
+ "grad_norm": 0.017078228294849396,
+ "learning_rate": 0.0006,
+ "loss": 7.227105140686035,
+ "step": 3039
+ },
+ {
+ "epoch": 42.223580786026204,
+ "grad_norm": 0.011759317480027676,
+ "learning_rate": 0.0006,
+ "loss": 7.275888442993164,
+ "step": 3040
+ },
+ {
+ "epoch": 42.237554585152836,
+ "grad_norm": 0.013019340112805367,
+ "learning_rate": 0.0006,
+ "loss": 7.2517008781433105,
+ "step": 3041
+ },
+ {
+ "epoch": 42.251528384279474,
+ "grad_norm": 0.01178441196680069,
+ "learning_rate": 0.0006,
+ "loss": 7.23576545715332,
+ "step": 3042
+ },
+ {
+ "epoch": 42.26550218340611,
+ "grad_norm": 0.014518891461193562,
+ "learning_rate": 0.0006,
+ "loss": 7.253895282745361,
+ "step": 3043
+ },
+ {
+ "epoch": 42.27947598253275,
+ "grad_norm": 0.018159950152039528,
+ "learning_rate": 0.0006,
+ "loss": 7.278058052062988,
+ "step": 3044
+ },
+ {
+ "epoch": 42.29344978165939,
+ "grad_norm": 0.007587796077132225,
+ "learning_rate": 0.0006,
+ "loss": 7.278578758239746,
+ "step": 3045
+ },
+ {
+ "epoch": 42.30742358078603,
+ "grad_norm": 0.015994979068636894,
+ "learning_rate": 0.0006,
+ "loss": 7.239841461181641,
+ "step": 3046
+ },
+ {
+ "epoch": 42.32139737991266,
+ "grad_norm": 0.017830168828368187,
+ "learning_rate": 0.0006,
+ "loss": 7.2443976402282715,
+ "step": 3047
+ },
+ {
+ "epoch": 42.3353711790393,
+ "grad_norm": 0.008738759905099869,
+ "learning_rate": 0.0006,
+ "loss": 7.232264518737793,
+ "step": 3048
+ },
+ {
+ "epoch": 42.34934497816594,
+ "grad_norm": 0.0175449438393116,
+ "learning_rate": 0.0006,
+ "loss": 7.228000640869141,
+ "step": 3049
+ },
+ {
+ "epoch": 42.36331877729258,
+ "grad_norm": 0.019611820578575134,
+ "learning_rate": 0.0006,
+ "loss": 7.234999179840088,
+ "step": 3050
+ },
+ {
+ "epoch": 42.377292576419215,
+ "grad_norm": 0.015302863903343678,
+ "learning_rate": 0.0006,
+ "loss": 7.242547988891602,
+ "step": 3051
+ },
+ {
+ "epoch": 42.391266375545854,
+ "grad_norm": 0.025919731706380844,
+ "learning_rate": 0.0006,
+ "loss": 7.216652870178223,
+ "step": 3052
+ },
+ {
+ "epoch": 42.40524017467249,
+ "grad_norm": 0.034587468951940536,
+ "learning_rate": 0.0006,
+ "loss": 7.245336532592773,
+ "step": 3053
+ },
+ {
+ "epoch": 42.419213973799124,
+ "grad_norm": 0.019198840484023094,
+ "learning_rate": 0.0006,
+ "loss": 7.256758689880371,
+ "step": 3054
+ },
+ {
+ "epoch": 42.43318777292576,
+ "grad_norm": 0.012316389009356499,
+ "learning_rate": 0.0006,
+ "loss": 7.209102630615234,
+ "step": 3055
+ },
+ {
+ "epoch": 42.4471615720524,
+ "grad_norm": 0.023512162268161774,
+ "learning_rate": 0.0006,
+ "loss": 7.250742435455322,
+ "step": 3056
+ },
+ {
+ "epoch": 42.46113537117904,
+ "grad_norm": 0.025457851588726044,
+ "learning_rate": 0.0006,
+ "loss": 7.263757705688477,
+ "step": 3057
+ },
+ {
+ "epoch": 42.47510917030568,
+ "grad_norm": 0.02192617766559124,
+ "learning_rate": 0.0006,
+ "loss": 7.246977806091309,
+ "step": 3058
+ },
+ {
+ "epoch": 42.48908296943232,
+ "grad_norm": 0.02541281282901764,
+ "learning_rate": 0.0006,
+ "loss": 7.232361316680908,
+ "step": 3059
+ },
+ {
+ "epoch": 42.50305676855895,
+ "grad_norm": 0.036353763192892075,
+ "learning_rate": 0.0006,
+ "loss": 7.242556571960449,
+ "step": 3060
+ },
+ {
+ "epoch": 42.51703056768559,
+ "grad_norm": 0.04555872827768326,
+ "learning_rate": 0.0006,
+ "loss": 7.228940963745117,
+ "step": 3061
+ },
+ {
+ "epoch": 42.531004366812226,
+ "grad_norm": 0.04563378915190697,
+ "learning_rate": 0.0006,
+ "loss": 7.235276222229004,
+ "step": 3062
+ },
+ {
+ "epoch": 42.544978165938865,
+ "grad_norm": 0.04275760427117348,
+ "learning_rate": 0.0006,
+ "loss": 7.238895416259766,
+ "step": 3063
+ },
+ {
+ "epoch": 42.5589519650655,
+ "grad_norm": 0.04535885155200958,
+ "learning_rate": 0.0006,
+ "loss": 7.205336570739746,
+ "step": 3064
+ },
+ {
+ "epoch": 42.57292576419214,
+ "grad_norm": 0.015534617006778717,
+ "learning_rate": 0.0006,
+ "loss": 7.238393783569336,
+ "step": 3065
+ },
+ {
+ "epoch": 42.58689956331878,
+ "grad_norm": 0.03286299481987953,
+ "learning_rate": 0.0006,
+ "loss": 7.216453552246094,
+ "step": 3066
+ },
+ {
+ "epoch": 42.60087336244541,
+ "grad_norm": 0.025453168898820877,
+ "learning_rate": 0.0006,
+ "loss": 7.203750133514404,
+ "step": 3067
+ },
+ {
+ "epoch": 42.61484716157205,
+ "grad_norm": 0.021501649171113968,
+ "learning_rate": 0.0006,
+ "loss": 7.241621971130371,
+ "step": 3068
+ },
+ {
+ "epoch": 42.62882096069869,
+ "grad_norm": 0.02743634767830372,
+ "learning_rate": 0.0006,
+ "loss": 7.22076940536499,
+ "step": 3069
+ },
+ {
+ "epoch": 42.64279475982533,
+ "grad_norm": 0.020890483632683754,
+ "learning_rate": 0.0006,
+ "loss": 7.218412399291992,
+ "step": 3070
+ },
+ {
+ "epoch": 42.65676855895197,
+ "grad_norm": 0.022499850019812584,
+ "learning_rate": 0.0006,
+ "loss": 7.217696189880371,
+ "step": 3071
+ },
+ {
+ "epoch": 42.670742358078606,
+ "grad_norm": 0.01653696969151497,
+ "learning_rate": 0.0006,
+ "loss": 7.156417369842529,
+ "step": 3072
+ },
+ {
+ "epoch": 42.68471615720524,
+ "grad_norm": 0.01834225095808506,
+ "learning_rate": 0.0006,
+ "loss": 7.211247444152832,
+ "step": 3073
+ },
+ {
+ "epoch": 42.698689956331876,
+ "grad_norm": 0.020137229934334755,
+ "learning_rate": 0.0006,
+ "loss": 7.232499599456787,
+ "step": 3074
+ },
+ {
+ "epoch": 42.712663755458514,
+ "grad_norm": 0.012671459466218948,
+ "learning_rate": 0.0006,
+ "loss": 7.172441482543945,
+ "step": 3075
+ },
+ {
+ "epoch": 42.72663755458515,
+ "grad_norm": 0.016803130507469177,
+ "learning_rate": 0.0006,
+ "loss": 7.152103424072266,
+ "step": 3076
+ },
+ {
+ "epoch": 42.74061135371179,
+ "grad_norm": 0.01623712107539177,
+ "learning_rate": 0.0006,
+ "loss": 7.187032699584961,
+ "step": 3077
+ },
+ {
+ "epoch": 42.75458515283843,
+ "grad_norm": 0.015938356518745422,
+ "learning_rate": 0.0006,
+ "loss": 7.171140670776367,
+ "step": 3078
+ },
+ {
+ "epoch": 42.76855895196506,
+ "grad_norm": 0.02580736204981804,
+ "learning_rate": 0.0006,
+ "loss": 7.204472541809082,
+ "step": 3079
+ },
+ {
+ "epoch": 42.7825327510917,
+ "grad_norm": 0.023372117429971695,
+ "learning_rate": 0.0006,
+ "loss": 7.165554523468018,
+ "step": 3080
+ },
+ {
+ "epoch": 42.79650655021834,
+ "grad_norm": 0.04440586268901825,
+ "learning_rate": 0.0006,
+ "loss": 7.169431209564209,
+ "step": 3081
+ },
+ {
+ "epoch": 42.81048034934498,
+ "grad_norm": 0.06797406077384949,
+ "learning_rate": 0.0006,
+ "loss": 7.168193340301514,
+ "step": 3082
+ },
+ {
+ "epoch": 42.82445414847162,
+ "grad_norm": 0.0914321169257164,
+ "learning_rate": 0.0006,
+ "loss": 7.204671859741211,
+ "step": 3083
+ },
+ {
+ "epoch": 42.838427947598255,
+ "grad_norm": 0.0506940595805645,
+ "learning_rate": 0.0006,
+ "loss": 7.15949821472168,
+ "step": 3084
+ },
+ {
+ "epoch": 42.852401746724894,
+ "grad_norm": 0.04819890484213829,
+ "learning_rate": 0.0006,
+ "loss": 7.192837238311768,
+ "step": 3085
+ },
+ {
+ "epoch": 42.866375545851525,
+ "grad_norm": 0.040044233202934265,
+ "learning_rate": 0.0006,
+ "loss": 7.1600165367126465,
+ "step": 3086
+ },
+ {
+ "epoch": 42.880349344978164,
+ "grad_norm": 0.02573299966752529,
+ "learning_rate": 0.0006,
+ "loss": 7.153427600860596,
+ "step": 3087
+ },
+ {
+ "epoch": 42.8943231441048,
+ "grad_norm": 0.046942535787820816,
+ "learning_rate": 0.0006,
+ "loss": 7.16767692565918,
+ "step": 3088
+ },
+ {
+ "epoch": 42.90829694323144,
+ "grad_norm": 0.02683161571621895,
+ "learning_rate": 0.0006,
+ "loss": 7.17144250869751,
+ "step": 3089
+ },
+ {
+ "epoch": 42.92227074235808,
+ "grad_norm": 0.05363674834370613,
+ "learning_rate": 0.0006,
+ "loss": 7.120128631591797,
+ "step": 3090
+ },
+ {
+ "epoch": 42.93624454148472,
+ "grad_norm": 0.023756247013807297,
+ "learning_rate": 0.0006,
+ "loss": 7.169703006744385,
+ "step": 3091
+ },
+ {
+ "epoch": 42.95021834061135,
+ "grad_norm": 0.04216783866286278,
+ "learning_rate": 0.0006,
+ "loss": 7.1206865310668945,
+ "step": 3092
+ },
+ {
+ "epoch": 42.96419213973799,
+ "grad_norm": 0.02219565585255623,
+ "learning_rate": 0.0006,
+ "loss": 7.157215118408203,
+ "step": 3093
+ },
+ {
+ "epoch": 42.97816593886463,
+ "grad_norm": 0.03527127206325531,
+ "learning_rate": 0.0006,
+ "loss": 7.166672706604004,
+ "step": 3094
+ },
+ {
+ "epoch": 42.992139737991266,
+ "grad_norm": 0.017748817801475525,
+ "learning_rate": 0.0006,
+ "loss": 7.154747009277344,
+ "step": 3095
+ },
+ {
+ "epoch": 43.0,
+ "grad_norm": 0.033165838569402695,
+ "learning_rate": 0.0006,
+ "loss": 7.150866508483887,
+ "step": 3096
+ },
+ {
+ "epoch": 43.0,
+ "eval_loss": 7.144403457641602,
+ "eval_runtime": 56.8853,
+ "eval_samples_per_second": 42.928,
+ "eval_steps_per_second": 1.354,
+ "step": 3096
+ },
+ {
+ "epoch": 43.01397379912664,
+ "grad_norm": 0.015199829824268818,
+ "learning_rate": 0.0006,
+ "loss": 7.122735023498535,
+ "step": 3097
+ },
+ {
+ "epoch": 43.02794759825328,
+ "grad_norm": 0.03220256045460701,
+ "learning_rate": 0.0006,
+ "loss": 7.143553733825684,
+ "step": 3098
+ },
+ {
+ "epoch": 43.041921397379916,
+ "grad_norm": 0.014687920920550823,
+ "learning_rate": 0.0006,
+ "loss": 7.080960750579834,
+ "step": 3099
+ },
+ {
+ "epoch": 43.05589519650655,
+ "grad_norm": 0.023522932082414627,
+ "learning_rate": 0.0006,
+ "loss": 7.137356758117676,
+ "step": 3100
+ },
+ {
+ "epoch": 43.069868995633186,
+ "grad_norm": 0.014808216132223606,
+ "learning_rate": 0.0006,
+ "loss": 7.09143590927124,
+ "step": 3101
+ },
+ {
+ "epoch": 43.083842794759825,
+ "grad_norm": 0.029339246451854706,
+ "learning_rate": 0.0006,
+ "loss": 7.122198104858398,
+ "step": 3102
+ },
+ {
+ "epoch": 43.09781659388646,
+ "grad_norm": 0.022953078150749207,
+ "learning_rate": 0.0006,
+ "loss": 7.092505931854248,
+ "step": 3103
+ },
+ {
+ "epoch": 43.1117903930131,
+ "grad_norm": 0.04590052738785744,
+ "learning_rate": 0.0006,
+ "loss": 7.113100051879883,
+ "step": 3104
+ },
+ {
+ "epoch": 43.12576419213974,
+ "grad_norm": 0.0524442233145237,
+ "learning_rate": 0.0006,
+ "loss": 7.147022247314453,
+ "step": 3105
+ },
+ {
+ "epoch": 43.13973799126637,
+ "grad_norm": 0.0315794013440609,
+ "learning_rate": 0.0006,
+ "loss": 7.094045162200928,
+ "step": 3106
+ },
+ {
+ "epoch": 43.15371179039301,
+ "grad_norm": 0.05112722888588905,
+ "learning_rate": 0.0006,
+ "loss": 7.054388523101807,
+ "step": 3107
+ },
+ {
+ "epoch": 43.16768558951965,
+ "grad_norm": 0.03146776929497719,
+ "learning_rate": 0.0006,
+ "loss": 7.116674423217773,
+ "step": 3108
+ },
+ {
+ "epoch": 43.18165938864629,
+ "grad_norm": 0.03391246497631073,
+ "learning_rate": 0.0006,
+ "loss": 7.150345802307129,
+ "step": 3109
+ },
+ {
+ "epoch": 43.19563318777293,
+ "grad_norm": 0.026957903057336807,
+ "learning_rate": 0.0006,
+ "loss": 7.058531284332275,
+ "step": 3110
+ },
+ {
+ "epoch": 43.209606986899566,
+ "grad_norm": 0.041532568633556366,
+ "learning_rate": 0.0006,
+ "loss": 7.085498809814453,
+ "step": 3111
+ },
+ {
+ "epoch": 43.223580786026204,
+ "grad_norm": 0.020001588389277458,
+ "learning_rate": 0.0006,
+ "loss": 7.131487846374512,
+ "step": 3112
+ },
+ {
+ "epoch": 43.237554585152836,
+ "grad_norm": 0.02268712781369686,
+ "learning_rate": 0.0006,
+ "loss": 7.0790863037109375,
+ "step": 3113
+ },
+ {
+ "epoch": 43.251528384279474,
+ "grad_norm": 0.023363925516605377,
+ "learning_rate": 0.0006,
+ "loss": 7.067816734313965,
+ "step": 3114
+ },
+ {
+ "epoch": 43.26550218340611,
+ "grad_norm": 0.019367508590221405,
+ "learning_rate": 0.0006,
+ "loss": 7.042041778564453,
+ "step": 3115
+ },
+ {
+ "epoch": 43.27947598253275,
+ "grad_norm": 0.02745090238749981,
+ "learning_rate": 0.0006,
+ "loss": 7.064733028411865,
+ "step": 3116
+ },
+ {
+ "epoch": 43.29344978165939,
+ "grad_norm": 0.02543802745640278,
+ "learning_rate": 0.0006,
+ "loss": 7.058948040008545,
+ "step": 3117
+ },
+ {
+ "epoch": 43.30742358078603,
+ "grad_norm": 0.018874092027544975,
+ "learning_rate": 0.0006,
+ "loss": 7.009946346282959,
+ "step": 3118
+ },
+ {
+ "epoch": 43.32139737991266,
+ "grad_norm": 0.024662822484970093,
+ "learning_rate": 0.0006,
+ "loss": 7.015913963317871,
+ "step": 3119
+ },
+ {
+ "epoch": 43.3353711790393,
+ "grad_norm": 0.02441416308283806,
+ "learning_rate": 0.0006,
+ "loss": 7.061622142791748,
+ "step": 3120
+ },
+ {
+ "epoch": 43.34934497816594,
+ "grad_norm": 0.031887322664260864,
+ "learning_rate": 0.0006,
+ "loss": 7.046883583068848,
+ "step": 3121
+ },
+ {
+ "epoch": 43.36331877729258,
+ "grad_norm": 0.021930739283561707,
+ "learning_rate": 0.0006,
+ "loss": 7.0309624671936035,
+ "step": 3122
+ },
+ {
+ "epoch": 43.377292576419215,
+ "grad_norm": 0.02722783200442791,
+ "learning_rate": 0.0006,
+ "loss": 7.072924613952637,
+ "step": 3123
+ },
+ {
+ "epoch": 43.391266375545854,
+ "grad_norm": 0.01483179721981287,
+ "learning_rate": 0.0006,
+ "loss": 7.028624534606934,
+ "step": 3124
+ },
+ {
+ "epoch": 43.40524017467249,
+ "grad_norm": 0.024793945252895355,
+ "learning_rate": 0.0006,
+ "loss": 7.001623153686523,
+ "step": 3125
+ },
+ {
+ "epoch": 43.419213973799124,
+ "grad_norm": 0.020148713141679764,
+ "learning_rate": 0.0006,
+ "loss": 6.9938812255859375,
+ "step": 3126
+ },
+ {
+ "epoch": 43.43318777292576,
+ "grad_norm": 0.02413850650191307,
+ "learning_rate": 0.0006,
+ "loss": 7.001243591308594,
+ "step": 3127
+ },
+ {
+ "epoch": 43.4471615720524,
+ "grad_norm": 0.028549203649163246,
+ "learning_rate": 0.0006,
+ "loss": 6.987671375274658,
+ "step": 3128
+ },
+ {
+ "epoch": 43.46113537117904,
+ "grad_norm": 0.025988785549998283,
+ "learning_rate": 0.0006,
+ "loss": 6.9551167488098145,
+ "step": 3129
+ },
+ {
+ "epoch": 43.47510917030568,
+ "grad_norm": 0.027140729129314423,
+ "learning_rate": 0.0006,
+ "loss": 6.9547810554504395,
+ "step": 3130
+ },
+ {
+ "epoch": 43.48908296943232,
+ "grad_norm": 0.030130082741379738,
+ "learning_rate": 0.0006,
+ "loss": 6.993186950683594,
+ "step": 3131
+ },
+ {
+ "epoch": 43.50305676855895,
+ "grad_norm": 0.025996994227170944,
+ "learning_rate": 0.0006,
+ "loss": 6.984624862670898,
+ "step": 3132
+ },
+ {
+ "epoch": 43.51703056768559,
+ "grad_norm": 0.02423008903861046,
+ "learning_rate": 0.0006,
+ "loss": 6.98813009262085,
+ "step": 3133
+ },
+ {
+ "epoch": 43.531004366812226,
+ "grad_norm": 0.029665837064385414,
+ "learning_rate": 0.0006,
+ "loss": 6.986610412597656,
+ "step": 3134
+ },
+ {
+ "epoch": 43.544978165938865,
+ "grad_norm": 0.028768595308065414,
+ "learning_rate": 0.0006,
+ "loss": 6.898797035217285,
+ "step": 3135
+ },
+ {
+ "epoch": 43.5589519650655,
+ "grad_norm": 0.03111208975315094,
+ "learning_rate": 0.0006,
+ "loss": 6.923741817474365,
+ "step": 3136
+ },
+ {
+ "epoch": 43.57292576419214,
+ "grad_norm": 0.0368187390267849,
+ "learning_rate": 0.0006,
+ "loss": 6.994343280792236,
+ "step": 3137
+ },
+ {
+ "epoch": 43.58689956331878,
+ "grad_norm": 0.04076763987541199,
+ "learning_rate": 0.0006,
+ "loss": 6.909458160400391,
+ "step": 3138
+ },
+ {
+ "epoch": 43.60087336244541,
+ "grad_norm": 0.03281061723828316,
+ "learning_rate": 0.0006,
+ "loss": 6.924851417541504,
+ "step": 3139
+ },
+ {
+ "epoch": 43.61484716157205,
+ "grad_norm": 0.025486988946795464,
+ "learning_rate": 0.0006,
+ "loss": 6.962926864624023,
+ "step": 3140
+ },
+ {
+ "epoch": 43.62882096069869,
+ "grad_norm": 0.04583485797047615,
+ "learning_rate": 0.0006,
+ "loss": 6.876555442810059,
+ "step": 3141
+ },
+ {
+ "epoch": 43.64279475982533,
+ "grad_norm": 0.05460785701870918,
+ "learning_rate": 0.0006,
+ "loss": 6.952932357788086,
+ "step": 3142
+ },
+ {
+ "epoch": 43.65676855895197,
+ "grad_norm": 0.040917906910181046,
+ "learning_rate": 0.0006,
+ "loss": 6.9624924659729,
+ "step": 3143
+ },
+ {
+ "epoch": 43.670742358078606,
+ "grad_norm": 0.04009668901562691,
+ "learning_rate": 0.0006,
+ "loss": 6.94033145904541,
+ "step": 3144
+ },
+ {
+ "epoch": 43.68471615720524,
+ "grad_norm": 0.04500625655055046,
+ "learning_rate": 0.0006,
+ "loss": 6.961839199066162,
+ "step": 3145
+ },
+ {
+ "epoch": 43.698689956331876,
+ "grad_norm": 0.03182258829474449,
+ "learning_rate": 0.0006,
+ "loss": 6.930580139160156,
+ "step": 3146
+ },
+ {
+ "epoch": 43.712663755458514,
+ "grad_norm": 0.053621504455804825,
+ "learning_rate": 0.0006,
+ "loss": 6.950235366821289,
+ "step": 3147
+ },
+ {
+ "epoch": 43.72663755458515,
+ "grad_norm": 0.031919289380311966,
+ "learning_rate": 0.0006,
+ "loss": 6.895111560821533,
+ "step": 3148
+ },
+ {
+ "epoch": 43.74061135371179,
+ "grad_norm": 0.042713891714811325,
+ "learning_rate": 0.0006,
+ "loss": 6.940885543823242,
+ "step": 3149
+ },
+ {
+ "epoch": 43.75458515283843,
+ "grad_norm": 0.02385653182864189,
+ "learning_rate": 0.0006,
+ "loss": 6.948801517486572,
+ "step": 3150
+ },
+ {
+ "epoch": 43.76855895196506,
+ "grad_norm": 0.04100663587450981,
+ "learning_rate": 0.0006,
+ "loss": 6.970411777496338,
+ "step": 3151
+ },
+ {
+ "epoch": 43.7825327510917,
+ "grad_norm": 0.02729583904147148,
+ "learning_rate": 0.0006,
+ "loss": 6.899379253387451,
+ "step": 3152
+ },
+ {
+ "epoch": 43.79650655021834,
+ "grad_norm": 0.03128620237112045,
+ "learning_rate": 0.0006,
+ "loss": 6.931319236755371,
+ "step": 3153
+ },
+ {
+ "epoch": 43.81048034934498,
+ "grad_norm": 0.025488659739494324,
+ "learning_rate": 0.0006,
+ "loss": 6.8486409187316895,
+ "step": 3154
+ },
+ {
+ "epoch": 43.82445414847162,
+ "grad_norm": 0.03679376840591431,
+ "learning_rate": 0.0006,
+ "loss": 6.840811252593994,
+ "step": 3155
+ },
+ {
+ "epoch": 43.838427947598255,
+ "grad_norm": 0.020578034222126007,
+ "learning_rate": 0.0006,
+ "loss": 6.9088616371154785,
+ "step": 3156
+ },
+ {
+ "epoch": 43.852401746724894,
+ "grad_norm": 0.02935841865837574,
+ "learning_rate": 0.0006,
+ "loss": 6.806676864624023,
+ "step": 3157
+ },
+ {
+ "epoch": 43.866375545851525,
+ "grad_norm": 0.023319244384765625,
+ "learning_rate": 0.0006,
+ "loss": 6.878457546234131,
+ "step": 3158
+ },
+ {
+ "epoch": 43.880349344978164,
+ "grad_norm": 0.02779589593410492,
+ "learning_rate": 0.0006,
+ "loss": 6.822447776794434,
+ "step": 3159
+ },
+ {
+ "epoch": 43.8943231441048,
+ "grad_norm": 0.01879825070500374,
+ "learning_rate": 0.0006,
+ "loss": 6.854121208190918,
+ "step": 3160
+ },
+ {
+ "epoch": 43.90829694323144,
+ "grad_norm": 0.023286139592528343,
+ "learning_rate": 0.0006,
+ "loss": 6.836827278137207,
+ "step": 3161
+ },
+ {
+ "epoch": 43.92227074235808,
+ "grad_norm": 0.030358297750353813,
+ "learning_rate": 0.0006,
+ "loss": 6.860037803649902,
+ "step": 3162
+ },
+ {
+ "epoch": 43.93624454148472,
+ "grad_norm": 0.029621105641126633,
+ "learning_rate": 0.0006,
+ "loss": 6.884025573730469,
+ "step": 3163
+ },
+ {
+ "epoch": 43.95021834061135,
+ "grad_norm": 0.029829824343323708,
+ "learning_rate": 0.0006,
+ "loss": 6.8702392578125,
+ "step": 3164
+ },
+ {
+ "epoch": 43.96419213973799,
+ "grad_norm": 0.036892473697662354,
+ "learning_rate": 0.0006,
+ "loss": 6.806652069091797,
+ "step": 3165
+ },
+ {
+ "epoch": 43.97816593886463,
+ "grad_norm": 0.03206709027290344,
+ "learning_rate": 0.0006,
+ "loss": 6.828495979309082,
+ "step": 3166
+ },
+ {
+ "epoch": 43.992139737991266,
+ "grad_norm": 0.02744000405073166,
+ "learning_rate": 0.0006,
+ "loss": 6.832002639770508,
+ "step": 3167
+ },
+ {
+ "epoch": 44.0,
+ "grad_norm": 0.02205660194158554,
+ "learning_rate": 0.0006,
+ "loss": 6.898828506469727,
+ "step": 3168
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 28800,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 400,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": true,
+ "should_log": false,
+ "should_save": false,
+ "should_training_stop": false
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 1.3506298789861786e+18,
+ "train_batch_size": 8,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-3168/training_args.bin b/runs/l2r50-i2-fulle-lm/checkpoint-3168/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..cad232831c222cc7f8b938b8bbe42f61e7d89c3a
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-3168/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:f7c40e3bfa2f5ac1450ca1243e3f2a895d44a06ed5995e6aefd97f78d229d8c8
+size 4792
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-3456/chat_template.jinja b/runs/l2r50-i2-fulle-lm/checkpoint-3456/chat_template.jinja
new file mode 100644
index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-3456/chat_template.jinja
@@ -0,0 +1,85 @@
+{%- if tools %}
+ {{- '<|im_start|>system\n' }}
+ {%- if messages[0].role == 'system' %}
+ {{- messages[0].content + '\n\n' }}
+ {%- endif %}
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }}
+ {%- for tool in tools %}
+ {{- "\n" }}
+ {{- tool | tojson }}
+ {%- endfor %}
+ {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }}
+{%- else %}
+ {%- if messages[0].role == 'system' %}
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
+ {%- endif %}
+{%- endif %}
+{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
+{%- for message in messages[::-1] %}
+ {%- set index = (messages|length - 1) - loop.index0 %}
+ {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %}
+ {%- set ns.multi_step_tool = false %}
+ {%- set ns.last_query_index = index %}
+ {%- endif %}
+{%- endfor %}
+{%- for message in messages %}
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
+ {%- elif message.role == "assistant" %}
+ {%- set content = message.content %}
+ {%- set reasoning_content = '' %}
+ {%- if message.reasoning_content is defined and message.reasoning_content is not none %}
+ {%- set reasoning_content = message.reasoning_content %}
+ {%- else %}
+ {%- if '' in message.content %}
+ {%- set content = message.content.split('')[-1].lstrip('\n') %}
+ {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %}
+ {%- endif %}
+ {%- endif %}
+ {%- if loop.index0 > ns.last_query_index %}
+ {%- if loop.last or (not loop.last and reasoning_content) %}
+ {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- if message.tool_calls %}
+ {%- for tool_call in message.tool_calls %}
+ {%- if (loop.first and content) or (not loop.first) %}
+ {{- '\n' }}
+ {%- endif %}
+ {%- if tool_call.function %}
+ {%- set tool_call = tool_call.function %}
+ {%- endif %}
+ {{- '\n{"name": "' }}
+ {{- tool_call.name }}
+ {{- '", "arguments": ' }}
+ {%- if tool_call.arguments is string %}
+ {{- tool_call.arguments }}
+ {%- else %}
+ {{- tool_call.arguments | tojson }}
+ {%- endif %}
+ {{- '}\n' }}
+ {%- endfor %}
+ {%- endif %}
+ {{- '<|im_end|>\n' }}
+ {%- elif message.role == "tool" %}
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
+ {{- '<|im_start|>user' }}
+ {%- endif %}
+ {{- '\n\n' }}
+ {{- message.content }}
+ {{- '\n' }}
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
+ {{- '<|im_end|>\n' }}
+ {%- endif %}
+ {%- endif %}
+{%- endfor %}
+{%- if add_generation_prompt %}
+ {{- '<|im_start|>assistant\n' }}
+ {%- if enable_thinking is defined and enable_thinking is false %}
+ {{- '\n\n\n\n' }}
+ {%- endif %}
+{%- endif %}
\ No newline at end of file
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-3456/config.json b/runs/l2r50-i2-fulle-lm/checkpoint-3456/config.json
new file mode 100644
index 0000000000000000000000000000000000000000..22f250953c6ce8fc91ee4411856dd73041a415e2
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-3456/config.json
@@ -0,0 +1,32 @@
+{
+ "architectures": [
+ "LlamaForCausalLM"
+ ],
+ "attention_bias": false,
+ "attention_dropout": 0.0,
+ "bos_token_id": null,
+ "dtype": "float32",
+ "eos_token_id": 151645,
+ "head_dim": 128,
+ "hidden_act": "silu",
+ "hidden_size": 512,
+ "initializer_range": 0.02,
+ "intermediate_size": 1536,
+ "max_position_embeddings": 2048,
+ "mlp_bias": false,
+ "model_type": "llama",
+ "num_attention_heads": 4,
+ "num_hidden_layers": 20,
+ "num_key_value_heads": 4,
+ "pad_token_id": 151645,
+ "pretraining_tp": 1,
+ "rms_norm_eps": 1e-06,
+ "rope_parameters": {
+ "rope_theta": 10000.0,
+ "rope_type": "default"
+ },
+ "tie_word_embeddings": true,
+ "transformers_version": "5.5.0",
+ "use_cache": false,
+ "vocab_size": 151673
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-3456/generation_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-3456/generation_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-3456/generation_config.json
@@ -0,0 +1,11 @@
+{
+ "_from_model_config": true,
+ "eos_token_id": [
+ 151645
+ ],
+ "output_attentions": false,
+ "output_hidden_states": false,
+ "pad_token_id": 151645,
+ "transformers_version": "5.5.0",
+ "use_cache": true
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-3456/model.safetensors b/runs/l2r50-i2-fulle-lm/checkpoint-3456/model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..dc71bb4ad1af2d7517c8abc07c9a2e8720061ef9
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-3456/model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:49654181507b5a62db885d2846536aee3275e38c88c34f4ebd1baf7d457f9ec6
+size 583360328
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-3456/optimizer.pt b/runs/l2r50-i2-fulle-lm/checkpoint-3456/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..11e11a544a685e991419998111d4098433f55c76
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-3456/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:01192f68bd9ec49948989234584a715cc3eb1e5cca3271bedf3d0144aa1e34a4
+size 1166833530
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-3456/rng_state_0.pth b/runs/l2r50-i2-fulle-lm/checkpoint-3456/rng_state_0.pth
new file mode 100644
index 0000000000000000000000000000000000000000..53ddb8e1764017752759402da1b95202b3baf1e9
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-3456/rng_state_0.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:3dcb6653f8376fc990944caeb4b9eb7aae01fb54a7598cdc206beb81b83cb2ff
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-3456/rng_state_1.pth b/runs/l2r50-i2-fulle-lm/checkpoint-3456/rng_state_1.pth
new file mode 100644
index 0000000000000000000000000000000000000000..cc84b3583e27ed3fb5357880884ee8b623a61531
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-3456/rng_state_1.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:4cd30828cea7b57651e4ff7a62b2441b5b0fe5521a3553e12bcdcf4d7da20691
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-3456/rng_state_2.pth b/runs/l2r50-i2-fulle-lm/checkpoint-3456/rng_state_2.pth
new file mode 100644
index 0000000000000000000000000000000000000000..7d556d66f69473a5f0c0dbde9bba5ed9b61a5c4f
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-3456/rng_state_2.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:614c6993219e973b42f4e87c341515ecd80e83ed9ee205cbf61430d4c2357d3c
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-3456/rng_state_3.pth b/runs/l2r50-i2-fulle-lm/checkpoint-3456/rng_state_3.pth
new file mode 100644
index 0000000000000000000000000000000000000000..29050e842dcc4976e1dd2e1877e62d2227ea3c2f
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-3456/rng_state_3.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:b2ca26b0a3f3294ddb7d6e30792aae3898215fe7ebe6d34ae1498c30ac0653ff
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-3456/scheduler.pt b/runs/l2r50-i2-fulle-lm/checkpoint-3456/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..6f299b34b59cabea3196797084c0dd37275dcaa8
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-3456/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:2a30063d3a42c95cfa4d505a809336bc072a917980dcfbeb28708fcb4baba115
+size 1064
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-3456/tokenizer.json b/runs/l2r50-i2-fulle-lm/checkpoint-3456/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..53786016f070e11184813f3c83101dc5e93e392e
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-3456/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:c3c63fda1646a555448e8f03662542e04b11d103721319e4d9b267d7da03ceb8
+size 11424463
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-3456/tokenizer_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-3456/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..98e2fb9b2cc54785dc9d871b8e503f8b8f2defe7
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-3456/tokenizer_config.json
@@ -0,0 +1,20 @@
+{
+ "add_prefix_space": false,
+ "backend": "tokenizers",
+ "bos_token": null,
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "<|im_end|>",
+ "errors": "replace",
+ "extra_special_tokens": [
+ "<|l2r_pred|>",
+ "<|r2l_pred|>",
+ "<|next_1_pred|>",
+ "<|next_2_pred|>"
+ ],
+ "is_local": false,
+ "model_max_length": 131072,
+ "pad_token": "<|im_end|>",
+ "split_special_tokens": false,
+ "tokenizer_class": "Qwen2Tokenizer",
+ "unk_token": null
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-3456/trainer_state.json b/runs/l2r50-i2-fulle-lm/checkpoint-3456/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..d01427abd2f15400b92d1e83ffb55f200a570544
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-3456/trainer_state.json
@@ -0,0 +1,24602 @@
+{
+ "best_global_step": null,
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 48.0,
+ "eval_steps": 500,
+ "global_step": 3456,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.013973799126637555,
+ "grad_norm": 0.1739506870508194,
+ "learning_rate": 0.0,
+ "loss": 11.99299430847168,
+ "step": 1
+ },
+ {
+ "epoch": 0.02794759825327511,
+ "grad_norm": 0.17216597497463226,
+ "learning_rate": 5.999999999999999e-06,
+ "loss": 11.988704681396484,
+ "step": 2
+ },
+ {
+ "epoch": 0.04192139737991266,
+ "grad_norm": 0.17459715902805328,
+ "learning_rate": 1.1999999999999999e-05,
+ "loss": 11.945581436157227,
+ "step": 3
+ },
+ {
+ "epoch": 0.05589519650655022,
+ "grad_norm": 0.17577414214611053,
+ "learning_rate": 1.7999999999999997e-05,
+ "loss": 11.859772682189941,
+ "step": 4
+ },
+ {
+ "epoch": 0.06986899563318777,
+ "grad_norm": 0.16163651645183563,
+ "learning_rate": 2.3999999999999997e-05,
+ "loss": 11.744415283203125,
+ "step": 5
+ },
+ {
+ "epoch": 0.08384279475982533,
+ "grad_norm": 0.14409932494163513,
+ "learning_rate": 2.9999999999999997e-05,
+ "loss": 11.618099212646484,
+ "step": 6
+ },
+ {
+ "epoch": 0.09781659388646288,
+ "grad_norm": 0.13079848885536194,
+ "learning_rate": 3.5999999999999994e-05,
+ "loss": 11.505935668945312,
+ "step": 7
+ },
+ {
+ "epoch": 0.11179039301310044,
+ "grad_norm": 0.11964613944292068,
+ "learning_rate": 4.2e-05,
+ "loss": 11.416532516479492,
+ "step": 8
+ },
+ {
+ "epoch": 0.125764192139738,
+ "grad_norm": 0.11345142871141434,
+ "learning_rate": 4.7999999999999994e-05,
+ "loss": 11.33206558227539,
+ "step": 9
+ },
+ {
+ "epoch": 0.13973799126637554,
+ "grad_norm": 0.11033952981233597,
+ "learning_rate": 5.399999999999999e-05,
+ "loss": 11.259052276611328,
+ "step": 10
+ },
+ {
+ "epoch": 0.1537117903930131,
+ "grad_norm": 0.10776489973068237,
+ "learning_rate": 5.9999999999999995e-05,
+ "loss": 11.203384399414062,
+ "step": 11
+ },
+ {
+ "epoch": 0.16768558951965065,
+ "grad_norm": 0.10721946507692337,
+ "learning_rate": 6.599999999999999e-05,
+ "loss": 11.153295516967773,
+ "step": 12
+ },
+ {
+ "epoch": 0.18165938864628822,
+ "grad_norm": 0.10673430562019348,
+ "learning_rate": 7.199999999999999e-05,
+ "loss": 11.110200881958008,
+ "step": 13
+ },
+ {
+ "epoch": 0.19563318777292577,
+ "grad_norm": 0.10630254447460175,
+ "learning_rate": 7.8e-05,
+ "loss": 11.067477226257324,
+ "step": 14
+ },
+ {
+ "epoch": 0.2096069868995633,
+ "grad_norm": 0.10631615668535233,
+ "learning_rate": 8.4e-05,
+ "loss": 11.022941589355469,
+ "step": 15
+ },
+ {
+ "epoch": 0.22358078602620088,
+ "grad_norm": 0.10609443485736847,
+ "learning_rate": 8.999999999999999e-05,
+ "loss": 10.970746040344238,
+ "step": 16
+ },
+ {
+ "epoch": 0.23755458515283842,
+ "grad_norm": 0.10648232698440552,
+ "learning_rate": 9.599999999999999e-05,
+ "loss": 10.91267204284668,
+ "step": 17
+ },
+ {
+ "epoch": 0.251528384279476,
+ "grad_norm": 0.10584589838981628,
+ "learning_rate": 0.000102,
+ "loss": 10.854547500610352,
+ "step": 18
+ },
+ {
+ "epoch": 0.26550218340611353,
+ "grad_norm": 0.10591720044612885,
+ "learning_rate": 0.00010799999999999998,
+ "loss": 10.788917541503906,
+ "step": 19
+ },
+ {
+ "epoch": 0.2794759825327511,
+ "grad_norm": 0.10611672699451447,
+ "learning_rate": 0.00011399999999999999,
+ "loss": 10.717556953430176,
+ "step": 20
+ },
+ {
+ "epoch": 0.2934497816593886,
+ "grad_norm": 0.10455359518527985,
+ "learning_rate": 0.00011999999999999999,
+ "loss": 10.650186538696289,
+ "step": 21
+ },
+ {
+ "epoch": 0.3074235807860262,
+ "grad_norm": 0.1047525629401207,
+ "learning_rate": 0.00012599999999999997,
+ "loss": 10.56808090209961,
+ "step": 22
+ },
+ {
+ "epoch": 0.32139737991266376,
+ "grad_norm": 0.10511605441570282,
+ "learning_rate": 0.00013199999999999998,
+ "loss": 10.484028816223145,
+ "step": 23
+ },
+ {
+ "epoch": 0.3353711790393013,
+ "grad_norm": 0.10467177629470825,
+ "learning_rate": 0.000138,
+ "loss": 10.399947166442871,
+ "step": 24
+ },
+ {
+ "epoch": 0.34934497816593885,
+ "grad_norm": 0.10432136058807373,
+ "learning_rate": 0.00014399999999999998,
+ "loss": 10.312714576721191,
+ "step": 25
+ },
+ {
+ "epoch": 0.36331877729257644,
+ "grad_norm": 0.10527262091636658,
+ "learning_rate": 0.00015,
+ "loss": 10.212261199951172,
+ "step": 26
+ },
+ {
+ "epoch": 0.377292576419214,
+ "grad_norm": 0.103610560297966,
+ "learning_rate": 0.000156,
+ "loss": 10.132034301757812,
+ "step": 27
+ },
+ {
+ "epoch": 0.39126637554585153,
+ "grad_norm": 0.10553059726953506,
+ "learning_rate": 0.000162,
+ "loss": 10.020434379577637,
+ "step": 28
+ },
+ {
+ "epoch": 0.4052401746724891,
+ "grad_norm": 0.1037299782037735,
+ "learning_rate": 0.000168,
+ "loss": 9.934466361999512,
+ "step": 29
+ },
+ {
+ "epoch": 0.4192139737991266,
+ "grad_norm": 0.1042739674448967,
+ "learning_rate": 0.00017399999999999997,
+ "loss": 9.822539329528809,
+ "step": 30
+ },
+ {
+ "epoch": 0.4331877729257642,
+ "grad_norm": 0.10300865024328232,
+ "learning_rate": 0.00017999999999999998,
+ "loss": 9.734994888305664,
+ "step": 31
+ },
+ {
+ "epoch": 0.44716157205240176,
+ "grad_norm": 0.10291654616594315,
+ "learning_rate": 0.000186,
+ "loss": 9.624505996704102,
+ "step": 32
+ },
+ {
+ "epoch": 0.4611353711790393,
+ "grad_norm": 0.10210851579904556,
+ "learning_rate": 0.00019199999999999998,
+ "loss": 9.518205642700195,
+ "step": 33
+ },
+ {
+ "epoch": 0.47510917030567684,
+ "grad_norm": 0.10222529619932175,
+ "learning_rate": 0.000198,
+ "loss": 9.412222862243652,
+ "step": 34
+ },
+ {
+ "epoch": 0.4890829694323144,
+ "grad_norm": 0.09992019087076187,
+ "learning_rate": 0.000204,
+ "loss": 9.326566696166992,
+ "step": 35
+ },
+ {
+ "epoch": 0.503056768558952,
+ "grad_norm": 0.09931084513664246,
+ "learning_rate": 0.00020999999999999998,
+ "loss": 9.210111618041992,
+ "step": 36
+ },
+ {
+ "epoch": 0.5170305676855895,
+ "grad_norm": 0.09928256273269653,
+ "learning_rate": 0.00021599999999999996,
+ "loss": 9.100493431091309,
+ "step": 37
+ },
+ {
+ "epoch": 0.5310043668122271,
+ "grad_norm": 0.09843816608190536,
+ "learning_rate": 0.00022199999999999998,
+ "loss": 8.996543884277344,
+ "step": 38
+ },
+ {
+ "epoch": 0.5449781659388646,
+ "grad_norm": 0.09825374186038971,
+ "learning_rate": 0.00022799999999999999,
+ "loss": 8.888991355895996,
+ "step": 39
+ },
+ {
+ "epoch": 0.5589519650655022,
+ "grad_norm": 0.09602385759353638,
+ "learning_rate": 0.000234,
+ "loss": 8.802907943725586,
+ "step": 40
+ },
+ {
+ "epoch": 0.5729257641921397,
+ "grad_norm": 0.09414064139127731,
+ "learning_rate": 0.00023999999999999998,
+ "loss": 8.715423583984375,
+ "step": 41
+ },
+ {
+ "epoch": 0.5868995633187772,
+ "grad_norm": 0.09273893386125565,
+ "learning_rate": 0.00024599999999999996,
+ "loss": 8.61374282836914,
+ "step": 42
+ },
+ {
+ "epoch": 0.6008733624454149,
+ "grad_norm": 0.09079599380493164,
+ "learning_rate": 0.00025199999999999995,
+ "loss": 8.502904891967773,
+ "step": 43
+ },
+ {
+ "epoch": 0.6148471615720524,
+ "grad_norm": 0.08605825901031494,
+ "learning_rate": 0.000258,
+ "loss": 8.453008651733398,
+ "step": 44
+ },
+ {
+ "epoch": 0.62882096069869,
+ "grad_norm": 0.08338996022939682,
+ "learning_rate": 0.00026399999999999997,
+ "loss": 8.365163803100586,
+ "step": 45
+ },
+ {
+ "epoch": 0.6427947598253275,
+ "grad_norm": 0.08150867372751236,
+ "learning_rate": 0.00027,
+ "loss": 8.285372734069824,
+ "step": 46
+ },
+ {
+ "epoch": 0.6567685589519651,
+ "grad_norm": 0.07794807851314545,
+ "learning_rate": 0.000276,
+ "loss": 8.185811996459961,
+ "step": 47
+ },
+ {
+ "epoch": 0.6707423580786026,
+ "grad_norm": 0.07592720538377762,
+ "learning_rate": 0.00028199999999999997,
+ "loss": 8.098482131958008,
+ "step": 48
+ },
+ {
+ "epoch": 0.6847161572052402,
+ "grad_norm": 0.0709897130727768,
+ "learning_rate": 0.00028799999999999995,
+ "loss": 8.042091369628906,
+ "step": 49
+ },
+ {
+ "epoch": 0.6986899563318777,
+ "grad_norm": 0.06582305580377579,
+ "learning_rate": 0.000294,
+ "loss": 7.972844123840332,
+ "step": 50
+ },
+ {
+ "epoch": 0.7126637554585152,
+ "grad_norm": 0.059452351182699203,
+ "learning_rate": 0.0003,
+ "loss": 7.924579620361328,
+ "step": 51
+ },
+ {
+ "epoch": 0.7266375545851529,
+ "grad_norm": 0.052426718175411224,
+ "learning_rate": 0.00030599999999999996,
+ "loss": 7.876336097717285,
+ "step": 52
+ },
+ {
+ "epoch": 0.7406113537117904,
+ "grad_norm": 0.04768150672316551,
+ "learning_rate": 0.000312,
+ "loss": 7.8102707862854,
+ "step": 53
+ },
+ {
+ "epoch": 0.754585152838428,
+ "grad_norm": 0.03970075398683548,
+ "learning_rate": 0.000318,
+ "loss": 7.791913986206055,
+ "step": 54
+ },
+ {
+ "epoch": 0.7685589519650655,
+ "grad_norm": 0.03559752553701401,
+ "learning_rate": 0.000324,
+ "loss": 7.756960868835449,
+ "step": 55
+ },
+ {
+ "epoch": 0.7825327510917031,
+ "grad_norm": 0.029321353882551193,
+ "learning_rate": 0.00033,
+ "loss": 7.724878787994385,
+ "step": 56
+ },
+ {
+ "epoch": 0.7965065502183406,
+ "grad_norm": 0.023485427722334862,
+ "learning_rate": 0.000336,
+ "loss": 7.72553014755249,
+ "step": 57
+ },
+ {
+ "epoch": 0.8104803493449781,
+ "grad_norm": 0.019319474697113037,
+ "learning_rate": 0.00034199999999999996,
+ "loss": 7.709493160247803,
+ "step": 58
+ },
+ {
+ "epoch": 0.8244541484716157,
+ "grad_norm": 0.016705691814422607,
+ "learning_rate": 0.00034799999999999995,
+ "loss": 7.6873297691345215,
+ "step": 59
+ },
+ {
+ "epoch": 0.8384279475982532,
+ "grad_norm": 0.017667734995484352,
+ "learning_rate": 0.00035399999999999993,
+ "loss": 7.677059173583984,
+ "step": 60
+ },
+ {
+ "epoch": 0.8524017467248908,
+ "grad_norm": 0.01671992614865303,
+ "learning_rate": 0.00035999999999999997,
+ "loss": 7.643512725830078,
+ "step": 61
+ },
+ {
+ "epoch": 0.8663755458515284,
+ "grad_norm": 0.012288172729313374,
+ "learning_rate": 0.00036599999999999995,
+ "loss": 7.658343315124512,
+ "step": 62
+ },
+ {
+ "epoch": 0.880349344978166,
+ "grad_norm": 0.012004776857793331,
+ "learning_rate": 0.000372,
+ "loss": 7.649759292602539,
+ "step": 63
+ },
+ {
+ "epoch": 0.8943231441048035,
+ "grad_norm": 0.012487445957958698,
+ "learning_rate": 0.00037799999999999997,
+ "loss": 7.631503105163574,
+ "step": 64
+ },
+ {
+ "epoch": 0.9082969432314411,
+ "grad_norm": 0.01489241048693657,
+ "learning_rate": 0.00038399999999999996,
+ "loss": 7.647110462188721,
+ "step": 65
+ },
+ {
+ "epoch": 0.9222707423580786,
+ "grad_norm": 0.01428980939090252,
+ "learning_rate": 0.00039,
+ "loss": 7.636983871459961,
+ "step": 66
+ },
+ {
+ "epoch": 0.9362445414847161,
+ "grad_norm": 0.010711144655942917,
+ "learning_rate": 0.000396,
+ "loss": 7.621130466461182,
+ "step": 67
+ },
+ {
+ "epoch": 0.9502183406113537,
+ "grad_norm": 0.010178886353969574,
+ "learning_rate": 0.000402,
+ "loss": 7.593924522399902,
+ "step": 68
+ },
+ {
+ "epoch": 0.9641921397379912,
+ "grad_norm": 0.010604874230921268,
+ "learning_rate": 0.000408,
+ "loss": 7.6015496253967285,
+ "step": 69
+ },
+ {
+ "epoch": 0.9781659388646288,
+ "grad_norm": 0.009751426056027412,
+ "learning_rate": 0.0004139999999999999,
+ "loss": 7.571964263916016,
+ "step": 70
+ },
+ {
+ "epoch": 0.9921397379912664,
+ "grad_norm": 0.009816481731832027,
+ "learning_rate": 0.00041999999999999996,
+ "loss": 7.57344913482666,
+ "step": 71
+ },
+ {
+ "epoch": 1.0,
+ "grad_norm": 0.008665637113153934,
+ "learning_rate": 0.00042599999999999995,
+ "loss": 7.579050064086914,
+ "step": 72
+ },
+ {
+ "epoch": 1.0,
+ "eval_loss": 7.581204891204834,
+ "eval_runtime": 59.0091,
+ "eval_samples_per_second": 41.383,
+ "eval_steps_per_second": 1.305,
+ "step": 72
+ },
+ {
+ "epoch": 1.0139737991266375,
+ "grad_norm": 0.009684201329946518,
+ "learning_rate": 0.00043199999999999993,
+ "loss": 7.568594932556152,
+ "step": 73
+ },
+ {
+ "epoch": 1.027947598253275,
+ "grad_norm": 0.025783710181713104,
+ "learning_rate": 0.00043799999999999997,
+ "loss": 7.578948497772217,
+ "step": 74
+ },
+ {
+ "epoch": 1.0419213973799126,
+ "grad_norm": 0.0327819399535656,
+ "learning_rate": 0.00044399999999999995,
+ "loss": 7.548173904418945,
+ "step": 75
+ },
+ {
+ "epoch": 1.0558951965065502,
+ "grad_norm": 0.021604498848319054,
+ "learning_rate": 0.00045,
+ "loss": 7.552953720092773,
+ "step": 76
+ },
+ {
+ "epoch": 1.0698689956331877,
+ "grad_norm": 0.010230105370283127,
+ "learning_rate": 0.00045599999999999997,
+ "loss": 7.5213212966918945,
+ "step": 77
+ },
+ {
+ "epoch": 1.0838427947598253,
+ "grad_norm": 0.023315589874982834,
+ "learning_rate": 0.00046199999999999995,
+ "loss": 7.501640319824219,
+ "step": 78
+ },
+ {
+ "epoch": 1.0978165938864628,
+ "grad_norm": 0.01109944935888052,
+ "learning_rate": 0.000468,
+ "loss": 7.539519786834717,
+ "step": 79
+ },
+ {
+ "epoch": 1.1117903930131003,
+ "grad_norm": 0.026093758642673492,
+ "learning_rate": 0.000474,
+ "loss": 7.525928974151611,
+ "step": 80
+ },
+ {
+ "epoch": 1.125764192139738,
+ "grad_norm": 0.03274895250797272,
+ "learning_rate": 0.00047999999999999996,
+ "loss": 7.535602569580078,
+ "step": 81
+ },
+ {
+ "epoch": 1.1397379912663754,
+ "grad_norm": 0.01543602254241705,
+ "learning_rate": 0.000486,
+ "loss": 7.513683795928955,
+ "step": 82
+ },
+ {
+ "epoch": 1.1537117903930132,
+ "grad_norm": 0.028377249836921692,
+ "learning_rate": 0.0004919999999999999,
+ "loss": 7.49384880065918,
+ "step": 83
+ },
+ {
+ "epoch": 1.1676855895196507,
+ "grad_norm": 0.04121481627225876,
+ "learning_rate": 0.000498,
+ "loss": 7.488447189331055,
+ "step": 84
+ },
+ {
+ "epoch": 1.1816593886462883,
+ "grad_norm": 0.013282302767038345,
+ "learning_rate": 0.0005039999999999999,
+ "loss": 7.463010787963867,
+ "step": 85
+ },
+ {
+ "epoch": 1.1956331877729258,
+ "grad_norm": 0.03200364485383034,
+ "learning_rate": 0.0005099999999999999,
+ "loss": 7.473515033721924,
+ "step": 86
+ },
+ {
+ "epoch": 1.2096069868995634,
+ "grad_norm": 0.020768029615283012,
+ "learning_rate": 0.000516,
+ "loss": 7.464688777923584,
+ "step": 87
+ },
+ {
+ "epoch": 1.223580786026201,
+ "grad_norm": 0.021715788170695305,
+ "learning_rate": 0.000522,
+ "loss": 7.4835028648376465,
+ "step": 88
+ },
+ {
+ "epoch": 1.2375545851528384,
+ "grad_norm": 0.026255754753947258,
+ "learning_rate": 0.0005279999999999999,
+ "loss": 7.443392753601074,
+ "step": 89
+ },
+ {
+ "epoch": 1.251528384279476,
+ "grad_norm": 0.022325951606035233,
+ "learning_rate": 0.000534,
+ "loss": 7.457373142242432,
+ "step": 90
+ },
+ {
+ "epoch": 1.2655021834061135,
+ "grad_norm": 0.009270387701690197,
+ "learning_rate": 0.00054,
+ "loss": 7.455986976623535,
+ "step": 91
+ },
+ {
+ "epoch": 1.279475982532751,
+ "grad_norm": 0.02369544468820095,
+ "learning_rate": 0.0005459999999999999,
+ "loss": 7.442216873168945,
+ "step": 92
+ },
+ {
+ "epoch": 1.2934497816593886,
+ "grad_norm": 0.008474506437778473,
+ "learning_rate": 0.000552,
+ "loss": 7.4294915199279785,
+ "step": 93
+ },
+ {
+ "epoch": 1.3074235807860262,
+ "grad_norm": 0.017043249681591988,
+ "learning_rate": 0.000558,
+ "loss": 7.406932830810547,
+ "step": 94
+ },
+ {
+ "epoch": 1.3213973799126637,
+ "grad_norm": 0.011813412420451641,
+ "learning_rate": 0.0005639999999999999,
+ "loss": 7.415900230407715,
+ "step": 95
+ },
+ {
+ "epoch": 1.3353711790393012,
+ "grad_norm": 0.01086883433163166,
+ "learning_rate": 0.00057,
+ "loss": 7.452998638153076,
+ "step": 96
+ },
+ {
+ "epoch": 1.3493449781659388,
+ "grad_norm": 0.014700951054692268,
+ "learning_rate": 0.0005759999999999999,
+ "loss": 7.410249710083008,
+ "step": 97
+ },
+ {
+ "epoch": 1.3633187772925766,
+ "grad_norm": 0.014512976631522179,
+ "learning_rate": 0.0005819999999999999,
+ "loss": 7.429132461547852,
+ "step": 98
+ },
+ {
+ "epoch": 1.3772925764192139,
+ "grad_norm": 0.012519920244812965,
+ "learning_rate": 0.000588,
+ "loss": 7.430817604064941,
+ "step": 99
+ },
+ {
+ "epoch": 1.3912663755458516,
+ "grad_norm": 0.007132664322853088,
+ "learning_rate": 0.0005939999999999999,
+ "loss": 7.405234336853027,
+ "step": 100
+ },
+ {
+ "epoch": 1.405240174672489,
+ "grad_norm": 0.019489388912916183,
+ "learning_rate": 0.0006,
+ "loss": 7.427791595458984,
+ "step": 101
+ },
+ {
+ "epoch": 1.4192139737991267,
+ "grad_norm": 0.027252664789557457,
+ "learning_rate": 0.0006,
+ "loss": 7.41761589050293,
+ "step": 102
+ },
+ {
+ "epoch": 1.4331877729257643,
+ "grad_norm": 0.02501637488603592,
+ "learning_rate": 0.0006,
+ "loss": 7.416735649108887,
+ "step": 103
+ },
+ {
+ "epoch": 1.4471615720524018,
+ "grad_norm": 0.00989772193133831,
+ "learning_rate": 0.0006,
+ "loss": 7.3756632804870605,
+ "step": 104
+ },
+ {
+ "epoch": 1.4611353711790394,
+ "grad_norm": 0.01622949168086052,
+ "learning_rate": 0.0006,
+ "loss": 7.400811672210693,
+ "step": 105
+ },
+ {
+ "epoch": 1.475109170305677,
+ "grad_norm": 0.012407763861119747,
+ "learning_rate": 0.0006,
+ "loss": 7.37675666809082,
+ "step": 106
+ },
+ {
+ "epoch": 1.4890829694323144,
+ "grad_norm": 0.010353565216064453,
+ "learning_rate": 0.0006,
+ "loss": 7.4013142585754395,
+ "step": 107
+ },
+ {
+ "epoch": 1.503056768558952,
+ "grad_norm": 0.0120553532615304,
+ "learning_rate": 0.0006,
+ "loss": 7.379873275756836,
+ "step": 108
+ },
+ {
+ "epoch": 1.5170305676855895,
+ "grad_norm": 0.01610225811600685,
+ "learning_rate": 0.0006,
+ "loss": 7.371631145477295,
+ "step": 109
+ },
+ {
+ "epoch": 1.531004366812227,
+ "grad_norm": 0.01159879844635725,
+ "learning_rate": 0.0006,
+ "loss": 7.389007568359375,
+ "step": 110
+ },
+ {
+ "epoch": 1.5449781659388646,
+ "grad_norm": 0.008013816550374031,
+ "learning_rate": 0.0006,
+ "loss": 7.395920753479004,
+ "step": 111
+ },
+ {
+ "epoch": 1.5589519650655022,
+ "grad_norm": 0.016334809362888336,
+ "learning_rate": 0.0006,
+ "loss": 7.393121719360352,
+ "step": 112
+ },
+ {
+ "epoch": 1.5729257641921397,
+ "grad_norm": 0.018821721896529198,
+ "learning_rate": 0.0006,
+ "loss": 7.367543697357178,
+ "step": 113
+ },
+ {
+ "epoch": 1.5868995633187772,
+ "grad_norm": 0.014088994823396206,
+ "learning_rate": 0.0006,
+ "loss": 7.385537624359131,
+ "step": 114
+ },
+ {
+ "epoch": 1.600873362445415,
+ "grad_norm": 0.005864050704985857,
+ "learning_rate": 0.0006,
+ "loss": 7.365346908569336,
+ "step": 115
+ },
+ {
+ "epoch": 1.6148471615720523,
+ "grad_norm": 0.011745430529117584,
+ "learning_rate": 0.0006,
+ "loss": 7.390227317810059,
+ "step": 116
+ },
+ {
+ "epoch": 1.62882096069869,
+ "grad_norm": 0.017205573618412018,
+ "learning_rate": 0.0006,
+ "loss": 7.3558549880981445,
+ "step": 117
+ },
+ {
+ "epoch": 1.6427947598253274,
+ "grad_norm": 0.021256346255540848,
+ "learning_rate": 0.0006,
+ "loss": 7.377346992492676,
+ "step": 118
+ },
+ {
+ "epoch": 1.6567685589519652,
+ "grad_norm": 0.019335204735398293,
+ "learning_rate": 0.0006,
+ "loss": 7.346194267272949,
+ "step": 119
+ },
+ {
+ "epoch": 1.6707423580786025,
+ "grad_norm": 0.008134003728628159,
+ "learning_rate": 0.0006,
+ "loss": 7.3063578605651855,
+ "step": 120
+ },
+ {
+ "epoch": 1.6847161572052403,
+ "grad_norm": 0.017613831907510757,
+ "learning_rate": 0.0006,
+ "loss": 7.338171005249023,
+ "step": 121
+ },
+ {
+ "epoch": 1.6986899563318776,
+ "grad_norm": 0.00890734139829874,
+ "learning_rate": 0.0006,
+ "loss": 7.355384826660156,
+ "step": 122
+ },
+ {
+ "epoch": 1.7126637554585153,
+ "grad_norm": 0.013448155485093594,
+ "learning_rate": 0.0006,
+ "loss": 7.344296455383301,
+ "step": 123
+ },
+ {
+ "epoch": 1.726637554585153,
+ "grad_norm": 0.009241633117198944,
+ "learning_rate": 0.0006,
+ "loss": 7.3454060554504395,
+ "step": 124
+ },
+ {
+ "epoch": 1.7406113537117904,
+ "grad_norm": 0.018561311066150665,
+ "learning_rate": 0.0006,
+ "loss": 7.32348108291626,
+ "step": 125
+ },
+ {
+ "epoch": 1.754585152838428,
+ "grad_norm": 0.025667704641819,
+ "learning_rate": 0.0006,
+ "loss": 7.356142520904541,
+ "step": 126
+ },
+ {
+ "epoch": 1.7685589519650655,
+ "grad_norm": 0.036232151091098785,
+ "learning_rate": 0.0006,
+ "loss": 7.363539695739746,
+ "step": 127
+ },
+ {
+ "epoch": 1.782532751091703,
+ "grad_norm": 0.059483714401721954,
+ "learning_rate": 0.0006,
+ "loss": 7.402331352233887,
+ "step": 128
+ },
+ {
+ "epoch": 1.7965065502183406,
+ "grad_norm": 0.05365009233355522,
+ "learning_rate": 0.0006,
+ "loss": 7.40457820892334,
+ "step": 129
+ },
+ {
+ "epoch": 1.8104803493449781,
+ "grad_norm": 0.02738415263593197,
+ "learning_rate": 0.0006,
+ "loss": 7.3756208419799805,
+ "step": 130
+ },
+ {
+ "epoch": 1.8244541484716157,
+ "grad_norm": 0.035017091780900955,
+ "learning_rate": 0.0006,
+ "loss": 7.379263877868652,
+ "step": 131
+ },
+ {
+ "epoch": 1.8384279475982532,
+ "grad_norm": 0.02819863148033619,
+ "learning_rate": 0.0006,
+ "loss": 7.327673435211182,
+ "step": 132
+ },
+ {
+ "epoch": 1.8524017467248908,
+ "grad_norm": 0.017754238098859787,
+ "learning_rate": 0.0006,
+ "loss": 7.333420276641846,
+ "step": 133
+ },
+ {
+ "epoch": 1.8663755458515285,
+ "grad_norm": 0.024685733020305634,
+ "learning_rate": 0.0006,
+ "loss": 7.34493350982666,
+ "step": 134
+ },
+ {
+ "epoch": 1.8803493449781659,
+ "grad_norm": 0.019426532089710236,
+ "learning_rate": 0.0006,
+ "loss": 7.328448295593262,
+ "step": 135
+ },
+ {
+ "epoch": 1.8943231441048036,
+ "grad_norm": 0.02106391079723835,
+ "learning_rate": 0.0006,
+ "loss": 7.327917098999023,
+ "step": 136
+ },
+ {
+ "epoch": 1.908296943231441,
+ "grad_norm": 0.02056879736483097,
+ "learning_rate": 0.0006,
+ "loss": 7.3311991691589355,
+ "step": 137
+ },
+ {
+ "epoch": 1.9222707423580787,
+ "grad_norm": 0.016865141689777374,
+ "learning_rate": 0.0006,
+ "loss": 7.312766075134277,
+ "step": 138
+ },
+ {
+ "epoch": 1.936244541484716,
+ "grad_norm": 0.017549855634570122,
+ "learning_rate": 0.0006,
+ "loss": 7.31191349029541,
+ "step": 139
+ },
+ {
+ "epoch": 1.9502183406113538,
+ "grad_norm": 0.015930157154798508,
+ "learning_rate": 0.0006,
+ "loss": 7.330901622772217,
+ "step": 140
+ },
+ {
+ "epoch": 1.9641921397379911,
+ "grad_norm": 0.013283558189868927,
+ "learning_rate": 0.0006,
+ "loss": 7.298318386077881,
+ "step": 141
+ },
+ {
+ "epoch": 1.9781659388646289,
+ "grad_norm": 0.012023469433188438,
+ "learning_rate": 0.0006,
+ "loss": 7.282337188720703,
+ "step": 142
+ },
+ {
+ "epoch": 1.9921397379912664,
+ "grad_norm": 0.01459511648863554,
+ "learning_rate": 0.0006,
+ "loss": 7.282924175262451,
+ "step": 143
+ },
+ {
+ "epoch": 2.0,
+ "grad_norm": 0.017474235966801643,
+ "learning_rate": 0.0006,
+ "loss": 7.254333972930908,
+ "step": 144
+ },
+ {
+ "epoch": 2.0,
+ "eval_loss": 7.318708896636963,
+ "eval_runtime": 59.6555,
+ "eval_samples_per_second": 40.935,
+ "eval_steps_per_second": 1.291,
+ "step": 144
+ },
+ {
+ "epoch": 2.0139737991266378,
+ "grad_norm": 0.019757507368922234,
+ "learning_rate": 0.0006,
+ "loss": 7.30288553237915,
+ "step": 145
+ },
+ {
+ "epoch": 2.027947598253275,
+ "grad_norm": 0.014571248553693295,
+ "learning_rate": 0.0006,
+ "loss": 7.274681568145752,
+ "step": 146
+ },
+ {
+ "epoch": 2.041921397379913,
+ "grad_norm": 0.015744198113679886,
+ "learning_rate": 0.0006,
+ "loss": 7.278632640838623,
+ "step": 147
+ },
+ {
+ "epoch": 2.05589519650655,
+ "grad_norm": 0.028048858046531677,
+ "learning_rate": 0.0006,
+ "loss": 7.26558780670166,
+ "step": 148
+ },
+ {
+ "epoch": 2.069868995633188,
+ "grad_norm": 0.06547307223081589,
+ "learning_rate": 0.0006,
+ "loss": 7.32567834854126,
+ "step": 149
+ },
+ {
+ "epoch": 2.0838427947598253,
+ "grad_norm": 0.05969763547182083,
+ "learning_rate": 0.0006,
+ "loss": 7.395836353302002,
+ "step": 150
+ },
+ {
+ "epoch": 2.097816593886463,
+ "grad_norm": 0.032574281096458435,
+ "learning_rate": 0.0006,
+ "loss": 7.318150997161865,
+ "step": 151
+ },
+ {
+ "epoch": 2.1117903930131003,
+ "grad_norm": 0.029452962800860405,
+ "learning_rate": 0.0006,
+ "loss": 7.308066368103027,
+ "step": 152
+ },
+ {
+ "epoch": 2.125764192139738,
+ "grad_norm": 0.03161991387605667,
+ "learning_rate": 0.0006,
+ "loss": 7.306290626525879,
+ "step": 153
+ },
+ {
+ "epoch": 2.1397379912663754,
+ "grad_norm": 0.027096295729279518,
+ "learning_rate": 0.0006,
+ "loss": 7.297796249389648,
+ "step": 154
+ },
+ {
+ "epoch": 2.153711790393013,
+ "grad_norm": 0.028657056391239166,
+ "learning_rate": 0.0006,
+ "loss": 7.317424774169922,
+ "step": 155
+ },
+ {
+ "epoch": 2.1676855895196505,
+ "grad_norm": 0.02046995982527733,
+ "learning_rate": 0.0006,
+ "loss": 7.297986030578613,
+ "step": 156
+ },
+ {
+ "epoch": 2.1816593886462883,
+ "grad_norm": 0.02220369316637516,
+ "learning_rate": 0.0006,
+ "loss": 7.268237590789795,
+ "step": 157
+ },
+ {
+ "epoch": 2.1956331877729256,
+ "grad_norm": 0.018381357192993164,
+ "learning_rate": 0.0006,
+ "loss": 7.259415626525879,
+ "step": 158
+ },
+ {
+ "epoch": 2.2096069868995634,
+ "grad_norm": 0.0204413291066885,
+ "learning_rate": 0.0006,
+ "loss": 7.258173942565918,
+ "step": 159
+ },
+ {
+ "epoch": 2.2235807860262007,
+ "grad_norm": 0.015874288976192474,
+ "learning_rate": 0.0006,
+ "loss": 7.281834602355957,
+ "step": 160
+ },
+ {
+ "epoch": 2.2375545851528384,
+ "grad_norm": 0.01634068600833416,
+ "learning_rate": 0.0006,
+ "loss": 7.2469868659973145,
+ "step": 161
+ },
+ {
+ "epoch": 2.251528384279476,
+ "grad_norm": 0.01728747971355915,
+ "learning_rate": 0.0006,
+ "loss": 7.280995845794678,
+ "step": 162
+ },
+ {
+ "epoch": 2.2655021834061135,
+ "grad_norm": 0.016341188922524452,
+ "learning_rate": 0.0006,
+ "loss": 7.273357391357422,
+ "step": 163
+ },
+ {
+ "epoch": 2.279475982532751,
+ "grad_norm": 0.016125807538628578,
+ "learning_rate": 0.0006,
+ "loss": 7.273510932922363,
+ "step": 164
+ },
+ {
+ "epoch": 2.2934497816593886,
+ "grad_norm": 0.01629287749528885,
+ "learning_rate": 0.0006,
+ "loss": 7.267029762268066,
+ "step": 165
+ },
+ {
+ "epoch": 2.3074235807860264,
+ "grad_norm": 0.015296131372451782,
+ "learning_rate": 0.0006,
+ "loss": 7.211989402770996,
+ "step": 166
+ },
+ {
+ "epoch": 2.3213973799126637,
+ "grad_norm": 0.01709570363163948,
+ "learning_rate": 0.0006,
+ "loss": 7.242851734161377,
+ "step": 167
+ },
+ {
+ "epoch": 2.3353711790393015,
+ "grad_norm": 0.012552388943731785,
+ "learning_rate": 0.0006,
+ "loss": 7.241696834564209,
+ "step": 168
+ },
+ {
+ "epoch": 2.349344978165939,
+ "grad_norm": 0.014571291394531727,
+ "learning_rate": 0.0006,
+ "loss": 7.270681381225586,
+ "step": 169
+ },
+ {
+ "epoch": 2.3633187772925766,
+ "grad_norm": 0.013389437459409237,
+ "learning_rate": 0.0006,
+ "loss": 7.191600799560547,
+ "step": 170
+ },
+ {
+ "epoch": 2.377292576419214,
+ "grad_norm": 0.011895314790308475,
+ "learning_rate": 0.0006,
+ "loss": 7.213610649108887,
+ "step": 171
+ },
+ {
+ "epoch": 2.3912663755458516,
+ "grad_norm": 0.01345642190426588,
+ "learning_rate": 0.0006,
+ "loss": 7.193787574768066,
+ "step": 172
+ },
+ {
+ "epoch": 2.405240174672489,
+ "grad_norm": 0.0091794328764081,
+ "learning_rate": 0.0006,
+ "loss": 7.233364105224609,
+ "step": 173
+ },
+ {
+ "epoch": 2.4192139737991267,
+ "grad_norm": 0.013767086900770664,
+ "learning_rate": 0.0006,
+ "loss": 7.217272758483887,
+ "step": 174
+ },
+ {
+ "epoch": 2.433187772925764,
+ "grad_norm": 0.015134960412979126,
+ "learning_rate": 0.0006,
+ "loss": 7.192791938781738,
+ "step": 175
+ },
+ {
+ "epoch": 2.447161572052402,
+ "grad_norm": 0.01790648326277733,
+ "learning_rate": 0.0006,
+ "loss": 7.229467391967773,
+ "step": 176
+ },
+ {
+ "epoch": 2.461135371179039,
+ "grad_norm": 0.015362377278506756,
+ "learning_rate": 0.0006,
+ "loss": 7.180346488952637,
+ "step": 177
+ },
+ {
+ "epoch": 2.475109170305677,
+ "grad_norm": 0.010505498386919498,
+ "learning_rate": 0.0006,
+ "loss": 7.188833236694336,
+ "step": 178
+ },
+ {
+ "epoch": 2.489082969432314,
+ "grad_norm": 0.009129747748374939,
+ "learning_rate": 0.0006,
+ "loss": 7.203153133392334,
+ "step": 179
+ },
+ {
+ "epoch": 2.503056768558952,
+ "grad_norm": 0.01147883478552103,
+ "learning_rate": 0.0006,
+ "loss": 7.163824558258057,
+ "step": 180
+ },
+ {
+ "epoch": 2.5170305676855893,
+ "grad_norm": 0.011338942684233189,
+ "learning_rate": 0.0006,
+ "loss": 7.16287899017334,
+ "step": 181
+ },
+ {
+ "epoch": 2.531004366812227,
+ "grad_norm": 0.013277675025165081,
+ "learning_rate": 0.0006,
+ "loss": 7.155905723571777,
+ "step": 182
+ },
+ {
+ "epoch": 2.544978165938865,
+ "grad_norm": 0.015269347466528416,
+ "learning_rate": 0.0006,
+ "loss": 7.189371109008789,
+ "step": 183
+ },
+ {
+ "epoch": 2.558951965065502,
+ "grad_norm": 0.017905596643686295,
+ "learning_rate": 0.0006,
+ "loss": 7.177547931671143,
+ "step": 184
+ },
+ {
+ "epoch": 2.5729257641921395,
+ "grad_norm": 0.024431610479950905,
+ "learning_rate": 0.0006,
+ "loss": 7.192169189453125,
+ "step": 185
+ },
+ {
+ "epoch": 2.5868995633187772,
+ "grad_norm": 0.03656961768865585,
+ "learning_rate": 0.0006,
+ "loss": 7.124387741088867,
+ "step": 186
+ },
+ {
+ "epoch": 2.600873362445415,
+ "grad_norm": 0.04337048903107643,
+ "learning_rate": 0.0006,
+ "loss": 7.140231609344482,
+ "step": 187
+ },
+ {
+ "epoch": 2.6148471615720523,
+ "grad_norm": 0.04047228768467903,
+ "learning_rate": 0.0006,
+ "loss": 7.171927452087402,
+ "step": 188
+ },
+ {
+ "epoch": 2.62882096069869,
+ "grad_norm": 0.07825770974159241,
+ "learning_rate": 0.0006,
+ "loss": 7.207631587982178,
+ "step": 189
+ },
+ {
+ "epoch": 2.6427947598253274,
+ "grad_norm": 0.07118063420057297,
+ "learning_rate": 0.0006,
+ "loss": 7.209488391876221,
+ "step": 190
+ },
+ {
+ "epoch": 2.656768558951965,
+ "grad_norm": 0.05153293535113335,
+ "learning_rate": 0.0006,
+ "loss": 7.14793062210083,
+ "step": 191
+ },
+ {
+ "epoch": 2.6707423580786025,
+ "grad_norm": 0.03318839520215988,
+ "learning_rate": 0.0006,
+ "loss": 7.163010597229004,
+ "step": 192
+ },
+ {
+ "epoch": 2.6847161572052403,
+ "grad_norm": 0.029333539307117462,
+ "learning_rate": 0.0006,
+ "loss": 7.159955978393555,
+ "step": 193
+ },
+ {
+ "epoch": 2.6986899563318776,
+ "grad_norm": 0.029405880719423294,
+ "learning_rate": 0.0006,
+ "loss": 7.136178493499756,
+ "step": 194
+ },
+ {
+ "epoch": 2.7126637554585153,
+ "grad_norm": 0.037815988063812256,
+ "learning_rate": 0.0006,
+ "loss": 7.183186054229736,
+ "step": 195
+ },
+ {
+ "epoch": 2.726637554585153,
+ "grad_norm": 0.020307250320911407,
+ "learning_rate": 0.0006,
+ "loss": 7.133164882659912,
+ "step": 196
+ },
+ {
+ "epoch": 2.7406113537117904,
+ "grad_norm": 0.026048069819808006,
+ "learning_rate": 0.0006,
+ "loss": 7.124977111816406,
+ "step": 197
+ },
+ {
+ "epoch": 2.7545851528384278,
+ "grad_norm": 0.027316724881529808,
+ "learning_rate": 0.0006,
+ "loss": 7.07265567779541,
+ "step": 198
+ },
+ {
+ "epoch": 2.7685589519650655,
+ "grad_norm": 0.021629009395837784,
+ "learning_rate": 0.0006,
+ "loss": 7.150020599365234,
+ "step": 199
+ },
+ {
+ "epoch": 2.7825327510917033,
+ "grad_norm": 0.01605929434299469,
+ "learning_rate": 0.0006,
+ "loss": 7.10919189453125,
+ "step": 200
+ },
+ {
+ "epoch": 2.7965065502183406,
+ "grad_norm": 0.020800089463591576,
+ "learning_rate": 0.0006,
+ "loss": 7.132045269012451,
+ "step": 201
+ },
+ {
+ "epoch": 2.810480349344978,
+ "grad_norm": 0.019225774332880974,
+ "learning_rate": 0.0006,
+ "loss": 7.128364562988281,
+ "step": 202
+ },
+ {
+ "epoch": 2.8244541484716157,
+ "grad_norm": 0.01561372634023428,
+ "learning_rate": 0.0006,
+ "loss": 7.083669662475586,
+ "step": 203
+ },
+ {
+ "epoch": 2.8384279475982535,
+ "grad_norm": 0.021969519555568695,
+ "learning_rate": 0.0006,
+ "loss": 7.050841331481934,
+ "step": 204
+ },
+ {
+ "epoch": 2.8524017467248908,
+ "grad_norm": 0.021453695371747017,
+ "learning_rate": 0.0006,
+ "loss": 7.097085475921631,
+ "step": 205
+ },
+ {
+ "epoch": 2.8663755458515285,
+ "grad_norm": 0.022836022078990936,
+ "learning_rate": 0.0006,
+ "loss": 7.10405969619751,
+ "step": 206
+ },
+ {
+ "epoch": 2.880349344978166,
+ "grad_norm": 0.02430851384997368,
+ "learning_rate": 0.0006,
+ "loss": 7.045181751251221,
+ "step": 207
+ },
+ {
+ "epoch": 2.8943231441048036,
+ "grad_norm": 0.020166993141174316,
+ "learning_rate": 0.0006,
+ "loss": 7.068183898925781,
+ "step": 208
+ },
+ {
+ "epoch": 2.908296943231441,
+ "grad_norm": 0.0118505684658885,
+ "learning_rate": 0.0006,
+ "loss": 7.07132625579834,
+ "step": 209
+ },
+ {
+ "epoch": 2.9222707423580787,
+ "grad_norm": 0.023772427812218666,
+ "learning_rate": 0.0006,
+ "loss": 7.055238723754883,
+ "step": 210
+ },
+ {
+ "epoch": 2.936244541484716,
+ "grad_norm": 0.02350712940096855,
+ "learning_rate": 0.0006,
+ "loss": 7.055898666381836,
+ "step": 211
+ },
+ {
+ "epoch": 2.950218340611354,
+ "grad_norm": 0.017244907096028328,
+ "learning_rate": 0.0006,
+ "loss": 7.0099711418151855,
+ "step": 212
+ },
+ {
+ "epoch": 2.964192139737991,
+ "grad_norm": 0.021565575152635574,
+ "learning_rate": 0.0006,
+ "loss": 6.991751670837402,
+ "step": 213
+ },
+ {
+ "epoch": 2.978165938864629,
+ "grad_norm": 0.03241860866546631,
+ "learning_rate": 0.0006,
+ "loss": 7.005980491638184,
+ "step": 214
+ },
+ {
+ "epoch": 2.992139737991266,
+ "grad_norm": 0.04613726586103439,
+ "learning_rate": 0.0006,
+ "loss": 7.027859687805176,
+ "step": 215
+ },
+ {
+ "epoch": 3.0,
+ "grad_norm": 0.06404894590377808,
+ "learning_rate": 0.0006,
+ "loss": 7.123725891113281,
+ "step": 216
+ },
+ {
+ "epoch": 3.0,
+ "eval_loss": 7.070647716522217,
+ "eval_runtime": 58.5801,
+ "eval_samples_per_second": 41.687,
+ "eval_steps_per_second": 1.314,
+ "step": 216
+ },
+ {
+ "epoch": 3.0139737991266378,
+ "grad_norm": 0.05542432889342308,
+ "learning_rate": 0.0006,
+ "loss": 7.078299522399902,
+ "step": 217
+ },
+ {
+ "epoch": 3.027947598253275,
+ "grad_norm": 0.026484526693820953,
+ "learning_rate": 0.0006,
+ "loss": 6.961159706115723,
+ "step": 218
+ },
+ {
+ "epoch": 3.041921397379913,
+ "grad_norm": 0.028276391327381134,
+ "learning_rate": 0.0006,
+ "loss": 7.019961833953857,
+ "step": 219
+ },
+ {
+ "epoch": 3.05589519650655,
+ "grad_norm": 0.024486595764756203,
+ "learning_rate": 0.0006,
+ "loss": 7.039677619934082,
+ "step": 220
+ },
+ {
+ "epoch": 3.069868995633188,
+ "grad_norm": 0.017620893195271492,
+ "learning_rate": 0.0006,
+ "loss": 6.9270782470703125,
+ "step": 221
+ },
+ {
+ "epoch": 3.0838427947598253,
+ "grad_norm": 0.02547302283346653,
+ "learning_rate": 0.0006,
+ "loss": 6.924233436584473,
+ "step": 222
+ },
+ {
+ "epoch": 3.097816593886463,
+ "grad_norm": 0.02539009042084217,
+ "learning_rate": 0.0006,
+ "loss": 6.973979949951172,
+ "step": 223
+ },
+ {
+ "epoch": 3.1117903930131003,
+ "grad_norm": 0.03256227448582649,
+ "learning_rate": 0.0006,
+ "loss": 6.933725357055664,
+ "step": 224
+ },
+ {
+ "epoch": 3.125764192139738,
+ "grad_norm": 0.04651800915598869,
+ "learning_rate": 0.0006,
+ "loss": 6.951044082641602,
+ "step": 225
+ },
+ {
+ "epoch": 3.1397379912663754,
+ "grad_norm": 0.04822975769639015,
+ "learning_rate": 0.0006,
+ "loss": 6.9614410400390625,
+ "step": 226
+ },
+ {
+ "epoch": 3.153711790393013,
+ "grad_norm": 0.031165020540356636,
+ "learning_rate": 0.0006,
+ "loss": 6.91193151473999,
+ "step": 227
+ },
+ {
+ "epoch": 3.1676855895196505,
+ "grad_norm": 0.029754292219877243,
+ "learning_rate": 0.0006,
+ "loss": 6.91802978515625,
+ "step": 228
+ },
+ {
+ "epoch": 3.1816593886462883,
+ "grad_norm": 0.02193152718245983,
+ "learning_rate": 0.0006,
+ "loss": 6.928090572357178,
+ "step": 229
+ },
+ {
+ "epoch": 3.1956331877729256,
+ "grad_norm": 0.02428470179438591,
+ "learning_rate": 0.0006,
+ "loss": 6.90396785736084,
+ "step": 230
+ },
+ {
+ "epoch": 3.2096069868995634,
+ "grad_norm": 0.023374345153570175,
+ "learning_rate": 0.0006,
+ "loss": 6.885858535766602,
+ "step": 231
+ },
+ {
+ "epoch": 3.2235807860262007,
+ "grad_norm": 0.023405000567436218,
+ "learning_rate": 0.0006,
+ "loss": 6.8855767250061035,
+ "step": 232
+ },
+ {
+ "epoch": 3.2375545851528384,
+ "grad_norm": 0.017277134582400322,
+ "learning_rate": 0.0006,
+ "loss": 6.901449203491211,
+ "step": 233
+ },
+ {
+ "epoch": 3.251528384279476,
+ "grad_norm": 0.019496750086545944,
+ "learning_rate": 0.0006,
+ "loss": 6.885687828063965,
+ "step": 234
+ },
+ {
+ "epoch": 3.2655021834061135,
+ "grad_norm": 0.016893276944756508,
+ "learning_rate": 0.0006,
+ "loss": 6.872369766235352,
+ "step": 235
+ },
+ {
+ "epoch": 3.279475982532751,
+ "grad_norm": 0.02058715932071209,
+ "learning_rate": 0.0006,
+ "loss": 6.823280334472656,
+ "step": 236
+ },
+ {
+ "epoch": 3.2934497816593886,
+ "grad_norm": 0.015530755743384361,
+ "learning_rate": 0.0006,
+ "loss": 6.860499858856201,
+ "step": 237
+ },
+ {
+ "epoch": 3.3074235807860264,
+ "grad_norm": 0.01934926211833954,
+ "learning_rate": 0.0006,
+ "loss": 6.817448616027832,
+ "step": 238
+ },
+ {
+ "epoch": 3.3213973799126637,
+ "grad_norm": 0.017413552850484848,
+ "learning_rate": 0.0006,
+ "loss": 6.881955146789551,
+ "step": 239
+ },
+ {
+ "epoch": 3.3353711790393015,
+ "grad_norm": 0.026478875428438187,
+ "learning_rate": 0.0006,
+ "loss": 6.827404499053955,
+ "step": 240
+ },
+ {
+ "epoch": 3.349344978165939,
+ "grad_norm": 0.032611701637506485,
+ "learning_rate": 0.0006,
+ "loss": 6.78285551071167,
+ "step": 241
+ },
+ {
+ "epoch": 3.3633187772925766,
+ "grad_norm": 0.041900306940078735,
+ "learning_rate": 0.0006,
+ "loss": 6.783053874969482,
+ "step": 242
+ },
+ {
+ "epoch": 3.377292576419214,
+ "grad_norm": 0.04149497672915459,
+ "learning_rate": 0.0006,
+ "loss": 6.745220184326172,
+ "step": 243
+ },
+ {
+ "epoch": 3.3912663755458516,
+ "grad_norm": 0.023949826136231422,
+ "learning_rate": 0.0006,
+ "loss": 6.841533184051514,
+ "step": 244
+ },
+ {
+ "epoch": 3.405240174672489,
+ "grad_norm": 0.028601842001080513,
+ "learning_rate": 0.0006,
+ "loss": 6.852852821350098,
+ "step": 245
+ },
+ {
+ "epoch": 3.4192139737991267,
+ "grad_norm": 0.027365155518054962,
+ "learning_rate": 0.0006,
+ "loss": 6.74376106262207,
+ "step": 246
+ },
+ {
+ "epoch": 3.433187772925764,
+ "grad_norm": 0.02856568619608879,
+ "learning_rate": 0.0006,
+ "loss": 6.74906063079834,
+ "step": 247
+ },
+ {
+ "epoch": 3.447161572052402,
+ "grad_norm": 0.019172416999936104,
+ "learning_rate": 0.0006,
+ "loss": 6.809206008911133,
+ "step": 248
+ },
+ {
+ "epoch": 3.461135371179039,
+ "grad_norm": 0.02161746844649315,
+ "learning_rate": 0.0006,
+ "loss": 6.789888381958008,
+ "step": 249
+ },
+ {
+ "epoch": 3.475109170305677,
+ "grad_norm": 0.020273666828870773,
+ "learning_rate": 0.0006,
+ "loss": 6.73512077331543,
+ "step": 250
+ },
+ {
+ "epoch": 3.489082969432314,
+ "grad_norm": 0.022329283878207207,
+ "learning_rate": 0.0006,
+ "loss": 6.826148509979248,
+ "step": 251
+ },
+ {
+ "epoch": 3.503056768558952,
+ "grad_norm": 0.02284284122288227,
+ "learning_rate": 0.0006,
+ "loss": 6.748521327972412,
+ "step": 252
+ },
+ {
+ "epoch": 3.5170305676855893,
+ "grad_norm": 0.026368247345089912,
+ "learning_rate": 0.0006,
+ "loss": 6.645879745483398,
+ "step": 253
+ },
+ {
+ "epoch": 3.531004366812227,
+ "grad_norm": 0.03627244383096695,
+ "learning_rate": 0.0006,
+ "loss": 6.666064739227295,
+ "step": 254
+ },
+ {
+ "epoch": 3.544978165938865,
+ "grad_norm": 0.04140935093164444,
+ "learning_rate": 0.0006,
+ "loss": 6.712275981903076,
+ "step": 255
+ },
+ {
+ "epoch": 3.558951965065502,
+ "grad_norm": 0.0390462800860405,
+ "learning_rate": 0.0006,
+ "loss": 6.665900230407715,
+ "step": 256
+ },
+ {
+ "epoch": 3.5729257641921395,
+ "grad_norm": 0.05178583785891533,
+ "learning_rate": 0.0006,
+ "loss": 6.753298759460449,
+ "step": 257
+ },
+ {
+ "epoch": 3.5868995633187772,
+ "grad_norm": 0.054328061640262604,
+ "learning_rate": 0.0006,
+ "loss": 6.732633113861084,
+ "step": 258
+ },
+ {
+ "epoch": 3.600873362445415,
+ "grad_norm": 0.04590460658073425,
+ "learning_rate": 0.0006,
+ "loss": 6.65250301361084,
+ "step": 259
+ },
+ {
+ "epoch": 3.6148471615720523,
+ "grad_norm": 0.044331666082143784,
+ "learning_rate": 0.0006,
+ "loss": 6.68654203414917,
+ "step": 260
+ },
+ {
+ "epoch": 3.62882096069869,
+ "grad_norm": 0.05308730527758598,
+ "learning_rate": 0.0006,
+ "loss": 6.642482757568359,
+ "step": 261
+ },
+ {
+ "epoch": 3.6427947598253274,
+ "grad_norm": 0.03999016433954239,
+ "learning_rate": 0.0006,
+ "loss": 6.625227451324463,
+ "step": 262
+ },
+ {
+ "epoch": 3.656768558951965,
+ "grad_norm": 0.04088086634874344,
+ "learning_rate": 0.0006,
+ "loss": 6.67623233795166,
+ "step": 263
+ },
+ {
+ "epoch": 3.6707423580786025,
+ "grad_norm": 0.03615636005997658,
+ "learning_rate": 0.0006,
+ "loss": 6.6869001388549805,
+ "step": 264
+ },
+ {
+ "epoch": 3.6847161572052403,
+ "grad_norm": 0.02817360684275627,
+ "learning_rate": 0.0006,
+ "loss": 6.674601078033447,
+ "step": 265
+ },
+ {
+ "epoch": 3.6986899563318776,
+ "grad_norm": 0.032586123794317245,
+ "learning_rate": 0.0006,
+ "loss": 6.612994194030762,
+ "step": 266
+ },
+ {
+ "epoch": 3.7126637554585153,
+ "grad_norm": 0.031117867678403854,
+ "learning_rate": 0.0006,
+ "loss": 6.6244354248046875,
+ "step": 267
+ },
+ {
+ "epoch": 3.726637554585153,
+ "grad_norm": 0.028330687433481216,
+ "learning_rate": 0.0006,
+ "loss": 6.669557094573975,
+ "step": 268
+ },
+ {
+ "epoch": 3.7406113537117904,
+ "grad_norm": 0.021489202976226807,
+ "learning_rate": 0.0006,
+ "loss": 6.632035732269287,
+ "step": 269
+ },
+ {
+ "epoch": 3.7545851528384278,
+ "grad_norm": 0.024269424378871918,
+ "learning_rate": 0.0006,
+ "loss": 6.660867691040039,
+ "step": 270
+ },
+ {
+ "epoch": 3.7685589519650655,
+ "grad_norm": 0.02237873338162899,
+ "learning_rate": 0.0006,
+ "loss": 6.602587699890137,
+ "step": 271
+ },
+ {
+ "epoch": 3.7825327510917033,
+ "grad_norm": 0.026503929868340492,
+ "learning_rate": 0.0006,
+ "loss": 6.599200248718262,
+ "step": 272
+ },
+ {
+ "epoch": 3.7965065502183406,
+ "grad_norm": 0.03687124326825142,
+ "learning_rate": 0.0006,
+ "loss": 6.567296504974365,
+ "step": 273
+ },
+ {
+ "epoch": 3.810480349344978,
+ "grad_norm": 0.04708952456712723,
+ "learning_rate": 0.0006,
+ "loss": 6.621265411376953,
+ "step": 274
+ },
+ {
+ "epoch": 3.8244541484716157,
+ "grad_norm": 0.041846614331007004,
+ "learning_rate": 0.0006,
+ "loss": 6.550536155700684,
+ "step": 275
+ },
+ {
+ "epoch": 3.8384279475982535,
+ "grad_norm": 0.031249675899744034,
+ "learning_rate": 0.0006,
+ "loss": 6.54278564453125,
+ "step": 276
+ },
+ {
+ "epoch": 3.8524017467248908,
+ "grad_norm": 0.023082256317138672,
+ "learning_rate": 0.0006,
+ "loss": 6.544787406921387,
+ "step": 277
+ },
+ {
+ "epoch": 3.8663755458515285,
+ "grad_norm": 0.020995570346713066,
+ "learning_rate": 0.0006,
+ "loss": 6.518143653869629,
+ "step": 278
+ },
+ {
+ "epoch": 3.880349344978166,
+ "grad_norm": 0.020839324221014977,
+ "learning_rate": 0.0006,
+ "loss": 6.546667098999023,
+ "step": 279
+ },
+ {
+ "epoch": 3.8943231441048036,
+ "grad_norm": 0.018377432599663734,
+ "learning_rate": 0.0006,
+ "loss": 6.478307723999023,
+ "step": 280
+ },
+ {
+ "epoch": 3.908296943231441,
+ "grad_norm": 0.021713724359869957,
+ "learning_rate": 0.0006,
+ "loss": 6.485574245452881,
+ "step": 281
+ },
+ {
+ "epoch": 3.9222707423580787,
+ "grad_norm": 0.01359301246702671,
+ "learning_rate": 0.0006,
+ "loss": 6.570268630981445,
+ "step": 282
+ },
+ {
+ "epoch": 3.936244541484716,
+ "grad_norm": 0.016233332455158234,
+ "learning_rate": 0.0006,
+ "loss": 6.57150936126709,
+ "step": 283
+ },
+ {
+ "epoch": 3.950218340611354,
+ "grad_norm": 0.020687608048319817,
+ "learning_rate": 0.0006,
+ "loss": 6.527956008911133,
+ "step": 284
+ },
+ {
+ "epoch": 3.964192139737991,
+ "grad_norm": 0.02032964862883091,
+ "learning_rate": 0.0006,
+ "loss": 6.468169689178467,
+ "step": 285
+ },
+ {
+ "epoch": 3.978165938864629,
+ "grad_norm": 0.016676288098096848,
+ "learning_rate": 0.0006,
+ "loss": 6.448418617248535,
+ "step": 286
+ },
+ {
+ "epoch": 3.992139737991266,
+ "grad_norm": 0.0168539360165596,
+ "learning_rate": 0.0006,
+ "loss": 6.461180686950684,
+ "step": 287
+ },
+ {
+ "epoch": 4.0,
+ "grad_norm": 0.02666233666241169,
+ "learning_rate": 0.0006,
+ "loss": 6.344893932342529,
+ "step": 288
+ },
+ {
+ "epoch": 4.0,
+ "eval_loss": 6.5384979248046875,
+ "eval_runtime": 58.814,
+ "eval_samples_per_second": 41.521,
+ "eval_steps_per_second": 1.309,
+ "step": 288
+ },
+ {
+ "epoch": 4.013973799126638,
+ "grad_norm": 0.04154708981513977,
+ "learning_rate": 0.0006,
+ "loss": 6.41855525970459,
+ "step": 289
+ },
+ {
+ "epoch": 4.0279475982532755,
+ "grad_norm": 0.04895668104290962,
+ "learning_rate": 0.0006,
+ "loss": 6.499302864074707,
+ "step": 290
+ },
+ {
+ "epoch": 4.041921397379912,
+ "grad_norm": 0.034464482218027115,
+ "learning_rate": 0.0006,
+ "loss": 6.438611030578613,
+ "step": 291
+ },
+ {
+ "epoch": 4.05589519650655,
+ "grad_norm": 0.06845773756504059,
+ "learning_rate": 0.0006,
+ "loss": 6.538361072540283,
+ "step": 292
+ },
+ {
+ "epoch": 4.069868995633188,
+ "grad_norm": 0.053768858313560486,
+ "learning_rate": 0.0006,
+ "loss": 6.527524948120117,
+ "step": 293
+ },
+ {
+ "epoch": 4.083842794759826,
+ "grad_norm": 0.039037931710481644,
+ "learning_rate": 0.0006,
+ "loss": 6.503847122192383,
+ "step": 294
+ },
+ {
+ "epoch": 4.097816593886463,
+ "grad_norm": 0.03962196782231331,
+ "learning_rate": 0.0006,
+ "loss": 6.475178241729736,
+ "step": 295
+ },
+ {
+ "epoch": 4.1117903930131,
+ "grad_norm": 0.049101535230875015,
+ "learning_rate": 0.0006,
+ "loss": 6.474587917327881,
+ "step": 296
+ },
+ {
+ "epoch": 4.125764192139738,
+ "grad_norm": 0.02830282226204872,
+ "learning_rate": 0.0006,
+ "loss": 6.45065975189209,
+ "step": 297
+ },
+ {
+ "epoch": 4.139737991266376,
+ "grad_norm": 0.023153482005000114,
+ "learning_rate": 0.0006,
+ "loss": 6.5218706130981445,
+ "step": 298
+ },
+ {
+ "epoch": 4.153711790393013,
+ "grad_norm": 0.025064002722501755,
+ "learning_rate": 0.0006,
+ "loss": 6.456475257873535,
+ "step": 299
+ },
+ {
+ "epoch": 4.1676855895196505,
+ "grad_norm": 0.02357092685997486,
+ "learning_rate": 0.0006,
+ "loss": 6.485952377319336,
+ "step": 300
+ },
+ {
+ "epoch": 4.181659388646288,
+ "grad_norm": 0.021832305938005447,
+ "learning_rate": 0.0006,
+ "loss": 6.425506591796875,
+ "step": 301
+ },
+ {
+ "epoch": 4.195633187772926,
+ "grad_norm": 0.0227900929749012,
+ "learning_rate": 0.0006,
+ "loss": 6.439155578613281,
+ "step": 302
+ },
+ {
+ "epoch": 4.209606986899563,
+ "grad_norm": 0.015961607918143272,
+ "learning_rate": 0.0006,
+ "loss": 6.460862636566162,
+ "step": 303
+ },
+ {
+ "epoch": 4.223580786026201,
+ "grad_norm": 0.017876390367746353,
+ "learning_rate": 0.0006,
+ "loss": 6.44564151763916,
+ "step": 304
+ },
+ {
+ "epoch": 4.2375545851528384,
+ "grad_norm": 0.013703204691410065,
+ "learning_rate": 0.0006,
+ "loss": 6.418414115905762,
+ "step": 305
+ },
+ {
+ "epoch": 4.251528384279476,
+ "grad_norm": 0.015240306034684181,
+ "learning_rate": 0.0006,
+ "loss": 6.411238670349121,
+ "step": 306
+ },
+ {
+ "epoch": 4.265502183406113,
+ "grad_norm": 0.01587662845849991,
+ "learning_rate": 0.0006,
+ "loss": 6.395642280578613,
+ "step": 307
+ },
+ {
+ "epoch": 4.279475982532751,
+ "grad_norm": 0.014650700613856316,
+ "learning_rate": 0.0006,
+ "loss": 6.302793025970459,
+ "step": 308
+ },
+ {
+ "epoch": 4.293449781659389,
+ "grad_norm": 0.016533225774765015,
+ "learning_rate": 0.0006,
+ "loss": 6.380588054656982,
+ "step": 309
+ },
+ {
+ "epoch": 4.307423580786026,
+ "grad_norm": 0.019963741302490234,
+ "learning_rate": 0.0006,
+ "loss": 6.345336437225342,
+ "step": 310
+ },
+ {
+ "epoch": 4.321397379912664,
+ "grad_norm": 0.02237936295568943,
+ "learning_rate": 0.0006,
+ "loss": 6.387770652770996,
+ "step": 311
+ },
+ {
+ "epoch": 4.335371179039301,
+ "grad_norm": 0.026658328250050545,
+ "learning_rate": 0.0006,
+ "loss": 6.323662757873535,
+ "step": 312
+ },
+ {
+ "epoch": 4.349344978165939,
+ "grad_norm": 0.02852284163236618,
+ "learning_rate": 0.0006,
+ "loss": 6.382204055786133,
+ "step": 313
+ },
+ {
+ "epoch": 4.3633187772925766,
+ "grad_norm": 0.028815090656280518,
+ "learning_rate": 0.0006,
+ "loss": 6.402390480041504,
+ "step": 314
+ },
+ {
+ "epoch": 4.377292576419214,
+ "grad_norm": 0.029393529519438744,
+ "learning_rate": 0.0006,
+ "loss": 6.280278205871582,
+ "step": 315
+ },
+ {
+ "epoch": 4.391266375545851,
+ "grad_norm": 0.02186041697859764,
+ "learning_rate": 0.0006,
+ "loss": 6.302707672119141,
+ "step": 316
+ },
+ {
+ "epoch": 4.405240174672489,
+ "grad_norm": 0.02071218006312847,
+ "learning_rate": 0.0006,
+ "loss": 6.3365020751953125,
+ "step": 317
+ },
+ {
+ "epoch": 4.419213973799127,
+ "grad_norm": 0.02779117226600647,
+ "learning_rate": 0.0006,
+ "loss": 6.252157211303711,
+ "step": 318
+ },
+ {
+ "epoch": 4.4331877729257645,
+ "grad_norm": 0.03470654785633087,
+ "learning_rate": 0.0006,
+ "loss": 6.319070339202881,
+ "step": 319
+ },
+ {
+ "epoch": 4.447161572052401,
+ "grad_norm": 0.037067804485559464,
+ "learning_rate": 0.0006,
+ "loss": 6.314116477966309,
+ "step": 320
+ },
+ {
+ "epoch": 4.461135371179039,
+ "grad_norm": 0.032182756811380386,
+ "learning_rate": 0.0006,
+ "loss": 6.267298698425293,
+ "step": 321
+ },
+ {
+ "epoch": 4.475109170305677,
+ "grad_norm": 0.026305217295885086,
+ "learning_rate": 0.0006,
+ "loss": 6.331688404083252,
+ "step": 322
+ },
+ {
+ "epoch": 4.489082969432315,
+ "grad_norm": 0.027228357270359993,
+ "learning_rate": 0.0006,
+ "loss": 6.31139612197876,
+ "step": 323
+ },
+ {
+ "epoch": 4.503056768558952,
+ "grad_norm": 0.025781169533729553,
+ "learning_rate": 0.0006,
+ "loss": 6.194684982299805,
+ "step": 324
+ },
+ {
+ "epoch": 4.517030567685589,
+ "grad_norm": 0.029186977073550224,
+ "learning_rate": 0.0006,
+ "loss": 6.31306266784668,
+ "step": 325
+ },
+ {
+ "epoch": 4.531004366812227,
+ "grad_norm": 0.023608777672052383,
+ "learning_rate": 0.0006,
+ "loss": 6.285243034362793,
+ "step": 326
+ },
+ {
+ "epoch": 4.544978165938865,
+ "grad_norm": 0.02267594076693058,
+ "learning_rate": 0.0006,
+ "loss": 6.30342960357666,
+ "step": 327
+ },
+ {
+ "epoch": 4.558951965065502,
+ "grad_norm": 0.01926310732960701,
+ "learning_rate": 0.0006,
+ "loss": 6.280606269836426,
+ "step": 328
+ },
+ {
+ "epoch": 4.5729257641921395,
+ "grad_norm": 0.026484837755560875,
+ "learning_rate": 0.0006,
+ "loss": 6.237980365753174,
+ "step": 329
+ },
+ {
+ "epoch": 4.586899563318777,
+ "grad_norm": 0.026228854432702065,
+ "learning_rate": 0.0006,
+ "loss": 6.3018293380737305,
+ "step": 330
+ },
+ {
+ "epoch": 4.600873362445415,
+ "grad_norm": 0.022096967324614525,
+ "learning_rate": 0.0006,
+ "loss": 6.253863334655762,
+ "step": 331
+ },
+ {
+ "epoch": 4.614847161572053,
+ "grad_norm": 0.027223890647292137,
+ "learning_rate": 0.0006,
+ "loss": 6.166882038116455,
+ "step": 332
+ },
+ {
+ "epoch": 4.62882096069869,
+ "grad_norm": 0.03160124272108078,
+ "learning_rate": 0.0006,
+ "loss": 6.230309963226318,
+ "step": 333
+ },
+ {
+ "epoch": 4.642794759825327,
+ "grad_norm": 0.04110539332032204,
+ "learning_rate": 0.0006,
+ "loss": 6.2334818840026855,
+ "step": 334
+ },
+ {
+ "epoch": 4.656768558951965,
+ "grad_norm": 0.04355933889746666,
+ "learning_rate": 0.0006,
+ "loss": 6.160956382751465,
+ "step": 335
+ },
+ {
+ "epoch": 4.670742358078603,
+ "grad_norm": 0.03611086308956146,
+ "learning_rate": 0.0006,
+ "loss": 6.23399543762207,
+ "step": 336
+ },
+ {
+ "epoch": 4.68471615720524,
+ "grad_norm": 0.051811400800943375,
+ "learning_rate": 0.0006,
+ "loss": 6.241855144500732,
+ "step": 337
+ },
+ {
+ "epoch": 4.698689956331878,
+ "grad_norm": 0.057434193789958954,
+ "learning_rate": 0.0006,
+ "loss": 6.220993995666504,
+ "step": 338
+ },
+ {
+ "epoch": 4.712663755458515,
+ "grad_norm": 0.04420953616499901,
+ "learning_rate": 0.0006,
+ "loss": 6.300461292266846,
+ "step": 339
+ },
+ {
+ "epoch": 4.726637554585153,
+ "grad_norm": 0.03515457361936569,
+ "learning_rate": 0.0006,
+ "loss": 6.269349098205566,
+ "step": 340
+ },
+ {
+ "epoch": 4.74061135371179,
+ "grad_norm": 0.03222600743174553,
+ "learning_rate": 0.0006,
+ "loss": 6.248613357543945,
+ "step": 341
+ },
+ {
+ "epoch": 4.754585152838428,
+ "grad_norm": 0.0384046845138073,
+ "learning_rate": 0.0006,
+ "loss": 6.266979217529297,
+ "step": 342
+ },
+ {
+ "epoch": 4.7685589519650655,
+ "grad_norm": 0.0514867827296257,
+ "learning_rate": 0.0006,
+ "loss": 6.234103202819824,
+ "step": 343
+ },
+ {
+ "epoch": 4.782532751091703,
+ "grad_norm": 0.050747793167829514,
+ "learning_rate": 0.0006,
+ "loss": 6.322582244873047,
+ "step": 344
+ },
+ {
+ "epoch": 4.796506550218341,
+ "grad_norm": 0.040878988802433014,
+ "learning_rate": 0.0006,
+ "loss": 6.212509632110596,
+ "step": 345
+ },
+ {
+ "epoch": 4.810480349344978,
+ "grad_norm": 0.034655820578336716,
+ "learning_rate": 0.0006,
+ "loss": 6.265879154205322,
+ "step": 346
+ },
+ {
+ "epoch": 4.824454148471616,
+ "grad_norm": 0.04202224686741829,
+ "learning_rate": 0.0006,
+ "loss": 6.2896409034729,
+ "step": 347
+ },
+ {
+ "epoch": 4.8384279475982535,
+ "grad_norm": 0.03632277995347977,
+ "learning_rate": 0.0006,
+ "loss": 6.253917217254639,
+ "step": 348
+ },
+ {
+ "epoch": 4.85240174672489,
+ "grad_norm": 0.028645765036344528,
+ "learning_rate": 0.0006,
+ "loss": 6.207965850830078,
+ "step": 349
+ },
+ {
+ "epoch": 4.866375545851528,
+ "grad_norm": 0.025305170565843582,
+ "learning_rate": 0.0006,
+ "loss": 6.168068885803223,
+ "step": 350
+ },
+ {
+ "epoch": 4.880349344978166,
+ "grad_norm": 0.022137803956866264,
+ "learning_rate": 0.0006,
+ "loss": 6.235283374786377,
+ "step": 351
+ },
+ {
+ "epoch": 4.894323144104804,
+ "grad_norm": 0.01814538985490799,
+ "learning_rate": 0.0006,
+ "loss": 6.259641647338867,
+ "step": 352
+ },
+ {
+ "epoch": 4.908296943231441,
+ "grad_norm": 0.018465086817741394,
+ "learning_rate": 0.0006,
+ "loss": 6.111738681793213,
+ "step": 353
+ },
+ {
+ "epoch": 4.922270742358078,
+ "grad_norm": 0.018638933077454567,
+ "learning_rate": 0.0006,
+ "loss": 6.225642204284668,
+ "step": 354
+ },
+ {
+ "epoch": 4.936244541484716,
+ "grad_norm": 0.016461260616779327,
+ "learning_rate": 0.0006,
+ "loss": 6.140647888183594,
+ "step": 355
+ },
+ {
+ "epoch": 4.950218340611354,
+ "grad_norm": 0.015458385460078716,
+ "learning_rate": 0.0006,
+ "loss": 6.155224800109863,
+ "step": 356
+ },
+ {
+ "epoch": 4.964192139737992,
+ "grad_norm": 0.014934048056602478,
+ "learning_rate": 0.0006,
+ "loss": 6.101555824279785,
+ "step": 357
+ },
+ {
+ "epoch": 4.978165938864628,
+ "grad_norm": 0.013979545794427395,
+ "learning_rate": 0.0006,
+ "loss": 6.161718845367432,
+ "step": 358
+ },
+ {
+ "epoch": 4.992139737991266,
+ "grad_norm": 0.01640394888818264,
+ "learning_rate": 0.0006,
+ "loss": 6.055768966674805,
+ "step": 359
+ },
+ {
+ "epoch": 5.0,
+ "grad_norm": 0.015135680325329304,
+ "learning_rate": 0.0006,
+ "loss": 6.169695854187012,
+ "step": 360
+ },
+ {
+ "epoch": 5.0,
+ "eval_loss": 6.16408109664917,
+ "eval_runtime": 59.3649,
+ "eval_samples_per_second": 41.135,
+ "eval_steps_per_second": 1.297,
+ "step": 360
+ },
+ {
+ "epoch": 5.013973799126638,
+ "grad_norm": 0.012849048711359501,
+ "learning_rate": 0.0006,
+ "loss": 6.055830001831055,
+ "step": 361
+ },
+ {
+ "epoch": 5.0279475982532755,
+ "grad_norm": 0.018169576302170753,
+ "learning_rate": 0.0006,
+ "loss": 6.082655429840088,
+ "step": 362
+ },
+ {
+ "epoch": 5.041921397379912,
+ "grad_norm": 0.024293430149555206,
+ "learning_rate": 0.0006,
+ "loss": 6.109032154083252,
+ "step": 363
+ },
+ {
+ "epoch": 5.05589519650655,
+ "grad_norm": 0.03266787901520729,
+ "learning_rate": 0.0006,
+ "loss": 6.030613899230957,
+ "step": 364
+ },
+ {
+ "epoch": 5.069868995633188,
+ "grad_norm": 0.03572266176342964,
+ "learning_rate": 0.0006,
+ "loss": 6.050390720367432,
+ "step": 365
+ },
+ {
+ "epoch": 5.083842794759826,
+ "grad_norm": 0.032625213265419006,
+ "learning_rate": 0.0006,
+ "loss": 6.169313430786133,
+ "step": 366
+ },
+ {
+ "epoch": 5.097816593886463,
+ "grad_norm": 0.04051872715353966,
+ "learning_rate": 0.0006,
+ "loss": 6.10398006439209,
+ "step": 367
+ },
+ {
+ "epoch": 5.1117903930131,
+ "grad_norm": 0.0351213738322258,
+ "learning_rate": 0.0006,
+ "loss": 6.121973514556885,
+ "step": 368
+ },
+ {
+ "epoch": 5.125764192139738,
+ "grad_norm": 0.03523759916424751,
+ "learning_rate": 0.0006,
+ "loss": 6.0626654624938965,
+ "step": 369
+ },
+ {
+ "epoch": 5.139737991266376,
+ "grad_norm": 0.02706182189285755,
+ "learning_rate": 0.0006,
+ "loss": 6.049354076385498,
+ "step": 370
+ },
+ {
+ "epoch": 5.153711790393013,
+ "grad_norm": 0.024212589487433434,
+ "learning_rate": 0.0006,
+ "loss": 6.034826278686523,
+ "step": 371
+ },
+ {
+ "epoch": 5.1676855895196505,
+ "grad_norm": 0.024798082187771797,
+ "learning_rate": 0.0006,
+ "loss": 6.004058837890625,
+ "step": 372
+ },
+ {
+ "epoch": 5.181659388646288,
+ "grad_norm": 0.02325567416846752,
+ "learning_rate": 0.0006,
+ "loss": 5.986461639404297,
+ "step": 373
+ },
+ {
+ "epoch": 5.195633187772926,
+ "grad_norm": 0.019060570746660233,
+ "learning_rate": 0.0006,
+ "loss": 6.059736251831055,
+ "step": 374
+ },
+ {
+ "epoch": 5.209606986899563,
+ "grad_norm": 0.016822976991534233,
+ "learning_rate": 0.0006,
+ "loss": 6.072957515716553,
+ "step": 375
+ },
+ {
+ "epoch": 5.223580786026201,
+ "grad_norm": 0.017218658700585365,
+ "learning_rate": 0.0006,
+ "loss": 6.031739234924316,
+ "step": 376
+ },
+ {
+ "epoch": 5.2375545851528384,
+ "grad_norm": 0.014889383688569069,
+ "learning_rate": 0.0006,
+ "loss": 6.025674343109131,
+ "step": 377
+ },
+ {
+ "epoch": 5.251528384279476,
+ "grad_norm": 0.01708107627928257,
+ "learning_rate": 0.0006,
+ "loss": 6.090451240539551,
+ "step": 378
+ },
+ {
+ "epoch": 5.265502183406113,
+ "grad_norm": 0.020260317251086235,
+ "learning_rate": 0.0006,
+ "loss": 6.094257354736328,
+ "step": 379
+ },
+ {
+ "epoch": 5.279475982532751,
+ "grad_norm": 0.020110400393605232,
+ "learning_rate": 0.0006,
+ "loss": 6.021188735961914,
+ "step": 380
+ },
+ {
+ "epoch": 5.293449781659389,
+ "grad_norm": 0.02027830481529236,
+ "learning_rate": 0.0006,
+ "loss": 6.022156238555908,
+ "step": 381
+ },
+ {
+ "epoch": 5.307423580786026,
+ "grad_norm": 0.022747062146663666,
+ "learning_rate": 0.0006,
+ "loss": 6.011836051940918,
+ "step": 382
+ },
+ {
+ "epoch": 5.321397379912664,
+ "grad_norm": 0.01990230567753315,
+ "learning_rate": 0.0006,
+ "loss": 6.003054618835449,
+ "step": 383
+ },
+ {
+ "epoch": 5.335371179039301,
+ "grad_norm": 0.01596238650381565,
+ "learning_rate": 0.0006,
+ "loss": 5.991410732269287,
+ "step": 384
+ },
+ {
+ "epoch": 5.349344978165939,
+ "grad_norm": 0.015847105532884598,
+ "learning_rate": 0.0006,
+ "loss": 6.100622177124023,
+ "step": 385
+ },
+ {
+ "epoch": 5.3633187772925766,
+ "grad_norm": 0.016179397702217102,
+ "learning_rate": 0.0006,
+ "loss": 6.032659530639648,
+ "step": 386
+ },
+ {
+ "epoch": 5.377292576419214,
+ "grad_norm": 0.018256602808833122,
+ "learning_rate": 0.0006,
+ "loss": 5.961983680725098,
+ "step": 387
+ },
+ {
+ "epoch": 5.391266375545851,
+ "grad_norm": 0.02805912122130394,
+ "learning_rate": 0.0006,
+ "loss": 5.983541965484619,
+ "step": 388
+ },
+ {
+ "epoch": 5.405240174672489,
+ "grad_norm": 0.039082255214452744,
+ "learning_rate": 0.0006,
+ "loss": 6.021474361419678,
+ "step": 389
+ },
+ {
+ "epoch": 5.419213973799127,
+ "grad_norm": 0.036757659167051315,
+ "learning_rate": 0.0006,
+ "loss": 6.005046367645264,
+ "step": 390
+ },
+ {
+ "epoch": 5.4331877729257645,
+ "grad_norm": 0.035277366638183594,
+ "learning_rate": 0.0006,
+ "loss": 6.036296844482422,
+ "step": 391
+ },
+ {
+ "epoch": 5.447161572052401,
+ "grad_norm": 0.034404635429382324,
+ "learning_rate": 0.0006,
+ "loss": 5.994539260864258,
+ "step": 392
+ },
+ {
+ "epoch": 5.461135371179039,
+ "grad_norm": 0.0377206988632679,
+ "learning_rate": 0.0006,
+ "loss": 5.941534519195557,
+ "step": 393
+ },
+ {
+ "epoch": 5.475109170305677,
+ "grad_norm": 0.0389922633767128,
+ "learning_rate": 0.0006,
+ "loss": 5.983644485473633,
+ "step": 394
+ },
+ {
+ "epoch": 5.489082969432315,
+ "grad_norm": 0.04176778718829155,
+ "learning_rate": 0.0006,
+ "loss": 6.030922889709473,
+ "step": 395
+ },
+ {
+ "epoch": 5.503056768558952,
+ "grad_norm": 0.0337153784930706,
+ "learning_rate": 0.0006,
+ "loss": 5.961367130279541,
+ "step": 396
+ },
+ {
+ "epoch": 5.517030567685589,
+ "grad_norm": 0.036116816103458405,
+ "learning_rate": 0.0006,
+ "loss": 5.819280624389648,
+ "step": 397
+ },
+ {
+ "epoch": 5.531004366812227,
+ "grad_norm": 0.030724041163921356,
+ "learning_rate": 0.0006,
+ "loss": 5.906380653381348,
+ "step": 398
+ },
+ {
+ "epoch": 5.544978165938865,
+ "grad_norm": 0.030264202505350113,
+ "learning_rate": 0.0006,
+ "loss": 5.975833415985107,
+ "step": 399
+ },
+ {
+ "epoch": 5.558951965065502,
+ "grad_norm": 0.032096318900585175,
+ "learning_rate": 0.0006,
+ "loss": 6.001348972320557,
+ "step": 400
+ },
+ {
+ "epoch": 5.5729257641921395,
+ "grad_norm": 0.030579356476664543,
+ "learning_rate": 0.0006,
+ "loss": 5.928768634796143,
+ "step": 401
+ },
+ {
+ "epoch": 5.586899563318777,
+ "grad_norm": 0.028241973370313644,
+ "learning_rate": 0.0006,
+ "loss": 5.92582893371582,
+ "step": 402
+ },
+ {
+ "epoch": 5.600873362445415,
+ "grad_norm": 0.023086953908205032,
+ "learning_rate": 0.0006,
+ "loss": 5.892926216125488,
+ "step": 403
+ },
+ {
+ "epoch": 5.614847161572053,
+ "grad_norm": 0.02547992393374443,
+ "learning_rate": 0.0006,
+ "loss": 5.903195381164551,
+ "step": 404
+ },
+ {
+ "epoch": 5.62882096069869,
+ "grad_norm": 0.023089831694960594,
+ "learning_rate": 0.0006,
+ "loss": 5.912033557891846,
+ "step": 405
+ },
+ {
+ "epoch": 5.642794759825327,
+ "grad_norm": 0.025446368381381035,
+ "learning_rate": 0.0006,
+ "loss": 5.926138401031494,
+ "step": 406
+ },
+ {
+ "epoch": 5.656768558951965,
+ "grad_norm": 0.03748747706413269,
+ "learning_rate": 0.0006,
+ "loss": 5.913451194763184,
+ "step": 407
+ },
+ {
+ "epoch": 5.670742358078603,
+ "grad_norm": 0.03493810072541237,
+ "learning_rate": 0.0006,
+ "loss": 5.97633695602417,
+ "step": 408
+ },
+ {
+ "epoch": 5.68471615720524,
+ "grad_norm": 0.023131275549530983,
+ "learning_rate": 0.0006,
+ "loss": 5.899571418762207,
+ "step": 409
+ },
+ {
+ "epoch": 5.698689956331878,
+ "grad_norm": 0.02986014075577259,
+ "learning_rate": 0.0006,
+ "loss": 5.894903182983398,
+ "step": 410
+ },
+ {
+ "epoch": 5.712663755458515,
+ "grad_norm": 0.030171144753694534,
+ "learning_rate": 0.0006,
+ "loss": 5.876595497131348,
+ "step": 411
+ },
+ {
+ "epoch": 5.726637554585153,
+ "grad_norm": 0.029377546161413193,
+ "learning_rate": 0.0006,
+ "loss": 5.911637306213379,
+ "step": 412
+ },
+ {
+ "epoch": 5.74061135371179,
+ "grad_norm": 0.029829490929841995,
+ "learning_rate": 0.0006,
+ "loss": 5.819629669189453,
+ "step": 413
+ },
+ {
+ "epoch": 5.754585152838428,
+ "grad_norm": 0.024139299988746643,
+ "learning_rate": 0.0006,
+ "loss": 5.81728458404541,
+ "step": 414
+ },
+ {
+ "epoch": 5.7685589519650655,
+ "grad_norm": 0.021453432738780975,
+ "learning_rate": 0.0006,
+ "loss": 5.987326145172119,
+ "step": 415
+ },
+ {
+ "epoch": 5.782532751091703,
+ "grad_norm": 0.017942246049642563,
+ "learning_rate": 0.0006,
+ "loss": 5.893008232116699,
+ "step": 416
+ },
+ {
+ "epoch": 5.796506550218341,
+ "grad_norm": 0.019724637269973755,
+ "learning_rate": 0.0006,
+ "loss": 5.912441253662109,
+ "step": 417
+ },
+ {
+ "epoch": 5.810480349344978,
+ "grad_norm": 0.018597912043333054,
+ "learning_rate": 0.0006,
+ "loss": 5.860394477844238,
+ "step": 418
+ },
+ {
+ "epoch": 5.824454148471616,
+ "grad_norm": 0.016232412308454514,
+ "learning_rate": 0.0006,
+ "loss": 5.94991397857666,
+ "step": 419
+ },
+ {
+ "epoch": 5.8384279475982535,
+ "grad_norm": 0.01667204685509205,
+ "learning_rate": 0.0006,
+ "loss": 5.8506364822387695,
+ "step": 420
+ },
+ {
+ "epoch": 5.85240174672489,
+ "grad_norm": 0.014232151210308075,
+ "learning_rate": 0.0006,
+ "loss": 5.928424835205078,
+ "step": 421
+ },
+ {
+ "epoch": 5.866375545851528,
+ "grad_norm": 0.01570476032793522,
+ "learning_rate": 0.0006,
+ "loss": 5.953692436218262,
+ "step": 422
+ },
+ {
+ "epoch": 5.880349344978166,
+ "grad_norm": 0.01586179807782173,
+ "learning_rate": 0.0006,
+ "loss": 5.820014476776123,
+ "step": 423
+ },
+ {
+ "epoch": 5.894323144104804,
+ "grad_norm": 0.014729145914316177,
+ "learning_rate": 0.0006,
+ "loss": 5.860154628753662,
+ "step": 424
+ },
+ {
+ "epoch": 5.908296943231441,
+ "grad_norm": 0.013111459091305733,
+ "learning_rate": 0.0006,
+ "loss": 5.85988187789917,
+ "step": 425
+ },
+ {
+ "epoch": 5.922270742358078,
+ "grad_norm": 0.012937922962009907,
+ "learning_rate": 0.0006,
+ "loss": 5.756265640258789,
+ "step": 426
+ },
+ {
+ "epoch": 5.936244541484716,
+ "grad_norm": 0.01377453189343214,
+ "learning_rate": 0.0006,
+ "loss": 5.8333048820495605,
+ "step": 427
+ },
+ {
+ "epoch": 5.950218340611354,
+ "grad_norm": 0.014054795727133751,
+ "learning_rate": 0.0006,
+ "loss": 5.803333759307861,
+ "step": 428
+ },
+ {
+ "epoch": 5.964192139737992,
+ "grad_norm": 0.01674959622323513,
+ "learning_rate": 0.0006,
+ "loss": 5.8718485832214355,
+ "step": 429
+ },
+ {
+ "epoch": 5.978165938864628,
+ "grad_norm": 0.018733017146587372,
+ "learning_rate": 0.0006,
+ "loss": 5.875979900360107,
+ "step": 430
+ },
+ {
+ "epoch": 5.992139737991266,
+ "grad_norm": 0.02088983915746212,
+ "learning_rate": 0.0006,
+ "loss": 5.763634204864502,
+ "step": 431
+ },
+ {
+ "epoch": 6.0,
+ "grad_norm": 0.026705985888838768,
+ "learning_rate": 0.0006,
+ "loss": 5.6922478675842285,
+ "step": 432
+ },
+ {
+ "epoch": 6.0,
+ "eval_loss": 5.847958087921143,
+ "eval_runtime": 58.5451,
+ "eval_samples_per_second": 41.711,
+ "eval_steps_per_second": 1.315,
+ "step": 432
+ },
+ {
+ "epoch": 6.013973799126638,
+ "grad_norm": 0.03181544691324234,
+ "learning_rate": 0.0006,
+ "loss": 5.786368370056152,
+ "step": 433
+ },
+ {
+ "epoch": 6.0279475982532755,
+ "grad_norm": 0.03238112851977348,
+ "learning_rate": 0.0006,
+ "loss": 5.810310363769531,
+ "step": 434
+ },
+ {
+ "epoch": 6.041921397379912,
+ "grad_norm": 0.03916756808757782,
+ "learning_rate": 0.0006,
+ "loss": 5.7761993408203125,
+ "step": 435
+ },
+ {
+ "epoch": 6.05589519650655,
+ "grad_norm": 0.044009730219841,
+ "learning_rate": 0.0006,
+ "loss": 5.898112773895264,
+ "step": 436
+ },
+ {
+ "epoch": 6.069868995633188,
+ "grad_norm": 0.045235246419906616,
+ "learning_rate": 0.0006,
+ "loss": 5.876371383666992,
+ "step": 437
+ },
+ {
+ "epoch": 6.083842794759826,
+ "grad_norm": 0.050218384712934494,
+ "learning_rate": 0.0006,
+ "loss": 5.805103302001953,
+ "step": 438
+ },
+ {
+ "epoch": 6.097816593886463,
+ "grad_norm": 0.05444490164518356,
+ "learning_rate": 0.0006,
+ "loss": 5.826424598693848,
+ "step": 439
+ },
+ {
+ "epoch": 6.1117903930131,
+ "grad_norm": 0.04890982061624527,
+ "learning_rate": 0.0006,
+ "loss": 5.7735395431518555,
+ "step": 440
+ },
+ {
+ "epoch": 6.125764192139738,
+ "grad_norm": 0.05478639155626297,
+ "learning_rate": 0.0006,
+ "loss": 5.823677062988281,
+ "step": 441
+ },
+ {
+ "epoch": 6.139737991266376,
+ "grad_norm": 0.057562313973903656,
+ "learning_rate": 0.0006,
+ "loss": 5.889334678649902,
+ "step": 442
+ },
+ {
+ "epoch": 6.153711790393013,
+ "grad_norm": 0.06447703391313553,
+ "learning_rate": 0.0006,
+ "loss": 5.84621524810791,
+ "step": 443
+ },
+ {
+ "epoch": 6.1676855895196505,
+ "grad_norm": 0.048861872404813766,
+ "learning_rate": 0.0006,
+ "loss": 5.931595802307129,
+ "step": 444
+ },
+ {
+ "epoch": 6.181659388646288,
+ "grad_norm": 0.05521339178085327,
+ "learning_rate": 0.0006,
+ "loss": 5.8229780197143555,
+ "step": 445
+ },
+ {
+ "epoch": 6.195633187772926,
+ "grad_norm": 0.053666990250349045,
+ "learning_rate": 0.0006,
+ "loss": 5.879191875457764,
+ "step": 446
+ },
+ {
+ "epoch": 6.209606986899563,
+ "grad_norm": 0.0451025515794754,
+ "learning_rate": 0.0006,
+ "loss": 5.801628112792969,
+ "step": 447
+ },
+ {
+ "epoch": 6.223580786026201,
+ "grad_norm": 0.04291655868291855,
+ "learning_rate": 0.0006,
+ "loss": 5.914680480957031,
+ "step": 448
+ },
+ {
+ "epoch": 6.2375545851528384,
+ "grad_norm": 0.035102490335702896,
+ "learning_rate": 0.0006,
+ "loss": 5.843682765960693,
+ "step": 449
+ },
+ {
+ "epoch": 6.251528384279476,
+ "grad_norm": 0.03403995931148529,
+ "learning_rate": 0.0006,
+ "loss": 5.827154159545898,
+ "step": 450
+ },
+ {
+ "epoch": 6.265502183406113,
+ "grad_norm": 0.03444375470280647,
+ "learning_rate": 0.0006,
+ "loss": 5.800136566162109,
+ "step": 451
+ },
+ {
+ "epoch": 6.279475982532751,
+ "grad_norm": 0.03165159001946449,
+ "learning_rate": 0.0006,
+ "loss": 5.906252384185791,
+ "step": 452
+ },
+ {
+ "epoch": 6.293449781659389,
+ "grad_norm": 0.026153093203902245,
+ "learning_rate": 0.0006,
+ "loss": 5.7423200607299805,
+ "step": 453
+ },
+ {
+ "epoch": 6.307423580786026,
+ "grad_norm": 0.024357657879590988,
+ "learning_rate": 0.0006,
+ "loss": 5.701364517211914,
+ "step": 454
+ },
+ {
+ "epoch": 6.321397379912664,
+ "grad_norm": 0.021508049219846725,
+ "learning_rate": 0.0006,
+ "loss": 5.8328094482421875,
+ "step": 455
+ },
+ {
+ "epoch": 6.335371179039301,
+ "grad_norm": 0.017313921824097633,
+ "learning_rate": 0.0006,
+ "loss": 5.792145252227783,
+ "step": 456
+ },
+ {
+ "epoch": 6.349344978165939,
+ "grad_norm": 0.018563397228717804,
+ "learning_rate": 0.0006,
+ "loss": 5.732672214508057,
+ "step": 457
+ },
+ {
+ "epoch": 6.3633187772925766,
+ "grad_norm": 0.016568679362535477,
+ "learning_rate": 0.0006,
+ "loss": 5.81948184967041,
+ "step": 458
+ },
+ {
+ "epoch": 6.377292576419214,
+ "grad_norm": 0.014133770950138569,
+ "learning_rate": 0.0006,
+ "loss": 5.734982490539551,
+ "step": 459
+ },
+ {
+ "epoch": 6.391266375545851,
+ "grad_norm": 0.014114447869360447,
+ "learning_rate": 0.0006,
+ "loss": 5.70073127746582,
+ "step": 460
+ },
+ {
+ "epoch": 6.405240174672489,
+ "grad_norm": 0.012907393276691437,
+ "learning_rate": 0.0006,
+ "loss": 5.758626937866211,
+ "step": 461
+ },
+ {
+ "epoch": 6.419213973799127,
+ "grad_norm": 0.012673444114625454,
+ "learning_rate": 0.0006,
+ "loss": 5.775138854980469,
+ "step": 462
+ },
+ {
+ "epoch": 6.4331877729257645,
+ "grad_norm": 0.01311410591006279,
+ "learning_rate": 0.0006,
+ "loss": 5.70920991897583,
+ "step": 463
+ },
+ {
+ "epoch": 6.447161572052401,
+ "grad_norm": 0.012935544364154339,
+ "learning_rate": 0.0006,
+ "loss": 5.781479835510254,
+ "step": 464
+ },
+ {
+ "epoch": 6.461135371179039,
+ "grad_norm": 0.01439738366752863,
+ "learning_rate": 0.0006,
+ "loss": 5.691315650939941,
+ "step": 465
+ },
+ {
+ "epoch": 6.475109170305677,
+ "grad_norm": 0.01205186452716589,
+ "learning_rate": 0.0006,
+ "loss": 5.662499904632568,
+ "step": 466
+ },
+ {
+ "epoch": 6.489082969432315,
+ "grad_norm": 0.011575652286410332,
+ "learning_rate": 0.0006,
+ "loss": 5.6271772384643555,
+ "step": 467
+ },
+ {
+ "epoch": 6.503056768558952,
+ "grad_norm": 0.011672502383589745,
+ "learning_rate": 0.0006,
+ "loss": 5.761662006378174,
+ "step": 468
+ },
+ {
+ "epoch": 6.517030567685589,
+ "grad_norm": 0.011541751213371754,
+ "learning_rate": 0.0006,
+ "loss": 5.762078285217285,
+ "step": 469
+ },
+ {
+ "epoch": 6.531004366812227,
+ "grad_norm": 0.011396365240216255,
+ "learning_rate": 0.0006,
+ "loss": 5.67873477935791,
+ "step": 470
+ },
+ {
+ "epoch": 6.544978165938865,
+ "grad_norm": 0.010688499547541142,
+ "learning_rate": 0.0006,
+ "loss": 5.717051982879639,
+ "step": 471
+ },
+ {
+ "epoch": 6.558951965065502,
+ "grad_norm": 0.012224317528307438,
+ "learning_rate": 0.0006,
+ "loss": 5.707948684692383,
+ "step": 472
+ },
+ {
+ "epoch": 6.5729257641921395,
+ "grad_norm": 0.011856825090944767,
+ "learning_rate": 0.0006,
+ "loss": 5.70878791809082,
+ "step": 473
+ },
+ {
+ "epoch": 6.586899563318777,
+ "grad_norm": 0.01199064590036869,
+ "learning_rate": 0.0006,
+ "loss": 5.708697319030762,
+ "step": 474
+ },
+ {
+ "epoch": 6.600873362445415,
+ "grad_norm": 0.01219446212053299,
+ "learning_rate": 0.0006,
+ "loss": 5.6061577796936035,
+ "step": 475
+ },
+ {
+ "epoch": 6.614847161572053,
+ "grad_norm": 0.013397484086453915,
+ "learning_rate": 0.0006,
+ "loss": 5.624789714813232,
+ "step": 476
+ },
+ {
+ "epoch": 6.62882096069869,
+ "grad_norm": 0.01483136136084795,
+ "learning_rate": 0.0006,
+ "loss": 5.721141338348389,
+ "step": 477
+ },
+ {
+ "epoch": 6.642794759825327,
+ "grad_norm": 0.0194488987326622,
+ "learning_rate": 0.0006,
+ "loss": 5.604279518127441,
+ "step": 478
+ },
+ {
+ "epoch": 6.656768558951965,
+ "grad_norm": 0.02079896256327629,
+ "learning_rate": 0.0006,
+ "loss": 5.586322784423828,
+ "step": 479
+ },
+ {
+ "epoch": 6.670742358078603,
+ "grad_norm": 0.018996229395270348,
+ "learning_rate": 0.0006,
+ "loss": 5.5987772941589355,
+ "step": 480
+ },
+ {
+ "epoch": 6.68471615720524,
+ "grad_norm": 0.015779603272676468,
+ "learning_rate": 0.0006,
+ "loss": 5.588602066040039,
+ "step": 481
+ },
+ {
+ "epoch": 6.698689956331878,
+ "grad_norm": 0.015322973020374775,
+ "learning_rate": 0.0006,
+ "loss": 5.686467170715332,
+ "step": 482
+ },
+ {
+ "epoch": 6.712663755458515,
+ "grad_norm": 0.015282213687896729,
+ "learning_rate": 0.0006,
+ "loss": 5.664676666259766,
+ "step": 483
+ },
+ {
+ "epoch": 6.726637554585153,
+ "grad_norm": 0.016589272767305374,
+ "learning_rate": 0.0006,
+ "loss": 5.666739463806152,
+ "step": 484
+ },
+ {
+ "epoch": 6.74061135371179,
+ "grad_norm": 0.01811421848833561,
+ "learning_rate": 0.0006,
+ "loss": 5.647593021392822,
+ "step": 485
+ },
+ {
+ "epoch": 6.754585152838428,
+ "grad_norm": 0.019652029499411583,
+ "learning_rate": 0.0006,
+ "loss": 5.628726959228516,
+ "step": 486
+ },
+ {
+ "epoch": 6.7685589519650655,
+ "grad_norm": 0.02118665538728237,
+ "learning_rate": 0.0006,
+ "loss": 5.6580705642700195,
+ "step": 487
+ },
+ {
+ "epoch": 6.782532751091703,
+ "grad_norm": 0.02237832546234131,
+ "learning_rate": 0.0006,
+ "loss": 5.635931968688965,
+ "step": 488
+ },
+ {
+ "epoch": 6.796506550218341,
+ "grad_norm": 0.023897631093859673,
+ "learning_rate": 0.0006,
+ "loss": 5.517949104309082,
+ "step": 489
+ },
+ {
+ "epoch": 6.810480349344978,
+ "grad_norm": 0.02442227490246296,
+ "learning_rate": 0.0006,
+ "loss": 5.600021839141846,
+ "step": 490
+ },
+ {
+ "epoch": 6.824454148471616,
+ "grad_norm": 0.024675287306308746,
+ "learning_rate": 0.0006,
+ "loss": 5.6639909744262695,
+ "step": 491
+ },
+ {
+ "epoch": 6.8384279475982535,
+ "grad_norm": 0.030372392386198044,
+ "learning_rate": 0.0006,
+ "loss": 5.582888603210449,
+ "step": 492
+ },
+ {
+ "epoch": 6.85240174672489,
+ "grad_norm": 0.03237690031528473,
+ "learning_rate": 0.0006,
+ "loss": 5.573896884918213,
+ "step": 493
+ },
+ {
+ "epoch": 6.866375545851528,
+ "grad_norm": 0.03511589393019676,
+ "learning_rate": 0.0006,
+ "loss": 5.565878868103027,
+ "step": 494
+ },
+ {
+ "epoch": 6.880349344978166,
+ "grad_norm": 0.03883938118815422,
+ "learning_rate": 0.0006,
+ "loss": 5.58540678024292,
+ "step": 495
+ },
+ {
+ "epoch": 6.894323144104804,
+ "grad_norm": 0.04175502806901932,
+ "learning_rate": 0.0006,
+ "loss": 5.6095194816589355,
+ "step": 496
+ },
+ {
+ "epoch": 6.908296943231441,
+ "grad_norm": 0.04013441503047943,
+ "learning_rate": 0.0006,
+ "loss": 5.654401779174805,
+ "step": 497
+ },
+ {
+ "epoch": 6.922270742358078,
+ "grad_norm": 0.04618770629167557,
+ "learning_rate": 0.0006,
+ "loss": 5.6483473777771,
+ "step": 498
+ },
+ {
+ "epoch": 6.936244541484716,
+ "grad_norm": 0.04507308453321457,
+ "learning_rate": 0.0006,
+ "loss": 5.5856218338012695,
+ "step": 499
+ },
+ {
+ "epoch": 6.950218340611354,
+ "grad_norm": 0.03151383996009827,
+ "learning_rate": 0.0006,
+ "loss": 5.594086647033691,
+ "step": 500
+ },
+ {
+ "epoch": 6.964192139737992,
+ "grad_norm": 0.026883797720074654,
+ "learning_rate": 0.0006,
+ "loss": 5.607676029205322,
+ "step": 501
+ },
+ {
+ "epoch": 6.978165938864628,
+ "grad_norm": 0.02981836162507534,
+ "learning_rate": 0.0006,
+ "loss": 5.563666820526123,
+ "step": 502
+ },
+ {
+ "epoch": 6.992139737991266,
+ "grad_norm": 0.03233477845788002,
+ "learning_rate": 0.0006,
+ "loss": 5.667543411254883,
+ "step": 503
+ },
+ {
+ "epoch": 7.0,
+ "grad_norm": 0.029927456751465797,
+ "learning_rate": 0.0006,
+ "loss": 5.733482837677002,
+ "step": 504
+ },
+ {
+ "epoch": 7.0,
+ "eval_loss": 5.679076671600342,
+ "eval_runtime": 59.5965,
+ "eval_samples_per_second": 40.976,
+ "eval_steps_per_second": 1.292,
+ "step": 504
+ },
+ {
+ "epoch": 7.013973799126638,
+ "grad_norm": 0.027081597596406937,
+ "learning_rate": 0.0006,
+ "loss": 5.634487152099609,
+ "step": 505
+ },
+ {
+ "epoch": 7.0279475982532755,
+ "grad_norm": 0.031959421932697296,
+ "learning_rate": 0.0006,
+ "loss": 5.643294811248779,
+ "step": 506
+ },
+ {
+ "epoch": 7.041921397379912,
+ "grad_norm": 0.02802063524723053,
+ "learning_rate": 0.0006,
+ "loss": 5.573239326477051,
+ "step": 507
+ },
+ {
+ "epoch": 7.05589519650655,
+ "grad_norm": 0.032579705119132996,
+ "learning_rate": 0.0006,
+ "loss": 5.627150058746338,
+ "step": 508
+ },
+ {
+ "epoch": 7.069868995633188,
+ "grad_norm": 0.02503928542137146,
+ "learning_rate": 0.0006,
+ "loss": 5.6045308113098145,
+ "step": 509
+ },
+ {
+ "epoch": 7.083842794759826,
+ "grad_norm": 0.02421317622065544,
+ "learning_rate": 0.0006,
+ "loss": 5.522153377532959,
+ "step": 510
+ },
+ {
+ "epoch": 7.097816593886463,
+ "grad_norm": 0.021983426064252853,
+ "learning_rate": 0.0006,
+ "loss": 5.632939338684082,
+ "step": 511
+ },
+ {
+ "epoch": 7.1117903930131,
+ "grad_norm": 0.021933183073997498,
+ "learning_rate": 0.0006,
+ "loss": 5.592185974121094,
+ "step": 512
+ },
+ {
+ "epoch": 7.125764192139738,
+ "grad_norm": 0.02287031151354313,
+ "learning_rate": 0.0006,
+ "loss": 5.613700866699219,
+ "step": 513
+ },
+ {
+ "epoch": 7.139737991266376,
+ "grad_norm": 0.02281602844595909,
+ "learning_rate": 0.0006,
+ "loss": 5.551383018493652,
+ "step": 514
+ },
+ {
+ "epoch": 7.153711790393013,
+ "grad_norm": 0.021050360053777695,
+ "learning_rate": 0.0006,
+ "loss": 5.55275821685791,
+ "step": 515
+ },
+ {
+ "epoch": 7.1676855895196505,
+ "grad_norm": 0.018299926072359085,
+ "learning_rate": 0.0006,
+ "loss": 5.528225421905518,
+ "step": 516
+ },
+ {
+ "epoch": 7.181659388646288,
+ "grad_norm": 0.02024853602051735,
+ "learning_rate": 0.0006,
+ "loss": 5.522453784942627,
+ "step": 517
+ },
+ {
+ "epoch": 7.195633187772926,
+ "grad_norm": 0.02135239914059639,
+ "learning_rate": 0.0006,
+ "loss": 5.527522563934326,
+ "step": 518
+ },
+ {
+ "epoch": 7.209606986899563,
+ "grad_norm": 0.023394184187054634,
+ "learning_rate": 0.0006,
+ "loss": 5.4535017013549805,
+ "step": 519
+ },
+ {
+ "epoch": 7.223580786026201,
+ "grad_norm": 0.025952080264687538,
+ "learning_rate": 0.0006,
+ "loss": 5.5374674797058105,
+ "step": 520
+ },
+ {
+ "epoch": 7.2375545851528384,
+ "grad_norm": 0.022287718951702118,
+ "learning_rate": 0.0006,
+ "loss": 5.493753433227539,
+ "step": 521
+ },
+ {
+ "epoch": 7.251528384279476,
+ "grad_norm": 0.02018447406589985,
+ "learning_rate": 0.0006,
+ "loss": 5.510575294494629,
+ "step": 522
+ },
+ {
+ "epoch": 7.265502183406113,
+ "grad_norm": 0.02050507813692093,
+ "learning_rate": 0.0006,
+ "loss": 5.466026782989502,
+ "step": 523
+ },
+ {
+ "epoch": 7.279475982532751,
+ "grad_norm": 0.023688802495598793,
+ "learning_rate": 0.0006,
+ "loss": 5.485815525054932,
+ "step": 524
+ },
+ {
+ "epoch": 7.293449781659389,
+ "grad_norm": 0.021078843623399734,
+ "learning_rate": 0.0006,
+ "loss": 5.415581703186035,
+ "step": 525
+ },
+ {
+ "epoch": 7.307423580786026,
+ "grad_norm": 0.018266011029481888,
+ "learning_rate": 0.0006,
+ "loss": 5.402815818786621,
+ "step": 526
+ },
+ {
+ "epoch": 7.321397379912664,
+ "grad_norm": 0.019693493843078613,
+ "learning_rate": 0.0006,
+ "loss": 5.504674911499023,
+ "step": 527
+ },
+ {
+ "epoch": 7.335371179039301,
+ "grad_norm": 0.02031373605132103,
+ "learning_rate": 0.0006,
+ "loss": 5.503837585449219,
+ "step": 528
+ },
+ {
+ "epoch": 7.349344978165939,
+ "grad_norm": 0.017636626958847046,
+ "learning_rate": 0.0006,
+ "loss": 5.530580997467041,
+ "step": 529
+ },
+ {
+ "epoch": 7.3633187772925766,
+ "grad_norm": 0.01787244901061058,
+ "learning_rate": 0.0006,
+ "loss": 5.450218677520752,
+ "step": 530
+ },
+ {
+ "epoch": 7.377292576419214,
+ "grad_norm": 0.0170669574290514,
+ "learning_rate": 0.0006,
+ "loss": 5.481570243835449,
+ "step": 531
+ },
+ {
+ "epoch": 7.391266375545851,
+ "grad_norm": 0.01802165061235428,
+ "learning_rate": 0.0006,
+ "loss": 5.386394500732422,
+ "step": 532
+ },
+ {
+ "epoch": 7.405240174672489,
+ "grad_norm": 0.022949472069740295,
+ "learning_rate": 0.0006,
+ "loss": 5.424929618835449,
+ "step": 533
+ },
+ {
+ "epoch": 7.419213973799127,
+ "grad_norm": 0.034213390201330185,
+ "learning_rate": 0.0006,
+ "loss": 5.346663475036621,
+ "step": 534
+ },
+ {
+ "epoch": 7.4331877729257645,
+ "grad_norm": 0.03847593814134598,
+ "learning_rate": 0.0006,
+ "loss": 5.418482780456543,
+ "step": 535
+ },
+ {
+ "epoch": 7.447161572052401,
+ "grad_norm": 0.029393676668405533,
+ "learning_rate": 0.0006,
+ "loss": 5.456090927124023,
+ "step": 536
+ },
+ {
+ "epoch": 7.461135371179039,
+ "grad_norm": 0.03388667106628418,
+ "learning_rate": 0.0006,
+ "loss": 5.572983741760254,
+ "step": 537
+ },
+ {
+ "epoch": 7.475109170305677,
+ "grad_norm": 0.039690613746643066,
+ "learning_rate": 0.0006,
+ "loss": 5.414067268371582,
+ "step": 538
+ },
+ {
+ "epoch": 7.489082969432315,
+ "grad_norm": 0.03630776330828667,
+ "learning_rate": 0.0006,
+ "loss": 5.523739814758301,
+ "step": 539
+ },
+ {
+ "epoch": 7.503056768558952,
+ "grad_norm": 0.03356870263814926,
+ "learning_rate": 0.0006,
+ "loss": 5.444767951965332,
+ "step": 540
+ },
+ {
+ "epoch": 7.517030567685589,
+ "grad_norm": 0.030812304466962814,
+ "learning_rate": 0.0006,
+ "loss": 5.461244106292725,
+ "step": 541
+ },
+ {
+ "epoch": 7.531004366812227,
+ "grad_norm": 0.03319217637181282,
+ "learning_rate": 0.0006,
+ "loss": 5.502161026000977,
+ "step": 542
+ },
+ {
+ "epoch": 7.544978165938865,
+ "grad_norm": 0.032997481524944305,
+ "learning_rate": 0.0006,
+ "loss": 5.4162278175354,
+ "step": 543
+ },
+ {
+ "epoch": 7.558951965065502,
+ "grad_norm": 0.03364962339401245,
+ "learning_rate": 0.0006,
+ "loss": 5.415736198425293,
+ "step": 544
+ },
+ {
+ "epoch": 7.5729257641921395,
+ "grad_norm": 0.0344221405684948,
+ "learning_rate": 0.0006,
+ "loss": 5.541967868804932,
+ "step": 545
+ },
+ {
+ "epoch": 7.586899563318777,
+ "grad_norm": 0.029609503224492073,
+ "learning_rate": 0.0006,
+ "loss": 5.440614223480225,
+ "step": 546
+ },
+ {
+ "epoch": 7.600873362445415,
+ "grad_norm": 0.029055926948785782,
+ "learning_rate": 0.0006,
+ "loss": 5.462865352630615,
+ "step": 547
+ },
+ {
+ "epoch": 7.614847161572053,
+ "grad_norm": 0.02658848837018013,
+ "learning_rate": 0.0006,
+ "loss": 5.444467544555664,
+ "step": 548
+ },
+ {
+ "epoch": 7.62882096069869,
+ "grad_norm": 0.026276404038071632,
+ "learning_rate": 0.0006,
+ "loss": 5.462278842926025,
+ "step": 549
+ },
+ {
+ "epoch": 7.642794759825327,
+ "grad_norm": 0.0286889486014843,
+ "learning_rate": 0.0006,
+ "loss": 5.4849534034729,
+ "step": 550
+ },
+ {
+ "epoch": 7.656768558951965,
+ "grad_norm": 0.026849817484617233,
+ "learning_rate": 0.0006,
+ "loss": 5.431473731994629,
+ "step": 551
+ },
+ {
+ "epoch": 7.670742358078603,
+ "grad_norm": 0.02312396466732025,
+ "learning_rate": 0.0006,
+ "loss": 5.45462703704834,
+ "step": 552
+ },
+ {
+ "epoch": 7.68471615720524,
+ "grad_norm": 0.026197485625743866,
+ "learning_rate": 0.0006,
+ "loss": 5.452552318572998,
+ "step": 553
+ },
+ {
+ "epoch": 7.698689956331878,
+ "grad_norm": 0.02747255191206932,
+ "learning_rate": 0.0006,
+ "loss": 5.447169780731201,
+ "step": 554
+ },
+ {
+ "epoch": 7.712663755458515,
+ "grad_norm": 0.028123432770371437,
+ "learning_rate": 0.0006,
+ "loss": 5.394400596618652,
+ "step": 555
+ },
+ {
+ "epoch": 7.726637554585153,
+ "grad_norm": 0.02599870041012764,
+ "learning_rate": 0.0006,
+ "loss": 5.398341178894043,
+ "step": 556
+ },
+ {
+ "epoch": 7.74061135371179,
+ "grad_norm": 0.022171657532453537,
+ "learning_rate": 0.0006,
+ "loss": 5.368820667266846,
+ "step": 557
+ },
+ {
+ "epoch": 7.754585152838428,
+ "grad_norm": 0.022309480234980583,
+ "learning_rate": 0.0006,
+ "loss": 5.416772365570068,
+ "step": 558
+ },
+ {
+ "epoch": 7.7685589519650655,
+ "grad_norm": 0.024074165150523186,
+ "learning_rate": 0.0006,
+ "loss": 5.359002113342285,
+ "step": 559
+ },
+ {
+ "epoch": 7.782532751091703,
+ "grad_norm": 0.02636653557419777,
+ "learning_rate": 0.0006,
+ "loss": 5.484368324279785,
+ "step": 560
+ },
+ {
+ "epoch": 7.796506550218341,
+ "grad_norm": 0.02203752100467682,
+ "learning_rate": 0.0006,
+ "loss": 5.375498294830322,
+ "step": 561
+ },
+ {
+ "epoch": 7.810480349344978,
+ "grad_norm": 0.019924819469451904,
+ "learning_rate": 0.0006,
+ "loss": 5.334723472595215,
+ "step": 562
+ },
+ {
+ "epoch": 7.824454148471616,
+ "grad_norm": 0.018755966797471046,
+ "learning_rate": 0.0006,
+ "loss": 5.348781108856201,
+ "step": 563
+ },
+ {
+ "epoch": 7.8384279475982535,
+ "grad_norm": 0.016747845336794853,
+ "learning_rate": 0.0006,
+ "loss": 5.433079719543457,
+ "step": 564
+ },
+ {
+ "epoch": 7.85240174672489,
+ "grad_norm": 0.0170395877212286,
+ "learning_rate": 0.0006,
+ "loss": 5.346794128417969,
+ "step": 565
+ },
+ {
+ "epoch": 7.866375545851528,
+ "grad_norm": 0.016074176877737045,
+ "learning_rate": 0.0006,
+ "loss": 5.3277740478515625,
+ "step": 566
+ },
+ {
+ "epoch": 7.880349344978166,
+ "grad_norm": 0.014119806699454784,
+ "learning_rate": 0.0006,
+ "loss": 5.483551979064941,
+ "step": 567
+ },
+ {
+ "epoch": 7.894323144104804,
+ "grad_norm": 0.015270394273102283,
+ "learning_rate": 0.0006,
+ "loss": 5.378519535064697,
+ "step": 568
+ },
+ {
+ "epoch": 7.908296943231441,
+ "grad_norm": 0.015677539631724358,
+ "learning_rate": 0.0006,
+ "loss": 5.290210247039795,
+ "step": 569
+ },
+ {
+ "epoch": 7.922270742358078,
+ "grad_norm": 0.015930423513054848,
+ "learning_rate": 0.0006,
+ "loss": 5.427360534667969,
+ "step": 570
+ },
+ {
+ "epoch": 7.936244541484716,
+ "grad_norm": 0.016641564667224884,
+ "learning_rate": 0.0006,
+ "loss": 5.301599979400635,
+ "step": 571
+ },
+ {
+ "epoch": 7.950218340611354,
+ "grad_norm": 0.018293552100658417,
+ "learning_rate": 0.0006,
+ "loss": 5.277889251708984,
+ "step": 572
+ },
+ {
+ "epoch": 7.964192139737992,
+ "grad_norm": 0.017618799582123756,
+ "learning_rate": 0.0006,
+ "loss": 5.297194480895996,
+ "step": 573
+ },
+ {
+ "epoch": 7.978165938864628,
+ "grad_norm": 0.015034242533147335,
+ "learning_rate": 0.0006,
+ "loss": 5.42428731918335,
+ "step": 574
+ },
+ {
+ "epoch": 7.992139737991266,
+ "grad_norm": 0.01617511734366417,
+ "learning_rate": 0.0006,
+ "loss": 5.420044898986816,
+ "step": 575
+ },
+ {
+ "epoch": 8.0,
+ "grad_norm": 0.01837257295846939,
+ "learning_rate": 0.0006,
+ "loss": 5.240387916564941,
+ "step": 576
+ },
+ {
+ "epoch": 8.0,
+ "eval_loss": 5.423932075500488,
+ "eval_runtime": 58.0841,
+ "eval_samples_per_second": 42.042,
+ "eval_steps_per_second": 1.326,
+ "step": 576
+ },
+ {
+ "epoch": 8.013973799126637,
+ "grad_norm": 0.020953809842467308,
+ "learning_rate": 0.0006,
+ "loss": 5.4320783615112305,
+ "step": 577
+ },
+ {
+ "epoch": 8.027947598253276,
+ "grad_norm": 0.0227745920419693,
+ "learning_rate": 0.0006,
+ "loss": 5.310197830200195,
+ "step": 578
+ },
+ {
+ "epoch": 8.041921397379912,
+ "grad_norm": 0.021372603252530098,
+ "learning_rate": 0.0006,
+ "loss": 5.339468002319336,
+ "step": 579
+ },
+ {
+ "epoch": 8.055895196506551,
+ "grad_norm": 0.022401731461286545,
+ "learning_rate": 0.0006,
+ "loss": 5.279829978942871,
+ "step": 580
+ },
+ {
+ "epoch": 8.069868995633188,
+ "grad_norm": 0.023942379280924797,
+ "learning_rate": 0.0006,
+ "loss": 5.372323989868164,
+ "step": 581
+ },
+ {
+ "epoch": 8.083842794759825,
+ "grad_norm": 0.024380534887313843,
+ "learning_rate": 0.0006,
+ "loss": 5.253486633300781,
+ "step": 582
+ },
+ {
+ "epoch": 8.097816593886463,
+ "grad_norm": 0.03246625140309334,
+ "learning_rate": 0.0006,
+ "loss": 5.305774211883545,
+ "step": 583
+ },
+ {
+ "epoch": 8.1117903930131,
+ "grad_norm": 0.030399736016988754,
+ "learning_rate": 0.0006,
+ "loss": 5.370976448059082,
+ "step": 584
+ },
+ {
+ "epoch": 8.125764192139737,
+ "grad_norm": 0.027148913592100143,
+ "learning_rate": 0.0006,
+ "loss": 5.368010520935059,
+ "step": 585
+ },
+ {
+ "epoch": 8.139737991266376,
+ "grad_norm": 0.02925540879368782,
+ "learning_rate": 0.0006,
+ "loss": 5.303013801574707,
+ "step": 586
+ },
+ {
+ "epoch": 8.153711790393013,
+ "grad_norm": 0.02915453538298607,
+ "learning_rate": 0.0006,
+ "loss": 5.329667091369629,
+ "step": 587
+ },
+ {
+ "epoch": 8.167685589519651,
+ "grad_norm": 0.031732626259326935,
+ "learning_rate": 0.0006,
+ "loss": 5.3634934425354,
+ "step": 588
+ },
+ {
+ "epoch": 8.181659388646288,
+ "grad_norm": 0.036166731268167496,
+ "learning_rate": 0.0006,
+ "loss": 5.223980903625488,
+ "step": 589
+ },
+ {
+ "epoch": 8.195633187772925,
+ "grad_norm": 0.035654108971357346,
+ "learning_rate": 0.0006,
+ "loss": 5.361059188842773,
+ "step": 590
+ },
+ {
+ "epoch": 8.209606986899564,
+ "grad_norm": 0.0365324430167675,
+ "learning_rate": 0.0006,
+ "loss": 5.336151599884033,
+ "step": 591
+ },
+ {
+ "epoch": 8.2235807860262,
+ "grad_norm": 0.03578880429267883,
+ "learning_rate": 0.0006,
+ "loss": 5.434076309204102,
+ "step": 592
+ },
+ {
+ "epoch": 8.237554585152838,
+ "grad_norm": 0.035734012722969055,
+ "learning_rate": 0.0006,
+ "loss": 5.330893516540527,
+ "step": 593
+ },
+ {
+ "epoch": 8.251528384279476,
+ "grad_norm": 0.03196857124567032,
+ "learning_rate": 0.0006,
+ "loss": 5.306009769439697,
+ "step": 594
+ },
+ {
+ "epoch": 8.265502183406113,
+ "grad_norm": 0.032042086124420166,
+ "learning_rate": 0.0006,
+ "loss": 5.343267917633057,
+ "step": 595
+ },
+ {
+ "epoch": 8.279475982532752,
+ "grad_norm": 0.03160746395587921,
+ "learning_rate": 0.0006,
+ "loss": 5.2353715896606445,
+ "step": 596
+ },
+ {
+ "epoch": 8.293449781659389,
+ "grad_norm": 0.026689818128943443,
+ "learning_rate": 0.0006,
+ "loss": 5.254042625427246,
+ "step": 597
+ },
+ {
+ "epoch": 8.307423580786025,
+ "grad_norm": 0.02880188450217247,
+ "learning_rate": 0.0006,
+ "loss": 5.326833248138428,
+ "step": 598
+ },
+ {
+ "epoch": 8.321397379912664,
+ "grad_norm": 0.027516381815075874,
+ "learning_rate": 0.0006,
+ "loss": 5.342336177825928,
+ "step": 599
+ },
+ {
+ "epoch": 8.335371179039301,
+ "grad_norm": 0.027875030413269997,
+ "learning_rate": 0.0006,
+ "loss": 5.200719833374023,
+ "step": 600
+ },
+ {
+ "epoch": 8.34934497816594,
+ "grad_norm": 0.0268265288323164,
+ "learning_rate": 0.0006,
+ "loss": 5.249449729919434,
+ "step": 601
+ },
+ {
+ "epoch": 8.363318777292577,
+ "grad_norm": 0.024824608117341995,
+ "learning_rate": 0.0006,
+ "loss": 5.318952560424805,
+ "step": 602
+ },
+ {
+ "epoch": 8.377292576419213,
+ "grad_norm": 0.022990627214312553,
+ "learning_rate": 0.0006,
+ "loss": 5.244993209838867,
+ "step": 603
+ },
+ {
+ "epoch": 8.391266375545852,
+ "grad_norm": 0.022804604843258858,
+ "learning_rate": 0.0006,
+ "loss": 5.3236846923828125,
+ "step": 604
+ },
+ {
+ "epoch": 8.405240174672489,
+ "grad_norm": 0.02419872209429741,
+ "learning_rate": 0.0006,
+ "loss": 5.255486488342285,
+ "step": 605
+ },
+ {
+ "epoch": 8.419213973799126,
+ "grad_norm": 0.021952059119939804,
+ "learning_rate": 0.0006,
+ "loss": 5.3511552810668945,
+ "step": 606
+ },
+ {
+ "epoch": 8.433187772925764,
+ "grad_norm": 0.022375497967004776,
+ "learning_rate": 0.0006,
+ "loss": 5.294790267944336,
+ "step": 607
+ },
+ {
+ "epoch": 8.447161572052401,
+ "grad_norm": 0.019242815673351288,
+ "learning_rate": 0.0006,
+ "loss": 5.252618789672852,
+ "step": 608
+ },
+ {
+ "epoch": 8.46113537117904,
+ "grad_norm": 0.0173486340790987,
+ "learning_rate": 0.0006,
+ "loss": 5.30147647857666,
+ "step": 609
+ },
+ {
+ "epoch": 8.475109170305677,
+ "grad_norm": 0.01612604409456253,
+ "learning_rate": 0.0006,
+ "loss": 5.23960542678833,
+ "step": 610
+ },
+ {
+ "epoch": 8.489082969432314,
+ "grad_norm": 0.014293976128101349,
+ "learning_rate": 0.0006,
+ "loss": 5.119932651519775,
+ "step": 611
+ },
+ {
+ "epoch": 8.503056768558952,
+ "grad_norm": 0.014453536830842495,
+ "learning_rate": 0.0006,
+ "loss": 5.166794776916504,
+ "step": 612
+ },
+ {
+ "epoch": 8.51703056768559,
+ "grad_norm": 0.014685769565403461,
+ "learning_rate": 0.0006,
+ "loss": 5.264822959899902,
+ "step": 613
+ },
+ {
+ "epoch": 8.531004366812226,
+ "grad_norm": 0.01427740603685379,
+ "learning_rate": 0.0006,
+ "loss": 5.278433799743652,
+ "step": 614
+ },
+ {
+ "epoch": 8.544978165938865,
+ "grad_norm": 0.014864951372146606,
+ "learning_rate": 0.0006,
+ "loss": 5.214512348175049,
+ "step": 615
+ },
+ {
+ "epoch": 8.558951965065502,
+ "grad_norm": 0.01565164513885975,
+ "learning_rate": 0.0006,
+ "loss": 5.183066368103027,
+ "step": 616
+ },
+ {
+ "epoch": 8.57292576419214,
+ "grad_norm": 0.01779816672205925,
+ "learning_rate": 0.0006,
+ "loss": 5.163185119628906,
+ "step": 617
+ },
+ {
+ "epoch": 8.586899563318777,
+ "grad_norm": 0.017254827544093132,
+ "learning_rate": 0.0006,
+ "loss": 5.243555068969727,
+ "step": 618
+ },
+ {
+ "epoch": 8.600873362445414,
+ "grad_norm": 0.01650584116578102,
+ "learning_rate": 0.0006,
+ "loss": 5.234884738922119,
+ "step": 619
+ },
+ {
+ "epoch": 8.614847161572053,
+ "grad_norm": 0.017021115869283676,
+ "learning_rate": 0.0006,
+ "loss": 5.12271785736084,
+ "step": 620
+ },
+ {
+ "epoch": 8.62882096069869,
+ "grad_norm": 0.01773759163916111,
+ "learning_rate": 0.0006,
+ "loss": 5.242153167724609,
+ "step": 621
+ },
+ {
+ "epoch": 8.642794759825328,
+ "grad_norm": 0.015679042786359787,
+ "learning_rate": 0.0006,
+ "loss": 5.170779228210449,
+ "step": 622
+ },
+ {
+ "epoch": 8.656768558951965,
+ "grad_norm": 0.013760549947619438,
+ "learning_rate": 0.0006,
+ "loss": 5.238644599914551,
+ "step": 623
+ },
+ {
+ "epoch": 8.670742358078602,
+ "grad_norm": 0.014571522362530231,
+ "learning_rate": 0.0006,
+ "loss": 5.177056312561035,
+ "step": 624
+ },
+ {
+ "epoch": 8.68471615720524,
+ "grad_norm": 0.016209015622735023,
+ "learning_rate": 0.0006,
+ "loss": 5.163750648498535,
+ "step": 625
+ },
+ {
+ "epoch": 8.698689956331878,
+ "grad_norm": 0.016813941299915314,
+ "learning_rate": 0.0006,
+ "loss": 5.183428764343262,
+ "step": 626
+ },
+ {
+ "epoch": 8.712663755458514,
+ "grad_norm": 0.020985357463359833,
+ "learning_rate": 0.0006,
+ "loss": 5.229465007781982,
+ "step": 627
+ },
+ {
+ "epoch": 8.726637554585153,
+ "grad_norm": 0.02679632417857647,
+ "learning_rate": 0.0006,
+ "loss": 5.257368087768555,
+ "step": 628
+ },
+ {
+ "epoch": 8.74061135371179,
+ "grad_norm": 0.025756070390343666,
+ "learning_rate": 0.0006,
+ "loss": 5.253736972808838,
+ "step": 629
+ },
+ {
+ "epoch": 8.754585152838429,
+ "grad_norm": 0.02643490768969059,
+ "learning_rate": 0.0006,
+ "loss": 5.228633403778076,
+ "step": 630
+ },
+ {
+ "epoch": 8.768558951965066,
+ "grad_norm": 0.03303210437297821,
+ "learning_rate": 0.0006,
+ "loss": 5.163540363311768,
+ "step": 631
+ },
+ {
+ "epoch": 8.782532751091702,
+ "grad_norm": 0.030432431027293205,
+ "learning_rate": 0.0006,
+ "loss": 5.243169784545898,
+ "step": 632
+ },
+ {
+ "epoch": 8.796506550218341,
+ "grad_norm": 0.028631288558244705,
+ "learning_rate": 0.0006,
+ "loss": 5.166398048400879,
+ "step": 633
+ },
+ {
+ "epoch": 8.810480349344978,
+ "grad_norm": 0.026188310235738754,
+ "learning_rate": 0.0006,
+ "loss": 5.128122329711914,
+ "step": 634
+ },
+ {
+ "epoch": 8.824454148471617,
+ "grad_norm": 0.025308528915047646,
+ "learning_rate": 0.0006,
+ "loss": 5.211284637451172,
+ "step": 635
+ },
+ {
+ "epoch": 8.838427947598253,
+ "grad_norm": 0.020657729357481003,
+ "learning_rate": 0.0006,
+ "loss": 5.161575794219971,
+ "step": 636
+ },
+ {
+ "epoch": 8.85240174672489,
+ "grad_norm": 0.021901234984397888,
+ "learning_rate": 0.0006,
+ "loss": 5.20050048828125,
+ "step": 637
+ },
+ {
+ "epoch": 8.866375545851529,
+ "grad_norm": 0.021219318732619286,
+ "learning_rate": 0.0006,
+ "loss": 5.211699485778809,
+ "step": 638
+ },
+ {
+ "epoch": 8.880349344978166,
+ "grad_norm": 0.0200974028557539,
+ "learning_rate": 0.0006,
+ "loss": 5.092503547668457,
+ "step": 639
+ },
+ {
+ "epoch": 8.894323144104803,
+ "grad_norm": 0.023804882541298866,
+ "learning_rate": 0.0006,
+ "loss": 5.216068267822266,
+ "step": 640
+ },
+ {
+ "epoch": 8.908296943231441,
+ "grad_norm": 0.026088010519742966,
+ "learning_rate": 0.0006,
+ "loss": 5.155592918395996,
+ "step": 641
+ },
+ {
+ "epoch": 8.922270742358078,
+ "grad_norm": 0.02601276896893978,
+ "learning_rate": 0.0006,
+ "loss": 5.197238922119141,
+ "step": 642
+ },
+ {
+ "epoch": 8.936244541484717,
+ "grad_norm": 0.020387211814522743,
+ "learning_rate": 0.0006,
+ "loss": 5.1955437660217285,
+ "step": 643
+ },
+ {
+ "epoch": 8.950218340611354,
+ "grad_norm": 0.019214622676372528,
+ "learning_rate": 0.0006,
+ "loss": 5.206346035003662,
+ "step": 644
+ },
+ {
+ "epoch": 8.96419213973799,
+ "grad_norm": 0.019674314185976982,
+ "learning_rate": 0.0006,
+ "loss": 5.219857215881348,
+ "step": 645
+ },
+ {
+ "epoch": 8.97816593886463,
+ "grad_norm": 0.020318234339356422,
+ "learning_rate": 0.0006,
+ "loss": 5.191132545471191,
+ "step": 646
+ },
+ {
+ "epoch": 8.992139737991266,
+ "grad_norm": 0.021428676322102547,
+ "learning_rate": 0.0006,
+ "loss": 5.214695930480957,
+ "step": 647
+ },
+ {
+ "epoch": 9.0,
+ "grad_norm": 0.023235054686665535,
+ "learning_rate": 0.0006,
+ "loss": 5.229091644287109,
+ "step": 648
+ },
+ {
+ "epoch": 9.0,
+ "eval_loss": 5.289593696594238,
+ "eval_runtime": 56.9256,
+ "eval_samples_per_second": 42.898,
+ "eval_steps_per_second": 1.353,
+ "step": 648
+ },
+ {
+ "epoch": 9.013973799126637,
+ "grad_norm": 0.02320289984345436,
+ "learning_rate": 0.0006,
+ "loss": 4.983669281005859,
+ "step": 649
+ },
+ {
+ "epoch": 9.027947598253276,
+ "grad_norm": 0.018521282821893692,
+ "learning_rate": 0.0006,
+ "loss": 5.200575828552246,
+ "step": 650
+ },
+ {
+ "epoch": 9.041921397379912,
+ "grad_norm": 0.0180169939994812,
+ "learning_rate": 0.0006,
+ "loss": 5.148033618927002,
+ "step": 651
+ },
+ {
+ "epoch": 9.055895196506551,
+ "grad_norm": 0.019969860091805458,
+ "learning_rate": 0.0006,
+ "loss": 5.099125385284424,
+ "step": 652
+ },
+ {
+ "epoch": 9.069868995633188,
+ "grad_norm": 0.01643305830657482,
+ "learning_rate": 0.0006,
+ "loss": 5.092347145080566,
+ "step": 653
+ },
+ {
+ "epoch": 9.083842794759825,
+ "grad_norm": 0.016022363677620888,
+ "learning_rate": 0.0006,
+ "loss": 5.129279613494873,
+ "step": 654
+ },
+ {
+ "epoch": 9.097816593886463,
+ "grad_norm": 0.01657041162252426,
+ "learning_rate": 0.0006,
+ "loss": 5.159562587738037,
+ "step": 655
+ },
+ {
+ "epoch": 9.1117903930131,
+ "grad_norm": 0.017924228683114052,
+ "learning_rate": 0.0006,
+ "loss": 5.165497779846191,
+ "step": 656
+ },
+ {
+ "epoch": 9.125764192139737,
+ "grad_norm": 0.021592361852526665,
+ "learning_rate": 0.0006,
+ "loss": 5.204775810241699,
+ "step": 657
+ },
+ {
+ "epoch": 9.139737991266376,
+ "grad_norm": 0.026924120262265205,
+ "learning_rate": 0.0006,
+ "loss": 5.128296852111816,
+ "step": 658
+ },
+ {
+ "epoch": 9.153711790393013,
+ "grad_norm": 0.023156899958848953,
+ "learning_rate": 0.0006,
+ "loss": 5.1255340576171875,
+ "step": 659
+ },
+ {
+ "epoch": 9.167685589519651,
+ "grad_norm": 0.020670367404818535,
+ "learning_rate": 0.0006,
+ "loss": 5.128479957580566,
+ "step": 660
+ },
+ {
+ "epoch": 9.181659388646288,
+ "grad_norm": 0.025447173044085503,
+ "learning_rate": 0.0006,
+ "loss": 5.154821872711182,
+ "step": 661
+ },
+ {
+ "epoch": 9.195633187772925,
+ "grad_norm": 0.027940068393945694,
+ "learning_rate": 0.0006,
+ "loss": 5.143466472625732,
+ "step": 662
+ },
+ {
+ "epoch": 9.209606986899564,
+ "grad_norm": 0.026847844943404198,
+ "learning_rate": 0.0006,
+ "loss": 5.045665740966797,
+ "step": 663
+ },
+ {
+ "epoch": 9.2235807860262,
+ "grad_norm": 0.02340601570904255,
+ "learning_rate": 0.0006,
+ "loss": 5.21933126449585,
+ "step": 664
+ },
+ {
+ "epoch": 9.237554585152838,
+ "grad_norm": 0.02340371161699295,
+ "learning_rate": 0.0006,
+ "loss": 4.942949295043945,
+ "step": 665
+ },
+ {
+ "epoch": 9.251528384279476,
+ "grad_norm": 0.02221992425620556,
+ "learning_rate": 0.0006,
+ "loss": 5.197099208831787,
+ "step": 666
+ },
+ {
+ "epoch": 9.265502183406113,
+ "grad_norm": 0.023508677259087563,
+ "learning_rate": 0.0006,
+ "loss": 5.022897720336914,
+ "step": 667
+ },
+ {
+ "epoch": 9.279475982532752,
+ "grad_norm": 0.026120394468307495,
+ "learning_rate": 0.0006,
+ "loss": 5.087725639343262,
+ "step": 668
+ },
+ {
+ "epoch": 9.293449781659389,
+ "grad_norm": 0.027273228392004967,
+ "learning_rate": 0.0006,
+ "loss": 5.090963840484619,
+ "step": 669
+ },
+ {
+ "epoch": 9.307423580786025,
+ "grad_norm": 0.03241586685180664,
+ "learning_rate": 0.0006,
+ "loss": 5.047842979431152,
+ "step": 670
+ },
+ {
+ "epoch": 9.321397379912664,
+ "grad_norm": 0.030245667323470116,
+ "learning_rate": 0.0006,
+ "loss": 5.104803085327148,
+ "step": 671
+ },
+ {
+ "epoch": 9.335371179039301,
+ "grad_norm": 0.027698364108800888,
+ "learning_rate": 0.0006,
+ "loss": 5.0644097328186035,
+ "step": 672
+ },
+ {
+ "epoch": 9.34934497816594,
+ "grad_norm": 0.029692554846405983,
+ "learning_rate": 0.0006,
+ "loss": 5.111942291259766,
+ "step": 673
+ },
+ {
+ "epoch": 9.363318777292577,
+ "grad_norm": 0.03328656405210495,
+ "learning_rate": 0.0006,
+ "loss": 5.1574554443359375,
+ "step": 674
+ },
+ {
+ "epoch": 9.377292576419213,
+ "grad_norm": 0.031938180327415466,
+ "learning_rate": 0.0006,
+ "loss": 5.190371990203857,
+ "step": 675
+ },
+ {
+ "epoch": 9.391266375545852,
+ "grad_norm": 0.024858945980668068,
+ "learning_rate": 0.0006,
+ "loss": 5.121420860290527,
+ "step": 676
+ },
+ {
+ "epoch": 9.405240174672489,
+ "grad_norm": 0.023033203557133675,
+ "learning_rate": 0.0006,
+ "loss": 5.0974345207214355,
+ "step": 677
+ },
+ {
+ "epoch": 9.419213973799126,
+ "grad_norm": 0.022083545103669167,
+ "learning_rate": 0.0006,
+ "loss": 5.046319961547852,
+ "step": 678
+ },
+ {
+ "epoch": 9.433187772925764,
+ "grad_norm": 0.0228437427431345,
+ "learning_rate": 0.0006,
+ "loss": 5.119963645935059,
+ "step": 679
+ },
+ {
+ "epoch": 9.447161572052401,
+ "grad_norm": 0.0242351982742548,
+ "learning_rate": 0.0006,
+ "loss": 5.0444560050964355,
+ "step": 680
+ },
+ {
+ "epoch": 9.46113537117904,
+ "grad_norm": 0.02401185780763626,
+ "learning_rate": 0.0006,
+ "loss": 5.063238620758057,
+ "step": 681
+ },
+ {
+ "epoch": 9.475109170305677,
+ "grad_norm": 0.02560959942638874,
+ "learning_rate": 0.0006,
+ "loss": 5.035224914550781,
+ "step": 682
+ },
+ {
+ "epoch": 9.489082969432314,
+ "grad_norm": 0.024985626339912415,
+ "learning_rate": 0.0006,
+ "loss": 5.049284934997559,
+ "step": 683
+ },
+ {
+ "epoch": 9.503056768558952,
+ "grad_norm": 0.023178115487098694,
+ "learning_rate": 0.0006,
+ "loss": 5.2266621589660645,
+ "step": 684
+ },
+ {
+ "epoch": 9.51703056768559,
+ "grad_norm": 0.0215512253344059,
+ "learning_rate": 0.0006,
+ "loss": 5.1875715255737305,
+ "step": 685
+ },
+ {
+ "epoch": 9.531004366812226,
+ "grad_norm": 0.01662837713956833,
+ "learning_rate": 0.0006,
+ "loss": 5.137012481689453,
+ "step": 686
+ },
+ {
+ "epoch": 9.544978165938865,
+ "grad_norm": 0.01952764391899109,
+ "learning_rate": 0.0006,
+ "loss": 5.072500228881836,
+ "step": 687
+ },
+ {
+ "epoch": 9.558951965065502,
+ "grad_norm": 0.017521077767014503,
+ "learning_rate": 0.0006,
+ "loss": 5.045711994171143,
+ "step": 688
+ },
+ {
+ "epoch": 9.57292576419214,
+ "grad_norm": 0.017355265095829964,
+ "learning_rate": 0.0006,
+ "loss": 5.02881383895874,
+ "step": 689
+ },
+ {
+ "epoch": 9.586899563318777,
+ "grad_norm": 0.015587719157338142,
+ "learning_rate": 0.0006,
+ "loss": 5.132787704467773,
+ "step": 690
+ },
+ {
+ "epoch": 9.600873362445414,
+ "grad_norm": 0.017469845712184906,
+ "learning_rate": 0.0006,
+ "loss": 5.105447292327881,
+ "step": 691
+ },
+ {
+ "epoch": 9.614847161572053,
+ "grad_norm": 0.019630100578069687,
+ "learning_rate": 0.0006,
+ "loss": 5.032464981079102,
+ "step": 692
+ },
+ {
+ "epoch": 9.62882096069869,
+ "grad_norm": 0.01784994639456272,
+ "learning_rate": 0.0006,
+ "loss": 5.081345558166504,
+ "step": 693
+ },
+ {
+ "epoch": 9.642794759825328,
+ "grad_norm": 0.01872754842042923,
+ "learning_rate": 0.0006,
+ "loss": 4.9772114753723145,
+ "step": 694
+ },
+ {
+ "epoch": 9.656768558951965,
+ "grad_norm": 0.01999032311141491,
+ "learning_rate": 0.0006,
+ "loss": 5.089740753173828,
+ "step": 695
+ },
+ {
+ "epoch": 9.670742358078602,
+ "grad_norm": 0.01710767112672329,
+ "learning_rate": 0.0006,
+ "loss": 4.9869890213012695,
+ "step": 696
+ },
+ {
+ "epoch": 9.68471615720524,
+ "grad_norm": 0.015144161880016327,
+ "learning_rate": 0.0006,
+ "loss": 5.022564888000488,
+ "step": 697
+ },
+ {
+ "epoch": 9.698689956331878,
+ "grad_norm": 0.01618543453514576,
+ "learning_rate": 0.0006,
+ "loss": 5.035921573638916,
+ "step": 698
+ },
+ {
+ "epoch": 9.712663755458514,
+ "grad_norm": 0.014653387479484081,
+ "learning_rate": 0.0006,
+ "loss": 4.914456367492676,
+ "step": 699
+ },
+ {
+ "epoch": 9.726637554585153,
+ "grad_norm": 0.013434977270662785,
+ "learning_rate": 0.0006,
+ "loss": 5.063764572143555,
+ "step": 700
+ },
+ {
+ "epoch": 9.74061135371179,
+ "grad_norm": 0.015128864906728268,
+ "learning_rate": 0.0006,
+ "loss": 5.052367687225342,
+ "step": 701
+ },
+ {
+ "epoch": 9.754585152838429,
+ "grad_norm": 0.014210768043994904,
+ "learning_rate": 0.0006,
+ "loss": 5.016489028930664,
+ "step": 702
+ },
+ {
+ "epoch": 9.768558951965066,
+ "grad_norm": 0.015191680751740932,
+ "learning_rate": 0.0006,
+ "loss": 4.990176200866699,
+ "step": 703
+ },
+ {
+ "epoch": 9.782532751091702,
+ "grad_norm": 0.018285127356648445,
+ "learning_rate": 0.0006,
+ "loss": 5.015164375305176,
+ "step": 704
+ },
+ {
+ "epoch": 9.796506550218341,
+ "grad_norm": 0.0204079058021307,
+ "learning_rate": 0.0006,
+ "loss": 5.091142654418945,
+ "step": 705
+ },
+ {
+ "epoch": 9.810480349344978,
+ "grad_norm": 0.01980586163699627,
+ "learning_rate": 0.0006,
+ "loss": 5.03350830078125,
+ "step": 706
+ },
+ {
+ "epoch": 9.824454148471617,
+ "grad_norm": 0.02048729732632637,
+ "learning_rate": 0.0006,
+ "loss": 5.03611421585083,
+ "step": 707
+ },
+ {
+ "epoch": 9.838427947598253,
+ "grad_norm": 0.02058163844048977,
+ "learning_rate": 0.0006,
+ "loss": 4.9613447189331055,
+ "step": 708
+ },
+ {
+ "epoch": 9.85240174672489,
+ "grad_norm": 0.018004924058914185,
+ "learning_rate": 0.0006,
+ "loss": 4.966541767120361,
+ "step": 709
+ },
+ {
+ "epoch": 9.866375545851529,
+ "grad_norm": 0.0198152307420969,
+ "learning_rate": 0.0006,
+ "loss": 5.01750373840332,
+ "step": 710
+ },
+ {
+ "epoch": 9.880349344978166,
+ "grad_norm": 0.018532969057559967,
+ "learning_rate": 0.0006,
+ "loss": 5.097424507141113,
+ "step": 711
+ },
+ {
+ "epoch": 9.894323144104803,
+ "grad_norm": 0.018939971923828125,
+ "learning_rate": 0.0006,
+ "loss": 5.177223205566406,
+ "step": 712
+ },
+ {
+ "epoch": 9.908296943231441,
+ "grad_norm": 0.017969170585274696,
+ "learning_rate": 0.0006,
+ "loss": 5.118013858795166,
+ "step": 713
+ },
+ {
+ "epoch": 9.922270742358078,
+ "grad_norm": 0.018021270632743835,
+ "learning_rate": 0.0006,
+ "loss": 5.034787178039551,
+ "step": 714
+ },
+ {
+ "epoch": 9.936244541484717,
+ "grad_norm": 0.017009710893034935,
+ "learning_rate": 0.0006,
+ "loss": 5.003920555114746,
+ "step": 715
+ },
+ {
+ "epoch": 9.950218340611354,
+ "grad_norm": 0.017722049728035927,
+ "learning_rate": 0.0006,
+ "loss": 5.031939506530762,
+ "step": 716
+ },
+ {
+ "epoch": 9.96419213973799,
+ "grad_norm": 0.017134087160229683,
+ "learning_rate": 0.0006,
+ "loss": 5.093474388122559,
+ "step": 717
+ },
+ {
+ "epoch": 9.97816593886463,
+ "grad_norm": 0.018511613830924034,
+ "learning_rate": 0.0006,
+ "loss": 5.078604698181152,
+ "step": 718
+ },
+ {
+ "epoch": 9.992139737991266,
+ "grad_norm": 0.021206334233283997,
+ "learning_rate": 0.0006,
+ "loss": 5.052572727203369,
+ "step": 719
+ },
+ {
+ "epoch": 10.0,
+ "grad_norm": 0.02239062264561653,
+ "learning_rate": 0.0006,
+ "loss": 5.063547134399414,
+ "step": 720
+ },
+ {
+ "epoch": 10.0,
+ "eval_loss": 5.173044681549072,
+ "eval_runtime": 56.7287,
+ "eval_samples_per_second": 43.047,
+ "eval_steps_per_second": 1.357,
+ "step": 720
+ },
+ {
+ "epoch": 10.013973799126637,
+ "grad_norm": 0.020757969468832016,
+ "learning_rate": 0.0006,
+ "loss": 5.045648574829102,
+ "step": 721
+ },
+ {
+ "epoch": 10.027947598253276,
+ "grad_norm": 0.022697702050209045,
+ "learning_rate": 0.0006,
+ "loss": 4.993833065032959,
+ "step": 722
+ },
+ {
+ "epoch": 10.041921397379912,
+ "grad_norm": 0.02434331737458706,
+ "learning_rate": 0.0006,
+ "loss": 5.073219299316406,
+ "step": 723
+ },
+ {
+ "epoch": 10.055895196506551,
+ "grad_norm": 0.025499660521745682,
+ "learning_rate": 0.0006,
+ "loss": 5.019072532653809,
+ "step": 724
+ },
+ {
+ "epoch": 10.069868995633188,
+ "grad_norm": 0.029281053692102432,
+ "learning_rate": 0.0006,
+ "loss": 4.959851264953613,
+ "step": 725
+ },
+ {
+ "epoch": 10.083842794759825,
+ "grad_norm": 0.033878110349178314,
+ "learning_rate": 0.0006,
+ "loss": 5.132717609405518,
+ "step": 726
+ },
+ {
+ "epoch": 10.097816593886463,
+ "grad_norm": 0.03385764732956886,
+ "learning_rate": 0.0006,
+ "loss": 5.0482892990112305,
+ "step": 727
+ },
+ {
+ "epoch": 10.1117903930131,
+ "grad_norm": 0.03466648980975151,
+ "learning_rate": 0.0006,
+ "loss": 4.946855545043945,
+ "step": 728
+ },
+ {
+ "epoch": 10.125764192139737,
+ "grad_norm": 0.028546305373311043,
+ "learning_rate": 0.0006,
+ "loss": 5.027774333953857,
+ "step": 729
+ },
+ {
+ "epoch": 10.139737991266376,
+ "grad_norm": 0.030575869604945183,
+ "learning_rate": 0.0006,
+ "loss": 4.9768476486206055,
+ "step": 730
+ },
+ {
+ "epoch": 10.153711790393013,
+ "grad_norm": 0.03614303469657898,
+ "learning_rate": 0.0006,
+ "loss": 5.078113555908203,
+ "step": 731
+ },
+ {
+ "epoch": 10.167685589519651,
+ "grad_norm": 0.040991660207509995,
+ "learning_rate": 0.0006,
+ "loss": 5.033185958862305,
+ "step": 732
+ },
+ {
+ "epoch": 10.181659388646288,
+ "grad_norm": 0.04551699757575989,
+ "learning_rate": 0.0006,
+ "loss": 5.0820746421813965,
+ "step": 733
+ },
+ {
+ "epoch": 10.195633187772925,
+ "grad_norm": 0.040193989872932434,
+ "learning_rate": 0.0006,
+ "loss": 5.059700965881348,
+ "step": 734
+ },
+ {
+ "epoch": 10.209606986899564,
+ "grad_norm": 0.035851605236530304,
+ "learning_rate": 0.0006,
+ "loss": 5.142467498779297,
+ "step": 735
+ },
+ {
+ "epoch": 10.2235807860262,
+ "grad_norm": 0.034200169146060944,
+ "learning_rate": 0.0006,
+ "loss": 5.072312355041504,
+ "step": 736
+ },
+ {
+ "epoch": 10.237554585152838,
+ "grad_norm": 0.033621110022068024,
+ "learning_rate": 0.0006,
+ "loss": 5.08037805557251,
+ "step": 737
+ },
+ {
+ "epoch": 10.251528384279476,
+ "grad_norm": 0.037372369319200516,
+ "learning_rate": 0.0006,
+ "loss": 4.998425483703613,
+ "step": 738
+ },
+ {
+ "epoch": 10.265502183406113,
+ "grad_norm": 0.037044707685709,
+ "learning_rate": 0.0006,
+ "loss": 4.923993110656738,
+ "step": 739
+ },
+ {
+ "epoch": 10.279475982532752,
+ "grad_norm": 0.030454808846116066,
+ "learning_rate": 0.0006,
+ "loss": 5.0889997482299805,
+ "step": 740
+ },
+ {
+ "epoch": 10.293449781659389,
+ "grad_norm": 0.030969638377428055,
+ "learning_rate": 0.0006,
+ "loss": 4.984112739562988,
+ "step": 741
+ },
+ {
+ "epoch": 10.307423580786025,
+ "grad_norm": 0.0276983380317688,
+ "learning_rate": 0.0006,
+ "loss": 4.955412864685059,
+ "step": 742
+ },
+ {
+ "epoch": 10.321397379912664,
+ "grad_norm": 0.02716052532196045,
+ "learning_rate": 0.0006,
+ "loss": 5.0098876953125,
+ "step": 743
+ },
+ {
+ "epoch": 10.335371179039301,
+ "grad_norm": 0.02403552085161209,
+ "learning_rate": 0.0006,
+ "loss": 5.165160655975342,
+ "step": 744
+ },
+ {
+ "epoch": 10.34934497816594,
+ "grad_norm": 0.0257862601429224,
+ "learning_rate": 0.0006,
+ "loss": 5.001547813415527,
+ "step": 745
+ },
+ {
+ "epoch": 10.363318777292577,
+ "grad_norm": 0.02300378680229187,
+ "learning_rate": 0.0006,
+ "loss": 5.038888931274414,
+ "step": 746
+ },
+ {
+ "epoch": 10.377292576419213,
+ "grad_norm": 0.019766854122281075,
+ "learning_rate": 0.0006,
+ "loss": 4.924291610717773,
+ "step": 747
+ },
+ {
+ "epoch": 10.391266375545852,
+ "grad_norm": 0.01888395845890045,
+ "learning_rate": 0.0006,
+ "loss": 5.059971332550049,
+ "step": 748
+ },
+ {
+ "epoch": 10.405240174672489,
+ "grad_norm": 0.015577499754726887,
+ "learning_rate": 0.0006,
+ "loss": 4.987686634063721,
+ "step": 749
+ },
+ {
+ "epoch": 10.419213973799126,
+ "grad_norm": 0.016012346372008324,
+ "learning_rate": 0.0006,
+ "loss": 5.015157699584961,
+ "step": 750
+ },
+ {
+ "epoch": 10.433187772925764,
+ "grad_norm": 0.016277998685836792,
+ "learning_rate": 0.0006,
+ "loss": 5.004931449890137,
+ "step": 751
+ },
+ {
+ "epoch": 10.447161572052401,
+ "grad_norm": 0.017504561692476273,
+ "learning_rate": 0.0006,
+ "loss": 4.902827739715576,
+ "step": 752
+ },
+ {
+ "epoch": 10.46113537117904,
+ "grad_norm": 0.01569022797048092,
+ "learning_rate": 0.0006,
+ "loss": 4.954863548278809,
+ "step": 753
+ },
+ {
+ "epoch": 10.475109170305677,
+ "grad_norm": 0.014114422723650932,
+ "learning_rate": 0.0006,
+ "loss": 5.00656795501709,
+ "step": 754
+ },
+ {
+ "epoch": 10.489082969432314,
+ "grad_norm": 0.01573832333087921,
+ "learning_rate": 0.0006,
+ "loss": 4.978353023529053,
+ "step": 755
+ },
+ {
+ "epoch": 10.503056768558952,
+ "grad_norm": 0.016480036079883575,
+ "learning_rate": 0.0006,
+ "loss": 4.953327655792236,
+ "step": 756
+ },
+ {
+ "epoch": 10.51703056768559,
+ "grad_norm": 0.014008025638759136,
+ "learning_rate": 0.0006,
+ "loss": 5.050840377807617,
+ "step": 757
+ },
+ {
+ "epoch": 10.531004366812226,
+ "grad_norm": 0.013147998601198196,
+ "learning_rate": 0.0006,
+ "loss": 4.974964141845703,
+ "step": 758
+ },
+ {
+ "epoch": 10.544978165938865,
+ "grad_norm": 0.013807603158056736,
+ "learning_rate": 0.0006,
+ "loss": 4.927907943725586,
+ "step": 759
+ },
+ {
+ "epoch": 10.558951965065502,
+ "grad_norm": 0.013555224984884262,
+ "learning_rate": 0.0006,
+ "loss": 4.980690956115723,
+ "step": 760
+ },
+ {
+ "epoch": 10.57292576419214,
+ "grad_norm": 0.013045408762991428,
+ "learning_rate": 0.0006,
+ "loss": 4.9436116218566895,
+ "step": 761
+ },
+ {
+ "epoch": 10.586899563318777,
+ "grad_norm": 0.012858862057328224,
+ "learning_rate": 0.0006,
+ "loss": 4.9595770835876465,
+ "step": 762
+ },
+ {
+ "epoch": 10.600873362445414,
+ "grad_norm": 0.01468253880739212,
+ "learning_rate": 0.0006,
+ "loss": 4.90725040435791,
+ "step": 763
+ },
+ {
+ "epoch": 10.614847161572053,
+ "grad_norm": 0.013224679045379162,
+ "learning_rate": 0.0006,
+ "loss": 4.877615928649902,
+ "step": 764
+ },
+ {
+ "epoch": 10.62882096069869,
+ "grad_norm": 0.013555348850786686,
+ "learning_rate": 0.0006,
+ "loss": 4.9456610679626465,
+ "step": 765
+ },
+ {
+ "epoch": 10.642794759825328,
+ "grad_norm": 0.014387160539627075,
+ "learning_rate": 0.0006,
+ "loss": 4.9233717918396,
+ "step": 766
+ },
+ {
+ "epoch": 10.656768558951965,
+ "grad_norm": 0.014354088343679905,
+ "learning_rate": 0.0006,
+ "loss": 4.983511924743652,
+ "step": 767
+ },
+ {
+ "epoch": 10.670742358078602,
+ "grad_norm": 0.01332700252532959,
+ "learning_rate": 0.0006,
+ "loss": 4.934416770935059,
+ "step": 768
+ },
+ {
+ "epoch": 10.68471615720524,
+ "grad_norm": 0.011543313041329384,
+ "learning_rate": 0.0006,
+ "loss": 4.96388053894043,
+ "step": 769
+ },
+ {
+ "epoch": 10.698689956331878,
+ "grad_norm": 0.012330878525972366,
+ "learning_rate": 0.0006,
+ "loss": 4.946396350860596,
+ "step": 770
+ },
+ {
+ "epoch": 10.712663755458514,
+ "grad_norm": 0.01379779726266861,
+ "learning_rate": 0.0006,
+ "loss": 4.981019020080566,
+ "step": 771
+ },
+ {
+ "epoch": 10.726637554585153,
+ "grad_norm": 0.014612431637942791,
+ "learning_rate": 0.0006,
+ "loss": 4.894567966461182,
+ "step": 772
+ },
+ {
+ "epoch": 10.74061135371179,
+ "grad_norm": 0.013062899932265282,
+ "learning_rate": 0.0006,
+ "loss": 5.033038139343262,
+ "step": 773
+ },
+ {
+ "epoch": 10.754585152838429,
+ "grad_norm": 0.013898049481213093,
+ "learning_rate": 0.0006,
+ "loss": 4.823636054992676,
+ "step": 774
+ },
+ {
+ "epoch": 10.768558951965066,
+ "grad_norm": 0.01502019353210926,
+ "learning_rate": 0.0006,
+ "loss": 5.020156383514404,
+ "step": 775
+ },
+ {
+ "epoch": 10.782532751091702,
+ "grad_norm": 0.013810204342007637,
+ "learning_rate": 0.0006,
+ "loss": 4.935647010803223,
+ "step": 776
+ },
+ {
+ "epoch": 10.796506550218341,
+ "grad_norm": 0.013616513460874557,
+ "learning_rate": 0.0006,
+ "loss": 4.901486396789551,
+ "step": 777
+ },
+ {
+ "epoch": 10.810480349344978,
+ "grad_norm": 0.013601024635136127,
+ "learning_rate": 0.0006,
+ "loss": 4.931873321533203,
+ "step": 778
+ },
+ {
+ "epoch": 10.824454148471617,
+ "grad_norm": 0.013668350875377655,
+ "learning_rate": 0.0006,
+ "loss": 4.937256813049316,
+ "step": 779
+ },
+ {
+ "epoch": 10.838427947598253,
+ "grad_norm": 0.01581823080778122,
+ "learning_rate": 0.0006,
+ "loss": 4.9611406326293945,
+ "step": 780
+ },
+ {
+ "epoch": 10.85240174672489,
+ "grad_norm": 0.019478484988212585,
+ "learning_rate": 0.0006,
+ "loss": 4.9212236404418945,
+ "step": 781
+ },
+ {
+ "epoch": 10.866375545851529,
+ "grad_norm": 0.026921333745121956,
+ "learning_rate": 0.0006,
+ "loss": 4.88501501083374,
+ "step": 782
+ },
+ {
+ "epoch": 10.880349344978166,
+ "grad_norm": 0.03155914694070816,
+ "learning_rate": 0.0006,
+ "loss": 4.989967346191406,
+ "step": 783
+ },
+ {
+ "epoch": 10.894323144104803,
+ "grad_norm": 0.029689429327845573,
+ "learning_rate": 0.0006,
+ "loss": 4.936000823974609,
+ "step": 784
+ },
+ {
+ "epoch": 10.908296943231441,
+ "grad_norm": 0.026626063510775566,
+ "learning_rate": 0.0006,
+ "loss": 4.984785079956055,
+ "step": 785
+ },
+ {
+ "epoch": 10.922270742358078,
+ "grad_norm": 0.026168212294578552,
+ "learning_rate": 0.0006,
+ "loss": 4.908282279968262,
+ "step": 786
+ },
+ {
+ "epoch": 10.936244541484717,
+ "grad_norm": 0.02828158251941204,
+ "learning_rate": 0.0006,
+ "loss": 4.926390647888184,
+ "step": 787
+ },
+ {
+ "epoch": 10.950218340611354,
+ "grad_norm": 0.02649078331887722,
+ "learning_rate": 0.0006,
+ "loss": 5.00400972366333,
+ "step": 788
+ },
+ {
+ "epoch": 10.96419213973799,
+ "grad_norm": 0.0256856270134449,
+ "learning_rate": 0.0006,
+ "loss": 5.0317487716674805,
+ "step": 789
+ },
+ {
+ "epoch": 10.97816593886463,
+ "grad_norm": 0.024728883057832718,
+ "learning_rate": 0.0006,
+ "loss": 4.9299635887146,
+ "step": 790
+ },
+ {
+ "epoch": 10.992139737991266,
+ "grad_norm": 0.026241116225719452,
+ "learning_rate": 0.0006,
+ "loss": 4.947319984436035,
+ "step": 791
+ },
+ {
+ "epoch": 11.0,
+ "grad_norm": 0.023526743054389954,
+ "learning_rate": 0.0006,
+ "loss": 4.94429349899292,
+ "step": 792
+ },
+ {
+ "epoch": 11.0,
+ "eval_loss": 5.104345798492432,
+ "eval_runtime": 57.0955,
+ "eval_samples_per_second": 42.77,
+ "eval_steps_per_second": 1.349,
+ "step": 792
+ },
+ {
+ "epoch": 11.013973799126637,
+ "grad_norm": 0.0216788612306118,
+ "learning_rate": 0.0006,
+ "loss": 4.881950855255127,
+ "step": 793
+ },
+ {
+ "epoch": 11.027947598253276,
+ "grad_norm": 0.01830463856458664,
+ "learning_rate": 0.0006,
+ "loss": 4.971016883850098,
+ "step": 794
+ },
+ {
+ "epoch": 11.041921397379912,
+ "grad_norm": 0.02204521745443344,
+ "learning_rate": 0.0006,
+ "loss": 4.894623756408691,
+ "step": 795
+ },
+ {
+ "epoch": 11.055895196506551,
+ "grad_norm": 0.0210255216807127,
+ "learning_rate": 0.0006,
+ "loss": 4.834009170532227,
+ "step": 796
+ },
+ {
+ "epoch": 11.069868995633188,
+ "grad_norm": 0.019797448068857193,
+ "learning_rate": 0.0006,
+ "loss": 5.086714744567871,
+ "step": 797
+ },
+ {
+ "epoch": 11.083842794759825,
+ "grad_norm": 0.02036641724407673,
+ "learning_rate": 0.0006,
+ "loss": 4.920056343078613,
+ "step": 798
+ },
+ {
+ "epoch": 11.097816593886463,
+ "grad_norm": 0.01966066285967827,
+ "learning_rate": 0.0006,
+ "loss": 4.879859447479248,
+ "step": 799
+ },
+ {
+ "epoch": 11.1117903930131,
+ "grad_norm": 0.0202149897813797,
+ "learning_rate": 0.0006,
+ "loss": 4.933640480041504,
+ "step": 800
+ },
+ {
+ "epoch": 11.125764192139737,
+ "grad_norm": 0.01893039606511593,
+ "learning_rate": 0.0006,
+ "loss": 4.901970863342285,
+ "step": 801
+ },
+ {
+ "epoch": 11.139737991266376,
+ "grad_norm": 0.016040155664086342,
+ "learning_rate": 0.0006,
+ "loss": 4.905491828918457,
+ "step": 802
+ },
+ {
+ "epoch": 11.153711790393013,
+ "grad_norm": 0.01720350608229637,
+ "learning_rate": 0.0006,
+ "loss": 4.798816680908203,
+ "step": 803
+ },
+ {
+ "epoch": 11.167685589519651,
+ "grad_norm": 0.017002852633595467,
+ "learning_rate": 0.0006,
+ "loss": 4.851859092712402,
+ "step": 804
+ },
+ {
+ "epoch": 11.181659388646288,
+ "grad_norm": 0.015594680793583393,
+ "learning_rate": 0.0006,
+ "loss": 4.90328311920166,
+ "step": 805
+ },
+ {
+ "epoch": 11.195633187772925,
+ "grad_norm": 0.014904462732374668,
+ "learning_rate": 0.0006,
+ "loss": 4.904215335845947,
+ "step": 806
+ },
+ {
+ "epoch": 11.209606986899564,
+ "grad_norm": 0.01366228237748146,
+ "learning_rate": 0.0006,
+ "loss": 4.9813055992126465,
+ "step": 807
+ },
+ {
+ "epoch": 11.2235807860262,
+ "grad_norm": 0.014719638973474503,
+ "learning_rate": 0.0006,
+ "loss": 4.927306175231934,
+ "step": 808
+ },
+ {
+ "epoch": 11.237554585152838,
+ "grad_norm": 0.015016036108136177,
+ "learning_rate": 0.0006,
+ "loss": 4.874215126037598,
+ "step": 809
+ },
+ {
+ "epoch": 11.251528384279476,
+ "grad_norm": 0.014925424009561539,
+ "learning_rate": 0.0006,
+ "loss": 4.838257789611816,
+ "step": 810
+ },
+ {
+ "epoch": 11.265502183406113,
+ "grad_norm": 0.014335823245346546,
+ "learning_rate": 0.0006,
+ "loss": 4.940608024597168,
+ "step": 811
+ },
+ {
+ "epoch": 11.279475982532752,
+ "grad_norm": 0.014987512491643429,
+ "learning_rate": 0.0006,
+ "loss": 4.831531524658203,
+ "step": 812
+ },
+ {
+ "epoch": 11.293449781659389,
+ "grad_norm": 0.016087457537651062,
+ "learning_rate": 0.0006,
+ "loss": 4.863227367401123,
+ "step": 813
+ },
+ {
+ "epoch": 11.307423580786025,
+ "grad_norm": 0.01671300269663334,
+ "learning_rate": 0.0006,
+ "loss": 4.8535661697387695,
+ "step": 814
+ },
+ {
+ "epoch": 11.321397379912664,
+ "grad_norm": 0.015229142270982265,
+ "learning_rate": 0.0006,
+ "loss": 4.948635101318359,
+ "step": 815
+ },
+ {
+ "epoch": 11.335371179039301,
+ "grad_norm": 0.01532587967813015,
+ "learning_rate": 0.0006,
+ "loss": 4.9035797119140625,
+ "step": 816
+ },
+ {
+ "epoch": 11.34934497816594,
+ "grad_norm": 0.015715476125478745,
+ "learning_rate": 0.0006,
+ "loss": 4.894709587097168,
+ "step": 817
+ },
+ {
+ "epoch": 11.363318777292577,
+ "grad_norm": 0.014156977646052837,
+ "learning_rate": 0.0006,
+ "loss": 4.960206985473633,
+ "step": 818
+ },
+ {
+ "epoch": 11.377292576419213,
+ "grad_norm": 0.017076566815376282,
+ "learning_rate": 0.0006,
+ "loss": 5.005850791931152,
+ "step": 819
+ },
+ {
+ "epoch": 11.391266375545852,
+ "grad_norm": 0.0172084029763937,
+ "learning_rate": 0.0006,
+ "loss": 4.903355598449707,
+ "step": 820
+ },
+ {
+ "epoch": 11.405240174672489,
+ "grad_norm": 0.0180258359760046,
+ "learning_rate": 0.0006,
+ "loss": 4.842667579650879,
+ "step": 821
+ },
+ {
+ "epoch": 11.419213973799126,
+ "grad_norm": 0.022130120545625687,
+ "learning_rate": 0.0006,
+ "loss": 4.799970626831055,
+ "step": 822
+ },
+ {
+ "epoch": 11.433187772925764,
+ "grad_norm": 0.02384897693991661,
+ "learning_rate": 0.0006,
+ "loss": 4.860746383666992,
+ "step": 823
+ },
+ {
+ "epoch": 11.447161572052401,
+ "grad_norm": 0.021881572902202606,
+ "learning_rate": 0.0006,
+ "loss": 4.918107032775879,
+ "step": 824
+ },
+ {
+ "epoch": 11.46113537117904,
+ "grad_norm": 0.023085080087184906,
+ "learning_rate": 0.0006,
+ "loss": 4.878946304321289,
+ "step": 825
+ },
+ {
+ "epoch": 11.475109170305677,
+ "grad_norm": 0.02291012555360794,
+ "learning_rate": 0.0006,
+ "loss": 4.958887100219727,
+ "step": 826
+ },
+ {
+ "epoch": 11.489082969432314,
+ "grad_norm": 0.0199968870729208,
+ "learning_rate": 0.0006,
+ "loss": 4.80955696105957,
+ "step": 827
+ },
+ {
+ "epoch": 11.503056768558952,
+ "grad_norm": 0.018702229484915733,
+ "learning_rate": 0.0006,
+ "loss": 5.0276408195495605,
+ "step": 828
+ },
+ {
+ "epoch": 11.51703056768559,
+ "grad_norm": 0.017787311226129532,
+ "learning_rate": 0.0006,
+ "loss": 4.96018123626709,
+ "step": 829
+ },
+ {
+ "epoch": 11.531004366812226,
+ "grad_norm": 0.01704501546919346,
+ "learning_rate": 0.0006,
+ "loss": 4.914553165435791,
+ "step": 830
+ },
+ {
+ "epoch": 11.544978165938865,
+ "grad_norm": 0.017519677057862282,
+ "learning_rate": 0.0006,
+ "loss": 4.798098087310791,
+ "step": 831
+ },
+ {
+ "epoch": 11.558951965065502,
+ "grad_norm": 0.018511991947889328,
+ "learning_rate": 0.0006,
+ "loss": 4.817785263061523,
+ "step": 832
+ },
+ {
+ "epoch": 11.57292576419214,
+ "grad_norm": 0.01876644790172577,
+ "learning_rate": 0.0006,
+ "loss": 4.818995475769043,
+ "step": 833
+ },
+ {
+ "epoch": 11.586899563318777,
+ "grad_norm": 0.01753568835556507,
+ "learning_rate": 0.0006,
+ "loss": 4.811938285827637,
+ "step": 834
+ },
+ {
+ "epoch": 11.600873362445414,
+ "grad_norm": 0.017046531662344933,
+ "learning_rate": 0.0006,
+ "loss": 4.914674282073975,
+ "step": 835
+ },
+ {
+ "epoch": 11.614847161572053,
+ "grad_norm": 0.01802094094455242,
+ "learning_rate": 0.0006,
+ "loss": 4.993284225463867,
+ "step": 836
+ },
+ {
+ "epoch": 11.62882096069869,
+ "grad_norm": 0.019493764266371727,
+ "learning_rate": 0.0006,
+ "loss": 4.895628452301025,
+ "step": 837
+ },
+ {
+ "epoch": 11.642794759825328,
+ "grad_norm": 0.019461210817098618,
+ "learning_rate": 0.0006,
+ "loss": 4.869269371032715,
+ "step": 838
+ },
+ {
+ "epoch": 11.656768558951965,
+ "grad_norm": 0.01981567218899727,
+ "learning_rate": 0.0006,
+ "loss": 4.863052845001221,
+ "step": 839
+ },
+ {
+ "epoch": 11.670742358078602,
+ "grad_norm": 0.019894491881132126,
+ "learning_rate": 0.0006,
+ "loss": 4.8775787353515625,
+ "step": 840
+ },
+ {
+ "epoch": 11.68471615720524,
+ "grad_norm": 0.01875786855816841,
+ "learning_rate": 0.0006,
+ "loss": 4.817763328552246,
+ "step": 841
+ },
+ {
+ "epoch": 11.698689956331878,
+ "grad_norm": 0.020926695317029953,
+ "learning_rate": 0.0006,
+ "loss": 4.7470526695251465,
+ "step": 842
+ },
+ {
+ "epoch": 11.712663755458514,
+ "grad_norm": 0.021477103233337402,
+ "learning_rate": 0.0006,
+ "loss": 4.762682914733887,
+ "step": 843
+ },
+ {
+ "epoch": 11.726637554585153,
+ "grad_norm": 0.020628711208701134,
+ "learning_rate": 0.0006,
+ "loss": 4.807985782623291,
+ "step": 844
+ },
+ {
+ "epoch": 11.74061135371179,
+ "grad_norm": 0.021489612758159637,
+ "learning_rate": 0.0006,
+ "loss": 4.843091011047363,
+ "step": 845
+ },
+ {
+ "epoch": 11.754585152838429,
+ "grad_norm": 0.021194491535425186,
+ "learning_rate": 0.0006,
+ "loss": 4.8300251960754395,
+ "step": 846
+ },
+ {
+ "epoch": 11.768558951965066,
+ "grad_norm": 0.01833350397646427,
+ "learning_rate": 0.0006,
+ "loss": 5.000202178955078,
+ "step": 847
+ },
+ {
+ "epoch": 11.782532751091702,
+ "grad_norm": 0.018096931278705597,
+ "learning_rate": 0.0006,
+ "loss": 4.880648136138916,
+ "step": 848
+ },
+ {
+ "epoch": 11.796506550218341,
+ "grad_norm": 0.0197161752730608,
+ "learning_rate": 0.0006,
+ "loss": 4.788316249847412,
+ "step": 849
+ },
+ {
+ "epoch": 11.810480349344978,
+ "grad_norm": 0.0203181654214859,
+ "learning_rate": 0.0006,
+ "loss": 4.870944976806641,
+ "step": 850
+ },
+ {
+ "epoch": 11.824454148471617,
+ "grad_norm": 0.01908455230295658,
+ "learning_rate": 0.0006,
+ "loss": 4.890144348144531,
+ "step": 851
+ },
+ {
+ "epoch": 11.838427947598253,
+ "grad_norm": 0.021281801164150238,
+ "learning_rate": 0.0006,
+ "loss": 4.89578914642334,
+ "step": 852
+ },
+ {
+ "epoch": 11.85240174672489,
+ "grad_norm": 0.02124273031949997,
+ "learning_rate": 0.0006,
+ "loss": 4.883749008178711,
+ "step": 853
+ },
+ {
+ "epoch": 11.866375545851529,
+ "grad_norm": 0.020969398319721222,
+ "learning_rate": 0.0006,
+ "loss": 4.809142589569092,
+ "step": 854
+ },
+ {
+ "epoch": 11.880349344978166,
+ "grad_norm": 0.01967657171189785,
+ "learning_rate": 0.0006,
+ "loss": 4.821864604949951,
+ "step": 855
+ },
+ {
+ "epoch": 11.894323144104803,
+ "grad_norm": 0.020887333899736404,
+ "learning_rate": 0.0006,
+ "loss": 4.917985916137695,
+ "step": 856
+ },
+ {
+ "epoch": 11.908296943231441,
+ "grad_norm": 0.019203083589673042,
+ "learning_rate": 0.0006,
+ "loss": 4.824960708618164,
+ "step": 857
+ },
+ {
+ "epoch": 11.922270742358078,
+ "grad_norm": 0.020162401720881462,
+ "learning_rate": 0.0006,
+ "loss": 4.772617816925049,
+ "step": 858
+ },
+ {
+ "epoch": 11.936244541484717,
+ "grad_norm": 0.02251487225294113,
+ "learning_rate": 0.0006,
+ "loss": 4.735418319702148,
+ "step": 859
+ },
+ {
+ "epoch": 11.950218340611354,
+ "grad_norm": 0.020313527435064316,
+ "learning_rate": 0.0006,
+ "loss": 4.939233779907227,
+ "step": 860
+ },
+ {
+ "epoch": 11.96419213973799,
+ "grad_norm": 0.021635599434375763,
+ "learning_rate": 0.0006,
+ "loss": 4.861807346343994,
+ "step": 861
+ },
+ {
+ "epoch": 11.97816593886463,
+ "grad_norm": 0.02285311557352543,
+ "learning_rate": 0.0006,
+ "loss": 4.932732582092285,
+ "step": 862
+ },
+ {
+ "epoch": 11.992139737991266,
+ "grad_norm": 0.02072535641491413,
+ "learning_rate": 0.0006,
+ "loss": 4.796767711639404,
+ "step": 863
+ },
+ {
+ "epoch": 12.0,
+ "grad_norm": 0.020588304847478867,
+ "learning_rate": 0.0006,
+ "loss": 4.997537612915039,
+ "step": 864
+ },
+ {
+ "epoch": 12.0,
+ "eval_loss": 5.044651985168457,
+ "eval_runtime": 56.9145,
+ "eval_samples_per_second": 42.906,
+ "eval_steps_per_second": 1.353,
+ "step": 864
+ },
+ {
+ "epoch": 12.013973799126637,
+ "grad_norm": 0.017597166821360588,
+ "learning_rate": 0.0006,
+ "loss": 4.830524444580078,
+ "step": 865
+ },
+ {
+ "epoch": 12.027947598253276,
+ "grad_norm": 0.01716788113117218,
+ "learning_rate": 0.0006,
+ "loss": 4.894325256347656,
+ "step": 866
+ },
+ {
+ "epoch": 12.041921397379912,
+ "grad_norm": 0.017712457105517387,
+ "learning_rate": 0.0006,
+ "loss": 4.868297576904297,
+ "step": 867
+ },
+ {
+ "epoch": 12.055895196506551,
+ "grad_norm": 0.021486390382051468,
+ "learning_rate": 0.0006,
+ "loss": 4.922049522399902,
+ "step": 868
+ },
+ {
+ "epoch": 12.069868995633188,
+ "grad_norm": 0.024560654535889626,
+ "learning_rate": 0.0006,
+ "loss": 4.734241008758545,
+ "step": 869
+ },
+ {
+ "epoch": 12.083842794759825,
+ "grad_norm": 0.025870252400636673,
+ "learning_rate": 0.0006,
+ "loss": 4.868566989898682,
+ "step": 870
+ },
+ {
+ "epoch": 12.097816593886463,
+ "grad_norm": 0.0277389045804739,
+ "learning_rate": 0.0006,
+ "loss": 4.755006790161133,
+ "step": 871
+ },
+ {
+ "epoch": 12.1117903930131,
+ "grad_norm": 0.029100844636559486,
+ "learning_rate": 0.0006,
+ "loss": 4.909667015075684,
+ "step": 872
+ },
+ {
+ "epoch": 12.125764192139737,
+ "grad_norm": 0.031833067536354065,
+ "learning_rate": 0.0006,
+ "loss": 4.858674049377441,
+ "step": 873
+ },
+ {
+ "epoch": 12.139737991266376,
+ "grad_norm": 0.032709237188100815,
+ "learning_rate": 0.0006,
+ "loss": 4.823696136474609,
+ "step": 874
+ },
+ {
+ "epoch": 12.153711790393013,
+ "grad_norm": 0.02995380386710167,
+ "learning_rate": 0.0006,
+ "loss": 4.850235462188721,
+ "step": 875
+ },
+ {
+ "epoch": 12.167685589519651,
+ "grad_norm": 0.034098610281944275,
+ "learning_rate": 0.0006,
+ "loss": 4.8169989585876465,
+ "step": 876
+ },
+ {
+ "epoch": 12.181659388646288,
+ "grad_norm": 0.03377080336213112,
+ "learning_rate": 0.0006,
+ "loss": 4.90738582611084,
+ "step": 877
+ },
+ {
+ "epoch": 12.195633187772925,
+ "grad_norm": 0.029780825600028038,
+ "learning_rate": 0.0006,
+ "loss": 4.811163902282715,
+ "step": 878
+ },
+ {
+ "epoch": 12.209606986899564,
+ "grad_norm": 0.03118024580180645,
+ "learning_rate": 0.0006,
+ "loss": 4.861286640167236,
+ "step": 879
+ },
+ {
+ "epoch": 12.2235807860262,
+ "grad_norm": 0.033052004873752594,
+ "learning_rate": 0.0006,
+ "loss": 4.865021228790283,
+ "step": 880
+ },
+ {
+ "epoch": 12.237554585152838,
+ "grad_norm": 0.029070032760500908,
+ "learning_rate": 0.0006,
+ "loss": 4.947303771972656,
+ "step": 881
+ },
+ {
+ "epoch": 12.251528384279476,
+ "grad_norm": 0.028355715796351433,
+ "learning_rate": 0.0006,
+ "loss": 4.804648399353027,
+ "step": 882
+ },
+ {
+ "epoch": 12.265502183406113,
+ "grad_norm": 0.031257227063179016,
+ "learning_rate": 0.0006,
+ "loss": 4.757718086242676,
+ "step": 883
+ },
+ {
+ "epoch": 12.279475982532752,
+ "grad_norm": 0.031438373029232025,
+ "learning_rate": 0.0006,
+ "loss": 4.851268768310547,
+ "step": 884
+ },
+ {
+ "epoch": 12.293449781659389,
+ "grad_norm": 0.030492138117551804,
+ "learning_rate": 0.0006,
+ "loss": 4.7891035079956055,
+ "step": 885
+ },
+ {
+ "epoch": 12.307423580786025,
+ "grad_norm": 0.026115991175174713,
+ "learning_rate": 0.0006,
+ "loss": 4.86474609375,
+ "step": 886
+ },
+ {
+ "epoch": 12.321397379912664,
+ "grad_norm": 0.024184465408325195,
+ "learning_rate": 0.0006,
+ "loss": 4.860456943511963,
+ "step": 887
+ },
+ {
+ "epoch": 12.335371179039301,
+ "grad_norm": 0.025437982752919197,
+ "learning_rate": 0.0006,
+ "loss": 4.806251049041748,
+ "step": 888
+ },
+ {
+ "epoch": 12.34934497816594,
+ "grad_norm": 0.022390231490135193,
+ "learning_rate": 0.0006,
+ "loss": 4.807086944580078,
+ "step": 889
+ },
+ {
+ "epoch": 12.363318777292577,
+ "grad_norm": 0.020174294710159302,
+ "learning_rate": 0.0006,
+ "loss": 4.78292179107666,
+ "step": 890
+ },
+ {
+ "epoch": 12.377292576419213,
+ "grad_norm": 0.019734324887394905,
+ "learning_rate": 0.0006,
+ "loss": 4.784509658813477,
+ "step": 891
+ },
+ {
+ "epoch": 12.391266375545852,
+ "grad_norm": 0.016733523458242416,
+ "learning_rate": 0.0006,
+ "loss": 4.811755180358887,
+ "step": 892
+ },
+ {
+ "epoch": 12.405240174672489,
+ "grad_norm": 0.01570146717131138,
+ "learning_rate": 0.0006,
+ "loss": 4.720088005065918,
+ "step": 893
+ },
+ {
+ "epoch": 12.419213973799126,
+ "grad_norm": 0.01601393334567547,
+ "learning_rate": 0.0006,
+ "loss": 4.783543109893799,
+ "step": 894
+ },
+ {
+ "epoch": 12.433187772925764,
+ "grad_norm": 0.014035487547516823,
+ "learning_rate": 0.0006,
+ "loss": 4.848340034484863,
+ "step": 895
+ },
+ {
+ "epoch": 12.447161572052401,
+ "grad_norm": 0.015248487703502178,
+ "learning_rate": 0.0006,
+ "loss": 4.773287773132324,
+ "step": 896
+ },
+ {
+ "epoch": 12.46113537117904,
+ "grad_norm": 0.01532268151640892,
+ "learning_rate": 0.0006,
+ "loss": 4.799646854400635,
+ "step": 897
+ },
+ {
+ "epoch": 12.475109170305677,
+ "grad_norm": 0.013982534408569336,
+ "learning_rate": 0.0006,
+ "loss": 4.8090500831604,
+ "step": 898
+ },
+ {
+ "epoch": 12.489082969432314,
+ "grad_norm": 0.014081164263188839,
+ "learning_rate": 0.0006,
+ "loss": 4.751888751983643,
+ "step": 899
+ },
+ {
+ "epoch": 12.503056768558952,
+ "grad_norm": 0.01380133256316185,
+ "learning_rate": 0.0006,
+ "loss": 4.78144645690918,
+ "step": 900
+ },
+ {
+ "epoch": 12.51703056768559,
+ "grad_norm": 0.013555269688367844,
+ "learning_rate": 0.0006,
+ "loss": 4.779257774353027,
+ "step": 901
+ },
+ {
+ "epoch": 12.531004366812226,
+ "grad_norm": 0.013075701892375946,
+ "learning_rate": 0.0006,
+ "loss": 4.820705413818359,
+ "step": 902
+ },
+ {
+ "epoch": 12.544978165938865,
+ "grad_norm": 0.01276316400617361,
+ "learning_rate": 0.0006,
+ "loss": 4.836733818054199,
+ "step": 903
+ },
+ {
+ "epoch": 12.558951965065502,
+ "grad_norm": 0.013033892959356308,
+ "learning_rate": 0.0006,
+ "loss": 4.764178276062012,
+ "step": 904
+ },
+ {
+ "epoch": 12.57292576419214,
+ "grad_norm": 0.013996980153024197,
+ "learning_rate": 0.0006,
+ "loss": 4.767983436584473,
+ "step": 905
+ },
+ {
+ "epoch": 12.586899563318777,
+ "grad_norm": 0.015222184360027313,
+ "learning_rate": 0.0006,
+ "loss": 4.759721755981445,
+ "step": 906
+ },
+ {
+ "epoch": 12.600873362445414,
+ "grad_norm": 0.01623941957950592,
+ "learning_rate": 0.0006,
+ "loss": 4.849811553955078,
+ "step": 907
+ },
+ {
+ "epoch": 12.614847161572053,
+ "grad_norm": 0.015711182728409767,
+ "learning_rate": 0.0006,
+ "loss": 4.7829084396362305,
+ "step": 908
+ },
+ {
+ "epoch": 12.62882096069869,
+ "grad_norm": 0.014137467369437218,
+ "learning_rate": 0.0006,
+ "loss": 4.73321533203125,
+ "step": 909
+ },
+ {
+ "epoch": 12.642794759825328,
+ "grad_norm": 0.012738462537527084,
+ "learning_rate": 0.0006,
+ "loss": 4.870052337646484,
+ "step": 910
+ },
+ {
+ "epoch": 12.656768558951965,
+ "grad_norm": 0.013683130033314228,
+ "learning_rate": 0.0006,
+ "loss": 4.82658052444458,
+ "step": 911
+ },
+ {
+ "epoch": 12.670742358078602,
+ "grad_norm": 0.014409830793738365,
+ "learning_rate": 0.0006,
+ "loss": 4.707494735717773,
+ "step": 912
+ },
+ {
+ "epoch": 12.68471615720524,
+ "grad_norm": 0.013806809671223164,
+ "learning_rate": 0.0006,
+ "loss": 4.778928279876709,
+ "step": 913
+ },
+ {
+ "epoch": 12.698689956331878,
+ "grad_norm": 0.012857912108302116,
+ "learning_rate": 0.0006,
+ "loss": 4.890717506408691,
+ "step": 914
+ },
+ {
+ "epoch": 12.712663755458514,
+ "grad_norm": 0.013828758150339127,
+ "learning_rate": 0.0006,
+ "loss": 4.8357439041137695,
+ "step": 915
+ },
+ {
+ "epoch": 12.726637554585153,
+ "grad_norm": 0.01478351280093193,
+ "learning_rate": 0.0006,
+ "loss": 4.869986534118652,
+ "step": 916
+ },
+ {
+ "epoch": 12.74061135371179,
+ "grad_norm": 0.01326421182602644,
+ "learning_rate": 0.0006,
+ "loss": 4.830646514892578,
+ "step": 917
+ },
+ {
+ "epoch": 12.754585152838429,
+ "grad_norm": 0.013335433788597584,
+ "learning_rate": 0.0006,
+ "loss": 4.7786149978637695,
+ "step": 918
+ },
+ {
+ "epoch": 12.768558951965066,
+ "grad_norm": 0.014261147007346153,
+ "learning_rate": 0.0006,
+ "loss": 4.788693428039551,
+ "step": 919
+ },
+ {
+ "epoch": 12.782532751091702,
+ "grad_norm": 0.014274783432483673,
+ "learning_rate": 0.0006,
+ "loss": 4.77295446395874,
+ "step": 920
+ },
+ {
+ "epoch": 12.796506550218341,
+ "grad_norm": 0.015074674971401691,
+ "learning_rate": 0.0006,
+ "loss": 4.830754280090332,
+ "step": 921
+ },
+ {
+ "epoch": 12.810480349344978,
+ "grad_norm": 0.015599515289068222,
+ "learning_rate": 0.0006,
+ "loss": 4.837759017944336,
+ "step": 922
+ },
+ {
+ "epoch": 12.824454148471617,
+ "grad_norm": 0.014595242217183113,
+ "learning_rate": 0.0006,
+ "loss": 4.882965564727783,
+ "step": 923
+ },
+ {
+ "epoch": 12.838427947598253,
+ "grad_norm": 0.014951467514038086,
+ "learning_rate": 0.0006,
+ "loss": 4.709748268127441,
+ "step": 924
+ },
+ {
+ "epoch": 12.85240174672489,
+ "grad_norm": 0.01539295818656683,
+ "learning_rate": 0.0006,
+ "loss": 4.791160583496094,
+ "step": 925
+ },
+ {
+ "epoch": 12.866375545851529,
+ "grad_norm": 0.015987906605005264,
+ "learning_rate": 0.0006,
+ "loss": 4.661876678466797,
+ "step": 926
+ },
+ {
+ "epoch": 12.880349344978166,
+ "grad_norm": 0.016527341678738594,
+ "learning_rate": 0.0006,
+ "loss": 4.721952438354492,
+ "step": 927
+ },
+ {
+ "epoch": 12.894323144104803,
+ "grad_norm": 0.01637161523103714,
+ "learning_rate": 0.0006,
+ "loss": 4.704446792602539,
+ "step": 928
+ },
+ {
+ "epoch": 12.908296943231441,
+ "grad_norm": 0.015221050009131432,
+ "learning_rate": 0.0006,
+ "loss": 4.771907806396484,
+ "step": 929
+ },
+ {
+ "epoch": 12.922270742358078,
+ "grad_norm": 0.016242166981101036,
+ "learning_rate": 0.0006,
+ "loss": 4.806267738342285,
+ "step": 930
+ },
+ {
+ "epoch": 12.936244541484717,
+ "grad_norm": 0.01628357358276844,
+ "learning_rate": 0.0006,
+ "loss": 4.7252326011657715,
+ "step": 931
+ },
+ {
+ "epoch": 12.950218340611354,
+ "grad_norm": 0.017428133636713028,
+ "learning_rate": 0.0006,
+ "loss": 4.765689849853516,
+ "step": 932
+ },
+ {
+ "epoch": 12.96419213973799,
+ "grad_norm": 0.01870061084628105,
+ "learning_rate": 0.0006,
+ "loss": 4.814488887786865,
+ "step": 933
+ },
+ {
+ "epoch": 12.97816593886463,
+ "grad_norm": 0.019796060398221016,
+ "learning_rate": 0.0006,
+ "loss": 4.739185810089111,
+ "step": 934
+ },
+ {
+ "epoch": 12.992139737991266,
+ "grad_norm": 0.01985820196568966,
+ "learning_rate": 0.0006,
+ "loss": 4.810608863830566,
+ "step": 935
+ },
+ {
+ "epoch": 13.0,
+ "grad_norm": 0.020834477618336678,
+ "learning_rate": 0.0006,
+ "loss": 4.727858066558838,
+ "step": 936
+ },
+ {
+ "epoch": 13.0,
+ "eval_loss": 4.979116916656494,
+ "eval_runtime": 57.2877,
+ "eval_samples_per_second": 42.627,
+ "eval_steps_per_second": 1.344,
+ "step": 936
+ },
+ {
+ "epoch": 13.013973799126637,
+ "grad_norm": 0.02025519125163555,
+ "learning_rate": 0.0006,
+ "loss": 4.628875732421875,
+ "step": 937
+ },
+ {
+ "epoch": 13.027947598253276,
+ "grad_norm": 0.019596368074417114,
+ "learning_rate": 0.0006,
+ "loss": 4.694699287414551,
+ "step": 938
+ },
+ {
+ "epoch": 13.041921397379912,
+ "grad_norm": 0.020494934171438217,
+ "learning_rate": 0.0006,
+ "loss": 4.691821098327637,
+ "step": 939
+ },
+ {
+ "epoch": 13.055895196506551,
+ "grad_norm": 0.020121760666370392,
+ "learning_rate": 0.0006,
+ "loss": 4.71478271484375,
+ "step": 940
+ },
+ {
+ "epoch": 13.069868995633188,
+ "grad_norm": 0.021091405302286148,
+ "learning_rate": 0.0006,
+ "loss": 4.695915222167969,
+ "step": 941
+ },
+ {
+ "epoch": 13.083842794759825,
+ "grad_norm": 0.019787423312664032,
+ "learning_rate": 0.0006,
+ "loss": 4.8652448654174805,
+ "step": 942
+ },
+ {
+ "epoch": 13.097816593886463,
+ "grad_norm": 0.0185833927243948,
+ "learning_rate": 0.0006,
+ "loss": 4.7065887451171875,
+ "step": 943
+ },
+ {
+ "epoch": 13.1117903930131,
+ "grad_norm": 0.021220475435256958,
+ "learning_rate": 0.0006,
+ "loss": 4.741232395172119,
+ "step": 944
+ },
+ {
+ "epoch": 13.125764192139737,
+ "grad_norm": 0.02136622555553913,
+ "learning_rate": 0.0006,
+ "loss": 4.705967903137207,
+ "step": 945
+ },
+ {
+ "epoch": 13.139737991266376,
+ "grad_norm": 0.0191587433218956,
+ "learning_rate": 0.0006,
+ "loss": 4.758667945861816,
+ "step": 946
+ },
+ {
+ "epoch": 13.153711790393013,
+ "grad_norm": 0.018290022388100624,
+ "learning_rate": 0.0006,
+ "loss": 4.760666847229004,
+ "step": 947
+ },
+ {
+ "epoch": 13.167685589519651,
+ "grad_norm": 0.021704409271478653,
+ "learning_rate": 0.0006,
+ "loss": 4.750555992126465,
+ "step": 948
+ },
+ {
+ "epoch": 13.181659388646288,
+ "grad_norm": 0.021668918430805206,
+ "learning_rate": 0.0006,
+ "loss": 4.840903282165527,
+ "step": 949
+ },
+ {
+ "epoch": 13.195633187772925,
+ "grad_norm": 0.019497141242027283,
+ "learning_rate": 0.0006,
+ "loss": 4.738969326019287,
+ "step": 950
+ },
+ {
+ "epoch": 13.209606986899564,
+ "grad_norm": 0.018500596284866333,
+ "learning_rate": 0.0006,
+ "loss": 4.75154972076416,
+ "step": 951
+ },
+ {
+ "epoch": 13.2235807860262,
+ "grad_norm": 0.01872239261865616,
+ "learning_rate": 0.0006,
+ "loss": 4.811738014221191,
+ "step": 952
+ },
+ {
+ "epoch": 13.237554585152838,
+ "grad_norm": 0.020321547985076904,
+ "learning_rate": 0.0006,
+ "loss": 4.815384387969971,
+ "step": 953
+ },
+ {
+ "epoch": 13.251528384279476,
+ "grad_norm": 0.02283575013279915,
+ "learning_rate": 0.0006,
+ "loss": 4.74582052230835,
+ "step": 954
+ },
+ {
+ "epoch": 13.265502183406113,
+ "grad_norm": 0.02474398910999298,
+ "learning_rate": 0.0006,
+ "loss": 4.738375186920166,
+ "step": 955
+ },
+ {
+ "epoch": 13.279475982532752,
+ "grad_norm": 0.02065764181315899,
+ "learning_rate": 0.0006,
+ "loss": 4.746875286102295,
+ "step": 956
+ },
+ {
+ "epoch": 13.293449781659389,
+ "grad_norm": 0.020534520968794823,
+ "learning_rate": 0.0006,
+ "loss": 4.656914710998535,
+ "step": 957
+ },
+ {
+ "epoch": 13.307423580786025,
+ "grad_norm": 0.0206378772854805,
+ "learning_rate": 0.0006,
+ "loss": 4.842615604400635,
+ "step": 958
+ },
+ {
+ "epoch": 13.321397379912664,
+ "grad_norm": 0.017935654148459435,
+ "learning_rate": 0.0006,
+ "loss": 4.717827796936035,
+ "step": 959
+ },
+ {
+ "epoch": 13.335371179039301,
+ "grad_norm": 0.016958363354206085,
+ "learning_rate": 0.0006,
+ "loss": 4.681070327758789,
+ "step": 960
+ },
+ {
+ "epoch": 13.34934497816594,
+ "grad_norm": 0.015735412016510963,
+ "learning_rate": 0.0006,
+ "loss": 4.7957048416137695,
+ "step": 961
+ },
+ {
+ "epoch": 13.363318777292577,
+ "grad_norm": 0.01563699170947075,
+ "learning_rate": 0.0006,
+ "loss": 4.701192378997803,
+ "step": 962
+ },
+ {
+ "epoch": 13.377292576419213,
+ "grad_norm": 0.015207760035991669,
+ "learning_rate": 0.0006,
+ "loss": 4.73256254196167,
+ "step": 963
+ },
+ {
+ "epoch": 13.391266375545852,
+ "grad_norm": 0.015396883711218834,
+ "learning_rate": 0.0006,
+ "loss": 4.76412296295166,
+ "step": 964
+ },
+ {
+ "epoch": 13.405240174672489,
+ "grad_norm": 0.015690583735704422,
+ "learning_rate": 0.0006,
+ "loss": 4.791157245635986,
+ "step": 965
+ },
+ {
+ "epoch": 13.419213973799126,
+ "grad_norm": 0.01755724661052227,
+ "learning_rate": 0.0006,
+ "loss": 4.845666885375977,
+ "step": 966
+ },
+ {
+ "epoch": 13.433187772925764,
+ "grad_norm": 0.019443422555923462,
+ "learning_rate": 0.0006,
+ "loss": 4.5902299880981445,
+ "step": 967
+ },
+ {
+ "epoch": 13.447161572052401,
+ "grad_norm": 0.02118338644504547,
+ "learning_rate": 0.0006,
+ "loss": 4.740627288818359,
+ "step": 968
+ },
+ {
+ "epoch": 13.46113537117904,
+ "grad_norm": 0.02174839936196804,
+ "learning_rate": 0.0006,
+ "loss": 4.814560890197754,
+ "step": 969
+ },
+ {
+ "epoch": 13.475109170305677,
+ "grad_norm": 0.022860374301671982,
+ "learning_rate": 0.0006,
+ "loss": 4.66440486907959,
+ "step": 970
+ },
+ {
+ "epoch": 13.489082969432314,
+ "grad_norm": 0.024742677807807922,
+ "learning_rate": 0.0006,
+ "loss": 4.7451653480529785,
+ "step": 971
+ },
+ {
+ "epoch": 13.503056768558952,
+ "grad_norm": 0.027068907395005226,
+ "learning_rate": 0.0006,
+ "loss": 4.802584171295166,
+ "step": 972
+ },
+ {
+ "epoch": 13.51703056768559,
+ "grad_norm": 0.026316625997424126,
+ "learning_rate": 0.0006,
+ "loss": 4.842397689819336,
+ "step": 973
+ },
+ {
+ "epoch": 13.531004366812226,
+ "grad_norm": 0.024610962718725204,
+ "learning_rate": 0.0006,
+ "loss": 4.650721073150635,
+ "step": 974
+ },
+ {
+ "epoch": 13.544978165938865,
+ "grad_norm": 0.024122603237628937,
+ "learning_rate": 0.0006,
+ "loss": 4.6984357833862305,
+ "step": 975
+ },
+ {
+ "epoch": 13.558951965065502,
+ "grad_norm": 0.022699708119034767,
+ "learning_rate": 0.0006,
+ "loss": 4.793007850646973,
+ "step": 976
+ },
+ {
+ "epoch": 13.57292576419214,
+ "grad_norm": 0.021706534549593925,
+ "learning_rate": 0.0006,
+ "loss": 4.691362380981445,
+ "step": 977
+ },
+ {
+ "epoch": 13.586899563318777,
+ "grad_norm": 0.021363291889429092,
+ "learning_rate": 0.0006,
+ "loss": 4.76762580871582,
+ "step": 978
+ },
+ {
+ "epoch": 13.600873362445414,
+ "grad_norm": 0.023476677015423775,
+ "learning_rate": 0.0006,
+ "loss": 4.777202606201172,
+ "step": 979
+ },
+ {
+ "epoch": 13.614847161572053,
+ "grad_norm": 0.02457100711762905,
+ "learning_rate": 0.0006,
+ "loss": 4.590703010559082,
+ "step": 980
+ },
+ {
+ "epoch": 13.62882096069869,
+ "grad_norm": 0.02469099499285221,
+ "learning_rate": 0.0006,
+ "loss": 4.640185356140137,
+ "step": 981
+ },
+ {
+ "epoch": 13.642794759825328,
+ "grad_norm": 0.024914603680372238,
+ "learning_rate": 0.0006,
+ "loss": 4.772201061248779,
+ "step": 982
+ },
+ {
+ "epoch": 13.656768558951965,
+ "grad_norm": 0.025133173912763596,
+ "learning_rate": 0.0006,
+ "loss": 4.743337154388428,
+ "step": 983
+ },
+ {
+ "epoch": 13.670742358078602,
+ "grad_norm": 0.02034137211740017,
+ "learning_rate": 0.0006,
+ "loss": 4.752127647399902,
+ "step": 984
+ },
+ {
+ "epoch": 13.68471615720524,
+ "grad_norm": 0.019645661115646362,
+ "learning_rate": 0.0006,
+ "loss": 4.838453769683838,
+ "step": 985
+ },
+ {
+ "epoch": 13.698689956331878,
+ "grad_norm": 0.01900297962129116,
+ "learning_rate": 0.0006,
+ "loss": 4.804868221282959,
+ "step": 986
+ },
+ {
+ "epoch": 13.712663755458514,
+ "grad_norm": 0.020292341709136963,
+ "learning_rate": 0.0006,
+ "loss": 4.609743118286133,
+ "step": 987
+ },
+ {
+ "epoch": 13.726637554585153,
+ "grad_norm": 0.017014775425195694,
+ "learning_rate": 0.0006,
+ "loss": 4.740612983703613,
+ "step": 988
+ },
+ {
+ "epoch": 13.74061135371179,
+ "grad_norm": 0.015199844725430012,
+ "learning_rate": 0.0006,
+ "loss": 4.719701766967773,
+ "step": 989
+ },
+ {
+ "epoch": 13.754585152838429,
+ "grad_norm": 0.015125414356589317,
+ "learning_rate": 0.0006,
+ "loss": 4.686114311218262,
+ "step": 990
+ },
+ {
+ "epoch": 13.768558951965066,
+ "grad_norm": 0.014779540710151196,
+ "learning_rate": 0.0006,
+ "loss": 4.674347877502441,
+ "step": 991
+ },
+ {
+ "epoch": 13.782532751091702,
+ "grad_norm": 0.015255163423717022,
+ "learning_rate": 0.0006,
+ "loss": 4.6309309005737305,
+ "step": 992
+ },
+ {
+ "epoch": 13.796506550218341,
+ "grad_norm": 0.013692197389900684,
+ "learning_rate": 0.0006,
+ "loss": 4.698497295379639,
+ "step": 993
+ },
+ {
+ "epoch": 13.810480349344978,
+ "grad_norm": 0.014529623091220856,
+ "learning_rate": 0.0006,
+ "loss": 4.628443717956543,
+ "step": 994
+ },
+ {
+ "epoch": 13.824454148471617,
+ "grad_norm": 0.01576659083366394,
+ "learning_rate": 0.0006,
+ "loss": 4.788015842437744,
+ "step": 995
+ },
+ {
+ "epoch": 13.838427947598253,
+ "grad_norm": 0.015834277495741844,
+ "learning_rate": 0.0006,
+ "loss": 4.672703742980957,
+ "step": 996
+ },
+ {
+ "epoch": 13.85240174672489,
+ "grad_norm": 0.015539503656327724,
+ "learning_rate": 0.0006,
+ "loss": 4.731503486633301,
+ "step": 997
+ },
+ {
+ "epoch": 13.866375545851529,
+ "grad_norm": 0.013463746756315231,
+ "learning_rate": 0.0006,
+ "loss": 4.7006330490112305,
+ "step": 998
+ },
+ {
+ "epoch": 13.880349344978166,
+ "grad_norm": 0.016377726569771767,
+ "learning_rate": 0.0006,
+ "loss": 4.636074066162109,
+ "step": 999
+ },
+ {
+ "epoch": 13.894323144104803,
+ "grad_norm": 0.016660034656524658,
+ "learning_rate": 0.0006,
+ "loss": 4.742693901062012,
+ "step": 1000
+ },
+ {
+ "epoch": 13.908296943231441,
+ "grad_norm": 0.014902938157320023,
+ "learning_rate": 0.0006,
+ "loss": 4.769811630249023,
+ "step": 1001
+ },
+ {
+ "epoch": 13.922270742358078,
+ "grad_norm": 0.014883476309478283,
+ "learning_rate": 0.0006,
+ "loss": 4.731159210205078,
+ "step": 1002
+ },
+ {
+ "epoch": 13.936244541484717,
+ "grad_norm": 0.014494217932224274,
+ "learning_rate": 0.0006,
+ "loss": 4.659485816955566,
+ "step": 1003
+ },
+ {
+ "epoch": 13.950218340611354,
+ "grad_norm": 0.015426991507411003,
+ "learning_rate": 0.0006,
+ "loss": 4.755229949951172,
+ "step": 1004
+ },
+ {
+ "epoch": 13.96419213973799,
+ "grad_norm": 0.016648586839437485,
+ "learning_rate": 0.0006,
+ "loss": 4.74111270904541,
+ "step": 1005
+ },
+ {
+ "epoch": 13.97816593886463,
+ "grad_norm": 0.016980541869997978,
+ "learning_rate": 0.0006,
+ "loss": 4.74072265625,
+ "step": 1006
+ },
+ {
+ "epoch": 13.992139737991266,
+ "grad_norm": 0.018070152029395103,
+ "learning_rate": 0.0006,
+ "loss": 4.732975006103516,
+ "step": 1007
+ },
+ {
+ "epoch": 14.0,
+ "grad_norm": 0.019910665228962898,
+ "learning_rate": 0.0006,
+ "loss": 4.634697914123535,
+ "step": 1008
+ },
+ {
+ "epoch": 14.0,
+ "eval_loss": 4.850640296936035,
+ "eval_runtime": 57.3039,
+ "eval_samples_per_second": 42.615,
+ "eval_steps_per_second": 1.344,
+ "step": 1008
+ },
+ {
+ "epoch": 14.013973799126637,
+ "grad_norm": 0.01980053074657917,
+ "learning_rate": 0.0006,
+ "loss": 4.628744602203369,
+ "step": 1009
+ },
+ {
+ "epoch": 14.027947598253276,
+ "grad_norm": 0.02497623674571514,
+ "learning_rate": 0.0006,
+ "loss": 4.75930118560791,
+ "step": 1010
+ },
+ {
+ "epoch": 14.041921397379912,
+ "grad_norm": 0.02794097363948822,
+ "learning_rate": 0.0006,
+ "loss": 4.691340446472168,
+ "step": 1011
+ },
+ {
+ "epoch": 14.055895196506551,
+ "grad_norm": 0.023725535720586777,
+ "learning_rate": 0.0006,
+ "loss": 4.733664512634277,
+ "step": 1012
+ },
+ {
+ "epoch": 14.069868995633188,
+ "grad_norm": 0.022966474294662476,
+ "learning_rate": 0.0006,
+ "loss": 4.664808750152588,
+ "step": 1013
+ },
+ {
+ "epoch": 14.083842794759825,
+ "grad_norm": 0.02530563250184059,
+ "learning_rate": 0.0006,
+ "loss": 4.8047285079956055,
+ "step": 1014
+ },
+ {
+ "epoch": 14.097816593886463,
+ "grad_norm": 0.02491082064807415,
+ "learning_rate": 0.0006,
+ "loss": 4.667956352233887,
+ "step": 1015
+ },
+ {
+ "epoch": 14.1117903930131,
+ "grad_norm": 0.022110294550657272,
+ "learning_rate": 0.0006,
+ "loss": 4.666140556335449,
+ "step": 1016
+ },
+ {
+ "epoch": 14.125764192139737,
+ "grad_norm": 0.02065849117934704,
+ "learning_rate": 0.0006,
+ "loss": 4.581960678100586,
+ "step": 1017
+ },
+ {
+ "epoch": 14.139737991266376,
+ "grad_norm": 0.01941298507153988,
+ "learning_rate": 0.0006,
+ "loss": 4.614226341247559,
+ "step": 1018
+ },
+ {
+ "epoch": 14.153711790393013,
+ "grad_norm": 0.022441856563091278,
+ "learning_rate": 0.0006,
+ "loss": 4.638352394104004,
+ "step": 1019
+ },
+ {
+ "epoch": 14.167685589519651,
+ "grad_norm": 0.021519694477319717,
+ "learning_rate": 0.0006,
+ "loss": 4.700160980224609,
+ "step": 1020
+ },
+ {
+ "epoch": 14.181659388646288,
+ "grad_norm": 0.018956486135721207,
+ "learning_rate": 0.0006,
+ "loss": 4.649593353271484,
+ "step": 1021
+ },
+ {
+ "epoch": 14.195633187772925,
+ "grad_norm": 0.016681140288710594,
+ "learning_rate": 0.0006,
+ "loss": 4.603952407836914,
+ "step": 1022
+ },
+ {
+ "epoch": 14.209606986899564,
+ "grad_norm": 0.01788809522986412,
+ "learning_rate": 0.0006,
+ "loss": 4.669160842895508,
+ "step": 1023
+ },
+ {
+ "epoch": 14.2235807860262,
+ "grad_norm": 0.016047228127717972,
+ "learning_rate": 0.0006,
+ "loss": 4.6367950439453125,
+ "step": 1024
+ },
+ {
+ "epoch": 14.237554585152838,
+ "grad_norm": 0.01614239625632763,
+ "learning_rate": 0.0006,
+ "loss": 4.664529323577881,
+ "step": 1025
+ },
+ {
+ "epoch": 14.251528384279476,
+ "grad_norm": 0.016362549737095833,
+ "learning_rate": 0.0006,
+ "loss": 4.702373504638672,
+ "step": 1026
+ },
+ {
+ "epoch": 14.265502183406113,
+ "grad_norm": 0.015800409018993378,
+ "learning_rate": 0.0006,
+ "loss": 4.7594428062438965,
+ "step": 1027
+ },
+ {
+ "epoch": 14.279475982532752,
+ "grad_norm": 0.016730139032006264,
+ "learning_rate": 0.0006,
+ "loss": 4.677244186401367,
+ "step": 1028
+ },
+ {
+ "epoch": 14.293449781659389,
+ "grad_norm": 0.01661759801208973,
+ "learning_rate": 0.0006,
+ "loss": 4.633679389953613,
+ "step": 1029
+ },
+ {
+ "epoch": 14.307423580786025,
+ "grad_norm": 0.017673347145318985,
+ "learning_rate": 0.0006,
+ "loss": 4.649107456207275,
+ "step": 1030
+ },
+ {
+ "epoch": 14.321397379912664,
+ "grad_norm": 0.019243352115154266,
+ "learning_rate": 0.0006,
+ "loss": 4.611257553100586,
+ "step": 1031
+ },
+ {
+ "epoch": 14.335371179039301,
+ "grad_norm": 0.019817164167761803,
+ "learning_rate": 0.0006,
+ "loss": 4.588086128234863,
+ "step": 1032
+ },
+ {
+ "epoch": 14.34934497816594,
+ "grad_norm": 0.022292211651802063,
+ "learning_rate": 0.0006,
+ "loss": 4.783066749572754,
+ "step": 1033
+ },
+ {
+ "epoch": 14.363318777292577,
+ "grad_norm": 0.02140096202492714,
+ "learning_rate": 0.0006,
+ "loss": 4.604718208312988,
+ "step": 1034
+ },
+ {
+ "epoch": 14.377292576419213,
+ "grad_norm": 0.0187817495316267,
+ "learning_rate": 0.0006,
+ "loss": 4.721773147583008,
+ "step": 1035
+ },
+ {
+ "epoch": 14.391266375545852,
+ "grad_norm": 0.019879210740327835,
+ "learning_rate": 0.0006,
+ "loss": 4.789286136627197,
+ "step": 1036
+ },
+ {
+ "epoch": 14.405240174672489,
+ "grad_norm": 0.01885785162448883,
+ "learning_rate": 0.0006,
+ "loss": 4.669743537902832,
+ "step": 1037
+ },
+ {
+ "epoch": 14.419213973799126,
+ "grad_norm": 0.018324270844459534,
+ "learning_rate": 0.0006,
+ "loss": 4.65901517868042,
+ "step": 1038
+ },
+ {
+ "epoch": 14.433187772925764,
+ "grad_norm": 0.017664821818470955,
+ "learning_rate": 0.0006,
+ "loss": 4.667839527130127,
+ "step": 1039
+ },
+ {
+ "epoch": 14.447161572052401,
+ "grad_norm": 0.015702905133366585,
+ "learning_rate": 0.0006,
+ "loss": 4.64700984954834,
+ "step": 1040
+ },
+ {
+ "epoch": 14.46113537117904,
+ "grad_norm": 0.016674194484949112,
+ "learning_rate": 0.0006,
+ "loss": 4.694093227386475,
+ "step": 1041
+ },
+ {
+ "epoch": 14.475109170305677,
+ "grad_norm": 0.016869477927684784,
+ "learning_rate": 0.0006,
+ "loss": 4.619544982910156,
+ "step": 1042
+ },
+ {
+ "epoch": 14.489082969432314,
+ "grad_norm": 0.016557540744543076,
+ "learning_rate": 0.0006,
+ "loss": 4.568667411804199,
+ "step": 1043
+ },
+ {
+ "epoch": 14.503056768558952,
+ "grad_norm": 0.014500945806503296,
+ "learning_rate": 0.0006,
+ "loss": 4.561741828918457,
+ "step": 1044
+ },
+ {
+ "epoch": 14.51703056768559,
+ "grad_norm": 0.014228739775717258,
+ "learning_rate": 0.0006,
+ "loss": 4.654277324676514,
+ "step": 1045
+ },
+ {
+ "epoch": 14.531004366812226,
+ "grad_norm": 0.013894238509237766,
+ "learning_rate": 0.0006,
+ "loss": 4.5867719650268555,
+ "step": 1046
+ },
+ {
+ "epoch": 14.544978165938865,
+ "grad_norm": 0.01550068985670805,
+ "learning_rate": 0.0006,
+ "loss": 4.61168098449707,
+ "step": 1047
+ },
+ {
+ "epoch": 14.558951965065502,
+ "grad_norm": 0.015940772369503975,
+ "learning_rate": 0.0006,
+ "loss": 4.691247940063477,
+ "step": 1048
+ },
+ {
+ "epoch": 14.57292576419214,
+ "grad_norm": 0.01714535616338253,
+ "learning_rate": 0.0006,
+ "loss": 4.61183500289917,
+ "step": 1049
+ },
+ {
+ "epoch": 14.586899563318777,
+ "grad_norm": 0.017925038933753967,
+ "learning_rate": 0.0006,
+ "loss": 4.719022750854492,
+ "step": 1050
+ },
+ {
+ "epoch": 14.600873362445414,
+ "grad_norm": 0.018141552805900574,
+ "learning_rate": 0.0006,
+ "loss": 4.760281562805176,
+ "step": 1051
+ },
+ {
+ "epoch": 14.614847161572053,
+ "grad_norm": 0.017131328582763672,
+ "learning_rate": 0.0006,
+ "loss": 4.709680080413818,
+ "step": 1052
+ },
+ {
+ "epoch": 14.62882096069869,
+ "grad_norm": 0.018730048090219498,
+ "learning_rate": 0.0006,
+ "loss": 4.5994720458984375,
+ "step": 1053
+ },
+ {
+ "epoch": 14.642794759825328,
+ "grad_norm": 0.018894298002123833,
+ "learning_rate": 0.0006,
+ "loss": 4.669070720672607,
+ "step": 1054
+ },
+ {
+ "epoch": 14.656768558951965,
+ "grad_norm": 0.017643198370933533,
+ "learning_rate": 0.0006,
+ "loss": 4.641790390014648,
+ "step": 1055
+ },
+ {
+ "epoch": 14.670742358078602,
+ "grad_norm": 0.017647244036197662,
+ "learning_rate": 0.0006,
+ "loss": 4.653899192810059,
+ "step": 1056
+ },
+ {
+ "epoch": 14.68471615720524,
+ "grad_norm": 0.017304198816418648,
+ "learning_rate": 0.0006,
+ "loss": 4.7487473487854,
+ "step": 1057
+ },
+ {
+ "epoch": 14.698689956331878,
+ "grad_norm": 0.01638447679579258,
+ "learning_rate": 0.0006,
+ "loss": 4.71945858001709,
+ "step": 1058
+ },
+ {
+ "epoch": 14.712663755458514,
+ "grad_norm": 0.015892071649432182,
+ "learning_rate": 0.0006,
+ "loss": 4.744903564453125,
+ "step": 1059
+ },
+ {
+ "epoch": 14.726637554585153,
+ "grad_norm": 0.017420614138245583,
+ "learning_rate": 0.0006,
+ "loss": 4.781563758850098,
+ "step": 1060
+ },
+ {
+ "epoch": 14.74061135371179,
+ "grad_norm": 0.016498520970344543,
+ "learning_rate": 0.0006,
+ "loss": 4.622779846191406,
+ "step": 1061
+ },
+ {
+ "epoch": 14.754585152838429,
+ "grad_norm": 0.017482127994298935,
+ "learning_rate": 0.0006,
+ "loss": 4.62825870513916,
+ "step": 1062
+ },
+ {
+ "epoch": 14.768558951965066,
+ "grad_norm": 0.019018379971385002,
+ "learning_rate": 0.0006,
+ "loss": 4.693194389343262,
+ "step": 1063
+ },
+ {
+ "epoch": 14.782532751091702,
+ "grad_norm": 0.01993914321064949,
+ "learning_rate": 0.0006,
+ "loss": 4.742903709411621,
+ "step": 1064
+ },
+ {
+ "epoch": 14.796506550218341,
+ "grad_norm": 0.02301335707306862,
+ "learning_rate": 0.0006,
+ "loss": 4.613251686096191,
+ "step": 1065
+ },
+ {
+ "epoch": 14.810480349344978,
+ "grad_norm": 0.024088222533464432,
+ "learning_rate": 0.0006,
+ "loss": 4.594305992126465,
+ "step": 1066
+ },
+ {
+ "epoch": 14.824454148471617,
+ "grad_norm": 0.02493320405483246,
+ "learning_rate": 0.0006,
+ "loss": 4.700588226318359,
+ "step": 1067
+ },
+ {
+ "epoch": 14.838427947598253,
+ "grad_norm": 0.02501937560737133,
+ "learning_rate": 0.0006,
+ "loss": 4.601742744445801,
+ "step": 1068
+ },
+ {
+ "epoch": 14.85240174672489,
+ "grad_norm": 0.023490751162171364,
+ "learning_rate": 0.0006,
+ "loss": 4.589211463928223,
+ "step": 1069
+ },
+ {
+ "epoch": 14.866375545851529,
+ "grad_norm": 0.019695665687322617,
+ "learning_rate": 0.0006,
+ "loss": 4.635081768035889,
+ "step": 1070
+ },
+ {
+ "epoch": 14.880349344978166,
+ "grad_norm": 0.019276980310678482,
+ "learning_rate": 0.0006,
+ "loss": 4.737002372741699,
+ "step": 1071
+ },
+ {
+ "epoch": 14.894323144104803,
+ "grad_norm": 0.022430511191487312,
+ "learning_rate": 0.0006,
+ "loss": 4.808295726776123,
+ "step": 1072
+ },
+ {
+ "epoch": 14.908296943231441,
+ "grad_norm": 0.021543916314840317,
+ "learning_rate": 0.0006,
+ "loss": 4.559475421905518,
+ "step": 1073
+ },
+ {
+ "epoch": 14.922270742358078,
+ "grad_norm": 0.019885241985321045,
+ "learning_rate": 0.0006,
+ "loss": 4.647948741912842,
+ "step": 1074
+ },
+ {
+ "epoch": 14.936244541484717,
+ "grad_norm": 0.02115175500512123,
+ "learning_rate": 0.0006,
+ "loss": 4.680505752563477,
+ "step": 1075
+ },
+ {
+ "epoch": 14.950218340611354,
+ "grad_norm": 0.02156151458621025,
+ "learning_rate": 0.0006,
+ "loss": 4.673477649688721,
+ "step": 1076
+ },
+ {
+ "epoch": 14.96419213973799,
+ "grad_norm": 0.02184601128101349,
+ "learning_rate": 0.0006,
+ "loss": 4.644402027130127,
+ "step": 1077
+ },
+ {
+ "epoch": 14.97816593886463,
+ "grad_norm": 0.0211983360350132,
+ "learning_rate": 0.0006,
+ "loss": 4.821558475494385,
+ "step": 1078
+ },
+ {
+ "epoch": 14.992139737991266,
+ "grad_norm": 0.019636567682027817,
+ "learning_rate": 0.0006,
+ "loss": 4.578868865966797,
+ "step": 1079
+ },
+ {
+ "epoch": 15.0,
+ "grad_norm": 0.019997350871562958,
+ "learning_rate": 0.0006,
+ "loss": 4.691307544708252,
+ "step": 1080
+ },
+ {
+ "epoch": 15.0,
+ "eval_loss": 4.872478008270264,
+ "eval_runtime": 56.8513,
+ "eval_samples_per_second": 42.954,
+ "eval_steps_per_second": 1.354,
+ "step": 1080
+ },
+ {
+ "epoch": 15.013973799126637,
+ "grad_norm": 0.018830662593245506,
+ "learning_rate": 0.0006,
+ "loss": 4.761779308319092,
+ "step": 1081
+ },
+ {
+ "epoch": 15.027947598253276,
+ "grad_norm": 0.016876481473445892,
+ "learning_rate": 0.0006,
+ "loss": 4.61148738861084,
+ "step": 1082
+ },
+ {
+ "epoch": 15.041921397379912,
+ "grad_norm": 0.015550863929092884,
+ "learning_rate": 0.0006,
+ "loss": 4.676124095916748,
+ "step": 1083
+ },
+ {
+ "epoch": 15.055895196506551,
+ "grad_norm": 0.017178896814584732,
+ "learning_rate": 0.0006,
+ "loss": 4.588935852050781,
+ "step": 1084
+ },
+ {
+ "epoch": 15.069868995633188,
+ "grad_norm": 0.015465447679162025,
+ "learning_rate": 0.0006,
+ "loss": 4.605499267578125,
+ "step": 1085
+ },
+ {
+ "epoch": 15.083842794759825,
+ "grad_norm": 0.01606195978820324,
+ "learning_rate": 0.0006,
+ "loss": 4.723373889923096,
+ "step": 1086
+ },
+ {
+ "epoch": 15.097816593886463,
+ "grad_norm": 0.016521936282515526,
+ "learning_rate": 0.0006,
+ "loss": 4.646608352661133,
+ "step": 1087
+ },
+ {
+ "epoch": 15.1117903930131,
+ "grad_norm": 0.01589205674827099,
+ "learning_rate": 0.0006,
+ "loss": 4.498138427734375,
+ "step": 1088
+ },
+ {
+ "epoch": 15.125764192139737,
+ "grad_norm": 0.01524856686592102,
+ "learning_rate": 0.0006,
+ "loss": 4.594562530517578,
+ "step": 1089
+ },
+ {
+ "epoch": 15.139737991266376,
+ "grad_norm": 0.014714641496539116,
+ "learning_rate": 0.0006,
+ "loss": 4.630744934082031,
+ "step": 1090
+ },
+ {
+ "epoch": 15.153711790393013,
+ "grad_norm": 0.015333331190049648,
+ "learning_rate": 0.0006,
+ "loss": 4.682388782501221,
+ "step": 1091
+ },
+ {
+ "epoch": 15.167685589519651,
+ "grad_norm": 0.01550073828548193,
+ "learning_rate": 0.0006,
+ "loss": 4.692652702331543,
+ "step": 1092
+ },
+ {
+ "epoch": 15.181659388646288,
+ "grad_norm": 0.014783318154513836,
+ "learning_rate": 0.0006,
+ "loss": 4.6223835945129395,
+ "step": 1093
+ },
+ {
+ "epoch": 15.195633187772925,
+ "grad_norm": 0.014372751116752625,
+ "learning_rate": 0.0006,
+ "loss": 4.52687406539917,
+ "step": 1094
+ },
+ {
+ "epoch": 15.209606986899564,
+ "grad_norm": 0.014779130928218365,
+ "learning_rate": 0.0006,
+ "loss": 4.637009620666504,
+ "step": 1095
+ },
+ {
+ "epoch": 15.2235807860262,
+ "grad_norm": 0.015697073191404343,
+ "learning_rate": 0.0006,
+ "loss": 4.682981014251709,
+ "step": 1096
+ },
+ {
+ "epoch": 15.237554585152838,
+ "grad_norm": 0.016523664817214012,
+ "learning_rate": 0.0006,
+ "loss": 4.705722808837891,
+ "step": 1097
+ },
+ {
+ "epoch": 15.251528384279476,
+ "grad_norm": 0.017633996903896332,
+ "learning_rate": 0.0006,
+ "loss": 4.574686050415039,
+ "step": 1098
+ },
+ {
+ "epoch": 15.265502183406113,
+ "grad_norm": 0.019955076277256012,
+ "learning_rate": 0.0006,
+ "loss": 4.634450912475586,
+ "step": 1099
+ },
+ {
+ "epoch": 15.279475982532752,
+ "grad_norm": 0.02132618986070156,
+ "learning_rate": 0.0006,
+ "loss": 4.56110143661499,
+ "step": 1100
+ },
+ {
+ "epoch": 15.293449781659389,
+ "grad_norm": 0.0207593385130167,
+ "learning_rate": 0.0006,
+ "loss": 4.629396438598633,
+ "step": 1101
+ },
+ {
+ "epoch": 15.307423580786025,
+ "grad_norm": 0.017882540822029114,
+ "learning_rate": 0.0006,
+ "loss": 4.730963230133057,
+ "step": 1102
+ },
+ {
+ "epoch": 15.321397379912664,
+ "grad_norm": 0.01917876861989498,
+ "learning_rate": 0.0006,
+ "loss": 4.64585542678833,
+ "step": 1103
+ },
+ {
+ "epoch": 15.335371179039301,
+ "grad_norm": 0.01943974196910858,
+ "learning_rate": 0.0006,
+ "loss": 4.656181335449219,
+ "step": 1104
+ },
+ {
+ "epoch": 15.34934497816594,
+ "grad_norm": 0.019904715940356255,
+ "learning_rate": 0.0006,
+ "loss": 4.619329452514648,
+ "step": 1105
+ },
+ {
+ "epoch": 15.363318777292577,
+ "grad_norm": 0.019174346700310707,
+ "learning_rate": 0.0006,
+ "loss": 4.647892475128174,
+ "step": 1106
+ },
+ {
+ "epoch": 15.377292576419213,
+ "grad_norm": 0.018244758248329163,
+ "learning_rate": 0.0006,
+ "loss": 4.56166934967041,
+ "step": 1107
+ },
+ {
+ "epoch": 15.391266375545852,
+ "grad_norm": 0.018914205953478813,
+ "learning_rate": 0.0006,
+ "loss": 4.648063659667969,
+ "step": 1108
+ },
+ {
+ "epoch": 15.405240174672489,
+ "grad_norm": 0.01961071416735649,
+ "learning_rate": 0.0006,
+ "loss": 4.665525436401367,
+ "step": 1109
+ },
+ {
+ "epoch": 15.419213973799126,
+ "grad_norm": 0.01841311901807785,
+ "learning_rate": 0.0006,
+ "loss": 4.611342430114746,
+ "step": 1110
+ },
+ {
+ "epoch": 15.433187772925764,
+ "grad_norm": 0.01863427273929119,
+ "learning_rate": 0.0006,
+ "loss": 4.664256572723389,
+ "step": 1111
+ },
+ {
+ "epoch": 15.447161572052401,
+ "grad_norm": 0.020506877452135086,
+ "learning_rate": 0.0006,
+ "loss": 4.631237506866455,
+ "step": 1112
+ },
+ {
+ "epoch": 15.46113537117904,
+ "grad_norm": 0.02223833277821541,
+ "learning_rate": 0.0006,
+ "loss": 4.668654441833496,
+ "step": 1113
+ },
+ {
+ "epoch": 15.475109170305677,
+ "grad_norm": 0.023336702957749367,
+ "learning_rate": 0.0006,
+ "loss": 4.562833309173584,
+ "step": 1114
+ },
+ {
+ "epoch": 15.489082969432314,
+ "grad_norm": 0.023052413016557693,
+ "learning_rate": 0.0006,
+ "loss": 4.621532917022705,
+ "step": 1115
+ },
+ {
+ "epoch": 15.503056768558952,
+ "grad_norm": 0.023420924320816994,
+ "learning_rate": 0.0006,
+ "loss": 4.665465354919434,
+ "step": 1116
+ },
+ {
+ "epoch": 15.51703056768559,
+ "grad_norm": 0.024071281775832176,
+ "learning_rate": 0.0006,
+ "loss": 4.595980644226074,
+ "step": 1117
+ },
+ {
+ "epoch": 15.531004366812226,
+ "grad_norm": 0.023406682536005974,
+ "learning_rate": 0.0006,
+ "loss": 4.642635345458984,
+ "step": 1118
+ },
+ {
+ "epoch": 15.544978165938865,
+ "grad_norm": 0.02138647809624672,
+ "learning_rate": 0.0006,
+ "loss": 4.704513072967529,
+ "step": 1119
+ },
+ {
+ "epoch": 15.558951965065502,
+ "grad_norm": 0.0221384447067976,
+ "learning_rate": 0.0006,
+ "loss": 4.567503452301025,
+ "step": 1120
+ },
+ {
+ "epoch": 15.57292576419214,
+ "grad_norm": 0.020879851654171944,
+ "learning_rate": 0.0006,
+ "loss": 4.664057731628418,
+ "step": 1121
+ },
+ {
+ "epoch": 15.586899563318777,
+ "grad_norm": 0.021478967741131783,
+ "learning_rate": 0.0006,
+ "loss": 4.539627552032471,
+ "step": 1122
+ },
+ {
+ "epoch": 15.600873362445414,
+ "grad_norm": 0.022748656570911407,
+ "learning_rate": 0.0006,
+ "loss": 4.5654497146606445,
+ "step": 1123
+ },
+ {
+ "epoch": 15.614847161572053,
+ "grad_norm": 0.021870248019695282,
+ "learning_rate": 0.0006,
+ "loss": 4.56899356842041,
+ "step": 1124
+ },
+ {
+ "epoch": 15.62882096069869,
+ "grad_norm": 0.021711355075240135,
+ "learning_rate": 0.0006,
+ "loss": 4.657808303833008,
+ "step": 1125
+ },
+ {
+ "epoch": 15.642794759825328,
+ "grad_norm": 0.022810909897089005,
+ "learning_rate": 0.0006,
+ "loss": 4.6526689529418945,
+ "step": 1126
+ },
+ {
+ "epoch": 15.656768558951965,
+ "grad_norm": 0.022561343386769295,
+ "learning_rate": 0.0006,
+ "loss": 4.534887790679932,
+ "step": 1127
+ },
+ {
+ "epoch": 15.670742358078602,
+ "grad_norm": 0.019909802824258804,
+ "learning_rate": 0.0006,
+ "loss": 4.600790023803711,
+ "step": 1128
+ },
+ {
+ "epoch": 15.68471615720524,
+ "grad_norm": 0.020092325285077095,
+ "learning_rate": 0.0006,
+ "loss": 4.773830413818359,
+ "step": 1129
+ },
+ {
+ "epoch": 15.698689956331878,
+ "grad_norm": 0.018834445625543594,
+ "learning_rate": 0.0006,
+ "loss": 4.628868103027344,
+ "step": 1130
+ },
+ {
+ "epoch": 15.712663755458514,
+ "grad_norm": 0.017552798613905907,
+ "learning_rate": 0.0006,
+ "loss": 4.647500038146973,
+ "step": 1131
+ },
+ {
+ "epoch": 15.726637554585153,
+ "grad_norm": 0.01751251146197319,
+ "learning_rate": 0.0006,
+ "loss": 4.590545654296875,
+ "step": 1132
+ },
+ {
+ "epoch": 15.74061135371179,
+ "grad_norm": 0.01699099875986576,
+ "learning_rate": 0.0006,
+ "loss": 4.509699821472168,
+ "step": 1133
+ },
+ {
+ "epoch": 15.754585152838429,
+ "grad_norm": 0.015808649361133575,
+ "learning_rate": 0.0006,
+ "loss": 4.606992244720459,
+ "step": 1134
+ },
+ {
+ "epoch": 15.768558951965066,
+ "grad_norm": 0.01655443385243416,
+ "learning_rate": 0.0006,
+ "loss": 4.585219860076904,
+ "step": 1135
+ },
+ {
+ "epoch": 15.782532751091702,
+ "grad_norm": 0.01708163693547249,
+ "learning_rate": 0.0006,
+ "loss": 4.668147087097168,
+ "step": 1136
+ },
+ {
+ "epoch": 15.796506550218341,
+ "grad_norm": 0.017161913216114044,
+ "learning_rate": 0.0006,
+ "loss": 4.7170634269714355,
+ "step": 1137
+ },
+ {
+ "epoch": 15.810480349344978,
+ "grad_norm": 0.017262674868106842,
+ "learning_rate": 0.0006,
+ "loss": 4.572132110595703,
+ "step": 1138
+ },
+ {
+ "epoch": 15.824454148471617,
+ "grad_norm": 0.017713576555252075,
+ "learning_rate": 0.0006,
+ "loss": 4.702538967132568,
+ "step": 1139
+ },
+ {
+ "epoch": 15.838427947598253,
+ "grad_norm": 0.018730733543634415,
+ "learning_rate": 0.0006,
+ "loss": 4.629721641540527,
+ "step": 1140
+ },
+ {
+ "epoch": 15.85240174672489,
+ "grad_norm": 0.017960187047719955,
+ "learning_rate": 0.0006,
+ "loss": 4.661039352416992,
+ "step": 1141
+ },
+ {
+ "epoch": 15.866375545851529,
+ "grad_norm": 0.01750079356133938,
+ "learning_rate": 0.0006,
+ "loss": 4.503710746765137,
+ "step": 1142
+ },
+ {
+ "epoch": 15.880349344978166,
+ "grad_norm": 0.016630422323942184,
+ "learning_rate": 0.0006,
+ "loss": 4.6582136154174805,
+ "step": 1143
+ },
+ {
+ "epoch": 15.894323144104803,
+ "grad_norm": 0.015896832570433617,
+ "learning_rate": 0.0006,
+ "loss": 4.596693992614746,
+ "step": 1144
+ },
+ {
+ "epoch": 15.908296943231441,
+ "grad_norm": 0.01566407084465027,
+ "learning_rate": 0.0006,
+ "loss": 4.6337199211120605,
+ "step": 1145
+ },
+ {
+ "epoch": 15.922270742358078,
+ "grad_norm": 0.016823330894112587,
+ "learning_rate": 0.0006,
+ "loss": 4.686994552612305,
+ "step": 1146
+ },
+ {
+ "epoch": 15.936244541484717,
+ "grad_norm": 0.016208255663514137,
+ "learning_rate": 0.0006,
+ "loss": 4.6205549240112305,
+ "step": 1147
+ },
+ {
+ "epoch": 15.950218340611354,
+ "grad_norm": 0.0171707421541214,
+ "learning_rate": 0.0006,
+ "loss": 4.582968711853027,
+ "step": 1148
+ },
+ {
+ "epoch": 15.96419213973799,
+ "grad_norm": 0.017009446397423744,
+ "learning_rate": 0.0006,
+ "loss": 4.510274887084961,
+ "step": 1149
+ },
+ {
+ "epoch": 15.97816593886463,
+ "grad_norm": 0.01565811224281788,
+ "learning_rate": 0.0006,
+ "loss": 4.55834436416626,
+ "step": 1150
+ },
+ {
+ "epoch": 15.992139737991266,
+ "grad_norm": 0.015276037156581879,
+ "learning_rate": 0.0006,
+ "loss": 4.537870407104492,
+ "step": 1151
+ },
+ {
+ "epoch": 16.0,
+ "grad_norm": 0.017305882647633553,
+ "learning_rate": 0.0006,
+ "loss": 4.611305236816406,
+ "step": 1152
+ },
+ {
+ "epoch": 16.0,
+ "eval_loss": 4.775513172149658,
+ "eval_runtime": 56.7341,
+ "eval_samples_per_second": 43.043,
+ "eval_steps_per_second": 1.357,
+ "step": 1152
+ },
+ {
+ "epoch": 16.01397379912664,
+ "grad_norm": 0.016752436757087708,
+ "learning_rate": 0.0006,
+ "loss": 4.541312217712402,
+ "step": 1153
+ },
+ {
+ "epoch": 16.027947598253274,
+ "grad_norm": 0.01643134094774723,
+ "learning_rate": 0.0006,
+ "loss": 4.585402965545654,
+ "step": 1154
+ },
+ {
+ "epoch": 16.041921397379912,
+ "grad_norm": 0.01888720504939556,
+ "learning_rate": 0.0006,
+ "loss": 4.597825050354004,
+ "step": 1155
+ },
+ {
+ "epoch": 16.05589519650655,
+ "grad_norm": 0.022044632583856583,
+ "learning_rate": 0.0006,
+ "loss": 4.532772064208984,
+ "step": 1156
+ },
+ {
+ "epoch": 16.069868995633186,
+ "grad_norm": 0.022986987605690956,
+ "learning_rate": 0.0006,
+ "loss": 4.551351547241211,
+ "step": 1157
+ },
+ {
+ "epoch": 16.083842794759825,
+ "grad_norm": 0.020116791129112244,
+ "learning_rate": 0.0006,
+ "loss": 4.484149932861328,
+ "step": 1158
+ },
+ {
+ "epoch": 16.097816593886463,
+ "grad_norm": 0.021442247554659843,
+ "learning_rate": 0.0006,
+ "loss": 4.515316486358643,
+ "step": 1159
+ },
+ {
+ "epoch": 16.111790393013102,
+ "grad_norm": 0.020974772050976753,
+ "learning_rate": 0.0006,
+ "loss": 4.51606559753418,
+ "step": 1160
+ },
+ {
+ "epoch": 16.125764192139737,
+ "grad_norm": 0.021274864673614502,
+ "learning_rate": 0.0006,
+ "loss": 4.587329864501953,
+ "step": 1161
+ },
+ {
+ "epoch": 16.139737991266376,
+ "grad_norm": 0.022799784317612648,
+ "learning_rate": 0.0006,
+ "loss": 4.568149089813232,
+ "step": 1162
+ },
+ {
+ "epoch": 16.153711790393015,
+ "grad_norm": 0.021784933283925056,
+ "learning_rate": 0.0006,
+ "loss": 4.515336990356445,
+ "step": 1163
+ },
+ {
+ "epoch": 16.16768558951965,
+ "grad_norm": 0.024261271581053734,
+ "learning_rate": 0.0006,
+ "loss": 4.5127081871032715,
+ "step": 1164
+ },
+ {
+ "epoch": 16.18165938864629,
+ "grad_norm": 0.02248375490307808,
+ "learning_rate": 0.0006,
+ "loss": 4.754251003265381,
+ "step": 1165
+ },
+ {
+ "epoch": 16.195633187772927,
+ "grad_norm": 0.02136492356657982,
+ "learning_rate": 0.0006,
+ "loss": 4.58193826675415,
+ "step": 1166
+ },
+ {
+ "epoch": 16.209606986899562,
+ "grad_norm": 0.020727066323161125,
+ "learning_rate": 0.0006,
+ "loss": 4.564209938049316,
+ "step": 1167
+ },
+ {
+ "epoch": 16.2235807860262,
+ "grad_norm": 0.017789119854569435,
+ "learning_rate": 0.0006,
+ "loss": 4.698680877685547,
+ "step": 1168
+ },
+ {
+ "epoch": 16.23755458515284,
+ "grad_norm": 0.017603034153580666,
+ "learning_rate": 0.0006,
+ "loss": 4.501171588897705,
+ "step": 1169
+ },
+ {
+ "epoch": 16.251528384279474,
+ "grad_norm": 0.017345862463116646,
+ "learning_rate": 0.0006,
+ "loss": 4.593590259552002,
+ "step": 1170
+ },
+ {
+ "epoch": 16.265502183406113,
+ "grad_norm": 0.018047425895929337,
+ "learning_rate": 0.0006,
+ "loss": 4.597467422485352,
+ "step": 1171
+ },
+ {
+ "epoch": 16.27947598253275,
+ "grad_norm": 0.01947222836315632,
+ "learning_rate": 0.0006,
+ "loss": 4.5072221755981445,
+ "step": 1172
+ },
+ {
+ "epoch": 16.29344978165939,
+ "grad_norm": 0.019598500803112984,
+ "learning_rate": 0.0006,
+ "loss": 4.601351737976074,
+ "step": 1173
+ },
+ {
+ "epoch": 16.307423580786025,
+ "grad_norm": 0.020002514123916626,
+ "learning_rate": 0.0006,
+ "loss": 4.580431938171387,
+ "step": 1174
+ },
+ {
+ "epoch": 16.321397379912664,
+ "grad_norm": 0.0216530729085207,
+ "learning_rate": 0.0006,
+ "loss": 4.424681663513184,
+ "step": 1175
+ },
+ {
+ "epoch": 16.335371179039303,
+ "grad_norm": 0.01949235238134861,
+ "learning_rate": 0.0006,
+ "loss": 4.751883029937744,
+ "step": 1176
+ },
+ {
+ "epoch": 16.349344978165938,
+ "grad_norm": 0.017779016867280006,
+ "learning_rate": 0.0006,
+ "loss": 4.541490077972412,
+ "step": 1177
+ },
+ {
+ "epoch": 16.363318777292577,
+ "grad_norm": 0.01909121684730053,
+ "learning_rate": 0.0006,
+ "loss": 4.594141006469727,
+ "step": 1178
+ },
+ {
+ "epoch": 16.377292576419215,
+ "grad_norm": 0.019894849509000778,
+ "learning_rate": 0.0006,
+ "loss": 4.555070400238037,
+ "step": 1179
+ },
+ {
+ "epoch": 16.39126637554585,
+ "grad_norm": 0.01993032731115818,
+ "learning_rate": 0.0006,
+ "loss": 4.521401405334473,
+ "step": 1180
+ },
+ {
+ "epoch": 16.40524017467249,
+ "grad_norm": 0.0210788045078516,
+ "learning_rate": 0.0006,
+ "loss": 4.672842979431152,
+ "step": 1181
+ },
+ {
+ "epoch": 16.419213973799128,
+ "grad_norm": 0.02274598926305771,
+ "learning_rate": 0.0006,
+ "loss": 4.629504203796387,
+ "step": 1182
+ },
+ {
+ "epoch": 16.433187772925763,
+ "grad_norm": 0.020448120310902596,
+ "learning_rate": 0.0006,
+ "loss": 4.548209190368652,
+ "step": 1183
+ },
+ {
+ "epoch": 16.4471615720524,
+ "grad_norm": 0.017565643414855003,
+ "learning_rate": 0.0006,
+ "loss": 4.529054641723633,
+ "step": 1184
+ },
+ {
+ "epoch": 16.46113537117904,
+ "grad_norm": 0.019504237920045853,
+ "learning_rate": 0.0006,
+ "loss": 4.48182487487793,
+ "step": 1185
+ },
+ {
+ "epoch": 16.475109170305675,
+ "grad_norm": 0.016823742538690567,
+ "learning_rate": 0.0006,
+ "loss": 4.608696937561035,
+ "step": 1186
+ },
+ {
+ "epoch": 16.489082969432314,
+ "grad_norm": 0.01602894812822342,
+ "learning_rate": 0.0006,
+ "loss": 4.620062351226807,
+ "step": 1187
+ },
+ {
+ "epoch": 16.503056768558952,
+ "grad_norm": 0.016615642234683037,
+ "learning_rate": 0.0006,
+ "loss": 4.69378662109375,
+ "step": 1188
+ },
+ {
+ "epoch": 16.51703056768559,
+ "grad_norm": 0.018576730042696,
+ "learning_rate": 0.0006,
+ "loss": 4.537041664123535,
+ "step": 1189
+ },
+ {
+ "epoch": 16.531004366812226,
+ "grad_norm": 0.018611151725053787,
+ "learning_rate": 0.0006,
+ "loss": 4.496169090270996,
+ "step": 1190
+ },
+ {
+ "epoch": 16.544978165938865,
+ "grad_norm": 0.018622413277626038,
+ "learning_rate": 0.0006,
+ "loss": 4.557358264923096,
+ "step": 1191
+ },
+ {
+ "epoch": 16.558951965065503,
+ "grad_norm": 0.01823544315993786,
+ "learning_rate": 0.0006,
+ "loss": 4.730792999267578,
+ "step": 1192
+ },
+ {
+ "epoch": 16.57292576419214,
+ "grad_norm": 0.018796846270561218,
+ "learning_rate": 0.0006,
+ "loss": 4.781791687011719,
+ "step": 1193
+ },
+ {
+ "epoch": 16.586899563318777,
+ "grad_norm": 0.01728292554616928,
+ "learning_rate": 0.0006,
+ "loss": 4.515257835388184,
+ "step": 1194
+ },
+ {
+ "epoch": 16.600873362445416,
+ "grad_norm": 0.01699589006602764,
+ "learning_rate": 0.0006,
+ "loss": 4.5996904373168945,
+ "step": 1195
+ },
+ {
+ "epoch": 16.61484716157205,
+ "grad_norm": 0.015905970707535744,
+ "learning_rate": 0.0006,
+ "loss": 4.601301193237305,
+ "step": 1196
+ },
+ {
+ "epoch": 16.62882096069869,
+ "grad_norm": 0.01338571310043335,
+ "learning_rate": 0.0006,
+ "loss": 4.557708740234375,
+ "step": 1197
+ },
+ {
+ "epoch": 16.64279475982533,
+ "grad_norm": 0.015298855490982533,
+ "learning_rate": 0.0006,
+ "loss": 4.542287349700928,
+ "step": 1198
+ },
+ {
+ "epoch": 16.656768558951963,
+ "grad_norm": 0.014753367751836777,
+ "learning_rate": 0.0006,
+ "loss": 4.595787048339844,
+ "step": 1199
+ },
+ {
+ "epoch": 16.670742358078602,
+ "grad_norm": 0.015527483075857162,
+ "learning_rate": 0.0006,
+ "loss": 4.607529163360596,
+ "step": 1200
+ },
+ {
+ "epoch": 16.68471615720524,
+ "grad_norm": 0.015042162500321865,
+ "learning_rate": 0.0006,
+ "loss": 4.686626434326172,
+ "step": 1201
+ },
+ {
+ "epoch": 16.69868995633188,
+ "grad_norm": 0.0159393772482872,
+ "learning_rate": 0.0006,
+ "loss": 4.47656774520874,
+ "step": 1202
+ },
+ {
+ "epoch": 16.712663755458514,
+ "grad_norm": 0.01532787922769785,
+ "learning_rate": 0.0006,
+ "loss": 4.55678653717041,
+ "step": 1203
+ },
+ {
+ "epoch": 16.726637554585153,
+ "grad_norm": 0.016332164406776428,
+ "learning_rate": 0.0006,
+ "loss": 4.51441764831543,
+ "step": 1204
+ },
+ {
+ "epoch": 16.74061135371179,
+ "grad_norm": 0.018901726230978966,
+ "learning_rate": 0.0006,
+ "loss": 4.57227897644043,
+ "step": 1205
+ },
+ {
+ "epoch": 16.754585152838427,
+ "grad_norm": 0.018589433282613754,
+ "learning_rate": 0.0006,
+ "loss": 4.550070762634277,
+ "step": 1206
+ },
+ {
+ "epoch": 16.768558951965066,
+ "grad_norm": 0.01634034886956215,
+ "learning_rate": 0.0006,
+ "loss": 4.586910724639893,
+ "step": 1207
+ },
+ {
+ "epoch": 16.782532751091704,
+ "grad_norm": 0.017599433660507202,
+ "learning_rate": 0.0006,
+ "loss": 4.5891923904418945,
+ "step": 1208
+ },
+ {
+ "epoch": 16.79650655021834,
+ "grad_norm": 0.01878446340560913,
+ "learning_rate": 0.0006,
+ "loss": 4.57227087020874,
+ "step": 1209
+ },
+ {
+ "epoch": 16.810480349344978,
+ "grad_norm": 0.01803469844162464,
+ "learning_rate": 0.0006,
+ "loss": 4.583845138549805,
+ "step": 1210
+ },
+ {
+ "epoch": 16.824454148471617,
+ "grad_norm": 0.018505144864320755,
+ "learning_rate": 0.0006,
+ "loss": 4.561151504516602,
+ "step": 1211
+ },
+ {
+ "epoch": 16.83842794759825,
+ "grad_norm": 0.01920081488788128,
+ "learning_rate": 0.0006,
+ "loss": 4.506834030151367,
+ "step": 1212
+ },
+ {
+ "epoch": 16.85240174672489,
+ "grad_norm": 0.0177712831646204,
+ "learning_rate": 0.0006,
+ "loss": 4.557195663452148,
+ "step": 1213
+ },
+ {
+ "epoch": 16.86637554585153,
+ "grad_norm": 0.019113918766379356,
+ "learning_rate": 0.0006,
+ "loss": 4.651227951049805,
+ "step": 1214
+ },
+ {
+ "epoch": 16.880349344978168,
+ "grad_norm": 0.02012539468705654,
+ "learning_rate": 0.0006,
+ "loss": 4.4952239990234375,
+ "step": 1215
+ },
+ {
+ "epoch": 16.894323144104803,
+ "grad_norm": 0.019993646070361137,
+ "learning_rate": 0.0006,
+ "loss": 4.517433166503906,
+ "step": 1216
+ },
+ {
+ "epoch": 16.90829694323144,
+ "grad_norm": 0.020041441544890404,
+ "learning_rate": 0.0006,
+ "loss": 4.489407539367676,
+ "step": 1217
+ },
+ {
+ "epoch": 16.92227074235808,
+ "grad_norm": 0.0183611661195755,
+ "learning_rate": 0.0006,
+ "loss": 4.5435638427734375,
+ "step": 1218
+ },
+ {
+ "epoch": 16.936244541484715,
+ "grad_norm": 0.019138429313898087,
+ "learning_rate": 0.0006,
+ "loss": 4.545949459075928,
+ "step": 1219
+ },
+ {
+ "epoch": 16.950218340611354,
+ "grad_norm": 0.018554266542196274,
+ "learning_rate": 0.0006,
+ "loss": 4.579620361328125,
+ "step": 1220
+ },
+ {
+ "epoch": 16.964192139737992,
+ "grad_norm": 0.0193545650690794,
+ "learning_rate": 0.0006,
+ "loss": 4.588255405426025,
+ "step": 1221
+ },
+ {
+ "epoch": 16.978165938864628,
+ "grad_norm": 0.019389165565371513,
+ "learning_rate": 0.0006,
+ "loss": 4.520698547363281,
+ "step": 1222
+ },
+ {
+ "epoch": 16.992139737991266,
+ "grad_norm": 0.016601379960775375,
+ "learning_rate": 0.0006,
+ "loss": 4.4809064865112305,
+ "step": 1223
+ },
+ {
+ "epoch": 17.0,
+ "grad_norm": 0.01771303080022335,
+ "learning_rate": 0.0006,
+ "loss": 4.721122741699219,
+ "step": 1224
+ },
+ {
+ "epoch": 17.0,
+ "eval_loss": 4.7903666496276855,
+ "eval_runtime": 56.9826,
+ "eval_samples_per_second": 42.855,
+ "eval_steps_per_second": 1.351,
+ "step": 1224
+ },
+ {
+ "epoch": 17.01397379912664,
+ "grad_norm": 0.017720289528369904,
+ "learning_rate": 0.0006,
+ "loss": 4.523834705352783,
+ "step": 1225
+ },
+ {
+ "epoch": 17.027947598253274,
+ "grad_norm": 0.020137161016464233,
+ "learning_rate": 0.0006,
+ "loss": 4.574061870574951,
+ "step": 1226
+ },
+ {
+ "epoch": 17.041921397379912,
+ "grad_norm": 0.02002314105629921,
+ "learning_rate": 0.0006,
+ "loss": 4.512407302856445,
+ "step": 1227
+ },
+ {
+ "epoch": 17.05589519650655,
+ "grad_norm": 0.019628183916211128,
+ "learning_rate": 0.0006,
+ "loss": 4.502533435821533,
+ "step": 1228
+ },
+ {
+ "epoch": 17.069868995633186,
+ "grad_norm": 0.019808458164334297,
+ "learning_rate": 0.0006,
+ "loss": 4.463156700134277,
+ "step": 1229
+ },
+ {
+ "epoch": 17.083842794759825,
+ "grad_norm": 0.021250786259770393,
+ "learning_rate": 0.0006,
+ "loss": 4.622682094573975,
+ "step": 1230
+ },
+ {
+ "epoch": 17.097816593886463,
+ "grad_norm": 0.022031694650650024,
+ "learning_rate": 0.0006,
+ "loss": 4.451882839202881,
+ "step": 1231
+ },
+ {
+ "epoch": 17.111790393013102,
+ "grad_norm": 0.01898987777531147,
+ "learning_rate": 0.0006,
+ "loss": 4.529331684112549,
+ "step": 1232
+ },
+ {
+ "epoch": 17.125764192139737,
+ "grad_norm": 0.018749834969639778,
+ "learning_rate": 0.0006,
+ "loss": 4.435815334320068,
+ "step": 1233
+ },
+ {
+ "epoch": 17.139737991266376,
+ "grad_norm": 0.01896674931049347,
+ "learning_rate": 0.0006,
+ "loss": 4.484029293060303,
+ "step": 1234
+ },
+ {
+ "epoch": 17.153711790393015,
+ "grad_norm": 0.018286997452378273,
+ "learning_rate": 0.0006,
+ "loss": 4.523510932922363,
+ "step": 1235
+ },
+ {
+ "epoch": 17.16768558951965,
+ "grad_norm": 0.017273059114813805,
+ "learning_rate": 0.0006,
+ "loss": 4.451767444610596,
+ "step": 1236
+ },
+ {
+ "epoch": 17.18165938864629,
+ "grad_norm": 0.016499150544404984,
+ "learning_rate": 0.0006,
+ "loss": 4.5568647384643555,
+ "step": 1237
+ },
+ {
+ "epoch": 17.195633187772927,
+ "grad_norm": 0.015934469178318977,
+ "learning_rate": 0.0006,
+ "loss": 4.4639997482299805,
+ "step": 1238
+ },
+ {
+ "epoch": 17.209606986899562,
+ "grad_norm": 0.01714209094643593,
+ "learning_rate": 0.0006,
+ "loss": 4.547614097595215,
+ "step": 1239
+ },
+ {
+ "epoch": 17.2235807860262,
+ "grad_norm": 0.019167495891451836,
+ "learning_rate": 0.0006,
+ "loss": 4.571097373962402,
+ "step": 1240
+ },
+ {
+ "epoch": 17.23755458515284,
+ "grad_norm": 0.018276996910572052,
+ "learning_rate": 0.0006,
+ "loss": 4.489764213562012,
+ "step": 1241
+ },
+ {
+ "epoch": 17.251528384279474,
+ "grad_norm": 0.021296866238117218,
+ "learning_rate": 0.0006,
+ "loss": 4.526913642883301,
+ "step": 1242
+ },
+ {
+ "epoch": 17.265502183406113,
+ "grad_norm": 0.022606629878282547,
+ "learning_rate": 0.0006,
+ "loss": 4.521624565124512,
+ "step": 1243
+ },
+ {
+ "epoch": 17.27947598253275,
+ "grad_norm": 0.021204711869359016,
+ "learning_rate": 0.0006,
+ "loss": 4.500864028930664,
+ "step": 1244
+ },
+ {
+ "epoch": 17.29344978165939,
+ "grad_norm": 0.01988849975168705,
+ "learning_rate": 0.0006,
+ "loss": 4.45162296295166,
+ "step": 1245
+ },
+ {
+ "epoch": 17.307423580786025,
+ "grad_norm": 0.020268384367227554,
+ "learning_rate": 0.0006,
+ "loss": 4.536198616027832,
+ "step": 1246
+ },
+ {
+ "epoch": 17.321397379912664,
+ "grad_norm": 0.020185677334666252,
+ "learning_rate": 0.0006,
+ "loss": 4.440262794494629,
+ "step": 1247
+ },
+ {
+ "epoch": 17.335371179039303,
+ "grad_norm": 0.020106853917241096,
+ "learning_rate": 0.0006,
+ "loss": 4.475907325744629,
+ "step": 1248
+ },
+ {
+ "epoch": 17.349344978165938,
+ "grad_norm": 0.018883297219872475,
+ "learning_rate": 0.0006,
+ "loss": 4.640661716461182,
+ "step": 1249
+ },
+ {
+ "epoch": 17.363318777292577,
+ "grad_norm": 0.01852620206773281,
+ "learning_rate": 0.0006,
+ "loss": 4.578243255615234,
+ "step": 1250
+ },
+ {
+ "epoch": 17.377292576419215,
+ "grad_norm": 0.01886451430618763,
+ "learning_rate": 0.0006,
+ "loss": 4.546026229858398,
+ "step": 1251
+ },
+ {
+ "epoch": 17.39126637554585,
+ "grad_norm": 0.01777452416718006,
+ "learning_rate": 0.0006,
+ "loss": 4.521542549133301,
+ "step": 1252
+ },
+ {
+ "epoch": 17.40524017467249,
+ "grad_norm": 0.016626447439193726,
+ "learning_rate": 0.0006,
+ "loss": 4.52629280090332,
+ "step": 1253
+ },
+ {
+ "epoch": 17.419213973799128,
+ "grad_norm": 0.01599477231502533,
+ "learning_rate": 0.0006,
+ "loss": 4.4796037673950195,
+ "step": 1254
+ },
+ {
+ "epoch": 17.433187772925763,
+ "grad_norm": 0.017021458595991135,
+ "learning_rate": 0.0006,
+ "loss": 4.468813419342041,
+ "step": 1255
+ },
+ {
+ "epoch": 17.4471615720524,
+ "grad_norm": 0.016925333067774773,
+ "learning_rate": 0.0006,
+ "loss": 4.584898948669434,
+ "step": 1256
+ },
+ {
+ "epoch": 17.46113537117904,
+ "grad_norm": 0.01714998669922352,
+ "learning_rate": 0.0006,
+ "loss": 4.523544788360596,
+ "step": 1257
+ },
+ {
+ "epoch": 17.475109170305675,
+ "grad_norm": 0.01802288182079792,
+ "learning_rate": 0.0006,
+ "loss": 4.595581531524658,
+ "step": 1258
+ },
+ {
+ "epoch": 17.489082969432314,
+ "grad_norm": 0.0180289875715971,
+ "learning_rate": 0.0006,
+ "loss": 4.394440650939941,
+ "step": 1259
+ },
+ {
+ "epoch": 17.503056768558952,
+ "grad_norm": 0.018504245206713676,
+ "learning_rate": 0.0006,
+ "loss": 4.4794416427612305,
+ "step": 1260
+ },
+ {
+ "epoch": 17.51703056768559,
+ "grad_norm": 0.019973233342170715,
+ "learning_rate": 0.0006,
+ "loss": 4.4241437911987305,
+ "step": 1261
+ },
+ {
+ "epoch": 17.531004366812226,
+ "grad_norm": 0.020063083618879318,
+ "learning_rate": 0.0006,
+ "loss": 4.48068380355835,
+ "step": 1262
+ },
+ {
+ "epoch": 17.544978165938865,
+ "grad_norm": 0.02066640742123127,
+ "learning_rate": 0.0006,
+ "loss": 4.628993034362793,
+ "step": 1263
+ },
+ {
+ "epoch": 17.558951965065503,
+ "grad_norm": 0.02206416428089142,
+ "learning_rate": 0.0006,
+ "loss": 4.564790725708008,
+ "step": 1264
+ },
+ {
+ "epoch": 17.57292576419214,
+ "grad_norm": 0.020640335977077484,
+ "learning_rate": 0.0006,
+ "loss": 4.578705787658691,
+ "step": 1265
+ },
+ {
+ "epoch": 17.586899563318777,
+ "grad_norm": 0.019093405455350876,
+ "learning_rate": 0.0006,
+ "loss": 4.5023603439331055,
+ "step": 1266
+ },
+ {
+ "epoch": 17.600873362445416,
+ "grad_norm": 0.019459193572402,
+ "learning_rate": 0.0006,
+ "loss": 4.594531059265137,
+ "step": 1267
+ },
+ {
+ "epoch": 17.61484716157205,
+ "grad_norm": 0.02021171525120735,
+ "learning_rate": 0.0006,
+ "loss": 4.770299911499023,
+ "step": 1268
+ },
+ {
+ "epoch": 17.62882096069869,
+ "grad_norm": 0.021161191165447235,
+ "learning_rate": 0.0006,
+ "loss": 4.5380401611328125,
+ "step": 1269
+ },
+ {
+ "epoch": 17.64279475982533,
+ "grad_norm": 0.021013403311371803,
+ "learning_rate": 0.0006,
+ "loss": 4.455524921417236,
+ "step": 1270
+ },
+ {
+ "epoch": 17.656768558951963,
+ "grad_norm": 0.020336592569947243,
+ "learning_rate": 0.0006,
+ "loss": 4.533672332763672,
+ "step": 1271
+ },
+ {
+ "epoch": 17.670742358078602,
+ "grad_norm": 0.020159434527158737,
+ "learning_rate": 0.0006,
+ "loss": 4.596151351928711,
+ "step": 1272
+ },
+ {
+ "epoch": 17.68471615720524,
+ "grad_norm": 0.020432960242033005,
+ "learning_rate": 0.0006,
+ "loss": 4.55531644821167,
+ "step": 1273
+ },
+ {
+ "epoch": 17.69868995633188,
+ "grad_norm": 0.025466229766607285,
+ "learning_rate": 0.0006,
+ "loss": 4.566912651062012,
+ "step": 1274
+ },
+ {
+ "epoch": 17.712663755458514,
+ "grad_norm": 0.026537558063864708,
+ "learning_rate": 0.0006,
+ "loss": 4.526003837585449,
+ "step": 1275
+ },
+ {
+ "epoch": 17.726637554585153,
+ "grad_norm": 0.024397607892751694,
+ "learning_rate": 0.0006,
+ "loss": 4.541755676269531,
+ "step": 1276
+ },
+ {
+ "epoch": 17.74061135371179,
+ "grad_norm": 0.022062312811613083,
+ "learning_rate": 0.0006,
+ "loss": 4.6532793045043945,
+ "step": 1277
+ },
+ {
+ "epoch": 17.754585152838427,
+ "grad_norm": 0.02204282023012638,
+ "learning_rate": 0.0006,
+ "loss": 4.582825660705566,
+ "step": 1278
+ },
+ {
+ "epoch": 17.768558951965066,
+ "grad_norm": 0.020844416692852974,
+ "learning_rate": 0.0006,
+ "loss": 4.522336483001709,
+ "step": 1279
+ },
+ {
+ "epoch": 17.782532751091704,
+ "grad_norm": 0.019815703853964806,
+ "learning_rate": 0.0006,
+ "loss": 4.607398509979248,
+ "step": 1280
+ },
+ {
+ "epoch": 17.79650655021834,
+ "grad_norm": 0.016676954925060272,
+ "learning_rate": 0.0006,
+ "loss": 4.446206092834473,
+ "step": 1281
+ },
+ {
+ "epoch": 17.810480349344978,
+ "grad_norm": 0.0166452769190073,
+ "learning_rate": 0.0006,
+ "loss": 4.505931854248047,
+ "step": 1282
+ },
+ {
+ "epoch": 17.824454148471617,
+ "grad_norm": 0.015928907319903374,
+ "learning_rate": 0.0006,
+ "loss": 4.455267429351807,
+ "step": 1283
+ },
+ {
+ "epoch": 17.83842794759825,
+ "grad_norm": 0.015219560824334621,
+ "learning_rate": 0.0006,
+ "loss": 4.5623674392700195,
+ "step": 1284
+ },
+ {
+ "epoch": 17.85240174672489,
+ "grad_norm": 0.015125355683267117,
+ "learning_rate": 0.0006,
+ "loss": 4.554104804992676,
+ "step": 1285
+ },
+ {
+ "epoch": 17.86637554585153,
+ "grad_norm": 0.015612718649208546,
+ "learning_rate": 0.0006,
+ "loss": 4.484847068786621,
+ "step": 1286
+ },
+ {
+ "epoch": 17.880349344978168,
+ "grad_norm": 0.015215662308037281,
+ "learning_rate": 0.0006,
+ "loss": 4.4802093505859375,
+ "step": 1287
+ },
+ {
+ "epoch": 17.894323144104803,
+ "grad_norm": 0.01511684525758028,
+ "learning_rate": 0.0006,
+ "loss": 4.537298679351807,
+ "step": 1288
+ },
+ {
+ "epoch": 17.90829694323144,
+ "grad_norm": 0.013841322623193264,
+ "learning_rate": 0.0006,
+ "loss": 4.427488327026367,
+ "step": 1289
+ },
+ {
+ "epoch": 17.92227074235808,
+ "grad_norm": 0.015115964226424694,
+ "learning_rate": 0.0006,
+ "loss": 4.4547119140625,
+ "step": 1290
+ },
+ {
+ "epoch": 17.936244541484715,
+ "grad_norm": 0.016411006450653076,
+ "learning_rate": 0.0006,
+ "loss": 4.52994966506958,
+ "step": 1291
+ },
+ {
+ "epoch": 17.950218340611354,
+ "grad_norm": 0.017313152551651,
+ "learning_rate": 0.0006,
+ "loss": 4.5337653160095215,
+ "step": 1292
+ },
+ {
+ "epoch": 17.964192139737992,
+ "grad_norm": 0.016559796407818794,
+ "learning_rate": 0.0006,
+ "loss": 4.5277910232543945,
+ "step": 1293
+ },
+ {
+ "epoch": 17.978165938864628,
+ "grad_norm": 0.015854761004447937,
+ "learning_rate": 0.0006,
+ "loss": 4.6334309577941895,
+ "step": 1294
+ },
+ {
+ "epoch": 17.992139737991266,
+ "grad_norm": 0.015658238902688026,
+ "learning_rate": 0.0006,
+ "loss": 4.613372802734375,
+ "step": 1295
+ },
+ {
+ "epoch": 18.0,
+ "grad_norm": 0.0169155802577734,
+ "learning_rate": 0.0006,
+ "loss": 4.527356147766113,
+ "step": 1296
+ },
+ {
+ "epoch": 18.0,
+ "eval_loss": 4.764543056488037,
+ "eval_runtime": 56.5998,
+ "eval_samples_per_second": 43.145,
+ "eval_steps_per_second": 1.36,
+ "step": 1296
+ },
+ {
+ "epoch": 18.01397379912664,
+ "grad_norm": 0.015348963439464569,
+ "learning_rate": 0.0006,
+ "loss": 4.466423988342285,
+ "step": 1297
+ },
+ {
+ "epoch": 18.027947598253274,
+ "grad_norm": 0.01768588088452816,
+ "learning_rate": 0.0006,
+ "loss": 4.545933246612549,
+ "step": 1298
+ },
+ {
+ "epoch": 18.041921397379912,
+ "grad_norm": 0.017381226643919945,
+ "learning_rate": 0.0006,
+ "loss": 4.494060039520264,
+ "step": 1299
+ },
+ {
+ "epoch": 18.05589519650655,
+ "grad_norm": 0.019002335146069527,
+ "learning_rate": 0.0006,
+ "loss": 4.395200729370117,
+ "step": 1300
+ },
+ {
+ "epoch": 18.069868995633186,
+ "grad_norm": 0.01827932707965374,
+ "learning_rate": 0.0006,
+ "loss": 4.424848556518555,
+ "step": 1301
+ },
+ {
+ "epoch": 18.083842794759825,
+ "grad_norm": 0.016383031383156776,
+ "learning_rate": 0.0006,
+ "loss": 4.494722366333008,
+ "step": 1302
+ },
+ {
+ "epoch": 18.097816593886463,
+ "grad_norm": 0.016265157610177994,
+ "learning_rate": 0.0006,
+ "loss": 4.416522979736328,
+ "step": 1303
+ },
+ {
+ "epoch": 18.111790393013102,
+ "grad_norm": 0.01702386513352394,
+ "learning_rate": 0.0006,
+ "loss": 4.600103378295898,
+ "step": 1304
+ },
+ {
+ "epoch": 18.125764192139737,
+ "grad_norm": 0.018495498225092888,
+ "learning_rate": 0.0006,
+ "loss": 4.588401794433594,
+ "step": 1305
+ },
+ {
+ "epoch": 18.139737991266376,
+ "grad_norm": 0.017921684309840202,
+ "learning_rate": 0.0006,
+ "loss": 4.478075981140137,
+ "step": 1306
+ },
+ {
+ "epoch": 18.153711790393015,
+ "grad_norm": 0.01801278069615364,
+ "learning_rate": 0.0006,
+ "loss": 4.413186550140381,
+ "step": 1307
+ },
+ {
+ "epoch": 18.16768558951965,
+ "grad_norm": 0.017211275175213814,
+ "learning_rate": 0.0006,
+ "loss": 4.461341857910156,
+ "step": 1308
+ },
+ {
+ "epoch": 18.18165938864629,
+ "grad_norm": 0.017589328810572624,
+ "learning_rate": 0.0006,
+ "loss": 4.546969413757324,
+ "step": 1309
+ },
+ {
+ "epoch": 18.195633187772927,
+ "grad_norm": 0.01797698810696602,
+ "learning_rate": 0.0006,
+ "loss": 4.530797004699707,
+ "step": 1310
+ },
+ {
+ "epoch": 18.209606986899562,
+ "grad_norm": 0.020408619195222855,
+ "learning_rate": 0.0006,
+ "loss": 4.62999153137207,
+ "step": 1311
+ },
+ {
+ "epoch": 18.2235807860262,
+ "grad_norm": 0.01994038000702858,
+ "learning_rate": 0.0006,
+ "loss": 4.586811065673828,
+ "step": 1312
+ },
+ {
+ "epoch": 18.23755458515284,
+ "grad_norm": 0.019381340593099594,
+ "learning_rate": 0.0006,
+ "loss": 4.4769368171691895,
+ "step": 1313
+ },
+ {
+ "epoch": 18.251528384279474,
+ "grad_norm": 0.020484555512666702,
+ "learning_rate": 0.0006,
+ "loss": 4.524302959442139,
+ "step": 1314
+ },
+ {
+ "epoch": 18.265502183406113,
+ "grad_norm": 0.02426416426897049,
+ "learning_rate": 0.0006,
+ "loss": 4.561705589294434,
+ "step": 1315
+ },
+ {
+ "epoch": 18.27947598253275,
+ "grad_norm": 0.024748101830482483,
+ "learning_rate": 0.0006,
+ "loss": 4.4654316902160645,
+ "step": 1316
+ },
+ {
+ "epoch": 18.29344978165939,
+ "grad_norm": 0.020703328773379326,
+ "learning_rate": 0.0006,
+ "loss": 4.53033447265625,
+ "step": 1317
+ },
+ {
+ "epoch": 18.307423580786025,
+ "grad_norm": 0.01765458658337593,
+ "learning_rate": 0.0006,
+ "loss": 4.5154266357421875,
+ "step": 1318
+ },
+ {
+ "epoch": 18.321397379912664,
+ "grad_norm": 0.018646420910954475,
+ "learning_rate": 0.0006,
+ "loss": 4.5224714279174805,
+ "step": 1319
+ },
+ {
+ "epoch": 18.335371179039303,
+ "grad_norm": 0.01907474733889103,
+ "learning_rate": 0.0006,
+ "loss": 4.5055437088012695,
+ "step": 1320
+ },
+ {
+ "epoch": 18.349344978165938,
+ "grad_norm": 0.018179846927523613,
+ "learning_rate": 0.0006,
+ "loss": 4.462161064147949,
+ "step": 1321
+ },
+ {
+ "epoch": 18.363318777292577,
+ "grad_norm": 0.016697878018021584,
+ "learning_rate": 0.0006,
+ "loss": 4.473718643188477,
+ "step": 1322
+ },
+ {
+ "epoch": 18.377292576419215,
+ "grad_norm": 0.017504019662737846,
+ "learning_rate": 0.0006,
+ "loss": 4.474348068237305,
+ "step": 1323
+ },
+ {
+ "epoch": 18.39126637554585,
+ "grad_norm": 0.018889913335442543,
+ "learning_rate": 0.0006,
+ "loss": 4.4531636238098145,
+ "step": 1324
+ },
+ {
+ "epoch": 18.40524017467249,
+ "grad_norm": 0.019167251884937286,
+ "learning_rate": 0.0006,
+ "loss": 4.4278364181518555,
+ "step": 1325
+ },
+ {
+ "epoch": 18.419213973799128,
+ "grad_norm": 0.015895741060376167,
+ "learning_rate": 0.0006,
+ "loss": 4.474592208862305,
+ "step": 1326
+ },
+ {
+ "epoch": 18.433187772925763,
+ "grad_norm": 0.01531882956624031,
+ "learning_rate": 0.0006,
+ "loss": 4.403757572174072,
+ "step": 1327
+ },
+ {
+ "epoch": 18.4471615720524,
+ "grad_norm": 0.01565849967300892,
+ "learning_rate": 0.0006,
+ "loss": 4.507664680480957,
+ "step": 1328
+ },
+ {
+ "epoch": 18.46113537117904,
+ "grad_norm": 0.01602315716445446,
+ "learning_rate": 0.0006,
+ "loss": 4.5648908615112305,
+ "step": 1329
+ },
+ {
+ "epoch": 18.475109170305675,
+ "grad_norm": 0.016467366367578506,
+ "learning_rate": 0.0006,
+ "loss": 4.428762912750244,
+ "step": 1330
+ },
+ {
+ "epoch": 18.489082969432314,
+ "grad_norm": 0.01604127697646618,
+ "learning_rate": 0.0006,
+ "loss": 4.527771472930908,
+ "step": 1331
+ },
+ {
+ "epoch": 18.503056768558952,
+ "grad_norm": 0.018825586885213852,
+ "learning_rate": 0.0006,
+ "loss": 4.485386848449707,
+ "step": 1332
+ },
+ {
+ "epoch": 18.51703056768559,
+ "grad_norm": 0.019258007407188416,
+ "learning_rate": 0.0006,
+ "loss": 4.414972305297852,
+ "step": 1333
+ },
+ {
+ "epoch": 18.531004366812226,
+ "grad_norm": 0.018192021176218987,
+ "learning_rate": 0.0006,
+ "loss": 4.412166595458984,
+ "step": 1334
+ },
+ {
+ "epoch": 18.544978165938865,
+ "grad_norm": 0.016120698302984238,
+ "learning_rate": 0.0006,
+ "loss": 4.484317779541016,
+ "step": 1335
+ },
+ {
+ "epoch": 18.558951965065503,
+ "grad_norm": 0.016496622934937477,
+ "learning_rate": 0.0006,
+ "loss": 4.514986515045166,
+ "step": 1336
+ },
+ {
+ "epoch": 18.57292576419214,
+ "grad_norm": 0.01795116998255253,
+ "learning_rate": 0.0006,
+ "loss": 4.41473388671875,
+ "step": 1337
+ },
+ {
+ "epoch": 18.586899563318777,
+ "grad_norm": 0.019062276929616928,
+ "learning_rate": 0.0006,
+ "loss": 4.601996898651123,
+ "step": 1338
+ },
+ {
+ "epoch": 18.600873362445416,
+ "grad_norm": 0.017174631357192993,
+ "learning_rate": 0.0006,
+ "loss": 4.571218490600586,
+ "step": 1339
+ },
+ {
+ "epoch": 18.61484716157205,
+ "grad_norm": 0.014015606604516506,
+ "learning_rate": 0.0006,
+ "loss": 4.526606559753418,
+ "step": 1340
+ },
+ {
+ "epoch": 18.62882096069869,
+ "grad_norm": 0.015358510427176952,
+ "learning_rate": 0.0006,
+ "loss": 4.604286193847656,
+ "step": 1341
+ },
+ {
+ "epoch": 18.64279475982533,
+ "grad_norm": 0.01619391329586506,
+ "learning_rate": 0.0006,
+ "loss": 4.560883522033691,
+ "step": 1342
+ },
+ {
+ "epoch": 18.656768558951963,
+ "grad_norm": 0.015972964465618134,
+ "learning_rate": 0.0006,
+ "loss": 4.479755401611328,
+ "step": 1343
+ },
+ {
+ "epoch": 18.670742358078602,
+ "grad_norm": 0.016861025243997574,
+ "learning_rate": 0.0006,
+ "loss": 4.474471092224121,
+ "step": 1344
+ },
+ {
+ "epoch": 18.68471615720524,
+ "grad_norm": 0.01822078227996826,
+ "learning_rate": 0.0006,
+ "loss": 4.509489059448242,
+ "step": 1345
+ },
+ {
+ "epoch": 18.69868995633188,
+ "grad_norm": 0.019415007904171944,
+ "learning_rate": 0.0006,
+ "loss": 4.508941173553467,
+ "step": 1346
+ },
+ {
+ "epoch": 18.712663755458514,
+ "grad_norm": 0.019856126978993416,
+ "learning_rate": 0.0006,
+ "loss": 4.446670055389404,
+ "step": 1347
+ },
+ {
+ "epoch": 18.726637554585153,
+ "grad_norm": 0.01842520199716091,
+ "learning_rate": 0.0006,
+ "loss": 4.524422645568848,
+ "step": 1348
+ },
+ {
+ "epoch": 18.74061135371179,
+ "grad_norm": 0.017196480184793472,
+ "learning_rate": 0.0006,
+ "loss": 4.498466968536377,
+ "step": 1349
+ },
+ {
+ "epoch": 18.754585152838427,
+ "grad_norm": 0.01823010854423046,
+ "learning_rate": 0.0006,
+ "loss": 4.534684181213379,
+ "step": 1350
+ },
+ {
+ "epoch": 18.768558951965066,
+ "grad_norm": 0.020398065447807312,
+ "learning_rate": 0.0006,
+ "loss": 4.456494331359863,
+ "step": 1351
+ },
+ {
+ "epoch": 18.782532751091704,
+ "grad_norm": 0.022992262616753578,
+ "learning_rate": 0.0006,
+ "loss": 4.476374626159668,
+ "step": 1352
+ },
+ {
+ "epoch": 18.79650655021834,
+ "grad_norm": 0.02235420234501362,
+ "learning_rate": 0.0006,
+ "loss": 4.587776184082031,
+ "step": 1353
+ },
+ {
+ "epoch": 18.810480349344978,
+ "grad_norm": 0.02093626745045185,
+ "learning_rate": 0.0006,
+ "loss": 4.424172401428223,
+ "step": 1354
+ },
+ {
+ "epoch": 18.824454148471617,
+ "grad_norm": 0.01997302658855915,
+ "learning_rate": 0.0006,
+ "loss": 4.448066711425781,
+ "step": 1355
+ },
+ {
+ "epoch": 18.83842794759825,
+ "grad_norm": 0.018265459686517715,
+ "learning_rate": 0.0006,
+ "loss": 4.436610698699951,
+ "step": 1356
+ },
+ {
+ "epoch": 18.85240174672489,
+ "grad_norm": 0.01673833839595318,
+ "learning_rate": 0.0006,
+ "loss": 4.5502214431762695,
+ "step": 1357
+ },
+ {
+ "epoch": 18.86637554585153,
+ "grad_norm": 0.017908833920955658,
+ "learning_rate": 0.0006,
+ "loss": 4.550530433654785,
+ "step": 1358
+ },
+ {
+ "epoch": 18.880349344978168,
+ "grad_norm": 0.01796654611825943,
+ "learning_rate": 0.0006,
+ "loss": 4.531463146209717,
+ "step": 1359
+ },
+ {
+ "epoch": 18.894323144104803,
+ "grad_norm": 0.01646183803677559,
+ "learning_rate": 0.0006,
+ "loss": 4.498582363128662,
+ "step": 1360
+ },
+ {
+ "epoch": 18.90829694323144,
+ "grad_norm": 0.015430327504873276,
+ "learning_rate": 0.0006,
+ "loss": 4.52571439743042,
+ "step": 1361
+ },
+ {
+ "epoch": 18.92227074235808,
+ "grad_norm": 0.016527952626347542,
+ "learning_rate": 0.0006,
+ "loss": 4.547013759613037,
+ "step": 1362
+ },
+ {
+ "epoch": 18.936244541484715,
+ "grad_norm": 0.016263660043478012,
+ "learning_rate": 0.0006,
+ "loss": 4.468754291534424,
+ "step": 1363
+ },
+ {
+ "epoch": 18.950218340611354,
+ "grad_norm": 0.016347210854291916,
+ "learning_rate": 0.0006,
+ "loss": 4.450741767883301,
+ "step": 1364
+ },
+ {
+ "epoch": 18.964192139737992,
+ "grad_norm": 0.01701398566365242,
+ "learning_rate": 0.0006,
+ "loss": 4.534899711608887,
+ "step": 1365
+ },
+ {
+ "epoch": 18.978165938864628,
+ "grad_norm": 0.017693081870675087,
+ "learning_rate": 0.0006,
+ "loss": 4.460553169250488,
+ "step": 1366
+ },
+ {
+ "epoch": 18.992139737991266,
+ "grad_norm": 0.018746308982372284,
+ "learning_rate": 0.0006,
+ "loss": 4.538792610168457,
+ "step": 1367
+ },
+ {
+ "epoch": 19.0,
+ "grad_norm": 0.020099416375160217,
+ "learning_rate": 0.0006,
+ "loss": 4.449617385864258,
+ "step": 1368
+ },
+ {
+ "epoch": 19.0,
+ "eval_loss": 4.7756571769714355,
+ "eval_runtime": 57.1804,
+ "eval_samples_per_second": 42.707,
+ "eval_steps_per_second": 1.347,
+ "step": 1368
+ },
+ {
+ "epoch": 19.01397379912664,
+ "grad_norm": 0.018726933747529984,
+ "learning_rate": 0.0006,
+ "loss": 4.446059703826904,
+ "step": 1369
+ },
+ {
+ "epoch": 19.027947598253274,
+ "grad_norm": 0.01956087350845337,
+ "learning_rate": 0.0006,
+ "loss": 4.4520039558410645,
+ "step": 1370
+ },
+ {
+ "epoch": 19.041921397379912,
+ "grad_norm": 0.020092811435461044,
+ "learning_rate": 0.0006,
+ "loss": 4.543343544006348,
+ "step": 1371
+ },
+ {
+ "epoch": 19.05589519650655,
+ "grad_norm": 0.01843622885644436,
+ "learning_rate": 0.0006,
+ "loss": 4.535579204559326,
+ "step": 1372
+ },
+ {
+ "epoch": 19.069868995633186,
+ "grad_norm": 0.016989488154649734,
+ "learning_rate": 0.0006,
+ "loss": 4.52316951751709,
+ "step": 1373
+ },
+ {
+ "epoch": 19.083842794759825,
+ "grad_norm": 0.01805422455072403,
+ "learning_rate": 0.0006,
+ "loss": 4.5393266677856445,
+ "step": 1374
+ },
+ {
+ "epoch": 19.097816593886463,
+ "grad_norm": 0.01874629408121109,
+ "learning_rate": 0.0006,
+ "loss": 4.422059059143066,
+ "step": 1375
+ },
+ {
+ "epoch": 19.111790393013102,
+ "grad_norm": 0.01804221048951149,
+ "learning_rate": 0.0006,
+ "loss": 4.50508451461792,
+ "step": 1376
+ },
+ {
+ "epoch": 19.125764192139737,
+ "grad_norm": 0.01785474829375744,
+ "learning_rate": 0.0006,
+ "loss": 4.412599563598633,
+ "step": 1377
+ },
+ {
+ "epoch": 19.139737991266376,
+ "grad_norm": 0.01971791312098503,
+ "learning_rate": 0.0006,
+ "loss": 4.518318176269531,
+ "step": 1378
+ },
+ {
+ "epoch": 19.153711790393015,
+ "grad_norm": 0.02061633951961994,
+ "learning_rate": 0.0006,
+ "loss": 4.332393169403076,
+ "step": 1379
+ },
+ {
+ "epoch": 19.16768558951965,
+ "grad_norm": 0.020563429221510887,
+ "learning_rate": 0.0006,
+ "loss": 4.445863723754883,
+ "step": 1380
+ },
+ {
+ "epoch": 19.18165938864629,
+ "grad_norm": 0.019995713606476784,
+ "learning_rate": 0.0006,
+ "loss": 4.497312068939209,
+ "step": 1381
+ },
+ {
+ "epoch": 19.195633187772927,
+ "grad_norm": 0.019823070615530014,
+ "learning_rate": 0.0006,
+ "loss": 4.475085258483887,
+ "step": 1382
+ },
+ {
+ "epoch": 19.209606986899562,
+ "grad_norm": 0.019039355218410492,
+ "learning_rate": 0.0006,
+ "loss": 4.50432014465332,
+ "step": 1383
+ },
+ {
+ "epoch": 19.2235807860262,
+ "grad_norm": 0.017876973375678062,
+ "learning_rate": 0.0006,
+ "loss": 4.384788513183594,
+ "step": 1384
+ },
+ {
+ "epoch": 19.23755458515284,
+ "grad_norm": 0.018697699531912804,
+ "learning_rate": 0.0006,
+ "loss": 4.410778045654297,
+ "step": 1385
+ },
+ {
+ "epoch": 19.251528384279474,
+ "grad_norm": 0.018826806917786598,
+ "learning_rate": 0.0006,
+ "loss": 4.524034023284912,
+ "step": 1386
+ },
+ {
+ "epoch": 19.265502183406113,
+ "grad_norm": 0.01824226602911949,
+ "learning_rate": 0.0006,
+ "loss": 4.526393890380859,
+ "step": 1387
+ },
+ {
+ "epoch": 19.27947598253275,
+ "grad_norm": 0.01753208599984646,
+ "learning_rate": 0.0006,
+ "loss": 4.493703365325928,
+ "step": 1388
+ },
+ {
+ "epoch": 19.29344978165939,
+ "grad_norm": 0.01742432825267315,
+ "learning_rate": 0.0006,
+ "loss": 4.392831802368164,
+ "step": 1389
+ },
+ {
+ "epoch": 19.307423580786025,
+ "grad_norm": 0.017099356278777122,
+ "learning_rate": 0.0006,
+ "loss": 4.578668594360352,
+ "step": 1390
+ },
+ {
+ "epoch": 19.321397379912664,
+ "grad_norm": 0.017946023494005203,
+ "learning_rate": 0.0006,
+ "loss": 4.456718444824219,
+ "step": 1391
+ },
+ {
+ "epoch": 19.335371179039303,
+ "grad_norm": 0.017705943435430527,
+ "learning_rate": 0.0006,
+ "loss": 4.553682327270508,
+ "step": 1392
+ },
+ {
+ "epoch": 19.349344978165938,
+ "grad_norm": 0.016647523269057274,
+ "learning_rate": 0.0006,
+ "loss": 4.525016784667969,
+ "step": 1393
+ },
+ {
+ "epoch": 19.363318777292577,
+ "grad_norm": 0.015407416969537735,
+ "learning_rate": 0.0006,
+ "loss": 4.500231742858887,
+ "step": 1394
+ },
+ {
+ "epoch": 19.377292576419215,
+ "grad_norm": 0.017264019697904587,
+ "learning_rate": 0.0006,
+ "loss": 4.4615349769592285,
+ "step": 1395
+ },
+ {
+ "epoch": 19.39126637554585,
+ "grad_norm": 0.01871555484831333,
+ "learning_rate": 0.0006,
+ "loss": 4.281026840209961,
+ "step": 1396
+ },
+ {
+ "epoch": 19.40524017467249,
+ "grad_norm": 0.017361650243401527,
+ "learning_rate": 0.0006,
+ "loss": 4.479382514953613,
+ "step": 1397
+ },
+ {
+ "epoch": 19.419213973799128,
+ "grad_norm": 0.019781362265348434,
+ "learning_rate": 0.0006,
+ "loss": 4.402320861816406,
+ "step": 1398
+ },
+ {
+ "epoch": 19.433187772925763,
+ "grad_norm": 0.019043298438191414,
+ "learning_rate": 0.0006,
+ "loss": 4.476983070373535,
+ "step": 1399
+ },
+ {
+ "epoch": 19.4471615720524,
+ "grad_norm": 0.017137985676527023,
+ "learning_rate": 0.0006,
+ "loss": 4.418689250946045,
+ "step": 1400
+ },
+ {
+ "epoch": 19.46113537117904,
+ "grad_norm": 0.016481833532452583,
+ "learning_rate": 0.0006,
+ "loss": 4.579145431518555,
+ "step": 1401
+ },
+ {
+ "epoch": 19.475109170305675,
+ "grad_norm": 0.016507714986801147,
+ "learning_rate": 0.0006,
+ "loss": 4.519158363342285,
+ "step": 1402
+ },
+ {
+ "epoch": 19.489082969432314,
+ "grad_norm": 0.018092872574925423,
+ "learning_rate": 0.0006,
+ "loss": 4.513748645782471,
+ "step": 1403
+ },
+ {
+ "epoch": 19.503056768558952,
+ "grad_norm": 0.018153440207242966,
+ "learning_rate": 0.0006,
+ "loss": 4.645384311676025,
+ "step": 1404
+ },
+ {
+ "epoch": 19.51703056768559,
+ "grad_norm": 0.01657119207084179,
+ "learning_rate": 0.0006,
+ "loss": 4.399757385253906,
+ "step": 1405
+ },
+ {
+ "epoch": 19.531004366812226,
+ "grad_norm": 0.0156280305236578,
+ "learning_rate": 0.0006,
+ "loss": 4.467836380004883,
+ "step": 1406
+ },
+ {
+ "epoch": 19.544978165938865,
+ "grad_norm": 0.016838403418660164,
+ "learning_rate": 0.0006,
+ "loss": 4.529176235198975,
+ "step": 1407
+ },
+ {
+ "epoch": 19.558951965065503,
+ "grad_norm": 0.01743236929178238,
+ "learning_rate": 0.0006,
+ "loss": 4.440032482147217,
+ "step": 1408
+ },
+ {
+ "epoch": 19.57292576419214,
+ "grad_norm": 0.01553610060364008,
+ "learning_rate": 0.0006,
+ "loss": 4.425588607788086,
+ "step": 1409
+ },
+ {
+ "epoch": 19.586899563318777,
+ "grad_norm": 0.017259271815419197,
+ "learning_rate": 0.0006,
+ "loss": 4.405452728271484,
+ "step": 1410
+ },
+ {
+ "epoch": 19.600873362445416,
+ "grad_norm": 0.015568481758236885,
+ "learning_rate": 0.0006,
+ "loss": 4.444279193878174,
+ "step": 1411
+ },
+ {
+ "epoch": 19.61484716157205,
+ "grad_norm": 0.015331248752772808,
+ "learning_rate": 0.0006,
+ "loss": 4.451846122741699,
+ "step": 1412
+ },
+ {
+ "epoch": 19.62882096069869,
+ "grad_norm": 0.015667378902435303,
+ "learning_rate": 0.0006,
+ "loss": 4.533931255340576,
+ "step": 1413
+ },
+ {
+ "epoch": 19.64279475982533,
+ "grad_norm": 0.016177602112293243,
+ "learning_rate": 0.0006,
+ "loss": 4.467313289642334,
+ "step": 1414
+ },
+ {
+ "epoch": 19.656768558951963,
+ "grad_norm": 0.01733129471540451,
+ "learning_rate": 0.0006,
+ "loss": 4.496893882751465,
+ "step": 1415
+ },
+ {
+ "epoch": 19.670742358078602,
+ "grad_norm": 0.016539594158530235,
+ "learning_rate": 0.0006,
+ "loss": 4.401961326599121,
+ "step": 1416
+ },
+ {
+ "epoch": 19.68471615720524,
+ "grad_norm": 0.016058551147580147,
+ "learning_rate": 0.0006,
+ "loss": 4.503375053405762,
+ "step": 1417
+ },
+ {
+ "epoch": 19.69868995633188,
+ "grad_norm": 0.01715777814388275,
+ "learning_rate": 0.0006,
+ "loss": 4.3808369636535645,
+ "step": 1418
+ },
+ {
+ "epoch": 19.712663755458514,
+ "grad_norm": 0.01681119203567505,
+ "learning_rate": 0.0006,
+ "loss": 4.5469560623168945,
+ "step": 1419
+ },
+ {
+ "epoch": 19.726637554585153,
+ "grad_norm": 0.01634550653398037,
+ "learning_rate": 0.0006,
+ "loss": 4.328197956085205,
+ "step": 1420
+ },
+ {
+ "epoch": 19.74061135371179,
+ "grad_norm": 0.01549347210675478,
+ "learning_rate": 0.0006,
+ "loss": 4.534621238708496,
+ "step": 1421
+ },
+ {
+ "epoch": 19.754585152838427,
+ "grad_norm": 0.01654665172100067,
+ "learning_rate": 0.0006,
+ "loss": 4.4724225997924805,
+ "step": 1422
+ },
+ {
+ "epoch": 19.768558951965066,
+ "grad_norm": 0.017119025811553,
+ "learning_rate": 0.0006,
+ "loss": 4.3897247314453125,
+ "step": 1423
+ },
+ {
+ "epoch": 19.782532751091704,
+ "grad_norm": 0.01603223942220211,
+ "learning_rate": 0.0006,
+ "loss": 4.412731170654297,
+ "step": 1424
+ },
+ {
+ "epoch": 19.79650655021834,
+ "grad_norm": 0.016774067655205727,
+ "learning_rate": 0.0006,
+ "loss": 4.54707145690918,
+ "step": 1425
+ },
+ {
+ "epoch": 19.810480349344978,
+ "grad_norm": 0.016453025862574577,
+ "learning_rate": 0.0006,
+ "loss": 4.462501049041748,
+ "step": 1426
+ },
+ {
+ "epoch": 19.824454148471617,
+ "grad_norm": 0.01597629487514496,
+ "learning_rate": 0.0006,
+ "loss": 4.229758262634277,
+ "step": 1427
+ },
+ {
+ "epoch": 19.83842794759825,
+ "grad_norm": 0.017999034374952316,
+ "learning_rate": 0.0006,
+ "loss": 4.389263153076172,
+ "step": 1428
+ },
+ {
+ "epoch": 19.85240174672489,
+ "grad_norm": 0.02165898121893406,
+ "learning_rate": 0.0006,
+ "loss": 4.342130661010742,
+ "step": 1429
+ },
+ {
+ "epoch": 19.86637554585153,
+ "grad_norm": 0.02222822792828083,
+ "learning_rate": 0.0006,
+ "loss": 4.492661476135254,
+ "step": 1430
+ },
+ {
+ "epoch": 19.880349344978168,
+ "grad_norm": 0.019378066062927246,
+ "learning_rate": 0.0006,
+ "loss": 4.4614105224609375,
+ "step": 1431
+ },
+ {
+ "epoch": 19.894323144104803,
+ "grad_norm": 0.019908083602786064,
+ "learning_rate": 0.0006,
+ "loss": 4.370929718017578,
+ "step": 1432
+ },
+ {
+ "epoch": 19.90829694323144,
+ "grad_norm": 0.022296471521258354,
+ "learning_rate": 0.0006,
+ "loss": 4.438320159912109,
+ "step": 1433
+ },
+ {
+ "epoch": 19.92227074235808,
+ "grad_norm": 0.021810343489050865,
+ "learning_rate": 0.0006,
+ "loss": 4.462067127227783,
+ "step": 1434
+ },
+ {
+ "epoch": 19.936244541484715,
+ "grad_norm": 0.02129209227859974,
+ "learning_rate": 0.0006,
+ "loss": 4.541261196136475,
+ "step": 1435
+ },
+ {
+ "epoch": 19.950218340611354,
+ "grad_norm": 0.02079629711806774,
+ "learning_rate": 0.0006,
+ "loss": 4.36712646484375,
+ "step": 1436
+ },
+ {
+ "epoch": 19.964192139737992,
+ "grad_norm": 0.022501740604639053,
+ "learning_rate": 0.0006,
+ "loss": 4.5279083251953125,
+ "step": 1437
+ },
+ {
+ "epoch": 19.978165938864628,
+ "grad_norm": 0.019121676683425903,
+ "learning_rate": 0.0006,
+ "loss": 4.477789878845215,
+ "step": 1438
+ },
+ {
+ "epoch": 19.992139737991266,
+ "grad_norm": 0.017429588362574577,
+ "learning_rate": 0.0006,
+ "loss": 4.4023847579956055,
+ "step": 1439
+ },
+ {
+ "epoch": 20.0,
+ "grad_norm": 0.018857665359973907,
+ "learning_rate": 0.0006,
+ "loss": 4.494743347167969,
+ "step": 1440
+ },
+ {
+ "epoch": 20.0,
+ "eval_loss": 4.694555759429932,
+ "eval_runtime": 56.4146,
+ "eval_samples_per_second": 43.287,
+ "eval_steps_per_second": 1.365,
+ "step": 1440
+ },
+ {
+ "epoch": 20.01397379912664,
+ "grad_norm": 0.016505861654877663,
+ "learning_rate": 0.0006,
+ "loss": 4.506229877471924,
+ "step": 1441
+ },
+ {
+ "epoch": 20.027947598253274,
+ "grad_norm": 0.017571842297911644,
+ "learning_rate": 0.0006,
+ "loss": 4.384073257446289,
+ "step": 1442
+ },
+ {
+ "epoch": 20.041921397379912,
+ "grad_norm": 0.019398365169763565,
+ "learning_rate": 0.0006,
+ "loss": 4.482177734375,
+ "step": 1443
+ },
+ {
+ "epoch": 20.05589519650655,
+ "grad_norm": 0.02011062018573284,
+ "learning_rate": 0.0006,
+ "loss": 4.440773963928223,
+ "step": 1444
+ },
+ {
+ "epoch": 20.069868995633186,
+ "grad_norm": 0.019026953727006912,
+ "learning_rate": 0.0006,
+ "loss": 4.507168769836426,
+ "step": 1445
+ },
+ {
+ "epoch": 20.083842794759825,
+ "grad_norm": 0.020506031811237335,
+ "learning_rate": 0.0006,
+ "loss": 4.469079971313477,
+ "step": 1446
+ },
+ {
+ "epoch": 20.097816593886463,
+ "grad_norm": 0.019608162343502045,
+ "learning_rate": 0.0006,
+ "loss": 4.395927429199219,
+ "step": 1447
+ },
+ {
+ "epoch": 20.111790393013102,
+ "grad_norm": 0.019419124349951744,
+ "learning_rate": 0.0006,
+ "loss": 4.484235763549805,
+ "step": 1448
+ },
+ {
+ "epoch": 20.125764192139737,
+ "grad_norm": 0.021775128319859505,
+ "learning_rate": 0.0006,
+ "loss": 4.333173751831055,
+ "step": 1449
+ },
+ {
+ "epoch": 20.139737991266376,
+ "grad_norm": 0.025512272492051125,
+ "learning_rate": 0.0006,
+ "loss": 4.460362434387207,
+ "step": 1450
+ },
+ {
+ "epoch": 20.153711790393015,
+ "grad_norm": 0.025157401338219643,
+ "learning_rate": 0.0006,
+ "loss": 4.4361395835876465,
+ "step": 1451
+ },
+ {
+ "epoch": 20.16768558951965,
+ "grad_norm": 0.019862616434693336,
+ "learning_rate": 0.0006,
+ "loss": 4.426807403564453,
+ "step": 1452
+ },
+ {
+ "epoch": 20.18165938864629,
+ "grad_norm": 0.01932152360677719,
+ "learning_rate": 0.0006,
+ "loss": 4.375677108764648,
+ "step": 1453
+ },
+ {
+ "epoch": 20.195633187772927,
+ "grad_norm": 0.01760999858379364,
+ "learning_rate": 0.0006,
+ "loss": 4.42259407043457,
+ "step": 1454
+ },
+ {
+ "epoch": 20.209606986899562,
+ "grad_norm": 0.016998112201690674,
+ "learning_rate": 0.0006,
+ "loss": 4.481280326843262,
+ "step": 1455
+ },
+ {
+ "epoch": 20.2235807860262,
+ "grad_norm": 0.017199190333485603,
+ "learning_rate": 0.0006,
+ "loss": 4.496712684631348,
+ "step": 1456
+ },
+ {
+ "epoch": 20.23755458515284,
+ "grad_norm": 0.016833679750561714,
+ "learning_rate": 0.0006,
+ "loss": 4.40736722946167,
+ "step": 1457
+ },
+ {
+ "epoch": 20.251528384279474,
+ "grad_norm": 0.017546730116009712,
+ "learning_rate": 0.0006,
+ "loss": 4.4915266036987305,
+ "step": 1458
+ },
+ {
+ "epoch": 20.265502183406113,
+ "grad_norm": 0.016595078632235527,
+ "learning_rate": 0.0006,
+ "loss": 4.460815906524658,
+ "step": 1459
+ },
+ {
+ "epoch": 20.27947598253275,
+ "grad_norm": 0.017728568986058235,
+ "learning_rate": 0.0006,
+ "loss": 4.346048831939697,
+ "step": 1460
+ },
+ {
+ "epoch": 20.29344978165939,
+ "grad_norm": 0.01849197782576084,
+ "learning_rate": 0.0006,
+ "loss": 4.399681091308594,
+ "step": 1461
+ },
+ {
+ "epoch": 20.307423580786025,
+ "grad_norm": 0.018336588516831398,
+ "learning_rate": 0.0006,
+ "loss": 4.414183616638184,
+ "step": 1462
+ },
+ {
+ "epoch": 20.321397379912664,
+ "grad_norm": 0.018086519092321396,
+ "learning_rate": 0.0006,
+ "loss": 4.346586227416992,
+ "step": 1463
+ },
+ {
+ "epoch": 20.335371179039303,
+ "grad_norm": 0.019697299227118492,
+ "learning_rate": 0.0006,
+ "loss": 4.4420952796936035,
+ "step": 1464
+ },
+ {
+ "epoch": 20.349344978165938,
+ "grad_norm": 0.02198074571788311,
+ "learning_rate": 0.0006,
+ "loss": 4.588729381561279,
+ "step": 1465
+ },
+ {
+ "epoch": 20.363318777292577,
+ "grad_norm": 0.022813329473137856,
+ "learning_rate": 0.0006,
+ "loss": 4.49729585647583,
+ "step": 1466
+ },
+ {
+ "epoch": 20.377292576419215,
+ "grad_norm": 0.022006649523973465,
+ "learning_rate": 0.0006,
+ "loss": 4.475641250610352,
+ "step": 1467
+ },
+ {
+ "epoch": 20.39126637554585,
+ "grad_norm": 0.023188438266515732,
+ "learning_rate": 0.0006,
+ "loss": 4.404322624206543,
+ "step": 1468
+ },
+ {
+ "epoch": 20.40524017467249,
+ "grad_norm": 0.022226877510547638,
+ "learning_rate": 0.0006,
+ "loss": 4.50904655456543,
+ "step": 1469
+ },
+ {
+ "epoch": 20.419213973799128,
+ "grad_norm": 0.021082637831568718,
+ "learning_rate": 0.0006,
+ "loss": 4.339011192321777,
+ "step": 1470
+ },
+ {
+ "epoch": 20.433187772925763,
+ "grad_norm": 0.02013542130589485,
+ "learning_rate": 0.0006,
+ "loss": 4.516942024230957,
+ "step": 1471
+ },
+ {
+ "epoch": 20.4471615720524,
+ "grad_norm": 0.01898968778550625,
+ "learning_rate": 0.0006,
+ "loss": 4.315804481506348,
+ "step": 1472
+ },
+ {
+ "epoch": 20.46113537117904,
+ "grad_norm": 0.019450880587100983,
+ "learning_rate": 0.0006,
+ "loss": 4.4794921875,
+ "step": 1473
+ },
+ {
+ "epoch": 20.475109170305675,
+ "grad_norm": 0.018577802926301956,
+ "learning_rate": 0.0006,
+ "loss": 4.422323703765869,
+ "step": 1474
+ },
+ {
+ "epoch": 20.489082969432314,
+ "grad_norm": 0.01778412237763405,
+ "learning_rate": 0.0006,
+ "loss": 4.429789066314697,
+ "step": 1475
+ },
+ {
+ "epoch": 20.503056768558952,
+ "grad_norm": 0.017134476453065872,
+ "learning_rate": 0.0006,
+ "loss": 4.416459083557129,
+ "step": 1476
+ },
+ {
+ "epoch": 20.51703056768559,
+ "grad_norm": 0.017153557389974594,
+ "learning_rate": 0.0006,
+ "loss": 4.446170806884766,
+ "step": 1477
+ },
+ {
+ "epoch": 20.531004366812226,
+ "grad_norm": 0.014998532831668854,
+ "learning_rate": 0.0006,
+ "loss": 4.419447898864746,
+ "step": 1478
+ },
+ {
+ "epoch": 20.544978165938865,
+ "grad_norm": 0.015877850353717804,
+ "learning_rate": 0.0006,
+ "loss": 4.312338829040527,
+ "step": 1479
+ },
+ {
+ "epoch": 20.558951965065503,
+ "grad_norm": 0.01569489948451519,
+ "learning_rate": 0.0006,
+ "loss": 4.572905540466309,
+ "step": 1480
+ },
+ {
+ "epoch": 20.57292576419214,
+ "grad_norm": 0.01591474376618862,
+ "learning_rate": 0.0006,
+ "loss": 4.414907455444336,
+ "step": 1481
+ },
+ {
+ "epoch": 20.586899563318777,
+ "grad_norm": 0.015395903028547764,
+ "learning_rate": 0.0006,
+ "loss": 4.549131393432617,
+ "step": 1482
+ },
+ {
+ "epoch": 20.600873362445416,
+ "grad_norm": 0.015715358778834343,
+ "learning_rate": 0.0006,
+ "loss": 4.530194282531738,
+ "step": 1483
+ },
+ {
+ "epoch": 20.61484716157205,
+ "grad_norm": 0.016042975708842278,
+ "learning_rate": 0.0006,
+ "loss": 4.322149276733398,
+ "step": 1484
+ },
+ {
+ "epoch": 20.62882096069869,
+ "grad_norm": 0.01664186641573906,
+ "learning_rate": 0.0006,
+ "loss": 4.424866676330566,
+ "step": 1485
+ },
+ {
+ "epoch": 20.64279475982533,
+ "grad_norm": 0.01631534844636917,
+ "learning_rate": 0.0006,
+ "loss": 4.441752910614014,
+ "step": 1486
+ },
+ {
+ "epoch": 20.656768558951963,
+ "grad_norm": 0.014057058840990067,
+ "learning_rate": 0.0006,
+ "loss": 4.399982452392578,
+ "step": 1487
+ },
+ {
+ "epoch": 20.670742358078602,
+ "grad_norm": 0.015076273120939732,
+ "learning_rate": 0.0006,
+ "loss": 4.441324234008789,
+ "step": 1488
+ },
+ {
+ "epoch": 20.68471615720524,
+ "grad_norm": 0.014138277620077133,
+ "learning_rate": 0.0006,
+ "loss": 4.4312543869018555,
+ "step": 1489
+ },
+ {
+ "epoch": 20.69868995633188,
+ "grad_norm": 0.013762440532445908,
+ "learning_rate": 0.0006,
+ "loss": 4.387807846069336,
+ "step": 1490
+ },
+ {
+ "epoch": 20.712663755458514,
+ "grad_norm": 0.015058411285281181,
+ "learning_rate": 0.0006,
+ "loss": 4.594747543334961,
+ "step": 1491
+ },
+ {
+ "epoch": 20.726637554585153,
+ "grad_norm": 0.013827123679220676,
+ "learning_rate": 0.0006,
+ "loss": 4.434604167938232,
+ "step": 1492
+ },
+ {
+ "epoch": 20.74061135371179,
+ "grad_norm": 0.015315295197069645,
+ "learning_rate": 0.0006,
+ "loss": 4.588066101074219,
+ "step": 1493
+ },
+ {
+ "epoch": 20.754585152838427,
+ "grad_norm": 0.018678732216358185,
+ "learning_rate": 0.0006,
+ "loss": 4.3514204025268555,
+ "step": 1494
+ },
+ {
+ "epoch": 20.768558951965066,
+ "grad_norm": 0.019521350041031837,
+ "learning_rate": 0.0006,
+ "loss": 4.437658786773682,
+ "step": 1495
+ },
+ {
+ "epoch": 20.782532751091704,
+ "grad_norm": 0.018363403156399727,
+ "learning_rate": 0.0006,
+ "loss": 4.495002269744873,
+ "step": 1496
+ },
+ {
+ "epoch": 20.79650655021834,
+ "grad_norm": 0.016830725595355034,
+ "learning_rate": 0.0006,
+ "loss": 4.402471542358398,
+ "step": 1497
+ },
+ {
+ "epoch": 20.810480349344978,
+ "grad_norm": 0.01722072623670101,
+ "learning_rate": 0.0006,
+ "loss": 4.437653541564941,
+ "step": 1498
+ },
+ {
+ "epoch": 20.824454148471617,
+ "grad_norm": 0.017459599301218987,
+ "learning_rate": 0.0006,
+ "loss": 4.366603851318359,
+ "step": 1499
+ },
+ {
+ "epoch": 20.83842794759825,
+ "grad_norm": 0.018902219831943512,
+ "learning_rate": 0.0006,
+ "loss": 4.501561164855957,
+ "step": 1500
+ },
+ {
+ "epoch": 20.85240174672489,
+ "grad_norm": 0.020814096555113792,
+ "learning_rate": 0.0006,
+ "loss": 4.535295486450195,
+ "step": 1501
+ },
+ {
+ "epoch": 20.86637554585153,
+ "grad_norm": 0.018990976735949516,
+ "learning_rate": 0.0006,
+ "loss": 4.333580017089844,
+ "step": 1502
+ },
+ {
+ "epoch": 20.880349344978168,
+ "grad_norm": 0.01837880350649357,
+ "learning_rate": 0.0006,
+ "loss": 4.313388824462891,
+ "step": 1503
+ },
+ {
+ "epoch": 20.894323144104803,
+ "grad_norm": 0.019820360466837883,
+ "learning_rate": 0.0006,
+ "loss": 4.388157844543457,
+ "step": 1504
+ },
+ {
+ "epoch": 20.90829694323144,
+ "grad_norm": 0.020181143656373024,
+ "learning_rate": 0.0006,
+ "loss": 4.572224140167236,
+ "step": 1505
+ },
+ {
+ "epoch": 20.92227074235808,
+ "grad_norm": 0.021206077188253403,
+ "learning_rate": 0.0006,
+ "loss": 4.3127970695495605,
+ "step": 1506
+ },
+ {
+ "epoch": 20.936244541484715,
+ "grad_norm": 0.017962060868740082,
+ "learning_rate": 0.0006,
+ "loss": 4.384734630584717,
+ "step": 1507
+ },
+ {
+ "epoch": 20.950218340611354,
+ "grad_norm": 0.01737220585346222,
+ "learning_rate": 0.0006,
+ "loss": 4.4075751304626465,
+ "step": 1508
+ },
+ {
+ "epoch": 20.964192139737992,
+ "grad_norm": 0.01839268207550049,
+ "learning_rate": 0.0006,
+ "loss": 4.553779125213623,
+ "step": 1509
+ },
+ {
+ "epoch": 20.978165938864628,
+ "grad_norm": 0.020102640613913536,
+ "learning_rate": 0.0006,
+ "loss": 4.399398326873779,
+ "step": 1510
+ },
+ {
+ "epoch": 20.992139737991266,
+ "grad_norm": 0.018831463530659676,
+ "learning_rate": 0.0006,
+ "loss": 4.358499526977539,
+ "step": 1511
+ },
+ {
+ "epoch": 21.0,
+ "grad_norm": 0.018254734575748444,
+ "learning_rate": 0.0006,
+ "loss": 4.384119510650635,
+ "step": 1512
+ },
+ {
+ "epoch": 21.0,
+ "eval_loss": 4.7524237632751465,
+ "eval_runtime": 57.8812,
+ "eval_samples_per_second": 42.19,
+ "eval_steps_per_second": 1.33,
+ "step": 1512
+ },
+ {
+ "epoch": 21.01397379912664,
+ "grad_norm": 0.019728587940335274,
+ "learning_rate": 0.0006,
+ "loss": 4.32363224029541,
+ "step": 1513
+ },
+ {
+ "epoch": 21.027947598253274,
+ "grad_norm": 0.02216625027358532,
+ "learning_rate": 0.0006,
+ "loss": 4.436988353729248,
+ "step": 1514
+ },
+ {
+ "epoch": 21.041921397379912,
+ "grad_norm": 0.02227606810629368,
+ "learning_rate": 0.0006,
+ "loss": 4.406063556671143,
+ "step": 1515
+ },
+ {
+ "epoch": 21.05589519650655,
+ "grad_norm": 0.020187009125947952,
+ "learning_rate": 0.0006,
+ "loss": 4.324402332305908,
+ "step": 1516
+ },
+ {
+ "epoch": 21.069868995633186,
+ "grad_norm": 0.02095157466828823,
+ "learning_rate": 0.0006,
+ "loss": 4.456544876098633,
+ "step": 1517
+ },
+ {
+ "epoch": 21.083842794759825,
+ "grad_norm": 0.02146013453602791,
+ "learning_rate": 0.0006,
+ "loss": 4.368387222290039,
+ "step": 1518
+ },
+ {
+ "epoch": 21.097816593886463,
+ "grad_norm": 0.02162035182118416,
+ "learning_rate": 0.0006,
+ "loss": 4.48838996887207,
+ "step": 1519
+ },
+ {
+ "epoch": 21.111790393013102,
+ "grad_norm": 0.021745676174759865,
+ "learning_rate": 0.0006,
+ "loss": 4.30778694152832,
+ "step": 1520
+ },
+ {
+ "epoch": 21.125764192139737,
+ "grad_norm": 0.02173588052392006,
+ "learning_rate": 0.0006,
+ "loss": 4.433149814605713,
+ "step": 1521
+ },
+ {
+ "epoch": 21.139737991266376,
+ "grad_norm": 0.020950648933649063,
+ "learning_rate": 0.0006,
+ "loss": 4.520089149475098,
+ "step": 1522
+ },
+ {
+ "epoch": 21.153711790393015,
+ "grad_norm": 0.020684752613306046,
+ "learning_rate": 0.0006,
+ "loss": 4.372437477111816,
+ "step": 1523
+ },
+ {
+ "epoch": 21.16768558951965,
+ "grad_norm": 0.019472869113087654,
+ "learning_rate": 0.0006,
+ "loss": 4.437015533447266,
+ "step": 1524
+ },
+ {
+ "epoch": 21.18165938864629,
+ "grad_norm": 0.01846962980926037,
+ "learning_rate": 0.0006,
+ "loss": 4.449531555175781,
+ "step": 1525
+ },
+ {
+ "epoch": 21.195633187772927,
+ "grad_norm": 0.017377803102135658,
+ "learning_rate": 0.0006,
+ "loss": 4.458279132843018,
+ "step": 1526
+ },
+ {
+ "epoch": 21.209606986899562,
+ "grad_norm": 0.016002200543880463,
+ "learning_rate": 0.0006,
+ "loss": 4.266297817230225,
+ "step": 1527
+ },
+ {
+ "epoch": 21.2235807860262,
+ "grad_norm": 0.015575903467833996,
+ "learning_rate": 0.0006,
+ "loss": 4.342632293701172,
+ "step": 1528
+ },
+ {
+ "epoch": 21.23755458515284,
+ "grad_norm": 0.015634698793292046,
+ "learning_rate": 0.0006,
+ "loss": 4.338623523712158,
+ "step": 1529
+ },
+ {
+ "epoch": 21.251528384279474,
+ "grad_norm": 0.01449581328779459,
+ "learning_rate": 0.0006,
+ "loss": 4.354330062866211,
+ "step": 1530
+ },
+ {
+ "epoch": 21.265502183406113,
+ "grad_norm": 0.01487616915255785,
+ "learning_rate": 0.0006,
+ "loss": 4.486835479736328,
+ "step": 1531
+ },
+ {
+ "epoch": 21.27947598253275,
+ "grad_norm": 0.016346555203199387,
+ "learning_rate": 0.0006,
+ "loss": 4.407098770141602,
+ "step": 1532
+ },
+ {
+ "epoch": 21.29344978165939,
+ "grad_norm": 0.01682182401418686,
+ "learning_rate": 0.0006,
+ "loss": 4.436809539794922,
+ "step": 1533
+ },
+ {
+ "epoch": 21.307423580786025,
+ "grad_norm": 0.017704004421830177,
+ "learning_rate": 0.0006,
+ "loss": 4.3854780197143555,
+ "step": 1534
+ },
+ {
+ "epoch": 21.321397379912664,
+ "grad_norm": 0.018437420949339867,
+ "learning_rate": 0.0006,
+ "loss": 4.413283824920654,
+ "step": 1535
+ },
+ {
+ "epoch": 21.335371179039303,
+ "grad_norm": 0.016911081969738007,
+ "learning_rate": 0.0006,
+ "loss": 4.357306003570557,
+ "step": 1536
+ },
+ {
+ "epoch": 21.349344978165938,
+ "grad_norm": 0.01642223633825779,
+ "learning_rate": 0.0006,
+ "loss": 4.422322750091553,
+ "step": 1537
+ },
+ {
+ "epoch": 21.363318777292577,
+ "grad_norm": 0.01780877821147442,
+ "learning_rate": 0.0006,
+ "loss": 4.387620449066162,
+ "step": 1538
+ },
+ {
+ "epoch": 21.377292576419215,
+ "grad_norm": 0.01653975248336792,
+ "learning_rate": 0.0006,
+ "loss": 4.327267646789551,
+ "step": 1539
+ },
+ {
+ "epoch": 21.39126637554585,
+ "grad_norm": 0.01572156324982643,
+ "learning_rate": 0.0006,
+ "loss": 4.475979804992676,
+ "step": 1540
+ },
+ {
+ "epoch": 21.40524017467249,
+ "grad_norm": 0.017156066372990608,
+ "learning_rate": 0.0006,
+ "loss": 4.4011125564575195,
+ "step": 1541
+ },
+ {
+ "epoch": 21.419213973799128,
+ "grad_norm": 0.01788829080760479,
+ "learning_rate": 0.0006,
+ "loss": 4.4174089431762695,
+ "step": 1542
+ },
+ {
+ "epoch": 21.433187772925763,
+ "grad_norm": 0.01728985086083412,
+ "learning_rate": 0.0006,
+ "loss": 4.429244041442871,
+ "step": 1543
+ },
+ {
+ "epoch": 21.4471615720524,
+ "grad_norm": 0.016534466296434402,
+ "learning_rate": 0.0006,
+ "loss": 4.326068878173828,
+ "step": 1544
+ },
+ {
+ "epoch": 21.46113537117904,
+ "grad_norm": 0.016582584008574486,
+ "learning_rate": 0.0006,
+ "loss": 4.2975969314575195,
+ "step": 1545
+ },
+ {
+ "epoch": 21.475109170305675,
+ "grad_norm": 0.018034178763628006,
+ "learning_rate": 0.0006,
+ "loss": 4.501319408416748,
+ "step": 1546
+ },
+ {
+ "epoch": 21.489082969432314,
+ "grad_norm": 0.017827702686190605,
+ "learning_rate": 0.0006,
+ "loss": 4.502284526824951,
+ "step": 1547
+ },
+ {
+ "epoch": 21.503056768558952,
+ "grad_norm": 0.017182299867272377,
+ "learning_rate": 0.0006,
+ "loss": 4.380828857421875,
+ "step": 1548
+ },
+ {
+ "epoch": 21.51703056768559,
+ "grad_norm": 0.019101588055491447,
+ "learning_rate": 0.0006,
+ "loss": 4.31564998626709,
+ "step": 1549
+ },
+ {
+ "epoch": 21.531004366812226,
+ "grad_norm": 0.0204677302390337,
+ "learning_rate": 0.0006,
+ "loss": 4.3840131759643555,
+ "step": 1550
+ },
+ {
+ "epoch": 21.544978165938865,
+ "grad_norm": 0.0188993439078331,
+ "learning_rate": 0.0006,
+ "loss": 4.4257731437683105,
+ "step": 1551
+ },
+ {
+ "epoch": 21.558951965065503,
+ "grad_norm": 0.01730230636894703,
+ "learning_rate": 0.0006,
+ "loss": 4.445132255554199,
+ "step": 1552
+ },
+ {
+ "epoch": 21.57292576419214,
+ "grad_norm": 0.01801574043929577,
+ "learning_rate": 0.0006,
+ "loss": 4.427064895629883,
+ "step": 1553
+ },
+ {
+ "epoch": 21.586899563318777,
+ "grad_norm": 0.018417565152049065,
+ "learning_rate": 0.0006,
+ "loss": 4.319595813751221,
+ "step": 1554
+ },
+ {
+ "epoch": 21.600873362445416,
+ "grad_norm": 0.01683160290122032,
+ "learning_rate": 0.0006,
+ "loss": 4.349218368530273,
+ "step": 1555
+ },
+ {
+ "epoch": 21.61484716157205,
+ "grad_norm": 0.016973106190562248,
+ "learning_rate": 0.0006,
+ "loss": 4.375990390777588,
+ "step": 1556
+ },
+ {
+ "epoch": 21.62882096069869,
+ "grad_norm": 0.015688583254814148,
+ "learning_rate": 0.0006,
+ "loss": 4.408170223236084,
+ "step": 1557
+ },
+ {
+ "epoch": 21.64279475982533,
+ "grad_norm": 0.01573043316602707,
+ "learning_rate": 0.0006,
+ "loss": 4.416255950927734,
+ "step": 1558
+ },
+ {
+ "epoch": 21.656768558951963,
+ "grad_norm": 0.01744293048977852,
+ "learning_rate": 0.0006,
+ "loss": 4.477141380310059,
+ "step": 1559
+ },
+ {
+ "epoch": 21.670742358078602,
+ "grad_norm": 0.017503513023257256,
+ "learning_rate": 0.0006,
+ "loss": 4.3367486000061035,
+ "step": 1560
+ },
+ {
+ "epoch": 21.68471615720524,
+ "grad_norm": 0.016592150554060936,
+ "learning_rate": 0.0006,
+ "loss": 4.314785480499268,
+ "step": 1561
+ },
+ {
+ "epoch": 21.69868995633188,
+ "grad_norm": 0.015774084255099297,
+ "learning_rate": 0.0006,
+ "loss": 4.448209285736084,
+ "step": 1562
+ },
+ {
+ "epoch": 21.712663755458514,
+ "grad_norm": 0.015779150649905205,
+ "learning_rate": 0.0006,
+ "loss": 4.392690658569336,
+ "step": 1563
+ },
+ {
+ "epoch": 21.726637554585153,
+ "grad_norm": 0.015392648056149483,
+ "learning_rate": 0.0006,
+ "loss": 4.437741756439209,
+ "step": 1564
+ },
+ {
+ "epoch": 21.74061135371179,
+ "grad_norm": 0.01530836895108223,
+ "learning_rate": 0.0006,
+ "loss": 4.424595832824707,
+ "step": 1565
+ },
+ {
+ "epoch": 21.754585152838427,
+ "grad_norm": 0.015600779093801975,
+ "learning_rate": 0.0006,
+ "loss": 4.335417747497559,
+ "step": 1566
+ },
+ {
+ "epoch": 21.768558951965066,
+ "grad_norm": 0.014646518044173717,
+ "learning_rate": 0.0006,
+ "loss": 4.279081344604492,
+ "step": 1567
+ },
+ {
+ "epoch": 21.782532751091704,
+ "grad_norm": 0.014132864773273468,
+ "learning_rate": 0.0006,
+ "loss": 4.345950126647949,
+ "step": 1568
+ },
+ {
+ "epoch": 21.79650655021834,
+ "grad_norm": 0.01425724383443594,
+ "learning_rate": 0.0006,
+ "loss": 4.277588844299316,
+ "step": 1569
+ },
+ {
+ "epoch": 21.810480349344978,
+ "grad_norm": 0.015396572649478912,
+ "learning_rate": 0.0006,
+ "loss": 4.326196670532227,
+ "step": 1570
+ },
+ {
+ "epoch": 21.824454148471617,
+ "grad_norm": 0.01683669537305832,
+ "learning_rate": 0.0006,
+ "loss": 4.3294172286987305,
+ "step": 1571
+ },
+ {
+ "epoch": 21.83842794759825,
+ "grad_norm": 0.016365647315979004,
+ "learning_rate": 0.0006,
+ "loss": 4.523321628570557,
+ "step": 1572
+ },
+ {
+ "epoch": 21.85240174672489,
+ "grad_norm": 0.017142774537205696,
+ "learning_rate": 0.0006,
+ "loss": 4.475842475891113,
+ "step": 1573
+ },
+ {
+ "epoch": 21.86637554585153,
+ "grad_norm": 0.015950839966535568,
+ "learning_rate": 0.0006,
+ "loss": 4.397187232971191,
+ "step": 1574
+ },
+ {
+ "epoch": 21.880349344978168,
+ "grad_norm": 0.016180362552404404,
+ "learning_rate": 0.0006,
+ "loss": 4.414778709411621,
+ "step": 1575
+ },
+ {
+ "epoch": 21.894323144104803,
+ "grad_norm": 0.017615556716918945,
+ "learning_rate": 0.0006,
+ "loss": 4.348567962646484,
+ "step": 1576
+ },
+ {
+ "epoch": 21.90829694323144,
+ "grad_norm": 0.01896277628839016,
+ "learning_rate": 0.0006,
+ "loss": 4.431762218475342,
+ "step": 1577
+ },
+ {
+ "epoch": 21.92227074235808,
+ "grad_norm": 0.01990971714258194,
+ "learning_rate": 0.0006,
+ "loss": 4.431285858154297,
+ "step": 1578
+ },
+ {
+ "epoch": 21.936244541484715,
+ "grad_norm": 0.0201258547604084,
+ "learning_rate": 0.0006,
+ "loss": 4.309872627258301,
+ "step": 1579
+ },
+ {
+ "epoch": 21.950218340611354,
+ "grad_norm": 0.019620198756456375,
+ "learning_rate": 0.0006,
+ "loss": 4.484044551849365,
+ "step": 1580
+ },
+ {
+ "epoch": 21.964192139737992,
+ "grad_norm": 0.01771123707294464,
+ "learning_rate": 0.0006,
+ "loss": 4.368093013763428,
+ "step": 1581
+ },
+ {
+ "epoch": 21.978165938864628,
+ "grad_norm": 0.017501654103398323,
+ "learning_rate": 0.0006,
+ "loss": 4.372651100158691,
+ "step": 1582
+ },
+ {
+ "epoch": 21.992139737991266,
+ "grad_norm": 0.01693608984351158,
+ "learning_rate": 0.0006,
+ "loss": 4.379167556762695,
+ "step": 1583
+ },
+ {
+ "epoch": 22.0,
+ "grad_norm": 0.02002849616110325,
+ "learning_rate": 0.0006,
+ "loss": 4.329627990722656,
+ "step": 1584
+ },
+ {
+ "epoch": 22.0,
+ "eval_loss": 4.723696231842041,
+ "eval_runtime": 56.1758,
+ "eval_samples_per_second": 43.471,
+ "eval_steps_per_second": 1.371,
+ "step": 1584
+ },
+ {
+ "epoch": 22.01397379912664,
+ "grad_norm": 0.02030305378139019,
+ "learning_rate": 0.0006,
+ "loss": 4.355566024780273,
+ "step": 1585
+ },
+ {
+ "epoch": 22.027947598253274,
+ "grad_norm": 0.018638934940099716,
+ "learning_rate": 0.0006,
+ "loss": 4.314009189605713,
+ "step": 1586
+ },
+ {
+ "epoch": 22.041921397379912,
+ "grad_norm": 0.019917072728276253,
+ "learning_rate": 0.0006,
+ "loss": 4.367525577545166,
+ "step": 1587
+ },
+ {
+ "epoch": 22.05589519650655,
+ "grad_norm": 0.020785167813301086,
+ "learning_rate": 0.0006,
+ "loss": 4.375561714172363,
+ "step": 1588
+ },
+ {
+ "epoch": 22.069868995633186,
+ "grad_norm": 0.021214453503489494,
+ "learning_rate": 0.0006,
+ "loss": 4.418305397033691,
+ "step": 1589
+ },
+ {
+ "epoch": 22.083842794759825,
+ "grad_norm": 0.019051678478717804,
+ "learning_rate": 0.0006,
+ "loss": 4.215126991271973,
+ "step": 1590
+ },
+ {
+ "epoch": 22.097816593886463,
+ "grad_norm": 0.02323366142809391,
+ "learning_rate": 0.0006,
+ "loss": 4.306807994842529,
+ "step": 1591
+ },
+ {
+ "epoch": 22.111790393013102,
+ "grad_norm": 0.026186738163232803,
+ "learning_rate": 0.0006,
+ "loss": 4.408508777618408,
+ "step": 1592
+ },
+ {
+ "epoch": 22.125764192139737,
+ "grad_norm": 0.027361227199435234,
+ "learning_rate": 0.0006,
+ "loss": 4.329322814941406,
+ "step": 1593
+ },
+ {
+ "epoch": 22.139737991266376,
+ "grad_norm": 0.026564646512269974,
+ "learning_rate": 0.0006,
+ "loss": 4.346642017364502,
+ "step": 1594
+ },
+ {
+ "epoch": 22.153711790393015,
+ "grad_norm": 0.025073667988181114,
+ "learning_rate": 0.0006,
+ "loss": 4.3984198570251465,
+ "step": 1595
+ },
+ {
+ "epoch": 22.16768558951965,
+ "grad_norm": 0.02347794733941555,
+ "learning_rate": 0.0006,
+ "loss": 4.352538108825684,
+ "step": 1596
+ },
+ {
+ "epoch": 22.18165938864629,
+ "grad_norm": 0.02203468047082424,
+ "learning_rate": 0.0006,
+ "loss": 4.298148155212402,
+ "step": 1597
+ },
+ {
+ "epoch": 22.195633187772927,
+ "grad_norm": 0.020621957257390022,
+ "learning_rate": 0.0006,
+ "loss": 4.480309963226318,
+ "step": 1598
+ },
+ {
+ "epoch": 22.209606986899562,
+ "grad_norm": 0.023710055276751518,
+ "learning_rate": 0.0006,
+ "loss": 4.339317321777344,
+ "step": 1599
+ },
+ {
+ "epoch": 22.2235807860262,
+ "grad_norm": 0.0233482476323843,
+ "learning_rate": 0.0006,
+ "loss": 4.436896324157715,
+ "step": 1600
+ },
+ {
+ "epoch": 22.23755458515284,
+ "grad_norm": 0.020122403278946877,
+ "learning_rate": 0.0006,
+ "loss": 4.236002445220947,
+ "step": 1601
+ },
+ {
+ "epoch": 22.251528384279474,
+ "grad_norm": 0.01957266591489315,
+ "learning_rate": 0.0006,
+ "loss": 4.497328281402588,
+ "step": 1602
+ },
+ {
+ "epoch": 22.265502183406113,
+ "grad_norm": 0.018964895978569984,
+ "learning_rate": 0.0006,
+ "loss": 4.36360502243042,
+ "step": 1603
+ },
+ {
+ "epoch": 22.27947598253275,
+ "grad_norm": 0.018902769312262535,
+ "learning_rate": 0.0006,
+ "loss": 4.427093505859375,
+ "step": 1604
+ },
+ {
+ "epoch": 22.29344978165939,
+ "grad_norm": 0.018918119370937347,
+ "learning_rate": 0.0006,
+ "loss": 4.294950485229492,
+ "step": 1605
+ },
+ {
+ "epoch": 22.307423580786025,
+ "grad_norm": 0.019315775483846664,
+ "learning_rate": 0.0006,
+ "loss": 4.427952289581299,
+ "step": 1606
+ },
+ {
+ "epoch": 22.321397379912664,
+ "grad_norm": 0.018350522965192795,
+ "learning_rate": 0.0006,
+ "loss": 4.272945404052734,
+ "step": 1607
+ },
+ {
+ "epoch": 22.335371179039303,
+ "grad_norm": 0.01759534701704979,
+ "learning_rate": 0.0006,
+ "loss": 4.348023891448975,
+ "step": 1608
+ },
+ {
+ "epoch": 22.349344978165938,
+ "grad_norm": 0.01646057888865471,
+ "learning_rate": 0.0006,
+ "loss": 4.278518199920654,
+ "step": 1609
+ },
+ {
+ "epoch": 22.363318777292577,
+ "grad_norm": 0.015875060111284256,
+ "learning_rate": 0.0006,
+ "loss": 4.423527240753174,
+ "step": 1610
+ },
+ {
+ "epoch": 22.377292576419215,
+ "grad_norm": 0.01638529635965824,
+ "learning_rate": 0.0006,
+ "loss": 4.321723937988281,
+ "step": 1611
+ },
+ {
+ "epoch": 22.39126637554585,
+ "grad_norm": 0.016942061483860016,
+ "learning_rate": 0.0006,
+ "loss": 4.346164703369141,
+ "step": 1612
+ },
+ {
+ "epoch": 22.40524017467249,
+ "grad_norm": 0.015209621749818325,
+ "learning_rate": 0.0006,
+ "loss": 4.434340000152588,
+ "step": 1613
+ },
+ {
+ "epoch": 22.419213973799128,
+ "grad_norm": 0.014423009008169174,
+ "learning_rate": 0.0006,
+ "loss": 4.342560768127441,
+ "step": 1614
+ },
+ {
+ "epoch": 22.433187772925763,
+ "grad_norm": 0.014369525946676731,
+ "learning_rate": 0.0006,
+ "loss": 4.296586990356445,
+ "step": 1615
+ },
+ {
+ "epoch": 22.4471615720524,
+ "grad_norm": 0.01392540242522955,
+ "learning_rate": 0.0006,
+ "loss": 4.433176040649414,
+ "step": 1616
+ },
+ {
+ "epoch": 22.46113537117904,
+ "grad_norm": 0.013206899166107178,
+ "learning_rate": 0.0006,
+ "loss": 4.32962703704834,
+ "step": 1617
+ },
+ {
+ "epoch": 22.475109170305675,
+ "grad_norm": 0.013427077792584896,
+ "learning_rate": 0.0006,
+ "loss": 4.420437335968018,
+ "step": 1618
+ },
+ {
+ "epoch": 22.489082969432314,
+ "grad_norm": 0.014194196090102196,
+ "learning_rate": 0.0006,
+ "loss": 4.410478591918945,
+ "step": 1619
+ },
+ {
+ "epoch": 22.503056768558952,
+ "grad_norm": 0.01362364087253809,
+ "learning_rate": 0.0006,
+ "loss": 4.338932037353516,
+ "step": 1620
+ },
+ {
+ "epoch": 22.51703056768559,
+ "grad_norm": 0.014299589209258556,
+ "learning_rate": 0.0006,
+ "loss": 4.34089994430542,
+ "step": 1621
+ },
+ {
+ "epoch": 22.531004366812226,
+ "grad_norm": 0.014061295427381992,
+ "learning_rate": 0.0006,
+ "loss": 4.29201602935791,
+ "step": 1622
+ },
+ {
+ "epoch": 22.544978165938865,
+ "grad_norm": 0.013945229351520538,
+ "learning_rate": 0.0006,
+ "loss": 4.326052188873291,
+ "step": 1623
+ },
+ {
+ "epoch": 22.558951965065503,
+ "grad_norm": 0.01387068536132574,
+ "learning_rate": 0.0006,
+ "loss": 4.393521308898926,
+ "step": 1624
+ },
+ {
+ "epoch": 22.57292576419214,
+ "grad_norm": 0.013931898400187492,
+ "learning_rate": 0.0006,
+ "loss": 4.3739752769470215,
+ "step": 1625
+ },
+ {
+ "epoch": 22.586899563318777,
+ "grad_norm": 0.01497740764170885,
+ "learning_rate": 0.0006,
+ "loss": 4.3991618156433105,
+ "step": 1626
+ },
+ {
+ "epoch": 22.600873362445416,
+ "grad_norm": 0.015842905268073082,
+ "learning_rate": 0.0006,
+ "loss": 4.405340671539307,
+ "step": 1627
+ },
+ {
+ "epoch": 22.61484716157205,
+ "grad_norm": 0.01580633595585823,
+ "learning_rate": 0.0006,
+ "loss": 4.280102729797363,
+ "step": 1628
+ },
+ {
+ "epoch": 22.62882096069869,
+ "grad_norm": 0.016567382961511612,
+ "learning_rate": 0.0006,
+ "loss": 4.407657146453857,
+ "step": 1629
+ },
+ {
+ "epoch": 22.64279475982533,
+ "grad_norm": 0.01648622378706932,
+ "learning_rate": 0.0006,
+ "loss": 4.392590522766113,
+ "step": 1630
+ },
+ {
+ "epoch": 22.656768558951963,
+ "grad_norm": 0.017066307365894318,
+ "learning_rate": 0.0006,
+ "loss": 4.324862003326416,
+ "step": 1631
+ },
+ {
+ "epoch": 22.670742358078602,
+ "grad_norm": 0.01637062057852745,
+ "learning_rate": 0.0006,
+ "loss": 4.34598445892334,
+ "step": 1632
+ },
+ {
+ "epoch": 22.68471615720524,
+ "grad_norm": 0.018038960173726082,
+ "learning_rate": 0.0006,
+ "loss": 4.320711612701416,
+ "step": 1633
+ },
+ {
+ "epoch": 22.69868995633188,
+ "grad_norm": 0.018096428364515305,
+ "learning_rate": 0.0006,
+ "loss": 4.536527156829834,
+ "step": 1634
+ },
+ {
+ "epoch": 22.712663755458514,
+ "grad_norm": 0.017889291048049927,
+ "learning_rate": 0.0006,
+ "loss": 4.388131141662598,
+ "step": 1635
+ },
+ {
+ "epoch": 22.726637554585153,
+ "grad_norm": 0.018771521747112274,
+ "learning_rate": 0.0006,
+ "loss": 4.357668876647949,
+ "step": 1636
+ },
+ {
+ "epoch": 22.74061135371179,
+ "grad_norm": 0.02100873365998268,
+ "learning_rate": 0.0006,
+ "loss": 4.293972015380859,
+ "step": 1637
+ },
+ {
+ "epoch": 22.754585152838427,
+ "grad_norm": 0.017715785652399063,
+ "learning_rate": 0.0006,
+ "loss": 4.349849224090576,
+ "step": 1638
+ },
+ {
+ "epoch": 22.768558951965066,
+ "grad_norm": 0.015546333976089954,
+ "learning_rate": 0.0006,
+ "loss": 4.311728000640869,
+ "step": 1639
+ },
+ {
+ "epoch": 22.782532751091704,
+ "grad_norm": 0.01605990156531334,
+ "learning_rate": 0.0006,
+ "loss": 4.499123573303223,
+ "step": 1640
+ },
+ {
+ "epoch": 22.79650655021834,
+ "grad_norm": 0.016642142087221146,
+ "learning_rate": 0.0006,
+ "loss": 4.4319257736206055,
+ "step": 1641
+ },
+ {
+ "epoch": 22.810480349344978,
+ "grad_norm": 0.016120387241244316,
+ "learning_rate": 0.0006,
+ "loss": 4.4373650550842285,
+ "step": 1642
+ },
+ {
+ "epoch": 22.824454148471617,
+ "grad_norm": 0.015174890868365765,
+ "learning_rate": 0.0006,
+ "loss": 4.314082145690918,
+ "step": 1643
+ },
+ {
+ "epoch": 22.83842794759825,
+ "grad_norm": 0.01551714539527893,
+ "learning_rate": 0.0006,
+ "loss": 4.360363960266113,
+ "step": 1644
+ },
+ {
+ "epoch": 22.85240174672489,
+ "grad_norm": 0.014966742135584354,
+ "learning_rate": 0.0006,
+ "loss": 4.346253395080566,
+ "step": 1645
+ },
+ {
+ "epoch": 22.86637554585153,
+ "grad_norm": 0.014288028702139854,
+ "learning_rate": 0.0006,
+ "loss": 4.464366436004639,
+ "step": 1646
+ },
+ {
+ "epoch": 22.880349344978168,
+ "grad_norm": 0.015291682444512844,
+ "learning_rate": 0.0006,
+ "loss": 4.504101753234863,
+ "step": 1647
+ },
+ {
+ "epoch": 22.894323144104803,
+ "grad_norm": 0.014534426853060722,
+ "learning_rate": 0.0006,
+ "loss": 4.462172031402588,
+ "step": 1648
+ },
+ {
+ "epoch": 22.90829694323144,
+ "grad_norm": 0.014129819348454475,
+ "learning_rate": 0.0006,
+ "loss": 4.350126266479492,
+ "step": 1649
+ },
+ {
+ "epoch": 22.92227074235808,
+ "grad_norm": 0.014932668767869473,
+ "learning_rate": 0.0006,
+ "loss": 4.374392032623291,
+ "step": 1650
+ },
+ {
+ "epoch": 22.936244541484715,
+ "grad_norm": 0.014647853560745716,
+ "learning_rate": 0.0006,
+ "loss": 4.252264499664307,
+ "step": 1651
+ },
+ {
+ "epoch": 22.950218340611354,
+ "grad_norm": 0.013553360477089882,
+ "learning_rate": 0.0006,
+ "loss": 4.379729270935059,
+ "step": 1652
+ },
+ {
+ "epoch": 22.964192139737992,
+ "grad_norm": 0.014606194570660591,
+ "learning_rate": 0.0006,
+ "loss": 4.363372802734375,
+ "step": 1653
+ },
+ {
+ "epoch": 22.978165938864628,
+ "grad_norm": 0.015000631101429462,
+ "learning_rate": 0.0006,
+ "loss": 4.408048629760742,
+ "step": 1654
+ },
+ {
+ "epoch": 22.992139737991266,
+ "grad_norm": 0.015370816923677921,
+ "learning_rate": 0.0006,
+ "loss": 4.297451019287109,
+ "step": 1655
+ },
+ {
+ "epoch": 23.0,
+ "grad_norm": 0.01709144562482834,
+ "learning_rate": 0.0006,
+ "loss": 4.294776916503906,
+ "step": 1656
+ },
+ {
+ "epoch": 23.0,
+ "eval_loss": 4.678618907928467,
+ "eval_runtime": 56.1668,
+ "eval_samples_per_second": 43.478,
+ "eval_steps_per_second": 1.371,
+ "step": 1656
+ },
+ {
+ "epoch": 23.01397379912664,
+ "grad_norm": 0.01727290451526642,
+ "learning_rate": 0.0006,
+ "loss": 4.311701774597168,
+ "step": 1657
+ },
+ {
+ "epoch": 23.027947598253274,
+ "grad_norm": 0.020645759999752045,
+ "learning_rate": 0.0006,
+ "loss": 4.429897308349609,
+ "step": 1658
+ },
+ {
+ "epoch": 23.041921397379912,
+ "grad_norm": 0.02204226516187191,
+ "learning_rate": 0.0006,
+ "loss": 4.365151882171631,
+ "step": 1659
+ },
+ {
+ "epoch": 23.05589519650655,
+ "grad_norm": 0.023069269955158234,
+ "learning_rate": 0.0006,
+ "loss": 4.3710832595825195,
+ "step": 1660
+ },
+ {
+ "epoch": 23.069868995633186,
+ "grad_norm": 0.024151179939508438,
+ "learning_rate": 0.0006,
+ "loss": 4.403108596801758,
+ "step": 1661
+ },
+ {
+ "epoch": 23.083842794759825,
+ "grad_norm": 0.02511233650147915,
+ "learning_rate": 0.0006,
+ "loss": 4.414128303527832,
+ "step": 1662
+ },
+ {
+ "epoch": 23.097816593886463,
+ "grad_norm": 0.025802113115787506,
+ "learning_rate": 0.0006,
+ "loss": 4.388265609741211,
+ "step": 1663
+ },
+ {
+ "epoch": 23.111790393013102,
+ "grad_norm": 0.02377348765730858,
+ "learning_rate": 0.0006,
+ "loss": 4.357505798339844,
+ "step": 1664
+ },
+ {
+ "epoch": 23.125764192139737,
+ "grad_norm": 0.020876651629805565,
+ "learning_rate": 0.0006,
+ "loss": 4.23043966293335,
+ "step": 1665
+ },
+ {
+ "epoch": 23.139737991266376,
+ "grad_norm": 0.01959727331995964,
+ "learning_rate": 0.0006,
+ "loss": 4.280970096588135,
+ "step": 1666
+ },
+ {
+ "epoch": 23.153711790393015,
+ "grad_norm": 0.018678616732358932,
+ "learning_rate": 0.0006,
+ "loss": 4.346716403961182,
+ "step": 1667
+ },
+ {
+ "epoch": 23.16768558951965,
+ "grad_norm": 0.02108944021165371,
+ "learning_rate": 0.0006,
+ "loss": 4.277904510498047,
+ "step": 1668
+ },
+ {
+ "epoch": 23.18165938864629,
+ "grad_norm": 0.020221566781401634,
+ "learning_rate": 0.0006,
+ "loss": 4.367693901062012,
+ "step": 1669
+ },
+ {
+ "epoch": 23.195633187772927,
+ "grad_norm": 0.019599711522459984,
+ "learning_rate": 0.0006,
+ "loss": 4.257092475891113,
+ "step": 1670
+ },
+ {
+ "epoch": 23.209606986899562,
+ "grad_norm": 0.019833361729979515,
+ "learning_rate": 0.0006,
+ "loss": 4.540737628936768,
+ "step": 1671
+ },
+ {
+ "epoch": 23.2235807860262,
+ "grad_norm": 0.020352095365524292,
+ "learning_rate": 0.0006,
+ "loss": 4.360923767089844,
+ "step": 1672
+ },
+ {
+ "epoch": 23.23755458515284,
+ "grad_norm": 0.01939748041331768,
+ "learning_rate": 0.0006,
+ "loss": 4.393110275268555,
+ "step": 1673
+ },
+ {
+ "epoch": 23.251528384279474,
+ "grad_norm": 0.01884354278445244,
+ "learning_rate": 0.0006,
+ "loss": 4.383938789367676,
+ "step": 1674
+ },
+ {
+ "epoch": 23.265502183406113,
+ "grad_norm": 0.02008470520377159,
+ "learning_rate": 0.0006,
+ "loss": 4.346344947814941,
+ "step": 1675
+ },
+ {
+ "epoch": 23.27947598253275,
+ "grad_norm": 0.019065195694565773,
+ "learning_rate": 0.0006,
+ "loss": 4.36611795425415,
+ "step": 1676
+ },
+ {
+ "epoch": 23.29344978165939,
+ "grad_norm": 0.01938541606068611,
+ "learning_rate": 0.0006,
+ "loss": 4.4009599685668945,
+ "step": 1677
+ },
+ {
+ "epoch": 23.307423580786025,
+ "grad_norm": 0.019044723361730576,
+ "learning_rate": 0.0006,
+ "loss": 4.340169906616211,
+ "step": 1678
+ },
+ {
+ "epoch": 23.321397379912664,
+ "grad_norm": 0.018850378692150116,
+ "learning_rate": 0.0006,
+ "loss": 4.37674617767334,
+ "step": 1679
+ },
+ {
+ "epoch": 23.335371179039303,
+ "grad_norm": 0.019076095893979073,
+ "learning_rate": 0.0006,
+ "loss": 4.226874351501465,
+ "step": 1680
+ },
+ {
+ "epoch": 23.349344978165938,
+ "grad_norm": 0.017543038353323936,
+ "learning_rate": 0.0006,
+ "loss": 4.337751388549805,
+ "step": 1681
+ },
+ {
+ "epoch": 23.363318777292577,
+ "grad_norm": 0.017170218750834465,
+ "learning_rate": 0.0006,
+ "loss": 4.282512664794922,
+ "step": 1682
+ },
+ {
+ "epoch": 23.377292576419215,
+ "grad_norm": 0.01574796624481678,
+ "learning_rate": 0.0006,
+ "loss": 4.30638313293457,
+ "step": 1683
+ },
+ {
+ "epoch": 23.39126637554585,
+ "grad_norm": 0.015160846523940563,
+ "learning_rate": 0.0006,
+ "loss": 4.336716175079346,
+ "step": 1684
+ },
+ {
+ "epoch": 23.40524017467249,
+ "grad_norm": 0.01534526702016592,
+ "learning_rate": 0.0006,
+ "loss": 4.2174787521362305,
+ "step": 1685
+ },
+ {
+ "epoch": 23.419213973799128,
+ "grad_norm": 0.015289999544620514,
+ "learning_rate": 0.0006,
+ "loss": 4.272546291351318,
+ "step": 1686
+ },
+ {
+ "epoch": 23.433187772925763,
+ "grad_norm": 0.014894185587763786,
+ "learning_rate": 0.0006,
+ "loss": 4.296045303344727,
+ "step": 1687
+ },
+ {
+ "epoch": 23.4471615720524,
+ "grad_norm": 0.014869502745568752,
+ "learning_rate": 0.0006,
+ "loss": 4.425461769104004,
+ "step": 1688
+ },
+ {
+ "epoch": 23.46113537117904,
+ "grad_norm": 0.01474483497440815,
+ "learning_rate": 0.0006,
+ "loss": 4.361867427825928,
+ "step": 1689
+ },
+ {
+ "epoch": 23.475109170305675,
+ "grad_norm": 0.015299245715141296,
+ "learning_rate": 0.0006,
+ "loss": 4.364523410797119,
+ "step": 1690
+ },
+ {
+ "epoch": 23.489082969432314,
+ "grad_norm": 0.015201340429484844,
+ "learning_rate": 0.0006,
+ "loss": 4.285182952880859,
+ "step": 1691
+ },
+ {
+ "epoch": 23.503056768558952,
+ "grad_norm": 0.016057293862104416,
+ "learning_rate": 0.0006,
+ "loss": 4.284873962402344,
+ "step": 1692
+ },
+ {
+ "epoch": 23.51703056768559,
+ "grad_norm": 0.015871062874794006,
+ "learning_rate": 0.0006,
+ "loss": 4.326470851898193,
+ "step": 1693
+ },
+ {
+ "epoch": 23.531004366812226,
+ "grad_norm": 0.01586288772523403,
+ "learning_rate": 0.0006,
+ "loss": 4.324883460998535,
+ "step": 1694
+ },
+ {
+ "epoch": 23.544978165938865,
+ "grad_norm": 0.016092827543616295,
+ "learning_rate": 0.0006,
+ "loss": 4.3258867263793945,
+ "step": 1695
+ },
+ {
+ "epoch": 23.558951965065503,
+ "grad_norm": 0.01582016795873642,
+ "learning_rate": 0.0006,
+ "loss": 4.358892440795898,
+ "step": 1696
+ },
+ {
+ "epoch": 23.57292576419214,
+ "grad_norm": 0.01688452623784542,
+ "learning_rate": 0.0006,
+ "loss": 4.280842304229736,
+ "step": 1697
+ },
+ {
+ "epoch": 23.586899563318777,
+ "grad_norm": 0.015865886583924294,
+ "learning_rate": 0.0006,
+ "loss": 4.245567321777344,
+ "step": 1698
+ },
+ {
+ "epoch": 23.600873362445416,
+ "grad_norm": 0.015303662046790123,
+ "learning_rate": 0.0006,
+ "loss": 4.383112907409668,
+ "step": 1699
+ },
+ {
+ "epoch": 23.61484716157205,
+ "grad_norm": 0.015649516135454178,
+ "learning_rate": 0.0006,
+ "loss": 4.519082069396973,
+ "step": 1700
+ },
+ {
+ "epoch": 23.62882096069869,
+ "grad_norm": 0.0167120099067688,
+ "learning_rate": 0.0006,
+ "loss": 4.447847366333008,
+ "step": 1701
+ },
+ {
+ "epoch": 23.64279475982533,
+ "grad_norm": 0.016368716955184937,
+ "learning_rate": 0.0006,
+ "loss": 4.434186935424805,
+ "step": 1702
+ },
+ {
+ "epoch": 23.656768558951963,
+ "grad_norm": 0.017379503697156906,
+ "learning_rate": 0.0006,
+ "loss": 4.317594528198242,
+ "step": 1703
+ },
+ {
+ "epoch": 23.670742358078602,
+ "grad_norm": 0.018573811277747154,
+ "learning_rate": 0.0006,
+ "loss": 4.205793857574463,
+ "step": 1704
+ },
+ {
+ "epoch": 23.68471615720524,
+ "grad_norm": 0.01749090477824211,
+ "learning_rate": 0.0006,
+ "loss": 4.3985700607299805,
+ "step": 1705
+ },
+ {
+ "epoch": 23.69868995633188,
+ "grad_norm": 0.0165668074041605,
+ "learning_rate": 0.0006,
+ "loss": 4.329226970672607,
+ "step": 1706
+ },
+ {
+ "epoch": 23.712663755458514,
+ "grad_norm": 0.016913846135139465,
+ "learning_rate": 0.0006,
+ "loss": 4.388485908508301,
+ "step": 1707
+ },
+ {
+ "epoch": 23.726637554585153,
+ "grad_norm": 0.016813859343528748,
+ "learning_rate": 0.0006,
+ "loss": 4.3395915031433105,
+ "step": 1708
+ },
+ {
+ "epoch": 23.74061135371179,
+ "grad_norm": 0.01676975190639496,
+ "learning_rate": 0.0006,
+ "loss": 4.254745960235596,
+ "step": 1709
+ },
+ {
+ "epoch": 23.754585152838427,
+ "grad_norm": 0.016370350494980812,
+ "learning_rate": 0.0006,
+ "loss": 4.2612223625183105,
+ "step": 1710
+ },
+ {
+ "epoch": 23.768558951965066,
+ "grad_norm": 0.01696198247373104,
+ "learning_rate": 0.0006,
+ "loss": 4.327524662017822,
+ "step": 1711
+ },
+ {
+ "epoch": 23.782532751091704,
+ "grad_norm": 0.0189108457416296,
+ "learning_rate": 0.0006,
+ "loss": 4.325839519500732,
+ "step": 1712
+ },
+ {
+ "epoch": 23.79650655021834,
+ "grad_norm": 0.01862351782619953,
+ "learning_rate": 0.0006,
+ "loss": 4.333662509918213,
+ "step": 1713
+ },
+ {
+ "epoch": 23.810480349344978,
+ "grad_norm": 0.01655014418065548,
+ "learning_rate": 0.0006,
+ "loss": 4.39305305480957,
+ "step": 1714
+ },
+ {
+ "epoch": 23.824454148471617,
+ "grad_norm": 0.015899376943707466,
+ "learning_rate": 0.0006,
+ "loss": 4.390053749084473,
+ "step": 1715
+ },
+ {
+ "epoch": 23.83842794759825,
+ "grad_norm": 0.016519056633114815,
+ "learning_rate": 0.0006,
+ "loss": 4.255903244018555,
+ "step": 1716
+ },
+ {
+ "epoch": 23.85240174672489,
+ "grad_norm": 0.0167539119720459,
+ "learning_rate": 0.0006,
+ "loss": 4.255727767944336,
+ "step": 1717
+ },
+ {
+ "epoch": 23.86637554585153,
+ "grad_norm": 0.0171353816986084,
+ "learning_rate": 0.0006,
+ "loss": 4.376740455627441,
+ "step": 1718
+ },
+ {
+ "epoch": 23.880349344978168,
+ "grad_norm": 0.016719117760658264,
+ "learning_rate": 0.0006,
+ "loss": 4.3376688957214355,
+ "step": 1719
+ },
+ {
+ "epoch": 23.894323144104803,
+ "grad_norm": 0.01561494916677475,
+ "learning_rate": 0.0006,
+ "loss": 4.3345746994018555,
+ "step": 1720
+ },
+ {
+ "epoch": 23.90829694323144,
+ "grad_norm": 0.016303174197673798,
+ "learning_rate": 0.0006,
+ "loss": 4.332371234893799,
+ "step": 1721
+ },
+ {
+ "epoch": 23.92227074235808,
+ "grad_norm": 0.016722865402698517,
+ "learning_rate": 0.0006,
+ "loss": 4.267125129699707,
+ "step": 1722
+ },
+ {
+ "epoch": 23.936244541484715,
+ "grad_norm": 0.017072124406695366,
+ "learning_rate": 0.0006,
+ "loss": 4.322751998901367,
+ "step": 1723
+ },
+ {
+ "epoch": 23.950218340611354,
+ "grad_norm": 0.01699932850897312,
+ "learning_rate": 0.0006,
+ "loss": 4.378422737121582,
+ "step": 1724
+ },
+ {
+ "epoch": 23.964192139737992,
+ "grad_norm": 0.01742720976471901,
+ "learning_rate": 0.0006,
+ "loss": 4.345083236694336,
+ "step": 1725
+ },
+ {
+ "epoch": 23.978165938864628,
+ "grad_norm": 0.018797501921653748,
+ "learning_rate": 0.0006,
+ "loss": 4.290443420410156,
+ "step": 1726
+ },
+ {
+ "epoch": 23.992139737991266,
+ "grad_norm": 0.018090683966875076,
+ "learning_rate": 0.0006,
+ "loss": 4.377580642700195,
+ "step": 1727
+ },
+ {
+ "epoch": 24.0,
+ "grad_norm": 0.017305459827184677,
+ "learning_rate": 0.0006,
+ "loss": 4.364426612854004,
+ "step": 1728
+ },
+ {
+ "epoch": 24.0,
+ "eval_loss": 4.647762298583984,
+ "eval_runtime": 56.5402,
+ "eval_samples_per_second": 43.191,
+ "eval_steps_per_second": 1.362,
+ "step": 1728
+ },
+ {
+ "epoch": 24.01397379912664,
+ "grad_norm": 0.016493607312440872,
+ "learning_rate": 0.0006,
+ "loss": 4.252255916595459,
+ "step": 1729
+ },
+ {
+ "epoch": 24.027947598253274,
+ "grad_norm": 0.016529450193047523,
+ "learning_rate": 0.0006,
+ "loss": 4.352746963500977,
+ "step": 1730
+ },
+ {
+ "epoch": 24.041921397379912,
+ "grad_norm": 0.016940191388130188,
+ "learning_rate": 0.0006,
+ "loss": 4.301870822906494,
+ "step": 1731
+ },
+ {
+ "epoch": 24.05589519650655,
+ "grad_norm": 0.01752558909356594,
+ "learning_rate": 0.0006,
+ "loss": 4.2762451171875,
+ "step": 1732
+ },
+ {
+ "epoch": 24.069868995633186,
+ "grad_norm": 0.017112884670495987,
+ "learning_rate": 0.0006,
+ "loss": 4.2659149169921875,
+ "step": 1733
+ },
+ {
+ "epoch": 24.083842794759825,
+ "grad_norm": 0.01900586113333702,
+ "learning_rate": 0.0006,
+ "loss": 4.306058883666992,
+ "step": 1734
+ },
+ {
+ "epoch": 24.097816593886463,
+ "grad_norm": 0.019447680562734604,
+ "learning_rate": 0.0006,
+ "loss": 4.450128555297852,
+ "step": 1735
+ },
+ {
+ "epoch": 24.111790393013102,
+ "grad_norm": 0.01948186755180359,
+ "learning_rate": 0.0006,
+ "loss": 4.202869415283203,
+ "step": 1736
+ },
+ {
+ "epoch": 24.125764192139737,
+ "grad_norm": 0.01716383546590805,
+ "learning_rate": 0.0006,
+ "loss": 4.318876266479492,
+ "step": 1737
+ },
+ {
+ "epoch": 24.139737991266376,
+ "grad_norm": 0.017049988731741905,
+ "learning_rate": 0.0006,
+ "loss": 4.237238883972168,
+ "step": 1738
+ },
+ {
+ "epoch": 24.153711790393015,
+ "grad_norm": 0.015169923193752766,
+ "learning_rate": 0.0006,
+ "loss": 4.289426803588867,
+ "step": 1739
+ },
+ {
+ "epoch": 24.16768558951965,
+ "grad_norm": 0.014760667458176613,
+ "learning_rate": 0.0006,
+ "loss": 4.356924057006836,
+ "step": 1740
+ },
+ {
+ "epoch": 24.18165938864629,
+ "grad_norm": 0.015127750113606453,
+ "learning_rate": 0.0006,
+ "loss": 4.142916679382324,
+ "step": 1741
+ },
+ {
+ "epoch": 24.195633187772927,
+ "grad_norm": 0.01549555640667677,
+ "learning_rate": 0.0006,
+ "loss": 4.345563888549805,
+ "step": 1742
+ },
+ {
+ "epoch": 24.209606986899562,
+ "grad_norm": 0.016393091529607773,
+ "learning_rate": 0.0006,
+ "loss": 4.417495250701904,
+ "step": 1743
+ },
+ {
+ "epoch": 24.2235807860262,
+ "grad_norm": 0.01785266026854515,
+ "learning_rate": 0.0006,
+ "loss": 4.3553547859191895,
+ "step": 1744
+ },
+ {
+ "epoch": 24.23755458515284,
+ "grad_norm": 0.017574617639183998,
+ "learning_rate": 0.0006,
+ "loss": 4.241024494171143,
+ "step": 1745
+ },
+ {
+ "epoch": 24.251528384279474,
+ "grad_norm": 0.019060306251049042,
+ "learning_rate": 0.0006,
+ "loss": 4.3685150146484375,
+ "step": 1746
+ },
+ {
+ "epoch": 24.265502183406113,
+ "grad_norm": 0.0214633010327816,
+ "learning_rate": 0.0006,
+ "loss": 4.323866844177246,
+ "step": 1747
+ },
+ {
+ "epoch": 24.27947598253275,
+ "grad_norm": 0.021055003628134727,
+ "learning_rate": 0.0006,
+ "loss": 4.278757095336914,
+ "step": 1748
+ },
+ {
+ "epoch": 24.29344978165939,
+ "grad_norm": 0.018618572503328323,
+ "learning_rate": 0.0006,
+ "loss": 4.28849983215332,
+ "step": 1749
+ },
+ {
+ "epoch": 24.307423580786025,
+ "grad_norm": 0.018666435033082962,
+ "learning_rate": 0.0006,
+ "loss": 4.368590354919434,
+ "step": 1750
+ },
+ {
+ "epoch": 24.321397379912664,
+ "grad_norm": 0.020626146346330643,
+ "learning_rate": 0.0006,
+ "loss": 4.2951836585998535,
+ "step": 1751
+ },
+ {
+ "epoch": 24.335371179039303,
+ "grad_norm": 0.01929759792983532,
+ "learning_rate": 0.0006,
+ "loss": 4.3664703369140625,
+ "step": 1752
+ },
+ {
+ "epoch": 24.349344978165938,
+ "grad_norm": 0.020956002175807953,
+ "learning_rate": 0.0006,
+ "loss": 4.345308303833008,
+ "step": 1753
+ },
+ {
+ "epoch": 24.363318777292577,
+ "grad_norm": 0.019227957352995872,
+ "learning_rate": 0.0006,
+ "loss": 4.349849224090576,
+ "step": 1754
+ },
+ {
+ "epoch": 24.377292576419215,
+ "grad_norm": 0.015970097854733467,
+ "learning_rate": 0.0006,
+ "loss": 4.223012447357178,
+ "step": 1755
+ },
+ {
+ "epoch": 24.39126637554585,
+ "grad_norm": 0.015951646491885185,
+ "learning_rate": 0.0006,
+ "loss": 4.236420631408691,
+ "step": 1756
+ },
+ {
+ "epoch": 24.40524017467249,
+ "grad_norm": 0.016032515093684196,
+ "learning_rate": 0.0006,
+ "loss": 4.226899147033691,
+ "step": 1757
+ },
+ {
+ "epoch": 24.419213973799128,
+ "grad_norm": 0.01573132909834385,
+ "learning_rate": 0.0006,
+ "loss": 4.315916061401367,
+ "step": 1758
+ },
+ {
+ "epoch": 24.433187772925763,
+ "grad_norm": 0.015367005951702595,
+ "learning_rate": 0.0006,
+ "loss": 4.358939170837402,
+ "step": 1759
+ },
+ {
+ "epoch": 24.4471615720524,
+ "grad_norm": 0.01669907011091709,
+ "learning_rate": 0.0006,
+ "loss": 4.232147216796875,
+ "step": 1760
+ },
+ {
+ "epoch": 24.46113537117904,
+ "grad_norm": 0.018708152696490288,
+ "learning_rate": 0.0006,
+ "loss": 4.329852104187012,
+ "step": 1761
+ },
+ {
+ "epoch": 24.475109170305675,
+ "grad_norm": 0.019035987555980682,
+ "learning_rate": 0.0006,
+ "loss": 4.260340213775635,
+ "step": 1762
+ },
+ {
+ "epoch": 24.489082969432314,
+ "grad_norm": 0.01832144521176815,
+ "learning_rate": 0.0006,
+ "loss": 4.359543800354004,
+ "step": 1763
+ },
+ {
+ "epoch": 24.503056768558952,
+ "grad_norm": 0.015313294716179371,
+ "learning_rate": 0.0006,
+ "loss": 4.279088020324707,
+ "step": 1764
+ },
+ {
+ "epoch": 24.51703056768559,
+ "grad_norm": 0.015803923830389977,
+ "learning_rate": 0.0006,
+ "loss": 4.230682849884033,
+ "step": 1765
+ },
+ {
+ "epoch": 24.531004366812226,
+ "grad_norm": 0.016748374328017235,
+ "learning_rate": 0.0006,
+ "loss": 4.349661827087402,
+ "step": 1766
+ },
+ {
+ "epoch": 24.544978165938865,
+ "grad_norm": 0.015732428058981895,
+ "learning_rate": 0.0006,
+ "loss": 4.372441291809082,
+ "step": 1767
+ },
+ {
+ "epoch": 24.558951965065503,
+ "grad_norm": 0.016564829275012016,
+ "learning_rate": 0.0006,
+ "loss": 4.323090553283691,
+ "step": 1768
+ },
+ {
+ "epoch": 24.57292576419214,
+ "grad_norm": 0.01603010483086109,
+ "learning_rate": 0.0006,
+ "loss": 4.347898006439209,
+ "step": 1769
+ },
+ {
+ "epoch": 24.586899563318777,
+ "grad_norm": 0.01674809865653515,
+ "learning_rate": 0.0006,
+ "loss": 4.269165515899658,
+ "step": 1770
+ },
+ {
+ "epoch": 24.600873362445416,
+ "grad_norm": 0.017333930358290672,
+ "learning_rate": 0.0006,
+ "loss": 4.353298187255859,
+ "step": 1771
+ },
+ {
+ "epoch": 24.61484716157205,
+ "grad_norm": 0.017519580200314522,
+ "learning_rate": 0.0006,
+ "loss": 4.265199661254883,
+ "step": 1772
+ },
+ {
+ "epoch": 24.62882096069869,
+ "grad_norm": 0.017675306648015976,
+ "learning_rate": 0.0006,
+ "loss": 4.322596549987793,
+ "step": 1773
+ },
+ {
+ "epoch": 24.64279475982533,
+ "grad_norm": 0.017166534438729286,
+ "learning_rate": 0.0006,
+ "loss": 4.413900375366211,
+ "step": 1774
+ },
+ {
+ "epoch": 24.656768558951963,
+ "grad_norm": 0.016362346708774567,
+ "learning_rate": 0.0006,
+ "loss": 4.307236671447754,
+ "step": 1775
+ },
+ {
+ "epoch": 24.670742358078602,
+ "grad_norm": 0.016086872667074203,
+ "learning_rate": 0.0006,
+ "loss": 4.235316276550293,
+ "step": 1776
+ },
+ {
+ "epoch": 24.68471615720524,
+ "grad_norm": 0.01689537614583969,
+ "learning_rate": 0.0006,
+ "loss": 4.390591621398926,
+ "step": 1777
+ },
+ {
+ "epoch": 24.69868995633188,
+ "grad_norm": 0.016836725175380707,
+ "learning_rate": 0.0006,
+ "loss": 4.2990031242370605,
+ "step": 1778
+ },
+ {
+ "epoch": 24.712663755458514,
+ "grad_norm": 0.015627335757017136,
+ "learning_rate": 0.0006,
+ "loss": 4.166745185852051,
+ "step": 1779
+ },
+ {
+ "epoch": 24.726637554585153,
+ "grad_norm": 0.015007016249001026,
+ "learning_rate": 0.0006,
+ "loss": 4.295273303985596,
+ "step": 1780
+ },
+ {
+ "epoch": 24.74061135371179,
+ "grad_norm": 0.016084209084510803,
+ "learning_rate": 0.0006,
+ "loss": 4.342424392700195,
+ "step": 1781
+ },
+ {
+ "epoch": 24.754585152838427,
+ "grad_norm": 0.01595073565840721,
+ "learning_rate": 0.0006,
+ "loss": 4.3060102462768555,
+ "step": 1782
+ },
+ {
+ "epoch": 24.768558951965066,
+ "grad_norm": 0.016113461926579475,
+ "learning_rate": 0.0006,
+ "loss": 4.267908096313477,
+ "step": 1783
+ },
+ {
+ "epoch": 24.782532751091704,
+ "grad_norm": 0.01564556173980236,
+ "learning_rate": 0.0006,
+ "loss": 4.324642658233643,
+ "step": 1784
+ },
+ {
+ "epoch": 24.79650655021834,
+ "grad_norm": 0.01575268618762493,
+ "learning_rate": 0.0006,
+ "loss": 4.328181743621826,
+ "step": 1785
+ },
+ {
+ "epoch": 24.810480349344978,
+ "grad_norm": 0.016830824315547943,
+ "learning_rate": 0.0006,
+ "loss": 4.247906684875488,
+ "step": 1786
+ },
+ {
+ "epoch": 24.824454148471617,
+ "grad_norm": 0.017989547923207283,
+ "learning_rate": 0.0006,
+ "loss": 4.228281021118164,
+ "step": 1787
+ },
+ {
+ "epoch": 24.83842794759825,
+ "grad_norm": 0.016783086583018303,
+ "learning_rate": 0.0006,
+ "loss": 4.2880964279174805,
+ "step": 1788
+ },
+ {
+ "epoch": 24.85240174672489,
+ "grad_norm": 0.016467561945319176,
+ "learning_rate": 0.0006,
+ "loss": 4.36118745803833,
+ "step": 1789
+ },
+ {
+ "epoch": 24.86637554585153,
+ "grad_norm": 0.01780523732304573,
+ "learning_rate": 0.0006,
+ "loss": 4.172321319580078,
+ "step": 1790
+ },
+ {
+ "epoch": 24.880349344978168,
+ "grad_norm": 0.017102031037211418,
+ "learning_rate": 0.0006,
+ "loss": 4.352826118469238,
+ "step": 1791
+ },
+ {
+ "epoch": 24.894323144104803,
+ "grad_norm": 0.016409218311309814,
+ "learning_rate": 0.0006,
+ "loss": 4.212162017822266,
+ "step": 1792
+ },
+ {
+ "epoch": 24.90829694323144,
+ "grad_norm": 0.018660584464669228,
+ "learning_rate": 0.0006,
+ "loss": 4.346684455871582,
+ "step": 1793
+ },
+ {
+ "epoch": 24.92227074235808,
+ "grad_norm": 0.017389440909028053,
+ "learning_rate": 0.0006,
+ "loss": 4.197913646697998,
+ "step": 1794
+ },
+ {
+ "epoch": 24.936244541484715,
+ "grad_norm": 0.01596485823392868,
+ "learning_rate": 0.0006,
+ "loss": 4.370617866516113,
+ "step": 1795
+ },
+ {
+ "epoch": 24.950218340611354,
+ "grad_norm": 0.01709776371717453,
+ "learning_rate": 0.0006,
+ "loss": 4.313847064971924,
+ "step": 1796
+ },
+ {
+ "epoch": 24.964192139737992,
+ "grad_norm": 0.016571134328842163,
+ "learning_rate": 0.0006,
+ "loss": 4.1895856857299805,
+ "step": 1797
+ },
+ {
+ "epoch": 24.978165938864628,
+ "grad_norm": 0.016851048916578293,
+ "learning_rate": 0.0006,
+ "loss": 4.350772857666016,
+ "step": 1798
+ },
+ {
+ "epoch": 24.992139737991266,
+ "grad_norm": 0.017007440328598022,
+ "learning_rate": 0.0006,
+ "loss": 4.399008750915527,
+ "step": 1799
+ },
+ {
+ "epoch": 25.0,
+ "grad_norm": 0.01791212148964405,
+ "learning_rate": 0.0006,
+ "loss": 4.306121826171875,
+ "step": 1800
+ },
+ {
+ "epoch": 25.0,
+ "eval_loss": 4.636417865753174,
+ "eval_runtime": 56.8296,
+ "eval_samples_per_second": 42.971,
+ "eval_steps_per_second": 1.355,
+ "step": 1800
+ },
+ {
+ "epoch": 25.01397379912664,
+ "grad_norm": 0.01729617640376091,
+ "learning_rate": 0.0006,
+ "loss": 4.257018089294434,
+ "step": 1801
+ },
+ {
+ "epoch": 25.027947598253274,
+ "grad_norm": 0.017466643825173378,
+ "learning_rate": 0.0006,
+ "loss": 4.307605266571045,
+ "step": 1802
+ },
+ {
+ "epoch": 25.041921397379912,
+ "grad_norm": 0.018614279106259346,
+ "learning_rate": 0.0006,
+ "loss": 4.360141277313232,
+ "step": 1803
+ },
+ {
+ "epoch": 25.05589519650655,
+ "grad_norm": 0.02010771445930004,
+ "learning_rate": 0.0006,
+ "loss": 4.196428298950195,
+ "step": 1804
+ },
+ {
+ "epoch": 25.069868995633186,
+ "grad_norm": 0.02090228535234928,
+ "learning_rate": 0.0006,
+ "loss": 4.288947105407715,
+ "step": 1805
+ },
+ {
+ "epoch": 25.083842794759825,
+ "grad_norm": 0.02050255797803402,
+ "learning_rate": 0.0006,
+ "loss": 4.259582042694092,
+ "step": 1806
+ },
+ {
+ "epoch": 25.097816593886463,
+ "grad_norm": 0.021844089031219482,
+ "learning_rate": 0.0006,
+ "loss": 4.405138969421387,
+ "step": 1807
+ },
+ {
+ "epoch": 25.111790393013102,
+ "grad_norm": 0.022287018597126007,
+ "learning_rate": 0.0006,
+ "loss": 4.333793640136719,
+ "step": 1808
+ },
+ {
+ "epoch": 25.125764192139737,
+ "grad_norm": 0.02288329415023327,
+ "learning_rate": 0.0006,
+ "loss": 4.3910980224609375,
+ "step": 1809
+ },
+ {
+ "epoch": 25.139737991266376,
+ "grad_norm": 0.02373199723660946,
+ "learning_rate": 0.0006,
+ "loss": 4.32921028137207,
+ "step": 1810
+ },
+ {
+ "epoch": 25.153711790393015,
+ "grad_norm": 0.024960234761238098,
+ "learning_rate": 0.0006,
+ "loss": 4.295609474182129,
+ "step": 1811
+ },
+ {
+ "epoch": 25.16768558951965,
+ "grad_norm": 0.026743892580270767,
+ "learning_rate": 0.0006,
+ "loss": 4.315308094024658,
+ "step": 1812
+ },
+ {
+ "epoch": 25.18165938864629,
+ "grad_norm": 0.022138185799121857,
+ "learning_rate": 0.0006,
+ "loss": 4.360208511352539,
+ "step": 1813
+ },
+ {
+ "epoch": 25.195633187772927,
+ "grad_norm": 0.022120574489235878,
+ "learning_rate": 0.0006,
+ "loss": 4.1973772048950195,
+ "step": 1814
+ },
+ {
+ "epoch": 25.209606986899562,
+ "grad_norm": 0.02230502851307392,
+ "learning_rate": 0.0006,
+ "loss": 4.285558700561523,
+ "step": 1815
+ },
+ {
+ "epoch": 25.2235807860262,
+ "grad_norm": 0.022301986813545227,
+ "learning_rate": 0.0006,
+ "loss": 4.238343238830566,
+ "step": 1816
+ },
+ {
+ "epoch": 25.23755458515284,
+ "grad_norm": 0.020050447434186935,
+ "learning_rate": 0.0006,
+ "loss": 4.298235893249512,
+ "step": 1817
+ },
+ {
+ "epoch": 25.251528384279474,
+ "grad_norm": 0.01987724006175995,
+ "learning_rate": 0.0006,
+ "loss": 4.301384449005127,
+ "step": 1818
+ },
+ {
+ "epoch": 25.265502183406113,
+ "grad_norm": 0.020440855994820595,
+ "learning_rate": 0.0006,
+ "loss": 4.269218444824219,
+ "step": 1819
+ },
+ {
+ "epoch": 25.27947598253275,
+ "grad_norm": 0.01891893707215786,
+ "learning_rate": 0.0006,
+ "loss": 4.2401957511901855,
+ "step": 1820
+ },
+ {
+ "epoch": 25.29344978165939,
+ "grad_norm": 0.01903688907623291,
+ "learning_rate": 0.0006,
+ "loss": 4.139670372009277,
+ "step": 1821
+ },
+ {
+ "epoch": 25.307423580786025,
+ "grad_norm": 0.018856938928365707,
+ "learning_rate": 0.0006,
+ "loss": 4.233214378356934,
+ "step": 1822
+ },
+ {
+ "epoch": 25.321397379912664,
+ "grad_norm": 0.017791323363780975,
+ "learning_rate": 0.0006,
+ "loss": 4.2342023849487305,
+ "step": 1823
+ },
+ {
+ "epoch": 25.335371179039303,
+ "grad_norm": 0.017304565757513046,
+ "learning_rate": 0.0006,
+ "loss": 4.355198860168457,
+ "step": 1824
+ },
+ {
+ "epoch": 25.349344978165938,
+ "grad_norm": 0.01576206088066101,
+ "learning_rate": 0.0006,
+ "loss": 4.114147186279297,
+ "step": 1825
+ },
+ {
+ "epoch": 25.363318777292577,
+ "grad_norm": 0.015105154365301132,
+ "learning_rate": 0.0006,
+ "loss": 4.313795566558838,
+ "step": 1826
+ },
+ {
+ "epoch": 25.377292576419215,
+ "grad_norm": 0.01563587784767151,
+ "learning_rate": 0.0006,
+ "loss": 4.308282852172852,
+ "step": 1827
+ },
+ {
+ "epoch": 25.39126637554585,
+ "grad_norm": 0.015980906784534454,
+ "learning_rate": 0.0006,
+ "loss": 4.345149040222168,
+ "step": 1828
+ },
+ {
+ "epoch": 25.40524017467249,
+ "grad_norm": 0.01669284701347351,
+ "learning_rate": 0.0006,
+ "loss": 4.237611293792725,
+ "step": 1829
+ },
+ {
+ "epoch": 25.419213973799128,
+ "grad_norm": 0.016202064231038094,
+ "learning_rate": 0.0006,
+ "loss": 4.3072309494018555,
+ "step": 1830
+ },
+ {
+ "epoch": 25.433187772925763,
+ "grad_norm": 0.01478662807494402,
+ "learning_rate": 0.0006,
+ "loss": 4.276215076446533,
+ "step": 1831
+ },
+ {
+ "epoch": 25.4471615720524,
+ "grad_norm": 0.015168731100857258,
+ "learning_rate": 0.0006,
+ "loss": 4.37509822845459,
+ "step": 1832
+ },
+ {
+ "epoch": 25.46113537117904,
+ "grad_norm": 0.01648980937898159,
+ "learning_rate": 0.0006,
+ "loss": 4.349167346954346,
+ "step": 1833
+ },
+ {
+ "epoch": 25.475109170305675,
+ "grad_norm": 0.01708398200571537,
+ "learning_rate": 0.0006,
+ "loss": 4.346580505371094,
+ "step": 1834
+ },
+ {
+ "epoch": 25.489082969432314,
+ "grad_norm": 0.017370784655213356,
+ "learning_rate": 0.0006,
+ "loss": 4.299466133117676,
+ "step": 1835
+ },
+ {
+ "epoch": 25.503056768558952,
+ "grad_norm": 0.016248228028416634,
+ "learning_rate": 0.0006,
+ "loss": 4.184610366821289,
+ "step": 1836
+ },
+ {
+ "epoch": 25.51703056768559,
+ "grad_norm": 0.017106913030147552,
+ "learning_rate": 0.0006,
+ "loss": 4.366856098175049,
+ "step": 1837
+ },
+ {
+ "epoch": 25.531004366812226,
+ "grad_norm": 0.016071965917944908,
+ "learning_rate": 0.0006,
+ "loss": 4.297691345214844,
+ "step": 1838
+ },
+ {
+ "epoch": 25.544978165938865,
+ "grad_norm": 0.016451986506581306,
+ "learning_rate": 0.0006,
+ "loss": 4.229983329772949,
+ "step": 1839
+ },
+ {
+ "epoch": 25.558951965065503,
+ "grad_norm": 0.016195081174373627,
+ "learning_rate": 0.0006,
+ "loss": 4.259893417358398,
+ "step": 1840
+ },
+ {
+ "epoch": 25.57292576419214,
+ "grad_norm": 0.0165175162255764,
+ "learning_rate": 0.0006,
+ "loss": 4.3348236083984375,
+ "step": 1841
+ },
+ {
+ "epoch": 25.586899563318777,
+ "grad_norm": 0.01651517115533352,
+ "learning_rate": 0.0006,
+ "loss": 4.319541931152344,
+ "step": 1842
+ },
+ {
+ "epoch": 25.600873362445416,
+ "grad_norm": 0.016143690794706345,
+ "learning_rate": 0.0006,
+ "loss": 4.345652103424072,
+ "step": 1843
+ },
+ {
+ "epoch": 25.61484716157205,
+ "grad_norm": 0.015140696428716183,
+ "learning_rate": 0.0006,
+ "loss": 4.212047576904297,
+ "step": 1844
+ },
+ {
+ "epoch": 25.62882096069869,
+ "grad_norm": 0.014637443237006664,
+ "learning_rate": 0.0006,
+ "loss": 4.187453269958496,
+ "step": 1845
+ },
+ {
+ "epoch": 25.64279475982533,
+ "grad_norm": 0.014585614204406738,
+ "learning_rate": 0.0006,
+ "loss": 4.253936767578125,
+ "step": 1846
+ },
+ {
+ "epoch": 25.656768558951963,
+ "grad_norm": 0.015158289112150669,
+ "learning_rate": 0.0006,
+ "loss": 4.34612512588501,
+ "step": 1847
+ },
+ {
+ "epoch": 25.670742358078602,
+ "grad_norm": 0.015202849172055721,
+ "learning_rate": 0.0006,
+ "loss": 4.270634651184082,
+ "step": 1848
+ },
+ {
+ "epoch": 25.68471615720524,
+ "grad_norm": 0.015363575890660286,
+ "learning_rate": 0.0006,
+ "loss": 4.271080493927002,
+ "step": 1849
+ },
+ {
+ "epoch": 25.69868995633188,
+ "grad_norm": 0.014360226690769196,
+ "learning_rate": 0.0006,
+ "loss": 4.375033855438232,
+ "step": 1850
+ },
+ {
+ "epoch": 25.712663755458514,
+ "grad_norm": 0.014263181947171688,
+ "learning_rate": 0.0006,
+ "loss": 4.424138069152832,
+ "step": 1851
+ },
+ {
+ "epoch": 25.726637554585153,
+ "grad_norm": 0.014398688450455666,
+ "learning_rate": 0.0006,
+ "loss": 4.410154819488525,
+ "step": 1852
+ },
+ {
+ "epoch": 25.74061135371179,
+ "grad_norm": 0.016131488606333733,
+ "learning_rate": 0.0006,
+ "loss": 4.324014663696289,
+ "step": 1853
+ },
+ {
+ "epoch": 25.754585152838427,
+ "grad_norm": 0.017568735405802727,
+ "learning_rate": 0.0006,
+ "loss": 4.367861747741699,
+ "step": 1854
+ },
+ {
+ "epoch": 25.768558951965066,
+ "grad_norm": 0.017907511442899704,
+ "learning_rate": 0.0006,
+ "loss": 4.352965354919434,
+ "step": 1855
+ },
+ {
+ "epoch": 25.782532751091704,
+ "grad_norm": 0.016918016597628593,
+ "learning_rate": 0.0006,
+ "loss": 4.252397537231445,
+ "step": 1856
+ },
+ {
+ "epoch": 25.79650655021834,
+ "grad_norm": 0.017274610698223114,
+ "learning_rate": 0.0006,
+ "loss": 4.342309951782227,
+ "step": 1857
+ },
+ {
+ "epoch": 25.810480349344978,
+ "grad_norm": 0.017879825085401535,
+ "learning_rate": 0.0006,
+ "loss": 4.330987453460693,
+ "step": 1858
+ },
+ {
+ "epoch": 25.824454148471617,
+ "grad_norm": 0.01736099272966385,
+ "learning_rate": 0.0006,
+ "loss": 4.307255744934082,
+ "step": 1859
+ },
+ {
+ "epoch": 25.83842794759825,
+ "grad_norm": 0.016575824469327927,
+ "learning_rate": 0.0006,
+ "loss": 4.262016773223877,
+ "step": 1860
+ },
+ {
+ "epoch": 25.85240174672489,
+ "grad_norm": 0.015181140042841434,
+ "learning_rate": 0.0006,
+ "loss": 4.291863441467285,
+ "step": 1861
+ },
+ {
+ "epoch": 25.86637554585153,
+ "grad_norm": 0.015965687111020088,
+ "learning_rate": 0.0006,
+ "loss": 4.424836158752441,
+ "step": 1862
+ },
+ {
+ "epoch": 25.880349344978168,
+ "grad_norm": 0.01687219925224781,
+ "learning_rate": 0.0006,
+ "loss": 4.343531131744385,
+ "step": 1863
+ },
+ {
+ "epoch": 25.894323144104803,
+ "grad_norm": 0.015115504153072834,
+ "learning_rate": 0.0006,
+ "loss": 4.357808589935303,
+ "step": 1864
+ },
+ {
+ "epoch": 25.90829694323144,
+ "grad_norm": 0.014150998555123806,
+ "learning_rate": 0.0006,
+ "loss": 4.324550151824951,
+ "step": 1865
+ },
+ {
+ "epoch": 25.92227074235808,
+ "grad_norm": 0.014989510178565979,
+ "learning_rate": 0.0006,
+ "loss": 4.380428314208984,
+ "step": 1866
+ },
+ {
+ "epoch": 25.936244541484715,
+ "grad_norm": 0.01527960691601038,
+ "learning_rate": 0.0006,
+ "loss": 4.399374961853027,
+ "step": 1867
+ },
+ {
+ "epoch": 25.950218340611354,
+ "grad_norm": 0.015434633940458298,
+ "learning_rate": 0.0006,
+ "loss": 4.260984420776367,
+ "step": 1868
+ },
+ {
+ "epoch": 25.964192139737992,
+ "grad_norm": 0.015370228327810764,
+ "learning_rate": 0.0006,
+ "loss": 4.3340253829956055,
+ "step": 1869
+ },
+ {
+ "epoch": 25.978165938864628,
+ "grad_norm": 0.015348542481660843,
+ "learning_rate": 0.0006,
+ "loss": 4.369132995605469,
+ "step": 1870
+ },
+ {
+ "epoch": 25.992139737991266,
+ "grad_norm": 0.015354936942458153,
+ "learning_rate": 0.0006,
+ "loss": 4.381937026977539,
+ "step": 1871
+ },
+ {
+ "epoch": 26.0,
+ "grad_norm": 0.01796272210776806,
+ "learning_rate": 0.0006,
+ "loss": 4.286768436431885,
+ "step": 1872
+ },
+ {
+ "epoch": 26.0,
+ "eval_loss": 4.576984405517578,
+ "eval_runtime": 57.2091,
+ "eval_samples_per_second": 42.685,
+ "eval_steps_per_second": 1.346,
+ "step": 1872
+ },
+ {
+ "epoch": 26.01397379912664,
+ "grad_norm": 0.01748845726251602,
+ "learning_rate": 0.0006,
+ "loss": 4.337504863739014,
+ "step": 1873
+ },
+ {
+ "epoch": 26.027947598253274,
+ "grad_norm": 0.019244296476244926,
+ "learning_rate": 0.0006,
+ "loss": 4.264280796051025,
+ "step": 1874
+ },
+ {
+ "epoch": 26.041921397379912,
+ "grad_norm": 0.0215692650526762,
+ "learning_rate": 0.0006,
+ "loss": 4.350830078125,
+ "step": 1875
+ },
+ {
+ "epoch": 26.05589519650655,
+ "grad_norm": 0.022189252078533173,
+ "learning_rate": 0.0006,
+ "loss": 4.299098968505859,
+ "step": 1876
+ },
+ {
+ "epoch": 26.069868995633186,
+ "grad_norm": 0.022384043782949448,
+ "learning_rate": 0.0006,
+ "loss": 4.1185173988342285,
+ "step": 1877
+ },
+ {
+ "epoch": 26.083842794759825,
+ "grad_norm": 0.02076014317572117,
+ "learning_rate": 0.0006,
+ "loss": 4.297612190246582,
+ "step": 1878
+ },
+ {
+ "epoch": 26.097816593886463,
+ "grad_norm": 0.020843051373958588,
+ "learning_rate": 0.0006,
+ "loss": 4.294098854064941,
+ "step": 1879
+ },
+ {
+ "epoch": 26.111790393013102,
+ "grad_norm": 0.020654456689953804,
+ "learning_rate": 0.0006,
+ "loss": 4.267856597900391,
+ "step": 1880
+ },
+ {
+ "epoch": 26.125764192139737,
+ "grad_norm": 0.01860472559928894,
+ "learning_rate": 0.0006,
+ "loss": 4.284232139587402,
+ "step": 1881
+ },
+ {
+ "epoch": 26.139737991266376,
+ "grad_norm": 0.018719421699643135,
+ "learning_rate": 0.0006,
+ "loss": 4.277214527130127,
+ "step": 1882
+ },
+ {
+ "epoch": 26.153711790393015,
+ "grad_norm": 0.020903829485177994,
+ "learning_rate": 0.0006,
+ "loss": 4.254947662353516,
+ "step": 1883
+ },
+ {
+ "epoch": 26.16768558951965,
+ "grad_norm": 0.02020149491727352,
+ "learning_rate": 0.0006,
+ "loss": 4.2526936531066895,
+ "step": 1884
+ },
+ {
+ "epoch": 26.18165938864629,
+ "grad_norm": 0.0180392786860466,
+ "learning_rate": 0.0006,
+ "loss": 4.309922695159912,
+ "step": 1885
+ },
+ {
+ "epoch": 26.195633187772927,
+ "grad_norm": 0.018163125962018967,
+ "learning_rate": 0.0006,
+ "loss": 4.13860559463501,
+ "step": 1886
+ },
+ {
+ "epoch": 26.209606986899562,
+ "grad_norm": 0.01853291131556034,
+ "learning_rate": 0.0006,
+ "loss": 4.190926551818848,
+ "step": 1887
+ },
+ {
+ "epoch": 26.2235807860262,
+ "grad_norm": 0.01757677271962166,
+ "learning_rate": 0.0006,
+ "loss": 4.163999557495117,
+ "step": 1888
+ },
+ {
+ "epoch": 26.23755458515284,
+ "grad_norm": 0.01767992228269577,
+ "learning_rate": 0.0006,
+ "loss": 4.184430122375488,
+ "step": 1889
+ },
+ {
+ "epoch": 26.251528384279474,
+ "grad_norm": 0.018139973282814026,
+ "learning_rate": 0.0006,
+ "loss": 4.300766944885254,
+ "step": 1890
+ },
+ {
+ "epoch": 26.265502183406113,
+ "grad_norm": 0.020184461027383804,
+ "learning_rate": 0.0006,
+ "loss": 4.232537269592285,
+ "step": 1891
+ },
+ {
+ "epoch": 26.27947598253275,
+ "grad_norm": 0.02028856985270977,
+ "learning_rate": 0.0006,
+ "loss": 4.324883460998535,
+ "step": 1892
+ },
+ {
+ "epoch": 26.29344978165939,
+ "grad_norm": 0.017120616510510445,
+ "learning_rate": 0.0006,
+ "loss": 4.23493766784668,
+ "step": 1893
+ },
+ {
+ "epoch": 26.307423580786025,
+ "grad_norm": 0.016202978789806366,
+ "learning_rate": 0.0006,
+ "loss": 4.197602272033691,
+ "step": 1894
+ },
+ {
+ "epoch": 26.321397379912664,
+ "grad_norm": 0.017231421545147896,
+ "learning_rate": 0.0006,
+ "loss": 4.302001953125,
+ "step": 1895
+ },
+ {
+ "epoch": 26.335371179039303,
+ "grad_norm": 0.01725655607879162,
+ "learning_rate": 0.0006,
+ "loss": 4.356800079345703,
+ "step": 1896
+ },
+ {
+ "epoch": 26.349344978165938,
+ "grad_norm": 0.016657443717122078,
+ "learning_rate": 0.0006,
+ "loss": 4.266307830810547,
+ "step": 1897
+ },
+ {
+ "epoch": 26.363318777292577,
+ "grad_norm": 0.01616556942462921,
+ "learning_rate": 0.0006,
+ "loss": 4.307868003845215,
+ "step": 1898
+ },
+ {
+ "epoch": 26.377292576419215,
+ "grad_norm": 0.016621306538581848,
+ "learning_rate": 0.0006,
+ "loss": 4.172905921936035,
+ "step": 1899
+ },
+ {
+ "epoch": 26.39126637554585,
+ "grad_norm": 0.016243206337094307,
+ "learning_rate": 0.0006,
+ "loss": 4.162896156311035,
+ "step": 1900
+ },
+ {
+ "epoch": 26.40524017467249,
+ "grad_norm": 0.016663808375597,
+ "learning_rate": 0.0006,
+ "loss": 4.258232116699219,
+ "step": 1901
+ },
+ {
+ "epoch": 26.419213973799128,
+ "grad_norm": 0.01677478663623333,
+ "learning_rate": 0.0006,
+ "loss": 4.336090087890625,
+ "step": 1902
+ },
+ {
+ "epoch": 26.433187772925763,
+ "grad_norm": 0.01641303487122059,
+ "learning_rate": 0.0006,
+ "loss": 4.191476821899414,
+ "step": 1903
+ },
+ {
+ "epoch": 26.4471615720524,
+ "grad_norm": 0.016700323671102524,
+ "learning_rate": 0.0006,
+ "loss": 4.299427509307861,
+ "step": 1904
+ },
+ {
+ "epoch": 26.46113537117904,
+ "grad_norm": 0.017595035955309868,
+ "learning_rate": 0.0006,
+ "loss": 4.312828063964844,
+ "step": 1905
+ },
+ {
+ "epoch": 26.475109170305675,
+ "grad_norm": 0.020300284028053284,
+ "learning_rate": 0.0006,
+ "loss": 4.28331184387207,
+ "step": 1906
+ },
+ {
+ "epoch": 26.489082969432314,
+ "grad_norm": 0.02198321744799614,
+ "learning_rate": 0.0006,
+ "loss": 4.3280839920043945,
+ "step": 1907
+ },
+ {
+ "epoch": 26.503056768558952,
+ "grad_norm": 0.02003200724720955,
+ "learning_rate": 0.0006,
+ "loss": 4.325623989105225,
+ "step": 1908
+ },
+ {
+ "epoch": 26.51703056768559,
+ "grad_norm": 0.01870287023484707,
+ "learning_rate": 0.0006,
+ "loss": 4.251203536987305,
+ "step": 1909
+ },
+ {
+ "epoch": 26.531004366812226,
+ "grad_norm": 0.01835448481142521,
+ "learning_rate": 0.0006,
+ "loss": 4.335144996643066,
+ "step": 1910
+ },
+ {
+ "epoch": 26.544978165938865,
+ "grad_norm": 0.02046327292919159,
+ "learning_rate": 0.0006,
+ "loss": 4.238023281097412,
+ "step": 1911
+ },
+ {
+ "epoch": 26.558951965065503,
+ "grad_norm": 0.019771311432123184,
+ "learning_rate": 0.0006,
+ "loss": 4.353594779968262,
+ "step": 1912
+ },
+ {
+ "epoch": 26.57292576419214,
+ "grad_norm": 0.018090544268488884,
+ "learning_rate": 0.0006,
+ "loss": 4.304343223571777,
+ "step": 1913
+ },
+ {
+ "epoch": 26.586899563318777,
+ "grad_norm": 0.017489777877926826,
+ "learning_rate": 0.0006,
+ "loss": 4.204375267028809,
+ "step": 1914
+ },
+ {
+ "epoch": 26.600873362445416,
+ "grad_norm": 0.016785891726613045,
+ "learning_rate": 0.0006,
+ "loss": 4.286669731140137,
+ "step": 1915
+ },
+ {
+ "epoch": 26.61484716157205,
+ "grad_norm": 0.016382789239287376,
+ "learning_rate": 0.0006,
+ "loss": 4.347662448883057,
+ "step": 1916
+ },
+ {
+ "epoch": 26.62882096069869,
+ "grad_norm": 0.016581876203417778,
+ "learning_rate": 0.0006,
+ "loss": 4.2716193199157715,
+ "step": 1917
+ },
+ {
+ "epoch": 26.64279475982533,
+ "grad_norm": 0.015627184882760048,
+ "learning_rate": 0.0006,
+ "loss": 4.311489105224609,
+ "step": 1918
+ },
+ {
+ "epoch": 26.656768558951963,
+ "grad_norm": 0.015908483415842056,
+ "learning_rate": 0.0006,
+ "loss": 4.262099742889404,
+ "step": 1919
+ },
+ {
+ "epoch": 26.670742358078602,
+ "grad_norm": 0.015576236881315708,
+ "learning_rate": 0.0006,
+ "loss": 4.275871276855469,
+ "step": 1920
+ },
+ {
+ "epoch": 26.68471615720524,
+ "grad_norm": 0.014621064998209476,
+ "learning_rate": 0.0006,
+ "loss": 4.184260845184326,
+ "step": 1921
+ },
+ {
+ "epoch": 26.69868995633188,
+ "grad_norm": 0.014797811396420002,
+ "learning_rate": 0.0006,
+ "loss": 4.3412628173828125,
+ "step": 1922
+ },
+ {
+ "epoch": 26.712663755458514,
+ "grad_norm": 0.015610570088028908,
+ "learning_rate": 0.0006,
+ "loss": 4.181094646453857,
+ "step": 1923
+ },
+ {
+ "epoch": 26.726637554585153,
+ "grad_norm": 0.016048265621066093,
+ "learning_rate": 0.0006,
+ "loss": 4.365290641784668,
+ "step": 1924
+ },
+ {
+ "epoch": 26.74061135371179,
+ "grad_norm": 0.016604071483016014,
+ "learning_rate": 0.0006,
+ "loss": 4.2628912925720215,
+ "step": 1925
+ },
+ {
+ "epoch": 26.754585152838427,
+ "grad_norm": 0.017044221982359886,
+ "learning_rate": 0.0006,
+ "loss": 4.378512382507324,
+ "step": 1926
+ },
+ {
+ "epoch": 26.768558951965066,
+ "grad_norm": 0.01746384985744953,
+ "learning_rate": 0.0006,
+ "loss": 4.3888773918151855,
+ "step": 1927
+ },
+ {
+ "epoch": 26.782532751091704,
+ "grad_norm": 0.01572711206972599,
+ "learning_rate": 0.0006,
+ "loss": 4.189935684204102,
+ "step": 1928
+ },
+ {
+ "epoch": 26.79650655021834,
+ "grad_norm": 0.014976629987359047,
+ "learning_rate": 0.0006,
+ "loss": 4.280525207519531,
+ "step": 1929
+ },
+ {
+ "epoch": 26.810480349344978,
+ "grad_norm": 0.014804664999246597,
+ "learning_rate": 0.0006,
+ "loss": 4.270630836486816,
+ "step": 1930
+ },
+ {
+ "epoch": 26.824454148471617,
+ "grad_norm": 0.014641920104622841,
+ "learning_rate": 0.0006,
+ "loss": 4.319505214691162,
+ "step": 1931
+ },
+ {
+ "epoch": 26.83842794759825,
+ "grad_norm": 0.014928505755960941,
+ "learning_rate": 0.0006,
+ "loss": 4.180147647857666,
+ "step": 1932
+ },
+ {
+ "epoch": 26.85240174672489,
+ "grad_norm": 0.016130365431308746,
+ "learning_rate": 0.0006,
+ "loss": 4.260660648345947,
+ "step": 1933
+ },
+ {
+ "epoch": 26.86637554585153,
+ "grad_norm": 0.015642493963241577,
+ "learning_rate": 0.0006,
+ "loss": 4.304064750671387,
+ "step": 1934
+ },
+ {
+ "epoch": 26.880349344978168,
+ "grad_norm": 0.015988217666745186,
+ "learning_rate": 0.0006,
+ "loss": 4.2548017501831055,
+ "step": 1935
+ },
+ {
+ "epoch": 26.894323144104803,
+ "grad_norm": 0.016501398757100105,
+ "learning_rate": 0.0006,
+ "loss": 4.260058879852295,
+ "step": 1936
+ },
+ {
+ "epoch": 26.90829694323144,
+ "grad_norm": 0.01645040698349476,
+ "learning_rate": 0.0006,
+ "loss": 4.483081340789795,
+ "step": 1937
+ },
+ {
+ "epoch": 26.92227074235808,
+ "grad_norm": 0.016140863299369812,
+ "learning_rate": 0.0006,
+ "loss": 4.1776838302612305,
+ "step": 1938
+ },
+ {
+ "epoch": 26.936244541484715,
+ "grad_norm": 0.016681019216775894,
+ "learning_rate": 0.0006,
+ "loss": 4.224747657775879,
+ "step": 1939
+ },
+ {
+ "epoch": 26.950218340611354,
+ "grad_norm": 0.01650378853082657,
+ "learning_rate": 0.0006,
+ "loss": 4.305703163146973,
+ "step": 1940
+ },
+ {
+ "epoch": 26.964192139737992,
+ "grad_norm": 0.017051683738827705,
+ "learning_rate": 0.0006,
+ "loss": 4.285835266113281,
+ "step": 1941
+ },
+ {
+ "epoch": 26.978165938864628,
+ "grad_norm": 0.016894327476620674,
+ "learning_rate": 0.0006,
+ "loss": 4.318901538848877,
+ "step": 1942
+ },
+ {
+ "epoch": 26.992139737991266,
+ "grad_norm": 0.017313100397586823,
+ "learning_rate": 0.0006,
+ "loss": 4.281434535980225,
+ "step": 1943
+ },
+ {
+ "epoch": 27.0,
+ "grad_norm": 0.018450884148478508,
+ "learning_rate": 0.0006,
+ "loss": 4.335104942321777,
+ "step": 1944
+ },
+ {
+ "epoch": 27.0,
+ "eval_loss": 4.640687465667725,
+ "eval_runtime": 56.9624,
+ "eval_samples_per_second": 42.87,
+ "eval_steps_per_second": 1.352,
+ "step": 1944
+ },
+ {
+ "epoch": 27.01397379912664,
+ "grad_norm": 0.017755698412656784,
+ "learning_rate": 0.0006,
+ "loss": 4.278824329376221,
+ "step": 1945
+ },
+ {
+ "epoch": 27.027947598253274,
+ "grad_norm": 0.01724310778081417,
+ "learning_rate": 0.0006,
+ "loss": 4.311049461364746,
+ "step": 1946
+ },
+ {
+ "epoch": 27.041921397379912,
+ "grad_norm": 0.01690123789012432,
+ "learning_rate": 0.0006,
+ "loss": 4.261994361877441,
+ "step": 1947
+ },
+ {
+ "epoch": 27.05589519650655,
+ "grad_norm": 0.018406696617603302,
+ "learning_rate": 0.0006,
+ "loss": 4.1925153732299805,
+ "step": 1948
+ },
+ {
+ "epoch": 27.069868995633186,
+ "grad_norm": 0.019945867359638214,
+ "learning_rate": 0.0006,
+ "loss": 4.108537197113037,
+ "step": 1949
+ },
+ {
+ "epoch": 27.083842794759825,
+ "grad_norm": 0.020316628739237785,
+ "learning_rate": 0.0006,
+ "loss": 4.209043025970459,
+ "step": 1950
+ },
+ {
+ "epoch": 27.097816593886463,
+ "grad_norm": 0.019690370187163353,
+ "learning_rate": 0.0006,
+ "loss": 4.195789337158203,
+ "step": 1951
+ },
+ {
+ "epoch": 27.111790393013102,
+ "grad_norm": 0.01785232499241829,
+ "learning_rate": 0.0006,
+ "loss": 4.21466064453125,
+ "step": 1952
+ },
+ {
+ "epoch": 27.125764192139737,
+ "grad_norm": 0.0164767038077116,
+ "learning_rate": 0.0006,
+ "loss": 4.270247459411621,
+ "step": 1953
+ },
+ {
+ "epoch": 27.139737991266376,
+ "grad_norm": 0.018008455634117126,
+ "learning_rate": 0.0006,
+ "loss": 4.201042175292969,
+ "step": 1954
+ },
+ {
+ "epoch": 27.153711790393015,
+ "grad_norm": 0.01807340793311596,
+ "learning_rate": 0.0006,
+ "loss": 4.25289249420166,
+ "step": 1955
+ },
+ {
+ "epoch": 27.16768558951965,
+ "grad_norm": 0.016246909275650978,
+ "learning_rate": 0.0006,
+ "loss": 4.116283416748047,
+ "step": 1956
+ },
+ {
+ "epoch": 27.18165938864629,
+ "grad_norm": 0.017545776441693306,
+ "learning_rate": 0.0006,
+ "loss": 4.316999435424805,
+ "step": 1957
+ },
+ {
+ "epoch": 27.195633187772927,
+ "grad_norm": 0.017915375530719757,
+ "learning_rate": 0.0006,
+ "loss": 4.253859519958496,
+ "step": 1958
+ },
+ {
+ "epoch": 27.209606986899562,
+ "grad_norm": 0.018112223595380783,
+ "learning_rate": 0.0006,
+ "loss": 4.2541022300720215,
+ "step": 1959
+ },
+ {
+ "epoch": 27.2235807860262,
+ "grad_norm": 0.01846139505505562,
+ "learning_rate": 0.0006,
+ "loss": 4.20438289642334,
+ "step": 1960
+ },
+ {
+ "epoch": 27.23755458515284,
+ "grad_norm": 0.01943897269666195,
+ "learning_rate": 0.0006,
+ "loss": 4.2799577713012695,
+ "step": 1961
+ },
+ {
+ "epoch": 27.251528384279474,
+ "grad_norm": 0.018253348767757416,
+ "learning_rate": 0.0006,
+ "loss": 4.1720499992370605,
+ "step": 1962
+ },
+ {
+ "epoch": 27.265502183406113,
+ "grad_norm": 0.018292246386408806,
+ "learning_rate": 0.0006,
+ "loss": 4.28231143951416,
+ "step": 1963
+ },
+ {
+ "epoch": 27.27947598253275,
+ "grad_norm": 0.01943155936896801,
+ "learning_rate": 0.0006,
+ "loss": 4.314976692199707,
+ "step": 1964
+ },
+ {
+ "epoch": 27.29344978165939,
+ "grad_norm": 0.019064994528889656,
+ "learning_rate": 0.0006,
+ "loss": 4.25328254699707,
+ "step": 1965
+ },
+ {
+ "epoch": 27.307423580786025,
+ "grad_norm": 0.017423273995518684,
+ "learning_rate": 0.0006,
+ "loss": 4.195992469787598,
+ "step": 1966
+ },
+ {
+ "epoch": 27.321397379912664,
+ "grad_norm": 0.01677924580872059,
+ "learning_rate": 0.0006,
+ "loss": 4.260043144226074,
+ "step": 1967
+ },
+ {
+ "epoch": 27.335371179039303,
+ "grad_norm": 0.01720673404633999,
+ "learning_rate": 0.0006,
+ "loss": 4.222780704498291,
+ "step": 1968
+ },
+ {
+ "epoch": 27.349344978165938,
+ "grad_norm": 0.017300743609666824,
+ "learning_rate": 0.0006,
+ "loss": 4.109524250030518,
+ "step": 1969
+ },
+ {
+ "epoch": 27.363318777292577,
+ "grad_norm": 0.016065871343016624,
+ "learning_rate": 0.0006,
+ "loss": 4.36108922958374,
+ "step": 1970
+ },
+ {
+ "epoch": 27.377292576419215,
+ "grad_norm": 0.016626616939902306,
+ "learning_rate": 0.0006,
+ "loss": 4.220630645751953,
+ "step": 1971
+ },
+ {
+ "epoch": 27.39126637554585,
+ "grad_norm": 0.016482602804899216,
+ "learning_rate": 0.0006,
+ "loss": 4.344797134399414,
+ "step": 1972
+ },
+ {
+ "epoch": 27.40524017467249,
+ "grad_norm": 0.01727953553199768,
+ "learning_rate": 0.0006,
+ "loss": 4.296645164489746,
+ "step": 1973
+ },
+ {
+ "epoch": 27.419213973799128,
+ "grad_norm": 0.01522731315344572,
+ "learning_rate": 0.0006,
+ "loss": 4.223751068115234,
+ "step": 1974
+ },
+ {
+ "epoch": 27.433187772925763,
+ "grad_norm": 0.01581776700913906,
+ "learning_rate": 0.0006,
+ "loss": 4.351379871368408,
+ "step": 1975
+ },
+ {
+ "epoch": 27.4471615720524,
+ "grad_norm": 0.016394605860114098,
+ "learning_rate": 0.0006,
+ "loss": 4.2412919998168945,
+ "step": 1976
+ },
+ {
+ "epoch": 27.46113537117904,
+ "grad_norm": 0.01751169003546238,
+ "learning_rate": 0.0006,
+ "loss": 4.257606506347656,
+ "step": 1977
+ },
+ {
+ "epoch": 27.475109170305675,
+ "grad_norm": 0.017452310770750046,
+ "learning_rate": 0.0006,
+ "loss": 4.380134582519531,
+ "step": 1978
+ },
+ {
+ "epoch": 27.489082969432314,
+ "grad_norm": 0.017741898074746132,
+ "learning_rate": 0.0006,
+ "loss": 4.279632568359375,
+ "step": 1979
+ },
+ {
+ "epoch": 27.503056768558952,
+ "grad_norm": 0.017680590972304344,
+ "learning_rate": 0.0006,
+ "loss": 4.273406028747559,
+ "step": 1980
+ },
+ {
+ "epoch": 27.51703056768559,
+ "grad_norm": 0.01864718459546566,
+ "learning_rate": 0.0006,
+ "loss": 4.278583526611328,
+ "step": 1981
+ },
+ {
+ "epoch": 27.531004366812226,
+ "grad_norm": 0.019558541476726532,
+ "learning_rate": 0.0006,
+ "loss": 4.211919784545898,
+ "step": 1982
+ },
+ {
+ "epoch": 27.544978165938865,
+ "grad_norm": 0.019467400386929512,
+ "learning_rate": 0.0006,
+ "loss": 4.276822566986084,
+ "step": 1983
+ },
+ {
+ "epoch": 27.558951965065503,
+ "grad_norm": 0.01756172813475132,
+ "learning_rate": 0.0006,
+ "loss": 4.165379047393799,
+ "step": 1984
+ },
+ {
+ "epoch": 27.57292576419214,
+ "grad_norm": 0.017458263784646988,
+ "learning_rate": 0.0006,
+ "loss": 4.2992401123046875,
+ "step": 1985
+ },
+ {
+ "epoch": 27.586899563318777,
+ "grad_norm": 0.018089665099978447,
+ "learning_rate": 0.0006,
+ "loss": 4.243378639221191,
+ "step": 1986
+ },
+ {
+ "epoch": 27.600873362445416,
+ "grad_norm": 0.019535524770617485,
+ "learning_rate": 0.0006,
+ "loss": 4.321477890014648,
+ "step": 1987
+ },
+ {
+ "epoch": 27.61484716157205,
+ "grad_norm": 0.01997970975935459,
+ "learning_rate": 0.0006,
+ "loss": 4.123082637786865,
+ "step": 1988
+ },
+ {
+ "epoch": 27.62882096069869,
+ "grad_norm": 0.01913067139685154,
+ "learning_rate": 0.0006,
+ "loss": 4.289896488189697,
+ "step": 1989
+ },
+ {
+ "epoch": 27.64279475982533,
+ "grad_norm": 0.0169806070625782,
+ "learning_rate": 0.0006,
+ "loss": 4.247004985809326,
+ "step": 1990
+ },
+ {
+ "epoch": 27.656768558951963,
+ "grad_norm": 0.017315007746219635,
+ "learning_rate": 0.0006,
+ "loss": 4.2095441818237305,
+ "step": 1991
+ },
+ {
+ "epoch": 27.670742358078602,
+ "grad_norm": 0.015098759904503822,
+ "learning_rate": 0.0006,
+ "loss": 4.145560264587402,
+ "step": 1992
+ },
+ {
+ "epoch": 27.68471615720524,
+ "grad_norm": 0.016240514814853668,
+ "learning_rate": 0.0006,
+ "loss": 4.254927635192871,
+ "step": 1993
+ },
+ {
+ "epoch": 27.69868995633188,
+ "grad_norm": 0.01621108129620552,
+ "learning_rate": 0.0006,
+ "loss": 4.219725608825684,
+ "step": 1994
+ },
+ {
+ "epoch": 27.712663755458514,
+ "grad_norm": 0.015191009268164635,
+ "learning_rate": 0.0006,
+ "loss": 4.194345474243164,
+ "step": 1995
+ },
+ {
+ "epoch": 27.726637554585153,
+ "grad_norm": 0.014387983828783035,
+ "learning_rate": 0.0006,
+ "loss": 4.211467742919922,
+ "step": 1996
+ },
+ {
+ "epoch": 27.74061135371179,
+ "grad_norm": 0.015460561029613018,
+ "learning_rate": 0.0006,
+ "loss": 4.358700752258301,
+ "step": 1997
+ },
+ {
+ "epoch": 27.754585152838427,
+ "grad_norm": 0.015100076794624329,
+ "learning_rate": 0.0006,
+ "loss": 4.171223163604736,
+ "step": 1998
+ },
+ {
+ "epoch": 27.768558951965066,
+ "grad_norm": 0.015607891604304314,
+ "learning_rate": 0.0006,
+ "loss": 4.217874526977539,
+ "step": 1999
+ },
+ {
+ "epoch": 27.782532751091704,
+ "grad_norm": 0.015650689601898193,
+ "learning_rate": 0.0006,
+ "loss": 4.280780792236328,
+ "step": 2000
+ },
+ {
+ "epoch": 27.79650655021834,
+ "grad_norm": 0.01580994389951229,
+ "learning_rate": 0.0006,
+ "loss": 4.159212112426758,
+ "step": 2001
+ },
+ {
+ "epoch": 27.810480349344978,
+ "grad_norm": 0.016607413068413734,
+ "learning_rate": 0.0006,
+ "loss": 4.238853454589844,
+ "step": 2002
+ },
+ {
+ "epoch": 27.824454148471617,
+ "grad_norm": 0.017383860424160957,
+ "learning_rate": 0.0006,
+ "loss": 4.233944892883301,
+ "step": 2003
+ },
+ {
+ "epoch": 27.83842794759825,
+ "grad_norm": 0.014966659247875214,
+ "learning_rate": 0.0006,
+ "loss": 4.167392730712891,
+ "step": 2004
+ },
+ {
+ "epoch": 27.85240174672489,
+ "grad_norm": 0.014589878730475903,
+ "learning_rate": 0.0006,
+ "loss": 4.2390899658203125,
+ "step": 2005
+ },
+ {
+ "epoch": 27.86637554585153,
+ "grad_norm": 0.017179450020194054,
+ "learning_rate": 0.0006,
+ "loss": 4.365785598754883,
+ "step": 2006
+ },
+ {
+ "epoch": 27.880349344978168,
+ "grad_norm": 0.01893901452422142,
+ "learning_rate": 0.0006,
+ "loss": 4.289430141448975,
+ "step": 2007
+ },
+ {
+ "epoch": 27.894323144104803,
+ "grad_norm": 0.018619712442159653,
+ "learning_rate": 0.0006,
+ "loss": 4.310665130615234,
+ "step": 2008
+ },
+ {
+ "epoch": 27.90829694323144,
+ "grad_norm": 0.019180385395884514,
+ "learning_rate": 0.0006,
+ "loss": 4.27454137802124,
+ "step": 2009
+ },
+ {
+ "epoch": 27.92227074235808,
+ "grad_norm": 0.017811523750424385,
+ "learning_rate": 0.0006,
+ "loss": 4.312562465667725,
+ "step": 2010
+ },
+ {
+ "epoch": 27.936244541484715,
+ "grad_norm": 0.016583573073148727,
+ "learning_rate": 0.0006,
+ "loss": 4.202253341674805,
+ "step": 2011
+ },
+ {
+ "epoch": 27.950218340611354,
+ "grad_norm": 0.01658615842461586,
+ "learning_rate": 0.0006,
+ "loss": 4.183433532714844,
+ "step": 2012
+ },
+ {
+ "epoch": 27.964192139737992,
+ "grad_norm": 0.015327401459217072,
+ "learning_rate": 0.0006,
+ "loss": 4.2125139236450195,
+ "step": 2013
+ },
+ {
+ "epoch": 27.978165938864628,
+ "grad_norm": 0.015004601329565048,
+ "learning_rate": 0.0006,
+ "loss": 4.257392883300781,
+ "step": 2014
+ },
+ {
+ "epoch": 27.992139737991266,
+ "grad_norm": 0.015259161591529846,
+ "learning_rate": 0.0006,
+ "loss": 4.241818428039551,
+ "step": 2015
+ },
+ {
+ "epoch": 28.0,
+ "grad_norm": 0.016808776184916496,
+ "learning_rate": 0.0006,
+ "loss": 4.146241188049316,
+ "step": 2016
+ },
+ {
+ "epoch": 28.0,
+ "eval_loss": 4.663090705871582,
+ "eval_runtime": 56.4559,
+ "eval_samples_per_second": 43.255,
+ "eval_steps_per_second": 1.364,
+ "step": 2016
+ },
+ {
+ "epoch": 28.01397379912664,
+ "grad_norm": 0.015784382820129395,
+ "learning_rate": 0.0006,
+ "loss": 4.044191360473633,
+ "step": 2017
+ },
+ {
+ "epoch": 28.027947598253274,
+ "grad_norm": 0.015636710450053215,
+ "learning_rate": 0.0006,
+ "loss": 4.128754615783691,
+ "step": 2018
+ },
+ {
+ "epoch": 28.041921397379912,
+ "grad_norm": 0.015364975668489933,
+ "learning_rate": 0.0006,
+ "loss": 4.191835403442383,
+ "step": 2019
+ },
+ {
+ "epoch": 28.05589519650655,
+ "grad_norm": 0.015015700832009315,
+ "learning_rate": 0.0006,
+ "loss": 4.120022773742676,
+ "step": 2020
+ },
+ {
+ "epoch": 28.069868995633186,
+ "grad_norm": 0.016666799783706665,
+ "learning_rate": 0.0006,
+ "loss": 4.33906364440918,
+ "step": 2021
+ },
+ {
+ "epoch": 28.083842794759825,
+ "grad_norm": 0.01780104823410511,
+ "learning_rate": 0.0006,
+ "loss": 4.279458999633789,
+ "step": 2022
+ },
+ {
+ "epoch": 28.097816593886463,
+ "grad_norm": 0.01870882883667946,
+ "learning_rate": 0.0006,
+ "loss": 4.34728479385376,
+ "step": 2023
+ },
+ {
+ "epoch": 28.111790393013102,
+ "grad_norm": 0.018406381830573082,
+ "learning_rate": 0.0006,
+ "loss": 4.211311340332031,
+ "step": 2024
+ },
+ {
+ "epoch": 28.125764192139737,
+ "grad_norm": 0.017833448946475983,
+ "learning_rate": 0.0006,
+ "loss": 4.223254203796387,
+ "step": 2025
+ },
+ {
+ "epoch": 28.139737991266376,
+ "grad_norm": 0.017552688717842102,
+ "learning_rate": 0.0006,
+ "loss": 4.208823204040527,
+ "step": 2026
+ },
+ {
+ "epoch": 28.153711790393015,
+ "grad_norm": 0.018500229343771935,
+ "learning_rate": 0.0006,
+ "loss": 4.198025226593018,
+ "step": 2027
+ },
+ {
+ "epoch": 28.16768558951965,
+ "grad_norm": 0.01782156340777874,
+ "learning_rate": 0.0006,
+ "loss": 4.179978370666504,
+ "step": 2028
+ },
+ {
+ "epoch": 28.18165938864629,
+ "grad_norm": 0.019580967724323273,
+ "learning_rate": 0.0006,
+ "loss": 4.291114807128906,
+ "step": 2029
+ },
+ {
+ "epoch": 28.195633187772927,
+ "grad_norm": 0.020804036408662796,
+ "learning_rate": 0.0006,
+ "loss": 4.220607280731201,
+ "step": 2030
+ },
+ {
+ "epoch": 28.209606986899562,
+ "grad_norm": 0.019595693796873093,
+ "learning_rate": 0.0006,
+ "loss": 4.2905731201171875,
+ "step": 2031
+ },
+ {
+ "epoch": 28.2235807860262,
+ "grad_norm": 0.019247086718678474,
+ "learning_rate": 0.0006,
+ "loss": 4.2467360496521,
+ "step": 2032
+ },
+ {
+ "epoch": 28.23755458515284,
+ "grad_norm": 0.01958036608994007,
+ "learning_rate": 0.0006,
+ "loss": 4.286137580871582,
+ "step": 2033
+ },
+ {
+ "epoch": 28.251528384279474,
+ "grad_norm": 0.021354753524065018,
+ "learning_rate": 0.0006,
+ "loss": 4.234755039215088,
+ "step": 2034
+ },
+ {
+ "epoch": 28.265502183406113,
+ "grad_norm": 0.02206375077366829,
+ "learning_rate": 0.0006,
+ "loss": 4.149641036987305,
+ "step": 2035
+ },
+ {
+ "epoch": 28.27947598253275,
+ "grad_norm": 0.02201310358941555,
+ "learning_rate": 0.0006,
+ "loss": 4.1928629875183105,
+ "step": 2036
+ },
+ {
+ "epoch": 28.29344978165939,
+ "grad_norm": 0.01961950585246086,
+ "learning_rate": 0.0006,
+ "loss": 4.211404323577881,
+ "step": 2037
+ },
+ {
+ "epoch": 28.307423580786025,
+ "grad_norm": 0.019211795181035995,
+ "learning_rate": 0.0006,
+ "loss": 4.200976848602295,
+ "step": 2038
+ },
+ {
+ "epoch": 28.321397379912664,
+ "grad_norm": 0.01779370754957199,
+ "learning_rate": 0.0006,
+ "loss": 4.249148845672607,
+ "step": 2039
+ },
+ {
+ "epoch": 28.335371179039303,
+ "grad_norm": 0.018635908141732216,
+ "learning_rate": 0.0006,
+ "loss": 4.11806058883667,
+ "step": 2040
+ },
+ {
+ "epoch": 28.349344978165938,
+ "grad_norm": 0.019792694598436356,
+ "learning_rate": 0.0006,
+ "loss": 4.335224151611328,
+ "step": 2041
+ },
+ {
+ "epoch": 28.363318777292577,
+ "grad_norm": 0.021731717512011528,
+ "learning_rate": 0.0006,
+ "loss": 4.302778244018555,
+ "step": 2042
+ },
+ {
+ "epoch": 28.377292576419215,
+ "grad_norm": 0.020823447033762932,
+ "learning_rate": 0.0006,
+ "loss": 4.260610580444336,
+ "step": 2043
+ },
+ {
+ "epoch": 28.39126637554585,
+ "grad_norm": 0.019385412335395813,
+ "learning_rate": 0.0006,
+ "loss": 4.2247724533081055,
+ "step": 2044
+ },
+ {
+ "epoch": 28.40524017467249,
+ "grad_norm": 0.019539108499884605,
+ "learning_rate": 0.0006,
+ "loss": 4.236655235290527,
+ "step": 2045
+ },
+ {
+ "epoch": 28.419213973799128,
+ "grad_norm": 0.01942690648138523,
+ "learning_rate": 0.0006,
+ "loss": 4.251850605010986,
+ "step": 2046
+ },
+ {
+ "epoch": 28.433187772925763,
+ "grad_norm": 0.019667886197566986,
+ "learning_rate": 0.0006,
+ "loss": 4.222312927246094,
+ "step": 2047
+ },
+ {
+ "epoch": 28.4471615720524,
+ "grad_norm": 0.02029012329876423,
+ "learning_rate": 0.0006,
+ "loss": 4.246252059936523,
+ "step": 2048
+ },
+ {
+ "epoch": 28.46113537117904,
+ "grad_norm": 0.01784469373524189,
+ "learning_rate": 0.0006,
+ "loss": 4.192128658294678,
+ "step": 2049
+ },
+ {
+ "epoch": 28.475109170305675,
+ "grad_norm": 0.0160287544131279,
+ "learning_rate": 0.0006,
+ "loss": 4.208607196807861,
+ "step": 2050
+ },
+ {
+ "epoch": 28.489082969432314,
+ "grad_norm": 0.015348346903920174,
+ "learning_rate": 0.0006,
+ "loss": 4.236598014831543,
+ "step": 2051
+ },
+ {
+ "epoch": 28.503056768558952,
+ "grad_norm": 0.015198206529021263,
+ "learning_rate": 0.0006,
+ "loss": 4.235552787780762,
+ "step": 2052
+ },
+ {
+ "epoch": 28.51703056768559,
+ "grad_norm": 0.01523754745721817,
+ "learning_rate": 0.0006,
+ "loss": 4.1941728591918945,
+ "step": 2053
+ },
+ {
+ "epoch": 28.531004366812226,
+ "grad_norm": 0.015559614636003971,
+ "learning_rate": 0.0006,
+ "loss": 4.294757843017578,
+ "step": 2054
+ },
+ {
+ "epoch": 28.544978165938865,
+ "grad_norm": 0.01568358950316906,
+ "learning_rate": 0.0006,
+ "loss": 4.213065147399902,
+ "step": 2055
+ },
+ {
+ "epoch": 28.558951965065503,
+ "grad_norm": 0.015336516313254833,
+ "learning_rate": 0.0006,
+ "loss": 4.240293979644775,
+ "step": 2056
+ },
+ {
+ "epoch": 28.57292576419214,
+ "grad_norm": 0.015041496604681015,
+ "learning_rate": 0.0006,
+ "loss": 4.273397445678711,
+ "step": 2057
+ },
+ {
+ "epoch": 28.586899563318777,
+ "grad_norm": 0.01610128954052925,
+ "learning_rate": 0.0006,
+ "loss": 4.125369071960449,
+ "step": 2058
+ },
+ {
+ "epoch": 28.600873362445416,
+ "grad_norm": 0.015974976122379303,
+ "learning_rate": 0.0006,
+ "loss": 4.133459091186523,
+ "step": 2059
+ },
+ {
+ "epoch": 28.61484716157205,
+ "grad_norm": 0.015936799347400665,
+ "learning_rate": 0.0006,
+ "loss": 4.306667327880859,
+ "step": 2060
+ },
+ {
+ "epoch": 28.62882096069869,
+ "grad_norm": 0.01795247197151184,
+ "learning_rate": 0.0006,
+ "loss": 4.205156326293945,
+ "step": 2061
+ },
+ {
+ "epoch": 28.64279475982533,
+ "grad_norm": 0.019639814272522926,
+ "learning_rate": 0.0006,
+ "loss": 4.277614593505859,
+ "step": 2062
+ },
+ {
+ "epoch": 28.656768558951963,
+ "grad_norm": 0.01722300611436367,
+ "learning_rate": 0.0006,
+ "loss": 4.191032886505127,
+ "step": 2063
+ },
+ {
+ "epoch": 28.670742358078602,
+ "grad_norm": 0.01612633652985096,
+ "learning_rate": 0.0006,
+ "loss": 4.20717191696167,
+ "step": 2064
+ },
+ {
+ "epoch": 28.68471615720524,
+ "grad_norm": 0.01603279449045658,
+ "learning_rate": 0.0006,
+ "loss": 4.228884220123291,
+ "step": 2065
+ },
+ {
+ "epoch": 28.69868995633188,
+ "grad_norm": 0.016405927017331123,
+ "learning_rate": 0.0006,
+ "loss": 4.201292991638184,
+ "step": 2066
+ },
+ {
+ "epoch": 28.712663755458514,
+ "grad_norm": 0.016644183546304703,
+ "learning_rate": 0.0006,
+ "loss": 4.239282131195068,
+ "step": 2067
+ },
+ {
+ "epoch": 28.726637554585153,
+ "grad_norm": 0.0175587497651577,
+ "learning_rate": 0.0006,
+ "loss": 4.289038181304932,
+ "step": 2068
+ },
+ {
+ "epoch": 28.74061135371179,
+ "grad_norm": 0.018223397433757782,
+ "learning_rate": 0.0006,
+ "loss": 4.2266998291015625,
+ "step": 2069
+ },
+ {
+ "epoch": 28.754585152838427,
+ "grad_norm": 0.018680868670344353,
+ "learning_rate": 0.0006,
+ "loss": 4.348849296569824,
+ "step": 2070
+ },
+ {
+ "epoch": 28.768558951965066,
+ "grad_norm": 0.016726728528738022,
+ "learning_rate": 0.0006,
+ "loss": 4.292080879211426,
+ "step": 2071
+ },
+ {
+ "epoch": 28.782532751091704,
+ "grad_norm": 0.017386795952916145,
+ "learning_rate": 0.0006,
+ "loss": 4.208870887756348,
+ "step": 2072
+ },
+ {
+ "epoch": 28.79650655021834,
+ "grad_norm": 0.01865178905427456,
+ "learning_rate": 0.0006,
+ "loss": 4.054150581359863,
+ "step": 2073
+ },
+ {
+ "epoch": 28.810480349344978,
+ "grad_norm": 0.01776815392076969,
+ "learning_rate": 0.0006,
+ "loss": 4.202404499053955,
+ "step": 2074
+ },
+ {
+ "epoch": 28.824454148471617,
+ "grad_norm": 0.01643497310578823,
+ "learning_rate": 0.0006,
+ "loss": 4.131855010986328,
+ "step": 2075
+ },
+ {
+ "epoch": 28.83842794759825,
+ "grad_norm": 0.01574273779988289,
+ "learning_rate": 0.0006,
+ "loss": 4.296788692474365,
+ "step": 2076
+ },
+ {
+ "epoch": 28.85240174672489,
+ "grad_norm": 0.015613908879458904,
+ "learning_rate": 0.0006,
+ "loss": 4.210949897766113,
+ "step": 2077
+ },
+ {
+ "epoch": 28.86637554585153,
+ "grad_norm": 0.015543212182819843,
+ "learning_rate": 0.0006,
+ "loss": 4.218136310577393,
+ "step": 2078
+ },
+ {
+ "epoch": 28.880349344978168,
+ "grad_norm": 0.014681251719594002,
+ "learning_rate": 0.0006,
+ "loss": 4.19401741027832,
+ "step": 2079
+ },
+ {
+ "epoch": 28.894323144104803,
+ "grad_norm": 0.014393828809261322,
+ "learning_rate": 0.0006,
+ "loss": 4.198468208312988,
+ "step": 2080
+ },
+ {
+ "epoch": 28.90829694323144,
+ "grad_norm": 0.01580170728266239,
+ "learning_rate": 0.0006,
+ "loss": 4.224969863891602,
+ "step": 2081
+ },
+ {
+ "epoch": 28.92227074235808,
+ "grad_norm": 0.01484165620058775,
+ "learning_rate": 0.0006,
+ "loss": 4.12236213684082,
+ "step": 2082
+ },
+ {
+ "epoch": 28.936244541484715,
+ "grad_norm": 0.01497623696923256,
+ "learning_rate": 0.0006,
+ "loss": 4.24141788482666,
+ "step": 2083
+ },
+ {
+ "epoch": 28.950218340611354,
+ "grad_norm": 0.015080046840012074,
+ "learning_rate": 0.0006,
+ "loss": 4.284151077270508,
+ "step": 2084
+ },
+ {
+ "epoch": 28.964192139737992,
+ "grad_norm": 0.016207212582230568,
+ "learning_rate": 0.0006,
+ "loss": 4.298693656921387,
+ "step": 2085
+ },
+ {
+ "epoch": 28.978165938864628,
+ "grad_norm": 0.015474417246878147,
+ "learning_rate": 0.0006,
+ "loss": 4.2412261962890625,
+ "step": 2086
+ },
+ {
+ "epoch": 28.992139737991266,
+ "grad_norm": 0.01515351701527834,
+ "learning_rate": 0.0006,
+ "loss": 4.3054423332214355,
+ "step": 2087
+ },
+ {
+ "epoch": 29.0,
+ "grad_norm": 0.016411345452070236,
+ "learning_rate": 0.0006,
+ "loss": 4.104830741882324,
+ "step": 2088
+ },
+ {
+ "epoch": 29.0,
+ "eval_loss": 4.661567687988281,
+ "eval_runtime": 57.1244,
+ "eval_samples_per_second": 42.749,
+ "eval_steps_per_second": 1.348,
+ "step": 2088
+ },
+ {
+ "epoch": 29.01397379912664,
+ "grad_norm": 0.01665342226624489,
+ "learning_rate": 0.0006,
+ "loss": 4.311391830444336,
+ "step": 2089
+ },
+ {
+ "epoch": 29.027947598253274,
+ "grad_norm": 0.018053214997053146,
+ "learning_rate": 0.0006,
+ "loss": 4.135597229003906,
+ "step": 2090
+ },
+ {
+ "epoch": 29.041921397379912,
+ "grad_norm": 0.018213748931884766,
+ "learning_rate": 0.0006,
+ "loss": 4.303299903869629,
+ "step": 2091
+ },
+ {
+ "epoch": 29.05589519650655,
+ "grad_norm": 0.018460242077708244,
+ "learning_rate": 0.0006,
+ "loss": 4.20850944519043,
+ "step": 2092
+ },
+ {
+ "epoch": 29.069868995633186,
+ "grad_norm": 0.018365809693932533,
+ "learning_rate": 0.0006,
+ "loss": 4.275847434997559,
+ "step": 2093
+ },
+ {
+ "epoch": 29.083842794759825,
+ "grad_norm": 0.018346186727285385,
+ "learning_rate": 0.0006,
+ "loss": 4.202091217041016,
+ "step": 2094
+ },
+ {
+ "epoch": 29.097816593886463,
+ "grad_norm": 0.017920587211847305,
+ "learning_rate": 0.0006,
+ "loss": 4.182092666625977,
+ "step": 2095
+ },
+ {
+ "epoch": 29.111790393013102,
+ "grad_norm": 0.01812003180384636,
+ "learning_rate": 0.0006,
+ "loss": 4.259721755981445,
+ "step": 2096
+ },
+ {
+ "epoch": 29.125764192139737,
+ "grad_norm": 0.01700986735522747,
+ "learning_rate": 0.0006,
+ "loss": 4.067573070526123,
+ "step": 2097
+ },
+ {
+ "epoch": 29.139737991266376,
+ "grad_norm": 0.018069760873913765,
+ "learning_rate": 0.0006,
+ "loss": 4.234400749206543,
+ "step": 2098
+ },
+ {
+ "epoch": 29.153711790393015,
+ "grad_norm": 0.01905428245663643,
+ "learning_rate": 0.0006,
+ "loss": 4.16609001159668,
+ "step": 2099
+ },
+ {
+ "epoch": 29.16768558951965,
+ "grad_norm": 0.018711242824792862,
+ "learning_rate": 0.0006,
+ "loss": 4.249088764190674,
+ "step": 2100
+ },
+ {
+ "epoch": 29.18165938864629,
+ "grad_norm": 0.018721306696534157,
+ "learning_rate": 0.0006,
+ "loss": 4.14540958404541,
+ "step": 2101
+ },
+ {
+ "epoch": 29.195633187772927,
+ "grad_norm": 0.020139874890446663,
+ "learning_rate": 0.0006,
+ "loss": 4.049248695373535,
+ "step": 2102
+ },
+ {
+ "epoch": 29.209606986899562,
+ "grad_norm": 0.020849574357271194,
+ "learning_rate": 0.0006,
+ "loss": 4.235224723815918,
+ "step": 2103
+ },
+ {
+ "epoch": 29.2235807860262,
+ "grad_norm": 0.018693549558520317,
+ "learning_rate": 0.0006,
+ "loss": 4.232961177825928,
+ "step": 2104
+ },
+ {
+ "epoch": 29.23755458515284,
+ "grad_norm": 0.017889857292175293,
+ "learning_rate": 0.0006,
+ "loss": 4.171422004699707,
+ "step": 2105
+ },
+ {
+ "epoch": 29.251528384279474,
+ "grad_norm": 0.016690224409103394,
+ "learning_rate": 0.0006,
+ "loss": 4.244717597961426,
+ "step": 2106
+ },
+ {
+ "epoch": 29.265502183406113,
+ "grad_norm": 0.01708616502583027,
+ "learning_rate": 0.0006,
+ "loss": 4.125068664550781,
+ "step": 2107
+ },
+ {
+ "epoch": 29.27947598253275,
+ "grad_norm": 0.017928237095475197,
+ "learning_rate": 0.0006,
+ "loss": 4.208292007446289,
+ "step": 2108
+ },
+ {
+ "epoch": 29.29344978165939,
+ "grad_norm": 0.018250394612550735,
+ "learning_rate": 0.0006,
+ "loss": 4.118630886077881,
+ "step": 2109
+ },
+ {
+ "epoch": 29.307423580786025,
+ "grad_norm": 0.020160246640443802,
+ "learning_rate": 0.0006,
+ "loss": 4.206025123596191,
+ "step": 2110
+ },
+ {
+ "epoch": 29.321397379912664,
+ "grad_norm": 0.020045453682541847,
+ "learning_rate": 0.0006,
+ "loss": 4.2758684158325195,
+ "step": 2111
+ },
+ {
+ "epoch": 29.335371179039303,
+ "grad_norm": 0.019556907936930656,
+ "learning_rate": 0.0006,
+ "loss": 4.188272953033447,
+ "step": 2112
+ },
+ {
+ "epoch": 29.349344978165938,
+ "grad_norm": 0.019185200333595276,
+ "learning_rate": 0.0006,
+ "loss": 4.17451286315918,
+ "step": 2113
+ },
+ {
+ "epoch": 29.363318777292577,
+ "grad_norm": 0.01816629432141781,
+ "learning_rate": 0.0006,
+ "loss": 4.165246963500977,
+ "step": 2114
+ },
+ {
+ "epoch": 29.377292576419215,
+ "grad_norm": 0.01865146867930889,
+ "learning_rate": 0.0006,
+ "loss": 4.148184776306152,
+ "step": 2115
+ },
+ {
+ "epoch": 29.39126637554585,
+ "grad_norm": 0.018338464200496674,
+ "learning_rate": 0.0006,
+ "loss": 4.288936614990234,
+ "step": 2116
+ },
+ {
+ "epoch": 29.40524017467249,
+ "grad_norm": 0.018986187875270844,
+ "learning_rate": 0.0006,
+ "loss": 4.183379650115967,
+ "step": 2117
+ },
+ {
+ "epoch": 29.419213973799128,
+ "grad_norm": 0.018409637734293938,
+ "learning_rate": 0.0006,
+ "loss": 4.057574272155762,
+ "step": 2118
+ },
+ {
+ "epoch": 29.433187772925763,
+ "grad_norm": 0.016664542257785797,
+ "learning_rate": 0.0006,
+ "loss": 4.18517541885376,
+ "step": 2119
+ },
+ {
+ "epoch": 29.4471615720524,
+ "grad_norm": 0.016486089676618576,
+ "learning_rate": 0.0006,
+ "loss": 4.179367542266846,
+ "step": 2120
+ },
+ {
+ "epoch": 29.46113537117904,
+ "grad_norm": 0.01787523552775383,
+ "learning_rate": 0.0006,
+ "loss": 4.184126377105713,
+ "step": 2121
+ },
+ {
+ "epoch": 29.475109170305675,
+ "grad_norm": 0.016574613749980927,
+ "learning_rate": 0.0006,
+ "loss": 4.290170192718506,
+ "step": 2122
+ },
+ {
+ "epoch": 29.489082969432314,
+ "grad_norm": 0.01643376611173153,
+ "learning_rate": 0.0006,
+ "loss": 4.052913665771484,
+ "step": 2123
+ },
+ {
+ "epoch": 29.503056768558952,
+ "grad_norm": 0.016503969207406044,
+ "learning_rate": 0.0006,
+ "loss": 4.335184097290039,
+ "step": 2124
+ },
+ {
+ "epoch": 29.51703056768559,
+ "grad_norm": 0.017596479505300522,
+ "learning_rate": 0.0006,
+ "loss": 4.165539741516113,
+ "step": 2125
+ },
+ {
+ "epoch": 29.531004366812226,
+ "grad_norm": 0.017468582838773727,
+ "learning_rate": 0.0006,
+ "loss": 4.315964698791504,
+ "step": 2126
+ },
+ {
+ "epoch": 29.544978165938865,
+ "grad_norm": 0.01724618673324585,
+ "learning_rate": 0.0006,
+ "loss": 4.243681907653809,
+ "step": 2127
+ },
+ {
+ "epoch": 29.558951965065503,
+ "grad_norm": 0.018384616822004318,
+ "learning_rate": 0.0006,
+ "loss": 4.112447261810303,
+ "step": 2128
+ },
+ {
+ "epoch": 29.57292576419214,
+ "grad_norm": 0.01824074238538742,
+ "learning_rate": 0.0006,
+ "loss": 4.236065864562988,
+ "step": 2129
+ },
+ {
+ "epoch": 29.586899563318777,
+ "grad_norm": 0.01625971868634224,
+ "learning_rate": 0.0006,
+ "loss": 4.18143892288208,
+ "step": 2130
+ },
+ {
+ "epoch": 29.600873362445416,
+ "grad_norm": 0.015537494793534279,
+ "learning_rate": 0.0006,
+ "loss": 4.225147247314453,
+ "step": 2131
+ },
+ {
+ "epoch": 29.61484716157205,
+ "grad_norm": 0.015818974003195763,
+ "learning_rate": 0.0006,
+ "loss": 4.3100175857543945,
+ "step": 2132
+ },
+ {
+ "epoch": 29.62882096069869,
+ "grad_norm": 0.015909671783447266,
+ "learning_rate": 0.0006,
+ "loss": 4.169775009155273,
+ "step": 2133
+ },
+ {
+ "epoch": 29.64279475982533,
+ "grad_norm": 0.01769658550620079,
+ "learning_rate": 0.0006,
+ "loss": 4.278904914855957,
+ "step": 2134
+ },
+ {
+ "epoch": 29.656768558951963,
+ "grad_norm": 0.015641704201698303,
+ "learning_rate": 0.0006,
+ "loss": 4.2426228523254395,
+ "step": 2135
+ },
+ {
+ "epoch": 29.670742358078602,
+ "grad_norm": 0.01565658301115036,
+ "learning_rate": 0.0006,
+ "loss": 4.232687950134277,
+ "step": 2136
+ },
+ {
+ "epoch": 29.68471615720524,
+ "grad_norm": 0.016485940665006638,
+ "learning_rate": 0.0006,
+ "loss": 4.209839820861816,
+ "step": 2137
+ },
+ {
+ "epoch": 29.69868995633188,
+ "grad_norm": 0.017622729763388634,
+ "learning_rate": 0.0006,
+ "loss": 4.196017265319824,
+ "step": 2138
+ },
+ {
+ "epoch": 29.712663755458514,
+ "grad_norm": 0.018320564180612564,
+ "learning_rate": 0.0006,
+ "loss": 4.225683689117432,
+ "step": 2139
+ },
+ {
+ "epoch": 29.726637554585153,
+ "grad_norm": 0.018140794709324837,
+ "learning_rate": 0.0006,
+ "loss": 4.311944961547852,
+ "step": 2140
+ },
+ {
+ "epoch": 29.74061135371179,
+ "grad_norm": 0.016227100044488907,
+ "learning_rate": 0.0006,
+ "loss": 4.230257511138916,
+ "step": 2141
+ },
+ {
+ "epoch": 29.754585152838427,
+ "grad_norm": 0.015759488567709923,
+ "learning_rate": 0.0006,
+ "loss": 4.119174003601074,
+ "step": 2142
+ },
+ {
+ "epoch": 29.768558951965066,
+ "grad_norm": 0.01757766678929329,
+ "learning_rate": 0.0006,
+ "loss": 4.165826797485352,
+ "step": 2143
+ },
+ {
+ "epoch": 29.782532751091704,
+ "grad_norm": 0.019661923870444298,
+ "learning_rate": 0.0006,
+ "loss": 4.392203330993652,
+ "step": 2144
+ },
+ {
+ "epoch": 29.79650655021834,
+ "grad_norm": 0.019478755071759224,
+ "learning_rate": 0.0006,
+ "loss": 4.348371982574463,
+ "step": 2145
+ },
+ {
+ "epoch": 29.810480349344978,
+ "grad_norm": 0.019149569794535637,
+ "learning_rate": 0.0006,
+ "loss": 4.34104061126709,
+ "step": 2146
+ },
+ {
+ "epoch": 29.824454148471617,
+ "grad_norm": 0.017707521095871925,
+ "learning_rate": 0.0006,
+ "loss": 4.228226661682129,
+ "step": 2147
+ },
+ {
+ "epoch": 29.83842794759825,
+ "grad_norm": 0.015615378506481647,
+ "learning_rate": 0.0006,
+ "loss": 4.175541877746582,
+ "step": 2148
+ },
+ {
+ "epoch": 29.85240174672489,
+ "grad_norm": 0.016111129894852638,
+ "learning_rate": 0.0006,
+ "loss": 4.267936706542969,
+ "step": 2149
+ },
+ {
+ "epoch": 29.86637554585153,
+ "grad_norm": 0.01528779324144125,
+ "learning_rate": 0.0006,
+ "loss": 4.186014175415039,
+ "step": 2150
+ },
+ {
+ "epoch": 29.880349344978168,
+ "grad_norm": 0.014925646595656872,
+ "learning_rate": 0.0006,
+ "loss": 4.174860000610352,
+ "step": 2151
+ },
+ {
+ "epoch": 29.894323144104803,
+ "grad_norm": 0.016917673870921135,
+ "learning_rate": 0.0006,
+ "loss": 4.33908748626709,
+ "step": 2152
+ },
+ {
+ "epoch": 29.90829694323144,
+ "grad_norm": 0.016007075086236,
+ "learning_rate": 0.0006,
+ "loss": 4.266141414642334,
+ "step": 2153
+ },
+ {
+ "epoch": 29.92227074235808,
+ "grad_norm": 0.016178956255316734,
+ "learning_rate": 0.0006,
+ "loss": 4.271215915679932,
+ "step": 2154
+ },
+ {
+ "epoch": 29.936244541484715,
+ "grad_norm": 0.016180871054530144,
+ "learning_rate": 0.0006,
+ "loss": 4.278801441192627,
+ "step": 2155
+ },
+ {
+ "epoch": 29.950218340611354,
+ "grad_norm": 0.015335503034293652,
+ "learning_rate": 0.0006,
+ "loss": 4.194978713989258,
+ "step": 2156
+ },
+ {
+ "epoch": 29.964192139737992,
+ "grad_norm": 0.016580764204263687,
+ "learning_rate": 0.0006,
+ "loss": 4.290809631347656,
+ "step": 2157
+ },
+ {
+ "epoch": 29.978165938864628,
+ "grad_norm": 0.01623128168284893,
+ "learning_rate": 0.0006,
+ "loss": 4.205451488494873,
+ "step": 2158
+ },
+ {
+ "epoch": 29.992139737991266,
+ "grad_norm": 0.014904248528182507,
+ "learning_rate": 0.0006,
+ "loss": 4.24924373626709,
+ "step": 2159
+ },
+ {
+ "epoch": 30.0,
+ "grad_norm": 0.017089299857616425,
+ "learning_rate": 0.0006,
+ "loss": 4.222072601318359,
+ "step": 2160
+ },
+ {
+ "epoch": 30.0,
+ "eval_loss": 4.585411071777344,
+ "eval_runtime": 57.1164,
+ "eval_samples_per_second": 42.755,
+ "eval_steps_per_second": 1.348,
+ "step": 2160
+ },
+ {
+ "epoch": 30.01397379912664,
+ "grad_norm": 0.017226383090019226,
+ "learning_rate": 0.0006,
+ "loss": 4.19467830657959,
+ "step": 2161
+ },
+ {
+ "epoch": 30.027947598253274,
+ "grad_norm": 0.01635066606104374,
+ "learning_rate": 0.0006,
+ "loss": 4.176875591278076,
+ "step": 2162
+ },
+ {
+ "epoch": 30.041921397379912,
+ "grad_norm": 0.01735362410545349,
+ "learning_rate": 0.0006,
+ "loss": 4.13667106628418,
+ "step": 2163
+ },
+ {
+ "epoch": 30.05589519650655,
+ "grad_norm": 0.017047109082341194,
+ "learning_rate": 0.0006,
+ "loss": 4.136329650878906,
+ "step": 2164
+ },
+ {
+ "epoch": 30.069868995633186,
+ "grad_norm": 0.018116053193807602,
+ "learning_rate": 0.0006,
+ "loss": 4.228706359863281,
+ "step": 2165
+ },
+ {
+ "epoch": 30.083842794759825,
+ "grad_norm": 0.01791389286518097,
+ "learning_rate": 0.0006,
+ "loss": 4.1597747802734375,
+ "step": 2166
+ },
+ {
+ "epoch": 30.097816593886463,
+ "grad_norm": 0.017793502658605576,
+ "learning_rate": 0.0006,
+ "loss": 4.0807976722717285,
+ "step": 2167
+ },
+ {
+ "epoch": 30.111790393013102,
+ "grad_norm": 0.01827503927052021,
+ "learning_rate": 0.0006,
+ "loss": 4.170646667480469,
+ "step": 2168
+ },
+ {
+ "epoch": 30.125764192139737,
+ "grad_norm": 0.017228564247488976,
+ "learning_rate": 0.0006,
+ "loss": 4.146934509277344,
+ "step": 2169
+ },
+ {
+ "epoch": 30.139737991266376,
+ "grad_norm": 0.01757437363266945,
+ "learning_rate": 0.0006,
+ "loss": 4.1859130859375,
+ "step": 2170
+ },
+ {
+ "epoch": 30.153711790393015,
+ "grad_norm": 0.01701093092560768,
+ "learning_rate": 0.0006,
+ "loss": 4.126981258392334,
+ "step": 2171
+ },
+ {
+ "epoch": 30.16768558951965,
+ "grad_norm": 0.015220489352941513,
+ "learning_rate": 0.0006,
+ "loss": 4.164018630981445,
+ "step": 2172
+ },
+ {
+ "epoch": 30.18165938864629,
+ "grad_norm": 0.016254238784313202,
+ "learning_rate": 0.0006,
+ "loss": 4.258275508880615,
+ "step": 2173
+ },
+ {
+ "epoch": 30.195633187772927,
+ "grad_norm": 0.016609614714980125,
+ "learning_rate": 0.0006,
+ "loss": 4.222233772277832,
+ "step": 2174
+ },
+ {
+ "epoch": 30.209606986899562,
+ "grad_norm": 0.016531318426132202,
+ "learning_rate": 0.0006,
+ "loss": 4.144489288330078,
+ "step": 2175
+ },
+ {
+ "epoch": 30.2235807860262,
+ "grad_norm": 0.01567668654024601,
+ "learning_rate": 0.0006,
+ "loss": 4.181916236877441,
+ "step": 2176
+ },
+ {
+ "epoch": 30.23755458515284,
+ "grad_norm": 0.017195681110024452,
+ "learning_rate": 0.0006,
+ "loss": 4.189693927764893,
+ "step": 2177
+ },
+ {
+ "epoch": 30.251528384279474,
+ "grad_norm": 0.016882948577404022,
+ "learning_rate": 0.0006,
+ "loss": 4.21604061126709,
+ "step": 2178
+ },
+ {
+ "epoch": 30.265502183406113,
+ "grad_norm": 0.016261622309684753,
+ "learning_rate": 0.0006,
+ "loss": 4.227639198303223,
+ "step": 2179
+ },
+ {
+ "epoch": 30.27947598253275,
+ "grad_norm": 0.016411812976002693,
+ "learning_rate": 0.0006,
+ "loss": 4.130362510681152,
+ "step": 2180
+ },
+ {
+ "epoch": 30.29344978165939,
+ "grad_norm": 0.01732019893825054,
+ "learning_rate": 0.0006,
+ "loss": 4.230443000793457,
+ "step": 2181
+ },
+ {
+ "epoch": 30.307423580786025,
+ "grad_norm": 0.01825905591249466,
+ "learning_rate": 0.0006,
+ "loss": 4.175817489624023,
+ "step": 2182
+ },
+ {
+ "epoch": 30.321397379912664,
+ "grad_norm": 0.01887713558971882,
+ "learning_rate": 0.0006,
+ "loss": 4.169020652770996,
+ "step": 2183
+ },
+ {
+ "epoch": 30.335371179039303,
+ "grad_norm": 0.019391661509871483,
+ "learning_rate": 0.0006,
+ "loss": 4.176318645477295,
+ "step": 2184
+ },
+ {
+ "epoch": 30.349344978165938,
+ "grad_norm": 0.020840002223849297,
+ "learning_rate": 0.0006,
+ "loss": 4.174228668212891,
+ "step": 2185
+ },
+ {
+ "epoch": 30.363318777292577,
+ "grad_norm": 0.020193178206682205,
+ "learning_rate": 0.0006,
+ "loss": 4.127919673919678,
+ "step": 2186
+ },
+ {
+ "epoch": 30.377292576419215,
+ "grad_norm": 0.018449561670422554,
+ "learning_rate": 0.0006,
+ "loss": 4.229487895965576,
+ "step": 2187
+ },
+ {
+ "epoch": 30.39126637554585,
+ "grad_norm": 0.01711983233690262,
+ "learning_rate": 0.0006,
+ "loss": 4.201651573181152,
+ "step": 2188
+ },
+ {
+ "epoch": 30.40524017467249,
+ "grad_norm": 0.017293395474553108,
+ "learning_rate": 0.0006,
+ "loss": 4.168704032897949,
+ "step": 2189
+ },
+ {
+ "epoch": 30.419213973799128,
+ "grad_norm": 0.018526358529925346,
+ "learning_rate": 0.0006,
+ "loss": 4.171864032745361,
+ "step": 2190
+ },
+ {
+ "epoch": 30.433187772925763,
+ "grad_norm": 0.019287291914224625,
+ "learning_rate": 0.0006,
+ "loss": 4.039534568786621,
+ "step": 2191
+ },
+ {
+ "epoch": 30.4471615720524,
+ "grad_norm": 0.016902444884181023,
+ "learning_rate": 0.0006,
+ "loss": 4.282787799835205,
+ "step": 2192
+ },
+ {
+ "epoch": 30.46113537117904,
+ "grad_norm": 0.01656750962138176,
+ "learning_rate": 0.0006,
+ "loss": 4.213098526000977,
+ "step": 2193
+ },
+ {
+ "epoch": 30.475109170305675,
+ "grad_norm": 0.016805065795779228,
+ "learning_rate": 0.0006,
+ "loss": 4.1583476066589355,
+ "step": 2194
+ },
+ {
+ "epoch": 30.489082969432314,
+ "grad_norm": 0.017166638746857643,
+ "learning_rate": 0.0006,
+ "loss": 4.188634395599365,
+ "step": 2195
+ },
+ {
+ "epoch": 30.503056768558952,
+ "grad_norm": 0.01592106744647026,
+ "learning_rate": 0.0006,
+ "loss": 4.234043121337891,
+ "step": 2196
+ },
+ {
+ "epoch": 30.51703056768559,
+ "grad_norm": 0.015270989388227463,
+ "learning_rate": 0.0006,
+ "loss": 4.301420211791992,
+ "step": 2197
+ },
+ {
+ "epoch": 30.531004366812226,
+ "grad_norm": 0.015194724313914776,
+ "learning_rate": 0.0006,
+ "loss": 4.18703031539917,
+ "step": 2198
+ },
+ {
+ "epoch": 30.544978165938865,
+ "grad_norm": 0.015587746165692806,
+ "learning_rate": 0.0006,
+ "loss": 4.138928413391113,
+ "step": 2199
+ },
+ {
+ "epoch": 30.558951965065503,
+ "grad_norm": 0.016029570251703262,
+ "learning_rate": 0.0006,
+ "loss": 4.290863990783691,
+ "step": 2200
+ },
+ {
+ "epoch": 30.57292576419214,
+ "grad_norm": 0.015651309862732887,
+ "learning_rate": 0.0006,
+ "loss": 4.272062301635742,
+ "step": 2201
+ },
+ {
+ "epoch": 30.586899563318777,
+ "grad_norm": 0.015992306172847748,
+ "learning_rate": 0.0006,
+ "loss": 4.2487406730651855,
+ "step": 2202
+ },
+ {
+ "epoch": 30.600873362445416,
+ "grad_norm": 0.014969157986342907,
+ "learning_rate": 0.0006,
+ "loss": 4.115347862243652,
+ "step": 2203
+ },
+ {
+ "epoch": 30.61484716157205,
+ "grad_norm": 0.014639027416706085,
+ "learning_rate": 0.0006,
+ "loss": 4.2207489013671875,
+ "step": 2204
+ },
+ {
+ "epoch": 30.62882096069869,
+ "grad_norm": 0.015059413388371468,
+ "learning_rate": 0.0006,
+ "loss": 4.274044990539551,
+ "step": 2205
+ },
+ {
+ "epoch": 30.64279475982533,
+ "grad_norm": 0.0166641678661108,
+ "learning_rate": 0.0006,
+ "loss": 4.191373348236084,
+ "step": 2206
+ },
+ {
+ "epoch": 30.656768558951963,
+ "grad_norm": 0.017806345596909523,
+ "learning_rate": 0.0006,
+ "loss": 4.139800548553467,
+ "step": 2207
+ },
+ {
+ "epoch": 30.670742358078602,
+ "grad_norm": 0.018712421879172325,
+ "learning_rate": 0.0006,
+ "loss": 4.035965442657471,
+ "step": 2208
+ },
+ {
+ "epoch": 30.68471615720524,
+ "grad_norm": 0.018457243219017982,
+ "learning_rate": 0.0006,
+ "loss": 4.2579240798950195,
+ "step": 2209
+ },
+ {
+ "epoch": 30.69868995633188,
+ "grad_norm": 0.01920999586582184,
+ "learning_rate": 0.0006,
+ "loss": 4.131112098693848,
+ "step": 2210
+ },
+ {
+ "epoch": 30.712663755458514,
+ "grad_norm": 0.01729944534599781,
+ "learning_rate": 0.0006,
+ "loss": 4.254059314727783,
+ "step": 2211
+ },
+ {
+ "epoch": 30.726637554585153,
+ "grad_norm": 0.016914231702685356,
+ "learning_rate": 0.0006,
+ "loss": 4.233428955078125,
+ "step": 2212
+ },
+ {
+ "epoch": 30.74061135371179,
+ "grad_norm": 0.01913098618388176,
+ "learning_rate": 0.0006,
+ "loss": 4.23525333404541,
+ "step": 2213
+ },
+ {
+ "epoch": 30.754585152838427,
+ "grad_norm": 0.019150281324982643,
+ "learning_rate": 0.0006,
+ "loss": 4.213695049285889,
+ "step": 2214
+ },
+ {
+ "epoch": 30.768558951965066,
+ "grad_norm": 0.018947452306747437,
+ "learning_rate": 0.0006,
+ "loss": 4.151697158813477,
+ "step": 2215
+ },
+ {
+ "epoch": 30.782532751091704,
+ "grad_norm": 0.016058508306741714,
+ "learning_rate": 0.0006,
+ "loss": 4.220280170440674,
+ "step": 2216
+ },
+ {
+ "epoch": 30.79650655021834,
+ "grad_norm": 0.017488619312644005,
+ "learning_rate": 0.0006,
+ "loss": 4.154428482055664,
+ "step": 2217
+ },
+ {
+ "epoch": 30.810480349344978,
+ "grad_norm": 0.017355090007185936,
+ "learning_rate": 0.0006,
+ "loss": 4.256086349487305,
+ "step": 2218
+ },
+ {
+ "epoch": 30.824454148471617,
+ "grad_norm": 0.016636032611131668,
+ "learning_rate": 0.0006,
+ "loss": 4.227564334869385,
+ "step": 2219
+ },
+ {
+ "epoch": 30.83842794759825,
+ "grad_norm": 0.016273891553282738,
+ "learning_rate": 0.0006,
+ "loss": 4.200466156005859,
+ "step": 2220
+ },
+ {
+ "epoch": 30.85240174672489,
+ "grad_norm": 0.016939911991357803,
+ "learning_rate": 0.0006,
+ "loss": 4.1571221351623535,
+ "step": 2221
+ },
+ {
+ "epoch": 30.86637554585153,
+ "grad_norm": 0.018332522362470627,
+ "learning_rate": 0.0006,
+ "loss": 4.228398323059082,
+ "step": 2222
+ },
+ {
+ "epoch": 30.880349344978168,
+ "grad_norm": 0.019508114084601402,
+ "learning_rate": 0.0006,
+ "loss": 4.222960472106934,
+ "step": 2223
+ },
+ {
+ "epoch": 30.894323144104803,
+ "grad_norm": 0.018440278246998787,
+ "learning_rate": 0.0006,
+ "loss": 4.261051177978516,
+ "step": 2224
+ },
+ {
+ "epoch": 30.90829694323144,
+ "grad_norm": 0.018113229423761368,
+ "learning_rate": 0.0006,
+ "loss": 4.109235763549805,
+ "step": 2225
+ },
+ {
+ "epoch": 30.92227074235808,
+ "grad_norm": 0.015392083674669266,
+ "learning_rate": 0.0006,
+ "loss": 4.21950101852417,
+ "step": 2226
+ },
+ {
+ "epoch": 30.936244541484715,
+ "grad_norm": 0.01633290946483612,
+ "learning_rate": 0.0006,
+ "loss": 4.176053047180176,
+ "step": 2227
+ },
+ {
+ "epoch": 30.950218340611354,
+ "grad_norm": 0.016236383467912674,
+ "learning_rate": 0.0006,
+ "loss": 4.314309120178223,
+ "step": 2228
+ },
+ {
+ "epoch": 30.964192139737992,
+ "grad_norm": 0.015284373424947262,
+ "learning_rate": 0.0006,
+ "loss": 4.303133487701416,
+ "step": 2229
+ },
+ {
+ "epoch": 30.978165938864628,
+ "grad_norm": 0.015982897952198982,
+ "learning_rate": 0.0006,
+ "loss": 4.25739860534668,
+ "step": 2230
+ },
+ {
+ "epoch": 30.992139737991266,
+ "grad_norm": 0.016682198271155357,
+ "learning_rate": 0.0006,
+ "loss": 4.292210578918457,
+ "step": 2231
+ },
+ {
+ "epoch": 31.0,
+ "grad_norm": 0.017521467059850693,
+ "learning_rate": 0.0006,
+ "loss": 4.170637130737305,
+ "step": 2232
+ },
+ {
+ "epoch": 31.0,
+ "eval_loss": 4.642786979675293,
+ "eval_runtime": 57.029,
+ "eval_samples_per_second": 42.82,
+ "eval_steps_per_second": 1.35,
+ "step": 2232
+ },
+ {
+ "epoch": 31.01397379912664,
+ "grad_norm": 0.016607852652668953,
+ "learning_rate": 0.0006,
+ "loss": 4.106224536895752,
+ "step": 2233
+ },
+ {
+ "epoch": 31.027947598253274,
+ "grad_norm": 0.018448904156684875,
+ "learning_rate": 0.0006,
+ "loss": 4.224459648132324,
+ "step": 2234
+ },
+ {
+ "epoch": 31.041921397379912,
+ "grad_norm": 0.020481228828430176,
+ "learning_rate": 0.0006,
+ "loss": 4.305548667907715,
+ "step": 2235
+ },
+ {
+ "epoch": 31.05589519650655,
+ "grad_norm": 0.01999707892537117,
+ "learning_rate": 0.0006,
+ "loss": 4.253663539886475,
+ "step": 2236
+ },
+ {
+ "epoch": 31.069868995633186,
+ "grad_norm": 0.020428303629159927,
+ "learning_rate": 0.0006,
+ "loss": 4.166954040527344,
+ "step": 2237
+ },
+ {
+ "epoch": 31.083842794759825,
+ "grad_norm": 0.022608213126659393,
+ "learning_rate": 0.0006,
+ "loss": 4.172819137573242,
+ "step": 2238
+ },
+ {
+ "epoch": 31.097816593886463,
+ "grad_norm": 0.024669520556926727,
+ "learning_rate": 0.0006,
+ "loss": 4.106586456298828,
+ "step": 2239
+ },
+ {
+ "epoch": 31.111790393013102,
+ "grad_norm": 0.02247351035475731,
+ "learning_rate": 0.0006,
+ "loss": 4.160271644592285,
+ "step": 2240
+ },
+ {
+ "epoch": 31.125764192139737,
+ "grad_norm": 0.0207810141146183,
+ "learning_rate": 0.0006,
+ "loss": 4.272491931915283,
+ "step": 2241
+ },
+ {
+ "epoch": 31.139737991266376,
+ "grad_norm": 0.021576499566435814,
+ "learning_rate": 0.0006,
+ "loss": 4.171474456787109,
+ "step": 2242
+ },
+ {
+ "epoch": 31.153711790393015,
+ "grad_norm": 0.02337106689810753,
+ "learning_rate": 0.0006,
+ "loss": 4.094712734222412,
+ "step": 2243
+ },
+ {
+ "epoch": 31.16768558951965,
+ "grad_norm": 0.02219875156879425,
+ "learning_rate": 0.0006,
+ "loss": 4.1800994873046875,
+ "step": 2244
+ },
+ {
+ "epoch": 31.18165938864629,
+ "grad_norm": 0.022680504247546196,
+ "learning_rate": 0.0006,
+ "loss": 4.2054362297058105,
+ "step": 2245
+ },
+ {
+ "epoch": 31.195633187772927,
+ "grad_norm": 0.0208734143525362,
+ "learning_rate": 0.0006,
+ "loss": 4.06068229675293,
+ "step": 2246
+ },
+ {
+ "epoch": 31.209606986899562,
+ "grad_norm": 0.01811066083610058,
+ "learning_rate": 0.0006,
+ "loss": 4.2465128898620605,
+ "step": 2247
+ },
+ {
+ "epoch": 31.2235807860262,
+ "grad_norm": 0.018571637570858,
+ "learning_rate": 0.0006,
+ "loss": 4.191955089569092,
+ "step": 2248
+ },
+ {
+ "epoch": 31.23755458515284,
+ "grad_norm": 0.018598807975649834,
+ "learning_rate": 0.0006,
+ "loss": 4.200872421264648,
+ "step": 2249
+ },
+ {
+ "epoch": 31.251528384279474,
+ "grad_norm": 0.017577625811100006,
+ "learning_rate": 0.0006,
+ "loss": 4.148700714111328,
+ "step": 2250
+ },
+ {
+ "epoch": 31.265502183406113,
+ "grad_norm": 0.01817980408668518,
+ "learning_rate": 0.0006,
+ "loss": 4.256978988647461,
+ "step": 2251
+ },
+ {
+ "epoch": 31.27947598253275,
+ "grad_norm": 0.017983486875891685,
+ "learning_rate": 0.0006,
+ "loss": 4.175005912780762,
+ "step": 2252
+ },
+ {
+ "epoch": 31.29344978165939,
+ "grad_norm": 0.016751961782574654,
+ "learning_rate": 0.0006,
+ "loss": 4.237518310546875,
+ "step": 2253
+ },
+ {
+ "epoch": 31.307423580786025,
+ "grad_norm": 0.016407202929258347,
+ "learning_rate": 0.0006,
+ "loss": 4.197269439697266,
+ "step": 2254
+ },
+ {
+ "epoch": 31.321397379912664,
+ "grad_norm": 0.01578645594418049,
+ "learning_rate": 0.0006,
+ "loss": 4.179751396179199,
+ "step": 2255
+ },
+ {
+ "epoch": 31.335371179039303,
+ "grad_norm": 0.016001226380467415,
+ "learning_rate": 0.0006,
+ "loss": 4.08821964263916,
+ "step": 2256
+ },
+ {
+ "epoch": 31.349344978165938,
+ "grad_norm": 0.015122811309993267,
+ "learning_rate": 0.0006,
+ "loss": 4.047828674316406,
+ "step": 2257
+ },
+ {
+ "epoch": 31.363318777292577,
+ "grad_norm": 0.015399965457618237,
+ "learning_rate": 0.0006,
+ "loss": 4.1506524085998535,
+ "step": 2258
+ },
+ {
+ "epoch": 31.377292576419215,
+ "grad_norm": 0.01534605398774147,
+ "learning_rate": 0.0006,
+ "loss": 4.27952766418457,
+ "step": 2259
+ },
+ {
+ "epoch": 31.39126637554585,
+ "grad_norm": 0.015461381524801254,
+ "learning_rate": 0.0006,
+ "loss": 4.27102518081665,
+ "step": 2260
+ },
+ {
+ "epoch": 31.40524017467249,
+ "grad_norm": 0.016769153997302055,
+ "learning_rate": 0.0006,
+ "loss": 4.191592216491699,
+ "step": 2261
+ },
+ {
+ "epoch": 31.419213973799128,
+ "grad_norm": 0.016557959839701653,
+ "learning_rate": 0.0006,
+ "loss": 4.159705638885498,
+ "step": 2262
+ },
+ {
+ "epoch": 31.433187772925763,
+ "grad_norm": 0.01659569889307022,
+ "learning_rate": 0.0006,
+ "loss": 4.2136664390563965,
+ "step": 2263
+ },
+ {
+ "epoch": 31.4471615720524,
+ "grad_norm": 0.017101937904953957,
+ "learning_rate": 0.0006,
+ "loss": 4.2716498374938965,
+ "step": 2264
+ },
+ {
+ "epoch": 31.46113537117904,
+ "grad_norm": 0.01722985878586769,
+ "learning_rate": 0.0006,
+ "loss": 4.300534248352051,
+ "step": 2265
+ },
+ {
+ "epoch": 31.475109170305675,
+ "grad_norm": 0.016800185665488243,
+ "learning_rate": 0.0006,
+ "loss": 4.113240718841553,
+ "step": 2266
+ },
+ {
+ "epoch": 31.489082969432314,
+ "grad_norm": 0.01603875495493412,
+ "learning_rate": 0.0006,
+ "loss": 4.157105445861816,
+ "step": 2267
+ },
+ {
+ "epoch": 31.503056768558952,
+ "grad_norm": 0.015903932973742485,
+ "learning_rate": 0.0006,
+ "loss": 4.233669757843018,
+ "step": 2268
+ },
+ {
+ "epoch": 31.51703056768559,
+ "grad_norm": 0.0160983856767416,
+ "learning_rate": 0.0006,
+ "loss": 4.147704124450684,
+ "step": 2269
+ },
+ {
+ "epoch": 31.531004366812226,
+ "grad_norm": 0.01604730263352394,
+ "learning_rate": 0.0006,
+ "loss": 4.1465911865234375,
+ "step": 2270
+ },
+ {
+ "epoch": 31.544978165938865,
+ "grad_norm": 0.016205785796046257,
+ "learning_rate": 0.0006,
+ "loss": 4.106704235076904,
+ "step": 2271
+ },
+ {
+ "epoch": 31.558951965065503,
+ "grad_norm": 0.017008036375045776,
+ "learning_rate": 0.0006,
+ "loss": 4.191000938415527,
+ "step": 2272
+ },
+ {
+ "epoch": 31.57292576419214,
+ "grad_norm": 0.01699635200202465,
+ "learning_rate": 0.0006,
+ "loss": 4.187042236328125,
+ "step": 2273
+ },
+ {
+ "epoch": 31.586899563318777,
+ "grad_norm": 0.01692967116832733,
+ "learning_rate": 0.0006,
+ "loss": 4.208805561065674,
+ "step": 2274
+ },
+ {
+ "epoch": 31.600873362445416,
+ "grad_norm": 0.017608756199479103,
+ "learning_rate": 0.0006,
+ "loss": 4.200512886047363,
+ "step": 2275
+ },
+ {
+ "epoch": 31.61484716157205,
+ "grad_norm": 0.016590990126132965,
+ "learning_rate": 0.0006,
+ "loss": 4.194705009460449,
+ "step": 2276
+ },
+ {
+ "epoch": 31.62882096069869,
+ "grad_norm": 0.0170235987752676,
+ "learning_rate": 0.0006,
+ "loss": 4.210326194763184,
+ "step": 2277
+ },
+ {
+ "epoch": 31.64279475982533,
+ "grad_norm": 0.01604801043868065,
+ "learning_rate": 0.0006,
+ "loss": 4.196024417877197,
+ "step": 2278
+ },
+ {
+ "epoch": 31.656768558951963,
+ "grad_norm": 0.015694202855229378,
+ "learning_rate": 0.0006,
+ "loss": 4.178236484527588,
+ "step": 2279
+ },
+ {
+ "epoch": 31.670742358078602,
+ "grad_norm": 0.015473959036171436,
+ "learning_rate": 0.0006,
+ "loss": 4.259339809417725,
+ "step": 2280
+ },
+ {
+ "epoch": 31.68471615720524,
+ "grad_norm": 0.014689362607896328,
+ "learning_rate": 0.0006,
+ "loss": 3.9702324867248535,
+ "step": 2281
+ },
+ {
+ "epoch": 31.69868995633188,
+ "grad_norm": 0.015245123766362667,
+ "learning_rate": 0.0006,
+ "loss": 4.197025775909424,
+ "step": 2282
+ },
+ {
+ "epoch": 31.712663755458514,
+ "grad_norm": 0.016148347407579422,
+ "learning_rate": 0.0006,
+ "loss": 4.231856346130371,
+ "step": 2283
+ },
+ {
+ "epoch": 31.726637554585153,
+ "grad_norm": 0.015169818885624409,
+ "learning_rate": 0.0006,
+ "loss": 4.230169296264648,
+ "step": 2284
+ },
+ {
+ "epoch": 31.74061135371179,
+ "grad_norm": 0.014843013137578964,
+ "learning_rate": 0.0006,
+ "loss": 4.195633411407471,
+ "step": 2285
+ },
+ {
+ "epoch": 31.754585152838427,
+ "grad_norm": 0.015229340642690659,
+ "learning_rate": 0.0006,
+ "loss": 4.231280326843262,
+ "step": 2286
+ },
+ {
+ "epoch": 31.768558951965066,
+ "grad_norm": 0.01551708485931158,
+ "learning_rate": 0.0006,
+ "loss": 4.17064094543457,
+ "step": 2287
+ },
+ {
+ "epoch": 31.782532751091704,
+ "grad_norm": 0.015863575041294098,
+ "learning_rate": 0.0006,
+ "loss": 4.1579484939575195,
+ "step": 2288
+ },
+ {
+ "epoch": 31.79650655021834,
+ "grad_norm": 0.015118198469281197,
+ "learning_rate": 0.0006,
+ "loss": 4.1626787185668945,
+ "step": 2289
+ },
+ {
+ "epoch": 31.810480349344978,
+ "grad_norm": 0.015220776200294495,
+ "learning_rate": 0.0006,
+ "loss": 4.135746955871582,
+ "step": 2290
+ },
+ {
+ "epoch": 31.824454148471617,
+ "grad_norm": 0.014529784210026264,
+ "learning_rate": 0.0006,
+ "loss": 4.164931297302246,
+ "step": 2291
+ },
+ {
+ "epoch": 31.83842794759825,
+ "grad_norm": 0.015533296391367912,
+ "learning_rate": 0.0006,
+ "loss": 4.112464904785156,
+ "step": 2292
+ },
+ {
+ "epoch": 31.85240174672489,
+ "grad_norm": 0.016562335193157196,
+ "learning_rate": 0.0006,
+ "loss": 4.164898872375488,
+ "step": 2293
+ },
+ {
+ "epoch": 31.86637554585153,
+ "grad_norm": 0.016258113086223602,
+ "learning_rate": 0.0006,
+ "loss": 4.1979265213012695,
+ "step": 2294
+ },
+ {
+ "epoch": 31.880349344978168,
+ "grad_norm": 0.01558110024780035,
+ "learning_rate": 0.0006,
+ "loss": 4.174675941467285,
+ "step": 2295
+ },
+ {
+ "epoch": 31.894323144104803,
+ "grad_norm": 0.016412515193223953,
+ "learning_rate": 0.0006,
+ "loss": 4.217909336090088,
+ "step": 2296
+ },
+ {
+ "epoch": 31.90829694323144,
+ "grad_norm": 0.015118442475795746,
+ "learning_rate": 0.0006,
+ "loss": 4.137143611907959,
+ "step": 2297
+ },
+ {
+ "epoch": 31.92227074235808,
+ "grad_norm": 0.01506161317229271,
+ "learning_rate": 0.0006,
+ "loss": 4.196122169494629,
+ "step": 2298
+ },
+ {
+ "epoch": 31.936244541484715,
+ "grad_norm": 0.017165271565318108,
+ "learning_rate": 0.0006,
+ "loss": 4.14834451675415,
+ "step": 2299
+ },
+ {
+ "epoch": 31.950218340611354,
+ "grad_norm": 0.01696440577507019,
+ "learning_rate": 0.0006,
+ "loss": 4.218868732452393,
+ "step": 2300
+ },
+ {
+ "epoch": 31.964192139737992,
+ "grad_norm": 0.017349394038319588,
+ "learning_rate": 0.0006,
+ "loss": 4.2754011154174805,
+ "step": 2301
+ },
+ {
+ "epoch": 31.978165938864628,
+ "grad_norm": 0.018387921154499054,
+ "learning_rate": 0.0006,
+ "loss": 4.282623291015625,
+ "step": 2302
+ },
+ {
+ "epoch": 31.992139737991266,
+ "grad_norm": 0.018837926909327507,
+ "learning_rate": 0.0006,
+ "loss": 4.185425281524658,
+ "step": 2303
+ },
+ {
+ "epoch": 32.0,
+ "grad_norm": 0.018998464569449425,
+ "learning_rate": 0.0006,
+ "loss": 4.151291847229004,
+ "step": 2304
+ },
+ {
+ "epoch": 32.0,
+ "eval_loss": 4.522614002227783,
+ "eval_runtime": 57.0106,
+ "eval_samples_per_second": 42.834,
+ "eval_steps_per_second": 1.351,
+ "step": 2304
+ },
+ {
+ "epoch": 32.01397379912664,
+ "grad_norm": 0.019036833196878433,
+ "learning_rate": 0.0006,
+ "loss": 4.183202266693115,
+ "step": 2305
+ },
+ {
+ "epoch": 32.02794759825328,
+ "grad_norm": 0.02207397110760212,
+ "learning_rate": 0.0006,
+ "loss": 4.17875862121582,
+ "step": 2306
+ },
+ {
+ "epoch": 32.041921397379916,
+ "grad_norm": 0.022008279338479042,
+ "learning_rate": 0.0006,
+ "loss": 4.08042049407959,
+ "step": 2307
+ },
+ {
+ "epoch": 32.05589519650655,
+ "grad_norm": 0.022407017648220062,
+ "learning_rate": 0.0006,
+ "loss": 4.112189769744873,
+ "step": 2308
+ },
+ {
+ "epoch": 32.069868995633186,
+ "grad_norm": 0.021599747240543365,
+ "learning_rate": 0.0006,
+ "loss": 4.126455307006836,
+ "step": 2309
+ },
+ {
+ "epoch": 32.083842794759825,
+ "grad_norm": 0.020340966060757637,
+ "learning_rate": 0.0006,
+ "loss": 4.080538272857666,
+ "step": 2310
+ },
+ {
+ "epoch": 32.09781659388646,
+ "grad_norm": 0.020956892520189285,
+ "learning_rate": 0.0006,
+ "loss": 4.171943664550781,
+ "step": 2311
+ },
+ {
+ "epoch": 32.1117903930131,
+ "grad_norm": 0.019966667518019676,
+ "learning_rate": 0.0006,
+ "loss": 4.144007682800293,
+ "step": 2312
+ },
+ {
+ "epoch": 32.12576419213974,
+ "grad_norm": 0.018256952986121178,
+ "learning_rate": 0.0006,
+ "loss": 4.208107948303223,
+ "step": 2313
+ },
+ {
+ "epoch": 32.13973799126637,
+ "grad_norm": 0.017289170995354652,
+ "learning_rate": 0.0006,
+ "loss": 4.182024955749512,
+ "step": 2314
+ },
+ {
+ "epoch": 32.15371179039301,
+ "grad_norm": 0.017656700685620308,
+ "learning_rate": 0.0006,
+ "loss": 4.068373680114746,
+ "step": 2315
+ },
+ {
+ "epoch": 32.16768558951965,
+ "grad_norm": 0.017903361469507217,
+ "learning_rate": 0.0006,
+ "loss": 4.2345428466796875,
+ "step": 2316
+ },
+ {
+ "epoch": 32.18165938864629,
+ "grad_norm": 0.01785109005868435,
+ "learning_rate": 0.0006,
+ "loss": 4.219776153564453,
+ "step": 2317
+ },
+ {
+ "epoch": 32.19563318777293,
+ "grad_norm": 0.01887257769703865,
+ "learning_rate": 0.0006,
+ "loss": 4.139206409454346,
+ "step": 2318
+ },
+ {
+ "epoch": 32.209606986899566,
+ "grad_norm": 0.01924685761332512,
+ "learning_rate": 0.0006,
+ "loss": 4.121052265167236,
+ "step": 2319
+ },
+ {
+ "epoch": 32.223580786026204,
+ "grad_norm": 0.020574072375893593,
+ "learning_rate": 0.0006,
+ "loss": 4.13262414932251,
+ "step": 2320
+ },
+ {
+ "epoch": 32.237554585152836,
+ "grad_norm": 0.023849064484238625,
+ "learning_rate": 0.0006,
+ "loss": 4.239119529724121,
+ "step": 2321
+ },
+ {
+ "epoch": 32.251528384279474,
+ "grad_norm": 0.026239361613988876,
+ "learning_rate": 0.0006,
+ "loss": 4.248993873596191,
+ "step": 2322
+ },
+ {
+ "epoch": 32.26550218340611,
+ "grad_norm": 0.023980243131518364,
+ "learning_rate": 0.0006,
+ "loss": 4.159852504730225,
+ "step": 2323
+ },
+ {
+ "epoch": 32.27947598253275,
+ "grad_norm": 0.020442405715584755,
+ "learning_rate": 0.0006,
+ "loss": 4.133546829223633,
+ "step": 2324
+ },
+ {
+ "epoch": 32.29344978165939,
+ "grad_norm": 0.020661186426877975,
+ "learning_rate": 0.0006,
+ "loss": 4.078243732452393,
+ "step": 2325
+ },
+ {
+ "epoch": 32.30742358078603,
+ "grad_norm": 0.02046387456357479,
+ "learning_rate": 0.0006,
+ "loss": 4.172012805938721,
+ "step": 2326
+ },
+ {
+ "epoch": 32.32139737991266,
+ "grad_norm": 0.02089597098529339,
+ "learning_rate": 0.0006,
+ "loss": 4.0604705810546875,
+ "step": 2327
+ },
+ {
+ "epoch": 32.3353711790393,
+ "grad_norm": 0.01875591278076172,
+ "learning_rate": 0.0006,
+ "loss": 4.125911235809326,
+ "step": 2328
+ },
+ {
+ "epoch": 32.34934497816594,
+ "grad_norm": 0.018606796860694885,
+ "learning_rate": 0.0006,
+ "loss": 4.097362518310547,
+ "step": 2329
+ },
+ {
+ "epoch": 32.36331877729258,
+ "grad_norm": 0.01766626164317131,
+ "learning_rate": 0.0006,
+ "loss": 4.118616580963135,
+ "step": 2330
+ },
+ {
+ "epoch": 32.377292576419215,
+ "grad_norm": 0.017692960798740387,
+ "learning_rate": 0.0006,
+ "loss": 4.1457672119140625,
+ "step": 2331
+ },
+ {
+ "epoch": 32.391266375545854,
+ "grad_norm": 0.017619704827666283,
+ "learning_rate": 0.0006,
+ "loss": 4.175961494445801,
+ "step": 2332
+ },
+ {
+ "epoch": 32.40524017467249,
+ "grad_norm": 0.01784413494169712,
+ "learning_rate": 0.0006,
+ "loss": 4.051291465759277,
+ "step": 2333
+ },
+ {
+ "epoch": 32.419213973799124,
+ "grad_norm": 0.018508171662688255,
+ "learning_rate": 0.0006,
+ "loss": 4.0638556480407715,
+ "step": 2334
+ },
+ {
+ "epoch": 32.43318777292576,
+ "grad_norm": 0.01731436885893345,
+ "learning_rate": 0.0006,
+ "loss": 4.1805419921875,
+ "step": 2335
+ },
+ {
+ "epoch": 32.4471615720524,
+ "grad_norm": 0.015886498615145683,
+ "learning_rate": 0.0006,
+ "loss": 4.151026725769043,
+ "step": 2336
+ },
+ {
+ "epoch": 32.46113537117904,
+ "grad_norm": 0.016583820804953575,
+ "learning_rate": 0.0006,
+ "loss": 4.203604698181152,
+ "step": 2337
+ },
+ {
+ "epoch": 32.47510917030568,
+ "grad_norm": 0.0176097322255373,
+ "learning_rate": 0.0006,
+ "loss": 4.135159015655518,
+ "step": 2338
+ },
+ {
+ "epoch": 32.48908296943232,
+ "grad_norm": 0.016657903790473938,
+ "learning_rate": 0.0006,
+ "loss": 4.126681804656982,
+ "step": 2339
+ },
+ {
+ "epoch": 32.50305676855895,
+ "grad_norm": 0.016866130754351616,
+ "learning_rate": 0.0006,
+ "loss": 4.162341117858887,
+ "step": 2340
+ },
+ {
+ "epoch": 32.51703056768559,
+ "grad_norm": 0.016286680474877357,
+ "learning_rate": 0.0006,
+ "loss": 4.261801242828369,
+ "step": 2341
+ },
+ {
+ "epoch": 32.531004366812226,
+ "grad_norm": 0.01481365505605936,
+ "learning_rate": 0.0006,
+ "loss": 4.110016822814941,
+ "step": 2342
+ },
+ {
+ "epoch": 32.544978165938865,
+ "grad_norm": 0.016605013981461525,
+ "learning_rate": 0.0006,
+ "loss": 4.230024814605713,
+ "step": 2343
+ },
+ {
+ "epoch": 32.5589519650655,
+ "grad_norm": 0.01645175740122795,
+ "learning_rate": 0.0006,
+ "loss": 4.174612045288086,
+ "step": 2344
+ },
+ {
+ "epoch": 32.57292576419214,
+ "grad_norm": 0.014725026674568653,
+ "learning_rate": 0.0006,
+ "loss": 4.066371440887451,
+ "step": 2345
+ },
+ {
+ "epoch": 32.58689956331878,
+ "grad_norm": 0.015003837645053864,
+ "learning_rate": 0.0006,
+ "loss": 4.193343639373779,
+ "step": 2346
+ },
+ {
+ "epoch": 32.60087336244541,
+ "grad_norm": 0.015346302650868893,
+ "learning_rate": 0.0006,
+ "loss": 4.210659980773926,
+ "step": 2347
+ },
+ {
+ "epoch": 32.61484716157205,
+ "grad_norm": 0.015135962516069412,
+ "learning_rate": 0.0006,
+ "loss": 4.204618453979492,
+ "step": 2348
+ },
+ {
+ "epoch": 32.62882096069869,
+ "grad_norm": 0.01601956970989704,
+ "learning_rate": 0.0006,
+ "loss": 4.145564556121826,
+ "step": 2349
+ },
+ {
+ "epoch": 32.64279475982533,
+ "grad_norm": 0.01517851185053587,
+ "learning_rate": 0.0006,
+ "loss": 4.051673889160156,
+ "step": 2350
+ },
+ {
+ "epoch": 32.65676855895197,
+ "grad_norm": 0.014894292689859867,
+ "learning_rate": 0.0006,
+ "loss": 4.184818267822266,
+ "step": 2351
+ },
+ {
+ "epoch": 32.670742358078606,
+ "grad_norm": 0.015720834955573082,
+ "learning_rate": 0.0006,
+ "loss": 4.220036506652832,
+ "step": 2352
+ },
+ {
+ "epoch": 32.68471615720524,
+ "grad_norm": 0.01573433168232441,
+ "learning_rate": 0.0006,
+ "loss": 4.076773166656494,
+ "step": 2353
+ },
+ {
+ "epoch": 32.698689956331876,
+ "grad_norm": 0.015828052535653114,
+ "learning_rate": 0.0006,
+ "loss": 4.116329669952393,
+ "step": 2354
+ },
+ {
+ "epoch": 32.712663755458514,
+ "grad_norm": 0.015462259761989117,
+ "learning_rate": 0.0006,
+ "loss": 4.1155242919921875,
+ "step": 2355
+ },
+ {
+ "epoch": 32.72663755458515,
+ "grad_norm": 0.015763849020004272,
+ "learning_rate": 0.0006,
+ "loss": 4.169687747955322,
+ "step": 2356
+ },
+ {
+ "epoch": 32.74061135371179,
+ "grad_norm": 0.014935145154595375,
+ "learning_rate": 0.0006,
+ "loss": 4.2090911865234375,
+ "step": 2357
+ },
+ {
+ "epoch": 32.75458515283843,
+ "grad_norm": 0.015119184739887714,
+ "learning_rate": 0.0006,
+ "loss": 4.131838798522949,
+ "step": 2358
+ },
+ {
+ "epoch": 32.76855895196506,
+ "grad_norm": 0.014833525754511356,
+ "learning_rate": 0.0006,
+ "loss": 4.17618465423584,
+ "step": 2359
+ },
+ {
+ "epoch": 32.7825327510917,
+ "grad_norm": 0.01588359661400318,
+ "learning_rate": 0.0006,
+ "loss": 4.163723945617676,
+ "step": 2360
+ },
+ {
+ "epoch": 32.79650655021834,
+ "grad_norm": 0.0168188214302063,
+ "learning_rate": 0.0006,
+ "loss": 4.178189277648926,
+ "step": 2361
+ },
+ {
+ "epoch": 32.81048034934498,
+ "grad_norm": 0.01726095750927925,
+ "learning_rate": 0.0006,
+ "loss": 4.144223690032959,
+ "step": 2362
+ },
+ {
+ "epoch": 32.82445414847162,
+ "grad_norm": 0.01701410301029682,
+ "learning_rate": 0.0006,
+ "loss": 4.209896087646484,
+ "step": 2363
+ },
+ {
+ "epoch": 32.838427947598255,
+ "grad_norm": 0.01682264916598797,
+ "learning_rate": 0.0006,
+ "loss": 4.26223087310791,
+ "step": 2364
+ },
+ {
+ "epoch": 32.852401746724894,
+ "grad_norm": 0.015597456134855747,
+ "learning_rate": 0.0006,
+ "loss": 4.292664051055908,
+ "step": 2365
+ },
+ {
+ "epoch": 32.866375545851525,
+ "grad_norm": 0.015258579514920712,
+ "learning_rate": 0.0006,
+ "loss": 4.165006637573242,
+ "step": 2366
+ },
+ {
+ "epoch": 32.880349344978164,
+ "grad_norm": 0.01662200503051281,
+ "learning_rate": 0.0006,
+ "loss": 4.309786319732666,
+ "step": 2367
+ },
+ {
+ "epoch": 32.8943231441048,
+ "grad_norm": 0.015691161155700684,
+ "learning_rate": 0.0006,
+ "loss": 4.167202949523926,
+ "step": 2368
+ },
+ {
+ "epoch": 32.90829694323144,
+ "grad_norm": 0.014870637096464634,
+ "learning_rate": 0.0006,
+ "loss": 4.142415523529053,
+ "step": 2369
+ },
+ {
+ "epoch": 32.92227074235808,
+ "grad_norm": 0.01567252166569233,
+ "learning_rate": 0.0006,
+ "loss": 4.2312092781066895,
+ "step": 2370
+ },
+ {
+ "epoch": 32.93624454148472,
+ "grad_norm": 0.014948616735637188,
+ "learning_rate": 0.0006,
+ "loss": 4.1514129638671875,
+ "step": 2371
+ },
+ {
+ "epoch": 32.95021834061135,
+ "grad_norm": 0.014654500409960747,
+ "learning_rate": 0.0006,
+ "loss": 4.205010890960693,
+ "step": 2372
+ },
+ {
+ "epoch": 32.96419213973799,
+ "grad_norm": 0.015059148892760277,
+ "learning_rate": 0.0006,
+ "loss": 4.148405075073242,
+ "step": 2373
+ },
+ {
+ "epoch": 32.97816593886463,
+ "grad_norm": 0.015026576817035675,
+ "learning_rate": 0.0006,
+ "loss": 4.182535171508789,
+ "step": 2374
+ },
+ {
+ "epoch": 32.992139737991266,
+ "grad_norm": 0.016394320875406265,
+ "learning_rate": 0.0006,
+ "loss": 4.129051208496094,
+ "step": 2375
+ },
+ {
+ "epoch": 33.0,
+ "grad_norm": 0.01867043413221836,
+ "learning_rate": 0.0006,
+ "loss": 4.185556411743164,
+ "step": 2376
+ },
+ {
+ "epoch": 33.0,
+ "eval_loss": 4.640405654907227,
+ "eval_runtime": 56.5861,
+ "eval_samples_per_second": 43.155,
+ "eval_steps_per_second": 1.361,
+ "step": 2376
+ },
+ {
+ "epoch": 33.01397379912664,
+ "grad_norm": 0.018626727163791656,
+ "learning_rate": 0.0006,
+ "loss": 4.143031597137451,
+ "step": 2377
+ },
+ {
+ "epoch": 33.02794759825328,
+ "grad_norm": 0.019283363595604897,
+ "learning_rate": 0.0006,
+ "loss": 4.117845058441162,
+ "step": 2378
+ },
+ {
+ "epoch": 33.041921397379916,
+ "grad_norm": 0.018185822293162346,
+ "learning_rate": 0.0006,
+ "loss": 4.076862812042236,
+ "step": 2379
+ },
+ {
+ "epoch": 33.05589519650655,
+ "grad_norm": 0.019411196932196617,
+ "learning_rate": 0.0006,
+ "loss": 4.124156475067139,
+ "step": 2380
+ },
+ {
+ "epoch": 33.069868995633186,
+ "grad_norm": 0.019248466938734055,
+ "learning_rate": 0.0006,
+ "loss": 4.201735019683838,
+ "step": 2381
+ },
+ {
+ "epoch": 33.083842794759825,
+ "grad_norm": 0.018856720998883247,
+ "learning_rate": 0.0006,
+ "loss": 4.00296688079834,
+ "step": 2382
+ },
+ {
+ "epoch": 33.09781659388646,
+ "grad_norm": 0.018554294481873512,
+ "learning_rate": 0.0006,
+ "loss": 4.270577430725098,
+ "step": 2383
+ },
+ {
+ "epoch": 33.1117903930131,
+ "grad_norm": 0.019946224987506866,
+ "learning_rate": 0.0006,
+ "loss": 4.119589328765869,
+ "step": 2384
+ },
+ {
+ "epoch": 33.12576419213974,
+ "grad_norm": 0.021395670250058174,
+ "learning_rate": 0.0006,
+ "loss": 4.128847122192383,
+ "step": 2385
+ },
+ {
+ "epoch": 33.13973799126637,
+ "grad_norm": 0.02009674906730652,
+ "learning_rate": 0.0006,
+ "loss": 4.183228969573975,
+ "step": 2386
+ },
+ {
+ "epoch": 33.15371179039301,
+ "grad_norm": 0.01820223033428192,
+ "learning_rate": 0.0006,
+ "loss": 4.131601810455322,
+ "step": 2387
+ },
+ {
+ "epoch": 33.16768558951965,
+ "grad_norm": 0.018653156235814095,
+ "learning_rate": 0.0006,
+ "loss": 4.116430282592773,
+ "step": 2388
+ },
+ {
+ "epoch": 33.18165938864629,
+ "grad_norm": 0.016828063875436783,
+ "learning_rate": 0.0006,
+ "loss": 4.054663181304932,
+ "step": 2389
+ },
+ {
+ "epoch": 33.19563318777293,
+ "grad_norm": 0.017025409266352654,
+ "learning_rate": 0.0006,
+ "loss": 4.110044956207275,
+ "step": 2390
+ },
+ {
+ "epoch": 33.209606986899566,
+ "grad_norm": 0.016608452424407005,
+ "learning_rate": 0.0006,
+ "loss": 4.102668285369873,
+ "step": 2391
+ },
+ {
+ "epoch": 33.223580786026204,
+ "grad_norm": 0.01743577979505062,
+ "learning_rate": 0.0006,
+ "loss": 4.166791915893555,
+ "step": 2392
+ },
+ {
+ "epoch": 33.237554585152836,
+ "grad_norm": 0.017253734171390533,
+ "learning_rate": 0.0006,
+ "loss": 4.217832088470459,
+ "step": 2393
+ },
+ {
+ "epoch": 33.251528384279474,
+ "grad_norm": 0.017937595024704933,
+ "learning_rate": 0.0006,
+ "loss": 4.2173967361450195,
+ "step": 2394
+ },
+ {
+ "epoch": 33.26550218340611,
+ "grad_norm": 0.01985993981361389,
+ "learning_rate": 0.0006,
+ "loss": 4.17849588394165,
+ "step": 2395
+ },
+ {
+ "epoch": 33.27947598253275,
+ "grad_norm": 0.019243355840444565,
+ "learning_rate": 0.0006,
+ "loss": 4.086523056030273,
+ "step": 2396
+ },
+ {
+ "epoch": 33.29344978165939,
+ "grad_norm": 0.01924416422843933,
+ "learning_rate": 0.0006,
+ "loss": 4.128574371337891,
+ "step": 2397
+ },
+ {
+ "epoch": 33.30742358078603,
+ "grad_norm": 0.01983758620917797,
+ "learning_rate": 0.0006,
+ "loss": 4.061581611633301,
+ "step": 2398
+ },
+ {
+ "epoch": 33.32139737991266,
+ "grad_norm": 0.01930462196469307,
+ "learning_rate": 0.0006,
+ "loss": 4.091731071472168,
+ "step": 2399
+ },
+ {
+ "epoch": 33.3353711790393,
+ "grad_norm": 0.017988424748182297,
+ "learning_rate": 0.0006,
+ "loss": 4.155599117279053,
+ "step": 2400
+ },
+ {
+ "epoch": 33.34934497816594,
+ "grad_norm": 0.015194767154753208,
+ "learning_rate": 0.0006,
+ "loss": 4.144134044647217,
+ "step": 2401
+ },
+ {
+ "epoch": 33.36331877729258,
+ "grad_norm": 0.01536081824451685,
+ "learning_rate": 0.0006,
+ "loss": 4.127272129058838,
+ "step": 2402
+ },
+ {
+ "epoch": 33.377292576419215,
+ "grad_norm": 0.015431297942996025,
+ "learning_rate": 0.0006,
+ "loss": 4.3213019371032715,
+ "step": 2403
+ },
+ {
+ "epoch": 33.391266375545854,
+ "grad_norm": 0.015740280970931053,
+ "learning_rate": 0.0006,
+ "loss": 4.121467590332031,
+ "step": 2404
+ },
+ {
+ "epoch": 33.40524017467249,
+ "grad_norm": 0.015379384160041809,
+ "learning_rate": 0.0006,
+ "loss": 4.095216751098633,
+ "step": 2405
+ },
+ {
+ "epoch": 33.419213973799124,
+ "grad_norm": 0.01528073474764824,
+ "learning_rate": 0.0006,
+ "loss": 4.13047981262207,
+ "step": 2406
+ },
+ {
+ "epoch": 33.43318777292576,
+ "grad_norm": 0.015739979222416878,
+ "learning_rate": 0.0006,
+ "loss": 4.2533135414123535,
+ "step": 2407
+ },
+ {
+ "epoch": 33.4471615720524,
+ "grad_norm": 0.01519764307886362,
+ "learning_rate": 0.0006,
+ "loss": 4.131782054901123,
+ "step": 2408
+ },
+ {
+ "epoch": 33.46113537117904,
+ "grad_norm": 0.015425996854901314,
+ "learning_rate": 0.0006,
+ "loss": 4.02811336517334,
+ "step": 2409
+ },
+ {
+ "epoch": 33.47510917030568,
+ "grad_norm": 0.015378830023109913,
+ "learning_rate": 0.0006,
+ "loss": 4.08652925491333,
+ "step": 2410
+ },
+ {
+ "epoch": 33.48908296943232,
+ "grad_norm": 0.01473075058311224,
+ "learning_rate": 0.0006,
+ "loss": 4.214364528656006,
+ "step": 2411
+ },
+ {
+ "epoch": 33.50305676855895,
+ "grad_norm": 0.015141637064516544,
+ "learning_rate": 0.0006,
+ "loss": 4.155712127685547,
+ "step": 2412
+ },
+ {
+ "epoch": 33.51703056768559,
+ "grad_norm": 0.01641593500971794,
+ "learning_rate": 0.0006,
+ "loss": 4.097907066345215,
+ "step": 2413
+ },
+ {
+ "epoch": 33.531004366812226,
+ "grad_norm": 0.015182029455900192,
+ "learning_rate": 0.0006,
+ "loss": 4.238499641418457,
+ "step": 2414
+ },
+ {
+ "epoch": 33.544978165938865,
+ "grad_norm": 0.01643732376396656,
+ "learning_rate": 0.0006,
+ "loss": 4.0661492347717285,
+ "step": 2415
+ },
+ {
+ "epoch": 33.5589519650655,
+ "grad_norm": 0.015614228323101997,
+ "learning_rate": 0.0006,
+ "loss": 4.120795249938965,
+ "step": 2416
+ },
+ {
+ "epoch": 33.57292576419214,
+ "grad_norm": 0.016066187992691994,
+ "learning_rate": 0.0006,
+ "loss": 4.096465110778809,
+ "step": 2417
+ },
+ {
+ "epoch": 33.58689956331878,
+ "grad_norm": 0.017214342951774597,
+ "learning_rate": 0.0006,
+ "loss": 4.242628574371338,
+ "step": 2418
+ },
+ {
+ "epoch": 33.60087336244541,
+ "grad_norm": 0.01787586510181427,
+ "learning_rate": 0.0006,
+ "loss": 4.149589538574219,
+ "step": 2419
+ },
+ {
+ "epoch": 33.61484716157205,
+ "grad_norm": 0.019455241039395332,
+ "learning_rate": 0.0006,
+ "loss": 4.267796039581299,
+ "step": 2420
+ },
+ {
+ "epoch": 33.62882096069869,
+ "grad_norm": 0.01984674669802189,
+ "learning_rate": 0.0006,
+ "loss": 4.175220966339111,
+ "step": 2421
+ },
+ {
+ "epoch": 33.64279475982533,
+ "grad_norm": 0.021933717653155327,
+ "learning_rate": 0.0006,
+ "loss": 4.106647491455078,
+ "step": 2422
+ },
+ {
+ "epoch": 33.65676855895197,
+ "grad_norm": 0.02123589813709259,
+ "learning_rate": 0.0006,
+ "loss": 4.204996109008789,
+ "step": 2423
+ },
+ {
+ "epoch": 33.670742358078606,
+ "grad_norm": 0.019814589992165565,
+ "learning_rate": 0.0006,
+ "loss": 4.186171531677246,
+ "step": 2424
+ },
+ {
+ "epoch": 33.68471615720524,
+ "grad_norm": 0.02059728465974331,
+ "learning_rate": 0.0006,
+ "loss": 4.170509338378906,
+ "step": 2425
+ },
+ {
+ "epoch": 33.698689956331876,
+ "grad_norm": 0.018483608961105347,
+ "learning_rate": 0.0006,
+ "loss": 4.085585594177246,
+ "step": 2426
+ },
+ {
+ "epoch": 33.712663755458514,
+ "grad_norm": 0.018403852358460426,
+ "learning_rate": 0.0006,
+ "loss": 4.2551774978637695,
+ "step": 2427
+ },
+ {
+ "epoch": 33.72663755458515,
+ "grad_norm": 0.018653830513358116,
+ "learning_rate": 0.0006,
+ "loss": 4.091766357421875,
+ "step": 2428
+ },
+ {
+ "epoch": 33.74061135371179,
+ "grad_norm": 0.017692334949970245,
+ "learning_rate": 0.0006,
+ "loss": 4.100838661193848,
+ "step": 2429
+ },
+ {
+ "epoch": 33.75458515283843,
+ "grad_norm": 0.016029126942157745,
+ "learning_rate": 0.0006,
+ "loss": 4.1530680656433105,
+ "step": 2430
+ },
+ {
+ "epoch": 33.76855895196506,
+ "grad_norm": 0.016095684841275215,
+ "learning_rate": 0.0006,
+ "loss": 4.20544958114624,
+ "step": 2431
+ },
+ {
+ "epoch": 33.7825327510917,
+ "grad_norm": 0.016375798732042313,
+ "learning_rate": 0.0006,
+ "loss": 4.199177265167236,
+ "step": 2432
+ },
+ {
+ "epoch": 33.79650655021834,
+ "grad_norm": 0.01590786688029766,
+ "learning_rate": 0.0006,
+ "loss": 4.200799465179443,
+ "step": 2433
+ },
+ {
+ "epoch": 33.81048034934498,
+ "grad_norm": 0.014896097593009472,
+ "learning_rate": 0.0006,
+ "loss": 4.185850143432617,
+ "step": 2434
+ },
+ {
+ "epoch": 33.82445414847162,
+ "grad_norm": 0.01438513956964016,
+ "learning_rate": 0.0006,
+ "loss": 4.214962005615234,
+ "step": 2435
+ },
+ {
+ "epoch": 33.838427947598255,
+ "grad_norm": 0.014459339901804924,
+ "learning_rate": 0.0006,
+ "loss": 4.196972370147705,
+ "step": 2436
+ },
+ {
+ "epoch": 33.852401746724894,
+ "grad_norm": 0.014782169833779335,
+ "learning_rate": 0.0006,
+ "loss": 4.0885910987854,
+ "step": 2437
+ },
+ {
+ "epoch": 33.866375545851525,
+ "grad_norm": 0.01520370040088892,
+ "learning_rate": 0.0006,
+ "loss": 4.0323262214660645,
+ "step": 2438
+ },
+ {
+ "epoch": 33.880349344978164,
+ "grad_norm": 0.014056282117962837,
+ "learning_rate": 0.0006,
+ "loss": 4.1386518478393555,
+ "step": 2439
+ },
+ {
+ "epoch": 33.8943231441048,
+ "grad_norm": 0.013567990623414516,
+ "learning_rate": 0.0006,
+ "loss": 4.095754623413086,
+ "step": 2440
+ },
+ {
+ "epoch": 33.90829694323144,
+ "grad_norm": 0.013306716457009315,
+ "learning_rate": 0.0006,
+ "loss": 4.1623077392578125,
+ "step": 2441
+ },
+ {
+ "epoch": 33.92227074235808,
+ "grad_norm": 0.014047150500118732,
+ "learning_rate": 0.0006,
+ "loss": 4.142541885375977,
+ "step": 2442
+ },
+ {
+ "epoch": 33.93624454148472,
+ "grad_norm": 0.01377617847174406,
+ "learning_rate": 0.0006,
+ "loss": 4.268551826477051,
+ "step": 2443
+ },
+ {
+ "epoch": 33.95021834061135,
+ "grad_norm": 0.014671621844172478,
+ "learning_rate": 0.0006,
+ "loss": 4.198196887969971,
+ "step": 2444
+ },
+ {
+ "epoch": 33.96419213973799,
+ "grad_norm": 0.014806744642555714,
+ "learning_rate": 0.0006,
+ "loss": 4.057899475097656,
+ "step": 2445
+ },
+ {
+ "epoch": 33.97816593886463,
+ "grad_norm": 0.015734272077679634,
+ "learning_rate": 0.0006,
+ "loss": 4.163505554199219,
+ "step": 2446
+ },
+ {
+ "epoch": 33.992139737991266,
+ "grad_norm": 0.015250146389007568,
+ "learning_rate": 0.0006,
+ "loss": 4.076294898986816,
+ "step": 2447
+ },
+ {
+ "epoch": 34.0,
+ "grad_norm": 0.016733702272176743,
+ "learning_rate": 0.0006,
+ "loss": 4.138420104980469,
+ "step": 2448
+ },
+ {
+ "epoch": 34.0,
+ "eval_loss": 4.542503356933594,
+ "eval_runtime": 56.983,
+ "eval_samples_per_second": 42.855,
+ "eval_steps_per_second": 1.351,
+ "step": 2448
+ },
+ {
+ "epoch": 34.01397379912664,
+ "grad_norm": 0.015866613015532494,
+ "learning_rate": 0.0006,
+ "loss": 4.032081127166748,
+ "step": 2449
+ },
+ {
+ "epoch": 34.02794759825328,
+ "grad_norm": 0.01946570910513401,
+ "learning_rate": 0.0006,
+ "loss": 4.122320175170898,
+ "step": 2450
+ },
+ {
+ "epoch": 34.041921397379916,
+ "grad_norm": 0.020265938714146614,
+ "learning_rate": 0.0006,
+ "loss": 4.06460428237915,
+ "step": 2451
+ },
+ {
+ "epoch": 34.05589519650655,
+ "grad_norm": 0.018887171521782875,
+ "learning_rate": 0.0006,
+ "loss": 4.096906661987305,
+ "step": 2452
+ },
+ {
+ "epoch": 34.069868995633186,
+ "grad_norm": 0.017304692417383194,
+ "learning_rate": 0.0006,
+ "loss": 4.137261390686035,
+ "step": 2453
+ },
+ {
+ "epoch": 34.083842794759825,
+ "grad_norm": 0.017029082402586937,
+ "learning_rate": 0.0006,
+ "loss": 4.081112861633301,
+ "step": 2454
+ },
+ {
+ "epoch": 34.09781659388646,
+ "grad_norm": 0.016414813697338104,
+ "learning_rate": 0.0006,
+ "loss": 4.140783309936523,
+ "step": 2455
+ },
+ {
+ "epoch": 34.1117903930131,
+ "grad_norm": 0.018238281831145287,
+ "learning_rate": 0.0006,
+ "loss": 4.161116600036621,
+ "step": 2456
+ },
+ {
+ "epoch": 34.12576419213974,
+ "grad_norm": 0.019929232075810432,
+ "learning_rate": 0.0006,
+ "loss": 4.243722438812256,
+ "step": 2457
+ },
+ {
+ "epoch": 34.13973799126637,
+ "grad_norm": 0.019375786185264587,
+ "learning_rate": 0.0006,
+ "loss": 4.134873390197754,
+ "step": 2458
+ },
+ {
+ "epoch": 34.15371179039301,
+ "grad_norm": 0.018151137977838516,
+ "learning_rate": 0.0006,
+ "loss": 4.098745346069336,
+ "step": 2459
+ },
+ {
+ "epoch": 34.16768558951965,
+ "grad_norm": 0.017054934054613113,
+ "learning_rate": 0.0006,
+ "loss": 4.194662570953369,
+ "step": 2460
+ },
+ {
+ "epoch": 34.18165938864629,
+ "grad_norm": 0.018220555037260056,
+ "learning_rate": 0.0006,
+ "loss": 4.128866195678711,
+ "step": 2461
+ },
+ {
+ "epoch": 34.19563318777293,
+ "grad_norm": 0.01692879945039749,
+ "learning_rate": 0.0006,
+ "loss": 4.062477111816406,
+ "step": 2462
+ },
+ {
+ "epoch": 34.209606986899566,
+ "grad_norm": 0.01674199476838112,
+ "learning_rate": 0.0006,
+ "loss": 4.186149597167969,
+ "step": 2463
+ },
+ {
+ "epoch": 34.223580786026204,
+ "grad_norm": 0.016527455300092697,
+ "learning_rate": 0.0006,
+ "loss": 4.0734028816223145,
+ "step": 2464
+ },
+ {
+ "epoch": 34.237554585152836,
+ "grad_norm": 0.016448386013507843,
+ "learning_rate": 0.0006,
+ "loss": 4.098896026611328,
+ "step": 2465
+ },
+ {
+ "epoch": 34.251528384279474,
+ "grad_norm": 0.014644953422248363,
+ "learning_rate": 0.0006,
+ "loss": 4.1031904220581055,
+ "step": 2466
+ },
+ {
+ "epoch": 34.26550218340611,
+ "grad_norm": 0.016649341210722923,
+ "learning_rate": 0.0006,
+ "loss": 4.148791313171387,
+ "step": 2467
+ },
+ {
+ "epoch": 34.27947598253275,
+ "grad_norm": 0.017442714422941208,
+ "learning_rate": 0.0006,
+ "loss": 4.070799827575684,
+ "step": 2468
+ },
+ {
+ "epoch": 34.29344978165939,
+ "grad_norm": 0.017852013930678368,
+ "learning_rate": 0.0006,
+ "loss": 4.031924247741699,
+ "step": 2469
+ },
+ {
+ "epoch": 34.30742358078603,
+ "grad_norm": 0.016371270641684532,
+ "learning_rate": 0.0006,
+ "loss": 4.149420738220215,
+ "step": 2470
+ },
+ {
+ "epoch": 34.32139737991266,
+ "grad_norm": 0.015846814960241318,
+ "learning_rate": 0.0006,
+ "loss": 4.15427303314209,
+ "step": 2471
+ },
+ {
+ "epoch": 34.3353711790393,
+ "grad_norm": 0.017411163076758385,
+ "learning_rate": 0.0006,
+ "loss": 4.1304192543029785,
+ "step": 2472
+ },
+ {
+ "epoch": 34.34934497816594,
+ "grad_norm": 0.01780046708881855,
+ "learning_rate": 0.0006,
+ "loss": 4.191732406616211,
+ "step": 2473
+ },
+ {
+ "epoch": 34.36331877729258,
+ "grad_norm": 0.01958383060991764,
+ "learning_rate": 0.0006,
+ "loss": 4.068355083465576,
+ "step": 2474
+ },
+ {
+ "epoch": 34.377292576419215,
+ "grad_norm": 0.022173317149281502,
+ "learning_rate": 0.0006,
+ "loss": 4.182929039001465,
+ "step": 2475
+ },
+ {
+ "epoch": 34.391266375545854,
+ "grad_norm": 0.022389156743884087,
+ "learning_rate": 0.0006,
+ "loss": 4.110218048095703,
+ "step": 2476
+ },
+ {
+ "epoch": 34.40524017467249,
+ "grad_norm": 0.022478511556982994,
+ "learning_rate": 0.0006,
+ "loss": 4.095479488372803,
+ "step": 2477
+ },
+ {
+ "epoch": 34.419213973799124,
+ "grad_norm": 0.019504647701978683,
+ "learning_rate": 0.0006,
+ "loss": 4.101964950561523,
+ "step": 2478
+ },
+ {
+ "epoch": 34.43318777292576,
+ "grad_norm": 0.019378576427698135,
+ "learning_rate": 0.0006,
+ "loss": 4.14620304107666,
+ "step": 2479
+ },
+ {
+ "epoch": 34.4471615720524,
+ "grad_norm": 0.017993204295635223,
+ "learning_rate": 0.0006,
+ "loss": 4.183492660522461,
+ "step": 2480
+ },
+ {
+ "epoch": 34.46113537117904,
+ "grad_norm": 0.01814805157482624,
+ "learning_rate": 0.0006,
+ "loss": 4.151247978210449,
+ "step": 2481
+ },
+ {
+ "epoch": 34.47510917030568,
+ "grad_norm": 0.018622448667883873,
+ "learning_rate": 0.0006,
+ "loss": 4.020938396453857,
+ "step": 2482
+ },
+ {
+ "epoch": 34.48908296943232,
+ "grad_norm": 0.018360132351517677,
+ "learning_rate": 0.0006,
+ "loss": 4.064734935760498,
+ "step": 2483
+ },
+ {
+ "epoch": 34.50305676855895,
+ "grad_norm": 0.017715711146593094,
+ "learning_rate": 0.0006,
+ "loss": 4.185687065124512,
+ "step": 2484
+ },
+ {
+ "epoch": 34.51703056768559,
+ "grad_norm": 0.01714295893907547,
+ "learning_rate": 0.0006,
+ "loss": 4.1044721603393555,
+ "step": 2485
+ },
+ {
+ "epoch": 34.531004366812226,
+ "grad_norm": 0.016943803057074547,
+ "learning_rate": 0.0006,
+ "loss": 4.106837272644043,
+ "step": 2486
+ },
+ {
+ "epoch": 34.544978165938865,
+ "grad_norm": 0.0182423684746027,
+ "learning_rate": 0.0006,
+ "loss": 4.1041340827941895,
+ "step": 2487
+ },
+ {
+ "epoch": 34.5589519650655,
+ "grad_norm": 0.01636587455868721,
+ "learning_rate": 0.0006,
+ "loss": 3.9735546112060547,
+ "step": 2488
+ },
+ {
+ "epoch": 34.57292576419214,
+ "grad_norm": 0.014593546278774738,
+ "learning_rate": 0.0006,
+ "loss": 4.041980266571045,
+ "step": 2489
+ },
+ {
+ "epoch": 34.58689956331878,
+ "grad_norm": 0.015436965972185135,
+ "learning_rate": 0.0006,
+ "loss": 4.301148414611816,
+ "step": 2490
+ },
+ {
+ "epoch": 34.60087336244541,
+ "grad_norm": 0.015075269155204296,
+ "learning_rate": 0.0006,
+ "loss": 4.118527412414551,
+ "step": 2491
+ },
+ {
+ "epoch": 34.61484716157205,
+ "grad_norm": 0.01701514609158039,
+ "learning_rate": 0.0006,
+ "loss": 4.245968818664551,
+ "step": 2492
+ },
+ {
+ "epoch": 34.62882096069869,
+ "grad_norm": 0.018047887831926346,
+ "learning_rate": 0.0006,
+ "loss": 4.227261066436768,
+ "step": 2493
+ },
+ {
+ "epoch": 34.64279475982533,
+ "grad_norm": 0.018499085679650307,
+ "learning_rate": 0.0006,
+ "loss": 4.121360778808594,
+ "step": 2494
+ },
+ {
+ "epoch": 34.65676855895197,
+ "grad_norm": 0.018987711519002914,
+ "learning_rate": 0.0006,
+ "loss": 4.131564617156982,
+ "step": 2495
+ },
+ {
+ "epoch": 34.670742358078606,
+ "grad_norm": 0.01834476925432682,
+ "learning_rate": 0.0006,
+ "loss": 4.224941253662109,
+ "step": 2496
+ },
+ {
+ "epoch": 34.68471615720524,
+ "grad_norm": 0.017507942393422127,
+ "learning_rate": 0.0006,
+ "loss": 4.1394548416137695,
+ "step": 2497
+ },
+ {
+ "epoch": 34.698689956331876,
+ "grad_norm": 0.01680580899119377,
+ "learning_rate": 0.0006,
+ "loss": 4.1073431968688965,
+ "step": 2498
+ },
+ {
+ "epoch": 34.712663755458514,
+ "grad_norm": 0.015233514830470085,
+ "learning_rate": 0.0006,
+ "loss": 4.136994361877441,
+ "step": 2499
+ },
+ {
+ "epoch": 34.72663755458515,
+ "grad_norm": 0.015482917428016663,
+ "learning_rate": 0.0006,
+ "loss": 4.110719680786133,
+ "step": 2500
+ },
+ {
+ "epoch": 34.74061135371179,
+ "grad_norm": 0.015940461307764053,
+ "learning_rate": 0.0006,
+ "loss": 4.207851886749268,
+ "step": 2501
+ },
+ {
+ "epoch": 34.75458515283843,
+ "grad_norm": 0.015355597250163555,
+ "learning_rate": 0.0006,
+ "loss": 4.169469833374023,
+ "step": 2502
+ },
+ {
+ "epoch": 34.76855895196506,
+ "grad_norm": 0.015974000096321106,
+ "learning_rate": 0.0006,
+ "loss": 4.110061168670654,
+ "step": 2503
+ },
+ {
+ "epoch": 34.7825327510917,
+ "grad_norm": 0.016088927164673805,
+ "learning_rate": 0.0006,
+ "loss": 4.155452728271484,
+ "step": 2504
+ },
+ {
+ "epoch": 34.79650655021834,
+ "grad_norm": 0.016156120225787163,
+ "learning_rate": 0.0006,
+ "loss": 4.178092956542969,
+ "step": 2505
+ },
+ {
+ "epoch": 34.81048034934498,
+ "grad_norm": 0.014928647316992283,
+ "learning_rate": 0.0006,
+ "loss": 4.05457878112793,
+ "step": 2506
+ },
+ {
+ "epoch": 34.82445414847162,
+ "grad_norm": 0.015289328061044216,
+ "learning_rate": 0.0006,
+ "loss": 4.18869686126709,
+ "step": 2507
+ },
+ {
+ "epoch": 34.838427947598255,
+ "grad_norm": 0.015799039974808693,
+ "learning_rate": 0.0006,
+ "loss": 4.106073379516602,
+ "step": 2508
+ },
+ {
+ "epoch": 34.852401746724894,
+ "grad_norm": 0.015824418514966965,
+ "learning_rate": 0.0006,
+ "loss": 4.0634589195251465,
+ "step": 2509
+ },
+ {
+ "epoch": 34.866375545851525,
+ "grad_norm": 0.015321656130254269,
+ "learning_rate": 0.0006,
+ "loss": 4.093090057373047,
+ "step": 2510
+ },
+ {
+ "epoch": 34.880349344978164,
+ "grad_norm": 0.014380166307091713,
+ "learning_rate": 0.0006,
+ "loss": 4.1750640869140625,
+ "step": 2511
+ },
+ {
+ "epoch": 34.8943231441048,
+ "grad_norm": 0.014962567947804928,
+ "learning_rate": 0.0006,
+ "loss": 4.185277938842773,
+ "step": 2512
+ },
+ {
+ "epoch": 34.90829694323144,
+ "grad_norm": 0.014467543922364712,
+ "learning_rate": 0.0006,
+ "loss": 4.1600799560546875,
+ "step": 2513
+ },
+ {
+ "epoch": 34.92227074235808,
+ "grad_norm": 0.01545181218534708,
+ "learning_rate": 0.0006,
+ "loss": 4.096299648284912,
+ "step": 2514
+ },
+ {
+ "epoch": 34.93624454148472,
+ "grad_norm": 0.01605340465903282,
+ "learning_rate": 0.0006,
+ "loss": 4.203323841094971,
+ "step": 2515
+ },
+ {
+ "epoch": 34.95021834061135,
+ "grad_norm": 0.015203815884888172,
+ "learning_rate": 0.0006,
+ "loss": 4.201385021209717,
+ "step": 2516
+ },
+ {
+ "epoch": 34.96419213973799,
+ "grad_norm": 0.015149409882724285,
+ "learning_rate": 0.0006,
+ "loss": 4.088164329528809,
+ "step": 2517
+ },
+ {
+ "epoch": 34.97816593886463,
+ "grad_norm": 0.014885558746755123,
+ "learning_rate": 0.0006,
+ "loss": 4.073911666870117,
+ "step": 2518
+ },
+ {
+ "epoch": 34.992139737991266,
+ "grad_norm": 0.015003419481217861,
+ "learning_rate": 0.0006,
+ "loss": 4.157113552093506,
+ "step": 2519
+ },
+ {
+ "epoch": 35.0,
+ "grad_norm": 0.01641916297376156,
+ "learning_rate": 0.0006,
+ "loss": 3.970052719116211,
+ "step": 2520
+ },
+ {
+ "epoch": 35.0,
+ "eval_loss": 4.641692161560059,
+ "eval_runtime": 56.8836,
+ "eval_samples_per_second": 42.93,
+ "eval_steps_per_second": 1.354,
+ "step": 2520
+ },
+ {
+ "epoch": 35.01397379912664,
+ "grad_norm": 0.015894673764705658,
+ "learning_rate": 0.0006,
+ "loss": 4.104437351226807,
+ "step": 2521
+ },
+ {
+ "epoch": 35.02794759825328,
+ "grad_norm": 0.0173867866396904,
+ "learning_rate": 0.0006,
+ "loss": 4.091726303100586,
+ "step": 2522
+ },
+ {
+ "epoch": 35.041921397379916,
+ "grad_norm": 0.019272832199931145,
+ "learning_rate": 0.0006,
+ "loss": 4.217905044555664,
+ "step": 2523
+ },
+ {
+ "epoch": 35.05589519650655,
+ "grad_norm": 0.021820371970534325,
+ "learning_rate": 0.0006,
+ "loss": 4.111428260803223,
+ "step": 2524
+ },
+ {
+ "epoch": 35.069868995633186,
+ "grad_norm": 0.02340741828083992,
+ "learning_rate": 0.0006,
+ "loss": 4.115771293640137,
+ "step": 2525
+ },
+ {
+ "epoch": 35.083842794759825,
+ "grad_norm": 0.023760691285133362,
+ "learning_rate": 0.0006,
+ "loss": 4.132383346557617,
+ "step": 2526
+ },
+ {
+ "epoch": 35.09781659388646,
+ "grad_norm": 0.02073468267917633,
+ "learning_rate": 0.0006,
+ "loss": 4.272691249847412,
+ "step": 2527
+ },
+ {
+ "epoch": 35.1117903930131,
+ "grad_norm": 0.02197718247771263,
+ "learning_rate": 0.0006,
+ "loss": 4.105586528778076,
+ "step": 2528
+ },
+ {
+ "epoch": 35.12576419213974,
+ "grad_norm": 0.0200254675000906,
+ "learning_rate": 0.0006,
+ "loss": 3.968820095062256,
+ "step": 2529
+ },
+ {
+ "epoch": 35.13973799126637,
+ "grad_norm": 0.01929587684571743,
+ "learning_rate": 0.0006,
+ "loss": 4.045468807220459,
+ "step": 2530
+ },
+ {
+ "epoch": 35.15371179039301,
+ "grad_norm": 0.01886162906885147,
+ "learning_rate": 0.0006,
+ "loss": 4.185571670532227,
+ "step": 2531
+ },
+ {
+ "epoch": 35.16768558951965,
+ "grad_norm": 0.019789839163422585,
+ "learning_rate": 0.0006,
+ "loss": 3.9734551906585693,
+ "step": 2532
+ },
+ {
+ "epoch": 35.18165938864629,
+ "grad_norm": 0.0198142658919096,
+ "learning_rate": 0.0006,
+ "loss": 4.2662506103515625,
+ "step": 2533
+ },
+ {
+ "epoch": 35.19563318777293,
+ "grad_norm": 0.02087012305855751,
+ "learning_rate": 0.0006,
+ "loss": 4.175931930541992,
+ "step": 2534
+ },
+ {
+ "epoch": 35.209606986899566,
+ "grad_norm": 0.020294535905122757,
+ "learning_rate": 0.0006,
+ "loss": 4.060324668884277,
+ "step": 2535
+ },
+ {
+ "epoch": 35.223580786026204,
+ "grad_norm": 0.01908983662724495,
+ "learning_rate": 0.0006,
+ "loss": 4.217939853668213,
+ "step": 2536
+ },
+ {
+ "epoch": 35.237554585152836,
+ "grad_norm": 0.0178892333060503,
+ "learning_rate": 0.0006,
+ "loss": 4.190216541290283,
+ "step": 2537
+ },
+ {
+ "epoch": 35.251528384279474,
+ "grad_norm": 0.017892315983772278,
+ "learning_rate": 0.0006,
+ "loss": 4.067419052124023,
+ "step": 2538
+ },
+ {
+ "epoch": 35.26550218340611,
+ "grad_norm": 0.018556250259280205,
+ "learning_rate": 0.0006,
+ "loss": 4.105958461761475,
+ "step": 2539
+ },
+ {
+ "epoch": 35.27947598253275,
+ "grad_norm": 0.01944422349333763,
+ "learning_rate": 0.0006,
+ "loss": 4.155775547027588,
+ "step": 2540
+ },
+ {
+ "epoch": 35.29344978165939,
+ "grad_norm": 0.017333146184682846,
+ "learning_rate": 0.0006,
+ "loss": 4.244513511657715,
+ "step": 2541
+ },
+ {
+ "epoch": 35.30742358078603,
+ "grad_norm": 0.01639244332909584,
+ "learning_rate": 0.0006,
+ "loss": 4.167815685272217,
+ "step": 2542
+ },
+ {
+ "epoch": 35.32139737991266,
+ "grad_norm": 0.01610143668949604,
+ "learning_rate": 0.0006,
+ "loss": 4.121486663818359,
+ "step": 2543
+ },
+ {
+ "epoch": 35.3353711790393,
+ "grad_norm": 0.01656411588191986,
+ "learning_rate": 0.0006,
+ "loss": 4.1755499839782715,
+ "step": 2544
+ },
+ {
+ "epoch": 35.34934497816594,
+ "grad_norm": 0.015075297094881535,
+ "learning_rate": 0.0006,
+ "loss": 4.102992534637451,
+ "step": 2545
+ },
+ {
+ "epoch": 35.36331877729258,
+ "grad_norm": 0.01538170501589775,
+ "learning_rate": 0.0006,
+ "loss": 4.161603927612305,
+ "step": 2546
+ },
+ {
+ "epoch": 35.377292576419215,
+ "grad_norm": 0.01483312901109457,
+ "learning_rate": 0.0006,
+ "loss": 4.085355758666992,
+ "step": 2547
+ },
+ {
+ "epoch": 35.391266375545854,
+ "grad_norm": 0.014894738793373108,
+ "learning_rate": 0.0006,
+ "loss": 4.083585262298584,
+ "step": 2548
+ },
+ {
+ "epoch": 35.40524017467249,
+ "grad_norm": 0.01475780550390482,
+ "learning_rate": 0.0006,
+ "loss": 4.012645721435547,
+ "step": 2549
+ },
+ {
+ "epoch": 35.419213973799124,
+ "grad_norm": 0.014531653374433517,
+ "learning_rate": 0.0006,
+ "loss": 4.197113990783691,
+ "step": 2550
+ },
+ {
+ "epoch": 35.43318777292576,
+ "grad_norm": 0.015316493809223175,
+ "learning_rate": 0.0006,
+ "loss": 4.227970123291016,
+ "step": 2551
+ },
+ {
+ "epoch": 35.4471615720524,
+ "grad_norm": 0.014685571193695068,
+ "learning_rate": 0.0006,
+ "loss": 4.089415550231934,
+ "step": 2552
+ },
+ {
+ "epoch": 35.46113537117904,
+ "grad_norm": 0.015337035991251469,
+ "learning_rate": 0.0006,
+ "loss": 4.080583572387695,
+ "step": 2553
+ },
+ {
+ "epoch": 35.47510917030568,
+ "grad_norm": 0.016754208132624626,
+ "learning_rate": 0.0006,
+ "loss": 4.163935661315918,
+ "step": 2554
+ },
+ {
+ "epoch": 35.48908296943232,
+ "grad_norm": 0.017535798251628876,
+ "learning_rate": 0.0006,
+ "loss": 4.093027591705322,
+ "step": 2555
+ },
+ {
+ "epoch": 35.50305676855895,
+ "grad_norm": 0.017336975783109665,
+ "learning_rate": 0.0006,
+ "loss": 4.127961158752441,
+ "step": 2556
+ },
+ {
+ "epoch": 35.51703056768559,
+ "grad_norm": 0.016242152079939842,
+ "learning_rate": 0.0006,
+ "loss": 4.155593395233154,
+ "step": 2557
+ },
+ {
+ "epoch": 35.531004366812226,
+ "grad_norm": 0.015121322125196457,
+ "learning_rate": 0.0006,
+ "loss": 3.984004020690918,
+ "step": 2558
+ },
+ {
+ "epoch": 35.544978165938865,
+ "grad_norm": 0.015178884379565716,
+ "learning_rate": 0.0006,
+ "loss": 4.111564636230469,
+ "step": 2559
+ },
+ {
+ "epoch": 35.5589519650655,
+ "grad_norm": 0.014777131378650665,
+ "learning_rate": 0.0006,
+ "loss": 4.041402816772461,
+ "step": 2560
+ },
+ {
+ "epoch": 35.57292576419214,
+ "grad_norm": 0.015197676606476307,
+ "learning_rate": 0.0006,
+ "loss": 4.133706092834473,
+ "step": 2561
+ },
+ {
+ "epoch": 35.58689956331878,
+ "grad_norm": 0.015905972570180893,
+ "learning_rate": 0.0006,
+ "loss": 4.1069135665893555,
+ "step": 2562
+ },
+ {
+ "epoch": 35.60087336244541,
+ "grad_norm": 0.015942877158522606,
+ "learning_rate": 0.0006,
+ "loss": 4.0720624923706055,
+ "step": 2563
+ },
+ {
+ "epoch": 35.61484716157205,
+ "grad_norm": 0.016697583720088005,
+ "learning_rate": 0.0006,
+ "loss": 4.167448043823242,
+ "step": 2564
+ },
+ {
+ "epoch": 35.62882096069869,
+ "grad_norm": 0.015603577718138695,
+ "learning_rate": 0.0006,
+ "loss": 4.127327919006348,
+ "step": 2565
+ },
+ {
+ "epoch": 35.64279475982533,
+ "grad_norm": 0.017153460532426834,
+ "learning_rate": 0.0006,
+ "loss": 4.056543350219727,
+ "step": 2566
+ },
+ {
+ "epoch": 35.65676855895197,
+ "grad_norm": 0.016731958836317062,
+ "learning_rate": 0.0006,
+ "loss": 4.0681843757629395,
+ "step": 2567
+ },
+ {
+ "epoch": 35.670742358078606,
+ "grad_norm": 0.016167359426617622,
+ "learning_rate": 0.0006,
+ "loss": 4.174992084503174,
+ "step": 2568
+ },
+ {
+ "epoch": 35.68471615720524,
+ "grad_norm": 0.01615108922123909,
+ "learning_rate": 0.0006,
+ "loss": 4.1383161544799805,
+ "step": 2569
+ },
+ {
+ "epoch": 35.698689956331876,
+ "grad_norm": 0.015750816091895103,
+ "learning_rate": 0.0006,
+ "loss": 4.031595706939697,
+ "step": 2570
+ },
+ {
+ "epoch": 35.712663755458514,
+ "grad_norm": 0.015416189096868038,
+ "learning_rate": 0.0006,
+ "loss": 4.067498207092285,
+ "step": 2571
+ },
+ {
+ "epoch": 35.72663755458515,
+ "grad_norm": 0.01570792682468891,
+ "learning_rate": 0.0006,
+ "loss": 4.1338300704956055,
+ "step": 2572
+ },
+ {
+ "epoch": 35.74061135371179,
+ "grad_norm": 0.016367752104997635,
+ "learning_rate": 0.0006,
+ "loss": 4.152523040771484,
+ "step": 2573
+ },
+ {
+ "epoch": 35.75458515283843,
+ "grad_norm": 0.016349803656339645,
+ "learning_rate": 0.0006,
+ "loss": 4.101985454559326,
+ "step": 2574
+ },
+ {
+ "epoch": 35.76855895196506,
+ "grad_norm": 0.01663660630583763,
+ "learning_rate": 0.0006,
+ "loss": 4.190130710601807,
+ "step": 2575
+ },
+ {
+ "epoch": 35.7825327510917,
+ "grad_norm": 0.015599170699715614,
+ "learning_rate": 0.0006,
+ "loss": 4.0792436599731445,
+ "step": 2576
+ },
+ {
+ "epoch": 35.79650655021834,
+ "grad_norm": 0.015472427010536194,
+ "learning_rate": 0.0006,
+ "loss": 4.186446666717529,
+ "step": 2577
+ },
+ {
+ "epoch": 35.81048034934498,
+ "grad_norm": 0.015520439483225346,
+ "learning_rate": 0.0006,
+ "loss": 4.246635437011719,
+ "step": 2578
+ },
+ {
+ "epoch": 35.82445414847162,
+ "grad_norm": 0.016998281702399254,
+ "learning_rate": 0.0006,
+ "loss": 4.13435173034668,
+ "step": 2579
+ },
+ {
+ "epoch": 35.838427947598255,
+ "grad_norm": 0.016290361061692238,
+ "learning_rate": 0.0006,
+ "loss": 4.107428073883057,
+ "step": 2580
+ },
+ {
+ "epoch": 35.852401746724894,
+ "grad_norm": 0.0165516659617424,
+ "learning_rate": 0.0006,
+ "loss": 4.1612138748168945,
+ "step": 2581
+ },
+ {
+ "epoch": 35.866375545851525,
+ "grad_norm": 0.01736091636121273,
+ "learning_rate": 0.0006,
+ "loss": 4.1072869300842285,
+ "step": 2582
+ },
+ {
+ "epoch": 35.880349344978164,
+ "grad_norm": 0.01787031814455986,
+ "learning_rate": 0.0006,
+ "loss": 4.1076812744140625,
+ "step": 2583
+ },
+ {
+ "epoch": 35.8943231441048,
+ "grad_norm": 0.015437985770404339,
+ "learning_rate": 0.0006,
+ "loss": 4.097488880157471,
+ "step": 2584
+ },
+ {
+ "epoch": 35.90829694323144,
+ "grad_norm": 0.016306741163134575,
+ "learning_rate": 0.0006,
+ "loss": 4.105017185211182,
+ "step": 2585
+ },
+ {
+ "epoch": 35.92227074235808,
+ "grad_norm": 0.015723245218396187,
+ "learning_rate": 0.0006,
+ "loss": 4.205283164978027,
+ "step": 2586
+ },
+ {
+ "epoch": 35.93624454148472,
+ "grad_norm": 0.014626456424593925,
+ "learning_rate": 0.0006,
+ "loss": 4.086552619934082,
+ "step": 2587
+ },
+ {
+ "epoch": 35.95021834061135,
+ "grad_norm": 0.014348454773426056,
+ "learning_rate": 0.0006,
+ "loss": 4.171329975128174,
+ "step": 2588
+ },
+ {
+ "epoch": 35.96419213973799,
+ "grad_norm": 0.015598480589687824,
+ "learning_rate": 0.0006,
+ "loss": 4.140815258026123,
+ "step": 2589
+ },
+ {
+ "epoch": 35.97816593886463,
+ "grad_norm": 0.017126062884926796,
+ "learning_rate": 0.0006,
+ "loss": 4.157176971435547,
+ "step": 2590
+ },
+ {
+ "epoch": 35.992139737991266,
+ "grad_norm": 0.01860199309885502,
+ "learning_rate": 0.0006,
+ "loss": 4.252880096435547,
+ "step": 2591
+ },
+ {
+ "epoch": 36.0,
+ "grad_norm": 0.021780220791697502,
+ "learning_rate": 0.0006,
+ "loss": 4.249861717224121,
+ "step": 2592
+ },
+ {
+ "epoch": 36.0,
+ "eval_loss": 4.606466293334961,
+ "eval_runtime": 56.6217,
+ "eval_samples_per_second": 43.128,
+ "eval_steps_per_second": 1.36,
+ "step": 2592
+ },
+ {
+ "epoch": 36.01397379912664,
+ "grad_norm": 0.020892828702926636,
+ "learning_rate": 0.0006,
+ "loss": 4.102055549621582,
+ "step": 2593
+ },
+ {
+ "epoch": 36.02794759825328,
+ "grad_norm": 0.020306255668401718,
+ "learning_rate": 0.0006,
+ "loss": 4.184308052062988,
+ "step": 2594
+ },
+ {
+ "epoch": 36.041921397379916,
+ "grad_norm": 0.02025097794830799,
+ "learning_rate": 0.0006,
+ "loss": 3.9691858291625977,
+ "step": 2595
+ },
+ {
+ "epoch": 36.05589519650655,
+ "grad_norm": 0.020254382863640785,
+ "learning_rate": 0.0006,
+ "loss": 4.02957820892334,
+ "step": 2596
+ },
+ {
+ "epoch": 36.069868995633186,
+ "grad_norm": 0.019307196140289307,
+ "learning_rate": 0.0006,
+ "loss": 4.033181667327881,
+ "step": 2597
+ },
+ {
+ "epoch": 36.083842794759825,
+ "grad_norm": 0.0198476854711771,
+ "learning_rate": 0.0006,
+ "loss": 4.0816874504089355,
+ "step": 2598
+ },
+ {
+ "epoch": 36.09781659388646,
+ "grad_norm": 0.019916515797376633,
+ "learning_rate": 0.0006,
+ "loss": 4.164889812469482,
+ "step": 2599
+ },
+ {
+ "epoch": 36.1117903930131,
+ "grad_norm": 0.0189144778996706,
+ "learning_rate": 0.0006,
+ "loss": 4.038427352905273,
+ "step": 2600
+ },
+ {
+ "epoch": 36.12576419213974,
+ "grad_norm": 0.01754768192768097,
+ "learning_rate": 0.0006,
+ "loss": 4.010427474975586,
+ "step": 2601
+ },
+ {
+ "epoch": 36.13973799126637,
+ "grad_norm": 0.016862990334630013,
+ "learning_rate": 0.0006,
+ "loss": 4.035970211029053,
+ "step": 2602
+ },
+ {
+ "epoch": 36.15371179039301,
+ "grad_norm": 0.019616752862930298,
+ "learning_rate": 0.0006,
+ "loss": 4.079898834228516,
+ "step": 2603
+ },
+ {
+ "epoch": 36.16768558951965,
+ "grad_norm": 0.020131465047597885,
+ "learning_rate": 0.0006,
+ "loss": 4.144684791564941,
+ "step": 2604
+ },
+ {
+ "epoch": 36.18165938864629,
+ "grad_norm": 0.02072521299123764,
+ "learning_rate": 0.0006,
+ "loss": 4.013660430908203,
+ "step": 2605
+ },
+ {
+ "epoch": 36.19563318777293,
+ "grad_norm": 0.01924615167081356,
+ "learning_rate": 0.0006,
+ "loss": 4.071398735046387,
+ "step": 2606
+ },
+ {
+ "epoch": 36.209606986899566,
+ "grad_norm": 0.02013954520225525,
+ "learning_rate": 0.0006,
+ "loss": 4.127094268798828,
+ "step": 2607
+ },
+ {
+ "epoch": 36.223580786026204,
+ "grad_norm": 0.02042008936405182,
+ "learning_rate": 0.0006,
+ "loss": 4.154261589050293,
+ "step": 2608
+ },
+ {
+ "epoch": 36.237554585152836,
+ "grad_norm": 0.021998876705765724,
+ "learning_rate": 0.0006,
+ "loss": 4.204150199890137,
+ "step": 2609
+ },
+ {
+ "epoch": 36.251528384279474,
+ "grad_norm": 0.022946683689951897,
+ "learning_rate": 0.0006,
+ "loss": 4.013632774353027,
+ "step": 2610
+ },
+ {
+ "epoch": 36.26550218340611,
+ "grad_norm": 0.02061830274760723,
+ "learning_rate": 0.0006,
+ "loss": 4.171084403991699,
+ "step": 2611
+ },
+ {
+ "epoch": 36.27947598253275,
+ "grad_norm": 0.01883152686059475,
+ "learning_rate": 0.0006,
+ "loss": 4.149848461151123,
+ "step": 2612
+ },
+ {
+ "epoch": 36.29344978165939,
+ "grad_norm": 0.019945282489061356,
+ "learning_rate": 0.0006,
+ "loss": 4.174656867980957,
+ "step": 2613
+ },
+ {
+ "epoch": 36.30742358078603,
+ "grad_norm": 0.020751051604747772,
+ "learning_rate": 0.0006,
+ "loss": 4.080747127532959,
+ "step": 2614
+ },
+ {
+ "epoch": 36.32139737991266,
+ "grad_norm": 0.01879691146314144,
+ "learning_rate": 0.0006,
+ "loss": 4.057356357574463,
+ "step": 2615
+ },
+ {
+ "epoch": 36.3353711790393,
+ "grad_norm": 0.019037071615457535,
+ "learning_rate": 0.0006,
+ "loss": 4.204607009887695,
+ "step": 2616
+ },
+ {
+ "epoch": 36.34934497816594,
+ "grad_norm": 0.01872078701853752,
+ "learning_rate": 0.0006,
+ "loss": 4.050586700439453,
+ "step": 2617
+ },
+ {
+ "epoch": 36.36331877729258,
+ "grad_norm": 0.018092811107635498,
+ "learning_rate": 0.0006,
+ "loss": 4.102832794189453,
+ "step": 2618
+ },
+ {
+ "epoch": 36.377292576419215,
+ "grad_norm": 0.017010487616062164,
+ "learning_rate": 0.0006,
+ "loss": 4.191788673400879,
+ "step": 2619
+ },
+ {
+ "epoch": 36.391266375545854,
+ "grad_norm": 0.01691143773496151,
+ "learning_rate": 0.0006,
+ "loss": 4.160456657409668,
+ "step": 2620
+ },
+ {
+ "epoch": 36.40524017467249,
+ "grad_norm": 0.017046816647052765,
+ "learning_rate": 0.0006,
+ "loss": 4.073828220367432,
+ "step": 2621
+ },
+ {
+ "epoch": 36.419213973799124,
+ "grad_norm": 0.01707981713116169,
+ "learning_rate": 0.0006,
+ "loss": 4.0980095863342285,
+ "step": 2622
+ },
+ {
+ "epoch": 36.43318777292576,
+ "grad_norm": 0.01630452275276184,
+ "learning_rate": 0.0006,
+ "loss": 4.101321220397949,
+ "step": 2623
+ },
+ {
+ "epoch": 36.4471615720524,
+ "grad_norm": 0.01576915942132473,
+ "learning_rate": 0.0006,
+ "loss": 4.158249855041504,
+ "step": 2624
+ },
+ {
+ "epoch": 36.46113537117904,
+ "grad_norm": 0.016179712489247322,
+ "learning_rate": 0.0006,
+ "loss": 4.077852249145508,
+ "step": 2625
+ },
+ {
+ "epoch": 36.47510917030568,
+ "grad_norm": 0.01609169878065586,
+ "learning_rate": 0.0006,
+ "loss": 4.187317848205566,
+ "step": 2626
+ },
+ {
+ "epoch": 36.48908296943232,
+ "grad_norm": 0.017148254439234734,
+ "learning_rate": 0.0006,
+ "loss": 4.109071731567383,
+ "step": 2627
+ },
+ {
+ "epoch": 36.50305676855895,
+ "grad_norm": 0.015842624008655548,
+ "learning_rate": 0.0006,
+ "loss": 4.145096302032471,
+ "step": 2628
+ },
+ {
+ "epoch": 36.51703056768559,
+ "grad_norm": 0.014906194061040878,
+ "learning_rate": 0.0006,
+ "loss": 4.028440475463867,
+ "step": 2629
+ },
+ {
+ "epoch": 36.531004366812226,
+ "grad_norm": 0.01485038734972477,
+ "learning_rate": 0.0006,
+ "loss": 4.175201416015625,
+ "step": 2630
+ },
+ {
+ "epoch": 36.544978165938865,
+ "grad_norm": 0.015834150835871696,
+ "learning_rate": 0.0006,
+ "loss": 4.076434135437012,
+ "step": 2631
+ },
+ {
+ "epoch": 36.5589519650655,
+ "grad_norm": 0.015411358326673508,
+ "learning_rate": 0.0006,
+ "loss": 4.03912878036499,
+ "step": 2632
+ },
+ {
+ "epoch": 36.57292576419214,
+ "grad_norm": 0.014762846753001213,
+ "learning_rate": 0.0006,
+ "loss": 4.019827365875244,
+ "step": 2633
+ },
+ {
+ "epoch": 36.58689956331878,
+ "grad_norm": 0.015269497409462929,
+ "learning_rate": 0.0006,
+ "loss": 4.116957187652588,
+ "step": 2634
+ },
+ {
+ "epoch": 36.60087336244541,
+ "grad_norm": 0.015171617269515991,
+ "learning_rate": 0.0006,
+ "loss": 4.2069926261901855,
+ "step": 2635
+ },
+ {
+ "epoch": 36.61484716157205,
+ "grad_norm": 0.014770902693271637,
+ "learning_rate": 0.0006,
+ "loss": 4.092021942138672,
+ "step": 2636
+ },
+ {
+ "epoch": 36.62882096069869,
+ "grad_norm": 0.014957531355321407,
+ "learning_rate": 0.0006,
+ "loss": 4.16819953918457,
+ "step": 2637
+ },
+ {
+ "epoch": 36.64279475982533,
+ "grad_norm": 0.014774451963603497,
+ "learning_rate": 0.0006,
+ "loss": 4.046073913574219,
+ "step": 2638
+ },
+ {
+ "epoch": 36.65676855895197,
+ "grad_norm": 0.014320445246994495,
+ "learning_rate": 0.0006,
+ "loss": 3.977792501449585,
+ "step": 2639
+ },
+ {
+ "epoch": 36.670742358078606,
+ "grad_norm": 0.01413130946457386,
+ "learning_rate": 0.0006,
+ "loss": 4.130649566650391,
+ "step": 2640
+ },
+ {
+ "epoch": 36.68471615720524,
+ "grad_norm": 0.014336307533085346,
+ "learning_rate": 0.0006,
+ "loss": 4.161357402801514,
+ "step": 2641
+ },
+ {
+ "epoch": 36.698689956331876,
+ "grad_norm": 0.014516931027173996,
+ "learning_rate": 0.0006,
+ "loss": 4.14248514175415,
+ "step": 2642
+ },
+ {
+ "epoch": 36.712663755458514,
+ "grad_norm": 0.014377126470208168,
+ "learning_rate": 0.0006,
+ "loss": 4.062711715698242,
+ "step": 2643
+ },
+ {
+ "epoch": 36.72663755458515,
+ "grad_norm": 0.014332668855786324,
+ "learning_rate": 0.0006,
+ "loss": 4.06423807144165,
+ "step": 2644
+ },
+ {
+ "epoch": 36.74061135371179,
+ "grad_norm": 0.014030571095645428,
+ "learning_rate": 0.0006,
+ "loss": 4.178613662719727,
+ "step": 2645
+ },
+ {
+ "epoch": 36.75458515283843,
+ "grad_norm": 0.014376776292920113,
+ "learning_rate": 0.0006,
+ "loss": 4.073130130767822,
+ "step": 2646
+ },
+ {
+ "epoch": 36.76855895196506,
+ "grad_norm": 0.013827716931700706,
+ "learning_rate": 0.0006,
+ "loss": 4.043869972229004,
+ "step": 2647
+ },
+ {
+ "epoch": 36.7825327510917,
+ "grad_norm": 0.015548762865364552,
+ "learning_rate": 0.0006,
+ "loss": 4.134049415588379,
+ "step": 2648
+ },
+ {
+ "epoch": 36.79650655021834,
+ "grad_norm": 0.017041970044374466,
+ "learning_rate": 0.0006,
+ "loss": 4.205024242401123,
+ "step": 2649
+ },
+ {
+ "epoch": 36.81048034934498,
+ "grad_norm": 0.017638275399804115,
+ "learning_rate": 0.0006,
+ "loss": 4.199924468994141,
+ "step": 2650
+ },
+ {
+ "epoch": 36.82445414847162,
+ "grad_norm": 0.017771627753973007,
+ "learning_rate": 0.0006,
+ "loss": 4.15706729888916,
+ "step": 2651
+ },
+ {
+ "epoch": 36.838427947598255,
+ "grad_norm": 0.015933562070131302,
+ "learning_rate": 0.0006,
+ "loss": 4.220742225646973,
+ "step": 2652
+ },
+ {
+ "epoch": 36.852401746724894,
+ "grad_norm": 0.01668144389986992,
+ "learning_rate": 0.0006,
+ "loss": 4.161426544189453,
+ "step": 2653
+ },
+ {
+ "epoch": 36.866375545851525,
+ "grad_norm": 0.018014680594205856,
+ "learning_rate": 0.0006,
+ "loss": 4.183046817779541,
+ "step": 2654
+ },
+ {
+ "epoch": 36.880349344978164,
+ "grad_norm": 0.018808670341968536,
+ "learning_rate": 0.0006,
+ "loss": 4.105293273925781,
+ "step": 2655
+ },
+ {
+ "epoch": 36.8943231441048,
+ "grad_norm": 0.018003476783633232,
+ "learning_rate": 0.0006,
+ "loss": 4.103171348571777,
+ "step": 2656
+ },
+ {
+ "epoch": 36.90829694323144,
+ "grad_norm": 0.016865914687514305,
+ "learning_rate": 0.0006,
+ "loss": 4.125634670257568,
+ "step": 2657
+ },
+ {
+ "epoch": 36.92227074235808,
+ "grad_norm": 0.018371153622865677,
+ "learning_rate": 0.0006,
+ "loss": 4.192461013793945,
+ "step": 2658
+ },
+ {
+ "epoch": 36.93624454148472,
+ "grad_norm": 0.019045377150177956,
+ "learning_rate": 0.0006,
+ "loss": 4.198105812072754,
+ "step": 2659
+ },
+ {
+ "epoch": 36.95021834061135,
+ "grad_norm": 0.01594168320298195,
+ "learning_rate": 0.0006,
+ "loss": 4.169788360595703,
+ "step": 2660
+ },
+ {
+ "epoch": 36.96419213973799,
+ "grad_norm": 0.015645820647478104,
+ "learning_rate": 0.0006,
+ "loss": 4.047277450561523,
+ "step": 2661
+ },
+ {
+ "epoch": 36.97816593886463,
+ "grad_norm": 0.016313156113028526,
+ "learning_rate": 0.0006,
+ "loss": 4.087392807006836,
+ "step": 2662
+ },
+ {
+ "epoch": 36.992139737991266,
+ "grad_norm": 0.017050625756382942,
+ "learning_rate": 0.0006,
+ "loss": 4.0908002853393555,
+ "step": 2663
+ },
+ {
+ "epoch": 37.0,
+ "grad_norm": 0.017730968073010445,
+ "learning_rate": 0.0006,
+ "loss": 4.110957622528076,
+ "step": 2664
+ },
+ {
+ "epoch": 37.0,
+ "eval_loss": 4.563793182373047,
+ "eval_runtime": 56.5922,
+ "eval_samples_per_second": 43.151,
+ "eval_steps_per_second": 1.361,
+ "step": 2664
+ },
+ {
+ "epoch": 37.01397379912664,
+ "grad_norm": 0.015687910839915276,
+ "learning_rate": 0.0006,
+ "loss": 4.143589019775391,
+ "step": 2665
+ },
+ {
+ "epoch": 37.02794759825328,
+ "grad_norm": 0.017451366409659386,
+ "learning_rate": 0.0006,
+ "loss": 4.0326666831970215,
+ "step": 2666
+ },
+ {
+ "epoch": 37.041921397379916,
+ "grad_norm": 0.01909538544714451,
+ "learning_rate": 0.0006,
+ "loss": 3.9706249237060547,
+ "step": 2667
+ },
+ {
+ "epoch": 37.05589519650655,
+ "grad_norm": 0.01988379657268524,
+ "learning_rate": 0.0006,
+ "loss": 3.9531009197235107,
+ "step": 2668
+ },
+ {
+ "epoch": 37.069868995633186,
+ "grad_norm": 0.02035115472972393,
+ "learning_rate": 0.0006,
+ "loss": 3.962214469909668,
+ "step": 2669
+ },
+ {
+ "epoch": 37.083842794759825,
+ "grad_norm": 0.01952001079916954,
+ "learning_rate": 0.0006,
+ "loss": 4.182262897491455,
+ "step": 2670
+ },
+ {
+ "epoch": 37.09781659388646,
+ "grad_norm": 0.01854325644671917,
+ "learning_rate": 0.0006,
+ "loss": 4.127663612365723,
+ "step": 2671
+ },
+ {
+ "epoch": 37.1117903930131,
+ "grad_norm": 0.017757616937160492,
+ "learning_rate": 0.0006,
+ "loss": 4.100804805755615,
+ "step": 2672
+ },
+ {
+ "epoch": 37.12576419213974,
+ "grad_norm": 0.016712479293346405,
+ "learning_rate": 0.0006,
+ "loss": 4.026284217834473,
+ "step": 2673
+ },
+ {
+ "epoch": 37.13973799126637,
+ "grad_norm": 0.0168139711022377,
+ "learning_rate": 0.0006,
+ "loss": 4.183882713317871,
+ "step": 2674
+ },
+ {
+ "epoch": 37.15371179039301,
+ "grad_norm": 0.017060423269867897,
+ "learning_rate": 0.0006,
+ "loss": 4.131832599639893,
+ "step": 2675
+ },
+ {
+ "epoch": 37.16768558951965,
+ "grad_norm": 0.01726631447672844,
+ "learning_rate": 0.0006,
+ "loss": 3.9156057834625244,
+ "step": 2676
+ },
+ {
+ "epoch": 37.18165938864629,
+ "grad_norm": 0.016348622739315033,
+ "learning_rate": 0.0006,
+ "loss": 4.1731367111206055,
+ "step": 2677
+ },
+ {
+ "epoch": 37.19563318777293,
+ "grad_norm": 0.016640059649944305,
+ "learning_rate": 0.0006,
+ "loss": 4.10997200012207,
+ "step": 2678
+ },
+ {
+ "epoch": 37.209606986899566,
+ "grad_norm": 0.01601342298090458,
+ "learning_rate": 0.0006,
+ "loss": 4.081145286560059,
+ "step": 2679
+ },
+ {
+ "epoch": 37.223580786026204,
+ "grad_norm": 0.01735631190240383,
+ "learning_rate": 0.0006,
+ "loss": 4.202108860015869,
+ "step": 2680
+ },
+ {
+ "epoch": 37.237554585152836,
+ "grad_norm": 0.019164685159921646,
+ "learning_rate": 0.0006,
+ "loss": 4.104109764099121,
+ "step": 2681
+ },
+ {
+ "epoch": 37.251528384279474,
+ "grad_norm": 0.01881759613752365,
+ "learning_rate": 0.0006,
+ "loss": 4.145532131195068,
+ "step": 2682
+ },
+ {
+ "epoch": 37.26550218340611,
+ "grad_norm": 0.017656555399298668,
+ "learning_rate": 0.0006,
+ "loss": 4.012782096862793,
+ "step": 2683
+ },
+ {
+ "epoch": 37.27947598253275,
+ "grad_norm": 0.01624823734164238,
+ "learning_rate": 0.0006,
+ "loss": 4.001727104187012,
+ "step": 2684
+ },
+ {
+ "epoch": 37.29344978165939,
+ "grad_norm": 0.016973139718174934,
+ "learning_rate": 0.0006,
+ "loss": 4.095399856567383,
+ "step": 2685
+ },
+ {
+ "epoch": 37.30742358078603,
+ "grad_norm": 0.015503671951591969,
+ "learning_rate": 0.0006,
+ "loss": 4.034431457519531,
+ "step": 2686
+ },
+ {
+ "epoch": 37.32139737991266,
+ "grad_norm": 0.015370250679552555,
+ "learning_rate": 0.0006,
+ "loss": 4.075836181640625,
+ "step": 2687
+ },
+ {
+ "epoch": 37.3353711790393,
+ "grad_norm": 0.01564021222293377,
+ "learning_rate": 0.0006,
+ "loss": 4.125202178955078,
+ "step": 2688
+ },
+ {
+ "epoch": 37.34934497816594,
+ "grad_norm": 0.016717975959181786,
+ "learning_rate": 0.0006,
+ "loss": 4.153017044067383,
+ "step": 2689
+ },
+ {
+ "epoch": 37.36331877729258,
+ "grad_norm": 0.015263617038726807,
+ "learning_rate": 0.0006,
+ "loss": 4.021814823150635,
+ "step": 2690
+ },
+ {
+ "epoch": 37.377292576419215,
+ "grad_norm": 0.016497142612934113,
+ "learning_rate": 0.0006,
+ "loss": 4.069095611572266,
+ "step": 2691
+ },
+ {
+ "epoch": 37.391266375545854,
+ "grad_norm": 0.01669355481863022,
+ "learning_rate": 0.0006,
+ "loss": 4.174587726593018,
+ "step": 2692
+ },
+ {
+ "epoch": 37.40524017467249,
+ "grad_norm": 0.019695773720741272,
+ "learning_rate": 0.0006,
+ "loss": 4.0073041915893555,
+ "step": 2693
+ },
+ {
+ "epoch": 37.419213973799124,
+ "grad_norm": 0.023720644414424896,
+ "learning_rate": 0.0006,
+ "loss": 4.111948013305664,
+ "step": 2694
+ },
+ {
+ "epoch": 37.43318777292576,
+ "grad_norm": 0.02399350143969059,
+ "learning_rate": 0.0006,
+ "loss": 4.081164360046387,
+ "step": 2695
+ },
+ {
+ "epoch": 37.4471615720524,
+ "grad_norm": 0.019630271941423416,
+ "learning_rate": 0.0006,
+ "loss": 4.065731048583984,
+ "step": 2696
+ },
+ {
+ "epoch": 37.46113537117904,
+ "grad_norm": 0.019679008051753044,
+ "learning_rate": 0.0006,
+ "loss": 3.993852376937866,
+ "step": 2697
+ },
+ {
+ "epoch": 37.47510917030568,
+ "grad_norm": 0.020597299560904503,
+ "learning_rate": 0.0006,
+ "loss": 4.275555610656738,
+ "step": 2698
+ },
+ {
+ "epoch": 37.48908296943232,
+ "grad_norm": 0.01983889564871788,
+ "learning_rate": 0.0006,
+ "loss": 4.254708290100098,
+ "step": 2699
+ },
+ {
+ "epoch": 37.50305676855895,
+ "grad_norm": 0.019150668755173683,
+ "learning_rate": 0.0006,
+ "loss": 4.097885608673096,
+ "step": 2700
+ },
+ {
+ "epoch": 37.51703056768559,
+ "grad_norm": 0.018847715109586716,
+ "learning_rate": 0.0006,
+ "loss": 4.054081439971924,
+ "step": 2701
+ },
+ {
+ "epoch": 37.531004366812226,
+ "grad_norm": 0.019607288762927055,
+ "learning_rate": 0.0006,
+ "loss": 3.9891440868377686,
+ "step": 2702
+ },
+ {
+ "epoch": 37.544978165938865,
+ "grad_norm": 0.018863150849938393,
+ "learning_rate": 0.0006,
+ "loss": 4.084877014160156,
+ "step": 2703
+ },
+ {
+ "epoch": 37.5589519650655,
+ "grad_norm": 0.017340239137411118,
+ "learning_rate": 0.0006,
+ "loss": 4.113290786743164,
+ "step": 2704
+ },
+ {
+ "epoch": 37.57292576419214,
+ "grad_norm": 0.015628693625330925,
+ "learning_rate": 0.0006,
+ "loss": 4.1483001708984375,
+ "step": 2705
+ },
+ {
+ "epoch": 37.58689956331878,
+ "grad_norm": 0.016493283212184906,
+ "learning_rate": 0.0006,
+ "loss": 4.1040143966674805,
+ "step": 2706
+ },
+ {
+ "epoch": 37.60087336244541,
+ "grad_norm": 0.016732443124055862,
+ "learning_rate": 0.0006,
+ "loss": 4.07205057144165,
+ "step": 2707
+ },
+ {
+ "epoch": 37.61484716157205,
+ "grad_norm": 0.017176156863570213,
+ "learning_rate": 0.0006,
+ "loss": 4.1279754638671875,
+ "step": 2708
+ },
+ {
+ "epoch": 37.62882096069869,
+ "grad_norm": 0.015730608254671097,
+ "learning_rate": 0.0006,
+ "loss": 4.086658000946045,
+ "step": 2709
+ },
+ {
+ "epoch": 37.64279475982533,
+ "grad_norm": 0.016111090779304504,
+ "learning_rate": 0.0006,
+ "loss": 4.064335823059082,
+ "step": 2710
+ },
+ {
+ "epoch": 37.65676855895197,
+ "grad_norm": 0.01769379712641239,
+ "learning_rate": 0.0006,
+ "loss": 4.116047382354736,
+ "step": 2711
+ },
+ {
+ "epoch": 37.670742358078606,
+ "grad_norm": 0.01683102548122406,
+ "learning_rate": 0.0006,
+ "loss": 4.095986843109131,
+ "step": 2712
+ },
+ {
+ "epoch": 37.68471615720524,
+ "grad_norm": 0.015726450830698013,
+ "learning_rate": 0.0006,
+ "loss": 4.10775089263916,
+ "step": 2713
+ },
+ {
+ "epoch": 37.698689956331876,
+ "grad_norm": 0.014760235324501991,
+ "learning_rate": 0.0006,
+ "loss": 4.1234893798828125,
+ "step": 2714
+ },
+ {
+ "epoch": 37.712663755458514,
+ "grad_norm": 0.014933438040316105,
+ "learning_rate": 0.0006,
+ "loss": 4.079200744628906,
+ "step": 2715
+ },
+ {
+ "epoch": 37.72663755458515,
+ "grad_norm": 0.015723997727036476,
+ "learning_rate": 0.0006,
+ "loss": 4.07865047454834,
+ "step": 2716
+ },
+ {
+ "epoch": 37.74061135371179,
+ "grad_norm": 0.014539001509547234,
+ "learning_rate": 0.0006,
+ "loss": 4.036242485046387,
+ "step": 2717
+ },
+ {
+ "epoch": 37.75458515283843,
+ "grad_norm": 0.014844979159533978,
+ "learning_rate": 0.0006,
+ "loss": 4.1051483154296875,
+ "step": 2718
+ },
+ {
+ "epoch": 37.76855895196506,
+ "grad_norm": 0.015939844772219658,
+ "learning_rate": 0.0006,
+ "loss": 4.017253875732422,
+ "step": 2719
+ },
+ {
+ "epoch": 37.7825327510917,
+ "grad_norm": 0.014771555550396442,
+ "learning_rate": 0.0006,
+ "loss": 4.0694661140441895,
+ "step": 2720
+ },
+ {
+ "epoch": 37.79650655021834,
+ "grad_norm": 0.015777921304106712,
+ "learning_rate": 0.0006,
+ "loss": 3.947587490081787,
+ "step": 2721
+ },
+ {
+ "epoch": 37.81048034934498,
+ "grad_norm": 0.017451291903853416,
+ "learning_rate": 0.0006,
+ "loss": 4.093625068664551,
+ "step": 2722
+ },
+ {
+ "epoch": 37.82445414847162,
+ "grad_norm": 0.016829246655106544,
+ "learning_rate": 0.0006,
+ "loss": 4.09030818939209,
+ "step": 2723
+ },
+ {
+ "epoch": 37.838427947598255,
+ "grad_norm": 0.016984397545456886,
+ "learning_rate": 0.0006,
+ "loss": 4.140626907348633,
+ "step": 2724
+ },
+ {
+ "epoch": 37.852401746724894,
+ "grad_norm": 0.01732603646814823,
+ "learning_rate": 0.0006,
+ "loss": 4.128276824951172,
+ "step": 2725
+ },
+ {
+ "epoch": 37.866375545851525,
+ "grad_norm": 0.01744619384407997,
+ "learning_rate": 0.0006,
+ "loss": 4.103130340576172,
+ "step": 2726
+ },
+ {
+ "epoch": 37.880349344978164,
+ "grad_norm": 0.016331320628523827,
+ "learning_rate": 0.0006,
+ "loss": 4.037951469421387,
+ "step": 2727
+ },
+ {
+ "epoch": 37.8943231441048,
+ "grad_norm": 0.015919430181384087,
+ "learning_rate": 0.0006,
+ "loss": 4.121513366699219,
+ "step": 2728
+ },
+ {
+ "epoch": 37.90829694323144,
+ "grad_norm": 0.016153695061802864,
+ "learning_rate": 0.0006,
+ "loss": 4.099384307861328,
+ "step": 2729
+ },
+ {
+ "epoch": 37.92227074235808,
+ "grad_norm": 0.01628800667822361,
+ "learning_rate": 0.0006,
+ "loss": 4.092325687408447,
+ "step": 2730
+ },
+ {
+ "epoch": 37.93624454148472,
+ "grad_norm": 0.017285684123635292,
+ "learning_rate": 0.0006,
+ "loss": 3.9828906059265137,
+ "step": 2731
+ },
+ {
+ "epoch": 37.95021834061135,
+ "grad_norm": 0.016047554090619087,
+ "learning_rate": 0.0006,
+ "loss": 4.100419044494629,
+ "step": 2732
+ },
+ {
+ "epoch": 37.96419213973799,
+ "grad_norm": 0.01562986709177494,
+ "learning_rate": 0.0006,
+ "loss": 4.065832614898682,
+ "step": 2733
+ },
+ {
+ "epoch": 37.97816593886463,
+ "grad_norm": 0.01677163504064083,
+ "learning_rate": 0.0006,
+ "loss": 4.077518463134766,
+ "step": 2734
+ },
+ {
+ "epoch": 37.992139737991266,
+ "grad_norm": 0.01755492016673088,
+ "learning_rate": 0.0006,
+ "loss": 4.1076154708862305,
+ "step": 2735
+ },
+ {
+ "epoch": 38.0,
+ "grad_norm": 0.01929762214422226,
+ "learning_rate": 0.0006,
+ "loss": 4.09603214263916,
+ "step": 2736
+ },
+ {
+ "epoch": 38.0,
+ "eval_loss": 4.582442283630371,
+ "eval_runtime": 62.8649,
+ "eval_samples_per_second": 38.845,
+ "eval_steps_per_second": 1.225,
+ "step": 2736
+ },
+ {
+ "epoch": 38.01397379912664,
+ "grad_norm": 0.019129136577248573,
+ "learning_rate": 0.0006,
+ "loss": 4.095005989074707,
+ "step": 2737
+ },
+ {
+ "epoch": 38.02794759825328,
+ "grad_norm": 0.019167408347129822,
+ "learning_rate": 0.0006,
+ "loss": 4.054426670074463,
+ "step": 2738
+ },
+ {
+ "epoch": 38.041921397379916,
+ "grad_norm": 0.01789460889995098,
+ "learning_rate": 0.0006,
+ "loss": 4.0434417724609375,
+ "step": 2739
+ },
+ {
+ "epoch": 38.05589519650655,
+ "grad_norm": 0.015947027131915092,
+ "learning_rate": 0.0006,
+ "loss": 4.074516296386719,
+ "step": 2740
+ },
+ {
+ "epoch": 38.069868995633186,
+ "grad_norm": 0.016667252406477928,
+ "learning_rate": 0.0006,
+ "loss": 4.108545303344727,
+ "step": 2741
+ },
+ {
+ "epoch": 38.083842794759825,
+ "grad_norm": 0.016014399006962776,
+ "learning_rate": 0.0006,
+ "loss": 4.026925086975098,
+ "step": 2742
+ },
+ {
+ "epoch": 38.09781659388646,
+ "grad_norm": 0.015397695824503899,
+ "learning_rate": 0.0006,
+ "loss": 4.163970470428467,
+ "step": 2743
+ },
+ {
+ "epoch": 38.1117903930131,
+ "grad_norm": 0.01563289947807789,
+ "learning_rate": 0.0006,
+ "loss": 4.116578578948975,
+ "step": 2744
+ },
+ {
+ "epoch": 38.12576419213974,
+ "grad_norm": 0.015512318350374699,
+ "learning_rate": 0.0006,
+ "loss": 4.096311569213867,
+ "step": 2745
+ },
+ {
+ "epoch": 38.13973799126637,
+ "grad_norm": 0.015426642261445522,
+ "learning_rate": 0.0006,
+ "loss": 4.017065048217773,
+ "step": 2746
+ },
+ {
+ "epoch": 38.15371179039301,
+ "grad_norm": 0.015115435235202312,
+ "learning_rate": 0.0006,
+ "loss": 4.005434989929199,
+ "step": 2747
+ },
+ {
+ "epoch": 38.16768558951965,
+ "grad_norm": 0.015235894359648228,
+ "learning_rate": 0.0006,
+ "loss": 3.982743263244629,
+ "step": 2748
+ },
+ {
+ "epoch": 38.18165938864629,
+ "grad_norm": 0.016095101833343506,
+ "learning_rate": 0.0006,
+ "loss": 4.0849385261535645,
+ "step": 2749
+ },
+ {
+ "epoch": 38.19563318777293,
+ "grad_norm": 0.015201197937130928,
+ "learning_rate": 0.0006,
+ "loss": 4.103863716125488,
+ "step": 2750
+ },
+ {
+ "epoch": 38.209606986899566,
+ "grad_norm": 0.015687478706240654,
+ "learning_rate": 0.0006,
+ "loss": 4.12074089050293,
+ "step": 2751
+ },
+ {
+ "epoch": 38.223580786026204,
+ "grad_norm": 0.015201589092612267,
+ "learning_rate": 0.0006,
+ "loss": 4.056872367858887,
+ "step": 2752
+ },
+ {
+ "epoch": 38.237554585152836,
+ "grad_norm": 0.015154894441366196,
+ "learning_rate": 0.0006,
+ "loss": 4.089670181274414,
+ "step": 2753
+ },
+ {
+ "epoch": 38.251528384279474,
+ "grad_norm": 0.01586608588695526,
+ "learning_rate": 0.0006,
+ "loss": 4.155317306518555,
+ "step": 2754
+ },
+ {
+ "epoch": 38.26550218340611,
+ "grad_norm": 0.01609654724597931,
+ "learning_rate": 0.0006,
+ "loss": 3.995800495147705,
+ "step": 2755
+ },
+ {
+ "epoch": 38.27947598253275,
+ "grad_norm": 0.015495185740292072,
+ "learning_rate": 0.0006,
+ "loss": 3.995769739151001,
+ "step": 2756
+ },
+ {
+ "epoch": 38.29344978165939,
+ "grad_norm": 0.017042381688952446,
+ "learning_rate": 0.0006,
+ "loss": 4.132832050323486,
+ "step": 2757
+ },
+ {
+ "epoch": 38.30742358078603,
+ "grad_norm": 0.018706614151597023,
+ "learning_rate": 0.0006,
+ "loss": 4.038820743560791,
+ "step": 2758
+ },
+ {
+ "epoch": 38.32139737991266,
+ "grad_norm": 0.018316026777029037,
+ "learning_rate": 0.0006,
+ "loss": 3.972052574157715,
+ "step": 2759
+ },
+ {
+ "epoch": 38.3353711790393,
+ "grad_norm": 0.018375389277935028,
+ "learning_rate": 0.0006,
+ "loss": 4.152097702026367,
+ "step": 2760
+ },
+ {
+ "epoch": 38.34934497816594,
+ "grad_norm": 0.016879886388778687,
+ "learning_rate": 0.0006,
+ "loss": 4.020869731903076,
+ "step": 2761
+ },
+ {
+ "epoch": 38.36331877729258,
+ "grad_norm": 0.015728183090686798,
+ "learning_rate": 0.0006,
+ "loss": 4.008016586303711,
+ "step": 2762
+ },
+ {
+ "epoch": 38.377292576419215,
+ "grad_norm": 0.016147758811712265,
+ "learning_rate": 0.0006,
+ "loss": 4.035854339599609,
+ "step": 2763
+ },
+ {
+ "epoch": 38.391266375545854,
+ "grad_norm": 0.017774797976017,
+ "learning_rate": 0.0006,
+ "loss": 4.122506141662598,
+ "step": 2764
+ },
+ {
+ "epoch": 38.40524017467249,
+ "grad_norm": 0.016590069979429245,
+ "learning_rate": 0.0006,
+ "loss": 4.082847595214844,
+ "step": 2765
+ },
+ {
+ "epoch": 38.419213973799124,
+ "grad_norm": 0.016544528305530548,
+ "learning_rate": 0.0006,
+ "loss": 4.184270858764648,
+ "step": 2766
+ },
+ {
+ "epoch": 38.43318777292576,
+ "grad_norm": 0.017972925677895546,
+ "learning_rate": 0.0006,
+ "loss": 3.987558603286743,
+ "step": 2767
+ },
+ {
+ "epoch": 38.4471615720524,
+ "grad_norm": 0.018562227487564087,
+ "learning_rate": 0.0006,
+ "loss": 4.1193461418151855,
+ "step": 2768
+ },
+ {
+ "epoch": 38.46113537117904,
+ "grad_norm": 0.018896685913205147,
+ "learning_rate": 0.0006,
+ "loss": 4.057808876037598,
+ "step": 2769
+ },
+ {
+ "epoch": 38.47510917030568,
+ "grad_norm": 0.016793809831142426,
+ "learning_rate": 0.0006,
+ "loss": 4.069523334503174,
+ "step": 2770
+ },
+ {
+ "epoch": 38.48908296943232,
+ "grad_norm": 0.01698676310479641,
+ "learning_rate": 0.0006,
+ "loss": 3.997624158859253,
+ "step": 2771
+ },
+ {
+ "epoch": 38.50305676855895,
+ "grad_norm": 0.016784202307462692,
+ "learning_rate": 0.0006,
+ "loss": 4.174795627593994,
+ "step": 2772
+ },
+ {
+ "epoch": 38.51703056768559,
+ "grad_norm": 0.016785014420747757,
+ "learning_rate": 0.0006,
+ "loss": 3.999321222305298,
+ "step": 2773
+ },
+ {
+ "epoch": 38.531004366812226,
+ "grad_norm": 0.018615184351801872,
+ "learning_rate": 0.0006,
+ "loss": 4.191237926483154,
+ "step": 2774
+ },
+ {
+ "epoch": 38.544978165938865,
+ "grad_norm": 0.019090790301561356,
+ "learning_rate": 0.0006,
+ "loss": 4.106966018676758,
+ "step": 2775
+ },
+ {
+ "epoch": 38.5589519650655,
+ "grad_norm": 0.018298320472240448,
+ "learning_rate": 0.0006,
+ "loss": 4.063363552093506,
+ "step": 2776
+ },
+ {
+ "epoch": 38.57292576419214,
+ "grad_norm": 0.018541090190410614,
+ "learning_rate": 0.0006,
+ "loss": 4.101193428039551,
+ "step": 2777
+ },
+ {
+ "epoch": 38.58689956331878,
+ "grad_norm": 0.01849970407783985,
+ "learning_rate": 0.0006,
+ "loss": 4.071335792541504,
+ "step": 2778
+ },
+ {
+ "epoch": 38.60087336244541,
+ "grad_norm": 0.018309077247977257,
+ "learning_rate": 0.0006,
+ "loss": 4.245305061340332,
+ "step": 2779
+ },
+ {
+ "epoch": 38.61484716157205,
+ "grad_norm": 0.017681840807199478,
+ "learning_rate": 0.0006,
+ "loss": 4.081173419952393,
+ "step": 2780
+ },
+ {
+ "epoch": 38.62882096069869,
+ "grad_norm": 0.016286224126815796,
+ "learning_rate": 0.0006,
+ "loss": 4.024351596832275,
+ "step": 2781
+ },
+ {
+ "epoch": 38.64279475982533,
+ "grad_norm": 0.015837235376238823,
+ "learning_rate": 0.0006,
+ "loss": 4.028205394744873,
+ "step": 2782
+ },
+ {
+ "epoch": 38.65676855895197,
+ "grad_norm": 0.016062596812844276,
+ "learning_rate": 0.0006,
+ "loss": 4.131122589111328,
+ "step": 2783
+ },
+ {
+ "epoch": 38.670742358078606,
+ "grad_norm": 0.01711239665746689,
+ "learning_rate": 0.0006,
+ "loss": 4.124416351318359,
+ "step": 2784
+ },
+ {
+ "epoch": 38.68471615720524,
+ "grad_norm": 0.015304913744330406,
+ "learning_rate": 0.0006,
+ "loss": 4.056122779846191,
+ "step": 2785
+ },
+ {
+ "epoch": 38.698689956331876,
+ "grad_norm": 0.015971887856721878,
+ "learning_rate": 0.0006,
+ "loss": 4.132555961608887,
+ "step": 2786
+ },
+ {
+ "epoch": 38.712663755458514,
+ "grad_norm": 0.015787221491336823,
+ "learning_rate": 0.0006,
+ "loss": 4.051596164703369,
+ "step": 2787
+ },
+ {
+ "epoch": 38.72663755458515,
+ "grad_norm": 0.01605444774031639,
+ "learning_rate": 0.0006,
+ "loss": 4.191808700561523,
+ "step": 2788
+ },
+ {
+ "epoch": 38.74061135371179,
+ "grad_norm": 0.016383009031414986,
+ "learning_rate": 0.0006,
+ "loss": 4.0648417472839355,
+ "step": 2789
+ },
+ {
+ "epoch": 38.75458515283843,
+ "grad_norm": 0.016135554760694504,
+ "learning_rate": 0.0006,
+ "loss": 3.988563060760498,
+ "step": 2790
+ },
+ {
+ "epoch": 38.76855895196506,
+ "grad_norm": 0.016634158790111542,
+ "learning_rate": 0.0006,
+ "loss": 4.133493423461914,
+ "step": 2791
+ },
+ {
+ "epoch": 38.7825327510917,
+ "grad_norm": 0.01688210666179657,
+ "learning_rate": 0.0006,
+ "loss": 4.091277122497559,
+ "step": 2792
+ },
+ {
+ "epoch": 38.79650655021834,
+ "grad_norm": 0.016102122142910957,
+ "learning_rate": 0.0006,
+ "loss": 4.051219940185547,
+ "step": 2793
+ },
+ {
+ "epoch": 38.81048034934498,
+ "grad_norm": 0.016585038974881172,
+ "learning_rate": 0.0006,
+ "loss": 4.058587074279785,
+ "step": 2794
+ },
+ {
+ "epoch": 38.82445414847162,
+ "grad_norm": 0.017700951546430588,
+ "learning_rate": 0.0006,
+ "loss": 4.016030311584473,
+ "step": 2795
+ },
+ {
+ "epoch": 38.838427947598255,
+ "grad_norm": 0.017922203987836838,
+ "learning_rate": 0.0006,
+ "loss": 4.079977989196777,
+ "step": 2796
+ },
+ {
+ "epoch": 38.852401746724894,
+ "grad_norm": 0.017299890518188477,
+ "learning_rate": 0.0006,
+ "loss": 4.062984943389893,
+ "step": 2797
+ },
+ {
+ "epoch": 38.866375545851525,
+ "grad_norm": 0.018648581579327583,
+ "learning_rate": 0.0006,
+ "loss": 4.065717697143555,
+ "step": 2798
+ },
+ {
+ "epoch": 38.880349344978164,
+ "grad_norm": 0.019057920202612877,
+ "learning_rate": 0.0006,
+ "loss": 4.227892875671387,
+ "step": 2799
+ },
+ {
+ "epoch": 38.8943231441048,
+ "grad_norm": 0.018824422731995583,
+ "learning_rate": 0.0006,
+ "loss": 4.116058349609375,
+ "step": 2800
+ },
+ {
+ "epoch": 38.90829694323144,
+ "grad_norm": 0.019084978848695755,
+ "learning_rate": 0.0006,
+ "loss": 4.061677932739258,
+ "step": 2801
+ },
+ {
+ "epoch": 38.92227074235808,
+ "grad_norm": 0.017842981964349747,
+ "learning_rate": 0.0006,
+ "loss": 4.083813667297363,
+ "step": 2802
+ },
+ {
+ "epoch": 38.93624454148472,
+ "grad_norm": 0.01880858838558197,
+ "learning_rate": 0.0006,
+ "loss": 4.152561187744141,
+ "step": 2803
+ },
+ {
+ "epoch": 38.95021834061135,
+ "grad_norm": 0.017400087788701057,
+ "learning_rate": 0.0006,
+ "loss": 4.154635429382324,
+ "step": 2804
+ },
+ {
+ "epoch": 38.96419213973799,
+ "grad_norm": 0.017682507634162903,
+ "learning_rate": 0.0006,
+ "loss": 4.151128768920898,
+ "step": 2805
+ },
+ {
+ "epoch": 38.97816593886463,
+ "grad_norm": 0.015803052112460136,
+ "learning_rate": 0.0006,
+ "loss": 4.10498571395874,
+ "step": 2806
+ },
+ {
+ "epoch": 38.992139737991266,
+ "grad_norm": 0.01582511141896248,
+ "learning_rate": 0.0006,
+ "loss": 4.140159606933594,
+ "step": 2807
+ },
+ {
+ "epoch": 39.0,
+ "grad_norm": 0.01925666816532612,
+ "learning_rate": 0.0006,
+ "loss": 4.126394271850586,
+ "step": 2808
+ },
+ {
+ "epoch": 39.0,
+ "eval_loss": 4.615617275238037,
+ "eval_runtime": 56.8895,
+ "eval_samples_per_second": 42.925,
+ "eval_steps_per_second": 1.354,
+ "step": 2808
+ },
+ {
+ "epoch": 39.01397379912664,
+ "grad_norm": 0.01677853614091873,
+ "learning_rate": 0.0006,
+ "loss": 4.102395057678223,
+ "step": 2809
+ },
+ {
+ "epoch": 39.02794759825328,
+ "grad_norm": 0.017901137471199036,
+ "learning_rate": 0.0006,
+ "loss": 3.976609230041504,
+ "step": 2810
+ },
+ {
+ "epoch": 39.041921397379916,
+ "grad_norm": 0.017978668212890625,
+ "learning_rate": 0.0006,
+ "loss": 3.944416046142578,
+ "step": 2811
+ },
+ {
+ "epoch": 39.05589519650655,
+ "grad_norm": 0.019967492669820786,
+ "learning_rate": 0.0006,
+ "loss": 3.992234468460083,
+ "step": 2812
+ },
+ {
+ "epoch": 39.069868995633186,
+ "grad_norm": 0.018937544897198677,
+ "learning_rate": 0.0006,
+ "loss": 4.028573036193848,
+ "step": 2813
+ },
+ {
+ "epoch": 39.083842794759825,
+ "grad_norm": 0.016852952539920807,
+ "learning_rate": 0.0006,
+ "loss": 4.062442779541016,
+ "step": 2814
+ },
+ {
+ "epoch": 39.09781659388646,
+ "grad_norm": 0.016643647104501724,
+ "learning_rate": 0.0006,
+ "loss": 4.08871603012085,
+ "step": 2815
+ },
+ {
+ "epoch": 39.1117903930131,
+ "grad_norm": 0.016872866079211235,
+ "learning_rate": 0.0006,
+ "loss": 3.9514245986938477,
+ "step": 2816
+ },
+ {
+ "epoch": 39.12576419213974,
+ "grad_norm": 0.018045669421553612,
+ "learning_rate": 0.0006,
+ "loss": 4.038842678070068,
+ "step": 2817
+ },
+ {
+ "epoch": 39.13973799126637,
+ "grad_norm": 0.017853403463959694,
+ "learning_rate": 0.0006,
+ "loss": 3.93709135055542,
+ "step": 2818
+ },
+ {
+ "epoch": 39.15371179039301,
+ "grad_norm": 0.017507443204522133,
+ "learning_rate": 0.0006,
+ "loss": 3.98665714263916,
+ "step": 2819
+ },
+ {
+ "epoch": 39.16768558951965,
+ "grad_norm": 0.01766853965818882,
+ "learning_rate": 0.0006,
+ "loss": 3.937126636505127,
+ "step": 2820
+ },
+ {
+ "epoch": 39.18165938864629,
+ "grad_norm": 0.016660021618008614,
+ "learning_rate": 0.0006,
+ "loss": 3.8397014141082764,
+ "step": 2821
+ },
+ {
+ "epoch": 39.19563318777293,
+ "grad_norm": 0.017719309777021408,
+ "learning_rate": 0.0006,
+ "loss": 4.003579616546631,
+ "step": 2822
+ },
+ {
+ "epoch": 39.209606986899566,
+ "grad_norm": 0.018234414979815483,
+ "learning_rate": 0.0006,
+ "loss": 4.053859710693359,
+ "step": 2823
+ },
+ {
+ "epoch": 39.223580786026204,
+ "grad_norm": 0.018777024000883102,
+ "learning_rate": 0.0006,
+ "loss": 4.056182861328125,
+ "step": 2824
+ },
+ {
+ "epoch": 39.237554585152836,
+ "grad_norm": 0.017443502321839333,
+ "learning_rate": 0.0006,
+ "loss": 4.025068283081055,
+ "step": 2825
+ },
+ {
+ "epoch": 39.251528384279474,
+ "grad_norm": 0.016500331461429596,
+ "learning_rate": 0.0006,
+ "loss": 4.101631164550781,
+ "step": 2826
+ },
+ {
+ "epoch": 39.26550218340611,
+ "grad_norm": 0.016101544722914696,
+ "learning_rate": 0.0006,
+ "loss": 4.027082443237305,
+ "step": 2827
+ },
+ {
+ "epoch": 39.27947598253275,
+ "grad_norm": 0.015471171587705612,
+ "learning_rate": 0.0006,
+ "loss": 4.022909641265869,
+ "step": 2828
+ },
+ {
+ "epoch": 39.29344978165939,
+ "grad_norm": 0.015779748558998108,
+ "learning_rate": 0.0006,
+ "loss": 4.18796968460083,
+ "step": 2829
+ },
+ {
+ "epoch": 39.30742358078603,
+ "grad_norm": 0.015522679314017296,
+ "learning_rate": 0.0006,
+ "loss": 4.005777359008789,
+ "step": 2830
+ },
+ {
+ "epoch": 39.32139737991266,
+ "grad_norm": 0.015226504765450954,
+ "learning_rate": 0.0006,
+ "loss": 4.108416557312012,
+ "step": 2831
+ },
+ {
+ "epoch": 39.3353711790393,
+ "grad_norm": 0.015095778740942478,
+ "learning_rate": 0.0006,
+ "loss": 4.08029842376709,
+ "step": 2832
+ },
+ {
+ "epoch": 39.34934497816594,
+ "grad_norm": 0.015601031482219696,
+ "learning_rate": 0.0006,
+ "loss": 3.989492416381836,
+ "step": 2833
+ },
+ {
+ "epoch": 39.36331877729258,
+ "grad_norm": 0.014825052581727505,
+ "learning_rate": 0.0006,
+ "loss": 3.998520612716675,
+ "step": 2834
+ },
+ {
+ "epoch": 39.377292576419215,
+ "grad_norm": 0.014821100980043411,
+ "learning_rate": 0.0006,
+ "loss": 4.001777172088623,
+ "step": 2835
+ },
+ {
+ "epoch": 39.391266375545854,
+ "grad_norm": 0.015333754941821098,
+ "learning_rate": 0.0006,
+ "loss": 4.11985969543457,
+ "step": 2836
+ },
+ {
+ "epoch": 39.40524017467249,
+ "grad_norm": 0.015940699726343155,
+ "learning_rate": 0.0006,
+ "loss": 4.025373935699463,
+ "step": 2837
+ },
+ {
+ "epoch": 39.419213973799124,
+ "grad_norm": 0.015096287243068218,
+ "learning_rate": 0.0006,
+ "loss": 4.068551063537598,
+ "step": 2838
+ },
+ {
+ "epoch": 39.43318777292576,
+ "grad_norm": 0.01533548254519701,
+ "learning_rate": 0.0006,
+ "loss": 3.9800801277160645,
+ "step": 2839
+ },
+ {
+ "epoch": 39.4471615720524,
+ "grad_norm": 0.014648732729256153,
+ "learning_rate": 0.0006,
+ "loss": 3.9670190811157227,
+ "step": 2840
+ },
+ {
+ "epoch": 39.46113537117904,
+ "grad_norm": 0.015759671106934547,
+ "learning_rate": 0.0006,
+ "loss": 4.050443649291992,
+ "step": 2841
+ },
+ {
+ "epoch": 39.47510917030568,
+ "grad_norm": 0.016054073348641396,
+ "learning_rate": 0.0006,
+ "loss": 4.22734260559082,
+ "step": 2842
+ },
+ {
+ "epoch": 39.48908296943232,
+ "grad_norm": 0.0160264540463686,
+ "learning_rate": 0.0006,
+ "loss": 4.116174221038818,
+ "step": 2843
+ },
+ {
+ "epoch": 39.50305676855895,
+ "grad_norm": 0.015426610596477985,
+ "learning_rate": 0.0006,
+ "loss": 4.000178337097168,
+ "step": 2844
+ },
+ {
+ "epoch": 39.51703056768559,
+ "grad_norm": 0.01520642638206482,
+ "learning_rate": 0.0006,
+ "loss": 4.218410968780518,
+ "step": 2845
+ },
+ {
+ "epoch": 39.531004366812226,
+ "grad_norm": 0.015305949375033379,
+ "learning_rate": 0.0006,
+ "loss": 4.025891304016113,
+ "step": 2846
+ },
+ {
+ "epoch": 39.544978165938865,
+ "grad_norm": 0.015461564064025879,
+ "learning_rate": 0.0006,
+ "loss": 4.100950717926025,
+ "step": 2847
+ },
+ {
+ "epoch": 39.5589519650655,
+ "grad_norm": 0.01712625101208687,
+ "learning_rate": 0.0006,
+ "loss": 4.0620856285095215,
+ "step": 2848
+ },
+ {
+ "epoch": 39.57292576419214,
+ "grad_norm": 0.017133301123976707,
+ "learning_rate": 0.0006,
+ "loss": 4.097138404846191,
+ "step": 2849
+ },
+ {
+ "epoch": 39.58689956331878,
+ "grad_norm": 0.01683625392615795,
+ "learning_rate": 0.0006,
+ "loss": 4.084364891052246,
+ "step": 2850
+ },
+ {
+ "epoch": 39.60087336244541,
+ "grad_norm": 0.0165710486471653,
+ "learning_rate": 0.0006,
+ "loss": 4.042361259460449,
+ "step": 2851
+ },
+ {
+ "epoch": 39.61484716157205,
+ "grad_norm": 0.01660393364727497,
+ "learning_rate": 0.0006,
+ "loss": 4.1395487785339355,
+ "step": 2852
+ },
+ {
+ "epoch": 39.62882096069869,
+ "grad_norm": 0.017025845125317574,
+ "learning_rate": 0.0006,
+ "loss": 4.1207075119018555,
+ "step": 2853
+ },
+ {
+ "epoch": 39.64279475982533,
+ "grad_norm": 0.01660497486591339,
+ "learning_rate": 0.0006,
+ "loss": 4.103846073150635,
+ "step": 2854
+ },
+ {
+ "epoch": 39.65676855895197,
+ "grad_norm": 0.01607946865260601,
+ "learning_rate": 0.0006,
+ "loss": 4.189016342163086,
+ "step": 2855
+ },
+ {
+ "epoch": 39.670742358078606,
+ "grad_norm": 0.01614505983889103,
+ "learning_rate": 0.0006,
+ "loss": 4.079860687255859,
+ "step": 2856
+ },
+ {
+ "epoch": 39.68471615720524,
+ "grad_norm": 0.017085086554288864,
+ "learning_rate": 0.0006,
+ "loss": 3.9774293899536133,
+ "step": 2857
+ },
+ {
+ "epoch": 39.698689956331876,
+ "grad_norm": 0.01816520094871521,
+ "learning_rate": 0.0006,
+ "loss": 4.113666534423828,
+ "step": 2858
+ },
+ {
+ "epoch": 39.712663755458514,
+ "grad_norm": 0.016597526147961617,
+ "learning_rate": 0.0006,
+ "loss": 4.068242073059082,
+ "step": 2859
+ },
+ {
+ "epoch": 39.72663755458515,
+ "grad_norm": 0.01672235131263733,
+ "learning_rate": 0.0006,
+ "loss": 4.024444580078125,
+ "step": 2860
+ },
+ {
+ "epoch": 39.74061135371179,
+ "grad_norm": 0.020093556493520737,
+ "learning_rate": 0.0006,
+ "loss": 4.064124584197998,
+ "step": 2861
+ },
+ {
+ "epoch": 39.75458515283843,
+ "grad_norm": 0.023141304031014442,
+ "learning_rate": 0.0006,
+ "loss": 4.130439758300781,
+ "step": 2862
+ },
+ {
+ "epoch": 39.76855895196506,
+ "grad_norm": 0.02289372682571411,
+ "learning_rate": 0.0006,
+ "loss": 4.14794921875,
+ "step": 2863
+ },
+ {
+ "epoch": 39.7825327510917,
+ "grad_norm": 0.01919020712375641,
+ "learning_rate": 0.0006,
+ "loss": 4.123961925506592,
+ "step": 2864
+ },
+ {
+ "epoch": 39.79650655021834,
+ "grad_norm": 0.016932329162955284,
+ "learning_rate": 0.0006,
+ "loss": 4.0944132804870605,
+ "step": 2865
+ },
+ {
+ "epoch": 39.81048034934498,
+ "grad_norm": 0.016644535586237907,
+ "learning_rate": 0.0006,
+ "loss": 4.124269962310791,
+ "step": 2866
+ },
+ {
+ "epoch": 39.82445414847162,
+ "grad_norm": 0.016302216798067093,
+ "learning_rate": 0.0006,
+ "loss": 4.0267205238342285,
+ "step": 2867
+ },
+ {
+ "epoch": 39.838427947598255,
+ "grad_norm": 0.01673300378024578,
+ "learning_rate": 0.0006,
+ "loss": 4.10650634765625,
+ "step": 2868
+ },
+ {
+ "epoch": 39.852401746724894,
+ "grad_norm": 0.016581356525421143,
+ "learning_rate": 0.0006,
+ "loss": 4.1547956466674805,
+ "step": 2869
+ },
+ {
+ "epoch": 39.866375545851525,
+ "grad_norm": 0.01691797934472561,
+ "learning_rate": 0.0006,
+ "loss": 4.056936740875244,
+ "step": 2870
+ },
+ {
+ "epoch": 39.880349344978164,
+ "grad_norm": 0.01596726104617119,
+ "learning_rate": 0.0006,
+ "loss": 3.9891414642333984,
+ "step": 2871
+ },
+ {
+ "epoch": 39.8943231441048,
+ "grad_norm": 0.015615091659128666,
+ "learning_rate": 0.0006,
+ "loss": 4.137948036193848,
+ "step": 2872
+ },
+ {
+ "epoch": 39.90829694323144,
+ "grad_norm": 0.015703413635492325,
+ "learning_rate": 0.0006,
+ "loss": 3.9844930171966553,
+ "step": 2873
+ },
+ {
+ "epoch": 39.92227074235808,
+ "grad_norm": 0.015453094616532326,
+ "learning_rate": 0.0006,
+ "loss": 4.098443508148193,
+ "step": 2874
+ },
+ {
+ "epoch": 39.93624454148472,
+ "grad_norm": 0.01449158601462841,
+ "learning_rate": 0.0006,
+ "loss": 4.090852737426758,
+ "step": 2875
+ },
+ {
+ "epoch": 39.95021834061135,
+ "grad_norm": 0.01537515688687563,
+ "learning_rate": 0.0006,
+ "loss": 4.239612579345703,
+ "step": 2876
+ },
+ {
+ "epoch": 39.96419213973799,
+ "grad_norm": 0.015989921987056732,
+ "learning_rate": 0.0006,
+ "loss": 4.142073154449463,
+ "step": 2877
+ },
+ {
+ "epoch": 39.97816593886463,
+ "grad_norm": 0.015985611826181412,
+ "learning_rate": 0.0006,
+ "loss": 4.100076675415039,
+ "step": 2878
+ },
+ {
+ "epoch": 39.992139737991266,
+ "grad_norm": 0.015260833315551281,
+ "learning_rate": 0.0006,
+ "loss": 4.068576335906982,
+ "step": 2879
+ },
+ {
+ "epoch": 40.0,
+ "grad_norm": 0.018047941848635674,
+ "learning_rate": 0.0006,
+ "loss": 4.005878925323486,
+ "step": 2880
+ },
+ {
+ "epoch": 40.0,
+ "eval_loss": 4.622848987579346,
+ "eval_runtime": 56.9618,
+ "eval_samples_per_second": 42.871,
+ "eval_steps_per_second": 1.352,
+ "step": 2880
+ },
+ {
+ "epoch": 40.01397379912664,
+ "grad_norm": 0.015976838767528534,
+ "learning_rate": 0.0006,
+ "loss": 4.035689353942871,
+ "step": 2881
+ },
+ {
+ "epoch": 40.02794759825328,
+ "grad_norm": 0.015176220797002316,
+ "learning_rate": 0.0006,
+ "loss": 3.932197332382202,
+ "step": 2882
+ },
+ {
+ "epoch": 40.041921397379916,
+ "grad_norm": 0.016676228493452072,
+ "learning_rate": 0.0006,
+ "loss": 4.014171123504639,
+ "step": 2883
+ },
+ {
+ "epoch": 40.05589519650655,
+ "grad_norm": 0.01698281429708004,
+ "learning_rate": 0.0006,
+ "loss": 4.117893218994141,
+ "step": 2884
+ },
+ {
+ "epoch": 40.069868995633186,
+ "grad_norm": 0.017872486263513565,
+ "learning_rate": 0.0006,
+ "loss": 4.132119178771973,
+ "step": 2885
+ },
+ {
+ "epoch": 40.083842794759825,
+ "grad_norm": 0.017615554854273796,
+ "learning_rate": 0.0006,
+ "loss": 4.072000980377197,
+ "step": 2886
+ },
+ {
+ "epoch": 40.09781659388646,
+ "grad_norm": 0.017469394952058792,
+ "learning_rate": 0.0006,
+ "loss": 4.017726421356201,
+ "step": 2887
+ },
+ {
+ "epoch": 40.1117903930131,
+ "grad_norm": 0.01954762637615204,
+ "learning_rate": 0.0006,
+ "loss": 4.014443874359131,
+ "step": 2888
+ },
+ {
+ "epoch": 40.12576419213974,
+ "grad_norm": 0.018935853615403175,
+ "learning_rate": 0.0006,
+ "loss": 4.0042948722839355,
+ "step": 2889
+ },
+ {
+ "epoch": 40.13973799126637,
+ "grad_norm": 0.019382692873477936,
+ "learning_rate": 0.0006,
+ "loss": 4.06160831451416,
+ "step": 2890
+ },
+ {
+ "epoch": 40.15371179039301,
+ "grad_norm": 0.019677115604281425,
+ "learning_rate": 0.0006,
+ "loss": 4.13466215133667,
+ "step": 2891
+ },
+ {
+ "epoch": 40.16768558951965,
+ "grad_norm": 0.020365122705698013,
+ "learning_rate": 0.0006,
+ "loss": 4.095000267028809,
+ "step": 2892
+ },
+ {
+ "epoch": 40.18165938864629,
+ "grad_norm": 0.021842138841748238,
+ "learning_rate": 0.0006,
+ "loss": 3.9877853393554688,
+ "step": 2893
+ },
+ {
+ "epoch": 40.19563318777293,
+ "grad_norm": 0.02354881353676319,
+ "learning_rate": 0.0006,
+ "loss": 3.9430112838745117,
+ "step": 2894
+ },
+ {
+ "epoch": 40.209606986899566,
+ "grad_norm": 0.026602735742926598,
+ "learning_rate": 0.0006,
+ "loss": 4.012076377868652,
+ "step": 2895
+ },
+ {
+ "epoch": 40.223580786026204,
+ "grad_norm": 0.023976484313607216,
+ "learning_rate": 0.0006,
+ "loss": 3.9768595695495605,
+ "step": 2896
+ },
+ {
+ "epoch": 40.237554585152836,
+ "grad_norm": 0.02336616814136505,
+ "learning_rate": 0.0006,
+ "loss": 4.1096038818359375,
+ "step": 2897
+ },
+ {
+ "epoch": 40.251528384279474,
+ "grad_norm": 0.02426023595035076,
+ "learning_rate": 0.0006,
+ "loss": 4.157189846038818,
+ "step": 2898
+ },
+ {
+ "epoch": 40.26550218340611,
+ "grad_norm": 0.021572696045041084,
+ "learning_rate": 0.0006,
+ "loss": 3.979175090789795,
+ "step": 2899
+ },
+ {
+ "epoch": 40.27947598253275,
+ "grad_norm": 0.08300432562828064,
+ "learning_rate": 0.0006,
+ "loss": 4.092336654663086,
+ "step": 2900
+ },
+ {
+ "epoch": 40.29344978165939,
+ "grad_norm": 0.7323967814445496,
+ "learning_rate": 0.0006,
+ "loss": 5.337990760803223,
+ "step": 2901
+ },
+ {
+ "epoch": 40.30742358078603,
+ "grad_norm": 0.3363337814807892,
+ "learning_rate": 0.0006,
+ "loss": 6.40043830871582,
+ "step": 2902
+ },
+ {
+ "epoch": 40.32139737991266,
+ "grad_norm": 0.598686933517456,
+ "learning_rate": 0.0006,
+ "loss": 7.004765510559082,
+ "step": 2903
+ },
+ {
+ "epoch": 40.3353711790393,
+ "grad_norm": 0.5998364686965942,
+ "learning_rate": 0.0006,
+ "loss": 8.184747695922852,
+ "step": 2904
+ },
+ {
+ "epoch": 40.34934497816594,
+ "grad_norm": 1.408208966255188,
+ "learning_rate": 0.0006,
+ "loss": 8.88978385925293,
+ "step": 2905
+ },
+ {
+ "epoch": 40.36331877729258,
+ "grad_norm": 0.7829785943031311,
+ "learning_rate": 0.0006,
+ "loss": 11.13601303100586,
+ "step": 2906
+ },
+ {
+ "epoch": 40.377292576419215,
+ "grad_norm": 0.7769215106964111,
+ "learning_rate": 0.0006,
+ "loss": 9.90416145324707,
+ "step": 2907
+ },
+ {
+ "epoch": 40.391266375545854,
+ "grad_norm": 0.46508079767227173,
+ "learning_rate": 0.0006,
+ "loss": 9.647209167480469,
+ "step": 2908
+ },
+ {
+ "epoch": 40.40524017467249,
+ "grad_norm": 1.0228968858718872,
+ "learning_rate": 0.0006,
+ "loss": 11.685434341430664,
+ "step": 2909
+ },
+ {
+ "epoch": 40.419213973799124,
+ "grad_norm": 0.20553778111934662,
+ "learning_rate": 0.0006,
+ "loss": 8.621708869934082,
+ "step": 2910
+ },
+ {
+ "epoch": 40.43318777292576,
+ "grad_norm": 0.18466424942016602,
+ "learning_rate": 0.0006,
+ "loss": 8.590787887573242,
+ "step": 2911
+ },
+ {
+ "epoch": 40.4471615720524,
+ "grad_norm": 0.10458594560623169,
+ "learning_rate": 0.0006,
+ "loss": 8.324458122253418,
+ "step": 2912
+ },
+ {
+ "epoch": 40.46113537117904,
+ "grad_norm": 0.1096842959523201,
+ "learning_rate": 0.0006,
+ "loss": 8.300851821899414,
+ "step": 2913
+ },
+ {
+ "epoch": 40.47510917030568,
+ "grad_norm": 0.10501226782798767,
+ "learning_rate": 0.0006,
+ "loss": 8.21009635925293,
+ "step": 2914
+ },
+ {
+ "epoch": 40.48908296943232,
+ "grad_norm": 0.13704165816307068,
+ "learning_rate": 0.0006,
+ "loss": 8.141355514526367,
+ "step": 2915
+ },
+ {
+ "epoch": 40.50305676855895,
+ "grad_norm": 0.12129108607769012,
+ "learning_rate": 0.0006,
+ "loss": 8.074612617492676,
+ "step": 2916
+ },
+ {
+ "epoch": 40.51703056768559,
+ "grad_norm": 0.10028897225856781,
+ "learning_rate": 0.0006,
+ "loss": 8.00384521484375,
+ "step": 2917
+ },
+ {
+ "epoch": 40.531004366812226,
+ "grad_norm": 0.11764246970415115,
+ "learning_rate": 0.0006,
+ "loss": 8.078598022460938,
+ "step": 2918
+ },
+ {
+ "epoch": 40.544978165938865,
+ "grad_norm": 0.10081959515810013,
+ "learning_rate": 0.0006,
+ "loss": 8.008938789367676,
+ "step": 2919
+ },
+ {
+ "epoch": 40.5589519650655,
+ "grad_norm": 0.1003170907497406,
+ "learning_rate": 0.0006,
+ "loss": 7.962940216064453,
+ "step": 2920
+ },
+ {
+ "epoch": 40.57292576419214,
+ "grad_norm": 0.08763492107391357,
+ "learning_rate": 0.0006,
+ "loss": 7.909419536590576,
+ "step": 2921
+ },
+ {
+ "epoch": 40.58689956331878,
+ "grad_norm": 0.0636807233095169,
+ "learning_rate": 0.0006,
+ "loss": 7.878905773162842,
+ "step": 2922
+ },
+ {
+ "epoch": 40.60087336244541,
+ "grad_norm": 0.06005936488509178,
+ "learning_rate": 0.0006,
+ "loss": 7.850980758666992,
+ "step": 2923
+ },
+ {
+ "epoch": 40.61484716157205,
+ "grad_norm": 0.06415296345949173,
+ "learning_rate": 0.0006,
+ "loss": 7.830353260040283,
+ "step": 2924
+ },
+ {
+ "epoch": 40.62882096069869,
+ "grad_norm": 0.0732007846236229,
+ "learning_rate": 0.0006,
+ "loss": 7.830746650695801,
+ "step": 2925
+ },
+ {
+ "epoch": 40.64279475982533,
+ "grad_norm": 0.06628110259771347,
+ "learning_rate": 0.0006,
+ "loss": 7.82661771774292,
+ "step": 2926
+ },
+ {
+ "epoch": 40.65676855895197,
+ "grad_norm": 0.05544218048453331,
+ "learning_rate": 0.0006,
+ "loss": 7.814305782318115,
+ "step": 2927
+ },
+ {
+ "epoch": 40.670742358078606,
+ "grad_norm": 0.050327103585004807,
+ "learning_rate": 0.0006,
+ "loss": 7.764055252075195,
+ "step": 2928
+ },
+ {
+ "epoch": 40.68471615720524,
+ "grad_norm": 0.04918425530195236,
+ "learning_rate": 0.0006,
+ "loss": 7.744625091552734,
+ "step": 2929
+ },
+ {
+ "epoch": 40.698689956331876,
+ "grad_norm": 0.042862266302108765,
+ "learning_rate": 0.0006,
+ "loss": 7.764426231384277,
+ "step": 2930
+ },
+ {
+ "epoch": 40.712663755458514,
+ "grad_norm": 0.04338089004158974,
+ "learning_rate": 0.0006,
+ "loss": 7.712530136108398,
+ "step": 2931
+ },
+ {
+ "epoch": 40.72663755458515,
+ "grad_norm": 0.04127737879753113,
+ "learning_rate": 0.0006,
+ "loss": 7.742180347442627,
+ "step": 2932
+ },
+ {
+ "epoch": 40.74061135371179,
+ "grad_norm": 0.03358631581068039,
+ "learning_rate": 0.0006,
+ "loss": 7.714834213256836,
+ "step": 2933
+ },
+ {
+ "epoch": 40.75458515283843,
+ "grad_norm": 0.02845790982246399,
+ "learning_rate": 0.0006,
+ "loss": 7.7021026611328125,
+ "step": 2934
+ },
+ {
+ "epoch": 40.76855895196506,
+ "grad_norm": 0.03012419492006302,
+ "learning_rate": 0.0006,
+ "loss": 7.712377548217773,
+ "step": 2935
+ },
+ {
+ "epoch": 40.7825327510917,
+ "grad_norm": 0.032981179654598236,
+ "learning_rate": 0.0006,
+ "loss": 7.666385650634766,
+ "step": 2936
+ },
+ {
+ "epoch": 40.79650655021834,
+ "grad_norm": 0.03357074037194252,
+ "learning_rate": 0.0006,
+ "loss": 7.649097919464111,
+ "step": 2937
+ },
+ {
+ "epoch": 40.81048034934498,
+ "grad_norm": 0.030914708971977234,
+ "learning_rate": 0.0006,
+ "loss": 7.649196624755859,
+ "step": 2938
+ },
+ {
+ "epoch": 40.82445414847162,
+ "grad_norm": 0.026296131312847137,
+ "learning_rate": 0.0006,
+ "loss": 7.629964828491211,
+ "step": 2939
+ },
+ {
+ "epoch": 40.838427947598255,
+ "grad_norm": 0.026677245274186134,
+ "learning_rate": 0.0006,
+ "loss": 7.599973678588867,
+ "step": 2940
+ },
+ {
+ "epoch": 40.852401746724894,
+ "grad_norm": 0.023208219558000565,
+ "learning_rate": 0.0006,
+ "loss": 7.601624488830566,
+ "step": 2941
+ },
+ {
+ "epoch": 40.866375545851525,
+ "grad_norm": 0.020741475746035576,
+ "learning_rate": 0.0006,
+ "loss": 7.603902339935303,
+ "step": 2942
+ },
+ {
+ "epoch": 40.880349344978164,
+ "grad_norm": 0.019221751019358635,
+ "learning_rate": 0.0006,
+ "loss": 7.5535888671875,
+ "step": 2943
+ },
+ {
+ "epoch": 40.8943231441048,
+ "grad_norm": 0.019491951912641525,
+ "learning_rate": 0.0006,
+ "loss": 7.54597282409668,
+ "step": 2944
+ },
+ {
+ "epoch": 40.90829694323144,
+ "grad_norm": 0.019702572375535965,
+ "learning_rate": 0.0006,
+ "loss": 7.537567138671875,
+ "step": 2945
+ },
+ {
+ "epoch": 40.92227074235808,
+ "grad_norm": 0.01768800988793373,
+ "learning_rate": 0.0006,
+ "loss": 7.512800216674805,
+ "step": 2946
+ },
+ {
+ "epoch": 40.93624454148472,
+ "grad_norm": 0.01926138624548912,
+ "learning_rate": 0.0006,
+ "loss": 7.509932518005371,
+ "step": 2947
+ },
+ {
+ "epoch": 40.95021834061135,
+ "grad_norm": 0.019030006602406502,
+ "learning_rate": 0.0006,
+ "loss": 7.497983932495117,
+ "step": 2948
+ },
+ {
+ "epoch": 40.96419213973799,
+ "grad_norm": 0.017313841730356216,
+ "learning_rate": 0.0006,
+ "loss": 7.463518142700195,
+ "step": 2949
+ },
+ {
+ "epoch": 40.97816593886463,
+ "grad_norm": 0.01310762483626604,
+ "learning_rate": 0.0006,
+ "loss": 7.474668502807617,
+ "step": 2950
+ },
+ {
+ "epoch": 40.992139737991266,
+ "grad_norm": 0.013008282519876957,
+ "learning_rate": 0.0006,
+ "loss": 7.45982551574707,
+ "step": 2951
+ },
+ {
+ "epoch": 41.0,
+ "grad_norm": 0.01441363524645567,
+ "learning_rate": 0.0006,
+ "loss": 7.446529388427734,
+ "step": 2952
+ },
+ {
+ "epoch": 41.0,
+ "eval_loss": 7.486120223999023,
+ "eval_runtime": 70.4986,
+ "eval_samples_per_second": 34.639,
+ "eval_steps_per_second": 1.092,
+ "step": 2952
+ },
+ {
+ "epoch": 41.01397379912664,
+ "grad_norm": 0.014061855152249336,
+ "learning_rate": 0.0006,
+ "loss": 7.4692182540893555,
+ "step": 2953
+ },
+ {
+ "epoch": 41.02794759825328,
+ "grad_norm": 0.01587500050663948,
+ "learning_rate": 0.0006,
+ "loss": 7.4182024002075195,
+ "step": 2954
+ },
+ {
+ "epoch": 41.041921397379916,
+ "grad_norm": 0.014783984050154686,
+ "learning_rate": 0.0006,
+ "loss": 7.466778755187988,
+ "step": 2955
+ },
+ {
+ "epoch": 41.05589519650655,
+ "grad_norm": 0.011560854502022266,
+ "learning_rate": 0.0006,
+ "loss": 7.438002586364746,
+ "step": 2956
+ },
+ {
+ "epoch": 41.069868995633186,
+ "grad_norm": 0.00970957800745964,
+ "learning_rate": 0.0006,
+ "loss": 7.442694187164307,
+ "step": 2957
+ },
+ {
+ "epoch": 41.083842794759825,
+ "grad_norm": 0.009633459150791168,
+ "learning_rate": 0.0006,
+ "loss": 7.458884239196777,
+ "step": 2958
+ },
+ {
+ "epoch": 41.09781659388646,
+ "grad_norm": 0.010664350353181362,
+ "learning_rate": 0.0006,
+ "loss": 7.48378324508667,
+ "step": 2959
+ },
+ {
+ "epoch": 41.1117903930131,
+ "grad_norm": 0.011440591886639595,
+ "learning_rate": 0.0006,
+ "loss": 7.474484920501709,
+ "step": 2960
+ },
+ {
+ "epoch": 41.12576419213974,
+ "grad_norm": 0.01150908600538969,
+ "learning_rate": 0.0006,
+ "loss": 7.417902946472168,
+ "step": 2961
+ },
+ {
+ "epoch": 41.13973799126637,
+ "grad_norm": 0.010375697165727615,
+ "learning_rate": 0.0006,
+ "loss": 7.403279781341553,
+ "step": 2962
+ },
+ {
+ "epoch": 41.15371179039301,
+ "grad_norm": 0.009075703099370003,
+ "learning_rate": 0.0006,
+ "loss": 7.447347164154053,
+ "step": 2963
+ },
+ {
+ "epoch": 41.16768558951965,
+ "grad_norm": 0.008892177604138851,
+ "learning_rate": 0.0006,
+ "loss": 7.402681350708008,
+ "step": 2964
+ },
+ {
+ "epoch": 41.18165938864629,
+ "grad_norm": 0.009492084383964539,
+ "learning_rate": 0.0006,
+ "loss": 7.3871355056762695,
+ "step": 2965
+ },
+ {
+ "epoch": 41.19563318777293,
+ "grad_norm": 0.009139460511505604,
+ "learning_rate": 0.0006,
+ "loss": 7.419617176055908,
+ "step": 2966
+ },
+ {
+ "epoch": 41.209606986899566,
+ "grad_norm": 0.007959138602018356,
+ "learning_rate": 0.0006,
+ "loss": 7.374949932098389,
+ "step": 2967
+ },
+ {
+ "epoch": 41.223580786026204,
+ "grad_norm": 0.007571548223495483,
+ "learning_rate": 0.0006,
+ "loss": 7.4205241203308105,
+ "step": 2968
+ },
+ {
+ "epoch": 41.237554585152836,
+ "grad_norm": 0.007885400205850601,
+ "learning_rate": 0.0006,
+ "loss": 7.416543006896973,
+ "step": 2969
+ },
+ {
+ "epoch": 41.251528384279474,
+ "grad_norm": 0.008650158531963825,
+ "learning_rate": 0.0006,
+ "loss": 7.357326984405518,
+ "step": 2970
+ },
+ {
+ "epoch": 41.26550218340611,
+ "grad_norm": 0.007483262103050947,
+ "learning_rate": 0.0006,
+ "loss": 7.410891056060791,
+ "step": 2971
+ },
+ {
+ "epoch": 41.27947598253275,
+ "grad_norm": 0.00869277399033308,
+ "learning_rate": 0.0006,
+ "loss": 7.392127990722656,
+ "step": 2972
+ },
+ {
+ "epoch": 41.29344978165939,
+ "grad_norm": 0.00885078776627779,
+ "learning_rate": 0.0006,
+ "loss": 7.405449867248535,
+ "step": 2973
+ },
+ {
+ "epoch": 41.30742358078603,
+ "grad_norm": 0.00958226341754198,
+ "learning_rate": 0.0006,
+ "loss": 7.399868488311768,
+ "step": 2974
+ },
+ {
+ "epoch": 41.32139737991266,
+ "grad_norm": 0.006991599686443806,
+ "learning_rate": 0.0006,
+ "loss": 7.35369873046875,
+ "step": 2975
+ },
+ {
+ "epoch": 41.3353711790393,
+ "grad_norm": 0.007415142375975847,
+ "learning_rate": 0.0006,
+ "loss": 7.358353137969971,
+ "step": 2976
+ },
+ {
+ "epoch": 41.34934497816594,
+ "grad_norm": 0.007024371065199375,
+ "learning_rate": 0.0006,
+ "loss": 7.4070892333984375,
+ "step": 2977
+ },
+ {
+ "epoch": 41.36331877729258,
+ "grad_norm": 0.007234332151710987,
+ "learning_rate": 0.0006,
+ "loss": 7.379120349884033,
+ "step": 2978
+ },
+ {
+ "epoch": 41.377292576419215,
+ "grad_norm": 0.00798892229795456,
+ "learning_rate": 0.0006,
+ "loss": 7.365455627441406,
+ "step": 2979
+ },
+ {
+ "epoch": 41.391266375545854,
+ "grad_norm": 0.007303398102521896,
+ "learning_rate": 0.0006,
+ "loss": 7.341798782348633,
+ "step": 2980
+ },
+ {
+ "epoch": 41.40524017467249,
+ "grad_norm": 0.007119669578969479,
+ "learning_rate": 0.0006,
+ "loss": 7.363280296325684,
+ "step": 2981
+ },
+ {
+ "epoch": 41.419213973799124,
+ "grad_norm": 0.007457619532942772,
+ "learning_rate": 0.0006,
+ "loss": 7.39388370513916,
+ "step": 2982
+ },
+ {
+ "epoch": 41.43318777292576,
+ "grad_norm": 0.007905688136816025,
+ "learning_rate": 0.0006,
+ "loss": 7.380652904510498,
+ "step": 2983
+ },
+ {
+ "epoch": 41.4471615720524,
+ "grad_norm": 0.006212855689227581,
+ "learning_rate": 0.0006,
+ "loss": 7.398769378662109,
+ "step": 2984
+ },
+ {
+ "epoch": 41.46113537117904,
+ "grad_norm": 0.0070641543716192245,
+ "learning_rate": 0.0006,
+ "loss": 7.3503546714782715,
+ "step": 2985
+ },
+ {
+ "epoch": 41.47510917030568,
+ "grad_norm": 0.00787805300205946,
+ "learning_rate": 0.0006,
+ "loss": 7.347678184509277,
+ "step": 2986
+ },
+ {
+ "epoch": 41.48908296943232,
+ "grad_norm": 0.006760426796972752,
+ "learning_rate": 0.0006,
+ "loss": 7.380006790161133,
+ "step": 2987
+ },
+ {
+ "epoch": 41.50305676855895,
+ "grad_norm": 0.00772385997697711,
+ "learning_rate": 0.0006,
+ "loss": 7.363680362701416,
+ "step": 2988
+ },
+ {
+ "epoch": 41.51703056768559,
+ "grad_norm": 0.007373564876616001,
+ "learning_rate": 0.0006,
+ "loss": 7.333014488220215,
+ "step": 2989
+ },
+ {
+ "epoch": 41.531004366812226,
+ "grad_norm": 0.006717653013765812,
+ "learning_rate": 0.0006,
+ "loss": 7.360076427459717,
+ "step": 2990
+ },
+ {
+ "epoch": 41.544978165938865,
+ "grad_norm": 0.007191413082182407,
+ "learning_rate": 0.0006,
+ "loss": 7.397030830383301,
+ "step": 2991
+ },
+ {
+ "epoch": 41.5589519650655,
+ "grad_norm": 0.007831960916519165,
+ "learning_rate": 0.0006,
+ "loss": 7.369959831237793,
+ "step": 2992
+ },
+ {
+ "epoch": 41.57292576419214,
+ "grad_norm": 0.006947481073439121,
+ "learning_rate": 0.0006,
+ "loss": 7.333254337310791,
+ "step": 2993
+ },
+ {
+ "epoch": 41.58689956331878,
+ "grad_norm": 0.006243606563657522,
+ "learning_rate": 0.0006,
+ "loss": 7.327378273010254,
+ "step": 2994
+ },
+ {
+ "epoch": 41.60087336244541,
+ "grad_norm": 0.006298339460045099,
+ "learning_rate": 0.0006,
+ "loss": 7.323482036590576,
+ "step": 2995
+ },
+ {
+ "epoch": 41.61484716157205,
+ "grad_norm": 0.007587412837892771,
+ "learning_rate": 0.0006,
+ "loss": 7.313789367675781,
+ "step": 2996
+ },
+ {
+ "epoch": 41.62882096069869,
+ "grad_norm": 0.007140746805816889,
+ "learning_rate": 0.0006,
+ "loss": 7.356067657470703,
+ "step": 2997
+ },
+ {
+ "epoch": 41.64279475982533,
+ "grad_norm": 0.008315625600516796,
+ "learning_rate": 0.0006,
+ "loss": 7.3339643478393555,
+ "step": 2998
+ },
+ {
+ "epoch": 41.65676855895197,
+ "grad_norm": 0.006547102238982916,
+ "learning_rate": 0.0006,
+ "loss": 7.342191696166992,
+ "step": 2999
+ },
+ {
+ "epoch": 41.670742358078606,
+ "grad_norm": 0.006068122107535601,
+ "learning_rate": 0.0006,
+ "loss": 7.321015357971191,
+ "step": 3000
+ },
+ {
+ "epoch": 41.68471615720524,
+ "grad_norm": 0.007316991221159697,
+ "learning_rate": 0.0006,
+ "loss": 7.346490859985352,
+ "step": 3001
+ },
+ {
+ "epoch": 41.698689956331876,
+ "grad_norm": 0.0067388564348220825,
+ "learning_rate": 0.0006,
+ "loss": 7.340723991394043,
+ "step": 3002
+ },
+ {
+ "epoch": 41.712663755458514,
+ "grad_norm": 0.006839963141828775,
+ "learning_rate": 0.0006,
+ "loss": 7.329365253448486,
+ "step": 3003
+ },
+ {
+ "epoch": 41.72663755458515,
+ "grad_norm": 0.00805041566491127,
+ "learning_rate": 0.0006,
+ "loss": 7.298934459686279,
+ "step": 3004
+ },
+ {
+ "epoch": 41.74061135371179,
+ "grad_norm": 0.007197503466159105,
+ "learning_rate": 0.0006,
+ "loss": 7.331496238708496,
+ "step": 3005
+ },
+ {
+ "epoch": 41.75458515283843,
+ "grad_norm": 0.007220805156975985,
+ "learning_rate": 0.0006,
+ "loss": 7.317695140838623,
+ "step": 3006
+ },
+ {
+ "epoch": 41.76855895196506,
+ "grad_norm": 0.00874276738613844,
+ "learning_rate": 0.0006,
+ "loss": 7.272614479064941,
+ "step": 3007
+ },
+ {
+ "epoch": 41.7825327510917,
+ "grad_norm": 0.008386565372347832,
+ "learning_rate": 0.0006,
+ "loss": 7.295244216918945,
+ "step": 3008
+ },
+ {
+ "epoch": 41.79650655021834,
+ "grad_norm": 0.0077730282209813595,
+ "learning_rate": 0.0006,
+ "loss": 7.3395490646362305,
+ "step": 3009
+ },
+ {
+ "epoch": 41.81048034934498,
+ "grad_norm": 0.008441480807960033,
+ "learning_rate": 0.0006,
+ "loss": 7.319921493530273,
+ "step": 3010
+ },
+ {
+ "epoch": 41.82445414847162,
+ "grad_norm": 0.007398269604891539,
+ "learning_rate": 0.0006,
+ "loss": 7.3029465675354,
+ "step": 3011
+ },
+ {
+ "epoch": 41.838427947598255,
+ "grad_norm": 0.006878677289932966,
+ "learning_rate": 0.0006,
+ "loss": 7.318467140197754,
+ "step": 3012
+ },
+ {
+ "epoch": 41.852401746724894,
+ "grad_norm": 0.007616598624736071,
+ "learning_rate": 0.0006,
+ "loss": 7.3112077713012695,
+ "step": 3013
+ },
+ {
+ "epoch": 41.866375545851525,
+ "grad_norm": 0.011629699729382992,
+ "learning_rate": 0.0006,
+ "loss": 7.32523250579834,
+ "step": 3014
+ },
+ {
+ "epoch": 41.880349344978164,
+ "grad_norm": 0.011011249385774136,
+ "learning_rate": 0.0006,
+ "loss": 7.287247657775879,
+ "step": 3015
+ },
+ {
+ "epoch": 41.8943231441048,
+ "grad_norm": 0.010208295658230782,
+ "learning_rate": 0.0006,
+ "loss": 7.292117118835449,
+ "step": 3016
+ },
+ {
+ "epoch": 41.90829694323144,
+ "grad_norm": 0.008651095442473888,
+ "learning_rate": 0.0006,
+ "loss": 7.302201271057129,
+ "step": 3017
+ },
+ {
+ "epoch": 41.92227074235808,
+ "grad_norm": 0.006692920345813036,
+ "learning_rate": 0.0006,
+ "loss": 7.319568634033203,
+ "step": 3018
+ },
+ {
+ "epoch": 41.93624454148472,
+ "grad_norm": 0.010033348575234413,
+ "learning_rate": 0.0006,
+ "loss": 7.2738566398620605,
+ "step": 3019
+ },
+ {
+ "epoch": 41.95021834061135,
+ "grad_norm": 0.011790832504630089,
+ "learning_rate": 0.0006,
+ "loss": 7.298540115356445,
+ "step": 3020
+ },
+ {
+ "epoch": 41.96419213973799,
+ "grad_norm": 0.018622662872076035,
+ "learning_rate": 0.0006,
+ "loss": 7.267182350158691,
+ "step": 3021
+ },
+ {
+ "epoch": 41.97816593886463,
+ "grad_norm": 0.02520257607102394,
+ "learning_rate": 0.0006,
+ "loss": 7.261566162109375,
+ "step": 3022
+ },
+ {
+ "epoch": 41.992139737991266,
+ "grad_norm": 0.036332644522190094,
+ "learning_rate": 0.0006,
+ "loss": 7.270719051361084,
+ "step": 3023
+ },
+ {
+ "epoch": 42.0,
+ "grad_norm": 0.04717305675148964,
+ "learning_rate": 0.0006,
+ "loss": 7.293332099914551,
+ "step": 3024
+ },
+ {
+ "epoch": 42.0,
+ "eval_loss": 7.354280471801758,
+ "eval_runtime": 56.4353,
+ "eval_samples_per_second": 43.271,
+ "eval_steps_per_second": 1.364,
+ "step": 3024
+ },
+ {
+ "epoch": 42.01397379912664,
+ "grad_norm": 0.06661626696586609,
+ "learning_rate": 0.0006,
+ "loss": 7.338189601898193,
+ "step": 3025
+ },
+ {
+ "epoch": 42.02794759825328,
+ "grad_norm": 0.05452054366469383,
+ "learning_rate": 0.0006,
+ "loss": 7.316764831542969,
+ "step": 3026
+ },
+ {
+ "epoch": 42.041921397379916,
+ "grad_norm": 0.025638775900006294,
+ "learning_rate": 0.0006,
+ "loss": 7.34204626083374,
+ "step": 3027
+ },
+ {
+ "epoch": 42.05589519650655,
+ "grad_norm": 0.022398799657821655,
+ "learning_rate": 0.0006,
+ "loss": 7.262149333953857,
+ "step": 3028
+ },
+ {
+ "epoch": 42.069868995633186,
+ "grad_norm": 0.03713499382138252,
+ "learning_rate": 0.0006,
+ "loss": 7.27695369720459,
+ "step": 3029
+ },
+ {
+ "epoch": 42.083842794759825,
+ "grad_norm": 0.014949247241020203,
+ "learning_rate": 0.0006,
+ "loss": 7.296508312225342,
+ "step": 3030
+ },
+ {
+ "epoch": 42.09781659388646,
+ "grad_norm": 0.02703995071351528,
+ "learning_rate": 0.0006,
+ "loss": 7.258018493652344,
+ "step": 3031
+ },
+ {
+ "epoch": 42.1117903930131,
+ "grad_norm": 0.020352816209197044,
+ "learning_rate": 0.0006,
+ "loss": 7.241450309753418,
+ "step": 3032
+ },
+ {
+ "epoch": 42.12576419213974,
+ "grad_norm": 0.014578479342162609,
+ "learning_rate": 0.0006,
+ "loss": 7.257323265075684,
+ "step": 3033
+ },
+ {
+ "epoch": 42.13973799126637,
+ "grad_norm": 0.026053044945001602,
+ "learning_rate": 0.0006,
+ "loss": 7.254680633544922,
+ "step": 3034
+ },
+ {
+ "epoch": 42.15371179039301,
+ "grad_norm": 0.015120275318622589,
+ "learning_rate": 0.0006,
+ "loss": 7.261074066162109,
+ "step": 3035
+ },
+ {
+ "epoch": 42.16768558951965,
+ "grad_norm": 0.016822177916765213,
+ "learning_rate": 0.0006,
+ "loss": 7.279064178466797,
+ "step": 3036
+ },
+ {
+ "epoch": 42.18165938864629,
+ "grad_norm": 0.016367588192224503,
+ "learning_rate": 0.0006,
+ "loss": 7.251429557800293,
+ "step": 3037
+ },
+ {
+ "epoch": 42.19563318777293,
+ "grad_norm": 0.014350418001413345,
+ "learning_rate": 0.0006,
+ "loss": 7.259264945983887,
+ "step": 3038
+ },
+ {
+ "epoch": 42.209606986899566,
+ "grad_norm": 0.017078228294849396,
+ "learning_rate": 0.0006,
+ "loss": 7.227105140686035,
+ "step": 3039
+ },
+ {
+ "epoch": 42.223580786026204,
+ "grad_norm": 0.011759317480027676,
+ "learning_rate": 0.0006,
+ "loss": 7.275888442993164,
+ "step": 3040
+ },
+ {
+ "epoch": 42.237554585152836,
+ "grad_norm": 0.013019340112805367,
+ "learning_rate": 0.0006,
+ "loss": 7.2517008781433105,
+ "step": 3041
+ },
+ {
+ "epoch": 42.251528384279474,
+ "grad_norm": 0.01178441196680069,
+ "learning_rate": 0.0006,
+ "loss": 7.23576545715332,
+ "step": 3042
+ },
+ {
+ "epoch": 42.26550218340611,
+ "grad_norm": 0.014518891461193562,
+ "learning_rate": 0.0006,
+ "loss": 7.253895282745361,
+ "step": 3043
+ },
+ {
+ "epoch": 42.27947598253275,
+ "grad_norm": 0.018159950152039528,
+ "learning_rate": 0.0006,
+ "loss": 7.278058052062988,
+ "step": 3044
+ },
+ {
+ "epoch": 42.29344978165939,
+ "grad_norm": 0.007587796077132225,
+ "learning_rate": 0.0006,
+ "loss": 7.278578758239746,
+ "step": 3045
+ },
+ {
+ "epoch": 42.30742358078603,
+ "grad_norm": 0.015994979068636894,
+ "learning_rate": 0.0006,
+ "loss": 7.239841461181641,
+ "step": 3046
+ },
+ {
+ "epoch": 42.32139737991266,
+ "grad_norm": 0.017830168828368187,
+ "learning_rate": 0.0006,
+ "loss": 7.2443976402282715,
+ "step": 3047
+ },
+ {
+ "epoch": 42.3353711790393,
+ "grad_norm": 0.008738759905099869,
+ "learning_rate": 0.0006,
+ "loss": 7.232264518737793,
+ "step": 3048
+ },
+ {
+ "epoch": 42.34934497816594,
+ "grad_norm": 0.0175449438393116,
+ "learning_rate": 0.0006,
+ "loss": 7.228000640869141,
+ "step": 3049
+ },
+ {
+ "epoch": 42.36331877729258,
+ "grad_norm": 0.019611820578575134,
+ "learning_rate": 0.0006,
+ "loss": 7.234999179840088,
+ "step": 3050
+ },
+ {
+ "epoch": 42.377292576419215,
+ "grad_norm": 0.015302863903343678,
+ "learning_rate": 0.0006,
+ "loss": 7.242547988891602,
+ "step": 3051
+ },
+ {
+ "epoch": 42.391266375545854,
+ "grad_norm": 0.025919731706380844,
+ "learning_rate": 0.0006,
+ "loss": 7.216652870178223,
+ "step": 3052
+ },
+ {
+ "epoch": 42.40524017467249,
+ "grad_norm": 0.034587468951940536,
+ "learning_rate": 0.0006,
+ "loss": 7.245336532592773,
+ "step": 3053
+ },
+ {
+ "epoch": 42.419213973799124,
+ "grad_norm": 0.019198840484023094,
+ "learning_rate": 0.0006,
+ "loss": 7.256758689880371,
+ "step": 3054
+ },
+ {
+ "epoch": 42.43318777292576,
+ "grad_norm": 0.012316389009356499,
+ "learning_rate": 0.0006,
+ "loss": 7.209102630615234,
+ "step": 3055
+ },
+ {
+ "epoch": 42.4471615720524,
+ "grad_norm": 0.023512162268161774,
+ "learning_rate": 0.0006,
+ "loss": 7.250742435455322,
+ "step": 3056
+ },
+ {
+ "epoch": 42.46113537117904,
+ "grad_norm": 0.025457851588726044,
+ "learning_rate": 0.0006,
+ "loss": 7.263757705688477,
+ "step": 3057
+ },
+ {
+ "epoch": 42.47510917030568,
+ "grad_norm": 0.02192617766559124,
+ "learning_rate": 0.0006,
+ "loss": 7.246977806091309,
+ "step": 3058
+ },
+ {
+ "epoch": 42.48908296943232,
+ "grad_norm": 0.02541281282901764,
+ "learning_rate": 0.0006,
+ "loss": 7.232361316680908,
+ "step": 3059
+ },
+ {
+ "epoch": 42.50305676855895,
+ "grad_norm": 0.036353763192892075,
+ "learning_rate": 0.0006,
+ "loss": 7.242556571960449,
+ "step": 3060
+ },
+ {
+ "epoch": 42.51703056768559,
+ "grad_norm": 0.04555872827768326,
+ "learning_rate": 0.0006,
+ "loss": 7.228940963745117,
+ "step": 3061
+ },
+ {
+ "epoch": 42.531004366812226,
+ "grad_norm": 0.04563378915190697,
+ "learning_rate": 0.0006,
+ "loss": 7.235276222229004,
+ "step": 3062
+ },
+ {
+ "epoch": 42.544978165938865,
+ "grad_norm": 0.04275760427117348,
+ "learning_rate": 0.0006,
+ "loss": 7.238895416259766,
+ "step": 3063
+ },
+ {
+ "epoch": 42.5589519650655,
+ "grad_norm": 0.04535885155200958,
+ "learning_rate": 0.0006,
+ "loss": 7.205336570739746,
+ "step": 3064
+ },
+ {
+ "epoch": 42.57292576419214,
+ "grad_norm": 0.015534617006778717,
+ "learning_rate": 0.0006,
+ "loss": 7.238393783569336,
+ "step": 3065
+ },
+ {
+ "epoch": 42.58689956331878,
+ "grad_norm": 0.03286299481987953,
+ "learning_rate": 0.0006,
+ "loss": 7.216453552246094,
+ "step": 3066
+ },
+ {
+ "epoch": 42.60087336244541,
+ "grad_norm": 0.025453168898820877,
+ "learning_rate": 0.0006,
+ "loss": 7.203750133514404,
+ "step": 3067
+ },
+ {
+ "epoch": 42.61484716157205,
+ "grad_norm": 0.021501649171113968,
+ "learning_rate": 0.0006,
+ "loss": 7.241621971130371,
+ "step": 3068
+ },
+ {
+ "epoch": 42.62882096069869,
+ "grad_norm": 0.02743634767830372,
+ "learning_rate": 0.0006,
+ "loss": 7.22076940536499,
+ "step": 3069
+ },
+ {
+ "epoch": 42.64279475982533,
+ "grad_norm": 0.020890483632683754,
+ "learning_rate": 0.0006,
+ "loss": 7.218412399291992,
+ "step": 3070
+ },
+ {
+ "epoch": 42.65676855895197,
+ "grad_norm": 0.022499850019812584,
+ "learning_rate": 0.0006,
+ "loss": 7.217696189880371,
+ "step": 3071
+ },
+ {
+ "epoch": 42.670742358078606,
+ "grad_norm": 0.01653696969151497,
+ "learning_rate": 0.0006,
+ "loss": 7.156417369842529,
+ "step": 3072
+ },
+ {
+ "epoch": 42.68471615720524,
+ "grad_norm": 0.01834225095808506,
+ "learning_rate": 0.0006,
+ "loss": 7.211247444152832,
+ "step": 3073
+ },
+ {
+ "epoch": 42.698689956331876,
+ "grad_norm": 0.020137229934334755,
+ "learning_rate": 0.0006,
+ "loss": 7.232499599456787,
+ "step": 3074
+ },
+ {
+ "epoch": 42.712663755458514,
+ "grad_norm": 0.012671459466218948,
+ "learning_rate": 0.0006,
+ "loss": 7.172441482543945,
+ "step": 3075
+ },
+ {
+ "epoch": 42.72663755458515,
+ "grad_norm": 0.016803130507469177,
+ "learning_rate": 0.0006,
+ "loss": 7.152103424072266,
+ "step": 3076
+ },
+ {
+ "epoch": 42.74061135371179,
+ "grad_norm": 0.01623712107539177,
+ "learning_rate": 0.0006,
+ "loss": 7.187032699584961,
+ "step": 3077
+ },
+ {
+ "epoch": 42.75458515283843,
+ "grad_norm": 0.015938356518745422,
+ "learning_rate": 0.0006,
+ "loss": 7.171140670776367,
+ "step": 3078
+ },
+ {
+ "epoch": 42.76855895196506,
+ "grad_norm": 0.02580736204981804,
+ "learning_rate": 0.0006,
+ "loss": 7.204472541809082,
+ "step": 3079
+ },
+ {
+ "epoch": 42.7825327510917,
+ "grad_norm": 0.023372117429971695,
+ "learning_rate": 0.0006,
+ "loss": 7.165554523468018,
+ "step": 3080
+ },
+ {
+ "epoch": 42.79650655021834,
+ "grad_norm": 0.04440586268901825,
+ "learning_rate": 0.0006,
+ "loss": 7.169431209564209,
+ "step": 3081
+ },
+ {
+ "epoch": 42.81048034934498,
+ "grad_norm": 0.06797406077384949,
+ "learning_rate": 0.0006,
+ "loss": 7.168193340301514,
+ "step": 3082
+ },
+ {
+ "epoch": 42.82445414847162,
+ "grad_norm": 0.0914321169257164,
+ "learning_rate": 0.0006,
+ "loss": 7.204671859741211,
+ "step": 3083
+ },
+ {
+ "epoch": 42.838427947598255,
+ "grad_norm": 0.0506940595805645,
+ "learning_rate": 0.0006,
+ "loss": 7.15949821472168,
+ "step": 3084
+ },
+ {
+ "epoch": 42.852401746724894,
+ "grad_norm": 0.04819890484213829,
+ "learning_rate": 0.0006,
+ "loss": 7.192837238311768,
+ "step": 3085
+ },
+ {
+ "epoch": 42.866375545851525,
+ "grad_norm": 0.040044233202934265,
+ "learning_rate": 0.0006,
+ "loss": 7.1600165367126465,
+ "step": 3086
+ },
+ {
+ "epoch": 42.880349344978164,
+ "grad_norm": 0.02573299966752529,
+ "learning_rate": 0.0006,
+ "loss": 7.153427600860596,
+ "step": 3087
+ },
+ {
+ "epoch": 42.8943231441048,
+ "grad_norm": 0.046942535787820816,
+ "learning_rate": 0.0006,
+ "loss": 7.16767692565918,
+ "step": 3088
+ },
+ {
+ "epoch": 42.90829694323144,
+ "grad_norm": 0.02683161571621895,
+ "learning_rate": 0.0006,
+ "loss": 7.17144250869751,
+ "step": 3089
+ },
+ {
+ "epoch": 42.92227074235808,
+ "grad_norm": 0.05363674834370613,
+ "learning_rate": 0.0006,
+ "loss": 7.120128631591797,
+ "step": 3090
+ },
+ {
+ "epoch": 42.93624454148472,
+ "grad_norm": 0.023756247013807297,
+ "learning_rate": 0.0006,
+ "loss": 7.169703006744385,
+ "step": 3091
+ },
+ {
+ "epoch": 42.95021834061135,
+ "grad_norm": 0.04216783866286278,
+ "learning_rate": 0.0006,
+ "loss": 7.1206865310668945,
+ "step": 3092
+ },
+ {
+ "epoch": 42.96419213973799,
+ "grad_norm": 0.02219565585255623,
+ "learning_rate": 0.0006,
+ "loss": 7.157215118408203,
+ "step": 3093
+ },
+ {
+ "epoch": 42.97816593886463,
+ "grad_norm": 0.03527127206325531,
+ "learning_rate": 0.0006,
+ "loss": 7.166672706604004,
+ "step": 3094
+ },
+ {
+ "epoch": 42.992139737991266,
+ "grad_norm": 0.017748817801475525,
+ "learning_rate": 0.0006,
+ "loss": 7.154747009277344,
+ "step": 3095
+ },
+ {
+ "epoch": 43.0,
+ "grad_norm": 0.033165838569402695,
+ "learning_rate": 0.0006,
+ "loss": 7.150866508483887,
+ "step": 3096
+ },
+ {
+ "epoch": 43.0,
+ "eval_loss": 7.144403457641602,
+ "eval_runtime": 56.8853,
+ "eval_samples_per_second": 42.928,
+ "eval_steps_per_second": 1.354,
+ "step": 3096
+ },
+ {
+ "epoch": 43.01397379912664,
+ "grad_norm": 0.015199829824268818,
+ "learning_rate": 0.0006,
+ "loss": 7.122735023498535,
+ "step": 3097
+ },
+ {
+ "epoch": 43.02794759825328,
+ "grad_norm": 0.03220256045460701,
+ "learning_rate": 0.0006,
+ "loss": 7.143553733825684,
+ "step": 3098
+ },
+ {
+ "epoch": 43.041921397379916,
+ "grad_norm": 0.014687920920550823,
+ "learning_rate": 0.0006,
+ "loss": 7.080960750579834,
+ "step": 3099
+ },
+ {
+ "epoch": 43.05589519650655,
+ "grad_norm": 0.023522932082414627,
+ "learning_rate": 0.0006,
+ "loss": 7.137356758117676,
+ "step": 3100
+ },
+ {
+ "epoch": 43.069868995633186,
+ "grad_norm": 0.014808216132223606,
+ "learning_rate": 0.0006,
+ "loss": 7.09143590927124,
+ "step": 3101
+ },
+ {
+ "epoch": 43.083842794759825,
+ "grad_norm": 0.029339246451854706,
+ "learning_rate": 0.0006,
+ "loss": 7.122198104858398,
+ "step": 3102
+ },
+ {
+ "epoch": 43.09781659388646,
+ "grad_norm": 0.022953078150749207,
+ "learning_rate": 0.0006,
+ "loss": 7.092505931854248,
+ "step": 3103
+ },
+ {
+ "epoch": 43.1117903930131,
+ "grad_norm": 0.04590052738785744,
+ "learning_rate": 0.0006,
+ "loss": 7.113100051879883,
+ "step": 3104
+ },
+ {
+ "epoch": 43.12576419213974,
+ "grad_norm": 0.0524442233145237,
+ "learning_rate": 0.0006,
+ "loss": 7.147022247314453,
+ "step": 3105
+ },
+ {
+ "epoch": 43.13973799126637,
+ "grad_norm": 0.0315794013440609,
+ "learning_rate": 0.0006,
+ "loss": 7.094045162200928,
+ "step": 3106
+ },
+ {
+ "epoch": 43.15371179039301,
+ "grad_norm": 0.05112722888588905,
+ "learning_rate": 0.0006,
+ "loss": 7.054388523101807,
+ "step": 3107
+ },
+ {
+ "epoch": 43.16768558951965,
+ "grad_norm": 0.03146776929497719,
+ "learning_rate": 0.0006,
+ "loss": 7.116674423217773,
+ "step": 3108
+ },
+ {
+ "epoch": 43.18165938864629,
+ "grad_norm": 0.03391246497631073,
+ "learning_rate": 0.0006,
+ "loss": 7.150345802307129,
+ "step": 3109
+ },
+ {
+ "epoch": 43.19563318777293,
+ "grad_norm": 0.026957903057336807,
+ "learning_rate": 0.0006,
+ "loss": 7.058531284332275,
+ "step": 3110
+ },
+ {
+ "epoch": 43.209606986899566,
+ "grad_norm": 0.041532568633556366,
+ "learning_rate": 0.0006,
+ "loss": 7.085498809814453,
+ "step": 3111
+ },
+ {
+ "epoch": 43.223580786026204,
+ "grad_norm": 0.020001588389277458,
+ "learning_rate": 0.0006,
+ "loss": 7.131487846374512,
+ "step": 3112
+ },
+ {
+ "epoch": 43.237554585152836,
+ "grad_norm": 0.02268712781369686,
+ "learning_rate": 0.0006,
+ "loss": 7.0790863037109375,
+ "step": 3113
+ },
+ {
+ "epoch": 43.251528384279474,
+ "grad_norm": 0.023363925516605377,
+ "learning_rate": 0.0006,
+ "loss": 7.067816734313965,
+ "step": 3114
+ },
+ {
+ "epoch": 43.26550218340611,
+ "grad_norm": 0.019367508590221405,
+ "learning_rate": 0.0006,
+ "loss": 7.042041778564453,
+ "step": 3115
+ },
+ {
+ "epoch": 43.27947598253275,
+ "grad_norm": 0.02745090238749981,
+ "learning_rate": 0.0006,
+ "loss": 7.064733028411865,
+ "step": 3116
+ },
+ {
+ "epoch": 43.29344978165939,
+ "grad_norm": 0.02543802745640278,
+ "learning_rate": 0.0006,
+ "loss": 7.058948040008545,
+ "step": 3117
+ },
+ {
+ "epoch": 43.30742358078603,
+ "grad_norm": 0.018874092027544975,
+ "learning_rate": 0.0006,
+ "loss": 7.009946346282959,
+ "step": 3118
+ },
+ {
+ "epoch": 43.32139737991266,
+ "grad_norm": 0.024662822484970093,
+ "learning_rate": 0.0006,
+ "loss": 7.015913963317871,
+ "step": 3119
+ },
+ {
+ "epoch": 43.3353711790393,
+ "grad_norm": 0.02441416308283806,
+ "learning_rate": 0.0006,
+ "loss": 7.061622142791748,
+ "step": 3120
+ },
+ {
+ "epoch": 43.34934497816594,
+ "grad_norm": 0.031887322664260864,
+ "learning_rate": 0.0006,
+ "loss": 7.046883583068848,
+ "step": 3121
+ },
+ {
+ "epoch": 43.36331877729258,
+ "grad_norm": 0.021930739283561707,
+ "learning_rate": 0.0006,
+ "loss": 7.0309624671936035,
+ "step": 3122
+ },
+ {
+ "epoch": 43.377292576419215,
+ "grad_norm": 0.02722783200442791,
+ "learning_rate": 0.0006,
+ "loss": 7.072924613952637,
+ "step": 3123
+ },
+ {
+ "epoch": 43.391266375545854,
+ "grad_norm": 0.01483179721981287,
+ "learning_rate": 0.0006,
+ "loss": 7.028624534606934,
+ "step": 3124
+ },
+ {
+ "epoch": 43.40524017467249,
+ "grad_norm": 0.024793945252895355,
+ "learning_rate": 0.0006,
+ "loss": 7.001623153686523,
+ "step": 3125
+ },
+ {
+ "epoch": 43.419213973799124,
+ "grad_norm": 0.020148713141679764,
+ "learning_rate": 0.0006,
+ "loss": 6.9938812255859375,
+ "step": 3126
+ },
+ {
+ "epoch": 43.43318777292576,
+ "grad_norm": 0.02413850650191307,
+ "learning_rate": 0.0006,
+ "loss": 7.001243591308594,
+ "step": 3127
+ },
+ {
+ "epoch": 43.4471615720524,
+ "grad_norm": 0.028549203649163246,
+ "learning_rate": 0.0006,
+ "loss": 6.987671375274658,
+ "step": 3128
+ },
+ {
+ "epoch": 43.46113537117904,
+ "grad_norm": 0.025988785549998283,
+ "learning_rate": 0.0006,
+ "loss": 6.9551167488098145,
+ "step": 3129
+ },
+ {
+ "epoch": 43.47510917030568,
+ "grad_norm": 0.027140729129314423,
+ "learning_rate": 0.0006,
+ "loss": 6.9547810554504395,
+ "step": 3130
+ },
+ {
+ "epoch": 43.48908296943232,
+ "grad_norm": 0.030130082741379738,
+ "learning_rate": 0.0006,
+ "loss": 6.993186950683594,
+ "step": 3131
+ },
+ {
+ "epoch": 43.50305676855895,
+ "grad_norm": 0.025996994227170944,
+ "learning_rate": 0.0006,
+ "loss": 6.984624862670898,
+ "step": 3132
+ },
+ {
+ "epoch": 43.51703056768559,
+ "grad_norm": 0.02423008903861046,
+ "learning_rate": 0.0006,
+ "loss": 6.98813009262085,
+ "step": 3133
+ },
+ {
+ "epoch": 43.531004366812226,
+ "grad_norm": 0.029665837064385414,
+ "learning_rate": 0.0006,
+ "loss": 6.986610412597656,
+ "step": 3134
+ },
+ {
+ "epoch": 43.544978165938865,
+ "grad_norm": 0.028768595308065414,
+ "learning_rate": 0.0006,
+ "loss": 6.898797035217285,
+ "step": 3135
+ },
+ {
+ "epoch": 43.5589519650655,
+ "grad_norm": 0.03111208975315094,
+ "learning_rate": 0.0006,
+ "loss": 6.923741817474365,
+ "step": 3136
+ },
+ {
+ "epoch": 43.57292576419214,
+ "grad_norm": 0.0368187390267849,
+ "learning_rate": 0.0006,
+ "loss": 6.994343280792236,
+ "step": 3137
+ },
+ {
+ "epoch": 43.58689956331878,
+ "grad_norm": 0.04076763987541199,
+ "learning_rate": 0.0006,
+ "loss": 6.909458160400391,
+ "step": 3138
+ },
+ {
+ "epoch": 43.60087336244541,
+ "grad_norm": 0.03281061723828316,
+ "learning_rate": 0.0006,
+ "loss": 6.924851417541504,
+ "step": 3139
+ },
+ {
+ "epoch": 43.61484716157205,
+ "grad_norm": 0.025486988946795464,
+ "learning_rate": 0.0006,
+ "loss": 6.962926864624023,
+ "step": 3140
+ },
+ {
+ "epoch": 43.62882096069869,
+ "grad_norm": 0.04583485797047615,
+ "learning_rate": 0.0006,
+ "loss": 6.876555442810059,
+ "step": 3141
+ },
+ {
+ "epoch": 43.64279475982533,
+ "grad_norm": 0.05460785701870918,
+ "learning_rate": 0.0006,
+ "loss": 6.952932357788086,
+ "step": 3142
+ },
+ {
+ "epoch": 43.65676855895197,
+ "grad_norm": 0.040917906910181046,
+ "learning_rate": 0.0006,
+ "loss": 6.9624924659729,
+ "step": 3143
+ },
+ {
+ "epoch": 43.670742358078606,
+ "grad_norm": 0.04009668901562691,
+ "learning_rate": 0.0006,
+ "loss": 6.94033145904541,
+ "step": 3144
+ },
+ {
+ "epoch": 43.68471615720524,
+ "grad_norm": 0.04500625655055046,
+ "learning_rate": 0.0006,
+ "loss": 6.961839199066162,
+ "step": 3145
+ },
+ {
+ "epoch": 43.698689956331876,
+ "grad_norm": 0.03182258829474449,
+ "learning_rate": 0.0006,
+ "loss": 6.930580139160156,
+ "step": 3146
+ },
+ {
+ "epoch": 43.712663755458514,
+ "grad_norm": 0.053621504455804825,
+ "learning_rate": 0.0006,
+ "loss": 6.950235366821289,
+ "step": 3147
+ },
+ {
+ "epoch": 43.72663755458515,
+ "grad_norm": 0.031919289380311966,
+ "learning_rate": 0.0006,
+ "loss": 6.895111560821533,
+ "step": 3148
+ },
+ {
+ "epoch": 43.74061135371179,
+ "grad_norm": 0.042713891714811325,
+ "learning_rate": 0.0006,
+ "loss": 6.940885543823242,
+ "step": 3149
+ },
+ {
+ "epoch": 43.75458515283843,
+ "grad_norm": 0.02385653182864189,
+ "learning_rate": 0.0006,
+ "loss": 6.948801517486572,
+ "step": 3150
+ },
+ {
+ "epoch": 43.76855895196506,
+ "grad_norm": 0.04100663587450981,
+ "learning_rate": 0.0006,
+ "loss": 6.970411777496338,
+ "step": 3151
+ },
+ {
+ "epoch": 43.7825327510917,
+ "grad_norm": 0.02729583904147148,
+ "learning_rate": 0.0006,
+ "loss": 6.899379253387451,
+ "step": 3152
+ },
+ {
+ "epoch": 43.79650655021834,
+ "grad_norm": 0.03128620237112045,
+ "learning_rate": 0.0006,
+ "loss": 6.931319236755371,
+ "step": 3153
+ },
+ {
+ "epoch": 43.81048034934498,
+ "grad_norm": 0.025488659739494324,
+ "learning_rate": 0.0006,
+ "loss": 6.8486409187316895,
+ "step": 3154
+ },
+ {
+ "epoch": 43.82445414847162,
+ "grad_norm": 0.03679376840591431,
+ "learning_rate": 0.0006,
+ "loss": 6.840811252593994,
+ "step": 3155
+ },
+ {
+ "epoch": 43.838427947598255,
+ "grad_norm": 0.020578034222126007,
+ "learning_rate": 0.0006,
+ "loss": 6.9088616371154785,
+ "step": 3156
+ },
+ {
+ "epoch": 43.852401746724894,
+ "grad_norm": 0.02935841865837574,
+ "learning_rate": 0.0006,
+ "loss": 6.806676864624023,
+ "step": 3157
+ },
+ {
+ "epoch": 43.866375545851525,
+ "grad_norm": 0.023319244384765625,
+ "learning_rate": 0.0006,
+ "loss": 6.878457546234131,
+ "step": 3158
+ },
+ {
+ "epoch": 43.880349344978164,
+ "grad_norm": 0.02779589593410492,
+ "learning_rate": 0.0006,
+ "loss": 6.822447776794434,
+ "step": 3159
+ },
+ {
+ "epoch": 43.8943231441048,
+ "grad_norm": 0.01879825070500374,
+ "learning_rate": 0.0006,
+ "loss": 6.854121208190918,
+ "step": 3160
+ },
+ {
+ "epoch": 43.90829694323144,
+ "grad_norm": 0.023286139592528343,
+ "learning_rate": 0.0006,
+ "loss": 6.836827278137207,
+ "step": 3161
+ },
+ {
+ "epoch": 43.92227074235808,
+ "grad_norm": 0.030358297750353813,
+ "learning_rate": 0.0006,
+ "loss": 6.860037803649902,
+ "step": 3162
+ },
+ {
+ "epoch": 43.93624454148472,
+ "grad_norm": 0.029621105641126633,
+ "learning_rate": 0.0006,
+ "loss": 6.884025573730469,
+ "step": 3163
+ },
+ {
+ "epoch": 43.95021834061135,
+ "grad_norm": 0.029829824343323708,
+ "learning_rate": 0.0006,
+ "loss": 6.8702392578125,
+ "step": 3164
+ },
+ {
+ "epoch": 43.96419213973799,
+ "grad_norm": 0.036892473697662354,
+ "learning_rate": 0.0006,
+ "loss": 6.806652069091797,
+ "step": 3165
+ },
+ {
+ "epoch": 43.97816593886463,
+ "grad_norm": 0.03206709027290344,
+ "learning_rate": 0.0006,
+ "loss": 6.828495979309082,
+ "step": 3166
+ },
+ {
+ "epoch": 43.992139737991266,
+ "grad_norm": 0.02744000405073166,
+ "learning_rate": 0.0006,
+ "loss": 6.832002639770508,
+ "step": 3167
+ },
+ {
+ "epoch": 44.0,
+ "grad_norm": 0.02205660194158554,
+ "learning_rate": 0.0006,
+ "loss": 6.898828506469727,
+ "step": 3168
+ },
+ {
+ "epoch": 44.0,
+ "eval_loss": 6.8150553703308105,
+ "eval_runtime": 56.784,
+ "eval_samples_per_second": 43.005,
+ "eval_steps_per_second": 1.356,
+ "step": 3168
+ },
+ {
+ "epoch": 44.01397379912664,
+ "grad_norm": 0.030421260744333267,
+ "learning_rate": 0.0006,
+ "loss": 6.775057792663574,
+ "step": 3169
+ },
+ {
+ "epoch": 44.02794759825328,
+ "grad_norm": 0.0465611033141613,
+ "learning_rate": 0.0006,
+ "loss": 6.849390506744385,
+ "step": 3170
+ },
+ {
+ "epoch": 44.041921397379916,
+ "grad_norm": 0.05977972596883774,
+ "learning_rate": 0.0006,
+ "loss": 6.725576400756836,
+ "step": 3171
+ },
+ {
+ "epoch": 44.05589519650655,
+ "grad_norm": 0.05432697385549545,
+ "learning_rate": 0.0006,
+ "loss": 6.768024444580078,
+ "step": 3172
+ },
+ {
+ "epoch": 44.069868995633186,
+ "grad_norm": 0.04850967228412628,
+ "learning_rate": 0.0006,
+ "loss": 6.799457550048828,
+ "step": 3173
+ },
+ {
+ "epoch": 44.083842794759825,
+ "grad_norm": 0.06201786920428276,
+ "learning_rate": 0.0006,
+ "loss": 6.897234916687012,
+ "step": 3174
+ },
+ {
+ "epoch": 44.09781659388646,
+ "grad_norm": 0.04695591703057289,
+ "learning_rate": 0.0006,
+ "loss": 6.826470375061035,
+ "step": 3175
+ },
+ {
+ "epoch": 44.1117903930131,
+ "grad_norm": 0.03850164636969566,
+ "learning_rate": 0.0006,
+ "loss": 6.835901260375977,
+ "step": 3176
+ },
+ {
+ "epoch": 44.12576419213974,
+ "grad_norm": 0.04250155761837959,
+ "learning_rate": 0.0006,
+ "loss": 6.782310485839844,
+ "step": 3177
+ },
+ {
+ "epoch": 44.13973799126637,
+ "grad_norm": 0.041465479880571365,
+ "learning_rate": 0.0006,
+ "loss": 6.785732746124268,
+ "step": 3178
+ },
+ {
+ "epoch": 44.15371179039301,
+ "grad_norm": 0.02527807280421257,
+ "learning_rate": 0.0006,
+ "loss": 6.78061580657959,
+ "step": 3179
+ },
+ {
+ "epoch": 44.16768558951965,
+ "grad_norm": 0.03660374879837036,
+ "learning_rate": 0.0006,
+ "loss": 6.783992767333984,
+ "step": 3180
+ },
+ {
+ "epoch": 44.18165938864629,
+ "grad_norm": 0.027567964047193527,
+ "learning_rate": 0.0006,
+ "loss": 6.765413284301758,
+ "step": 3181
+ },
+ {
+ "epoch": 44.19563318777293,
+ "grad_norm": 0.024301424622535706,
+ "learning_rate": 0.0006,
+ "loss": 6.760368347167969,
+ "step": 3182
+ },
+ {
+ "epoch": 44.209606986899566,
+ "grad_norm": 0.0266821701079607,
+ "learning_rate": 0.0006,
+ "loss": 6.71417236328125,
+ "step": 3183
+ },
+ {
+ "epoch": 44.223580786026204,
+ "grad_norm": 0.028125090524554253,
+ "learning_rate": 0.0006,
+ "loss": 6.646126747131348,
+ "step": 3184
+ },
+ {
+ "epoch": 44.237554585152836,
+ "grad_norm": 0.0185737032443285,
+ "learning_rate": 0.0006,
+ "loss": 6.712944030761719,
+ "step": 3185
+ },
+ {
+ "epoch": 44.251528384279474,
+ "grad_norm": 0.025836674496531487,
+ "learning_rate": 0.0006,
+ "loss": 6.707030296325684,
+ "step": 3186
+ },
+ {
+ "epoch": 44.26550218340611,
+ "grad_norm": 0.029744870960712433,
+ "learning_rate": 0.0006,
+ "loss": 6.6271071434021,
+ "step": 3187
+ },
+ {
+ "epoch": 44.27947598253275,
+ "grad_norm": 0.03021426685154438,
+ "learning_rate": 0.0006,
+ "loss": 6.677826404571533,
+ "step": 3188
+ },
+ {
+ "epoch": 44.29344978165939,
+ "grad_norm": 0.032727014273405075,
+ "learning_rate": 0.0006,
+ "loss": 6.657764434814453,
+ "step": 3189
+ },
+ {
+ "epoch": 44.30742358078603,
+ "grad_norm": 0.03696778789162636,
+ "learning_rate": 0.0006,
+ "loss": 6.632883548736572,
+ "step": 3190
+ },
+ {
+ "epoch": 44.32139737991266,
+ "grad_norm": 0.034630827605724335,
+ "learning_rate": 0.0006,
+ "loss": 6.759583473205566,
+ "step": 3191
+ },
+ {
+ "epoch": 44.3353711790393,
+ "grad_norm": 0.030517758801579475,
+ "learning_rate": 0.0006,
+ "loss": 6.702268600463867,
+ "step": 3192
+ },
+ {
+ "epoch": 44.34934497816594,
+ "grad_norm": 0.04105190932750702,
+ "learning_rate": 0.0006,
+ "loss": 6.705792427062988,
+ "step": 3193
+ },
+ {
+ "epoch": 44.36331877729258,
+ "grad_norm": 0.06659174710512161,
+ "learning_rate": 0.0006,
+ "loss": 6.73862361907959,
+ "step": 3194
+ },
+ {
+ "epoch": 44.377292576419215,
+ "grad_norm": 0.08013619482517242,
+ "learning_rate": 0.0006,
+ "loss": 6.710282802581787,
+ "step": 3195
+ },
+ {
+ "epoch": 44.391266375545854,
+ "grad_norm": 0.049850963056087494,
+ "learning_rate": 0.0006,
+ "loss": 6.684974670410156,
+ "step": 3196
+ },
+ {
+ "epoch": 44.40524017467249,
+ "grad_norm": 0.061247408390045166,
+ "learning_rate": 0.0006,
+ "loss": 6.716034412384033,
+ "step": 3197
+ },
+ {
+ "epoch": 44.419213973799124,
+ "grad_norm": 0.05868857726454735,
+ "learning_rate": 0.0006,
+ "loss": 6.589107513427734,
+ "step": 3198
+ },
+ {
+ "epoch": 44.43318777292576,
+ "grad_norm": 0.048878561705350876,
+ "learning_rate": 0.0006,
+ "loss": 6.694855690002441,
+ "step": 3199
+ },
+ {
+ "epoch": 44.4471615720524,
+ "grad_norm": 0.028748324140906334,
+ "learning_rate": 0.0006,
+ "loss": 6.629969596862793,
+ "step": 3200
+ },
+ {
+ "epoch": 44.46113537117904,
+ "grad_norm": 0.04060306400060654,
+ "learning_rate": 0.0006,
+ "loss": 6.721431255340576,
+ "step": 3201
+ },
+ {
+ "epoch": 44.47510917030568,
+ "grad_norm": 0.03833824768662453,
+ "learning_rate": 0.0006,
+ "loss": 6.625029563903809,
+ "step": 3202
+ },
+ {
+ "epoch": 44.48908296943232,
+ "grad_norm": 0.03791997581720352,
+ "learning_rate": 0.0006,
+ "loss": 6.669187545776367,
+ "step": 3203
+ },
+ {
+ "epoch": 44.50305676855895,
+ "grad_norm": 0.03259684890508652,
+ "learning_rate": 0.0006,
+ "loss": 6.665643692016602,
+ "step": 3204
+ },
+ {
+ "epoch": 44.51703056768559,
+ "grad_norm": 0.040658626705408096,
+ "learning_rate": 0.0006,
+ "loss": 6.639796257019043,
+ "step": 3205
+ },
+ {
+ "epoch": 44.531004366812226,
+ "grad_norm": 0.03691193461418152,
+ "learning_rate": 0.0006,
+ "loss": 6.692244529724121,
+ "step": 3206
+ },
+ {
+ "epoch": 44.544978165938865,
+ "grad_norm": 0.02814812771975994,
+ "learning_rate": 0.0006,
+ "loss": 6.61978006362915,
+ "step": 3207
+ },
+ {
+ "epoch": 44.5589519650655,
+ "grad_norm": 0.022155074402689934,
+ "learning_rate": 0.0006,
+ "loss": 6.625905990600586,
+ "step": 3208
+ },
+ {
+ "epoch": 44.57292576419214,
+ "grad_norm": 0.030794885009527206,
+ "learning_rate": 0.0006,
+ "loss": 6.678500652313232,
+ "step": 3209
+ },
+ {
+ "epoch": 44.58689956331878,
+ "grad_norm": 0.03201768547296524,
+ "learning_rate": 0.0006,
+ "loss": 6.6015729904174805,
+ "step": 3210
+ },
+ {
+ "epoch": 44.60087336244541,
+ "grad_norm": 0.025138361379504204,
+ "learning_rate": 0.0006,
+ "loss": 6.568234443664551,
+ "step": 3211
+ },
+ {
+ "epoch": 44.61484716157205,
+ "grad_norm": 0.02969353087246418,
+ "learning_rate": 0.0006,
+ "loss": 6.588366508483887,
+ "step": 3212
+ },
+ {
+ "epoch": 44.62882096069869,
+ "grad_norm": 0.03136962279677391,
+ "learning_rate": 0.0006,
+ "loss": 6.479961395263672,
+ "step": 3213
+ },
+ {
+ "epoch": 44.64279475982533,
+ "grad_norm": 0.030178170651197433,
+ "learning_rate": 0.0006,
+ "loss": 6.622910976409912,
+ "step": 3214
+ },
+ {
+ "epoch": 44.65676855895197,
+ "grad_norm": 0.02682245895266533,
+ "learning_rate": 0.0006,
+ "loss": 6.597707748413086,
+ "step": 3215
+ },
+ {
+ "epoch": 44.670742358078606,
+ "grad_norm": 0.025154557079076767,
+ "learning_rate": 0.0006,
+ "loss": 6.630541801452637,
+ "step": 3216
+ },
+ {
+ "epoch": 44.68471615720524,
+ "grad_norm": 0.03564463183283806,
+ "learning_rate": 0.0006,
+ "loss": 6.542823791503906,
+ "step": 3217
+ },
+ {
+ "epoch": 44.698689956331876,
+ "grad_norm": 0.049477651715278625,
+ "learning_rate": 0.0006,
+ "loss": 6.646030426025391,
+ "step": 3218
+ },
+ {
+ "epoch": 44.712663755458514,
+ "grad_norm": 0.06536535918712616,
+ "learning_rate": 0.0006,
+ "loss": 6.595463275909424,
+ "step": 3219
+ },
+ {
+ "epoch": 44.72663755458515,
+ "grad_norm": 0.07710316777229309,
+ "learning_rate": 0.0006,
+ "loss": 6.55524206161499,
+ "step": 3220
+ },
+ {
+ "epoch": 44.74061135371179,
+ "grad_norm": 0.09571163356304169,
+ "learning_rate": 0.0006,
+ "loss": 6.659764289855957,
+ "step": 3221
+ },
+ {
+ "epoch": 44.75458515283843,
+ "grad_norm": 0.055076830089092255,
+ "learning_rate": 0.0006,
+ "loss": 6.658882141113281,
+ "step": 3222
+ },
+ {
+ "epoch": 44.76855895196506,
+ "grad_norm": 0.07671099156141281,
+ "learning_rate": 0.0006,
+ "loss": 6.605809211730957,
+ "step": 3223
+ },
+ {
+ "epoch": 44.7825327510917,
+ "grad_norm": 0.05295584350824356,
+ "learning_rate": 0.0006,
+ "loss": 6.641455173492432,
+ "step": 3224
+ },
+ {
+ "epoch": 44.79650655021834,
+ "grad_norm": 0.09128042310476303,
+ "learning_rate": 0.0006,
+ "loss": 6.691066741943359,
+ "step": 3225
+ },
+ {
+ "epoch": 44.81048034934498,
+ "grad_norm": 0.05362958461046219,
+ "learning_rate": 0.0006,
+ "loss": 6.573406219482422,
+ "step": 3226
+ },
+ {
+ "epoch": 44.82445414847162,
+ "grad_norm": 0.046910062432289124,
+ "learning_rate": 0.0006,
+ "loss": 6.550876617431641,
+ "step": 3227
+ },
+ {
+ "epoch": 44.838427947598255,
+ "grad_norm": 0.03494197130203247,
+ "learning_rate": 0.0006,
+ "loss": 6.540179252624512,
+ "step": 3228
+ },
+ {
+ "epoch": 44.852401746724894,
+ "grad_norm": 0.04716387391090393,
+ "learning_rate": 0.0006,
+ "loss": 6.490459442138672,
+ "step": 3229
+ },
+ {
+ "epoch": 44.866375545851525,
+ "grad_norm": 0.04190506786108017,
+ "learning_rate": 0.0006,
+ "loss": 6.509925365447998,
+ "step": 3230
+ },
+ {
+ "epoch": 44.880349344978164,
+ "grad_norm": 0.04749064892530441,
+ "learning_rate": 0.0006,
+ "loss": 6.478779315948486,
+ "step": 3231
+ },
+ {
+ "epoch": 44.8943231441048,
+ "grad_norm": 0.028002185747027397,
+ "learning_rate": 0.0006,
+ "loss": 6.528904914855957,
+ "step": 3232
+ },
+ {
+ "epoch": 44.90829694323144,
+ "grad_norm": 0.036473166197538376,
+ "learning_rate": 0.0006,
+ "loss": 6.549042701721191,
+ "step": 3233
+ },
+ {
+ "epoch": 44.92227074235808,
+ "grad_norm": 0.02543315477669239,
+ "learning_rate": 0.0006,
+ "loss": 6.498875141143799,
+ "step": 3234
+ },
+ {
+ "epoch": 44.93624454148472,
+ "grad_norm": 0.029906807467341423,
+ "learning_rate": 0.0006,
+ "loss": 6.453019142150879,
+ "step": 3235
+ },
+ {
+ "epoch": 44.95021834061135,
+ "grad_norm": 0.027231857180595398,
+ "learning_rate": 0.0006,
+ "loss": 6.428829193115234,
+ "step": 3236
+ },
+ {
+ "epoch": 44.96419213973799,
+ "grad_norm": 0.02506154775619507,
+ "learning_rate": 0.0006,
+ "loss": 6.420858383178711,
+ "step": 3237
+ },
+ {
+ "epoch": 44.97816593886463,
+ "grad_norm": 0.0269472673535347,
+ "learning_rate": 0.0006,
+ "loss": 6.484989643096924,
+ "step": 3238
+ },
+ {
+ "epoch": 44.992139737991266,
+ "grad_norm": 0.023410487920045853,
+ "learning_rate": 0.0006,
+ "loss": 6.454830169677734,
+ "step": 3239
+ },
+ {
+ "epoch": 45.0,
+ "grad_norm": 0.034175023436546326,
+ "learning_rate": 0.0006,
+ "loss": 6.430975437164307,
+ "step": 3240
+ },
+ {
+ "epoch": 45.0,
+ "eval_loss": 6.458896160125732,
+ "eval_runtime": 56.6934,
+ "eval_samples_per_second": 43.074,
+ "eval_steps_per_second": 1.358,
+ "step": 3240
+ },
+ {
+ "epoch": 45.01397379912664,
+ "grad_norm": 0.045031629502773285,
+ "learning_rate": 0.0006,
+ "loss": 6.4243011474609375,
+ "step": 3241
+ },
+ {
+ "epoch": 45.02794759825328,
+ "grad_norm": 0.04968734085559845,
+ "learning_rate": 0.0006,
+ "loss": 6.380091667175293,
+ "step": 3242
+ },
+ {
+ "epoch": 45.041921397379916,
+ "grad_norm": 0.047145187854766846,
+ "learning_rate": 0.0006,
+ "loss": 6.477178573608398,
+ "step": 3243
+ },
+ {
+ "epoch": 45.05589519650655,
+ "grad_norm": 0.04304948449134827,
+ "learning_rate": 0.0006,
+ "loss": 6.399726867675781,
+ "step": 3244
+ },
+ {
+ "epoch": 45.069868995633186,
+ "grad_norm": 0.04466010630130768,
+ "learning_rate": 0.0006,
+ "loss": 6.3415679931640625,
+ "step": 3245
+ },
+ {
+ "epoch": 45.083842794759825,
+ "grad_norm": 0.04305306077003479,
+ "learning_rate": 0.0006,
+ "loss": 6.3413310050964355,
+ "step": 3246
+ },
+ {
+ "epoch": 45.09781659388646,
+ "grad_norm": 0.043064575642347336,
+ "learning_rate": 0.0006,
+ "loss": 6.4838714599609375,
+ "step": 3247
+ },
+ {
+ "epoch": 45.1117903930131,
+ "grad_norm": 0.049503277987241745,
+ "learning_rate": 0.0006,
+ "loss": 6.38473653793335,
+ "step": 3248
+ },
+ {
+ "epoch": 45.12576419213974,
+ "grad_norm": 0.04613238573074341,
+ "learning_rate": 0.0006,
+ "loss": 6.370769500732422,
+ "step": 3249
+ },
+ {
+ "epoch": 45.13973799126637,
+ "grad_norm": 0.052119478583335876,
+ "learning_rate": 0.0006,
+ "loss": 6.362321853637695,
+ "step": 3250
+ },
+ {
+ "epoch": 45.15371179039301,
+ "grad_norm": 0.045110076665878296,
+ "learning_rate": 0.0006,
+ "loss": 6.369844913482666,
+ "step": 3251
+ },
+ {
+ "epoch": 45.16768558951965,
+ "grad_norm": 0.03798583149909973,
+ "learning_rate": 0.0006,
+ "loss": 6.324633598327637,
+ "step": 3252
+ },
+ {
+ "epoch": 45.18165938864629,
+ "grad_norm": 0.03744659945368767,
+ "learning_rate": 0.0006,
+ "loss": 6.349421501159668,
+ "step": 3253
+ },
+ {
+ "epoch": 45.19563318777293,
+ "grad_norm": 0.03767292574048042,
+ "learning_rate": 0.0006,
+ "loss": 6.334238052368164,
+ "step": 3254
+ },
+ {
+ "epoch": 45.209606986899566,
+ "grad_norm": 0.03577509894967079,
+ "learning_rate": 0.0006,
+ "loss": 6.3453826904296875,
+ "step": 3255
+ },
+ {
+ "epoch": 45.223580786026204,
+ "grad_norm": 0.03403240442276001,
+ "learning_rate": 0.0006,
+ "loss": 6.2836174964904785,
+ "step": 3256
+ },
+ {
+ "epoch": 45.237554585152836,
+ "grad_norm": 0.03517331928014755,
+ "learning_rate": 0.0006,
+ "loss": 6.281718730926514,
+ "step": 3257
+ },
+ {
+ "epoch": 45.251528384279474,
+ "grad_norm": 0.04409139230847359,
+ "learning_rate": 0.0006,
+ "loss": 6.320815086364746,
+ "step": 3258
+ },
+ {
+ "epoch": 45.26550218340611,
+ "grad_norm": 0.054750364273786545,
+ "learning_rate": 0.0006,
+ "loss": 6.343618392944336,
+ "step": 3259
+ },
+ {
+ "epoch": 45.27947598253275,
+ "grad_norm": 0.07612067461013794,
+ "learning_rate": 0.0006,
+ "loss": 6.310614585876465,
+ "step": 3260
+ },
+ {
+ "epoch": 45.29344978165939,
+ "grad_norm": 0.08438985049724579,
+ "learning_rate": 0.0006,
+ "loss": 6.2369184494018555,
+ "step": 3261
+ },
+ {
+ "epoch": 45.30742358078603,
+ "grad_norm": 0.0782201811671257,
+ "learning_rate": 0.0006,
+ "loss": 6.254452705383301,
+ "step": 3262
+ },
+ {
+ "epoch": 45.32139737991266,
+ "grad_norm": 0.09928355365991592,
+ "learning_rate": 0.0006,
+ "loss": 6.3112077713012695,
+ "step": 3263
+ },
+ {
+ "epoch": 45.3353711790393,
+ "grad_norm": 0.07233133912086487,
+ "learning_rate": 0.0006,
+ "loss": 6.408087253570557,
+ "step": 3264
+ },
+ {
+ "epoch": 45.34934497816594,
+ "grad_norm": 0.07478684186935425,
+ "learning_rate": 0.0006,
+ "loss": 6.3770036697387695,
+ "step": 3265
+ },
+ {
+ "epoch": 45.36331877729258,
+ "grad_norm": 0.06036846712231636,
+ "learning_rate": 0.0006,
+ "loss": 6.33865213394165,
+ "step": 3266
+ },
+ {
+ "epoch": 45.377292576419215,
+ "grad_norm": 0.05783458054065704,
+ "learning_rate": 0.0006,
+ "loss": 6.363203048706055,
+ "step": 3267
+ },
+ {
+ "epoch": 45.391266375545854,
+ "grad_norm": 0.05443909019231796,
+ "learning_rate": 0.0006,
+ "loss": 6.287069797515869,
+ "step": 3268
+ },
+ {
+ "epoch": 45.40524017467249,
+ "grad_norm": 0.0557909794151783,
+ "learning_rate": 0.0006,
+ "loss": 6.297784805297852,
+ "step": 3269
+ },
+ {
+ "epoch": 45.419213973799124,
+ "grad_norm": 0.03483372926712036,
+ "learning_rate": 0.0006,
+ "loss": 6.280734062194824,
+ "step": 3270
+ },
+ {
+ "epoch": 45.43318777292576,
+ "grad_norm": 0.0425543375313282,
+ "learning_rate": 0.0006,
+ "loss": 6.33565616607666,
+ "step": 3271
+ },
+ {
+ "epoch": 45.4471615720524,
+ "grad_norm": 0.03230159357190132,
+ "learning_rate": 0.0006,
+ "loss": 6.197698593139648,
+ "step": 3272
+ },
+ {
+ "epoch": 45.46113537117904,
+ "grad_norm": 0.036897290498018265,
+ "learning_rate": 0.0006,
+ "loss": 6.241239547729492,
+ "step": 3273
+ },
+ {
+ "epoch": 45.47510917030568,
+ "grad_norm": 0.031103838235139847,
+ "learning_rate": 0.0006,
+ "loss": 6.159435749053955,
+ "step": 3274
+ },
+ {
+ "epoch": 45.48908296943232,
+ "grad_norm": 0.027116449549794197,
+ "learning_rate": 0.0006,
+ "loss": 6.212225914001465,
+ "step": 3275
+ },
+ {
+ "epoch": 45.50305676855895,
+ "grad_norm": 0.029923981055617332,
+ "learning_rate": 0.0006,
+ "loss": 6.157269477844238,
+ "step": 3276
+ },
+ {
+ "epoch": 45.51703056768559,
+ "grad_norm": 0.02521538734436035,
+ "learning_rate": 0.0006,
+ "loss": 6.146918773651123,
+ "step": 3277
+ },
+ {
+ "epoch": 45.531004366812226,
+ "grad_norm": 0.02758920192718506,
+ "learning_rate": 0.0006,
+ "loss": 6.0446977615356445,
+ "step": 3278
+ },
+ {
+ "epoch": 45.544978165938865,
+ "grad_norm": 0.024328220635652542,
+ "learning_rate": 0.0006,
+ "loss": 6.099508285522461,
+ "step": 3279
+ },
+ {
+ "epoch": 45.5589519650655,
+ "grad_norm": 0.027425771579146385,
+ "learning_rate": 0.0006,
+ "loss": 6.084299087524414,
+ "step": 3280
+ },
+ {
+ "epoch": 45.57292576419214,
+ "grad_norm": 0.026927651837468147,
+ "learning_rate": 0.0006,
+ "loss": 6.146334171295166,
+ "step": 3281
+ },
+ {
+ "epoch": 45.58689956331878,
+ "grad_norm": 0.04054379090666771,
+ "learning_rate": 0.0006,
+ "loss": 6.1301493644714355,
+ "step": 3282
+ },
+ {
+ "epoch": 45.60087336244541,
+ "grad_norm": 0.04456290975213051,
+ "learning_rate": 0.0006,
+ "loss": 6.152511119842529,
+ "step": 3283
+ },
+ {
+ "epoch": 45.61484716157205,
+ "grad_norm": 0.052821505814790726,
+ "learning_rate": 0.0006,
+ "loss": 6.113544940948486,
+ "step": 3284
+ },
+ {
+ "epoch": 45.62882096069869,
+ "grad_norm": 0.07251739501953125,
+ "learning_rate": 0.0006,
+ "loss": 6.118298530578613,
+ "step": 3285
+ },
+ {
+ "epoch": 45.64279475982533,
+ "grad_norm": 0.09433556348085403,
+ "learning_rate": 0.0006,
+ "loss": 6.170897483825684,
+ "step": 3286
+ },
+ {
+ "epoch": 45.65676855895197,
+ "grad_norm": 0.057683493942022324,
+ "learning_rate": 0.0006,
+ "loss": 6.1747822761535645,
+ "step": 3287
+ },
+ {
+ "epoch": 45.670742358078606,
+ "grad_norm": 0.06363270431756973,
+ "learning_rate": 0.0006,
+ "loss": 6.1617889404296875,
+ "step": 3288
+ },
+ {
+ "epoch": 45.68471615720524,
+ "grad_norm": 0.07528169453144073,
+ "learning_rate": 0.0006,
+ "loss": 6.230060577392578,
+ "step": 3289
+ },
+ {
+ "epoch": 45.698689956331876,
+ "grad_norm": 0.07816170156002045,
+ "learning_rate": 0.0006,
+ "loss": 6.171257019042969,
+ "step": 3290
+ },
+ {
+ "epoch": 45.712663755458514,
+ "grad_norm": 0.05860164389014244,
+ "learning_rate": 0.0006,
+ "loss": 6.095484256744385,
+ "step": 3291
+ },
+ {
+ "epoch": 45.72663755458515,
+ "grad_norm": 0.04747429117560387,
+ "learning_rate": 0.0006,
+ "loss": 6.060067176818848,
+ "step": 3292
+ },
+ {
+ "epoch": 45.74061135371179,
+ "grad_norm": 0.04514474794268608,
+ "learning_rate": 0.0006,
+ "loss": 6.100218772888184,
+ "step": 3293
+ },
+ {
+ "epoch": 45.75458515283843,
+ "grad_norm": 0.05168681964278221,
+ "learning_rate": 0.0006,
+ "loss": 6.0621867179870605,
+ "step": 3294
+ },
+ {
+ "epoch": 45.76855895196506,
+ "grad_norm": 0.04665876924991608,
+ "learning_rate": 0.0006,
+ "loss": 6.118292808532715,
+ "step": 3295
+ },
+ {
+ "epoch": 45.7825327510917,
+ "grad_norm": 0.049374762922525406,
+ "learning_rate": 0.0006,
+ "loss": 6.098394393920898,
+ "step": 3296
+ },
+ {
+ "epoch": 45.79650655021834,
+ "grad_norm": 0.03793549910187721,
+ "learning_rate": 0.0006,
+ "loss": 6.069541931152344,
+ "step": 3297
+ },
+ {
+ "epoch": 45.81048034934498,
+ "grad_norm": 0.036362383514642715,
+ "learning_rate": 0.0006,
+ "loss": 6.1188435554504395,
+ "step": 3298
+ },
+ {
+ "epoch": 45.82445414847162,
+ "grad_norm": 0.03397097811102867,
+ "learning_rate": 0.0006,
+ "loss": 6.117589950561523,
+ "step": 3299
+ },
+ {
+ "epoch": 45.838427947598255,
+ "grad_norm": 0.030977753922343254,
+ "learning_rate": 0.0006,
+ "loss": 6.0236711502075195,
+ "step": 3300
+ },
+ {
+ "epoch": 45.852401746724894,
+ "grad_norm": 0.02800217643380165,
+ "learning_rate": 0.0006,
+ "loss": 6.034825325012207,
+ "step": 3301
+ },
+ {
+ "epoch": 45.866375545851525,
+ "grad_norm": 0.03129778802394867,
+ "learning_rate": 0.0006,
+ "loss": 5.980271339416504,
+ "step": 3302
+ },
+ {
+ "epoch": 45.880349344978164,
+ "grad_norm": 0.025886094197630882,
+ "learning_rate": 0.0006,
+ "loss": 5.966419219970703,
+ "step": 3303
+ },
+ {
+ "epoch": 45.8943231441048,
+ "grad_norm": 0.025712328031659126,
+ "learning_rate": 0.0006,
+ "loss": 5.938652515411377,
+ "step": 3304
+ },
+ {
+ "epoch": 45.90829694323144,
+ "grad_norm": 0.022508298978209496,
+ "learning_rate": 0.0006,
+ "loss": 5.982312202453613,
+ "step": 3305
+ },
+ {
+ "epoch": 45.92227074235808,
+ "grad_norm": 0.024880951270461082,
+ "learning_rate": 0.0006,
+ "loss": 5.931150436401367,
+ "step": 3306
+ },
+ {
+ "epoch": 45.93624454148472,
+ "grad_norm": 0.030007576569914818,
+ "learning_rate": 0.0006,
+ "loss": 5.936639308929443,
+ "step": 3307
+ },
+ {
+ "epoch": 45.95021834061135,
+ "grad_norm": 0.033314041793346405,
+ "learning_rate": 0.0006,
+ "loss": 5.944248676300049,
+ "step": 3308
+ },
+ {
+ "epoch": 45.96419213973799,
+ "grad_norm": 0.043078381568193436,
+ "learning_rate": 0.0006,
+ "loss": 5.971286773681641,
+ "step": 3309
+ },
+ {
+ "epoch": 45.97816593886463,
+ "grad_norm": 0.059296440333127975,
+ "learning_rate": 0.0006,
+ "loss": 5.910497665405273,
+ "step": 3310
+ },
+ {
+ "epoch": 45.992139737991266,
+ "grad_norm": 0.06232666224241257,
+ "learning_rate": 0.0006,
+ "loss": 5.929817199707031,
+ "step": 3311
+ },
+ {
+ "epoch": 46.0,
+ "grad_norm": 0.05303940549492836,
+ "learning_rate": 0.0006,
+ "loss": 5.910822868347168,
+ "step": 3312
+ },
+ {
+ "epoch": 46.0,
+ "eval_loss": 5.977851390838623,
+ "eval_runtime": 56.2435,
+ "eval_samples_per_second": 43.418,
+ "eval_steps_per_second": 1.369,
+ "step": 3312
+ },
+ {
+ "epoch": 46.01397379912664,
+ "grad_norm": 0.06943798810243607,
+ "learning_rate": 0.0006,
+ "loss": 5.854922294616699,
+ "step": 3313
+ },
+ {
+ "epoch": 46.02794759825328,
+ "grad_norm": 0.05236278846859932,
+ "learning_rate": 0.0006,
+ "loss": 5.876611232757568,
+ "step": 3314
+ },
+ {
+ "epoch": 46.041921397379916,
+ "grad_norm": 0.06240001320838928,
+ "learning_rate": 0.0006,
+ "loss": 5.878912448883057,
+ "step": 3315
+ },
+ {
+ "epoch": 46.05589519650655,
+ "grad_norm": 0.07401192933320999,
+ "learning_rate": 0.0006,
+ "loss": 5.985136032104492,
+ "step": 3316
+ },
+ {
+ "epoch": 46.069868995633186,
+ "grad_norm": 0.0826529711484909,
+ "learning_rate": 0.0006,
+ "loss": 5.939439296722412,
+ "step": 3317
+ },
+ {
+ "epoch": 46.083842794759825,
+ "grad_norm": 0.07999005168676376,
+ "learning_rate": 0.0006,
+ "loss": 5.982336044311523,
+ "step": 3318
+ },
+ {
+ "epoch": 46.09781659388646,
+ "grad_norm": 0.07014869153499603,
+ "learning_rate": 0.0006,
+ "loss": 5.896534442901611,
+ "step": 3319
+ },
+ {
+ "epoch": 46.1117903930131,
+ "grad_norm": 0.05795128643512726,
+ "learning_rate": 0.0006,
+ "loss": 5.843111991882324,
+ "step": 3320
+ },
+ {
+ "epoch": 46.12576419213974,
+ "grad_norm": 0.06374258548021317,
+ "learning_rate": 0.0006,
+ "loss": 5.78752326965332,
+ "step": 3321
+ },
+ {
+ "epoch": 46.13973799126637,
+ "grad_norm": 0.045524805784225464,
+ "learning_rate": 0.0006,
+ "loss": 5.86755895614624,
+ "step": 3322
+ },
+ {
+ "epoch": 46.15371179039301,
+ "grad_norm": 0.03773504123091698,
+ "learning_rate": 0.0006,
+ "loss": 5.83627462387085,
+ "step": 3323
+ },
+ {
+ "epoch": 46.16768558951965,
+ "grad_norm": 0.035815201699733734,
+ "learning_rate": 0.0006,
+ "loss": 5.798238277435303,
+ "step": 3324
+ },
+ {
+ "epoch": 46.18165938864629,
+ "grad_norm": 0.03756783902645111,
+ "learning_rate": 0.0006,
+ "loss": 5.845197677612305,
+ "step": 3325
+ },
+ {
+ "epoch": 46.19563318777293,
+ "grad_norm": 0.03569113463163376,
+ "learning_rate": 0.0006,
+ "loss": 5.888932704925537,
+ "step": 3326
+ },
+ {
+ "epoch": 46.209606986899566,
+ "grad_norm": 0.029467813670635223,
+ "learning_rate": 0.0006,
+ "loss": 5.792047500610352,
+ "step": 3327
+ },
+ {
+ "epoch": 46.223580786026204,
+ "grad_norm": 0.025698406621813774,
+ "learning_rate": 0.0006,
+ "loss": 5.783734321594238,
+ "step": 3328
+ },
+ {
+ "epoch": 46.237554585152836,
+ "grad_norm": 0.028426671400666237,
+ "learning_rate": 0.0006,
+ "loss": 5.762084007263184,
+ "step": 3329
+ },
+ {
+ "epoch": 46.251528384279474,
+ "grad_norm": 0.03099622018635273,
+ "learning_rate": 0.0006,
+ "loss": 5.719109058380127,
+ "step": 3330
+ },
+ {
+ "epoch": 46.26550218340611,
+ "grad_norm": 0.0310219656676054,
+ "learning_rate": 0.0006,
+ "loss": 5.737057209014893,
+ "step": 3331
+ },
+ {
+ "epoch": 46.27947598253275,
+ "grad_norm": 0.02713446132838726,
+ "learning_rate": 0.0006,
+ "loss": 5.758657455444336,
+ "step": 3332
+ },
+ {
+ "epoch": 46.29344978165939,
+ "grad_norm": 0.03099844604730606,
+ "learning_rate": 0.0006,
+ "loss": 5.742086410522461,
+ "step": 3333
+ },
+ {
+ "epoch": 46.30742358078603,
+ "grad_norm": 0.030258730053901672,
+ "learning_rate": 0.0006,
+ "loss": 5.767426490783691,
+ "step": 3334
+ },
+ {
+ "epoch": 46.32139737991266,
+ "grad_norm": 0.029034852981567383,
+ "learning_rate": 0.0006,
+ "loss": 5.651792049407959,
+ "step": 3335
+ },
+ {
+ "epoch": 46.3353711790393,
+ "grad_norm": 0.03138665854930878,
+ "learning_rate": 0.0006,
+ "loss": 5.664980888366699,
+ "step": 3336
+ },
+ {
+ "epoch": 46.34934497816594,
+ "grad_norm": 0.040546901524066925,
+ "learning_rate": 0.0006,
+ "loss": 5.665165901184082,
+ "step": 3337
+ },
+ {
+ "epoch": 46.36331877729258,
+ "grad_norm": 0.06395485252141953,
+ "learning_rate": 0.0006,
+ "loss": 5.7776713371276855,
+ "step": 3338
+ },
+ {
+ "epoch": 46.377292576419215,
+ "grad_norm": 0.10088056325912476,
+ "learning_rate": 0.0006,
+ "loss": 5.755645751953125,
+ "step": 3339
+ },
+ {
+ "epoch": 46.391266375545854,
+ "grad_norm": 0.09527938067913055,
+ "learning_rate": 0.0006,
+ "loss": 5.838809490203857,
+ "step": 3340
+ },
+ {
+ "epoch": 46.40524017467249,
+ "grad_norm": 0.0849648267030716,
+ "learning_rate": 0.0006,
+ "loss": 5.781932830810547,
+ "step": 3341
+ },
+ {
+ "epoch": 46.419213973799124,
+ "grad_norm": 0.10142666101455688,
+ "learning_rate": 0.0006,
+ "loss": 5.838970184326172,
+ "step": 3342
+ },
+ {
+ "epoch": 46.43318777292576,
+ "grad_norm": 0.42731454968452454,
+ "learning_rate": 0.0006,
+ "loss": 6.071902275085449,
+ "step": 3343
+ },
+ {
+ "epoch": 46.4471615720524,
+ "grad_norm": 0.13102488219738007,
+ "learning_rate": 0.0006,
+ "loss": 6.8635993003845215,
+ "step": 3344
+ },
+ {
+ "epoch": 46.46113537117904,
+ "grad_norm": 0.23691938817501068,
+ "learning_rate": 0.0006,
+ "loss": 7.313158988952637,
+ "step": 3345
+ },
+ {
+ "epoch": 46.47510917030568,
+ "grad_norm": 0.11670669168233871,
+ "learning_rate": 0.0006,
+ "loss": 7.240363121032715,
+ "step": 3346
+ },
+ {
+ "epoch": 46.48908296943232,
+ "grad_norm": 0.09985820949077606,
+ "learning_rate": 0.0006,
+ "loss": 7.113332271575928,
+ "step": 3347
+ },
+ {
+ "epoch": 46.50305676855895,
+ "grad_norm": 0.11406408250331879,
+ "learning_rate": 0.0006,
+ "loss": 7.051285743713379,
+ "step": 3348
+ },
+ {
+ "epoch": 46.51703056768559,
+ "grad_norm": 0.07900336384773254,
+ "learning_rate": 0.0006,
+ "loss": 7.014365196228027,
+ "step": 3349
+ },
+ {
+ "epoch": 46.531004366812226,
+ "grad_norm": 0.05380333960056305,
+ "learning_rate": 0.0006,
+ "loss": 6.8868818283081055,
+ "step": 3350
+ },
+ {
+ "epoch": 46.544978165938865,
+ "grad_norm": 0.05868449807167053,
+ "learning_rate": 0.0006,
+ "loss": 6.783599853515625,
+ "step": 3351
+ },
+ {
+ "epoch": 46.5589519650655,
+ "grad_norm": 0.04428785294294357,
+ "learning_rate": 0.0006,
+ "loss": 6.798511505126953,
+ "step": 3352
+ },
+ {
+ "epoch": 46.57292576419214,
+ "grad_norm": 0.055527277290821075,
+ "learning_rate": 0.0006,
+ "loss": 6.694507598876953,
+ "step": 3353
+ },
+ {
+ "epoch": 46.58689956331878,
+ "grad_norm": 0.03872432932257652,
+ "learning_rate": 0.0006,
+ "loss": 6.736631393432617,
+ "step": 3354
+ },
+ {
+ "epoch": 46.60087336244541,
+ "grad_norm": 0.04330779239535332,
+ "learning_rate": 0.0006,
+ "loss": 6.576806545257568,
+ "step": 3355
+ },
+ {
+ "epoch": 46.61484716157205,
+ "grad_norm": 0.036767736077308655,
+ "learning_rate": 0.0006,
+ "loss": 6.559871196746826,
+ "step": 3356
+ },
+ {
+ "epoch": 46.62882096069869,
+ "grad_norm": 0.039857108145952225,
+ "learning_rate": 0.0006,
+ "loss": 6.5749831199646,
+ "step": 3357
+ },
+ {
+ "epoch": 46.64279475982533,
+ "grad_norm": 0.048719972372055054,
+ "learning_rate": 0.0006,
+ "loss": 6.4915618896484375,
+ "step": 3358
+ },
+ {
+ "epoch": 46.65676855895197,
+ "grad_norm": 0.042849015444517136,
+ "learning_rate": 0.0006,
+ "loss": 6.413516044616699,
+ "step": 3359
+ },
+ {
+ "epoch": 46.670742358078606,
+ "grad_norm": 0.02871009334921837,
+ "learning_rate": 0.0006,
+ "loss": 6.320952415466309,
+ "step": 3360
+ },
+ {
+ "epoch": 46.68471615720524,
+ "grad_norm": 0.03193620592355728,
+ "learning_rate": 0.0006,
+ "loss": 6.4176435470581055,
+ "step": 3361
+ },
+ {
+ "epoch": 46.698689956331876,
+ "grad_norm": 0.027417024597525597,
+ "learning_rate": 0.0006,
+ "loss": 6.342319488525391,
+ "step": 3362
+ },
+ {
+ "epoch": 46.712663755458514,
+ "grad_norm": 0.03305937349796295,
+ "learning_rate": 0.0006,
+ "loss": 6.246112823486328,
+ "step": 3363
+ },
+ {
+ "epoch": 46.72663755458515,
+ "grad_norm": 0.031403183937072754,
+ "learning_rate": 0.0006,
+ "loss": 6.149503231048584,
+ "step": 3364
+ },
+ {
+ "epoch": 46.74061135371179,
+ "grad_norm": 0.031563788652420044,
+ "learning_rate": 0.0006,
+ "loss": 6.278808116912842,
+ "step": 3365
+ },
+ {
+ "epoch": 46.75458515283843,
+ "grad_norm": 0.028366589918732643,
+ "learning_rate": 0.0006,
+ "loss": 6.057036876678467,
+ "step": 3366
+ },
+ {
+ "epoch": 46.76855895196506,
+ "grad_norm": 0.023715803399682045,
+ "learning_rate": 0.0006,
+ "loss": 6.111324310302734,
+ "step": 3367
+ },
+ {
+ "epoch": 46.7825327510917,
+ "grad_norm": 0.026546811684966087,
+ "learning_rate": 0.0006,
+ "loss": 6.105669975280762,
+ "step": 3368
+ },
+ {
+ "epoch": 46.79650655021834,
+ "grad_norm": 0.023531019687652588,
+ "learning_rate": 0.0006,
+ "loss": 6.068072319030762,
+ "step": 3369
+ },
+ {
+ "epoch": 46.81048034934498,
+ "grad_norm": 0.024480128660798073,
+ "learning_rate": 0.0006,
+ "loss": 5.947896957397461,
+ "step": 3370
+ },
+ {
+ "epoch": 46.82445414847162,
+ "grad_norm": 0.02517310529947281,
+ "learning_rate": 0.0006,
+ "loss": 6.030037879943848,
+ "step": 3371
+ },
+ {
+ "epoch": 46.838427947598255,
+ "grad_norm": 0.026723403483629227,
+ "learning_rate": 0.0006,
+ "loss": 5.963776111602783,
+ "step": 3372
+ },
+ {
+ "epoch": 46.852401746724894,
+ "grad_norm": 0.03163504600524902,
+ "learning_rate": 0.0006,
+ "loss": 5.893466949462891,
+ "step": 3373
+ },
+ {
+ "epoch": 46.866375545851525,
+ "grad_norm": 0.04058683291077614,
+ "learning_rate": 0.0006,
+ "loss": 5.9202799797058105,
+ "step": 3374
+ },
+ {
+ "epoch": 46.880349344978164,
+ "grad_norm": 0.04964117333292961,
+ "learning_rate": 0.0006,
+ "loss": 5.9366559982299805,
+ "step": 3375
+ },
+ {
+ "epoch": 46.8943231441048,
+ "grad_norm": 0.055406033992767334,
+ "learning_rate": 0.0006,
+ "loss": 5.912230968475342,
+ "step": 3376
+ },
+ {
+ "epoch": 46.90829694323144,
+ "grad_norm": 0.04643625393509865,
+ "learning_rate": 0.0006,
+ "loss": 5.87260103225708,
+ "step": 3377
+ },
+ {
+ "epoch": 46.92227074235808,
+ "grad_norm": 0.029325876384973526,
+ "learning_rate": 0.0006,
+ "loss": 5.854763984680176,
+ "step": 3378
+ },
+ {
+ "epoch": 46.93624454148472,
+ "grad_norm": 0.03159382566809654,
+ "learning_rate": 0.0006,
+ "loss": 5.880882263183594,
+ "step": 3379
+ },
+ {
+ "epoch": 46.95021834061135,
+ "grad_norm": 0.030224012210965157,
+ "learning_rate": 0.0006,
+ "loss": 5.787360191345215,
+ "step": 3380
+ },
+ {
+ "epoch": 46.96419213973799,
+ "grad_norm": 0.028368130326271057,
+ "learning_rate": 0.0006,
+ "loss": 5.817195892333984,
+ "step": 3381
+ },
+ {
+ "epoch": 46.97816593886463,
+ "grad_norm": 0.029411399737000465,
+ "learning_rate": 0.0006,
+ "loss": 5.783403396606445,
+ "step": 3382
+ },
+ {
+ "epoch": 46.992139737991266,
+ "grad_norm": 0.029927462339401245,
+ "learning_rate": 0.0006,
+ "loss": 5.662444591522217,
+ "step": 3383
+ },
+ {
+ "epoch": 47.0,
+ "grad_norm": 0.02872961387038231,
+ "learning_rate": 0.0006,
+ "loss": 5.742781639099121,
+ "step": 3384
+ },
+ {
+ "epoch": 47.0,
+ "eval_loss": 5.870260238647461,
+ "eval_runtime": 57.3734,
+ "eval_samples_per_second": 42.563,
+ "eval_steps_per_second": 1.342,
+ "step": 3384
+ },
+ {
+ "epoch": 47.01397379912664,
+ "grad_norm": 0.028542131185531616,
+ "learning_rate": 0.0006,
+ "loss": 5.753256797790527,
+ "step": 3385
+ },
+ {
+ "epoch": 47.02794759825328,
+ "grad_norm": 0.03227119892835617,
+ "learning_rate": 0.0006,
+ "loss": 5.722246170043945,
+ "step": 3386
+ },
+ {
+ "epoch": 47.041921397379916,
+ "grad_norm": 0.031461723148822784,
+ "learning_rate": 0.0006,
+ "loss": 5.740520000457764,
+ "step": 3387
+ },
+ {
+ "epoch": 47.05589519650655,
+ "grad_norm": 0.03184220567345619,
+ "learning_rate": 0.0006,
+ "loss": 5.720890045166016,
+ "step": 3388
+ },
+ {
+ "epoch": 47.069868995633186,
+ "grad_norm": 0.030015580356121063,
+ "learning_rate": 0.0006,
+ "loss": 5.597814559936523,
+ "step": 3389
+ },
+ {
+ "epoch": 47.083842794759825,
+ "grad_norm": 0.029931779950857162,
+ "learning_rate": 0.0006,
+ "loss": 5.5412468910217285,
+ "step": 3390
+ },
+ {
+ "epoch": 47.09781659388646,
+ "grad_norm": 0.03206505626440048,
+ "learning_rate": 0.0006,
+ "loss": 5.6967597007751465,
+ "step": 3391
+ },
+ {
+ "epoch": 47.1117903930131,
+ "grad_norm": 0.03978161886334419,
+ "learning_rate": 0.0006,
+ "loss": 5.59661340713501,
+ "step": 3392
+ },
+ {
+ "epoch": 47.12576419213974,
+ "grad_norm": 0.04619080200791359,
+ "learning_rate": 0.0006,
+ "loss": 5.795788764953613,
+ "step": 3393
+ },
+ {
+ "epoch": 47.13973799126637,
+ "grad_norm": 0.049085382372140884,
+ "learning_rate": 0.0006,
+ "loss": 5.667335510253906,
+ "step": 3394
+ },
+ {
+ "epoch": 47.15371179039301,
+ "grad_norm": 0.04830858111381531,
+ "learning_rate": 0.0006,
+ "loss": 5.668806552886963,
+ "step": 3395
+ },
+ {
+ "epoch": 47.16768558951965,
+ "grad_norm": 0.0583178736269474,
+ "learning_rate": 0.0006,
+ "loss": 5.692285537719727,
+ "step": 3396
+ },
+ {
+ "epoch": 47.18165938864629,
+ "grad_norm": 0.05182253196835518,
+ "learning_rate": 0.0006,
+ "loss": 5.678645133972168,
+ "step": 3397
+ },
+ {
+ "epoch": 47.19563318777293,
+ "grad_norm": 0.04011905938386917,
+ "learning_rate": 0.0006,
+ "loss": 5.661550998687744,
+ "step": 3398
+ },
+ {
+ "epoch": 47.209606986899566,
+ "grad_norm": 0.03888486698269844,
+ "learning_rate": 0.0006,
+ "loss": 5.643001556396484,
+ "step": 3399
+ },
+ {
+ "epoch": 47.223580786026204,
+ "grad_norm": 0.030658265575766563,
+ "learning_rate": 0.0006,
+ "loss": 5.659090995788574,
+ "step": 3400
+ },
+ {
+ "epoch": 47.237554585152836,
+ "grad_norm": 0.03609256446361542,
+ "learning_rate": 0.0006,
+ "loss": 5.544501781463623,
+ "step": 3401
+ },
+ {
+ "epoch": 47.251528384279474,
+ "grad_norm": 0.033589769154787064,
+ "learning_rate": 0.0006,
+ "loss": 5.463878631591797,
+ "step": 3402
+ },
+ {
+ "epoch": 47.26550218340611,
+ "grad_norm": 0.035128287971019745,
+ "learning_rate": 0.0006,
+ "loss": 5.5680060386657715,
+ "step": 3403
+ },
+ {
+ "epoch": 47.27947598253275,
+ "grad_norm": 0.038684550672769547,
+ "learning_rate": 0.0006,
+ "loss": 5.587898254394531,
+ "step": 3404
+ },
+ {
+ "epoch": 47.29344978165939,
+ "grad_norm": 0.043566226959228516,
+ "learning_rate": 0.0006,
+ "loss": 5.580896854400635,
+ "step": 3405
+ },
+ {
+ "epoch": 47.30742358078603,
+ "grad_norm": 0.04428374394774437,
+ "learning_rate": 0.0006,
+ "loss": 5.603460311889648,
+ "step": 3406
+ },
+ {
+ "epoch": 47.32139737991266,
+ "grad_norm": 0.037368934601545334,
+ "learning_rate": 0.0006,
+ "loss": 5.564740180969238,
+ "step": 3407
+ },
+ {
+ "epoch": 47.3353711790393,
+ "grad_norm": 0.04092350974678993,
+ "learning_rate": 0.0006,
+ "loss": 5.584153652191162,
+ "step": 3408
+ },
+ {
+ "epoch": 47.34934497816594,
+ "grad_norm": 0.045031506568193436,
+ "learning_rate": 0.0006,
+ "loss": 5.601186275482178,
+ "step": 3409
+ },
+ {
+ "epoch": 47.36331877729258,
+ "grad_norm": 0.041830409318208694,
+ "learning_rate": 0.0006,
+ "loss": 5.512788772583008,
+ "step": 3410
+ },
+ {
+ "epoch": 47.377292576419215,
+ "grad_norm": 0.037633974105119705,
+ "learning_rate": 0.0006,
+ "loss": 5.503828525543213,
+ "step": 3411
+ },
+ {
+ "epoch": 47.391266375545854,
+ "grad_norm": 0.03473588079214096,
+ "learning_rate": 0.0006,
+ "loss": 5.570557594299316,
+ "step": 3412
+ },
+ {
+ "epoch": 47.40524017467249,
+ "grad_norm": 0.039325181394815445,
+ "learning_rate": 0.0006,
+ "loss": 5.427783966064453,
+ "step": 3413
+ },
+ {
+ "epoch": 47.419213973799124,
+ "grad_norm": 0.04034624621272087,
+ "learning_rate": 0.0006,
+ "loss": 5.572591304779053,
+ "step": 3414
+ },
+ {
+ "epoch": 47.43318777292576,
+ "grad_norm": 0.05245184153318405,
+ "learning_rate": 0.0006,
+ "loss": 5.505208969116211,
+ "step": 3415
+ },
+ {
+ "epoch": 47.4471615720524,
+ "grad_norm": 0.0568675734102726,
+ "learning_rate": 0.0006,
+ "loss": 5.589373588562012,
+ "step": 3416
+ },
+ {
+ "epoch": 47.46113537117904,
+ "grad_norm": 0.048336587846279144,
+ "learning_rate": 0.0006,
+ "loss": 5.522005081176758,
+ "step": 3417
+ },
+ {
+ "epoch": 47.47510917030568,
+ "grad_norm": 0.04194975271821022,
+ "learning_rate": 0.0006,
+ "loss": 5.387576103210449,
+ "step": 3418
+ },
+ {
+ "epoch": 47.48908296943232,
+ "grad_norm": 0.04019433259963989,
+ "learning_rate": 0.0006,
+ "loss": 5.522615432739258,
+ "step": 3419
+ },
+ {
+ "epoch": 47.50305676855895,
+ "grad_norm": 0.05709933117032051,
+ "learning_rate": 0.0006,
+ "loss": 5.420286178588867,
+ "step": 3420
+ },
+ {
+ "epoch": 47.51703056768559,
+ "grad_norm": 0.09130338579416275,
+ "learning_rate": 0.0006,
+ "loss": 5.484622478485107,
+ "step": 3421
+ },
+ {
+ "epoch": 47.531004366812226,
+ "grad_norm": 0.2757541835308075,
+ "learning_rate": 0.0006,
+ "loss": 5.821832656860352,
+ "step": 3422
+ },
+ {
+ "epoch": 47.544978165938865,
+ "grad_norm": 0.28292080760002136,
+ "learning_rate": 0.0006,
+ "loss": 6.579934120178223,
+ "step": 3423
+ },
+ {
+ "epoch": 47.5589519650655,
+ "grad_norm": 0.15293604135513306,
+ "learning_rate": 0.0006,
+ "loss": 6.478213310241699,
+ "step": 3424
+ },
+ {
+ "epoch": 47.57292576419214,
+ "grad_norm": 0.11284393817186356,
+ "learning_rate": 0.0006,
+ "loss": 6.4066386222839355,
+ "step": 3425
+ },
+ {
+ "epoch": 47.58689956331878,
+ "grad_norm": 0.10933073610067368,
+ "learning_rate": 0.0006,
+ "loss": 6.3306169509887695,
+ "step": 3426
+ },
+ {
+ "epoch": 47.60087336244541,
+ "grad_norm": 0.05837641656398773,
+ "learning_rate": 0.0006,
+ "loss": 6.250000953674316,
+ "step": 3427
+ },
+ {
+ "epoch": 47.61484716157205,
+ "grad_norm": 0.05987229570746422,
+ "learning_rate": 0.0006,
+ "loss": 6.152463436126709,
+ "step": 3428
+ },
+ {
+ "epoch": 47.62882096069869,
+ "grad_norm": 0.04064938426017761,
+ "learning_rate": 0.0006,
+ "loss": 6.117897033691406,
+ "step": 3429
+ },
+ {
+ "epoch": 47.64279475982533,
+ "grad_norm": 0.04565057530999184,
+ "learning_rate": 0.0006,
+ "loss": 6.042291641235352,
+ "step": 3430
+ },
+ {
+ "epoch": 47.65676855895197,
+ "grad_norm": 0.03475063294172287,
+ "learning_rate": 0.0006,
+ "loss": 6.0071868896484375,
+ "step": 3431
+ },
+ {
+ "epoch": 47.670742358078606,
+ "grad_norm": 0.03809220343828201,
+ "learning_rate": 0.0006,
+ "loss": 6.0128021240234375,
+ "step": 3432
+ },
+ {
+ "epoch": 47.68471615720524,
+ "grad_norm": 0.03307219594717026,
+ "learning_rate": 0.0006,
+ "loss": 5.894664287567139,
+ "step": 3433
+ },
+ {
+ "epoch": 47.698689956331876,
+ "grad_norm": 0.028824346140027046,
+ "learning_rate": 0.0006,
+ "loss": 5.914865493774414,
+ "step": 3434
+ },
+ {
+ "epoch": 47.712663755458514,
+ "grad_norm": 0.027923690155148506,
+ "learning_rate": 0.0006,
+ "loss": 5.851759910583496,
+ "step": 3435
+ },
+ {
+ "epoch": 47.72663755458515,
+ "grad_norm": 0.025492684915661812,
+ "learning_rate": 0.0006,
+ "loss": 5.749186038970947,
+ "step": 3436
+ },
+ {
+ "epoch": 47.74061135371179,
+ "grad_norm": 0.025105174630880356,
+ "learning_rate": 0.0006,
+ "loss": 5.803566932678223,
+ "step": 3437
+ },
+ {
+ "epoch": 47.75458515283843,
+ "grad_norm": 0.023298880085349083,
+ "learning_rate": 0.0006,
+ "loss": 5.76276159286499,
+ "step": 3438
+ },
+ {
+ "epoch": 47.76855895196506,
+ "grad_norm": 0.02468576282262802,
+ "learning_rate": 0.0006,
+ "loss": 5.833852767944336,
+ "step": 3439
+ },
+ {
+ "epoch": 47.7825327510917,
+ "grad_norm": 0.022730689495801926,
+ "learning_rate": 0.0006,
+ "loss": 5.736724376678467,
+ "step": 3440
+ },
+ {
+ "epoch": 47.79650655021834,
+ "grad_norm": 0.021335354074835777,
+ "learning_rate": 0.0006,
+ "loss": 5.644248008728027,
+ "step": 3441
+ },
+ {
+ "epoch": 47.81048034934498,
+ "grad_norm": 0.02288101427257061,
+ "learning_rate": 0.0006,
+ "loss": 5.7011942863464355,
+ "step": 3442
+ },
+ {
+ "epoch": 47.82445414847162,
+ "grad_norm": 0.01961585320532322,
+ "learning_rate": 0.0006,
+ "loss": 5.63215446472168,
+ "step": 3443
+ },
+ {
+ "epoch": 47.838427947598255,
+ "grad_norm": 0.017659205943346024,
+ "learning_rate": 0.0006,
+ "loss": 5.744845390319824,
+ "step": 3444
+ },
+ {
+ "epoch": 47.852401746724894,
+ "grad_norm": 0.02281784638762474,
+ "learning_rate": 0.0006,
+ "loss": 5.696993350982666,
+ "step": 3445
+ },
+ {
+ "epoch": 47.866375545851525,
+ "grad_norm": 0.016394242644309998,
+ "learning_rate": 0.0006,
+ "loss": 5.659870147705078,
+ "step": 3446
+ },
+ {
+ "epoch": 47.880349344978164,
+ "grad_norm": 0.01967368647456169,
+ "learning_rate": 0.0006,
+ "loss": 5.638072490692139,
+ "step": 3447
+ },
+ {
+ "epoch": 47.8943231441048,
+ "grad_norm": 0.017326267436146736,
+ "learning_rate": 0.0006,
+ "loss": 5.638836860656738,
+ "step": 3448
+ },
+ {
+ "epoch": 47.90829694323144,
+ "grad_norm": 0.017778463661670685,
+ "learning_rate": 0.0006,
+ "loss": 5.634466648101807,
+ "step": 3449
+ },
+ {
+ "epoch": 47.92227074235808,
+ "grad_norm": 0.017269758507609367,
+ "learning_rate": 0.0006,
+ "loss": 5.506277561187744,
+ "step": 3450
+ },
+ {
+ "epoch": 47.93624454148472,
+ "grad_norm": 0.01733289286494255,
+ "learning_rate": 0.0006,
+ "loss": 5.544455528259277,
+ "step": 3451
+ },
+ {
+ "epoch": 47.95021834061135,
+ "grad_norm": 0.017787795513868332,
+ "learning_rate": 0.0006,
+ "loss": 5.498943328857422,
+ "step": 3452
+ },
+ {
+ "epoch": 47.96419213973799,
+ "grad_norm": 0.015770932659506798,
+ "learning_rate": 0.0006,
+ "loss": 5.537227630615234,
+ "step": 3453
+ },
+ {
+ "epoch": 47.97816593886463,
+ "grad_norm": 0.016132809221744537,
+ "learning_rate": 0.0006,
+ "loss": 5.489081382751465,
+ "step": 3454
+ },
+ {
+ "epoch": 47.992139737991266,
+ "grad_norm": 0.01749664731323719,
+ "learning_rate": 0.0006,
+ "loss": 5.454263687133789,
+ "step": 3455
+ },
+ {
+ "epoch": 48.0,
+ "grad_norm": 0.016711296513676643,
+ "learning_rate": 0.0006,
+ "loss": 5.530783653259277,
+ "step": 3456
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 28800,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 400,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": true,
+ "should_log": false,
+ "should_save": false,
+ "should_training_stop": false
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 1.4734144134394675e+18,
+ "train_batch_size": 8,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-3456/training_args.bin b/runs/l2r50-i2-fulle-lm/checkpoint-3456/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..cad232831c222cc7f8b938b8bbe42f61e7d89c3a
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-3456/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:f7c40e3bfa2f5ac1450ca1243e3f2a895d44a06ed5995e6aefd97f78d229d8c8
+size 4792
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-3744/chat_template.jinja b/runs/l2r50-i2-fulle-lm/checkpoint-3744/chat_template.jinja
new file mode 100644
index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-3744/chat_template.jinja
@@ -0,0 +1,85 @@
+{%- if tools %}
+ {{- '<|im_start|>system\n' }}
+ {%- if messages[0].role == 'system' %}
+ {{- messages[0].content + '\n\n' }}
+ {%- endif %}
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }}
+ {%- for tool in tools %}
+ {{- "\n" }}
+ {{- tool | tojson }}
+ {%- endfor %}
+ {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }}
+{%- else %}
+ {%- if messages[0].role == 'system' %}
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
+ {%- endif %}
+{%- endif %}
+{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
+{%- for message in messages[::-1] %}
+ {%- set index = (messages|length - 1) - loop.index0 %}
+ {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %}
+ {%- set ns.multi_step_tool = false %}
+ {%- set ns.last_query_index = index %}
+ {%- endif %}
+{%- endfor %}
+{%- for message in messages %}
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
+ {%- elif message.role == "assistant" %}
+ {%- set content = message.content %}
+ {%- set reasoning_content = '' %}
+ {%- if message.reasoning_content is defined and message.reasoning_content is not none %}
+ {%- set reasoning_content = message.reasoning_content %}
+ {%- else %}
+ {%- if '' in message.content %}
+ {%- set content = message.content.split('')[-1].lstrip('\n') %}
+ {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %}
+ {%- endif %}
+ {%- endif %}
+ {%- if loop.index0 > ns.last_query_index %}
+ {%- if loop.last or (not loop.last and reasoning_content) %}
+ {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- if message.tool_calls %}
+ {%- for tool_call in message.tool_calls %}
+ {%- if (loop.first and content) or (not loop.first) %}
+ {{- '\n' }}
+ {%- endif %}
+ {%- if tool_call.function %}
+ {%- set tool_call = tool_call.function %}
+ {%- endif %}
+ {{- '\n{"name": "' }}
+ {{- tool_call.name }}
+ {{- '", "arguments": ' }}
+ {%- if tool_call.arguments is string %}
+ {{- tool_call.arguments }}
+ {%- else %}
+ {{- tool_call.arguments | tojson }}
+ {%- endif %}
+ {{- '}\n' }}
+ {%- endfor %}
+ {%- endif %}
+ {{- '<|im_end|>\n' }}
+ {%- elif message.role == "tool" %}
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
+ {{- '<|im_start|>user' }}
+ {%- endif %}
+ {{- '\n\n' }}
+ {{- message.content }}
+ {{- '\n' }}
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
+ {{- '<|im_end|>\n' }}
+ {%- endif %}
+ {%- endif %}
+{%- endfor %}
+{%- if add_generation_prompt %}
+ {{- '<|im_start|>assistant\n' }}
+ {%- if enable_thinking is defined and enable_thinking is false %}
+ {{- '\n\n\n\n' }}
+ {%- endif %}
+{%- endif %}
\ No newline at end of file
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-3744/config.json b/runs/l2r50-i2-fulle-lm/checkpoint-3744/config.json
new file mode 100644
index 0000000000000000000000000000000000000000..22f250953c6ce8fc91ee4411856dd73041a415e2
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-3744/config.json
@@ -0,0 +1,32 @@
+{
+ "architectures": [
+ "LlamaForCausalLM"
+ ],
+ "attention_bias": false,
+ "attention_dropout": 0.0,
+ "bos_token_id": null,
+ "dtype": "float32",
+ "eos_token_id": 151645,
+ "head_dim": 128,
+ "hidden_act": "silu",
+ "hidden_size": 512,
+ "initializer_range": 0.02,
+ "intermediate_size": 1536,
+ "max_position_embeddings": 2048,
+ "mlp_bias": false,
+ "model_type": "llama",
+ "num_attention_heads": 4,
+ "num_hidden_layers": 20,
+ "num_key_value_heads": 4,
+ "pad_token_id": 151645,
+ "pretraining_tp": 1,
+ "rms_norm_eps": 1e-06,
+ "rope_parameters": {
+ "rope_theta": 10000.0,
+ "rope_type": "default"
+ },
+ "tie_word_embeddings": true,
+ "transformers_version": "5.5.0",
+ "use_cache": false,
+ "vocab_size": 151673
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-3744/generation_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-3744/generation_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-3744/generation_config.json
@@ -0,0 +1,11 @@
+{
+ "_from_model_config": true,
+ "eos_token_id": [
+ 151645
+ ],
+ "output_attentions": false,
+ "output_hidden_states": false,
+ "pad_token_id": 151645,
+ "transformers_version": "5.5.0",
+ "use_cache": true
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-3744/model.safetensors b/runs/l2r50-i2-fulle-lm/checkpoint-3744/model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..2ec81ab0377664c10aabba3af0bf3b69f93ae2cd
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-3744/model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:9bc2b51e2915002f43ce9df305c5014f83c702527028ecdf03c2c77dec88feee
+size 583360328
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-3744/optimizer.pt b/runs/l2r50-i2-fulle-lm/checkpoint-3744/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..3ffbfe0662338dce32c7ee31990f1a1a1ea93d1c
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-3744/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:9c7b20cb99576253b87386125af4bbb9ad1121eb8df6cea4e611b07ed54ea615
+size 1166833530
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-3744/rng_state_0.pth b/runs/l2r50-i2-fulle-lm/checkpoint-3744/rng_state_0.pth
new file mode 100644
index 0000000000000000000000000000000000000000..c097e0911dd435e7778913d466a4c8e8b90d9b84
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-3744/rng_state_0.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:1176d244adaa134fc1c33fc3d5f23966bc402da63e8d71cc7436503fabd30d43
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-3744/rng_state_1.pth b/runs/l2r50-i2-fulle-lm/checkpoint-3744/rng_state_1.pth
new file mode 100644
index 0000000000000000000000000000000000000000..fb77759bb4236393d9b44cac5502638baa61429e
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-3744/rng_state_1.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:752f121ae614cecac88411665432466278a4bb5b3fc46ca2ef8287271608db27
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-3744/rng_state_2.pth b/runs/l2r50-i2-fulle-lm/checkpoint-3744/rng_state_2.pth
new file mode 100644
index 0000000000000000000000000000000000000000..c4e6038a511b50d7a59a55c7ef57633938b53fd0
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-3744/rng_state_2.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:350de1630844c162a2e13816d52724a1a18ced3552347a5ca57068d1ee6c50b7
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-3744/rng_state_3.pth b/runs/l2r50-i2-fulle-lm/checkpoint-3744/rng_state_3.pth
new file mode 100644
index 0000000000000000000000000000000000000000..b675dc9296475644de03638efd8b8109d32ca1ef
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-3744/rng_state_3.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:8923c51953991a359e189cde40248a1821c5e36a0f0dc6ee11bd455b08aba509
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-3744/scheduler.pt b/runs/l2r50-i2-fulle-lm/checkpoint-3744/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..20b086465f72e318302a946aa7b4256ef216a798
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-3744/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:7fdc869be493cb85f416c60210527a1c69abda3f8a295fcb1ff6f1bec552aedc
+size 1064
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-3744/tokenizer.json b/runs/l2r50-i2-fulle-lm/checkpoint-3744/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..53786016f070e11184813f3c83101dc5e93e392e
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-3744/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:c3c63fda1646a555448e8f03662542e04b11d103721319e4d9b267d7da03ceb8
+size 11424463
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-3744/tokenizer_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-3744/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..98e2fb9b2cc54785dc9d871b8e503f8b8f2defe7
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-3744/tokenizer_config.json
@@ -0,0 +1,20 @@
+{
+ "add_prefix_space": false,
+ "backend": "tokenizers",
+ "bos_token": null,
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "<|im_end|>",
+ "errors": "replace",
+ "extra_special_tokens": [
+ "<|l2r_pred|>",
+ "<|r2l_pred|>",
+ "<|next_1_pred|>",
+ "<|next_2_pred|>"
+ ],
+ "is_local": false,
+ "model_max_length": 131072,
+ "pad_token": "<|im_end|>",
+ "split_special_tokens": false,
+ "tokenizer_class": "Qwen2Tokenizer",
+ "unk_token": null
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-3744/trainer_state.json b/runs/l2r50-i2-fulle-lm/checkpoint-3744/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..3605246a0545a878f31c4eacd16e77ceb81c0800
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-3744/trainer_state.json
@@ -0,0 +1,26650 @@
+{
+ "best_global_step": null,
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 52.0,
+ "eval_steps": 500,
+ "global_step": 3744,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.013973799126637555,
+ "grad_norm": 0.1739506870508194,
+ "learning_rate": 0.0,
+ "loss": 11.99299430847168,
+ "step": 1
+ },
+ {
+ "epoch": 0.02794759825327511,
+ "grad_norm": 0.17216597497463226,
+ "learning_rate": 5.999999999999999e-06,
+ "loss": 11.988704681396484,
+ "step": 2
+ },
+ {
+ "epoch": 0.04192139737991266,
+ "grad_norm": 0.17459715902805328,
+ "learning_rate": 1.1999999999999999e-05,
+ "loss": 11.945581436157227,
+ "step": 3
+ },
+ {
+ "epoch": 0.05589519650655022,
+ "grad_norm": 0.17577414214611053,
+ "learning_rate": 1.7999999999999997e-05,
+ "loss": 11.859772682189941,
+ "step": 4
+ },
+ {
+ "epoch": 0.06986899563318777,
+ "grad_norm": 0.16163651645183563,
+ "learning_rate": 2.3999999999999997e-05,
+ "loss": 11.744415283203125,
+ "step": 5
+ },
+ {
+ "epoch": 0.08384279475982533,
+ "grad_norm": 0.14409932494163513,
+ "learning_rate": 2.9999999999999997e-05,
+ "loss": 11.618099212646484,
+ "step": 6
+ },
+ {
+ "epoch": 0.09781659388646288,
+ "grad_norm": 0.13079848885536194,
+ "learning_rate": 3.5999999999999994e-05,
+ "loss": 11.505935668945312,
+ "step": 7
+ },
+ {
+ "epoch": 0.11179039301310044,
+ "grad_norm": 0.11964613944292068,
+ "learning_rate": 4.2e-05,
+ "loss": 11.416532516479492,
+ "step": 8
+ },
+ {
+ "epoch": 0.125764192139738,
+ "grad_norm": 0.11345142871141434,
+ "learning_rate": 4.7999999999999994e-05,
+ "loss": 11.33206558227539,
+ "step": 9
+ },
+ {
+ "epoch": 0.13973799126637554,
+ "grad_norm": 0.11033952981233597,
+ "learning_rate": 5.399999999999999e-05,
+ "loss": 11.259052276611328,
+ "step": 10
+ },
+ {
+ "epoch": 0.1537117903930131,
+ "grad_norm": 0.10776489973068237,
+ "learning_rate": 5.9999999999999995e-05,
+ "loss": 11.203384399414062,
+ "step": 11
+ },
+ {
+ "epoch": 0.16768558951965065,
+ "grad_norm": 0.10721946507692337,
+ "learning_rate": 6.599999999999999e-05,
+ "loss": 11.153295516967773,
+ "step": 12
+ },
+ {
+ "epoch": 0.18165938864628822,
+ "grad_norm": 0.10673430562019348,
+ "learning_rate": 7.199999999999999e-05,
+ "loss": 11.110200881958008,
+ "step": 13
+ },
+ {
+ "epoch": 0.19563318777292577,
+ "grad_norm": 0.10630254447460175,
+ "learning_rate": 7.8e-05,
+ "loss": 11.067477226257324,
+ "step": 14
+ },
+ {
+ "epoch": 0.2096069868995633,
+ "grad_norm": 0.10631615668535233,
+ "learning_rate": 8.4e-05,
+ "loss": 11.022941589355469,
+ "step": 15
+ },
+ {
+ "epoch": 0.22358078602620088,
+ "grad_norm": 0.10609443485736847,
+ "learning_rate": 8.999999999999999e-05,
+ "loss": 10.970746040344238,
+ "step": 16
+ },
+ {
+ "epoch": 0.23755458515283842,
+ "grad_norm": 0.10648232698440552,
+ "learning_rate": 9.599999999999999e-05,
+ "loss": 10.91267204284668,
+ "step": 17
+ },
+ {
+ "epoch": 0.251528384279476,
+ "grad_norm": 0.10584589838981628,
+ "learning_rate": 0.000102,
+ "loss": 10.854547500610352,
+ "step": 18
+ },
+ {
+ "epoch": 0.26550218340611353,
+ "grad_norm": 0.10591720044612885,
+ "learning_rate": 0.00010799999999999998,
+ "loss": 10.788917541503906,
+ "step": 19
+ },
+ {
+ "epoch": 0.2794759825327511,
+ "grad_norm": 0.10611672699451447,
+ "learning_rate": 0.00011399999999999999,
+ "loss": 10.717556953430176,
+ "step": 20
+ },
+ {
+ "epoch": 0.2934497816593886,
+ "grad_norm": 0.10455359518527985,
+ "learning_rate": 0.00011999999999999999,
+ "loss": 10.650186538696289,
+ "step": 21
+ },
+ {
+ "epoch": 0.3074235807860262,
+ "grad_norm": 0.1047525629401207,
+ "learning_rate": 0.00012599999999999997,
+ "loss": 10.56808090209961,
+ "step": 22
+ },
+ {
+ "epoch": 0.32139737991266376,
+ "grad_norm": 0.10511605441570282,
+ "learning_rate": 0.00013199999999999998,
+ "loss": 10.484028816223145,
+ "step": 23
+ },
+ {
+ "epoch": 0.3353711790393013,
+ "grad_norm": 0.10467177629470825,
+ "learning_rate": 0.000138,
+ "loss": 10.399947166442871,
+ "step": 24
+ },
+ {
+ "epoch": 0.34934497816593885,
+ "grad_norm": 0.10432136058807373,
+ "learning_rate": 0.00014399999999999998,
+ "loss": 10.312714576721191,
+ "step": 25
+ },
+ {
+ "epoch": 0.36331877729257644,
+ "grad_norm": 0.10527262091636658,
+ "learning_rate": 0.00015,
+ "loss": 10.212261199951172,
+ "step": 26
+ },
+ {
+ "epoch": 0.377292576419214,
+ "grad_norm": 0.103610560297966,
+ "learning_rate": 0.000156,
+ "loss": 10.132034301757812,
+ "step": 27
+ },
+ {
+ "epoch": 0.39126637554585153,
+ "grad_norm": 0.10553059726953506,
+ "learning_rate": 0.000162,
+ "loss": 10.020434379577637,
+ "step": 28
+ },
+ {
+ "epoch": 0.4052401746724891,
+ "grad_norm": 0.1037299782037735,
+ "learning_rate": 0.000168,
+ "loss": 9.934466361999512,
+ "step": 29
+ },
+ {
+ "epoch": 0.4192139737991266,
+ "grad_norm": 0.1042739674448967,
+ "learning_rate": 0.00017399999999999997,
+ "loss": 9.822539329528809,
+ "step": 30
+ },
+ {
+ "epoch": 0.4331877729257642,
+ "grad_norm": 0.10300865024328232,
+ "learning_rate": 0.00017999999999999998,
+ "loss": 9.734994888305664,
+ "step": 31
+ },
+ {
+ "epoch": 0.44716157205240176,
+ "grad_norm": 0.10291654616594315,
+ "learning_rate": 0.000186,
+ "loss": 9.624505996704102,
+ "step": 32
+ },
+ {
+ "epoch": 0.4611353711790393,
+ "grad_norm": 0.10210851579904556,
+ "learning_rate": 0.00019199999999999998,
+ "loss": 9.518205642700195,
+ "step": 33
+ },
+ {
+ "epoch": 0.47510917030567684,
+ "grad_norm": 0.10222529619932175,
+ "learning_rate": 0.000198,
+ "loss": 9.412222862243652,
+ "step": 34
+ },
+ {
+ "epoch": 0.4890829694323144,
+ "grad_norm": 0.09992019087076187,
+ "learning_rate": 0.000204,
+ "loss": 9.326566696166992,
+ "step": 35
+ },
+ {
+ "epoch": 0.503056768558952,
+ "grad_norm": 0.09931084513664246,
+ "learning_rate": 0.00020999999999999998,
+ "loss": 9.210111618041992,
+ "step": 36
+ },
+ {
+ "epoch": 0.5170305676855895,
+ "grad_norm": 0.09928256273269653,
+ "learning_rate": 0.00021599999999999996,
+ "loss": 9.100493431091309,
+ "step": 37
+ },
+ {
+ "epoch": 0.5310043668122271,
+ "grad_norm": 0.09843816608190536,
+ "learning_rate": 0.00022199999999999998,
+ "loss": 8.996543884277344,
+ "step": 38
+ },
+ {
+ "epoch": 0.5449781659388646,
+ "grad_norm": 0.09825374186038971,
+ "learning_rate": 0.00022799999999999999,
+ "loss": 8.888991355895996,
+ "step": 39
+ },
+ {
+ "epoch": 0.5589519650655022,
+ "grad_norm": 0.09602385759353638,
+ "learning_rate": 0.000234,
+ "loss": 8.802907943725586,
+ "step": 40
+ },
+ {
+ "epoch": 0.5729257641921397,
+ "grad_norm": 0.09414064139127731,
+ "learning_rate": 0.00023999999999999998,
+ "loss": 8.715423583984375,
+ "step": 41
+ },
+ {
+ "epoch": 0.5868995633187772,
+ "grad_norm": 0.09273893386125565,
+ "learning_rate": 0.00024599999999999996,
+ "loss": 8.61374282836914,
+ "step": 42
+ },
+ {
+ "epoch": 0.6008733624454149,
+ "grad_norm": 0.09079599380493164,
+ "learning_rate": 0.00025199999999999995,
+ "loss": 8.502904891967773,
+ "step": 43
+ },
+ {
+ "epoch": 0.6148471615720524,
+ "grad_norm": 0.08605825901031494,
+ "learning_rate": 0.000258,
+ "loss": 8.453008651733398,
+ "step": 44
+ },
+ {
+ "epoch": 0.62882096069869,
+ "grad_norm": 0.08338996022939682,
+ "learning_rate": 0.00026399999999999997,
+ "loss": 8.365163803100586,
+ "step": 45
+ },
+ {
+ "epoch": 0.6427947598253275,
+ "grad_norm": 0.08150867372751236,
+ "learning_rate": 0.00027,
+ "loss": 8.285372734069824,
+ "step": 46
+ },
+ {
+ "epoch": 0.6567685589519651,
+ "grad_norm": 0.07794807851314545,
+ "learning_rate": 0.000276,
+ "loss": 8.185811996459961,
+ "step": 47
+ },
+ {
+ "epoch": 0.6707423580786026,
+ "grad_norm": 0.07592720538377762,
+ "learning_rate": 0.00028199999999999997,
+ "loss": 8.098482131958008,
+ "step": 48
+ },
+ {
+ "epoch": 0.6847161572052402,
+ "grad_norm": 0.0709897130727768,
+ "learning_rate": 0.00028799999999999995,
+ "loss": 8.042091369628906,
+ "step": 49
+ },
+ {
+ "epoch": 0.6986899563318777,
+ "grad_norm": 0.06582305580377579,
+ "learning_rate": 0.000294,
+ "loss": 7.972844123840332,
+ "step": 50
+ },
+ {
+ "epoch": 0.7126637554585152,
+ "grad_norm": 0.059452351182699203,
+ "learning_rate": 0.0003,
+ "loss": 7.924579620361328,
+ "step": 51
+ },
+ {
+ "epoch": 0.7266375545851529,
+ "grad_norm": 0.052426718175411224,
+ "learning_rate": 0.00030599999999999996,
+ "loss": 7.876336097717285,
+ "step": 52
+ },
+ {
+ "epoch": 0.7406113537117904,
+ "grad_norm": 0.04768150672316551,
+ "learning_rate": 0.000312,
+ "loss": 7.8102707862854,
+ "step": 53
+ },
+ {
+ "epoch": 0.754585152838428,
+ "grad_norm": 0.03970075398683548,
+ "learning_rate": 0.000318,
+ "loss": 7.791913986206055,
+ "step": 54
+ },
+ {
+ "epoch": 0.7685589519650655,
+ "grad_norm": 0.03559752553701401,
+ "learning_rate": 0.000324,
+ "loss": 7.756960868835449,
+ "step": 55
+ },
+ {
+ "epoch": 0.7825327510917031,
+ "grad_norm": 0.029321353882551193,
+ "learning_rate": 0.00033,
+ "loss": 7.724878787994385,
+ "step": 56
+ },
+ {
+ "epoch": 0.7965065502183406,
+ "grad_norm": 0.023485427722334862,
+ "learning_rate": 0.000336,
+ "loss": 7.72553014755249,
+ "step": 57
+ },
+ {
+ "epoch": 0.8104803493449781,
+ "grad_norm": 0.019319474697113037,
+ "learning_rate": 0.00034199999999999996,
+ "loss": 7.709493160247803,
+ "step": 58
+ },
+ {
+ "epoch": 0.8244541484716157,
+ "grad_norm": 0.016705691814422607,
+ "learning_rate": 0.00034799999999999995,
+ "loss": 7.6873297691345215,
+ "step": 59
+ },
+ {
+ "epoch": 0.8384279475982532,
+ "grad_norm": 0.017667734995484352,
+ "learning_rate": 0.00035399999999999993,
+ "loss": 7.677059173583984,
+ "step": 60
+ },
+ {
+ "epoch": 0.8524017467248908,
+ "grad_norm": 0.01671992614865303,
+ "learning_rate": 0.00035999999999999997,
+ "loss": 7.643512725830078,
+ "step": 61
+ },
+ {
+ "epoch": 0.8663755458515284,
+ "grad_norm": 0.012288172729313374,
+ "learning_rate": 0.00036599999999999995,
+ "loss": 7.658343315124512,
+ "step": 62
+ },
+ {
+ "epoch": 0.880349344978166,
+ "grad_norm": 0.012004776857793331,
+ "learning_rate": 0.000372,
+ "loss": 7.649759292602539,
+ "step": 63
+ },
+ {
+ "epoch": 0.8943231441048035,
+ "grad_norm": 0.012487445957958698,
+ "learning_rate": 0.00037799999999999997,
+ "loss": 7.631503105163574,
+ "step": 64
+ },
+ {
+ "epoch": 0.9082969432314411,
+ "grad_norm": 0.01489241048693657,
+ "learning_rate": 0.00038399999999999996,
+ "loss": 7.647110462188721,
+ "step": 65
+ },
+ {
+ "epoch": 0.9222707423580786,
+ "grad_norm": 0.01428980939090252,
+ "learning_rate": 0.00039,
+ "loss": 7.636983871459961,
+ "step": 66
+ },
+ {
+ "epoch": 0.9362445414847161,
+ "grad_norm": 0.010711144655942917,
+ "learning_rate": 0.000396,
+ "loss": 7.621130466461182,
+ "step": 67
+ },
+ {
+ "epoch": 0.9502183406113537,
+ "grad_norm": 0.010178886353969574,
+ "learning_rate": 0.000402,
+ "loss": 7.593924522399902,
+ "step": 68
+ },
+ {
+ "epoch": 0.9641921397379912,
+ "grad_norm": 0.010604874230921268,
+ "learning_rate": 0.000408,
+ "loss": 7.6015496253967285,
+ "step": 69
+ },
+ {
+ "epoch": 0.9781659388646288,
+ "grad_norm": 0.009751426056027412,
+ "learning_rate": 0.0004139999999999999,
+ "loss": 7.571964263916016,
+ "step": 70
+ },
+ {
+ "epoch": 0.9921397379912664,
+ "grad_norm": 0.009816481731832027,
+ "learning_rate": 0.00041999999999999996,
+ "loss": 7.57344913482666,
+ "step": 71
+ },
+ {
+ "epoch": 1.0,
+ "grad_norm": 0.008665637113153934,
+ "learning_rate": 0.00042599999999999995,
+ "loss": 7.579050064086914,
+ "step": 72
+ },
+ {
+ "epoch": 1.0,
+ "eval_loss": 7.581204891204834,
+ "eval_runtime": 59.0091,
+ "eval_samples_per_second": 41.383,
+ "eval_steps_per_second": 1.305,
+ "step": 72
+ },
+ {
+ "epoch": 1.0139737991266375,
+ "grad_norm": 0.009684201329946518,
+ "learning_rate": 0.00043199999999999993,
+ "loss": 7.568594932556152,
+ "step": 73
+ },
+ {
+ "epoch": 1.027947598253275,
+ "grad_norm": 0.025783710181713104,
+ "learning_rate": 0.00043799999999999997,
+ "loss": 7.578948497772217,
+ "step": 74
+ },
+ {
+ "epoch": 1.0419213973799126,
+ "grad_norm": 0.0327819399535656,
+ "learning_rate": 0.00044399999999999995,
+ "loss": 7.548173904418945,
+ "step": 75
+ },
+ {
+ "epoch": 1.0558951965065502,
+ "grad_norm": 0.021604498848319054,
+ "learning_rate": 0.00045,
+ "loss": 7.552953720092773,
+ "step": 76
+ },
+ {
+ "epoch": 1.0698689956331877,
+ "grad_norm": 0.010230105370283127,
+ "learning_rate": 0.00045599999999999997,
+ "loss": 7.5213212966918945,
+ "step": 77
+ },
+ {
+ "epoch": 1.0838427947598253,
+ "grad_norm": 0.023315589874982834,
+ "learning_rate": 0.00046199999999999995,
+ "loss": 7.501640319824219,
+ "step": 78
+ },
+ {
+ "epoch": 1.0978165938864628,
+ "grad_norm": 0.01109944935888052,
+ "learning_rate": 0.000468,
+ "loss": 7.539519786834717,
+ "step": 79
+ },
+ {
+ "epoch": 1.1117903930131003,
+ "grad_norm": 0.026093758642673492,
+ "learning_rate": 0.000474,
+ "loss": 7.525928974151611,
+ "step": 80
+ },
+ {
+ "epoch": 1.125764192139738,
+ "grad_norm": 0.03274895250797272,
+ "learning_rate": 0.00047999999999999996,
+ "loss": 7.535602569580078,
+ "step": 81
+ },
+ {
+ "epoch": 1.1397379912663754,
+ "grad_norm": 0.01543602254241705,
+ "learning_rate": 0.000486,
+ "loss": 7.513683795928955,
+ "step": 82
+ },
+ {
+ "epoch": 1.1537117903930132,
+ "grad_norm": 0.028377249836921692,
+ "learning_rate": 0.0004919999999999999,
+ "loss": 7.49384880065918,
+ "step": 83
+ },
+ {
+ "epoch": 1.1676855895196507,
+ "grad_norm": 0.04121481627225876,
+ "learning_rate": 0.000498,
+ "loss": 7.488447189331055,
+ "step": 84
+ },
+ {
+ "epoch": 1.1816593886462883,
+ "grad_norm": 0.013282302767038345,
+ "learning_rate": 0.0005039999999999999,
+ "loss": 7.463010787963867,
+ "step": 85
+ },
+ {
+ "epoch": 1.1956331877729258,
+ "grad_norm": 0.03200364485383034,
+ "learning_rate": 0.0005099999999999999,
+ "loss": 7.473515033721924,
+ "step": 86
+ },
+ {
+ "epoch": 1.2096069868995634,
+ "grad_norm": 0.020768029615283012,
+ "learning_rate": 0.000516,
+ "loss": 7.464688777923584,
+ "step": 87
+ },
+ {
+ "epoch": 1.223580786026201,
+ "grad_norm": 0.021715788170695305,
+ "learning_rate": 0.000522,
+ "loss": 7.4835028648376465,
+ "step": 88
+ },
+ {
+ "epoch": 1.2375545851528384,
+ "grad_norm": 0.026255754753947258,
+ "learning_rate": 0.0005279999999999999,
+ "loss": 7.443392753601074,
+ "step": 89
+ },
+ {
+ "epoch": 1.251528384279476,
+ "grad_norm": 0.022325951606035233,
+ "learning_rate": 0.000534,
+ "loss": 7.457373142242432,
+ "step": 90
+ },
+ {
+ "epoch": 1.2655021834061135,
+ "grad_norm": 0.009270387701690197,
+ "learning_rate": 0.00054,
+ "loss": 7.455986976623535,
+ "step": 91
+ },
+ {
+ "epoch": 1.279475982532751,
+ "grad_norm": 0.02369544468820095,
+ "learning_rate": 0.0005459999999999999,
+ "loss": 7.442216873168945,
+ "step": 92
+ },
+ {
+ "epoch": 1.2934497816593886,
+ "grad_norm": 0.008474506437778473,
+ "learning_rate": 0.000552,
+ "loss": 7.4294915199279785,
+ "step": 93
+ },
+ {
+ "epoch": 1.3074235807860262,
+ "grad_norm": 0.017043249681591988,
+ "learning_rate": 0.000558,
+ "loss": 7.406932830810547,
+ "step": 94
+ },
+ {
+ "epoch": 1.3213973799126637,
+ "grad_norm": 0.011813412420451641,
+ "learning_rate": 0.0005639999999999999,
+ "loss": 7.415900230407715,
+ "step": 95
+ },
+ {
+ "epoch": 1.3353711790393012,
+ "grad_norm": 0.01086883433163166,
+ "learning_rate": 0.00057,
+ "loss": 7.452998638153076,
+ "step": 96
+ },
+ {
+ "epoch": 1.3493449781659388,
+ "grad_norm": 0.014700951054692268,
+ "learning_rate": 0.0005759999999999999,
+ "loss": 7.410249710083008,
+ "step": 97
+ },
+ {
+ "epoch": 1.3633187772925766,
+ "grad_norm": 0.014512976631522179,
+ "learning_rate": 0.0005819999999999999,
+ "loss": 7.429132461547852,
+ "step": 98
+ },
+ {
+ "epoch": 1.3772925764192139,
+ "grad_norm": 0.012519920244812965,
+ "learning_rate": 0.000588,
+ "loss": 7.430817604064941,
+ "step": 99
+ },
+ {
+ "epoch": 1.3912663755458516,
+ "grad_norm": 0.007132664322853088,
+ "learning_rate": 0.0005939999999999999,
+ "loss": 7.405234336853027,
+ "step": 100
+ },
+ {
+ "epoch": 1.405240174672489,
+ "grad_norm": 0.019489388912916183,
+ "learning_rate": 0.0006,
+ "loss": 7.427791595458984,
+ "step": 101
+ },
+ {
+ "epoch": 1.4192139737991267,
+ "grad_norm": 0.027252664789557457,
+ "learning_rate": 0.0006,
+ "loss": 7.41761589050293,
+ "step": 102
+ },
+ {
+ "epoch": 1.4331877729257643,
+ "grad_norm": 0.02501637488603592,
+ "learning_rate": 0.0006,
+ "loss": 7.416735649108887,
+ "step": 103
+ },
+ {
+ "epoch": 1.4471615720524018,
+ "grad_norm": 0.00989772193133831,
+ "learning_rate": 0.0006,
+ "loss": 7.3756632804870605,
+ "step": 104
+ },
+ {
+ "epoch": 1.4611353711790394,
+ "grad_norm": 0.01622949168086052,
+ "learning_rate": 0.0006,
+ "loss": 7.400811672210693,
+ "step": 105
+ },
+ {
+ "epoch": 1.475109170305677,
+ "grad_norm": 0.012407763861119747,
+ "learning_rate": 0.0006,
+ "loss": 7.37675666809082,
+ "step": 106
+ },
+ {
+ "epoch": 1.4890829694323144,
+ "grad_norm": 0.010353565216064453,
+ "learning_rate": 0.0006,
+ "loss": 7.4013142585754395,
+ "step": 107
+ },
+ {
+ "epoch": 1.503056768558952,
+ "grad_norm": 0.0120553532615304,
+ "learning_rate": 0.0006,
+ "loss": 7.379873275756836,
+ "step": 108
+ },
+ {
+ "epoch": 1.5170305676855895,
+ "grad_norm": 0.01610225811600685,
+ "learning_rate": 0.0006,
+ "loss": 7.371631145477295,
+ "step": 109
+ },
+ {
+ "epoch": 1.531004366812227,
+ "grad_norm": 0.01159879844635725,
+ "learning_rate": 0.0006,
+ "loss": 7.389007568359375,
+ "step": 110
+ },
+ {
+ "epoch": 1.5449781659388646,
+ "grad_norm": 0.008013816550374031,
+ "learning_rate": 0.0006,
+ "loss": 7.395920753479004,
+ "step": 111
+ },
+ {
+ "epoch": 1.5589519650655022,
+ "grad_norm": 0.016334809362888336,
+ "learning_rate": 0.0006,
+ "loss": 7.393121719360352,
+ "step": 112
+ },
+ {
+ "epoch": 1.5729257641921397,
+ "grad_norm": 0.018821721896529198,
+ "learning_rate": 0.0006,
+ "loss": 7.367543697357178,
+ "step": 113
+ },
+ {
+ "epoch": 1.5868995633187772,
+ "grad_norm": 0.014088994823396206,
+ "learning_rate": 0.0006,
+ "loss": 7.385537624359131,
+ "step": 114
+ },
+ {
+ "epoch": 1.600873362445415,
+ "grad_norm": 0.005864050704985857,
+ "learning_rate": 0.0006,
+ "loss": 7.365346908569336,
+ "step": 115
+ },
+ {
+ "epoch": 1.6148471615720523,
+ "grad_norm": 0.011745430529117584,
+ "learning_rate": 0.0006,
+ "loss": 7.390227317810059,
+ "step": 116
+ },
+ {
+ "epoch": 1.62882096069869,
+ "grad_norm": 0.017205573618412018,
+ "learning_rate": 0.0006,
+ "loss": 7.3558549880981445,
+ "step": 117
+ },
+ {
+ "epoch": 1.6427947598253274,
+ "grad_norm": 0.021256346255540848,
+ "learning_rate": 0.0006,
+ "loss": 7.377346992492676,
+ "step": 118
+ },
+ {
+ "epoch": 1.6567685589519652,
+ "grad_norm": 0.019335204735398293,
+ "learning_rate": 0.0006,
+ "loss": 7.346194267272949,
+ "step": 119
+ },
+ {
+ "epoch": 1.6707423580786025,
+ "grad_norm": 0.008134003728628159,
+ "learning_rate": 0.0006,
+ "loss": 7.3063578605651855,
+ "step": 120
+ },
+ {
+ "epoch": 1.6847161572052403,
+ "grad_norm": 0.017613831907510757,
+ "learning_rate": 0.0006,
+ "loss": 7.338171005249023,
+ "step": 121
+ },
+ {
+ "epoch": 1.6986899563318776,
+ "grad_norm": 0.00890734139829874,
+ "learning_rate": 0.0006,
+ "loss": 7.355384826660156,
+ "step": 122
+ },
+ {
+ "epoch": 1.7126637554585153,
+ "grad_norm": 0.013448155485093594,
+ "learning_rate": 0.0006,
+ "loss": 7.344296455383301,
+ "step": 123
+ },
+ {
+ "epoch": 1.726637554585153,
+ "grad_norm": 0.009241633117198944,
+ "learning_rate": 0.0006,
+ "loss": 7.3454060554504395,
+ "step": 124
+ },
+ {
+ "epoch": 1.7406113537117904,
+ "grad_norm": 0.018561311066150665,
+ "learning_rate": 0.0006,
+ "loss": 7.32348108291626,
+ "step": 125
+ },
+ {
+ "epoch": 1.754585152838428,
+ "grad_norm": 0.025667704641819,
+ "learning_rate": 0.0006,
+ "loss": 7.356142520904541,
+ "step": 126
+ },
+ {
+ "epoch": 1.7685589519650655,
+ "grad_norm": 0.036232151091098785,
+ "learning_rate": 0.0006,
+ "loss": 7.363539695739746,
+ "step": 127
+ },
+ {
+ "epoch": 1.782532751091703,
+ "grad_norm": 0.059483714401721954,
+ "learning_rate": 0.0006,
+ "loss": 7.402331352233887,
+ "step": 128
+ },
+ {
+ "epoch": 1.7965065502183406,
+ "grad_norm": 0.05365009233355522,
+ "learning_rate": 0.0006,
+ "loss": 7.40457820892334,
+ "step": 129
+ },
+ {
+ "epoch": 1.8104803493449781,
+ "grad_norm": 0.02738415263593197,
+ "learning_rate": 0.0006,
+ "loss": 7.3756208419799805,
+ "step": 130
+ },
+ {
+ "epoch": 1.8244541484716157,
+ "grad_norm": 0.035017091780900955,
+ "learning_rate": 0.0006,
+ "loss": 7.379263877868652,
+ "step": 131
+ },
+ {
+ "epoch": 1.8384279475982532,
+ "grad_norm": 0.02819863148033619,
+ "learning_rate": 0.0006,
+ "loss": 7.327673435211182,
+ "step": 132
+ },
+ {
+ "epoch": 1.8524017467248908,
+ "grad_norm": 0.017754238098859787,
+ "learning_rate": 0.0006,
+ "loss": 7.333420276641846,
+ "step": 133
+ },
+ {
+ "epoch": 1.8663755458515285,
+ "grad_norm": 0.024685733020305634,
+ "learning_rate": 0.0006,
+ "loss": 7.34493350982666,
+ "step": 134
+ },
+ {
+ "epoch": 1.8803493449781659,
+ "grad_norm": 0.019426532089710236,
+ "learning_rate": 0.0006,
+ "loss": 7.328448295593262,
+ "step": 135
+ },
+ {
+ "epoch": 1.8943231441048036,
+ "grad_norm": 0.02106391079723835,
+ "learning_rate": 0.0006,
+ "loss": 7.327917098999023,
+ "step": 136
+ },
+ {
+ "epoch": 1.908296943231441,
+ "grad_norm": 0.02056879736483097,
+ "learning_rate": 0.0006,
+ "loss": 7.3311991691589355,
+ "step": 137
+ },
+ {
+ "epoch": 1.9222707423580787,
+ "grad_norm": 0.016865141689777374,
+ "learning_rate": 0.0006,
+ "loss": 7.312766075134277,
+ "step": 138
+ },
+ {
+ "epoch": 1.936244541484716,
+ "grad_norm": 0.017549855634570122,
+ "learning_rate": 0.0006,
+ "loss": 7.31191349029541,
+ "step": 139
+ },
+ {
+ "epoch": 1.9502183406113538,
+ "grad_norm": 0.015930157154798508,
+ "learning_rate": 0.0006,
+ "loss": 7.330901622772217,
+ "step": 140
+ },
+ {
+ "epoch": 1.9641921397379911,
+ "grad_norm": 0.013283558189868927,
+ "learning_rate": 0.0006,
+ "loss": 7.298318386077881,
+ "step": 141
+ },
+ {
+ "epoch": 1.9781659388646289,
+ "grad_norm": 0.012023469433188438,
+ "learning_rate": 0.0006,
+ "loss": 7.282337188720703,
+ "step": 142
+ },
+ {
+ "epoch": 1.9921397379912664,
+ "grad_norm": 0.01459511648863554,
+ "learning_rate": 0.0006,
+ "loss": 7.282924175262451,
+ "step": 143
+ },
+ {
+ "epoch": 2.0,
+ "grad_norm": 0.017474235966801643,
+ "learning_rate": 0.0006,
+ "loss": 7.254333972930908,
+ "step": 144
+ },
+ {
+ "epoch": 2.0,
+ "eval_loss": 7.318708896636963,
+ "eval_runtime": 59.6555,
+ "eval_samples_per_second": 40.935,
+ "eval_steps_per_second": 1.291,
+ "step": 144
+ },
+ {
+ "epoch": 2.0139737991266378,
+ "grad_norm": 0.019757507368922234,
+ "learning_rate": 0.0006,
+ "loss": 7.30288553237915,
+ "step": 145
+ },
+ {
+ "epoch": 2.027947598253275,
+ "grad_norm": 0.014571248553693295,
+ "learning_rate": 0.0006,
+ "loss": 7.274681568145752,
+ "step": 146
+ },
+ {
+ "epoch": 2.041921397379913,
+ "grad_norm": 0.015744198113679886,
+ "learning_rate": 0.0006,
+ "loss": 7.278632640838623,
+ "step": 147
+ },
+ {
+ "epoch": 2.05589519650655,
+ "grad_norm": 0.028048858046531677,
+ "learning_rate": 0.0006,
+ "loss": 7.26558780670166,
+ "step": 148
+ },
+ {
+ "epoch": 2.069868995633188,
+ "grad_norm": 0.06547307223081589,
+ "learning_rate": 0.0006,
+ "loss": 7.32567834854126,
+ "step": 149
+ },
+ {
+ "epoch": 2.0838427947598253,
+ "grad_norm": 0.05969763547182083,
+ "learning_rate": 0.0006,
+ "loss": 7.395836353302002,
+ "step": 150
+ },
+ {
+ "epoch": 2.097816593886463,
+ "grad_norm": 0.032574281096458435,
+ "learning_rate": 0.0006,
+ "loss": 7.318150997161865,
+ "step": 151
+ },
+ {
+ "epoch": 2.1117903930131003,
+ "grad_norm": 0.029452962800860405,
+ "learning_rate": 0.0006,
+ "loss": 7.308066368103027,
+ "step": 152
+ },
+ {
+ "epoch": 2.125764192139738,
+ "grad_norm": 0.03161991387605667,
+ "learning_rate": 0.0006,
+ "loss": 7.306290626525879,
+ "step": 153
+ },
+ {
+ "epoch": 2.1397379912663754,
+ "grad_norm": 0.027096295729279518,
+ "learning_rate": 0.0006,
+ "loss": 7.297796249389648,
+ "step": 154
+ },
+ {
+ "epoch": 2.153711790393013,
+ "grad_norm": 0.028657056391239166,
+ "learning_rate": 0.0006,
+ "loss": 7.317424774169922,
+ "step": 155
+ },
+ {
+ "epoch": 2.1676855895196505,
+ "grad_norm": 0.02046995982527733,
+ "learning_rate": 0.0006,
+ "loss": 7.297986030578613,
+ "step": 156
+ },
+ {
+ "epoch": 2.1816593886462883,
+ "grad_norm": 0.02220369316637516,
+ "learning_rate": 0.0006,
+ "loss": 7.268237590789795,
+ "step": 157
+ },
+ {
+ "epoch": 2.1956331877729256,
+ "grad_norm": 0.018381357192993164,
+ "learning_rate": 0.0006,
+ "loss": 7.259415626525879,
+ "step": 158
+ },
+ {
+ "epoch": 2.2096069868995634,
+ "grad_norm": 0.0204413291066885,
+ "learning_rate": 0.0006,
+ "loss": 7.258173942565918,
+ "step": 159
+ },
+ {
+ "epoch": 2.2235807860262007,
+ "grad_norm": 0.015874288976192474,
+ "learning_rate": 0.0006,
+ "loss": 7.281834602355957,
+ "step": 160
+ },
+ {
+ "epoch": 2.2375545851528384,
+ "grad_norm": 0.01634068600833416,
+ "learning_rate": 0.0006,
+ "loss": 7.2469868659973145,
+ "step": 161
+ },
+ {
+ "epoch": 2.251528384279476,
+ "grad_norm": 0.01728747971355915,
+ "learning_rate": 0.0006,
+ "loss": 7.280995845794678,
+ "step": 162
+ },
+ {
+ "epoch": 2.2655021834061135,
+ "grad_norm": 0.016341188922524452,
+ "learning_rate": 0.0006,
+ "loss": 7.273357391357422,
+ "step": 163
+ },
+ {
+ "epoch": 2.279475982532751,
+ "grad_norm": 0.016125807538628578,
+ "learning_rate": 0.0006,
+ "loss": 7.273510932922363,
+ "step": 164
+ },
+ {
+ "epoch": 2.2934497816593886,
+ "grad_norm": 0.01629287749528885,
+ "learning_rate": 0.0006,
+ "loss": 7.267029762268066,
+ "step": 165
+ },
+ {
+ "epoch": 2.3074235807860264,
+ "grad_norm": 0.015296131372451782,
+ "learning_rate": 0.0006,
+ "loss": 7.211989402770996,
+ "step": 166
+ },
+ {
+ "epoch": 2.3213973799126637,
+ "grad_norm": 0.01709570363163948,
+ "learning_rate": 0.0006,
+ "loss": 7.242851734161377,
+ "step": 167
+ },
+ {
+ "epoch": 2.3353711790393015,
+ "grad_norm": 0.012552388943731785,
+ "learning_rate": 0.0006,
+ "loss": 7.241696834564209,
+ "step": 168
+ },
+ {
+ "epoch": 2.349344978165939,
+ "grad_norm": 0.014571291394531727,
+ "learning_rate": 0.0006,
+ "loss": 7.270681381225586,
+ "step": 169
+ },
+ {
+ "epoch": 2.3633187772925766,
+ "grad_norm": 0.013389437459409237,
+ "learning_rate": 0.0006,
+ "loss": 7.191600799560547,
+ "step": 170
+ },
+ {
+ "epoch": 2.377292576419214,
+ "grad_norm": 0.011895314790308475,
+ "learning_rate": 0.0006,
+ "loss": 7.213610649108887,
+ "step": 171
+ },
+ {
+ "epoch": 2.3912663755458516,
+ "grad_norm": 0.01345642190426588,
+ "learning_rate": 0.0006,
+ "loss": 7.193787574768066,
+ "step": 172
+ },
+ {
+ "epoch": 2.405240174672489,
+ "grad_norm": 0.0091794328764081,
+ "learning_rate": 0.0006,
+ "loss": 7.233364105224609,
+ "step": 173
+ },
+ {
+ "epoch": 2.4192139737991267,
+ "grad_norm": 0.013767086900770664,
+ "learning_rate": 0.0006,
+ "loss": 7.217272758483887,
+ "step": 174
+ },
+ {
+ "epoch": 2.433187772925764,
+ "grad_norm": 0.015134960412979126,
+ "learning_rate": 0.0006,
+ "loss": 7.192791938781738,
+ "step": 175
+ },
+ {
+ "epoch": 2.447161572052402,
+ "grad_norm": 0.01790648326277733,
+ "learning_rate": 0.0006,
+ "loss": 7.229467391967773,
+ "step": 176
+ },
+ {
+ "epoch": 2.461135371179039,
+ "grad_norm": 0.015362377278506756,
+ "learning_rate": 0.0006,
+ "loss": 7.180346488952637,
+ "step": 177
+ },
+ {
+ "epoch": 2.475109170305677,
+ "grad_norm": 0.010505498386919498,
+ "learning_rate": 0.0006,
+ "loss": 7.188833236694336,
+ "step": 178
+ },
+ {
+ "epoch": 2.489082969432314,
+ "grad_norm": 0.009129747748374939,
+ "learning_rate": 0.0006,
+ "loss": 7.203153133392334,
+ "step": 179
+ },
+ {
+ "epoch": 2.503056768558952,
+ "grad_norm": 0.01147883478552103,
+ "learning_rate": 0.0006,
+ "loss": 7.163824558258057,
+ "step": 180
+ },
+ {
+ "epoch": 2.5170305676855893,
+ "grad_norm": 0.011338942684233189,
+ "learning_rate": 0.0006,
+ "loss": 7.16287899017334,
+ "step": 181
+ },
+ {
+ "epoch": 2.531004366812227,
+ "grad_norm": 0.013277675025165081,
+ "learning_rate": 0.0006,
+ "loss": 7.155905723571777,
+ "step": 182
+ },
+ {
+ "epoch": 2.544978165938865,
+ "grad_norm": 0.015269347466528416,
+ "learning_rate": 0.0006,
+ "loss": 7.189371109008789,
+ "step": 183
+ },
+ {
+ "epoch": 2.558951965065502,
+ "grad_norm": 0.017905596643686295,
+ "learning_rate": 0.0006,
+ "loss": 7.177547931671143,
+ "step": 184
+ },
+ {
+ "epoch": 2.5729257641921395,
+ "grad_norm": 0.024431610479950905,
+ "learning_rate": 0.0006,
+ "loss": 7.192169189453125,
+ "step": 185
+ },
+ {
+ "epoch": 2.5868995633187772,
+ "grad_norm": 0.03656961768865585,
+ "learning_rate": 0.0006,
+ "loss": 7.124387741088867,
+ "step": 186
+ },
+ {
+ "epoch": 2.600873362445415,
+ "grad_norm": 0.04337048903107643,
+ "learning_rate": 0.0006,
+ "loss": 7.140231609344482,
+ "step": 187
+ },
+ {
+ "epoch": 2.6148471615720523,
+ "grad_norm": 0.04047228768467903,
+ "learning_rate": 0.0006,
+ "loss": 7.171927452087402,
+ "step": 188
+ },
+ {
+ "epoch": 2.62882096069869,
+ "grad_norm": 0.07825770974159241,
+ "learning_rate": 0.0006,
+ "loss": 7.207631587982178,
+ "step": 189
+ },
+ {
+ "epoch": 2.6427947598253274,
+ "grad_norm": 0.07118063420057297,
+ "learning_rate": 0.0006,
+ "loss": 7.209488391876221,
+ "step": 190
+ },
+ {
+ "epoch": 2.656768558951965,
+ "grad_norm": 0.05153293535113335,
+ "learning_rate": 0.0006,
+ "loss": 7.14793062210083,
+ "step": 191
+ },
+ {
+ "epoch": 2.6707423580786025,
+ "grad_norm": 0.03318839520215988,
+ "learning_rate": 0.0006,
+ "loss": 7.163010597229004,
+ "step": 192
+ },
+ {
+ "epoch": 2.6847161572052403,
+ "grad_norm": 0.029333539307117462,
+ "learning_rate": 0.0006,
+ "loss": 7.159955978393555,
+ "step": 193
+ },
+ {
+ "epoch": 2.6986899563318776,
+ "grad_norm": 0.029405880719423294,
+ "learning_rate": 0.0006,
+ "loss": 7.136178493499756,
+ "step": 194
+ },
+ {
+ "epoch": 2.7126637554585153,
+ "grad_norm": 0.037815988063812256,
+ "learning_rate": 0.0006,
+ "loss": 7.183186054229736,
+ "step": 195
+ },
+ {
+ "epoch": 2.726637554585153,
+ "grad_norm": 0.020307250320911407,
+ "learning_rate": 0.0006,
+ "loss": 7.133164882659912,
+ "step": 196
+ },
+ {
+ "epoch": 2.7406113537117904,
+ "grad_norm": 0.026048069819808006,
+ "learning_rate": 0.0006,
+ "loss": 7.124977111816406,
+ "step": 197
+ },
+ {
+ "epoch": 2.7545851528384278,
+ "grad_norm": 0.027316724881529808,
+ "learning_rate": 0.0006,
+ "loss": 7.07265567779541,
+ "step": 198
+ },
+ {
+ "epoch": 2.7685589519650655,
+ "grad_norm": 0.021629009395837784,
+ "learning_rate": 0.0006,
+ "loss": 7.150020599365234,
+ "step": 199
+ },
+ {
+ "epoch": 2.7825327510917033,
+ "grad_norm": 0.01605929434299469,
+ "learning_rate": 0.0006,
+ "loss": 7.10919189453125,
+ "step": 200
+ },
+ {
+ "epoch": 2.7965065502183406,
+ "grad_norm": 0.020800089463591576,
+ "learning_rate": 0.0006,
+ "loss": 7.132045269012451,
+ "step": 201
+ },
+ {
+ "epoch": 2.810480349344978,
+ "grad_norm": 0.019225774332880974,
+ "learning_rate": 0.0006,
+ "loss": 7.128364562988281,
+ "step": 202
+ },
+ {
+ "epoch": 2.8244541484716157,
+ "grad_norm": 0.01561372634023428,
+ "learning_rate": 0.0006,
+ "loss": 7.083669662475586,
+ "step": 203
+ },
+ {
+ "epoch": 2.8384279475982535,
+ "grad_norm": 0.021969519555568695,
+ "learning_rate": 0.0006,
+ "loss": 7.050841331481934,
+ "step": 204
+ },
+ {
+ "epoch": 2.8524017467248908,
+ "grad_norm": 0.021453695371747017,
+ "learning_rate": 0.0006,
+ "loss": 7.097085475921631,
+ "step": 205
+ },
+ {
+ "epoch": 2.8663755458515285,
+ "grad_norm": 0.022836022078990936,
+ "learning_rate": 0.0006,
+ "loss": 7.10405969619751,
+ "step": 206
+ },
+ {
+ "epoch": 2.880349344978166,
+ "grad_norm": 0.02430851384997368,
+ "learning_rate": 0.0006,
+ "loss": 7.045181751251221,
+ "step": 207
+ },
+ {
+ "epoch": 2.8943231441048036,
+ "grad_norm": 0.020166993141174316,
+ "learning_rate": 0.0006,
+ "loss": 7.068183898925781,
+ "step": 208
+ },
+ {
+ "epoch": 2.908296943231441,
+ "grad_norm": 0.0118505684658885,
+ "learning_rate": 0.0006,
+ "loss": 7.07132625579834,
+ "step": 209
+ },
+ {
+ "epoch": 2.9222707423580787,
+ "grad_norm": 0.023772427812218666,
+ "learning_rate": 0.0006,
+ "loss": 7.055238723754883,
+ "step": 210
+ },
+ {
+ "epoch": 2.936244541484716,
+ "grad_norm": 0.02350712940096855,
+ "learning_rate": 0.0006,
+ "loss": 7.055898666381836,
+ "step": 211
+ },
+ {
+ "epoch": 2.950218340611354,
+ "grad_norm": 0.017244907096028328,
+ "learning_rate": 0.0006,
+ "loss": 7.0099711418151855,
+ "step": 212
+ },
+ {
+ "epoch": 2.964192139737991,
+ "grad_norm": 0.021565575152635574,
+ "learning_rate": 0.0006,
+ "loss": 6.991751670837402,
+ "step": 213
+ },
+ {
+ "epoch": 2.978165938864629,
+ "grad_norm": 0.03241860866546631,
+ "learning_rate": 0.0006,
+ "loss": 7.005980491638184,
+ "step": 214
+ },
+ {
+ "epoch": 2.992139737991266,
+ "grad_norm": 0.04613726586103439,
+ "learning_rate": 0.0006,
+ "loss": 7.027859687805176,
+ "step": 215
+ },
+ {
+ "epoch": 3.0,
+ "grad_norm": 0.06404894590377808,
+ "learning_rate": 0.0006,
+ "loss": 7.123725891113281,
+ "step": 216
+ },
+ {
+ "epoch": 3.0,
+ "eval_loss": 7.070647716522217,
+ "eval_runtime": 58.5801,
+ "eval_samples_per_second": 41.687,
+ "eval_steps_per_second": 1.314,
+ "step": 216
+ },
+ {
+ "epoch": 3.0139737991266378,
+ "grad_norm": 0.05542432889342308,
+ "learning_rate": 0.0006,
+ "loss": 7.078299522399902,
+ "step": 217
+ },
+ {
+ "epoch": 3.027947598253275,
+ "grad_norm": 0.026484526693820953,
+ "learning_rate": 0.0006,
+ "loss": 6.961159706115723,
+ "step": 218
+ },
+ {
+ "epoch": 3.041921397379913,
+ "grad_norm": 0.028276391327381134,
+ "learning_rate": 0.0006,
+ "loss": 7.019961833953857,
+ "step": 219
+ },
+ {
+ "epoch": 3.05589519650655,
+ "grad_norm": 0.024486595764756203,
+ "learning_rate": 0.0006,
+ "loss": 7.039677619934082,
+ "step": 220
+ },
+ {
+ "epoch": 3.069868995633188,
+ "grad_norm": 0.017620893195271492,
+ "learning_rate": 0.0006,
+ "loss": 6.9270782470703125,
+ "step": 221
+ },
+ {
+ "epoch": 3.0838427947598253,
+ "grad_norm": 0.02547302283346653,
+ "learning_rate": 0.0006,
+ "loss": 6.924233436584473,
+ "step": 222
+ },
+ {
+ "epoch": 3.097816593886463,
+ "grad_norm": 0.02539009042084217,
+ "learning_rate": 0.0006,
+ "loss": 6.973979949951172,
+ "step": 223
+ },
+ {
+ "epoch": 3.1117903930131003,
+ "grad_norm": 0.03256227448582649,
+ "learning_rate": 0.0006,
+ "loss": 6.933725357055664,
+ "step": 224
+ },
+ {
+ "epoch": 3.125764192139738,
+ "grad_norm": 0.04651800915598869,
+ "learning_rate": 0.0006,
+ "loss": 6.951044082641602,
+ "step": 225
+ },
+ {
+ "epoch": 3.1397379912663754,
+ "grad_norm": 0.04822975769639015,
+ "learning_rate": 0.0006,
+ "loss": 6.9614410400390625,
+ "step": 226
+ },
+ {
+ "epoch": 3.153711790393013,
+ "grad_norm": 0.031165020540356636,
+ "learning_rate": 0.0006,
+ "loss": 6.91193151473999,
+ "step": 227
+ },
+ {
+ "epoch": 3.1676855895196505,
+ "grad_norm": 0.029754292219877243,
+ "learning_rate": 0.0006,
+ "loss": 6.91802978515625,
+ "step": 228
+ },
+ {
+ "epoch": 3.1816593886462883,
+ "grad_norm": 0.02193152718245983,
+ "learning_rate": 0.0006,
+ "loss": 6.928090572357178,
+ "step": 229
+ },
+ {
+ "epoch": 3.1956331877729256,
+ "grad_norm": 0.02428470179438591,
+ "learning_rate": 0.0006,
+ "loss": 6.90396785736084,
+ "step": 230
+ },
+ {
+ "epoch": 3.2096069868995634,
+ "grad_norm": 0.023374345153570175,
+ "learning_rate": 0.0006,
+ "loss": 6.885858535766602,
+ "step": 231
+ },
+ {
+ "epoch": 3.2235807860262007,
+ "grad_norm": 0.023405000567436218,
+ "learning_rate": 0.0006,
+ "loss": 6.8855767250061035,
+ "step": 232
+ },
+ {
+ "epoch": 3.2375545851528384,
+ "grad_norm": 0.017277134582400322,
+ "learning_rate": 0.0006,
+ "loss": 6.901449203491211,
+ "step": 233
+ },
+ {
+ "epoch": 3.251528384279476,
+ "grad_norm": 0.019496750086545944,
+ "learning_rate": 0.0006,
+ "loss": 6.885687828063965,
+ "step": 234
+ },
+ {
+ "epoch": 3.2655021834061135,
+ "grad_norm": 0.016893276944756508,
+ "learning_rate": 0.0006,
+ "loss": 6.872369766235352,
+ "step": 235
+ },
+ {
+ "epoch": 3.279475982532751,
+ "grad_norm": 0.02058715932071209,
+ "learning_rate": 0.0006,
+ "loss": 6.823280334472656,
+ "step": 236
+ },
+ {
+ "epoch": 3.2934497816593886,
+ "grad_norm": 0.015530755743384361,
+ "learning_rate": 0.0006,
+ "loss": 6.860499858856201,
+ "step": 237
+ },
+ {
+ "epoch": 3.3074235807860264,
+ "grad_norm": 0.01934926211833954,
+ "learning_rate": 0.0006,
+ "loss": 6.817448616027832,
+ "step": 238
+ },
+ {
+ "epoch": 3.3213973799126637,
+ "grad_norm": 0.017413552850484848,
+ "learning_rate": 0.0006,
+ "loss": 6.881955146789551,
+ "step": 239
+ },
+ {
+ "epoch": 3.3353711790393015,
+ "grad_norm": 0.026478875428438187,
+ "learning_rate": 0.0006,
+ "loss": 6.827404499053955,
+ "step": 240
+ },
+ {
+ "epoch": 3.349344978165939,
+ "grad_norm": 0.032611701637506485,
+ "learning_rate": 0.0006,
+ "loss": 6.78285551071167,
+ "step": 241
+ },
+ {
+ "epoch": 3.3633187772925766,
+ "grad_norm": 0.041900306940078735,
+ "learning_rate": 0.0006,
+ "loss": 6.783053874969482,
+ "step": 242
+ },
+ {
+ "epoch": 3.377292576419214,
+ "grad_norm": 0.04149497672915459,
+ "learning_rate": 0.0006,
+ "loss": 6.745220184326172,
+ "step": 243
+ },
+ {
+ "epoch": 3.3912663755458516,
+ "grad_norm": 0.023949826136231422,
+ "learning_rate": 0.0006,
+ "loss": 6.841533184051514,
+ "step": 244
+ },
+ {
+ "epoch": 3.405240174672489,
+ "grad_norm": 0.028601842001080513,
+ "learning_rate": 0.0006,
+ "loss": 6.852852821350098,
+ "step": 245
+ },
+ {
+ "epoch": 3.4192139737991267,
+ "grad_norm": 0.027365155518054962,
+ "learning_rate": 0.0006,
+ "loss": 6.74376106262207,
+ "step": 246
+ },
+ {
+ "epoch": 3.433187772925764,
+ "grad_norm": 0.02856568619608879,
+ "learning_rate": 0.0006,
+ "loss": 6.74906063079834,
+ "step": 247
+ },
+ {
+ "epoch": 3.447161572052402,
+ "grad_norm": 0.019172416999936104,
+ "learning_rate": 0.0006,
+ "loss": 6.809206008911133,
+ "step": 248
+ },
+ {
+ "epoch": 3.461135371179039,
+ "grad_norm": 0.02161746844649315,
+ "learning_rate": 0.0006,
+ "loss": 6.789888381958008,
+ "step": 249
+ },
+ {
+ "epoch": 3.475109170305677,
+ "grad_norm": 0.020273666828870773,
+ "learning_rate": 0.0006,
+ "loss": 6.73512077331543,
+ "step": 250
+ },
+ {
+ "epoch": 3.489082969432314,
+ "grad_norm": 0.022329283878207207,
+ "learning_rate": 0.0006,
+ "loss": 6.826148509979248,
+ "step": 251
+ },
+ {
+ "epoch": 3.503056768558952,
+ "grad_norm": 0.02284284122288227,
+ "learning_rate": 0.0006,
+ "loss": 6.748521327972412,
+ "step": 252
+ },
+ {
+ "epoch": 3.5170305676855893,
+ "grad_norm": 0.026368247345089912,
+ "learning_rate": 0.0006,
+ "loss": 6.645879745483398,
+ "step": 253
+ },
+ {
+ "epoch": 3.531004366812227,
+ "grad_norm": 0.03627244383096695,
+ "learning_rate": 0.0006,
+ "loss": 6.666064739227295,
+ "step": 254
+ },
+ {
+ "epoch": 3.544978165938865,
+ "grad_norm": 0.04140935093164444,
+ "learning_rate": 0.0006,
+ "loss": 6.712275981903076,
+ "step": 255
+ },
+ {
+ "epoch": 3.558951965065502,
+ "grad_norm": 0.0390462800860405,
+ "learning_rate": 0.0006,
+ "loss": 6.665900230407715,
+ "step": 256
+ },
+ {
+ "epoch": 3.5729257641921395,
+ "grad_norm": 0.05178583785891533,
+ "learning_rate": 0.0006,
+ "loss": 6.753298759460449,
+ "step": 257
+ },
+ {
+ "epoch": 3.5868995633187772,
+ "grad_norm": 0.054328061640262604,
+ "learning_rate": 0.0006,
+ "loss": 6.732633113861084,
+ "step": 258
+ },
+ {
+ "epoch": 3.600873362445415,
+ "grad_norm": 0.04590460658073425,
+ "learning_rate": 0.0006,
+ "loss": 6.65250301361084,
+ "step": 259
+ },
+ {
+ "epoch": 3.6148471615720523,
+ "grad_norm": 0.044331666082143784,
+ "learning_rate": 0.0006,
+ "loss": 6.68654203414917,
+ "step": 260
+ },
+ {
+ "epoch": 3.62882096069869,
+ "grad_norm": 0.05308730527758598,
+ "learning_rate": 0.0006,
+ "loss": 6.642482757568359,
+ "step": 261
+ },
+ {
+ "epoch": 3.6427947598253274,
+ "grad_norm": 0.03999016433954239,
+ "learning_rate": 0.0006,
+ "loss": 6.625227451324463,
+ "step": 262
+ },
+ {
+ "epoch": 3.656768558951965,
+ "grad_norm": 0.04088086634874344,
+ "learning_rate": 0.0006,
+ "loss": 6.67623233795166,
+ "step": 263
+ },
+ {
+ "epoch": 3.6707423580786025,
+ "grad_norm": 0.03615636005997658,
+ "learning_rate": 0.0006,
+ "loss": 6.6869001388549805,
+ "step": 264
+ },
+ {
+ "epoch": 3.6847161572052403,
+ "grad_norm": 0.02817360684275627,
+ "learning_rate": 0.0006,
+ "loss": 6.674601078033447,
+ "step": 265
+ },
+ {
+ "epoch": 3.6986899563318776,
+ "grad_norm": 0.032586123794317245,
+ "learning_rate": 0.0006,
+ "loss": 6.612994194030762,
+ "step": 266
+ },
+ {
+ "epoch": 3.7126637554585153,
+ "grad_norm": 0.031117867678403854,
+ "learning_rate": 0.0006,
+ "loss": 6.6244354248046875,
+ "step": 267
+ },
+ {
+ "epoch": 3.726637554585153,
+ "grad_norm": 0.028330687433481216,
+ "learning_rate": 0.0006,
+ "loss": 6.669557094573975,
+ "step": 268
+ },
+ {
+ "epoch": 3.7406113537117904,
+ "grad_norm": 0.021489202976226807,
+ "learning_rate": 0.0006,
+ "loss": 6.632035732269287,
+ "step": 269
+ },
+ {
+ "epoch": 3.7545851528384278,
+ "grad_norm": 0.024269424378871918,
+ "learning_rate": 0.0006,
+ "loss": 6.660867691040039,
+ "step": 270
+ },
+ {
+ "epoch": 3.7685589519650655,
+ "grad_norm": 0.02237873338162899,
+ "learning_rate": 0.0006,
+ "loss": 6.602587699890137,
+ "step": 271
+ },
+ {
+ "epoch": 3.7825327510917033,
+ "grad_norm": 0.026503929868340492,
+ "learning_rate": 0.0006,
+ "loss": 6.599200248718262,
+ "step": 272
+ },
+ {
+ "epoch": 3.7965065502183406,
+ "grad_norm": 0.03687124326825142,
+ "learning_rate": 0.0006,
+ "loss": 6.567296504974365,
+ "step": 273
+ },
+ {
+ "epoch": 3.810480349344978,
+ "grad_norm": 0.04708952456712723,
+ "learning_rate": 0.0006,
+ "loss": 6.621265411376953,
+ "step": 274
+ },
+ {
+ "epoch": 3.8244541484716157,
+ "grad_norm": 0.041846614331007004,
+ "learning_rate": 0.0006,
+ "loss": 6.550536155700684,
+ "step": 275
+ },
+ {
+ "epoch": 3.8384279475982535,
+ "grad_norm": 0.031249675899744034,
+ "learning_rate": 0.0006,
+ "loss": 6.54278564453125,
+ "step": 276
+ },
+ {
+ "epoch": 3.8524017467248908,
+ "grad_norm": 0.023082256317138672,
+ "learning_rate": 0.0006,
+ "loss": 6.544787406921387,
+ "step": 277
+ },
+ {
+ "epoch": 3.8663755458515285,
+ "grad_norm": 0.020995570346713066,
+ "learning_rate": 0.0006,
+ "loss": 6.518143653869629,
+ "step": 278
+ },
+ {
+ "epoch": 3.880349344978166,
+ "grad_norm": 0.020839324221014977,
+ "learning_rate": 0.0006,
+ "loss": 6.546667098999023,
+ "step": 279
+ },
+ {
+ "epoch": 3.8943231441048036,
+ "grad_norm": 0.018377432599663734,
+ "learning_rate": 0.0006,
+ "loss": 6.478307723999023,
+ "step": 280
+ },
+ {
+ "epoch": 3.908296943231441,
+ "grad_norm": 0.021713724359869957,
+ "learning_rate": 0.0006,
+ "loss": 6.485574245452881,
+ "step": 281
+ },
+ {
+ "epoch": 3.9222707423580787,
+ "grad_norm": 0.01359301246702671,
+ "learning_rate": 0.0006,
+ "loss": 6.570268630981445,
+ "step": 282
+ },
+ {
+ "epoch": 3.936244541484716,
+ "grad_norm": 0.016233332455158234,
+ "learning_rate": 0.0006,
+ "loss": 6.57150936126709,
+ "step": 283
+ },
+ {
+ "epoch": 3.950218340611354,
+ "grad_norm": 0.020687608048319817,
+ "learning_rate": 0.0006,
+ "loss": 6.527956008911133,
+ "step": 284
+ },
+ {
+ "epoch": 3.964192139737991,
+ "grad_norm": 0.02032964862883091,
+ "learning_rate": 0.0006,
+ "loss": 6.468169689178467,
+ "step": 285
+ },
+ {
+ "epoch": 3.978165938864629,
+ "grad_norm": 0.016676288098096848,
+ "learning_rate": 0.0006,
+ "loss": 6.448418617248535,
+ "step": 286
+ },
+ {
+ "epoch": 3.992139737991266,
+ "grad_norm": 0.0168539360165596,
+ "learning_rate": 0.0006,
+ "loss": 6.461180686950684,
+ "step": 287
+ },
+ {
+ "epoch": 4.0,
+ "grad_norm": 0.02666233666241169,
+ "learning_rate": 0.0006,
+ "loss": 6.344893932342529,
+ "step": 288
+ },
+ {
+ "epoch": 4.0,
+ "eval_loss": 6.5384979248046875,
+ "eval_runtime": 58.814,
+ "eval_samples_per_second": 41.521,
+ "eval_steps_per_second": 1.309,
+ "step": 288
+ },
+ {
+ "epoch": 4.013973799126638,
+ "grad_norm": 0.04154708981513977,
+ "learning_rate": 0.0006,
+ "loss": 6.41855525970459,
+ "step": 289
+ },
+ {
+ "epoch": 4.0279475982532755,
+ "grad_norm": 0.04895668104290962,
+ "learning_rate": 0.0006,
+ "loss": 6.499302864074707,
+ "step": 290
+ },
+ {
+ "epoch": 4.041921397379912,
+ "grad_norm": 0.034464482218027115,
+ "learning_rate": 0.0006,
+ "loss": 6.438611030578613,
+ "step": 291
+ },
+ {
+ "epoch": 4.05589519650655,
+ "grad_norm": 0.06845773756504059,
+ "learning_rate": 0.0006,
+ "loss": 6.538361072540283,
+ "step": 292
+ },
+ {
+ "epoch": 4.069868995633188,
+ "grad_norm": 0.053768858313560486,
+ "learning_rate": 0.0006,
+ "loss": 6.527524948120117,
+ "step": 293
+ },
+ {
+ "epoch": 4.083842794759826,
+ "grad_norm": 0.039037931710481644,
+ "learning_rate": 0.0006,
+ "loss": 6.503847122192383,
+ "step": 294
+ },
+ {
+ "epoch": 4.097816593886463,
+ "grad_norm": 0.03962196782231331,
+ "learning_rate": 0.0006,
+ "loss": 6.475178241729736,
+ "step": 295
+ },
+ {
+ "epoch": 4.1117903930131,
+ "grad_norm": 0.049101535230875015,
+ "learning_rate": 0.0006,
+ "loss": 6.474587917327881,
+ "step": 296
+ },
+ {
+ "epoch": 4.125764192139738,
+ "grad_norm": 0.02830282226204872,
+ "learning_rate": 0.0006,
+ "loss": 6.45065975189209,
+ "step": 297
+ },
+ {
+ "epoch": 4.139737991266376,
+ "grad_norm": 0.023153482005000114,
+ "learning_rate": 0.0006,
+ "loss": 6.5218706130981445,
+ "step": 298
+ },
+ {
+ "epoch": 4.153711790393013,
+ "grad_norm": 0.025064002722501755,
+ "learning_rate": 0.0006,
+ "loss": 6.456475257873535,
+ "step": 299
+ },
+ {
+ "epoch": 4.1676855895196505,
+ "grad_norm": 0.02357092685997486,
+ "learning_rate": 0.0006,
+ "loss": 6.485952377319336,
+ "step": 300
+ },
+ {
+ "epoch": 4.181659388646288,
+ "grad_norm": 0.021832305938005447,
+ "learning_rate": 0.0006,
+ "loss": 6.425506591796875,
+ "step": 301
+ },
+ {
+ "epoch": 4.195633187772926,
+ "grad_norm": 0.0227900929749012,
+ "learning_rate": 0.0006,
+ "loss": 6.439155578613281,
+ "step": 302
+ },
+ {
+ "epoch": 4.209606986899563,
+ "grad_norm": 0.015961607918143272,
+ "learning_rate": 0.0006,
+ "loss": 6.460862636566162,
+ "step": 303
+ },
+ {
+ "epoch": 4.223580786026201,
+ "grad_norm": 0.017876390367746353,
+ "learning_rate": 0.0006,
+ "loss": 6.44564151763916,
+ "step": 304
+ },
+ {
+ "epoch": 4.2375545851528384,
+ "grad_norm": 0.013703204691410065,
+ "learning_rate": 0.0006,
+ "loss": 6.418414115905762,
+ "step": 305
+ },
+ {
+ "epoch": 4.251528384279476,
+ "grad_norm": 0.015240306034684181,
+ "learning_rate": 0.0006,
+ "loss": 6.411238670349121,
+ "step": 306
+ },
+ {
+ "epoch": 4.265502183406113,
+ "grad_norm": 0.01587662845849991,
+ "learning_rate": 0.0006,
+ "loss": 6.395642280578613,
+ "step": 307
+ },
+ {
+ "epoch": 4.279475982532751,
+ "grad_norm": 0.014650700613856316,
+ "learning_rate": 0.0006,
+ "loss": 6.302793025970459,
+ "step": 308
+ },
+ {
+ "epoch": 4.293449781659389,
+ "grad_norm": 0.016533225774765015,
+ "learning_rate": 0.0006,
+ "loss": 6.380588054656982,
+ "step": 309
+ },
+ {
+ "epoch": 4.307423580786026,
+ "grad_norm": 0.019963741302490234,
+ "learning_rate": 0.0006,
+ "loss": 6.345336437225342,
+ "step": 310
+ },
+ {
+ "epoch": 4.321397379912664,
+ "grad_norm": 0.02237936295568943,
+ "learning_rate": 0.0006,
+ "loss": 6.387770652770996,
+ "step": 311
+ },
+ {
+ "epoch": 4.335371179039301,
+ "grad_norm": 0.026658328250050545,
+ "learning_rate": 0.0006,
+ "loss": 6.323662757873535,
+ "step": 312
+ },
+ {
+ "epoch": 4.349344978165939,
+ "grad_norm": 0.02852284163236618,
+ "learning_rate": 0.0006,
+ "loss": 6.382204055786133,
+ "step": 313
+ },
+ {
+ "epoch": 4.3633187772925766,
+ "grad_norm": 0.028815090656280518,
+ "learning_rate": 0.0006,
+ "loss": 6.402390480041504,
+ "step": 314
+ },
+ {
+ "epoch": 4.377292576419214,
+ "grad_norm": 0.029393529519438744,
+ "learning_rate": 0.0006,
+ "loss": 6.280278205871582,
+ "step": 315
+ },
+ {
+ "epoch": 4.391266375545851,
+ "grad_norm": 0.02186041697859764,
+ "learning_rate": 0.0006,
+ "loss": 6.302707672119141,
+ "step": 316
+ },
+ {
+ "epoch": 4.405240174672489,
+ "grad_norm": 0.02071218006312847,
+ "learning_rate": 0.0006,
+ "loss": 6.3365020751953125,
+ "step": 317
+ },
+ {
+ "epoch": 4.419213973799127,
+ "grad_norm": 0.02779117226600647,
+ "learning_rate": 0.0006,
+ "loss": 6.252157211303711,
+ "step": 318
+ },
+ {
+ "epoch": 4.4331877729257645,
+ "grad_norm": 0.03470654785633087,
+ "learning_rate": 0.0006,
+ "loss": 6.319070339202881,
+ "step": 319
+ },
+ {
+ "epoch": 4.447161572052401,
+ "grad_norm": 0.037067804485559464,
+ "learning_rate": 0.0006,
+ "loss": 6.314116477966309,
+ "step": 320
+ },
+ {
+ "epoch": 4.461135371179039,
+ "grad_norm": 0.032182756811380386,
+ "learning_rate": 0.0006,
+ "loss": 6.267298698425293,
+ "step": 321
+ },
+ {
+ "epoch": 4.475109170305677,
+ "grad_norm": 0.026305217295885086,
+ "learning_rate": 0.0006,
+ "loss": 6.331688404083252,
+ "step": 322
+ },
+ {
+ "epoch": 4.489082969432315,
+ "grad_norm": 0.027228357270359993,
+ "learning_rate": 0.0006,
+ "loss": 6.31139612197876,
+ "step": 323
+ },
+ {
+ "epoch": 4.503056768558952,
+ "grad_norm": 0.025781169533729553,
+ "learning_rate": 0.0006,
+ "loss": 6.194684982299805,
+ "step": 324
+ },
+ {
+ "epoch": 4.517030567685589,
+ "grad_norm": 0.029186977073550224,
+ "learning_rate": 0.0006,
+ "loss": 6.31306266784668,
+ "step": 325
+ },
+ {
+ "epoch": 4.531004366812227,
+ "grad_norm": 0.023608777672052383,
+ "learning_rate": 0.0006,
+ "loss": 6.285243034362793,
+ "step": 326
+ },
+ {
+ "epoch": 4.544978165938865,
+ "grad_norm": 0.02267594076693058,
+ "learning_rate": 0.0006,
+ "loss": 6.30342960357666,
+ "step": 327
+ },
+ {
+ "epoch": 4.558951965065502,
+ "grad_norm": 0.01926310732960701,
+ "learning_rate": 0.0006,
+ "loss": 6.280606269836426,
+ "step": 328
+ },
+ {
+ "epoch": 4.5729257641921395,
+ "grad_norm": 0.026484837755560875,
+ "learning_rate": 0.0006,
+ "loss": 6.237980365753174,
+ "step": 329
+ },
+ {
+ "epoch": 4.586899563318777,
+ "grad_norm": 0.026228854432702065,
+ "learning_rate": 0.0006,
+ "loss": 6.3018293380737305,
+ "step": 330
+ },
+ {
+ "epoch": 4.600873362445415,
+ "grad_norm": 0.022096967324614525,
+ "learning_rate": 0.0006,
+ "loss": 6.253863334655762,
+ "step": 331
+ },
+ {
+ "epoch": 4.614847161572053,
+ "grad_norm": 0.027223890647292137,
+ "learning_rate": 0.0006,
+ "loss": 6.166882038116455,
+ "step": 332
+ },
+ {
+ "epoch": 4.62882096069869,
+ "grad_norm": 0.03160124272108078,
+ "learning_rate": 0.0006,
+ "loss": 6.230309963226318,
+ "step": 333
+ },
+ {
+ "epoch": 4.642794759825327,
+ "grad_norm": 0.04110539332032204,
+ "learning_rate": 0.0006,
+ "loss": 6.2334818840026855,
+ "step": 334
+ },
+ {
+ "epoch": 4.656768558951965,
+ "grad_norm": 0.04355933889746666,
+ "learning_rate": 0.0006,
+ "loss": 6.160956382751465,
+ "step": 335
+ },
+ {
+ "epoch": 4.670742358078603,
+ "grad_norm": 0.03611086308956146,
+ "learning_rate": 0.0006,
+ "loss": 6.23399543762207,
+ "step": 336
+ },
+ {
+ "epoch": 4.68471615720524,
+ "grad_norm": 0.051811400800943375,
+ "learning_rate": 0.0006,
+ "loss": 6.241855144500732,
+ "step": 337
+ },
+ {
+ "epoch": 4.698689956331878,
+ "grad_norm": 0.057434193789958954,
+ "learning_rate": 0.0006,
+ "loss": 6.220993995666504,
+ "step": 338
+ },
+ {
+ "epoch": 4.712663755458515,
+ "grad_norm": 0.04420953616499901,
+ "learning_rate": 0.0006,
+ "loss": 6.300461292266846,
+ "step": 339
+ },
+ {
+ "epoch": 4.726637554585153,
+ "grad_norm": 0.03515457361936569,
+ "learning_rate": 0.0006,
+ "loss": 6.269349098205566,
+ "step": 340
+ },
+ {
+ "epoch": 4.74061135371179,
+ "grad_norm": 0.03222600743174553,
+ "learning_rate": 0.0006,
+ "loss": 6.248613357543945,
+ "step": 341
+ },
+ {
+ "epoch": 4.754585152838428,
+ "grad_norm": 0.0384046845138073,
+ "learning_rate": 0.0006,
+ "loss": 6.266979217529297,
+ "step": 342
+ },
+ {
+ "epoch": 4.7685589519650655,
+ "grad_norm": 0.0514867827296257,
+ "learning_rate": 0.0006,
+ "loss": 6.234103202819824,
+ "step": 343
+ },
+ {
+ "epoch": 4.782532751091703,
+ "grad_norm": 0.050747793167829514,
+ "learning_rate": 0.0006,
+ "loss": 6.322582244873047,
+ "step": 344
+ },
+ {
+ "epoch": 4.796506550218341,
+ "grad_norm": 0.040878988802433014,
+ "learning_rate": 0.0006,
+ "loss": 6.212509632110596,
+ "step": 345
+ },
+ {
+ "epoch": 4.810480349344978,
+ "grad_norm": 0.034655820578336716,
+ "learning_rate": 0.0006,
+ "loss": 6.265879154205322,
+ "step": 346
+ },
+ {
+ "epoch": 4.824454148471616,
+ "grad_norm": 0.04202224686741829,
+ "learning_rate": 0.0006,
+ "loss": 6.2896409034729,
+ "step": 347
+ },
+ {
+ "epoch": 4.8384279475982535,
+ "grad_norm": 0.03632277995347977,
+ "learning_rate": 0.0006,
+ "loss": 6.253917217254639,
+ "step": 348
+ },
+ {
+ "epoch": 4.85240174672489,
+ "grad_norm": 0.028645765036344528,
+ "learning_rate": 0.0006,
+ "loss": 6.207965850830078,
+ "step": 349
+ },
+ {
+ "epoch": 4.866375545851528,
+ "grad_norm": 0.025305170565843582,
+ "learning_rate": 0.0006,
+ "loss": 6.168068885803223,
+ "step": 350
+ },
+ {
+ "epoch": 4.880349344978166,
+ "grad_norm": 0.022137803956866264,
+ "learning_rate": 0.0006,
+ "loss": 6.235283374786377,
+ "step": 351
+ },
+ {
+ "epoch": 4.894323144104804,
+ "grad_norm": 0.01814538985490799,
+ "learning_rate": 0.0006,
+ "loss": 6.259641647338867,
+ "step": 352
+ },
+ {
+ "epoch": 4.908296943231441,
+ "grad_norm": 0.018465086817741394,
+ "learning_rate": 0.0006,
+ "loss": 6.111738681793213,
+ "step": 353
+ },
+ {
+ "epoch": 4.922270742358078,
+ "grad_norm": 0.018638933077454567,
+ "learning_rate": 0.0006,
+ "loss": 6.225642204284668,
+ "step": 354
+ },
+ {
+ "epoch": 4.936244541484716,
+ "grad_norm": 0.016461260616779327,
+ "learning_rate": 0.0006,
+ "loss": 6.140647888183594,
+ "step": 355
+ },
+ {
+ "epoch": 4.950218340611354,
+ "grad_norm": 0.015458385460078716,
+ "learning_rate": 0.0006,
+ "loss": 6.155224800109863,
+ "step": 356
+ },
+ {
+ "epoch": 4.964192139737992,
+ "grad_norm": 0.014934048056602478,
+ "learning_rate": 0.0006,
+ "loss": 6.101555824279785,
+ "step": 357
+ },
+ {
+ "epoch": 4.978165938864628,
+ "grad_norm": 0.013979545794427395,
+ "learning_rate": 0.0006,
+ "loss": 6.161718845367432,
+ "step": 358
+ },
+ {
+ "epoch": 4.992139737991266,
+ "grad_norm": 0.01640394888818264,
+ "learning_rate": 0.0006,
+ "loss": 6.055768966674805,
+ "step": 359
+ },
+ {
+ "epoch": 5.0,
+ "grad_norm": 0.015135680325329304,
+ "learning_rate": 0.0006,
+ "loss": 6.169695854187012,
+ "step": 360
+ },
+ {
+ "epoch": 5.0,
+ "eval_loss": 6.16408109664917,
+ "eval_runtime": 59.3649,
+ "eval_samples_per_second": 41.135,
+ "eval_steps_per_second": 1.297,
+ "step": 360
+ },
+ {
+ "epoch": 5.013973799126638,
+ "grad_norm": 0.012849048711359501,
+ "learning_rate": 0.0006,
+ "loss": 6.055830001831055,
+ "step": 361
+ },
+ {
+ "epoch": 5.0279475982532755,
+ "grad_norm": 0.018169576302170753,
+ "learning_rate": 0.0006,
+ "loss": 6.082655429840088,
+ "step": 362
+ },
+ {
+ "epoch": 5.041921397379912,
+ "grad_norm": 0.024293430149555206,
+ "learning_rate": 0.0006,
+ "loss": 6.109032154083252,
+ "step": 363
+ },
+ {
+ "epoch": 5.05589519650655,
+ "grad_norm": 0.03266787901520729,
+ "learning_rate": 0.0006,
+ "loss": 6.030613899230957,
+ "step": 364
+ },
+ {
+ "epoch": 5.069868995633188,
+ "grad_norm": 0.03572266176342964,
+ "learning_rate": 0.0006,
+ "loss": 6.050390720367432,
+ "step": 365
+ },
+ {
+ "epoch": 5.083842794759826,
+ "grad_norm": 0.032625213265419006,
+ "learning_rate": 0.0006,
+ "loss": 6.169313430786133,
+ "step": 366
+ },
+ {
+ "epoch": 5.097816593886463,
+ "grad_norm": 0.04051872715353966,
+ "learning_rate": 0.0006,
+ "loss": 6.10398006439209,
+ "step": 367
+ },
+ {
+ "epoch": 5.1117903930131,
+ "grad_norm": 0.0351213738322258,
+ "learning_rate": 0.0006,
+ "loss": 6.121973514556885,
+ "step": 368
+ },
+ {
+ "epoch": 5.125764192139738,
+ "grad_norm": 0.03523759916424751,
+ "learning_rate": 0.0006,
+ "loss": 6.0626654624938965,
+ "step": 369
+ },
+ {
+ "epoch": 5.139737991266376,
+ "grad_norm": 0.02706182189285755,
+ "learning_rate": 0.0006,
+ "loss": 6.049354076385498,
+ "step": 370
+ },
+ {
+ "epoch": 5.153711790393013,
+ "grad_norm": 0.024212589487433434,
+ "learning_rate": 0.0006,
+ "loss": 6.034826278686523,
+ "step": 371
+ },
+ {
+ "epoch": 5.1676855895196505,
+ "grad_norm": 0.024798082187771797,
+ "learning_rate": 0.0006,
+ "loss": 6.004058837890625,
+ "step": 372
+ },
+ {
+ "epoch": 5.181659388646288,
+ "grad_norm": 0.02325567416846752,
+ "learning_rate": 0.0006,
+ "loss": 5.986461639404297,
+ "step": 373
+ },
+ {
+ "epoch": 5.195633187772926,
+ "grad_norm": 0.019060570746660233,
+ "learning_rate": 0.0006,
+ "loss": 6.059736251831055,
+ "step": 374
+ },
+ {
+ "epoch": 5.209606986899563,
+ "grad_norm": 0.016822976991534233,
+ "learning_rate": 0.0006,
+ "loss": 6.072957515716553,
+ "step": 375
+ },
+ {
+ "epoch": 5.223580786026201,
+ "grad_norm": 0.017218658700585365,
+ "learning_rate": 0.0006,
+ "loss": 6.031739234924316,
+ "step": 376
+ },
+ {
+ "epoch": 5.2375545851528384,
+ "grad_norm": 0.014889383688569069,
+ "learning_rate": 0.0006,
+ "loss": 6.025674343109131,
+ "step": 377
+ },
+ {
+ "epoch": 5.251528384279476,
+ "grad_norm": 0.01708107627928257,
+ "learning_rate": 0.0006,
+ "loss": 6.090451240539551,
+ "step": 378
+ },
+ {
+ "epoch": 5.265502183406113,
+ "grad_norm": 0.020260317251086235,
+ "learning_rate": 0.0006,
+ "loss": 6.094257354736328,
+ "step": 379
+ },
+ {
+ "epoch": 5.279475982532751,
+ "grad_norm": 0.020110400393605232,
+ "learning_rate": 0.0006,
+ "loss": 6.021188735961914,
+ "step": 380
+ },
+ {
+ "epoch": 5.293449781659389,
+ "grad_norm": 0.02027830481529236,
+ "learning_rate": 0.0006,
+ "loss": 6.022156238555908,
+ "step": 381
+ },
+ {
+ "epoch": 5.307423580786026,
+ "grad_norm": 0.022747062146663666,
+ "learning_rate": 0.0006,
+ "loss": 6.011836051940918,
+ "step": 382
+ },
+ {
+ "epoch": 5.321397379912664,
+ "grad_norm": 0.01990230567753315,
+ "learning_rate": 0.0006,
+ "loss": 6.003054618835449,
+ "step": 383
+ },
+ {
+ "epoch": 5.335371179039301,
+ "grad_norm": 0.01596238650381565,
+ "learning_rate": 0.0006,
+ "loss": 5.991410732269287,
+ "step": 384
+ },
+ {
+ "epoch": 5.349344978165939,
+ "grad_norm": 0.015847105532884598,
+ "learning_rate": 0.0006,
+ "loss": 6.100622177124023,
+ "step": 385
+ },
+ {
+ "epoch": 5.3633187772925766,
+ "grad_norm": 0.016179397702217102,
+ "learning_rate": 0.0006,
+ "loss": 6.032659530639648,
+ "step": 386
+ },
+ {
+ "epoch": 5.377292576419214,
+ "grad_norm": 0.018256602808833122,
+ "learning_rate": 0.0006,
+ "loss": 5.961983680725098,
+ "step": 387
+ },
+ {
+ "epoch": 5.391266375545851,
+ "grad_norm": 0.02805912122130394,
+ "learning_rate": 0.0006,
+ "loss": 5.983541965484619,
+ "step": 388
+ },
+ {
+ "epoch": 5.405240174672489,
+ "grad_norm": 0.039082255214452744,
+ "learning_rate": 0.0006,
+ "loss": 6.021474361419678,
+ "step": 389
+ },
+ {
+ "epoch": 5.419213973799127,
+ "grad_norm": 0.036757659167051315,
+ "learning_rate": 0.0006,
+ "loss": 6.005046367645264,
+ "step": 390
+ },
+ {
+ "epoch": 5.4331877729257645,
+ "grad_norm": 0.035277366638183594,
+ "learning_rate": 0.0006,
+ "loss": 6.036296844482422,
+ "step": 391
+ },
+ {
+ "epoch": 5.447161572052401,
+ "grad_norm": 0.034404635429382324,
+ "learning_rate": 0.0006,
+ "loss": 5.994539260864258,
+ "step": 392
+ },
+ {
+ "epoch": 5.461135371179039,
+ "grad_norm": 0.0377206988632679,
+ "learning_rate": 0.0006,
+ "loss": 5.941534519195557,
+ "step": 393
+ },
+ {
+ "epoch": 5.475109170305677,
+ "grad_norm": 0.0389922633767128,
+ "learning_rate": 0.0006,
+ "loss": 5.983644485473633,
+ "step": 394
+ },
+ {
+ "epoch": 5.489082969432315,
+ "grad_norm": 0.04176778718829155,
+ "learning_rate": 0.0006,
+ "loss": 6.030922889709473,
+ "step": 395
+ },
+ {
+ "epoch": 5.503056768558952,
+ "grad_norm": 0.0337153784930706,
+ "learning_rate": 0.0006,
+ "loss": 5.961367130279541,
+ "step": 396
+ },
+ {
+ "epoch": 5.517030567685589,
+ "grad_norm": 0.036116816103458405,
+ "learning_rate": 0.0006,
+ "loss": 5.819280624389648,
+ "step": 397
+ },
+ {
+ "epoch": 5.531004366812227,
+ "grad_norm": 0.030724041163921356,
+ "learning_rate": 0.0006,
+ "loss": 5.906380653381348,
+ "step": 398
+ },
+ {
+ "epoch": 5.544978165938865,
+ "grad_norm": 0.030264202505350113,
+ "learning_rate": 0.0006,
+ "loss": 5.975833415985107,
+ "step": 399
+ },
+ {
+ "epoch": 5.558951965065502,
+ "grad_norm": 0.032096318900585175,
+ "learning_rate": 0.0006,
+ "loss": 6.001348972320557,
+ "step": 400
+ },
+ {
+ "epoch": 5.5729257641921395,
+ "grad_norm": 0.030579356476664543,
+ "learning_rate": 0.0006,
+ "loss": 5.928768634796143,
+ "step": 401
+ },
+ {
+ "epoch": 5.586899563318777,
+ "grad_norm": 0.028241973370313644,
+ "learning_rate": 0.0006,
+ "loss": 5.92582893371582,
+ "step": 402
+ },
+ {
+ "epoch": 5.600873362445415,
+ "grad_norm": 0.023086953908205032,
+ "learning_rate": 0.0006,
+ "loss": 5.892926216125488,
+ "step": 403
+ },
+ {
+ "epoch": 5.614847161572053,
+ "grad_norm": 0.02547992393374443,
+ "learning_rate": 0.0006,
+ "loss": 5.903195381164551,
+ "step": 404
+ },
+ {
+ "epoch": 5.62882096069869,
+ "grad_norm": 0.023089831694960594,
+ "learning_rate": 0.0006,
+ "loss": 5.912033557891846,
+ "step": 405
+ },
+ {
+ "epoch": 5.642794759825327,
+ "grad_norm": 0.025446368381381035,
+ "learning_rate": 0.0006,
+ "loss": 5.926138401031494,
+ "step": 406
+ },
+ {
+ "epoch": 5.656768558951965,
+ "grad_norm": 0.03748747706413269,
+ "learning_rate": 0.0006,
+ "loss": 5.913451194763184,
+ "step": 407
+ },
+ {
+ "epoch": 5.670742358078603,
+ "grad_norm": 0.03493810072541237,
+ "learning_rate": 0.0006,
+ "loss": 5.97633695602417,
+ "step": 408
+ },
+ {
+ "epoch": 5.68471615720524,
+ "grad_norm": 0.023131275549530983,
+ "learning_rate": 0.0006,
+ "loss": 5.899571418762207,
+ "step": 409
+ },
+ {
+ "epoch": 5.698689956331878,
+ "grad_norm": 0.02986014075577259,
+ "learning_rate": 0.0006,
+ "loss": 5.894903182983398,
+ "step": 410
+ },
+ {
+ "epoch": 5.712663755458515,
+ "grad_norm": 0.030171144753694534,
+ "learning_rate": 0.0006,
+ "loss": 5.876595497131348,
+ "step": 411
+ },
+ {
+ "epoch": 5.726637554585153,
+ "grad_norm": 0.029377546161413193,
+ "learning_rate": 0.0006,
+ "loss": 5.911637306213379,
+ "step": 412
+ },
+ {
+ "epoch": 5.74061135371179,
+ "grad_norm": 0.029829490929841995,
+ "learning_rate": 0.0006,
+ "loss": 5.819629669189453,
+ "step": 413
+ },
+ {
+ "epoch": 5.754585152838428,
+ "grad_norm": 0.024139299988746643,
+ "learning_rate": 0.0006,
+ "loss": 5.81728458404541,
+ "step": 414
+ },
+ {
+ "epoch": 5.7685589519650655,
+ "grad_norm": 0.021453432738780975,
+ "learning_rate": 0.0006,
+ "loss": 5.987326145172119,
+ "step": 415
+ },
+ {
+ "epoch": 5.782532751091703,
+ "grad_norm": 0.017942246049642563,
+ "learning_rate": 0.0006,
+ "loss": 5.893008232116699,
+ "step": 416
+ },
+ {
+ "epoch": 5.796506550218341,
+ "grad_norm": 0.019724637269973755,
+ "learning_rate": 0.0006,
+ "loss": 5.912441253662109,
+ "step": 417
+ },
+ {
+ "epoch": 5.810480349344978,
+ "grad_norm": 0.018597912043333054,
+ "learning_rate": 0.0006,
+ "loss": 5.860394477844238,
+ "step": 418
+ },
+ {
+ "epoch": 5.824454148471616,
+ "grad_norm": 0.016232412308454514,
+ "learning_rate": 0.0006,
+ "loss": 5.94991397857666,
+ "step": 419
+ },
+ {
+ "epoch": 5.8384279475982535,
+ "grad_norm": 0.01667204685509205,
+ "learning_rate": 0.0006,
+ "loss": 5.8506364822387695,
+ "step": 420
+ },
+ {
+ "epoch": 5.85240174672489,
+ "grad_norm": 0.014232151210308075,
+ "learning_rate": 0.0006,
+ "loss": 5.928424835205078,
+ "step": 421
+ },
+ {
+ "epoch": 5.866375545851528,
+ "grad_norm": 0.01570476032793522,
+ "learning_rate": 0.0006,
+ "loss": 5.953692436218262,
+ "step": 422
+ },
+ {
+ "epoch": 5.880349344978166,
+ "grad_norm": 0.01586179807782173,
+ "learning_rate": 0.0006,
+ "loss": 5.820014476776123,
+ "step": 423
+ },
+ {
+ "epoch": 5.894323144104804,
+ "grad_norm": 0.014729145914316177,
+ "learning_rate": 0.0006,
+ "loss": 5.860154628753662,
+ "step": 424
+ },
+ {
+ "epoch": 5.908296943231441,
+ "grad_norm": 0.013111459091305733,
+ "learning_rate": 0.0006,
+ "loss": 5.85988187789917,
+ "step": 425
+ },
+ {
+ "epoch": 5.922270742358078,
+ "grad_norm": 0.012937922962009907,
+ "learning_rate": 0.0006,
+ "loss": 5.756265640258789,
+ "step": 426
+ },
+ {
+ "epoch": 5.936244541484716,
+ "grad_norm": 0.01377453189343214,
+ "learning_rate": 0.0006,
+ "loss": 5.8333048820495605,
+ "step": 427
+ },
+ {
+ "epoch": 5.950218340611354,
+ "grad_norm": 0.014054795727133751,
+ "learning_rate": 0.0006,
+ "loss": 5.803333759307861,
+ "step": 428
+ },
+ {
+ "epoch": 5.964192139737992,
+ "grad_norm": 0.01674959622323513,
+ "learning_rate": 0.0006,
+ "loss": 5.8718485832214355,
+ "step": 429
+ },
+ {
+ "epoch": 5.978165938864628,
+ "grad_norm": 0.018733017146587372,
+ "learning_rate": 0.0006,
+ "loss": 5.875979900360107,
+ "step": 430
+ },
+ {
+ "epoch": 5.992139737991266,
+ "grad_norm": 0.02088983915746212,
+ "learning_rate": 0.0006,
+ "loss": 5.763634204864502,
+ "step": 431
+ },
+ {
+ "epoch": 6.0,
+ "grad_norm": 0.026705985888838768,
+ "learning_rate": 0.0006,
+ "loss": 5.6922478675842285,
+ "step": 432
+ },
+ {
+ "epoch": 6.0,
+ "eval_loss": 5.847958087921143,
+ "eval_runtime": 58.5451,
+ "eval_samples_per_second": 41.711,
+ "eval_steps_per_second": 1.315,
+ "step": 432
+ },
+ {
+ "epoch": 6.013973799126638,
+ "grad_norm": 0.03181544691324234,
+ "learning_rate": 0.0006,
+ "loss": 5.786368370056152,
+ "step": 433
+ },
+ {
+ "epoch": 6.0279475982532755,
+ "grad_norm": 0.03238112851977348,
+ "learning_rate": 0.0006,
+ "loss": 5.810310363769531,
+ "step": 434
+ },
+ {
+ "epoch": 6.041921397379912,
+ "grad_norm": 0.03916756808757782,
+ "learning_rate": 0.0006,
+ "loss": 5.7761993408203125,
+ "step": 435
+ },
+ {
+ "epoch": 6.05589519650655,
+ "grad_norm": 0.044009730219841,
+ "learning_rate": 0.0006,
+ "loss": 5.898112773895264,
+ "step": 436
+ },
+ {
+ "epoch": 6.069868995633188,
+ "grad_norm": 0.045235246419906616,
+ "learning_rate": 0.0006,
+ "loss": 5.876371383666992,
+ "step": 437
+ },
+ {
+ "epoch": 6.083842794759826,
+ "grad_norm": 0.050218384712934494,
+ "learning_rate": 0.0006,
+ "loss": 5.805103302001953,
+ "step": 438
+ },
+ {
+ "epoch": 6.097816593886463,
+ "grad_norm": 0.05444490164518356,
+ "learning_rate": 0.0006,
+ "loss": 5.826424598693848,
+ "step": 439
+ },
+ {
+ "epoch": 6.1117903930131,
+ "grad_norm": 0.04890982061624527,
+ "learning_rate": 0.0006,
+ "loss": 5.7735395431518555,
+ "step": 440
+ },
+ {
+ "epoch": 6.125764192139738,
+ "grad_norm": 0.05478639155626297,
+ "learning_rate": 0.0006,
+ "loss": 5.823677062988281,
+ "step": 441
+ },
+ {
+ "epoch": 6.139737991266376,
+ "grad_norm": 0.057562313973903656,
+ "learning_rate": 0.0006,
+ "loss": 5.889334678649902,
+ "step": 442
+ },
+ {
+ "epoch": 6.153711790393013,
+ "grad_norm": 0.06447703391313553,
+ "learning_rate": 0.0006,
+ "loss": 5.84621524810791,
+ "step": 443
+ },
+ {
+ "epoch": 6.1676855895196505,
+ "grad_norm": 0.048861872404813766,
+ "learning_rate": 0.0006,
+ "loss": 5.931595802307129,
+ "step": 444
+ },
+ {
+ "epoch": 6.181659388646288,
+ "grad_norm": 0.05521339178085327,
+ "learning_rate": 0.0006,
+ "loss": 5.8229780197143555,
+ "step": 445
+ },
+ {
+ "epoch": 6.195633187772926,
+ "grad_norm": 0.053666990250349045,
+ "learning_rate": 0.0006,
+ "loss": 5.879191875457764,
+ "step": 446
+ },
+ {
+ "epoch": 6.209606986899563,
+ "grad_norm": 0.0451025515794754,
+ "learning_rate": 0.0006,
+ "loss": 5.801628112792969,
+ "step": 447
+ },
+ {
+ "epoch": 6.223580786026201,
+ "grad_norm": 0.04291655868291855,
+ "learning_rate": 0.0006,
+ "loss": 5.914680480957031,
+ "step": 448
+ },
+ {
+ "epoch": 6.2375545851528384,
+ "grad_norm": 0.035102490335702896,
+ "learning_rate": 0.0006,
+ "loss": 5.843682765960693,
+ "step": 449
+ },
+ {
+ "epoch": 6.251528384279476,
+ "grad_norm": 0.03403995931148529,
+ "learning_rate": 0.0006,
+ "loss": 5.827154159545898,
+ "step": 450
+ },
+ {
+ "epoch": 6.265502183406113,
+ "grad_norm": 0.03444375470280647,
+ "learning_rate": 0.0006,
+ "loss": 5.800136566162109,
+ "step": 451
+ },
+ {
+ "epoch": 6.279475982532751,
+ "grad_norm": 0.03165159001946449,
+ "learning_rate": 0.0006,
+ "loss": 5.906252384185791,
+ "step": 452
+ },
+ {
+ "epoch": 6.293449781659389,
+ "grad_norm": 0.026153093203902245,
+ "learning_rate": 0.0006,
+ "loss": 5.7423200607299805,
+ "step": 453
+ },
+ {
+ "epoch": 6.307423580786026,
+ "grad_norm": 0.024357657879590988,
+ "learning_rate": 0.0006,
+ "loss": 5.701364517211914,
+ "step": 454
+ },
+ {
+ "epoch": 6.321397379912664,
+ "grad_norm": 0.021508049219846725,
+ "learning_rate": 0.0006,
+ "loss": 5.8328094482421875,
+ "step": 455
+ },
+ {
+ "epoch": 6.335371179039301,
+ "grad_norm": 0.017313921824097633,
+ "learning_rate": 0.0006,
+ "loss": 5.792145252227783,
+ "step": 456
+ },
+ {
+ "epoch": 6.349344978165939,
+ "grad_norm": 0.018563397228717804,
+ "learning_rate": 0.0006,
+ "loss": 5.732672214508057,
+ "step": 457
+ },
+ {
+ "epoch": 6.3633187772925766,
+ "grad_norm": 0.016568679362535477,
+ "learning_rate": 0.0006,
+ "loss": 5.81948184967041,
+ "step": 458
+ },
+ {
+ "epoch": 6.377292576419214,
+ "grad_norm": 0.014133770950138569,
+ "learning_rate": 0.0006,
+ "loss": 5.734982490539551,
+ "step": 459
+ },
+ {
+ "epoch": 6.391266375545851,
+ "grad_norm": 0.014114447869360447,
+ "learning_rate": 0.0006,
+ "loss": 5.70073127746582,
+ "step": 460
+ },
+ {
+ "epoch": 6.405240174672489,
+ "grad_norm": 0.012907393276691437,
+ "learning_rate": 0.0006,
+ "loss": 5.758626937866211,
+ "step": 461
+ },
+ {
+ "epoch": 6.419213973799127,
+ "grad_norm": 0.012673444114625454,
+ "learning_rate": 0.0006,
+ "loss": 5.775138854980469,
+ "step": 462
+ },
+ {
+ "epoch": 6.4331877729257645,
+ "grad_norm": 0.01311410591006279,
+ "learning_rate": 0.0006,
+ "loss": 5.70920991897583,
+ "step": 463
+ },
+ {
+ "epoch": 6.447161572052401,
+ "grad_norm": 0.012935544364154339,
+ "learning_rate": 0.0006,
+ "loss": 5.781479835510254,
+ "step": 464
+ },
+ {
+ "epoch": 6.461135371179039,
+ "grad_norm": 0.01439738366752863,
+ "learning_rate": 0.0006,
+ "loss": 5.691315650939941,
+ "step": 465
+ },
+ {
+ "epoch": 6.475109170305677,
+ "grad_norm": 0.01205186452716589,
+ "learning_rate": 0.0006,
+ "loss": 5.662499904632568,
+ "step": 466
+ },
+ {
+ "epoch": 6.489082969432315,
+ "grad_norm": 0.011575652286410332,
+ "learning_rate": 0.0006,
+ "loss": 5.6271772384643555,
+ "step": 467
+ },
+ {
+ "epoch": 6.503056768558952,
+ "grad_norm": 0.011672502383589745,
+ "learning_rate": 0.0006,
+ "loss": 5.761662006378174,
+ "step": 468
+ },
+ {
+ "epoch": 6.517030567685589,
+ "grad_norm": 0.011541751213371754,
+ "learning_rate": 0.0006,
+ "loss": 5.762078285217285,
+ "step": 469
+ },
+ {
+ "epoch": 6.531004366812227,
+ "grad_norm": 0.011396365240216255,
+ "learning_rate": 0.0006,
+ "loss": 5.67873477935791,
+ "step": 470
+ },
+ {
+ "epoch": 6.544978165938865,
+ "grad_norm": 0.010688499547541142,
+ "learning_rate": 0.0006,
+ "loss": 5.717051982879639,
+ "step": 471
+ },
+ {
+ "epoch": 6.558951965065502,
+ "grad_norm": 0.012224317528307438,
+ "learning_rate": 0.0006,
+ "loss": 5.707948684692383,
+ "step": 472
+ },
+ {
+ "epoch": 6.5729257641921395,
+ "grad_norm": 0.011856825090944767,
+ "learning_rate": 0.0006,
+ "loss": 5.70878791809082,
+ "step": 473
+ },
+ {
+ "epoch": 6.586899563318777,
+ "grad_norm": 0.01199064590036869,
+ "learning_rate": 0.0006,
+ "loss": 5.708697319030762,
+ "step": 474
+ },
+ {
+ "epoch": 6.600873362445415,
+ "grad_norm": 0.01219446212053299,
+ "learning_rate": 0.0006,
+ "loss": 5.6061577796936035,
+ "step": 475
+ },
+ {
+ "epoch": 6.614847161572053,
+ "grad_norm": 0.013397484086453915,
+ "learning_rate": 0.0006,
+ "loss": 5.624789714813232,
+ "step": 476
+ },
+ {
+ "epoch": 6.62882096069869,
+ "grad_norm": 0.01483136136084795,
+ "learning_rate": 0.0006,
+ "loss": 5.721141338348389,
+ "step": 477
+ },
+ {
+ "epoch": 6.642794759825327,
+ "grad_norm": 0.0194488987326622,
+ "learning_rate": 0.0006,
+ "loss": 5.604279518127441,
+ "step": 478
+ },
+ {
+ "epoch": 6.656768558951965,
+ "grad_norm": 0.02079896256327629,
+ "learning_rate": 0.0006,
+ "loss": 5.586322784423828,
+ "step": 479
+ },
+ {
+ "epoch": 6.670742358078603,
+ "grad_norm": 0.018996229395270348,
+ "learning_rate": 0.0006,
+ "loss": 5.5987772941589355,
+ "step": 480
+ },
+ {
+ "epoch": 6.68471615720524,
+ "grad_norm": 0.015779603272676468,
+ "learning_rate": 0.0006,
+ "loss": 5.588602066040039,
+ "step": 481
+ },
+ {
+ "epoch": 6.698689956331878,
+ "grad_norm": 0.015322973020374775,
+ "learning_rate": 0.0006,
+ "loss": 5.686467170715332,
+ "step": 482
+ },
+ {
+ "epoch": 6.712663755458515,
+ "grad_norm": 0.015282213687896729,
+ "learning_rate": 0.0006,
+ "loss": 5.664676666259766,
+ "step": 483
+ },
+ {
+ "epoch": 6.726637554585153,
+ "grad_norm": 0.016589272767305374,
+ "learning_rate": 0.0006,
+ "loss": 5.666739463806152,
+ "step": 484
+ },
+ {
+ "epoch": 6.74061135371179,
+ "grad_norm": 0.01811421848833561,
+ "learning_rate": 0.0006,
+ "loss": 5.647593021392822,
+ "step": 485
+ },
+ {
+ "epoch": 6.754585152838428,
+ "grad_norm": 0.019652029499411583,
+ "learning_rate": 0.0006,
+ "loss": 5.628726959228516,
+ "step": 486
+ },
+ {
+ "epoch": 6.7685589519650655,
+ "grad_norm": 0.02118665538728237,
+ "learning_rate": 0.0006,
+ "loss": 5.6580705642700195,
+ "step": 487
+ },
+ {
+ "epoch": 6.782532751091703,
+ "grad_norm": 0.02237832546234131,
+ "learning_rate": 0.0006,
+ "loss": 5.635931968688965,
+ "step": 488
+ },
+ {
+ "epoch": 6.796506550218341,
+ "grad_norm": 0.023897631093859673,
+ "learning_rate": 0.0006,
+ "loss": 5.517949104309082,
+ "step": 489
+ },
+ {
+ "epoch": 6.810480349344978,
+ "grad_norm": 0.02442227490246296,
+ "learning_rate": 0.0006,
+ "loss": 5.600021839141846,
+ "step": 490
+ },
+ {
+ "epoch": 6.824454148471616,
+ "grad_norm": 0.024675287306308746,
+ "learning_rate": 0.0006,
+ "loss": 5.6639909744262695,
+ "step": 491
+ },
+ {
+ "epoch": 6.8384279475982535,
+ "grad_norm": 0.030372392386198044,
+ "learning_rate": 0.0006,
+ "loss": 5.582888603210449,
+ "step": 492
+ },
+ {
+ "epoch": 6.85240174672489,
+ "grad_norm": 0.03237690031528473,
+ "learning_rate": 0.0006,
+ "loss": 5.573896884918213,
+ "step": 493
+ },
+ {
+ "epoch": 6.866375545851528,
+ "grad_norm": 0.03511589393019676,
+ "learning_rate": 0.0006,
+ "loss": 5.565878868103027,
+ "step": 494
+ },
+ {
+ "epoch": 6.880349344978166,
+ "grad_norm": 0.03883938118815422,
+ "learning_rate": 0.0006,
+ "loss": 5.58540678024292,
+ "step": 495
+ },
+ {
+ "epoch": 6.894323144104804,
+ "grad_norm": 0.04175502806901932,
+ "learning_rate": 0.0006,
+ "loss": 5.6095194816589355,
+ "step": 496
+ },
+ {
+ "epoch": 6.908296943231441,
+ "grad_norm": 0.04013441503047943,
+ "learning_rate": 0.0006,
+ "loss": 5.654401779174805,
+ "step": 497
+ },
+ {
+ "epoch": 6.922270742358078,
+ "grad_norm": 0.04618770629167557,
+ "learning_rate": 0.0006,
+ "loss": 5.6483473777771,
+ "step": 498
+ },
+ {
+ "epoch": 6.936244541484716,
+ "grad_norm": 0.04507308453321457,
+ "learning_rate": 0.0006,
+ "loss": 5.5856218338012695,
+ "step": 499
+ },
+ {
+ "epoch": 6.950218340611354,
+ "grad_norm": 0.03151383996009827,
+ "learning_rate": 0.0006,
+ "loss": 5.594086647033691,
+ "step": 500
+ },
+ {
+ "epoch": 6.964192139737992,
+ "grad_norm": 0.026883797720074654,
+ "learning_rate": 0.0006,
+ "loss": 5.607676029205322,
+ "step": 501
+ },
+ {
+ "epoch": 6.978165938864628,
+ "grad_norm": 0.02981836162507534,
+ "learning_rate": 0.0006,
+ "loss": 5.563666820526123,
+ "step": 502
+ },
+ {
+ "epoch": 6.992139737991266,
+ "grad_norm": 0.03233477845788002,
+ "learning_rate": 0.0006,
+ "loss": 5.667543411254883,
+ "step": 503
+ },
+ {
+ "epoch": 7.0,
+ "grad_norm": 0.029927456751465797,
+ "learning_rate": 0.0006,
+ "loss": 5.733482837677002,
+ "step": 504
+ },
+ {
+ "epoch": 7.0,
+ "eval_loss": 5.679076671600342,
+ "eval_runtime": 59.5965,
+ "eval_samples_per_second": 40.976,
+ "eval_steps_per_second": 1.292,
+ "step": 504
+ },
+ {
+ "epoch": 7.013973799126638,
+ "grad_norm": 0.027081597596406937,
+ "learning_rate": 0.0006,
+ "loss": 5.634487152099609,
+ "step": 505
+ },
+ {
+ "epoch": 7.0279475982532755,
+ "grad_norm": 0.031959421932697296,
+ "learning_rate": 0.0006,
+ "loss": 5.643294811248779,
+ "step": 506
+ },
+ {
+ "epoch": 7.041921397379912,
+ "grad_norm": 0.02802063524723053,
+ "learning_rate": 0.0006,
+ "loss": 5.573239326477051,
+ "step": 507
+ },
+ {
+ "epoch": 7.05589519650655,
+ "grad_norm": 0.032579705119132996,
+ "learning_rate": 0.0006,
+ "loss": 5.627150058746338,
+ "step": 508
+ },
+ {
+ "epoch": 7.069868995633188,
+ "grad_norm": 0.02503928542137146,
+ "learning_rate": 0.0006,
+ "loss": 5.6045308113098145,
+ "step": 509
+ },
+ {
+ "epoch": 7.083842794759826,
+ "grad_norm": 0.02421317622065544,
+ "learning_rate": 0.0006,
+ "loss": 5.522153377532959,
+ "step": 510
+ },
+ {
+ "epoch": 7.097816593886463,
+ "grad_norm": 0.021983426064252853,
+ "learning_rate": 0.0006,
+ "loss": 5.632939338684082,
+ "step": 511
+ },
+ {
+ "epoch": 7.1117903930131,
+ "grad_norm": 0.021933183073997498,
+ "learning_rate": 0.0006,
+ "loss": 5.592185974121094,
+ "step": 512
+ },
+ {
+ "epoch": 7.125764192139738,
+ "grad_norm": 0.02287031151354313,
+ "learning_rate": 0.0006,
+ "loss": 5.613700866699219,
+ "step": 513
+ },
+ {
+ "epoch": 7.139737991266376,
+ "grad_norm": 0.02281602844595909,
+ "learning_rate": 0.0006,
+ "loss": 5.551383018493652,
+ "step": 514
+ },
+ {
+ "epoch": 7.153711790393013,
+ "grad_norm": 0.021050360053777695,
+ "learning_rate": 0.0006,
+ "loss": 5.55275821685791,
+ "step": 515
+ },
+ {
+ "epoch": 7.1676855895196505,
+ "grad_norm": 0.018299926072359085,
+ "learning_rate": 0.0006,
+ "loss": 5.528225421905518,
+ "step": 516
+ },
+ {
+ "epoch": 7.181659388646288,
+ "grad_norm": 0.02024853602051735,
+ "learning_rate": 0.0006,
+ "loss": 5.522453784942627,
+ "step": 517
+ },
+ {
+ "epoch": 7.195633187772926,
+ "grad_norm": 0.02135239914059639,
+ "learning_rate": 0.0006,
+ "loss": 5.527522563934326,
+ "step": 518
+ },
+ {
+ "epoch": 7.209606986899563,
+ "grad_norm": 0.023394184187054634,
+ "learning_rate": 0.0006,
+ "loss": 5.4535017013549805,
+ "step": 519
+ },
+ {
+ "epoch": 7.223580786026201,
+ "grad_norm": 0.025952080264687538,
+ "learning_rate": 0.0006,
+ "loss": 5.5374674797058105,
+ "step": 520
+ },
+ {
+ "epoch": 7.2375545851528384,
+ "grad_norm": 0.022287718951702118,
+ "learning_rate": 0.0006,
+ "loss": 5.493753433227539,
+ "step": 521
+ },
+ {
+ "epoch": 7.251528384279476,
+ "grad_norm": 0.02018447406589985,
+ "learning_rate": 0.0006,
+ "loss": 5.510575294494629,
+ "step": 522
+ },
+ {
+ "epoch": 7.265502183406113,
+ "grad_norm": 0.02050507813692093,
+ "learning_rate": 0.0006,
+ "loss": 5.466026782989502,
+ "step": 523
+ },
+ {
+ "epoch": 7.279475982532751,
+ "grad_norm": 0.023688802495598793,
+ "learning_rate": 0.0006,
+ "loss": 5.485815525054932,
+ "step": 524
+ },
+ {
+ "epoch": 7.293449781659389,
+ "grad_norm": 0.021078843623399734,
+ "learning_rate": 0.0006,
+ "loss": 5.415581703186035,
+ "step": 525
+ },
+ {
+ "epoch": 7.307423580786026,
+ "grad_norm": 0.018266011029481888,
+ "learning_rate": 0.0006,
+ "loss": 5.402815818786621,
+ "step": 526
+ },
+ {
+ "epoch": 7.321397379912664,
+ "grad_norm": 0.019693493843078613,
+ "learning_rate": 0.0006,
+ "loss": 5.504674911499023,
+ "step": 527
+ },
+ {
+ "epoch": 7.335371179039301,
+ "grad_norm": 0.02031373605132103,
+ "learning_rate": 0.0006,
+ "loss": 5.503837585449219,
+ "step": 528
+ },
+ {
+ "epoch": 7.349344978165939,
+ "grad_norm": 0.017636626958847046,
+ "learning_rate": 0.0006,
+ "loss": 5.530580997467041,
+ "step": 529
+ },
+ {
+ "epoch": 7.3633187772925766,
+ "grad_norm": 0.01787244901061058,
+ "learning_rate": 0.0006,
+ "loss": 5.450218677520752,
+ "step": 530
+ },
+ {
+ "epoch": 7.377292576419214,
+ "grad_norm": 0.0170669574290514,
+ "learning_rate": 0.0006,
+ "loss": 5.481570243835449,
+ "step": 531
+ },
+ {
+ "epoch": 7.391266375545851,
+ "grad_norm": 0.01802165061235428,
+ "learning_rate": 0.0006,
+ "loss": 5.386394500732422,
+ "step": 532
+ },
+ {
+ "epoch": 7.405240174672489,
+ "grad_norm": 0.022949472069740295,
+ "learning_rate": 0.0006,
+ "loss": 5.424929618835449,
+ "step": 533
+ },
+ {
+ "epoch": 7.419213973799127,
+ "grad_norm": 0.034213390201330185,
+ "learning_rate": 0.0006,
+ "loss": 5.346663475036621,
+ "step": 534
+ },
+ {
+ "epoch": 7.4331877729257645,
+ "grad_norm": 0.03847593814134598,
+ "learning_rate": 0.0006,
+ "loss": 5.418482780456543,
+ "step": 535
+ },
+ {
+ "epoch": 7.447161572052401,
+ "grad_norm": 0.029393676668405533,
+ "learning_rate": 0.0006,
+ "loss": 5.456090927124023,
+ "step": 536
+ },
+ {
+ "epoch": 7.461135371179039,
+ "grad_norm": 0.03388667106628418,
+ "learning_rate": 0.0006,
+ "loss": 5.572983741760254,
+ "step": 537
+ },
+ {
+ "epoch": 7.475109170305677,
+ "grad_norm": 0.039690613746643066,
+ "learning_rate": 0.0006,
+ "loss": 5.414067268371582,
+ "step": 538
+ },
+ {
+ "epoch": 7.489082969432315,
+ "grad_norm": 0.03630776330828667,
+ "learning_rate": 0.0006,
+ "loss": 5.523739814758301,
+ "step": 539
+ },
+ {
+ "epoch": 7.503056768558952,
+ "grad_norm": 0.03356870263814926,
+ "learning_rate": 0.0006,
+ "loss": 5.444767951965332,
+ "step": 540
+ },
+ {
+ "epoch": 7.517030567685589,
+ "grad_norm": 0.030812304466962814,
+ "learning_rate": 0.0006,
+ "loss": 5.461244106292725,
+ "step": 541
+ },
+ {
+ "epoch": 7.531004366812227,
+ "grad_norm": 0.03319217637181282,
+ "learning_rate": 0.0006,
+ "loss": 5.502161026000977,
+ "step": 542
+ },
+ {
+ "epoch": 7.544978165938865,
+ "grad_norm": 0.032997481524944305,
+ "learning_rate": 0.0006,
+ "loss": 5.4162278175354,
+ "step": 543
+ },
+ {
+ "epoch": 7.558951965065502,
+ "grad_norm": 0.03364962339401245,
+ "learning_rate": 0.0006,
+ "loss": 5.415736198425293,
+ "step": 544
+ },
+ {
+ "epoch": 7.5729257641921395,
+ "grad_norm": 0.0344221405684948,
+ "learning_rate": 0.0006,
+ "loss": 5.541967868804932,
+ "step": 545
+ },
+ {
+ "epoch": 7.586899563318777,
+ "grad_norm": 0.029609503224492073,
+ "learning_rate": 0.0006,
+ "loss": 5.440614223480225,
+ "step": 546
+ },
+ {
+ "epoch": 7.600873362445415,
+ "grad_norm": 0.029055926948785782,
+ "learning_rate": 0.0006,
+ "loss": 5.462865352630615,
+ "step": 547
+ },
+ {
+ "epoch": 7.614847161572053,
+ "grad_norm": 0.02658848837018013,
+ "learning_rate": 0.0006,
+ "loss": 5.444467544555664,
+ "step": 548
+ },
+ {
+ "epoch": 7.62882096069869,
+ "grad_norm": 0.026276404038071632,
+ "learning_rate": 0.0006,
+ "loss": 5.462278842926025,
+ "step": 549
+ },
+ {
+ "epoch": 7.642794759825327,
+ "grad_norm": 0.0286889486014843,
+ "learning_rate": 0.0006,
+ "loss": 5.4849534034729,
+ "step": 550
+ },
+ {
+ "epoch": 7.656768558951965,
+ "grad_norm": 0.026849817484617233,
+ "learning_rate": 0.0006,
+ "loss": 5.431473731994629,
+ "step": 551
+ },
+ {
+ "epoch": 7.670742358078603,
+ "grad_norm": 0.02312396466732025,
+ "learning_rate": 0.0006,
+ "loss": 5.45462703704834,
+ "step": 552
+ },
+ {
+ "epoch": 7.68471615720524,
+ "grad_norm": 0.026197485625743866,
+ "learning_rate": 0.0006,
+ "loss": 5.452552318572998,
+ "step": 553
+ },
+ {
+ "epoch": 7.698689956331878,
+ "grad_norm": 0.02747255191206932,
+ "learning_rate": 0.0006,
+ "loss": 5.447169780731201,
+ "step": 554
+ },
+ {
+ "epoch": 7.712663755458515,
+ "grad_norm": 0.028123432770371437,
+ "learning_rate": 0.0006,
+ "loss": 5.394400596618652,
+ "step": 555
+ },
+ {
+ "epoch": 7.726637554585153,
+ "grad_norm": 0.02599870041012764,
+ "learning_rate": 0.0006,
+ "loss": 5.398341178894043,
+ "step": 556
+ },
+ {
+ "epoch": 7.74061135371179,
+ "grad_norm": 0.022171657532453537,
+ "learning_rate": 0.0006,
+ "loss": 5.368820667266846,
+ "step": 557
+ },
+ {
+ "epoch": 7.754585152838428,
+ "grad_norm": 0.022309480234980583,
+ "learning_rate": 0.0006,
+ "loss": 5.416772365570068,
+ "step": 558
+ },
+ {
+ "epoch": 7.7685589519650655,
+ "grad_norm": 0.024074165150523186,
+ "learning_rate": 0.0006,
+ "loss": 5.359002113342285,
+ "step": 559
+ },
+ {
+ "epoch": 7.782532751091703,
+ "grad_norm": 0.02636653557419777,
+ "learning_rate": 0.0006,
+ "loss": 5.484368324279785,
+ "step": 560
+ },
+ {
+ "epoch": 7.796506550218341,
+ "grad_norm": 0.02203752100467682,
+ "learning_rate": 0.0006,
+ "loss": 5.375498294830322,
+ "step": 561
+ },
+ {
+ "epoch": 7.810480349344978,
+ "grad_norm": 0.019924819469451904,
+ "learning_rate": 0.0006,
+ "loss": 5.334723472595215,
+ "step": 562
+ },
+ {
+ "epoch": 7.824454148471616,
+ "grad_norm": 0.018755966797471046,
+ "learning_rate": 0.0006,
+ "loss": 5.348781108856201,
+ "step": 563
+ },
+ {
+ "epoch": 7.8384279475982535,
+ "grad_norm": 0.016747845336794853,
+ "learning_rate": 0.0006,
+ "loss": 5.433079719543457,
+ "step": 564
+ },
+ {
+ "epoch": 7.85240174672489,
+ "grad_norm": 0.0170395877212286,
+ "learning_rate": 0.0006,
+ "loss": 5.346794128417969,
+ "step": 565
+ },
+ {
+ "epoch": 7.866375545851528,
+ "grad_norm": 0.016074176877737045,
+ "learning_rate": 0.0006,
+ "loss": 5.3277740478515625,
+ "step": 566
+ },
+ {
+ "epoch": 7.880349344978166,
+ "grad_norm": 0.014119806699454784,
+ "learning_rate": 0.0006,
+ "loss": 5.483551979064941,
+ "step": 567
+ },
+ {
+ "epoch": 7.894323144104804,
+ "grad_norm": 0.015270394273102283,
+ "learning_rate": 0.0006,
+ "loss": 5.378519535064697,
+ "step": 568
+ },
+ {
+ "epoch": 7.908296943231441,
+ "grad_norm": 0.015677539631724358,
+ "learning_rate": 0.0006,
+ "loss": 5.290210247039795,
+ "step": 569
+ },
+ {
+ "epoch": 7.922270742358078,
+ "grad_norm": 0.015930423513054848,
+ "learning_rate": 0.0006,
+ "loss": 5.427360534667969,
+ "step": 570
+ },
+ {
+ "epoch": 7.936244541484716,
+ "grad_norm": 0.016641564667224884,
+ "learning_rate": 0.0006,
+ "loss": 5.301599979400635,
+ "step": 571
+ },
+ {
+ "epoch": 7.950218340611354,
+ "grad_norm": 0.018293552100658417,
+ "learning_rate": 0.0006,
+ "loss": 5.277889251708984,
+ "step": 572
+ },
+ {
+ "epoch": 7.964192139737992,
+ "grad_norm": 0.017618799582123756,
+ "learning_rate": 0.0006,
+ "loss": 5.297194480895996,
+ "step": 573
+ },
+ {
+ "epoch": 7.978165938864628,
+ "grad_norm": 0.015034242533147335,
+ "learning_rate": 0.0006,
+ "loss": 5.42428731918335,
+ "step": 574
+ },
+ {
+ "epoch": 7.992139737991266,
+ "grad_norm": 0.01617511734366417,
+ "learning_rate": 0.0006,
+ "loss": 5.420044898986816,
+ "step": 575
+ },
+ {
+ "epoch": 8.0,
+ "grad_norm": 0.01837257295846939,
+ "learning_rate": 0.0006,
+ "loss": 5.240387916564941,
+ "step": 576
+ },
+ {
+ "epoch": 8.0,
+ "eval_loss": 5.423932075500488,
+ "eval_runtime": 58.0841,
+ "eval_samples_per_second": 42.042,
+ "eval_steps_per_second": 1.326,
+ "step": 576
+ },
+ {
+ "epoch": 8.013973799126637,
+ "grad_norm": 0.020953809842467308,
+ "learning_rate": 0.0006,
+ "loss": 5.4320783615112305,
+ "step": 577
+ },
+ {
+ "epoch": 8.027947598253276,
+ "grad_norm": 0.0227745920419693,
+ "learning_rate": 0.0006,
+ "loss": 5.310197830200195,
+ "step": 578
+ },
+ {
+ "epoch": 8.041921397379912,
+ "grad_norm": 0.021372603252530098,
+ "learning_rate": 0.0006,
+ "loss": 5.339468002319336,
+ "step": 579
+ },
+ {
+ "epoch": 8.055895196506551,
+ "grad_norm": 0.022401731461286545,
+ "learning_rate": 0.0006,
+ "loss": 5.279829978942871,
+ "step": 580
+ },
+ {
+ "epoch": 8.069868995633188,
+ "grad_norm": 0.023942379280924797,
+ "learning_rate": 0.0006,
+ "loss": 5.372323989868164,
+ "step": 581
+ },
+ {
+ "epoch": 8.083842794759825,
+ "grad_norm": 0.024380534887313843,
+ "learning_rate": 0.0006,
+ "loss": 5.253486633300781,
+ "step": 582
+ },
+ {
+ "epoch": 8.097816593886463,
+ "grad_norm": 0.03246625140309334,
+ "learning_rate": 0.0006,
+ "loss": 5.305774211883545,
+ "step": 583
+ },
+ {
+ "epoch": 8.1117903930131,
+ "grad_norm": 0.030399736016988754,
+ "learning_rate": 0.0006,
+ "loss": 5.370976448059082,
+ "step": 584
+ },
+ {
+ "epoch": 8.125764192139737,
+ "grad_norm": 0.027148913592100143,
+ "learning_rate": 0.0006,
+ "loss": 5.368010520935059,
+ "step": 585
+ },
+ {
+ "epoch": 8.139737991266376,
+ "grad_norm": 0.02925540879368782,
+ "learning_rate": 0.0006,
+ "loss": 5.303013801574707,
+ "step": 586
+ },
+ {
+ "epoch": 8.153711790393013,
+ "grad_norm": 0.02915453538298607,
+ "learning_rate": 0.0006,
+ "loss": 5.329667091369629,
+ "step": 587
+ },
+ {
+ "epoch": 8.167685589519651,
+ "grad_norm": 0.031732626259326935,
+ "learning_rate": 0.0006,
+ "loss": 5.3634934425354,
+ "step": 588
+ },
+ {
+ "epoch": 8.181659388646288,
+ "grad_norm": 0.036166731268167496,
+ "learning_rate": 0.0006,
+ "loss": 5.223980903625488,
+ "step": 589
+ },
+ {
+ "epoch": 8.195633187772925,
+ "grad_norm": 0.035654108971357346,
+ "learning_rate": 0.0006,
+ "loss": 5.361059188842773,
+ "step": 590
+ },
+ {
+ "epoch": 8.209606986899564,
+ "grad_norm": 0.0365324430167675,
+ "learning_rate": 0.0006,
+ "loss": 5.336151599884033,
+ "step": 591
+ },
+ {
+ "epoch": 8.2235807860262,
+ "grad_norm": 0.03578880429267883,
+ "learning_rate": 0.0006,
+ "loss": 5.434076309204102,
+ "step": 592
+ },
+ {
+ "epoch": 8.237554585152838,
+ "grad_norm": 0.035734012722969055,
+ "learning_rate": 0.0006,
+ "loss": 5.330893516540527,
+ "step": 593
+ },
+ {
+ "epoch": 8.251528384279476,
+ "grad_norm": 0.03196857124567032,
+ "learning_rate": 0.0006,
+ "loss": 5.306009769439697,
+ "step": 594
+ },
+ {
+ "epoch": 8.265502183406113,
+ "grad_norm": 0.032042086124420166,
+ "learning_rate": 0.0006,
+ "loss": 5.343267917633057,
+ "step": 595
+ },
+ {
+ "epoch": 8.279475982532752,
+ "grad_norm": 0.03160746395587921,
+ "learning_rate": 0.0006,
+ "loss": 5.2353715896606445,
+ "step": 596
+ },
+ {
+ "epoch": 8.293449781659389,
+ "grad_norm": 0.026689818128943443,
+ "learning_rate": 0.0006,
+ "loss": 5.254042625427246,
+ "step": 597
+ },
+ {
+ "epoch": 8.307423580786025,
+ "grad_norm": 0.02880188450217247,
+ "learning_rate": 0.0006,
+ "loss": 5.326833248138428,
+ "step": 598
+ },
+ {
+ "epoch": 8.321397379912664,
+ "grad_norm": 0.027516381815075874,
+ "learning_rate": 0.0006,
+ "loss": 5.342336177825928,
+ "step": 599
+ },
+ {
+ "epoch": 8.335371179039301,
+ "grad_norm": 0.027875030413269997,
+ "learning_rate": 0.0006,
+ "loss": 5.200719833374023,
+ "step": 600
+ },
+ {
+ "epoch": 8.34934497816594,
+ "grad_norm": 0.0268265288323164,
+ "learning_rate": 0.0006,
+ "loss": 5.249449729919434,
+ "step": 601
+ },
+ {
+ "epoch": 8.363318777292577,
+ "grad_norm": 0.024824608117341995,
+ "learning_rate": 0.0006,
+ "loss": 5.318952560424805,
+ "step": 602
+ },
+ {
+ "epoch": 8.377292576419213,
+ "grad_norm": 0.022990627214312553,
+ "learning_rate": 0.0006,
+ "loss": 5.244993209838867,
+ "step": 603
+ },
+ {
+ "epoch": 8.391266375545852,
+ "grad_norm": 0.022804604843258858,
+ "learning_rate": 0.0006,
+ "loss": 5.3236846923828125,
+ "step": 604
+ },
+ {
+ "epoch": 8.405240174672489,
+ "grad_norm": 0.02419872209429741,
+ "learning_rate": 0.0006,
+ "loss": 5.255486488342285,
+ "step": 605
+ },
+ {
+ "epoch": 8.419213973799126,
+ "grad_norm": 0.021952059119939804,
+ "learning_rate": 0.0006,
+ "loss": 5.3511552810668945,
+ "step": 606
+ },
+ {
+ "epoch": 8.433187772925764,
+ "grad_norm": 0.022375497967004776,
+ "learning_rate": 0.0006,
+ "loss": 5.294790267944336,
+ "step": 607
+ },
+ {
+ "epoch": 8.447161572052401,
+ "grad_norm": 0.019242815673351288,
+ "learning_rate": 0.0006,
+ "loss": 5.252618789672852,
+ "step": 608
+ },
+ {
+ "epoch": 8.46113537117904,
+ "grad_norm": 0.0173486340790987,
+ "learning_rate": 0.0006,
+ "loss": 5.30147647857666,
+ "step": 609
+ },
+ {
+ "epoch": 8.475109170305677,
+ "grad_norm": 0.01612604409456253,
+ "learning_rate": 0.0006,
+ "loss": 5.23960542678833,
+ "step": 610
+ },
+ {
+ "epoch": 8.489082969432314,
+ "grad_norm": 0.014293976128101349,
+ "learning_rate": 0.0006,
+ "loss": 5.119932651519775,
+ "step": 611
+ },
+ {
+ "epoch": 8.503056768558952,
+ "grad_norm": 0.014453536830842495,
+ "learning_rate": 0.0006,
+ "loss": 5.166794776916504,
+ "step": 612
+ },
+ {
+ "epoch": 8.51703056768559,
+ "grad_norm": 0.014685769565403461,
+ "learning_rate": 0.0006,
+ "loss": 5.264822959899902,
+ "step": 613
+ },
+ {
+ "epoch": 8.531004366812226,
+ "grad_norm": 0.01427740603685379,
+ "learning_rate": 0.0006,
+ "loss": 5.278433799743652,
+ "step": 614
+ },
+ {
+ "epoch": 8.544978165938865,
+ "grad_norm": 0.014864951372146606,
+ "learning_rate": 0.0006,
+ "loss": 5.214512348175049,
+ "step": 615
+ },
+ {
+ "epoch": 8.558951965065502,
+ "grad_norm": 0.01565164513885975,
+ "learning_rate": 0.0006,
+ "loss": 5.183066368103027,
+ "step": 616
+ },
+ {
+ "epoch": 8.57292576419214,
+ "grad_norm": 0.01779816672205925,
+ "learning_rate": 0.0006,
+ "loss": 5.163185119628906,
+ "step": 617
+ },
+ {
+ "epoch": 8.586899563318777,
+ "grad_norm": 0.017254827544093132,
+ "learning_rate": 0.0006,
+ "loss": 5.243555068969727,
+ "step": 618
+ },
+ {
+ "epoch": 8.600873362445414,
+ "grad_norm": 0.01650584116578102,
+ "learning_rate": 0.0006,
+ "loss": 5.234884738922119,
+ "step": 619
+ },
+ {
+ "epoch": 8.614847161572053,
+ "grad_norm": 0.017021115869283676,
+ "learning_rate": 0.0006,
+ "loss": 5.12271785736084,
+ "step": 620
+ },
+ {
+ "epoch": 8.62882096069869,
+ "grad_norm": 0.01773759163916111,
+ "learning_rate": 0.0006,
+ "loss": 5.242153167724609,
+ "step": 621
+ },
+ {
+ "epoch": 8.642794759825328,
+ "grad_norm": 0.015679042786359787,
+ "learning_rate": 0.0006,
+ "loss": 5.170779228210449,
+ "step": 622
+ },
+ {
+ "epoch": 8.656768558951965,
+ "grad_norm": 0.013760549947619438,
+ "learning_rate": 0.0006,
+ "loss": 5.238644599914551,
+ "step": 623
+ },
+ {
+ "epoch": 8.670742358078602,
+ "grad_norm": 0.014571522362530231,
+ "learning_rate": 0.0006,
+ "loss": 5.177056312561035,
+ "step": 624
+ },
+ {
+ "epoch": 8.68471615720524,
+ "grad_norm": 0.016209015622735023,
+ "learning_rate": 0.0006,
+ "loss": 5.163750648498535,
+ "step": 625
+ },
+ {
+ "epoch": 8.698689956331878,
+ "grad_norm": 0.016813941299915314,
+ "learning_rate": 0.0006,
+ "loss": 5.183428764343262,
+ "step": 626
+ },
+ {
+ "epoch": 8.712663755458514,
+ "grad_norm": 0.020985357463359833,
+ "learning_rate": 0.0006,
+ "loss": 5.229465007781982,
+ "step": 627
+ },
+ {
+ "epoch": 8.726637554585153,
+ "grad_norm": 0.02679632417857647,
+ "learning_rate": 0.0006,
+ "loss": 5.257368087768555,
+ "step": 628
+ },
+ {
+ "epoch": 8.74061135371179,
+ "grad_norm": 0.025756070390343666,
+ "learning_rate": 0.0006,
+ "loss": 5.253736972808838,
+ "step": 629
+ },
+ {
+ "epoch": 8.754585152838429,
+ "grad_norm": 0.02643490768969059,
+ "learning_rate": 0.0006,
+ "loss": 5.228633403778076,
+ "step": 630
+ },
+ {
+ "epoch": 8.768558951965066,
+ "grad_norm": 0.03303210437297821,
+ "learning_rate": 0.0006,
+ "loss": 5.163540363311768,
+ "step": 631
+ },
+ {
+ "epoch": 8.782532751091702,
+ "grad_norm": 0.030432431027293205,
+ "learning_rate": 0.0006,
+ "loss": 5.243169784545898,
+ "step": 632
+ },
+ {
+ "epoch": 8.796506550218341,
+ "grad_norm": 0.028631288558244705,
+ "learning_rate": 0.0006,
+ "loss": 5.166398048400879,
+ "step": 633
+ },
+ {
+ "epoch": 8.810480349344978,
+ "grad_norm": 0.026188310235738754,
+ "learning_rate": 0.0006,
+ "loss": 5.128122329711914,
+ "step": 634
+ },
+ {
+ "epoch": 8.824454148471617,
+ "grad_norm": 0.025308528915047646,
+ "learning_rate": 0.0006,
+ "loss": 5.211284637451172,
+ "step": 635
+ },
+ {
+ "epoch": 8.838427947598253,
+ "grad_norm": 0.020657729357481003,
+ "learning_rate": 0.0006,
+ "loss": 5.161575794219971,
+ "step": 636
+ },
+ {
+ "epoch": 8.85240174672489,
+ "grad_norm": 0.021901234984397888,
+ "learning_rate": 0.0006,
+ "loss": 5.20050048828125,
+ "step": 637
+ },
+ {
+ "epoch": 8.866375545851529,
+ "grad_norm": 0.021219318732619286,
+ "learning_rate": 0.0006,
+ "loss": 5.211699485778809,
+ "step": 638
+ },
+ {
+ "epoch": 8.880349344978166,
+ "grad_norm": 0.0200974028557539,
+ "learning_rate": 0.0006,
+ "loss": 5.092503547668457,
+ "step": 639
+ },
+ {
+ "epoch": 8.894323144104803,
+ "grad_norm": 0.023804882541298866,
+ "learning_rate": 0.0006,
+ "loss": 5.216068267822266,
+ "step": 640
+ },
+ {
+ "epoch": 8.908296943231441,
+ "grad_norm": 0.026088010519742966,
+ "learning_rate": 0.0006,
+ "loss": 5.155592918395996,
+ "step": 641
+ },
+ {
+ "epoch": 8.922270742358078,
+ "grad_norm": 0.02601276896893978,
+ "learning_rate": 0.0006,
+ "loss": 5.197238922119141,
+ "step": 642
+ },
+ {
+ "epoch": 8.936244541484717,
+ "grad_norm": 0.020387211814522743,
+ "learning_rate": 0.0006,
+ "loss": 5.1955437660217285,
+ "step": 643
+ },
+ {
+ "epoch": 8.950218340611354,
+ "grad_norm": 0.019214622676372528,
+ "learning_rate": 0.0006,
+ "loss": 5.206346035003662,
+ "step": 644
+ },
+ {
+ "epoch": 8.96419213973799,
+ "grad_norm": 0.019674314185976982,
+ "learning_rate": 0.0006,
+ "loss": 5.219857215881348,
+ "step": 645
+ },
+ {
+ "epoch": 8.97816593886463,
+ "grad_norm": 0.020318234339356422,
+ "learning_rate": 0.0006,
+ "loss": 5.191132545471191,
+ "step": 646
+ },
+ {
+ "epoch": 8.992139737991266,
+ "grad_norm": 0.021428676322102547,
+ "learning_rate": 0.0006,
+ "loss": 5.214695930480957,
+ "step": 647
+ },
+ {
+ "epoch": 9.0,
+ "grad_norm": 0.023235054686665535,
+ "learning_rate": 0.0006,
+ "loss": 5.229091644287109,
+ "step": 648
+ },
+ {
+ "epoch": 9.0,
+ "eval_loss": 5.289593696594238,
+ "eval_runtime": 56.9256,
+ "eval_samples_per_second": 42.898,
+ "eval_steps_per_second": 1.353,
+ "step": 648
+ },
+ {
+ "epoch": 9.013973799126637,
+ "grad_norm": 0.02320289984345436,
+ "learning_rate": 0.0006,
+ "loss": 4.983669281005859,
+ "step": 649
+ },
+ {
+ "epoch": 9.027947598253276,
+ "grad_norm": 0.018521282821893692,
+ "learning_rate": 0.0006,
+ "loss": 5.200575828552246,
+ "step": 650
+ },
+ {
+ "epoch": 9.041921397379912,
+ "grad_norm": 0.0180169939994812,
+ "learning_rate": 0.0006,
+ "loss": 5.148033618927002,
+ "step": 651
+ },
+ {
+ "epoch": 9.055895196506551,
+ "grad_norm": 0.019969860091805458,
+ "learning_rate": 0.0006,
+ "loss": 5.099125385284424,
+ "step": 652
+ },
+ {
+ "epoch": 9.069868995633188,
+ "grad_norm": 0.01643305830657482,
+ "learning_rate": 0.0006,
+ "loss": 5.092347145080566,
+ "step": 653
+ },
+ {
+ "epoch": 9.083842794759825,
+ "grad_norm": 0.016022363677620888,
+ "learning_rate": 0.0006,
+ "loss": 5.129279613494873,
+ "step": 654
+ },
+ {
+ "epoch": 9.097816593886463,
+ "grad_norm": 0.01657041162252426,
+ "learning_rate": 0.0006,
+ "loss": 5.159562587738037,
+ "step": 655
+ },
+ {
+ "epoch": 9.1117903930131,
+ "grad_norm": 0.017924228683114052,
+ "learning_rate": 0.0006,
+ "loss": 5.165497779846191,
+ "step": 656
+ },
+ {
+ "epoch": 9.125764192139737,
+ "grad_norm": 0.021592361852526665,
+ "learning_rate": 0.0006,
+ "loss": 5.204775810241699,
+ "step": 657
+ },
+ {
+ "epoch": 9.139737991266376,
+ "grad_norm": 0.026924120262265205,
+ "learning_rate": 0.0006,
+ "loss": 5.128296852111816,
+ "step": 658
+ },
+ {
+ "epoch": 9.153711790393013,
+ "grad_norm": 0.023156899958848953,
+ "learning_rate": 0.0006,
+ "loss": 5.1255340576171875,
+ "step": 659
+ },
+ {
+ "epoch": 9.167685589519651,
+ "grad_norm": 0.020670367404818535,
+ "learning_rate": 0.0006,
+ "loss": 5.128479957580566,
+ "step": 660
+ },
+ {
+ "epoch": 9.181659388646288,
+ "grad_norm": 0.025447173044085503,
+ "learning_rate": 0.0006,
+ "loss": 5.154821872711182,
+ "step": 661
+ },
+ {
+ "epoch": 9.195633187772925,
+ "grad_norm": 0.027940068393945694,
+ "learning_rate": 0.0006,
+ "loss": 5.143466472625732,
+ "step": 662
+ },
+ {
+ "epoch": 9.209606986899564,
+ "grad_norm": 0.026847844943404198,
+ "learning_rate": 0.0006,
+ "loss": 5.045665740966797,
+ "step": 663
+ },
+ {
+ "epoch": 9.2235807860262,
+ "grad_norm": 0.02340601570904255,
+ "learning_rate": 0.0006,
+ "loss": 5.21933126449585,
+ "step": 664
+ },
+ {
+ "epoch": 9.237554585152838,
+ "grad_norm": 0.02340371161699295,
+ "learning_rate": 0.0006,
+ "loss": 4.942949295043945,
+ "step": 665
+ },
+ {
+ "epoch": 9.251528384279476,
+ "grad_norm": 0.02221992425620556,
+ "learning_rate": 0.0006,
+ "loss": 5.197099208831787,
+ "step": 666
+ },
+ {
+ "epoch": 9.265502183406113,
+ "grad_norm": 0.023508677259087563,
+ "learning_rate": 0.0006,
+ "loss": 5.022897720336914,
+ "step": 667
+ },
+ {
+ "epoch": 9.279475982532752,
+ "grad_norm": 0.026120394468307495,
+ "learning_rate": 0.0006,
+ "loss": 5.087725639343262,
+ "step": 668
+ },
+ {
+ "epoch": 9.293449781659389,
+ "grad_norm": 0.027273228392004967,
+ "learning_rate": 0.0006,
+ "loss": 5.090963840484619,
+ "step": 669
+ },
+ {
+ "epoch": 9.307423580786025,
+ "grad_norm": 0.03241586685180664,
+ "learning_rate": 0.0006,
+ "loss": 5.047842979431152,
+ "step": 670
+ },
+ {
+ "epoch": 9.321397379912664,
+ "grad_norm": 0.030245667323470116,
+ "learning_rate": 0.0006,
+ "loss": 5.104803085327148,
+ "step": 671
+ },
+ {
+ "epoch": 9.335371179039301,
+ "grad_norm": 0.027698364108800888,
+ "learning_rate": 0.0006,
+ "loss": 5.0644097328186035,
+ "step": 672
+ },
+ {
+ "epoch": 9.34934497816594,
+ "grad_norm": 0.029692554846405983,
+ "learning_rate": 0.0006,
+ "loss": 5.111942291259766,
+ "step": 673
+ },
+ {
+ "epoch": 9.363318777292577,
+ "grad_norm": 0.03328656405210495,
+ "learning_rate": 0.0006,
+ "loss": 5.1574554443359375,
+ "step": 674
+ },
+ {
+ "epoch": 9.377292576419213,
+ "grad_norm": 0.031938180327415466,
+ "learning_rate": 0.0006,
+ "loss": 5.190371990203857,
+ "step": 675
+ },
+ {
+ "epoch": 9.391266375545852,
+ "grad_norm": 0.024858945980668068,
+ "learning_rate": 0.0006,
+ "loss": 5.121420860290527,
+ "step": 676
+ },
+ {
+ "epoch": 9.405240174672489,
+ "grad_norm": 0.023033203557133675,
+ "learning_rate": 0.0006,
+ "loss": 5.0974345207214355,
+ "step": 677
+ },
+ {
+ "epoch": 9.419213973799126,
+ "grad_norm": 0.022083545103669167,
+ "learning_rate": 0.0006,
+ "loss": 5.046319961547852,
+ "step": 678
+ },
+ {
+ "epoch": 9.433187772925764,
+ "grad_norm": 0.0228437427431345,
+ "learning_rate": 0.0006,
+ "loss": 5.119963645935059,
+ "step": 679
+ },
+ {
+ "epoch": 9.447161572052401,
+ "grad_norm": 0.0242351982742548,
+ "learning_rate": 0.0006,
+ "loss": 5.0444560050964355,
+ "step": 680
+ },
+ {
+ "epoch": 9.46113537117904,
+ "grad_norm": 0.02401185780763626,
+ "learning_rate": 0.0006,
+ "loss": 5.063238620758057,
+ "step": 681
+ },
+ {
+ "epoch": 9.475109170305677,
+ "grad_norm": 0.02560959942638874,
+ "learning_rate": 0.0006,
+ "loss": 5.035224914550781,
+ "step": 682
+ },
+ {
+ "epoch": 9.489082969432314,
+ "grad_norm": 0.024985626339912415,
+ "learning_rate": 0.0006,
+ "loss": 5.049284934997559,
+ "step": 683
+ },
+ {
+ "epoch": 9.503056768558952,
+ "grad_norm": 0.023178115487098694,
+ "learning_rate": 0.0006,
+ "loss": 5.2266621589660645,
+ "step": 684
+ },
+ {
+ "epoch": 9.51703056768559,
+ "grad_norm": 0.0215512253344059,
+ "learning_rate": 0.0006,
+ "loss": 5.1875715255737305,
+ "step": 685
+ },
+ {
+ "epoch": 9.531004366812226,
+ "grad_norm": 0.01662837713956833,
+ "learning_rate": 0.0006,
+ "loss": 5.137012481689453,
+ "step": 686
+ },
+ {
+ "epoch": 9.544978165938865,
+ "grad_norm": 0.01952764391899109,
+ "learning_rate": 0.0006,
+ "loss": 5.072500228881836,
+ "step": 687
+ },
+ {
+ "epoch": 9.558951965065502,
+ "grad_norm": 0.017521077767014503,
+ "learning_rate": 0.0006,
+ "loss": 5.045711994171143,
+ "step": 688
+ },
+ {
+ "epoch": 9.57292576419214,
+ "grad_norm": 0.017355265095829964,
+ "learning_rate": 0.0006,
+ "loss": 5.02881383895874,
+ "step": 689
+ },
+ {
+ "epoch": 9.586899563318777,
+ "grad_norm": 0.015587719157338142,
+ "learning_rate": 0.0006,
+ "loss": 5.132787704467773,
+ "step": 690
+ },
+ {
+ "epoch": 9.600873362445414,
+ "grad_norm": 0.017469845712184906,
+ "learning_rate": 0.0006,
+ "loss": 5.105447292327881,
+ "step": 691
+ },
+ {
+ "epoch": 9.614847161572053,
+ "grad_norm": 0.019630100578069687,
+ "learning_rate": 0.0006,
+ "loss": 5.032464981079102,
+ "step": 692
+ },
+ {
+ "epoch": 9.62882096069869,
+ "grad_norm": 0.01784994639456272,
+ "learning_rate": 0.0006,
+ "loss": 5.081345558166504,
+ "step": 693
+ },
+ {
+ "epoch": 9.642794759825328,
+ "grad_norm": 0.01872754842042923,
+ "learning_rate": 0.0006,
+ "loss": 4.9772114753723145,
+ "step": 694
+ },
+ {
+ "epoch": 9.656768558951965,
+ "grad_norm": 0.01999032311141491,
+ "learning_rate": 0.0006,
+ "loss": 5.089740753173828,
+ "step": 695
+ },
+ {
+ "epoch": 9.670742358078602,
+ "grad_norm": 0.01710767112672329,
+ "learning_rate": 0.0006,
+ "loss": 4.9869890213012695,
+ "step": 696
+ },
+ {
+ "epoch": 9.68471615720524,
+ "grad_norm": 0.015144161880016327,
+ "learning_rate": 0.0006,
+ "loss": 5.022564888000488,
+ "step": 697
+ },
+ {
+ "epoch": 9.698689956331878,
+ "grad_norm": 0.01618543453514576,
+ "learning_rate": 0.0006,
+ "loss": 5.035921573638916,
+ "step": 698
+ },
+ {
+ "epoch": 9.712663755458514,
+ "grad_norm": 0.014653387479484081,
+ "learning_rate": 0.0006,
+ "loss": 4.914456367492676,
+ "step": 699
+ },
+ {
+ "epoch": 9.726637554585153,
+ "grad_norm": 0.013434977270662785,
+ "learning_rate": 0.0006,
+ "loss": 5.063764572143555,
+ "step": 700
+ },
+ {
+ "epoch": 9.74061135371179,
+ "grad_norm": 0.015128864906728268,
+ "learning_rate": 0.0006,
+ "loss": 5.052367687225342,
+ "step": 701
+ },
+ {
+ "epoch": 9.754585152838429,
+ "grad_norm": 0.014210768043994904,
+ "learning_rate": 0.0006,
+ "loss": 5.016489028930664,
+ "step": 702
+ },
+ {
+ "epoch": 9.768558951965066,
+ "grad_norm": 0.015191680751740932,
+ "learning_rate": 0.0006,
+ "loss": 4.990176200866699,
+ "step": 703
+ },
+ {
+ "epoch": 9.782532751091702,
+ "grad_norm": 0.018285127356648445,
+ "learning_rate": 0.0006,
+ "loss": 5.015164375305176,
+ "step": 704
+ },
+ {
+ "epoch": 9.796506550218341,
+ "grad_norm": 0.0204079058021307,
+ "learning_rate": 0.0006,
+ "loss": 5.091142654418945,
+ "step": 705
+ },
+ {
+ "epoch": 9.810480349344978,
+ "grad_norm": 0.01980586163699627,
+ "learning_rate": 0.0006,
+ "loss": 5.03350830078125,
+ "step": 706
+ },
+ {
+ "epoch": 9.824454148471617,
+ "grad_norm": 0.02048729732632637,
+ "learning_rate": 0.0006,
+ "loss": 5.03611421585083,
+ "step": 707
+ },
+ {
+ "epoch": 9.838427947598253,
+ "grad_norm": 0.02058163844048977,
+ "learning_rate": 0.0006,
+ "loss": 4.9613447189331055,
+ "step": 708
+ },
+ {
+ "epoch": 9.85240174672489,
+ "grad_norm": 0.018004924058914185,
+ "learning_rate": 0.0006,
+ "loss": 4.966541767120361,
+ "step": 709
+ },
+ {
+ "epoch": 9.866375545851529,
+ "grad_norm": 0.0198152307420969,
+ "learning_rate": 0.0006,
+ "loss": 5.01750373840332,
+ "step": 710
+ },
+ {
+ "epoch": 9.880349344978166,
+ "grad_norm": 0.018532969057559967,
+ "learning_rate": 0.0006,
+ "loss": 5.097424507141113,
+ "step": 711
+ },
+ {
+ "epoch": 9.894323144104803,
+ "grad_norm": 0.018939971923828125,
+ "learning_rate": 0.0006,
+ "loss": 5.177223205566406,
+ "step": 712
+ },
+ {
+ "epoch": 9.908296943231441,
+ "grad_norm": 0.017969170585274696,
+ "learning_rate": 0.0006,
+ "loss": 5.118013858795166,
+ "step": 713
+ },
+ {
+ "epoch": 9.922270742358078,
+ "grad_norm": 0.018021270632743835,
+ "learning_rate": 0.0006,
+ "loss": 5.034787178039551,
+ "step": 714
+ },
+ {
+ "epoch": 9.936244541484717,
+ "grad_norm": 0.017009710893034935,
+ "learning_rate": 0.0006,
+ "loss": 5.003920555114746,
+ "step": 715
+ },
+ {
+ "epoch": 9.950218340611354,
+ "grad_norm": 0.017722049728035927,
+ "learning_rate": 0.0006,
+ "loss": 5.031939506530762,
+ "step": 716
+ },
+ {
+ "epoch": 9.96419213973799,
+ "grad_norm": 0.017134087160229683,
+ "learning_rate": 0.0006,
+ "loss": 5.093474388122559,
+ "step": 717
+ },
+ {
+ "epoch": 9.97816593886463,
+ "grad_norm": 0.018511613830924034,
+ "learning_rate": 0.0006,
+ "loss": 5.078604698181152,
+ "step": 718
+ },
+ {
+ "epoch": 9.992139737991266,
+ "grad_norm": 0.021206334233283997,
+ "learning_rate": 0.0006,
+ "loss": 5.052572727203369,
+ "step": 719
+ },
+ {
+ "epoch": 10.0,
+ "grad_norm": 0.02239062264561653,
+ "learning_rate": 0.0006,
+ "loss": 5.063547134399414,
+ "step": 720
+ },
+ {
+ "epoch": 10.0,
+ "eval_loss": 5.173044681549072,
+ "eval_runtime": 56.7287,
+ "eval_samples_per_second": 43.047,
+ "eval_steps_per_second": 1.357,
+ "step": 720
+ },
+ {
+ "epoch": 10.013973799126637,
+ "grad_norm": 0.020757969468832016,
+ "learning_rate": 0.0006,
+ "loss": 5.045648574829102,
+ "step": 721
+ },
+ {
+ "epoch": 10.027947598253276,
+ "grad_norm": 0.022697702050209045,
+ "learning_rate": 0.0006,
+ "loss": 4.993833065032959,
+ "step": 722
+ },
+ {
+ "epoch": 10.041921397379912,
+ "grad_norm": 0.02434331737458706,
+ "learning_rate": 0.0006,
+ "loss": 5.073219299316406,
+ "step": 723
+ },
+ {
+ "epoch": 10.055895196506551,
+ "grad_norm": 0.025499660521745682,
+ "learning_rate": 0.0006,
+ "loss": 5.019072532653809,
+ "step": 724
+ },
+ {
+ "epoch": 10.069868995633188,
+ "grad_norm": 0.029281053692102432,
+ "learning_rate": 0.0006,
+ "loss": 4.959851264953613,
+ "step": 725
+ },
+ {
+ "epoch": 10.083842794759825,
+ "grad_norm": 0.033878110349178314,
+ "learning_rate": 0.0006,
+ "loss": 5.132717609405518,
+ "step": 726
+ },
+ {
+ "epoch": 10.097816593886463,
+ "grad_norm": 0.03385764732956886,
+ "learning_rate": 0.0006,
+ "loss": 5.0482892990112305,
+ "step": 727
+ },
+ {
+ "epoch": 10.1117903930131,
+ "grad_norm": 0.03466648980975151,
+ "learning_rate": 0.0006,
+ "loss": 4.946855545043945,
+ "step": 728
+ },
+ {
+ "epoch": 10.125764192139737,
+ "grad_norm": 0.028546305373311043,
+ "learning_rate": 0.0006,
+ "loss": 5.027774333953857,
+ "step": 729
+ },
+ {
+ "epoch": 10.139737991266376,
+ "grad_norm": 0.030575869604945183,
+ "learning_rate": 0.0006,
+ "loss": 4.9768476486206055,
+ "step": 730
+ },
+ {
+ "epoch": 10.153711790393013,
+ "grad_norm": 0.03614303469657898,
+ "learning_rate": 0.0006,
+ "loss": 5.078113555908203,
+ "step": 731
+ },
+ {
+ "epoch": 10.167685589519651,
+ "grad_norm": 0.040991660207509995,
+ "learning_rate": 0.0006,
+ "loss": 5.033185958862305,
+ "step": 732
+ },
+ {
+ "epoch": 10.181659388646288,
+ "grad_norm": 0.04551699757575989,
+ "learning_rate": 0.0006,
+ "loss": 5.0820746421813965,
+ "step": 733
+ },
+ {
+ "epoch": 10.195633187772925,
+ "grad_norm": 0.040193989872932434,
+ "learning_rate": 0.0006,
+ "loss": 5.059700965881348,
+ "step": 734
+ },
+ {
+ "epoch": 10.209606986899564,
+ "grad_norm": 0.035851605236530304,
+ "learning_rate": 0.0006,
+ "loss": 5.142467498779297,
+ "step": 735
+ },
+ {
+ "epoch": 10.2235807860262,
+ "grad_norm": 0.034200169146060944,
+ "learning_rate": 0.0006,
+ "loss": 5.072312355041504,
+ "step": 736
+ },
+ {
+ "epoch": 10.237554585152838,
+ "grad_norm": 0.033621110022068024,
+ "learning_rate": 0.0006,
+ "loss": 5.08037805557251,
+ "step": 737
+ },
+ {
+ "epoch": 10.251528384279476,
+ "grad_norm": 0.037372369319200516,
+ "learning_rate": 0.0006,
+ "loss": 4.998425483703613,
+ "step": 738
+ },
+ {
+ "epoch": 10.265502183406113,
+ "grad_norm": 0.037044707685709,
+ "learning_rate": 0.0006,
+ "loss": 4.923993110656738,
+ "step": 739
+ },
+ {
+ "epoch": 10.279475982532752,
+ "grad_norm": 0.030454808846116066,
+ "learning_rate": 0.0006,
+ "loss": 5.0889997482299805,
+ "step": 740
+ },
+ {
+ "epoch": 10.293449781659389,
+ "grad_norm": 0.030969638377428055,
+ "learning_rate": 0.0006,
+ "loss": 4.984112739562988,
+ "step": 741
+ },
+ {
+ "epoch": 10.307423580786025,
+ "grad_norm": 0.0276983380317688,
+ "learning_rate": 0.0006,
+ "loss": 4.955412864685059,
+ "step": 742
+ },
+ {
+ "epoch": 10.321397379912664,
+ "grad_norm": 0.02716052532196045,
+ "learning_rate": 0.0006,
+ "loss": 5.0098876953125,
+ "step": 743
+ },
+ {
+ "epoch": 10.335371179039301,
+ "grad_norm": 0.02403552085161209,
+ "learning_rate": 0.0006,
+ "loss": 5.165160655975342,
+ "step": 744
+ },
+ {
+ "epoch": 10.34934497816594,
+ "grad_norm": 0.0257862601429224,
+ "learning_rate": 0.0006,
+ "loss": 5.001547813415527,
+ "step": 745
+ },
+ {
+ "epoch": 10.363318777292577,
+ "grad_norm": 0.02300378680229187,
+ "learning_rate": 0.0006,
+ "loss": 5.038888931274414,
+ "step": 746
+ },
+ {
+ "epoch": 10.377292576419213,
+ "grad_norm": 0.019766854122281075,
+ "learning_rate": 0.0006,
+ "loss": 4.924291610717773,
+ "step": 747
+ },
+ {
+ "epoch": 10.391266375545852,
+ "grad_norm": 0.01888395845890045,
+ "learning_rate": 0.0006,
+ "loss": 5.059971332550049,
+ "step": 748
+ },
+ {
+ "epoch": 10.405240174672489,
+ "grad_norm": 0.015577499754726887,
+ "learning_rate": 0.0006,
+ "loss": 4.987686634063721,
+ "step": 749
+ },
+ {
+ "epoch": 10.419213973799126,
+ "grad_norm": 0.016012346372008324,
+ "learning_rate": 0.0006,
+ "loss": 5.015157699584961,
+ "step": 750
+ },
+ {
+ "epoch": 10.433187772925764,
+ "grad_norm": 0.016277998685836792,
+ "learning_rate": 0.0006,
+ "loss": 5.004931449890137,
+ "step": 751
+ },
+ {
+ "epoch": 10.447161572052401,
+ "grad_norm": 0.017504561692476273,
+ "learning_rate": 0.0006,
+ "loss": 4.902827739715576,
+ "step": 752
+ },
+ {
+ "epoch": 10.46113537117904,
+ "grad_norm": 0.01569022797048092,
+ "learning_rate": 0.0006,
+ "loss": 4.954863548278809,
+ "step": 753
+ },
+ {
+ "epoch": 10.475109170305677,
+ "grad_norm": 0.014114422723650932,
+ "learning_rate": 0.0006,
+ "loss": 5.00656795501709,
+ "step": 754
+ },
+ {
+ "epoch": 10.489082969432314,
+ "grad_norm": 0.01573832333087921,
+ "learning_rate": 0.0006,
+ "loss": 4.978353023529053,
+ "step": 755
+ },
+ {
+ "epoch": 10.503056768558952,
+ "grad_norm": 0.016480036079883575,
+ "learning_rate": 0.0006,
+ "loss": 4.953327655792236,
+ "step": 756
+ },
+ {
+ "epoch": 10.51703056768559,
+ "grad_norm": 0.014008025638759136,
+ "learning_rate": 0.0006,
+ "loss": 5.050840377807617,
+ "step": 757
+ },
+ {
+ "epoch": 10.531004366812226,
+ "grad_norm": 0.013147998601198196,
+ "learning_rate": 0.0006,
+ "loss": 4.974964141845703,
+ "step": 758
+ },
+ {
+ "epoch": 10.544978165938865,
+ "grad_norm": 0.013807603158056736,
+ "learning_rate": 0.0006,
+ "loss": 4.927907943725586,
+ "step": 759
+ },
+ {
+ "epoch": 10.558951965065502,
+ "grad_norm": 0.013555224984884262,
+ "learning_rate": 0.0006,
+ "loss": 4.980690956115723,
+ "step": 760
+ },
+ {
+ "epoch": 10.57292576419214,
+ "grad_norm": 0.013045408762991428,
+ "learning_rate": 0.0006,
+ "loss": 4.9436116218566895,
+ "step": 761
+ },
+ {
+ "epoch": 10.586899563318777,
+ "grad_norm": 0.012858862057328224,
+ "learning_rate": 0.0006,
+ "loss": 4.9595770835876465,
+ "step": 762
+ },
+ {
+ "epoch": 10.600873362445414,
+ "grad_norm": 0.01468253880739212,
+ "learning_rate": 0.0006,
+ "loss": 4.90725040435791,
+ "step": 763
+ },
+ {
+ "epoch": 10.614847161572053,
+ "grad_norm": 0.013224679045379162,
+ "learning_rate": 0.0006,
+ "loss": 4.877615928649902,
+ "step": 764
+ },
+ {
+ "epoch": 10.62882096069869,
+ "grad_norm": 0.013555348850786686,
+ "learning_rate": 0.0006,
+ "loss": 4.9456610679626465,
+ "step": 765
+ },
+ {
+ "epoch": 10.642794759825328,
+ "grad_norm": 0.014387160539627075,
+ "learning_rate": 0.0006,
+ "loss": 4.9233717918396,
+ "step": 766
+ },
+ {
+ "epoch": 10.656768558951965,
+ "grad_norm": 0.014354088343679905,
+ "learning_rate": 0.0006,
+ "loss": 4.983511924743652,
+ "step": 767
+ },
+ {
+ "epoch": 10.670742358078602,
+ "grad_norm": 0.01332700252532959,
+ "learning_rate": 0.0006,
+ "loss": 4.934416770935059,
+ "step": 768
+ },
+ {
+ "epoch": 10.68471615720524,
+ "grad_norm": 0.011543313041329384,
+ "learning_rate": 0.0006,
+ "loss": 4.96388053894043,
+ "step": 769
+ },
+ {
+ "epoch": 10.698689956331878,
+ "grad_norm": 0.012330878525972366,
+ "learning_rate": 0.0006,
+ "loss": 4.946396350860596,
+ "step": 770
+ },
+ {
+ "epoch": 10.712663755458514,
+ "grad_norm": 0.01379779726266861,
+ "learning_rate": 0.0006,
+ "loss": 4.981019020080566,
+ "step": 771
+ },
+ {
+ "epoch": 10.726637554585153,
+ "grad_norm": 0.014612431637942791,
+ "learning_rate": 0.0006,
+ "loss": 4.894567966461182,
+ "step": 772
+ },
+ {
+ "epoch": 10.74061135371179,
+ "grad_norm": 0.013062899932265282,
+ "learning_rate": 0.0006,
+ "loss": 5.033038139343262,
+ "step": 773
+ },
+ {
+ "epoch": 10.754585152838429,
+ "grad_norm": 0.013898049481213093,
+ "learning_rate": 0.0006,
+ "loss": 4.823636054992676,
+ "step": 774
+ },
+ {
+ "epoch": 10.768558951965066,
+ "grad_norm": 0.01502019353210926,
+ "learning_rate": 0.0006,
+ "loss": 5.020156383514404,
+ "step": 775
+ },
+ {
+ "epoch": 10.782532751091702,
+ "grad_norm": 0.013810204342007637,
+ "learning_rate": 0.0006,
+ "loss": 4.935647010803223,
+ "step": 776
+ },
+ {
+ "epoch": 10.796506550218341,
+ "grad_norm": 0.013616513460874557,
+ "learning_rate": 0.0006,
+ "loss": 4.901486396789551,
+ "step": 777
+ },
+ {
+ "epoch": 10.810480349344978,
+ "grad_norm": 0.013601024635136127,
+ "learning_rate": 0.0006,
+ "loss": 4.931873321533203,
+ "step": 778
+ },
+ {
+ "epoch": 10.824454148471617,
+ "grad_norm": 0.013668350875377655,
+ "learning_rate": 0.0006,
+ "loss": 4.937256813049316,
+ "step": 779
+ },
+ {
+ "epoch": 10.838427947598253,
+ "grad_norm": 0.01581823080778122,
+ "learning_rate": 0.0006,
+ "loss": 4.9611406326293945,
+ "step": 780
+ },
+ {
+ "epoch": 10.85240174672489,
+ "grad_norm": 0.019478484988212585,
+ "learning_rate": 0.0006,
+ "loss": 4.9212236404418945,
+ "step": 781
+ },
+ {
+ "epoch": 10.866375545851529,
+ "grad_norm": 0.026921333745121956,
+ "learning_rate": 0.0006,
+ "loss": 4.88501501083374,
+ "step": 782
+ },
+ {
+ "epoch": 10.880349344978166,
+ "grad_norm": 0.03155914694070816,
+ "learning_rate": 0.0006,
+ "loss": 4.989967346191406,
+ "step": 783
+ },
+ {
+ "epoch": 10.894323144104803,
+ "grad_norm": 0.029689429327845573,
+ "learning_rate": 0.0006,
+ "loss": 4.936000823974609,
+ "step": 784
+ },
+ {
+ "epoch": 10.908296943231441,
+ "grad_norm": 0.026626063510775566,
+ "learning_rate": 0.0006,
+ "loss": 4.984785079956055,
+ "step": 785
+ },
+ {
+ "epoch": 10.922270742358078,
+ "grad_norm": 0.026168212294578552,
+ "learning_rate": 0.0006,
+ "loss": 4.908282279968262,
+ "step": 786
+ },
+ {
+ "epoch": 10.936244541484717,
+ "grad_norm": 0.02828158251941204,
+ "learning_rate": 0.0006,
+ "loss": 4.926390647888184,
+ "step": 787
+ },
+ {
+ "epoch": 10.950218340611354,
+ "grad_norm": 0.02649078331887722,
+ "learning_rate": 0.0006,
+ "loss": 5.00400972366333,
+ "step": 788
+ },
+ {
+ "epoch": 10.96419213973799,
+ "grad_norm": 0.0256856270134449,
+ "learning_rate": 0.0006,
+ "loss": 5.0317487716674805,
+ "step": 789
+ },
+ {
+ "epoch": 10.97816593886463,
+ "grad_norm": 0.024728883057832718,
+ "learning_rate": 0.0006,
+ "loss": 4.9299635887146,
+ "step": 790
+ },
+ {
+ "epoch": 10.992139737991266,
+ "grad_norm": 0.026241116225719452,
+ "learning_rate": 0.0006,
+ "loss": 4.947319984436035,
+ "step": 791
+ },
+ {
+ "epoch": 11.0,
+ "grad_norm": 0.023526743054389954,
+ "learning_rate": 0.0006,
+ "loss": 4.94429349899292,
+ "step": 792
+ },
+ {
+ "epoch": 11.0,
+ "eval_loss": 5.104345798492432,
+ "eval_runtime": 57.0955,
+ "eval_samples_per_second": 42.77,
+ "eval_steps_per_second": 1.349,
+ "step": 792
+ },
+ {
+ "epoch": 11.013973799126637,
+ "grad_norm": 0.0216788612306118,
+ "learning_rate": 0.0006,
+ "loss": 4.881950855255127,
+ "step": 793
+ },
+ {
+ "epoch": 11.027947598253276,
+ "grad_norm": 0.01830463856458664,
+ "learning_rate": 0.0006,
+ "loss": 4.971016883850098,
+ "step": 794
+ },
+ {
+ "epoch": 11.041921397379912,
+ "grad_norm": 0.02204521745443344,
+ "learning_rate": 0.0006,
+ "loss": 4.894623756408691,
+ "step": 795
+ },
+ {
+ "epoch": 11.055895196506551,
+ "grad_norm": 0.0210255216807127,
+ "learning_rate": 0.0006,
+ "loss": 4.834009170532227,
+ "step": 796
+ },
+ {
+ "epoch": 11.069868995633188,
+ "grad_norm": 0.019797448068857193,
+ "learning_rate": 0.0006,
+ "loss": 5.086714744567871,
+ "step": 797
+ },
+ {
+ "epoch": 11.083842794759825,
+ "grad_norm": 0.02036641724407673,
+ "learning_rate": 0.0006,
+ "loss": 4.920056343078613,
+ "step": 798
+ },
+ {
+ "epoch": 11.097816593886463,
+ "grad_norm": 0.01966066285967827,
+ "learning_rate": 0.0006,
+ "loss": 4.879859447479248,
+ "step": 799
+ },
+ {
+ "epoch": 11.1117903930131,
+ "grad_norm": 0.0202149897813797,
+ "learning_rate": 0.0006,
+ "loss": 4.933640480041504,
+ "step": 800
+ },
+ {
+ "epoch": 11.125764192139737,
+ "grad_norm": 0.01893039606511593,
+ "learning_rate": 0.0006,
+ "loss": 4.901970863342285,
+ "step": 801
+ },
+ {
+ "epoch": 11.139737991266376,
+ "grad_norm": 0.016040155664086342,
+ "learning_rate": 0.0006,
+ "loss": 4.905491828918457,
+ "step": 802
+ },
+ {
+ "epoch": 11.153711790393013,
+ "grad_norm": 0.01720350608229637,
+ "learning_rate": 0.0006,
+ "loss": 4.798816680908203,
+ "step": 803
+ },
+ {
+ "epoch": 11.167685589519651,
+ "grad_norm": 0.017002852633595467,
+ "learning_rate": 0.0006,
+ "loss": 4.851859092712402,
+ "step": 804
+ },
+ {
+ "epoch": 11.181659388646288,
+ "grad_norm": 0.015594680793583393,
+ "learning_rate": 0.0006,
+ "loss": 4.90328311920166,
+ "step": 805
+ },
+ {
+ "epoch": 11.195633187772925,
+ "grad_norm": 0.014904462732374668,
+ "learning_rate": 0.0006,
+ "loss": 4.904215335845947,
+ "step": 806
+ },
+ {
+ "epoch": 11.209606986899564,
+ "grad_norm": 0.01366228237748146,
+ "learning_rate": 0.0006,
+ "loss": 4.9813055992126465,
+ "step": 807
+ },
+ {
+ "epoch": 11.2235807860262,
+ "grad_norm": 0.014719638973474503,
+ "learning_rate": 0.0006,
+ "loss": 4.927306175231934,
+ "step": 808
+ },
+ {
+ "epoch": 11.237554585152838,
+ "grad_norm": 0.015016036108136177,
+ "learning_rate": 0.0006,
+ "loss": 4.874215126037598,
+ "step": 809
+ },
+ {
+ "epoch": 11.251528384279476,
+ "grad_norm": 0.014925424009561539,
+ "learning_rate": 0.0006,
+ "loss": 4.838257789611816,
+ "step": 810
+ },
+ {
+ "epoch": 11.265502183406113,
+ "grad_norm": 0.014335823245346546,
+ "learning_rate": 0.0006,
+ "loss": 4.940608024597168,
+ "step": 811
+ },
+ {
+ "epoch": 11.279475982532752,
+ "grad_norm": 0.014987512491643429,
+ "learning_rate": 0.0006,
+ "loss": 4.831531524658203,
+ "step": 812
+ },
+ {
+ "epoch": 11.293449781659389,
+ "grad_norm": 0.016087457537651062,
+ "learning_rate": 0.0006,
+ "loss": 4.863227367401123,
+ "step": 813
+ },
+ {
+ "epoch": 11.307423580786025,
+ "grad_norm": 0.01671300269663334,
+ "learning_rate": 0.0006,
+ "loss": 4.8535661697387695,
+ "step": 814
+ },
+ {
+ "epoch": 11.321397379912664,
+ "grad_norm": 0.015229142270982265,
+ "learning_rate": 0.0006,
+ "loss": 4.948635101318359,
+ "step": 815
+ },
+ {
+ "epoch": 11.335371179039301,
+ "grad_norm": 0.01532587967813015,
+ "learning_rate": 0.0006,
+ "loss": 4.9035797119140625,
+ "step": 816
+ },
+ {
+ "epoch": 11.34934497816594,
+ "grad_norm": 0.015715476125478745,
+ "learning_rate": 0.0006,
+ "loss": 4.894709587097168,
+ "step": 817
+ },
+ {
+ "epoch": 11.363318777292577,
+ "grad_norm": 0.014156977646052837,
+ "learning_rate": 0.0006,
+ "loss": 4.960206985473633,
+ "step": 818
+ },
+ {
+ "epoch": 11.377292576419213,
+ "grad_norm": 0.017076566815376282,
+ "learning_rate": 0.0006,
+ "loss": 5.005850791931152,
+ "step": 819
+ },
+ {
+ "epoch": 11.391266375545852,
+ "grad_norm": 0.0172084029763937,
+ "learning_rate": 0.0006,
+ "loss": 4.903355598449707,
+ "step": 820
+ },
+ {
+ "epoch": 11.405240174672489,
+ "grad_norm": 0.0180258359760046,
+ "learning_rate": 0.0006,
+ "loss": 4.842667579650879,
+ "step": 821
+ },
+ {
+ "epoch": 11.419213973799126,
+ "grad_norm": 0.022130120545625687,
+ "learning_rate": 0.0006,
+ "loss": 4.799970626831055,
+ "step": 822
+ },
+ {
+ "epoch": 11.433187772925764,
+ "grad_norm": 0.02384897693991661,
+ "learning_rate": 0.0006,
+ "loss": 4.860746383666992,
+ "step": 823
+ },
+ {
+ "epoch": 11.447161572052401,
+ "grad_norm": 0.021881572902202606,
+ "learning_rate": 0.0006,
+ "loss": 4.918107032775879,
+ "step": 824
+ },
+ {
+ "epoch": 11.46113537117904,
+ "grad_norm": 0.023085080087184906,
+ "learning_rate": 0.0006,
+ "loss": 4.878946304321289,
+ "step": 825
+ },
+ {
+ "epoch": 11.475109170305677,
+ "grad_norm": 0.02291012555360794,
+ "learning_rate": 0.0006,
+ "loss": 4.958887100219727,
+ "step": 826
+ },
+ {
+ "epoch": 11.489082969432314,
+ "grad_norm": 0.0199968870729208,
+ "learning_rate": 0.0006,
+ "loss": 4.80955696105957,
+ "step": 827
+ },
+ {
+ "epoch": 11.503056768558952,
+ "grad_norm": 0.018702229484915733,
+ "learning_rate": 0.0006,
+ "loss": 5.0276408195495605,
+ "step": 828
+ },
+ {
+ "epoch": 11.51703056768559,
+ "grad_norm": 0.017787311226129532,
+ "learning_rate": 0.0006,
+ "loss": 4.96018123626709,
+ "step": 829
+ },
+ {
+ "epoch": 11.531004366812226,
+ "grad_norm": 0.01704501546919346,
+ "learning_rate": 0.0006,
+ "loss": 4.914553165435791,
+ "step": 830
+ },
+ {
+ "epoch": 11.544978165938865,
+ "grad_norm": 0.017519677057862282,
+ "learning_rate": 0.0006,
+ "loss": 4.798098087310791,
+ "step": 831
+ },
+ {
+ "epoch": 11.558951965065502,
+ "grad_norm": 0.018511991947889328,
+ "learning_rate": 0.0006,
+ "loss": 4.817785263061523,
+ "step": 832
+ },
+ {
+ "epoch": 11.57292576419214,
+ "grad_norm": 0.01876644790172577,
+ "learning_rate": 0.0006,
+ "loss": 4.818995475769043,
+ "step": 833
+ },
+ {
+ "epoch": 11.586899563318777,
+ "grad_norm": 0.01753568835556507,
+ "learning_rate": 0.0006,
+ "loss": 4.811938285827637,
+ "step": 834
+ },
+ {
+ "epoch": 11.600873362445414,
+ "grad_norm": 0.017046531662344933,
+ "learning_rate": 0.0006,
+ "loss": 4.914674282073975,
+ "step": 835
+ },
+ {
+ "epoch": 11.614847161572053,
+ "grad_norm": 0.01802094094455242,
+ "learning_rate": 0.0006,
+ "loss": 4.993284225463867,
+ "step": 836
+ },
+ {
+ "epoch": 11.62882096069869,
+ "grad_norm": 0.019493764266371727,
+ "learning_rate": 0.0006,
+ "loss": 4.895628452301025,
+ "step": 837
+ },
+ {
+ "epoch": 11.642794759825328,
+ "grad_norm": 0.019461210817098618,
+ "learning_rate": 0.0006,
+ "loss": 4.869269371032715,
+ "step": 838
+ },
+ {
+ "epoch": 11.656768558951965,
+ "grad_norm": 0.01981567218899727,
+ "learning_rate": 0.0006,
+ "loss": 4.863052845001221,
+ "step": 839
+ },
+ {
+ "epoch": 11.670742358078602,
+ "grad_norm": 0.019894491881132126,
+ "learning_rate": 0.0006,
+ "loss": 4.8775787353515625,
+ "step": 840
+ },
+ {
+ "epoch": 11.68471615720524,
+ "grad_norm": 0.01875786855816841,
+ "learning_rate": 0.0006,
+ "loss": 4.817763328552246,
+ "step": 841
+ },
+ {
+ "epoch": 11.698689956331878,
+ "grad_norm": 0.020926695317029953,
+ "learning_rate": 0.0006,
+ "loss": 4.7470526695251465,
+ "step": 842
+ },
+ {
+ "epoch": 11.712663755458514,
+ "grad_norm": 0.021477103233337402,
+ "learning_rate": 0.0006,
+ "loss": 4.762682914733887,
+ "step": 843
+ },
+ {
+ "epoch": 11.726637554585153,
+ "grad_norm": 0.020628711208701134,
+ "learning_rate": 0.0006,
+ "loss": 4.807985782623291,
+ "step": 844
+ },
+ {
+ "epoch": 11.74061135371179,
+ "grad_norm": 0.021489612758159637,
+ "learning_rate": 0.0006,
+ "loss": 4.843091011047363,
+ "step": 845
+ },
+ {
+ "epoch": 11.754585152838429,
+ "grad_norm": 0.021194491535425186,
+ "learning_rate": 0.0006,
+ "loss": 4.8300251960754395,
+ "step": 846
+ },
+ {
+ "epoch": 11.768558951965066,
+ "grad_norm": 0.01833350397646427,
+ "learning_rate": 0.0006,
+ "loss": 5.000202178955078,
+ "step": 847
+ },
+ {
+ "epoch": 11.782532751091702,
+ "grad_norm": 0.018096931278705597,
+ "learning_rate": 0.0006,
+ "loss": 4.880648136138916,
+ "step": 848
+ },
+ {
+ "epoch": 11.796506550218341,
+ "grad_norm": 0.0197161752730608,
+ "learning_rate": 0.0006,
+ "loss": 4.788316249847412,
+ "step": 849
+ },
+ {
+ "epoch": 11.810480349344978,
+ "grad_norm": 0.0203181654214859,
+ "learning_rate": 0.0006,
+ "loss": 4.870944976806641,
+ "step": 850
+ },
+ {
+ "epoch": 11.824454148471617,
+ "grad_norm": 0.01908455230295658,
+ "learning_rate": 0.0006,
+ "loss": 4.890144348144531,
+ "step": 851
+ },
+ {
+ "epoch": 11.838427947598253,
+ "grad_norm": 0.021281801164150238,
+ "learning_rate": 0.0006,
+ "loss": 4.89578914642334,
+ "step": 852
+ },
+ {
+ "epoch": 11.85240174672489,
+ "grad_norm": 0.02124273031949997,
+ "learning_rate": 0.0006,
+ "loss": 4.883749008178711,
+ "step": 853
+ },
+ {
+ "epoch": 11.866375545851529,
+ "grad_norm": 0.020969398319721222,
+ "learning_rate": 0.0006,
+ "loss": 4.809142589569092,
+ "step": 854
+ },
+ {
+ "epoch": 11.880349344978166,
+ "grad_norm": 0.01967657171189785,
+ "learning_rate": 0.0006,
+ "loss": 4.821864604949951,
+ "step": 855
+ },
+ {
+ "epoch": 11.894323144104803,
+ "grad_norm": 0.020887333899736404,
+ "learning_rate": 0.0006,
+ "loss": 4.917985916137695,
+ "step": 856
+ },
+ {
+ "epoch": 11.908296943231441,
+ "grad_norm": 0.019203083589673042,
+ "learning_rate": 0.0006,
+ "loss": 4.824960708618164,
+ "step": 857
+ },
+ {
+ "epoch": 11.922270742358078,
+ "grad_norm": 0.020162401720881462,
+ "learning_rate": 0.0006,
+ "loss": 4.772617816925049,
+ "step": 858
+ },
+ {
+ "epoch": 11.936244541484717,
+ "grad_norm": 0.02251487225294113,
+ "learning_rate": 0.0006,
+ "loss": 4.735418319702148,
+ "step": 859
+ },
+ {
+ "epoch": 11.950218340611354,
+ "grad_norm": 0.020313527435064316,
+ "learning_rate": 0.0006,
+ "loss": 4.939233779907227,
+ "step": 860
+ },
+ {
+ "epoch": 11.96419213973799,
+ "grad_norm": 0.021635599434375763,
+ "learning_rate": 0.0006,
+ "loss": 4.861807346343994,
+ "step": 861
+ },
+ {
+ "epoch": 11.97816593886463,
+ "grad_norm": 0.02285311557352543,
+ "learning_rate": 0.0006,
+ "loss": 4.932732582092285,
+ "step": 862
+ },
+ {
+ "epoch": 11.992139737991266,
+ "grad_norm": 0.02072535641491413,
+ "learning_rate": 0.0006,
+ "loss": 4.796767711639404,
+ "step": 863
+ },
+ {
+ "epoch": 12.0,
+ "grad_norm": 0.020588304847478867,
+ "learning_rate": 0.0006,
+ "loss": 4.997537612915039,
+ "step": 864
+ },
+ {
+ "epoch": 12.0,
+ "eval_loss": 5.044651985168457,
+ "eval_runtime": 56.9145,
+ "eval_samples_per_second": 42.906,
+ "eval_steps_per_second": 1.353,
+ "step": 864
+ },
+ {
+ "epoch": 12.013973799126637,
+ "grad_norm": 0.017597166821360588,
+ "learning_rate": 0.0006,
+ "loss": 4.830524444580078,
+ "step": 865
+ },
+ {
+ "epoch": 12.027947598253276,
+ "grad_norm": 0.01716788113117218,
+ "learning_rate": 0.0006,
+ "loss": 4.894325256347656,
+ "step": 866
+ },
+ {
+ "epoch": 12.041921397379912,
+ "grad_norm": 0.017712457105517387,
+ "learning_rate": 0.0006,
+ "loss": 4.868297576904297,
+ "step": 867
+ },
+ {
+ "epoch": 12.055895196506551,
+ "grad_norm": 0.021486390382051468,
+ "learning_rate": 0.0006,
+ "loss": 4.922049522399902,
+ "step": 868
+ },
+ {
+ "epoch": 12.069868995633188,
+ "grad_norm": 0.024560654535889626,
+ "learning_rate": 0.0006,
+ "loss": 4.734241008758545,
+ "step": 869
+ },
+ {
+ "epoch": 12.083842794759825,
+ "grad_norm": 0.025870252400636673,
+ "learning_rate": 0.0006,
+ "loss": 4.868566989898682,
+ "step": 870
+ },
+ {
+ "epoch": 12.097816593886463,
+ "grad_norm": 0.0277389045804739,
+ "learning_rate": 0.0006,
+ "loss": 4.755006790161133,
+ "step": 871
+ },
+ {
+ "epoch": 12.1117903930131,
+ "grad_norm": 0.029100844636559486,
+ "learning_rate": 0.0006,
+ "loss": 4.909667015075684,
+ "step": 872
+ },
+ {
+ "epoch": 12.125764192139737,
+ "grad_norm": 0.031833067536354065,
+ "learning_rate": 0.0006,
+ "loss": 4.858674049377441,
+ "step": 873
+ },
+ {
+ "epoch": 12.139737991266376,
+ "grad_norm": 0.032709237188100815,
+ "learning_rate": 0.0006,
+ "loss": 4.823696136474609,
+ "step": 874
+ },
+ {
+ "epoch": 12.153711790393013,
+ "grad_norm": 0.02995380386710167,
+ "learning_rate": 0.0006,
+ "loss": 4.850235462188721,
+ "step": 875
+ },
+ {
+ "epoch": 12.167685589519651,
+ "grad_norm": 0.034098610281944275,
+ "learning_rate": 0.0006,
+ "loss": 4.8169989585876465,
+ "step": 876
+ },
+ {
+ "epoch": 12.181659388646288,
+ "grad_norm": 0.03377080336213112,
+ "learning_rate": 0.0006,
+ "loss": 4.90738582611084,
+ "step": 877
+ },
+ {
+ "epoch": 12.195633187772925,
+ "grad_norm": 0.029780825600028038,
+ "learning_rate": 0.0006,
+ "loss": 4.811163902282715,
+ "step": 878
+ },
+ {
+ "epoch": 12.209606986899564,
+ "grad_norm": 0.03118024580180645,
+ "learning_rate": 0.0006,
+ "loss": 4.861286640167236,
+ "step": 879
+ },
+ {
+ "epoch": 12.2235807860262,
+ "grad_norm": 0.033052004873752594,
+ "learning_rate": 0.0006,
+ "loss": 4.865021228790283,
+ "step": 880
+ },
+ {
+ "epoch": 12.237554585152838,
+ "grad_norm": 0.029070032760500908,
+ "learning_rate": 0.0006,
+ "loss": 4.947303771972656,
+ "step": 881
+ },
+ {
+ "epoch": 12.251528384279476,
+ "grad_norm": 0.028355715796351433,
+ "learning_rate": 0.0006,
+ "loss": 4.804648399353027,
+ "step": 882
+ },
+ {
+ "epoch": 12.265502183406113,
+ "grad_norm": 0.031257227063179016,
+ "learning_rate": 0.0006,
+ "loss": 4.757718086242676,
+ "step": 883
+ },
+ {
+ "epoch": 12.279475982532752,
+ "grad_norm": 0.031438373029232025,
+ "learning_rate": 0.0006,
+ "loss": 4.851268768310547,
+ "step": 884
+ },
+ {
+ "epoch": 12.293449781659389,
+ "grad_norm": 0.030492138117551804,
+ "learning_rate": 0.0006,
+ "loss": 4.7891035079956055,
+ "step": 885
+ },
+ {
+ "epoch": 12.307423580786025,
+ "grad_norm": 0.026115991175174713,
+ "learning_rate": 0.0006,
+ "loss": 4.86474609375,
+ "step": 886
+ },
+ {
+ "epoch": 12.321397379912664,
+ "grad_norm": 0.024184465408325195,
+ "learning_rate": 0.0006,
+ "loss": 4.860456943511963,
+ "step": 887
+ },
+ {
+ "epoch": 12.335371179039301,
+ "grad_norm": 0.025437982752919197,
+ "learning_rate": 0.0006,
+ "loss": 4.806251049041748,
+ "step": 888
+ },
+ {
+ "epoch": 12.34934497816594,
+ "grad_norm": 0.022390231490135193,
+ "learning_rate": 0.0006,
+ "loss": 4.807086944580078,
+ "step": 889
+ },
+ {
+ "epoch": 12.363318777292577,
+ "grad_norm": 0.020174294710159302,
+ "learning_rate": 0.0006,
+ "loss": 4.78292179107666,
+ "step": 890
+ },
+ {
+ "epoch": 12.377292576419213,
+ "grad_norm": 0.019734324887394905,
+ "learning_rate": 0.0006,
+ "loss": 4.784509658813477,
+ "step": 891
+ },
+ {
+ "epoch": 12.391266375545852,
+ "grad_norm": 0.016733523458242416,
+ "learning_rate": 0.0006,
+ "loss": 4.811755180358887,
+ "step": 892
+ },
+ {
+ "epoch": 12.405240174672489,
+ "grad_norm": 0.01570146717131138,
+ "learning_rate": 0.0006,
+ "loss": 4.720088005065918,
+ "step": 893
+ },
+ {
+ "epoch": 12.419213973799126,
+ "grad_norm": 0.01601393334567547,
+ "learning_rate": 0.0006,
+ "loss": 4.783543109893799,
+ "step": 894
+ },
+ {
+ "epoch": 12.433187772925764,
+ "grad_norm": 0.014035487547516823,
+ "learning_rate": 0.0006,
+ "loss": 4.848340034484863,
+ "step": 895
+ },
+ {
+ "epoch": 12.447161572052401,
+ "grad_norm": 0.015248487703502178,
+ "learning_rate": 0.0006,
+ "loss": 4.773287773132324,
+ "step": 896
+ },
+ {
+ "epoch": 12.46113537117904,
+ "grad_norm": 0.01532268151640892,
+ "learning_rate": 0.0006,
+ "loss": 4.799646854400635,
+ "step": 897
+ },
+ {
+ "epoch": 12.475109170305677,
+ "grad_norm": 0.013982534408569336,
+ "learning_rate": 0.0006,
+ "loss": 4.8090500831604,
+ "step": 898
+ },
+ {
+ "epoch": 12.489082969432314,
+ "grad_norm": 0.014081164263188839,
+ "learning_rate": 0.0006,
+ "loss": 4.751888751983643,
+ "step": 899
+ },
+ {
+ "epoch": 12.503056768558952,
+ "grad_norm": 0.01380133256316185,
+ "learning_rate": 0.0006,
+ "loss": 4.78144645690918,
+ "step": 900
+ },
+ {
+ "epoch": 12.51703056768559,
+ "grad_norm": 0.013555269688367844,
+ "learning_rate": 0.0006,
+ "loss": 4.779257774353027,
+ "step": 901
+ },
+ {
+ "epoch": 12.531004366812226,
+ "grad_norm": 0.013075701892375946,
+ "learning_rate": 0.0006,
+ "loss": 4.820705413818359,
+ "step": 902
+ },
+ {
+ "epoch": 12.544978165938865,
+ "grad_norm": 0.01276316400617361,
+ "learning_rate": 0.0006,
+ "loss": 4.836733818054199,
+ "step": 903
+ },
+ {
+ "epoch": 12.558951965065502,
+ "grad_norm": 0.013033892959356308,
+ "learning_rate": 0.0006,
+ "loss": 4.764178276062012,
+ "step": 904
+ },
+ {
+ "epoch": 12.57292576419214,
+ "grad_norm": 0.013996980153024197,
+ "learning_rate": 0.0006,
+ "loss": 4.767983436584473,
+ "step": 905
+ },
+ {
+ "epoch": 12.586899563318777,
+ "grad_norm": 0.015222184360027313,
+ "learning_rate": 0.0006,
+ "loss": 4.759721755981445,
+ "step": 906
+ },
+ {
+ "epoch": 12.600873362445414,
+ "grad_norm": 0.01623941957950592,
+ "learning_rate": 0.0006,
+ "loss": 4.849811553955078,
+ "step": 907
+ },
+ {
+ "epoch": 12.614847161572053,
+ "grad_norm": 0.015711182728409767,
+ "learning_rate": 0.0006,
+ "loss": 4.7829084396362305,
+ "step": 908
+ },
+ {
+ "epoch": 12.62882096069869,
+ "grad_norm": 0.014137467369437218,
+ "learning_rate": 0.0006,
+ "loss": 4.73321533203125,
+ "step": 909
+ },
+ {
+ "epoch": 12.642794759825328,
+ "grad_norm": 0.012738462537527084,
+ "learning_rate": 0.0006,
+ "loss": 4.870052337646484,
+ "step": 910
+ },
+ {
+ "epoch": 12.656768558951965,
+ "grad_norm": 0.013683130033314228,
+ "learning_rate": 0.0006,
+ "loss": 4.82658052444458,
+ "step": 911
+ },
+ {
+ "epoch": 12.670742358078602,
+ "grad_norm": 0.014409830793738365,
+ "learning_rate": 0.0006,
+ "loss": 4.707494735717773,
+ "step": 912
+ },
+ {
+ "epoch": 12.68471615720524,
+ "grad_norm": 0.013806809671223164,
+ "learning_rate": 0.0006,
+ "loss": 4.778928279876709,
+ "step": 913
+ },
+ {
+ "epoch": 12.698689956331878,
+ "grad_norm": 0.012857912108302116,
+ "learning_rate": 0.0006,
+ "loss": 4.890717506408691,
+ "step": 914
+ },
+ {
+ "epoch": 12.712663755458514,
+ "grad_norm": 0.013828758150339127,
+ "learning_rate": 0.0006,
+ "loss": 4.8357439041137695,
+ "step": 915
+ },
+ {
+ "epoch": 12.726637554585153,
+ "grad_norm": 0.01478351280093193,
+ "learning_rate": 0.0006,
+ "loss": 4.869986534118652,
+ "step": 916
+ },
+ {
+ "epoch": 12.74061135371179,
+ "grad_norm": 0.01326421182602644,
+ "learning_rate": 0.0006,
+ "loss": 4.830646514892578,
+ "step": 917
+ },
+ {
+ "epoch": 12.754585152838429,
+ "grad_norm": 0.013335433788597584,
+ "learning_rate": 0.0006,
+ "loss": 4.7786149978637695,
+ "step": 918
+ },
+ {
+ "epoch": 12.768558951965066,
+ "grad_norm": 0.014261147007346153,
+ "learning_rate": 0.0006,
+ "loss": 4.788693428039551,
+ "step": 919
+ },
+ {
+ "epoch": 12.782532751091702,
+ "grad_norm": 0.014274783432483673,
+ "learning_rate": 0.0006,
+ "loss": 4.77295446395874,
+ "step": 920
+ },
+ {
+ "epoch": 12.796506550218341,
+ "grad_norm": 0.015074674971401691,
+ "learning_rate": 0.0006,
+ "loss": 4.830754280090332,
+ "step": 921
+ },
+ {
+ "epoch": 12.810480349344978,
+ "grad_norm": 0.015599515289068222,
+ "learning_rate": 0.0006,
+ "loss": 4.837759017944336,
+ "step": 922
+ },
+ {
+ "epoch": 12.824454148471617,
+ "grad_norm": 0.014595242217183113,
+ "learning_rate": 0.0006,
+ "loss": 4.882965564727783,
+ "step": 923
+ },
+ {
+ "epoch": 12.838427947598253,
+ "grad_norm": 0.014951467514038086,
+ "learning_rate": 0.0006,
+ "loss": 4.709748268127441,
+ "step": 924
+ },
+ {
+ "epoch": 12.85240174672489,
+ "grad_norm": 0.01539295818656683,
+ "learning_rate": 0.0006,
+ "loss": 4.791160583496094,
+ "step": 925
+ },
+ {
+ "epoch": 12.866375545851529,
+ "grad_norm": 0.015987906605005264,
+ "learning_rate": 0.0006,
+ "loss": 4.661876678466797,
+ "step": 926
+ },
+ {
+ "epoch": 12.880349344978166,
+ "grad_norm": 0.016527341678738594,
+ "learning_rate": 0.0006,
+ "loss": 4.721952438354492,
+ "step": 927
+ },
+ {
+ "epoch": 12.894323144104803,
+ "grad_norm": 0.01637161523103714,
+ "learning_rate": 0.0006,
+ "loss": 4.704446792602539,
+ "step": 928
+ },
+ {
+ "epoch": 12.908296943231441,
+ "grad_norm": 0.015221050009131432,
+ "learning_rate": 0.0006,
+ "loss": 4.771907806396484,
+ "step": 929
+ },
+ {
+ "epoch": 12.922270742358078,
+ "grad_norm": 0.016242166981101036,
+ "learning_rate": 0.0006,
+ "loss": 4.806267738342285,
+ "step": 930
+ },
+ {
+ "epoch": 12.936244541484717,
+ "grad_norm": 0.01628357358276844,
+ "learning_rate": 0.0006,
+ "loss": 4.7252326011657715,
+ "step": 931
+ },
+ {
+ "epoch": 12.950218340611354,
+ "grad_norm": 0.017428133636713028,
+ "learning_rate": 0.0006,
+ "loss": 4.765689849853516,
+ "step": 932
+ },
+ {
+ "epoch": 12.96419213973799,
+ "grad_norm": 0.01870061084628105,
+ "learning_rate": 0.0006,
+ "loss": 4.814488887786865,
+ "step": 933
+ },
+ {
+ "epoch": 12.97816593886463,
+ "grad_norm": 0.019796060398221016,
+ "learning_rate": 0.0006,
+ "loss": 4.739185810089111,
+ "step": 934
+ },
+ {
+ "epoch": 12.992139737991266,
+ "grad_norm": 0.01985820196568966,
+ "learning_rate": 0.0006,
+ "loss": 4.810608863830566,
+ "step": 935
+ },
+ {
+ "epoch": 13.0,
+ "grad_norm": 0.020834477618336678,
+ "learning_rate": 0.0006,
+ "loss": 4.727858066558838,
+ "step": 936
+ },
+ {
+ "epoch": 13.0,
+ "eval_loss": 4.979116916656494,
+ "eval_runtime": 57.2877,
+ "eval_samples_per_second": 42.627,
+ "eval_steps_per_second": 1.344,
+ "step": 936
+ },
+ {
+ "epoch": 13.013973799126637,
+ "grad_norm": 0.02025519125163555,
+ "learning_rate": 0.0006,
+ "loss": 4.628875732421875,
+ "step": 937
+ },
+ {
+ "epoch": 13.027947598253276,
+ "grad_norm": 0.019596368074417114,
+ "learning_rate": 0.0006,
+ "loss": 4.694699287414551,
+ "step": 938
+ },
+ {
+ "epoch": 13.041921397379912,
+ "grad_norm": 0.020494934171438217,
+ "learning_rate": 0.0006,
+ "loss": 4.691821098327637,
+ "step": 939
+ },
+ {
+ "epoch": 13.055895196506551,
+ "grad_norm": 0.020121760666370392,
+ "learning_rate": 0.0006,
+ "loss": 4.71478271484375,
+ "step": 940
+ },
+ {
+ "epoch": 13.069868995633188,
+ "grad_norm": 0.021091405302286148,
+ "learning_rate": 0.0006,
+ "loss": 4.695915222167969,
+ "step": 941
+ },
+ {
+ "epoch": 13.083842794759825,
+ "grad_norm": 0.019787423312664032,
+ "learning_rate": 0.0006,
+ "loss": 4.8652448654174805,
+ "step": 942
+ },
+ {
+ "epoch": 13.097816593886463,
+ "grad_norm": 0.0185833927243948,
+ "learning_rate": 0.0006,
+ "loss": 4.7065887451171875,
+ "step": 943
+ },
+ {
+ "epoch": 13.1117903930131,
+ "grad_norm": 0.021220475435256958,
+ "learning_rate": 0.0006,
+ "loss": 4.741232395172119,
+ "step": 944
+ },
+ {
+ "epoch": 13.125764192139737,
+ "grad_norm": 0.02136622555553913,
+ "learning_rate": 0.0006,
+ "loss": 4.705967903137207,
+ "step": 945
+ },
+ {
+ "epoch": 13.139737991266376,
+ "grad_norm": 0.0191587433218956,
+ "learning_rate": 0.0006,
+ "loss": 4.758667945861816,
+ "step": 946
+ },
+ {
+ "epoch": 13.153711790393013,
+ "grad_norm": 0.018290022388100624,
+ "learning_rate": 0.0006,
+ "loss": 4.760666847229004,
+ "step": 947
+ },
+ {
+ "epoch": 13.167685589519651,
+ "grad_norm": 0.021704409271478653,
+ "learning_rate": 0.0006,
+ "loss": 4.750555992126465,
+ "step": 948
+ },
+ {
+ "epoch": 13.181659388646288,
+ "grad_norm": 0.021668918430805206,
+ "learning_rate": 0.0006,
+ "loss": 4.840903282165527,
+ "step": 949
+ },
+ {
+ "epoch": 13.195633187772925,
+ "grad_norm": 0.019497141242027283,
+ "learning_rate": 0.0006,
+ "loss": 4.738969326019287,
+ "step": 950
+ },
+ {
+ "epoch": 13.209606986899564,
+ "grad_norm": 0.018500596284866333,
+ "learning_rate": 0.0006,
+ "loss": 4.75154972076416,
+ "step": 951
+ },
+ {
+ "epoch": 13.2235807860262,
+ "grad_norm": 0.01872239261865616,
+ "learning_rate": 0.0006,
+ "loss": 4.811738014221191,
+ "step": 952
+ },
+ {
+ "epoch": 13.237554585152838,
+ "grad_norm": 0.020321547985076904,
+ "learning_rate": 0.0006,
+ "loss": 4.815384387969971,
+ "step": 953
+ },
+ {
+ "epoch": 13.251528384279476,
+ "grad_norm": 0.02283575013279915,
+ "learning_rate": 0.0006,
+ "loss": 4.74582052230835,
+ "step": 954
+ },
+ {
+ "epoch": 13.265502183406113,
+ "grad_norm": 0.02474398910999298,
+ "learning_rate": 0.0006,
+ "loss": 4.738375186920166,
+ "step": 955
+ },
+ {
+ "epoch": 13.279475982532752,
+ "grad_norm": 0.02065764181315899,
+ "learning_rate": 0.0006,
+ "loss": 4.746875286102295,
+ "step": 956
+ },
+ {
+ "epoch": 13.293449781659389,
+ "grad_norm": 0.020534520968794823,
+ "learning_rate": 0.0006,
+ "loss": 4.656914710998535,
+ "step": 957
+ },
+ {
+ "epoch": 13.307423580786025,
+ "grad_norm": 0.0206378772854805,
+ "learning_rate": 0.0006,
+ "loss": 4.842615604400635,
+ "step": 958
+ },
+ {
+ "epoch": 13.321397379912664,
+ "grad_norm": 0.017935654148459435,
+ "learning_rate": 0.0006,
+ "loss": 4.717827796936035,
+ "step": 959
+ },
+ {
+ "epoch": 13.335371179039301,
+ "grad_norm": 0.016958363354206085,
+ "learning_rate": 0.0006,
+ "loss": 4.681070327758789,
+ "step": 960
+ },
+ {
+ "epoch": 13.34934497816594,
+ "grad_norm": 0.015735412016510963,
+ "learning_rate": 0.0006,
+ "loss": 4.7957048416137695,
+ "step": 961
+ },
+ {
+ "epoch": 13.363318777292577,
+ "grad_norm": 0.01563699170947075,
+ "learning_rate": 0.0006,
+ "loss": 4.701192378997803,
+ "step": 962
+ },
+ {
+ "epoch": 13.377292576419213,
+ "grad_norm": 0.015207760035991669,
+ "learning_rate": 0.0006,
+ "loss": 4.73256254196167,
+ "step": 963
+ },
+ {
+ "epoch": 13.391266375545852,
+ "grad_norm": 0.015396883711218834,
+ "learning_rate": 0.0006,
+ "loss": 4.76412296295166,
+ "step": 964
+ },
+ {
+ "epoch": 13.405240174672489,
+ "grad_norm": 0.015690583735704422,
+ "learning_rate": 0.0006,
+ "loss": 4.791157245635986,
+ "step": 965
+ },
+ {
+ "epoch": 13.419213973799126,
+ "grad_norm": 0.01755724661052227,
+ "learning_rate": 0.0006,
+ "loss": 4.845666885375977,
+ "step": 966
+ },
+ {
+ "epoch": 13.433187772925764,
+ "grad_norm": 0.019443422555923462,
+ "learning_rate": 0.0006,
+ "loss": 4.5902299880981445,
+ "step": 967
+ },
+ {
+ "epoch": 13.447161572052401,
+ "grad_norm": 0.02118338644504547,
+ "learning_rate": 0.0006,
+ "loss": 4.740627288818359,
+ "step": 968
+ },
+ {
+ "epoch": 13.46113537117904,
+ "grad_norm": 0.02174839936196804,
+ "learning_rate": 0.0006,
+ "loss": 4.814560890197754,
+ "step": 969
+ },
+ {
+ "epoch": 13.475109170305677,
+ "grad_norm": 0.022860374301671982,
+ "learning_rate": 0.0006,
+ "loss": 4.66440486907959,
+ "step": 970
+ },
+ {
+ "epoch": 13.489082969432314,
+ "grad_norm": 0.024742677807807922,
+ "learning_rate": 0.0006,
+ "loss": 4.7451653480529785,
+ "step": 971
+ },
+ {
+ "epoch": 13.503056768558952,
+ "grad_norm": 0.027068907395005226,
+ "learning_rate": 0.0006,
+ "loss": 4.802584171295166,
+ "step": 972
+ },
+ {
+ "epoch": 13.51703056768559,
+ "grad_norm": 0.026316625997424126,
+ "learning_rate": 0.0006,
+ "loss": 4.842397689819336,
+ "step": 973
+ },
+ {
+ "epoch": 13.531004366812226,
+ "grad_norm": 0.024610962718725204,
+ "learning_rate": 0.0006,
+ "loss": 4.650721073150635,
+ "step": 974
+ },
+ {
+ "epoch": 13.544978165938865,
+ "grad_norm": 0.024122603237628937,
+ "learning_rate": 0.0006,
+ "loss": 4.6984357833862305,
+ "step": 975
+ },
+ {
+ "epoch": 13.558951965065502,
+ "grad_norm": 0.022699708119034767,
+ "learning_rate": 0.0006,
+ "loss": 4.793007850646973,
+ "step": 976
+ },
+ {
+ "epoch": 13.57292576419214,
+ "grad_norm": 0.021706534549593925,
+ "learning_rate": 0.0006,
+ "loss": 4.691362380981445,
+ "step": 977
+ },
+ {
+ "epoch": 13.586899563318777,
+ "grad_norm": 0.021363291889429092,
+ "learning_rate": 0.0006,
+ "loss": 4.76762580871582,
+ "step": 978
+ },
+ {
+ "epoch": 13.600873362445414,
+ "grad_norm": 0.023476677015423775,
+ "learning_rate": 0.0006,
+ "loss": 4.777202606201172,
+ "step": 979
+ },
+ {
+ "epoch": 13.614847161572053,
+ "grad_norm": 0.02457100711762905,
+ "learning_rate": 0.0006,
+ "loss": 4.590703010559082,
+ "step": 980
+ },
+ {
+ "epoch": 13.62882096069869,
+ "grad_norm": 0.02469099499285221,
+ "learning_rate": 0.0006,
+ "loss": 4.640185356140137,
+ "step": 981
+ },
+ {
+ "epoch": 13.642794759825328,
+ "grad_norm": 0.024914603680372238,
+ "learning_rate": 0.0006,
+ "loss": 4.772201061248779,
+ "step": 982
+ },
+ {
+ "epoch": 13.656768558951965,
+ "grad_norm": 0.025133173912763596,
+ "learning_rate": 0.0006,
+ "loss": 4.743337154388428,
+ "step": 983
+ },
+ {
+ "epoch": 13.670742358078602,
+ "grad_norm": 0.02034137211740017,
+ "learning_rate": 0.0006,
+ "loss": 4.752127647399902,
+ "step": 984
+ },
+ {
+ "epoch": 13.68471615720524,
+ "grad_norm": 0.019645661115646362,
+ "learning_rate": 0.0006,
+ "loss": 4.838453769683838,
+ "step": 985
+ },
+ {
+ "epoch": 13.698689956331878,
+ "grad_norm": 0.01900297962129116,
+ "learning_rate": 0.0006,
+ "loss": 4.804868221282959,
+ "step": 986
+ },
+ {
+ "epoch": 13.712663755458514,
+ "grad_norm": 0.020292341709136963,
+ "learning_rate": 0.0006,
+ "loss": 4.609743118286133,
+ "step": 987
+ },
+ {
+ "epoch": 13.726637554585153,
+ "grad_norm": 0.017014775425195694,
+ "learning_rate": 0.0006,
+ "loss": 4.740612983703613,
+ "step": 988
+ },
+ {
+ "epoch": 13.74061135371179,
+ "grad_norm": 0.015199844725430012,
+ "learning_rate": 0.0006,
+ "loss": 4.719701766967773,
+ "step": 989
+ },
+ {
+ "epoch": 13.754585152838429,
+ "grad_norm": 0.015125414356589317,
+ "learning_rate": 0.0006,
+ "loss": 4.686114311218262,
+ "step": 990
+ },
+ {
+ "epoch": 13.768558951965066,
+ "grad_norm": 0.014779540710151196,
+ "learning_rate": 0.0006,
+ "loss": 4.674347877502441,
+ "step": 991
+ },
+ {
+ "epoch": 13.782532751091702,
+ "grad_norm": 0.015255163423717022,
+ "learning_rate": 0.0006,
+ "loss": 4.6309309005737305,
+ "step": 992
+ },
+ {
+ "epoch": 13.796506550218341,
+ "grad_norm": 0.013692197389900684,
+ "learning_rate": 0.0006,
+ "loss": 4.698497295379639,
+ "step": 993
+ },
+ {
+ "epoch": 13.810480349344978,
+ "grad_norm": 0.014529623091220856,
+ "learning_rate": 0.0006,
+ "loss": 4.628443717956543,
+ "step": 994
+ },
+ {
+ "epoch": 13.824454148471617,
+ "grad_norm": 0.01576659083366394,
+ "learning_rate": 0.0006,
+ "loss": 4.788015842437744,
+ "step": 995
+ },
+ {
+ "epoch": 13.838427947598253,
+ "grad_norm": 0.015834277495741844,
+ "learning_rate": 0.0006,
+ "loss": 4.672703742980957,
+ "step": 996
+ },
+ {
+ "epoch": 13.85240174672489,
+ "grad_norm": 0.015539503656327724,
+ "learning_rate": 0.0006,
+ "loss": 4.731503486633301,
+ "step": 997
+ },
+ {
+ "epoch": 13.866375545851529,
+ "grad_norm": 0.013463746756315231,
+ "learning_rate": 0.0006,
+ "loss": 4.7006330490112305,
+ "step": 998
+ },
+ {
+ "epoch": 13.880349344978166,
+ "grad_norm": 0.016377726569771767,
+ "learning_rate": 0.0006,
+ "loss": 4.636074066162109,
+ "step": 999
+ },
+ {
+ "epoch": 13.894323144104803,
+ "grad_norm": 0.016660034656524658,
+ "learning_rate": 0.0006,
+ "loss": 4.742693901062012,
+ "step": 1000
+ },
+ {
+ "epoch": 13.908296943231441,
+ "grad_norm": 0.014902938157320023,
+ "learning_rate": 0.0006,
+ "loss": 4.769811630249023,
+ "step": 1001
+ },
+ {
+ "epoch": 13.922270742358078,
+ "grad_norm": 0.014883476309478283,
+ "learning_rate": 0.0006,
+ "loss": 4.731159210205078,
+ "step": 1002
+ },
+ {
+ "epoch": 13.936244541484717,
+ "grad_norm": 0.014494217932224274,
+ "learning_rate": 0.0006,
+ "loss": 4.659485816955566,
+ "step": 1003
+ },
+ {
+ "epoch": 13.950218340611354,
+ "grad_norm": 0.015426991507411003,
+ "learning_rate": 0.0006,
+ "loss": 4.755229949951172,
+ "step": 1004
+ },
+ {
+ "epoch": 13.96419213973799,
+ "grad_norm": 0.016648586839437485,
+ "learning_rate": 0.0006,
+ "loss": 4.74111270904541,
+ "step": 1005
+ },
+ {
+ "epoch": 13.97816593886463,
+ "grad_norm": 0.016980541869997978,
+ "learning_rate": 0.0006,
+ "loss": 4.74072265625,
+ "step": 1006
+ },
+ {
+ "epoch": 13.992139737991266,
+ "grad_norm": 0.018070152029395103,
+ "learning_rate": 0.0006,
+ "loss": 4.732975006103516,
+ "step": 1007
+ },
+ {
+ "epoch": 14.0,
+ "grad_norm": 0.019910665228962898,
+ "learning_rate": 0.0006,
+ "loss": 4.634697914123535,
+ "step": 1008
+ },
+ {
+ "epoch": 14.0,
+ "eval_loss": 4.850640296936035,
+ "eval_runtime": 57.3039,
+ "eval_samples_per_second": 42.615,
+ "eval_steps_per_second": 1.344,
+ "step": 1008
+ },
+ {
+ "epoch": 14.013973799126637,
+ "grad_norm": 0.01980053074657917,
+ "learning_rate": 0.0006,
+ "loss": 4.628744602203369,
+ "step": 1009
+ },
+ {
+ "epoch": 14.027947598253276,
+ "grad_norm": 0.02497623674571514,
+ "learning_rate": 0.0006,
+ "loss": 4.75930118560791,
+ "step": 1010
+ },
+ {
+ "epoch": 14.041921397379912,
+ "grad_norm": 0.02794097363948822,
+ "learning_rate": 0.0006,
+ "loss": 4.691340446472168,
+ "step": 1011
+ },
+ {
+ "epoch": 14.055895196506551,
+ "grad_norm": 0.023725535720586777,
+ "learning_rate": 0.0006,
+ "loss": 4.733664512634277,
+ "step": 1012
+ },
+ {
+ "epoch": 14.069868995633188,
+ "grad_norm": 0.022966474294662476,
+ "learning_rate": 0.0006,
+ "loss": 4.664808750152588,
+ "step": 1013
+ },
+ {
+ "epoch": 14.083842794759825,
+ "grad_norm": 0.02530563250184059,
+ "learning_rate": 0.0006,
+ "loss": 4.8047285079956055,
+ "step": 1014
+ },
+ {
+ "epoch": 14.097816593886463,
+ "grad_norm": 0.02491082064807415,
+ "learning_rate": 0.0006,
+ "loss": 4.667956352233887,
+ "step": 1015
+ },
+ {
+ "epoch": 14.1117903930131,
+ "grad_norm": 0.022110294550657272,
+ "learning_rate": 0.0006,
+ "loss": 4.666140556335449,
+ "step": 1016
+ },
+ {
+ "epoch": 14.125764192139737,
+ "grad_norm": 0.02065849117934704,
+ "learning_rate": 0.0006,
+ "loss": 4.581960678100586,
+ "step": 1017
+ },
+ {
+ "epoch": 14.139737991266376,
+ "grad_norm": 0.01941298507153988,
+ "learning_rate": 0.0006,
+ "loss": 4.614226341247559,
+ "step": 1018
+ },
+ {
+ "epoch": 14.153711790393013,
+ "grad_norm": 0.022441856563091278,
+ "learning_rate": 0.0006,
+ "loss": 4.638352394104004,
+ "step": 1019
+ },
+ {
+ "epoch": 14.167685589519651,
+ "grad_norm": 0.021519694477319717,
+ "learning_rate": 0.0006,
+ "loss": 4.700160980224609,
+ "step": 1020
+ },
+ {
+ "epoch": 14.181659388646288,
+ "grad_norm": 0.018956486135721207,
+ "learning_rate": 0.0006,
+ "loss": 4.649593353271484,
+ "step": 1021
+ },
+ {
+ "epoch": 14.195633187772925,
+ "grad_norm": 0.016681140288710594,
+ "learning_rate": 0.0006,
+ "loss": 4.603952407836914,
+ "step": 1022
+ },
+ {
+ "epoch": 14.209606986899564,
+ "grad_norm": 0.01788809522986412,
+ "learning_rate": 0.0006,
+ "loss": 4.669160842895508,
+ "step": 1023
+ },
+ {
+ "epoch": 14.2235807860262,
+ "grad_norm": 0.016047228127717972,
+ "learning_rate": 0.0006,
+ "loss": 4.6367950439453125,
+ "step": 1024
+ },
+ {
+ "epoch": 14.237554585152838,
+ "grad_norm": 0.01614239625632763,
+ "learning_rate": 0.0006,
+ "loss": 4.664529323577881,
+ "step": 1025
+ },
+ {
+ "epoch": 14.251528384279476,
+ "grad_norm": 0.016362549737095833,
+ "learning_rate": 0.0006,
+ "loss": 4.702373504638672,
+ "step": 1026
+ },
+ {
+ "epoch": 14.265502183406113,
+ "grad_norm": 0.015800409018993378,
+ "learning_rate": 0.0006,
+ "loss": 4.7594428062438965,
+ "step": 1027
+ },
+ {
+ "epoch": 14.279475982532752,
+ "grad_norm": 0.016730139032006264,
+ "learning_rate": 0.0006,
+ "loss": 4.677244186401367,
+ "step": 1028
+ },
+ {
+ "epoch": 14.293449781659389,
+ "grad_norm": 0.01661759801208973,
+ "learning_rate": 0.0006,
+ "loss": 4.633679389953613,
+ "step": 1029
+ },
+ {
+ "epoch": 14.307423580786025,
+ "grad_norm": 0.017673347145318985,
+ "learning_rate": 0.0006,
+ "loss": 4.649107456207275,
+ "step": 1030
+ },
+ {
+ "epoch": 14.321397379912664,
+ "grad_norm": 0.019243352115154266,
+ "learning_rate": 0.0006,
+ "loss": 4.611257553100586,
+ "step": 1031
+ },
+ {
+ "epoch": 14.335371179039301,
+ "grad_norm": 0.019817164167761803,
+ "learning_rate": 0.0006,
+ "loss": 4.588086128234863,
+ "step": 1032
+ },
+ {
+ "epoch": 14.34934497816594,
+ "grad_norm": 0.022292211651802063,
+ "learning_rate": 0.0006,
+ "loss": 4.783066749572754,
+ "step": 1033
+ },
+ {
+ "epoch": 14.363318777292577,
+ "grad_norm": 0.02140096202492714,
+ "learning_rate": 0.0006,
+ "loss": 4.604718208312988,
+ "step": 1034
+ },
+ {
+ "epoch": 14.377292576419213,
+ "grad_norm": 0.0187817495316267,
+ "learning_rate": 0.0006,
+ "loss": 4.721773147583008,
+ "step": 1035
+ },
+ {
+ "epoch": 14.391266375545852,
+ "grad_norm": 0.019879210740327835,
+ "learning_rate": 0.0006,
+ "loss": 4.789286136627197,
+ "step": 1036
+ },
+ {
+ "epoch": 14.405240174672489,
+ "grad_norm": 0.01885785162448883,
+ "learning_rate": 0.0006,
+ "loss": 4.669743537902832,
+ "step": 1037
+ },
+ {
+ "epoch": 14.419213973799126,
+ "grad_norm": 0.018324270844459534,
+ "learning_rate": 0.0006,
+ "loss": 4.65901517868042,
+ "step": 1038
+ },
+ {
+ "epoch": 14.433187772925764,
+ "grad_norm": 0.017664821818470955,
+ "learning_rate": 0.0006,
+ "loss": 4.667839527130127,
+ "step": 1039
+ },
+ {
+ "epoch": 14.447161572052401,
+ "grad_norm": 0.015702905133366585,
+ "learning_rate": 0.0006,
+ "loss": 4.64700984954834,
+ "step": 1040
+ },
+ {
+ "epoch": 14.46113537117904,
+ "grad_norm": 0.016674194484949112,
+ "learning_rate": 0.0006,
+ "loss": 4.694093227386475,
+ "step": 1041
+ },
+ {
+ "epoch": 14.475109170305677,
+ "grad_norm": 0.016869477927684784,
+ "learning_rate": 0.0006,
+ "loss": 4.619544982910156,
+ "step": 1042
+ },
+ {
+ "epoch": 14.489082969432314,
+ "grad_norm": 0.016557540744543076,
+ "learning_rate": 0.0006,
+ "loss": 4.568667411804199,
+ "step": 1043
+ },
+ {
+ "epoch": 14.503056768558952,
+ "grad_norm": 0.014500945806503296,
+ "learning_rate": 0.0006,
+ "loss": 4.561741828918457,
+ "step": 1044
+ },
+ {
+ "epoch": 14.51703056768559,
+ "grad_norm": 0.014228739775717258,
+ "learning_rate": 0.0006,
+ "loss": 4.654277324676514,
+ "step": 1045
+ },
+ {
+ "epoch": 14.531004366812226,
+ "grad_norm": 0.013894238509237766,
+ "learning_rate": 0.0006,
+ "loss": 4.5867719650268555,
+ "step": 1046
+ },
+ {
+ "epoch": 14.544978165938865,
+ "grad_norm": 0.01550068985670805,
+ "learning_rate": 0.0006,
+ "loss": 4.61168098449707,
+ "step": 1047
+ },
+ {
+ "epoch": 14.558951965065502,
+ "grad_norm": 0.015940772369503975,
+ "learning_rate": 0.0006,
+ "loss": 4.691247940063477,
+ "step": 1048
+ },
+ {
+ "epoch": 14.57292576419214,
+ "grad_norm": 0.01714535616338253,
+ "learning_rate": 0.0006,
+ "loss": 4.61183500289917,
+ "step": 1049
+ },
+ {
+ "epoch": 14.586899563318777,
+ "grad_norm": 0.017925038933753967,
+ "learning_rate": 0.0006,
+ "loss": 4.719022750854492,
+ "step": 1050
+ },
+ {
+ "epoch": 14.600873362445414,
+ "grad_norm": 0.018141552805900574,
+ "learning_rate": 0.0006,
+ "loss": 4.760281562805176,
+ "step": 1051
+ },
+ {
+ "epoch": 14.614847161572053,
+ "grad_norm": 0.017131328582763672,
+ "learning_rate": 0.0006,
+ "loss": 4.709680080413818,
+ "step": 1052
+ },
+ {
+ "epoch": 14.62882096069869,
+ "grad_norm": 0.018730048090219498,
+ "learning_rate": 0.0006,
+ "loss": 4.5994720458984375,
+ "step": 1053
+ },
+ {
+ "epoch": 14.642794759825328,
+ "grad_norm": 0.018894298002123833,
+ "learning_rate": 0.0006,
+ "loss": 4.669070720672607,
+ "step": 1054
+ },
+ {
+ "epoch": 14.656768558951965,
+ "grad_norm": 0.017643198370933533,
+ "learning_rate": 0.0006,
+ "loss": 4.641790390014648,
+ "step": 1055
+ },
+ {
+ "epoch": 14.670742358078602,
+ "grad_norm": 0.017647244036197662,
+ "learning_rate": 0.0006,
+ "loss": 4.653899192810059,
+ "step": 1056
+ },
+ {
+ "epoch": 14.68471615720524,
+ "grad_norm": 0.017304198816418648,
+ "learning_rate": 0.0006,
+ "loss": 4.7487473487854,
+ "step": 1057
+ },
+ {
+ "epoch": 14.698689956331878,
+ "grad_norm": 0.01638447679579258,
+ "learning_rate": 0.0006,
+ "loss": 4.71945858001709,
+ "step": 1058
+ },
+ {
+ "epoch": 14.712663755458514,
+ "grad_norm": 0.015892071649432182,
+ "learning_rate": 0.0006,
+ "loss": 4.744903564453125,
+ "step": 1059
+ },
+ {
+ "epoch": 14.726637554585153,
+ "grad_norm": 0.017420614138245583,
+ "learning_rate": 0.0006,
+ "loss": 4.781563758850098,
+ "step": 1060
+ },
+ {
+ "epoch": 14.74061135371179,
+ "grad_norm": 0.016498520970344543,
+ "learning_rate": 0.0006,
+ "loss": 4.622779846191406,
+ "step": 1061
+ },
+ {
+ "epoch": 14.754585152838429,
+ "grad_norm": 0.017482127994298935,
+ "learning_rate": 0.0006,
+ "loss": 4.62825870513916,
+ "step": 1062
+ },
+ {
+ "epoch": 14.768558951965066,
+ "grad_norm": 0.019018379971385002,
+ "learning_rate": 0.0006,
+ "loss": 4.693194389343262,
+ "step": 1063
+ },
+ {
+ "epoch": 14.782532751091702,
+ "grad_norm": 0.01993914321064949,
+ "learning_rate": 0.0006,
+ "loss": 4.742903709411621,
+ "step": 1064
+ },
+ {
+ "epoch": 14.796506550218341,
+ "grad_norm": 0.02301335707306862,
+ "learning_rate": 0.0006,
+ "loss": 4.613251686096191,
+ "step": 1065
+ },
+ {
+ "epoch": 14.810480349344978,
+ "grad_norm": 0.024088222533464432,
+ "learning_rate": 0.0006,
+ "loss": 4.594305992126465,
+ "step": 1066
+ },
+ {
+ "epoch": 14.824454148471617,
+ "grad_norm": 0.02493320405483246,
+ "learning_rate": 0.0006,
+ "loss": 4.700588226318359,
+ "step": 1067
+ },
+ {
+ "epoch": 14.838427947598253,
+ "grad_norm": 0.02501937560737133,
+ "learning_rate": 0.0006,
+ "loss": 4.601742744445801,
+ "step": 1068
+ },
+ {
+ "epoch": 14.85240174672489,
+ "grad_norm": 0.023490751162171364,
+ "learning_rate": 0.0006,
+ "loss": 4.589211463928223,
+ "step": 1069
+ },
+ {
+ "epoch": 14.866375545851529,
+ "grad_norm": 0.019695665687322617,
+ "learning_rate": 0.0006,
+ "loss": 4.635081768035889,
+ "step": 1070
+ },
+ {
+ "epoch": 14.880349344978166,
+ "grad_norm": 0.019276980310678482,
+ "learning_rate": 0.0006,
+ "loss": 4.737002372741699,
+ "step": 1071
+ },
+ {
+ "epoch": 14.894323144104803,
+ "grad_norm": 0.022430511191487312,
+ "learning_rate": 0.0006,
+ "loss": 4.808295726776123,
+ "step": 1072
+ },
+ {
+ "epoch": 14.908296943231441,
+ "grad_norm": 0.021543916314840317,
+ "learning_rate": 0.0006,
+ "loss": 4.559475421905518,
+ "step": 1073
+ },
+ {
+ "epoch": 14.922270742358078,
+ "grad_norm": 0.019885241985321045,
+ "learning_rate": 0.0006,
+ "loss": 4.647948741912842,
+ "step": 1074
+ },
+ {
+ "epoch": 14.936244541484717,
+ "grad_norm": 0.02115175500512123,
+ "learning_rate": 0.0006,
+ "loss": 4.680505752563477,
+ "step": 1075
+ },
+ {
+ "epoch": 14.950218340611354,
+ "grad_norm": 0.02156151458621025,
+ "learning_rate": 0.0006,
+ "loss": 4.673477649688721,
+ "step": 1076
+ },
+ {
+ "epoch": 14.96419213973799,
+ "grad_norm": 0.02184601128101349,
+ "learning_rate": 0.0006,
+ "loss": 4.644402027130127,
+ "step": 1077
+ },
+ {
+ "epoch": 14.97816593886463,
+ "grad_norm": 0.0211983360350132,
+ "learning_rate": 0.0006,
+ "loss": 4.821558475494385,
+ "step": 1078
+ },
+ {
+ "epoch": 14.992139737991266,
+ "grad_norm": 0.019636567682027817,
+ "learning_rate": 0.0006,
+ "loss": 4.578868865966797,
+ "step": 1079
+ },
+ {
+ "epoch": 15.0,
+ "grad_norm": 0.019997350871562958,
+ "learning_rate": 0.0006,
+ "loss": 4.691307544708252,
+ "step": 1080
+ },
+ {
+ "epoch": 15.0,
+ "eval_loss": 4.872478008270264,
+ "eval_runtime": 56.8513,
+ "eval_samples_per_second": 42.954,
+ "eval_steps_per_second": 1.354,
+ "step": 1080
+ },
+ {
+ "epoch": 15.013973799126637,
+ "grad_norm": 0.018830662593245506,
+ "learning_rate": 0.0006,
+ "loss": 4.761779308319092,
+ "step": 1081
+ },
+ {
+ "epoch": 15.027947598253276,
+ "grad_norm": 0.016876481473445892,
+ "learning_rate": 0.0006,
+ "loss": 4.61148738861084,
+ "step": 1082
+ },
+ {
+ "epoch": 15.041921397379912,
+ "grad_norm": 0.015550863929092884,
+ "learning_rate": 0.0006,
+ "loss": 4.676124095916748,
+ "step": 1083
+ },
+ {
+ "epoch": 15.055895196506551,
+ "grad_norm": 0.017178896814584732,
+ "learning_rate": 0.0006,
+ "loss": 4.588935852050781,
+ "step": 1084
+ },
+ {
+ "epoch": 15.069868995633188,
+ "grad_norm": 0.015465447679162025,
+ "learning_rate": 0.0006,
+ "loss": 4.605499267578125,
+ "step": 1085
+ },
+ {
+ "epoch": 15.083842794759825,
+ "grad_norm": 0.01606195978820324,
+ "learning_rate": 0.0006,
+ "loss": 4.723373889923096,
+ "step": 1086
+ },
+ {
+ "epoch": 15.097816593886463,
+ "grad_norm": 0.016521936282515526,
+ "learning_rate": 0.0006,
+ "loss": 4.646608352661133,
+ "step": 1087
+ },
+ {
+ "epoch": 15.1117903930131,
+ "grad_norm": 0.01589205674827099,
+ "learning_rate": 0.0006,
+ "loss": 4.498138427734375,
+ "step": 1088
+ },
+ {
+ "epoch": 15.125764192139737,
+ "grad_norm": 0.01524856686592102,
+ "learning_rate": 0.0006,
+ "loss": 4.594562530517578,
+ "step": 1089
+ },
+ {
+ "epoch": 15.139737991266376,
+ "grad_norm": 0.014714641496539116,
+ "learning_rate": 0.0006,
+ "loss": 4.630744934082031,
+ "step": 1090
+ },
+ {
+ "epoch": 15.153711790393013,
+ "grad_norm": 0.015333331190049648,
+ "learning_rate": 0.0006,
+ "loss": 4.682388782501221,
+ "step": 1091
+ },
+ {
+ "epoch": 15.167685589519651,
+ "grad_norm": 0.01550073828548193,
+ "learning_rate": 0.0006,
+ "loss": 4.692652702331543,
+ "step": 1092
+ },
+ {
+ "epoch": 15.181659388646288,
+ "grad_norm": 0.014783318154513836,
+ "learning_rate": 0.0006,
+ "loss": 4.6223835945129395,
+ "step": 1093
+ },
+ {
+ "epoch": 15.195633187772925,
+ "grad_norm": 0.014372751116752625,
+ "learning_rate": 0.0006,
+ "loss": 4.52687406539917,
+ "step": 1094
+ },
+ {
+ "epoch": 15.209606986899564,
+ "grad_norm": 0.014779130928218365,
+ "learning_rate": 0.0006,
+ "loss": 4.637009620666504,
+ "step": 1095
+ },
+ {
+ "epoch": 15.2235807860262,
+ "grad_norm": 0.015697073191404343,
+ "learning_rate": 0.0006,
+ "loss": 4.682981014251709,
+ "step": 1096
+ },
+ {
+ "epoch": 15.237554585152838,
+ "grad_norm": 0.016523664817214012,
+ "learning_rate": 0.0006,
+ "loss": 4.705722808837891,
+ "step": 1097
+ },
+ {
+ "epoch": 15.251528384279476,
+ "grad_norm": 0.017633996903896332,
+ "learning_rate": 0.0006,
+ "loss": 4.574686050415039,
+ "step": 1098
+ },
+ {
+ "epoch": 15.265502183406113,
+ "grad_norm": 0.019955076277256012,
+ "learning_rate": 0.0006,
+ "loss": 4.634450912475586,
+ "step": 1099
+ },
+ {
+ "epoch": 15.279475982532752,
+ "grad_norm": 0.02132618986070156,
+ "learning_rate": 0.0006,
+ "loss": 4.56110143661499,
+ "step": 1100
+ },
+ {
+ "epoch": 15.293449781659389,
+ "grad_norm": 0.0207593385130167,
+ "learning_rate": 0.0006,
+ "loss": 4.629396438598633,
+ "step": 1101
+ },
+ {
+ "epoch": 15.307423580786025,
+ "grad_norm": 0.017882540822029114,
+ "learning_rate": 0.0006,
+ "loss": 4.730963230133057,
+ "step": 1102
+ },
+ {
+ "epoch": 15.321397379912664,
+ "grad_norm": 0.01917876861989498,
+ "learning_rate": 0.0006,
+ "loss": 4.64585542678833,
+ "step": 1103
+ },
+ {
+ "epoch": 15.335371179039301,
+ "grad_norm": 0.01943974196910858,
+ "learning_rate": 0.0006,
+ "loss": 4.656181335449219,
+ "step": 1104
+ },
+ {
+ "epoch": 15.34934497816594,
+ "grad_norm": 0.019904715940356255,
+ "learning_rate": 0.0006,
+ "loss": 4.619329452514648,
+ "step": 1105
+ },
+ {
+ "epoch": 15.363318777292577,
+ "grad_norm": 0.019174346700310707,
+ "learning_rate": 0.0006,
+ "loss": 4.647892475128174,
+ "step": 1106
+ },
+ {
+ "epoch": 15.377292576419213,
+ "grad_norm": 0.018244758248329163,
+ "learning_rate": 0.0006,
+ "loss": 4.56166934967041,
+ "step": 1107
+ },
+ {
+ "epoch": 15.391266375545852,
+ "grad_norm": 0.018914205953478813,
+ "learning_rate": 0.0006,
+ "loss": 4.648063659667969,
+ "step": 1108
+ },
+ {
+ "epoch": 15.405240174672489,
+ "grad_norm": 0.01961071416735649,
+ "learning_rate": 0.0006,
+ "loss": 4.665525436401367,
+ "step": 1109
+ },
+ {
+ "epoch": 15.419213973799126,
+ "grad_norm": 0.01841311901807785,
+ "learning_rate": 0.0006,
+ "loss": 4.611342430114746,
+ "step": 1110
+ },
+ {
+ "epoch": 15.433187772925764,
+ "grad_norm": 0.01863427273929119,
+ "learning_rate": 0.0006,
+ "loss": 4.664256572723389,
+ "step": 1111
+ },
+ {
+ "epoch": 15.447161572052401,
+ "grad_norm": 0.020506877452135086,
+ "learning_rate": 0.0006,
+ "loss": 4.631237506866455,
+ "step": 1112
+ },
+ {
+ "epoch": 15.46113537117904,
+ "grad_norm": 0.02223833277821541,
+ "learning_rate": 0.0006,
+ "loss": 4.668654441833496,
+ "step": 1113
+ },
+ {
+ "epoch": 15.475109170305677,
+ "grad_norm": 0.023336702957749367,
+ "learning_rate": 0.0006,
+ "loss": 4.562833309173584,
+ "step": 1114
+ },
+ {
+ "epoch": 15.489082969432314,
+ "grad_norm": 0.023052413016557693,
+ "learning_rate": 0.0006,
+ "loss": 4.621532917022705,
+ "step": 1115
+ },
+ {
+ "epoch": 15.503056768558952,
+ "grad_norm": 0.023420924320816994,
+ "learning_rate": 0.0006,
+ "loss": 4.665465354919434,
+ "step": 1116
+ },
+ {
+ "epoch": 15.51703056768559,
+ "grad_norm": 0.024071281775832176,
+ "learning_rate": 0.0006,
+ "loss": 4.595980644226074,
+ "step": 1117
+ },
+ {
+ "epoch": 15.531004366812226,
+ "grad_norm": 0.023406682536005974,
+ "learning_rate": 0.0006,
+ "loss": 4.642635345458984,
+ "step": 1118
+ },
+ {
+ "epoch": 15.544978165938865,
+ "grad_norm": 0.02138647809624672,
+ "learning_rate": 0.0006,
+ "loss": 4.704513072967529,
+ "step": 1119
+ },
+ {
+ "epoch": 15.558951965065502,
+ "grad_norm": 0.0221384447067976,
+ "learning_rate": 0.0006,
+ "loss": 4.567503452301025,
+ "step": 1120
+ },
+ {
+ "epoch": 15.57292576419214,
+ "grad_norm": 0.020879851654171944,
+ "learning_rate": 0.0006,
+ "loss": 4.664057731628418,
+ "step": 1121
+ },
+ {
+ "epoch": 15.586899563318777,
+ "grad_norm": 0.021478967741131783,
+ "learning_rate": 0.0006,
+ "loss": 4.539627552032471,
+ "step": 1122
+ },
+ {
+ "epoch": 15.600873362445414,
+ "grad_norm": 0.022748656570911407,
+ "learning_rate": 0.0006,
+ "loss": 4.5654497146606445,
+ "step": 1123
+ },
+ {
+ "epoch": 15.614847161572053,
+ "grad_norm": 0.021870248019695282,
+ "learning_rate": 0.0006,
+ "loss": 4.56899356842041,
+ "step": 1124
+ },
+ {
+ "epoch": 15.62882096069869,
+ "grad_norm": 0.021711355075240135,
+ "learning_rate": 0.0006,
+ "loss": 4.657808303833008,
+ "step": 1125
+ },
+ {
+ "epoch": 15.642794759825328,
+ "grad_norm": 0.022810909897089005,
+ "learning_rate": 0.0006,
+ "loss": 4.6526689529418945,
+ "step": 1126
+ },
+ {
+ "epoch": 15.656768558951965,
+ "grad_norm": 0.022561343386769295,
+ "learning_rate": 0.0006,
+ "loss": 4.534887790679932,
+ "step": 1127
+ },
+ {
+ "epoch": 15.670742358078602,
+ "grad_norm": 0.019909802824258804,
+ "learning_rate": 0.0006,
+ "loss": 4.600790023803711,
+ "step": 1128
+ },
+ {
+ "epoch": 15.68471615720524,
+ "grad_norm": 0.020092325285077095,
+ "learning_rate": 0.0006,
+ "loss": 4.773830413818359,
+ "step": 1129
+ },
+ {
+ "epoch": 15.698689956331878,
+ "grad_norm": 0.018834445625543594,
+ "learning_rate": 0.0006,
+ "loss": 4.628868103027344,
+ "step": 1130
+ },
+ {
+ "epoch": 15.712663755458514,
+ "grad_norm": 0.017552798613905907,
+ "learning_rate": 0.0006,
+ "loss": 4.647500038146973,
+ "step": 1131
+ },
+ {
+ "epoch": 15.726637554585153,
+ "grad_norm": 0.01751251146197319,
+ "learning_rate": 0.0006,
+ "loss": 4.590545654296875,
+ "step": 1132
+ },
+ {
+ "epoch": 15.74061135371179,
+ "grad_norm": 0.01699099875986576,
+ "learning_rate": 0.0006,
+ "loss": 4.509699821472168,
+ "step": 1133
+ },
+ {
+ "epoch": 15.754585152838429,
+ "grad_norm": 0.015808649361133575,
+ "learning_rate": 0.0006,
+ "loss": 4.606992244720459,
+ "step": 1134
+ },
+ {
+ "epoch": 15.768558951965066,
+ "grad_norm": 0.01655443385243416,
+ "learning_rate": 0.0006,
+ "loss": 4.585219860076904,
+ "step": 1135
+ },
+ {
+ "epoch": 15.782532751091702,
+ "grad_norm": 0.01708163693547249,
+ "learning_rate": 0.0006,
+ "loss": 4.668147087097168,
+ "step": 1136
+ },
+ {
+ "epoch": 15.796506550218341,
+ "grad_norm": 0.017161913216114044,
+ "learning_rate": 0.0006,
+ "loss": 4.7170634269714355,
+ "step": 1137
+ },
+ {
+ "epoch": 15.810480349344978,
+ "grad_norm": 0.017262674868106842,
+ "learning_rate": 0.0006,
+ "loss": 4.572132110595703,
+ "step": 1138
+ },
+ {
+ "epoch": 15.824454148471617,
+ "grad_norm": 0.017713576555252075,
+ "learning_rate": 0.0006,
+ "loss": 4.702538967132568,
+ "step": 1139
+ },
+ {
+ "epoch": 15.838427947598253,
+ "grad_norm": 0.018730733543634415,
+ "learning_rate": 0.0006,
+ "loss": 4.629721641540527,
+ "step": 1140
+ },
+ {
+ "epoch": 15.85240174672489,
+ "grad_norm": 0.017960187047719955,
+ "learning_rate": 0.0006,
+ "loss": 4.661039352416992,
+ "step": 1141
+ },
+ {
+ "epoch": 15.866375545851529,
+ "grad_norm": 0.01750079356133938,
+ "learning_rate": 0.0006,
+ "loss": 4.503710746765137,
+ "step": 1142
+ },
+ {
+ "epoch": 15.880349344978166,
+ "grad_norm": 0.016630422323942184,
+ "learning_rate": 0.0006,
+ "loss": 4.6582136154174805,
+ "step": 1143
+ },
+ {
+ "epoch": 15.894323144104803,
+ "grad_norm": 0.015896832570433617,
+ "learning_rate": 0.0006,
+ "loss": 4.596693992614746,
+ "step": 1144
+ },
+ {
+ "epoch": 15.908296943231441,
+ "grad_norm": 0.01566407084465027,
+ "learning_rate": 0.0006,
+ "loss": 4.6337199211120605,
+ "step": 1145
+ },
+ {
+ "epoch": 15.922270742358078,
+ "grad_norm": 0.016823330894112587,
+ "learning_rate": 0.0006,
+ "loss": 4.686994552612305,
+ "step": 1146
+ },
+ {
+ "epoch": 15.936244541484717,
+ "grad_norm": 0.016208255663514137,
+ "learning_rate": 0.0006,
+ "loss": 4.6205549240112305,
+ "step": 1147
+ },
+ {
+ "epoch": 15.950218340611354,
+ "grad_norm": 0.0171707421541214,
+ "learning_rate": 0.0006,
+ "loss": 4.582968711853027,
+ "step": 1148
+ },
+ {
+ "epoch": 15.96419213973799,
+ "grad_norm": 0.017009446397423744,
+ "learning_rate": 0.0006,
+ "loss": 4.510274887084961,
+ "step": 1149
+ },
+ {
+ "epoch": 15.97816593886463,
+ "grad_norm": 0.01565811224281788,
+ "learning_rate": 0.0006,
+ "loss": 4.55834436416626,
+ "step": 1150
+ },
+ {
+ "epoch": 15.992139737991266,
+ "grad_norm": 0.015276037156581879,
+ "learning_rate": 0.0006,
+ "loss": 4.537870407104492,
+ "step": 1151
+ },
+ {
+ "epoch": 16.0,
+ "grad_norm": 0.017305882647633553,
+ "learning_rate": 0.0006,
+ "loss": 4.611305236816406,
+ "step": 1152
+ },
+ {
+ "epoch": 16.0,
+ "eval_loss": 4.775513172149658,
+ "eval_runtime": 56.7341,
+ "eval_samples_per_second": 43.043,
+ "eval_steps_per_second": 1.357,
+ "step": 1152
+ },
+ {
+ "epoch": 16.01397379912664,
+ "grad_norm": 0.016752436757087708,
+ "learning_rate": 0.0006,
+ "loss": 4.541312217712402,
+ "step": 1153
+ },
+ {
+ "epoch": 16.027947598253274,
+ "grad_norm": 0.01643134094774723,
+ "learning_rate": 0.0006,
+ "loss": 4.585402965545654,
+ "step": 1154
+ },
+ {
+ "epoch": 16.041921397379912,
+ "grad_norm": 0.01888720504939556,
+ "learning_rate": 0.0006,
+ "loss": 4.597825050354004,
+ "step": 1155
+ },
+ {
+ "epoch": 16.05589519650655,
+ "grad_norm": 0.022044632583856583,
+ "learning_rate": 0.0006,
+ "loss": 4.532772064208984,
+ "step": 1156
+ },
+ {
+ "epoch": 16.069868995633186,
+ "grad_norm": 0.022986987605690956,
+ "learning_rate": 0.0006,
+ "loss": 4.551351547241211,
+ "step": 1157
+ },
+ {
+ "epoch": 16.083842794759825,
+ "grad_norm": 0.020116791129112244,
+ "learning_rate": 0.0006,
+ "loss": 4.484149932861328,
+ "step": 1158
+ },
+ {
+ "epoch": 16.097816593886463,
+ "grad_norm": 0.021442247554659843,
+ "learning_rate": 0.0006,
+ "loss": 4.515316486358643,
+ "step": 1159
+ },
+ {
+ "epoch": 16.111790393013102,
+ "grad_norm": 0.020974772050976753,
+ "learning_rate": 0.0006,
+ "loss": 4.51606559753418,
+ "step": 1160
+ },
+ {
+ "epoch": 16.125764192139737,
+ "grad_norm": 0.021274864673614502,
+ "learning_rate": 0.0006,
+ "loss": 4.587329864501953,
+ "step": 1161
+ },
+ {
+ "epoch": 16.139737991266376,
+ "grad_norm": 0.022799784317612648,
+ "learning_rate": 0.0006,
+ "loss": 4.568149089813232,
+ "step": 1162
+ },
+ {
+ "epoch": 16.153711790393015,
+ "grad_norm": 0.021784933283925056,
+ "learning_rate": 0.0006,
+ "loss": 4.515336990356445,
+ "step": 1163
+ },
+ {
+ "epoch": 16.16768558951965,
+ "grad_norm": 0.024261271581053734,
+ "learning_rate": 0.0006,
+ "loss": 4.5127081871032715,
+ "step": 1164
+ },
+ {
+ "epoch": 16.18165938864629,
+ "grad_norm": 0.02248375490307808,
+ "learning_rate": 0.0006,
+ "loss": 4.754251003265381,
+ "step": 1165
+ },
+ {
+ "epoch": 16.195633187772927,
+ "grad_norm": 0.02136492356657982,
+ "learning_rate": 0.0006,
+ "loss": 4.58193826675415,
+ "step": 1166
+ },
+ {
+ "epoch": 16.209606986899562,
+ "grad_norm": 0.020727066323161125,
+ "learning_rate": 0.0006,
+ "loss": 4.564209938049316,
+ "step": 1167
+ },
+ {
+ "epoch": 16.2235807860262,
+ "grad_norm": 0.017789119854569435,
+ "learning_rate": 0.0006,
+ "loss": 4.698680877685547,
+ "step": 1168
+ },
+ {
+ "epoch": 16.23755458515284,
+ "grad_norm": 0.017603034153580666,
+ "learning_rate": 0.0006,
+ "loss": 4.501171588897705,
+ "step": 1169
+ },
+ {
+ "epoch": 16.251528384279474,
+ "grad_norm": 0.017345862463116646,
+ "learning_rate": 0.0006,
+ "loss": 4.593590259552002,
+ "step": 1170
+ },
+ {
+ "epoch": 16.265502183406113,
+ "grad_norm": 0.018047425895929337,
+ "learning_rate": 0.0006,
+ "loss": 4.597467422485352,
+ "step": 1171
+ },
+ {
+ "epoch": 16.27947598253275,
+ "grad_norm": 0.01947222836315632,
+ "learning_rate": 0.0006,
+ "loss": 4.5072221755981445,
+ "step": 1172
+ },
+ {
+ "epoch": 16.29344978165939,
+ "grad_norm": 0.019598500803112984,
+ "learning_rate": 0.0006,
+ "loss": 4.601351737976074,
+ "step": 1173
+ },
+ {
+ "epoch": 16.307423580786025,
+ "grad_norm": 0.020002514123916626,
+ "learning_rate": 0.0006,
+ "loss": 4.580431938171387,
+ "step": 1174
+ },
+ {
+ "epoch": 16.321397379912664,
+ "grad_norm": 0.0216530729085207,
+ "learning_rate": 0.0006,
+ "loss": 4.424681663513184,
+ "step": 1175
+ },
+ {
+ "epoch": 16.335371179039303,
+ "grad_norm": 0.01949235238134861,
+ "learning_rate": 0.0006,
+ "loss": 4.751883029937744,
+ "step": 1176
+ },
+ {
+ "epoch": 16.349344978165938,
+ "grad_norm": 0.017779016867280006,
+ "learning_rate": 0.0006,
+ "loss": 4.541490077972412,
+ "step": 1177
+ },
+ {
+ "epoch": 16.363318777292577,
+ "grad_norm": 0.01909121684730053,
+ "learning_rate": 0.0006,
+ "loss": 4.594141006469727,
+ "step": 1178
+ },
+ {
+ "epoch": 16.377292576419215,
+ "grad_norm": 0.019894849509000778,
+ "learning_rate": 0.0006,
+ "loss": 4.555070400238037,
+ "step": 1179
+ },
+ {
+ "epoch": 16.39126637554585,
+ "grad_norm": 0.01993032731115818,
+ "learning_rate": 0.0006,
+ "loss": 4.521401405334473,
+ "step": 1180
+ },
+ {
+ "epoch": 16.40524017467249,
+ "grad_norm": 0.0210788045078516,
+ "learning_rate": 0.0006,
+ "loss": 4.672842979431152,
+ "step": 1181
+ },
+ {
+ "epoch": 16.419213973799128,
+ "grad_norm": 0.02274598926305771,
+ "learning_rate": 0.0006,
+ "loss": 4.629504203796387,
+ "step": 1182
+ },
+ {
+ "epoch": 16.433187772925763,
+ "grad_norm": 0.020448120310902596,
+ "learning_rate": 0.0006,
+ "loss": 4.548209190368652,
+ "step": 1183
+ },
+ {
+ "epoch": 16.4471615720524,
+ "grad_norm": 0.017565643414855003,
+ "learning_rate": 0.0006,
+ "loss": 4.529054641723633,
+ "step": 1184
+ },
+ {
+ "epoch": 16.46113537117904,
+ "grad_norm": 0.019504237920045853,
+ "learning_rate": 0.0006,
+ "loss": 4.48182487487793,
+ "step": 1185
+ },
+ {
+ "epoch": 16.475109170305675,
+ "grad_norm": 0.016823742538690567,
+ "learning_rate": 0.0006,
+ "loss": 4.608696937561035,
+ "step": 1186
+ },
+ {
+ "epoch": 16.489082969432314,
+ "grad_norm": 0.01602894812822342,
+ "learning_rate": 0.0006,
+ "loss": 4.620062351226807,
+ "step": 1187
+ },
+ {
+ "epoch": 16.503056768558952,
+ "grad_norm": 0.016615642234683037,
+ "learning_rate": 0.0006,
+ "loss": 4.69378662109375,
+ "step": 1188
+ },
+ {
+ "epoch": 16.51703056768559,
+ "grad_norm": 0.018576730042696,
+ "learning_rate": 0.0006,
+ "loss": 4.537041664123535,
+ "step": 1189
+ },
+ {
+ "epoch": 16.531004366812226,
+ "grad_norm": 0.018611151725053787,
+ "learning_rate": 0.0006,
+ "loss": 4.496169090270996,
+ "step": 1190
+ },
+ {
+ "epoch": 16.544978165938865,
+ "grad_norm": 0.018622413277626038,
+ "learning_rate": 0.0006,
+ "loss": 4.557358264923096,
+ "step": 1191
+ },
+ {
+ "epoch": 16.558951965065503,
+ "grad_norm": 0.01823544315993786,
+ "learning_rate": 0.0006,
+ "loss": 4.730792999267578,
+ "step": 1192
+ },
+ {
+ "epoch": 16.57292576419214,
+ "grad_norm": 0.018796846270561218,
+ "learning_rate": 0.0006,
+ "loss": 4.781791687011719,
+ "step": 1193
+ },
+ {
+ "epoch": 16.586899563318777,
+ "grad_norm": 0.01728292554616928,
+ "learning_rate": 0.0006,
+ "loss": 4.515257835388184,
+ "step": 1194
+ },
+ {
+ "epoch": 16.600873362445416,
+ "grad_norm": 0.01699589006602764,
+ "learning_rate": 0.0006,
+ "loss": 4.5996904373168945,
+ "step": 1195
+ },
+ {
+ "epoch": 16.61484716157205,
+ "grad_norm": 0.015905970707535744,
+ "learning_rate": 0.0006,
+ "loss": 4.601301193237305,
+ "step": 1196
+ },
+ {
+ "epoch": 16.62882096069869,
+ "grad_norm": 0.01338571310043335,
+ "learning_rate": 0.0006,
+ "loss": 4.557708740234375,
+ "step": 1197
+ },
+ {
+ "epoch": 16.64279475982533,
+ "grad_norm": 0.015298855490982533,
+ "learning_rate": 0.0006,
+ "loss": 4.542287349700928,
+ "step": 1198
+ },
+ {
+ "epoch": 16.656768558951963,
+ "grad_norm": 0.014753367751836777,
+ "learning_rate": 0.0006,
+ "loss": 4.595787048339844,
+ "step": 1199
+ },
+ {
+ "epoch": 16.670742358078602,
+ "grad_norm": 0.015527483075857162,
+ "learning_rate": 0.0006,
+ "loss": 4.607529163360596,
+ "step": 1200
+ },
+ {
+ "epoch": 16.68471615720524,
+ "grad_norm": 0.015042162500321865,
+ "learning_rate": 0.0006,
+ "loss": 4.686626434326172,
+ "step": 1201
+ },
+ {
+ "epoch": 16.69868995633188,
+ "grad_norm": 0.0159393772482872,
+ "learning_rate": 0.0006,
+ "loss": 4.47656774520874,
+ "step": 1202
+ },
+ {
+ "epoch": 16.712663755458514,
+ "grad_norm": 0.01532787922769785,
+ "learning_rate": 0.0006,
+ "loss": 4.55678653717041,
+ "step": 1203
+ },
+ {
+ "epoch": 16.726637554585153,
+ "grad_norm": 0.016332164406776428,
+ "learning_rate": 0.0006,
+ "loss": 4.51441764831543,
+ "step": 1204
+ },
+ {
+ "epoch": 16.74061135371179,
+ "grad_norm": 0.018901726230978966,
+ "learning_rate": 0.0006,
+ "loss": 4.57227897644043,
+ "step": 1205
+ },
+ {
+ "epoch": 16.754585152838427,
+ "grad_norm": 0.018589433282613754,
+ "learning_rate": 0.0006,
+ "loss": 4.550070762634277,
+ "step": 1206
+ },
+ {
+ "epoch": 16.768558951965066,
+ "grad_norm": 0.01634034886956215,
+ "learning_rate": 0.0006,
+ "loss": 4.586910724639893,
+ "step": 1207
+ },
+ {
+ "epoch": 16.782532751091704,
+ "grad_norm": 0.017599433660507202,
+ "learning_rate": 0.0006,
+ "loss": 4.5891923904418945,
+ "step": 1208
+ },
+ {
+ "epoch": 16.79650655021834,
+ "grad_norm": 0.01878446340560913,
+ "learning_rate": 0.0006,
+ "loss": 4.57227087020874,
+ "step": 1209
+ },
+ {
+ "epoch": 16.810480349344978,
+ "grad_norm": 0.01803469844162464,
+ "learning_rate": 0.0006,
+ "loss": 4.583845138549805,
+ "step": 1210
+ },
+ {
+ "epoch": 16.824454148471617,
+ "grad_norm": 0.018505144864320755,
+ "learning_rate": 0.0006,
+ "loss": 4.561151504516602,
+ "step": 1211
+ },
+ {
+ "epoch": 16.83842794759825,
+ "grad_norm": 0.01920081488788128,
+ "learning_rate": 0.0006,
+ "loss": 4.506834030151367,
+ "step": 1212
+ },
+ {
+ "epoch": 16.85240174672489,
+ "grad_norm": 0.0177712831646204,
+ "learning_rate": 0.0006,
+ "loss": 4.557195663452148,
+ "step": 1213
+ },
+ {
+ "epoch": 16.86637554585153,
+ "grad_norm": 0.019113918766379356,
+ "learning_rate": 0.0006,
+ "loss": 4.651227951049805,
+ "step": 1214
+ },
+ {
+ "epoch": 16.880349344978168,
+ "grad_norm": 0.02012539468705654,
+ "learning_rate": 0.0006,
+ "loss": 4.4952239990234375,
+ "step": 1215
+ },
+ {
+ "epoch": 16.894323144104803,
+ "grad_norm": 0.019993646070361137,
+ "learning_rate": 0.0006,
+ "loss": 4.517433166503906,
+ "step": 1216
+ },
+ {
+ "epoch": 16.90829694323144,
+ "grad_norm": 0.020041441544890404,
+ "learning_rate": 0.0006,
+ "loss": 4.489407539367676,
+ "step": 1217
+ },
+ {
+ "epoch": 16.92227074235808,
+ "grad_norm": 0.0183611661195755,
+ "learning_rate": 0.0006,
+ "loss": 4.5435638427734375,
+ "step": 1218
+ },
+ {
+ "epoch": 16.936244541484715,
+ "grad_norm": 0.019138429313898087,
+ "learning_rate": 0.0006,
+ "loss": 4.545949459075928,
+ "step": 1219
+ },
+ {
+ "epoch": 16.950218340611354,
+ "grad_norm": 0.018554266542196274,
+ "learning_rate": 0.0006,
+ "loss": 4.579620361328125,
+ "step": 1220
+ },
+ {
+ "epoch": 16.964192139737992,
+ "grad_norm": 0.0193545650690794,
+ "learning_rate": 0.0006,
+ "loss": 4.588255405426025,
+ "step": 1221
+ },
+ {
+ "epoch": 16.978165938864628,
+ "grad_norm": 0.019389165565371513,
+ "learning_rate": 0.0006,
+ "loss": 4.520698547363281,
+ "step": 1222
+ },
+ {
+ "epoch": 16.992139737991266,
+ "grad_norm": 0.016601379960775375,
+ "learning_rate": 0.0006,
+ "loss": 4.4809064865112305,
+ "step": 1223
+ },
+ {
+ "epoch": 17.0,
+ "grad_norm": 0.01771303080022335,
+ "learning_rate": 0.0006,
+ "loss": 4.721122741699219,
+ "step": 1224
+ },
+ {
+ "epoch": 17.0,
+ "eval_loss": 4.7903666496276855,
+ "eval_runtime": 56.9826,
+ "eval_samples_per_second": 42.855,
+ "eval_steps_per_second": 1.351,
+ "step": 1224
+ },
+ {
+ "epoch": 17.01397379912664,
+ "grad_norm": 0.017720289528369904,
+ "learning_rate": 0.0006,
+ "loss": 4.523834705352783,
+ "step": 1225
+ },
+ {
+ "epoch": 17.027947598253274,
+ "grad_norm": 0.020137161016464233,
+ "learning_rate": 0.0006,
+ "loss": 4.574061870574951,
+ "step": 1226
+ },
+ {
+ "epoch": 17.041921397379912,
+ "grad_norm": 0.02002314105629921,
+ "learning_rate": 0.0006,
+ "loss": 4.512407302856445,
+ "step": 1227
+ },
+ {
+ "epoch": 17.05589519650655,
+ "grad_norm": 0.019628183916211128,
+ "learning_rate": 0.0006,
+ "loss": 4.502533435821533,
+ "step": 1228
+ },
+ {
+ "epoch": 17.069868995633186,
+ "grad_norm": 0.019808458164334297,
+ "learning_rate": 0.0006,
+ "loss": 4.463156700134277,
+ "step": 1229
+ },
+ {
+ "epoch": 17.083842794759825,
+ "grad_norm": 0.021250786259770393,
+ "learning_rate": 0.0006,
+ "loss": 4.622682094573975,
+ "step": 1230
+ },
+ {
+ "epoch": 17.097816593886463,
+ "grad_norm": 0.022031694650650024,
+ "learning_rate": 0.0006,
+ "loss": 4.451882839202881,
+ "step": 1231
+ },
+ {
+ "epoch": 17.111790393013102,
+ "grad_norm": 0.01898987777531147,
+ "learning_rate": 0.0006,
+ "loss": 4.529331684112549,
+ "step": 1232
+ },
+ {
+ "epoch": 17.125764192139737,
+ "grad_norm": 0.018749834969639778,
+ "learning_rate": 0.0006,
+ "loss": 4.435815334320068,
+ "step": 1233
+ },
+ {
+ "epoch": 17.139737991266376,
+ "grad_norm": 0.01896674931049347,
+ "learning_rate": 0.0006,
+ "loss": 4.484029293060303,
+ "step": 1234
+ },
+ {
+ "epoch": 17.153711790393015,
+ "grad_norm": 0.018286997452378273,
+ "learning_rate": 0.0006,
+ "loss": 4.523510932922363,
+ "step": 1235
+ },
+ {
+ "epoch": 17.16768558951965,
+ "grad_norm": 0.017273059114813805,
+ "learning_rate": 0.0006,
+ "loss": 4.451767444610596,
+ "step": 1236
+ },
+ {
+ "epoch": 17.18165938864629,
+ "grad_norm": 0.016499150544404984,
+ "learning_rate": 0.0006,
+ "loss": 4.5568647384643555,
+ "step": 1237
+ },
+ {
+ "epoch": 17.195633187772927,
+ "grad_norm": 0.015934469178318977,
+ "learning_rate": 0.0006,
+ "loss": 4.4639997482299805,
+ "step": 1238
+ },
+ {
+ "epoch": 17.209606986899562,
+ "grad_norm": 0.01714209094643593,
+ "learning_rate": 0.0006,
+ "loss": 4.547614097595215,
+ "step": 1239
+ },
+ {
+ "epoch": 17.2235807860262,
+ "grad_norm": 0.019167495891451836,
+ "learning_rate": 0.0006,
+ "loss": 4.571097373962402,
+ "step": 1240
+ },
+ {
+ "epoch": 17.23755458515284,
+ "grad_norm": 0.018276996910572052,
+ "learning_rate": 0.0006,
+ "loss": 4.489764213562012,
+ "step": 1241
+ },
+ {
+ "epoch": 17.251528384279474,
+ "grad_norm": 0.021296866238117218,
+ "learning_rate": 0.0006,
+ "loss": 4.526913642883301,
+ "step": 1242
+ },
+ {
+ "epoch": 17.265502183406113,
+ "grad_norm": 0.022606629878282547,
+ "learning_rate": 0.0006,
+ "loss": 4.521624565124512,
+ "step": 1243
+ },
+ {
+ "epoch": 17.27947598253275,
+ "grad_norm": 0.021204711869359016,
+ "learning_rate": 0.0006,
+ "loss": 4.500864028930664,
+ "step": 1244
+ },
+ {
+ "epoch": 17.29344978165939,
+ "grad_norm": 0.01988849975168705,
+ "learning_rate": 0.0006,
+ "loss": 4.45162296295166,
+ "step": 1245
+ },
+ {
+ "epoch": 17.307423580786025,
+ "grad_norm": 0.020268384367227554,
+ "learning_rate": 0.0006,
+ "loss": 4.536198616027832,
+ "step": 1246
+ },
+ {
+ "epoch": 17.321397379912664,
+ "grad_norm": 0.020185677334666252,
+ "learning_rate": 0.0006,
+ "loss": 4.440262794494629,
+ "step": 1247
+ },
+ {
+ "epoch": 17.335371179039303,
+ "grad_norm": 0.020106853917241096,
+ "learning_rate": 0.0006,
+ "loss": 4.475907325744629,
+ "step": 1248
+ },
+ {
+ "epoch": 17.349344978165938,
+ "grad_norm": 0.018883297219872475,
+ "learning_rate": 0.0006,
+ "loss": 4.640661716461182,
+ "step": 1249
+ },
+ {
+ "epoch": 17.363318777292577,
+ "grad_norm": 0.01852620206773281,
+ "learning_rate": 0.0006,
+ "loss": 4.578243255615234,
+ "step": 1250
+ },
+ {
+ "epoch": 17.377292576419215,
+ "grad_norm": 0.01886451430618763,
+ "learning_rate": 0.0006,
+ "loss": 4.546026229858398,
+ "step": 1251
+ },
+ {
+ "epoch": 17.39126637554585,
+ "grad_norm": 0.01777452416718006,
+ "learning_rate": 0.0006,
+ "loss": 4.521542549133301,
+ "step": 1252
+ },
+ {
+ "epoch": 17.40524017467249,
+ "grad_norm": 0.016626447439193726,
+ "learning_rate": 0.0006,
+ "loss": 4.52629280090332,
+ "step": 1253
+ },
+ {
+ "epoch": 17.419213973799128,
+ "grad_norm": 0.01599477231502533,
+ "learning_rate": 0.0006,
+ "loss": 4.4796037673950195,
+ "step": 1254
+ },
+ {
+ "epoch": 17.433187772925763,
+ "grad_norm": 0.017021458595991135,
+ "learning_rate": 0.0006,
+ "loss": 4.468813419342041,
+ "step": 1255
+ },
+ {
+ "epoch": 17.4471615720524,
+ "grad_norm": 0.016925333067774773,
+ "learning_rate": 0.0006,
+ "loss": 4.584898948669434,
+ "step": 1256
+ },
+ {
+ "epoch": 17.46113537117904,
+ "grad_norm": 0.01714998669922352,
+ "learning_rate": 0.0006,
+ "loss": 4.523544788360596,
+ "step": 1257
+ },
+ {
+ "epoch": 17.475109170305675,
+ "grad_norm": 0.01802288182079792,
+ "learning_rate": 0.0006,
+ "loss": 4.595581531524658,
+ "step": 1258
+ },
+ {
+ "epoch": 17.489082969432314,
+ "grad_norm": 0.0180289875715971,
+ "learning_rate": 0.0006,
+ "loss": 4.394440650939941,
+ "step": 1259
+ },
+ {
+ "epoch": 17.503056768558952,
+ "grad_norm": 0.018504245206713676,
+ "learning_rate": 0.0006,
+ "loss": 4.4794416427612305,
+ "step": 1260
+ },
+ {
+ "epoch": 17.51703056768559,
+ "grad_norm": 0.019973233342170715,
+ "learning_rate": 0.0006,
+ "loss": 4.4241437911987305,
+ "step": 1261
+ },
+ {
+ "epoch": 17.531004366812226,
+ "grad_norm": 0.020063083618879318,
+ "learning_rate": 0.0006,
+ "loss": 4.48068380355835,
+ "step": 1262
+ },
+ {
+ "epoch": 17.544978165938865,
+ "grad_norm": 0.02066640742123127,
+ "learning_rate": 0.0006,
+ "loss": 4.628993034362793,
+ "step": 1263
+ },
+ {
+ "epoch": 17.558951965065503,
+ "grad_norm": 0.02206416428089142,
+ "learning_rate": 0.0006,
+ "loss": 4.564790725708008,
+ "step": 1264
+ },
+ {
+ "epoch": 17.57292576419214,
+ "grad_norm": 0.020640335977077484,
+ "learning_rate": 0.0006,
+ "loss": 4.578705787658691,
+ "step": 1265
+ },
+ {
+ "epoch": 17.586899563318777,
+ "grad_norm": 0.019093405455350876,
+ "learning_rate": 0.0006,
+ "loss": 4.5023603439331055,
+ "step": 1266
+ },
+ {
+ "epoch": 17.600873362445416,
+ "grad_norm": 0.019459193572402,
+ "learning_rate": 0.0006,
+ "loss": 4.594531059265137,
+ "step": 1267
+ },
+ {
+ "epoch": 17.61484716157205,
+ "grad_norm": 0.02021171525120735,
+ "learning_rate": 0.0006,
+ "loss": 4.770299911499023,
+ "step": 1268
+ },
+ {
+ "epoch": 17.62882096069869,
+ "grad_norm": 0.021161191165447235,
+ "learning_rate": 0.0006,
+ "loss": 4.5380401611328125,
+ "step": 1269
+ },
+ {
+ "epoch": 17.64279475982533,
+ "grad_norm": 0.021013403311371803,
+ "learning_rate": 0.0006,
+ "loss": 4.455524921417236,
+ "step": 1270
+ },
+ {
+ "epoch": 17.656768558951963,
+ "grad_norm": 0.020336592569947243,
+ "learning_rate": 0.0006,
+ "loss": 4.533672332763672,
+ "step": 1271
+ },
+ {
+ "epoch": 17.670742358078602,
+ "grad_norm": 0.020159434527158737,
+ "learning_rate": 0.0006,
+ "loss": 4.596151351928711,
+ "step": 1272
+ },
+ {
+ "epoch": 17.68471615720524,
+ "grad_norm": 0.020432960242033005,
+ "learning_rate": 0.0006,
+ "loss": 4.55531644821167,
+ "step": 1273
+ },
+ {
+ "epoch": 17.69868995633188,
+ "grad_norm": 0.025466229766607285,
+ "learning_rate": 0.0006,
+ "loss": 4.566912651062012,
+ "step": 1274
+ },
+ {
+ "epoch": 17.712663755458514,
+ "grad_norm": 0.026537558063864708,
+ "learning_rate": 0.0006,
+ "loss": 4.526003837585449,
+ "step": 1275
+ },
+ {
+ "epoch": 17.726637554585153,
+ "grad_norm": 0.024397607892751694,
+ "learning_rate": 0.0006,
+ "loss": 4.541755676269531,
+ "step": 1276
+ },
+ {
+ "epoch": 17.74061135371179,
+ "grad_norm": 0.022062312811613083,
+ "learning_rate": 0.0006,
+ "loss": 4.6532793045043945,
+ "step": 1277
+ },
+ {
+ "epoch": 17.754585152838427,
+ "grad_norm": 0.02204282023012638,
+ "learning_rate": 0.0006,
+ "loss": 4.582825660705566,
+ "step": 1278
+ },
+ {
+ "epoch": 17.768558951965066,
+ "grad_norm": 0.020844416692852974,
+ "learning_rate": 0.0006,
+ "loss": 4.522336483001709,
+ "step": 1279
+ },
+ {
+ "epoch": 17.782532751091704,
+ "grad_norm": 0.019815703853964806,
+ "learning_rate": 0.0006,
+ "loss": 4.607398509979248,
+ "step": 1280
+ },
+ {
+ "epoch": 17.79650655021834,
+ "grad_norm": 0.016676954925060272,
+ "learning_rate": 0.0006,
+ "loss": 4.446206092834473,
+ "step": 1281
+ },
+ {
+ "epoch": 17.810480349344978,
+ "grad_norm": 0.0166452769190073,
+ "learning_rate": 0.0006,
+ "loss": 4.505931854248047,
+ "step": 1282
+ },
+ {
+ "epoch": 17.824454148471617,
+ "grad_norm": 0.015928907319903374,
+ "learning_rate": 0.0006,
+ "loss": 4.455267429351807,
+ "step": 1283
+ },
+ {
+ "epoch": 17.83842794759825,
+ "grad_norm": 0.015219560824334621,
+ "learning_rate": 0.0006,
+ "loss": 4.5623674392700195,
+ "step": 1284
+ },
+ {
+ "epoch": 17.85240174672489,
+ "grad_norm": 0.015125355683267117,
+ "learning_rate": 0.0006,
+ "loss": 4.554104804992676,
+ "step": 1285
+ },
+ {
+ "epoch": 17.86637554585153,
+ "grad_norm": 0.015612718649208546,
+ "learning_rate": 0.0006,
+ "loss": 4.484847068786621,
+ "step": 1286
+ },
+ {
+ "epoch": 17.880349344978168,
+ "grad_norm": 0.015215662308037281,
+ "learning_rate": 0.0006,
+ "loss": 4.4802093505859375,
+ "step": 1287
+ },
+ {
+ "epoch": 17.894323144104803,
+ "grad_norm": 0.01511684525758028,
+ "learning_rate": 0.0006,
+ "loss": 4.537298679351807,
+ "step": 1288
+ },
+ {
+ "epoch": 17.90829694323144,
+ "grad_norm": 0.013841322623193264,
+ "learning_rate": 0.0006,
+ "loss": 4.427488327026367,
+ "step": 1289
+ },
+ {
+ "epoch": 17.92227074235808,
+ "grad_norm": 0.015115964226424694,
+ "learning_rate": 0.0006,
+ "loss": 4.4547119140625,
+ "step": 1290
+ },
+ {
+ "epoch": 17.936244541484715,
+ "grad_norm": 0.016411006450653076,
+ "learning_rate": 0.0006,
+ "loss": 4.52994966506958,
+ "step": 1291
+ },
+ {
+ "epoch": 17.950218340611354,
+ "grad_norm": 0.017313152551651,
+ "learning_rate": 0.0006,
+ "loss": 4.5337653160095215,
+ "step": 1292
+ },
+ {
+ "epoch": 17.964192139737992,
+ "grad_norm": 0.016559796407818794,
+ "learning_rate": 0.0006,
+ "loss": 4.5277910232543945,
+ "step": 1293
+ },
+ {
+ "epoch": 17.978165938864628,
+ "grad_norm": 0.015854761004447937,
+ "learning_rate": 0.0006,
+ "loss": 4.6334309577941895,
+ "step": 1294
+ },
+ {
+ "epoch": 17.992139737991266,
+ "grad_norm": 0.015658238902688026,
+ "learning_rate": 0.0006,
+ "loss": 4.613372802734375,
+ "step": 1295
+ },
+ {
+ "epoch": 18.0,
+ "grad_norm": 0.0169155802577734,
+ "learning_rate": 0.0006,
+ "loss": 4.527356147766113,
+ "step": 1296
+ },
+ {
+ "epoch": 18.0,
+ "eval_loss": 4.764543056488037,
+ "eval_runtime": 56.5998,
+ "eval_samples_per_second": 43.145,
+ "eval_steps_per_second": 1.36,
+ "step": 1296
+ },
+ {
+ "epoch": 18.01397379912664,
+ "grad_norm": 0.015348963439464569,
+ "learning_rate": 0.0006,
+ "loss": 4.466423988342285,
+ "step": 1297
+ },
+ {
+ "epoch": 18.027947598253274,
+ "grad_norm": 0.01768588088452816,
+ "learning_rate": 0.0006,
+ "loss": 4.545933246612549,
+ "step": 1298
+ },
+ {
+ "epoch": 18.041921397379912,
+ "grad_norm": 0.017381226643919945,
+ "learning_rate": 0.0006,
+ "loss": 4.494060039520264,
+ "step": 1299
+ },
+ {
+ "epoch": 18.05589519650655,
+ "grad_norm": 0.019002335146069527,
+ "learning_rate": 0.0006,
+ "loss": 4.395200729370117,
+ "step": 1300
+ },
+ {
+ "epoch": 18.069868995633186,
+ "grad_norm": 0.01827932707965374,
+ "learning_rate": 0.0006,
+ "loss": 4.424848556518555,
+ "step": 1301
+ },
+ {
+ "epoch": 18.083842794759825,
+ "grad_norm": 0.016383031383156776,
+ "learning_rate": 0.0006,
+ "loss": 4.494722366333008,
+ "step": 1302
+ },
+ {
+ "epoch": 18.097816593886463,
+ "grad_norm": 0.016265157610177994,
+ "learning_rate": 0.0006,
+ "loss": 4.416522979736328,
+ "step": 1303
+ },
+ {
+ "epoch": 18.111790393013102,
+ "grad_norm": 0.01702386513352394,
+ "learning_rate": 0.0006,
+ "loss": 4.600103378295898,
+ "step": 1304
+ },
+ {
+ "epoch": 18.125764192139737,
+ "grad_norm": 0.018495498225092888,
+ "learning_rate": 0.0006,
+ "loss": 4.588401794433594,
+ "step": 1305
+ },
+ {
+ "epoch": 18.139737991266376,
+ "grad_norm": 0.017921684309840202,
+ "learning_rate": 0.0006,
+ "loss": 4.478075981140137,
+ "step": 1306
+ },
+ {
+ "epoch": 18.153711790393015,
+ "grad_norm": 0.01801278069615364,
+ "learning_rate": 0.0006,
+ "loss": 4.413186550140381,
+ "step": 1307
+ },
+ {
+ "epoch": 18.16768558951965,
+ "grad_norm": 0.017211275175213814,
+ "learning_rate": 0.0006,
+ "loss": 4.461341857910156,
+ "step": 1308
+ },
+ {
+ "epoch": 18.18165938864629,
+ "grad_norm": 0.017589328810572624,
+ "learning_rate": 0.0006,
+ "loss": 4.546969413757324,
+ "step": 1309
+ },
+ {
+ "epoch": 18.195633187772927,
+ "grad_norm": 0.01797698810696602,
+ "learning_rate": 0.0006,
+ "loss": 4.530797004699707,
+ "step": 1310
+ },
+ {
+ "epoch": 18.209606986899562,
+ "grad_norm": 0.020408619195222855,
+ "learning_rate": 0.0006,
+ "loss": 4.62999153137207,
+ "step": 1311
+ },
+ {
+ "epoch": 18.2235807860262,
+ "grad_norm": 0.01994038000702858,
+ "learning_rate": 0.0006,
+ "loss": 4.586811065673828,
+ "step": 1312
+ },
+ {
+ "epoch": 18.23755458515284,
+ "grad_norm": 0.019381340593099594,
+ "learning_rate": 0.0006,
+ "loss": 4.4769368171691895,
+ "step": 1313
+ },
+ {
+ "epoch": 18.251528384279474,
+ "grad_norm": 0.020484555512666702,
+ "learning_rate": 0.0006,
+ "loss": 4.524302959442139,
+ "step": 1314
+ },
+ {
+ "epoch": 18.265502183406113,
+ "grad_norm": 0.02426416426897049,
+ "learning_rate": 0.0006,
+ "loss": 4.561705589294434,
+ "step": 1315
+ },
+ {
+ "epoch": 18.27947598253275,
+ "grad_norm": 0.024748101830482483,
+ "learning_rate": 0.0006,
+ "loss": 4.4654316902160645,
+ "step": 1316
+ },
+ {
+ "epoch": 18.29344978165939,
+ "grad_norm": 0.020703328773379326,
+ "learning_rate": 0.0006,
+ "loss": 4.53033447265625,
+ "step": 1317
+ },
+ {
+ "epoch": 18.307423580786025,
+ "grad_norm": 0.01765458658337593,
+ "learning_rate": 0.0006,
+ "loss": 4.5154266357421875,
+ "step": 1318
+ },
+ {
+ "epoch": 18.321397379912664,
+ "grad_norm": 0.018646420910954475,
+ "learning_rate": 0.0006,
+ "loss": 4.5224714279174805,
+ "step": 1319
+ },
+ {
+ "epoch": 18.335371179039303,
+ "grad_norm": 0.01907474733889103,
+ "learning_rate": 0.0006,
+ "loss": 4.5055437088012695,
+ "step": 1320
+ },
+ {
+ "epoch": 18.349344978165938,
+ "grad_norm": 0.018179846927523613,
+ "learning_rate": 0.0006,
+ "loss": 4.462161064147949,
+ "step": 1321
+ },
+ {
+ "epoch": 18.363318777292577,
+ "grad_norm": 0.016697878018021584,
+ "learning_rate": 0.0006,
+ "loss": 4.473718643188477,
+ "step": 1322
+ },
+ {
+ "epoch": 18.377292576419215,
+ "grad_norm": 0.017504019662737846,
+ "learning_rate": 0.0006,
+ "loss": 4.474348068237305,
+ "step": 1323
+ },
+ {
+ "epoch": 18.39126637554585,
+ "grad_norm": 0.018889913335442543,
+ "learning_rate": 0.0006,
+ "loss": 4.4531636238098145,
+ "step": 1324
+ },
+ {
+ "epoch": 18.40524017467249,
+ "grad_norm": 0.019167251884937286,
+ "learning_rate": 0.0006,
+ "loss": 4.4278364181518555,
+ "step": 1325
+ },
+ {
+ "epoch": 18.419213973799128,
+ "grad_norm": 0.015895741060376167,
+ "learning_rate": 0.0006,
+ "loss": 4.474592208862305,
+ "step": 1326
+ },
+ {
+ "epoch": 18.433187772925763,
+ "grad_norm": 0.01531882956624031,
+ "learning_rate": 0.0006,
+ "loss": 4.403757572174072,
+ "step": 1327
+ },
+ {
+ "epoch": 18.4471615720524,
+ "grad_norm": 0.01565849967300892,
+ "learning_rate": 0.0006,
+ "loss": 4.507664680480957,
+ "step": 1328
+ },
+ {
+ "epoch": 18.46113537117904,
+ "grad_norm": 0.01602315716445446,
+ "learning_rate": 0.0006,
+ "loss": 4.5648908615112305,
+ "step": 1329
+ },
+ {
+ "epoch": 18.475109170305675,
+ "grad_norm": 0.016467366367578506,
+ "learning_rate": 0.0006,
+ "loss": 4.428762912750244,
+ "step": 1330
+ },
+ {
+ "epoch": 18.489082969432314,
+ "grad_norm": 0.01604127697646618,
+ "learning_rate": 0.0006,
+ "loss": 4.527771472930908,
+ "step": 1331
+ },
+ {
+ "epoch": 18.503056768558952,
+ "grad_norm": 0.018825586885213852,
+ "learning_rate": 0.0006,
+ "loss": 4.485386848449707,
+ "step": 1332
+ },
+ {
+ "epoch": 18.51703056768559,
+ "grad_norm": 0.019258007407188416,
+ "learning_rate": 0.0006,
+ "loss": 4.414972305297852,
+ "step": 1333
+ },
+ {
+ "epoch": 18.531004366812226,
+ "grad_norm": 0.018192021176218987,
+ "learning_rate": 0.0006,
+ "loss": 4.412166595458984,
+ "step": 1334
+ },
+ {
+ "epoch": 18.544978165938865,
+ "grad_norm": 0.016120698302984238,
+ "learning_rate": 0.0006,
+ "loss": 4.484317779541016,
+ "step": 1335
+ },
+ {
+ "epoch": 18.558951965065503,
+ "grad_norm": 0.016496622934937477,
+ "learning_rate": 0.0006,
+ "loss": 4.514986515045166,
+ "step": 1336
+ },
+ {
+ "epoch": 18.57292576419214,
+ "grad_norm": 0.01795116998255253,
+ "learning_rate": 0.0006,
+ "loss": 4.41473388671875,
+ "step": 1337
+ },
+ {
+ "epoch": 18.586899563318777,
+ "grad_norm": 0.019062276929616928,
+ "learning_rate": 0.0006,
+ "loss": 4.601996898651123,
+ "step": 1338
+ },
+ {
+ "epoch": 18.600873362445416,
+ "grad_norm": 0.017174631357192993,
+ "learning_rate": 0.0006,
+ "loss": 4.571218490600586,
+ "step": 1339
+ },
+ {
+ "epoch": 18.61484716157205,
+ "grad_norm": 0.014015606604516506,
+ "learning_rate": 0.0006,
+ "loss": 4.526606559753418,
+ "step": 1340
+ },
+ {
+ "epoch": 18.62882096069869,
+ "grad_norm": 0.015358510427176952,
+ "learning_rate": 0.0006,
+ "loss": 4.604286193847656,
+ "step": 1341
+ },
+ {
+ "epoch": 18.64279475982533,
+ "grad_norm": 0.01619391329586506,
+ "learning_rate": 0.0006,
+ "loss": 4.560883522033691,
+ "step": 1342
+ },
+ {
+ "epoch": 18.656768558951963,
+ "grad_norm": 0.015972964465618134,
+ "learning_rate": 0.0006,
+ "loss": 4.479755401611328,
+ "step": 1343
+ },
+ {
+ "epoch": 18.670742358078602,
+ "grad_norm": 0.016861025243997574,
+ "learning_rate": 0.0006,
+ "loss": 4.474471092224121,
+ "step": 1344
+ },
+ {
+ "epoch": 18.68471615720524,
+ "grad_norm": 0.01822078227996826,
+ "learning_rate": 0.0006,
+ "loss": 4.509489059448242,
+ "step": 1345
+ },
+ {
+ "epoch": 18.69868995633188,
+ "grad_norm": 0.019415007904171944,
+ "learning_rate": 0.0006,
+ "loss": 4.508941173553467,
+ "step": 1346
+ },
+ {
+ "epoch": 18.712663755458514,
+ "grad_norm": 0.019856126978993416,
+ "learning_rate": 0.0006,
+ "loss": 4.446670055389404,
+ "step": 1347
+ },
+ {
+ "epoch": 18.726637554585153,
+ "grad_norm": 0.01842520199716091,
+ "learning_rate": 0.0006,
+ "loss": 4.524422645568848,
+ "step": 1348
+ },
+ {
+ "epoch": 18.74061135371179,
+ "grad_norm": 0.017196480184793472,
+ "learning_rate": 0.0006,
+ "loss": 4.498466968536377,
+ "step": 1349
+ },
+ {
+ "epoch": 18.754585152838427,
+ "grad_norm": 0.01823010854423046,
+ "learning_rate": 0.0006,
+ "loss": 4.534684181213379,
+ "step": 1350
+ },
+ {
+ "epoch": 18.768558951965066,
+ "grad_norm": 0.020398065447807312,
+ "learning_rate": 0.0006,
+ "loss": 4.456494331359863,
+ "step": 1351
+ },
+ {
+ "epoch": 18.782532751091704,
+ "grad_norm": 0.022992262616753578,
+ "learning_rate": 0.0006,
+ "loss": 4.476374626159668,
+ "step": 1352
+ },
+ {
+ "epoch": 18.79650655021834,
+ "grad_norm": 0.02235420234501362,
+ "learning_rate": 0.0006,
+ "loss": 4.587776184082031,
+ "step": 1353
+ },
+ {
+ "epoch": 18.810480349344978,
+ "grad_norm": 0.02093626745045185,
+ "learning_rate": 0.0006,
+ "loss": 4.424172401428223,
+ "step": 1354
+ },
+ {
+ "epoch": 18.824454148471617,
+ "grad_norm": 0.01997302658855915,
+ "learning_rate": 0.0006,
+ "loss": 4.448066711425781,
+ "step": 1355
+ },
+ {
+ "epoch": 18.83842794759825,
+ "grad_norm": 0.018265459686517715,
+ "learning_rate": 0.0006,
+ "loss": 4.436610698699951,
+ "step": 1356
+ },
+ {
+ "epoch": 18.85240174672489,
+ "grad_norm": 0.01673833839595318,
+ "learning_rate": 0.0006,
+ "loss": 4.5502214431762695,
+ "step": 1357
+ },
+ {
+ "epoch": 18.86637554585153,
+ "grad_norm": 0.017908833920955658,
+ "learning_rate": 0.0006,
+ "loss": 4.550530433654785,
+ "step": 1358
+ },
+ {
+ "epoch": 18.880349344978168,
+ "grad_norm": 0.01796654611825943,
+ "learning_rate": 0.0006,
+ "loss": 4.531463146209717,
+ "step": 1359
+ },
+ {
+ "epoch": 18.894323144104803,
+ "grad_norm": 0.01646183803677559,
+ "learning_rate": 0.0006,
+ "loss": 4.498582363128662,
+ "step": 1360
+ },
+ {
+ "epoch": 18.90829694323144,
+ "grad_norm": 0.015430327504873276,
+ "learning_rate": 0.0006,
+ "loss": 4.52571439743042,
+ "step": 1361
+ },
+ {
+ "epoch": 18.92227074235808,
+ "grad_norm": 0.016527952626347542,
+ "learning_rate": 0.0006,
+ "loss": 4.547013759613037,
+ "step": 1362
+ },
+ {
+ "epoch": 18.936244541484715,
+ "grad_norm": 0.016263660043478012,
+ "learning_rate": 0.0006,
+ "loss": 4.468754291534424,
+ "step": 1363
+ },
+ {
+ "epoch": 18.950218340611354,
+ "grad_norm": 0.016347210854291916,
+ "learning_rate": 0.0006,
+ "loss": 4.450741767883301,
+ "step": 1364
+ },
+ {
+ "epoch": 18.964192139737992,
+ "grad_norm": 0.01701398566365242,
+ "learning_rate": 0.0006,
+ "loss": 4.534899711608887,
+ "step": 1365
+ },
+ {
+ "epoch": 18.978165938864628,
+ "grad_norm": 0.017693081870675087,
+ "learning_rate": 0.0006,
+ "loss": 4.460553169250488,
+ "step": 1366
+ },
+ {
+ "epoch": 18.992139737991266,
+ "grad_norm": 0.018746308982372284,
+ "learning_rate": 0.0006,
+ "loss": 4.538792610168457,
+ "step": 1367
+ },
+ {
+ "epoch": 19.0,
+ "grad_norm": 0.020099416375160217,
+ "learning_rate": 0.0006,
+ "loss": 4.449617385864258,
+ "step": 1368
+ },
+ {
+ "epoch": 19.0,
+ "eval_loss": 4.7756571769714355,
+ "eval_runtime": 57.1804,
+ "eval_samples_per_second": 42.707,
+ "eval_steps_per_second": 1.347,
+ "step": 1368
+ },
+ {
+ "epoch": 19.01397379912664,
+ "grad_norm": 0.018726933747529984,
+ "learning_rate": 0.0006,
+ "loss": 4.446059703826904,
+ "step": 1369
+ },
+ {
+ "epoch": 19.027947598253274,
+ "grad_norm": 0.01956087350845337,
+ "learning_rate": 0.0006,
+ "loss": 4.4520039558410645,
+ "step": 1370
+ },
+ {
+ "epoch": 19.041921397379912,
+ "grad_norm": 0.020092811435461044,
+ "learning_rate": 0.0006,
+ "loss": 4.543343544006348,
+ "step": 1371
+ },
+ {
+ "epoch": 19.05589519650655,
+ "grad_norm": 0.01843622885644436,
+ "learning_rate": 0.0006,
+ "loss": 4.535579204559326,
+ "step": 1372
+ },
+ {
+ "epoch": 19.069868995633186,
+ "grad_norm": 0.016989488154649734,
+ "learning_rate": 0.0006,
+ "loss": 4.52316951751709,
+ "step": 1373
+ },
+ {
+ "epoch": 19.083842794759825,
+ "grad_norm": 0.01805422455072403,
+ "learning_rate": 0.0006,
+ "loss": 4.5393266677856445,
+ "step": 1374
+ },
+ {
+ "epoch": 19.097816593886463,
+ "grad_norm": 0.01874629408121109,
+ "learning_rate": 0.0006,
+ "loss": 4.422059059143066,
+ "step": 1375
+ },
+ {
+ "epoch": 19.111790393013102,
+ "grad_norm": 0.01804221048951149,
+ "learning_rate": 0.0006,
+ "loss": 4.50508451461792,
+ "step": 1376
+ },
+ {
+ "epoch": 19.125764192139737,
+ "grad_norm": 0.01785474829375744,
+ "learning_rate": 0.0006,
+ "loss": 4.412599563598633,
+ "step": 1377
+ },
+ {
+ "epoch": 19.139737991266376,
+ "grad_norm": 0.01971791312098503,
+ "learning_rate": 0.0006,
+ "loss": 4.518318176269531,
+ "step": 1378
+ },
+ {
+ "epoch": 19.153711790393015,
+ "grad_norm": 0.02061633951961994,
+ "learning_rate": 0.0006,
+ "loss": 4.332393169403076,
+ "step": 1379
+ },
+ {
+ "epoch": 19.16768558951965,
+ "grad_norm": 0.020563429221510887,
+ "learning_rate": 0.0006,
+ "loss": 4.445863723754883,
+ "step": 1380
+ },
+ {
+ "epoch": 19.18165938864629,
+ "grad_norm": 0.019995713606476784,
+ "learning_rate": 0.0006,
+ "loss": 4.497312068939209,
+ "step": 1381
+ },
+ {
+ "epoch": 19.195633187772927,
+ "grad_norm": 0.019823070615530014,
+ "learning_rate": 0.0006,
+ "loss": 4.475085258483887,
+ "step": 1382
+ },
+ {
+ "epoch": 19.209606986899562,
+ "grad_norm": 0.019039355218410492,
+ "learning_rate": 0.0006,
+ "loss": 4.50432014465332,
+ "step": 1383
+ },
+ {
+ "epoch": 19.2235807860262,
+ "grad_norm": 0.017876973375678062,
+ "learning_rate": 0.0006,
+ "loss": 4.384788513183594,
+ "step": 1384
+ },
+ {
+ "epoch": 19.23755458515284,
+ "grad_norm": 0.018697699531912804,
+ "learning_rate": 0.0006,
+ "loss": 4.410778045654297,
+ "step": 1385
+ },
+ {
+ "epoch": 19.251528384279474,
+ "grad_norm": 0.018826806917786598,
+ "learning_rate": 0.0006,
+ "loss": 4.524034023284912,
+ "step": 1386
+ },
+ {
+ "epoch": 19.265502183406113,
+ "grad_norm": 0.01824226602911949,
+ "learning_rate": 0.0006,
+ "loss": 4.526393890380859,
+ "step": 1387
+ },
+ {
+ "epoch": 19.27947598253275,
+ "grad_norm": 0.01753208599984646,
+ "learning_rate": 0.0006,
+ "loss": 4.493703365325928,
+ "step": 1388
+ },
+ {
+ "epoch": 19.29344978165939,
+ "grad_norm": 0.01742432825267315,
+ "learning_rate": 0.0006,
+ "loss": 4.392831802368164,
+ "step": 1389
+ },
+ {
+ "epoch": 19.307423580786025,
+ "grad_norm": 0.017099356278777122,
+ "learning_rate": 0.0006,
+ "loss": 4.578668594360352,
+ "step": 1390
+ },
+ {
+ "epoch": 19.321397379912664,
+ "grad_norm": 0.017946023494005203,
+ "learning_rate": 0.0006,
+ "loss": 4.456718444824219,
+ "step": 1391
+ },
+ {
+ "epoch": 19.335371179039303,
+ "grad_norm": 0.017705943435430527,
+ "learning_rate": 0.0006,
+ "loss": 4.553682327270508,
+ "step": 1392
+ },
+ {
+ "epoch": 19.349344978165938,
+ "grad_norm": 0.016647523269057274,
+ "learning_rate": 0.0006,
+ "loss": 4.525016784667969,
+ "step": 1393
+ },
+ {
+ "epoch": 19.363318777292577,
+ "grad_norm": 0.015407416969537735,
+ "learning_rate": 0.0006,
+ "loss": 4.500231742858887,
+ "step": 1394
+ },
+ {
+ "epoch": 19.377292576419215,
+ "grad_norm": 0.017264019697904587,
+ "learning_rate": 0.0006,
+ "loss": 4.4615349769592285,
+ "step": 1395
+ },
+ {
+ "epoch": 19.39126637554585,
+ "grad_norm": 0.01871555484831333,
+ "learning_rate": 0.0006,
+ "loss": 4.281026840209961,
+ "step": 1396
+ },
+ {
+ "epoch": 19.40524017467249,
+ "grad_norm": 0.017361650243401527,
+ "learning_rate": 0.0006,
+ "loss": 4.479382514953613,
+ "step": 1397
+ },
+ {
+ "epoch": 19.419213973799128,
+ "grad_norm": 0.019781362265348434,
+ "learning_rate": 0.0006,
+ "loss": 4.402320861816406,
+ "step": 1398
+ },
+ {
+ "epoch": 19.433187772925763,
+ "grad_norm": 0.019043298438191414,
+ "learning_rate": 0.0006,
+ "loss": 4.476983070373535,
+ "step": 1399
+ },
+ {
+ "epoch": 19.4471615720524,
+ "grad_norm": 0.017137985676527023,
+ "learning_rate": 0.0006,
+ "loss": 4.418689250946045,
+ "step": 1400
+ },
+ {
+ "epoch": 19.46113537117904,
+ "grad_norm": 0.016481833532452583,
+ "learning_rate": 0.0006,
+ "loss": 4.579145431518555,
+ "step": 1401
+ },
+ {
+ "epoch": 19.475109170305675,
+ "grad_norm": 0.016507714986801147,
+ "learning_rate": 0.0006,
+ "loss": 4.519158363342285,
+ "step": 1402
+ },
+ {
+ "epoch": 19.489082969432314,
+ "grad_norm": 0.018092872574925423,
+ "learning_rate": 0.0006,
+ "loss": 4.513748645782471,
+ "step": 1403
+ },
+ {
+ "epoch": 19.503056768558952,
+ "grad_norm": 0.018153440207242966,
+ "learning_rate": 0.0006,
+ "loss": 4.645384311676025,
+ "step": 1404
+ },
+ {
+ "epoch": 19.51703056768559,
+ "grad_norm": 0.01657119207084179,
+ "learning_rate": 0.0006,
+ "loss": 4.399757385253906,
+ "step": 1405
+ },
+ {
+ "epoch": 19.531004366812226,
+ "grad_norm": 0.0156280305236578,
+ "learning_rate": 0.0006,
+ "loss": 4.467836380004883,
+ "step": 1406
+ },
+ {
+ "epoch": 19.544978165938865,
+ "grad_norm": 0.016838403418660164,
+ "learning_rate": 0.0006,
+ "loss": 4.529176235198975,
+ "step": 1407
+ },
+ {
+ "epoch": 19.558951965065503,
+ "grad_norm": 0.01743236929178238,
+ "learning_rate": 0.0006,
+ "loss": 4.440032482147217,
+ "step": 1408
+ },
+ {
+ "epoch": 19.57292576419214,
+ "grad_norm": 0.01553610060364008,
+ "learning_rate": 0.0006,
+ "loss": 4.425588607788086,
+ "step": 1409
+ },
+ {
+ "epoch": 19.586899563318777,
+ "grad_norm": 0.017259271815419197,
+ "learning_rate": 0.0006,
+ "loss": 4.405452728271484,
+ "step": 1410
+ },
+ {
+ "epoch": 19.600873362445416,
+ "grad_norm": 0.015568481758236885,
+ "learning_rate": 0.0006,
+ "loss": 4.444279193878174,
+ "step": 1411
+ },
+ {
+ "epoch": 19.61484716157205,
+ "grad_norm": 0.015331248752772808,
+ "learning_rate": 0.0006,
+ "loss": 4.451846122741699,
+ "step": 1412
+ },
+ {
+ "epoch": 19.62882096069869,
+ "grad_norm": 0.015667378902435303,
+ "learning_rate": 0.0006,
+ "loss": 4.533931255340576,
+ "step": 1413
+ },
+ {
+ "epoch": 19.64279475982533,
+ "grad_norm": 0.016177602112293243,
+ "learning_rate": 0.0006,
+ "loss": 4.467313289642334,
+ "step": 1414
+ },
+ {
+ "epoch": 19.656768558951963,
+ "grad_norm": 0.01733129471540451,
+ "learning_rate": 0.0006,
+ "loss": 4.496893882751465,
+ "step": 1415
+ },
+ {
+ "epoch": 19.670742358078602,
+ "grad_norm": 0.016539594158530235,
+ "learning_rate": 0.0006,
+ "loss": 4.401961326599121,
+ "step": 1416
+ },
+ {
+ "epoch": 19.68471615720524,
+ "grad_norm": 0.016058551147580147,
+ "learning_rate": 0.0006,
+ "loss": 4.503375053405762,
+ "step": 1417
+ },
+ {
+ "epoch": 19.69868995633188,
+ "grad_norm": 0.01715777814388275,
+ "learning_rate": 0.0006,
+ "loss": 4.3808369636535645,
+ "step": 1418
+ },
+ {
+ "epoch": 19.712663755458514,
+ "grad_norm": 0.01681119203567505,
+ "learning_rate": 0.0006,
+ "loss": 4.5469560623168945,
+ "step": 1419
+ },
+ {
+ "epoch": 19.726637554585153,
+ "grad_norm": 0.01634550653398037,
+ "learning_rate": 0.0006,
+ "loss": 4.328197956085205,
+ "step": 1420
+ },
+ {
+ "epoch": 19.74061135371179,
+ "grad_norm": 0.01549347210675478,
+ "learning_rate": 0.0006,
+ "loss": 4.534621238708496,
+ "step": 1421
+ },
+ {
+ "epoch": 19.754585152838427,
+ "grad_norm": 0.01654665172100067,
+ "learning_rate": 0.0006,
+ "loss": 4.4724225997924805,
+ "step": 1422
+ },
+ {
+ "epoch": 19.768558951965066,
+ "grad_norm": 0.017119025811553,
+ "learning_rate": 0.0006,
+ "loss": 4.3897247314453125,
+ "step": 1423
+ },
+ {
+ "epoch": 19.782532751091704,
+ "grad_norm": 0.01603223942220211,
+ "learning_rate": 0.0006,
+ "loss": 4.412731170654297,
+ "step": 1424
+ },
+ {
+ "epoch": 19.79650655021834,
+ "grad_norm": 0.016774067655205727,
+ "learning_rate": 0.0006,
+ "loss": 4.54707145690918,
+ "step": 1425
+ },
+ {
+ "epoch": 19.810480349344978,
+ "grad_norm": 0.016453025862574577,
+ "learning_rate": 0.0006,
+ "loss": 4.462501049041748,
+ "step": 1426
+ },
+ {
+ "epoch": 19.824454148471617,
+ "grad_norm": 0.01597629487514496,
+ "learning_rate": 0.0006,
+ "loss": 4.229758262634277,
+ "step": 1427
+ },
+ {
+ "epoch": 19.83842794759825,
+ "grad_norm": 0.017999034374952316,
+ "learning_rate": 0.0006,
+ "loss": 4.389263153076172,
+ "step": 1428
+ },
+ {
+ "epoch": 19.85240174672489,
+ "grad_norm": 0.02165898121893406,
+ "learning_rate": 0.0006,
+ "loss": 4.342130661010742,
+ "step": 1429
+ },
+ {
+ "epoch": 19.86637554585153,
+ "grad_norm": 0.02222822792828083,
+ "learning_rate": 0.0006,
+ "loss": 4.492661476135254,
+ "step": 1430
+ },
+ {
+ "epoch": 19.880349344978168,
+ "grad_norm": 0.019378066062927246,
+ "learning_rate": 0.0006,
+ "loss": 4.4614105224609375,
+ "step": 1431
+ },
+ {
+ "epoch": 19.894323144104803,
+ "grad_norm": 0.019908083602786064,
+ "learning_rate": 0.0006,
+ "loss": 4.370929718017578,
+ "step": 1432
+ },
+ {
+ "epoch": 19.90829694323144,
+ "grad_norm": 0.022296471521258354,
+ "learning_rate": 0.0006,
+ "loss": 4.438320159912109,
+ "step": 1433
+ },
+ {
+ "epoch": 19.92227074235808,
+ "grad_norm": 0.021810343489050865,
+ "learning_rate": 0.0006,
+ "loss": 4.462067127227783,
+ "step": 1434
+ },
+ {
+ "epoch": 19.936244541484715,
+ "grad_norm": 0.02129209227859974,
+ "learning_rate": 0.0006,
+ "loss": 4.541261196136475,
+ "step": 1435
+ },
+ {
+ "epoch": 19.950218340611354,
+ "grad_norm": 0.02079629711806774,
+ "learning_rate": 0.0006,
+ "loss": 4.36712646484375,
+ "step": 1436
+ },
+ {
+ "epoch": 19.964192139737992,
+ "grad_norm": 0.022501740604639053,
+ "learning_rate": 0.0006,
+ "loss": 4.5279083251953125,
+ "step": 1437
+ },
+ {
+ "epoch": 19.978165938864628,
+ "grad_norm": 0.019121676683425903,
+ "learning_rate": 0.0006,
+ "loss": 4.477789878845215,
+ "step": 1438
+ },
+ {
+ "epoch": 19.992139737991266,
+ "grad_norm": 0.017429588362574577,
+ "learning_rate": 0.0006,
+ "loss": 4.4023847579956055,
+ "step": 1439
+ },
+ {
+ "epoch": 20.0,
+ "grad_norm": 0.018857665359973907,
+ "learning_rate": 0.0006,
+ "loss": 4.494743347167969,
+ "step": 1440
+ },
+ {
+ "epoch": 20.0,
+ "eval_loss": 4.694555759429932,
+ "eval_runtime": 56.4146,
+ "eval_samples_per_second": 43.287,
+ "eval_steps_per_second": 1.365,
+ "step": 1440
+ },
+ {
+ "epoch": 20.01397379912664,
+ "grad_norm": 0.016505861654877663,
+ "learning_rate": 0.0006,
+ "loss": 4.506229877471924,
+ "step": 1441
+ },
+ {
+ "epoch": 20.027947598253274,
+ "grad_norm": 0.017571842297911644,
+ "learning_rate": 0.0006,
+ "loss": 4.384073257446289,
+ "step": 1442
+ },
+ {
+ "epoch": 20.041921397379912,
+ "grad_norm": 0.019398365169763565,
+ "learning_rate": 0.0006,
+ "loss": 4.482177734375,
+ "step": 1443
+ },
+ {
+ "epoch": 20.05589519650655,
+ "grad_norm": 0.02011062018573284,
+ "learning_rate": 0.0006,
+ "loss": 4.440773963928223,
+ "step": 1444
+ },
+ {
+ "epoch": 20.069868995633186,
+ "grad_norm": 0.019026953727006912,
+ "learning_rate": 0.0006,
+ "loss": 4.507168769836426,
+ "step": 1445
+ },
+ {
+ "epoch": 20.083842794759825,
+ "grad_norm": 0.020506031811237335,
+ "learning_rate": 0.0006,
+ "loss": 4.469079971313477,
+ "step": 1446
+ },
+ {
+ "epoch": 20.097816593886463,
+ "grad_norm": 0.019608162343502045,
+ "learning_rate": 0.0006,
+ "loss": 4.395927429199219,
+ "step": 1447
+ },
+ {
+ "epoch": 20.111790393013102,
+ "grad_norm": 0.019419124349951744,
+ "learning_rate": 0.0006,
+ "loss": 4.484235763549805,
+ "step": 1448
+ },
+ {
+ "epoch": 20.125764192139737,
+ "grad_norm": 0.021775128319859505,
+ "learning_rate": 0.0006,
+ "loss": 4.333173751831055,
+ "step": 1449
+ },
+ {
+ "epoch": 20.139737991266376,
+ "grad_norm": 0.025512272492051125,
+ "learning_rate": 0.0006,
+ "loss": 4.460362434387207,
+ "step": 1450
+ },
+ {
+ "epoch": 20.153711790393015,
+ "grad_norm": 0.025157401338219643,
+ "learning_rate": 0.0006,
+ "loss": 4.4361395835876465,
+ "step": 1451
+ },
+ {
+ "epoch": 20.16768558951965,
+ "grad_norm": 0.019862616434693336,
+ "learning_rate": 0.0006,
+ "loss": 4.426807403564453,
+ "step": 1452
+ },
+ {
+ "epoch": 20.18165938864629,
+ "grad_norm": 0.01932152360677719,
+ "learning_rate": 0.0006,
+ "loss": 4.375677108764648,
+ "step": 1453
+ },
+ {
+ "epoch": 20.195633187772927,
+ "grad_norm": 0.01760999858379364,
+ "learning_rate": 0.0006,
+ "loss": 4.42259407043457,
+ "step": 1454
+ },
+ {
+ "epoch": 20.209606986899562,
+ "grad_norm": 0.016998112201690674,
+ "learning_rate": 0.0006,
+ "loss": 4.481280326843262,
+ "step": 1455
+ },
+ {
+ "epoch": 20.2235807860262,
+ "grad_norm": 0.017199190333485603,
+ "learning_rate": 0.0006,
+ "loss": 4.496712684631348,
+ "step": 1456
+ },
+ {
+ "epoch": 20.23755458515284,
+ "grad_norm": 0.016833679750561714,
+ "learning_rate": 0.0006,
+ "loss": 4.40736722946167,
+ "step": 1457
+ },
+ {
+ "epoch": 20.251528384279474,
+ "grad_norm": 0.017546730116009712,
+ "learning_rate": 0.0006,
+ "loss": 4.4915266036987305,
+ "step": 1458
+ },
+ {
+ "epoch": 20.265502183406113,
+ "grad_norm": 0.016595078632235527,
+ "learning_rate": 0.0006,
+ "loss": 4.460815906524658,
+ "step": 1459
+ },
+ {
+ "epoch": 20.27947598253275,
+ "grad_norm": 0.017728568986058235,
+ "learning_rate": 0.0006,
+ "loss": 4.346048831939697,
+ "step": 1460
+ },
+ {
+ "epoch": 20.29344978165939,
+ "grad_norm": 0.01849197782576084,
+ "learning_rate": 0.0006,
+ "loss": 4.399681091308594,
+ "step": 1461
+ },
+ {
+ "epoch": 20.307423580786025,
+ "grad_norm": 0.018336588516831398,
+ "learning_rate": 0.0006,
+ "loss": 4.414183616638184,
+ "step": 1462
+ },
+ {
+ "epoch": 20.321397379912664,
+ "grad_norm": 0.018086519092321396,
+ "learning_rate": 0.0006,
+ "loss": 4.346586227416992,
+ "step": 1463
+ },
+ {
+ "epoch": 20.335371179039303,
+ "grad_norm": 0.019697299227118492,
+ "learning_rate": 0.0006,
+ "loss": 4.4420952796936035,
+ "step": 1464
+ },
+ {
+ "epoch": 20.349344978165938,
+ "grad_norm": 0.02198074571788311,
+ "learning_rate": 0.0006,
+ "loss": 4.588729381561279,
+ "step": 1465
+ },
+ {
+ "epoch": 20.363318777292577,
+ "grad_norm": 0.022813329473137856,
+ "learning_rate": 0.0006,
+ "loss": 4.49729585647583,
+ "step": 1466
+ },
+ {
+ "epoch": 20.377292576419215,
+ "grad_norm": 0.022006649523973465,
+ "learning_rate": 0.0006,
+ "loss": 4.475641250610352,
+ "step": 1467
+ },
+ {
+ "epoch": 20.39126637554585,
+ "grad_norm": 0.023188438266515732,
+ "learning_rate": 0.0006,
+ "loss": 4.404322624206543,
+ "step": 1468
+ },
+ {
+ "epoch": 20.40524017467249,
+ "grad_norm": 0.022226877510547638,
+ "learning_rate": 0.0006,
+ "loss": 4.50904655456543,
+ "step": 1469
+ },
+ {
+ "epoch": 20.419213973799128,
+ "grad_norm": 0.021082637831568718,
+ "learning_rate": 0.0006,
+ "loss": 4.339011192321777,
+ "step": 1470
+ },
+ {
+ "epoch": 20.433187772925763,
+ "grad_norm": 0.02013542130589485,
+ "learning_rate": 0.0006,
+ "loss": 4.516942024230957,
+ "step": 1471
+ },
+ {
+ "epoch": 20.4471615720524,
+ "grad_norm": 0.01898968778550625,
+ "learning_rate": 0.0006,
+ "loss": 4.315804481506348,
+ "step": 1472
+ },
+ {
+ "epoch": 20.46113537117904,
+ "grad_norm": 0.019450880587100983,
+ "learning_rate": 0.0006,
+ "loss": 4.4794921875,
+ "step": 1473
+ },
+ {
+ "epoch": 20.475109170305675,
+ "grad_norm": 0.018577802926301956,
+ "learning_rate": 0.0006,
+ "loss": 4.422323703765869,
+ "step": 1474
+ },
+ {
+ "epoch": 20.489082969432314,
+ "grad_norm": 0.01778412237763405,
+ "learning_rate": 0.0006,
+ "loss": 4.429789066314697,
+ "step": 1475
+ },
+ {
+ "epoch": 20.503056768558952,
+ "grad_norm": 0.017134476453065872,
+ "learning_rate": 0.0006,
+ "loss": 4.416459083557129,
+ "step": 1476
+ },
+ {
+ "epoch": 20.51703056768559,
+ "grad_norm": 0.017153557389974594,
+ "learning_rate": 0.0006,
+ "loss": 4.446170806884766,
+ "step": 1477
+ },
+ {
+ "epoch": 20.531004366812226,
+ "grad_norm": 0.014998532831668854,
+ "learning_rate": 0.0006,
+ "loss": 4.419447898864746,
+ "step": 1478
+ },
+ {
+ "epoch": 20.544978165938865,
+ "grad_norm": 0.015877850353717804,
+ "learning_rate": 0.0006,
+ "loss": 4.312338829040527,
+ "step": 1479
+ },
+ {
+ "epoch": 20.558951965065503,
+ "grad_norm": 0.01569489948451519,
+ "learning_rate": 0.0006,
+ "loss": 4.572905540466309,
+ "step": 1480
+ },
+ {
+ "epoch": 20.57292576419214,
+ "grad_norm": 0.01591474376618862,
+ "learning_rate": 0.0006,
+ "loss": 4.414907455444336,
+ "step": 1481
+ },
+ {
+ "epoch": 20.586899563318777,
+ "grad_norm": 0.015395903028547764,
+ "learning_rate": 0.0006,
+ "loss": 4.549131393432617,
+ "step": 1482
+ },
+ {
+ "epoch": 20.600873362445416,
+ "grad_norm": 0.015715358778834343,
+ "learning_rate": 0.0006,
+ "loss": 4.530194282531738,
+ "step": 1483
+ },
+ {
+ "epoch": 20.61484716157205,
+ "grad_norm": 0.016042975708842278,
+ "learning_rate": 0.0006,
+ "loss": 4.322149276733398,
+ "step": 1484
+ },
+ {
+ "epoch": 20.62882096069869,
+ "grad_norm": 0.01664186641573906,
+ "learning_rate": 0.0006,
+ "loss": 4.424866676330566,
+ "step": 1485
+ },
+ {
+ "epoch": 20.64279475982533,
+ "grad_norm": 0.01631534844636917,
+ "learning_rate": 0.0006,
+ "loss": 4.441752910614014,
+ "step": 1486
+ },
+ {
+ "epoch": 20.656768558951963,
+ "grad_norm": 0.014057058840990067,
+ "learning_rate": 0.0006,
+ "loss": 4.399982452392578,
+ "step": 1487
+ },
+ {
+ "epoch": 20.670742358078602,
+ "grad_norm": 0.015076273120939732,
+ "learning_rate": 0.0006,
+ "loss": 4.441324234008789,
+ "step": 1488
+ },
+ {
+ "epoch": 20.68471615720524,
+ "grad_norm": 0.014138277620077133,
+ "learning_rate": 0.0006,
+ "loss": 4.4312543869018555,
+ "step": 1489
+ },
+ {
+ "epoch": 20.69868995633188,
+ "grad_norm": 0.013762440532445908,
+ "learning_rate": 0.0006,
+ "loss": 4.387807846069336,
+ "step": 1490
+ },
+ {
+ "epoch": 20.712663755458514,
+ "grad_norm": 0.015058411285281181,
+ "learning_rate": 0.0006,
+ "loss": 4.594747543334961,
+ "step": 1491
+ },
+ {
+ "epoch": 20.726637554585153,
+ "grad_norm": 0.013827123679220676,
+ "learning_rate": 0.0006,
+ "loss": 4.434604167938232,
+ "step": 1492
+ },
+ {
+ "epoch": 20.74061135371179,
+ "grad_norm": 0.015315295197069645,
+ "learning_rate": 0.0006,
+ "loss": 4.588066101074219,
+ "step": 1493
+ },
+ {
+ "epoch": 20.754585152838427,
+ "grad_norm": 0.018678732216358185,
+ "learning_rate": 0.0006,
+ "loss": 4.3514204025268555,
+ "step": 1494
+ },
+ {
+ "epoch": 20.768558951965066,
+ "grad_norm": 0.019521350041031837,
+ "learning_rate": 0.0006,
+ "loss": 4.437658786773682,
+ "step": 1495
+ },
+ {
+ "epoch": 20.782532751091704,
+ "grad_norm": 0.018363403156399727,
+ "learning_rate": 0.0006,
+ "loss": 4.495002269744873,
+ "step": 1496
+ },
+ {
+ "epoch": 20.79650655021834,
+ "grad_norm": 0.016830725595355034,
+ "learning_rate": 0.0006,
+ "loss": 4.402471542358398,
+ "step": 1497
+ },
+ {
+ "epoch": 20.810480349344978,
+ "grad_norm": 0.01722072623670101,
+ "learning_rate": 0.0006,
+ "loss": 4.437653541564941,
+ "step": 1498
+ },
+ {
+ "epoch": 20.824454148471617,
+ "grad_norm": 0.017459599301218987,
+ "learning_rate": 0.0006,
+ "loss": 4.366603851318359,
+ "step": 1499
+ },
+ {
+ "epoch": 20.83842794759825,
+ "grad_norm": 0.018902219831943512,
+ "learning_rate": 0.0006,
+ "loss": 4.501561164855957,
+ "step": 1500
+ },
+ {
+ "epoch": 20.85240174672489,
+ "grad_norm": 0.020814096555113792,
+ "learning_rate": 0.0006,
+ "loss": 4.535295486450195,
+ "step": 1501
+ },
+ {
+ "epoch": 20.86637554585153,
+ "grad_norm": 0.018990976735949516,
+ "learning_rate": 0.0006,
+ "loss": 4.333580017089844,
+ "step": 1502
+ },
+ {
+ "epoch": 20.880349344978168,
+ "grad_norm": 0.01837880350649357,
+ "learning_rate": 0.0006,
+ "loss": 4.313388824462891,
+ "step": 1503
+ },
+ {
+ "epoch": 20.894323144104803,
+ "grad_norm": 0.019820360466837883,
+ "learning_rate": 0.0006,
+ "loss": 4.388157844543457,
+ "step": 1504
+ },
+ {
+ "epoch": 20.90829694323144,
+ "grad_norm": 0.020181143656373024,
+ "learning_rate": 0.0006,
+ "loss": 4.572224140167236,
+ "step": 1505
+ },
+ {
+ "epoch": 20.92227074235808,
+ "grad_norm": 0.021206077188253403,
+ "learning_rate": 0.0006,
+ "loss": 4.3127970695495605,
+ "step": 1506
+ },
+ {
+ "epoch": 20.936244541484715,
+ "grad_norm": 0.017962060868740082,
+ "learning_rate": 0.0006,
+ "loss": 4.384734630584717,
+ "step": 1507
+ },
+ {
+ "epoch": 20.950218340611354,
+ "grad_norm": 0.01737220585346222,
+ "learning_rate": 0.0006,
+ "loss": 4.4075751304626465,
+ "step": 1508
+ },
+ {
+ "epoch": 20.964192139737992,
+ "grad_norm": 0.01839268207550049,
+ "learning_rate": 0.0006,
+ "loss": 4.553779125213623,
+ "step": 1509
+ },
+ {
+ "epoch": 20.978165938864628,
+ "grad_norm": 0.020102640613913536,
+ "learning_rate": 0.0006,
+ "loss": 4.399398326873779,
+ "step": 1510
+ },
+ {
+ "epoch": 20.992139737991266,
+ "grad_norm": 0.018831463530659676,
+ "learning_rate": 0.0006,
+ "loss": 4.358499526977539,
+ "step": 1511
+ },
+ {
+ "epoch": 21.0,
+ "grad_norm": 0.018254734575748444,
+ "learning_rate": 0.0006,
+ "loss": 4.384119510650635,
+ "step": 1512
+ },
+ {
+ "epoch": 21.0,
+ "eval_loss": 4.7524237632751465,
+ "eval_runtime": 57.8812,
+ "eval_samples_per_second": 42.19,
+ "eval_steps_per_second": 1.33,
+ "step": 1512
+ },
+ {
+ "epoch": 21.01397379912664,
+ "grad_norm": 0.019728587940335274,
+ "learning_rate": 0.0006,
+ "loss": 4.32363224029541,
+ "step": 1513
+ },
+ {
+ "epoch": 21.027947598253274,
+ "grad_norm": 0.02216625027358532,
+ "learning_rate": 0.0006,
+ "loss": 4.436988353729248,
+ "step": 1514
+ },
+ {
+ "epoch": 21.041921397379912,
+ "grad_norm": 0.02227606810629368,
+ "learning_rate": 0.0006,
+ "loss": 4.406063556671143,
+ "step": 1515
+ },
+ {
+ "epoch": 21.05589519650655,
+ "grad_norm": 0.020187009125947952,
+ "learning_rate": 0.0006,
+ "loss": 4.324402332305908,
+ "step": 1516
+ },
+ {
+ "epoch": 21.069868995633186,
+ "grad_norm": 0.02095157466828823,
+ "learning_rate": 0.0006,
+ "loss": 4.456544876098633,
+ "step": 1517
+ },
+ {
+ "epoch": 21.083842794759825,
+ "grad_norm": 0.02146013453602791,
+ "learning_rate": 0.0006,
+ "loss": 4.368387222290039,
+ "step": 1518
+ },
+ {
+ "epoch": 21.097816593886463,
+ "grad_norm": 0.02162035182118416,
+ "learning_rate": 0.0006,
+ "loss": 4.48838996887207,
+ "step": 1519
+ },
+ {
+ "epoch": 21.111790393013102,
+ "grad_norm": 0.021745676174759865,
+ "learning_rate": 0.0006,
+ "loss": 4.30778694152832,
+ "step": 1520
+ },
+ {
+ "epoch": 21.125764192139737,
+ "grad_norm": 0.02173588052392006,
+ "learning_rate": 0.0006,
+ "loss": 4.433149814605713,
+ "step": 1521
+ },
+ {
+ "epoch": 21.139737991266376,
+ "grad_norm": 0.020950648933649063,
+ "learning_rate": 0.0006,
+ "loss": 4.520089149475098,
+ "step": 1522
+ },
+ {
+ "epoch": 21.153711790393015,
+ "grad_norm": 0.020684752613306046,
+ "learning_rate": 0.0006,
+ "loss": 4.372437477111816,
+ "step": 1523
+ },
+ {
+ "epoch": 21.16768558951965,
+ "grad_norm": 0.019472869113087654,
+ "learning_rate": 0.0006,
+ "loss": 4.437015533447266,
+ "step": 1524
+ },
+ {
+ "epoch": 21.18165938864629,
+ "grad_norm": 0.01846962980926037,
+ "learning_rate": 0.0006,
+ "loss": 4.449531555175781,
+ "step": 1525
+ },
+ {
+ "epoch": 21.195633187772927,
+ "grad_norm": 0.017377803102135658,
+ "learning_rate": 0.0006,
+ "loss": 4.458279132843018,
+ "step": 1526
+ },
+ {
+ "epoch": 21.209606986899562,
+ "grad_norm": 0.016002200543880463,
+ "learning_rate": 0.0006,
+ "loss": 4.266297817230225,
+ "step": 1527
+ },
+ {
+ "epoch": 21.2235807860262,
+ "grad_norm": 0.015575903467833996,
+ "learning_rate": 0.0006,
+ "loss": 4.342632293701172,
+ "step": 1528
+ },
+ {
+ "epoch": 21.23755458515284,
+ "grad_norm": 0.015634698793292046,
+ "learning_rate": 0.0006,
+ "loss": 4.338623523712158,
+ "step": 1529
+ },
+ {
+ "epoch": 21.251528384279474,
+ "grad_norm": 0.01449581328779459,
+ "learning_rate": 0.0006,
+ "loss": 4.354330062866211,
+ "step": 1530
+ },
+ {
+ "epoch": 21.265502183406113,
+ "grad_norm": 0.01487616915255785,
+ "learning_rate": 0.0006,
+ "loss": 4.486835479736328,
+ "step": 1531
+ },
+ {
+ "epoch": 21.27947598253275,
+ "grad_norm": 0.016346555203199387,
+ "learning_rate": 0.0006,
+ "loss": 4.407098770141602,
+ "step": 1532
+ },
+ {
+ "epoch": 21.29344978165939,
+ "grad_norm": 0.01682182401418686,
+ "learning_rate": 0.0006,
+ "loss": 4.436809539794922,
+ "step": 1533
+ },
+ {
+ "epoch": 21.307423580786025,
+ "grad_norm": 0.017704004421830177,
+ "learning_rate": 0.0006,
+ "loss": 4.3854780197143555,
+ "step": 1534
+ },
+ {
+ "epoch": 21.321397379912664,
+ "grad_norm": 0.018437420949339867,
+ "learning_rate": 0.0006,
+ "loss": 4.413283824920654,
+ "step": 1535
+ },
+ {
+ "epoch": 21.335371179039303,
+ "grad_norm": 0.016911081969738007,
+ "learning_rate": 0.0006,
+ "loss": 4.357306003570557,
+ "step": 1536
+ },
+ {
+ "epoch": 21.349344978165938,
+ "grad_norm": 0.01642223633825779,
+ "learning_rate": 0.0006,
+ "loss": 4.422322750091553,
+ "step": 1537
+ },
+ {
+ "epoch": 21.363318777292577,
+ "grad_norm": 0.01780877821147442,
+ "learning_rate": 0.0006,
+ "loss": 4.387620449066162,
+ "step": 1538
+ },
+ {
+ "epoch": 21.377292576419215,
+ "grad_norm": 0.01653975248336792,
+ "learning_rate": 0.0006,
+ "loss": 4.327267646789551,
+ "step": 1539
+ },
+ {
+ "epoch": 21.39126637554585,
+ "grad_norm": 0.01572156324982643,
+ "learning_rate": 0.0006,
+ "loss": 4.475979804992676,
+ "step": 1540
+ },
+ {
+ "epoch": 21.40524017467249,
+ "grad_norm": 0.017156066372990608,
+ "learning_rate": 0.0006,
+ "loss": 4.4011125564575195,
+ "step": 1541
+ },
+ {
+ "epoch": 21.419213973799128,
+ "grad_norm": 0.01788829080760479,
+ "learning_rate": 0.0006,
+ "loss": 4.4174089431762695,
+ "step": 1542
+ },
+ {
+ "epoch": 21.433187772925763,
+ "grad_norm": 0.01728985086083412,
+ "learning_rate": 0.0006,
+ "loss": 4.429244041442871,
+ "step": 1543
+ },
+ {
+ "epoch": 21.4471615720524,
+ "grad_norm": 0.016534466296434402,
+ "learning_rate": 0.0006,
+ "loss": 4.326068878173828,
+ "step": 1544
+ },
+ {
+ "epoch": 21.46113537117904,
+ "grad_norm": 0.016582584008574486,
+ "learning_rate": 0.0006,
+ "loss": 4.2975969314575195,
+ "step": 1545
+ },
+ {
+ "epoch": 21.475109170305675,
+ "grad_norm": 0.018034178763628006,
+ "learning_rate": 0.0006,
+ "loss": 4.501319408416748,
+ "step": 1546
+ },
+ {
+ "epoch": 21.489082969432314,
+ "grad_norm": 0.017827702686190605,
+ "learning_rate": 0.0006,
+ "loss": 4.502284526824951,
+ "step": 1547
+ },
+ {
+ "epoch": 21.503056768558952,
+ "grad_norm": 0.017182299867272377,
+ "learning_rate": 0.0006,
+ "loss": 4.380828857421875,
+ "step": 1548
+ },
+ {
+ "epoch": 21.51703056768559,
+ "grad_norm": 0.019101588055491447,
+ "learning_rate": 0.0006,
+ "loss": 4.31564998626709,
+ "step": 1549
+ },
+ {
+ "epoch": 21.531004366812226,
+ "grad_norm": 0.0204677302390337,
+ "learning_rate": 0.0006,
+ "loss": 4.3840131759643555,
+ "step": 1550
+ },
+ {
+ "epoch": 21.544978165938865,
+ "grad_norm": 0.0188993439078331,
+ "learning_rate": 0.0006,
+ "loss": 4.4257731437683105,
+ "step": 1551
+ },
+ {
+ "epoch": 21.558951965065503,
+ "grad_norm": 0.01730230636894703,
+ "learning_rate": 0.0006,
+ "loss": 4.445132255554199,
+ "step": 1552
+ },
+ {
+ "epoch": 21.57292576419214,
+ "grad_norm": 0.01801574043929577,
+ "learning_rate": 0.0006,
+ "loss": 4.427064895629883,
+ "step": 1553
+ },
+ {
+ "epoch": 21.586899563318777,
+ "grad_norm": 0.018417565152049065,
+ "learning_rate": 0.0006,
+ "loss": 4.319595813751221,
+ "step": 1554
+ },
+ {
+ "epoch": 21.600873362445416,
+ "grad_norm": 0.01683160290122032,
+ "learning_rate": 0.0006,
+ "loss": 4.349218368530273,
+ "step": 1555
+ },
+ {
+ "epoch": 21.61484716157205,
+ "grad_norm": 0.016973106190562248,
+ "learning_rate": 0.0006,
+ "loss": 4.375990390777588,
+ "step": 1556
+ },
+ {
+ "epoch": 21.62882096069869,
+ "grad_norm": 0.015688583254814148,
+ "learning_rate": 0.0006,
+ "loss": 4.408170223236084,
+ "step": 1557
+ },
+ {
+ "epoch": 21.64279475982533,
+ "grad_norm": 0.01573043316602707,
+ "learning_rate": 0.0006,
+ "loss": 4.416255950927734,
+ "step": 1558
+ },
+ {
+ "epoch": 21.656768558951963,
+ "grad_norm": 0.01744293048977852,
+ "learning_rate": 0.0006,
+ "loss": 4.477141380310059,
+ "step": 1559
+ },
+ {
+ "epoch": 21.670742358078602,
+ "grad_norm": 0.017503513023257256,
+ "learning_rate": 0.0006,
+ "loss": 4.3367486000061035,
+ "step": 1560
+ },
+ {
+ "epoch": 21.68471615720524,
+ "grad_norm": 0.016592150554060936,
+ "learning_rate": 0.0006,
+ "loss": 4.314785480499268,
+ "step": 1561
+ },
+ {
+ "epoch": 21.69868995633188,
+ "grad_norm": 0.015774084255099297,
+ "learning_rate": 0.0006,
+ "loss": 4.448209285736084,
+ "step": 1562
+ },
+ {
+ "epoch": 21.712663755458514,
+ "grad_norm": 0.015779150649905205,
+ "learning_rate": 0.0006,
+ "loss": 4.392690658569336,
+ "step": 1563
+ },
+ {
+ "epoch": 21.726637554585153,
+ "grad_norm": 0.015392648056149483,
+ "learning_rate": 0.0006,
+ "loss": 4.437741756439209,
+ "step": 1564
+ },
+ {
+ "epoch": 21.74061135371179,
+ "grad_norm": 0.01530836895108223,
+ "learning_rate": 0.0006,
+ "loss": 4.424595832824707,
+ "step": 1565
+ },
+ {
+ "epoch": 21.754585152838427,
+ "grad_norm": 0.015600779093801975,
+ "learning_rate": 0.0006,
+ "loss": 4.335417747497559,
+ "step": 1566
+ },
+ {
+ "epoch": 21.768558951965066,
+ "grad_norm": 0.014646518044173717,
+ "learning_rate": 0.0006,
+ "loss": 4.279081344604492,
+ "step": 1567
+ },
+ {
+ "epoch": 21.782532751091704,
+ "grad_norm": 0.014132864773273468,
+ "learning_rate": 0.0006,
+ "loss": 4.345950126647949,
+ "step": 1568
+ },
+ {
+ "epoch": 21.79650655021834,
+ "grad_norm": 0.01425724383443594,
+ "learning_rate": 0.0006,
+ "loss": 4.277588844299316,
+ "step": 1569
+ },
+ {
+ "epoch": 21.810480349344978,
+ "grad_norm": 0.015396572649478912,
+ "learning_rate": 0.0006,
+ "loss": 4.326196670532227,
+ "step": 1570
+ },
+ {
+ "epoch": 21.824454148471617,
+ "grad_norm": 0.01683669537305832,
+ "learning_rate": 0.0006,
+ "loss": 4.3294172286987305,
+ "step": 1571
+ },
+ {
+ "epoch": 21.83842794759825,
+ "grad_norm": 0.016365647315979004,
+ "learning_rate": 0.0006,
+ "loss": 4.523321628570557,
+ "step": 1572
+ },
+ {
+ "epoch": 21.85240174672489,
+ "grad_norm": 0.017142774537205696,
+ "learning_rate": 0.0006,
+ "loss": 4.475842475891113,
+ "step": 1573
+ },
+ {
+ "epoch": 21.86637554585153,
+ "grad_norm": 0.015950839966535568,
+ "learning_rate": 0.0006,
+ "loss": 4.397187232971191,
+ "step": 1574
+ },
+ {
+ "epoch": 21.880349344978168,
+ "grad_norm": 0.016180362552404404,
+ "learning_rate": 0.0006,
+ "loss": 4.414778709411621,
+ "step": 1575
+ },
+ {
+ "epoch": 21.894323144104803,
+ "grad_norm": 0.017615556716918945,
+ "learning_rate": 0.0006,
+ "loss": 4.348567962646484,
+ "step": 1576
+ },
+ {
+ "epoch": 21.90829694323144,
+ "grad_norm": 0.01896277628839016,
+ "learning_rate": 0.0006,
+ "loss": 4.431762218475342,
+ "step": 1577
+ },
+ {
+ "epoch": 21.92227074235808,
+ "grad_norm": 0.01990971714258194,
+ "learning_rate": 0.0006,
+ "loss": 4.431285858154297,
+ "step": 1578
+ },
+ {
+ "epoch": 21.936244541484715,
+ "grad_norm": 0.0201258547604084,
+ "learning_rate": 0.0006,
+ "loss": 4.309872627258301,
+ "step": 1579
+ },
+ {
+ "epoch": 21.950218340611354,
+ "grad_norm": 0.019620198756456375,
+ "learning_rate": 0.0006,
+ "loss": 4.484044551849365,
+ "step": 1580
+ },
+ {
+ "epoch": 21.964192139737992,
+ "grad_norm": 0.01771123707294464,
+ "learning_rate": 0.0006,
+ "loss": 4.368093013763428,
+ "step": 1581
+ },
+ {
+ "epoch": 21.978165938864628,
+ "grad_norm": 0.017501654103398323,
+ "learning_rate": 0.0006,
+ "loss": 4.372651100158691,
+ "step": 1582
+ },
+ {
+ "epoch": 21.992139737991266,
+ "grad_norm": 0.01693608984351158,
+ "learning_rate": 0.0006,
+ "loss": 4.379167556762695,
+ "step": 1583
+ },
+ {
+ "epoch": 22.0,
+ "grad_norm": 0.02002849616110325,
+ "learning_rate": 0.0006,
+ "loss": 4.329627990722656,
+ "step": 1584
+ },
+ {
+ "epoch": 22.0,
+ "eval_loss": 4.723696231842041,
+ "eval_runtime": 56.1758,
+ "eval_samples_per_second": 43.471,
+ "eval_steps_per_second": 1.371,
+ "step": 1584
+ },
+ {
+ "epoch": 22.01397379912664,
+ "grad_norm": 0.02030305378139019,
+ "learning_rate": 0.0006,
+ "loss": 4.355566024780273,
+ "step": 1585
+ },
+ {
+ "epoch": 22.027947598253274,
+ "grad_norm": 0.018638934940099716,
+ "learning_rate": 0.0006,
+ "loss": 4.314009189605713,
+ "step": 1586
+ },
+ {
+ "epoch": 22.041921397379912,
+ "grad_norm": 0.019917072728276253,
+ "learning_rate": 0.0006,
+ "loss": 4.367525577545166,
+ "step": 1587
+ },
+ {
+ "epoch": 22.05589519650655,
+ "grad_norm": 0.020785167813301086,
+ "learning_rate": 0.0006,
+ "loss": 4.375561714172363,
+ "step": 1588
+ },
+ {
+ "epoch": 22.069868995633186,
+ "grad_norm": 0.021214453503489494,
+ "learning_rate": 0.0006,
+ "loss": 4.418305397033691,
+ "step": 1589
+ },
+ {
+ "epoch": 22.083842794759825,
+ "grad_norm": 0.019051678478717804,
+ "learning_rate": 0.0006,
+ "loss": 4.215126991271973,
+ "step": 1590
+ },
+ {
+ "epoch": 22.097816593886463,
+ "grad_norm": 0.02323366142809391,
+ "learning_rate": 0.0006,
+ "loss": 4.306807994842529,
+ "step": 1591
+ },
+ {
+ "epoch": 22.111790393013102,
+ "grad_norm": 0.026186738163232803,
+ "learning_rate": 0.0006,
+ "loss": 4.408508777618408,
+ "step": 1592
+ },
+ {
+ "epoch": 22.125764192139737,
+ "grad_norm": 0.027361227199435234,
+ "learning_rate": 0.0006,
+ "loss": 4.329322814941406,
+ "step": 1593
+ },
+ {
+ "epoch": 22.139737991266376,
+ "grad_norm": 0.026564646512269974,
+ "learning_rate": 0.0006,
+ "loss": 4.346642017364502,
+ "step": 1594
+ },
+ {
+ "epoch": 22.153711790393015,
+ "grad_norm": 0.025073667988181114,
+ "learning_rate": 0.0006,
+ "loss": 4.3984198570251465,
+ "step": 1595
+ },
+ {
+ "epoch": 22.16768558951965,
+ "grad_norm": 0.02347794733941555,
+ "learning_rate": 0.0006,
+ "loss": 4.352538108825684,
+ "step": 1596
+ },
+ {
+ "epoch": 22.18165938864629,
+ "grad_norm": 0.02203468047082424,
+ "learning_rate": 0.0006,
+ "loss": 4.298148155212402,
+ "step": 1597
+ },
+ {
+ "epoch": 22.195633187772927,
+ "grad_norm": 0.020621957257390022,
+ "learning_rate": 0.0006,
+ "loss": 4.480309963226318,
+ "step": 1598
+ },
+ {
+ "epoch": 22.209606986899562,
+ "grad_norm": 0.023710055276751518,
+ "learning_rate": 0.0006,
+ "loss": 4.339317321777344,
+ "step": 1599
+ },
+ {
+ "epoch": 22.2235807860262,
+ "grad_norm": 0.0233482476323843,
+ "learning_rate": 0.0006,
+ "loss": 4.436896324157715,
+ "step": 1600
+ },
+ {
+ "epoch": 22.23755458515284,
+ "grad_norm": 0.020122403278946877,
+ "learning_rate": 0.0006,
+ "loss": 4.236002445220947,
+ "step": 1601
+ },
+ {
+ "epoch": 22.251528384279474,
+ "grad_norm": 0.01957266591489315,
+ "learning_rate": 0.0006,
+ "loss": 4.497328281402588,
+ "step": 1602
+ },
+ {
+ "epoch": 22.265502183406113,
+ "grad_norm": 0.018964895978569984,
+ "learning_rate": 0.0006,
+ "loss": 4.36360502243042,
+ "step": 1603
+ },
+ {
+ "epoch": 22.27947598253275,
+ "grad_norm": 0.018902769312262535,
+ "learning_rate": 0.0006,
+ "loss": 4.427093505859375,
+ "step": 1604
+ },
+ {
+ "epoch": 22.29344978165939,
+ "grad_norm": 0.018918119370937347,
+ "learning_rate": 0.0006,
+ "loss": 4.294950485229492,
+ "step": 1605
+ },
+ {
+ "epoch": 22.307423580786025,
+ "grad_norm": 0.019315775483846664,
+ "learning_rate": 0.0006,
+ "loss": 4.427952289581299,
+ "step": 1606
+ },
+ {
+ "epoch": 22.321397379912664,
+ "grad_norm": 0.018350522965192795,
+ "learning_rate": 0.0006,
+ "loss": 4.272945404052734,
+ "step": 1607
+ },
+ {
+ "epoch": 22.335371179039303,
+ "grad_norm": 0.01759534701704979,
+ "learning_rate": 0.0006,
+ "loss": 4.348023891448975,
+ "step": 1608
+ },
+ {
+ "epoch": 22.349344978165938,
+ "grad_norm": 0.01646057888865471,
+ "learning_rate": 0.0006,
+ "loss": 4.278518199920654,
+ "step": 1609
+ },
+ {
+ "epoch": 22.363318777292577,
+ "grad_norm": 0.015875060111284256,
+ "learning_rate": 0.0006,
+ "loss": 4.423527240753174,
+ "step": 1610
+ },
+ {
+ "epoch": 22.377292576419215,
+ "grad_norm": 0.01638529635965824,
+ "learning_rate": 0.0006,
+ "loss": 4.321723937988281,
+ "step": 1611
+ },
+ {
+ "epoch": 22.39126637554585,
+ "grad_norm": 0.016942061483860016,
+ "learning_rate": 0.0006,
+ "loss": 4.346164703369141,
+ "step": 1612
+ },
+ {
+ "epoch": 22.40524017467249,
+ "grad_norm": 0.015209621749818325,
+ "learning_rate": 0.0006,
+ "loss": 4.434340000152588,
+ "step": 1613
+ },
+ {
+ "epoch": 22.419213973799128,
+ "grad_norm": 0.014423009008169174,
+ "learning_rate": 0.0006,
+ "loss": 4.342560768127441,
+ "step": 1614
+ },
+ {
+ "epoch": 22.433187772925763,
+ "grad_norm": 0.014369525946676731,
+ "learning_rate": 0.0006,
+ "loss": 4.296586990356445,
+ "step": 1615
+ },
+ {
+ "epoch": 22.4471615720524,
+ "grad_norm": 0.01392540242522955,
+ "learning_rate": 0.0006,
+ "loss": 4.433176040649414,
+ "step": 1616
+ },
+ {
+ "epoch": 22.46113537117904,
+ "grad_norm": 0.013206899166107178,
+ "learning_rate": 0.0006,
+ "loss": 4.32962703704834,
+ "step": 1617
+ },
+ {
+ "epoch": 22.475109170305675,
+ "grad_norm": 0.013427077792584896,
+ "learning_rate": 0.0006,
+ "loss": 4.420437335968018,
+ "step": 1618
+ },
+ {
+ "epoch": 22.489082969432314,
+ "grad_norm": 0.014194196090102196,
+ "learning_rate": 0.0006,
+ "loss": 4.410478591918945,
+ "step": 1619
+ },
+ {
+ "epoch": 22.503056768558952,
+ "grad_norm": 0.01362364087253809,
+ "learning_rate": 0.0006,
+ "loss": 4.338932037353516,
+ "step": 1620
+ },
+ {
+ "epoch": 22.51703056768559,
+ "grad_norm": 0.014299589209258556,
+ "learning_rate": 0.0006,
+ "loss": 4.34089994430542,
+ "step": 1621
+ },
+ {
+ "epoch": 22.531004366812226,
+ "grad_norm": 0.014061295427381992,
+ "learning_rate": 0.0006,
+ "loss": 4.29201602935791,
+ "step": 1622
+ },
+ {
+ "epoch": 22.544978165938865,
+ "grad_norm": 0.013945229351520538,
+ "learning_rate": 0.0006,
+ "loss": 4.326052188873291,
+ "step": 1623
+ },
+ {
+ "epoch": 22.558951965065503,
+ "grad_norm": 0.01387068536132574,
+ "learning_rate": 0.0006,
+ "loss": 4.393521308898926,
+ "step": 1624
+ },
+ {
+ "epoch": 22.57292576419214,
+ "grad_norm": 0.013931898400187492,
+ "learning_rate": 0.0006,
+ "loss": 4.3739752769470215,
+ "step": 1625
+ },
+ {
+ "epoch": 22.586899563318777,
+ "grad_norm": 0.01497740764170885,
+ "learning_rate": 0.0006,
+ "loss": 4.3991618156433105,
+ "step": 1626
+ },
+ {
+ "epoch": 22.600873362445416,
+ "grad_norm": 0.015842905268073082,
+ "learning_rate": 0.0006,
+ "loss": 4.405340671539307,
+ "step": 1627
+ },
+ {
+ "epoch": 22.61484716157205,
+ "grad_norm": 0.01580633595585823,
+ "learning_rate": 0.0006,
+ "loss": 4.280102729797363,
+ "step": 1628
+ },
+ {
+ "epoch": 22.62882096069869,
+ "grad_norm": 0.016567382961511612,
+ "learning_rate": 0.0006,
+ "loss": 4.407657146453857,
+ "step": 1629
+ },
+ {
+ "epoch": 22.64279475982533,
+ "grad_norm": 0.01648622378706932,
+ "learning_rate": 0.0006,
+ "loss": 4.392590522766113,
+ "step": 1630
+ },
+ {
+ "epoch": 22.656768558951963,
+ "grad_norm": 0.017066307365894318,
+ "learning_rate": 0.0006,
+ "loss": 4.324862003326416,
+ "step": 1631
+ },
+ {
+ "epoch": 22.670742358078602,
+ "grad_norm": 0.01637062057852745,
+ "learning_rate": 0.0006,
+ "loss": 4.34598445892334,
+ "step": 1632
+ },
+ {
+ "epoch": 22.68471615720524,
+ "grad_norm": 0.018038960173726082,
+ "learning_rate": 0.0006,
+ "loss": 4.320711612701416,
+ "step": 1633
+ },
+ {
+ "epoch": 22.69868995633188,
+ "grad_norm": 0.018096428364515305,
+ "learning_rate": 0.0006,
+ "loss": 4.536527156829834,
+ "step": 1634
+ },
+ {
+ "epoch": 22.712663755458514,
+ "grad_norm": 0.017889291048049927,
+ "learning_rate": 0.0006,
+ "loss": 4.388131141662598,
+ "step": 1635
+ },
+ {
+ "epoch": 22.726637554585153,
+ "grad_norm": 0.018771521747112274,
+ "learning_rate": 0.0006,
+ "loss": 4.357668876647949,
+ "step": 1636
+ },
+ {
+ "epoch": 22.74061135371179,
+ "grad_norm": 0.02100873365998268,
+ "learning_rate": 0.0006,
+ "loss": 4.293972015380859,
+ "step": 1637
+ },
+ {
+ "epoch": 22.754585152838427,
+ "grad_norm": 0.017715785652399063,
+ "learning_rate": 0.0006,
+ "loss": 4.349849224090576,
+ "step": 1638
+ },
+ {
+ "epoch": 22.768558951965066,
+ "grad_norm": 0.015546333976089954,
+ "learning_rate": 0.0006,
+ "loss": 4.311728000640869,
+ "step": 1639
+ },
+ {
+ "epoch": 22.782532751091704,
+ "grad_norm": 0.01605990156531334,
+ "learning_rate": 0.0006,
+ "loss": 4.499123573303223,
+ "step": 1640
+ },
+ {
+ "epoch": 22.79650655021834,
+ "grad_norm": 0.016642142087221146,
+ "learning_rate": 0.0006,
+ "loss": 4.4319257736206055,
+ "step": 1641
+ },
+ {
+ "epoch": 22.810480349344978,
+ "grad_norm": 0.016120387241244316,
+ "learning_rate": 0.0006,
+ "loss": 4.4373650550842285,
+ "step": 1642
+ },
+ {
+ "epoch": 22.824454148471617,
+ "grad_norm": 0.015174890868365765,
+ "learning_rate": 0.0006,
+ "loss": 4.314082145690918,
+ "step": 1643
+ },
+ {
+ "epoch": 22.83842794759825,
+ "grad_norm": 0.01551714539527893,
+ "learning_rate": 0.0006,
+ "loss": 4.360363960266113,
+ "step": 1644
+ },
+ {
+ "epoch": 22.85240174672489,
+ "grad_norm": 0.014966742135584354,
+ "learning_rate": 0.0006,
+ "loss": 4.346253395080566,
+ "step": 1645
+ },
+ {
+ "epoch": 22.86637554585153,
+ "grad_norm": 0.014288028702139854,
+ "learning_rate": 0.0006,
+ "loss": 4.464366436004639,
+ "step": 1646
+ },
+ {
+ "epoch": 22.880349344978168,
+ "grad_norm": 0.015291682444512844,
+ "learning_rate": 0.0006,
+ "loss": 4.504101753234863,
+ "step": 1647
+ },
+ {
+ "epoch": 22.894323144104803,
+ "grad_norm": 0.014534426853060722,
+ "learning_rate": 0.0006,
+ "loss": 4.462172031402588,
+ "step": 1648
+ },
+ {
+ "epoch": 22.90829694323144,
+ "grad_norm": 0.014129819348454475,
+ "learning_rate": 0.0006,
+ "loss": 4.350126266479492,
+ "step": 1649
+ },
+ {
+ "epoch": 22.92227074235808,
+ "grad_norm": 0.014932668767869473,
+ "learning_rate": 0.0006,
+ "loss": 4.374392032623291,
+ "step": 1650
+ },
+ {
+ "epoch": 22.936244541484715,
+ "grad_norm": 0.014647853560745716,
+ "learning_rate": 0.0006,
+ "loss": 4.252264499664307,
+ "step": 1651
+ },
+ {
+ "epoch": 22.950218340611354,
+ "grad_norm": 0.013553360477089882,
+ "learning_rate": 0.0006,
+ "loss": 4.379729270935059,
+ "step": 1652
+ },
+ {
+ "epoch": 22.964192139737992,
+ "grad_norm": 0.014606194570660591,
+ "learning_rate": 0.0006,
+ "loss": 4.363372802734375,
+ "step": 1653
+ },
+ {
+ "epoch": 22.978165938864628,
+ "grad_norm": 0.015000631101429462,
+ "learning_rate": 0.0006,
+ "loss": 4.408048629760742,
+ "step": 1654
+ },
+ {
+ "epoch": 22.992139737991266,
+ "grad_norm": 0.015370816923677921,
+ "learning_rate": 0.0006,
+ "loss": 4.297451019287109,
+ "step": 1655
+ },
+ {
+ "epoch": 23.0,
+ "grad_norm": 0.01709144562482834,
+ "learning_rate": 0.0006,
+ "loss": 4.294776916503906,
+ "step": 1656
+ },
+ {
+ "epoch": 23.0,
+ "eval_loss": 4.678618907928467,
+ "eval_runtime": 56.1668,
+ "eval_samples_per_second": 43.478,
+ "eval_steps_per_second": 1.371,
+ "step": 1656
+ },
+ {
+ "epoch": 23.01397379912664,
+ "grad_norm": 0.01727290451526642,
+ "learning_rate": 0.0006,
+ "loss": 4.311701774597168,
+ "step": 1657
+ },
+ {
+ "epoch": 23.027947598253274,
+ "grad_norm": 0.020645759999752045,
+ "learning_rate": 0.0006,
+ "loss": 4.429897308349609,
+ "step": 1658
+ },
+ {
+ "epoch": 23.041921397379912,
+ "grad_norm": 0.02204226516187191,
+ "learning_rate": 0.0006,
+ "loss": 4.365151882171631,
+ "step": 1659
+ },
+ {
+ "epoch": 23.05589519650655,
+ "grad_norm": 0.023069269955158234,
+ "learning_rate": 0.0006,
+ "loss": 4.3710832595825195,
+ "step": 1660
+ },
+ {
+ "epoch": 23.069868995633186,
+ "grad_norm": 0.024151179939508438,
+ "learning_rate": 0.0006,
+ "loss": 4.403108596801758,
+ "step": 1661
+ },
+ {
+ "epoch": 23.083842794759825,
+ "grad_norm": 0.02511233650147915,
+ "learning_rate": 0.0006,
+ "loss": 4.414128303527832,
+ "step": 1662
+ },
+ {
+ "epoch": 23.097816593886463,
+ "grad_norm": 0.025802113115787506,
+ "learning_rate": 0.0006,
+ "loss": 4.388265609741211,
+ "step": 1663
+ },
+ {
+ "epoch": 23.111790393013102,
+ "grad_norm": 0.02377348765730858,
+ "learning_rate": 0.0006,
+ "loss": 4.357505798339844,
+ "step": 1664
+ },
+ {
+ "epoch": 23.125764192139737,
+ "grad_norm": 0.020876651629805565,
+ "learning_rate": 0.0006,
+ "loss": 4.23043966293335,
+ "step": 1665
+ },
+ {
+ "epoch": 23.139737991266376,
+ "grad_norm": 0.01959727331995964,
+ "learning_rate": 0.0006,
+ "loss": 4.280970096588135,
+ "step": 1666
+ },
+ {
+ "epoch": 23.153711790393015,
+ "grad_norm": 0.018678616732358932,
+ "learning_rate": 0.0006,
+ "loss": 4.346716403961182,
+ "step": 1667
+ },
+ {
+ "epoch": 23.16768558951965,
+ "grad_norm": 0.02108944021165371,
+ "learning_rate": 0.0006,
+ "loss": 4.277904510498047,
+ "step": 1668
+ },
+ {
+ "epoch": 23.18165938864629,
+ "grad_norm": 0.020221566781401634,
+ "learning_rate": 0.0006,
+ "loss": 4.367693901062012,
+ "step": 1669
+ },
+ {
+ "epoch": 23.195633187772927,
+ "grad_norm": 0.019599711522459984,
+ "learning_rate": 0.0006,
+ "loss": 4.257092475891113,
+ "step": 1670
+ },
+ {
+ "epoch": 23.209606986899562,
+ "grad_norm": 0.019833361729979515,
+ "learning_rate": 0.0006,
+ "loss": 4.540737628936768,
+ "step": 1671
+ },
+ {
+ "epoch": 23.2235807860262,
+ "grad_norm": 0.020352095365524292,
+ "learning_rate": 0.0006,
+ "loss": 4.360923767089844,
+ "step": 1672
+ },
+ {
+ "epoch": 23.23755458515284,
+ "grad_norm": 0.01939748041331768,
+ "learning_rate": 0.0006,
+ "loss": 4.393110275268555,
+ "step": 1673
+ },
+ {
+ "epoch": 23.251528384279474,
+ "grad_norm": 0.01884354278445244,
+ "learning_rate": 0.0006,
+ "loss": 4.383938789367676,
+ "step": 1674
+ },
+ {
+ "epoch": 23.265502183406113,
+ "grad_norm": 0.02008470520377159,
+ "learning_rate": 0.0006,
+ "loss": 4.346344947814941,
+ "step": 1675
+ },
+ {
+ "epoch": 23.27947598253275,
+ "grad_norm": 0.019065195694565773,
+ "learning_rate": 0.0006,
+ "loss": 4.36611795425415,
+ "step": 1676
+ },
+ {
+ "epoch": 23.29344978165939,
+ "grad_norm": 0.01938541606068611,
+ "learning_rate": 0.0006,
+ "loss": 4.4009599685668945,
+ "step": 1677
+ },
+ {
+ "epoch": 23.307423580786025,
+ "grad_norm": 0.019044723361730576,
+ "learning_rate": 0.0006,
+ "loss": 4.340169906616211,
+ "step": 1678
+ },
+ {
+ "epoch": 23.321397379912664,
+ "grad_norm": 0.018850378692150116,
+ "learning_rate": 0.0006,
+ "loss": 4.37674617767334,
+ "step": 1679
+ },
+ {
+ "epoch": 23.335371179039303,
+ "grad_norm": 0.019076095893979073,
+ "learning_rate": 0.0006,
+ "loss": 4.226874351501465,
+ "step": 1680
+ },
+ {
+ "epoch": 23.349344978165938,
+ "grad_norm": 0.017543038353323936,
+ "learning_rate": 0.0006,
+ "loss": 4.337751388549805,
+ "step": 1681
+ },
+ {
+ "epoch": 23.363318777292577,
+ "grad_norm": 0.017170218750834465,
+ "learning_rate": 0.0006,
+ "loss": 4.282512664794922,
+ "step": 1682
+ },
+ {
+ "epoch": 23.377292576419215,
+ "grad_norm": 0.01574796624481678,
+ "learning_rate": 0.0006,
+ "loss": 4.30638313293457,
+ "step": 1683
+ },
+ {
+ "epoch": 23.39126637554585,
+ "grad_norm": 0.015160846523940563,
+ "learning_rate": 0.0006,
+ "loss": 4.336716175079346,
+ "step": 1684
+ },
+ {
+ "epoch": 23.40524017467249,
+ "grad_norm": 0.01534526702016592,
+ "learning_rate": 0.0006,
+ "loss": 4.2174787521362305,
+ "step": 1685
+ },
+ {
+ "epoch": 23.419213973799128,
+ "grad_norm": 0.015289999544620514,
+ "learning_rate": 0.0006,
+ "loss": 4.272546291351318,
+ "step": 1686
+ },
+ {
+ "epoch": 23.433187772925763,
+ "grad_norm": 0.014894185587763786,
+ "learning_rate": 0.0006,
+ "loss": 4.296045303344727,
+ "step": 1687
+ },
+ {
+ "epoch": 23.4471615720524,
+ "grad_norm": 0.014869502745568752,
+ "learning_rate": 0.0006,
+ "loss": 4.425461769104004,
+ "step": 1688
+ },
+ {
+ "epoch": 23.46113537117904,
+ "grad_norm": 0.01474483497440815,
+ "learning_rate": 0.0006,
+ "loss": 4.361867427825928,
+ "step": 1689
+ },
+ {
+ "epoch": 23.475109170305675,
+ "grad_norm": 0.015299245715141296,
+ "learning_rate": 0.0006,
+ "loss": 4.364523410797119,
+ "step": 1690
+ },
+ {
+ "epoch": 23.489082969432314,
+ "grad_norm": 0.015201340429484844,
+ "learning_rate": 0.0006,
+ "loss": 4.285182952880859,
+ "step": 1691
+ },
+ {
+ "epoch": 23.503056768558952,
+ "grad_norm": 0.016057293862104416,
+ "learning_rate": 0.0006,
+ "loss": 4.284873962402344,
+ "step": 1692
+ },
+ {
+ "epoch": 23.51703056768559,
+ "grad_norm": 0.015871062874794006,
+ "learning_rate": 0.0006,
+ "loss": 4.326470851898193,
+ "step": 1693
+ },
+ {
+ "epoch": 23.531004366812226,
+ "grad_norm": 0.01586288772523403,
+ "learning_rate": 0.0006,
+ "loss": 4.324883460998535,
+ "step": 1694
+ },
+ {
+ "epoch": 23.544978165938865,
+ "grad_norm": 0.016092827543616295,
+ "learning_rate": 0.0006,
+ "loss": 4.3258867263793945,
+ "step": 1695
+ },
+ {
+ "epoch": 23.558951965065503,
+ "grad_norm": 0.01582016795873642,
+ "learning_rate": 0.0006,
+ "loss": 4.358892440795898,
+ "step": 1696
+ },
+ {
+ "epoch": 23.57292576419214,
+ "grad_norm": 0.01688452623784542,
+ "learning_rate": 0.0006,
+ "loss": 4.280842304229736,
+ "step": 1697
+ },
+ {
+ "epoch": 23.586899563318777,
+ "grad_norm": 0.015865886583924294,
+ "learning_rate": 0.0006,
+ "loss": 4.245567321777344,
+ "step": 1698
+ },
+ {
+ "epoch": 23.600873362445416,
+ "grad_norm": 0.015303662046790123,
+ "learning_rate": 0.0006,
+ "loss": 4.383112907409668,
+ "step": 1699
+ },
+ {
+ "epoch": 23.61484716157205,
+ "grad_norm": 0.015649516135454178,
+ "learning_rate": 0.0006,
+ "loss": 4.519082069396973,
+ "step": 1700
+ },
+ {
+ "epoch": 23.62882096069869,
+ "grad_norm": 0.0167120099067688,
+ "learning_rate": 0.0006,
+ "loss": 4.447847366333008,
+ "step": 1701
+ },
+ {
+ "epoch": 23.64279475982533,
+ "grad_norm": 0.016368716955184937,
+ "learning_rate": 0.0006,
+ "loss": 4.434186935424805,
+ "step": 1702
+ },
+ {
+ "epoch": 23.656768558951963,
+ "grad_norm": 0.017379503697156906,
+ "learning_rate": 0.0006,
+ "loss": 4.317594528198242,
+ "step": 1703
+ },
+ {
+ "epoch": 23.670742358078602,
+ "grad_norm": 0.018573811277747154,
+ "learning_rate": 0.0006,
+ "loss": 4.205793857574463,
+ "step": 1704
+ },
+ {
+ "epoch": 23.68471615720524,
+ "grad_norm": 0.01749090477824211,
+ "learning_rate": 0.0006,
+ "loss": 4.3985700607299805,
+ "step": 1705
+ },
+ {
+ "epoch": 23.69868995633188,
+ "grad_norm": 0.0165668074041605,
+ "learning_rate": 0.0006,
+ "loss": 4.329226970672607,
+ "step": 1706
+ },
+ {
+ "epoch": 23.712663755458514,
+ "grad_norm": 0.016913846135139465,
+ "learning_rate": 0.0006,
+ "loss": 4.388485908508301,
+ "step": 1707
+ },
+ {
+ "epoch": 23.726637554585153,
+ "grad_norm": 0.016813859343528748,
+ "learning_rate": 0.0006,
+ "loss": 4.3395915031433105,
+ "step": 1708
+ },
+ {
+ "epoch": 23.74061135371179,
+ "grad_norm": 0.01676975190639496,
+ "learning_rate": 0.0006,
+ "loss": 4.254745960235596,
+ "step": 1709
+ },
+ {
+ "epoch": 23.754585152838427,
+ "grad_norm": 0.016370350494980812,
+ "learning_rate": 0.0006,
+ "loss": 4.2612223625183105,
+ "step": 1710
+ },
+ {
+ "epoch": 23.768558951965066,
+ "grad_norm": 0.01696198247373104,
+ "learning_rate": 0.0006,
+ "loss": 4.327524662017822,
+ "step": 1711
+ },
+ {
+ "epoch": 23.782532751091704,
+ "grad_norm": 0.0189108457416296,
+ "learning_rate": 0.0006,
+ "loss": 4.325839519500732,
+ "step": 1712
+ },
+ {
+ "epoch": 23.79650655021834,
+ "grad_norm": 0.01862351782619953,
+ "learning_rate": 0.0006,
+ "loss": 4.333662509918213,
+ "step": 1713
+ },
+ {
+ "epoch": 23.810480349344978,
+ "grad_norm": 0.01655014418065548,
+ "learning_rate": 0.0006,
+ "loss": 4.39305305480957,
+ "step": 1714
+ },
+ {
+ "epoch": 23.824454148471617,
+ "grad_norm": 0.015899376943707466,
+ "learning_rate": 0.0006,
+ "loss": 4.390053749084473,
+ "step": 1715
+ },
+ {
+ "epoch": 23.83842794759825,
+ "grad_norm": 0.016519056633114815,
+ "learning_rate": 0.0006,
+ "loss": 4.255903244018555,
+ "step": 1716
+ },
+ {
+ "epoch": 23.85240174672489,
+ "grad_norm": 0.0167539119720459,
+ "learning_rate": 0.0006,
+ "loss": 4.255727767944336,
+ "step": 1717
+ },
+ {
+ "epoch": 23.86637554585153,
+ "grad_norm": 0.0171353816986084,
+ "learning_rate": 0.0006,
+ "loss": 4.376740455627441,
+ "step": 1718
+ },
+ {
+ "epoch": 23.880349344978168,
+ "grad_norm": 0.016719117760658264,
+ "learning_rate": 0.0006,
+ "loss": 4.3376688957214355,
+ "step": 1719
+ },
+ {
+ "epoch": 23.894323144104803,
+ "grad_norm": 0.01561494916677475,
+ "learning_rate": 0.0006,
+ "loss": 4.3345746994018555,
+ "step": 1720
+ },
+ {
+ "epoch": 23.90829694323144,
+ "grad_norm": 0.016303174197673798,
+ "learning_rate": 0.0006,
+ "loss": 4.332371234893799,
+ "step": 1721
+ },
+ {
+ "epoch": 23.92227074235808,
+ "grad_norm": 0.016722865402698517,
+ "learning_rate": 0.0006,
+ "loss": 4.267125129699707,
+ "step": 1722
+ },
+ {
+ "epoch": 23.936244541484715,
+ "grad_norm": 0.017072124406695366,
+ "learning_rate": 0.0006,
+ "loss": 4.322751998901367,
+ "step": 1723
+ },
+ {
+ "epoch": 23.950218340611354,
+ "grad_norm": 0.01699932850897312,
+ "learning_rate": 0.0006,
+ "loss": 4.378422737121582,
+ "step": 1724
+ },
+ {
+ "epoch": 23.964192139737992,
+ "grad_norm": 0.01742720976471901,
+ "learning_rate": 0.0006,
+ "loss": 4.345083236694336,
+ "step": 1725
+ },
+ {
+ "epoch": 23.978165938864628,
+ "grad_norm": 0.018797501921653748,
+ "learning_rate": 0.0006,
+ "loss": 4.290443420410156,
+ "step": 1726
+ },
+ {
+ "epoch": 23.992139737991266,
+ "grad_norm": 0.018090683966875076,
+ "learning_rate": 0.0006,
+ "loss": 4.377580642700195,
+ "step": 1727
+ },
+ {
+ "epoch": 24.0,
+ "grad_norm": 0.017305459827184677,
+ "learning_rate": 0.0006,
+ "loss": 4.364426612854004,
+ "step": 1728
+ },
+ {
+ "epoch": 24.0,
+ "eval_loss": 4.647762298583984,
+ "eval_runtime": 56.5402,
+ "eval_samples_per_second": 43.191,
+ "eval_steps_per_second": 1.362,
+ "step": 1728
+ },
+ {
+ "epoch": 24.01397379912664,
+ "grad_norm": 0.016493607312440872,
+ "learning_rate": 0.0006,
+ "loss": 4.252255916595459,
+ "step": 1729
+ },
+ {
+ "epoch": 24.027947598253274,
+ "grad_norm": 0.016529450193047523,
+ "learning_rate": 0.0006,
+ "loss": 4.352746963500977,
+ "step": 1730
+ },
+ {
+ "epoch": 24.041921397379912,
+ "grad_norm": 0.016940191388130188,
+ "learning_rate": 0.0006,
+ "loss": 4.301870822906494,
+ "step": 1731
+ },
+ {
+ "epoch": 24.05589519650655,
+ "grad_norm": 0.01752558909356594,
+ "learning_rate": 0.0006,
+ "loss": 4.2762451171875,
+ "step": 1732
+ },
+ {
+ "epoch": 24.069868995633186,
+ "grad_norm": 0.017112884670495987,
+ "learning_rate": 0.0006,
+ "loss": 4.2659149169921875,
+ "step": 1733
+ },
+ {
+ "epoch": 24.083842794759825,
+ "grad_norm": 0.01900586113333702,
+ "learning_rate": 0.0006,
+ "loss": 4.306058883666992,
+ "step": 1734
+ },
+ {
+ "epoch": 24.097816593886463,
+ "grad_norm": 0.019447680562734604,
+ "learning_rate": 0.0006,
+ "loss": 4.450128555297852,
+ "step": 1735
+ },
+ {
+ "epoch": 24.111790393013102,
+ "grad_norm": 0.01948186755180359,
+ "learning_rate": 0.0006,
+ "loss": 4.202869415283203,
+ "step": 1736
+ },
+ {
+ "epoch": 24.125764192139737,
+ "grad_norm": 0.01716383546590805,
+ "learning_rate": 0.0006,
+ "loss": 4.318876266479492,
+ "step": 1737
+ },
+ {
+ "epoch": 24.139737991266376,
+ "grad_norm": 0.017049988731741905,
+ "learning_rate": 0.0006,
+ "loss": 4.237238883972168,
+ "step": 1738
+ },
+ {
+ "epoch": 24.153711790393015,
+ "grad_norm": 0.015169923193752766,
+ "learning_rate": 0.0006,
+ "loss": 4.289426803588867,
+ "step": 1739
+ },
+ {
+ "epoch": 24.16768558951965,
+ "grad_norm": 0.014760667458176613,
+ "learning_rate": 0.0006,
+ "loss": 4.356924057006836,
+ "step": 1740
+ },
+ {
+ "epoch": 24.18165938864629,
+ "grad_norm": 0.015127750113606453,
+ "learning_rate": 0.0006,
+ "loss": 4.142916679382324,
+ "step": 1741
+ },
+ {
+ "epoch": 24.195633187772927,
+ "grad_norm": 0.01549555640667677,
+ "learning_rate": 0.0006,
+ "loss": 4.345563888549805,
+ "step": 1742
+ },
+ {
+ "epoch": 24.209606986899562,
+ "grad_norm": 0.016393091529607773,
+ "learning_rate": 0.0006,
+ "loss": 4.417495250701904,
+ "step": 1743
+ },
+ {
+ "epoch": 24.2235807860262,
+ "grad_norm": 0.01785266026854515,
+ "learning_rate": 0.0006,
+ "loss": 4.3553547859191895,
+ "step": 1744
+ },
+ {
+ "epoch": 24.23755458515284,
+ "grad_norm": 0.017574617639183998,
+ "learning_rate": 0.0006,
+ "loss": 4.241024494171143,
+ "step": 1745
+ },
+ {
+ "epoch": 24.251528384279474,
+ "grad_norm": 0.019060306251049042,
+ "learning_rate": 0.0006,
+ "loss": 4.3685150146484375,
+ "step": 1746
+ },
+ {
+ "epoch": 24.265502183406113,
+ "grad_norm": 0.0214633010327816,
+ "learning_rate": 0.0006,
+ "loss": 4.323866844177246,
+ "step": 1747
+ },
+ {
+ "epoch": 24.27947598253275,
+ "grad_norm": 0.021055003628134727,
+ "learning_rate": 0.0006,
+ "loss": 4.278757095336914,
+ "step": 1748
+ },
+ {
+ "epoch": 24.29344978165939,
+ "grad_norm": 0.018618572503328323,
+ "learning_rate": 0.0006,
+ "loss": 4.28849983215332,
+ "step": 1749
+ },
+ {
+ "epoch": 24.307423580786025,
+ "grad_norm": 0.018666435033082962,
+ "learning_rate": 0.0006,
+ "loss": 4.368590354919434,
+ "step": 1750
+ },
+ {
+ "epoch": 24.321397379912664,
+ "grad_norm": 0.020626146346330643,
+ "learning_rate": 0.0006,
+ "loss": 4.2951836585998535,
+ "step": 1751
+ },
+ {
+ "epoch": 24.335371179039303,
+ "grad_norm": 0.01929759792983532,
+ "learning_rate": 0.0006,
+ "loss": 4.3664703369140625,
+ "step": 1752
+ },
+ {
+ "epoch": 24.349344978165938,
+ "grad_norm": 0.020956002175807953,
+ "learning_rate": 0.0006,
+ "loss": 4.345308303833008,
+ "step": 1753
+ },
+ {
+ "epoch": 24.363318777292577,
+ "grad_norm": 0.019227957352995872,
+ "learning_rate": 0.0006,
+ "loss": 4.349849224090576,
+ "step": 1754
+ },
+ {
+ "epoch": 24.377292576419215,
+ "grad_norm": 0.015970097854733467,
+ "learning_rate": 0.0006,
+ "loss": 4.223012447357178,
+ "step": 1755
+ },
+ {
+ "epoch": 24.39126637554585,
+ "grad_norm": 0.015951646491885185,
+ "learning_rate": 0.0006,
+ "loss": 4.236420631408691,
+ "step": 1756
+ },
+ {
+ "epoch": 24.40524017467249,
+ "grad_norm": 0.016032515093684196,
+ "learning_rate": 0.0006,
+ "loss": 4.226899147033691,
+ "step": 1757
+ },
+ {
+ "epoch": 24.419213973799128,
+ "grad_norm": 0.01573132909834385,
+ "learning_rate": 0.0006,
+ "loss": 4.315916061401367,
+ "step": 1758
+ },
+ {
+ "epoch": 24.433187772925763,
+ "grad_norm": 0.015367005951702595,
+ "learning_rate": 0.0006,
+ "loss": 4.358939170837402,
+ "step": 1759
+ },
+ {
+ "epoch": 24.4471615720524,
+ "grad_norm": 0.01669907011091709,
+ "learning_rate": 0.0006,
+ "loss": 4.232147216796875,
+ "step": 1760
+ },
+ {
+ "epoch": 24.46113537117904,
+ "grad_norm": 0.018708152696490288,
+ "learning_rate": 0.0006,
+ "loss": 4.329852104187012,
+ "step": 1761
+ },
+ {
+ "epoch": 24.475109170305675,
+ "grad_norm": 0.019035987555980682,
+ "learning_rate": 0.0006,
+ "loss": 4.260340213775635,
+ "step": 1762
+ },
+ {
+ "epoch": 24.489082969432314,
+ "grad_norm": 0.01832144521176815,
+ "learning_rate": 0.0006,
+ "loss": 4.359543800354004,
+ "step": 1763
+ },
+ {
+ "epoch": 24.503056768558952,
+ "grad_norm": 0.015313294716179371,
+ "learning_rate": 0.0006,
+ "loss": 4.279088020324707,
+ "step": 1764
+ },
+ {
+ "epoch": 24.51703056768559,
+ "grad_norm": 0.015803923830389977,
+ "learning_rate": 0.0006,
+ "loss": 4.230682849884033,
+ "step": 1765
+ },
+ {
+ "epoch": 24.531004366812226,
+ "grad_norm": 0.016748374328017235,
+ "learning_rate": 0.0006,
+ "loss": 4.349661827087402,
+ "step": 1766
+ },
+ {
+ "epoch": 24.544978165938865,
+ "grad_norm": 0.015732428058981895,
+ "learning_rate": 0.0006,
+ "loss": 4.372441291809082,
+ "step": 1767
+ },
+ {
+ "epoch": 24.558951965065503,
+ "grad_norm": 0.016564829275012016,
+ "learning_rate": 0.0006,
+ "loss": 4.323090553283691,
+ "step": 1768
+ },
+ {
+ "epoch": 24.57292576419214,
+ "grad_norm": 0.01603010483086109,
+ "learning_rate": 0.0006,
+ "loss": 4.347898006439209,
+ "step": 1769
+ },
+ {
+ "epoch": 24.586899563318777,
+ "grad_norm": 0.01674809865653515,
+ "learning_rate": 0.0006,
+ "loss": 4.269165515899658,
+ "step": 1770
+ },
+ {
+ "epoch": 24.600873362445416,
+ "grad_norm": 0.017333930358290672,
+ "learning_rate": 0.0006,
+ "loss": 4.353298187255859,
+ "step": 1771
+ },
+ {
+ "epoch": 24.61484716157205,
+ "grad_norm": 0.017519580200314522,
+ "learning_rate": 0.0006,
+ "loss": 4.265199661254883,
+ "step": 1772
+ },
+ {
+ "epoch": 24.62882096069869,
+ "grad_norm": 0.017675306648015976,
+ "learning_rate": 0.0006,
+ "loss": 4.322596549987793,
+ "step": 1773
+ },
+ {
+ "epoch": 24.64279475982533,
+ "grad_norm": 0.017166534438729286,
+ "learning_rate": 0.0006,
+ "loss": 4.413900375366211,
+ "step": 1774
+ },
+ {
+ "epoch": 24.656768558951963,
+ "grad_norm": 0.016362346708774567,
+ "learning_rate": 0.0006,
+ "loss": 4.307236671447754,
+ "step": 1775
+ },
+ {
+ "epoch": 24.670742358078602,
+ "grad_norm": 0.016086872667074203,
+ "learning_rate": 0.0006,
+ "loss": 4.235316276550293,
+ "step": 1776
+ },
+ {
+ "epoch": 24.68471615720524,
+ "grad_norm": 0.01689537614583969,
+ "learning_rate": 0.0006,
+ "loss": 4.390591621398926,
+ "step": 1777
+ },
+ {
+ "epoch": 24.69868995633188,
+ "grad_norm": 0.016836725175380707,
+ "learning_rate": 0.0006,
+ "loss": 4.2990031242370605,
+ "step": 1778
+ },
+ {
+ "epoch": 24.712663755458514,
+ "grad_norm": 0.015627335757017136,
+ "learning_rate": 0.0006,
+ "loss": 4.166745185852051,
+ "step": 1779
+ },
+ {
+ "epoch": 24.726637554585153,
+ "grad_norm": 0.015007016249001026,
+ "learning_rate": 0.0006,
+ "loss": 4.295273303985596,
+ "step": 1780
+ },
+ {
+ "epoch": 24.74061135371179,
+ "grad_norm": 0.016084209084510803,
+ "learning_rate": 0.0006,
+ "loss": 4.342424392700195,
+ "step": 1781
+ },
+ {
+ "epoch": 24.754585152838427,
+ "grad_norm": 0.01595073565840721,
+ "learning_rate": 0.0006,
+ "loss": 4.3060102462768555,
+ "step": 1782
+ },
+ {
+ "epoch": 24.768558951965066,
+ "grad_norm": 0.016113461926579475,
+ "learning_rate": 0.0006,
+ "loss": 4.267908096313477,
+ "step": 1783
+ },
+ {
+ "epoch": 24.782532751091704,
+ "grad_norm": 0.01564556173980236,
+ "learning_rate": 0.0006,
+ "loss": 4.324642658233643,
+ "step": 1784
+ },
+ {
+ "epoch": 24.79650655021834,
+ "grad_norm": 0.01575268618762493,
+ "learning_rate": 0.0006,
+ "loss": 4.328181743621826,
+ "step": 1785
+ },
+ {
+ "epoch": 24.810480349344978,
+ "grad_norm": 0.016830824315547943,
+ "learning_rate": 0.0006,
+ "loss": 4.247906684875488,
+ "step": 1786
+ },
+ {
+ "epoch": 24.824454148471617,
+ "grad_norm": 0.017989547923207283,
+ "learning_rate": 0.0006,
+ "loss": 4.228281021118164,
+ "step": 1787
+ },
+ {
+ "epoch": 24.83842794759825,
+ "grad_norm": 0.016783086583018303,
+ "learning_rate": 0.0006,
+ "loss": 4.2880964279174805,
+ "step": 1788
+ },
+ {
+ "epoch": 24.85240174672489,
+ "grad_norm": 0.016467561945319176,
+ "learning_rate": 0.0006,
+ "loss": 4.36118745803833,
+ "step": 1789
+ },
+ {
+ "epoch": 24.86637554585153,
+ "grad_norm": 0.01780523732304573,
+ "learning_rate": 0.0006,
+ "loss": 4.172321319580078,
+ "step": 1790
+ },
+ {
+ "epoch": 24.880349344978168,
+ "grad_norm": 0.017102031037211418,
+ "learning_rate": 0.0006,
+ "loss": 4.352826118469238,
+ "step": 1791
+ },
+ {
+ "epoch": 24.894323144104803,
+ "grad_norm": 0.016409218311309814,
+ "learning_rate": 0.0006,
+ "loss": 4.212162017822266,
+ "step": 1792
+ },
+ {
+ "epoch": 24.90829694323144,
+ "grad_norm": 0.018660584464669228,
+ "learning_rate": 0.0006,
+ "loss": 4.346684455871582,
+ "step": 1793
+ },
+ {
+ "epoch": 24.92227074235808,
+ "grad_norm": 0.017389440909028053,
+ "learning_rate": 0.0006,
+ "loss": 4.197913646697998,
+ "step": 1794
+ },
+ {
+ "epoch": 24.936244541484715,
+ "grad_norm": 0.01596485823392868,
+ "learning_rate": 0.0006,
+ "loss": 4.370617866516113,
+ "step": 1795
+ },
+ {
+ "epoch": 24.950218340611354,
+ "grad_norm": 0.01709776371717453,
+ "learning_rate": 0.0006,
+ "loss": 4.313847064971924,
+ "step": 1796
+ },
+ {
+ "epoch": 24.964192139737992,
+ "grad_norm": 0.016571134328842163,
+ "learning_rate": 0.0006,
+ "loss": 4.1895856857299805,
+ "step": 1797
+ },
+ {
+ "epoch": 24.978165938864628,
+ "grad_norm": 0.016851048916578293,
+ "learning_rate": 0.0006,
+ "loss": 4.350772857666016,
+ "step": 1798
+ },
+ {
+ "epoch": 24.992139737991266,
+ "grad_norm": 0.017007440328598022,
+ "learning_rate": 0.0006,
+ "loss": 4.399008750915527,
+ "step": 1799
+ },
+ {
+ "epoch": 25.0,
+ "grad_norm": 0.01791212148964405,
+ "learning_rate": 0.0006,
+ "loss": 4.306121826171875,
+ "step": 1800
+ },
+ {
+ "epoch": 25.0,
+ "eval_loss": 4.636417865753174,
+ "eval_runtime": 56.8296,
+ "eval_samples_per_second": 42.971,
+ "eval_steps_per_second": 1.355,
+ "step": 1800
+ },
+ {
+ "epoch": 25.01397379912664,
+ "grad_norm": 0.01729617640376091,
+ "learning_rate": 0.0006,
+ "loss": 4.257018089294434,
+ "step": 1801
+ },
+ {
+ "epoch": 25.027947598253274,
+ "grad_norm": 0.017466643825173378,
+ "learning_rate": 0.0006,
+ "loss": 4.307605266571045,
+ "step": 1802
+ },
+ {
+ "epoch": 25.041921397379912,
+ "grad_norm": 0.018614279106259346,
+ "learning_rate": 0.0006,
+ "loss": 4.360141277313232,
+ "step": 1803
+ },
+ {
+ "epoch": 25.05589519650655,
+ "grad_norm": 0.02010771445930004,
+ "learning_rate": 0.0006,
+ "loss": 4.196428298950195,
+ "step": 1804
+ },
+ {
+ "epoch": 25.069868995633186,
+ "grad_norm": 0.02090228535234928,
+ "learning_rate": 0.0006,
+ "loss": 4.288947105407715,
+ "step": 1805
+ },
+ {
+ "epoch": 25.083842794759825,
+ "grad_norm": 0.02050255797803402,
+ "learning_rate": 0.0006,
+ "loss": 4.259582042694092,
+ "step": 1806
+ },
+ {
+ "epoch": 25.097816593886463,
+ "grad_norm": 0.021844089031219482,
+ "learning_rate": 0.0006,
+ "loss": 4.405138969421387,
+ "step": 1807
+ },
+ {
+ "epoch": 25.111790393013102,
+ "grad_norm": 0.022287018597126007,
+ "learning_rate": 0.0006,
+ "loss": 4.333793640136719,
+ "step": 1808
+ },
+ {
+ "epoch": 25.125764192139737,
+ "grad_norm": 0.02288329415023327,
+ "learning_rate": 0.0006,
+ "loss": 4.3910980224609375,
+ "step": 1809
+ },
+ {
+ "epoch": 25.139737991266376,
+ "grad_norm": 0.02373199723660946,
+ "learning_rate": 0.0006,
+ "loss": 4.32921028137207,
+ "step": 1810
+ },
+ {
+ "epoch": 25.153711790393015,
+ "grad_norm": 0.024960234761238098,
+ "learning_rate": 0.0006,
+ "loss": 4.295609474182129,
+ "step": 1811
+ },
+ {
+ "epoch": 25.16768558951965,
+ "grad_norm": 0.026743892580270767,
+ "learning_rate": 0.0006,
+ "loss": 4.315308094024658,
+ "step": 1812
+ },
+ {
+ "epoch": 25.18165938864629,
+ "grad_norm": 0.022138185799121857,
+ "learning_rate": 0.0006,
+ "loss": 4.360208511352539,
+ "step": 1813
+ },
+ {
+ "epoch": 25.195633187772927,
+ "grad_norm": 0.022120574489235878,
+ "learning_rate": 0.0006,
+ "loss": 4.1973772048950195,
+ "step": 1814
+ },
+ {
+ "epoch": 25.209606986899562,
+ "grad_norm": 0.02230502851307392,
+ "learning_rate": 0.0006,
+ "loss": 4.285558700561523,
+ "step": 1815
+ },
+ {
+ "epoch": 25.2235807860262,
+ "grad_norm": 0.022301986813545227,
+ "learning_rate": 0.0006,
+ "loss": 4.238343238830566,
+ "step": 1816
+ },
+ {
+ "epoch": 25.23755458515284,
+ "grad_norm": 0.020050447434186935,
+ "learning_rate": 0.0006,
+ "loss": 4.298235893249512,
+ "step": 1817
+ },
+ {
+ "epoch": 25.251528384279474,
+ "grad_norm": 0.01987724006175995,
+ "learning_rate": 0.0006,
+ "loss": 4.301384449005127,
+ "step": 1818
+ },
+ {
+ "epoch": 25.265502183406113,
+ "grad_norm": 0.020440855994820595,
+ "learning_rate": 0.0006,
+ "loss": 4.269218444824219,
+ "step": 1819
+ },
+ {
+ "epoch": 25.27947598253275,
+ "grad_norm": 0.01891893707215786,
+ "learning_rate": 0.0006,
+ "loss": 4.2401957511901855,
+ "step": 1820
+ },
+ {
+ "epoch": 25.29344978165939,
+ "grad_norm": 0.01903688907623291,
+ "learning_rate": 0.0006,
+ "loss": 4.139670372009277,
+ "step": 1821
+ },
+ {
+ "epoch": 25.307423580786025,
+ "grad_norm": 0.018856938928365707,
+ "learning_rate": 0.0006,
+ "loss": 4.233214378356934,
+ "step": 1822
+ },
+ {
+ "epoch": 25.321397379912664,
+ "grad_norm": 0.017791323363780975,
+ "learning_rate": 0.0006,
+ "loss": 4.2342023849487305,
+ "step": 1823
+ },
+ {
+ "epoch": 25.335371179039303,
+ "grad_norm": 0.017304565757513046,
+ "learning_rate": 0.0006,
+ "loss": 4.355198860168457,
+ "step": 1824
+ },
+ {
+ "epoch": 25.349344978165938,
+ "grad_norm": 0.01576206088066101,
+ "learning_rate": 0.0006,
+ "loss": 4.114147186279297,
+ "step": 1825
+ },
+ {
+ "epoch": 25.363318777292577,
+ "grad_norm": 0.015105154365301132,
+ "learning_rate": 0.0006,
+ "loss": 4.313795566558838,
+ "step": 1826
+ },
+ {
+ "epoch": 25.377292576419215,
+ "grad_norm": 0.01563587784767151,
+ "learning_rate": 0.0006,
+ "loss": 4.308282852172852,
+ "step": 1827
+ },
+ {
+ "epoch": 25.39126637554585,
+ "grad_norm": 0.015980906784534454,
+ "learning_rate": 0.0006,
+ "loss": 4.345149040222168,
+ "step": 1828
+ },
+ {
+ "epoch": 25.40524017467249,
+ "grad_norm": 0.01669284701347351,
+ "learning_rate": 0.0006,
+ "loss": 4.237611293792725,
+ "step": 1829
+ },
+ {
+ "epoch": 25.419213973799128,
+ "grad_norm": 0.016202064231038094,
+ "learning_rate": 0.0006,
+ "loss": 4.3072309494018555,
+ "step": 1830
+ },
+ {
+ "epoch": 25.433187772925763,
+ "grad_norm": 0.01478662807494402,
+ "learning_rate": 0.0006,
+ "loss": 4.276215076446533,
+ "step": 1831
+ },
+ {
+ "epoch": 25.4471615720524,
+ "grad_norm": 0.015168731100857258,
+ "learning_rate": 0.0006,
+ "loss": 4.37509822845459,
+ "step": 1832
+ },
+ {
+ "epoch": 25.46113537117904,
+ "grad_norm": 0.01648980937898159,
+ "learning_rate": 0.0006,
+ "loss": 4.349167346954346,
+ "step": 1833
+ },
+ {
+ "epoch": 25.475109170305675,
+ "grad_norm": 0.01708398200571537,
+ "learning_rate": 0.0006,
+ "loss": 4.346580505371094,
+ "step": 1834
+ },
+ {
+ "epoch": 25.489082969432314,
+ "grad_norm": 0.017370784655213356,
+ "learning_rate": 0.0006,
+ "loss": 4.299466133117676,
+ "step": 1835
+ },
+ {
+ "epoch": 25.503056768558952,
+ "grad_norm": 0.016248228028416634,
+ "learning_rate": 0.0006,
+ "loss": 4.184610366821289,
+ "step": 1836
+ },
+ {
+ "epoch": 25.51703056768559,
+ "grad_norm": 0.017106913030147552,
+ "learning_rate": 0.0006,
+ "loss": 4.366856098175049,
+ "step": 1837
+ },
+ {
+ "epoch": 25.531004366812226,
+ "grad_norm": 0.016071965917944908,
+ "learning_rate": 0.0006,
+ "loss": 4.297691345214844,
+ "step": 1838
+ },
+ {
+ "epoch": 25.544978165938865,
+ "grad_norm": 0.016451986506581306,
+ "learning_rate": 0.0006,
+ "loss": 4.229983329772949,
+ "step": 1839
+ },
+ {
+ "epoch": 25.558951965065503,
+ "grad_norm": 0.016195081174373627,
+ "learning_rate": 0.0006,
+ "loss": 4.259893417358398,
+ "step": 1840
+ },
+ {
+ "epoch": 25.57292576419214,
+ "grad_norm": 0.0165175162255764,
+ "learning_rate": 0.0006,
+ "loss": 4.3348236083984375,
+ "step": 1841
+ },
+ {
+ "epoch": 25.586899563318777,
+ "grad_norm": 0.01651517115533352,
+ "learning_rate": 0.0006,
+ "loss": 4.319541931152344,
+ "step": 1842
+ },
+ {
+ "epoch": 25.600873362445416,
+ "grad_norm": 0.016143690794706345,
+ "learning_rate": 0.0006,
+ "loss": 4.345652103424072,
+ "step": 1843
+ },
+ {
+ "epoch": 25.61484716157205,
+ "grad_norm": 0.015140696428716183,
+ "learning_rate": 0.0006,
+ "loss": 4.212047576904297,
+ "step": 1844
+ },
+ {
+ "epoch": 25.62882096069869,
+ "grad_norm": 0.014637443237006664,
+ "learning_rate": 0.0006,
+ "loss": 4.187453269958496,
+ "step": 1845
+ },
+ {
+ "epoch": 25.64279475982533,
+ "grad_norm": 0.014585614204406738,
+ "learning_rate": 0.0006,
+ "loss": 4.253936767578125,
+ "step": 1846
+ },
+ {
+ "epoch": 25.656768558951963,
+ "grad_norm": 0.015158289112150669,
+ "learning_rate": 0.0006,
+ "loss": 4.34612512588501,
+ "step": 1847
+ },
+ {
+ "epoch": 25.670742358078602,
+ "grad_norm": 0.015202849172055721,
+ "learning_rate": 0.0006,
+ "loss": 4.270634651184082,
+ "step": 1848
+ },
+ {
+ "epoch": 25.68471615720524,
+ "grad_norm": 0.015363575890660286,
+ "learning_rate": 0.0006,
+ "loss": 4.271080493927002,
+ "step": 1849
+ },
+ {
+ "epoch": 25.69868995633188,
+ "grad_norm": 0.014360226690769196,
+ "learning_rate": 0.0006,
+ "loss": 4.375033855438232,
+ "step": 1850
+ },
+ {
+ "epoch": 25.712663755458514,
+ "grad_norm": 0.014263181947171688,
+ "learning_rate": 0.0006,
+ "loss": 4.424138069152832,
+ "step": 1851
+ },
+ {
+ "epoch": 25.726637554585153,
+ "grad_norm": 0.014398688450455666,
+ "learning_rate": 0.0006,
+ "loss": 4.410154819488525,
+ "step": 1852
+ },
+ {
+ "epoch": 25.74061135371179,
+ "grad_norm": 0.016131488606333733,
+ "learning_rate": 0.0006,
+ "loss": 4.324014663696289,
+ "step": 1853
+ },
+ {
+ "epoch": 25.754585152838427,
+ "grad_norm": 0.017568735405802727,
+ "learning_rate": 0.0006,
+ "loss": 4.367861747741699,
+ "step": 1854
+ },
+ {
+ "epoch": 25.768558951965066,
+ "grad_norm": 0.017907511442899704,
+ "learning_rate": 0.0006,
+ "loss": 4.352965354919434,
+ "step": 1855
+ },
+ {
+ "epoch": 25.782532751091704,
+ "grad_norm": 0.016918016597628593,
+ "learning_rate": 0.0006,
+ "loss": 4.252397537231445,
+ "step": 1856
+ },
+ {
+ "epoch": 25.79650655021834,
+ "grad_norm": 0.017274610698223114,
+ "learning_rate": 0.0006,
+ "loss": 4.342309951782227,
+ "step": 1857
+ },
+ {
+ "epoch": 25.810480349344978,
+ "grad_norm": 0.017879825085401535,
+ "learning_rate": 0.0006,
+ "loss": 4.330987453460693,
+ "step": 1858
+ },
+ {
+ "epoch": 25.824454148471617,
+ "grad_norm": 0.01736099272966385,
+ "learning_rate": 0.0006,
+ "loss": 4.307255744934082,
+ "step": 1859
+ },
+ {
+ "epoch": 25.83842794759825,
+ "grad_norm": 0.016575824469327927,
+ "learning_rate": 0.0006,
+ "loss": 4.262016773223877,
+ "step": 1860
+ },
+ {
+ "epoch": 25.85240174672489,
+ "grad_norm": 0.015181140042841434,
+ "learning_rate": 0.0006,
+ "loss": 4.291863441467285,
+ "step": 1861
+ },
+ {
+ "epoch": 25.86637554585153,
+ "grad_norm": 0.015965687111020088,
+ "learning_rate": 0.0006,
+ "loss": 4.424836158752441,
+ "step": 1862
+ },
+ {
+ "epoch": 25.880349344978168,
+ "grad_norm": 0.01687219925224781,
+ "learning_rate": 0.0006,
+ "loss": 4.343531131744385,
+ "step": 1863
+ },
+ {
+ "epoch": 25.894323144104803,
+ "grad_norm": 0.015115504153072834,
+ "learning_rate": 0.0006,
+ "loss": 4.357808589935303,
+ "step": 1864
+ },
+ {
+ "epoch": 25.90829694323144,
+ "grad_norm": 0.014150998555123806,
+ "learning_rate": 0.0006,
+ "loss": 4.324550151824951,
+ "step": 1865
+ },
+ {
+ "epoch": 25.92227074235808,
+ "grad_norm": 0.014989510178565979,
+ "learning_rate": 0.0006,
+ "loss": 4.380428314208984,
+ "step": 1866
+ },
+ {
+ "epoch": 25.936244541484715,
+ "grad_norm": 0.01527960691601038,
+ "learning_rate": 0.0006,
+ "loss": 4.399374961853027,
+ "step": 1867
+ },
+ {
+ "epoch": 25.950218340611354,
+ "grad_norm": 0.015434633940458298,
+ "learning_rate": 0.0006,
+ "loss": 4.260984420776367,
+ "step": 1868
+ },
+ {
+ "epoch": 25.964192139737992,
+ "grad_norm": 0.015370228327810764,
+ "learning_rate": 0.0006,
+ "loss": 4.3340253829956055,
+ "step": 1869
+ },
+ {
+ "epoch": 25.978165938864628,
+ "grad_norm": 0.015348542481660843,
+ "learning_rate": 0.0006,
+ "loss": 4.369132995605469,
+ "step": 1870
+ },
+ {
+ "epoch": 25.992139737991266,
+ "grad_norm": 0.015354936942458153,
+ "learning_rate": 0.0006,
+ "loss": 4.381937026977539,
+ "step": 1871
+ },
+ {
+ "epoch": 26.0,
+ "grad_norm": 0.01796272210776806,
+ "learning_rate": 0.0006,
+ "loss": 4.286768436431885,
+ "step": 1872
+ },
+ {
+ "epoch": 26.0,
+ "eval_loss": 4.576984405517578,
+ "eval_runtime": 57.2091,
+ "eval_samples_per_second": 42.685,
+ "eval_steps_per_second": 1.346,
+ "step": 1872
+ },
+ {
+ "epoch": 26.01397379912664,
+ "grad_norm": 0.01748845726251602,
+ "learning_rate": 0.0006,
+ "loss": 4.337504863739014,
+ "step": 1873
+ },
+ {
+ "epoch": 26.027947598253274,
+ "grad_norm": 0.019244296476244926,
+ "learning_rate": 0.0006,
+ "loss": 4.264280796051025,
+ "step": 1874
+ },
+ {
+ "epoch": 26.041921397379912,
+ "grad_norm": 0.0215692650526762,
+ "learning_rate": 0.0006,
+ "loss": 4.350830078125,
+ "step": 1875
+ },
+ {
+ "epoch": 26.05589519650655,
+ "grad_norm": 0.022189252078533173,
+ "learning_rate": 0.0006,
+ "loss": 4.299098968505859,
+ "step": 1876
+ },
+ {
+ "epoch": 26.069868995633186,
+ "grad_norm": 0.022384043782949448,
+ "learning_rate": 0.0006,
+ "loss": 4.1185173988342285,
+ "step": 1877
+ },
+ {
+ "epoch": 26.083842794759825,
+ "grad_norm": 0.02076014317572117,
+ "learning_rate": 0.0006,
+ "loss": 4.297612190246582,
+ "step": 1878
+ },
+ {
+ "epoch": 26.097816593886463,
+ "grad_norm": 0.020843051373958588,
+ "learning_rate": 0.0006,
+ "loss": 4.294098854064941,
+ "step": 1879
+ },
+ {
+ "epoch": 26.111790393013102,
+ "grad_norm": 0.020654456689953804,
+ "learning_rate": 0.0006,
+ "loss": 4.267856597900391,
+ "step": 1880
+ },
+ {
+ "epoch": 26.125764192139737,
+ "grad_norm": 0.01860472559928894,
+ "learning_rate": 0.0006,
+ "loss": 4.284232139587402,
+ "step": 1881
+ },
+ {
+ "epoch": 26.139737991266376,
+ "grad_norm": 0.018719421699643135,
+ "learning_rate": 0.0006,
+ "loss": 4.277214527130127,
+ "step": 1882
+ },
+ {
+ "epoch": 26.153711790393015,
+ "grad_norm": 0.020903829485177994,
+ "learning_rate": 0.0006,
+ "loss": 4.254947662353516,
+ "step": 1883
+ },
+ {
+ "epoch": 26.16768558951965,
+ "grad_norm": 0.02020149491727352,
+ "learning_rate": 0.0006,
+ "loss": 4.2526936531066895,
+ "step": 1884
+ },
+ {
+ "epoch": 26.18165938864629,
+ "grad_norm": 0.0180392786860466,
+ "learning_rate": 0.0006,
+ "loss": 4.309922695159912,
+ "step": 1885
+ },
+ {
+ "epoch": 26.195633187772927,
+ "grad_norm": 0.018163125962018967,
+ "learning_rate": 0.0006,
+ "loss": 4.13860559463501,
+ "step": 1886
+ },
+ {
+ "epoch": 26.209606986899562,
+ "grad_norm": 0.01853291131556034,
+ "learning_rate": 0.0006,
+ "loss": 4.190926551818848,
+ "step": 1887
+ },
+ {
+ "epoch": 26.2235807860262,
+ "grad_norm": 0.01757677271962166,
+ "learning_rate": 0.0006,
+ "loss": 4.163999557495117,
+ "step": 1888
+ },
+ {
+ "epoch": 26.23755458515284,
+ "grad_norm": 0.01767992228269577,
+ "learning_rate": 0.0006,
+ "loss": 4.184430122375488,
+ "step": 1889
+ },
+ {
+ "epoch": 26.251528384279474,
+ "grad_norm": 0.018139973282814026,
+ "learning_rate": 0.0006,
+ "loss": 4.300766944885254,
+ "step": 1890
+ },
+ {
+ "epoch": 26.265502183406113,
+ "grad_norm": 0.020184461027383804,
+ "learning_rate": 0.0006,
+ "loss": 4.232537269592285,
+ "step": 1891
+ },
+ {
+ "epoch": 26.27947598253275,
+ "grad_norm": 0.02028856985270977,
+ "learning_rate": 0.0006,
+ "loss": 4.324883460998535,
+ "step": 1892
+ },
+ {
+ "epoch": 26.29344978165939,
+ "grad_norm": 0.017120616510510445,
+ "learning_rate": 0.0006,
+ "loss": 4.23493766784668,
+ "step": 1893
+ },
+ {
+ "epoch": 26.307423580786025,
+ "grad_norm": 0.016202978789806366,
+ "learning_rate": 0.0006,
+ "loss": 4.197602272033691,
+ "step": 1894
+ },
+ {
+ "epoch": 26.321397379912664,
+ "grad_norm": 0.017231421545147896,
+ "learning_rate": 0.0006,
+ "loss": 4.302001953125,
+ "step": 1895
+ },
+ {
+ "epoch": 26.335371179039303,
+ "grad_norm": 0.01725655607879162,
+ "learning_rate": 0.0006,
+ "loss": 4.356800079345703,
+ "step": 1896
+ },
+ {
+ "epoch": 26.349344978165938,
+ "grad_norm": 0.016657443717122078,
+ "learning_rate": 0.0006,
+ "loss": 4.266307830810547,
+ "step": 1897
+ },
+ {
+ "epoch": 26.363318777292577,
+ "grad_norm": 0.01616556942462921,
+ "learning_rate": 0.0006,
+ "loss": 4.307868003845215,
+ "step": 1898
+ },
+ {
+ "epoch": 26.377292576419215,
+ "grad_norm": 0.016621306538581848,
+ "learning_rate": 0.0006,
+ "loss": 4.172905921936035,
+ "step": 1899
+ },
+ {
+ "epoch": 26.39126637554585,
+ "grad_norm": 0.016243206337094307,
+ "learning_rate": 0.0006,
+ "loss": 4.162896156311035,
+ "step": 1900
+ },
+ {
+ "epoch": 26.40524017467249,
+ "grad_norm": 0.016663808375597,
+ "learning_rate": 0.0006,
+ "loss": 4.258232116699219,
+ "step": 1901
+ },
+ {
+ "epoch": 26.419213973799128,
+ "grad_norm": 0.01677478663623333,
+ "learning_rate": 0.0006,
+ "loss": 4.336090087890625,
+ "step": 1902
+ },
+ {
+ "epoch": 26.433187772925763,
+ "grad_norm": 0.01641303487122059,
+ "learning_rate": 0.0006,
+ "loss": 4.191476821899414,
+ "step": 1903
+ },
+ {
+ "epoch": 26.4471615720524,
+ "grad_norm": 0.016700323671102524,
+ "learning_rate": 0.0006,
+ "loss": 4.299427509307861,
+ "step": 1904
+ },
+ {
+ "epoch": 26.46113537117904,
+ "grad_norm": 0.017595035955309868,
+ "learning_rate": 0.0006,
+ "loss": 4.312828063964844,
+ "step": 1905
+ },
+ {
+ "epoch": 26.475109170305675,
+ "grad_norm": 0.020300284028053284,
+ "learning_rate": 0.0006,
+ "loss": 4.28331184387207,
+ "step": 1906
+ },
+ {
+ "epoch": 26.489082969432314,
+ "grad_norm": 0.02198321744799614,
+ "learning_rate": 0.0006,
+ "loss": 4.3280839920043945,
+ "step": 1907
+ },
+ {
+ "epoch": 26.503056768558952,
+ "grad_norm": 0.02003200724720955,
+ "learning_rate": 0.0006,
+ "loss": 4.325623989105225,
+ "step": 1908
+ },
+ {
+ "epoch": 26.51703056768559,
+ "grad_norm": 0.01870287023484707,
+ "learning_rate": 0.0006,
+ "loss": 4.251203536987305,
+ "step": 1909
+ },
+ {
+ "epoch": 26.531004366812226,
+ "grad_norm": 0.01835448481142521,
+ "learning_rate": 0.0006,
+ "loss": 4.335144996643066,
+ "step": 1910
+ },
+ {
+ "epoch": 26.544978165938865,
+ "grad_norm": 0.02046327292919159,
+ "learning_rate": 0.0006,
+ "loss": 4.238023281097412,
+ "step": 1911
+ },
+ {
+ "epoch": 26.558951965065503,
+ "grad_norm": 0.019771311432123184,
+ "learning_rate": 0.0006,
+ "loss": 4.353594779968262,
+ "step": 1912
+ },
+ {
+ "epoch": 26.57292576419214,
+ "grad_norm": 0.018090544268488884,
+ "learning_rate": 0.0006,
+ "loss": 4.304343223571777,
+ "step": 1913
+ },
+ {
+ "epoch": 26.586899563318777,
+ "grad_norm": 0.017489777877926826,
+ "learning_rate": 0.0006,
+ "loss": 4.204375267028809,
+ "step": 1914
+ },
+ {
+ "epoch": 26.600873362445416,
+ "grad_norm": 0.016785891726613045,
+ "learning_rate": 0.0006,
+ "loss": 4.286669731140137,
+ "step": 1915
+ },
+ {
+ "epoch": 26.61484716157205,
+ "grad_norm": 0.016382789239287376,
+ "learning_rate": 0.0006,
+ "loss": 4.347662448883057,
+ "step": 1916
+ },
+ {
+ "epoch": 26.62882096069869,
+ "grad_norm": 0.016581876203417778,
+ "learning_rate": 0.0006,
+ "loss": 4.2716193199157715,
+ "step": 1917
+ },
+ {
+ "epoch": 26.64279475982533,
+ "grad_norm": 0.015627184882760048,
+ "learning_rate": 0.0006,
+ "loss": 4.311489105224609,
+ "step": 1918
+ },
+ {
+ "epoch": 26.656768558951963,
+ "grad_norm": 0.015908483415842056,
+ "learning_rate": 0.0006,
+ "loss": 4.262099742889404,
+ "step": 1919
+ },
+ {
+ "epoch": 26.670742358078602,
+ "grad_norm": 0.015576236881315708,
+ "learning_rate": 0.0006,
+ "loss": 4.275871276855469,
+ "step": 1920
+ },
+ {
+ "epoch": 26.68471615720524,
+ "grad_norm": 0.014621064998209476,
+ "learning_rate": 0.0006,
+ "loss": 4.184260845184326,
+ "step": 1921
+ },
+ {
+ "epoch": 26.69868995633188,
+ "grad_norm": 0.014797811396420002,
+ "learning_rate": 0.0006,
+ "loss": 4.3412628173828125,
+ "step": 1922
+ },
+ {
+ "epoch": 26.712663755458514,
+ "grad_norm": 0.015610570088028908,
+ "learning_rate": 0.0006,
+ "loss": 4.181094646453857,
+ "step": 1923
+ },
+ {
+ "epoch": 26.726637554585153,
+ "grad_norm": 0.016048265621066093,
+ "learning_rate": 0.0006,
+ "loss": 4.365290641784668,
+ "step": 1924
+ },
+ {
+ "epoch": 26.74061135371179,
+ "grad_norm": 0.016604071483016014,
+ "learning_rate": 0.0006,
+ "loss": 4.2628912925720215,
+ "step": 1925
+ },
+ {
+ "epoch": 26.754585152838427,
+ "grad_norm": 0.017044221982359886,
+ "learning_rate": 0.0006,
+ "loss": 4.378512382507324,
+ "step": 1926
+ },
+ {
+ "epoch": 26.768558951965066,
+ "grad_norm": 0.01746384985744953,
+ "learning_rate": 0.0006,
+ "loss": 4.3888773918151855,
+ "step": 1927
+ },
+ {
+ "epoch": 26.782532751091704,
+ "grad_norm": 0.01572711206972599,
+ "learning_rate": 0.0006,
+ "loss": 4.189935684204102,
+ "step": 1928
+ },
+ {
+ "epoch": 26.79650655021834,
+ "grad_norm": 0.014976629987359047,
+ "learning_rate": 0.0006,
+ "loss": 4.280525207519531,
+ "step": 1929
+ },
+ {
+ "epoch": 26.810480349344978,
+ "grad_norm": 0.014804664999246597,
+ "learning_rate": 0.0006,
+ "loss": 4.270630836486816,
+ "step": 1930
+ },
+ {
+ "epoch": 26.824454148471617,
+ "grad_norm": 0.014641920104622841,
+ "learning_rate": 0.0006,
+ "loss": 4.319505214691162,
+ "step": 1931
+ },
+ {
+ "epoch": 26.83842794759825,
+ "grad_norm": 0.014928505755960941,
+ "learning_rate": 0.0006,
+ "loss": 4.180147647857666,
+ "step": 1932
+ },
+ {
+ "epoch": 26.85240174672489,
+ "grad_norm": 0.016130365431308746,
+ "learning_rate": 0.0006,
+ "loss": 4.260660648345947,
+ "step": 1933
+ },
+ {
+ "epoch": 26.86637554585153,
+ "grad_norm": 0.015642493963241577,
+ "learning_rate": 0.0006,
+ "loss": 4.304064750671387,
+ "step": 1934
+ },
+ {
+ "epoch": 26.880349344978168,
+ "grad_norm": 0.015988217666745186,
+ "learning_rate": 0.0006,
+ "loss": 4.2548017501831055,
+ "step": 1935
+ },
+ {
+ "epoch": 26.894323144104803,
+ "grad_norm": 0.016501398757100105,
+ "learning_rate": 0.0006,
+ "loss": 4.260058879852295,
+ "step": 1936
+ },
+ {
+ "epoch": 26.90829694323144,
+ "grad_norm": 0.01645040698349476,
+ "learning_rate": 0.0006,
+ "loss": 4.483081340789795,
+ "step": 1937
+ },
+ {
+ "epoch": 26.92227074235808,
+ "grad_norm": 0.016140863299369812,
+ "learning_rate": 0.0006,
+ "loss": 4.1776838302612305,
+ "step": 1938
+ },
+ {
+ "epoch": 26.936244541484715,
+ "grad_norm": 0.016681019216775894,
+ "learning_rate": 0.0006,
+ "loss": 4.224747657775879,
+ "step": 1939
+ },
+ {
+ "epoch": 26.950218340611354,
+ "grad_norm": 0.01650378853082657,
+ "learning_rate": 0.0006,
+ "loss": 4.305703163146973,
+ "step": 1940
+ },
+ {
+ "epoch": 26.964192139737992,
+ "grad_norm": 0.017051683738827705,
+ "learning_rate": 0.0006,
+ "loss": 4.285835266113281,
+ "step": 1941
+ },
+ {
+ "epoch": 26.978165938864628,
+ "grad_norm": 0.016894327476620674,
+ "learning_rate": 0.0006,
+ "loss": 4.318901538848877,
+ "step": 1942
+ },
+ {
+ "epoch": 26.992139737991266,
+ "grad_norm": 0.017313100397586823,
+ "learning_rate": 0.0006,
+ "loss": 4.281434535980225,
+ "step": 1943
+ },
+ {
+ "epoch": 27.0,
+ "grad_norm": 0.018450884148478508,
+ "learning_rate": 0.0006,
+ "loss": 4.335104942321777,
+ "step": 1944
+ },
+ {
+ "epoch": 27.0,
+ "eval_loss": 4.640687465667725,
+ "eval_runtime": 56.9624,
+ "eval_samples_per_second": 42.87,
+ "eval_steps_per_second": 1.352,
+ "step": 1944
+ },
+ {
+ "epoch": 27.01397379912664,
+ "grad_norm": 0.017755698412656784,
+ "learning_rate": 0.0006,
+ "loss": 4.278824329376221,
+ "step": 1945
+ },
+ {
+ "epoch": 27.027947598253274,
+ "grad_norm": 0.01724310778081417,
+ "learning_rate": 0.0006,
+ "loss": 4.311049461364746,
+ "step": 1946
+ },
+ {
+ "epoch": 27.041921397379912,
+ "grad_norm": 0.01690123789012432,
+ "learning_rate": 0.0006,
+ "loss": 4.261994361877441,
+ "step": 1947
+ },
+ {
+ "epoch": 27.05589519650655,
+ "grad_norm": 0.018406696617603302,
+ "learning_rate": 0.0006,
+ "loss": 4.1925153732299805,
+ "step": 1948
+ },
+ {
+ "epoch": 27.069868995633186,
+ "grad_norm": 0.019945867359638214,
+ "learning_rate": 0.0006,
+ "loss": 4.108537197113037,
+ "step": 1949
+ },
+ {
+ "epoch": 27.083842794759825,
+ "grad_norm": 0.020316628739237785,
+ "learning_rate": 0.0006,
+ "loss": 4.209043025970459,
+ "step": 1950
+ },
+ {
+ "epoch": 27.097816593886463,
+ "grad_norm": 0.019690370187163353,
+ "learning_rate": 0.0006,
+ "loss": 4.195789337158203,
+ "step": 1951
+ },
+ {
+ "epoch": 27.111790393013102,
+ "grad_norm": 0.01785232499241829,
+ "learning_rate": 0.0006,
+ "loss": 4.21466064453125,
+ "step": 1952
+ },
+ {
+ "epoch": 27.125764192139737,
+ "grad_norm": 0.0164767038077116,
+ "learning_rate": 0.0006,
+ "loss": 4.270247459411621,
+ "step": 1953
+ },
+ {
+ "epoch": 27.139737991266376,
+ "grad_norm": 0.018008455634117126,
+ "learning_rate": 0.0006,
+ "loss": 4.201042175292969,
+ "step": 1954
+ },
+ {
+ "epoch": 27.153711790393015,
+ "grad_norm": 0.01807340793311596,
+ "learning_rate": 0.0006,
+ "loss": 4.25289249420166,
+ "step": 1955
+ },
+ {
+ "epoch": 27.16768558951965,
+ "grad_norm": 0.016246909275650978,
+ "learning_rate": 0.0006,
+ "loss": 4.116283416748047,
+ "step": 1956
+ },
+ {
+ "epoch": 27.18165938864629,
+ "grad_norm": 0.017545776441693306,
+ "learning_rate": 0.0006,
+ "loss": 4.316999435424805,
+ "step": 1957
+ },
+ {
+ "epoch": 27.195633187772927,
+ "grad_norm": 0.017915375530719757,
+ "learning_rate": 0.0006,
+ "loss": 4.253859519958496,
+ "step": 1958
+ },
+ {
+ "epoch": 27.209606986899562,
+ "grad_norm": 0.018112223595380783,
+ "learning_rate": 0.0006,
+ "loss": 4.2541022300720215,
+ "step": 1959
+ },
+ {
+ "epoch": 27.2235807860262,
+ "grad_norm": 0.01846139505505562,
+ "learning_rate": 0.0006,
+ "loss": 4.20438289642334,
+ "step": 1960
+ },
+ {
+ "epoch": 27.23755458515284,
+ "grad_norm": 0.01943897269666195,
+ "learning_rate": 0.0006,
+ "loss": 4.2799577713012695,
+ "step": 1961
+ },
+ {
+ "epoch": 27.251528384279474,
+ "grad_norm": 0.018253348767757416,
+ "learning_rate": 0.0006,
+ "loss": 4.1720499992370605,
+ "step": 1962
+ },
+ {
+ "epoch": 27.265502183406113,
+ "grad_norm": 0.018292246386408806,
+ "learning_rate": 0.0006,
+ "loss": 4.28231143951416,
+ "step": 1963
+ },
+ {
+ "epoch": 27.27947598253275,
+ "grad_norm": 0.01943155936896801,
+ "learning_rate": 0.0006,
+ "loss": 4.314976692199707,
+ "step": 1964
+ },
+ {
+ "epoch": 27.29344978165939,
+ "grad_norm": 0.019064994528889656,
+ "learning_rate": 0.0006,
+ "loss": 4.25328254699707,
+ "step": 1965
+ },
+ {
+ "epoch": 27.307423580786025,
+ "grad_norm": 0.017423273995518684,
+ "learning_rate": 0.0006,
+ "loss": 4.195992469787598,
+ "step": 1966
+ },
+ {
+ "epoch": 27.321397379912664,
+ "grad_norm": 0.01677924580872059,
+ "learning_rate": 0.0006,
+ "loss": 4.260043144226074,
+ "step": 1967
+ },
+ {
+ "epoch": 27.335371179039303,
+ "grad_norm": 0.01720673404633999,
+ "learning_rate": 0.0006,
+ "loss": 4.222780704498291,
+ "step": 1968
+ },
+ {
+ "epoch": 27.349344978165938,
+ "grad_norm": 0.017300743609666824,
+ "learning_rate": 0.0006,
+ "loss": 4.109524250030518,
+ "step": 1969
+ },
+ {
+ "epoch": 27.363318777292577,
+ "grad_norm": 0.016065871343016624,
+ "learning_rate": 0.0006,
+ "loss": 4.36108922958374,
+ "step": 1970
+ },
+ {
+ "epoch": 27.377292576419215,
+ "grad_norm": 0.016626616939902306,
+ "learning_rate": 0.0006,
+ "loss": 4.220630645751953,
+ "step": 1971
+ },
+ {
+ "epoch": 27.39126637554585,
+ "grad_norm": 0.016482602804899216,
+ "learning_rate": 0.0006,
+ "loss": 4.344797134399414,
+ "step": 1972
+ },
+ {
+ "epoch": 27.40524017467249,
+ "grad_norm": 0.01727953553199768,
+ "learning_rate": 0.0006,
+ "loss": 4.296645164489746,
+ "step": 1973
+ },
+ {
+ "epoch": 27.419213973799128,
+ "grad_norm": 0.01522731315344572,
+ "learning_rate": 0.0006,
+ "loss": 4.223751068115234,
+ "step": 1974
+ },
+ {
+ "epoch": 27.433187772925763,
+ "grad_norm": 0.01581776700913906,
+ "learning_rate": 0.0006,
+ "loss": 4.351379871368408,
+ "step": 1975
+ },
+ {
+ "epoch": 27.4471615720524,
+ "grad_norm": 0.016394605860114098,
+ "learning_rate": 0.0006,
+ "loss": 4.2412919998168945,
+ "step": 1976
+ },
+ {
+ "epoch": 27.46113537117904,
+ "grad_norm": 0.01751169003546238,
+ "learning_rate": 0.0006,
+ "loss": 4.257606506347656,
+ "step": 1977
+ },
+ {
+ "epoch": 27.475109170305675,
+ "grad_norm": 0.017452310770750046,
+ "learning_rate": 0.0006,
+ "loss": 4.380134582519531,
+ "step": 1978
+ },
+ {
+ "epoch": 27.489082969432314,
+ "grad_norm": 0.017741898074746132,
+ "learning_rate": 0.0006,
+ "loss": 4.279632568359375,
+ "step": 1979
+ },
+ {
+ "epoch": 27.503056768558952,
+ "grad_norm": 0.017680590972304344,
+ "learning_rate": 0.0006,
+ "loss": 4.273406028747559,
+ "step": 1980
+ },
+ {
+ "epoch": 27.51703056768559,
+ "grad_norm": 0.01864718459546566,
+ "learning_rate": 0.0006,
+ "loss": 4.278583526611328,
+ "step": 1981
+ },
+ {
+ "epoch": 27.531004366812226,
+ "grad_norm": 0.019558541476726532,
+ "learning_rate": 0.0006,
+ "loss": 4.211919784545898,
+ "step": 1982
+ },
+ {
+ "epoch": 27.544978165938865,
+ "grad_norm": 0.019467400386929512,
+ "learning_rate": 0.0006,
+ "loss": 4.276822566986084,
+ "step": 1983
+ },
+ {
+ "epoch": 27.558951965065503,
+ "grad_norm": 0.01756172813475132,
+ "learning_rate": 0.0006,
+ "loss": 4.165379047393799,
+ "step": 1984
+ },
+ {
+ "epoch": 27.57292576419214,
+ "grad_norm": 0.017458263784646988,
+ "learning_rate": 0.0006,
+ "loss": 4.2992401123046875,
+ "step": 1985
+ },
+ {
+ "epoch": 27.586899563318777,
+ "grad_norm": 0.018089665099978447,
+ "learning_rate": 0.0006,
+ "loss": 4.243378639221191,
+ "step": 1986
+ },
+ {
+ "epoch": 27.600873362445416,
+ "grad_norm": 0.019535524770617485,
+ "learning_rate": 0.0006,
+ "loss": 4.321477890014648,
+ "step": 1987
+ },
+ {
+ "epoch": 27.61484716157205,
+ "grad_norm": 0.01997970975935459,
+ "learning_rate": 0.0006,
+ "loss": 4.123082637786865,
+ "step": 1988
+ },
+ {
+ "epoch": 27.62882096069869,
+ "grad_norm": 0.01913067139685154,
+ "learning_rate": 0.0006,
+ "loss": 4.289896488189697,
+ "step": 1989
+ },
+ {
+ "epoch": 27.64279475982533,
+ "grad_norm": 0.0169806070625782,
+ "learning_rate": 0.0006,
+ "loss": 4.247004985809326,
+ "step": 1990
+ },
+ {
+ "epoch": 27.656768558951963,
+ "grad_norm": 0.017315007746219635,
+ "learning_rate": 0.0006,
+ "loss": 4.2095441818237305,
+ "step": 1991
+ },
+ {
+ "epoch": 27.670742358078602,
+ "grad_norm": 0.015098759904503822,
+ "learning_rate": 0.0006,
+ "loss": 4.145560264587402,
+ "step": 1992
+ },
+ {
+ "epoch": 27.68471615720524,
+ "grad_norm": 0.016240514814853668,
+ "learning_rate": 0.0006,
+ "loss": 4.254927635192871,
+ "step": 1993
+ },
+ {
+ "epoch": 27.69868995633188,
+ "grad_norm": 0.01621108129620552,
+ "learning_rate": 0.0006,
+ "loss": 4.219725608825684,
+ "step": 1994
+ },
+ {
+ "epoch": 27.712663755458514,
+ "grad_norm": 0.015191009268164635,
+ "learning_rate": 0.0006,
+ "loss": 4.194345474243164,
+ "step": 1995
+ },
+ {
+ "epoch": 27.726637554585153,
+ "grad_norm": 0.014387983828783035,
+ "learning_rate": 0.0006,
+ "loss": 4.211467742919922,
+ "step": 1996
+ },
+ {
+ "epoch": 27.74061135371179,
+ "grad_norm": 0.015460561029613018,
+ "learning_rate": 0.0006,
+ "loss": 4.358700752258301,
+ "step": 1997
+ },
+ {
+ "epoch": 27.754585152838427,
+ "grad_norm": 0.015100076794624329,
+ "learning_rate": 0.0006,
+ "loss": 4.171223163604736,
+ "step": 1998
+ },
+ {
+ "epoch": 27.768558951965066,
+ "grad_norm": 0.015607891604304314,
+ "learning_rate": 0.0006,
+ "loss": 4.217874526977539,
+ "step": 1999
+ },
+ {
+ "epoch": 27.782532751091704,
+ "grad_norm": 0.015650689601898193,
+ "learning_rate": 0.0006,
+ "loss": 4.280780792236328,
+ "step": 2000
+ },
+ {
+ "epoch": 27.79650655021834,
+ "grad_norm": 0.01580994389951229,
+ "learning_rate": 0.0006,
+ "loss": 4.159212112426758,
+ "step": 2001
+ },
+ {
+ "epoch": 27.810480349344978,
+ "grad_norm": 0.016607413068413734,
+ "learning_rate": 0.0006,
+ "loss": 4.238853454589844,
+ "step": 2002
+ },
+ {
+ "epoch": 27.824454148471617,
+ "grad_norm": 0.017383860424160957,
+ "learning_rate": 0.0006,
+ "loss": 4.233944892883301,
+ "step": 2003
+ },
+ {
+ "epoch": 27.83842794759825,
+ "grad_norm": 0.014966659247875214,
+ "learning_rate": 0.0006,
+ "loss": 4.167392730712891,
+ "step": 2004
+ },
+ {
+ "epoch": 27.85240174672489,
+ "grad_norm": 0.014589878730475903,
+ "learning_rate": 0.0006,
+ "loss": 4.2390899658203125,
+ "step": 2005
+ },
+ {
+ "epoch": 27.86637554585153,
+ "grad_norm": 0.017179450020194054,
+ "learning_rate": 0.0006,
+ "loss": 4.365785598754883,
+ "step": 2006
+ },
+ {
+ "epoch": 27.880349344978168,
+ "grad_norm": 0.01893901452422142,
+ "learning_rate": 0.0006,
+ "loss": 4.289430141448975,
+ "step": 2007
+ },
+ {
+ "epoch": 27.894323144104803,
+ "grad_norm": 0.018619712442159653,
+ "learning_rate": 0.0006,
+ "loss": 4.310665130615234,
+ "step": 2008
+ },
+ {
+ "epoch": 27.90829694323144,
+ "grad_norm": 0.019180385395884514,
+ "learning_rate": 0.0006,
+ "loss": 4.27454137802124,
+ "step": 2009
+ },
+ {
+ "epoch": 27.92227074235808,
+ "grad_norm": 0.017811523750424385,
+ "learning_rate": 0.0006,
+ "loss": 4.312562465667725,
+ "step": 2010
+ },
+ {
+ "epoch": 27.936244541484715,
+ "grad_norm": 0.016583573073148727,
+ "learning_rate": 0.0006,
+ "loss": 4.202253341674805,
+ "step": 2011
+ },
+ {
+ "epoch": 27.950218340611354,
+ "grad_norm": 0.01658615842461586,
+ "learning_rate": 0.0006,
+ "loss": 4.183433532714844,
+ "step": 2012
+ },
+ {
+ "epoch": 27.964192139737992,
+ "grad_norm": 0.015327401459217072,
+ "learning_rate": 0.0006,
+ "loss": 4.2125139236450195,
+ "step": 2013
+ },
+ {
+ "epoch": 27.978165938864628,
+ "grad_norm": 0.015004601329565048,
+ "learning_rate": 0.0006,
+ "loss": 4.257392883300781,
+ "step": 2014
+ },
+ {
+ "epoch": 27.992139737991266,
+ "grad_norm": 0.015259161591529846,
+ "learning_rate": 0.0006,
+ "loss": 4.241818428039551,
+ "step": 2015
+ },
+ {
+ "epoch": 28.0,
+ "grad_norm": 0.016808776184916496,
+ "learning_rate": 0.0006,
+ "loss": 4.146241188049316,
+ "step": 2016
+ },
+ {
+ "epoch": 28.0,
+ "eval_loss": 4.663090705871582,
+ "eval_runtime": 56.4559,
+ "eval_samples_per_second": 43.255,
+ "eval_steps_per_second": 1.364,
+ "step": 2016
+ },
+ {
+ "epoch": 28.01397379912664,
+ "grad_norm": 0.015784382820129395,
+ "learning_rate": 0.0006,
+ "loss": 4.044191360473633,
+ "step": 2017
+ },
+ {
+ "epoch": 28.027947598253274,
+ "grad_norm": 0.015636710450053215,
+ "learning_rate": 0.0006,
+ "loss": 4.128754615783691,
+ "step": 2018
+ },
+ {
+ "epoch": 28.041921397379912,
+ "grad_norm": 0.015364975668489933,
+ "learning_rate": 0.0006,
+ "loss": 4.191835403442383,
+ "step": 2019
+ },
+ {
+ "epoch": 28.05589519650655,
+ "grad_norm": 0.015015700832009315,
+ "learning_rate": 0.0006,
+ "loss": 4.120022773742676,
+ "step": 2020
+ },
+ {
+ "epoch": 28.069868995633186,
+ "grad_norm": 0.016666799783706665,
+ "learning_rate": 0.0006,
+ "loss": 4.33906364440918,
+ "step": 2021
+ },
+ {
+ "epoch": 28.083842794759825,
+ "grad_norm": 0.01780104823410511,
+ "learning_rate": 0.0006,
+ "loss": 4.279458999633789,
+ "step": 2022
+ },
+ {
+ "epoch": 28.097816593886463,
+ "grad_norm": 0.01870882883667946,
+ "learning_rate": 0.0006,
+ "loss": 4.34728479385376,
+ "step": 2023
+ },
+ {
+ "epoch": 28.111790393013102,
+ "grad_norm": 0.018406381830573082,
+ "learning_rate": 0.0006,
+ "loss": 4.211311340332031,
+ "step": 2024
+ },
+ {
+ "epoch": 28.125764192139737,
+ "grad_norm": 0.017833448946475983,
+ "learning_rate": 0.0006,
+ "loss": 4.223254203796387,
+ "step": 2025
+ },
+ {
+ "epoch": 28.139737991266376,
+ "grad_norm": 0.017552688717842102,
+ "learning_rate": 0.0006,
+ "loss": 4.208823204040527,
+ "step": 2026
+ },
+ {
+ "epoch": 28.153711790393015,
+ "grad_norm": 0.018500229343771935,
+ "learning_rate": 0.0006,
+ "loss": 4.198025226593018,
+ "step": 2027
+ },
+ {
+ "epoch": 28.16768558951965,
+ "grad_norm": 0.01782156340777874,
+ "learning_rate": 0.0006,
+ "loss": 4.179978370666504,
+ "step": 2028
+ },
+ {
+ "epoch": 28.18165938864629,
+ "grad_norm": 0.019580967724323273,
+ "learning_rate": 0.0006,
+ "loss": 4.291114807128906,
+ "step": 2029
+ },
+ {
+ "epoch": 28.195633187772927,
+ "grad_norm": 0.020804036408662796,
+ "learning_rate": 0.0006,
+ "loss": 4.220607280731201,
+ "step": 2030
+ },
+ {
+ "epoch": 28.209606986899562,
+ "grad_norm": 0.019595693796873093,
+ "learning_rate": 0.0006,
+ "loss": 4.2905731201171875,
+ "step": 2031
+ },
+ {
+ "epoch": 28.2235807860262,
+ "grad_norm": 0.019247086718678474,
+ "learning_rate": 0.0006,
+ "loss": 4.2467360496521,
+ "step": 2032
+ },
+ {
+ "epoch": 28.23755458515284,
+ "grad_norm": 0.01958036608994007,
+ "learning_rate": 0.0006,
+ "loss": 4.286137580871582,
+ "step": 2033
+ },
+ {
+ "epoch": 28.251528384279474,
+ "grad_norm": 0.021354753524065018,
+ "learning_rate": 0.0006,
+ "loss": 4.234755039215088,
+ "step": 2034
+ },
+ {
+ "epoch": 28.265502183406113,
+ "grad_norm": 0.02206375077366829,
+ "learning_rate": 0.0006,
+ "loss": 4.149641036987305,
+ "step": 2035
+ },
+ {
+ "epoch": 28.27947598253275,
+ "grad_norm": 0.02201310358941555,
+ "learning_rate": 0.0006,
+ "loss": 4.1928629875183105,
+ "step": 2036
+ },
+ {
+ "epoch": 28.29344978165939,
+ "grad_norm": 0.01961950585246086,
+ "learning_rate": 0.0006,
+ "loss": 4.211404323577881,
+ "step": 2037
+ },
+ {
+ "epoch": 28.307423580786025,
+ "grad_norm": 0.019211795181035995,
+ "learning_rate": 0.0006,
+ "loss": 4.200976848602295,
+ "step": 2038
+ },
+ {
+ "epoch": 28.321397379912664,
+ "grad_norm": 0.01779370754957199,
+ "learning_rate": 0.0006,
+ "loss": 4.249148845672607,
+ "step": 2039
+ },
+ {
+ "epoch": 28.335371179039303,
+ "grad_norm": 0.018635908141732216,
+ "learning_rate": 0.0006,
+ "loss": 4.11806058883667,
+ "step": 2040
+ },
+ {
+ "epoch": 28.349344978165938,
+ "grad_norm": 0.019792694598436356,
+ "learning_rate": 0.0006,
+ "loss": 4.335224151611328,
+ "step": 2041
+ },
+ {
+ "epoch": 28.363318777292577,
+ "grad_norm": 0.021731717512011528,
+ "learning_rate": 0.0006,
+ "loss": 4.302778244018555,
+ "step": 2042
+ },
+ {
+ "epoch": 28.377292576419215,
+ "grad_norm": 0.020823447033762932,
+ "learning_rate": 0.0006,
+ "loss": 4.260610580444336,
+ "step": 2043
+ },
+ {
+ "epoch": 28.39126637554585,
+ "grad_norm": 0.019385412335395813,
+ "learning_rate": 0.0006,
+ "loss": 4.2247724533081055,
+ "step": 2044
+ },
+ {
+ "epoch": 28.40524017467249,
+ "grad_norm": 0.019539108499884605,
+ "learning_rate": 0.0006,
+ "loss": 4.236655235290527,
+ "step": 2045
+ },
+ {
+ "epoch": 28.419213973799128,
+ "grad_norm": 0.01942690648138523,
+ "learning_rate": 0.0006,
+ "loss": 4.251850605010986,
+ "step": 2046
+ },
+ {
+ "epoch": 28.433187772925763,
+ "grad_norm": 0.019667886197566986,
+ "learning_rate": 0.0006,
+ "loss": 4.222312927246094,
+ "step": 2047
+ },
+ {
+ "epoch": 28.4471615720524,
+ "grad_norm": 0.02029012329876423,
+ "learning_rate": 0.0006,
+ "loss": 4.246252059936523,
+ "step": 2048
+ },
+ {
+ "epoch": 28.46113537117904,
+ "grad_norm": 0.01784469373524189,
+ "learning_rate": 0.0006,
+ "loss": 4.192128658294678,
+ "step": 2049
+ },
+ {
+ "epoch": 28.475109170305675,
+ "grad_norm": 0.0160287544131279,
+ "learning_rate": 0.0006,
+ "loss": 4.208607196807861,
+ "step": 2050
+ },
+ {
+ "epoch": 28.489082969432314,
+ "grad_norm": 0.015348346903920174,
+ "learning_rate": 0.0006,
+ "loss": 4.236598014831543,
+ "step": 2051
+ },
+ {
+ "epoch": 28.503056768558952,
+ "grad_norm": 0.015198206529021263,
+ "learning_rate": 0.0006,
+ "loss": 4.235552787780762,
+ "step": 2052
+ },
+ {
+ "epoch": 28.51703056768559,
+ "grad_norm": 0.01523754745721817,
+ "learning_rate": 0.0006,
+ "loss": 4.1941728591918945,
+ "step": 2053
+ },
+ {
+ "epoch": 28.531004366812226,
+ "grad_norm": 0.015559614636003971,
+ "learning_rate": 0.0006,
+ "loss": 4.294757843017578,
+ "step": 2054
+ },
+ {
+ "epoch": 28.544978165938865,
+ "grad_norm": 0.01568358950316906,
+ "learning_rate": 0.0006,
+ "loss": 4.213065147399902,
+ "step": 2055
+ },
+ {
+ "epoch": 28.558951965065503,
+ "grad_norm": 0.015336516313254833,
+ "learning_rate": 0.0006,
+ "loss": 4.240293979644775,
+ "step": 2056
+ },
+ {
+ "epoch": 28.57292576419214,
+ "grad_norm": 0.015041496604681015,
+ "learning_rate": 0.0006,
+ "loss": 4.273397445678711,
+ "step": 2057
+ },
+ {
+ "epoch": 28.586899563318777,
+ "grad_norm": 0.01610128954052925,
+ "learning_rate": 0.0006,
+ "loss": 4.125369071960449,
+ "step": 2058
+ },
+ {
+ "epoch": 28.600873362445416,
+ "grad_norm": 0.015974976122379303,
+ "learning_rate": 0.0006,
+ "loss": 4.133459091186523,
+ "step": 2059
+ },
+ {
+ "epoch": 28.61484716157205,
+ "grad_norm": 0.015936799347400665,
+ "learning_rate": 0.0006,
+ "loss": 4.306667327880859,
+ "step": 2060
+ },
+ {
+ "epoch": 28.62882096069869,
+ "grad_norm": 0.01795247197151184,
+ "learning_rate": 0.0006,
+ "loss": 4.205156326293945,
+ "step": 2061
+ },
+ {
+ "epoch": 28.64279475982533,
+ "grad_norm": 0.019639814272522926,
+ "learning_rate": 0.0006,
+ "loss": 4.277614593505859,
+ "step": 2062
+ },
+ {
+ "epoch": 28.656768558951963,
+ "grad_norm": 0.01722300611436367,
+ "learning_rate": 0.0006,
+ "loss": 4.191032886505127,
+ "step": 2063
+ },
+ {
+ "epoch": 28.670742358078602,
+ "grad_norm": 0.01612633652985096,
+ "learning_rate": 0.0006,
+ "loss": 4.20717191696167,
+ "step": 2064
+ },
+ {
+ "epoch": 28.68471615720524,
+ "grad_norm": 0.01603279449045658,
+ "learning_rate": 0.0006,
+ "loss": 4.228884220123291,
+ "step": 2065
+ },
+ {
+ "epoch": 28.69868995633188,
+ "grad_norm": 0.016405927017331123,
+ "learning_rate": 0.0006,
+ "loss": 4.201292991638184,
+ "step": 2066
+ },
+ {
+ "epoch": 28.712663755458514,
+ "grad_norm": 0.016644183546304703,
+ "learning_rate": 0.0006,
+ "loss": 4.239282131195068,
+ "step": 2067
+ },
+ {
+ "epoch": 28.726637554585153,
+ "grad_norm": 0.0175587497651577,
+ "learning_rate": 0.0006,
+ "loss": 4.289038181304932,
+ "step": 2068
+ },
+ {
+ "epoch": 28.74061135371179,
+ "grad_norm": 0.018223397433757782,
+ "learning_rate": 0.0006,
+ "loss": 4.2266998291015625,
+ "step": 2069
+ },
+ {
+ "epoch": 28.754585152838427,
+ "grad_norm": 0.018680868670344353,
+ "learning_rate": 0.0006,
+ "loss": 4.348849296569824,
+ "step": 2070
+ },
+ {
+ "epoch": 28.768558951965066,
+ "grad_norm": 0.016726728528738022,
+ "learning_rate": 0.0006,
+ "loss": 4.292080879211426,
+ "step": 2071
+ },
+ {
+ "epoch": 28.782532751091704,
+ "grad_norm": 0.017386795952916145,
+ "learning_rate": 0.0006,
+ "loss": 4.208870887756348,
+ "step": 2072
+ },
+ {
+ "epoch": 28.79650655021834,
+ "grad_norm": 0.01865178905427456,
+ "learning_rate": 0.0006,
+ "loss": 4.054150581359863,
+ "step": 2073
+ },
+ {
+ "epoch": 28.810480349344978,
+ "grad_norm": 0.01776815392076969,
+ "learning_rate": 0.0006,
+ "loss": 4.202404499053955,
+ "step": 2074
+ },
+ {
+ "epoch": 28.824454148471617,
+ "grad_norm": 0.01643497310578823,
+ "learning_rate": 0.0006,
+ "loss": 4.131855010986328,
+ "step": 2075
+ },
+ {
+ "epoch": 28.83842794759825,
+ "grad_norm": 0.01574273779988289,
+ "learning_rate": 0.0006,
+ "loss": 4.296788692474365,
+ "step": 2076
+ },
+ {
+ "epoch": 28.85240174672489,
+ "grad_norm": 0.015613908879458904,
+ "learning_rate": 0.0006,
+ "loss": 4.210949897766113,
+ "step": 2077
+ },
+ {
+ "epoch": 28.86637554585153,
+ "grad_norm": 0.015543212182819843,
+ "learning_rate": 0.0006,
+ "loss": 4.218136310577393,
+ "step": 2078
+ },
+ {
+ "epoch": 28.880349344978168,
+ "grad_norm": 0.014681251719594002,
+ "learning_rate": 0.0006,
+ "loss": 4.19401741027832,
+ "step": 2079
+ },
+ {
+ "epoch": 28.894323144104803,
+ "grad_norm": 0.014393828809261322,
+ "learning_rate": 0.0006,
+ "loss": 4.198468208312988,
+ "step": 2080
+ },
+ {
+ "epoch": 28.90829694323144,
+ "grad_norm": 0.01580170728266239,
+ "learning_rate": 0.0006,
+ "loss": 4.224969863891602,
+ "step": 2081
+ },
+ {
+ "epoch": 28.92227074235808,
+ "grad_norm": 0.01484165620058775,
+ "learning_rate": 0.0006,
+ "loss": 4.12236213684082,
+ "step": 2082
+ },
+ {
+ "epoch": 28.936244541484715,
+ "grad_norm": 0.01497623696923256,
+ "learning_rate": 0.0006,
+ "loss": 4.24141788482666,
+ "step": 2083
+ },
+ {
+ "epoch": 28.950218340611354,
+ "grad_norm": 0.015080046840012074,
+ "learning_rate": 0.0006,
+ "loss": 4.284151077270508,
+ "step": 2084
+ },
+ {
+ "epoch": 28.964192139737992,
+ "grad_norm": 0.016207212582230568,
+ "learning_rate": 0.0006,
+ "loss": 4.298693656921387,
+ "step": 2085
+ },
+ {
+ "epoch": 28.978165938864628,
+ "grad_norm": 0.015474417246878147,
+ "learning_rate": 0.0006,
+ "loss": 4.2412261962890625,
+ "step": 2086
+ },
+ {
+ "epoch": 28.992139737991266,
+ "grad_norm": 0.01515351701527834,
+ "learning_rate": 0.0006,
+ "loss": 4.3054423332214355,
+ "step": 2087
+ },
+ {
+ "epoch": 29.0,
+ "grad_norm": 0.016411345452070236,
+ "learning_rate": 0.0006,
+ "loss": 4.104830741882324,
+ "step": 2088
+ },
+ {
+ "epoch": 29.0,
+ "eval_loss": 4.661567687988281,
+ "eval_runtime": 57.1244,
+ "eval_samples_per_second": 42.749,
+ "eval_steps_per_second": 1.348,
+ "step": 2088
+ },
+ {
+ "epoch": 29.01397379912664,
+ "grad_norm": 0.01665342226624489,
+ "learning_rate": 0.0006,
+ "loss": 4.311391830444336,
+ "step": 2089
+ },
+ {
+ "epoch": 29.027947598253274,
+ "grad_norm": 0.018053214997053146,
+ "learning_rate": 0.0006,
+ "loss": 4.135597229003906,
+ "step": 2090
+ },
+ {
+ "epoch": 29.041921397379912,
+ "grad_norm": 0.018213748931884766,
+ "learning_rate": 0.0006,
+ "loss": 4.303299903869629,
+ "step": 2091
+ },
+ {
+ "epoch": 29.05589519650655,
+ "grad_norm": 0.018460242077708244,
+ "learning_rate": 0.0006,
+ "loss": 4.20850944519043,
+ "step": 2092
+ },
+ {
+ "epoch": 29.069868995633186,
+ "grad_norm": 0.018365809693932533,
+ "learning_rate": 0.0006,
+ "loss": 4.275847434997559,
+ "step": 2093
+ },
+ {
+ "epoch": 29.083842794759825,
+ "grad_norm": 0.018346186727285385,
+ "learning_rate": 0.0006,
+ "loss": 4.202091217041016,
+ "step": 2094
+ },
+ {
+ "epoch": 29.097816593886463,
+ "grad_norm": 0.017920587211847305,
+ "learning_rate": 0.0006,
+ "loss": 4.182092666625977,
+ "step": 2095
+ },
+ {
+ "epoch": 29.111790393013102,
+ "grad_norm": 0.01812003180384636,
+ "learning_rate": 0.0006,
+ "loss": 4.259721755981445,
+ "step": 2096
+ },
+ {
+ "epoch": 29.125764192139737,
+ "grad_norm": 0.01700986735522747,
+ "learning_rate": 0.0006,
+ "loss": 4.067573070526123,
+ "step": 2097
+ },
+ {
+ "epoch": 29.139737991266376,
+ "grad_norm": 0.018069760873913765,
+ "learning_rate": 0.0006,
+ "loss": 4.234400749206543,
+ "step": 2098
+ },
+ {
+ "epoch": 29.153711790393015,
+ "grad_norm": 0.01905428245663643,
+ "learning_rate": 0.0006,
+ "loss": 4.16609001159668,
+ "step": 2099
+ },
+ {
+ "epoch": 29.16768558951965,
+ "grad_norm": 0.018711242824792862,
+ "learning_rate": 0.0006,
+ "loss": 4.249088764190674,
+ "step": 2100
+ },
+ {
+ "epoch": 29.18165938864629,
+ "grad_norm": 0.018721306696534157,
+ "learning_rate": 0.0006,
+ "loss": 4.14540958404541,
+ "step": 2101
+ },
+ {
+ "epoch": 29.195633187772927,
+ "grad_norm": 0.020139874890446663,
+ "learning_rate": 0.0006,
+ "loss": 4.049248695373535,
+ "step": 2102
+ },
+ {
+ "epoch": 29.209606986899562,
+ "grad_norm": 0.020849574357271194,
+ "learning_rate": 0.0006,
+ "loss": 4.235224723815918,
+ "step": 2103
+ },
+ {
+ "epoch": 29.2235807860262,
+ "grad_norm": 0.018693549558520317,
+ "learning_rate": 0.0006,
+ "loss": 4.232961177825928,
+ "step": 2104
+ },
+ {
+ "epoch": 29.23755458515284,
+ "grad_norm": 0.017889857292175293,
+ "learning_rate": 0.0006,
+ "loss": 4.171422004699707,
+ "step": 2105
+ },
+ {
+ "epoch": 29.251528384279474,
+ "grad_norm": 0.016690224409103394,
+ "learning_rate": 0.0006,
+ "loss": 4.244717597961426,
+ "step": 2106
+ },
+ {
+ "epoch": 29.265502183406113,
+ "grad_norm": 0.01708616502583027,
+ "learning_rate": 0.0006,
+ "loss": 4.125068664550781,
+ "step": 2107
+ },
+ {
+ "epoch": 29.27947598253275,
+ "grad_norm": 0.017928237095475197,
+ "learning_rate": 0.0006,
+ "loss": 4.208292007446289,
+ "step": 2108
+ },
+ {
+ "epoch": 29.29344978165939,
+ "grad_norm": 0.018250394612550735,
+ "learning_rate": 0.0006,
+ "loss": 4.118630886077881,
+ "step": 2109
+ },
+ {
+ "epoch": 29.307423580786025,
+ "grad_norm": 0.020160246640443802,
+ "learning_rate": 0.0006,
+ "loss": 4.206025123596191,
+ "step": 2110
+ },
+ {
+ "epoch": 29.321397379912664,
+ "grad_norm": 0.020045453682541847,
+ "learning_rate": 0.0006,
+ "loss": 4.2758684158325195,
+ "step": 2111
+ },
+ {
+ "epoch": 29.335371179039303,
+ "grad_norm": 0.019556907936930656,
+ "learning_rate": 0.0006,
+ "loss": 4.188272953033447,
+ "step": 2112
+ },
+ {
+ "epoch": 29.349344978165938,
+ "grad_norm": 0.019185200333595276,
+ "learning_rate": 0.0006,
+ "loss": 4.17451286315918,
+ "step": 2113
+ },
+ {
+ "epoch": 29.363318777292577,
+ "grad_norm": 0.01816629432141781,
+ "learning_rate": 0.0006,
+ "loss": 4.165246963500977,
+ "step": 2114
+ },
+ {
+ "epoch": 29.377292576419215,
+ "grad_norm": 0.01865146867930889,
+ "learning_rate": 0.0006,
+ "loss": 4.148184776306152,
+ "step": 2115
+ },
+ {
+ "epoch": 29.39126637554585,
+ "grad_norm": 0.018338464200496674,
+ "learning_rate": 0.0006,
+ "loss": 4.288936614990234,
+ "step": 2116
+ },
+ {
+ "epoch": 29.40524017467249,
+ "grad_norm": 0.018986187875270844,
+ "learning_rate": 0.0006,
+ "loss": 4.183379650115967,
+ "step": 2117
+ },
+ {
+ "epoch": 29.419213973799128,
+ "grad_norm": 0.018409637734293938,
+ "learning_rate": 0.0006,
+ "loss": 4.057574272155762,
+ "step": 2118
+ },
+ {
+ "epoch": 29.433187772925763,
+ "grad_norm": 0.016664542257785797,
+ "learning_rate": 0.0006,
+ "loss": 4.18517541885376,
+ "step": 2119
+ },
+ {
+ "epoch": 29.4471615720524,
+ "grad_norm": 0.016486089676618576,
+ "learning_rate": 0.0006,
+ "loss": 4.179367542266846,
+ "step": 2120
+ },
+ {
+ "epoch": 29.46113537117904,
+ "grad_norm": 0.01787523552775383,
+ "learning_rate": 0.0006,
+ "loss": 4.184126377105713,
+ "step": 2121
+ },
+ {
+ "epoch": 29.475109170305675,
+ "grad_norm": 0.016574613749980927,
+ "learning_rate": 0.0006,
+ "loss": 4.290170192718506,
+ "step": 2122
+ },
+ {
+ "epoch": 29.489082969432314,
+ "grad_norm": 0.01643376611173153,
+ "learning_rate": 0.0006,
+ "loss": 4.052913665771484,
+ "step": 2123
+ },
+ {
+ "epoch": 29.503056768558952,
+ "grad_norm": 0.016503969207406044,
+ "learning_rate": 0.0006,
+ "loss": 4.335184097290039,
+ "step": 2124
+ },
+ {
+ "epoch": 29.51703056768559,
+ "grad_norm": 0.017596479505300522,
+ "learning_rate": 0.0006,
+ "loss": 4.165539741516113,
+ "step": 2125
+ },
+ {
+ "epoch": 29.531004366812226,
+ "grad_norm": 0.017468582838773727,
+ "learning_rate": 0.0006,
+ "loss": 4.315964698791504,
+ "step": 2126
+ },
+ {
+ "epoch": 29.544978165938865,
+ "grad_norm": 0.01724618673324585,
+ "learning_rate": 0.0006,
+ "loss": 4.243681907653809,
+ "step": 2127
+ },
+ {
+ "epoch": 29.558951965065503,
+ "grad_norm": 0.018384616822004318,
+ "learning_rate": 0.0006,
+ "loss": 4.112447261810303,
+ "step": 2128
+ },
+ {
+ "epoch": 29.57292576419214,
+ "grad_norm": 0.01824074238538742,
+ "learning_rate": 0.0006,
+ "loss": 4.236065864562988,
+ "step": 2129
+ },
+ {
+ "epoch": 29.586899563318777,
+ "grad_norm": 0.01625971868634224,
+ "learning_rate": 0.0006,
+ "loss": 4.18143892288208,
+ "step": 2130
+ },
+ {
+ "epoch": 29.600873362445416,
+ "grad_norm": 0.015537494793534279,
+ "learning_rate": 0.0006,
+ "loss": 4.225147247314453,
+ "step": 2131
+ },
+ {
+ "epoch": 29.61484716157205,
+ "grad_norm": 0.015818974003195763,
+ "learning_rate": 0.0006,
+ "loss": 4.3100175857543945,
+ "step": 2132
+ },
+ {
+ "epoch": 29.62882096069869,
+ "grad_norm": 0.015909671783447266,
+ "learning_rate": 0.0006,
+ "loss": 4.169775009155273,
+ "step": 2133
+ },
+ {
+ "epoch": 29.64279475982533,
+ "grad_norm": 0.01769658550620079,
+ "learning_rate": 0.0006,
+ "loss": 4.278904914855957,
+ "step": 2134
+ },
+ {
+ "epoch": 29.656768558951963,
+ "grad_norm": 0.015641704201698303,
+ "learning_rate": 0.0006,
+ "loss": 4.2426228523254395,
+ "step": 2135
+ },
+ {
+ "epoch": 29.670742358078602,
+ "grad_norm": 0.01565658301115036,
+ "learning_rate": 0.0006,
+ "loss": 4.232687950134277,
+ "step": 2136
+ },
+ {
+ "epoch": 29.68471615720524,
+ "grad_norm": 0.016485940665006638,
+ "learning_rate": 0.0006,
+ "loss": 4.209839820861816,
+ "step": 2137
+ },
+ {
+ "epoch": 29.69868995633188,
+ "grad_norm": 0.017622729763388634,
+ "learning_rate": 0.0006,
+ "loss": 4.196017265319824,
+ "step": 2138
+ },
+ {
+ "epoch": 29.712663755458514,
+ "grad_norm": 0.018320564180612564,
+ "learning_rate": 0.0006,
+ "loss": 4.225683689117432,
+ "step": 2139
+ },
+ {
+ "epoch": 29.726637554585153,
+ "grad_norm": 0.018140794709324837,
+ "learning_rate": 0.0006,
+ "loss": 4.311944961547852,
+ "step": 2140
+ },
+ {
+ "epoch": 29.74061135371179,
+ "grad_norm": 0.016227100044488907,
+ "learning_rate": 0.0006,
+ "loss": 4.230257511138916,
+ "step": 2141
+ },
+ {
+ "epoch": 29.754585152838427,
+ "grad_norm": 0.015759488567709923,
+ "learning_rate": 0.0006,
+ "loss": 4.119174003601074,
+ "step": 2142
+ },
+ {
+ "epoch": 29.768558951965066,
+ "grad_norm": 0.01757766678929329,
+ "learning_rate": 0.0006,
+ "loss": 4.165826797485352,
+ "step": 2143
+ },
+ {
+ "epoch": 29.782532751091704,
+ "grad_norm": 0.019661923870444298,
+ "learning_rate": 0.0006,
+ "loss": 4.392203330993652,
+ "step": 2144
+ },
+ {
+ "epoch": 29.79650655021834,
+ "grad_norm": 0.019478755071759224,
+ "learning_rate": 0.0006,
+ "loss": 4.348371982574463,
+ "step": 2145
+ },
+ {
+ "epoch": 29.810480349344978,
+ "grad_norm": 0.019149569794535637,
+ "learning_rate": 0.0006,
+ "loss": 4.34104061126709,
+ "step": 2146
+ },
+ {
+ "epoch": 29.824454148471617,
+ "grad_norm": 0.017707521095871925,
+ "learning_rate": 0.0006,
+ "loss": 4.228226661682129,
+ "step": 2147
+ },
+ {
+ "epoch": 29.83842794759825,
+ "grad_norm": 0.015615378506481647,
+ "learning_rate": 0.0006,
+ "loss": 4.175541877746582,
+ "step": 2148
+ },
+ {
+ "epoch": 29.85240174672489,
+ "grad_norm": 0.016111129894852638,
+ "learning_rate": 0.0006,
+ "loss": 4.267936706542969,
+ "step": 2149
+ },
+ {
+ "epoch": 29.86637554585153,
+ "grad_norm": 0.01528779324144125,
+ "learning_rate": 0.0006,
+ "loss": 4.186014175415039,
+ "step": 2150
+ },
+ {
+ "epoch": 29.880349344978168,
+ "grad_norm": 0.014925646595656872,
+ "learning_rate": 0.0006,
+ "loss": 4.174860000610352,
+ "step": 2151
+ },
+ {
+ "epoch": 29.894323144104803,
+ "grad_norm": 0.016917673870921135,
+ "learning_rate": 0.0006,
+ "loss": 4.33908748626709,
+ "step": 2152
+ },
+ {
+ "epoch": 29.90829694323144,
+ "grad_norm": 0.016007075086236,
+ "learning_rate": 0.0006,
+ "loss": 4.266141414642334,
+ "step": 2153
+ },
+ {
+ "epoch": 29.92227074235808,
+ "grad_norm": 0.016178956255316734,
+ "learning_rate": 0.0006,
+ "loss": 4.271215915679932,
+ "step": 2154
+ },
+ {
+ "epoch": 29.936244541484715,
+ "grad_norm": 0.016180871054530144,
+ "learning_rate": 0.0006,
+ "loss": 4.278801441192627,
+ "step": 2155
+ },
+ {
+ "epoch": 29.950218340611354,
+ "grad_norm": 0.015335503034293652,
+ "learning_rate": 0.0006,
+ "loss": 4.194978713989258,
+ "step": 2156
+ },
+ {
+ "epoch": 29.964192139737992,
+ "grad_norm": 0.016580764204263687,
+ "learning_rate": 0.0006,
+ "loss": 4.290809631347656,
+ "step": 2157
+ },
+ {
+ "epoch": 29.978165938864628,
+ "grad_norm": 0.01623128168284893,
+ "learning_rate": 0.0006,
+ "loss": 4.205451488494873,
+ "step": 2158
+ },
+ {
+ "epoch": 29.992139737991266,
+ "grad_norm": 0.014904248528182507,
+ "learning_rate": 0.0006,
+ "loss": 4.24924373626709,
+ "step": 2159
+ },
+ {
+ "epoch": 30.0,
+ "grad_norm": 0.017089299857616425,
+ "learning_rate": 0.0006,
+ "loss": 4.222072601318359,
+ "step": 2160
+ },
+ {
+ "epoch": 30.0,
+ "eval_loss": 4.585411071777344,
+ "eval_runtime": 57.1164,
+ "eval_samples_per_second": 42.755,
+ "eval_steps_per_second": 1.348,
+ "step": 2160
+ },
+ {
+ "epoch": 30.01397379912664,
+ "grad_norm": 0.017226383090019226,
+ "learning_rate": 0.0006,
+ "loss": 4.19467830657959,
+ "step": 2161
+ },
+ {
+ "epoch": 30.027947598253274,
+ "grad_norm": 0.01635066606104374,
+ "learning_rate": 0.0006,
+ "loss": 4.176875591278076,
+ "step": 2162
+ },
+ {
+ "epoch": 30.041921397379912,
+ "grad_norm": 0.01735362410545349,
+ "learning_rate": 0.0006,
+ "loss": 4.13667106628418,
+ "step": 2163
+ },
+ {
+ "epoch": 30.05589519650655,
+ "grad_norm": 0.017047109082341194,
+ "learning_rate": 0.0006,
+ "loss": 4.136329650878906,
+ "step": 2164
+ },
+ {
+ "epoch": 30.069868995633186,
+ "grad_norm": 0.018116053193807602,
+ "learning_rate": 0.0006,
+ "loss": 4.228706359863281,
+ "step": 2165
+ },
+ {
+ "epoch": 30.083842794759825,
+ "grad_norm": 0.01791389286518097,
+ "learning_rate": 0.0006,
+ "loss": 4.1597747802734375,
+ "step": 2166
+ },
+ {
+ "epoch": 30.097816593886463,
+ "grad_norm": 0.017793502658605576,
+ "learning_rate": 0.0006,
+ "loss": 4.0807976722717285,
+ "step": 2167
+ },
+ {
+ "epoch": 30.111790393013102,
+ "grad_norm": 0.01827503927052021,
+ "learning_rate": 0.0006,
+ "loss": 4.170646667480469,
+ "step": 2168
+ },
+ {
+ "epoch": 30.125764192139737,
+ "grad_norm": 0.017228564247488976,
+ "learning_rate": 0.0006,
+ "loss": 4.146934509277344,
+ "step": 2169
+ },
+ {
+ "epoch": 30.139737991266376,
+ "grad_norm": 0.01757437363266945,
+ "learning_rate": 0.0006,
+ "loss": 4.1859130859375,
+ "step": 2170
+ },
+ {
+ "epoch": 30.153711790393015,
+ "grad_norm": 0.01701093092560768,
+ "learning_rate": 0.0006,
+ "loss": 4.126981258392334,
+ "step": 2171
+ },
+ {
+ "epoch": 30.16768558951965,
+ "grad_norm": 0.015220489352941513,
+ "learning_rate": 0.0006,
+ "loss": 4.164018630981445,
+ "step": 2172
+ },
+ {
+ "epoch": 30.18165938864629,
+ "grad_norm": 0.016254238784313202,
+ "learning_rate": 0.0006,
+ "loss": 4.258275508880615,
+ "step": 2173
+ },
+ {
+ "epoch": 30.195633187772927,
+ "grad_norm": 0.016609614714980125,
+ "learning_rate": 0.0006,
+ "loss": 4.222233772277832,
+ "step": 2174
+ },
+ {
+ "epoch": 30.209606986899562,
+ "grad_norm": 0.016531318426132202,
+ "learning_rate": 0.0006,
+ "loss": 4.144489288330078,
+ "step": 2175
+ },
+ {
+ "epoch": 30.2235807860262,
+ "grad_norm": 0.01567668654024601,
+ "learning_rate": 0.0006,
+ "loss": 4.181916236877441,
+ "step": 2176
+ },
+ {
+ "epoch": 30.23755458515284,
+ "grad_norm": 0.017195681110024452,
+ "learning_rate": 0.0006,
+ "loss": 4.189693927764893,
+ "step": 2177
+ },
+ {
+ "epoch": 30.251528384279474,
+ "grad_norm": 0.016882948577404022,
+ "learning_rate": 0.0006,
+ "loss": 4.21604061126709,
+ "step": 2178
+ },
+ {
+ "epoch": 30.265502183406113,
+ "grad_norm": 0.016261622309684753,
+ "learning_rate": 0.0006,
+ "loss": 4.227639198303223,
+ "step": 2179
+ },
+ {
+ "epoch": 30.27947598253275,
+ "grad_norm": 0.016411812976002693,
+ "learning_rate": 0.0006,
+ "loss": 4.130362510681152,
+ "step": 2180
+ },
+ {
+ "epoch": 30.29344978165939,
+ "grad_norm": 0.01732019893825054,
+ "learning_rate": 0.0006,
+ "loss": 4.230443000793457,
+ "step": 2181
+ },
+ {
+ "epoch": 30.307423580786025,
+ "grad_norm": 0.01825905591249466,
+ "learning_rate": 0.0006,
+ "loss": 4.175817489624023,
+ "step": 2182
+ },
+ {
+ "epoch": 30.321397379912664,
+ "grad_norm": 0.01887713558971882,
+ "learning_rate": 0.0006,
+ "loss": 4.169020652770996,
+ "step": 2183
+ },
+ {
+ "epoch": 30.335371179039303,
+ "grad_norm": 0.019391661509871483,
+ "learning_rate": 0.0006,
+ "loss": 4.176318645477295,
+ "step": 2184
+ },
+ {
+ "epoch": 30.349344978165938,
+ "grad_norm": 0.020840002223849297,
+ "learning_rate": 0.0006,
+ "loss": 4.174228668212891,
+ "step": 2185
+ },
+ {
+ "epoch": 30.363318777292577,
+ "grad_norm": 0.020193178206682205,
+ "learning_rate": 0.0006,
+ "loss": 4.127919673919678,
+ "step": 2186
+ },
+ {
+ "epoch": 30.377292576419215,
+ "grad_norm": 0.018449561670422554,
+ "learning_rate": 0.0006,
+ "loss": 4.229487895965576,
+ "step": 2187
+ },
+ {
+ "epoch": 30.39126637554585,
+ "grad_norm": 0.01711983233690262,
+ "learning_rate": 0.0006,
+ "loss": 4.201651573181152,
+ "step": 2188
+ },
+ {
+ "epoch": 30.40524017467249,
+ "grad_norm": 0.017293395474553108,
+ "learning_rate": 0.0006,
+ "loss": 4.168704032897949,
+ "step": 2189
+ },
+ {
+ "epoch": 30.419213973799128,
+ "grad_norm": 0.018526358529925346,
+ "learning_rate": 0.0006,
+ "loss": 4.171864032745361,
+ "step": 2190
+ },
+ {
+ "epoch": 30.433187772925763,
+ "grad_norm": 0.019287291914224625,
+ "learning_rate": 0.0006,
+ "loss": 4.039534568786621,
+ "step": 2191
+ },
+ {
+ "epoch": 30.4471615720524,
+ "grad_norm": 0.016902444884181023,
+ "learning_rate": 0.0006,
+ "loss": 4.282787799835205,
+ "step": 2192
+ },
+ {
+ "epoch": 30.46113537117904,
+ "grad_norm": 0.01656750962138176,
+ "learning_rate": 0.0006,
+ "loss": 4.213098526000977,
+ "step": 2193
+ },
+ {
+ "epoch": 30.475109170305675,
+ "grad_norm": 0.016805065795779228,
+ "learning_rate": 0.0006,
+ "loss": 4.1583476066589355,
+ "step": 2194
+ },
+ {
+ "epoch": 30.489082969432314,
+ "grad_norm": 0.017166638746857643,
+ "learning_rate": 0.0006,
+ "loss": 4.188634395599365,
+ "step": 2195
+ },
+ {
+ "epoch": 30.503056768558952,
+ "grad_norm": 0.01592106744647026,
+ "learning_rate": 0.0006,
+ "loss": 4.234043121337891,
+ "step": 2196
+ },
+ {
+ "epoch": 30.51703056768559,
+ "grad_norm": 0.015270989388227463,
+ "learning_rate": 0.0006,
+ "loss": 4.301420211791992,
+ "step": 2197
+ },
+ {
+ "epoch": 30.531004366812226,
+ "grad_norm": 0.015194724313914776,
+ "learning_rate": 0.0006,
+ "loss": 4.18703031539917,
+ "step": 2198
+ },
+ {
+ "epoch": 30.544978165938865,
+ "grad_norm": 0.015587746165692806,
+ "learning_rate": 0.0006,
+ "loss": 4.138928413391113,
+ "step": 2199
+ },
+ {
+ "epoch": 30.558951965065503,
+ "grad_norm": 0.016029570251703262,
+ "learning_rate": 0.0006,
+ "loss": 4.290863990783691,
+ "step": 2200
+ },
+ {
+ "epoch": 30.57292576419214,
+ "grad_norm": 0.015651309862732887,
+ "learning_rate": 0.0006,
+ "loss": 4.272062301635742,
+ "step": 2201
+ },
+ {
+ "epoch": 30.586899563318777,
+ "grad_norm": 0.015992306172847748,
+ "learning_rate": 0.0006,
+ "loss": 4.2487406730651855,
+ "step": 2202
+ },
+ {
+ "epoch": 30.600873362445416,
+ "grad_norm": 0.014969157986342907,
+ "learning_rate": 0.0006,
+ "loss": 4.115347862243652,
+ "step": 2203
+ },
+ {
+ "epoch": 30.61484716157205,
+ "grad_norm": 0.014639027416706085,
+ "learning_rate": 0.0006,
+ "loss": 4.2207489013671875,
+ "step": 2204
+ },
+ {
+ "epoch": 30.62882096069869,
+ "grad_norm": 0.015059413388371468,
+ "learning_rate": 0.0006,
+ "loss": 4.274044990539551,
+ "step": 2205
+ },
+ {
+ "epoch": 30.64279475982533,
+ "grad_norm": 0.0166641678661108,
+ "learning_rate": 0.0006,
+ "loss": 4.191373348236084,
+ "step": 2206
+ },
+ {
+ "epoch": 30.656768558951963,
+ "grad_norm": 0.017806345596909523,
+ "learning_rate": 0.0006,
+ "loss": 4.139800548553467,
+ "step": 2207
+ },
+ {
+ "epoch": 30.670742358078602,
+ "grad_norm": 0.018712421879172325,
+ "learning_rate": 0.0006,
+ "loss": 4.035965442657471,
+ "step": 2208
+ },
+ {
+ "epoch": 30.68471615720524,
+ "grad_norm": 0.018457243219017982,
+ "learning_rate": 0.0006,
+ "loss": 4.2579240798950195,
+ "step": 2209
+ },
+ {
+ "epoch": 30.69868995633188,
+ "grad_norm": 0.01920999586582184,
+ "learning_rate": 0.0006,
+ "loss": 4.131112098693848,
+ "step": 2210
+ },
+ {
+ "epoch": 30.712663755458514,
+ "grad_norm": 0.01729944534599781,
+ "learning_rate": 0.0006,
+ "loss": 4.254059314727783,
+ "step": 2211
+ },
+ {
+ "epoch": 30.726637554585153,
+ "grad_norm": 0.016914231702685356,
+ "learning_rate": 0.0006,
+ "loss": 4.233428955078125,
+ "step": 2212
+ },
+ {
+ "epoch": 30.74061135371179,
+ "grad_norm": 0.01913098618388176,
+ "learning_rate": 0.0006,
+ "loss": 4.23525333404541,
+ "step": 2213
+ },
+ {
+ "epoch": 30.754585152838427,
+ "grad_norm": 0.019150281324982643,
+ "learning_rate": 0.0006,
+ "loss": 4.213695049285889,
+ "step": 2214
+ },
+ {
+ "epoch": 30.768558951965066,
+ "grad_norm": 0.018947452306747437,
+ "learning_rate": 0.0006,
+ "loss": 4.151697158813477,
+ "step": 2215
+ },
+ {
+ "epoch": 30.782532751091704,
+ "grad_norm": 0.016058508306741714,
+ "learning_rate": 0.0006,
+ "loss": 4.220280170440674,
+ "step": 2216
+ },
+ {
+ "epoch": 30.79650655021834,
+ "grad_norm": 0.017488619312644005,
+ "learning_rate": 0.0006,
+ "loss": 4.154428482055664,
+ "step": 2217
+ },
+ {
+ "epoch": 30.810480349344978,
+ "grad_norm": 0.017355090007185936,
+ "learning_rate": 0.0006,
+ "loss": 4.256086349487305,
+ "step": 2218
+ },
+ {
+ "epoch": 30.824454148471617,
+ "grad_norm": 0.016636032611131668,
+ "learning_rate": 0.0006,
+ "loss": 4.227564334869385,
+ "step": 2219
+ },
+ {
+ "epoch": 30.83842794759825,
+ "grad_norm": 0.016273891553282738,
+ "learning_rate": 0.0006,
+ "loss": 4.200466156005859,
+ "step": 2220
+ },
+ {
+ "epoch": 30.85240174672489,
+ "grad_norm": 0.016939911991357803,
+ "learning_rate": 0.0006,
+ "loss": 4.1571221351623535,
+ "step": 2221
+ },
+ {
+ "epoch": 30.86637554585153,
+ "grad_norm": 0.018332522362470627,
+ "learning_rate": 0.0006,
+ "loss": 4.228398323059082,
+ "step": 2222
+ },
+ {
+ "epoch": 30.880349344978168,
+ "grad_norm": 0.019508114084601402,
+ "learning_rate": 0.0006,
+ "loss": 4.222960472106934,
+ "step": 2223
+ },
+ {
+ "epoch": 30.894323144104803,
+ "grad_norm": 0.018440278246998787,
+ "learning_rate": 0.0006,
+ "loss": 4.261051177978516,
+ "step": 2224
+ },
+ {
+ "epoch": 30.90829694323144,
+ "grad_norm": 0.018113229423761368,
+ "learning_rate": 0.0006,
+ "loss": 4.109235763549805,
+ "step": 2225
+ },
+ {
+ "epoch": 30.92227074235808,
+ "grad_norm": 0.015392083674669266,
+ "learning_rate": 0.0006,
+ "loss": 4.21950101852417,
+ "step": 2226
+ },
+ {
+ "epoch": 30.936244541484715,
+ "grad_norm": 0.01633290946483612,
+ "learning_rate": 0.0006,
+ "loss": 4.176053047180176,
+ "step": 2227
+ },
+ {
+ "epoch": 30.950218340611354,
+ "grad_norm": 0.016236383467912674,
+ "learning_rate": 0.0006,
+ "loss": 4.314309120178223,
+ "step": 2228
+ },
+ {
+ "epoch": 30.964192139737992,
+ "grad_norm": 0.015284373424947262,
+ "learning_rate": 0.0006,
+ "loss": 4.303133487701416,
+ "step": 2229
+ },
+ {
+ "epoch": 30.978165938864628,
+ "grad_norm": 0.015982897952198982,
+ "learning_rate": 0.0006,
+ "loss": 4.25739860534668,
+ "step": 2230
+ },
+ {
+ "epoch": 30.992139737991266,
+ "grad_norm": 0.016682198271155357,
+ "learning_rate": 0.0006,
+ "loss": 4.292210578918457,
+ "step": 2231
+ },
+ {
+ "epoch": 31.0,
+ "grad_norm": 0.017521467059850693,
+ "learning_rate": 0.0006,
+ "loss": 4.170637130737305,
+ "step": 2232
+ },
+ {
+ "epoch": 31.0,
+ "eval_loss": 4.642786979675293,
+ "eval_runtime": 57.029,
+ "eval_samples_per_second": 42.82,
+ "eval_steps_per_second": 1.35,
+ "step": 2232
+ },
+ {
+ "epoch": 31.01397379912664,
+ "grad_norm": 0.016607852652668953,
+ "learning_rate": 0.0006,
+ "loss": 4.106224536895752,
+ "step": 2233
+ },
+ {
+ "epoch": 31.027947598253274,
+ "grad_norm": 0.018448904156684875,
+ "learning_rate": 0.0006,
+ "loss": 4.224459648132324,
+ "step": 2234
+ },
+ {
+ "epoch": 31.041921397379912,
+ "grad_norm": 0.020481228828430176,
+ "learning_rate": 0.0006,
+ "loss": 4.305548667907715,
+ "step": 2235
+ },
+ {
+ "epoch": 31.05589519650655,
+ "grad_norm": 0.01999707892537117,
+ "learning_rate": 0.0006,
+ "loss": 4.253663539886475,
+ "step": 2236
+ },
+ {
+ "epoch": 31.069868995633186,
+ "grad_norm": 0.020428303629159927,
+ "learning_rate": 0.0006,
+ "loss": 4.166954040527344,
+ "step": 2237
+ },
+ {
+ "epoch": 31.083842794759825,
+ "grad_norm": 0.022608213126659393,
+ "learning_rate": 0.0006,
+ "loss": 4.172819137573242,
+ "step": 2238
+ },
+ {
+ "epoch": 31.097816593886463,
+ "grad_norm": 0.024669520556926727,
+ "learning_rate": 0.0006,
+ "loss": 4.106586456298828,
+ "step": 2239
+ },
+ {
+ "epoch": 31.111790393013102,
+ "grad_norm": 0.02247351035475731,
+ "learning_rate": 0.0006,
+ "loss": 4.160271644592285,
+ "step": 2240
+ },
+ {
+ "epoch": 31.125764192139737,
+ "grad_norm": 0.0207810141146183,
+ "learning_rate": 0.0006,
+ "loss": 4.272491931915283,
+ "step": 2241
+ },
+ {
+ "epoch": 31.139737991266376,
+ "grad_norm": 0.021576499566435814,
+ "learning_rate": 0.0006,
+ "loss": 4.171474456787109,
+ "step": 2242
+ },
+ {
+ "epoch": 31.153711790393015,
+ "grad_norm": 0.02337106689810753,
+ "learning_rate": 0.0006,
+ "loss": 4.094712734222412,
+ "step": 2243
+ },
+ {
+ "epoch": 31.16768558951965,
+ "grad_norm": 0.02219875156879425,
+ "learning_rate": 0.0006,
+ "loss": 4.1800994873046875,
+ "step": 2244
+ },
+ {
+ "epoch": 31.18165938864629,
+ "grad_norm": 0.022680504247546196,
+ "learning_rate": 0.0006,
+ "loss": 4.2054362297058105,
+ "step": 2245
+ },
+ {
+ "epoch": 31.195633187772927,
+ "grad_norm": 0.0208734143525362,
+ "learning_rate": 0.0006,
+ "loss": 4.06068229675293,
+ "step": 2246
+ },
+ {
+ "epoch": 31.209606986899562,
+ "grad_norm": 0.01811066083610058,
+ "learning_rate": 0.0006,
+ "loss": 4.2465128898620605,
+ "step": 2247
+ },
+ {
+ "epoch": 31.2235807860262,
+ "grad_norm": 0.018571637570858,
+ "learning_rate": 0.0006,
+ "loss": 4.191955089569092,
+ "step": 2248
+ },
+ {
+ "epoch": 31.23755458515284,
+ "grad_norm": 0.018598807975649834,
+ "learning_rate": 0.0006,
+ "loss": 4.200872421264648,
+ "step": 2249
+ },
+ {
+ "epoch": 31.251528384279474,
+ "grad_norm": 0.017577625811100006,
+ "learning_rate": 0.0006,
+ "loss": 4.148700714111328,
+ "step": 2250
+ },
+ {
+ "epoch": 31.265502183406113,
+ "grad_norm": 0.01817980408668518,
+ "learning_rate": 0.0006,
+ "loss": 4.256978988647461,
+ "step": 2251
+ },
+ {
+ "epoch": 31.27947598253275,
+ "grad_norm": 0.017983486875891685,
+ "learning_rate": 0.0006,
+ "loss": 4.175005912780762,
+ "step": 2252
+ },
+ {
+ "epoch": 31.29344978165939,
+ "grad_norm": 0.016751961782574654,
+ "learning_rate": 0.0006,
+ "loss": 4.237518310546875,
+ "step": 2253
+ },
+ {
+ "epoch": 31.307423580786025,
+ "grad_norm": 0.016407202929258347,
+ "learning_rate": 0.0006,
+ "loss": 4.197269439697266,
+ "step": 2254
+ },
+ {
+ "epoch": 31.321397379912664,
+ "grad_norm": 0.01578645594418049,
+ "learning_rate": 0.0006,
+ "loss": 4.179751396179199,
+ "step": 2255
+ },
+ {
+ "epoch": 31.335371179039303,
+ "grad_norm": 0.016001226380467415,
+ "learning_rate": 0.0006,
+ "loss": 4.08821964263916,
+ "step": 2256
+ },
+ {
+ "epoch": 31.349344978165938,
+ "grad_norm": 0.015122811309993267,
+ "learning_rate": 0.0006,
+ "loss": 4.047828674316406,
+ "step": 2257
+ },
+ {
+ "epoch": 31.363318777292577,
+ "grad_norm": 0.015399965457618237,
+ "learning_rate": 0.0006,
+ "loss": 4.1506524085998535,
+ "step": 2258
+ },
+ {
+ "epoch": 31.377292576419215,
+ "grad_norm": 0.01534605398774147,
+ "learning_rate": 0.0006,
+ "loss": 4.27952766418457,
+ "step": 2259
+ },
+ {
+ "epoch": 31.39126637554585,
+ "grad_norm": 0.015461381524801254,
+ "learning_rate": 0.0006,
+ "loss": 4.27102518081665,
+ "step": 2260
+ },
+ {
+ "epoch": 31.40524017467249,
+ "grad_norm": 0.016769153997302055,
+ "learning_rate": 0.0006,
+ "loss": 4.191592216491699,
+ "step": 2261
+ },
+ {
+ "epoch": 31.419213973799128,
+ "grad_norm": 0.016557959839701653,
+ "learning_rate": 0.0006,
+ "loss": 4.159705638885498,
+ "step": 2262
+ },
+ {
+ "epoch": 31.433187772925763,
+ "grad_norm": 0.01659569889307022,
+ "learning_rate": 0.0006,
+ "loss": 4.2136664390563965,
+ "step": 2263
+ },
+ {
+ "epoch": 31.4471615720524,
+ "grad_norm": 0.017101937904953957,
+ "learning_rate": 0.0006,
+ "loss": 4.2716498374938965,
+ "step": 2264
+ },
+ {
+ "epoch": 31.46113537117904,
+ "grad_norm": 0.01722985878586769,
+ "learning_rate": 0.0006,
+ "loss": 4.300534248352051,
+ "step": 2265
+ },
+ {
+ "epoch": 31.475109170305675,
+ "grad_norm": 0.016800185665488243,
+ "learning_rate": 0.0006,
+ "loss": 4.113240718841553,
+ "step": 2266
+ },
+ {
+ "epoch": 31.489082969432314,
+ "grad_norm": 0.01603875495493412,
+ "learning_rate": 0.0006,
+ "loss": 4.157105445861816,
+ "step": 2267
+ },
+ {
+ "epoch": 31.503056768558952,
+ "grad_norm": 0.015903932973742485,
+ "learning_rate": 0.0006,
+ "loss": 4.233669757843018,
+ "step": 2268
+ },
+ {
+ "epoch": 31.51703056768559,
+ "grad_norm": 0.0160983856767416,
+ "learning_rate": 0.0006,
+ "loss": 4.147704124450684,
+ "step": 2269
+ },
+ {
+ "epoch": 31.531004366812226,
+ "grad_norm": 0.01604730263352394,
+ "learning_rate": 0.0006,
+ "loss": 4.1465911865234375,
+ "step": 2270
+ },
+ {
+ "epoch": 31.544978165938865,
+ "grad_norm": 0.016205785796046257,
+ "learning_rate": 0.0006,
+ "loss": 4.106704235076904,
+ "step": 2271
+ },
+ {
+ "epoch": 31.558951965065503,
+ "grad_norm": 0.017008036375045776,
+ "learning_rate": 0.0006,
+ "loss": 4.191000938415527,
+ "step": 2272
+ },
+ {
+ "epoch": 31.57292576419214,
+ "grad_norm": 0.01699635200202465,
+ "learning_rate": 0.0006,
+ "loss": 4.187042236328125,
+ "step": 2273
+ },
+ {
+ "epoch": 31.586899563318777,
+ "grad_norm": 0.01692967116832733,
+ "learning_rate": 0.0006,
+ "loss": 4.208805561065674,
+ "step": 2274
+ },
+ {
+ "epoch": 31.600873362445416,
+ "grad_norm": 0.017608756199479103,
+ "learning_rate": 0.0006,
+ "loss": 4.200512886047363,
+ "step": 2275
+ },
+ {
+ "epoch": 31.61484716157205,
+ "grad_norm": 0.016590990126132965,
+ "learning_rate": 0.0006,
+ "loss": 4.194705009460449,
+ "step": 2276
+ },
+ {
+ "epoch": 31.62882096069869,
+ "grad_norm": 0.0170235987752676,
+ "learning_rate": 0.0006,
+ "loss": 4.210326194763184,
+ "step": 2277
+ },
+ {
+ "epoch": 31.64279475982533,
+ "grad_norm": 0.01604801043868065,
+ "learning_rate": 0.0006,
+ "loss": 4.196024417877197,
+ "step": 2278
+ },
+ {
+ "epoch": 31.656768558951963,
+ "grad_norm": 0.015694202855229378,
+ "learning_rate": 0.0006,
+ "loss": 4.178236484527588,
+ "step": 2279
+ },
+ {
+ "epoch": 31.670742358078602,
+ "grad_norm": 0.015473959036171436,
+ "learning_rate": 0.0006,
+ "loss": 4.259339809417725,
+ "step": 2280
+ },
+ {
+ "epoch": 31.68471615720524,
+ "grad_norm": 0.014689362607896328,
+ "learning_rate": 0.0006,
+ "loss": 3.9702324867248535,
+ "step": 2281
+ },
+ {
+ "epoch": 31.69868995633188,
+ "grad_norm": 0.015245123766362667,
+ "learning_rate": 0.0006,
+ "loss": 4.197025775909424,
+ "step": 2282
+ },
+ {
+ "epoch": 31.712663755458514,
+ "grad_norm": 0.016148347407579422,
+ "learning_rate": 0.0006,
+ "loss": 4.231856346130371,
+ "step": 2283
+ },
+ {
+ "epoch": 31.726637554585153,
+ "grad_norm": 0.015169818885624409,
+ "learning_rate": 0.0006,
+ "loss": 4.230169296264648,
+ "step": 2284
+ },
+ {
+ "epoch": 31.74061135371179,
+ "grad_norm": 0.014843013137578964,
+ "learning_rate": 0.0006,
+ "loss": 4.195633411407471,
+ "step": 2285
+ },
+ {
+ "epoch": 31.754585152838427,
+ "grad_norm": 0.015229340642690659,
+ "learning_rate": 0.0006,
+ "loss": 4.231280326843262,
+ "step": 2286
+ },
+ {
+ "epoch": 31.768558951965066,
+ "grad_norm": 0.01551708485931158,
+ "learning_rate": 0.0006,
+ "loss": 4.17064094543457,
+ "step": 2287
+ },
+ {
+ "epoch": 31.782532751091704,
+ "grad_norm": 0.015863575041294098,
+ "learning_rate": 0.0006,
+ "loss": 4.1579484939575195,
+ "step": 2288
+ },
+ {
+ "epoch": 31.79650655021834,
+ "grad_norm": 0.015118198469281197,
+ "learning_rate": 0.0006,
+ "loss": 4.1626787185668945,
+ "step": 2289
+ },
+ {
+ "epoch": 31.810480349344978,
+ "grad_norm": 0.015220776200294495,
+ "learning_rate": 0.0006,
+ "loss": 4.135746955871582,
+ "step": 2290
+ },
+ {
+ "epoch": 31.824454148471617,
+ "grad_norm": 0.014529784210026264,
+ "learning_rate": 0.0006,
+ "loss": 4.164931297302246,
+ "step": 2291
+ },
+ {
+ "epoch": 31.83842794759825,
+ "grad_norm": 0.015533296391367912,
+ "learning_rate": 0.0006,
+ "loss": 4.112464904785156,
+ "step": 2292
+ },
+ {
+ "epoch": 31.85240174672489,
+ "grad_norm": 0.016562335193157196,
+ "learning_rate": 0.0006,
+ "loss": 4.164898872375488,
+ "step": 2293
+ },
+ {
+ "epoch": 31.86637554585153,
+ "grad_norm": 0.016258113086223602,
+ "learning_rate": 0.0006,
+ "loss": 4.1979265213012695,
+ "step": 2294
+ },
+ {
+ "epoch": 31.880349344978168,
+ "grad_norm": 0.01558110024780035,
+ "learning_rate": 0.0006,
+ "loss": 4.174675941467285,
+ "step": 2295
+ },
+ {
+ "epoch": 31.894323144104803,
+ "grad_norm": 0.016412515193223953,
+ "learning_rate": 0.0006,
+ "loss": 4.217909336090088,
+ "step": 2296
+ },
+ {
+ "epoch": 31.90829694323144,
+ "grad_norm": 0.015118442475795746,
+ "learning_rate": 0.0006,
+ "loss": 4.137143611907959,
+ "step": 2297
+ },
+ {
+ "epoch": 31.92227074235808,
+ "grad_norm": 0.01506161317229271,
+ "learning_rate": 0.0006,
+ "loss": 4.196122169494629,
+ "step": 2298
+ },
+ {
+ "epoch": 31.936244541484715,
+ "grad_norm": 0.017165271565318108,
+ "learning_rate": 0.0006,
+ "loss": 4.14834451675415,
+ "step": 2299
+ },
+ {
+ "epoch": 31.950218340611354,
+ "grad_norm": 0.01696440577507019,
+ "learning_rate": 0.0006,
+ "loss": 4.218868732452393,
+ "step": 2300
+ },
+ {
+ "epoch": 31.964192139737992,
+ "grad_norm": 0.017349394038319588,
+ "learning_rate": 0.0006,
+ "loss": 4.2754011154174805,
+ "step": 2301
+ },
+ {
+ "epoch": 31.978165938864628,
+ "grad_norm": 0.018387921154499054,
+ "learning_rate": 0.0006,
+ "loss": 4.282623291015625,
+ "step": 2302
+ },
+ {
+ "epoch": 31.992139737991266,
+ "grad_norm": 0.018837926909327507,
+ "learning_rate": 0.0006,
+ "loss": 4.185425281524658,
+ "step": 2303
+ },
+ {
+ "epoch": 32.0,
+ "grad_norm": 0.018998464569449425,
+ "learning_rate": 0.0006,
+ "loss": 4.151291847229004,
+ "step": 2304
+ },
+ {
+ "epoch": 32.0,
+ "eval_loss": 4.522614002227783,
+ "eval_runtime": 57.0106,
+ "eval_samples_per_second": 42.834,
+ "eval_steps_per_second": 1.351,
+ "step": 2304
+ },
+ {
+ "epoch": 32.01397379912664,
+ "grad_norm": 0.019036833196878433,
+ "learning_rate": 0.0006,
+ "loss": 4.183202266693115,
+ "step": 2305
+ },
+ {
+ "epoch": 32.02794759825328,
+ "grad_norm": 0.02207397110760212,
+ "learning_rate": 0.0006,
+ "loss": 4.17875862121582,
+ "step": 2306
+ },
+ {
+ "epoch": 32.041921397379916,
+ "grad_norm": 0.022008279338479042,
+ "learning_rate": 0.0006,
+ "loss": 4.08042049407959,
+ "step": 2307
+ },
+ {
+ "epoch": 32.05589519650655,
+ "grad_norm": 0.022407017648220062,
+ "learning_rate": 0.0006,
+ "loss": 4.112189769744873,
+ "step": 2308
+ },
+ {
+ "epoch": 32.069868995633186,
+ "grad_norm": 0.021599747240543365,
+ "learning_rate": 0.0006,
+ "loss": 4.126455307006836,
+ "step": 2309
+ },
+ {
+ "epoch": 32.083842794759825,
+ "grad_norm": 0.020340966060757637,
+ "learning_rate": 0.0006,
+ "loss": 4.080538272857666,
+ "step": 2310
+ },
+ {
+ "epoch": 32.09781659388646,
+ "grad_norm": 0.020956892520189285,
+ "learning_rate": 0.0006,
+ "loss": 4.171943664550781,
+ "step": 2311
+ },
+ {
+ "epoch": 32.1117903930131,
+ "grad_norm": 0.019966667518019676,
+ "learning_rate": 0.0006,
+ "loss": 4.144007682800293,
+ "step": 2312
+ },
+ {
+ "epoch": 32.12576419213974,
+ "grad_norm": 0.018256952986121178,
+ "learning_rate": 0.0006,
+ "loss": 4.208107948303223,
+ "step": 2313
+ },
+ {
+ "epoch": 32.13973799126637,
+ "grad_norm": 0.017289170995354652,
+ "learning_rate": 0.0006,
+ "loss": 4.182024955749512,
+ "step": 2314
+ },
+ {
+ "epoch": 32.15371179039301,
+ "grad_norm": 0.017656700685620308,
+ "learning_rate": 0.0006,
+ "loss": 4.068373680114746,
+ "step": 2315
+ },
+ {
+ "epoch": 32.16768558951965,
+ "grad_norm": 0.017903361469507217,
+ "learning_rate": 0.0006,
+ "loss": 4.2345428466796875,
+ "step": 2316
+ },
+ {
+ "epoch": 32.18165938864629,
+ "grad_norm": 0.01785109005868435,
+ "learning_rate": 0.0006,
+ "loss": 4.219776153564453,
+ "step": 2317
+ },
+ {
+ "epoch": 32.19563318777293,
+ "grad_norm": 0.01887257769703865,
+ "learning_rate": 0.0006,
+ "loss": 4.139206409454346,
+ "step": 2318
+ },
+ {
+ "epoch": 32.209606986899566,
+ "grad_norm": 0.01924685761332512,
+ "learning_rate": 0.0006,
+ "loss": 4.121052265167236,
+ "step": 2319
+ },
+ {
+ "epoch": 32.223580786026204,
+ "grad_norm": 0.020574072375893593,
+ "learning_rate": 0.0006,
+ "loss": 4.13262414932251,
+ "step": 2320
+ },
+ {
+ "epoch": 32.237554585152836,
+ "grad_norm": 0.023849064484238625,
+ "learning_rate": 0.0006,
+ "loss": 4.239119529724121,
+ "step": 2321
+ },
+ {
+ "epoch": 32.251528384279474,
+ "grad_norm": 0.026239361613988876,
+ "learning_rate": 0.0006,
+ "loss": 4.248993873596191,
+ "step": 2322
+ },
+ {
+ "epoch": 32.26550218340611,
+ "grad_norm": 0.023980243131518364,
+ "learning_rate": 0.0006,
+ "loss": 4.159852504730225,
+ "step": 2323
+ },
+ {
+ "epoch": 32.27947598253275,
+ "grad_norm": 0.020442405715584755,
+ "learning_rate": 0.0006,
+ "loss": 4.133546829223633,
+ "step": 2324
+ },
+ {
+ "epoch": 32.29344978165939,
+ "grad_norm": 0.020661186426877975,
+ "learning_rate": 0.0006,
+ "loss": 4.078243732452393,
+ "step": 2325
+ },
+ {
+ "epoch": 32.30742358078603,
+ "grad_norm": 0.02046387456357479,
+ "learning_rate": 0.0006,
+ "loss": 4.172012805938721,
+ "step": 2326
+ },
+ {
+ "epoch": 32.32139737991266,
+ "grad_norm": 0.02089597098529339,
+ "learning_rate": 0.0006,
+ "loss": 4.0604705810546875,
+ "step": 2327
+ },
+ {
+ "epoch": 32.3353711790393,
+ "grad_norm": 0.01875591278076172,
+ "learning_rate": 0.0006,
+ "loss": 4.125911235809326,
+ "step": 2328
+ },
+ {
+ "epoch": 32.34934497816594,
+ "grad_norm": 0.018606796860694885,
+ "learning_rate": 0.0006,
+ "loss": 4.097362518310547,
+ "step": 2329
+ },
+ {
+ "epoch": 32.36331877729258,
+ "grad_norm": 0.01766626164317131,
+ "learning_rate": 0.0006,
+ "loss": 4.118616580963135,
+ "step": 2330
+ },
+ {
+ "epoch": 32.377292576419215,
+ "grad_norm": 0.017692960798740387,
+ "learning_rate": 0.0006,
+ "loss": 4.1457672119140625,
+ "step": 2331
+ },
+ {
+ "epoch": 32.391266375545854,
+ "grad_norm": 0.017619704827666283,
+ "learning_rate": 0.0006,
+ "loss": 4.175961494445801,
+ "step": 2332
+ },
+ {
+ "epoch": 32.40524017467249,
+ "grad_norm": 0.01784413494169712,
+ "learning_rate": 0.0006,
+ "loss": 4.051291465759277,
+ "step": 2333
+ },
+ {
+ "epoch": 32.419213973799124,
+ "grad_norm": 0.018508171662688255,
+ "learning_rate": 0.0006,
+ "loss": 4.0638556480407715,
+ "step": 2334
+ },
+ {
+ "epoch": 32.43318777292576,
+ "grad_norm": 0.01731436885893345,
+ "learning_rate": 0.0006,
+ "loss": 4.1805419921875,
+ "step": 2335
+ },
+ {
+ "epoch": 32.4471615720524,
+ "grad_norm": 0.015886498615145683,
+ "learning_rate": 0.0006,
+ "loss": 4.151026725769043,
+ "step": 2336
+ },
+ {
+ "epoch": 32.46113537117904,
+ "grad_norm": 0.016583820804953575,
+ "learning_rate": 0.0006,
+ "loss": 4.203604698181152,
+ "step": 2337
+ },
+ {
+ "epoch": 32.47510917030568,
+ "grad_norm": 0.0176097322255373,
+ "learning_rate": 0.0006,
+ "loss": 4.135159015655518,
+ "step": 2338
+ },
+ {
+ "epoch": 32.48908296943232,
+ "grad_norm": 0.016657903790473938,
+ "learning_rate": 0.0006,
+ "loss": 4.126681804656982,
+ "step": 2339
+ },
+ {
+ "epoch": 32.50305676855895,
+ "grad_norm": 0.016866130754351616,
+ "learning_rate": 0.0006,
+ "loss": 4.162341117858887,
+ "step": 2340
+ },
+ {
+ "epoch": 32.51703056768559,
+ "grad_norm": 0.016286680474877357,
+ "learning_rate": 0.0006,
+ "loss": 4.261801242828369,
+ "step": 2341
+ },
+ {
+ "epoch": 32.531004366812226,
+ "grad_norm": 0.01481365505605936,
+ "learning_rate": 0.0006,
+ "loss": 4.110016822814941,
+ "step": 2342
+ },
+ {
+ "epoch": 32.544978165938865,
+ "grad_norm": 0.016605013981461525,
+ "learning_rate": 0.0006,
+ "loss": 4.230024814605713,
+ "step": 2343
+ },
+ {
+ "epoch": 32.5589519650655,
+ "grad_norm": 0.01645175740122795,
+ "learning_rate": 0.0006,
+ "loss": 4.174612045288086,
+ "step": 2344
+ },
+ {
+ "epoch": 32.57292576419214,
+ "grad_norm": 0.014725026674568653,
+ "learning_rate": 0.0006,
+ "loss": 4.066371440887451,
+ "step": 2345
+ },
+ {
+ "epoch": 32.58689956331878,
+ "grad_norm": 0.015003837645053864,
+ "learning_rate": 0.0006,
+ "loss": 4.193343639373779,
+ "step": 2346
+ },
+ {
+ "epoch": 32.60087336244541,
+ "grad_norm": 0.015346302650868893,
+ "learning_rate": 0.0006,
+ "loss": 4.210659980773926,
+ "step": 2347
+ },
+ {
+ "epoch": 32.61484716157205,
+ "grad_norm": 0.015135962516069412,
+ "learning_rate": 0.0006,
+ "loss": 4.204618453979492,
+ "step": 2348
+ },
+ {
+ "epoch": 32.62882096069869,
+ "grad_norm": 0.01601956970989704,
+ "learning_rate": 0.0006,
+ "loss": 4.145564556121826,
+ "step": 2349
+ },
+ {
+ "epoch": 32.64279475982533,
+ "grad_norm": 0.01517851185053587,
+ "learning_rate": 0.0006,
+ "loss": 4.051673889160156,
+ "step": 2350
+ },
+ {
+ "epoch": 32.65676855895197,
+ "grad_norm": 0.014894292689859867,
+ "learning_rate": 0.0006,
+ "loss": 4.184818267822266,
+ "step": 2351
+ },
+ {
+ "epoch": 32.670742358078606,
+ "grad_norm": 0.015720834955573082,
+ "learning_rate": 0.0006,
+ "loss": 4.220036506652832,
+ "step": 2352
+ },
+ {
+ "epoch": 32.68471615720524,
+ "grad_norm": 0.01573433168232441,
+ "learning_rate": 0.0006,
+ "loss": 4.076773166656494,
+ "step": 2353
+ },
+ {
+ "epoch": 32.698689956331876,
+ "grad_norm": 0.015828052535653114,
+ "learning_rate": 0.0006,
+ "loss": 4.116329669952393,
+ "step": 2354
+ },
+ {
+ "epoch": 32.712663755458514,
+ "grad_norm": 0.015462259761989117,
+ "learning_rate": 0.0006,
+ "loss": 4.1155242919921875,
+ "step": 2355
+ },
+ {
+ "epoch": 32.72663755458515,
+ "grad_norm": 0.015763849020004272,
+ "learning_rate": 0.0006,
+ "loss": 4.169687747955322,
+ "step": 2356
+ },
+ {
+ "epoch": 32.74061135371179,
+ "grad_norm": 0.014935145154595375,
+ "learning_rate": 0.0006,
+ "loss": 4.2090911865234375,
+ "step": 2357
+ },
+ {
+ "epoch": 32.75458515283843,
+ "grad_norm": 0.015119184739887714,
+ "learning_rate": 0.0006,
+ "loss": 4.131838798522949,
+ "step": 2358
+ },
+ {
+ "epoch": 32.76855895196506,
+ "grad_norm": 0.014833525754511356,
+ "learning_rate": 0.0006,
+ "loss": 4.17618465423584,
+ "step": 2359
+ },
+ {
+ "epoch": 32.7825327510917,
+ "grad_norm": 0.01588359661400318,
+ "learning_rate": 0.0006,
+ "loss": 4.163723945617676,
+ "step": 2360
+ },
+ {
+ "epoch": 32.79650655021834,
+ "grad_norm": 0.0168188214302063,
+ "learning_rate": 0.0006,
+ "loss": 4.178189277648926,
+ "step": 2361
+ },
+ {
+ "epoch": 32.81048034934498,
+ "grad_norm": 0.01726095750927925,
+ "learning_rate": 0.0006,
+ "loss": 4.144223690032959,
+ "step": 2362
+ },
+ {
+ "epoch": 32.82445414847162,
+ "grad_norm": 0.01701410301029682,
+ "learning_rate": 0.0006,
+ "loss": 4.209896087646484,
+ "step": 2363
+ },
+ {
+ "epoch": 32.838427947598255,
+ "grad_norm": 0.01682264916598797,
+ "learning_rate": 0.0006,
+ "loss": 4.26223087310791,
+ "step": 2364
+ },
+ {
+ "epoch": 32.852401746724894,
+ "grad_norm": 0.015597456134855747,
+ "learning_rate": 0.0006,
+ "loss": 4.292664051055908,
+ "step": 2365
+ },
+ {
+ "epoch": 32.866375545851525,
+ "grad_norm": 0.015258579514920712,
+ "learning_rate": 0.0006,
+ "loss": 4.165006637573242,
+ "step": 2366
+ },
+ {
+ "epoch": 32.880349344978164,
+ "grad_norm": 0.01662200503051281,
+ "learning_rate": 0.0006,
+ "loss": 4.309786319732666,
+ "step": 2367
+ },
+ {
+ "epoch": 32.8943231441048,
+ "grad_norm": 0.015691161155700684,
+ "learning_rate": 0.0006,
+ "loss": 4.167202949523926,
+ "step": 2368
+ },
+ {
+ "epoch": 32.90829694323144,
+ "grad_norm": 0.014870637096464634,
+ "learning_rate": 0.0006,
+ "loss": 4.142415523529053,
+ "step": 2369
+ },
+ {
+ "epoch": 32.92227074235808,
+ "grad_norm": 0.01567252166569233,
+ "learning_rate": 0.0006,
+ "loss": 4.2312092781066895,
+ "step": 2370
+ },
+ {
+ "epoch": 32.93624454148472,
+ "grad_norm": 0.014948616735637188,
+ "learning_rate": 0.0006,
+ "loss": 4.1514129638671875,
+ "step": 2371
+ },
+ {
+ "epoch": 32.95021834061135,
+ "grad_norm": 0.014654500409960747,
+ "learning_rate": 0.0006,
+ "loss": 4.205010890960693,
+ "step": 2372
+ },
+ {
+ "epoch": 32.96419213973799,
+ "grad_norm": 0.015059148892760277,
+ "learning_rate": 0.0006,
+ "loss": 4.148405075073242,
+ "step": 2373
+ },
+ {
+ "epoch": 32.97816593886463,
+ "grad_norm": 0.015026576817035675,
+ "learning_rate": 0.0006,
+ "loss": 4.182535171508789,
+ "step": 2374
+ },
+ {
+ "epoch": 32.992139737991266,
+ "grad_norm": 0.016394320875406265,
+ "learning_rate": 0.0006,
+ "loss": 4.129051208496094,
+ "step": 2375
+ },
+ {
+ "epoch": 33.0,
+ "grad_norm": 0.01867043413221836,
+ "learning_rate": 0.0006,
+ "loss": 4.185556411743164,
+ "step": 2376
+ },
+ {
+ "epoch": 33.0,
+ "eval_loss": 4.640405654907227,
+ "eval_runtime": 56.5861,
+ "eval_samples_per_second": 43.155,
+ "eval_steps_per_second": 1.361,
+ "step": 2376
+ },
+ {
+ "epoch": 33.01397379912664,
+ "grad_norm": 0.018626727163791656,
+ "learning_rate": 0.0006,
+ "loss": 4.143031597137451,
+ "step": 2377
+ },
+ {
+ "epoch": 33.02794759825328,
+ "grad_norm": 0.019283363595604897,
+ "learning_rate": 0.0006,
+ "loss": 4.117845058441162,
+ "step": 2378
+ },
+ {
+ "epoch": 33.041921397379916,
+ "grad_norm": 0.018185822293162346,
+ "learning_rate": 0.0006,
+ "loss": 4.076862812042236,
+ "step": 2379
+ },
+ {
+ "epoch": 33.05589519650655,
+ "grad_norm": 0.019411196932196617,
+ "learning_rate": 0.0006,
+ "loss": 4.124156475067139,
+ "step": 2380
+ },
+ {
+ "epoch": 33.069868995633186,
+ "grad_norm": 0.019248466938734055,
+ "learning_rate": 0.0006,
+ "loss": 4.201735019683838,
+ "step": 2381
+ },
+ {
+ "epoch": 33.083842794759825,
+ "grad_norm": 0.018856720998883247,
+ "learning_rate": 0.0006,
+ "loss": 4.00296688079834,
+ "step": 2382
+ },
+ {
+ "epoch": 33.09781659388646,
+ "grad_norm": 0.018554294481873512,
+ "learning_rate": 0.0006,
+ "loss": 4.270577430725098,
+ "step": 2383
+ },
+ {
+ "epoch": 33.1117903930131,
+ "grad_norm": 0.019946224987506866,
+ "learning_rate": 0.0006,
+ "loss": 4.119589328765869,
+ "step": 2384
+ },
+ {
+ "epoch": 33.12576419213974,
+ "grad_norm": 0.021395670250058174,
+ "learning_rate": 0.0006,
+ "loss": 4.128847122192383,
+ "step": 2385
+ },
+ {
+ "epoch": 33.13973799126637,
+ "grad_norm": 0.02009674906730652,
+ "learning_rate": 0.0006,
+ "loss": 4.183228969573975,
+ "step": 2386
+ },
+ {
+ "epoch": 33.15371179039301,
+ "grad_norm": 0.01820223033428192,
+ "learning_rate": 0.0006,
+ "loss": 4.131601810455322,
+ "step": 2387
+ },
+ {
+ "epoch": 33.16768558951965,
+ "grad_norm": 0.018653156235814095,
+ "learning_rate": 0.0006,
+ "loss": 4.116430282592773,
+ "step": 2388
+ },
+ {
+ "epoch": 33.18165938864629,
+ "grad_norm": 0.016828063875436783,
+ "learning_rate": 0.0006,
+ "loss": 4.054663181304932,
+ "step": 2389
+ },
+ {
+ "epoch": 33.19563318777293,
+ "grad_norm": 0.017025409266352654,
+ "learning_rate": 0.0006,
+ "loss": 4.110044956207275,
+ "step": 2390
+ },
+ {
+ "epoch": 33.209606986899566,
+ "grad_norm": 0.016608452424407005,
+ "learning_rate": 0.0006,
+ "loss": 4.102668285369873,
+ "step": 2391
+ },
+ {
+ "epoch": 33.223580786026204,
+ "grad_norm": 0.01743577979505062,
+ "learning_rate": 0.0006,
+ "loss": 4.166791915893555,
+ "step": 2392
+ },
+ {
+ "epoch": 33.237554585152836,
+ "grad_norm": 0.017253734171390533,
+ "learning_rate": 0.0006,
+ "loss": 4.217832088470459,
+ "step": 2393
+ },
+ {
+ "epoch": 33.251528384279474,
+ "grad_norm": 0.017937595024704933,
+ "learning_rate": 0.0006,
+ "loss": 4.2173967361450195,
+ "step": 2394
+ },
+ {
+ "epoch": 33.26550218340611,
+ "grad_norm": 0.01985993981361389,
+ "learning_rate": 0.0006,
+ "loss": 4.17849588394165,
+ "step": 2395
+ },
+ {
+ "epoch": 33.27947598253275,
+ "grad_norm": 0.019243355840444565,
+ "learning_rate": 0.0006,
+ "loss": 4.086523056030273,
+ "step": 2396
+ },
+ {
+ "epoch": 33.29344978165939,
+ "grad_norm": 0.01924416422843933,
+ "learning_rate": 0.0006,
+ "loss": 4.128574371337891,
+ "step": 2397
+ },
+ {
+ "epoch": 33.30742358078603,
+ "grad_norm": 0.01983758620917797,
+ "learning_rate": 0.0006,
+ "loss": 4.061581611633301,
+ "step": 2398
+ },
+ {
+ "epoch": 33.32139737991266,
+ "grad_norm": 0.01930462196469307,
+ "learning_rate": 0.0006,
+ "loss": 4.091731071472168,
+ "step": 2399
+ },
+ {
+ "epoch": 33.3353711790393,
+ "grad_norm": 0.017988424748182297,
+ "learning_rate": 0.0006,
+ "loss": 4.155599117279053,
+ "step": 2400
+ },
+ {
+ "epoch": 33.34934497816594,
+ "grad_norm": 0.015194767154753208,
+ "learning_rate": 0.0006,
+ "loss": 4.144134044647217,
+ "step": 2401
+ },
+ {
+ "epoch": 33.36331877729258,
+ "grad_norm": 0.01536081824451685,
+ "learning_rate": 0.0006,
+ "loss": 4.127272129058838,
+ "step": 2402
+ },
+ {
+ "epoch": 33.377292576419215,
+ "grad_norm": 0.015431297942996025,
+ "learning_rate": 0.0006,
+ "loss": 4.3213019371032715,
+ "step": 2403
+ },
+ {
+ "epoch": 33.391266375545854,
+ "grad_norm": 0.015740280970931053,
+ "learning_rate": 0.0006,
+ "loss": 4.121467590332031,
+ "step": 2404
+ },
+ {
+ "epoch": 33.40524017467249,
+ "grad_norm": 0.015379384160041809,
+ "learning_rate": 0.0006,
+ "loss": 4.095216751098633,
+ "step": 2405
+ },
+ {
+ "epoch": 33.419213973799124,
+ "grad_norm": 0.01528073474764824,
+ "learning_rate": 0.0006,
+ "loss": 4.13047981262207,
+ "step": 2406
+ },
+ {
+ "epoch": 33.43318777292576,
+ "grad_norm": 0.015739979222416878,
+ "learning_rate": 0.0006,
+ "loss": 4.2533135414123535,
+ "step": 2407
+ },
+ {
+ "epoch": 33.4471615720524,
+ "grad_norm": 0.01519764307886362,
+ "learning_rate": 0.0006,
+ "loss": 4.131782054901123,
+ "step": 2408
+ },
+ {
+ "epoch": 33.46113537117904,
+ "grad_norm": 0.015425996854901314,
+ "learning_rate": 0.0006,
+ "loss": 4.02811336517334,
+ "step": 2409
+ },
+ {
+ "epoch": 33.47510917030568,
+ "grad_norm": 0.015378830023109913,
+ "learning_rate": 0.0006,
+ "loss": 4.08652925491333,
+ "step": 2410
+ },
+ {
+ "epoch": 33.48908296943232,
+ "grad_norm": 0.01473075058311224,
+ "learning_rate": 0.0006,
+ "loss": 4.214364528656006,
+ "step": 2411
+ },
+ {
+ "epoch": 33.50305676855895,
+ "grad_norm": 0.015141637064516544,
+ "learning_rate": 0.0006,
+ "loss": 4.155712127685547,
+ "step": 2412
+ },
+ {
+ "epoch": 33.51703056768559,
+ "grad_norm": 0.01641593500971794,
+ "learning_rate": 0.0006,
+ "loss": 4.097907066345215,
+ "step": 2413
+ },
+ {
+ "epoch": 33.531004366812226,
+ "grad_norm": 0.015182029455900192,
+ "learning_rate": 0.0006,
+ "loss": 4.238499641418457,
+ "step": 2414
+ },
+ {
+ "epoch": 33.544978165938865,
+ "grad_norm": 0.01643732376396656,
+ "learning_rate": 0.0006,
+ "loss": 4.0661492347717285,
+ "step": 2415
+ },
+ {
+ "epoch": 33.5589519650655,
+ "grad_norm": 0.015614228323101997,
+ "learning_rate": 0.0006,
+ "loss": 4.120795249938965,
+ "step": 2416
+ },
+ {
+ "epoch": 33.57292576419214,
+ "grad_norm": 0.016066187992691994,
+ "learning_rate": 0.0006,
+ "loss": 4.096465110778809,
+ "step": 2417
+ },
+ {
+ "epoch": 33.58689956331878,
+ "grad_norm": 0.017214342951774597,
+ "learning_rate": 0.0006,
+ "loss": 4.242628574371338,
+ "step": 2418
+ },
+ {
+ "epoch": 33.60087336244541,
+ "grad_norm": 0.01787586510181427,
+ "learning_rate": 0.0006,
+ "loss": 4.149589538574219,
+ "step": 2419
+ },
+ {
+ "epoch": 33.61484716157205,
+ "grad_norm": 0.019455241039395332,
+ "learning_rate": 0.0006,
+ "loss": 4.267796039581299,
+ "step": 2420
+ },
+ {
+ "epoch": 33.62882096069869,
+ "grad_norm": 0.01984674669802189,
+ "learning_rate": 0.0006,
+ "loss": 4.175220966339111,
+ "step": 2421
+ },
+ {
+ "epoch": 33.64279475982533,
+ "grad_norm": 0.021933717653155327,
+ "learning_rate": 0.0006,
+ "loss": 4.106647491455078,
+ "step": 2422
+ },
+ {
+ "epoch": 33.65676855895197,
+ "grad_norm": 0.02123589813709259,
+ "learning_rate": 0.0006,
+ "loss": 4.204996109008789,
+ "step": 2423
+ },
+ {
+ "epoch": 33.670742358078606,
+ "grad_norm": 0.019814589992165565,
+ "learning_rate": 0.0006,
+ "loss": 4.186171531677246,
+ "step": 2424
+ },
+ {
+ "epoch": 33.68471615720524,
+ "grad_norm": 0.02059728465974331,
+ "learning_rate": 0.0006,
+ "loss": 4.170509338378906,
+ "step": 2425
+ },
+ {
+ "epoch": 33.698689956331876,
+ "grad_norm": 0.018483608961105347,
+ "learning_rate": 0.0006,
+ "loss": 4.085585594177246,
+ "step": 2426
+ },
+ {
+ "epoch": 33.712663755458514,
+ "grad_norm": 0.018403852358460426,
+ "learning_rate": 0.0006,
+ "loss": 4.2551774978637695,
+ "step": 2427
+ },
+ {
+ "epoch": 33.72663755458515,
+ "grad_norm": 0.018653830513358116,
+ "learning_rate": 0.0006,
+ "loss": 4.091766357421875,
+ "step": 2428
+ },
+ {
+ "epoch": 33.74061135371179,
+ "grad_norm": 0.017692334949970245,
+ "learning_rate": 0.0006,
+ "loss": 4.100838661193848,
+ "step": 2429
+ },
+ {
+ "epoch": 33.75458515283843,
+ "grad_norm": 0.016029126942157745,
+ "learning_rate": 0.0006,
+ "loss": 4.1530680656433105,
+ "step": 2430
+ },
+ {
+ "epoch": 33.76855895196506,
+ "grad_norm": 0.016095684841275215,
+ "learning_rate": 0.0006,
+ "loss": 4.20544958114624,
+ "step": 2431
+ },
+ {
+ "epoch": 33.7825327510917,
+ "grad_norm": 0.016375798732042313,
+ "learning_rate": 0.0006,
+ "loss": 4.199177265167236,
+ "step": 2432
+ },
+ {
+ "epoch": 33.79650655021834,
+ "grad_norm": 0.01590786688029766,
+ "learning_rate": 0.0006,
+ "loss": 4.200799465179443,
+ "step": 2433
+ },
+ {
+ "epoch": 33.81048034934498,
+ "grad_norm": 0.014896097593009472,
+ "learning_rate": 0.0006,
+ "loss": 4.185850143432617,
+ "step": 2434
+ },
+ {
+ "epoch": 33.82445414847162,
+ "grad_norm": 0.01438513956964016,
+ "learning_rate": 0.0006,
+ "loss": 4.214962005615234,
+ "step": 2435
+ },
+ {
+ "epoch": 33.838427947598255,
+ "grad_norm": 0.014459339901804924,
+ "learning_rate": 0.0006,
+ "loss": 4.196972370147705,
+ "step": 2436
+ },
+ {
+ "epoch": 33.852401746724894,
+ "grad_norm": 0.014782169833779335,
+ "learning_rate": 0.0006,
+ "loss": 4.0885910987854,
+ "step": 2437
+ },
+ {
+ "epoch": 33.866375545851525,
+ "grad_norm": 0.01520370040088892,
+ "learning_rate": 0.0006,
+ "loss": 4.0323262214660645,
+ "step": 2438
+ },
+ {
+ "epoch": 33.880349344978164,
+ "grad_norm": 0.014056282117962837,
+ "learning_rate": 0.0006,
+ "loss": 4.1386518478393555,
+ "step": 2439
+ },
+ {
+ "epoch": 33.8943231441048,
+ "grad_norm": 0.013567990623414516,
+ "learning_rate": 0.0006,
+ "loss": 4.095754623413086,
+ "step": 2440
+ },
+ {
+ "epoch": 33.90829694323144,
+ "grad_norm": 0.013306716457009315,
+ "learning_rate": 0.0006,
+ "loss": 4.1623077392578125,
+ "step": 2441
+ },
+ {
+ "epoch": 33.92227074235808,
+ "grad_norm": 0.014047150500118732,
+ "learning_rate": 0.0006,
+ "loss": 4.142541885375977,
+ "step": 2442
+ },
+ {
+ "epoch": 33.93624454148472,
+ "grad_norm": 0.01377617847174406,
+ "learning_rate": 0.0006,
+ "loss": 4.268551826477051,
+ "step": 2443
+ },
+ {
+ "epoch": 33.95021834061135,
+ "grad_norm": 0.014671621844172478,
+ "learning_rate": 0.0006,
+ "loss": 4.198196887969971,
+ "step": 2444
+ },
+ {
+ "epoch": 33.96419213973799,
+ "grad_norm": 0.014806744642555714,
+ "learning_rate": 0.0006,
+ "loss": 4.057899475097656,
+ "step": 2445
+ },
+ {
+ "epoch": 33.97816593886463,
+ "grad_norm": 0.015734272077679634,
+ "learning_rate": 0.0006,
+ "loss": 4.163505554199219,
+ "step": 2446
+ },
+ {
+ "epoch": 33.992139737991266,
+ "grad_norm": 0.015250146389007568,
+ "learning_rate": 0.0006,
+ "loss": 4.076294898986816,
+ "step": 2447
+ },
+ {
+ "epoch": 34.0,
+ "grad_norm": 0.016733702272176743,
+ "learning_rate": 0.0006,
+ "loss": 4.138420104980469,
+ "step": 2448
+ },
+ {
+ "epoch": 34.0,
+ "eval_loss": 4.542503356933594,
+ "eval_runtime": 56.983,
+ "eval_samples_per_second": 42.855,
+ "eval_steps_per_second": 1.351,
+ "step": 2448
+ },
+ {
+ "epoch": 34.01397379912664,
+ "grad_norm": 0.015866613015532494,
+ "learning_rate": 0.0006,
+ "loss": 4.032081127166748,
+ "step": 2449
+ },
+ {
+ "epoch": 34.02794759825328,
+ "grad_norm": 0.01946570910513401,
+ "learning_rate": 0.0006,
+ "loss": 4.122320175170898,
+ "step": 2450
+ },
+ {
+ "epoch": 34.041921397379916,
+ "grad_norm": 0.020265938714146614,
+ "learning_rate": 0.0006,
+ "loss": 4.06460428237915,
+ "step": 2451
+ },
+ {
+ "epoch": 34.05589519650655,
+ "grad_norm": 0.018887171521782875,
+ "learning_rate": 0.0006,
+ "loss": 4.096906661987305,
+ "step": 2452
+ },
+ {
+ "epoch": 34.069868995633186,
+ "grad_norm": 0.017304692417383194,
+ "learning_rate": 0.0006,
+ "loss": 4.137261390686035,
+ "step": 2453
+ },
+ {
+ "epoch": 34.083842794759825,
+ "grad_norm": 0.017029082402586937,
+ "learning_rate": 0.0006,
+ "loss": 4.081112861633301,
+ "step": 2454
+ },
+ {
+ "epoch": 34.09781659388646,
+ "grad_norm": 0.016414813697338104,
+ "learning_rate": 0.0006,
+ "loss": 4.140783309936523,
+ "step": 2455
+ },
+ {
+ "epoch": 34.1117903930131,
+ "grad_norm": 0.018238281831145287,
+ "learning_rate": 0.0006,
+ "loss": 4.161116600036621,
+ "step": 2456
+ },
+ {
+ "epoch": 34.12576419213974,
+ "grad_norm": 0.019929232075810432,
+ "learning_rate": 0.0006,
+ "loss": 4.243722438812256,
+ "step": 2457
+ },
+ {
+ "epoch": 34.13973799126637,
+ "grad_norm": 0.019375786185264587,
+ "learning_rate": 0.0006,
+ "loss": 4.134873390197754,
+ "step": 2458
+ },
+ {
+ "epoch": 34.15371179039301,
+ "grad_norm": 0.018151137977838516,
+ "learning_rate": 0.0006,
+ "loss": 4.098745346069336,
+ "step": 2459
+ },
+ {
+ "epoch": 34.16768558951965,
+ "grad_norm": 0.017054934054613113,
+ "learning_rate": 0.0006,
+ "loss": 4.194662570953369,
+ "step": 2460
+ },
+ {
+ "epoch": 34.18165938864629,
+ "grad_norm": 0.018220555037260056,
+ "learning_rate": 0.0006,
+ "loss": 4.128866195678711,
+ "step": 2461
+ },
+ {
+ "epoch": 34.19563318777293,
+ "grad_norm": 0.01692879945039749,
+ "learning_rate": 0.0006,
+ "loss": 4.062477111816406,
+ "step": 2462
+ },
+ {
+ "epoch": 34.209606986899566,
+ "grad_norm": 0.01674199476838112,
+ "learning_rate": 0.0006,
+ "loss": 4.186149597167969,
+ "step": 2463
+ },
+ {
+ "epoch": 34.223580786026204,
+ "grad_norm": 0.016527455300092697,
+ "learning_rate": 0.0006,
+ "loss": 4.0734028816223145,
+ "step": 2464
+ },
+ {
+ "epoch": 34.237554585152836,
+ "grad_norm": 0.016448386013507843,
+ "learning_rate": 0.0006,
+ "loss": 4.098896026611328,
+ "step": 2465
+ },
+ {
+ "epoch": 34.251528384279474,
+ "grad_norm": 0.014644953422248363,
+ "learning_rate": 0.0006,
+ "loss": 4.1031904220581055,
+ "step": 2466
+ },
+ {
+ "epoch": 34.26550218340611,
+ "grad_norm": 0.016649341210722923,
+ "learning_rate": 0.0006,
+ "loss": 4.148791313171387,
+ "step": 2467
+ },
+ {
+ "epoch": 34.27947598253275,
+ "grad_norm": 0.017442714422941208,
+ "learning_rate": 0.0006,
+ "loss": 4.070799827575684,
+ "step": 2468
+ },
+ {
+ "epoch": 34.29344978165939,
+ "grad_norm": 0.017852013930678368,
+ "learning_rate": 0.0006,
+ "loss": 4.031924247741699,
+ "step": 2469
+ },
+ {
+ "epoch": 34.30742358078603,
+ "grad_norm": 0.016371270641684532,
+ "learning_rate": 0.0006,
+ "loss": 4.149420738220215,
+ "step": 2470
+ },
+ {
+ "epoch": 34.32139737991266,
+ "grad_norm": 0.015846814960241318,
+ "learning_rate": 0.0006,
+ "loss": 4.15427303314209,
+ "step": 2471
+ },
+ {
+ "epoch": 34.3353711790393,
+ "grad_norm": 0.017411163076758385,
+ "learning_rate": 0.0006,
+ "loss": 4.1304192543029785,
+ "step": 2472
+ },
+ {
+ "epoch": 34.34934497816594,
+ "grad_norm": 0.01780046708881855,
+ "learning_rate": 0.0006,
+ "loss": 4.191732406616211,
+ "step": 2473
+ },
+ {
+ "epoch": 34.36331877729258,
+ "grad_norm": 0.01958383060991764,
+ "learning_rate": 0.0006,
+ "loss": 4.068355083465576,
+ "step": 2474
+ },
+ {
+ "epoch": 34.377292576419215,
+ "grad_norm": 0.022173317149281502,
+ "learning_rate": 0.0006,
+ "loss": 4.182929039001465,
+ "step": 2475
+ },
+ {
+ "epoch": 34.391266375545854,
+ "grad_norm": 0.022389156743884087,
+ "learning_rate": 0.0006,
+ "loss": 4.110218048095703,
+ "step": 2476
+ },
+ {
+ "epoch": 34.40524017467249,
+ "grad_norm": 0.022478511556982994,
+ "learning_rate": 0.0006,
+ "loss": 4.095479488372803,
+ "step": 2477
+ },
+ {
+ "epoch": 34.419213973799124,
+ "grad_norm": 0.019504647701978683,
+ "learning_rate": 0.0006,
+ "loss": 4.101964950561523,
+ "step": 2478
+ },
+ {
+ "epoch": 34.43318777292576,
+ "grad_norm": 0.019378576427698135,
+ "learning_rate": 0.0006,
+ "loss": 4.14620304107666,
+ "step": 2479
+ },
+ {
+ "epoch": 34.4471615720524,
+ "grad_norm": 0.017993204295635223,
+ "learning_rate": 0.0006,
+ "loss": 4.183492660522461,
+ "step": 2480
+ },
+ {
+ "epoch": 34.46113537117904,
+ "grad_norm": 0.01814805157482624,
+ "learning_rate": 0.0006,
+ "loss": 4.151247978210449,
+ "step": 2481
+ },
+ {
+ "epoch": 34.47510917030568,
+ "grad_norm": 0.018622448667883873,
+ "learning_rate": 0.0006,
+ "loss": 4.020938396453857,
+ "step": 2482
+ },
+ {
+ "epoch": 34.48908296943232,
+ "grad_norm": 0.018360132351517677,
+ "learning_rate": 0.0006,
+ "loss": 4.064734935760498,
+ "step": 2483
+ },
+ {
+ "epoch": 34.50305676855895,
+ "grad_norm": 0.017715711146593094,
+ "learning_rate": 0.0006,
+ "loss": 4.185687065124512,
+ "step": 2484
+ },
+ {
+ "epoch": 34.51703056768559,
+ "grad_norm": 0.01714295893907547,
+ "learning_rate": 0.0006,
+ "loss": 4.1044721603393555,
+ "step": 2485
+ },
+ {
+ "epoch": 34.531004366812226,
+ "grad_norm": 0.016943803057074547,
+ "learning_rate": 0.0006,
+ "loss": 4.106837272644043,
+ "step": 2486
+ },
+ {
+ "epoch": 34.544978165938865,
+ "grad_norm": 0.0182423684746027,
+ "learning_rate": 0.0006,
+ "loss": 4.1041340827941895,
+ "step": 2487
+ },
+ {
+ "epoch": 34.5589519650655,
+ "grad_norm": 0.01636587455868721,
+ "learning_rate": 0.0006,
+ "loss": 3.9735546112060547,
+ "step": 2488
+ },
+ {
+ "epoch": 34.57292576419214,
+ "grad_norm": 0.014593546278774738,
+ "learning_rate": 0.0006,
+ "loss": 4.041980266571045,
+ "step": 2489
+ },
+ {
+ "epoch": 34.58689956331878,
+ "grad_norm": 0.015436965972185135,
+ "learning_rate": 0.0006,
+ "loss": 4.301148414611816,
+ "step": 2490
+ },
+ {
+ "epoch": 34.60087336244541,
+ "grad_norm": 0.015075269155204296,
+ "learning_rate": 0.0006,
+ "loss": 4.118527412414551,
+ "step": 2491
+ },
+ {
+ "epoch": 34.61484716157205,
+ "grad_norm": 0.01701514609158039,
+ "learning_rate": 0.0006,
+ "loss": 4.245968818664551,
+ "step": 2492
+ },
+ {
+ "epoch": 34.62882096069869,
+ "grad_norm": 0.018047887831926346,
+ "learning_rate": 0.0006,
+ "loss": 4.227261066436768,
+ "step": 2493
+ },
+ {
+ "epoch": 34.64279475982533,
+ "grad_norm": 0.018499085679650307,
+ "learning_rate": 0.0006,
+ "loss": 4.121360778808594,
+ "step": 2494
+ },
+ {
+ "epoch": 34.65676855895197,
+ "grad_norm": 0.018987711519002914,
+ "learning_rate": 0.0006,
+ "loss": 4.131564617156982,
+ "step": 2495
+ },
+ {
+ "epoch": 34.670742358078606,
+ "grad_norm": 0.01834476925432682,
+ "learning_rate": 0.0006,
+ "loss": 4.224941253662109,
+ "step": 2496
+ },
+ {
+ "epoch": 34.68471615720524,
+ "grad_norm": 0.017507942393422127,
+ "learning_rate": 0.0006,
+ "loss": 4.1394548416137695,
+ "step": 2497
+ },
+ {
+ "epoch": 34.698689956331876,
+ "grad_norm": 0.01680580899119377,
+ "learning_rate": 0.0006,
+ "loss": 4.1073431968688965,
+ "step": 2498
+ },
+ {
+ "epoch": 34.712663755458514,
+ "grad_norm": 0.015233514830470085,
+ "learning_rate": 0.0006,
+ "loss": 4.136994361877441,
+ "step": 2499
+ },
+ {
+ "epoch": 34.72663755458515,
+ "grad_norm": 0.015482917428016663,
+ "learning_rate": 0.0006,
+ "loss": 4.110719680786133,
+ "step": 2500
+ },
+ {
+ "epoch": 34.74061135371179,
+ "grad_norm": 0.015940461307764053,
+ "learning_rate": 0.0006,
+ "loss": 4.207851886749268,
+ "step": 2501
+ },
+ {
+ "epoch": 34.75458515283843,
+ "grad_norm": 0.015355597250163555,
+ "learning_rate": 0.0006,
+ "loss": 4.169469833374023,
+ "step": 2502
+ },
+ {
+ "epoch": 34.76855895196506,
+ "grad_norm": 0.015974000096321106,
+ "learning_rate": 0.0006,
+ "loss": 4.110061168670654,
+ "step": 2503
+ },
+ {
+ "epoch": 34.7825327510917,
+ "grad_norm": 0.016088927164673805,
+ "learning_rate": 0.0006,
+ "loss": 4.155452728271484,
+ "step": 2504
+ },
+ {
+ "epoch": 34.79650655021834,
+ "grad_norm": 0.016156120225787163,
+ "learning_rate": 0.0006,
+ "loss": 4.178092956542969,
+ "step": 2505
+ },
+ {
+ "epoch": 34.81048034934498,
+ "grad_norm": 0.014928647316992283,
+ "learning_rate": 0.0006,
+ "loss": 4.05457878112793,
+ "step": 2506
+ },
+ {
+ "epoch": 34.82445414847162,
+ "grad_norm": 0.015289328061044216,
+ "learning_rate": 0.0006,
+ "loss": 4.18869686126709,
+ "step": 2507
+ },
+ {
+ "epoch": 34.838427947598255,
+ "grad_norm": 0.015799039974808693,
+ "learning_rate": 0.0006,
+ "loss": 4.106073379516602,
+ "step": 2508
+ },
+ {
+ "epoch": 34.852401746724894,
+ "grad_norm": 0.015824418514966965,
+ "learning_rate": 0.0006,
+ "loss": 4.0634589195251465,
+ "step": 2509
+ },
+ {
+ "epoch": 34.866375545851525,
+ "grad_norm": 0.015321656130254269,
+ "learning_rate": 0.0006,
+ "loss": 4.093090057373047,
+ "step": 2510
+ },
+ {
+ "epoch": 34.880349344978164,
+ "grad_norm": 0.014380166307091713,
+ "learning_rate": 0.0006,
+ "loss": 4.1750640869140625,
+ "step": 2511
+ },
+ {
+ "epoch": 34.8943231441048,
+ "grad_norm": 0.014962567947804928,
+ "learning_rate": 0.0006,
+ "loss": 4.185277938842773,
+ "step": 2512
+ },
+ {
+ "epoch": 34.90829694323144,
+ "grad_norm": 0.014467543922364712,
+ "learning_rate": 0.0006,
+ "loss": 4.1600799560546875,
+ "step": 2513
+ },
+ {
+ "epoch": 34.92227074235808,
+ "grad_norm": 0.01545181218534708,
+ "learning_rate": 0.0006,
+ "loss": 4.096299648284912,
+ "step": 2514
+ },
+ {
+ "epoch": 34.93624454148472,
+ "grad_norm": 0.01605340465903282,
+ "learning_rate": 0.0006,
+ "loss": 4.203323841094971,
+ "step": 2515
+ },
+ {
+ "epoch": 34.95021834061135,
+ "grad_norm": 0.015203815884888172,
+ "learning_rate": 0.0006,
+ "loss": 4.201385021209717,
+ "step": 2516
+ },
+ {
+ "epoch": 34.96419213973799,
+ "grad_norm": 0.015149409882724285,
+ "learning_rate": 0.0006,
+ "loss": 4.088164329528809,
+ "step": 2517
+ },
+ {
+ "epoch": 34.97816593886463,
+ "grad_norm": 0.014885558746755123,
+ "learning_rate": 0.0006,
+ "loss": 4.073911666870117,
+ "step": 2518
+ },
+ {
+ "epoch": 34.992139737991266,
+ "grad_norm": 0.015003419481217861,
+ "learning_rate": 0.0006,
+ "loss": 4.157113552093506,
+ "step": 2519
+ },
+ {
+ "epoch": 35.0,
+ "grad_norm": 0.01641916297376156,
+ "learning_rate": 0.0006,
+ "loss": 3.970052719116211,
+ "step": 2520
+ },
+ {
+ "epoch": 35.0,
+ "eval_loss": 4.641692161560059,
+ "eval_runtime": 56.8836,
+ "eval_samples_per_second": 42.93,
+ "eval_steps_per_second": 1.354,
+ "step": 2520
+ },
+ {
+ "epoch": 35.01397379912664,
+ "grad_norm": 0.015894673764705658,
+ "learning_rate": 0.0006,
+ "loss": 4.104437351226807,
+ "step": 2521
+ },
+ {
+ "epoch": 35.02794759825328,
+ "grad_norm": 0.0173867866396904,
+ "learning_rate": 0.0006,
+ "loss": 4.091726303100586,
+ "step": 2522
+ },
+ {
+ "epoch": 35.041921397379916,
+ "grad_norm": 0.019272832199931145,
+ "learning_rate": 0.0006,
+ "loss": 4.217905044555664,
+ "step": 2523
+ },
+ {
+ "epoch": 35.05589519650655,
+ "grad_norm": 0.021820371970534325,
+ "learning_rate": 0.0006,
+ "loss": 4.111428260803223,
+ "step": 2524
+ },
+ {
+ "epoch": 35.069868995633186,
+ "grad_norm": 0.02340741828083992,
+ "learning_rate": 0.0006,
+ "loss": 4.115771293640137,
+ "step": 2525
+ },
+ {
+ "epoch": 35.083842794759825,
+ "grad_norm": 0.023760691285133362,
+ "learning_rate": 0.0006,
+ "loss": 4.132383346557617,
+ "step": 2526
+ },
+ {
+ "epoch": 35.09781659388646,
+ "grad_norm": 0.02073468267917633,
+ "learning_rate": 0.0006,
+ "loss": 4.272691249847412,
+ "step": 2527
+ },
+ {
+ "epoch": 35.1117903930131,
+ "grad_norm": 0.02197718247771263,
+ "learning_rate": 0.0006,
+ "loss": 4.105586528778076,
+ "step": 2528
+ },
+ {
+ "epoch": 35.12576419213974,
+ "grad_norm": 0.0200254675000906,
+ "learning_rate": 0.0006,
+ "loss": 3.968820095062256,
+ "step": 2529
+ },
+ {
+ "epoch": 35.13973799126637,
+ "grad_norm": 0.01929587684571743,
+ "learning_rate": 0.0006,
+ "loss": 4.045468807220459,
+ "step": 2530
+ },
+ {
+ "epoch": 35.15371179039301,
+ "grad_norm": 0.01886162906885147,
+ "learning_rate": 0.0006,
+ "loss": 4.185571670532227,
+ "step": 2531
+ },
+ {
+ "epoch": 35.16768558951965,
+ "grad_norm": 0.019789839163422585,
+ "learning_rate": 0.0006,
+ "loss": 3.9734551906585693,
+ "step": 2532
+ },
+ {
+ "epoch": 35.18165938864629,
+ "grad_norm": 0.0198142658919096,
+ "learning_rate": 0.0006,
+ "loss": 4.2662506103515625,
+ "step": 2533
+ },
+ {
+ "epoch": 35.19563318777293,
+ "grad_norm": 0.02087012305855751,
+ "learning_rate": 0.0006,
+ "loss": 4.175931930541992,
+ "step": 2534
+ },
+ {
+ "epoch": 35.209606986899566,
+ "grad_norm": 0.020294535905122757,
+ "learning_rate": 0.0006,
+ "loss": 4.060324668884277,
+ "step": 2535
+ },
+ {
+ "epoch": 35.223580786026204,
+ "grad_norm": 0.01908983662724495,
+ "learning_rate": 0.0006,
+ "loss": 4.217939853668213,
+ "step": 2536
+ },
+ {
+ "epoch": 35.237554585152836,
+ "grad_norm": 0.0178892333060503,
+ "learning_rate": 0.0006,
+ "loss": 4.190216541290283,
+ "step": 2537
+ },
+ {
+ "epoch": 35.251528384279474,
+ "grad_norm": 0.017892315983772278,
+ "learning_rate": 0.0006,
+ "loss": 4.067419052124023,
+ "step": 2538
+ },
+ {
+ "epoch": 35.26550218340611,
+ "grad_norm": 0.018556250259280205,
+ "learning_rate": 0.0006,
+ "loss": 4.105958461761475,
+ "step": 2539
+ },
+ {
+ "epoch": 35.27947598253275,
+ "grad_norm": 0.01944422349333763,
+ "learning_rate": 0.0006,
+ "loss": 4.155775547027588,
+ "step": 2540
+ },
+ {
+ "epoch": 35.29344978165939,
+ "grad_norm": 0.017333146184682846,
+ "learning_rate": 0.0006,
+ "loss": 4.244513511657715,
+ "step": 2541
+ },
+ {
+ "epoch": 35.30742358078603,
+ "grad_norm": 0.01639244332909584,
+ "learning_rate": 0.0006,
+ "loss": 4.167815685272217,
+ "step": 2542
+ },
+ {
+ "epoch": 35.32139737991266,
+ "grad_norm": 0.01610143668949604,
+ "learning_rate": 0.0006,
+ "loss": 4.121486663818359,
+ "step": 2543
+ },
+ {
+ "epoch": 35.3353711790393,
+ "grad_norm": 0.01656411588191986,
+ "learning_rate": 0.0006,
+ "loss": 4.1755499839782715,
+ "step": 2544
+ },
+ {
+ "epoch": 35.34934497816594,
+ "grad_norm": 0.015075297094881535,
+ "learning_rate": 0.0006,
+ "loss": 4.102992534637451,
+ "step": 2545
+ },
+ {
+ "epoch": 35.36331877729258,
+ "grad_norm": 0.01538170501589775,
+ "learning_rate": 0.0006,
+ "loss": 4.161603927612305,
+ "step": 2546
+ },
+ {
+ "epoch": 35.377292576419215,
+ "grad_norm": 0.01483312901109457,
+ "learning_rate": 0.0006,
+ "loss": 4.085355758666992,
+ "step": 2547
+ },
+ {
+ "epoch": 35.391266375545854,
+ "grad_norm": 0.014894738793373108,
+ "learning_rate": 0.0006,
+ "loss": 4.083585262298584,
+ "step": 2548
+ },
+ {
+ "epoch": 35.40524017467249,
+ "grad_norm": 0.01475780550390482,
+ "learning_rate": 0.0006,
+ "loss": 4.012645721435547,
+ "step": 2549
+ },
+ {
+ "epoch": 35.419213973799124,
+ "grad_norm": 0.014531653374433517,
+ "learning_rate": 0.0006,
+ "loss": 4.197113990783691,
+ "step": 2550
+ },
+ {
+ "epoch": 35.43318777292576,
+ "grad_norm": 0.015316493809223175,
+ "learning_rate": 0.0006,
+ "loss": 4.227970123291016,
+ "step": 2551
+ },
+ {
+ "epoch": 35.4471615720524,
+ "grad_norm": 0.014685571193695068,
+ "learning_rate": 0.0006,
+ "loss": 4.089415550231934,
+ "step": 2552
+ },
+ {
+ "epoch": 35.46113537117904,
+ "grad_norm": 0.015337035991251469,
+ "learning_rate": 0.0006,
+ "loss": 4.080583572387695,
+ "step": 2553
+ },
+ {
+ "epoch": 35.47510917030568,
+ "grad_norm": 0.016754208132624626,
+ "learning_rate": 0.0006,
+ "loss": 4.163935661315918,
+ "step": 2554
+ },
+ {
+ "epoch": 35.48908296943232,
+ "grad_norm": 0.017535798251628876,
+ "learning_rate": 0.0006,
+ "loss": 4.093027591705322,
+ "step": 2555
+ },
+ {
+ "epoch": 35.50305676855895,
+ "grad_norm": 0.017336975783109665,
+ "learning_rate": 0.0006,
+ "loss": 4.127961158752441,
+ "step": 2556
+ },
+ {
+ "epoch": 35.51703056768559,
+ "grad_norm": 0.016242152079939842,
+ "learning_rate": 0.0006,
+ "loss": 4.155593395233154,
+ "step": 2557
+ },
+ {
+ "epoch": 35.531004366812226,
+ "grad_norm": 0.015121322125196457,
+ "learning_rate": 0.0006,
+ "loss": 3.984004020690918,
+ "step": 2558
+ },
+ {
+ "epoch": 35.544978165938865,
+ "grad_norm": 0.015178884379565716,
+ "learning_rate": 0.0006,
+ "loss": 4.111564636230469,
+ "step": 2559
+ },
+ {
+ "epoch": 35.5589519650655,
+ "grad_norm": 0.014777131378650665,
+ "learning_rate": 0.0006,
+ "loss": 4.041402816772461,
+ "step": 2560
+ },
+ {
+ "epoch": 35.57292576419214,
+ "grad_norm": 0.015197676606476307,
+ "learning_rate": 0.0006,
+ "loss": 4.133706092834473,
+ "step": 2561
+ },
+ {
+ "epoch": 35.58689956331878,
+ "grad_norm": 0.015905972570180893,
+ "learning_rate": 0.0006,
+ "loss": 4.1069135665893555,
+ "step": 2562
+ },
+ {
+ "epoch": 35.60087336244541,
+ "grad_norm": 0.015942877158522606,
+ "learning_rate": 0.0006,
+ "loss": 4.0720624923706055,
+ "step": 2563
+ },
+ {
+ "epoch": 35.61484716157205,
+ "grad_norm": 0.016697583720088005,
+ "learning_rate": 0.0006,
+ "loss": 4.167448043823242,
+ "step": 2564
+ },
+ {
+ "epoch": 35.62882096069869,
+ "grad_norm": 0.015603577718138695,
+ "learning_rate": 0.0006,
+ "loss": 4.127327919006348,
+ "step": 2565
+ },
+ {
+ "epoch": 35.64279475982533,
+ "grad_norm": 0.017153460532426834,
+ "learning_rate": 0.0006,
+ "loss": 4.056543350219727,
+ "step": 2566
+ },
+ {
+ "epoch": 35.65676855895197,
+ "grad_norm": 0.016731958836317062,
+ "learning_rate": 0.0006,
+ "loss": 4.0681843757629395,
+ "step": 2567
+ },
+ {
+ "epoch": 35.670742358078606,
+ "grad_norm": 0.016167359426617622,
+ "learning_rate": 0.0006,
+ "loss": 4.174992084503174,
+ "step": 2568
+ },
+ {
+ "epoch": 35.68471615720524,
+ "grad_norm": 0.01615108922123909,
+ "learning_rate": 0.0006,
+ "loss": 4.1383161544799805,
+ "step": 2569
+ },
+ {
+ "epoch": 35.698689956331876,
+ "grad_norm": 0.015750816091895103,
+ "learning_rate": 0.0006,
+ "loss": 4.031595706939697,
+ "step": 2570
+ },
+ {
+ "epoch": 35.712663755458514,
+ "grad_norm": 0.015416189096868038,
+ "learning_rate": 0.0006,
+ "loss": 4.067498207092285,
+ "step": 2571
+ },
+ {
+ "epoch": 35.72663755458515,
+ "grad_norm": 0.01570792682468891,
+ "learning_rate": 0.0006,
+ "loss": 4.1338300704956055,
+ "step": 2572
+ },
+ {
+ "epoch": 35.74061135371179,
+ "grad_norm": 0.016367752104997635,
+ "learning_rate": 0.0006,
+ "loss": 4.152523040771484,
+ "step": 2573
+ },
+ {
+ "epoch": 35.75458515283843,
+ "grad_norm": 0.016349803656339645,
+ "learning_rate": 0.0006,
+ "loss": 4.101985454559326,
+ "step": 2574
+ },
+ {
+ "epoch": 35.76855895196506,
+ "grad_norm": 0.01663660630583763,
+ "learning_rate": 0.0006,
+ "loss": 4.190130710601807,
+ "step": 2575
+ },
+ {
+ "epoch": 35.7825327510917,
+ "grad_norm": 0.015599170699715614,
+ "learning_rate": 0.0006,
+ "loss": 4.0792436599731445,
+ "step": 2576
+ },
+ {
+ "epoch": 35.79650655021834,
+ "grad_norm": 0.015472427010536194,
+ "learning_rate": 0.0006,
+ "loss": 4.186446666717529,
+ "step": 2577
+ },
+ {
+ "epoch": 35.81048034934498,
+ "grad_norm": 0.015520439483225346,
+ "learning_rate": 0.0006,
+ "loss": 4.246635437011719,
+ "step": 2578
+ },
+ {
+ "epoch": 35.82445414847162,
+ "grad_norm": 0.016998281702399254,
+ "learning_rate": 0.0006,
+ "loss": 4.13435173034668,
+ "step": 2579
+ },
+ {
+ "epoch": 35.838427947598255,
+ "grad_norm": 0.016290361061692238,
+ "learning_rate": 0.0006,
+ "loss": 4.107428073883057,
+ "step": 2580
+ },
+ {
+ "epoch": 35.852401746724894,
+ "grad_norm": 0.0165516659617424,
+ "learning_rate": 0.0006,
+ "loss": 4.1612138748168945,
+ "step": 2581
+ },
+ {
+ "epoch": 35.866375545851525,
+ "grad_norm": 0.01736091636121273,
+ "learning_rate": 0.0006,
+ "loss": 4.1072869300842285,
+ "step": 2582
+ },
+ {
+ "epoch": 35.880349344978164,
+ "grad_norm": 0.01787031814455986,
+ "learning_rate": 0.0006,
+ "loss": 4.1076812744140625,
+ "step": 2583
+ },
+ {
+ "epoch": 35.8943231441048,
+ "grad_norm": 0.015437985770404339,
+ "learning_rate": 0.0006,
+ "loss": 4.097488880157471,
+ "step": 2584
+ },
+ {
+ "epoch": 35.90829694323144,
+ "grad_norm": 0.016306741163134575,
+ "learning_rate": 0.0006,
+ "loss": 4.105017185211182,
+ "step": 2585
+ },
+ {
+ "epoch": 35.92227074235808,
+ "grad_norm": 0.015723245218396187,
+ "learning_rate": 0.0006,
+ "loss": 4.205283164978027,
+ "step": 2586
+ },
+ {
+ "epoch": 35.93624454148472,
+ "grad_norm": 0.014626456424593925,
+ "learning_rate": 0.0006,
+ "loss": 4.086552619934082,
+ "step": 2587
+ },
+ {
+ "epoch": 35.95021834061135,
+ "grad_norm": 0.014348454773426056,
+ "learning_rate": 0.0006,
+ "loss": 4.171329975128174,
+ "step": 2588
+ },
+ {
+ "epoch": 35.96419213973799,
+ "grad_norm": 0.015598480589687824,
+ "learning_rate": 0.0006,
+ "loss": 4.140815258026123,
+ "step": 2589
+ },
+ {
+ "epoch": 35.97816593886463,
+ "grad_norm": 0.017126062884926796,
+ "learning_rate": 0.0006,
+ "loss": 4.157176971435547,
+ "step": 2590
+ },
+ {
+ "epoch": 35.992139737991266,
+ "grad_norm": 0.01860199309885502,
+ "learning_rate": 0.0006,
+ "loss": 4.252880096435547,
+ "step": 2591
+ },
+ {
+ "epoch": 36.0,
+ "grad_norm": 0.021780220791697502,
+ "learning_rate": 0.0006,
+ "loss": 4.249861717224121,
+ "step": 2592
+ },
+ {
+ "epoch": 36.0,
+ "eval_loss": 4.606466293334961,
+ "eval_runtime": 56.6217,
+ "eval_samples_per_second": 43.128,
+ "eval_steps_per_second": 1.36,
+ "step": 2592
+ },
+ {
+ "epoch": 36.01397379912664,
+ "grad_norm": 0.020892828702926636,
+ "learning_rate": 0.0006,
+ "loss": 4.102055549621582,
+ "step": 2593
+ },
+ {
+ "epoch": 36.02794759825328,
+ "grad_norm": 0.020306255668401718,
+ "learning_rate": 0.0006,
+ "loss": 4.184308052062988,
+ "step": 2594
+ },
+ {
+ "epoch": 36.041921397379916,
+ "grad_norm": 0.02025097794830799,
+ "learning_rate": 0.0006,
+ "loss": 3.9691858291625977,
+ "step": 2595
+ },
+ {
+ "epoch": 36.05589519650655,
+ "grad_norm": 0.020254382863640785,
+ "learning_rate": 0.0006,
+ "loss": 4.02957820892334,
+ "step": 2596
+ },
+ {
+ "epoch": 36.069868995633186,
+ "grad_norm": 0.019307196140289307,
+ "learning_rate": 0.0006,
+ "loss": 4.033181667327881,
+ "step": 2597
+ },
+ {
+ "epoch": 36.083842794759825,
+ "grad_norm": 0.0198476854711771,
+ "learning_rate": 0.0006,
+ "loss": 4.0816874504089355,
+ "step": 2598
+ },
+ {
+ "epoch": 36.09781659388646,
+ "grad_norm": 0.019916515797376633,
+ "learning_rate": 0.0006,
+ "loss": 4.164889812469482,
+ "step": 2599
+ },
+ {
+ "epoch": 36.1117903930131,
+ "grad_norm": 0.0189144778996706,
+ "learning_rate": 0.0006,
+ "loss": 4.038427352905273,
+ "step": 2600
+ },
+ {
+ "epoch": 36.12576419213974,
+ "grad_norm": 0.01754768192768097,
+ "learning_rate": 0.0006,
+ "loss": 4.010427474975586,
+ "step": 2601
+ },
+ {
+ "epoch": 36.13973799126637,
+ "grad_norm": 0.016862990334630013,
+ "learning_rate": 0.0006,
+ "loss": 4.035970211029053,
+ "step": 2602
+ },
+ {
+ "epoch": 36.15371179039301,
+ "grad_norm": 0.019616752862930298,
+ "learning_rate": 0.0006,
+ "loss": 4.079898834228516,
+ "step": 2603
+ },
+ {
+ "epoch": 36.16768558951965,
+ "grad_norm": 0.020131465047597885,
+ "learning_rate": 0.0006,
+ "loss": 4.144684791564941,
+ "step": 2604
+ },
+ {
+ "epoch": 36.18165938864629,
+ "grad_norm": 0.02072521299123764,
+ "learning_rate": 0.0006,
+ "loss": 4.013660430908203,
+ "step": 2605
+ },
+ {
+ "epoch": 36.19563318777293,
+ "grad_norm": 0.01924615167081356,
+ "learning_rate": 0.0006,
+ "loss": 4.071398735046387,
+ "step": 2606
+ },
+ {
+ "epoch": 36.209606986899566,
+ "grad_norm": 0.02013954520225525,
+ "learning_rate": 0.0006,
+ "loss": 4.127094268798828,
+ "step": 2607
+ },
+ {
+ "epoch": 36.223580786026204,
+ "grad_norm": 0.02042008936405182,
+ "learning_rate": 0.0006,
+ "loss": 4.154261589050293,
+ "step": 2608
+ },
+ {
+ "epoch": 36.237554585152836,
+ "grad_norm": 0.021998876705765724,
+ "learning_rate": 0.0006,
+ "loss": 4.204150199890137,
+ "step": 2609
+ },
+ {
+ "epoch": 36.251528384279474,
+ "grad_norm": 0.022946683689951897,
+ "learning_rate": 0.0006,
+ "loss": 4.013632774353027,
+ "step": 2610
+ },
+ {
+ "epoch": 36.26550218340611,
+ "grad_norm": 0.02061830274760723,
+ "learning_rate": 0.0006,
+ "loss": 4.171084403991699,
+ "step": 2611
+ },
+ {
+ "epoch": 36.27947598253275,
+ "grad_norm": 0.01883152686059475,
+ "learning_rate": 0.0006,
+ "loss": 4.149848461151123,
+ "step": 2612
+ },
+ {
+ "epoch": 36.29344978165939,
+ "grad_norm": 0.019945282489061356,
+ "learning_rate": 0.0006,
+ "loss": 4.174656867980957,
+ "step": 2613
+ },
+ {
+ "epoch": 36.30742358078603,
+ "grad_norm": 0.020751051604747772,
+ "learning_rate": 0.0006,
+ "loss": 4.080747127532959,
+ "step": 2614
+ },
+ {
+ "epoch": 36.32139737991266,
+ "grad_norm": 0.01879691146314144,
+ "learning_rate": 0.0006,
+ "loss": 4.057356357574463,
+ "step": 2615
+ },
+ {
+ "epoch": 36.3353711790393,
+ "grad_norm": 0.019037071615457535,
+ "learning_rate": 0.0006,
+ "loss": 4.204607009887695,
+ "step": 2616
+ },
+ {
+ "epoch": 36.34934497816594,
+ "grad_norm": 0.01872078701853752,
+ "learning_rate": 0.0006,
+ "loss": 4.050586700439453,
+ "step": 2617
+ },
+ {
+ "epoch": 36.36331877729258,
+ "grad_norm": 0.018092811107635498,
+ "learning_rate": 0.0006,
+ "loss": 4.102832794189453,
+ "step": 2618
+ },
+ {
+ "epoch": 36.377292576419215,
+ "grad_norm": 0.017010487616062164,
+ "learning_rate": 0.0006,
+ "loss": 4.191788673400879,
+ "step": 2619
+ },
+ {
+ "epoch": 36.391266375545854,
+ "grad_norm": 0.01691143773496151,
+ "learning_rate": 0.0006,
+ "loss": 4.160456657409668,
+ "step": 2620
+ },
+ {
+ "epoch": 36.40524017467249,
+ "grad_norm": 0.017046816647052765,
+ "learning_rate": 0.0006,
+ "loss": 4.073828220367432,
+ "step": 2621
+ },
+ {
+ "epoch": 36.419213973799124,
+ "grad_norm": 0.01707981713116169,
+ "learning_rate": 0.0006,
+ "loss": 4.0980095863342285,
+ "step": 2622
+ },
+ {
+ "epoch": 36.43318777292576,
+ "grad_norm": 0.01630452275276184,
+ "learning_rate": 0.0006,
+ "loss": 4.101321220397949,
+ "step": 2623
+ },
+ {
+ "epoch": 36.4471615720524,
+ "grad_norm": 0.01576915942132473,
+ "learning_rate": 0.0006,
+ "loss": 4.158249855041504,
+ "step": 2624
+ },
+ {
+ "epoch": 36.46113537117904,
+ "grad_norm": 0.016179712489247322,
+ "learning_rate": 0.0006,
+ "loss": 4.077852249145508,
+ "step": 2625
+ },
+ {
+ "epoch": 36.47510917030568,
+ "grad_norm": 0.01609169878065586,
+ "learning_rate": 0.0006,
+ "loss": 4.187317848205566,
+ "step": 2626
+ },
+ {
+ "epoch": 36.48908296943232,
+ "grad_norm": 0.017148254439234734,
+ "learning_rate": 0.0006,
+ "loss": 4.109071731567383,
+ "step": 2627
+ },
+ {
+ "epoch": 36.50305676855895,
+ "grad_norm": 0.015842624008655548,
+ "learning_rate": 0.0006,
+ "loss": 4.145096302032471,
+ "step": 2628
+ },
+ {
+ "epoch": 36.51703056768559,
+ "grad_norm": 0.014906194061040878,
+ "learning_rate": 0.0006,
+ "loss": 4.028440475463867,
+ "step": 2629
+ },
+ {
+ "epoch": 36.531004366812226,
+ "grad_norm": 0.01485038734972477,
+ "learning_rate": 0.0006,
+ "loss": 4.175201416015625,
+ "step": 2630
+ },
+ {
+ "epoch": 36.544978165938865,
+ "grad_norm": 0.015834150835871696,
+ "learning_rate": 0.0006,
+ "loss": 4.076434135437012,
+ "step": 2631
+ },
+ {
+ "epoch": 36.5589519650655,
+ "grad_norm": 0.015411358326673508,
+ "learning_rate": 0.0006,
+ "loss": 4.03912878036499,
+ "step": 2632
+ },
+ {
+ "epoch": 36.57292576419214,
+ "grad_norm": 0.014762846753001213,
+ "learning_rate": 0.0006,
+ "loss": 4.019827365875244,
+ "step": 2633
+ },
+ {
+ "epoch": 36.58689956331878,
+ "grad_norm": 0.015269497409462929,
+ "learning_rate": 0.0006,
+ "loss": 4.116957187652588,
+ "step": 2634
+ },
+ {
+ "epoch": 36.60087336244541,
+ "grad_norm": 0.015171617269515991,
+ "learning_rate": 0.0006,
+ "loss": 4.2069926261901855,
+ "step": 2635
+ },
+ {
+ "epoch": 36.61484716157205,
+ "grad_norm": 0.014770902693271637,
+ "learning_rate": 0.0006,
+ "loss": 4.092021942138672,
+ "step": 2636
+ },
+ {
+ "epoch": 36.62882096069869,
+ "grad_norm": 0.014957531355321407,
+ "learning_rate": 0.0006,
+ "loss": 4.16819953918457,
+ "step": 2637
+ },
+ {
+ "epoch": 36.64279475982533,
+ "grad_norm": 0.014774451963603497,
+ "learning_rate": 0.0006,
+ "loss": 4.046073913574219,
+ "step": 2638
+ },
+ {
+ "epoch": 36.65676855895197,
+ "grad_norm": 0.014320445246994495,
+ "learning_rate": 0.0006,
+ "loss": 3.977792501449585,
+ "step": 2639
+ },
+ {
+ "epoch": 36.670742358078606,
+ "grad_norm": 0.01413130946457386,
+ "learning_rate": 0.0006,
+ "loss": 4.130649566650391,
+ "step": 2640
+ },
+ {
+ "epoch": 36.68471615720524,
+ "grad_norm": 0.014336307533085346,
+ "learning_rate": 0.0006,
+ "loss": 4.161357402801514,
+ "step": 2641
+ },
+ {
+ "epoch": 36.698689956331876,
+ "grad_norm": 0.014516931027173996,
+ "learning_rate": 0.0006,
+ "loss": 4.14248514175415,
+ "step": 2642
+ },
+ {
+ "epoch": 36.712663755458514,
+ "grad_norm": 0.014377126470208168,
+ "learning_rate": 0.0006,
+ "loss": 4.062711715698242,
+ "step": 2643
+ },
+ {
+ "epoch": 36.72663755458515,
+ "grad_norm": 0.014332668855786324,
+ "learning_rate": 0.0006,
+ "loss": 4.06423807144165,
+ "step": 2644
+ },
+ {
+ "epoch": 36.74061135371179,
+ "grad_norm": 0.014030571095645428,
+ "learning_rate": 0.0006,
+ "loss": 4.178613662719727,
+ "step": 2645
+ },
+ {
+ "epoch": 36.75458515283843,
+ "grad_norm": 0.014376776292920113,
+ "learning_rate": 0.0006,
+ "loss": 4.073130130767822,
+ "step": 2646
+ },
+ {
+ "epoch": 36.76855895196506,
+ "grad_norm": 0.013827716931700706,
+ "learning_rate": 0.0006,
+ "loss": 4.043869972229004,
+ "step": 2647
+ },
+ {
+ "epoch": 36.7825327510917,
+ "grad_norm": 0.015548762865364552,
+ "learning_rate": 0.0006,
+ "loss": 4.134049415588379,
+ "step": 2648
+ },
+ {
+ "epoch": 36.79650655021834,
+ "grad_norm": 0.017041970044374466,
+ "learning_rate": 0.0006,
+ "loss": 4.205024242401123,
+ "step": 2649
+ },
+ {
+ "epoch": 36.81048034934498,
+ "grad_norm": 0.017638275399804115,
+ "learning_rate": 0.0006,
+ "loss": 4.199924468994141,
+ "step": 2650
+ },
+ {
+ "epoch": 36.82445414847162,
+ "grad_norm": 0.017771627753973007,
+ "learning_rate": 0.0006,
+ "loss": 4.15706729888916,
+ "step": 2651
+ },
+ {
+ "epoch": 36.838427947598255,
+ "grad_norm": 0.015933562070131302,
+ "learning_rate": 0.0006,
+ "loss": 4.220742225646973,
+ "step": 2652
+ },
+ {
+ "epoch": 36.852401746724894,
+ "grad_norm": 0.01668144389986992,
+ "learning_rate": 0.0006,
+ "loss": 4.161426544189453,
+ "step": 2653
+ },
+ {
+ "epoch": 36.866375545851525,
+ "grad_norm": 0.018014680594205856,
+ "learning_rate": 0.0006,
+ "loss": 4.183046817779541,
+ "step": 2654
+ },
+ {
+ "epoch": 36.880349344978164,
+ "grad_norm": 0.018808670341968536,
+ "learning_rate": 0.0006,
+ "loss": 4.105293273925781,
+ "step": 2655
+ },
+ {
+ "epoch": 36.8943231441048,
+ "grad_norm": 0.018003476783633232,
+ "learning_rate": 0.0006,
+ "loss": 4.103171348571777,
+ "step": 2656
+ },
+ {
+ "epoch": 36.90829694323144,
+ "grad_norm": 0.016865914687514305,
+ "learning_rate": 0.0006,
+ "loss": 4.125634670257568,
+ "step": 2657
+ },
+ {
+ "epoch": 36.92227074235808,
+ "grad_norm": 0.018371153622865677,
+ "learning_rate": 0.0006,
+ "loss": 4.192461013793945,
+ "step": 2658
+ },
+ {
+ "epoch": 36.93624454148472,
+ "grad_norm": 0.019045377150177956,
+ "learning_rate": 0.0006,
+ "loss": 4.198105812072754,
+ "step": 2659
+ },
+ {
+ "epoch": 36.95021834061135,
+ "grad_norm": 0.01594168320298195,
+ "learning_rate": 0.0006,
+ "loss": 4.169788360595703,
+ "step": 2660
+ },
+ {
+ "epoch": 36.96419213973799,
+ "grad_norm": 0.015645820647478104,
+ "learning_rate": 0.0006,
+ "loss": 4.047277450561523,
+ "step": 2661
+ },
+ {
+ "epoch": 36.97816593886463,
+ "grad_norm": 0.016313156113028526,
+ "learning_rate": 0.0006,
+ "loss": 4.087392807006836,
+ "step": 2662
+ },
+ {
+ "epoch": 36.992139737991266,
+ "grad_norm": 0.017050625756382942,
+ "learning_rate": 0.0006,
+ "loss": 4.0908002853393555,
+ "step": 2663
+ },
+ {
+ "epoch": 37.0,
+ "grad_norm": 0.017730968073010445,
+ "learning_rate": 0.0006,
+ "loss": 4.110957622528076,
+ "step": 2664
+ },
+ {
+ "epoch": 37.0,
+ "eval_loss": 4.563793182373047,
+ "eval_runtime": 56.5922,
+ "eval_samples_per_second": 43.151,
+ "eval_steps_per_second": 1.361,
+ "step": 2664
+ },
+ {
+ "epoch": 37.01397379912664,
+ "grad_norm": 0.015687910839915276,
+ "learning_rate": 0.0006,
+ "loss": 4.143589019775391,
+ "step": 2665
+ },
+ {
+ "epoch": 37.02794759825328,
+ "grad_norm": 0.017451366409659386,
+ "learning_rate": 0.0006,
+ "loss": 4.0326666831970215,
+ "step": 2666
+ },
+ {
+ "epoch": 37.041921397379916,
+ "grad_norm": 0.01909538544714451,
+ "learning_rate": 0.0006,
+ "loss": 3.9706249237060547,
+ "step": 2667
+ },
+ {
+ "epoch": 37.05589519650655,
+ "grad_norm": 0.01988379657268524,
+ "learning_rate": 0.0006,
+ "loss": 3.9531009197235107,
+ "step": 2668
+ },
+ {
+ "epoch": 37.069868995633186,
+ "grad_norm": 0.02035115472972393,
+ "learning_rate": 0.0006,
+ "loss": 3.962214469909668,
+ "step": 2669
+ },
+ {
+ "epoch": 37.083842794759825,
+ "grad_norm": 0.01952001079916954,
+ "learning_rate": 0.0006,
+ "loss": 4.182262897491455,
+ "step": 2670
+ },
+ {
+ "epoch": 37.09781659388646,
+ "grad_norm": 0.01854325644671917,
+ "learning_rate": 0.0006,
+ "loss": 4.127663612365723,
+ "step": 2671
+ },
+ {
+ "epoch": 37.1117903930131,
+ "grad_norm": 0.017757616937160492,
+ "learning_rate": 0.0006,
+ "loss": 4.100804805755615,
+ "step": 2672
+ },
+ {
+ "epoch": 37.12576419213974,
+ "grad_norm": 0.016712479293346405,
+ "learning_rate": 0.0006,
+ "loss": 4.026284217834473,
+ "step": 2673
+ },
+ {
+ "epoch": 37.13973799126637,
+ "grad_norm": 0.0168139711022377,
+ "learning_rate": 0.0006,
+ "loss": 4.183882713317871,
+ "step": 2674
+ },
+ {
+ "epoch": 37.15371179039301,
+ "grad_norm": 0.017060423269867897,
+ "learning_rate": 0.0006,
+ "loss": 4.131832599639893,
+ "step": 2675
+ },
+ {
+ "epoch": 37.16768558951965,
+ "grad_norm": 0.01726631447672844,
+ "learning_rate": 0.0006,
+ "loss": 3.9156057834625244,
+ "step": 2676
+ },
+ {
+ "epoch": 37.18165938864629,
+ "grad_norm": 0.016348622739315033,
+ "learning_rate": 0.0006,
+ "loss": 4.1731367111206055,
+ "step": 2677
+ },
+ {
+ "epoch": 37.19563318777293,
+ "grad_norm": 0.016640059649944305,
+ "learning_rate": 0.0006,
+ "loss": 4.10997200012207,
+ "step": 2678
+ },
+ {
+ "epoch": 37.209606986899566,
+ "grad_norm": 0.01601342298090458,
+ "learning_rate": 0.0006,
+ "loss": 4.081145286560059,
+ "step": 2679
+ },
+ {
+ "epoch": 37.223580786026204,
+ "grad_norm": 0.01735631190240383,
+ "learning_rate": 0.0006,
+ "loss": 4.202108860015869,
+ "step": 2680
+ },
+ {
+ "epoch": 37.237554585152836,
+ "grad_norm": 0.019164685159921646,
+ "learning_rate": 0.0006,
+ "loss": 4.104109764099121,
+ "step": 2681
+ },
+ {
+ "epoch": 37.251528384279474,
+ "grad_norm": 0.01881759613752365,
+ "learning_rate": 0.0006,
+ "loss": 4.145532131195068,
+ "step": 2682
+ },
+ {
+ "epoch": 37.26550218340611,
+ "grad_norm": 0.017656555399298668,
+ "learning_rate": 0.0006,
+ "loss": 4.012782096862793,
+ "step": 2683
+ },
+ {
+ "epoch": 37.27947598253275,
+ "grad_norm": 0.01624823734164238,
+ "learning_rate": 0.0006,
+ "loss": 4.001727104187012,
+ "step": 2684
+ },
+ {
+ "epoch": 37.29344978165939,
+ "grad_norm": 0.016973139718174934,
+ "learning_rate": 0.0006,
+ "loss": 4.095399856567383,
+ "step": 2685
+ },
+ {
+ "epoch": 37.30742358078603,
+ "grad_norm": 0.015503671951591969,
+ "learning_rate": 0.0006,
+ "loss": 4.034431457519531,
+ "step": 2686
+ },
+ {
+ "epoch": 37.32139737991266,
+ "grad_norm": 0.015370250679552555,
+ "learning_rate": 0.0006,
+ "loss": 4.075836181640625,
+ "step": 2687
+ },
+ {
+ "epoch": 37.3353711790393,
+ "grad_norm": 0.01564021222293377,
+ "learning_rate": 0.0006,
+ "loss": 4.125202178955078,
+ "step": 2688
+ },
+ {
+ "epoch": 37.34934497816594,
+ "grad_norm": 0.016717975959181786,
+ "learning_rate": 0.0006,
+ "loss": 4.153017044067383,
+ "step": 2689
+ },
+ {
+ "epoch": 37.36331877729258,
+ "grad_norm": 0.015263617038726807,
+ "learning_rate": 0.0006,
+ "loss": 4.021814823150635,
+ "step": 2690
+ },
+ {
+ "epoch": 37.377292576419215,
+ "grad_norm": 0.016497142612934113,
+ "learning_rate": 0.0006,
+ "loss": 4.069095611572266,
+ "step": 2691
+ },
+ {
+ "epoch": 37.391266375545854,
+ "grad_norm": 0.01669355481863022,
+ "learning_rate": 0.0006,
+ "loss": 4.174587726593018,
+ "step": 2692
+ },
+ {
+ "epoch": 37.40524017467249,
+ "grad_norm": 0.019695773720741272,
+ "learning_rate": 0.0006,
+ "loss": 4.0073041915893555,
+ "step": 2693
+ },
+ {
+ "epoch": 37.419213973799124,
+ "grad_norm": 0.023720644414424896,
+ "learning_rate": 0.0006,
+ "loss": 4.111948013305664,
+ "step": 2694
+ },
+ {
+ "epoch": 37.43318777292576,
+ "grad_norm": 0.02399350143969059,
+ "learning_rate": 0.0006,
+ "loss": 4.081164360046387,
+ "step": 2695
+ },
+ {
+ "epoch": 37.4471615720524,
+ "grad_norm": 0.019630271941423416,
+ "learning_rate": 0.0006,
+ "loss": 4.065731048583984,
+ "step": 2696
+ },
+ {
+ "epoch": 37.46113537117904,
+ "grad_norm": 0.019679008051753044,
+ "learning_rate": 0.0006,
+ "loss": 3.993852376937866,
+ "step": 2697
+ },
+ {
+ "epoch": 37.47510917030568,
+ "grad_norm": 0.020597299560904503,
+ "learning_rate": 0.0006,
+ "loss": 4.275555610656738,
+ "step": 2698
+ },
+ {
+ "epoch": 37.48908296943232,
+ "grad_norm": 0.01983889564871788,
+ "learning_rate": 0.0006,
+ "loss": 4.254708290100098,
+ "step": 2699
+ },
+ {
+ "epoch": 37.50305676855895,
+ "grad_norm": 0.019150668755173683,
+ "learning_rate": 0.0006,
+ "loss": 4.097885608673096,
+ "step": 2700
+ },
+ {
+ "epoch": 37.51703056768559,
+ "grad_norm": 0.018847715109586716,
+ "learning_rate": 0.0006,
+ "loss": 4.054081439971924,
+ "step": 2701
+ },
+ {
+ "epoch": 37.531004366812226,
+ "grad_norm": 0.019607288762927055,
+ "learning_rate": 0.0006,
+ "loss": 3.9891440868377686,
+ "step": 2702
+ },
+ {
+ "epoch": 37.544978165938865,
+ "grad_norm": 0.018863150849938393,
+ "learning_rate": 0.0006,
+ "loss": 4.084877014160156,
+ "step": 2703
+ },
+ {
+ "epoch": 37.5589519650655,
+ "grad_norm": 0.017340239137411118,
+ "learning_rate": 0.0006,
+ "loss": 4.113290786743164,
+ "step": 2704
+ },
+ {
+ "epoch": 37.57292576419214,
+ "grad_norm": 0.015628693625330925,
+ "learning_rate": 0.0006,
+ "loss": 4.1483001708984375,
+ "step": 2705
+ },
+ {
+ "epoch": 37.58689956331878,
+ "grad_norm": 0.016493283212184906,
+ "learning_rate": 0.0006,
+ "loss": 4.1040143966674805,
+ "step": 2706
+ },
+ {
+ "epoch": 37.60087336244541,
+ "grad_norm": 0.016732443124055862,
+ "learning_rate": 0.0006,
+ "loss": 4.07205057144165,
+ "step": 2707
+ },
+ {
+ "epoch": 37.61484716157205,
+ "grad_norm": 0.017176156863570213,
+ "learning_rate": 0.0006,
+ "loss": 4.1279754638671875,
+ "step": 2708
+ },
+ {
+ "epoch": 37.62882096069869,
+ "grad_norm": 0.015730608254671097,
+ "learning_rate": 0.0006,
+ "loss": 4.086658000946045,
+ "step": 2709
+ },
+ {
+ "epoch": 37.64279475982533,
+ "grad_norm": 0.016111090779304504,
+ "learning_rate": 0.0006,
+ "loss": 4.064335823059082,
+ "step": 2710
+ },
+ {
+ "epoch": 37.65676855895197,
+ "grad_norm": 0.01769379712641239,
+ "learning_rate": 0.0006,
+ "loss": 4.116047382354736,
+ "step": 2711
+ },
+ {
+ "epoch": 37.670742358078606,
+ "grad_norm": 0.01683102548122406,
+ "learning_rate": 0.0006,
+ "loss": 4.095986843109131,
+ "step": 2712
+ },
+ {
+ "epoch": 37.68471615720524,
+ "grad_norm": 0.015726450830698013,
+ "learning_rate": 0.0006,
+ "loss": 4.10775089263916,
+ "step": 2713
+ },
+ {
+ "epoch": 37.698689956331876,
+ "grad_norm": 0.014760235324501991,
+ "learning_rate": 0.0006,
+ "loss": 4.1234893798828125,
+ "step": 2714
+ },
+ {
+ "epoch": 37.712663755458514,
+ "grad_norm": 0.014933438040316105,
+ "learning_rate": 0.0006,
+ "loss": 4.079200744628906,
+ "step": 2715
+ },
+ {
+ "epoch": 37.72663755458515,
+ "grad_norm": 0.015723997727036476,
+ "learning_rate": 0.0006,
+ "loss": 4.07865047454834,
+ "step": 2716
+ },
+ {
+ "epoch": 37.74061135371179,
+ "grad_norm": 0.014539001509547234,
+ "learning_rate": 0.0006,
+ "loss": 4.036242485046387,
+ "step": 2717
+ },
+ {
+ "epoch": 37.75458515283843,
+ "grad_norm": 0.014844979159533978,
+ "learning_rate": 0.0006,
+ "loss": 4.1051483154296875,
+ "step": 2718
+ },
+ {
+ "epoch": 37.76855895196506,
+ "grad_norm": 0.015939844772219658,
+ "learning_rate": 0.0006,
+ "loss": 4.017253875732422,
+ "step": 2719
+ },
+ {
+ "epoch": 37.7825327510917,
+ "grad_norm": 0.014771555550396442,
+ "learning_rate": 0.0006,
+ "loss": 4.0694661140441895,
+ "step": 2720
+ },
+ {
+ "epoch": 37.79650655021834,
+ "grad_norm": 0.015777921304106712,
+ "learning_rate": 0.0006,
+ "loss": 3.947587490081787,
+ "step": 2721
+ },
+ {
+ "epoch": 37.81048034934498,
+ "grad_norm": 0.017451291903853416,
+ "learning_rate": 0.0006,
+ "loss": 4.093625068664551,
+ "step": 2722
+ },
+ {
+ "epoch": 37.82445414847162,
+ "grad_norm": 0.016829246655106544,
+ "learning_rate": 0.0006,
+ "loss": 4.09030818939209,
+ "step": 2723
+ },
+ {
+ "epoch": 37.838427947598255,
+ "grad_norm": 0.016984397545456886,
+ "learning_rate": 0.0006,
+ "loss": 4.140626907348633,
+ "step": 2724
+ },
+ {
+ "epoch": 37.852401746724894,
+ "grad_norm": 0.01732603646814823,
+ "learning_rate": 0.0006,
+ "loss": 4.128276824951172,
+ "step": 2725
+ },
+ {
+ "epoch": 37.866375545851525,
+ "grad_norm": 0.01744619384407997,
+ "learning_rate": 0.0006,
+ "loss": 4.103130340576172,
+ "step": 2726
+ },
+ {
+ "epoch": 37.880349344978164,
+ "grad_norm": 0.016331320628523827,
+ "learning_rate": 0.0006,
+ "loss": 4.037951469421387,
+ "step": 2727
+ },
+ {
+ "epoch": 37.8943231441048,
+ "grad_norm": 0.015919430181384087,
+ "learning_rate": 0.0006,
+ "loss": 4.121513366699219,
+ "step": 2728
+ },
+ {
+ "epoch": 37.90829694323144,
+ "grad_norm": 0.016153695061802864,
+ "learning_rate": 0.0006,
+ "loss": 4.099384307861328,
+ "step": 2729
+ },
+ {
+ "epoch": 37.92227074235808,
+ "grad_norm": 0.01628800667822361,
+ "learning_rate": 0.0006,
+ "loss": 4.092325687408447,
+ "step": 2730
+ },
+ {
+ "epoch": 37.93624454148472,
+ "grad_norm": 0.017285684123635292,
+ "learning_rate": 0.0006,
+ "loss": 3.9828906059265137,
+ "step": 2731
+ },
+ {
+ "epoch": 37.95021834061135,
+ "grad_norm": 0.016047554090619087,
+ "learning_rate": 0.0006,
+ "loss": 4.100419044494629,
+ "step": 2732
+ },
+ {
+ "epoch": 37.96419213973799,
+ "grad_norm": 0.01562986709177494,
+ "learning_rate": 0.0006,
+ "loss": 4.065832614898682,
+ "step": 2733
+ },
+ {
+ "epoch": 37.97816593886463,
+ "grad_norm": 0.01677163504064083,
+ "learning_rate": 0.0006,
+ "loss": 4.077518463134766,
+ "step": 2734
+ },
+ {
+ "epoch": 37.992139737991266,
+ "grad_norm": 0.01755492016673088,
+ "learning_rate": 0.0006,
+ "loss": 4.1076154708862305,
+ "step": 2735
+ },
+ {
+ "epoch": 38.0,
+ "grad_norm": 0.01929762214422226,
+ "learning_rate": 0.0006,
+ "loss": 4.09603214263916,
+ "step": 2736
+ },
+ {
+ "epoch": 38.0,
+ "eval_loss": 4.582442283630371,
+ "eval_runtime": 62.8649,
+ "eval_samples_per_second": 38.845,
+ "eval_steps_per_second": 1.225,
+ "step": 2736
+ },
+ {
+ "epoch": 38.01397379912664,
+ "grad_norm": 0.019129136577248573,
+ "learning_rate": 0.0006,
+ "loss": 4.095005989074707,
+ "step": 2737
+ },
+ {
+ "epoch": 38.02794759825328,
+ "grad_norm": 0.019167408347129822,
+ "learning_rate": 0.0006,
+ "loss": 4.054426670074463,
+ "step": 2738
+ },
+ {
+ "epoch": 38.041921397379916,
+ "grad_norm": 0.01789460889995098,
+ "learning_rate": 0.0006,
+ "loss": 4.0434417724609375,
+ "step": 2739
+ },
+ {
+ "epoch": 38.05589519650655,
+ "grad_norm": 0.015947027131915092,
+ "learning_rate": 0.0006,
+ "loss": 4.074516296386719,
+ "step": 2740
+ },
+ {
+ "epoch": 38.069868995633186,
+ "grad_norm": 0.016667252406477928,
+ "learning_rate": 0.0006,
+ "loss": 4.108545303344727,
+ "step": 2741
+ },
+ {
+ "epoch": 38.083842794759825,
+ "grad_norm": 0.016014399006962776,
+ "learning_rate": 0.0006,
+ "loss": 4.026925086975098,
+ "step": 2742
+ },
+ {
+ "epoch": 38.09781659388646,
+ "grad_norm": 0.015397695824503899,
+ "learning_rate": 0.0006,
+ "loss": 4.163970470428467,
+ "step": 2743
+ },
+ {
+ "epoch": 38.1117903930131,
+ "grad_norm": 0.01563289947807789,
+ "learning_rate": 0.0006,
+ "loss": 4.116578578948975,
+ "step": 2744
+ },
+ {
+ "epoch": 38.12576419213974,
+ "grad_norm": 0.015512318350374699,
+ "learning_rate": 0.0006,
+ "loss": 4.096311569213867,
+ "step": 2745
+ },
+ {
+ "epoch": 38.13973799126637,
+ "grad_norm": 0.015426642261445522,
+ "learning_rate": 0.0006,
+ "loss": 4.017065048217773,
+ "step": 2746
+ },
+ {
+ "epoch": 38.15371179039301,
+ "grad_norm": 0.015115435235202312,
+ "learning_rate": 0.0006,
+ "loss": 4.005434989929199,
+ "step": 2747
+ },
+ {
+ "epoch": 38.16768558951965,
+ "grad_norm": 0.015235894359648228,
+ "learning_rate": 0.0006,
+ "loss": 3.982743263244629,
+ "step": 2748
+ },
+ {
+ "epoch": 38.18165938864629,
+ "grad_norm": 0.016095101833343506,
+ "learning_rate": 0.0006,
+ "loss": 4.0849385261535645,
+ "step": 2749
+ },
+ {
+ "epoch": 38.19563318777293,
+ "grad_norm": 0.015201197937130928,
+ "learning_rate": 0.0006,
+ "loss": 4.103863716125488,
+ "step": 2750
+ },
+ {
+ "epoch": 38.209606986899566,
+ "grad_norm": 0.015687478706240654,
+ "learning_rate": 0.0006,
+ "loss": 4.12074089050293,
+ "step": 2751
+ },
+ {
+ "epoch": 38.223580786026204,
+ "grad_norm": 0.015201589092612267,
+ "learning_rate": 0.0006,
+ "loss": 4.056872367858887,
+ "step": 2752
+ },
+ {
+ "epoch": 38.237554585152836,
+ "grad_norm": 0.015154894441366196,
+ "learning_rate": 0.0006,
+ "loss": 4.089670181274414,
+ "step": 2753
+ },
+ {
+ "epoch": 38.251528384279474,
+ "grad_norm": 0.01586608588695526,
+ "learning_rate": 0.0006,
+ "loss": 4.155317306518555,
+ "step": 2754
+ },
+ {
+ "epoch": 38.26550218340611,
+ "grad_norm": 0.01609654724597931,
+ "learning_rate": 0.0006,
+ "loss": 3.995800495147705,
+ "step": 2755
+ },
+ {
+ "epoch": 38.27947598253275,
+ "grad_norm": 0.015495185740292072,
+ "learning_rate": 0.0006,
+ "loss": 3.995769739151001,
+ "step": 2756
+ },
+ {
+ "epoch": 38.29344978165939,
+ "grad_norm": 0.017042381688952446,
+ "learning_rate": 0.0006,
+ "loss": 4.132832050323486,
+ "step": 2757
+ },
+ {
+ "epoch": 38.30742358078603,
+ "grad_norm": 0.018706614151597023,
+ "learning_rate": 0.0006,
+ "loss": 4.038820743560791,
+ "step": 2758
+ },
+ {
+ "epoch": 38.32139737991266,
+ "grad_norm": 0.018316026777029037,
+ "learning_rate": 0.0006,
+ "loss": 3.972052574157715,
+ "step": 2759
+ },
+ {
+ "epoch": 38.3353711790393,
+ "grad_norm": 0.018375389277935028,
+ "learning_rate": 0.0006,
+ "loss": 4.152097702026367,
+ "step": 2760
+ },
+ {
+ "epoch": 38.34934497816594,
+ "grad_norm": 0.016879886388778687,
+ "learning_rate": 0.0006,
+ "loss": 4.020869731903076,
+ "step": 2761
+ },
+ {
+ "epoch": 38.36331877729258,
+ "grad_norm": 0.015728183090686798,
+ "learning_rate": 0.0006,
+ "loss": 4.008016586303711,
+ "step": 2762
+ },
+ {
+ "epoch": 38.377292576419215,
+ "grad_norm": 0.016147758811712265,
+ "learning_rate": 0.0006,
+ "loss": 4.035854339599609,
+ "step": 2763
+ },
+ {
+ "epoch": 38.391266375545854,
+ "grad_norm": 0.017774797976017,
+ "learning_rate": 0.0006,
+ "loss": 4.122506141662598,
+ "step": 2764
+ },
+ {
+ "epoch": 38.40524017467249,
+ "grad_norm": 0.016590069979429245,
+ "learning_rate": 0.0006,
+ "loss": 4.082847595214844,
+ "step": 2765
+ },
+ {
+ "epoch": 38.419213973799124,
+ "grad_norm": 0.016544528305530548,
+ "learning_rate": 0.0006,
+ "loss": 4.184270858764648,
+ "step": 2766
+ },
+ {
+ "epoch": 38.43318777292576,
+ "grad_norm": 0.017972925677895546,
+ "learning_rate": 0.0006,
+ "loss": 3.987558603286743,
+ "step": 2767
+ },
+ {
+ "epoch": 38.4471615720524,
+ "grad_norm": 0.018562227487564087,
+ "learning_rate": 0.0006,
+ "loss": 4.1193461418151855,
+ "step": 2768
+ },
+ {
+ "epoch": 38.46113537117904,
+ "grad_norm": 0.018896685913205147,
+ "learning_rate": 0.0006,
+ "loss": 4.057808876037598,
+ "step": 2769
+ },
+ {
+ "epoch": 38.47510917030568,
+ "grad_norm": 0.016793809831142426,
+ "learning_rate": 0.0006,
+ "loss": 4.069523334503174,
+ "step": 2770
+ },
+ {
+ "epoch": 38.48908296943232,
+ "grad_norm": 0.01698676310479641,
+ "learning_rate": 0.0006,
+ "loss": 3.997624158859253,
+ "step": 2771
+ },
+ {
+ "epoch": 38.50305676855895,
+ "grad_norm": 0.016784202307462692,
+ "learning_rate": 0.0006,
+ "loss": 4.174795627593994,
+ "step": 2772
+ },
+ {
+ "epoch": 38.51703056768559,
+ "grad_norm": 0.016785014420747757,
+ "learning_rate": 0.0006,
+ "loss": 3.999321222305298,
+ "step": 2773
+ },
+ {
+ "epoch": 38.531004366812226,
+ "grad_norm": 0.018615184351801872,
+ "learning_rate": 0.0006,
+ "loss": 4.191237926483154,
+ "step": 2774
+ },
+ {
+ "epoch": 38.544978165938865,
+ "grad_norm": 0.019090790301561356,
+ "learning_rate": 0.0006,
+ "loss": 4.106966018676758,
+ "step": 2775
+ },
+ {
+ "epoch": 38.5589519650655,
+ "grad_norm": 0.018298320472240448,
+ "learning_rate": 0.0006,
+ "loss": 4.063363552093506,
+ "step": 2776
+ },
+ {
+ "epoch": 38.57292576419214,
+ "grad_norm": 0.018541090190410614,
+ "learning_rate": 0.0006,
+ "loss": 4.101193428039551,
+ "step": 2777
+ },
+ {
+ "epoch": 38.58689956331878,
+ "grad_norm": 0.01849970407783985,
+ "learning_rate": 0.0006,
+ "loss": 4.071335792541504,
+ "step": 2778
+ },
+ {
+ "epoch": 38.60087336244541,
+ "grad_norm": 0.018309077247977257,
+ "learning_rate": 0.0006,
+ "loss": 4.245305061340332,
+ "step": 2779
+ },
+ {
+ "epoch": 38.61484716157205,
+ "grad_norm": 0.017681840807199478,
+ "learning_rate": 0.0006,
+ "loss": 4.081173419952393,
+ "step": 2780
+ },
+ {
+ "epoch": 38.62882096069869,
+ "grad_norm": 0.016286224126815796,
+ "learning_rate": 0.0006,
+ "loss": 4.024351596832275,
+ "step": 2781
+ },
+ {
+ "epoch": 38.64279475982533,
+ "grad_norm": 0.015837235376238823,
+ "learning_rate": 0.0006,
+ "loss": 4.028205394744873,
+ "step": 2782
+ },
+ {
+ "epoch": 38.65676855895197,
+ "grad_norm": 0.016062596812844276,
+ "learning_rate": 0.0006,
+ "loss": 4.131122589111328,
+ "step": 2783
+ },
+ {
+ "epoch": 38.670742358078606,
+ "grad_norm": 0.01711239665746689,
+ "learning_rate": 0.0006,
+ "loss": 4.124416351318359,
+ "step": 2784
+ },
+ {
+ "epoch": 38.68471615720524,
+ "grad_norm": 0.015304913744330406,
+ "learning_rate": 0.0006,
+ "loss": 4.056122779846191,
+ "step": 2785
+ },
+ {
+ "epoch": 38.698689956331876,
+ "grad_norm": 0.015971887856721878,
+ "learning_rate": 0.0006,
+ "loss": 4.132555961608887,
+ "step": 2786
+ },
+ {
+ "epoch": 38.712663755458514,
+ "grad_norm": 0.015787221491336823,
+ "learning_rate": 0.0006,
+ "loss": 4.051596164703369,
+ "step": 2787
+ },
+ {
+ "epoch": 38.72663755458515,
+ "grad_norm": 0.01605444774031639,
+ "learning_rate": 0.0006,
+ "loss": 4.191808700561523,
+ "step": 2788
+ },
+ {
+ "epoch": 38.74061135371179,
+ "grad_norm": 0.016383009031414986,
+ "learning_rate": 0.0006,
+ "loss": 4.0648417472839355,
+ "step": 2789
+ },
+ {
+ "epoch": 38.75458515283843,
+ "grad_norm": 0.016135554760694504,
+ "learning_rate": 0.0006,
+ "loss": 3.988563060760498,
+ "step": 2790
+ },
+ {
+ "epoch": 38.76855895196506,
+ "grad_norm": 0.016634158790111542,
+ "learning_rate": 0.0006,
+ "loss": 4.133493423461914,
+ "step": 2791
+ },
+ {
+ "epoch": 38.7825327510917,
+ "grad_norm": 0.01688210666179657,
+ "learning_rate": 0.0006,
+ "loss": 4.091277122497559,
+ "step": 2792
+ },
+ {
+ "epoch": 38.79650655021834,
+ "grad_norm": 0.016102122142910957,
+ "learning_rate": 0.0006,
+ "loss": 4.051219940185547,
+ "step": 2793
+ },
+ {
+ "epoch": 38.81048034934498,
+ "grad_norm": 0.016585038974881172,
+ "learning_rate": 0.0006,
+ "loss": 4.058587074279785,
+ "step": 2794
+ },
+ {
+ "epoch": 38.82445414847162,
+ "grad_norm": 0.017700951546430588,
+ "learning_rate": 0.0006,
+ "loss": 4.016030311584473,
+ "step": 2795
+ },
+ {
+ "epoch": 38.838427947598255,
+ "grad_norm": 0.017922203987836838,
+ "learning_rate": 0.0006,
+ "loss": 4.079977989196777,
+ "step": 2796
+ },
+ {
+ "epoch": 38.852401746724894,
+ "grad_norm": 0.017299890518188477,
+ "learning_rate": 0.0006,
+ "loss": 4.062984943389893,
+ "step": 2797
+ },
+ {
+ "epoch": 38.866375545851525,
+ "grad_norm": 0.018648581579327583,
+ "learning_rate": 0.0006,
+ "loss": 4.065717697143555,
+ "step": 2798
+ },
+ {
+ "epoch": 38.880349344978164,
+ "grad_norm": 0.019057920202612877,
+ "learning_rate": 0.0006,
+ "loss": 4.227892875671387,
+ "step": 2799
+ },
+ {
+ "epoch": 38.8943231441048,
+ "grad_norm": 0.018824422731995583,
+ "learning_rate": 0.0006,
+ "loss": 4.116058349609375,
+ "step": 2800
+ },
+ {
+ "epoch": 38.90829694323144,
+ "grad_norm": 0.019084978848695755,
+ "learning_rate": 0.0006,
+ "loss": 4.061677932739258,
+ "step": 2801
+ },
+ {
+ "epoch": 38.92227074235808,
+ "grad_norm": 0.017842981964349747,
+ "learning_rate": 0.0006,
+ "loss": 4.083813667297363,
+ "step": 2802
+ },
+ {
+ "epoch": 38.93624454148472,
+ "grad_norm": 0.01880858838558197,
+ "learning_rate": 0.0006,
+ "loss": 4.152561187744141,
+ "step": 2803
+ },
+ {
+ "epoch": 38.95021834061135,
+ "grad_norm": 0.017400087788701057,
+ "learning_rate": 0.0006,
+ "loss": 4.154635429382324,
+ "step": 2804
+ },
+ {
+ "epoch": 38.96419213973799,
+ "grad_norm": 0.017682507634162903,
+ "learning_rate": 0.0006,
+ "loss": 4.151128768920898,
+ "step": 2805
+ },
+ {
+ "epoch": 38.97816593886463,
+ "grad_norm": 0.015803052112460136,
+ "learning_rate": 0.0006,
+ "loss": 4.10498571395874,
+ "step": 2806
+ },
+ {
+ "epoch": 38.992139737991266,
+ "grad_norm": 0.01582511141896248,
+ "learning_rate": 0.0006,
+ "loss": 4.140159606933594,
+ "step": 2807
+ },
+ {
+ "epoch": 39.0,
+ "grad_norm": 0.01925666816532612,
+ "learning_rate": 0.0006,
+ "loss": 4.126394271850586,
+ "step": 2808
+ },
+ {
+ "epoch": 39.0,
+ "eval_loss": 4.615617275238037,
+ "eval_runtime": 56.8895,
+ "eval_samples_per_second": 42.925,
+ "eval_steps_per_second": 1.354,
+ "step": 2808
+ },
+ {
+ "epoch": 39.01397379912664,
+ "grad_norm": 0.01677853614091873,
+ "learning_rate": 0.0006,
+ "loss": 4.102395057678223,
+ "step": 2809
+ },
+ {
+ "epoch": 39.02794759825328,
+ "grad_norm": 0.017901137471199036,
+ "learning_rate": 0.0006,
+ "loss": 3.976609230041504,
+ "step": 2810
+ },
+ {
+ "epoch": 39.041921397379916,
+ "grad_norm": 0.017978668212890625,
+ "learning_rate": 0.0006,
+ "loss": 3.944416046142578,
+ "step": 2811
+ },
+ {
+ "epoch": 39.05589519650655,
+ "grad_norm": 0.019967492669820786,
+ "learning_rate": 0.0006,
+ "loss": 3.992234468460083,
+ "step": 2812
+ },
+ {
+ "epoch": 39.069868995633186,
+ "grad_norm": 0.018937544897198677,
+ "learning_rate": 0.0006,
+ "loss": 4.028573036193848,
+ "step": 2813
+ },
+ {
+ "epoch": 39.083842794759825,
+ "grad_norm": 0.016852952539920807,
+ "learning_rate": 0.0006,
+ "loss": 4.062442779541016,
+ "step": 2814
+ },
+ {
+ "epoch": 39.09781659388646,
+ "grad_norm": 0.016643647104501724,
+ "learning_rate": 0.0006,
+ "loss": 4.08871603012085,
+ "step": 2815
+ },
+ {
+ "epoch": 39.1117903930131,
+ "grad_norm": 0.016872866079211235,
+ "learning_rate": 0.0006,
+ "loss": 3.9514245986938477,
+ "step": 2816
+ },
+ {
+ "epoch": 39.12576419213974,
+ "grad_norm": 0.018045669421553612,
+ "learning_rate": 0.0006,
+ "loss": 4.038842678070068,
+ "step": 2817
+ },
+ {
+ "epoch": 39.13973799126637,
+ "grad_norm": 0.017853403463959694,
+ "learning_rate": 0.0006,
+ "loss": 3.93709135055542,
+ "step": 2818
+ },
+ {
+ "epoch": 39.15371179039301,
+ "grad_norm": 0.017507443204522133,
+ "learning_rate": 0.0006,
+ "loss": 3.98665714263916,
+ "step": 2819
+ },
+ {
+ "epoch": 39.16768558951965,
+ "grad_norm": 0.01766853965818882,
+ "learning_rate": 0.0006,
+ "loss": 3.937126636505127,
+ "step": 2820
+ },
+ {
+ "epoch": 39.18165938864629,
+ "grad_norm": 0.016660021618008614,
+ "learning_rate": 0.0006,
+ "loss": 3.8397014141082764,
+ "step": 2821
+ },
+ {
+ "epoch": 39.19563318777293,
+ "grad_norm": 0.017719309777021408,
+ "learning_rate": 0.0006,
+ "loss": 4.003579616546631,
+ "step": 2822
+ },
+ {
+ "epoch": 39.209606986899566,
+ "grad_norm": 0.018234414979815483,
+ "learning_rate": 0.0006,
+ "loss": 4.053859710693359,
+ "step": 2823
+ },
+ {
+ "epoch": 39.223580786026204,
+ "grad_norm": 0.018777024000883102,
+ "learning_rate": 0.0006,
+ "loss": 4.056182861328125,
+ "step": 2824
+ },
+ {
+ "epoch": 39.237554585152836,
+ "grad_norm": 0.017443502321839333,
+ "learning_rate": 0.0006,
+ "loss": 4.025068283081055,
+ "step": 2825
+ },
+ {
+ "epoch": 39.251528384279474,
+ "grad_norm": 0.016500331461429596,
+ "learning_rate": 0.0006,
+ "loss": 4.101631164550781,
+ "step": 2826
+ },
+ {
+ "epoch": 39.26550218340611,
+ "grad_norm": 0.016101544722914696,
+ "learning_rate": 0.0006,
+ "loss": 4.027082443237305,
+ "step": 2827
+ },
+ {
+ "epoch": 39.27947598253275,
+ "grad_norm": 0.015471171587705612,
+ "learning_rate": 0.0006,
+ "loss": 4.022909641265869,
+ "step": 2828
+ },
+ {
+ "epoch": 39.29344978165939,
+ "grad_norm": 0.015779748558998108,
+ "learning_rate": 0.0006,
+ "loss": 4.18796968460083,
+ "step": 2829
+ },
+ {
+ "epoch": 39.30742358078603,
+ "grad_norm": 0.015522679314017296,
+ "learning_rate": 0.0006,
+ "loss": 4.005777359008789,
+ "step": 2830
+ },
+ {
+ "epoch": 39.32139737991266,
+ "grad_norm": 0.015226504765450954,
+ "learning_rate": 0.0006,
+ "loss": 4.108416557312012,
+ "step": 2831
+ },
+ {
+ "epoch": 39.3353711790393,
+ "grad_norm": 0.015095778740942478,
+ "learning_rate": 0.0006,
+ "loss": 4.08029842376709,
+ "step": 2832
+ },
+ {
+ "epoch": 39.34934497816594,
+ "grad_norm": 0.015601031482219696,
+ "learning_rate": 0.0006,
+ "loss": 3.989492416381836,
+ "step": 2833
+ },
+ {
+ "epoch": 39.36331877729258,
+ "grad_norm": 0.014825052581727505,
+ "learning_rate": 0.0006,
+ "loss": 3.998520612716675,
+ "step": 2834
+ },
+ {
+ "epoch": 39.377292576419215,
+ "grad_norm": 0.014821100980043411,
+ "learning_rate": 0.0006,
+ "loss": 4.001777172088623,
+ "step": 2835
+ },
+ {
+ "epoch": 39.391266375545854,
+ "grad_norm": 0.015333754941821098,
+ "learning_rate": 0.0006,
+ "loss": 4.11985969543457,
+ "step": 2836
+ },
+ {
+ "epoch": 39.40524017467249,
+ "grad_norm": 0.015940699726343155,
+ "learning_rate": 0.0006,
+ "loss": 4.025373935699463,
+ "step": 2837
+ },
+ {
+ "epoch": 39.419213973799124,
+ "grad_norm": 0.015096287243068218,
+ "learning_rate": 0.0006,
+ "loss": 4.068551063537598,
+ "step": 2838
+ },
+ {
+ "epoch": 39.43318777292576,
+ "grad_norm": 0.01533548254519701,
+ "learning_rate": 0.0006,
+ "loss": 3.9800801277160645,
+ "step": 2839
+ },
+ {
+ "epoch": 39.4471615720524,
+ "grad_norm": 0.014648732729256153,
+ "learning_rate": 0.0006,
+ "loss": 3.9670190811157227,
+ "step": 2840
+ },
+ {
+ "epoch": 39.46113537117904,
+ "grad_norm": 0.015759671106934547,
+ "learning_rate": 0.0006,
+ "loss": 4.050443649291992,
+ "step": 2841
+ },
+ {
+ "epoch": 39.47510917030568,
+ "grad_norm": 0.016054073348641396,
+ "learning_rate": 0.0006,
+ "loss": 4.22734260559082,
+ "step": 2842
+ },
+ {
+ "epoch": 39.48908296943232,
+ "grad_norm": 0.0160264540463686,
+ "learning_rate": 0.0006,
+ "loss": 4.116174221038818,
+ "step": 2843
+ },
+ {
+ "epoch": 39.50305676855895,
+ "grad_norm": 0.015426610596477985,
+ "learning_rate": 0.0006,
+ "loss": 4.000178337097168,
+ "step": 2844
+ },
+ {
+ "epoch": 39.51703056768559,
+ "grad_norm": 0.01520642638206482,
+ "learning_rate": 0.0006,
+ "loss": 4.218410968780518,
+ "step": 2845
+ },
+ {
+ "epoch": 39.531004366812226,
+ "grad_norm": 0.015305949375033379,
+ "learning_rate": 0.0006,
+ "loss": 4.025891304016113,
+ "step": 2846
+ },
+ {
+ "epoch": 39.544978165938865,
+ "grad_norm": 0.015461564064025879,
+ "learning_rate": 0.0006,
+ "loss": 4.100950717926025,
+ "step": 2847
+ },
+ {
+ "epoch": 39.5589519650655,
+ "grad_norm": 0.01712625101208687,
+ "learning_rate": 0.0006,
+ "loss": 4.0620856285095215,
+ "step": 2848
+ },
+ {
+ "epoch": 39.57292576419214,
+ "grad_norm": 0.017133301123976707,
+ "learning_rate": 0.0006,
+ "loss": 4.097138404846191,
+ "step": 2849
+ },
+ {
+ "epoch": 39.58689956331878,
+ "grad_norm": 0.01683625392615795,
+ "learning_rate": 0.0006,
+ "loss": 4.084364891052246,
+ "step": 2850
+ },
+ {
+ "epoch": 39.60087336244541,
+ "grad_norm": 0.0165710486471653,
+ "learning_rate": 0.0006,
+ "loss": 4.042361259460449,
+ "step": 2851
+ },
+ {
+ "epoch": 39.61484716157205,
+ "grad_norm": 0.01660393364727497,
+ "learning_rate": 0.0006,
+ "loss": 4.1395487785339355,
+ "step": 2852
+ },
+ {
+ "epoch": 39.62882096069869,
+ "grad_norm": 0.017025845125317574,
+ "learning_rate": 0.0006,
+ "loss": 4.1207075119018555,
+ "step": 2853
+ },
+ {
+ "epoch": 39.64279475982533,
+ "grad_norm": 0.01660497486591339,
+ "learning_rate": 0.0006,
+ "loss": 4.103846073150635,
+ "step": 2854
+ },
+ {
+ "epoch": 39.65676855895197,
+ "grad_norm": 0.01607946865260601,
+ "learning_rate": 0.0006,
+ "loss": 4.189016342163086,
+ "step": 2855
+ },
+ {
+ "epoch": 39.670742358078606,
+ "grad_norm": 0.01614505983889103,
+ "learning_rate": 0.0006,
+ "loss": 4.079860687255859,
+ "step": 2856
+ },
+ {
+ "epoch": 39.68471615720524,
+ "grad_norm": 0.017085086554288864,
+ "learning_rate": 0.0006,
+ "loss": 3.9774293899536133,
+ "step": 2857
+ },
+ {
+ "epoch": 39.698689956331876,
+ "grad_norm": 0.01816520094871521,
+ "learning_rate": 0.0006,
+ "loss": 4.113666534423828,
+ "step": 2858
+ },
+ {
+ "epoch": 39.712663755458514,
+ "grad_norm": 0.016597526147961617,
+ "learning_rate": 0.0006,
+ "loss": 4.068242073059082,
+ "step": 2859
+ },
+ {
+ "epoch": 39.72663755458515,
+ "grad_norm": 0.01672235131263733,
+ "learning_rate": 0.0006,
+ "loss": 4.024444580078125,
+ "step": 2860
+ },
+ {
+ "epoch": 39.74061135371179,
+ "grad_norm": 0.020093556493520737,
+ "learning_rate": 0.0006,
+ "loss": 4.064124584197998,
+ "step": 2861
+ },
+ {
+ "epoch": 39.75458515283843,
+ "grad_norm": 0.023141304031014442,
+ "learning_rate": 0.0006,
+ "loss": 4.130439758300781,
+ "step": 2862
+ },
+ {
+ "epoch": 39.76855895196506,
+ "grad_norm": 0.02289372682571411,
+ "learning_rate": 0.0006,
+ "loss": 4.14794921875,
+ "step": 2863
+ },
+ {
+ "epoch": 39.7825327510917,
+ "grad_norm": 0.01919020712375641,
+ "learning_rate": 0.0006,
+ "loss": 4.123961925506592,
+ "step": 2864
+ },
+ {
+ "epoch": 39.79650655021834,
+ "grad_norm": 0.016932329162955284,
+ "learning_rate": 0.0006,
+ "loss": 4.0944132804870605,
+ "step": 2865
+ },
+ {
+ "epoch": 39.81048034934498,
+ "grad_norm": 0.016644535586237907,
+ "learning_rate": 0.0006,
+ "loss": 4.124269962310791,
+ "step": 2866
+ },
+ {
+ "epoch": 39.82445414847162,
+ "grad_norm": 0.016302216798067093,
+ "learning_rate": 0.0006,
+ "loss": 4.0267205238342285,
+ "step": 2867
+ },
+ {
+ "epoch": 39.838427947598255,
+ "grad_norm": 0.01673300378024578,
+ "learning_rate": 0.0006,
+ "loss": 4.10650634765625,
+ "step": 2868
+ },
+ {
+ "epoch": 39.852401746724894,
+ "grad_norm": 0.016581356525421143,
+ "learning_rate": 0.0006,
+ "loss": 4.1547956466674805,
+ "step": 2869
+ },
+ {
+ "epoch": 39.866375545851525,
+ "grad_norm": 0.01691797934472561,
+ "learning_rate": 0.0006,
+ "loss": 4.056936740875244,
+ "step": 2870
+ },
+ {
+ "epoch": 39.880349344978164,
+ "grad_norm": 0.01596726104617119,
+ "learning_rate": 0.0006,
+ "loss": 3.9891414642333984,
+ "step": 2871
+ },
+ {
+ "epoch": 39.8943231441048,
+ "grad_norm": 0.015615091659128666,
+ "learning_rate": 0.0006,
+ "loss": 4.137948036193848,
+ "step": 2872
+ },
+ {
+ "epoch": 39.90829694323144,
+ "grad_norm": 0.015703413635492325,
+ "learning_rate": 0.0006,
+ "loss": 3.9844930171966553,
+ "step": 2873
+ },
+ {
+ "epoch": 39.92227074235808,
+ "grad_norm": 0.015453094616532326,
+ "learning_rate": 0.0006,
+ "loss": 4.098443508148193,
+ "step": 2874
+ },
+ {
+ "epoch": 39.93624454148472,
+ "grad_norm": 0.01449158601462841,
+ "learning_rate": 0.0006,
+ "loss": 4.090852737426758,
+ "step": 2875
+ },
+ {
+ "epoch": 39.95021834061135,
+ "grad_norm": 0.01537515688687563,
+ "learning_rate": 0.0006,
+ "loss": 4.239612579345703,
+ "step": 2876
+ },
+ {
+ "epoch": 39.96419213973799,
+ "grad_norm": 0.015989921987056732,
+ "learning_rate": 0.0006,
+ "loss": 4.142073154449463,
+ "step": 2877
+ },
+ {
+ "epoch": 39.97816593886463,
+ "grad_norm": 0.015985611826181412,
+ "learning_rate": 0.0006,
+ "loss": 4.100076675415039,
+ "step": 2878
+ },
+ {
+ "epoch": 39.992139737991266,
+ "grad_norm": 0.015260833315551281,
+ "learning_rate": 0.0006,
+ "loss": 4.068576335906982,
+ "step": 2879
+ },
+ {
+ "epoch": 40.0,
+ "grad_norm": 0.018047941848635674,
+ "learning_rate": 0.0006,
+ "loss": 4.005878925323486,
+ "step": 2880
+ },
+ {
+ "epoch": 40.0,
+ "eval_loss": 4.622848987579346,
+ "eval_runtime": 56.9618,
+ "eval_samples_per_second": 42.871,
+ "eval_steps_per_second": 1.352,
+ "step": 2880
+ },
+ {
+ "epoch": 40.01397379912664,
+ "grad_norm": 0.015976838767528534,
+ "learning_rate": 0.0006,
+ "loss": 4.035689353942871,
+ "step": 2881
+ },
+ {
+ "epoch": 40.02794759825328,
+ "grad_norm": 0.015176220797002316,
+ "learning_rate": 0.0006,
+ "loss": 3.932197332382202,
+ "step": 2882
+ },
+ {
+ "epoch": 40.041921397379916,
+ "grad_norm": 0.016676228493452072,
+ "learning_rate": 0.0006,
+ "loss": 4.014171123504639,
+ "step": 2883
+ },
+ {
+ "epoch": 40.05589519650655,
+ "grad_norm": 0.01698281429708004,
+ "learning_rate": 0.0006,
+ "loss": 4.117893218994141,
+ "step": 2884
+ },
+ {
+ "epoch": 40.069868995633186,
+ "grad_norm": 0.017872486263513565,
+ "learning_rate": 0.0006,
+ "loss": 4.132119178771973,
+ "step": 2885
+ },
+ {
+ "epoch": 40.083842794759825,
+ "grad_norm": 0.017615554854273796,
+ "learning_rate": 0.0006,
+ "loss": 4.072000980377197,
+ "step": 2886
+ },
+ {
+ "epoch": 40.09781659388646,
+ "grad_norm": 0.017469394952058792,
+ "learning_rate": 0.0006,
+ "loss": 4.017726421356201,
+ "step": 2887
+ },
+ {
+ "epoch": 40.1117903930131,
+ "grad_norm": 0.01954762637615204,
+ "learning_rate": 0.0006,
+ "loss": 4.014443874359131,
+ "step": 2888
+ },
+ {
+ "epoch": 40.12576419213974,
+ "grad_norm": 0.018935853615403175,
+ "learning_rate": 0.0006,
+ "loss": 4.0042948722839355,
+ "step": 2889
+ },
+ {
+ "epoch": 40.13973799126637,
+ "grad_norm": 0.019382692873477936,
+ "learning_rate": 0.0006,
+ "loss": 4.06160831451416,
+ "step": 2890
+ },
+ {
+ "epoch": 40.15371179039301,
+ "grad_norm": 0.019677115604281425,
+ "learning_rate": 0.0006,
+ "loss": 4.13466215133667,
+ "step": 2891
+ },
+ {
+ "epoch": 40.16768558951965,
+ "grad_norm": 0.020365122705698013,
+ "learning_rate": 0.0006,
+ "loss": 4.095000267028809,
+ "step": 2892
+ },
+ {
+ "epoch": 40.18165938864629,
+ "grad_norm": 0.021842138841748238,
+ "learning_rate": 0.0006,
+ "loss": 3.9877853393554688,
+ "step": 2893
+ },
+ {
+ "epoch": 40.19563318777293,
+ "grad_norm": 0.02354881353676319,
+ "learning_rate": 0.0006,
+ "loss": 3.9430112838745117,
+ "step": 2894
+ },
+ {
+ "epoch": 40.209606986899566,
+ "grad_norm": 0.026602735742926598,
+ "learning_rate": 0.0006,
+ "loss": 4.012076377868652,
+ "step": 2895
+ },
+ {
+ "epoch": 40.223580786026204,
+ "grad_norm": 0.023976484313607216,
+ "learning_rate": 0.0006,
+ "loss": 3.9768595695495605,
+ "step": 2896
+ },
+ {
+ "epoch": 40.237554585152836,
+ "grad_norm": 0.02336616814136505,
+ "learning_rate": 0.0006,
+ "loss": 4.1096038818359375,
+ "step": 2897
+ },
+ {
+ "epoch": 40.251528384279474,
+ "grad_norm": 0.02426023595035076,
+ "learning_rate": 0.0006,
+ "loss": 4.157189846038818,
+ "step": 2898
+ },
+ {
+ "epoch": 40.26550218340611,
+ "grad_norm": 0.021572696045041084,
+ "learning_rate": 0.0006,
+ "loss": 3.979175090789795,
+ "step": 2899
+ },
+ {
+ "epoch": 40.27947598253275,
+ "grad_norm": 0.08300432562828064,
+ "learning_rate": 0.0006,
+ "loss": 4.092336654663086,
+ "step": 2900
+ },
+ {
+ "epoch": 40.29344978165939,
+ "grad_norm": 0.7323967814445496,
+ "learning_rate": 0.0006,
+ "loss": 5.337990760803223,
+ "step": 2901
+ },
+ {
+ "epoch": 40.30742358078603,
+ "grad_norm": 0.3363337814807892,
+ "learning_rate": 0.0006,
+ "loss": 6.40043830871582,
+ "step": 2902
+ },
+ {
+ "epoch": 40.32139737991266,
+ "grad_norm": 0.598686933517456,
+ "learning_rate": 0.0006,
+ "loss": 7.004765510559082,
+ "step": 2903
+ },
+ {
+ "epoch": 40.3353711790393,
+ "grad_norm": 0.5998364686965942,
+ "learning_rate": 0.0006,
+ "loss": 8.184747695922852,
+ "step": 2904
+ },
+ {
+ "epoch": 40.34934497816594,
+ "grad_norm": 1.408208966255188,
+ "learning_rate": 0.0006,
+ "loss": 8.88978385925293,
+ "step": 2905
+ },
+ {
+ "epoch": 40.36331877729258,
+ "grad_norm": 0.7829785943031311,
+ "learning_rate": 0.0006,
+ "loss": 11.13601303100586,
+ "step": 2906
+ },
+ {
+ "epoch": 40.377292576419215,
+ "grad_norm": 0.7769215106964111,
+ "learning_rate": 0.0006,
+ "loss": 9.90416145324707,
+ "step": 2907
+ },
+ {
+ "epoch": 40.391266375545854,
+ "grad_norm": 0.46508079767227173,
+ "learning_rate": 0.0006,
+ "loss": 9.647209167480469,
+ "step": 2908
+ },
+ {
+ "epoch": 40.40524017467249,
+ "grad_norm": 1.0228968858718872,
+ "learning_rate": 0.0006,
+ "loss": 11.685434341430664,
+ "step": 2909
+ },
+ {
+ "epoch": 40.419213973799124,
+ "grad_norm": 0.20553778111934662,
+ "learning_rate": 0.0006,
+ "loss": 8.621708869934082,
+ "step": 2910
+ },
+ {
+ "epoch": 40.43318777292576,
+ "grad_norm": 0.18466424942016602,
+ "learning_rate": 0.0006,
+ "loss": 8.590787887573242,
+ "step": 2911
+ },
+ {
+ "epoch": 40.4471615720524,
+ "grad_norm": 0.10458594560623169,
+ "learning_rate": 0.0006,
+ "loss": 8.324458122253418,
+ "step": 2912
+ },
+ {
+ "epoch": 40.46113537117904,
+ "grad_norm": 0.1096842959523201,
+ "learning_rate": 0.0006,
+ "loss": 8.300851821899414,
+ "step": 2913
+ },
+ {
+ "epoch": 40.47510917030568,
+ "grad_norm": 0.10501226782798767,
+ "learning_rate": 0.0006,
+ "loss": 8.21009635925293,
+ "step": 2914
+ },
+ {
+ "epoch": 40.48908296943232,
+ "grad_norm": 0.13704165816307068,
+ "learning_rate": 0.0006,
+ "loss": 8.141355514526367,
+ "step": 2915
+ },
+ {
+ "epoch": 40.50305676855895,
+ "grad_norm": 0.12129108607769012,
+ "learning_rate": 0.0006,
+ "loss": 8.074612617492676,
+ "step": 2916
+ },
+ {
+ "epoch": 40.51703056768559,
+ "grad_norm": 0.10028897225856781,
+ "learning_rate": 0.0006,
+ "loss": 8.00384521484375,
+ "step": 2917
+ },
+ {
+ "epoch": 40.531004366812226,
+ "grad_norm": 0.11764246970415115,
+ "learning_rate": 0.0006,
+ "loss": 8.078598022460938,
+ "step": 2918
+ },
+ {
+ "epoch": 40.544978165938865,
+ "grad_norm": 0.10081959515810013,
+ "learning_rate": 0.0006,
+ "loss": 8.008938789367676,
+ "step": 2919
+ },
+ {
+ "epoch": 40.5589519650655,
+ "grad_norm": 0.1003170907497406,
+ "learning_rate": 0.0006,
+ "loss": 7.962940216064453,
+ "step": 2920
+ },
+ {
+ "epoch": 40.57292576419214,
+ "grad_norm": 0.08763492107391357,
+ "learning_rate": 0.0006,
+ "loss": 7.909419536590576,
+ "step": 2921
+ },
+ {
+ "epoch": 40.58689956331878,
+ "grad_norm": 0.0636807233095169,
+ "learning_rate": 0.0006,
+ "loss": 7.878905773162842,
+ "step": 2922
+ },
+ {
+ "epoch": 40.60087336244541,
+ "grad_norm": 0.06005936488509178,
+ "learning_rate": 0.0006,
+ "loss": 7.850980758666992,
+ "step": 2923
+ },
+ {
+ "epoch": 40.61484716157205,
+ "grad_norm": 0.06415296345949173,
+ "learning_rate": 0.0006,
+ "loss": 7.830353260040283,
+ "step": 2924
+ },
+ {
+ "epoch": 40.62882096069869,
+ "grad_norm": 0.0732007846236229,
+ "learning_rate": 0.0006,
+ "loss": 7.830746650695801,
+ "step": 2925
+ },
+ {
+ "epoch": 40.64279475982533,
+ "grad_norm": 0.06628110259771347,
+ "learning_rate": 0.0006,
+ "loss": 7.82661771774292,
+ "step": 2926
+ },
+ {
+ "epoch": 40.65676855895197,
+ "grad_norm": 0.05544218048453331,
+ "learning_rate": 0.0006,
+ "loss": 7.814305782318115,
+ "step": 2927
+ },
+ {
+ "epoch": 40.670742358078606,
+ "grad_norm": 0.050327103585004807,
+ "learning_rate": 0.0006,
+ "loss": 7.764055252075195,
+ "step": 2928
+ },
+ {
+ "epoch": 40.68471615720524,
+ "grad_norm": 0.04918425530195236,
+ "learning_rate": 0.0006,
+ "loss": 7.744625091552734,
+ "step": 2929
+ },
+ {
+ "epoch": 40.698689956331876,
+ "grad_norm": 0.042862266302108765,
+ "learning_rate": 0.0006,
+ "loss": 7.764426231384277,
+ "step": 2930
+ },
+ {
+ "epoch": 40.712663755458514,
+ "grad_norm": 0.04338089004158974,
+ "learning_rate": 0.0006,
+ "loss": 7.712530136108398,
+ "step": 2931
+ },
+ {
+ "epoch": 40.72663755458515,
+ "grad_norm": 0.04127737879753113,
+ "learning_rate": 0.0006,
+ "loss": 7.742180347442627,
+ "step": 2932
+ },
+ {
+ "epoch": 40.74061135371179,
+ "grad_norm": 0.03358631581068039,
+ "learning_rate": 0.0006,
+ "loss": 7.714834213256836,
+ "step": 2933
+ },
+ {
+ "epoch": 40.75458515283843,
+ "grad_norm": 0.02845790982246399,
+ "learning_rate": 0.0006,
+ "loss": 7.7021026611328125,
+ "step": 2934
+ },
+ {
+ "epoch": 40.76855895196506,
+ "grad_norm": 0.03012419492006302,
+ "learning_rate": 0.0006,
+ "loss": 7.712377548217773,
+ "step": 2935
+ },
+ {
+ "epoch": 40.7825327510917,
+ "grad_norm": 0.032981179654598236,
+ "learning_rate": 0.0006,
+ "loss": 7.666385650634766,
+ "step": 2936
+ },
+ {
+ "epoch": 40.79650655021834,
+ "grad_norm": 0.03357074037194252,
+ "learning_rate": 0.0006,
+ "loss": 7.649097919464111,
+ "step": 2937
+ },
+ {
+ "epoch": 40.81048034934498,
+ "grad_norm": 0.030914708971977234,
+ "learning_rate": 0.0006,
+ "loss": 7.649196624755859,
+ "step": 2938
+ },
+ {
+ "epoch": 40.82445414847162,
+ "grad_norm": 0.026296131312847137,
+ "learning_rate": 0.0006,
+ "loss": 7.629964828491211,
+ "step": 2939
+ },
+ {
+ "epoch": 40.838427947598255,
+ "grad_norm": 0.026677245274186134,
+ "learning_rate": 0.0006,
+ "loss": 7.599973678588867,
+ "step": 2940
+ },
+ {
+ "epoch": 40.852401746724894,
+ "grad_norm": 0.023208219558000565,
+ "learning_rate": 0.0006,
+ "loss": 7.601624488830566,
+ "step": 2941
+ },
+ {
+ "epoch": 40.866375545851525,
+ "grad_norm": 0.020741475746035576,
+ "learning_rate": 0.0006,
+ "loss": 7.603902339935303,
+ "step": 2942
+ },
+ {
+ "epoch": 40.880349344978164,
+ "grad_norm": 0.019221751019358635,
+ "learning_rate": 0.0006,
+ "loss": 7.5535888671875,
+ "step": 2943
+ },
+ {
+ "epoch": 40.8943231441048,
+ "grad_norm": 0.019491951912641525,
+ "learning_rate": 0.0006,
+ "loss": 7.54597282409668,
+ "step": 2944
+ },
+ {
+ "epoch": 40.90829694323144,
+ "grad_norm": 0.019702572375535965,
+ "learning_rate": 0.0006,
+ "loss": 7.537567138671875,
+ "step": 2945
+ },
+ {
+ "epoch": 40.92227074235808,
+ "grad_norm": 0.01768800988793373,
+ "learning_rate": 0.0006,
+ "loss": 7.512800216674805,
+ "step": 2946
+ },
+ {
+ "epoch": 40.93624454148472,
+ "grad_norm": 0.01926138624548912,
+ "learning_rate": 0.0006,
+ "loss": 7.509932518005371,
+ "step": 2947
+ },
+ {
+ "epoch": 40.95021834061135,
+ "grad_norm": 0.019030006602406502,
+ "learning_rate": 0.0006,
+ "loss": 7.497983932495117,
+ "step": 2948
+ },
+ {
+ "epoch": 40.96419213973799,
+ "grad_norm": 0.017313841730356216,
+ "learning_rate": 0.0006,
+ "loss": 7.463518142700195,
+ "step": 2949
+ },
+ {
+ "epoch": 40.97816593886463,
+ "grad_norm": 0.01310762483626604,
+ "learning_rate": 0.0006,
+ "loss": 7.474668502807617,
+ "step": 2950
+ },
+ {
+ "epoch": 40.992139737991266,
+ "grad_norm": 0.013008282519876957,
+ "learning_rate": 0.0006,
+ "loss": 7.45982551574707,
+ "step": 2951
+ },
+ {
+ "epoch": 41.0,
+ "grad_norm": 0.01441363524645567,
+ "learning_rate": 0.0006,
+ "loss": 7.446529388427734,
+ "step": 2952
+ },
+ {
+ "epoch": 41.0,
+ "eval_loss": 7.486120223999023,
+ "eval_runtime": 70.4986,
+ "eval_samples_per_second": 34.639,
+ "eval_steps_per_second": 1.092,
+ "step": 2952
+ },
+ {
+ "epoch": 41.01397379912664,
+ "grad_norm": 0.014061855152249336,
+ "learning_rate": 0.0006,
+ "loss": 7.4692182540893555,
+ "step": 2953
+ },
+ {
+ "epoch": 41.02794759825328,
+ "grad_norm": 0.01587500050663948,
+ "learning_rate": 0.0006,
+ "loss": 7.4182024002075195,
+ "step": 2954
+ },
+ {
+ "epoch": 41.041921397379916,
+ "grad_norm": 0.014783984050154686,
+ "learning_rate": 0.0006,
+ "loss": 7.466778755187988,
+ "step": 2955
+ },
+ {
+ "epoch": 41.05589519650655,
+ "grad_norm": 0.011560854502022266,
+ "learning_rate": 0.0006,
+ "loss": 7.438002586364746,
+ "step": 2956
+ },
+ {
+ "epoch": 41.069868995633186,
+ "grad_norm": 0.00970957800745964,
+ "learning_rate": 0.0006,
+ "loss": 7.442694187164307,
+ "step": 2957
+ },
+ {
+ "epoch": 41.083842794759825,
+ "grad_norm": 0.009633459150791168,
+ "learning_rate": 0.0006,
+ "loss": 7.458884239196777,
+ "step": 2958
+ },
+ {
+ "epoch": 41.09781659388646,
+ "grad_norm": 0.010664350353181362,
+ "learning_rate": 0.0006,
+ "loss": 7.48378324508667,
+ "step": 2959
+ },
+ {
+ "epoch": 41.1117903930131,
+ "grad_norm": 0.011440591886639595,
+ "learning_rate": 0.0006,
+ "loss": 7.474484920501709,
+ "step": 2960
+ },
+ {
+ "epoch": 41.12576419213974,
+ "grad_norm": 0.01150908600538969,
+ "learning_rate": 0.0006,
+ "loss": 7.417902946472168,
+ "step": 2961
+ },
+ {
+ "epoch": 41.13973799126637,
+ "grad_norm": 0.010375697165727615,
+ "learning_rate": 0.0006,
+ "loss": 7.403279781341553,
+ "step": 2962
+ },
+ {
+ "epoch": 41.15371179039301,
+ "grad_norm": 0.009075703099370003,
+ "learning_rate": 0.0006,
+ "loss": 7.447347164154053,
+ "step": 2963
+ },
+ {
+ "epoch": 41.16768558951965,
+ "grad_norm": 0.008892177604138851,
+ "learning_rate": 0.0006,
+ "loss": 7.402681350708008,
+ "step": 2964
+ },
+ {
+ "epoch": 41.18165938864629,
+ "grad_norm": 0.009492084383964539,
+ "learning_rate": 0.0006,
+ "loss": 7.3871355056762695,
+ "step": 2965
+ },
+ {
+ "epoch": 41.19563318777293,
+ "grad_norm": 0.009139460511505604,
+ "learning_rate": 0.0006,
+ "loss": 7.419617176055908,
+ "step": 2966
+ },
+ {
+ "epoch": 41.209606986899566,
+ "grad_norm": 0.007959138602018356,
+ "learning_rate": 0.0006,
+ "loss": 7.374949932098389,
+ "step": 2967
+ },
+ {
+ "epoch": 41.223580786026204,
+ "grad_norm": 0.007571548223495483,
+ "learning_rate": 0.0006,
+ "loss": 7.4205241203308105,
+ "step": 2968
+ },
+ {
+ "epoch": 41.237554585152836,
+ "grad_norm": 0.007885400205850601,
+ "learning_rate": 0.0006,
+ "loss": 7.416543006896973,
+ "step": 2969
+ },
+ {
+ "epoch": 41.251528384279474,
+ "grad_norm": 0.008650158531963825,
+ "learning_rate": 0.0006,
+ "loss": 7.357326984405518,
+ "step": 2970
+ },
+ {
+ "epoch": 41.26550218340611,
+ "grad_norm": 0.007483262103050947,
+ "learning_rate": 0.0006,
+ "loss": 7.410891056060791,
+ "step": 2971
+ },
+ {
+ "epoch": 41.27947598253275,
+ "grad_norm": 0.00869277399033308,
+ "learning_rate": 0.0006,
+ "loss": 7.392127990722656,
+ "step": 2972
+ },
+ {
+ "epoch": 41.29344978165939,
+ "grad_norm": 0.00885078776627779,
+ "learning_rate": 0.0006,
+ "loss": 7.405449867248535,
+ "step": 2973
+ },
+ {
+ "epoch": 41.30742358078603,
+ "grad_norm": 0.00958226341754198,
+ "learning_rate": 0.0006,
+ "loss": 7.399868488311768,
+ "step": 2974
+ },
+ {
+ "epoch": 41.32139737991266,
+ "grad_norm": 0.006991599686443806,
+ "learning_rate": 0.0006,
+ "loss": 7.35369873046875,
+ "step": 2975
+ },
+ {
+ "epoch": 41.3353711790393,
+ "grad_norm": 0.007415142375975847,
+ "learning_rate": 0.0006,
+ "loss": 7.358353137969971,
+ "step": 2976
+ },
+ {
+ "epoch": 41.34934497816594,
+ "grad_norm": 0.007024371065199375,
+ "learning_rate": 0.0006,
+ "loss": 7.4070892333984375,
+ "step": 2977
+ },
+ {
+ "epoch": 41.36331877729258,
+ "grad_norm": 0.007234332151710987,
+ "learning_rate": 0.0006,
+ "loss": 7.379120349884033,
+ "step": 2978
+ },
+ {
+ "epoch": 41.377292576419215,
+ "grad_norm": 0.00798892229795456,
+ "learning_rate": 0.0006,
+ "loss": 7.365455627441406,
+ "step": 2979
+ },
+ {
+ "epoch": 41.391266375545854,
+ "grad_norm": 0.007303398102521896,
+ "learning_rate": 0.0006,
+ "loss": 7.341798782348633,
+ "step": 2980
+ },
+ {
+ "epoch": 41.40524017467249,
+ "grad_norm": 0.007119669578969479,
+ "learning_rate": 0.0006,
+ "loss": 7.363280296325684,
+ "step": 2981
+ },
+ {
+ "epoch": 41.419213973799124,
+ "grad_norm": 0.007457619532942772,
+ "learning_rate": 0.0006,
+ "loss": 7.39388370513916,
+ "step": 2982
+ },
+ {
+ "epoch": 41.43318777292576,
+ "grad_norm": 0.007905688136816025,
+ "learning_rate": 0.0006,
+ "loss": 7.380652904510498,
+ "step": 2983
+ },
+ {
+ "epoch": 41.4471615720524,
+ "grad_norm": 0.006212855689227581,
+ "learning_rate": 0.0006,
+ "loss": 7.398769378662109,
+ "step": 2984
+ },
+ {
+ "epoch": 41.46113537117904,
+ "grad_norm": 0.0070641543716192245,
+ "learning_rate": 0.0006,
+ "loss": 7.3503546714782715,
+ "step": 2985
+ },
+ {
+ "epoch": 41.47510917030568,
+ "grad_norm": 0.00787805300205946,
+ "learning_rate": 0.0006,
+ "loss": 7.347678184509277,
+ "step": 2986
+ },
+ {
+ "epoch": 41.48908296943232,
+ "grad_norm": 0.006760426796972752,
+ "learning_rate": 0.0006,
+ "loss": 7.380006790161133,
+ "step": 2987
+ },
+ {
+ "epoch": 41.50305676855895,
+ "grad_norm": 0.00772385997697711,
+ "learning_rate": 0.0006,
+ "loss": 7.363680362701416,
+ "step": 2988
+ },
+ {
+ "epoch": 41.51703056768559,
+ "grad_norm": 0.007373564876616001,
+ "learning_rate": 0.0006,
+ "loss": 7.333014488220215,
+ "step": 2989
+ },
+ {
+ "epoch": 41.531004366812226,
+ "grad_norm": 0.006717653013765812,
+ "learning_rate": 0.0006,
+ "loss": 7.360076427459717,
+ "step": 2990
+ },
+ {
+ "epoch": 41.544978165938865,
+ "grad_norm": 0.007191413082182407,
+ "learning_rate": 0.0006,
+ "loss": 7.397030830383301,
+ "step": 2991
+ },
+ {
+ "epoch": 41.5589519650655,
+ "grad_norm": 0.007831960916519165,
+ "learning_rate": 0.0006,
+ "loss": 7.369959831237793,
+ "step": 2992
+ },
+ {
+ "epoch": 41.57292576419214,
+ "grad_norm": 0.006947481073439121,
+ "learning_rate": 0.0006,
+ "loss": 7.333254337310791,
+ "step": 2993
+ },
+ {
+ "epoch": 41.58689956331878,
+ "grad_norm": 0.006243606563657522,
+ "learning_rate": 0.0006,
+ "loss": 7.327378273010254,
+ "step": 2994
+ },
+ {
+ "epoch": 41.60087336244541,
+ "grad_norm": 0.006298339460045099,
+ "learning_rate": 0.0006,
+ "loss": 7.323482036590576,
+ "step": 2995
+ },
+ {
+ "epoch": 41.61484716157205,
+ "grad_norm": 0.007587412837892771,
+ "learning_rate": 0.0006,
+ "loss": 7.313789367675781,
+ "step": 2996
+ },
+ {
+ "epoch": 41.62882096069869,
+ "grad_norm": 0.007140746805816889,
+ "learning_rate": 0.0006,
+ "loss": 7.356067657470703,
+ "step": 2997
+ },
+ {
+ "epoch": 41.64279475982533,
+ "grad_norm": 0.008315625600516796,
+ "learning_rate": 0.0006,
+ "loss": 7.3339643478393555,
+ "step": 2998
+ },
+ {
+ "epoch": 41.65676855895197,
+ "grad_norm": 0.006547102238982916,
+ "learning_rate": 0.0006,
+ "loss": 7.342191696166992,
+ "step": 2999
+ },
+ {
+ "epoch": 41.670742358078606,
+ "grad_norm": 0.006068122107535601,
+ "learning_rate": 0.0006,
+ "loss": 7.321015357971191,
+ "step": 3000
+ },
+ {
+ "epoch": 41.68471615720524,
+ "grad_norm": 0.007316991221159697,
+ "learning_rate": 0.0006,
+ "loss": 7.346490859985352,
+ "step": 3001
+ },
+ {
+ "epoch": 41.698689956331876,
+ "grad_norm": 0.0067388564348220825,
+ "learning_rate": 0.0006,
+ "loss": 7.340723991394043,
+ "step": 3002
+ },
+ {
+ "epoch": 41.712663755458514,
+ "grad_norm": 0.006839963141828775,
+ "learning_rate": 0.0006,
+ "loss": 7.329365253448486,
+ "step": 3003
+ },
+ {
+ "epoch": 41.72663755458515,
+ "grad_norm": 0.00805041566491127,
+ "learning_rate": 0.0006,
+ "loss": 7.298934459686279,
+ "step": 3004
+ },
+ {
+ "epoch": 41.74061135371179,
+ "grad_norm": 0.007197503466159105,
+ "learning_rate": 0.0006,
+ "loss": 7.331496238708496,
+ "step": 3005
+ },
+ {
+ "epoch": 41.75458515283843,
+ "grad_norm": 0.007220805156975985,
+ "learning_rate": 0.0006,
+ "loss": 7.317695140838623,
+ "step": 3006
+ },
+ {
+ "epoch": 41.76855895196506,
+ "grad_norm": 0.00874276738613844,
+ "learning_rate": 0.0006,
+ "loss": 7.272614479064941,
+ "step": 3007
+ },
+ {
+ "epoch": 41.7825327510917,
+ "grad_norm": 0.008386565372347832,
+ "learning_rate": 0.0006,
+ "loss": 7.295244216918945,
+ "step": 3008
+ },
+ {
+ "epoch": 41.79650655021834,
+ "grad_norm": 0.0077730282209813595,
+ "learning_rate": 0.0006,
+ "loss": 7.3395490646362305,
+ "step": 3009
+ },
+ {
+ "epoch": 41.81048034934498,
+ "grad_norm": 0.008441480807960033,
+ "learning_rate": 0.0006,
+ "loss": 7.319921493530273,
+ "step": 3010
+ },
+ {
+ "epoch": 41.82445414847162,
+ "grad_norm": 0.007398269604891539,
+ "learning_rate": 0.0006,
+ "loss": 7.3029465675354,
+ "step": 3011
+ },
+ {
+ "epoch": 41.838427947598255,
+ "grad_norm": 0.006878677289932966,
+ "learning_rate": 0.0006,
+ "loss": 7.318467140197754,
+ "step": 3012
+ },
+ {
+ "epoch": 41.852401746724894,
+ "grad_norm": 0.007616598624736071,
+ "learning_rate": 0.0006,
+ "loss": 7.3112077713012695,
+ "step": 3013
+ },
+ {
+ "epoch": 41.866375545851525,
+ "grad_norm": 0.011629699729382992,
+ "learning_rate": 0.0006,
+ "loss": 7.32523250579834,
+ "step": 3014
+ },
+ {
+ "epoch": 41.880349344978164,
+ "grad_norm": 0.011011249385774136,
+ "learning_rate": 0.0006,
+ "loss": 7.287247657775879,
+ "step": 3015
+ },
+ {
+ "epoch": 41.8943231441048,
+ "grad_norm": 0.010208295658230782,
+ "learning_rate": 0.0006,
+ "loss": 7.292117118835449,
+ "step": 3016
+ },
+ {
+ "epoch": 41.90829694323144,
+ "grad_norm": 0.008651095442473888,
+ "learning_rate": 0.0006,
+ "loss": 7.302201271057129,
+ "step": 3017
+ },
+ {
+ "epoch": 41.92227074235808,
+ "grad_norm": 0.006692920345813036,
+ "learning_rate": 0.0006,
+ "loss": 7.319568634033203,
+ "step": 3018
+ },
+ {
+ "epoch": 41.93624454148472,
+ "grad_norm": 0.010033348575234413,
+ "learning_rate": 0.0006,
+ "loss": 7.2738566398620605,
+ "step": 3019
+ },
+ {
+ "epoch": 41.95021834061135,
+ "grad_norm": 0.011790832504630089,
+ "learning_rate": 0.0006,
+ "loss": 7.298540115356445,
+ "step": 3020
+ },
+ {
+ "epoch": 41.96419213973799,
+ "grad_norm": 0.018622662872076035,
+ "learning_rate": 0.0006,
+ "loss": 7.267182350158691,
+ "step": 3021
+ },
+ {
+ "epoch": 41.97816593886463,
+ "grad_norm": 0.02520257607102394,
+ "learning_rate": 0.0006,
+ "loss": 7.261566162109375,
+ "step": 3022
+ },
+ {
+ "epoch": 41.992139737991266,
+ "grad_norm": 0.036332644522190094,
+ "learning_rate": 0.0006,
+ "loss": 7.270719051361084,
+ "step": 3023
+ },
+ {
+ "epoch": 42.0,
+ "grad_norm": 0.04717305675148964,
+ "learning_rate": 0.0006,
+ "loss": 7.293332099914551,
+ "step": 3024
+ },
+ {
+ "epoch": 42.0,
+ "eval_loss": 7.354280471801758,
+ "eval_runtime": 56.4353,
+ "eval_samples_per_second": 43.271,
+ "eval_steps_per_second": 1.364,
+ "step": 3024
+ },
+ {
+ "epoch": 42.01397379912664,
+ "grad_norm": 0.06661626696586609,
+ "learning_rate": 0.0006,
+ "loss": 7.338189601898193,
+ "step": 3025
+ },
+ {
+ "epoch": 42.02794759825328,
+ "grad_norm": 0.05452054366469383,
+ "learning_rate": 0.0006,
+ "loss": 7.316764831542969,
+ "step": 3026
+ },
+ {
+ "epoch": 42.041921397379916,
+ "grad_norm": 0.025638775900006294,
+ "learning_rate": 0.0006,
+ "loss": 7.34204626083374,
+ "step": 3027
+ },
+ {
+ "epoch": 42.05589519650655,
+ "grad_norm": 0.022398799657821655,
+ "learning_rate": 0.0006,
+ "loss": 7.262149333953857,
+ "step": 3028
+ },
+ {
+ "epoch": 42.069868995633186,
+ "grad_norm": 0.03713499382138252,
+ "learning_rate": 0.0006,
+ "loss": 7.27695369720459,
+ "step": 3029
+ },
+ {
+ "epoch": 42.083842794759825,
+ "grad_norm": 0.014949247241020203,
+ "learning_rate": 0.0006,
+ "loss": 7.296508312225342,
+ "step": 3030
+ },
+ {
+ "epoch": 42.09781659388646,
+ "grad_norm": 0.02703995071351528,
+ "learning_rate": 0.0006,
+ "loss": 7.258018493652344,
+ "step": 3031
+ },
+ {
+ "epoch": 42.1117903930131,
+ "grad_norm": 0.020352816209197044,
+ "learning_rate": 0.0006,
+ "loss": 7.241450309753418,
+ "step": 3032
+ },
+ {
+ "epoch": 42.12576419213974,
+ "grad_norm": 0.014578479342162609,
+ "learning_rate": 0.0006,
+ "loss": 7.257323265075684,
+ "step": 3033
+ },
+ {
+ "epoch": 42.13973799126637,
+ "grad_norm": 0.026053044945001602,
+ "learning_rate": 0.0006,
+ "loss": 7.254680633544922,
+ "step": 3034
+ },
+ {
+ "epoch": 42.15371179039301,
+ "grad_norm": 0.015120275318622589,
+ "learning_rate": 0.0006,
+ "loss": 7.261074066162109,
+ "step": 3035
+ },
+ {
+ "epoch": 42.16768558951965,
+ "grad_norm": 0.016822177916765213,
+ "learning_rate": 0.0006,
+ "loss": 7.279064178466797,
+ "step": 3036
+ },
+ {
+ "epoch": 42.18165938864629,
+ "grad_norm": 0.016367588192224503,
+ "learning_rate": 0.0006,
+ "loss": 7.251429557800293,
+ "step": 3037
+ },
+ {
+ "epoch": 42.19563318777293,
+ "grad_norm": 0.014350418001413345,
+ "learning_rate": 0.0006,
+ "loss": 7.259264945983887,
+ "step": 3038
+ },
+ {
+ "epoch": 42.209606986899566,
+ "grad_norm": 0.017078228294849396,
+ "learning_rate": 0.0006,
+ "loss": 7.227105140686035,
+ "step": 3039
+ },
+ {
+ "epoch": 42.223580786026204,
+ "grad_norm": 0.011759317480027676,
+ "learning_rate": 0.0006,
+ "loss": 7.275888442993164,
+ "step": 3040
+ },
+ {
+ "epoch": 42.237554585152836,
+ "grad_norm": 0.013019340112805367,
+ "learning_rate": 0.0006,
+ "loss": 7.2517008781433105,
+ "step": 3041
+ },
+ {
+ "epoch": 42.251528384279474,
+ "grad_norm": 0.01178441196680069,
+ "learning_rate": 0.0006,
+ "loss": 7.23576545715332,
+ "step": 3042
+ },
+ {
+ "epoch": 42.26550218340611,
+ "grad_norm": 0.014518891461193562,
+ "learning_rate": 0.0006,
+ "loss": 7.253895282745361,
+ "step": 3043
+ },
+ {
+ "epoch": 42.27947598253275,
+ "grad_norm": 0.018159950152039528,
+ "learning_rate": 0.0006,
+ "loss": 7.278058052062988,
+ "step": 3044
+ },
+ {
+ "epoch": 42.29344978165939,
+ "grad_norm": 0.007587796077132225,
+ "learning_rate": 0.0006,
+ "loss": 7.278578758239746,
+ "step": 3045
+ },
+ {
+ "epoch": 42.30742358078603,
+ "grad_norm": 0.015994979068636894,
+ "learning_rate": 0.0006,
+ "loss": 7.239841461181641,
+ "step": 3046
+ },
+ {
+ "epoch": 42.32139737991266,
+ "grad_norm": 0.017830168828368187,
+ "learning_rate": 0.0006,
+ "loss": 7.2443976402282715,
+ "step": 3047
+ },
+ {
+ "epoch": 42.3353711790393,
+ "grad_norm": 0.008738759905099869,
+ "learning_rate": 0.0006,
+ "loss": 7.232264518737793,
+ "step": 3048
+ },
+ {
+ "epoch": 42.34934497816594,
+ "grad_norm": 0.0175449438393116,
+ "learning_rate": 0.0006,
+ "loss": 7.228000640869141,
+ "step": 3049
+ },
+ {
+ "epoch": 42.36331877729258,
+ "grad_norm": 0.019611820578575134,
+ "learning_rate": 0.0006,
+ "loss": 7.234999179840088,
+ "step": 3050
+ },
+ {
+ "epoch": 42.377292576419215,
+ "grad_norm": 0.015302863903343678,
+ "learning_rate": 0.0006,
+ "loss": 7.242547988891602,
+ "step": 3051
+ },
+ {
+ "epoch": 42.391266375545854,
+ "grad_norm": 0.025919731706380844,
+ "learning_rate": 0.0006,
+ "loss": 7.216652870178223,
+ "step": 3052
+ },
+ {
+ "epoch": 42.40524017467249,
+ "grad_norm": 0.034587468951940536,
+ "learning_rate": 0.0006,
+ "loss": 7.245336532592773,
+ "step": 3053
+ },
+ {
+ "epoch": 42.419213973799124,
+ "grad_norm": 0.019198840484023094,
+ "learning_rate": 0.0006,
+ "loss": 7.256758689880371,
+ "step": 3054
+ },
+ {
+ "epoch": 42.43318777292576,
+ "grad_norm": 0.012316389009356499,
+ "learning_rate": 0.0006,
+ "loss": 7.209102630615234,
+ "step": 3055
+ },
+ {
+ "epoch": 42.4471615720524,
+ "grad_norm": 0.023512162268161774,
+ "learning_rate": 0.0006,
+ "loss": 7.250742435455322,
+ "step": 3056
+ },
+ {
+ "epoch": 42.46113537117904,
+ "grad_norm": 0.025457851588726044,
+ "learning_rate": 0.0006,
+ "loss": 7.263757705688477,
+ "step": 3057
+ },
+ {
+ "epoch": 42.47510917030568,
+ "grad_norm": 0.02192617766559124,
+ "learning_rate": 0.0006,
+ "loss": 7.246977806091309,
+ "step": 3058
+ },
+ {
+ "epoch": 42.48908296943232,
+ "grad_norm": 0.02541281282901764,
+ "learning_rate": 0.0006,
+ "loss": 7.232361316680908,
+ "step": 3059
+ },
+ {
+ "epoch": 42.50305676855895,
+ "grad_norm": 0.036353763192892075,
+ "learning_rate": 0.0006,
+ "loss": 7.242556571960449,
+ "step": 3060
+ },
+ {
+ "epoch": 42.51703056768559,
+ "grad_norm": 0.04555872827768326,
+ "learning_rate": 0.0006,
+ "loss": 7.228940963745117,
+ "step": 3061
+ },
+ {
+ "epoch": 42.531004366812226,
+ "grad_norm": 0.04563378915190697,
+ "learning_rate": 0.0006,
+ "loss": 7.235276222229004,
+ "step": 3062
+ },
+ {
+ "epoch": 42.544978165938865,
+ "grad_norm": 0.04275760427117348,
+ "learning_rate": 0.0006,
+ "loss": 7.238895416259766,
+ "step": 3063
+ },
+ {
+ "epoch": 42.5589519650655,
+ "grad_norm": 0.04535885155200958,
+ "learning_rate": 0.0006,
+ "loss": 7.205336570739746,
+ "step": 3064
+ },
+ {
+ "epoch": 42.57292576419214,
+ "grad_norm": 0.015534617006778717,
+ "learning_rate": 0.0006,
+ "loss": 7.238393783569336,
+ "step": 3065
+ },
+ {
+ "epoch": 42.58689956331878,
+ "grad_norm": 0.03286299481987953,
+ "learning_rate": 0.0006,
+ "loss": 7.216453552246094,
+ "step": 3066
+ },
+ {
+ "epoch": 42.60087336244541,
+ "grad_norm": 0.025453168898820877,
+ "learning_rate": 0.0006,
+ "loss": 7.203750133514404,
+ "step": 3067
+ },
+ {
+ "epoch": 42.61484716157205,
+ "grad_norm": 0.021501649171113968,
+ "learning_rate": 0.0006,
+ "loss": 7.241621971130371,
+ "step": 3068
+ },
+ {
+ "epoch": 42.62882096069869,
+ "grad_norm": 0.02743634767830372,
+ "learning_rate": 0.0006,
+ "loss": 7.22076940536499,
+ "step": 3069
+ },
+ {
+ "epoch": 42.64279475982533,
+ "grad_norm": 0.020890483632683754,
+ "learning_rate": 0.0006,
+ "loss": 7.218412399291992,
+ "step": 3070
+ },
+ {
+ "epoch": 42.65676855895197,
+ "grad_norm": 0.022499850019812584,
+ "learning_rate": 0.0006,
+ "loss": 7.217696189880371,
+ "step": 3071
+ },
+ {
+ "epoch": 42.670742358078606,
+ "grad_norm": 0.01653696969151497,
+ "learning_rate": 0.0006,
+ "loss": 7.156417369842529,
+ "step": 3072
+ },
+ {
+ "epoch": 42.68471615720524,
+ "grad_norm": 0.01834225095808506,
+ "learning_rate": 0.0006,
+ "loss": 7.211247444152832,
+ "step": 3073
+ },
+ {
+ "epoch": 42.698689956331876,
+ "grad_norm": 0.020137229934334755,
+ "learning_rate": 0.0006,
+ "loss": 7.232499599456787,
+ "step": 3074
+ },
+ {
+ "epoch": 42.712663755458514,
+ "grad_norm": 0.012671459466218948,
+ "learning_rate": 0.0006,
+ "loss": 7.172441482543945,
+ "step": 3075
+ },
+ {
+ "epoch": 42.72663755458515,
+ "grad_norm": 0.016803130507469177,
+ "learning_rate": 0.0006,
+ "loss": 7.152103424072266,
+ "step": 3076
+ },
+ {
+ "epoch": 42.74061135371179,
+ "grad_norm": 0.01623712107539177,
+ "learning_rate": 0.0006,
+ "loss": 7.187032699584961,
+ "step": 3077
+ },
+ {
+ "epoch": 42.75458515283843,
+ "grad_norm": 0.015938356518745422,
+ "learning_rate": 0.0006,
+ "loss": 7.171140670776367,
+ "step": 3078
+ },
+ {
+ "epoch": 42.76855895196506,
+ "grad_norm": 0.02580736204981804,
+ "learning_rate": 0.0006,
+ "loss": 7.204472541809082,
+ "step": 3079
+ },
+ {
+ "epoch": 42.7825327510917,
+ "grad_norm": 0.023372117429971695,
+ "learning_rate": 0.0006,
+ "loss": 7.165554523468018,
+ "step": 3080
+ },
+ {
+ "epoch": 42.79650655021834,
+ "grad_norm": 0.04440586268901825,
+ "learning_rate": 0.0006,
+ "loss": 7.169431209564209,
+ "step": 3081
+ },
+ {
+ "epoch": 42.81048034934498,
+ "grad_norm": 0.06797406077384949,
+ "learning_rate": 0.0006,
+ "loss": 7.168193340301514,
+ "step": 3082
+ },
+ {
+ "epoch": 42.82445414847162,
+ "grad_norm": 0.0914321169257164,
+ "learning_rate": 0.0006,
+ "loss": 7.204671859741211,
+ "step": 3083
+ },
+ {
+ "epoch": 42.838427947598255,
+ "grad_norm": 0.0506940595805645,
+ "learning_rate": 0.0006,
+ "loss": 7.15949821472168,
+ "step": 3084
+ },
+ {
+ "epoch": 42.852401746724894,
+ "grad_norm": 0.04819890484213829,
+ "learning_rate": 0.0006,
+ "loss": 7.192837238311768,
+ "step": 3085
+ },
+ {
+ "epoch": 42.866375545851525,
+ "grad_norm": 0.040044233202934265,
+ "learning_rate": 0.0006,
+ "loss": 7.1600165367126465,
+ "step": 3086
+ },
+ {
+ "epoch": 42.880349344978164,
+ "grad_norm": 0.02573299966752529,
+ "learning_rate": 0.0006,
+ "loss": 7.153427600860596,
+ "step": 3087
+ },
+ {
+ "epoch": 42.8943231441048,
+ "grad_norm": 0.046942535787820816,
+ "learning_rate": 0.0006,
+ "loss": 7.16767692565918,
+ "step": 3088
+ },
+ {
+ "epoch": 42.90829694323144,
+ "grad_norm": 0.02683161571621895,
+ "learning_rate": 0.0006,
+ "loss": 7.17144250869751,
+ "step": 3089
+ },
+ {
+ "epoch": 42.92227074235808,
+ "grad_norm": 0.05363674834370613,
+ "learning_rate": 0.0006,
+ "loss": 7.120128631591797,
+ "step": 3090
+ },
+ {
+ "epoch": 42.93624454148472,
+ "grad_norm": 0.023756247013807297,
+ "learning_rate": 0.0006,
+ "loss": 7.169703006744385,
+ "step": 3091
+ },
+ {
+ "epoch": 42.95021834061135,
+ "grad_norm": 0.04216783866286278,
+ "learning_rate": 0.0006,
+ "loss": 7.1206865310668945,
+ "step": 3092
+ },
+ {
+ "epoch": 42.96419213973799,
+ "grad_norm": 0.02219565585255623,
+ "learning_rate": 0.0006,
+ "loss": 7.157215118408203,
+ "step": 3093
+ },
+ {
+ "epoch": 42.97816593886463,
+ "grad_norm": 0.03527127206325531,
+ "learning_rate": 0.0006,
+ "loss": 7.166672706604004,
+ "step": 3094
+ },
+ {
+ "epoch": 42.992139737991266,
+ "grad_norm": 0.017748817801475525,
+ "learning_rate": 0.0006,
+ "loss": 7.154747009277344,
+ "step": 3095
+ },
+ {
+ "epoch": 43.0,
+ "grad_norm": 0.033165838569402695,
+ "learning_rate": 0.0006,
+ "loss": 7.150866508483887,
+ "step": 3096
+ },
+ {
+ "epoch": 43.0,
+ "eval_loss": 7.144403457641602,
+ "eval_runtime": 56.8853,
+ "eval_samples_per_second": 42.928,
+ "eval_steps_per_second": 1.354,
+ "step": 3096
+ },
+ {
+ "epoch": 43.01397379912664,
+ "grad_norm": 0.015199829824268818,
+ "learning_rate": 0.0006,
+ "loss": 7.122735023498535,
+ "step": 3097
+ },
+ {
+ "epoch": 43.02794759825328,
+ "grad_norm": 0.03220256045460701,
+ "learning_rate": 0.0006,
+ "loss": 7.143553733825684,
+ "step": 3098
+ },
+ {
+ "epoch": 43.041921397379916,
+ "grad_norm": 0.014687920920550823,
+ "learning_rate": 0.0006,
+ "loss": 7.080960750579834,
+ "step": 3099
+ },
+ {
+ "epoch": 43.05589519650655,
+ "grad_norm": 0.023522932082414627,
+ "learning_rate": 0.0006,
+ "loss": 7.137356758117676,
+ "step": 3100
+ },
+ {
+ "epoch": 43.069868995633186,
+ "grad_norm": 0.014808216132223606,
+ "learning_rate": 0.0006,
+ "loss": 7.09143590927124,
+ "step": 3101
+ },
+ {
+ "epoch": 43.083842794759825,
+ "grad_norm": 0.029339246451854706,
+ "learning_rate": 0.0006,
+ "loss": 7.122198104858398,
+ "step": 3102
+ },
+ {
+ "epoch": 43.09781659388646,
+ "grad_norm": 0.022953078150749207,
+ "learning_rate": 0.0006,
+ "loss": 7.092505931854248,
+ "step": 3103
+ },
+ {
+ "epoch": 43.1117903930131,
+ "grad_norm": 0.04590052738785744,
+ "learning_rate": 0.0006,
+ "loss": 7.113100051879883,
+ "step": 3104
+ },
+ {
+ "epoch": 43.12576419213974,
+ "grad_norm": 0.0524442233145237,
+ "learning_rate": 0.0006,
+ "loss": 7.147022247314453,
+ "step": 3105
+ },
+ {
+ "epoch": 43.13973799126637,
+ "grad_norm": 0.0315794013440609,
+ "learning_rate": 0.0006,
+ "loss": 7.094045162200928,
+ "step": 3106
+ },
+ {
+ "epoch": 43.15371179039301,
+ "grad_norm": 0.05112722888588905,
+ "learning_rate": 0.0006,
+ "loss": 7.054388523101807,
+ "step": 3107
+ },
+ {
+ "epoch": 43.16768558951965,
+ "grad_norm": 0.03146776929497719,
+ "learning_rate": 0.0006,
+ "loss": 7.116674423217773,
+ "step": 3108
+ },
+ {
+ "epoch": 43.18165938864629,
+ "grad_norm": 0.03391246497631073,
+ "learning_rate": 0.0006,
+ "loss": 7.150345802307129,
+ "step": 3109
+ },
+ {
+ "epoch": 43.19563318777293,
+ "grad_norm": 0.026957903057336807,
+ "learning_rate": 0.0006,
+ "loss": 7.058531284332275,
+ "step": 3110
+ },
+ {
+ "epoch": 43.209606986899566,
+ "grad_norm": 0.041532568633556366,
+ "learning_rate": 0.0006,
+ "loss": 7.085498809814453,
+ "step": 3111
+ },
+ {
+ "epoch": 43.223580786026204,
+ "grad_norm": 0.020001588389277458,
+ "learning_rate": 0.0006,
+ "loss": 7.131487846374512,
+ "step": 3112
+ },
+ {
+ "epoch": 43.237554585152836,
+ "grad_norm": 0.02268712781369686,
+ "learning_rate": 0.0006,
+ "loss": 7.0790863037109375,
+ "step": 3113
+ },
+ {
+ "epoch": 43.251528384279474,
+ "grad_norm": 0.023363925516605377,
+ "learning_rate": 0.0006,
+ "loss": 7.067816734313965,
+ "step": 3114
+ },
+ {
+ "epoch": 43.26550218340611,
+ "grad_norm": 0.019367508590221405,
+ "learning_rate": 0.0006,
+ "loss": 7.042041778564453,
+ "step": 3115
+ },
+ {
+ "epoch": 43.27947598253275,
+ "grad_norm": 0.02745090238749981,
+ "learning_rate": 0.0006,
+ "loss": 7.064733028411865,
+ "step": 3116
+ },
+ {
+ "epoch": 43.29344978165939,
+ "grad_norm": 0.02543802745640278,
+ "learning_rate": 0.0006,
+ "loss": 7.058948040008545,
+ "step": 3117
+ },
+ {
+ "epoch": 43.30742358078603,
+ "grad_norm": 0.018874092027544975,
+ "learning_rate": 0.0006,
+ "loss": 7.009946346282959,
+ "step": 3118
+ },
+ {
+ "epoch": 43.32139737991266,
+ "grad_norm": 0.024662822484970093,
+ "learning_rate": 0.0006,
+ "loss": 7.015913963317871,
+ "step": 3119
+ },
+ {
+ "epoch": 43.3353711790393,
+ "grad_norm": 0.02441416308283806,
+ "learning_rate": 0.0006,
+ "loss": 7.061622142791748,
+ "step": 3120
+ },
+ {
+ "epoch": 43.34934497816594,
+ "grad_norm": 0.031887322664260864,
+ "learning_rate": 0.0006,
+ "loss": 7.046883583068848,
+ "step": 3121
+ },
+ {
+ "epoch": 43.36331877729258,
+ "grad_norm": 0.021930739283561707,
+ "learning_rate": 0.0006,
+ "loss": 7.0309624671936035,
+ "step": 3122
+ },
+ {
+ "epoch": 43.377292576419215,
+ "grad_norm": 0.02722783200442791,
+ "learning_rate": 0.0006,
+ "loss": 7.072924613952637,
+ "step": 3123
+ },
+ {
+ "epoch": 43.391266375545854,
+ "grad_norm": 0.01483179721981287,
+ "learning_rate": 0.0006,
+ "loss": 7.028624534606934,
+ "step": 3124
+ },
+ {
+ "epoch": 43.40524017467249,
+ "grad_norm": 0.024793945252895355,
+ "learning_rate": 0.0006,
+ "loss": 7.001623153686523,
+ "step": 3125
+ },
+ {
+ "epoch": 43.419213973799124,
+ "grad_norm": 0.020148713141679764,
+ "learning_rate": 0.0006,
+ "loss": 6.9938812255859375,
+ "step": 3126
+ },
+ {
+ "epoch": 43.43318777292576,
+ "grad_norm": 0.02413850650191307,
+ "learning_rate": 0.0006,
+ "loss": 7.001243591308594,
+ "step": 3127
+ },
+ {
+ "epoch": 43.4471615720524,
+ "grad_norm": 0.028549203649163246,
+ "learning_rate": 0.0006,
+ "loss": 6.987671375274658,
+ "step": 3128
+ },
+ {
+ "epoch": 43.46113537117904,
+ "grad_norm": 0.025988785549998283,
+ "learning_rate": 0.0006,
+ "loss": 6.9551167488098145,
+ "step": 3129
+ },
+ {
+ "epoch": 43.47510917030568,
+ "grad_norm": 0.027140729129314423,
+ "learning_rate": 0.0006,
+ "loss": 6.9547810554504395,
+ "step": 3130
+ },
+ {
+ "epoch": 43.48908296943232,
+ "grad_norm": 0.030130082741379738,
+ "learning_rate": 0.0006,
+ "loss": 6.993186950683594,
+ "step": 3131
+ },
+ {
+ "epoch": 43.50305676855895,
+ "grad_norm": 0.025996994227170944,
+ "learning_rate": 0.0006,
+ "loss": 6.984624862670898,
+ "step": 3132
+ },
+ {
+ "epoch": 43.51703056768559,
+ "grad_norm": 0.02423008903861046,
+ "learning_rate": 0.0006,
+ "loss": 6.98813009262085,
+ "step": 3133
+ },
+ {
+ "epoch": 43.531004366812226,
+ "grad_norm": 0.029665837064385414,
+ "learning_rate": 0.0006,
+ "loss": 6.986610412597656,
+ "step": 3134
+ },
+ {
+ "epoch": 43.544978165938865,
+ "grad_norm": 0.028768595308065414,
+ "learning_rate": 0.0006,
+ "loss": 6.898797035217285,
+ "step": 3135
+ },
+ {
+ "epoch": 43.5589519650655,
+ "grad_norm": 0.03111208975315094,
+ "learning_rate": 0.0006,
+ "loss": 6.923741817474365,
+ "step": 3136
+ },
+ {
+ "epoch": 43.57292576419214,
+ "grad_norm": 0.0368187390267849,
+ "learning_rate": 0.0006,
+ "loss": 6.994343280792236,
+ "step": 3137
+ },
+ {
+ "epoch": 43.58689956331878,
+ "grad_norm": 0.04076763987541199,
+ "learning_rate": 0.0006,
+ "loss": 6.909458160400391,
+ "step": 3138
+ },
+ {
+ "epoch": 43.60087336244541,
+ "grad_norm": 0.03281061723828316,
+ "learning_rate": 0.0006,
+ "loss": 6.924851417541504,
+ "step": 3139
+ },
+ {
+ "epoch": 43.61484716157205,
+ "grad_norm": 0.025486988946795464,
+ "learning_rate": 0.0006,
+ "loss": 6.962926864624023,
+ "step": 3140
+ },
+ {
+ "epoch": 43.62882096069869,
+ "grad_norm": 0.04583485797047615,
+ "learning_rate": 0.0006,
+ "loss": 6.876555442810059,
+ "step": 3141
+ },
+ {
+ "epoch": 43.64279475982533,
+ "grad_norm": 0.05460785701870918,
+ "learning_rate": 0.0006,
+ "loss": 6.952932357788086,
+ "step": 3142
+ },
+ {
+ "epoch": 43.65676855895197,
+ "grad_norm": 0.040917906910181046,
+ "learning_rate": 0.0006,
+ "loss": 6.9624924659729,
+ "step": 3143
+ },
+ {
+ "epoch": 43.670742358078606,
+ "grad_norm": 0.04009668901562691,
+ "learning_rate": 0.0006,
+ "loss": 6.94033145904541,
+ "step": 3144
+ },
+ {
+ "epoch": 43.68471615720524,
+ "grad_norm": 0.04500625655055046,
+ "learning_rate": 0.0006,
+ "loss": 6.961839199066162,
+ "step": 3145
+ },
+ {
+ "epoch": 43.698689956331876,
+ "grad_norm": 0.03182258829474449,
+ "learning_rate": 0.0006,
+ "loss": 6.930580139160156,
+ "step": 3146
+ },
+ {
+ "epoch": 43.712663755458514,
+ "grad_norm": 0.053621504455804825,
+ "learning_rate": 0.0006,
+ "loss": 6.950235366821289,
+ "step": 3147
+ },
+ {
+ "epoch": 43.72663755458515,
+ "grad_norm": 0.031919289380311966,
+ "learning_rate": 0.0006,
+ "loss": 6.895111560821533,
+ "step": 3148
+ },
+ {
+ "epoch": 43.74061135371179,
+ "grad_norm": 0.042713891714811325,
+ "learning_rate": 0.0006,
+ "loss": 6.940885543823242,
+ "step": 3149
+ },
+ {
+ "epoch": 43.75458515283843,
+ "grad_norm": 0.02385653182864189,
+ "learning_rate": 0.0006,
+ "loss": 6.948801517486572,
+ "step": 3150
+ },
+ {
+ "epoch": 43.76855895196506,
+ "grad_norm": 0.04100663587450981,
+ "learning_rate": 0.0006,
+ "loss": 6.970411777496338,
+ "step": 3151
+ },
+ {
+ "epoch": 43.7825327510917,
+ "grad_norm": 0.02729583904147148,
+ "learning_rate": 0.0006,
+ "loss": 6.899379253387451,
+ "step": 3152
+ },
+ {
+ "epoch": 43.79650655021834,
+ "grad_norm": 0.03128620237112045,
+ "learning_rate": 0.0006,
+ "loss": 6.931319236755371,
+ "step": 3153
+ },
+ {
+ "epoch": 43.81048034934498,
+ "grad_norm": 0.025488659739494324,
+ "learning_rate": 0.0006,
+ "loss": 6.8486409187316895,
+ "step": 3154
+ },
+ {
+ "epoch": 43.82445414847162,
+ "grad_norm": 0.03679376840591431,
+ "learning_rate": 0.0006,
+ "loss": 6.840811252593994,
+ "step": 3155
+ },
+ {
+ "epoch": 43.838427947598255,
+ "grad_norm": 0.020578034222126007,
+ "learning_rate": 0.0006,
+ "loss": 6.9088616371154785,
+ "step": 3156
+ },
+ {
+ "epoch": 43.852401746724894,
+ "grad_norm": 0.02935841865837574,
+ "learning_rate": 0.0006,
+ "loss": 6.806676864624023,
+ "step": 3157
+ },
+ {
+ "epoch": 43.866375545851525,
+ "grad_norm": 0.023319244384765625,
+ "learning_rate": 0.0006,
+ "loss": 6.878457546234131,
+ "step": 3158
+ },
+ {
+ "epoch": 43.880349344978164,
+ "grad_norm": 0.02779589593410492,
+ "learning_rate": 0.0006,
+ "loss": 6.822447776794434,
+ "step": 3159
+ },
+ {
+ "epoch": 43.8943231441048,
+ "grad_norm": 0.01879825070500374,
+ "learning_rate": 0.0006,
+ "loss": 6.854121208190918,
+ "step": 3160
+ },
+ {
+ "epoch": 43.90829694323144,
+ "grad_norm": 0.023286139592528343,
+ "learning_rate": 0.0006,
+ "loss": 6.836827278137207,
+ "step": 3161
+ },
+ {
+ "epoch": 43.92227074235808,
+ "grad_norm": 0.030358297750353813,
+ "learning_rate": 0.0006,
+ "loss": 6.860037803649902,
+ "step": 3162
+ },
+ {
+ "epoch": 43.93624454148472,
+ "grad_norm": 0.029621105641126633,
+ "learning_rate": 0.0006,
+ "loss": 6.884025573730469,
+ "step": 3163
+ },
+ {
+ "epoch": 43.95021834061135,
+ "grad_norm": 0.029829824343323708,
+ "learning_rate": 0.0006,
+ "loss": 6.8702392578125,
+ "step": 3164
+ },
+ {
+ "epoch": 43.96419213973799,
+ "grad_norm": 0.036892473697662354,
+ "learning_rate": 0.0006,
+ "loss": 6.806652069091797,
+ "step": 3165
+ },
+ {
+ "epoch": 43.97816593886463,
+ "grad_norm": 0.03206709027290344,
+ "learning_rate": 0.0006,
+ "loss": 6.828495979309082,
+ "step": 3166
+ },
+ {
+ "epoch": 43.992139737991266,
+ "grad_norm": 0.02744000405073166,
+ "learning_rate": 0.0006,
+ "loss": 6.832002639770508,
+ "step": 3167
+ },
+ {
+ "epoch": 44.0,
+ "grad_norm": 0.02205660194158554,
+ "learning_rate": 0.0006,
+ "loss": 6.898828506469727,
+ "step": 3168
+ },
+ {
+ "epoch": 44.0,
+ "eval_loss": 6.8150553703308105,
+ "eval_runtime": 56.784,
+ "eval_samples_per_second": 43.005,
+ "eval_steps_per_second": 1.356,
+ "step": 3168
+ },
+ {
+ "epoch": 44.01397379912664,
+ "grad_norm": 0.030421260744333267,
+ "learning_rate": 0.0006,
+ "loss": 6.775057792663574,
+ "step": 3169
+ },
+ {
+ "epoch": 44.02794759825328,
+ "grad_norm": 0.0465611033141613,
+ "learning_rate": 0.0006,
+ "loss": 6.849390506744385,
+ "step": 3170
+ },
+ {
+ "epoch": 44.041921397379916,
+ "grad_norm": 0.05977972596883774,
+ "learning_rate": 0.0006,
+ "loss": 6.725576400756836,
+ "step": 3171
+ },
+ {
+ "epoch": 44.05589519650655,
+ "grad_norm": 0.05432697385549545,
+ "learning_rate": 0.0006,
+ "loss": 6.768024444580078,
+ "step": 3172
+ },
+ {
+ "epoch": 44.069868995633186,
+ "grad_norm": 0.04850967228412628,
+ "learning_rate": 0.0006,
+ "loss": 6.799457550048828,
+ "step": 3173
+ },
+ {
+ "epoch": 44.083842794759825,
+ "grad_norm": 0.06201786920428276,
+ "learning_rate": 0.0006,
+ "loss": 6.897234916687012,
+ "step": 3174
+ },
+ {
+ "epoch": 44.09781659388646,
+ "grad_norm": 0.04695591703057289,
+ "learning_rate": 0.0006,
+ "loss": 6.826470375061035,
+ "step": 3175
+ },
+ {
+ "epoch": 44.1117903930131,
+ "grad_norm": 0.03850164636969566,
+ "learning_rate": 0.0006,
+ "loss": 6.835901260375977,
+ "step": 3176
+ },
+ {
+ "epoch": 44.12576419213974,
+ "grad_norm": 0.04250155761837959,
+ "learning_rate": 0.0006,
+ "loss": 6.782310485839844,
+ "step": 3177
+ },
+ {
+ "epoch": 44.13973799126637,
+ "grad_norm": 0.041465479880571365,
+ "learning_rate": 0.0006,
+ "loss": 6.785732746124268,
+ "step": 3178
+ },
+ {
+ "epoch": 44.15371179039301,
+ "grad_norm": 0.02527807280421257,
+ "learning_rate": 0.0006,
+ "loss": 6.78061580657959,
+ "step": 3179
+ },
+ {
+ "epoch": 44.16768558951965,
+ "grad_norm": 0.03660374879837036,
+ "learning_rate": 0.0006,
+ "loss": 6.783992767333984,
+ "step": 3180
+ },
+ {
+ "epoch": 44.18165938864629,
+ "grad_norm": 0.027567964047193527,
+ "learning_rate": 0.0006,
+ "loss": 6.765413284301758,
+ "step": 3181
+ },
+ {
+ "epoch": 44.19563318777293,
+ "grad_norm": 0.024301424622535706,
+ "learning_rate": 0.0006,
+ "loss": 6.760368347167969,
+ "step": 3182
+ },
+ {
+ "epoch": 44.209606986899566,
+ "grad_norm": 0.0266821701079607,
+ "learning_rate": 0.0006,
+ "loss": 6.71417236328125,
+ "step": 3183
+ },
+ {
+ "epoch": 44.223580786026204,
+ "grad_norm": 0.028125090524554253,
+ "learning_rate": 0.0006,
+ "loss": 6.646126747131348,
+ "step": 3184
+ },
+ {
+ "epoch": 44.237554585152836,
+ "grad_norm": 0.0185737032443285,
+ "learning_rate": 0.0006,
+ "loss": 6.712944030761719,
+ "step": 3185
+ },
+ {
+ "epoch": 44.251528384279474,
+ "grad_norm": 0.025836674496531487,
+ "learning_rate": 0.0006,
+ "loss": 6.707030296325684,
+ "step": 3186
+ },
+ {
+ "epoch": 44.26550218340611,
+ "grad_norm": 0.029744870960712433,
+ "learning_rate": 0.0006,
+ "loss": 6.6271071434021,
+ "step": 3187
+ },
+ {
+ "epoch": 44.27947598253275,
+ "grad_norm": 0.03021426685154438,
+ "learning_rate": 0.0006,
+ "loss": 6.677826404571533,
+ "step": 3188
+ },
+ {
+ "epoch": 44.29344978165939,
+ "grad_norm": 0.032727014273405075,
+ "learning_rate": 0.0006,
+ "loss": 6.657764434814453,
+ "step": 3189
+ },
+ {
+ "epoch": 44.30742358078603,
+ "grad_norm": 0.03696778789162636,
+ "learning_rate": 0.0006,
+ "loss": 6.632883548736572,
+ "step": 3190
+ },
+ {
+ "epoch": 44.32139737991266,
+ "grad_norm": 0.034630827605724335,
+ "learning_rate": 0.0006,
+ "loss": 6.759583473205566,
+ "step": 3191
+ },
+ {
+ "epoch": 44.3353711790393,
+ "grad_norm": 0.030517758801579475,
+ "learning_rate": 0.0006,
+ "loss": 6.702268600463867,
+ "step": 3192
+ },
+ {
+ "epoch": 44.34934497816594,
+ "grad_norm": 0.04105190932750702,
+ "learning_rate": 0.0006,
+ "loss": 6.705792427062988,
+ "step": 3193
+ },
+ {
+ "epoch": 44.36331877729258,
+ "grad_norm": 0.06659174710512161,
+ "learning_rate": 0.0006,
+ "loss": 6.73862361907959,
+ "step": 3194
+ },
+ {
+ "epoch": 44.377292576419215,
+ "grad_norm": 0.08013619482517242,
+ "learning_rate": 0.0006,
+ "loss": 6.710282802581787,
+ "step": 3195
+ },
+ {
+ "epoch": 44.391266375545854,
+ "grad_norm": 0.049850963056087494,
+ "learning_rate": 0.0006,
+ "loss": 6.684974670410156,
+ "step": 3196
+ },
+ {
+ "epoch": 44.40524017467249,
+ "grad_norm": 0.061247408390045166,
+ "learning_rate": 0.0006,
+ "loss": 6.716034412384033,
+ "step": 3197
+ },
+ {
+ "epoch": 44.419213973799124,
+ "grad_norm": 0.05868857726454735,
+ "learning_rate": 0.0006,
+ "loss": 6.589107513427734,
+ "step": 3198
+ },
+ {
+ "epoch": 44.43318777292576,
+ "grad_norm": 0.048878561705350876,
+ "learning_rate": 0.0006,
+ "loss": 6.694855690002441,
+ "step": 3199
+ },
+ {
+ "epoch": 44.4471615720524,
+ "grad_norm": 0.028748324140906334,
+ "learning_rate": 0.0006,
+ "loss": 6.629969596862793,
+ "step": 3200
+ },
+ {
+ "epoch": 44.46113537117904,
+ "grad_norm": 0.04060306400060654,
+ "learning_rate": 0.0006,
+ "loss": 6.721431255340576,
+ "step": 3201
+ },
+ {
+ "epoch": 44.47510917030568,
+ "grad_norm": 0.03833824768662453,
+ "learning_rate": 0.0006,
+ "loss": 6.625029563903809,
+ "step": 3202
+ },
+ {
+ "epoch": 44.48908296943232,
+ "grad_norm": 0.03791997581720352,
+ "learning_rate": 0.0006,
+ "loss": 6.669187545776367,
+ "step": 3203
+ },
+ {
+ "epoch": 44.50305676855895,
+ "grad_norm": 0.03259684890508652,
+ "learning_rate": 0.0006,
+ "loss": 6.665643692016602,
+ "step": 3204
+ },
+ {
+ "epoch": 44.51703056768559,
+ "grad_norm": 0.040658626705408096,
+ "learning_rate": 0.0006,
+ "loss": 6.639796257019043,
+ "step": 3205
+ },
+ {
+ "epoch": 44.531004366812226,
+ "grad_norm": 0.03691193461418152,
+ "learning_rate": 0.0006,
+ "loss": 6.692244529724121,
+ "step": 3206
+ },
+ {
+ "epoch": 44.544978165938865,
+ "grad_norm": 0.02814812771975994,
+ "learning_rate": 0.0006,
+ "loss": 6.61978006362915,
+ "step": 3207
+ },
+ {
+ "epoch": 44.5589519650655,
+ "grad_norm": 0.022155074402689934,
+ "learning_rate": 0.0006,
+ "loss": 6.625905990600586,
+ "step": 3208
+ },
+ {
+ "epoch": 44.57292576419214,
+ "grad_norm": 0.030794885009527206,
+ "learning_rate": 0.0006,
+ "loss": 6.678500652313232,
+ "step": 3209
+ },
+ {
+ "epoch": 44.58689956331878,
+ "grad_norm": 0.03201768547296524,
+ "learning_rate": 0.0006,
+ "loss": 6.6015729904174805,
+ "step": 3210
+ },
+ {
+ "epoch": 44.60087336244541,
+ "grad_norm": 0.025138361379504204,
+ "learning_rate": 0.0006,
+ "loss": 6.568234443664551,
+ "step": 3211
+ },
+ {
+ "epoch": 44.61484716157205,
+ "grad_norm": 0.02969353087246418,
+ "learning_rate": 0.0006,
+ "loss": 6.588366508483887,
+ "step": 3212
+ },
+ {
+ "epoch": 44.62882096069869,
+ "grad_norm": 0.03136962279677391,
+ "learning_rate": 0.0006,
+ "loss": 6.479961395263672,
+ "step": 3213
+ },
+ {
+ "epoch": 44.64279475982533,
+ "grad_norm": 0.030178170651197433,
+ "learning_rate": 0.0006,
+ "loss": 6.622910976409912,
+ "step": 3214
+ },
+ {
+ "epoch": 44.65676855895197,
+ "grad_norm": 0.02682245895266533,
+ "learning_rate": 0.0006,
+ "loss": 6.597707748413086,
+ "step": 3215
+ },
+ {
+ "epoch": 44.670742358078606,
+ "grad_norm": 0.025154557079076767,
+ "learning_rate": 0.0006,
+ "loss": 6.630541801452637,
+ "step": 3216
+ },
+ {
+ "epoch": 44.68471615720524,
+ "grad_norm": 0.03564463183283806,
+ "learning_rate": 0.0006,
+ "loss": 6.542823791503906,
+ "step": 3217
+ },
+ {
+ "epoch": 44.698689956331876,
+ "grad_norm": 0.049477651715278625,
+ "learning_rate": 0.0006,
+ "loss": 6.646030426025391,
+ "step": 3218
+ },
+ {
+ "epoch": 44.712663755458514,
+ "grad_norm": 0.06536535918712616,
+ "learning_rate": 0.0006,
+ "loss": 6.595463275909424,
+ "step": 3219
+ },
+ {
+ "epoch": 44.72663755458515,
+ "grad_norm": 0.07710316777229309,
+ "learning_rate": 0.0006,
+ "loss": 6.55524206161499,
+ "step": 3220
+ },
+ {
+ "epoch": 44.74061135371179,
+ "grad_norm": 0.09571163356304169,
+ "learning_rate": 0.0006,
+ "loss": 6.659764289855957,
+ "step": 3221
+ },
+ {
+ "epoch": 44.75458515283843,
+ "grad_norm": 0.055076830089092255,
+ "learning_rate": 0.0006,
+ "loss": 6.658882141113281,
+ "step": 3222
+ },
+ {
+ "epoch": 44.76855895196506,
+ "grad_norm": 0.07671099156141281,
+ "learning_rate": 0.0006,
+ "loss": 6.605809211730957,
+ "step": 3223
+ },
+ {
+ "epoch": 44.7825327510917,
+ "grad_norm": 0.05295584350824356,
+ "learning_rate": 0.0006,
+ "loss": 6.641455173492432,
+ "step": 3224
+ },
+ {
+ "epoch": 44.79650655021834,
+ "grad_norm": 0.09128042310476303,
+ "learning_rate": 0.0006,
+ "loss": 6.691066741943359,
+ "step": 3225
+ },
+ {
+ "epoch": 44.81048034934498,
+ "grad_norm": 0.05362958461046219,
+ "learning_rate": 0.0006,
+ "loss": 6.573406219482422,
+ "step": 3226
+ },
+ {
+ "epoch": 44.82445414847162,
+ "grad_norm": 0.046910062432289124,
+ "learning_rate": 0.0006,
+ "loss": 6.550876617431641,
+ "step": 3227
+ },
+ {
+ "epoch": 44.838427947598255,
+ "grad_norm": 0.03494197130203247,
+ "learning_rate": 0.0006,
+ "loss": 6.540179252624512,
+ "step": 3228
+ },
+ {
+ "epoch": 44.852401746724894,
+ "grad_norm": 0.04716387391090393,
+ "learning_rate": 0.0006,
+ "loss": 6.490459442138672,
+ "step": 3229
+ },
+ {
+ "epoch": 44.866375545851525,
+ "grad_norm": 0.04190506786108017,
+ "learning_rate": 0.0006,
+ "loss": 6.509925365447998,
+ "step": 3230
+ },
+ {
+ "epoch": 44.880349344978164,
+ "grad_norm": 0.04749064892530441,
+ "learning_rate": 0.0006,
+ "loss": 6.478779315948486,
+ "step": 3231
+ },
+ {
+ "epoch": 44.8943231441048,
+ "grad_norm": 0.028002185747027397,
+ "learning_rate": 0.0006,
+ "loss": 6.528904914855957,
+ "step": 3232
+ },
+ {
+ "epoch": 44.90829694323144,
+ "grad_norm": 0.036473166197538376,
+ "learning_rate": 0.0006,
+ "loss": 6.549042701721191,
+ "step": 3233
+ },
+ {
+ "epoch": 44.92227074235808,
+ "grad_norm": 0.02543315477669239,
+ "learning_rate": 0.0006,
+ "loss": 6.498875141143799,
+ "step": 3234
+ },
+ {
+ "epoch": 44.93624454148472,
+ "grad_norm": 0.029906807467341423,
+ "learning_rate": 0.0006,
+ "loss": 6.453019142150879,
+ "step": 3235
+ },
+ {
+ "epoch": 44.95021834061135,
+ "grad_norm": 0.027231857180595398,
+ "learning_rate": 0.0006,
+ "loss": 6.428829193115234,
+ "step": 3236
+ },
+ {
+ "epoch": 44.96419213973799,
+ "grad_norm": 0.02506154775619507,
+ "learning_rate": 0.0006,
+ "loss": 6.420858383178711,
+ "step": 3237
+ },
+ {
+ "epoch": 44.97816593886463,
+ "grad_norm": 0.0269472673535347,
+ "learning_rate": 0.0006,
+ "loss": 6.484989643096924,
+ "step": 3238
+ },
+ {
+ "epoch": 44.992139737991266,
+ "grad_norm": 0.023410487920045853,
+ "learning_rate": 0.0006,
+ "loss": 6.454830169677734,
+ "step": 3239
+ },
+ {
+ "epoch": 45.0,
+ "grad_norm": 0.034175023436546326,
+ "learning_rate": 0.0006,
+ "loss": 6.430975437164307,
+ "step": 3240
+ },
+ {
+ "epoch": 45.0,
+ "eval_loss": 6.458896160125732,
+ "eval_runtime": 56.6934,
+ "eval_samples_per_second": 43.074,
+ "eval_steps_per_second": 1.358,
+ "step": 3240
+ },
+ {
+ "epoch": 45.01397379912664,
+ "grad_norm": 0.045031629502773285,
+ "learning_rate": 0.0006,
+ "loss": 6.4243011474609375,
+ "step": 3241
+ },
+ {
+ "epoch": 45.02794759825328,
+ "grad_norm": 0.04968734085559845,
+ "learning_rate": 0.0006,
+ "loss": 6.380091667175293,
+ "step": 3242
+ },
+ {
+ "epoch": 45.041921397379916,
+ "grad_norm": 0.047145187854766846,
+ "learning_rate": 0.0006,
+ "loss": 6.477178573608398,
+ "step": 3243
+ },
+ {
+ "epoch": 45.05589519650655,
+ "grad_norm": 0.04304948449134827,
+ "learning_rate": 0.0006,
+ "loss": 6.399726867675781,
+ "step": 3244
+ },
+ {
+ "epoch": 45.069868995633186,
+ "grad_norm": 0.04466010630130768,
+ "learning_rate": 0.0006,
+ "loss": 6.3415679931640625,
+ "step": 3245
+ },
+ {
+ "epoch": 45.083842794759825,
+ "grad_norm": 0.04305306077003479,
+ "learning_rate": 0.0006,
+ "loss": 6.3413310050964355,
+ "step": 3246
+ },
+ {
+ "epoch": 45.09781659388646,
+ "grad_norm": 0.043064575642347336,
+ "learning_rate": 0.0006,
+ "loss": 6.4838714599609375,
+ "step": 3247
+ },
+ {
+ "epoch": 45.1117903930131,
+ "grad_norm": 0.049503277987241745,
+ "learning_rate": 0.0006,
+ "loss": 6.38473653793335,
+ "step": 3248
+ },
+ {
+ "epoch": 45.12576419213974,
+ "grad_norm": 0.04613238573074341,
+ "learning_rate": 0.0006,
+ "loss": 6.370769500732422,
+ "step": 3249
+ },
+ {
+ "epoch": 45.13973799126637,
+ "grad_norm": 0.052119478583335876,
+ "learning_rate": 0.0006,
+ "loss": 6.362321853637695,
+ "step": 3250
+ },
+ {
+ "epoch": 45.15371179039301,
+ "grad_norm": 0.045110076665878296,
+ "learning_rate": 0.0006,
+ "loss": 6.369844913482666,
+ "step": 3251
+ },
+ {
+ "epoch": 45.16768558951965,
+ "grad_norm": 0.03798583149909973,
+ "learning_rate": 0.0006,
+ "loss": 6.324633598327637,
+ "step": 3252
+ },
+ {
+ "epoch": 45.18165938864629,
+ "grad_norm": 0.03744659945368767,
+ "learning_rate": 0.0006,
+ "loss": 6.349421501159668,
+ "step": 3253
+ },
+ {
+ "epoch": 45.19563318777293,
+ "grad_norm": 0.03767292574048042,
+ "learning_rate": 0.0006,
+ "loss": 6.334238052368164,
+ "step": 3254
+ },
+ {
+ "epoch": 45.209606986899566,
+ "grad_norm": 0.03577509894967079,
+ "learning_rate": 0.0006,
+ "loss": 6.3453826904296875,
+ "step": 3255
+ },
+ {
+ "epoch": 45.223580786026204,
+ "grad_norm": 0.03403240442276001,
+ "learning_rate": 0.0006,
+ "loss": 6.2836174964904785,
+ "step": 3256
+ },
+ {
+ "epoch": 45.237554585152836,
+ "grad_norm": 0.03517331928014755,
+ "learning_rate": 0.0006,
+ "loss": 6.281718730926514,
+ "step": 3257
+ },
+ {
+ "epoch": 45.251528384279474,
+ "grad_norm": 0.04409139230847359,
+ "learning_rate": 0.0006,
+ "loss": 6.320815086364746,
+ "step": 3258
+ },
+ {
+ "epoch": 45.26550218340611,
+ "grad_norm": 0.054750364273786545,
+ "learning_rate": 0.0006,
+ "loss": 6.343618392944336,
+ "step": 3259
+ },
+ {
+ "epoch": 45.27947598253275,
+ "grad_norm": 0.07612067461013794,
+ "learning_rate": 0.0006,
+ "loss": 6.310614585876465,
+ "step": 3260
+ },
+ {
+ "epoch": 45.29344978165939,
+ "grad_norm": 0.08438985049724579,
+ "learning_rate": 0.0006,
+ "loss": 6.2369184494018555,
+ "step": 3261
+ },
+ {
+ "epoch": 45.30742358078603,
+ "grad_norm": 0.0782201811671257,
+ "learning_rate": 0.0006,
+ "loss": 6.254452705383301,
+ "step": 3262
+ },
+ {
+ "epoch": 45.32139737991266,
+ "grad_norm": 0.09928355365991592,
+ "learning_rate": 0.0006,
+ "loss": 6.3112077713012695,
+ "step": 3263
+ },
+ {
+ "epoch": 45.3353711790393,
+ "grad_norm": 0.07233133912086487,
+ "learning_rate": 0.0006,
+ "loss": 6.408087253570557,
+ "step": 3264
+ },
+ {
+ "epoch": 45.34934497816594,
+ "grad_norm": 0.07478684186935425,
+ "learning_rate": 0.0006,
+ "loss": 6.3770036697387695,
+ "step": 3265
+ },
+ {
+ "epoch": 45.36331877729258,
+ "grad_norm": 0.06036846712231636,
+ "learning_rate": 0.0006,
+ "loss": 6.33865213394165,
+ "step": 3266
+ },
+ {
+ "epoch": 45.377292576419215,
+ "grad_norm": 0.05783458054065704,
+ "learning_rate": 0.0006,
+ "loss": 6.363203048706055,
+ "step": 3267
+ },
+ {
+ "epoch": 45.391266375545854,
+ "grad_norm": 0.05443909019231796,
+ "learning_rate": 0.0006,
+ "loss": 6.287069797515869,
+ "step": 3268
+ },
+ {
+ "epoch": 45.40524017467249,
+ "grad_norm": 0.0557909794151783,
+ "learning_rate": 0.0006,
+ "loss": 6.297784805297852,
+ "step": 3269
+ },
+ {
+ "epoch": 45.419213973799124,
+ "grad_norm": 0.03483372926712036,
+ "learning_rate": 0.0006,
+ "loss": 6.280734062194824,
+ "step": 3270
+ },
+ {
+ "epoch": 45.43318777292576,
+ "grad_norm": 0.0425543375313282,
+ "learning_rate": 0.0006,
+ "loss": 6.33565616607666,
+ "step": 3271
+ },
+ {
+ "epoch": 45.4471615720524,
+ "grad_norm": 0.03230159357190132,
+ "learning_rate": 0.0006,
+ "loss": 6.197698593139648,
+ "step": 3272
+ },
+ {
+ "epoch": 45.46113537117904,
+ "grad_norm": 0.036897290498018265,
+ "learning_rate": 0.0006,
+ "loss": 6.241239547729492,
+ "step": 3273
+ },
+ {
+ "epoch": 45.47510917030568,
+ "grad_norm": 0.031103838235139847,
+ "learning_rate": 0.0006,
+ "loss": 6.159435749053955,
+ "step": 3274
+ },
+ {
+ "epoch": 45.48908296943232,
+ "grad_norm": 0.027116449549794197,
+ "learning_rate": 0.0006,
+ "loss": 6.212225914001465,
+ "step": 3275
+ },
+ {
+ "epoch": 45.50305676855895,
+ "grad_norm": 0.029923981055617332,
+ "learning_rate": 0.0006,
+ "loss": 6.157269477844238,
+ "step": 3276
+ },
+ {
+ "epoch": 45.51703056768559,
+ "grad_norm": 0.02521538734436035,
+ "learning_rate": 0.0006,
+ "loss": 6.146918773651123,
+ "step": 3277
+ },
+ {
+ "epoch": 45.531004366812226,
+ "grad_norm": 0.02758920192718506,
+ "learning_rate": 0.0006,
+ "loss": 6.0446977615356445,
+ "step": 3278
+ },
+ {
+ "epoch": 45.544978165938865,
+ "grad_norm": 0.024328220635652542,
+ "learning_rate": 0.0006,
+ "loss": 6.099508285522461,
+ "step": 3279
+ },
+ {
+ "epoch": 45.5589519650655,
+ "grad_norm": 0.027425771579146385,
+ "learning_rate": 0.0006,
+ "loss": 6.084299087524414,
+ "step": 3280
+ },
+ {
+ "epoch": 45.57292576419214,
+ "grad_norm": 0.026927651837468147,
+ "learning_rate": 0.0006,
+ "loss": 6.146334171295166,
+ "step": 3281
+ },
+ {
+ "epoch": 45.58689956331878,
+ "grad_norm": 0.04054379090666771,
+ "learning_rate": 0.0006,
+ "loss": 6.1301493644714355,
+ "step": 3282
+ },
+ {
+ "epoch": 45.60087336244541,
+ "grad_norm": 0.04456290975213051,
+ "learning_rate": 0.0006,
+ "loss": 6.152511119842529,
+ "step": 3283
+ },
+ {
+ "epoch": 45.61484716157205,
+ "grad_norm": 0.052821505814790726,
+ "learning_rate": 0.0006,
+ "loss": 6.113544940948486,
+ "step": 3284
+ },
+ {
+ "epoch": 45.62882096069869,
+ "grad_norm": 0.07251739501953125,
+ "learning_rate": 0.0006,
+ "loss": 6.118298530578613,
+ "step": 3285
+ },
+ {
+ "epoch": 45.64279475982533,
+ "grad_norm": 0.09433556348085403,
+ "learning_rate": 0.0006,
+ "loss": 6.170897483825684,
+ "step": 3286
+ },
+ {
+ "epoch": 45.65676855895197,
+ "grad_norm": 0.057683493942022324,
+ "learning_rate": 0.0006,
+ "loss": 6.1747822761535645,
+ "step": 3287
+ },
+ {
+ "epoch": 45.670742358078606,
+ "grad_norm": 0.06363270431756973,
+ "learning_rate": 0.0006,
+ "loss": 6.1617889404296875,
+ "step": 3288
+ },
+ {
+ "epoch": 45.68471615720524,
+ "grad_norm": 0.07528169453144073,
+ "learning_rate": 0.0006,
+ "loss": 6.230060577392578,
+ "step": 3289
+ },
+ {
+ "epoch": 45.698689956331876,
+ "grad_norm": 0.07816170156002045,
+ "learning_rate": 0.0006,
+ "loss": 6.171257019042969,
+ "step": 3290
+ },
+ {
+ "epoch": 45.712663755458514,
+ "grad_norm": 0.05860164389014244,
+ "learning_rate": 0.0006,
+ "loss": 6.095484256744385,
+ "step": 3291
+ },
+ {
+ "epoch": 45.72663755458515,
+ "grad_norm": 0.04747429117560387,
+ "learning_rate": 0.0006,
+ "loss": 6.060067176818848,
+ "step": 3292
+ },
+ {
+ "epoch": 45.74061135371179,
+ "grad_norm": 0.04514474794268608,
+ "learning_rate": 0.0006,
+ "loss": 6.100218772888184,
+ "step": 3293
+ },
+ {
+ "epoch": 45.75458515283843,
+ "grad_norm": 0.05168681964278221,
+ "learning_rate": 0.0006,
+ "loss": 6.0621867179870605,
+ "step": 3294
+ },
+ {
+ "epoch": 45.76855895196506,
+ "grad_norm": 0.04665876924991608,
+ "learning_rate": 0.0006,
+ "loss": 6.118292808532715,
+ "step": 3295
+ },
+ {
+ "epoch": 45.7825327510917,
+ "grad_norm": 0.049374762922525406,
+ "learning_rate": 0.0006,
+ "loss": 6.098394393920898,
+ "step": 3296
+ },
+ {
+ "epoch": 45.79650655021834,
+ "grad_norm": 0.03793549910187721,
+ "learning_rate": 0.0006,
+ "loss": 6.069541931152344,
+ "step": 3297
+ },
+ {
+ "epoch": 45.81048034934498,
+ "grad_norm": 0.036362383514642715,
+ "learning_rate": 0.0006,
+ "loss": 6.1188435554504395,
+ "step": 3298
+ },
+ {
+ "epoch": 45.82445414847162,
+ "grad_norm": 0.03397097811102867,
+ "learning_rate": 0.0006,
+ "loss": 6.117589950561523,
+ "step": 3299
+ },
+ {
+ "epoch": 45.838427947598255,
+ "grad_norm": 0.030977753922343254,
+ "learning_rate": 0.0006,
+ "loss": 6.0236711502075195,
+ "step": 3300
+ },
+ {
+ "epoch": 45.852401746724894,
+ "grad_norm": 0.02800217643380165,
+ "learning_rate": 0.0006,
+ "loss": 6.034825325012207,
+ "step": 3301
+ },
+ {
+ "epoch": 45.866375545851525,
+ "grad_norm": 0.03129778802394867,
+ "learning_rate": 0.0006,
+ "loss": 5.980271339416504,
+ "step": 3302
+ },
+ {
+ "epoch": 45.880349344978164,
+ "grad_norm": 0.025886094197630882,
+ "learning_rate": 0.0006,
+ "loss": 5.966419219970703,
+ "step": 3303
+ },
+ {
+ "epoch": 45.8943231441048,
+ "grad_norm": 0.025712328031659126,
+ "learning_rate": 0.0006,
+ "loss": 5.938652515411377,
+ "step": 3304
+ },
+ {
+ "epoch": 45.90829694323144,
+ "grad_norm": 0.022508298978209496,
+ "learning_rate": 0.0006,
+ "loss": 5.982312202453613,
+ "step": 3305
+ },
+ {
+ "epoch": 45.92227074235808,
+ "grad_norm": 0.024880951270461082,
+ "learning_rate": 0.0006,
+ "loss": 5.931150436401367,
+ "step": 3306
+ },
+ {
+ "epoch": 45.93624454148472,
+ "grad_norm": 0.030007576569914818,
+ "learning_rate": 0.0006,
+ "loss": 5.936639308929443,
+ "step": 3307
+ },
+ {
+ "epoch": 45.95021834061135,
+ "grad_norm": 0.033314041793346405,
+ "learning_rate": 0.0006,
+ "loss": 5.944248676300049,
+ "step": 3308
+ },
+ {
+ "epoch": 45.96419213973799,
+ "grad_norm": 0.043078381568193436,
+ "learning_rate": 0.0006,
+ "loss": 5.971286773681641,
+ "step": 3309
+ },
+ {
+ "epoch": 45.97816593886463,
+ "grad_norm": 0.059296440333127975,
+ "learning_rate": 0.0006,
+ "loss": 5.910497665405273,
+ "step": 3310
+ },
+ {
+ "epoch": 45.992139737991266,
+ "grad_norm": 0.06232666224241257,
+ "learning_rate": 0.0006,
+ "loss": 5.929817199707031,
+ "step": 3311
+ },
+ {
+ "epoch": 46.0,
+ "grad_norm": 0.05303940549492836,
+ "learning_rate": 0.0006,
+ "loss": 5.910822868347168,
+ "step": 3312
+ },
+ {
+ "epoch": 46.0,
+ "eval_loss": 5.977851390838623,
+ "eval_runtime": 56.2435,
+ "eval_samples_per_second": 43.418,
+ "eval_steps_per_second": 1.369,
+ "step": 3312
+ },
+ {
+ "epoch": 46.01397379912664,
+ "grad_norm": 0.06943798810243607,
+ "learning_rate": 0.0006,
+ "loss": 5.854922294616699,
+ "step": 3313
+ },
+ {
+ "epoch": 46.02794759825328,
+ "grad_norm": 0.05236278846859932,
+ "learning_rate": 0.0006,
+ "loss": 5.876611232757568,
+ "step": 3314
+ },
+ {
+ "epoch": 46.041921397379916,
+ "grad_norm": 0.06240001320838928,
+ "learning_rate": 0.0006,
+ "loss": 5.878912448883057,
+ "step": 3315
+ },
+ {
+ "epoch": 46.05589519650655,
+ "grad_norm": 0.07401192933320999,
+ "learning_rate": 0.0006,
+ "loss": 5.985136032104492,
+ "step": 3316
+ },
+ {
+ "epoch": 46.069868995633186,
+ "grad_norm": 0.0826529711484909,
+ "learning_rate": 0.0006,
+ "loss": 5.939439296722412,
+ "step": 3317
+ },
+ {
+ "epoch": 46.083842794759825,
+ "grad_norm": 0.07999005168676376,
+ "learning_rate": 0.0006,
+ "loss": 5.982336044311523,
+ "step": 3318
+ },
+ {
+ "epoch": 46.09781659388646,
+ "grad_norm": 0.07014869153499603,
+ "learning_rate": 0.0006,
+ "loss": 5.896534442901611,
+ "step": 3319
+ },
+ {
+ "epoch": 46.1117903930131,
+ "grad_norm": 0.05795128643512726,
+ "learning_rate": 0.0006,
+ "loss": 5.843111991882324,
+ "step": 3320
+ },
+ {
+ "epoch": 46.12576419213974,
+ "grad_norm": 0.06374258548021317,
+ "learning_rate": 0.0006,
+ "loss": 5.78752326965332,
+ "step": 3321
+ },
+ {
+ "epoch": 46.13973799126637,
+ "grad_norm": 0.045524805784225464,
+ "learning_rate": 0.0006,
+ "loss": 5.86755895614624,
+ "step": 3322
+ },
+ {
+ "epoch": 46.15371179039301,
+ "grad_norm": 0.03773504123091698,
+ "learning_rate": 0.0006,
+ "loss": 5.83627462387085,
+ "step": 3323
+ },
+ {
+ "epoch": 46.16768558951965,
+ "grad_norm": 0.035815201699733734,
+ "learning_rate": 0.0006,
+ "loss": 5.798238277435303,
+ "step": 3324
+ },
+ {
+ "epoch": 46.18165938864629,
+ "grad_norm": 0.03756783902645111,
+ "learning_rate": 0.0006,
+ "loss": 5.845197677612305,
+ "step": 3325
+ },
+ {
+ "epoch": 46.19563318777293,
+ "grad_norm": 0.03569113463163376,
+ "learning_rate": 0.0006,
+ "loss": 5.888932704925537,
+ "step": 3326
+ },
+ {
+ "epoch": 46.209606986899566,
+ "grad_norm": 0.029467813670635223,
+ "learning_rate": 0.0006,
+ "loss": 5.792047500610352,
+ "step": 3327
+ },
+ {
+ "epoch": 46.223580786026204,
+ "grad_norm": 0.025698406621813774,
+ "learning_rate": 0.0006,
+ "loss": 5.783734321594238,
+ "step": 3328
+ },
+ {
+ "epoch": 46.237554585152836,
+ "grad_norm": 0.028426671400666237,
+ "learning_rate": 0.0006,
+ "loss": 5.762084007263184,
+ "step": 3329
+ },
+ {
+ "epoch": 46.251528384279474,
+ "grad_norm": 0.03099622018635273,
+ "learning_rate": 0.0006,
+ "loss": 5.719109058380127,
+ "step": 3330
+ },
+ {
+ "epoch": 46.26550218340611,
+ "grad_norm": 0.0310219656676054,
+ "learning_rate": 0.0006,
+ "loss": 5.737057209014893,
+ "step": 3331
+ },
+ {
+ "epoch": 46.27947598253275,
+ "grad_norm": 0.02713446132838726,
+ "learning_rate": 0.0006,
+ "loss": 5.758657455444336,
+ "step": 3332
+ },
+ {
+ "epoch": 46.29344978165939,
+ "grad_norm": 0.03099844604730606,
+ "learning_rate": 0.0006,
+ "loss": 5.742086410522461,
+ "step": 3333
+ },
+ {
+ "epoch": 46.30742358078603,
+ "grad_norm": 0.030258730053901672,
+ "learning_rate": 0.0006,
+ "loss": 5.767426490783691,
+ "step": 3334
+ },
+ {
+ "epoch": 46.32139737991266,
+ "grad_norm": 0.029034852981567383,
+ "learning_rate": 0.0006,
+ "loss": 5.651792049407959,
+ "step": 3335
+ },
+ {
+ "epoch": 46.3353711790393,
+ "grad_norm": 0.03138665854930878,
+ "learning_rate": 0.0006,
+ "loss": 5.664980888366699,
+ "step": 3336
+ },
+ {
+ "epoch": 46.34934497816594,
+ "grad_norm": 0.040546901524066925,
+ "learning_rate": 0.0006,
+ "loss": 5.665165901184082,
+ "step": 3337
+ },
+ {
+ "epoch": 46.36331877729258,
+ "grad_norm": 0.06395485252141953,
+ "learning_rate": 0.0006,
+ "loss": 5.7776713371276855,
+ "step": 3338
+ },
+ {
+ "epoch": 46.377292576419215,
+ "grad_norm": 0.10088056325912476,
+ "learning_rate": 0.0006,
+ "loss": 5.755645751953125,
+ "step": 3339
+ },
+ {
+ "epoch": 46.391266375545854,
+ "grad_norm": 0.09527938067913055,
+ "learning_rate": 0.0006,
+ "loss": 5.838809490203857,
+ "step": 3340
+ },
+ {
+ "epoch": 46.40524017467249,
+ "grad_norm": 0.0849648267030716,
+ "learning_rate": 0.0006,
+ "loss": 5.781932830810547,
+ "step": 3341
+ },
+ {
+ "epoch": 46.419213973799124,
+ "grad_norm": 0.10142666101455688,
+ "learning_rate": 0.0006,
+ "loss": 5.838970184326172,
+ "step": 3342
+ },
+ {
+ "epoch": 46.43318777292576,
+ "grad_norm": 0.42731454968452454,
+ "learning_rate": 0.0006,
+ "loss": 6.071902275085449,
+ "step": 3343
+ },
+ {
+ "epoch": 46.4471615720524,
+ "grad_norm": 0.13102488219738007,
+ "learning_rate": 0.0006,
+ "loss": 6.8635993003845215,
+ "step": 3344
+ },
+ {
+ "epoch": 46.46113537117904,
+ "grad_norm": 0.23691938817501068,
+ "learning_rate": 0.0006,
+ "loss": 7.313158988952637,
+ "step": 3345
+ },
+ {
+ "epoch": 46.47510917030568,
+ "grad_norm": 0.11670669168233871,
+ "learning_rate": 0.0006,
+ "loss": 7.240363121032715,
+ "step": 3346
+ },
+ {
+ "epoch": 46.48908296943232,
+ "grad_norm": 0.09985820949077606,
+ "learning_rate": 0.0006,
+ "loss": 7.113332271575928,
+ "step": 3347
+ },
+ {
+ "epoch": 46.50305676855895,
+ "grad_norm": 0.11406408250331879,
+ "learning_rate": 0.0006,
+ "loss": 7.051285743713379,
+ "step": 3348
+ },
+ {
+ "epoch": 46.51703056768559,
+ "grad_norm": 0.07900336384773254,
+ "learning_rate": 0.0006,
+ "loss": 7.014365196228027,
+ "step": 3349
+ },
+ {
+ "epoch": 46.531004366812226,
+ "grad_norm": 0.05380333960056305,
+ "learning_rate": 0.0006,
+ "loss": 6.8868818283081055,
+ "step": 3350
+ },
+ {
+ "epoch": 46.544978165938865,
+ "grad_norm": 0.05868449807167053,
+ "learning_rate": 0.0006,
+ "loss": 6.783599853515625,
+ "step": 3351
+ },
+ {
+ "epoch": 46.5589519650655,
+ "grad_norm": 0.04428785294294357,
+ "learning_rate": 0.0006,
+ "loss": 6.798511505126953,
+ "step": 3352
+ },
+ {
+ "epoch": 46.57292576419214,
+ "grad_norm": 0.055527277290821075,
+ "learning_rate": 0.0006,
+ "loss": 6.694507598876953,
+ "step": 3353
+ },
+ {
+ "epoch": 46.58689956331878,
+ "grad_norm": 0.03872432932257652,
+ "learning_rate": 0.0006,
+ "loss": 6.736631393432617,
+ "step": 3354
+ },
+ {
+ "epoch": 46.60087336244541,
+ "grad_norm": 0.04330779239535332,
+ "learning_rate": 0.0006,
+ "loss": 6.576806545257568,
+ "step": 3355
+ },
+ {
+ "epoch": 46.61484716157205,
+ "grad_norm": 0.036767736077308655,
+ "learning_rate": 0.0006,
+ "loss": 6.559871196746826,
+ "step": 3356
+ },
+ {
+ "epoch": 46.62882096069869,
+ "grad_norm": 0.039857108145952225,
+ "learning_rate": 0.0006,
+ "loss": 6.5749831199646,
+ "step": 3357
+ },
+ {
+ "epoch": 46.64279475982533,
+ "grad_norm": 0.048719972372055054,
+ "learning_rate": 0.0006,
+ "loss": 6.4915618896484375,
+ "step": 3358
+ },
+ {
+ "epoch": 46.65676855895197,
+ "grad_norm": 0.042849015444517136,
+ "learning_rate": 0.0006,
+ "loss": 6.413516044616699,
+ "step": 3359
+ },
+ {
+ "epoch": 46.670742358078606,
+ "grad_norm": 0.02871009334921837,
+ "learning_rate": 0.0006,
+ "loss": 6.320952415466309,
+ "step": 3360
+ },
+ {
+ "epoch": 46.68471615720524,
+ "grad_norm": 0.03193620592355728,
+ "learning_rate": 0.0006,
+ "loss": 6.4176435470581055,
+ "step": 3361
+ },
+ {
+ "epoch": 46.698689956331876,
+ "grad_norm": 0.027417024597525597,
+ "learning_rate": 0.0006,
+ "loss": 6.342319488525391,
+ "step": 3362
+ },
+ {
+ "epoch": 46.712663755458514,
+ "grad_norm": 0.03305937349796295,
+ "learning_rate": 0.0006,
+ "loss": 6.246112823486328,
+ "step": 3363
+ },
+ {
+ "epoch": 46.72663755458515,
+ "grad_norm": 0.031403183937072754,
+ "learning_rate": 0.0006,
+ "loss": 6.149503231048584,
+ "step": 3364
+ },
+ {
+ "epoch": 46.74061135371179,
+ "grad_norm": 0.031563788652420044,
+ "learning_rate": 0.0006,
+ "loss": 6.278808116912842,
+ "step": 3365
+ },
+ {
+ "epoch": 46.75458515283843,
+ "grad_norm": 0.028366589918732643,
+ "learning_rate": 0.0006,
+ "loss": 6.057036876678467,
+ "step": 3366
+ },
+ {
+ "epoch": 46.76855895196506,
+ "grad_norm": 0.023715803399682045,
+ "learning_rate": 0.0006,
+ "loss": 6.111324310302734,
+ "step": 3367
+ },
+ {
+ "epoch": 46.7825327510917,
+ "grad_norm": 0.026546811684966087,
+ "learning_rate": 0.0006,
+ "loss": 6.105669975280762,
+ "step": 3368
+ },
+ {
+ "epoch": 46.79650655021834,
+ "grad_norm": 0.023531019687652588,
+ "learning_rate": 0.0006,
+ "loss": 6.068072319030762,
+ "step": 3369
+ },
+ {
+ "epoch": 46.81048034934498,
+ "grad_norm": 0.024480128660798073,
+ "learning_rate": 0.0006,
+ "loss": 5.947896957397461,
+ "step": 3370
+ },
+ {
+ "epoch": 46.82445414847162,
+ "grad_norm": 0.02517310529947281,
+ "learning_rate": 0.0006,
+ "loss": 6.030037879943848,
+ "step": 3371
+ },
+ {
+ "epoch": 46.838427947598255,
+ "grad_norm": 0.026723403483629227,
+ "learning_rate": 0.0006,
+ "loss": 5.963776111602783,
+ "step": 3372
+ },
+ {
+ "epoch": 46.852401746724894,
+ "grad_norm": 0.03163504600524902,
+ "learning_rate": 0.0006,
+ "loss": 5.893466949462891,
+ "step": 3373
+ },
+ {
+ "epoch": 46.866375545851525,
+ "grad_norm": 0.04058683291077614,
+ "learning_rate": 0.0006,
+ "loss": 5.9202799797058105,
+ "step": 3374
+ },
+ {
+ "epoch": 46.880349344978164,
+ "grad_norm": 0.04964117333292961,
+ "learning_rate": 0.0006,
+ "loss": 5.9366559982299805,
+ "step": 3375
+ },
+ {
+ "epoch": 46.8943231441048,
+ "grad_norm": 0.055406033992767334,
+ "learning_rate": 0.0006,
+ "loss": 5.912230968475342,
+ "step": 3376
+ },
+ {
+ "epoch": 46.90829694323144,
+ "grad_norm": 0.04643625393509865,
+ "learning_rate": 0.0006,
+ "loss": 5.87260103225708,
+ "step": 3377
+ },
+ {
+ "epoch": 46.92227074235808,
+ "grad_norm": 0.029325876384973526,
+ "learning_rate": 0.0006,
+ "loss": 5.854763984680176,
+ "step": 3378
+ },
+ {
+ "epoch": 46.93624454148472,
+ "grad_norm": 0.03159382566809654,
+ "learning_rate": 0.0006,
+ "loss": 5.880882263183594,
+ "step": 3379
+ },
+ {
+ "epoch": 46.95021834061135,
+ "grad_norm": 0.030224012210965157,
+ "learning_rate": 0.0006,
+ "loss": 5.787360191345215,
+ "step": 3380
+ },
+ {
+ "epoch": 46.96419213973799,
+ "grad_norm": 0.028368130326271057,
+ "learning_rate": 0.0006,
+ "loss": 5.817195892333984,
+ "step": 3381
+ },
+ {
+ "epoch": 46.97816593886463,
+ "grad_norm": 0.029411399737000465,
+ "learning_rate": 0.0006,
+ "loss": 5.783403396606445,
+ "step": 3382
+ },
+ {
+ "epoch": 46.992139737991266,
+ "grad_norm": 0.029927462339401245,
+ "learning_rate": 0.0006,
+ "loss": 5.662444591522217,
+ "step": 3383
+ },
+ {
+ "epoch": 47.0,
+ "grad_norm": 0.02872961387038231,
+ "learning_rate": 0.0006,
+ "loss": 5.742781639099121,
+ "step": 3384
+ },
+ {
+ "epoch": 47.0,
+ "eval_loss": 5.870260238647461,
+ "eval_runtime": 57.3734,
+ "eval_samples_per_second": 42.563,
+ "eval_steps_per_second": 1.342,
+ "step": 3384
+ },
+ {
+ "epoch": 47.01397379912664,
+ "grad_norm": 0.028542131185531616,
+ "learning_rate": 0.0006,
+ "loss": 5.753256797790527,
+ "step": 3385
+ },
+ {
+ "epoch": 47.02794759825328,
+ "grad_norm": 0.03227119892835617,
+ "learning_rate": 0.0006,
+ "loss": 5.722246170043945,
+ "step": 3386
+ },
+ {
+ "epoch": 47.041921397379916,
+ "grad_norm": 0.031461723148822784,
+ "learning_rate": 0.0006,
+ "loss": 5.740520000457764,
+ "step": 3387
+ },
+ {
+ "epoch": 47.05589519650655,
+ "grad_norm": 0.03184220567345619,
+ "learning_rate": 0.0006,
+ "loss": 5.720890045166016,
+ "step": 3388
+ },
+ {
+ "epoch": 47.069868995633186,
+ "grad_norm": 0.030015580356121063,
+ "learning_rate": 0.0006,
+ "loss": 5.597814559936523,
+ "step": 3389
+ },
+ {
+ "epoch": 47.083842794759825,
+ "grad_norm": 0.029931779950857162,
+ "learning_rate": 0.0006,
+ "loss": 5.5412468910217285,
+ "step": 3390
+ },
+ {
+ "epoch": 47.09781659388646,
+ "grad_norm": 0.03206505626440048,
+ "learning_rate": 0.0006,
+ "loss": 5.6967597007751465,
+ "step": 3391
+ },
+ {
+ "epoch": 47.1117903930131,
+ "grad_norm": 0.03978161886334419,
+ "learning_rate": 0.0006,
+ "loss": 5.59661340713501,
+ "step": 3392
+ },
+ {
+ "epoch": 47.12576419213974,
+ "grad_norm": 0.04619080200791359,
+ "learning_rate": 0.0006,
+ "loss": 5.795788764953613,
+ "step": 3393
+ },
+ {
+ "epoch": 47.13973799126637,
+ "grad_norm": 0.049085382372140884,
+ "learning_rate": 0.0006,
+ "loss": 5.667335510253906,
+ "step": 3394
+ },
+ {
+ "epoch": 47.15371179039301,
+ "grad_norm": 0.04830858111381531,
+ "learning_rate": 0.0006,
+ "loss": 5.668806552886963,
+ "step": 3395
+ },
+ {
+ "epoch": 47.16768558951965,
+ "grad_norm": 0.0583178736269474,
+ "learning_rate": 0.0006,
+ "loss": 5.692285537719727,
+ "step": 3396
+ },
+ {
+ "epoch": 47.18165938864629,
+ "grad_norm": 0.05182253196835518,
+ "learning_rate": 0.0006,
+ "loss": 5.678645133972168,
+ "step": 3397
+ },
+ {
+ "epoch": 47.19563318777293,
+ "grad_norm": 0.04011905938386917,
+ "learning_rate": 0.0006,
+ "loss": 5.661550998687744,
+ "step": 3398
+ },
+ {
+ "epoch": 47.209606986899566,
+ "grad_norm": 0.03888486698269844,
+ "learning_rate": 0.0006,
+ "loss": 5.643001556396484,
+ "step": 3399
+ },
+ {
+ "epoch": 47.223580786026204,
+ "grad_norm": 0.030658265575766563,
+ "learning_rate": 0.0006,
+ "loss": 5.659090995788574,
+ "step": 3400
+ },
+ {
+ "epoch": 47.237554585152836,
+ "grad_norm": 0.03609256446361542,
+ "learning_rate": 0.0006,
+ "loss": 5.544501781463623,
+ "step": 3401
+ },
+ {
+ "epoch": 47.251528384279474,
+ "grad_norm": 0.033589769154787064,
+ "learning_rate": 0.0006,
+ "loss": 5.463878631591797,
+ "step": 3402
+ },
+ {
+ "epoch": 47.26550218340611,
+ "grad_norm": 0.035128287971019745,
+ "learning_rate": 0.0006,
+ "loss": 5.5680060386657715,
+ "step": 3403
+ },
+ {
+ "epoch": 47.27947598253275,
+ "grad_norm": 0.038684550672769547,
+ "learning_rate": 0.0006,
+ "loss": 5.587898254394531,
+ "step": 3404
+ },
+ {
+ "epoch": 47.29344978165939,
+ "grad_norm": 0.043566226959228516,
+ "learning_rate": 0.0006,
+ "loss": 5.580896854400635,
+ "step": 3405
+ },
+ {
+ "epoch": 47.30742358078603,
+ "grad_norm": 0.04428374394774437,
+ "learning_rate": 0.0006,
+ "loss": 5.603460311889648,
+ "step": 3406
+ },
+ {
+ "epoch": 47.32139737991266,
+ "grad_norm": 0.037368934601545334,
+ "learning_rate": 0.0006,
+ "loss": 5.564740180969238,
+ "step": 3407
+ },
+ {
+ "epoch": 47.3353711790393,
+ "grad_norm": 0.04092350974678993,
+ "learning_rate": 0.0006,
+ "loss": 5.584153652191162,
+ "step": 3408
+ },
+ {
+ "epoch": 47.34934497816594,
+ "grad_norm": 0.045031506568193436,
+ "learning_rate": 0.0006,
+ "loss": 5.601186275482178,
+ "step": 3409
+ },
+ {
+ "epoch": 47.36331877729258,
+ "grad_norm": 0.041830409318208694,
+ "learning_rate": 0.0006,
+ "loss": 5.512788772583008,
+ "step": 3410
+ },
+ {
+ "epoch": 47.377292576419215,
+ "grad_norm": 0.037633974105119705,
+ "learning_rate": 0.0006,
+ "loss": 5.503828525543213,
+ "step": 3411
+ },
+ {
+ "epoch": 47.391266375545854,
+ "grad_norm": 0.03473588079214096,
+ "learning_rate": 0.0006,
+ "loss": 5.570557594299316,
+ "step": 3412
+ },
+ {
+ "epoch": 47.40524017467249,
+ "grad_norm": 0.039325181394815445,
+ "learning_rate": 0.0006,
+ "loss": 5.427783966064453,
+ "step": 3413
+ },
+ {
+ "epoch": 47.419213973799124,
+ "grad_norm": 0.04034624621272087,
+ "learning_rate": 0.0006,
+ "loss": 5.572591304779053,
+ "step": 3414
+ },
+ {
+ "epoch": 47.43318777292576,
+ "grad_norm": 0.05245184153318405,
+ "learning_rate": 0.0006,
+ "loss": 5.505208969116211,
+ "step": 3415
+ },
+ {
+ "epoch": 47.4471615720524,
+ "grad_norm": 0.0568675734102726,
+ "learning_rate": 0.0006,
+ "loss": 5.589373588562012,
+ "step": 3416
+ },
+ {
+ "epoch": 47.46113537117904,
+ "grad_norm": 0.048336587846279144,
+ "learning_rate": 0.0006,
+ "loss": 5.522005081176758,
+ "step": 3417
+ },
+ {
+ "epoch": 47.47510917030568,
+ "grad_norm": 0.04194975271821022,
+ "learning_rate": 0.0006,
+ "loss": 5.387576103210449,
+ "step": 3418
+ },
+ {
+ "epoch": 47.48908296943232,
+ "grad_norm": 0.04019433259963989,
+ "learning_rate": 0.0006,
+ "loss": 5.522615432739258,
+ "step": 3419
+ },
+ {
+ "epoch": 47.50305676855895,
+ "grad_norm": 0.05709933117032051,
+ "learning_rate": 0.0006,
+ "loss": 5.420286178588867,
+ "step": 3420
+ },
+ {
+ "epoch": 47.51703056768559,
+ "grad_norm": 0.09130338579416275,
+ "learning_rate": 0.0006,
+ "loss": 5.484622478485107,
+ "step": 3421
+ },
+ {
+ "epoch": 47.531004366812226,
+ "grad_norm": 0.2757541835308075,
+ "learning_rate": 0.0006,
+ "loss": 5.821832656860352,
+ "step": 3422
+ },
+ {
+ "epoch": 47.544978165938865,
+ "grad_norm": 0.28292080760002136,
+ "learning_rate": 0.0006,
+ "loss": 6.579934120178223,
+ "step": 3423
+ },
+ {
+ "epoch": 47.5589519650655,
+ "grad_norm": 0.15293604135513306,
+ "learning_rate": 0.0006,
+ "loss": 6.478213310241699,
+ "step": 3424
+ },
+ {
+ "epoch": 47.57292576419214,
+ "grad_norm": 0.11284393817186356,
+ "learning_rate": 0.0006,
+ "loss": 6.4066386222839355,
+ "step": 3425
+ },
+ {
+ "epoch": 47.58689956331878,
+ "grad_norm": 0.10933073610067368,
+ "learning_rate": 0.0006,
+ "loss": 6.3306169509887695,
+ "step": 3426
+ },
+ {
+ "epoch": 47.60087336244541,
+ "grad_norm": 0.05837641656398773,
+ "learning_rate": 0.0006,
+ "loss": 6.250000953674316,
+ "step": 3427
+ },
+ {
+ "epoch": 47.61484716157205,
+ "grad_norm": 0.05987229570746422,
+ "learning_rate": 0.0006,
+ "loss": 6.152463436126709,
+ "step": 3428
+ },
+ {
+ "epoch": 47.62882096069869,
+ "grad_norm": 0.04064938426017761,
+ "learning_rate": 0.0006,
+ "loss": 6.117897033691406,
+ "step": 3429
+ },
+ {
+ "epoch": 47.64279475982533,
+ "grad_norm": 0.04565057530999184,
+ "learning_rate": 0.0006,
+ "loss": 6.042291641235352,
+ "step": 3430
+ },
+ {
+ "epoch": 47.65676855895197,
+ "grad_norm": 0.03475063294172287,
+ "learning_rate": 0.0006,
+ "loss": 6.0071868896484375,
+ "step": 3431
+ },
+ {
+ "epoch": 47.670742358078606,
+ "grad_norm": 0.03809220343828201,
+ "learning_rate": 0.0006,
+ "loss": 6.0128021240234375,
+ "step": 3432
+ },
+ {
+ "epoch": 47.68471615720524,
+ "grad_norm": 0.03307219594717026,
+ "learning_rate": 0.0006,
+ "loss": 5.894664287567139,
+ "step": 3433
+ },
+ {
+ "epoch": 47.698689956331876,
+ "grad_norm": 0.028824346140027046,
+ "learning_rate": 0.0006,
+ "loss": 5.914865493774414,
+ "step": 3434
+ },
+ {
+ "epoch": 47.712663755458514,
+ "grad_norm": 0.027923690155148506,
+ "learning_rate": 0.0006,
+ "loss": 5.851759910583496,
+ "step": 3435
+ },
+ {
+ "epoch": 47.72663755458515,
+ "grad_norm": 0.025492684915661812,
+ "learning_rate": 0.0006,
+ "loss": 5.749186038970947,
+ "step": 3436
+ },
+ {
+ "epoch": 47.74061135371179,
+ "grad_norm": 0.025105174630880356,
+ "learning_rate": 0.0006,
+ "loss": 5.803566932678223,
+ "step": 3437
+ },
+ {
+ "epoch": 47.75458515283843,
+ "grad_norm": 0.023298880085349083,
+ "learning_rate": 0.0006,
+ "loss": 5.76276159286499,
+ "step": 3438
+ },
+ {
+ "epoch": 47.76855895196506,
+ "grad_norm": 0.02468576282262802,
+ "learning_rate": 0.0006,
+ "loss": 5.833852767944336,
+ "step": 3439
+ },
+ {
+ "epoch": 47.7825327510917,
+ "grad_norm": 0.022730689495801926,
+ "learning_rate": 0.0006,
+ "loss": 5.736724376678467,
+ "step": 3440
+ },
+ {
+ "epoch": 47.79650655021834,
+ "grad_norm": 0.021335354074835777,
+ "learning_rate": 0.0006,
+ "loss": 5.644248008728027,
+ "step": 3441
+ },
+ {
+ "epoch": 47.81048034934498,
+ "grad_norm": 0.02288101427257061,
+ "learning_rate": 0.0006,
+ "loss": 5.7011942863464355,
+ "step": 3442
+ },
+ {
+ "epoch": 47.82445414847162,
+ "grad_norm": 0.01961585320532322,
+ "learning_rate": 0.0006,
+ "loss": 5.63215446472168,
+ "step": 3443
+ },
+ {
+ "epoch": 47.838427947598255,
+ "grad_norm": 0.017659205943346024,
+ "learning_rate": 0.0006,
+ "loss": 5.744845390319824,
+ "step": 3444
+ },
+ {
+ "epoch": 47.852401746724894,
+ "grad_norm": 0.02281784638762474,
+ "learning_rate": 0.0006,
+ "loss": 5.696993350982666,
+ "step": 3445
+ },
+ {
+ "epoch": 47.866375545851525,
+ "grad_norm": 0.016394242644309998,
+ "learning_rate": 0.0006,
+ "loss": 5.659870147705078,
+ "step": 3446
+ },
+ {
+ "epoch": 47.880349344978164,
+ "grad_norm": 0.01967368647456169,
+ "learning_rate": 0.0006,
+ "loss": 5.638072490692139,
+ "step": 3447
+ },
+ {
+ "epoch": 47.8943231441048,
+ "grad_norm": 0.017326267436146736,
+ "learning_rate": 0.0006,
+ "loss": 5.638836860656738,
+ "step": 3448
+ },
+ {
+ "epoch": 47.90829694323144,
+ "grad_norm": 0.017778463661670685,
+ "learning_rate": 0.0006,
+ "loss": 5.634466648101807,
+ "step": 3449
+ },
+ {
+ "epoch": 47.92227074235808,
+ "grad_norm": 0.017269758507609367,
+ "learning_rate": 0.0006,
+ "loss": 5.506277561187744,
+ "step": 3450
+ },
+ {
+ "epoch": 47.93624454148472,
+ "grad_norm": 0.01733289286494255,
+ "learning_rate": 0.0006,
+ "loss": 5.544455528259277,
+ "step": 3451
+ },
+ {
+ "epoch": 47.95021834061135,
+ "grad_norm": 0.017787795513868332,
+ "learning_rate": 0.0006,
+ "loss": 5.498943328857422,
+ "step": 3452
+ },
+ {
+ "epoch": 47.96419213973799,
+ "grad_norm": 0.015770932659506798,
+ "learning_rate": 0.0006,
+ "loss": 5.537227630615234,
+ "step": 3453
+ },
+ {
+ "epoch": 47.97816593886463,
+ "grad_norm": 0.016132809221744537,
+ "learning_rate": 0.0006,
+ "loss": 5.489081382751465,
+ "step": 3454
+ },
+ {
+ "epoch": 47.992139737991266,
+ "grad_norm": 0.01749664731323719,
+ "learning_rate": 0.0006,
+ "loss": 5.454263687133789,
+ "step": 3455
+ },
+ {
+ "epoch": 48.0,
+ "grad_norm": 0.016711296513676643,
+ "learning_rate": 0.0006,
+ "loss": 5.530783653259277,
+ "step": 3456
+ },
+ {
+ "epoch": 48.0,
+ "eval_loss": 5.596925258636475,
+ "eval_runtime": 56.9612,
+ "eval_samples_per_second": 42.871,
+ "eval_steps_per_second": 1.352,
+ "step": 3456
+ },
+ {
+ "epoch": 48.01397379912664,
+ "grad_norm": 0.01867641694843769,
+ "learning_rate": 0.0006,
+ "loss": 5.506853103637695,
+ "step": 3457
+ },
+ {
+ "epoch": 48.02794759825328,
+ "grad_norm": 0.023250792175531387,
+ "learning_rate": 0.0006,
+ "loss": 5.4074177742004395,
+ "step": 3458
+ },
+ {
+ "epoch": 48.041921397379916,
+ "grad_norm": 0.01970936730504036,
+ "learning_rate": 0.0006,
+ "loss": 5.508777618408203,
+ "step": 3459
+ },
+ {
+ "epoch": 48.05589519650655,
+ "grad_norm": 0.01879260502755642,
+ "learning_rate": 0.0006,
+ "loss": 5.412424087524414,
+ "step": 3460
+ },
+ {
+ "epoch": 48.069868995633186,
+ "grad_norm": 0.017420092597603798,
+ "learning_rate": 0.0006,
+ "loss": 5.43361759185791,
+ "step": 3461
+ },
+ {
+ "epoch": 48.083842794759825,
+ "grad_norm": 0.018707862123847008,
+ "learning_rate": 0.0006,
+ "loss": 5.476800918579102,
+ "step": 3462
+ },
+ {
+ "epoch": 48.09781659388646,
+ "grad_norm": 0.023185882717370987,
+ "learning_rate": 0.0006,
+ "loss": 5.465272426605225,
+ "step": 3463
+ },
+ {
+ "epoch": 48.1117903930131,
+ "grad_norm": 0.023570368066430092,
+ "learning_rate": 0.0006,
+ "loss": 5.391471862792969,
+ "step": 3464
+ },
+ {
+ "epoch": 48.12576419213974,
+ "grad_norm": 0.02651144750416279,
+ "learning_rate": 0.0006,
+ "loss": 5.422799110412598,
+ "step": 3465
+ },
+ {
+ "epoch": 48.13973799126637,
+ "grad_norm": 0.03316329047083855,
+ "learning_rate": 0.0006,
+ "loss": 5.3952202796936035,
+ "step": 3466
+ },
+ {
+ "epoch": 48.15371179039301,
+ "grad_norm": 0.04012278467416763,
+ "learning_rate": 0.0006,
+ "loss": 5.540853023529053,
+ "step": 3467
+ },
+ {
+ "epoch": 48.16768558951965,
+ "grad_norm": 0.04763203114271164,
+ "learning_rate": 0.0006,
+ "loss": 5.387890338897705,
+ "step": 3468
+ },
+ {
+ "epoch": 48.18165938864629,
+ "grad_norm": 0.05129165202379227,
+ "learning_rate": 0.0006,
+ "loss": 5.414722442626953,
+ "step": 3469
+ },
+ {
+ "epoch": 48.19563318777293,
+ "grad_norm": 0.04438071325421333,
+ "learning_rate": 0.0006,
+ "loss": 5.386244773864746,
+ "step": 3470
+ },
+ {
+ "epoch": 48.209606986899566,
+ "grad_norm": 0.040234919637441635,
+ "learning_rate": 0.0006,
+ "loss": 5.431617736816406,
+ "step": 3471
+ },
+ {
+ "epoch": 48.223580786026204,
+ "grad_norm": 0.03217008337378502,
+ "learning_rate": 0.0006,
+ "loss": 5.30168342590332,
+ "step": 3472
+ },
+ {
+ "epoch": 48.237554585152836,
+ "grad_norm": 0.03087974339723587,
+ "learning_rate": 0.0006,
+ "loss": 5.316389083862305,
+ "step": 3473
+ },
+ {
+ "epoch": 48.251528384279474,
+ "grad_norm": 0.02789347991347313,
+ "learning_rate": 0.0006,
+ "loss": 5.385970592498779,
+ "step": 3474
+ },
+ {
+ "epoch": 48.26550218340611,
+ "grad_norm": 0.024792097508907318,
+ "learning_rate": 0.0006,
+ "loss": 5.362329959869385,
+ "step": 3475
+ },
+ {
+ "epoch": 48.27947598253275,
+ "grad_norm": 0.025869522243738174,
+ "learning_rate": 0.0006,
+ "loss": 5.28400182723999,
+ "step": 3476
+ },
+ {
+ "epoch": 48.29344978165939,
+ "grad_norm": 0.021339882165193558,
+ "learning_rate": 0.0006,
+ "loss": 5.36669921875,
+ "step": 3477
+ },
+ {
+ "epoch": 48.30742358078603,
+ "grad_norm": 0.021462520584464073,
+ "learning_rate": 0.0006,
+ "loss": 5.311222076416016,
+ "step": 3478
+ },
+ {
+ "epoch": 48.32139737991266,
+ "grad_norm": 0.02193879708647728,
+ "learning_rate": 0.0006,
+ "loss": 5.291441917419434,
+ "step": 3479
+ },
+ {
+ "epoch": 48.3353711790393,
+ "grad_norm": 0.01963093690574169,
+ "learning_rate": 0.0006,
+ "loss": 5.342780113220215,
+ "step": 3480
+ },
+ {
+ "epoch": 48.34934497816594,
+ "grad_norm": 0.020634736865758896,
+ "learning_rate": 0.0006,
+ "loss": 5.144141674041748,
+ "step": 3481
+ },
+ {
+ "epoch": 48.36331877729258,
+ "grad_norm": 0.02126968279480934,
+ "learning_rate": 0.0006,
+ "loss": 5.232624530792236,
+ "step": 3482
+ },
+ {
+ "epoch": 48.377292576419215,
+ "grad_norm": 0.021203994750976562,
+ "learning_rate": 0.0006,
+ "loss": 5.290219783782959,
+ "step": 3483
+ },
+ {
+ "epoch": 48.391266375545854,
+ "grad_norm": 0.02193225920200348,
+ "learning_rate": 0.0006,
+ "loss": 5.29197359085083,
+ "step": 3484
+ },
+ {
+ "epoch": 48.40524017467249,
+ "grad_norm": 0.023174211382865906,
+ "learning_rate": 0.0006,
+ "loss": 5.204349517822266,
+ "step": 3485
+ },
+ {
+ "epoch": 48.419213973799124,
+ "grad_norm": 0.024161914363503456,
+ "learning_rate": 0.0006,
+ "loss": 5.339351654052734,
+ "step": 3486
+ },
+ {
+ "epoch": 48.43318777292576,
+ "grad_norm": 0.02221524342894554,
+ "learning_rate": 0.0006,
+ "loss": 5.263001441955566,
+ "step": 3487
+ },
+ {
+ "epoch": 48.4471615720524,
+ "grad_norm": 0.026632830500602722,
+ "learning_rate": 0.0006,
+ "loss": 5.221281051635742,
+ "step": 3488
+ },
+ {
+ "epoch": 48.46113537117904,
+ "grad_norm": 0.034923870116472244,
+ "learning_rate": 0.0006,
+ "loss": 5.296039581298828,
+ "step": 3489
+ },
+ {
+ "epoch": 48.47510917030568,
+ "grad_norm": 0.046184979379177094,
+ "learning_rate": 0.0006,
+ "loss": 5.212769985198975,
+ "step": 3490
+ },
+ {
+ "epoch": 48.48908296943232,
+ "grad_norm": 0.05154098570346832,
+ "learning_rate": 0.0006,
+ "loss": 5.337864875793457,
+ "step": 3491
+ },
+ {
+ "epoch": 48.50305676855895,
+ "grad_norm": 0.042967259883880615,
+ "learning_rate": 0.0006,
+ "loss": 5.267958641052246,
+ "step": 3492
+ },
+ {
+ "epoch": 48.51703056768559,
+ "grad_norm": 0.05356583371758461,
+ "learning_rate": 0.0006,
+ "loss": 5.1404218673706055,
+ "step": 3493
+ },
+ {
+ "epoch": 48.531004366812226,
+ "grad_norm": 0.08219032734632492,
+ "learning_rate": 0.0006,
+ "loss": 5.300687789916992,
+ "step": 3494
+ },
+ {
+ "epoch": 48.544978165938865,
+ "grad_norm": 0.19204089045524597,
+ "learning_rate": 0.0006,
+ "loss": 5.456025123596191,
+ "step": 3495
+ },
+ {
+ "epoch": 48.5589519650655,
+ "grad_norm": 0.25917696952819824,
+ "learning_rate": 0.0006,
+ "loss": 6.156761169433594,
+ "step": 3496
+ },
+ {
+ "epoch": 48.57292576419214,
+ "grad_norm": 0.1566283404827118,
+ "learning_rate": 0.0006,
+ "loss": 5.862346649169922,
+ "step": 3497
+ },
+ {
+ "epoch": 48.58689956331878,
+ "grad_norm": 0.1761486828327179,
+ "learning_rate": 0.0006,
+ "loss": 6.012795448303223,
+ "step": 3498
+ },
+ {
+ "epoch": 48.60087336244541,
+ "grad_norm": 0.09857621043920517,
+ "learning_rate": 0.0006,
+ "loss": 5.9248151779174805,
+ "step": 3499
+ },
+ {
+ "epoch": 48.61484716157205,
+ "grad_norm": 0.05968055501580238,
+ "learning_rate": 0.0006,
+ "loss": 5.780706882476807,
+ "step": 3500
+ },
+ {
+ "epoch": 48.62882096069869,
+ "grad_norm": 0.057433564215898514,
+ "learning_rate": 0.0006,
+ "loss": 5.645388603210449,
+ "step": 3501
+ },
+ {
+ "epoch": 48.64279475982533,
+ "grad_norm": 0.039659615606069565,
+ "learning_rate": 0.0006,
+ "loss": 5.695138931274414,
+ "step": 3502
+ },
+ {
+ "epoch": 48.65676855895197,
+ "grad_norm": 0.04019695147871971,
+ "learning_rate": 0.0006,
+ "loss": 5.672435283660889,
+ "step": 3503
+ },
+ {
+ "epoch": 48.670742358078606,
+ "grad_norm": 0.03033018670976162,
+ "learning_rate": 0.0006,
+ "loss": 5.682260513305664,
+ "step": 3504
+ },
+ {
+ "epoch": 48.68471615720524,
+ "grad_norm": 0.03215787187218666,
+ "learning_rate": 0.0006,
+ "loss": 5.604182243347168,
+ "step": 3505
+ },
+ {
+ "epoch": 48.698689956331876,
+ "grad_norm": 0.026865126565098763,
+ "learning_rate": 0.0006,
+ "loss": 5.64040470123291,
+ "step": 3506
+ },
+ {
+ "epoch": 48.712663755458514,
+ "grad_norm": 0.02831609919667244,
+ "learning_rate": 0.0006,
+ "loss": 5.552720069885254,
+ "step": 3507
+ },
+ {
+ "epoch": 48.72663755458515,
+ "grad_norm": 0.024226613342761993,
+ "learning_rate": 0.0006,
+ "loss": 5.556946277618408,
+ "step": 3508
+ },
+ {
+ "epoch": 48.74061135371179,
+ "grad_norm": 0.025393934920430183,
+ "learning_rate": 0.0006,
+ "loss": 5.476178169250488,
+ "step": 3509
+ },
+ {
+ "epoch": 48.75458515283843,
+ "grad_norm": 0.01984412968158722,
+ "learning_rate": 0.0006,
+ "loss": 5.471088409423828,
+ "step": 3510
+ },
+ {
+ "epoch": 48.76855895196506,
+ "grad_norm": 0.02035129815340042,
+ "learning_rate": 0.0006,
+ "loss": 5.438812732696533,
+ "step": 3511
+ },
+ {
+ "epoch": 48.7825327510917,
+ "grad_norm": 0.0208350270986557,
+ "learning_rate": 0.0006,
+ "loss": 5.381746292114258,
+ "step": 3512
+ },
+ {
+ "epoch": 48.79650655021834,
+ "grad_norm": 0.017783882096409798,
+ "learning_rate": 0.0006,
+ "loss": 5.351675510406494,
+ "step": 3513
+ },
+ {
+ "epoch": 48.81048034934498,
+ "grad_norm": 0.01671770215034485,
+ "learning_rate": 0.0006,
+ "loss": 5.329428672790527,
+ "step": 3514
+ },
+ {
+ "epoch": 48.82445414847162,
+ "grad_norm": 0.017891215160489082,
+ "learning_rate": 0.0006,
+ "loss": 5.212074279785156,
+ "step": 3515
+ },
+ {
+ "epoch": 48.838427947598255,
+ "grad_norm": 0.016847791150212288,
+ "learning_rate": 0.0006,
+ "loss": 5.299534797668457,
+ "step": 3516
+ },
+ {
+ "epoch": 48.852401746724894,
+ "grad_norm": 0.01593637838959694,
+ "learning_rate": 0.0006,
+ "loss": 5.30466890335083,
+ "step": 3517
+ },
+ {
+ "epoch": 48.866375545851525,
+ "grad_norm": 0.014713102020323277,
+ "learning_rate": 0.0006,
+ "loss": 5.294345378875732,
+ "step": 3518
+ },
+ {
+ "epoch": 48.880349344978164,
+ "grad_norm": 0.01433857623487711,
+ "learning_rate": 0.0006,
+ "loss": 5.2148542404174805,
+ "step": 3519
+ },
+ {
+ "epoch": 48.8943231441048,
+ "grad_norm": 0.014979379251599312,
+ "learning_rate": 0.0006,
+ "loss": 5.298582077026367,
+ "step": 3520
+ },
+ {
+ "epoch": 48.90829694323144,
+ "grad_norm": 0.013494192622601986,
+ "learning_rate": 0.0006,
+ "loss": 5.309783935546875,
+ "step": 3521
+ },
+ {
+ "epoch": 48.92227074235808,
+ "grad_norm": 0.014020012691617012,
+ "learning_rate": 0.0006,
+ "loss": 5.284879684448242,
+ "step": 3522
+ },
+ {
+ "epoch": 48.93624454148472,
+ "grad_norm": 0.01348058134317398,
+ "learning_rate": 0.0006,
+ "loss": 5.249116897583008,
+ "step": 3523
+ },
+ {
+ "epoch": 48.95021834061135,
+ "grad_norm": 0.013280218467116356,
+ "learning_rate": 0.0006,
+ "loss": 5.273489952087402,
+ "step": 3524
+ },
+ {
+ "epoch": 48.96419213973799,
+ "grad_norm": 0.012647110037505627,
+ "learning_rate": 0.0006,
+ "loss": 5.306652069091797,
+ "step": 3525
+ },
+ {
+ "epoch": 48.97816593886463,
+ "grad_norm": 0.012847724370658398,
+ "learning_rate": 0.0006,
+ "loss": 5.235745429992676,
+ "step": 3526
+ },
+ {
+ "epoch": 48.992139737991266,
+ "grad_norm": 0.011820166371762753,
+ "learning_rate": 0.0006,
+ "loss": 5.238091468811035,
+ "step": 3527
+ },
+ {
+ "epoch": 49.0,
+ "grad_norm": 0.01255585066974163,
+ "learning_rate": 0.0006,
+ "loss": 5.231295585632324,
+ "step": 3528
+ },
+ {
+ "epoch": 49.0,
+ "eval_loss": 5.394587993621826,
+ "eval_runtime": 56.3734,
+ "eval_samples_per_second": 43.318,
+ "eval_steps_per_second": 1.366,
+ "step": 3528
+ },
+ {
+ "epoch": 49.01397379912664,
+ "grad_norm": 0.011031972244381905,
+ "learning_rate": 0.0006,
+ "loss": 5.1354570388793945,
+ "step": 3529
+ },
+ {
+ "epoch": 49.02794759825328,
+ "grad_norm": 0.010631263256072998,
+ "learning_rate": 0.0006,
+ "loss": 5.177639961242676,
+ "step": 3530
+ },
+ {
+ "epoch": 49.041921397379916,
+ "grad_norm": 0.01133775431662798,
+ "learning_rate": 0.0006,
+ "loss": 5.283428192138672,
+ "step": 3531
+ },
+ {
+ "epoch": 49.05589519650655,
+ "grad_norm": 0.01154172234237194,
+ "learning_rate": 0.0006,
+ "loss": 5.206364154815674,
+ "step": 3532
+ },
+ {
+ "epoch": 49.069868995633186,
+ "grad_norm": 0.010192356072366238,
+ "learning_rate": 0.0006,
+ "loss": 5.2385735511779785,
+ "step": 3533
+ },
+ {
+ "epoch": 49.083842794759825,
+ "grad_norm": 0.012789200060069561,
+ "learning_rate": 0.0006,
+ "loss": 5.071800231933594,
+ "step": 3534
+ },
+ {
+ "epoch": 49.09781659388646,
+ "grad_norm": 0.011282820254564285,
+ "learning_rate": 0.0006,
+ "loss": 5.218020439147949,
+ "step": 3535
+ },
+ {
+ "epoch": 49.1117903930131,
+ "grad_norm": 0.012153448536992073,
+ "learning_rate": 0.0006,
+ "loss": 5.134698867797852,
+ "step": 3536
+ },
+ {
+ "epoch": 49.12576419213974,
+ "grad_norm": 0.011920543387532234,
+ "learning_rate": 0.0006,
+ "loss": 5.141620635986328,
+ "step": 3537
+ },
+ {
+ "epoch": 49.13973799126637,
+ "grad_norm": 0.010949329473078251,
+ "learning_rate": 0.0006,
+ "loss": 5.116259574890137,
+ "step": 3538
+ },
+ {
+ "epoch": 49.15371179039301,
+ "grad_norm": 0.012362448498606682,
+ "learning_rate": 0.0006,
+ "loss": 5.090177059173584,
+ "step": 3539
+ },
+ {
+ "epoch": 49.16768558951965,
+ "grad_norm": 0.011575071141123772,
+ "learning_rate": 0.0006,
+ "loss": 5.093441009521484,
+ "step": 3540
+ },
+ {
+ "epoch": 49.18165938864629,
+ "grad_norm": 0.011126236990094185,
+ "learning_rate": 0.0006,
+ "loss": 5.060235023498535,
+ "step": 3541
+ },
+ {
+ "epoch": 49.19563318777293,
+ "grad_norm": 0.0107401292771101,
+ "learning_rate": 0.0006,
+ "loss": 5.140704154968262,
+ "step": 3542
+ },
+ {
+ "epoch": 49.209606986899566,
+ "grad_norm": 0.011202906258404255,
+ "learning_rate": 0.0006,
+ "loss": 5.10855770111084,
+ "step": 3543
+ },
+ {
+ "epoch": 49.223580786026204,
+ "grad_norm": 0.011083454824984074,
+ "learning_rate": 0.0006,
+ "loss": 5.092751502990723,
+ "step": 3544
+ },
+ {
+ "epoch": 49.237554585152836,
+ "grad_norm": 0.009523800574243069,
+ "learning_rate": 0.0006,
+ "loss": 5.13733434677124,
+ "step": 3545
+ },
+ {
+ "epoch": 49.251528384279474,
+ "grad_norm": 0.011298353783786297,
+ "learning_rate": 0.0006,
+ "loss": 5.036048889160156,
+ "step": 3546
+ },
+ {
+ "epoch": 49.26550218340611,
+ "grad_norm": 0.010676483623683453,
+ "learning_rate": 0.0006,
+ "loss": 5.051901340484619,
+ "step": 3547
+ },
+ {
+ "epoch": 49.27947598253275,
+ "grad_norm": 0.011463399045169353,
+ "learning_rate": 0.0006,
+ "loss": 4.983170986175537,
+ "step": 3548
+ },
+ {
+ "epoch": 49.29344978165939,
+ "grad_norm": 0.010603402741253376,
+ "learning_rate": 0.0006,
+ "loss": 5.112311363220215,
+ "step": 3549
+ },
+ {
+ "epoch": 49.30742358078603,
+ "grad_norm": 0.010535670444369316,
+ "learning_rate": 0.0006,
+ "loss": 5.085227966308594,
+ "step": 3550
+ },
+ {
+ "epoch": 49.32139737991266,
+ "grad_norm": 0.011843928135931492,
+ "learning_rate": 0.0006,
+ "loss": 5.07575798034668,
+ "step": 3551
+ },
+ {
+ "epoch": 49.3353711790393,
+ "grad_norm": 0.010587952099740505,
+ "learning_rate": 0.0006,
+ "loss": 5.019237041473389,
+ "step": 3552
+ },
+ {
+ "epoch": 49.34934497816594,
+ "grad_norm": 0.012285659089684486,
+ "learning_rate": 0.0006,
+ "loss": 5.144603729248047,
+ "step": 3553
+ },
+ {
+ "epoch": 49.36331877729258,
+ "grad_norm": 0.013119431212544441,
+ "learning_rate": 0.0006,
+ "loss": 5.035256862640381,
+ "step": 3554
+ },
+ {
+ "epoch": 49.377292576419215,
+ "grad_norm": 0.012138537131249905,
+ "learning_rate": 0.0006,
+ "loss": 4.996667861938477,
+ "step": 3555
+ },
+ {
+ "epoch": 49.391266375545854,
+ "grad_norm": 0.012142804451286793,
+ "learning_rate": 0.0006,
+ "loss": 4.95058536529541,
+ "step": 3556
+ },
+ {
+ "epoch": 49.40524017467249,
+ "grad_norm": 0.012205676175653934,
+ "learning_rate": 0.0006,
+ "loss": 5.096931457519531,
+ "step": 3557
+ },
+ {
+ "epoch": 49.419213973799124,
+ "grad_norm": 0.012086744420230389,
+ "learning_rate": 0.0006,
+ "loss": 4.99384880065918,
+ "step": 3558
+ },
+ {
+ "epoch": 49.43318777292576,
+ "grad_norm": 0.012918103486299515,
+ "learning_rate": 0.0006,
+ "loss": 5.031294822692871,
+ "step": 3559
+ },
+ {
+ "epoch": 49.4471615720524,
+ "grad_norm": 0.012006713077425957,
+ "learning_rate": 0.0006,
+ "loss": 5.0547943115234375,
+ "step": 3560
+ },
+ {
+ "epoch": 49.46113537117904,
+ "grad_norm": 0.0120399734005332,
+ "learning_rate": 0.0006,
+ "loss": 5.014238357543945,
+ "step": 3561
+ },
+ {
+ "epoch": 49.47510917030568,
+ "grad_norm": 0.013306990265846252,
+ "learning_rate": 0.0006,
+ "loss": 4.972925186157227,
+ "step": 3562
+ },
+ {
+ "epoch": 49.48908296943232,
+ "grad_norm": 0.013878636993467808,
+ "learning_rate": 0.0006,
+ "loss": 5.079033851623535,
+ "step": 3563
+ },
+ {
+ "epoch": 49.50305676855895,
+ "grad_norm": 0.012454566545784473,
+ "learning_rate": 0.0006,
+ "loss": 4.93708610534668,
+ "step": 3564
+ },
+ {
+ "epoch": 49.51703056768559,
+ "grad_norm": 0.014075061306357384,
+ "learning_rate": 0.0006,
+ "loss": 4.965582370758057,
+ "step": 3565
+ },
+ {
+ "epoch": 49.531004366812226,
+ "grad_norm": 0.012525617145001888,
+ "learning_rate": 0.0006,
+ "loss": 4.968148708343506,
+ "step": 3566
+ },
+ {
+ "epoch": 49.544978165938865,
+ "grad_norm": 0.012418225407600403,
+ "learning_rate": 0.0006,
+ "loss": 5.06901741027832,
+ "step": 3567
+ },
+ {
+ "epoch": 49.5589519650655,
+ "grad_norm": 0.015034239739179611,
+ "learning_rate": 0.0006,
+ "loss": 4.984277725219727,
+ "step": 3568
+ },
+ {
+ "epoch": 49.57292576419214,
+ "grad_norm": 0.014329628087580204,
+ "learning_rate": 0.0006,
+ "loss": 4.951558589935303,
+ "step": 3569
+ },
+ {
+ "epoch": 49.58689956331878,
+ "grad_norm": 0.017769070342183113,
+ "learning_rate": 0.0006,
+ "loss": 4.966532230377197,
+ "step": 3570
+ },
+ {
+ "epoch": 49.60087336244541,
+ "grad_norm": 0.01909453794360161,
+ "learning_rate": 0.0006,
+ "loss": 4.919890880584717,
+ "step": 3571
+ },
+ {
+ "epoch": 49.61484716157205,
+ "grad_norm": 0.017519688233733177,
+ "learning_rate": 0.0006,
+ "loss": 5.031286716461182,
+ "step": 3572
+ },
+ {
+ "epoch": 49.62882096069869,
+ "grad_norm": 0.017143897712230682,
+ "learning_rate": 0.0006,
+ "loss": 4.969764709472656,
+ "step": 3573
+ },
+ {
+ "epoch": 49.64279475982533,
+ "grad_norm": 0.01688285917043686,
+ "learning_rate": 0.0006,
+ "loss": 5.0733442306518555,
+ "step": 3574
+ },
+ {
+ "epoch": 49.65676855895197,
+ "grad_norm": 0.018547311425209045,
+ "learning_rate": 0.0006,
+ "loss": 4.980863571166992,
+ "step": 3575
+ },
+ {
+ "epoch": 49.670742358078606,
+ "grad_norm": 0.018877137452363968,
+ "learning_rate": 0.0006,
+ "loss": 4.954646110534668,
+ "step": 3576
+ },
+ {
+ "epoch": 49.68471615720524,
+ "grad_norm": 0.018681600689888,
+ "learning_rate": 0.0006,
+ "loss": 4.959702014923096,
+ "step": 3577
+ },
+ {
+ "epoch": 49.698689956331876,
+ "grad_norm": 0.018947336822748184,
+ "learning_rate": 0.0006,
+ "loss": 5.070502758026123,
+ "step": 3578
+ },
+ {
+ "epoch": 49.712663755458514,
+ "grad_norm": 0.018817484378814697,
+ "learning_rate": 0.0006,
+ "loss": 5.056508541107178,
+ "step": 3579
+ },
+ {
+ "epoch": 49.72663755458515,
+ "grad_norm": 0.01931724324822426,
+ "learning_rate": 0.0006,
+ "loss": 4.922451972961426,
+ "step": 3580
+ },
+ {
+ "epoch": 49.74061135371179,
+ "grad_norm": 0.022047819569706917,
+ "learning_rate": 0.0006,
+ "loss": 5.053170204162598,
+ "step": 3581
+ },
+ {
+ "epoch": 49.75458515283843,
+ "grad_norm": 0.02279801294207573,
+ "learning_rate": 0.0006,
+ "loss": 4.8768439292907715,
+ "step": 3582
+ },
+ {
+ "epoch": 49.76855895196506,
+ "grad_norm": 0.020661601796746254,
+ "learning_rate": 0.0006,
+ "loss": 4.907928466796875,
+ "step": 3583
+ },
+ {
+ "epoch": 49.7825327510917,
+ "grad_norm": 0.020479509606957436,
+ "learning_rate": 0.0006,
+ "loss": 5.040694236755371,
+ "step": 3584
+ },
+ {
+ "epoch": 49.79650655021834,
+ "grad_norm": 0.021654162555933,
+ "learning_rate": 0.0006,
+ "loss": 5.018434524536133,
+ "step": 3585
+ },
+ {
+ "epoch": 49.81048034934498,
+ "grad_norm": 0.022486824542284012,
+ "learning_rate": 0.0006,
+ "loss": 5.086538314819336,
+ "step": 3586
+ },
+ {
+ "epoch": 49.82445414847162,
+ "grad_norm": 0.02141369879245758,
+ "learning_rate": 0.0006,
+ "loss": 5.035694122314453,
+ "step": 3587
+ },
+ {
+ "epoch": 49.838427947598255,
+ "grad_norm": 0.015926677733659744,
+ "learning_rate": 0.0006,
+ "loss": 4.905038833618164,
+ "step": 3588
+ },
+ {
+ "epoch": 49.852401746724894,
+ "grad_norm": 0.015775298699736595,
+ "learning_rate": 0.0006,
+ "loss": 5.003902435302734,
+ "step": 3589
+ },
+ {
+ "epoch": 49.866375545851525,
+ "grad_norm": 0.01872118003666401,
+ "learning_rate": 0.0006,
+ "loss": 4.891060829162598,
+ "step": 3590
+ },
+ {
+ "epoch": 49.880349344978164,
+ "grad_norm": 0.02066192403435707,
+ "learning_rate": 0.0006,
+ "loss": 5.002897262573242,
+ "step": 3591
+ },
+ {
+ "epoch": 49.8943231441048,
+ "grad_norm": 0.02348714880645275,
+ "learning_rate": 0.0006,
+ "loss": 4.9189958572387695,
+ "step": 3592
+ },
+ {
+ "epoch": 49.90829694323144,
+ "grad_norm": 0.020402103662490845,
+ "learning_rate": 0.0006,
+ "loss": 4.992661476135254,
+ "step": 3593
+ },
+ {
+ "epoch": 49.92227074235808,
+ "grad_norm": 0.020466072484850883,
+ "learning_rate": 0.0006,
+ "loss": 4.962968826293945,
+ "step": 3594
+ },
+ {
+ "epoch": 49.93624454148472,
+ "grad_norm": 0.01922314241528511,
+ "learning_rate": 0.0006,
+ "loss": 4.897730827331543,
+ "step": 3595
+ },
+ {
+ "epoch": 49.95021834061135,
+ "grad_norm": 0.017228754237294197,
+ "learning_rate": 0.0006,
+ "loss": 4.964842796325684,
+ "step": 3596
+ },
+ {
+ "epoch": 49.96419213973799,
+ "grad_norm": 0.016021743416786194,
+ "learning_rate": 0.0006,
+ "loss": 4.910758972167969,
+ "step": 3597
+ },
+ {
+ "epoch": 49.97816593886463,
+ "grad_norm": 0.017985103651881218,
+ "learning_rate": 0.0006,
+ "loss": 4.916072368621826,
+ "step": 3598
+ },
+ {
+ "epoch": 49.992139737991266,
+ "grad_norm": 0.018780240789055824,
+ "learning_rate": 0.0006,
+ "loss": 4.94658088684082,
+ "step": 3599
+ },
+ {
+ "epoch": 50.0,
+ "grad_norm": 0.01776314526796341,
+ "learning_rate": 0.0006,
+ "loss": 4.969357967376709,
+ "step": 3600
+ },
+ {
+ "epoch": 50.0,
+ "eval_loss": 5.141158580780029,
+ "eval_runtime": 56.2332,
+ "eval_samples_per_second": 43.426,
+ "eval_steps_per_second": 1.369,
+ "step": 3600
+ },
+ {
+ "epoch": 50.01397379912664,
+ "grad_norm": 0.016308696940541267,
+ "learning_rate": 0.0006,
+ "loss": 4.924664497375488,
+ "step": 3601
+ },
+ {
+ "epoch": 50.02794759825328,
+ "grad_norm": 0.018810199573636055,
+ "learning_rate": 0.0006,
+ "loss": 4.82293701171875,
+ "step": 3602
+ },
+ {
+ "epoch": 50.041921397379916,
+ "grad_norm": 0.02227160707116127,
+ "learning_rate": 0.0006,
+ "loss": 4.9612298011779785,
+ "step": 3603
+ },
+ {
+ "epoch": 50.05589519650655,
+ "grad_norm": 0.025266462936997414,
+ "learning_rate": 0.0006,
+ "loss": 4.963835716247559,
+ "step": 3604
+ },
+ {
+ "epoch": 50.069868995633186,
+ "grad_norm": 0.028588181361556053,
+ "learning_rate": 0.0006,
+ "loss": 4.992118835449219,
+ "step": 3605
+ },
+ {
+ "epoch": 50.083842794759825,
+ "grad_norm": 0.02739487960934639,
+ "learning_rate": 0.0006,
+ "loss": 4.915547847747803,
+ "step": 3606
+ },
+ {
+ "epoch": 50.09781659388646,
+ "grad_norm": 0.027000300586223602,
+ "learning_rate": 0.0006,
+ "loss": 4.883801460266113,
+ "step": 3607
+ },
+ {
+ "epoch": 50.1117903930131,
+ "grad_norm": 0.023468483239412308,
+ "learning_rate": 0.0006,
+ "loss": 5.006168365478516,
+ "step": 3608
+ },
+ {
+ "epoch": 50.12576419213974,
+ "grad_norm": 0.026753634214401245,
+ "learning_rate": 0.0006,
+ "loss": 4.92526912689209,
+ "step": 3609
+ },
+ {
+ "epoch": 50.13973799126637,
+ "grad_norm": 0.026816733181476593,
+ "learning_rate": 0.0006,
+ "loss": 4.847546100616455,
+ "step": 3610
+ },
+ {
+ "epoch": 50.15371179039301,
+ "grad_norm": 0.027362391352653503,
+ "learning_rate": 0.0006,
+ "loss": 4.862548351287842,
+ "step": 3611
+ },
+ {
+ "epoch": 50.16768558951965,
+ "grad_norm": 0.03450845181941986,
+ "learning_rate": 0.0006,
+ "loss": 4.963202953338623,
+ "step": 3612
+ },
+ {
+ "epoch": 50.18165938864629,
+ "grad_norm": 0.03801408037543297,
+ "learning_rate": 0.0006,
+ "loss": 4.770240306854248,
+ "step": 3613
+ },
+ {
+ "epoch": 50.19563318777293,
+ "grad_norm": 0.03579140082001686,
+ "learning_rate": 0.0006,
+ "loss": 4.797981262207031,
+ "step": 3614
+ },
+ {
+ "epoch": 50.209606986899566,
+ "grad_norm": 0.029393760487437248,
+ "learning_rate": 0.0006,
+ "loss": 4.798581600189209,
+ "step": 3615
+ },
+ {
+ "epoch": 50.223580786026204,
+ "grad_norm": 0.026111142709851265,
+ "learning_rate": 0.0006,
+ "loss": 4.990712642669678,
+ "step": 3616
+ },
+ {
+ "epoch": 50.237554585152836,
+ "grad_norm": 0.02213387005031109,
+ "learning_rate": 0.0006,
+ "loss": 4.865842342376709,
+ "step": 3617
+ },
+ {
+ "epoch": 50.251528384279474,
+ "grad_norm": 0.025450684130191803,
+ "learning_rate": 0.0006,
+ "loss": 4.903746128082275,
+ "step": 3618
+ },
+ {
+ "epoch": 50.26550218340611,
+ "grad_norm": 0.020915117114782333,
+ "learning_rate": 0.0006,
+ "loss": 4.951366424560547,
+ "step": 3619
+ },
+ {
+ "epoch": 50.27947598253275,
+ "grad_norm": 0.020972849801182747,
+ "learning_rate": 0.0006,
+ "loss": 4.783311367034912,
+ "step": 3620
+ },
+ {
+ "epoch": 50.29344978165939,
+ "grad_norm": 0.016956273466348648,
+ "learning_rate": 0.0006,
+ "loss": 4.955767631530762,
+ "step": 3621
+ },
+ {
+ "epoch": 50.30742358078603,
+ "grad_norm": 0.018045252189040184,
+ "learning_rate": 0.0006,
+ "loss": 4.869322776794434,
+ "step": 3622
+ },
+ {
+ "epoch": 50.32139737991266,
+ "grad_norm": 0.018199188634753227,
+ "learning_rate": 0.0006,
+ "loss": 4.95680046081543,
+ "step": 3623
+ },
+ {
+ "epoch": 50.3353711790393,
+ "grad_norm": 0.016688082367181778,
+ "learning_rate": 0.0006,
+ "loss": 4.783535957336426,
+ "step": 3624
+ },
+ {
+ "epoch": 50.34934497816594,
+ "grad_norm": 0.017540352419018745,
+ "learning_rate": 0.0006,
+ "loss": 4.8236083984375,
+ "step": 3625
+ },
+ {
+ "epoch": 50.36331877729258,
+ "grad_norm": 0.018054187297821045,
+ "learning_rate": 0.0006,
+ "loss": 4.764655113220215,
+ "step": 3626
+ },
+ {
+ "epoch": 50.377292576419215,
+ "grad_norm": 0.01741131953895092,
+ "learning_rate": 0.0006,
+ "loss": 4.875613689422607,
+ "step": 3627
+ },
+ {
+ "epoch": 50.391266375545854,
+ "grad_norm": 0.01682269014418125,
+ "learning_rate": 0.0006,
+ "loss": 4.818509101867676,
+ "step": 3628
+ },
+ {
+ "epoch": 50.40524017467249,
+ "grad_norm": 0.016652341932058334,
+ "learning_rate": 0.0006,
+ "loss": 4.974020481109619,
+ "step": 3629
+ },
+ {
+ "epoch": 50.419213973799124,
+ "grad_norm": 0.02035282365977764,
+ "learning_rate": 0.0006,
+ "loss": 4.893037796020508,
+ "step": 3630
+ },
+ {
+ "epoch": 50.43318777292576,
+ "grad_norm": 0.019907061010599136,
+ "learning_rate": 0.0006,
+ "loss": 4.848119735717773,
+ "step": 3631
+ },
+ {
+ "epoch": 50.4471615720524,
+ "grad_norm": 0.023861493915319443,
+ "learning_rate": 0.0006,
+ "loss": 4.819353103637695,
+ "step": 3632
+ },
+ {
+ "epoch": 50.46113537117904,
+ "grad_norm": 0.025254026055336,
+ "learning_rate": 0.0006,
+ "loss": 4.769312858581543,
+ "step": 3633
+ },
+ {
+ "epoch": 50.47510917030568,
+ "grad_norm": 0.026154937222599983,
+ "learning_rate": 0.0006,
+ "loss": 4.7586774826049805,
+ "step": 3634
+ },
+ {
+ "epoch": 50.48908296943232,
+ "grad_norm": 0.026342090219259262,
+ "learning_rate": 0.0006,
+ "loss": 4.861218452453613,
+ "step": 3635
+ },
+ {
+ "epoch": 50.50305676855895,
+ "grad_norm": 0.028169559314846992,
+ "learning_rate": 0.0006,
+ "loss": 4.818380355834961,
+ "step": 3636
+ },
+ {
+ "epoch": 50.51703056768559,
+ "grad_norm": 0.02820264920592308,
+ "learning_rate": 0.0006,
+ "loss": 4.892871856689453,
+ "step": 3637
+ },
+ {
+ "epoch": 50.531004366812226,
+ "grad_norm": 0.03345080465078354,
+ "learning_rate": 0.0006,
+ "loss": 4.811084747314453,
+ "step": 3638
+ },
+ {
+ "epoch": 50.544978165938865,
+ "grad_norm": 0.04005736857652664,
+ "learning_rate": 0.0006,
+ "loss": 4.765410900115967,
+ "step": 3639
+ },
+ {
+ "epoch": 50.5589519650655,
+ "grad_norm": 0.04333870857954025,
+ "learning_rate": 0.0006,
+ "loss": 4.956520080566406,
+ "step": 3640
+ },
+ {
+ "epoch": 50.57292576419214,
+ "grad_norm": 0.04216952994465828,
+ "learning_rate": 0.0006,
+ "loss": 4.793441295623779,
+ "step": 3641
+ },
+ {
+ "epoch": 50.58689956331878,
+ "grad_norm": 0.02936951443552971,
+ "learning_rate": 0.0006,
+ "loss": 4.8447957038879395,
+ "step": 3642
+ },
+ {
+ "epoch": 50.60087336244541,
+ "grad_norm": 0.026937026530504227,
+ "learning_rate": 0.0006,
+ "loss": 4.962215423583984,
+ "step": 3643
+ },
+ {
+ "epoch": 50.61484716157205,
+ "grad_norm": 0.02628890983760357,
+ "learning_rate": 0.0006,
+ "loss": 4.883061408996582,
+ "step": 3644
+ },
+ {
+ "epoch": 50.62882096069869,
+ "grad_norm": 0.025984834879636765,
+ "learning_rate": 0.0006,
+ "loss": 4.768662452697754,
+ "step": 3645
+ },
+ {
+ "epoch": 50.64279475982533,
+ "grad_norm": 0.023998012766242027,
+ "learning_rate": 0.0006,
+ "loss": 4.851325035095215,
+ "step": 3646
+ },
+ {
+ "epoch": 50.65676855895197,
+ "grad_norm": 0.02143184095621109,
+ "learning_rate": 0.0006,
+ "loss": 4.8274688720703125,
+ "step": 3647
+ },
+ {
+ "epoch": 50.670742358078606,
+ "grad_norm": 0.02488900162279606,
+ "learning_rate": 0.0006,
+ "loss": 4.831140518188477,
+ "step": 3648
+ },
+ {
+ "epoch": 50.68471615720524,
+ "grad_norm": 0.02330533228814602,
+ "learning_rate": 0.0006,
+ "loss": 4.850462913513184,
+ "step": 3649
+ },
+ {
+ "epoch": 50.698689956331876,
+ "grad_norm": 0.02121943049132824,
+ "learning_rate": 0.0006,
+ "loss": 4.870253086090088,
+ "step": 3650
+ },
+ {
+ "epoch": 50.712663755458514,
+ "grad_norm": 0.02080935798585415,
+ "learning_rate": 0.0006,
+ "loss": 4.798770904541016,
+ "step": 3651
+ },
+ {
+ "epoch": 50.72663755458515,
+ "grad_norm": 0.019047029316425323,
+ "learning_rate": 0.0006,
+ "loss": 4.8427300453186035,
+ "step": 3652
+ },
+ {
+ "epoch": 50.74061135371179,
+ "grad_norm": 0.01824103854596615,
+ "learning_rate": 0.0006,
+ "loss": 4.820980548858643,
+ "step": 3653
+ },
+ {
+ "epoch": 50.75458515283843,
+ "grad_norm": 0.018129967153072357,
+ "learning_rate": 0.0006,
+ "loss": 4.946732997894287,
+ "step": 3654
+ },
+ {
+ "epoch": 50.76855895196506,
+ "grad_norm": 0.01784815452992916,
+ "learning_rate": 0.0006,
+ "loss": 4.829516410827637,
+ "step": 3655
+ },
+ {
+ "epoch": 50.7825327510917,
+ "grad_norm": 0.020021555945277214,
+ "learning_rate": 0.0006,
+ "loss": 4.780428409576416,
+ "step": 3656
+ },
+ {
+ "epoch": 50.79650655021834,
+ "grad_norm": 0.01863253302872181,
+ "learning_rate": 0.0006,
+ "loss": 4.791567802429199,
+ "step": 3657
+ },
+ {
+ "epoch": 50.81048034934498,
+ "grad_norm": 0.019151683896780014,
+ "learning_rate": 0.0006,
+ "loss": 4.824892997741699,
+ "step": 3658
+ },
+ {
+ "epoch": 50.82445414847162,
+ "grad_norm": 0.01891648955643177,
+ "learning_rate": 0.0006,
+ "loss": 4.789605140686035,
+ "step": 3659
+ },
+ {
+ "epoch": 50.838427947598255,
+ "grad_norm": 0.016019510105252266,
+ "learning_rate": 0.0006,
+ "loss": 4.779555320739746,
+ "step": 3660
+ },
+ {
+ "epoch": 50.852401746724894,
+ "grad_norm": 0.016096491366624832,
+ "learning_rate": 0.0006,
+ "loss": 4.753842353820801,
+ "step": 3661
+ },
+ {
+ "epoch": 50.866375545851525,
+ "grad_norm": 0.01693476364016533,
+ "learning_rate": 0.0006,
+ "loss": 4.768477439880371,
+ "step": 3662
+ },
+ {
+ "epoch": 50.880349344978164,
+ "grad_norm": 0.01749575324356556,
+ "learning_rate": 0.0006,
+ "loss": 4.779143333435059,
+ "step": 3663
+ },
+ {
+ "epoch": 50.8943231441048,
+ "grad_norm": 0.016851043328642845,
+ "learning_rate": 0.0006,
+ "loss": 4.780387878417969,
+ "step": 3664
+ },
+ {
+ "epoch": 50.90829694323144,
+ "grad_norm": 0.019045202061533928,
+ "learning_rate": 0.0006,
+ "loss": 4.769913673400879,
+ "step": 3665
+ },
+ {
+ "epoch": 50.92227074235808,
+ "grad_norm": 0.01954123005270958,
+ "learning_rate": 0.0006,
+ "loss": 4.769009590148926,
+ "step": 3666
+ },
+ {
+ "epoch": 50.93624454148472,
+ "grad_norm": 0.022527096793055534,
+ "learning_rate": 0.0006,
+ "loss": 4.789215087890625,
+ "step": 3667
+ },
+ {
+ "epoch": 50.95021834061135,
+ "grad_norm": 0.02871634066104889,
+ "learning_rate": 0.0006,
+ "loss": 4.749727725982666,
+ "step": 3668
+ },
+ {
+ "epoch": 50.96419213973799,
+ "grad_norm": 0.04045053571462631,
+ "learning_rate": 0.0006,
+ "loss": 4.909038543701172,
+ "step": 3669
+ },
+ {
+ "epoch": 50.97816593886463,
+ "grad_norm": 0.04156879335641861,
+ "learning_rate": 0.0006,
+ "loss": 4.876831531524658,
+ "step": 3670
+ },
+ {
+ "epoch": 50.992139737991266,
+ "grad_norm": 0.02621397376060486,
+ "learning_rate": 0.0006,
+ "loss": 4.797173500061035,
+ "step": 3671
+ },
+ {
+ "epoch": 51.0,
+ "grad_norm": 0.03389597684144974,
+ "learning_rate": 0.0006,
+ "loss": 4.8428850173950195,
+ "step": 3672
+ },
+ {
+ "epoch": 51.0,
+ "eval_loss": 4.989261627197266,
+ "eval_runtime": 56.358,
+ "eval_samples_per_second": 43.33,
+ "eval_steps_per_second": 1.366,
+ "step": 3672
+ },
+ {
+ "epoch": 51.01397379912664,
+ "grad_norm": 0.03216972202062607,
+ "learning_rate": 0.0006,
+ "loss": 4.755102634429932,
+ "step": 3673
+ },
+ {
+ "epoch": 51.02794759825328,
+ "grad_norm": 0.03190374746918678,
+ "learning_rate": 0.0006,
+ "loss": 4.720424652099609,
+ "step": 3674
+ },
+ {
+ "epoch": 51.041921397379916,
+ "grad_norm": 0.031727299094200134,
+ "learning_rate": 0.0006,
+ "loss": 4.803277969360352,
+ "step": 3675
+ },
+ {
+ "epoch": 51.05589519650655,
+ "grad_norm": 0.02473079040646553,
+ "learning_rate": 0.0006,
+ "loss": 4.761087417602539,
+ "step": 3676
+ },
+ {
+ "epoch": 51.069868995633186,
+ "grad_norm": 0.020360209047794342,
+ "learning_rate": 0.0006,
+ "loss": 4.753884792327881,
+ "step": 3677
+ },
+ {
+ "epoch": 51.083842794759825,
+ "grad_norm": 0.02078923024237156,
+ "learning_rate": 0.0006,
+ "loss": 4.647130489349365,
+ "step": 3678
+ },
+ {
+ "epoch": 51.09781659388646,
+ "grad_norm": 0.019387200474739075,
+ "learning_rate": 0.0006,
+ "loss": 4.74135160446167,
+ "step": 3679
+ },
+ {
+ "epoch": 51.1117903930131,
+ "grad_norm": 0.01874949410557747,
+ "learning_rate": 0.0006,
+ "loss": 4.795329570770264,
+ "step": 3680
+ },
+ {
+ "epoch": 51.12576419213974,
+ "grad_norm": 0.019331837072968483,
+ "learning_rate": 0.0006,
+ "loss": 4.897967338562012,
+ "step": 3681
+ },
+ {
+ "epoch": 51.13973799126637,
+ "grad_norm": 0.017066050320863724,
+ "learning_rate": 0.0006,
+ "loss": 4.825351715087891,
+ "step": 3682
+ },
+ {
+ "epoch": 51.15371179039301,
+ "grad_norm": 0.01757810264825821,
+ "learning_rate": 0.0006,
+ "loss": 4.719335079193115,
+ "step": 3683
+ },
+ {
+ "epoch": 51.16768558951965,
+ "grad_norm": 0.018174266442656517,
+ "learning_rate": 0.0006,
+ "loss": 4.732680320739746,
+ "step": 3684
+ },
+ {
+ "epoch": 51.18165938864629,
+ "grad_norm": 0.018497727811336517,
+ "learning_rate": 0.0006,
+ "loss": 4.731050491333008,
+ "step": 3685
+ },
+ {
+ "epoch": 51.19563318777293,
+ "grad_norm": 0.0186338871717453,
+ "learning_rate": 0.0006,
+ "loss": 4.748035430908203,
+ "step": 3686
+ },
+ {
+ "epoch": 51.209606986899566,
+ "grad_norm": 0.018531711772084236,
+ "learning_rate": 0.0006,
+ "loss": 4.9007463455200195,
+ "step": 3687
+ },
+ {
+ "epoch": 51.223580786026204,
+ "grad_norm": 0.018221354112029076,
+ "learning_rate": 0.0006,
+ "loss": 4.84077262878418,
+ "step": 3688
+ },
+ {
+ "epoch": 51.237554585152836,
+ "grad_norm": 0.0203656405210495,
+ "learning_rate": 0.0006,
+ "loss": 4.73009729385376,
+ "step": 3689
+ },
+ {
+ "epoch": 51.251528384279474,
+ "grad_norm": 0.01964893378317356,
+ "learning_rate": 0.0006,
+ "loss": 4.772583961486816,
+ "step": 3690
+ },
+ {
+ "epoch": 51.26550218340611,
+ "grad_norm": 0.019606510177254677,
+ "learning_rate": 0.0006,
+ "loss": 4.599812030792236,
+ "step": 3691
+ },
+ {
+ "epoch": 51.27947598253275,
+ "grad_norm": 0.020501255989074707,
+ "learning_rate": 0.0006,
+ "loss": 4.809898853302002,
+ "step": 3692
+ },
+ {
+ "epoch": 51.29344978165939,
+ "grad_norm": 0.01870838552713394,
+ "learning_rate": 0.0006,
+ "loss": 4.741898059844971,
+ "step": 3693
+ },
+ {
+ "epoch": 51.30742358078603,
+ "grad_norm": 0.01802559196949005,
+ "learning_rate": 0.0006,
+ "loss": 4.746562480926514,
+ "step": 3694
+ },
+ {
+ "epoch": 51.32139737991266,
+ "grad_norm": 0.02053048275411129,
+ "learning_rate": 0.0006,
+ "loss": 4.69932746887207,
+ "step": 3695
+ },
+ {
+ "epoch": 51.3353711790393,
+ "grad_norm": 0.02661825157701969,
+ "learning_rate": 0.0006,
+ "loss": 4.68748664855957,
+ "step": 3696
+ },
+ {
+ "epoch": 51.34934497816594,
+ "grad_norm": 0.028269147500395775,
+ "learning_rate": 0.0006,
+ "loss": 4.731565475463867,
+ "step": 3697
+ },
+ {
+ "epoch": 51.36331877729258,
+ "grad_norm": 0.02650964818894863,
+ "learning_rate": 0.0006,
+ "loss": 4.733857154846191,
+ "step": 3698
+ },
+ {
+ "epoch": 51.377292576419215,
+ "grad_norm": 0.023696066811680794,
+ "learning_rate": 0.0006,
+ "loss": 4.812362194061279,
+ "step": 3699
+ },
+ {
+ "epoch": 51.391266375545854,
+ "grad_norm": 0.02365880273282528,
+ "learning_rate": 0.0006,
+ "loss": 4.733426094055176,
+ "step": 3700
+ },
+ {
+ "epoch": 51.40524017467249,
+ "grad_norm": 0.026249447837471962,
+ "learning_rate": 0.0006,
+ "loss": 4.715622901916504,
+ "step": 3701
+ },
+ {
+ "epoch": 51.419213973799124,
+ "grad_norm": 0.03301899880170822,
+ "learning_rate": 0.0006,
+ "loss": 4.7223663330078125,
+ "step": 3702
+ },
+ {
+ "epoch": 51.43318777292576,
+ "grad_norm": 0.044686660170555115,
+ "learning_rate": 0.0006,
+ "loss": 4.8121161460876465,
+ "step": 3703
+ },
+ {
+ "epoch": 51.4471615720524,
+ "grad_norm": 0.04733522981405258,
+ "learning_rate": 0.0006,
+ "loss": 4.841394424438477,
+ "step": 3704
+ },
+ {
+ "epoch": 51.46113537117904,
+ "grad_norm": 0.0395209975540638,
+ "learning_rate": 0.0006,
+ "loss": 4.773837089538574,
+ "step": 3705
+ },
+ {
+ "epoch": 51.47510917030568,
+ "grad_norm": 0.032790958881378174,
+ "learning_rate": 0.0006,
+ "loss": 4.833066463470459,
+ "step": 3706
+ },
+ {
+ "epoch": 51.48908296943232,
+ "grad_norm": 0.030578427016735077,
+ "learning_rate": 0.0006,
+ "loss": 4.685190200805664,
+ "step": 3707
+ },
+ {
+ "epoch": 51.50305676855895,
+ "grad_norm": 0.031619083136320114,
+ "learning_rate": 0.0006,
+ "loss": 4.740378379821777,
+ "step": 3708
+ },
+ {
+ "epoch": 51.51703056768559,
+ "grad_norm": 0.02385825850069523,
+ "learning_rate": 0.0006,
+ "loss": 4.812911510467529,
+ "step": 3709
+ },
+ {
+ "epoch": 51.531004366812226,
+ "grad_norm": 0.023523425683379173,
+ "learning_rate": 0.0006,
+ "loss": 4.669580459594727,
+ "step": 3710
+ },
+ {
+ "epoch": 51.544978165938865,
+ "grad_norm": 0.021316280588507652,
+ "learning_rate": 0.0006,
+ "loss": 4.76888370513916,
+ "step": 3711
+ },
+ {
+ "epoch": 51.5589519650655,
+ "grad_norm": 0.024663880467414856,
+ "learning_rate": 0.0006,
+ "loss": 4.702592372894287,
+ "step": 3712
+ },
+ {
+ "epoch": 51.57292576419214,
+ "grad_norm": 0.024173280224204063,
+ "learning_rate": 0.0006,
+ "loss": 4.710604667663574,
+ "step": 3713
+ },
+ {
+ "epoch": 51.58689956331878,
+ "grad_norm": 0.02754114381968975,
+ "learning_rate": 0.0006,
+ "loss": 4.746356964111328,
+ "step": 3714
+ },
+ {
+ "epoch": 51.60087336244541,
+ "grad_norm": 0.027903622016310692,
+ "learning_rate": 0.0006,
+ "loss": 4.7546491622924805,
+ "step": 3715
+ },
+ {
+ "epoch": 51.61484716157205,
+ "grad_norm": 0.024871617555618286,
+ "learning_rate": 0.0006,
+ "loss": 4.605746269226074,
+ "step": 3716
+ },
+ {
+ "epoch": 51.62882096069869,
+ "grad_norm": 0.021239031106233597,
+ "learning_rate": 0.0006,
+ "loss": 4.6942458152771,
+ "step": 3717
+ },
+ {
+ "epoch": 51.64279475982533,
+ "grad_norm": 0.01925622671842575,
+ "learning_rate": 0.0006,
+ "loss": 4.593558311462402,
+ "step": 3718
+ },
+ {
+ "epoch": 51.65676855895197,
+ "grad_norm": 0.01956283673644066,
+ "learning_rate": 0.0006,
+ "loss": 4.631172180175781,
+ "step": 3719
+ },
+ {
+ "epoch": 51.670742358078606,
+ "grad_norm": 0.02269871160387993,
+ "learning_rate": 0.0006,
+ "loss": 4.729236602783203,
+ "step": 3720
+ },
+ {
+ "epoch": 51.68471615720524,
+ "grad_norm": 0.02253916673362255,
+ "learning_rate": 0.0006,
+ "loss": 4.7191033363342285,
+ "step": 3721
+ },
+ {
+ "epoch": 51.698689956331876,
+ "grad_norm": 0.026231050491333008,
+ "learning_rate": 0.0006,
+ "loss": 4.811850547790527,
+ "step": 3722
+ },
+ {
+ "epoch": 51.712663755458514,
+ "grad_norm": 0.02558821812272072,
+ "learning_rate": 0.0006,
+ "loss": 4.681466102600098,
+ "step": 3723
+ },
+ {
+ "epoch": 51.72663755458515,
+ "grad_norm": 0.02124209702014923,
+ "learning_rate": 0.0006,
+ "loss": 4.693664073944092,
+ "step": 3724
+ },
+ {
+ "epoch": 51.74061135371179,
+ "grad_norm": 0.016638319939374924,
+ "learning_rate": 0.0006,
+ "loss": 4.771440029144287,
+ "step": 3725
+ },
+ {
+ "epoch": 51.75458515283843,
+ "grad_norm": 0.018048347905278206,
+ "learning_rate": 0.0006,
+ "loss": 4.643252372741699,
+ "step": 3726
+ },
+ {
+ "epoch": 51.76855895196506,
+ "grad_norm": 0.019904717803001404,
+ "learning_rate": 0.0006,
+ "loss": 4.66530704498291,
+ "step": 3727
+ },
+ {
+ "epoch": 51.7825327510917,
+ "grad_norm": 0.0194614976644516,
+ "learning_rate": 0.0006,
+ "loss": 4.711943626403809,
+ "step": 3728
+ },
+ {
+ "epoch": 51.79650655021834,
+ "grad_norm": 0.019175555557012558,
+ "learning_rate": 0.0006,
+ "loss": 4.721931457519531,
+ "step": 3729
+ },
+ {
+ "epoch": 51.81048034934498,
+ "grad_norm": 0.02093447372317314,
+ "learning_rate": 0.0006,
+ "loss": 4.633059024810791,
+ "step": 3730
+ },
+ {
+ "epoch": 51.82445414847162,
+ "grad_norm": 0.01930283196270466,
+ "learning_rate": 0.0006,
+ "loss": 4.659373760223389,
+ "step": 3731
+ },
+ {
+ "epoch": 51.838427947598255,
+ "grad_norm": 0.015350676141679287,
+ "learning_rate": 0.0006,
+ "loss": 4.763708114624023,
+ "step": 3732
+ },
+ {
+ "epoch": 51.852401746724894,
+ "grad_norm": 0.014139448292553425,
+ "learning_rate": 0.0006,
+ "loss": 4.753891468048096,
+ "step": 3733
+ },
+ {
+ "epoch": 51.866375545851525,
+ "grad_norm": 0.016215872019529343,
+ "learning_rate": 0.0006,
+ "loss": 4.658967018127441,
+ "step": 3734
+ },
+ {
+ "epoch": 51.880349344978164,
+ "grad_norm": 0.017328619956970215,
+ "learning_rate": 0.0006,
+ "loss": 4.696579933166504,
+ "step": 3735
+ },
+ {
+ "epoch": 51.8943231441048,
+ "grad_norm": 0.018103396520018578,
+ "learning_rate": 0.0006,
+ "loss": 4.627464771270752,
+ "step": 3736
+ },
+ {
+ "epoch": 51.90829694323144,
+ "grad_norm": 0.01715235412120819,
+ "learning_rate": 0.0006,
+ "loss": 4.7368011474609375,
+ "step": 3737
+ },
+ {
+ "epoch": 51.92227074235808,
+ "grad_norm": 0.014790094457566738,
+ "learning_rate": 0.0006,
+ "loss": 4.629441738128662,
+ "step": 3738
+ },
+ {
+ "epoch": 51.93624454148472,
+ "grad_norm": 0.017114296555519104,
+ "learning_rate": 0.0006,
+ "loss": 4.761256694793701,
+ "step": 3739
+ },
+ {
+ "epoch": 51.95021834061135,
+ "grad_norm": 0.019009845331311226,
+ "learning_rate": 0.0006,
+ "loss": 4.789155006408691,
+ "step": 3740
+ },
+ {
+ "epoch": 51.96419213973799,
+ "grad_norm": 0.019696692004799843,
+ "learning_rate": 0.0006,
+ "loss": 4.721269607543945,
+ "step": 3741
+ },
+ {
+ "epoch": 51.97816593886463,
+ "grad_norm": 0.020542822778224945,
+ "learning_rate": 0.0006,
+ "loss": 4.687038898468018,
+ "step": 3742
+ },
+ {
+ "epoch": 51.992139737991266,
+ "grad_norm": 0.021267160773277283,
+ "learning_rate": 0.0006,
+ "loss": 4.747860908508301,
+ "step": 3743
+ },
+ {
+ "epoch": 52.0,
+ "grad_norm": 0.025049567222595215,
+ "learning_rate": 0.0006,
+ "loss": 4.8473100662231445,
+ "step": 3744
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 28800,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 400,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": true,
+ "should_log": false,
+ "should_save": false,
+ "should_training_stop": false
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 1.5961989478927565e+18,
+ "train_batch_size": 8,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-3744/training_args.bin b/runs/l2r50-i2-fulle-lm/checkpoint-3744/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..cad232831c222cc7f8b938b8bbe42f61e7d89c3a
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-3744/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:f7c40e3bfa2f5ac1450ca1243e3f2a895d44a06ed5995e6aefd97f78d229d8c8
+size 4792
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4032/chat_template.jinja b/runs/l2r50-i2-fulle-lm/checkpoint-4032/chat_template.jinja
new file mode 100644
index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-4032/chat_template.jinja
@@ -0,0 +1,85 @@
+{%- if tools %}
+ {{- '<|im_start|>system\n' }}
+ {%- if messages[0].role == 'system' %}
+ {{- messages[0].content + '\n\n' }}
+ {%- endif %}
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }}
+ {%- for tool in tools %}
+ {{- "\n" }}
+ {{- tool | tojson }}
+ {%- endfor %}
+ {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }}
+{%- else %}
+ {%- if messages[0].role == 'system' %}
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
+ {%- endif %}
+{%- endif %}
+{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
+{%- for message in messages[::-1] %}
+ {%- set index = (messages|length - 1) - loop.index0 %}
+ {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %}
+ {%- set ns.multi_step_tool = false %}
+ {%- set ns.last_query_index = index %}
+ {%- endif %}
+{%- endfor %}
+{%- for message in messages %}
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
+ {%- elif message.role == "assistant" %}
+ {%- set content = message.content %}
+ {%- set reasoning_content = '' %}
+ {%- if message.reasoning_content is defined and message.reasoning_content is not none %}
+ {%- set reasoning_content = message.reasoning_content %}
+ {%- else %}
+ {%- if '' in message.content %}
+ {%- set content = message.content.split('')[-1].lstrip('\n') %}
+ {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %}
+ {%- endif %}
+ {%- endif %}
+ {%- if loop.index0 > ns.last_query_index %}
+ {%- if loop.last or (not loop.last and reasoning_content) %}
+ {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- if message.tool_calls %}
+ {%- for tool_call in message.tool_calls %}
+ {%- if (loop.first and content) or (not loop.first) %}
+ {{- '\n' }}
+ {%- endif %}
+ {%- if tool_call.function %}
+ {%- set tool_call = tool_call.function %}
+ {%- endif %}
+ {{- '\n{"name": "' }}
+ {{- tool_call.name }}
+ {{- '", "arguments": ' }}
+ {%- if tool_call.arguments is string %}
+ {{- tool_call.arguments }}
+ {%- else %}
+ {{- tool_call.arguments | tojson }}
+ {%- endif %}
+ {{- '}\n' }}
+ {%- endfor %}
+ {%- endif %}
+ {{- '<|im_end|>\n' }}
+ {%- elif message.role == "tool" %}
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
+ {{- '<|im_start|>user' }}
+ {%- endif %}
+ {{- '\n\n' }}
+ {{- message.content }}
+ {{- '\n' }}
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
+ {{- '<|im_end|>\n' }}
+ {%- endif %}
+ {%- endif %}
+{%- endfor %}
+{%- if add_generation_prompt %}
+ {{- '<|im_start|>assistant\n' }}
+ {%- if enable_thinking is defined and enable_thinking is false %}
+ {{- '\n\n\n\n' }}
+ {%- endif %}
+{%- endif %}
\ No newline at end of file
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4032/config.json b/runs/l2r50-i2-fulle-lm/checkpoint-4032/config.json
new file mode 100644
index 0000000000000000000000000000000000000000..22f250953c6ce8fc91ee4411856dd73041a415e2
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-4032/config.json
@@ -0,0 +1,32 @@
+{
+ "architectures": [
+ "LlamaForCausalLM"
+ ],
+ "attention_bias": false,
+ "attention_dropout": 0.0,
+ "bos_token_id": null,
+ "dtype": "float32",
+ "eos_token_id": 151645,
+ "head_dim": 128,
+ "hidden_act": "silu",
+ "hidden_size": 512,
+ "initializer_range": 0.02,
+ "intermediate_size": 1536,
+ "max_position_embeddings": 2048,
+ "mlp_bias": false,
+ "model_type": "llama",
+ "num_attention_heads": 4,
+ "num_hidden_layers": 20,
+ "num_key_value_heads": 4,
+ "pad_token_id": 151645,
+ "pretraining_tp": 1,
+ "rms_norm_eps": 1e-06,
+ "rope_parameters": {
+ "rope_theta": 10000.0,
+ "rope_type": "default"
+ },
+ "tie_word_embeddings": true,
+ "transformers_version": "5.5.0",
+ "use_cache": false,
+ "vocab_size": 151673
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4032/generation_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-4032/generation_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-4032/generation_config.json
@@ -0,0 +1,11 @@
+{
+ "_from_model_config": true,
+ "eos_token_id": [
+ 151645
+ ],
+ "output_attentions": false,
+ "output_hidden_states": false,
+ "pad_token_id": 151645,
+ "transformers_version": "5.5.0",
+ "use_cache": true
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4032/model.safetensors b/runs/l2r50-i2-fulle-lm/checkpoint-4032/model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..e100d59b7aaacd71434636a1c89885f21cc5fa50
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-4032/model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:412f049b6362640e7d8533ad087f506ac4b0493f32fa49f321584598b50f890d
+size 583360328
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4032/optimizer.pt b/runs/l2r50-i2-fulle-lm/checkpoint-4032/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..e80be47e4f07320bb653502a98217c84235dd68c
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-4032/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:93e3cc6b6dfeb78481954b5f343786b73d0665da20dfebd9ecc6a71153e60439
+size 1166833530
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4032/rng_state_0.pth b/runs/l2r50-i2-fulle-lm/checkpoint-4032/rng_state_0.pth
new file mode 100644
index 0000000000000000000000000000000000000000..ad303665e91bd8f463c96f8ed5c5df78f614a9fa
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-4032/rng_state_0.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:aa32010f36bc8768606dd76bbf2cc108c36257b13d01a3f21156b47945e64631
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4032/rng_state_1.pth b/runs/l2r50-i2-fulle-lm/checkpoint-4032/rng_state_1.pth
new file mode 100644
index 0000000000000000000000000000000000000000..4bac4005ebb4b7f2f74832f578c276bcb71f334a
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-4032/rng_state_1.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:a78e70ff4356ac6656e8073aba4a1df6fa153f4e3114fafaf49b3cc0a03475bf
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4032/rng_state_2.pth b/runs/l2r50-i2-fulle-lm/checkpoint-4032/rng_state_2.pth
new file mode 100644
index 0000000000000000000000000000000000000000..0db8e27e712c3dd6632afa8e2d230a3da16e1364
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-4032/rng_state_2.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:ff498817004558cbfa2b8135a52cbe67932c55151d0c17d095105acdf1951053
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4032/rng_state_3.pth b/runs/l2r50-i2-fulle-lm/checkpoint-4032/rng_state_3.pth
new file mode 100644
index 0000000000000000000000000000000000000000..951c60de2fb5f641f5a18648f8761a3af42de82f
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-4032/rng_state_3.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:863bc081e419560c57a86100786cd6de3749e7265ca91548d5cd0f11e6d052b5
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4032/scheduler.pt b/runs/l2r50-i2-fulle-lm/checkpoint-4032/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..5da96a3852815ab990f23f4b91093acc83e76469
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-4032/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:9dd17660f6a4c270971d4c998ea51235588dc2ab05504036a1351d00b0a57d2c
+size 1064
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4032/tokenizer.json b/runs/l2r50-i2-fulle-lm/checkpoint-4032/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..53786016f070e11184813f3c83101dc5e93e392e
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-4032/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:c3c63fda1646a555448e8f03662542e04b11d103721319e4d9b267d7da03ceb8
+size 11424463
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4032/tokenizer_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-4032/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..98e2fb9b2cc54785dc9d871b8e503f8b8f2defe7
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-4032/tokenizer_config.json
@@ -0,0 +1,20 @@
+{
+ "add_prefix_space": false,
+ "backend": "tokenizers",
+ "bos_token": null,
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "<|im_end|>",
+ "errors": "replace",
+ "extra_special_tokens": [
+ "<|l2r_pred|>",
+ "<|r2l_pred|>",
+ "<|next_1_pred|>",
+ "<|next_2_pred|>"
+ ],
+ "is_local": false,
+ "model_max_length": 131072,
+ "pad_token": "<|im_end|>",
+ "split_special_tokens": false,
+ "tokenizer_class": "Qwen2Tokenizer",
+ "unk_token": null
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4032/trainer_state.json b/runs/l2r50-i2-fulle-lm/checkpoint-4032/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..dc148a0e461de1f5c1ee5b27a570890ef7358abe
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-4032/trainer_state.json
@@ -0,0 +1,28698 @@
+{
+ "best_global_step": null,
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 56.0,
+ "eval_steps": 500,
+ "global_step": 4032,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.013973799126637555,
+ "grad_norm": 0.1739506870508194,
+ "learning_rate": 0.0,
+ "loss": 11.99299430847168,
+ "step": 1
+ },
+ {
+ "epoch": 0.02794759825327511,
+ "grad_norm": 0.17216597497463226,
+ "learning_rate": 5.999999999999999e-06,
+ "loss": 11.988704681396484,
+ "step": 2
+ },
+ {
+ "epoch": 0.04192139737991266,
+ "grad_norm": 0.17459715902805328,
+ "learning_rate": 1.1999999999999999e-05,
+ "loss": 11.945581436157227,
+ "step": 3
+ },
+ {
+ "epoch": 0.05589519650655022,
+ "grad_norm": 0.17577414214611053,
+ "learning_rate": 1.7999999999999997e-05,
+ "loss": 11.859772682189941,
+ "step": 4
+ },
+ {
+ "epoch": 0.06986899563318777,
+ "grad_norm": 0.16163651645183563,
+ "learning_rate": 2.3999999999999997e-05,
+ "loss": 11.744415283203125,
+ "step": 5
+ },
+ {
+ "epoch": 0.08384279475982533,
+ "grad_norm": 0.14409932494163513,
+ "learning_rate": 2.9999999999999997e-05,
+ "loss": 11.618099212646484,
+ "step": 6
+ },
+ {
+ "epoch": 0.09781659388646288,
+ "grad_norm": 0.13079848885536194,
+ "learning_rate": 3.5999999999999994e-05,
+ "loss": 11.505935668945312,
+ "step": 7
+ },
+ {
+ "epoch": 0.11179039301310044,
+ "grad_norm": 0.11964613944292068,
+ "learning_rate": 4.2e-05,
+ "loss": 11.416532516479492,
+ "step": 8
+ },
+ {
+ "epoch": 0.125764192139738,
+ "grad_norm": 0.11345142871141434,
+ "learning_rate": 4.7999999999999994e-05,
+ "loss": 11.33206558227539,
+ "step": 9
+ },
+ {
+ "epoch": 0.13973799126637554,
+ "grad_norm": 0.11033952981233597,
+ "learning_rate": 5.399999999999999e-05,
+ "loss": 11.259052276611328,
+ "step": 10
+ },
+ {
+ "epoch": 0.1537117903930131,
+ "grad_norm": 0.10776489973068237,
+ "learning_rate": 5.9999999999999995e-05,
+ "loss": 11.203384399414062,
+ "step": 11
+ },
+ {
+ "epoch": 0.16768558951965065,
+ "grad_norm": 0.10721946507692337,
+ "learning_rate": 6.599999999999999e-05,
+ "loss": 11.153295516967773,
+ "step": 12
+ },
+ {
+ "epoch": 0.18165938864628822,
+ "grad_norm": 0.10673430562019348,
+ "learning_rate": 7.199999999999999e-05,
+ "loss": 11.110200881958008,
+ "step": 13
+ },
+ {
+ "epoch": 0.19563318777292577,
+ "grad_norm": 0.10630254447460175,
+ "learning_rate": 7.8e-05,
+ "loss": 11.067477226257324,
+ "step": 14
+ },
+ {
+ "epoch": 0.2096069868995633,
+ "grad_norm": 0.10631615668535233,
+ "learning_rate": 8.4e-05,
+ "loss": 11.022941589355469,
+ "step": 15
+ },
+ {
+ "epoch": 0.22358078602620088,
+ "grad_norm": 0.10609443485736847,
+ "learning_rate": 8.999999999999999e-05,
+ "loss": 10.970746040344238,
+ "step": 16
+ },
+ {
+ "epoch": 0.23755458515283842,
+ "grad_norm": 0.10648232698440552,
+ "learning_rate": 9.599999999999999e-05,
+ "loss": 10.91267204284668,
+ "step": 17
+ },
+ {
+ "epoch": 0.251528384279476,
+ "grad_norm": 0.10584589838981628,
+ "learning_rate": 0.000102,
+ "loss": 10.854547500610352,
+ "step": 18
+ },
+ {
+ "epoch": 0.26550218340611353,
+ "grad_norm": 0.10591720044612885,
+ "learning_rate": 0.00010799999999999998,
+ "loss": 10.788917541503906,
+ "step": 19
+ },
+ {
+ "epoch": 0.2794759825327511,
+ "grad_norm": 0.10611672699451447,
+ "learning_rate": 0.00011399999999999999,
+ "loss": 10.717556953430176,
+ "step": 20
+ },
+ {
+ "epoch": 0.2934497816593886,
+ "grad_norm": 0.10455359518527985,
+ "learning_rate": 0.00011999999999999999,
+ "loss": 10.650186538696289,
+ "step": 21
+ },
+ {
+ "epoch": 0.3074235807860262,
+ "grad_norm": 0.1047525629401207,
+ "learning_rate": 0.00012599999999999997,
+ "loss": 10.56808090209961,
+ "step": 22
+ },
+ {
+ "epoch": 0.32139737991266376,
+ "grad_norm": 0.10511605441570282,
+ "learning_rate": 0.00013199999999999998,
+ "loss": 10.484028816223145,
+ "step": 23
+ },
+ {
+ "epoch": 0.3353711790393013,
+ "grad_norm": 0.10467177629470825,
+ "learning_rate": 0.000138,
+ "loss": 10.399947166442871,
+ "step": 24
+ },
+ {
+ "epoch": 0.34934497816593885,
+ "grad_norm": 0.10432136058807373,
+ "learning_rate": 0.00014399999999999998,
+ "loss": 10.312714576721191,
+ "step": 25
+ },
+ {
+ "epoch": 0.36331877729257644,
+ "grad_norm": 0.10527262091636658,
+ "learning_rate": 0.00015,
+ "loss": 10.212261199951172,
+ "step": 26
+ },
+ {
+ "epoch": 0.377292576419214,
+ "grad_norm": 0.103610560297966,
+ "learning_rate": 0.000156,
+ "loss": 10.132034301757812,
+ "step": 27
+ },
+ {
+ "epoch": 0.39126637554585153,
+ "grad_norm": 0.10553059726953506,
+ "learning_rate": 0.000162,
+ "loss": 10.020434379577637,
+ "step": 28
+ },
+ {
+ "epoch": 0.4052401746724891,
+ "grad_norm": 0.1037299782037735,
+ "learning_rate": 0.000168,
+ "loss": 9.934466361999512,
+ "step": 29
+ },
+ {
+ "epoch": 0.4192139737991266,
+ "grad_norm": 0.1042739674448967,
+ "learning_rate": 0.00017399999999999997,
+ "loss": 9.822539329528809,
+ "step": 30
+ },
+ {
+ "epoch": 0.4331877729257642,
+ "grad_norm": 0.10300865024328232,
+ "learning_rate": 0.00017999999999999998,
+ "loss": 9.734994888305664,
+ "step": 31
+ },
+ {
+ "epoch": 0.44716157205240176,
+ "grad_norm": 0.10291654616594315,
+ "learning_rate": 0.000186,
+ "loss": 9.624505996704102,
+ "step": 32
+ },
+ {
+ "epoch": 0.4611353711790393,
+ "grad_norm": 0.10210851579904556,
+ "learning_rate": 0.00019199999999999998,
+ "loss": 9.518205642700195,
+ "step": 33
+ },
+ {
+ "epoch": 0.47510917030567684,
+ "grad_norm": 0.10222529619932175,
+ "learning_rate": 0.000198,
+ "loss": 9.412222862243652,
+ "step": 34
+ },
+ {
+ "epoch": 0.4890829694323144,
+ "grad_norm": 0.09992019087076187,
+ "learning_rate": 0.000204,
+ "loss": 9.326566696166992,
+ "step": 35
+ },
+ {
+ "epoch": 0.503056768558952,
+ "grad_norm": 0.09931084513664246,
+ "learning_rate": 0.00020999999999999998,
+ "loss": 9.210111618041992,
+ "step": 36
+ },
+ {
+ "epoch": 0.5170305676855895,
+ "grad_norm": 0.09928256273269653,
+ "learning_rate": 0.00021599999999999996,
+ "loss": 9.100493431091309,
+ "step": 37
+ },
+ {
+ "epoch": 0.5310043668122271,
+ "grad_norm": 0.09843816608190536,
+ "learning_rate": 0.00022199999999999998,
+ "loss": 8.996543884277344,
+ "step": 38
+ },
+ {
+ "epoch": 0.5449781659388646,
+ "grad_norm": 0.09825374186038971,
+ "learning_rate": 0.00022799999999999999,
+ "loss": 8.888991355895996,
+ "step": 39
+ },
+ {
+ "epoch": 0.5589519650655022,
+ "grad_norm": 0.09602385759353638,
+ "learning_rate": 0.000234,
+ "loss": 8.802907943725586,
+ "step": 40
+ },
+ {
+ "epoch": 0.5729257641921397,
+ "grad_norm": 0.09414064139127731,
+ "learning_rate": 0.00023999999999999998,
+ "loss": 8.715423583984375,
+ "step": 41
+ },
+ {
+ "epoch": 0.5868995633187772,
+ "grad_norm": 0.09273893386125565,
+ "learning_rate": 0.00024599999999999996,
+ "loss": 8.61374282836914,
+ "step": 42
+ },
+ {
+ "epoch": 0.6008733624454149,
+ "grad_norm": 0.09079599380493164,
+ "learning_rate": 0.00025199999999999995,
+ "loss": 8.502904891967773,
+ "step": 43
+ },
+ {
+ "epoch": 0.6148471615720524,
+ "grad_norm": 0.08605825901031494,
+ "learning_rate": 0.000258,
+ "loss": 8.453008651733398,
+ "step": 44
+ },
+ {
+ "epoch": 0.62882096069869,
+ "grad_norm": 0.08338996022939682,
+ "learning_rate": 0.00026399999999999997,
+ "loss": 8.365163803100586,
+ "step": 45
+ },
+ {
+ "epoch": 0.6427947598253275,
+ "grad_norm": 0.08150867372751236,
+ "learning_rate": 0.00027,
+ "loss": 8.285372734069824,
+ "step": 46
+ },
+ {
+ "epoch": 0.6567685589519651,
+ "grad_norm": 0.07794807851314545,
+ "learning_rate": 0.000276,
+ "loss": 8.185811996459961,
+ "step": 47
+ },
+ {
+ "epoch": 0.6707423580786026,
+ "grad_norm": 0.07592720538377762,
+ "learning_rate": 0.00028199999999999997,
+ "loss": 8.098482131958008,
+ "step": 48
+ },
+ {
+ "epoch": 0.6847161572052402,
+ "grad_norm": 0.0709897130727768,
+ "learning_rate": 0.00028799999999999995,
+ "loss": 8.042091369628906,
+ "step": 49
+ },
+ {
+ "epoch": 0.6986899563318777,
+ "grad_norm": 0.06582305580377579,
+ "learning_rate": 0.000294,
+ "loss": 7.972844123840332,
+ "step": 50
+ },
+ {
+ "epoch": 0.7126637554585152,
+ "grad_norm": 0.059452351182699203,
+ "learning_rate": 0.0003,
+ "loss": 7.924579620361328,
+ "step": 51
+ },
+ {
+ "epoch": 0.7266375545851529,
+ "grad_norm": 0.052426718175411224,
+ "learning_rate": 0.00030599999999999996,
+ "loss": 7.876336097717285,
+ "step": 52
+ },
+ {
+ "epoch": 0.7406113537117904,
+ "grad_norm": 0.04768150672316551,
+ "learning_rate": 0.000312,
+ "loss": 7.8102707862854,
+ "step": 53
+ },
+ {
+ "epoch": 0.754585152838428,
+ "grad_norm": 0.03970075398683548,
+ "learning_rate": 0.000318,
+ "loss": 7.791913986206055,
+ "step": 54
+ },
+ {
+ "epoch": 0.7685589519650655,
+ "grad_norm": 0.03559752553701401,
+ "learning_rate": 0.000324,
+ "loss": 7.756960868835449,
+ "step": 55
+ },
+ {
+ "epoch": 0.7825327510917031,
+ "grad_norm": 0.029321353882551193,
+ "learning_rate": 0.00033,
+ "loss": 7.724878787994385,
+ "step": 56
+ },
+ {
+ "epoch": 0.7965065502183406,
+ "grad_norm": 0.023485427722334862,
+ "learning_rate": 0.000336,
+ "loss": 7.72553014755249,
+ "step": 57
+ },
+ {
+ "epoch": 0.8104803493449781,
+ "grad_norm": 0.019319474697113037,
+ "learning_rate": 0.00034199999999999996,
+ "loss": 7.709493160247803,
+ "step": 58
+ },
+ {
+ "epoch": 0.8244541484716157,
+ "grad_norm": 0.016705691814422607,
+ "learning_rate": 0.00034799999999999995,
+ "loss": 7.6873297691345215,
+ "step": 59
+ },
+ {
+ "epoch": 0.8384279475982532,
+ "grad_norm": 0.017667734995484352,
+ "learning_rate": 0.00035399999999999993,
+ "loss": 7.677059173583984,
+ "step": 60
+ },
+ {
+ "epoch": 0.8524017467248908,
+ "grad_norm": 0.01671992614865303,
+ "learning_rate": 0.00035999999999999997,
+ "loss": 7.643512725830078,
+ "step": 61
+ },
+ {
+ "epoch": 0.8663755458515284,
+ "grad_norm": 0.012288172729313374,
+ "learning_rate": 0.00036599999999999995,
+ "loss": 7.658343315124512,
+ "step": 62
+ },
+ {
+ "epoch": 0.880349344978166,
+ "grad_norm": 0.012004776857793331,
+ "learning_rate": 0.000372,
+ "loss": 7.649759292602539,
+ "step": 63
+ },
+ {
+ "epoch": 0.8943231441048035,
+ "grad_norm": 0.012487445957958698,
+ "learning_rate": 0.00037799999999999997,
+ "loss": 7.631503105163574,
+ "step": 64
+ },
+ {
+ "epoch": 0.9082969432314411,
+ "grad_norm": 0.01489241048693657,
+ "learning_rate": 0.00038399999999999996,
+ "loss": 7.647110462188721,
+ "step": 65
+ },
+ {
+ "epoch": 0.9222707423580786,
+ "grad_norm": 0.01428980939090252,
+ "learning_rate": 0.00039,
+ "loss": 7.636983871459961,
+ "step": 66
+ },
+ {
+ "epoch": 0.9362445414847161,
+ "grad_norm": 0.010711144655942917,
+ "learning_rate": 0.000396,
+ "loss": 7.621130466461182,
+ "step": 67
+ },
+ {
+ "epoch": 0.9502183406113537,
+ "grad_norm": 0.010178886353969574,
+ "learning_rate": 0.000402,
+ "loss": 7.593924522399902,
+ "step": 68
+ },
+ {
+ "epoch": 0.9641921397379912,
+ "grad_norm": 0.010604874230921268,
+ "learning_rate": 0.000408,
+ "loss": 7.6015496253967285,
+ "step": 69
+ },
+ {
+ "epoch": 0.9781659388646288,
+ "grad_norm": 0.009751426056027412,
+ "learning_rate": 0.0004139999999999999,
+ "loss": 7.571964263916016,
+ "step": 70
+ },
+ {
+ "epoch": 0.9921397379912664,
+ "grad_norm": 0.009816481731832027,
+ "learning_rate": 0.00041999999999999996,
+ "loss": 7.57344913482666,
+ "step": 71
+ },
+ {
+ "epoch": 1.0,
+ "grad_norm": 0.008665637113153934,
+ "learning_rate": 0.00042599999999999995,
+ "loss": 7.579050064086914,
+ "step": 72
+ },
+ {
+ "epoch": 1.0,
+ "eval_loss": 7.581204891204834,
+ "eval_runtime": 59.0091,
+ "eval_samples_per_second": 41.383,
+ "eval_steps_per_second": 1.305,
+ "step": 72
+ },
+ {
+ "epoch": 1.0139737991266375,
+ "grad_norm": 0.009684201329946518,
+ "learning_rate": 0.00043199999999999993,
+ "loss": 7.568594932556152,
+ "step": 73
+ },
+ {
+ "epoch": 1.027947598253275,
+ "grad_norm": 0.025783710181713104,
+ "learning_rate": 0.00043799999999999997,
+ "loss": 7.578948497772217,
+ "step": 74
+ },
+ {
+ "epoch": 1.0419213973799126,
+ "grad_norm": 0.0327819399535656,
+ "learning_rate": 0.00044399999999999995,
+ "loss": 7.548173904418945,
+ "step": 75
+ },
+ {
+ "epoch": 1.0558951965065502,
+ "grad_norm": 0.021604498848319054,
+ "learning_rate": 0.00045,
+ "loss": 7.552953720092773,
+ "step": 76
+ },
+ {
+ "epoch": 1.0698689956331877,
+ "grad_norm": 0.010230105370283127,
+ "learning_rate": 0.00045599999999999997,
+ "loss": 7.5213212966918945,
+ "step": 77
+ },
+ {
+ "epoch": 1.0838427947598253,
+ "grad_norm": 0.023315589874982834,
+ "learning_rate": 0.00046199999999999995,
+ "loss": 7.501640319824219,
+ "step": 78
+ },
+ {
+ "epoch": 1.0978165938864628,
+ "grad_norm": 0.01109944935888052,
+ "learning_rate": 0.000468,
+ "loss": 7.539519786834717,
+ "step": 79
+ },
+ {
+ "epoch": 1.1117903930131003,
+ "grad_norm": 0.026093758642673492,
+ "learning_rate": 0.000474,
+ "loss": 7.525928974151611,
+ "step": 80
+ },
+ {
+ "epoch": 1.125764192139738,
+ "grad_norm": 0.03274895250797272,
+ "learning_rate": 0.00047999999999999996,
+ "loss": 7.535602569580078,
+ "step": 81
+ },
+ {
+ "epoch": 1.1397379912663754,
+ "grad_norm": 0.01543602254241705,
+ "learning_rate": 0.000486,
+ "loss": 7.513683795928955,
+ "step": 82
+ },
+ {
+ "epoch": 1.1537117903930132,
+ "grad_norm": 0.028377249836921692,
+ "learning_rate": 0.0004919999999999999,
+ "loss": 7.49384880065918,
+ "step": 83
+ },
+ {
+ "epoch": 1.1676855895196507,
+ "grad_norm": 0.04121481627225876,
+ "learning_rate": 0.000498,
+ "loss": 7.488447189331055,
+ "step": 84
+ },
+ {
+ "epoch": 1.1816593886462883,
+ "grad_norm": 0.013282302767038345,
+ "learning_rate": 0.0005039999999999999,
+ "loss": 7.463010787963867,
+ "step": 85
+ },
+ {
+ "epoch": 1.1956331877729258,
+ "grad_norm": 0.03200364485383034,
+ "learning_rate": 0.0005099999999999999,
+ "loss": 7.473515033721924,
+ "step": 86
+ },
+ {
+ "epoch": 1.2096069868995634,
+ "grad_norm": 0.020768029615283012,
+ "learning_rate": 0.000516,
+ "loss": 7.464688777923584,
+ "step": 87
+ },
+ {
+ "epoch": 1.223580786026201,
+ "grad_norm": 0.021715788170695305,
+ "learning_rate": 0.000522,
+ "loss": 7.4835028648376465,
+ "step": 88
+ },
+ {
+ "epoch": 1.2375545851528384,
+ "grad_norm": 0.026255754753947258,
+ "learning_rate": 0.0005279999999999999,
+ "loss": 7.443392753601074,
+ "step": 89
+ },
+ {
+ "epoch": 1.251528384279476,
+ "grad_norm": 0.022325951606035233,
+ "learning_rate": 0.000534,
+ "loss": 7.457373142242432,
+ "step": 90
+ },
+ {
+ "epoch": 1.2655021834061135,
+ "grad_norm": 0.009270387701690197,
+ "learning_rate": 0.00054,
+ "loss": 7.455986976623535,
+ "step": 91
+ },
+ {
+ "epoch": 1.279475982532751,
+ "grad_norm": 0.02369544468820095,
+ "learning_rate": 0.0005459999999999999,
+ "loss": 7.442216873168945,
+ "step": 92
+ },
+ {
+ "epoch": 1.2934497816593886,
+ "grad_norm": 0.008474506437778473,
+ "learning_rate": 0.000552,
+ "loss": 7.4294915199279785,
+ "step": 93
+ },
+ {
+ "epoch": 1.3074235807860262,
+ "grad_norm": 0.017043249681591988,
+ "learning_rate": 0.000558,
+ "loss": 7.406932830810547,
+ "step": 94
+ },
+ {
+ "epoch": 1.3213973799126637,
+ "grad_norm": 0.011813412420451641,
+ "learning_rate": 0.0005639999999999999,
+ "loss": 7.415900230407715,
+ "step": 95
+ },
+ {
+ "epoch": 1.3353711790393012,
+ "grad_norm": 0.01086883433163166,
+ "learning_rate": 0.00057,
+ "loss": 7.452998638153076,
+ "step": 96
+ },
+ {
+ "epoch": 1.3493449781659388,
+ "grad_norm": 0.014700951054692268,
+ "learning_rate": 0.0005759999999999999,
+ "loss": 7.410249710083008,
+ "step": 97
+ },
+ {
+ "epoch": 1.3633187772925766,
+ "grad_norm": 0.014512976631522179,
+ "learning_rate": 0.0005819999999999999,
+ "loss": 7.429132461547852,
+ "step": 98
+ },
+ {
+ "epoch": 1.3772925764192139,
+ "grad_norm": 0.012519920244812965,
+ "learning_rate": 0.000588,
+ "loss": 7.430817604064941,
+ "step": 99
+ },
+ {
+ "epoch": 1.3912663755458516,
+ "grad_norm": 0.007132664322853088,
+ "learning_rate": 0.0005939999999999999,
+ "loss": 7.405234336853027,
+ "step": 100
+ },
+ {
+ "epoch": 1.405240174672489,
+ "grad_norm": 0.019489388912916183,
+ "learning_rate": 0.0006,
+ "loss": 7.427791595458984,
+ "step": 101
+ },
+ {
+ "epoch": 1.4192139737991267,
+ "grad_norm": 0.027252664789557457,
+ "learning_rate": 0.0006,
+ "loss": 7.41761589050293,
+ "step": 102
+ },
+ {
+ "epoch": 1.4331877729257643,
+ "grad_norm": 0.02501637488603592,
+ "learning_rate": 0.0006,
+ "loss": 7.416735649108887,
+ "step": 103
+ },
+ {
+ "epoch": 1.4471615720524018,
+ "grad_norm": 0.00989772193133831,
+ "learning_rate": 0.0006,
+ "loss": 7.3756632804870605,
+ "step": 104
+ },
+ {
+ "epoch": 1.4611353711790394,
+ "grad_norm": 0.01622949168086052,
+ "learning_rate": 0.0006,
+ "loss": 7.400811672210693,
+ "step": 105
+ },
+ {
+ "epoch": 1.475109170305677,
+ "grad_norm": 0.012407763861119747,
+ "learning_rate": 0.0006,
+ "loss": 7.37675666809082,
+ "step": 106
+ },
+ {
+ "epoch": 1.4890829694323144,
+ "grad_norm": 0.010353565216064453,
+ "learning_rate": 0.0006,
+ "loss": 7.4013142585754395,
+ "step": 107
+ },
+ {
+ "epoch": 1.503056768558952,
+ "grad_norm": 0.0120553532615304,
+ "learning_rate": 0.0006,
+ "loss": 7.379873275756836,
+ "step": 108
+ },
+ {
+ "epoch": 1.5170305676855895,
+ "grad_norm": 0.01610225811600685,
+ "learning_rate": 0.0006,
+ "loss": 7.371631145477295,
+ "step": 109
+ },
+ {
+ "epoch": 1.531004366812227,
+ "grad_norm": 0.01159879844635725,
+ "learning_rate": 0.0006,
+ "loss": 7.389007568359375,
+ "step": 110
+ },
+ {
+ "epoch": 1.5449781659388646,
+ "grad_norm": 0.008013816550374031,
+ "learning_rate": 0.0006,
+ "loss": 7.395920753479004,
+ "step": 111
+ },
+ {
+ "epoch": 1.5589519650655022,
+ "grad_norm": 0.016334809362888336,
+ "learning_rate": 0.0006,
+ "loss": 7.393121719360352,
+ "step": 112
+ },
+ {
+ "epoch": 1.5729257641921397,
+ "grad_norm": 0.018821721896529198,
+ "learning_rate": 0.0006,
+ "loss": 7.367543697357178,
+ "step": 113
+ },
+ {
+ "epoch": 1.5868995633187772,
+ "grad_norm": 0.014088994823396206,
+ "learning_rate": 0.0006,
+ "loss": 7.385537624359131,
+ "step": 114
+ },
+ {
+ "epoch": 1.600873362445415,
+ "grad_norm": 0.005864050704985857,
+ "learning_rate": 0.0006,
+ "loss": 7.365346908569336,
+ "step": 115
+ },
+ {
+ "epoch": 1.6148471615720523,
+ "grad_norm": 0.011745430529117584,
+ "learning_rate": 0.0006,
+ "loss": 7.390227317810059,
+ "step": 116
+ },
+ {
+ "epoch": 1.62882096069869,
+ "grad_norm": 0.017205573618412018,
+ "learning_rate": 0.0006,
+ "loss": 7.3558549880981445,
+ "step": 117
+ },
+ {
+ "epoch": 1.6427947598253274,
+ "grad_norm": 0.021256346255540848,
+ "learning_rate": 0.0006,
+ "loss": 7.377346992492676,
+ "step": 118
+ },
+ {
+ "epoch": 1.6567685589519652,
+ "grad_norm": 0.019335204735398293,
+ "learning_rate": 0.0006,
+ "loss": 7.346194267272949,
+ "step": 119
+ },
+ {
+ "epoch": 1.6707423580786025,
+ "grad_norm": 0.008134003728628159,
+ "learning_rate": 0.0006,
+ "loss": 7.3063578605651855,
+ "step": 120
+ },
+ {
+ "epoch": 1.6847161572052403,
+ "grad_norm": 0.017613831907510757,
+ "learning_rate": 0.0006,
+ "loss": 7.338171005249023,
+ "step": 121
+ },
+ {
+ "epoch": 1.6986899563318776,
+ "grad_norm": 0.00890734139829874,
+ "learning_rate": 0.0006,
+ "loss": 7.355384826660156,
+ "step": 122
+ },
+ {
+ "epoch": 1.7126637554585153,
+ "grad_norm": 0.013448155485093594,
+ "learning_rate": 0.0006,
+ "loss": 7.344296455383301,
+ "step": 123
+ },
+ {
+ "epoch": 1.726637554585153,
+ "grad_norm": 0.009241633117198944,
+ "learning_rate": 0.0006,
+ "loss": 7.3454060554504395,
+ "step": 124
+ },
+ {
+ "epoch": 1.7406113537117904,
+ "grad_norm": 0.018561311066150665,
+ "learning_rate": 0.0006,
+ "loss": 7.32348108291626,
+ "step": 125
+ },
+ {
+ "epoch": 1.754585152838428,
+ "grad_norm": 0.025667704641819,
+ "learning_rate": 0.0006,
+ "loss": 7.356142520904541,
+ "step": 126
+ },
+ {
+ "epoch": 1.7685589519650655,
+ "grad_norm": 0.036232151091098785,
+ "learning_rate": 0.0006,
+ "loss": 7.363539695739746,
+ "step": 127
+ },
+ {
+ "epoch": 1.782532751091703,
+ "grad_norm": 0.059483714401721954,
+ "learning_rate": 0.0006,
+ "loss": 7.402331352233887,
+ "step": 128
+ },
+ {
+ "epoch": 1.7965065502183406,
+ "grad_norm": 0.05365009233355522,
+ "learning_rate": 0.0006,
+ "loss": 7.40457820892334,
+ "step": 129
+ },
+ {
+ "epoch": 1.8104803493449781,
+ "grad_norm": 0.02738415263593197,
+ "learning_rate": 0.0006,
+ "loss": 7.3756208419799805,
+ "step": 130
+ },
+ {
+ "epoch": 1.8244541484716157,
+ "grad_norm": 0.035017091780900955,
+ "learning_rate": 0.0006,
+ "loss": 7.379263877868652,
+ "step": 131
+ },
+ {
+ "epoch": 1.8384279475982532,
+ "grad_norm": 0.02819863148033619,
+ "learning_rate": 0.0006,
+ "loss": 7.327673435211182,
+ "step": 132
+ },
+ {
+ "epoch": 1.8524017467248908,
+ "grad_norm": 0.017754238098859787,
+ "learning_rate": 0.0006,
+ "loss": 7.333420276641846,
+ "step": 133
+ },
+ {
+ "epoch": 1.8663755458515285,
+ "grad_norm": 0.024685733020305634,
+ "learning_rate": 0.0006,
+ "loss": 7.34493350982666,
+ "step": 134
+ },
+ {
+ "epoch": 1.8803493449781659,
+ "grad_norm": 0.019426532089710236,
+ "learning_rate": 0.0006,
+ "loss": 7.328448295593262,
+ "step": 135
+ },
+ {
+ "epoch": 1.8943231441048036,
+ "grad_norm": 0.02106391079723835,
+ "learning_rate": 0.0006,
+ "loss": 7.327917098999023,
+ "step": 136
+ },
+ {
+ "epoch": 1.908296943231441,
+ "grad_norm": 0.02056879736483097,
+ "learning_rate": 0.0006,
+ "loss": 7.3311991691589355,
+ "step": 137
+ },
+ {
+ "epoch": 1.9222707423580787,
+ "grad_norm": 0.016865141689777374,
+ "learning_rate": 0.0006,
+ "loss": 7.312766075134277,
+ "step": 138
+ },
+ {
+ "epoch": 1.936244541484716,
+ "grad_norm": 0.017549855634570122,
+ "learning_rate": 0.0006,
+ "loss": 7.31191349029541,
+ "step": 139
+ },
+ {
+ "epoch": 1.9502183406113538,
+ "grad_norm": 0.015930157154798508,
+ "learning_rate": 0.0006,
+ "loss": 7.330901622772217,
+ "step": 140
+ },
+ {
+ "epoch": 1.9641921397379911,
+ "grad_norm": 0.013283558189868927,
+ "learning_rate": 0.0006,
+ "loss": 7.298318386077881,
+ "step": 141
+ },
+ {
+ "epoch": 1.9781659388646289,
+ "grad_norm": 0.012023469433188438,
+ "learning_rate": 0.0006,
+ "loss": 7.282337188720703,
+ "step": 142
+ },
+ {
+ "epoch": 1.9921397379912664,
+ "grad_norm": 0.01459511648863554,
+ "learning_rate": 0.0006,
+ "loss": 7.282924175262451,
+ "step": 143
+ },
+ {
+ "epoch": 2.0,
+ "grad_norm": 0.017474235966801643,
+ "learning_rate": 0.0006,
+ "loss": 7.254333972930908,
+ "step": 144
+ },
+ {
+ "epoch": 2.0,
+ "eval_loss": 7.318708896636963,
+ "eval_runtime": 59.6555,
+ "eval_samples_per_second": 40.935,
+ "eval_steps_per_second": 1.291,
+ "step": 144
+ },
+ {
+ "epoch": 2.0139737991266378,
+ "grad_norm": 0.019757507368922234,
+ "learning_rate": 0.0006,
+ "loss": 7.30288553237915,
+ "step": 145
+ },
+ {
+ "epoch": 2.027947598253275,
+ "grad_norm": 0.014571248553693295,
+ "learning_rate": 0.0006,
+ "loss": 7.274681568145752,
+ "step": 146
+ },
+ {
+ "epoch": 2.041921397379913,
+ "grad_norm": 0.015744198113679886,
+ "learning_rate": 0.0006,
+ "loss": 7.278632640838623,
+ "step": 147
+ },
+ {
+ "epoch": 2.05589519650655,
+ "grad_norm": 0.028048858046531677,
+ "learning_rate": 0.0006,
+ "loss": 7.26558780670166,
+ "step": 148
+ },
+ {
+ "epoch": 2.069868995633188,
+ "grad_norm": 0.06547307223081589,
+ "learning_rate": 0.0006,
+ "loss": 7.32567834854126,
+ "step": 149
+ },
+ {
+ "epoch": 2.0838427947598253,
+ "grad_norm": 0.05969763547182083,
+ "learning_rate": 0.0006,
+ "loss": 7.395836353302002,
+ "step": 150
+ },
+ {
+ "epoch": 2.097816593886463,
+ "grad_norm": 0.032574281096458435,
+ "learning_rate": 0.0006,
+ "loss": 7.318150997161865,
+ "step": 151
+ },
+ {
+ "epoch": 2.1117903930131003,
+ "grad_norm": 0.029452962800860405,
+ "learning_rate": 0.0006,
+ "loss": 7.308066368103027,
+ "step": 152
+ },
+ {
+ "epoch": 2.125764192139738,
+ "grad_norm": 0.03161991387605667,
+ "learning_rate": 0.0006,
+ "loss": 7.306290626525879,
+ "step": 153
+ },
+ {
+ "epoch": 2.1397379912663754,
+ "grad_norm": 0.027096295729279518,
+ "learning_rate": 0.0006,
+ "loss": 7.297796249389648,
+ "step": 154
+ },
+ {
+ "epoch": 2.153711790393013,
+ "grad_norm": 0.028657056391239166,
+ "learning_rate": 0.0006,
+ "loss": 7.317424774169922,
+ "step": 155
+ },
+ {
+ "epoch": 2.1676855895196505,
+ "grad_norm": 0.02046995982527733,
+ "learning_rate": 0.0006,
+ "loss": 7.297986030578613,
+ "step": 156
+ },
+ {
+ "epoch": 2.1816593886462883,
+ "grad_norm": 0.02220369316637516,
+ "learning_rate": 0.0006,
+ "loss": 7.268237590789795,
+ "step": 157
+ },
+ {
+ "epoch": 2.1956331877729256,
+ "grad_norm": 0.018381357192993164,
+ "learning_rate": 0.0006,
+ "loss": 7.259415626525879,
+ "step": 158
+ },
+ {
+ "epoch": 2.2096069868995634,
+ "grad_norm": 0.0204413291066885,
+ "learning_rate": 0.0006,
+ "loss": 7.258173942565918,
+ "step": 159
+ },
+ {
+ "epoch": 2.2235807860262007,
+ "grad_norm": 0.015874288976192474,
+ "learning_rate": 0.0006,
+ "loss": 7.281834602355957,
+ "step": 160
+ },
+ {
+ "epoch": 2.2375545851528384,
+ "grad_norm": 0.01634068600833416,
+ "learning_rate": 0.0006,
+ "loss": 7.2469868659973145,
+ "step": 161
+ },
+ {
+ "epoch": 2.251528384279476,
+ "grad_norm": 0.01728747971355915,
+ "learning_rate": 0.0006,
+ "loss": 7.280995845794678,
+ "step": 162
+ },
+ {
+ "epoch": 2.2655021834061135,
+ "grad_norm": 0.016341188922524452,
+ "learning_rate": 0.0006,
+ "loss": 7.273357391357422,
+ "step": 163
+ },
+ {
+ "epoch": 2.279475982532751,
+ "grad_norm": 0.016125807538628578,
+ "learning_rate": 0.0006,
+ "loss": 7.273510932922363,
+ "step": 164
+ },
+ {
+ "epoch": 2.2934497816593886,
+ "grad_norm": 0.01629287749528885,
+ "learning_rate": 0.0006,
+ "loss": 7.267029762268066,
+ "step": 165
+ },
+ {
+ "epoch": 2.3074235807860264,
+ "grad_norm": 0.015296131372451782,
+ "learning_rate": 0.0006,
+ "loss": 7.211989402770996,
+ "step": 166
+ },
+ {
+ "epoch": 2.3213973799126637,
+ "grad_norm": 0.01709570363163948,
+ "learning_rate": 0.0006,
+ "loss": 7.242851734161377,
+ "step": 167
+ },
+ {
+ "epoch": 2.3353711790393015,
+ "grad_norm": 0.012552388943731785,
+ "learning_rate": 0.0006,
+ "loss": 7.241696834564209,
+ "step": 168
+ },
+ {
+ "epoch": 2.349344978165939,
+ "grad_norm": 0.014571291394531727,
+ "learning_rate": 0.0006,
+ "loss": 7.270681381225586,
+ "step": 169
+ },
+ {
+ "epoch": 2.3633187772925766,
+ "grad_norm": 0.013389437459409237,
+ "learning_rate": 0.0006,
+ "loss": 7.191600799560547,
+ "step": 170
+ },
+ {
+ "epoch": 2.377292576419214,
+ "grad_norm": 0.011895314790308475,
+ "learning_rate": 0.0006,
+ "loss": 7.213610649108887,
+ "step": 171
+ },
+ {
+ "epoch": 2.3912663755458516,
+ "grad_norm": 0.01345642190426588,
+ "learning_rate": 0.0006,
+ "loss": 7.193787574768066,
+ "step": 172
+ },
+ {
+ "epoch": 2.405240174672489,
+ "grad_norm": 0.0091794328764081,
+ "learning_rate": 0.0006,
+ "loss": 7.233364105224609,
+ "step": 173
+ },
+ {
+ "epoch": 2.4192139737991267,
+ "grad_norm": 0.013767086900770664,
+ "learning_rate": 0.0006,
+ "loss": 7.217272758483887,
+ "step": 174
+ },
+ {
+ "epoch": 2.433187772925764,
+ "grad_norm": 0.015134960412979126,
+ "learning_rate": 0.0006,
+ "loss": 7.192791938781738,
+ "step": 175
+ },
+ {
+ "epoch": 2.447161572052402,
+ "grad_norm": 0.01790648326277733,
+ "learning_rate": 0.0006,
+ "loss": 7.229467391967773,
+ "step": 176
+ },
+ {
+ "epoch": 2.461135371179039,
+ "grad_norm": 0.015362377278506756,
+ "learning_rate": 0.0006,
+ "loss": 7.180346488952637,
+ "step": 177
+ },
+ {
+ "epoch": 2.475109170305677,
+ "grad_norm": 0.010505498386919498,
+ "learning_rate": 0.0006,
+ "loss": 7.188833236694336,
+ "step": 178
+ },
+ {
+ "epoch": 2.489082969432314,
+ "grad_norm": 0.009129747748374939,
+ "learning_rate": 0.0006,
+ "loss": 7.203153133392334,
+ "step": 179
+ },
+ {
+ "epoch": 2.503056768558952,
+ "grad_norm": 0.01147883478552103,
+ "learning_rate": 0.0006,
+ "loss": 7.163824558258057,
+ "step": 180
+ },
+ {
+ "epoch": 2.5170305676855893,
+ "grad_norm": 0.011338942684233189,
+ "learning_rate": 0.0006,
+ "loss": 7.16287899017334,
+ "step": 181
+ },
+ {
+ "epoch": 2.531004366812227,
+ "grad_norm": 0.013277675025165081,
+ "learning_rate": 0.0006,
+ "loss": 7.155905723571777,
+ "step": 182
+ },
+ {
+ "epoch": 2.544978165938865,
+ "grad_norm": 0.015269347466528416,
+ "learning_rate": 0.0006,
+ "loss": 7.189371109008789,
+ "step": 183
+ },
+ {
+ "epoch": 2.558951965065502,
+ "grad_norm": 0.017905596643686295,
+ "learning_rate": 0.0006,
+ "loss": 7.177547931671143,
+ "step": 184
+ },
+ {
+ "epoch": 2.5729257641921395,
+ "grad_norm": 0.024431610479950905,
+ "learning_rate": 0.0006,
+ "loss": 7.192169189453125,
+ "step": 185
+ },
+ {
+ "epoch": 2.5868995633187772,
+ "grad_norm": 0.03656961768865585,
+ "learning_rate": 0.0006,
+ "loss": 7.124387741088867,
+ "step": 186
+ },
+ {
+ "epoch": 2.600873362445415,
+ "grad_norm": 0.04337048903107643,
+ "learning_rate": 0.0006,
+ "loss": 7.140231609344482,
+ "step": 187
+ },
+ {
+ "epoch": 2.6148471615720523,
+ "grad_norm": 0.04047228768467903,
+ "learning_rate": 0.0006,
+ "loss": 7.171927452087402,
+ "step": 188
+ },
+ {
+ "epoch": 2.62882096069869,
+ "grad_norm": 0.07825770974159241,
+ "learning_rate": 0.0006,
+ "loss": 7.207631587982178,
+ "step": 189
+ },
+ {
+ "epoch": 2.6427947598253274,
+ "grad_norm": 0.07118063420057297,
+ "learning_rate": 0.0006,
+ "loss": 7.209488391876221,
+ "step": 190
+ },
+ {
+ "epoch": 2.656768558951965,
+ "grad_norm": 0.05153293535113335,
+ "learning_rate": 0.0006,
+ "loss": 7.14793062210083,
+ "step": 191
+ },
+ {
+ "epoch": 2.6707423580786025,
+ "grad_norm": 0.03318839520215988,
+ "learning_rate": 0.0006,
+ "loss": 7.163010597229004,
+ "step": 192
+ },
+ {
+ "epoch": 2.6847161572052403,
+ "grad_norm": 0.029333539307117462,
+ "learning_rate": 0.0006,
+ "loss": 7.159955978393555,
+ "step": 193
+ },
+ {
+ "epoch": 2.6986899563318776,
+ "grad_norm": 0.029405880719423294,
+ "learning_rate": 0.0006,
+ "loss": 7.136178493499756,
+ "step": 194
+ },
+ {
+ "epoch": 2.7126637554585153,
+ "grad_norm": 0.037815988063812256,
+ "learning_rate": 0.0006,
+ "loss": 7.183186054229736,
+ "step": 195
+ },
+ {
+ "epoch": 2.726637554585153,
+ "grad_norm": 0.020307250320911407,
+ "learning_rate": 0.0006,
+ "loss": 7.133164882659912,
+ "step": 196
+ },
+ {
+ "epoch": 2.7406113537117904,
+ "grad_norm": 0.026048069819808006,
+ "learning_rate": 0.0006,
+ "loss": 7.124977111816406,
+ "step": 197
+ },
+ {
+ "epoch": 2.7545851528384278,
+ "grad_norm": 0.027316724881529808,
+ "learning_rate": 0.0006,
+ "loss": 7.07265567779541,
+ "step": 198
+ },
+ {
+ "epoch": 2.7685589519650655,
+ "grad_norm": 0.021629009395837784,
+ "learning_rate": 0.0006,
+ "loss": 7.150020599365234,
+ "step": 199
+ },
+ {
+ "epoch": 2.7825327510917033,
+ "grad_norm": 0.01605929434299469,
+ "learning_rate": 0.0006,
+ "loss": 7.10919189453125,
+ "step": 200
+ },
+ {
+ "epoch": 2.7965065502183406,
+ "grad_norm": 0.020800089463591576,
+ "learning_rate": 0.0006,
+ "loss": 7.132045269012451,
+ "step": 201
+ },
+ {
+ "epoch": 2.810480349344978,
+ "grad_norm": 0.019225774332880974,
+ "learning_rate": 0.0006,
+ "loss": 7.128364562988281,
+ "step": 202
+ },
+ {
+ "epoch": 2.8244541484716157,
+ "grad_norm": 0.01561372634023428,
+ "learning_rate": 0.0006,
+ "loss": 7.083669662475586,
+ "step": 203
+ },
+ {
+ "epoch": 2.8384279475982535,
+ "grad_norm": 0.021969519555568695,
+ "learning_rate": 0.0006,
+ "loss": 7.050841331481934,
+ "step": 204
+ },
+ {
+ "epoch": 2.8524017467248908,
+ "grad_norm": 0.021453695371747017,
+ "learning_rate": 0.0006,
+ "loss": 7.097085475921631,
+ "step": 205
+ },
+ {
+ "epoch": 2.8663755458515285,
+ "grad_norm": 0.022836022078990936,
+ "learning_rate": 0.0006,
+ "loss": 7.10405969619751,
+ "step": 206
+ },
+ {
+ "epoch": 2.880349344978166,
+ "grad_norm": 0.02430851384997368,
+ "learning_rate": 0.0006,
+ "loss": 7.045181751251221,
+ "step": 207
+ },
+ {
+ "epoch": 2.8943231441048036,
+ "grad_norm": 0.020166993141174316,
+ "learning_rate": 0.0006,
+ "loss": 7.068183898925781,
+ "step": 208
+ },
+ {
+ "epoch": 2.908296943231441,
+ "grad_norm": 0.0118505684658885,
+ "learning_rate": 0.0006,
+ "loss": 7.07132625579834,
+ "step": 209
+ },
+ {
+ "epoch": 2.9222707423580787,
+ "grad_norm": 0.023772427812218666,
+ "learning_rate": 0.0006,
+ "loss": 7.055238723754883,
+ "step": 210
+ },
+ {
+ "epoch": 2.936244541484716,
+ "grad_norm": 0.02350712940096855,
+ "learning_rate": 0.0006,
+ "loss": 7.055898666381836,
+ "step": 211
+ },
+ {
+ "epoch": 2.950218340611354,
+ "grad_norm": 0.017244907096028328,
+ "learning_rate": 0.0006,
+ "loss": 7.0099711418151855,
+ "step": 212
+ },
+ {
+ "epoch": 2.964192139737991,
+ "grad_norm": 0.021565575152635574,
+ "learning_rate": 0.0006,
+ "loss": 6.991751670837402,
+ "step": 213
+ },
+ {
+ "epoch": 2.978165938864629,
+ "grad_norm": 0.03241860866546631,
+ "learning_rate": 0.0006,
+ "loss": 7.005980491638184,
+ "step": 214
+ },
+ {
+ "epoch": 2.992139737991266,
+ "grad_norm": 0.04613726586103439,
+ "learning_rate": 0.0006,
+ "loss": 7.027859687805176,
+ "step": 215
+ },
+ {
+ "epoch": 3.0,
+ "grad_norm": 0.06404894590377808,
+ "learning_rate": 0.0006,
+ "loss": 7.123725891113281,
+ "step": 216
+ },
+ {
+ "epoch": 3.0,
+ "eval_loss": 7.070647716522217,
+ "eval_runtime": 58.5801,
+ "eval_samples_per_second": 41.687,
+ "eval_steps_per_second": 1.314,
+ "step": 216
+ },
+ {
+ "epoch": 3.0139737991266378,
+ "grad_norm": 0.05542432889342308,
+ "learning_rate": 0.0006,
+ "loss": 7.078299522399902,
+ "step": 217
+ },
+ {
+ "epoch": 3.027947598253275,
+ "grad_norm": 0.026484526693820953,
+ "learning_rate": 0.0006,
+ "loss": 6.961159706115723,
+ "step": 218
+ },
+ {
+ "epoch": 3.041921397379913,
+ "grad_norm": 0.028276391327381134,
+ "learning_rate": 0.0006,
+ "loss": 7.019961833953857,
+ "step": 219
+ },
+ {
+ "epoch": 3.05589519650655,
+ "grad_norm": 0.024486595764756203,
+ "learning_rate": 0.0006,
+ "loss": 7.039677619934082,
+ "step": 220
+ },
+ {
+ "epoch": 3.069868995633188,
+ "grad_norm": 0.017620893195271492,
+ "learning_rate": 0.0006,
+ "loss": 6.9270782470703125,
+ "step": 221
+ },
+ {
+ "epoch": 3.0838427947598253,
+ "grad_norm": 0.02547302283346653,
+ "learning_rate": 0.0006,
+ "loss": 6.924233436584473,
+ "step": 222
+ },
+ {
+ "epoch": 3.097816593886463,
+ "grad_norm": 0.02539009042084217,
+ "learning_rate": 0.0006,
+ "loss": 6.973979949951172,
+ "step": 223
+ },
+ {
+ "epoch": 3.1117903930131003,
+ "grad_norm": 0.03256227448582649,
+ "learning_rate": 0.0006,
+ "loss": 6.933725357055664,
+ "step": 224
+ },
+ {
+ "epoch": 3.125764192139738,
+ "grad_norm": 0.04651800915598869,
+ "learning_rate": 0.0006,
+ "loss": 6.951044082641602,
+ "step": 225
+ },
+ {
+ "epoch": 3.1397379912663754,
+ "grad_norm": 0.04822975769639015,
+ "learning_rate": 0.0006,
+ "loss": 6.9614410400390625,
+ "step": 226
+ },
+ {
+ "epoch": 3.153711790393013,
+ "grad_norm": 0.031165020540356636,
+ "learning_rate": 0.0006,
+ "loss": 6.91193151473999,
+ "step": 227
+ },
+ {
+ "epoch": 3.1676855895196505,
+ "grad_norm": 0.029754292219877243,
+ "learning_rate": 0.0006,
+ "loss": 6.91802978515625,
+ "step": 228
+ },
+ {
+ "epoch": 3.1816593886462883,
+ "grad_norm": 0.02193152718245983,
+ "learning_rate": 0.0006,
+ "loss": 6.928090572357178,
+ "step": 229
+ },
+ {
+ "epoch": 3.1956331877729256,
+ "grad_norm": 0.02428470179438591,
+ "learning_rate": 0.0006,
+ "loss": 6.90396785736084,
+ "step": 230
+ },
+ {
+ "epoch": 3.2096069868995634,
+ "grad_norm": 0.023374345153570175,
+ "learning_rate": 0.0006,
+ "loss": 6.885858535766602,
+ "step": 231
+ },
+ {
+ "epoch": 3.2235807860262007,
+ "grad_norm": 0.023405000567436218,
+ "learning_rate": 0.0006,
+ "loss": 6.8855767250061035,
+ "step": 232
+ },
+ {
+ "epoch": 3.2375545851528384,
+ "grad_norm": 0.017277134582400322,
+ "learning_rate": 0.0006,
+ "loss": 6.901449203491211,
+ "step": 233
+ },
+ {
+ "epoch": 3.251528384279476,
+ "grad_norm": 0.019496750086545944,
+ "learning_rate": 0.0006,
+ "loss": 6.885687828063965,
+ "step": 234
+ },
+ {
+ "epoch": 3.2655021834061135,
+ "grad_norm": 0.016893276944756508,
+ "learning_rate": 0.0006,
+ "loss": 6.872369766235352,
+ "step": 235
+ },
+ {
+ "epoch": 3.279475982532751,
+ "grad_norm": 0.02058715932071209,
+ "learning_rate": 0.0006,
+ "loss": 6.823280334472656,
+ "step": 236
+ },
+ {
+ "epoch": 3.2934497816593886,
+ "grad_norm": 0.015530755743384361,
+ "learning_rate": 0.0006,
+ "loss": 6.860499858856201,
+ "step": 237
+ },
+ {
+ "epoch": 3.3074235807860264,
+ "grad_norm": 0.01934926211833954,
+ "learning_rate": 0.0006,
+ "loss": 6.817448616027832,
+ "step": 238
+ },
+ {
+ "epoch": 3.3213973799126637,
+ "grad_norm": 0.017413552850484848,
+ "learning_rate": 0.0006,
+ "loss": 6.881955146789551,
+ "step": 239
+ },
+ {
+ "epoch": 3.3353711790393015,
+ "grad_norm": 0.026478875428438187,
+ "learning_rate": 0.0006,
+ "loss": 6.827404499053955,
+ "step": 240
+ },
+ {
+ "epoch": 3.349344978165939,
+ "grad_norm": 0.032611701637506485,
+ "learning_rate": 0.0006,
+ "loss": 6.78285551071167,
+ "step": 241
+ },
+ {
+ "epoch": 3.3633187772925766,
+ "grad_norm": 0.041900306940078735,
+ "learning_rate": 0.0006,
+ "loss": 6.783053874969482,
+ "step": 242
+ },
+ {
+ "epoch": 3.377292576419214,
+ "grad_norm": 0.04149497672915459,
+ "learning_rate": 0.0006,
+ "loss": 6.745220184326172,
+ "step": 243
+ },
+ {
+ "epoch": 3.3912663755458516,
+ "grad_norm": 0.023949826136231422,
+ "learning_rate": 0.0006,
+ "loss": 6.841533184051514,
+ "step": 244
+ },
+ {
+ "epoch": 3.405240174672489,
+ "grad_norm": 0.028601842001080513,
+ "learning_rate": 0.0006,
+ "loss": 6.852852821350098,
+ "step": 245
+ },
+ {
+ "epoch": 3.4192139737991267,
+ "grad_norm": 0.027365155518054962,
+ "learning_rate": 0.0006,
+ "loss": 6.74376106262207,
+ "step": 246
+ },
+ {
+ "epoch": 3.433187772925764,
+ "grad_norm": 0.02856568619608879,
+ "learning_rate": 0.0006,
+ "loss": 6.74906063079834,
+ "step": 247
+ },
+ {
+ "epoch": 3.447161572052402,
+ "grad_norm": 0.019172416999936104,
+ "learning_rate": 0.0006,
+ "loss": 6.809206008911133,
+ "step": 248
+ },
+ {
+ "epoch": 3.461135371179039,
+ "grad_norm": 0.02161746844649315,
+ "learning_rate": 0.0006,
+ "loss": 6.789888381958008,
+ "step": 249
+ },
+ {
+ "epoch": 3.475109170305677,
+ "grad_norm": 0.020273666828870773,
+ "learning_rate": 0.0006,
+ "loss": 6.73512077331543,
+ "step": 250
+ },
+ {
+ "epoch": 3.489082969432314,
+ "grad_norm": 0.022329283878207207,
+ "learning_rate": 0.0006,
+ "loss": 6.826148509979248,
+ "step": 251
+ },
+ {
+ "epoch": 3.503056768558952,
+ "grad_norm": 0.02284284122288227,
+ "learning_rate": 0.0006,
+ "loss": 6.748521327972412,
+ "step": 252
+ },
+ {
+ "epoch": 3.5170305676855893,
+ "grad_norm": 0.026368247345089912,
+ "learning_rate": 0.0006,
+ "loss": 6.645879745483398,
+ "step": 253
+ },
+ {
+ "epoch": 3.531004366812227,
+ "grad_norm": 0.03627244383096695,
+ "learning_rate": 0.0006,
+ "loss": 6.666064739227295,
+ "step": 254
+ },
+ {
+ "epoch": 3.544978165938865,
+ "grad_norm": 0.04140935093164444,
+ "learning_rate": 0.0006,
+ "loss": 6.712275981903076,
+ "step": 255
+ },
+ {
+ "epoch": 3.558951965065502,
+ "grad_norm": 0.0390462800860405,
+ "learning_rate": 0.0006,
+ "loss": 6.665900230407715,
+ "step": 256
+ },
+ {
+ "epoch": 3.5729257641921395,
+ "grad_norm": 0.05178583785891533,
+ "learning_rate": 0.0006,
+ "loss": 6.753298759460449,
+ "step": 257
+ },
+ {
+ "epoch": 3.5868995633187772,
+ "grad_norm": 0.054328061640262604,
+ "learning_rate": 0.0006,
+ "loss": 6.732633113861084,
+ "step": 258
+ },
+ {
+ "epoch": 3.600873362445415,
+ "grad_norm": 0.04590460658073425,
+ "learning_rate": 0.0006,
+ "loss": 6.65250301361084,
+ "step": 259
+ },
+ {
+ "epoch": 3.6148471615720523,
+ "grad_norm": 0.044331666082143784,
+ "learning_rate": 0.0006,
+ "loss": 6.68654203414917,
+ "step": 260
+ },
+ {
+ "epoch": 3.62882096069869,
+ "grad_norm": 0.05308730527758598,
+ "learning_rate": 0.0006,
+ "loss": 6.642482757568359,
+ "step": 261
+ },
+ {
+ "epoch": 3.6427947598253274,
+ "grad_norm": 0.03999016433954239,
+ "learning_rate": 0.0006,
+ "loss": 6.625227451324463,
+ "step": 262
+ },
+ {
+ "epoch": 3.656768558951965,
+ "grad_norm": 0.04088086634874344,
+ "learning_rate": 0.0006,
+ "loss": 6.67623233795166,
+ "step": 263
+ },
+ {
+ "epoch": 3.6707423580786025,
+ "grad_norm": 0.03615636005997658,
+ "learning_rate": 0.0006,
+ "loss": 6.6869001388549805,
+ "step": 264
+ },
+ {
+ "epoch": 3.6847161572052403,
+ "grad_norm": 0.02817360684275627,
+ "learning_rate": 0.0006,
+ "loss": 6.674601078033447,
+ "step": 265
+ },
+ {
+ "epoch": 3.6986899563318776,
+ "grad_norm": 0.032586123794317245,
+ "learning_rate": 0.0006,
+ "loss": 6.612994194030762,
+ "step": 266
+ },
+ {
+ "epoch": 3.7126637554585153,
+ "grad_norm": 0.031117867678403854,
+ "learning_rate": 0.0006,
+ "loss": 6.6244354248046875,
+ "step": 267
+ },
+ {
+ "epoch": 3.726637554585153,
+ "grad_norm": 0.028330687433481216,
+ "learning_rate": 0.0006,
+ "loss": 6.669557094573975,
+ "step": 268
+ },
+ {
+ "epoch": 3.7406113537117904,
+ "grad_norm": 0.021489202976226807,
+ "learning_rate": 0.0006,
+ "loss": 6.632035732269287,
+ "step": 269
+ },
+ {
+ "epoch": 3.7545851528384278,
+ "grad_norm": 0.024269424378871918,
+ "learning_rate": 0.0006,
+ "loss": 6.660867691040039,
+ "step": 270
+ },
+ {
+ "epoch": 3.7685589519650655,
+ "grad_norm": 0.02237873338162899,
+ "learning_rate": 0.0006,
+ "loss": 6.602587699890137,
+ "step": 271
+ },
+ {
+ "epoch": 3.7825327510917033,
+ "grad_norm": 0.026503929868340492,
+ "learning_rate": 0.0006,
+ "loss": 6.599200248718262,
+ "step": 272
+ },
+ {
+ "epoch": 3.7965065502183406,
+ "grad_norm": 0.03687124326825142,
+ "learning_rate": 0.0006,
+ "loss": 6.567296504974365,
+ "step": 273
+ },
+ {
+ "epoch": 3.810480349344978,
+ "grad_norm": 0.04708952456712723,
+ "learning_rate": 0.0006,
+ "loss": 6.621265411376953,
+ "step": 274
+ },
+ {
+ "epoch": 3.8244541484716157,
+ "grad_norm": 0.041846614331007004,
+ "learning_rate": 0.0006,
+ "loss": 6.550536155700684,
+ "step": 275
+ },
+ {
+ "epoch": 3.8384279475982535,
+ "grad_norm": 0.031249675899744034,
+ "learning_rate": 0.0006,
+ "loss": 6.54278564453125,
+ "step": 276
+ },
+ {
+ "epoch": 3.8524017467248908,
+ "grad_norm": 0.023082256317138672,
+ "learning_rate": 0.0006,
+ "loss": 6.544787406921387,
+ "step": 277
+ },
+ {
+ "epoch": 3.8663755458515285,
+ "grad_norm": 0.020995570346713066,
+ "learning_rate": 0.0006,
+ "loss": 6.518143653869629,
+ "step": 278
+ },
+ {
+ "epoch": 3.880349344978166,
+ "grad_norm": 0.020839324221014977,
+ "learning_rate": 0.0006,
+ "loss": 6.546667098999023,
+ "step": 279
+ },
+ {
+ "epoch": 3.8943231441048036,
+ "grad_norm": 0.018377432599663734,
+ "learning_rate": 0.0006,
+ "loss": 6.478307723999023,
+ "step": 280
+ },
+ {
+ "epoch": 3.908296943231441,
+ "grad_norm": 0.021713724359869957,
+ "learning_rate": 0.0006,
+ "loss": 6.485574245452881,
+ "step": 281
+ },
+ {
+ "epoch": 3.9222707423580787,
+ "grad_norm": 0.01359301246702671,
+ "learning_rate": 0.0006,
+ "loss": 6.570268630981445,
+ "step": 282
+ },
+ {
+ "epoch": 3.936244541484716,
+ "grad_norm": 0.016233332455158234,
+ "learning_rate": 0.0006,
+ "loss": 6.57150936126709,
+ "step": 283
+ },
+ {
+ "epoch": 3.950218340611354,
+ "grad_norm": 0.020687608048319817,
+ "learning_rate": 0.0006,
+ "loss": 6.527956008911133,
+ "step": 284
+ },
+ {
+ "epoch": 3.964192139737991,
+ "grad_norm": 0.02032964862883091,
+ "learning_rate": 0.0006,
+ "loss": 6.468169689178467,
+ "step": 285
+ },
+ {
+ "epoch": 3.978165938864629,
+ "grad_norm": 0.016676288098096848,
+ "learning_rate": 0.0006,
+ "loss": 6.448418617248535,
+ "step": 286
+ },
+ {
+ "epoch": 3.992139737991266,
+ "grad_norm": 0.0168539360165596,
+ "learning_rate": 0.0006,
+ "loss": 6.461180686950684,
+ "step": 287
+ },
+ {
+ "epoch": 4.0,
+ "grad_norm": 0.02666233666241169,
+ "learning_rate": 0.0006,
+ "loss": 6.344893932342529,
+ "step": 288
+ },
+ {
+ "epoch": 4.0,
+ "eval_loss": 6.5384979248046875,
+ "eval_runtime": 58.814,
+ "eval_samples_per_second": 41.521,
+ "eval_steps_per_second": 1.309,
+ "step": 288
+ },
+ {
+ "epoch": 4.013973799126638,
+ "grad_norm": 0.04154708981513977,
+ "learning_rate": 0.0006,
+ "loss": 6.41855525970459,
+ "step": 289
+ },
+ {
+ "epoch": 4.0279475982532755,
+ "grad_norm": 0.04895668104290962,
+ "learning_rate": 0.0006,
+ "loss": 6.499302864074707,
+ "step": 290
+ },
+ {
+ "epoch": 4.041921397379912,
+ "grad_norm": 0.034464482218027115,
+ "learning_rate": 0.0006,
+ "loss": 6.438611030578613,
+ "step": 291
+ },
+ {
+ "epoch": 4.05589519650655,
+ "grad_norm": 0.06845773756504059,
+ "learning_rate": 0.0006,
+ "loss": 6.538361072540283,
+ "step": 292
+ },
+ {
+ "epoch": 4.069868995633188,
+ "grad_norm": 0.053768858313560486,
+ "learning_rate": 0.0006,
+ "loss": 6.527524948120117,
+ "step": 293
+ },
+ {
+ "epoch": 4.083842794759826,
+ "grad_norm": 0.039037931710481644,
+ "learning_rate": 0.0006,
+ "loss": 6.503847122192383,
+ "step": 294
+ },
+ {
+ "epoch": 4.097816593886463,
+ "grad_norm": 0.03962196782231331,
+ "learning_rate": 0.0006,
+ "loss": 6.475178241729736,
+ "step": 295
+ },
+ {
+ "epoch": 4.1117903930131,
+ "grad_norm": 0.049101535230875015,
+ "learning_rate": 0.0006,
+ "loss": 6.474587917327881,
+ "step": 296
+ },
+ {
+ "epoch": 4.125764192139738,
+ "grad_norm": 0.02830282226204872,
+ "learning_rate": 0.0006,
+ "loss": 6.45065975189209,
+ "step": 297
+ },
+ {
+ "epoch": 4.139737991266376,
+ "grad_norm": 0.023153482005000114,
+ "learning_rate": 0.0006,
+ "loss": 6.5218706130981445,
+ "step": 298
+ },
+ {
+ "epoch": 4.153711790393013,
+ "grad_norm": 0.025064002722501755,
+ "learning_rate": 0.0006,
+ "loss": 6.456475257873535,
+ "step": 299
+ },
+ {
+ "epoch": 4.1676855895196505,
+ "grad_norm": 0.02357092685997486,
+ "learning_rate": 0.0006,
+ "loss": 6.485952377319336,
+ "step": 300
+ },
+ {
+ "epoch": 4.181659388646288,
+ "grad_norm": 0.021832305938005447,
+ "learning_rate": 0.0006,
+ "loss": 6.425506591796875,
+ "step": 301
+ },
+ {
+ "epoch": 4.195633187772926,
+ "grad_norm": 0.0227900929749012,
+ "learning_rate": 0.0006,
+ "loss": 6.439155578613281,
+ "step": 302
+ },
+ {
+ "epoch": 4.209606986899563,
+ "grad_norm": 0.015961607918143272,
+ "learning_rate": 0.0006,
+ "loss": 6.460862636566162,
+ "step": 303
+ },
+ {
+ "epoch": 4.223580786026201,
+ "grad_norm": 0.017876390367746353,
+ "learning_rate": 0.0006,
+ "loss": 6.44564151763916,
+ "step": 304
+ },
+ {
+ "epoch": 4.2375545851528384,
+ "grad_norm": 0.013703204691410065,
+ "learning_rate": 0.0006,
+ "loss": 6.418414115905762,
+ "step": 305
+ },
+ {
+ "epoch": 4.251528384279476,
+ "grad_norm": 0.015240306034684181,
+ "learning_rate": 0.0006,
+ "loss": 6.411238670349121,
+ "step": 306
+ },
+ {
+ "epoch": 4.265502183406113,
+ "grad_norm": 0.01587662845849991,
+ "learning_rate": 0.0006,
+ "loss": 6.395642280578613,
+ "step": 307
+ },
+ {
+ "epoch": 4.279475982532751,
+ "grad_norm": 0.014650700613856316,
+ "learning_rate": 0.0006,
+ "loss": 6.302793025970459,
+ "step": 308
+ },
+ {
+ "epoch": 4.293449781659389,
+ "grad_norm": 0.016533225774765015,
+ "learning_rate": 0.0006,
+ "loss": 6.380588054656982,
+ "step": 309
+ },
+ {
+ "epoch": 4.307423580786026,
+ "grad_norm": 0.019963741302490234,
+ "learning_rate": 0.0006,
+ "loss": 6.345336437225342,
+ "step": 310
+ },
+ {
+ "epoch": 4.321397379912664,
+ "grad_norm": 0.02237936295568943,
+ "learning_rate": 0.0006,
+ "loss": 6.387770652770996,
+ "step": 311
+ },
+ {
+ "epoch": 4.335371179039301,
+ "grad_norm": 0.026658328250050545,
+ "learning_rate": 0.0006,
+ "loss": 6.323662757873535,
+ "step": 312
+ },
+ {
+ "epoch": 4.349344978165939,
+ "grad_norm": 0.02852284163236618,
+ "learning_rate": 0.0006,
+ "loss": 6.382204055786133,
+ "step": 313
+ },
+ {
+ "epoch": 4.3633187772925766,
+ "grad_norm": 0.028815090656280518,
+ "learning_rate": 0.0006,
+ "loss": 6.402390480041504,
+ "step": 314
+ },
+ {
+ "epoch": 4.377292576419214,
+ "grad_norm": 0.029393529519438744,
+ "learning_rate": 0.0006,
+ "loss": 6.280278205871582,
+ "step": 315
+ },
+ {
+ "epoch": 4.391266375545851,
+ "grad_norm": 0.02186041697859764,
+ "learning_rate": 0.0006,
+ "loss": 6.302707672119141,
+ "step": 316
+ },
+ {
+ "epoch": 4.405240174672489,
+ "grad_norm": 0.02071218006312847,
+ "learning_rate": 0.0006,
+ "loss": 6.3365020751953125,
+ "step": 317
+ },
+ {
+ "epoch": 4.419213973799127,
+ "grad_norm": 0.02779117226600647,
+ "learning_rate": 0.0006,
+ "loss": 6.252157211303711,
+ "step": 318
+ },
+ {
+ "epoch": 4.4331877729257645,
+ "grad_norm": 0.03470654785633087,
+ "learning_rate": 0.0006,
+ "loss": 6.319070339202881,
+ "step": 319
+ },
+ {
+ "epoch": 4.447161572052401,
+ "grad_norm": 0.037067804485559464,
+ "learning_rate": 0.0006,
+ "loss": 6.314116477966309,
+ "step": 320
+ },
+ {
+ "epoch": 4.461135371179039,
+ "grad_norm": 0.032182756811380386,
+ "learning_rate": 0.0006,
+ "loss": 6.267298698425293,
+ "step": 321
+ },
+ {
+ "epoch": 4.475109170305677,
+ "grad_norm": 0.026305217295885086,
+ "learning_rate": 0.0006,
+ "loss": 6.331688404083252,
+ "step": 322
+ },
+ {
+ "epoch": 4.489082969432315,
+ "grad_norm": 0.027228357270359993,
+ "learning_rate": 0.0006,
+ "loss": 6.31139612197876,
+ "step": 323
+ },
+ {
+ "epoch": 4.503056768558952,
+ "grad_norm": 0.025781169533729553,
+ "learning_rate": 0.0006,
+ "loss": 6.194684982299805,
+ "step": 324
+ },
+ {
+ "epoch": 4.517030567685589,
+ "grad_norm": 0.029186977073550224,
+ "learning_rate": 0.0006,
+ "loss": 6.31306266784668,
+ "step": 325
+ },
+ {
+ "epoch": 4.531004366812227,
+ "grad_norm": 0.023608777672052383,
+ "learning_rate": 0.0006,
+ "loss": 6.285243034362793,
+ "step": 326
+ },
+ {
+ "epoch": 4.544978165938865,
+ "grad_norm": 0.02267594076693058,
+ "learning_rate": 0.0006,
+ "loss": 6.30342960357666,
+ "step": 327
+ },
+ {
+ "epoch": 4.558951965065502,
+ "grad_norm": 0.01926310732960701,
+ "learning_rate": 0.0006,
+ "loss": 6.280606269836426,
+ "step": 328
+ },
+ {
+ "epoch": 4.5729257641921395,
+ "grad_norm": 0.026484837755560875,
+ "learning_rate": 0.0006,
+ "loss": 6.237980365753174,
+ "step": 329
+ },
+ {
+ "epoch": 4.586899563318777,
+ "grad_norm": 0.026228854432702065,
+ "learning_rate": 0.0006,
+ "loss": 6.3018293380737305,
+ "step": 330
+ },
+ {
+ "epoch": 4.600873362445415,
+ "grad_norm": 0.022096967324614525,
+ "learning_rate": 0.0006,
+ "loss": 6.253863334655762,
+ "step": 331
+ },
+ {
+ "epoch": 4.614847161572053,
+ "grad_norm": 0.027223890647292137,
+ "learning_rate": 0.0006,
+ "loss": 6.166882038116455,
+ "step": 332
+ },
+ {
+ "epoch": 4.62882096069869,
+ "grad_norm": 0.03160124272108078,
+ "learning_rate": 0.0006,
+ "loss": 6.230309963226318,
+ "step": 333
+ },
+ {
+ "epoch": 4.642794759825327,
+ "grad_norm": 0.04110539332032204,
+ "learning_rate": 0.0006,
+ "loss": 6.2334818840026855,
+ "step": 334
+ },
+ {
+ "epoch": 4.656768558951965,
+ "grad_norm": 0.04355933889746666,
+ "learning_rate": 0.0006,
+ "loss": 6.160956382751465,
+ "step": 335
+ },
+ {
+ "epoch": 4.670742358078603,
+ "grad_norm": 0.03611086308956146,
+ "learning_rate": 0.0006,
+ "loss": 6.23399543762207,
+ "step": 336
+ },
+ {
+ "epoch": 4.68471615720524,
+ "grad_norm": 0.051811400800943375,
+ "learning_rate": 0.0006,
+ "loss": 6.241855144500732,
+ "step": 337
+ },
+ {
+ "epoch": 4.698689956331878,
+ "grad_norm": 0.057434193789958954,
+ "learning_rate": 0.0006,
+ "loss": 6.220993995666504,
+ "step": 338
+ },
+ {
+ "epoch": 4.712663755458515,
+ "grad_norm": 0.04420953616499901,
+ "learning_rate": 0.0006,
+ "loss": 6.300461292266846,
+ "step": 339
+ },
+ {
+ "epoch": 4.726637554585153,
+ "grad_norm": 0.03515457361936569,
+ "learning_rate": 0.0006,
+ "loss": 6.269349098205566,
+ "step": 340
+ },
+ {
+ "epoch": 4.74061135371179,
+ "grad_norm": 0.03222600743174553,
+ "learning_rate": 0.0006,
+ "loss": 6.248613357543945,
+ "step": 341
+ },
+ {
+ "epoch": 4.754585152838428,
+ "grad_norm": 0.0384046845138073,
+ "learning_rate": 0.0006,
+ "loss": 6.266979217529297,
+ "step": 342
+ },
+ {
+ "epoch": 4.7685589519650655,
+ "grad_norm": 0.0514867827296257,
+ "learning_rate": 0.0006,
+ "loss": 6.234103202819824,
+ "step": 343
+ },
+ {
+ "epoch": 4.782532751091703,
+ "grad_norm": 0.050747793167829514,
+ "learning_rate": 0.0006,
+ "loss": 6.322582244873047,
+ "step": 344
+ },
+ {
+ "epoch": 4.796506550218341,
+ "grad_norm": 0.040878988802433014,
+ "learning_rate": 0.0006,
+ "loss": 6.212509632110596,
+ "step": 345
+ },
+ {
+ "epoch": 4.810480349344978,
+ "grad_norm": 0.034655820578336716,
+ "learning_rate": 0.0006,
+ "loss": 6.265879154205322,
+ "step": 346
+ },
+ {
+ "epoch": 4.824454148471616,
+ "grad_norm": 0.04202224686741829,
+ "learning_rate": 0.0006,
+ "loss": 6.2896409034729,
+ "step": 347
+ },
+ {
+ "epoch": 4.8384279475982535,
+ "grad_norm": 0.03632277995347977,
+ "learning_rate": 0.0006,
+ "loss": 6.253917217254639,
+ "step": 348
+ },
+ {
+ "epoch": 4.85240174672489,
+ "grad_norm": 0.028645765036344528,
+ "learning_rate": 0.0006,
+ "loss": 6.207965850830078,
+ "step": 349
+ },
+ {
+ "epoch": 4.866375545851528,
+ "grad_norm": 0.025305170565843582,
+ "learning_rate": 0.0006,
+ "loss": 6.168068885803223,
+ "step": 350
+ },
+ {
+ "epoch": 4.880349344978166,
+ "grad_norm": 0.022137803956866264,
+ "learning_rate": 0.0006,
+ "loss": 6.235283374786377,
+ "step": 351
+ },
+ {
+ "epoch": 4.894323144104804,
+ "grad_norm": 0.01814538985490799,
+ "learning_rate": 0.0006,
+ "loss": 6.259641647338867,
+ "step": 352
+ },
+ {
+ "epoch": 4.908296943231441,
+ "grad_norm": 0.018465086817741394,
+ "learning_rate": 0.0006,
+ "loss": 6.111738681793213,
+ "step": 353
+ },
+ {
+ "epoch": 4.922270742358078,
+ "grad_norm": 0.018638933077454567,
+ "learning_rate": 0.0006,
+ "loss": 6.225642204284668,
+ "step": 354
+ },
+ {
+ "epoch": 4.936244541484716,
+ "grad_norm": 0.016461260616779327,
+ "learning_rate": 0.0006,
+ "loss": 6.140647888183594,
+ "step": 355
+ },
+ {
+ "epoch": 4.950218340611354,
+ "grad_norm": 0.015458385460078716,
+ "learning_rate": 0.0006,
+ "loss": 6.155224800109863,
+ "step": 356
+ },
+ {
+ "epoch": 4.964192139737992,
+ "grad_norm": 0.014934048056602478,
+ "learning_rate": 0.0006,
+ "loss": 6.101555824279785,
+ "step": 357
+ },
+ {
+ "epoch": 4.978165938864628,
+ "grad_norm": 0.013979545794427395,
+ "learning_rate": 0.0006,
+ "loss": 6.161718845367432,
+ "step": 358
+ },
+ {
+ "epoch": 4.992139737991266,
+ "grad_norm": 0.01640394888818264,
+ "learning_rate": 0.0006,
+ "loss": 6.055768966674805,
+ "step": 359
+ },
+ {
+ "epoch": 5.0,
+ "grad_norm": 0.015135680325329304,
+ "learning_rate": 0.0006,
+ "loss": 6.169695854187012,
+ "step": 360
+ },
+ {
+ "epoch": 5.0,
+ "eval_loss": 6.16408109664917,
+ "eval_runtime": 59.3649,
+ "eval_samples_per_second": 41.135,
+ "eval_steps_per_second": 1.297,
+ "step": 360
+ },
+ {
+ "epoch": 5.013973799126638,
+ "grad_norm": 0.012849048711359501,
+ "learning_rate": 0.0006,
+ "loss": 6.055830001831055,
+ "step": 361
+ },
+ {
+ "epoch": 5.0279475982532755,
+ "grad_norm": 0.018169576302170753,
+ "learning_rate": 0.0006,
+ "loss": 6.082655429840088,
+ "step": 362
+ },
+ {
+ "epoch": 5.041921397379912,
+ "grad_norm": 0.024293430149555206,
+ "learning_rate": 0.0006,
+ "loss": 6.109032154083252,
+ "step": 363
+ },
+ {
+ "epoch": 5.05589519650655,
+ "grad_norm": 0.03266787901520729,
+ "learning_rate": 0.0006,
+ "loss": 6.030613899230957,
+ "step": 364
+ },
+ {
+ "epoch": 5.069868995633188,
+ "grad_norm": 0.03572266176342964,
+ "learning_rate": 0.0006,
+ "loss": 6.050390720367432,
+ "step": 365
+ },
+ {
+ "epoch": 5.083842794759826,
+ "grad_norm": 0.032625213265419006,
+ "learning_rate": 0.0006,
+ "loss": 6.169313430786133,
+ "step": 366
+ },
+ {
+ "epoch": 5.097816593886463,
+ "grad_norm": 0.04051872715353966,
+ "learning_rate": 0.0006,
+ "loss": 6.10398006439209,
+ "step": 367
+ },
+ {
+ "epoch": 5.1117903930131,
+ "grad_norm": 0.0351213738322258,
+ "learning_rate": 0.0006,
+ "loss": 6.121973514556885,
+ "step": 368
+ },
+ {
+ "epoch": 5.125764192139738,
+ "grad_norm": 0.03523759916424751,
+ "learning_rate": 0.0006,
+ "loss": 6.0626654624938965,
+ "step": 369
+ },
+ {
+ "epoch": 5.139737991266376,
+ "grad_norm": 0.02706182189285755,
+ "learning_rate": 0.0006,
+ "loss": 6.049354076385498,
+ "step": 370
+ },
+ {
+ "epoch": 5.153711790393013,
+ "grad_norm": 0.024212589487433434,
+ "learning_rate": 0.0006,
+ "loss": 6.034826278686523,
+ "step": 371
+ },
+ {
+ "epoch": 5.1676855895196505,
+ "grad_norm": 0.024798082187771797,
+ "learning_rate": 0.0006,
+ "loss": 6.004058837890625,
+ "step": 372
+ },
+ {
+ "epoch": 5.181659388646288,
+ "grad_norm": 0.02325567416846752,
+ "learning_rate": 0.0006,
+ "loss": 5.986461639404297,
+ "step": 373
+ },
+ {
+ "epoch": 5.195633187772926,
+ "grad_norm": 0.019060570746660233,
+ "learning_rate": 0.0006,
+ "loss": 6.059736251831055,
+ "step": 374
+ },
+ {
+ "epoch": 5.209606986899563,
+ "grad_norm": 0.016822976991534233,
+ "learning_rate": 0.0006,
+ "loss": 6.072957515716553,
+ "step": 375
+ },
+ {
+ "epoch": 5.223580786026201,
+ "grad_norm": 0.017218658700585365,
+ "learning_rate": 0.0006,
+ "loss": 6.031739234924316,
+ "step": 376
+ },
+ {
+ "epoch": 5.2375545851528384,
+ "grad_norm": 0.014889383688569069,
+ "learning_rate": 0.0006,
+ "loss": 6.025674343109131,
+ "step": 377
+ },
+ {
+ "epoch": 5.251528384279476,
+ "grad_norm": 0.01708107627928257,
+ "learning_rate": 0.0006,
+ "loss": 6.090451240539551,
+ "step": 378
+ },
+ {
+ "epoch": 5.265502183406113,
+ "grad_norm": 0.020260317251086235,
+ "learning_rate": 0.0006,
+ "loss": 6.094257354736328,
+ "step": 379
+ },
+ {
+ "epoch": 5.279475982532751,
+ "grad_norm": 0.020110400393605232,
+ "learning_rate": 0.0006,
+ "loss": 6.021188735961914,
+ "step": 380
+ },
+ {
+ "epoch": 5.293449781659389,
+ "grad_norm": 0.02027830481529236,
+ "learning_rate": 0.0006,
+ "loss": 6.022156238555908,
+ "step": 381
+ },
+ {
+ "epoch": 5.307423580786026,
+ "grad_norm": 0.022747062146663666,
+ "learning_rate": 0.0006,
+ "loss": 6.011836051940918,
+ "step": 382
+ },
+ {
+ "epoch": 5.321397379912664,
+ "grad_norm": 0.01990230567753315,
+ "learning_rate": 0.0006,
+ "loss": 6.003054618835449,
+ "step": 383
+ },
+ {
+ "epoch": 5.335371179039301,
+ "grad_norm": 0.01596238650381565,
+ "learning_rate": 0.0006,
+ "loss": 5.991410732269287,
+ "step": 384
+ },
+ {
+ "epoch": 5.349344978165939,
+ "grad_norm": 0.015847105532884598,
+ "learning_rate": 0.0006,
+ "loss": 6.100622177124023,
+ "step": 385
+ },
+ {
+ "epoch": 5.3633187772925766,
+ "grad_norm": 0.016179397702217102,
+ "learning_rate": 0.0006,
+ "loss": 6.032659530639648,
+ "step": 386
+ },
+ {
+ "epoch": 5.377292576419214,
+ "grad_norm": 0.018256602808833122,
+ "learning_rate": 0.0006,
+ "loss": 5.961983680725098,
+ "step": 387
+ },
+ {
+ "epoch": 5.391266375545851,
+ "grad_norm": 0.02805912122130394,
+ "learning_rate": 0.0006,
+ "loss": 5.983541965484619,
+ "step": 388
+ },
+ {
+ "epoch": 5.405240174672489,
+ "grad_norm": 0.039082255214452744,
+ "learning_rate": 0.0006,
+ "loss": 6.021474361419678,
+ "step": 389
+ },
+ {
+ "epoch": 5.419213973799127,
+ "grad_norm": 0.036757659167051315,
+ "learning_rate": 0.0006,
+ "loss": 6.005046367645264,
+ "step": 390
+ },
+ {
+ "epoch": 5.4331877729257645,
+ "grad_norm": 0.035277366638183594,
+ "learning_rate": 0.0006,
+ "loss": 6.036296844482422,
+ "step": 391
+ },
+ {
+ "epoch": 5.447161572052401,
+ "grad_norm": 0.034404635429382324,
+ "learning_rate": 0.0006,
+ "loss": 5.994539260864258,
+ "step": 392
+ },
+ {
+ "epoch": 5.461135371179039,
+ "grad_norm": 0.0377206988632679,
+ "learning_rate": 0.0006,
+ "loss": 5.941534519195557,
+ "step": 393
+ },
+ {
+ "epoch": 5.475109170305677,
+ "grad_norm": 0.0389922633767128,
+ "learning_rate": 0.0006,
+ "loss": 5.983644485473633,
+ "step": 394
+ },
+ {
+ "epoch": 5.489082969432315,
+ "grad_norm": 0.04176778718829155,
+ "learning_rate": 0.0006,
+ "loss": 6.030922889709473,
+ "step": 395
+ },
+ {
+ "epoch": 5.503056768558952,
+ "grad_norm": 0.0337153784930706,
+ "learning_rate": 0.0006,
+ "loss": 5.961367130279541,
+ "step": 396
+ },
+ {
+ "epoch": 5.517030567685589,
+ "grad_norm": 0.036116816103458405,
+ "learning_rate": 0.0006,
+ "loss": 5.819280624389648,
+ "step": 397
+ },
+ {
+ "epoch": 5.531004366812227,
+ "grad_norm": 0.030724041163921356,
+ "learning_rate": 0.0006,
+ "loss": 5.906380653381348,
+ "step": 398
+ },
+ {
+ "epoch": 5.544978165938865,
+ "grad_norm": 0.030264202505350113,
+ "learning_rate": 0.0006,
+ "loss": 5.975833415985107,
+ "step": 399
+ },
+ {
+ "epoch": 5.558951965065502,
+ "grad_norm": 0.032096318900585175,
+ "learning_rate": 0.0006,
+ "loss": 6.001348972320557,
+ "step": 400
+ },
+ {
+ "epoch": 5.5729257641921395,
+ "grad_norm": 0.030579356476664543,
+ "learning_rate": 0.0006,
+ "loss": 5.928768634796143,
+ "step": 401
+ },
+ {
+ "epoch": 5.586899563318777,
+ "grad_norm": 0.028241973370313644,
+ "learning_rate": 0.0006,
+ "loss": 5.92582893371582,
+ "step": 402
+ },
+ {
+ "epoch": 5.600873362445415,
+ "grad_norm": 0.023086953908205032,
+ "learning_rate": 0.0006,
+ "loss": 5.892926216125488,
+ "step": 403
+ },
+ {
+ "epoch": 5.614847161572053,
+ "grad_norm": 0.02547992393374443,
+ "learning_rate": 0.0006,
+ "loss": 5.903195381164551,
+ "step": 404
+ },
+ {
+ "epoch": 5.62882096069869,
+ "grad_norm": 0.023089831694960594,
+ "learning_rate": 0.0006,
+ "loss": 5.912033557891846,
+ "step": 405
+ },
+ {
+ "epoch": 5.642794759825327,
+ "grad_norm": 0.025446368381381035,
+ "learning_rate": 0.0006,
+ "loss": 5.926138401031494,
+ "step": 406
+ },
+ {
+ "epoch": 5.656768558951965,
+ "grad_norm": 0.03748747706413269,
+ "learning_rate": 0.0006,
+ "loss": 5.913451194763184,
+ "step": 407
+ },
+ {
+ "epoch": 5.670742358078603,
+ "grad_norm": 0.03493810072541237,
+ "learning_rate": 0.0006,
+ "loss": 5.97633695602417,
+ "step": 408
+ },
+ {
+ "epoch": 5.68471615720524,
+ "grad_norm": 0.023131275549530983,
+ "learning_rate": 0.0006,
+ "loss": 5.899571418762207,
+ "step": 409
+ },
+ {
+ "epoch": 5.698689956331878,
+ "grad_norm": 0.02986014075577259,
+ "learning_rate": 0.0006,
+ "loss": 5.894903182983398,
+ "step": 410
+ },
+ {
+ "epoch": 5.712663755458515,
+ "grad_norm": 0.030171144753694534,
+ "learning_rate": 0.0006,
+ "loss": 5.876595497131348,
+ "step": 411
+ },
+ {
+ "epoch": 5.726637554585153,
+ "grad_norm": 0.029377546161413193,
+ "learning_rate": 0.0006,
+ "loss": 5.911637306213379,
+ "step": 412
+ },
+ {
+ "epoch": 5.74061135371179,
+ "grad_norm": 0.029829490929841995,
+ "learning_rate": 0.0006,
+ "loss": 5.819629669189453,
+ "step": 413
+ },
+ {
+ "epoch": 5.754585152838428,
+ "grad_norm": 0.024139299988746643,
+ "learning_rate": 0.0006,
+ "loss": 5.81728458404541,
+ "step": 414
+ },
+ {
+ "epoch": 5.7685589519650655,
+ "grad_norm": 0.021453432738780975,
+ "learning_rate": 0.0006,
+ "loss": 5.987326145172119,
+ "step": 415
+ },
+ {
+ "epoch": 5.782532751091703,
+ "grad_norm": 0.017942246049642563,
+ "learning_rate": 0.0006,
+ "loss": 5.893008232116699,
+ "step": 416
+ },
+ {
+ "epoch": 5.796506550218341,
+ "grad_norm": 0.019724637269973755,
+ "learning_rate": 0.0006,
+ "loss": 5.912441253662109,
+ "step": 417
+ },
+ {
+ "epoch": 5.810480349344978,
+ "grad_norm": 0.018597912043333054,
+ "learning_rate": 0.0006,
+ "loss": 5.860394477844238,
+ "step": 418
+ },
+ {
+ "epoch": 5.824454148471616,
+ "grad_norm": 0.016232412308454514,
+ "learning_rate": 0.0006,
+ "loss": 5.94991397857666,
+ "step": 419
+ },
+ {
+ "epoch": 5.8384279475982535,
+ "grad_norm": 0.01667204685509205,
+ "learning_rate": 0.0006,
+ "loss": 5.8506364822387695,
+ "step": 420
+ },
+ {
+ "epoch": 5.85240174672489,
+ "grad_norm": 0.014232151210308075,
+ "learning_rate": 0.0006,
+ "loss": 5.928424835205078,
+ "step": 421
+ },
+ {
+ "epoch": 5.866375545851528,
+ "grad_norm": 0.01570476032793522,
+ "learning_rate": 0.0006,
+ "loss": 5.953692436218262,
+ "step": 422
+ },
+ {
+ "epoch": 5.880349344978166,
+ "grad_norm": 0.01586179807782173,
+ "learning_rate": 0.0006,
+ "loss": 5.820014476776123,
+ "step": 423
+ },
+ {
+ "epoch": 5.894323144104804,
+ "grad_norm": 0.014729145914316177,
+ "learning_rate": 0.0006,
+ "loss": 5.860154628753662,
+ "step": 424
+ },
+ {
+ "epoch": 5.908296943231441,
+ "grad_norm": 0.013111459091305733,
+ "learning_rate": 0.0006,
+ "loss": 5.85988187789917,
+ "step": 425
+ },
+ {
+ "epoch": 5.922270742358078,
+ "grad_norm": 0.012937922962009907,
+ "learning_rate": 0.0006,
+ "loss": 5.756265640258789,
+ "step": 426
+ },
+ {
+ "epoch": 5.936244541484716,
+ "grad_norm": 0.01377453189343214,
+ "learning_rate": 0.0006,
+ "loss": 5.8333048820495605,
+ "step": 427
+ },
+ {
+ "epoch": 5.950218340611354,
+ "grad_norm": 0.014054795727133751,
+ "learning_rate": 0.0006,
+ "loss": 5.803333759307861,
+ "step": 428
+ },
+ {
+ "epoch": 5.964192139737992,
+ "grad_norm": 0.01674959622323513,
+ "learning_rate": 0.0006,
+ "loss": 5.8718485832214355,
+ "step": 429
+ },
+ {
+ "epoch": 5.978165938864628,
+ "grad_norm": 0.018733017146587372,
+ "learning_rate": 0.0006,
+ "loss": 5.875979900360107,
+ "step": 430
+ },
+ {
+ "epoch": 5.992139737991266,
+ "grad_norm": 0.02088983915746212,
+ "learning_rate": 0.0006,
+ "loss": 5.763634204864502,
+ "step": 431
+ },
+ {
+ "epoch": 6.0,
+ "grad_norm": 0.026705985888838768,
+ "learning_rate": 0.0006,
+ "loss": 5.6922478675842285,
+ "step": 432
+ },
+ {
+ "epoch": 6.0,
+ "eval_loss": 5.847958087921143,
+ "eval_runtime": 58.5451,
+ "eval_samples_per_second": 41.711,
+ "eval_steps_per_second": 1.315,
+ "step": 432
+ },
+ {
+ "epoch": 6.013973799126638,
+ "grad_norm": 0.03181544691324234,
+ "learning_rate": 0.0006,
+ "loss": 5.786368370056152,
+ "step": 433
+ },
+ {
+ "epoch": 6.0279475982532755,
+ "grad_norm": 0.03238112851977348,
+ "learning_rate": 0.0006,
+ "loss": 5.810310363769531,
+ "step": 434
+ },
+ {
+ "epoch": 6.041921397379912,
+ "grad_norm": 0.03916756808757782,
+ "learning_rate": 0.0006,
+ "loss": 5.7761993408203125,
+ "step": 435
+ },
+ {
+ "epoch": 6.05589519650655,
+ "grad_norm": 0.044009730219841,
+ "learning_rate": 0.0006,
+ "loss": 5.898112773895264,
+ "step": 436
+ },
+ {
+ "epoch": 6.069868995633188,
+ "grad_norm": 0.045235246419906616,
+ "learning_rate": 0.0006,
+ "loss": 5.876371383666992,
+ "step": 437
+ },
+ {
+ "epoch": 6.083842794759826,
+ "grad_norm": 0.050218384712934494,
+ "learning_rate": 0.0006,
+ "loss": 5.805103302001953,
+ "step": 438
+ },
+ {
+ "epoch": 6.097816593886463,
+ "grad_norm": 0.05444490164518356,
+ "learning_rate": 0.0006,
+ "loss": 5.826424598693848,
+ "step": 439
+ },
+ {
+ "epoch": 6.1117903930131,
+ "grad_norm": 0.04890982061624527,
+ "learning_rate": 0.0006,
+ "loss": 5.7735395431518555,
+ "step": 440
+ },
+ {
+ "epoch": 6.125764192139738,
+ "grad_norm": 0.05478639155626297,
+ "learning_rate": 0.0006,
+ "loss": 5.823677062988281,
+ "step": 441
+ },
+ {
+ "epoch": 6.139737991266376,
+ "grad_norm": 0.057562313973903656,
+ "learning_rate": 0.0006,
+ "loss": 5.889334678649902,
+ "step": 442
+ },
+ {
+ "epoch": 6.153711790393013,
+ "grad_norm": 0.06447703391313553,
+ "learning_rate": 0.0006,
+ "loss": 5.84621524810791,
+ "step": 443
+ },
+ {
+ "epoch": 6.1676855895196505,
+ "grad_norm": 0.048861872404813766,
+ "learning_rate": 0.0006,
+ "loss": 5.931595802307129,
+ "step": 444
+ },
+ {
+ "epoch": 6.181659388646288,
+ "grad_norm": 0.05521339178085327,
+ "learning_rate": 0.0006,
+ "loss": 5.8229780197143555,
+ "step": 445
+ },
+ {
+ "epoch": 6.195633187772926,
+ "grad_norm": 0.053666990250349045,
+ "learning_rate": 0.0006,
+ "loss": 5.879191875457764,
+ "step": 446
+ },
+ {
+ "epoch": 6.209606986899563,
+ "grad_norm": 0.0451025515794754,
+ "learning_rate": 0.0006,
+ "loss": 5.801628112792969,
+ "step": 447
+ },
+ {
+ "epoch": 6.223580786026201,
+ "grad_norm": 0.04291655868291855,
+ "learning_rate": 0.0006,
+ "loss": 5.914680480957031,
+ "step": 448
+ },
+ {
+ "epoch": 6.2375545851528384,
+ "grad_norm": 0.035102490335702896,
+ "learning_rate": 0.0006,
+ "loss": 5.843682765960693,
+ "step": 449
+ },
+ {
+ "epoch": 6.251528384279476,
+ "grad_norm": 0.03403995931148529,
+ "learning_rate": 0.0006,
+ "loss": 5.827154159545898,
+ "step": 450
+ },
+ {
+ "epoch": 6.265502183406113,
+ "grad_norm": 0.03444375470280647,
+ "learning_rate": 0.0006,
+ "loss": 5.800136566162109,
+ "step": 451
+ },
+ {
+ "epoch": 6.279475982532751,
+ "grad_norm": 0.03165159001946449,
+ "learning_rate": 0.0006,
+ "loss": 5.906252384185791,
+ "step": 452
+ },
+ {
+ "epoch": 6.293449781659389,
+ "grad_norm": 0.026153093203902245,
+ "learning_rate": 0.0006,
+ "loss": 5.7423200607299805,
+ "step": 453
+ },
+ {
+ "epoch": 6.307423580786026,
+ "grad_norm": 0.024357657879590988,
+ "learning_rate": 0.0006,
+ "loss": 5.701364517211914,
+ "step": 454
+ },
+ {
+ "epoch": 6.321397379912664,
+ "grad_norm": 0.021508049219846725,
+ "learning_rate": 0.0006,
+ "loss": 5.8328094482421875,
+ "step": 455
+ },
+ {
+ "epoch": 6.335371179039301,
+ "grad_norm": 0.017313921824097633,
+ "learning_rate": 0.0006,
+ "loss": 5.792145252227783,
+ "step": 456
+ },
+ {
+ "epoch": 6.349344978165939,
+ "grad_norm": 0.018563397228717804,
+ "learning_rate": 0.0006,
+ "loss": 5.732672214508057,
+ "step": 457
+ },
+ {
+ "epoch": 6.3633187772925766,
+ "grad_norm": 0.016568679362535477,
+ "learning_rate": 0.0006,
+ "loss": 5.81948184967041,
+ "step": 458
+ },
+ {
+ "epoch": 6.377292576419214,
+ "grad_norm": 0.014133770950138569,
+ "learning_rate": 0.0006,
+ "loss": 5.734982490539551,
+ "step": 459
+ },
+ {
+ "epoch": 6.391266375545851,
+ "grad_norm": 0.014114447869360447,
+ "learning_rate": 0.0006,
+ "loss": 5.70073127746582,
+ "step": 460
+ },
+ {
+ "epoch": 6.405240174672489,
+ "grad_norm": 0.012907393276691437,
+ "learning_rate": 0.0006,
+ "loss": 5.758626937866211,
+ "step": 461
+ },
+ {
+ "epoch": 6.419213973799127,
+ "grad_norm": 0.012673444114625454,
+ "learning_rate": 0.0006,
+ "loss": 5.775138854980469,
+ "step": 462
+ },
+ {
+ "epoch": 6.4331877729257645,
+ "grad_norm": 0.01311410591006279,
+ "learning_rate": 0.0006,
+ "loss": 5.70920991897583,
+ "step": 463
+ },
+ {
+ "epoch": 6.447161572052401,
+ "grad_norm": 0.012935544364154339,
+ "learning_rate": 0.0006,
+ "loss": 5.781479835510254,
+ "step": 464
+ },
+ {
+ "epoch": 6.461135371179039,
+ "grad_norm": 0.01439738366752863,
+ "learning_rate": 0.0006,
+ "loss": 5.691315650939941,
+ "step": 465
+ },
+ {
+ "epoch": 6.475109170305677,
+ "grad_norm": 0.01205186452716589,
+ "learning_rate": 0.0006,
+ "loss": 5.662499904632568,
+ "step": 466
+ },
+ {
+ "epoch": 6.489082969432315,
+ "grad_norm": 0.011575652286410332,
+ "learning_rate": 0.0006,
+ "loss": 5.6271772384643555,
+ "step": 467
+ },
+ {
+ "epoch": 6.503056768558952,
+ "grad_norm": 0.011672502383589745,
+ "learning_rate": 0.0006,
+ "loss": 5.761662006378174,
+ "step": 468
+ },
+ {
+ "epoch": 6.517030567685589,
+ "grad_norm": 0.011541751213371754,
+ "learning_rate": 0.0006,
+ "loss": 5.762078285217285,
+ "step": 469
+ },
+ {
+ "epoch": 6.531004366812227,
+ "grad_norm": 0.011396365240216255,
+ "learning_rate": 0.0006,
+ "loss": 5.67873477935791,
+ "step": 470
+ },
+ {
+ "epoch": 6.544978165938865,
+ "grad_norm": 0.010688499547541142,
+ "learning_rate": 0.0006,
+ "loss": 5.717051982879639,
+ "step": 471
+ },
+ {
+ "epoch": 6.558951965065502,
+ "grad_norm": 0.012224317528307438,
+ "learning_rate": 0.0006,
+ "loss": 5.707948684692383,
+ "step": 472
+ },
+ {
+ "epoch": 6.5729257641921395,
+ "grad_norm": 0.011856825090944767,
+ "learning_rate": 0.0006,
+ "loss": 5.70878791809082,
+ "step": 473
+ },
+ {
+ "epoch": 6.586899563318777,
+ "grad_norm": 0.01199064590036869,
+ "learning_rate": 0.0006,
+ "loss": 5.708697319030762,
+ "step": 474
+ },
+ {
+ "epoch": 6.600873362445415,
+ "grad_norm": 0.01219446212053299,
+ "learning_rate": 0.0006,
+ "loss": 5.6061577796936035,
+ "step": 475
+ },
+ {
+ "epoch": 6.614847161572053,
+ "grad_norm": 0.013397484086453915,
+ "learning_rate": 0.0006,
+ "loss": 5.624789714813232,
+ "step": 476
+ },
+ {
+ "epoch": 6.62882096069869,
+ "grad_norm": 0.01483136136084795,
+ "learning_rate": 0.0006,
+ "loss": 5.721141338348389,
+ "step": 477
+ },
+ {
+ "epoch": 6.642794759825327,
+ "grad_norm": 0.0194488987326622,
+ "learning_rate": 0.0006,
+ "loss": 5.604279518127441,
+ "step": 478
+ },
+ {
+ "epoch": 6.656768558951965,
+ "grad_norm": 0.02079896256327629,
+ "learning_rate": 0.0006,
+ "loss": 5.586322784423828,
+ "step": 479
+ },
+ {
+ "epoch": 6.670742358078603,
+ "grad_norm": 0.018996229395270348,
+ "learning_rate": 0.0006,
+ "loss": 5.5987772941589355,
+ "step": 480
+ },
+ {
+ "epoch": 6.68471615720524,
+ "grad_norm": 0.015779603272676468,
+ "learning_rate": 0.0006,
+ "loss": 5.588602066040039,
+ "step": 481
+ },
+ {
+ "epoch": 6.698689956331878,
+ "grad_norm": 0.015322973020374775,
+ "learning_rate": 0.0006,
+ "loss": 5.686467170715332,
+ "step": 482
+ },
+ {
+ "epoch": 6.712663755458515,
+ "grad_norm": 0.015282213687896729,
+ "learning_rate": 0.0006,
+ "loss": 5.664676666259766,
+ "step": 483
+ },
+ {
+ "epoch": 6.726637554585153,
+ "grad_norm": 0.016589272767305374,
+ "learning_rate": 0.0006,
+ "loss": 5.666739463806152,
+ "step": 484
+ },
+ {
+ "epoch": 6.74061135371179,
+ "grad_norm": 0.01811421848833561,
+ "learning_rate": 0.0006,
+ "loss": 5.647593021392822,
+ "step": 485
+ },
+ {
+ "epoch": 6.754585152838428,
+ "grad_norm": 0.019652029499411583,
+ "learning_rate": 0.0006,
+ "loss": 5.628726959228516,
+ "step": 486
+ },
+ {
+ "epoch": 6.7685589519650655,
+ "grad_norm": 0.02118665538728237,
+ "learning_rate": 0.0006,
+ "loss": 5.6580705642700195,
+ "step": 487
+ },
+ {
+ "epoch": 6.782532751091703,
+ "grad_norm": 0.02237832546234131,
+ "learning_rate": 0.0006,
+ "loss": 5.635931968688965,
+ "step": 488
+ },
+ {
+ "epoch": 6.796506550218341,
+ "grad_norm": 0.023897631093859673,
+ "learning_rate": 0.0006,
+ "loss": 5.517949104309082,
+ "step": 489
+ },
+ {
+ "epoch": 6.810480349344978,
+ "grad_norm": 0.02442227490246296,
+ "learning_rate": 0.0006,
+ "loss": 5.600021839141846,
+ "step": 490
+ },
+ {
+ "epoch": 6.824454148471616,
+ "grad_norm": 0.024675287306308746,
+ "learning_rate": 0.0006,
+ "loss": 5.6639909744262695,
+ "step": 491
+ },
+ {
+ "epoch": 6.8384279475982535,
+ "grad_norm": 0.030372392386198044,
+ "learning_rate": 0.0006,
+ "loss": 5.582888603210449,
+ "step": 492
+ },
+ {
+ "epoch": 6.85240174672489,
+ "grad_norm": 0.03237690031528473,
+ "learning_rate": 0.0006,
+ "loss": 5.573896884918213,
+ "step": 493
+ },
+ {
+ "epoch": 6.866375545851528,
+ "grad_norm": 0.03511589393019676,
+ "learning_rate": 0.0006,
+ "loss": 5.565878868103027,
+ "step": 494
+ },
+ {
+ "epoch": 6.880349344978166,
+ "grad_norm": 0.03883938118815422,
+ "learning_rate": 0.0006,
+ "loss": 5.58540678024292,
+ "step": 495
+ },
+ {
+ "epoch": 6.894323144104804,
+ "grad_norm": 0.04175502806901932,
+ "learning_rate": 0.0006,
+ "loss": 5.6095194816589355,
+ "step": 496
+ },
+ {
+ "epoch": 6.908296943231441,
+ "grad_norm": 0.04013441503047943,
+ "learning_rate": 0.0006,
+ "loss": 5.654401779174805,
+ "step": 497
+ },
+ {
+ "epoch": 6.922270742358078,
+ "grad_norm": 0.04618770629167557,
+ "learning_rate": 0.0006,
+ "loss": 5.6483473777771,
+ "step": 498
+ },
+ {
+ "epoch": 6.936244541484716,
+ "grad_norm": 0.04507308453321457,
+ "learning_rate": 0.0006,
+ "loss": 5.5856218338012695,
+ "step": 499
+ },
+ {
+ "epoch": 6.950218340611354,
+ "grad_norm": 0.03151383996009827,
+ "learning_rate": 0.0006,
+ "loss": 5.594086647033691,
+ "step": 500
+ },
+ {
+ "epoch": 6.964192139737992,
+ "grad_norm": 0.026883797720074654,
+ "learning_rate": 0.0006,
+ "loss": 5.607676029205322,
+ "step": 501
+ },
+ {
+ "epoch": 6.978165938864628,
+ "grad_norm": 0.02981836162507534,
+ "learning_rate": 0.0006,
+ "loss": 5.563666820526123,
+ "step": 502
+ },
+ {
+ "epoch": 6.992139737991266,
+ "grad_norm": 0.03233477845788002,
+ "learning_rate": 0.0006,
+ "loss": 5.667543411254883,
+ "step": 503
+ },
+ {
+ "epoch": 7.0,
+ "grad_norm": 0.029927456751465797,
+ "learning_rate": 0.0006,
+ "loss": 5.733482837677002,
+ "step": 504
+ },
+ {
+ "epoch": 7.0,
+ "eval_loss": 5.679076671600342,
+ "eval_runtime": 59.5965,
+ "eval_samples_per_second": 40.976,
+ "eval_steps_per_second": 1.292,
+ "step": 504
+ },
+ {
+ "epoch": 7.013973799126638,
+ "grad_norm": 0.027081597596406937,
+ "learning_rate": 0.0006,
+ "loss": 5.634487152099609,
+ "step": 505
+ },
+ {
+ "epoch": 7.0279475982532755,
+ "grad_norm": 0.031959421932697296,
+ "learning_rate": 0.0006,
+ "loss": 5.643294811248779,
+ "step": 506
+ },
+ {
+ "epoch": 7.041921397379912,
+ "grad_norm": 0.02802063524723053,
+ "learning_rate": 0.0006,
+ "loss": 5.573239326477051,
+ "step": 507
+ },
+ {
+ "epoch": 7.05589519650655,
+ "grad_norm": 0.032579705119132996,
+ "learning_rate": 0.0006,
+ "loss": 5.627150058746338,
+ "step": 508
+ },
+ {
+ "epoch": 7.069868995633188,
+ "grad_norm": 0.02503928542137146,
+ "learning_rate": 0.0006,
+ "loss": 5.6045308113098145,
+ "step": 509
+ },
+ {
+ "epoch": 7.083842794759826,
+ "grad_norm": 0.02421317622065544,
+ "learning_rate": 0.0006,
+ "loss": 5.522153377532959,
+ "step": 510
+ },
+ {
+ "epoch": 7.097816593886463,
+ "grad_norm": 0.021983426064252853,
+ "learning_rate": 0.0006,
+ "loss": 5.632939338684082,
+ "step": 511
+ },
+ {
+ "epoch": 7.1117903930131,
+ "grad_norm": 0.021933183073997498,
+ "learning_rate": 0.0006,
+ "loss": 5.592185974121094,
+ "step": 512
+ },
+ {
+ "epoch": 7.125764192139738,
+ "grad_norm": 0.02287031151354313,
+ "learning_rate": 0.0006,
+ "loss": 5.613700866699219,
+ "step": 513
+ },
+ {
+ "epoch": 7.139737991266376,
+ "grad_norm": 0.02281602844595909,
+ "learning_rate": 0.0006,
+ "loss": 5.551383018493652,
+ "step": 514
+ },
+ {
+ "epoch": 7.153711790393013,
+ "grad_norm": 0.021050360053777695,
+ "learning_rate": 0.0006,
+ "loss": 5.55275821685791,
+ "step": 515
+ },
+ {
+ "epoch": 7.1676855895196505,
+ "grad_norm": 0.018299926072359085,
+ "learning_rate": 0.0006,
+ "loss": 5.528225421905518,
+ "step": 516
+ },
+ {
+ "epoch": 7.181659388646288,
+ "grad_norm": 0.02024853602051735,
+ "learning_rate": 0.0006,
+ "loss": 5.522453784942627,
+ "step": 517
+ },
+ {
+ "epoch": 7.195633187772926,
+ "grad_norm": 0.02135239914059639,
+ "learning_rate": 0.0006,
+ "loss": 5.527522563934326,
+ "step": 518
+ },
+ {
+ "epoch": 7.209606986899563,
+ "grad_norm": 0.023394184187054634,
+ "learning_rate": 0.0006,
+ "loss": 5.4535017013549805,
+ "step": 519
+ },
+ {
+ "epoch": 7.223580786026201,
+ "grad_norm": 0.025952080264687538,
+ "learning_rate": 0.0006,
+ "loss": 5.5374674797058105,
+ "step": 520
+ },
+ {
+ "epoch": 7.2375545851528384,
+ "grad_norm": 0.022287718951702118,
+ "learning_rate": 0.0006,
+ "loss": 5.493753433227539,
+ "step": 521
+ },
+ {
+ "epoch": 7.251528384279476,
+ "grad_norm": 0.02018447406589985,
+ "learning_rate": 0.0006,
+ "loss": 5.510575294494629,
+ "step": 522
+ },
+ {
+ "epoch": 7.265502183406113,
+ "grad_norm": 0.02050507813692093,
+ "learning_rate": 0.0006,
+ "loss": 5.466026782989502,
+ "step": 523
+ },
+ {
+ "epoch": 7.279475982532751,
+ "grad_norm": 0.023688802495598793,
+ "learning_rate": 0.0006,
+ "loss": 5.485815525054932,
+ "step": 524
+ },
+ {
+ "epoch": 7.293449781659389,
+ "grad_norm": 0.021078843623399734,
+ "learning_rate": 0.0006,
+ "loss": 5.415581703186035,
+ "step": 525
+ },
+ {
+ "epoch": 7.307423580786026,
+ "grad_norm": 0.018266011029481888,
+ "learning_rate": 0.0006,
+ "loss": 5.402815818786621,
+ "step": 526
+ },
+ {
+ "epoch": 7.321397379912664,
+ "grad_norm": 0.019693493843078613,
+ "learning_rate": 0.0006,
+ "loss": 5.504674911499023,
+ "step": 527
+ },
+ {
+ "epoch": 7.335371179039301,
+ "grad_norm": 0.02031373605132103,
+ "learning_rate": 0.0006,
+ "loss": 5.503837585449219,
+ "step": 528
+ },
+ {
+ "epoch": 7.349344978165939,
+ "grad_norm": 0.017636626958847046,
+ "learning_rate": 0.0006,
+ "loss": 5.530580997467041,
+ "step": 529
+ },
+ {
+ "epoch": 7.3633187772925766,
+ "grad_norm": 0.01787244901061058,
+ "learning_rate": 0.0006,
+ "loss": 5.450218677520752,
+ "step": 530
+ },
+ {
+ "epoch": 7.377292576419214,
+ "grad_norm": 0.0170669574290514,
+ "learning_rate": 0.0006,
+ "loss": 5.481570243835449,
+ "step": 531
+ },
+ {
+ "epoch": 7.391266375545851,
+ "grad_norm": 0.01802165061235428,
+ "learning_rate": 0.0006,
+ "loss": 5.386394500732422,
+ "step": 532
+ },
+ {
+ "epoch": 7.405240174672489,
+ "grad_norm": 0.022949472069740295,
+ "learning_rate": 0.0006,
+ "loss": 5.424929618835449,
+ "step": 533
+ },
+ {
+ "epoch": 7.419213973799127,
+ "grad_norm": 0.034213390201330185,
+ "learning_rate": 0.0006,
+ "loss": 5.346663475036621,
+ "step": 534
+ },
+ {
+ "epoch": 7.4331877729257645,
+ "grad_norm": 0.03847593814134598,
+ "learning_rate": 0.0006,
+ "loss": 5.418482780456543,
+ "step": 535
+ },
+ {
+ "epoch": 7.447161572052401,
+ "grad_norm": 0.029393676668405533,
+ "learning_rate": 0.0006,
+ "loss": 5.456090927124023,
+ "step": 536
+ },
+ {
+ "epoch": 7.461135371179039,
+ "grad_norm": 0.03388667106628418,
+ "learning_rate": 0.0006,
+ "loss": 5.572983741760254,
+ "step": 537
+ },
+ {
+ "epoch": 7.475109170305677,
+ "grad_norm": 0.039690613746643066,
+ "learning_rate": 0.0006,
+ "loss": 5.414067268371582,
+ "step": 538
+ },
+ {
+ "epoch": 7.489082969432315,
+ "grad_norm": 0.03630776330828667,
+ "learning_rate": 0.0006,
+ "loss": 5.523739814758301,
+ "step": 539
+ },
+ {
+ "epoch": 7.503056768558952,
+ "grad_norm": 0.03356870263814926,
+ "learning_rate": 0.0006,
+ "loss": 5.444767951965332,
+ "step": 540
+ },
+ {
+ "epoch": 7.517030567685589,
+ "grad_norm": 0.030812304466962814,
+ "learning_rate": 0.0006,
+ "loss": 5.461244106292725,
+ "step": 541
+ },
+ {
+ "epoch": 7.531004366812227,
+ "grad_norm": 0.03319217637181282,
+ "learning_rate": 0.0006,
+ "loss": 5.502161026000977,
+ "step": 542
+ },
+ {
+ "epoch": 7.544978165938865,
+ "grad_norm": 0.032997481524944305,
+ "learning_rate": 0.0006,
+ "loss": 5.4162278175354,
+ "step": 543
+ },
+ {
+ "epoch": 7.558951965065502,
+ "grad_norm": 0.03364962339401245,
+ "learning_rate": 0.0006,
+ "loss": 5.415736198425293,
+ "step": 544
+ },
+ {
+ "epoch": 7.5729257641921395,
+ "grad_norm": 0.0344221405684948,
+ "learning_rate": 0.0006,
+ "loss": 5.541967868804932,
+ "step": 545
+ },
+ {
+ "epoch": 7.586899563318777,
+ "grad_norm": 0.029609503224492073,
+ "learning_rate": 0.0006,
+ "loss": 5.440614223480225,
+ "step": 546
+ },
+ {
+ "epoch": 7.600873362445415,
+ "grad_norm": 0.029055926948785782,
+ "learning_rate": 0.0006,
+ "loss": 5.462865352630615,
+ "step": 547
+ },
+ {
+ "epoch": 7.614847161572053,
+ "grad_norm": 0.02658848837018013,
+ "learning_rate": 0.0006,
+ "loss": 5.444467544555664,
+ "step": 548
+ },
+ {
+ "epoch": 7.62882096069869,
+ "grad_norm": 0.026276404038071632,
+ "learning_rate": 0.0006,
+ "loss": 5.462278842926025,
+ "step": 549
+ },
+ {
+ "epoch": 7.642794759825327,
+ "grad_norm": 0.0286889486014843,
+ "learning_rate": 0.0006,
+ "loss": 5.4849534034729,
+ "step": 550
+ },
+ {
+ "epoch": 7.656768558951965,
+ "grad_norm": 0.026849817484617233,
+ "learning_rate": 0.0006,
+ "loss": 5.431473731994629,
+ "step": 551
+ },
+ {
+ "epoch": 7.670742358078603,
+ "grad_norm": 0.02312396466732025,
+ "learning_rate": 0.0006,
+ "loss": 5.45462703704834,
+ "step": 552
+ },
+ {
+ "epoch": 7.68471615720524,
+ "grad_norm": 0.026197485625743866,
+ "learning_rate": 0.0006,
+ "loss": 5.452552318572998,
+ "step": 553
+ },
+ {
+ "epoch": 7.698689956331878,
+ "grad_norm": 0.02747255191206932,
+ "learning_rate": 0.0006,
+ "loss": 5.447169780731201,
+ "step": 554
+ },
+ {
+ "epoch": 7.712663755458515,
+ "grad_norm": 0.028123432770371437,
+ "learning_rate": 0.0006,
+ "loss": 5.394400596618652,
+ "step": 555
+ },
+ {
+ "epoch": 7.726637554585153,
+ "grad_norm": 0.02599870041012764,
+ "learning_rate": 0.0006,
+ "loss": 5.398341178894043,
+ "step": 556
+ },
+ {
+ "epoch": 7.74061135371179,
+ "grad_norm": 0.022171657532453537,
+ "learning_rate": 0.0006,
+ "loss": 5.368820667266846,
+ "step": 557
+ },
+ {
+ "epoch": 7.754585152838428,
+ "grad_norm": 0.022309480234980583,
+ "learning_rate": 0.0006,
+ "loss": 5.416772365570068,
+ "step": 558
+ },
+ {
+ "epoch": 7.7685589519650655,
+ "grad_norm": 0.024074165150523186,
+ "learning_rate": 0.0006,
+ "loss": 5.359002113342285,
+ "step": 559
+ },
+ {
+ "epoch": 7.782532751091703,
+ "grad_norm": 0.02636653557419777,
+ "learning_rate": 0.0006,
+ "loss": 5.484368324279785,
+ "step": 560
+ },
+ {
+ "epoch": 7.796506550218341,
+ "grad_norm": 0.02203752100467682,
+ "learning_rate": 0.0006,
+ "loss": 5.375498294830322,
+ "step": 561
+ },
+ {
+ "epoch": 7.810480349344978,
+ "grad_norm": 0.019924819469451904,
+ "learning_rate": 0.0006,
+ "loss": 5.334723472595215,
+ "step": 562
+ },
+ {
+ "epoch": 7.824454148471616,
+ "grad_norm": 0.018755966797471046,
+ "learning_rate": 0.0006,
+ "loss": 5.348781108856201,
+ "step": 563
+ },
+ {
+ "epoch": 7.8384279475982535,
+ "grad_norm": 0.016747845336794853,
+ "learning_rate": 0.0006,
+ "loss": 5.433079719543457,
+ "step": 564
+ },
+ {
+ "epoch": 7.85240174672489,
+ "grad_norm": 0.0170395877212286,
+ "learning_rate": 0.0006,
+ "loss": 5.346794128417969,
+ "step": 565
+ },
+ {
+ "epoch": 7.866375545851528,
+ "grad_norm": 0.016074176877737045,
+ "learning_rate": 0.0006,
+ "loss": 5.3277740478515625,
+ "step": 566
+ },
+ {
+ "epoch": 7.880349344978166,
+ "grad_norm": 0.014119806699454784,
+ "learning_rate": 0.0006,
+ "loss": 5.483551979064941,
+ "step": 567
+ },
+ {
+ "epoch": 7.894323144104804,
+ "grad_norm": 0.015270394273102283,
+ "learning_rate": 0.0006,
+ "loss": 5.378519535064697,
+ "step": 568
+ },
+ {
+ "epoch": 7.908296943231441,
+ "grad_norm": 0.015677539631724358,
+ "learning_rate": 0.0006,
+ "loss": 5.290210247039795,
+ "step": 569
+ },
+ {
+ "epoch": 7.922270742358078,
+ "grad_norm": 0.015930423513054848,
+ "learning_rate": 0.0006,
+ "loss": 5.427360534667969,
+ "step": 570
+ },
+ {
+ "epoch": 7.936244541484716,
+ "grad_norm": 0.016641564667224884,
+ "learning_rate": 0.0006,
+ "loss": 5.301599979400635,
+ "step": 571
+ },
+ {
+ "epoch": 7.950218340611354,
+ "grad_norm": 0.018293552100658417,
+ "learning_rate": 0.0006,
+ "loss": 5.277889251708984,
+ "step": 572
+ },
+ {
+ "epoch": 7.964192139737992,
+ "grad_norm": 0.017618799582123756,
+ "learning_rate": 0.0006,
+ "loss": 5.297194480895996,
+ "step": 573
+ },
+ {
+ "epoch": 7.978165938864628,
+ "grad_norm": 0.015034242533147335,
+ "learning_rate": 0.0006,
+ "loss": 5.42428731918335,
+ "step": 574
+ },
+ {
+ "epoch": 7.992139737991266,
+ "grad_norm": 0.01617511734366417,
+ "learning_rate": 0.0006,
+ "loss": 5.420044898986816,
+ "step": 575
+ },
+ {
+ "epoch": 8.0,
+ "grad_norm": 0.01837257295846939,
+ "learning_rate": 0.0006,
+ "loss": 5.240387916564941,
+ "step": 576
+ },
+ {
+ "epoch": 8.0,
+ "eval_loss": 5.423932075500488,
+ "eval_runtime": 58.0841,
+ "eval_samples_per_second": 42.042,
+ "eval_steps_per_second": 1.326,
+ "step": 576
+ },
+ {
+ "epoch": 8.013973799126637,
+ "grad_norm": 0.020953809842467308,
+ "learning_rate": 0.0006,
+ "loss": 5.4320783615112305,
+ "step": 577
+ },
+ {
+ "epoch": 8.027947598253276,
+ "grad_norm": 0.0227745920419693,
+ "learning_rate": 0.0006,
+ "loss": 5.310197830200195,
+ "step": 578
+ },
+ {
+ "epoch": 8.041921397379912,
+ "grad_norm": 0.021372603252530098,
+ "learning_rate": 0.0006,
+ "loss": 5.339468002319336,
+ "step": 579
+ },
+ {
+ "epoch": 8.055895196506551,
+ "grad_norm": 0.022401731461286545,
+ "learning_rate": 0.0006,
+ "loss": 5.279829978942871,
+ "step": 580
+ },
+ {
+ "epoch": 8.069868995633188,
+ "grad_norm": 0.023942379280924797,
+ "learning_rate": 0.0006,
+ "loss": 5.372323989868164,
+ "step": 581
+ },
+ {
+ "epoch": 8.083842794759825,
+ "grad_norm": 0.024380534887313843,
+ "learning_rate": 0.0006,
+ "loss": 5.253486633300781,
+ "step": 582
+ },
+ {
+ "epoch": 8.097816593886463,
+ "grad_norm": 0.03246625140309334,
+ "learning_rate": 0.0006,
+ "loss": 5.305774211883545,
+ "step": 583
+ },
+ {
+ "epoch": 8.1117903930131,
+ "grad_norm": 0.030399736016988754,
+ "learning_rate": 0.0006,
+ "loss": 5.370976448059082,
+ "step": 584
+ },
+ {
+ "epoch": 8.125764192139737,
+ "grad_norm": 0.027148913592100143,
+ "learning_rate": 0.0006,
+ "loss": 5.368010520935059,
+ "step": 585
+ },
+ {
+ "epoch": 8.139737991266376,
+ "grad_norm": 0.02925540879368782,
+ "learning_rate": 0.0006,
+ "loss": 5.303013801574707,
+ "step": 586
+ },
+ {
+ "epoch": 8.153711790393013,
+ "grad_norm": 0.02915453538298607,
+ "learning_rate": 0.0006,
+ "loss": 5.329667091369629,
+ "step": 587
+ },
+ {
+ "epoch": 8.167685589519651,
+ "grad_norm": 0.031732626259326935,
+ "learning_rate": 0.0006,
+ "loss": 5.3634934425354,
+ "step": 588
+ },
+ {
+ "epoch": 8.181659388646288,
+ "grad_norm": 0.036166731268167496,
+ "learning_rate": 0.0006,
+ "loss": 5.223980903625488,
+ "step": 589
+ },
+ {
+ "epoch": 8.195633187772925,
+ "grad_norm": 0.035654108971357346,
+ "learning_rate": 0.0006,
+ "loss": 5.361059188842773,
+ "step": 590
+ },
+ {
+ "epoch": 8.209606986899564,
+ "grad_norm": 0.0365324430167675,
+ "learning_rate": 0.0006,
+ "loss": 5.336151599884033,
+ "step": 591
+ },
+ {
+ "epoch": 8.2235807860262,
+ "grad_norm": 0.03578880429267883,
+ "learning_rate": 0.0006,
+ "loss": 5.434076309204102,
+ "step": 592
+ },
+ {
+ "epoch": 8.237554585152838,
+ "grad_norm": 0.035734012722969055,
+ "learning_rate": 0.0006,
+ "loss": 5.330893516540527,
+ "step": 593
+ },
+ {
+ "epoch": 8.251528384279476,
+ "grad_norm": 0.03196857124567032,
+ "learning_rate": 0.0006,
+ "loss": 5.306009769439697,
+ "step": 594
+ },
+ {
+ "epoch": 8.265502183406113,
+ "grad_norm": 0.032042086124420166,
+ "learning_rate": 0.0006,
+ "loss": 5.343267917633057,
+ "step": 595
+ },
+ {
+ "epoch": 8.279475982532752,
+ "grad_norm": 0.03160746395587921,
+ "learning_rate": 0.0006,
+ "loss": 5.2353715896606445,
+ "step": 596
+ },
+ {
+ "epoch": 8.293449781659389,
+ "grad_norm": 0.026689818128943443,
+ "learning_rate": 0.0006,
+ "loss": 5.254042625427246,
+ "step": 597
+ },
+ {
+ "epoch": 8.307423580786025,
+ "grad_norm": 0.02880188450217247,
+ "learning_rate": 0.0006,
+ "loss": 5.326833248138428,
+ "step": 598
+ },
+ {
+ "epoch": 8.321397379912664,
+ "grad_norm": 0.027516381815075874,
+ "learning_rate": 0.0006,
+ "loss": 5.342336177825928,
+ "step": 599
+ },
+ {
+ "epoch": 8.335371179039301,
+ "grad_norm": 0.027875030413269997,
+ "learning_rate": 0.0006,
+ "loss": 5.200719833374023,
+ "step": 600
+ },
+ {
+ "epoch": 8.34934497816594,
+ "grad_norm": 0.0268265288323164,
+ "learning_rate": 0.0006,
+ "loss": 5.249449729919434,
+ "step": 601
+ },
+ {
+ "epoch": 8.363318777292577,
+ "grad_norm": 0.024824608117341995,
+ "learning_rate": 0.0006,
+ "loss": 5.318952560424805,
+ "step": 602
+ },
+ {
+ "epoch": 8.377292576419213,
+ "grad_norm": 0.022990627214312553,
+ "learning_rate": 0.0006,
+ "loss": 5.244993209838867,
+ "step": 603
+ },
+ {
+ "epoch": 8.391266375545852,
+ "grad_norm": 0.022804604843258858,
+ "learning_rate": 0.0006,
+ "loss": 5.3236846923828125,
+ "step": 604
+ },
+ {
+ "epoch": 8.405240174672489,
+ "grad_norm": 0.02419872209429741,
+ "learning_rate": 0.0006,
+ "loss": 5.255486488342285,
+ "step": 605
+ },
+ {
+ "epoch": 8.419213973799126,
+ "grad_norm": 0.021952059119939804,
+ "learning_rate": 0.0006,
+ "loss": 5.3511552810668945,
+ "step": 606
+ },
+ {
+ "epoch": 8.433187772925764,
+ "grad_norm": 0.022375497967004776,
+ "learning_rate": 0.0006,
+ "loss": 5.294790267944336,
+ "step": 607
+ },
+ {
+ "epoch": 8.447161572052401,
+ "grad_norm": 0.019242815673351288,
+ "learning_rate": 0.0006,
+ "loss": 5.252618789672852,
+ "step": 608
+ },
+ {
+ "epoch": 8.46113537117904,
+ "grad_norm": 0.0173486340790987,
+ "learning_rate": 0.0006,
+ "loss": 5.30147647857666,
+ "step": 609
+ },
+ {
+ "epoch": 8.475109170305677,
+ "grad_norm": 0.01612604409456253,
+ "learning_rate": 0.0006,
+ "loss": 5.23960542678833,
+ "step": 610
+ },
+ {
+ "epoch": 8.489082969432314,
+ "grad_norm": 0.014293976128101349,
+ "learning_rate": 0.0006,
+ "loss": 5.119932651519775,
+ "step": 611
+ },
+ {
+ "epoch": 8.503056768558952,
+ "grad_norm": 0.014453536830842495,
+ "learning_rate": 0.0006,
+ "loss": 5.166794776916504,
+ "step": 612
+ },
+ {
+ "epoch": 8.51703056768559,
+ "grad_norm": 0.014685769565403461,
+ "learning_rate": 0.0006,
+ "loss": 5.264822959899902,
+ "step": 613
+ },
+ {
+ "epoch": 8.531004366812226,
+ "grad_norm": 0.01427740603685379,
+ "learning_rate": 0.0006,
+ "loss": 5.278433799743652,
+ "step": 614
+ },
+ {
+ "epoch": 8.544978165938865,
+ "grad_norm": 0.014864951372146606,
+ "learning_rate": 0.0006,
+ "loss": 5.214512348175049,
+ "step": 615
+ },
+ {
+ "epoch": 8.558951965065502,
+ "grad_norm": 0.01565164513885975,
+ "learning_rate": 0.0006,
+ "loss": 5.183066368103027,
+ "step": 616
+ },
+ {
+ "epoch": 8.57292576419214,
+ "grad_norm": 0.01779816672205925,
+ "learning_rate": 0.0006,
+ "loss": 5.163185119628906,
+ "step": 617
+ },
+ {
+ "epoch": 8.586899563318777,
+ "grad_norm": 0.017254827544093132,
+ "learning_rate": 0.0006,
+ "loss": 5.243555068969727,
+ "step": 618
+ },
+ {
+ "epoch": 8.600873362445414,
+ "grad_norm": 0.01650584116578102,
+ "learning_rate": 0.0006,
+ "loss": 5.234884738922119,
+ "step": 619
+ },
+ {
+ "epoch": 8.614847161572053,
+ "grad_norm": 0.017021115869283676,
+ "learning_rate": 0.0006,
+ "loss": 5.12271785736084,
+ "step": 620
+ },
+ {
+ "epoch": 8.62882096069869,
+ "grad_norm": 0.01773759163916111,
+ "learning_rate": 0.0006,
+ "loss": 5.242153167724609,
+ "step": 621
+ },
+ {
+ "epoch": 8.642794759825328,
+ "grad_norm": 0.015679042786359787,
+ "learning_rate": 0.0006,
+ "loss": 5.170779228210449,
+ "step": 622
+ },
+ {
+ "epoch": 8.656768558951965,
+ "grad_norm": 0.013760549947619438,
+ "learning_rate": 0.0006,
+ "loss": 5.238644599914551,
+ "step": 623
+ },
+ {
+ "epoch": 8.670742358078602,
+ "grad_norm": 0.014571522362530231,
+ "learning_rate": 0.0006,
+ "loss": 5.177056312561035,
+ "step": 624
+ },
+ {
+ "epoch": 8.68471615720524,
+ "grad_norm": 0.016209015622735023,
+ "learning_rate": 0.0006,
+ "loss": 5.163750648498535,
+ "step": 625
+ },
+ {
+ "epoch": 8.698689956331878,
+ "grad_norm": 0.016813941299915314,
+ "learning_rate": 0.0006,
+ "loss": 5.183428764343262,
+ "step": 626
+ },
+ {
+ "epoch": 8.712663755458514,
+ "grad_norm": 0.020985357463359833,
+ "learning_rate": 0.0006,
+ "loss": 5.229465007781982,
+ "step": 627
+ },
+ {
+ "epoch": 8.726637554585153,
+ "grad_norm": 0.02679632417857647,
+ "learning_rate": 0.0006,
+ "loss": 5.257368087768555,
+ "step": 628
+ },
+ {
+ "epoch": 8.74061135371179,
+ "grad_norm": 0.025756070390343666,
+ "learning_rate": 0.0006,
+ "loss": 5.253736972808838,
+ "step": 629
+ },
+ {
+ "epoch": 8.754585152838429,
+ "grad_norm": 0.02643490768969059,
+ "learning_rate": 0.0006,
+ "loss": 5.228633403778076,
+ "step": 630
+ },
+ {
+ "epoch": 8.768558951965066,
+ "grad_norm": 0.03303210437297821,
+ "learning_rate": 0.0006,
+ "loss": 5.163540363311768,
+ "step": 631
+ },
+ {
+ "epoch": 8.782532751091702,
+ "grad_norm": 0.030432431027293205,
+ "learning_rate": 0.0006,
+ "loss": 5.243169784545898,
+ "step": 632
+ },
+ {
+ "epoch": 8.796506550218341,
+ "grad_norm": 0.028631288558244705,
+ "learning_rate": 0.0006,
+ "loss": 5.166398048400879,
+ "step": 633
+ },
+ {
+ "epoch": 8.810480349344978,
+ "grad_norm": 0.026188310235738754,
+ "learning_rate": 0.0006,
+ "loss": 5.128122329711914,
+ "step": 634
+ },
+ {
+ "epoch": 8.824454148471617,
+ "grad_norm": 0.025308528915047646,
+ "learning_rate": 0.0006,
+ "loss": 5.211284637451172,
+ "step": 635
+ },
+ {
+ "epoch": 8.838427947598253,
+ "grad_norm": 0.020657729357481003,
+ "learning_rate": 0.0006,
+ "loss": 5.161575794219971,
+ "step": 636
+ },
+ {
+ "epoch": 8.85240174672489,
+ "grad_norm": 0.021901234984397888,
+ "learning_rate": 0.0006,
+ "loss": 5.20050048828125,
+ "step": 637
+ },
+ {
+ "epoch": 8.866375545851529,
+ "grad_norm": 0.021219318732619286,
+ "learning_rate": 0.0006,
+ "loss": 5.211699485778809,
+ "step": 638
+ },
+ {
+ "epoch": 8.880349344978166,
+ "grad_norm": 0.0200974028557539,
+ "learning_rate": 0.0006,
+ "loss": 5.092503547668457,
+ "step": 639
+ },
+ {
+ "epoch": 8.894323144104803,
+ "grad_norm": 0.023804882541298866,
+ "learning_rate": 0.0006,
+ "loss": 5.216068267822266,
+ "step": 640
+ },
+ {
+ "epoch": 8.908296943231441,
+ "grad_norm": 0.026088010519742966,
+ "learning_rate": 0.0006,
+ "loss": 5.155592918395996,
+ "step": 641
+ },
+ {
+ "epoch": 8.922270742358078,
+ "grad_norm": 0.02601276896893978,
+ "learning_rate": 0.0006,
+ "loss": 5.197238922119141,
+ "step": 642
+ },
+ {
+ "epoch": 8.936244541484717,
+ "grad_norm": 0.020387211814522743,
+ "learning_rate": 0.0006,
+ "loss": 5.1955437660217285,
+ "step": 643
+ },
+ {
+ "epoch": 8.950218340611354,
+ "grad_norm": 0.019214622676372528,
+ "learning_rate": 0.0006,
+ "loss": 5.206346035003662,
+ "step": 644
+ },
+ {
+ "epoch": 8.96419213973799,
+ "grad_norm": 0.019674314185976982,
+ "learning_rate": 0.0006,
+ "loss": 5.219857215881348,
+ "step": 645
+ },
+ {
+ "epoch": 8.97816593886463,
+ "grad_norm": 0.020318234339356422,
+ "learning_rate": 0.0006,
+ "loss": 5.191132545471191,
+ "step": 646
+ },
+ {
+ "epoch": 8.992139737991266,
+ "grad_norm": 0.021428676322102547,
+ "learning_rate": 0.0006,
+ "loss": 5.214695930480957,
+ "step": 647
+ },
+ {
+ "epoch": 9.0,
+ "grad_norm": 0.023235054686665535,
+ "learning_rate": 0.0006,
+ "loss": 5.229091644287109,
+ "step": 648
+ },
+ {
+ "epoch": 9.0,
+ "eval_loss": 5.289593696594238,
+ "eval_runtime": 56.9256,
+ "eval_samples_per_second": 42.898,
+ "eval_steps_per_second": 1.353,
+ "step": 648
+ },
+ {
+ "epoch": 9.013973799126637,
+ "grad_norm": 0.02320289984345436,
+ "learning_rate": 0.0006,
+ "loss": 4.983669281005859,
+ "step": 649
+ },
+ {
+ "epoch": 9.027947598253276,
+ "grad_norm": 0.018521282821893692,
+ "learning_rate": 0.0006,
+ "loss": 5.200575828552246,
+ "step": 650
+ },
+ {
+ "epoch": 9.041921397379912,
+ "grad_norm": 0.0180169939994812,
+ "learning_rate": 0.0006,
+ "loss": 5.148033618927002,
+ "step": 651
+ },
+ {
+ "epoch": 9.055895196506551,
+ "grad_norm": 0.019969860091805458,
+ "learning_rate": 0.0006,
+ "loss": 5.099125385284424,
+ "step": 652
+ },
+ {
+ "epoch": 9.069868995633188,
+ "grad_norm": 0.01643305830657482,
+ "learning_rate": 0.0006,
+ "loss": 5.092347145080566,
+ "step": 653
+ },
+ {
+ "epoch": 9.083842794759825,
+ "grad_norm": 0.016022363677620888,
+ "learning_rate": 0.0006,
+ "loss": 5.129279613494873,
+ "step": 654
+ },
+ {
+ "epoch": 9.097816593886463,
+ "grad_norm": 0.01657041162252426,
+ "learning_rate": 0.0006,
+ "loss": 5.159562587738037,
+ "step": 655
+ },
+ {
+ "epoch": 9.1117903930131,
+ "grad_norm": 0.017924228683114052,
+ "learning_rate": 0.0006,
+ "loss": 5.165497779846191,
+ "step": 656
+ },
+ {
+ "epoch": 9.125764192139737,
+ "grad_norm": 0.021592361852526665,
+ "learning_rate": 0.0006,
+ "loss": 5.204775810241699,
+ "step": 657
+ },
+ {
+ "epoch": 9.139737991266376,
+ "grad_norm": 0.026924120262265205,
+ "learning_rate": 0.0006,
+ "loss": 5.128296852111816,
+ "step": 658
+ },
+ {
+ "epoch": 9.153711790393013,
+ "grad_norm": 0.023156899958848953,
+ "learning_rate": 0.0006,
+ "loss": 5.1255340576171875,
+ "step": 659
+ },
+ {
+ "epoch": 9.167685589519651,
+ "grad_norm": 0.020670367404818535,
+ "learning_rate": 0.0006,
+ "loss": 5.128479957580566,
+ "step": 660
+ },
+ {
+ "epoch": 9.181659388646288,
+ "grad_norm": 0.025447173044085503,
+ "learning_rate": 0.0006,
+ "loss": 5.154821872711182,
+ "step": 661
+ },
+ {
+ "epoch": 9.195633187772925,
+ "grad_norm": 0.027940068393945694,
+ "learning_rate": 0.0006,
+ "loss": 5.143466472625732,
+ "step": 662
+ },
+ {
+ "epoch": 9.209606986899564,
+ "grad_norm": 0.026847844943404198,
+ "learning_rate": 0.0006,
+ "loss": 5.045665740966797,
+ "step": 663
+ },
+ {
+ "epoch": 9.2235807860262,
+ "grad_norm": 0.02340601570904255,
+ "learning_rate": 0.0006,
+ "loss": 5.21933126449585,
+ "step": 664
+ },
+ {
+ "epoch": 9.237554585152838,
+ "grad_norm": 0.02340371161699295,
+ "learning_rate": 0.0006,
+ "loss": 4.942949295043945,
+ "step": 665
+ },
+ {
+ "epoch": 9.251528384279476,
+ "grad_norm": 0.02221992425620556,
+ "learning_rate": 0.0006,
+ "loss": 5.197099208831787,
+ "step": 666
+ },
+ {
+ "epoch": 9.265502183406113,
+ "grad_norm": 0.023508677259087563,
+ "learning_rate": 0.0006,
+ "loss": 5.022897720336914,
+ "step": 667
+ },
+ {
+ "epoch": 9.279475982532752,
+ "grad_norm": 0.026120394468307495,
+ "learning_rate": 0.0006,
+ "loss": 5.087725639343262,
+ "step": 668
+ },
+ {
+ "epoch": 9.293449781659389,
+ "grad_norm": 0.027273228392004967,
+ "learning_rate": 0.0006,
+ "loss": 5.090963840484619,
+ "step": 669
+ },
+ {
+ "epoch": 9.307423580786025,
+ "grad_norm": 0.03241586685180664,
+ "learning_rate": 0.0006,
+ "loss": 5.047842979431152,
+ "step": 670
+ },
+ {
+ "epoch": 9.321397379912664,
+ "grad_norm": 0.030245667323470116,
+ "learning_rate": 0.0006,
+ "loss": 5.104803085327148,
+ "step": 671
+ },
+ {
+ "epoch": 9.335371179039301,
+ "grad_norm": 0.027698364108800888,
+ "learning_rate": 0.0006,
+ "loss": 5.0644097328186035,
+ "step": 672
+ },
+ {
+ "epoch": 9.34934497816594,
+ "grad_norm": 0.029692554846405983,
+ "learning_rate": 0.0006,
+ "loss": 5.111942291259766,
+ "step": 673
+ },
+ {
+ "epoch": 9.363318777292577,
+ "grad_norm": 0.03328656405210495,
+ "learning_rate": 0.0006,
+ "loss": 5.1574554443359375,
+ "step": 674
+ },
+ {
+ "epoch": 9.377292576419213,
+ "grad_norm": 0.031938180327415466,
+ "learning_rate": 0.0006,
+ "loss": 5.190371990203857,
+ "step": 675
+ },
+ {
+ "epoch": 9.391266375545852,
+ "grad_norm": 0.024858945980668068,
+ "learning_rate": 0.0006,
+ "loss": 5.121420860290527,
+ "step": 676
+ },
+ {
+ "epoch": 9.405240174672489,
+ "grad_norm": 0.023033203557133675,
+ "learning_rate": 0.0006,
+ "loss": 5.0974345207214355,
+ "step": 677
+ },
+ {
+ "epoch": 9.419213973799126,
+ "grad_norm": 0.022083545103669167,
+ "learning_rate": 0.0006,
+ "loss": 5.046319961547852,
+ "step": 678
+ },
+ {
+ "epoch": 9.433187772925764,
+ "grad_norm": 0.0228437427431345,
+ "learning_rate": 0.0006,
+ "loss": 5.119963645935059,
+ "step": 679
+ },
+ {
+ "epoch": 9.447161572052401,
+ "grad_norm": 0.0242351982742548,
+ "learning_rate": 0.0006,
+ "loss": 5.0444560050964355,
+ "step": 680
+ },
+ {
+ "epoch": 9.46113537117904,
+ "grad_norm": 0.02401185780763626,
+ "learning_rate": 0.0006,
+ "loss": 5.063238620758057,
+ "step": 681
+ },
+ {
+ "epoch": 9.475109170305677,
+ "grad_norm": 0.02560959942638874,
+ "learning_rate": 0.0006,
+ "loss": 5.035224914550781,
+ "step": 682
+ },
+ {
+ "epoch": 9.489082969432314,
+ "grad_norm": 0.024985626339912415,
+ "learning_rate": 0.0006,
+ "loss": 5.049284934997559,
+ "step": 683
+ },
+ {
+ "epoch": 9.503056768558952,
+ "grad_norm": 0.023178115487098694,
+ "learning_rate": 0.0006,
+ "loss": 5.2266621589660645,
+ "step": 684
+ },
+ {
+ "epoch": 9.51703056768559,
+ "grad_norm": 0.0215512253344059,
+ "learning_rate": 0.0006,
+ "loss": 5.1875715255737305,
+ "step": 685
+ },
+ {
+ "epoch": 9.531004366812226,
+ "grad_norm": 0.01662837713956833,
+ "learning_rate": 0.0006,
+ "loss": 5.137012481689453,
+ "step": 686
+ },
+ {
+ "epoch": 9.544978165938865,
+ "grad_norm": 0.01952764391899109,
+ "learning_rate": 0.0006,
+ "loss": 5.072500228881836,
+ "step": 687
+ },
+ {
+ "epoch": 9.558951965065502,
+ "grad_norm": 0.017521077767014503,
+ "learning_rate": 0.0006,
+ "loss": 5.045711994171143,
+ "step": 688
+ },
+ {
+ "epoch": 9.57292576419214,
+ "grad_norm": 0.017355265095829964,
+ "learning_rate": 0.0006,
+ "loss": 5.02881383895874,
+ "step": 689
+ },
+ {
+ "epoch": 9.586899563318777,
+ "grad_norm": 0.015587719157338142,
+ "learning_rate": 0.0006,
+ "loss": 5.132787704467773,
+ "step": 690
+ },
+ {
+ "epoch": 9.600873362445414,
+ "grad_norm": 0.017469845712184906,
+ "learning_rate": 0.0006,
+ "loss": 5.105447292327881,
+ "step": 691
+ },
+ {
+ "epoch": 9.614847161572053,
+ "grad_norm": 0.019630100578069687,
+ "learning_rate": 0.0006,
+ "loss": 5.032464981079102,
+ "step": 692
+ },
+ {
+ "epoch": 9.62882096069869,
+ "grad_norm": 0.01784994639456272,
+ "learning_rate": 0.0006,
+ "loss": 5.081345558166504,
+ "step": 693
+ },
+ {
+ "epoch": 9.642794759825328,
+ "grad_norm": 0.01872754842042923,
+ "learning_rate": 0.0006,
+ "loss": 4.9772114753723145,
+ "step": 694
+ },
+ {
+ "epoch": 9.656768558951965,
+ "grad_norm": 0.01999032311141491,
+ "learning_rate": 0.0006,
+ "loss": 5.089740753173828,
+ "step": 695
+ },
+ {
+ "epoch": 9.670742358078602,
+ "grad_norm": 0.01710767112672329,
+ "learning_rate": 0.0006,
+ "loss": 4.9869890213012695,
+ "step": 696
+ },
+ {
+ "epoch": 9.68471615720524,
+ "grad_norm": 0.015144161880016327,
+ "learning_rate": 0.0006,
+ "loss": 5.022564888000488,
+ "step": 697
+ },
+ {
+ "epoch": 9.698689956331878,
+ "grad_norm": 0.01618543453514576,
+ "learning_rate": 0.0006,
+ "loss": 5.035921573638916,
+ "step": 698
+ },
+ {
+ "epoch": 9.712663755458514,
+ "grad_norm": 0.014653387479484081,
+ "learning_rate": 0.0006,
+ "loss": 4.914456367492676,
+ "step": 699
+ },
+ {
+ "epoch": 9.726637554585153,
+ "grad_norm": 0.013434977270662785,
+ "learning_rate": 0.0006,
+ "loss": 5.063764572143555,
+ "step": 700
+ },
+ {
+ "epoch": 9.74061135371179,
+ "grad_norm": 0.015128864906728268,
+ "learning_rate": 0.0006,
+ "loss": 5.052367687225342,
+ "step": 701
+ },
+ {
+ "epoch": 9.754585152838429,
+ "grad_norm": 0.014210768043994904,
+ "learning_rate": 0.0006,
+ "loss": 5.016489028930664,
+ "step": 702
+ },
+ {
+ "epoch": 9.768558951965066,
+ "grad_norm": 0.015191680751740932,
+ "learning_rate": 0.0006,
+ "loss": 4.990176200866699,
+ "step": 703
+ },
+ {
+ "epoch": 9.782532751091702,
+ "grad_norm": 0.018285127356648445,
+ "learning_rate": 0.0006,
+ "loss": 5.015164375305176,
+ "step": 704
+ },
+ {
+ "epoch": 9.796506550218341,
+ "grad_norm": 0.0204079058021307,
+ "learning_rate": 0.0006,
+ "loss": 5.091142654418945,
+ "step": 705
+ },
+ {
+ "epoch": 9.810480349344978,
+ "grad_norm": 0.01980586163699627,
+ "learning_rate": 0.0006,
+ "loss": 5.03350830078125,
+ "step": 706
+ },
+ {
+ "epoch": 9.824454148471617,
+ "grad_norm": 0.02048729732632637,
+ "learning_rate": 0.0006,
+ "loss": 5.03611421585083,
+ "step": 707
+ },
+ {
+ "epoch": 9.838427947598253,
+ "grad_norm": 0.02058163844048977,
+ "learning_rate": 0.0006,
+ "loss": 4.9613447189331055,
+ "step": 708
+ },
+ {
+ "epoch": 9.85240174672489,
+ "grad_norm": 0.018004924058914185,
+ "learning_rate": 0.0006,
+ "loss": 4.966541767120361,
+ "step": 709
+ },
+ {
+ "epoch": 9.866375545851529,
+ "grad_norm": 0.0198152307420969,
+ "learning_rate": 0.0006,
+ "loss": 5.01750373840332,
+ "step": 710
+ },
+ {
+ "epoch": 9.880349344978166,
+ "grad_norm": 0.018532969057559967,
+ "learning_rate": 0.0006,
+ "loss": 5.097424507141113,
+ "step": 711
+ },
+ {
+ "epoch": 9.894323144104803,
+ "grad_norm": 0.018939971923828125,
+ "learning_rate": 0.0006,
+ "loss": 5.177223205566406,
+ "step": 712
+ },
+ {
+ "epoch": 9.908296943231441,
+ "grad_norm": 0.017969170585274696,
+ "learning_rate": 0.0006,
+ "loss": 5.118013858795166,
+ "step": 713
+ },
+ {
+ "epoch": 9.922270742358078,
+ "grad_norm": 0.018021270632743835,
+ "learning_rate": 0.0006,
+ "loss": 5.034787178039551,
+ "step": 714
+ },
+ {
+ "epoch": 9.936244541484717,
+ "grad_norm": 0.017009710893034935,
+ "learning_rate": 0.0006,
+ "loss": 5.003920555114746,
+ "step": 715
+ },
+ {
+ "epoch": 9.950218340611354,
+ "grad_norm": 0.017722049728035927,
+ "learning_rate": 0.0006,
+ "loss": 5.031939506530762,
+ "step": 716
+ },
+ {
+ "epoch": 9.96419213973799,
+ "grad_norm": 0.017134087160229683,
+ "learning_rate": 0.0006,
+ "loss": 5.093474388122559,
+ "step": 717
+ },
+ {
+ "epoch": 9.97816593886463,
+ "grad_norm": 0.018511613830924034,
+ "learning_rate": 0.0006,
+ "loss": 5.078604698181152,
+ "step": 718
+ },
+ {
+ "epoch": 9.992139737991266,
+ "grad_norm": 0.021206334233283997,
+ "learning_rate": 0.0006,
+ "loss": 5.052572727203369,
+ "step": 719
+ },
+ {
+ "epoch": 10.0,
+ "grad_norm": 0.02239062264561653,
+ "learning_rate": 0.0006,
+ "loss": 5.063547134399414,
+ "step": 720
+ },
+ {
+ "epoch": 10.0,
+ "eval_loss": 5.173044681549072,
+ "eval_runtime": 56.7287,
+ "eval_samples_per_second": 43.047,
+ "eval_steps_per_second": 1.357,
+ "step": 720
+ },
+ {
+ "epoch": 10.013973799126637,
+ "grad_norm": 0.020757969468832016,
+ "learning_rate": 0.0006,
+ "loss": 5.045648574829102,
+ "step": 721
+ },
+ {
+ "epoch": 10.027947598253276,
+ "grad_norm": 0.022697702050209045,
+ "learning_rate": 0.0006,
+ "loss": 4.993833065032959,
+ "step": 722
+ },
+ {
+ "epoch": 10.041921397379912,
+ "grad_norm": 0.02434331737458706,
+ "learning_rate": 0.0006,
+ "loss": 5.073219299316406,
+ "step": 723
+ },
+ {
+ "epoch": 10.055895196506551,
+ "grad_norm": 0.025499660521745682,
+ "learning_rate": 0.0006,
+ "loss": 5.019072532653809,
+ "step": 724
+ },
+ {
+ "epoch": 10.069868995633188,
+ "grad_norm": 0.029281053692102432,
+ "learning_rate": 0.0006,
+ "loss": 4.959851264953613,
+ "step": 725
+ },
+ {
+ "epoch": 10.083842794759825,
+ "grad_norm": 0.033878110349178314,
+ "learning_rate": 0.0006,
+ "loss": 5.132717609405518,
+ "step": 726
+ },
+ {
+ "epoch": 10.097816593886463,
+ "grad_norm": 0.03385764732956886,
+ "learning_rate": 0.0006,
+ "loss": 5.0482892990112305,
+ "step": 727
+ },
+ {
+ "epoch": 10.1117903930131,
+ "grad_norm": 0.03466648980975151,
+ "learning_rate": 0.0006,
+ "loss": 4.946855545043945,
+ "step": 728
+ },
+ {
+ "epoch": 10.125764192139737,
+ "grad_norm": 0.028546305373311043,
+ "learning_rate": 0.0006,
+ "loss": 5.027774333953857,
+ "step": 729
+ },
+ {
+ "epoch": 10.139737991266376,
+ "grad_norm": 0.030575869604945183,
+ "learning_rate": 0.0006,
+ "loss": 4.9768476486206055,
+ "step": 730
+ },
+ {
+ "epoch": 10.153711790393013,
+ "grad_norm": 0.03614303469657898,
+ "learning_rate": 0.0006,
+ "loss": 5.078113555908203,
+ "step": 731
+ },
+ {
+ "epoch": 10.167685589519651,
+ "grad_norm": 0.040991660207509995,
+ "learning_rate": 0.0006,
+ "loss": 5.033185958862305,
+ "step": 732
+ },
+ {
+ "epoch": 10.181659388646288,
+ "grad_norm": 0.04551699757575989,
+ "learning_rate": 0.0006,
+ "loss": 5.0820746421813965,
+ "step": 733
+ },
+ {
+ "epoch": 10.195633187772925,
+ "grad_norm": 0.040193989872932434,
+ "learning_rate": 0.0006,
+ "loss": 5.059700965881348,
+ "step": 734
+ },
+ {
+ "epoch": 10.209606986899564,
+ "grad_norm": 0.035851605236530304,
+ "learning_rate": 0.0006,
+ "loss": 5.142467498779297,
+ "step": 735
+ },
+ {
+ "epoch": 10.2235807860262,
+ "grad_norm": 0.034200169146060944,
+ "learning_rate": 0.0006,
+ "loss": 5.072312355041504,
+ "step": 736
+ },
+ {
+ "epoch": 10.237554585152838,
+ "grad_norm": 0.033621110022068024,
+ "learning_rate": 0.0006,
+ "loss": 5.08037805557251,
+ "step": 737
+ },
+ {
+ "epoch": 10.251528384279476,
+ "grad_norm": 0.037372369319200516,
+ "learning_rate": 0.0006,
+ "loss": 4.998425483703613,
+ "step": 738
+ },
+ {
+ "epoch": 10.265502183406113,
+ "grad_norm": 0.037044707685709,
+ "learning_rate": 0.0006,
+ "loss": 4.923993110656738,
+ "step": 739
+ },
+ {
+ "epoch": 10.279475982532752,
+ "grad_norm": 0.030454808846116066,
+ "learning_rate": 0.0006,
+ "loss": 5.0889997482299805,
+ "step": 740
+ },
+ {
+ "epoch": 10.293449781659389,
+ "grad_norm": 0.030969638377428055,
+ "learning_rate": 0.0006,
+ "loss": 4.984112739562988,
+ "step": 741
+ },
+ {
+ "epoch": 10.307423580786025,
+ "grad_norm": 0.0276983380317688,
+ "learning_rate": 0.0006,
+ "loss": 4.955412864685059,
+ "step": 742
+ },
+ {
+ "epoch": 10.321397379912664,
+ "grad_norm": 0.02716052532196045,
+ "learning_rate": 0.0006,
+ "loss": 5.0098876953125,
+ "step": 743
+ },
+ {
+ "epoch": 10.335371179039301,
+ "grad_norm": 0.02403552085161209,
+ "learning_rate": 0.0006,
+ "loss": 5.165160655975342,
+ "step": 744
+ },
+ {
+ "epoch": 10.34934497816594,
+ "grad_norm": 0.0257862601429224,
+ "learning_rate": 0.0006,
+ "loss": 5.001547813415527,
+ "step": 745
+ },
+ {
+ "epoch": 10.363318777292577,
+ "grad_norm": 0.02300378680229187,
+ "learning_rate": 0.0006,
+ "loss": 5.038888931274414,
+ "step": 746
+ },
+ {
+ "epoch": 10.377292576419213,
+ "grad_norm": 0.019766854122281075,
+ "learning_rate": 0.0006,
+ "loss": 4.924291610717773,
+ "step": 747
+ },
+ {
+ "epoch": 10.391266375545852,
+ "grad_norm": 0.01888395845890045,
+ "learning_rate": 0.0006,
+ "loss": 5.059971332550049,
+ "step": 748
+ },
+ {
+ "epoch": 10.405240174672489,
+ "grad_norm": 0.015577499754726887,
+ "learning_rate": 0.0006,
+ "loss": 4.987686634063721,
+ "step": 749
+ },
+ {
+ "epoch": 10.419213973799126,
+ "grad_norm": 0.016012346372008324,
+ "learning_rate": 0.0006,
+ "loss": 5.015157699584961,
+ "step": 750
+ },
+ {
+ "epoch": 10.433187772925764,
+ "grad_norm": 0.016277998685836792,
+ "learning_rate": 0.0006,
+ "loss": 5.004931449890137,
+ "step": 751
+ },
+ {
+ "epoch": 10.447161572052401,
+ "grad_norm": 0.017504561692476273,
+ "learning_rate": 0.0006,
+ "loss": 4.902827739715576,
+ "step": 752
+ },
+ {
+ "epoch": 10.46113537117904,
+ "grad_norm": 0.01569022797048092,
+ "learning_rate": 0.0006,
+ "loss": 4.954863548278809,
+ "step": 753
+ },
+ {
+ "epoch": 10.475109170305677,
+ "grad_norm": 0.014114422723650932,
+ "learning_rate": 0.0006,
+ "loss": 5.00656795501709,
+ "step": 754
+ },
+ {
+ "epoch": 10.489082969432314,
+ "grad_norm": 0.01573832333087921,
+ "learning_rate": 0.0006,
+ "loss": 4.978353023529053,
+ "step": 755
+ },
+ {
+ "epoch": 10.503056768558952,
+ "grad_norm": 0.016480036079883575,
+ "learning_rate": 0.0006,
+ "loss": 4.953327655792236,
+ "step": 756
+ },
+ {
+ "epoch": 10.51703056768559,
+ "grad_norm": 0.014008025638759136,
+ "learning_rate": 0.0006,
+ "loss": 5.050840377807617,
+ "step": 757
+ },
+ {
+ "epoch": 10.531004366812226,
+ "grad_norm": 0.013147998601198196,
+ "learning_rate": 0.0006,
+ "loss": 4.974964141845703,
+ "step": 758
+ },
+ {
+ "epoch": 10.544978165938865,
+ "grad_norm": 0.013807603158056736,
+ "learning_rate": 0.0006,
+ "loss": 4.927907943725586,
+ "step": 759
+ },
+ {
+ "epoch": 10.558951965065502,
+ "grad_norm": 0.013555224984884262,
+ "learning_rate": 0.0006,
+ "loss": 4.980690956115723,
+ "step": 760
+ },
+ {
+ "epoch": 10.57292576419214,
+ "grad_norm": 0.013045408762991428,
+ "learning_rate": 0.0006,
+ "loss": 4.9436116218566895,
+ "step": 761
+ },
+ {
+ "epoch": 10.586899563318777,
+ "grad_norm": 0.012858862057328224,
+ "learning_rate": 0.0006,
+ "loss": 4.9595770835876465,
+ "step": 762
+ },
+ {
+ "epoch": 10.600873362445414,
+ "grad_norm": 0.01468253880739212,
+ "learning_rate": 0.0006,
+ "loss": 4.90725040435791,
+ "step": 763
+ },
+ {
+ "epoch": 10.614847161572053,
+ "grad_norm": 0.013224679045379162,
+ "learning_rate": 0.0006,
+ "loss": 4.877615928649902,
+ "step": 764
+ },
+ {
+ "epoch": 10.62882096069869,
+ "grad_norm": 0.013555348850786686,
+ "learning_rate": 0.0006,
+ "loss": 4.9456610679626465,
+ "step": 765
+ },
+ {
+ "epoch": 10.642794759825328,
+ "grad_norm": 0.014387160539627075,
+ "learning_rate": 0.0006,
+ "loss": 4.9233717918396,
+ "step": 766
+ },
+ {
+ "epoch": 10.656768558951965,
+ "grad_norm": 0.014354088343679905,
+ "learning_rate": 0.0006,
+ "loss": 4.983511924743652,
+ "step": 767
+ },
+ {
+ "epoch": 10.670742358078602,
+ "grad_norm": 0.01332700252532959,
+ "learning_rate": 0.0006,
+ "loss": 4.934416770935059,
+ "step": 768
+ },
+ {
+ "epoch": 10.68471615720524,
+ "grad_norm": 0.011543313041329384,
+ "learning_rate": 0.0006,
+ "loss": 4.96388053894043,
+ "step": 769
+ },
+ {
+ "epoch": 10.698689956331878,
+ "grad_norm": 0.012330878525972366,
+ "learning_rate": 0.0006,
+ "loss": 4.946396350860596,
+ "step": 770
+ },
+ {
+ "epoch": 10.712663755458514,
+ "grad_norm": 0.01379779726266861,
+ "learning_rate": 0.0006,
+ "loss": 4.981019020080566,
+ "step": 771
+ },
+ {
+ "epoch": 10.726637554585153,
+ "grad_norm": 0.014612431637942791,
+ "learning_rate": 0.0006,
+ "loss": 4.894567966461182,
+ "step": 772
+ },
+ {
+ "epoch": 10.74061135371179,
+ "grad_norm": 0.013062899932265282,
+ "learning_rate": 0.0006,
+ "loss": 5.033038139343262,
+ "step": 773
+ },
+ {
+ "epoch": 10.754585152838429,
+ "grad_norm": 0.013898049481213093,
+ "learning_rate": 0.0006,
+ "loss": 4.823636054992676,
+ "step": 774
+ },
+ {
+ "epoch": 10.768558951965066,
+ "grad_norm": 0.01502019353210926,
+ "learning_rate": 0.0006,
+ "loss": 5.020156383514404,
+ "step": 775
+ },
+ {
+ "epoch": 10.782532751091702,
+ "grad_norm": 0.013810204342007637,
+ "learning_rate": 0.0006,
+ "loss": 4.935647010803223,
+ "step": 776
+ },
+ {
+ "epoch": 10.796506550218341,
+ "grad_norm": 0.013616513460874557,
+ "learning_rate": 0.0006,
+ "loss": 4.901486396789551,
+ "step": 777
+ },
+ {
+ "epoch": 10.810480349344978,
+ "grad_norm": 0.013601024635136127,
+ "learning_rate": 0.0006,
+ "loss": 4.931873321533203,
+ "step": 778
+ },
+ {
+ "epoch": 10.824454148471617,
+ "grad_norm": 0.013668350875377655,
+ "learning_rate": 0.0006,
+ "loss": 4.937256813049316,
+ "step": 779
+ },
+ {
+ "epoch": 10.838427947598253,
+ "grad_norm": 0.01581823080778122,
+ "learning_rate": 0.0006,
+ "loss": 4.9611406326293945,
+ "step": 780
+ },
+ {
+ "epoch": 10.85240174672489,
+ "grad_norm": 0.019478484988212585,
+ "learning_rate": 0.0006,
+ "loss": 4.9212236404418945,
+ "step": 781
+ },
+ {
+ "epoch": 10.866375545851529,
+ "grad_norm": 0.026921333745121956,
+ "learning_rate": 0.0006,
+ "loss": 4.88501501083374,
+ "step": 782
+ },
+ {
+ "epoch": 10.880349344978166,
+ "grad_norm": 0.03155914694070816,
+ "learning_rate": 0.0006,
+ "loss": 4.989967346191406,
+ "step": 783
+ },
+ {
+ "epoch": 10.894323144104803,
+ "grad_norm": 0.029689429327845573,
+ "learning_rate": 0.0006,
+ "loss": 4.936000823974609,
+ "step": 784
+ },
+ {
+ "epoch": 10.908296943231441,
+ "grad_norm": 0.026626063510775566,
+ "learning_rate": 0.0006,
+ "loss": 4.984785079956055,
+ "step": 785
+ },
+ {
+ "epoch": 10.922270742358078,
+ "grad_norm": 0.026168212294578552,
+ "learning_rate": 0.0006,
+ "loss": 4.908282279968262,
+ "step": 786
+ },
+ {
+ "epoch": 10.936244541484717,
+ "grad_norm": 0.02828158251941204,
+ "learning_rate": 0.0006,
+ "loss": 4.926390647888184,
+ "step": 787
+ },
+ {
+ "epoch": 10.950218340611354,
+ "grad_norm": 0.02649078331887722,
+ "learning_rate": 0.0006,
+ "loss": 5.00400972366333,
+ "step": 788
+ },
+ {
+ "epoch": 10.96419213973799,
+ "grad_norm": 0.0256856270134449,
+ "learning_rate": 0.0006,
+ "loss": 5.0317487716674805,
+ "step": 789
+ },
+ {
+ "epoch": 10.97816593886463,
+ "grad_norm": 0.024728883057832718,
+ "learning_rate": 0.0006,
+ "loss": 4.9299635887146,
+ "step": 790
+ },
+ {
+ "epoch": 10.992139737991266,
+ "grad_norm": 0.026241116225719452,
+ "learning_rate": 0.0006,
+ "loss": 4.947319984436035,
+ "step": 791
+ },
+ {
+ "epoch": 11.0,
+ "grad_norm": 0.023526743054389954,
+ "learning_rate": 0.0006,
+ "loss": 4.94429349899292,
+ "step": 792
+ },
+ {
+ "epoch": 11.0,
+ "eval_loss": 5.104345798492432,
+ "eval_runtime": 57.0955,
+ "eval_samples_per_second": 42.77,
+ "eval_steps_per_second": 1.349,
+ "step": 792
+ },
+ {
+ "epoch": 11.013973799126637,
+ "grad_norm": 0.0216788612306118,
+ "learning_rate": 0.0006,
+ "loss": 4.881950855255127,
+ "step": 793
+ },
+ {
+ "epoch": 11.027947598253276,
+ "grad_norm": 0.01830463856458664,
+ "learning_rate": 0.0006,
+ "loss": 4.971016883850098,
+ "step": 794
+ },
+ {
+ "epoch": 11.041921397379912,
+ "grad_norm": 0.02204521745443344,
+ "learning_rate": 0.0006,
+ "loss": 4.894623756408691,
+ "step": 795
+ },
+ {
+ "epoch": 11.055895196506551,
+ "grad_norm": 0.0210255216807127,
+ "learning_rate": 0.0006,
+ "loss": 4.834009170532227,
+ "step": 796
+ },
+ {
+ "epoch": 11.069868995633188,
+ "grad_norm": 0.019797448068857193,
+ "learning_rate": 0.0006,
+ "loss": 5.086714744567871,
+ "step": 797
+ },
+ {
+ "epoch": 11.083842794759825,
+ "grad_norm": 0.02036641724407673,
+ "learning_rate": 0.0006,
+ "loss": 4.920056343078613,
+ "step": 798
+ },
+ {
+ "epoch": 11.097816593886463,
+ "grad_norm": 0.01966066285967827,
+ "learning_rate": 0.0006,
+ "loss": 4.879859447479248,
+ "step": 799
+ },
+ {
+ "epoch": 11.1117903930131,
+ "grad_norm": 0.0202149897813797,
+ "learning_rate": 0.0006,
+ "loss": 4.933640480041504,
+ "step": 800
+ },
+ {
+ "epoch": 11.125764192139737,
+ "grad_norm": 0.01893039606511593,
+ "learning_rate": 0.0006,
+ "loss": 4.901970863342285,
+ "step": 801
+ },
+ {
+ "epoch": 11.139737991266376,
+ "grad_norm": 0.016040155664086342,
+ "learning_rate": 0.0006,
+ "loss": 4.905491828918457,
+ "step": 802
+ },
+ {
+ "epoch": 11.153711790393013,
+ "grad_norm": 0.01720350608229637,
+ "learning_rate": 0.0006,
+ "loss": 4.798816680908203,
+ "step": 803
+ },
+ {
+ "epoch": 11.167685589519651,
+ "grad_norm": 0.017002852633595467,
+ "learning_rate": 0.0006,
+ "loss": 4.851859092712402,
+ "step": 804
+ },
+ {
+ "epoch": 11.181659388646288,
+ "grad_norm": 0.015594680793583393,
+ "learning_rate": 0.0006,
+ "loss": 4.90328311920166,
+ "step": 805
+ },
+ {
+ "epoch": 11.195633187772925,
+ "grad_norm": 0.014904462732374668,
+ "learning_rate": 0.0006,
+ "loss": 4.904215335845947,
+ "step": 806
+ },
+ {
+ "epoch": 11.209606986899564,
+ "grad_norm": 0.01366228237748146,
+ "learning_rate": 0.0006,
+ "loss": 4.9813055992126465,
+ "step": 807
+ },
+ {
+ "epoch": 11.2235807860262,
+ "grad_norm": 0.014719638973474503,
+ "learning_rate": 0.0006,
+ "loss": 4.927306175231934,
+ "step": 808
+ },
+ {
+ "epoch": 11.237554585152838,
+ "grad_norm": 0.015016036108136177,
+ "learning_rate": 0.0006,
+ "loss": 4.874215126037598,
+ "step": 809
+ },
+ {
+ "epoch": 11.251528384279476,
+ "grad_norm": 0.014925424009561539,
+ "learning_rate": 0.0006,
+ "loss": 4.838257789611816,
+ "step": 810
+ },
+ {
+ "epoch": 11.265502183406113,
+ "grad_norm": 0.014335823245346546,
+ "learning_rate": 0.0006,
+ "loss": 4.940608024597168,
+ "step": 811
+ },
+ {
+ "epoch": 11.279475982532752,
+ "grad_norm": 0.014987512491643429,
+ "learning_rate": 0.0006,
+ "loss": 4.831531524658203,
+ "step": 812
+ },
+ {
+ "epoch": 11.293449781659389,
+ "grad_norm": 0.016087457537651062,
+ "learning_rate": 0.0006,
+ "loss": 4.863227367401123,
+ "step": 813
+ },
+ {
+ "epoch": 11.307423580786025,
+ "grad_norm": 0.01671300269663334,
+ "learning_rate": 0.0006,
+ "loss": 4.8535661697387695,
+ "step": 814
+ },
+ {
+ "epoch": 11.321397379912664,
+ "grad_norm": 0.015229142270982265,
+ "learning_rate": 0.0006,
+ "loss": 4.948635101318359,
+ "step": 815
+ },
+ {
+ "epoch": 11.335371179039301,
+ "grad_norm": 0.01532587967813015,
+ "learning_rate": 0.0006,
+ "loss": 4.9035797119140625,
+ "step": 816
+ },
+ {
+ "epoch": 11.34934497816594,
+ "grad_norm": 0.015715476125478745,
+ "learning_rate": 0.0006,
+ "loss": 4.894709587097168,
+ "step": 817
+ },
+ {
+ "epoch": 11.363318777292577,
+ "grad_norm": 0.014156977646052837,
+ "learning_rate": 0.0006,
+ "loss": 4.960206985473633,
+ "step": 818
+ },
+ {
+ "epoch": 11.377292576419213,
+ "grad_norm": 0.017076566815376282,
+ "learning_rate": 0.0006,
+ "loss": 5.005850791931152,
+ "step": 819
+ },
+ {
+ "epoch": 11.391266375545852,
+ "grad_norm": 0.0172084029763937,
+ "learning_rate": 0.0006,
+ "loss": 4.903355598449707,
+ "step": 820
+ },
+ {
+ "epoch": 11.405240174672489,
+ "grad_norm": 0.0180258359760046,
+ "learning_rate": 0.0006,
+ "loss": 4.842667579650879,
+ "step": 821
+ },
+ {
+ "epoch": 11.419213973799126,
+ "grad_norm": 0.022130120545625687,
+ "learning_rate": 0.0006,
+ "loss": 4.799970626831055,
+ "step": 822
+ },
+ {
+ "epoch": 11.433187772925764,
+ "grad_norm": 0.02384897693991661,
+ "learning_rate": 0.0006,
+ "loss": 4.860746383666992,
+ "step": 823
+ },
+ {
+ "epoch": 11.447161572052401,
+ "grad_norm": 0.021881572902202606,
+ "learning_rate": 0.0006,
+ "loss": 4.918107032775879,
+ "step": 824
+ },
+ {
+ "epoch": 11.46113537117904,
+ "grad_norm": 0.023085080087184906,
+ "learning_rate": 0.0006,
+ "loss": 4.878946304321289,
+ "step": 825
+ },
+ {
+ "epoch": 11.475109170305677,
+ "grad_norm": 0.02291012555360794,
+ "learning_rate": 0.0006,
+ "loss": 4.958887100219727,
+ "step": 826
+ },
+ {
+ "epoch": 11.489082969432314,
+ "grad_norm": 0.0199968870729208,
+ "learning_rate": 0.0006,
+ "loss": 4.80955696105957,
+ "step": 827
+ },
+ {
+ "epoch": 11.503056768558952,
+ "grad_norm": 0.018702229484915733,
+ "learning_rate": 0.0006,
+ "loss": 5.0276408195495605,
+ "step": 828
+ },
+ {
+ "epoch": 11.51703056768559,
+ "grad_norm": 0.017787311226129532,
+ "learning_rate": 0.0006,
+ "loss": 4.96018123626709,
+ "step": 829
+ },
+ {
+ "epoch": 11.531004366812226,
+ "grad_norm": 0.01704501546919346,
+ "learning_rate": 0.0006,
+ "loss": 4.914553165435791,
+ "step": 830
+ },
+ {
+ "epoch": 11.544978165938865,
+ "grad_norm": 0.017519677057862282,
+ "learning_rate": 0.0006,
+ "loss": 4.798098087310791,
+ "step": 831
+ },
+ {
+ "epoch": 11.558951965065502,
+ "grad_norm": 0.018511991947889328,
+ "learning_rate": 0.0006,
+ "loss": 4.817785263061523,
+ "step": 832
+ },
+ {
+ "epoch": 11.57292576419214,
+ "grad_norm": 0.01876644790172577,
+ "learning_rate": 0.0006,
+ "loss": 4.818995475769043,
+ "step": 833
+ },
+ {
+ "epoch": 11.586899563318777,
+ "grad_norm": 0.01753568835556507,
+ "learning_rate": 0.0006,
+ "loss": 4.811938285827637,
+ "step": 834
+ },
+ {
+ "epoch": 11.600873362445414,
+ "grad_norm": 0.017046531662344933,
+ "learning_rate": 0.0006,
+ "loss": 4.914674282073975,
+ "step": 835
+ },
+ {
+ "epoch": 11.614847161572053,
+ "grad_norm": 0.01802094094455242,
+ "learning_rate": 0.0006,
+ "loss": 4.993284225463867,
+ "step": 836
+ },
+ {
+ "epoch": 11.62882096069869,
+ "grad_norm": 0.019493764266371727,
+ "learning_rate": 0.0006,
+ "loss": 4.895628452301025,
+ "step": 837
+ },
+ {
+ "epoch": 11.642794759825328,
+ "grad_norm": 0.019461210817098618,
+ "learning_rate": 0.0006,
+ "loss": 4.869269371032715,
+ "step": 838
+ },
+ {
+ "epoch": 11.656768558951965,
+ "grad_norm": 0.01981567218899727,
+ "learning_rate": 0.0006,
+ "loss": 4.863052845001221,
+ "step": 839
+ },
+ {
+ "epoch": 11.670742358078602,
+ "grad_norm": 0.019894491881132126,
+ "learning_rate": 0.0006,
+ "loss": 4.8775787353515625,
+ "step": 840
+ },
+ {
+ "epoch": 11.68471615720524,
+ "grad_norm": 0.01875786855816841,
+ "learning_rate": 0.0006,
+ "loss": 4.817763328552246,
+ "step": 841
+ },
+ {
+ "epoch": 11.698689956331878,
+ "grad_norm": 0.020926695317029953,
+ "learning_rate": 0.0006,
+ "loss": 4.7470526695251465,
+ "step": 842
+ },
+ {
+ "epoch": 11.712663755458514,
+ "grad_norm": 0.021477103233337402,
+ "learning_rate": 0.0006,
+ "loss": 4.762682914733887,
+ "step": 843
+ },
+ {
+ "epoch": 11.726637554585153,
+ "grad_norm": 0.020628711208701134,
+ "learning_rate": 0.0006,
+ "loss": 4.807985782623291,
+ "step": 844
+ },
+ {
+ "epoch": 11.74061135371179,
+ "grad_norm": 0.021489612758159637,
+ "learning_rate": 0.0006,
+ "loss": 4.843091011047363,
+ "step": 845
+ },
+ {
+ "epoch": 11.754585152838429,
+ "grad_norm": 0.021194491535425186,
+ "learning_rate": 0.0006,
+ "loss": 4.8300251960754395,
+ "step": 846
+ },
+ {
+ "epoch": 11.768558951965066,
+ "grad_norm": 0.01833350397646427,
+ "learning_rate": 0.0006,
+ "loss": 5.000202178955078,
+ "step": 847
+ },
+ {
+ "epoch": 11.782532751091702,
+ "grad_norm": 0.018096931278705597,
+ "learning_rate": 0.0006,
+ "loss": 4.880648136138916,
+ "step": 848
+ },
+ {
+ "epoch": 11.796506550218341,
+ "grad_norm": 0.0197161752730608,
+ "learning_rate": 0.0006,
+ "loss": 4.788316249847412,
+ "step": 849
+ },
+ {
+ "epoch": 11.810480349344978,
+ "grad_norm": 0.0203181654214859,
+ "learning_rate": 0.0006,
+ "loss": 4.870944976806641,
+ "step": 850
+ },
+ {
+ "epoch": 11.824454148471617,
+ "grad_norm": 0.01908455230295658,
+ "learning_rate": 0.0006,
+ "loss": 4.890144348144531,
+ "step": 851
+ },
+ {
+ "epoch": 11.838427947598253,
+ "grad_norm": 0.021281801164150238,
+ "learning_rate": 0.0006,
+ "loss": 4.89578914642334,
+ "step": 852
+ },
+ {
+ "epoch": 11.85240174672489,
+ "grad_norm": 0.02124273031949997,
+ "learning_rate": 0.0006,
+ "loss": 4.883749008178711,
+ "step": 853
+ },
+ {
+ "epoch": 11.866375545851529,
+ "grad_norm": 0.020969398319721222,
+ "learning_rate": 0.0006,
+ "loss": 4.809142589569092,
+ "step": 854
+ },
+ {
+ "epoch": 11.880349344978166,
+ "grad_norm": 0.01967657171189785,
+ "learning_rate": 0.0006,
+ "loss": 4.821864604949951,
+ "step": 855
+ },
+ {
+ "epoch": 11.894323144104803,
+ "grad_norm": 0.020887333899736404,
+ "learning_rate": 0.0006,
+ "loss": 4.917985916137695,
+ "step": 856
+ },
+ {
+ "epoch": 11.908296943231441,
+ "grad_norm": 0.019203083589673042,
+ "learning_rate": 0.0006,
+ "loss": 4.824960708618164,
+ "step": 857
+ },
+ {
+ "epoch": 11.922270742358078,
+ "grad_norm": 0.020162401720881462,
+ "learning_rate": 0.0006,
+ "loss": 4.772617816925049,
+ "step": 858
+ },
+ {
+ "epoch": 11.936244541484717,
+ "grad_norm": 0.02251487225294113,
+ "learning_rate": 0.0006,
+ "loss": 4.735418319702148,
+ "step": 859
+ },
+ {
+ "epoch": 11.950218340611354,
+ "grad_norm": 0.020313527435064316,
+ "learning_rate": 0.0006,
+ "loss": 4.939233779907227,
+ "step": 860
+ },
+ {
+ "epoch": 11.96419213973799,
+ "grad_norm": 0.021635599434375763,
+ "learning_rate": 0.0006,
+ "loss": 4.861807346343994,
+ "step": 861
+ },
+ {
+ "epoch": 11.97816593886463,
+ "grad_norm": 0.02285311557352543,
+ "learning_rate": 0.0006,
+ "loss": 4.932732582092285,
+ "step": 862
+ },
+ {
+ "epoch": 11.992139737991266,
+ "grad_norm": 0.02072535641491413,
+ "learning_rate": 0.0006,
+ "loss": 4.796767711639404,
+ "step": 863
+ },
+ {
+ "epoch": 12.0,
+ "grad_norm": 0.020588304847478867,
+ "learning_rate": 0.0006,
+ "loss": 4.997537612915039,
+ "step": 864
+ },
+ {
+ "epoch": 12.0,
+ "eval_loss": 5.044651985168457,
+ "eval_runtime": 56.9145,
+ "eval_samples_per_second": 42.906,
+ "eval_steps_per_second": 1.353,
+ "step": 864
+ },
+ {
+ "epoch": 12.013973799126637,
+ "grad_norm": 0.017597166821360588,
+ "learning_rate": 0.0006,
+ "loss": 4.830524444580078,
+ "step": 865
+ },
+ {
+ "epoch": 12.027947598253276,
+ "grad_norm": 0.01716788113117218,
+ "learning_rate": 0.0006,
+ "loss": 4.894325256347656,
+ "step": 866
+ },
+ {
+ "epoch": 12.041921397379912,
+ "grad_norm": 0.017712457105517387,
+ "learning_rate": 0.0006,
+ "loss": 4.868297576904297,
+ "step": 867
+ },
+ {
+ "epoch": 12.055895196506551,
+ "grad_norm": 0.021486390382051468,
+ "learning_rate": 0.0006,
+ "loss": 4.922049522399902,
+ "step": 868
+ },
+ {
+ "epoch": 12.069868995633188,
+ "grad_norm": 0.024560654535889626,
+ "learning_rate": 0.0006,
+ "loss": 4.734241008758545,
+ "step": 869
+ },
+ {
+ "epoch": 12.083842794759825,
+ "grad_norm": 0.025870252400636673,
+ "learning_rate": 0.0006,
+ "loss": 4.868566989898682,
+ "step": 870
+ },
+ {
+ "epoch": 12.097816593886463,
+ "grad_norm": 0.0277389045804739,
+ "learning_rate": 0.0006,
+ "loss": 4.755006790161133,
+ "step": 871
+ },
+ {
+ "epoch": 12.1117903930131,
+ "grad_norm": 0.029100844636559486,
+ "learning_rate": 0.0006,
+ "loss": 4.909667015075684,
+ "step": 872
+ },
+ {
+ "epoch": 12.125764192139737,
+ "grad_norm": 0.031833067536354065,
+ "learning_rate": 0.0006,
+ "loss": 4.858674049377441,
+ "step": 873
+ },
+ {
+ "epoch": 12.139737991266376,
+ "grad_norm": 0.032709237188100815,
+ "learning_rate": 0.0006,
+ "loss": 4.823696136474609,
+ "step": 874
+ },
+ {
+ "epoch": 12.153711790393013,
+ "grad_norm": 0.02995380386710167,
+ "learning_rate": 0.0006,
+ "loss": 4.850235462188721,
+ "step": 875
+ },
+ {
+ "epoch": 12.167685589519651,
+ "grad_norm": 0.034098610281944275,
+ "learning_rate": 0.0006,
+ "loss": 4.8169989585876465,
+ "step": 876
+ },
+ {
+ "epoch": 12.181659388646288,
+ "grad_norm": 0.03377080336213112,
+ "learning_rate": 0.0006,
+ "loss": 4.90738582611084,
+ "step": 877
+ },
+ {
+ "epoch": 12.195633187772925,
+ "grad_norm": 0.029780825600028038,
+ "learning_rate": 0.0006,
+ "loss": 4.811163902282715,
+ "step": 878
+ },
+ {
+ "epoch": 12.209606986899564,
+ "grad_norm": 0.03118024580180645,
+ "learning_rate": 0.0006,
+ "loss": 4.861286640167236,
+ "step": 879
+ },
+ {
+ "epoch": 12.2235807860262,
+ "grad_norm": 0.033052004873752594,
+ "learning_rate": 0.0006,
+ "loss": 4.865021228790283,
+ "step": 880
+ },
+ {
+ "epoch": 12.237554585152838,
+ "grad_norm": 0.029070032760500908,
+ "learning_rate": 0.0006,
+ "loss": 4.947303771972656,
+ "step": 881
+ },
+ {
+ "epoch": 12.251528384279476,
+ "grad_norm": 0.028355715796351433,
+ "learning_rate": 0.0006,
+ "loss": 4.804648399353027,
+ "step": 882
+ },
+ {
+ "epoch": 12.265502183406113,
+ "grad_norm": 0.031257227063179016,
+ "learning_rate": 0.0006,
+ "loss": 4.757718086242676,
+ "step": 883
+ },
+ {
+ "epoch": 12.279475982532752,
+ "grad_norm": 0.031438373029232025,
+ "learning_rate": 0.0006,
+ "loss": 4.851268768310547,
+ "step": 884
+ },
+ {
+ "epoch": 12.293449781659389,
+ "grad_norm": 0.030492138117551804,
+ "learning_rate": 0.0006,
+ "loss": 4.7891035079956055,
+ "step": 885
+ },
+ {
+ "epoch": 12.307423580786025,
+ "grad_norm": 0.026115991175174713,
+ "learning_rate": 0.0006,
+ "loss": 4.86474609375,
+ "step": 886
+ },
+ {
+ "epoch": 12.321397379912664,
+ "grad_norm": 0.024184465408325195,
+ "learning_rate": 0.0006,
+ "loss": 4.860456943511963,
+ "step": 887
+ },
+ {
+ "epoch": 12.335371179039301,
+ "grad_norm": 0.025437982752919197,
+ "learning_rate": 0.0006,
+ "loss": 4.806251049041748,
+ "step": 888
+ },
+ {
+ "epoch": 12.34934497816594,
+ "grad_norm": 0.022390231490135193,
+ "learning_rate": 0.0006,
+ "loss": 4.807086944580078,
+ "step": 889
+ },
+ {
+ "epoch": 12.363318777292577,
+ "grad_norm": 0.020174294710159302,
+ "learning_rate": 0.0006,
+ "loss": 4.78292179107666,
+ "step": 890
+ },
+ {
+ "epoch": 12.377292576419213,
+ "grad_norm": 0.019734324887394905,
+ "learning_rate": 0.0006,
+ "loss": 4.784509658813477,
+ "step": 891
+ },
+ {
+ "epoch": 12.391266375545852,
+ "grad_norm": 0.016733523458242416,
+ "learning_rate": 0.0006,
+ "loss": 4.811755180358887,
+ "step": 892
+ },
+ {
+ "epoch": 12.405240174672489,
+ "grad_norm": 0.01570146717131138,
+ "learning_rate": 0.0006,
+ "loss": 4.720088005065918,
+ "step": 893
+ },
+ {
+ "epoch": 12.419213973799126,
+ "grad_norm": 0.01601393334567547,
+ "learning_rate": 0.0006,
+ "loss": 4.783543109893799,
+ "step": 894
+ },
+ {
+ "epoch": 12.433187772925764,
+ "grad_norm": 0.014035487547516823,
+ "learning_rate": 0.0006,
+ "loss": 4.848340034484863,
+ "step": 895
+ },
+ {
+ "epoch": 12.447161572052401,
+ "grad_norm": 0.015248487703502178,
+ "learning_rate": 0.0006,
+ "loss": 4.773287773132324,
+ "step": 896
+ },
+ {
+ "epoch": 12.46113537117904,
+ "grad_norm": 0.01532268151640892,
+ "learning_rate": 0.0006,
+ "loss": 4.799646854400635,
+ "step": 897
+ },
+ {
+ "epoch": 12.475109170305677,
+ "grad_norm": 0.013982534408569336,
+ "learning_rate": 0.0006,
+ "loss": 4.8090500831604,
+ "step": 898
+ },
+ {
+ "epoch": 12.489082969432314,
+ "grad_norm": 0.014081164263188839,
+ "learning_rate": 0.0006,
+ "loss": 4.751888751983643,
+ "step": 899
+ },
+ {
+ "epoch": 12.503056768558952,
+ "grad_norm": 0.01380133256316185,
+ "learning_rate": 0.0006,
+ "loss": 4.78144645690918,
+ "step": 900
+ },
+ {
+ "epoch": 12.51703056768559,
+ "grad_norm": 0.013555269688367844,
+ "learning_rate": 0.0006,
+ "loss": 4.779257774353027,
+ "step": 901
+ },
+ {
+ "epoch": 12.531004366812226,
+ "grad_norm": 0.013075701892375946,
+ "learning_rate": 0.0006,
+ "loss": 4.820705413818359,
+ "step": 902
+ },
+ {
+ "epoch": 12.544978165938865,
+ "grad_norm": 0.01276316400617361,
+ "learning_rate": 0.0006,
+ "loss": 4.836733818054199,
+ "step": 903
+ },
+ {
+ "epoch": 12.558951965065502,
+ "grad_norm": 0.013033892959356308,
+ "learning_rate": 0.0006,
+ "loss": 4.764178276062012,
+ "step": 904
+ },
+ {
+ "epoch": 12.57292576419214,
+ "grad_norm": 0.013996980153024197,
+ "learning_rate": 0.0006,
+ "loss": 4.767983436584473,
+ "step": 905
+ },
+ {
+ "epoch": 12.586899563318777,
+ "grad_norm": 0.015222184360027313,
+ "learning_rate": 0.0006,
+ "loss": 4.759721755981445,
+ "step": 906
+ },
+ {
+ "epoch": 12.600873362445414,
+ "grad_norm": 0.01623941957950592,
+ "learning_rate": 0.0006,
+ "loss": 4.849811553955078,
+ "step": 907
+ },
+ {
+ "epoch": 12.614847161572053,
+ "grad_norm": 0.015711182728409767,
+ "learning_rate": 0.0006,
+ "loss": 4.7829084396362305,
+ "step": 908
+ },
+ {
+ "epoch": 12.62882096069869,
+ "grad_norm": 0.014137467369437218,
+ "learning_rate": 0.0006,
+ "loss": 4.73321533203125,
+ "step": 909
+ },
+ {
+ "epoch": 12.642794759825328,
+ "grad_norm": 0.012738462537527084,
+ "learning_rate": 0.0006,
+ "loss": 4.870052337646484,
+ "step": 910
+ },
+ {
+ "epoch": 12.656768558951965,
+ "grad_norm": 0.013683130033314228,
+ "learning_rate": 0.0006,
+ "loss": 4.82658052444458,
+ "step": 911
+ },
+ {
+ "epoch": 12.670742358078602,
+ "grad_norm": 0.014409830793738365,
+ "learning_rate": 0.0006,
+ "loss": 4.707494735717773,
+ "step": 912
+ },
+ {
+ "epoch": 12.68471615720524,
+ "grad_norm": 0.013806809671223164,
+ "learning_rate": 0.0006,
+ "loss": 4.778928279876709,
+ "step": 913
+ },
+ {
+ "epoch": 12.698689956331878,
+ "grad_norm": 0.012857912108302116,
+ "learning_rate": 0.0006,
+ "loss": 4.890717506408691,
+ "step": 914
+ },
+ {
+ "epoch": 12.712663755458514,
+ "grad_norm": 0.013828758150339127,
+ "learning_rate": 0.0006,
+ "loss": 4.8357439041137695,
+ "step": 915
+ },
+ {
+ "epoch": 12.726637554585153,
+ "grad_norm": 0.01478351280093193,
+ "learning_rate": 0.0006,
+ "loss": 4.869986534118652,
+ "step": 916
+ },
+ {
+ "epoch": 12.74061135371179,
+ "grad_norm": 0.01326421182602644,
+ "learning_rate": 0.0006,
+ "loss": 4.830646514892578,
+ "step": 917
+ },
+ {
+ "epoch": 12.754585152838429,
+ "grad_norm": 0.013335433788597584,
+ "learning_rate": 0.0006,
+ "loss": 4.7786149978637695,
+ "step": 918
+ },
+ {
+ "epoch": 12.768558951965066,
+ "grad_norm": 0.014261147007346153,
+ "learning_rate": 0.0006,
+ "loss": 4.788693428039551,
+ "step": 919
+ },
+ {
+ "epoch": 12.782532751091702,
+ "grad_norm": 0.014274783432483673,
+ "learning_rate": 0.0006,
+ "loss": 4.77295446395874,
+ "step": 920
+ },
+ {
+ "epoch": 12.796506550218341,
+ "grad_norm": 0.015074674971401691,
+ "learning_rate": 0.0006,
+ "loss": 4.830754280090332,
+ "step": 921
+ },
+ {
+ "epoch": 12.810480349344978,
+ "grad_norm": 0.015599515289068222,
+ "learning_rate": 0.0006,
+ "loss": 4.837759017944336,
+ "step": 922
+ },
+ {
+ "epoch": 12.824454148471617,
+ "grad_norm": 0.014595242217183113,
+ "learning_rate": 0.0006,
+ "loss": 4.882965564727783,
+ "step": 923
+ },
+ {
+ "epoch": 12.838427947598253,
+ "grad_norm": 0.014951467514038086,
+ "learning_rate": 0.0006,
+ "loss": 4.709748268127441,
+ "step": 924
+ },
+ {
+ "epoch": 12.85240174672489,
+ "grad_norm": 0.01539295818656683,
+ "learning_rate": 0.0006,
+ "loss": 4.791160583496094,
+ "step": 925
+ },
+ {
+ "epoch": 12.866375545851529,
+ "grad_norm": 0.015987906605005264,
+ "learning_rate": 0.0006,
+ "loss": 4.661876678466797,
+ "step": 926
+ },
+ {
+ "epoch": 12.880349344978166,
+ "grad_norm": 0.016527341678738594,
+ "learning_rate": 0.0006,
+ "loss": 4.721952438354492,
+ "step": 927
+ },
+ {
+ "epoch": 12.894323144104803,
+ "grad_norm": 0.01637161523103714,
+ "learning_rate": 0.0006,
+ "loss": 4.704446792602539,
+ "step": 928
+ },
+ {
+ "epoch": 12.908296943231441,
+ "grad_norm": 0.015221050009131432,
+ "learning_rate": 0.0006,
+ "loss": 4.771907806396484,
+ "step": 929
+ },
+ {
+ "epoch": 12.922270742358078,
+ "grad_norm": 0.016242166981101036,
+ "learning_rate": 0.0006,
+ "loss": 4.806267738342285,
+ "step": 930
+ },
+ {
+ "epoch": 12.936244541484717,
+ "grad_norm": 0.01628357358276844,
+ "learning_rate": 0.0006,
+ "loss": 4.7252326011657715,
+ "step": 931
+ },
+ {
+ "epoch": 12.950218340611354,
+ "grad_norm": 0.017428133636713028,
+ "learning_rate": 0.0006,
+ "loss": 4.765689849853516,
+ "step": 932
+ },
+ {
+ "epoch": 12.96419213973799,
+ "grad_norm": 0.01870061084628105,
+ "learning_rate": 0.0006,
+ "loss": 4.814488887786865,
+ "step": 933
+ },
+ {
+ "epoch": 12.97816593886463,
+ "grad_norm": 0.019796060398221016,
+ "learning_rate": 0.0006,
+ "loss": 4.739185810089111,
+ "step": 934
+ },
+ {
+ "epoch": 12.992139737991266,
+ "grad_norm": 0.01985820196568966,
+ "learning_rate": 0.0006,
+ "loss": 4.810608863830566,
+ "step": 935
+ },
+ {
+ "epoch": 13.0,
+ "grad_norm": 0.020834477618336678,
+ "learning_rate": 0.0006,
+ "loss": 4.727858066558838,
+ "step": 936
+ },
+ {
+ "epoch": 13.0,
+ "eval_loss": 4.979116916656494,
+ "eval_runtime": 57.2877,
+ "eval_samples_per_second": 42.627,
+ "eval_steps_per_second": 1.344,
+ "step": 936
+ },
+ {
+ "epoch": 13.013973799126637,
+ "grad_norm": 0.02025519125163555,
+ "learning_rate": 0.0006,
+ "loss": 4.628875732421875,
+ "step": 937
+ },
+ {
+ "epoch": 13.027947598253276,
+ "grad_norm": 0.019596368074417114,
+ "learning_rate": 0.0006,
+ "loss": 4.694699287414551,
+ "step": 938
+ },
+ {
+ "epoch": 13.041921397379912,
+ "grad_norm": 0.020494934171438217,
+ "learning_rate": 0.0006,
+ "loss": 4.691821098327637,
+ "step": 939
+ },
+ {
+ "epoch": 13.055895196506551,
+ "grad_norm": 0.020121760666370392,
+ "learning_rate": 0.0006,
+ "loss": 4.71478271484375,
+ "step": 940
+ },
+ {
+ "epoch": 13.069868995633188,
+ "grad_norm": 0.021091405302286148,
+ "learning_rate": 0.0006,
+ "loss": 4.695915222167969,
+ "step": 941
+ },
+ {
+ "epoch": 13.083842794759825,
+ "grad_norm": 0.019787423312664032,
+ "learning_rate": 0.0006,
+ "loss": 4.8652448654174805,
+ "step": 942
+ },
+ {
+ "epoch": 13.097816593886463,
+ "grad_norm": 0.0185833927243948,
+ "learning_rate": 0.0006,
+ "loss": 4.7065887451171875,
+ "step": 943
+ },
+ {
+ "epoch": 13.1117903930131,
+ "grad_norm": 0.021220475435256958,
+ "learning_rate": 0.0006,
+ "loss": 4.741232395172119,
+ "step": 944
+ },
+ {
+ "epoch": 13.125764192139737,
+ "grad_norm": 0.02136622555553913,
+ "learning_rate": 0.0006,
+ "loss": 4.705967903137207,
+ "step": 945
+ },
+ {
+ "epoch": 13.139737991266376,
+ "grad_norm": 0.0191587433218956,
+ "learning_rate": 0.0006,
+ "loss": 4.758667945861816,
+ "step": 946
+ },
+ {
+ "epoch": 13.153711790393013,
+ "grad_norm": 0.018290022388100624,
+ "learning_rate": 0.0006,
+ "loss": 4.760666847229004,
+ "step": 947
+ },
+ {
+ "epoch": 13.167685589519651,
+ "grad_norm": 0.021704409271478653,
+ "learning_rate": 0.0006,
+ "loss": 4.750555992126465,
+ "step": 948
+ },
+ {
+ "epoch": 13.181659388646288,
+ "grad_norm": 0.021668918430805206,
+ "learning_rate": 0.0006,
+ "loss": 4.840903282165527,
+ "step": 949
+ },
+ {
+ "epoch": 13.195633187772925,
+ "grad_norm": 0.019497141242027283,
+ "learning_rate": 0.0006,
+ "loss": 4.738969326019287,
+ "step": 950
+ },
+ {
+ "epoch": 13.209606986899564,
+ "grad_norm": 0.018500596284866333,
+ "learning_rate": 0.0006,
+ "loss": 4.75154972076416,
+ "step": 951
+ },
+ {
+ "epoch": 13.2235807860262,
+ "grad_norm": 0.01872239261865616,
+ "learning_rate": 0.0006,
+ "loss": 4.811738014221191,
+ "step": 952
+ },
+ {
+ "epoch": 13.237554585152838,
+ "grad_norm": 0.020321547985076904,
+ "learning_rate": 0.0006,
+ "loss": 4.815384387969971,
+ "step": 953
+ },
+ {
+ "epoch": 13.251528384279476,
+ "grad_norm": 0.02283575013279915,
+ "learning_rate": 0.0006,
+ "loss": 4.74582052230835,
+ "step": 954
+ },
+ {
+ "epoch": 13.265502183406113,
+ "grad_norm": 0.02474398910999298,
+ "learning_rate": 0.0006,
+ "loss": 4.738375186920166,
+ "step": 955
+ },
+ {
+ "epoch": 13.279475982532752,
+ "grad_norm": 0.02065764181315899,
+ "learning_rate": 0.0006,
+ "loss": 4.746875286102295,
+ "step": 956
+ },
+ {
+ "epoch": 13.293449781659389,
+ "grad_norm": 0.020534520968794823,
+ "learning_rate": 0.0006,
+ "loss": 4.656914710998535,
+ "step": 957
+ },
+ {
+ "epoch": 13.307423580786025,
+ "grad_norm": 0.0206378772854805,
+ "learning_rate": 0.0006,
+ "loss": 4.842615604400635,
+ "step": 958
+ },
+ {
+ "epoch": 13.321397379912664,
+ "grad_norm": 0.017935654148459435,
+ "learning_rate": 0.0006,
+ "loss": 4.717827796936035,
+ "step": 959
+ },
+ {
+ "epoch": 13.335371179039301,
+ "grad_norm": 0.016958363354206085,
+ "learning_rate": 0.0006,
+ "loss": 4.681070327758789,
+ "step": 960
+ },
+ {
+ "epoch": 13.34934497816594,
+ "grad_norm": 0.015735412016510963,
+ "learning_rate": 0.0006,
+ "loss": 4.7957048416137695,
+ "step": 961
+ },
+ {
+ "epoch": 13.363318777292577,
+ "grad_norm": 0.01563699170947075,
+ "learning_rate": 0.0006,
+ "loss": 4.701192378997803,
+ "step": 962
+ },
+ {
+ "epoch": 13.377292576419213,
+ "grad_norm": 0.015207760035991669,
+ "learning_rate": 0.0006,
+ "loss": 4.73256254196167,
+ "step": 963
+ },
+ {
+ "epoch": 13.391266375545852,
+ "grad_norm": 0.015396883711218834,
+ "learning_rate": 0.0006,
+ "loss": 4.76412296295166,
+ "step": 964
+ },
+ {
+ "epoch": 13.405240174672489,
+ "grad_norm": 0.015690583735704422,
+ "learning_rate": 0.0006,
+ "loss": 4.791157245635986,
+ "step": 965
+ },
+ {
+ "epoch": 13.419213973799126,
+ "grad_norm": 0.01755724661052227,
+ "learning_rate": 0.0006,
+ "loss": 4.845666885375977,
+ "step": 966
+ },
+ {
+ "epoch": 13.433187772925764,
+ "grad_norm": 0.019443422555923462,
+ "learning_rate": 0.0006,
+ "loss": 4.5902299880981445,
+ "step": 967
+ },
+ {
+ "epoch": 13.447161572052401,
+ "grad_norm": 0.02118338644504547,
+ "learning_rate": 0.0006,
+ "loss": 4.740627288818359,
+ "step": 968
+ },
+ {
+ "epoch": 13.46113537117904,
+ "grad_norm": 0.02174839936196804,
+ "learning_rate": 0.0006,
+ "loss": 4.814560890197754,
+ "step": 969
+ },
+ {
+ "epoch": 13.475109170305677,
+ "grad_norm": 0.022860374301671982,
+ "learning_rate": 0.0006,
+ "loss": 4.66440486907959,
+ "step": 970
+ },
+ {
+ "epoch": 13.489082969432314,
+ "grad_norm": 0.024742677807807922,
+ "learning_rate": 0.0006,
+ "loss": 4.7451653480529785,
+ "step": 971
+ },
+ {
+ "epoch": 13.503056768558952,
+ "grad_norm": 0.027068907395005226,
+ "learning_rate": 0.0006,
+ "loss": 4.802584171295166,
+ "step": 972
+ },
+ {
+ "epoch": 13.51703056768559,
+ "grad_norm": 0.026316625997424126,
+ "learning_rate": 0.0006,
+ "loss": 4.842397689819336,
+ "step": 973
+ },
+ {
+ "epoch": 13.531004366812226,
+ "grad_norm": 0.024610962718725204,
+ "learning_rate": 0.0006,
+ "loss": 4.650721073150635,
+ "step": 974
+ },
+ {
+ "epoch": 13.544978165938865,
+ "grad_norm": 0.024122603237628937,
+ "learning_rate": 0.0006,
+ "loss": 4.6984357833862305,
+ "step": 975
+ },
+ {
+ "epoch": 13.558951965065502,
+ "grad_norm": 0.022699708119034767,
+ "learning_rate": 0.0006,
+ "loss": 4.793007850646973,
+ "step": 976
+ },
+ {
+ "epoch": 13.57292576419214,
+ "grad_norm": 0.021706534549593925,
+ "learning_rate": 0.0006,
+ "loss": 4.691362380981445,
+ "step": 977
+ },
+ {
+ "epoch": 13.586899563318777,
+ "grad_norm": 0.021363291889429092,
+ "learning_rate": 0.0006,
+ "loss": 4.76762580871582,
+ "step": 978
+ },
+ {
+ "epoch": 13.600873362445414,
+ "grad_norm": 0.023476677015423775,
+ "learning_rate": 0.0006,
+ "loss": 4.777202606201172,
+ "step": 979
+ },
+ {
+ "epoch": 13.614847161572053,
+ "grad_norm": 0.02457100711762905,
+ "learning_rate": 0.0006,
+ "loss": 4.590703010559082,
+ "step": 980
+ },
+ {
+ "epoch": 13.62882096069869,
+ "grad_norm": 0.02469099499285221,
+ "learning_rate": 0.0006,
+ "loss": 4.640185356140137,
+ "step": 981
+ },
+ {
+ "epoch": 13.642794759825328,
+ "grad_norm": 0.024914603680372238,
+ "learning_rate": 0.0006,
+ "loss": 4.772201061248779,
+ "step": 982
+ },
+ {
+ "epoch": 13.656768558951965,
+ "grad_norm": 0.025133173912763596,
+ "learning_rate": 0.0006,
+ "loss": 4.743337154388428,
+ "step": 983
+ },
+ {
+ "epoch": 13.670742358078602,
+ "grad_norm": 0.02034137211740017,
+ "learning_rate": 0.0006,
+ "loss": 4.752127647399902,
+ "step": 984
+ },
+ {
+ "epoch": 13.68471615720524,
+ "grad_norm": 0.019645661115646362,
+ "learning_rate": 0.0006,
+ "loss": 4.838453769683838,
+ "step": 985
+ },
+ {
+ "epoch": 13.698689956331878,
+ "grad_norm": 0.01900297962129116,
+ "learning_rate": 0.0006,
+ "loss": 4.804868221282959,
+ "step": 986
+ },
+ {
+ "epoch": 13.712663755458514,
+ "grad_norm": 0.020292341709136963,
+ "learning_rate": 0.0006,
+ "loss": 4.609743118286133,
+ "step": 987
+ },
+ {
+ "epoch": 13.726637554585153,
+ "grad_norm": 0.017014775425195694,
+ "learning_rate": 0.0006,
+ "loss": 4.740612983703613,
+ "step": 988
+ },
+ {
+ "epoch": 13.74061135371179,
+ "grad_norm": 0.015199844725430012,
+ "learning_rate": 0.0006,
+ "loss": 4.719701766967773,
+ "step": 989
+ },
+ {
+ "epoch": 13.754585152838429,
+ "grad_norm": 0.015125414356589317,
+ "learning_rate": 0.0006,
+ "loss": 4.686114311218262,
+ "step": 990
+ },
+ {
+ "epoch": 13.768558951965066,
+ "grad_norm": 0.014779540710151196,
+ "learning_rate": 0.0006,
+ "loss": 4.674347877502441,
+ "step": 991
+ },
+ {
+ "epoch": 13.782532751091702,
+ "grad_norm": 0.015255163423717022,
+ "learning_rate": 0.0006,
+ "loss": 4.6309309005737305,
+ "step": 992
+ },
+ {
+ "epoch": 13.796506550218341,
+ "grad_norm": 0.013692197389900684,
+ "learning_rate": 0.0006,
+ "loss": 4.698497295379639,
+ "step": 993
+ },
+ {
+ "epoch": 13.810480349344978,
+ "grad_norm": 0.014529623091220856,
+ "learning_rate": 0.0006,
+ "loss": 4.628443717956543,
+ "step": 994
+ },
+ {
+ "epoch": 13.824454148471617,
+ "grad_norm": 0.01576659083366394,
+ "learning_rate": 0.0006,
+ "loss": 4.788015842437744,
+ "step": 995
+ },
+ {
+ "epoch": 13.838427947598253,
+ "grad_norm": 0.015834277495741844,
+ "learning_rate": 0.0006,
+ "loss": 4.672703742980957,
+ "step": 996
+ },
+ {
+ "epoch": 13.85240174672489,
+ "grad_norm": 0.015539503656327724,
+ "learning_rate": 0.0006,
+ "loss": 4.731503486633301,
+ "step": 997
+ },
+ {
+ "epoch": 13.866375545851529,
+ "grad_norm": 0.013463746756315231,
+ "learning_rate": 0.0006,
+ "loss": 4.7006330490112305,
+ "step": 998
+ },
+ {
+ "epoch": 13.880349344978166,
+ "grad_norm": 0.016377726569771767,
+ "learning_rate": 0.0006,
+ "loss": 4.636074066162109,
+ "step": 999
+ },
+ {
+ "epoch": 13.894323144104803,
+ "grad_norm": 0.016660034656524658,
+ "learning_rate": 0.0006,
+ "loss": 4.742693901062012,
+ "step": 1000
+ },
+ {
+ "epoch": 13.908296943231441,
+ "grad_norm": 0.014902938157320023,
+ "learning_rate": 0.0006,
+ "loss": 4.769811630249023,
+ "step": 1001
+ },
+ {
+ "epoch": 13.922270742358078,
+ "grad_norm": 0.014883476309478283,
+ "learning_rate": 0.0006,
+ "loss": 4.731159210205078,
+ "step": 1002
+ },
+ {
+ "epoch": 13.936244541484717,
+ "grad_norm": 0.014494217932224274,
+ "learning_rate": 0.0006,
+ "loss": 4.659485816955566,
+ "step": 1003
+ },
+ {
+ "epoch": 13.950218340611354,
+ "grad_norm": 0.015426991507411003,
+ "learning_rate": 0.0006,
+ "loss": 4.755229949951172,
+ "step": 1004
+ },
+ {
+ "epoch": 13.96419213973799,
+ "grad_norm": 0.016648586839437485,
+ "learning_rate": 0.0006,
+ "loss": 4.74111270904541,
+ "step": 1005
+ },
+ {
+ "epoch": 13.97816593886463,
+ "grad_norm": 0.016980541869997978,
+ "learning_rate": 0.0006,
+ "loss": 4.74072265625,
+ "step": 1006
+ },
+ {
+ "epoch": 13.992139737991266,
+ "grad_norm": 0.018070152029395103,
+ "learning_rate": 0.0006,
+ "loss": 4.732975006103516,
+ "step": 1007
+ },
+ {
+ "epoch": 14.0,
+ "grad_norm": 0.019910665228962898,
+ "learning_rate": 0.0006,
+ "loss": 4.634697914123535,
+ "step": 1008
+ },
+ {
+ "epoch": 14.0,
+ "eval_loss": 4.850640296936035,
+ "eval_runtime": 57.3039,
+ "eval_samples_per_second": 42.615,
+ "eval_steps_per_second": 1.344,
+ "step": 1008
+ },
+ {
+ "epoch": 14.013973799126637,
+ "grad_norm": 0.01980053074657917,
+ "learning_rate": 0.0006,
+ "loss": 4.628744602203369,
+ "step": 1009
+ },
+ {
+ "epoch": 14.027947598253276,
+ "grad_norm": 0.02497623674571514,
+ "learning_rate": 0.0006,
+ "loss": 4.75930118560791,
+ "step": 1010
+ },
+ {
+ "epoch": 14.041921397379912,
+ "grad_norm": 0.02794097363948822,
+ "learning_rate": 0.0006,
+ "loss": 4.691340446472168,
+ "step": 1011
+ },
+ {
+ "epoch": 14.055895196506551,
+ "grad_norm": 0.023725535720586777,
+ "learning_rate": 0.0006,
+ "loss": 4.733664512634277,
+ "step": 1012
+ },
+ {
+ "epoch": 14.069868995633188,
+ "grad_norm": 0.022966474294662476,
+ "learning_rate": 0.0006,
+ "loss": 4.664808750152588,
+ "step": 1013
+ },
+ {
+ "epoch": 14.083842794759825,
+ "grad_norm": 0.02530563250184059,
+ "learning_rate": 0.0006,
+ "loss": 4.8047285079956055,
+ "step": 1014
+ },
+ {
+ "epoch": 14.097816593886463,
+ "grad_norm": 0.02491082064807415,
+ "learning_rate": 0.0006,
+ "loss": 4.667956352233887,
+ "step": 1015
+ },
+ {
+ "epoch": 14.1117903930131,
+ "grad_norm": 0.022110294550657272,
+ "learning_rate": 0.0006,
+ "loss": 4.666140556335449,
+ "step": 1016
+ },
+ {
+ "epoch": 14.125764192139737,
+ "grad_norm": 0.02065849117934704,
+ "learning_rate": 0.0006,
+ "loss": 4.581960678100586,
+ "step": 1017
+ },
+ {
+ "epoch": 14.139737991266376,
+ "grad_norm": 0.01941298507153988,
+ "learning_rate": 0.0006,
+ "loss": 4.614226341247559,
+ "step": 1018
+ },
+ {
+ "epoch": 14.153711790393013,
+ "grad_norm": 0.022441856563091278,
+ "learning_rate": 0.0006,
+ "loss": 4.638352394104004,
+ "step": 1019
+ },
+ {
+ "epoch": 14.167685589519651,
+ "grad_norm": 0.021519694477319717,
+ "learning_rate": 0.0006,
+ "loss": 4.700160980224609,
+ "step": 1020
+ },
+ {
+ "epoch": 14.181659388646288,
+ "grad_norm": 0.018956486135721207,
+ "learning_rate": 0.0006,
+ "loss": 4.649593353271484,
+ "step": 1021
+ },
+ {
+ "epoch": 14.195633187772925,
+ "grad_norm": 0.016681140288710594,
+ "learning_rate": 0.0006,
+ "loss": 4.603952407836914,
+ "step": 1022
+ },
+ {
+ "epoch": 14.209606986899564,
+ "grad_norm": 0.01788809522986412,
+ "learning_rate": 0.0006,
+ "loss": 4.669160842895508,
+ "step": 1023
+ },
+ {
+ "epoch": 14.2235807860262,
+ "grad_norm": 0.016047228127717972,
+ "learning_rate": 0.0006,
+ "loss": 4.6367950439453125,
+ "step": 1024
+ },
+ {
+ "epoch": 14.237554585152838,
+ "grad_norm": 0.01614239625632763,
+ "learning_rate": 0.0006,
+ "loss": 4.664529323577881,
+ "step": 1025
+ },
+ {
+ "epoch": 14.251528384279476,
+ "grad_norm": 0.016362549737095833,
+ "learning_rate": 0.0006,
+ "loss": 4.702373504638672,
+ "step": 1026
+ },
+ {
+ "epoch": 14.265502183406113,
+ "grad_norm": 0.015800409018993378,
+ "learning_rate": 0.0006,
+ "loss": 4.7594428062438965,
+ "step": 1027
+ },
+ {
+ "epoch": 14.279475982532752,
+ "grad_norm": 0.016730139032006264,
+ "learning_rate": 0.0006,
+ "loss": 4.677244186401367,
+ "step": 1028
+ },
+ {
+ "epoch": 14.293449781659389,
+ "grad_norm": 0.01661759801208973,
+ "learning_rate": 0.0006,
+ "loss": 4.633679389953613,
+ "step": 1029
+ },
+ {
+ "epoch": 14.307423580786025,
+ "grad_norm": 0.017673347145318985,
+ "learning_rate": 0.0006,
+ "loss": 4.649107456207275,
+ "step": 1030
+ },
+ {
+ "epoch": 14.321397379912664,
+ "grad_norm": 0.019243352115154266,
+ "learning_rate": 0.0006,
+ "loss": 4.611257553100586,
+ "step": 1031
+ },
+ {
+ "epoch": 14.335371179039301,
+ "grad_norm": 0.019817164167761803,
+ "learning_rate": 0.0006,
+ "loss": 4.588086128234863,
+ "step": 1032
+ },
+ {
+ "epoch": 14.34934497816594,
+ "grad_norm": 0.022292211651802063,
+ "learning_rate": 0.0006,
+ "loss": 4.783066749572754,
+ "step": 1033
+ },
+ {
+ "epoch": 14.363318777292577,
+ "grad_norm": 0.02140096202492714,
+ "learning_rate": 0.0006,
+ "loss": 4.604718208312988,
+ "step": 1034
+ },
+ {
+ "epoch": 14.377292576419213,
+ "grad_norm": 0.0187817495316267,
+ "learning_rate": 0.0006,
+ "loss": 4.721773147583008,
+ "step": 1035
+ },
+ {
+ "epoch": 14.391266375545852,
+ "grad_norm": 0.019879210740327835,
+ "learning_rate": 0.0006,
+ "loss": 4.789286136627197,
+ "step": 1036
+ },
+ {
+ "epoch": 14.405240174672489,
+ "grad_norm": 0.01885785162448883,
+ "learning_rate": 0.0006,
+ "loss": 4.669743537902832,
+ "step": 1037
+ },
+ {
+ "epoch": 14.419213973799126,
+ "grad_norm": 0.018324270844459534,
+ "learning_rate": 0.0006,
+ "loss": 4.65901517868042,
+ "step": 1038
+ },
+ {
+ "epoch": 14.433187772925764,
+ "grad_norm": 0.017664821818470955,
+ "learning_rate": 0.0006,
+ "loss": 4.667839527130127,
+ "step": 1039
+ },
+ {
+ "epoch": 14.447161572052401,
+ "grad_norm": 0.015702905133366585,
+ "learning_rate": 0.0006,
+ "loss": 4.64700984954834,
+ "step": 1040
+ },
+ {
+ "epoch": 14.46113537117904,
+ "grad_norm": 0.016674194484949112,
+ "learning_rate": 0.0006,
+ "loss": 4.694093227386475,
+ "step": 1041
+ },
+ {
+ "epoch": 14.475109170305677,
+ "grad_norm": 0.016869477927684784,
+ "learning_rate": 0.0006,
+ "loss": 4.619544982910156,
+ "step": 1042
+ },
+ {
+ "epoch": 14.489082969432314,
+ "grad_norm": 0.016557540744543076,
+ "learning_rate": 0.0006,
+ "loss": 4.568667411804199,
+ "step": 1043
+ },
+ {
+ "epoch": 14.503056768558952,
+ "grad_norm": 0.014500945806503296,
+ "learning_rate": 0.0006,
+ "loss": 4.561741828918457,
+ "step": 1044
+ },
+ {
+ "epoch": 14.51703056768559,
+ "grad_norm": 0.014228739775717258,
+ "learning_rate": 0.0006,
+ "loss": 4.654277324676514,
+ "step": 1045
+ },
+ {
+ "epoch": 14.531004366812226,
+ "grad_norm": 0.013894238509237766,
+ "learning_rate": 0.0006,
+ "loss": 4.5867719650268555,
+ "step": 1046
+ },
+ {
+ "epoch": 14.544978165938865,
+ "grad_norm": 0.01550068985670805,
+ "learning_rate": 0.0006,
+ "loss": 4.61168098449707,
+ "step": 1047
+ },
+ {
+ "epoch": 14.558951965065502,
+ "grad_norm": 0.015940772369503975,
+ "learning_rate": 0.0006,
+ "loss": 4.691247940063477,
+ "step": 1048
+ },
+ {
+ "epoch": 14.57292576419214,
+ "grad_norm": 0.01714535616338253,
+ "learning_rate": 0.0006,
+ "loss": 4.61183500289917,
+ "step": 1049
+ },
+ {
+ "epoch": 14.586899563318777,
+ "grad_norm": 0.017925038933753967,
+ "learning_rate": 0.0006,
+ "loss": 4.719022750854492,
+ "step": 1050
+ },
+ {
+ "epoch": 14.600873362445414,
+ "grad_norm": 0.018141552805900574,
+ "learning_rate": 0.0006,
+ "loss": 4.760281562805176,
+ "step": 1051
+ },
+ {
+ "epoch": 14.614847161572053,
+ "grad_norm": 0.017131328582763672,
+ "learning_rate": 0.0006,
+ "loss": 4.709680080413818,
+ "step": 1052
+ },
+ {
+ "epoch": 14.62882096069869,
+ "grad_norm": 0.018730048090219498,
+ "learning_rate": 0.0006,
+ "loss": 4.5994720458984375,
+ "step": 1053
+ },
+ {
+ "epoch": 14.642794759825328,
+ "grad_norm": 0.018894298002123833,
+ "learning_rate": 0.0006,
+ "loss": 4.669070720672607,
+ "step": 1054
+ },
+ {
+ "epoch": 14.656768558951965,
+ "grad_norm": 0.017643198370933533,
+ "learning_rate": 0.0006,
+ "loss": 4.641790390014648,
+ "step": 1055
+ },
+ {
+ "epoch": 14.670742358078602,
+ "grad_norm": 0.017647244036197662,
+ "learning_rate": 0.0006,
+ "loss": 4.653899192810059,
+ "step": 1056
+ },
+ {
+ "epoch": 14.68471615720524,
+ "grad_norm": 0.017304198816418648,
+ "learning_rate": 0.0006,
+ "loss": 4.7487473487854,
+ "step": 1057
+ },
+ {
+ "epoch": 14.698689956331878,
+ "grad_norm": 0.01638447679579258,
+ "learning_rate": 0.0006,
+ "loss": 4.71945858001709,
+ "step": 1058
+ },
+ {
+ "epoch": 14.712663755458514,
+ "grad_norm": 0.015892071649432182,
+ "learning_rate": 0.0006,
+ "loss": 4.744903564453125,
+ "step": 1059
+ },
+ {
+ "epoch": 14.726637554585153,
+ "grad_norm": 0.017420614138245583,
+ "learning_rate": 0.0006,
+ "loss": 4.781563758850098,
+ "step": 1060
+ },
+ {
+ "epoch": 14.74061135371179,
+ "grad_norm": 0.016498520970344543,
+ "learning_rate": 0.0006,
+ "loss": 4.622779846191406,
+ "step": 1061
+ },
+ {
+ "epoch": 14.754585152838429,
+ "grad_norm": 0.017482127994298935,
+ "learning_rate": 0.0006,
+ "loss": 4.62825870513916,
+ "step": 1062
+ },
+ {
+ "epoch": 14.768558951965066,
+ "grad_norm": 0.019018379971385002,
+ "learning_rate": 0.0006,
+ "loss": 4.693194389343262,
+ "step": 1063
+ },
+ {
+ "epoch": 14.782532751091702,
+ "grad_norm": 0.01993914321064949,
+ "learning_rate": 0.0006,
+ "loss": 4.742903709411621,
+ "step": 1064
+ },
+ {
+ "epoch": 14.796506550218341,
+ "grad_norm": 0.02301335707306862,
+ "learning_rate": 0.0006,
+ "loss": 4.613251686096191,
+ "step": 1065
+ },
+ {
+ "epoch": 14.810480349344978,
+ "grad_norm": 0.024088222533464432,
+ "learning_rate": 0.0006,
+ "loss": 4.594305992126465,
+ "step": 1066
+ },
+ {
+ "epoch": 14.824454148471617,
+ "grad_norm": 0.02493320405483246,
+ "learning_rate": 0.0006,
+ "loss": 4.700588226318359,
+ "step": 1067
+ },
+ {
+ "epoch": 14.838427947598253,
+ "grad_norm": 0.02501937560737133,
+ "learning_rate": 0.0006,
+ "loss": 4.601742744445801,
+ "step": 1068
+ },
+ {
+ "epoch": 14.85240174672489,
+ "grad_norm": 0.023490751162171364,
+ "learning_rate": 0.0006,
+ "loss": 4.589211463928223,
+ "step": 1069
+ },
+ {
+ "epoch": 14.866375545851529,
+ "grad_norm": 0.019695665687322617,
+ "learning_rate": 0.0006,
+ "loss": 4.635081768035889,
+ "step": 1070
+ },
+ {
+ "epoch": 14.880349344978166,
+ "grad_norm": 0.019276980310678482,
+ "learning_rate": 0.0006,
+ "loss": 4.737002372741699,
+ "step": 1071
+ },
+ {
+ "epoch": 14.894323144104803,
+ "grad_norm": 0.022430511191487312,
+ "learning_rate": 0.0006,
+ "loss": 4.808295726776123,
+ "step": 1072
+ },
+ {
+ "epoch": 14.908296943231441,
+ "grad_norm": 0.021543916314840317,
+ "learning_rate": 0.0006,
+ "loss": 4.559475421905518,
+ "step": 1073
+ },
+ {
+ "epoch": 14.922270742358078,
+ "grad_norm": 0.019885241985321045,
+ "learning_rate": 0.0006,
+ "loss": 4.647948741912842,
+ "step": 1074
+ },
+ {
+ "epoch": 14.936244541484717,
+ "grad_norm": 0.02115175500512123,
+ "learning_rate": 0.0006,
+ "loss": 4.680505752563477,
+ "step": 1075
+ },
+ {
+ "epoch": 14.950218340611354,
+ "grad_norm": 0.02156151458621025,
+ "learning_rate": 0.0006,
+ "loss": 4.673477649688721,
+ "step": 1076
+ },
+ {
+ "epoch": 14.96419213973799,
+ "grad_norm": 0.02184601128101349,
+ "learning_rate": 0.0006,
+ "loss": 4.644402027130127,
+ "step": 1077
+ },
+ {
+ "epoch": 14.97816593886463,
+ "grad_norm": 0.0211983360350132,
+ "learning_rate": 0.0006,
+ "loss": 4.821558475494385,
+ "step": 1078
+ },
+ {
+ "epoch": 14.992139737991266,
+ "grad_norm": 0.019636567682027817,
+ "learning_rate": 0.0006,
+ "loss": 4.578868865966797,
+ "step": 1079
+ },
+ {
+ "epoch": 15.0,
+ "grad_norm": 0.019997350871562958,
+ "learning_rate": 0.0006,
+ "loss": 4.691307544708252,
+ "step": 1080
+ },
+ {
+ "epoch": 15.0,
+ "eval_loss": 4.872478008270264,
+ "eval_runtime": 56.8513,
+ "eval_samples_per_second": 42.954,
+ "eval_steps_per_second": 1.354,
+ "step": 1080
+ },
+ {
+ "epoch": 15.013973799126637,
+ "grad_norm": 0.018830662593245506,
+ "learning_rate": 0.0006,
+ "loss": 4.761779308319092,
+ "step": 1081
+ },
+ {
+ "epoch": 15.027947598253276,
+ "grad_norm": 0.016876481473445892,
+ "learning_rate": 0.0006,
+ "loss": 4.61148738861084,
+ "step": 1082
+ },
+ {
+ "epoch": 15.041921397379912,
+ "grad_norm": 0.015550863929092884,
+ "learning_rate": 0.0006,
+ "loss": 4.676124095916748,
+ "step": 1083
+ },
+ {
+ "epoch": 15.055895196506551,
+ "grad_norm": 0.017178896814584732,
+ "learning_rate": 0.0006,
+ "loss": 4.588935852050781,
+ "step": 1084
+ },
+ {
+ "epoch": 15.069868995633188,
+ "grad_norm": 0.015465447679162025,
+ "learning_rate": 0.0006,
+ "loss": 4.605499267578125,
+ "step": 1085
+ },
+ {
+ "epoch": 15.083842794759825,
+ "grad_norm": 0.01606195978820324,
+ "learning_rate": 0.0006,
+ "loss": 4.723373889923096,
+ "step": 1086
+ },
+ {
+ "epoch": 15.097816593886463,
+ "grad_norm": 0.016521936282515526,
+ "learning_rate": 0.0006,
+ "loss": 4.646608352661133,
+ "step": 1087
+ },
+ {
+ "epoch": 15.1117903930131,
+ "grad_norm": 0.01589205674827099,
+ "learning_rate": 0.0006,
+ "loss": 4.498138427734375,
+ "step": 1088
+ },
+ {
+ "epoch": 15.125764192139737,
+ "grad_norm": 0.01524856686592102,
+ "learning_rate": 0.0006,
+ "loss": 4.594562530517578,
+ "step": 1089
+ },
+ {
+ "epoch": 15.139737991266376,
+ "grad_norm": 0.014714641496539116,
+ "learning_rate": 0.0006,
+ "loss": 4.630744934082031,
+ "step": 1090
+ },
+ {
+ "epoch": 15.153711790393013,
+ "grad_norm": 0.015333331190049648,
+ "learning_rate": 0.0006,
+ "loss": 4.682388782501221,
+ "step": 1091
+ },
+ {
+ "epoch": 15.167685589519651,
+ "grad_norm": 0.01550073828548193,
+ "learning_rate": 0.0006,
+ "loss": 4.692652702331543,
+ "step": 1092
+ },
+ {
+ "epoch": 15.181659388646288,
+ "grad_norm": 0.014783318154513836,
+ "learning_rate": 0.0006,
+ "loss": 4.6223835945129395,
+ "step": 1093
+ },
+ {
+ "epoch": 15.195633187772925,
+ "grad_norm": 0.014372751116752625,
+ "learning_rate": 0.0006,
+ "loss": 4.52687406539917,
+ "step": 1094
+ },
+ {
+ "epoch": 15.209606986899564,
+ "grad_norm": 0.014779130928218365,
+ "learning_rate": 0.0006,
+ "loss": 4.637009620666504,
+ "step": 1095
+ },
+ {
+ "epoch": 15.2235807860262,
+ "grad_norm": 0.015697073191404343,
+ "learning_rate": 0.0006,
+ "loss": 4.682981014251709,
+ "step": 1096
+ },
+ {
+ "epoch": 15.237554585152838,
+ "grad_norm": 0.016523664817214012,
+ "learning_rate": 0.0006,
+ "loss": 4.705722808837891,
+ "step": 1097
+ },
+ {
+ "epoch": 15.251528384279476,
+ "grad_norm": 0.017633996903896332,
+ "learning_rate": 0.0006,
+ "loss": 4.574686050415039,
+ "step": 1098
+ },
+ {
+ "epoch": 15.265502183406113,
+ "grad_norm": 0.019955076277256012,
+ "learning_rate": 0.0006,
+ "loss": 4.634450912475586,
+ "step": 1099
+ },
+ {
+ "epoch": 15.279475982532752,
+ "grad_norm": 0.02132618986070156,
+ "learning_rate": 0.0006,
+ "loss": 4.56110143661499,
+ "step": 1100
+ },
+ {
+ "epoch": 15.293449781659389,
+ "grad_norm": 0.0207593385130167,
+ "learning_rate": 0.0006,
+ "loss": 4.629396438598633,
+ "step": 1101
+ },
+ {
+ "epoch": 15.307423580786025,
+ "grad_norm": 0.017882540822029114,
+ "learning_rate": 0.0006,
+ "loss": 4.730963230133057,
+ "step": 1102
+ },
+ {
+ "epoch": 15.321397379912664,
+ "grad_norm": 0.01917876861989498,
+ "learning_rate": 0.0006,
+ "loss": 4.64585542678833,
+ "step": 1103
+ },
+ {
+ "epoch": 15.335371179039301,
+ "grad_norm": 0.01943974196910858,
+ "learning_rate": 0.0006,
+ "loss": 4.656181335449219,
+ "step": 1104
+ },
+ {
+ "epoch": 15.34934497816594,
+ "grad_norm": 0.019904715940356255,
+ "learning_rate": 0.0006,
+ "loss": 4.619329452514648,
+ "step": 1105
+ },
+ {
+ "epoch": 15.363318777292577,
+ "grad_norm": 0.019174346700310707,
+ "learning_rate": 0.0006,
+ "loss": 4.647892475128174,
+ "step": 1106
+ },
+ {
+ "epoch": 15.377292576419213,
+ "grad_norm": 0.018244758248329163,
+ "learning_rate": 0.0006,
+ "loss": 4.56166934967041,
+ "step": 1107
+ },
+ {
+ "epoch": 15.391266375545852,
+ "grad_norm": 0.018914205953478813,
+ "learning_rate": 0.0006,
+ "loss": 4.648063659667969,
+ "step": 1108
+ },
+ {
+ "epoch": 15.405240174672489,
+ "grad_norm": 0.01961071416735649,
+ "learning_rate": 0.0006,
+ "loss": 4.665525436401367,
+ "step": 1109
+ },
+ {
+ "epoch": 15.419213973799126,
+ "grad_norm": 0.01841311901807785,
+ "learning_rate": 0.0006,
+ "loss": 4.611342430114746,
+ "step": 1110
+ },
+ {
+ "epoch": 15.433187772925764,
+ "grad_norm": 0.01863427273929119,
+ "learning_rate": 0.0006,
+ "loss": 4.664256572723389,
+ "step": 1111
+ },
+ {
+ "epoch": 15.447161572052401,
+ "grad_norm": 0.020506877452135086,
+ "learning_rate": 0.0006,
+ "loss": 4.631237506866455,
+ "step": 1112
+ },
+ {
+ "epoch": 15.46113537117904,
+ "grad_norm": 0.02223833277821541,
+ "learning_rate": 0.0006,
+ "loss": 4.668654441833496,
+ "step": 1113
+ },
+ {
+ "epoch": 15.475109170305677,
+ "grad_norm": 0.023336702957749367,
+ "learning_rate": 0.0006,
+ "loss": 4.562833309173584,
+ "step": 1114
+ },
+ {
+ "epoch": 15.489082969432314,
+ "grad_norm": 0.023052413016557693,
+ "learning_rate": 0.0006,
+ "loss": 4.621532917022705,
+ "step": 1115
+ },
+ {
+ "epoch": 15.503056768558952,
+ "grad_norm": 0.023420924320816994,
+ "learning_rate": 0.0006,
+ "loss": 4.665465354919434,
+ "step": 1116
+ },
+ {
+ "epoch": 15.51703056768559,
+ "grad_norm": 0.024071281775832176,
+ "learning_rate": 0.0006,
+ "loss": 4.595980644226074,
+ "step": 1117
+ },
+ {
+ "epoch": 15.531004366812226,
+ "grad_norm": 0.023406682536005974,
+ "learning_rate": 0.0006,
+ "loss": 4.642635345458984,
+ "step": 1118
+ },
+ {
+ "epoch": 15.544978165938865,
+ "grad_norm": 0.02138647809624672,
+ "learning_rate": 0.0006,
+ "loss": 4.704513072967529,
+ "step": 1119
+ },
+ {
+ "epoch": 15.558951965065502,
+ "grad_norm": 0.0221384447067976,
+ "learning_rate": 0.0006,
+ "loss": 4.567503452301025,
+ "step": 1120
+ },
+ {
+ "epoch": 15.57292576419214,
+ "grad_norm": 0.020879851654171944,
+ "learning_rate": 0.0006,
+ "loss": 4.664057731628418,
+ "step": 1121
+ },
+ {
+ "epoch": 15.586899563318777,
+ "grad_norm": 0.021478967741131783,
+ "learning_rate": 0.0006,
+ "loss": 4.539627552032471,
+ "step": 1122
+ },
+ {
+ "epoch": 15.600873362445414,
+ "grad_norm": 0.022748656570911407,
+ "learning_rate": 0.0006,
+ "loss": 4.5654497146606445,
+ "step": 1123
+ },
+ {
+ "epoch": 15.614847161572053,
+ "grad_norm": 0.021870248019695282,
+ "learning_rate": 0.0006,
+ "loss": 4.56899356842041,
+ "step": 1124
+ },
+ {
+ "epoch": 15.62882096069869,
+ "grad_norm": 0.021711355075240135,
+ "learning_rate": 0.0006,
+ "loss": 4.657808303833008,
+ "step": 1125
+ },
+ {
+ "epoch": 15.642794759825328,
+ "grad_norm": 0.022810909897089005,
+ "learning_rate": 0.0006,
+ "loss": 4.6526689529418945,
+ "step": 1126
+ },
+ {
+ "epoch": 15.656768558951965,
+ "grad_norm": 0.022561343386769295,
+ "learning_rate": 0.0006,
+ "loss": 4.534887790679932,
+ "step": 1127
+ },
+ {
+ "epoch": 15.670742358078602,
+ "grad_norm": 0.019909802824258804,
+ "learning_rate": 0.0006,
+ "loss": 4.600790023803711,
+ "step": 1128
+ },
+ {
+ "epoch": 15.68471615720524,
+ "grad_norm": 0.020092325285077095,
+ "learning_rate": 0.0006,
+ "loss": 4.773830413818359,
+ "step": 1129
+ },
+ {
+ "epoch": 15.698689956331878,
+ "grad_norm": 0.018834445625543594,
+ "learning_rate": 0.0006,
+ "loss": 4.628868103027344,
+ "step": 1130
+ },
+ {
+ "epoch": 15.712663755458514,
+ "grad_norm": 0.017552798613905907,
+ "learning_rate": 0.0006,
+ "loss": 4.647500038146973,
+ "step": 1131
+ },
+ {
+ "epoch": 15.726637554585153,
+ "grad_norm": 0.01751251146197319,
+ "learning_rate": 0.0006,
+ "loss": 4.590545654296875,
+ "step": 1132
+ },
+ {
+ "epoch": 15.74061135371179,
+ "grad_norm": 0.01699099875986576,
+ "learning_rate": 0.0006,
+ "loss": 4.509699821472168,
+ "step": 1133
+ },
+ {
+ "epoch": 15.754585152838429,
+ "grad_norm": 0.015808649361133575,
+ "learning_rate": 0.0006,
+ "loss": 4.606992244720459,
+ "step": 1134
+ },
+ {
+ "epoch": 15.768558951965066,
+ "grad_norm": 0.01655443385243416,
+ "learning_rate": 0.0006,
+ "loss": 4.585219860076904,
+ "step": 1135
+ },
+ {
+ "epoch": 15.782532751091702,
+ "grad_norm": 0.01708163693547249,
+ "learning_rate": 0.0006,
+ "loss": 4.668147087097168,
+ "step": 1136
+ },
+ {
+ "epoch": 15.796506550218341,
+ "grad_norm": 0.017161913216114044,
+ "learning_rate": 0.0006,
+ "loss": 4.7170634269714355,
+ "step": 1137
+ },
+ {
+ "epoch": 15.810480349344978,
+ "grad_norm": 0.017262674868106842,
+ "learning_rate": 0.0006,
+ "loss": 4.572132110595703,
+ "step": 1138
+ },
+ {
+ "epoch": 15.824454148471617,
+ "grad_norm": 0.017713576555252075,
+ "learning_rate": 0.0006,
+ "loss": 4.702538967132568,
+ "step": 1139
+ },
+ {
+ "epoch": 15.838427947598253,
+ "grad_norm": 0.018730733543634415,
+ "learning_rate": 0.0006,
+ "loss": 4.629721641540527,
+ "step": 1140
+ },
+ {
+ "epoch": 15.85240174672489,
+ "grad_norm": 0.017960187047719955,
+ "learning_rate": 0.0006,
+ "loss": 4.661039352416992,
+ "step": 1141
+ },
+ {
+ "epoch": 15.866375545851529,
+ "grad_norm": 0.01750079356133938,
+ "learning_rate": 0.0006,
+ "loss": 4.503710746765137,
+ "step": 1142
+ },
+ {
+ "epoch": 15.880349344978166,
+ "grad_norm": 0.016630422323942184,
+ "learning_rate": 0.0006,
+ "loss": 4.6582136154174805,
+ "step": 1143
+ },
+ {
+ "epoch": 15.894323144104803,
+ "grad_norm": 0.015896832570433617,
+ "learning_rate": 0.0006,
+ "loss": 4.596693992614746,
+ "step": 1144
+ },
+ {
+ "epoch": 15.908296943231441,
+ "grad_norm": 0.01566407084465027,
+ "learning_rate": 0.0006,
+ "loss": 4.6337199211120605,
+ "step": 1145
+ },
+ {
+ "epoch": 15.922270742358078,
+ "grad_norm": 0.016823330894112587,
+ "learning_rate": 0.0006,
+ "loss": 4.686994552612305,
+ "step": 1146
+ },
+ {
+ "epoch": 15.936244541484717,
+ "grad_norm": 0.016208255663514137,
+ "learning_rate": 0.0006,
+ "loss": 4.6205549240112305,
+ "step": 1147
+ },
+ {
+ "epoch": 15.950218340611354,
+ "grad_norm": 0.0171707421541214,
+ "learning_rate": 0.0006,
+ "loss": 4.582968711853027,
+ "step": 1148
+ },
+ {
+ "epoch": 15.96419213973799,
+ "grad_norm": 0.017009446397423744,
+ "learning_rate": 0.0006,
+ "loss": 4.510274887084961,
+ "step": 1149
+ },
+ {
+ "epoch": 15.97816593886463,
+ "grad_norm": 0.01565811224281788,
+ "learning_rate": 0.0006,
+ "loss": 4.55834436416626,
+ "step": 1150
+ },
+ {
+ "epoch": 15.992139737991266,
+ "grad_norm": 0.015276037156581879,
+ "learning_rate": 0.0006,
+ "loss": 4.537870407104492,
+ "step": 1151
+ },
+ {
+ "epoch": 16.0,
+ "grad_norm": 0.017305882647633553,
+ "learning_rate": 0.0006,
+ "loss": 4.611305236816406,
+ "step": 1152
+ },
+ {
+ "epoch": 16.0,
+ "eval_loss": 4.775513172149658,
+ "eval_runtime": 56.7341,
+ "eval_samples_per_second": 43.043,
+ "eval_steps_per_second": 1.357,
+ "step": 1152
+ },
+ {
+ "epoch": 16.01397379912664,
+ "grad_norm": 0.016752436757087708,
+ "learning_rate": 0.0006,
+ "loss": 4.541312217712402,
+ "step": 1153
+ },
+ {
+ "epoch": 16.027947598253274,
+ "grad_norm": 0.01643134094774723,
+ "learning_rate": 0.0006,
+ "loss": 4.585402965545654,
+ "step": 1154
+ },
+ {
+ "epoch": 16.041921397379912,
+ "grad_norm": 0.01888720504939556,
+ "learning_rate": 0.0006,
+ "loss": 4.597825050354004,
+ "step": 1155
+ },
+ {
+ "epoch": 16.05589519650655,
+ "grad_norm": 0.022044632583856583,
+ "learning_rate": 0.0006,
+ "loss": 4.532772064208984,
+ "step": 1156
+ },
+ {
+ "epoch": 16.069868995633186,
+ "grad_norm": 0.022986987605690956,
+ "learning_rate": 0.0006,
+ "loss": 4.551351547241211,
+ "step": 1157
+ },
+ {
+ "epoch": 16.083842794759825,
+ "grad_norm": 0.020116791129112244,
+ "learning_rate": 0.0006,
+ "loss": 4.484149932861328,
+ "step": 1158
+ },
+ {
+ "epoch": 16.097816593886463,
+ "grad_norm": 0.021442247554659843,
+ "learning_rate": 0.0006,
+ "loss": 4.515316486358643,
+ "step": 1159
+ },
+ {
+ "epoch": 16.111790393013102,
+ "grad_norm": 0.020974772050976753,
+ "learning_rate": 0.0006,
+ "loss": 4.51606559753418,
+ "step": 1160
+ },
+ {
+ "epoch": 16.125764192139737,
+ "grad_norm": 0.021274864673614502,
+ "learning_rate": 0.0006,
+ "loss": 4.587329864501953,
+ "step": 1161
+ },
+ {
+ "epoch": 16.139737991266376,
+ "grad_norm": 0.022799784317612648,
+ "learning_rate": 0.0006,
+ "loss": 4.568149089813232,
+ "step": 1162
+ },
+ {
+ "epoch": 16.153711790393015,
+ "grad_norm": 0.021784933283925056,
+ "learning_rate": 0.0006,
+ "loss": 4.515336990356445,
+ "step": 1163
+ },
+ {
+ "epoch": 16.16768558951965,
+ "grad_norm": 0.024261271581053734,
+ "learning_rate": 0.0006,
+ "loss": 4.5127081871032715,
+ "step": 1164
+ },
+ {
+ "epoch": 16.18165938864629,
+ "grad_norm": 0.02248375490307808,
+ "learning_rate": 0.0006,
+ "loss": 4.754251003265381,
+ "step": 1165
+ },
+ {
+ "epoch": 16.195633187772927,
+ "grad_norm": 0.02136492356657982,
+ "learning_rate": 0.0006,
+ "loss": 4.58193826675415,
+ "step": 1166
+ },
+ {
+ "epoch": 16.209606986899562,
+ "grad_norm": 0.020727066323161125,
+ "learning_rate": 0.0006,
+ "loss": 4.564209938049316,
+ "step": 1167
+ },
+ {
+ "epoch": 16.2235807860262,
+ "grad_norm": 0.017789119854569435,
+ "learning_rate": 0.0006,
+ "loss": 4.698680877685547,
+ "step": 1168
+ },
+ {
+ "epoch": 16.23755458515284,
+ "grad_norm": 0.017603034153580666,
+ "learning_rate": 0.0006,
+ "loss": 4.501171588897705,
+ "step": 1169
+ },
+ {
+ "epoch": 16.251528384279474,
+ "grad_norm": 0.017345862463116646,
+ "learning_rate": 0.0006,
+ "loss": 4.593590259552002,
+ "step": 1170
+ },
+ {
+ "epoch": 16.265502183406113,
+ "grad_norm": 0.018047425895929337,
+ "learning_rate": 0.0006,
+ "loss": 4.597467422485352,
+ "step": 1171
+ },
+ {
+ "epoch": 16.27947598253275,
+ "grad_norm": 0.01947222836315632,
+ "learning_rate": 0.0006,
+ "loss": 4.5072221755981445,
+ "step": 1172
+ },
+ {
+ "epoch": 16.29344978165939,
+ "grad_norm": 0.019598500803112984,
+ "learning_rate": 0.0006,
+ "loss": 4.601351737976074,
+ "step": 1173
+ },
+ {
+ "epoch": 16.307423580786025,
+ "grad_norm": 0.020002514123916626,
+ "learning_rate": 0.0006,
+ "loss": 4.580431938171387,
+ "step": 1174
+ },
+ {
+ "epoch": 16.321397379912664,
+ "grad_norm": 0.0216530729085207,
+ "learning_rate": 0.0006,
+ "loss": 4.424681663513184,
+ "step": 1175
+ },
+ {
+ "epoch": 16.335371179039303,
+ "grad_norm": 0.01949235238134861,
+ "learning_rate": 0.0006,
+ "loss": 4.751883029937744,
+ "step": 1176
+ },
+ {
+ "epoch": 16.349344978165938,
+ "grad_norm": 0.017779016867280006,
+ "learning_rate": 0.0006,
+ "loss": 4.541490077972412,
+ "step": 1177
+ },
+ {
+ "epoch": 16.363318777292577,
+ "grad_norm": 0.01909121684730053,
+ "learning_rate": 0.0006,
+ "loss": 4.594141006469727,
+ "step": 1178
+ },
+ {
+ "epoch": 16.377292576419215,
+ "grad_norm": 0.019894849509000778,
+ "learning_rate": 0.0006,
+ "loss": 4.555070400238037,
+ "step": 1179
+ },
+ {
+ "epoch": 16.39126637554585,
+ "grad_norm": 0.01993032731115818,
+ "learning_rate": 0.0006,
+ "loss": 4.521401405334473,
+ "step": 1180
+ },
+ {
+ "epoch": 16.40524017467249,
+ "grad_norm": 0.0210788045078516,
+ "learning_rate": 0.0006,
+ "loss": 4.672842979431152,
+ "step": 1181
+ },
+ {
+ "epoch": 16.419213973799128,
+ "grad_norm": 0.02274598926305771,
+ "learning_rate": 0.0006,
+ "loss": 4.629504203796387,
+ "step": 1182
+ },
+ {
+ "epoch": 16.433187772925763,
+ "grad_norm": 0.020448120310902596,
+ "learning_rate": 0.0006,
+ "loss": 4.548209190368652,
+ "step": 1183
+ },
+ {
+ "epoch": 16.4471615720524,
+ "grad_norm": 0.017565643414855003,
+ "learning_rate": 0.0006,
+ "loss": 4.529054641723633,
+ "step": 1184
+ },
+ {
+ "epoch": 16.46113537117904,
+ "grad_norm": 0.019504237920045853,
+ "learning_rate": 0.0006,
+ "loss": 4.48182487487793,
+ "step": 1185
+ },
+ {
+ "epoch": 16.475109170305675,
+ "grad_norm": 0.016823742538690567,
+ "learning_rate": 0.0006,
+ "loss": 4.608696937561035,
+ "step": 1186
+ },
+ {
+ "epoch": 16.489082969432314,
+ "grad_norm": 0.01602894812822342,
+ "learning_rate": 0.0006,
+ "loss": 4.620062351226807,
+ "step": 1187
+ },
+ {
+ "epoch": 16.503056768558952,
+ "grad_norm": 0.016615642234683037,
+ "learning_rate": 0.0006,
+ "loss": 4.69378662109375,
+ "step": 1188
+ },
+ {
+ "epoch": 16.51703056768559,
+ "grad_norm": 0.018576730042696,
+ "learning_rate": 0.0006,
+ "loss": 4.537041664123535,
+ "step": 1189
+ },
+ {
+ "epoch": 16.531004366812226,
+ "grad_norm": 0.018611151725053787,
+ "learning_rate": 0.0006,
+ "loss": 4.496169090270996,
+ "step": 1190
+ },
+ {
+ "epoch": 16.544978165938865,
+ "grad_norm": 0.018622413277626038,
+ "learning_rate": 0.0006,
+ "loss": 4.557358264923096,
+ "step": 1191
+ },
+ {
+ "epoch": 16.558951965065503,
+ "grad_norm": 0.01823544315993786,
+ "learning_rate": 0.0006,
+ "loss": 4.730792999267578,
+ "step": 1192
+ },
+ {
+ "epoch": 16.57292576419214,
+ "grad_norm": 0.018796846270561218,
+ "learning_rate": 0.0006,
+ "loss": 4.781791687011719,
+ "step": 1193
+ },
+ {
+ "epoch": 16.586899563318777,
+ "grad_norm": 0.01728292554616928,
+ "learning_rate": 0.0006,
+ "loss": 4.515257835388184,
+ "step": 1194
+ },
+ {
+ "epoch": 16.600873362445416,
+ "grad_norm": 0.01699589006602764,
+ "learning_rate": 0.0006,
+ "loss": 4.5996904373168945,
+ "step": 1195
+ },
+ {
+ "epoch": 16.61484716157205,
+ "grad_norm": 0.015905970707535744,
+ "learning_rate": 0.0006,
+ "loss": 4.601301193237305,
+ "step": 1196
+ },
+ {
+ "epoch": 16.62882096069869,
+ "grad_norm": 0.01338571310043335,
+ "learning_rate": 0.0006,
+ "loss": 4.557708740234375,
+ "step": 1197
+ },
+ {
+ "epoch": 16.64279475982533,
+ "grad_norm": 0.015298855490982533,
+ "learning_rate": 0.0006,
+ "loss": 4.542287349700928,
+ "step": 1198
+ },
+ {
+ "epoch": 16.656768558951963,
+ "grad_norm": 0.014753367751836777,
+ "learning_rate": 0.0006,
+ "loss": 4.595787048339844,
+ "step": 1199
+ },
+ {
+ "epoch": 16.670742358078602,
+ "grad_norm": 0.015527483075857162,
+ "learning_rate": 0.0006,
+ "loss": 4.607529163360596,
+ "step": 1200
+ },
+ {
+ "epoch": 16.68471615720524,
+ "grad_norm": 0.015042162500321865,
+ "learning_rate": 0.0006,
+ "loss": 4.686626434326172,
+ "step": 1201
+ },
+ {
+ "epoch": 16.69868995633188,
+ "grad_norm": 0.0159393772482872,
+ "learning_rate": 0.0006,
+ "loss": 4.47656774520874,
+ "step": 1202
+ },
+ {
+ "epoch": 16.712663755458514,
+ "grad_norm": 0.01532787922769785,
+ "learning_rate": 0.0006,
+ "loss": 4.55678653717041,
+ "step": 1203
+ },
+ {
+ "epoch": 16.726637554585153,
+ "grad_norm": 0.016332164406776428,
+ "learning_rate": 0.0006,
+ "loss": 4.51441764831543,
+ "step": 1204
+ },
+ {
+ "epoch": 16.74061135371179,
+ "grad_norm": 0.018901726230978966,
+ "learning_rate": 0.0006,
+ "loss": 4.57227897644043,
+ "step": 1205
+ },
+ {
+ "epoch": 16.754585152838427,
+ "grad_norm": 0.018589433282613754,
+ "learning_rate": 0.0006,
+ "loss": 4.550070762634277,
+ "step": 1206
+ },
+ {
+ "epoch": 16.768558951965066,
+ "grad_norm": 0.01634034886956215,
+ "learning_rate": 0.0006,
+ "loss": 4.586910724639893,
+ "step": 1207
+ },
+ {
+ "epoch": 16.782532751091704,
+ "grad_norm": 0.017599433660507202,
+ "learning_rate": 0.0006,
+ "loss": 4.5891923904418945,
+ "step": 1208
+ },
+ {
+ "epoch": 16.79650655021834,
+ "grad_norm": 0.01878446340560913,
+ "learning_rate": 0.0006,
+ "loss": 4.57227087020874,
+ "step": 1209
+ },
+ {
+ "epoch": 16.810480349344978,
+ "grad_norm": 0.01803469844162464,
+ "learning_rate": 0.0006,
+ "loss": 4.583845138549805,
+ "step": 1210
+ },
+ {
+ "epoch": 16.824454148471617,
+ "grad_norm": 0.018505144864320755,
+ "learning_rate": 0.0006,
+ "loss": 4.561151504516602,
+ "step": 1211
+ },
+ {
+ "epoch": 16.83842794759825,
+ "grad_norm": 0.01920081488788128,
+ "learning_rate": 0.0006,
+ "loss": 4.506834030151367,
+ "step": 1212
+ },
+ {
+ "epoch": 16.85240174672489,
+ "grad_norm": 0.0177712831646204,
+ "learning_rate": 0.0006,
+ "loss": 4.557195663452148,
+ "step": 1213
+ },
+ {
+ "epoch": 16.86637554585153,
+ "grad_norm": 0.019113918766379356,
+ "learning_rate": 0.0006,
+ "loss": 4.651227951049805,
+ "step": 1214
+ },
+ {
+ "epoch": 16.880349344978168,
+ "grad_norm": 0.02012539468705654,
+ "learning_rate": 0.0006,
+ "loss": 4.4952239990234375,
+ "step": 1215
+ },
+ {
+ "epoch": 16.894323144104803,
+ "grad_norm": 0.019993646070361137,
+ "learning_rate": 0.0006,
+ "loss": 4.517433166503906,
+ "step": 1216
+ },
+ {
+ "epoch": 16.90829694323144,
+ "grad_norm": 0.020041441544890404,
+ "learning_rate": 0.0006,
+ "loss": 4.489407539367676,
+ "step": 1217
+ },
+ {
+ "epoch": 16.92227074235808,
+ "grad_norm": 0.0183611661195755,
+ "learning_rate": 0.0006,
+ "loss": 4.5435638427734375,
+ "step": 1218
+ },
+ {
+ "epoch": 16.936244541484715,
+ "grad_norm": 0.019138429313898087,
+ "learning_rate": 0.0006,
+ "loss": 4.545949459075928,
+ "step": 1219
+ },
+ {
+ "epoch": 16.950218340611354,
+ "grad_norm": 0.018554266542196274,
+ "learning_rate": 0.0006,
+ "loss": 4.579620361328125,
+ "step": 1220
+ },
+ {
+ "epoch": 16.964192139737992,
+ "grad_norm": 0.0193545650690794,
+ "learning_rate": 0.0006,
+ "loss": 4.588255405426025,
+ "step": 1221
+ },
+ {
+ "epoch": 16.978165938864628,
+ "grad_norm": 0.019389165565371513,
+ "learning_rate": 0.0006,
+ "loss": 4.520698547363281,
+ "step": 1222
+ },
+ {
+ "epoch": 16.992139737991266,
+ "grad_norm": 0.016601379960775375,
+ "learning_rate": 0.0006,
+ "loss": 4.4809064865112305,
+ "step": 1223
+ },
+ {
+ "epoch": 17.0,
+ "grad_norm": 0.01771303080022335,
+ "learning_rate": 0.0006,
+ "loss": 4.721122741699219,
+ "step": 1224
+ },
+ {
+ "epoch": 17.0,
+ "eval_loss": 4.7903666496276855,
+ "eval_runtime": 56.9826,
+ "eval_samples_per_second": 42.855,
+ "eval_steps_per_second": 1.351,
+ "step": 1224
+ },
+ {
+ "epoch": 17.01397379912664,
+ "grad_norm": 0.017720289528369904,
+ "learning_rate": 0.0006,
+ "loss": 4.523834705352783,
+ "step": 1225
+ },
+ {
+ "epoch": 17.027947598253274,
+ "grad_norm": 0.020137161016464233,
+ "learning_rate": 0.0006,
+ "loss": 4.574061870574951,
+ "step": 1226
+ },
+ {
+ "epoch": 17.041921397379912,
+ "grad_norm": 0.02002314105629921,
+ "learning_rate": 0.0006,
+ "loss": 4.512407302856445,
+ "step": 1227
+ },
+ {
+ "epoch": 17.05589519650655,
+ "grad_norm": 0.019628183916211128,
+ "learning_rate": 0.0006,
+ "loss": 4.502533435821533,
+ "step": 1228
+ },
+ {
+ "epoch": 17.069868995633186,
+ "grad_norm": 0.019808458164334297,
+ "learning_rate": 0.0006,
+ "loss": 4.463156700134277,
+ "step": 1229
+ },
+ {
+ "epoch": 17.083842794759825,
+ "grad_norm": 0.021250786259770393,
+ "learning_rate": 0.0006,
+ "loss": 4.622682094573975,
+ "step": 1230
+ },
+ {
+ "epoch": 17.097816593886463,
+ "grad_norm": 0.022031694650650024,
+ "learning_rate": 0.0006,
+ "loss": 4.451882839202881,
+ "step": 1231
+ },
+ {
+ "epoch": 17.111790393013102,
+ "grad_norm": 0.01898987777531147,
+ "learning_rate": 0.0006,
+ "loss": 4.529331684112549,
+ "step": 1232
+ },
+ {
+ "epoch": 17.125764192139737,
+ "grad_norm": 0.018749834969639778,
+ "learning_rate": 0.0006,
+ "loss": 4.435815334320068,
+ "step": 1233
+ },
+ {
+ "epoch": 17.139737991266376,
+ "grad_norm": 0.01896674931049347,
+ "learning_rate": 0.0006,
+ "loss": 4.484029293060303,
+ "step": 1234
+ },
+ {
+ "epoch": 17.153711790393015,
+ "grad_norm": 0.018286997452378273,
+ "learning_rate": 0.0006,
+ "loss": 4.523510932922363,
+ "step": 1235
+ },
+ {
+ "epoch": 17.16768558951965,
+ "grad_norm": 0.017273059114813805,
+ "learning_rate": 0.0006,
+ "loss": 4.451767444610596,
+ "step": 1236
+ },
+ {
+ "epoch": 17.18165938864629,
+ "grad_norm": 0.016499150544404984,
+ "learning_rate": 0.0006,
+ "loss": 4.5568647384643555,
+ "step": 1237
+ },
+ {
+ "epoch": 17.195633187772927,
+ "grad_norm": 0.015934469178318977,
+ "learning_rate": 0.0006,
+ "loss": 4.4639997482299805,
+ "step": 1238
+ },
+ {
+ "epoch": 17.209606986899562,
+ "grad_norm": 0.01714209094643593,
+ "learning_rate": 0.0006,
+ "loss": 4.547614097595215,
+ "step": 1239
+ },
+ {
+ "epoch": 17.2235807860262,
+ "grad_norm": 0.019167495891451836,
+ "learning_rate": 0.0006,
+ "loss": 4.571097373962402,
+ "step": 1240
+ },
+ {
+ "epoch": 17.23755458515284,
+ "grad_norm": 0.018276996910572052,
+ "learning_rate": 0.0006,
+ "loss": 4.489764213562012,
+ "step": 1241
+ },
+ {
+ "epoch": 17.251528384279474,
+ "grad_norm": 0.021296866238117218,
+ "learning_rate": 0.0006,
+ "loss": 4.526913642883301,
+ "step": 1242
+ },
+ {
+ "epoch": 17.265502183406113,
+ "grad_norm": 0.022606629878282547,
+ "learning_rate": 0.0006,
+ "loss": 4.521624565124512,
+ "step": 1243
+ },
+ {
+ "epoch": 17.27947598253275,
+ "grad_norm": 0.021204711869359016,
+ "learning_rate": 0.0006,
+ "loss": 4.500864028930664,
+ "step": 1244
+ },
+ {
+ "epoch": 17.29344978165939,
+ "grad_norm": 0.01988849975168705,
+ "learning_rate": 0.0006,
+ "loss": 4.45162296295166,
+ "step": 1245
+ },
+ {
+ "epoch": 17.307423580786025,
+ "grad_norm": 0.020268384367227554,
+ "learning_rate": 0.0006,
+ "loss": 4.536198616027832,
+ "step": 1246
+ },
+ {
+ "epoch": 17.321397379912664,
+ "grad_norm": 0.020185677334666252,
+ "learning_rate": 0.0006,
+ "loss": 4.440262794494629,
+ "step": 1247
+ },
+ {
+ "epoch": 17.335371179039303,
+ "grad_norm": 0.020106853917241096,
+ "learning_rate": 0.0006,
+ "loss": 4.475907325744629,
+ "step": 1248
+ },
+ {
+ "epoch": 17.349344978165938,
+ "grad_norm": 0.018883297219872475,
+ "learning_rate": 0.0006,
+ "loss": 4.640661716461182,
+ "step": 1249
+ },
+ {
+ "epoch": 17.363318777292577,
+ "grad_norm": 0.01852620206773281,
+ "learning_rate": 0.0006,
+ "loss": 4.578243255615234,
+ "step": 1250
+ },
+ {
+ "epoch": 17.377292576419215,
+ "grad_norm": 0.01886451430618763,
+ "learning_rate": 0.0006,
+ "loss": 4.546026229858398,
+ "step": 1251
+ },
+ {
+ "epoch": 17.39126637554585,
+ "grad_norm": 0.01777452416718006,
+ "learning_rate": 0.0006,
+ "loss": 4.521542549133301,
+ "step": 1252
+ },
+ {
+ "epoch": 17.40524017467249,
+ "grad_norm": 0.016626447439193726,
+ "learning_rate": 0.0006,
+ "loss": 4.52629280090332,
+ "step": 1253
+ },
+ {
+ "epoch": 17.419213973799128,
+ "grad_norm": 0.01599477231502533,
+ "learning_rate": 0.0006,
+ "loss": 4.4796037673950195,
+ "step": 1254
+ },
+ {
+ "epoch": 17.433187772925763,
+ "grad_norm": 0.017021458595991135,
+ "learning_rate": 0.0006,
+ "loss": 4.468813419342041,
+ "step": 1255
+ },
+ {
+ "epoch": 17.4471615720524,
+ "grad_norm": 0.016925333067774773,
+ "learning_rate": 0.0006,
+ "loss": 4.584898948669434,
+ "step": 1256
+ },
+ {
+ "epoch": 17.46113537117904,
+ "grad_norm": 0.01714998669922352,
+ "learning_rate": 0.0006,
+ "loss": 4.523544788360596,
+ "step": 1257
+ },
+ {
+ "epoch": 17.475109170305675,
+ "grad_norm": 0.01802288182079792,
+ "learning_rate": 0.0006,
+ "loss": 4.595581531524658,
+ "step": 1258
+ },
+ {
+ "epoch": 17.489082969432314,
+ "grad_norm": 0.0180289875715971,
+ "learning_rate": 0.0006,
+ "loss": 4.394440650939941,
+ "step": 1259
+ },
+ {
+ "epoch": 17.503056768558952,
+ "grad_norm": 0.018504245206713676,
+ "learning_rate": 0.0006,
+ "loss": 4.4794416427612305,
+ "step": 1260
+ },
+ {
+ "epoch": 17.51703056768559,
+ "grad_norm": 0.019973233342170715,
+ "learning_rate": 0.0006,
+ "loss": 4.4241437911987305,
+ "step": 1261
+ },
+ {
+ "epoch": 17.531004366812226,
+ "grad_norm": 0.020063083618879318,
+ "learning_rate": 0.0006,
+ "loss": 4.48068380355835,
+ "step": 1262
+ },
+ {
+ "epoch": 17.544978165938865,
+ "grad_norm": 0.02066640742123127,
+ "learning_rate": 0.0006,
+ "loss": 4.628993034362793,
+ "step": 1263
+ },
+ {
+ "epoch": 17.558951965065503,
+ "grad_norm": 0.02206416428089142,
+ "learning_rate": 0.0006,
+ "loss": 4.564790725708008,
+ "step": 1264
+ },
+ {
+ "epoch": 17.57292576419214,
+ "grad_norm": 0.020640335977077484,
+ "learning_rate": 0.0006,
+ "loss": 4.578705787658691,
+ "step": 1265
+ },
+ {
+ "epoch": 17.586899563318777,
+ "grad_norm": 0.019093405455350876,
+ "learning_rate": 0.0006,
+ "loss": 4.5023603439331055,
+ "step": 1266
+ },
+ {
+ "epoch": 17.600873362445416,
+ "grad_norm": 0.019459193572402,
+ "learning_rate": 0.0006,
+ "loss": 4.594531059265137,
+ "step": 1267
+ },
+ {
+ "epoch": 17.61484716157205,
+ "grad_norm": 0.02021171525120735,
+ "learning_rate": 0.0006,
+ "loss": 4.770299911499023,
+ "step": 1268
+ },
+ {
+ "epoch": 17.62882096069869,
+ "grad_norm": 0.021161191165447235,
+ "learning_rate": 0.0006,
+ "loss": 4.5380401611328125,
+ "step": 1269
+ },
+ {
+ "epoch": 17.64279475982533,
+ "grad_norm": 0.021013403311371803,
+ "learning_rate": 0.0006,
+ "loss": 4.455524921417236,
+ "step": 1270
+ },
+ {
+ "epoch": 17.656768558951963,
+ "grad_norm": 0.020336592569947243,
+ "learning_rate": 0.0006,
+ "loss": 4.533672332763672,
+ "step": 1271
+ },
+ {
+ "epoch": 17.670742358078602,
+ "grad_norm": 0.020159434527158737,
+ "learning_rate": 0.0006,
+ "loss": 4.596151351928711,
+ "step": 1272
+ },
+ {
+ "epoch": 17.68471615720524,
+ "grad_norm": 0.020432960242033005,
+ "learning_rate": 0.0006,
+ "loss": 4.55531644821167,
+ "step": 1273
+ },
+ {
+ "epoch": 17.69868995633188,
+ "grad_norm": 0.025466229766607285,
+ "learning_rate": 0.0006,
+ "loss": 4.566912651062012,
+ "step": 1274
+ },
+ {
+ "epoch": 17.712663755458514,
+ "grad_norm": 0.026537558063864708,
+ "learning_rate": 0.0006,
+ "loss": 4.526003837585449,
+ "step": 1275
+ },
+ {
+ "epoch": 17.726637554585153,
+ "grad_norm": 0.024397607892751694,
+ "learning_rate": 0.0006,
+ "loss": 4.541755676269531,
+ "step": 1276
+ },
+ {
+ "epoch": 17.74061135371179,
+ "grad_norm": 0.022062312811613083,
+ "learning_rate": 0.0006,
+ "loss": 4.6532793045043945,
+ "step": 1277
+ },
+ {
+ "epoch": 17.754585152838427,
+ "grad_norm": 0.02204282023012638,
+ "learning_rate": 0.0006,
+ "loss": 4.582825660705566,
+ "step": 1278
+ },
+ {
+ "epoch": 17.768558951965066,
+ "grad_norm": 0.020844416692852974,
+ "learning_rate": 0.0006,
+ "loss": 4.522336483001709,
+ "step": 1279
+ },
+ {
+ "epoch": 17.782532751091704,
+ "grad_norm": 0.019815703853964806,
+ "learning_rate": 0.0006,
+ "loss": 4.607398509979248,
+ "step": 1280
+ },
+ {
+ "epoch": 17.79650655021834,
+ "grad_norm": 0.016676954925060272,
+ "learning_rate": 0.0006,
+ "loss": 4.446206092834473,
+ "step": 1281
+ },
+ {
+ "epoch": 17.810480349344978,
+ "grad_norm": 0.0166452769190073,
+ "learning_rate": 0.0006,
+ "loss": 4.505931854248047,
+ "step": 1282
+ },
+ {
+ "epoch": 17.824454148471617,
+ "grad_norm": 0.015928907319903374,
+ "learning_rate": 0.0006,
+ "loss": 4.455267429351807,
+ "step": 1283
+ },
+ {
+ "epoch": 17.83842794759825,
+ "grad_norm": 0.015219560824334621,
+ "learning_rate": 0.0006,
+ "loss": 4.5623674392700195,
+ "step": 1284
+ },
+ {
+ "epoch": 17.85240174672489,
+ "grad_norm": 0.015125355683267117,
+ "learning_rate": 0.0006,
+ "loss": 4.554104804992676,
+ "step": 1285
+ },
+ {
+ "epoch": 17.86637554585153,
+ "grad_norm": 0.015612718649208546,
+ "learning_rate": 0.0006,
+ "loss": 4.484847068786621,
+ "step": 1286
+ },
+ {
+ "epoch": 17.880349344978168,
+ "grad_norm": 0.015215662308037281,
+ "learning_rate": 0.0006,
+ "loss": 4.4802093505859375,
+ "step": 1287
+ },
+ {
+ "epoch": 17.894323144104803,
+ "grad_norm": 0.01511684525758028,
+ "learning_rate": 0.0006,
+ "loss": 4.537298679351807,
+ "step": 1288
+ },
+ {
+ "epoch": 17.90829694323144,
+ "grad_norm": 0.013841322623193264,
+ "learning_rate": 0.0006,
+ "loss": 4.427488327026367,
+ "step": 1289
+ },
+ {
+ "epoch": 17.92227074235808,
+ "grad_norm": 0.015115964226424694,
+ "learning_rate": 0.0006,
+ "loss": 4.4547119140625,
+ "step": 1290
+ },
+ {
+ "epoch": 17.936244541484715,
+ "grad_norm": 0.016411006450653076,
+ "learning_rate": 0.0006,
+ "loss": 4.52994966506958,
+ "step": 1291
+ },
+ {
+ "epoch": 17.950218340611354,
+ "grad_norm": 0.017313152551651,
+ "learning_rate": 0.0006,
+ "loss": 4.5337653160095215,
+ "step": 1292
+ },
+ {
+ "epoch": 17.964192139737992,
+ "grad_norm": 0.016559796407818794,
+ "learning_rate": 0.0006,
+ "loss": 4.5277910232543945,
+ "step": 1293
+ },
+ {
+ "epoch": 17.978165938864628,
+ "grad_norm": 0.015854761004447937,
+ "learning_rate": 0.0006,
+ "loss": 4.6334309577941895,
+ "step": 1294
+ },
+ {
+ "epoch": 17.992139737991266,
+ "grad_norm": 0.015658238902688026,
+ "learning_rate": 0.0006,
+ "loss": 4.613372802734375,
+ "step": 1295
+ },
+ {
+ "epoch": 18.0,
+ "grad_norm": 0.0169155802577734,
+ "learning_rate": 0.0006,
+ "loss": 4.527356147766113,
+ "step": 1296
+ },
+ {
+ "epoch": 18.0,
+ "eval_loss": 4.764543056488037,
+ "eval_runtime": 56.5998,
+ "eval_samples_per_second": 43.145,
+ "eval_steps_per_second": 1.36,
+ "step": 1296
+ },
+ {
+ "epoch": 18.01397379912664,
+ "grad_norm": 0.015348963439464569,
+ "learning_rate": 0.0006,
+ "loss": 4.466423988342285,
+ "step": 1297
+ },
+ {
+ "epoch": 18.027947598253274,
+ "grad_norm": 0.01768588088452816,
+ "learning_rate": 0.0006,
+ "loss": 4.545933246612549,
+ "step": 1298
+ },
+ {
+ "epoch": 18.041921397379912,
+ "grad_norm": 0.017381226643919945,
+ "learning_rate": 0.0006,
+ "loss": 4.494060039520264,
+ "step": 1299
+ },
+ {
+ "epoch": 18.05589519650655,
+ "grad_norm": 0.019002335146069527,
+ "learning_rate": 0.0006,
+ "loss": 4.395200729370117,
+ "step": 1300
+ },
+ {
+ "epoch": 18.069868995633186,
+ "grad_norm": 0.01827932707965374,
+ "learning_rate": 0.0006,
+ "loss": 4.424848556518555,
+ "step": 1301
+ },
+ {
+ "epoch": 18.083842794759825,
+ "grad_norm": 0.016383031383156776,
+ "learning_rate": 0.0006,
+ "loss": 4.494722366333008,
+ "step": 1302
+ },
+ {
+ "epoch": 18.097816593886463,
+ "grad_norm": 0.016265157610177994,
+ "learning_rate": 0.0006,
+ "loss": 4.416522979736328,
+ "step": 1303
+ },
+ {
+ "epoch": 18.111790393013102,
+ "grad_norm": 0.01702386513352394,
+ "learning_rate": 0.0006,
+ "loss": 4.600103378295898,
+ "step": 1304
+ },
+ {
+ "epoch": 18.125764192139737,
+ "grad_norm": 0.018495498225092888,
+ "learning_rate": 0.0006,
+ "loss": 4.588401794433594,
+ "step": 1305
+ },
+ {
+ "epoch": 18.139737991266376,
+ "grad_norm": 0.017921684309840202,
+ "learning_rate": 0.0006,
+ "loss": 4.478075981140137,
+ "step": 1306
+ },
+ {
+ "epoch": 18.153711790393015,
+ "grad_norm": 0.01801278069615364,
+ "learning_rate": 0.0006,
+ "loss": 4.413186550140381,
+ "step": 1307
+ },
+ {
+ "epoch": 18.16768558951965,
+ "grad_norm": 0.017211275175213814,
+ "learning_rate": 0.0006,
+ "loss": 4.461341857910156,
+ "step": 1308
+ },
+ {
+ "epoch": 18.18165938864629,
+ "grad_norm": 0.017589328810572624,
+ "learning_rate": 0.0006,
+ "loss": 4.546969413757324,
+ "step": 1309
+ },
+ {
+ "epoch": 18.195633187772927,
+ "grad_norm": 0.01797698810696602,
+ "learning_rate": 0.0006,
+ "loss": 4.530797004699707,
+ "step": 1310
+ },
+ {
+ "epoch": 18.209606986899562,
+ "grad_norm": 0.020408619195222855,
+ "learning_rate": 0.0006,
+ "loss": 4.62999153137207,
+ "step": 1311
+ },
+ {
+ "epoch": 18.2235807860262,
+ "grad_norm": 0.01994038000702858,
+ "learning_rate": 0.0006,
+ "loss": 4.586811065673828,
+ "step": 1312
+ },
+ {
+ "epoch": 18.23755458515284,
+ "grad_norm": 0.019381340593099594,
+ "learning_rate": 0.0006,
+ "loss": 4.4769368171691895,
+ "step": 1313
+ },
+ {
+ "epoch": 18.251528384279474,
+ "grad_norm": 0.020484555512666702,
+ "learning_rate": 0.0006,
+ "loss": 4.524302959442139,
+ "step": 1314
+ },
+ {
+ "epoch": 18.265502183406113,
+ "grad_norm": 0.02426416426897049,
+ "learning_rate": 0.0006,
+ "loss": 4.561705589294434,
+ "step": 1315
+ },
+ {
+ "epoch": 18.27947598253275,
+ "grad_norm": 0.024748101830482483,
+ "learning_rate": 0.0006,
+ "loss": 4.4654316902160645,
+ "step": 1316
+ },
+ {
+ "epoch": 18.29344978165939,
+ "grad_norm": 0.020703328773379326,
+ "learning_rate": 0.0006,
+ "loss": 4.53033447265625,
+ "step": 1317
+ },
+ {
+ "epoch": 18.307423580786025,
+ "grad_norm": 0.01765458658337593,
+ "learning_rate": 0.0006,
+ "loss": 4.5154266357421875,
+ "step": 1318
+ },
+ {
+ "epoch": 18.321397379912664,
+ "grad_norm": 0.018646420910954475,
+ "learning_rate": 0.0006,
+ "loss": 4.5224714279174805,
+ "step": 1319
+ },
+ {
+ "epoch": 18.335371179039303,
+ "grad_norm": 0.01907474733889103,
+ "learning_rate": 0.0006,
+ "loss": 4.5055437088012695,
+ "step": 1320
+ },
+ {
+ "epoch": 18.349344978165938,
+ "grad_norm": 0.018179846927523613,
+ "learning_rate": 0.0006,
+ "loss": 4.462161064147949,
+ "step": 1321
+ },
+ {
+ "epoch": 18.363318777292577,
+ "grad_norm": 0.016697878018021584,
+ "learning_rate": 0.0006,
+ "loss": 4.473718643188477,
+ "step": 1322
+ },
+ {
+ "epoch": 18.377292576419215,
+ "grad_norm": 0.017504019662737846,
+ "learning_rate": 0.0006,
+ "loss": 4.474348068237305,
+ "step": 1323
+ },
+ {
+ "epoch": 18.39126637554585,
+ "grad_norm": 0.018889913335442543,
+ "learning_rate": 0.0006,
+ "loss": 4.4531636238098145,
+ "step": 1324
+ },
+ {
+ "epoch": 18.40524017467249,
+ "grad_norm": 0.019167251884937286,
+ "learning_rate": 0.0006,
+ "loss": 4.4278364181518555,
+ "step": 1325
+ },
+ {
+ "epoch": 18.419213973799128,
+ "grad_norm": 0.015895741060376167,
+ "learning_rate": 0.0006,
+ "loss": 4.474592208862305,
+ "step": 1326
+ },
+ {
+ "epoch": 18.433187772925763,
+ "grad_norm": 0.01531882956624031,
+ "learning_rate": 0.0006,
+ "loss": 4.403757572174072,
+ "step": 1327
+ },
+ {
+ "epoch": 18.4471615720524,
+ "grad_norm": 0.01565849967300892,
+ "learning_rate": 0.0006,
+ "loss": 4.507664680480957,
+ "step": 1328
+ },
+ {
+ "epoch": 18.46113537117904,
+ "grad_norm": 0.01602315716445446,
+ "learning_rate": 0.0006,
+ "loss": 4.5648908615112305,
+ "step": 1329
+ },
+ {
+ "epoch": 18.475109170305675,
+ "grad_norm": 0.016467366367578506,
+ "learning_rate": 0.0006,
+ "loss": 4.428762912750244,
+ "step": 1330
+ },
+ {
+ "epoch": 18.489082969432314,
+ "grad_norm": 0.01604127697646618,
+ "learning_rate": 0.0006,
+ "loss": 4.527771472930908,
+ "step": 1331
+ },
+ {
+ "epoch": 18.503056768558952,
+ "grad_norm": 0.018825586885213852,
+ "learning_rate": 0.0006,
+ "loss": 4.485386848449707,
+ "step": 1332
+ },
+ {
+ "epoch": 18.51703056768559,
+ "grad_norm": 0.019258007407188416,
+ "learning_rate": 0.0006,
+ "loss": 4.414972305297852,
+ "step": 1333
+ },
+ {
+ "epoch": 18.531004366812226,
+ "grad_norm": 0.018192021176218987,
+ "learning_rate": 0.0006,
+ "loss": 4.412166595458984,
+ "step": 1334
+ },
+ {
+ "epoch": 18.544978165938865,
+ "grad_norm": 0.016120698302984238,
+ "learning_rate": 0.0006,
+ "loss": 4.484317779541016,
+ "step": 1335
+ },
+ {
+ "epoch": 18.558951965065503,
+ "grad_norm": 0.016496622934937477,
+ "learning_rate": 0.0006,
+ "loss": 4.514986515045166,
+ "step": 1336
+ },
+ {
+ "epoch": 18.57292576419214,
+ "grad_norm": 0.01795116998255253,
+ "learning_rate": 0.0006,
+ "loss": 4.41473388671875,
+ "step": 1337
+ },
+ {
+ "epoch": 18.586899563318777,
+ "grad_norm": 0.019062276929616928,
+ "learning_rate": 0.0006,
+ "loss": 4.601996898651123,
+ "step": 1338
+ },
+ {
+ "epoch": 18.600873362445416,
+ "grad_norm": 0.017174631357192993,
+ "learning_rate": 0.0006,
+ "loss": 4.571218490600586,
+ "step": 1339
+ },
+ {
+ "epoch": 18.61484716157205,
+ "grad_norm": 0.014015606604516506,
+ "learning_rate": 0.0006,
+ "loss": 4.526606559753418,
+ "step": 1340
+ },
+ {
+ "epoch": 18.62882096069869,
+ "grad_norm": 0.015358510427176952,
+ "learning_rate": 0.0006,
+ "loss": 4.604286193847656,
+ "step": 1341
+ },
+ {
+ "epoch": 18.64279475982533,
+ "grad_norm": 0.01619391329586506,
+ "learning_rate": 0.0006,
+ "loss": 4.560883522033691,
+ "step": 1342
+ },
+ {
+ "epoch": 18.656768558951963,
+ "grad_norm": 0.015972964465618134,
+ "learning_rate": 0.0006,
+ "loss": 4.479755401611328,
+ "step": 1343
+ },
+ {
+ "epoch": 18.670742358078602,
+ "grad_norm": 0.016861025243997574,
+ "learning_rate": 0.0006,
+ "loss": 4.474471092224121,
+ "step": 1344
+ },
+ {
+ "epoch": 18.68471615720524,
+ "grad_norm": 0.01822078227996826,
+ "learning_rate": 0.0006,
+ "loss": 4.509489059448242,
+ "step": 1345
+ },
+ {
+ "epoch": 18.69868995633188,
+ "grad_norm": 0.019415007904171944,
+ "learning_rate": 0.0006,
+ "loss": 4.508941173553467,
+ "step": 1346
+ },
+ {
+ "epoch": 18.712663755458514,
+ "grad_norm": 0.019856126978993416,
+ "learning_rate": 0.0006,
+ "loss": 4.446670055389404,
+ "step": 1347
+ },
+ {
+ "epoch": 18.726637554585153,
+ "grad_norm": 0.01842520199716091,
+ "learning_rate": 0.0006,
+ "loss": 4.524422645568848,
+ "step": 1348
+ },
+ {
+ "epoch": 18.74061135371179,
+ "grad_norm": 0.017196480184793472,
+ "learning_rate": 0.0006,
+ "loss": 4.498466968536377,
+ "step": 1349
+ },
+ {
+ "epoch": 18.754585152838427,
+ "grad_norm": 0.01823010854423046,
+ "learning_rate": 0.0006,
+ "loss": 4.534684181213379,
+ "step": 1350
+ },
+ {
+ "epoch": 18.768558951965066,
+ "grad_norm": 0.020398065447807312,
+ "learning_rate": 0.0006,
+ "loss": 4.456494331359863,
+ "step": 1351
+ },
+ {
+ "epoch": 18.782532751091704,
+ "grad_norm": 0.022992262616753578,
+ "learning_rate": 0.0006,
+ "loss": 4.476374626159668,
+ "step": 1352
+ },
+ {
+ "epoch": 18.79650655021834,
+ "grad_norm": 0.02235420234501362,
+ "learning_rate": 0.0006,
+ "loss": 4.587776184082031,
+ "step": 1353
+ },
+ {
+ "epoch": 18.810480349344978,
+ "grad_norm": 0.02093626745045185,
+ "learning_rate": 0.0006,
+ "loss": 4.424172401428223,
+ "step": 1354
+ },
+ {
+ "epoch": 18.824454148471617,
+ "grad_norm": 0.01997302658855915,
+ "learning_rate": 0.0006,
+ "loss": 4.448066711425781,
+ "step": 1355
+ },
+ {
+ "epoch": 18.83842794759825,
+ "grad_norm": 0.018265459686517715,
+ "learning_rate": 0.0006,
+ "loss": 4.436610698699951,
+ "step": 1356
+ },
+ {
+ "epoch": 18.85240174672489,
+ "grad_norm": 0.01673833839595318,
+ "learning_rate": 0.0006,
+ "loss": 4.5502214431762695,
+ "step": 1357
+ },
+ {
+ "epoch": 18.86637554585153,
+ "grad_norm": 0.017908833920955658,
+ "learning_rate": 0.0006,
+ "loss": 4.550530433654785,
+ "step": 1358
+ },
+ {
+ "epoch": 18.880349344978168,
+ "grad_norm": 0.01796654611825943,
+ "learning_rate": 0.0006,
+ "loss": 4.531463146209717,
+ "step": 1359
+ },
+ {
+ "epoch": 18.894323144104803,
+ "grad_norm": 0.01646183803677559,
+ "learning_rate": 0.0006,
+ "loss": 4.498582363128662,
+ "step": 1360
+ },
+ {
+ "epoch": 18.90829694323144,
+ "grad_norm": 0.015430327504873276,
+ "learning_rate": 0.0006,
+ "loss": 4.52571439743042,
+ "step": 1361
+ },
+ {
+ "epoch": 18.92227074235808,
+ "grad_norm": 0.016527952626347542,
+ "learning_rate": 0.0006,
+ "loss": 4.547013759613037,
+ "step": 1362
+ },
+ {
+ "epoch": 18.936244541484715,
+ "grad_norm": 0.016263660043478012,
+ "learning_rate": 0.0006,
+ "loss": 4.468754291534424,
+ "step": 1363
+ },
+ {
+ "epoch": 18.950218340611354,
+ "grad_norm": 0.016347210854291916,
+ "learning_rate": 0.0006,
+ "loss": 4.450741767883301,
+ "step": 1364
+ },
+ {
+ "epoch": 18.964192139737992,
+ "grad_norm": 0.01701398566365242,
+ "learning_rate": 0.0006,
+ "loss": 4.534899711608887,
+ "step": 1365
+ },
+ {
+ "epoch": 18.978165938864628,
+ "grad_norm": 0.017693081870675087,
+ "learning_rate": 0.0006,
+ "loss": 4.460553169250488,
+ "step": 1366
+ },
+ {
+ "epoch": 18.992139737991266,
+ "grad_norm": 0.018746308982372284,
+ "learning_rate": 0.0006,
+ "loss": 4.538792610168457,
+ "step": 1367
+ },
+ {
+ "epoch": 19.0,
+ "grad_norm": 0.020099416375160217,
+ "learning_rate": 0.0006,
+ "loss": 4.449617385864258,
+ "step": 1368
+ },
+ {
+ "epoch": 19.0,
+ "eval_loss": 4.7756571769714355,
+ "eval_runtime": 57.1804,
+ "eval_samples_per_second": 42.707,
+ "eval_steps_per_second": 1.347,
+ "step": 1368
+ },
+ {
+ "epoch": 19.01397379912664,
+ "grad_norm": 0.018726933747529984,
+ "learning_rate": 0.0006,
+ "loss": 4.446059703826904,
+ "step": 1369
+ },
+ {
+ "epoch": 19.027947598253274,
+ "grad_norm": 0.01956087350845337,
+ "learning_rate": 0.0006,
+ "loss": 4.4520039558410645,
+ "step": 1370
+ },
+ {
+ "epoch": 19.041921397379912,
+ "grad_norm": 0.020092811435461044,
+ "learning_rate": 0.0006,
+ "loss": 4.543343544006348,
+ "step": 1371
+ },
+ {
+ "epoch": 19.05589519650655,
+ "grad_norm": 0.01843622885644436,
+ "learning_rate": 0.0006,
+ "loss": 4.535579204559326,
+ "step": 1372
+ },
+ {
+ "epoch": 19.069868995633186,
+ "grad_norm": 0.016989488154649734,
+ "learning_rate": 0.0006,
+ "loss": 4.52316951751709,
+ "step": 1373
+ },
+ {
+ "epoch": 19.083842794759825,
+ "grad_norm": 0.01805422455072403,
+ "learning_rate": 0.0006,
+ "loss": 4.5393266677856445,
+ "step": 1374
+ },
+ {
+ "epoch": 19.097816593886463,
+ "grad_norm": 0.01874629408121109,
+ "learning_rate": 0.0006,
+ "loss": 4.422059059143066,
+ "step": 1375
+ },
+ {
+ "epoch": 19.111790393013102,
+ "grad_norm": 0.01804221048951149,
+ "learning_rate": 0.0006,
+ "loss": 4.50508451461792,
+ "step": 1376
+ },
+ {
+ "epoch": 19.125764192139737,
+ "grad_norm": 0.01785474829375744,
+ "learning_rate": 0.0006,
+ "loss": 4.412599563598633,
+ "step": 1377
+ },
+ {
+ "epoch": 19.139737991266376,
+ "grad_norm": 0.01971791312098503,
+ "learning_rate": 0.0006,
+ "loss": 4.518318176269531,
+ "step": 1378
+ },
+ {
+ "epoch": 19.153711790393015,
+ "grad_norm": 0.02061633951961994,
+ "learning_rate": 0.0006,
+ "loss": 4.332393169403076,
+ "step": 1379
+ },
+ {
+ "epoch": 19.16768558951965,
+ "grad_norm": 0.020563429221510887,
+ "learning_rate": 0.0006,
+ "loss": 4.445863723754883,
+ "step": 1380
+ },
+ {
+ "epoch": 19.18165938864629,
+ "grad_norm": 0.019995713606476784,
+ "learning_rate": 0.0006,
+ "loss": 4.497312068939209,
+ "step": 1381
+ },
+ {
+ "epoch": 19.195633187772927,
+ "grad_norm": 0.019823070615530014,
+ "learning_rate": 0.0006,
+ "loss": 4.475085258483887,
+ "step": 1382
+ },
+ {
+ "epoch": 19.209606986899562,
+ "grad_norm": 0.019039355218410492,
+ "learning_rate": 0.0006,
+ "loss": 4.50432014465332,
+ "step": 1383
+ },
+ {
+ "epoch": 19.2235807860262,
+ "grad_norm": 0.017876973375678062,
+ "learning_rate": 0.0006,
+ "loss": 4.384788513183594,
+ "step": 1384
+ },
+ {
+ "epoch": 19.23755458515284,
+ "grad_norm": 0.018697699531912804,
+ "learning_rate": 0.0006,
+ "loss": 4.410778045654297,
+ "step": 1385
+ },
+ {
+ "epoch": 19.251528384279474,
+ "grad_norm": 0.018826806917786598,
+ "learning_rate": 0.0006,
+ "loss": 4.524034023284912,
+ "step": 1386
+ },
+ {
+ "epoch": 19.265502183406113,
+ "grad_norm": 0.01824226602911949,
+ "learning_rate": 0.0006,
+ "loss": 4.526393890380859,
+ "step": 1387
+ },
+ {
+ "epoch": 19.27947598253275,
+ "grad_norm": 0.01753208599984646,
+ "learning_rate": 0.0006,
+ "loss": 4.493703365325928,
+ "step": 1388
+ },
+ {
+ "epoch": 19.29344978165939,
+ "grad_norm": 0.01742432825267315,
+ "learning_rate": 0.0006,
+ "loss": 4.392831802368164,
+ "step": 1389
+ },
+ {
+ "epoch": 19.307423580786025,
+ "grad_norm": 0.017099356278777122,
+ "learning_rate": 0.0006,
+ "loss": 4.578668594360352,
+ "step": 1390
+ },
+ {
+ "epoch": 19.321397379912664,
+ "grad_norm": 0.017946023494005203,
+ "learning_rate": 0.0006,
+ "loss": 4.456718444824219,
+ "step": 1391
+ },
+ {
+ "epoch": 19.335371179039303,
+ "grad_norm": 0.017705943435430527,
+ "learning_rate": 0.0006,
+ "loss": 4.553682327270508,
+ "step": 1392
+ },
+ {
+ "epoch": 19.349344978165938,
+ "grad_norm": 0.016647523269057274,
+ "learning_rate": 0.0006,
+ "loss": 4.525016784667969,
+ "step": 1393
+ },
+ {
+ "epoch": 19.363318777292577,
+ "grad_norm": 0.015407416969537735,
+ "learning_rate": 0.0006,
+ "loss": 4.500231742858887,
+ "step": 1394
+ },
+ {
+ "epoch": 19.377292576419215,
+ "grad_norm": 0.017264019697904587,
+ "learning_rate": 0.0006,
+ "loss": 4.4615349769592285,
+ "step": 1395
+ },
+ {
+ "epoch": 19.39126637554585,
+ "grad_norm": 0.01871555484831333,
+ "learning_rate": 0.0006,
+ "loss": 4.281026840209961,
+ "step": 1396
+ },
+ {
+ "epoch": 19.40524017467249,
+ "grad_norm": 0.017361650243401527,
+ "learning_rate": 0.0006,
+ "loss": 4.479382514953613,
+ "step": 1397
+ },
+ {
+ "epoch": 19.419213973799128,
+ "grad_norm": 0.019781362265348434,
+ "learning_rate": 0.0006,
+ "loss": 4.402320861816406,
+ "step": 1398
+ },
+ {
+ "epoch": 19.433187772925763,
+ "grad_norm": 0.019043298438191414,
+ "learning_rate": 0.0006,
+ "loss": 4.476983070373535,
+ "step": 1399
+ },
+ {
+ "epoch": 19.4471615720524,
+ "grad_norm": 0.017137985676527023,
+ "learning_rate": 0.0006,
+ "loss": 4.418689250946045,
+ "step": 1400
+ },
+ {
+ "epoch": 19.46113537117904,
+ "grad_norm": 0.016481833532452583,
+ "learning_rate": 0.0006,
+ "loss": 4.579145431518555,
+ "step": 1401
+ },
+ {
+ "epoch": 19.475109170305675,
+ "grad_norm": 0.016507714986801147,
+ "learning_rate": 0.0006,
+ "loss": 4.519158363342285,
+ "step": 1402
+ },
+ {
+ "epoch": 19.489082969432314,
+ "grad_norm": 0.018092872574925423,
+ "learning_rate": 0.0006,
+ "loss": 4.513748645782471,
+ "step": 1403
+ },
+ {
+ "epoch": 19.503056768558952,
+ "grad_norm": 0.018153440207242966,
+ "learning_rate": 0.0006,
+ "loss": 4.645384311676025,
+ "step": 1404
+ },
+ {
+ "epoch": 19.51703056768559,
+ "grad_norm": 0.01657119207084179,
+ "learning_rate": 0.0006,
+ "loss": 4.399757385253906,
+ "step": 1405
+ },
+ {
+ "epoch": 19.531004366812226,
+ "grad_norm": 0.0156280305236578,
+ "learning_rate": 0.0006,
+ "loss": 4.467836380004883,
+ "step": 1406
+ },
+ {
+ "epoch": 19.544978165938865,
+ "grad_norm": 0.016838403418660164,
+ "learning_rate": 0.0006,
+ "loss": 4.529176235198975,
+ "step": 1407
+ },
+ {
+ "epoch": 19.558951965065503,
+ "grad_norm": 0.01743236929178238,
+ "learning_rate": 0.0006,
+ "loss": 4.440032482147217,
+ "step": 1408
+ },
+ {
+ "epoch": 19.57292576419214,
+ "grad_norm": 0.01553610060364008,
+ "learning_rate": 0.0006,
+ "loss": 4.425588607788086,
+ "step": 1409
+ },
+ {
+ "epoch": 19.586899563318777,
+ "grad_norm": 0.017259271815419197,
+ "learning_rate": 0.0006,
+ "loss": 4.405452728271484,
+ "step": 1410
+ },
+ {
+ "epoch": 19.600873362445416,
+ "grad_norm": 0.015568481758236885,
+ "learning_rate": 0.0006,
+ "loss": 4.444279193878174,
+ "step": 1411
+ },
+ {
+ "epoch": 19.61484716157205,
+ "grad_norm": 0.015331248752772808,
+ "learning_rate": 0.0006,
+ "loss": 4.451846122741699,
+ "step": 1412
+ },
+ {
+ "epoch": 19.62882096069869,
+ "grad_norm": 0.015667378902435303,
+ "learning_rate": 0.0006,
+ "loss": 4.533931255340576,
+ "step": 1413
+ },
+ {
+ "epoch": 19.64279475982533,
+ "grad_norm": 0.016177602112293243,
+ "learning_rate": 0.0006,
+ "loss": 4.467313289642334,
+ "step": 1414
+ },
+ {
+ "epoch": 19.656768558951963,
+ "grad_norm": 0.01733129471540451,
+ "learning_rate": 0.0006,
+ "loss": 4.496893882751465,
+ "step": 1415
+ },
+ {
+ "epoch": 19.670742358078602,
+ "grad_norm": 0.016539594158530235,
+ "learning_rate": 0.0006,
+ "loss": 4.401961326599121,
+ "step": 1416
+ },
+ {
+ "epoch": 19.68471615720524,
+ "grad_norm": 0.016058551147580147,
+ "learning_rate": 0.0006,
+ "loss": 4.503375053405762,
+ "step": 1417
+ },
+ {
+ "epoch": 19.69868995633188,
+ "grad_norm": 0.01715777814388275,
+ "learning_rate": 0.0006,
+ "loss": 4.3808369636535645,
+ "step": 1418
+ },
+ {
+ "epoch": 19.712663755458514,
+ "grad_norm": 0.01681119203567505,
+ "learning_rate": 0.0006,
+ "loss": 4.5469560623168945,
+ "step": 1419
+ },
+ {
+ "epoch": 19.726637554585153,
+ "grad_norm": 0.01634550653398037,
+ "learning_rate": 0.0006,
+ "loss": 4.328197956085205,
+ "step": 1420
+ },
+ {
+ "epoch": 19.74061135371179,
+ "grad_norm": 0.01549347210675478,
+ "learning_rate": 0.0006,
+ "loss": 4.534621238708496,
+ "step": 1421
+ },
+ {
+ "epoch": 19.754585152838427,
+ "grad_norm": 0.01654665172100067,
+ "learning_rate": 0.0006,
+ "loss": 4.4724225997924805,
+ "step": 1422
+ },
+ {
+ "epoch": 19.768558951965066,
+ "grad_norm": 0.017119025811553,
+ "learning_rate": 0.0006,
+ "loss": 4.3897247314453125,
+ "step": 1423
+ },
+ {
+ "epoch": 19.782532751091704,
+ "grad_norm": 0.01603223942220211,
+ "learning_rate": 0.0006,
+ "loss": 4.412731170654297,
+ "step": 1424
+ },
+ {
+ "epoch": 19.79650655021834,
+ "grad_norm": 0.016774067655205727,
+ "learning_rate": 0.0006,
+ "loss": 4.54707145690918,
+ "step": 1425
+ },
+ {
+ "epoch": 19.810480349344978,
+ "grad_norm": 0.016453025862574577,
+ "learning_rate": 0.0006,
+ "loss": 4.462501049041748,
+ "step": 1426
+ },
+ {
+ "epoch": 19.824454148471617,
+ "grad_norm": 0.01597629487514496,
+ "learning_rate": 0.0006,
+ "loss": 4.229758262634277,
+ "step": 1427
+ },
+ {
+ "epoch": 19.83842794759825,
+ "grad_norm": 0.017999034374952316,
+ "learning_rate": 0.0006,
+ "loss": 4.389263153076172,
+ "step": 1428
+ },
+ {
+ "epoch": 19.85240174672489,
+ "grad_norm": 0.02165898121893406,
+ "learning_rate": 0.0006,
+ "loss": 4.342130661010742,
+ "step": 1429
+ },
+ {
+ "epoch": 19.86637554585153,
+ "grad_norm": 0.02222822792828083,
+ "learning_rate": 0.0006,
+ "loss": 4.492661476135254,
+ "step": 1430
+ },
+ {
+ "epoch": 19.880349344978168,
+ "grad_norm": 0.019378066062927246,
+ "learning_rate": 0.0006,
+ "loss": 4.4614105224609375,
+ "step": 1431
+ },
+ {
+ "epoch": 19.894323144104803,
+ "grad_norm": 0.019908083602786064,
+ "learning_rate": 0.0006,
+ "loss": 4.370929718017578,
+ "step": 1432
+ },
+ {
+ "epoch": 19.90829694323144,
+ "grad_norm": 0.022296471521258354,
+ "learning_rate": 0.0006,
+ "loss": 4.438320159912109,
+ "step": 1433
+ },
+ {
+ "epoch": 19.92227074235808,
+ "grad_norm": 0.021810343489050865,
+ "learning_rate": 0.0006,
+ "loss": 4.462067127227783,
+ "step": 1434
+ },
+ {
+ "epoch": 19.936244541484715,
+ "grad_norm": 0.02129209227859974,
+ "learning_rate": 0.0006,
+ "loss": 4.541261196136475,
+ "step": 1435
+ },
+ {
+ "epoch": 19.950218340611354,
+ "grad_norm": 0.02079629711806774,
+ "learning_rate": 0.0006,
+ "loss": 4.36712646484375,
+ "step": 1436
+ },
+ {
+ "epoch": 19.964192139737992,
+ "grad_norm": 0.022501740604639053,
+ "learning_rate": 0.0006,
+ "loss": 4.5279083251953125,
+ "step": 1437
+ },
+ {
+ "epoch": 19.978165938864628,
+ "grad_norm": 0.019121676683425903,
+ "learning_rate": 0.0006,
+ "loss": 4.477789878845215,
+ "step": 1438
+ },
+ {
+ "epoch": 19.992139737991266,
+ "grad_norm": 0.017429588362574577,
+ "learning_rate": 0.0006,
+ "loss": 4.4023847579956055,
+ "step": 1439
+ },
+ {
+ "epoch": 20.0,
+ "grad_norm": 0.018857665359973907,
+ "learning_rate": 0.0006,
+ "loss": 4.494743347167969,
+ "step": 1440
+ },
+ {
+ "epoch": 20.0,
+ "eval_loss": 4.694555759429932,
+ "eval_runtime": 56.4146,
+ "eval_samples_per_second": 43.287,
+ "eval_steps_per_second": 1.365,
+ "step": 1440
+ },
+ {
+ "epoch": 20.01397379912664,
+ "grad_norm": 0.016505861654877663,
+ "learning_rate": 0.0006,
+ "loss": 4.506229877471924,
+ "step": 1441
+ },
+ {
+ "epoch": 20.027947598253274,
+ "grad_norm": 0.017571842297911644,
+ "learning_rate": 0.0006,
+ "loss": 4.384073257446289,
+ "step": 1442
+ },
+ {
+ "epoch": 20.041921397379912,
+ "grad_norm": 0.019398365169763565,
+ "learning_rate": 0.0006,
+ "loss": 4.482177734375,
+ "step": 1443
+ },
+ {
+ "epoch": 20.05589519650655,
+ "grad_norm": 0.02011062018573284,
+ "learning_rate": 0.0006,
+ "loss": 4.440773963928223,
+ "step": 1444
+ },
+ {
+ "epoch": 20.069868995633186,
+ "grad_norm": 0.019026953727006912,
+ "learning_rate": 0.0006,
+ "loss": 4.507168769836426,
+ "step": 1445
+ },
+ {
+ "epoch": 20.083842794759825,
+ "grad_norm": 0.020506031811237335,
+ "learning_rate": 0.0006,
+ "loss": 4.469079971313477,
+ "step": 1446
+ },
+ {
+ "epoch": 20.097816593886463,
+ "grad_norm": 0.019608162343502045,
+ "learning_rate": 0.0006,
+ "loss": 4.395927429199219,
+ "step": 1447
+ },
+ {
+ "epoch": 20.111790393013102,
+ "grad_norm": 0.019419124349951744,
+ "learning_rate": 0.0006,
+ "loss": 4.484235763549805,
+ "step": 1448
+ },
+ {
+ "epoch": 20.125764192139737,
+ "grad_norm": 0.021775128319859505,
+ "learning_rate": 0.0006,
+ "loss": 4.333173751831055,
+ "step": 1449
+ },
+ {
+ "epoch": 20.139737991266376,
+ "grad_norm": 0.025512272492051125,
+ "learning_rate": 0.0006,
+ "loss": 4.460362434387207,
+ "step": 1450
+ },
+ {
+ "epoch": 20.153711790393015,
+ "grad_norm": 0.025157401338219643,
+ "learning_rate": 0.0006,
+ "loss": 4.4361395835876465,
+ "step": 1451
+ },
+ {
+ "epoch": 20.16768558951965,
+ "grad_norm": 0.019862616434693336,
+ "learning_rate": 0.0006,
+ "loss": 4.426807403564453,
+ "step": 1452
+ },
+ {
+ "epoch": 20.18165938864629,
+ "grad_norm": 0.01932152360677719,
+ "learning_rate": 0.0006,
+ "loss": 4.375677108764648,
+ "step": 1453
+ },
+ {
+ "epoch": 20.195633187772927,
+ "grad_norm": 0.01760999858379364,
+ "learning_rate": 0.0006,
+ "loss": 4.42259407043457,
+ "step": 1454
+ },
+ {
+ "epoch": 20.209606986899562,
+ "grad_norm": 0.016998112201690674,
+ "learning_rate": 0.0006,
+ "loss": 4.481280326843262,
+ "step": 1455
+ },
+ {
+ "epoch": 20.2235807860262,
+ "grad_norm": 0.017199190333485603,
+ "learning_rate": 0.0006,
+ "loss": 4.496712684631348,
+ "step": 1456
+ },
+ {
+ "epoch": 20.23755458515284,
+ "grad_norm": 0.016833679750561714,
+ "learning_rate": 0.0006,
+ "loss": 4.40736722946167,
+ "step": 1457
+ },
+ {
+ "epoch": 20.251528384279474,
+ "grad_norm": 0.017546730116009712,
+ "learning_rate": 0.0006,
+ "loss": 4.4915266036987305,
+ "step": 1458
+ },
+ {
+ "epoch": 20.265502183406113,
+ "grad_norm": 0.016595078632235527,
+ "learning_rate": 0.0006,
+ "loss": 4.460815906524658,
+ "step": 1459
+ },
+ {
+ "epoch": 20.27947598253275,
+ "grad_norm": 0.017728568986058235,
+ "learning_rate": 0.0006,
+ "loss": 4.346048831939697,
+ "step": 1460
+ },
+ {
+ "epoch": 20.29344978165939,
+ "grad_norm": 0.01849197782576084,
+ "learning_rate": 0.0006,
+ "loss": 4.399681091308594,
+ "step": 1461
+ },
+ {
+ "epoch": 20.307423580786025,
+ "grad_norm": 0.018336588516831398,
+ "learning_rate": 0.0006,
+ "loss": 4.414183616638184,
+ "step": 1462
+ },
+ {
+ "epoch": 20.321397379912664,
+ "grad_norm": 0.018086519092321396,
+ "learning_rate": 0.0006,
+ "loss": 4.346586227416992,
+ "step": 1463
+ },
+ {
+ "epoch": 20.335371179039303,
+ "grad_norm": 0.019697299227118492,
+ "learning_rate": 0.0006,
+ "loss": 4.4420952796936035,
+ "step": 1464
+ },
+ {
+ "epoch": 20.349344978165938,
+ "grad_norm": 0.02198074571788311,
+ "learning_rate": 0.0006,
+ "loss": 4.588729381561279,
+ "step": 1465
+ },
+ {
+ "epoch": 20.363318777292577,
+ "grad_norm": 0.022813329473137856,
+ "learning_rate": 0.0006,
+ "loss": 4.49729585647583,
+ "step": 1466
+ },
+ {
+ "epoch": 20.377292576419215,
+ "grad_norm": 0.022006649523973465,
+ "learning_rate": 0.0006,
+ "loss": 4.475641250610352,
+ "step": 1467
+ },
+ {
+ "epoch": 20.39126637554585,
+ "grad_norm": 0.023188438266515732,
+ "learning_rate": 0.0006,
+ "loss": 4.404322624206543,
+ "step": 1468
+ },
+ {
+ "epoch": 20.40524017467249,
+ "grad_norm": 0.022226877510547638,
+ "learning_rate": 0.0006,
+ "loss": 4.50904655456543,
+ "step": 1469
+ },
+ {
+ "epoch": 20.419213973799128,
+ "grad_norm": 0.021082637831568718,
+ "learning_rate": 0.0006,
+ "loss": 4.339011192321777,
+ "step": 1470
+ },
+ {
+ "epoch": 20.433187772925763,
+ "grad_norm": 0.02013542130589485,
+ "learning_rate": 0.0006,
+ "loss": 4.516942024230957,
+ "step": 1471
+ },
+ {
+ "epoch": 20.4471615720524,
+ "grad_norm": 0.01898968778550625,
+ "learning_rate": 0.0006,
+ "loss": 4.315804481506348,
+ "step": 1472
+ },
+ {
+ "epoch": 20.46113537117904,
+ "grad_norm": 0.019450880587100983,
+ "learning_rate": 0.0006,
+ "loss": 4.4794921875,
+ "step": 1473
+ },
+ {
+ "epoch": 20.475109170305675,
+ "grad_norm": 0.018577802926301956,
+ "learning_rate": 0.0006,
+ "loss": 4.422323703765869,
+ "step": 1474
+ },
+ {
+ "epoch": 20.489082969432314,
+ "grad_norm": 0.01778412237763405,
+ "learning_rate": 0.0006,
+ "loss": 4.429789066314697,
+ "step": 1475
+ },
+ {
+ "epoch": 20.503056768558952,
+ "grad_norm": 0.017134476453065872,
+ "learning_rate": 0.0006,
+ "loss": 4.416459083557129,
+ "step": 1476
+ },
+ {
+ "epoch": 20.51703056768559,
+ "grad_norm": 0.017153557389974594,
+ "learning_rate": 0.0006,
+ "loss": 4.446170806884766,
+ "step": 1477
+ },
+ {
+ "epoch": 20.531004366812226,
+ "grad_norm": 0.014998532831668854,
+ "learning_rate": 0.0006,
+ "loss": 4.419447898864746,
+ "step": 1478
+ },
+ {
+ "epoch": 20.544978165938865,
+ "grad_norm": 0.015877850353717804,
+ "learning_rate": 0.0006,
+ "loss": 4.312338829040527,
+ "step": 1479
+ },
+ {
+ "epoch": 20.558951965065503,
+ "grad_norm": 0.01569489948451519,
+ "learning_rate": 0.0006,
+ "loss": 4.572905540466309,
+ "step": 1480
+ },
+ {
+ "epoch": 20.57292576419214,
+ "grad_norm": 0.01591474376618862,
+ "learning_rate": 0.0006,
+ "loss": 4.414907455444336,
+ "step": 1481
+ },
+ {
+ "epoch": 20.586899563318777,
+ "grad_norm": 0.015395903028547764,
+ "learning_rate": 0.0006,
+ "loss": 4.549131393432617,
+ "step": 1482
+ },
+ {
+ "epoch": 20.600873362445416,
+ "grad_norm": 0.015715358778834343,
+ "learning_rate": 0.0006,
+ "loss": 4.530194282531738,
+ "step": 1483
+ },
+ {
+ "epoch": 20.61484716157205,
+ "grad_norm": 0.016042975708842278,
+ "learning_rate": 0.0006,
+ "loss": 4.322149276733398,
+ "step": 1484
+ },
+ {
+ "epoch": 20.62882096069869,
+ "grad_norm": 0.01664186641573906,
+ "learning_rate": 0.0006,
+ "loss": 4.424866676330566,
+ "step": 1485
+ },
+ {
+ "epoch": 20.64279475982533,
+ "grad_norm": 0.01631534844636917,
+ "learning_rate": 0.0006,
+ "loss": 4.441752910614014,
+ "step": 1486
+ },
+ {
+ "epoch": 20.656768558951963,
+ "grad_norm": 0.014057058840990067,
+ "learning_rate": 0.0006,
+ "loss": 4.399982452392578,
+ "step": 1487
+ },
+ {
+ "epoch": 20.670742358078602,
+ "grad_norm": 0.015076273120939732,
+ "learning_rate": 0.0006,
+ "loss": 4.441324234008789,
+ "step": 1488
+ },
+ {
+ "epoch": 20.68471615720524,
+ "grad_norm": 0.014138277620077133,
+ "learning_rate": 0.0006,
+ "loss": 4.4312543869018555,
+ "step": 1489
+ },
+ {
+ "epoch": 20.69868995633188,
+ "grad_norm": 0.013762440532445908,
+ "learning_rate": 0.0006,
+ "loss": 4.387807846069336,
+ "step": 1490
+ },
+ {
+ "epoch": 20.712663755458514,
+ "grad_norm": 0.015058411285281181,
+ "learning_rate": 0.0006,
+ "loss": 4.594747543334961,
+ "step": 1491
+ },
+ {
+ "epoch": 20.726637554585153,
+ "grad_norm": 0.013827123679220676,
+ "learning_rate": 0.0006,
+ "loss": 4.434604167938232,
+ "step": 1492
+ },
+ {
+ "epoch": 20.74061135371179,
+ "grad_norm": 0.015315295197069645,
+ "learning_rate": 0.0006,
+ "loss": 4.588066101074219,
+ "step": 1493
+ },
+ {
+ "epoch": 20.754585152838427,
+ "grad_norm": 0.018678732216358185,
+ "learning_rate": 0.0006,
+ "loss": 4.3514204025268555,
+ "step": 1494
+ },
+ {
+ "epoch": 20.768558951965066,
+ "grad_norm": 0.019521350041031837,
+ "learning_rate": 0.0006,
+ "loss": 4.437658786773682,
+ "step": 1495
+ },
+ {
+ "epoch": 20.782532751091704,
+ "grad_norm": 0.018363403156399727,
+ "learning_rate": 0.0006,
+ "loss": 4.495002269744873,
+ "step": 1496
+ },
+ {
+ "epoch": 20.79650655021834,
+ "grad_norm": 0.016830725595355034,
+ "learning_rate": 0.0006,
+ "loss": 4.402471542358398,
+ "step": 1497
+ },
+ {
+ "epoch": 20.810480349344978,
+ "grad_norm": 0.01722072623670101,
+ "learning_rate": 0.0006,
+ "loss": 4.437653541564941,
+ "step": 1498
+ },
+ {
+ "epoch": 20.824454148471617,
+ "grad_norm": 0.017459599301218987,
+ "learning_rate": 0.0006,
+ "loss": 4.366603851318359,
+ "step": 1499
+ },
+ {
+ "epoch": 20.83842794759825,
+ "grad_norm": 0.018902219831943512,
+ "learning_rate": 0.0006,
+ "loss": 4.501561164855957,
+ "step": 1500
+ },
+ {
+ "epoch": 20.85240174672489,
+ "grad_norm": 0.020814096555113792,
+ "learning_rate": 0.0006,
+ "loss": 4.535295486450195,
+ "step": 1501
+ },
+ {
+ "epoch": 20.86637554585153,
+ "grad_norm": 0.018990976735949516,
+ "learning_rate": 0.0006,
+ "loss": 4.333580017089844,
+ "step": 1502
+ },
+ {
+ "epoch": 20.880349344978168,
+ "grad_norm": 0.01837880350649357,
+ "learning_rate": 0.0006,
+ "loss": 4.313388824462891,
+ "step": 1503
+ },
+ {
+ "epoch": 20.894323144104803,
+ "grad_norm": 0.019820360466837883,
+ "learning_rate": 0.0006,
+ "loss": 4.388157844543457,
+ "step": 1504
+ },
+ {
+ "epoch": 20.90829694323144,
+ "grad_norm": 0.020181143656373024,
+ "learning_rate": 0.0006,
+ "loss": 4.572224140167236,
+ "step": 1505
+ },
+ {
+ "epoch": 20.92227074235808,
+ "grad_norm": 0.021206077188253403,
+ "learning_rate": 0.0006,
+ "loss": 4.3127970695495605,
+ "step": 1506
+ },
+ {
+ "epoch": 20.936244541484715,
+ "grad_norm": 0.017962060868740082,
+ "learning_rate": 0.0006,
+ "loss": 4.384734630584717,
+ "step": 1507
+ },
+ {
+ "epoch": 20.950218340611354,
+ "grad_norm": 0.01737220585346222,
+ "learning_rate": 0.0006,
+ "loss": 4.4075751304626465,
+ "step": 1508
+ },
+ {
+ "epoch": 20.964192139737992,
+ "grad_norm": 0.01839268207550049,
+ "learning_rate": 0.0006,
+ "loss": 4.553779125213623,
+ "step": 1509
+ },
+ {
+ "epoch": 20.978165938864628,
+ "grad_norm": 0.020102640613913536,
+ "learning_rate": 0.0006,
+ "loss": 4.399398326873779,
+ "step": 1510
+ },
+ {
+ "epoch": 20.992139737991266,
+ "grad_norm": 0.018831463530659676,
+ "learning_rate": 0.0006,
+ "loss": 4.358499526977539,
+ "step": 1511
+ },
+ {
+ "epoch": 21.0,
+ "grad_norm": 0.018254734575748444,
+ "learning_rate": 0.0006,
+ "loss": 4.384119510650635,
+ "step": 1512
+ },
+ {
+ "epoch": 21.0,
+ "eval_loss": 4.7524237632751465,
+ "eval_runtime": 57.8812,
+ "eval_samples_per_second": 42.19,
+ "eval_steps_per_second": 1.33,
+ "step": 1512
+ },
+ {
+ "epoch": 21.01397379912664,
+ "grad_norm": 0.019728587940335274,
+ "learning_rate": 0.0006,
+ "loss": 4.32363224029541,
+ "step": 1513
+ },
+ {
+ "epoch": 21.027947598253274,
+ "grad_norm": 0.02216625027358532,
+ "learning_rate": 0.0006,
+ "loss": 4.436988353729248,
+ "step": 1514
+ },
+ {
+ "epoch": 21.041921397379912,
+ "grad_norm": 0.02227606810629368,
+ "learning_rate": 0.0006,
+ "loss": 4.406063556671143,
+ "step": 1515
+ },
+ {
+ "epoch": 21.05589519650655,
+ "grad_norm": 0.020187009125947952,
+ "learning_rate": 0.0006,
+ "loss": 4.324402332305908,
+ "step": 1516
+ },
+ {
+ "epoch": 21.069868995633186,
+ "grad_norm": 0.02095157466828823,
+ "learning_rate": 0.0006,
+ "loss": 4.456544876098633,
+ "step": 1517
+ },
+ {
+ "epoch": 21.083842794759825,
+ "grad_norm": 0.02146013453602791,
+ "learning_rate": 0.0006,
+ "loss": 4.368387222290039,
+ "step": 1518
+ },
+ {
+ "epoch": 21.097816593886463,
+ "grad_norm": 0.02162035182118416,
+ "learning_rate": 0.0006,
+ "loss": 4.48838996887207,
+ "step": 1519
+ },
+ {
+ "epoch": 21.111790393013102,
+ "grad_norm": 0.021745676174759865,
+ "learning_rate": 0.0006,
+ "loss": 4.30778694152832,
+ "step": 1520
+ },
+ {
+ "epoch": 21.125764192139737,
+ "grad_norm": 0.02173588052392006,
+ "learning_rate": 0.0006,
+ "loss": 4.433149814605713,
+ "step": 1521
+ },
+ {
+ "epoch": 21.139737991266376,
+ "grad_norm": 0.020950648933649063,
+ "learning_rate": 0.0006,
+ "loss": 4.520089149475098,
+ "step": 1522
+ },
+ {
+ "epoch": 21.153711790393015,
+ "grad_norm": 0.020684752613306046,
+ "learning_rate": 0.0006,
+ "loss": 4.372437477111816,
+ "step": 1523
+ },
+ {
+ "epoch": 21.16768558951965,
+ "grad_norm": 0.019472869113087654,
+ "learning_rate": 0.0006,
+ "loss": 4.437015533447266,
+ "step": 1524
+ },
+ {
+ "epoch": 21.18165938864629,
+ "grad_norm": 0.01846962980926037,
+ "learning_rate": 0.0006,
+ "loss": 4.449531555175781,
+ "step": 1525
+ },
+ {
+ "epoch": 21.195633187772927,
+ "grad_norm": 0.017377803102135658,
+ "learning_rate": 0.0006,
+ "loss": 4.458279132843018,
+ "step": 1526
+ },
+ {
+ "epoch": 21.209606986899562,
+ "grad_norm": 0.016002200543880463,
+ "learning_rate": 0.0006,
+ "loss": 4.266297817230225,
+ "step": 1527
+ },
+ {
+ "epoch": 21.2235807860262,
+ "grad_norm": 0.015575903467833996,
+ "learning_rate": 0.0006,
+ "loss": 4.342632293701172,
+ "step": 1528
+ },
+ {
+ "epoch": 21.23755458515284,
+ "grad_norm": 0.015634698793292046,
+ "learning_rate": 0.0006,
+ "loss": 4.338623523712158,
+ "step": 1529
+ },
+ {
+ "epoch": 21.251528384279474,
+ "grad_norm": 0.01449581328779459,
+ "learning_rate": 0.0006,
+ "loss": 4.354330062866211,
+ "step": 1530
+ },
+ {
+ "epoch": 21.265502183406113,
+ "grad_norm": 0.01487616915255785,
+ "learning_rate": 0.0006,
+ "loss": 4.486835479736328,
+ "step": 1531
+ },
+ {
+ "epoch": 21.27947598253275,
+ "grad_norm": 0.016346555203199387,
+ "learning_rate": 0.0006,
+ "loss": 4.407098770141602,
+ "step": 1532
+ },
+ {
+ "epoch": 21.29344978165939,
+ "grad_norm": 0.01682182401418686,
+ "learning_rate": 0.0006,
+ "loss": 4.436809539794922,
+ "step": 1533
+ },
+ {
+ "epoch": 21.307423580786025,
+ "grad_norm": 0.017704004421830177,
+ "learning_rate": 0.0006,
+ "loss": 4.3854780197143555,
+ "step": 1534
+ },
+ {
+ "epoch": 21.321397379912664,
+ "grad_norm": 0.018437420949339867,
+ "learning_rate": 0.0006,
+ "loss": 4.413283824920654,
+ "step": 1535
+ },
+ {
+ "epoch": 21.335371179039303,
+ "grad_norm": 0.016911081969738007,
+ "learning_rate": 0.0006,
+ "loss": 4.357306003570557,
+ "step": 1536
+ },
+ {
+ "epoch": 21.349344978165938,
+ "grad_norm": 0.01642223633825779,
+ "learning_rate": 0.0006,
+ "loss": 4.422322750091553,
+ "step": 1537
+ },
+ {
+ "epoch": 21.363318777292577,
+ "grad_norm": 0.01780877821147442,
+ "learning_rate": 0.0006,
+ "loss": 4.387620449066162,
+ "step": 1538
+ },
+ {
+ "epoch": 21.377292576419215,
+ "grad_norm": 0.01653975248336792,
+ "learning_rate": 0.0006,
+ "loss": 4.327267646789551,
+ "step": 1539
+ },
+ {
+ "epoch": 21.39126637554585,
+ "grad_norm": 0.01572156324982643,
+ "learning_rate": 0.0006,
+ "loss": 4.475979804992676,
+ "step": 1540
+ },
+ {
+ "epoch": 21.40524017467249,
+ "grad_norm": 0.017156066372990608,
+ "learning_rate": 0.0006,
+ "loss": 4.4011125564575195,
+ "step": 1541
+ },
+ {
+ "epoch": 21.419213973799128,
+ "grad_norm": 0.01788829080760479,
+ "learning_rate": 0.0006,
+ "loss": 4.4174089431762695,
+ "step": 1542
+ },
+ {
+ "epoch": 21.433187772925763,
+ "grad_norm": 0.01728985086083412,
+ "learning_rate": 0.0006,
+ "loss": 4.429244041442871,
+ "step": 1543
+ },
+ {
+ "epoch": 21.4471615720524,
+ "grad_norm": 0.016534466296434402,
+ "learning_rate": 0.0006,
+ "loss": 4.326068878173828,
+ "step": 1544
+ },
+ {
+ "epoch": 21.46113537117904,
+ "grad_norm": 0.016582584008574486,
+ "learning_rate": 0.0006,
+ "loss": 4.2975969314575195,
+ "step": 1545
+ },
+ {
+ "epoch": 21.475109170305675,
+ "grad_norm": 0.018034178763628006,
+ "learning_rate": 0.0006,
+ "loss": 4.501319408416748,
+ "step": 1546
+ },
+ {
+ "epoch": 21.489082969432314,
+ "grad_norm": 0.017827702686190605,
+ "learning_rate": 0.0006,
+ "loss": 4.502284526824951,
+ "step": 1547
+ },
+ {
+ "epoch": 21.503056768558952,
+ "grad_norm": 0.017182299867272377,
+ "learning_rate": 0.0006,
+ "loss": 4.380828857421875,
+ "step": 1548
+ },
+ {
+ "epoch": 21.51703056768559,
+ "grad_norm": 0.019101588055491447,
+ "learning_rate": 0.0006,
+ "loss": 4.31564998626709,
+ "step": 1549
+ },
+ {
+ "epoch": 21.531004366812226,
+ "grad_norm": 0.0204677302390337,
+ "learning_rate": 0.0006,
+ "loss": 4.3840131759643555,
+ "step": 1550
+ },
+ {
+ "epoch": 21.544978165938865,
+ "grad_norm": 0.0188993439078331,
+ "learning_rate": 0.0006,
+ "loss": 4.4257731437683105,
+ "step": 1551
+ },
+ {
+ "epoch": 21.558951965065503,
+ "grad_norm": 0.01730230636894703,
+ "learning_rate": 0.0006,
+ "loss": 4.445132255554199,
+ "step": 1552
+ },
+ {
+ "epoch": 21.57292576419214,
+ "grad_norm": 0.01801574043929577,
+ "learning_rate": 0.0006,
+ "loss": 4.427064895629883,
+ "step": 1553
+ },
+ {
+ "epoch": 21.586899563318777,
+ "grad_norm": 0.018417565152049065,
+ "learning_rate": 0.0006,
+ "loss": 4.319595813751221,
+ "step": 1554
+ },
+ {
+ "epoch": 21.600873362445416,
+ "grad_norm": 0.01683160290122032,
+ "learning_rate": 0.0006,
+ "loss": 4.349218368530273,
+ "step": 1555
+ },
+ {
+ "epoch": 21.61484716157205,
+ "grad_norm": 0.016973106190562248,
+ "learning_rate": 0.0006,
+ "loss": 4.375990390777588,
+ "step": 1556
+ },
+ {
+ "epoch": 21.62882096069869,
+ "grad_norm": 0.015688583254814148,
+ "learning_rate": 0.0006,
+ "loss": 4.408170223236084,
+ "step": 1557
+ },
+ {
+ "epoch": 21.64279475982533,
+ "grad_norm": 0.01573043316602707,
+ "learning_rate": 0.0006,
+ "loss": 4.416255950927734,
+ "step": 1558
+ },
+ {
+ "epoch": 21.656768558951963,
+ "grad_norm": 0.01744293048977852,
+ "learning_rate": 0.0006,
+ "loss": 4.477141380310059,
+ "step": 1559
+ },
+ {
+ "epoch": 21.670742358078602,
+ "grad_norm": 0.017503513023257256,
+ "learning_rate": 0.0006,
+ "loss": 4.3367486000061035,
+ "step": 1560
+ },
+ {
+ "epoch": 21.68471615720524,
+ "grad_norm": 0.016592150554060936,
+ "learning_rate": 0.0006,
+ "loss": 4.314785480499268,
+ "step": 1561
+ },
+ {
+ "epoch": 21.69868995633188,
+ "grad_norm": 0.015774084255099297,
+ "learning_rate": 0.0006,
+ "loss": 4.448209285736084,
+ "step": 1562
+ },
+ {
+ "epoch": 21.712663755458514,
+ "grad_norm": 0.015779150649905205,
+ "learning_rate": 0.0006,
+ "loss": 4.392690658569336,
+ "step": 1563
+ },
+ {
+ "epoch": 21.726637554585153,
+ "grad_norm": 0.015392648056149483,
+ "learning_rate": 0.0006,
+ "loss": 4.437741756439209,
+ "step": 1564
+ },
+ {
+ "epoch": 21.74061135371179,
+ "grad_norm": 0.01530836895108223,
+ "learning_rate": 0.0006,
+ "loss": 4.424595832824707,
+ "step": 1565
+ },
+ {
+ "epoch": 21.754585152838427,
+ "grad_norm": 0.015600779093801975,
+ "learning_rate": 0.0006,
+ "loss": 4.335417747497559,
+ "step": 1566
+ },
+ {
+ "epoch": 21.768558951965066,
+ "grad_norm": 0.014646518044173717,
+ "learning_rate": 0.0006,
+ "loss": 4.279081344604492,
+ "step": 1567
+ },
+ {
+ "epoch": 21.782532751091704,
+ "grad_norm": 0.014132864773273468,
+ "learning_rate": 0.0006,
+ "loss": 4.345950126647949,
+ "step": 1568
+ },
+ {
+ "epoch": 21.79650655021834,
+ "grad_norm": 0.01425724383443594,
+ "learning_rate": 0.0006,
+ "loss": 4.277588844299316,
+ "step": 1569
+ },
+ {
+ "epoch": 21.810480349344978,
+ "grad_norm": 0.015396572649478912,
+ "learning_rate": 0.0006,
+ "loss": 4.326196670532227,
+ "step": 1570
+ },
+ {
+ "epoch": 21.824454148471617,
+ "grad_norm": 0.01683669537305832,
+ "learning_rate": 0.0006,
+ "loss": 4.3294172286987305,
+ "step": 1571
+ },
+ {
+ "epoch": 21.83842794759825,
+ "grad_norm": 0.016365647315979004,
+ "learning_rate": 0.0006,
+ "loss": 4.523321628570557,
+ "step": 1572
+ },
+ {
+ "epoch": 21.85240174672489,
+ "grad_norm": 0.017142774537205696,
+ "learning_rate": 0.0006,
+ "loss": 4.475842475891113,
+ "step": 1573
+ },
+ {
+ "epoch": 21.86637554585153,
+ "grad_norm": 0.015950839966535568,
+ "learning_rate": 0.0006,
+ "loss": 4.397187232971191,
+ "step": 1574
+ },
+ {
+ "epoch": 21.880349344978168,
+ "grad_norm": 0.016180362552404404,
+ "learning_rate": 0.0006,
+ "loss": 4.414778709411621,
+ "step": 1575
+ },
+ {
+ "epoch": 21.894323144104803,
+ "grad_norm": 0.017615556716918945,
+ "learning_rate": 0.0006,
+ "loss": 4.348567962646484,
+ "step": 1576
+ },
+ {
+ "epoch": 21.90829694323144,
+ "grad_norm": 0.01896277628839016,
+ "learning_rate": 0.0006,
+ "loss": 4.431762218475342,
+ "step": 1577
+ },
+ {
+ "epoch": 21.92227074235808,
+ "grad_norm": 0.01990971714258194,
+ "learning_rate": 0.0006,
+ "loss": 4.431285858154297,
+ "step": 1578
+ },
+ {
+ "epoch": 21.936244541484715,
+ "grad_norm": 0.0201258547604084,
+ "learning_rate": 0.0006,
+ "loss": 4.309872627258301,
+ "step": 1579
+ },
+ {
+ "epoch": 21.950218340611354,
+ "grad_norm": 0.019620198756456375,
+ "learning_rate": 0.0006,
+ "loss": 4.484044551849365,
+ "step": 1580
+ },
+ {
+ "epoch": 21.964192139737992,
+ "grad_norm": 0.01771123707294464,
+ "learning_rate": 0.0006,
+ "loss": 4.368093013763428,
+ "step": 1581
+ },
+ {
+ "epoch": 21.978165938864628,
+ "grad_norm": 0.017501654103398323,
+ "learning_rate": 0.0006,
+ "loss": 4.372651100158691,
+ "step": 1582
+ },
+ {
+ "epoch": 21.992139737991266,
+ "grad_norm": 0.01693608984351158,
+ "learning_rate": 0.0006,
+ "loss": 4.379167556762695,
+ "step": 1583
+ },
+ {
+ "epoch": 22.0,
+ "grad_norm": 0.02002849616110325,
+ "learning_rate": 0.0006,
+ "loss": 4.329627990722656,
+ "step": 1584
+ },
+ {
+ "epoch": 22.0,
+ "eval_loss": 4.723696231842041,
+ "eval_runtime": 56.1758,
+ "eval_samples_per_second": 43.471,
+ "eval_steps_per_second": 1.371,
+ "step": 1584
+ },
+ {
+ "epoch": 22.01397379912664,
+ "grad_norm": 0.02030305378139019,
+ "learning_rate": 0.0006,
+ "loss": 4.355566024780273,
+ "step": 1585
+ },
+ {
+ "epoch": 22.027947598253274,
+ "grad_norm": 0.018638934940099716,
+ "learning_rate": 0.0006,
+ "loss": 4.314009189605713,
+ "step": 1586
+ },
+ {
+ "epoch": 22.041921397379912,
+ "grad_norm": 0.019917072728276253,
+ "learning_rate": 0.0006,
+ "loss": 4.367525577545166,
+ "step": 1587
+ },
+ {
+ "epoch": 22.05589519650655,
+ "grad_norm": 0.020785167813301086,
+ "learning_rate": 0.0006,
+ "loss": 4.375561714172363,
+ "step": 1588
+ },
+ {
+ "epoch": 22.069868995633186,
+ "grad_norm": 0.021214453503489494,
+ "learning_rate": 0.0006,
+ "loss": 4.418305397033691,
+ "step": 1589
+ },
+ {
+ "epoch": 22.083842794759825,
+ "grad_norm": 0.019051678478717804,
+ "learning_rate": 0.0006,
+ "loss": 4.215126991271973,
+ "step": 1590
+ },
+ {
+ "epoch": 22.097816593886463,
+ "grad_norm": 0.02323366142809391,
+ "learning_rate": 0.0006,
+ "loss": 4.306807994842529,
+ "step": 1591
+ },
+ {
+ "epoch": 22.111790393013102,
+ "grad_norm": 0.026186738163232803,
+ "learning_rate": 0.0006,
+ "loss": 4.408508777618408,
+ "step": 1592
+ },
+ {
+ "epoch": 22.125764192139737,
+ "grad_norm": 0.027361227199435234,
+ "learning_rate": 0.0006,
+ "loss": 4.329322814941406,
+ "step": 1593
+ },
+ {
+ "epoch": 22.139737991266376,
+ "grad_norm": 0.026564646512269974,
+ "learning_rate": 0.0006,
+ "loss": 4.346642017364502,
+ "step": 1594
+ },
+ {
+ "epoch": 22.153711790393015,
+ "grad_norm": 0.025073667988181114,
+ "learning_rate": 0.0006,
+ "loss": 4.3984198570251465,
+ "step": 1595
+ },
+ {
+ "epoch": 22.16768558951965,
+ "grad_norm": 0.02347794733941555,
+ "learning_rate": 0.0006,
+ "loss": 4.352538108825684,
+ "step": 1596
+ },
+ {
+ "epoch": 22.18165938864629,
+ "grad_norm": 0.02203468047082424,
+ "learning_rate": 0.0006,
+ "loss": 4.298148155212402,
+ "step": 1597
+ },
+ {
+ "epoch": 22.195633187772927,
+ "grad_norm": 0.020621957257390022,
+ "learning_rate": 0.0006,
+ "loss": 4.480309963226318,
+ "step": 1598
+ },
+ {
+ "epoch": 22.209606986899562,
+ "grad_norm": 0.023710055276751518,
+ "learning_rate": 0.0006,
+ "loss": 4.339317321777344,
+ "step": 1599
+ },
+ {
+ "epoch": 22.2235807860262,
+ "grad_norm": 0.0233482476323843,
+ "learning_rate": 0.0006,
+ "loss": 4.436896324157715,
+ "step": 1600
+ },
+ {
+ "epoch": 22.23755458515284,
+ "grad_norm": 0.020122403278946877,
+ "learning_rate": 0.0006,
+ "loss": 4.236002445220947,
+ "step": 1601
+ },
+ {
+ "epoch": 22.251528384279474,
+ "grad_norm": 0.01957266591489315,
+ "learning_rate": 0.0006,
+ "loss": 4.497328281402588,
+ "step": 1602
+ },
+ {
+ "epoch": 22.265502183406113,
+ "grad_norm": 0.018964895978569984,
+ "learning_rate": 0.0006,
+ "loss": 4.36360502243042,
+ "step": 1603
+ },
+ {
+ "epoch": 22.27947598253275,
+ "grad_norm": 0.018902769312262535,
+ "learning_rate": 0.0006,
+ "loss": 4.427093505859375,
+ "step": 1604
+ },
+ {
+ "epoch": 22.29344978165939,
+ "grad_norm": 0.018918119370937347,
+ "learning_rate": 0.0006,
+ "loss": 4.294950485229492,
+ "step": 1605
+ },
+ {
+ "epoch": 22.307423580786025,
+ "grad_norm": 0.019315775483846664,
+ "learning_rate": 0.0006,
+ "loss": 4.427952289581299,
+ "step": 1606
+ },
+ {
+ "epoch": 22.321397379912664,
+ "grad_norm": 0.018350522965192795,
+ "learning_rate": 0.0006,
+ "loss": 4.272945404052734,
+ "step": 1607
+ },
+ {
+ "epoch": 22.335371179039303,
+ "grad_norm": 0.01759534701704979,
+ "learning_rate": 0.0006,
+ "loss": 4.348023891448975,
+ "step": 1608
+ },
+ {
+ "epoch": 22.349344978165938,
+ "grad_norm": 0.01646057888865471,
+ "learning_rate": 0.0006,
+ "loss": 4.278518199920654,
+ "step": 1609
+ },
+ {
+ "epoch": 22.363318777292577,
+ "grad_norm": 0.015875060111284256,
+ "learning_rate": 0.0006,
+ "loss": 4.423527240753174,
+ "step": 1610
+ },
+ {
+ "epoch": 22.377292576419215,
+ "grad_norm": 0.01638529635965824,
+ "learning_rate": 0.0006,
+ "loss": 4.321723937988281,
+ "step": 1611
+ },
+ {
+ "epoch": 22.39126637554585,
+ "grad_norm": 0.016942061483860016,
+ "learning_rate": 0.0006,
+ "loss": 4.346164703369141,
+ "step": 1612
+ },
+ {
+ "epoch": 22.40524017467249,
+ "grad_norm": 0.015209621749818325,
+ "learning_rate": 0.0006,
+ "loss": 4.434340000152588,
+ "step": 1613
+ },
+ {
+ "epoch": 22.419213973799128,
+ "grad_norm": 0.014423009008169174,
+ "learning_rate": 0.0006,
+ "loss": 4.342560768127441,
+ "step": 1614
+ },
+ {
+ "epoch": 22.433187772925763,
+ "grad_norm": 0.014369525946676731,
+ "learning_rate": 0.0006,
+ "loss": 4.296586990356445,
+ "step": 1615
+ },
+ {
+ "epoch": 22.4471615720524,
+ "grad_norm": 0.01392540242522955,
+ "learning_rate": 0.0006,
+ "loss": 4.433176040649414,
+ "step": 1616
+ },
+ {
+ "epoch": 22.46113537117904,
+ "grad_norm": 0.013206899166107178,
+ "learning_rate": 0.0006,
+ "loss": 4.32962703704834,
+ "step": 1617
+ },
+ {
+ "epoch": 22.475109170305675,
+ "grad_norm": 0.013427077792584896,
+ "learning_rate": 0.0006,
+ "loss": 4.420437335968018,
+ "step": 1618
+ },
+ {
+ "epoch": 22.489082969432314,
+ "grad_norm": 0.014194196090102196,
+ "learning_rate": 0.0006,
+ "loss": 4.410478591918945,
+ "step": 1619
+ },
+ {
+ "epoch": 22.503056768558952,
+ "grad_norm": 0.01362364087253809,
+ "learning_rate": 0.0006,
+ "loss": 4.338932037353516,
+ "step": 1620
+ },
+ {
+ "epoch": 22.51703056768559,
+ "grad_norm": 0.014299589209258556,
+ "learning_rate": 0.0006,
+ "loss": 4.34089994430542,
+ "step": 1621
+ },
+ {
+ "epoch": 22.531004366812226,
+ "grad_norm": 0.014061295427381992,
+ "learning_rate": 0.0006,
+ "loss": 4.29201602935791,
+ "step": 1622
+ },
+ {
+ "epoch": 22.544978165938865,
+ "grad_norm": 0.013945229351520538,
+ "learning_rate": 0.0006,
+ "loss": 4.326052188873291,
+ "step": 1623
+ },
+ {
+ "epoch": 22.558951965065503,
+ "grad_norm": 0.01387068536132574,
+ "learning_rate": 0.0006,
+ "loss": 4.393521308898926,
+ "step": 1624
+ },
+ {
+ "epoch": 22.57292576419214,
+ "grad_norm": 0.013931898400187492,
+ "learning_rate": 0.0006,
+ "loss": 4.3739752769470215,
+ "step": 1625
+ },
+ {
+ "epoch": 22.586899563318777,
+ "grad_norm": 0.01497740764170885,
+ "learning_rate": 0.0006,
+ "loss": 4.3991618156433105,
+ "step": 1626
+ },
+ {
+ "epoch": 22.600873362445416,
+ "grad_norm": 0.015842905268073082,
+ "learning_rate": 0.0006,
+ "loss": 4.405340671539307,
+ "step": 1627
+ },
+ {
+ "epoch": 22.61484716157205,
+ "grad_norm": 0.01580633595585823,
+ "learning_rate": 0.0006,
+ "loss": 4.280102729797363,
+ "step": 1628
+ },
+ {
+ "epoch": 22.62882096069869,
+ "grad_norm": 0.016567382961511612,
+ "learning_rate": 0.0006,
+ "loss": 4.407657146453857,
+ "step": 1629
+ },
+ {
+ "epoch": 22.64279475982533,
+ "grad_norm": 0.01648622378706932,
+ "learning_rate": 0.0006,
+ "loss": 4.392590522766113,
+ "step": 1630
+ },
+ {
+ "epoch": 22.656768558951963,
+ "grad_norm": 0.017066307365894318,
+ "learning_rate": 0.0006,
+ "loss": 4.324862003326416,
+ "step": 1631
+ },
+ {
+ "epoch": 22.670742358078602,
+ "grad_norm": 0.01637062057852745,
+ "learning_rate": 0.0006,
+ "loss": 4.34598445892334,
+ "step": 1632
+ },
+ {
+ "epoch": 22.68471615720524,
+ "grad_norm": 0.018038960173726082,
+ "learning_rate": 0.0006,
+ "loss": 4.320711612701416,
+ "step": 1633
+ },
+ {
+ "epoch": 22.69868995633188,
+ "grad_norm": 0.018096428364515305,
+ "learning_rate": 0.0006,
+ "loss": 4.536527156829834,
+ "step": 1634
+ },
+ {
+ "epoch": 22.712663755458514,
+ "grad_norm": 0.017889291048049927,
+ "learning_rate": 0.0006,
+ "loss": 4.388131141662598,
+ "step": 1635
+ },
+ {
+ "epoch": 22.726637554585153,
+ "grad_norm": 0.018771521747112274,
+ "learning_rate": 0.0006,
+ "loss": 4.357668876647949,
+ "step": 1636
+ },
+ {
+ "epoch": 22.74061135371179,
+ "grad_norm": 0.02100873365998268,
+ "learning_rate": 0.0006,
+ "loss": 4.293972015380859,
+ "step": 1637
+ },
+ {
+ "epoch": 22.754585152838427,
+ "grad_norm": 0.017715785652399063,
+ "learning_rate": 0.0006,
+ "loss": 4.349849224090576,
+ "step": 1638
+ },
+ {
+ "epoch": 22.768558951965066,
+ "grad_norm": 0.015546333976089954,
+ "learning_rate": 0.0006,
+ "loss": 4.311728000640869,
+ "step": 1639
+ },
+ {
+ "epoch": 22.782532751091704,
+ "grad_norm": 0.01605990156531334,
+ "learning_rate": 0.0006,
+ "loss": 4.499123573303223,
+ "step": 1640
+ },
+ {
+ "epoch": 22.79650655021834,
+ "grad_norm": 0.016642142087221146,
+ "learning_rate": 0.0006,
+ "loss": 4.4319257736206055,
+ "step": 1641
+ },
+ {
+ "epoch": 22.810480349344978,
+ "grad_norm": 0.016120387241244316,
+ "learning_rate": 0.0006,
+ "loss": 4.4373650550842285,
+ "step": 1642
+ },
+ {
+ "epoch": 22.824454148471617,
+ "grad_norm": 0.015174890868365765,
+ "learning_rate": 0.0006,
+ "loss": 4.314082145690918,
+ "step": 1643
+ },
+ {
+ "epoch": 22.83842794759825,
+ "grad_norm": 0.01551714539527893,
+ "learning_rate": 0.0006,
+ "loss": 4.360363960266113,
+ "step": 1644
+ },
+ {
+ "epoch": 22.85240174672489,
+ "grad_norm": 0.014966742135584354,
+ "learning_rate": 0.0006,
+ "loss": 4.346253395080566,
+ "step": 1645
+ },
+ {
+ "epoch": 22.86637554585153,
+ "grad_norm": 0.014288028702139854,
+ "learning_rate": 0.0006,
+ "loss": 4.464366436004639,
+ "step": 1646
+ },
+ {
+ "epoch": 22.880349344978168,
+ "grad_norm": 0.015291682444512844,
+ "learning_rate": 0.0006,
+ "loss": 4.504101753234863,
+ "step": 1647
+ },
+ {
+ "epoch": 22.894323144104803,
+ "grad_norm": 0.014534426853060722,
+ "learning_rate": 0.0006,
+ "loss": 4.462172031402588,
+ "step": 1648
+ },
+ {
+ "epoch": 22.90829694323144,
+ "grad_norm": 0.014129819348454475,
+ "learning_rate": 0.0006,
+ "loss": 4.350126266479492,
+ "step": 1649
+ },
+ {
+ "epoch": 22.92227074235808,
+ "grad_norm": 0.014932668767869473,
+ "learning_rate": 0.0006,
+ "loss": 4.374392032623291,
+ "step": 1650
+ },
+ {
+ "epoch": 22.936244541484715,
+ "grad_norm": 0.014647853560745716,
+ "learning_rate": 0.0006,
+ "loss": 4.252264499664307,
+ "step": 1651
+ },
+ {
+ "epoch": 22.950218340611354,
+ "grad_norm": 0.013553360477089882,
+ "learning_rate": 0.0006,
+ "loss": 4.379729270935059,
+ "step": 1652
+ },
+ {
+ "epoch": 22.964192139737992,
+ "grad_norm": 0.014606194570660591,
+ "learning_rate": 0.0006,
+ "loss": 4.363372802734375,
+ "step": 1653
+ },
+ {
+ "epoch": 22.978165938864628,
+ "grad_norm": 0.015000631101429462,
+ "learning_rate": 0.0006,
+ "loss": 4.408048629760742,
+ "step": 1654
+ },
+ {
+ "epoch": 22.992139737991266,
+ "grad_norm": 0.015370816923677921,
+ "learning_rate": 0.0006,
+ "loss": 4.297451019287109,
+ "step": 1655
+ },
+ {
+ "epoch": 23.0,
+ "grad_norm": 0.01709144562482834,
+ "learning_rate": 0.0006,
+ "loss": 4.294776916503906,
+ "step": 1656
+ },
+ {
+ "epoch": 23.0,
+ "eval_loss": 4.678618907928467,
+ "eval_runtime": 56.1668,
+ "eval_samples_per_second": 43.478,
+ "eval_steps_per_second": 1.371,
+ "step": 1656
+ },
+ {
+ "epoch": 23.01397379912664,
+ "grad_norm": 0.01727290451526642,
+ "learning_rate": 0.0006,
+ "loss": 4.311701774597168,
+ "step": 1657
+ },
+ {
+ "epoch": 23.027947598253274,
+ "grad_norm": 0.020645759999752045,
+ "learning_rate": 0.0006,
+ "loss": 4.429897308349609,
+ "step": 1658
+ },
+ {
+ "epoch": 23.041921397379912,
+ "grad_norm": 0.02204226516187191,
+ "learning_rate": 0.0006,
+ "loss": 4.365151882171631,
+ "step": 1659
+ },
+ {
+ "epoch": 23.05589519650655,
+ "grad_norm": 0.023069269955158234,
+ "learning_rate": 0.0006,
+ "loss": 4.3710832595825195,
+ "step": 1660
+ },
+ {
+ "epoch": 23.069868995633186,
+ "grad_norm": 0.024151179939508438,
+ "learning_rate": 0.0006,
+ "loss": 4.403108596801758,
+ "step": 1661
+ },
+ {
+ "epoch": 23.083842794759825,
+ "grad_norm": 0.02511233650147915,
+ "learning_rate": 0.0006,
+ "loss": 4.414128303527832,
+ "step": 1662
+ },
+ {
+ "epoch": 23.097816593886463,
+ "grad_norm": 0.025802113115787506,
+ "learning_rate": 0.0006,
+ "loss": 4.388265609741211,
+ "step": 1663
+ },
+ {
+ "epoch": 23.111790393013102,
+ "grad_norm": 0.02377348765730858,
+ "learning_rate": 0.0006,
+ "loss": 4.357505798339844,
+ "step": 1664
+ },
+ {
+ "epoch": 23.125764192139737,
+ "grad_norm": 0.020876651629805565,
+ "learning_rate": 0.0006,
+ "loss": 4.23043966293335,
+ "step": 1665
+ },
+ {
+ "epoch": 23.139737991266376,
+ "grad_norm": 0.01959727331995964,
+ "learning_rate": 0.0006,
+ "loss": 4.280970096588135,
+ "step": 1666
+ },
+ {
+ "epoch": 23.153711790393015,
+ "grad_norm": 0.018678616732358932,
+ "learning_rate": 0.0006,
+ "loss": 4.346716403961182,
+ "step": 1667
+ },
+ {
+ "epoch": 23.16768558951965,
+ "grad_norm": 0.02108944021165371,
+ "learning_rate": 0.0006,
+ "loss": 4.277904510498047,
+ "step": 1668
+ },
+ {
+ "epoch": 23.18165938864629,
+ "grad_norm": 0.020221566781401634,
+ "learning_rate": 0.0006,
+ "loss": 4.367693901062012,
+ "step": 1669
+ },
+ {
+ "epoch": 23.195633187772927,
+ "grad_norm": 0.019599711522459984,
+ "learning_rate": 0.0006,
+ "loss": 4.257092475891113,
+ "step": 1670
+ },
+ {
+ "epoch": 23.209606986899562,
+ "grad_norm": 0.019833361729979515,
+ "learning_rate": 0.0006,
+ "loss": 4.540737628936768,
+ "step": 1671
+ },
+ {
+ "epoch": 23.2235807860262,
+ "grad_norm": 0.020352095365524292,
+ "learning_rate": 0.0006,
+ "loss": 4.360923767089844,
+ "step": 1672
+ },
+ {
+ "epoch": 23.23755458515284,
+ "grad_norm": 0.01939748041331768,
+ "learning_rate": 0.0006,
+ "loss": 4.393110275268555,
+ "step": 1673
+ },
+ {
+ "epoch": 23.251528384279474,
+ "grad_norm": 0.01884354278445244,
+ "learning_rate": 0.0006,
+ "loss": 4.383938789367676,
+ "step": 1674
+ },
+ {
+ "epoch": 23.265502183406113,
+ "grad_norm": 0.02008470520377159,
+ "learning_rate": 0.0006,
+ "loss": 4.346344947814941,
+ "step": 1675
+ },
+ {
+ "epoch": 23.27947598253275,
+ "grad_norm": 0.019065195694565773,
+ "learning_rate": 0.0006,
+ "loss": 4.36611795425415,
+ "step": 1676
+ },
+ {
+ "epoch": 23.29344978165939,
+ "grad_norm": 0.01938541606068611,
+ "learning_rate": 0.0006,
+ "loss": 4.4009599685668945,
+ "step": 1677
+ },
+ {
+ "epoch": 23.307423580786025,
+ "grad_norm": 0.019044723361730576,
+ "learning_rate": 0.0006,
+ "loss": 4.340169906616211,
+ "step": 1678
+ },
+ {
+ "epoch": 23.321397379912664,
+ "grad_norm": 0.018850378692150116,
+ "learning_rate": 0.0006,
+ "loss": 4.37674617767334,
+ "step": 1679
+ },
+ {
+ "epoch": 23.335371179039303,
+ "grad_norm": 0.019076095893979073,
+ "learning_rate": 0.0006,
+ "loss": 4.226874351501465,
+ "step": 1680
+ },
+ {
+ "epoch": 23.349344978165938,
+ "grad_norm": 0.017543038353323936,
+ "learning_rate": 0.0006,
+ "loss": 4.337751388549805,
+ "step": 1681
+ },
+ {
+ "epoch": 23.363318777292577,
+ "grad_norm": 0.017170218750834465,
+ "learning_rate": 0.0006,
+ "loss": 4.282512664794922,
+ "step": 1682
+ },
+ {
+ "epoch": 23.377292576419215,
+ "grad_norm": 0.01574796624481678,
+ "learning_rate": 0.0006,
+ "loss": 4.30638313293457,
+ "step": 1683
+ },
+ {
+ "epoch": 23.39126637554585,
+ "grad_norm": 0.015160846523940563,
+ "learning_rate": 0.0006,
+ "loss": 4.336716175079346,
+ "step": 1684
+ },
+ {
+ "epoch": 23.40524017467249,
+ "grad_norm": 0.01534526702016592,
+ "learning_rate": 0.0006,
+ "loss": 4.2174787521362305,
+ "step": 1685
+ },
+ {
+ "epoch": 23.419213973799128,
+ "grad_norm": 0.015289999544620514,
+ "learning_rate": 0.0006,
+ "loss": 4.272546291351318,
+ "step": 1686
+ },
+ {
+ "epoch": 23.433187772925763,
+ "grad_norm": 0.014894185587763786,
+ "learning_rate": 0.0006,
+ "loss": 4.296045303344727,
+ "step": 1687
+ },
+ {
+ "epoch": 23.4471615720524,
+ "grad_norm": 0.014869502745568752,
+ "learning_rate": 0.0006,
+ "loss": 4.425461769104004,
+ "step": 1688
+ },
+ {
+ "epoch": 23.46113537117904,
+ "grad_norm": 0.01474483497440815,
+ "learning_rate": 0.0006,
+ "loss": 4.361867427825928,
+ "step": 1689
+ },
+ {
+ "epoch": 23.475109170305675,
+ "grad_norm": 0.015299245715141296,
+ "learning_rate": 0.0006,
+ "loss": 4.364523410797119,
+ "step": 1690
+ },
+ {
+ "epoch": 23.489082969432314,
+ "grad_norm": 0.015201340429484844,
+ "learning_rate": 0.0006,
+ "loss": 4.285182952880859,
+ "step": 1691
+ },
+ {
+ "epoch": 23.503056768558952,
+ "grad_norm": 0.016057293862104416,
+ "learning_rate": 0.0006,
+ "loss": 4.284873962402344,
+ "step": 1692
+ },
+ {
+ "epoch": 23.51703056768559,
+ "grad_norm": 0.015871062874794006,
+ "learning_rate": 0.0006,
+ "loss": 4.326470851898193,
+ "step": 1693
+ },
+ {
+ "epoch": 23.531004366812226,
+ "grad_norm": 0.01586288772523403,
+ "learning_rate": 0.0006,
+ "loss": 4.324883460998535,
+ "step": 1694
+ },
+ {
+ "epoch": 23.544978165938865,
+ "grad_norm": 0.016092827543616295,
+ "learning_rate": 0.0006,
+ "loss": 4.3258867263793945,
+ "step": 1695
+ },
+ {
+ "epoch": 23.558951965065503,
+ "grad_norm": 0.01582016795873642,
+ "learning_rate": 0.0006,
+ "loss": 4.358892440795898,
+ "step": 1696
+ },
+ {
+ "epoch": 23.57292576419214,
+ "grad_norm": 0.01688452623784542,
+ "learning_rate": 0.0006,
+ "loss": 4.280842304229736,
+ "step": 1697
+ },
+ {
+ "epoch": 23.586899563318777,
+ "grad_norm": 0.015865886583924294,
+ "learning_rate": 0.0006,
+ "loss": 4.245567321777344,
+ "step": 1698
+ },
+ {
+ "epoch": 23.600873362445416,
+ "grad_norm": 0.015303662046790123,
+ "learning_rate": 0.0006,
+ "loss": 4.383112907409668,
+ "step": 1699
+ },
+ {
+ "epoch": 23.61484716157205,
+ "grad_norm": 0.015649516135454178,
+ "learning_rate": 0.0006,
+ "loss": 4.519082069396973,
+ "step": 1700
+ },
+ {
+ "epoch": 23.62882096069869,
+ "grad_norm": 0.0167120099067688,
+ "learning_rate": 0.0006,
+ "loss": 4.447847366333008,
+ "step": 1701
+ },
+ {
+ "epoch": 23.64279475982533,
+ "grad_norm": 0.016368716955184937,
+ "learning_rate": 0.0006,
+ "loss": 4.434186935424805,
+ "step": 1702
+ },
+ {
+ "epoch": 23.656768558951963,
+ "grad_norm": 0.017379503697156906,
+ "learning_rate": 0.0006,
+ "loss": 4.317594528198242,
+ "step": 1703
+ },
+ {
+ "epoch": 23.670742358078602,
+ "grad_norm": 0.018573811277747154,
+ "learning_rate": 0.0006,
+ "loss": 4.205793857574463,
+ "step": 1704
+ },
+ {
+ "epoch": 23.68471615720524,
+ "grad_norm": 0.01749090477824211,
+ "learning_rate": 0.0006,
+ "loss": 4.3985700607299805,
+ "step": 1705
+ },
+ {
+ "epoch": 23.69868995633188,
+ "grad_norm": 0.0165668074041605,
+ "learning_rate": 0.0006,
+ "loss": 4.329226970672607,
+ "step": 1706
+ },
+ {
+ "epoch": 23.712663755458514,
+ "grad_norm": 0.016913846135139465,
+ "learning_rate": 0.0006,
+ "loss": 4.388485908508301,
+ "step": 1707
+ },
+ {
+ "epoch": 23.726637554585153,
+ "grad_norm": 0.016813859343528748,
+ "learning_rate": 0.0006,
+ "loss": 4.3395915031433105,
+ "step": 1708
+ },
+ {
+ "epoch": 23.74061135371179,
+ "grad_norm": 0.01676975190639496,
+ "learning_rate": 0.0006,
+ "loss": 4.254745960235596,
+ "step": 1709
+ },
+ {
+ "epoch": 23.754585152838427,
+ "grad_norm": 0.016370350494980812,
+ "learning_rate": 0.0006,
+ "loss": 4.2612223625183105,
+ "step": 1710
+ },
+ {
+ "epoch": 23.768558951965066,
+ "grad_norm": 0.01696198247373104,
+ "learning_rate": 0.0006,
+ "loss": 4.327524662017822,
+ "step": 1711
+ },
+ {
+ "epoch": 23.782532751091704,
+ "grad_norm": 0.0189108457416296,
+ "learning_rate": 0.0006,
+ "loss": 4.325839519500732,
+ "step": 1712
+ },
+ {
+ "epoch": 23.79650655021834,
+ "grad_norm": 0.01862351782619953,
+ "learning_rate": 0.0006,
+ "loss": 4.333662509918213,
+ "step": 1713
+ },
+ {
+ "epoch": 23.810480349344978,
+ "grad_norm": 0.01655014418065548,
+ "learning_rate": 0.0006,
+ "loss": 4.39305305480957,
+ "step": 1714
+ },
+ {
+ "epoch": 23.824454148471617,
+ "grad_norm": 0.015899376943707466,
+ "learning_rate": 0.0006,
+ "loss": 4.390053749084473,
+ "step": 1715
+ },
+ {
+ "epoch": 23.83842794759825,
+ "grad_norm": 0.016519056633114815,
+ "learning_rate": 0.0006,
+ "loss": 4.255903244018555,
+ "step": 1716
+ },
+ {
+ "epoch": 23.85240174672489,
+ "grad_norm": 0.0167539119720459,
+ "learning_rate": 0.0006,
+ "loss": 4.255727767944336,
+ "step": 1717
+ },
+ {
+ "epoch": 23.86637554585153,
+ "grad_norm": 0.0171353816986084,
+ "learning_rate": 0.0006,
+ "loss": 4.376740455627441,
+ "step": 1718
+ },
+ {
+ "epoch": 23.880349344978168,
+ "grad_norm": 0.016719117760658264,
+ "learning_rate": 0.0006,
+ "loss": 4.3376688957214355,
+ "step": 1719
+ },
+ {
+ "epoch": 23.894323144104803,
+ "grad_norm": 0.01561494916677475,
+ "learning_rate": 0.0006,
+ "loss": 4.3345746994018555,
+ "step": 1720
+ },
+ {
+ "epoch": 23.90829694323144,
+ "grad_norm": 0.016303174197673798,
+ "learning_rate": 0.0006,
+ "loss": 4.332371234893799,
+ "step": 1721
+ },
+ {
+ "epoch": 23.92227074235808,
+ "grad_norm": 0.016722865402698517,
+ "learning_rate": 0.0006,
+ "loss": 4.267125129699707,
+ "step": 1722
+ },
+ {
+ "epoch": 23.936244541484715,
+ "grad_norm": 0.017072124406695366,
+ "learning_rate": 0.0006,
+ "loss": 4.322751998901367,
+ "step": 1723
+ },
+ {
+ "epoch": 23.950218340611354,
+ "grad_norm": 0.01699932850897312,
+ "learning_rate": 0.0006,
+ "loss": 4.378422737121582,
+ "step": 1724
+ },
+ {
+ "epoch": 23.964192139737992,
+ "grad_norm": 0.01742720976471901,
+ "learning_rate": 0.0006,
+ "loss": 4.345083236694336,
+ "step": 1725
+ },
+ {
+ "epoch": 23.978165938864628,
+ "grad_norm": 0.018797501921653748,
+ "learning_rate": 0.0006,
+ "loss": 4.290443420410156,
+ "step": 1726
+ },
+ {
+ "epoch": 23.992139737991266,
+ "grad_norm": 0.018090683966875076,
+ "learning_rate": 0.0006,
+ "loss": 4.377580642700195,
+ "step": 1727
+ },
+ {
+ "epoch": 24.0,
+ "grad_norm": 0.017305459827184677,
+ "learning_rate": 0.0006,
+ "loss": 4.364426612854004,
+ "step": 1728
+ },
+ {
+ "epoch": 24.0,
+ "eval_loss": 4.647762298583984,
+ "eval_runtime": 56.5402,
+ "eval_samples_per_second": 43.191,
+ "eval_steps_per_second": 1.362,
+ "step": 1728
+ },
+ {
+ "epoch": 24.01397379912664,
+ "grad_norm": 0.016493607312440872,
+ "learning_rate": 0.0006,
+ "loss": 4.252255916595459,
+ "step": 1729
+ },
+ {
+ "epoch": 24.027947598253274,
+ "grad_norm": 0.016529450193047523,
+ "learning_rate": 0.0006,
+ "loss": 4.352746963500977,
+ "step": 1730
+ },
+ {
+ "epoch": 24.041921397379912,
+ "grad_norm": 0.016940191388130188,
+ "learning_rate": 0.0006,
+ "loss": 4.301870822906494,
+ "step": 1731
+ },
+ {
+ "epoch": 24.05589519650655,
+ "grad_norm": 0.01752558909356594,
+ "learning_rate": 0.0006,
+ "loss": 4.2762451171875,
+ "step": 1732
+ },
+ {
+ "epoch": 24.069868995633186,
+ "grad_norm": 0.017112884670495987,
+ "learning_rate": 0.0006,
+ "loss": 4.2659149169921875,
+ "step": 1733
+ },
+ {
+ "epoch": 24.083842794759825,
+ "grad_norm": 0.01900586113333702,
+ "learning_rate": 0.0006,
+ "loss": 4.306058883666992,
+ "step": 1734
+ },
+ {
+ "epoch": 24.097816593886463,
+ "grad_norm": 0.019447680562734604,
+ "learning_rate": 0.0006,
+ "loss": 4.450128555297852,
+ "step": 1735
+ },
+ {
+ "epoch": 24.111790393013102,
+ "grad_norm": 0.01948186755180359,
+ "learning_rate": 0.0006,
+ "loss": 4.202869415283203,
+ "step": 1736
+ },
+ {
+ "epoch": 24.125764192139737,
+ "grad_norm": 0.01716383546590805,
+ "learning_rate": 0.0006,
+ "loss": 4.318876266479492,
+ "step": 1737
+ },
+ {
+ "epoch": 24.139737991266376,
+ "grad_norm": 0.017049988731741905,
+ "learning_rate": 0.0006,
+ "loss": 4.237238883972168,
+ "step": 1738
+ },
+ {
+ "epoch": 24.153711790393015,
+ "grad_norm": 0.015169923193752766,
+ "learning_rate": 0.0006,
+ "loss": 4.289426803588867,
+ "step": 1739
+ },
+ {
+ "epoch": 24.16768558951965,
+ "grad_norm": 0.014760667458176613,
+ "learning_rate": 0.0006,
+ "loss": 4.356924057006836,
+ "step": 1740
+ },
+ {
+ "epoch": 24.18165938864629,
+ "grad_norm": 0.015127750113606453,
+ "learning_rate": 0.0006,
+ "loss": 4.142916679382324,
+ "step": 1741
+ },
+ {
+ "epoch": 24.195633187772927,
+ "grad_norm": 0.01549555640667677,
+ "learning_rate": 0.0006,
+ "loss": 4.345563888549805,
+ "step": 1742
+ },
+ {
+ "epoch": 24.209606986899562,
+ "grad_norm": 0.016393091529607773,
+ "learning_rate": 0.0006,
+ "loss": 4.417495250701904,
+ "step": 1743
+ },
+ {
+ "epoch": 24.2235807860262,
+ "grad_norm": 0.01785266026854515,
+ "learning_rate": 0.0006,
+ "loss": 4.3553547859191895,
+ "step": 1744
+ },
+ {
+ "epoch": 24.23755458515284,
+ "grad_norm": 0.017574617639183998,
+ "learning_rate": 0.0006,
+ "loss": 4.241024494171143,
+ "step": 1745
+ },
+ {
+ "epoch": 24.251528384279474,
+ "grad_norm": 0.019060306251049042,
+ "learning_rate": 0.0006,
+ "loss": 4.3685150146484375,
+ "step": 1746
+ },
+ {
+ "epoch": 24.265502183406113,
+ "grad_norm": 0.0214633010327816,
+ "learning_rate": 0.0006,
+ "loss": 4.323866844177246,
+ "step": 1747
+ },
+ {
+ "epoch": 24.27947598253275,
+ "grad_norm": 0.021055003628134727,
+ "learning_rate": 0.0006,
+ "loss": 4.278757095336914,
+ "step": 1748
+ },
+ {
+ "epoch": 24.29344978165939,
+ "grad_norm": 0.018618572503328323,
+ "learning_rate": 0.0006,
+ "loss": 4.28849983215332,
+ "step": 1749
+ },
+ {
+ "epoch": 24.307423580786025,
+ "grad_norm": 0.018666435033082962,
+ "learning_rate": 0.0006,
+ "loss": 4.368590354919434,
+ "step": 1750
+ },
+ {
+ "epoch": 24.321397379912664,
+ "grad_norm": 0.020626146346330643,
+ "learning_rate": 0.0006,
+ "loss": 4.2951836585998535,
+ "step": 1751
+ },
+ {
+ "epoch": 24.335371179039303,
+ "grad_norm": 0.01929759792983532,
+ "learning_rate": 0.0006,
+ "loss": 4.3664703369140625,
+ "step": 1752
+ },
+ {
+ "epoch": 24.349344978165938,
+ "grad_norm": 0.020956002175807953,
+ "learning_rate": 0.0006,
+ "loss": 4.345308303833008,
+ "step": 1753
+ },
+ {
+ "epoch": 24.363318777292577,
+ "grad_norm": 0.019227957352995872,
+ "learning_rate": 0.0006,
+ "loss": 4.349849224090576,
+ "step": 1754
+ },
+ {
+ "epoch": 24.377292576419215,
+ "grad_norm": 0.015970097854733467,
+ "learning_rate": 0.0006,
+ "loss": 4.223012447357178,
+ "step": 1755
+ },
+ {
+ "epoch": 24.39126637554585,
+ "grad_norm": 0.015951646491885185,
+ "learning_rate": 0.0006,
+ "loss": 4.236420631408691,
+ "step": 1756
+ },
+ {
+ "epoch": 24.40524017467249,
+ "grad_norm": 0.016032515093684196,
+ "learning_rate": 0.0006,
+ "loss": 4.226899147033691,
+ "step": 1757
+ },
+ {
+ "epoch": 24.419213973799128,
+ "grad_norm": 0.01573132909834385,
+ "learning_rate": 0.0006,
+ "loss": 4.315916061401367,
+ "step": 1758
+ },
+ {
+ "epoch": 24.433187772925763,
+ "grad_norm": 0.015367005951702595,
+ "learning_rate": 0.0006,
+ "loss": 4.358939170837402,
+ "step": 1759
+ },
+ {
+ "epoch": 24.4471615720524,
+ "grad_norm": 0.01669907011091709,
+ "learning_rate": 0.0006,
+ "loss": 4.232147216796875,
+ "step": 1760
+ },
+ {
+ "epoch": 24.46113537117904,
+ "grad_norm": 0.018708152696490288,
+ "learning_rate": 0.0006,
+ "loss": 4.329852104187012,
+ "step": 1761
+ },
+ {
+ "epoch": 24.475109170305675,
+ "grad_norm": 0.019035987555980682,
+ "learning_rate": 0.0006,
+ "loss": 4.260340213775635,
+ "step": 1762
+ },
+ {
+ "epoch": 24.489082969432314,
+ "grad_norm": 0.01832144521176815,
+ "learning_rate": 0.0006,
+ "loss": 4.359543800354004,
+ "step": 1763
+ },
+ {
+ "epoch": 24.503056768558952,
+ "grad_norm": 0.015313294716179371,
+ "learning_rate": 0.0006,
+ "loss": 4.279088020324707,
+ "step": 1764
+ },
+ {
+ "epoch": 24.51703056768559,
+ "grad_norm": 0.015803923830389977,
+ "learning_rate": 0.0006,
+ "loss": 4.230682849884033,
+ "step": 1765
+ },
+ {
+ "epoch": 24.531004366812226,
+ "grad_norm": 0.016748374328017235,
+ "learning_rate": 0.0006,
+ "loss": 4.349661827087402,
+ "step": 1766
+ },
+ {
+ "epoch": 24.544978165938865,
+ "grad_norm": 0.015732428058981895,
+ "learning_rate": 0.0006,
+ "loss": 4.372441291809082,
+ "step": 1767
+ },
+ {
+ "epoch": 24.558951965065503,
+ "grad_norm": 0.016564829275012016,
+ "learning_rate": 0.0006,
+ "loss": 4.323090553283691,
+ "step": 1768
+ },
+ {
+ "epoch": 24.57292576419214,
+ "grad_norm": 0.01603010483086109,
+ "learning_rate": 0.0006,
+ "loss": 4.347898006439209,
+ "step": 1769
+ },
+ {
+ "epoch": 24.586899563318777,
+ "grad_norm": 0.01674809865653515,
+ "learning_rate": 0.0006,
+ "loss": 4.269165515899658,
+ "step": 1770
+ },
+ {
+ "epoch": 24.600873362445416,
+ "grad_norm": 0.017333930358290672,
+ "learning_rate": 0.0006,
+ "loss": 4.353298187255859,
+ "step": 1771
+ },
+ {
+ "epoch": 24.61484716157205,
+ "grad_norm": 0.017519580200314522,
+ "learning_rate": 0.0006,
+ "loss": 4.265199661254883,
+ "step": 1772
+ },
+ {
+ "epoch": 24.62882096069869,
+ "grad_norm": 0.017675306648015976,
+ "learning_rate": 0.0006,
+ "loss": 4.322596549987793,
+ "step": 1773
+ },
+ {
+ "epoch": 24.64279475982533,
+ "grad_norm": 0.017166534438729286,
+ "learning_rate": 0.0006,
+ "loss": 4.413900375366211,
+ "step": 1774
+ },
+ {
+ "epoch": 24.656768558951963,
+ "grad_norm": 0.016362346708774567,
+ "learning_rate": 0.0006,
+ "loss": 4.307236671447754,
+ "step": 1775
+ },
+ {
+ "epoch": 24.670742358078602,
+ "grad_norm": 0.016086872667074203,
+ "learning_rate": 0.0006,
+ "loss": 4.235316276550293,
+ "step": 1776
+ },
+ {
+ "epoch": 24.68471615720524,
+ "grad_norm": 0.01689537614583969,
+ "learning_rate": 0.0006,
+ "loss": 4.390591621398926,
+ "step": 1777
+ },
+ {
+ "epoch": 24.69868995633188,
+ "grad_norm": 0.016836725175380707,
+ "learning_rate": 0.0006,
+ "loss": 4.2990031242370605,
+ "step": 1778
+ },
+ {
+ "epoch": 24.712663755458514,
+ "grad_norm": 0.015627335757017136,
+ "learning_rate": 0.0006,
+ "loss": 4.166745185852051,
+ "step": 1779
+ },
+ {
+ "epoch": 24.726637554585153,
+ "grad_norm": 0.015007016249001026,
+ "learning_rate": 0.0006,
+ "loss": 4.295273303985596,
+ "step": 1780
+ },
+ {
+ "epoch": 24.74061135371179,
+ "grad_norm": 0.016084209084510803,
+ "learning_rate": 0.0006,
+ "loss": 4.342424392700195,
+ "step": 1781
+ },
+ {
+ "epoch": 24.754585152838427,
+ "grad_norm": 0.01595073565840721,
+ "learning_rate": 0.0006,
+ "loss": 4.3060102462768555,
+ "step": 1782
+ },
+ {
+ "epoch": 24.768558951965066,
+ "grad_norm": 0.016113461926579475,
+ "learning_rate": 0.0006,
+ "loss": 4.267908096313477,
+ "step": 1783
+ },
+ {
+ "epoch": 24.782532751091704,
+ "grad_norm": 0.01564556173980236,
+ "learning_rate": 0.0006,
+ "loss": 4.324642658233643,
+ "step": 1784
+ },
+ {
+ "epoch": 24.79650655021834,
+ "grad_norm": 0.01575268618762493,
+ "learning_rate": 0.0006,
+ "loss": 4.328181743621826,
+ "step": 1785
+ },
+ {
+ "epoch": 24.810480349344978,
+ "grad_norm": 0.016830824315547943,
+ "learning_rate": 0.0006,
+ "loss": 4.247906684875488,
+ "step": 1786
+ },
+ {
+ "epoch": 24.824454148471617,
+ "grad_norm": 0.017989547923207283,
+ "learning_rate": 0.0006,
+ "loss": 4.228281021118164,
+ "step": 1787
+ },
+ {
+ "epoch": 24.83842794759825,
+ "grad_norm": 0.016783086583018303,
+ "learning_rate": 0.0006,
+ "loss": 4.2880964279174805,
+ "step": 1788
+ },
+ {
+ "epoch": 24.85240174672489,
+ "grad_norm": 0.016467561945319176,
+ "learning_rate": 0.0006,
+ "loss": 4.36118745803833,
+ "step": 1789
+ },
+ {
+ "epoch": 24.86637554585153,
+ "grad_norm": 0.01780523732304573,
+ "learning_rate": 0.0006,
+ "loss": 4.172321319580078,
+ "step": 1790
+ },
+ {
+ "epoch": 24.880349344978168,
+ "grad_norm": 0.017102031037211418,
+ "learning_rate": 0.0006,
+ "loss": 4.352826118469238,
+ "step": 1791
+ },
+ {
+ "epoch": 24.894323144104803,
+ "grad_norm": 0.016409218311309814,
+ "learning_rate": 0.0006,
+ "loss": 4.212162017822266,
+ "step": 1792
+ },
+ {
+ "epoch": 24.90829694323144,
+ "grad_norm": 0.018660584464669228,
+ "learning_rate": 0.0006,
+ "loss": 4.346684455871582,
+ "step": 1793
+ },
+ {
+ "epoch": 24.92227074235808,
+ "grad_norm": 0.017389440909028053,
+ "learning_rate": 0.0006,
+ "loss": 4.197913646697998,
+ "step": 1794
+ },
+ {
+ "epoch": 24.936244541484715,
+ "grad_norm": 0.01596485823392868,
+ "learning_rate": 0.0006,
+ "loss": 4.370617866516113,
+ "step": 1795
+ },
+ {
+ "epoch": 24.950218340611354,
+ "grad_norm": 0.01709776371717453,
+ "learning_rate": 0.0006,
+ "loss": 4.313847064971924,
+ "step": 1796
+ },
+ {
+ "epoch": 24.964192139737992,
+ "grad_norm": 0.016571134328842163,
+ "learning_rate": 0.0006,
+ "loss": 4.1895856857299805,
+ "step": 1797
+ },
+ {
+ "epoch": 24.978165938864628,
+ "grad_norm": 0.016851048916578293,
+ "learning_rate": 0.0006,
+ "loss": 4.350772857666016,
+ "step": 1798
+ },
+ {
+ "epoch": 24.992139737991266,
+ "grad_norm": 0.017007440328598022,
+ "learning_rate": 0.0006,
+ "loss": 4.399008750915527,
+ "step": 1799
+ },
+ {
+ "epoch": 25.0,
+ "grad_norm": 0.01791212148964405,
+ "learning_rate": 0.0006,
+ "loss": 4.306121826171875,
+ "step": 1800
+ },
+ {
+ "epoch": 25.0,
+ "eval_loss": 4.636417865753174,
+ "eval_runtime": 56.8296,
+ "eval_samples_per_second": 42.971,
+ "eval_steps_per_second": 1.355,
+ "step": 1800
+ },
+ {
+ "epoch": 25.01397379912664,
+ "grad_norm": 0.01729617640376091,
+ "learning_rate": 0.0006,
+ "loss": 4.257018089294434,
+ "step": 1801
+ },
+ {
+ "epoch": 25.027947598253274,
+ "grad_norm": 0.017466643825173378,
+ "learning_rate": 0.0006,
+ "loss": 4.307605266571045,
+ "step": 1802
+ },
+ {
+ "epoch": 25.041921397379912,
+ "grad_norm": 0.018614279106259346,
+ "learning_rate": 0.0006,
+ "loss": 4.360141277313232,
+ "step": 1803
+ },
+ {
+ "epoch": 25.05589519650655,
+ "grad_norm": 0.02010771445930004,
+ "learning_rate": 0.0006,
+ "loss": 4.196428298950195,
+ "step": 1804
+ },
+ {
+ "epoch": 25.069868995633186,
+ "grad_norm": 0.02090228535234928,
+ "learning_rate": 0.0006,
+ "loss": 4.288947105407715,
+ "step": 1805
+ },
+ {
+ "epoch": 25.083842794759825,
+ "grad_norm": 0.02050255797803402,
+ "learning_rate": 0.0006,
+ "loss": 4.259582042694092,
+ "step": 1806
+ },
+ {
+ "epoch": 25.097816593886463,
+ "grad_norm": 0.021844089031219482,
+ "learning_rate": 0.0006,
+ "loss": 4.405138969421387,
+ "step": 1807
+ },
+ {
+ "epoch": 25.111790393013102,
+ "grad_norm": 0.022287018597126007,
+ "learning_rate": 0.0006,
+ "loss": 4.333793640136719,
+ "step": 1808
+ },
+ {
+ "epoch": 25.125764192139737,
+ "grad_norm": 0.02288329415023327,
+ "learning_rate": 0.0006,
+ "loss": 4.3910980224609375,
+ "step": 1809
+ },
+ {
+ "epoch": 25.139737991266376,
+ "grad_norm": 0.02373199723660946,
+ "learning_rate": 0.0006,
+ "loss": 4.32921028137207,
+ "step": 1810
+ },
+ {
+ "epoch": 25.153711790393015,
+ "grad_norm": 0.024960234761238098,
+ "learning_rate": 0.0006,
+ "loss": 4.295609474182129,
+ "step": 1811
+ },
+ {
+ "epoch": 25.16768558951965,
+ "grad_norm": 0.026743892580270767,
+ "learning_rate": 0.0006,
+ "loss": 4.315308094024658,
+ "step": 1812
+ },
+ {
+ "epoch": 25.18165938864629,
+ "grad_norm": 0.022138185799121857,
+ "learning_rate": 0.0006,
+ "loss": 4.360208511352539,
+ "step": 1813
+ },
+ {
+ "epoch": 25.195633187772927,
+ "grad_norm": 0.022120574489235878,
+ "learning_rate": 0.0006,
+ "loss": 4.1973772048950195,
+ "step": 1814
+ },
+ {
+ "epoch": 25.209606986899562,
+ "grad_norm": 0.02230502851307392,
+ "learning_rate": 0.0006,
+ "loss": 4.285558700561523,
+ "step": 1815
+ },
+ {
+ "epoch": 25.2235807860262,
+ "grad_norm": 0.022301986813545227,
+ "learning_rate": 0.0006,
+ "loss": 4.238343238830566,
+ "step": 1816
+ },
+ {
+ "epoch": 25.23755458515284,
+ "grad_norm": 0.020050447434186935,
+ "learning_rate": 0.0006,
+ "loss": 4.298235893249512,
+ "step": 1817
+ },
+ {
+ "epoch": 25.251528384279474,
+ "grad_norm": 0.01987724006175995,
+ "learning_rate": 0.0006,
+ "loss": 4.301384449005127,
+ "step": 1818
+ },
+ {
+ "epoch": 25.265502183406113,
+ "grad_norm": 0.020440855994820595,
+ "learning_rate": 0.0006,
+ "loss": 4.269218444824219,
+ "step": 1819
+ },
+ {
+ "epoch": 25.27947598253275,
+ "grad_norm": 0.01891893707215786,
+ "learning_rate": 0.0006,
+ "loss": 4.2401957511901855,
+ "step": 1820
+ },
+ {
+ "epoch": 25.29344978165939,
+ "grad_norm": 0.01903688907623291,
+ "learning_rate": 0.0006,
+ "loss": 4.139670372009277,
+ "step": 1821
+ },
+ {
+ "epoch": 25.307423580786025,
+ "grad_norm": 0.018856938928365707,
+ "learning_rate": 0.0006,
+ "loss": 4.233214378356934,
+ "step": 1822
+ },
+ {
+ "epoch": 25.321397379912664,
+ "grad_norm": 0.017791323363780975,
+ "learning_rate": 0.0006,
+ "loss": 4.2342023849487305,
+ "step": 1823
+ },
+ {
+ "epoch": 25.335371179039303,
+ "grad_norm": 0.017304565757513046,
+ "learning_rate": 0.0006,
+ "loss": 4.355198860168457,
+ "step": 1824
+ },
+ {
+ "epoch": 25.349344978165938,
+ "grad_norm": 0.01576206088066101,
+ "learning_rate": 0.0006,
+ "loss": 4.114147186279297,
+ "step": 1825
+ },
+ {
+ "epoch": 25.363318777292577,
+ "grad_norm": 0.015105154365301132,
+ "learning_rate": 0.0006,
+ "loss": 4.313795566558838,
+ "step": 1826
+ },
+ {
+ "epoch": 25.377292576419215,
+ "grad_norm": 0.01563587784767151,
+ "learning_rate": 0.0006,
+ "loss": 4.308282852172852,
+ "step": 1827
+ },
+ {
+ "epoch": 25.39126637554585,
+ "grad_norm": 0.015980906784534454,
+ "learning_rate": 0.0006,
+ "loss": 4.345149040222168,
+ "step": 1828
+ },
+ {
+ "epoch": 25.40524017467249,
+ "grad_norm": 0.01669284701347351,
+ "learning_rate": 0.0006,
+ "loss": 4.237611293792725,
+ "step": 1829
+ },
+ {
+ "epoch": 25.419213973799128,
+ "grad_norm": 0.016202064231038094,
+ "learning_rate": 0.0006,
+ "loss": 4.3072309494018555,
+ "step": 1830
+ },
+ {
+ "epoch": 25.433187772925763,
+ "grad_norm": 0.01478662807494402,
+ "learning_rate": 0.0006,
+ "loss": 4.276215076446533,
+ "step": 1831
+ },
+ {
+ "epoch": 25.4471615720524,
+ "grad_norm": 0.015168731100857258,
+ "learning_rate": 0.0006,
+ "loss": 4.37509822845459,
+ "step": 1832
+ },
+ {
+ "epoch": 25.46113537117904,
+ "grad_norm": 0.01648980937898159,
+ "learning_rate": 0.0006,
+ "loss": 4.349167346954346,
+ "step": 1833
+ },
+ {
+ "epoch": 25.475109170305675,
+ "grad_norm": 0.01708398200571537,
+ "learning_rate": 0.0006,
+ "loss": 4.346580505371094,
+ "step": 1834
+ },
+ {
+ "epoch": 25.489082969432314,
+ "grad_norm": 0.017370784655213356,
+ "learning_rate": 0.0006,
+ "loss": 4.299466133117676,
+ "step": 1835
+ },
+ {
+ "epoch": 25.503056768558952,
+ "grad_norm": 0.016248228028416634,
+ "learning_rate": 0.0006,
+ "loss": 4.184610366821289,
+ "step": 1836
+ },
+ {
+ "epoch": 25.51703056768559,
+ "grad_norm": 0.017106913030147552,
+ "learning_rate": 0.0006,
+ "loss": 4.366856098175049,
+ "step": 1837
+ },
+ {
+ "epoch": 25.531004366812226,
+ "grad_norm": 0.016071965917944908,
+ "learning_rate": 0.0006,
+ "loss": 4.297691345214844,
+ "step": 1838
+ },
+ {
+ "epoch": 25.544978165938865,
+ "grad_norm": 0.016451986506581306,
+ "learning_rate": 0.0006,
+ "loss": 4.229983329772949,
+ "step": 1839
+ },
+ {
+ "epoch": 25.558951965065503,
+ "grad_norm": 0.016195081174373627,
+ "learning_rate": 0.0006,
+ "loss": 4.259893417358398,
+ "step": 1840
+ },
+ {
+ "epoch": 25.57292576419214,
+ "grad_norm": 0.0165175162255764,
+ "learning_rate": 0.0006,
+ "loss": 4.3348236083984375,
+ "step": 1841
+ },
+ {
+ "epoch": 25.586899563318777,
+ "grad_norm": 0.01651517115533352,
+ "learning_rate": 0.0006,
+ "loss": 4.319541931152344,
+ "step": 1842
+ },
+ {
+ "epoch": 25.600873362445416,
+ "grad_norm": 0.016143690794706345,
+ "learning_rate": 0.0006,
+ "loss": 4.345652103424072,
+ "step": 1843
+ },
+ {
+ "epoch": 25.61484716157205,
+ "grad_norm": 0.015140696428716183,
+ "learning_rate": 0.0006,
+ "loss": 4.212047576904297,
+ "step": 1844
+ },
+ {
+ "epoch": 25.62882096069869,
+ "grad_norm": 0.014637443237006664,
+ "learning_rate": 0.0006,
+ "loss": 4.187453269958496,
+ "step": 1845
+ },
+ {
+ "epoch": 25.64279475982533,
+ "grad_norm": 0.014585614204406738,
+ "learning_rate": 0.0006,
+ "loss": 4.253936767578125,
+ "step": 1846
+ },
+ {
+ "epoch": 25.656768558951963,
+ "grad_norm": 0.015158289112150669,
+ "learning_rate": 0.0006,
+ "loss": 4.34612512588501,
+ "step": 1847
+ },
+ {
+ "epoch": 25.670742358078602,
+ "grad_norm": 0.015202849172055721,
+ "learning_rate": 0.0006,
+ "loss": 4.270634651184082,
+ "step": 1848
+ },
+ {
+ "epoch": 25.68471615720524,
+ "grad_norm": 0.015363575890660286,
+ "learning_rate": 0.0006,
+ "loss": 4.271080493927002,
+ "step": 1849
+ },
+ {
+ "epoch": 25.69868995633188,
+ "grad_norm": 0.014360226690769196,
+ "learning_rate": 0.0006,
+ "loss": 4.375033855438232,
+ "step": 1850
+ },
+ {
+ "epoch": 25.712663755458514,
+ "grad_norm": 0.014263181947171688,
+ "learning_rate": 0.0006,
+ "loss": 4.424138069152832,
+ "step": 1851
+ },
+ {
+ "epoch": 25.726637554585153,
+ "grad_norm": 0.014398688450455666,
+ "learning_rate": 0.0006,
+ "loss": 4.410154819488525,
+ "step": 1852
+ },
+ {
+ "epoch": 25.74061135371179,
+ "grad_norm": 0.016131488606333733,
+ "learning_rate": 0.0006,
+ "loss": 4.324014663696289,
+ "step": 1853
+ },
+ {
+ "epoch": 25.754585152838427,
+ "grad_norm": 0.017568735405802727,
+ "learning_rate": 0.0006,
+ "loss": 4.367861747741699,
+ "step": 1854
+ },
+ {
+ "epoch": 25.768558951965066,
+ "grad_norm": 0.017907511442899704,
+ "learning_rate": 0.0006,
+ "loss": 4.352965354919434,
+ "step": 1855
+ },
+ {
+ "epoch": 25.782532751091704,
+ "grad_norm": 0.016918016597628593,
+ "learning_rate": 0.0006,
+ "loss": 4.252397537231445,
+ "step": 1856
+ },
+ {
+ "epoch": 25.79650655021834,
+ "grad_norm": 0.017274610698223114,
+ "learning_rate": 0.0006,
+ "loss": 4.342309951782227,
+ "step": 1857
+ },
+ {
+ "epoch": 25.810480349344978,
+ "grad_norm": 0.017879825085401535,
+ "learning_rate": 0.0006,
+ "loss": 4.330987453460693,
+ "step": 1858
+ },
+ {
+ "epoch": 25.824454148471617,
+ "grad_norm": 0.01736099272966385,
+ "learning_rate": 0.0006,
+ "loss": 4.307255744934082,
+ "step": 1859
+ },
+ {
+ "epoch": 25.83842794759825,
+ "grad_norm": 0.016575824469327927,
+ "learning_rate": 0.0006,
+ "loss": 4.262016773223877,
+ "step": 1860
+ },
+ {
+ "epoch": 25.85240174672489,
+ "grad_norm": 0.015181140042841434,
+ "learning_rate": 0.0006,
+ "loss": 4.291863441467285,
+ "step": 1861
+ },
+ {
+ "epoch": 25.86637554585153,
+ "grad_norm": 0.015965687111020088,
+ "learning_rate": 0.0006,
+ "loss": 4.424836158752441,
+ "step": 1862
+ },
+ {
+ "epoch": 25.880349344978168,
+ "grad_norm": 0.01687219925224781,
+ "learning_rate": 0.0006,
+ "loss": 4.343531131744385,
+ "step": 1863
+ },
+ {
+ "epoch": 25.894323144104803,
+ "grad_norm": 0.015115504153072834,
+ "learning_rate": 0.0006,
+ "loss": 4.357808589935303,
+ "step": 1864
+ },
+ {
+ "epoch": 25.90829694323144,
+ "grad_norm": 0.014150998555123806,
+ "learning_rate": 0.0006,
+ "loss": 4.324550151824951,
+ "step": 1865
+ },
+ {
+ "epoch": 25.92227074235808,
+ "grad_norm": 0.014989510178565979,
+ "learning_rate": 0.0006,
+ "loss": 4.380428314208984,
+ "step": 1866
+ },
+ {
+ "epoch": 25.936244541484715,
+ "grad_norm": 0.01527960691601038,
+ "learning_rate": 0.0006,
+ "loss": 4.399374961853027,
+ "step": 1867
+ },
+ {
+ "epoch": 25.950218340611354,
+ "grad_norm": 0.015434633940458298,
+ "learning_rate": 0.0006,
+ "loss": 4.260984420776367,
+ "step": 1868
+ },
+ {
+ "epoch": 25.964192139737992,
+ "grad_norm": 0.015370228327810764,
+ "learning_rate": 0.0006,
+ "loss": 4.3340253829956055,
+ "step": 1869
+ },
+ {
+ "epoch": 25.978165938864628,
+ "grad_norm": 0.015348542481660843,
+ "learning_rate": 0.0006,
+ "loss": 4.369132995605469,
+ "step": 1870
+ },
+ {
+ "epoch": 25.992139737991266,
+ "grad_norm": 0.015354936942458153,
+ "learning_rate": 0.0006,
+ "loss": 4.381937026977539,
+ "step": 1871
+ },
+ {
+ "epoch": 26.0,
+ "grad_norm": 0.01796272210776806,
+ "learning_rate": 0.0006,
+ "loss": 4.286768436431885,
+ "step": 1872
+ },
+ {
+ "epoch": 26.0,
+ "eval_loss": 4.576984405517578,
+ "eval_runtime": 57.2091,
+ "eval_samples_per_second": 42.685,
+ "eval_steps_per_second": 1.346,
+ "step": 1872
+ },
+ {
+ "epoch": 26.01397379912664,
+ "grad_norm": 0.01748845726251602,
+ "learning_rate": 0.0006,
+ "loss": 4.337504863739014,
+ "step": 1873
+ },
+ {
+ "epoch": 26.027947598253274,
+ "grad_norm": 0.019244296476244926,
+ "learning_rate": 0.0006,
+ "loss": 4.264280796051025,
+ "step": 1874
+ },
+ {
+ "epoch": 26.041921397379912,
+ "grad_norm": 0.0215692650526762,
+ "learning_rate": 0.0006,
+ "loss": 4.350830078125,
+ "step": 1875
+ },
+ {
+ "epoch": 26.05589519650655,
+ "grad_norm": 0.022189252078533173,
+ "learning_rate": 0.0006,
+ "loss": 4.299098968505859,
+ "step": 1876
+ },
+ {
+ "epoch": 26.069868995633186,
+ "grad_norm": 0.022384043782949448,
+ "learning_rate": 0.0006,
+ "loss": 4.1185173988342285,
+ "step": 1877
+ },
+ {
+ "epoch": 26.083842794759825,
+ "grad_norm": 0.02076014317572117,
+ "learning_rate": 0.0006,
+ "loss": 4.297612190246582,
+ "step": 1878
+ },
+ {
+ "epoch": 26.097816593886463,
+ "grad_norm": 0.020843051373958588,
+ "learning_rate": 0.0006,
+ "loss": 4.294098854064941,
+ "step": 1879
+ },
+ {
+ "epoch": 26.111790393013102,
+ "grad_norm": 0.020654456689953804,
+ "learning_rate": 0.0006,
+ "loss": 4.267856597900391,
+ "step": 1880
+ },
+ {
+ "epoch": 26.125764192139737,
+ "grad_norm": 0.01860472559928894,
+ "learning_rate": 0.0006,
+ "loss": 4.284232139587402,
+ "step": 1881
+ },
+ {
+ "epoch": 26.139737991266376,
+ "grad_norm": 0.018719421699643135,
+ "learning_rate": 0.0006,
+ "loss": 4.277214527130127,
+ "step": 1882
+ },
+ {
+ "epoch": 26.153711790393015,
+ "grad_norm": 0.020903829485177994,
+ "learning_rate": 0.0006,
+ "loss": 4.254947662353516,
+ "step": 1883
+ },
+ {
+ "epoch": 26.16768558951965,
+ "grad_norm": 0.02020149491727352,
+ "learning_rate": 0.0006,
+ "loss": 4.2526936531066895,
+ "step": 1884
+ },
+ {
+ "epoch": 26.18165938864629,
+ "grad_norm": 0.0180392786860466,
+ "learning_rate": 0.0006,
+ "loss": 4.309922695159912,
+ "step": 1885
+ },
+ {
+ "epoch": 26.195633187772927,
+ "grad_norm": 0.018163125962018967,
+ "learning_rate": 0.0006,
+ "loss": 4.13860559463501,
+ "step": 1886
+ },
+ {
+ "epoch": 26.209606986899562,
+ "grad_norm": 0.01853291131556034,
+ "learning_rate": 0.0006,
+ "loss": 4.190926551818848,
+ "step": 1887
+ },
+ {
+ "epoch": 26.2235807860262,
+ "grad_norm": 0.01757677271962166,
+ "learning_rate": 0.0006,
+ "loss": 4.163999557495117,
+ "step": 1888
+ },
+ {
+ "epoch": 26.23755458515284,
+ "grad_norm": 0.01767992228269577,
+ "learning_rate": 0.0006,
+ "loss": 4.184430122375488,
+ "step": 1889
+ },
+ {
+ "epoch": 26.251528384279474,
+ "grad_norm": 0.018139973282814026,
+ "learning_rate": 0.0006,
+ "loss": 4.300766944885254,
+ "step": 1890
+ },
+ {
+ "epoch": 26.265502183406113,
+ "grad_norm": 0.020184461027383804,
+ "learning_rate": 0.0006,
+ "loss": 4.232537269592285,
+ "step": 1891
+ },
+ {
+ "epoch": 26.27947598253275,
+ "grad_norm": 0.02028856985270977,
+ "learning_rate": 0.0006,
+ "loss": 4.324883460998535,
+ "step": 1892
+ },
+ {
+ "epoch": 26.29344978165939,
+ "grad_norm": 0.017120616510510445,
+ "learning_rate": 0.0006,
+ "loss": 4.23493766784668,
+ "step": 1893
+ },
+ {
+ "epoch": 26.307423580786025,
+ "grad_norm": 0.016202978789806366,
+ "learning_rate": 0.0006,
+ "loss": 4.197602272033691,
+ "step": 1894
+ },
+ {
+ "epoch": 26.321397379912664,
+ "grad_norm": 0.017231421545147896,
+ "learning_rate": 0.0006,
+ "loss": 4.302001953125,
+ "step": 1895
+ },
+ {
+ "epoch": 26.335371179039303,
+ "grad_norm": 0.01725655607879162,
+ "learning_rate": 0.0006,
+ "loss": 4.356800079345703,
+ "step": 1896
+ },
+ {
+ "epoch": 26.349344978165938,
+ "grad_norm": 0.016657443717122078,
+ "learning_rate": 0.0006,
+ "loss": 4.266307830810547,
+ "step": 1897
+ },
+ {
+ "epoch": 26.363318777292577,
+ "grad_norm": 0.01616556942462921,
+ "learning_rate": 0.0006,
+ "loss": 4.307868003845215,
+ "step": 1898
+ },
+ {
+ "epoch": 26.377292576419215,
+ "grad_norm": 0.016621306538581848,
+ "learning_rate": 0.0006,
+ "loss": 4.172905921936035,
+ "step": 1899
+ },
+ {
+ "epoch": 26.39126637554585,
+ "grad_norm": 0.016243206337094307,
+ "learning_rate": 0.0006,
+ "loss": 4.162896156311035,
+ "step": 1900
+ },
+ {
+ "epoch": 26.40524017467249,
+ "grad_norm": 0.016663808375597,
+ "learning_rate": 0.0006,
+ "loss": 4.258232116699219,
+ "step": 1901
+ },
+ {
+ "epoch": 26.419213973799128,
+ "grad_norm": 0.01677478663623333,
+ "learning_rate": 0.0006,
+ "loss": 4.336090087890625,
+ "step": 1902
+ },
+ {
+ "epoch": 26.433187772925763,
+ "grad_norm": 0.01641303487122059,
+ "learning_rate": 0.0006,
+ "loss": 4.191476821899414,
+ "step": 1903
+ },
+ {
+ "epoch": 26.4471615720524,
+ "grad_norm": 0.016700323671102524,
+ "learning_rate": 0.0006,
+ "loss": 4.299427509307861,
+ "step": 1904
+ },
+ {
+ "epoch": 26.46113537117904,
+ "grad_norm": 0.017595035955309868,
+ "learning_rate": 0.0006,
+ "loss": 4.312828063964844,
+ "step": 1905
+ },
+ {
+ "epoch": 26.475109170305675,
+ "grad_norm": 0.020300284028053284,
+ "learning_rate": 0.0006,
+ "loss": 4.28331184387207,
+ "step": 1906
+ },
+ {
+ "epoch": 26.489082969432314,
+ "grad_norm": 0.02198321744799614,
+ "learning_rate": 0.0006,
+ "loss": 4.3280839920043945,
+ "step": 1907
+ },
+ {
+ "epoch": 26.503056768558952,
+ "grad_norm": 0.02003200724720955,
+ "learning_rate": 0.0006,
+ "loss": 4.325623989105225,
+ "step": 1908
+ },
+ {
+ "epoch": 26.51703056768559,
+ "grad_norm": 0.01870287023484707,
+ "learning_rate": 0.0006,
+ "loss": 4.251203536987305,
+ "step": 1909
+ },
+ {
+ "epoch": 26.531004366812226,
+ "grad_norm": 0.01835448481142521,
+ "learning_rate": 0.0006,
+ "loss": 4.335144996643066,
+ "step": 1910
+ },
+ {
+ "epoch": 26.544978165938865,
+ "grad_norm": 0.02046327292919159,
+ "learning_rate": 0.0006,
+ "loss": 4.238023281097412,
+ "step": 1911
+ },
+ {
+ "epoch": 26.558951965065503,
+ "grad_norm": 0.019771311432123184,
+ "learning_rate": 0.0006,
+ "loss": 4.353594779968262,
+ "step": 1912
+ },
+ {
+ "epoch": 26.57292576419214,
+ "grad_norm": 0.018090544268488884,
+ "learning_rate": 0.0006,
+ "loss": 4.304343223571777,
+ "step": 1913
+ },
+ {
+ "epoch": 26.586899563318777,
+ "grad_norm": 0.017489777877926826,
+ "learning_rate": 0.0006,
+ "loss": 4.204375267028809,
+ "step": 1914
+ },
+ {
+ "epoch": 26.600873362445416,
+ "grad_norm": 0.016785891726613045,
+ "learning_rate": 0.0006,
+ "loss": 4.286669731140137,
+ "step": 1915
+ },
+ {
+ "epoch": 26.61484716157205,
+ "grad_norm": 0.016382789239287376,
+ "learning_rate": 0.0006,
+ "loss": 4.347662448883057,
+ "step": 1916
+ },
+ {
+ "epoch": 26.62882096069869,
+ "grad_norm": 0.016581876203417778,
+ "learning_rate": 0.0006,
+ "loss": 4.2716193199157715,
+ "step": 1917
+ },
+ {
+ "epoch": 26.64279475982533,
+ "grad_norm": 0.015627184882760048,
+ "learning_rate": 0.0006,
+ "loss": 4.311489105224609,
+ "step": 1918
+ },
+ {
+ "epoch": 26.656768558951963,
+ "grad_norm": 0.015908483415842056,
+ "learning_rate": 0.0006,
+ "loss": 4.262099742889404,
+ "step": 1919
+ },
+ {
+ "epoch": 26.670742358078602,
+ "grad_norm": 0.015576236881315708,
+ "learning_rate": 0.0006,
+ "loss": 4.275871276855469,
+ "step": 1920
+ },
+ {
+ "epoch": 26.68471615720524,
+ "grad_norm": 0.014621064998209476,
+ "learning_rate": 0.0006,
+ "loss": 4.184260845184326,
+ "step": 1921
+ },
+ {
+ "epoch": 26.69868995633188,
+ "grad_norm": 0.014797811396420002,
+ "learning_rate": 0.0006,
+ "loss": 4.3412628173828125,
+ "step": 1922
+ },
+ {
+ "epoch": 26.712663755458514,
+ "grad_norm": 0.015610570088028908,
+ "learning_rate": 0.0006,
+ "loss": 4.181094646453857,
+ "step": 1923
+ },
+ {
+ "epoch": 26.726637554585153,
+ "grad_norm": 0.016048265621066093,
+ "learning_rate": 0.0006,
+ "loss": 4.365290641784668,
+ "step": 1924
+ },
+ {
+ "epoch": 26.74061135371179,
+ "grad_norm": 0.016604071483016014,
+ "learning_rate": 0.0006,
+ "loss": 4.2628912925720215,
+ "step": 1925
+ },
+ {
+ "epoch": 26.754585152838427,
+ "grad_norm": 0.017044221982359886,
+ "learning_rate": 0.0006,
+ "loss": 4.378512382507324,
+ "step": 1926
+ },
+ {
+ "epoch": 26.768558951965066,
+ "grad_norm": 0.01746384985744953,
+ "learning_rate": 0.0006,
+ "loss": 4.3888773918151855,
+ "step": 1927
+ },
+ {
+ "epoch": 26.782532751091704,
+ "grad_norm": 0.01572711206972599,
+ "learning_rate": 0.0006,
+ "loss": 4.189935684204102,
+ "step": 1928
+ },
+ {
+ "epoch": 26.79650655021834,
+ "grad_norm": 0.014976629987359047,
+ "learning_rate": 0.0006,
+ "loss": 4.280525207519531,
+ "step": 1929
+ },
+ {
+ "epoch": 26.810480349344978,
+ "grad_norm": 0.014804664999246597,
+ "learning_rate": 0.0006,
+ "loss": 4.270630836486816,
+ "step": 1930
+ },
+ {
+ "epoch": 26.824454148471617,
+ "grad_norm": 0.014641920104622841,
+ "learning_rate": 0.0006,
+ "loss": 4.319505214691162,
+ "step": 1931
+ },
+ {
+ "epoch": 26.83842794759825,
+ "grad_norm": 0.014928505755960941,
+ "learning_rate": 0.0006,
+ "loss": 4.180147647857666,
+ "step": 1932
+ },
+ {
+ "epoch": 26.85240174672489,
+ "grad_norm": 0.016130365431308746,
+ "learning_rate": 0.0006,
+ "loss": 4.260660648345947,
+ "step": 1933
+ },
+ {
+ "epoch": 26.86637554585153,
+ "grad_norm": 0.015642493963241577,
+ "learning_rate": 0.0006,
+ "loss": 4.304064750671387,
+ "step": 1934
+ },
+ {
+ "epoch": 26.880349344978168,
+ "grad_norm": 0.015988217666745186,
+ "learning_rate": 0.0006,
+ "loss": 4.2548017501831055,
+ "step": 1935
+ },
+ {
+ "epoch": 26.894323144104803,
+ "grad_norm": 0.016501398757100105,
+ "learning_rate": 0.0006,
+ "loss": 4.260058879852295,
+ "step": 1936
+ },
+ {
+ "epoch": 26.90829694323144,
+ "grad_norm": 0.01645040698349476,
+ "learning_rate": 0.0006,
+ "loss": 4.483081340789795,
+ "step": 1937
+ },
+ {
+ "epoch": 26.92227074235808,
+ "grad_norm": 0.016140863299369812,
+ "learning_rate": 0.0006,
+ "loss": 4.1776838302612305,
+ "step": 1938
+ },
+ {
+ "epoch": 26.936244541484715,
+ "grad_norm": 0.016681019216775894,
+ "learning_rate": 0.0006,
+ "loss": 4.224747657775879,
+ "step": 1939
+ },
+ {
+ "epoch": 26.950218340611354,
+ "grad_norm": 0.01650378853082657,
+ "learning_rate": 0.0006,
+ "loss": 4.305703163146973,
+ "step": 1940
+ },
+ {
+ "epoch": 26.964192139737992,
+ "grad_norm": 0.017051683738827705,
+ "learning_rate": 0.0006,
+ "loss": 4.285835266113281,
+ "step": 1941
+ },
+ {
+ "epoch": 26.978165938864628,
+ "grad_norm": 0.016894327476620674,
+ "learning_rate": 0.0006,
+ "loss": 4.318901538848877,
+ "step": 1942
+ },
+ {
+ "epoch": 26.992139737991266,
+ "grad_norm": 0.017313100397586823,
+ "learning_rate": 0.0006,
+ "loss": 4.281434535980225,
+ "step": 1943
+ },
+ {
+ "epoch": 27.0,
+ "grad_norm": 0.018450884148478508,
+ "learning_rate": 0.0006,
+ "loss": 4.335104942321777,
+ "step": 1944
+ },
+ {
+ "epoch": 27.0,
+ "eval_loss": 4.640687465667725,
+ "eval_runtime": 56.9624,
+ "eval_samples_per_second": 42.87,
+ "eval_steps_per_second": 1.352,
+ "step": 1944
+ },
+ {
+ "epoch": 27.01397379912664,
+ "grad_norm": 0.017755698412656784,
+ "learning_rate": 0.0006,
+ "loss": 4.278824329376221,
+ "step": 1945
+ },
+ {
+ "epoch": 27.027947598253274,
+ "grad_norm": 0.01724310778081417,
+ "learning_rate": 0.0006,
+ "loss": 4.311049461364746,
+ "step": 1946
+ },
+ {
+ "epoch": 27.041921397379912,
+ "grad_norm": 0.01690123789012432,
+ "learning_rate": 0.0006,
+ "loss": 4.261994361877441,
+ "step": 1947
+ },
+ {
+ "epoch": 27.05589519650655,
+ "grad_norm": 0.018406696617603302,
+ "learning_rate": 0.0006,
+ "loss": 4.1925153732299805,
+ "step": 1948
+ },
+ {
+ "epoch": 27.069868995633186,
+ "grad_norm": 0.019945867359638214,
+ "learning_rate": 0.0006,
+ "loss": 4.108537197113037,
+ "step": 1949
+ },
+ {
+ "epoch": 27.083842794759825,
+ "grad_norm": 0.020316628739237785,
+ "learning_rate": 0.0006,
+ "loss": 4.209043025970459,
+ "step": 1950
+ },
+ {
+ "epoch": 27.097816593886463,
+ "grad_norm": 0.019690370187163353,
+ "learning_rate": 0.0006,
+ "loss": 4.195789337158203,
+ "step": 1951
+ },
+ {
+ "epoch": 27.111790393013102,
+ "grad_norm": 0.01785232499241829,
+ "learning_rate": 0.0006,
+ "loss": 4.21466064453125,
+ "step": 1952
+ },
+ {
+ "epoch": 27.125764192139737,
+ "grad_norm": 0.0164767038077116,
+ "learning_rate": 0.0006,
+ "loss": 4.270247459411621,
+ "step": 1953
+ },
+ {
+ "epoch": 27.139737991266376,
+ "grad_norm": 0.018008455634117126,
+ "learning_rate": 0.0006,
+ "loss": 4.201042175292969,
+ "step": 1954
+ },
+ {
+ "epoch": 27.153711790393015,
+ "grad_norm": 0.01807340793311596,
+ "learning_rate": 0.0006,
+ "loss": 4.25289249420166,
+ "step": 1955
+ },
+ {
+ "epoch": 27.16768558951965,
+ "grad_norm": 0.016246909275650978,
+ "learning_rate": 0.0006,
+ "loss": 4.116283416748047,
+ "step": 1956
+ },
+ {
+ "epoch": 27.18165938864629,
+ "grad_norm": 0.017545776441693306,
+ "learning_rate": 0.0006,
+ "loss": 4.316999435424805,
+ "step": 1957
+ },
+ {
+ "epoch": 27.195633187772927,
+ "grad_norm": 0.017915375530719757,
+ "learning_rate": 0.0006,
+ "loss": 4.253859519958496,
+ "step": 1958
+ },
+ {
+ "epoch": 27.209606986899562,
+ "grad_norm": 0.018112223595380783,
+ "learning_rate": 0.0006,
+ "loss": 4.2541022300720215,
+ "step": 1959
+ },
+ {
+ "epoch": 27.2235807860262,
+ "grad_norm": 0.01846139505505562,
+ "learning_rate": 0.0006,
+ "loss": 4.20438289642334,
+ "step": 1960
+ },
+ {
+ "epoch": 27.23755458515284,
+ "grad_norm": 0.01943897269666195,
+ "learning_rate": 0.0006,
+ "loss": 4.2799577713012695,
+ "step": 1961
+ },
+ {
+ "epoch": 27.251528384279474,
+ "grad_norm": 0.018253348767757416,
+ "learning_rate": 0.0006,
+ "loss": 4.1720499992370605,
+ "step": 1962
+ },
+ {
+ "epoch": 27.265502183406113,
+ "grad_norm": 0.018292246386408806,
+ "learning_rate": 0.0006,
+ "loss": 4.28231143951416,
+ "step": 1963
+ },
+ {
+ "epoch": 27.27947598253275,
+ "grad_norm": 0.01943155936896801,
+ "learning_rate": 0.0006,
+ "loss": 4.314976692199707,
+ "step": 1964
+ },
+ {
+ "epoch": 27.29344978165939,
+ "grad_norm": 0.019064994528889656,
+ "learning_rate": 0.0006,
+ "loss": 4.25328254699707,
+ "step": 1965
+ },
+ {
+ "epoch": 27.307423580786025,
+ "grad_norm": 0.017423273995518684,
+ "learning_rate": 0.0006,
+ "loss": 4.195992469787598,
+ "step": 1966
+ },
+ {
+ "epoch": 27.321397379912664,
+ "grad_norm": 0.01677924580872059,
+ "learning_rate": 0.0006,
+ "loss": 4.260043144226074,
+ "step": 1967
+ },
+ {
+ "epoch": 27.335371179039303,
+ "grad_norm": 0.01720673404633999,
+ "learning_rate": 0.0006,
+ "loss": 4.222780704498291,
+ "step": 1968
+ },
+ {
+ "epoch": 27.349344978165938,
+ "grad_norm": 0.017300743609666824,
+ "learning_rate": 0.0006,
+ "loss": 4.109524250030518,
+ "step": 1969
+ },
+ {
+ "epoch": 27.363318777292577,
+ "grad_norm": 0.016065871343016624,
+ "learning_rate": 0.0006,
+ "loss": 4.36108922958374,
+ "step": 1970
+ },
+ {
+ "epoch": 27.377292576419215,
+ "grad_norm": 0.016626616939902306,
+ "learning_rate": 0.0006,
+ "loss": 4.220630645751953,
+ "step": 1971
+ },
+ {
+ "epoch": 27.39126637554585,
+ "grad_norm": 0.016482602804899216,
+ "learning_rate": 0.0006,
+ "loss": 4.344797134399414,
+ "step": 1972
+ },
+ {
+ "epoch": 27.40524017467249,
+ "grad_norm": 0.01727953553199768,
+ "learning_rate": 0.0006,
+ "loss": 4.296645164489746,
+ "step": 1973
+ },
+ {
+ "epoch": 27.419213973799128,
+ "grad_norm": 0.01522731315344572,
+ "learning_rate": 0.0006,
+ "loss": 4.223751068115234,
+ "step": 1974
+ },
+ {
+ "epoch": 27.433187772925763,
+ "grad_norm": 0.01581776700913906,
+ "learning_rate": 0.0006,
+ "loss": 4.351379871368408,
+ "step": 1975
+ },
+ {
+ "epoch": 27.4471615720524,
+ "grad_norm": 0.016394605860114098,
+ "learning_rate": 0.0006,
+ "loss": 4.2412919998168945,
+ "step": 1976
+ },
+ {
+ "epoch": 27.46113537117904,
+ "grad_norm": 0.01751169003546238,
+ "learning_rate": 0.0006,
+ "loss": 4.257606506347656,
+ "step": 1977
+ },
+ {
+ "epoch": 27.475109170305675,
+ "grad_norm": 0.017452310770750046,
+ "learning_rate": 0.0006,
+ "loss": 4.380134582519531,
+ "step": 1978
+ },
+ {
+ "epoch": 27.489082969432314,
+ "grad_norm": 0.017741898074746132,
+ "learning_rate": 0.0006,
+ "loss": 4.279632568359375,
+ "step": 1979
+ },
+ {
+ "epoch": 27.503056768558952,
+ "grad_norm": 0.017680590972304344,
+ "learning_rate": 0.0006,
+ "loss": 4.273406028747559,
+ "step": 1980
+ },
+ {
+ "epoch": 27.51703056768559,
+ "grad_norm": 0.01864718459546566,
+ "learning_rate": 0.0006,
+ "loss": 4.278583526611328,
+ "step": 1981
+ },
+ {
+ "epoch": 27.531004366812226,
+ "grad_norm": 0.019558541476726532,
+ "learning_rate": 0.0006,
+ "loss": 4.211919784545898,
+ "step": 1982
+ },
+ {
+ "epoch": 27.544978165938865,
+ "grad_norm": 0.019467400386929512,
+ "learning_rate": 0.0006,
+ "loss": 4.276822566986084,
+ "step": 1983
+ },
+ {
+ "epoch": 27.558951965065503,
+ "grad_norm": 0.01756172813475132,
+ "learning_rate": 0.0006,
+ "loss": 4.165379047393799,
+ "step": 1984
+ },
+ {
+ "epoch": 27.57292576419214,
+ "grad_norm": 0.017458263784646988,
+ "learning_rate": 0.0006,
+ "loss": 4.2992401123046875,
+ "step": 1985
+ },
+ {
+ "epoch": 27.586899563318777,
+ "grad_norm": 0.018089665099978447,
+ "learning_rate": 0.0006,
+ "loss": 4.243378639221191,
+ "step": 1986
+ },
+ {
+ "epoch": 27.600873362445416,
+ "grad_norm": 0.019535524770617485,
+ "learning_rate": 0.0006,
+ "loss": 4.321477890014648,
+ "step": 1987
+ },
+ {
+ "epoch": 27.61484716157205,
+ "grad_norm": 0.01997970975935459,
+ "learning_rate": 0.0006,
+ "loss": 4.123082637786865,
+ "step": 1988
+ },
+ {
+ "epoch": 27.62882096069869,
+ "grad_norm": 0.01913067139685154,
+ "learning_rate": 0.0006,
+ "loss": 4.289896488189697,
+ "step": 1989
+ },
+ {
+ "epoch": 27.64279475982533,
+ "grad_norm": 0.0169806070625782,
+ "learning_rate": 0.0006,
+ "loss": 4.247004985809326,
+ "step": 1990
+ },
+ {
+ "epoch": 27.656768558951963,
+ "grad_norm": 0.017315007746219635,
+ "learning_rate": 0.0006,
+ "loss": 4.2095441818237305,
+ "step": 1991
+ },
+ {
+ "epoch": 27.670742358078602,
+ "grad_norm": 0.015098759904503822,
+ "learning_rate": 0.0006,
+ "loss": 4.145560264587402,
+ "step": 1992
+ },
+ {
+ "epoch": 27.68471615720524,
+ "grad_norm": 0.016240514814853668,
+ "learning_rate": 0.0006,
+ "loss": 4.254927635192871,
+ "step": 1993
+ },
+ {
+ "epoch": 27.69868995633188,
+ "grad_norm": 0.01621108129620552,
+ "learning_rate": 0.0006,
+ "loss": 4.219725608825684,
+ "step": 1994
+ },
+ {
+ "epoch": 27.712663755458514,
+ "grad_norm": 0.015191009268164635,
+ "learning_rate": 0.0006,
+ "loss": 4.194345474243164,
+ "step": 1995
+ },
+ {
+ "epoch": 27.726637554585153,
+ "grad_norm": 0.014387983828783035,
+ "learning_rate": 0.0006,
+ "loss": 4.211467742919922,
+ "step": 1996
+ },
+ {
+ "epoch": 27.74061135371179,
+ "grad_norm": 0.015460561029613018,
+ "learning_rate": 0.0006,
+ "loss": 4.358700752258301,
+ "step": 1997
+ },
+ {
+ "epoch": 27.754585152838427,
+ "grad_norm": 0.015100076794624329,
+ "learning_rate": 0.0006,
+ "loss": 4.171223163604736,
+ "step": 1998
+ },
+ {
+ "epoch": 27.768558951965066,
+ "grad_norm": 0.015607891604304314,
+ "learning_rate": 0.0006,
+ "loss": 4.217874526977539,
+ "step": 1999
+ },
+ {
+ "epoch": 27.782532751091704,
+ "grad_norm": 0.015650689601898193,
+ "learning_rate": 0.0006,
+ "loss": 4.280780792236328,
+ "step": 2000
+ },
+ {
+ "epoch": 27.79650655021834,
+ "grad_norm": 0.01580994389951229,
+ "learning_rate": 0.0006,
+ "loss": 4.159212112426758,
+ "step": 2001
+ },
+ {
+ "epoch": 27.810480349344978,
+ "grad_norm": 0.016607413068413734,
+ "learning_rate": 0.0006,
+ "loss": 4.238853454589844,
+ "step": 2002
+ },
+ {
+ "epoch": 27.824454148471617,
+ "grad_norm": 0.017383860424160957,
+ "learning_rate": 0.0006,
+ "loss": 4.233944892883301,
+ "step": 2003
+ },
+ {
+ "epoch": 27.83842794759825,
+ "grad_norm": 0.014966659247875214,
+ "learning_rate": 0.0006,
+ "loss": 4.167392730712891,
+ "step": 2004
+ },
+ {
+ "epoch": 27.85240174672489,
+ "grad_norm": 0.014589878730475903,
+ "learning_rate": 0.0006,
+ "loss": 4.2390899658203125,
+ "step": 2005
+ },
+ {
+ "epoch": 27.86637554585153,
+ "grad_norm": 0.017179450020194054,
+ "learning_rate": 0.0006,
+ "loss": 4.365785598754883,
+ "step": 2006
+ },
+ {
+ "epoch": 27.880349344978168,
+ "grad_norm": 0.01893901452422142,
+ "learning_rate": 0.0006,
+ "loss": 4.289430141448975,
+ "step": 2007
+ },
+ {
+ "epoch": 27.894323144104803,
+ "grad_norm": 0.018619712442159653,
+ "learning_rate": 0.0006,
+ "loss": 4.310665130615234,
+ "step": 2008
+ },
+ {
+ "epoch": 27.90829694323144,
+ "grad_norm": 0.019180385395884514,
+ "learning_rate": 0.0006,
+ "loss": 4.27454137802124,
+ "step": 2009
+ },
+ {
+ "epoch": 27.92227074235808,
+ "grad_norm": 0.017811523750424385,
+ "learning_rate": 0.0006,
+ "loss": 4.312562465667725,
+ "step": 2010
+ },
+ {
+ "epoch": 27.936244541484715,
+ "grad_norm": 0.016583573073148727,
+ "learning_rate": 0.0006,
+ "loss": 4.202253341674805,
+ "step": 2011
+ },
+ {
+ "epoch": 27.950218340611354,
+ "grad_norm": 0.01658615842461586,
+ "learning_rate": 0.0006,
+ "loss": 4.183433532714844,
+ "step": 2012
+ },
+ {
+ "epoch": 27.964192139737992,
+ "grad_norm": 0.015327401459217072,
+ "learning_rate": 0.0006,
+ "loss": 4.2125139236450195,
+ "step": 2013
+ },
+ {
+ "epoch": 27.978165938864628,
+ "grad_norm": 0.015004601329565048,
+ "learning_rate": 0.0006,
+ "loss": 4.257392883300781,
+ "step": 2014
+ },
+ {
+ "epoch": 27.992139737991266,
+ "grad_norm": 0.015259161591529846,
+ "learning_rate": 0.0006,
+ "loss": 4.241818428039551,
+ "step": 2015
+ },
+ {
+ "epoch": 28.0,
+ "grad_norm": 0.016808776184916496,
+ "learning_rate": 0.0006,
+ "loss": 4.146241188049316,
+ "step": 2016
+ },
+ {
+ "epoch": 28.0,
+ "eval_loss": 4.663090705871582,
+ "eval_runtime": 56.4559,
+ "eval_samples_per_second": 43.255,
+ "eval_steps_per_second": 1.364,
+ "step": 2016
+ },
+ {
+ "epoch": 28.01397379912664,
+ "grad_norm": 0.015784382820129395,
+ "learning_rate": 0.0006,
+ "loss": 4.044191360473633,
+ "step": 2017
+ },
+ {
+ "epoch": 28.027947598253274,
+ "grad_norm": 0.015636710450053215,
+ "learning_rate": 0.0006,
+ "loss": 4.128754615783691,
+ "step": 2018
+ },
+ {
+ "epoch": 28.041921397379912,
+ "grad_norm": 0.015364975668489933,
+ "learning_rate": 0.0006,
+ "loss": 4.191835403442383,
+ "step": 2019
+ },
+ {
+ "epoch": 28.05589519650655,
+ "grad_norm": 0.015015700832009315,
+ "learning_rate": 0.0006,
+ "loss": 4.120022773742676,
+ "step": 2020
+ },
+ {
+ "epoch": 28.069868995633186,
+ "grad_norm": 0.016666799783706665,
+ "learning_rate": 0.0006,
+ "loss": 4.33906364440918,
+ "step": 2021
+ },
+ {
+ "epoch": 28.083842794759825,
+ "grad_norm": 0.01780104823410511,
+ "learning_rate": 0.0006,
+ "loss": 4.279458999633789,
+ "step": 2022
+ },
+ {
+ "epoch": 28.097816593886463,
+ "grad_norm": 0.01870882883667946,
+ "learning_rate": 0.0006,
+ "loss": 4.34728479385376,
+ "step": 2023
+ },
+ {
+ "epoch": 28.111790393013102,
+ "grad_norm": 0.018406381830573082,
+ "learning_rate": 0.0006,
+ "loss": 4.211311340332031,
+ "step": 2024
+ },
+ {
+ "epoch": 28.125764192139737,
+ "grad_norm": 0.017833448946475983,
+ "learning_rate": 0.0006,
+ "loss": 4.223254203796387,
+ "step": 2025
+ },
+ {
+ "epoch": 28.139737991266376,
+ "grad_norm": 0.017552688717842102,
+ "learning_rate": 0.0006,
+ "loss": 4.208823204040527,
+ "step": 2026
+ },
+ {
+ "epoch": 28.153711790393015,
+ "grad_norm": 0.018500229343771935,
+ "learning_rate": 0.0006,
+ "loss": 4.198025226593018,
+ "step": 2027
+ },
+ {
+ "epoch": 28.16768558951965,
+ "grad_norm": 0.01782156340777874,
+ "learning_rate": 0.0006,
+ "loss": 4.179978370666504,
+ "step": 2028
+ },
+ {
+ "epoch": 28.18165938864629,
+ "grad_norm": 0.019580967724323273,
+ "learning_rate": 0.0006,
+ "loss": 4.291114807128906,
+ "step": 2029
+ },
+ {
+ "epoch": 28.195633187772927,
+ "grad_norm": 0.020804036408662796,
+ "learning_rate": 0.0006,
+ "loss": 4.220607280731201,
+ "step": 2030
+ },
+ {
+ "epoch": 28.209606986899562,
+ "grad_norm": 0.019595693796873093,
+ "learning_rate": 0.0006,
+ "loss": 4.2905731201171875,
+ "step": 2031
+ },
+ {
+ "epoch": 28.2235807860262,
+ "grad_norm": 0.019247086718678474,
+ "learning_rate": 0.0006,
+ "loss": 4.2467360496521,
+ "step": 2032
+ },
+ {
+ "epoch": 28.23755458515284,
+ "grad_norm": 0.01958036608994007,
+ "learning_rate": 0.0006,
+ "loss": 4.286137580871582,
+ "step": 2033
+ },
+ {
+ "epoch": 28.251528384279474,
+ "grad_norm": 0.021354753524065018,
+ "learning_rate": 0.0006,
+ "loss": 4.234755039215088,
+ "step": 2034
+ },
+ {
+ "epoch": 28.265502183406113,
+ "grad_norm": 0.02206375077366829,
+ "learning_rate": 0.0006,
+ "loss": 4.149641036987305,
+ "step": 2035
+ },
+ {
+ "epoch": 28.27947598253275,
+ "grad_norm": 0.02201310358941555,
+ "learning_rate": 0.0006,
+ "loss": 4.1928629875183105,
+ "step": 2036
+ },
+ {
+ "epoch": 28.29344978165939,
+ "grad_norm": 0.01961950585246086,
+ "learning_rate": 0.0006,
+ "loss": 4.211404323577881,
+ "step": 2037
+ },
+ {
+ "epoch": 28.307423580786025,
+ "grad_norm": 0.019211795181035995,
+ "learning_rate": 0.0006,
+ "loss": 4.200976848602295,
+ "step": 2038
+ },
+ {
+ "epoch": 28.321397379912664,
+ "grad_norm": 0.01779370754957199,
+ "learning_rate": 0.0006,
+ "loss": 4.249148845672607,
+ "step": 2039
+ },
+ {
+ "epoch": 28.335371179039303,
+ "grad_norm": 0.018635908141732216,
+ "learning_rate": 0.0006,
+ "loss": 4.11806058883667,
+ "step": 2040
+ },
+ {
+ "epoch": 28.349344978165938,
+ "grad_norm": 0.019792694598436356,
+ "learning_rate": 0.0006,
+ "loss": 4.335224151611328,
+ "step": 2041
+ },
+ {
+ "epoch": 28.363318777292577,
+ "grad_norm": 0.021731717512011528,
+ "learning_rate": 0.0006,
+ "loss": 4.302778244018555,
+ "step": 2042
+ },
+ {
+ "epoch": 28.377292576419215,
+ "grad_norm": 0.020823447033762932,
+ "learning_rate": 0.0006,
+ "loss": 4.260610580444336,
+ "step": 2043
+ },
+ {
+ "epoch": 28.39126637554585,
+ "grad_norm": 0.019385412335395813,
+ "learning_rate": 0.0006,
+ "loss": 4.2247724533081055,
+ "step": 2044
+ },
+ {
+ "epoch": 28.40524017467249,
+ "grad_norm": 0.019539108499884605,
+ "learning_rate": 0.0006,
+ "loss": 4.236655235290527,
+ "step": 2045
+ },
+ {
+ "epoch": 28.419213973799128,
+ "grad_norm": 0.01942690648138523,
+ "learning_rate": 0.0006,
+ "loss": 4.251850605010986,
+ "step": 2046
+ },
+ {
+ "epoch": 28.433187772925763,
+ "grad_norm": 0.019667886197566986,
+ "learning_rate": 0.0006,
+ "loss": 4.222312927246094,
+ "step": 2047
+ },
+ {
+ "epoch": 28.4471615720524,
+ "grad_norm": 0.02029012329876423,
+ "learning_rate": 0.0006,
+ "loss": 4.246252059936523,
+ "step": 2048
+ },
+ {
+ "epoch": 28.46113537117904,
+ "grad_norm": 0.01784469373524189,
+ "learning_rate": 0.0006,
+ "loss": 4.192128658294678,
+ "step": 2049
+ },
+ {
+ "epoch": 28.475109170305675,
+ "grad_norm": 0.0160287544131279,
+ "learning_rate": 0.0006,
+ "loss": 4.208607196807861,
+ "step": 2050
+ },
+ {
+ "epoch": 28.489082969432314,
+ "grad_norm": 0.015348346903920174,
+ "learning_rate": 0.0006,
+ "loss": 4.236598014831543,
+ "step": 2051
+ },
+ {
+ "epoch": 28.503056768558952,
+ "grad_norm": 0.015198206529021263,
+ "learning_rate": 0.0006,
+ "loss": 4.235552787780762,
+ "step": 2052
+ },
+ {
+ "epoch": 28.51703056768559,
+ "grad_norm": 0.01523754745721817,
+ "learning_rate": 0.0006,
+ "loss": 4.1941728591918945,
+ "step": 2053
+ },
+ {
+ "epoch": 28.531004366812226,
+ "grad_norm": 0.015559614636003971,
+ "learning_rate": 0.0006,
+ "loss": 4.294757843017578,
+ "step": 2054
+ },
+ {
+ "epoch": 28.544978165938865,
+ "grad_norm": 0.01568358950316906,
+ "learning_rate": 0.0006,
+ "loss": 4.213065147399902,
+ "step": 2055
+ },
+ {
+ "epoch": 28.558951965065503,
+ "grad_norm": 0.015336516313254833,
+ "learning_rate": 0.0006,
+ "loss": 4.240293979644775,
+ "step": 2056
+ },
+ {
+ "epoch": 28.57292576419214,
+ "grad_norm": 0.015041496604681015,
+ "learning_rate": 0.0006,
+ "loss": 4.273397445678711,
+ "step": 2057
+ },
+ {
+ "epoch": 28.586899563318777,
+ "grad_norm": 0.01610128954052925,
+ "learning_rate": 0.0006,
+ "loss": 4.125369071960449,
+ "step": 2058
+ },
+ {
+ "epoch": 28.600873362445416,
+ "grad_norm": 0.015974976122379303,
+ "learning_rate": 0.0006,
+ "loss": 4.133459091186523,
+ "step": 2059
+ },
+ {
+ "epoch": 28.61484716157205,
+ "grad_norm": 0.015936799347400665,
+ "learning_rate": 0.0006,
+ "loss": 4.306667327880859,
+ "step": 2060
+ },
+ {
+ "epoch": 28.62882096069869,
+ "grad_norm": 0.01795247197151184,
+ "learning_rate": 0.0006,
+ "loss": 4.205156326293945,
+ "step": 2061
+ },
+ {
+ "epoch": 28.64279475982533,
+ "grad_norm": 0.019639814272522926,
+ "learning_rate": 0.0006,
+ "loss": 4.277614593505859,
+ "step": 2062
+ },
+ {
+ "epoch": 28.656768558951963,
+ "grad_norm": 0.01722300611436367,
+ "learning_rate": 0.0006,
+ "loss": 4.191032886505127,
+ "step": 2063
+ },
+ {
+ "epoch": 28.670742358078602,
+ "grad_norm": 0.01612633652985096,
+ "learning_rate": 0.0006,
+ "loss": 4.20717191696167,
+ "step": 2064
+ },
+ {
+ "epoch": 28.68471615720524,
+ "grad_norm": 0.01603279449045658,
+ "learning_rate": 0.0006,
+ "loss": 4.228884220123291,
+ "step": 2065
+ },
+ {
+ "epoch": 28.69868995633188,
+ "grad_norm": 0.016405927017331123,
+ "learning_rate": 0.0006,
+ "loss": 4.201292991638184,
+ "step": 2066
+ },
+ {
+ "epoch": 28.712663755458514,
+ "grad_norm": 0.016644183546304703,
+ "learning_rate": 0.0006,
+ "loss": 4.239282131195068,
+ "step": 2067
+ },
+ {
+ "epoch": 28.726637554585153,
+ "grad_norm": 0.0175587497651577,
+ "learning_rate": 0.0006,
+ "loss": 4.289038181304932,
+ "step": 2068
+ },
+ {
+ "epoch": 28.74061135371179,
+ "grad_norm": 0.018223397433757782,
+ "learning_rate": 0.0006,
+ "loss": 4.2266998291015625,
+ "step": 2069
+ },
+ {
+ "epoch": 28.754585152838427,
+ "grad_norm": 0.018680868670344353,
+ "learning_rate": 0.0006,
+ "loss": 4.348849296569824,
+ "step": 2070
+ },
+ {
+ "epoch": 28.768558951965066,
+ "grad_norm": 0.016726728528738022,
+ "learning_rate": 0.0006,
+ "loss": 4.292080879211426,
+ "step": 2071
+ },
+ {
+ "epoch": 28.782532751091704,
+ "grad_norm": 0.017386795952916145,
+ "learning_rate": 0.0006,
+ "loss": 4.208870887756348,
+ "step": 2072
+ },
+ {
+ "epoch": 28.79650655021834,
+ "grad_norm": 0.01865178905427456,
+ "learning_rate": 0.0006,
+ "loss": 4.054150581359863,
+ "step": 2073
+ },
+ {
+ "epoch": 28.810480349344978,
+ "grad_norm": 0.01776815392076969,
+ "learning_rate": 0.0006,
+ "loss": 4.202404499053955,
+ "step": 2074
+ },
+ {
+ "epoch": 28.824454148471617,
+ "grad_norm": 0.01643497310578823,
+ "learning_rate": 0.0006,
+ "loss": 4.131855010986328,
+ "step": 2075
+ },
+ {
+ "epoch": 28.83842794759825,
+ "grad_norm": 0.01574273779988289,
+ "learning_rate": 0.0006,
+ "loss": 4.296788692474365,
+ "step": 2076
+ },
+ {
+ "epoch": 28.85240174672489,
+ "grad_norm": 0.015613908879458904,
+ "learning_rate": 0.0006,
+ "loss": 4.210949897766113,
+ "step": 2077
+ },
+ {
+ "epoch": 28.86637554585153,
+ "grad_norm": 0.015543212182819843,
+ "learning_rate": 0.0006,
+ "loss": 4.218136310577393,
+ "step": 2078
+ },
+ {
+ "epoch": 28.880349344978168,
+ "grad_norm": 0.014681251719594002,
+ "learning_rate": 0.0006,
+ "loss": 4.19401741027832,
+ "step": 2079
+ },
+ {
+ "epoch": 28.894323144104803,
+ "grad_norm": 0.014393828809261322,
+ "learning_rate": 0.0006,
+ "loss": 4.198468208312988,
+ "step": 2080
+ },
+ {
+ "epoch": 28.90829694323144,
+ "grad_norm": 0.01580170728266239,
+ "learning_rate": 0.0006,
+ "loss": 4.224969863891602,
+ "step": 2081
+ },
+ {
+ "epoch": 28.92227074235808,
+ "grad_norm": 0.01484165620058775,
+ "learning_rate": 0.0006,
+ "loss": 4.12236213684082,
+ "step": 2082
+ },
+ {
+ "epoch": 28.936244541484715,
+ "grad_norm": 0.01497623696923256,
+ "learning_rate": 0.0006,
+ "loss": 4.24141788482666,
+ "step": 2083
+ },
+ {
+ "epoch": 28.950218340611354,
+ "grad_norm": 0.015080046840012074,
+ "learning_rate": 0.0006,
+ "loss": 4.284151077270508,
+ "step": 2084
+ },
+ {
+ "epoch": 28.964192139737992,
+ "grad_norm": 0.016207212582230568,
+ "learning_rate": 0.0006,
+ "loss": 4.298693656921387,
+ "step": 2085
+ },
+ {
+ "epoch": 28.978165938864628,
+ "grad_norm": 0.015474417246878147,
+ "learning_rate": 0.0006,
+ "loss": 4.2412261962890625,
+ "step": 2086
+ },
+ {
+ "epoch": 28.992139737991266,
+ "grad_norm": 0.01515351701527834,
+ "learning_rate": 0.0006,
+ "loss": 4.3054423332214355,
+ "step": 2087
+ },
+ {
+ "epoch": 29.0,
+ "grad_norm": 0.016411345452070236,
+ "learning_rate": 0.0006,
+ "loss": 4.104830741882324,
+ "step": 2088
+ },
+ {
+ "epoch": 29.0,
+ "eval_loss": 4.661567687988281,
+ "eval_runtime": 57.1244,
+ "eval_samples_per_second": 42.749,
+ "eval_steps_per_second": 1.348,
+ "step": 2088
+ },
+ {
+ "epoch": 29.01397379912664,
+ "grad_norm": 0.01665342226624489,
+ "learning_rate": 0.0006,
+ "loss": 4.311391830444336,
+ "step": 2089
+ },
+ {
+ "epoch": 29.027947598253274,
+ "grad_norm": 0.018053214997053146,
+ "learning_rate": 0.0006,
+ "loss": 4.135597229003906,
+ "step": 2090
+ },
+ {
+ "epoch": 29.041921397379912,
+ "grad_norm": 0.018213748931884766,
+ "learning_rate": 0.0006,
+ "loss": 4.303299903869629,
+ "step": 2091
+ },
+ {
+ "epoch": 29.05589519650655,
+ "grad_norm": 0.018460242077708244,
+ "learning_rate": 0.0006,
+ "loss": 4.20850944519043,
+ "step": 2092
+ },
+ {
+ "epoch": 29.069868995633186,
+ "grad_norm": 0.018365809693932533,
+ "learning_rate": 0.0006,
+ "loss": 4.275847434997559,
+ "step": 2093
+ },
+ {
+ "epoch": 29.083842794759825,
+ "grad_norm": 0.018346186727285385,
+ "learning_rate": 0.0006,
+ "loss": 4.202091217041016,
+ "step": 2094
+ },
+ {
+ "epoch": 29.097816593886463,
+ "grad_norm": 0.017920587211847305,
+ "learning_rate": 0.0006,
+ "loss": 4.182092666625977,
+ "step": 2095
+ },
+ {
+ "epoch": 29.111790393013102,
+ "grad_norm": 0.01812003180384636,
+ "learning_rate": 0.0006,
+ "loss": 4.259721755981445,
+ "step": 2096
+ },
+ {
+ "epoch": 29.125764192139737,
+ "grad_norm": 0.01700986735522747,
+ "learning_rate": 0.0006,
+ "loss": 4.067573070526123,
+ "step": 2097
+ },
+ {
+ "epoch": 29.139737991266376,
+ "grad_norm": 0.018069760873913765,
+ "learning_rate": 0.0006,
+ "loss": 4.234400749206543,
+ "step": 2098
+ },
+ {
+ "epoch": 29.153711790393015,
+ "grad_norm": 0.01905428245663643,
+ "learning_rate": 0.0006,
+ "loss": 4.16609001159668,
+ "step": 2099
+ },
+ {
+ "epoch": 29.16768558951965,
+ "grad_norm": 0.018711242824792862,
+ "learning_rate": 0.0006,
+ "loss": 4.249088764190674,
+ "step": 2100
+ },
+ {
+ "epoch": 29.18165938864629,
+ "grad_norm": 0.018721306696534157,
+ "learning_rate": 0.0006,
+ "loss": 4.14540958404541,
+ "step": 2101
+ },
+ {
+ "epoch": 29.195633187772927,
+ "grad_norm": 0.020139874890446663,
+ "learning_rate": 0.0006,
+ "loss": 4.049248695373535,
+ "step": 2102
+ },
+ {
+ "epoch": 29.209606986899562,
+ "grad_norm": 0.020849574357271194,
+ "learning_rate": 0.0006,
+ "loss": 4.235224723815918,
+ "step": 2103
+ },
+ {
+ "epoch": 29.2235807860262,
+ "grad_norm": 0.018693549558520317,
+ "learning_rate": 0.0006,
+ "loss": 4.232961177825928,
+ "step": 2104
+ },
+ {
+ "epoch": 29.23755458515284,
+ "grad_norm": 0.017889857292175293,
+ "learning_rate": 0.0006,
+ "loss": 4.171422004699707,
+ "step": 2105
+ },
+ {
+ "epoch": 29.251528384279474,
+ "grad_norm": 0.016690224409103394,
+ "learning_rate": 0.0006,
+ "loss": 4.244717597961426,
+ "step": 2106
+ },
+ {
+ "epoch": 29.265502183406113,
+ "grad_norm": 0.01708616502583027,
+ "learning_rate": 0.0006,
+ "loss": 4.125068664550781,
+ "step": 2107
+ },
+ {
+ "epoch": 29.27947598253275,
+ "grad_norm": 0.017928237095475197,
+ "learning_rate": 0.0006,
+ "loss": 4.208292007446289,
+ "step": 2108
+ },
+ {
+ "epoch": 29.29344978165939,
+ "grad_norm": 0.018250394612550735,
+ "learning_rate": 0.0006,
+ "loss": 4.118630886077881,
+ "step": 2109
+ },
+ {
+ "epoch": 29.307423580786025,
+ "grad_norm": 0.020160246640443802,
+ "learning_rate": 0.0006,
+ "loss": 4.206025123596191,
+ "step": 2110
+ },
+ {
+ "epoch": 29.321397379912664,
+ "grad_norm": 0.020045453682541847,
+ "learning_rate": 0.0006,
+ "loss": 4.2758684158325195,
+ "step": 2111
+ },
+ {
+ "epoch": 29.335371179039303,
+ "grad_norm": 0.019556907936930656,
+ "learning_rate": 0.0006,
+ "loss": 4.188272953033447,
+ "step": 2112
+ },
+ {
+ "epoch": 29.349344978165938,
+ "grad_norm": 0.019185200333595276,
+ "learning_rate": 0.0006,
+ "loss": 4.17451286315918,
+ "step": 2113
+ },
+ {
+ "epoch": 29.363318777292577,
+ "grad_norm": 0.01816629432141781,
+ "learning_rate": 0.0006,
+ "loss": 4.165246963500977,
+ "step": 2114
+ },
+ {
+ "epoch": 29.377292576419215,
+ "grad_norm": 0.01865146867930889,
+ "learning_rate": 0.0006,
+ "loss": 4.148184776306152,
+ "step": 2115
+ },
+ {
+ "epoch": 29.39126637554585,
+ "grad_norm": 0.018338464200496674,
+ "learning_rate": 0.0006,
+ "loss": 4.288936614990234,
+ "step": 2116
+ },
+ {
+ "epoch": 29.40524017467249,
+ "grad_norm": 0.018986187875270844,
+ "learning_rate": 0.0006,
+ "loss": 4.183379650115967,
+ "step": 2117
+ },
+ {
+ "epoch": 29.419213973799128,
+ "grad_norm": 0.018409637734293938,
+ "learning_rate": 0.0006,
+ "loss": 4.057574272155762,
+ "step": 2118
+ },
+ {
+ "epoch": 29.433187772925763,
+ "grad_norm": 0.016664542257785797,
+ "learning_rate": 0.0006,
+ "loss": 4.18517541885376,
+ "step": 2119
+ },
+ {
+ "epoch": 29.4471615720524,
+ "grad_norm": 0.016486089676618576,
+ "learning_rate": 0.0006,
+ "loss": 4.179367542266846,
+ "step": 2120
+ },
+ {
+ "epoch": 29.46113537117904,
+ "grad_norm": 0.01787523552775383,
+ "learning_rate": 0.0006,
+ "loss": 4.184126377105713,
+ "step": 2121
+ },
+ {
+ "epoch": 29.475109170305675,
+ "grad_norm": 0.016574613749980927,
+ "learning_rate": 0.0006,
+ "loss": 4.290170192718506,
+ "step": 2122
+ },
+ {
+ "epoch": 29.489082969432314,
+ "grad_norm": 0.01643376611173153,
+ "learning_rate": 0.0006,
+ "loss": 4.052913665771484,
+ "step": 2123
+ },
+ {
+ "epoch": 29.503056768558952,
+ "grad_norm": 0.016503969207406044,
+ "learning_rate": 0.0006,
+ "loss": 4.335184097290039,
+ "step": 2124
+ },
+ {
+ "epoch": 29.51703056768559,
+ "grad_norm": 0.017596479505300522,
+ "learning_rate": 0.0006,
+ "loss": 4.165539741516113,
+ "step": 2125
+ },
+ {
+ "epoch": 29.531004366812226,
+ "grad_norm": 0.017468582838773727,
+ "learning_rate": 0.0006,
+ "loss": 4.315964698791504,
+ "step": 2126
+ },
+ {
+ "epoch": 29.544978165938865,
+ "grad_norm": 0.01724618673324585,
+ "learning_rate": 0.0006,
+ "loss": 4.243681907653809,
+ "step": 2127
+ },
+ {
+ "epoch": 29.558951965065503,
+ "grad_norm": 0.018384616822004318,
+ "learning_rate": 0.0006,
+ "loss": 4.112447261810303,
+ "step": 2128
+ },
+ {
+ "epoch": 29.57292576419214,
+ "grad_norm": 0.01824074238538742,
+ "learning_rate": 0.0006,
+ "loss": 4.236065864562988,
+ "step": 2129
+ },
+ {
+ "epoch": 29.586899563318777,
+ "grad_norm": 0.01625971868634224,
+ "learning_rate": 0.0006,
+ "loss": 4.18143892288208,
+ "step": 2130
+ },
+ {
+ "epoch": 29.600873362445416,
+ "grad_norm": 0.015537494793534279,
+ "learning_rate": 0.0006,
+ "loss": 4.225147247314453,
+ "step": 2131
+ },
+ {
+ "epoch": 29.61484716157205,
+ "grad_norm": 0.015818974003195763,
+ "learning_rate": 0.0006,
+ "loss": 4.3100175857543945,
+ "step": 2132
+ },
+ {
+ "epoch": 29.62882096069869,
+ "grad_norm": 0.015909671783447266,
+ "learning_rate": 0.0006,
+ "loss": 4.169775009155273,
+ "step": 2133
+ },
+ {
+ "epoch": 29.64279475982533,
+ "grad_norm": 0.01769658550620079,
+ "learning_rate": 0.0006,
+ "loss": 4.278904914855957,
+ "step": 2134
+ },
+ {
+ "epoch": 29.656768558951963,
+ "grad_norm": 0.015641704201698303,
+ "learning_rate": 0.0006,
+ "loss": 4.2426228523254395,
+ "step": 2135
+ },
+ {
+ "epoch": 29.670742358078602,
+ "grad_norm": 0.01565658301115036,
+ "learning_rate": 0.0006,
+ "loss": 4.232687950134277,
+ "step": 2136
+ },
+ {
+ "epoch": 29.68471615720524,
+ "grad_norm": 0.016485940665006638,
+ "learning_rate": 0.0006,
+ "loss": 4.209839820861816,
+ "step": 2137
+ },
+ {
+ "epoch": 29.69868995633188,
+ "grad_norm": 0.017622729763388634,
+ "learning_rate": 0.0006,
+ "loss": 4.196017265319824,
+ "step": 2138
+ },
+ {
+ "epoch": 29.712663755458514,
+ "grad_norm": 0.018320564180612564,
+ "learning_rate": 0.0006,
+ "loss": 4.225683689117432,
+ "step": 2139
+ },
+ {
+ "epoch": 29.726637554585153,
+ "grad_norm": 0.018140794709324837,
+ "learning_rate": 0.0006,
+ "loss": 4.311944961547852,
+ "step": 2140
+ },
+ {
+ "epoch": 29.74061135371179,
+ "grad_norm": 0.016227100044488907,
+ "learning_rate": 0.0006,
+ "loss": 4.230257511138916,
+ "step": 2141
+ },
+ {
+ "epoch": 29.754585152838427,
+ "grad_norm": 0.015759488567709923,
+ "learning_rate": 0.0006,
+ "loss": 4.119174003601074,
+ "step": 2142
+ },
+ {
+ "epoch": 29.768558951965066,
+ "grad_norm": 0.01757766678929329,
+ "learning_rate": 0.0006,
+ "loss": 4.165826797485352,
+ "step": 2143
+ },
+ {
+ "epoch": 29.782532751091704,
+ "grad_norm": 0.019661923870444298,
+ "learning_rate": 0.0006,
+ "loss": 4.392203330993652,
+ "step": 2144
+ },
+ {
+ "epoch": 29.79650655021834,
+ "grad_norm": 0.019478755071759224,
+ "learning_rate": 0.0006,
+ "loss": 4.348371982574463,
+ "step": 2145
+ },
+ {
+ "epoch": 29.810480349344978,
+ "grad_norm": 0.019149569794535637,
+ "learning_rate": 0.0006,
+ "loss": 4.34104061126709,
+ "step": 2146
+ },
+ {
+ "epoch": 29.824454148471617,
+ "grad_norm": 0.017707521095871925,
+ "learning_rate": 0.0006,
+ "loss": 4.228226661682129,
+ "step": 2147
+ },
+ {
+ "epoch": 29.83842794759825,
+ "grad_norm": 0.015615378506481647,
+ "learning_rate": 0.0006,
+ "loss": 4.175541877746582,
+ "step": 2148
+ },
+ {
+ "epoch": 29.85240174672489,
+ "grad_norm": 0.016111129894852638,
+ "learning_rate": 0.0006,
+ "loss": 4.267936706542969,
+ "step": 2149
+ },
+ {
+ "epoch": 29.86637554585153,
+ "grad_norm": 0.01528779324144125,
+ "learning_rate": 0.0006,
+ "loss": 4.186014175415039,
+ "step": 2150
+ },
+ {
+ "epoch": 29.880349344978168,
+ "grad_norm": 0.014925646595656872,
+ "learning_rate": 0.0006,
+ "loss": 4.174860000610352,
+ "step": 2151
+ },
+ {
+ "epoch": 29.894323144104803,
+ "grad_norm": 0.016917673870921135,
+ "learning_rate": 0.0006,
+ "loss": 4.33908748626709,
+ "step": 2152
+ },
+ {
+ "epoch": 29.90829694323144,
+ "grad_norm": 0.016007075086236,
+ "learning_rate": 0.0006,
+ "loss": 4.266141414642334,
+ "step": 2153
+ },
+ {
+ "epoch": 29.92227074235808,
+ "grad_norm": 0.016178956255316734,
+ "learning_rate": 0.0006,
+ "loss": 4.271215915679932,
+ "step": 2154
+ },
+ {
+ "epoch": 29.936244541484715,
+ "grad_norm": 0.016180871054530144,
+ "learning_rate": 0.0006,
+ "loss": 4.278801441192627,
+ "step": 2155
+ },
+ {
+ "epoch": 29.950218340611354,
+ "grad_norm": 0.015335503034293652,
+ "learning_rate": 0.0006,
+ "loss": 4.194978713989258,
+ "step": 2156
+ },
+ {
+ "epoch": 29.964192139737992,
+ "grad_norm": 0.016580764204263687,
+ "learning_rate": 0.0006,
+ "loss": 4.290809631347656,
+ "step": 2157
+ },
+ {
+ "epoch": 29.978165938864628,
+ "grad_norm": 0.01623128168284893,
+ "learning_rate": 0.0006,
+ "loss": 4.205451488494873,
+ "step": 2158
+ },
+ {
+ "epoch": 29.992139737991266,
+ "grad_norm": 0.014904248528182507,
+ "learning_rate": 0.0006,
+ "loss": 4.24924373626709,
+ "step": 2159
+ },
+ {
+ "epoch": 30.0,
+ "grad_norm": 0.017089299857616425,
+ "learning_rate": 0.0006,
+ "loss": 4.222072601318359,
+ "step": 2160
+ },
+ {
+ "epoch": 30.0,
+ "eval_loss": 4.585411071777344,
+ "eval_runtime": 57.1164,
+ "eval_samples_per_second": 42.755,
+ "eval_steps_per_second": 1.348,
+ "step": 2160
+ },
+ {
+ "epoch": 30.01397379912664,
+ "grad_norm": 0.017226383090019226,
+ "learning_rate": 0.0006,
+ "loss": 4.19467830657959,
+ "step": 2161
+ },
+ {
+ "epoch": 30.027947598253274,
+ "grad_norm": 0.01635066606104374,
+ "learning_rate": 0.0006,
+ "loss": 4.176875591278076,
+ "step": 2162
+ },
+ {
+ "epoch": 30.041921397379912,
+ "grad_norm": 0.01735362410545349,
+ "learning_rate": 0.0006,
+ "loss": 4.13667106628418,
+ "step": 2163
+ },
+ {
+ "epoch": 30.05589519650655,
+ "grad_norm": 0.017047109082341194,
+ "learning_rate": 0.0006,
+ "loss": 4.136329650878906,
+ "step": 2164
+ },
+ {
+ "epoch": 30.069868995633186,
+ "grad_norm": 0.018116053193807602,
+ "learning_rate": 0.0006,
+ "loss": 4.228706359863281,
+ "step": 2165
+ },
+ {
+ "epoch": 30.083842794759825,
+ "grad_norm": 0.01791389286518097,
+ "learning_rate": 0.0006,
+ "loss": 4.1597747802734375,
+ "step": 2166
+ },
+ {
+ "epoch": 30.097816593886463,
+ "grad_norm": 0.017793502658605576,
+ "learning_rate": 0.0006,
+ "loss": 4.0807976722717285,
+ "step": 2167
+ },
+ {
+ "epoch": 30.111790393013102,
+ "grad_norm": 0.01827503927052021,
+ "learning_rate": 0.0006,
+ "loss": 4.170646667480469,
+ "step": 2168
+ },
+ {
+ "epoch": 30.125764192139737,
+ "grad_norm": 0.017228564247488976,
+ "learning_rate": 0.0006,
+ "loss": 4.146934509277344,
+ "step": 2169
+ },
+ {
+ "epoch": 30.139737991266376,
+ "grad_norm": 0.01757437363266945,
+ "learning_rate": 0.0006,
+ "loss": 4.1859130859375,
+ "step": 2170
+ },
+ {
+ "epoch": 30.153711790393015,
+ "grad_norm": 0.01701093092560768,
+ "learning_rate": 0.0006,
+ "loss": 4.126981258392334,
+ "step": 2171
+ },
+ {
+ "epoch": 30.16768558951965,
+ "grad_norm": 0.015220489352941513,
+ "learning_rate": 0.0006,
+ "loss": 4.164018630981445,
+ "step": 2172
+ },
+ {
+ "epoch": 30.18165938864629,
+ "grad_norm": 0.016254238784313202,
+ "learning_rate": 0.0006,
+ "loss": 4.258275508880615,
+ "step": 2173
+ },
+ {
+ "epoch": 30.195633187772927,
+ "grad_norm": 0.016609614714980125,
+ "learning_rate": 0.0006,
+ "loss": 4.222233772277832,
+ "step": 2174
+ },
+ {
+ "epoch": 30.209606986899562,
+ "grad_norm": 0.016531318426132202,
+ "learning_rate": 0.0006,
+ "loss": 4.144489288330078,
+ "step": 2175
+ },
+ {
+ "epoch": 30.2235807860262,
+ "grad_norm": 0.01567668654024601,
+ "learning_rate": 0.0006,
+ "loss": 4.181916236877441,
+ "step": 2176
+ },
+ {
+ "epoch": 30.23755458515284,
+ "grad_norm": 0.017195681110024452,
+ "learning_rate": 0.0006,
+ "loss": 4.189693927764893,
+ "step": 2177
+ },
+ {
+ "epoch": 30.251528384279474,
+ "grad_norm": 0.016882948577404022,
+ "learning_rate": 0.0006,
+ "loss": 4.21604061126709,
+ "step": 2178
+ },
+ {
+ "epoch": 30.265502183406113,
+ "grad_norm": 0.016261622309684753,
+ "learning_rate": 0.0006,
+ "loss": 4.227639198303223,
+ "step": 2179
+ },
+ {
+ "epoch": 30.27947598253275,
+ "grad_norm": 0.016411812976002693,
+ "learning_rate": 0.0006,
+ "loss": 4.130362510681152,
+ "step": 2180
+ },
+ {
+ "epoch": 30.29344978165939,
+ "grad_norm": 0.01732019893825054,
+ "learning_rate": 0.0006,
+ "loss": 4.230443000793457,
+ "step": 2181
+ },
+ {
+ "epoch": 30.307423580786025,
+ "grad_norm": 0.01825905591249466,
+ "learning_rate": 0.0006,
+ "loss": 4.175817489624023,
+ "step": 2182
+ },
+ {
+ "epoch": 30.321397379912664,
+ "grad_norm": 0.01887713558971882,
+ "learning_rate": 0.0006,
+ "loss": 4.169020652770996,
+ "step": 2183
+ },
+ {
+ "epoch": 30.335371179039303,
+ "grad_norm": 0.019391661509871483,
+ "learning_rate": 0.0006,
+ "loss": 4.176318645477295,
+ "step": 2184
+ },
+ {
+ "epoch": 30.349344978165938,
+ "grad_norm": 0.020840002223849297,
+ "learning_rate": 0.0006,
+ "loss": 4.174228668212891,
+ "step": 2185
+ },
+ {
+ "epoch": 30.363318777292577,
+ "grad_norm": 0.020193178206682205,
+ "learning_rate": 0.0006,
+ "loss": 4.127919673919678,
+ "step": 2186
+ },
+ {
+ "epoch": 30.377292576419215,
+ "grad_norm": 0.018449561670422554,
+ "learning_rate": 0.0006,
+ "loss": 4.229487895965576,
+ "step": 2187
+ },
+ {
+ "epoch": 30.39126637554585,
+ "grad_norm": 0.01711983233690262,
+ "learning_rate": 0.0006,
+ "loss": 4.201651573181152,
+ "step": 2188
+ },
+ {
+ "epoch": 30.40524017467249,
+ "grad_norm": 0.017293395474553108,
+ "learning_rate": 0.0006,
+ "loss": 4.168704032897949,
+ "step": 2189
+ },
+ {
+ "epoch": 30.419213973799128,
+ "grad_norm": 0.018526358529925346,
+ "learning_rate": 0.0006,
+ "loss": 4.171864032745361,
+ "step": 2190
+ },
+ {
+ "epoch": 30.433187772925763,
+ "grad_norm": 0.019287291914224625,
+ "learning_rate": 0.0006,
+ "loss": 4.039534568786621,
+ "step": 2191
+ },
+ {
+ "epoch": 30.4471615720524,
+ "grad_norm": 0.016902444884181023,
+ "learning_rate": 0.0006,
+ "loss": 4.282787799835205,
+ "step": 2192
+ },
+ {
+ "epoch": 30.46113537117904,
+ "grad_norm": 0.01656750962138176,
+ "learning_rate": 0.0006,
+ "loss": 4.213098526000977,
+ "step": 2193
+ },
+ {
+ "epoch": 30.475109170305675,
+ "grad_norm": 0.016805065795779228,
+ "learning_rate": 0.0006,
+ "loss": 4.1583476066589355,
+ "step": 2194
+ },
+ {
+ "epoch": 30.489082969432314,
+ "grad_norm": 0.017166638746857643,
+ "learning_rate": 0.0006,
+ "loss": 4.188634395599365,
+ "step": 2195
+ },
+ {
+ "epoch": 30.503056768558952,
+ "grad_norm": 0.01592106744647026,
+ "learning_rate": 0.0006,
+ "loss": 4.234043121337891,
+ "step": 2196
+ },
+ {
+ "epoch": 30.51703056768559,
+ "grad_norm": 0.015270989388227463,
+ "learning_rate": 0.0006,
+ "loss": 4.301420211791992,
+ "step": 2197
+ },
+ {
+ "epoch": 30.531004366812226,
+ "grad_norm": 0.015194724313914776,
+ "learning_rate": 0.0006,
+ "loss": 4.18703031539917,
+ "step": 2198
+ },
+ {
+ "epoch": 30.544978165938865,
+ "grad_norm": 0.015587746165692806,
+ "learning_rate": 0.0006,
+ "loss": 4.138928413391113,
+ "step": 2199
+ },
+ {
+ "epoch": 30.558951965065503,
+ "grad_norm": 0.016029570251703262,
+ "learning_rate": 0.0006,
+ "loss": 4.290863990783691,
+ "step": 2200
+ },
+ {
+ "epoch": 30.57292576419214,
+ "grad_norm": 0.015651309862732887,
+ "learning_rate": 0.0006,
+ "loss": 4.272062301635742,
+ "step": 2201
+ },
+ {
+ "epoch": 30.586899563318777,
+ "grad_norm": 0.015992306172847748,
+ "learning_rate": 0.0006,
+ "loss": 4.2487406730651855,
+ "step": 2202
+ },
+ {
+ "epoch": 30.600873362445416,
+ "grad_norm": 0.014969157986342907,
+ "learning_rate": 0.0006,
+ "loss": 4.115347862243652,
+ "step": 2203
+ },
+ {
+ "epoch": 30.61484716157205,
+ "grad_norm": 0.014639027416706085,
+ "learning_rate": 0.0006,
+ "loss": 4.2207489013671875,
+ "step": 2204
+ },
+ {
+ "epoch": 30.62882096069869,
+ "grad_norm": 0.015059413388371468,
+ "learning_rate": 0.0006,
+ "loss": 4.274044990539551,
+ "step": 2205
+ },
+ {
+ "epoch": 30.64279475982533,
+ "grad_norm": 0.0166641678661108,
+ "learning_rate": 0.0006,
+ "loss": 4.191373348236084,
+ "step": 2206
+ },
+ {
+ "epoch": 30.656768558951963,
+ "grad_norm": 0.017806345596909523,
+ "learning_rate": 0.0006,
+ "loss": 4.139800548553467,
+ "step": 2207
+ },
+ {
+ "epoch": 30.670742358078602,
+ "grad_norm": 0.018712421879172325,
+ "learning_rate": 0.0006,
+ "loss": 4.035965442657471,
+ "step": 2208
+ },
+ {
+ "epoch": 30.68471615720524,
+ "grad_norm": 0.018457243219017982,
+ "learning_rate": 0.0006,
+ "loss": 4.2579240798950195,
+ "step": 2209
+ },
+ {
+ "epoch": 30.69868995633188,
+ "grad_norm": 0.01920999586582184,
+ "learning_rate": 0.0006,
+ "loss": 4.131112098693848,
+ "step": 2210
+ },
+ {
+ "epoch": 30.712663755458514,
+ "grad_norm": 0.01729944534599781,
+ "learning_rate": 0.0006,
+ "loss": 4.254059314727783,
+ "step": 2211
+ },
+ {
+ "epoch": 30.726637554585153,
+ "grad_norm": 0.016914231702685356,
+ "learning_rate": 0.0006,
+ "loss": 4.233428955078125,
+ "step": 2212
+ },
+ {
+ "epoch": 30.74061135371179,
+ "grad_norm": 0.01913098618388176,
+ "learning_rate": 0.0006,
+ "loss": 4.23525333404541,
+ "step": 2213
+ },
+ {
+ "epoch": 30.754585152838427,
+ "grad_norm": 0.019150281324982643,
+ "learning_rate": 0.0006,
+ "loss": 4.213695049285889,
+ "step": 2214
+ },
+ {
+ "epoch": 30.768558951965066,
+ "grad_norm": 0.018947452306747437,
+ "learning_rate": 0.0006,
+ "loss": 4.151697158813477,
+ "step": 2215
+ },
+ {
+ "epoch": 30.782532751091704,
+ "grad_norm": 0.016058508306741714,
+ "learning_rate": 0.0006,
+ "loss": 4.220280170440674,
+ "step": 2216
+ },
+ {
+ "epoch": 30.79650655021834,
+ "grad_norm": 0.017488619312644005,
+ "learning_rate": 0.0006,
+ "loss": 4.154428482055664,
+ "step": 2217
+ },
+ {
+ "epoch": 30.810480349344978,
+ "grad_norm": 0.017355090007185936,
+ "learning_rate": 0.0006,
+ "loss": 4.256086349487305,
+ "step": 2218
+ },
+ {
+ "epoch": 30.824454148471617,
+ "grad_norm": 0.016636032611131668,
+ "learning_rate": 0.0006,
+ "loss": 4.227564334869385,
+ "step": 2219
+ },
+ {
+ "epoch": 30.83842794759825,
+ "grad_norm": 0.016273891553282738,
+ "learning_rate": 0.0006,
+ "loss": 4.200466156005859,
+ "step": 2220
+ },
+ {
+ "epoch": 30.85240174672489,
+ "grad_norm": 0.016939911991357803,
+ "learning_rate": 0.0006,
+ "loss": 4.1571221351623535,
+ "step": 2221
+ },
+ {
+ "epoch": 30.86637554585153,
+ "grad_norm": 0.018332522362470627,
+ "learning_rate": 0.0006,
+ "loss": 4.228398323059082,
+ "step": 2222
+ },
+ {
+ "epoch": 30.880349344978168,
+ "grad_norm": 0.019508114084601402,
+ "learning_rate": 0.0006,
+ "loss": 4.222960472106934,
+ "step": 2223
+ },
+ {
+ "epoch": 30.894323144104803,
+ "grad_norm": 0.018440278246998787,
+ "learning_rate": 0.0006,
+ "loss": 4.261051177978516,
+ "step": 2224
+ },
+ {
+ "epoch": 30.90829694323144,
+ "grad_norm": 0.018113229423761368,
+ "learning_rate": 0.0006,
+ "loss": 4.109235763549805,
+ "step": 2225
+ },
+ {
+ "epoch": 30.92227074235808,
+ "grad_norm": 0.015392083674669266,
+ "learning_rate": 0.0006,
+ "loss": 4.21950101852417,
+ "step": 2226
+ },
+ {
+ "epoch": 30.936244541484715,
+ "grad_norm": 0.01633290946483612,
+ "learning_rate": 0.0006,
+ "loss": 4.176053047180176,
+ "step": 2227
+ },
+ {
+ "epoch": 30.950218340611354,
+ "grad_norm": 0.016236383467912674,
+ "learning_rate": 0.0006,
+ "loss": 4.314309120178223,
+ "step": 2228
+ },
+ {
+ "epoch": 30.964192139737992,
+ "grad_norm": 0.015284373424947262,
+ "learning_rate": 0.0006,
+ "loss": 4.303133487701416,
+ "step": 2229
+ },
+ {
+ "epoch": 30.978165938864628,
+ "grad_norm": 0.015982897952198982,
+ "learning_rate": 0.0006,
+ "loss": 4.25739860534668,
+ "step": 2230
+ },
+ {
+ "epoch": 30.992139737991266,
+ "grad_norm": 0.016682198271155357,
+ "learning_rate": 0.0006,
+ "loss": 4.292210578918457,
+ "step": 2231
+ },
+ {
+ "epoch": 31.0,
+ "grad_norm": 0.017521467059850693,
+ "learning_rate": 0.0006,
+ "loss": 4.170637130737305,
+ "step": 2232
+ },
+ {
+ "epoch": 31.0,
+ "eval_loss": 4.642786979675293,
+ "eval_runtime": 57.029,
+ "eval_samples_per_second": 42.82,
+ "eval_steps_per_second": 1.35,
+ "step": 2232
+ },
+ {
+ "epoch": 31.01397379912664,
+ "grad_norm": 0.016607852652668953,
+ "learning_rate": 0.0006,
+ "loss": 4.106224536895752,
+ "step": 2233
+ },
+ {
+ "epoch": 31.027947598253274,
+ "grad_norm": 0.018448904156684875,
+ "learning_rate": 0.0006,
+ "loss": 4.224459648132324,
+ "step": 2234
+ },
+ {
+ "epoch": 31.041921397379912,
+ "grad_norm": 0.020481228828430176,
+ "learning_rate": 0.0006,
+ "loss": 4.305548667907715,
+ "step": 2235
+ },
+ {
+ "epoch": 31.05589519650655,
+ "grad_norm": 0.01999707892537117,
+ "learning_rate": 0.0006,
+ "loss": 4.253663539886475,
+ "step": 2236
+ },
+ {
+ "epoch": 31.069868995633186,
+ "grad_norm": 0.020428303629159927,
+ "learning_rate": 0.0006,
+ "loss": 4.166954040527344,
+ "step": 2237
+ },
+ {
+ "epoch": 31.083842794759825,
+ "grad_norm": 0.022608213126659393,
+ "learning_rate": 0.0006,
+ "loss": 4.172819137573242,
+ "step": 2238
+ },
+ {
+ "epoch": 31.097816593886463,
+ "grad_norm": 0.024669520556926727,
+ "learning_rate": 0.0006,
+ "loss": 4.106586456298828,
+ "step": 2239
+ },
+ {
+ "epoch": 31.111790393013102,
+ "grad_norm": 0.02247351035475731,
+ "learning_rate": 0.0006,
+ "loss": 4.160271644592285,
+ "step": 2240
+ },
+ {
+ "epoch": 31.125764192139737,
+ "grad_norm": 0.0207810141146183,
+ "learning_rate": 0.0006,
+ "loss": 4.272491931915283,
+ "step": 2241
+ },
+ {
+ "epoch": 31.139737991266376,
+ "grad_norm": 0.021576499566435814,
+ "learning_rate": 0.0006,
+ "loss": 4.171474456787109,
+ "step": 2242
+ },
+ {
+ "epoch": 31.153711790393015,
+ "grad_norm": 0.02337106689810753,
+ "learning_rate": 0.0006,
+ "loss": 4.094712734222412,
+ "step": 2243
+ },
+ {
+ "epoch": 31.16768558951965,
+ "grad_norm": 0.02219875156879425,
+ "learning_rate": 0.0006,
+ "loss": 4.1800994873046875,
+ "step": 2244
+ },
+ {
+ "epoch": 31.18165938864629,
+ "grad_norm": 0.022680504247546196,
+ "learning_rate": 0.0006,
+ "loss": 4.2054362297058105,
+ "step": 2245
+ },
+ {
+ "epoch": 31.195633187772927,
+ "grad_norm": 0.0208734143525362,
+ "learning_rate": 0.0006,
+ "loss": 4.06068229675293,
+ "step": 2246
+ },
+ {
+ "epoch": 31.209606986899562,
+ "grad_norm": 0.01811066083610058,
+ "learning_rate": 0.0006,
+ "loss": 4.2465128898620605,
+ "step": 2247
+ },
+ {
+ "epoch": 31.2235807860262,
+ "grad_norm": 0.018571637570858,
+ "learning_rate": 0.0006,
+ "loss": 4.191955089569092,
+ "step": 2248
+ },
+ {
+ "epoch": 31.23755458515284,
+ "grad_norm": 0.018598807975649834,
+ "learning_rate": 0.0006,
+ "loss": 4.200872421264648,
+ "step": 2249
+ },
+ {
+ "epoch": 31.251528384279474,
+ "grad_norm": 0.017577625811100006,
+ "learning_rate": 0.0006,
+ "loss": 4.148700714111328,
+ "step": 2250
+ },
+ {
+ "epoch": 31.265502183406113,
+ "grad_norm": 0.01817980408668518,
+ "learning_rate": 0.0006,
+ "loss": 4.256978988647461,
+ "step": 2251
+ },
+ {
+ "epoch": 31.27947598253275,
+ "grad_norm": 0.017983486875891685,
+ "learning_rate": 0.0006,
+ "loss": 4.175005912780762,
+ "step": 2252
+ },
+ {
+ "epoch": 31.29344978165939,
+ "grad_norm": 0.016751961782574654,
+ "learning_rate": 0.0006,
+ "loss": 4.237518310546875,
+ "step": 2253
+ },
+ {
+ "epoch": 31.307423580786025,
+ "grad_norm": 0.016407202929258347,
+ "learning_rate": 0.0006,
+ "loss": 4.197269439697266,
+ "step": 2254
+ },
+ {
+ "epoch": 31.321397379912664,
+ "grad_norm": 0.01578645594418049,
+ "learning_rate": 0.0006,
+ "loss": 4.179751396179199,
+ "step": 2255
+ },
+ {
+ "epoch": 31.335371179039303,
+ "grad_norm": 0.016001226380467415,
+ "learning_rate": 0.0006,
+ "loss": 4.08821964263916,
+ "step": 2256
+ },
+ {
+ "epoch": 31.349344978165938,
+ "grad_norm": 0.015122811309993267,
+ "learning_rate": 0.0006,
+ "loss": 4.047828674316406,
+ "step": 2257
+ },
+ {
+ "epoch": 31.363318777292577,
+ "grad_norm": 0.015399965457618237,
+ "learning_rate": 0.0006,
+ "loss": 4.1506524085998535,
+ "step": 2258
+ },
+ {
+ "epoch": 31.377292576419215,
+ "grad_norm": 0.01534605398774147,
+ "learning_rate": 0.0006,
+ "loss": 4.27952766418457,
+ "step": 2259
+ },
+ {
+ "epoch": 31.39126637554585,
+ "grad_norm": 0.015461381524801254,
+ "learning_rate": 0.0006,
+ "loss": 4.27102518081665,
+ "step": 2260
+ },
+ {
+ "epoch": 31.40524017467249,
+ "grad_norm": 0.016769153997302055,
+ "learning_rate": 0.0006,
+ "loss": 4.191592216491699,
+ "step": 2261
+ },
+ {
+ "epoch": 31.419213973799128,
+ "grad_norm": 0.016557959839701653,
+ "learning_rate": 0.0006,
+ "loss": 4.159705638885498,
+ "step": 2262
+ },
+ {
+ "epoch": 31.433187772925763,
+ "grad_norm": 0.01659569889307022,
+ "learning_rate": 0.0006,
+ "loss": 4.2136664390563965,
+ "step": 2263
+ },
+ {
+ "epoch": 31.4471615720524,
+ "grad_norm": 0.017101937904953957,
+ "learning_rate": 0.0006,
+ "loss": 4.2716498374938965,
+ "step": 2264
+ },
+ {
+ "epoch": 31.46113537117904,
+ "grad_norm": 0.01722985878586769,
+ "learning_rate": 0.0006,
+ "loss": 4.300534248352051,
+ "step": 2265
+ },
+ {
+ "epoch": 31.475109170305675,
+ "grad_norm": 0.016800185665488243,
+ "learning_rate": 0.0006,
+ "loss": 4.113240718841553,
+ "step": 2266
+ },
+ {
+ "epoch": 31.489082969432314,
+ "grad_norm": 0.01603875495493412,
+ "learning_rate": 0.0006,
+ "loss": 4.157105445861816,
+ "step": 2267
+ },
+ {
+ "epoch": 31.503056768558952,
+ "grad_norm": 0.015903932973742485,
+ "learning_rate": 0.0006,
+ "loss": 4.233669757843018,
+ "step": 2268
+ },
+ {
+ "epoch": 31.51703056768559,
+ "grad_norm": 0.0160983856767416,
+ "learning_rate": 0.0006,
+ "loss": 4.147704124450684,
+ "step": 2269
+ },
+ {
+ "epoch": 31.531004366812226,
+ "grad_norm": 0.01604730263352394,
+ "learning_rate": 0.0006,
+ "loss": 4.1465911865234375,
+ "step": 2270
+ },
+ {
+ "epoch": 31.544978165938865,
+ "grad_norm": 0.016205785796046257,
+ "learning_rate": 0.0006,
+ "loss": 4.106704235076904,
+ "step": 2271
+ },
+ {
+ "epoch": 31.558951965065503,
+ "grad_norm": 0.017008036375045776,
+ "learning_rate": 0.0006,
+ "loss": 4.191000938415527,
+ "step": 2272
+ },
+ {
+ "epoch": 31.57292576419214,
+ "grad_norm": 0.01699635200202465,
+ "learning_rate": 0.0006,
+ "loss": 4.187042236328125,
+ "step": 2273
+ },
+ {
+ "epoch": 31.586899563318777,
+ "grad_norm": 0.01692967116832733,
+ "learning_rate": 0.0006,
+ "loss": 4.208805561065674,
+ "step": 2274
+ },
+ {
+ "epoch": 31.600873362445416,
+ "grad_norm": 0.017608756199479103,
+ "learning_rate": 0.0006,
+ "loss": 4.200512886047363,
+ "step": 2275
+ },
+ {
+ "epoch": 31.61484716157205,
+ "grad_norm": 0.016590990126132965,
+ "learning_rate": 0.0006,
+ "loss": 4.194705009460449,
+ "step": 2276
+ },
+ {
+ "epoch": 31.62882096069869,
+ "grad_norm": 0.0170235987752676,
+ "learning_rate": 0.0006,
+ "loss": 4.210326194763184,
+ "step": 2277
+ },
+ {
+ "epoch": 31.64279475982533,
+ "grad_norm": 0.01604801043868065,
+ "learning_rate": 0.0006,
+ "loss": 4.196024417877197,
+ "step": 2278
+ },
+ {
+ "epoch": 31.656768558951963,
+ "grad_norm": 0.015694202855229378,
+ "learning_rate": 0.0006,
+ "loss": 4.178236484527588,
+ "step": 2279
+ },
+ {
+ "epoch": 31.670742358078602,
+ "grad_norm": 0.015473959036171436,
+ "learning_rate": 0.0006,
+ "loss": 4.259339809417725,
+ "step": 2280
+ },
+ {
+ "epoch": 31.68471615720524,
+ "grad_norm": 0.014689362607896328,
+ "learning_rate": 0.0006,
+ "loss": 3.9702324867248535,
+ "step": 2281
+ },
+ {
+ "epoch": 31.69868995633188,
+ "grad_norm": 0.015245123766362667,
+ "learning_rate": 0.0006,
+ "loss": 4.197025775909424,
+ "step": 2282
+ },
+ {
+ "epoch": 31.712663755458514,
+ "grad_norm": 0.016148347407579422,
+ "learning_rate": 0.0006,
+ "loss": 4.231856346130371,
+ "step": 2283
+ },
+ {
+ "epoch": 31.726637554585153,
+ "grad_norm": 0.015169818885624409,
+ "learning_rate": 0.0006,
+ "loss": 4.230169296264648,
+ "step": 2284
+ },
+ {
+ "epoch": 31.74061135371179,
+ "grad_norm": 0.014843013137578964,
+ "learning_rate": 0.0006,
+ "loss": 4.195633411407471,
+ "step": 2285
+ },
+ {
+ "epoch": 31.754585152838427,
+ "grad_norm": 0.015229340642690659,
+ "learning_rate": 0.0006,
+ "loss": 4.231280326843262,
+ "step": 2286
+ },
+ {
+ "epoch": 31.768558951965066,
+ "grad_norm": 0.01551708485931158,
+ "learning_rate": 0.0006,
+ "loss": 4.17064094543457,
+ "step": 2287
+ },
+ {
+ "epoch": 31.782532751091704,
+ "grad_norm": 0.015863575041294098,
+ "learning_rate": 0.0006,
+ "loss": 4.1579484939575195,
+ "step": 2288
+ },
+ {
+ "epoch": 31.79650655021834,
+ "grad_norm": 0.015118198469281197,
+ "learning_rate": 0.0006,
+ "loss": 4.1626787185668945,
+ "step": 2289
+ },
+ {
+ "epoch": 31.810480349344978,
+ "grad_norm": 0.015220776200294495,
+ "learning_rate": 0.0006,
+ "loss": 4.135746955871582,
+ "step": 2290
+ },
+ {
+ "epoch": 31.824454148471617,
+ "grad_norm": 0.014529784210026264,
+ "learning_rate": 0.0006,
+ "loss": 4.164931297302246,
+ "step": 2291
+ },
+ {
+ "epoch": 31.83842794759825,
+ "grad_norm": 0.015533296391367912,
+ "learning_rate": 0.0006,
+ "loss": 4.112464904785156,
+ "step": 2292
+ },
+ {
+ "epoch": 31.85240174672489,
+ "grad_norm": 0.016562335193157196,
+ "learning_rate": 0.0006,
+ "loss": 4.164898872375488,
+ "step": 2293
+ },
+ {
+ "epoch": 31.86637554585153,
+ "grad_norm": 0.016258113086223602,
+ "learning_rate": 0.0006,
+ "loss": 4.1979265213012695,
+ "step": 2294
+ },
+ {
+ "epoch": 31.880349344978168,
+ "grad_norm": 0.01558110024780035,
+ "learning_rate": 0.0006,
+ "loss": 4.174675941467285,
+ "step": 2295
+ },
+ {
+ "epoch": 31.894323144104803,
+ "grad_norm": 0.016412515193223953,
+ "learning_rate": 0.0006,
+ "loss": 4.217909336090088,
+ "step": 2296
+ },
+ {
+ "epoch": 31.90829694323144,
+ "grad_norm": 0.015118442475795746,
+ "learning_rate": 0.0006,
+ "loss": 4.137143611907959,
+ "step": 2297
+ },
+ {
+ "epoch": 31.92227074235808,
+ "grad_norm": 0.01506161317229271,
+ "learning_rate": 0.0006,
+ "loss": 4.196122169494629,
+ "step": 2298
+ },
+ {
+ "epoch": 31.936244541484715,
+ "grad_norm": 0.017165271565318108,
+ "learning_rate": 0.0006,
+ "loss": 4.14834451675415,
+ "step": 2299
+ },
+ {
+ "epoch": 31.950218340611354,
+ "grad_norm": 0.01696440577507019,
+ "learning_rate": 0.0006,
+ "loss": 4.218868732452393,
+ "step": 2300
+ },
+ {
+ "epoch": 31.964192139737992,
+ "grad_norm": 0.017349394038319588,
+ "learning_rate": 0.0006,
+ "loss": 4.2754011154174805,
+ "step": 2301
+ },
+ {
+ "epoch": 31.978165938864628,
+ "grad_norm": 0.018387921154499054,
+ "learning_rate": 0.0006,
+ "loss": 4.282623291015625,
+ "step": 2302
+ },
+ {
+ "epoch": 31.992139737991266,
+ "grad_norm": 0.018837926909327507,
+ "learning_rate": 0.0006,
+ "loss": 4.185425281524658,
+ "step": 2303
+ },
+ {
+ "epoch": 32.0,
+ "grad_norm": 0.018998464569449425,
+ "learning_rate": 0.0006,
+ "loss": 4.151291847229004,
+ "step": 2304
+ },
+ {
+ "epoch": 32.0,
+ "eval_loss": 4.522614002227783,
+ "eval_runtime": 57.0106,
+ "eval_samples_per_second": 42.834,
+ "eval_steps_per_second": 1.351,
+ "step": 2304
+ },
+ {
+ "epoch": 32.01397379912664,
+ "grad_norm": 0.019036833196878433,
+ "learning_rate": 0.0006,
+ "loss": 4.183202266693115,
+ "step": 2305
+ },
+ {
+ "epoch": 32.02794759825328,
+ "grad_norm": 0.02207397110760212,
+ "learning_rate": 0.0006,
+ "loss": 4.17875862121582,
+ "step": 2306
+ },
+ {
+ "epoch": 32.041921397379916,
+ "grad_norm": 0.022008279338479042,
+ "learning_rate": 0.0006,
+ "loss": 4.08042049407959,
+ "step": 2307
+ },
+ {
+ "epoch": 32.05589519650655,
+ "grad_norm": 0.022407017648220062,
+ "learning_rate": 0.0006,
+ "loss": 4.112189769744873,
+ "step": 2308
+ },
+ {
+ "epoch": 32.069868995633186,
+ "grad_norm": 0.021599747240543365,
+ "learning_rate": 0.0006,
+ "loss": 4.126455307006836,
+ "step": 2309
+ },
+ {
+ "epoch": 32.083842794759825,
+ "grad_norm": 0.020340966060757637,
+ "learning_rate": 0.0006,
+ "loss": 4.080538272857666,
+ "step": 2310
+ },
+ {
+ "epoch": 32.09781659388646,
+ "grad_norm": 0.020956892520189285,
+ "learning_rate": 0.0006,
+ "loss": 4.171943664550781,
+ "step": 2311
+ },
+ {
+ "epoch": 32.1117903930131,
+ "grad_norm": 0.019966667518019676,
+ "learning_rate": 0.0006,
+ "loss": 4.144007682800293,
+ "step": 2312
+ },
+ {
+ "epoch": 32.12576419213974,
+ "grad_norm": 0.018256952986121178,
+ "learning_rate": 0.0006,
+ "loss": 4.208107948303223,
+ "step": 2313
+ },
+ {
+ "epoch": 32.13973799126637,
+ "grad_norm": 0.017289170995354652,
+ "learning_rate": 0.0006,
+ "loss": 4.182024955749512,
+ "step": 2314
+ },
+ {
+ "epoch": 32.15371179039301,
+ "grad_norm": 0.017656700685620308,
+ "learning_rate": 0.0006,
+ "loss": 4.068373680114746,
+ "step": 2315
+ },
+ {
+ "epoch": 32.16768558951965,
+ "grad_norm": 0.017903361469507217,
+ "learning_rate": 0.0006,
+ "loss": 4.2345428466796875,
+ "step": 2316
+ },
+ {
+ "epoch": 32.18165938864629,
+ "grad_norm": 0.01785109005868435,
+ "learning_rate": 0.0006,
+ "loss": 4.219776153564453,
+ "step": 2317
+ },
+ {
+ "epoch": 32.19563318777293,
+ "grad_norm": 0.01887257769703865,
+ "learning_rate": 0.0006,
+ "loss": 4.139206409454346,
+ "step": 2318
+ },
+ {
+ "epoch": 32.209606986899566,
+ "grad_norm": 0.01924685761332512,
+ "learning_rate": 0.0006,
+ "loss": 4.121052265167236,
+ "step": 2319
+ },
+ {
+ "epoch": 32.223580786026204,
+ "grad_norm": 0.020574072375893593,
+ "learning_rate": 0.0006,
+ "loss": 4.13262414932251,
+ "step": 2320
+ },
+ {
+ "epoch": 32.237554585152836,
+ "grad_norm": 0.023849064484238625,
+ "learning_rate": 0.0006,
+ "loss": 4.239119529724121,
+ "step": 2321
+ },
+ {
+ "epoch": 32.251528384279474,
+ "grad_norm": 0.026239361613988876,
+ "learning_rate": 0.0006,
+ "loss": 4.248993873596191,
+ "step": 2322
+ },
+ {
+ "epoch": 32.26550218340611,
+ "grad_norm": 0.023980243131518364,
+ "learning_rate": 0.0006,
+ "loss": 4.159852504730225,
+ "step": 2323
+ },
+ {
+ "epoch": 32.27947598253275,
+ "grad_norm": 0.020442405715584755,
+ "learning_rate": 0.0006,
+ "loss": 4.133546829223633,
+ "step": 2324
+ },
+ {
+ "epoch": 32.29344978165939,
+ "grad_norm": 0.020661186426877975,
+ "learning_rate": 0.0006,
+ "loss": 4.078243732452393,
+ "step": 2325
+ },
+ {
+ "epoch": 32.30742358078603,
+ "grad_norm": 0.02046387456357479,
+ "learning_rate": 0.0006,
+ "loss": 4.172012805938721,
+ "step": 2326
+ },
+ {
+ "epoch": 32.32139737991266,
+ "grad_norm": 0.02089597098529339,
+ "learning_rate": 0.0006,
+ "loss": 4.0604705810546875,
+ "step": 2327
+ },
+ {
+ "epoch": 32.3353711790393,
+ "grad_norm": 0.01875591278076172,
+ "learning_rate": 0.0006,
+ "loss": 4.125911235809326,
+ "step": 2328
+ },
+ {
+ "epoch": 32.34934497816594,
+ "grad_norm": 0.018606796860694885,
+ "learning_rate": 0.0006,
+ "loss": 4.097362518310547,
+ "step": 2329
+ },
+ {
+ "epoch": 32.36331877729258,
+ "grad_norm": 0.01766626164317131,
+ "learning_rate": 0.0006,
+ "loss": 4.118616580963135,
+ "step": 2330
+ },
+ {
+ "epoch": 32.377292576419215,
+ "grad_norm": 0.017692960798740387,
+ "learning_rate": 0.0006,
+ "loss": 4.1457672119140625,
+ "step": 2331
+ },
+ {
+ "epoch": 32.391266375545854,
+ "grad_norm": 0.017619704827666283,
+ "learning_rate": 0.0006,
+ "loss": 4.175961494445801,
+ "step": 2332
+ },
+ {
+ "epoch": 32.40524017467249,
+ "grad_norm": 0.01784413494169712,
+ "learning_rate": 0.0006,
+ "loss": 4.051291465759277,
+ "step": 2333
+ },
+ {
+ "epoch": 32.419213973799124,
+ "grad_norm": 0.018508171662688255,
+ "learning_rate": 0.0006,
+ "loss": 4.0638556480407715,
+ "step": 2334
+ },
+ {
+ "epoch": 32.43318777292576,
+ "grad_norm": 0.01731436885893345,
+ "learning_rate": 0.0006,
+ "loss": 4.1805419921875,
+ "step": 2335
+ },
+ {
+ "epoch": 32.4471615720524,
+ "grad_norm": 0.015886498615145683,
+ "learning_rate": 0.0006,
+ "loss": 4.151026725769043,
+ "step": 2336
+ },
+ {
+ "epoch": 32.46113537117904,
+ "grad_norm": 0.016583820804953575,
+ "learning_rate": 0.0006,
+ "loss": 4.203604698181152,
+ "step": 2337
+ },
+ {
+ "epoch": 32.47510917030568,
+ "grad_norm": 0.0176097322255373,
+ "learning_rate": 0.0006,
+ "loss": 4.135159015655518,
+ "step": 2338
+ },
+ {
+ "epoch": 32.48908296943232,
+ "grad_norm": 0.016657903790473938,
+ "learning_rate": 0.0006,
+ "loss": 4.126681804656982,
+ "step": 2339
+ },
+ {
+ "epoch": 32.50305676855895,
+ "grad_norm": 0.016866130754351616,
+ "learning_rate": 0.0006,
+ "loss": 4.162341117858887,
+ "step": 2340
+ },
+ {
+ "epoch": 32.51703056768559,
+ "grad_norm": 0.016286680474877357,
+ "learning_rate": 0.0006,
+ "loss": 4.261801242828369,
+ "step": 2341
+ },
+ {
+ "epoch": 32.531004366812226,
+ "grad_norm": 0.01481365505605936,
+ "learning_rate": 0.0006,
+ "loss": 4.110016822814941,
+ "step": 2342
+ },
+ {
+ "epoch": 32.544978165938865,
+ "grad_norm": 0.016605013981461525,
+ "learning_rate": 0.0006,
+ "loss": 4.230024814605713,
+ "step": 2343
+ },
+ {
+ "epoch": 32.5589519650655,
+ "grad_norm": 0.01645175740122795,
+ "learning_rate": 0.0006,
+ "loss": 4.174612045288086,
+ "step": 2344
+ },
+ {
+ "epoch": 32.57292576419214,
+ "grad_norm": 0.014725026674568653,
+ "learning_rate": 0.0006,
+ "loss": 4.066371440887451,
+ "step": 2345
+ },
+ {
+ "epoch": 32.58689956331878,
+ "grad_norm": 0.015003837645053864,
+ "learning_rate": 0.0006,
+ "loss": 4.193343639373779,
+ "step": 2346
+ },
+ {
+ "epoch": 32.60087336244541,
+ "grad_norm": 0.015346302650868893,
+ "learning_rate": 0.0006,
+ "loss": 4.210659980773926,
+ "step": 2347
+ },
+ {
+ "epoch": 32.61484716157205,
+ "grad_norm": 0.015135962516069412,
+ "learning_rate": 0.0006,
+ "loss": 4.204618453979492,
+ "step": 2348
+ },
+ {
+ "epoch": 32.62882096069869,
+ "grad_norm": 0.01601956970989704,
+ "learning_rate": 0.0006,
+ "loss": 4.145564556121826,
+ "step": 2349
+ },
+ {
+ "epoch": 32.64279475982533,
+ "grad_norm": 0.01517851185053587,
+ "learning_rate": 0.0006,
+ "loss": 4.051673889160156,
+ "step": 2350
+ },
+ {
+ "epoch": 32.65676855895197,
+ "grad_norm": 0.014894292689859867,
+ "learning_rate": 0.0006,
+ "loss": 4.184818267822266,
+ "step": 2351
+ },
+ {
+ "epoch": 32.670742358078606,
+ "grad_norm": 0.015720834955573082,
+ "learning_rate": 0.0006,
+ "loss": 4.220036506652832,
+ "step": 2352
+ },
+ {
+ "epoch": 32.68471615720524,
+ "grad_norm": 0.01573433168232441,
+ "learning_rate": 0.0006,
+ "loss": 4.076773166656494,
+ "step": 2353
+ },
+ {
+ "epoch": 32.698689956331876,
+ "grad_norm": 0.015828052535653114,
+ "learning_rate": 0.0006,
+ "loss": 4.116329669952393,
+ "step": 2354
+ },
+ {
+ "epoch": 32.712663755458514,
+ "grad_norm": 0.015462259761989117,
+ "learning_rate": 0.0006,
+ "loss": 4.1155242919921875,
+ "step": 2355
+ },
+ {
+ "epoch": 32.72663755458515,
+ "grad_norm": 0.015763849020004272,
+ "learning_rate": 0.0006,
+ "loss": 4.169687747955322,
+ "step": 2356
+ },
+ {
+ "epoch": 32.74061135371179,
+ "grad_norm": 0.014935145154595375,
+ "learning_rate": 0.0006,
+ "loss": 4.2090911865234375,
+ "step": 2357
+ },
+ {
+ "epoch": 32.75458515283843,
+ "grad_norm": 0.015119184739887714,
+ "learning_rate": 0.0006,
+ "loss": 4.131838798522949,
+ "step": 2358
+ },
+ {
+ "epoch": 32.76855895196506,
+ "grad_norm": 0.014833525754511356,
+ "learning_rate": 0.0006,
+ "loss": 4.17618465423584,
+ "step": 2359
+ },
+ {
+ "epoch": 32.7825327510917,
+ "grad_norm": 0.01588359661400318,
+ "learning_rate": 0.0006,
+ "loss": 4.163723945617676,
+ "step": 2360
+ },
+ {
+ "epoch": 32.79650655021834,
+ "grad_norm": 0.0168188214302063,
+ "learning_rate": 0.0006,
+ "loss": 4.178189277648926,
+ "step": 2361
+ },
+ {
+ "epoch": 32.81048034934498,
+ "grad_norm": 0.01726095750927925,
+ "learning_rate": 0.0006,
+ "loss": 4.144223690032959,
+ "step": 2362
+ },
+ {
+ "epoch": 32.82445414847162,
+ "grad_norm": 0.01701410301029682,
+ "learning_rate": 0.0006,
+ "loss": 4.209896087646484,
+ "step": 2363
+ },
+ {
+ "epoch": 32.838427947598255,
+ "grad_norm": 0.01682264916598797,
+ "learning_rate": 0.0006,
+ "loss": 4.26223087310791,
+ "step": 2364
+ },
+ {
+ "epoch": 32.852401746724894,
+ "grad_norm": 0.015597456134855747,
+ "learning_rate": 0.0006,
+ "loss": 4.292664051055908,
+ "step": 2365
+ },
+ {
+ "epoch": 32.866375545851525,
+ "grad_norm": 0.015258579514920712,
+ "learning_rate": 0.0006,
+ "loss": 4.165006637573242,
+ "step": 2366
+ },
+ {
+ "epoch": 32.880349344978164,
+ "grad_norm": 0.01662200503051281,
+ "learning_rate": 0.0006,
+ "loss": 4.309786319732666,
+ "step": 2367
+ },
+ {
+ "epoch": 32.8943231441048,
+ "grad_norm": 0.015691161155700684,
+ "learning_rate": 0.0006,
+ "loss": 4.167202949523926,
+ "step": 2368
+ },
+ {
+ "epoch": 32.90829694323144,
+ "grad_norm": 0.014870637096464634,
+ "learning_rate": 0.0006,
+ "loss": 4.142415523529053,
+ "step": 2369
+ },
+ {
+ "epoch": 32.92227074235808,
+ "grad_norm": 0.01567252166569233,
+ "learning_rate": 0.0006,
+ "loss": 4.2312092781066895,
+ "step": 2370
+ },
+ {
+ "epoch": 32.93624454148472,
+ "grad_norm": 0.014948616735637188,
+ "learning_rate": 0.0006,
+ "loss": 4.1514129638671875,
+ "step": 2371
+ },
+ {
+ "epoch": 32.95021834061135,
+ "grad_norm": 0.014654500409960747,
+ "learning_rate": 0.0006,
+ "loss": 4.205010890960693,
+ "step": 2372
+ },
+ {
+ "epoch": 32.96419213973799,
+ "grad_norm": 0.015059148892760277,
+ "learning_rate": 0.0006,
+ "loss": 4.148405075073242,
+ "step": 2373
+ },
+ {
+ "epoch": 32.97816593886463,
+ "grad_norm": 0.015026576817035675,
+ "learning_rate": 0.0006,
+ "loss": 4.182535171508789,
+ "step": 2374
+ },
+ {
+ "epoch": 32.992139737991266,
+ "grad_norm": 0.016394320875406265,
+ "learning_rate": 0.0006,
+ "loss": 4.129051208496094,
+ "step": 2375
+ },
+ {
+ "epoch": 33.0,
+ "grad_norm": 0.01867043413221836,
+ "learning_rate": 0.0006,
+ "loss": 4.185556411743164,
+ "step": 2376
+ },
+ {
+ "epoch": 33.0,
+ "eval_loss": 4.640405654907227,
+ "eval_runtime": 56.5861,
+ "eval_samples_per_second": 43.155,
+ "eval_steps_per_second": 1.361,
+ "step": 2376
+ },
+ {
+ "epoch": 33.01397379912664,
+ "grad_norm": 0.018626727163791656,
+ "learning_rate": 0.0006,
+ "loss": 4.143031597137451,
+ "step": 2377
+ },
+ {
+ "epoch": 33.02794759825328,
+ "grad_norm": 0.019283363595604897,
+ "learning_rate": 0.0006,
+ "loss": 4.117845058441162,
+ "step": 2378
+ },
+ {
+ "epoch": 33.041921397379916,
+ "grad_norm": 0.018185822293162346,
+ "learning_rate": 0.0006,
+ "loss": 4.076862812042236,
+ "step": 2379
+ },
+ {
+ "epoch": 33.05589519650655,
+ "grad_norm": 0.019411196932196617,
+ "learning_rate": 0.0006,
+ "loss": 4.124156475067139,
+ "step": 2380
+ },
+ {
+ "epoch": 33.069868995633186,
+ "grad_norm": 0.019248466938734055,
+ "learning_rate": 0.0006,
+ "loss": 4.201735019683838,
+ "step": 2381
+ },
+ {
+ "epoch": 33.083842794759825,
+ "grad_norm": 0.018856720998883247,
+ "learning_rate": 0.0006,
+ "loss": 4.00296688079834,
+ "step": 2382
+ },
+ {
+ "epoch": 33.09781659388646,
+ "grad_norm": 0.018554294481873512,
+ "learning_rate": 0.0006,
+ "loss": 4.270577430725098,
+ "step": 2383
+ },
+ {
+ "epoch": 33.1117903930131,
+ "grad_norm": 0.019946224987506866,
+ "learning_rate": 0.0006,
+ "loss": 4.119589328765869,
+ "step": 2384
+ },
+ {
+ "epoch": 33.12576419213974,
+ "grad_norm": 0.021395670250058174,
+ "learning_rate": 0.0006,
+ "loss": 4.128847122192383,
+ "step": 2385
+ },
+ {
+ "epoch": 33.13973799126637,
+ "grad_norm": 0.02009674906730652,
+ "learning_rate": 0.0006,
+ "loss": 4.183228969573975,
+ "step": 2386
+ },
+ {
+ "epoch": 33.15371179039301,
+ "grad_norm": 0.01820223033428192,
+ "learning_rate": 0.0006,
+ "loss": 4.131601810455322,
+ "step": 2387
+ },
+ {
+ "epoch": 33.16768558951965,
+ "grad_norm": 0.018653156235814095,
+ "learning_rate": 0.0006,
+ "loss": 4.116430282592773,
+ "step": 2388
+ },
+ {
+ "epoch": 33.18165938864629,
+ "grad_norm": 0.016828063875436783,
+ "learning_rate": 0.0006,
+ "loss": 4.054663181304932,
+ "step": 2389
+ },
+ {
+ "epoch": 33.19563318777293,
+ "grad_norm": 0.017025409266352654,
+ "learning_rate": 0.0006,
+ "loss": 4.110044956207275,
+ "step": 2390
+ },
+ {
+ "epoch": 33.209606986899566,
+ "grad_norm": 0.016608452424407005,
+ "learning_rate": 0.0006,
+ "loss": 4.102668285369873,
+ "step": 2391
+ },
+ {
+ "epoch": 33.223580786026204,
+ "grad_norm": 0.01743577979505062,
+ "learning_rate": 0.0006,
+ "loss": 4.166791915893555,
+ "step": 2392
+ },
+ {
+ "epoch": 33.237554585152836,
+ "grad_norm": 0.017253734171390533,
+ "learning_rate": 0.0006,
+ "loss": 4.217832088470459,
+ "step": 2393
+ },
+ {
+ "epoch": 33.251528384279474,
+ "grad_norm": 0.017937595024704933,
+ "learning_rate": 0.0006,
+ "loss": 4.2173967361450195,
+ "step": 2394
+ },
+ {
+ "epoch": 33.26550218340611,
+ "grad_norm": 0.01985993981361389,
+ "learning_rate": 0.0006,
+ "loss": 4.17849588394165,
+ "step": 2395
+ },
+ {
+ "epoch": 33.27947598253275,
+ "grad_norm": 0.019243355840444565,
+ "learning_rate": 0.0006,
+ "loss": 4.086523056030273,
+ "step": 2396
+ },
+ {
+ "epoch": 33.29344978165939,
+ "grad_norm": 0.01924416422843933,
+ "learning_rate": 0.0006,
+ "loss": 4.128574371337891,
+ "step": 2397
+ },
+ {
+ "epoch": 33.30742358078603,
+ "grad_norm": 0.01983758620917797,
+ "learning_rate": 0.0006,
+ "loss": 4.061581611633301,
+ "step": 2398
+ },
+ {
+ "epoch": 33.32139737991266,
+ "grad_norm": 0.01930462196469307,
+ "learning_rate": 0.0006,
+ "loss": 4.091731071472168,
+ "step": 2399
+ },
+ {
+ "epoch": 33.3353711790393,
+ "grad_norm": 0.017988424748182297,
+ "learning_rate": 0.0006,
+ "loss": 4.155599117279053,
+ "step": 2400
+ },
+ {
+ "epoch": 33.34934497816594,
+ "grad_norm": 0.015194767154753208,
+ "learning_rate": 0.0006,
+ "loss": 4.144134044647217,
+ "step": 2401
+ },
+ {
+ "epoch": 33.36331877729258,
+ "grad_norm": 0.01536081824451685,
+ "learning_rate": 0.0006,
+ "loss": 4.127272129058838,
+ "step": 2402
+ },
+ {
+ "epoch": 33.377292576419215,
+ "grad_norm": 0.015431297942996025,
+ "learning_rate": 0.0006,
+ "loss": 4.3213019371032715,
+ "step": 2403
+ },
+ {
+ "epoch": 33.391266375545854,
+ "grad_norm": 0.015740280970931053,
+ "learning_rate": 0.0006,
+ "loss": 4.121467590332031,
+ "step": 2404
+ },
+ {
+ "epoch": 33.40524017467249,
+ "grad_norm": 0.015379384160041809,
+ "learning_rate": 0.0006,
+ "loss": 4.095216751098633,
+ "step": 2405
+ },
+ {
+ "epoch": 33.419213973799124,
+ "grad_norm": 0.01528073474764824,
+ "learning_rate": 0.0006,
+ "loss": 4.13047981262207,
+ "step": 2406
+ },
+ {
+ "epoch": 33.43318777292576,
+ "grad_norm": 0.015739979222416878,
+ "learning_rate": 0.0006,
+ "loss": 4.2533135414123535,
+ "step": 2407
+ },
+ {
+ "epoch": 33.4471615720524,
+ "grad_norm": 0.01519764307886362,
+ "learning_rate": 0.0006,
+ "loss": 4.131782054901123,
+ "step": 2408
+ },
+ {
+ "epoch": 33.46113537117904,
+ "grad_norm": 0.015425996854901314,
+ "learning_rate": 0.0006,
+ "loss": 4.02811336517334,
+ "step": 2409
+ },
+ {
+ "epoch": 33.47510917030568,
+ "grad_norm": 0.015378830023109913,
+ "learning_rate": 0.0006,
+ "loss": 4.08652925491333,
+ "step": 2410
+ },
+ {
+ "epoch": 33.48908296943232,
+ "grad_norm": 0.01473075058311224,
+ "learning_rate": 0.0006,
+ "loss": 4.214364528656006,
+ "step": 2411
+ },
+ {
+ "epoch": 33.50305676855895,
+ "grad_norm": 0.015141637064516544,
+ "learning_rate": 0.0006,
+ "loss": 4.155712127685547,
+ "step": 2412
+ },
+ {
+ "epoch": 33.51703056768559,
+ "grad_norm": 0.01641593500971794,
+ "learning_rate": 0.0006,
+ "loss": 4.097907066345215,
+ "step": 2413
+ },
+ {
+ "epoch": 33.531004366812226,
+ "grad_norm": 0.015182029455900192,
+ "learning_rate": 0.0006,
+ "loss": 4.238499641418457,
+ "step": 2414
+ },
+ {
+ "epoch": 33.544978165938865,
+ "grad_norm": 0.01643732376396656,
+ "learning_rate": 0.0006,
+ "loss": 4.0661492347717285,
+ "step": 2415
+ },
+ {
+ "epoch": 33.5589519650655,
+ "grad_norm": 0.015614228323101997,
+ "learning_rate": 0.0006,
+ "loss": 4.120795249938965,
+ "step": 2416
+ },
+ {
+ "epoch": 33.57292576419214,
+ "grad_norm": 0.016066187992691994,
+ "learning_rate": 0.0006,
+ "loss": 4.096465110778809,
+ "step": 2417
+ },
+ {
+ "epoch": 33.58689956331878,
+ "grad_norm": 0.017214342951774597,
+ "learning_rate": 0.0006,
+ "loss": 4.242628574371338,
+ "step": 2418
+ },
+ {
+ "epoch": 33.60087336244541,
+ "grad_norm": 0.01787586510181427,
+ "learning_rate": 0.0006,
+ "loss": 4.149589538574219,
+ "step": 2419
+ },
+ {
+ "epoch": 33.61484716157205,
+ "grad_norm": 0.019455241039395332,
+ "learning_rate": 0.0006,
+ "loss": 4.267796039581299,
+ "step": 2420
+ },
+ {
+ "epoch": 33.62882096069869,
+ "grad_norm": 0.01984674669802189,
+ "learning_rate": 0.0006,
+ "loss": 4.175220966339111,
+ "step": 2421
+ },
+ {
+ "epoch": 33.64279475982533,
+ "grad_norm": 0.021933717653155327,
+ "learning_rate": 0.0006,
+ "loss": 4.106647491455078,
+ "step": 2422
+ },
+ {
+ "epoch": 33.65676855895197,
+ "grad_norm": 0.02123589813709259,
+ "learning_rate": 0.0006,
+ "loss": 4.204996109008789,
+ "step": 2423
+ },
+ {
+ "epoch": 33.670742358078606,
+ "grad_norm": 0.019814589992165565,
+ "learning_rate": 0.0006,
+ "loss": 4.186171531677246,
+ "step": 2424
+ },
+ {
+ "epoch": 33.68471615720524,
+ "grad_norm": 0.02059728465974331,
+ "learning_rate": 0.0006,
+ "loss": 4.170509338378906,
+ "step": 2425
+ },
+ {
+ "epoch": 33.698689956331876,
+ "grad_norm": 0.018483608961105347,
+ "learning_rate": 0.0006,
+ "loss": 4.085585594177246,
+ "step": 2426
+ },
+ {
+ "epoch": 33.712663755458514,
+ "grad_norm": 0.018403852358460426,
+ "learning_rate": 0.0006,
+ "loss": 4.2551774978637695,
+ "step": 2427
+ },
+ {
+ "epoch": 33.72663755458515,
+ "grad_norm": 0.018653830513358116,
+ "learning_rate": 0.0006,
+ "loss": 4.091766357421875,
+ "step": 2428
+ },
+ {
+ "epoch": 33.74061135371179,
+ "grad_norm": 0.017692334949970245,
+ "learning_rate": 0.0006,
+ "loss": 4.100838661193848,
+ "step": 2429
+ },
+ {
+ "epoch": 33.75458515283843,
+ "grad_norm": 0.016029126942157745,
+ "learning_rate": 0.0006,
+ "loss": 4.1530680656433105,
+ "step": 2430
+ },
+ {
+ "epoch": 33.76855895196506,
+ "grad_norm": 0.016095684841275215,
+ "learning_rate": 0.0006,
+ "loss": 4.20544958114624,
+ "step": 2431
+ },
+ {
+ "epoch": 33.7825327510917,
+ "grad_norm": 0.016375798732042313,
+ "learning_rate": 0.0006,
+ "loss": 4.199177265167236,
+ "step": 2432
+ },
+ {
+ "epoch": 33.79650655021834,
+ "grad_norm": 0.01590786688029766,
+ "learning_rate": 0.0006,
+ "loss": 4.200799465179443,
+ "step": 2433
+ },
+ {
+ "epoch": 33.81048034934498,
+ "grad_norm": 0.014896097593009472,
+ "learning_rate": 0.0006,
+ "loss": 4.185850143432617,
+ "step": 2434
+ },
+ {
+ "epoch": 33.82445414847162,
+ "grad_norm": 0.01438513956964016,
+ "learning_rate": 0.0006,
+ "loss": 4.214962005615234,
+ "step": 2435
+ },
+ {
+ "epoch": 33.838427947598255,
+ "grad_norm": 0.014459339901804924,
+ "learning_rate": 0.0006,
+ "loss": 4.196972370147705,
+ "step": 2436
+ },
+ {
+ "epoch": 33.852401746724894,
+ "grad_norm": 0.014782169833779335,
+ "learning_rate": 0.0006,
+ "loss": 4.0885910987854,
+ "step": 2437
+ },
+ {
+ "epoch": 33.866375545851525,
+ "grad_norm": 0.01520370040088892,
+ "learning_rate": 0.0006,
+ "loss": 4.0323262214660645,
+ "step": 2438
+ },
+ {
+ "epoch": 33.880349344978164,
+ "grad_norm": 0.014056282117962837,
+ "learning_rate": 0.0006,
+ "loss": 4.1386518478393555,
+ "step": 2439
+ },
+ {
+ "epoch": 33.8943231441048,
+ "grad_norm": 0.013567990623414516,
+ "learning_rate": 0.0006,
+ "loss": 4.095754623413086,
+ "step": 2440
+ },
+ {
+ "epoch": 33.90829694323144,
+ "grad_norm": 0.013306716457009315,
+ "learning_rate": 0.0006,
+ "loss": 4.1623077392578125,
+ "step": 2441
+ },
+ {
+ "epoch": 33.92227074235808,
+ "grad_norm": 0.014047150500118732,
+ "learning_rate": 0.0006,
+ "loss": 4.142541885375977,
+ "step": 2442
+ },
+ {
+ "epoch": 33.93624454148472,
+ "grad_norm": 0.01377617847174406,
+ "learning_rate": 0.0006,
+ "loss": 4.268551826477051,
+ "step": 2443
+ },
+ {
+ "epoch": 33.95021834061135,
+ "grad_norm": 0.014671621844172478,
+ "learning_rate": 0.0006,
+ "loss": 4.198196887969971,
+ "step": 2444
+ },
+ {
+ "epoch": 33.96419213973799,
+ "grad_norm": 0.014806744642555714,
+ "learning_rate": 0.0006,
+ "loss": 4.057899475097656,
+ "step": 2445
+ },
+ {
+ "epoch": 33.97816593886463,
+ "grad_norm": 0.015734272077679634,
+ "learning_rate": 0.0006,
+ "loss": 4.163505554199219,
+ "step": 2446
+ },
+ {
+ "epoch": 33.992139737991266,
+ "grad_norm": 0.015250146389007568,
+ "learning_rate": 0.0006,
+ "loss": 4.076294898986816,
+ "step": 2447
+ },
+ {
+ "epoch": 34.0,
+ "grad_norm": 0.016733702272176743,
+ "learning_rate": 0.0006,
+ "loss": 4.138420104980469,
+ "step": 2448
+ },
+ {
+ "epoch": 34.0,
+ "eval_loss": 4.542503356933594,
+ "eval_runtime": 56.983,
+ "eval_samples_per_second": 42.855,
+ "eval_steps_per_second": 1.351,
+ "step": 2448
+ },
+ {
+ "epoch": 34.01397379912664,
+ "grad_norm": 0.015866613015532494,
+ "learning_rate": 0.0006,
+ "loss": 4.032081127166748,
+ "step": 2449
+ },
+ {
+ "epoch": 34.02794759825328,
+ "grad_norm": 0.01946570910513401,
+ "learning_rate": 0.0006,
+ "loss": 4.122320175170898,
+ "step": 2450
+ },
+ {
+ "epoch": 34.041921397379916,
+ "grad_norm": 0.020265938714146614,
+ "learning_rate": 0.0006,
+ "loss": 4.06460428237915,
+ "step": 2451
+ },
+ {
+ "epoch": 34.05589519650655,
+ "grad_norm": 0.018887171521782875,
+ "learning_rate": 0.0006,
+ "loss": 4.096906661987305,
+ "step": 2452
+ },
+ {
+ "epoch": 34.069868995633186,
+ "grad_norm": 0.017304692417383194,
+ "learning_rate": 0.0006,
+ "loss": 4.137261390686035,
+ "step": 2453
+ },
+ {
+ "epoch": 34.083842794759825,
+ "grad_norm": 0.017029082402586937,
+ "learning_rate": 0.0006,
+ "loss": 4.081112861633301,
+ "step": 2454
+ },
+ {
+ "epoch": 34.09781659388646,
+ "grad_norm": 0.016414813697338104,
+ "learning_rate": 0.0006,
+ "loss": 4.140783309936523,
+ "step": 2455
+ },
+ {
+ "epoch": 34.1117903930131,
+ "grad_norm": 0.018238281831145287,
+ "learning_rate": 0.0006,
+ "loss": 4.161116600036621,
+ "step": 2456
+ },
+ {
+ "epoch": 34.12576419213974,
+ "grad_norm": 0.019929232075810432,
+ "learning_rate": 0.0006,
+ "loss": 4.243722438812256,
+ "step": 2457
+ },
+ {
+ "epoch": 34.13973799126637,
+ "grad_norm": 0.019375786185264587,
+ "learning_rate": 0.0006,
+ "loss": 4.134873390197754,
+ "step": 2458
+ },
+ {
+ "epoch": 34.15371179039301,
+ "grad_norm": 0.018151137977838516,
+ "learning_rate": 0.0006,
+ "loss": 4.098745346069336,
+ "step": 2459
+ },
+ {
+ "epoch": 34.16768558951965,
+ "grad_norm": 0.017054934054613113,
+ "learning_rate": 0.0006,
+ "loss": 4.194662570953369,
+ "step": 2460
+ },
+ {
+ "epoch": 34.18165938864629,
+ "grad_norm": 0.018220555037260056,
+ "learning_rate": 0.0006,
+ "loss": 4.128866195678711,
+ "step": 2461
+ },
+ {
+ "epoch": 34.19563318777293,
+ "grad_norm": 0.01692879945039749,
+ "learning_rate": 0.0006,
+ "loss": 4.062477111816406,
+ "step": 2462
+ },
+ {
+ "epoch": 34.209606986899566,
+ "grad_norm": 0.01674199476838112,
+ "learning_rate": 0.0006,
+ "loss": 4.186149597167969,
+ "step": 2463
+ },
+ {
+ "epoch": 34.223580786026204,
+ "grad_norm": 0.016527455300092697,
+ "learning_rate": 0.0006,
+ "loss": 4.0734028816223145,
+ "step": 2464
+ },
+ {
+ "epoch": 34.237554585152836,
+ "grad_norm": 0.016448386013507843,
+ "learning_rate": 0.0006,
+ "loss": 4.098896026611328,
+ "step": 2465
+ },
+ {
+ "epoch": 34.251528384279474,
+ "grad_norm": 0.014644953422248363,
+ "learning_rate": 0.0006,
+ "loss": 4.1031904220581055,
+ "step": 2466
+ },
+ {
+ "epoch": 34.26550218340611,
+ "grad_norm": 0.016649341210722923,
+ "learning_rate": 0.0006,
+ "loss": 4.148791313171387,
+ "step": 2467
+ },
+ {
+ "epoch": 34.27947598253275,
+ "grad_norm": 0.017442714422941208,
+ "learning_rate": 0.0006,
+ "loss": 4.070799827575684,
+ "step": 2468
+ },
+ {
+ "epoch": 34.29344978165939,
+ "grad_norm": 0.017852013930678368,
+ "learning_rate": 0.0006,
+ "loss": 4.031924247741699,
+ "step": 2469
+ },
+ {
+ "epoch": 34.30742358078603,
+ "grad_norm": 0.016371270641684532,
+ "learning_rate": 0.0006,
+ "loss": 4.149420738220215,
+ "step": 2470
+ },
+ {
+ "epoch": 34.32139737991266,
+ "grad_norm": 0.015846814960241318,
+ "learning_rate": 0.0006,
+ "loss": 4.15427303314209,
+ "step": 2471
+ },
+ {
+ "epoch": 34.3353711790393,
+ "grad_norm": 0.017411163076758385,
+ "learning_rate": 0.0006,
+ "loss": 4.1304192543029785,
+ "step": 2472
+ },
+ {
+ "epoch": 34.34934497816594,
+ "grad_norm": 0.01780046708881855,
+ "learning_rate": 0.0006,
+ "loss": 4.191732406616211,
+ "step": 2473
+ },
+ {
+ "epoch": 34.36331877729258,
+ "grad_norm": 0.01958383060991764,
+ "learning_rate": 0.0006,
+ "loss": 4.068355083465576,
+ "step": 2474
+ },
+ {
+ "epoch": 34.377292576419215,
+ "grad_norm": 0.022173317149281502,
+ "learning_rate": 0.0006,
+ "loss": 4.182929039001465,
+ "step": 2475
+ },
+ {
+ "epoch": 34.391266375545854,
+ "grad_norm": 0.022389156743884087,
+ "learning_rate": 0.0006,
+ "loss": 4.110218048095703,
+ "step": 2476
+ },
+ {
+ "epoch": 34.40524017467249,
+ "grad_norm": 0.022478511556982994,
+ "learning_rate": 0.0006,
+ "loss": 4.095479488372803,
+ "step": 2477
+ },
+ {
+ "epoch": 34.419213973799124,
+ "grad_norm": 0.019504647701978683,
+ "learning_rate": 0.0006,
+ "loss": 4.101964950561523,
+ "step": 2478
+ },
+ {
+ "epoch": 34.43318777292576,
+ "grad_norm": 0.019378576427698135,
+ "learning_rate": 0.0006,
+ "loss": 4.14620304107666,
+ "step": 2479
+ },
+ {
+ "epoch": 34.4471615720524,
+ "grad_norm": 0.017993204295635223,
+ "learning_rate": 0.0006,
+ "loss": 4.183492660522461,
+ "step": 2480
+ },
+ {
+ "epoch": 34.46113537117904,
+ "grad_norm": 0.01814805157482624,
+ "learning_rate": 0.0006,
+ "loss": 4.151247978210449,
+ "step": 2481
+ },
+ {
+ "epoch": 34.47510917030568,
+ "grad_norm": 0.018622448667883873,
+ "learning_rate": 0.0006,
+ "loss": 4.020938396453857,
+ "step": 2482
+ },
+ {
+ "epoch": 34.48908296943232,
+ "grad_norm": 0.018360132351517677,
+ "learning_rate": 0.0006,
+ "loss": 4.064734935760498,
+ "step": 2483
+ },
+ {
+ "epoch": 34.50305676855895,
+ "grad_norm": 0.017715711146593094,
+ "learning_rate": 0.0006,
+ "loss": 4.185687065124512,
+ "step": 2484
+ },
+ {
+ "epoch": 34.51703056768559,
+ "grad_norm": 0.01714295893907547,
+ "learning_rate": 0.0006,
+ "loss": 4.1044721603393555,
+ "step": 2485
+ },
+ {
+ "epoch": 34.531004366812226,
+ "grad_norm": 0.016943803057074547,
+ "learning_rate": 0.0006,
+ "loss": 4.106837272644043,
+ "step": 2486
+ },
+ {
+ "epoch": 34.544978165938865,
+ "grad_norm": 0.0182423684746027,
+ "learning_rate": 0.0006,
+ "loss": 4.1041340827941895,
+ "step": 2487
+ },
+ {
+ "epoch": 34.5589519650655,
+ "grad_norm": 0.01636587455868721,
+ "learning_rate": 0.0006,
+ "loss": 3.9735546112060547,
+ "step": 2488
+ },
+ {
+ "epoch": 34.57292576419214,
+ "grad_norm": 0.014593546278774738,
+ "learning_rate": 0.0006,
+ "loss": 4.041980266571045,
+ "step": 2489
+ },
+ {
+ "epoch": 34.58689956331878,
+ "grad_norm": 0.015436965972185135,
+ "learning_rate": 0.0006,
+ "loss": 4.301148414611816,
+ "step": 2490
+ },
+ {
+ "epoch": 34.60087336244541,
+ "grad_norm": 0.015075269155204296,
+ "learning_rate": 0.0006,
+ "loss": 4.118527412414551,
+ "step": 2491
+ },
+ {
+ "epoch": 34.61484716157205,
+ "grad_norm": 0.01701514609158039,
+ "learning_rate": 0.0006,
+ "loss": 4.245968818664551,
+ "step": 2492
+ },
+ {
+ "epoch": 34.62882096069869,
+ "grad_norm": 0.018047887831926346,
+ "learning_rate": 0.0006,
+ "loss": 4.227261066436768,
+ "step": 2493
+ },
+ {
+ "epoch": 34.64279475982533,
+ "grad_norm": 0.018499085679650307,
+ "learning_rate": 0.0006,
+ "loss": 4.121360778808594,
+ "step": 2494
+ },
+ {
+ "epoch": 34.65676855895197,
+ "grad_norm": 0.018987711519002914,
+ "learning_rate": 0.0006,
+ "loss": 4.131564617156982,
+ "step": 2495
+ },
+ {
+ "epoch": 34.670742358078606,
+ "grad_norm": 0.01834476925432682,
+ "learning_rate": 0.0006,
+ "loss": 4.224941253662109,
+ "step": 2496
+ },
+ {
+ "epoch": 34.68471615720524,
+ "grad_norm": 0.017507942393422127,
+ "learning_rate": 0.0006,
+ "loss": 4.1394548416137695,
+ "step": 2497
+ },
+ {
+ "epoch": 34.698689956331876,
+ "grad_norm": 0.01680580899119377,
+ "learning_rate": 0.0006,
+ "loss": 4.1073431968688965,
+ "step": 2498
+ },
+ {
+ "epoch": 34.712663755458514,
+ "grad_norm": 0.015233514830470085,
+ "learning_rate": 0.0006,
+ "loss": 4.136994361877441,
+ "step": 2499
+ },
+ {
+ "epoch": 34.72663755458515,
+ "grad_norm": 0.015482917428016663,
+ "learning_rate": 0.0006,
+ "loss": 4.110719680786133,
+ "step": 2500
+ },
+ {
+ "epoch": 34.74061135371179,
+ "grad_norm": 0.015940461307764053,
+ "learning_rate": 0.0006,
+ "loss": 4.207851886749268,
+ "step": 2501
+ },
+ {
+ "epoch": 34.75458515283843,
+ "grad_norm": 0.015355597250163555,
+ "learning_rate": 0.0006,
+ "loss": 4.169469833374023,
+ "step": 2502
+ },
+ {
+ "epoch": 34.76855895196506,
+ "grad_norm": 0.015974000096321106,
+ "learning_rate": 0.0006,
+ "loss": 4.110061168670654,
+ "step": 2503
+ },
+ {
+ "epoch": 34.7825327510917,
+ "grad_norm": 0.016088927164673805,
+ "learning_rate": 0.0006,
+ "loss": 4.155452728271484,
+ "step": 2504
+ },
+ {
+ "epoch": 34.79650655021834,
+ "grad_norm": 0.016156120225787163,
+ "learning_rate": 0.0006,
+ "loss": 4.178092956542969,
+ "step": 2505
+ },
+ {
+ "epoch": 34.81048034934498,
+ "grad_norm": 0.014928647316992283,
+ "learning_rate": 0.0006,
+ "loss": 4.05457878112793,
+ "step": 2506
+ },
+ {
+ "epoch": 34.82445414847162,
+ "grad_norm": 0.015289328061044216,
+ "learning_rate": 0.0006,
+ "loss": 4.18869686126709,
+ "step": 2507
+ },
+ {
+ "epoch": 34.838427947598255,
+ "grad_norm": 0.015799039974808693,
+ "learning_rate": 0.0006,
+ "loss": 4.106073379516602,
+ "step": 2508
+ },
+ {
+ "epoch": 34.852401746724894,
+ "grad_norm": 0.015824418514966965,
+ "learning_rate": 0.0006,
+ "loss": 4.0634589195251465,
+ "step": 2509
+ },
+ {
+ "epoch": 34.866375545851525,
+ "grad_norm": 0.015321656130254269,
+ "learning_rate": 0.0006,
+ "loss": 4.093090057373047,
+ "step": 2510
+ },
+ {
+ "epoch": 34.880349344978164,
+ "grad_norm": 0.014380166307091713,
+ "learning_rate": 0.0006,
+ "loss": 4.1750640869140625,
+ "step": 2511
+ },
+ {
+ "epoch": 34.8943231441048,
+ "grad_norm": 0.014962567947804928,
+ "learning_rate": 0.0006,
+ "loss": 4.185277938842773,
+ "step": 2512
+ },
+ {
+ "epoch": 34.90829694323144,
+ "grad_norm": 0.014467543922364712,
+ "learning_rate": 0.0006,
+ "loss": 4.1600799560546875,
+ "step": 2513
+ },
+ {
+ "epoch": 34.92227074235808,
+ "grad_norm": 0.01545181218534708,
+ "learning_rate": 0.0006,
+ "loss": 4.096299648284912,
+ "step": 2514
+ },
+ {
+ "epoch": 34.93624454148472,
+ "grad_norm": 0.01605340465903282,
+ "learning_rate": 0.0006,
+ "loss": 4.203323841094971,
+ "step": 2515
+ },
+ {
+ "epoch": 34.95021834061135,
+ "grad_norm": 0.015203815884888172,
+ "learning_rate": 0.0006,
+ "loss": 4.201385021209717,
+ "step": 2516
+ },
+ {
+ "epoch": 34.96419213973799,
+ "grad_norm": 0.015149409882724285,
+ "learning_rate": 0.0006,
+ "loss": 4.088164329528809,
+ "step": 2517
+ },
+ {
+ "epoch": 34.97816593886463,
+ "grad_norm": 0.014885558746755123,
+ "learning_rate": 0.0006,
+ "loss": 4.073911666870117,
+ "step": 2518
+ },
+ {
+ "epoch": 34.992139737991266,
+ "grad_norm": 0.015003419481217861,
+ "learning_rate": 0.0006,
+ "loss": 4.157113552093506,
+ "step": 2519
+ },
+ {
+ "epoch": 35.0,
+ "grad_norm": 0.01641916297376156,
+ "learning_rate": 0.0006,
+ "loss": 3.970052719116211,
+ "step": 2520
+ },
+ {
+ "epoch": 35.0,
+ "eval_loss": 4.641692161560059,
+ "eval_runtime": 56.8836,
+ "eval_samples_per_second": 42.93,
+ "eval_steps_per_second": 1.354,
+ "step": 2520
+ },
+ {
+ "epoch": 35.01397379912664,
+ "grad_norm": 0.015894673764705658,
+ "learning_rate": 0.0006,
+ "loss": 4.104437351226807,
+ "step": 2521
+ },
+ {
+ "epoch": 35.02794759825328,
+ "grad_norm": 0.0173867866396904,
+ "learning_rate": 0.0006,
+ "loss": 4.091726303100586,
+ "step": 2522
+ },
+ {
+ "epoch": 35.041921397379916,
+ "grad_norm": 0.019272832199931145,
+ "learning_rate": 0.0006,
+ "loss": 4.217905044555664,
+ "step": 2523
+ },
+ {
+ "epoch": 35.05589519650655,
+ "grad_norm": 0.021820371970534325,
+ "learning_rate": 0.0006,
+ "loss": 4.111428260803223,
+ "step": 2524
+ },
+ {
+ "epoch": 35.069868995633186,
+ "grad_norm": 0.02340741828083992,
+ "learning_rate": 0.0006,
+ "loss": 4.115771293640137,
+ "step": 2525
+ },
+ {
+ "epoch": 35.083842794759825,
+ "grad_norm": 0.023760691285133362,
+ "learning_rate": 0.0006,
+ "loss": 4.132383346557617,
+ "step": 2526
+ },
+ {
+ "epoch": 35.09781659388646,
+ "grad_norm": 0.02073468267917633,
+ "learning_rate": 0.0006,
+ "loss": 4.272691249847412,
+ "step": 2527
+ },
+ {
+ "epoch": 35.1117903930131,
+ "grad_norm": 0.02197718247771263,
+ "learning_rate": 0.0006,
+ "loss": 4.105586528778076,
+ "step": 2528
+ },
+ {
+ "epoch": 35.12576419213974,
+ "grad_norm": 0.0200254675000906,
+ "learning_rate": 0.0006,
+ "loss": 3.968820095062256,
+ "step": 2529
+ },
+ {
+ "epoch": 35.13973799126637,
+ "grad_norm": 0.01929587684571743,
+ "learning_rate": 0.0006,
+ "loss": 4.045468807220459,
+ "step": 2530
+ },
+ {
+ "epoch": 35.15371179039301,
+ "grad_norm": 0.01886162906885147,
+ "learning_rate": 0.0006,
+ "loss": 4.185571670532227,
+ "step": 2531
+ },
+ {
+ "epoch": 35.16768558951965,
+ "grad_norm": 0.019789839163422585,
+ "learning_rate": 0.0006,
+ "loss": 3.9734551906585693,
+ "step": 2532
+ },
+ {
+ "epoch": 35.18165938864629,
+ "grad_norm": 0.0198142658919096,
+ "learning_rate": 0.0006,
+ "loss": 4.2662506103515625,
+ "step": 2533
+ },
+ {
+ "epoch": 35.19563318777293,
+ "grad_norm": 0.02087012305855751,
+ "learning_rate": 0.0006,
+ "loss": 4.175931930541992,
+ "step": 2534
+ },
+ {
+ "epoch": 35.209606986899566,
+ "grad_norm": 0.020294535905122757,
+ "learning_rate": 0.0006,
+ "loss": 4.060324668884277,
+ "step": 2535
+ },
+ {
+ "epoch": 35.223580786026204,
+ "grad_norm": 0.01908983662724495,
+ "learning_rate": 0.0006,
+ "loss": 4.217939853668213,
+ "step": 2536
+ },
+ {
+ "epoch": 35.237554585152836,
+ "grad_norm": 0.0178892333060503,
+ "learning_rate": 0.0006,
+ "loss": 4.190216541290283,
+ "step": 2537
+ },
+ {
+ "epoch": 35.251528384279474,
+ "grad_norm": 0.017892315983772278,
+ "learning_rate": 0.0006,
+ "loss": 4.067419052124023,
+ "step": 2538
+ },
+ {
+ "epoch": 35.26550218340611,
+ "grad_norm": 0.018556250259280205,
+ "learning_rate": 0.0006,
+ "loss": 4.105958461761475,
+ "step": 2539
+ },
+ {
+ "epoch": 35.27947598253275,
+ "grad_norm": 0.01944422349333763,
+ "learning_rate": 0.0006,
+ "loss": 4.155775547027588,
+ "step": 2540
+ },
+ {
+ "epoch": 35.29344978165939,
+ "grad_norm": 0.017333146184682846,
+ "learning_rate": 0.0006,
+ "loss": 4.244513511657715,
+ "step": 2541
+ },
+ {
+ "epoch": 35.30742358078603,
+ "grad_norm": 0.01639244332909584,
+ "learning_rate": 0.0006,
+ "loss": 4.167815685272217,
+ "step": 2542
+ },
+ {
+ "epoch": 35.32139737991266,
+ "grad_norm": 0.01610143668949604,
+ "learning_rate": 0.0006,
+ "loss": 4.121486663818359,
+ "step": 2543
+ },
+ {
+ "epoch": 35.3353711790393,
+ "grad_norm": 0.01656411588191986,
+ "learning_rate": 0.0006,
+ "loss": 4.1755499839782715,
+ "step": 2544
+ },
+ {
+ "epoch": 35.34934497816594,
+ "grad_norm": 0.015075297094881535,
+ "learning_rate": 0.0006,
+ "loss": 4.102992534637451,
+ "step": 2545
+ },
+ {
+ "epoch": 35.36331877729258,
+ "grad_norm": 0.01538170501589775,
+ "learning_rate": 0.0006,
+ "loss": 4.161603927612305,
+ "step": 2546
+ },
+ {
+ "epoch": 35.377292576419215,
+ "grad_norm": 0.01483312901109457,
+ "learning_rate": 0.0006,
+ "loss": 4.085355758666992,
+ "step": 2547
+ },
+ {
+ "epoch": 35.391266375545854,
+ "grad_norm": 0.014894738793373108,
+ "learning_rate": 0.0006,
+ "loss": 4.083585262298584,
+ "step": 2548
+ },
+ {
+ "epoch": 35.40524017467249,
+ "grad_norm": 0.01475780550390482,
+ "learning_rate": 0.0006,
+ "loss": 4.012645721435547,
+ "step": 2549
+ },
+ {
+ "epoch": 35.419213973799124,
+ "grad_norm": 0.014531653374433517,
+ "learning_rate": 0.0006,
+ "loss": 4.197113990783691,
+ "step": 2550
+ },
+ {
+ "epoch": 35.43318777292576,
+ "grad_norm": 0.015316493809223175,
+ "learning_rate": 0.0006,
+ "loss": 4.227970123291016,
+ "step": 2551
+ },
+ {
+ "epoch": 35.4471615720524,
+ "grad_norm": 0.014685571193695068,
+ "learning_rate": 0.0006,
+ "loss": 4.089415550231934,
+ "step": 2552
+ },
+ {
+ "epoch": 35.46113537117904,
+ "grad_norm": 0.015337035991251469,
+ "learning_rate": 0.0006,
+ "loss": 4.080583572387695,
+ "step": 2553
+ },
+ {
+ "epoch": 35.47510917030568,
+ "grad_norm": 0.016754208132624626,
+ "learning_rate": 0.0006,
+ "loss": 4.163935661315918,
+ "step": 2554
+ },
+ {
+ "epoch": 35.48908296943232,
+ "grad_norm": 0.017535798251628876,
+ "learning_rate": 0.0006,
+ "loss": 4.093027591705322,
+ "step": 2555
+ },
+ {
+ "epoch": 35.50305676855895,
+ "grad_norm": 0.017336975783109665,
+ "learning_rate": 0.0006,
+ "loss": 4.127961158752441,
+ "step": 2556
+ },
+ {
+ "epoch": 35.51703056768559,
+ "grad_norm": 0.016242152079939842,
+ "learning_rate": 0.0006,
+ "loss": 4.155593395233154,
+ "step": 2557
+ },
+ {
+ "epoch": 35.531004366812226,
+ "grad_norm": 0.015121322125196457,
+ "learning_rate": 0.0006,
+ "loss": 3.984004020690918,
+ "step": 2558
+ },
+ {
+ "epoch": 35.544978165938865,
+ "grad_norm": 0.015178884379565716,
+ "learning_rate": 0.0006,
+ "loss": 4.111564636230469,
+ "step": 2559
+ },
+ {
+ "epoch": 35.5589519650655,
+ "grad_norm": 0.014777131378650665,
+ "learning_rate": 0.0006,
+ "loss": 4.041402816772461,
+ "step": 2560
+ },
+ {
+ "epoch": 35.57292576419214,
+ "grad_norm": 0.015197676606476307,
+ "learning_rate": 0.0006,
+ "loss": 4.133706092834473,
+ "step": 2561
+ },
+ {
+ "epoch": 35.58689956331878,
+ "grad_norm": 0.015905972570180893,
+ "learning_rate": 0.0006,
+ "loss": 4.1069135665893555,
+ "step": 2562
+ },
+ {
+ "epoch": 35.60087336244541,
+ "grad_norm": 0.015942877158522606,
+ "learning_rate": 0.0006,
+ "loss": 4.0720624923706055,
+ "step": 2563
+ },
+ {
+ "epoch": 35.61484716157205,
+ "grad_norm": 0.016697583720088005,
+ "learning_rate": 0.0006,
+ "loss": 4.167448043823242,
+ "step": 2564
+ },
+ {
+ "epoch": 35.62882096069869,
+ "grad_norm": 0.015603577718138695,
+ "learning_rate": 0.0006,
+ "loss": 4.127327919006348,
+ "step": 2565
+ },
+ {
+ "epoch": 35.64279475982533,
+ "grad_norm": 0.017153460532426834,
+ "learning_rate": 0.0006,
+ "loss": 4.056543350219727,
+ "step": 2566
+ },
+ {
+ "epoch": 35.65676855895197,
+ "grad_norm": 0.016731958836317062,
+ "learning_rate": 0.0006,
+ "loss": 4.0681843757629395,
+ "step": 2567
+ },
+ {
+ "epoch": 35.670742358078606,
+ "grad_norm": 0.016167359426617622,
+ "learning_rate": 0.0006,
+ "loss": 4.174992084503174,
+ "step": 2568
+ },
+ {
+ "epoch": 35.68471615720524,
+ "grad_norm": 0.01615108922123909,
+ "learning_rate": 0.0006,
+ "loss": 4.1383161544799805,
+ "step": 2569
+ },
+ {
+ "epoch": 35.698689956331876,
+ "grad_norm": 0.015750816091895103,
+ "learning_rate": 0.0006,
+ "loss": 4.031595706939697,
+ "step": 2570
+ },
+ {
+ "epoch": 35.712663755458514,
+ "grad_norm": 0.015416189096868038,
+ "learning_rate": 0.0006,
+ "loss": 4.067498207092285,
+ "step": 2571
+ },
+ {
+ "epoch": 35.72663755458515,
+ "grad_norm": 0.01570792682468891,
+ "learning_rate": 0.0006,
+ "loss": 4.1338300704956055,
+ "step": 2572
+ },
+ {
+ "epoch": 35.74061135371179,
+ "grad_norm": 0.016367752104997635,
+ "learning_rate": 0.0006,
+ "loss": 4.152523040771484,
+ "step": 2573
+ },
+ {
+ "epoch": 35.75458515283843,
+ "grad_norm": 0.016349803656339645,
+ "learning_rate": 0.0006,
+ "loss": 4.101985454559326,
+ "step": 2574
+ },
+ {
+ "epoch": 35.76855895196506,
+ "grad_norm": 0.01663660630583763,
+ "learning_rate": 0.0006,
+ "loss": 4.190130710601807,
+ "step": 2575
+ },
+ {
+ "epoch": 35.7825327510917,
+ "grad_norm": 0.015599170699715614,
+ "learning_rate": 0.0006,
+ "loss": 4.0792436599731445,
+ "step": 2576
+ },
+ {
+ "epoch": 35.79650655021834,
+ "grad_norm": 0.015472427010536194,
+ "learning_rate": 0.0006,
+ "loss": 4.186446666717529,
+ "step": 2577
+ },
+ {
+ "epoch": 35.81048034934498,
+ "grad_norm": 0.015520439483225346,
+ "learning_rate": 0.0006,
+ "loss": 4.246635437011719,
+ "step": 2578
+ },
+ {
+ "epoch": 35.82445414847162,
+ "grad_norm": 0.016998281702399254,
+ "learning_rate": 0.0006,
+ "loss": 4.13435173034668,
+ "step": 2579
+ },
+ {
+ "epoch": 35.838427947598255,
+ "grad_norm": 0.016290361061692238,
+ "learning_rate": 0.0006,
+ "loss": 4.107428073883057,
+ "step": 2580
+ },
+ {
+ "epoch": 35.852401746724894,
+ "grad_norm": 0.0165516659617424,
+ "learning_rate": 0.0006,
+ "loss": 4.1612138748168945,
+ "step": 2581
+ },
+ {
+ "epoch": 35.866375545851525,
+ "grad_norm": 0.01736091636121273,
+ "learning_rate": 0.0006,
+ "loss": 4.1072869300842285,
+ "step": 2582
+ },
+ {
+ "epoch": 35.880349344978164,
+ "grad_norm": 0.01787031814455986,
+ "learning_rate": 0.0006,
+ "loss": 4.1076812744140625,
+ "step": 2583
+ },
+ {
+ "epoch": 35.8943231441048,
+ "grad_norm": 0.015437985770404339,
+ "learning_rate": 0.0006,
+ "loss": 4.097488880157471,
+ "step": 2584
+ },
+ {
+ "epoch": 35.90829694323144,
+ "grad_norm": 0.016306741163134575,
+ "learning_rate": 0.0006,
+ "loss": 4.105017185211182,
+ "step": 2585
+ },
+ {
+ "epoch": 35.92227074235808,
+ "grad_norm": 0.015723245218396187,
+ "learning_rate": 0.0006,
+ "loss": 4.205283164978027,
+ "step": 2586
+ },
+ {
+ "epoch": 35.93624454148472,
+ "grad_norm": 0.014626456424593925,
+ "learning_rate": 0.0006,
+ "loss": 4.086552619934082,
+ "step": 2587
+ },
+ {
+ "epoch": 35.95021834061135,
+ "grad_norm": 0.014348454773426056,
+ "learning_rate": 0.0006,
+ "loss": 4.171329975128174,
+ "step": 2588
+ },
+ {
+ "epoch": 35.96419213973799,
+ "grad_norm": 0.015598480589687824,
+ "learning_rate": 0.0006,
+ "loss": 4.140815258026123,
+ "step": 2589
+ },
+ {
+ "epoch": 35.97816593886463,
+ "grad_norm": 0.017126062884926796,
+ "learning_rate": 0.0006,
+ "loss": 4.157176971435547,
+ "step": 2590
+ },
+ {
+ "epoch": 35.992139737991266,
+ "grad_norm": 0.01860199309885502,
+ "learning_rate": 0.0006,
+ "loss": 4.252880096435547,
+ "step": 2591
+ },
+ {
+ "epoch": 36.0,
+ "grad_norm": 0.021780220791697502,
+ "learning_rate": 0.0006,
+ "loss": 4.249861717224121,
+ "step": 2592
+ },
+ {
+ "epoch": 36.0,
+ "eval_loss": 4.606466293334961,
+ "eval_runtime": 56.6217,
+ "eval_samples_per_second": 43.128,
+ "eval_steps_per_second": 1.36,
+ "step": 2592
+ },
+ {
+ "epoch": 36.01397379912664,
+ "grad_norm": 0.020892828702926636,
+ "learning_rate": 0.0006,
+ "loss": 4.102055549621582,
+ "step": 2593
+ },
+ {
+ "epoch": 36.02794759825328,
+ "grad_norm": 0.020306255668401718,
+ "learning_rate": 0.0006,
+ "loss": 4.184308052062988,
+ "step": 2594
+ },
+ {
+ "epoch": 36.041921397379916,
+ "grad_norm": 0.02025097794830799,
+ "learning_rate": 0.0006,
+ "loss": 3.9691858291625977,
+ "step": 2595
+ },
+ {
+ "epoch": 36.05589519650655,
+ "grad_norm": 0.020254382863640785,
+ "learning_rate": 0.0006,
+ "loss": 4.02957820892334,
+ "step": 2596
+ },
+ {
+ "epoch": 36.069868995633186,
+ "grad_norm": 0.019307196140289307,
+ "learning_rate": 0.0006,
+ "loss": 4.033181667327881,
+ "step": 2597
+ },
+ {
+ "epoch": 36.083842794759825,
+ "grad_norm": 0.0198476854711771,
+ "learning_rate": 0.0006,
+ "loss": 4.0816874504089355,
+ "step": 2598
+ },
+ {
+ "epoch": 36.09781659388646,
+ "grad_norm": 0.019916515797376633,
+ "learning_rate": 0.0006,
+ "loss": 4.164889812469482,
+ "step": 2599
+ },
+ {
+ "epoch": 36.1117903930131,
+ "grad_norm": 0.0189144778996706,
+ "learning_rate": 0.0006,
+ "loss": 4.038427352905273,
+ "step": 2600
+ },
+ {
+ "epoch": 36.12576419213974,
+ "grad_norm": 0.01754768192768097,
+ "learning_rate": 0.0006,
+ "loss": 4.010427474975586,
+ "step": 2601
+ },
+ {
+ "epoch": 36.13973799126637,
+ "grad_norm": 0.016862990334630013,
+ "learning_rate": 0.0006,
+ "loss": 4.035970211029053,
+ "step": 2602
+ },
+ {
+ "epoch": 36.15371179039301,
+ "grad_norm": 0.019616752862930298,
+ "learning_rate": 0.0006,
+ "loss": 4.079898834228516,
+ "step": 2603
+ },
+ {
+ "epoch": 36.16768558951965,
+ "grad_norm": 0.020131465047597885,
+ "learning_rate": 0.0006,
+ "loss": 4.144684791564941,
+ "step": 2604
+ },
+ {
+ "epoch": 36.18165938864629,
+ "grad_norm": 0.02072521299123764,
+ "learning_rate": 0.0006,
+ "loss": 4.013660430908203,
+ "step": 2605
+ },
+ {
+ "epoch": 36.19563318777293,
+ "grad_norm": 0.01924615167081356,
+ "learning_rate": 0.0006,
+ "loss": 4.071398735046387,
+ "step": 2606
+ },
+ {
+ "epoch": 36.209606986899566,
+ "grad_norm": 0.02013954520225525,
+ "learning_rate": 0.0006,
+ "loss": 4.127094268798828,
+ "step": 2607
+ },
+ {
+ "epoch": 36.223580786026204,
+ "grad_norm": 0.02042008936405182,
+ "learning_rate": 0.0006,
+ "loss": 4.154261589050293,
+ "step": 2608
+ },
+ {
+ "epoch": 36.237554585152836,
+ "grad_norm": 0.021998876705765724,
+ "learning_rate": 0.0006,
+ "loss": 4.204150199890137,
+ "step": 2609
+ },
+ {
+ "epoch": 36.251528384279474,
+ "grad_norm": 0.022946683689951897,
+ "learning_rate": 0.0006,
+ "loss": 4.013632774353027,
+ "step": 2610
+ },
+ {
+ "epoch": 36.26550218340611,
+ "grad_norm": 0.02061830274760723,
+ "learning_rate": 0.0006,
+ "loss": 4.171084403991699,
+ "step": 2611
+ },
+ {
+ "epoch": 36.27947598253275,
+ "grad_norm": 0.01883152686059475,
+ "learning_rate": 0.0006,
+ "loss": 4.149848461151123,
+ "step": 2612
+ },
+ {
+ "epoch": 36.29344978165939,
+ "grad_norm": 0.019945282489061356,
+ "learning_rate": 0.0006,
+ "loss": 4.174656867980957,
+ "step": 2613
+ },
+ {
+ "epoch": 36.30742358078603,
+ "grad_norm": 0.020751051604747772,
+ "learning_rate": 0.0006,
+ "loss": 4.080747127532959,
+ "step": 2614
+ },
+ {
+ "epoch": 36.32139737991266,
+ "grad_norm": 0.01879691146314144,
+ "learning_rate": 0.0006,
+ "loss": 4.057356357574463,
+ "step": 2615
+ },
+ {
+ "epoch": 36.3353711790393,
+ "grad_norm": 0.019037071615457535,
+ "learning_rate": 0.0006,
+ "loss": 4.204607009887695,
+ "step": 2616
+ },
+ {
+ "epoch": 36.34934497816594,
+ "grad_norm": 0.01872078701853752,
+ "learning_rate": 0.0006,
+ "loss": 4.050586700439453,
+ "step": 2617
+ },
+ {
+ "epoch": 36.36331877729258,
+ "grad_norm": 0.018092811107635498,
+ "learning_rate": 0.0006,
+ "loss": 4.102832794189453,
+ "step": 2618
+ },
+ {
+ "epoch": 36.377292576419215,
+ "grad_norm": 0.017010487616062164,
+ "learning_rate": 0.0006,
+ "loss": 4.191788673400879,
+ "step": 2619
+ },
+ {
+ "epoch": 36.391266375545854,
+ "grad_norm": 0.01691143773496151,
+ "learning_rate": 0.0006,
+ "loss": 4.160456657409668,
+ "step": 2620
+ },
+ {
+ "epoch": 36.40524017467249,
+ "grad_norm": 0.017046816647052765,
+ "learning_rate": 0.0006,
+ "loss": 4.073828220367432,
+ "step": 2621
+ },
+ {
+ "epoch": 36.419213973799124,
+ "grad_norm": 0.01707981713116169,
+ "learning_rate": 0.0006,
+ "loss": 4.0980095863342285,
+ "step": 2622
+ },
+ {
+ "epoch": 36.43318777292576,
+ "grad_norm": 0.01630452275276184,
+ "learning_rate": 0.0006,
+ "loss": 4.101321220397949,
+ "step": 2623
+ },
+ {
+ "epoch": 36.4471615720524,
+ "grad_norm": 0.01576915942132473,
+ "learning_rate": 0.0006,
+ "loss": 4.158249855041504,
+ "step": 2624
+ },
+ {
+ "epoch": 36.46113537117904,
+ "grad_norm": 0.016179712489247322,
+ "learning_rate": 0.0006,
+ "loss": 4.077852249145508,
+ "step": 2625
+ },
+ {
+ "epoch": 36.47510917030568,
+ "grad_norm": 0.01609169878065586,
+ "learning_rate": 0.0006,
+ "loss": 4.187317848205566,
+ "step": 2626
+ },
+ {
+ "epoch": 36.48908296943232,
+ "grad_norm": 0.017148254439234734,
+ "learning_rate": 0.0006,
+ "loss": 4.109071731567383,
+ "step": 2627
+ },
+ {
+ "epoch": 36.50305676855895,
+ "grad_norm": 0.015842624008655548,
+ "learning_rate": 0.0006,
+ "loss": 4.145096302032471,
+ "step": 2628
+ },
+ {
+ "epoch": 36.51703056768559,
+ "grad_norm": 0.014906194061040878,
+ "learning_rate": 0.0006,
+ "loss": 4.028440475463867,
+ "step": 2629
+ },
+ {
+ "epoch": 36.531004366812226,
+ "grad_norm": 0.01485038734972477,
+ "learning_rate": 0.0006,
+ "loss": 4.175201416015625,
+ "step": 2630
+ },
+ {
+ "epoch": 36.544978165938865,
+ "grad_norm": 0.015834150835871696,
+ "learning_rate": 0.0006,
+ "loss": 4.076434135437012,
+ "step": 2631
+ },
+ {
+ "epoch": 36.5589519650655,
+ "grad_norm": 0.015411358326673508,
+ "learning_rate": 0.0006,
+ "loss": 4.03912878036499,
+ "step": 2632
+ },
+ {
+ "epoch": 36.57292576419214,
+ "grad_norm": 0.014762846753001213,
+ "learning_rate": 0.0006,
+ "loss": 4.019827365875244,
+ "step": 2633
+ },
+ {
+ "epoch": 36.58689956331878,
+ "grad_norm": 0.015269497409462929,
+ "learning_rate": 0.0006,
+ "loss": 4.116957187652588,
+ "step": 2634
+ },
+ {
+ "epoch": 36.60087336244541,
+ "grad_norm": 0.015171617269515991,
+ "learning_rate": 0.0006,
+ "loss": 4.2069926261901855,
+ "step": 2635
+ },
+ {
+ "epoch": 36.61484716157205,
+ "grad_norm": 0.014770902693271637,
+ "learning_rate": 0.0006,
+ "loss": 4.092021942138672,
+ "step": 2636
+ },
+ {
+ "epoch": 36.62882096069869,
+ "grad_norm": 0.014957531355321407,
+ "learning_rate": 0.0006,
+ "loss": 4.16819953918457,
+ "step": 2637
+ },
+ {
+ "epoch": 36.64279475982533,
+ "grad_norm": 0.014774451963603497,
+ "learning_rate": 0.0006,
+ "loss": 4.046073913574219,
+ "step": 2638
+ },
+ {
+ "epoch": 36.65676855895197,
+ "grad_norm": 0.014320445246994495,
+ "learning_rate": 0.0006,
+ "loss": 3.977792501449585,
+ "step": 2639
+ },
+ {
+ "epoch": 36.670742358078606,
+ "grad_norm": 0.01413130946457386,
+ "learning_rate": 0.0006,
+ "loss": 4.130649566650391,
+ "step": 2640
+ },
+ {
+ "epoch": 36.68471615720524,
+ "grad_norm": 0.014336307533085346,
+ "learning_rate": 0.0006,
+ "loss": 4.161357402801514,
+ "step": 2641
+ },
+ {
+ "epoch": 36.698689956331876,
+ "grad_norm": 0.014516931027173996,
+ "learning_rate": 0.0006,
+ "loss": 4.14248514175415,
+ "step": 2642
+ },
+ {
+ "epoch": 36.712663755458514,
+ "grad_norm": 0.014377126470208168,
+ "learning_rate": 0.0006,
+ "loss": 4.062711715698242,
+ "step": 2643
+ },
+ {
+ "epoch": 36.72663755458515,
+ "grad_norm": 0.014332668855786324,
+ "learning_rate": 0.0006,
+ "loss": 4.06423807144165,
+ "step": 2644
+ },
+ {
+ "epoch": 36.74061135371179,
+ "grad_norm": 0.014030571095645428,
+ "learning_rate": 0.0006,
+ "loss": 4.178613662719727,
+ "step": 2645
+ },
+ {
+ "epoch": 36.75458515283843,
+ "grad_norm": 0.014376776292920113,
+ "learning_rate": 0.0006,
+ "loss": 4.073130130767822,
+ "step": 2646
+ },
+ {
+ "epoch": 36.76855895196506,
+ "grad_norm": 0.013827716931700706,
+ "learning_rate": 0.0006,
+ "loss": 4.043869972229004,
+ "step": 2647
+ },
+ {
+ "epoch": 36.7825327510917,
+ "grad_norm": 0.015548762865364552,
+ "learning_rate": 0.0006,
+ "loss": 4.134049415588379,
+ "step": 2648
+ },
+ {
+ "epoch": 36.79650655021834,
+ "grad_norm": 0.017041970044374466,
+ "learning_rate": 0.0006,
+ "loss": 4.205024242401123,
+ "step": 2649
+ },
+ {
+ "epoch": 36.81048034934498,
+ "grad_norm": 0.017638275399804115,
+ "learning_rate": 0.0006,
+ "loss": 4.199924468994141,
+ "step": 2650
+ },
+ {
+ "epoch": 36.82445414847162,
+ "grad_norm": 0.017771627753973007,
+ "learning_rate": 0.0006,
+ "loss": 4.15706729888916,
+ "step": 2651
+ },
+ {
+ "epoch": 36.838427947598255,
+ "grad_norm": 0.015933562070131302,
+ "learning_rate": 0.0006,
+ "loss": 4.220742225646973,
+ "step": 2652
+ },
+ {
+ "epoch": 36.852401746724894,
+ "grad_norm": 0.01668144389986992,
+ "learning_rate": 0.0006,
+ "loss": 4.161426544189453,
+ "step": 2653
+ },
+ {
+ "epoch": 36.866375545851525,
+ "grad_norm": 0.018014680594205856,
+ "learning_rate": 0.0006,
+ "loss": 4.183046817779541,
+ "step": 2654
+ },
+ {
+ "epoch": 36.880349344978164,
+ "grad_norm": 0.018808670341968536,
+ "learning_rate": 0.0006,
+ "loss": 4.105293273925781,
+ "step": 2655
+ },
+ {
+ "epoch": 36.8943231441048,
+ "grad_norm": 0.018003476783633232,
+ "learning_rate": 0.0006,
+ "loss": 4.103171348571777,
+ "step": 2656
+ },
+ {
+ "epoch": 36.90829694323144,
+ "grad_norm": 0.016865914687514305,
+ "learning_rate": 0.0006,
+ "loss": 4.125634670257568,
+ "step": 2657
+ },
+ {
+ "epoch": 36.92227074235808,
+ "grad_norm": 0.018371153622865677,
+ "learning_rate": 0.0006,
+ "loss": 4.192461013793945,
+ "step": 2658
+ },
+ {
+ "epoch": 36.93624454148472,
+ "grad_norm": 0.019045377150177956,
+ "learning_rate": 0.0006,
+ "loss": 4.198105812072754,
+ "step": 2659
+ },
+ {
+ "epoch": 36.95021834061135,
+ "grad_norm": 0.01594168320298195,
+ "learning_rate": 0.0006,
+ "loss": 4.169788360595703,
+ "step": 2660
+ },
+ {
+ "epoch": 36.96419213973799,
+ "grad_norm": 0.015645820647478104,
+ "learning_rate": 0.0006,
+ "loss": 4.047277450561523,
+ "step": 2661
+ },
+ {
+ "epoch": 36.97816593886463,
+ "grad_norm": 0.016313156113028526,
+ "learning_rate": 0.0006,
+ "loss": 4.087392807006836,
+ "step": 2662
+ },
+ {
+ "epoch": 36.992139737991266,
+ "grad_norm": 0.017050625756382942,
+ "learning_rate": 0.0006,
+ "loss": 4.0908002853393555,
+ "step": 2663
+ },
+ {
+ "epoch": 37.0,
+ "grad_norm": 0.017730968073010445,
+ "learning_rate": 0.0006,
+ "loss": 4.110957622528076,
+ "step": 2664
+ },
+ {
+ "epoch": 37.0,
+ "eval_loss": 4.563793182373047,
+ "eval_runtime": 56.5922,
+ "eval_samples_per_second": 43.151,
+ "eval_steps_per_second": 1.361,
+ "step": 2664
+ },
+ {
+ "epoch": 37.01397379912664,
+ "grad_norm": 0.015687910839915276,
+ "learning_rate": 0.0006,
+ "loss": 4.143589019775391,
+ "step": 2665
+ },
+ {
+ "epoch": 37.02794759825328,
+ "grad_norm": 0.017451366409659386,
+ "learning_rate": 0.0006,
+ "loss": 4.0326666831970215,
+ "step": 2666
+ },
+ {
+ "epoch": 37.041921397379916,
+ "grad_norm": 0.01909538544714451,
+ "learning_rate": 0.0006,
+ "loss": 3.9706249237060547,
+ "step": 2667
+ },
+ {
+ "epoch": 37.05589519650655,
+ "grad_norm": 0.01988379657268524,
+ "learning_rate": 0.0006,
+ "loss": 3.9531009197235107,
+ "step": 2668
+ },
+ {
+ "epoch": 37.069868995633186,
+ "grad_norm": 0.02035115472972393,
+ "learning_rate": 0.0006,
+ "loss": 3.962214469909668,
+ "step": 2669
+ },
+ {
+ "epoch": 37.083842794759825,
+ "grad_norm": 0.01952001079916954,
+ "learning_rate": 0.0006,
+ "loss": 4.182262897491455,
+ "step": 2670
+ },
+ {
+ "epoch": 37.09781659388646,
+ "grad_norm": 0.01854325644671917,
+ "learning_rate": 0.0006,
+ "loss": 4.127663612365723,
+ "step": 2671
+ },
+ {
+ "epoch": 37.1117903930131,
+ "grad_norm": 0.017757616937160492,
+ "learning_rate": 0.0006,
+ "loss": 4.100804805755615,
+ "step": 2672
+ },
+ {
+ "epoch": 37.12576419213974,
+ "grad_norm": 0.016712479293346405,
+ "learning_rate": 0.0006,
+ "loss": 4.026284217834473,
+ "step": 2673
+ },
+ {
+ "epoch": 37.13973799126637,
+ "grad_norm": 0.0168139711022377,
+ "learning_rate": 0.0006,
+ "loss": 4.183882713317871,
+ "step": 2674
+ },
+ {
+ "epoch": 37.15371179039301,
+ "grad_norm": 0.017060423269867897,
+ "learning_rate": 0.0006,
+ "loss": 4.131832599639893,
+ "step": 2675
+ },
+ {
+ "epoch": 37.16768558951965,
+ "grad_norm": 0.01726631447672844,
+ "learning_rate": 0.0006,
+ "loss": 3.9156057834625244,
+ "step": 2676
+ },
+ {
+ "epoch": 37.18165938864629,
+ "grad_norm": 0.016348622739315033,
+ "learning_rate": 0.0006,
+ "loss": 4.1731367111206055,
+ "step": 2677
+ },
+ {
+ "epoch": 37.19563318777293,
+ "grad_norm": 0.016640059649944305,
+ "learning_rate": 0.0006,
+ "loss": 4.10997200012207,
+ "step": 2678
+ },
+ {
+ "epoch": 37.209606986899566,
+ "grad_norm": 0.01601342298090458,
+ "learning_rate": 0.0006,
+ "loss": 4.081145286560059,
+ "step": 2679
+ },
+ {
+ "epoch": 37.223580786026204,
+ "grad_norm": 0.01735631190240383,
+ "learning_rate": 0.0006,
+ "loss": 4.202108860015869,
+ "step": 2680
+ },
+ {
+ "epoch": 37.237554585152836,
+ "grad_norm": 0.019164685159921646,
+ "learning_rate": 0.0006,
+ "loss": 4.104109764099121,
+ "step": 2681
+ },
+ {
+ "epoch": 37.251528384279474,
+ "grad_norm": 0.01881759613752365,
+ "learning_rate": 0.0006,
+ "loss": 4.145532131195068,
+ "step": 2682
+ },
+ {
+ "epoch": 37.26550218340611,
+ "grad_norm": 0.017656555399298668,
+ "learning_rate": 0.0006,
+ "loss": 4.012782096862793,
+ "step": 2683
+ },
+ {
+ "epoch": 37.27947598253275,
+ "grad_norm": 0.01624823734164238,
+ "learning_rate": 0.0006,
+ "loss": 4.001727104187012,
+ "step": 2684
+ },
+ {
+ "epoch": 37.29344978165939,
+ "grad_norm": 0.016973139718174934,
+ "learning_rate": 0.0006,
+ "loss": 4.095399856567383,
+ "step": 2685
+ },
+ {
+ "epoch": 37.30742358078603,
+ "grad_norm": 0.015503671951591969,
+ "learning_rate": 0.0006,
+ "loss": 4.034431457519531,
+ "step": 2686
+ },
+ {
+ "epoch": 37.32139737991266,
+ "grad_norm": 0.015370250679552555,
+ "learning_rate": 0.0006,
+ "loss": 4.075836181640625,
+ "step": 2687
+ },
+ {
+ "epoch": 37.3353711790393,
+ "grad_norm": 0.01564021222293377,
+ "learning_rate": 0.0006,
+ "loss": 4.125202178955078,
+ "step": 2688
+ },
+ {
+ "epoch": 37.34934497816594,
+ "grad_norm": 0.016717975959181786,
+ "learning_rate": 0.0006,
+ "loss": 4.153017044067383,
+ "step": 2689
+ },
+ {
+ "epoch": 37.36331877729258,
+ "grad_norm": 0.015263617038726807,
+ "learning_rate": 0.0006,
+ "loss": 4.021814823150635,
+ "step": 2690
+ },
+ {
+ "epoch": 37.377292576419215,
+ "grad_norm": 0.016497142612934113,
+ "learning_rate": 0.0006,
+ "loss": 4.069095611572266,
+ "step": 2691
+ },
+ {
+ "epoch": 37.391266375545854,
+ "grad_norm": 0.01669355481863022,
+ "learning_rate": 0.0006,
+ "loss": 4.174587726593018,
+ "step": 2692
+ },
+ {
+ "epoch": 37.40524017467249,
+ "grad_norm": 0.019695773720741272,
+ "learning_rate": 0.0006,
+ "loss": 4.0073041915893555,
+ "step": 2693
+ },
+ {
+ "epoch": 37.419213973799124,
+ "grad_norm": 0.023720644414424896,
+ "learning_rate": 0.0006,
+ "loss": 4.111948013305664,
+ "step": 2694
+ },
+ {
+ "epoch": 37.43318777292576,
+ "grad_norm": 0.02399350143969059,
+ "learning_rate": 0.0006,
+ "loss": 4.081164360046387,
+ "step": 2695
+ },
+ {
+ "epoch": 37.4471615720524,
+ "grad_norm": 0.019630271941423416,
+ "learning_rate": 0.0006,
+ "loss": 4.065731048583984,
+ "step": 2696
+ },
+ {
+ "epoch": 37.46113537117904,
+ "grad_norm": 0.019679008051753044,
+ "learning_rate": 0.0006,
+ "loss": 3.993852376937866,
+ "step": 2697
+ },
+ {
+ "epoch": 37.47510917030568,
+ "grad_norm": 0.020597299560904503,
+ "learning_rate": 0.0006,
+ "loss": 4.275555610656738,
+ "step": 2698
+ },
+ {
+ "epoch": 37.48908296943232,
+ "grad_norm": 0.01983889564871788,
+ "learning_rate": 0.0006,
+ "loss": 4.254708290100098,
+ "step": 2699
+ },
+ {
+ "epoch": 37.50305676855895,
+ "grad_norm": 0.019150668755173683,
+ "learning_rate": 0.0006,
+ "loss": 4.097885608673096,
+ "step": 2700
+ },
+ {
+ "epoch": 37.51703056768559,
+ "grad_norm": 0.018847715109586716,
+ "learning_rate": 0.0006,
+ "loss": 4.054081439971924,
+ "step": 2701
+ },
+ {
+ "epoch": 37.531004366812226,
+ "grad_norm": 0.019607288762927055,
+ "learning_rate": 0.0006,
+ "loss": 3.9891440868377686,
+ "step": 2702
+ },
+ {
+ "epoch": 37.544978165938865,
+ "grad_norm": 0.018863150849938393,
+ "learning_rate": 0.0006,
+ "loss": 4.084877014160156,
+ "step": 2703
+ },
+ {
+ "epoch": 37.5589519650655,
+ "grad_norm": 0.017340239137411118,
+ "learning_rate": 0.0006,
+ "loss": 4.113290786743164,
+ "step": 2704
+ },
+ {
+ "epoch": 37.57292576419214,
+ "grad_norm": 0.015628693625330925,
+ "learning_rate": 0.0006,
+ "loss": 4.1483001708984375,
+ "step": 2705
+ },
+ {
+ "epoch": 37.58689956331878,
+ "grad_norm": 0.016493283212184906,
+ "learning_rate": 0.0006,
+ "loss": 4.1040143966674805,
+ "step": 2706
+ },
+ {
+ "epoch": 37.60087336244541,
+ "grad_norm": 0.016732443124055862,
+ "learning_rate": 0.0006,
+ "loss": 4.07205057144165,
+ "step": 2707
+ },
+ {
+ "epoch": 37.61484716157205,
+ "grad_norm": 0.017176156863570213,
+ "learning_rate": 0.0006,
+ "loss": 4.1279754638671875,
+ "step": 2708
+ },
+ {
+ "epoch": 37.62882096069869,
+ "grad_norm": 0.015730608254671097,
+ "learning_rate": 0.0006,
+ "loss": 4.086658000946045,
+ "step": 2709
+ },
+ {
+ "epoch": 37.64279475982533,
+ "grad_norm": 0.016111090779304504,
+ "learning_rate": 0.0006,
+ "loss": 4.064335823059082,
+ "step": 2710
+ },
+ {
+ "epoch": 37.65676855895197,
+ "grad_norm": 0.01769379712641239,
+ "learning_rate": 0.0006,
+ "loss": 4.116047382354736,
+ "step": 2711
+ },
+ {
+ "epoch": 37.670742358078606,
+ "grad_norm": 0.01683102548122406,
+ "learning_rate": 0.0006,
+ "loss": 4.095986843109131,
+ "step": 2712
+ },
+ {
+ "epoch": 37.68471615720524,
+ "grad_norm": 0.015726450830698013,
+ "learning_rate": 0.0006,
+ "loss": 4.10775089263916,
+ "step": 2713
+ },
+ {
+ "epoch": 37.698689956331876,
+ "grad_norm": 0.014760235324501991,
+ "learning_rate": 0.0006,
+ "loss": 4.1234893798828125,
+ "step": 2714
+ },
+ {
+ "epoch": 37.712663755458514,
+ "grad_norm": 0.014933438040316105,
+ "learning_rate": 0.0006,
+ "loss": 4.079200744628906,
+ "step": 2715
+ },
+ {
+ "epoch": 37.72663755458515,
+ "grad_norm": 0.015723997727036476,
+ "learning_rate": 0.0006,
+ "loss": 4.07865047454834,
+ "step": 2716
+ },
+ {
+ "epoch": 37.74061135371179,
+ "grad_norm": 0.014539001509547234,
+ "learning_rate": 0.0006,
+ "loss": 4.036242485046387,
+ "step": 2717
+ },
+ {
+ "epoch": 37.75458515283843,
+ "grad_norm": 0.014844979159533978,
+ "learning_rate": 0.0006,
+ "loss": 4.1051483154296875,
+ "step": 2718
+ },
+ {
+ "epoch": 37.76855895196506,
+ "grad_norm": 0.015939844772219658,
+ "learning_rate": 0.0006,
+ "loss": 4.017253875732422,
+ "step": 2719
+ },
+ {
+ "epoch": 37.7825327510917,
+ "grad_norm": 0.014771555550396442,
+ "learning_rate": 0.0006,
+ "loss": 4.0694661140441895,
+ "step": 2720
+ },
+ {
+ "epoch": 37.79650655021834,
+ "grad_norm": 0.015777921304106712,
+ "learning_rate": 0.0006,
+ "loss": 3.947587490081787,
+ "step": 2721
+ },
+ {
+ "epoch": 37.81048034934498,
+ "grad_norm": 0.017451291903853416,
+ "learning_rate": 0.0006,
+ "loss": 4.093625068664551,
+ "step": 2722
+ },
+ {
+ "epoch": 37.82445414847162,
+ "grad_norm": 0.016829246655106544,
+ "learning_rate": 0.0006,
+ "loss": 4.09030818939209,
+ "step": 2723
+ },
+ {
+ "epoch": 37.838427947598255,
+ "grad_norm": 0.016984397545456886,
+ "learning_rate": 0.0006,
+ "loss": 4.140626907348633,
+ "step": 2724
+ },
+ {
+ "epoch": 37.852401746724894,
+ "grad_norm": 0.01732603646814823,
+ "learning_rate": 0.0006,
+ "loss": 4.128276824951172,
+ "step": 2725
+ },
+ {
+ "epoch": 37.866375545851525,
+ "grad_norm": 0.01744619384407997,
+ "learning_rate": 0.0006,
+ "loss": 4.103130340576172,
+ "step": 2726
+ },
+ {
+ "epoch": 37.880349344978164,
+ "grad_norm": 0.016331320628523827,
+ "learning_rate": 0.0006,
+ "loss": 4.037951469421387,
+ "step": 2727
+ },
+ {
+ "epoch": 37.8943231441048,
+ "grad_norm": 0.015919430181384087,
+ "learning_rate": 0.0006,
+ "loss": 4.121513366699219,
+ "step": 2728
+ },
+ {
+ "epoch": 37.90829694323144,
+ "grad_norm": 0.016153695061802864,
+ "learning_rate": 0.0006,
+ "loss": 4.099384307861328,
+ "step": 2729
+ },
+ {
+ "epoch": 37.92227074235808,
+ "grad_norm": 0.01628800667822361,
+ "learning_rate": 0.0006,
+ "loss": 4.092325687408447,
+ "step": 2730
+ },
+ {
+ "epoch": 37.93624454148472,
+ "grad_norm": 0.017285684123635292,
+ "learning_rate": 0.0006,
+ "loss": 3.9828906059265137,
+ "step": 2731
+ },
+ {
+ "epoch": 37.95021834061135,
+ "grad_norm": 0.016047554090619087,
+ "learning_rate": 0.0006,
+ "loss": 4.100419044494629,
+ "step": 2732
+ },
+ {
+ "epoch": 37.96419213973799,
+ "grad_norm": 0.01562986709177494,
+ "learning_rate": 0.0006,
+ "loss": 4.065832614898682,
+ "step": 2733
+ },
+ {
+ "epoch": 37.97816593886463,
+ "grad_norm": 0.01677163504064083,
+ "learning_rate": 0.0006,
+ "loss": 4.077518463134766,
+ "step": 2734
+ },
+ {
+ "epoch": 37.992139737991266,
+ "grad_norm": 0.01755492016673088,
+ "learning_rate": 0.0006,
+ "loss": 4.1076154708862305,
+ "step": 2735
+ },
+ {
+ "epoch": 38.0,
+ "grad_norm": 0.01929762214422226,
+ "learning_rate": 0.0006,
+ "loss": 4.09603214263916,
+ "step": 2736
+ },
+ {
+ "epoch": 38.0,
+ "eval_loss": 4.582442283630371,
+ "eval_runtime": 62.8649,
+ "eval_samples_per_second": 38.845,
+ "eval_steps_per_second": 1.225,
+ "step": 2736
+ },
+ {
+ "epoch": 38.01397379912664,
+ "grad_norm": 0.019129136577248573,
+ "learning_rate": 0.0006,
+ "loss": 4.095005989074707,
+ "step": 2737
+ },
+ {
+ "epoch": 38.02794759825328,
+ "grad_norm": 0.019167408347129822,
+ "learning_rate": 0.0006,
+ "loss": 4.054426670074463,
+ "step": 2738
+ },
+ {
+ "epoch": 38.041921397379916,
+ "grad_norm": 0.01789460889995098,
+ "learning_rate": 0.0006,
+ "loss": 4.0434417724609375,
+ "step": 2739
+ },
+ {
+ "epoch": 38.05589519650655,
+ "grad_norm": 0.015947027131915092,
+ "learning_rate": 0.0006,
+ "loss": 4.074516296386719,
+ "step": 2740
+ },
+ {
+ "epoch": 38.069868995633186,
+ "grad_norm": 0.016667252406477928,
+ "learning_rate": 0.0006,
+ "loss": 4.108545303344727,
+ "step": 2741
+ },
+ {
+ "epoch": 38.083842794759825,
+ "grad_norm": 0.016014399006962776,
+ "learning_rate": 0.0006,
+ "loss": 4.026925086975098,
+ "step": 2742
+ },
+ {
+ "epoch": 38.09781659388646,
+ "grad_norm": 0.015397695824503899,
+ "learning_rate": 0.0006,
+ "loss": 4.163970470428467,
+ "step": 2743
+ },
+ {
+ "epoch": 38.1117903930131,
+ "grad_norm": 0.01563289947807789,
+ "learning_rate": 0.0006,
+ "loss": 4.116578578948975,
+ "step": 2744
+ },
+ {
+ "epoch": 38.12576419213974,
+ "grad_norm": 0.015512318350374699,
+ "learning_rate": 0.0006,
+ "loss": 4.096311569213867,
+ "step": 2745
+ },
+ {
+ "epoch": 38.13973799126637,
+ "grad_norm": 0.015426642261445522,
+ "learning_rate": 0.0006,
+ "loss": 4.017065048217773,
+ "step": 2746
+ },
+ {
+ "epoch": 38.15371179039301,
+ "grad_norm": 0.015115435235202312,
+ "learning_rate": 0.0006,
+ "loss": 4.005434989929199,
+ "step": 2747
+ },
+ {
+ "epoch": 38.16768558951965,
+ "grad_norm": 0.015235894359648228,
+ "learning_rate": 0.0006,
+ "loss": 3.982743263244629,
+ "step": 2748
+ },
+ {
+ "epoch": 38.18165938864629,
+ "grad_norm": 0.016095101833343506,
+ "learning_rate": 0.0006,
+ "loss": 4.0849385261535645,
+ "step": 2749
+ },
+ {
+ "epoch": 38.19563318777293,
+ "grad_norm": 0.015201197937130928,
+ "learning_rate": 0.0006,
+ "loss": 4.103863716125488,
+ "step": 2750
+ },
+ {
+ "epoch": 38.209606986899566,
+ "grad_norm": 0.015687478706240654,
+ "learning_rate": 0.0006,
+ "loss": 4.12074089050293,
+ "step": 2751
+ },
+ {
+ "epoch": 38.223580786026204,
+ "grad_norm": 0.015201589092612267,
+ "learning_rate": 0.0006,
+ "loss": 4.056872367858887,
+ "step": 2752
+ },
+ {
+ "epoch": 38.237554585152836,
+ "grad_norm": 0.015154894441366196,
+ "learning_rate": 0.0006,
+ "loss": 4.089670181274414,
+ "step": 2753
+ },
+ {
+ "epoch": 38.251528384279474,
+ "grad_norm": 0.01586608588695526,
+ "learning_rate": 0.0006,
+ "loss": 4.155317306518555,
+ "step": 2754
+ },
+ {
+ "epoch": 38.26550218340611,
+ "grad_norm": 0.01609654724597931,
+ "learning_rate": 0.0006,
+ "loss": 3.995800495147705,
+ "step": 2755
+ },
+ {
+ "epoch": 38.27947598253275,
+ "grad_norm": 0.015495185740292072,
+ "learning_rate": 0.0006,
+ "loss": 3.995769739151001,
+ "step": 2756
+ },
+ {
+ "epoch": 38.29344978165939,
+ "grad_norm": 0.017042381688952446,
+ "learning_rate": 0.0006,
+ "loss": 4.132832050323486,
+ "step": 2757
+ },
+ {
+ "epoch": 38.30742358078603,
+ "grad_norm": 0.018706614151597023,
+ "learning_rate": 0.0006,
+ "loss": 4.038820743560791,
+ "step": 2758
+ },
+ {
+ "epoch": 38.32139737991266,
+ "grad_norm": 0.018316026777029037,
+ "learning_rate": 0.0006,
+ "loss": 3.972052574157715,
+ "step": 2759
+ },
+ {
+ "epoch": 38.3353711790393,
+ "grad_norm": 0.018375389277935028,
+ "learning_rate": 0.0006,
+ "loss": 4.152097702026367,
+ "step": 2760
+ },
+ {
+ "epoch": 38.34934497816594,
+ "grad_norm": 0.016879886388778687,
+ "learning_rate": 0.0006,
+ "loss": 4.020869731903076,
+ "step": 2761
+ },
+ {
+ "epoch": 38.36331877729258,
+ "grad_norm": 0.015728183090686798,
+ "learning_rate": 0.0006,
+ "loss": 4.008016586303711,
+ "step": 2762
+ },
+ {
+ "epoch": 38.377292576419215,
+ "grad_norm": 0.016147758811712265,
+ "learning_rate": 0.0006,
+ "loss": 4.035854339599609,
+ "step": 2763
+ },
+ {
+ "epoch": 38.391266375545854,
+ "grad_norm": 0.017774797976017,
+ "learning_rate": 0.0006,
+ "loss": 4.122506141662598,
+ "step": 2764
+ },
+ {
+ "epoch": 38.40524017467249,
+ "grad_norm": 0.016590069979429245,
+ "learning_rate": 0.0006,
+ "loss": 4.082847595214844,
+ "step": 2765
+ },
+ {
+ "epoch": 38.419213973799124,
+ "grad_norm": 0.016544528305530548,
+ "learning_rate": 0.0006,
+ "loss": 4.184270858764648,
+ "step": 2766
+ },
+ {
+ "epoch": 38.43318777292576,
+ "grad_norm": 0.017972925677895546,
+ "learning_rate": 0.0006,
+ "loss": 3.987558603286743,
+ "step": 2767
+ },
+ {
+ "epoch": 38.4471615720524,
+ "grad_norm": 0.018562227487564087,
+ "learning_rate": 0.0006,
+ "loss": 4.1193461418151855,
+ "step": 2768
+ },
+ {
+ "epoch": 38.46113537117904,
+ "grad_norm": 0.018896685913205147,
+ "learning_rate": 0.0006,
+ "loss": 4.057808876037598,
+ "step": 2769
+ },
+ {
+ "epoch": 38.47510917030568,
+ "grad_norm": 0.016793809831142426,
+ "learning_rate": 0.0006,
+ "loss": 4.069523334503174,
+ "step": 2770
+ },
+ {
+ "epoch": 38.48908296943232,
+ "grad_norm": 0.01698676310479641,
+ "learning_rate": 0.0006,
+ "loss": 3.997624158859253,
+ "step": 2771
+ },
+ {
+ "epoch": 38.50305676855895,
+ "grad_norm": 0.016784202307462692,
+ "learning_rate": 0.0006,
+ "loss": 4.174795627593994,
+ "step": 2772
+ },
+ {
+ "epoch": 38.51703056768559,
+ "grad_norm": 0.016785014420747757,
+ "learning_rate": 0.0006,
+ "loss": 3.999321222305298,
+ "step": 2773
+ },
+ {
+ "epoch": 38.531004366812226,
+ "grad_norm": 0.018615184351801872,
+ "learning_rate": 0.0006,
+ "loss": 4.191237926483154,
+ "step": 2774
+ },
+ {
+ "epoch": 38.544978165938865,
+ "grad_norm": 0.019090790301561356,
+ "learning_rate": 0.0006,
+ "loss": 4.106966018676758,
+ "step": 2775
+ },
+ {
+ "epoch": 38.5589519650655,
+ "grad_norm": 0.018298320472240448,
+ "learning_rate": 0.0006,
+ "loss": 4.063363552093506,
+ "step": 2776
+ },
+ {
+ "epoch": 38.57292576419214,
+ "grad_norm": 0.018541090190410614,
+ "learning_rate": 0.0006,
+ "loss": 4.101193428039551,
+ "step": 2777
+ },
+ {
+ "epoch": 38.58689956331878,
+ "grad_norm": 0.01849970407783985,
+ "learning_rate": 0.0006,
+ "loss": 4.071335792541504,
+ "step": 2778
+ },
+ {
+ "epoch": 38.60087336244541,
+ "grad_norm": 0.018309077247977257,
+ "learning_rate": 0.0006,
+ "loss": 4.245305061340332,
+ "step": 2779
+ },
+ {
+ "epoch": 38.61484716157205,
+ "grad_norm": 0.017681840807199478,
+ "learning_rate": 0.0006,
+ "loss": 4.081173419952393,
+ "step": 2780
+ },
+ {
+ "epoch": 38.62882096069869,
+ "grad_norm": 0.016286224126815796,
+ "learning_rate": 0.0006,
+ "loss": 4.024351596832275,
+ "step": 2781
+ },
+ {
+ "epoch": 38.64279475982533,
+ "grad_norm": 0.015837235376238823,
+ "learning_rate": 0.0006,
+ "loss": 4.028205394744873,
+ "step": 2782
+ },
+ {
+ "epoch": 38.65676855895197,
+ "grad_norm": 0.016062596812844276,
+ "learning_rate": 0.0006,
+ "loss": 4.131122589111328,
+ "step": 2783
+ },
+ {
+ "epoch": 38.670742358078606,
+ "grad_norm": 0.01711239665746689,
+ "learning_rate": 0.0006,
+ "loss": 4.124416351318359,
+ "step": 2784
+ },
+ {
+ "epoch": 38.68471615720524,
+ "grad_norm": 0.015304913744330406,
+ "learning_rate": 0.0006,
+ "loss": 4.056122779846191,
+ "step": 2785
+ },
+ {
+ "epoch": 38.698689956331876,
+ "grad_norm": 0.015971887856721878,
+ "learning_rate": 0.0006,
+ "loss": 4.132555961608887,
+ "step": 2786
+ },
+ {
+ "epoch": 38.712663755458514,
+ "grad_norm": 0.015787221491336823,
+ "learning_rate": 0.0006,
+ "loss": 4.051596164703369,
+ "step": 2787
+ },
+ {
+ "epoch": 38.72663755458515,
+ "grad_norm": 0.01605444774031639,
+ "learning_rate": 0.0006,
+ "loss": 4.191808700561523,
+ "step": 2788
+ },
+ {
+ "epoch": 38.74061135371179,
+ "grad_norm": 0.016383009031414986,
+ "learning_rate": 0.0006,
+ "loss": 4.0648417472839355,
+ "step": 2789
+ },
+ {
+ "epoch": 38.75458515283843,
+ "grad_norm": 0.016135554760694504,
+ "learning_rate": 0.0006,
+ "loss": 3.988563060760498,
+ "step": 2790
+ },
+ {
+ "epoch": 38.76855895196506,
+ "grad_norm": 0.016634158790111542,
+ "learning_rate": 0.0006,
+ "loss": 4.133493423461914,
+ "step": 2791
+ },
+ {
+ "epoch": 38.7825327510917,
+ "grad_norm": 0.01688210666179657,
+ "learning_rate": 0.0006,
+ "loss": 4.091277122497559,
+ "step": 2792
+ },
+ {
+ "epoch": 38.79650655021834,
+ "grad_norm": 0.016102122142910957,
+ "learning_rate": 0.0006,
+ "loss": 4.051219940185547,
+ "step": 2793
+ },
+ {
+ "epoch": 38.81048034934498,
+ "grad_norm": 0.016585038974881172,
+ "learning_rate": 0.0006,
+ "loss": 4.058587074279785,
+ "step": 2794
+ },
+ {
+ "epoch": 38.82445414847162,
+ "grad_norm": 0.017700951546430588,
+ "learning_rate": 0.0006,
+ "loss": 4.016030311584473,
+ "step": 2795
+ },
+ {
+ "epoch": 38.838427947598255,
+ "grad_norm": 0.017922203987836838,
+ "learning_rate": 0.0006,
+ "loss": 4.079977989196777,
+ "step": 2796
+ },
+ {
+ "epoch": 38.852401746724894,
+ "grad_norm": 0.017299890518188477,
+ "learning_rate": 0.0006,
+ "loss": 4.062984943389893,
+ "step": 2797
+ },
+ {
+ "epoch": 38.866375545851525,
+ "grad_norm": 0.018648581579327583,
+ "learning_rate": 0.0006,
+ "loss": 4.065717697143555,
+ "step": 2798
+ },
+ {
+ "epoch": 38.880349344978164,
+ "grad_norm": 0.019057920202612877,
+ "learning_rate": 0.0006,
+ "loss": 4.227892875671387,
+ "step": 2799
+ },
+ {
+ "epoch": 38.8943231441048,
+ "grad_norm": 0.018824422731995583,
+ "learning_rate": 0.0006,
+ "loss": 4.116058349609375,
+ "step": 2800
+ },
+ {
+ "epoch": 38.90829694323144,
+ "grad_norm": 0.019084978848695755,
+ "learning_rate": 0.0006,
+ "loss": 4.061677932739258,
+ "step": 2801
+ },
+ {
+ "epoch": 38.92227074235808,
+ "grad_norm": 0.017842981964349747,
+ "learning_rate": 0.0006,
+ "loss": 4.083813667297363,
+ "step": 2802
+ },
+ {
+ "epoch": 38.93624454148472,
+ "grad_norm": 0.01880858838558197,
+ "learning_rate": 0.0006,
+ "loss": 4.152561187744141,
+ "step": 2803
+ },
+ {
+ "epoch": 38.95021834061135,
+ "grad_norm": 0.017400087788701057,
+ "learning_rate": 0.0006,
+ "loss": 4.154635429382324,
+ "step": 2804
+ },
+ {
+ "epoch": 38.96419213973799,
+ "grad_norm": 0.017682507634162903,
+ "learning_rate": 0.0006,
+ "loss": 4.151128768920898,
+ "step": 2805
+ },
+ {
+ "epoch": 38.97816593886463,
+ "grad_norm": 0.015803052112460136,
+ "learning_rate": 0.0006,
+ "loss": 4.10498571395874,
+ "step": 2806
+ },
+ {
+ "epoch": 38.992139737991266,
+ "grad_norm": 0.01582511141896248,
+ "learning_rate": 0.0006,
+ "loss": 4.140159606933594,
+ "step": 2807
+ },
+ {
+ "epoch": 39.0,
+ "grad_norm": 0.01925666816532612,
+ "learning_rate": 0.0006,
+ "loss": 4.126394271850586,
+ "step": 2808
+ },
+ {
+ "epoch": 39.0,
+ "eval_loss": 4.615617275238037,
+ "eval_runtime": 56.8895,
+ "eval_samples_per_second": 42.925,
+ "eval_steps_per_second": 1.354,
+ "step": 2808
+ },
+ {
+ "epoch": 39.01397379912664,
+ "grad_norm": 0.01677853614091873,
+ "learning_rate": 0.0006,
+ "loss": 4.102395057678223,
+ "step": 2809
+ },
+ {
+ "epoch": 39.02794759825328,
+ "grad_norm": 0.017901137471199036,
+ "learning_rate": 0.0006,
+ "loss": 3.976609230041504,
+ "step": 2810
+ },
+ {
+ "epoch": 39.041921397379916,
+ "grad_norm": 0.017978668212890625,
+ "learning_rate": 0.0006,
+ "loss": 3.944416046142578,
+ "step": 2811
+ },
+ {
+ "epoch": 39.05589519650655,
+ "grad_norm": 0.019967492669820786,
+ "learning_rate": 0.0006,
+ "loss": 3.992234468460083,
+ "step": 2812
+ },
+ {
+ "epoch": 39.069868995633186,
+ "grad_norm": 0.018937544897198677,
+ "learning_rate": 0.0006,
+ "loss": 4.028573036193848,
+ "step": 2813
+ },
+ {
+ "epoch": 39.083842794759825,
+ "grad_norm": 0.016852952539920807,
+ "learning_rate": 0.0006,
+ "loss": 4.062442779541016,
+ "step": 2814
+ },
+ {
+ "epoch": 39.09781659388646,
+ "grad_norm": 0.016643647104501724,
+ "learning_rate": 0.0006,
+ "loss": 4.08871603012085,
+ "step": 2815
+ },
+ {
+ "epoch": 39.1117903930131,
+ "grad_norm": 0.016872866079211235,
+ "learning_rate": 0.0006,
+ "loss": 3.9514245986938477,
+ "step": 2816
+ },
+ {
+ "epoch": 39.12576419213974,
+ "grad_norm": 0.018045669421553612,
+ "learning_rate": 0.0006,
+ "loss": 4.038842678070068,
+ "step": 2817
+ },
+ {
+ "epoch": 39.13973799126637,
+ "grad_norm": 0.017853403463959694,
+ "learning_rate": 0.0006,
+ "loss": 3.93709135055542,
+ "step": 2818
+ },
+ {
+ "epoch": 39.15371179039301,
+ "grad_norm": 0.017507443204522133,
+ "learning_rate": 0.0006,
+ "loss": 3.98665714263916,
+ "step": 2819
+ },
+ {
+ "epoch": 39.16768558951965,
+ "grad_norm": 0.01766853965818882,
+ "learning_rate": 0.0006,
+ "loss": 3.937126636505127,
+ "step": 2820
+ },
+ {
+ "epoch": 39.18165938864629,
+ "grad_norm": 0.016660021618008614,
+ "learning_rate": 0.0006,
+ "loss": 3.8397014141082764,
+ "step": 2821
+ },
+ {
+ "epoch": 39.19563318777293,
+ "grad_norm": 0.017719309777021408,
+ "learning_rate": 0.0006,
+ "loss": 4.003579616546631,
+ "step": 2822
+ },
+ {
+ "epoch": 39.209606986899566,
+ "grad_norm": 0.018234414979815483,
+ "learning_rate": 0.0006,
+ "loss": 4.053859710693359,
+ "step": 2823
+ },
+ {
+ "epoch": 39.223580786026204,
+ "grad_norm": 0.018777024000883102,
+ "learning_rate": 0.0006,
+ "loss": 4.056182861328125,
+ "step": 2824
+ },
+ {
+ "epoch": 39.237554585152836,
+ "grad_norm": 0.017443502321839333,
+ "learning_rate": 0.0006,
+ "loss": 4.025068283081055,
+ "step": 2825
+ },
+ {
+ "epoch": 39.251528384279474,
+ "grad_norm": 0.016500331461429596,
+ "learning_rate": 0.0006,
+ "loss": 4.101631164550781,
+ "step": 2826
+ },
+ {
+ "epoch": 39.26550218340611,
+ "grad_norm": 0.016101544722914696,
+ "learning_rate": 0.0006,
+ "loss": 4.027082443237305,
+ "step": 2827
+ },
+ {
+ "epoch": 39.27947598253275,
+ "grad_norm": 0.015471171587705612,
+ "learning_rate": 0.0006,
+ "loss": 4.022909641265869,
+ "step": 2828
+ },
+ {
+ "epoch": 39.29344978165939,
+ "grad_norm": 0.015779748558998108,
+ "learning_rate": 0.0006,
+ "loss": 4.18796968460083,
+ "step": 2829
+ },
+ {
+ "epoch": 39.30742358078603,
+ "grad_norm": 0.015522679314017296,
+ "learning_rate": 0.0006,
+ "loss": 4.005777359008789,
+ "step": 2830
+ },
+ {
+ "epoch": 39.32139737991266,
+ "grad_norm": 0.015226504765450954,
+ "learning_rate": 0.0006,
+ "loss": 4.108416557312012,
+ "step": 2831
+ },
+ {
+ "epoch": 39.3353711790393,
+ "grad_norm": 0.015095778740942478,
+ "learning_rate": 0.0006,
+ "loss": 4.08029842376709,
+ "step": 2832
+ },
+ {
+ "epoch": 39.34934497816594,
+ "grad_norm": 0.015601031482219696,
+ "learning_rate": 0.0006,
+ "loss": 3.989492416381836,
+ "step": 2833
+ },
+ {
+ "epoch": 39.36331877729258,
+ "grad_norm": 0.014825052581727505,
+ "learning_rate": 0.0006,
+ "loss": 3.998520612716675,
+ "step": 2834
+ },
+ {
+ "epoch": 39.377292576419215,
+ "grad_norm": 0.014821100980043411,
+ "learning_rate": 0.0006,
+ "loss": 4.001777172088623,
+ "step": 2835
+ },
+ {
+ "epoch": 39.391266375545854,
+ "grad_norm": 0.015333754941821098,
+ "learning_rate": 0.0006,
+ "loss": 4.11985969543457,
+ "step": 2836
+ },
+ {
+ "epoch": 39.40524017467249,
+ "grad_norm": 0.015940699726343155,
+ "learning_rate": 0.0006,
+ "loss": 4.025373935699463,
+ "step": 2837
+ },
+ {
+ "epoch": 39.419213973799124,
+ "grad_norm": 0.015096287243068218,
+ "learning_rate": 0.0006,
+ "loss": 4.068551063537598,
+ "step": 2838
+ },
+ {
+ "epoch": 39.43318777292576,
+ "grad_norm": 0.01533548254519701,
+ "learning_rate": 0.0006,
+ "loss": 3.9800801277160645,
+ "step": 2839
+ },
+ {
+ "epoch": 39.4471615720524,
+ "grad_norm": 0.014648732729256153,
+ "learning_rate": 0.0006,
+ "loss": 3.9670190811157227,
+ "step": 2840
+ },
+ {
+ "epoch": 39.46113537117904,
+ "grad_norm": 0.015759671106934547,
+ "learning_rate": 0.0006,
+ "loss": 4.050443649291992,
+ "step": 2841
+ },
+ {
+ "epoch": 39.47510917030568,
+ "grad_norm": 0.016054073348641396,
+ "learning_rate": 0.0006,
+ "loss": 4.22734260559082,
+ "step": 2842
+ },
+ {
+ "epoch": 39.48908296943232,
+ "grad_norm": 0.0160264540463686,
+ "learning_rate": 0.0006,
+ "loss": 4.116174221038818,
+ "step": 2843
+ },
+ {
+ "epoch": 39.50305676855895,
+ "grad_norm": 0.015426610596477985,
+ "learning_rate": 0.0006,
+ "loss": 4.000178337097168,
+ "step": 2844
+ },
+ {
+ "epoch": 39.51703056768559,
+ "grad_norm": 0.01520642638206482,
+ "learning_rate": 0.0006,
+ "loss": 4.218410968780518,
+ "step": 2845
+ },
+ {
+ "epoch": 39.531004366812226,
+ "grad_norm": 0.015305949375033379,
+ "learning_rate": 0.0006,
+ "loss": 4.025891304016113,
+ "step": 2846
+ },
+ {
+ "epoch": 39.544978165938865,
+ "grad_norm": 0.015461564064025879,
+ "learning_rate": 0.0006,
+ "loss": 4.100950717926025,
+ "step": 2847
+ },
+ {
+ "epoch": 39.5589519650655,
+ "grad_norm": 0.01712625101208687,
+ "learning_rate": 0.0006,
+ "loss": 4.0620856285095215,
+ "step": 2848
+ },
+ {
+ "epoch": 39.57292576419214,
+ "grad_norm": 0.017133301123976707,
+ "learning_rate": 0.0006,
+ "loss": 4.097138404846191,
+ "step": 2849
+ },
+ {
+ "epoch": 39.58689956331878,
+ "grad_norm": 0.01683625392615795,
+ "learning_rate": 0.0006,
+ "loss": 4.084364891052246,
+ "step": 2850
+ },
+ {
+ "epoch": 39.60087336244541,
+ "grad_norm": 0.0165710486471653,
+ "learning_rate": 0.0006,
+ "loss": 4.042361259460449,
+ "step": 2851
+ },
+ {
+ "epoch": 39.61484716157205,
+ "grad_norm": 0.01660393364727497,
+ "learning_rate": 0.0006,
+ "loss": 4.1395487785339355,
+ "step": 2852
+ },
+ {
+ "epoch": 39.62882096069869,
+ "grad_norm": 0.017025845125317574,
+ "learning_rate": 0.0006,
+ "loss": 4.1207075119018555,
+ "step": 2853
+ },
+ {
+ "epoch": 39.64279475982533,
+ "grad_norm": 0.01660497486591339,
+ "learning_rate": 0.0006,
+ "loss": 4.103846073150635,
+ "step": 2854
+ },
+ {
+ "epoch": 39.65676855895197,
+ "grad_norm": 0.01607946865260601,
+ "learning_rate": 0.0006,
+ "loss": 4.189016342163086,
+ "step": 2855
+ },
+ {
+ "epoch": 39.670742358078606,
+ "grad_norm": 0.01614505983889103,
+ "learning_rate": 0.0006,
+ "loss": 4.079860687255859,
+ "step": 2856
+ },
+ {
+ "epoch": 39.68471615720524,
+ "grad_norm": 0.017085086554288864,
+ "learning_rate": 0.0006,
+ "loss": 3.9774293899536133,
+ "step": 2857
+ },
+ {
+ "epoch": 39.698689956331876,
+ "grad_norm": 0.01816520094871521,
+ "learning_rate": 0.0006,
+ "loss": 4.113666534423828,
+ "step": 2858
+ },
+ {
+ "epoch": 39.712663755458514,
+ "grad_norm": 0.016597526147961617,
+ "learning_rate": 0.0006,
+ "loss": 4.068242073059082,
+ "step": 2859
+ },
+ {
+ "epoch": 39.72663755458515,
+ "grad_norm": 0.01672235131263733,
+ "learning_rate": 0.0006,
+ "loss": 4.024444580078125,
+ "step": 2860
+ },
+ {
+ "epoch": 39.74061135371179,
+ "grad_norm": 0.020093556493520737,
+ "learning_rate": 0.0006,
+ "loss": 4.064124584197998,
+ "step": 2861
+ },
+ {
+ "epoch": 39.75458515283843,
+ "grad_norm": 0.023141304031014442,
+ "learning_rate": 0.0006,
+ "loss": 4.130439758300781,
+ "step": 2862
+ },
+ {
+ "epoch": 39.76855895196506,
+ "grad_norm": 0.02289372682571411,
+ "learning_rate": 0.0006,
+ "loss": 4.14794921875,
+ "step": 2863
+ },
+ {
+ "epoch": 39.7825327510917,
+ "grad_norm": 0.01919020712375641,
+ "learning_rate": 0.0006,
+ "loss": 4.123961925506592,
+ "step": 2864
+ },
+ {
+ "epoch": 39.79650655021834,
+ "grad_norm": 0.016932329162955284,
+ "learning_rate": 0.0006,
+ "loss": 4.0944132804870605,
+ "step": 2865
+ },
+ {
+ "epoch": 39.81048034934498,
+ "grad_norm": 0.016644535586237907,
+ "learning_rate": 0.0006,
+ "loss": 4.124269962310791,
+ "step": 2866
+ },
+ {
+ "epoch": 39.82445414847162,
+ "grad_norm": 0.016302216798067093,
+ "learning_rate": 0.0006,
+ "loss": 4.0267205238342285,
+ "step": 2867
+ },
+ {
+ "epoch": 39.838427947598255,
+ "grad_norm": 0.01673300378024578,
+ "learning_rate": 0.0006,
+ "loss": 4.10650634765625,
+ "step": 2868
+ },
+ {
+ "epoch": 39.852401746724894,
+ "grad_norm": 0.016581356525421143,
+ "learning_rate": 0.0006,
+ "loss": 4.1547956466674805,
+ "step": 2869
+ },
+ {
+ "epoch": 39.866375545851525,
+ "grad_norm": 0.01691797934472561,
+ "learning_rate": 0.0006,
+ "loss": 4.056936740875244,
+ "step": 2870
+ },
+ {
+ "epoch": 39.880349344978164,
+ "grad_norm": 0.01596726104617119,
+ "learning_rate": 0.0006,
+ "loss": 3.9891414642333984,
+ "step": 2871
+ },
+ {
+ "epoch": 39.8943231441048,
+ "grad_norm": 0.015615091659128666,
+ "learning_rate": 0.0006,
+ "loss": 4.137948036193848,
+ "step": 2872
+ },
+ {
+ "epoch": 39.90829694323144,
+ "grad_norm": 0.015703413635492325,
+ "learning_rate": 0.0006,
+ "loss": 3.9844930171966553,
+ "step": 2873
+ },
+ {
+ "epoch": 39.92227074235808,
+ "grad_norm": 0.015453094616532326,
+ "learning_rate": 0.0006,
+ "loss": 4.098443508148193,
+ "step": 2874
+ },
+ {
+ "epoch": 39.93624454148472,
+ "grad_norm": 0.01449158601462841,
+ "learning_rate": 0.0006,
+ "loss": 4.090852737426758,
+ "step": 2875
+ },
+ {
+ "epoch": 39.95021834061135,
+ "grad_norm": 0.01537515688687563,
+ "learning_rate": 0.0006,
+ "loss": 4.239612579345703,
+ "step": 2876
+ },
+ {
+ "epoch": 39.96419213973799,
+ "grad_norm": 0.015989921987056732,
+ "learning_rate": 0.0006,
+ "loss": 4.142073154449463,
+ "step": 2877
+ },
+ {
+ "epoch": 39.97816593886463,
+ "grad_norm": 0.015985611826181412,
+ "learning_rate": 0.0006,
+ "loss": 4.100076675415039,
+ "step": 2878
+ },
+ {
+ "epoch": 39.992139737991266,
+ "grad_norm": 0.015260833315551281,
+ "learning_rate": 0.0006,
+ "loss": 4.068576335906982,
+ "step": 2879
+ },
+ {
+ "epoch": 40.0,
+ "grad_norm": 0.018047941848635674,
+ "learning_rate": 0.0006,
+ "loss": 4.005878925323486,
+ "step": 2880
+ },
+ {
+ "epoch": 40.0,
+ "eval_loss": 4.622848987579346,
+ "eval_runtime": 56.9618,
+ "eval_samples_per_second": 42.871,
+ "eval_steps_per_second": 1.352,
+ "step": 2880
+ },
+ {
+ "epoch": 40.01397379912664,
+ "grad_norm": 0.015976838767528534,
+ "learning_rate": 0.0006,
+ "loss": 4.035689353942871,
+ "step": 2881
+ },
+ {
+ "epoch": 40.02794759825328,
+ "grad_norm": 0.015176220797002316,
+ "learning_rate": 0.0006,
+ "loss": 3.932197332382202,
+ "step": 2882
+ },
+ {
+ "epoch": 40.041921397379916,
+ "grad_norm": 0.016676228493452072,
+ "learning_rate": 0.0006,
+ "loss": 4.014171123504639,
+ "step": 2883
+ },
+ {
+ "epoch": 40.05589519650655,
+ "grad_norm": 0.01698281429708004,
+ "learning_rate": 0.0006,
+ "loss": 4.117893218994141,
+ "step": 2884
+ },
+ {
+ "epoch": 40.069868995633186,
+ "grad_norm": 0.017872486263513565,
+ "learning_rate": 0.0006,
+ "loss": 4.132119178771973,
+ "step": 2885
+ },
+ {
+ "epoch": 40.083842794759825,
+ "grad_norm": 0.017615554854273796,
+ "learning_rate": 0.0006,
+ "loss": 4.072000980377197,
+ "step": 2886
+ },
+ {
+ "epoch": 40.09781659388646,
+ "grad_norm": 0.017469394952058792,
+ "learning_rate": 0.0006,
+ "loss": 4.017726421356201,
+ "step": 2887
+ },
+ {
+ "epoch": 40.1117903930131,
+ "grad_norm": 0.01954762637615204,
+ "learning_rate": 0.0006,
+ "loss": 4.014443874359131,
+ "step": 2888
+ },
+ {
+ "epoch": 40.12576419213974,
+ "grad_norm": 0.018935853615403175,
+ "learning_rate": 0.0006,
+ "loss": 4.0042948722839355,
+ "step": 2889
+ },
+ {
+ "epoch": 40.13973799126637,
+ "grad_norm": 0.019382692873477936,
+ "learning_rate": 0.0006,
+ "loss": 4.06160831451416,
+ "step": 2890
+ },
+ {
+ "epoch": 40.15371179039301,
+ "grad_norm": 0.019677115604281425,
+ "learning_rate": 0.0006,
+ "loss": 4.13466215133667,
+ "step": 2891
+ },
+ {
+ "epoch": 40.16768558951965,
+ "grad_norm": 0.020365122705698013,
+ "learning_rate": 0.0006,
+ "loss": 4.095000267028809,
+ "step": 2892
+ },
+ {
+ "epoch": 40.18165938864629,
+ "grad_norm": 0.021842138841748238,
+ "learning_rate": 0.0006,
+ "loss": 3.9877853393554688,
+ "step": 2893
+ },
+ {
+ "epoch": 40.19563318777293,
+ "grad_norm": 0.02354881353676319,
+ "learning_rate": 0.0006,
+ "loss": 3.9430112838745117,
+ "step": 2894
+ },
+ {
+ "epoch": 40.209606986899566,
+ "grad_norm": 0.026602735742926598,
+ "learning_rate": 0.0006,
+ "loss": 4.012076377868652,
+ "step": 2895
+ },
+ {
+ "epoch": 40.223580786026204,
+ "grad_norm": 0.023976484313607216,
+ "learning_rate": 0.0006,
+ "loss": 3.9768595695495605,
+ "step": 2896
+ },
+ {
+ "epoch": 40.237554585152836,
+ "grad_norm": 0.02336616814136505,
+ "learning_rate": 0.0006,
+ "loss": 4.1096038818359375,
+ "step": 2897
+ },
+ {
+ "epoch": 40.251528384279474,
+ "grad_norm": 0.02426023595035076,
+ "learning_rate": 0.0006,
+ "loss": 4.157189846038818,
+ "step": 2898
+ },
+ {
+ "epoch": 40.26550218340611,
+ "grad_norm": 0.021572696045041084,
+ "learning_rate": 0.0006,
+ "loss": 3.979175090789795,
+ "step": 2899
+ },
+ {
+ "epoch": 40.27947598253275,
+ "grad_norm": 0.08300432562828064,
+ "learning_rate": 0.0006,
+ "loss": 4.092336654663086,
+ "step": 2900
+ },
+ {
+ "epoch": 40.29344978165939,
+ "grad_norm": 0.7323967814445496,
+ "learning_rate": 0.0006,
+ "loss": 5.337990760803223,
+ "step": 2901
+ },
+ {
+ "epoch": 40.30742358078603,
+ "grad_norm": 0.3363337814807892,
+ "learning_rate": 0.0006,
+ "loss": 6.40043830871582,
+ "step": 2902
+ },
+ {
+ "epoch": 40.32139737991266,
+ "grad_norm": 0.598686933517456,
+ "learning_rate": 0.0006,
+ "loss": 7.004765510559082,
+ "step": 2903
+ },
+ {
+ "epoch": 40.3353711790393,
+ "grad_norm": 0.5998364686965942,
+ "learning_rate": 0.0006,
+ "loss": 8.184747695922852,
+ "step": 2904
+ },
+ {
+ "epoch": 40.34934497816594,
+ "grad_norm": 1.408208966255188,
+ "learning_rate": 0.0006,
+ "loss": 8.88978385925293,
+ "step": 2905
+ },
+ {
+ "epoch": 40.36331877729258,
+ "grad_norm": 0.7829785943031311,
+ "learning_rate": 0.0006,
+ "loss": 11.13601303100586,
+ "step": 2906
+ },
+ {
+ "epoch": 40.377292576419215,
+ "grad_norm": 0.7769215106964111,
+ "learning_rate": 0.0006,
+ "loss": 9.90416145324707,
+ "step": 2907
+ },
+ {
+ "epoch": 40.391266375545854,
+ "grad_norm": 0.46508079767227173,
+ "learning_rate": 0.0006,
+ "loss": 9.647209167480469,
+ "step": 2908
+ },
+ {
+ "epoch": 40.40524017467249,
+ "grad_norm": 1.0228968858718872,
+ "learning_rate": 0.0006,
+ "loss": 11.685434341430664,
+ "step": 2909
+ },
+ {
+ "epoch": 40.419213973799124,
+ "grad_norm": 0.20553778111934662,
+ "learning_rate": 0.0006,
+ "loss": 8.621708869934082,
+ "step": 2910
+ },
+ {
+ "epoch": 40.43318777292576,
+ "grad_norm": 0.18466424942016602,
+ "learning_rate": 0.0006,
+ "loss": 8.590787887573242,
+ "step": 2911
+ },
+ {
+ "epoch": 40.4471615720524,
+ "grad_norm": 0.10458594560623169,
+ "learning_rate": 0.0006,
+ "loss": 8.324458122253418,
+ "step": 2912
+ },
+ {
+ "epoch": 40.46113537117904,
+ "grad_norm": 0.1096842959523201,
+ "learning_rate": 0.0006,
+ "loss": 8.300851821899414,
+ "step": 2913
+ },
+ {
+ "epoch": 40.47510917030568,
+ "grad_norm": 0.10501226782798767,
+ "learning_rate": 0.0006,
+ "loss": 8.21009635925293,
+ "step": 2914
+ },
+ {
+ "epoch": 40.48908296943232,
+ "grad_norm": 0.13704165816307068,
+ "learning_rate": 0.0006,
+ "loss": 8.141355514526367,
+ "step": 2915
+ },
+ {
+ "epoch": 40.50305676855895,
+ "grad_norm": 0.12129108607769012,
+ "learning_rate": 0.0006,
+ "loss": 8.074612617492676,
+ "step": 2916
+ },
+ {
+ "epoch": 40.51703056768559,
+ "grad_norm": 0.10028897225856781,
+ "learning_rate": 0.0006,
+ "loss": 8.00384521484375,
+ "step": 2917
+ },
+ {
+ "epoch": 40.531004366812226,
+ "grad_norm": 0.11764246970415115,
+ "learning_rate": 0.0006,
+ "loss": 8.078598022460938,
+ "step": 2918
+ },
+ {
+ "epoch": 40.544978165938865,
+ "grad_norm": 0.10081959515810013,
+ "learning_rate": 0.0006,
+ "loss": 8.008938789367676,
+ "step": 2919
+ },
+ {
+ "epoch": 40.5589519650655,
+ "grad_norm": 0.1003170907497406,
+ "learning_rate": 0.0006,
+ "loss": 7.962940216064453,
+ "step": 2920
+ },
+ {
+ "epoch": 40.57292576419214,
+ "grad_norm": 0.08763492107391357,
+ "learning_rate": 0.0006,
+ "loss": 7.909419536590576,
+ "step": 2921
+ },
+ {
+ "epoch": 40.58689956331878,
+ "grad_norm": 0.0636807233095169,
+ "learning_rate": 0.0006,
+ "loss": 7.878905773162842,
+ "step": 2922
+ },
+ {
+ "epoch": 40.60087336244541,
+ "grad_norm": 0.06005936488509178,
+ "learning_rate": 0.0006,
+ "loss": 7.850980758666992,
+ "step": 2923
+ },
+ {
+ "epoch": 40.61484716157205,
+ "grad_norm": 0.06415296345949173,
+ "learning_rate": 0.0006,
+ "loss": 7.830353260040283,
+ "step": 2924
+ },
+ {
+ "epoch": 40.62882096069869,
+ "grad_norm": 0.0732007846236229,
+ "learning_rate": 0.0006,
+ "loss": 7.830746650695801,
+ "step": 2925
+ },
+ {
+ "epoch": 40.64279475982533,
+ "grad_norm": 0.06628110259771347,
+ "learning_rate": 0.0006,
+ "loss": 7.82661771774292,
+ "step": 2926
+ },
+ {
+ "epoch": 40.65676855895197,
+ "grad_norm": 0.05544218048453331,
+ "learning_rate": 0.0006,
+ "loss": 7.814305782318115,
+ "step": 2927
+ },
+ {
+ "epoch": 40.670742358078606,
+ "grad_norm": 0.050327103585004807,
+ "learning_rate": 0.0006,
+ "loss": 7.764055252075195,
+ "step": 2928
+ },
+ {
+ "epoch": 40.68471615720524,
+ "grad_norm": 0.04918425530195236,
+ "learning_rate": 0.0006,
+ "loss": 7.744625091552734,
+ "step": 2929
+ },
+ {
+ "epoch": 40.698689956331876,
+ "grad_norm": 0.042862266302108765,
+ "learning_rate": 0.0006,
+ "loss": 7.764426231384277,
+ "step": 2930
+ },
+ {
+ "epoch": 40.712663755458514,
+ "grad_norm": 0.04338089004158974,
+ "learning_rate": 0.0006,
+ "loss": 7.712530136108398,
+ "step": 2931
+ },
+ {
+ "epoch": 40.72663755458515,
+ "grad_norm": 0.04127737879753113,
+ "learning_rate": 0.0006,
+ "loss": 7.742180347442627,
+ "step": 2932
+ },
+ {
+ "epoch": 40.74061135371179,
+ "grad_norm": 0.03358631581068039,
+ "learning_rate": 0.0006,
+ "loss": 7.714834213256836,
+ "step": 2933
+ },
+ {
+ "epoch": 40.75458515283843,
+ "grad_norm": 0.02845790982246399,
+ "learning_rate": 0.0006,
+ "loss": 7.7021026611328125,
+ "step": 2934
+ },
+ {
+ "epoch": 40.76855895196506,
+ "grad_norm": 0.03012419492006302,
+ "learning_rate": 0.0006,
+ "loss": 7.712377548217773,
+ "step": 2935
+ },
+ {
+ "epoch": 40.7825327510917,
+ "grad_norm": 0.032981179654598236,
+ "learning_rate": 0.0006,
+ "loss": 7.666385650634766,
+ "step": 2936
+ },
+ {
+ "epoch": 40.79650655021834,
+ "grad_norm": 0.03357074037194252,
+ "learning_rate": 0.0006,
+ "loss": 7.649097919464111,
+ "step": 2937
+ },
+ {
+ "epoch": 40.81048034934498,
+ "grad_norm": 0.030914708971977234,
+ "learning_rate": 0.0006,
+ "loss": 7.649196624755859,
+ "step": 2938
+ },
+ {
+ "epoch": 40.82445414847162,
+ "grad_norm": 0.026296131312847137,
+ "learning_rate": 0.0006,
+ "loss": 7.629964828491211,
+ "step": 2939
+ },
+ {
+ "epoch": 40.838427947598255,
+ "grad_norm": 0.026677245274186134,
+ "learning_rate": 0.0006,
+ "loss": 7.599973678588867,
+ "step": 2940
+ },
+ {
+ "epoch": 40.852401746724894,
+ "grad_norm": 0.023208219558000565,
+ "learning_rate": 0.0006,
+ "loss": 7.601624488830566,
+ "step": 2941
+ },
+ {
+ "epoch": 40.866375545851525,
+ "grad_norm": 0.020741475746035576,
+ "learning_rate": 0.0006,
+ "loss": 7.603902339935303,
+ "step": 2942
+ },
+ {
+ "epoch": 40.880349344978164,
+ "grad_norm": 0.019221751019358635,
+ "learning_rate": 0.0006,
+ "loss": 7.5535888671875,
+ "step": 2943
+ },
+ {
+ "epoch": 40.8943231441048,
+ "grad_norm": 0.019491951912641525,
+ "learning_rate": 0.0006,
+ "loss": 7.54597282409668,
+ "step": 2944
+ },
+ {
+ "epoch": 40.90829694323144,
+ "grad_norm": 0.019702572375535965,
+ "learning_rate": 0.0006,
+ "loss": 7.537567138671875,
+ "step": 2945
+ },
+ {
+ "epoch": 40.92227074235808,
+ "grad_norm": 0.01768800988793373,
+ "learning_rate": 0.0006,
+ "loss": 7.512800216674805,
+ "step": 2946
+ },
+ {
+ "epoch": 40.93624454148472,
+ "grad_norm": 0.01926138624548912,
+ "learning_rate": 0.0006,
+ "loss": 7.509932518005371,
+ "step": 2947
+ },
+ {
+ "epoch": 40.95021834061135,
+ "grad_norm": 0.019030006602406502,
+ "learning_rate": 0.0006,
+ "loss": 7.497983932495117,
+ "step": 2948
+ },
+ {
+ "epoch": 40.96419213973799,
+ "grad_norm": 0.017313841730356216,
+ "learning_rate": 0.0006,
+ "loss": 7.463518142700195,
+ "step": 2949
+ },
+ {
+ "epoch": 40.97816593886463,
+ "grad_norm": 0.01310762483626604,
+ "learning_rate": 0.0006,
+ "loss": 7.474668502807617,
+ "step": 2950
+ },
+ {
+ "epoch": 40.992139737991266,
+ "grad_norm": 0.013008282519876957,
+ "learning_rate": 0.0006,
+ "loss": 7.45982551574707,
+ "step": 2951
+ },
+ {
+ "epoch": 41.0,
+ "grad_norm": 0.01441363524645567,
+ "learning_rate": 0.0006,
+ "loss": 7.446529388427734,
+ "step": 2952
+ },
+ {
+ "epoch": 41.0,
+ "eval_loss": 7.486120223999023,
+ "eval_runtime": 70.4986,
+ "eval_samples_per_second": 34.639,
+ "eval_steps_per_second": 1.092,
+ "step": 2952
+ },
+ {
+ "epoch": 41.01397379912664,
+ "grad_norm": 0.014061855152249336,
+ "learning_rate": 0.0006,
+ "loss": 7.4692182540893555,
+ "step": 2953
+ },
+ {
+ "epoch": 41.02794759825328,
+ "grad_norm": 0.01587500050663948,
+ "learning_rate": 0.0006,
+ "loss": 7.4182024002075195,
+ "step": 2954
+ },
+ {
+ "epoch": 41.041921397379916,
+ "grad_norm": 0.014783984050154686,
+ "learning_rate": 0.0006,
+ "loss": 7.466778755187988,
+ "step": 2955
+ },
+ {
+ "epoch": 41.05589519650655,
+ "grad_norm": 0.011560854502022266,
+ "learning_rate": 0.0006,
+ "loss": 7.438002586364746,
+ "step": 2956
+ },
+ {
+ "epoch": 41.069868995633186,
+ "grad_norm": 0.00970957800745964,
+ "learning_rate": 0.0006,
+ "loss": 7.442694187164307,
+ "step": 2957
+ },
+ {
+ "epoch": 41.083842794759825,
+ "grad_norm": 0.009633459150791168,
+ "learning_rate": 0.0006,
+ "loss": 7.458884239196777,
+ "step": 2958
+ },
+ {
+ "epoch": 41.09781659388646,
+ "grad_norm": 0.010664350353181362,
+ "learning_rate": 0.0006,
+ "loss": 7.48378324508667,
+ "step": 2959
+ },
+ {
+ "epoch": 41.1117903930131,
+ "grad_norm": 0.011440591886639595,
+ "learning_rate": 0.0006,
+ "loss": 7.474484920501709,
+ "step": 2960
+ },
+ {
+ "epoch": 41.12576419213974,
+ "grad_norm": 0.01150908600538969,
+ "learning_rate": 0.0006,
+ "loss": 7.417902946472168,
+ "step": 2961
+ },
+ {
+ "epoch": 41.13973799126637,
+ "grad_norm": 0.010375697165727615,
+ "learning_rate": 0.0006,
+ "loss": 7.403279781341553,
+ "step": 2962
+ },
+ {
+ "epoch": 41.15371179039301,
+ "grad_norm": 0.009075703099370003,
+ "learning_rate": 0.0006,
+ "loss": 7.447347164154053,
+ "step": 2963
+ },
+ {
+ "epoch": 41.16768558951965,
+ "grad_norm": 0.008892177604138851,
+ "learning_rate": 0.0006,
+ "loss": 7.402681350708008,
+ "step": 2964
+ },
+ {
+ "epoch": 41.18165938864629,
+ "grad_norm": 0.009492084383964539,
+ "learning_rate": 0.0006,
+ "loss": 7.3871355056762695,
+ "step": 2965
+ },
+ {
+ "epoch": 41.19563318777293,
+ "grad_norm": 0.009139460511505604,
+ "learning_rate": 0.0006,
+ "loss": 7.419617176055908,
+ "step": 2966
+ },
+ {
+ "epoch": 41.209606986899566,
+ "grad_norm": 0.007959138602018356,
+ "learning_rate": 0.0006,
+ "loss": 7.374949932098389,
+ "step": 2967
+ },
+ {
+ "epoch": 41.223580786026204,
+ "grad_norm": 0.007571548223495483,
+ "learning_rate": 0.0006,
+ "loss": 7.4205241203308105,
+ "step": 2968
+ },
+ {
+ "epoch": 41.237554585152836,
+ "grad_norm": 0.007885400205850601,
+ "learning_rate": 0.0006,
+ "loss": 7.416543006896973,
+ "step": 2969
+ },
+ {
+ "epoch": 41.251528384279474,
+ "grad_norm": 0.008650158531963825,
+ "learning_rate": 0.0006,
+ "loss": 7.357326984405518,
+ "step": 2970
+ },
+ {
+ "epoch": 41.26550218340611,
+ "grad_norm": 0.007483262103050947,
+ "learning_rate": 0.0006,
+ "loss": 7.410891056060791,
+ "step": 2971
+ },
+ {
+ "epoch": 41.27947598253275,
+ "grad_norm": 0.00869277399033308,
+ "learning_rate": 0.0006,
+ "loss": 7.392127990722656,
+ "step": 2972
+ },
+ {
+ "epoch": 41.29344978165939,
+ "grad_norm": 0.00885078776627779,
+ "learning_rate": 0.0006,
+ "loss": 7.405449867248535,
+ "step": 2973
+ },
+ {
+ "epoch": 41.30742358078603,
+ "grad_norm": 0.00958226341754198,
+ "learning_rate": 0.0006,
+ "loss": 7.399868488311768,
+ "step": 2974
+ },
+ {
+ "epoch": 41.32139737991266,
+ "grad_norm": 0.006991599686443806,
+ "learning_rate": 0.0006,
+ "loss": 7.35369873046875,
+ "step": 2975
+ },
+ {
+ "epoch": 41.3353711790393,
+ "grad_norm": 0.007415142375975847,
+ "learning_rate": 0.0006,
+ "loss": 7.358353137969971,
+ "step": 2976
+ },
+ {
+ "epoch": 41.34934497816594,
+ "grad_norm": 0.007024371065199375,
+ "learning_rate": 0.0006,
+ "loss": 7.4070892333984375,
+ "step": 2977
+ },
+ {
+ "epoch": 41.36331877729258,
+ "grad_norm": 0.007234332151710987,
+ "learning_rate": 0.0006,
+ "loss": 7.379120349884033,
+ "step": 2978
+ },
+ {
+ "epoch": 41.377292576419215,
+ "grad_norm": 0.00798892229795456,
+ "learning_rate": 0.0006,
+ "loss": 7.365455627441406,
+ "step": 2979
+ },
+ {
+ "epoch": 41.391266375545854,
+ "grad_norm": 0.007303398102521896,
+ "learning_rate": 0.0006,
+ "loss": 7.341798782348633,
+ "step": 2980
+ },
+ {
+ "epoch": 41.40524017467249,
+ "grad_norm": 0.007119669578969479,
+ "learning_rate": 0.0006,
+ "loss": 7.363280296325684,
+ "step": 2981
+ },
+ {
+ "epoch": 41.419213973799124,
+ "grad_norm": 0.007457619532942772,
+ "learning_rate": 0.0006,
+ "loss": 7.39388370513916,
+ "step": 2982
+ },
+ {
+ "epoch": 41.43318777292576,
+ "grad_norm": 0.007905688136816025,
+ "learning_rate": 0.0006,
+ "loss": 7.380652904510498,
+ "step": 2983
+ },
+ {
+ "epoch": 41.4471615720524,
+ "grad_norm": 0.006212855689227581,
+ "learning_rate": 0.0006,
+ "loss": 7.398769378662109,
+ "step": 2984
+ },
+ {
+ "epoch": 41.46113537117904,
+ "grad_norm": 0.0070641543716192245,
+ "learning_rate": 0.0006,
+ "loss": 7.3503546714782715,
+ "step": 2985
+ },
+ {
+ "epoch": 41.47510917030568,
+ "grad_norm": 0.00787805300205946,
+ "learning_rate": 0.0006,
+ "loss": 7.347678184509277,
+ "step": 2986
+ },
+ {
+ "epoch": 41.48908296943232,
+ "grad_norm": 0.006760426796972752,
+ "learning_rate": 0.0006,
+ "loss": 7.380006790161133,
+ "step": 2987
+ },
+ {
+ "epoch": 41.50305676855895,
+ "grad_norm": 0.00772385997697711,
+ "learning_rate": 0.0006,
+ "loss": 7.363680362701416,
+ "step": 2988
+ },
+ {
+ "epoch": 41.51703056768559,
+ "grad_norm": 0.007373564876616001,
+ "learning_rate": 0.0006,
+ "loss": 7.333014488220215,
+ "step": 2989
+ },
+ {
+ "epoch": 41.531004366812226,
+ "grad_norm": 0.006717653013765812,
+ "learning_rate": 0.0006,
+ "loss": 7.360076427459717,
+ "step": 2990
+ },
+ {
+ "epoch": 41.544978165938865,
+ "grad_norm": 0.007191413082182407,
+ "learning_rate": 0.0006,
+ "loss": 7.397030830383301,
+ "step": 2991
+ },
+ {
+ "epoch": 41.5589519650655,
+ "grad_norm": 0.007831960916519165,
+ "learning_rate": 0.0006,
+ "loss": 7.369959831237793,
+ "step": 2992
+ },
+ {
+ "epoch": 41.57292576419214,
+ "grad_norm": 0.006947481073439121,
+ "learning_rate": 0.0006,
+ "loss": 7.333254337310791,
+ "step": 2993
+ },
+ {
+ "epoch": 41.58689956331878,
+ "grad_norm": 0.006243606563657522,
+ "learning_rate": 0.0006,
+ "loss": 7.327378273010254,
+ "step": 2994
+ },
+ {
+ "epoch": 41.60087336244541,
+ "grad_norm": 0.006298339460045099,
+ "learning_rate": 0.0006,
+ "loss": 7.323482036590576,
+ "step": 2995
+ },
+ {
+ "epoch": 41.61484716157205,
+ "grad_norm": 0.007587412837892771,
+ "learning_rate": 0.0006,
+ "loss": 7.313789367675781,
+ "step": 2996
+ },
+ {
+ "epoch": 41.62882096069869,
+ "grad_norm": 0.007140746805816889,
+ "learning_rate": 0.0006,
+ "loss": 7.356067657470703,
+ "step": 2997
+ },
+ {
+ "epoch": 41.64279475982533,
+ "grad_norm": 0.008315625600516796,
+ "learning_rate": 0.0006,
+ "loss": 7.3339643478393555,
+ "step": 2998
+ },
+ {
+ "epoch": 41.65676855895197,
+ "grad_norm": 0.006547102238982916,
+ "learning_rate": 0.0006,
+ "loss": 7.342191696166992,
+ "step": 2999
+ },
+ {
+ "epoch": 41.670742358078606,
+ "grad_norm": 0.006068122107535601,
+ "learning_rate": 0.0006,
+ "loss": 7.321015357971191,
+ "step": 3000
+ },
+ {
+ "epoch": 41.68471615720524,
+ "grad_norm": 0.007316991221159697,
+ "learning_rate": 0.0006,
+ "loss": 7.346490859985352,
+ "step": 3001
+ },
+ {
+ "epoch": 41.698689956331876,
+ "grad_norm": 0.0067388564348220825,
+ "learning_rate": 0.0006,
+ "loss": 7.340723991394043,
+ "step": 3002
+ },
+ {
+ "epoch": 41.712663755458514,
+ "grad_norm": 0.006839963141828775,
+ "learning_rate": 0.0006,
+ "loss": 7.329365253448486,
+ "step": 3003
+ },
+ {
+ "epoch": 41.72663755458515,
+ "grad_norm": 0.00805041566491127,
+ "learning_rate": 0.0006,
+ "loss": 7.298934459686279,
+ "step": 3004
+ },
+ {
+ "epoch": 41.74061135371179,
+ "grad_norm": 0.007197503466159105,
+ "learning_rate": 0.0006,
+ "loss": 7.331496238708496,
+ "step": 3005
+ },
+ {
+ "epoch": 41.75458515283843,
+ "grad_norm": 0.007220805156975985,
+ "learning_rate": 0.0006,
+ "loss": 7.317695140838623,
+ "step": 3006
+ },
+ {
+ "epoch": 41.76855895196506,
+ "grad_norm": 0.00874276738613844,
+ "learning_rate": 0.0006,
+ "loss": 7.272614479064941,
+ "step": 3007
+ },
+ {
+ "epoch": 41.7825327510917,
+ "grad_norm": 0.008386565372347832,
+ "learning_rate": 0.0006,
+ "loss": 7.295244216918945,
+ "step": 3008
+ },
+ {
+ "epoch": 41.79650655021834,
+ "grad_norm": 0.0077730282209813595,
+ "learning_rate": 0.0006,
+ "loss": 7.3395490646362305,
+ "step": 3009
+ },
+ {
+ "epoch": 41.81048034934498,
+ "grad_norm": 0.008441480807960033,
+ "learning_rate": 0.0006,
+ "loss": 7.319921493530273,
+ "step": 3010
+ },
+ {
+ "epoch": 41.82445414847162,
+ "grad_norm": 0.007398269604891539,
+ "learning_rate": 0.0006,
+ "loss": 7.3029465675354,
+ "step": 3011
+ },
+ {
+ "epoch": 41.838427947598255,
+ "grad_norm": 0.006878677289932966,
+ "learning_rate": 0.0006,
+ "loss": 7.318467140197754,
+ "step": 3012
+ },
+ {
+ "epoch": 41.852401746724894,
+ "grad_norm": 0.007616598624736071,
+ "learning_rate": 0.0006,
+ "loss": 7.3112077713012695,
+ "step": 3013
+ },
+ {
+ "epoch": 41.866375545851525,
+ "grad_norm": 0.011629699729382992,
+ "learning_rate": 0.0006,
+ "loss": 7.32523250579834,
+ "step": 3014
+ },
+ {
+ "epoch": 41.880349344978164,
+ "grad_norm": 0.011011249385774136,
+ "learning_rate": 0.0006,
+ "loss": 7.287247657775879,
+ "step": 3015
+ },
+ {
+ "epoch": 41.8943231441048,
+ "grad_norm": 0.010208295658230782,
+ "learning_rate": 0.0006,
+ "loss": 7.292117118835449,
+ "step": 3016
+ },
+ {
+ "epoch": 41.90829694323144,
+ "grad_norm": 0.008651095442473888,
+ "learning_rate": 0.0006,
+ "loss": 7.302201271057129,
+ "step": 3017
+ },
+ {
+ "epoch": 41.92227074235808,
+ "grad_norm": 0.006692920345813036,
+ "learning_rate": 0.0006,
+ "loss": 7.319568634033203,
+ "step": 3018
+ },
+ {
+ "epoch": 41.93624454148472,
+ "grad_norm": 0.010033348575234413,
+ "learning_rate": 0.0006,
+ "loss": 7.2738566398620605,
+ "step": 3019
+ },
+ {
+ "epoch": 41.95021834061135,
+ "grad_norm": 0.011790832504630089,
+ "learning_rate": 0.0006,
+ "loss": 7.298540115356445,
+ "step": 3020
+ },
+ {
+ "epoch": 41.96419213973799,
+ "grad_norm": 0.018622662872076035,
+ "learning_rate": 0.0006,
+ "loss": 7.267182350158691,
+ "step": 3021
+ },
+ {
+ "epoch": 41.97816593886463,
+ "grad_norm": 0.02520257607102394,
+ "learning_rate": 0.0006,
+ "loss": 7.261566162109375,
+ "step": 3022
+ },
+ {
+ "epoch": 41.992139737991266,
+ "grad_norm": 0.036332644522190094,
+ "learning_rate": 0.0006,
+ "loss": 7.270719051361084,
+ "step": 3023
+ },
+ {
+ "epoch": 42.0,
+ "grad_norm": 0.04717305675148964,
+ "learning_rate": 0.0006,
+ "loss": 7.293332099914551,
+ "step": 3024
+ },
+ {
+ "epoch": 42.0,
+ "eval_loss": 7.354280471801758,
+ "eval_runtime": 56.4353,
+ "eval_samples_per_second": 43.271,
+ "eval_steps_per_second": 1.364,
+ "step": 3024
+ },
+ {
+ "epoch": 42.01397379912664,
+ "grad_norm": 0.06661626696586609,
+ "learning_rate": 0.0006,
+ "loss": 7.338189601898193,
+ "step": 3025
+ },
+ {
+ "epoch": 42.02794759825328,
+ "grad_norm": 0.05452054366469383,
+ "learning_rate": 0.0006,
+ "loss": 7.316764831542969,
+ "step": 3026
+ },
+ {
+ "epoch": 42.041921397379916,
+ "grad_norm": 0.025638775900006294,
+ "learning_rate": 0.0006,
+ "loss": 7.34204626083374,
+ "step": 3027
+ },
+ {
+ "epoch": 42.05589519650655,
+ "grad_norm": 0.022398799657821655,
+ "learning_rate": 0.0006,
+ "loss": 7.262149333953857,
+ "step": 3028
+ },
+ {
+ "epoch": 42.069868995633186,
+ "grad_norm": 0.03713499382138252,
+ "learning_rate": 0.0006,
+ "loss": 7.27695369720459,
+ "step": 3029
+ },
+ {
+ "epoch": 42.083842794759825,
+ "grad_norm": 0.014949247241020203,
+ "learning_rate": 0.0006,
+ "loss": 7.296508312225342,
+ "step": 3030
+ },
+ {
+ "epoch": 42.09781659388646,
+ "grad_norm": 0.02703995071351528,
+ "learning_rate": 0.0006,
+ "loss": 7.258018493652344,
+ "step": 3031
+ },
+ {
+ "epoch": 42.1117903930131,
+ "grad_norm": 0.020352816209197044,
+ "learning_rate": 0.0006,
+ "loss": 7.241450309753418,
+ "step": 3032
+ },
+ {
+ "epoch": 42.12576419213974,
+ "grad_norm": 0.014578479342162609,
+ "learning_rate": 0.0006,
+ "loss": 7.257323265075684,
+ "step": 3033
+ },
+ {
+ "epoch": 42.13973799126637,
+ "grad_norm": 0.026053044945001602,
+ "learning_rate": 0.0006,
+ "loss": 7.254680633544922,
+ "step": 3034
+ },
+ {
+ "epoch": 42.15371179039301,
+ "grad_norm": 0.015120275318622589,
+ "learning_rate": 0.0006,
+ "loss": 7.261074066162109,
+ "step": 3035
+ },
+ {
+ "epoch": 42.16768558951965,
+ "grad_norm": 0.016822177916765213,
+ "learning_rate": 0.0006,
+ "loss": 7.279064178466797,
+ "step": 3036
+ },
+ {
+ "epoch": 42.18165938864629,
+ "grad_norm": 0.016367588192224503,
+ "learning_rate": 0.0006,
+ "loss": 7.251429557800293,
+ "step": 3037
+ },
+ {
+ "epoch": 42.19563318777293,
+ "grad_norm": 0.014350418001413345,
+ "learning_rate": 0.0006,
+ "loss": 7.259264945983887,
+ "step": 3038
+ },
+ {
+ "epoch": 42.209606986899566,
+ "grad_norm": 0.017078228294849396,
+ "learning_rate": 0.0006,
+ "loss": 7.227105140686035,
+ "step": 3039
+ },
+ {
+ "epoch": 42.223580786026204,
+ "grad_norm": 0.011759317480027676,
+ "learning_rate": 0.0006,
+ "loss": 7.275888442993164,
+ "step": 3040
+ },
+ {
+ "epoch": 42.237554585152836,
+ "grad_norm": 0.013019340112805367,
+ "learning_rate": 0.0006,
+ "loss": 7.2517008781433105,
+ "step": 3041
+ },
+ {
+ "epoch": 42.251528384279474,
+ "grad_norm": 0.01178441196680069,
+ "learning_rate": 0.0006,
+ "loss": 7.23576545715332,
+ "step": 3042
+ },
+ {
+ "epoch": 42.26550218340611,
+ "grad_norm": 0.014518891461193562,
+ "learning_rate": 0.0006,
+ "loss": 7.253895282745361,
+ "step": 3043
+ },
+ {
+ "epoch": 42.27947598253275,
+ "grad_norm": 0.018159950152039528,
+ "learning_rate": 0.0006,
+ "loss": 7.278058052062988,
+ "step": 3044
+ },
+ {
+ "epoch": 42.29344978165939,
+ "grad_norm": 0.007587796077132225,
+ "learning_rate": 0.0006,
+ "loss": 7.278578758239746,
+ "step": 3045
+ },
+ {
+ "epoch": 42.30742358078603,
+ "grad_norm": 0.015994979068636894,
+ "learning_rate": 0.0006,
+ "loss": 7.239841461181641,
+ "step": 3046
+ },
+ {
+ "epoch": 42.32139737991266,
+ "grad_norm": 0.017830168828368187,
+ "learning_rate": 0.0006,
+ "loss": 7.2443976402282715,
+ "step": 3047
+ },
+ {
+ "epoch": 42.3353711790393,
+ "grad_norm": 0.008738759905099869,
+ "learning_rate": 0.0006,
+ "loss": 7.232264518737793,
+ "step": 3048
+ },
+ {
+ "epoch": 42.34934497816594,
+ "grad_norm": 0.0175449438393116,
+ "learning_rate": 0.0006,
+ "loss": 7.228000640869141,
+ "step": 3049
+ },
+ {
+ "epoch": 42.36331877729258,
+ "grad_norm": 0.019611820578575134,
+ "learning_rate": 0.0006,
+ "loss": 7.234999179840088,
+ "step": 3050
+ },
+ {
+ "epoch": 42.377292576419215,
+ "grad_norm": 0.015302863903343678,
+ "learning_rate": 0.0006,
+ "loss": 7.242547988891602,
+ "step": 3051
+ },
+ {
+ "epoch": 42.391266375545854,
+ "grad_norm": 0.025919731706380844,
+ "learning_rate": 0.0006,
+ "loss": 7.216652870178223,
+ "step": 3052
+ },
+ {
+ "epoch": 42.40524017467249,
+ "grad_norm": 0.034587468951940536,
+ "learning_rate": 0.0006,
+ "loss": 7.245336532592773,
+ "step": 3053
+ },
+ {
+ "epoch": 42.419213973799124,
+ "grad_norm": 0.019198840484023094,
+ "learning_rate": 0.0006,
+ "loss": 7.256758689880371,
+ "step": 3054
+ },
+ {
+ "epoch": 42.43318777292576,
+ "grad_norm": 0.012316389009356499,
+ "learning_rate": 0.0006,
+ "loss": 7.209102630615234,
+ "step": 3055
+ },
+ {
+ "epoch": 42.4471615720524,
+ "grad_norm": 0.023512162268161774,
+ "learning_rate": 0.0006,
+ "loss": 7.250742435455322,
+ "step": 3056
+ },
+ {
+ "epoch": 42.46113537117904,
+ "grad_norm": 0.025457851588726044,
+ "learning_rate": 0.0006,
+ "loss": 7.263757705688477,
+ "step": 3057
+ },
+ {
+ "epoch": 42.47510917030568,
+ "grad_norm": 0.02192617766559124,
+ "learning_rate": 0.0006,
+ "loss": 7.246977806091309,
+ "step": 3058
+ },
+ {
+ "epoch": 42.48908296943232,
+ "grad_norm": 0.02541281282901764,
+ "learning_rate": 0.0006,
+ "loss": 7.232361316680908,
+ "step": 3059
+ },
+ {
+ "epoch": 42.50305676855895,
+ "grad_norm": 0.036353763192892075,
+ "learning_rate": 0.0006,
+ "loss": 7.242556571960449,
+ "step": 3060
+ },
+ {
+ "epoch": 42.51703056768559,
+ "grad_norm": 0.04555872827768326,
+ "learning_rate": 0.0006,
+ "loss": 7.228940963745117,
+ "step": 3061
+ },
+ {
+ "epoch": 42.531004366812226,
+ "grad_norm": 0.04563378915190697,
+ "learning_rate": 0.0006,
+ "loss": 7.235276222229004,
+ "step": 3062
+ },
+ {
+ "epoch": 42.544978165938865,
+ "grad_norm": 0.04275760427117348,
+ "learning_rate": 0.0006,
+ "loss": 7.238895416259766,
+ "step": 3063
+ },
+ {
+ "epoch": 42.5589519650655,
+ "grad_norm": 0.04535885155200958,
+ "learning_rate": 0.0006,
+ "loss": 7.205336570739746,
+ "step": 3064
+ },
+ {
+ "epoch": 42.57292576419214,
+ "grad_norm": 0.015534617006778717,
+ "learning_rate": 0.0006,
+ "loss": 7.238393783569336,
+ "step": 3065
+ },
+ {
+ "epoch": 42.58689956331878,
+ "grad_norm": 0.03286299481987953,
+ "learning_rate": 0.0006,
+ "loss": 7.216453552246094,
+ "step": 3066
+ },
+ {
+ "epoch": 42.60087336244541,
+ "grad_norm": 0.025453168898820877,
+ "learning_rate": 0.0006,
+ "loss": 7.203750133514404,
+ "step": 3067
+ },
+ {
+ "epoch": 42.61484716157205,
+ "grad_norm": 0.021501649171113968,
+ "learning_rate": 0.0006,
+ "loss": 7.241621971130371,
+ "step": 3068
+ },
+ {
+ "epoch": 42.62882096069869,
+ "grad_norm": 0.02743634767830372,
+ "learning_rate": 0.0006,
+ "loss": 7.22076940536499,
+ "step": 3069
+ },
+ {
+ "epoch": 42.64279475982533,
+ "grad_norm": 0.020890483632683754,
+ "learning_rate": 0.0006,
+ "loss": 7.218412399291992,
+ "step": 3070
+ },
+ {
+ "epoch": 42.65676855895197,
+ "grad_norm": 0.022499850019812584,
+ "learning_rate": 0.0006,
+ "loss": 7.217696189880371,
+ "step": 3071
+ },
+ {
+ "epoch": 42.670742358078606,
+ "grad_norm": 0.01653696969151497,
+ "learning_rate": 0.0006,
+ "loss": 7.156417369842529,
+ "step": 3072
+ },
+ {
+ "epoch": 42.68471615720524,
+ "grad_norm": 0.01834225095808506,
+ "learning_rate": 0.0006,
+ "loss": 7.211247444152832,
+ "step": 3073
+ },
+ {
+ "epoch": 42.698689956331876,
+ "grad_norm": 0.020137229934334755,
+ "learning_rate": 0.0006,
+ "loss": 7.232499599456787,
+ "step": 3074
+ },
+ {
+ "epoch": 42.712663755458514,
+ "grad_norm": 0.012671459466218948,
+ "learning_rate": 0.0006,
+ "loss": 7.172441482543945,
+ "step": 3075
+ },
+ {
+ "epoch": 42.72663755458515,
+ "grad_norm": 0.016803130507469177,
+ "learning_rate": 0.0006,
+ "loss": 7.152103424072266,
+ "step": 3076
+ },
+ {
+ "epoch": 42.74061135371179,
+ "grad_norm": 0.01623712107539177,
+ "learning_rate": 0.0006,
+ "loss": 7.187032699584961,
+ "step": 3077
+ },
+ {
+ "epoch": 42.75458515283843,
+ "grad_norm": 0.015938356518745422,
+ "learning_rate": 0.0006,
+ "loss": 7.171140670776367,
+ "step": 3078
+ },
+ {
+ "epoch": 42.76855895196506,
+ "grad_norm": 0.02580736204981804,
+ "learning_rate": 0.0006,
+ "loss": 7.204472541809082,
+ "step": 3079
+ },
+ {
+ "epoch": 42.7825327510917,
+ "grad_norm": 0.023372117429971695,
+ "learning_rate": 0.0006,
+ "loss": 7.165554523468018,
+ "step": 3080
+ },
+ {
+ "epoch": 42.79650655021834,
+ "grad_norm": 0.04440586268901825,
+ "learning_rate": 0.0006,
+ "loss": 7.169431209564209,
+ "step": 3081
+ },
+ {
+ "epoch": 42.81048034934498,
+ "grad_norm": 0.06797406077384949,
+ "learning_rate": 0.0006,
+ "loss": 7.168193340301514,
+ "step": 3082
+ },
+ {
+ "epoch": 42.82445414847162,
+ "grad_norm": 0.0914321169257164,
+ "learning_rate": 0.0006,
+ "loss": 7.204671859741211,
+ "step": 3083
+ },
+ {
+ "epoch": 42.838427947598255,
+ "grad_norm": 0.0506940595805645,
+ "learning_rate": 0.0006,
+ "loss": 7.15949821472168,
+ "step": 3084
+ },
+ {
+ "epoch": 42.852401746724894,
+ "grad_norm": 0.04819890484213829,
+ "learning_rate": 0.0006,
+ "loss": 7.192837238311768,
+ "step": 3085
+ },
+ {
+ "epoch": 42.866375545851525,
+ "grad_norm": 0.040044233202934265,
+ "learning_rate": 0.0006,
+ "loss": 7.1600165367126465,
+ "step": 3086
+ },
+ {
+ "epoch": 42.880349344978164,
+ "grad_norm": 0.02573299966752529,
+ "learning_rate": 0.0006,
+ "loss": 7.153427600860596,
+ "step": 3087
+ },
+ {
+ "epoch": 42.8943231441048,
+ "grad_norm": 0.046942535787820816,
+ "learning_rate": 0.0006,
+ "loss": 7.16767692565918,
+ "step": 3088
+ },
+ {
+ "epoch": 42.90829694323144,
+ "grad_norm": 0.02683161571621895,
+ "learning_rate": 0.0006,
+ "loss": 7.17144250869751,
+ "step": 3089
+ },
+ {
+ "epoch": 42.92227074235808,
+ "grad_norm": 0.05363674834370613,
+ "learning_rate": 0.0006,
+ "loss": 7.120128631591797,
+ "step": 3090
+ },
+ {
+ "epoch": 42.93624454148472,
+ "grad_norm": 0.023756247013807297,
+ "learning_rate": 0.0006,
+ "loss": 7.169703006744385,
+ "step": 3091
+ },
+ {
+ "epoch": 42.95021834061135,
+ "grad_norm": 0.04216783866286278,
+ "learning_rate": 0.0006,
+ "loss": 7.1206865310668945,
+ "step": 3092
+ },
+ {
+ "epoch": 42.96419213973799,
+ "grad_norm": 0.02219565585255623,
+ "learning_rate": 0.0006,
+ "loss": 7.157215118408203,
+ "step": 3093
+ },
+ {
+ "epoch": 42.97816593886463,
+ "grad_norm": 0.03527127206325531,
+ "learning_rate": 0.0006,
+ "loss": 7.166672706604004,
+ "step": 3094
+ },
+ {
+ "epoch": 42.992139737991266,
+ "grad_norm": 0.017748817801475525,
+ "learning_rate": 0.0006,
+ "loss": 7.154747009277344,
+ "step": 3095
+ },
+ {
+ "epoch": 43.0,
+ "grad_norm": 0.033165838569402695,
+ "learning_rate": 0.0006,
+ "loss": 7.150866508483887,
+ "step": 3096
+ },
+ {
+ "epoch": 43.0,
+ "eval_loss": 7.144403457641602,
+ "eval_runtime": 56.8853,
+ "eval_samples_per_second": 42.928,
+ "eval_steps_per_second": 1.354,
+ "step": 3096
+ },
+ {
+ "epoch": 43.01397379912664,
+ "grad_norm": 0.015199829824268818,
+ "learning_rate": 0.0006,
+ "loss": 7.122735023498535,
+ "step": 3097
+ },
+ {
+ "epoch": 43.02794759825328,
+ "grad_norm": 0.03220256045460701,
+ "learning_rate": 0.0006,
+ "loss": 7.143553733825684,
+ "step": 3098
+ },
+ {
+ "epoch": 43.041921397379916,
+ "grad_norm": 0.014687920920550823,
+ "learning_rate": 0.0006,
+ "loss": 7.080960750579834,
+ "step": 3099
+ },
+ {
+ "epoch": 43.05589519650655,
+ "grad_norm": 0.023522932082414627,
+ "learning_rate": 0.0006,
+ "loss": 7.137356758117676,
+ "step": 3100
+ },
+ {
+ "epoch": 43.069868995633186,
+ "grad_norm": 0.014808216132223606,
+ "learning_rate": 0.0006,
+ "loss": 7.09143590927124,
+ "step": 3101
+ },
+ {
+ "epoch": 43.083842794759825,
+ "grad_norm": 0.029339246451854706,
+ "learning_rate": 0.0006,
+ "loss": 7.122198104858398,
+ "step": 3102
+ },
+ {
+ "epoch": 43.09781659388646,
+ "grad_norm": 0.022953078150749207,
+ "learning_rate": 0.0006,
+ "loss": 7.092505931854248,
+ "step": 3103
+ },
+ {
+ "epoch": 43.1117903930131,
+ "grad_norm": 0.04590052738785744,
+ "learning_rate": 0.0006,
+ "loss": 7.113100051879883,
+ "step": 3104
+ },
+ {
+ "epoch": 43.12576419213974,
+ "grad_norm": 0.0524442233145237,
+ "learning_rate": 0.0006,
+ "loss": 7.147022247314453,
+ "step": 3105
+ },
+ {
+ "epoch": 43.13973799126637,
+ "grad_norm": 0.0315794013440609,
+ "learning_rate": 0.0006,
+ "loss": 7.094045162200928,
+ "step": 3106
+ },
+ {
+ "epoch": 43.15371179039301,
+ "grad_norm": 0.05112722888588905,
+ "learning_rate": 0.0006,
+ "loss": 7.054388523101807,
+ "step": 3107
+ },
+ {
+ "epoch": 43.16768558951965,
+ "grad_norm": 0.03146776929497719,
+ "learning_rate": 0.0006,
+ "loss": 7.116674423217773,
+ "step": 3108
+ },
+ {
+ "epoch": 43.18165938864629,
+ "grad_norm": 0.03391246497631073,
+ "learning_rate": 0.0006,
+ "loss": 7.150345802307129,
+ "step": 3109
+ },
+ {
+ "epoch": 43.19563318777293,
+ "grad_norm": 0.026957903057336807,
+ "learning_rate": 0.0006,
+ "loss": 7.058531284332275,
+ "step": 3110
+ },
+ {
+ "epoch": 43.209606986899566,
+ "grad_norm": 0.041532568633556366,
+ "learning_rate": 0.0006,
+ "loss": 7.085498809814453,
+ "step": 3111
+ },
+ {
+ "epoch": 43.223580786026204,
+ "grad_norm": 0.020001588389277458,
+ "learning_rate": 0.0006,
+ "loss": 7.131487846374512,
+ "step": 3112
+ },
+ {
+ "epoch": 43.237554585152836,
+ "grad_norm": 0.02268712781369686,
+ "learning_rate": 0.0006,
+ "loss": 7.0790863037109375,
+ "step": 3113
+ },
+ {
+ "epoch": 43.251528384279474,
+ "grad_norm": 0.023363925516605377,
+ "learning_rate": 0.0006,
+ "loss": 7.067816734313965,
+ "step": 3114
+ },
+ {
+ "epoch": 43.26550218340611,
+ "grad_norm": 0.019367508590221405,
+ "learning_rate": 0.0006,
+ "loss": 7.042041778564453,
+ "step": 3115
+ },
+ {
+ "epoch": 43.27947598253275,
+ "grad_norm": 0.02745090238749981,
+ "learning_rate": 0.0006,
+ "loss": 7.064733028411865,
+ "step": 3116
+ },
+ {
+ "epoch": 43.29344978165939,
+ "grad_norm": 0.02543802745640278,
+ "learning_rate": 0.0006,
+ "loss": 7.058948040008545,
+ "step": 3117
+ },
+ {
+ "epoch": 43.30742358078603,
+ "grad_norm": 0.018874092027544975,
+ "learning_rate": 0.0006,
+ "loss": 7.009946346282959,
+ "step": 3118
+ },
+ {
+ "epoch": 43.32139737991266,
+ "grad_norm": 0.024662822484970093,
+ "learning_rate": 0.0006,
+ "loss": 7.015913963317871,
+ "step": 3119
+ },
+ {
+ "epoch": 43.3353711790393,
+ "grad_norm": 0.02441416308283806,
+ "learning_rate": 0.0006,
+ "loss": 7.061622142791748,
+ "step": 3120
+ },
+ {
+ "epoch": 43.34934497816594,
+ "grad_norm": 0.031887322664260864,
+ "learning_rate": 0.0006,
+ "loss": 7.046883583068848,
+ "step": 3121
+ },
+ {
+ "epoch": 43.36331877729258,
+ "grad_norm": 0.021930739283561707,
+ "learning_rate": 0.0006,
+ "loss": 7.0309624671936035,
+ "step": 3122
+ },
+ {
+ "epoch": 43.377292576419215,
+ "grad_norm": 0.02722783200442791,
+ "learning_rate": 0.0006,
+ "loss": 7.072924613952637,
+ "step": 3123
+ },
+ {
+ "epoch": 43.391266375545854,
+ "grad_norm": 0.01483179721981287,
+ "learning_rate": 0.0006,
+ "loss": 7.028624534606934,
+ "step": 3124
+ },
+ {
+ "epoch": 43.40524017467249,
+ "grad_norm": 0.024793945252895355,
+ "learning_rate": 0.0006,
+ "loss": 7.001623153686523,
+ "step": 3125
+ },
+ {
+ "epoch": 43.419213973799124,
+ "grad_norm": 0.020148713141679764,
+ "learning_rate": 0.0006,
+ "loss": 6.9938812255859375,
+ "step": 3126
+ },
+ {
+ "epoch": 43.43318777292576,
+ "grad_norm": 0.02413850650191307,
+ "learning_rate": 0.0006,
+ "loss": 7.001243591308594,
+ "step": 3127
+ },
+ {
+ "epoch": 43.4471615720524,
+ "grad_norm": 0.028549203649163246,
+ "learning_rate": 0.0006,
+ "loss": 6.987671375274658,
+ "step": 3128
+ },
+ {
+ "epoch": 43.46113537117904,
+ "grad_norm": 0.025988785549998283,
+ "learning_rate": 0.0006,
+ "loss": 6.9551167488098145,
+ "step": 3129
+ },
+ {
+ "epoch": 43.47510917030568,
+ "grad_norm": 0.027140729129314423,
+ "learning_rate": 0.0006,
+ "loss": 6.9547810554504395,
+ "step": 3130
+ },
+ {
+ "epoch": 43.48908296943232,
+ "grad_norm": 0.030130082741379738,
+ "learning_rate": 0.0006,
+ "loss": 6.993186950683594,
+ "step": 3131
+ },
+ {
+ "epoch": 43.50305676855895,
+ "grad_norm": 0.025996994227170944,
+ "learning_rate": 0.0006,
+ "loss": 6.984624862670898,
+ "step": 3132
+ },
+ {
+ "epoch": 43.51703056768559,
+ "grad_norm": 0.02423008903861046,
+ "learning_rate": 0.0006,
+ "loss": 6.98813009262085,
+ "step": 3133
+ },
+ {
+ "epoch": 43.531004366812226,
+ "grad_norm": 0.029665837064385414,
+ "learning_rate": 0.0006,
+ "loss": 6.986610412597656,
+ "step": 3134
+ },
+ {
+ "epoch": 43.544978165938865,
+ "grad_norm": 0.028768595308065414,
+ "learning_rate": 0.0006,
+ "loss": 6.898797035217285,
+ "step": 3135
+ },
+ {
+ "epoch": 43.5589519650655,
+ "grad_norm": 0.03111208975315094,
+ "learning_rate": 0.0006,
+ "loss": 6.923741817474365,
+ "step": 3136
+ },
+ {
+ "epoch": 43.57292576419214,
+ "grad_norm": 0.0368187390267849,
+ "learning_rate": 0.0006,
+ "loss": 6.994343280792236,
+ "step": 3137
+ },
+ {
+ "epoch": 43.58689956331878,
+ "grad_norm": 0.04076763987541199,
+ "learning_rate": 0.0006,
+ "loss": 6.909458160400391,
+ "step": 3138
+ },
+ {
+ "epoch": 43.60087336244541,
+ "grad_norm": 0.03281061723828316,
+ "learning_rate": 0.0006,
+ "loss": 6.924851417541504,
+ "step": 3139
+ },
+ {
+ "epoch": 43.61484716157205,
+ "grad_norm": 0.025486988946795464,
+ "learning_rate": 0.0006,
+ "loss": 6.962926864624023,
+ "step": 3140
+ },
+ {
+ "epoch": 43.62882096069869,
+ "grad_norm": 0.04583485797047615,
+ "learning_rate": 0.0006,
+ "loss": 6.876555442810059,
+ "step": 3141
+ },
+ {
+ "epoch": 43.64279475982533,
+ "grad_norm": 0.05460785701870918,
+ "learning_rate": 0.0006,
+ "loss": 6.952932357788086,
+ "step": 3142
+ },
+ {
+ "epoch": 43.65676855895197,
+ "grad_norm": 0.040917906910181046,
+ "learning_rate": 0.0006,
+ "loss": 6.9624924659729,
+ "step": 3143
+ },
+ {
+ "epoch": 43.670742358078606,
+ "grad_norm": 0.04009668901562691,
+ "learning_rate": 0.0006,
+ "loss": 6.94033145904541,
+ "step": 3144
+ },
+ {
+ "epoch": 43.68471615720524,
+ "grad_norm": 0.04500625655055046,
+ "learning_rate": 0.0006,
+ "loss": 6.961839199066162,
+ "step": 3145
+ },
+ {
+ "epoch": 43.698689956331876,
+ "grad_norm": 0.03182258829474449,
+ "learning_rate": 0.0006,
+ "loss": 6.930580139160156,
+ "step": 3146
+ },
+ {
+ "epoch": 43.712663755458514,
+ "grad_norm": 0.053621504455804825,
+ "learning_rate": 0.0006,
+ "loss": 6.950235366821289,
+ "step": 3147
+ },
+ {
+ "epoch": 43.72663755458515,
+ "grad_norm": 0.031919289380311966,
+ "learning_rate": 0.0006,
+ "loss": 6.895111560821533,
+ "step": 3148
+ },
+ {
+ "epoch": 43.74061135371179,
+ "grad_norm": 0.042713891714811325,
+ "learning_rate": 0.0006,
+ "loss": 6.940885543823242,
+ "step": 3149
+ },
+ {
+ "epoch": 43.75458515283843,
+ "grad_norm": 0.02385653182864189,
+ "learning_rate": 0.0006,
+ "loss": 6.948801517486572,
+ "step": 3150
+ },
+ {
+ "epoch": 43.76855895196506,
+ "grad_norm": 0.04100663587450981,
+ "learning_rate": 0.0006,
+ "loss": 6.970411777496338,
+ "step": 3151
+ },
+ {
+ "epoch": 43.7825327510917,
+ "grad_norm": 0.02729583904147148,
+ "learning_rate": 0.0006,
+ "loss": 6.899379253387451,
+ "step": 3152
+ },
+ {
+ "epoch": 43.79650655021834,
+ "grad_norm": 0.03128620237112045,
+ "learning_rate": 0.0006,
+ "loss": 6.931319236755371,
+ "step": 3153
+ },
+ {
+ "epoch": 43.81048034934498,
+ "grad_norm": 0.025488659739494324,
+ "learning_rate": 0.0006,
+ "loss": 6.8486409187316895,
+ "step": 3154
+ },
+ {
+ "epoch": 43.82445414847162,
+ "grad_norm": 0.03679376840591431,
+ "learning_rate": 0.0006,
+ "loss": 6.840811252593994,
+ "step": 3155
+ },
+ {
+ "epoch": 43.838427947598255,
+ "grad_norm": 0.020578034222126007,
+ "learning_rate": 0.0006,
+ "loss": 6.9088616371154785,
+ "step": 3156
+ },
+ {
+ "epoch": 43.852401746724894,
+ "grad_norm": 0.02935841865837574,
+ "learning_rate": 0.0006,
+ "loss": 6.806676864624023,
+ "step": 3157
+ },
+ {
+ "epoch": 43.866375545851525,
+ "grad_norm": 0.023319244384765625,
+ "learning_rate": 0.0006,
+ "loss": 6.878457546234131,
+ "step": 3158
+ },
+ {
+ "epoch": 43.880349344978164,
+ "grad_norm": 0.02779589593410492,
+ "learning_rate": 0.0006,
+ "loss": 6.822447776794434,
+ "step": 3159
+ },
+ {
+ "epoch": 43.8943231441048,
+ "grad_norm": 0.01879825070500374,
+ "learning_rate": 0.0006,
+ "loss": 6.854121208190918,
+ "step": 3160
+ },
+ {
+ "epoch": 43.90829694323144,
+ "grad_norm": 0.023286139592528343,
+ "learning_rate": 0.0006,
+ "loss": 6.836827278137207,
+ "step": 3161
+ },
+ {
+ "epoch": 43.92227074235808,
+ "grad_norm": 0.030358297750353813,
+ "learning_rate": 0.0006,
+ "loss": 6.860037803649902,
+ "step": 3162
+ },
+ {
+ "epoch": 43.93624454148472,
+ "grad_norm": 0.029621105641126633,
+ "learning_rate": 0.0006,
+ "loss": 6.884025573730469,
+ "step": 3163
+ },
+ {
+ "epoch": 43.95021834061135,
+ "grad_norm": 0.029829824343323708,
+ "learning_rate": 0.0006,
+ "loss": 6.8702392578125,
+ "step": 3164
+ },
+ {
+ "epoch": 43.96419213973799,
+ "grad_norm": 0.036892473697662354,
+ "learning_rate": 0.0006,
+ "loss": 6.806652069091797,
+ "step": 3165
+ },
+ {
+ "epoch": 43.97816593886463,
+ "grad_norm": 0.03206709027290344,
+ "learning_rate": 0.0006,
+ "loss": 6.828495979309082,
+ "step": 3166
+ },
+ {
+ "epoch": 43.992139737991266,
+ "grad_norm": 0.02744000405073166,
+ "learning_rate": 0.0006,
+ "loss": 6.832002639770508,
+ "step": 3167
+ },
+ {
+ "epoch": 44.0,
+ "grad_norm": 0.02205660194158554,
+ "learning_rate": 0.0006,
+ "loss": 6.898828506469727,
+ "step": 3168
+ },
+ {
+ "epoch": 44.0,
+ "eval_loss": 6.8150553703308105,
+ "eval_runtime": 56.784,
+ "eval_samples_per_second": 43.005,
+ "eval_steps_per_second": 1.356,
+ "step": 3168
+ },
+ {
+ "epoch": 44.01397379912664,
+ "grad_norm": 0.030421260744333267,
+ "learning_rate": 0.0006,
+ "loss": 6.775057792663574,
+ "step": 3169
+ },
+ {
+ "epoch": 44.02794759825328,
+ "grad_norm": 0.0465611033141613,
+ "learning_rate": 0.0006,
+ "loss": 6.849390506744385,
+ "step": 3170
+ },
+ {
+ "epoch": 44.041921397379916,
+ "grad_norm": 0.05977972596883774,
+ "learning_rate": 0.0006,
+ "loss": 6.725576400756836,
+ "step": 3171
+ },
+ {
+ "epoch": 44.05589519650655,
+ "grad_norm": 0.05432697385549545,
+ "learning_rate": 0.0006,
+ "loss": 6.768024444580078,
+ "step": 3172
+ },
+ {
+ "epoch": 44.069868995633186,
+ "grad_norm": 0.04850967228412628,
+ "learning_rate": 0.0006,
+ "loss": 6.799457550048828,
+ "step": 3173
+ },
+ {
+ "epoch": 44.083842794759825,
+ "grad_norm": 0.06201786920428276,
+ "learning_rate": 0.0006,
+ "loss": 6.897234916687012,
+ "step": 3174
+ },
+ {
+ "epoch": 44.09781659388646,
+ "grad_norm": 0.04695591703057289,
+ "learning_rate": 0.0006,
+ "loss": 6.826470375061035,
+ "step": 3175
+ },
+ {
+ "epoch": 44.1117903930131,
+ "grad_norm": 0.03850164636969566,
+ "learning_rate": 0.0006,
+ "loss": 6.835901260375977,
+ "step": 3176
+ },
+ {
+ "epoch": 44.12576419213974,
+ "grad_norm": 0.04250155761837959,
+ "learning_rate": 0.0006,
+ "loss": 6.782310485839844,
+ "step": 3177
+ },
+ {
+ "epoch": 44.13973799126637,
+ "grad_norm": 0.041465479880571365,
+ "learning_rate": 0.0006,
+ "loss": 6.785732746124268,
+ "step": 3178
+ },
+ {
+ "epoch": 44.15371179039301,
+ "grad_norm": 0.02527807280421257,
+ "learning_rate": 0.0006,
+ "loss": 6.78061580657959,
+ "step": 3179
+ },
+ {
+ "epoch": 44.16768558951965,
+ "grad_norm": 0.03660374879837036,
+ "learning_rate": 0.0006,
+ "loss": 6.783992767333984,
+ "step": 3180
+ },
+ {
+ "epoch": 44.18165938864629,
+ "grad_norm": 0.027567964047193527,
+ "learning_rate": 0.0006,
+ "loss": 6.765413284301758,
+ "step": 3181
+ },
+ {
+ "epoch": 44.19563318777293,
+ "grad_norm": 0.024301424622535706,
+ "learning_rate": 0.0006,
+ "loss": 6.760368347167969,
+ "step": 3182
+ },
+ {
+ "epoch": 44.209606986899566,
+ "grad_norm": 0.0266821701079607,
+ "learning_rate": 0.0006,
+ "loss": 6.71417236328125,
+ "step": 3183
+ },
+ {
+ "epoch": 44.223580786026204,
+ "grad_norm": 0.028125090524554253,
+ "learning_rate": 0.0006,
+ "loss": 6.646126747131348,
+ "step": 3184
+ },
+ {
+ "epoch": 44.237554585152836,
+ "grad_norm": 0.0185737032443285,
+ "learning_rate": 0.0006,
+ "loss": 6.712944030761719,
+ "step": 3185
+ },
+ {
+ "epoch": 44.251528384279474,
+ "grad_norm": 0.025836674496531487,
+ "learning_rate": 0.0006,
+ "loss": 6.707030296325684,
+ "step": 3186
+ },
+ {
+ "epoch": 44.26550218340611,
+ "grad_norm": 0.029744870960712433,
+ "learning_rate": 0.0006,
+ "loss": 6.6271071434021,
+ "step": 3187
+ },
+ {
+ "epoch": 44.27947598253275,
+ "grad_norm": 0.03021426685154438,
+ "learning_rate": 0.0006,
+ "loss": 6.677826404571533,
+ "step": 3188
+ },
+ {
+ "epoch": 44.29344978165939,
+ "grad_norm": 0.032727014273405075,
+ "learning_rate": 0.0006,
+ "loss": 6.657764434814453,
+ "step": 3189
+ },
+ {
+ "epoch": 44.30742358078603,
+ "grad_norm": 0.03696778789162636,
+ "learning_rate": 0.0006,
+ "loss": 6.632883548736572,
+ "step": 3190
+ },
+ {
+ "epoch": 44.32139737991266,
+ "grad_norm": 0.034630827605724335,
+ "learning_rate": 0.0006,
+ "loss": 6.759583473205566,
+ "step": 3191
+ },
+ {
+ "epoch": 44.3353711790393,
+ "grad_norm": 0.030517758801579475,
+ "learning_rate": 0.0006,
+ "loss": 6.702268600463867,
+ "step": 3192
+ },
+ {
+ "epoch": 44.34934497816594,
+ "grad_norm": 0.04105190932750702,
+ "learning_rate": 0.0006,
+ "loss": 6.705792427062988,
+ "step": 3193
+ },
+ {
+ "epoch": 44.36331877729258,
+ "grad_norm": 0.06659174710512161,
+ "learning_rate": 0.0006,
+ "loss": 6.73862361907959,
+ "step": 3194
+ },
+ {
+ "epoch": 44.377292576419215,
+ "grad_norm": 0.08013619482517242,
+ "learning_rate": 0.0006,
+ "loss": 6.710282802581787,
+ "step": 3195
+ },
+ {
+ "epoch": 44.391266375545854,
+ "grad_norm": 0.049850963056087494,
+ "learning_rate": 0.0006,
+ "loss": 6.684974670410156,
+ "step": 3196
+ },
+ {
+ "epoch": 44.40524017467249,
+ "grad_norm": 0.061247408390045166,
+ "learning_rate": 0.0006,
+ "loss": 6.716034412384033,
+ "step": 3197
+ },
+ {
+ "epoch": 44.419213973799124,
+ "grad_norm": 0.05868857726454735,
+ "learning_rate": 0.0006,
+ "loss": 6.589107513427734,
+ "step": 3198
+ },
+ {
+ "epoch": 44.43318777292576,
+ "grad_norm": 0.048878561705350876,
+ "learning_rate": 0.0006,
+ "loss": 6.694855690002441,
+ "step": 3199
+ },
+ {
+ "epoch": 44.4471615720524,
+ "grad_norm": 0.028748324140906334,
+ "learning_rate": 0.0006,
+ "loss": 6.629969596862793,
+ "step": 3200
+ },
+ {
+ "epoch": 44.46113537117904,
+ "grad_norm": 0.04060306400060654,
+ "learning_rate": 0.0006,
+ "loss": 6.721431255340576,
+ "step": 3201
+ },
+ {
+ "epoch": 44.47510917030568,
+ "grad_norm": 0.03833824768662453,
+ "learning_rate": 0.0006,
+ "loss": 6.625029563903809,
+ "step": 3202
+ },
+ {
+ "epoch": 44.48908296943232,
+ "grad_norm": 0.03791997581720352,
+ "learning_rate": 0.0006,
+ "loss": 6.669187545776367,
+ "step": 3203
+ },
+ {
+ "epoch": 44.50305676855895,
+ "grad_norm": 0.03259684890508652,
+ "learning_rate": 0.0006,
+ "loss": 6.665643692016602,
+ "step": 3204
+ },
+ {
+ "epoch": 44.51703056768559,
+ "grad_norm": 0.040658626705408096,
+ "learning_rate": 0.0006,
+ "loss": 6.639796257019043,
+ "step": 3205
+ },
+ {
+ "epoch": 44.531004366812226,
+ "grad_norm": 0.03691193461418152,
+ "learning_rate": 0.0006,
+ "loss": 6.692244529724121,
+ "step": 3206
+ },
+ {
+ "epoch": 44.544978165938865,
+ "grad_norm": 0.02814812771975994,
+ "learning_rate": 0.0006,
+ "loss": 6.61978006362915,
+ "step": 3207
+ },
+ {
+ "epoch": 44.5589519650655,
+ "grad_norm": 0.022155074402689934,
+ "learning_rate": 0.0006,
+ "loss": 6.625905990600586,
+ "step": 3208
+ },
+ {
+ "epoch": 44.57292576419214,
+ "grad_norm": 0.030794885009527206,
+ "learning_rate": 0.0006,
+ "loss": 6.678500652313232,
+ "step": 3209
+ },
+ {
+ "epoch": 44.58689956331878,
+ "grad_norm": 0.03201768547296524,
+ "learning_rate": 0.0006,
+ "loss": 6.6015729904174805,
+ "step": 3210
+ },
+ {
+ "epoch": 44.60087336244541,
+ "grad_norm": 0.025138361379504204,
+ "learning_rate": 0.0006,
+ "loss": 6.568234443664551,
+ "step": 3211
+ },
+ {
+ "epoch": 44.61484716157205,
+ "grad_norm": 0.02969353087246418,
+ "learning_rate": 0.0006,
+ "loss": 6.588366508483887,
+ "step": 3212
+ },
+ {
+ "epoch": 44.62882096069869,
+ "grad_norm": 0.03136962279677391,
+ "learning_rate": 0.0006,
+ "loss": 6.479961395263672,
+ "step": 3213
+ },
+ {
+ "epoch": 44.64279475982533,
+ "grad_norm": 0.030178170651197433,
+ "learning_rate": 0.0006,
+ "loss": 6.622910976409912,
+ "step": 3214
+ },
+ {
+ "epoch": 44.65676855895197,
+ "grad_norm": 0.02682245895266533,
+ "learning_rate": 0.0006,
+ "loss": 6.597707748413086,
+ "step": 3215
+ },
+ {
+ "epoch": 44.670742358078606,
+ "grad_norm": 0.025154557079076767,
+ "learning_rate": 0.0006,
+ "loss": 6.630541801452637,
+ "step": 3216
+ },
+ {
+ "epoch": 44.68471615720524,
+ "grad_norm": 0.03564463183283806,
+ "learning_rate": 0.0006,
+ "loss": 6.542823791503906,
+ "step": 3217
+ },
+ {
+ "epoch": 44.698689956331876,
+ "grad_norm": 0.049477651715278625,
+ "learning_rate": 0.0006,
+ "loss": 6.646030426025391,
+ "step": 3218
+ },
+ {
+ "epoch": 44.712663755458514,
+ "grad_norm": 0.06536535918712616,
+ "learning_rate": 0.0006,
+ "loss": 6.595463275909424,
+ "step": 3219
+ },
+ {
+ "epoch": 44.72663755458515,
+ "grad_norm": 0.07710316777229309,
+ "learning_rate": 0.0006,
+ "loss": 6.55524206161499,
+ "step": 3220
+ },
+ {
+ "epoch": 44.74061135371179,
+ "grad_norm": 0.09571163356304169,
+ "learning_rate": 0.0006,
+ "loss": 6.659764289855957,
+ "step": 3221
+ },
+ {
+ "epoch": 44.75458515283843,
+ "grad_norm": 0.055076830089092255,
+ "learning_rate": 0.0006,
+ "loss": 6.658882141113281,
+ "step": 3222
+ },
+ {
+ "epoch": 44.76855895196506,
+ "grad_norm": 0.07671099156141281,
+ "learning_rate": 0.0006,
+ "loss": 6.605809211730957,
+ "step": 3223
+ },
+ {
+ "epoch": 44.7825327510917,
+ "grad_norm": 0.05295584350824356,
+ "learning_rate": 0.0006,
+ "loss": 6.641455173492432,
+ "step": 3224
+ },
+ {
+ "epoch": 44.79650655021834,
+ "grad_norm": 0.09128042310476303,
+ "learning_rate": 0.0006,
+ "loss": 6.691066741943359,
+ "step": 3225
+ },
+ {
+ "epoch": 44.81048034934498,
+ "grad_norm": 0.05362958461046219,
+ "learning_rate": 0.0006,
+ "loss": 6.573406219482422,
+ "step": 3226
+ },
+ {
+ "epoch": 44.82445414847162,
+ "grad_norm": 0.046910062432289124,
+ "learning_rate": 0.0006,
+ "loss": 6.550876617431641,
+ "step": 3227
+ },
+ {
+ "epoch": 44.838427947598255,
+ "grad_norm": 0.03494197130203247,
+ "learning_rate": 0.0006,
+ "loss": 6.540179252624512,
+ "step": 3228
+ },
+ {
+ "epoch": 44.852401746724894,
+ "grad_norm": 0.04716387391090393,
+ "learning_rate": 0.0006,
+ "loss": 6.490459442138672,
+ "step": 3229
+ },
+ {
+ "epoch": 44.866375545851525,
+ "grad_norm": 0.04190506786108017,
+ "learning_rate": 0.0006,
+ "loss": 6.509925365447998,
+ "step": 3230
+ },
+ {
+ "epoch": 44.880349344978164,
+ "grad_norm": 0.04749064892530441,
+ "learning_rate": 0.0006,
+ "loss": 6.478779315948486,
+ "step": 3231
+ },
+ {
+ "epoch": 44.8943231441048,
+ "grad_norm": 0.028002185747027397,
+ "learning_rate": 0.0006,
+ "loss": 6.528904914855957,
+ "step": 3232
+ },
+ {
+ "epoch": 44.90829694323144,
+ "grad_norm": 0.036473166197538376,
+ "learning_rate": 0.0006,
+ "loss": 6.549042701721191,
+ "step": 3233
+ },
+ {
+ "epoch": 44.92227074235808,
+ "grad_norm": 0.02543315477669239,
+ "learning_rate": 0.0006,
+ "loss": 6.498875141143799,
+ "step": 3234
+ },
+ {
+ "epoch": 44.93624454148472,
+ "grad_norm": 0.029906807467341423,
+ "learning_rate": 0.0006,
+ "loss": 6.453019142150879,
+ "step": 3235
+ },
+ {
+ "epoch": 44.95021834061135,
+ "grad_norm": 0.027231857180595398,
+ "learning_rate": 0.0006,
+ "loss": 6.428829193115234,
+ "step": 3236
+ },
+ {
+ "epoch": 44.96419213973799,
+ "grad_norm": 0.02506154775619507,
+ "learning_rate": 0.0006,
+ "loss": 6.420858383178711,
+ "step": 3237
+ },
+ {
+ "epoch": 44.97816593886463,
+ "grad_norm": 0.0269472673535347,
+ "learning_rate": 0.0006,
+ "loss": 6.484989643096924,
+ "step": 3238
+ },
+ {
+ "epoch": 44.992139737991266,
+ "grad_norm": 0.023410487920045853,
+ "learning_rate": 0.0006,
+ "loss": 6.454830169677734,
+ "step": 3239
+ },
+ {
+ "epoch": 45.0,
+ "grad_norm": 0.034175023436546326,
+ "learning_rate": 0.0006,
+ "loss": 6.430975437164307,
+ "step": 3240
+ },
+ {
+ "epoch": 45.0,
+ "eval_loss": 6.458896160125732,
+ "eval_runtime": 56.6934,
+ "eval_samples_per_second": 43.074,
+ "eval_steps_per_second": 1.358,
+ "step": 3240
+ },
+ {
+ "epoch": 45.01397379912664,
+ "grad_norm": 0.045031629502773285,
+ "learning_rate": 0.0006,
+ "loss": 6.4243011474609375,
+ "step": 3241
+ },
+ {
+ "epoch": 45.02794759825328,
+ "grad_norm": 0.04968734085559845,
+ "learning_rate": 0.0006,
+ "loss": 6.380091667175293,
+ "step": 3242
+ },
+ {
+ "epoch": 45.041921397379916,
+ "grad_norm": 0.047145187854766846,
+ "learning_rate": 0.0006,
+ "loss": 6.477178573608398,
+ "step": 3243
+ },
+ {
+ "epoch": 45.05589519650655,
+ "grad_norm": 0.04304948449134827,
+ "learning_rate": 0.0006,
+ "loss": 6.399726867675781,
+ "step": 3244
+ },
+ {
+ "epoch": 45.069868995633186,
+ "grad_norm": 0.04466010630130768,
+ "learning_rate": 0.0006,
+ "loss": 6.3415679931640625,
+ "step": 3245
+ },
+ {
+ "epoch": 45.083842794759825,
+ "grad_norm": 0.04305306077003479,
+ "learning_rate": 0.0006,
+ "loss": 6.3413310050964355,
+ "step": 3246
+ },
+ {
+ "epoch": 45.09781659388646,
+ "grad_norm": 0.043064575642347336,
+ "learning_rate": 0.0006,
+ "loss": 6.4838714599609375,
+ "step": 3247
+ },
+ {
+ "epoch": 45.1117903930131,
+ "grad_norm": 0.049503277987241745,
+ "learning_rate": 0.0006,
+ "loss": 6.38473653793335,
+ "step": 3248
+ },
+ {
+ "epoch": 45.12576419213974,
+ "grad_norm": 0.04613238573074341,
+ "learning_rate": 0.0006,
+ "loss": 6.370769500732422,
+ "step": 3249
+ },
+ {
+ "epoch": 45.13973799126637,
+ "grad_norm": 0.052119478583335876,
+ "learning_rate": 0.0006,
+ "loss": 6.362321853637695,
+ "step": 3250
+ },
+ {
+ "epoch": 45.15371179039301,
+ "grad_norm": 0.045110076665878296,
+ "learning_rate": 0.0006,
+ "loss": 6.369844913482666,
+ "step": 3251
+ },
+ {
+ "epoch": 45.16768558951965,
+ "grad_norm": 0.03798583149909973,
+ "learning_rate": 0.0006,
+ "loss": 6.324633598327637,
+ "step": 3252
+ },
+ {
+ "epoch": 45.18165938864629,
+ "grad_norm": 0.03744659945368767,
+ "learning_rate": 0.0006,
+ "loss": 6.349421501159668,
+ "step": 3253
+ },
+ {
+ "epoch": 45.19563318777293,
+ "grad_norm": 0.03767292574048042,
+ "learning_rate": 0.0006,
+ "loss": 6.334238052368164,
+ "step": 3254
+ },
+ {
+ "epoch": 45.209606986899566,
+ "grad_norm": 0.03577509894967079,
+ "learning_rate": 0.0006,
+ "loss": 6.3453826904296875,
+ "step": 3255
+ },
+ {
+ "epoch": 45.223580786026204,
+ "grad_norm": 0.03403240442276001,
+ "learning_rate": 0.0006,
+ "loss": 6.2836174964904785,
+ "step": 3256
+ },
+ {
+ "epoch": 45.237554585152836,
+ "grad_norm": 0.03517331928014755,
+ "learning_rate": 0.0006,
+ "loss": 6.281718730926514,
+ "step": 3257
+ },
+ {
+ "epoch": 45.251528384279474,
+ "grad_norm": 0.04409139230847359,
+ "learning_rate": 0.0006,
+ "loss": 6.320815086364746,
+ "step": 3258
+ },
+ {
+ "epoch": 45.26550218340611,
+ "grad_norm": 0.054750364273786545,
+ "learning_rate": 0.0006,
+ "loss": 6.343618392944336,
+ "step": 3259
+ },
+ {
+ "epoch": 45.27947598253275,
+ "grad_norm": 0.07612067461013794,
+ "learning_rate": 0.0006,
+ "loss": 6.310614585876465,
+ "step": 3260
+ },
+ {
+ "epoch": 45.29344978165939,
+ "grad_norm": 0.08438985049724579,
+ "learning_rate": 0.0006,
+ "loss": 6.2369184494018555,
+ "step": 3261
+ },
+ {
+ "epoch": 45.30742358078603,
+ "grad_norm": 0.0782201811671257,
+ "learning_rate": 0.0006,
+ "loss": 6.254452705383301,
+ "step": 3262
+ },
+ {
+ "epoch": 45.32139737991266,
+ "grad_norm": 0.09928355365991592,
+ "learning_rate": 0.0006,
+ "loss": 6.3112077713012695,
+ "step": 3263
+ },
+ {
+ "epoch": 45.3353711790393,
+ "grad_norm": 0.07233133912086487,
+ "learning_rate": 0.0006,
+ "loss": 6.408087253570557,
+ "step": 3264
+ },
+ {
+ "epoch": 45.34934497816594,
+ "grad_norm": 0.07478684186935425,
+ "learning_rate": 0.0006,
+ "loss": 6.3770036697387695,
+ "step": 3265
+ },
+ {
+ "epoch": 45.36331877729258,
+ "grad_norm": 0.06036846712231636,
+ "learning_rate": 0.0006,
+ "loss": 6.33865213394165,
+ "step": 3266
+ },
+ {
+ "epoch": 45.377292576419215,
+ "grad_norm": 0.05783458054065704,
+ "learning_rate": 0.0006,
+ "loss": 6.363203048706055,
+ "step": 3267
+ },
+ {
+ "epoch": 45.391266375545854,
+ "grad_norm": 0.05443909019231796,
+ "learning_rate": 0.0006,
+ "loss": 6.287069797515869,
+ "step": 3268
+ },
+ {
+ "epoch": 45.40524017467249,
+ "grad_norm": 0.0557909794151783,
+ "learning_rate": 0.0006,
+ "loss": 6.297784805297852,
+ "step": 3269
+ },
+ {
+ "epoch": 45.419213973799124,
+ "grad_norm": 0.03483372926712036,
+ "learning_rate": 0.0006,
+ "loss": 6.280734062194824,
+ "step": 3270
+ },
+ {
+ "epoch": 45.43318777292576,
+ "grad_norm": 0.0425543375313282,
+ "learning_rate": 0.0006,
+ "loss": 6.33565616607666,
+ "step": 3271
+ },
+ {
+ "epoch": 45.4471615720524,
+ "grad_norm": 0.03230159357190132,
+ "learning_rate": 0.0006,
+ "loss": 6.197698593139648,
+ "step": 3272
+ },
+ {
+ "epoch": 45.46113537117904,
+ "grad_norm": 0.036897290498018265,
+ "learning_rate": 0.0006,
+ "loss": 6.241239547729492,
+ "step": 3273
+ },
+ {
+ "epoch": 45.47510917030568,
+ "grad_norm": 0.031103838235139847,
+ "learning_rate": 0.0006,
+ "loss": 6.159435749053955,
+ "step": 3274
+ },
+ {
+ "epoch": 45.48908296943232,
+ "grad_norm": 0.027116449549794197,
+ "learning_rate": 0.0006,
+ "loss": 6.212225914001465,
+ "step": 3275
+ },
+ {
+ "epoch": 45.50305676855895,
+ "grad_norm": 0.029923981055617332,
+ "learning_rate": 0.0006,
+ "loss": 6.157269477844238,
+ "step": 3276
+ },
+ {
+ "epoch": 45.51703056768559,
+ "grad_norm": 0.02521538734436035,
+ "learning_rate": 0.0006,
+ "loss": 6.146918773651123,
+ "step": 3277
+ },
+ {
+ "epoch": 45.531004366812226,
+ "grad_norm": 0.02758920192718506,
+ "learning_rate": 0.0006,
+ "loss": 6.0446977615356445,
+ "step": 3278
+ },
+ {
+ "epoch": 45.544978165938865,
+ "grad_norm": 0.024328220635652542,
+ "learning_rate": 0.0006,
+ "loss": 6.099508285522461,
+ "step": 3279
+ },
+ {
+ "epoch": 45.5589519650655,
+ "grad_norm": 0.027425771579146385,
+ "learning_rate": 0.0006,
+ "loss": 6.084299087524414,
+ "step": 3280
+ },
+ {
+ "epoch": 45.57292576419214,
+ "grad_norm": 0.026927651837468147,
+ "learning_rate": 0.0006,
+ "loss": 6.146334171295166,
+ "step": 3281
+ },
+ {
+ "epoch": 45.58689956331878,
+ "grad_norm": 0.04054379090666771,
+ "learning_rate": 0.0006,
+ "loss": 6.1301493644714355,
+ "step": 3282
+ },
+ {
+ "epoch": 45.60087336244541,
+ "grad_norm": 0.04456290975213051,
+ "learning_rate": 0.0006,
+ "loss": 6.152511119842529,
+ "step": 3283
+ },
+ {
+ "epoch": 45.61484716157205,
+ "grad_norm": 0.052821505814790726,
+ "learning_rate": 0.0006,
+ "loss": 6.113544940948486,
+ "step": 3284
+ },
+ {
+ "epoch": 45.62882096069869,
+ "grad_norm": 0.07251739501953125,
+ "learning_rate": 0.0006,
+ "loss": 6.118298530578613,
+ "step": 3285
+ },
+ {
+ "epoch": 45.64279475982533,
+ "grad_norm": 0.09433556348085403,
+ "learning_rate": 0.0006,
+ "loss": 6.170897483825684,
+ "step": 3286
+ },
+ {
+ "epoch": 45.65676855895197,
+ "grad_norm": 0.057683493942022324,
+ "learning_rate": 0.0006,
+ "loss": 6.1747822761535645,
+ "step": 3287
+ },
+ {
+ "epoch": 45.670742358078606,
+ "grad_norm": 0.06363270431756973,
+ "learning_rate": 0.0006,
+ "loss": 6.1617889404296875,
+ "step": 3288
+ },
+ {
+ "epoch": 45.68471615720524,
+ "grad_norm": 0.07528169453144073,
+ "learning_rate": 0.0006,
+ "loss": 6.230060577392578,
+ "step": 3289
+ },
+ {
+ "epoch": 45.698689956331876,
+ "grad_norm": 0.07816170156002045,
+ "learning_rate": 0.0006,
+ "loss": 6.171257019042969,
+ "step": 3290
+ },
+ {
+ "epoch": 45.712663755458514,
+ "grad_norm": 0.05860164389014244,
+ "learning_rate": 0.0006,
+ "loss": 6.095484256744385,
+ "step": 3291
+ },
+ {
+ "epoch": 45.72663755458515,
+ "grad_norm": 0.04747429117560387,
+ "learning_rate": 0.0006,
+ "loss": 6.060067176818848,
+ "step": 3292
+ },
+ {
+ "epoch": 45.74061135371179,
+ "grad_norm": 0.04514474794268608,
+ "learning_rate": 0.0006,
+ "loss": 6.100218772888184,
+ "step": 3293
+ },
+ {
+ "epoch": 45.75458515283843,
+ "grad_norm": 0.05168681964278221,
+ "learning_rate": 0.0006,
+ "loss": 6.0621867179870605,
+ "step": 3294
+ },
+ {
+ "epoch": 45.76855895196506,
+ "grad_norm": 0.04665876924991608,
+ "learning_rate": 0.0006,
+ "loss": 6.118292808532715,
+ "step": 3295
+ },
+ {
+ "epoch": 45.7825327510917,
+ "grad_norm": 0.049374762922525406,
+ "learning_rate": 0.0006,
+ "loss": 6.098394393920898,
+ "step": 3296
+ },
+ {
+ "epoch": 45.79650655021834,
+ "grad_norm": 0.03793549910187721,
+ "learning_rate": 0.0006,
+ "loss": 6.069541931152344,
+ "step": 3297
+ },
+ {
+ "epoch": 45.81048034934498,
+ "grad_norm": 0.036362383514642715,
+ "learning_rate": 0.0006,
+ "loss": 6.1188435554504395,
+ "step": 3298
+ },
+ {
+ "epoch": 45.82445414847162,
+ "grad_norm": 0.03397097811102867,
+ "learning_rate": 0.0006,
+ "loss": 6.117589950561523,
+ "step": 3299
+ },
+ {
+ "epoch": 45.838427947598255,
+ "grad_norm": 0.030977753922343254,
+ "learning_rate": 0.0006,
+ "loss": 6.0236711502075195,
+ "step": 3300
+ },
+ {
+ "epoch": 45.852401746724894,
+ "grad_norm": 0.02800217643380165,
+ "learning_rate": 0.0006,
+ "loss": 6.034825325012207,
+ "step": 3301
+ },
+ {
+ "epoch": 45.866375545851525,
+ "grad_norm": 0.03129778802394867,
+ "learning_rate": 0.0006,
+ "loss": 5.980271339416504,
+ "step": 3302
+ },
+ {
+ "epoch": 45.880349344978164,
+ "grad_norm": 0.025886094197630882,
+ "learning_rate": 0.0006,
+ "loss": 5.966419219970703,
+ "step": 3303
+ },
+ {
+ "epoch": 45.8943231441048,
+ "grad_norm": 0.025712328031659126,
+ "learning_rate": 0.0006,
+ "loss": 5.938652515411377,
+ "step": 3304
+ },
+ {
+ "epoch": 45.90829694323144,
+ "grad_norm": 0.022508298978209496,
+ "learning_rate": 0.0006,
+ "loss": 5.982312202453613,
+ "step": 3305
+ },
+ {
+ "epoch": 45.92227074235808,
+ "grad_norm": 0.024880951270461082,
+ "learning_rate": 0.0006,
+ "loss": 5.931150436401367,
+ "step": 3306
+ },
+ {
+ "epoch": 45.93624454148472,
+ "grad_norm": 0.030007576569914818,
+ "learning_rate": 0.0006,
+ "loss": 5.936639308929443,
+ "step": 3307
+ },
+ {
+ "epoch": 45.95021834061135,
+ "grad_norm": 0.033314041793346405,
+ "learning_rate": 0.0006,
+ "loss": 5.944248676300049,
+ "step": 3308
+ },
+ {
+ "epoch": 45.96419213973799,
+ "grad_norm": 0.043078381568193436,
+ "learning_rate": 0.0006,
+ "loss": 5.971286773681641,
+ "step": 3309
+ },
+ {
+ "epoch": 45.97816593886463,
+ "grad_norm": 0.059296440333127975,
+ "learning_rate": 0.0006,
+ "loss": 5.910497665405273,
+ "step": 3310
+ },
+ {
+ "epoch": 45.992139737991266,
+ "grad_norm": 0.06232666224241257,
+ "learning_rate": 0.0006,
+ "loss": 5.929817199707031,
+ "step": 3311
+ },
+ {
+ "epoch": 46.0,
+ "grad_norm": 0.05303940549492836,
+ "learning_rate": 0.0006,
+ "loss": 5.910822868347168,
+ "step": 3312
+ },
+ {
+ "epoch": 46.0,
+ "eval_loss": 5.977851390838623,
+ "eval_runtime": 56.2435,
+ "eval_samples_per_second": 43.418,
+ "eval_steps_per_second": 1.369,
+ "step": 3312
+ },
+ {
+ "epoch": 46.01397379912664,
+ "grad_norm": 0.06943798810243607,
+ "learning_rate": 0.0006,
+ "loss": 5.854922294616699,
+ "step": 3313
+ },
+ {
+ "epoch": 46.02794759825328,
+ "grad_norm": 0.05236278846859932,
+ "learning_rate": 0.0006,
+ "loss": 5.876611232757568,
+ "step": 3314
+ },
+ {
+ "epoch": 46.041921397379916,
+ "grad_norm": 0.06240001320838928,
+ "learning_rate": 0.0006,
+ "loss": 5.878912448883057,
+ "step": 3315
+ },
+ {
+ "epoch": 46.05589519650655,
+ "grad_norm": 0.07401192933320999,
+ "learning_rate": 0.0006,
+ "loss": 5.985136032104492,
+ "step": 3316
+ },
+ {
+ "epoch": 46.069868995633186,
+ "grad_norm": 0.0826529711484909,
+ "learning_rate": 0.0006,
+ "loss": 5.939439296722412,
+ "step": 3317
+ },
+ {
+ "epoch": 46.083842794759825,
+ "grad_norm": 0.07999005168676376,
+ "learning_rate": 0.0006,
+ "loss": 5.982336044311523,
+ "step": 3318
+ },
+ {
+ "epoch": 46.09781659388646,
+ "grad_norm": 0.07014869153499603,
+ "learning_rate": 0.0006,
+ "loss": 5.896534442901611,
+ "step": 3319
+ },
+ {
+ "epoch": 46.1117903930131,
+ "grad_norm": 0.05795128643512726,
+ "learning_rate": 0.0006,
+ "loss": 5.843111991882324,
+ "step": 3320
+ },
+ {
+ "epoch": 46.12576419213974,
+ "grad_norm": 0.06374258548021317,
+ "learning_rate": 0.0006,
+ "loss": 5.78752326965332,
+ "step": 3321
+ },
+ {
+ "epoch": 46.13973799126637,
+ "grad_norm": 0.045524805784225464,
+ "learning_rate": 0.0006,
+ "loss": 5.86755895614624,
+ "step": 3322
+ },
+ {
+ "epoch": 46.15371179039301,
+ "grad_norm": 0.03773504123091698,
+ "learning_rate": 0.0006,
+ "loss": 5.83627462387085,
+ "step": 3323
+ },
+ {
+ "epoch": 46.16768558951965,
+ "grad_norm": 0.035815201699733734,
+ "learning_rate": 0.0006,
+ "loss": 5.798238277435303,
+ "step": 3324
+ },
+ {
+ "epoch": 46.18165938864629,
+ "grad_norm": 0.03756783902645111,
+ "learning_rate": 0.0006,
+ "loss": 5.845197677612305,
+ "step": 3325
+ },
+ {
+ "epoch": 46.19563318777293,
+ "grad_norm": 0.03569113463163376,
+ "learning_rate": 0.0006,
+ "loss": 5.888932704925537,
+ "step": 3326
+ },
+ {
+ "epoch": 46.209606986899566,
+ "grad_norm": 0.029467813670635223,
+ "learning_rate": 0.0006,
+ "loss": 5.792047500610352,
+ "step": 3327
+ },
+ {
+ "epoch": 46.223580786026204,
+ "grad_norm": 0.025698406621813774,
+ "learning_rate": 0.0006,
+ "loss": 5.783734321594238,
+ "step": 3328
+ },
+ {
+ "epoch": 46.237554585152836,
+ "grad_norm": 0.028426671400666237,
+ "learning_rate": 0.0006,
+ "loss": 5.762084007263184,
+ "step": 3329
+ },
+ {
+ "epoch": 46.251528384279474,
+ "grad_norm": 0.03099622018635273,
+ "learning_rate": 0.0006,
+ "loss": 5.719109058380127,
+ "step": 3330
+ },
+ {
+ "epoch": 46.26550218340611,
+ "grad_norm": 0.0310219656676054,
+ "learning_rate": 0.0006,
+ "loss": 5.737057209014893,
+ "step": 3331
+ },
+ {
+ "epoch": 46.27947598253275,
+ "grad_norm": 0.02713446132838726,
+ "learning_rate": 0.0006,
+ "loss": 5.758657455444336,
+ "step": 3332
+ },
+ {
+ "epoch": 46.29344978165939,
+ "grad_norm": 0.03099844604730606,
+ "learning_rate": 0.0006,
+ "loss": 5.742086410522461,
+ "step": 3333
+ },
+ {
+ "epoch": 46.30742358078603,
+ "grad_norm": 0.030258730053901672,
+ "learning_rate": 0.0006,
+ "loss": 5.767426490783691,
+ "step": 3334
+ },
+ {
+ "epoch": 46.32139737991266,
+ "grad_norm": 0.029034852981567383,
+ "learning_rate": 0.0006,
+ "loss": 5.651792049407959,
+ "step": 3335
+ },
+ {
+ "epoch": 46.3353711790393,
+ "grad_norm": 0.03138665854930878,
+ "learning_rate": 0.0006,
+ "loss": 5.664980888366699,
+ "step": 3336
+ },
+ {
+ "epoch": 46.34934497816594,
+ "grad_norm": 0.040546901524066925,
+ "learning_rate": 0.0006,
+ "loss": 5.665165901184082,
+ "step": 3337
+ },
+ {
+ "epoch": 46.36331877729258,
+ "grad_norm": 0.06395485252141953,
+ "learning_rate": 0.0006,
+ "loss": 5.7776713371276855,
+ "step": 3338
+ },
+ {
+ "epoch": 46.377292576419215,
+ "grad_norm": 0.10088056325912476,
+ "learning_rate": 0.0006,
+ "loss": 5.755645751953125,
+ "step": 3339
+ },
+ {
+ "epoch": 46.391266375545854,
+ "grad_norm": 0.09527938067913055,
+ "learning_rate": 0.0006,
+ "loss": 5.838809490203857,
+ "step": 3340
+ },
+ {
+ "epoch": 46.40524017467249,
+ "grad_norm": 0.0849648267030716,
+ "learning_rate": 0.0006,
+ "loss": 5.781932830810547,
+ "step": 3341
+ },
+ {
+ "epoch": 46.419213973799124,
+ "grad_norm": 0.10142666101455688,
+ "learning_rate": 0.0006,
+ "loss": 5.838970184326172,
+ "step": 3342
+ },
+ {
+ "epoch": 46.43318777292576,
+ "grad_norm": 0.42731454968452454,
+ "learning_rate": 0.0006,
+ "loss": 6.071902275085449,
+ "step": 3343
+ },
+ {
+ "epoch": 46.4471615720524,
+ "grad_norm": 0.13102488219738007,
+ "learning_rate": 0.0006,
+ "loss": 6.8635993003845215,
+ "step": 3344
+ },
+ {
+ "epoch": 46.46113537117904,
+ "grad_norm": 0.23691938817501068,
+ "learning_rate": 0.0006,
+ "loss": 7.313158988952637,
+ "step": 3345
+ },
+ {
+ "epoch": 46.47510917030568,
+ "grad_norm": 0.11670669168233871,
+ "learning_rate": 0.0006,
+ "loss": 7.240363121032715,
+ "step": 3346
+ },
+ {
+ "epoch": 46.48908296943232,
+ "grad_norm": 0.09985820949077606,
+ "learning_rate": 0.0006,
+ "loss": 7.113332271575928,
+ "step": 3347
+ },
+ {
+ "epoch": 46.50305676855895,
+ "grad_norm": 0.11406408250331879,
+ "learning_rate": 0.0006,
+ "loss": 7.051285743713379,
+ "step": 3348
+ },
+ {
+ "epoch": 46.51703056768559,
+ "grad_norm": 0.07900336384773254,
+ "learning_rate": 0.0006,
+ "loss": 7.014365196228027,
+ "step": 3349
+ },
+ {
+ "epoch": 46.531004366812226,
+ "grad_norm": 0.05380333960056305,
+ "learning_rate": 0.0006,
+ "loss": 6.8868818283081055,
+ "step": 3350
+ },
+ {
+ "epoch": 46.544978165938865,
+ "grad_norm": 0.05868449807167053,
+ "learning_rate": 0.0006,
+ "loss": 6.783599853515625,
+ "step": 3351
+ },
+ {
+ "epoch": 46.5589519650655,
+ "grad_norm": 0.04428785294294357,
+ "learning_rate": 0.0006,
+ "loss": 6.798511505126953,
+ "step": 3352
+ },
+ {
+ "epoch": 46.57292576419214,
+ "grad_norm": 0.055527277290821075,
+ "learning_rate": 0.0006,
+ "loss": 6.694507598876953,
+ "step": 3353
+ },
+ {
+ "epoch": 46.58689956331878,
+ "grad_norm": 0.03872432932257652,
+ "learning_rate": 0.0006,
+ "loss": 6.736631393432617,
+ "step": 3354
+ },
+ {
+ "epoch": 46.60087336244541,
+ "grad_norm": 0.04330779239535332,
+ "learning_rate": 0.0006,
+ "loss": 6.576806545257568,
+ "step": 3355
+ },
+ {
+ "epoch": 46.61484716157205,
+ "grad_norm": 0.036767736077308655,
+ "learning_rate": 0.0006,
+ "loss": 6.559871196746826,
+ "step": 3356
+ },
+ {
+ "epoch": 46.62882096069869,
+ "grad_norm": 0.039857108145952225,
+ "learning_rate": 0.0006,
+ "loss": 6.5749831199646,
+ "step": 3357
+ },
+ {
+ "epoch": 46.64279475982533,
+ "grad_norm": 0.048719972372055054,
+ "learning_rate": 0.0006,
+ "loss": 6.4915618896484375,
+ "step": 3358
+ },
+ {
+ "epoch": 46.65676855895197,
+ "grad_norm": 0.042849015444517136,
+ "learning_rate": 0.0006,
+ "loss": 6.413516044616699,
+ "step": 3359
+ },
+ {
+ "epoch": 46.670742358078606,
+ "grad_norm": 0.02871009334921837,
+ "learning_rate": 0.0006,
+ "loss": 6.320952415466309,
+ "step": 3360
+ },
+ {
+ "epoch": 46.68471615720524,
+ "grad_norm": 0.03193620592355728,
+ "learning_rate": 0.0006,
+ "loss": 6.4176435470581055,
+ "step": 3361
+ },
+ {
+ "epoch": 46.698689956331876,
+ "grad_norm": 0.027417024597525597,
+ "learning_rate": 0.0006,
+ "loss": 6.342319488525391,
+ "step": 3362
+ },
+ {
+ "epoch": 46.712663755458514,
+ "grad_norm": 0.03305937349796295,
+ "learning_rate": 0.0006,
+ "loss": 6.246112823486328,
+ "step": 3363
+ },
+ {
+ "epoch": 46.72663755458515,
+ "grad_norm": 0.031403183937072754,
+ "learning_rate": 0.0006,
+ "loss": 6.149503231048584,
+ "step": 3364
+ },
+ {
+ "epoch": 46.74061135371179,
+ "grad_norm": 0.031563788652420044,
+ "learning_rate": 0.0006,
+ "loss": 6.278808116912842,
+ "step": 3365
+ },
+ {
+ "epoch": 46.75458515283843,
+ "grad_norm": 0.028366589918732643,
+ "learning_rate": 0.0006,
+ "loss": 6.057036876678467,
+ "step": 3366
+ },
+ {
+ "epoch": 46.76855895196506,
+ "grad_norm": 0.023715803399682045,
+ "learning_rate": 0.0006,
+ "loss": 6.111324310302734,
+ "step": 3367
+ },
+ {
+ "epoch": 46.7825327510917,
+ "grad_norm": 0.026546811684966087,
+ "learning_rate": 0.0006,
+ "loss": 6.105669975280762,
+ "step": 3368
+ },
+ {
+ "epoch": 46.79650655021834,
+ "grad_norm": 0.023531019687652588,
+ "learning_rate": 0.0006,
+ "loss": 6.068072319030762,
+ "step": 3369
+ },
+ {
+ "epoch": 46.81048034934498,
+ "grad_norm": 0.024480128660798073,
+ "learning_rate": 0.0006,
+ "loss": 5.947896957397461,
+ "step": 3370
+ },
+ {
+ "epoch": 46.82445414847162,
+ "grad_norm": 0.02517310529947281,
+ "learning_rate": 0.0006,
+ "loss": 6.030037879943848,
+ "step": 3371
+ },
+ {
+ "epoch": 46.838427947598255,
+ "grad_norm": 0.026723403483629227,
+ "learning_rate": 0.0006,
+ "loss": 5.963776111602783,
+ "step": 3372
+ },
+ {
+ "epoch": 46.852401746724894,
+ "grad_norm": 0.03163504600524902,
+ "learning_rate": 0.0006,
+ "loss": 5.893466949462891,
+ "step": 3373
+ },
+ {
+ "epoch": 46.866375545851525,
+ "grad_norm": 0.04058683291077614,
+ "learning_rate": 0.0006,
+ "loss": 5.9202799797058105,
+ "step": 3374
+ },
+ {
+ "epoch": 46.880349344978164,
+ "grad_norm": 0.04964117333292961,
+ "learning_rate": 0.0006,
+ "loss": 5.9366559982299805,
+ "step": 3375
+ },
+ {
+ "epoch": 46.8943231441048,
+ "grad_norm": 0.055406033992767334,
+ "learning_rate": 0.0006,
+ "loss": 5.912230968475342,
+ "step": 3376
+ },
+ {
+ "epoch": 46.90829694323144,
+ "grad_norm": 0.04643625393509865,
+ "learning_rate": 0.0006,
+ "loss": 5.87260103225708,
+ "step": 3377
+ },
+ {
+ "epoch": 46.92227074235808,
+ "grad_norm": 0.029325876384973526,
+ "learning_rate": 0.0006,
+ "loss": 5.854763984680176,
+ "step": 3378
+ },
+ {
+ "epoch": 46.93624454148472,
+ "grad_norm": 0.03159382566809654,
+ "learning_rate": 0.0006,
+ "loss": 5.880882263183594,
+ "step": 3379
+ },
+ {
+ "epoch": 46.95021834061135,
+ "grad_norm": 0.030224012210965157,
+ "learning_rate": 0.0006,
+ "loss": 5.787360191345215,
+ "step": 3380
+ },
+ {
+ "epoch": 46.96419213973799,
+ "grad_norm": 0.028368130326271057,
+ "learning_rate": 0.0006,
+ "loss": 5.817195892333984,
+ "step": 3381
+ },
+ {
+ "epoch": 46.97816593886463,
+ "grad_norm": 0.029411399737000465,
+ "learning_rate": 0.0006,
+ "loss": 5.783403396606445,
+ "step": 3382
+ },
+ {
+ "epoch": 46.992139737991266,
+ "grad_norm": 0.029927462339401245,
+ "learning_rate": 0.0006,
+ "loss": 5.662444591522217,
+ "step": 3383
+ },
+ {
+ "epoch": 47.0,
+ "grad_norm": 0.02872961387038231,
+ "learning_rate": 0.0006,
+ "loss": 5.742781639099121,
+ "step": 3384
+ },
+ {
+ "epoch": 47.0,
+ "eval_loss": 5.870260238647461,
+ "eval_runtime": 57.3734,
+ "eval_samples_per_second": 42.563,
+ "eval_steps_per_second": 1.342,
+ "step": 3384
+ },
+ {
+ "epoch": 47.01397379912664,
+ "grad_norm": 0.028542131185531616,
+ "learning_rate": 0.0006,
+ "loss": 5.753256797790527,
+ "step": 3385
+ },
+ {
+ "epoch": 47.02794759825328,
+ "grad_norm": 0.03227119892835617,
+ "learning_rate": 0.0006,
+ "loss": 5.722246170043945,
+ "step": 3386
+ },
+ {
+ "epoch": 47.041921397379916,
+ "grad_norm": 0.031461723148822784,
+ "learning_rate": 0.0006,
+ "loss": 5.740520000457764,
+ "step": 3387
+ },
+ {
+ "epoch": 47.05589519650655,
+ "grad_norm": 0.03184220567345619,
+ "learning_rate": 0.0006,
+ "loss": 5.720890045166016,
+ "step": 3388
+ },
+ {
+ "epoch": 47.069868995633186,
+ "grad_norm": 0.030015580356121063,
+ "learning_rate": 0.0006,
+ "loss": 5.597814559936523,
+ "step": 3389
+ },
+ {
+ "epoch": 47.083842794759825,
+ "grad_norm": 0.029931779950857162,
+ "learning_rate": 0.0006,
+ "loss": 5.5412468910217285,
+ "step": 3390
+ },
+ {
+ "epoch": 47.09781659388646,
+ "grad_norm": 0.03206505626440048,
+ "learning_rate": 0.0006,
+ "loss": 5.6967597007751465,
+ "step": 3391
+ },
+ {
+ "epoch": 47.1117903930131,
+ "grad_norm": 0.03978161886334419,
+ "learning_rate": 0.0006,
+ "loss": 5.59661340713501,
+ "step": 3392
+ },
+ {
+ "epoch": 47.12576419213974,
+ "grad_norm": 0.04619080200791359,
+ "learning_rate": 0.0006,
+ "loss": 5.795788764953613,
+ "step": 3393
+ },
+ {
+ "epoch": 47.13973799126637,
+ "grad_norm": 0.049085382372140884,
+ "learning_rate": 0.0006,
+ "loss": 5.667335510253906,
+ "step": 3394
+ },
+ {
+ "epoch": 47.15371179039301,
+ "grad_norm": 0.04830858111381531,
+ "learning_rate": 0.0006,
+ "loss": 5.668806552886963,
+ "step": 3395
+ },
+ {
+ "epoch": 47.16768558951965,
+ "grad_norm": 0.0583178736269474,
+ "learning_rate": 0.0006,
+ "loss": 5.692285537719727,
+ "step": 3396
+ },
+ {
+ "epoch": 47.18165938864629,
+ "grad_norm": 0.05182253196835518,
+ "learning_rate": 0.0006,
+ "loss": 5.678645133972168,
+ "step": 3397
+ },
+ {
+ "epoch": 47.19563318777293,
+ "grad_norm": 0.04011905938386917,
+ "learning_rate": 0.0006,
+ "loss": 5.661550998687744,
+ "step": 3398
+ },
+ {
+ "epoch": 47.209606986899566,
+ "grad_norm": 0.03888486698269844,
+ "learning_rate": 0.0006,
+ "loss": 5.643001556396484,
+ "step": 3399
+ },
+ {
+ "epoch": 47.223580786026204,
+ "grad_norm": 0.030658265575766563,
+ "learning_rate": 0.0006,
+ "loss": 5.659090995788574,
+ "step": 3400
+ },
+ {
+ "epoch": 47.237554585152836,
+ "grad_norm": 0.03609256446361542,
+ "learning_rate": 0.0006,
+ "loss": 5.544501781463623,
+ "step": 3401
+ },
+ {
+ "epoch": 47.251528384279474,
+ "grad_norm": 0.033589769154787064,
+ "learning_rate": 0.0006,
+ "loss": 5.463878631591797,
+ "step": 3402
+ },
+ {
+ "epoch": 47.26550218340611,
+ "grad_norm": 0.035128287971019745,
+ "learning_rate": 0.0006,
+ "loss": 5.5680060386657715,
+ "step": 3403
+ },
+ {
+ "epoch": 47.27947598253275,
+ "grad_norm": 0.038684550672769547,
+ "learning_rate": 0.0006,
+ "loss": 5.587898254394531,
+ "step": 3404
+ },
+ {
+ "epoch": 47.29344978165939,
+ "grad_norm": 0.043566226959228516,
+ "learning_rate": 0.0006,
+ "loss": 5.580896854400635,
+ "step": 3405
+ },
+ {
+ "epoch": 47.30742358078603,
+ "grad_norm": 0.04428374394774437,
+ "learning_rate": 0.0006,
+ "loss": 5.603460311889648,
+ "step": 3406
+ },
+ {
+ "epoch": 47.32139737991266,
+ "grad_norm": 0.037368934601545334,
+ "learning_rate": 0.0006,
+ "loss": 5.564740180969238,
+ "step": 3407
+ },
+ {
+ "epoch": 47.3353711790393,
+ "grad_norm": 0.04092350974678993,
+ "learning_rate": 0.0006,
+ "loss": 5.584153652191162,
+ "step": 3408
+ },
+ {
+ "epoch": 47.34934497816594,
+ "grad_norm": 0.045031506568193436,
+ "learning_rate": 0.0006,
+ "loss": 5.601186275482178,
+ "step": 3409
+ },
+ {
+ "epoch": 47.36331877729258,
+ "grad_norm": 0.041830409318208694,
+ "learning_rate": 0.0006,
+ "loss": 5.512788772583008,
+ "step": 3410
+ },
+ {
+ "epoch": 47.377292576419215,
+ "grad_norm": 0.037633974105119705,
+ "learning_rate": 0.0006,
+ "loss": 5.503828525543213,
+ "step": 3411
+ },
+ {
+ "epoch": 47.391266375545854,
+ "grad_norm": 0.03473588079214096,
+ "learning_rate": 0.0006,
+ "loss": 5.570557594299316,
+ "step": 3412
+ },
+ {
+ "epoch": 47.40524017467249,
+ "grad_norm": 0.039325181394815445,
+ "learning_rate": 0.0006,
+ "loss": 5.427783966064453,
+ "step": 3413
+ },
+ {
+ "epoch": 47.419213973799124,
+ "grad_norm": 0.04034624621272087,
+ "learning_rate": 0.0006,
+ "loss": 5.572591304779053,
+ "step": 3414
+ },
+ {
+ "epoch": 47.43318777292576,
+ "grad_norm": 0.05245184153318405,
+ "learning_rate": 0.0006,
+ "loss": 5.505208969116211,
+ "step": 3415
+ },
+ {
+ "epoch": 47.4471615720524,
+ "grad_norm": 0.0568675734102726,
+ "learning_rate": 0.0006,
+ "loss": 5.589373588562012,
+ "step": 3416
+ },
+ {
+ "epoch": 47.46113537117904,
+ "grad_norm": 0.048336587846279144,
+ "learning_rate": 0.0006,
+ "loss": 5.522005081176758,
+ "step": 3417
+ },
+ {
+ "epoch": 47.47510917030568,
+ "grad_norm": 0.04194975271821022,
+ "learning_rate": 0.0006,
+ "loss": 5.387576103210449,
+ "step": 3418
+ },
+ {
+ "epoch": 47.48908296943232,
+ "grad_norm": 0.04019433259963989,
+ "learning_rate": 0.0006,
+ "loss": 5.522615432739258,
+ "step": 3419
+ },
+ {
+ "epoch": 47.50305676855895,
+ "grad_norm": 0.05709933117032051,
+ "learning_rate": 0.0006,
+ "loss": 5.420286178588867,
+ "step": 3420
+ },
+ {
+ "epoch": 47.51703056768559,
+ "grad_norm": 0.09130338579416275,
+ "learning_rate": 0.0006,
+ "loss": 5.484622478485107,
+ "step": 3421
+ },
+ {
+ "epoch": 47.531004366812226,
+ "grad_norm": 0.2757541835308075,
+ "learning_rate": 0.0006,
+ "loss": 5.821832656860352,
+ "step": 3422
+ },
+ {
+ "epoch": 47.544978165938865,
+ "grad_norm": 0.28292080760002136,
+ "learning_rate": 0.0006,
+ "loss": 6.579934120178223,
+ "step": 3423
+ },
+ {
+ "epoch": 47.5589519650655,
+ "grad_norm": 0.15293604135513306,
+ "learning_rate": 0.0006,
+ "loss": 6.478213310241699,
+ "step": 3424
+ },
+ {
+ "epoch": 47.57292576419214,
+ "grad_norm": 0.11284393817186356,
+ "learning_rate": 0.0006,
+ "loss": 6.4066386222839355,
+ "step": 3425
+ },
+ {
+ "epoch": 47.58689956331878,
+ "grad_norm": 0.10933073610067368,
+ "learning_rate": 0.0006,
+ "loss": 6.3306169509887695,
+ "step": 3426
+ },
+ {
+ "epoch": 47.60087336244541,
+ "grad_norm": 0.05837641656398773,
+ "learning_rate": 0.0006,
+ "loss": 6.250000953674316,
+ "step": 3427
+ },
+ {
+ "epoch": 47.61484716157205,
+ "grad_norm": 0.05987229570746422,
+ "learning_rate": 0.0006,
+ "loss": 6.152463436126709,
+ "step": 3428
+ },
+ {
+ "epoch": 47.62882096069869,
+ "grad_norm": 0.04064938426017761,
+ "learning_rate": 0.0006,
+ "loss": 6.117897033691406,
+ "step": 3429
+ },
+ {
+ "epoch": 47.64279475982533,
+ "grad_norm": 0.04565057530999184,
+ "learning_rate": 0.0006,
+ "loss": 6.042291641235352,
+ "step": 3430
+ },
+ {
+ "epoch": 47.65676855895197,
+ "grad_norm": 0.03475063294172287,
+ "learning_rate": 0.0006,
+ "loss": 6.0071868896484375,
+ "step": 3431
+ },
+ {
+ "epoch": 47.670742358078606,
+ "grad_norm": 0.03809220343828201,
+ "learning_rate": 0.0006,
+ "loss": 6.0128021240234375,
+ "step": 3432
+ },
+ {
+ "epoch": 47.68471615720524,
+ "grad_norm": 0.03307219594717026,
+ "learning_rate": 0.0006,
+ "loss": 5.894664287567139,
+ "step": 3433
+ },
+ {
+ "epoch": 47.698689956331876,
+ "grad_norm": 0.028824346140027046,
+ "learning_rate": 0.0006,
+ "loss": 5.914865493774414,
+ "step": 3434
+ },
+ {
+ "epoch": 47.712663755458514,
+ "grad_norm": 0.027923690155148506,
+ "learning_rate": 0.0006,
+ "loss": 5.851759910583496,
+ "step": 3435
+ },
+ {
+ "epoch": 47.72663755458515,
+ "grad_norm": 0.025492684915661812,
+ "learning_rate": 0.0006,
+ "loss": 5.749186038970947,
+ "step": 3436
+ },
+ {
+ "epoch": 47.74061135371179,
+ "grad_norm": 0.025105174630880356,
+ "learning_rate": 0.0006,
+ "loss": 5.803566932678223,
+ "step": 3437
+ },
+ {
+ "epoch": 47.75458515283843,
+ "grad_norm": 0.023298880085349083,
+ "learning_rate": 0.0006,
+ "loss": 5.76276159286499,
+ "step": 3438
+ },
+ {
+ "epoch": 47.76855895196506,
+ "grad_norm": 0.02468576282262802,
+ "learning_rate": 0.0006,
+ "loss": 5.833852767944336,
+ "step": 3439
+ },
+ {
+ "epoch": 47.7825327510917,
+ "grad_norm": 0.022730689495801926,
+ "learning_rate": 0.0006,
+ "loss": 5.736724376678467,
+ "step": 3440
+ },
+ {
+ "epoch": 47.79650655021834,
+ "grad_norm": 0.021335354074835777,
+ "learning_rate": 0.0006,
+ "loss": 5.644248008728027,
+ "step": 3441
+ },
+ {
+ "epoch": 47.81048034934498,
+ "grad_norm": 0.02288101427257061,
+ "learning_rate": 0.0006,
+ "loss": 5.7011942863464355,
+ "step": 3442
+ },
+ {
+ "epoch": 47.82445414847162,
+ "grad_norm": 0.01961585320532322,
+ "learning_rate": 0.0006,
+ "loss": 5.63215446472168,
+ "step": 3443
+ },
+ {
+ "epoch": 47.838427947598255,
+ "grad_norm": 0.017659205943346024,
+ "learning_rate": 0.0006,
+ "loss": 5.744845390319824,
+ "step": 3444
+ },
+ {
+ "epoch": 47.852401746724894,
+ "grad_norm": 0.02281784638762474,
+ "learning_rate": 0.0006,
+ "loss": 5.696993350982666,
+ "step": 3445
+ },
+ {
+ "epoch": 47.866375545851525,
+ "grad_norm": 0.016394242644309998,
+ "learning_rate": 0.0006,
+ "loss": 5.659870147705078,
+ "step": 3446
+ },
+ {
+ "epoch": 47.880349344978164,
+ "grad_norm": 0.01967368647456169,
+ "learning_rate": 0.0006,
+ "loss": 5.638072490692139,
+ "step": 3447
+ },
+ {
+ "epoch": 47.8943231441048,
+ "grad_norm": 0.017326267436146736,
+ "learning_rate": 0.0006,
+ "loss": 5.638836860656738,
+ "step": 3448
+ },
+ {
+ "epoch": 47.90829694323144,
+ "grad_norm": 0.017778463661670685,
+ "learning_rate": 0.0006,
+ "loss": 5.634466648101807,
+ "step": 3449
+ },
+ {
+ "epoch": 47.92227074235808,
+ "grad_norm": 0.017269758507609367,
+ "learning_rate": 0.0006,
+ "loss": 5.506277561187744,
+ "step": 3450
+ },
+ {
+ "epoch": 47.93624454148472,
+ "grad_norm": 0.01733289286494255,
+ "learning_rate": 0.0006,
+ "loss": 5.544455528259277,
+ "step": 3451
+ },
+ {
+ "epoch": 47.95021834061135,
+ "grad_norm": 0.017787795513868332,
+ "learning_rate": 0.0006,
+ "loss": 5.498943328857422,
+ "step": 3452
+ },
+ {
+ "epoch": 47.96419213973799,
+ "grad_norm": 0.015770932659506798,
+ "learning_rate": 0.0006,
+ "loss": 5.537227630615234,
+ "step": 3453
+ },
+ {
+ "epoch": 47.97816593886463,
+ "grad_norm": 0.016132809221744537,
+ "learning_rate": 0.0006,
+ "loss": 5.489081382751465,
+ "step": 3454
+ },
+ {
+ "epoch": 47.992139737991266,
+ "grad_norm": 0.01749664731323719,
+ "learning_rate": 0.0006,
+ "loss": 5.454263687133789,
+ "step": 3455
+ },
+ {
+ "epoch": 48.0,
+ "grad_norm": 0.016711296513676643,
+ "learning_rate": 0.0006,
+ "loss": 5.530783653259277,
+ "step": 3456
+ },
+ {
+ "epoch": 48.0,
+ "eval_loss": 5.596925258636475,
+ "eval_runtime": 56.9612,
+ "eval_samples_per_second": 42.871,
+ "eval_steps_per_second": 1.352,
+ "step": 3456
+ },
+ {
+ "epoch": 48.01397379912664,
+ "grad_norm": 0.01867641694843769,
+ "learning_rate": 0.0006,
+ "loss": 5.506853103637695,
+ "step": 3457
+ },
+ {
+ "epoch": 48.02794759825328,
+ "grad_norm": 0.023250792175531387,
+ "learning_rate": 0.0006,
+ "loss": 5.4074177742004395,
+ "step": 3458
+ },
+ {
+ "epoch": 48.041921397379916,
+ "grad_norm": 0.01970936730504036,
+ "learning_rate": 0.0006,
+ "loss": 5.508777618408203,
+ "step": 3459
+ },
+ {
+ "epoch": 48.05589519650655,
+ "grad_norm": 0.01879260502755642,
+ "learning_rate": 0.0006,
+ "loss": 5.412424087524414,
+ "step": 3460
+ },
+ {
+ "epoch": 48.069868995633186,
+ "grad_norm": 0.017420092597603798,
+ "learning_rate": 0.0006,
+ "loss": 5.43361759185791,
+ "step": 3461
+ },
+ {
+ "epoch": 48.083842794759825,
+ "grad_norm": 0.018707862123847008,
+ "learning_rate": 0.0006,
+ "loss": 5.476800918579102,
+ "step": 3462
+ },
+ {
+ "epoch": 48.09781659388646,
+ "grad_norm": 0.023185882717370987,
+ "learning_rate": 0.0006,
+ "loss": 5.465272426605225,
+ "step": 3463
+ },
+ {
+ "epoch": 48.1117903930131,
+ "grad_norm": 0.023570368066430092,
+ "learning_rate": 0.0006,
+ "loss": 5.391471862792969,
+ "step": 3464
+ },
+ {
+ "epoch": 48.12576419213974,
+ "grad_norm": 0.02651144750416279,
+ "learning_rate": 0.0006,
+ "loss": 5.422799110412598,
+ "step": 3465
+ },
+ {
+ "epoch": 48.13973799126637,
+ "grad_norm": 0.03316329047083855,
+ "learning_rate": 0.0006,
+ "loss": 5.3952202796936035,
+ "step": 3466
+ },
+ {
+ "epoch": 48.15371179039301,
+ "grad_norm": 0.04012278467416763,
+ "learning_rate": 0.0006,
+ "loss": 5.540853023529053,
+ "step": 3467
+ },
+ {
+ "epoch": 48.16768558951965,
+ "grad_norm": 0.04763203114271164,
+ "learning_rate": 0.0006,
+ "loss": 5.387890338897705,
+ "step": 3468
+ },
+ {
+ "epoch": 48.18165938864629,
+ "grad_norm": 0.05129165202379227,
+ "learning_rate": 0.0006,
+ "loss": 5.414722442626953,
+ "step": 3469
+ },
+ {
+ "epoch": 48.19563318777293,
+ "grad_norm": 0.04438071325421333,
+ "learning_rate": 0.0006,
+ "loss": 5.386244773864746,
+ "step": 3470
+ },
+ {
+ "epoch": 48.209606986899566,
+ "grad_norm": 0.040234919637441635,
+ "learning_rate": 0.0006,
+ "loss": 5.431617736816406,
+ "step": 3471
+ },
+ {
+ "epoch": 48.223580786026204,
+ "grad_norm": 0.03217008337378502,
+ "learning_rate": 0.0006,
+ "loss": 5.30168342590332,
+ "step": 3472
+ },
+ {
+ "epoch": 48.237554585152836,
+ "grad_norm": 0.03087974339723587,
+ "learning_rate": 0.0006,
+ "loss": 5.316389083862305,
+ "step": 3473
+ },
+ {
+ "epoch": 48.251528384279474,
+ "grad_norm": 0.02789347991347313,
+ "learning_rate": 0.0006,
+ "loss": 5.385970592498779,
+ "step": 3474
+ },
+ {
+ "epoch": 48.26550218340611,
+ "grad_norm": 0.024792097508907318,
+ "learning_rate": 0.0006,
+ "loss": 5.362329959869385,
+ "step": 3475
+ },
+ {
+ "epoch": 48.27947598253275,
+ "grad_norm": 0.025869522243738174,
+ "learning_rate": 0.0006,
+ "loss": 5.28400182723999,
+ "step": 3476
+ },
+ {
+ "epoch": 48.29344978165939,
+ "grad_norm": 0.021339882165193558,
+ "learning_rate": 0.0006,
+ "loss": 5.36669921875,
+ "step": 3477
+ },
+ {
+ "epoch": 48.30742358078603,
+ "grad_norm": 0.021462520584464073,
+ "learning_rate": 0.0006,
+ "loss": 5.311222076416016,
+ "step": 3478
+ },
+ {
+ "epoch": 48.32139737991266,
+ "grad_norm": 0.02193879708647728,
+ "learning_rate": 0.0006,
+ "loss": 5.291441917419434,
+ "step": 3479
+ },
+ {
+ "epoch": 48.3353711790393,
+ "grad_norm": 0.01963093690574169,
+ "learning_rate": 0.0006,
+ "loss": 5.342780113220215,
+ "step": 3480
+ },
+ {
+ "epoch": 48.34934497816594,
+ "grad_norm": 0.020634736865758896,
+ "learning_rate": 0.0006,
+ "loss": 5.144141674041748,
+ "step": 3481
+ },
+ {
+ "epoch": 48.36331877729258,
+ "grad_norm": 0.02126968279480934,
+ "learning_rate": 0.0006,
+ "loss": 5.232624530792236,
+ "step": 3482
+ },
+ {
+ "epoch": 48.377292576419215,
+ "grad_norm": 0.021203994750976562,
+ "learning_rate": 0.0006,
+ "loss": 5.290219783782959,
+ "step": 3483
+ },
+ {
+ "epoch": 48.391266375545854,
+ "grad_norm": 0.02193225920200348,
+ "learning_rate": 0.0006,
+ "loss": 5.29197359085083,
+ "step": 3484
+ },
+ {
+ "epoch": 48.40524017467249,
+ "grad_norm": 0.023174211382865906,
+ "learning_rate": 0.0006,
+ "loss": 5.204349517822266,
+ "step": 3485
+ },
+ {
+ "epoch": 48.419213973799124,
+ "grad_norm": 0.024161914363503456,
+ "learning_rate": 0.0006,
+ "loss": 5.339351654052734,
+ "step": 3486
+ },
+ {
+ "epoch": 48.43318777292576,
+ "grad_norm": 0.02221524342894554,
+ "learning_rate": 0.0006,
+ "loss": 5.263001441955566,
+ "step": 3487
+ },
+ {
+ "epoch": 48.4471615720524,
+ "grad_norm": 0.026632830500602722,
+ "learning_rate": 0.0006,
+ "loss": 5.221281051635742,
+ "step": 3488
+ },
+ {
+ "epoch": 48.46113537117904,
+ "grad_norm": 0.034923870116472244,
+ "learning_rate": 0.0006,
+ "loss": 5.296039581298828,
+ "step": 3489
+ },
+ {
+ "epoch": 48.47510917030568,
+ "grad_norm": 0.046184979379177094,
+ "learning_rate": 0.0006,
+ "loss": 5.212769985198975,
+ "step": 3490
+ },
+ {
+ "epoch": 48.48908296943232,
+ "grad_norm": 0.05154098570346832,
+ "learning_rate": 0.0006,
+ "loss": 5.337864875793457,
+ "step": 3491
+ },
+ {
+ "epoch": 48.50305676855895,
+ "grad_norm": 0.042967259883880615,
+ "learning_rate": 0.0006,
+ "loss": 5.267958641052246,
+ "step": 3492
+ },
+ {
+ "epoch": 48.51703056768559,
+ "grad_norm": 0.05356583371758461,
+ "learning_rate": 0.0006,
+ "loss": 5.1404218673706055,
+ "step": 3493
+ },
+ {
+ "epoch": 48.531004366812226,
+ "grad_norm": 0.08219032734632492,
+ "learning_rate": 0.0006,
+ "loss": 5.300687789916992,
+ "step": 3494
+ },
+ {
+ "epoch": 48.544978165938865,
+ "grad_norm": 0.19204089045524597,
+ "learning_rate": 0.0006,
+ "loss": 5.456025123596191,
+ "step": 3495
+ },
+ {
+ "epoch": 48.5589519650655,
+ "grad_norm": 0.25917696952819824,
+ "learning_rate": 0.0006,
+ "loss": 6.156761169433594,
+ "step": 3496
+ },
+ {
+ "epoch": 48.57292576419214,
+ "grad_norm": 0.1566283404827118,
+ "learning_rate": 0.0006,
+ "loss": 5.862346649169922,
+ "step": 3497
+ },
+ {
+ "epoch": 48.58689956331878,
+ "grad_norm": 0.1761486828327179,
+ "learning_rate": 0.0006,
+ "loss": 6.012795448303223,
+ "step": 3498
+ },
+ {
+ "epoch": 48.60087336244541,
+ "grad_norm": 0.09857621043920517,
+ "learning_rate": 0.0006,
+ "loss": 5.9248151779174805,
+ "step": 3499
+ },
+ {
+ "epoch": 48.61484716157205,
+ "grad_norm": 0.05968055501580238,
+ "learning_rate": 0.0006,
+ "loss": 5.780706882476807,
+ "step": 3500
+ },
+ {
+ "epoch": 48.62882096069869,
+ "grad_norm": 0.057433564215898514,
+ "learning_rate": 0.0006,
+ "loss": 5.645388603210449,
+ "step": 3501
+ },
+ {
+ "epoch": 48.64279475982533,
+ "grad_norm": 0.039659615606069565,
+ "learning_rate": 0.0006,
+ "loss": 5.695138931274414,
+ "step": 3502
+ },
+ {
+ "epoch": 48.65676855895197,
+ "grad_norm": 0.04019695147871971,
+ "learning_rate": 0.0006,
+ "loss": 5.672435283660889,
+ "step": 3503
+ },
+ {
+ "epoch": 48.670742358078606,
+ "grad_norm": 0.03033018670976162,
+ "learning_rate": 0.0006,
+ "loss": 5.682260513305664,
+ "step": 3504
+ },
+ {
+ "epoch": 48.68471615720524,
+ "grad_norm": 0.03215787187218666,
+ "learning_rate": 0.0006,
+ "loss": 5.604182243347168,
+ "step": 3505
+ },
+ {
+ "epoch": 48.698689956331876,
+ "grad_norm": 0.026865126565098763,
+ "learning_rate": 0.0006,
+ "loss": 5.64040470123291,
+ "step": 3506
+ },
+ {
+ "epoch": 48.712663755458514,
+ "grad_norm": 0.02831609919667244,
+ "learning_rate": 0.0006,
+ "loss": 5.552720069885254,
+ "step": 3507
+ },
+ {
+ "epoch": 48.72663755458515,
+ "grad_norm": 0.024226613342761993,
+ "learning_rate": 0.0006,
+ "loss": 5.556946277618408,
+ "step": 3508
+ },
+ {
+ "epoch": 48.74061135371179,
+ "grad_norm": 0.025393934920430183,
+ "learning_rate": 0.0006,
+ "loss": 5.476178169250488,
+ "step": 3509
+ },
+ {
+ "epoch": 48.75458515283843,
+ "grad_norm": 0.01984412968158722,
+ "learning_rate": 0.0006,
+ "loss": 5.471088409423828,
+ "step": 3510
+ },
+ {
+ "epoch": 48.76855895196506,
+ "grad_norm": 0.02035129815340042,
+ "learning_rate": 0.0006,
+ "loss": 5.438812732696533,
+ "step": 3511
+ },
+ {
+ "epoch": 48.7825327510917,
+ "grad_norm": 0.0208350270986557,
+ "learning_rate": 0.0006,
+ "loss": 5.381746292114258,
+ "step": 3512
+ },
+ {
+ "epoch": 48.79650655021834,
+ "grad_norm": 0.017783882096409798,
+ "learning_rate": 0.0006,
+ "loss": 5.351675510406494,
+ "step": 3513
+ },
+ {
+ "epoch": 48.81048034934498,
+ "grad_norm": 0.01671770215034485,
+ "learning_rate": 0.0006,
+ "loss": 5.329428672790527,
+ "step": 3514
+ },
+ {
+ "epoch": 48.82445414847162,
+ "grad_norm": 0.017891215160489082,
+ "learning_rate": 0.0006,
+ "loss": 5.212074279785156,
+ "step": 3515
+ },
+ {
+ "epoch": 48.838427947598255,
+ "grad_norm": 0.016847791150212288,
+ "learning_rate": 0.0006,
+ "loss": 5.299534797668457,
+ "step": 3516
+ },
+ {
+ "epoch": 48.852401746724894,
+ "grad_norm": 0.01593637838959694,
+ "learning_rate": 0.0006,
+ "loss": 5.30466890335083,
+ "step": 3517
+ },
+ {
+ "epoch": 48.866375545851525,
+ "grad_norm": 0.014713102020323277,
+ "learning_rate": 0.0006,
+ "loss": 5.294345378875732,
+ "step": 3518
+ },
+ {
+ "epoch": 48.880349344978164,
+ "grad_norm": 0.01433857623487711,
+ "learning_rate": 0.0006,
+ "loss": 5.2148542404174805,
+ "step": 3519
+ },
+ {
+ "epoch": 48.8943231441048,
+ "grad_norm": 0.014979379251599312,
+ "learning_rate": 0.0006,
+ "loss": 5.298582077026367,
+ "step": 3520
+ },
+ {
+ "epoch": 48.90829694323144,
+ "grad_norm": 0.013494192622601986,
+ "learning_rate": 0.0006,
+ "loss": 5.309783935546875,
+ "step": 3521
+ },
+ {
+ "epoch": 48.92227074235808,
+ "grad_norm": 0.014020012691617012,
+ "learning_rate": 0.0006,
+ "loss": 5.284879684448242,
+ "step": 3522
+ },
+ {
+ "epoch": 48.93624454148472,
+ "grad_norm": 0.01348058134317398,
+ "learning_rate": 0.0006,
+ "loss": 5.249116897583008,
+ "step": 3523
+ },
+ {
+ "epoch": 48.95021834061135,
+ "grad_norm": 0.013280218467116356,
+ "learning_rate": 0.0006,
+ "loss": 5.273489952087402,
+ "step": 3524
+ },
+ {
+ "epoch": 48.96419213973799,
+ "grad_norm": 0.012647110037505627,
+ "learning_rate": 0.0006,
+ "loss": 5.306652069091797,
+ "step": 3525
+ },
+ {
+ "epoch": 48.97816593886463,
+ "grad_norm": 0.012847724370658398,
+ "learning_rate": 0.0006,
+ "loss": 5.235745429992676,
+ "step": 3526
+ },
+ {
+ "epoch": 48.992139737991266,
+ "grad_norm": 0.011820166371762753,
+ "learning_rate": 0.0006,
+ "loss": 5.238091468811035,
+ "step": 3527
+ },
+ {
+ "epoch": 49.0,
+ "grad_norm": 0.01255585066974163,
+ "learning_rate": 0.0006,
+ "loss": 5.231295585632324,
+ "step": 3528
+ },
+ {
+ "epoch": 49.0,
+ "eval_loss": 5.394587993621826,
+ "eval_runtime": 56.3734,
+ "eval_samples_per_second": 43.318,
+ "eval_steps_per_second": 1.366,
+ "step": 3528
+ },
+ {
+ "epoch": 49.01397379912664,
+ "grad_norm": 0.011031972244381905,
+ "learning_rate": 0.0006,
+ "loss": 5.1354570388793945,
+ "step": 3529
+ },
+ {
+ "epoch": 49.02794759825328,
+ "grad_norm": 0.010631263256072998,
+ "learning_rate": 0.0006,
+ "loss": 5.177639961242676,
+ "step": 3530
+ },
+ {
+ "epoch": 49.041921397379916,
+ "grad_norm": 0.01133775431662798,
+ "learning_rate": 0.0006,
+ "loss": 5.283428192138672,
+ "step": 3531
+ },
+ {
+ "epoch": 49.05589519650655,
+ "grad_norm": 0.01154172234237194,
+ "learning_rate": 0.0006,
+ "loss": 5.206364154815674,
+ "step": 3532
+ },
+ {
+ "epoch": 49.069868995633186,
+ "grad_norm": 0.010192356072366238,
+ "learning_rate": 0.0006,
+ "loss": 5.2385735511779785,
+ "step": 3533
+ },
+ {
+ "epoch": 49.083842794759825,
+ "grad_norm": 0.012789200060069561,
+ "learning_rate": 0.0006,
+ "loss": 5.071800231933594,
+ "step": 3534
+ },
+ {
+ "epoch": 49.09781659388646,
+ "grad_norm": 0.011282820254564285,
+ "learning_rate": 0.0006,
+ "loss": 5.218020439147949,
+ "step": 3535
+ },
+ {
+ "epoch": 49.1117903930131,
+ "grad_norm": 0.012153448536992073,
+ "learning_rate": 0.0006,
+ "loss": 5.134698867797852,
+ "step": 3536
+ },
+ {
+ "epoch": 49.12576419213974,
+ "grad_norm": 0.011920543387532234,
+ "learning_rate": 0.0006,
+ "loss": 5.141620635986328,
+ "step": 3537
+ },
+ {
+ "epoch": 49.13973799126637,
+ "grad_norm": 0.010949329473078251,
+ "learning_rate": 0.0006,
+ "loss": 5.116259574890137,
+ "step": 3538
+ },
+ {
+ "epoch": 49.15371179039301,
+ "grad_norm": 0.012362448498606682,
+ "learning_rate": 0.0006,
+ "loss": 5.090177059173584,
+ "step": 3539
+ },
+ {
+ "epoch": 49.16768558951965,
+ "grad_norm": 0.011575071141123772,
+ "learning_rate": 0.0006,
+ "loss": 5.093441009521484,
+ "step": 3540
+ },
+ {
+ "epoch": 49.18165938864629,
+ "grad_norm": 0.011126236990094185,
+ "learning_rate": 0.0006,
+ "loss": 5.060235023498535,
+ "step": 3541
+ },
+ {
+ "epoch": 49.19563318777293,
+ "grad_norm": 0.0107401292771101,
+ "learning_rate": 0.0006,
+ "loss": 5.140704154968262,
+ "step": 3542
+ },
+ {
+ "epoch": 49.209606986899566,
+ "grad_norm": 0.011202906258404255,
+ "learning_rate": 0.0006,
+ "loss": 5.10855770111084,
+ "step": 3543
+ },
+ {
+ "epoch": 49.223580786026204,
+ "grad_norm": 0.011083454824984074,
+ "learning_rate": 0.0006,
+ "loss": 5.092751502990723,
+ "step": 3544
+ },
+ {
+ "epoch": 49.237554585152836,
+ "grad_norm": 0.009523800574243069,
+ "learning_rate": 0.0006,
+ "loss": 5.13733434677124,
+ "step": 3545
+ },
+ {
+ "epoch": 49.251528384279474,
+ "grad_norm": 0.011298353783786297,
+ "learning_rate": 0.0006,
+ "loss": 5.036048889160156,
+ "step": 3546
+ },
+ {
+ "epoch": 49.26550218340611,
+ "grad_norm": 0.010676483623683453,
+ "learning_rate": 0.0006,
+ "loss": 5.051901340484619,
+ "step": 3547
+ },
+ {
+ "epoch": 49.27947598253275,
+ "grad_norm": 0.011463399045169353,
+ "learning_rate": 0.0006,
+ "loss": 4.983170986175537,
+ "step": 3548
+ },
+ {
+ "epoch": 49.29344978165939,
+ "grad_norm": 0.010603402741253376,
+ "learning_rate": 0.0006,
+ "loss": 5.112311363220215,
+ "step": 3549
+ },
+ {
+ "epoch": 49.30742358078603,
+ "grad_norm": 0.010535670444369316,
+ "learning_rate": 0.0006,
+ "loss": 5.085227966308594,
+ "step": 3550
+ },
+ {
+ "epoch": 49.32139737991266,
+ "grad_norm": 0.011843928135931492,
+ "learning_rate": 0.0006,
+ "loss": 5.07575798034668,
+ "step": 3551
+ },
+ {
+ "epoch": 49.3353711790393,
+ "grad_norm": 0.010587952099740505,
+ "learning_rate": 0.0006,
+ "loss": 5.019237041473389,
+ "step": 3552
+ },
+ {
+ "epoch": 49.34934497816594,
+ "grad_norm": 0.012285659089684486,
+ "learning_rate": 0.0006,
+ "loss": 5.144603729248047,
+ "step": 3553
+ },
+ {
+ "epoch": 49.36331877729258,
+ "grad_norm": 0.013119431212544441,
+ "learning_rate": 0.0006,
+ "loss": 5.035256862640381,
+ "step": 3554
+ },
+ {
+ "epoch": 49.377292576419215,
+ "grad_norm": 0.012138537131249905,
+ "learning_rate": 0.0006,
+ "loss": 4.996667861938477,
+ "step": 3555
+ },
+ {
+ "epoch": 49.391266375545854,
+ "grad_norm": 0.012142804451286793,
+ "learning_rate": 0.0006,
+ "loss": 4.95058536529541,
+ "step": 3556
+ },
+ {
+ "epoch": 49.40524017467249,
+ "grad_norm": 0.012205676175653934,
+ "learning_rate": 0.0006,
+ "loss": 5.096931457519531,
+ "step": 3557
+ },
+ {
+ "epoch": 49.419213973799124,
+ "grad_norm": 0.012086744420230389,
+ "learning_rate": 0.0006,
+ "loss": 4.99384880065918,
+ "step": 3558
+ },
+ {
+ "epoch": 49.43318777292576,
+ "grad_norm": 0.012918103486299515,
+ "learning_rate": 0.0006,
+ "loss": 5.031294822692871,
+ "step": 3559
+ },
+ {
+ "epoch": 49.4471615720524,
+ "grad_norm": 0.012006713077425957,
+ "learning_rate": 0.0006,
+ "loss": 5.0547943115234375,
+ "step": 3560
+ },
+ {
+ "epoch": 49.46113537117904,
+ "grad_norm": 0.0120399734005332,
+ "learning_rate": 0.0006,
+ "loss": 5.014238357543945,
+ "step": 3561
+ },
+ {
+ "epoch": 49.47510917030568,
+ "grad_norm": 0.013306990265846252,
+ "learning_rate": 0.0006,
+ "loss": 4.972925186157227,
+ "step": 3562
+ },
+ {
+ "epoch": 49.48908296943232,
+ "grad_norm": 0.013878636993467808,
+ "learning_rate": 0.0006,
+ "loss": 5.079033851623535,
+ "step": 3563
+ },
+ {
+ "epoch": 49.50305676855895,
+ "grad_norm": 0.012454566545784473,
+ "learning_rate": 0.0006,
+ "loss": 4.93708610534668,
+ "step": 3564
+ },
+ {
+ "epoch": 49.51703056768559,
+ "grad_norm": 0.014075061306357384,
+ "learning_rate": 0.0006,
+ "loss": 4.965582370758057,
+ "step": 3565
+ },
+ {
+ "epoch": 49.531004366812226,
+ "grad_norm": 0.012525617145001888,
+ "learning_rate": 0.0006,
+ "loss": 4.968148708343506,
+ "step": 3566
+ },
+ {
+ "epoch": 49.544978165938865,
+ "grad_norm": 0.012418225407600403,
+ "learning_rate": 0.0006,
+ "loss": 5.06901741027832,
+ "step": 3567
+ },
+ {
+ "epoch": 49.5589519650655,
+ "grad_norm": 0.015034239739179611,
+ "learning_rate": 0.0006,
+ "loss": 4.984277725219727,
+ "step": 3568
+ },
+ {
+ "epoch": 49.57292576419214,
+ "grad_norm": 0.014329628087580204,
+ "learning_rate": 0.0006,
+ "loss": 4.951558589935303,
+ "step": 3569
+ },
+ {
+ "epoch": 49.58689956331878,
+ "grad_norm": 0.017769070342183113,
+ "learning_rate": 0.0006,
+ "loss": 4.966532230377197,
+ "step": 3570
+ },
+ {
+ "epoch": 49.60087336244541,
+ "grad_norm": 0.01909453794360161,
+ "learning_rate": 0.0006,
+ "loss": 4.919890880584717,
+ "step": 3571
+ },
+ {
+ "epoch": 49.61484716157205,
+ "grad_norm": 0.017519688233733177,
+ "learning_rate": 0.0006,
+ "loss": 5.031286716461182,
+ "step": 3572
+ },
+ {
+ "epoch": 49.62882096069869,
+ "grad_norm": 0.017143897712230682,
+ "learning_rate": 0.0006,
+ "loss": 4.969764709472656,
+ "step": 3573
+ },
+ {
+ "epoch": 49.64279475982533,
+ "grad_norm": 0.01688285917043686,
+ "learning_rate": 0.0006,
+ "loss": 5.0733442306518555,
+ "step": 3574
+ },
+ {
+ "epoch": 49.65676855895197,
+ "grad_norm": 0.018547311425209045,
+ "learning_rate": 0.0006,
+ "loss": 4.980863571166992,
+ "step": 3575
+ },
+ {
+ "epoch": 49.670742358078606,
+ "grad_norm": 0.018877137452363968,
+ "learning_rate": 0.0006,
+ "loss": 4.954646110534668,
+ "step": 3576
+ },
+ {
+ "epoch": 49.68471615720524,
+ "grad_norm": 0.018681600689888,
+ "learning_rate": 0.0006,
+ "loss": 4.959702014923096,
+ "step": 3577
+ },
+ {
+ "epoch": 49.698689956331876,
+ "grad_norm": 0.018947336822748184,
+ "learning_rate": 0.0006,
+ "loss": 5.070502758026123,
+ "step": 3578
+ },
+ {
+ "epoch": 49.712663755458514,
+ "grad_norm": 0.018817484378814697,
+ "learning_rate": 0.0006,
+ "loss": 5.056508541107178,
+ "step": 3579
+ },
+ {
+ "epoch": 49.72663755458515,
+ "grad_norm": 0.01931724324822426,
+ "learning_rate": 0.0006,
+ "loss": 4.922451972961426,
+ "step": 3580
+ },
+ {
+ "epoch": 49.74061135371179,
+ "grad_norm": 0.022047819569706917,
+ "learning_rate": 0.0006,
+ "loss": 5.053170204162598,
+ "step": 3581
+ },
+ {
+ "epoch": 49.75458515283843,
+ "grad_norm": 0.02279801294207573,
+ "learning_rate": 0.0006,
+ "loss": 4.8768439292907715,
+ "step": 3582
+ },
+ {
+ "epoch": 49.76855895196506,
+ "grad_norm": 0.020661601796746254,
+ "learning_rate": 0.0006,
+ "loss": 4.907928466796875,
+ "step": 3583
+ },
+ {
+ "epoch": 49.7825327510917,
+ "grad_norm": 0.020479509606957436,
+ "learning_rate": 0.0006,
+ "loss": 5.040694236755371,
+ "step": 3584
+ },
+ {
+ "epoch": 49.79650655021834,
+ "grad_norm": 0.021654162555933,
+ "learning_rate": 0.0006,
+ "loss": 5.018434524536133,
+ "step": 3585
+ },
+ {
+ "epoch": 49.81048034934498,
+ "grad_norm": 0.022486824542284012,
+ "learning_rate": 0.0006,
+ "loss": 5.086538314819336,
+ "step": 3586
+ },
+ {
+ "epoch": 49.82445414847162,
+ "grad_norm": 0.02141369879245758,
+ "learning_rate": 0.0006,
+ "loss": 5.035694122314453,
+ "step": 3587
+ },
+ {
+ "epoch": 49.838427947598255,
+ "grad_norm": 0.015926677733659744,
+ "learning_rate": 0.0006,
+ "loss": 4.905038833618164,
+ "step": 3588
+ },
+ {
+ "epoch": 49.852401746724894,
+ "grad_norm": 0.015775298699736595,
+ "learning_rate": 0.0006,
+ "loss": 5.003902435302734,
+ "step": 3589
+ },
+ {
+ "epoch": 49.866375545851525,
+ "grad_norm": 0.01872118003666401,
+ "learning_rate": 0.0006,
+ "loss": 4.891060829162598,
+ "step": 3590
+ },
+ {
+ "epoch": 49.880349344978164,
+ "grad_norm": 0.02066192403435707,
+ "learning_rate": 0.0006,
+ "loss": 5.002897262573242,
+ "step": 3591
+ },
+ {
+ "epoch": 49.8943231441048,
+ "grad_norm": 0.02348714880645275,
+ "learning_rate": 0.0006,
+ "loss": 4.9189958572387695,
+ "step": 3592
+ },
+ {
+ "epoch": 49.90829694323144,
+ "grad_norm": 0.020402103662490845,
+ "learning_rate": 0.0006,
+ "loss": 4.992661476135254,
+ "step": 3593
+ },
+ {
+ "epoch": 49.92227074235808,
+ "grad_norm": 0.020466072484850883,
+ "learning_rate": 0.0006,
+ "loss": 4.962968826293945,
+ "step": 3594
+ },
+ {
+ "epoch": 49.93624454148472,
+ "grad_norm": 0.01922314241528511,
+ "learning_rate": 0.0006,
+ "loss": 4.897730827331543,
+ "step": 3595
+ },
+ {
+ "epoch": 49.95021834061135,
+ "grad_norm": 0.017228754237294197,
+ "learning_rate": 0.0006,
+ "loss": 4.964842796325684,
+ "step": 3596
+ },
+ {
+ "epoch": 49.96419213973799,
+ "grad_norm": 0.016021743416786194,
+ "learning_rate": 0.0006,
+ "loss": 4.910758972167969,
+ "step": 3597
+ },
+ {
+ "epoch": 49.97816593886463,
+ "grad_norm": 0.017985103651881218,
+ "learning_rate": 0.0006,
+ "loss": 4.916072368621826,
+ "step": 3598
+ },
+ {
+ "epoch": 49.992139737991266,
+ "grad_norm": 0.018780240789055824,
+ "learning_rate": 0.0006,
+ "loss": 4.94658088684082,
+ "step": 3599
+ },
+ {
+ "epoch": 50.0,
+ "grad_norm": 0.01776314526796341,
+ "learning_rate": 0.0006,
+ "loss": 4.969357967376709,
+ "step": 3600
+ },
+ {
+ "epoch": 50.0,
+ "eval_loss": 5.141158580780029,
+ "eval_runtime": 56.2332,
+ "eval_samples_per_second": 43.426,
+ "eval_steps_per_second": 1.369,
+ "step": 3600
+ },
+ {
+ "epoch": 50.01397379912664,
+ "grad_norm": 0.016308696940541267,
+ "learning_rate": 0.0006,
+ "loss": 4.924664497375488,
+ "step": 3601
+ },
+ {
+ "epoch": 50.02794759825328,
+ "grad_norm": 0.018810199573636055,
+ "learning_rate": 0.0006,
+ "loss": 4.82293701171875,
+ "step": 3602
+ },
+ {
+ "epoch": 50.041921397379916,
+ "grad_norm": 0.02227160707116127,
+ "learning_rate": 0.0006,
+ "loss": 4.9612298011779785,
+ "step": 3603
+ },
+ {
+ "epoch": 50.05589519650655,
+ "grad_norm": 0.025266462936997414,
+ "learning_rate": 0.0006,
+ "loss": 4.963835716247559,
+ "step": 3604
+ },
+ {
+ "epoch": 50.069868995633186,
+ "grad_norm": 0.028588181361556053,
+ "learning_rate": 0.0006,
+ "loss": 4.992118835449219,
+ "step": 3605
+ },
+ {
+ "epoch": 50.083842794759825,
+ "grad_norm": 0.02739487960934639,
+ "learning_rate": 0.0006,
+ "loss": 4.915547847747803,
+ "step": 3606
+ },
+ {
+ "epoch": 50.09781659388646,
+ "grad_norm": 0.027000300586223602,
+ "learning_rate": 0.0006,
+ "loss": 4.883801460266113,
+ "step": 3607
+ },
+ {
+ "epoch": 50.1117903930131,
+ "grad_norm": 0.023468483239412308,
+ "learning_rate": 0.0006,
+ "loss": 5.006168365478516,
+ "step": 3608
+ },
+ {
+ "epoch": 50.12576419213974,
+ "grad_norm": 0.026753634214401245,
+ "learning_rate": 0.0006,
+ "loss": 4.92526912689209,
+ "step": 3609
+ },
+ {
+ "epoch": 50.13973799126637,
+ "grad_norm": 0.026816733181476593,
+ "learning_rate": 0.0006,
+ "loss": 4.847546100616455,
+ "step": 3610
+ },
+ {
+ "epoch": 50.15371179039301,
+ "grad_norm": 0.027362391352653503,
+ "learning_rate": 0.0006,
+ "loss": 4.862548351287842,
+ "step": 3611
+ },
+ {
+ "epoch": 50.16768558951965,
+ "grad_norm": 0.03450845181941986,
+ "learning_rate": 0.0006,
+ "loss": 4.963202953338623,
+ "step": 3612
+ },
+ {
+ "epoch": 50.18165938864629,
+ "grad_norm": 0.03801408037543297,
+ "learning_rate": 0.0006,
+ "loss": 4.770240306854248,
+ "step": 3613
+ },
+ {
+ "epoch": 50.19563318777293,
+ "grad_norm": 0.03579140082001686,
+ "learning_rate": 0.0006,
+ "loss": 4.797981262207031,
+ "step": 3614
+ },
+ {
+ "epoch": 50.209606986899566,
+ "grad_norm": 0.029393760487437248,
+ "learning_rate": 0.0006,
+ "loss": 4.798581600189209,
+ "step": 3615
+ },
+ {
+ "epoch": 50.223580786026204,
+ "grad_norm": 0.026111142709851265,
+ "learning_rate": 0.0006,
+ "loss": 4.990712642669678,
+ "step": 3616
+ },
+ {
+ "epoch": 50.237554585152836,
+ "grad_norm": 0.02213387005031109,
+ "learning_rate": 0.0006,
+ "loss": 4.865842342376709,
+ "step": 3617
+ },
+ {
+ "epoch": 50.251528384279474,
+ "grad_norm": 0.025450684130191803,
+ "learning_rate": 0.0006,
+ "loss": 4.903746128082275,
+ "step": 3618
+ },
+ {
+ "epoch": 50.26550218340611,
+ "grad_norm": 0.020915117114782333,
+ "learning_rate": 0.0006,
+ "loss": 4.951366424560547,
+ "step": 3619
+ },
+ {
+ "epoch": 50.27947598253275,
+ "grad_norm": 0.020972849801182747,
+ "learning_rate": 0.0006,
+ "loss": 4.783311367034912,
+ "step": 3620
+ },
+ {
+ "epoch": 50.29344978165939,
+ "grad_norm": 0.016956273466348648,
+ "learning_rate": 0.0006,
+ "loss": 4.955767631530762,
+ "step": 3621
+ },
+ {
+ "epoch": 50.30742358078603,
+ "grad_norm": 0.018045252189040184,
+ "learning_rate": 0.0006,
+ "loss": 4.869322776794434,
+ "step": 3622
+ },
+ {
+ "epoch": 50.32139737991266,
+ "grad_norm": 0.018199188634753227,
+ "learning_rate": 0.0006,
+ "loss": 4.95680046081543,
+ "step": 3623
+ },
+ {
+ "epoch": 50.3353711790393,
+ "grad_norm": 0.016688082367181778,
+ "learning_rate": 0.0006,
+ "loss": 4.783535957336426,
+ "step": 3624
+ },
+ {
+ "epoch": 50.34934497816594,
+ "grad_norm": 0.017540352419018745,
+ "learning_rate": 0.0006,
+ "loss": 4.8236083984375,
+ "step": 3625
+ },
+ {
+ "epoch": 50.36331877729258,
+ "grad_norm": 0.018054187297821045,
+ "learning_rate": 0.0006,
+ "loss": 4.764655113220215,
+ "step": 3626
+ },
+ {
+ "epoch": 50.377292576419215,
+ "grad_norm": 0.01741131953895092,
+ "learning_rate": 0.0006,
+ "loss": 4.875613689422607,
+ "step": 3627
+ },
+ {
+ "epoch": 50.391266375545854,
+ "grad_norm": 0.01682269014418125,
+ "learning_rate": 0.0006,
+ "loss": 4.818509101867676,
+ "step": 3628
+ },
+ {
+ "epoch": 50.40524017467249,
+ "grad_norm": 0.016652341932058334,
+ "learning_rate": 0.0006,
+ "loss": 4.974020481109619,
+ "step": 3629
+ },
+ {
+ "epoch": 50.419213973799124,
+ "grad_norm": 0.02035282365977764,
+ "learning_rate": 0.0006,
+ "loss": 4.893037796020508,
+ "step": 3630
+ },
+ {
+ "epoch": 50.43318777292576,
+ "grad_norm": 0.019907061010599136,
+ "learning_rate": 0.0006,
+ "loss": 4.848119735717773,
+ "step": 3631
+ },
+ {
+ "epoch": 50.4471615720524,
+ "grad_norm": 0.023861493915319443,
+ "learning_rate": 0.0006,
+ "loss": 4.819353103637695,
+ "step": 3632
+ },
+ {
+ "epoch": 50.46113537117904,
+ "grad_norm": 0.025254026055336,
+ "learning_rate": 0.0006,
+ "loss": 4.769312858581543,
+ "step": 3633
+ },
+ {
+ "epoch": 50.47510917030568,
+ "grad_norm": 0.026154937222599983,
+ "learning_rate": 0.0006,
+ "loss": 4.7586774826049805,
+ "step": 3634
+ },
+ {
+ "epoch": 50.48908296943232,
+ "grad_norm": 0.026342090219259262,
+ "learning_rate": 0.0006,
+ "loss": 4.861218452453613,
+ "step": 3635
+ },
+ {
+ "epoch": 50.50305676855895,
+ "grad_norm": 0.028169559314846992,
+ "learning_rate": 0.0006,
+ "loss": 4.818380355834961,
+ "step": 3636
+ },
+ {
+ "epoch": 50.51703056768559,
+ "grad_norm": 0.02820264920592308,
+ "learning_rate": 0.0006,
+ "loss": 4.892871856689453,
+ "step": 3637
+ },
+ {
+ "epoch": 50.531004366812226,
+ "grad_norm": 0.03345080465078354,
+ "learning_rate": 0.0006,
+ "loss": 4.811084747314453,
+ "step": 3638
+ },
+ {
+ "epoch": 50.544978165938865,
+ "grad_norm": 0.04005736857652664,
+ "learning_rate": 0.0006,
+ "loss": 4.765410900115967,
+ "step": 3639
+ },
+ {
+ "epoch": 50.5589519650655,
+ "grad_norm": 0.04333870857954025,
+ "learning_rate": 0.0006,
+ "loss": 4.956520080566406,
+ "step": 3640
+ },
+ {
+ "epoch": 50.57292576419214,
+ "grad_norm": 0.04216952994465828,
+ "learning_rate": 0.0006,
+ "loss": 4.793441295623779,
+ "step": 3641
+ },
+ {
+ "epoch": 50.58689956331878,
+ "grad_norm": 0.02936951443552971,
+ "learning_rate": 0.0006,
+ "loss": 4.8447957038879395,
+ "step": 3642
+ },
+ {
+ "epoch": 50.60087336244541,
+ "grad_norm": 0.026937026530504227,
+ "learning_rate": 0.0006,
+ "loss": 4.962215423583984,
+ "step": 3643
+ },
+ {
+ "epoch": 50.61484716157205,
+ "grad_norm": 0.02628890983760357,
+ "learning_rate": 0.0006,
+ "loss": 4.883061408996582,
+ "step": 3644
+ },
+ {
+ "epoch": 50.62882096069869,
+ "grad_norm": 0.025984834879636765,
+ "learning_rate": 0.0006,
+ "loss": 4.768662452697754,
+ "step": 3645
+ },
+ {
+ "epoch": 50.64279475982533,
+ "grad_norm": 0.023998012766242027,
+ "learning_rate": 0.0006,
+ "loss": 4.851325035095215,
+ "step": 3646
+ },
+ {
+ "epoch": 50.65676855895197,
+ "grad_norm": 0.02143184095621109,
+ "learning_rate": 0.0006,
+ "loss": 4.8274688720703125,
+ "step": 3647
+ },
+ {
+ "epoch": 50.670742358078606,
+ "grad_norm": 0.02488900162279606,
+ "learning_rate": 0.0006,
+ "loss": 4.831140518188477,
+ "step": 3648
+ },
+ {
+ "epoch": 50.68471615720524,
+ "grad_norm": 0.02330533228814602,
+ "learning_rate": 0.0006,
+ "loss": 4.850462913513184,
+ "step": 3649
+ },
+ {
+ "epoch": 50.698689956331876,
+ "grad_norm": 0.02121943049132824,
+ "learning_rate": 0.0006,
+ "loss": 4.870253086090088,
+ "step": 3650
+ },
+ {
+ "epoch": 50.712663755458514,
+ "grad_norm": 0.02080935798585415,
+ "learning_rate": 0.0006,
+ "loss": 4.798770904541016,
+ "step": 3651
+ },
+ {
+ "epoch": 50.72663755458515,
+ "grad_norm": 0.019047029316425323,
+ "learning_rate": 0.0006,
+ "loss": 4.8427300453186035,
+ "step": 3652
+ },
+ {
+ "epoch": 50.74061135371179,
+ "grad_norm": 0.01824103854596615,
+ "learning_rate": 0.0006,
+ "loss": 4.820980548858643,
+ "step": 3653
+ },
+ {
+ "epoch": 50.75458515283843,
+ "grad_norm": 0.018129967153072357,
+ "learning_rate": 0.0006,
+ "loss": 4.946732997894287,
+ "step": 3654
+ },
+ {
+ "epoch": 50.76855895196506,
+ "grad_norm": 0.01784815452992916,
+ "learning_rate": 0.0006,
+ "loss": 4.829516410827637,
+ "step": 3655
+ },
+ {
+ "epoch": 50.7825327510917,
+ "grad_norm": 0.020021555945277214,
+ "learning_rate": 0.0006,
+ "loss": 4.780428409576416,
+ "step": 3656
+ },
+ {
+ "epoch": 50.79650655021834,
+ "grad_norm": 0.01863253302872181,
+ "learning_rate": 0.0006,
+ "loss": 4.791567802429199,
+ "step": 3657
+ },
+ {
+ "epoch": 50.81048034934498,
+ "grad_norm": 0.019151683896780014,
+ "learning_rate": 0.0006,
+ "loss": 4.824892997741699,
+ "step": 3658
+ },
+ {
+ "epoch": 50.82445414847162,
+ "grad_norm": 0.01891648955643177,
+ "learning_rate": 0.0006,
+ "loss": 4.789605140686035,
+ "step": 3659
+ },
+ {
+ "epoch": 50.838427947598255,
+ "grad_norm": 0.016019510105252266,
+ "learning_rate": 0.0006,
+ "loss": 4.779555320739746,
+ "step": 3660
+ },
+ {
+ "epoch": 50.852401746724894,
+ "grad_norm": 0.016096491366624832,
+ "learning_rate": 0.0006,
+ "loss": 4.753842353820801,
+ "step": 3661
+ },
+ {
+ "epoch": 50.866375545851525,
+ "grad_norm": 0.01693476364016533,
+ "learning_rate": 0.0006,
+ "loss": 4.768477439880371,
+ "step": 3662
+ },
+ {
+ "epoch": 50.880349344978164,
+ "grad_norm": 0.01749575324356556,
+ "learning_rate": 0.0006,
+ "loss": 4.779143333435059,
+ "step": 3663
+ },
+ {
+ "epoch": 50.8943231441048,
+ "grad_norm": 0.016851043328642845,
+ "learning_rate": 0.0006,
+ "loss": 4.780387878417969,
+ "step": 3664
+ },
+ {
+ "epoch": 50.90829694323144,
+ "grad_norm": 0.019045202061533928,
+ "learning_rate": 0.0006,
+ "loss": 4.769913673400879,
+ "step": 3665
+ },
+ {
+ "epoch": 50.92227074235808,
+ "grad_norm": 0.01954123005270958,
+ "learning_rate": 0.0006,
+ "loss": 4.769009590148926,
+ "step": 3666
+ },
+ {
+ "epoch": 50.93624454148472,
+ "grad_norm": 0.022527096793055534,
+ "learning_rate": 0.0006,
+ "loss": 4.789215087890625,
+ "step": 3667
+ },
+ {
+ "epoch": 50.95021834061135,
+ "grad_norm": 0.02871634066104889,
+ "learning_rate": 0.0006,
+ "loss": 4.749727725982666,
+ "step": 3668
+ },
+ {
+ "epoch": 50.96419213973799,
+ "grad_norm": 0.04045053571462631,
+ "learning_rate": 0.0006,
+ "loss": 4.909038543701172,
+ "step": 3669
+ },
+ {
+ "epoch": 50.97816593886463,
+ "grad_norm": 0.04156879335641861,
+ "learning_rate": 0.0006,
+ "loss": 4.876831531524658,
+ "step": 3670
+ },
+ {
+ "epoch": 50.992139737991266,
+ "grad_norm": 0.02621397376060486,
+ "learning_rate": 0.0006,
+ "loss": 4.797173500061035,
+ "step": 3671
+ },
+ {
+ "epoch": 51.0,
+ "grad_norm": 0.03389597684144974,
+ "learning_rate": 0.0006,
+ "loss": 4.8428850173950195,
+ "step": 3672
+ },
+ {
+ "epoch": 51.0,
+ "eval_loss": 4.989261627197266,
+ "eval_runtime": 56.358,
+ "eval_samples_per_second": 43.33,
+ "eval_steps_per_second": 1.366,
+ "step": 3672
+ },
+ {
+ "epoch": 51.01397379912664,
+ "grad_norm": 0.03216972202062607,
+ "learning_rate": 0.0006,
+ "loss": 4.755102634429932,
+ "step": 3673
+ },
+ {
+ "epoch": 51.02794759825328,
+ "grad_norm": 0.03190374746918678,
+ "learning_rate": 0.0006,
+ "loss": 4.720424652099609,
+ "step": 3674
+ },
+ {
+ "epoch": 51.041921397379916,
+ "grad_norm": 0.031727299094200134,
+ "learning_rate": 0.0006,
+ "loss": 4.803277969360352,
+ "step": 3675
+ },
+ {
+ "epoch": 51.05589519650655,
+ "grad_norm": 0.02473079040646553,
+ "learning_rate": 0.0006,
+ "loss": 4.761087417602539,
+ "step": 3676
+ },
+ {
+ "epoch": 51.069868995633186,
+ "grad_norm": 0.020360209047794342,
+ "learning_rate": 0.0006,
+ "loss": 4.753884792327881,
+ "step": 3677
+ },
+ {
+ "epoch": 51.083842794759825,
+ "grad_norm": 0.02078923024237156,
+ "learning_rate": 0.0006,
+ "loss": 4.647130489349365,
+ "step": 3678
+ },
+ {
+ "epoch": 51.09781659388646,
+ "grad_norm": 0.019387200474739075,
+ "learning_rate": 0.0006,
+ "loss": 4.74135160446167,
+ "step": 3679
+ },
+ {
+ "epoch": 51.1117903930131,
+ "grad_norm": 0.01874949410557747,
+ "learning_rate": 0.0006,
+ "loss": 4.795329570770264,
+ "step": 3680
+ },
+ {
+ "epoch": 51.12576419213974,
+ "grad_norm": 0.019331837072968483,
+ "learning_rate": 0.0006,
+ "loss": 4.897967338562012,
+ "step": 3681
+ },
+ {
+ "epoch": 51.13973799126637,
+ "grad_norm": 0.017066050320863724,
+ "learning_rate": 0.0006,
+ "loss": 4.825351715087891,
+ "step": 3682
+ },
+ {
+ "epoch": 51.15371179039301,
+ "grad_norm": 0.01757810264825821,
+ "learning_rate": 0.0006,
+ "loss": 4.719335079193115,
+ "step": 3683
+ },
+ {
+ "epoch": 51.16768558951965,
+ "grad_norm": 0.018174266442656517,
+ "learning_rate": 0.0006,
+ "loss": 4.732680320739746,
+ "step": 3684
+ },
+ {
+ "epoch": 51.18165938864629,
+ "grad_norm": 0.018497727811336517,
+ "learning_rate": 0.0006,
+ "loss": 4.731050491333008,
+ "step": 3685
+ },
+ {
+ "epoch": 51.19563318777293,
+ "grad_norm": 0.0186338871717453,
+ "learning_rate": 0.0006,
+ "loss": 4.748035430908203,
+ "step": 3686
+ },
+ {
+ "epoch": 51.209606986899566,
+ "grad_norm": 0.018531711772084236,
+ "learning_rate": 0.0006,
+ "loss": 4.9007463455200195,
+ "step": 3687
+ },
+ {
+ "epoch": 51.223580786026204,
+ "grad_norm": 0.018221354112029076,
+ "learning_rate": 0.0006,
+ "loss": 4.84077262878418,
+ "step": 3688
+ },
+ {
+ "epoch": 51.237554585152836,
+ "grad_norm": 0.0203656405210495,
+ "learning_rate": 0.0006,
+ "loss": 4.73009729385376,
+ "step": 3689
+ },
+ {
+ "epoch": 51.251528384279474,
+ "grad_norm": 0.01964893378317356,
+ "learning_rate": 0.0006,
+ "loss": 4.772583961486816,
+ "step": 3690
+ },
+ {
+ "epoch": 51.26550218340611,
+ "grad_norm": 0.019606510177254677,
+ "learning_rate": 0.0006,
+ "loss": 4.599812030792236,
+ "step": 3691
+ },
+ {
+ "epoch": 51.27947598253275,
+ "grad_norm": 0.020501255989074707,
+ "learning_rate": 0.0006,
+ "loss": 4.809898853302002,
+ "step": 3692
+ },
+ {
+ "epoch": 51.29344978165939,
+ "grad_norm": 0.01870838552713394,
+ "learning_rate": 0.0006,
+ "loss": 4.741898059844971,
+ "step": 3693
+ },
+ {
+ "epoch": 51.30742358078603,
+ "grad_norm": 0.01802559196949005,
+ "learning_rate": 0.0006,
+ "loss": 4.746562480926514,
+ "step": 3694
+ },
+ {
+ "epoch": 51.32139737991266,
+ "grad_norm": 0.02053048275411129,
+ "learning_rate": 0.0006,
+ "loss": 4.69932746887207,
+ "step": 3695
+ },
+ {
+ "epoch": 51.3353711790393,
+ "grad_norm": 0.02661825157701969,
+ "learning_rate": 0.0006,
+ "loss": 4.68748664855957,
+ "step": 3696
+ },
+ {
+ "epoch": 51.34934497816594,
+ "grad_norm": 0.028269147500395775,
+ "learning_rate": 0.0006,
+ "loss": 4.731565475463867,
+ "step": 3697
+ },
+ {
+ "epoch": 51.36331877729258,
+ "grad_norm": 0.02650964818894863,
+ "learning_rate": 0.0006,
+ "loss": 4.733857154846191,
+ "step": 3698
+ },
+ {
+ "epoch": 51.377292576419215,
+ "grad_norm": 0.023696066811680794,
+ "learning_rate": 0.0006,
+ "loss": 4.812362194061279,
+ "step": 3699
+ },
+ {
+ "epoch": 51.391266375545854,
+ "grad_norm": 0.02365880273282528,
+ "learning_rate": 0.0006,
+ "loss": 4.733426094055176,
+ "step": 3700
+ },
+ {
+ "epoch": 51.40524017467249,
+ "grad_norm": 0.026249447837471962,
+ "learning_rate": 0.0006,
+ "loss": 4.715622901916504,
+ "step": 3701
+ },
+ {
+ "epoch": 51.419213973799124,
+ "grad_norm": 0.03301899880170822,
+ "learning_rate": 0.0006,
+ "loss": 4.7223663330078125,
+ "step": 3702
+ },
+ {
+ "epoch": 51.43318777292576,
+ "grad_norm": 0.044686660170555115,
+ "learning_rate": 0.0006,
+ "loss": 4.8121161460876465,
+ "step": 3703
+ },
+ {
+ "epoch": 51.4471615720524,
+ "grad_norm": 0.04733522981405258,
+ "learning_rate": 0.0006,
+ "loss": 4.841394424438477,
+ "step": 3704
+ },
+ {
+ "epoch": 51.46113537117904,
+ "grad_norm": 0.0395209975540638,
+ "learning_rate": 0.0006,
+ "loss": 4.773837089538574,
+ "step": 3705
+ },
+ {
+ "epoch": 51.47510917030568,
+ "grad_norm": 0.032790958881378174,
+ "learning_rate": 0.0006,
+ "loss": 4.833066463470459,
+ "step": 3706
+ },
+ {
+ "epoch": 51.48908296943232,
+ "grad_norm": 0.030578427016735077,
+ "learning_rate": 0.0006,
+ "loss": 4.685190200805664,
+ "step": 3707
+ },
+ {
+ "epoch": 51.50305676855895,
+ "grad_norm": 0.031619083136320114,
+ "learning_rate": 0.0006,
+ "loss": 4.740378379821777,
+ "step": 3708
+ },
+ {
+ "epoch": 51.51703056768559,
+ "grad_norm": 0.02385825850069523,
+ "learning_rate": 0.0006,
+ "loss": 4.812911510467529,
+ "step": 3709
+ },
+ {
+ "epoch": 51.531004366812226,
+ "grad_norm": 0.023523425683379173,
+ "learning_rate": 0.0006,
+ "loss": 4.669580459594727,
+ "step": 3710
+ },
+ {
+ "epoch": 51.544978165938865,
+ "grad_norm": 0.021316280588507652,
+ "learning_rate": 0.0006,
+ "loss": 4.76888370513916,
+ "step": 3711
+ },
+ {
+ "epoch": 51.5589519650655,
+ "grad_norm": 0.024663880467414856,
+ "learning_rate": 0.0006,
+ "loss": 4.702592372894287,
+ "step": 3712
+ },
+ {
+ "epoch": 51.57292576419214,
+ "grad_norm": 0.024173280224204063,
+ "learning_rate": 0.0006,
+ "loss": 4.710604667663574,
+ "step": 3713
+ },
+ {
+ "epoch": 51.58689956331878,
+ "grad_norm": 0.02754114381968975,
+ "learning_rate": 0.0006,
+ "loss": 4.746356964111328,
+ "step": 3714
+ },
+ {
+ "epoch": 51.60087336244541,
+ "grad_norm": 0.027903622016310692,
+ "learning_rate": 0.0006,
+ "loss": 4.7546491622924805,
+ "step": 3715
+ },
+ {
+ "epoch": 51.61484716157205,
+ "grad_norm": 0.024871617555618286,
+ "learning_rate": 0.0006,
+ "loss": 4.605746269226074,
+ "step": 3716
+ },
+ {
+ "epoch": 51.62882096069869,
+ "grad_norm": 0.021239031106233597,
+ "learning_rate": 0.0006,
+ "loss": 4.6942458152771,
+ "step": 3717
+ },
+ {
+ "epoch": 51.64279475982533,
+ "grad_norm": 0.01925622671842575,
+ "learning_rate": 0.0006,
+ "loss": 4.593558311462402,
+ "step": 3718
+ },
+ {
+ "epoch": 51.65676855895197,
+ "grad_norm": 0.01956283673644066,
+ "learning_rate": 0.0006,
+ "loss": 4.631172180175781,
+ "step": 3719
+ },
+ {
+ "epoch": 51.670742358078606,
+ "grad_norm": 0.02269871160387993,
+ "learning_rate": 0.0006,
+ "loss": 4.729236602783203,
+ "step": 3720
+ },
+ {
+ "epoch": 51.68471615720524,
+ "grad_norm": 0.02253916673362255,
+ "learning_rate": 0.0006,
+ "loss": 4.7191033363342285,
+ "step": 3721
+ },
+ {
+ "epoch": 51.698689956331876,
+ "grad_norm": 0.026231050491333008,
+ "learning_rate": 0.0006,
+ "loss": 4.811850547790527,
+ "step": 3722
+ },
+ {
+ "epoch": 51.712663755458514,
+ "grad_norm": 0.02558821812272072,
+ "learning_rate": 0.0006,
+ "loss": 4.681466102600098,
+ "step": 3723
+ },
+ {
+ "epoch": 51.72663755458515,
+ "grad_norm": 0.02124209702014923,
+ "learning_rate": 0.0006,
+ "loss": 4.693664073944092,
+ "step": 3724
+ },
+ {
+ "epoch": 51.74061135371179,
+ "grad_norm": 0.016638319939374924,
+ "learning_rate": 0.0006,
+ "loss": 4.771440029144287,
+ "step": 3725
+ },
+ {
+ "epoch": 51.75458515283843,
+ "grad_norm": 0.018048347905278206,
+ "learning_rate": 0.0006,
+ "loss": 4.643252372741699,
+ "step": 3726
+ },
+ {
+ "epoch": 51.76855895196506,
+ "grad_norm": 0.019904717803001404,
+ "learning_rate": 0.0006,
+ "loss": 4.66530704498291,
+ "step": 3727
+ },
+ {
+ "epoch": 51.7825327510917,
+ "grad_norm": 0.0194614976644516,
+ "learning_rate": 0.0006,
+ "loss": 4.711943626403809,
+ "step": 3728
+ },
+ {
+ "epoch": 51.79650655021834,
+ "grad_norm": 0.019175555557012558,
+ "learning_rate": 0.0006,
+ "loss": 4.721931457519531,
+ "step": 3729
+ },
+ {
+ "epoch": 51.81048034934498,
+ "grad_norm": 0.02093447372317314,
+ "learning_rate": 0.0006,
+ "loss": 4.633059024810791,
+ "step": 3730
+ },
+ {
+ "epoch": 51.82445414847162,
+ "grad_norm": 0.01930283196270466,
+ "learning_rate": 0.0006,
+ "loss": 4.659373760223389,
+ "step": 3731
+ },
+ {
+ "epoch": 51.838427947598255,
+ "grad_norm": 0.015350676141679287,
+ "learning_rate": 0.0006,
+ "loss": 4.763708114624023,
+ "step": 3732
+ },
+ {
+ "epoch": 51.852401746724894,
+ "grad_norm": 0.014139448292553425,
+ "learning_rate": 0.0006,
+ "loss": 4.753891468048096,
+ "step": 3733
+ },
+ {
+ "epoch": 51.866375545851525,
+ "grad_norm": 0.016215872019529343,
+ "learning_rate": 0.0006,
+ "loss": 4.658967018127441,
+ "step": 3734
+ },
+ {
+ "epoch": 51.880349344978164,
+ "grad_norm": 0.017328619956970215,
+ "learning_rate": 0.0006,
+ "loss": 4.696579933166504,
+ "step": 3735
+ },
+ {
+ "epoch": 51.8943231441048,
+ "grad_norm": 0.018103396520018578,
+ "learning_rate": 0.0006,
+ "loss": 4.627464771270752,
+ "step": 3736
+ },
+ {
+ "epoch": 51.90829694323144,
+ "grad_norm": 0.01715235412120819,
+ "learning_rate": 0.0006,
+ "loss": 4.7368011474609375,
+ "step": 3737
+ },
+ {
+ "epoch": 51.92227074235808,
+ "grad_norm": 0.014790094457566738,
+ "learning_rate": 0.0006,
+ "loss": 4.629441738128662,
+ "step": 3738
+ },
+ {
+ "epoch": 51.93624454148472,
+ "grad_norm": 0.017114296555519104,
+ "learning_rate": 0.0006,
+ "loss": 4.761256694793701,
+ "step": 3739
+ },
+ {
+ "epoch": 51.95021834061135,
+ "grad_norm": 0.019009845331311226,
+ "learning_rate": 0.0006,
+ "loss": 4.789155006408691,
+ "step": 3740
+ },
+ {
+ "epoch": 51.96419213973799,
+ "grad_norm": 0.019696692004799843,
+ "learning_rate": 0.0006,
+ "loss": 4.721269607543945,
+ "step": 3741
+ },
+ {
+ "epoch": 51.97816593886463,
+ "grad_norm": 0.020542822778224945,
+ "learning_rate": 0.0006,
+ "loss": 4.687038898468018,
+ "step": 3742
+ },
+ {
+ "epoch": 51.992139737991266,
+ "grad_norm": 0.021267160773277283,
+ "learning_rate": 0.0006,
+ "loss": 4.747860908508301,
+ "step": 3743
+ },
+ {
+ "epoch": 52.0,
+ "grad_norm": 0.025049567222595215,
+ "learning_rate": 0.0006,
+ "loss": 4.8473100662231445,
+ "step": 3744
+ },
+ {
+ "epoch": 52.0,
+ "eval_loss": 4.943043231964111,
+ "eval_runtime": 56.4929,
+ "eval_samples_per_second": 43.227,
+ "eval_steps_per_second": 1.363,
+ "step": 3744
+ },
+ {
+ "epoch": 52.01397379912664,
+ "grad_norm": 0.03054986707866192,
+ "learning_rate": 0.0006,
+ "loss": 4.856931686401367,
+ "step": 3745
+ },
+ {
+ "epoch": 52.02794759825328,
+ "grad_norm": 0.03838375583291054,
+ "learning_rate": 0.0006,
+ "loss": 4.699807167053223,
+ "step": 3746
+ },
+ {
+ "epoch": 52.041921397379916,
+ "grad_norm": 0.03383488208055496,
+ "learning_rate": 0.0006,
+ "loss": 4.6626787185668945,
+ "step": 3747
+ },
+ {
+ "epoch": 52.05589519650655,
+ "grad_norm": 0.0201561376452446,
+ "learning_rate": 0.0006,
+ "loss": 4.621831893920898,
+ "step": 3748
+ },
+ {
+ "epoch": 52.069868995633186,
+ "grad_norm": 0.028749438002705574,
+ "learning_rate": 0.0006,
+ "loss": 4.682540416717529,
+ "step": 3749
+ },
+ {
+ "epoch": 52.083842794759825,
+ "grad_norm": 0.02681174874305725,
+ "learning_rate": 0.0006,
+ "loss": 4.629997253417969,
+ "step": 3750
+ },
+ {
+ "epoch": 52.09781659388646,
+ "grad_norm": 0.024063318967819214,
+ "learning_rate": 0.0006,
+ "loss": 4.573021411895752,
+ "step": 3751
+ },
+ {
+ "epoch": 52.1117903930131,
+ "grad_norm": 0.02324330434203148,
+ "learning_rate": 0.0006,
+ "loss": 4.739465236663818,
+ "step": 3752
+ },
+ {
+ "epoch": 52.12576419213974,
+ "grad_norm": 0.0225426834076643,
+ "learning_rate": 0.0006,
+ "loss": 4.7335333824157715,
+ "step": 3753
+ },
+ {
+ "epoch": 52.13973799126637,
+ "grad_norm": 0.019593428820371628,
+ "learning_rate": 0.0006,
+ "loss": 4.704958438873291,
+ "step": 3754
+ },
+ {
+ "epoch": 52.15371179039301,
+ "grad_norm": 0.020777082070708275,
+ "learning_rate": 0.0006,
+ "loss": 4.638696670532227,
+ "step": 3755
+ },
+ {
+ "epoch": 52.16768558951965,
+ "grad_norm": 0.021525604650378227,
+ "learning_rate": 0.0006,
+ "loss": 4.643706798553467,
+ "step": 3756
+ },
+ {
+ "epoch": 52.18165938864629,
+ "grad_norm": 0.018943654373288155,
+ "learning_rate": 0.0006,
+ "loss": 4.5173211097717285,
+ "step": 3757
+ },
+ {
+ "epoch": 52.19563318777293,
+ "grad_norm": 0.01856580562889576,
+ "learning_rate": 0.0006,
+ "loss": 4.678788661956787,
+ "step": 3758
+ },
+ {
+ "epoch": 52.209606986899566,
+ "grad_norm": 0.019417259842157364,
+ "learning_rate": 0.0006,
+ "loss": 4.587244987487793,
+ "step": 3759
+ },
+ {
+ "epoch": 52.223580786026204,
+ "grad_norm": 0.01911911740899086,
+ "learning_rate": 0.0006,
+ "loss": 4.688503265380859,
+ "step": 3760
+ },
+ {
+ "epoch": 52.237554585152836,
+ "grad_norm": 0.021017879247665405,
+ "learning_rate": 0.0006,
+ "loss": 4.65829610824585,
+ "step": 3761
+ },
+ {
+ "epoch": 52.251528384279474,
+ "grad_norm": 0.02048875205218792,
+ "learning_rate": 0.0006,
+ "loss": 4.624055862426758,
+ "step": 3762
+ },
+ {
+ "epoch": 52.26550218340611,
+ "grad_norm": 0.01932772807776928,
+ "learning_rate": 0.0006,
+ "loss": 4.556562900543213,
+ "step": 3763
+ },
+ {
+ "epoch": 52.27947598253275,
+ "grad_norm": 0.01871614344418049,
+ "learning_rate": 0.0006,
+ "loss": 4.581201076507568,
+ "step": 3764
+ },
+ {
+ "epoch": 52.29344978165939,
+ "grad_norm": 0.017253106459975243,
+ "learning_rate": 0.0006,
+ "loss": 4.605219841003418,
+ "step": 3765
+ },
+ {
+ "epoch": 52.30742358078603,
+ "grad_norm": 0.01831238530576229,
+ "learning_rate": 0.0006,
+ "loss": 4.759900093078613,
+ "step": 3766
+ },
+ {
+ "epoch": 52.32139737991266,
+ "grad_norm": 0.017946207895874977,
+ "learning_rate": 0.0006,
+ "loss": 4.5598344802856445,
+ "step": 3767
+ },
+ {
+ "epoch": 52.3353711790393,
+ "grad_norm": 0.01680106669664383,
+ "learning_rate": 0.0006,
+ "loss": 4.605034828186035,
+ "step": 3768
+ },
+ {
+ "epoch": 52.34934497816594,
+ "grad_norm": 0.021482540294528008,
+ "learning_rate": 0.0006,
+ "loss": 4.620411396026611,
+ "step": 3769
+ },
+ {
+ "epoch": 52.36331877729258,
+ "grad_norm": 0.02210252732038498,
+ "learning_rate": 0.0006,
+ "loss": 4.667577743530273,
+ "step": 3770
+ },
+ {
+ "epoch": 52.377292576419215,
+ "grad_norm": 0.02237485907971859,
+ "learning_rate": 0.0006,
+ "loss": 4.628812789916992,
+ "step": 3771
+ },
+ {
+ "epoch": 52.391266375545854,
+ "grad_norm": 0.02196541801095009,
+ "learning_rate": 0.0006,
+ "loss": 4.572988510131836,
+ "step": 3772
+ },
+ {
+ "epoch": 52.40524017467249,
+ "grad_norm": 0.01930229924619198,
+ "learning_rate": 0.0006,
+ "loss": 4.588339805603027,
+ "step": 3773
+ },
+ {
+ "epoch": 52.419213973799124,
+ "grad_norm": 0.02222434990108013,
+ "learning_rate": 0.0006,
+ "loss": 4.766482353210449,
+ "step": 3774
+ },
+ {
+ "epoch": 52.43318777292576,
+ "grad_norm": 0.028806306421756744,
+ "learning_rate": 0.0006,
+ "loss": 4.617892742156982,
+ "step": 3775
+ },
+ {
+ "epoch": 52.4471615720524,
+ "grad_norm": 0.03306042402982712,
+ "learning_rate": 0.0006,
+ "loss": 4.665334701538086,
+ "step": 3776
+ },
+ {
+ "epoch": 52.46113537117904,
+ "grad_norm": 0.0358956903219223,
+ "learning_rate": 0.0006,
+ "loss": 4.566668510437012,
+ "step": 3777
+ },
+ {
+ "epoch": 52.47510917030568,
+ "grad_norm": 0.030406497418880463,
+ "learning_rate": 0.0006,
+ "loss": 4.513612270355225,
+ "step": 3778
+ },
+ {
+ "epoch": 52.48908296943232,
+ "grad_norm": 0.030295712873339653,
+ "learning_rate": 0.0006,
+ "loss": 4.640189170837402,
+ "step": 3779
+ },
+ {
+ "epoch": 52.50305676855895,
+ "grad_norm": 0.024981003254652023,
+ "learning_rate": 0.0006,
+ "loss": 4.57558012008667,
+ "step": 3780
+ },
+ {
+ "epoch": 52.51703056768559,
+ "grad_norm": 0.024990713223814964,
+ "learning_rate": 0.0006,
+ "loss": 4.708589553833008,
+ "step": 3781
+ },
+ {
+ "epoch": 52.531004366812226,
+ "grad_norm": 0.025034775957465172,
+ "learning_rate": 0.0006,
+ "loss": 4.70468282699585,
+ "step": 3782
+ },
+ {
+ "epoch": 52.544978165938865,
+ "grad_norm": 0.028243759647011757,
+ "learning_rate": 0.0006,
+ "loss": 4.712060451507568,
+ "step": 3783
+ },
+ {
+ "epoch": 52.5589519650655,
+ "grad_norm": 0.026984544470906258,
+ "learning_rate": 0.0006,
+ "loss": 4.505004405975342,
+ "step": 3784
+ },
+ {
+ "epoch": 52.57292576419214,
+ "grad_norm": 0.050843529403209686,
+ "learning_rate": 0.0006,
+ "loss": 4.593271255493164,
+ "step": 3785
+ },
+ {
+ "epoch": 52.58689956331878,
+ "grad_norm": 0.25422802567481995,
+ "learning_rate": 0.0006,
+ "loss": 4.7178802490234375,
+ "step": 3786
+ },
+ {
+ "epoch": 52.60087336244541,
+ "grad_norm": 0.6078406572341919,
+ "learning_rate": 0.0006,
+ "loss": 6.144372940063477,
+ "step": 3787
+ },
+ {
+ "epoch": 52.61484716157205,
+ "grad_norm": 0.25933101773262024,
+ "learning_rate": 0.0006,
+ "loss": 5.971083641052246,
+ "step": 3788
+ },
+ {
+ "epoch": 52.62882096069869,
+ "grad_norm": 0.294657826423645,
+ "learning_rate": 0.0006,
+ "loss": 6.204090118408203,
+ "step": 3789
+ },
+ {
+ "epoch": 52.64279475982533,
+ "grad_norm": 0.23641522228717804,
+ "learning_rate": 0.0006,
+ "loss": 6.869574546813965,
+ "step": 3790
+ },
+ {
+ "epoch": 52.65676855895197,
+ "grad_norm": 0.23717311024665833,
+ "learning_rate": 0.0006,
+ "loss": 6.739034175872803,
+ "step": 3791
+ },
+ {
+ "epoch": 52.670742358078606,
+ "grad_norm": 0.10300362855195999,
+ "learning_rate": 0.0006,
+ "loss": 6.49191951751709,
+ "step": 3792
+ },
+ {
+ "epoch": 52.68471615720524,
+ "grad_norm": 0.1484968513250351,
+ "learning_rate": 0.0006,
+ "loss": 6.6021270751953125,
+ "step": 3793
+ },
+ {
+ "epoch": 52.698689956331876,
+ "grad_norm": 0.08052115142345428,
+ "learning_rate": 0.0006,
+ "loss": 6.432796478271484,
+ "step": 3794
+ },
+ {
+ "epoch": 52.712663755458514,
+ "grad_norm": 0.06869253516197205,
+ "learning_rate": 0.0006,
+ "loss": 6.212784767150879,
+ "step": 3795
+ },
+ {
+ "epoch": 52.72663755458515,
+ "grad_norm": 0.05589582398533821,
+ "learning_rate": 0.0006,
+ "loss": 6.168288707733154,
+ "step": 3796
+ },
+ {
+ "epoch": 52.74061135371179,
+ "grad_norm": 0.04840940609574318,
+ "learning_rate": 0.0006,
+ "loss": 6.163174629211426,
+ "step": 3797
+ },
+ {
+ "epoch": 52.75458515283843,
+ "grad_norm": 0.04578601196408272,
+ "learning_rate": 0.0006,
+ "loss": 6.038390636444092,
+ "step": 3798
+ },
+ {
+ "epoch": 52.76855895196506,
+ "grad_norm": 0.04102053865790367,
+ "learning_rate": 0.0006,
+ "loss": 5.929539680480957,
+ "step": 3799
+ },
+ {
+ "epoch": 52.7825327510917,
+ "grad_norm": 0.038398485630750656,
+ "learning_rate": 0.0006,
+ "loss": 6.007378101348877,
+ "step": 3800
+ },
+ {
+ "epoch": 52.79650655021834,
+ "grad_norm": 0.03466642275452614,
+ "learning_rate": 0.0006,
+ "loss": 5.767291069030762,
+ "step": 3801
+ },
+ {
+ "epoch": 52.81048034934498,
+ "grad_norm": 0.03328905999660492,
+ "learning_rate": 0.0006,
+ "loss": 5.785032272338867,
+ "step": 3802
+ },
+ {
+ "epoch": 52.82445414847162,
+ "grad_norm": 0.03186681121587753,
+ "learning_rate": 0.0006,
+ "loss": 5.6942057609558105,
+ "step": 3803
+ },
+ {
+ "epoch": 52.838427947598255,
+ "grad_norm": 0.030553551390767097,
+ "learning_rate": 0.0006,
+ "loss": 5.680774688720703,
+ "step": 3804
+ },
+ {
+ "epoch": 52.852401746724894,
+ "grad_norm": 0.02578839473426342,
+ "learning_rate": 0.0006,
+ "loss": 5.703461647033691,
+ "step": 3805
+ },
+ {
+ "epoch": 52.866375545851525,
+ "grad_norm": 0.026261312887072563,
+ "learning_rate": 0.0006,
+ "loss": 5.642549514770508,
+ "step": 3806
+ },
+ {
+ "epoch": 52.880349344978164,
+ "grad_norm": 0.02782430872321129,
+ "learning_rate": 0.0006,
+ "loss": 5.563597679138184,
+ "step": 3807
+ },
+ {
+ "epoch": 52.8943231441048,
+ "grad_norm": 0.024467799812555313,
+ "learning_rate": 0.0006,
+ "loss": 5.583395957946777,
+ "step": 3808
+ },
+ {
+ "epoch": 52.90829694323144,
+ "grad_norm": 0.022602897137403488,
+ "learning_rate": 0.0006,
+ "loss": 5.50742769241333,
+ "step": 3809
+ },
+ {
+ "epoch": 52.92227074235808,
+ "grad_norm": 0.024292029440402985,
+ "learning_rate": 0.0006,
+ "loss": 5.367804527282715,
+ "step": 3810
+ },
+ {
+ "epoch": 52.93624454148472,
+ "grad_norm": 0.021062295883893967,
+ "learning_rate": 0.0006,
+ "loss": 5.423705101013184,
+ "step": 3811
+ },
+ {
+ "epoch": 52.95021834061135,
+ "grad_norm": 0.022226469591259956,
+ "learning_rate": 0.0006,
+ "loss": 5.344039440155029,
+ "step": 3812
+ },
+ {
+ "epoch": 52.96419213973799,
+ "grad_norm": 0.023691752925515175,
+ "learning_rate": 0.0006,
+ "loss": 5.2581892013549805,
+ "step": 3813
+ },
+ {
+ "epoch": 52.97816593886463,
+ "grad_norm": 0.02181999757885933,
+ "learning_rate": 0.0006,
+ "loss": 5.2710113525390625,
+ "step": 3814
+ },
+ {
+ "epoch": 52.992139737991266,
+ "grad_norm": 0.025295857340097427,
+ "learning_rate": 0.0006,
+ "loss": 5.278355598449707,
+ "step": 3815
+ },
+ {
+ "epoch": 53.0,
+ "grad_norm": 0.02216828241944313,
+ "learning_rate": 0.0006,
+ "loss": 5.127945899963379,
+ "step": 3816
+ },
+ {
+ "epoch": 53.0,
+ "eval_loss": 5.3411736488342285,
+ "eval_runtime": 56.2837,
+ "eval_samples_per_second": 43.387,
+ "eval_steps_per_second": 1.368,
+ "step": 3816
+ },
+ {
+ "epoch": 53.01397379912664,
+ "grad_norm": 0.022496026009321213,
+ "learning_rate": 0.0006,
+ "loss": 5.19655704498291,
+ "step": 3817
+ },
+ {
+ "epoch": 53.02794759825328,
+ "grad_norm": 0.03201168775558472,
+ "learning_rate": 0.0006,
+ "loss": 5.094621181488037,
+ "step": 3818
+ },
+ {
+ "epoch": 53.041921397379916,
+ "grad_norm": 0.02683178335428238,
+ "learning_rate": 0.0006,
+ "loss": 5.19899845123291,
+ "step": 3819
+ },
+ {
+ "epoch": 53.05589519650655,
+ "grad_norm": 0.022659169510006905,
+ "learning_rate": 0.0006,
+ "loss": 5.184103012084961,
+ "step": 3820
+ },
+ {
+ "epoch": 53.069868995633186,
+ "grad_norm": 0.024784443899989128,
+ "learning_rate": 0.0006,
+ "loss": 5.124286651611328,
+ "step": 3821
+ },
+ {
+ "epoch": 53.083842794759825,
+ "grad_norm": 0.028710370883345604,
+ "learning_rate": 0.0006,
+ "loss": 5.130426406860352,
+ "step": 3822
+ },
+ {
+ "epoch": 53.09781659388646,
+ "grad_norm": 0.02366088517010212,
+ "learning_rate": 0.0006,
+ "loss": 5.1453399658203125,
+ "step": 3823
+ },
+ {
+ "epoch": 53.1117903930131,
+ "grad_norm": 0.01813938096165657,
+ "learning_rate": 0.0006,
+ "loss": 5.05271053314209,
+ "step": 3824
+ },
+ {
+ "epoch": 53.12576419213974,
+ "grad_norm": 0.02051435597240925,
+ "learning_rate": 0.0006,
+ "loss": 5.023038864135742,
+ "step": 3825
+ },
+ {
+ "epoch": 53.13973799126637,
+ "grad_norm": 0.02185754105448723,
+ "learning_rate": 0.0006,
+ "loss": 4.8950018882751465,
+ "step": 3826
+ },
+ {
+ "epoch": 53.15371179039301,
+ "grad_norm": 0.017223743721842766,
+ "learning_rate": 0.0006,
+ "loss": 5.078621864318848,
+ "step": 3827
+ },
+ {
+ "epoch": 53.16768558951965,
+ "grad_norm": 0.01832217164337635,
+ "learning_rate": 0.0006,
+ "loss": 4.921341419219971,
+ "step": 3828
+ },
+ {
+ "epoch": 53.18165938864629,
+ "grad_norm": 0.02114005573093891,
+ "learning_rate": 0.0006,
+ "loss": 4.997865676879883,
+ "step": 3829
+ },
+ {
+ "epoch": 53.19563318777293,
+ "grad_norm": 0.021426431834697723,
+ "learning_rate": 0.0006,
+ "loss": 4.877785682678223,
+ "step": 3830
+ },
+ {
+ "epoch": 53.209606986899566,
+ "grad_norm": 0.021327560767531395,
+ "learning_rate": 0.0006,
+ "loss": 4.847512245178223,
+ "step": 3831
+ },
+ {
+ "epoch": 53.223580786026204,
+ "grad_norm": 0.027825530618429184,
+ "learning_rate": 0.0006,
+ "loss": 5.004823207855225,
+ "step": 3832
+ },
+ {
+ "epoch": 53.237554585152836,
+ "grad_norm": 0.03756481409072876,
+ "learning_rate": 0.0006,
+ "loss": 4.817290306091309,
+ "step": 3833
+ },
+ {
+ "epoch": 53.251528384279474,
+ "grad_norm": 0.04933721572160721,
+ "learning_rate": 0.0006,
+ "loss": 4.863317489624023,
+ "step": 3834
+ },
+ {
+ "epoch": 53.26550218340611,
+ "grad_norm": 0.04588694870471954,
+ "learning_rate": 0.0006,
+ "loss": 4.864742755889893,
+ "step": 3835
+ },
+ {
+ "epoch": 53.27947598253275,
+ "grad_norm": 0.02728264033794403,
+ "learning_rate": 0.0006,
+ "loss": 4.9068756103515625,
+ "step": 3836
+ },
+ {
+ "epoch": 53.29344978165939,
+ "grad_norm": 0.036240726709365845,
+ "learning_rate": 0.0006,
+ "loss": 4.902045249938965,
+ "step": 3837
+ },
+ {
+ "epoch": 53.30742358078603,
+ "grad_norm": 0.02703195810317993,
+ "learning_rate": 0.0006,
+ "loss": 4.908980369567871,
+ "step": 3838
+ },
+ {
+ "epoch": 53.32139737991266,
+ "grad_norm": 0.024894049391150475,
+ "learning_rate": 0.0006,
+ "loss": 4.794848442077637,
+ "step": 3839
+ },
+ {
+ "epoch": 53.3353711790393,
+ "grad_norm": 0.02389993891119957,
+ "learning_rate": 0.0006,
+ "loss": 4.8234734535217285,
+ "step": 3840
+ },
+ {
+ "epoch": 53.34934497816594,
+ "grad_norm": 0.02486400306224823,
+ "learning_rate": 0.0006,
+ "loss": 4.851463794708252,
+ "step": 3841
+ },
+ {
+ "epoch": 53.36331877729258,
+ "grad_norm": 0.02613813802599907,
+ "learning_rate": 0.0006,
+ "loss": 4.860444068908691,
+ "step": 3842
+ },
+ {
+ "epoch": 53.377292576419215,
+ "grad_norm": 0.0229744054377079,
+ "learning_rate": 0.0006,
+ "loss": 4.709606170654297,
+ "step": 3843
+ },
+ {
+ "epoch": 53.391266375545854,
+ "grad_norm": 0.02123042568564415,
+ "learning_rate": 0.0006,
+ "loss": 4.654613494873047,
+ "step": 3844
+ },
+ {
+ "epoch": 53.40524017467249,
+ "grad_norm": 0.017352338880300522,
+ "learning_rate": 0.0006,
+ "loss": 4.769761085510254,
+ "step": 3845
+ },
+ {
+ "epoch": 53.419213973799124,
+ "grad_norm": 0.019808724522590637,
+ "learning_rate": 0.0006,
+ "loss": 4.98514461517334,
+ "step": 3846
+ },
+ {
+ "epoch": 53.43318777292576,
+ "grad_norm": 0.018776152282953262,
+ "learning_rate": 0.0006,
+ "loss": 4.749990463256836,
+ "step": 3847
+ },
+ {
+ "epoch": 53.4471615720524,
+ "grad_norm": 0.02017521858215332,
+ "learning_rate": 0.0006,
+ "loss": 4.757381439208984,
+ "step": 3848
+ },
+ {
+ "epoch": 53.46113537117904,
+ "grad_norm": 0.019509416073560715,
+ "learning_rate": 0.0006,
+ "loss": 4.8111572265625,
+ "step": 3849
+ },
+ {
+ "epoch": 53.47510917030568,
+ "grad_norm": 0.015543580986559391,
+ "learning_rate": 0.0006,
+ "loss": 4.7501678466796875,
+ "step": 3850
+ },
+ {
+ "epoch": 53.48908296943232,
+ "grad_norm": 0.017941651865839958,
+ "learning_rate": 0.0006,
+ "loss": 4.701838493347168,
+ "step": 3851
+ },
+ {
+ "epoch": 53.50305676855895,
+ "grad_norm": 0.01678212359547615,
+ "learning_rate": 0.0006,
+ "loss": 4.787798881530762,
+ "step": 3852
+ },
+ {
+ "epoch": 53.51703056768559,
+ "grad_norm": 0.019050654023885727,
+ "learning_rate": 0.0006,
+ "loss": 4.754859924316406,
+ "step": 3853
+ },
+ {
+ "epoch": 53.531004366812226,
+ "grad_norm": 0.017406078055500984,
+ "learning_rate": 0.0006,
+ "loss": 4.752772808074951,
+ "step": 3854
+ },
+ {
+ "epoch": 53.544978165938865,
+ "grad_norm": 0.01773344725370407,
+ "learning_rate": 0.0006,
+ "loss": 4.651812553405762,
+ "step": 3855
+ },
+ {
+ "epoch": 53.5589519650655,
+ "grad_norm": 0.017221346497535706,
+ "learning_rate": 0.0006,
+ "loss": 4.792244911193848,
+ "step": 3856
+ },
+ {
+ "epoch": 53.57292576419214,
+ "grad_norm": 0.01722758449614048,
+ "learning_rate": 0.0006,
+ "loss": 4.7902960777282715,
+ "step": 3857
+ },
+ {
+ "epoch": 53.58689956331878,
+ "grad_norm": 0.020445192232728004,
+ "learning_rate": 0.0006,
+ "loss": 4.67392110824585,
+ "step": 3858
+ },
+ {
+ "epoch": 53.60087336244541,
+ "grad_norm": 0.02267114631831646,
+ "learning_rate": 0.0006,
+ "loss": 4.796089172363281,
+ "step": 3859
+ },
+ {
+ "epoch": 53.61484716157205,
+ "grad_norm": 0.027129942551255226,
+ "learning_rate": 0.0006,
+ "loss": 4.774303913116455,
+ "step": 3860
+ },
+ {
+ "epoch": 53.62882096069869,
+ "grad_norm": 0.025454839691519737,
+ "learning_rate": 0.0006,
+ "loss": 4.708813190460205,
+ "step": 3861
+ },
+ {
+ "epoch": 53.64279475982533,
+ "grad_norm": 0.019248757511377335,
+ "learning_rate": 0.0006,
+ "loss": 4.760467529296875,
+ "step": 3862
+ },
+ {
+ "epoch": 53.65676855895197,
+ "grad_norm": 0.01791176199913025,
+ "learning_rate": 0.0006,
+ "loss": 4.789203643798828,
+ "step": 3863
+ },
+ {
+ "epoch": 53.670742358078606,
+ "grad_norm": 0.016306612640619278,
+ "learning_rate": 0.0006,
+ "loss": 4.732605934143066,
+ "step": 3864
+ },
+ {
+ "epoch": 53.68471615720524,
+ "grad_norm": 0.017836976796388626,
+ "learning_rate": 0.0006,
+ "loss": 4.691871643066406,
+ "step": 3865
+ },
+ {
+ "epoch": 53.698689956331876,
+ "grad_norm": 0.015102861449122429,
+ "learning_rate": 0.0006,
+ "loss": 4.724156856536865,
+ "step": 3866
+ },
+ {
+ "epoch": 53.712663755458514,
+ "grad_norm": 0.017458142712712288,
+ "learning_rate": 0.0006,
+ "loss": 4.73524808883667,
+ "step": 3867
+ },
+ {
+ "epoch": 53.72663755458515,
+ "grad_norm": 0.01591932773590088,
+ "learning_rate": 0.0006,
+ "loss": 4.705888748168945,
+ "step": 3868
+ },
+ {
+ "epoch": 53.74061135371179,
+ "grad_norm": 0.01713593676686287,
+ "learning_rate": 0.0006,
+ "loss": 4.644516944885254,
+ "step": 3869
+ },
+ {
+ "epoch": 53.75458515283843,
+ "grad_norm": 0.01905885338783264,
+ "learning_rate": 0.0006,
+ "loss": 4.698416233062744,
+ "step": 3870
+ },
+ {
+ "epoch": 53.76855895196506,
+ "grad_norm": 0.01766437292098999,
+ "learning_rate": 0.0006,
+ "loss": 4.6796555519104,
+ "step": 3871
+ },
+ {
+ "epoch": 53.7825327510917,
+ "grad_norm": 0.01552243810147047,
+ "learning_rate": 0.0006,
+ "loss": 4.78138542175293,
+ "step": 3872
+ },
+ {
+ "epoch": 53.79650655021834,
+ "grad_norm": 0.014191331341862679,
+ "learning_rate": 0.0006,
+ "loss": 4.703588485717773,
+ "step": 3873
+ },
+ {
+ "epoch": 53.81048034934498,
+ "grad_norm": 0.015872279182076454,
+ "learning_rate": 0.0006,
+ "loss": 4.705611228942871,
+ "step": 3874
+ },
+ {
+ "epoch": 53.82445414847162,
+ "grad_norm": 0.01787560060620308,
+ "learning_rate": 0.0006,
+ "loss": 4.8227949142456055,
+ "step": 3875
+ },
+ {
+ "epoch": 53.838427947598255,
+ "grad_norm": 0.0158209390938282,
+ "learning_rate": 0.0006,
+ "loss": 4.645227432250977,
+ "step": 3876
+ },
+ {
+ "epoch": 53.852401746724894,
+ "grad_norm": 0.014362678863108158,
+ "learning_rate": 0.0006,
+ "loss": 4.7225422859191895,
+ "step": 3877
+ },
+ {
+ "epoch": 53.866375545851525,
+ "grad_norm": 0.014899161644279957,
+ "learning_rate": 0.0006,
+ "loss": 4.685914039611816,
+ "step": 3878
+ },
+ {
+ "epoch": 53.880349344978164,
+ "grad_norm": 0.01628505438566208,
+ "learning_rate": 0.0006,
+ "loss": 4.674065589904785,
+ "step": 3879
+ },
+ {
+ "epoch": 53.8943231441048,
+ "grad_norm": 0.01873331144452095,
+ "learning_rate": 0.0006,
+ "loss": 4.811744213104248,
+ "step": 3880
+ },
+ {
+ "epoch": 53.90829694323144,
+ "grad_norm": 0.02229069173336029,
+ "learning_rate": 0.0006,
+ "loss": 4.721514701843262,
+ "step": 3881
+ },
+ {
+ "epoch": 53.92227074235808,
+ "grad_norm": 0.028012916445732117,
+ "learning_rate": 0.0006,
+ "loss": 4.773603439331055,
+ "step": 3882
+ },
+ {
+ "epoch": 53.93624454148472,
+ "grad_norm": 0.036322783678770065,
+ "learning_rate": 0.0006,
+ "loss": 4.787134170532227,
+ "step": 3883
+ },
+ {
+ "epoch": 53.95021834061135,
+ "grad_norm": 0.048149142414331436,
+ "learning_rate": 0.0006,
+ "loss": 4.736196517944336,
+ "step": 3884
+ },
+ {
+ "epoch": 53.96419213973799,
+ "grad_norm": 0.051918916404247284,
+ "learning_rate": 0.0006,
+ "loss": 4.686728000640869,
+ "step": 3885
+ },
+ {
+ "epoch": 53.97816593886463,
+ "grad_norm": 0.0481342189013958,
+ "learning_rate": 0.0006,
+ "loss": 4.733730792999268,
+ "step": 3886
+ },
+ {
+ "epoch": 53.992139737991266,
+ "grad_norm": 0.045424818992614746,
+ "learning_rate": 0.0006,
+ "loss": 4.748723983764648,
+ "step": 3887
+ },
+ {
+ "epoch": 54.0,
+ "grad_norm": 0.04153316840529442,
+ "learning_rate": 0.0006,
+ "loss": 4.715220928192139,
+ "step": 3888
+ },
+ {
+ "epoch": 54.0,
+ "eval_loss": 4.964215278625488,
+ "eval_runtime": 56.7757,
+ "eval_samples_per_second": 43.011,
+ "eval_steps_per_second": 1.356,
+ "step": 3888
+ },
+ {
+ "epoch": 54.01397379912664,
+ "grad_norm": 0.03459750860929489,
+ "learning_rate": 0.0006,
+ "loss": 4.661557674407959,
+ "step": 3889
+ },
+ {
+ "epoch": 54.02794759825328,
+ "grad_norm": 0.025409208610653877,
+ "learning_rate": 0.0006,
+ "loss": 4.7065582275390625,
+ "step": 3890
+ },
+ {
+ "epoch": 54.041921397379916,
+ "grad_norm": 0.03202550485730171,
+ "learning_rate": 0.0006,
+ "loss": 4.751070976257324,
+ "step": 3891
+ },
+ {
+ "epoch": 54.05589519650655,
+ "grad_norm": 0.02968018501996994,
+ "learning_rate": 0.0006,
+ "loss": 4.635388374328613,
+ "step": 3892
+ },
+ {
+ "epoch": 54.069868995633186,
+ "grad_norm": 0.02757406234741211,
+ "learning_rate": 0.0006,
+ "loss": 4.706214427947998,
+ "step": 3893
+ },
+ {
+ "epoch": 54.083842794759825,
+ "grad_norm": 0.025223620235919952,
+ "learning_rate": 0.0006,
+ "loss": 4.703792572021484,
+ "step": 3894
+ },
+ {
+ "epoch": 54.09781659388646,
+ "grad_norm": 0.02431778982281685,
+ "learning_rate": 0.0006,
+ "loss": 4.658116340637207,
+ "step": 3895
+ },
+ {
+ "epoch": 54.1117903930131,
+ "grad_norm": 0.02792172320187092,
+ "learning_rate": 0.0006,
+ "loss": 4.638895034790039,
+ "step": 3896
+ },
+ {
+ "epoch": 54.12576419213974,
+ "grad_norm": 0.032253511250019073,
+ "learning_rate": 0.0006,
+ "loss": 4.693902492523193,
+ "step": 3897
+ },
+ {
+ "epoch": 54.13973799126637,
+ "grad_norm": 0.027044931426644325,
+ "learning_rate": 0.0006,
+ "loss": 4.575805187225342,
+ "step": 3898
+ },
+ {
+ "epoch": 54.15371179039301,
+ "grad_norm": 0.023270361125469208,
+ "learning_rate": 0.0006,
+ "loss": 4.67645263671875,
+ "step": 3899
+ },
+ {
+ "epoch": 54.16768558951965,
+ "grad_norm": 0.01929079182446003,
+ "learning_rate": 0.0006,
+ "loss": 4.68062686920166,
+ "step": 3900
+ },
+ {
+ "epoch": 54.18165938864629,
+ "grad_norm": 0.020343473181128502,
+ "learning_rate": 0.0006,
+ "loss": 4.708086967468262,
+ "step": 3901
+ },
+ {
+ "epoch": 54.19563318777293,
+ "grad_norm": 0.0211793202906847,
+ "learning_rate": 0.0006,
+ "loss": 4.677323341369629,
+ "step": 3902
+ },
+ {
+ "epoch": 54.209606986899566,
+ "grad_norm": 0.020341217517852783,
+ "learning_rate": 0.0006,
+ "loss": 4.760937690734863,
+ "step": 3903
+ },
+ {
+ "epoch": 54.223580786026204,
+ "grad_norm": 0.019421054050326347,
+ "learning_rate": 0.0006,
+ "loss": 4.508834362030029,
+ "step": 3904
+ },
+ {
+ "epoch": 54.237554585152836,
+ "grad_norm": 0.016701679676771164,
+ "learning_rate": 0.0006,
+ "loss": 4.664399147033691,
+ "step": 3905
+ },
+ {
+ "epoch": 54.251528384279474,
+ "grad_norm": 0.016561470925807953,
+ "learning_rate": 0.0006,
+ "loss": 4.742990493774414,
+ "step": 3906
+ },
+ {
+ "epoch": 54.26550218340611,
+ "grad_norm": 0.017446085810661316,
+ "learning_rate": 0.0006,
+ "loss": 4.63038444519043,
+ "step": 3907
+ },
+ {
+ "epoch": 54.27947598253275,
+ "grad_norm": 0.014340086840093136,
+ "learning_rate": 0.0006,
+ "loss": 4.688791275024414,
+ "step": 3908
+ },
+ {
+ "epoch": 54.29344978165939,
+ "grad_norm": 0.013580698519945145,
+ "learning_rate": 0.0006,
+ "loss": 4.625302314758301,
+ "step": 3909
+ },
+ {
+ "epoch": 54.30742358078603,
+ "grad_norm": 0.014931539073586464,
+ "learning_rate": 0.0006,
+ "loss": 4.6167497634887695,
+ "step": 3910
+ },
+ {
+ "epoch": 54.32139737991266,
+ "grad_norm": 0.013492150232195854,
+ "learning_rate": 0.0006,
+ "loss": 4.657521724700928,
+ "step": 3911
+ },
+ {
+ "epoch": 54.3353711790393,
+ "grad_norm": 0.015151413157582283,
+ "learning_rate": 0.0006,
+ "loss": 4.634521484375,
+ "step": 3912
+ },
+ {
+ "epoch": 54.34934497816594,
+ "grad_norm": 0.01302040833979845,
+ "learning_rate": 0.0006,
+ "loss": 4.718928337097168,
+ "step": 3913
+ },
+ {
+ "epoch": 54.36331877729258,
+ "grad_norm": 0.013433529995381832,
+ "learning_rate": 0.0006,
+ "loss": 4.584803581237793,
+ "step": 3914
+ },
+ {
+ "epoch": 54.377292576419215,
+ "grad_norm": 0.012987039051949978,
+ "learning_rate": 0.0006,
+ "loss": 4.6393585205078125,
+ "step": 3915
+ },
+ {
+ "epoch": 54.391266375545854,
+ "grad_norm": 0.013695075176656246,
+ "learning_rate": 0.0006,
+ "loss": 4.556313514709473,
+ "step": 3916
+ },
+ {
+ "epoch": 54.40524017467249,
+ "grad_norm": 0.013718255795538425,
+ "learning_rate": 0.0006,
+ "loss": 4.576223850250244,
+ "step": 3917
+ },
+ {
+ "epoch": 54.419213973799124,
+ "grad_norm": 0.013699112460017204,
+ "learning_rate": 0.0006,
+ "loss": 4.63889217376709,
+ "step": 3918
+ },
+ {
+ "epoch": 54.43318777292576,
+ "grad_norm": 0.013516264036297798,
+ "learning_rate": 0.0006,
+ "loss": 4.669532299041748,
+ "step": 3919
+ },
+ {
+ "epoch": 54.4471615720524,
+ "grad_norm": 0.013597294688224792,
+ "learning_rate": 0.0006,
+ "loss": 4.6644287109375,
+ "step": 3920
+ },
+ {
+ "epoch": 54.46113537117904,
+ "grad_norm": 0.013185849413275719,
+ "learning_rate": 0.0006,
+ "loss": 4.636900901794434,
+ "step": 3921
+ },
+ {
+ "epoch": 54.47510917030568,
+ "grad_norm": 0.014311990700662136,
+ "learning_rate": 0.0006,
+ "loss": 4.6047587394714355,
+ "step": 3922
+ },
+ {
+ "epoch": 54.48908296943232,
+ "grad_norm": 0.012426027096807957,
+ "learning_rate": 0.0006,
+ "loss": 4.6161394119262695,
+ "step": 3923
+ },
+ {
+ "epoch": 54.50305676855895,
+ "grad_norm": 0.01283493172377348,
+ "learning_rate": 0.0006,
+ "loss": 4.584190368652344,
+ "step": 3924
+ },
+ {
+ "epoch": 54.51703056768559,
+ "grad_norm": 0.01339943427592516,
+ "learning_rate": 0.0006,
+ "loss": 4.624184608459473,
+ "step": 3925
+ },
+ {
+ "epoch": 54.531004366812226,
+ "grad_norm": 0.012859724462032318,
+ "learning_rate": 0.0006,
+ "loss": 4.6174635887146,
+ "step": 3926
+ },
+ {
+ "epoch": 54.544978165938865,
+ "grad_norm": 0.014420023187994957,
+ "learning_rate": 0.0006,
+ "loss": 4.536544322967529,
+ "step": 3927
+ },
+ {
+ "epoch": 54.5589519650655,
+ "grad_norm": 0.014237185008823872,
+ "learning_rate": 0.0006,
+ "loss": 4.681514739990234,
+ "step": 3928
+ },
+ {
+ "epoch": 54.57292576419214,
+ "grad_norm": 0.013340328820049763,
+ "learning_rate": 0.0006,
+ "loss": 4.579689979553223,
+ "step": 3929
+ },
+ {
+ "epoch": 54.58689956331878,
+ "grad_norm": 0.013894530944526196,
+ "learning_rate": 0.0006,
+ "loss": 4.638958930969238,
+ "step": 3930
+ },
+ {
+ "epoch": 54.60087336244541,
+ "grad_norm": 0.016181141138076782,
+ "learning_rate": 0.0006,
+ "loss": 4.573508262634277,
+ "step": 3931
+ },
+ {
+ "epoch": 54.61484716157205,
+ "grad_norm": 0.013221992179751396,
+ "learning_rate": 0.0006,
+ "loss": 4.6229047775268555,
+ "step": 3932
+ },
+ {
+ "epoch": 54.62882096069869,
+ "grad_norm": 0.013241561129689217,
+ "learning_rate": 0.0006,
+ "loss": 4.641152858734131,
+ "step": 3933
+ },
+ {
+ "epoch": 54.64279475982533,
+ "grad_norm": 0.016237538307905197,
+ "learning_rate": 0.0006,
+ "loss": 4.6421217918396,
+ "step": 3934
+ },
+ {
+ "epoch": 54.65676855895197,
+ "grad_norm": 0.020866507664322853,
+ "learning_rate": 0.0006,
+ "loss": 4.646058082580566,
+ "step": 3935
+ },
+ {
+ "epoch": 54.670742358078606,
+ "grad_norm": 0.022182893007993698,
+ "learning_rate": 0.0006,
+ "loss": 4.554501533508301,
+ "step": 3936
+ },
+ {
+ "epoch": 54.68471615720524,
+ "grad_norm": 0.02254868485033512,
+ "learning_rate": 0.0006,
+ "loss": 4.563869953155518,
+ "step": 3937
+ },
+ {
+ "epoch": 54.698689956331876,
+ "grad_norm": 0.019730940461158752,
+ "learning_rate": 0.0006,
+ "loss": 4.592165946960449,
+ "step": 3938
+ },
+ {
+ "epoch": 54.712663755458514,
+ "grad_norm": 0.015146561898291111,
+ "learning_rate": 0.0006,
+ "loss": 4.6228742599487305,
+ "step": 3939
+ },
+ {
+ "epoch": 54.72663755458515,
+ "grad_norm": 0.01824716106057167,
+ "learning_rate": 0.0006,
+ "loss": 4.716236114501953,
+ "step": 3940
+ },
+ {
+ "epoch": 54.74061135371179,
+ "grad_norm": 0.017730550840497017,
+ "learning_rate": 0.0006,
+ "loss": 4.650232315063477,
+ "step": 3941
+ },
+ {
+ "epoch": 54.75458515283843,
+ "grad_norm": 0.014875595457851887,
+ "learning_rate": 0.0006,
+ "loss": 4.6377458572387695,
+ "step": 3942
+ },
+ {
+ "epoch": 54.76855895196506,
+ "grad_norm": 0.01613209955394268,
+ "learning_rate": 0.0006,
+ "loss": 4.5768537521362305,
+ "step": 3943
+ },
+ {
+ "epoch": 54.7825327510917,
+ "grad_norm": 0.017311280593276024,
+ "learning_rate": 0.0006,
+ "loss": 4.625744342803955,
+ "step": 3944
+ },
+ {
+ "epoch": 54.79650655021834,
+ "grad_norm": 0.018770085647702217,
+ "learning_rate": 0.0006,
+ "loss": 4.613787651062012,
+ "step": 3945
+ },
+ {
+ "epoch": 54.81048034934498,
+ "grad_norm": 0.0165257528424263,
+ "learning_rate": 0.0006,
+ "loss": 4.5991315841674805,
+ "step": 3946
+ },
+ {
+ "epoch": 54.82445414847162,
+ "grad_norm": 0.017433857545256615,
+ "learning_rate": 0.0006,
+ "loss": 4.513872146606445,
+ "step": 3947
+ },
+ {
+ "epoch": 54.838427947598255,
+ "grad_norm": 0.020126905292272568,
+ "learning_rate": 0.0006,
+ "loss": 4.616185188293457,
+ "step": 3948
+ },
+ {
+ "epoch": 54.852401746724894,
+ "grad_norm": 0.01819358579814434,
+ "learning_rate": 0.0006,
+ "loss": 4.535691261291504,
+ "step": 3949
+ },
+ {
+ "epoch": 54.866375545851525,
+ "grad_norm": 0.01899927295744419,
+ "learning_rate": 0.0006,
+ "loss": 4.595199108123779,
+ "step": 3950
+ },
+ {
+ "epoch": 54.880349344978164,
+ "grad_norm": 0.02005566656589508,
+ "learning_rate": 0.0006,
+ "loss": 4.633289337158203,
+ "step": 3951
+ },
+ {
+ "epoch": 54.8943231441048,
+ "grad_norm": 0.021570220589637756,
+ "learning_rate": 0.0006,
+ "loss": 4.592503547668457,
+ "step": 3952
+ },
+ {
+ "epoch": 54.90829694323144,
+ "grad_norm": 0.01883506216108799,
+ "learning_rate": 0.0006,
+ "loss": 4.6597795486450195,
+ "step": 3953
+ },
+ {
+ "epoch": 54.92227074235808,
+ "grad_norm": 0.018070021644234657,
+ "learning_rate": 0.0006,
+ "loss": 4.601474285125732,
+ "step": 3954
+ },
+ {
+ "epoch": 54.93624454148472,
+ "grad_norm": 0.018090378493070602,
+ "learning_rate": 0.0006,
+ "loss": 4.569846153259277,
+ "step": 3955
+ },
+ {
+ "epoch": 54.95021834061135,
+ "grad_norm": 0.018751563504338264,
+ "learning_rate": 0.0006,
+ "loss": 4.646064281463623,
+ "step": 3956
+ },
+ {
+ "epoch": 54.96419213973799,
+ "grad_norm": 0.017512254416942596,
+ "learning_rate": 0.0006,
+ "loss": 4.606670379638672,
+ "step": 3957
+ },
+ {
+ "epoch": 54.97816593886463,
+ "grad_norm": 0.018771110102534294,
+ "learning_rate": 0.0006,
+ "loss": 4.604315757751465,
+ "step": 3958
+ },
+ {
+ "epoch": 54.992139737991266,
+ "grad_norm": 0.01611141301691532,
+ "learning_rate": 0.0006,
+ "loss": 4.643245220184326,
+ "step": 3959
+ },
+ {
+ "epoch": 55.0,
+ "grad_norm": 0.0156728383153677,
+ "learning_rate": 0.0006,
+ "loss": 4.6739044189453125,
+ "step": 3960
+ },
+ {
+ "epoch": 55.0,
+ "eval_loss": 4.851032257080078,
+ "eval_runtime": 56.8699,
+ "eval_samples_per_second": 42.94,
+ "eval_steps_per_second": 1.354,
+ "step": 3960
+ },
+ {
+ "epoch": 55.01397379912664,
+ "grad_norm": 0.016214672476053238,
+ "learning_rate": 0.0006,
+ "loss": 4.571749687194824,
+ "step": 3961
+ },
+ {
+ "epoch": 55.02794759825328,
+ "grad_norm": 0.01649155467748642,
+ "learning_rate": 0.0006,
+ "loss": 4.553592205047607,
+ "step": 3962
+ },
+ {
+ "epoch": 55.041921397379916,
+ "grad_norm": 0.016793379560112953,
+ "learning_rate": 0.0006,
+ "loss": 4.650628566741943,
+ "step": 3963
+ },
+ {
+ "epoch": 55.05589519650655,
+ "grad_norm": 0.0191937405616045,
+ "learning_rate": 0.0006,
+ "loss": 4.620532989501953,
+ "step": 3964
+ },
+ {
+ "epoch": 55.069868995633186,
+ "grad_norm": 0.020976556465029716,
+ "learning_rate": 0.0006,
+ "loss": 4.6329545974731445,
+ "step": 3965
+ },
+ {
+ "epoch": 55.083842794759825,
+ "grad_norm": 0.023450477048754692,
+ "learning_rate": 0.0006,
+ "loss": 4.60828971862793,
+ "step": 3966
+ },
+ {
+ "epoch": 55.09781659388646,
+ "grad_norm": 0.026778312399983406,
+ "learning_rate": 0.0006,
+ "loss": 4.617481708526611,
+ "step": 3967
+ },
+ {
+ "epoch": 55.1117903930131,
+ "grad_norm": 0.02333568036556244,
+ "learning_rate": 0.0006,
+ "loss": 4.580722332000732,
+ "step": 3968
+ },
+ {
+ "epoch": 55.12576419213974,
+ "grad_norm": 0.018491310998797417,
+ "learning_rate": 0.0006,
+ "loss": 4.590953350067139,
+ "step": 3969
+ },
+ {
+ "epoch": 55.13973799126637,
+ "grad_norm": 0.020339803770184517,
+ "learning_rate": 0.0006,
+ "loss": 4.518374443054199,
+ "step": 3970
+ },
+ {
+ "epoch": 55.15371179039301,
+ "grad_norm": 0.01759706623852253,
+ "learning_rate": 0.0006,
+ "loss": 4.560522079467773,
+ "step": 3971
+ },
+ {
+ "epoch": 55.16768558951965,
+ "grad_norm": 0.015875980257987976,
+ "learning_rate": 0.0006,
+ "loss": 4.4795708656311035,
+ "step": 3972
+ },
+ {
+ "epoch": 55.18165938864629,
+ "grad_norm": 0.01650134287774563,
+ "learning_rate": 0.0006,
+ "loss": 4.57435941696167,
+ "step": 3973
+ },
+ {
+ "epoch": 55.19563318777293,
+ "grad_norm": 0.01734331250190735,
+ "learning_rate": 0.0006,
+ "loss": 4.661007881164551,
+ "step": 3974
+ },
+ {
+ "epoch": 55.209606986899566,
+ "grad_norm": 0.01651303470134735,
+ "learning_rate": 0.0006,
+ "loss": 4.511148452758789,
+ "step": 3975
+ },
+ {
+ "epoch": 55.223580786026204,
+ "grad_norm": 0.016189761459827423,
+ "learning_rate": 0.0006,
+ "loss": 4.592350006103516,
+ "step": 3976
+ },
+ {
+ "epoch": 55.237554585152836,
+ "grad_norm": 0.01833222061395645,
+ "learning_rate": 0.0006,
+ "loss": 4.506649971008301,
+ "step": 3977
+ },
+ {
+ "epoch": 55.251528384279474,
+ "grad_norm": 0.02433410845696926,
+ "learning_rate": 0.0006,
+ "loss": 4.49500036239624,
+ "step": 3978
+ },
+ {
+ "epoch": 55.26550218340611,
+ "grad_norm": 0.0314289927482605,
+ "learning_rate": 0.0006,
+ "loss": 4.507624626159668,
+ "step": 3979
+ },
+ {
+ "epoch": 55.27947598253275,
+ "grad_norm": 0.03013679012656212,
+ "learning_rate": 0.0006,
+ "loss": 4.533641815185547,
+ "step": 3980
+ },
+ {
+ "epoch": 55.29344978165939,
+ "grad_norm": 0.024524061009287834,
+ "learning_rate": 0.0006,
+ "loss": 4.5721116065979,
+ "step": 3981
+ },
+ {
+ "epoch": 55.30742358078603,
+ "grad_norm": 0.02640336938202381,
+ "learning_rate": 0.0006,
+ "loss": 4.652521133422852,
+ "step": 3982
+ },
+ {
+ "epoch": 55.32139737991266,
+ "grad_norm": 0.021309101954102516,
+ "learning_rate": 0.0006,
+ "loss": 4.600375175476074,
+ "step": 3983
+ },
+ {
+ "epoch": 55.3353711790393,
+ "grad_norm": 0.023008549585938454,
+ "learning_rate": 0.0006,
+ "loss": 4.53914737701416,
+ "step": 3984
+ },
+ {
+ "epoch": 55.34934497816594,
+ "grad_norm": 0.022491468116641045,
+ "learning_rate": 0.0006,
+ "loss": 4.661163330078125,
+ "step": 3985
+ },
+ {
+ "epoch": 55.36331877729258,
+ "grad_norm": 0.01904081553220749,
+ "learning_rate": 0.0006,
+ "loss": 4.5426177978515625,
+ "step": 3986
+ },
+ {
+ "epoch": 55.377292576419215,
+ "grad_norm": 0.019229518249630928,
+ "learning_rate": 0.0006,
+ "loss": 4.606066703796387,
+ "step": 3987
+ },
+ {
+ "epoch": 55.391266375545854,
+ "grad_norm": 0.01781383901834488,
+ "learning_rate": 0.0006,
+ "loss": 4.54642391204834,
+ "step": 3988
+ },
+ {
+ "epoch": 55.40524017467249,
+ "grad_norm": 0.0174600500613451,
+ "learning_rate": 0.0006,
+ "loss": 4.711891174316406,
+ "step": 3989
+ },
+ {
+ "epoch": 55.419213973799124,
+ "grad_norm": 0.018154339864850044,
+ "learning_rate": 0.0006,
+ "loss": 4.527122974395752,
+ "step": 3990
+ },
+ {
+ "epoch": 55.43318777292576,
+ "grad_norm": 0.01914648339152336,
+ "learning_rate": 0.0006,
+ "loss": 4.641975402832031,
+ "step": 3991
+ },
+ {
+ "epoch": 55.4471615720524,
+ "grad_norm": 0.017118625342845917,
+ "learning_rate": 0.0006,
+ "loss": 4.511425971984863,
+ "step": 3992
+ },
+ {
+ "epoch": 55.46113537117904,
+ "grad_norm": 0.01551489345729351,
+ "learning_rate": 0.0006,
+ "loss": 4.551141738891602,
+ "step": 3993
+ },
+ {
+ "epoch": 55.47510917030568,
+ "grad_norm": 0.015072065405547619,
+ "learning_rate": 0.0006,
+ "loss": 4.651115894317627,
+ "step": 3994
+ },
+ {
+ "epoch": 55.48908296943232,
+ "grad_norm": 0.017010986804962158,
+ "learning_rate": 0.0006,
+ "loss": 4.583282470703125,
+ "step": 3995
+ },
+ {
+ "epoch": 55.50305676855895,
+ "grad_norm": 0.01764831133186817,
+ "learning_rate": 0.0006,
+ "loss": 4.593364238739014,
+ "step": 3996
+ },
+ {
+ "epoch": 55.51703056768559,
+ "grad_norm": 0.01762884296476841,
+ "learning_rate": 0.0006,
+ "loss": 4.45671272277832,
+ "step": 3997
+ },
+ {
+ "epoch": 55.531004366812226,
+ "grad_norm": 0.01600457727909088,
+ "learning_rate": 0.0006,
+ "loss": 4.511260032653809,
+ "step": 3998
+ },
+ {
+ "epoch": 55.544978165938865,
+ "grad_norm": 0.015706021338701248,
+ "learning_rate": 0.0006,
+ "loss": 4.47593355178833,
+ "step": 3999
+ },
+ {
+ "epoch": 55.5589519650655,
+ "grad_norm": 0.017718996852636337,
+ "learning_rate": 0.0006,
+ "loss": 4.610543251037598,
+ "step": 4000
+ },
+ {
+ "epoch": 55.57292576419214,
+ "grad_norm": 0.015073757618665695,
+ "learning_rate": 0.0006,
+ "loss": 4.466620445251465,
+ "step": 4001
+ },
+ {
+ "epoch": 55.58689956331878,
+ "grad_norm": 0.01789247989654541,
+ "learning_rate": 0.0006,
+ "loss": 4.503409385681152,
+ "step": 4002
+ },
+ {
+ "epoch": 55.60087336244541,
+ "grad_norm": 0.022329598665237427,
+ "learning_rate": 0.0006,
+ "loss": 4.62099552154541,
+ "step": 4003
+ },
+ {
+ "epoch": 55.61484716157205,
+ "grad_norm": 0.02343829534947872,
+ "learning_rate": 0.0006,
+ "loss": 4.620813369750977,
+ "step": 4004
+ },
+ {
+ "epoch": 55.62882096069869,
+ "grad_norm": 0.023279687389731407,
+ "learning_rate": 0.0006,
+ "loss": 4.545765399932861,
+ "step": 4005
+ },
+ {
+ "epoch": 55.64279475982533,
+ "grad_norm": 0.020817464217543602,
+ "learning_rate": 0.0006,
+ "loss": 4.561470031738281,
+ "step": 4006
+ },
+ {
+ "epoch": 55.65676855895197,
+ "grad_norm": 0.020270058885216713,
+ "learning_rate": 0.0006,
+ "loss": 4.5875020027160645,
+ "step": 4007
+ },
+ {
+ "epoch": 55.670742358078606,
+ "grad_norm": 0.021085843443870544,
+ "learning_rate": 0.0006,
+ "loss": 4.504560470581055,
+ "step": 4008
+ },
+ {
+ "epoch": 55.68471615720524,
+ "grad_norm": 0.018507730215787888,
+ "learning_rate": 0.0006,
+ "loss": 4.520644664764404,
+ "step": 4009
+ },
+ {
+ "epoch": 55.698689956331876,
+ "grad_norm": 0.020571516826748848,
+ "learning_rate": 0.0006,
+ "loss": 4.539643287658691,
+ "step": 4010
+ },
+ {
+ "epoch": 55.712663755458514,
+ "grad_norm": 0.021484510973095894,
+ "learning_rate": 0.0006,
+ "loss": 4.628833293914795,
+ "step": 4011
+ },
+ {
+ "epoch": 55.72663755458515,
+ "grad_norm": 0.017693819478154182,
+ "learning_rate": 0.0006,
+ "loss": 4.6023783683776855,
+ "step": 4012
+ },
+ {
+ "epoch": 55.74061135371179,
+ "grad_norm": 0.020186059176921844,
+ "learning_rate": 0.0006,
+ "loss": 4.634884834289551,
+ "step": 4013
+ },
+ {
+ "epoch": 55.75458515283843,
+ "grad_norm": 0.01835780404508114,
+ "learning_rate": 0.0006,
+ "loss": 4.577791213989258,
+ "step": 4014
+ },
+ {
+ "epoch": 55.76855895196506,
+ "grad_norm": 0.018162988126277924,
+ "learning_rate": 0.0006,
+ "loss": 4.641597270965576,
+ "step": 4015
+ },
+ {
+ "epoch": 55.7825327510917,
+ "grad_norm": 0.018909504637122154,
+ "learning_rate": 0.0006,
+ "loss": 4.487447738647461,
+ "step": 4016
+ },
+ {
+ "epoch": 55.79650655021834,
+ "grad_norm": 0.018972251564264297,
+ "learning_rate": 0.0006,
+ "loss": 4.590317726135254,
+ "step": 4017
+ },
+ {
+ "epoch": 55.81048034934498,
+ "grad_norm": 0.01793607883155346,
+ "learning_rate": 0.0006,
+ "loss": 4.4994611740112305,
+ "step": 4018
+ },
+ {
+ "epoch": 55.82445414847162,
+ "grad_norm": 0.02123514749109745,
+ "learning_rate": 0.0006,
+ "loss": 4.479006767272949,
+ "step": 4019
+ },
+ {
+ "epoch": 55.838427947598255,
+ "grad_norm": 0.021246599033474922,
+ "learning_rate": 0.0006,
+ "loss": 4.4738640785217285,
+ "step": 4020
+ },
+ {
+ "epoch": 55.852401746724894,
+ "grad_norm": 0.020131602883338928,
+ "learning_rate": 0.0006,
+ "loss": 4.600308418273926,
+ "step": 4021
+ },
+ {
+ "epoch": 55.866375545851525,
+ "grad_norm": 0.019159460440278053,
+ "learning_rate": 0.0006,
+ "loss": 4.540908336639404,
+ "step": 4022
+ },
+ {
+ "epoch": 55.880349344978164,
+ "grad_norm": 0.023114705458283424,
+ "learning_rate": 0.0006,
+ "loss": 4.641247749328613,
+ "step": 4023
+ },
+ {
+ "epoch": 55.8943231441048,
+ "grad_norm": 0.02296135015785694,
+ "learning_rate": 0.0006,
+ "loss": 4.444624423980713,
+ "step": 4024
+ },
+ {
+ "epoch": 55.90829694323144,
+ "grad_norm": 0.02204655110836029,
+ "learning_rate": 0.0006,
+ "loss": 4.506349563598633,
+ "step": 4025
+ },
+ {
+ "epoch": 55.92227074235808,
+ "grad_norm": 0.019338225945830345,
+ "learning_rate": 0.0006,
+ "loss": 4.616047382354736,
+ "step": 4026
+ },
+ {
+ "epoch": 55.93624454148472,
+ "grad_norm": 0.020357884466648102,
+ "learning_rate": 0.0006,
+ "loss": 4.48170280456543,
+ "step": 4027
+ },
+ {
+ "epoch": 55.95021834061135,
+ "grad_norm": 0.02426772564649582,
+ "learning_rate": 0.0006,
+ "loss": 4.576877117156982,
+ "step": 4028
+ },
+ {
+ "epoch": 55.96419213973799,
+ "grad_norm": 0.023351816460490227,
+ "learning_rate": 0.0006,
+ "loss": 4.458885192871094,
+ "step": 4029
+ },
+ {
+ "epoch": 55.97816593886463,
+ "grad_norm": 0.019892243668437004,
+ "learning_rate": 0.0006,
+ "loss": 4.520229816436768,
+ "step": 4030
+ },
+ {
+ "epoch": 55.992139737991266,
+ "grad_norm": 0.018331089988350868,
+ "learning_rate": 0.0006,
+ "loss": 4.432201862335205,
+ "step": 4031
+ },
+ {
+ "epoch": 56.0,
+ "grad_norm": 0.021300828084349632,
+ "learning_rate": 0.0006,
+ "loss": 4.617072582244873,
+ "step": 4032
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 28800,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 400,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": true,
+ "should_log": false,
+ "should_save": false,
+ "should_training_stop": false
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 1.7189834823460454e+18,
+ "train_batch_size": 8,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4032/training_args.bin b/runs/l2r50-i2-fulle-lm/checkpoint-4032/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..cad232831c222cc7f8b938b8bbe42f61e7d89c3a
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-4032/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:f7c40e3bfa2f5ac1450ca1243e3f2a895d44a06ed5995e6aefd97f78d229d8c8
+size 4792
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4320/chat_template.jinja b/runs/l2r50-i2-fulle-lm/checkpoint-4320/chat_template.jinja
new file mode 100644
index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-4320/chat_template.jinja
@@ -0,0 +1,85 @@
+{%- if tools %}
+ {{- '<|im_start|>system\n' }}
+ {%- if messages[0].role == 'system' %}
+ {{- messages[0].content + '\n\n' }}
+ {%- endif %}
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }}
+ {%- for tool in tools %}
+ {{- "\n" }}
+ {{- tool | tojson }}
+ {%- endfor %}
+ {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }}
+{%- else %}
+ {%- if messages[0].role == 'system' %}
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
+ {%- endif %}
+{%- endif %}
+{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
+{%- for message in messages[::-1] %}
+ {%- set index = (messages|length - 1) - loop.index0 %}
+ {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %}
+ {%- set ns.multi_step_tool = false %}
+ {%- set ns.last_query_index = index %}
+ {%- endif %}
+{%- endfor %}
+{%- for message in messages %}
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
+ {%- elif message.role == "assistant" %}
+ {%- set content = message.content %}
+ {%- set reasoning_content = '' %}
+ {%- if message.reasoning_content is defined and message.reasoning_content is not none %}
+ {%- set reasoning_content = message.reasoning_content %}
+ {%- else %}
+ {%- if '' in message.content %}
+ {%- set content = message.content.split('')[-1].lstrip('\n') %}
+ {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %}
+ {%- endif %}
+ {%- endif %}
+ {%- if loop.index0 > ns.last_query_index %}
+ {%- if loop.last or (not loop.last and reasoning_content) %}
+ {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- if message.tool_calls %}
+ {%- for tool_call in message.tool_calls %}
+ {%- if (loop.first and content) or (not loop.first) %}
+ {{- '\n' }}
+ {%- endif %}
+ {%- if tool_call.function %}
+ {%- set tool_call = tool_call.function %}
+ {%- endif %}
+ {{- '\n{"name": "' }}
+ {{- tool_call.name }}
+ {{- '", "arguments": ' }}
+ {%- if tool_call.arguments is string %}
+ {{- tool_call.arguments }}
+ {%- else %}
+ {{- tool_call.arguments | tojson }}
+ {%- endif %}
+ {{- '}\n' }}
+ {%- endfor %}
+ {%- endif %}
+ {{- '<|im_end|>\n' }}
+ {%- elif message.role == "tool" %}
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
+ {{- '<|im_start|>user' }}
+ {%- endif %}
+ {{- '\n\n' }}
+ {{- message.content }}
+ {{- '\n' }}
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
+ {{- '<|im_end|>\n' }}
+ {%- endif %}
+ {%- endif %}
+{%- endfor %}
+{%- if add_generation_prompt %}
+ {{- '<|im_start|>assistant\n' }}
+ {%- if enable_thinking is defined and enable_thinking is false %}
+ {{- '\n\n\n\n' }}
+ {%- endif %}
+{%- endif %}
\ No newline at end of file
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4320/config.json b/runs/l2r50-i2-fulle-lm/checkpoint-4320/config.json
new file mode 100644
index 0000000000000000000000000000000000000000..22f250953c6ce8fc91ee4411856dd73041a415e2
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-4320/config.json
@@ -0,0 +1,32 @@
+{
+ "architectures": [
+ "LlamaForCausalLM"
+ ],
+ "attention_bias": false,
+ "attention_dropout": 0.0,
+ "bos_token_id": null,
+ "dtype": "float32",
+ "eos_token_id": 151645,
+ "head_dim": 128,
+ "hidden_act": "silu",
+ "hidden_size": 512,
+ "initializer_range": 0.02,
+ "intermediate_size": 1536,
+ "max_position_embeddings": 2048,
+ "mlp_bias": false,
+ "model_type": "llama",
+ "num_attention_heads": 4,
+ "num_hidden_layers": 20,
+ "num_key_value_heads": 4,
+ "pad_token_id": 151645,
+ "pretraining_tp": 1,
+ "rms_norm_eps": 1e-06,
+ "rope_parameters": {
+ "rope_theta": 10000.0,
+ "rope_type": "default"
+ },
+ "tie_word_embeddings": true,
+ "transformers_version": "5.5.0",
+ "use_cache": false,
+ "vocab_size": 151673
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4320/generation_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-4320/generation_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-4320/generation_config.json
@@ -0,0 +1,11 @@
+{
+ "_from_model_config": true,
+ "eos_token_id": [
+ 151645
+ ],
+ "output_attentions": false,
+ "output_hidden_states": false,
+ "pad_token_id": 151645,
+ "transformers_version": "5.5.0",
+ "use_cache": true
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4320/model.safetensors b/runs/l2r50-i2-fulle-lm/checkpoint-4320/model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..81ef3fb2d97e60fe0e421a3ef81f0e91b4cb0c09
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-4320/model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:d95648a6b9afbdcbd09ea254944ed25587d460ffb84366de3d076e2721ed01f2
+size 583360328
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4320/optimizer.pt b/runs/l2r50-i2-fulle-lm/checkpoint-4320/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..7d2d5d07f4e70ded0e545b44c4e704baad78f9d8
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-4320/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:3bedc1e957284cd8f730f74e70592d488d054ec89f70926baad9378c399c3cb7
+size 1166833530
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4320/rng_state_0.pth b/runs/l2r50-i2-fulle-lm/checkpoint-4320/rng_state_0.pth
new file mode 100644
index 0000000000000000000000000000000000000000..857a2f67a20363a76b04e398c4b43929a2876378
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-4320/rng_state_0.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:7b9c43b8c954569c82e76a9f2d993225b6b7dbf6b329599acfe5545ac32c8f22
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4320/rng_state_1.pth b/runs/l2r50-i2-fulle-lm/checkpoint-4320/rng_state_1.pth
new file mode 100644
index 0000000000000000000000000000000000000000..d69ea8592d9d7a0ef2bbaa6e8f106b074b9c90f6
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-4320/rng_state_1.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:26c963656cfadaeec7edb40d9c8f021310d2d6343fbbbe402907b79bfb5ada91
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4320/rng_state_2.pth b/runs/l2r50-i2-fulle-lm/checkpoint-4320/rng_state_2.pth
new file mode 100644
index 0000000000000000000000000000000000000000..61a4d72e3ecceb047507833e7b22210803abc58e
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-4320/rng_state_2.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:6f58ab0058bd189237a3838eb734c7827116ce23174e1b3a6a5c22a152fad5b3
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4320/rng_state_3.pth b/runs/l2r50-i2-fulle-lm/checkpoint-4320/rng_state_3.pth
new file mode 100644
index 0000000000000000000000000000000000000000..2302ff29e4b6b17d0e81ad6016f44d9c82e028e9
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-4320/rng_state_3.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:12cbf2c759fd06b8222f1c95276e256443f6c608d42df51778c153c282006dac
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4320/scheduler.pt b/runs/l2r50-i2-fulle-lm/checkpoint-4320/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..8304bf8d1ccccae077ecc7e6e66bba8073475ac5
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-4320/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:9c504cbc081ac24362a030317097aa4f7d2e9075cfc12b05dc64f4ee58e8558c
+size 1064
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4320/tokenizer.json b/runs/l2r50-i2-fulle-lm/checkpoint-4320/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..53786016f070e11184813f3c83101dc5e93e392e
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-4320/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:c3c63fda1646a555448e8f03662542e04b11d103721319e4d9b267d7da03ceb8
+size 11424463
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4320/tokenizer_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-4320/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..98e2fb9b2cc54785dc9d871b8e503f8b8f2defe7
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-4320/tokenizer_config.json
@@ -0,0 +1,20 @@
+{
+ "add_prefix_space": false,
+ "backend": "tokenizers",
+ "bos_token": null,
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "<|im_end|>",
+ "errors": "replace",
+ "extra_special_tokens": [
+ "<|l2r_pred|>",
+ "<|r2l_pred|>",
+ "<|next_1_pred|>",
+ "<|next_2_pred|>"
+ ],
+ "is_local": false,
+ "model_max_length": 131072,
+ "pad_token": "<|im_end|>",
+ "split_special_tokens": false,
+ "tokenizer_class": "Qwen2Tokenizer",
+ "unk_token": null
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4320/trainer_state.json b/runs/l2r50-i2-fulle-lm/checkpoint-4320/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..0d1ae75e409b2a7f2fdecaeb6a16c540c9d2987b
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-4320/trainer_state.json
@@ -0,0 +1,30746 @@
+{
+ "best_global_step": null,
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 60.0,
+ "eval_steps": 500,
+ "global_step": 4320,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.013973799126637555,
+ "grad_norm": 0.1739506870508194,
+ "learning_rate": 0.0,
+ "loss": 11.99299430847168,
+ "step": 1
+ },
+ {
+ "epoch": 0.02794759825327511,
+ "grad_norm": 0.17216597497463226,
+ "learning_rate": 5.999999999999999e-06,
+ "loss": 11.988704681396484,
+ "step": 2
+ },
+ {
+ "epoch": 0.04192139737991266,
+ "grad_norm": 0.17459715902805328,
+ "learning_rate": 1.1999999999999999e-05,
+ "loss": 11.945581436157227,
+ "step": 3
+ },
+ {
+ "epoch": 0.05589519650655022,
+ "grad_norm": 0.17577414214611053,
+ "learning_rate": 1.7999999999999997e-05,
+ "loss": 11.859772682189941,
+ "step": 4
+ },
+ {
+ "epoch": 0.06986899563318777,
+ "grad_norm": 0.16163651645183563,
+ "learning_rate": 2.3999999999999997e-05,
+ "loss": 11.744415283203125,
+ "step": 5
+ },
+ {
+ "epoch": 0.08384279475982533,
+ "grad_norm": 0.14409932494163513,
+ "learning_rate": 2.9999999999999997e-05,
+ "loss": 11.618099212646484,
+ "step": 6
+ },
+ {
+ "epoch": 0.09781659388646288,
+ "grad_norm": 0.13079848885536194,
+ "learning_rate": 3.5999999999999994e-05,
+ "loss": 11.505935668945312,
+ "step": 7
+ },
+ {
+ "epoch": 0.11179039301310044,
+ "grad_norm": 0.11964613944292068,
+ "learning_rate": 4.2e-05,
+ "loss": 11.416532516479492,
+ "step": 8
+ },
+ {
+ "epoch": 0.125764192139738,
+ "grad_norm": 0.11345142871141434,
+ "learning_rate": 4.7999999999999994e-05,
+ "loss": 11.33206558227539,
+ "step": 9
+ },
+ {
+ "epoch": 0.13973799126637554,
+ "grad_norm": 0.11033952981233597,
+ "learning_rate": 5.399999999999999e-05,
+ "loss": 11.259052276611328,
+ "step": 10
+ },
+ {
+ "epoch": 0.1537117903930131,
+ "grad_norm": 0.10776489973068237,
+ "learning_rate": 5.9999999999999995e-05,
+ "loss": 11.203384399414062,
+ "step": 11
+ },
+ {
+ "epoch": 0.16768558951965065,
+ "grad_norm": 0.10721946507692337,
+ "learning_rate": 6.599999999999999e-05,
+ "loss": 11.153295516967773,
+ "step": 12
+ },
+ {
+ "epoch": 0.18165938864628822,
+ "grad_norm": 0.10673430562019348,
+ "learning_rate": 7.199999999999999e-05,
+ "loss": 11.110200881958008,
+ "step": 13
+ },
+ {
+ "epoch": 0.19563318777292577,
+ "grad_norm": 0.10630254447460175,
+ "learning_rate": 7.8e-05,
+ "loss": 11.067477226257324,
+ "step": 14
+ },
+ {
+ "epoch": 0.2096069868995633,
+ "grad_norm": 0.10631615668535233,
+ "learning_rate": 8.4e-05,
+ "loss": 11.022941589355469,
+ "step": 15
+ },
+ {
+ "epoch": 0.22358078602620088,
+ "grad_norm": 0.10609443485736847,
+ "learning_rate": 8.999999999999999e-05,
+ "loss": 10.970746040344238,
+ "step": 16
+ },
+ {
+ "epoch": 0.23755458515283842,
+ "grad_norm": 0.10648232698440552,
+ "learning_rate": 9.599999999999999e-05,
+ "loss": 10.91267204284668,
+ "step": 17
+ },
+ {
+ "epoch": 0.251528384279476,
+ "grad_norm": 0.10584589838981628,
+ "learning_rate": 0.000102,
+ "loss": 10.854547500610352,
+ "step": 18
+ },
+ {
+ "epoch": 0.26550218340611353,
+ "grad_norm": 0.10591720044612885,
+ "learning_rate": 0.00010799999999999998,
+ "loss": 10.788917541503906,
+ "step": 19
+ },
+ {
+ "epoch": 0.2794759825327511,
+ "grad_norm": 0.10611672699451447,
+ "learning_rate": 0.00011399999999999999,
+ "loss": 10.717556953430176,
+ "step": 20
+ },
+ {
+ "epoch": 0.2934497816593886,
+ "grad_norm": 0.10455359518527985,
+ "learning_rate": 0.00011999999999999999,
+ "loss": 10.650186538696289,
+ "step": 21
+ },
+ {
+ "epoch": 0.3074235807860262,
+ "grad_norm": 0.1047525629401207,
+ "learning_rate": 0.00012599999999999997,
+ "loss": 10.56808090209961,
+ "step": 22
+ },
+ {
+ "epoch": 0.32139737991266376,
+ "grad_norm": 0.10511605441570282,
+ "learning_rate": 0.00013199999999999998,
+ "loss": 10.484028816223145,
+ "step": 23
+ },
+ {
+ "epoch": 0.3353711790393013,
+ "grad_norm": 0.10467177629470825,
+ "learning_rate": 0.000138,
+ "loss": 10.399947166442871,
+ "step": 24
+ },
+ {
+ "epoch": 0.34934497816593885,
+ "grad_norm": 0.10432136058807373,
+ "learning_rate": 0.00014399999999999998,
+ "loss": 10.312714576721191,
+ "step": 25
+ },
+ {
+ "epoch": 0.36331877729257644,
+ "grad_norm": 0.10527262091636658,
+ "learning_rate": 0.00015,
+ "loss": 10.212261199951172,
+ "step": 26
+ },
+ {
+ "epoch": 0.377292576419214,
+ "grad_norm": 0.103610560297966,
+ "learning_rate": 0.000156,
+ "loss": 10.132034301757812,
+ "step": 27
+ },
+ {
+ "epoch": 0.39126637554585153,
+ "grad_norm": 0.10553059726953506,
+ "learning_rate": 0.000162,
+ "loss": 10.020434379577637,
+ "step": 28
+ },
+ {
+ "epoch": 0.4052401746724891,
+ "grad_norm": 0.1037299782037735,
+ "learning_rate": 0.000168,
+ "loss": 9.934466361999512,
+ "step": 29
+ },
+ {
+ "epoch": 0.4192139737991266,
+ "grad_norm": 0.1042739674448967,
+ "learning_rate": 0.00017399999999999997,
+ "loss": 9.822539329528809,
+ "step": 30
+ },
+ {
+ "epoch": 0.4331877729257642,
+ "grad_norm": 0.10300865024328232,
+ "learning_rate": 0.00017999999999999998,
+ "loss": 9.734994888305664,
+ "step": 31
+ },
+ {
+ "epoch": 0.44716157205240176,
+ "grad_norm": 0.10291654616594315,
+ "learning_rate": 0.000186,
+ "loss": 9.624505996704102,
+ "step": 32
+ },
+ {
+ "epoch": 0.4611353711790393,
+ "grad_norm": 0.10210851579904556,
+ "learning_rate": 0.00019199999999999998,
+ "loss": 9.518205642700195,
+ "step": 33
+ },
+ {
+ "epoch": 0.47510917030567684,
+ "grad_norm": 0.10222529619932175,
+ "learning_rate": 0.000198,
+ "loss": 9.412222862243652,
+ "step": 34
+ },
+ {
+ "epoch": 0.4890829694323144,
+ "grad_norm": 0.09992019087076187,
+ "learning_rate": 0.000204,
+ "loss": 9.326566696166992,
+ "step": 35
+ },
+ {
+ "epoch": 0.503056768558952,
+ "grad_norm": 0.09931084513664246,
+ "learning_rate": 0.00020999999999999998,
+ "loss": 9.210111618041992,
+ "step": 36
+ },
+ {
+ "epoch": 0.5170305676855895,
+ "grad_norm": 0.09928256273269653,
+ "learning_rate": 0.00021599999999999996,
+ "loss": 9.100493431091309,
+ "step": 37
+ },
+ {
+ "epoch": 0.5310043668122271,
+ "grad_norm": 0.09843816608190536,
+ "learning_rate": 0.00022199999999999998,
+ "loss": 8.996543884277344,
+ "step": 38
+ },
+ {
+ "epoch": 0.5449781659388646,
+ "grad_norm": 0.09825374186038971,
+ "learning_rate": 0.00022799999999999999,
+ "loss": 8.888991355895996,
+ "step": 39
+ },
+ {
+ "epoch": 0.5589519650655022,
+ "grad_norm": 0.09602385759353638,
+ "learning_rate": 0.000234,
+ "loss": 8.802907943725586,
+ "step": 40
+ },
+ {
+ "epoch": 0.5729257641921397,
+ "grad_norm": 0.09414064139127731,
+ "learning_rate": 0.00023999999999999998,
+ "loss": 8.715423583984375,
+ "step": 41
+ },
+ {
+ "epoch": 0.5868995633187772,
+ "grad_norm": 0.09273893386125565,
+ "learning_rate": 0.00024599999999999996,
+ "loss": 8.61374282836914,
+ "step": 42
+ },
+ {
+ "epoch": 0.6008733624454149,
+ "grad_norm": 0.09079599380493164,
+ "learning_rate": 0.00025199999999999995,
+ "loss": 8.502904891967773,
+ "step": 43
+ },
+ {
+ "epoch": 0.6148471615720524,
+ "grad_norm": 0.08605825901031494,
+ "learning_rate": 0.000258,
+ "loss": 8.453008651733398,
+ "step": 44
+ },
+ {
+ "epoch": 0.62882096069869,
+ "grad_norm": 0.08338996022939682,
+ "learning_rate": 0.00026399999999999997,
+ "loss": 8.365163803100586,
+ "step": 45
+ },
+ {
+ "epoch": 0.6427947598253275,
+ "grad_norm": 0.08150867372751236,
+ "learning_rate": 0.00027,
+ "loss": 8.285372734069824,
+ "step": 46
+ },
+ {
+ "epoch": 0.6567685589519651,
+ "grad_norm": 0.07794807851314545,
+ "learning_rate": 0.000276,
+ "loss": 8.185811996459961,
+ "step": 47
+ },
+ {
+ "epoch": 0.6707423580786026,
+ "grad_norm": 0.07592720538377762,
+ "learning_rate": 0.00028199999999999997,
+ "loss": 8.098482131958008,
+ "step": 48
+ },
+ {
+ "epoch": 0.6847161572052402,
+ "grad_norm": 0.0709897130727768,
+ "learning_rate": 0.00028799999999999995,
+ "loss": 8.042091369628906,
+ "step": 49
+ },
+ {
+ "epoch": 0.6986899563318777,
+ "grad_norm": 0.06582305580377579,
+ "learning_rate": 0.000294,
+ "loss": 7.972844123840332,
+ "step": 50
+ },
+ {
+ "epoch": 0.7126637554585152,
+ "grad_norm": 0.059452351182699203,
+ "learning_rate": 0.0003,
+ "loss": 7.924579620361328,
+ "step": 51
+ },
+ {
+ "epoch": 0.7266375545851529,
+ "grad_norm": 0.052426718175411224,
+ "learning_rate": 0.00030599999999999996,
+ "loss": 7.876336097717285,
+ "step": 52
+ },
+ {
+ "epoch": 0.7406113537117904,
+ "grad_norm": 0.04768150672316551,
+ "learning_rate": 0.000312,
+ "loss": 7.8102707862854,
+ "step": 53
+ },
+ {
+ "epoch": 0.754585152838428,
+ "grad_norm": 0.03970075398683548,
+ "learning_rate": 0.000318,
+ "loss": 7.791913986206055,
+ "step": 54
+ },
+ {
+ "epoch": 0.7685589519650655,
+ "grad_norm": 0.03559752553701401,
+ "learning_rate": 0.000324,
+ "loss": 7.756960868835449,
+ "step": 55
+ },
+ {
+ "epoch": 0.7825327510917031,
+ "grad_norm": 0.029321353882551193,
+ "learning_rate": 0.00033,
+ "loss": 7.724878787994385,
+ "step": 56
+ },
+ {
+ "epoch": 0.7965065502183406,
+ "grad_norm": 0.023485427722334862,
+ "learning_rate": 0.000336,
+ "loss": 7.72553014755249,
+ "step": 57
+ },
+ {
+ "epoch": 0.8104803493449781,
+ "grad_norm": 0.019319474697113037,
+ "learning_rate": 0.00034199999999999996,
+ "loss": 7.709493160247803,
+ "step": 58
+ },
+ {
+ "epoch": 0.8244541484716157,
+ "grad_norm": 0.016705691814422607,
+ "learning_rate": 0.00034799999999999995,
+ "loss": 7.6873297691345215,
+ "step": 59
+ },
+ {
+ "epoch": 0.8384279475982532,
+ "grad_norm": 0.017667734995484352,
+ "learning_rate": 0.00035399999999999993,
+ "loss": 7.677059173583984,
+ "step": 60
+ },
+ {
+ "epoch": 0.8524017467248908,
+ "grad_norm": 0.01671992614865303,
+ "learning_rate": 0.00035999999999999997,
+ "loss": 7.643512725830078,
+ "step": 61
+ },
+ {
+ "epoch": 0.8663755458515284,
+ "grad_norm": 0.012288172729313374,
+ "learning_rate": 0.00036599999999999995,
+ "loss": 7.658343315124512,
+ "step": 62
+ },
+ {
+ "epoch": 0.880349344978166,
+ "grad_norm": 0.012004776857793331,
+ "learning_rate": 0.000372,
+ "loss": 7.649759292602539,
+ "step": 63
+ },
+ {
+ "epoch": 0.8943231441048035,
+ "grad_norm": 0.012487445957958698,
+ "learning_rate": 0.00037799999999999997,
+ "loss": 7.631503105163574,
+ "step": 64
+ },
+ {
+ "epoch": 0.9082969432314411,
+ "grad_norm": 0.01489241048693657,
+ "learning_rate": 0.00038399999999999996,
+ "loss": 7.647110462188721,
+ "step": 65
+ },
+ {
+ "epoch": 0.9222707423580786,
+ "grad_norm": 0.01428980939090252,
+ "learning_rate": 0.00039,
+ "loss": 7.636983871459961,
+ "step": 66
+ },
+ {
+ "epoch": 0.9362445414847161,
+ "grad_norm": 0.010711144655942917,
+ "learning_rate": 0.000396,
+ "loss": 7.621130466461182,
+ "step": 67
+ },
+ {
+ "epoch": 0.9502183406113537,
+ "grad_norm": 0.010178886353969574,
+ "learning_rate": 0.000402,
+ "loss": 7.593924522399902,
+ "step": 68
+ },
+ {
+ "epoch": 0.9641921397379912,
+ "grad_norm": 0.010604874230921268,
+ "learning_rate": 0.000408,
+ "loss": 7.6015496253967285,
+ "step": 69
+ },
+ {
+ "epoch": 0.9781659388646288,
+ "grad_norm": 0.009751426056027412,
+ "learning_rate": 0.0004139999999999999,
+ "loss": 7.571964263916016,
+ "step": 70
+ },
+ {
+ "epoch": 0.9921397379912664,
+ "grad_norm": 0.009816481731832027,
+ "learning_rate": 0.00041999999999999996,
+ "loss": 7.57344913482666,
+ "step": 71
+ },
+ {
+ "epoch": 1.0,
+ "grad_norm": 0.008665637113153934,
+ "learning_rate": 0.00042599999999999995,
+ "loss": 7.579050064086914,
+ "step": 72
+ },
+ {
+ "epoch": 1.0,
+ "eval_loss": 7.581204891204834,
+ "eval_runtime": 59.0091,
+ "eval_samples_per_second": 41.383,
+ "eval_steps_per_second": 1.305,
+ "step": 72
+ },
+ {
+ "epoch": 1.0139737991266375,
+ "grad_norm": 0.009684201329946518,
+ "learning_rate": 0.00043199999999999993,
+ "loss": 7.568594932556152,
+ "step": 73
+ },
+ {
+ "epoch": 1.027947598253275,
+ "grad_norm": 0.025783710181713104,
+ "learning_rate": 0.00043799999999999997,
+ "loss": 7.578948497772217,
+ "step": 74
+ },
+ {
+ "epoch": 1.0419213973799126,
+ "grad_norm": 0.0327819399535656,
+ "learning_rate": 0.00044399999999999995,
+ "loss": 7.548173904418945,
+ "step": 75
+ },
+ {
+ "epoch": 1.0558951965065502,
+ "grad_norm": 0.021604498848319054,
+ "learning_rate": 0.00045,
+ "loss": 7.552953720092773,
+ "step": 76
+ },
+ {
+ "epoch": 1.0698689956331877,
+ "grad_norm": 0.010230105370283127,
+ "learning_rate": 0.00045599999999999997,
+ "loss": 7.5213212966918945,
+ "step": 77
+ },
+ {
+ "epoch": 1.0838427947598253,
+ "grad_norm": 0.023315589874982834,
+ "learning_rate": 0.00046199999999999995,
+ "loss": 7.501640319824219,
+ "step": 78
+ },
+ {
+ "epoch": 1.0978165938864628,
+ "grad_norm": 0.01109944935888052,
+ "learning_rate": 0.000468,
+ "loss": 7.539519786834717,
+ "step": 79
+ },
+ {
+ "epoch": 1.1117903930131003,
+ "grad_norm": 0.026093758642673492,
+ "learning_rate": 0.000474,
+ "loss": 7.525928974151611,
+ "step": 80
+ },
+ {
+ "epoch": 1.125764192139738,
+ "grad_norm": 0.03274895250797272,
+ "learning_rate": 0.00047999999999999996,
+ "loss": 7.535602569580078,
+ "step": 81
+ },
+ {
+ "epoch": 1.1397379912663754,
+ "grad_norm": 0.01543602254241705,
+ "learning_rate": 0.000486,
+ "loss": 7.513683795928955,
+ "step": 82
+ },
+ {
+ "epoch": 1.1537117903930132,
+ "grad_norm": 0.028377249836921692,
+ "learning_rate": 0.0004919999999999999,
+ "loss": 7.49384880065918,
+ "step": 83
+ },
+ {
+ "epoch": 1.1676855895196507,
+ "grad_norm": 0.04121481627225876,
+ "learning_rate": 0.000498,
+ "loss": 7.488447189331055,
+ "step": 84
+ },
+ {
+ "epoch": 1.1816593886462883,
+ "grad_norm": 0.013282302767038345,
+ "learning_rate": 0.0005039999999999999,
+ "loss": 7.463010787963867,
+ "step": 85
+ },
+ {
+ "epoch": 1.1956331877729258,
+ "grad_norm": 0.03200364485383034,
+ "learning_rate": 0.0005099999999999999,
+ "loss": 7.473515033721924,
+ "step": 86
+ },
+ {
+ "epoch": 1.2096069868995634,
+ "grad_norm": 0.020768029615283012,
+ "learning_rate": 0.000516,
+ "loss": 7.464688777923584,
+ "step": 87
+ },
+ {
+ "epoch": 1.223580786026201,
+ "grad_norm": 0.021715788170695305,
+ "learning_rate": 0.000522,
+ "loss": 7.4835028648376465,
+ "step": 88
+ },
+ {
+ "epoch": 1.2375545851528384,
+ "grad_norm": 0.026255754753947258,
+ "learning_rate": 0.0005279999999999999,
+ "loss": 7.443392753601074,
+ "step": 89
+ },
+ {
+ "epoch": 1.251528384279476,
+ "grad_norm": 0.022325951606035233,
+ "learning_rate": 0.000534,
+ "loss": 7.457373142242432,
+ "step": 90
+ },
+ {
+ "epoch": 1.2655021834061135,
+ "grad_norm": 0.009270387701690197,
+ "learning_rate": 0.00054,
+ "loss": 7.455986976623535,
+ "step": 91
+ },
+ {
+ "epoch": 1.279475982532751,
+ "grad_norm": 0.02369544468820095,
+ "learning_rate": 0.0005459999999999999,
+ "loss": 7.442216873168945,
+ "step": 92
+ },
+ {
+ "epoch": 1.2934497816593886,
+ "grad_norm": 0.008474506437778473,
+ "learning_rate": 0.000552,
+ "loss": 7.4294915199279785,
+ "step": 93
+ },
+ {
+ "epoch": 1.3074235807860262,
+ "grad_norm": 0.017043249681591988,
+ "learning_rate": 0.000558,
+ "loss": 7.406932830810547,
+ "step": 94
+ },
+ {
+ "epoch": 1.3213973799126637,
+ "grad_norm": 0.011813412420451641,
+ "learning_rate": 0.0005639999999999999,
+ "loss": 7.415900230407715,
+ "step": 95
+ },
+ {
+ "epoch": 1.3353711790393012,
+ "grad_norm": 0.01086883433163166,
+ "learning_rate": 0.00057,
+ "loss": 7.452998638153076,
+ "step": 96
+ },
+ {
+ "epoch": 1.3493449781659388,
+ "grad_norm": 0.014700951054692268,
+ "learning_rate": 0.0005759999999999999,
+ "loss": 7.410249710083008,
+ "step": 97
+ },
+ {
+ "epoch": 1.3633187772925766,
+ "grad_norm": 0.014512976631522179,
+ "learning_rate": 0.0005819999999999999,
+ "loss": 7.429132461547852,
+ "step": 98
+ },
+ {
+ "epoch": 1.3772925764192139,
+ "grad_norm": 0.012519920244812965,
+ "learning_rate": 0.000588,
+ "loss": 7.430817604064941,
+ "step": 99
+ },
+ {
+ "epoch": 1.3912663755458516,
+ "grad_norm": 0.007132664322853088,
+ "learning_rate": 0.0005939999999999999,
+ "loss": 7.405234336853027,
+ "step": 100
+ },
+ {
+ "epoch": 1.405240174672489,
+ "grad_norm": 0.019489388912916183,
+ "learning_rate": 0.0006,
+ "loss": 7.427791595458984,
+ "step": 101
+ },
+ {
+ "epoch": 1.4192139737991267,
+ "grad_norm": 0.027252664789557457,
+ "learning_rate": 0.0006,
+ "loss": 7.41761589050293,
+ "step": 102
+ },
+ {
+ "epoch": 1.4331877729257643,
+ "grad_norm": 0.02501637488603592,
+ "learning_rate": 0.0006,
+ "loss": 7.416735649108887,
+ "step": 103
+ },
+ {
+ "epoch": 1.4471615720524018,
+ "grad_norm": 0.00989772193133831,
+ "learning_rate": 0.0006,
+ "loss": 7.3756632804870605,
+ "step": 104
+ },
+ {
+ "epoch": 1.4611353711790394,
+ "grad_norm": 0.01622949168086052,
+ "learning_rate": 0.0006,
+ "loss": 7.400811672210693,
+ "step": 105
+ },
+ {
+ "epoch": 1.475109170305677,
+ "grad_norm": 0.012407763861119747,
+ "learning_rate": 0.0006,
+ "loss": 7.37675666809082,
+ "step": 106
+ },
+ {
+ "epoch": 1.4890829694323144,
+ "grad_norm": 0.010353565216064453,
+ "learning_rate": 0.0006,
+ "loss": 7.4013142585754395,
+ "step": 107
+ },
+ {
+ "epoch": 1.503056768558952,
+ "grad_norm": 0.0120553532615304,
+ "learning_rate": 0.0006,
+ "loss": 7.379873275756836,
+ "step": 108
+ },
+ {
+ "epoch": 1.5170305676855895,
+ "grad_norm": 0.01610225811600685,
+ "learning_rate": 0.0006,
+ "loss": 7.371631145477295,
+ "step": 109
+ },
+ {
+ "epoch": 1.531004366812227,
+ "grad_norm": 0.01159879844635725,
+ "learning_rate": 0.0006,
+ "loss": 7.389007568359375,
+ "step": 110
+ },
+ {
+ "epoch": 1.5449781659388646,
+ "grad_norm": 0.008013816550374031,
+ "learning_rate": 0.0006,
+ "loss": 7.395920753479004,
+ "step": 111
+ },
+ {
+ "epoch": 1.5589519650655022,
+ "grad_norm": 0.016334809362888336,
+ "learning_rate": 0.0006,
+ "loss": 7.393121719360352,
+ "step": 112
+ },
+ {
+ "epoch": 1.5729257641921397,
+ "grad_norm": 0.018821721896529198,
+ "learning_rate": 0.0006,
+ "loss": 7.367543697357178,
+ "step": 113
+ },
+ {
+ "epoch": 1.5868995633187772,
+ "grad_norm": 0.014088994823396206,
+ "learning_rate": 0.0006,
+ "loss": 7.385537624359131,
+ "step": 114
+ },
+ {
+ "epoch": 1.600873362445415,
+ "grad_norm": 0.005864050704985857,
+ "learning_rate": 0.0006,
+ "loss": 7.365346908569336,
+ "step": 115
+ },
+ {
+ "epoch": 1.6148471615720523,
+ "grad_norm": 0.011745430529117584,
+ "learning_rate": 0.0006,
+ "loss": 7.390227317810059,
+ "step": 116
+ },
+ {
+ "epoch": 1.62882096069869,
+ "grad_norm": 0.017205573618412018,
+ "learning_rate": 0.0006,
+ "loss": 7.3558549880981445,
+ "step": 117
+ },
+ {
+ "epoch": 1.6427947598253274,
+ "grad_norm": 0.021256346255540848,
+ "learning_rate": 0.0006,
+ "loss": 7.377346992492676,
+ "step": 118
+ },
+ {
+ "epoch": 1.6567685589519652,
+ "grad_norm": 0.019335204735398293,
+ "learning_rate": 0.0006,
+ "loss": 7.346194267272949,
+ "step": 119
+ },
+ {
+ "epoch": 1.6707423580786025,
+ "grad_norm": 0.008134003728628159,
+ "learning_rate": 0.0006,
+ "loss": 7.3063578605651855,
+ "step": 120
+ },
+ {
+ "epoch": 1.6847161572052403,
+ "grad_norm": 0.017613831907510757,
+ "learning_rate": 0.0006,
+ "loss": 7.338171005249023,
+ "step": 121
+ },
+ {
+ "epoch": 1.6986899563318776,
+ "grad_norm": 0.00890734139829874,
+ "learning_rate": 0.0006,
+ "loss": 7.355384826660156,
+ "step": 122
+ },
+ {
+ "epoch": 1.7126637554585153,
+ "grad_norm": 0.013448155485093594,
+ "learning_rate": 0.0006,
+ "loss": 7.344296455383301,
+ "step": 123
+ },
+ {
+ "epoch": 1.726637554585153,
+ "grad_norm": 0.009241633117198944,
+ "learning_rate": 0.0006,
+ "loss": 7.3454060554504395,
+ "step": 124
+ },
+ {
+ "epoch": 1.7406113537117904,
+ "grad_norm": 0.018561311066150665,
+ "learning_rate": 0.0006,
+ "loss": 7.32348108291626,
+ "step": 125
+ },
+ {
+ "epoch": 1.754585152838428,
+ "grad_norm": 0.025667704641819,
+ "learning_rate": 0.0006,
+ "loss": 7.356142520904541,
+ "step": 126
+ },
+ {
+ "epoch": 1.7685589519650655,
+ "grad_norm": 0.036232151091098785,
+ "learning_rate": 0.0006,
+ "loss": 7.363539695739746,
+ "step": 127
+ },
+ {
+ "epoch": 1.782532751091703,
+ "grad_norm": 0.059483714401721954,
+ "learning_rate": 0.0006,
+ "loss": 7.402331352233887,
+ "step": 128
+ },
+ {
+ "epoch": 1.7965065502183406,
+ "grad_norm": 0.05365009233355522,
+ "learning_rate": 0.0006,
+ "loss": 7.40457820892334,
+ "step": 129
+ },
+ {
+ "epoch": 1.8104803493449781,
+ "grad_norm": 0.02738415263593197,
+ "learning_rate": 0.0006,
+ "loss": 7.3756208419799805,
+ "step": 130
+ },
+ {
+ "epoch": 1.8244541484716157,
+ "grad_norm": 0.035017091780900955,
+ "learning_rate": 0.0006,
+ "loss": 7.379263877868652,
+ "step": 131
+ },
+ {
+ "epoch": 1.8384279475982532,
+ "grad_norm": 0.02819863148033619,
+ "learning_rate": 0.0006,
+ "loss": 7.327673435211182,
+ "step": 132
+ },
+ {
+ "epoch": 1.8524017467248908,
+ "grad_norm": 0.017754238098859787,
+ "learning_rate": 0.0006,
+ "loss": 7.333420276641846,
+ "step": 133
+ },
+ {
+ "epoch": 1.8663755458515285,
+ "grad_norm": 0.024685733020305634,
+ "learning_rate": 0.0006,
+ "loss": 7.34493350982666,
+ "step": 134
+ },
+ {
+ "epoch": 1.8803493449781659,
+ "grad_norm": 0.019426532089710236,
+ "learning_rate": 0.0006,
+ "loss": 7.328448295593262,
+ "step": 135
+ },
+ {
+ "epoch": 1.8943231441048036,
+ "grad_norm": 0.02106391079723835,
+ "learning_rate": 0.0006,
+ "loss": 7.327917098999023,
+ "step": 136
+ },
+ {
+ "epoch": 1.908296943231441,
+ "grad_norm": 0.02056879736483097,
+ "learning_rate": 0.0006,
+ "loss": 7.3311991691589355,
+ "step": 137
+ },
+ {
+ "epoch": 1.9222707423580787,
+ "grad_norm": 0.016865141689777374,
+ "learning_rate": 0.0006,
+ "loss": 7.312766075134277,
+ "step": 138
+ },
+ {
+ "epoch": 1.936244541484716,
+ "grad_norm": 0.017549855634570122,
+ "learning_rate": 0.0006,
+ "loss": 7.31191349029541,
+ "step": 139
+ },
+ {
+ "epoch": 1.9502183406113538,
+ "grad_norm": 0.015930157154798508,
+ "learning_rate": 0.0006,
+ "loss": 7.330901622772217,
+ "step": 140
+ },
+ {
+ "epoch": 1.9641921397379911,
+ "grad_norm": 0.013283558189868927,
+ "learning_rate": 0.0006,
+ "loss": 7.298318386077881,
+ "step": 141
+ },
+ {
+ "epoch": 1.9781659388646289,
+ "grad_norm": 0.012023469433188438,
+ "learning_rate": 0.0006,
+ "loss": 7.282337188720703,
+ "step": 142
+ },
+ {
+ "epoch": 1.9921397379912664,
+ "grad_norm": 0.01459511648863554,
+ "learning_rate": 0.0006,
+ "loss": 7.282924175262451,
+ "step": 143
+ },
+ {
+ "epoch": 2.0,
+ "grad_norm": 0.017474235966801643,
+ "learning_rate": 0.0006,
+ "loss": 7.254333972930908,
+ "step": 144
+ },
+ {
+ "epoch": 2.0,
+ "eval_loss": 7.318708896636963,
+ "eval_runtime": 59.6555,
+ "eval_samples_per_second": 40.935,
+ "eval_steps_per_second": 1.291,
+ "step": 144
+ },
+ {
+ "epoch": 2.0139737991266378,
+ "grad_norm": 0.019757507368922234,
+ "learning_rate": 0.0006,
+ "loss": 7.30288553237915,
+ "step": 145
+ },
+ {
+ "epoch": 2.027947598253275,
+ "grad_norm": 0.014571248553693295,
+ "learning_rate": 0.0006,
+ "loss": 7.274681568145752,
+ "step": 146
+ },
+ {
+ "epoch": 2.041921397379913,
+ "grad_norm": 0.015744198113679886,
+ "learning_rate": 0.0006,
+ "loss": 7.278632640838623,
+ "step": 147
+ },
+ {
+ "epoch": 2.05589519650655,
+ "grad_norm": 0.028048858046531677,
+ "learning_rate": 0.0006,
+ "loss": 7.26558780670166,
+ "step": 148
+ },
+ {
+ "epoch": 2.069868995633188,
+ "grad_norm": 0.06547307223081589,
+ "learning_rate": 0.0006,
+ "loss": 7.32567834854126,
+ "step": 149
+ },
+ {
+ "epoch": 2.0838427947598253,
+ "grad_norm": 0.05969763547182083,
+ "learning_rate": 0.0006,
+ "loss": 7.395836353302002,
+ "step": 150
+ },
+ {
+ "epoch": 2.097816593886463,
+ "grad_norm": 0.032574281096458435,
+ "learning_rate": 0.0006,
+ "loss": 7.318150997161865,
+ "step": 151
+ },
+ {
+ "epoch": 2.1117903930131003,
+ "grad_norm": 0.029452962800860405,
+ "learning_rate": 0.0006,
+ "loss": 7.308066368103027,
+ "step": 152
+ },
+ {
+ "epoch": 2.125764192139738,
+ "grad_norm": 0.03161991387605667,
+ "learning_rate": 0.0006,
+ "loss": 7.306290626525879,
+ "step": 153
+ },
+ {
+ "epoch": 2.1397379912663754,
+ "grad_norm": 0.027096295729279518,
+ "learning_rate": 0.0006,
+ "loss": 7.297796249389648,
+ "step": 154
+ },
+ {
+ "epoch": 2.153711790393013,
+ "grad_norm": 0.028657056391239166,
+ "learning_rate": 0.0006,
+ "loss": 7.317424774169922,
+ "step": 155
+ },
+ {
+ "epoch": 2.1676855895196505,
+ "grad_norm": 0.02046995982527733,
+ "learning_rate": 0.0006,
+ "loss": 7.297986030578613,
+ "step": 156
+ },
+ {
+ "epoch": 2.1816593886462883,
+ "grad_norm": 0.02220369316637516,
+ "learning_rate": 0.0006,
+ "loss": 7.268237590789795,
+ "step": 157
+ },
+ {
+ "epoch": 2.1956331877729256,
+ "grad_norm": 0.018381357192993164,
+ "learning_rate": 0.0006,
+ "loss": 7.259415626525879,
+ "step": 158
+ },
+ {
+ "epoch": 2.2096069868995634,
+ "grad_norm": 0.0204413291066885,
+ "learning_rate": 0.0006,
+ "loss": 7.258173942565918,
+ "step": 159
+ },
+ {
+ "epoch": 2.2235807860262007,
+ "grad_norm": 0.015874288976192474,
+ "learning_rate": 0.0006,
+ "loss": 7.281834602355957,
+ "step": 160
+ },
+ {
+ "epoch": 2.2375545851528384,
+ "grad_norm": 0.01634068600833416,
+ "learning_rate": 0.0006,
+ "loss": 7.2469868659973145,
+ "step": 161
+ },
+ {
+ "epoch": 2.251528384279476,
+ "grad_norm": 0.01728747971355915,
+ "learning_rate": 0.0006,
+ "loss": 7.280995845794678,
+ "step": 162
+ },
+ {
+ "epoch": 2.2655021834061135,
+ "grad_norm": 0.016341188922524452,
+ "learning_rate": 0.0006,
+ "loss": 7.273357391357422,
+ "step": 163
+ },
+ {
+ "epoch": 2.279475982532751,
+ "grad_norm": 0.016125807538628578,
+ "learning_rate": 0.0006,
+ "loss": 7.273510932922363,
+ "step": 164
+ },
+ {
+ "epoch": 2.2934497816593886,
+ "grad_norm": 0.01629287749528885,
+ "learning_rate": 0.0006,
+ "loss": 7.267029762268066,
+ "step": 165
+ },
+ {
+ "epoch": 2.3074235807860264,
+ "grad_norm": 0.015296131372451782,
+ "learning_rate": 0.0006,
+ "loss": 7.211989402770996,
+ "step": 166
+ },
+ {
+ "epoch": 2.3213973799126637,
+ "grad_norm": 0.01709570363163948,
+ "learning_rate": 0.0006,
+ "loss": 7.242851734161377,
+ "step": 167
+ },
+ {
+ "epoch": 2.3353711790393015,
+ "grad_norm": 0.012552388943731785,
+ "learning_rate": 0.0006,
+ "loss": 7.241696834564209,
+ "step": 168
+ },
+ {
+ "epoch": 2.349344978165939,
+ "grad_norm": 0.014571291394531727,
+ "learning_rate": 0.0006,
+ "loss": 7.270681381225586,
+ "step": 169
+ },
+ {
+ "epoch": 2.3633187772925766,
+ "grad_norm": 0.013389437459409237,
+ "learning_rate": 0.0006,
+ "loss": 7.191600799560547,
+ "step": 170
+ },
+ {
+ "epoch": 2.377292576419214,
+ "grad_norm": 0.011895314790308475,
+ "learning_rate": 0.0006,
+ "loss": 7.213610649108887,
+ "step": 171
+ },
+ {
+ "epoch": 2.3912663755458516,
+ "grad_norm": 0.01345642190426588,
+ "learning_rate": 0.0006,
+ "loss": 7.193787574768066,
+ "step": 172
+ },
+ {
+ "epoch": 2.405240174672489,
+ "grad_norm": 0.0091794328764081,
+ "learning_rate": 0.0006,
+ "loss": 7.233364105224609,
+ "step": 173
+ },
+ {
+ "epoch": 2.4192139737991267,
+ "grad_norm": 0.013767086900770664,
+ "learning_rate": 0.0006,
+ "loss": 7.217272758483887,
+ "step": 174
+ },
+ {
+ "epoch": 2.433187772925764,
+ "grad_norm": 0.015134960412979126,
+ "learning_rate": 0.0006,
+ "loss": 7.192791938781738,
+ "step": 175
+ },
+ {
+ "epoch": 2.447161572052402,
+ "grad_norm": 0.01790648326277733,
+ "learning_rate": 0.0006,
+ "loss": 7.229467391967773,
+ "step": 176
+ },
+ {
+ "epoch": 2.461135371179039,
+ "grad_norm": 0.015362377278506756,
+ "learning_rate": 0.0006,
+ "loss": 7.180346488952637,
+ "step": 177
+ },
+ {
+ "epoch": 2.475109170305677,
+ "grad_norm": 0.010505498386919498,
+ "learning_rate": 0.0006,
+ "loss": 7.188833236694336,
+ "step": 178
+ },
+ {
+ "epoch": 2.489082969432314,
+ "grad_norm": 0.009129747748374939,
+ "learning_rate": 0.0006,
+ "loss": 7.203153133392334,
+ "step": 179
+ },
+ {
+ "epoch": 2.503056768558952,
+ "grad_norm": 0.01147883478552103,
+ "learning_rate": 0.0006,
+ "loss": 7.163824558258057,
+ "step": 180
+ },
+ {
+ "epoch": 2.5170305676855893,
+ "grad_norm": 0.011338942684233189,
+ "learning_rate": 0.0006,
+ "loss": 7.16287899017334,
+ "step": 181
+ },
+ {
+ "epoch": 2.531004366812227,
+ "grad_norm": 0.013277675025165081,
+ "learning_rate": 0.0006,
+ "loss": 7.155905723571777,
+ "step": 182
+ },
+ {
+ "epoch": 2.544978165938865,
+ "grad_norm": 0.015269347466528416,
+ "learning_rate": 0.0006,
+ "loss": 7.189371109008789,
+ "step": 183
+ },
+ {
+ "epoch": 2.558951965065502,
+ "grad_norm": 0.017905596643686295,
+ "learning_rate": 0.0006,
+ "loss": 7.177547931671143,
+ "step": 184
+ },
+ {
+ "epoch": 2.5729257641921395,
+ "grad_norm": 0.024431610479950905,
+ "learning_rate": 0.0006,
+ "loss": 7.192169189453125,
+ "step": 185
+ },
+ {
+ "epoch": 2.5868995633187772,
+ "grad_norm": 0.03656961768865585,
+ "learning_rate": 0.0006,
+ "loss": 7.124387741088867,
+ "step": 186
+ },
+ {
+ "epoch": 2.600873362445415,
+ "grad_norm": 0.04337048903107643,
+ "learning_rate": 0.0006,
+ "loss": 7.140231609344482,
+ "step": 187
+ },
+ {
+ "epoch": 2.6148471615720523,
+ "grad_norm": 0.04047228768467903,
+ "learning_rate": 0.0006,
+ "loss": 7.171927452087402,
+ "step": 188
+ },
+ {
+ "epoch": 2.62882096069869,
+ "grad_norm": 0.07825770974159241,
+ "learning_rate": 0.0006,
+ "loss": 7.207631587982178,
+ "step": 189
+ },
+ {
+ "epoch": 2.6427947598253274,
+ "grad_norm": 0.07118063420057297,
+ "learning_rate": 0.0006,
+ "loss": 7.209488391876221,
+ "step": 190
+ },
+ {
+ "epoch": 2.656768558951965,
+ "grad_norm": 0.05153293535113335,
+ "learning_rate": 0.0006,
+ "loss": 7.14793062210083,
+ "step": 191
+ },
+ {
+ "epoch": 2.6707423580786025,
+ "grad_norm": 0.03318839520215988,
+ "learning_rate": 0.0006,
+ "loss": 7.163010597229004,
+ "step": 192
+ },
+ {
+ "epoch": 2.6847161572052403,
+ "grad_norm": 0.029333539307117462,
+ "learning_rate": 0.0006,
+ "loss": 7.159955978393555,
+ "step": 193
+ },
+ {
+ "epoch": 2.6986899563318776,
+ "grad_norm": 0.029405880719423294,
+ "learning_rate": 0.0006,
+ "loss": 7.136178493499756,
+ "step": 194
+ },
+ {
+ "epoch": 2.7126637554585153,
+ "grad_norm": 0.037815988063812256,
+ "learning_rate": 0.0006,
+ "loss": 7.183186054229736,
+ "step": 195
+ },
+ {
+ "epoch": 2.726637554585153,
+ "grad_norm": 0.020307250320911407,
+ "learning_rate": 0.0006,
+ "loss": 7.133164882659912,
+ "step": 196
+ },
+ {
+ "epoch": 2.7406113537117904,
+ "grad_norm": 0.026048069819808006,
+ "learning_rate": 0.0006,
+ "loss": 7.124977111816406,
+ "step": 197
+ },
+ {
+ "epoch": 2.7545851528384278,
+ "grad_norm": 0.027316724881529808,
+ "learning_rate": 0.0006,
+ "loss": 7.07265567779541,
+ "step": 198
+ },
+ {
+ "epoch": 2.7685589519650655,
+ "grad_norm": 0.021629009395837784,
+ "learning_rate": 0.0006,
+ "loss": 7.150020599365234,
+ "step": 199
+ },
+ {
+ "epoch": 2.7825327510917033,
+ "grad_norm": 0.01605929434299469,
+ "learning_rate": 0.0006,
+ "loss": 7.10919189453125,
+ "step": 200
+ },
+ {
+ "epoch": 2.7965065502183406,
+ "grad_norm": 0.020800089463591576,
+ "learning_rate": 0.0006,
+ "loss": 7.132045269012451,
+ "step": 201
+ },
+ {
+ "epoch": 2.810480349344978,
+ "grad_norm": 0.019225774332880974,
+ "learning_rate": 0.0006,
+ "loss": 7.128364562988281,
+ "step": 202
+ },
+ {
+ "epoch": 2.8244541484716157,
+ "grad_norm": 0.01561372634023428,
+ "learning_rate": 0.0006,
+ "loss": 7.083669662475586,
+ "step": 203
+ },
+ {
+ "epoch": 2.8384279475982535,
+ "grad_norm": 0.021969519555568695,
+ "learning_rate": 0.0006,
+ "loss": 7.050841331481934,
+ "step": 204
+ },
+ {
+ "epoch": 2.8524017467248908,
+ "grad_norm": 0.021453695371747017,
+ "learning_rate": 0.0006,
+ "loss": 7.097085475921631,
+ "step": 205
+ },
+ {
+ "epoch": 2.8663755458515285,
+ "grad_norm": 0.022836022078990936,
+ "learning_rate": 0.0006,
+ "loss": 7.10405969619751,
+ "step": 206
+ },
+ {
+ "epoch": 2.880349344978166,
+ "grad_norm": 0.02430851384997368,
+ "learning_rate": 0.0006,
+ "loss": 7.045181751251221,
+ "step": 207
+ },
+ {
+ "epoch": 2.8943231441048036,
+ "grad_norm": 0.020166993141174316,
+ "learning_rate": 0.0006,
+ "loss": 7.068183898925781,
+ "step": 208
+ },
+ {
+ "epoch": 2.908296943231441,
+ "grad_norm": 0.0118505684658885,
+ "learning_rate": 0.0006,
+ "loss": 7.07132625579834,
+ "step": 209
+ },
+ {
+ "epoch": 2.9222707423580787,
+ "grad_norm": 0.023772427812218666,
+ "learning_rate": 0.0006,
+ "loss": 7.055238723754883,
+ "step": 210
+ },
+ {
+ "epoch": 2.936244541484716,
+ "grad_norm": 0.02350712940096855,
+ "learning_rate": 0.0006,
+ "loss": 7.055898666381836,
+ "step": 211
+ },
+ {
+ "epoch": 2.950218340611354,
+ "grad_norm": 0.017244907096028328,
+ "learning_rate": 0.0006,
+ "loss": 7.0099711418151855,
+ "step": 212
+ },
+ {
+ "epoch": 2.964192139737991,
+ "grad_norm": 0.021565575152635574,
+ "learning_rate": 0.0006,
+ "loss": 6.991751670837402,
+ "step": 213
+ },
+ {
+ "epoch": 2.978165938864629,
+ "grad_norm": 0.03241860866546631,
+ "learning_rate": 0.0006,
+ "loss": 7.005980491638184,
+ "step": 214
+ },
+ {
+ "epoch": 2.992139737991266,
+ "grad_norm": 0.04613726586103439,
+ "learning_rate": 0.0006,
+ "loss": 7.027859687805176,
+ "step": 215
+ },
+ {
+ "epoch": 3.0,
+ "grad_norm": 0.06404894590377808,
+ "learning_rate": 0.0006,
+ "loss": 7.123725891113281,
+ "step": 216
+ },
+ {
+ "epoch": 3.0,
+ "eval_loss": 7.070647716522217,
+ "eval_runtime": 58.5801,
+ "eval_samples_per_second": 41.687,
+ "eval_steps_per_second": 1.314,
+ "step": 216
+ },
+ {
+ "epoch": 3.0139737991266378,
+ "grad_norm": 0.05542432889342308,
+ "learning_rate": 0.0006,
+ "loss": 7.078299522399902,
+ "step": 217
+ },
+ {
+ "epoch": 3.027947598253275,
+ "grad_norm": 0.026484526693820953,
+ "learning_rate": 0.0006,
+ "loss": 6.961159706115723,
+ "step": 218
+ },
+ {
+ "epoch": 3.041921397379913,
+ "grad_norm": 0.028276391327381134,
+ "learning_rate": 0.0006,
+ "loss": 7.019961833953857,
+ "step": 219
+ },
+ {
+ "epoch": 3.05589519650655,
+ "grad_norm": 0.024486595764756203,
+ "learning_rate": 0.0006,
+ "loss": 7.039677619934082,
+ "step": 220
+ },
+ {
+ "epoch": 3.069868995633188,
+ "grad_norm": 0.017620893195271492,
+ "learning_rate": 0.0006,
+ "loss": 6.9270782470703125,
+ "step": 221
+ },
+ {
+ "epoch": 3.0838427947598253,
+ "grad_norm": 0.02547302283346653,
+ "learning_rate": 0.0006,
+ "loss": 6.924233436584473,
+ "step": 222
+ },
+ {
+ "epoch": 3.097816593886463,
+ "grad_norm": 0.02539009042084217,
+ "learning_rate": 0.0006,
+ "loss": 6.973979949951172,
+ "step": 223
+ },
+ {
+ "epoch": 3.1117903930131003,
+ "grad_norm": 0.03256227448582649,
+ "learning_rate": 0.0006,
+ "loss": 6.933725357055664,
+ "step": 224
+ },
+ {
+ "epoch": 3.125764192139738,
+ "grad_norm": 0.04651800915598869,
+ "learning_rate": 0.0006,
+ "loss": 6.951044082641602,
+ "step": 225
+ },
+ {
+ "epoch": 3.1397379912663754,
+ "grad_norm": 0.04822975769639015,
+ "learning_rate": 0.0006,
+ "loss": 6.9614410400390625,
+ "step": 226
+ },
+ {
+ "epoch": 3.153711790393013,
+ "grad_norm": 0.031165020540356636,
+ "learning_rate": 0.0006,
+ "loss": 6.91193151473999,
+ "step": 227
+ },
+ {
+ "epoch": 3.1676855895196505,
+ "grad_norm": 0.029754292219877243,
+ "learning_rate": 0.0006,
+ "loss": 6.91802978515625,
+ "step": 228
+ },
+ {
+ "epoch": 3.1816593886462883,
+ "grad_norm": 0.02193152718245983,
+ "learning_rate": 0.0006,
+ "loss": 6.928090572357178,
+ "step": 229
+ },
+ {
+ "epoch": 3.1956331877729256,
+ "grad_norm": 0.02428470179438591,
+ "learning_rate": 0.0006,
+ "loss": 6.90396785736084,
+ "step": 230
+ },
+ {
+ "epoch": 3.2096069868995634,
+ "grad_norm": 0.023374345153570175,
+ "learning_rate": 0.0006,
+ "loss": 6.885858535766602,
+ "step": 231
+ },
+ {
+ "epoch": 3.2235807860262007,
+ "grad_norm": 0.023405000567436218,
+ "learning_rate": 0.0006,
+ "loss": 6.8855767250061035,
+ "step": 232
+ },
+ {
+ "epoch": 3.2375545851528384,
+ "grad_norm": 0.017277134582400322,
+ "learning_rate": 0.0006,
+ "loss": 6.901449203491211,
+ "step": 233
+ },
+ {
+ "epoch": 3.251528384279476,
+ "grad_norm": 0.019496750086545944,
+ "learning_rate": 0.0006,
+ "loss": 6.885687828063965,
+ "step": 234
+ },
+ {
+ "epoch": 3.2655021834061135,
+ "grad_norm": 0.016893276944756508,
+ "learning_rate": 0.0006,
+ "loss": 6.872369766235352,
+ "step": 235
+ },
+ {
+ "epoch": 3.279475982532751,
+ "grad_norm": 0.02058715932071209,
+ "learning_rate": 0.0006,
+ "loss": 6.823280334472656,
+ "step": 236
+ },
+ {
+ "epoch": 3.2934497816593886,
+ "grad_norm": 0.015530755743384361,
+ "learning_rate": 0.0006,
+ "loss": 6.860499858856201,
+ "step": 237
+ },
+ {
+ "epoch": 3.3074235807860264,
+ "grad_norm": 0.01934926211833954,
+ "learning_rate": 0.0006,
+ "loss": 6.817448616027832,
+ "step": 238
+ },
+ {
+ "epoch": 3.3213973799126637,
+ "grad_norm": 0.017413552850484848,
+ "learning_rate": 0.0006,
+ "loss": 6.881955146789551,
+ "step": 239
+ },
+ {
+ "epoch": 3.3353711790393015,
+ "grad_norm": 0.026478875428438187,
+ "learning_rate": 0.0006,
+ "loss": 6.827404499053955,
+ "step": 240
+ },
+ {
+ "epoch": 3.349344978165939,
+ "grad_norm": 0.032611701637506485,
+ "learning_rate": 0.0006,
+ "loss": 6.78285551071167,
+ "step": 241
+ },
+ {
+ "epoch": 3.3633187772925766,
+ "grad_norm": 0.041900306940078735,
+ "learning_rate": 0.0006,
+ "loss": 6.783053874969482,
+ "step": 242
+ },
+ {
+ "epoch": 3.377292576419214,
+ "grad_norm": 0.04149497672915459,
+ "learning_rate": 0.0006,
+ "loss": 6.745220184326172,
+ "step": 243
+ },
+ {
+ "epoch": 3.3912663755458516,
+ "grad_norm": 0.023949826136231422,
+ "learning_rate": 0.0006,
+ "loss": 6.841533184051514,
+ "step": 244
+ },
+ {
+ "epoch": 3.405240174672489,
+ "grad_norm": 0.028601842001080513,
+ "learning_rate": 0.0006,
+ "loss": 6.852852821350098,
+ "step": 245
+ },
+ {
+ "epoch": 3.4192139737991267,
+ "grad_norm": 0.027365155518054962,
+ "learning_rate": 0.0006,
+ "loss": 6.74376106262207,
+ "step": 246
+ },
+ {
+ "epoch": 3.433187772925764,
+ "grad_norm": 0.02856568619608879,
+ "learning_rate": 0.0006,
+ "loss": 6.74906063079834,
+ "step": 247
+ },
+ {
+ "epoch": 3.447161572052402,
+ "grad_norm": 0.019172416999936104,
+ "learning_rate": 0.0006,
+ "loss": 6.809206008911133,
+ "step": 248
+ },
+ {
+ "epoch": 3.461135371179039,
+ "grad_norm": 0.02161746844649315,
+ "learning_rate": 0.0006,
+ "loss": 6.789888381958008,
+ "step": 249
+ },
+ {
+ "epoch": 3.475109170305677,
+ "grad_norm": 0.020273666828870773,
+ "learning_rate": 0.0006,
+ "loss": 6.73512077331543,
+ "step": 250
+ },
+ {
+ "epoch": 3.489082969432314,
+ "grad_norm": 0.022329283878207207,
+ "learning_rate": 0.0006,
+ "loss": 6.826148509979248,
+ "step": 251
+ },
+ {
+ "epoch": 3.503056768558952,
+ "grad_norm": 0.02284284122288227,
+ "learning_rate": 0.0006,
+ "loss": 6.748521327972412,
+ "step": 252
+ },
+ {
+ "epoch": 3.5170305676855893,
+ "grad_norm": 0.026368247345089912,
+ "learning_rate": 0.0006,
+ "loss": 6.645879745483398,
+ "step": 253
+ },
+ {
+ "epoch": 3.531004366812227,
+ "grad_norm": 0.03627244383096695,
+ "learning_rate": 0.0006,
+ "loss": 6.666064739227295,
+ "step": 254
+ },
+ {
+ "epoch": 3.544978165938865,
+ "grad_norm": 0.04140935093164444,
+ "learning_rate": 0.0006,
+ "loss": 6.712275981903076,
+ "step": 255
+ },
+ {
+ "epoch": 3.558951965065502,
+ "grad_norm": 0.0390462800860405,
+ "learning_rate": 0.0006,
+ "loss": 6.665900230407715,
+ "step": 256
+ },
+ {
+ "epoch": 3.5729257641921395,
+ "grad_norm": 0.05178583785891533,
+ "learning_rate": 0.0006,
+ "loss": 6.753298759460449,
+ "step": 257
+ },
+ {
+ "epoch": 3.5868995633187772,
+ "grad_norm": 0.054328061640262604,
+ "learning_rate": 0.0006,
+ "loss": 6.732633113861084,
+ "step": 258
+ },
+ {
+ "epoch": 3.600873362445415,
+ "grad_norm": 0.04590460658073425,
+ "learning_rate": 0.0006,
+ "loss": 6.65250301361084,
+ "step": 259
+ },
+ {
+ "epoch": 3.6148471615720523,
+ "grad_norm": 0.044331666082143784,
+ "learning_rate": 0.0006,
+ "loss": 6.68654203414917,
+ "step": 260
+ },
+ {
+ "epoch": 3.62882096069869,
+ "grad_norm": 0.05308730527758598,
+ "learning_rate": 0.0006,
+ "loss": 6.642482757568359,
+ "step": 261
+ },
+ {
+ "epoch": 3.6427947598253274,
+ "grad_norm": 0.03999016433954239,
+ "learning_rate": 0.0006,
+ "loss": 6.625227451324463,
+ "step": 262
+ },
+ {
+ "epoch": 3.656768558951965,
+ "grad_norm": 0.04088086634874344,
+ "learning_rate": 0.0006,
+ "loss": 6.67623233795166,
+ "step": 263
+ },
+ {
+ "epoch": 3.6707423580786025,
+ "grad_norm": 0.03615636005997658,
+ "learning_rate": 0.0006,
+ "loss": 6.6869001388549805,
+ "step": 264
+ },
+ {
+ "epoch": 3.6847161572052403,
+ "grad_norm": 0.02817360684275627,
+ "learning_rate": 0.0006,
+ "loss": 6.674601078033447,
+ "step": 265
+ },
+ {
+ "epoch": 3.6986899563318776,
+ "grad_norm": 0.032586123794317245,
+ "learning_rate": 0.0006,
+ "loss": 6.612994194030762,
+ "step": 266
+ },
+ {
+ "epoch": 3.7126637554585153,
+ "grad_norm": 0.031117867678403854,
+ "learning_rate": 0.0006,
+ "loss": 6.6244354248046875,
+ "step": 267
+ },
+ {
+ "epoch": 3.726637554585153,
+ "grad_norm": 0.028330687433481216,
+ "learning_rate": 0.0006,
+ "loss": 6.669557094573975,
+ "step": 268
+ },
+ {
+ "epoch": 3.7406113537117904,
+ "grad_norm": 0.021489202976226807,
+ "learning_rate": 0.0006,
+ "loss": 6.632035732269287,
+ "step": 269
+ },
+ {
+ "epoch": 3.7545851528384278,
+ "grad_norm": 0.024269424378871918,
+ "learning_rate": 0.0006,
+ "loss": 6.660867691040039,
+ "step": 270
+ },
+ {
+ "epoch": 3.7685589519650655,
+ "grad_norm": 0.02237873338162899,
+ "learning_rate": 0.0006,
+ "loss": 6.602587699890137,
+ "step": 271
+ },
+ {
+ "epoch": 3.7825327510917033,
+ "grad_norm": 0.026503929868340492,
+ "learning_rate": 0.0006,
+ "loss": 6.599200248718262,
+ "step": 272
+ },
+ {
+ "epoch": 3.7965065502183406,
+ "grad_norm": 0.03687124326825142,
+ "learning_rate": 0.0006,
+ "loss": 6.567296504974365,
+ "step": 273
+ },
+ {
+ "epoch": 3.810480349344978,
+ "grad_norm": 0.04708952456712723,
+ "learning_rate": 0.0006,
+ "loss": 6.621265411376953,
+ "step": 274
+ },
+ {
+ "epoch": 3.8244541484716157,
+ "grad_norm": 0.041846614331007004,
+ "learning_rate": 0.0006,
+ "loss": 6.550536155700684,
+ "step": 275
+ },
+ {
+ "epoch": 3.8384279475982535,
+ "grad_norm": 0.031249675899744034,
+ "learning_rate": 0.0006,
+ "loss": 6.54278564453125,
+ "step": 276
+ },
+ {
+ "epoch": 3.8524017467248908,
+ "grad_norm": 0.023082256317138672,
+ "learning_rate": 0.0006,
+ "loss": 6.544787406921387,
+ "step": 277
+ },
+ {
+ "epoch": 3.8663755458515285,
+ "grad_norm": 0.020995570346713066,
+ "learning_rate": 0.0006,
+ "loss": 6.518143653869629,
+ "step": 278
+ },
+ {
+ "epoch": 3.880349344978166,
+ "grad_norm": 0.020839324221014977,
+ "learning_rate": 0.0006,
+ "loss": 6.546667098999023,
+ "step": 279
+ },
+ {
+ "epoch": 3.8943231441048036,
+ "grad_norm": 0.018377432599663734,
+ "learning_rate": 0.0006,
+ "loss": 6.478307723999023,
+ "step": 280
+ },
+ {
+ "epoch": 3.908296943231441,
+ "grad_norm": 0.021713724359869957,
+ "learning_rate": 0.0006,
+ "loss": 6.485574245452881,
+ "step": 281
+ },
+ {
+ "epoch": 3.9222707423580787,
+ "grad_norm": 0.01359301246702671,
+ "learning_rate": 0.0006,
+ "loss": 6.570268630981445,
+ "step": 282
+ },
+ {
+ "epoch": 3.936244541484716,
+ "grad_norm": 0.016233332455158234,
+ "learning_rate": 0.0006,
+ "loss": 6.57150936126709,
+ "step": 283
+ },
+ {
+ "epoch": 3.950218340611354,
+ "grad_norm": 0.020687608048319817,
+ "learning_rate": 0.0006,
+ "loss": 6.527956008911133,
+ "step": 284
+ },
+ {
+ "epoch": 3.964192139737991,
+ "grad_norm": 0.02032964862883091,
+ "learning_rate": 0.0006,
+ "loss": 6.468169689178467,
+ "step": 285
+ },
+ {
+ "epoch": 3.978165938864629,
+ "grad_norm": 0.016676288098096848,
+ "learning_rate": 0.0006,
+ "loss": 6.448418617248535,
+ "step": 286
+ },
+ {
+ "epoch": 3.992139737991266,
+ "grad_norm": 0.0168539360165596,
+ "learning_rate": 0.0006,
+ "loss": 6.461180686950684,
+ "step": 287
+ },
+ {
+ "epoch": 4.0,
+ "grad_norm": 0.02666233666241169,
+ "learning_rate": 0.0006,
+ "loss": 6.344893932342529,
+ "step": 288
+ },
+ {
+ "epoch": 4.0,
+ "eval_loss": 6.5384979248046875,
+ "eval_runtime": 58.814,
+ "eval_samples_per_second": 41.521,
+ "eval_steps_per_second": 1.309,
+ "step": 288
+ },
+ {
+ "epoch": 4.013973799126638,
+ "grad_norm": 0.04154708981513977,
+ "learning_rate": 0.0006,
+ "loss": 6.41855525970459,
+ "step": 289
+ },
+ {
+ "epoch": 4.0279475982532755,
+ "grad_norm": 0.04895668104290962,
+ "learning_rate": 0.0006,
+ "loss": 6.499302864074707,
+ "step": 290
+ },
+ {
+ "epoch": 4.041921397379912,
+ "grad_norm": 0.034464482218027115,
+ "learning_rate": 0.0006,
+ "loss": 6.438611030578613,
+ "step": 291
+ },
+ {
+ "epoch": 4.05589519650655,
+ "grad_norm": 0.06845773756504059,
+ "learning_rate": 0.0006,
+ "loss": 6.538361072540283,
+ "step": 292
+ },
+ {
+ "epoch": 4.069868995633188,
+ "grad_norm": 0.053768858313560486,
+ "learning_rate": 0.0006,
+ "loss": 6.527524948120117,
+ "step": 293
+ },
+ {
+ "epoch": 4.083842794759826,
+ "grad_norm": 0.039037931710481644,
+ "learning_rate": 0.0006,
+ "loss": 6.503847122192383,
+ "step": 294
+ },
+ {
+ "epoch": 4.097816593886463,
+ "grad_norm": 0.03962196782231331,
+ "learning_rate": 0.0006,
+ "loss": 6.475178241729736,
+ "step": 295
+ },
+ {
+ "epoch": 4.1117903930131,
+ "grad_norm": 0.049101535230875015,
+ "learning_rate": 0.0006,
+ "loss": 6.474587917327881,
+ "step": 296
+ },
+ {
+ "epoch": 4.125764192139738,
+ "grad_norm": 0.02830282226204872,
+ "learning_rate": 0.0006,
+ "loss": 6.45065975189209,
+ "step": 297
+ },
+ {
+ "epoch": 4.139737991266376,
+ "grad_norm": 0.023153482005000114,
+ "learning_rate": 0.0006,
+ "loss": 6.5218706130981445,
+ "step": 298
+ },
+ {
+ "epoch": 4.153711790393013,
+ "grad_norm": 0.025064002722501755,
+ "learning_rate": 0.0006,
+ "loss": 6.456475257873535,
+ "step": 299
+ },
+ {
+ "epoch": 4.1676855895196505,
+ "grad_norm": 0.02357092685997486,
+ "learning_rate": 0.0006,
+ "loss": 6.485952377319336,
+ "step": 300
+ },
+ {
+ "epoch": 4.181659388646288,
+ "grad_norm": 0.021832305938005447,
+ "learning_rate": 0.0006,
+ "loss": 6.425506591796875,
+ "step": 301
+ },
+ {
+ "epoch": 4.195633187772926,
+ "grad_norm": 0.0227900929749012,
+ "learning_rate": 0.0006,
+ "loss": 6.439155578613281,
+ "step": 302
+ },
+ {
+ "epoch": 4.209606986899563,
+ "grad_norm": 0.015961607918143272,
+ "learning_rate": 0.0006,
+ "loss": 6.460862636566162,
+ "step": 303
+ },
+ {
+ "epoch": 4.223580786026201,
+ "grad_norm": 0.017876390367746353,
+ "learning_rate": 0.0006,
+ "loss": 6.44564151763916,
+ "step": 304
+ },
+ {
+ "epoch": 4.2375545851528384,
+ "grad_norm": 0.013703204691410065,
+ "learning_rate": 0.0006,
+ "loss": 6.418414115905762,
+ "step": 305
+ },
+ {
+ "epoch": 4.251528384279476,
+ "grad_norm": 0.015240306034684181,
+ "learning_rate": 0.0006,
+ "loss": 6.411238670349121,
+ "step": 306
+ },
+ {
+ "epoch": 4.265502183406113,
+ "grad_norm": 0.01587662845849991,
+ "learning_rate": 0.0006,
+ "loss": 6.395642280578613,
+ "step": 307
+ },
+ {
+ "epoch": 4.279475982532751,
+ "grad_norm": 0.014650700613856316,
+ "learning_rate": 0.0006,
+ "loss": 6.302793025970459,
+ "step": 308
+ },
+ {
+ "epoch": 4.293449781659389,
+ "grad_norm": 0.016533225774765015,
+ "learning_rate": 0.0006,
+ "loss": 6.380588054656982,
+ "step": 309
+ },
+ {
+ "epoch": 4.307423580786026,
+ "grad_norm": 0.019963741302490234,
+ "learning_rate": 0.0006,
+ "loss": 6.345336437225342,
+ "step": 310
+ },
+ {
+ "epoch": 4.321397379912664,
+ "grad_norm": 0.02237936295568943,
+ "learning_rate": 0.0006,
+ "loss": 6.387770652770996,
+ "step": 311
+ },
+ {
+ "epoch": 4.335371179039301,
+ "grad_norm": 0.026658328250050545,
+ "learning_rate": 0.0006,
+ "loss": 6.323662757873535,
+ "step": 312
+ },
+ {
+ "epoch": 4.349344978165939,
+ "grad_norm": 0.02852284163236618,
+ "learning_rate": 0.0006,
+ "loss": 6.382204055786133,
+ "step": 313
+ },
+ {
+ "epoch": 4.3633187772925766,
+ "grad_norm": 0.028815090656280518,
+ "learning_rate": 0.0006,
+ "loss": 6.402390480041504,
+ "step": 314
+ },
+ {
+ "epoch": 4.377292576419214,
+ "grad_norm": 0.029393529519438744,
+ "learning_rate": 0.0006,
+ "loss": 6.280278205871582,
+ "step": 315
+ },
+ {
+ "epoch": 4.391266375545851,
+ "grad_norm": 0.02186041697859764,
+ "learning_rate": 0.0006,
+ "loss": 6.302707672119141,
+ "step": 316
+ },
+ {
+ "epoch": 4.405240174672489,
+ "grad_norm": 0.02071218006312847,
+ "learning_rate": 0.0006,
+ "loss": 6.3365020751953125,
+ "step": 317
+ },
+ {
+ "epoch": 4.419213973799127,
+ "grad_norm": 0.02779117226600647,
+ "learning_rate": 0.0006,
+ "loss": 6.252157211303711,
+ "step": 318
+ },
+ {
+ "epoch": 4.4331877729257645,
+ "grad_norm": 0.03470654785633087,
+ "learning_rate": 0.0006,
+ "loss": 6.319070339202881,
+ "step": 319
+ },
+ {
+ "epoch": 4.447161572052401,
+ "grad_norm": 0.037067804485559464,
+ "learning_rate": 0.0006,
+ "loss": 6.314116477966309,
+ "step": 320
+ },
+ {
+ "epoch": 4.461135371179039,
+ "grad_norm": 0.032182756811380386,
+ "learning_rate": 0.0006,
+ "loss": 6.267298698425293,
+ "step": 321
+ },
+ {
+ "epoch": 4.475109170305677,
+ "grad_norm": 0.026305217295885086,
+ "learning_rate": 0.0006,
+ "loss": 6.331688404083252,
+ "step": 322
+ },
+ {
+ "epoch": 4.489082969432315,
+ "grad_norm": 0.027228357270359993,
+ "learning_rate": 0.0006,
+ "loss": 6.31139612197876,
+ "step": 323
+ },
+ {
+ "epoch": 4.503056768558952,
+ "grad_norm": 0.025781169533729553,
+ "learning_rate": 0.0006,
+ "loss": 6.194684982299805,
+ "step": 324
+ },
+ {
+ "epoch": 4.517030567685589,
+ "grad_norm": 0.029186977073550224,
+ "learning_rate": 0.0006,
+ "loss": 6.31306266784668,
+ "step": 325
+ },
+ {
+ "epoch": 4.531004366812227,
+ "grad_norm": 0.023608777672052383,
+ "learning_rate": 0.0006,
+ "loss": 6.285243034362793,
+ "step": 326
+ },
+ {
+ "epoch": 4.544978165938865,
+ "grad_norm": 0.02267594076693058,
+ "learning_rate": 0.0006,
+ "loss": 6.30342960357666,
+ "step": 327
+ },
+ {
+ "epoch": 4.558951965065502,
+ "grad_norm": 0.01926310732960701,
+ "learning_rate": 0.0006,
+ "loss": 6.280606269836426,
+ "step": 328
+ },
+ {
+ "epoch": 4.5729257641921395,
+ "grad_norm": 0.026484837755560875,
+ "learning_rate": 0.0006,
+ "loss": 6.237980365753174,
+ "step": 329
+ },
+ {
+ "epoch": 4.586899563318777,
+ "grad_norm": 0.026228854432702065,
+ "learning_rate": 0.0006,
+ "loss": 6.3018293380737305,
+ "step": 330
+ },
+ {
+ "epoch": 4.600873362445415,
+ "grad_norm": 0.022096967324614525,
+ "learning_rate": 0.0006,
+ "loss": 6.253863334655762,
+ "step": 331
+ },
+ {
+ "epoch": 4.614847161572053,
+ "grad_norm": 0.027223890647292137,
+ "learning_rate": 0.0006,
+ "loss": 6.166882038116455,
+ "step": 332
+ },
+ {
+ "epoch": 4.62882096069869,
+ "grad_norm": 0.03160124272108078,
+ "learning_rate": 0.0006,
+ "loss": 6.230309963226318,
+ "step": 333
+ },
+ {
+ "epoch": 4.642794759825327,
+ "grad_norm": 0.04110539332032204,
+ "learning_rate": 0.0006,
+ "loss": 6.2334818840026855,
+ "step": 334
+ },
+ {
+ "epoch": 4.656768558951965,
+ "grad_norm": 0.04355933889746666,
+ "learning_rate": 0.0006,
+ "loss": 6.160956382751465,
+ "step": 335
+ },
+ {
+ "epoch": 4.670742358078603,
+ "grad_norm": 0.03611086308956146,
+ "learning_rate": 0.0006,
+ "loss": 6.23399543762207,
+ "step": 336
+ },
+ {
+ "epoch": 4.68471615720524,
+ "grad_norm": 0.051811400800943375,
+ "learning_rate": 0.0006,
+ "loss": 6.241855144500732,
+ "step": 337
+ },
+ {
+ "epoch": 4.698689956331878,
+ "grad_norm": 0.057434193789958954,
+ "learning_rate": 0.0006,
+ "loss": 6.220993995666504,
+ "step": 338
+ },
+ {
+ "epoch": 4.712663755458515,
+ "grad_norm": 0.04420953616499901,
+ "learning_rate": 0.0006,
+ "loss": 6.300461292266846,
+ "step": 339
+ },
+ {
+ "epoch": 4.726637554585153,
+ "grad_norm": 0.03515457361936569,
+ "learning_rate": 0.0006,
+ "loss": 6.269349098205566,
+ "step": 340
+ },
+ {
+ "epoch": 4.74061135371179,
+ "grad_norm": 0.03222600743174553,
+ "learning_rate": 0.0006,
+ "loss": 6.248613357543945,
+ "step": 341
+ },
+ {
+ "epoch": 4.754585152838428,
+ "grad_norm": 0.0384046845138073,
+ "learning_rate": 0.0006,
+ "loss": 6.266979217529297,
+ "step": 342
+ },
+ {
+ "epoch": 4.7685589519650655,
+ "grad_norm": 0.0514867827296257,
+ "learning_rate": 0.0006,
+ "loss": 6.234103202819824,
+ "step": 343
+ },
+ {
+ "epoch": 4.782532751091703,
+ "grad_norm": 0.050747793167829514,
+ "learning_rate": 0.0006,
+ "loss": 6.322582244873047,
+ "step": 344
+ },
+ {
+ "epoch": 4.796506550218341,
+ "grad_norm": 0.040878988802433014,
+ "learning_rate": 0.0006,
+ "loss": 6.212509632110596,
+ "step": 345
+ },
+ {
+ "epoch": 4.810480349344978,
+ "grad_norm": 0.034655820578336716,
+ "learning_rate": 0.0006,
+ "loss": 6.265879154205322,
+ "step": 346
+ },
+ {
+ "epoch": 4.824454148471616,
+ "grad_norm": 0.04202224686741829,
+ "learning_rate": 0.0006,
+ "loss": 6.2896409034729,
+ "step": 347
+ },
+ {
+ "epoch": 4.8384279475982535,
+ "grad_norm": 0.03632277995347977,
+ "learning_rate": 0.0006,
+ "loss": 6.253917217254639,
+ "step": 348
+ },
+ {
+ "epoch": 4.85240174672489,
+ "grad_norm": 0.028645765036344528,
+ "learning_rate": 0.0006,
+ "loss": 6.207965850830078,
+ "step": 349
+ },
+ {
+ "epoch": 4.866375545851528,
+ "grad_norm": 0.025305170565843582,
+ "learning_rate": 0.0006,
+ "loss": 6.168068885803223,
+ "step": 350
+ },
+ {
+ "epoch": 4.880349344978166,
+ "grad_norm": 0.022137803956866264,
+ "learning_rate": 0.0006,
+ "loss": 6.235283374786377,
+ "step": 351
+ },
+ {
+ "epoch": 4.894323144104804,
+ "grad_norm": 0.01814538985490799,
+ "learning_rate": 0.0006,
+ "loss": 6.259641647338867,
+ "step": 352
+ },
+ {
+ "epoch": 4.908296943231441,
+ "grad_norm": 0.018465086817741394,
+ "learning_rate": 0.0006,
+ "loss": 6.111738681793213,
+ "step": 353
+ },
+ {
+ "epoch": 4.922270742358078,
+ "grad_norm": 0.018638933077454567,
+ "learning_rate": 0.0006,
+ "loss": 6.225642204284668,
+ "step": 354
+ },
+ {
+ "epoch": 4.936244541484716,
+ "grad_norm": 0.016461260616779327,
+ "learning_rate": 0.0006,
+ "loss": 6.140647888183594,
+ "step": 355
+ },
+ {
+ "epoch": 4.950218340611354,
+ "grad_norm": 0.015458385460078716,
+ "learning_rate": 0.0006,
+ "loss": 6.155224800109863,
+ "step": 356
+ },
+ {
+ "epoch": 4.964192139737992,
+ "grad_norm": 0.014934048056602478,
+ "learning_rate": 0.0006,
+ "loss": 6.101555824279785,
+ "step": 357
+ },
+ {
+ "epoch": 4.978165938864628,
+ "grad_norm": 0.013979545794427395,
+ "learning_rate": 0.0006,
+ "loss": 6.161718845367432,
+ "step": 358
+ },
+ {
+ "epoch": 4.992139737991266,
+ "grad_norm": 0.01640394888818264,
+ "learning_rate": 0.0006,
+ "loss": 6.055768966674805,
+ "step": 359
+ },
+ {
+ "epoch": 5.0,
+ "grad_norm": 0.015135680325329304,
+ "learning_rate": 0.0006,
+ "loss": 6.169695854187012,
+ "step": 360
+ },
+ {
+ "epoch": 5.0,
+ "eval_loss": 6.16408109664917,
+ "eval_runtime": 59.3649,
+ "eval_samples_per_second": 41.135,
+ "eval_steps_per_second": 1.297,
+ "step": 360
+ },
+ {
+ "epoch": 5.013973799126638,
+ "grad_norm": 0.012849048711359501,
+ "learning_rate": 0.0006,
+ "loss": 6.055830001831055,
+ "step": 361
+ },
+ {
+ "epoch": 5.0279475982532755,
+ "grad_norm": 0.018169576302170753,
+ "learning_rate": 0.0006,
+ "loss": 6.082655429840088,
+ "step": 362
+ },
+ {
+ "epoch": 5.041921397379912,
+ "grad_norm": 0.024293430149555206,
+ "learning_rate": 0.0006,
+ "loss": 6.109032154083252,
+ "step": 363
+ },
+ {
+ "epoch": 5.05589519650655,
+ "grad_norm": 0.03266787901520729,
+ "learning_rate": 0.0006,
+ "loss": 6.030613899230957,
+ "step": 364
+ },
+ {
+ "epoch": 5.069868995633188,
+ "grad_norm": 0.03572266176342964,
+ "learning_rate": 0.0006,
+ "loss": 6.050390720367432,
+ "step": 365
+ },
+ {
+ "epoch": 5.083842794759826,
+ "grad_norm": 0.032625213265419006,
+ "learning_rate": 0.0006,
+ "loss": 6.169313430786133,
+ "step": 366
+ },
+ {
+ "epoch": 5.097816593886463,
+ "grad_norm": 0.04051872715353966,
+ "learning_rate": 0.0006,
+ "loss": 6.10398006439209,
+ "step": 367
+ },
+ {
+ "epoch": 5.1117903930131,
+ "grad_norm": 0.0351213738322258,
+ "learning_rate": 0.0006,
+ "loss": 6.121973514556885,
+ "step": 368
+ },
+ {
+ "epoch": 5.125764192139738,
+ "grad_norm": 0.03523759916424751,
+ "learning_rate": 0.0006,
+ "loss": 6.0626654624938965,
+ "step": 369
+ },
+ {
+ "epoch": 5.139737991266376,
+ "grad_norm": 0.02706182189285755,
+ "learning_rate": 0.0006,
+ "loss": 6.049354076385498,
+ "step": 370
+ },
+ {
+ "epoch": 5.153711790393013,
+ "grad_norm": 0.024212589487433434,
+ "learning_rate": 0.0006,
+ "loss": 6.034826278686523,
+ "step": 371
+ },
+ {
+ "epoch": 5.1676855895196505,
+ "grad_norm": 0.024798082187771797,
+ "learning_rate": 0.0006,
+ "loss": 6.004058837890625,
+ "step": 372
+ },
+ {
+ "epoch": 5.181659388646288,
+ "grad_norm": 0.02325567416846752,
+ "learning_rate": 0.0006,
+ "loss": 5.986461639404297,
+ "step": 373
+ },
+ {
+ "epoch": 5.195633187772926,
+ "grad_norm": 0.019060570746660233,
+ "learning_rate": 0.0006,
+ "loss": 6.059736251831055,
+ "step": 374
+ },
+ {
+ "epoch": 5.209606986899563,
+ "grad_norm": 0.016822976991534233,
+ "learning_rate": 0.0006,
+ "loss": 6.072957515716553,
+ "step": 375
+ },
+ {
+ "epoch": 5.223580786026201,
+ "grad_norm": 0.017218658700585365,
+ "learning_rate": 0.0006,
+ "loss": 6.031739234924316,
+ "step": 376
+ },
+ {
+ "epoch": 5.2375545851528384,
+ "grad_norm": 0.014889383688569069,
+ "learning_rate": 0.0006,
+ "loss": 6.025674343109131,
+ "step": 377
+ },
+ {
+ "epoch": 5.251528384279476,
+ "grad_norm": 0.01708107627928257,
+ "learning_rate": 0.0006,
+ "loss": 6.090451240539551,
+ "step": 378
+ },
+ {
+ "epoch": 5.265502183406113,
+ "grad_norm": 0.020260317251086235,
+ "learning_rate": 0.0006,
+ "loss": 6.094257354736328,
+ "step": 379
+ },
+ {
+ "epoch": 5.279475982532751,
+ "grad_norm": 0.020110400393605232,
+ "learning_rate": 0.0006,
+ "loss": 6.021188735961914,
+ "step": 380
+ },
+ {
+ "epoch": 5.293449781659389,
+ "grad_norm": 0.02027830481529236,
+ "learning_rate": 0.0006,
+ "loss": 6.022156238555908,
+ "step": 381
+ },
+ {
+ "epoch": 5.307423580786026,
+ "grad_norm": 0.022747062146663666,
+ "learning_rate": 0.0006,
+ "loss": 6.011836051940918,
+ "step": 382
+ },
+ {
+ "epoch": 5.321397379912664,
+ "grad_norm": 0.01990230567753315,
+ "learning_rate": 0.0006,
+ "loss": 6.003054618835449,
+ "step": 383
+ },
+ {
+ "epoch": 5.335371179039301,
+ "grad_norm": 0.01596238650381565,
+ "learning_rate": 0.0006,
+ "loss": 5.991410732269287,
+ "step": 384
+ },
+ {
+ "epoch": 5.349344978165939,
+ "grad_norm": 0.015847105532884598,
+ "learning_rate": 0.0006,
+ "loss": 6.100622177124023,
+ "step": 385
+ },
+ {
+ "epoch": 5.3633187772925766,
+ "grad_norm": 0.016179397702217102,
+ "learning_rate": 0.0006,
+ "loss": 6.032659530639648,
+ "step": 386
+ },
+ {
+ "epoch": 5.377292576419214,
+ "grad_norm": 0.018256602808833122,
+ "learning_rate": 0.0006,
+ "loss": 5.961983680725098,
+ "step": 387
+ },
+ {
+ "epoch": 5.391266375545851,
+ "grad_norm": 0.02805912122130394,
+ "learning_rate": 0.0006,
+ "loss": 5.983541965484619,
+ "step": 388
+ },
+ {
+ "epoch": 5.405240174672489,
+ "grad_norm": 0.039082255214452744,
+ "learning_rate": 0.0006,
+ "loss": 6.021474361419678,
+ "step": 389
+ },
+ {
+ "epoch": 5.419213973799127,
+ "grad_norm": 0.036757659167051315,
+ "learning_rate": 0.0006,
+ "loss": 6.005046367645264,
+ "step": 390
+ },
+ {
+ "epoch": 5.4331877729257645,
+ "grad_norm": 0.035277366638183594,
+ "learning_rate": 0.0006,
+ "loss": 6.036296844482422,
+ "step": 391
+ },
+ {
+ "epoch": 5.447161572052401,
+ "grad_norm": 0.034404635429382324,
+ "learning_rate": 0.0006,
+ "loss": 5.994539260864258,
+ "step": 392
+ },
+ {
+ "epoch": 5.461135371179039,
+ "grad_norm": 0.0377206988632679,
+ "learning_rate": 0.0006,
+ "loss": 5.941534519195557,
+ "step": 393
+ },
+ {
+ "epoch": 5.475109170305677,
+ "grad_norm": 0.0389922633767128,
+ "learning_rate": 0.0006,
+ "loss": 5.983644485473633,
+ "step": 394
+ },
+ {
+ "epoch": 5.489082969432315,
+ "grad_norm": 0.04176778718829155,
+ "learning_rate": 0.0006,
+ "loss": 6.030922889709473,
+ "step": 395
+ },
+ {
+ "epoch": 5.503056768558952,
+ "grad_norm": 0.0337153784930706,
+ "learning_rate": 0.0006,
+ "loss": 5.961367130279541,
+ "step": 396
+ },
+ {
+ "epoch": 5.517030567685589,
+ "grad_norm": 0.036116816103458405,
+ "learning_rate": 0.0006,
+ "loss": 5.819280624389648,
+ "step": 397
+ },
+ {
+ "epoch": 5.531004366812227,
+ "grad_norm": 0.030724041163921356,
+ "learning_rate": 0.0006,
+ "loss": 5.906380653381348,
+ "step": 398
+ },
+ {
+ "epoch": 5.544978165938865,
+ "grad_norm": 0.030264202505350113,
+ "learning_rate": 0.0006,
+ "loss": 5.975833415985107,
+ "step": 399
+ },
+ {
+ "epoch": 5.558951965065502,
+ "grad_norm": 0.032096318900585175,
+ "learning_rate": 0.0006,
+ "loss": 6.001348972320557,
+ "step": 400
+ },
+ {
+ "epoch": 5.5729257641921395,
+ "grad_norm": 0.030579356476664543,
+ "learning_rate": 0.0006,
+ "loss": 5.928768634796143,
+ "step": 401
+ },
+ {
+ "epoch": 5.586899563318777,
+ "grad_norm": 0.028241973370313644,
+ "learning_rate": 0.0006,
+ "loss": 5.92582893371582,
+ "step": 402
+ },
+ {
+ "epoch": 5.600873362445415,
+ "grad_norm": 0.023086953908205032,
+ "learning_rate": 0.0006,
+ "loss": 5.892926216125488,
+ "step": 403
+ },
+ {
+ "epoch": 5.614847161572053,
+ "grad_norm": 0.02547992393374443,
+ "learning_rate": 0.0006,
+ "loss": 5.903195381164551,
+ "step": 404
+ },
+ {
+ "epoch": 5.62882096069869,
+ "grad_norm": 0.023089831694960594,
+ "learning_rate": 0.0006,
+ "loss": 5.912033557891846,
+ "step": 405
+ },
+ {
+ "epoch": 5.642794759825327,
+ "grad_norm": 0.025446368381381035,
+ "learning_rate": 0.0006,
+ "loss": 5.926138401031494,
+ "step": 406
+ },
+ {
+ "epoch": 5.656768558951965,
+ "grad_norm": 0.03748747706413269,
+ "learning_rate": 0.0006,
+ "loss": 5.913451194763184,
+ "step": 407
+ },
+ {
+ "epoch": 5.670742358078603,
+ "grad_norm": 0.03493810072541237,
+ "learning_rate": 0.0006,
+ "loss": 5.97633695602417,
+ "step": 408
+ },
+ {
+ "epoch": 5.68471615720524,
+ "grad_norm": 0.023131275549530983,
+ "learning_rate": 0.0006,
+ "loss": 5.899571418762207,
+ "step": 409
+ },
+ {
+ "epoch": 5.698689956331878,
+ "grad_norm": 0.02986014075577259,
+ "learning_rate": 0.0006,
+ "loss": 5.894903182983398,
+ "step": 410
+ },
+ {
+ "epoch": 5.712663755458515,
+ "grad_norm": 0.030171144753694534,
+ "learning_rate": 0.0006,
+ "loss": 5.876595497131348,
+ "step": 411
+ },
+ {
+ "epoch": 5.726637554585153,
+ "grad_norm": 0.029377546161413193,
+ "learning_rate": 0.0006,
+ "loss": 5.911637306213379,
+ "step": 412
+ },
+ {
+ "epoch": 5.74061135371179,
+ "grad_norm": 0.029829490929841995,
+ "learning_rate": 0.0006,
+ "loss": 5.819629669189453,
+ "step": 413
+ },
+ {
+ "epoch": 5.754585152838428,
+ "grad_norm": 0.024139299988746643,
+ "learning_rate": 0.0006,
+ "loss": 5.81728458404541,
+ "step": 414
+ },
+ {
+ "epoch": 5.7685589519650655,
+ "grad_norm": 0.021453432738780975,
+ "learning_rate": 0.0006,
+ "loss": 5.987326145172119,
+ "step": 415
+ },
+ {
+ "epoch": 5.782532751091703,
+ "grad_norm": 0.017942246049642563,
+ "learning_rate": 0.0006,
+ "loss": 5.893008232116699,
+ "step": 416
+ },
+ {
+ "epoch": 5.796506550218341,
+ "grad_norm": 0.019724637269973755,
+ "learning_rate": 0.0006,
+ "loss": 5.912441253662109,
+ "step": 417
+ },
+ {
+ "epoch": 5.810480349344978,
+ "grad_norm": 0.018597912043333054,
+ "learning_rate": 0.0006,
+ "loss": 5.860394477844238,
+ "step": 418
+ },
+ {
+ "epoch": 5.824454148471616,
+ "grad_norm": 0.016232412308454514,
+ "learning_rate": 0.0006,
+ "loss": 5.94991397857666,
+ "step": 419
+ },
+ {
+ "epoch": 5.8384279475982535,
+ "grad_norm": 0.01667204685509205,
+ "learning_rate": 0.0006,
+ "loss": 5.8506364822387695,
+ "step": 420
+ },
+ {
+ "epoch": 5.85240174672489,
+ "grad_norm": 0.014232151210308075,
+ "learning_rate": 0.0006,
+ "loss": 5.928424835205078,
+ "step": 421
+ },
+ {
+ "epoch": 5.866375545851528,
+ "grad_norm": 0.01570476032793522,
+ "learning_rate": 0.0006,
+ "loss": 5.953692436218262,
+ "step": 422
+ },
+ {
+ "epoch": 5.880349344978166,
+ "grad_norm": 0.01586179807782173,
+ "learning_rate": 0.0006,
+ "loss": 5.820014476776123,
+ "step": 423
+ },
+ {
+ "epoch": 5.894323144104804,
+ "grad_norm": 0.014729145914316177,
+ "learning_rate": 0.0006,
+ "loss": 5.860154628753662,
+ "step": 424
+ },
+ {
+ "epoch": 5.908296943231441,
+ "grad_norm": 0.013111459091305733,
+ "learning_rate": 0.0006,
+ "loss": 5.85988187789917,
+ "step": 425
+ },
+ {
+ "epoch": 5.922270742358078,
+ "grad_norm": 0.012937922962009907,
+ "learning_rate": 0.0006,
+ "loss": 5.756265640258789,
+ "step": 426
+ },
+ {
+ "epoch": 5.936244541484716,
+ "grad_norm": 0.01377453189343214,
+ "learning_rate": 0.0006,
+ "loss": 5.8333048820495605,
+ "step": 427
+ },
+ {
+ "epoch": 5.950218340611354,
+ "grad_norm": 0.014054795727133751,
+ "learning_rate": 0.0006,
+ "loss": 5.803333759307861,
+ "step": 428
+ },
+ {
+ "epoch": 5.964192139737992,
+ "grad_norm": 0.01674959622323513,
+ "learning_rate": 0.0006,
+ "loss": 5.8718485832214355,
+ "step": 429
+ },
+ {
+ "epoch": 5.978165938864628,
+ "grad_norm": 0.018733017146587372,
+ "learning_rate": 0.0006,
+ "loss": 5.875979900360107,
+ "step": 430
+ },
+ {
+ "epoch": 5.992139737991266,
+ "grad_norm": 0.02088983915746212,
+ "learning_rate": 0.0006,
+ "loss": 5.763634204864502,
+ "step": 431
+ },
+ {
+ "epoch": 6.0,
+ "grad_norm": 0.026705985888838768,
+ "learning_rate": 0.0006,
+ "loss": 5.6922478675842285,
+ "step": 432
+ },
+ {
+ "epoch": 6.0,
+ "eval_loss": 5.847958087921143,
+ "eval_runtime": 58.5451,
+ "eval_samples_per_second": 41.711,
+ "eval_steps_per_second": 1.315,
+ "step": 432
+ },
+ {
+ "epoch": 6.013973799126638,
+ "grad_norm": 0.03181544691324234,
+ "learning_rate": 0.0006,
+ "loss": 5.786368370056152,
+ "step": 433
+ },
+ {
+ "epoch": 6.0279475982532755,
+ "grad_norm": 0.03238112851977348,
+ "learning_rate": 0.0006,
+ "loss": 5.810310363769531,
+ "step": 434
+ },
+ {
+ "epoch": 6.041921397379912,
+ "grad_norm": 0.03916756808757782,
+ "learning_rate": 0.0006,
+ "loss": 5.7761993408203125,
+ "step": 435
+ },
+ {
+ "epoch": 6.05589519650655,
+ "grad_norm": 0.044009730219841,
+ "learning_rate": 0.0006,
+ "loss": 5.898112773895264,
+ "step": 436
+ },
+ {
+ "epoch": 6.069868995633188,
+ "grad_norm": 0.045235246419906616,
+ "learning_rate": 0.0006,
+ "loss": 5.876371383666992,
+ "step": 437
+ },
+ {
+ "epoch": 6.083842794759826,
+ "grad_norm": 0.050218384712934494,
+ "learning_rate": 0.0006,
+ "loss": 5.805103302001953,
+ "step": 438
+ },
+ {
+ "epoch": 6.097816593886463,
+ "grad_norm": 0.05444490164518356,
+ "learning_rate": 0.0006,
+ "loss": 5.826424598693848,
+ "step": 439
+ },
+ {
+ "epoch": 6.1117903930131,
+ "grad_norm": 0.04890982061624527,
+ "learning_rate": 0.0006,
+ "loss": 5.7735395431518555,
+ "step": 440
+ },
+ {
+ "epoch": 6.125764192139738,
+ "grad_norm": 0.05478639155626297,
+ "learning_rate": 0.0006,
+ "loss": 5.823677062988281,
+ "step": 441
+ },
+ {
+ "epoch": 6.139737991266376,
+ "grad_norm": 0.057562313973903656,
+ "learning_rate": 0.0006,
+ "loss": 5.889334678649902,
+ "step": 442
+ },
+ {
+ "epoch": 6.153711790393013,
+ "grad_norm": 0.06447703391313553,
+ "learning_rate": 0.0006,
+ "loss": 5.84621524810791,
+ "step": 443
+ },
+ {
+ "epoch": 6.1676855895196505,
+ "grad_norm": 0.048861872404813766,
+ "learning_rate": 0.0006,
+ "loss": 5.931595802307129,
+ "step": 444
+ },
+ {
+ "epoch": 6.181659388646288,
+ "grad_norm": 0.05521339178085327,
+ "learning_rate": 0.0006,
+ "loss": 5.8229780197143555,
+ "step": 445
+ },
+ {
+ "epoch": 6.195633187772926,
+ "grad_norm": 0.053666990250349045,
+ "learning_rate": 0.0006,
+ "loss": 5.879191875457764,
+ "step": 446
+ },
+ {
+ "epoch": 6.209606986899563,
+ "grad_norm": 0.0451025515794754,
+ "learning_rate": 0.0006,
+ "loss": 5.801628112792969,
+ "step": 447
+ },
+ {
+ "epoch": 6.223580786026201,
+ "grad_norm": 0.04291655868291855,
+ "learning_rate": 0.0006,
+ "loss": 5.914680480957031,
+ "step": 448
+ },
+ {
+ "epoch": 6.2375545851528384,
+ "grad_norm": 0.035102490335702896,
+ "learning_rate": 0.0006,
+ "loss": 5.843682765960693,
+ "step": 449
+ },
+ {
+ "epoch": 6.251528384279476,
+ "grad_norm": 0.03403995931148529,
+ "learning_rate": 0.0006,
+ "loss": 5.827154159545898,
+ "step": 450
+ },
+ {
+ "epoch": 6.265502183406113,
+ "grad_norm": 0.03444375470280647,
+ "learning_rate": 0.0006,
+ "loss": 5.800136566162109,
+ "step": 451
+ },
+ {
+ "epoch": 6.279475982532751,
+ "grad_norm": 0.03165159001946449,
+ "learning_rate": 0.0006,
+ "loss": 5.906252384185791,
+ "step": 452
+ },
+ {
+ "epoch": 6.293449781659389,
+ "grad_norm": 0.026153093203902245,
+ "learning_rate": 0.0006,
+ "loss": 5.7423200607299805,
+ "step": 453
+ },
+ {
+ "epoch": 6.307423580786026,
+ "grad_norm": 0.024357657879590988,
+ "learning_rate": 0.0006,
+ "loss": 5.701364517211914,
+ "step": 454
+ },
+ {
+ "epoch": 6.321397379912664,
+ "grad_norm": 0.021508049219846725,
+ "learning_rate": 0.0006,
+ "loss": 5.8328094482421875,
+ "step": 455
+ },
+ {
+ "epoch": 6.335371179039301,
+ "grad_norm": 0.017313921824097633,
+ "learning_rate": 0.0006,
+ "loss": 5.792145252227783,
+ "step": 456
+ },
+ {
+ "epoch": 6.349344978165939,
+ "grad_norm": 0.018563397228717804,
+ "learning_rate": 0.0006,
+ "loss": 5.732672214508057,
+ "step": 457
+ },
+ {
+ "epoch": 6.3633187772925766,
+ "grad_norm": 0.016568679362535477,
+ "learning_rate": 0.0006,
+ "loss": 5.81948184967041,
+ "step": 458
+ },
+ {
+ "epoch": 6.377292576419214,
+ "grad_norm": 0.014133770950138569,
+ "learning_rate": 0.0006,
+ "loss": 5.734982490539551,
+ "step": 459
+ },
+ {
+ "epoch": 6.391266375545851,
+ "grad_norm": 0.014114447869360447,
+ "learning_rate": 0.0006,
+ "loss": 5.70073127746582,
+ "step": 460
+ },
+ {
+ "epoch": 6.405240174672489,
+ "grad_norm": 0.012907393276691437,
+ "learning_rate": 0.0006,
+ "loss": 5.758626937866211,
+ "step": 461
+ },
+ {
+ "epoch": 6.419213973799127,
+ "grad_norm": 0.012673444114625454,
+ "learning_rate": 0.0006,
+ "loss": 5.775138854980469,
+ "step": 462
+ },
+ {
+ "epoch": 6.4331877729257645,
+ "grad_norm": 0.01311410591006279,
+ "learning_rate": 0.0006,
+ "loss": 5.70920991897583,
+ "step": 463
+ },
+ {
+ "epoch": 6.447161572052401,
+ "grad_norm": 0.012935544364154339,
+ "learning_rate": 0.0006,
+ "loss": 5.781479835510254,
+ "step": 464
+ },
+ {
+ "epoch": 6.461135371179039,
+ "grad_norm": 0.01439738366752863,
+ "learning_rate": 0.0006,
+ "loss": 5.691315650939941,
+ "step": 465
+ },
+ {
+ "epoch": 6.475109170305677,
+ "grad_norm": 0.01205186452716589,
+ "learning_rate": 0.0006,
+ "loss": 5.662499904632568,
+ "step": 466
+ },
+ {
+ "epoch": 6.489082969432315,
+ "grad_norm": 0.011575652286410332,
+ "learning_rate": 0.0006,
+ "loss": 5.6271772384643555,
+ "step": 467
+ },
+ {
+ "epoch": 6.503056768558952,
+ "grad_norm": 0.011672502383589745,
+ "learning_rate": 0.0006,
+ "loss": 5.761662006378174,
+ "step": 468
+ },
+ {
+ "epoch": 6.517030567685589,
+ "grad_norm": 0.011541751213371754,
+ "learning_rate": 0.0006,
+ "loss": 5.762078285217285,
+ "step": 469
+ },
+ {
+ "epoch": 6.531004366812227,
+ "grad_norm": 0.011396365240216255,
+ "learning_rate": 0.0006,
+ "loss": 5.67873477935791,
+ "step": 470
+ },
+ {
+ "epoch": 6.544978165938865,
+ "grad_norm": 0.010688499547541142,
+ "learning_rate": 0.0006,
+ "loss": 5.717051982879639,
+ "step": 471
+ },
+ {
+ "epoch": 6.558951965065502,
+ "grad_norm": 0.012224317528307438,
+ "learning_rate": 0.0006,
+ "loss": 5.707948684692383,
+ "step": 472
+ },
+ {
+ "epoch": 6.5729257641921395,
+ "grad_norm": 0.011856825090944767,
+ "learning_rate": 0.0006,
+ "loss": 5.70878791809082,
+ "step": 473
+ },
+ {
+ "epoch": 6.586899563318777,
+ "grad_norm": 0.01199064590036869,
+ "learning_rate": 0.0006,
+ "loss": 5.708697319030762,
+ "step": 474
+ },
+ {
+ "epoch": 6.600873362445415,
+ "grad_norm": 0.01219446212053299,
+ "learning_rate": 0.0006,
+ "loss": 5.6061577796936035,
+ "step": 475
+ },
+ {
+ "epoch": 6.614847161572053,
+ "grad_norm": 0.013397484086453915,
+ "learning_rate": 0.0006,
+ "loss": 5.624789714813232,
+ "step": 476
+ },
+ {
+ "epoch": 6.62882096069869,
+ "grad_norm": 0.01483136136084795,
+ "learning_rate": 0.0006,
+ "loss": 5.721141338348389,
+ "step": 477
+ },
+ {
+ "epoch": 6.642794759825327,
+ "grad_norm": 0.0194488987326622,
+ "learning_rate": 0.0006,
+ "loss": 5.604279518127441,
+ "step": 478
+ },
+ {
+ "epoch": 6.656768558951965,
+ "grad_norm": 0.02079896256327629,
+ "learning_rate": 0.0006,
+ "loss": 5.586322784423828,
+ "step": 479
+ },
+ {
+ "epoch": 6.670742358078603,
+ "grad_norm": 0.018996229395270348,
+ "learning_rate": 0.0006,
+ "loss": 5.5987772941589355,
+ "step": 480
+ },
+ {
+ "epoch": 6.68471615720524,
+ "grad_norm": 0.015779603272676468,
+ "learning_rate": 0.0006,
+ "loss": 5.588602066040039,
+ "step": 481
+ },
+ {
+ "epoch": 6.698689956331878,
+ "grad_norm": 0.015322973020374775,
+ "learning_rate": 0.0006,
+ "loss": 5.686467170715332,
+ "step": 482
+ },
+ {
+ "epoch": 6.712663755458515,
+ "grad_norm": 0.015282213687896729,
+ "learning_rate": 0.0006,
+ "loss": 5.664676666259766,
+ "step": 483
+ },
+ {
+ "epoch": 6.726637554585153,
+ "grad_norm": 0.016589272767305374,
+ "learning_rate": 0.0006,
+ "loss": 5.666739463806152,
+ "step": 484
+ },
+ {
+ "epoch": 6.74061135371179,
+ "grad_norm": 0.01811421848833561,
+ "learning_rate": 0.0006,
+ "loss": 5.647593021392822,
+ "step": 485
+ },
+ {
+ "epoch": 6.754585152838428,
+ "grad_norm": 0.019652029499411583,
+ "learning_rate": 0.0006,
+ "loss": 5.628726959228516,
+ "step": 486
+ },
+ {
+ "epoch": 6.7685589519650655,
+ "grad_norm": 0.02118665538728237,
+ "learning_rate": 0.0006,
+ "loss": 5.6580705642700195,
+ "step": 487
+ },
+ {
+ "epoch": 6.782532751091703,
+ "grad_norm": 0.02237832546234131,
+ "learning_rate": 0.0006,
+ "loss": 5.635931968688965,
+ "step": 488
+ },
+ {
+ "epoch": 6.796506550218341,
+ "grad_norm": 0.023897631093859673,
+ "learning_rate": 0.0006,
+ "loss": 5.517949104309082,
+ "step": 489
+ },
+ {
+ "epoch": 6.810480349344978,
+ "grad_norm": 0.02442227490246296,
+ "learning_rate": 0.0006,
+ "loss": 5.600021839141846,
+ "step": 490
+ },
+ {
+ "epoch": 6.824454148471616,
+ "grad_norm": 0.024675287306308746,
+ "learning_rate": 0.0006,
+ "loss": 5.6639909744262695,
+ "step": 491
+ },
+ {
+ "epoch": 6.8384279475982535,
+ "grad_norm": 0.030372392386198044,
+ "learning_rate": 0.0006,
+ "loss": 5.582888603210449,
+ "step": 492
+ },
+ {
+ "epoch": 6.85240174672489,
+ "grad_norm": 0.03237690031528473,
+ "learning_rate": 0.0006,
+ "loss": 5.573896884918213,
+ "step": 493
+ },
+ {
+ "epoch": 6.866375545851528,
+ "grad_norm": 0.03511589393019676,
+ "learning_rate": 0.0006,
+ "loss": 5.565878868103027,
+ "step": 494
+ },
+ {
+ "epoch": 6.880349344978166,
+ "grad_norm": 0.03883938118815422,
+ "learning_rate": 0.0006,
+ "loss": 5.58540678024292,
+ "step": 495
+ },
+ {
+ "epoch": 6.894323144104804,
+ "grad_norm": 0.04175502806901932,
+ "learning_rate": 0.0006,
+ "loss": 5.6095194816589355,
+ "step": 496
+ },
+ {
+ "epoch": 6.908296943231441,
+ "grad_norm": 0.04013441503047943,
+ "learning_rate": 0.0006,
+ "loss": 5.654401779174805,
+ "step": 497
+ },
+ {
+ "epoch": 6.922270742358078,
+ "grad_norm": 0.04618770629167557,
+ "learning_rate": 0.0006,
+ "loss": 5.6483473777771,
+ "step": 498
+ },
+ {
+ "epoch": 6.936244541484716,
+ "grad_norm": 0.04507308453321457,
+ "learning_rate": 0.0006,
+ "loss": 5.5856218338012695,
+ "step": 499
+ },
+ {
+ "epoch": 6.950218340611354,
+ "grad_norm": 0.03151383996009827,
+ "learning_rate": 0.0006,
+ "loss": 5.594086647033691,
+ "step": 500
+ },
+ {
+ "epoch": 6.964192139737992,
+ "grad_norm": 0.026883797720074654,
+ "learning_rate": 0.0006,
+ "loss": 5.607676029205322,
+ "step": 501
+ },
+ {
+ "epoch": 6.978165938864628,
+ "grad_norm": 0.02981836162507534,
+ "learning_rate": 0.0006,
+ "loss": 5.563666820526123,
+ "step": 502
+ },
+ {
+ "epoch": 6.992139737991266,
+ "grad_norm": 0.03233477845788002,
+ "learning_rate": 0.0006,
+ "loss": 5.667543411254883,
+ "step": 503
+ },
+ {
+ "epoch": 7.0,
+ "grad_norm": 0.029927456751465797,
+ "learning_rate": 0.0006,
+ "loss": 5.733482837677002,
+ "step": 504
+ },
+ {
+ "epoch": 7.0,
+ "eval_loss": 5.679076671600342,
+ "eval_runtime": 59.5965,
+ "eval_samples_per_second": 40.976,
+ "eval_steps_per_second": 1.292,
+ "step": 504
+ },
+ {
+ "epoch": 7.013973799126638,
+ "grad_norm": 0.027081597596406937,
+ "learning_rate": 0.0006,
+ "loss": 5.634487152099609,
+ "step": 505
+ },
+ {
+ "epoch": 7.0279475982532755,
+ "grad_norm": 0.031959421932697296,
+ "learning_rate": 0.0006,
+ "loss": 5.643294811248779,
+ "step": 506
+ },
+ {
+ "epoch": 7.041921397379912,
+ "grad_norm": 0.02802063524723053,
+ "learning_rate": 0.0006,
+ "loss": 5.573239326477051,
+ "step": 507
+ },
+ {
+ "epoch": 7.05589519650655,
+ "grad_norm": 0.032579705119132996,
+ "learning_rate": 0.0006,
+ "loss": 5.627150058746338,
+ "step": 508
+ },
+ {
+ "epoch": 7.069868995633188,
+ "grad_norm": 0.02503928542137146,
+ "learning_rate": 0.0006,
+ "loss": 5.6045308113098145,
+ "step": 509
+ },
+ {
+ "epoch": 7.083842794759826,
+ "grad_norm": 0.02421317622065544,
+ "learning_rate": 0.0006,
+ "loss": 5.522153377532959,
+ "step": 510
+ },
+ {
+ "epoch": 7.097816593886463,
+ "grad_norm": 0.021983426064252853,
+ "learning_rate": 0.0006,
+ "loss": 5.632939338684082,
+ "step": 511
+ },
+ {
+ "epoch": 7.1117903930131,
+ "grad_norm": 0.021933183073997498,
+ "learning_rate": 0.0006,
+ "loss": 5.592185974121094,
+ "step": 512
+ },
+ {
+ "epoch": 7.125764192139738,
+ "grad_norm": 0.02287031151354313,
+ "learning_rate": 0.0006,
+ "loss": 5.613700866699219,
+ "step": 513
+ },
+ {
+ "epoch": 7.139737991266376,
+ "grad_norm": 0.02281602844595909,
+ "learning_rate": 0.0006,
+ "loss": 5.551383018493652,
+ "step": 514
+ },
+ {
+ "epoch": 7.153711790393013,
+ "grad_norm": 0.021050360053777695,
+ "learning_rate": 0.0006,
+ "loss": 5.55275821685791,
+ "step": 515
+ },
+ {
+ "epoch": 7.1676855895196505,
+ "grad_norm": 0.018299926072359085,
+ "learning_rate": 0.0006,
+ "loss": 5.528225421905518,
+ "step": 516
+ },
+ {
+ "epoch": 7.181659388646288,
+ "grad_norm": 0.02024853602051735,
+ "learning_rate": 0.0006,
+ "loss": 5.522453784942627,
+ "step": 517
+ },
+ {
+ "epoch": 7.195633187772926,
+ "grad_norm": 0.02135239914059639,
+ "learning_rate": 0.0006,
+ "loss": 5.527522563934326,
+ "step": 518
+ },
+ {
+ "epoch": 7.209606986899563,
+ "grad_norm": 0.023394184187054634,
+ "learning_rate": 0.0006,
+ "loss": 5.4535017013549805,
+ "step": 519
+ },
+ {
+ "epoch": 7.223580786026201,
+ "grad_norm": 0.025952080264687538,
+ "learning_rate": 0.0006,
+ "loss": 5.5374674797058105,
+ "step": 520
+ },
+ {
+ "epoch": 7.2375545851528384,
+ "grad_norm": 0.022287718951702118,
+ "learning_rate": 0.0006,
+ "loss": 5.493753433227539,
+ "step": 521
+ },
+ {
+ "epoch": 7.251528384279476,
+ "grad_norm": 0.02018447406589985,
+ "learning_rate": 0.0006,
+ "loss": 5.510575294494629,
+ "step": 522
+ },
+ {
+ "epoch": 7.265502183406113,
+ "grad_norm": 0.02050507813692093,
+ "learning_rate": 0.0006,
+ "loss": 5.466026782989502,
+ "step": 523
+ },
+ {
+ "epoch": 7.279475982532751,
+ "grad_norm": 0.023688802495598793,
+ "learning_rate": 0.0006,
+ "loss": 5.485815525054932,
+ "step": 524
+ },
+ {
+ "epoch": 7.293449781659389,
+ "grad_norm": 0.021078843623399734,
+ "learning_rate": 0.0006,
+ "loss": 5.415581703186035,
+ "step": 525
+ },
+ {
+ "epoch": 7.307423580786026,
+ "grad_norm": 0.018266011029481888,
+ "learning_rate": 0.0006,
+ "loss": 5.402815818786621,
+ "step": 526
+ },
+ {
+ "epoch": 7.321397379912664,
+ "grad_norm": 0.019693493843078613,
+ "learning_rate": 0.0006,
+ "loss": 5.504674911499023,
+ "step": 527
+ },
+ {
+ "epoch": 7.335371179039301,
+ "grad_norm": 0.02031373605132103,
+ "learning_rate": 0.0006,
+ "loss": 5.503837585449219,
+ "step": 528
+ },
+ {
+ "epoch": 7.349344978165939,
+ "grad_norm": 0.017636626958847046,
+ "learning_rate": 0.0006,
+ "loss": 5.530580997467041,
+ "step": 529
+ },
+ {
+ "epoch": 7.3633187772925766,
+ "grad_norm": 0.01787244901061058,
+ "learning_rate": 0.0006,
+ "loss": 5.450218677520752,
+ "step": 530
+ },
+ {
+ "epoch": 7.377292576419214,
+ "grad_norm": 0.0170669574290514,
+ "learning_rate": 0.0006,
+ "loss": 5.481570243835449,
+ "step": 531
+ },
+ {
+ "epoch": 7.391266375545851,
+ "grad_norm": 0.01802165061235428,
+ "learning_rate": 0.0006,
+ "loss": 5.386394500732422,
+ "step": 532
+ },
+ {
+ "epoch": 7.405240174672489,
+ "grad_norm": 0.022949472069740295,
+ "learning_rate": 0.0006,
+ "loss": 5.424929618835449,
+ "step": 533
+ },
+ {
+ "epoch": 7.419213973799127,
+ "grad_norm": 0.034213390201330185,
+ "learning_rate": 0.0006,
+ "loss": 5.346663475036621,
+ "step": 534
+ },
+ {
+ "epoch": 7.4331877729257645,
+ "grad_norm": 0.03847593814134598,
+ "learning_rate": 0.0006,
+ "loss": 5.418482780456543,
+ "step": 535
+ },
+ {
+ "epoch": 7.447161572052401,
+ "grad_norm": 0.029393676668405533,
+ "learning_rate": 0.0006,
+ "loss": 5.456090927124023,
+ "step": 536
+ },
+ {
+ "epoch": 7.461135371179039,
+ "grad_norm": 0.03388667106628418,
+ "learning_rate": 0.0006,
+ "loss": 5.572983741760254,
+ "step": 537
+ },
+ {
+ "epoch": 7.475109170305677,
+ "grad_norm": 0.039690613746643066,
+ "learning_rate": 0.0006,
+ "loss": 5.414067268371582,
+ "step": 538
+ },
+ {
+ "epoch": 7.489082969432315,
+ "grad_norm": 0.03630776330828667,
+ "learning_rate": 0.0006,
+ "loss": 5.523739814758301,
+ "step": 539
+ },
+ {
+ "epoch": 7.503056768558952,
+ "grad_norm": 0.03356870263814926,
+ "learning_rate": 0.0006,
+ "loss": 5.444767951965332,
+ "step": 540
+ },
+ {
+ "epoch": 7.517030567685589,
+ "grad_norm": 0.030812304466962814,
+ "learning_rate": 0.0006,
+ "loss": 5.461244106292725,
+ "step": 541
+ },
+ {
+ "epoch": 7.531004366812227,
+ "grad_norm": 0.03319217637181282,
+ "learning_rate": 0.0006,
+ "loss": 5.502161026000977,
+ "step": 542
+ },
+ {
+ "epoch": 7.544978165938865,
+ "grad_norm": 0.032997481524944305,
+ "learning_rate": 0.0006,
+ "loss": 5.4162278175354,
+ "step": 543
+ },
+ {
+ "epoch": 7.558951965065502,
+ "grad_norm": 0.03364962339401245,
+ "learning_rate": 0.0006,
+ "loss": 5.415736198425293,
+ "step": 544
+ },
+ {
+ "epoch": 7.5729257641921395,
+ "grad_norm": 0.0344221405684948,
+ "learning_rate": 0.0006,
+ "loss": 5.541967868804932,
+ "step": 545
+ },
+ {
+ "epoch": 7.586899563318777,
+ "grad_norm": 0.029609503224492073,
+ "learning_rate": 0.0006,
+ "loss": 5.440614223480225,
+ "step": 546
+ },
+ {
+ "epoch": 7.600873362445415,
+ "grad_norm": 0.029055926948785782,
+ "learning_rate": 0.0006,
+ "loss": 5.462865352630615,
+ "step": 547
+ },
+ {
+ "epoch": 7.614847161572053,
+ "grad_norm": 0.02658848837018013,
+ "learning_rate": 0.0006,
+ "loss": 5.444467544555664,
+ "step": 548
+ },
+ {
+ "epoch": 7.62882096069869,
+ "grad_norm": 0.026276404038071632,
+ "learning_rate": 0.0006,
+ "loss": 5.462278842926025,
+ "step": 549
+ },
+ {
+ "epoch": 7.642794759825327,
+ "grad_norm": 0.0286889486014843,
+ "learning_rate": 0.0006,
+ "loss": 5.4849534034729,
+ "step": 550
+ },
+ {
+ "epoch": 7.656768558951965,
+ "grad_norm": 0.026849817484617233,
+ "learning_rate": 0.0006,
+ "loss": 5.431473731994629,
+ "step": 551
+ },
+ {
+ "epoch": 7.670742358078603,
+ "grad_norm": 0.02312396466732025,
+ "learning_rate": 0.0006,
+ "loss": 5.45462703704834,
+ "step": 552
+ },
+ {
+ "epoch": 7.68471615720524,
+ "grad_norm": 0.026197485625743866,
+ "learning_rate": 0.0006,
+ "loss": 5.452552318572998,
+ "step": 553
+ },
+ {
+ "epoch": 7.698689956331878,
+ "grad_norm": 0.02747255191206932,
+ "learning_rate": 0.0006,
+ "loss": 5.447169780731201,
+ "step": 554
+ },
+ {
+ "epoch": 7.712663755458515,
+ "grad_norm": 0.028123432770371437,
+ "learning_rate": 0.0006,
+ "loss": 5.394400596618652,
+ "step": 555
+ },
+ {
+ "epoch": 7.726637554585153,
+ "grad_norm": 0.02599870041012764,
+ "learning_rate": 0.0006,
+ "loss": 5.398341178894043,
+ "step": 556
+ },
+ {
+ "epoch": 7.74061135371179,
+ "grad_norm": 0.022171657532453537,
+ "learning_rate": 0.0006,
+ "loss": 5.368820667266846,
+ "step": 557
+ },
+ {
+ "epoch": 7.754585152838428,
+ "grad_norm": 0.022309480234980583,
+ "learning_rate": 0.0006,
+ "loss": 5.416772365570068,
+ "step": 558
+ },
+ {
+ "epoch": 7.7685589519650655,
+ "grad_norm": 0.024074165150523186,
+ "learning_rate": 0.0006,
+ "loss": 5.359002113342285,
+ "step": 559
+ },
+ {
+ "epoch": 7.782532751091703,
+ "grad_norm": 0.02636653557419777,
+ "learning_rate": 0.0006,
+ "loss": 5.484368324279785,
+ "step": 560
+ },
+ {
+ "epoch": 7.796506550218341,
+ "grad_norm": 0.02203752100467682,
+ "learning_rate": 0.0006,
+ "loss": 5.375498294830322,
+ "step": 561
+ },
+ {
+ "epoch": 7.810480349344978,
+ "grad_norm": 0.019924819469451904,
+ "learning_rate": 0.0006,
+ "loss": 5.334723472595215,
+ "step": 562
+ },
+ {
+ "epoch": 7.824454148471616,
+ "grad_norm": 0.018755966797471046,
+ "learning_rate": 0.0006,
+ "loss": 5.348781108856201,
+ "step": 563
+ },
+ {
+ "epoch": 7.8384279475982535,
+ "grad_norm": 0.016747845336794853,
+ "learning_rate": 0.0006,
+ "loss": 5.433079719543457,
+ "step": 564
+ },
+ {
+ "epoch": 7.85240174672489,
+ "grad_norm": 0.0170395877212286,
+ "learning_rate": 0.0006,
+ "loss": 5.346794128417969,
+ "step": 565
+ },
+ {
+ "epoch": 7.866375545851528,
+ "grad_norm": 0.016074176877737045,
+ "learning_rate": 0.0006,
+ "loss": 5.3277740478515625,
+ "step": 566
+ },
+ {
+ "epoch": 7.880349344978166,
+ "grad_norm": 0.014119806699454784,
+ "learning_rate": 0.0006,
+ "loss": 5.483551979064941,
+ "step": 567
+ },
+ {
+ "epoch": 7.894323144104804,
+ "grad_norm": 0.015270394273102283,
+ "learning_rate": 0.0006,
+ "loss": 5.378519535064697,
+ "step": 568
+ },
+ {
+ "epoch": 7.908296943231441,
+ "grad_norm": 0.015677539631724358,
+ "learning_rate": 0.0006,
+ "loss": 5.290210247039795,
+ "step": 569
+ },
+ {
+ "epoch": 7.922270742358078,
+ "grad_norm": 0.015930423513054848,
+ "learning_rate": 0.0006,
+ "loss": 5.427360534667969,
+ "step": 570
+ },
+ {
+ "epoch": 7.936244541484716,
+ "grad_norm": 0.016641564667224884,
+ "learning_rate": 0.0006,
+ "loss": 5.301599979400635,
+ "step": 571
+ },
+ {
+ "epoch": 7.950218340611354,
+ "grad_norm": 0.018293552100658417,
+ "learning_rate": 0.0006,
+ "loss": 5.277889251708984,
+ "step": 572
+ },
+ {
+ "epoch": 7.964192139737992,
+ "grad_norm": 0.017618799582123756,
+ "learning_rate": 0.0006,
+ "loss": 5.297194480895996,
+ "step": 573
+ },
+ {
+ "epoch": 7.978165938864628,
+ "grad_norm": 0.015034242533147335,
+ "learning_rate": 0.0006,
+ "loss": 5.42428731918335,
+ "step": 574
+ },
+ {
+ "epoch": 7.992139737991266,
+ "grad_norm": 0.01617511734366417,
+ "learning_rate": 0.0006,
+ "loss": 5.420044898986816,
+ "step": 575
+ },
+ {
+ "epoch": 8.0,
+ "grad_norm": 0.01837257295846939,
+ "learning_rate": 0.0006,
+ "loss": 5.240387916564941,
+ "step": 576
+ },
+ {
+ "epoch": 8.0,
+ "eval_loss": 5.423932075500488,
+ "eval_runtime": 58.0841,
+ "eval_samples_per_second": 42.042,
+ "eval_steps_per_second": 1.326,
+ "step": 576
+ },
+ {
+ "epoch": 8.013973799126637,
+ "grad_norm": 0.020953809842467308,
+ "learning_rate": 0.0006,
+ "loss": 5.4320783615112305,
+ "step": 577
+ },
+ {
+ "epoch": 8.027947598253276,
+ "grad_norm": 0.0227745920419693,
+ "learning_rate": 0.0006,
+ "loss": 5.310197830200195,
+ "step": 578
+ },
+ {
+ "epoch": 8.041921397379912,
+ "grad_norm": 0.021372603252530098,
+ "learning_rate": 0.0006,
+ "loss": 5.339468002319336,
+ "step": 579
+ },
+ {
+ "epoch": 8.055895196506551,
+ "grad_norm": 0.022401731461286545,
+ "learning_rate": 0.0006,
+ "loss": 5.279829978942871,
+ "step": 580
+ },
+ {
+ "epoch": 8.069868995633188,
+ "grad_norm": 0.023942379280924797,
+ "learning_rate": 0.0006,
+ "loss": 5.372323989868164,
+ "step": 581
+ },
+ {
+ "epoch": 8.083842794759825,
+ "grad_norm": 0.024380534887313843,
+ "learning_rate": 0.0006,
+ "loss": 5.253486633300781,
+ "step": 582
+ },
+ {
+ "epoch": 8.097816593886463,
+ "grad_norm": 0.03246625140309334,
+ "learning_rate": 0.0006,
+ "loss": 5.305774211883545,
+ "step": 583
+ },
+ {
+ "epoch": 8.1117903930131,
+ "grad_norm": 0.030399736016988754,
+ "learning_rate": 0.0006,
+ "loss": 5.370976448059082,
+ "step": 584
+ },
+ {
+ "epoch": 8.125764192139737,
+ "grad_norm": 0.027148913592100143,
+ "learning_rate": 0.0006,
+ "loss": 5.368010520935059,
+ "step": 585
+ },
+ {
+ "epoch": 8.139737991266376,
+ "grad_norm": 0.02925540879368782,
+ "learning_rate": 0.0006,
+ "loss": 5.303013801574707,
+ "step": 586
+ },
+ {
+ "epoch": 8.153711790393013,
+ "grad_norm": 0.02915453538298607,
+ "learning_rate": 0.0006,
+ "loss": 5.329667091369629,
+ "step": 587
+ },
+ {
+ "epoch": 8.167685589519651,
+ "grad_norm": 0.031732626259326935,
+ "learning_rate": 0.0006,
+ "loss": 5.3634934425354,
+ "step": 588
+ },
+ {
+ "epoch": 8.181659388646288,
+ "grad_norm": 0.036166731268167496,
+ "learning_rate": 0.0006,
+ "loss": 5.223980903625488,
+ "step": 589
+ },
+ {
+ "epoch": 8.195633187772925,
+ "grad_norm": 0.035654108971357346,
+ "learning_rate": 0.0006,
+ "loss": 5.361059188842773,
+ "step": 590
+ },
+ {
+ "epoch": 8.209606986899564,
+ "grad_norm": 0.0365324430167675,
+ "learning_rate": 0.0006,
+ "loss": 5.336151599884033,
+ "step": 591
+ },
+ {
+ "epoch": 8.2235807860262,
+ "grad_norm": 0.03578880429267883,
+ "learning_rate": 0.0006,
+ "loss": 5.434076309204102,
+ "step": 592
+ },
+ {
+ "epoch": 8.237554585152838,
+ "grad_norm": 0.035734012722969055,
+ "learning_rate": 0.0006,
+ "loss": 5.330893516540527,
+ "step": 593
+ },
+ {
+ "epoch": 8.251528384279476,
+ "grad_norm": 0.03196857124567032,
+ "learning_rate": 0.0006,
+ "loss": 5.306009769439697,
+ "step": 594
+ },
+ {
+ "epoch": 8.265502183406113,
+ "grad_norm": 0.032042086124420166,
+ "learning_rate": 0.0006,
+ "loss": 5.343267917633057,
+ "step": 595
+ },
+ {
+ "epoch": 8.279475982532752,
+ "grad_norm": 0.03160746395587921,
+ "learning_rate": 0.0006,
+ "loss": 5.2353715896606445,
+ "step": 596
+ },
+ {
+ "epoch": 8.293449781659389,
+ "grad_norm": 0.026689818128943443,
+ "learning_rate": 0.0006,
+ "loss": 5.254042625427246,
+ "step": 597
+ },
+ {
+ "epoch": 8.307423580786025,
+ "grad_norm": 0.02880188450217247,
+ "learning_rate": 0.0006,
+ "loss": 5.326833248138428,
+ "step": 598
+ },
+ {
+ "epoch": 8.321397379912664,
+ "grad_norm": 0.027516381815075874,
+ "learning_rate": 0.0006,
+ "loss": 5.342336177825928,
+ "step": 599
+ },
+ {
+ "epoch": 8.335371179039301,
+ "grad_norm": 0.027875030413269997,
+ "learning_rate": 0.0006,
+ "loss": 5.200719833374023,
+ "step": 600
+ },
+ {
+ "epoch": 8.34934497816594,
+ "grad_norm": 0.0268265288323164,
+ "learning_rate": 0.0006,
+ "loss": 5.249449729919434,
+ "step": 601
+ },
+ {
+ "epoch": 8.363318777292577,
+ "grad_norm": 0.024824608117341995,
+ "learning_rate": 0.0006,
+ "loss": 5.318952560424805,
+ "step": 602
+ },
+ {
+ "epoch": 8.377292576419213,
+ "grad_norm": 0.022990627214312553,
+ "learning_rate": 0.0006,
+ "loss": 5.244993209838867,
+ "step": 603
+ },
+ {
+ "epoch": 8.391266375545852,
+ "grad_norm": 0.022804604843258858,
+ "learning_rate": 0.0006,
+ "loss": 5.3236846923828125,
+ "step": 604
+ },
+ {
+ "epoch": 8.405240174672489,
+ "grad_norm": 0.02419872209429741,
+ "learning_rate": 0.0006,
+ "loss": 5.255486488342285,
+ "step": 605
+ },
+ {
+ "epoch": 8.419213973799126,
+ "grad_norm": 0.021952059119939804,
+ "learning_rate": 0.0006,
+ "loss": 5.3511552810668945,
+ "step": 606
+ },
+ {
+ "epoch": 8.433187772925764,
+ "grad_norm": 0.022375497967004776,
+ "learning_rate": 0.0006,
+ "loss": 5.294790267944336,
+ "step": 607
+ },
+ {
+ "epoch": 8.447161572052401,
+ "grad_norm": 0.019242815673351288,
+ "learning_rate": 0.0006,
+ "loss": 5.252618789672852,
+ "step": 608
+ },
+ {
+ "epoch": 8.46113537117904,
+ "grad_norm": 0.0173486340790987,
+ "learning_rate": 0.0006,
+ "loss": 5.30147647857666,
+ "step": 609
+ },
+ {
+ "epoch": 8.475109170305677,
+ "grad_norm": 0.01612604409456253,
+ "learning_rate": 0.0006,
+ "loss": 5.23960542678833,
+ "step": 610
+ },
+ {
+ "epoch": 8.489082969432314,
+ "grad_norm": 0.014293976128101349,
+ "learning_rate": 0.0006,
+ "loss": 5.119932651519775,
+ "step": 611
+ },
+ {
+ "epoch": 8.503056768558952,
+ "grad_norm": 0.014453536830842495,
+ "learning_rate": 0.0006,
+ "loss": 5.166794776916504,
+ "step": 612
+ },
+ {
+ "epoch": 8.51703056768559,
+ "grad_norm": 0.014685769565403461,
+ "learning_rate": 0.0006,
+ "loss": 5.264822959899902,
+ "step": 613
+ },
+ {
+ "epoch": 8.531004366812226,
+ "grad_norm": 0.01427740603685379,
+ "learning_rate": 0.0006,
+ "loss": 5.278433799743652,
+ "step": 614
+ },
+ {
+ "epoch": 8.544978165938865,
+ "grad_norm": 0.014864951372146606,
+ "learning_rate": 0.0006,
+ "loss": 5.214512348175049,
+ "step": 615
+ },
+ {
+ "epoch": 8.558951965065502,
+ "grad_norm": 0.01565164513885975,
+ "learning_rate": 0.0006,
+ "loss": 5.183066368103027,
+ "step": 616
+ },
+ {
+ "epoch": 8.57292576419214,
+ "grad_norm": 0.01779816672205925,
+ "learning_rate": 0.0006,
+ "loss": 5.163185119628906,
+ "step": 617
+ },
+ {
+ "epoch": 8.586899563318777,
+ "grad_norm": 0.017254827544093132,
+ "learning_rate": 0.0006,
+ "loss": 5.243555068969727,
+ "step": 618
+ },
+ {
+ "epoch": 8.600873362445414,
+ "grad_norm": 0.01650584116578102,
+ "learning_rate": 0.0006,
+ "loss": 5.234884738922119,
+ "step": 619
+ },
+ {
+ "epoch": 8.614847161572053,
+ "grad_norm": 0.017021115869283676,
+ "learning_rate": 0.0006,
+ "loss": 5.12271785736084,
+ "step": 620
+ },
+ {
+ "epoch": 8.62882096069869,
+ "grad_norm": 0.01773759163916111,
+ "learning_rate": 0.0006,
+ "loss": 5.242153167724609,
+ "step": 621
+ },
+ {
+ "epoch": 8.642794759825328,
+ "grad_norm": 0.015679042786359787,
+ "learning_rate": 0.0006,
+ "loss": 5.170779228210449,
+ "step": 622
+ },
+ {
+ "epoch": 8.656768558951965,
+ "grad_norm": 0.013760549947619438,
+ "learning_rate": 0.0006,
+ "loss": 5.238644599914551,
+ "step": 623
+ },
+ {
+ "epoch": 8.670742358078602,
+ "grad_norm": 0.014571522362530231,
+ "learning_rate": 0.0006,
+ "loss": 5.177056312561035,
+ "step": 624
+ },
+ {
+ "epoch": 8.68471615720524,
+ "grad_norm": 0.016209015622735023,
+ "learning_rate": 0.0006,
+ "loss": 5.163750648498535,
+ "step": 625
+ },
+ {
+ "epoch": 8.698689956331878,
+ "grad_norm": 0.016813941299915314,
+ "learning_rate": 0.0006,
+ "loss": 5.183428764343262,
+ "step": 626
+ },
+ {
+ "epoch": 8.712663755458514,
+ "grad_norm": 0.020985357463359833,
+ "learning_rate": 0.0006,
+ "loss": 5.229465007781982,
+ "step": 627
+ },
+ {
+ "epoch": 8.726637554585153,
+ "grad_norm": 0.02679632417857647,
+ "learning_rate": 0.0006,
+ "loss": 5.257368087768555,
+ "step": 628
+ },
+ {
+ "epoch": 8.74061135371179,
+ "grad_norm": 0.025756070390343666,
+ "learning_rate": 0.0006,
+ "loss": 5.253736972808838,
+ "step": 629
+ },
+ {
+ "epoch": 8.754585152838429,
+ "grad_norm": 0.02643490768969059,
+ "learning_rate": 0.0006,
+ "loss": 5.228633403778076,
+ "step": 630
+ },
+ {
+ "epoch": 8.768558951965066,
+ "grad_norm": 0.03303210437297821,
+ "learning_rate": 0.0006,
+ "loss": 5.163540363311768,
+ "step": 631
+ },
+ {
+ "epoch": 8.782532751091702,
+ "grad_norm": 0.030432431027293205,
+ "learning_rate": 0.0006,
+ "loss": 5.243169784545898,
+ "step": 632
+ },
+ {
+ "epoch": 8.796506550218341,
+ "grad_norm": 0.028631288558244705,
+ "learning_rate": 0.0006,
+ "loss": 5.166398048400879,
+ "step": 633
+ },
+ {
+ "epoch": 8.810480349344978,
+ "grad_norm": 0.026188310235738754,
+ "learning_rate": 0.0006,
+ "loss": 5.128122329711914,
+ "step": 634
+ },
+ {
+ "epoch": 8.824454148471617,
+ "grad_norm": 0.025308528915047646,
+ "learning_rate": 0.0006,
+ "loss": 5.211284637451172,
+ "step": 635
+ },
+ {
+ "epoch": 8.838427947598253,
+ "grad_norm": 0.020657729357481003,
+ "learning_rate": 0.0006,
+ "loss": 5.161575794219971,
+ "step": 636
+ },
+ {
+ "epoch": 8.85240174672489,
+ "grad_norm": 0.021901234984397888,
+ "learning_rate": 0.0006,
+ "loss": 5.20050048828125,
+ "step": 637
+ },
+ {
+ "epoch": 8.866375545851529,
+ "grad_norm": 0.021219318732619286,
+ "learning_rate": 0.0006,
+ "loss": 5.211699485778809,
+ "step": 638
+ },
+ {
+ "epoch": 8.880349344978166,
+ "grad_norm": 0.0200974028557539,
+ "learning_rate": 0.0006,
+ "loss": 5.092503547668457,
+ "step": 639
+ },
+ {
+ "epoch": 8.894323144104803,
+ "grad_norm": 0.023804882541298866,
+ "learning_rate": 0.0006,
+ "loss": 5.216068267822266,
+ "step": 640
+ },
+ {
+ "epoch": 8.908296943231441,
+ "grad_norm": 0.026088010519742966,
+ "learning_rate": 0.0006,
+ "loss": 5.155592918395996,
+ "step": 641
+ },
+ {
+ "epoch": 8.922270742358078,
+ "grad_norm": 0.02601276896893978,
+ "learning_rate": 0.0006,
+ "loss": 5.197238922119141,
+ "step": 642
+ },
+ {
+ "epoch": 8.936244541484717,
+ "grad_norm": 0.020387211814522743,
+ "learning_rate": 0.0006,
+ "loss": 5.1955437660217285,
+ "step": 643
+ },
+ {
+ "epoch": 8.950218340611354,
+ "grad_norm": 0.019214622676372528,
+ "learning_rate": 0.0006,
+ "loss": 5.206346035003662,
+ "step": 644
+ },
+ {
+ "epoch": 8.96419213973799,
+ "grad_norm": 0.019674314185976982,
+ "learning_rate": 0.0006,
+ "loss": 5.219857215881348,
+ "step": 645
+ },
+ {
+ "epoch": 8.97816593886463,
+ "grad_norm": 0.020318234339356422,
+ "learning_rate": 0.0006,
+ "loss": 5.191132545471191,
+ "step": 646
+ },
+ {
+ "epoch": 8.992139737991266,
+ "grad_norm": 0.021428676322102547,
+ "learning_rate": 0.0006,
+ "loss": 5.214695930480957,
+ "step": 647
+ },
+ {
+ "epoch": 9.0,
+ "grad_norm": 0.023235054686665535,
+ "learning_rate": 0.0006,
+ "loss": 5.229091644287109,
+ "step": 648
+ },
+ {
+ "epoch": 9.0,
+ "eval_loss": 5.289593696594238,
+ "eval_runtime": 56.9256,
+ "eval_samples_per_second": 42.898,
+ "eval_steps_per_second": 1.353,
+ "step": 648
+ },
+ {
+ "epoch": 9.013973799126637,
+ "grad_norm": 0.02320289984345436,
+ "learning_rate": 0.0006,
+ "loss": 4.983669281005859,
+ "step": 649
+ },
+ {
+ "epoch": 9.027947598253276,
+ "grad_norm": 0.018521282821893692,
+ "learning_rate": 0.0006,
+ "loss": 5.200575828552246,
+ "step": 650
+ },
+ {
+ "epoch": 9.041921397379912,
+ "grad_norm": 0.0180169939994812,
+ "learning_rate": 0.0006,
+ "loss": 5.148033618927002,
+ "step": 651
+ },
+ {
+ "epoch": 9.055895196506551,
+ "grad_norm": 0.019969860091805458,
+ "learning_rate": 0.0006,
+ "loss": 5.099125385284424,
+ "step": 652
+ },
+ {
+ "epoch": 9.069868995633188,
+ "grad_norm": 0.01643305830657482,
+ "learning_rate": 0.0006,
+ "loss": 5.092347145080566,
+ "step": 653
+ },
+ {
+ "epoch": 9.083842794759825,
+ "grad_norm": 0.016022363677620888,
+ "learning_rate": 0.0006,
+ "loss": 5.129279613494873,
+ "step": 654
+ },
+ {
+ "epoch": 9.097816593886463,
+ "grad_norm": 0.01657041162252426,
+ "learning_rate": 0.0006,
+ "loss": 5.159562587738037,
+ "step": 655
+ },
+ {
+ "epoch": 9.1117903930131,
+ "grad_norm": 0.017924228683114052,
+ "learning_rate": 0.0006,
+ "loss": 5.165497779846191,
+ "step": 656
+ },
+ {
+ "epoch": 9.125764192139737,
+ "grad_norm": 0.021592361852526665,
+ "learning_rate": 0.0006,
+ "loss": 5.204775810241699,
+ "step": 657
+ },
+ {
+ "epoch": 9.139737991266376,
+ "grad_norm": 0.026924120262265205,
+ "learning_rate": 0.0006,
+ "loss": 5.128296852111816,
+ "step": 658
+ },
+ {
+ "epoch": 9.153711790393013,
+ "grad_norm": 0.023156899958848953,
+ "learning_rate": 0.0006,
+ "loss": 5.1255340576171875,
+ "step": 659
+ },
+ {
+ "epoch": 9.167685589519651,
+ "grad_norm": 0.020670367404818535,
+ "learning_rate": 0.0006,
+ "loss": 5.128479957580566,
+ "step": 660
+ },
+ {
+ "epoch": 9.181659388646288,
+ "grad_norm": 0.025447173044085503,
+ "learning_rate": 0.0006,
+ "loss": 5.154821872711182,
+ "step": 661
+ },
+ {
+ "epoch": 9.195633187772925,
+ "grad_norm": 0.027940068393945694,
+ "learning_rate": 0.0006,
+ "loss": 5.143466472625732,
+ "step": 662
+ },
+ {
+ "epoch": 9.209606986899564,
+ "grad_norm": 0.026847844943404198,
+ "learning_rate": 0.0006,
+ "loss": 5.045665740966797,
+ "step": 663
+ },
+ {
+ "epoch": 9.2235807860262,
+ "grad_norm": 0.02340601570904255,
+ "learning_rate": 0.0006,
+ "loss": 5.21933126449585,
+ "step": 664
+ },
+ {
+ "epoch": 9.237554585152838,
+ "grad_norm": 0.02340371161699295,
+ "learning_rate": 0.0006,
+ "loss": 4.942949295043945,
+ "step": 665
+ },
+ {
+ "epoch": 9.251528384279476,
+ "grad_norm": 0.02221992425620556,
+ "learning_rate": 0.0006,
+ "loss": 5.197099208831787,
+ "step": 666
+ },
+ {
+ "epoch": 9.265502183406113,
+ "grad_norm": 0.023508677259087563,
+ "learning_rate": 0.0006,
+ "loss": 5.022897720336914,
+ "step": 667
+ },
+ {
+ "epoch": 9.279475982532752,
+ "grad_norm": 0.026120394468307495,
+ "learning_rate": 0.0006,
+ "loss": 5.087725639343262,
+ "step": 668
+ },
+ {
+ "epoch": 9.293449781659389,
+ "grad_norm": 0.027273228392004967,
+ "learning_rate": 0.0006,
+ "loss": 5.090963840484619,
+ "step": 669
+ },
+ {
+ "epoch": 9.307423580786025,
+ "grad_norm": 0.03241586685180664,
+ "learning_rate": 0.0006,
+ "loss": 5.047842979431152,
+ "step": 670
+ },
+ {
+ "epoch": 9.321397379912664,
+ "grad_norm": 0.030245667323470116,
+ "learning_rate": 0.0006,
+ "loss": 5.104803085327148,
+ "step": 671
+ },
+ {
+ "epoch": 9.335371179039301,
+ "grad_norm": 0.027698364108800888,
+ "learning_rate": 0.0006,
+ "loss": 5.0644097328186035,
+ "step": 672
+ },
+ {
+ "epoch": 9.34934497816594,
+ "grad_norm": 0.029692554846405983,
+ "learning_rate": 0.0006,
+ "loss": 5.111942291259766,
+ "step": 673
+ },
+ {
+ "epoch": 9.363318777292577,
+ "grad_norm": 0.03328656405210495,
+ "learning_rate": 0.0006,
+ "loss": 5.1574554443359375,
+ "step": 674
+ },
+ {
+ "epoch": 9.377292576419213,
+ "grad_norm": 0.031938180327415466,
+ "learning_rate": 0.0006,
+ "loss": 5.190371990203857,
+ "step": 675
+ },
+ {
+ "epoch": 9.391266375545852,
+ "grad_norm": 0.024858945980668068,
+ "learning_rate": 0.0006,
+ "loss": 5.121420860290527,
+ "step": 676
+ },
+ {
+ "epoch": 9.405240174672489,
+ "grad_norm": 0.023033203557133675,
+ "learning_rate": 0.0006,
+ "loss": 5.0974345207214355,
+ "step": 677
+ },
+ {
+ "epoch": 9.419213973799126,
+ "grad_norm": 0.022083545103669167,
+ "learning_rate": 0.0006,
+ "loss": 5.046319961547852,
+ "step": 678
+ },
+ {
+ "epoch": 9.433187772925764,
+ "grad_norm": 0.0228437427431345,
+ "learning_rate": 0.0006,
+ "loss": 5.119963645935059,
+ "step": 679
+ },
+ {
+ "epoch": 9.447161572052401,
+ "grad_norm": 0.0242351982742548,
+ "learning_rate": 0.0006,
+ "loss": 5.0444560050964355,
+ "step": 680
+ },
+ {
+ "epoch": 9.46113537117904,
+ "grad_norm": 0.02401185780763626,
+ "learning_rate": 0.0006,
+ "loss": 5.063238620758057,
+ "step": 681
+ },
+ {
+ "epoch": 9.475109170305677,
+ "grad_norm": 0.02560959942638874,
+ "learning_rate": 0.0006,
+ "loss": 5.035224914550781,
+ "step": 682
+ },
+ {
+ "epoch": 9.489082969432314,
+ "grad_norm": 0.024985626339912415,
+ "learning_rate": 0.0006,
+ "loss": 5.049284934997559,
+ "step": 683
+ },
+ {
+ "epoch": 9.503056768558952,
+ "grad_norm": 0.023178115487098694,
+ "learning_rate": 0.0006,
+ "loss": 5.2266621589660645,
+ "step": 684
+ },
+ {
+ "epoch": 9.51703056768559,
+ "grad_norm": 0.0215512253344059,
+ "learning_rate": 0.0006,
+ "loss": 5.1875715255737305,
+ "step": 685
+ },
+ {
+ "epoch": 9.531004366812226,
+ "grad_norm": 0.01662837713956833,
+ "learning_rate": 0.0006,
+ "loss": 5.137012481689453,
+ "step": 686
+ },
+ {
+ "epoch": 9.544978165938865,
+ "grad_norm": 0.01952764391899109,
+ "learning_rate": 0.0006,
+ "loss": 5.072500228881836,
+ "step": 687
+ },
+ {
+ "epoch": 9.558951965065502,
+ "grad_norm": 0.017521077767014503,
+ "learning_rate": 0.0006,
+ "loss": 5.045711994171143,
+ "step": 688
+ },
+ {
+ "epoch": 9.57292576419214,
+ "grad_norm": 0.017355265095829964,
+ "learning_rate": 0.0006,
+ "loss": 5.02881383895874,
+ "step": 689
+ },
+ {
+ "epoch": 9.586899563318777,
+ "grad_norm": 0.015587719157338142,
+ "learning_rate": 0.0006,
+ "loss": 5.132787704467773,
+ "step": 690
+ },
+ {
+ "epoch": 9.600873362445414,
+ "grad_norm": 0.017469845712184906,
+ "learning_rate": 0.0006,
+ "loss": 5.105447292327881,
+ "step": 691
+ },
+ {
+ "epoch": 9.614847161572053,
+ "grad_norm": 0.019630100578069687,
+ "learning_rate": 0.0006,
+ "loss": 5.032464981079102,
+ "step": 692
+ },
+ {
+ "epoch": 9.62882096069869,
+ "grad_norm": 0.01784994639456272,
+ "learning_rate": 0.0006,
+ "loss": 5.081345558166504,
+ "step": 693
+ },
+ {
+ "epoch": 9.642794759825328,
+ "grad_norm": 0.01872754842042923,
+ "learning_rate": 0.0006,
+ "loss": 4.9772114753723145,
+ "step": 694
+ },
+ {
+ "epoch": 9.656768558951965,
+ "grad_norm": 0.01999032311141491,
+ "learning_rate": 0.0006,
+ "loss": 5.089740753173828,
+ "step": 695
+ },
+ {
+ "epoch": 9.670742358078602,
+ "grad_norm": 0.01710767112672329,
+ "learning_rate": 0.0006,
+ "loss": 4.9869890213012695,
+ "step": 696
+ },
+ {
+ "epoch": 9.68471615720524,
+ "grad_norm": 0.015144161880016327,
+ "learning_rate": 0.0006,
+ "loss": 5.022564888000488,
+ "step": 697
+ },
+ {
+ "epoch": 9.698689956331878,
+ "grad_norm": 0.01618543453514576,
+ "learning_rate": 0.0006,
+ "loss": 5.035921573638916,
+ "step": 698
+ },
+ {
+ "epoch": 9.712663755458514,
+ "grad_norm": 0.014653387479484081,
+ "learning_rate": 0.0006,
+ "loss": 4.914456367492676,
+ "step": 699
+ },
+ {
+ "epoch": 9.726637554585153,
+ "grad_norm": 0.013434977270662785,
+ "learning_rate": 0.0006,
+ "loss": 5.063764572143555,
+ "step": 700
+ },
+ {
+ "epoch": 9.74061135371179,
+ "grad_norm": 0.015128864906728268,
+ "learning_rate": 0.0006,
+ "loss": 5.052367687225342,
+ "step": 701
+ },
+ {
+ "epoch": 9.754585152838429,
+ "grad_norm": 0.014210768043994904,
+ "learning_rate": 0.0006,
+ "loss": 5.016489028930664,
+ "step": 702
+ },
+ {
+ "epoch": 9.768558951965066,
+ "grad_norm": 0.015191680751740932,
+ "learning_rate": 0.0006,
+ "loss": 4.990176200866699,
+ "step": 703
+ },
+ {
+ "epoch": 9.782532751091702,
+ "grad_norm": 0.018285127356648445,
+ "learning_rate": 0.0006,
+ "loss": 5.015164375305176,
+ "step": 704
+ },
+ {
+ "epoch": 9.796506550218341,
+ "grad_norm": 0.0204079058021307,
+ "learning_rate": 0.0006,
+ "loss": 5.091142654418945,
+ "step": 705
+ },
+ {
+ "epoch": 9.810480349344978,
+ "grad_norm": 0.01980586163699627,
+ "learning_rate": 0.0006,
+ "loss": 5.03350830078125,
+ "step": 706
+ },
+ {
+ "epoch": 9.824454148471617,
+ "grad_norm": 0.02048729732632637,
+ "learning_rate": 0.0006,
+ "loss": 5.03611421585083,
+ "step": 707
+ },
+ {
+ "epoch": 9.838427947598253,
+ "grad_norm": 0.02058163844048977,
+ "learning_rate": 0.0006,
+ "loss": 4.9613447189331055,
+ "step": 708
+ },
+ {
+ "epoch": 9.85240174672489,
+ "grad_norm": 0.018004924058914185,
+ "learning_rate": 0.0006,
+ "loss": 4.966541767120361,
+ "step": 709
+ },
+ {
+ "epoch": 9.866375545851529,
+ "grad_norm": 0.0198152307420969,
+ "learning_rate": 0.0006,
+ "loss": 5.01750373840332,
+ "step": 710
+ },
+ {
+ "epoch": 9.880349344978166,
+ "grad_norm": 0.018532969057559967,
+ "learning_rate": 0.0006,
+ "loss": 5.097424507141113,
+ "step": 711
+ },
+ {
+ "epoch": 9.894323144104803,
+ "grad_norm": 0.018939971923828125,
+ "learning_rate": 0.0006,
+ "loss": 5.177223205566406,
+ "step": 712
+ },
+ {
+ "epoch": 9.908296943231441,
+ "grad_norm": 0.017969170585274696,
+ "learning_rate": 0.0006,
+ "loss": 5.118013858795166,
+ "step": 713
+ },
+ {
+ "epoch": 9.922270742358078,
+ "grad_norm": 0.018021270632743835,
+ "learning_rate": 0.0006,
+ "loss": 5.034787178039551,
+ "step": 714
+ },
+ {
+ "epoch": 9.936244541484717,
+ "grad_norm": 0.017009710893034935,
+ "learning_rate": 0.0006,
+ "loss": 5.003920555114746,
+ "step": 715
+ },
+ {
+ "epoch": 9.950218340611354,
+ "grad_norm": 0.017722049728035927,
+ "learning_rate": 0.0006,
+ "loss": 5.031939506530762,
+ "step": 716
+ },
+ {
+ "epoch": 9.96419213973799,
+ "grad_norm": 0.017134087160229683,
+ "learning_rate": 0.0006,
+ "loss": 5.093474388122559,
+ "step": 717
+ },
+ {
+ "epoch": 9.97816593886463,
+ "grad_norm": 0.018511613830924034,
+ "learning_rate": 0.0006,
+ "loss": 5.078604698181152,
+ "step": 718
+ },
+ {
+ "epoch": 9.992139737991266,
+ "grad_norm": 0.021206334233283997,
+ "learning_rate": 0.0006,
+ "loss": 5.052572727203369,
+ "step": 719
+ },
+ {
+ "epoch": 10.0,
+ "grad_norm": 0.02239062264561653,
+ "learning_rate": 0.0006,
+ "loss": 5.063547134399414,
+ "step": 720
+ },
+ {
+ "epoch": 10.0,
+ "eval_loss": 5.173044681549072,
+ "eval_runtime": 56.7287,
+ "eval_samples_per_second": 43.047,
+ "eval_steps_per_second": 1.357,
+ "step": 720
+ },
+ {
+ "epoch": 10.013973799126637,
+ "grad_norm": 0.020757969468832016,
+ "learning_rate": 0.0006,
+ "loss": 5.045648574829102,
+ "step": 721
+ },
+ {
+ "epoch": 10.027947598253276,
+ "grad_norm": 0.022697702050209045,
+ "learning_rate": 0.0006,
+ "loss": 4.993833065032959,
+ "step": 722
+ },
+ {
+ "epoch": 10.041921397379912,
+ "grad_norm": 0.02434331737458706,
+ "learning_rate": 0.0006,
+ "loss": 5.073219299316406,
+ "step": 723
+ },
+ {
+ "epoch": 10.055895196506551,
+ "grad_norm": 0.025499660521745682,
+ "learning_rate": 0.0006,
+ "loss": 5.019072532653809,
+ "step": 724
+ },
+ {
+ "epoch": 10.069868995633188,
+ "grad_norm": 0.029281053692102432,
+ "learning_rate": 0.0006,
+ "loss": 4.959851264953613,
+ "step": 725
+ },
+ {
+ "epoch": 10.083842794759825,
+ "grad_norm": 0.033878110349178314,
+ "learning_rate": 0.0006,
+ "loss": 5.132717609405518,
+ "step": 726
+ },
+ {
+ "epoch": 10.097816593886463,
+ "grad_norm": 0.03385764732956886,
+ "learning_rate": 0.0006,
+ "loss": 5.0482892990112305,
+ "step": 727
+ },
+ {
+ "epoch": 10.1117903930131,
+ "grad_norm": 0.03466648980975151,
+ "learning_rate": 0.0006,
+ "loss": 4.946855545043945,
+ "step": 728
+ },
+ {
+ "epoch": 10.125764192139737,
+ "grad_norm": 0.028546305373311043,
+ "learning_rate": 0.0006,
+ "loss": 5.027774333953857,
+ "step": 729
+ },
+ {
+ "epoch": 10.139737991266376,
+ "grad_norm": 0.030575869604945183,
+ "learning_rate": 0.0006,
+ "loss": 4.9768476486206055,
+ "step": 730
+ },
+ {
+ "epoch": 10.153711790393013,
+ "grad_norm": 0.03614303469657898,
+ "learning_rate": 0.0006,
+ "loss": 5.078113555908203,
+ "step": 731
+ },
+ {
+ "epoch": 10.167685589519651,
+ "grad_norm": 0.040991660207509995,
+ "learning_rate": 0.0006,
+ "loss": 5.033185958862305,
+ "step": 732
+ },
+ {
+ "epoch": 10.181659388646288,
+ "grad_norm": 0.04551699757575989,
+ "learning_rate": 0.0006,
+ "loss": 5.0820746421813965,
+ "step": 733
+ },
+ {
+ "epoch": 10.195633187772925,
+ "grad_norm": 0.040193989872932434,
+ "learning_rate": 0.0006,
+ "loss": 5.059700965881348,
+ "step": 734
+ },
+ {
+ "epoch": 10.209606986899564,
+ "grad_norm": 0.035851605236530304,
+ "learning_rate": 0.0006,
+ "loss": 5.142467498779297,
+ "step": 735
+ },
+ {
+ "epoch": 10.2235807860262,
+ "grad_norm": 0.034200169146060944,
+ "learning_rate": 0.0006,
+ "loss": 5.072312355041504,
+ "step": 736
+ },
+ {
+ "epoch": 10.237554585152838,
+ "grad_norm": 0.033621110022068024,
+ "learning_rate": 0.0006,
+ "loss": 5.08037805557251,
+ "step": 737
+ },
+ {
+ "epoch": 10.251528384279476,
+ "grad_norm": 0.037372369319200516,
+ "learning_rate": 0.0006,
+ "loss": 4.998425483703613,
+ "step": 738
+ },
+ {
+ "epoch": 10.265502183406113,
+ "grad_norm": 0.037044707685709,
+ "learning_rate": 0.0006,
+ "loss": 4.923993110656738,
+ "step": 739
+ },
+ {
+ "epoch": 10.279475982532752,
+ "grad_norm": 0.030454808846116066,
+ "learning_rate": 0.0006,
+ "loss": 5.0889997482299805,
+ "step": 740
+ },
+ {
+ "epoch": 10.293449781659389,
+ "grad_norm": 0.030969638377428055,
+ "learning_rate": 0.0006,
+ "loss": 4.984112739562988,
+ "step": 741
+ },
+ {
+ "epoch": 10.307423580786025,
+ "grad_norm": 0.0276983380317688,
+ "learning_rate": 0.0006,
+ "loss": 4.955412864685059,
+ "step": 742
+ },
+ {
+ "epoch": 10.321397379912664,
+ "grad_norm": 0.02716052532196045,
+ "learning_rate": 0.0006,
+ "loss": 5.0098876953125,
+ "step": 743
+ },
+ {
+ "epoch": 10.335371179039301,
+ "grad_norm": 0.02403552085161209,
+ "learning_rate": 0.0006,
+ "loss": 5.165160655975342,
+ "step": 744
+ },
+ {
+ "epoch": 10.34934497816594,
+ "grad_norm": 0.0257862601429224,
+ "learning_rate": 0.0006,
+ "loss": 5.001547813415527,
+ "step": 745
+ },
+ {
+ "epoch": 10.363318777292577,
+ "grad_norm": 0.02300378680229187,
+ "learning_rate": 0.0006,
+ "loss": 5.038888931274414,
+ "step": 746
+ },
+ {
+ "epoch": 10.377292576419213,
+ "grad_norm": 0.019766854122281075,
+ "learning_rate": 0.0006,
+ "loss": 4.924291610717773,
+ "step": 747
+ },
+ {
+ "epoch": 10.391266375545852,
+ "grad_norm": 0.01888395845890045,
+ "learning_rate": 0.0006,
+ "loss": 5.059971332550049,
+ "step": 748
+ },
+ {
+ "epoch": 10.405240174672489,
+ "grad_norm": 0.015577499754726887,
+ "learning_rate": 0.0006,
+ "loss": 4.987686634063721,
+ "step": 749
+ },
+ {
+ "epoch": 10.419213973799126,
+ "grad_norm": 0.016012346372008324,
+ "learning_rate": 0.0006,
+ "loss": 5.015157699584961,
+ "step": 750
+ },
+ {
+ "epoch": 10.433187772925764,
+ "grad_norm": 0.016277998685836792,
+ "learning_rate": 0.0006,
+ "loss": 5.004931449890137,
+ "step": 751
+ },
+ {
+ "epoch": 10.447161572052401,
+ "grad_norm": 0.017504561692476273,
+ "learning_rate": 0.0006,
+ "loss": 4.902827739715576,
+ "step": 752
+ },
+ {
+ "epoch": 10.46113537117904,
+ "grad_norm": 0.01569022797048092,
+ "learning_rate": 0.0006,
+ "loss": 4.954863548278809,
+ "step": 753
+ },
+ {
+ "epoch": 10.475109170305677,
+ "grad_norm": 0.014114422723650932,
+ "learning_rate": 0.0006,
+ "loss": 5.00656795501709,
+ "step": 754
+ },
+ {
+ "epoch": 10.489082969432314,
+ "grad_norm": 0.01573832333087921,
+ "learning_rate": 0.0006,
+ "loss": 4.978353023529053,
+ "step": 755
+ },
+ {
+ "epoch": 10.503056768558952,
+ "grad_norm": 0.016480036079883575,
+ "learning_rate": 0.0006,
+ "loss": 4.953327655792236,
+ "step": 756
+ },
+ {
+ "epoch": 10.51703056768559,
+ "grad_norm": 0.014008025638759136,
+ "learning_rate": 0.0006,
+ "loss": 5.050840377807617,
+ "step": 757
+ },
+ {
+ "epoch": 10.531004366812226,
+ "grad_norm": 0.013147998601198196,
+ "learning_rate": 0.0006,
+ "loss": 4.974964141845703,
+ "step": 758
+ },
+ {
+ "epoch": 10.544978165938865,
+ "grad_norm": 0.013807603158056736,
+ "learning_rate": 0.0006,
+ "loss": 4.927907943725586,
+ "step": 759
+ },
+ {
+ "epoch": 10.558951965065502,
+ "grad_norm": 0.013555224984884262,
+ "learning_rate": 0.0006,
+ "loss": 4.980690956115723,
+ "step": 760
+ },
+ {
+ "epoch": 10.57292576419214,
+ "grad_norm": 0.013045408762991428,
+ "learning_rate": 0.0006,
+ "loss": 4.9436116218566895,
+ "step": 761
+ },
+ {
+ "epoch": 10.586899563318777,
+ "grad_norm": 0.012858862057328224,
+ "learning_rate": 0.0006,
+ "loss": 4.9595770835876465,
+ "step": 762
+ },
+ {
+ "epoch": 10.600873362445414,
+ "grad_norm": 0.01468253880739212,
+ "learning_rate": 0.0006,
+ "loss": 4.90725040435791,
+ "step": 763
+ },
+ {
+ "epoch": 10.614847161572053,
+ "grad_norm": 0.013224679045379162,
+ "learning_rate": 0.0006,
+ "loss": 4.877615928649902,
+ "step": 764
+ },
+ {
+ "epoch": 10.62882096069869,
+ "grad_norm": 0.013555348850786686,
+ "learning_rate": 0.0006,
+ "loss": 4.9456610679626465,
+ "step": 765
+ },
+ {
+ "epoch": 10.642794759825328,
+ "grad_norm": 0.014387160539627075,
+ "learning_rate": 0.0006,
+ "loss": 4.9233717918396,
+ "step": 766
+ },
+ {
+ "epoch": 10.656768558951965,
+ "grad_norm": 0.014354088343679905,
+ "learning_rate": 0.0006,
+ "loss": 4.983511924743652,
+ "step": 767
+ },
+ {
+ "epoch": 10.670742358078602,
+ "grad_norm": 0.01332700252532959,
+ "learning_rate": 0.0006,
+ "loss": 4.934416770935059,
+ "step": 768
+ },
+ {
+ "epoch": 10.68471615720524,
+ "grad_norm": 0.011543313041329384,
+ "learning_rate": 0.0006,
+ "loss": 4.96388053894043,
+ "step": 769
+ },
+ {
+ "epoch": 10.698689956331878,
+ "grad_norm": 0.012330878525972366,
+ "learning_rate": 0.0006,
+ "loss": 4.946396350860596,
+ "step": 770
+ },
+ {
+ "epoch": 10.712663755458514,
+ "grad_norm": 0.01379779726266861,
+ "learning_rate": 0.0006,
+ "loss": 4.981019020080566,
+ "step": 771
+ },
+ {
+ "epoch": 10.726637554585153,
+ "grad_norm": 0.014612431637942791,
+ "learning_rate": 0.0006,
+ "loss": 4.894567966461182,
+ "step": 772
+ },
+ {
+ "epoch": 10.74061135371179,
+ "grad_norm": 0.013062899932265282,
+ "learning_rate": 0.0006,
+ "loss": 5.033038139343262,
+ "step": 773
+ },
+ {
+ "epoch": 10.754585152838429,
+ "grad_norm": 0.013898049481213093,
+ "learning_rate": 0.0006,
+ "loss": 4.823636054992676,
+ "step": 774
+ },
+ {
+ "epoch": 10.768558951965066,
+ "grad_norm": 0.01502019353210926,
+ "learning_rate": 0.0006,
+ "loss": 5.020156383514404,
+ "step": 775
+ },
+ {
+ "epoch": 10.782532751091702,
+ "grad_norm": 0.013810204342007637,
+ "learning_rate": 0.0006,
+ "loss": 4.935647010803223,
+ "step": 776
+ },
+ {
+ "epoch": 10.796506550218341,
+ "grad_norm": 0.013616513460874557,
+ "learning_rate": 0.0006,
+ "loss": 4.901486396789551,
+ "step": 777
+ },
+ {
+ "epoch": 10.810480349344978,
+ "grad_norm": 0.013601024635136127,
+ "learning_rate": 0.0006,
+ "loss": 4.931873321533203,
+ "step": 778
+ },
+ {
+ "epoch": 10.824454148471617,
+ "grad_norm": 0.013668350875377655,
+ "learning_rate": 0.0006,
+ "loss": 4.937256813049316,
+ "step": 779
+ },
+ {
+ "epoch": 10.838427947598253,
+ "grad_norm": 0.01581823080778122,
+ "learning_rate": 0.0006,
+ "loss": 4.9611406326293945,
+ "step": 780
+ },
+ {
+ "epoch": 10.85240174672489,
+ "grad_norm": 0.019478484988212585,
+ "learning_rate": 0.0006,
+ "loss": 4.9212236404418945,
+ "step": 781
+ },
+ {
+ "epoch": 10.866375545851529,
+ "grad_norm": 0.026921333745121956,
+ "learning_rate": 0.0006,
+ "loss": 4.88501501083374,
+ "step": 782
+ },
+ {
+ "epoch": 10.880349344978166,
+ "grad_norm": 0.03155914694070816,
+ "learning_rate": 0.0006,
+ "loss": 4.989967346191406,
+ "step": 783
+ },
+ {
+ "epoch": 10.894323144104803,
+ "grad_norm": 0.029689429327845573,
+ "learning_rate": 0.0006,
+ "loss": 4.936000823974609,
+ "step": 784
+ },
+ {
+ "epoch": 10.908296943231441,
+ "grad_norm": 0.026626063510775566,
+ "learning_rate": 0.0006,
+ "loss": 4.984785079956055,
+ "step": 785
+ },
+ {
+ "epoch": 10.922270742358078,
+ "grad_norm": 0.026168212294578552,
+ "learning_rate": 0.0006,
+ "loss": 4.908282279968262,
+ "step": 786
+ },
+ {
+ "epoch": 10.936244541484717,
+ "grad_norm": 0.02828158251941204,
+ "learning_rate": 0.0006,
+ "loss": 4.926390647888184,
+ "step": 787
+ },
+ {
+ "epoch": 10.950218340611354,
+ "grad_norm": 0.02649078331887722,
+ "learning_rate": 0.0006,
+ "loss": 5.00400972366333,
+ "step": 788
+ },
+ {
+ "epoch": 10.96419213973799,
+ "grad_norm": 0.0256856270134449,
+ "learning_rate": 0.0006,
+ "loss": 5.0317487716674805,
+ "step": 789
+ },
+ {
+ "epoch": 10.97816593886463,
+ "grad_norm": 0.024728883057832718,
+ "learning_rate": 0.0006,
+ "loss": 4.9299635887146,
+ "step": 790
+ },
+ {
+ "epoch": 10.992139737991266,
+ "grad_norm": 0.026241116225719452,
+ "learning_rate": 0.0006,
+ "loss": 4.947319984436035,
+ "step": 791
+ },
+ {
+ "epoch": 11.0,
+ "grad_norm": 0.023526743054389954,
+ "learning_rate": 0.0006,
+ "loss": 4.94429349899292,
+ "step": 792
+ },
+ {
+ "epoch": 11.0,
+ "eval_loss": 5.104345798492432,
+ "eval_runtime": 57.0955,
+ "eval_samples_per_second": 42.77,
+ "eval_steps_per_second": 1.349,
+ "step": 792
+ },
+ {
+ "epoch": 11.013973799126637,
+ "grad_norm": 0.0216788612306118,
+ "learning_rate": 0.0006,
+ "loss": 4.881950855255127,
+ "step": 793
+ },
+ {
+ "epoch": 11.027947598253276,
+ "grad_norm": 0.01830463856458664,
+ "learning_rate": 0.0006,
+ "loss": 4.971016883850098,
+ "step": 794
+ },
+ {
+ "epoch": 11.041921397379912,
+ "grad_norm": 0.02204521745443344,
+ "learning_rate": 0.0006,
+ "loss": 4.894623756408691,
+ "step": 795
+ },
+ {
+ "epoch": 11.055895196506551,
+ "grad_norm": 0.0210255216807127,
+ "learning_rate": 0.0006,
+ "loss": 4.834009170532227,
+ "step": 796
+ },
+ {
+ "epoch": 11.069868995633188,
+ "grad_norm": 0.019797448068857193,
+ "learning_rate": 0.0006,
+ "loss": 5.086714744567871,
+ "step": 797
+ },
+ {
+ "epoch": 11.083842794759825,
+ "grad_norm": 0.02036641724407673,
+ "learning_rate": 0.0006,
+ "loss": 4.920056343078613,
+ "step": 798
+ },
+ {
+ "epoch": 11.097816593886463,
+ "grad_norm": 0.01966066285967827,
+ "learning_rate": 0.0006,
+ "loss": 4.879859447479248,
+ "step": 799
+ },
+ {
+ "epoch": 11.1117903930131,
+ "grad_norm": 0.0202149897813797,
+ "learning_rate": 0.0006,
+ "loss": 4.933640480041504,
+ "step": 800
+ },
+ {
+ "epoch": 11.125764192139737,
+ "grad_norm": 0.01893039606511593,
+ "learning_rate": 0.0006,
+ "loss": 4.901970863342285,
+ "step": 801
+ },
+ {
+ "epoch": 11.139737991266376,
+ "grad_norm": 0.016040155664086342,
+ "learning_rate": 0.0006,
+ "loss": 4.905491828918457,
+ "step": 802
+ },
+ {
+ "epoch": 11.153711790393013,
+ "grad_norm": 0.01720350608229637,
+ "learning_rate": 0.0006,
+ "loss": 4.798816680908203,
+ "step": 803
+ },
+ {
+ "epoch": 11.167685589519651,
+ "grad_norm": 0.017002852633595467,
+ "learning_rate": 0.0006,
+ "loss": 4.851859092712402,
+ "step": 804
+ },
+ {
+ "epoch": 11.181659388646288,
+ "grad_norm": 0.015594680793583393,
+ "learning_rate": 0.0006,
+ "loss": 4.90328311920166,
+ "step": 805
+ },
+ {
+ "epoch": 11.195633187772925,
+ "grad_norm": 0.014904462732374668,
+ "learning_rate": 0.0006,
+ "loss": 4.904215335845947,
+ "step": 806
+ },
+ {
+ "epoch": 11.209606986899564,
+ "grad_norm": 0.01366228237748146,
+ "learning_rate": 0.0006,
+ "loss": 4.9813055992126465,
+ "step": 807
+ },
+ {
+ "epoch": 11.2235807860262,
+ "grad_norm": 0.014719638973474503,
+ "learning_rate": 0.0006,
+ "loss": 4.927306175231934,
+ "step": 808
+ },
+ {
+ "epoch": 11.237554585152838,
+ "grad_norm": 0.015016036108136177,
+ "learning_rate": 0.0006,
+ "loss": 4.874215126037598,
+ "step": 809
+ },
+ {
+ "epoch": 11.251528384279476,
+ "grad_norm": 0.014925424009561539,
+ "learning_rate": 0.0006,
+ "loss": 4.838257789611816,
+ "step": 810
+ },
+ {
+ "epoch": 11.265502183406113,
+ "grad_norm": 0.014335823245346546,
+ "learning_rate": 0.0006,
+ "loss": 4.940608024597168,
+ "step": 811
+ },
+ {
+ "epoch": 11.279475982532752,
+ "grad_norm": 0.014987512491643429,
+ "learning_rate": 0.0006,
+ "loss": 4.831531524658203,
+ "step": 812
+ },
+ {
+ "epoch": 11.293449781659389,
+ "grad_norm": 0.016087457537651062,
+ "learning_rate": 0.0006,
+ "loss": 4.863227367401123,
+ "step": 813
+ },
+ {
+ "epoch": 11.307423580786025,
+ "grad_norm": 0.01671300269663334,
+ "learning_rate": 0.0006,
+ "loss": 4.8535661697387695,
+ "step": 814
+ },
+ {
+ "epoch": 11.321397379912664,
+ "grad_norm": 0.015229142270982265,
+ "learning_rate": 0.0006,
+ "loss": 4.948635101318359,
+ "step": 815
+ },
+ {
+ "epoch": 11.335371179039301,
+ "grad_norm": 0.01532587967813015,
+ "learning_rate": 0.0006,
+ "loss": 4.9035797119140625,
+ "step": 816
+ },
+ {
+ "epoch": 11.34934497816594,
+ "grad_norm": 0.015715476125478745,
+ "learning_rate": 0.0006,
+ "loss": 4.894709587097168,
+ "step": 817
+ },
+ {
+ "epoch": 11.363318777292577,
+ "grad_norm": 0.014156977646052837,
+ "learning_rate": 0.0006,
+ "loss": 4.960206985473633,
+ "step": 818
+ },
+ {
+ "epoch": 11.377292576419213,
+ "grad_norm": 0.017076566815376282,
+ "learning_rate": 0.0006,
+ "loss": 5.005850791931152,
+ "step": 819
+ },
+ {
+ "epoch": 11.391266375545852,
+ "grad_norm": 0.0172084029763937,
+ "learning_rate": 0.0006,
+ "loss": 4.903355598449707,
+ "step": 820
+ },
+ {
+ "epoch": 11.405240174672489,
+ "grad_norm": 0.0180258359760046,
+ "learning_rate": 0.0006,
+ "loss": 4.842667579650879,
+ "step": 821
+ },
+ {
+ "epoch": 11.419213973799126,
+ "grad_norm": 0.022130120545625687,
+ "learning_rate": 0.0006,
+ "loss": 4.799970626831055,
+ "step": 822
+ },
+ {
+ "epoch": 11.433187772925764,
+ "grad_norm": 0.02384897693991661,
+ "learning_rate": 0.0006,
+ "loss": 4.860746383666992,
+ "step": 823
+ },
+ {
+ "epoch": 11.447161572052401,
+ "grad_norm": 0.021881572902202606,
+ "learning_rate": 0.0006,
+ "loss": 4.918107032775879,
+ "step": 824
+ },
+ {
+ "epoch": 11.46113537117904,
+ "grad_norm": 0.023085080087184906,
+ "learning_rate": 0.0006,
+ "loss": 4.878946304321289,
+ "step": 825
+ },
+ {
+ "epoch": 11.475109170305677,
+ "grad_norm": 0.02291012555360794,
+ "learning_rate": 0.0006,
+ "loss": 4.958887100219727,
+ "step": 826
+ },
+ {
+ "epoch": 11.489082969432314,
+ "grad_norm": 0.0199968870729208,
+ "learning_rate": 0.0006,
+ "loss": 4.80955696105957,
+ "step": 827
+ },
+ {
+ "epoch": 11.503056768558952,
+ "grad_norm": 0.018702229484915733,
+ "learning_rate": 0.0006,
+ "loss": 5.0276408195495605,
+ "step": 828
+ },
+ {
+ "epoch": 11.51703056768559,
+ "grad_norm": 0.017787311226129532,
+ "learning_rate": 0.0006,
+ "loss": 4.96018123626709,
+ "step": 829
+ },
+ {
+ "epoch": 11.531004366812226,
+ "grad_norm": 0.01704501546919346,
+ "learning_rate": 0.0006,
+ "loss": 4.914553165435791,
+ "step": 830
+ },
+ {
+ "epoch": 11.544978165938865,
+ "grad_norm": 0.017519677057862282,
+ "learning_rate": 0.0006,
+ "loss": 4.798098087310791,
+ "step": 831
+ },
+ {
+ "epoch": 11.558951965065502,
+ "grad_norm": 0.018511991947889328,
+ "learning_rate": 0.0006,
+ "loss": 4.817785263061523,
+ "step": 832
+ },
+ {
+ "epoch": 11.57292576419214,
+ "grad_norm": 0.01876644790172577,
+ "learning_rate": 0.0006,
+ "loss": 4.818995475769043,
+ "step": 833
+ },
+ {
+ "epoch": 11.586899563318777,
+ "grad_norm": 0.01753568835556507,
+ "learning_rate": 0.0006,
+ "loss": 4.811938285827637,
+ "step": 834
+ },
+ {
+ "epoch": 11.600873362445414,
+ "grad_norm": 0.017046531662344933,
+ "learning_rate": 0.0006,
+ "loss": 4.914674282073975,
+ "step": 835
+ },
+ {
+ "epoch": 11.614847161572053,
+ "grad_norm": 0.01802094094455242,
+ "learning_rate": 0.0006,
+ "loss": 4.993284225463867,
+ "step": 836
+ },
+ {
+ "epoch": 11.62882096069869,
+ "grad_norm": 0.019493764266371727,
+ "learning_rate": 0.0006,
+ "loss": 4.895628452301025,
+ "step": 837
+ },
+ {
+ "epoch": 11.642794759825328,
+ "grad_norm": 0.019461210817098618,
+ "learning_rate": 0.0006,
+ "loss": 4.869269371032715,
+ "step": 838
+ },
+ {
+ "epoch": 11.656768558951965,
+ "grad_norm": 0.01981567218899727,
+ "learning_rate": 0.0006,
+ "loss": 4.863052845001221,
+ "step": 839
+ },
+ {
+ "epoch": 11.670742358078602,
+ "grad_norm": 0.019894491881132126,
+ "learning_rate": 0.0006,
+ "loss": 4.8775787353515625,
+ "step": 840
+ },
+ {
+ "epoch": 11.68471615720524,
+ "grad_norm": 0.01875786855816841,
+ "learning_rate": 0.0006,
+ "loss": 4.817763328552246,
+ "step": 841
+ },
+ {
+ "epoch": 11.698689956331878,
+ "grad_norm": 0.020926695317029953,
+ "learning_rate": 0.0006,
+ "loss": 4.7470526695251465,
+ "step": 842
+ },
+ {
+ "epoch": 11.712663755458514,
+ "grad_norm": 0.021477103233337402,
+ "learning_rate": 0.0006,
+ "loss": 4.762682914733887,
+ "step": 843
+ },
+ {
+ "epoch": 11.726637554585153,
+ "grad_norm": 0.020628711208701134,
+ "learning_rate": 0.0006,
+ "loss": 4.807985782623291,
+ "step": 844
+ },
+ {
+ "epoch": 11.74061135371179,
+ "grad_norm": 0.021489612758159637,
+ "learning_rate": 0.0006,
+ "loss": 4.843091011047363,
+ "step": 845
+ },
+ {
+ "epoch": 11.754585152838429,
+ "grad_norm": 0.021194491535425186,
+ "learning_rate": 0.0006,
+ "loss": 4.8300251960754395,
+ "step": 846
+ },
+ {
+ "epoch": 11.768558951965066,
+ "grad_norm": 0.01833350397646427,
+ "learning_rate": 0.0006,
+ "loss": 5.000202178955078,
+ "step": 847
+ },
+ {
+ "epoch": 11.782532751091702,
+ "grad_norm": 0.018096931278705597,
+ "learning_rate": 0.0006,
+ "loss": 4.880648136138916,
+ "step": 848
+ },
+ {
+ "epoch": 11.796506550218341,
+ "grad_norm": 0.0197161752730608,
+ "learning_rate": 0.0006,
+ "loss": 4.788316249847412,
+ "step": 849
+ },
+ {
+ "epoch": 11.810480349344978,
+ "grad_norm": 0.0203181654214859,
+ "learning_rate": 0.0006,
+ "loss": 4.870944976806641,
+ "step": 850
+ },
+ {
+ "epoch": 11.824454148471617,
+ "grad_norm": 0.01908455230295658,
+ "learning_rate": 0.0006,
+ "loss": 4.890144348144531,
+ "step": 851
+ },
+ {
+ "epoch": 11.838427947598253,
+ "grad_norm": 0.021281801164150238,
+ "learning_rate": 0.0006,
+ "loss": 4.89578914642334,
+ "step": 852
+ },
+ {
+ "epoch": 11.85240174672489,
+ "grad_norm": 0.02124273031949997,
+ "learning_rate": 0.0006,
+ "loss": 4.883749008178711,
+ "step": 853
+ },
+ {
+ "epoch": 11.866375545851529,
+ "grad_norm": 0.020969398319721222,
+ "learning_rate": 0.0006,
+ "loss": 4.809142589569092,
+ "step": 854
+ },
+ {
+ "epoch": 11.880349344978166,
+ "grad_norm": 0.01967657171189785,
+ "learning_rate": 0.0006,
+ "loss": 4.821864604949951,
+ "step": 855
+ },
+ {
+ "epoch": 11.894323144104803,
+ "grad_norm": 0.020887333899736404,
+ "learning_rate": 0.0006,
+ "loss": 4.917985916137695,
+ "step": 856
+ },
+ {
+ "epoch": 11.908296943231441,
+ "grad_norm": 0.019203083589673042,
+ "learning_rate": 0.0006,
+ "loss": 4.824960708618164,
+ "step": 857
+ },
+ {
+ "epoch": 11.922270742358078,
+ "grad_norm": 0.020162401720881462,
+ "learning_rate": 0.0006,
+ "loss": 4.772617816925049,
+ "step": 858
+ },
+ {
+ "epoch": 11.936244541484717,
+ "grad_norm": 0.02251487225294113,
+ "learning_rate": 0.0006,
+ "loss": 4.735418319702148,
+ "step": 859
+ },
+ {
+ "epoch": 11.950218340611354,
+ "grad_norm": 0.020313527435064316,
+ "learning_rate": 0.0006,
+ "loss": 4.939233779907227,
+ "step": 860
+ },
+ {
+ "epoch": 11.96419213973799,
+ "grad_norm": 0.021635599434375763,
+ "learning_rate": 0.0006,
+ "loss": 4.861807346343994,
+ "step": 861
+ },
+ {
+ "epoch": 11.97816593886463,
+ "grad_norm": 0.02285311557352543,
+ "learning_rate": 0.0006,
+ "loss": 4.932732582092285,
+ "step": 862
+ },
+ {
+ "epoch": 11.992139737991266,
+ "grad_norm": 0.02072535641491413,
+ "learning_rate": 0.0006,
+ "loss": 4.796767711639404,
+ "step": 863
+ },
+ {
+ "epoch": 12.0,
+ "grad_norm": 0.020588304847478867,
+ "learning_rate": 0.0006,
+ "loss": 4.997537612915039,
+ "step": 864
+ },
+ {
+ "epoch": 12.0,
+ "eval_loss": 5.044651985168457,
+ "eval_runtime": 56.9145,
+ "eval_samples_per_second": 42.906,
+ "eval_steps_per_second": 1.353,
+ "step": 864
+ },
+ {
+ "epoch": 12.013973799126637,
+ "grad_norm": 0.017597166821360588,
+ "learning_rate": 0.0006,
+ "loss": 4.830524444580078,
+ "step": 865
+ },
+ {
+ "epoch": 12.027947598253276,
+ "grad_norm": 0.01716788113117218,
+ "learning_rate": 0.0006,
+ "loss": 4.894325256347656,
+ "step": 866
+ },
+ {
+ "epoch": 12.041921397379912,
+ "grad_norm": 0.017712457105517387,
+ "learning_rate": 0.0006,
+ "loss": 4.868297576904297,
+ "step": 867
+ },
+ {
+ "epoch": 12.055895196506551,
+ "grad_norm": 0.021486390382051468,
+ "learning_rate": 0.0006,
+ "loss": 4.922049522399902,
+ "step": 868
+ },
+ {
+ "epoch": 12.069868995633188,
+ "grad_norm": 0.024560654535889626,
+ "learning_rate": 0.0006,
+ "loss": 4.734241008758545,
+ "step": 869
+ },
+ {
+ "epoch": 12.083842794759825,
+ "grad_norm": 0.025870252400636673,
+ "learning_rate": 0.0006,
+ "loss": 4.868566989898682,
+ "step": 870
+ },
+ {
+ "epoch": 12.097816593886463,
+ "grad_norm": 0.0277389045804739,
+ "learning_rate": 0.0006,
+ "loss": 4.755006790161133,
+ "step": 871
+ },
+ {
+ "epoch": 12.1117903930131,
+ "grad_norm": 0.029100844636559486,
+ "learning_rate": 0.0006,
+ "loss": 4.909667015075684,
+ "step": 872
+ },
+ {
+ "epoch": 12.125764192139737,
+ "grad_norm": 0.031833067536354065,
+ "learning_rate": 0.0006,
+ "loss": 4.858674049377441,
+ "step": 873
+ },
+ {
+ "epoch": 12.139737991266376,
+ "grad_norm": 0.032709237188100815,
+ "learning_rate": 0.0006,
+ "loss": 4.823696136474609,
+ "step": 874
+ },
+ {
+ "epoch": 12.153711790393013,
+ "grad_norm": 0.02995380386710167,
+ "learning_rate": 0.0006,
+ "loss": 4.850235462188721,
+ "step": 875
+ },
+ {
+ "epoch": 12.167685589519651,
+ "grad_norm": 0.034098610281944275,
+ "learning_rate": 0.0006,
+ "loss": 4.8169989585876465,
+ "step": 876
+ },
+ {
+ "epoch": 12.181659388646288,
+ "grad_norm": 0.03377080336213112,
+ "learning_rate": 0.0006,
+ "loss": 4.90738582611084,
+ "step": 877
+ },
+ {
+ "epoch": 12.195633187772925,
+ "grad_norm": 0.029780825600028038,
+ "learning_rate": 0.0006,
+ "loss": 4.811163902282715,
+ "step": 878
+ },
+ {
+ "epoch": 12.209606986899564,
+ "grad_norm": 0.03118024580180645,
+ "learning_rate": 0.0006,
+ "loss": 4.861286640167236,
+ "step": 879
+ },
+ {
+ "epoch": 12.2235807860262,
+ "grad_norm": 0.033052004873752594,
+ "learning_rate": 0.0006,
+ "loss": 4.865021228790283,
+ "step": 880
+ },
+ {
+ "epoch": 12.237554585152838,
+ "grad_norm": 0.029070032760500908,
+ "learning_rate": 0.0006,
+ "loss": 4.947303771972656,
+ "step": 881
+ },
+ {
+ "epoch": 12.251528384279476,
+ "grad_norm": 0.028355715796351433,
+ "learning_rate": 0.0006,
+ "loss": 4.804648399353027,
+ "step": 882
+ },
+ {
+ "epoch": 12.265502183406113,
+ "grad_norm": 0.031257227063179016,
+ "learning_rate": 0.0006,
+ "loss": 4.757718086242676,
+ "step": 883
+ },
+ {
+ "epoch": 12.279475982532752,
+ "grad_norm": 0.031438373029232025,
+ "learning_rate": 0.0006,
+ "loss": 4.851268768310547,
+ "step": 884
+ },
+ {
+ "epoch": 12.293449781659389,
+ "grad_norm": 0.030492138117551804,
+ "learning_rate": 0.0006,
+ "loss": 4.7891035079956055,
+ "step": 885
+ },
+ {
+ "epoch": 12.307423580786025,
+ "grad_norm": 0.026115991175174713,
+ "learning_rate": 0.0006,
+ "loss": 4.86474609375,
+ "step": 886
+ },
+ {
+ "epoch": 12.321397379912664,
+ "grad_norm": 0.024184465408325195,
+ "learning_rate": 0.0006,
+ "loss": 4.860456943511963,
+ "step": 887
+ },
+ {
+ "epoch": 12.335371179039301,
+ "grad_norm": 0.025437982752919197,
+ "learning_rate": 0.0006,
+ "loss": 4.806251049041748,
+ "step": 888
+ },
+ {
+ "epoch": 12.34934497816594,
+ "grad_norm": 0.022390231490135193,
+ "learning_rate": 0.0006,
+ "loss": 4.807086944580078,
+ "step": 889
+ },
+ {
+ "epoch": 12.363318777292577,
+ "grad_norm": 0.020174294710159302,
+ "learning_rate": 0.0006,
+ "loss": 4.78292179107666,
+ "step": 890
+ },
+ {
+ "epoch": 12.377292576419213,
+ "grad_norm": 0.019734324887394905,
+ "learning_rate": 0.0006,
+ "loss": 4.784509658813477,
+ "step": 891
+ },
+ {
+ "epoch": 12.391266375545852,
+ "grad_norm": 0.016733523458242416,
+ "learning_rate": 0.0006,
+ "loss": 4.811755180358887,
+ "step": 892
+ },
+ {
+ "epoch": 12.405240174672489,
+ "grad_norm": 0.01570146717131138,
+ "learning_rate": 0.0006,
+ "loss": 4.720088005065918,
+ "step": 893
+ },
+ {
+ "epoch": 12.419213973799126,
+ "grad_norm": 0.01601393334567547,
+ "learning_rate": 0.0006,
+ "loss": 4.783543109893799,
+ "step": 894
+ },
+ {
+ "epoch": 12.433187772925764,
+ "grad_norm": 0.014035487547516823,
+ "learning_rate": 0.0006,
+ "loss": 4.848340034484863,
+ "step": 895
+ },
+ {
+ "epoch": 12.447161572052401,
+ "grad_norm": 0.015248487703502178,
+ "learning_rate": 0.0006,
+ "loss": 4.773287773132324,
+ "step": 896
+ },
+ {
+ "epoch": 12.46113537117904,
+ "grad_norm": 0.01532268151640892,
+ "learning_rate": 0.0006,
+ "loss": 4.799646854400635,
+ "step": 897
+ },
+ {
+ "epoch": 12.475109170305677,
+ "grad_norm": 0.013982534408569336,
+ "learning_rate": 0.0006,
+ "loss": 4.8090500831604,
+ "step": 898
+ },
+ {
+ "epoch": 12.489082969432314,
+ "grad_norm": 0.014081164263188839,
+ "learning_rate": 0.0006,
+ "loss": 4.751888751983643,
+ "step": 899
+ },
+ {
+ "epoch": 12.503056768558952,
+ "grad_norm": 0.01380133256316185,
+ "learning_rate": 0.0006,
+ "loss": 4.78144645690918,
+ "step": 900
+ },
+ {
+ "epoch": 12.51703056768559,
+ "grad_norm": 0.013555269688367844,
+ "learning_rate": 0.0006,
+ "loss": 4.779257774353027,
+ "step": 901
+ },
+ {
+ "epoch": 12.531004366812226,
+ "grad_norm": 0.013075701892375946,
+ "learning_rate": 0.0006,
+ "loss": 4.820705413818359,
+ "step": 902
+ },
+ {
+ "epoch": 12.544978165938865,
+ "grad_norm": 0.01276316400617361,
+ "learning_rate": 0.0006,
+ "loss": 4.836733818054199,
+ "step": 903
+ },
+ {
+ "epoch": 12.558951965065502,
+ "grad_norm": 0.013033892959356308,
+ "learning_rate": 0.0006,
+ "loss": 4.764178276062012,
+ "step": 904
+ },
+ {
+ "epoch": 12.57292576419214,
+ "grad_norm": 0.013996980153024197,
+ "learning_rate": 0.0006,
+ "loss": 4.767983436584473,
+ "step": 905
+ },
+ {
+ "epoch": 12.586899563318777,
+ "grad_norm": 0.015222184360027313,
+ "learning_rate": 0.0006,
+ "loss": 4.759721755981445,
+ "step": 906
+ },
+ {
+ "epoch": 12.600873362445414,
+ "grad_norm": 0.01623941957950592,
+ "learning_rate": 0.0006,
+ "loss": 4.849811553955078,
+ "step": 907
+ },
+ {
+ "epoch": 12.614847161572053,
+ "grad_norm": 0.015711182728409767,
+ "learning_rate": 0.0006,
+ "loss": 4.7829084396362305,
+ "step": 908
+ },
+ {
+ "epoch": 12.62882096069869,
+ "grad_norm": 0.014137467369437218,
+ "learning_rate": 0.0006,
+ "loss": 4.73321533203125,
+ "step": 909
+ },
+ {
+ "epoch": 12.642794759825328,
+ "grad_norm": 0.012738462537527084,
+ "learning_rate": 0.0006,
+ "loss": 4.870052337646484,
+ "step": 910
+ },
+ {
+ "epoch": 12.656768558951965,
+ "grad_norm": 0.013683130033314228,
+ "learning_rate": 0.0006,
+ "loss": 4.82658052444458,
+ "step": 911
+ },
+ {
+ "epoch": 12.670742358078602,
+ "grad_norm": 0.014409830793738365,
+ "learning_rate": 0.0006,
+ "loss": 4.707494735717773,
+ "step": 912
+ },
+ {
+ "epoch": 12.68471615720524,
+ "grad_norm": 0.013806809671223164,
+ "learning_rate": 0.0006,
+ "loss": 4.778928279876709,
+ "step": 913
+ },
+ {
+ "epoch": 12.698689956331878,
+ "grad_norm": 0.012857912108302116,
+ "learning_rate": 0.0006,
+ "loss": 4.890717506408691,
+ "step": 914
+ },
+ {
+ "epoch": 12.712663755458514,
+ "grad_norm": 0.013828758150339127,
+ "learning_rate": 0.0006,
+ "loss": 4.8357439041137695,
+ "step": 915
+ },
+ {
+ "epoch": 12.726637554585153,
+ "grad_norm": 0.01478351280093193,
+ "learning_rate": 0.0006,
+ "loss": 4.869986534118652,
+ "step": 916
+ },
+ {
+ "epoch": 12.74061135371179,
+ "grad_norm": 0.01326421182602644,
+ "learning_rate": 0.0006,
+ "loss": 4.830646514892578,
+ "step": 917
+ },
+ {
+ "epoch": 12.754585152838429,
+ "grad_norm": 0.013335433788597584,
+ "learning_rate": 0.0006,
+ "loss": 4.7786149978637695,
+ "step": 918
+ },
+ {
+ "epoch": 12.768558951965066,
+ "grad_norm": 0.014261147007346153,
+ "learning_rate": 0.0006,
+ "loss": 4.788693428039551,
+ "step": 919
+ },
+ {
+ "epoch": 12.782532751091702,
+ "grad_norm": 0.014274783432483673,
+ "learning_rate": 0.0006,
+ "loss": 4.77295446395874,
+ "step": 920
+ },
+ {
+ "epoch": 12.796506550218341,
+ "grad_norm": 0.015074674971401691,
+ "learning_rate": 0.0006,
+ "loss": 4.830754280090332,
+ "step": 921
+ },
+ {
+ "epoch": 12.810480349344978,
+ "grad_norm": 0.015599515289068222,
+ "learning_rate": 0.0006,
+ "loss": 4.837759017944336,
+ "step": 922
+ },
+ {
+ "epoch": 12.824454148471617,
+ "grad_norm": 0.014595242217183113,
+ "learning_rate": 0.0006,
+ "loss": 4.882965564727783,
+ "step": 923
+ },
+ {
+ "epoch": 12.838427947598253,
+ "grad_norm": 0.014951467514038086,
+ "learning_rate": 0.0006,
+ "loss": 4.709748268127441,
+ "step": 924
+ },
+ {
+ "epoch": 12.85240174672489,
+ "grad_norm": 0.01539295818656683,
+ "learning_rate": 0.0006,
+ "loss": 4.791160583496094,
+ "step": 925
+ },
+ {
+ "epoch": 12.866375545851529,
+ "grad_norm": 0.015987906605005264,
+ "learning_rate": 0.0006,
+ "loss": 4.661876678466797,
+ "step": 926
+ },
+ {
+ "epoch": 12.880349344978166,
+ "grad_norm": 0.016527341678738594,
+ "learning_rate": 0.0006,
+ "loss": 4.721952438354492,
+ "step": 927
+ },
+ {
+ "epoch": 12.894323144104803,
+ "grad_norm": 0.01637161523103714,
+ "learning_rate": 0.0006,
+ "loss": 4.704446792602539,
+ "step": 928
+ },
+ {
+ "epoch": 12.908296943231441,
+ "grad_norm": 0.015221050009131432,
+ "learning_rate": 0.0006,
+ "loss": 4.771907806396484,
+ "step": 929
+ },
+ {
+ "epoch": 12.922270742358078,
+ "grad_norm": 0.016242166981101036,
+ "learning_rate": 0.0006,
+ "loss": 4.806267738342285,
+ "step": 930
+ },
+ {
+ "epoch": 12.936244541484717,
+ "grad_norm": 0.01628357358276844,
+ "learning_rate": 0.0006,
+ "loss": 4.7252326011657715,
+ "step": 931
+ },
+ {
+ "epoch": 12.950218340611354,
+ "grad_norm": 0.017428133636713028,
+ "learning_rate": 0.0006,
+ "loss": 4.765689849853516,
+ "step": 932
+ },
+ {
+ "epoch": 12.96419213973799,
+ "grad_norm": 0.01870061084628105,
+ "learning_rate": 0.0006,
+ "loss": 4.814488887786865,
+ "step": 933
+ },
+ {
+ "epoch": 12.97816593886463,
+ "grad_norm": 0.019796060398221016,
+ "learning_rate": 0.0006,
+ "loss": 4.739185810089111,
+ "step": 934
+ },
+ {
+ "epoch": 12.992139737991266,
+ "grad_norm": 0.01985820196568966,
+ "learning_rate": 0.0006,
+ "loss": 4.810608863830566,
+ "step": 935
+ },
+ {
+ "epoch": 13.0,
+ "grad_norm": 0.020834477618336678,
+ "learning_rate": 0.0006,
+ "loss": 4.727858066558838,
+ "step": 936
+ },
+ {
+ "epoch": 13.0,
+ "eval_loss": 4.979116916656494,
+ "eval_runtime": 57.2877,
+ "eval_samples_per_second": 42.627,
+ "eval_steps_per_second": 1.344,
+ "step": 936
+ },
+ {
+ "epoch": 13.013973799126637,
+ "grad_norm": 0.02025519125163555,
+ "learning_rate": 0.0006,
+ "loss": 4.628875732421875,
+ "step": 937
+ },
+ {
+ "epoch": 13.027947598253276,
+ "grad_norm": 0.019596368074417114,
+ "learning_rate": 0.0006,
+ "loss": 4.694699287414551,
+ "step": 938
+ },
+ {
+ "epoch": 13.041921397379912,
+ "grad_norm": 0.020494934171438217,
+ "learning_rate": 0.0006,
+ "loss": 4.691821098327637,
+ "step": 939
+ },
+ {
+ "epoch": 13.055895196506551,
+ "grad_norm": 0.020121760666370392,
+ "learning_rate": 0.0006,
+ "loss": 4.71478271484375,
+ "step": 940
+ },
+ {
+ "epoch": 13.069868995633188,
+ "grad_norm": 0.021091405302286148,
+ "learning_rate": 0.0006,
+ "loss": 4.695915222167969,
+ "step": 941
+ },
+ {
+ "epoch": 13.083842794759825,
+ "grad_norm": 0.019787423312664032,
+ "learning_rate": 0.0006,
+ "loss": 4.8652448654174805,
+ "step": 942
+ },
+ {
+ "epoch": 13.097816593886463,
+ "grad_norm": 0.0185833927243948,
+ "learning_rate": 0.0006,
+ "loss": 4.7065887451171875,
+ "step": 943
+ },
+ {
+ "epoch": 13.1117903930131,
+ "grad_norm": 0.021220475435256958,
+ "learning_rate": 0.0006,
+ "loss": 4.741232395172119,
+ "step": 944
+ },
+ {
+ "epoch": 13.125764192139737,
+ "grad_norm": 0.02136622555553913,
+ "learning_rate": 0.0006,
+ "loss": 4.705967903137207,
+ "step": 945
+ },
+ {
+ "epoch": 13.139737991266376,
+ "grad_norm": 0.0191587433218956,
+ "learning_rate": 0.0006,
+ "loss": 4.758667945861816,
+ "step": 946
+ },
+ {
+ "epoch": 13.153711790393013,
+ "grad_norm": 0.018290022388100624,
+ "learning_rate": 0.0006,
+ "loss": 4.760666847229004,
+ "step": 947
+ },
+ {
+ "epoch": 13.167685589519651,
+ "grad_norm": 0.021704409271478653,
+ "learning_rate": 0.0006,
+ "loss": 4.750555992126465,
+ "step": 948
+ },
+ {
+ "epoch": 13.181659388646288,
+ "grad_norm": 0.021668918430805206,
+ "learning_rate": 0.0006,
+ "loss": 4.840903282165527,
+ "step": 949
+ },
+ {
+ "epoch": 13.195633187772925,
+ "grad_norm": 0.019497141242027283,
+ "learning_rate": 0.0006,
+ "loss": 4.738969326019287,
+ "step": 950
+ },
+ {
+ "epoch": 13.209606986899564,
+ "grad_norm": 0.018500596284866333,
+ "learning_rate": 0.0006,
+ "loss": 4.75154972076416,
+ "step": 951
+ },
+ {
+ "epoch": 13.2235807860262,
+ "grad_norm": 0.01872239261865616,
+ "learning_rate": 0.0006,
+ "loss": 4.811738014221191,
+ "step": 952
+ },
+ {
+ "epoch": 13.237554585152838,
+ "grad_norm": 0.020321547985076904,
+ "learning_rate": 0.0006,
+ "loss": 4.815384387969971,
+ "step": 953
+ },
+ {
+ "epoch": 13.251528384279476,
+ "grad_norm": 0.02283575013279915,
+ "learning_rate": 0.0006,
+ "loss": 4.74582052230835,
+ "step": 954
+ },
+ {
+ "epoch": 13.265502183406113,
+ "grad_norm": 0.02474398910999298,
+ "learning_rate": 0.0006,
+ "loss": 4.738375186920166,
+ "step": 955
+ },
+ {
+ "epoch": 13.279475982532752,
+ "grad_norm": 0.02065764181315899,
+ "learning_rate": 0.0006,
+ "loss": 4.746875286102295,
+ "step": 956
+ },
+ {
+ "epoch": 13.293449781659389,
+ "grad_norm": 0.020534520968794823,
+ "learning_rate": 0.0006,
+ "loss": 4.656914710998535,
+ "step": 957
+ },
+ {
+ "epoch": 13.307423580786025,
+ "grad_norm": 0.0206378772854805,
+ "learning_rate": 0.0006,
+ "loss": 4.842615604400635,
+ "step": 958
+ },
+ {
+ "epoch": 13.321397379912664,
+ "grad_norm": 0.017935654148459435,
+ "learning_rate": 0.0006,
+ "loss": 4.717827796936035,
+ "step": 959
+ },
+ {
+ "epoch": 13.335371179039301,
+ "grad_norm": 0.016958363354206085,
+ "learning_rate": 0.0006,
+ "loss": 4.681070327758789,
+ "step": 960
+ },
+ {
+ "epoch": 13.34934497816594,
+ "grad_norm": 0.015735412016510963,
+ "learning_rate": 0.0006,
+ "loss": 4.7957048416137695,
+ "step": 961
+ },
+ {
+ "epoch": 13.363318777292577,
+ "grad_norm": 0.01563699170947075,
+ "learning_rate": 0.0006,
+ "loss": 4.701192378997803,
+ "step": 962
+ },
+ {
+ "epoch": 13.377292576419213,
+ "grad_norm": 0.015207760035991669,
+ "learning_rate": 0.0006,
+ "loss": 4.73256254196167,
+ "step": 963
+ },
+ {
+ "epoch": 13.391266375545852,
+ "grad_norm": 0.015396883711218834,
+ "learning_rate": 0.0006,
+ "loss": 4.76412296295166,
+ "step": 964
+ },
+ {
+ "epoch": 13.405240174672489,
+ "grad_norm": 0.015690583735704422,
+ "learning_rate": 0.0006,
+ "loss": 4.791157245635986,
+ "step": 965
+ },
+ {
+ "epoch": 13.419213973799126,
+ "grad_norm": 0.01755724661052227,
+ "learning_rate": 0.0006,
+ "loss": 4.845666885375977,
+ "step": 966
+ },
+ {
+ "epoch": 13.433187772925764,
+ "grad_norm": 0.019443422555923462,
+ "learning_rate": 0.0006,
+ "loss": 4.5902299880981445,
+ "step": 967
+ },
+ {
+ "epoch": 13.447161572052401,
+ "grad_norm": 0.02118338644504547,
+ "learning_rate": 0.0006,
+ "loss": 4.740627288818359,
+ "step": 968
+ },
+ {
+ "epoch": 13.46113537117904,
+ "grad_norm": 0.02174839936196804,
+ "learning_rate": 0.0006,
+ "loss": 4.814560890197754,
+ "step": 969
+ },
+ {
+ "epoch": 13.475109170305677,
+ "grad_norm": 0.022860374301671982,
+ "learning_rate": 0.0006,
+ "loss": 4.66440486907959,
+ "step": 970
+ },
+ {
+ "epoch": 13.489082969432314,
+ "grad_norm": 0.024742677807807922,
+ "learning_rate": 0.0006,
+ "loss": 4.7451653480529785,
+ "step": 971
+ },
+ {
+ "epoch": 13.503056768558952,
+ "grad_norm": 0.027068907395005226,
+ "learning_rate": 0.0006,
+ "loss": 4.802584171295166,
+ "step": 972
+ },
+ {
+ "epoch": 13.51703056768559,
+ "grad_norm": 0.026316625997424126,
+ "learning_rate": 0.0006,
+ "loss": 4.842397689819336,
+ "step": 973
+ },
+ {
+ "epoch": 13.531004366812226,
+ "grad_norm": 0.024610962718725204,
+ "learning_rate": 0.0006,
+ "loss": 4.650721073150635,
+ "step": 974
+ },
+ {
+ "epoch": 13.544978165938865,
+ "grad_norm": 0.024122603237628937,
+ "learning_rate": 0.0006,
+ "loss": 4.6984357833862305,
+ "step": 975
+ },
+ {
+ "epoch": 13.558951965065502,
+ "grad_norm": 0.022699708119034767,
+ "learning_rate": 0.0006,
+ "loss": 4.793007850646973,
+ "step": 976
+ },
+ {
+ "epoch": 13.57292576419214,
+ "grad_norm": 0.021706534549593925,
+ "learning_rate": 0.0006,
+ "loss": 4.691362380981445,
+ "step": 977
+ },
+ {
+ "epoch": 13.586899563318777,
+ "grad_norm": 0.021363291889429092,
+ "learning_rate": 0.0006,
+ "loss": 4.76762580871582,
+ "step": 978
+ },
+ {
+ "epoch": 13.600873362445414,
+ "grad_norm": 0.023476677015423775,
+ "learning_rate": 0.0006,
+ "loss": 4.777202606201172,
+ "step": 979
+ },
+ {
+ "epoch": 13.614847161572053,
+ "grad_norm": 0.02457100711762905,
+ "learning_rate": 0.0006,
+ "loss": 4.590703010559082,
+ "step": 980
+ },
+ {
+ "epoch": 13.62882096069869,
+ "grad_norm": 0.02469099499285221,
+ "learning_rate": 0.0006,
+ "loss": 4.640185356140137,
+ "step": 981
+ },
+ {
+ "epoch": 13.642794759825328,
+ "grad_norm": 0.024914603680372238,
+ "learning_rate": 0.0006,
+ "loss": 4.772201061248779,
+ "step": 982
+ },
+ {
+ "epoch": 13.656768558951965,
+ "grad_norm": 0.025133173912763596,
+ "learning_rate": 0.0006,
+ "loss": 4.743337154388428,
+ "step": 983
+ },
+ {
+ "epoch": 13.670742358078602,
+ "grad_norm": 0.02034137211740017,
+ "learning_rate": 0.0006,
+ "loss": 4.752127647399902,
+ "step": 984
+ },
+ {
+ "epoch": 13.68471615720524,
+ "grad_norm": 0.019645661115646362,
+ "learning_rate": 0.0006,
+ "loss": 4.838453769683838,
+ "step": 985
+ },
+ {
+ "epoch": 13.698689956331878,
+ "grad_norm": 0.01900297962129116,
+ "learning_rate": 0.0006,
+ "loss": 4.804868221282959,
+ "step": 986
+ },
+ {
+ "epoch": 13.712663755458514,
+ "grad_norm": 0.020292341709136963,
+ "learning_rate": 0.0006,
+ "loss": 4.609743118286133,
+ "step": 987
+ },
+ {
+ "epoch": 13.726637554585153,
+ "grad_norm": 0.017014775425195694,
+ "learning_rate": 0.0006,
+ "loss": 4.740612983703613,
+ "step": 988
+ },
+ {
+ "epoch": 13.74061135371179,
+ "grad_norm": 0.015199844725430012,
+ "learning_rate": 0.0006,
+ "loss": 4.719701766967773,
+ "step": 989
+ },
+ {
+ "epoch": 13.754585152838429,
+ "grad_norm": 0.015125414356589317,
+ "learning_rate": 0.0006,
+ "loss": 4.686114311218262,
+ "step": 990
+ },
+ {
+ "epoch": 13.768558951965066,
+ "grad_norm": 0.014779540710151196,
+ "learning_rate": 0.0006,
+ "loss": 4.674347877502441,
+ "step": 991
+ },
+ {
+ "epoch": 13.782532751091702,
+ "grad_norm": 0.015255163423717022,
+ "learning_rate": 0.0006,
+ "loss": 4.6309309005737305,
+ "step": 992
+ },
+ {
+ "epoch": 13.796506550218341,
+ "grad_norm": 0.013692197389900684,
+ "learning_rate": 0.0006,
+ "loss": 4.698497295379639,
+ "step": 993
+ },
+ {
+ "epoch": 13.810480349344978,
+ "grad_norm": 0.014529623091220856,
+ "learning_rate": 0.0006,
+ "loss": 4.628443717956543,
+ "step": 994
+ },
+ {
+ "epoch": 13.824454148471617,
+ "grad_norm": 0.01576659083366394,
+ "learning_rate": 0.0006,
+ "loss": 4.788015842437744,
+ "step": 995
+ },
+ {
+ "epoch": 13.838427947598253,
+ "grad_norm": 0.015834277495741844,
+ "learning_rate": 0.0006,
+ "loss": 4.672703742980957,
+ "step": 996
+ },
+ {
+ "epoch": 13.85240174672489,
+ "grad_norm": 0.015539503656327724,
+ "learning_rate": 0.0006,
+ "loss": 4.731503486633301,
+ "step": 997
+ },
+ {
+ "epoch": 13.866375545851529,
+ "grad_norm": 0.013463746756315231,
+ "learning_rate": 0.0006,
+ "loss": 4.7006330490112305,
+ "step": 998
+ },
+ {
+ "epoch": 13.880349344978166,
+ "grad_norm": 0.016377726569771767,
+ "learning_rate": 0.0006,
+ "loss": 4.636074066162109,
+ "step": 999
+ },
+ {
+ "epoch": 13.894323144104803,
+ "grad_norm": 0.016660034656524658,
+ "learning_rate": 0.0006,
+ "loss": 4.742693901062012,
+ "step": 1000
+ },
+ {
+ "epoch": 13.908296943231441,
+ "grad_norm": 0.014902938157320023,
+ "learning_rate": 0.0006,
+ "loss": 4.769811630249023,
+ "step": 1001
+ },
+ {
+ "epoch": 13.922270742358078,
+ "grad_norm": 0.014883476309478283,
+ "learning_rate": 0.0006,
+ "loss": 4.731159210205078,
+ "step": 1002
+ },
+ {
+ "epoch": 13.936244541484717,
+ "grad_norm": 0.014494217932224274,
+ "learning_rate": 0.0006,
+ "loss": 4.659485816955566,
+ "step": 1003
+ },
+ {
+ "epoch": 13.950218340611354,
+ "grad_norm": 0.015426991507411003,
+ "learning_rate": 0.0006,
+ "loss": 4.755229949951172,
+ "step": 1004
+ },
+ {
+ "epoch": 13.96419213973799,
+ "grad_norm": 0.016648586839437485,
+ "learning_rate": 0.0006,
+ "loss": 4.74111270904541,
+ "step": 1005
+ },
+ {
+ "epoch": 13.97816593886463,
+ "grad_norm": 0.016980541869997978,
+ "learning_rate": 0.0006,
+ "loss": 4.74072265625,
+ "step": 1006
+ },
+ {
+ "epoch": 13.992139737991266,
+ "grad_norm": 0.018070152029395103,
+ "learning_rate": 0.0006,
+ "loss": 4.732975006103516,
+ "step": 1007
+ },
+ {
+ "epoch": 14.0,
+ "grad_norm": 0.019910665228962898,
+ "learning_rate": 0.0006,
+ "loss": 4.634697914123535,
+ "step": 1008
+ },
+ {
+ "epoch": 14.0,
+ "eval_loss": 4.850640296936035,
+ "eval_runtime": 57.3039,
+ "eval_samples_per_second": 42.615,
+ "eval_steps_per_second": 1.344,
+ "step": 1008
+ },
+ {
+ "epoch": 14.013973799126637,
+ "grad_norm": 0.01980053074657917,
+ "learning_rate": 0.0006,
+ "loss": 4.628744602203369,
+ "step": 1009
+ },
+ {
+ "epoch": 14.027947598253276,
+ "grad_norm": 0.02497623674571514,
+ "learning_rate": 0.0006,
+ "loss": 4.75930118560791,
+ "step": 1010
+ },
+ {
+ "epoch": 14.041921397379912,
+ "grad_norm": 0.02794097363948822,
+ "learning_rate": 0.0006,
+ "loss": 4.691340446472168,
+ "step": 1011
+ },
+ {
+ "epoch": 14.055895196506551,
+ "grad_norm": 0.023725535720586777,
+ "learning_rate": 0.0006,
+ "loss": 4.733664512634277,
+ "step": 1012
+ },
+ {
+ "epoch": 14.069868995633188,
+ "grad_norm": 0.022966474294662476,
+ "learning_rate": 0.0006,
+ "loss": 4.664808750152588,
+ "step": 1013
+ },
+ {
+ "epoch": 14.083842794759825,
+ "grad_norm": 0.02530563250184059,
+ "learning_rate": 0.0006,
+ "loss": 4.8047285079956055,
+ "step": 1014
+ },
+ {
+ "epoch": 14.097816593886463,
+ "grad_norm": 0.02491082064807415,
+ "learning_rate": 0.0006,
+ "loss": 4.667956352233887,
+ "step": 1015
+ },
+ {
+ "epoch": 14.1117903930131,
+ "grad_norm": 0.022110294550657272,
+ "learning_rate": 0.0006,
+ "loss": 4.666140556335449,
+ "step": 1016
+ },
+ {
+ "epoch": 14.125764192139737,
+ "grad_norm": 0.02065849117934704,
+ "learning_rate": 0.0006,
+ "loss": 4.581960678100586,
+ "step": 1017
+ },
+ {
+ "epoch": 14.139737991266376,
+ "grad_norm": 0.01941298507153988,
+ "learning_rate": 0.0006,
+ "loss": 4.614226341247559,
+ "step": 1018
+ },
+ {
+ "epoch": 14.153711790393013,
+ "grad_norm": 0.022441856563091278,
+ "learning_rate": 0.0006,
+ "loss": 4.638352394104004,
+ "step": 1019
+ },
+ {
+ "epoch": 14.167685589519651,
+ "grad_norm": 0.021519694477319717,
+ "learning_rate": 0.0006,
+ "loss": 4.700160980224609,
+ "step": 1020
+ },
+ {
+ "epoch": 14.181659388646288,
+ "grad_norm": 0.018956486135721207,
+ "learning_rate": 0.0006,
+ "loss": 4.649593353271484,
+ "step": 1021
+ },
+ {
+ "epoch": 14.195633187772925,
+ "grad_norm": 0.016681140288710594,
+ "learning_rate": 0.0006,
+ "loss": 4.603952407836914,
+ "step": 1022
+ },
+ {
+ "epoch": 14.209606986899564,
+ "grad_norm": 0.01788809522986412,
+ "learning_rate": 0.0006,
+ "loss": 4.669160842895508,
+ "step": 1023
+ },
+ {
+ "epoch": 14.2235807860262,
+ "grad_norm": 0.016047228127717972,
+ "learning_rate": 0.0006,
+ "loss": 4.6367950439453125,
+ "step": 1024
+ },
+ {
+ "epoch": 14.237554585152838,
+ "grad_norm": 0.01614239625632763,
+ "learning_rate": 0.0006,
+ "loss": 4.664529323577881,
+ "step": 1025
+ },
+ {
+ "epoch": 14.251528384279476,
+ "grad_norm": 0.016362549737095833,
+ "learning_rate": 0.0006,
+ "loss": 4.702373504638672,
+ "step": 1026
+ },
+ {
+ "epoch": 14.265502183406113,
+ "grad_norm": 0.015800409018993378,
+ "learning_rate": 0.0006,
+ "loss": 4.7594428062438965,
+ "step": 1027
+ },
+ {
+ "epoch": 14.279475982532752,
+ "grad_norm": 0.016730139032006264,
+ "learning_rate": 0.0006,
+ "loss": 4.677244186401367,
+ "step": 1028
+ },
+ {
+ "epoch": 14.293449781659389,
+ "grad_norm": 0.01661759801208973,
+ "learning_rate": 0.0006,
+ "loss": 4.633679389953613,
+ "step": 1029
+ },
+ {
+ "epoch": 14.307423580786025,
+ "grad_norm": 0.017673347145318985,
+ "learning_rate": 0.0006,
+ "loss": 4.649107456207275,
+ "step": 1030
+ },
+ {
+ "epoch": 14.321397379912664,
+ "grad_norm": 0.019243352115154266,
+ "learning_rate": 0.0006,
+ "loss": 4.611257553100586,
+ "step": 1031
+ },
+ {
+ "epoch": 14.335371179039301,
+ "grad_norm": 0.019817164167761803,
+ "learning_rate": 0.0006,
+ "loss": 4.588086128234863,
+ "step": 1032
+ },
+ {
+ "epoch": 14.34934497816594,
+ "grad_norm": 0.022292211651802063,
+ "learning_rate": 0.0006,
+ "loss": 4.783066749572754,
+ "step": 1033
+ },
+ {
+ "epoch": 14.363318777292577,
+ "grad_norm": 0.02140096202492714,
+ "learning_rate": 0.0006,
+ "loss": 4.604718208312988,
+ "step": 1034
+ },
+ {
+ "epoch": 14.377292576419213,
+ "grad_norm": 0.0187817495316267,
+ "learning_rate": 0.0006,
+ "loss": 4.721773147583008,
+ "step": 1035
+ },
+ {
+ "epoch": 14.391266375545852,
+ "grad_norm": 0.019879210740327835,
+ "learning_rate": 0.0006,
+ "loss": 4.789286136627197,
+ "step": 1036
+ },
+ {
+ "epoch": 14.405240174672489,
+ "grad_norm": 0.01885785162448883,
+ "learning_rate": 0.0006,
+ "loss": 4.669743537902832,
+ "step": 1037
+ },
+ {
+ "epoch": 14.419213973799126,
+ "grad_norm": 0.018324270844459534,
+ "learning_rate": 0.0006,
+ "loss": 4.65901517868042,
+ "step": 1038
+ },
+ {
+ "epoch": 14.433187772925764,
+ "grad_norm": 0.017664821818470955,
+ "learning_rate": 0.0006,
+ "loss": 4.667839527130127,
+ "step": 1039
+ },
+ {
+ "epoch": 14.447161572052401,
+ "grad_norm": 0.015702905133366585,
+ "learning_rate": 0.0006,
+ "loss": 4.64700984954834,
+ "step": 1040
+ },
+ {
+ "epoch": 14.46113537117904,
+ "grad_norm": 0.016674194484949112,
+ "learning_rate": 0.0006,
+ "loss": 4.694093227386475,
+ "step": 1041
+ },
+ {
+ "epoch": 14.475109170305677,
+ "grad_norm": 0.016869477927684784,
+ "learning_rate": 0.0006,
+ "loss": 4.619544982910156,
+ "step": 1042
+ },
+ {
+ "epoch": 14.489082969432314,
+ "grad_norm": 0.016557540744543076,
+ "learning_rate": 0.0006,
+ "loss": 4.568667411804199,
+ "step": 1043
+ },
+ {
+ "epoch": 14.503056768558952,
+ "grad_norm": 0.014500945806503296,
+ "learning_rate": 0.0006,
+ "loss": 4.561741828918457,
+ "step": 1044
+ },
+ {
+ "epoch": 14.51703056768559,
+ "grad_norm": 0.014228739775717258,
+ "learning_rate": 0.0006,
+ "loss": 4.654277324676514,
+ "step": 1045
+ },
+ {
+ "epoch": 14.531004366812226,
+ "grad_norm": 0.013894238509237766,
+ "learning_rate": 0.0006,
+ "loss": 4.5867719650268555,
+ "step": 1046
+ },
+ {
+ "epoch": 14.544978165938865,
+ "grad_norm": 0.01550068985670805,
+ "learning_rate": 0.0006,
+ "loss": 4.61168098449707,
+ "step": 1047
+ },
+ {
+ "epoch": 14.558951965065502,
+ "grad_norm": 0.015940772369503975,
+ "learning_rate": 0.0006,
+ "loss": 4.691247940063477,
+ "step": 1048
+ },
+ {
+ "epoch": 14.57292576419214,
+ "grad_norm": 0.01714535616338253,
+ "learning_rate": 0.0006,
+ "loss": 4.61183500289917,
+ "step": 1049
+ },
+ {
+ "epoch": 14.586899563318777,
+ "grad_norm": 0.017925038933753967,
+ "learning_rate": 0.0006,
+ "loss": 4.719022750854492,
+ "step": 1050
+ },
+ {
+ "epoch": 14.600873362445414,
+ "grad_norm": 0.018141552805900574,
+ "learning_rate": 0.0006,
+ "loss": 4.760281562805176,
+ "step": 1051
+ },
+ {
+ "epoch": 14.614847161572053,
+ "grad_norm": 0.017131328582763672,
+ "learning_rate": 0.0006,
+ "loss": 4.709680080413818,
+ "step": 1052
+ },
+ {
+ "epoch": 14.62882096069869,
+ "grad_norm": 0.018730048090219498,
+ "learning_rate": 0.0006,
+ "loss": 4.5994720458984375,
+ "step": 1053
+ },
+ {
+ "epoch": 14.642794759825328,
+ "grad_norm": 0.018894298002123833,
+ "learning_rate": 0.0006,
+ "loss": 4.669070720672607,
+ "step": 1054
+ },
+ {
+ "epoch": 14.656768558951965,
+ "grad_norm": 0.017643198370933533,
+ "learning_rate": 0.0006,
+ "loss": 4.641790390014648,
+ "step": 1055
+ },
+ {
+ "epoch": 14.670742358078602,
+ "grad_norm": 0.017647244036197662,
+ "learning_rate": 0.0006,
+ "loss": 4.653899192810059,
+ "step": 1056
+ },
+ {
+ "epoch": 14.68471615720524,
+ "grad_norm": 0.017304198816418648,
+ "learning_rate": 0.0006,
+ "loss": 4.7487473487854,
+ "step": 1057
+ },
+ {
+ "epoch": 14.698689956331878,
+ "grad_norm": 0.01638447679579258,
+ "learning_rate": 0.0006,
+ "loss": 4.71945858001709,
+ "step": 1058
+ },
+ {
+ "epoch": 14.712663755458514,
+ "grad_norm": 0.015892071649432182,
+ "learning_rate": 0.0006,
+ "loss": 4.744903564453125,
+ "step": 1059
+ },
+ {
+ "epoch": 14.726637554585153,
+ "grad_norm": 0.017420614138245583,
+ "learning_rate": 0.0006,
+ "loss": 4.781563758850098,
+ "step": 1060
+ },
+ {
+ "epoch": 14.74061135371179,
+ "grad_norm": 0.016498520970344543,
+ "learning_rate": 0.0006,
+ "loss": 4.622779846191406,
+ "step": 1061
+ },
+ {
+ "epoch": 14.754585152838429,
+ "grad_norm": 0.017482127994298935,
+ "learning_rate": 0.0006,
+ "loss": 4.62825870513916,
+ "step": 1062
+ },
+ {
+ "epoch": 14.768558951965066,
+ "grad_norm": 0.019018379971385002,
+ "learning_rate": 0.0006,
+ "loss": 4.693194389343262,
+ "step": 1063
+ },
+ {
+ "epoch": 14.782532751091702,
+ "grad_norm": 0.01993914321064949,
+ "learning_rate": 0.0006,
+ "loss": 4.742903709411621,
+ "step": 1064
+ },
+ {
+ "epoch": 14.796506550218341,
+ "grad_norm": 0.02301335707306862,
+ "learning_rate": 0.0006,
+ "loss": 4.613251686096191,
+ "step": 1065
+ },
+ {
+ "epoch": 14.810480349344978,
+ "grad_norm": 0.024088222533464432,
+ "learning_rate": 0.0006,
+ "loss": 4.594305992126465,
+ "step": 1066
+ },
+ {
+ "epoch": 14.824454148471617,
+ "grad_norm": 0.02493320405483246,
+ "learning_rate": 0.0006,
+ "loss": 4.700588226318359,
+ "step": 1067
+ },
+ {
+ "epoch": 14.838427947598253,
+ "grad_norm": 0.02501937560737133,
+ "learning_rate": 0.0006,
+ "loss": 4.601742744445801,
+ "step": 1068
+ },
+ {
+ "epoch": 14.85240174672489,
+ "grad_norm": 0.023490751162171364,
+ "learning_rate": 0.0006,
+ "loss": 4.589211463928223,
+ "step": 1069
+ },
+ {
+ "epoch": 14.866375545851529,
+ "grad_norm": 0.019695665687322617,
+ "learning_rate": 0.0006,
+ "loss": 4.635081768035889,
+ "step": 1070
+ },
+ {
+ "epoch": 14.880349344978166,
+ "grad_norm": 0.019276980310678482,
+ "learning_rate": 0.0006,
+ "loss": 4.737002372741699,
+ "step": 1071
+ },
+ {
+ "epoch": 14.894323144104803,
+ "grad_norm": 0.022430511191487312,
+ "learning_rate": 0.0006,
+ "loss": 4.808295726776123,
+ "step": 1072
+ },
+ {
+ "epoch": 14.908296943231441,
+ "grad_norm": 0.021543916314840317,
+ "learning_rate": 0.0006,
+ "loss": 4.559475421905518,
+ "step": 1073
+ },
+ {
+ "epoch": 14.922270742358078,
+ "grad_norm": 0.019885241985321045,
+ "learning_rate": 0.0006,
+ "loss": 4.647948741912842,
+ "step": 1074
+ },
+ {
+ "epoch": 14.936244541484717,
+ "grad_norm": 0.02115175500512123,
+ "learning_rate": 0.0006,
+ "loss": 4.680505752563477,
+ "step": 1075
+ },
+ {
+ "epoch": 14.950218340611354,
+ "grad_norm": 0.02156151458621025,
+ "learning_rate": 0.0006,
+ "loss": 4.673477649688721,
+ "step": 1076
+ },
+ {
+ "epoch": 14.96419213973799,
+ "grad_norm": 0.02184601128101349,
+ "learning_rate": 0.0006,
+ "loss": 4.644402027130127,
+ "step": 1077
+ },
+ {
+ "epoch": 14.97816593886463,
+ "grad_norm": 0.0211983360350132,
+ "learning_rate": 0.0006,
+ "loss": 4.821558475494385,
+ "step": 1078
+ },
+ {
+ "epoch": 14.992139737991266,
+ "grad_norm": 0.019636567682027817,
+ "learning_rate": 0.0006,
+ "loss": 4.578868865966797,
+ "step": 1079
+ },
+ {
+ "epoch": 15.0,
+ "grad_norm": 0.019997350871562958,
+ "learning_rate": 0.0006,
+ "loss": 4.691307544708252,
+ "step": 1080
+ },
+ {
+ "epoch": 15.0,
+ "eval_loss": 4.872478008270264,
+ "eval_runtime": 56.8513,
+ "eval_samples_per_second": 42.954,
+ "eval_steps_per_second": 1.354,
+ "step": 1080
+ },
+ {
+ "epoch": 15.013973799126637,
+ "grad_norm": 0.018830662593245506,
+ "learning_rate": 0.0006,
+ "loss": 4.761779308319092,
+ "step": 1081
+ },
+ {
+ "epoch": 15.027947598253276,
+ "grad_norm": 0.016876481473445892,
+ "learning_rate": 0.0006,
+ "loss": 4.61148738861084,
+ "step": 1082
+ },
+ {
+ "epoch": 15.041921397379912,
+ "grad_norm": 0.015550863929092884,
+ "learning_rate": 0.0006,
+ "loss": 4.676124095916748,
+ "step": 1083
+ },
+ {
+ "epoch": 15.055895196506551,
+ "grad_norm": 0.017178896814584732,
+ "learning_rate": 0.0006,
+ "loss": 4.588935852050781,
+ "step": 1084
+ },
+ {
+ "epoch": 15.069868995633188,
+ "grad_norm": 0.015465447679162025,
+ "learning_rate": 0.0006,
+ "loss": 4.605499267578125,
+ "step": 1085
+ },
+ {
+ "epoch": 15.083842794759825,
+ "grad_norm": 0.01606195978820324,
+ "learning_rate": 0.0006,
+ "loss": 4.723373889923096,
+ "step": 1086
+ },
+ {
+ "epoch": 15.097816593886463,
+ "grad_norm": 0.016521936282515526,
+ "learning_rate": 0.0006,
+ "loss": 4.646608352661133,
+ "step": 1087
+ },
+ {
+ "epoch": 15.1117903930131,
+ "grad_norm": 0.01589205674827099,
+ "learning_rate": 0.0006,
+ "loss": 4.498138427734375,
+ "step": 1088
+ },
+ {
+ "epoch": 15.125764192139737,
+ "grad_norm": 0.01524856686592102,
+ "learning_rate": 0.0006,
+ "loss": 4.594562530517578,
+ "step": 1089
+ },
+ {
+ "epoch": 15.139737991266376,
+ "grad_norm": 0.014714641496539116,
+ "learning_rate": 0.0006,
+ "loss": 4.630744934082031,
+ "step": 1090
+ },
+ {
+ "epoch": 15.153711790393013,
+ "grad_norm": 0.015333331190049648,
+ "learning_rate": 0.0006,
+ "loss": 4.682388782501221,
+ "step": 1091
+ },
+ {
+ "epoch": 15.167685589519651,
+ "grad_norm": 0.01550073828548193,
+ "learning_rate": 0.0006,
+ "loss": 4.692652702331543,
+ "step": 1092
+ },
+ {
+ "epoch": 15.181659388646288,
+ "grad_norm": 0.014783318154513836,
+ "learning_rate": 0.0006,
+ "loss": 4.6223835945129395,
+ "step": 1093
+ },
+ {
+ "epoch": 15.195633187772925,
+ "grad_norm": 0.014372751116752625,
+ "learning_rate": 0.0006,
+ "loss": 4.52687406539917,
+ "step": 1094
+ },
+ {
+ "epoch": 15.209606986899564,
+ "grad_norm": 0.014779130928218365,
+ "learning_rate": 0.0006,
+ "loss": 4.637009620666504,
+ "step": 1095
+ },
+ {
+ "epoch": 15.2235807860262,
+ "grad_norm": 0.015697073191404343,
+ "learning_rate": 0.0006,
+ "loss": 4.682981014251709,
+ "step": 1096
+ },
+ {
+ "epoch": 15.237554585152838,
+ "grad_norm": 0.016523664817214012,
+ "learning_rate": 0.0006,
+ "loss": 4.705722808837891,
+ "step": 1097
+ },
+ {
+ "epoch": 15.251528384279476,
+ "grad_norm": 0.017633996903896332,
+ "learning_rate": 0.0006,
+ "loss": 4.574686050415039,
+ "step": 1098
+ },
+ {
+ "epoch": 15.265502183406113,
+ "grad_norm": 0.019955076277256012,
+ "learning_rate": 0.0006,
+ "loss": 4.634450912475586,
+ "step": 1099
+ },
+ {
+ "epoch": 15.279475982532752,
+ "grad_norm": 0.02132618986070156,
+ "learning_rate": 0.0006,
+ "loss": 4.56110143661499,
+ "step": 1100
+ },
+ {
+ "epoch": 15.293449781659389,
+ "grad_norm": 0.0207593385130167,
+ "learning_rate": 0.0006,
+ "loss": 4.629396438598633,
+ "step": 1101
+ },
+ {
+ "epoch": 15.307423580786025,
+ "grad_norm": 0.017882540822029114,
+ "learning_rate": 0.0006,
+ "loss": 4.730963230133057,
+ "step": 1102
+ },
+ {
+ "epoch": 15.321397379912664,
+ "grad_norm": 0.01917876861989498,
+ "learning_rate": 0.0006,
+ "loss": 4.64585542678833,
+ "step": 1103
+ },
+ {
+ "epoch": 15.335371179039301,
+ "grad_norm": 0.01943974196910858,
+ "learning_rate": 0.0006,
+ "loss": 4.656181335449219,
+ "step": 1104
+ },
+ {
+ "epoch": 15.34934497816594,
+ "grad_norm": 0.019904715940356255,
+ "learning_rate": 0.0006,
+ "loss": 4.619329452514648,
+ "step": 1105
+ },
+ {
+ "epoch": 15.363318777292577,
+ "grad_norm": 0.019174346700310707,
+ "learning_rate": 0.0006,
+ "loss": 4.647892475128174,
+ "step": 1106
+ },
+ {
+ "epoch": 15.377292576419213,
+ "grad_norm": 0.018244758248329163,
+ "learning_rate": 0.0006,
+ "loss": 4.56166934967041,
+ "step": 1107
+ },
+ {
+ "epoch": 15.391266375545852,
+ "grad_norm": 0.018914205953478813,
+ "learning_rate": 0.0006,
+ "loss": 4.648063659667969,
+ "step": 1108
+ },
+ {
+ "epoch": 15.405240174672489,
+ "grad_norm": 0.01961071416735649,
+ "learning_rate": 0.0006,
+ "loss": 4.665525436401367,
+ "step": 1109
+ },
+ {
+ "epoch": 15.419213973799126,
+ "grad_norm": 0.01841311901807785,
+ "learning_rate": 0.0006,
+ "loss": 4.611342430114746,
+ "step": 1110
+ },
+ {
+ "epoch": 15.433187772925764,
+ "grad_norm": 0.01863427273929119,
+ "learning_rate": 0.0006,
+ "loss": 4.664256572723389,
+ "step": 1111
+ },
+ {
+ "epoch": 15.447161572052401,
+ "grad_norm": 0.020506877452135086,
+ "learning_rate": 0.0006,
+ "loss": 4.631237506866455,
+ "step": 1112
+ },
+ {
+ "epoch": 15.46113537117904,
+ "grad_norm": 0.02223833277821541,
+ "learning_rate": 0.0006,
+ "loss": 4.668654441833496,
+ "step": 1113
+ },
+ {
+ "epoch": 15.475109170305677,
+ "grad_norm": 0.023336702957749367,
+ "learning_rate": 0.0006,
+ "loss": 4.562833309173584,
+ "step": 1114
+ },
+ {
+ "epoch": 15.489082969432314,
+ "grad_norm": 0.023052413016557693,
+ "learning_rate": 0.0006,
+ "loss": 4.621532917022705,
+ "step": 1115
+ },
+ {
+ "epoch": 15.503056768558952,
+ "grad_norm": 0.023420924320816994,
+ "learning_rate": 0.0006,
+ "loss": 4.665465354919434,
+ "step": 1116
+ },
+ {
+ "epoch": 15.51703056768559,
+ "grad_norm": 0.024071281775832176,
+ "learning_rate": 0.0006,
+ "loss": 4.595980644226074,
+ "step": 1117
+ },
+ {
+ "epoch": 15.531004366812226,
+ "grad_norm": 0.023406682536005974,
+ "learning_rate": 0.0006,
+ "loss": 4.642635345458984,
+ "step": 1118
+ },
+ {
+ "epoch": 15.544978165938865,
+ "grad_norm": 0.02138647809624672,
+ "learning_rate": 0.0006,
+ "loss": 4.704513072967529,
+ "step": 1119
+ },
+ {
+ "epoch": 15.558951965065502,
+ "grad_norm": 0.0221384447067976,
+ "learning_rate": 0.0006,
+ "loss": 4.567503452301025,
+ "step": 1120
+ },
+ {
+ "epoch": 15.57292576419214,
+ "grad_norm": 0.020879851654171944,
+ "learning_rate": 0.0006,
+ "loss": 4.664057731628418,
+ "step": 1121
+ },
+ {
+ "epoch": 15.586899563318777,
+ "grad_norm": 0.021478967741131783,
+ "learning_rate": 0.0006,
+ "loss": 4.539627552032471,
+ "step": 1122
+ },
+ {
+ "epoch": 15.600873362445414,
+ "grad_norm": 0.022748656570911407,
+ "learning_rate": 0.0006,
+ "loss": 4.5654497146606445,
+ "step": 1123
+ },
+ {
+ "epoch": 15.614847161572053,
+ "grad_norm": 0.021870248019695282,
+ "learning_rate": 0.0006,
+ "loss": 4.56899356842041,
+ "step": 1124
+ },
+ {
+ "epoch": 15.62882096069869,
+ "grad_norm": 0.021711355075240135,
+ "learning_rate": 0.0006,
+ "loss": 4.657808303833008,
+ "step": 1125
+ },
+ {
+ "epoch": 15.642794759825328,
+ "grad_norm": 0.022810909897089005,
+ "learning_rate": 0.0006,
+ "loss": 4.6526689529418945,
+ "step": 1126
+ },
+ {
+ "epoch": 15.656768558951965,
+ "grad_norm": 0.022561343386769295,
+ "learning_rate": 0.0006,
+ "loss": 4.534887790679932,
+ "step": 1127
+ },
+ {
+ "epoch": 15.670742358078602,
+ "grad_norm": 0.019909802824258804,
+ "learning_rate": 0.0006,
+ "loss": 4.600790023803711,
+ "step": 1128
+ },
+ {
+ "epoch": 15.68471615720524,
+ "grad_norm": 0.020092325285077095,
+ "learning_rate": 0.0006,
+ "loss": 4.773830413818359,
+ "step": 1129
+ },
+ {
+ "epoch": 15.698689956331878,
+ "grad_norm": 0.018834445625543594,
+ "learning_rate": 0.0006,
+ "loss": 4.628868103027344,
+ "step": 1130
+ },
+ {
+ "epoch": 15.712663755458514,
+ "grad_norm": 0.017552798613905907,
+ "learning_rate": 0.0006,
+ "loss": 4.647500038146973,
+ "step": 1131
+ },
+ {
+ "epoch": 15.726637554585153,
+ "grad_norm": 0.01751251146197319,
+ "learning_rate": 0.0006,
+ "loss": 4.590545654296875,
+ "step": 1132
+ },
+ {
+ "epoch": 15.74061135371179,
+ "grad_norm": 0.01699099875986576,
+ "learning_rate": 0.0006,
+ "loss": 4.509699821472168,
+ "step": 1133
+ },
+ {
+ "epoch": 15.754585152838429,
+ "grad_norm": 0.015808649361133575,
+ "learning_rate": 0.0006,
+ "loss": 4.606992244720459,
+ "step": 1134
+ },
+ {
+ "epoch": 15.768558951965066,
+ "grad_norm": 0.01655443385243416,
+ "learning_rate": 0.0006,
+ "loss": 4.585219860076904,
+ "step": 1135
+ },
+ {
+ "epoch": 15.782532751091702,
+ "grad_norm": 0.01708163693547249,
+ "learning_rate": 0.0006,
+ "loss": 4.668147087097168,
+ "step": 1136
+ },
+ {
+ "epoch": 15.796506550218341,
+ "grad_norm": 0.017161913216114044,
+ "learning_rate": 0.0006,
+ "loss": 4.7170634269714355,
+ "step": 1137
+ },
+ {
+ "epoch": 15.810480349344978,
+ "grad_norm": 0.017262674868106842,
+ "learning_rate": 0.0006,
+ "loss": 4.572132110595703,
+ "step": 1138
+ },
+ {
+ "epoch": 15.824454148471617,
+ "grad_norm": 0.017713576555252075,
+ "learning_rate": 0.0006,
+ "loss": 4.702538967132568,
+ "step": 1139
+ },
+ {
+ "epoch": 15.838427947598253,
+ "grad_norm": 0.018730733543634415,
+ "learning_rate": 0.0006,
+ "loss": 4.629721641540527,
+ "step": 1140
+ },
+ {
+ "epoch": 15.85240174672489,
+ "grad_norm": 0.017960187047719955,
+ "learning_rate": 0.0006,
+ "loss": 4.661039352416992,
+ "step": 1141
+ },
+ {
+ "epoch": 15.866375545851529,
+ "grad_norm": 0.01750079356133938,
+ "learning_rate": 0.0006,
+ "loss": 4.503710746765137,
+ "step": 1142
+ },
+ {
+ "epoch": 15.880349344978166,
+ "grad_norm": 0.016630422323942184,
+ "learning_rate": 0.0006,
+ "loss": 4.6582136154174805,
+ "step": 1143
+ },
+ {
+ "epoch": 15.894323144104803,
+ "grad_norm": 0.015896832570433617,
+ "learning_rate": 0.0006,
+ "loss": 4.596693992614746,
+ "step": 1144
+ },
+ {
+ "epoch": 15.908296943231441,
+ "grad_norm": 0.01566407084465027,
+ "learning_rate": 0.0006,
+ "loss": 4.6337199211120605,
+ "step": 1145
+ },
+ {
+ "epoch": 15.922270742358078,
+ "grad_norm": 0.016823330894112587,
+ "learning_rate": 0.0006,
+ "loss": 4.686994552612305,
+ "step": 1146
+ },
+ {
+ "epoch": 15.936244541484717,
+ "grad_norm": 0.016208255663514137,
+ "learning_rate": 0.0006,
+ "loss": 4.6205549240112305,
+ "step": 1147
+ },
+ {
+ "epoch": 15.950218340611354,
+ "grad_norm": 0.0171707421541214,
+ "learning_rate": 0.0006,
+ "loss": 4.582968711853027,
+ "step": 1148
+ },
+ {
+ "epoch": 15.96419213973799,
+ "grad_norm": 0.017009446397423744,
+ "learning_rate": 0.0006,
+ "loss": 4.510274887084961,
+ "step": 1149
+ },
+ {
+ "epoch": 15.97816593886463,
+ "grad_norm": 0.01565811224281788,
+ "learning_rate": 0.0006,
+ "loss": 4.55834436416626,
+ "step": 1150
+ },
+ {
+ "epoch": 15.992139737991266,
+ "grad_norm": 0.015276037156581879,
+ "learning_rate": 0.0006,
+ "loss": 4.537870407104492,
+ "step": 1151
+ },
+ {
+ "epoch": 16.0,
+ "grad_norm": 0.017305882647633553,
+ "learning_rate": 0.0006,
+ "loss": 4.611305236816406,
+ "step": 1152
+ },
+ {
+ "epoch": 16.0,
+ "eval_loss": 4.775513172149658,
+ "eval_runtime": 56.7341,
+ "eval_samples_per_second": 43.043,
+ "eval_steps_per_second": 1.357,
+ "step": 1152
+ },
+ {
+ "epoch": 16.01397379912664,
+ "grad_norm": 0.016752436757087708,
+ "learning_rate": 0.0006,
+ "loss": 4.541312217712402,
+ "step": 1153
+ },
+ {
+ "epoch": 16.027947598253274,
+ "grad_norm": 0.01643134094774723,
+ "learning_rate": 0.0006,
+ "loss": 4.585402965545654,
+ "step": 1154
+ },
+ {
+ "epoch": 16.041921397379912,
+ "grad_norm": 0.01888720504939556,
+ "learning_rate": 0.0006,
+ "loss": 4.597825050354004,
+ "step": 1155
+ },
+ {
+ "epoch": 16.05589519650655,
+ "grad_norm": 0.022044632583856583,
+ "learning_rate": 0.0006,
+ "loss": 4.532772064208984,
+ "step": 1156
+ },
+ {
+ "epoch": 16.069868995633186,
+ "grad_norm": 0.022986987605690956,
+ "learning_rate": 0.0006,
+ "loss": 4.551351547241211,
+ "step": 1157
+ },
+ {
+ "epoch": 16.083842794759825,
+ "grad_norm": 0.020116791129112244,
+ "learning_rate": 0.0006,
+ "loss": 4.484149932861328,
+ "step": 1158
+ },
+ {
+ "epoch": 16.097816593886463,
+ "grad_norm": 0.021442247554659843,
+ "learning_rate": 0.0006,
+ "loss": 4.515316486358643,
+ "step": 1159
+ },
+ {
+ "epoch": 16.111790393013102,
+ "grad_norm": 0.020974772050976753,
+ "learning_rate": 0.0006,
+ "loss": 4.51606559753418,
+ "step": 1160
+ },
+ {
+ "epoch": 16.125764192139737,
+ "grad_norm": 0.021274864673614502,
+ "learning_rate": 0.0006,
+ "loss": 4.587329864501953,
+ "step": 1161
+ },
+ {
+ "epoch": 16.139737991266376,
+ "grad_norm": 0.022799784317612648,
+ "learning_rate": 0.0006,
+ "loss": 4.568149089813232,
+ "step": 1162
+ },
+ {
+ "epoch": 16.153711790393015,
+ "grad_norm": 0.021784933283925056,
+ "learning_rate": 0.0006,
+ "loss": 4.515336990356445,
+ "step": 1163
+ },
+ {
+ "epoch": 16.16768558951965,
+ "grad_norm": 0.024261271581053734,
+ "learning_rate": 0.0006,
+ "loss": 4.5127081871032715,
+ "step": 1164
+ },
+ {
+ "epoch": 16.18165938864629,
+ "grad_norm": 0.02248375490307808,
+ "learning_rate": 0.0006,
+ "loss": 4.754251003265381,
+ "step": 1165
+ },
+ {
+ "epoch": 16.195633187772927,
+ "grad_norm": 0.02136492356657982,
+ "learning_rate": 0.0006,
+ "loss": 4.58193826675415,
+ "step": 1166
+ },
+ {
+ "epoch": 16.209606986899562,
+ "grad_norm": 0.020727066323161125,
+ "learning_rate": 0.0006,
+ "loss": 4.564209938049316,
+ "step": 1167
+ },
+ {
+ "epoch": 16.2235807860262,
+ "grad_norm": 0.017789119854569435,
+ "learning_rate": 0.0006,
+ "loss": 4.698680877685547,
+ "step": 1168
+ },
+ {
+ "epoch": 16.23755458515284,
+ "grad_norm": 0.017603034153580666,
+ "learning_rate": 0.0006,
+ "loss": 4.501171588897705,
+ "step": 1169
+ },
+ {
+ "epoch": 16.251528384279474,
+ "grad_norm": 0.017345862463116646,
+ "learning_rate": 0.0006,
+ "loss": 4.593590259552002,
+ "step": 1170
+ },
+ {
+ "epoch": 16.265502183406113,
+ "grad_norm": 0.018047425895929337,
+ "learning_rate": 0.0006,
+ "loss": 4.597467422485352,
+ "step": 1171
+ },
+ {
+ "epoch": 16.27947598253275,
+ "grad_norm": 0.01947222836315632,
+ "learning_rate": 0.0006,
+ "loss": 4.5072221755981445,
+ "step": 1172
+ },
+ {
+ "epoch": 16.29344978165939,
+ "grad_norm": 0.019598500803112984,
+ "learning_rate": 0.0006,
+ "loss": 4.601351737976074,
+ "step": 1173
+ },
+ {
+ "epoch": 16.307423580786025,
+ "grad_norm": 0.020002514123916626,
+ "learning_rate": 0.0006,
+ "loss": 4.580431938171387,
+ "step": 1174
+ },
+ {
+ "epoch": 16.321397379912664,
+ "grad_norm": 0.0216530729085207,
+ "learning_rate": 0.0006,
+ "loss": 4.424681663513184,
+ "step": 1175
+ },
+ {
+ "epoch": 16.335371179039303,
+ "grad_norm": 0.01949235238134861,
+ "learning_rate": 0.0006,
+ "loss": 4.751883029937744,
+ "step": 1176
+ },
+ {
+ "epoch": 16.349344978165938,
+ "grad_norm": 0.017779016867280006,
+ "learning_rate": 0.0006,
+ "loss": 4.541490077972412,
+ "step": 1177
+ },
+ {
+ "epoch": 16.363318777292577,
+ "grad_norm": 0.01909121684730053,
+ "learning_rate": 0.0006,
+ "loss": 4.594141006469727,
+ "step": 1178
+ },
+ {
+ "epoch": 16.377292576419215,
+ "grad_norm": 0.019894849509000778,
+ "learning_rate": 0.0006,
+ "loss": 4.555070400238037,
+ "step": 1179
+ },
+ {
+ "epoch": 16.39126637554585,
+ "grad_norm": 0.01993032731115818,
+ "learning_rate": 0.0006,
+ "loss": 4.521401405334473,
+ "step": 1180
+ },
+ {
+ "epoch": 16.40524017467249,
+ "grad_norm": 0.0210788045078516,
+ "learning_rate": 0.0006,
+ "loss": 4.672842979431152,
+ "step": 1181
+ },
+ {
+ "epoch": 16.419213973799128,
+ "grad_norm": 0.02274598926305771,
+ "learning_rate": 0.0006,
+ "loss": 4.629504203796387,
+ "step": 1182
+ },
+ {
+ "epoch": 16.433187772925763,
+ "grad_norm": 0.020448120310902596,
+ "learning_rate": 0.0006,
+ "loss": 4.548209190368652,
+ "step": 1183
+ },
+ {
+ "epoch": 16.4471615720524,
+ "grad_norm": 0.017565643414855003,
+ "learning_rate": 0.0006,
+ "loss": 4.529054641723633,
+ "step": 1184
+ },
+ {
+ "epoch": 16.46113537117904,
+ "grad_norm": 0.019504237920045853,
+ "learning_rate": 0.0006,
+ "loss": 4.48182487487793,
+ "step": 1185
+ },
+ {
+ "epoch": 16.475109170305675,
+ "grad_norm": 0.016823742538690567,
+ "learning_rate": 0.0006,
+ "loss": 4.608696937561035,
+ "step": 1186
+ },
+ {
+ "epoch": 16.489082969432314,
+ "grad_norm": 0.01602894812822342,
+ "learning_rate": 0.0006,
+ "loss": 4.620062351226807,
+ "step": 1187
+ },
+ {
+ "epoch": 16.503056768558952,
+ "grad_norm": 0.016615642234683037,
+ "learning_rate": 0.0006,
+ "loss": 4.69378662109375,
+ "step": 1188
+ },
+ {
+ "epoch": 16.51703056768559,
+ "grad_norm": 0.018576730042696,
+ "learning_rate": 0.0006,
+ "loss": 4.537041664123535,
+ "step": 1189
+ },
+ {
+ "epoch": 16.531004366812226,
+ "grad_norm": 0.018611151725053787,
+ "learning_rate": 0.0006,
+ "loss": 4.496169090270996,
+ "step": 1190
+ },
+ {
+ "epoch": 16.544978165938865,
+ "grad_norm": 0.018622413277626038,
+ "learning_rate": 0.0006,
+ "loss": 4.557358264923096,
+ "step": 1191
+ },
+ {
+ "epoch": 16.558951965065503,
+ "grad_norm": 0.01823544315993786,
+ "learning_rate": 0.0006,
+ "loss": 4.730792999267578,
+ "step": 1192
+ },
+ {
+ "epoch": 16.57292576419214,
+ "grad_norm": 0.018796846270561218,
+ "learning_rate": 0.0006,
+ "loss": 4.781791687011719,
+ "step": 1193
+ },
+ {
+ "epoch": 16.586899563318777,
+ "grad_norm": 0.01728292554616928,
+ "learning_rate": 0.0006,
+ "loss": 4.515257835388184,
+ "step": 1194
+ },
+ {
+ "epoch": 16.600873362445416,
+ "grad_norm": 0.01699589006602764,
+ "learning_rate": 0.0006,
+ "loss": 4.5996904373168945,
+ "step": 1195
+ },
+ {
+ "epoch": 16.61484716157205,
+ "grad_norm": 0.015905970707535744,
+ "learning_rate": 0.0006,
+ "loss": 4.601301193237305,
+ "step": 1196
+ },
+ {
+ "epoch": 16.62882096069869,
+ "grad_norm": 0.01338571310043335,
+ "learning_rate": 0.0006,
+ "loss": 4.557708740234375,
+ "step": 1197
+ },
+ {
+ "epoch": 16.64279475982533,
+ "grad_norm": 0.015298855490982533,
+ "learning_rate": 0.0006,
+ "loss": 4.542287349700928,
+ "step": 1198
+ },
+ {
+ "epoch": 16.656768558951963,
+ "grad_norm": 0.014753367751836777,
+ "learning_rate": 0.0006,
+ "loss": 4.595787048339844,
+ "step": 1199
+ },
+ {
+ "epoch": 16.670742358078602,
+ "grad_norm": 0.015527483075857162,
+ "learning_rate": 0.0006,
+ "loss": 4.607529163360596,
+ "step": 1200
+ },
+ {
+ "epoch": 16.68471615720524,
+ "grad_norm": 0.015042162500321865,
+ "learning_rate": 0.0006,
+ "loss": 4.686626434326172,
+ "step": 1201
+ },
+ {
+ "epoch": 16.69868995633188,
+ "grad_norm": 0.0159393772482872,
+ "learning_rate": 0.0006,
+ "loss": 4.47656774520874,
+ "step": 1202
+ },
+ {
+ "epoch": 16.712663755458514,
+ "grad_norm": 0.01532787922769785,
+ "learning_rate": 0.0006,
+ "loss": 4.55678653717041,
+ "step": 1203
+ },
+ {
+ "epoch": 16.726637554585153,
+ "grad_norm": 0.016332164406776428,
+ "learning_rate": 0.0006,
+ "loss": 4.51441764831543,
+ "step": 1204
+ },
+ {
+ "epoch": 16.74061135371179,
+ "grad_norm": 0.018901726230978966,
+ "learning_rate": 0.0006,
+ "loss": 4.57227897644043,
+ "step": 1205
+ },
+ {
+ "epoch": 16.754585152838427,
+ "grad_norm": 0.018589433282613754,
+ "learning_rate": 0.0006,
+ "loss": 4.550070762634277,
+ "step": 1206
+ },
+ {
+ "epoch": 16.768558951965066,
+ "grad_norm": 0.01634034886956215,
+ "learning_rate": 0.0006,
+ "loss": 4.586910724639893,
+ "step": 1207
+ },
+ {
+ "epoch": 16.782532751091704,
+ "grad_norm": 0.017599433660507202,
+ "learning_rate": 0.0006,
+ "loss": 4.5891923904418945,
+ "step": 1208
+ },
+ {
+ "epoch": 16.79650655021834,
+ "grad_norm": 0.01878446340560913,
+ "learning_rate": 0.0006,
+ "loss": 4.57227087020874,
+ "step": 1209
+ },
+ {
+ "epoch": 16.810480349344978,
+ "grad_norm": 0.01803469844162464,
+ "learning_rate": 0.0006,
+ "loss": 4.583845138549805,
+ "step": 1210
+ },
+ {
+ "epoch": 16.824454148471617,
+ "grad_norm": 0.018505144864320755,
+ "learning_rate": 0.0006,
+ "loss": 4.561151504516602,
+ "step": 1211
+ },
+ {
+ "epoch": 16.83842794759825,
+ "grad_norm": 0.01920081488788128,
+ "learning_rate": 0.0006,
+ "loss": 4.506834030151367,
+ "step": 1212
+ },
+ {
+ "epoch": 16.85240174672489,
+ "grad_norm": 0.0177712831646204,
+ "learning_rate": 0.0006,
+ "loss": 4.557195663452148,
+ "step": 1213
+ },
+ {
+ "epoch": 16.86637554585153,
+ "grad_norm": 0.019113918766379356,
+ "learning_rate": 0.0006,
+ "loss": 4.651227951049805,
+ "step": 1214
+ },
+ {
+ "epoch": 16.880349344978168,
+ "grad_norm": 0.02012539468705654,
+ "learning_rate": 0.0006,
+ "loss": 4.4952239990234375,
+ "step": 1215
+ },
+ {
+ "epoch": 16.894323144104803,
+ "grad_norm": 0.019993646070361137,
+ "learning_rate": 0.0006,
+ "loss": 4.517433166503906,
+ "step": 1216
+ },
+ {
+ "epoch": 16.90829694323144,
+ "grad_norm": 0.020041441544890404,
+ "learning_rate": 0.0006,
+ "loss": 4.489407539367676,
+ "step": 1217
+ },
+ {
+ "epoch": 16.92227074235808,
+ "grad_norm": 0.0183611661195755,
+ "learning_rate": 0.0006,
+ "loss": 4.5435638427734375,
+ "step": 1218
+ },
+ {
+ "epoch": 16.936244541484715,
+ "grad_norm": 0.019138429313898087,
+ "learning_rate": 0.0006,
+ "loss": 4.545949459075928,
+ "step": 1219
+ },
+ {
+ "epoch": 16.950218340611354,
+ "grad_norm": 0.018554266542196274,
+ "learning_rate": 0.0006,
+ "loss": 4.579620361328125,
+ "step": 1220
+ },
+ {
+ "epoch": 16.964192139737992,
+ "grad_norm": 0.0193545650690794,
+ "learning_rate": 0.0006,
+ "loss": 4.588255405426025,
+ "step": 1221
+ },
+ {
+ "epoch": 16.978165938864628,
+ "grad_norm": 0.019389165565371513,
+ "learning_rate": 0.0006,
+ "loss": 4.520698547363281,
+ "step": 1222
+ },
+ {
+ "epoch": 16.992139737991266,
+ "grad_norm": 0.016601379960775375,
+ "learning_rate": 0.0006,
+ "loss": 4.4809064865112305,
+ "step": 1223
+ },
+ {
+ "epoch": 17.0,
+ "grad_norm": 0.01771303080022335,
+ "learning_rate": 0.0006,
+ "loss": 4.721122741699219,
+ "step": 1224
+ },
+ {
+ "epoch": 17.0,
+ "eval_loss": 4.7903666496276855,
+ "eval_runtime": 56.9826,
+ "eval_samples_per_second": 42.855,
+ "eval_steps_per_second": 1.351,
+ "step": 1224
+ },
+ {
+ "epoch": 17.01397379912664,
+ "grad_norm": 0.017720289528369904,
+ "learning_rate": 0.0006,
+ "loss": 4.523834705352783,
+ "step": 1225
+ },
+ {
+ "epoch": 17.027947598253274,
+ "grad_norm": 0.020137161016464233,
+ "learning_rate": 0.0006,
+ "loss": 4.574061870574951,
+ "step": 1226
+ },
+ {
+ "epoch": 17.041921397379912,
+ "grad_norm": 0.02002314105629921,
+ "learning_rate": 0.0006,
+ "loss": 4.512407302856445,
+ "step": 1227
+ },
+ {
+ "epoch": 17.05589519650655,
+ "grad_norm": 0.019628183916211128,
+ "learning_rate": 0.0006,
+ "loss": 4.502533435821533,
+ "step": 1228
+ },
+ {
+ "epoch": 17.069868995633186,
+ "grad_norm": 0.019808458164334297,
+ "learning_rate": 0.0006,
+ "loss": 4.463156700134277,
+ "step": 1229
+ },
+ {
+ "epoch": 17.083842794759825,
+ "grad_norm": 0.021250786259770393,
+ "learning_rate": 0.0006,
+ "loss": 4.622682094573975,
+ "step": 1230
+ },
+ {
+ "epoch": 17.097816593886463,
+ "grad_norm": 0.022031694650650024,
+ "learning_rate": 0.0006,
+ "loss": 4.451882839202881,
+ "step": 1231
+ },
+ {
+ "epoch": 17.111790393013102,
+ "grad_norm": 0.01898987777531147,
+ "learning_rate": 0.0006,
+ "loss": 4.529331684112549,
+ "step": 1232
+ },
+ {
+ "epoch": 17.125764192139737,
+ "grad_norm": 0.018749834969639778,
+ "learning_rate": 0.0006,
+ "loss": 4.435815334320068,
+ "step": 1233
+ },
+ {
+ "epoch": 17.139737991266376,
+ "grad_norm": 0.01896674931049347,
+ "learning_rate": 0.0006,
+ "loss": 4.484029293060303,
+ "step": 1234
+ },
+ {
+ "epoch": 17.153711790393015,
+ "grad_norm": 0.018286997452378273,
+ "learning_rate": 0.0006,
+ "loss": 4.523510932922363,
+ "step": 1235
+ },
+ {
+ "epoch": 17.16768558951965,
+ "grad_norm": 0.017273059114813805,
+ "learning_rate": 0.0006,
+ "loss": 4.451767444610596,
+ "step": 1236
+ },
+ {
+ "epoch": 17.18165938864629,
+ "grad_norm": 0.016499150544404984,
+ "learning_rate": 0.0006,
+ "loss": 4.5568647384643555,
+ "step": 1237
+ },
+ {
+ "epoch": 17.195633187772927,
+ "grad_norm": 0.015934469178318977,
+ "learning_rate": 0.0006,
+ "loss": 4.4639997482299805,
+ "step": 1238
+ },
+ {
+ "epoch": 17.209606986899562,
+ "grad_norm": 0.01714209094643593,
+ "learning_rate": 0.0006,
+ "loss": 4.547614097595215,
+ "step": 1239
+ },
+ {
+ "epoch": 17.2235807860262,
+ "grad_norm": 0.019167495891451836,
+ "learning_rate": 0.0006,
+ "loss": 4.571097373962402,
+ "step": 1240
+ },
+ {
+ "epoch": 17.23755458515284,
+ "grad_norm": 0.018276996910572052,
+ "learning_rate": 0.0006,
+ "loss": 4.489764213562012,
+ "step": 1241
+ },
+ {
+ "epoch": 17.251528384279474,
+ "grad_norm": 0.021296866238117218,
+ "learning_rate": 0.0006,
+ "loss": 4.526913642883301,
+ "step": 1242
+ },
+ {
+ "epoch": 17.265502183406113,
+ "grad_norm": 0.022606629878282547,
+ "learning_rate": 0.0006,
+ "loss": 4.521624565124512,
+ "step": 1243
+ },
+ {
+ "epoch": 17.27947598253275,
+ "grad_norm": 0.021204711869359016,
+ "learning_rate": 0.0006,
+ "loss": 4.500864028930664,
+ "step": 1244
+ },
+ {
+ "epoch": 17.29344978165939,
+ "grad_norm": 0.01988849975168705,
+ "learning_rate": 0.0006,
+ "loss": 4.45162296295166,
+ "step": 1245
+ },
+ {
+ "epoch": 17.307423580786025,
+ "grad_norm": 0.020268384367227554,
+ "learning_rate": 0.0006,
+ "loss": 4.536198616027832,
+ "step": 1246
+ },
+ {
+ "epoch": 17.321397379912664,
+ "grad_norm": 0.020185677334666252,
+ "learning_rate": 0.0006,
+ "loss": 4.440262794494629,
+ "step": 1247
+ },
+ {
+ "epoch": 17.335371179039303,
+ "grad_norm": 0.020106853917241096,
+ "learning_rate": 0.0006,
+ "loss": 4.475907325744629,
+ "step": 1248
+ },
+ {
+ "epoch": 17.349344978165938,
+ "grad_norm": 0.018883297219872475,
+ "learning_rate": 0.0006,
+ "loss": 4.640661716461182,
+ "step": 1249
+ },
+ {
+ "epoch": 17.363318777292577,
+ "grad_norm": 0.01852620206773281,
+ "learning_rate": 0.0006,
+ "loss": 4.578243255615234,
+ "step": 1250
+ },
+ {
+ "epoch": 17.377292576419215,
+ "grad_norm": 0.01886451430618763,
+ "learning_rate": 0.0006,
+ "loss": 4.546026229858398,
+ "step": 1251
+ },
+ {
+ "epoch": 17.39126637554585,
+ "grad_norm": 0.01777452416718006,
+ "learning_rate": 0.0006,
+ "loss": 4.521542549133301,
+ "step": 1252
+ },
+ {
+ "epoch": 17.40524017467249,
+ "grad_norm": 0.016626447439193726,
+ "learning_rate": 0.0006,
+ "loss": 4.52629280090332,
+ "step": 1253
+ },
+ {
+ "epoch": 17.419213973799128,
+ "grad_norm": 0.01599477231502533,
+ "learning_rate": 0.0006,
+ "loss": 4.4796037673950195,
+ "step": 1254
+ },
+ {
+ "epoch": 17.433187772925763,
+ "grad_norm": 0.017021458595991135,
+ "learning_rate": 0.0006,
+ "loss": 4.468813419342041,
+ "step": 1255
+ },
+ {
+ "epoch": 17.4471615720524,
+ "grad_norm": 0.016925333067774773,
+ "learning_rate": 0.0006,
+ "loss": 4.584898948669434,
+ "step": 1256
+ },
+ {
+ "epoch": 17.46113537117904,
+ "grad_norm": 0.01714998669922352,
+ "learning_rate": 0.0006,
+ "loss": 4.523544788360596,
+ "step": 1257
+ },
+ {
+ "epoch": 17.475109170305675,
+ "grad_norm": 0.01802288182079792,
+ "learning_rate": 0.0006,
+ "loss": 4.595581531524658,
+ "step": 1258
+ },
+ {
+ "epoch": 17.489082969432314,
+ "grad_norm": 0.0180289875715971,
+ "learning_rate": 0.0006,
+ "loss": 4.394440650939941,
+ "step": 1259
+ },
+ {
+ "epoch": 17.503056768558952,
+ "grad_norm": 0.018504245206713676,
+ "learning_rate": 0.0006,
+ "loss": 4.4794416427612305,
+ "step": 1260
+ },
+ {
+ "epoch": 17.51703056768559,
+ "grad_norm": 0.019973233342170715,
+ "learning_rate": 0.0006,
+ "loss": 4.4241437911987305,
+ "step": 1261
+ },
+ {
+ "epoch": 17.531004366812226,
+ "grad_norm": 0.020063083618879318,
+ "learning_rate": 0.0006,
+ "loss": 4.48068380355835,
+ "step": 1262
+ },
+ {
+ "epoch": 17.544978165938865,
+ "grad_norm": 0.02066640742123127,
+ "learning_rate": 0.0006,
+ "loss": 4.628993034362793,
+ "step": 1263
+ },
+ {
+ "epoch": 17.558951965065503,
+ "grad_norm": 0.02206416428089142,
+ "learning_rate": 0.0006,
+ "loss": 4.564790725708008,
+ "step": 1264
+ },
+ {
+ "epoch": 17.57292576419214,
+ "grad_norm": 0.020640335977077484,
+ "learning_rate": 0.0006,
+ "loss": 4.578705787658691,
+ "step": 1265
+ },
+ {
+ "epoch": 17.586899563318777,
+ "grad_norm": 0.019093405455350876,
+ "learning_rate": 0.0006,
+ "loss": 4.5023603439331055,
+ "step": 1266
+ },
+ {
+ "epoch": 17.600873362445416,
+ "grad_norm": 0.019459193572402,
+ "learning_rate": 0.0006,
+ "loss": 4.594531059265137,
+ "step": 1267
+ },
+ {
+ "epoch": 17.61484716157205,
+ "grad_norm": 0.02021171525120735,
+ "learning_rate": 0.0006,
+ "loss": 4.770299911499023,
+ "step": 1268
+ },
+ {
+ "epoch": 17.62882096069869,
+ "grad_norm": 0.021161191165447235,
+ "learning_rate": 0.0006,
+ "loss": 4.5380401611328125,
+ "step": 1269
+ },
+ {
+ "epoch": 17.64279475982533,
+ "grad_norm": 0.021013403311371803,
+ "learning_rate": 0.0006,
+ "loss": 4.455524921417236,
+ "step": 1270
+ },
+ {
+ "epoch": 17.656768558951963,
+ "grad_norm": 0.020336592569947243,
+ "learning_rate": 0.0006,
+ "loss": 4.533672332763672,
+ "step": 1271
+ },
+ {
+ "epoch": 17.670742358078602,
+ "grad_norm": 0.020159434527158737,
+ "learning_rate": 0.0006,
+ "loss": 4.596151351928711,
+ "step": 1272
+ },
+ {
+ "epoch": 17.68471615720524,
+ "grad_norm": 0.020432960242033005,
+ "learning_rate": 0.0006,
+ "loss": 4.55531644821167,
+ "step": 1273
+ },
+ {
+ "epoch": 17.69868995633188,
+ "grad_norm": 0.025466229766607285,
+ "learning_rate": 0.0006,
+ "loss": 4.566912651062012,
+ "step": 1274
+ },
+ {
+ "epoch": 17.712663755458514,
+ "grad_norm": 0.026537558063864708,
+ "learning_rate": 0.0006,
+ "loss": 4.526003837585449,
+ "step": 1275
+ },
+ {
+ "epoch": 17.726637554585153,
+ "grad_norm": 0.024397607892751694,
+ "learning_rate": 0.0006,
+ "loss": 4.541755676269531,
+ "step": 1276
+ },
+ {
+ "epoch": 17.74061135371179,
+ "grad_norm": 0.022062312811613083,
+ "learning_rate": 0.0006,
+ "loss": 4.6532793045043945,
+ "step": 1277
+ },
+ {
+ "epoch": 17.754585152838427,
+ "grad_norm": 0.02204282023012638,
+ "learning_rate": 0.0006,
+ "loss": 4.582825660705566,
+ "step": 1278
+ },
+ {
+ "epoch": 17.768558951965066,
+ "grad_norm": 0.020844416692852974,
+ "learning_rate": 0.0006,
+ "loss": 4.522336483001709,
+ "step": 1279
+ },
+ {
+ "epoch": 17.782532751091704,
+ "grad_norm": 0.019815703853964806,
+ "learning_rate": 0.0006,
+ "loss": 4.607398509979248,
+ "step": 1280
+ },
+ {
+ "epoch": 17.79650655021834,
+ "grad_norm": 0.016676954925060272,
+ "learning_rate": 0.0006,
+ "loss": 4.446206092834473,
+ "step": 1281
+ },
+ {
+ "epoch": 17.810480349344978,
+ "grad_norm": 0.0166452769190073,
+ "learning_rate": 0.0006,
+ "loss": 4.505931854248047,
+ "step": 1282
+ },
+ {
+ "epoch": 17.824454148471617,
+ "grad_norm": 0.015928907319903374,
+ "learning_rate": 0.0006,
+ "loss": 4.455267429351807,
+ "step": 1283
+ },
+ {
+ "epoch": 17.83842794759825,
+ "grad_norm": 0.015219560824334621,
+ "learning_rate": 0.0006,
+ "loss": 4.5623674392700195,
+ "step": 1284
+ },
+ {
+ "epoch": 17.85240174672489,
+ "grad_norm": 0.015125355683267117,
+ "learning_rate": 0.0006,
+ "loss": 4.554104804992676,
+ "step": 1285
+ },
+ {
+ "epoch": 17.86637554585153,
+ "grad_norm": 0.015612718649208546,
+ "learning_rate": 0.0006,
+ "loss": 4.484847068786621,
+ "step": 1286
+ },
+ {
+ "epoch": 17.880349344978168,
+ "grad_norm": 0.015215662308037281,
+ "learning_rate": 0.0006,
+ "loss": 4.4802093505859375,
+ "step": 1287
+ },
+ {
+ "epoch": 17.894323144104803,
+ "grad_norm": 0.01511684525758028,
+ "learning_rate": 0.0006,
+ "loss": 4.537298679351807,
+ "step": 1288
+ },
+ {
+ "epoch": 17.90829694323144,
+ "grad_norm": 0.013841322623193264,
+ "learning_rate": 0.0006,
+ "loss": 4.427488327026367,
+ "step": 1289
+ },
+ {
+ "epoch": 17.92227074235808,
+ "grad_norm": 0.015115964226424694,
+ "learning_rate": 0.0006,
+ "loss": 4.4547119140625,
+ "step": 1290
+ },
+ {
+ "epoch": 17.936244541484715,
+ "grad_norm": 0.016411006450653076,
+ "learning_rate": 0.0006,
+ "loss": 4.52994966506958,
+ "step": 1291
+ },
+ {
+ "epoch": 17.950218340611354,
+ "grad_norm": 0.017313152551651,
+ "learning_rate": 0.0006,
+ "loss": 4.5337653160095215,
+ "step": 1292
+ },
+ {
+ "epoch": 17.964192139737992,
+ "grad_norm": 0.016559796407818794,
+ "learning_rate": 0.0006,
+ "loss": 4.5277910232543945,
+ "step": 1293
+ },
+ {
+ "epoch": 17.978165938864628,
+ "grad_norm": 0.015854761004447937,
+ "learning_rate": 0.0006,
+ "loss": 4.6334309577941895,
+ "step": 1294
+ },
+ {
+ "epoch": 17.992139737991266,
+ "grad_norm": 0.015658238902688026,
+ "learning_rate": 0.0006,
+ "loss": 4.613372802734375,
+ "step": 1295
+ },
+ {
+ "epoch": 18.0,
+ "grad_norm": 0.0169155802577734,
+ "learning_rate": 0.0006,
+ "loss": 4.527356147766113,
+ "step": 1296
+ },
+ {
+ "epoch": 18.0,
+ "eval_loss": 4.764543056488037,
+ "eval_runtime": 56.5998,
+ "eval_samples_per_second": 43.145,
+ "eval_steps_per_second": 1.36,
+ "step": 1296
+ },
+ {
+ "epoch": 18.01397379912664,
+ "grad_norm": 0.015348963439464569,
+ "learning_rate": 0.0006,
+ "loss": 4.466423988342285,
+ "step": 1297
+ },
+ {
+ "epoch": 18.027947598253274,
+ "grad_norm": 0.01768588088452816,
+ "learning_rate": 0.0006,
+ "loss": 4.545933246612549,
+ "step": 1298
+ },
+ {
+ "epoch": 18.041921397379912,
+ "grad_norm": 0.017381226643919945,
+ "learning_rate": 0.0006,
+ "loss": 4.494060039520264,
+ "step": 1299
+ },
+ {
+ "epoch": 18.05589519650655,
+ "grad_norm": 0.019002335146069527,
+ "learning_rate": 0.0006,
+ "loss": 4.395200729370117,
+ "step": 1300
+ },
+ {
+ "epoch": 18.069868995633186,
+ "grad_norm": 0.01827932707965374,
+ "learning_rate": 0.0006,
+ "loss": 4.424848556518555,
+ "step": 1301
+ },
+ {
+ "epoch": 18.083842794759825,
+ "grad_norm": 0.016383031383156776,
+ "learning_rate": 0.0006,
+ "loss": 4.494722366333008,
+ "step": 1302
+ },
+ {
+ "epoch": 18.097816593886463,
+ "grad_norm": 0.016265157610177994,
+ "learning_rate": 0.0006,
+ "loss": 4.416522979736328,
+ "step": 1303
+ },
+ {
+ "epoch": 18.111790393013102,
+ "grad_norm": 0.01702386513352394,
+ "learning_rate": 0.0006,
+ "loss": 4.600103378295898,
+ "step": 1304
+ },
+ {
+ "epoch": 18.125764192139737,
+ "grad_norm": 0.018495498225092888,
+ "learning_rate": 0.0006,
+ "loss": 4.588401794433594,
+ "step": 1305
+ },
+ {
+ "epoch": 18.139737991266376,
+ "grad_norm": 0.017921684309840202,
+ "learning_rate": 0.0006,
+ "loss": 4.478075981140137,
+ "step": 1306
+ },
+ {
+ "epoch": 18.153711790393015,
+ "grad_norm": 0.01801278069615364,
+ "learning_rate": 0.0006,
+ "loss": 4.413186550140381,
+ "step": 1307
+ },
+ {
+ "epoch": 18.16768558951965,
+ "grad_norm": 0.017211275175213814,
+ "learning_rate": 0.0006,
+ "loss": 4.461341857910156,
+ "step": 1308
+ },
+ {
+ "epoch": 18.18165938864629,
+ "grad_norm": 0.017589328810572624,
+ "learning_rate": 0.0006,
+ "loss": 4.546969413757324,
+ "step": 1309
+ },
+ {
+ "epoch": 18.195633187772927,
+ "grad_norm": 0.01797698810696602,
+ "learning_rate": 0.0006,
+ "loss": 4.530797004699707,
+ "step": 1310
+ },
+ {
+ "epoch": 18.209606986899562,
+ "grad_norm": 0.020408619195222855,
+ "learning_rate": 0.0006,
+ "loss": 4.62999153137207,
+ "step": 1311
+ },
+ {
+ "epoch": 18.2235807860262,
+ "grad_norm": 0.01994038000702858,
+ "learning_rate": 0.0006,
+ "loss": 4.586811065673828,
+ "step": 1312
+ },
+ {
+ "epoch": 18.23755458515284,
+ "grad_norm": 0.019381340593099594,
+ "learning_rate": 0.0006,
+ "loss": 4.4769368171691895,
+ "step": 1313
+ },
+ {
+ "epoch": 18.251528384279474,
+ "grad_norm": 0.020484555512666702,
+ "learning_rate": 0.0006,
+ "loss": 4.524302959442139,
+ "step": 1314
+ },
+ {
+ "epoch": 18.265502183406113,
+ "grad_norm": 0.02426416426897049,
+ "learning_rate": 0.0006,
+ "loss": 4.561705589294434,
+ "step": 1315
+ },
+ {
+ "epoch": 18.27947598253275,
+ "grad_norm": 0.024748101830482483,
+ "learning_rate": 0.0006,
+ "loss": 4.4654316902160645,
+ "step": 1316
+ },
+ {
+ "epoch": 18.29344978165939,
+ "grad_norm": 0.020703328773379326,
+ "learning_rate": 0.0006,
+ "loss": 4.53033447265625,
+ "step": 1317
+ },
+ {
+ "epoch": 18.307423580786025,
+ "grad_norm": 0.01765458658337593,
+ "learning_rate": 0.0006,
+ "loss": 4.5154266357421875,
+ "step": 1318
+ },
+ {
+ "epoch": 18.321397379912664,
+ "grad_norm": 0.018646420910954475,
+ "learning_rate": 0.0006,
+ "loss": 4.5224714279174805,
+ "step": 1319
+ },
+ {
+ "epoch": 18.335371179039303,
+ "grad_norm": 0.01907474733889103,
+ "learning_rate": 0.0006,
+ "loss": 4.5055437088012695,
+ "step": 1320
+ },
+ {
+ "epoch": 18.349344978165938,
+ "grad_norm": 0.018179846927523613,
+ "learning_rate": 0.0006,
+ "loss": 4.462161064147949,
+ "step": 1321
+ },
+ {
+ "epoch": 18.363318777292577,
+ "grad_norm": 0.016697878018021584,
+ "learning_rate": 0.0006,
+ "loss": 4.473718643188477,
+ "step": 1322
+ },
+ {
+ "epoch": 18.377292576419215,
+ "grad_norm": 0.017504019662737846,
+ "learning_rate": 0.0006,
+ "loss": 4.474348068237305,
+ "step": 1323
+ },
+ {
+ "epoch": 18.39126637554585,
+ "grad_norm": 0.018889913335442543,
+ "learning_rate": 0.0006,
+ "loss": 4.4531636238098145,
+ "step": 1324
+ },
+ {
+ "epoch": 18.40524017467249,
+ "grad_norm": 0.019167251884937286,
+ "learning_rate": 0.0006,
+ "loss": 4.4278364181518555,
+ "step": 1325
+ },
+ {
+ "epoch": 18.419213973799128,
+ "grad_norm": 0.015895741060376167,
+ "learning_rate": 0.0006,
+ "loss": 4.474592208862305,
+ "step": 1326
+ },
+ {
+ "epoch": 18.433187772925763,
+ "grad_norm": 0.01531882956624031,
+ "learning_rate": 0.0006,
+ "loss": 4.403757572174072,
+ "step": 1327
+ },
+ {
+ "epoch": 18.4471615720524,
+ "grad_norm": 0.01565849967300892,
+ "learning_rate": 0.0006,
+ "loss": 4.507664680480957,
+ "step": 1328
+ },
+ {
+ "epoch": 18.46113537117904,
+ "grad_norm": 0.01602315716445446,
+ "learning_rate": 0.0006,
+ "loss": 4.5648908615112305,
+ "step": 1329
+ },
+ {
+ "epoch": 18.475109170305675,
+ "grad_norm": 0.016467366367578506,
+ "learning_rate": 0.0006,
+ "loss": 4.428762912750244,
+ "step": 1330
+ },
+ {
+ "epoch": 18.489082969432314,
+ "grad_norm": 0.01604127697646618,
+ "learning_rate": 0.0006,
+ "loss": 4.527771472930908,
+ "step": 1331
+ },
+ {
+ "epoch": 18.503056768558952,
+ "grad_norm": 0.018825586885213852,
+ "learning_rate": 0.0006,
+ "loss": 4.485386848449707,
+ "step": 1332
+ },
+ {
+ "epoch": 18.51703056768559,
+ "grad_norm": 0.019258007407188416,
+ "learning_rate": 0.0006,
+ "loss": 4.414972305297852,
+ "step": 1333
+ },
+ {
+ "epoch": 18.531004366812226,
+ "grad_norm": 0.018192021176218987,
+ "learning_rate": 0.0006,
+ "loss": 4.412166595458984,
+ "step": 1334
+ },
+ {
+ "epoch": 18.544978165938865,
+ "grad_norm": 0.016120698302984238,
+ "learning_rate": 0.0006,
+ "loss": 4.484317779541016,
+ "step": 1335
+ },
+ {
+ "epoch": 18.558951965065503,
+ "grad_norm": 0.016496622934937477,
+ "learning_rate": 0.0006,
+ "loss": 4.514986515045166,
+ "step": 1336
+ },
+ {
+ "epoch": 18.57292576419214,
+ "grad_norm": 0.01795116998255253,
+ "learning_rate": 0.0006,
+ "loss": 4.41473388671875,
+ "step": 1337
+ },
+ {
+ "epoch": 18.586899563318777,
+ "grad_norm": 0.019062276929616928,
+ "learning_rate": 0.0006,
+ "loss": 4.601996898651123,
+ "step": 1338
+ },
+ {
+ "epoch": 18.600873362445416,
+ "grad_norm": 0.017174631357192993,
+ "learning_rate": 0.0006,
+ "loss": 4.571218490600586,
+ "step": 1339
+ },
+ {
+ "epoch": 18.61484716157205,
+ "grad_norm": 0.014015606604516506,
+ "learning_rate": 0.0006,
+ "loss": 4.526606559753418,
+ "step": 1340
+ },
+ {
+ "epoch": 18.62882096069869,
+ "grad_norm": 0.015358510427176952,
+ "learning_rate": 0.0006,
+ "loss": 4.604286193847656,
+ "step": 1341
+ },
+ {
+ "epoch": 18.64279475982533,
+ "grad_norm": 0.01619391329586506,
+ "learning_rate": 0.0006,
+ "loss": 4.560883522033691,
+ "step": 1342
+ },
+ {
+ "epoch": 18.656768558951963,
+ "grad_norm": 0.015972964465618134,
+ "learning_rate": 0.0006,
+ "loss": 4.479755401611328,
+ "step": 1343
+ },
+ {
+ "epoch": 18.670742358078602,
+ "grad_norm": 0.016861025243997574,
+ "learning_rate": 0.0006,
+ "loss": 4.474471092224121,
+ "step": 1344
+ },
+ {
+ "epoch": 18.68471615720524,
+ "grad_norm": 0.01822078227996826,
+ "learning_rate": 0.0006,
+ "loss": 4.509489059448242,
+ "step": 1345
+ },
+ {
+ "epoch": 18.69868995633188,
+ "grad_norm": 0.019415007904171944,
+ "learning_rate": 0.0006,
+ "loss": 4.508941173553467,
+ "step": 1346
+ },
+ {
+ "epoch": 18.712663755458514,
+ "grad_norm": 0.019856126978993416,
+ "learning_rate": 0.0006,
+ "loss": 4.446670055389404,
+ "step": 1347
+ },
+ {
+ "epoch": 18.726637554585153,
+ "grad_norm": 0.01842520199716091,
+ "learning_rate": 0.0006,
+ "loss": 4.524422645568848,
+ "step": 1348
+ },
+ {
+ "epoch": 18.74061135371179,
+ "grad_norm": 0.017196480184793472,
+ "learning_rate": 0.0006,
+ "loss": 4.498466968536377,
+ "step": 1349
+ },
+ {
+ "epoch": 18.754585152838427,
+ "grad_norm": 0.01823010854423046,
+ "learning_rate": 0.0006,
+ "loss": 4.534684181213379,
+ "step": 1350
+ },
+ {
+ "epoch": 18.768558951965066,
+ "grad_norm": 0.020398065447807312,
+ "learning_rate": 0.0006,
+ "loss": 4.456494331359863,
+ "step": 1351
+ },
+ {
+ "epoch": 18.782532751091704,
+ "grad_norm": 0.022992262616753578,
+ "learning_rate": 0.0006,
+ "loss": 4.476374626159668,
+ "step": 1352
+ },
+ {
+ "epoch": 18.79650655021834,
+ "grad_norm": 0.02235420234501362,
+ "learning_rate": 0.0006,
+ "loss": 4.587776184082031,
+ "step": 1353
+ },
+ {
+ "epoch": 18.810480349344978,
+ "grad_norm": 0.02093626745045185,
+ "learning_rate": 0.0006,
+ "loss": 4.424172401428223,
+ "step": 1354
+ },
+ {
+ "epoch": 18.824454148471617,
+ "grad_norm": 0.01997302658855915,
+ "learning_rate": 0.0006,
+ "loss": 4.448066711425781,
+ "step": 1355
+ },
+ {
+ "epoch": 18.83842794759825,
+ "grad_norm": 0.018265459686517715,
+ "learning_rate": 0.0006,
+ "loss": 4.436610698699951,
+ "step": 1356
+ },
+ {
+ "epoch": 18.85240174672489,
+ "grad_norm": 0.01673833839595318,
+ "learning_rate": 0.0006,
+ "loss": 4.5502214431762695,
+ "step": 1357
+ },
+ {
+ "epoch": 18.86637554585153,
+ "grad_norm": 0.017908833920955658,
+ "learning_rate": 0.0006,
+ "loss": 4.550530433654785,
+ "step": 1358
+ },
+ {
+ "epoch": 18.880349344978168,
+ "grad_norm": 0.01796654611825943,
+ "learning_rate": 0.0006,
+ "loss": 4.531463146209717,
+ "step": 1359
+ },
+ {
+ "epoch": 18.894323144104803,
+ "grad_norm": 0.01646183803677559,
+ "learning_rate": 0.0006,
+ "loss": 4.498582363128662,
+ "step": 1360
+ },
+ {
+ "epoch": 18.90829694323144,
+ "grad_norm": 0.015430327504873276,
+ "learning_rate": 0.0006,
+ "loss": 4.52571439743042,
+ "step": 1361
+ },
+ {
+ "epoch": 18.92227074235808,
+ "grad_norm": 0.016527952626347542,
+ "learning_rate": 0.0006,
+ "loss": 4.547013759613037,
+ "step": 1362
+ },
+ {
+ "epoch": 18.936244541484715,
+ "grad_norm": 0.016263660043478012,
+ "learning_rate": 0.0006,
+ "loss": 4.468754291534424,
+ "step": 1363
+ },
+ {
+ "epoch": 18.950218340611354,
+ "grad_norm": 0.016347210854291916,
+ "learning_rate": 0.0006,
+ "loss": 4.450741767883301,
+ "step": 1364
+ },
+ {
+ "epoch": 18.964192139737992,
+ "grad_norm": 0.01701398566365242,
+ "learning_rate": 0.0006,
+ "loss": 4.534899711608887,
+ "step": 1365
+ },
+ {
+ "epoch": 18.978165938864628,
+ "grad_norm": 0.017693081870675087,
+ "learning_rate": 0.0006,
+ "loss": 4.460553169250488,
+ "step": 1366
+ },
+ {
+ "epoch": 18.992139737991266,
+ "grad_norm": 0.018746308982372284,
+ "learning_rate": 0.0006,
+ "loss": 4.538792610168457,
+ "step": 1367
+ },
+ {
+ "epoch": 19.0,
+ "grad_norm": 0.020099416375160217,
+ "learning_rate": 0.0006,
+ "loss": 4.449617385864258,
+ "step": 1368
+ },
+ {
+ "epoch": 19.0,
+ "eval_loss": 4.7756571769714355,
+ "eval_runtime": 57.1804,
+ "eval_samples_per_second": 42.707,
+ "eval_steps_per_second": 1.347,
+ "step": 1368
+ },
+ {
+ "epoch": 19.01397379912664,
+ "grad_norm": 0.018726933747529984,
+ "learning_rate": 0.0006,
+ "loss": 4.446059703826904,
+ "step": 1369
+ },
+ {
+ "epoch": 19.027947598253274,
+ "grad_norm": 0.01956087350845337,
+ "learning_rate": 0.0006,
+ "loss": 4.4520039558410645,
+ "step": 1370
+ },
+ {
+ "epoch": 19.041921397379912,
+ "grad_norm": 0.020092811435461044,
+ "learning_rate": 0.0006,
+ "loss": 4.543343544006348,
+ "step": 1371
+ },
+ {
+ "epoch": 19.05589519650655,
+ "grad_norm": 0.01843622885644436,
+ "learning_rate": 0.0006,
+ "loss": 4.535579204559326,
+ "step": 1372
+ },
+ {
+ "epoch": 19.069868995633186,
+ "grad_norm": 0.016989488154649734,
+ "learning_rate": 0.0006,
+ "loss": 4.52316951751709,
+ "step": 1373
+ },
+ {
+ "epoch": 19.083842794759825,
+ "grad_norm": 0.01805422455072403,
+ "learning_rate": 0.0006,
+ "loss": 4.5393266677856445,
+ "step": 1374
+ },
+ {
+ "epoch": 19.097816593886463,
+ "grad_norm": 0.01874629408121109,
+ "learning_rate": 0.0006,
+ "loss": 4.422059059143066,
+ "step": 1375
+ },
+ {
+ "epoch": 19.111790393013102,
+ "grad_norm": 0.01804221048951149,
+ "learning_rate": 0.0006,
+ "loss": 4.50508451461792,
+ "step": 1376
+ },
+ {
+ "epoch": 19.125764192139737,
+ "grad_norm": 0.01785474829375744,
+ "learning_rate": 0.0006,
+ "loss": 4.412599563598633,
+ "step": 1377
+ },
+ {
+ "epoch": 19.139737991266376,
+ "grad_norm": 0.01971791312098503,
+ "learning_rate": 0.0006,
+ "loss": 4.518318176269531,
+ "step": 1378
+ },
+ {
+ "epoch": 19.153711790393015,
+ "grad_norm": 0.02061633951961994,
+ "learning_rate": 0.0006,
+ "loss": 4.332393169403076,
+ "step": 1379
+ },
+ {
+ "epoch": 19.16768558951965,
+ "grad_norm": 0.020563429221510887,
+ "learning_rate": 0.0006,
+ "loss": 4.445863723754883,
+ "step": 1380
+ },
+ {
+ "epoch": 19.18165938864629,
+ "grad_norm": 0.019995713606476784,
+ "learning_rate": 0.0006,
+ "loss": 4.497312068939209,
+ "step": 1381
+ },
+ {
+ "epoch": 19.195633187772927,
+ "grad_norm": 0.019823070615530014,
+ "learning_rate": 0.0006,
+ "loss": 4.475085258483887,
+ "step": 1382
+ },
+ {
+ "epoch": 19.209606986899562,
+ "grad_norm": 0.019039355218410492,
+ "learning_rate": 0.0006,
+ "loss": 4.50432014465332,
+ "step": 1383
+ },
+ {
+ "epoch": 19.2235807860262,
+ "grad_norm": 0.017876973375678062,
+ "learning_rate": 0.0006,
+ "loss": 4.384788513183594,
+ "step": 1384
+ },
+ {
+ "epoch": 19.23755458515284,
+ "grad_norm": 0.018697699531912804,
+ "learning_rate": 0.0006,
+ "loss": 4.410778045654297,
+ "step": 1385
+ },
+ {
+ "epoch": 19.251528384279474,
+ "grad_norm": 0.018826806917786598,
+ "learning_rate": 0.0006,
+ "loss": 4.524034023284912,
+ "step": 1386
+ },
+ {
+ "epoch": 19.265502183406113,
+ "grad_norm": 0.01824226602911949,
+ "learning_rate": 0.0006,
+ "loss": 4.526393890380859,
+ "step": 1387
+ },
+ {
+ "epoch": 19.27947598253275,
+ "grad_norm": 0.01753208599984646,
+ "learning_rate": 0.0006,
+ "loss": 4.493703365325928,
+ "step": 1388
+ },
+ {
+ "epoch": 19.29344978165939,
+ "grad_norm": 0.01742432825267315,
+ "learning_rate": 0.0006,
+ "loss": 4.392831802368164,
+ "step": 1389
+ },
+ {
+ "epoch": 19.307423580786025,
+ "grad_norm": 0.017099356278777122,
+ "learning_rate": 0.0006,
+ "loss": 4.578668594360352,
+ "step": 1390
+ },
+ {
+ "epoch": 19.321397379912664,
+ "grad_norm": 0.017946023494005203,
+ "learning_rate": 0.0006,
+ "loss": 4.456718444824219,
+ "step": 1391
+ },
+ {
+ "epoch": 19.335371179039303,
+ "grad_norm": 0.017705943435430527,
+ "learning_rate": 0.0006,
+ "loss": 4.553682327270508,
+ "step": 1392
+ },
+ {
+ "epoch": 19.349344978165938,
+ "grad_norm": 0.016647523269057274,
+ "learning_rate": 0.0006,
+ "loss": 4.525016784667969,
+ "step": 1393
+ },
+ {
+ "epoch": 19.363318777292577,
+ "grad_norm": 0.015407416969537735,
+ "learning_rate": 0.0006,
+ "loss": 4.500231742858887,
+ "step": 1394
+ },
+ {
+ "epoch": 19.377292576419215,
+ "grad_norm": 0.017264019697904587,
+ "learning_rate": 0.0006,
+ "loss": 4.4615349769592285,
+ "step": 1395
+ },
+ {
+ "epoch": 19.39126637554585,
+ "grad_norm": 0.01871555484831333,
+ "learning_rate": 0.0006,
+ "loss": 4.281026840209961,
+ "step": 1396
+ },
+ {
+ "epoch": 19.40524017467249,
+ "grad_norm": 0.017361650243401527,
+ "learning_rate": 0.0006,
+ "loss": 4.479382514953613,
+ "step": 1397
+ },
+ {
+ "epoch": 19.419213973799128,
+ "grad_norm": 0.019781362265348434,
+ "learning_rate": 0.0006,
+ "loss": 4.402320861816406,
+ "step": 1398
+ },
+ {
+ "epoch": 19.433187772925763,
+ "grad_norm": 0.019043298438191414,
+ "learning_rate": 0.0006,
+ "loss": 4.476983070373535,
+ "step": 1399
+ },
+ {
+ "epoch": 19.4471615720524,
+ "grad_norm": 0.017137985676527023,
+ "learning_rate": 0.0006,
+ "loss": 4.418689250946045,
+ "step": 1400
+ },
+ {
+ "epoch": 19.46113537117904,
+ "grad_norm": 0.016481833532452583,
+ "learning_rate": 0.0006,
+ "loss": 4.579145431518555,
+ "step": 1401
+ },
+ {
+ "epoch": 19.475109170305675,
+ "grad_norm": 0.016507714986801147,
+ "learning_rate": 0.0006,
+ "loss": 4.519158363342285,
+ "step": 1402
+ },
+ {
+ "epoch": 19.489082969432314,
+ "grad_norm": 0.018092872574925423,
+ "learning_rate": 0.0006,
+ "loss": 4.513748645782471,
+ "step": 1403
+ },
+ {
+ "epoch": 19.503056768558952,
+ "grad_norm": 0.018153440207242966,
+ "learning_rate": 0.0006,
+ "loss": 4.645384311676025,
+ "step": 1404
+ },
+ {
+ "epoch": 19.51703056768559,
+ "grad_norm": 0.01657119207084179,
+ "learning_rate": 0.0006,
+ "loss": 4.399757385253906,
+ "step": 1405
+ },
+ {
+ "epoch": 19.531004366812226,
+ "grad_norm": 0.0156280305236578,
+ "learning_rate": 0.0006,
+ "loss": 4.467836380004883,
+ "step": 1406
+ },
+ {
+ "epoch": 19.544978165938865,
+ "grad_norm": 0.016838403418660164,
+ "learning_rate": 0.0006,
+ "loss": 4.529176235198975,
+ "step": 1407
+ },
+ {
+ "epoch": 19.558951965065503,
+ "grad_norm": 0.01743236929178238,
+ "learning_rate": 0.0006,
+ "loss": 4.440032482147217,
+ "step": 1408
+ },
+ {
+ "epoch": 19.57292576419214,
+ "grad_norm": 0.01553610060364008,
+ "learning_rate": 0.0006,
+ "loss": 4.425588607788086,
+ "step": 1409
+ },
+ {
+ "epoch": 19.586899563318777,
+ "grad_norm": 0.017259271815419197,
+ "learning_rate": 0.0006,
+ "loss": 4.405452728271484,
+ "step": 1410
+ },
+ {
+ "epoch": 19.600873362445416,
+ "grad_norm": 0.015568481758236885,
+ "learning_rate": 0.0006,
+ "loss": 4.444279193878174,
+ "step": 1411
+ },
+ {
+ "epoch": 19.61484716157205,
+ "grad_norm": 0.015331248752772808,
+ "learning_rate": 0.0006,
+ "loss": 4.451846122741699,
+ "step": 1412
+ },
+ {
+ "epoch": 19.62882096069869,
+ "grad_norm": 0.015667378902435303,
+ "learning_rate": 0.0006,
+ "loss": 4.533931255340576,
+ "step": 1413
+ },
+ {
+ "epoch": 19.64279475982533,
+ "grad_norm": 0.016177602112293243,
+ "learning_rate": 0.0006,
+ "loss": 4.467313289642334,
+ "step": 1414
+ },
+ {
+ "epoch": 19.656768558951963,
+ "grad_norm": 0.01733129471540451,
+ "learning_rate": 0.0006,
+ "loss": 4.496893882751465,
+ "step": 1415
+ },
+ {
+ "epoch": 19.670742358078602,
+ "grad_norm": 0.016539594158530235,
+ "learning_rate": 0.0006,
+ "loss": 4.401961326599121,
+ "step": 1416
+ },
+ {
+ "epoch": 19.68471615720524,
+ "grad_norm": 0.016058551147580147,
+ "learning_rate": 0.0006,
+ "loss": 4.503375053405762,
+ "step": 1417
+ },
+ {
+ "epoch": 19.69868995633188,
+ "grad_norm": 0.01715777814388275,
+ "learning_rate": 0.0006,
+ "loss": 4.3808369636535645,
+ "step": 1418
+ },
+ {
+ "epoch": 19.712663755458514,
+ "grad_norm": 0.01681119203567505,
+ "learning_rate": 0.0006,
+ "loss": 4.5469560623168945,
+ "step": 1419
+ },
+ {
+ "epoch": 19.726637554585153,
+ "grad_norm": 0.01634550653398037,
+ "learning_rate": 0.0006,
+ "loss": 4.328197956085205,
+ "step": 1420
+ },
+ {
+ "epoch": 19.74061135371179,
+ "grad_norm": 0.01549347210675478,
+ "learning_rate": 0.0006,
+ "loss": 4.534621238708496,
+ "step": 1421
+ },
+ {
+ "epoch": 19.754585152838427,
+ "grad_norm": 0.01654665172100067,
+ "learning_rate": 0.0006,
+ "loss": 4.4724225997924805,
+ "step": 1422
+ },
+ {
+ "epoch": 19.768558951965066,
+ "grad_norm": 0.017119025811553,
+ "learning_rate": 0.0006,
+ "loss": 4.3897247314453125,
+ "step": 1423
+ },
+ {
+ "epoch": 19.782532751091704,
+ "grad_norm": 0.01603223942220211,
+ "learning_rate": 0.0006,
+ "loss": 4.412731170654297,
+ "step": 1424
+ },
+ {
+ "epoch": 19.79650655021834,
+ "grad_norm": 0.016774067655205727,
+ "learning_rate": 0.0006,
+ "loss": 4.54707145690918,
+ "step": 1425
+ },
+ {
+ "epoch": 19.810480349344978,
+ "grad_norm": 0.016453025862574577,
+ "learning_rate": 0.0006,
+ "loss": 4.462501049041748,
+ "step": 1426
+ },
+ {
+ "epoch": 19.824454148471617,
+ "grad_norm": 0.01597629487514496,
+ "learning_rate": 0.0006,
+ "loss": 4.229758262634277,
+ "step": 1427
+ },
+ {
+ "epoch": 19.83842794759825,
+ "grad_norm": 0.017999034374952316,
+ "learning_rate": 0.0006,
+ "loss": 4.389263153076172,
+ "step": 1428
+ },
+ {
+ "epoch": 19.85240174672489,
+ "grad_norm": 0.02165898121893406,
+ "learning_rate": 0.0006,
+ "loss": 4.342130661010742,
+ "step": 1429
+ },
+ {
+ "epoch": 19.86637554585153,
+ "grad_norm": 0.02222822792828083,
+ "learning_rate": 0.0006,
+ "loss": 4.492661476135254,
+ "step": 1430
+ },
+ {
+ "epoch": 19.880349344978168,
+ "grad_norm": 0.019378066062927246,
+ "learning_rate": 0.0006,
+ "loss": 4.4614105224609375,
+ "step": 1431
+ },
+ {
+ "epoch": 19.894323144104803,
+ "grad_norm": 0.019908083602786064,
+ "learning_rate": 0.0006,
+ "loss": 4.370929718017578,
+ "step": 1432
+ },
+ {
+ "epoch": 19.90829694323144,
+ "grad_norm": 0.022296471521258354,
+ "learning_rate": 0.0006,
+ "loss": 4.438320159912109,
+ "step": 1433
+ },
+ {
+ "epoch": 19.92227074235808,
+ "grad_norm": 0.021810343489050865,
+ "learning_rate": 0.0006,
+ "loss": 4.462067127227783,
+ "step": 1434
+ },
+ {
+ "epoch": 19.936244541484715,
+ "grad_norm": 0.02129209227859974,
+ "learning_rate": 0.0006,
+ "loss": 4.541261196136475,
+ "step": 1435
+ },
+ {
+ "epoch": 19.950218340611354,
+ "grad_norm": 0.02079629711806774,
+ "learning_rate": 0.0006,
+ "loss": 4.36712646484375,
+ "step": 1436
+ },
+ {
+ "epoch": 19.964192139737992,
+ "grad_norm": 0.022501740604639053,
+ "learning_rate": 0.0006,
+ "loss": 4.5279083251953125,
+ "step": 1437
+ },
+ {
+ "epoch": 19.978165938864628,
+ "grad_norm": 0.019121676683425903,
+ "learning_rate": 0.0006,
+ "loss": 4.477789878845215,
+ "step": 1438
+ },
+ {
+ "epoch": 19.992139737991266,
+ "grad_norm": 0.017429588362574577,
+ "learning_rate": 0.0006,
+ "loss": 4.4023847579956055,
+ "step": 1439
+ },
+ {
+ "epoch": 20.0,
+ "grad_norm": 0.018857665359973907,
+ "learning_rate": 0.0006,
+ "loss": 4.494743347167969,
+ "step": 1440
+ },
+ {
+ "epoch": 20.0,
+ "eval_loss": 4.694555759429932,
+ "eval_runtime": 56.4146,
+ "eval_samples_per_second": 43.287,
+ "eval_steps_per_second": 1.365,
+ "step": 1440
+ },
+ {
+ "epoch": 20.01397379912664,
+ "grad_norm": 0.016505861654877663,
+ "learning_rate": 0.0006,
+ "loss": 4.506229877471924,
+ "step": 1441
+ },
+ {
+ "epoch": 20.027947598253274,
+ "grad_norm": 0.017571842297911644,
+ "learning_rate": 0.0006,
+ "loss": 4.384073257446289,
+ "step": 1442
+ },
+ {
+ "epoch": 20.041921397379912,
+ "grad_norm": 0.019398365169763565,
+ "learning_rate": 0.0006,
+ "loss": 4.482177734375,
+ "step": 1443
+ },
+ {
+ "epoch": 20.05589519650655,
+ "grad_norm": 0.02011062018573284,
+ "learning_rate": 0.0006,
+ "loss": 4.440773963928223,
+ "step": 1444
+ },
+ {
+ "epoch": 20.069868995633186,
+ "grad_norm": 0.019026953727006912,
+ "learning_rate": 0.0006,
+ "loss": 4.507168769836426,
+ "step": 1445
+ },
+ {
+ "epoch": 20.083842794759825,
+ "grad_norm": 0.020506031811237335,
+ "learning_rate": 0.0006,
+ "loss": 4.469079971313477,
+ "step": 1446
+ },
+ {
+ "epoch": 20.097816593886463,
+ "grad_norm": 0.019608162343502045,
+ "learning_rate": 0.0006,
+ "loss": 4.395927429199219,
+ "step": 1447
+ },
+ {
+ "epoch": 20.111790393013102,
+ "grad_norm": 0.019419124349951744,
+ "learning_rate": 0.0006,
+ "loss": 4.484235763549805,
+ "step": 1448
+ },
+ {
+ "epoch": 20.125764192139737,
+ "grad_norm": 0.021775128319859505,
+ "learning_rate": 0.0006,
+ "loss": 4.333173751831055,
+ "step": 1449
+ },
+ {
+ "epoch": 20.139737991266376,
+ "grad_norm": 0.025512272492051125,
+ "learning_rate": 0.0006,
+ "loss": 4.460362434387207,
+ "step": 1450
+ },
+ {
+ "epoch": 20.153711790393015,
+ "grad_norm": 0.025157401338219643,
+ "learning_rate": 0.0006,
+ "loss": 4.4361395835876465,
+ "step": 1451
+ },
+ {
+ "epoch": 20.16768558951965,
+ "grad_norm": 0.019862616434693336,
+ "learning_rate": 0.0006,
+ "loss": 4.426807403564453,
+ "step": 1452
+ },
+ {
+ "epoch": 20.18165938864629,
+ "grad_norm": 0.01932152360677719,
+ "learning_rate": 0.0006,
+ "loss": 4.375677108764648,
+ "step": 1453
+ },
+ {
+ "epoch": 20.195633187772927,
+ "grad_norm": 0.01760999858379364,
+ "learning_rate": 0.0006,
+ "loss": 4.42259407043457,
+ "step": 1454
+ },
+ {
+ "epoch": 20.209606986899562,
+ "grad_norm": 0.016998112201690674,
+ "learning_rate": 0.0006,
+ "loss": 4.481280326843262,
+ "step": 1455
+ },
+ {
+ "epoch": 20.2235807860262,
+ "grad_norm": 0.017199190333485603,
+ "learning_rate": 0.0006,
+ "loss": 4.496712684631348,
+ "step": 1456
+ },
+ {
+ "epoch": 20.23755458515284,
+ "grad_norm": 0.016833679750561714,
+ "learning_rate": 0.0006,
+ "loss": 4.40736722946167,
+ "step": 1457
+ },
+ {
+ "epoch": 20.251528384279474,
+ "grad_norm": 0.017546730116009712,
+ "learning_rate": 0.0006,
+ "loss": 4.4915266036987305,
+ "step": 1458
+ },
+ {
+ "epoch": 20.265502183406113,
+ "grad_norm": 0.016595078632235527,
+ "learning_rate": 0.0006,
+ "loss": 4.460815906524658,
+ "step": 1459
+ },
+ {
+ "epoch": 20.27947598253275,
+ "grad_norm": 0.017728568986058235,
+ "learning_rate": 0.0006,
+ "loss": 4.346048831939697,
+ "step": 1460
+ },
+ {
+ "epoch": 20.29344978165939,
+ "grad_norm": 0.01849197782576084,
+ "learning_rate": 0.0006,
+ "loss": 4.399681091308594,
+ "step": 1461
+ },
+ {
+ "epoch": 20.307423580786025,
+ "grad_norm": 0.018336588516831398,
+ "learning_rate": 0.0006,
+ "loss": 4.414183616638184,
+ "step": 1462
+ },
+ {
+ "epoch": 20.321397379912664,
+ "grad_norm": 0.018086519092321396,
+ "learning_rate": 0.0006,
+ "loss": 4.346586227416992,
+ "step": 1463
+ },
+ {
+ "epoch": 20.335371179039303,
+ "grad_norm": 0.019697299227118492,
+ "learning_rate": 0.0006,
+ "loss": 4.4420952796936035,
+ "step": 1464
+ },
+ {
+ "epoch": 20.349344978165938,
+ "grad_norm": 0.02198074571788311,
+ "learning_rate": 0.0006,
+ "loss": 4.588729381561279,
+ "step": 1465
+ },
+ {
+ "epoch": 20.363318777292577,
+ "grad_norm": 0.022813329473137856,
+ "learning_rate": 0.0006,
+ "loss": 4.49729585647583,
+ "step": 1466
+ },
+ {
+ "epoch": 20.377292576419215,
+ "grad_norm": 0.022006649523973465,
+ "learning_rate": 0.0006,
+ "loss": 4.475641250610352,
+ "step": 1467
+ },
+ {
+ "epoch": 20.39126637554585,
+ "grad_norm": 0.023188438266515732,
+ "learning_rate": 0.0006,
+ "loss": 4.404322624206543,
+ "step": 1468
+ },
+ {
+ "epoch": 20.40524017467249,
+ "grad_norm": 0.022226877510547638,
+ "learning_rate": 0.0006,
+ "loss": 4.50904655456543,
+ "step": 1469
+ },
+ {
+ "epoch": 20.419213973799128,
+ "grad_norm": 0.021082637831568718,
+ "learning_rate": 0.0006,
+ "loss": 4.339011192321777,
+ "step": 1470
+ },
+ {
+ "epoch": 20.433187772925763,
+ "grad_norm": 0.02013542130589485,
+ "learning_rate": 0.0006,
+ "loss": 4.516942024230957,
+ "step": 1471
+ },
+ {
+ "epoch": 20.4471615720524,
+ "grad_norm": 0.01898968778550625,
+ "learning_rate": 0.0006,
+ "loss": 4.315804481506348,
+ "step": 1472
+ },
+ {
+ "epoch": 20.46113537117904,
+ "grad_norm": 0.019450880587100983,
+ "learning_rate": 0.0006,
+ "loss": 4.4794921875,
+ "step": 1473
+ },
+ {
+ "epoch": 20.475109170305675,
+ "grad_norm": 0.018577802926301956,
+ "learning_rate": 0.0006,
+ "loss": 4.422323703765869,
+ "step": 1474
+ },
+ {
+ "epoch": 20.489082969432314,
+ "grad_norm": 0.01778412237763405,
+ "learning_rate": 0.0006,
+ "loss": 4.429789066314697,
+ "step": 1475
+ },
+ {
+ "epoch": 20.503056768558952,
+ "grad_norm": 0.017134476453065872,
+ "learning_rate": 0.0006,
+ "loss": 4.416459083557129,
+ "step": 1476
+ },
+ {
+ "epoch": 20.51703056768559,
+ "grad_norm": 0.017153557389974594,
+ "learning_rate": 0.0006,
+ "loss": 4.446170806884766,
+ "step": 1477
+ },
+ {
+ "epoch": 20.531004366812226,
+ "grad_norm": 0.014998532831668854,
+ "learning_rate": 0.0006,
+ "loss": 4.419447898864746,
+ "step": 1478
+ },
+ {
+ "epoch": 20.544978165938865,
+ "grad_norm": 0.015877850353717804,
+ "learning_rate": 0.0006,
+ "loss": 4.312338829040527,
+ "step": 1479
+ },
+ {
+ "epoch": 20.558951965065503,
+ "grad_norm": 0.01569489948451519,
+ "learning_rate": 0.0006,
+ "loss": 4.572905540466309,
+ "step": 1480
+ },
+ {
+ "epoch": 20.57292576419214,
+ "grad_norm": 0.01591474376618862,
+ "learning_rate": 0.0006,
+ "loss": 4.414907455444336,
+ "step": 1481
+ },
+ {
+ "epoch": 20.586899563318777,
+ "grad_norm": 0.015395903028547764,
+ "learning_rate": 0.0006,
+ "loss": 4.549131393432617,
+ "step": 1482
+ },
+ {
+ "epoch": 20.600873362445416,
+ "grad_norm": 0.015715358778834343,
+ "learning_rate": 0.0006,
+ "loss": 4.530194282531738,
+ "step": 1483
+ },
+ {
+ "epoch": 20.61484716157205,
+ "grad_norm": 0.016042975708842278,
+ "learning_rate": 0.0006,
+ "loss": 4.322149276733398,
+ "step": 1484
+ },
+ {
+ "epoch": 20.62882096069869,
+ "grad_norm": 0.01664186641573906,
+ "learning_rate": 0.0006,
+ "loss": 4.424866676330566,
+ "step": 1485
+ },
+ {
+ "epoch": 20.64279475982533,
+ "grad_norm": 0.01631534844636917,
+ "learning_rate": 0.0006,
+ "loss": 4.441752910614014,
+ "step": 1486
+ },
+ {
+ "epoch": 20.656768558951963,
+ "grad_norm": 0.014057058840990067,
+ "learning_rate": 0.0006,
+ "loss": 4.399982452392578,
+ "step": 1487
+ },
+ {
+ "epoch": 20.670742358078602,
+ "grad_norm": 0.015076273120939732,
+ "learning_rate": 0.0006,
+ "loss": 4.441324234008789,
+ "step": 1488
+ },
+ {
+ "epoch": 20.68471615720524,
+ "grad_norm": 0.014138277620077133,
+ "learning_rate": 0.0006,
+ "loss": 4.4312543869018555,
+ "step": 1489
+ },
+ {
+ "epoch": 20.69868995633188,
+ "grad_norm": 0.013762440532445908,
+ "learning_rate": 0.0006,
+ "loss": 4.387807846069336,
+ "step": 1490
+ },
+ {
+ "epoch": 20.712663755458514,
+ "grad_norm": 0.015058411285281181,
+ "learning_rate": 0.0006,
+ "loss": 4.594747543334961,
+ "step": 1491
+ },
+ {
+ "epoch": 20.726637554585153,
+ "grad_norm": 0.013827123679220676,
+ "learning_rate": 0.0006,
+ "loss": 4.434604167938232,
+ "step": 1492
+ },
+ {
+ "epoch": 20.74061135371179,
+ "grad_norm": 0.015315295197069645,
+ "learning_rate": 0.0006,
+ "loss": 4.588066101074219,
+ "step": 1493
+ },
+ {
+ "epoch": 20.754585152838427,
+ "grad_norm": 0.018678732216358185,
+ "learning_rate": 0.0006,
+ "loss": 4.3514204025268555,
+ "step": 1494
+ },
+ {
+ "epoch": 20.768558951965066,
+ "grad_norm": 0.019521350041031837,
+ "learning_rate": 0.0006,
+ "loss": 4.437658786773682,
+ "step": 1495
+ },
+ {
+ "epoch": 20.782532751091704,
+ "grad_norm": 0.018363403156399727,
+ "learning_rate": 0.0006,
+ "loss": 4.495002269744873,
+ "step": 1496
+ },
+ {
+ "epoch": 20.79650655021834,
+ "grad_norm": 0.016830725595355034,
+ "learning_rate": 0.0006,
+ "loss": 4.402471542358398,
+ "step": 1497
+ },
+ {
+ "epoch": 20.810480349344978,
+ "grad_norm": 0.01722072623670101,
+ "learning_rate": 0.0006,
+ "loss": 4.437653541564941,
+ "step": 1498
+ },
+ {
+ "epoch": 20.824454148471617,
+ "grad_norm": 0.017459599301218987,
+ "learning_rate": 0.0006,
+ "loss": 4.366603851318359,
+ "step": 1499
+ },
+ {
+ "epoch": 20.83842794759825,
+ "grad_norm": 0.018902219831943512,
+ "learning_rate": 0.0006,
+ "loss": 4.501561164855957,
+ "step": 1500
+ },
+ {
+ "epoch": 20.85240174672489,
+ "grad_norm": 0.020814096555113792,
+ "learning_rate": 0.0006,
+ "loss": 4.535295486450195,
+ "step": 1501
+ },
+ {
+ "epoch": 20.86637554585153,
+ "grad_norm": 0.018990976735949516,
+ "learning_rate": 0.0006,
+ "loss": 4.333580017089844,
+ "step": 1502
+ },
+ {
+ "epoch": 20.880349344978168,
+ "grad_norm": 0.01837880350649357,
+ "learning_rate": 0.0006,
+ "loss": 4.313388824462891,
+ "step": 1503
+ },
+ {
+ "epoch": 20.894323144104803,
+ "grad_norm": 0.019820360466837883,
+ "learning_rate": 0.0006,
+ "loss": 4.388157844543457,
+ "step": 1504
+ },
+ {
+ "epoch": 20.90829694323144,
+ "grad_norm": 0.020181143656373024,
+ "learning_rate": 0.0006,
+ "loss": 4.572224140167236,
+ "step": 1505
+ },
+ {
+ "epoch": 20.92227074235808,
+ "grad_norm": 0.021206077188253403,
+ "learning_rate": 0.0006,
+ "loss": 4.3127970695495605,
+ "step": 1506
+ },
+ {
+ "epoch": 20.936244541484715,
+ "grad_norm": 0.017962060868740082,
+ "learning_rate": 0.0006,
+ "loss": 4.384734630584717,
+ "step": 1507
+ },
+ {
+ "epoch": 20.950218340611354,
+ "grad_norm": 0.01737220585346222,
+ "learning_rate": 0.0006,
+ "loss": 4.4075751304626465,
+ "step": 1508
+ },
+ {
+ "epoch": 20.964192139737992,
+ "grad_norm": 0.01839268207550049,
+ "learning_rate": 0.0006,
+ "loss": 4.553779125213623,
+ "step": 1509
+ },
+ {
+ "epoch": 20.978165938864628,
+ "grad_norm": 0.020102640613913536,
+ "learning_rate": 0.0006,
+ "loss": 4.399398326873779,
+ "step": 1510
+ },
+ {
+ "epoch": 20.992139737991266,
+ "grad_norm": 0.018831463530659676,
+ "learning_rate": 0.0006,
+ "loss": 4.358499526977539,
+ "step": 1511
+ },
+ {
+ "epoch": 21.0,
+ "grad_norm": 0.018254734575748444,
+ "learning_rate": 0.0006,
+ "loss": 4.384119510650635,
+ "step": 1512
+ },
+ {
+ "epoch": 21.0,
+ "eval_loss": 4.7524237632751465,
+ "eval_runtime": 57.8812,
+ "eval_samples_per_second": 42.19,
+ "eval_steps_per_second": 1.33,
+ "step": 1512
+ },
+ {
+ "epoch": 21.01397379912664,
+ "grad_norm": 0.019728587940335274,
+ "learning_rate": 0.0006,
+ "loss": 4.32363224029541,
+ "step": 1513
+ },
+ {
+ "epoch": 21.027947598253274,
+ "grad_norm": 0.02216625027358532,
+ "learning_rate": 0.0006,
+ "loss": 4.436988353729248,
+ "step": 1514
+ },
+ {
+ "epoch": 21.041921397379912,
+ "grad_norm": 0.02227606810629368,
+ "learning_rate": 0.0006,
+ "loss": 4.406063556671143,
+ "step": 1515
+ },
+ {
+ "epoch": 21.05589519650655,
+ "grad_norm": 0.020187009125947952,
+ "learning_rate": 0.0006,
+ "loss": 4.324402332305908,
+ "step": 1516
+ },
+ {
+ "epoch": 21.069868995633186,
+ "grad_norm": 0.02095157466828823,
+ "learning_rate": 0.0006,
+ "loss": 4.456544876098633,
+ "step": 1517
+ },
+ {
+ "epoch": 21.083842794759825,
+ "grad_norm": 0.02146013453602791,
+ "learning_rate": 0.0006,
+ "loss": 4.368387222290039,
+ "step": 1518
+ },
+ {
+ "epoch": 21.097816593886463,
+ "grad_norm": 0.02162035182118416,
+ "learning_rate": 0.0006,
+ "loss": 4.48838996887207,
+ "step": 1519
+ },
+ {
+ "epoch": 21.111790393013102,
+ "grad_norm": 0.021745676174759865,
+ "learning_rate": 0.0006,
+ "loss": 4.30778694152832,
+ "step": 1520
+ },
+ {
+ "epoch": 21.125764192139737,
+ "grad_norm": 0.02173588052392006,
+ "learning_rate": 0.0006,
+ "loss": 4.433149814605713,
+ "step": 1521
+ },
+ {
+ "epoch": 21.139737991266376,
+ "grad_norm": 0.020950648933649063,
+ "learning_rate": 0.0006,
+ "loss": 4.520089149475098,
+ "step": 1522
+ },
+ {
+ "epoch": 21.153711790393015,
+ "grad_norm": 0.020684752613306046,
+ "learning_rate": 0.0006,
+ "loss": 4.372437477111816,
+ "step": 1523
+ },
+ {
+ "epoch": 21.16768558951965,
+ "grad_norm": 0.019472869113087654,
+ "learning_rate": 0.0006,
+ "loss": 4.437015533447266,
+ "step": 1524
+ },
+ {
+ "epoch": 21.18165938864629,
+ "grad_norm": 0.01846962980926037,
+ "learning_rate": 0.0006,
+ "loss": 4.449531555175781,
+ "step": 1525
+ },
+ {
+ "epoch": 21.195633187772927,
+ "grad_norm": 0.017377803102135658,
+ "learning_rate": 0.0006,
+ "loss": 4.458279132843018,
+ "step": 1526
+ },
+ {
+ "epoch": 21.209606986899562,
+ "grad_norm": 0.016002200543880463,
+ "learning_rate": 0.0006,
+ "loss": 4.266297817230225,
+ "step": 1527
+ },
+ {
+ "epoch": 21.2235807860262,
+ "grad_norm": 0.015575903467833996,
+ "learning_rate": 0.0006,
+ "loss": 4.342632293701172,
+ "step": 1528
+ },
+ {
+ "epoch": 21.23755458515284,
+ "grad_norm": 0.015634698793292046,
+ "learning_rate": 0.0006,
+ "loss": 4.338623523712158,
+ "step": 1529
+ },
+ {
+ "epoch": 21.251528384279474,
+ "grad_norm": 0.01449581328779459,
+ "learning_rate": 0.0006,
+ "loss": 4.354330062866211,
+ "step": 1530
+ },
+ {
+ "epoch": 21.265502183406113,
+ "grad_norm": 0.01487616915255785,
+ "learning_rate": 0.0006,
+ "loss": 4.486835479736328,
+ "step": 1531
+ },
+ {
+ "epoch": 21.27947598253275,
+ "grad_norm": 0.016346555203199387,
+ "learning_rate": 0.0006,
+ "loss": 4.407098770141602,
+ "step": 1532
+ },
+ {
+ "epoch": 21.29344978165939,
+ "grad_norm": 0.01682182401418686,
+ "learning_rate": 0.0006,
+ "loss": 4.436809539794922,
+ "step": 1533
+ },
+ {
+ "epoch": 21.307423580786025,
+ "grad_norm": 0.017704004421830177,
+ "learning_rate": 0.0006,
+ "loss": 4.3854780197143555,
+ "step": 1534
+ },
+ {
+ "epoch": 21.321397379912664,
+ "grad_norm": 0.018437420949339867,
+ "learning_rate": 0.0006,
+ "loss": 4.413283824920654,
+ "step": 1535
+ },
+ {
+ "epoch": 21.335371179039303,
+ "grad_norm": 0.016911081969738007,
+ "learning_rate": 0.0006,
+ "loss": 4.357306003570557,
+ "step": 1536
+ },
+ {
+ "epoch": 21.349344978165938,
+ "grad_norm": 0.01642223633825779,
+ "learning_rate": 0.0006,
+ "loss": 4.422322750091553,
+ "step": 1537
+ },
+ {
+ "epoch": 21.363318777292577,
+ "grad_norm": 0.01780877821147442,
+ "learning_rate": 0.0006,
+ "loss": 4.387620449066162,
+ "step": 1538
+ },
+ {
+ "epoch": 21.377292576419215,
+ "grad_norm": 0.01653975248336792,
+ "learning_rate": 0.0006,
+ "loss": 4.327267646789551,
+ "step": 1539
+ },
+ {
+ "epoch": 21.39126637554585,
+ "grad_norm": 0.01572156324982643,
+ "learning_rate": 0.0006,
+ "loss": 4.475979804992676,
+ "step": 1540
+ },
+ {
+ "epoch": 21.40524017467249,
+ "grad_norm": 0.017156066372990608,
+ "learning_rate": 0.0006,
+ "loss": 4.4011125564575195,
+ "step": 1541
+ },
+ {
+ "epoch": 21.419213973799128,
+ "grad_norm": 0.01788829080760479,
+ "learning_rate": 0.0006,
+ "loss": 4.4174089431762695,
+ "step": 1542
+ },
+ {
+ "epoch": 21.433187772925763,
+ "grad_norm": 0.01728985086083412,
+ "learning_rate": 0.0006,
+ "loss": 4.429244041442871,
+ "step": 1543
+ },
+ {
+ "epoch": 21.4471615720524,
+ "grad_norm": 0.016534466296434402,
+ "learning_rate": 0.0006,
+ "loss": 4.326068878173828,
+ "step": 1544
+ },
+ {
+ "epoch": 21.46113537117904,
+ "grad_norm": 0.016582584008574486,
+ "learning_rate": 0.0006,
+ "loss": 4.2975969314575195,
+ "step": 1545
+ },
+ {
+ "epoch": 21.475109170305675,
+ "grad_norm": 0.018034178763628006,
+ "learning_rate": 0.0006,
+ "loss": 4.501319408416748,
+ "step": 1546
+ },
+ {
+ "epoch": 21.489082969432314,
+ "grad_norm": 0.017827702686190605,
+ "learning_rate": 0.0006,
+ "loss": 4.502284526824951,
+ "step": 1547
+ },
+ {
+ "epoch": 21.503056768558952,
+ "grad_norm": 0.017182299867272377,
+ "learning_rate": 0.0006,
+ "loss": 4.380828857421875,
+ "step": 1548
+ },
+ {
+ "epoch": 21.51703056768559,
+ "grad_norm": 0.019101588055491447,
+ "learning_rate": 0.0006,
+ "loss": 4.31564998626709,
+ "step": 1549
+ },
+ {
+ "epoch": 21.531004366812226,
+ "grad_norm": 0.0204677302390337,
+ "learning_rate": 0.0006,
+ "loss": 4.3840131759643555,
+ "step": 1550
+ },
+ {
+ "epoch": 21.544978165938865,
+ "grad_norm": 0.0188993439078331,
+ "learning_rate": 0.0006,
+ "loss": 4.4257731437683105,
+ "step": 1551
+ },
+ {
+ "epoch": 21.558951965065503,
+ "grad_norm": 0.01730230636894703,
+ "learning_rate": 0.0006,
+ "loss": 4.445132255554199,
+ "step": 1552
+ },
+ {
+ "epoch": 21.57292576419214,
+ "grad_norm": 0.01801574043929577,
+ "learning_rate": 0.0006,
+ "loss": 4.427064895629883,
+ "step": 1553
+ },
+ {
+ "epoch": 21.586899563318777,
+ "grad_norm": 0.018417565152049065,
+ "learning_rate": 0.0006,
+ "loss": 4.319595813751221,
+ "step": 1554
+ },
+ {
+ "epoch": 21.600873362445416,
+ "grad_norm": 0.01683160290122032,
+ "learning_rate": 0.0006,
+ "loss": 4.349218368530273,
+ "step": 1555
+ },
+ {
+ "epoch": 21.61484716157205,
+ "grad_norm": 0.016973106190562248,
+ "learning_rate": 0.0006,
+ "loss": 4.375990390777588,
+ "step": 1556
+ },
+ {
+ "epoch": 21.62882096069869,
+ "grad_norm": 0.015688583254814148,
+ "learning_rate": 0.0006,
+ "loss": 4.408170223236084,
+ "step": 1557
+ },
+ {
+ "epoch": 21.64279475982533,
+ "grad_norm": 0.01573043316602707,
+ "learning_rate": 0.0006,
+ "loss": 4.416255950927734,
+ "step": 1558
+ },
+ {
+ "epoch": 21.656768558951963,
+ "grad_norm": 0.01744293048977852,
+ "learning_rate": 0.0006,
+ "loss": 4.477141380310059,
+ "step": 1559
+ },
+ {
+ "epoch": 21.670742358078602,
+ "grad_norm": 0.017503513023257256,
+ "learning_rate": 0.0006,
+ "loss": 4.3367486000061035,
+ "step": 1560
+ },
+ {
+ "epoch": 21.68471615720524,
+ "grad_norm": 0.016592150554060936,
+ "learning_rate": 0.0006,
+ "loss": 4.314785480499268,
+ "step": 1561
+ },
+ {
+ "epoch": 21.69868995633188,
+ "grad_norm": 0.015774084255099297,
+ "learning_rate": 0.0006,
+ "loss": 4.448209285736084,
+ "step": 1562
+ },
+ {
+ "epoch": 21.712663755458514,
+ "grad_norm": 0.015779150649905205,
+ "learning_rate": 0.0006,
+ "loss": 4.392690658569336,
+ "step": 1563
+ },
+ {
+ "epoch": 21.726637554585153,
+ "grad_norm": 0.015392648056149483,
+ "learning_rate": 0.0006,
+ "loss": 4.437741756439209,
+ "step": 1564
+ },
+ {
+ "epoch": 21.74061135371179,
+ "grad_norm": 0.01530836895108223,
+ "learning_rate": 0.0006,
+ "loss": 4.424595832824707,
+ "step": 1565
+ },
+ {
+ "epoch": 21.754585152838427,
+ "grad_norm": 0.015600779093801975,
+ "learning_rate": 0.0006,
+ "loss": 4.335417747497559,
+ "step": 1566
+ },
+ {
+ "epoch": 21.768558951965066,
+ "grad_norm": 0.014646518044173717,
+ "learning_rate": 0.0006,
+ "loss": 4.279081344604492,
+ "step": 1567
+ },
+ {
+ "epoch": 21.782532751091704,
+ "grad_norm": 0.014132864773273468,
+ "learning_rate": 0.0006,
+ "loss": 4.345950126647949,
+ "step": 1568
+ },
+ {
+ "epoch": 21.79650655021834,
+ "grad_norm": 0.01425724383443594,
+ "learning_rate": 0.0006,
+ "loss": 4.277588844299316,
+ "step": 1569
+ },
+ {
+ "epoch": 21.810480349344978,
+ "grad_norm": 0.015396572649478912,
+ "learning_rate": 0.0006,
+ "loss": 4.326196670532227,
+ "step": 1570
+ },
+ {
+ "epoch": 21.824454148471617,
+ "grad_norm": 0.01683669537305832,
+ "learning_rate": 0.0006,
+ "loss": 4.3294172286987305,
+ "step": 1571
+ },
+ {
+ "epoch": 21.83842794759825,
+ "grad_norm": 0.016365647315979004,
+ "learning_rate": 0.0006,
+ "loss": 4.523321628570557,
+ "step": 1572
+ },
+ {
+ "epoch": 21.85240174672489,
+ "grad_norm": 0.017142774537205696,
+ "learning_rate": 0.0006,
+ "loss": 4.475842475891113,
+ "step": 1573
+ },
+ {
+ "epoch": 21.86637554585153,
+ "grad_norm": 0.015950839966535568,
+ "learning_rate": 0.0006,
+ "loss": 4.397187232971191,
+ "step": 1574
+ },
+ {
+ "epoch": 21.880349344978168,
+ "grad_norm": 0.016180362552404404,
+ "learning_rate": 0.0006,
+ "loss": 4.414778709411621,
+ "step": 1575
+ },
+ {
+ "epoch": 21.894323144104803,
+ "grad_norm": 0.017615556716918945,
+ "learning_rate": 0.0006,
+ "loss": 4.348567962646484,
+ "step": 1576
+ },
+ {
+ "epoch": 21.90829694323144,
+ "grad_norm": 0.01896277628839016,
+ "learning_rate": 0.0006,
+ "loss": 4.431762218475342,
+ "step": 1577
+ },
+ {
+ "epoch": 21.92227074235808,
+ "grad_norm": 0.01990971714258194,
+ "learning_rate": 0.0006,
+ "loss": 4.431285858154297,
+ "step": 1578
+ },
+ {
+ "epoch": 21.936244541484715,
+ "grad_norm": 0.0201258547604084,
+ "learning_rate": 0.0006,
+ "loss": 4.309872627258301,
+ "step": 1579
+ },
+ {
+ "epoch": 21.950218340611354,
+ "grad_norm": 0.019620198756456375,
+ "learning_rate": 0.0006,
+ "loss": 4.484044551849365,
+ "step": 1580
+ },
+ {
+ "epoch": 21.964192139737992,
+ "grad_norm": 0.01771123707294464,
+ "learning_rate": 0.0006,
+ "loss": 4.368093013763428,
+ "step": 1581
+ },
+ {
+ "epoch": 21.978165938864628,
+ "grad_norm": 0.017501654103398323,
+ "learning_rate": 0.0006,
+ "loss": 4.372651100158691,
+ "step": 1582
+ },
+ {
+ "epoch": 21.992139737991266,
+ "grad_norm": 0.01693608984351158,
+ "learning_rate": 0.0006,
+ "loss": 4.379167556762695,
+ "step": 1583
+ },
+ {
+ "epoch": 22.0,
+ "grad_norm": 0.02002849616110325,
+ "learning_rate": 0.0006,
+ "loss": 4.329627990722656,
+ "step": 1584
+ },
+ {
+ "epoch": 22.0,
+ "eval_loss": 4.723696231842041,
+ "eval_runtime": 56.1758,
+ "eval_samples_per_second": 43.471,
+ "eval_steps_per_second": 1.371,
+ "step": 1584
+ },
+ {
+ "epoch": 22.01397379912664,
+ "grad_norm": 0.02030305378139019,
+ "learning_rate": 0.0006,
+ "loss": 4.355566024780273,
+ "step": 1585
+ },
+ {
+ "epoch": 22.027947598253274,
+ "grad_norm": 0.018638934940099716,
+ "learning_rate": 0.0006,
+ "loss": 4.314009189605713,
+ "step": 1586
+ },
+ {
+ "epoch": 22.041921397379912,
+ "grad_norm": 0.019917072728276253,
+ "learning_rate": 0.0006,
+ "loss": 4.367525577545166,
+ "step": 1587
+ },
+ {
+ "epoch": 22.05589519650655,
+ "grad_norm": 0.020785167813301086,
+ "learning_rate": 0.0006,
+ "loss": 4.375561714172363,
+ "step": 1588
+ },
+ {
+ "epoch": 22.069868995633186,
+ "grad_norm": 0.021214453503489494,
+ "learning_rate": 0.0006,
+ "loss": 4.418305397033691,
+ "step": 1589
+ },
+ {
+ "epoch": 22.083842794759825,
+ "grad_norm": 0.019051678478717804,
+ "learning_rate": 0.0006,
+ "loss": 4.215126991271973,
+ "step": 1590
+ },
+ {
+ "epoch": 22.097816593886463,
+ "grad_norm": 0.02323366142809391,
+ "learning_rate": 0.0006,
+ "loss": 4.306807994842529,
+ "step": 1591
+ },
+ {
+ "epoch": 22.111790393013102,
+ "grad_norm": 0.026186738163232803,
+ "learning_rate": 0.0006,
+ "loss": 4.408508777618408,
+ "step": 1592
+ },
+ {
+ "epoch": 22.125764192139737,
+ "grad_norm": 0.027361227199435234,
+ "learning_rate": 0.0006,
+ "loss": 4.329322814941406,
+ "step": 1593
+ },
+ {
+ "epoch": 22.139737991266376,
+ "grad_norm": 0.026564646512269974,
+ "learning_rate": 0.0006,
+ "loss": 4.346642017364502,
+ "step": 1594
+ },
+ {
+ "epoch": 22.153711790393015,
+ "grad_norm": 0.025073667988181114,
+ "learning_rate": 0.0006,
+ "loss": 4.3984198570251465,
+ "step": 1595
+ },
+ {
+ "epoch": 22.16768558951965,
+ "grad_norm": 0.02347794733941555,
+ "learning_rate": 0.0006,
+ "loss": 4.352538108825684,
+ "step": 1596
+ },
+ {
+ "epoch": 22.18165938864629,
+ "grad_norm": 0.02203468047082424,
+ "learning_rate": 0.0006,
+ "loss": 4.298148155212402,
+ "step": 1597
+ },
+ {
+ "epoch": 22.195633187772927,
+ "grad_norm": 0.020621957257390022,
+ "learning_rate": 0.0006,
+ "loss": 4.480309963226318,
+ "step": 1598
+ },
+ {
+ "epoch": 22.209606986899562,
+ "grad_norm": 0.023710055276751518,
+ "learning_rate": 0.0006,
+ "loss": 4.339317321777344,
+ "step": 1599
+ },
+ {
+ "epoch": 22.2235807860262,
+ "grad_norm": 0.0233482476323843,
+ "learning_rate": 0.0006,
+ "loss": 4.436896324157715,
+ "step": 1600
+ },
+ {
+ "epoch": 22.23755458515284,
+ "grad_norm": 0.020122403278946877,
+ "learning_rate": 0.0006,
+ "loss": 4.236002445220947,
+ "step": 1601
+ },
+ {
+ "epoch": 22.251528384279474,
+ "grad_norm": 0.01957266591489315,
+ "learning_rate": 0.0006,
+ "loss": 4.497328281402588,
+ "step": 1602
+ },
+ {
+ "epoch": 22.265502183406113,
+ "grad_norm": 0.018964895978569984,
+ "learning_rate": 0.0006,
+ "loss": 4.36360502243042,
+ "step": 1603
+ },
+ {
+ "epoch": 22.27947598253275,
+ "grad_norm": 0.018902769312262535,
+ "learning_rate": 0.0006,
+ "loss": 4.427093505859375,
+ "step": 1604
+ },
+ {
+ "epoch": 22.29344978165939,
+ "grad_norm": 0.018918119370937347,
+ "learning_rate": 0.0006,
+ "loss": 4.294950485229492,
+ "step": 1605
+ },
+ {
+ "epoch": 22.307423580786025,
+ "grad_norm": 0.019315775483846664,
+ "learning_rate": 0.0006,
+ "loss": 4.427952289581299,
+ "step": 1606
+ },
+ {
+ "epoch": 22.321397379912664,
+ "grad_norm": 0.018350522965192795,
+ "learning_rate": 0.0006,
+ "loss": 4.272945404052734,
+ "step": 1607
+ },
+ {
+ "epoch": 22.335371179039303,
+ "grad_norm": 0.01759534701704979,
+ "learning_rate": 0.0006,
+ "loss": 4.348023891448975,
+ "step": 1608
+ },
+ {
+ "epoch": 22.349344978165938,
+ "grad_norm": 0.01646057888865471,
+ "learning_rate": 0.0006,
+ "loss": 4.278518199920654,
+ "step": 1609
+ },
+ {
+ "epoch": 22.363318777292577,
+ "grad_norm": 0.015875060111284256,
+ "learning_rate": 0.0006,
+ "loss": 4.423527240753174,
+ "step": 1610
+ },
+ {
+ "epoch": 22.377292576419215,
+ "grad_norm": 0.01638529635965824,
+ "learning_rate": 0.0006,
+ "loss": 4.321723937988281,
+ "step": 1611
+ },
+ {
+ "epoch": 22.39126637554585,
+ "grad_norm": 0.016942061483860016,
+ "learning_rate": 0.0006,
+ "loss": 4.346164703369141,
+ "step": 1612
+ },
+ {
+ "epoch": 22.40524017467249,
+ "grad_norm": 0.015209621749818325,
+ "learning_rate": 0.0006,
+ "loss": 4.434340000152588,
+ "step": 1613
+ },
+ {
+ "epoch": 22.419213973799128,
+ "grad_norm": 0.014423009008169174,
+ "learning_rate": 0.0006,
+ "loss": 4.342560768127441,
+ "step": 1614
+ },
+ {
+ "epoch": 22.433187772925763,
+ "grad_norm": 0.014369525946676731,
+ "learning_rate": 0.0006,
+ "loss": 4.296586990356445,
+ "step": 1615
+ },
+ {
+ "epoch": 22.4471615720524,
+ "grad_norm": 0.01392540242522955,
+ "learning_rate": 0.0006,
+ "loss": 4.433176040649414,
+ "step": 1616
+ },
+ {
+ "epoch": 22.46113537117904,
+ "grad_norm": 0.013206899166107178,
+ "learning_rate": 0.0006,
+ "loss": 4.32962703704834,
+ "step": 1617
+ },
+ {
+ "epoch": 22.475109170305675,
+ "grad_norm": 0.013427077792584896,
+ "learning_rate": 0.0006,
+ "loss": 4.420437335968018,
+ "step": 1618
+ },
+ {
+ "epoch": 22.489082969432314,
+ "grad_norm": 0.014194196090102196,
+ "learning_rate": 0.0006,
+ "loss": 4.410478591918945,
+ "step": 1619
+ },
+ {
+ "epoch": 22.503056768558952,
+ "grad_norm": 0.01362364087253809,
+ "learning_rate": 0.0006,
+ "loss": 4.338932037353516,
+ "step": 1620
+ },
+ {
+ "epoch": 22.51703056768559,
+ "grad_norm": 0.014299589209258556,
+ "learning_rate": 0.0006,
+ "loss": 4.34089994430542,
+ "step": 1621
+ },
+ {
+ "epoch": 22.531004366812226,
+ "grad_norm": 0.014061295427381992,
+ "learning_rate": 0.0006,
+ "loss": 4.29201602935791,
+ "step": 1622
+ },
+ {
+ "epoch": 22.544978165938865,
+ "grad_norm": 0.013945229351520538,
+ "learning_rate": 0.0006,
+ "loss": 4.326052188873291,
+ "step": 1623
+ },
+ {
+ "epoch": 22.558951965065503,
+ "grad_norm": 0.01387068536132574,
+ "learning_rate": 0.0006,
+ "loss": 4.393521308898926,
+ "step": 1624
+ },
+ {
+ "epoch": 22.57292576419214,
+ "grad_norm": 0.013931898400187492,
+ "learning_rate": 0.0006,
+ "loss": 4.3739752769470215,
+ "step": 1625
+ },
+ {
+ "epoch": 22.586899563318777,
+ "grad_norm": 0.01497740764170885,
+ "learning_rate": 0.0006,
+ "loss": 4.3991618156433105,
+ "step": 1626
+ },
+ {
+ "epoch": 22.600873362445416,
+ "grad_norm": 0.015842905268073082,
+ "learning_rate": 0.0006,
+ "loss": 4.405340671539307,
+ "step": 1627
+ },
+ {
+ "epoch": 22.61484716157205,
+ "grad_norm": 0.01580633595585823,
+ "learning_rate": 0.0006,
+ "loss": 4.280102729797363,
+ "step": 1628
+ },
+ {
+ "epoch": 22.62882096069869,
+ "grad_norm": 0.016567382961511612,
+ "learning_rate": 0.0006,
+ "loss": 4.407657146453857,
+ "step": 1629
+ },
+ {
+ "epoch": 22.64279475982533,
+ "grad_norm": 0.01648622378706932,
+ "learning_rate": 0.0006,
+ "loss": 4.392590522766113,
+ "step": 1630
+ },
+ {
+ "epoch": 22.656768558951963,
+ "grad_norm": 0.017066307365894318,
+ "learning_rate": 0.0006,
+ "loss": 4.324862003326416,
+ "step": 1631
+ },
+ {
+ "epoch": 22.670742358078602,
+ "grad_norm": 0.01637062057852745,
+ "learning_rate": 0.0006,
+ "loss": 4.34598445892334,
+ "step": 1632
+ },
+ {
+ "epoch": 22.68471615720524,
+ "grad_norm": 0.018038960173726082,
+ "learning_rate": 0.0006,
+ "loss": 4.320711612701416,
+ "step": 1633
+ },
+ {
+ "epoch": 22.69868995633188,
+ "grad_norm": 0.018096428364515305,
+ "learning_rate": 0.0006,
+ "loss": 4.536527156829834,
+ "step": 1634
+ },
+ {
+ "epoch": 22.712663755458514,
+ "grad_norm": 0.017889291048049927,
+ "learning_rate": 0.0006,
+ "loss": 4.388131141662598,
+ "step": 1635
+ },
+ {
+ "epoch": 22.726637554585153,
+ "grad_norm": 0.018771521747112274,
+ "learning_rate": 0.0006,
+ "loss": 4.357668876647949,
+ "step": 1636
+ },
+ {
+ "epoch": 22.74061135371179,
+ "grad_norm": 0.02100873365998268,
+ "learning_rate": 0.0006,
+ "loss": 4.293972015380859,
+ "step": 1637
+ },
+ {
+ "epoch": 22.754585152838427,
+ "grad_norm": 0.017715785652399063,
+ "learning_rate": 0.0006,
+ "loss": 4.349849224090576,
+ "step": 1638
+ },
+ {
+ "epoch": 22.768558951965066,
+ "grad_norm": 0.015546333976089954,
+ "learning_rate": 0.0006,
+ "loss": 4.311728000640869,
+ "step": 1639
+ },
+ {
+ "epoch": 22.782532751091704,
+ "grad_norm": 0.01605990156531334,
+ "learning_rate": 0.0006,
+ "loss": 4.499123573303223,
+ "step": 1640
+ },
+ {
+ "epoch": 22.79650655021834,
+ "grad_norm": 0.016642142087221146,
+ "learning_rate": 0.0006,
+ "loss": 4.4319257736206055,
+ "step": 1641
+ },
+ {
+ "epoch": 22.810480349344978,
+ "grad_norm": 0.016120387241244316,
+ "learning_rate": 0.0006,
+ "loss": 4.4373650550842285,
+ "step": 1642
+ },
+ {
+ "epoch": 22.824454148471617,
+ "grad_norm": 0.015174890868365765,
+ "learning_rate": 0.0006,
+ "loss": 4.314082145690918,
+ "step": 1643
+ },
+ {
+ "epoch": 22.83842794759825,
+ "grad_norm": 0.01551714539527893,
+ "learning_rate": 0.0006,
+ "loss": 4.360363960266113,
+ "step": 1644
+ },
+ {
+ "epoch": 22.85240174672489,
+ "grad_norm": 0.014966742135584354,
+ "learning_rate": 0.0006,
+ "loss": 4.346253395080566,
+ "step": 1645
+ },
+ {
+ "epoch": 22.86637554585153,
+ "grad_norm": 0.014288028702139854,
+ "learning_rate": 0.0006,
+ "loss": 4.464366436004639,
+ "step": 1646
+ },
+ {
+ "epoch": 22.880349344978168,
+ "grad_norm": 0.015291682444512844,
+ "learning_rate": 0.0006,
+ "loss": 4.504101753234863,
+ "step": 1647
+ },
+ {
+ "epoch": 22.894323144104803,
+ "grad_norm": 0.014534426853060722,
+ "learning_rate": 0.0006,
+ "loss": 4.462172031402588,
+ "step": 1648
+ },
+ {
+ "epoch": 22.90829694323144,
+ "grad_norm": 0.014129819348454475,
+ "learning_rate": 0.0006,
+ "loss": 4.350126266479492,
+ "step": 1649
+ },
+ {
+ "epoch": 22.92227074235808,
+ "grad_norm": 0.014932668767869473,
+ "learning_rate": 0.0006,
+ "loss": 4.374392032623291,
+ "step": 1650
+ },
+ {
+ "epoch": 22.936244541484715,
+ "grad_norm": 0.014647853560745716,
+ "learning_rate": 0.0006,
+ "loss": 4.252264499664307,
+ "step": 1651
+ },
+ {
+ "epoch": 22.950218340611354,
+ "grad_norm": 0.013553360477089882,
+ "learning_rate": 0.0006,
+ "loss": 4.379729270935059,
+ "step": 1652
+ },
+ {
+ "epoch": 22.964192139737992,
+ "grad_norm": 0.014606194570660591,
+ "learning_rate": 0.0006,
+ "loss": 4.363372802734375,
+ "step": 1653
+ },
+ {
+ "epoch": 22.978165938864628,
+ "grad_norm": 0.015000631101429462,
+ "learning_rate": 0.0006,
+ "loss": 4.408048629760742,
+ "step": 1654
+ },
+ {
+ "epoch": 22.992139737991266,
+ "grad_norm": 0.015370816923677921,
+ "learning_rate": 0.0006,
+ "loss": 4.297451019287109,
+ "step": 1655
+ },
+ {
+ "epoch": 23.0,
+ "grad_norm": 0.01709144562482834,
+ "learning_rate": 0.0006,
+ "loss": 4.294776916503906,
+ "step": 1656
+ },
+ {
+ "epoch": 23.0,
+ "eval_loss": 4.678618907928467,
+ "eval_runtime": 56.1668,
+ "eval_samples_per_second": 43.478,
+ "eval_steps_per_second": 1.371,
+ "step": 1656
+ },
+ {
+ "epoch": 23.01397379912664,
+ "grad_norm": 0.01727290451526642,
+ "learning_rate": 0.0006,
+ "loss": 4.311701774597168,
+ "step": 1657
+ },
+ {
+ "epoch": 23.027947598253274,
+ "grad_norm": 0.020645759999752045,
+ "learning_rate": 0.0006,
+ "loss": 4.429897308349609,
+ "step": 1658
+ },
+ {
+ "epoch": 23.041921397379912,
+ "grad_norm": 0.02204226516187191,
+ "learning_rate": 0.0006,
+ "loss": 4.365151882171631,
+ "step": 1659
+ },
+ {
+ "epoch": 23.05589519650655,
+ "grad_norm": 0.023069269955158234,
+ "learning_rate": 0.0006,
+ "loss": 4.3710832595825195,
+ "step": 1660
+ },
+ {
+ "epoch": 23.069868995633186,
+ "grad_norm": 0.024151179939508438,
+ "learning_rate": 0.0006,
+ "loss": 4.403108596801758,
+ "step": 1661
+ },
+ {
+ "epoch": 23.083842794759825,
+ "grad_norm": 0.02511233650147915,
+ "learning_rate": 0.0006,
+ "loss": 4.414128303527832,
+ "step": 1662
+ },
+ {
+ "epoch": 23.097816593886463,
+ "grad_norm": 0.025802113115787506,
+ "learning_rate": 0.0006,
+ "loss": 4.388265609741211,
+ "step": 1663
+ },
+ {
+ "epoch": 23.111790393013102,
+ "grad_norm": 0.02377348765730858,
+ "learning_rate": 0.0006,
+ "loss": 4.357505798339844,
+ "step": 1664
+ },
+ {
+ "epoch": 23.125764192139737,
+ "grad_norm": 0.020876651629805565,
+ "learning_rate": 0.0006,
+ "loss": 4.23043966293335,
+ "step": 1665
+ },
+ {
+ "epoch": 23.139737991266376,
+ "grad_norm": 0.01959727331995964,
+ "learning_rate": 0.0006,
+ "loss": 4.280970096588135,
+ "step": 1666
+ },
+ {
+ "epoch": 23.153711790393015,
+ "grad_norm": 0.018678616732358932,
+ "learning_rate": 0.0006,
+ "loss": 4.346716403961182,
+ "step": 1667
+ },
+ {
+ "epoch": 23.16768558951965,
+ "grad_norm": 0.02108944021165371,
+ "learning_rate": 0.0006,
+ "loss": 4.277904510498047,
+ "step": 1668
+ },
+ {
+ "epoch": 23.18165938864629,
+ "grad_norm": 0.020221566781401634,
+ "learning_rate": 0.0006,
+ "loss": 4.367693901062012,
+ "step": 1669
+ },
+ {
+ "epoch": 23.195633187772927,
+ "grad_norm": 0.019599711522459984,
+ "learning_rate": 0.0006,
+ "loss": 4.257092475891113,
+ "step": 1670
+ },
+ {
+ "epoch": 23.209606986899562,
+ "grad_norm": 0.019833361729979515,
+ "learning_rate": 0.0006,
+ "loss": 4.540737628936768,
+ "step": 1671
+ },
+ {
+ "epoch": 23.2235807860262,
+ "grad_norm": 0.020352095365524292,
+ "learning_rate": 0.0006,
+ "loss": 4.360923767089844,
+ "step": 1672
+ },
+ {
+ "epoch": 23.23755458515284,
+ "grad_norm": 0.01939748041331768,
+ "learning_rate": 0.0006,
+ "loss": 4.393110275268555,
+ "step": 1673
+ },
+ {
+ "epoch": 23.251528384279474,
+ "grad_norm": 0.01884354278445244,
+ "learning_rate": 0.0006,
+ "loss": 4.383938789367676,
+ "step": 1674
+ },
+ {
+ "epoch": 23.265502183406113,
+ "grad_norm": 0.02008470520377159,
+ "learning_rate": 0.0006,
+ "loss": 4.346344947814941,
+ "step": 1675
+ },
+ {
+ "epoch": 23.27947598253275,
+ "grad_norm": 0.019065195694565773,
+ "learning_rate": 0.0006,
+ "loss": 4.36611795425415,
+ "step": 1676
+ },
+ {
+ "epoch": 23.29344978165939,
+ "grad_norm": 0.01938541606068611,
+ "learning_rate": 0.0006,
+ "loss": 4.4009599685668945,
+ "step": 1677
+ },
+ {
+ "epoch": 23.307423580786025,
+ "grad_norm": 0.019044723361730576,
+ "learning_rate": 0.0006,
+ "loss": 4.340169906616211,
+ "step": 1678
+ },
+ {
+ "epoch": 23.321397379912664,
+ "grad_norm": 0.018850378692150116,
+ "learning_rate": 0.0006,
+ "loss": 4.37674617767334,
+ "step": 1679
+ },
+ {
+ "epoch": 23.335371179039303,
+ "grad_norm": 0.019076095893979073,
+ "learning_rate": 0.0006,
+ "loss": 4.226874351501465,
+ "step": 1680
+ },
+ {
+ "epoch": 23.349344978165938,
+ "grad_norm": 0.017543038353323936,
+ "learning_rate": 0.0006,
+ "loss": 4.337751388549805,
+ "step": 1681
+ },
+ {
+ "epoch": 23.363318777292577,
+ "grad_norm": 0.017170218750834465,
+ "learning_rate": 0.0006,
+ "loss": 4.282512664794922,
+ "step": 1682
+ },
+ {
+ "epoch": 23.377292576419215,
+ "grad_norm": 0.01574796624481678,
+ "learning_rate": 0.0006,
+ "loss": 4.30638313293457,
+ "step": 1683
+ },
+ {
+ "epoch": 23.39126637554585,
+ "grad_norm": 0.015160846523940563,
+ "learning_rate": 0.0006,
+ "loss": 4.336716175079346,
+ "step": 1684
+ },
+ {
+ "epoch": 23.40524017467249,
+ "grad_norm": 0.01534526702016592,
+ "learning_rate": 0.0006,
+ "loss": 4.2174787521362305,
+ "step": 1685
+ },
+ {
+ "epoch": 23.419213973799128,
+ "grad_norm": 0.015289999544620514,
+ "learning_rate": 0.0006,
+ "loss": 4.272546291351318,
+ "step": 1686
+ },
+ {
+ "epoch": 23.433187772925763,
+ "grad_norm": 0.014894185587763786,
+ "learning_rate": 0.0006,
+ "loss": 4.296045303344727,
+ "step": 1687
+ },
+ {
+ "epoch": 23.4471615720524,
+ "grad_norm": 0.014869502745568752,
+ "learning_rate": 0.0006,
+ "loss": 4.425461769104004,
+ "step": 1688
+ },
+ {
+ "epoch": 23.46113537117904,
+ "grad_norm": 0.01474483497440815,
+ "learning_rate": 0.0006,
+ "loss": 4.361867427825928,
+ "step": 1689
+ },
+ {
+ "epoch": 23.475109170305675,
+ "grad_norm": 0.015299245715141296,
+ "learning_rate": 0.0006,
+ "loss": 4.364523410797119,
+ "step": 1690
+ },
+ {
+ "epoch": 23.489082969432314,
+ "grad_norm": 0.015201340429484844,
+ "learning_rate": 0.0006,
+ "loss": 4.285182952880859,
+ "step": 1691
+ },
+ {
+ "epoch": 23.503056768558952,
+ "grad_norm": 0.016057293862104416,
+ "learning_rate": 0.0006,
+ "loss": 4.284873962402344,
+ "step": 1692
+ },
+ {
+ "epoch": 23.51703056768559,
+ "grad_norm": 0.015871062874794006,
+ "learning_rate": 0.0006,
+ "loss": 4.326470851898193,
+ "step": 1693
+ },
+ {
+ "epoch": 23.531004366812226,
+ "grad_norm": 0.01586288772523403,
+ "learning_rate": 0.0006,
+ "loss": 4.324883460998535,
+ "step": 1694
+ },
+ {
+ "epoch": 23.544978165938865,
+ "grad_norm": 0.016092827543616295,
+ "learning_rate": 0.0006,
+ "loss": 4.3258867263793945,
+ "step": 1695
+ },
+ {
+ "epoch": 23.558951965065503,
+ "grad_norm": 0.01582016795873642,
+ "learning_rate": 0.0006,
+ "loss": 4.358892440795898,
+ "step": 1696
+ },
+ {
+ "epoch": 23.57292576419214,
+ "grad_norm": 0.01688452623784542,
+ "learning_rate": 0.0006,
+ "loss": 4.280842304229736,
+ "step": 1697
+ },
+ {
+ "epoch": 23.586899563318777,
+ "grad_norm": 0.015865886583924294,
+ "learning_rate": 0.0006,
+ "loss": 4.245567321777344,
+ "step": 1698
+ },
+ {
+ "epoch": 23.600873362445416,
+ "grad_norm": 0.015303662046790123,
+ "learning_rate": 0.0006,
+ "loss": 4.383112907409668,
+ "step": 1699
+ },
+ {
+ "epoch": 23.61484716157205,
+ "grad_norm": 0.015649516135454178,
+ "learning_rate": 0.0006,
+ "loss": 4.519082069396973,
+ "step": 1700
+ },
+ {
+ "epoch": 23.62882096069869,
+ "grad_norm": 0.0167120099067688,
+ "learning_rate": 0.0006,
+ "loss": 4.447847366333008,
+ "step": 1701
+ },
+ {
+ "epoch": 23.64279475982533,
+ "grad_norm": 0.016368716955184937,
+ "learning_rate": 0.0006,
+ "loss": 4.434186935424805,
+ "step": 1702
+ },
+ {
+ "epoch": 23.656768558951963,
+ "grad_norm": 0.017379503697156906,
+ "learning_rate": 0.0006,
+ "loss": 4.317594528198242,
+ "step": 1703
+ },
+ {
+ "epoch": 23.670742358078602,
+ "grad_norm": 0.018573811277747154,
+ "learning_rate": 0.0006,
+ "loss": 4.205793857574463,
+ "step": 1704
+ },
+ {
+ "epoch": 23.68471615720524,
+ "grad_norm": 0.01749090477824211,
+ "learning_rate": 0.0006,
+ "loss": 4.3985700607299805,
+ "step": 1705
+ },
+ {
+ "epoch": 23.69868995633188,
+ "grad_norm": 0.0165668074041605,
+ "learning_rate": 0.0006,
+ "loss": 4.329226970672607,
+ "step": 1706
+ },
+ {
+ "epoch": 23.712663755458514,
+ "grad_norm": 0.016913846135139465,
+ "learning_rate": 0.0006,
+ "loss": 4.388485908508301,
+ "step": 1707
+ },
+ {
+ "epoch": 23.726637554585153,
+ "grad_norm": 0.016813859343528748,
+ "learning_rate": 0.0006,
+ "loss": 4.3395915031433105,
+ "step": 1708
+ },
+ {
+ "epoch": 23.74061135371179,
+ "grad_norm": 0.01676975190639496,
+ "learning_rate": 0.0006,
+ "loss": 4.254745960235596,
+ "step": 1709
+ },
+ {
+ "epoch": 23.754585152838427,
+ "grad_norm": 0.016370350494980812,
+ "learning_rate": 0.0006,
+ "loss": 4.2612223625183105,
+ "step": 1710
+ },
+ {
+ "epoch": 23.768558951965066,
+ "grad_norm": 0.01696198247373104,
+ "learning_rate": 0.0006,
+ "loss": 4.327524662017822,
+ "step": 1711
+ },
+ {
+ "epoch": 23.782532751091704,
+ "grad_norm": 0.0189108457416296,
+ "learning_rate": 0.0006,
+ "loss": 4.325839519500732,
+ "step": 1712
+ },
+ {
+ "epoch": 23.79650655021834,
+ "grad_norm": 0.01862351782619953,
+ "learning_rate": 0.0006,
+ "loss": 4.333662509918213,
+ "step": 1713
+ },
+ {
+ "epoch": 23.810480349344978,
+ "grad_norm": 0.01655014418065548,
+ "learning_rate": 0.0006,
+ "loss": 4.39305305480957,
+ "step": 1714
+ },
+ {
+ "epoch": 23.824454148471617,
+ "grad_norm": 0.015899376943707466,
+ "learning_rate": 0.0006,
+ "loss": 4.390053749084473,
+ "step": 1715
+ },
+ {
+ "epoch": 23.83842794759825,
+ "grad_norm": 0.016519056633114815,
+ "learning_rate": 0.0006,
+ "loss": 4.255903244018555,
+ "step": 1716
+ },
+ {
+ "epoch": 23.85240174672489,
+ "grad_norm": 0.0167539119720459,
+ "learning_rate": 0.0006,
+ "loss": 4.255727767944336,
+ "step": 1717
+ },
+ {
+ "epoch": 23.86637554585153,
+ "grad_norm": 0.0171353816986084,
+ "learning_rate": 0.0006,
+ "loss": 4.376740455627441,
+ "step": 1718
+ },
+ {
+ "epoch": 23.880349344978168,
+ "grad_norm": 0.016719117760658264,
+ "learning_rate": 0.0006,
+ "loss": 4.3376688957214355,
+ "step": 1719
+ },
+ {
+ "epoch": 23.894323144104803,
+ "grad_norm": 0.01561494916677475,
+ "learning_rate": 0.0006,
+ "loss": 4.3345746994018555,
+ "step": 1720
+ },
+ {
+ "epoch": 23.90829694323144,
+ "grad_norm": 0.016303174197673798,
+ "learning_rate": 0.0006,
+ "loss": 4.332371234893799,
+ "step": 1721
+ },
+ {
+ "epoch": 23.92227074235808,
+ "grad_norm": 0.016722865402698517,
+ "learning_rate": 0.0006,
+ "loss": 4.267125129699707,
+ "step": 1722
+ },
+ {
+ "epoch": 23.936244541484715,
+ "grad_norm": 0.017072124406695366,
+ "learning_rate": 0.0006,
+ "loss": 4.322751998901367,
+ "step": 1723
+ },
+ {
+ "epoch": 23.950218340611354,
+ "grad_norm": 0.01699932850897312,
+ "learning_rate": 0.0006,
+ "loss": 4.378422737121582,
+ "step": 1724
+ },
+ {
+ "epoch": 23.964192139737992,
+ "grad_norm": 0.01742720976471901,
+ "learning_rate": 0.0006,
+ "loss": 4.345083236694336,
+ "step": 1725
+ },
+ {
+ "epoch": 23.978165938864628,
+ "grad_norm": 0.018797501921653748,
+ "learning_rate": 0.0006,
+ "loss": 4.290443420410156,
+ "step": 1726
+ },
+ {
+ "epoch": 23.992139737991266,
+ "grad_norm": 0.018090683966875076,
+ "learning_rate": 0.0006,
+ "loss": 4.377580642700195,
+ "step": 1727
+ },
+ {
+ "epoch": 24.0,
+ "grad_norm": 0.017305459827184677,
+ "learning_rate": 0.0006,
+ "loss": 4.364426612854004,
+ "step": 1728
+ },
+ {
+ "epoch": 24.0,
+ "eval_loss": 4.647762298583984,
+ "eval_runtime": 56.5402,
+ "eval_samples_per_second": 43.191,
+ "eval_steps_per_second": 1.362,
+ "step": 1728
+ },
+ {
+ "epoch": 24.01397379912664,
+ "grad_norm": 0.016493607312440872,
+ "learning_rate": 0.0006,
+ "loss": 4.252255916595459,
+ "step": 1729
+ },
+ {
+ "epoch": 24.027947598253274,
+ "grad_norm": 0.016529450193047523,
+ "learning_rate": 0.0006,
+ "loss": 4.352746963500977,
+ "step": 1730
+ },
+ {
+ "epoch": 24.041921397379912,
+ "grad_norm": 0.016940191388130188,
+ "learning_rate": 0.0006,
+ "loss": 4.301870822906494,
+ "step": 1731
+ },
+ {
+ "epoch": 24.05589519650655,
+ "grad_norm": 0.01752558909356594,
+ "learning_rate": 0.0006,
+ "loss": 4.2762451171875,
+ "step": 1732
+ },
+ {
+ "epoch": 24.069868995633186,
+ "grad_norm": 0.017112884670495987,
+ "learning_rate": 0.0006,
+ "loss": 4.2659149169921875,
+ "step": 1733
+ },
+ {
+ "epoch": 24.083842794759825,
+ "grad_norm": 0.01900586113333702,
+ "learning_rate": 0.0006,
+ "loss": 4.306058883666992,
+ "step": 1734
+ },
+ {
+ "epoch": 24.097816593886463,
+ "grad_norm": 0.019447680562734604,
+ "learning_rate": 0.0006,
+ "loss": 4.450128555297852,
+ "step": 1735
+ },
+ {
+ "epoch": 24.111790393013102,
+ "grad_norm": 0.01948186755180359,
+ "learning_rate": 0.0006,
+ "loss": 4.202869415283203,
+ "step": 1736
+ },
+ {
+ "epoch": 24.125764192139737,
+ "grad_norm": 0.01716383546590805,
+ "learning_rate": 0.0006,
+ "loss": 4.318876266479492,
+ "step": 1737
+ },
+ {
+ "epoch": 24.139737991266376,
+ "grad_norm": 0.017049988731741905,
+ "learning_rate": 0.0006,
+ "loss": 4.237238883972168,
+ "step": 1738
+ },
+ {
+ "epoch": 24.153711790393015,
+ "grad_norm": 0.015169923193752766,
+ "learning_rate": 0.0006,
+ "loss": 4.289426803588867,
+ "step": 1739
+ },
+ {
+ "epoch": 24.16768558951965,
+ "grad_norm": 0.014760667458176613,
+ "learning_rate": 0.0006,
+ "loss": 4.356924057006836,
+ "step": 1740
+ },
+ {
+ "epoch": 24.18165938864629,
+ "grad_norm": 0.015127750113606453,
+ "learning_rate": 0.0006,
+ "loss": 4.142916679382324,
+ "step": 1741
+ },
+ {
+ "epoch": 24.195633187772927,
+ "grad_norm": 0.01549555640667677,
+ "learning_rate": 0.0006,
+ "loss": 4.345563888549805,
+ "step": 1742
+ },
+ {
+ "epoch": 24.209606986899562,
+ "grad_norm": 0.016393091529607773,
+ "learning_rate": 0.0006,
+ "loss": 4.417495250701904,
+ "step": 1743
+ },
+ {
+ "epoch": 24.2235807860262,
+ "grad_norm": 0.01785266026854515,
+ "learning_rate": 0.0006,
+ "loss": 4.3553547859191895,
+ "step": 1744
+ },
+ {
+ "epoch": 24.23755458515284,
+ "grad_norm": 0.017574617639183998,
+ "learning_rate": 0.0006,
+ "loss": 4.241024494171143,
+ "step": 1745
+ },
+ {
+ "epoch": 24.251528384279474,
+ "grad_norm": 0.019060306251049042,
+ "learning_rate": 0.0006,
+ "loss": 4.3685150146484375,
+ "step": 1746
+ },
+ {
+ "epoch": 24.265502183406113,
+ "grad_norm": 0.0214633010327816,
+ "learning_rate": 0.0006,
+ "loss": 4.323866844177246,
+ "step": 1747
+ },
+ {
+ "epoch": 24.27947598253275,
+ "grad_norm": 0.021055003628134727,
+ "learning_rate": 0.0006,
+ "loss": 4.278757095336914,
+ "step": 1748
+ },
+ {
+ "epoch": 24.29344978165939,
+ "grad_norm": 0.018618572503328323,
+ "learning_rate": 0.0006,
+ "loss": 4.28849983215332,
+ "step": 1749
+ },
+ {
+ "epoch": 24.307423580786025,
+ "grad_norm": 0.018666435033082962,
+ "learning_rate": 0.0006,
+ "loss": 4.368590354919434,
+ "step": 1750
+ },
+ {
+ "epoch": 24.321397379912664,
+ "grad_norm": 0.020626146346330643,
+ "learning_rate": 0.0006,
+ "loss": 4.2951836585998535,
+ "step": 1751
+ },
+ {
+ "epoch": 24.335371179039303,
+ "grad_norm": 0.01929759792983532,
+ "learning_rate": 0.0006,
+ "loss": 4.3664703369140625,
+ "step": 1752
+ },
+ {
+ "epoch": 24.349344978165938,
+ "grad_norm": 0.020956002175807953,
+ "learning_rate": 0.0006,
+ "loss": 4.345308303833008,
+ "step": 1753
+ },
+ {
+ "epoch": 24.363318777292577,
+ "grad_norm": 0.019227957352995872,
+ "learning_rate": 0.0006,
+ "loss": 4.349849224090576,
+ "step": 1754
+ },
+ {
+ "epoch": 24.377292576419215,
+ "grad_norm": 0.015970097854733467,
+ "learning_rate": 0.0006,
+ "loss": 4.223012447357178,
+ "step": 1755
+ },
+ {
+ "epoch": 24.39126637554585,
+ "grad_norm": 0.015951646491885185,
+ "learning_rate": 0.0006,
+ "loss": 4.236420631408691,
+ "step": 1756
+ },
+ {
+ "epoch": 24.40524017467249,
+ "grad_norm": 0.016032515093684196,
+ "learning_rate": 0.0006,
+ "loss": 4.226899147033691,
+ "step": 1757
+ },
+ {
+ "epoch": 24.419213973799128,
+ "grad_norm": 0.01573132909834385,
+ "learning_rate": 0.0006,
+ "loss": 4.315916061401367,
+ "step": 1758
+ },
+ {
+ "epoch": 24.433187772925763,
+ "grad_norm": 0.015367005951702595,
+ "learning_rate": 0.0006,
+ "loss": 4.358939170837402,
+ "step": 1759
+ },
+ {
+ "epoch": 24.4471615720524,
+ "grad_norm": 0.01669907011091709,
+ "learning_rate": 0.0006,
+ "loss": 4.232147216796875,
+ "step": 1760
+ },
+ {
+ "epoch": 24.46113537117904,
+ "grad_norm": 0.018708152696490288,
+ "learning_rate": 0.0006,
+ "loss": 4.329852104187012,
+ "step": 1761
+ },
+ {
+ "epoch": 24.475109170305675,
+ "grad_norm": 0.019035987555980682,
+ "learning_rate": 0.0006,
+ "loss": 4.260340213775635,
+ "step": 1762
+ },
+ {
+ "epoch": 24.489082969432314,
+ "grad_norm": 0.01832144521176815,
+ "learning_rate": 0.0006,
+ "loss": 4.359543800354004,
+ "step": 1763
+ },
+ {
+ "epoch": 24.503056768558952,
+ "grad_norm": 0.015313294716179371,
+ "learning_rate": 0.0006,
+ "loss": 4.279088020324707,
+ "step": 1764
+ },
+ {
+ "epoch": 24.51703056768559,
+ "grad_norm": 0.015803923830389977,
+ "learning_rate": 0.0006,
+ "loss": 4.230682849884033,
+ "step": 1765
+ },
+ {
+ "epoch": 24.531004366812226,
+ "grad_norm": 0.016748374328017235,
+ "learning_rate": 0.0006,
+ "loss": 4.349661827087402,
+ "step": 1766
+ },
+ {
+ "epoch": 24.544978165938865,
+ "grad_norm": 0.015732428058981895,
+ "learning_rate": 0.0006,
+ "loss": 4.372441291809082,
+ "step": 1767
+ },
+ {
+ "epoch": 24.558951965065503,
+ "grad_norm": 0.016564829275012016,
+ "learning_rate": 0.0006,
+ "loss": 4.323090553283691,
+ "step": 1768
+ },
+ {
+ "epoch": 24.57292576419214,
+ "grad_norm": 0.01603010483086109,
+ "learning_rate": 0.0006,
+ "loss": 4.347898006439209,
+ "step": 1769
+ },
+ {
+ "epoch": 24.586899563318777,
+ "grad_norm": 0.01674809865653515,
+ "learning_rate": 0.0006,
+ "loss": 4.269165515899658,
+ "step": 1770
+ },
+ {
+ "epoch": 24.600873362445416,
+ "grad_norm": 0.017333930358290672,
+ "learning_rate": 0.0006,
+ "loss": 4.353298187255859,
+ "step": 1771
+ },
+ {
+ "epoch": 24.61484716157205,
+ "grad_norm": 0.017519580200314522,
+ "learning_rate": 0.0006,
+ "loss": 4.265199661254883,
+ "step": 1772
+ },
+ {
+ "epoch": 24.62882096069869,
+ "grad_norm": 0.017675306648015976,
+ "learning_rate": 0.0006,
+ "loss": 4.322596549987793,
+ "step": 1773
+ },
+ {
+ "epoch": 24.64279475982533,
+ "grad_norm": 0.017166534438729286,
+ "learning_rate": 0.0006,
+ "loss": 4.413900375366211,
+ "step": 1774
+ },
+ {
+ "epoch": 24.656768558951963,
+ "grad_norm": 0.016362346708774567,
+ "learning_rate": 0.0006,
+ "loss": 4.307236671447754,
+ "step": 1775
+ },
+ {
+ "epoch": 24.670742358078602,
+ "grad_norm": 0.016086872667074203,
+ "learning_rate": 0.0006,
+ "loss": 4.235316276550293,
+ "step": 1776
+ },
+ {
+ "epoch": 24.68471615720524,
+ "grad_norm": 0.01689537614583969,
+ "learning_rate": 0.0006,
+ "loss": 4.390591621398926,
+ "step": 1777
+ },
+ {
+ "epoch": 24.69868995633188,
+ "grad_norm": 0.016836725175380707,
+ "learning_rate": 0.0006,
+ "loss": 4.2990031242370605,
+ "step": 1778
+ },
+ {
+ "epoch": 24.712663755458514,
+ "grad_norm": 0.015627335757017136,
+ "learning_rate": 0.0006,
+ "loss": 4.166745185852051,
+ "step": 1779
+ },
+ {
+ "epoch": 24.726637554585153,
+ "grad_norm": 0.015007016249001026,
+ "learning_rate": 0.0006,
+ "loss": 4.295273303985596,
+ "step": 1780
+ },
+ {
+ "epoch": 24.74061135371179,
+ "grad_norm": 0.016084209084510803,
+ "learning_rate": 0.0006,
+ "loss": 4.342424392700195,
+ "step": 1781
+ },
+ {
+ "epoch": 24.754585152838427,
+ "grad_norm": 0.01595073565840721,
+ "learning_rate": 0.0006,
+ "loss": 4.3060102462768555,
+ "step": 1782
+ },
+ {
+ "epoch": 24.768558951965066,
+ "grad_norm": 0.016113461926579475,
+ "learning_rate": 0.0006,
+ "loss": 4.267908096313477,
+ "step": 1783
+ },
+ {
+ "epoch": 24.782532751091704,
+ "grad_norm": 0.01564556173980236,
+ "learning_rate": 0.0006,
+ "loss": 4.324642658233643,
+ "step": 1784
+ },
+ {
+ "epoch": 24.79650655021834,
+ "grad_norm": 0.01575268618762493,
+ "learning_rate": 0.0006,
+ "loss": 4.328181743621826,
+ "step": 1785
+ },
+ {
+ "epoch": 24.810480349344978,
+ "grad_norm": 0.016830824315547943,
+ "learning_rate": 0.0006,
+ "loss": 4.247906684875488,
+ "step": 1786
+ },
+ {
+ "epoch": 24.824454148471617,
+ "grad_norm": 0.017989547923207283,
+ "learning_rate": 0.0006,
+ "loss": 4.228281021118164,
+ "step": 1787
+ },
+ {
+ "epoch": 24.83842794759825,
+ "grad_norm": 0.016783086583018303,
+ "learning_rate": 0.0006,
+ "loss": 4.2880964279174805,
+ "step": 1788
+ },
+ {
+ "epoch": 24.85240174672489,
+ "grad_norm": 0.016467561945319176,
+ "learning_rate": 0.0006,
+ "loss": 4.36118745803833,
+ "step": 1789
+ },
+ {
+ "epoch": 24.86637554585153,
+ "grad_norm": 0.01780523732304573,
+ "learning_rate": 0.0006,
+ "loss": 4.172321319580078,
+ "step": 1790
+ },
+ {
+ "epoch": 24.880349344978168,
+ "grad_norm": 0.017102031037211418,
+ "learning_rate": 0.0006,
+ "loss": 4.352826118469238,
+ "step": 1791
+ },
+ {
+ "epoch": 24.894323144104803,
+ "grad_norm": 0.016409218311309814,
+ "learning_rate": 0.0006,
+ "loss": 4.212162017822266,
+ "step": 1792
+ },
+ {
+ "epoch": 24.90829694323144,
+ "grad_norm": 0.018660584464669228,
+ "learning_rate": 0.0006,
+ "loss": 4.346684455871582,
+ "step": 1793
+ },
+ {
+ "epoch": 24.92227074235808,
+ "grad_norm": 0.017389440909028053,
+ "learning_rate": 0.0006,
+ "loss": 4.197913646697998,
+ "step": 1794
+ },
+ {
+ "epoch": 24.936244541484715,
+ "grad_norm": 0.01596485823392868,
+ "learning_rate": 0.0006,
+ "loss": 4.370617866516113,
+ "step": 1795
+ },
+ {
+ "epoch": 24.950218340611354,
+ "grad_norm": 0.01709776371717453,
+ "learning_rate": 0.0006,
+ "loss": 4.313847064971924,
+ "step": 1796
+ },
+ {
+ "epoch": 24.964192139737992,
+ "grad_norm": 0.016571134328842163,
+ "learning_rate": 0.0006,
+ "loss": 4.1895856857299805,
+ "step": 1797
+ },
+ {
+ "epoch": 24.978165938864628,
+ "grad_norm": 0.016851048916578293,
+ "learning_rate": 0.0006,
+ "loss": 4.350772857666016,
+ "step": 1798
+ },
+ {
+ "epoch": 24.992139737991266,
+ "grad_norm": 0.017007440328598022,
+ "learning_rate": 0.0006,
+ "loss": 4.399008750915527,
+ "step": 1799
+ },
+ {
+ "epoch": 25.0,
+ "grad_norm": 0.01791212148964405,
+ "learning_rate": 0.0006,
+ "loss": 4.306121826171875,
+ "step": 1800
+ },
+ {
+ "epoch": 25.0,
+ "eval_loss": 4.636417865753174,
+ "eval_runtime": 56.8296,
+ "eval_samples_per_second": 42.971,
+ "eval_steps_per_second": 1.355,
+ "step": 1800
+ },
+ {
+ "epoch": 25.01397379912664,
+ "grad_norm": 0.01729617640376091,
+ "learning_rate": 0.0006,
+ "loss": 4.257018089294434,
+ "step": 1801
+ },
+ {
+ "epoch": 25.027947598253274,
+ "grad_norm": 0.017466643825173378,
+ "learning_rate": 0.0006,
+ "loss": 4.307605266571045,
+ "step": 1802
+ },
+ {
+ "epoch": 25.041921397379912,
+ "grad_norm": 0.018614279106259346,
+ "learning_rate": 0.0006,
+ "loss": 4.360141277313232,
+ "step": 1803
+ },
+ {
+ "epoch": 25.05589519650655,
+ "grad_norm": 0.02010771445930004,
+ "learning_rate": 0.0006,
+ "loss": 4.196428298950195,
+ "step": 1804
+ },
+ {
+ "epoch": 25.069868995633186,
+ "grad_norm": 0.02090228535234928,
+ "learning_rate": 0.0006,
+ "loss": 4.288947105407715,
+ "step": 1805
+ },
+ {
+ "epoch": 25.083842794759825,
+ "grad_norm": 0.02050255797803402,
+ "learning_rate": 0.0006,
+ "loss": 4.259582042694092,
+ "step": 1806
+ },
+ {
+ "epoch": 25.097816593886463,
+ "grad_norm": 0.021844089031219482,
+ "learning_rate": 0.0006,
+ "loss": 4.405138969421387,
+ "step": 1807
+ },
+ {
+ "epoch": 25.111790393013102,
+ "grad_norm": 0.022287018597126007,
+ "learning_rate": 0.0006,
+ "loss": 4.333793640136719,
+ "step": 1808
+ },
+ {
+ "epoch": 25.125764192139737,
+ "grad_norm": 0.02288329415023327,
+ "learning_rate": 0.0006,
+ "loss": 4.3910980224609375,
+ "step": 1809
+ },
+ {
+ "epoch": 25.139737991266376,
+ "grad_norm": 0.02373199723660946,
+ "learning_rate": 0.0006,
+ "loss": 4.32921028137207,
+ "step": 1810
+ },
+ {
+ "epoch": 25.153711790393015,
+ "grad_norm": 0.024960234761238098,
+ "learning_rate": 0.0006,
+ "loss": 4.295609474182129,
+ "step": 1811
+ },
+ {
+ "epoch": 25.16768558951965,
+ "grad_norm": 0.026743892580270767,
+ "learning_rate": 0.0006,
+ "loss": 4.315308094024658,
+ "step": 1812
+ },
+ {
+ "epoch": 25.18165938864629,
+ "grad_norm": 0.022138185799121857,
+ "learning_rate": 0.0006,
+ "loss": 4.360208511352539,
+ "step": 1813
+ },
+ {
+ "epoch": 25.195633187772927,
+ "grad_norm": 0.022120574489235878,
+ "learning_rate": 0.0006,
+ "loss": 4.1973772048950195,
+ "step": 1814
+ },
+ {
+ "epoch": 25.209606986899562,
+ "grad_norm": 0.02230502851307392,
+ "learning_rate": 0.0006,
+ "loss": 4.285558700561523,
+ "step": 1815
+ },
+ {
+ "epoch": 25.2235807860262,
+ "grad_norm": 0.022301986813545227,
+ "learning_rate": 0.0006,
+ "loss": 4.238343238830566,
+ "step": 1816
+ },
+ {
+ "epoch": 25.23755458515284,
+ "grad_norm": 0.020050447434186935,
+ "learning_rate": 0.0006,
+ "loss": 4.298235893249512,
+ "step": 1817
+ },
+ {
+ "epoch": 25.251528384279474,
+ "grad_norm": 0.01987724006175995,
+ "learning_rate": 0.0006,
+ "loss": 4.301384449005127,
+ "step": 1818
+ },
+ {
+ "epoch": 25.265502183406113,
+ "grad_norm": 0.020440855994820595,
+ "learning_rate": 0.0006,
+ "loss": 4.269218444824219,
+ "step": 1819
+ },
+ {
+ "epoch": 25.27947598253275,
+ "grad_norm": 0.01891893707215786,
+ "learning_rate": 0.0006,
+ "loss": 4.2401957511901855,
+ "step": 1820
+ },
+ {
+ "epoch": 25.29344978165939,
+ "grad_norm": 0.01903688907623291,
+ "learning_rate": 0.0006,
+ "loss": 4.139670372009277,
+ "step": 1821
+ },
+ {
+ "epoch": 25.307423580786025,
+ "grad_norm": 0.018856938928365707,
+ "learning_rate": 0.0006,
+ "loss": 4.233214378356934,
+ "step": 1822
+ },
+ {
+ "epoch": 25.321397379912664,
+ "grad_norm": 0.017791323363780975,
+ "learning_rate": 0.0006,
+ "loss": 4.2342023849487305,
+ "step": 1823
+ },
+ {
+ "epoch": 25.335371179039303,
+ "grad_norm": 0.017304565757513046,
+ "learning_rate": 0.0006,
+ "loss": 4.355198860168457,
+ "step": 1824
+ },
+ {
+ "epoch": 25.349344978165938,
+ "grad_norm": 0.01576206088066101,
+ "learning_rate": 0.0006,
+ "loss": 4.114147186279297,
+ "step": 1825
+ },
+ {
+ "epoch": 25.363318777292577,
+ "grad_norm": 0.015105154365301132,
+ "learning_rate": 0.0006,
+ "loss": 4.313795566558838,
+ "step": 1826
+ },
+ {
+ "epoch": 25.377292576419215,
+ "grad_norm": 0.01563587784767151,
+ "learning_rate": 0.0006,
+ "loss": 4.308282852172852,
+ "step": 1827
+ },
+ {
+ "epoch": 25.39126637554585,
+ "grad_norm": 0.015980906784534454,
+ "learning_rate": 0.0006,
+ "loss": 4.345149040222168,
+ "step": 1828
+ },
+ {
+ "epoch": 25.40524017467249,
+ "grad_norm": 0.01669284701347351,
+ "learning_rate": 0.0006,
+ "loss": 4.237611293792725,
+ "step": 1829
+ },
+ {
+ "epoch": 25.419213973799128,
+ "grad_norm": 0.016202064231038094,
+ "learning_rate": 0.0006,
+ "loss": 4.3072309494018555,
+ "step": 1830
+ },
+ {
+ "epoch": 25.433187772925763,
+ "grad_norm": 0.01478662807494402,
+ "learning_rate": 0.0006,
+ "loss": 4.276215076446533,
+ "step": 1831
+ },
+ {
+ "epoch": 25.4471615720524,
+ "grad_norm": 0.015168731100857258,
+ "learning_rate": 0.0006,
+ "loss": 4.37509822845459,
+ "step": 1832
+ },
+ {
+ "epoch": 25.46113537117904,
+ "grad_norm": 0.01648980937898159,
+ "learning_rate": 0.0006,
+ "loss": 4.349167346954346,
+ "step": 1833
+ },
+ {
+ "epoch": 25.475109170305675,
+ "grad_norm": 0.01708398200571537,
+ "learning_rate": 0.0006,
+ "loss": 4.346580505371094,
+ "step": 1834
+ },
+ {
+ "epoch": 25.489082969432314,
+ "grad_norm": 0.017370784655213356,
+ "learning_rate": 0.0006,
+ "loss": 4.299466133117676,
+ "step": 1835
+ },
+ {
+ "epoch": 25.503056768558952,
+ "grad_norm": 0.016248228028416634,
+ "learning_rate": 0.0006,
+ "loss": 4.184610366821289,
+ "step": 1836
+ },
+ {
+ "epoch": 25.51703056768559,
+ "grad_norm": 0.017106913030147552,
+ "learning_rate": 0.0006,
+ "loss": 4.366856098175049,
+ "step": 1837
+ },
+ {
+ "epoch": 25.531004366812226,
+ "grad_norm": 0.016071965917944908,
+ "learning_rate": 0.0006,
+ "loss": 4.297691345214844,
+ "step": 1838
+ },
+ {
+ "epoch": 25.544978165938865,
+ "grad_norm": 0.016451986506581306,
+ "learning_rate": 0.0006,
+ "loss": 4.229983329772949,
+ "step": 1839
+ },
+ {
+ "epoch": 25.558951965065503,
+ "grad_norm": 0.016195081174373627,
+ "learning_rate": 0.0006,
+ "loss": 4.259893417358398,
+ "step": 1840
+ },
+ {
+ "epoch": 25.57292576419214,
+ "grad_norm": 0.0165175162255764,
+ "learning_rate": 0.0006,
+ "loss": 4.3348236083984375,
+ "step": 1841
+ },
+ {
+ "epoch": 25.586899563318777,
+ "grad_norm": 0.01651517115533352,
+ "learning_rate": 0.0006,
+ "loss": 4.319541931152344,
+ "step": 1842
+ },
+ {
+ "epoch": 25.600873362445416,
+ "grad_norm": 0.016143690794706345,
+ "learning_rate": 0.0006,
+ "loss": 4.345652103424072,
+ "step": 1843
+ },
+ {
+ "epoch": 25.61484716157205,
+ "grad_norm": 0.015140696428716183,
+ "learning_rate": 0.0006,
+ "loss": 4.212047576904297,
+ "step": 1844
+ },
+ {
+ "epoch": 25.62882096069869,
+ "grad_norm": 0.014637443237006664,
+ "learning_rate": 0.0006,
+ "loss": 4.187453269958496,
+ "step": 1845
+ },
+ {
+ "epoch": 25.64279475982533,
+ "grad_norm": 0.014585614204406738,
+ "learning_rate": 0.0006,
+ "loss": 4.253936767578125,
+ "step": 1846
+ },
+ {
+ "epoch": 25.656768558951963,
+ "grad_norm": 0.015158289112150669,
+ "learning_rate": 0.0006,
+ "loss": 4.34612512588501,
+ "step": 1847
+ },
+ {
+ "epoch": 25.670742358078602,
+ "grad_norm": 0.015202849172055721,
+ "learning_rate": 0.0006,
+ "loss": 4.270634651184082,
+ "step": 1848
+ },
+ {
+ "epoch": 25.68471615720524,
+ "grad_norm": 0.015363575890660286,
+ "learning_rate": 0.0006,
+ "loss": 4.271080493927002,
+ "step": 1849
+ },
+ {
+ "epoch": 25.69868995633188,
+ "grad_norm": 0.014360226690769196,
+ "learning_rate": 0.0006,
+ "loss": 4.375033855438232,
+ "step": 1850
+ },
+ {
+ "epoch": 25.712663755458514,
+ "grad_norm": 0.014263181947171688,
+ "learning_rate": 0.0006,
+ "loss": 4.424138069152832,
+ "step": 1851
+ },
+ {
+ "epoch": 25.726637554585153,
+ "grad_norm": 0.014398688450455666,
+ "learning_rate": 0.0006,
+ "loss": 4.410154819488525,
+ "step": 1852
+ },
+ {
+ "epoch": 25.74061135371179,
+ "grad_norm": 0.016131488606333733,
+ "learning_rate": 0.0006,
+ "loss": 4.324014663696289,
+ "step": 1853
+ },
+ {
+ "epoch": 25.754585152838427,
+ "grad_norm": 0.017568735405802727,
+ "learning_rate": 0.0006,
+ "loss": 4.367861747741699,
+ "step": 1854
+ },
+ {
+ "epoch": 25.768558951965066,
+ "grad_norm": 0.017907511442899704,
+ "learning_rate": 0.0006,
+ "loss": 4.352965354919434,
+ "step": 1855
+ },
+ {
+ "epoch": 25.782532751091704,
+ "grad_norm": 0.016918016597628593,
+ "learning_rate": 0.0006,
+ "loss": 4.252397537231445,
+ "step": 1856
+ },
+ {
+ "epoch": 25.79650655021834,
+ "grad_norm": 0.017274610698223114,
+ "learning_rate": 0.0006,
+ "loss": 4.342309951782227,
+ "step": 1857
+ },
+ {
+ "epoch": 25.810480349344978,
+ "grad_norm": 0.017879825085401535,
+ "learning_rate": 0.0006,
+ "loss": 4.330987453460693,
+ "step": 1858
+ },
+ {
+ "epoch": 25.824454148471617,
+ "grad_norm": 0.01736099272966385,
+ "learning_rate": 0.0006,
+ "loss": 4.307255744934082,
+ "step": 1859
+ },
+ {
+ "epoch": 25.83842794759825,
+ "grad_norm": 0.016575824469327927,
+ "learning_rate": 0.0006,
+ "loss": 4.262016773223877,
+ "step": 1860
+ },
+ {
+ "epoch": 25.85240174672489,
+ "grad_norm": 0.015181140042841434,
+ "learning_rate": 0.0006,
+ "loss": 4.291863441467285,
+ "step": 1861
+ },
+ {
+ "epoch": 25.86637554585153,
+ "grad_norm": 0.015965687111020088,
+ "learning_rate": 0.0006,
+ "loss": 4.424836158752441,
+ "step": 1862
+ },
+ {
+ "epoch": 25.880349344978168,
+ "grad_norm": 0.01687219925224781,
+ "learning_rate": 0.0006,
+ "loss": 4.343531131744385,
+ "step": 1863
+ },
+ {
+ "epoch": 25.894323144104803,
+ "grad_norm": 0.015115504153072834,
+ "learning_rate": 0.0006,
+ "loss": 4.357808589935303,
+ "step": 1864
+ },
+ {
+ "epoch": 25.90829694323144,
+ "grad_norm": 0.014150998555123806,
+ "learning_rate": 0.0006,
+ "loss": 4.324550151824951,
+ "step": 1865
+ },
+ {
+ "epoch": 25.92227074235808,
+ "grad_norm": 0.014989510178565979,
+ "learning_rate": 0.0006,
+ "loss": 4.380428314208984,
+ "step": 1866
+ },
+ {
+ "epoch": 25.936244541484715,
+ "grad_norm": 0.01527960691601038,
+ "learning_rate": 0.0006,
+ "loss": 4.399374961853027,
+ "step": 1867
+ },
+ {
+ "epoch": 25.950218340611354,
+ "grad_norm": 0.015434633940458298,
+ "learning_rate": 0.0006,
+ "loss": 4.260984420776367,
+ "step": 1868
+ },
+ {
+ "epoch": 25.964192139737992,
+ "grad_norm": 0.015370228327810764,
+ "learning_rate": 0.0006,
+ "loss": 4.3340253829956055,
+ "step": 1869
+ },
+ {
+ "epoch": 25.978165938864628,
+ "grad_norm": 0.015348542481660843,
+ "learning_rate": 0.0006,
+ "loss": 4.369132995605469,
+ "step": 1870
+ },
+ {
+ "epoch": 25.992139737991266,
+ "grad_norm": 0.015354936942458153,
+ "learning_rate": 0.0006,
+ "loss": 4.381937026977539,
+ "step": 1871
+ },
+ {
+ "epoch": 26.0,
+ "grad_norm": 0.01796272210776806,
+ "learning_rate": 0.0006,
+ "loss": 4.286768436431885,
+ "step": 1872
+ },
+ {
+ "epoch": 26.0,
+ "eval_loss": 4.576984405517578,
+ "eval_runtime": 57.2091,
+ "eval_samples_per_second": 42.685,
+ "eval_steps_per_second": 1.346,
+ "step": 1872
+ },
+ {
+ "epoch": 26.01397379912664,
+ "grad_norm": 0.01748845726251602,
+ "learning_rate": 0.0006,
+ "loss": 4.337504863739014,
+ "step": 1873
+ },
+ {
+ "epoch": 26.027947598253274,
+ "grad_norm": 0.019244296476244926,
+ "learning_rate": 0.0006,
+ "loss": 4.264280796051025,
+ "step": 1874
+ },
+ {
+ "epoch": 26.041921397379912,
+ "grad_norm": 0.0215692650526762,
+ "learning_rate": 0.0006,
+ "loss": 4.350830078125,
+ "step": 1875
+ },
+ {
+ "epoch": 26.05589519650655,
+ "grad_norm": 0.022189252078533173,
+ "learning_rate": 0.0006,
+ "loss": 4.299098968505859,
+ "step": 1876
+ },
+ {
+ "epoch": 26.069868995633186,
+ "grad_norm": 0.022384043782949448,
+ "learning_rate": 0.0006,
+ "loss": 4.1185173988342285,
+ "step": 1877
+ },
+ {
+ "epoch": 26.083842794759825,
+ "grad_norm": 0.02076014317572117,
+ "learning_rate": 0.0006,
+ "loss": 4.297612190246582,
+ "step": 1878
+ },
+ {
+ "epoch": 26.097816593886463,
+ "grad_norm": 0.020843051373958588,
+ "learning_rate": 0.0006,
+ "loss": 4.294098854064941,
+ "step": 1879
+ },
+ {
+ "epoch": 26.111790393013102,
+ "grad_norm": 0.020654456689953804,
+ "learning_rate": 0.0006,
+ "loss": 4.267856597900391,
+ "step": 1880
+ },
+ {
+ "epoch": 26.125764192139737,
+ "grad_norm": 0.01860472559928894,
+ "learning_rate": 0.0006,
+ "loss": 4.284232139587402,
+ "step": 1881
+ },
+ {
+ "epoch": 26.139737991266376,
+ "grad_norm": 0.018719421699643135,
+ "learning_rate": 0.0006,
+ "loss": 4.277214527130127,
+ "step": 1882
+ },
+ {
+ "epoch": 26.153711790393015,
+ "grad_norm": 0.020903829485177994,
+ "learning_rate": 0.0006,
+ "loss": 4.254947662353516,
+ "step": 1883
+ },
+ {
+ "epoch": 26.16768558951965,
+ "grad_norm": 0.02020149491727352,
+ "learning_rate": 0.0006,
+ "loss": 4.2526936531066895,
+ "step": 1884
+ },
+ {
+ "epoch": 26.18165938864629,
+ "grad_norm": 0.0180392786860466,
+ "learning_rate": 0.0006,
+ "loss": 4.309922695159912,
+ "step": 1885
+ },
+ {
+ "epoch": 26.195633187772927,
+ "grad_norm": 0.018163125962018967,
+ "learning_rate": 0.0006,
+ "loss": 4.13860559463501,
+ "step": 1886
+ },
+ {
+ "epoch": 26.209606986899562,
+ "grad_norm": 0.01853291131556034,
+ "learning_rate": 0.0006,
+ "loss": 4.190926551818848,
+ "step": 1887
+ },
+ {
+ "epoch": 26.2235807860262,
+ "grad_norm": 0.01757677271962166,
+ "learning_rate": 0.0006,
+ "loss": 4.163999557495117,
+ "step": 1888
+ },
+ {
+ "epoch": 26.23755458515284,
+ "grad_norm": 0.01767992228269577,
+ "learning_rate": 0.0006,
+ "loss": 4.184430122375488,
+ "step": 1889
+ },
+ {
+ "epoch": 26.251528384279474,
+ "grad_norm": 0.018139973282814026,
+ "learning_rate": 0.0006,
+ "loss": 4.300766944885254,
+ "step": 1890
+ },
+ {
+ "epoch": 26.265502183406113,
+ "grad_norm": 0.020184461027383804,
+ "learning_rate": 0.0006,
+ "loss": 4.232537269592285,
+ "step": 1891
+ },
+ {
+ "epoch": 26.27947598253275,
+ "grad_norm": 0.02028856985270977,
+ "learning_rate": 0.0006,
+ "loss": 4.324883460998535,
+ "step": 1892
+ },
+ {
+ "epoch": 26.29344978165939,
+ "grad_norm": 0.017120616510510445,
+ "learning_rate": 0.0006,
+ "loss": 4.23493766784668,
+ "step": 1893
+ },
+ {
+ "epoch": 26.307423580786025,
+ "grad_norm": 0.016202978789806366,
+ "learning_rate": 0.0006,
+ "loss": 4.197602272033691,
+ "step": 1894
+ },
+ {
+ "epoch": 26.321397379912664,
+ "grad_norm": 0.017231421545147896,
+ "learning_rate": 0.0006,
+ "loss": 4.302001953125,
+ "step": 1895
+ },
+ {
+ "epoch": 26.335371179039303,
+ "grad_norm": 0.01725655607879162,
+ "learning_rate": 0.0006,
+ "loss": 4.356800079345703,
+ "step": 1896
+ },
+ {
+ "epoch": 26.349344978165938,
+ "grad_norm": 0.016657443717122078,
+ "learning_rate": 0.0006,
+ "loss": 4.266307830810547,
+ "step": 1897
+ },
+ {
+ "epoch": 26.363318777292577,
+ "grad_norm": 0.01616556942462921,
+ "learning_rate": 0.0006,
+ "loss": 4.307868003845215,
+ "step": 1898
+ },
+ {
+ "epoch": 26.377292576419215,
+ "grad_norm": 0.016621306538581848,
+ "learning_rate": 0.0006,
+ "loss": 4.172905921936035,
+ "step": 1899
+ },
+ {
+ "epoch": 26.39126637554585,
+ "grad_norm": 0.016243206337094307,
+ "learning_rate": 0.0006,
+ "loss": 4.162896156311035,
+ "step": 1900
+ },
+ {
+ "epoch": 26.40524017467249,
+ "grad_norm": 0.016663808375597,
+ "learning_rate": 0.0006,
+ "loss": 4.258232116699219,
+ "step": 1901
+ },
+ {
+ "epoch": 26.419213973799128,
+ "grad_norm": 0.01677478663623333,
+ "learning_rate": 0.0006,
+ "loss": 4.336090087890625,
+ "step": 1902
+ },
+ {
+ "epoch": 26.433187772925763,
+ "grad_norm": 0.01641303487122059,
+ "learning_rate": 0.0006,
+ "loss": 4.191476821899414,
+ "step": 1903
+ },
+ {
+ "epoch": 26.4471615720524,
+ "grad_norm": 0.016700323671102524,
+ "learning_rate": 0.0006,
+ "loss": 4.299427509307861,
+ "step": 1904
+ },
+ {
+ "epoch": 26.46113537117904,
+ "grad_norm": 0.017595035955309868,
+ "learning_rate": 0.0006,
+ "loss": 4.312828063964844,
+ "step": 1905
+ },
+ {
+ "epoch": 26.475109170305675,
+ "grad_norm": 0.020300284028053284,
+ "learning_rate": 0.0006,
+ "loss": 4.28331184387207,
+ "step": 1906
+ },
+ {
+ "epoch": 26.489082969432314,
+ "grad_norm": 0.02198321744799614,
+ "learning_rate": 0.0006,
+ "loss": 4.3280839920043945,
+ "step": 1907
+ },
+ {
+ "epoch": 26.503056768558952,
+ "grad_norm": 0.02003200724720955,
+ "learning_rate": 0.0006,
+ "loss": 4.325623989105225,
+ "step": 1908
+ },
+ {
+ "epoch": 26.51703056768559,
+ "grad_norm": 0.01870287023484707,
+ "learning_rate": 0.0006,
+ "loss": 4.251203536987305,
+ "step": 1909
+ },
+ {
+ "epoch": 26.531004366812226,
+ "grad_norm": 0.01835448481142521,
+ "learning_rate": 0.0006,
+ "loss": 4.335144996643066,
+ "step": 1910
+ },
+ {
+ "epoch": 26.544978165938865,
+ "grad_norm": 0.02046327292919159,
+ "learning_rate": 0.0006,
+ "loss": 4.238023281097412,
+ "step": 1911
+ },
+ {
+ "epoch": 26.558951965065503,
+ "grad_norm": 0.019771311432123184,
+ "learning_rate": 0.0006,
+ "loss": 4.353594779968262,
+ "step": 1912
+ },
+ {
+ "epoch": 26.57292576419214,
+ "grad_norm": 0.018090544268488884,
+ "learning_rate": 0.0006,
+ "loss": 4.304343223571777,
+ "step": 1913
+ },
+ {
+ "epoch": 26.586899563318777,
+ "grad_norm": 0.017489777877926826,
+ "learning_rate": 0.0006,
+ "loss": 4.204375267028809,
+ "step": 1914
+ },
+ {
+ "epoch": 26.600873362445416,
+ "grad_norm": 0.016785891726613045,
+ "learning_rate": 0.0006,
+ "loss": 4.286669731140137,
+ "step": 1915
+ },
+ {
+ "epoch": 26.61484716157205,
+ "grad_norm": 0.016382789239287376,
+ "learning_rate": 0.0006,
+ "loss": 4.347662448883057,
+ "step": 1916
+ },
+ {
+ "epoch": 26.62882096069869,
+ "grad_norm": 0.016581876203417778,
+ "learning_rate": 0.0006,
+ "loss": 4.2716193199157715,
+ "step": 1917
+ },
+ {
+ "epoch": 26.64279475982533,
+ "grad_norm": 0.015627184882760048,
+ "learning_rate": 0.0006,
+ "loss": 4.311489105224609,
+ "step": 1918
+ },
+ {
+ "epoch": 26.656768558951963,
+ "grad_norm": 0.015908483415842056,
+ "learning_rate": 0.0006,
+ "loss": 4.262099742889404,
+ "step": 1919
+ },
+ {
+ "epoch": 26.670742358078602,
+ "grad_norm": 0.015576236881315708,
+ "learning_rate": 0.0006,
+ "loss": 4.275871276855469,
+ "step": 1920
+ },
+ {
+ "epoch": 26.68471615720524,
+ "grad_norm": 0.014621064998209476,
+ "learning_rate": 0.0006,
+ "loss": 4.184260845184326,
+ "step": 1921
+ },
+ {
+ "epoch": 26.69868995633188,
+ "grad_norm": 0.014797811396420002,
+ "learning_rate": 0.0006,
+ "loss": 4.3412628173828125,
+ "step": 1922
+ },
+ {
+ "epoch": 26.712663755458514,
+ "grad_norm": 0.015610570088028908,
+ "learning_rate": 0.0006,
+ "loss": 4.181094646453857,
+ "step": 1923
+ },
+ {
+ "epoch": 26.726637554585153,
+ "grad_norm": 0.016048265621066093,
+ "learning_rate": 0.0006,
+ "loss": 4.365290641784668,
+ "step": 1924
+ },
+ {
+ "epoch": 26.74061135371179,
+ "grad_norm": 0.016604071483016014,
+ "learning_rate": 0.0006,
+ "loss": 4.2628912925720215,
+ "step": 1925
+ },
+ {
+ "epoch": 26.754585152838427,
+ "grad_norm": 0.017044221982359886,
+ "learning_rate": 0.0006,
+ "loss": 4.378512382507324,
+ "step": 1926
+ },
+ {
+ "epoch": 26.768558951965066,
+ "grad_norm": 0.01746384985744953,
+ "learning_rate": 0.0006,
+ "loss": 4.3888773918151855,
+ "step": 1927
+ },
+ {
+ "epoch": 26.782532751091704,
+ "grad_norm": 0.01572711206972599,
+ "learning_rate": 0.0006,
+ "loss": 4.189935684204102,
+ "step": 1928
+ },
+ {
+ "epoch": 26.79650655021834,
+ "grad_norm": 0.014976629987359047,
+ "learning_rate": 0.0006,
+ "loss": 4.280525207519531,
+ "step": 1929
+ },
+ {
+ "epoch": 26.810480349344978,
+ "grad_norm": 0.014804664999246597,
+ "learning_rate": 0.0006,
+ "loss": 4.270630836486816,
+ "step": 1930
+ },
+ {
+ "epoch": 26.824454148471617,
+ "grad_norm": 0.014641920104622841,
+ "learning_rate": 0.0006,
+ "loss": 4.319505214691162,
+ "step": 1931
+ },
+ {
+ "epoch": 26.83842794759825,
+ "grad_norm": 0.014928505755960941,
+ "learning_rate": 0.0006,
+ "loss": 4.180147647857666,
+ "step": 1932
+ },
+ {
+ "epoch": 26.85240174672489,
+ "grad_norm": 0.016130365431308746,
+ "learning_rate": 0.0006,
+ "loss": 4.260660648345947,
+ "step": 1933
+ },
+ {
+ "epoch": 26.86637554585153,
+ "grad_norm": 0.015642493963241577,
+ "learning_rate": 0.0006,
+ "loss": 4.304064750671387,
+ "step": 1934
+ },
+ {
+ "epoch": 26.880349344978168,
+ "grad_norm": 0.015988217666745186,
+ "learning_rate": 0.0006,
+ "loss": 4.2548017501831055,
+ "step": 1935
+ },
+ {
+ "epoch": 26.894323144104803,
+ "grad_norm": 0.016501398757100105,
+ "learning_rate": 0.0006,
+ "loss": 4.260058879852295,
+ "step": 1936
+ },
+ {
+ "epoch": 26.90829694323144,
+ "grad_norm": 0.01645040698349476,
+ "learning_rate": 0.0006,
+ "loss": 4.483081340789795,
+ "step": 1937
+ },
+ {
+ "epoch": 26.92227074235808,
+ "grad_norm": 0.016140863299369812,
+ "learning_rate": 0.0006,
+ "loss": 4.1776838302612305,
+ "step": 1938
+ },
+ {
+ "epoch": 26.936244541484715,
+ "grad_norm": 0.016681019216775894,
+ "learning_rate": 0.0006,
+ "loss": 4.224747657775879,
+ "step": 1939
+ },
+ {
+ "epoch": 26.950218340611354,
+ "grad_norm": 0.01650378853082657,
+ "learning_rate": 0.0006,
+ "loss": 4.305703163146973,
+ "step": 1940
+ },
+ {
+ "epoch": 26.964192139737992,
+ "grad_norm": 0.017051683738827705,
+ "learning_rate": 0.0006,
+ "loss": 4.285835266113281,
+ "step": 1941
+ },
+ {
+ "epoch": 26.978165938864628,
+ "grad_norm": 0.016894327476620674,
+ "learning_rate": 0.0006,
+ "loss": 4.318901538848877,
+ "step": 1942
+ },
+ {
+ "epoch": 26.992139737991266,
+ "grad_norm": 0.017313100397586823,
+ "learning_rate": 0.0006,
+ "loss": 4.281434535980225,
+ "step": 1943
+ },
+ {
+ "epoch": 27.0,
+ "grad_norm": 0.018450884148478508,
+ "learning_rate": 0.0006,
+ "loss": 4.335104942321777,
+ "step": 1944
+ },
+ {
+ "epoch": 27.0,
+ "eval_loss": 4.640687465667725,
+ "eval_runtime": 56.9624,
+ "eval_samples_per_second": 42.87,
+ "eval_steps_per_second": 1.352,
+ "step": 1944
+ },
+ {
+ "epoch": 27.01397379912664,
+ "grad_norm": 0.017755698412656784,
+ "learning_rate": 0.0006,
+ "loss": 4.278824329376221,
+ "step": 1945
+ },
+ {
+ "epoch": 27.027947598253274,
+ "grad_norm": 0.01724310778081417,
+ "learning_rate": 0.0006,
+ "loss": 4.311049461364746,
+ "step": 1946
+ },
+ {
+ "epoch": 27.041921397379912,
+ "grad_norm": 0.01690123789012432,
+ "learning_rate": 0.0006,
+ "loss": 4.261994361877441,
+ "step": 1947
+ },
+ {
+ "epoch": 27.05589519650655,
+ "grad_norm": 0.018406696617603302,
+ "learning_rate": 0.0006,
+ "loss": 4.1925153732299805,
+ "step": 1948
+ },
+ {
+ "epoch": 27.069868995633186,
+ "grad_norm": 0.019945867359638214,
+ "learning_rate": 0.0006,
+ "loss": 4.108537197113037,
+ "step": 1949
+ },
+ {
+ "epoch": 27.083842794759825,
+ "grad_norm": 0.020316628739237785,
+ "learning_rate": 0.0006,
+ "loss": 4.209043025970459,
+ "step": 1950
+ },
+ {
+ "epoch": 27.097816593886463,
+ "grad_norm": 0.019690370187163353,
+ "learning_rate": 0.0006,
+ "loss": 4.195789337158203,
+ "step": 1951
+ },
+ {
+ "epoch": 27.111790393013102,
+ "grad_norm": 0.01785232499241829,
+ "learning_rate": 0.0006,
+ "loss": 4.21466064453125,
+ "step": 1952
+ },
+ {
+ "epoch": 27.125764192139737,
+ "grad_norm": 0.0164767038077116,
+ "learning_rate": 0.0006,
+ "loss": 4.270247459411621,
+ "step": 1953
+ },
+ {
+ "epoch": 27.139737991266376,
+ "grad_norm": 0.018008455634117126,
+ "learning_rate": 0.0006,
+ "loss": 4.201042175292969,
+ "step": 1954
+ },
+ {
+ "epoch": 27.153711790393015,
+ "grad_norm": 0.01807340793311596,
+ "learning_rate": 0.0006,
+ "loss": 4.25289249420166,
+ "step": 1955
+ },
+ {
+ "epoch": 27.16768558951965,
+ "grad_norm": 0.016246909275650978,
+ "learning_rate": 0.0006,
+ "loss": 4.116283416748047,
+ "step": 1956
+ },
+ {
+ "epoch": 27.18165938864629,
+ "grad_norm": 0.017545776441693306,
+ "learning_rate": 0.0006,
+ "loss": 4.316999435424805,
+ "step": 1957
+ },
+ {
+ "epoch": 27.195633187772927,
+ "grad_norm": 0.017915375530719757,
+ "learning_rate": 0.0006,
+ "loss": 4.253859519958496,
+ "step": 1958
+ },
+ {
+ "epoch": 27.209606986899562,
+ "grad_norm": 0.018112223595380783,
+ "learning_rate": 0.0006,
+ "loss": 4.2541022300720215,
+ "step": 1959
+ },
+ {
+ "epoch": 27.2235807860262,
+ "grad_norm": 0.01846139505505562,
+ "learning_rate": 0.0006,
+ "loss": 4.20438289642334,
+ "step": 1960
+ },
+ {
+ "epoch": 27.23755458515284,
+ "grad_norm": 0.01943897269666195,
+ "learning_rate": 0.0006,
+ "loss": 4.2799577713012695,
+ "step": 1961
+ },
+ {
+ "epoch": 27.251528384279474,
+ "grad_norm": 0.018253348767757416,
+ "learning_rate": 0.0006,
+ "loss": 4.1720499992370605,
+ "step": 1962
+ },
+ {
+ "epoch": 27.265502183406113,
+ "grad_norm": 0.018292246386408806,
+ "learning_rate": 0.0006,
+ "loss": 4.28231143951416,
+ "step": 1963
+ },
+ {
+ "epoch": 27.27947598253275,
+ "grad_norm": 0.01943155936896801,
+ "learning_rate": 0.0006,
+ "loss": 4.314976692199707,
+ "step": 1964
+ },
+ {
+ "epoch": 27.29344978165939,
+ "grad_norm": 0.019064994528889656,
+ "learning_rate": 0.0006,
+ "loss": 4.25328254699707,
+ "step": 1965
+ },
+ {
+ "epoch": 27.307423580786025,
+ "grad_norm": 0.017423273995518684,
+ "learning_rate": 0.0006,
+ "loss": 4.195992469787598,
+ "step": 1966
+ },
+ {
+ "epoch": 27.321397379912664,
+ "grad_norm": 0.01677924580872059,
+ "learning_rate": 0.0006,
+ "loss": 4.260043144226074,
+ "step": 1967
+ },
+ {
+ "epoch": 27.335371179039303,
+ "grad_norm": 0.01720673404633999,
+ "learning_rate": 0.0006,
+ "loss": 4.222780704498291,
+ "step": 1968
+ },
+ {
+ "epoch": 27.349344978165938,
+ "grad_norm": 0.017300743609666824,
+ "learning_rate": 0.0006,
+ "loss": 4.109524250030518,
+ "step": 1969
+ },
+ {
+ "epoch": 27.363318777292577,
+ "grad_norm": 0.016065871343016624,
+ "learning_rate": 0.0006,
+ "loss": 4.36108922958374,
+ "step": 1970
+ },
+ {
+ "epoch": 27.377292576419215,
+ "grad_norm": 0.016626616939902306,
+ "learning_rate": 0.0006,
+ "loss": 4.220630645751953,
+ "step": 1971
+ },
+ {
+ "epoch": 27.39126637554585,
+ "grad_norm": 0.016482602804899216,
+ "learning_rate": 0.0006,
+ "loss": 4.344797134399414,
+ "step": 1972
+ },
+ {
+ "epoch": 27.40524017467249,
+ "grad_norm": 0.01727953553199768,
+ "learning_rate": 0.0006,
+ "loss": 4.296645164489746,
+ "step": 1973
+ },
+ {
+ "epoch": 27.419213973799128,
+ "grad_norm": 0.01522731315344572,
+ "learning_rate": 0.0006,
+ "loss": 4.223751068115234,
+ "step": 1974
+ },
+ {
+ "epoch": 27.433187772925763,
+ "grad_norm": 0.01581776700913906,
+ "learning_rate": 0.0006,
+ "loss": 4.351379871368408,
+ "step": 1975
+ },
+ {
+ "epoch": 27.4471615720524,
+ "grad_norm": 0.016394605860114098,
+ "learning_rate": 0.0006,
+ "loss": 4.2412919998168945,
+ "step": 1976
+ },
+ {
+ "epoch": 27.46113537117904,
+ "grad_norm": 0.01751169003546238,
+ "learning_rate": 0.0006,
+ "loss": 4.257606506347656,
+ "step": 1977
+ },
+ {
+ "epoch": 27.475109170305675,
+ "grad_norm": 0.017452310770750046,
+ "learning_rate": 0.0006,
+ "loss": 4.380134582519531,
+ "step": 1978
+ },
+ {
+ "epoch": 27.489082969432314,
+ "grad_norm": 0.017741898074746132,
+ "learning_rate": 0.0006,
+ "loss": 4.279632568359375,
+ "step": 1979
+ },
+ {
+ "epoch": 27.503056768558952,
+ "grad_norm": 0.017680590972304344,
+ "learning_rate": 0.0006,
+ "loss": 4.273406028747559,
+ "step": 1980
+ },
+ {
+ "epoch": 27.51703056768559,
+ "grad_norm": 0.01864718459546566,
+ "learning_rate": 0.0006,
+ "loss": 4.278583526611328,
+ "step": 1981
+ },
+ {
+ "epoch": 27.531004366812226,
+ "grad_norm": 0.019558541476726532,
+ "learning_rate": 0.0006,
+ "loss": 4.211919784545898,
+ "step": 1982
+ },
+ {
+ "epoch": 27.544978165938865,
+ "grad_norm": 0.019467400386929512,
+ "learning_rate": 0.0006,
+ "loss": 4.276822566986084,
+ "step": 1983
+ },
+ {
+ "epoch": 27.558951965065503,
+ "grad_norm": 0.01756172813475132,
+ "learning_rate": 0.0006,
+ "loss": 4.165379047393799,
+ "step": 1984
+ },
+ {
+ "epoch": 27.57292576419214,
+ "grad_norm": 0.017458263784646988,
+ "learning_rate": 0.0006,
+ "loss": 4.2992401123046875,
+ "step": 1985
+ },
+ {
+ "epoch": 27.586899563318777,
+ "grad_norm": 0.018089665099978447,
+ "learning_rate": 0.0006,
+ "loss": 4.243378639221191,
+ "step": 1986
+ },
+ {
+ "epoch": 27.600873362445416,
+ "grad_norm": 0.019535524770617485,
+ "learning_rate": 0.0006,
+ "loss": 4.321477890014648,
+ "step": 1987
+ },
+ {
+ "epoch": 27.61484716157205,
+ "grad_norm": 0.01997970975935459,
+ "learning_rate": 0.0006,
+ "loss": 4.123082637786865,
+ "step": 1988
+ },
+ {
+ "epoch": 27.62882096069869,
+ "grad_norm": 0.01913067139685154,
+ "learning_rate": 0.0006,
+ "loss": 4.289896488189697,
+ "step": 1989
+ },
+ {
+ "epoch": 27.64279475982533,
+ "grad_norm": 0.0169806070625782,
+ "learning_rate": 0.0006,
+ "loss": 4.247004985809326,
+ "step": 1990
+ },
+ {
+ "epoch": 27.656768558951963,
+ "grad_norm": 0.017315007746219635,
+ "learning_rate": 0.0006,
+ "loss": 4.2095441818237305,
+ "step": 1991
+ },
+ {
+ "epoch": 27.670742358078602,
+ "grad_norm": 0.015098759904503822,
+ "learning_rate": 0.0006,
+ "loss": 4.145560264587402,
+ "step": 1992
+ },
+ {
+ "epoch": 27.68471615720524,
+ "grad_norm": 0.016240514814853668,
+ "learning_rate": 0.0006,
+ "loss": 4.254927635192871,
+ "step": 1993
+ },
+ {
+ "epoch": 27.69868995633188,
+ "grad_norm": 0.01621108129620552,
+ "learning_rate": 0.0006,
+ "loss": 4.219725608825684,
+ "step": 1994
+ },
+ {
+ "epoch": 27.712663755458514,
+ "grad_norm": 0.015191009268164635,
+ "learning_rate": 0.0006,
+ "loss": 4.194345474243164,
+ "step": 1995
+ },
+ {
+ "epoch": 27.726637554585153,
+ "grad_norm": 0.014387983828783035,
+ "learning_rate": 0.0006,
+ "loss": 4.211467742919922,
+ "step": 1996
+ },
+ {
+ "epoch": 27.74061135371179,
+ "grad_norm": 0.015460561029613018,
+ "learning_rate": 0.0006,
+ "loss": 4.358700752258301,
+ "step": 1997
+ },
+ {
+ "epoch": 27.754585152838427,
+ "grad_norm": 0.015100076794624329,
+ "learning_rate": 0.0006,
+ "loss": 4.171223163604736,
+ "step": 1998
+ },
+ {
+ "epoch": 27.768558951965066,
+ "grad_norm": 0.015607891604304314,
+ "learning_rate": 0.0006,
+ "loss": 4.217874526977539,
+ "step": 1999
+ },
+ {
+ "epoch": 27.782532751091704,
+ "grad_norm": 0.015650689601898193,
+ "learning_rate": 0.0006,
+ "loss": 4.280780792236328,
+ "step": 2000
+ },
+ {
+ "epoch": 27.79650655021834,
+ "grad_norm": 0.01580994389951229,
+ "learning_rate": 0.0006,
+ "loss": 4.159212112426758,
+ "step": 2001
+ },
+ {
+ "epoch": 27.810480349344978,
+ "grad_norm": 0.016607413068413734,
+ "learning_rate": 0.0006,
+ "loss": 4.238853454589844,
+ "step": 2002
+ },
+ {
+ "epoch": 27.824454148471617,
+ "grad_norm": 0.017383860424160957,
+ "learning_rate": 0.0006,
+ "loss": 4.233944892883301,
+ "step": 2003
+ },
+ {
+ "epoch": 27.83842794759825,
+ "grad_norm": 0.014966659247875214,
+ "learning_rate": 0.0006,
+ "loss": 4.167392730712891,
+ "step": 2004
+ },
+ {
+ "epoch": 27.85240174672489,
+ "grad_norm": 0.014589878730475903,
+ "learning_rate": 0.0006,
+ "loss": 4.2390899658203125,
+ "step": 2005
+ },
+ {
+ "epoch": 27.86637554585153,
+ "grad_norm": 0.017179450020194054,
+ "learning_rate": 0.0006,
+ "loss": 4.365785598754883,
+ "step": 2006
+ },
+ {
+ "epoch": 27.880349344978168,
+ "grad_norm": 0.01893901452422142,
+ "learning_rate": 0.0006,
+ "loss": 4.289430141448975,
+ "step": 2007
+ },
+ {
+ "epoch": 27.894323144104803,
+ "grad_norm": 0.018619712442159653,
+ "learning_rate": 0.0006,
+ "loss": 4.310665130615234,
+ "step": 2008
+ },
+ {
+ "epoch": 27.90829694323144,
+ "grad_norm": 0.019180385395884514,
+ "learning_rate": 0.0006,
+ "loss": 4.27454137802124,
+ "step": 2009
+ },
+ {
+ "epoch": 27.92227074235808,
+ "grad_norm": 0.017811523750424385,
+ "learning_rate": 0.0006,
+ "loss": 4.312562465667725,
+ "step": 2010
+ },
+ {
+ "epoch": 27.936244541484715,
+ "grad_norm": 0.016583573073148727,
+ "learning_rate": 0.0006,
+ "loss": 4.202253341674805,
+ "step": 2011
+ },
+ {
+ "epoch": 27.950218340611354,
+ "grad_norm": 0.01658615842461586,
+ "learning_rate": 0.0006,
+ "loss": 4.183433532714844,
+ "step": 2012
+ },
+ {
+ "epoch": 27.964192139737992,
+ "grad_norm": 0.015327401459217072,
+ "learning_rate": 0.0006,
+ "loss": 4.2125139236450195,
+ "step": 2013
+ },
+ {
+ "epoch": 27.978165938864628,
+ "grad_norm": 0.015004601329565048,
+ "learning_rate": 0.0006,
+ "loss": 4.257392883300781,
+ "step": 2014
+ },
+ {
+ "epoch": 27.992139737991266,
+ "grad_norm": 0.015259161591529846,
+ "learning_rate": 0.0006,
+ "loss": 4.241818428039551,
+ "step": 2015
+ },
+ {
+ "epoch": 28.0,
+ "grad_norm": 0.016808776184916496,
+ "learning_rate": 0.0006,
+ "loss": 4.146241188049316,
+ "step": 2016
+ },
+ {
+ "epoch": 28.0,
+ "eval_loss": 4.663090705871582,
+ "eval_runtime": 56.4559,
+ "eval_samples_per_second": 43.255,
+ "eval_steps_per_second": 1.364,
+ "step": 2016
+ },
+ {
+ "epoch": 28.01397379912664,
+ "grad_norm": 0.015784382820129395,
+ "learning_rate": 0.0006,
+ "loss": 4.044191360473633,
+ "step": 2017
+ },
+ {
+ "epoch": 28.027947598253274,
+ "grad_norm": 0.015636710450053215,
+ "learning_rate": 0.0006,
+ "loss": 4.128754615783691,
+ "step": 2018
+ },
+ {
+ "epoch": 28.041921397379912,
+ "grad_norm": 0.015364975668489933,
+ "learning_rate": 0.0006,
+ "loss": 4.191835403442383,
+ "step": 2019
+ },
+ {
+ "epoch": 28.05589519650655,
+ "grad_norm": 0.015015700832009315,
+ "learning_rate": 0.0006,
+ "loss": 4.120022773742676,
+ "step": 2020
+ },
+ {
+ "epoch": 28.069868995633186,
+ "grad_norm": 0.016666799783706665,
+ "learning_rate": 0.0006,
+ "loss": 4.33906364440918,
+ "step": 2021
+ },
+ {
+ "epoch": 28.083842794759825,
+ "grad_norm": 0.01780104823410511,
+ "learning_rate": 0.0006,
+ "loss": 4.279458999633789,
+ "step": 2022
+ },
+ {
+ "epoch": 28.097816593886463,
+ "grad_norm": 0.01870882883667946,
+ "learning_rate": 0.0006,
+ "loss": 4.34728479385376,
+ "step": 2023
+ },
+ {
+ "epoch": 28.111790393013102,
+ "grad_norm": 0.018406381830573082,
+ "learning_rate": 0.0006,
+ "loss": 4.211311340332031,
+ "step": 2024
+ },
+ {
+ "epoch": 28.125764192139737,
+ "grad_norm": 0.017833448946475983,
+ "learning_rate": 0.0006,
+ "loss": 4.223254203796387,
+ "step": 2025
+ },
+ {
+ "epoch": 28.139737991266376,
+ "grad_norm": 0.017552688717842102,
+ "learning_rate": 0.0006,
+ "loss": 4.208823204040527,
+ "step": 2026
+ },
+ {
+ "epoch": 28.153711790393015,
+ "grad_norm": 0.018500229343771935,
+ "learning_rate": 0.0006,
+ "loss": 4.198025226593018,
+ "step": 2027
+ },
+ {
+ "epoch": 28.16768558951965,
+ "grad_norm": 0.01782156340777874,
+ "learning_rate": 0.0006,
+ "loss": 4.179978370666504,
+ "step": 2028
+ },
+ {
+ "epoch": 28.18165938864629,
+ "grad_norm": 0.019580967724323273,
+ "learning_rate": 0.0006,
+ "loss": 4.291114807128906,
+ "step": 2029
+ },
+ {
+ "epoch": 28.195633187772927,
+ "grad_norm": 0.020804036408662796,
+ "learning_rate": 0.0006,
+ "loss": 4.220607280731201,
+ "step": 2030
+ },
+ {
+ "epoch": 28.209606986899562,
+ "grad_norm": 0.019595693796873093,
+ "learning_rate": 0.0006,
+ "loss": 4.2905731201171875,
+ "step": 2031
+ },
+ {
+ "epoch": 28.2235807860262,
+ "grad_norm": 0.019247086718678474,
+ "learning_rate": 0.0006,
+ "loss": 4.2467360496521,
+ "step": 2032
+ },
+ {
+ "epoch": 28.23755458515284,
+ "grad_norm": 0.01958036608994007,
+ "learning_rate": 0.0006,
+ "loss": 4.286137580871582,
+ "step": 2033
+ },
+ {
+ "epoch": 28.251528384279474,
+ "grad_norm": 0.021354753524065018,
+ "learning_rate": 0.0006,
+ "loss": 4.234755039215088,
+ "step": 2034
+ },
+ {
+ "epoch": 28.265502183406113,
+ "grad_norm": 0.02206375077366829,
+ "learning_rate": 0.0006,
+ "loss": 4.149641036987305,
+ "step": 2035
+ },
+ {
+ "epoch": 28.27947598253275,
+ "grad_norm": 0.02201310358941555,
+ "learning_rate": 0.0006,
+ "loss": 4.1928629875183105,
+ "step": 2036
+ },
+ {
+ "epoch": 28.29344978165939,
+ "grad_norm": 0.01961950585246086,
+ "learning_rate": 0.0006,
+ "loss": 4.211404323577881,
+ "step": 2037
+ },
+ {
+ "epoch": 28.307423580786025,
+ "grad_norm": 0.019211795181035995,
+ "learning_rate": 0.0006,
+ "loss": 4.200976848602295,
+ "step": 2038
+ },
+ {
+ "epoch": 28.321397379912664,
+ "grad_norm": 0.01779370754957199,
+ "learning_rate": 0.0006,
+ "loss": 4.249148845672607,
+ "step": 2039
+ },
+ {
+ "epoch": 28.335371179039303,
+ "grad_norm": 0.018635908141732216,
+ "learning_rate": 0.0006,
+ "loss": 4.11806058883667,
+ "step": 2040
+ },
+ {
+ "epoch": 28.349344978165938,
+ "grad_norm": 0.019792694598436356,
+ "learning_rate": 0.0006,
+ "loss": 4.335224151611328,
+ "step": 2041
+ },
+ {
+ "epoch": 28.363318777292577,
+ "grad_norm": 0.021731717512011528,
+ "learning_rate": 0.0006,
+ "loss": 4.302778244018555,
+ "step": 2042
+ },
+ {
+ "epoch": 28.377292576419215,
+ "grad_norm": 0.020823447033762932,
+ "learning_rate": 0.0006,
+ "loss": 4.260610580444336,
+ "step": 2043
+ },
+ {
+ "epoch": 28.39126637554585,
+ "grad_norm": 0.019385412335395813,
+ "learning_rate": 0.0006,
+ "loss": 4.2247724533081055,
+ "step": 2044
+ },
+ {
+ "epoch": 28.40524017467249,
+ "grad_norm": 0.019539108499884605,
+ "learning_rate": 0.0006,
+ "loss": 4.236655235290527,
+ "step": 2045
+ },
+ {
+ "epoch": 28.419213973799128,
+ "grad_norm": 0.01942690648138523,
+ "learning_rate": 0.0006,
+ "loss": 4.251850605010986,
+ "step": 2046
+ },
+ {
+ "epoch": 28.433187772925763,
+ "grad_norm": 0.019667886197566986,
+ "learning_rate": 0.0006,
+ "loss": 4.222312927246094,
+ "step": 2047
+ },
+ {
+ "epoch": 28.4471615720524,
+ "grad_norm": 0.02029012329876423,
+ "learning_rate": 0.0006,
+ "loss": 4.246252059936523,
+ "step": 2048
+ },
+ {
+ "epoch": 28.46113537117904,
+ "grad_norm": 0.01784469373524189,
+ "learning_rate": 0.0006,
+ "loss": 4.192128658294678,
+ "step": 2049
+ },
+ {
+ "epoch": 28.475109170305675,
+ "grad_norm": 0.0160287544131279,
+ "learning_rate": 0.0006,
+ "loss": 4.208607196807861,
+ "step": 2050
+ },
+ {
+ "epoch": 28.489082969432314,
+ "grad_norm": 0.015348346903920174,
+ "learning_rate": 0.0006,
+ "loss": 4.236598014831543,
+ "step": 2051
+ },
+ {
+ "epoch": 28.503056768558952,
+ "grad_norm": 0.015198206529021263,
+ "learning_rate": 0.0006,
+ "loss": 4.235552787780762,
+ "step": 2052
+ },
+ {
+ "epoch": 28.51703056768559,
+ "grad_norm": 0.01523754745721817,
+ "learning_rate": 0.0006,
+ "loss": 4.1941728591918945,
+ "step": 2053
+ },
+ {
+ "epoch": 28.531004366812226,
+ "grad_norm": 0.015559614636003971,
+ "learning_rate": 0.0006,
+ "loss": 4.294757843017578,
+ "step": 2054
+ },
+ {
+ "epoch": 28.544978165938865,
+ "grad_norm": 0.01568358950316906,
+ "learning_rate": 0.0006,
+ "loss": 4.213065147399902,
+ "step": 2055
+ },
+ {
+ "epoch": 28.558951965065503,
+ "grad_norm": 0.015336516313254833,
+ "learning_rate": 0.0006,
+ "loss": 4.240293979644775,
+ "step": 2056
+ },
+ {
+ "epoch": 28.57292576419214,
+ "grad_norm": 0.015041496604681015,
+ "learning_rate": 0.0006,
+ "loss": 4.273397445678711,
+ "step": 2057
+ },
+ {
+ "epoch": 28.586899563318777,
+ "grad_norm": 0.01610128954052925,
+ "learning_rate": 0.0006,
+ "loss": 4.125369071960449,
+ "step": 2058
+ },
+ {
+ "epoch": 28.600873362445416,
+ "grad_norm": 0.015974976122379303,
+ "learning_rate": 0.0006,
+ "loss": 4.133459091186523,
+ "step": 2059
+ },
+ {
+ "epoch": 28.61484716157205,
+ "grad_norm": 0.015936799347400665,
+ "learning_rate": 0.0006,
+ "loss": 4.306667327880859,
+ "step": 2060
+ },
+ {
+ "epoch": 28.62882096069869,
+ "grad_norm": 0.01795247197151184,
+ "learning_rate": 0.0006,
+ "loss": 4.205156326293945,
+ "step": 2061
+ },
+ {
+ "epoch": 28.64279475982533,
+ "grad_norm": 0.019639814272522926,
+ "learning_rate": 0.0006,
+ "loss": 4.277614593505859,
+ "step": 2062
+ },
+ {
+ "epoch": 28.656768558951963,
+ "grad_norm": 0.01722300611436367,
+ "learning_rate": 0.0006,
+ "loss": 4.191032886505127,
+ "step": 2063
+ },
+ {
+ "epoch": 28.670742358078602,
+ "grad_norm": 0.01612633652985096,
+ "learning_rate": 0.0006,
+ "loss": 4.20717191696167,
+ "step": 2064
+ },
+ {
+ "epoch": 28.68471615720524,
+ "grad_norm": 0.01603279449045658,
+ "learning_rate": 0.0006,
+ "loss": 4.228884220123291,
+ "step": 2065
+ },
+ {
+ "epoch": 28.69868995633188,
+ "grad_norm": 0.016405927017331123,
+ "learning_rate": 0.0006,
+ "loss": 4.201292991638184,
+ "step": 2066
+ },
+ {
+ "epoch": 28.712663755458514,
+ "grad_norm": 0.016644183546304703,
+ "learning_rate": 0.0006,
+ "loss": 4.239282131195068,
+ "step": 2067
+ },
+ {
+ "epoch": 28.726637554585153,
+ "grad_norm": 0.0175587497651577,
+ "learning_rate": 0.0006,
+ "loss": 4.289038181304932,
+ "step": 2068
+ },
+ {
+ "epoch": 28.74061135371179,
+ "grad_norm": 0.018223397433757782,
+ "learning_rate": 0.0006,
+ "loss": 4.2266998291015625,
+ "step": 2069
+ },
+ {
+ "epoch": 28.754585152838427,
+ "grad_norm": 0.018680868670344353,
+ "learning_rate": 0.0006,
+ "loss": 4.348849296569824,
+ "step": 2070
+ },
+ {
+ "epoch": 28.768558951965066,
+ "grad_norm": 0.016726728528738022,
+ "learning_rate": 0.0006,
+ "loss": 4.292080879211426,
+ "step": 2071
+ },
+ {
+ "epoch": 28.782532751091704,
+ "grad_norm": 0.017386795952916145,
+ "learning_rate": 0.0006,
+ "loss": 4.208870887756348,
+ "step": 2072
+ },
+ {
+ "epoch": 28.79650655021834,
+ "grad_norm": 0.01865178905427456,
+ "learning_rate": 0.0006,
+ "loss": 4.054150581359863,
+ "step": 2073
+ },
+ {
+ "epoch": 28.810480349344978,
+ "grad_norm": 0.01776815392076969,
+ "learning_rate": 0.0006,
+ "loss": 4.202404499053955,
+ "step": 2074
+ },
+ {
+ "epoch": 28.824454148471617,
+ "grad_norm": 0.01643497310578823,
+ "learning_rate": 0.0006,
+ "loss": 4.131855010986328,
+ "step": 2075
+ },
+ {
+ "epoch": 28.83842794759825,
+ "grad_norm": 0.01574273779988289,
+ "learning_rate": 0.0006,
+ "loss": 4.296788692474365,
+ "step": 2076
+ },
+ {
+ "epoch": 28.85240174672489,
+ "grad_norm": 0.015613908879458904,
+ "learning_rate": 0.0006,
+ "loss": 4.210949897766113,
+ "step": 2077
+ },
+ {
+ "epoch": 28.86637554585153,
+ "grad_norm": 0.015543212182819843,
+ "learning_rate": 0.0006,
+ "loss": 4.218136310577393,
+ "step": 2078
+ },
+ {
+ "epoch": 28.880349344978168,
+ "grad_norm": 0.014681251719594002,
+ "learning_rate": 0.0006,
+ "loss": 4.19401741027832,
+ "step": 2079
+ },
+ {
+ "epoch": 28.894323144104803,
+ "grad_norm": 0.014393828809261322,
+ "learning_rate": 0.0006,
+ "loss": 4.198468208312988,
+ "step": 2080
+ },
+ {
+ "epoch": 28.90829694323144,
+ "grad_norm": 0.01580170728266239,
+ "learning_rate": 0.0006,
+ "loss": 4.224969863891602,
+ "step": 2081
+ },
+ {
+ "epoch": 28.92227074235808,
+ "grad_norm": 0.01484165620058775,
+ "learning_rate": 0.0006,
+ "loss": 4.12236213684082,
+ "step": 2082
+ },
+ {
+ "epoch": 28.936244541484715,
+ "grad_norm": 0.01497623696923256,
+ "learning_rate": 0.0006,
+ "loss": 4.24141788482666,
+ "step": 2083
+ },
+ {
+ "epoch": 28.950218340611354,
+ "grad_norm": 0.015080046840012074,
+ "learning_rate": 0.0006,
+ "loss": 4.284151077270508,
+ "step": 2084
+ },
+ {
+ "epoch": 28.964192139737992,
+ "grad_norm": 0.016207212582230568,
+ "learning_rate": 0.0006,
+ "loss": 4.298693656921387,
+ "step": 2085
+ },
+ {
+ "epoch": 28.978165938864628,
+ "grad_norm": 0.015474417246878147,
+ "learning_rate": 0.0006,
+ "loss": 4.2412261962890625,
+ "step": 2086
+ },
+ {
+ "epoch": 28.992139737991266,
+ "grad_norm": 0.01515351701527834,
+ "learning_rate": 0.0006,
+ "loss": 4.3054423332214355,
+ "step": 2087
+ },
+ {
+ "epoch": 29.0,
+ "grad_norm": 0.016411345452070236,
+ "learning_rate": 0.0006,
+ "loss": 4.104830741882324,
+ "step": 2088
+ },
+ {
+ "epoch": 29.0,
+ "eval_loss": 4.661567687988281,
+ "eval_runtime": 57.1244,
+ "eval_samples_per_second": 42.749,
+ "eval_steps_per_second": 1.348,
+ "step": 2088
+ },
+ {
+ "epoch": 29.01397379912664,
+ "grad_norm": 0.01665342226624489,
+ "learning_rate": 0.0006,
+ "loss": 4.311391830444336,
+ "step": 2089
+ },
+ {
+ "epoch": 29.027947598253274,
+ "grad_norm": 0.018053214997053146,
+ "learning_rate": 0.0006,
+ "loss": 4.135597229003906,
+ "step": 2090
+ },
+ {
+ "epoch": 29.041921397379912,
+ "grad_norm": 0.018213748931884766,
+ "learning_rate": 0.0006,
+ "loss": 4.303299903869629,
+ "step": 2091
+ },
+ {
+ "epoch": 29.05589519650655,
+ "grad_norm": 0.018460242077708244,
+ "learning_rate": 0.0006,
+ "loss": 4.20850944519043,
+ "step": 2092
+ },
+ {
+ "epoch": 29.069868995633186,
+ "grad_norm": 0.018365809693932533,
+ "learning_rate": 0.0006,
+ "loss": 4.275847434997559,
+ "step": 2093
+ },
+ {
+ "epoch": 29.083842794759825,
+ "grad_norm": 0.018346186727285385,
+ "learning_rate": 0.0006,
+ "loss": 4.202091217041016,
+ "step": 2094
+ },
+ {
+ "epoch": 29.097816593886463,
+ "grad_norm": 0.017920587211847305,
+ "learning_rate": 0.0006,
+ "loss": 4.182092666625977,
+ "step": 2095
+ },
+ {
+ "epoch": 29.111790393013102,
+ "grad_norm": 0.01812003180384636,
+ "learning_rate": 0.0006,
+ "loss": 4.259721755981445,
+ "step": 2096
+ },
+ {
+ "epoch": 29.125764192139737,
+ "grad_norm": 0.01700986735522747,
+ "learning_rate": 0.0006,
+ "loss": 4.067573070526123,
+ "step": 2097
+ },
+ {
+ "epoch": 29.139737991266376,
+ "grad_norm": 0.018069760873913765,
+ "learning_rate": 0.0006,
+ "loss": 4.234400749206543,
+ "step": 2098
+ },
+ {
+ "epoch": 29.153711790393015,
+ "grad_norm": 0.01905428245663643,
+ "learning_rate": 0.0006,
+ "loss": 4.16609001159668,
+ "step": 2099
+ },
+ {
+ "epoch": 29.16768558951965,
+ "grad_norm": 0.018711242824792862,
+ "learning_rate": 0.0006,
+ "loss": 4.249088764190674,
+ "step": 2100
+ },
+ {
+ "epoch": 29.18165938864629,
+ "grad_norm": 0.018721306696534157,
+ "learning_rate": 0.0006,
+ "loss": 4.14540958404541,
+ "step": 2101
+ },
+ {
+ "epoch": 29.195633187772927,
+ "grad_norm": 0.020139874890446663,
+ "learning_rate": 0.0006,
+ "loss": 4.049248695373535,
+ "step": 2102
+ },
+ {
+ "epoch": 29.209606986899562,
+ "grad_norm": 0.020849574357271194,
+ "learning_rate": 0.0006,
+ "loss": 4.235224723815918,
+ "step": 2103
+ },
+ {
+ "epoch": 29.2235807860262,
+ "grad_norm": 0.018693549558520317,
+ "learning_rate": 0.0006,
+ "loss": 4.232961177825928,
+ "step": 2104
+ },
+ {
+ "epoch": 29.23755458515284,
+ "grad_norm": 0.017889857292175293,
+ "learning_rate": 0.0006,
+ "loss": 4.171422004699707,
+ "step": 2105
+ },
+ {
+ "epoch": 29.251528384279474,
+ "grad_norm": 0.016690224409103394,
+ "learning_rate": 0.0006,
+ "loss": 4.244717597961426,
+ "step": 2106
+ },
+ {
+ "epoch": 29.265502183406113,
+ "grad_norm": 0.01708616502583027,
+ "learning_rate": 0.0006,
+ "loss": 4.125068664550781,
+ "step": 2107
+ },
+ {
+ "epoch": 29.27947598253275,
+ "grad_norm": 0.017928237095475197,
+ "learning_rate": 0.0006,
+ "loss": 4.208292007446289,
+ "step": 2108
+ },
+ {
+ "epoch": 29.29344978165939,
+ "grad_norm": 0.018250394612550735,
+ "learning_rate": 0.0006,
+ "loss": 4.118630886077881,
+ "step": 2109
+ },
+ {
+ "epoch": 29.307423580786025,
+ "grad_norm": 0.020160246640443802,
+ "learning_rate": 0.0006,
+ "loss": 4.206025123596191,
+ "step": 2110
+ },
+ {
+ "epoch": 29.321397379912664,
+ "grad_norm": 0.020045453682541847,
+ "learning_rate": 0.0006,
+ "loss": 4.2758684158325195,
+ "step": 2111
+ },
+ {
+ "epoch": 29.335371179039303,
+ "grad_norm": 0.019556907936930656,
+ "learning_rate": 0.0006,
+ "loss": 4.188272953033447,
+ "step": 2112
+ },
+ {
+ "epoch": 29.349344978165938,
+ "grad_norm": 0.019185200333595276,
+ "learning_rate": 0.0006,
+ "loss": 4.17451286315918,
+ "step": 2113
+ },
+ {
+ "epoch": 29.363318777292577,
+ "grad_norm": 0.01816629432141781,
+ "learning_rate": 0.0006,
+ "loss": 4.165246963500977,
+ "step": 2114
+ },
+ {
+ "epoch": 29.377292576419215,
+ "grad_norm": 0.01865146867930889,
+ "learning_rate": 0.0006,
+ "loss": 4.148184776306152,
+ "step": 2115
+ },
+ {
+ "epoch": 29.39126637554585,
+ "grad_norm": 0.018338464200496674,
+ "learning_rate": 0.0006,
+ "loss": 4.288936614990234,
+ "step": 2116
+ },
+ {
+ "epoch": 29.40524017467249,
+ "grad_norm": 0.018986187875270844,
+ "learning_rate": 0.0006,
+ "loss": 4.183379650115967,
+ "step": 2117
+ },
+ {
+ "epoch": 29.419213973799128,
+ "grad_norm": 0.018409637734293938,
+ "learning_rate": 0.0006,
+ "loss": 4.057574272155762,
+ "step": 2118
+ },
+ {
+ "epoch": 29.433187772925763,
+ "grad_norm": 0.016664542257785797,
+ "learning_rate": 0.0006,
+ "loss": 4.18517541885376,
+ "step": 2119
+ },
+ {
+ "epoch": 29.4471615720524,
+ "grad_norm": 0.016486089676618576,
+ "learning_rate": 0.0006,
+ "loss": 4.179367542266846,
+ "step": 2120
+ },
+ {
+ "epoch": 29.46113537117904,
+ "grad_norm": 0.01787523552775383,
+ "learning_rate": 0.0006,
+ "loss": 4.184126377105713,
+ "step": 2121
+ },
+ {
+ "epoch": 29.475109170305675,
+ "grad_norm": 0.016574613749980927,
+ "learning_rate": 0.0006,
+ "loss": 4.290170192718506,
+ "step": 2122
+ },
+ {
+ "epoch": 29.489082969432314,
+ "grad_norm": 0.01643376611173153,
+ "learning_rate": 0.0006,
+ "loss": 4.052913665771484,
+ "step": 2123
+ },
+ {
+ "epoch": 29.503056768558952,
+ "grad_norm": 0.016503969207406044,
+ "learning_rate": 0.0006,
+ "loss": 4.335184097290039,
+ "step": 2124
+ },
+ {
+ "epoch": 29.51703056768559,
+ "grad_norm": 0.017596479505300522,
+ "learning_rate": 0.0006,
+ "loss": 4.165539741516113,
+ "step": 2125
+ },
+ {
+ "epoch": 29.531004366812226,
+ "grad_norm": 0.017468582838773727,
+ "learning_rate": 0.0006,
+ "loss": 4.315964698791504,
+ "step": 2126
+ },
+ {
+ "epoch": 29.544978165938865,
+ "grad_norm": 0.01724618673324585,
+ "learning_rate": 0.0006,
+ "loss": 4.243681907653809,
+ "step": 2127
+ },
+ {
+ "epoch": 29.558951965065503,
+ "grad_norm": 0.018384616822004318,
+ "learning_rate": 0.0006,
+ "loss": 4.112447261810303,
+ "step": 2128
+ },
+ {
+ "epoch": 29.57292576419214,
+ "grad_norm": 0.01824074238538742,
+ "learning_rate": 0.0006,
+ "loss": 4.236065864562988,
+ "step": 2129
+ },
+ {
+ "epoch": 29.586899563318777,
+ "grad_norm": 0.01625971868634224,
+ "learning_rate": 0.0006,
+ "loss": 4.18143892288208,
+ "step": 2130
+ },
+ {
+ "epoch": 29.600873362445416,
+ "grad_norm": 0.015537494793534279,
+ "learning_rate": 0.0006,
+ "loss": 4.225147247314453,
+ "step": 2131
+ },
+ {
+ "epoch": 29.61484716157205,
+ "grad_norm": 0.015818974003195763,
+ "learning_rate": 0.0006,
+ "loss": 4.3100175857543945,
+ "step": 2132
+ },
+ {
+ "epoch": 29.62882096069869,
+ "grad_norm": 0.015909671783447266,
+ "learning_rate": 0.0006,
+ "loss": 4.169775009155273,
+ "step": 2133
+ },
+ {
+ "epoch": 29.64279475982533,
+ "grad_norm": 0.01769658550620079,
+ "learning_rate": 0.0006,
+ "loss": 4.278904914855957,
+ "step": 2134
+ },
+ {
+ "epoch": 29.656768558951963,
+ "grad_norm": 0.015641704201698303,
+ "learning_rate": 0.0006,
+ "loss": 4.2426228523254395,
+ "step": 2135
+ },
+ {
+ "epoch": 29.670742358078602,
+ "grad_norm": 0.01565658301115036,
+ "learning_rate": 0.0006,
+ "loss": 4.232687950134277,
+ "step": 2136
+ },
+ {
+ "epoch": 29.68471615720524,
+ "grad_norm": 0.016485940665006638,
+ "learning_rate": 0.0006,
+ "loss": 4.209839820861816,
+ "step": 2137
+ },
+ {
+ "epoch": 29.69868995633188,
+ "grad_norm": 0.017622729763388634,
+ "learning_rate": 0.0006,
+ "loss": 4.196017265319824,
+ "step": 2138
+ },
+ {
+ "epoch": 29.712663755458514,
+ "grad_norm": 0.018320564180612564,
+ "learning_rate": 0.0006,
+ "loss": 4.225683689117432,
+ "step": 2139
+ },
+ {
+ "epoch": 29.726637554585153,
+ "grad_norm": 0.018140794709324837,
+ "learning_rate": 0.0006,
+ "loss": 4.311944961547852,
+ "step": 2140
+ },
+ {
+ "epoch": 29.74061135371179,
+ "grad_norm": 0.016227100044488907,
+ "learning_rate": 0.0006,
+ "loss": 4.230257511138916,
+ "step": 2141
+ },
+ {
+ "epoch": 29.754585152838427,
+ "grad_norm": 0.015759488567709923,
+ "learning_rate": 0.0006,
+ "loss": 4.119174003601074,
+ "step": 2142
+ },
+ {
+ "epoch": 29.768558951965066,
+ "grad_norm": 0.01757766678929329,
+ "learning_rate": 0.0006,
+ "loss": 4.165826797485352,
+ "step": 2143
+ },
+ {
+ "epoch": 29.782532751091704,
+ "grad_norm": 0.019661923870444298,
+ "learning_rate": 0.0006,
+ "loss": 4.392203330993652,
+ "step": 2144
+ },
+ {
+ "epoch": 29.79650655021834,
+ "grad_norm": 0.019478755071759224,
+ "learning_rate": 0.0006,
+ "loss": 4.348371982574463,
+ "step": 2145
+ },
+ {
+ "epoch": 29.810480349344978,
+ "grad_norm": 0.019149569794535637,
+ "learning_rate": 0.0006,
+ "loss": 4.34104061126709,
+ "step": 2146
+ },
+ {
+ "epoch": 29.824454148471617,
+ "grad_norm": 0.017707521095871925,
+ "learning_rate": 0.0006,
+ "loss": 4.228226661682129,
+ "step": 2147
+ },
+ {
+ "epoch": 29.83842794759825,
+ "grad_norm": 0.015615378506481647,
+ "learning_rate": 0.0006,
+ "loss": 4.175541877746582,
+ "step": 2148
+ },
+ {
+ "epoch": 29.85240174672489,
+ "grad_norm": 0.016111129894852638,
+ "learning_rate": 0.0006,
+ "loss": 4.267936706542969,
+ "step": 2149
+ },
+ {
+ "epoch": 29.86637554585153,
+ "grad_norm": 0.01528779324144125,
+ "learning_rate": 0.0006,
+ "loss": 4.186014175415039,
+ "step": 2150
+ },
+ {
+ "epoch": 29.880349344978168,
+ "grad_norm": 0.014925646595656872,
+ "learning_rate": 0.0006,
+ "loss": 4.174860000610352,
+ "step": 2151
+ },
+ {
+ "epoch": 29.894323144104803,
+ "grad_norm": 0.016917673870921135,
+ "learning_rate": 0.0006,
+ "loss": 4.33908748626709,
+ "step": 2152
+ },
+ {
+ "epoch": 29.90829694323144,
+ "grad_norm": 0.016007075086236,
+ "learning_rate": 0.0006,
+ "loss": 4.266141414642334,
+ "step": 2153
+ },
+ {
+ "epoch": 29.92227074235808,
+ "grad_norm": 0.016178956255316734,
+ "learning_rate": 0.0006,
+ "loss": 4.271215915679932,
+ "step": 2154
+ },
+ {
+ "epoch": 29.936244541484715,
+ "grad_norm": 0.016180871054530144,
+ "learning_rate": 0.0006,
+ "loss": 4.278801441192627,
+ "step": 2155
+ },
+ {
+ "epoch": 29.950218340611354,
+ "grad_norm": 0.015335503034293652,
+ "learning_rate": 0.0006,
+ "loss": 4.194978713989258,
+ "step": 2156
+ },
+ {
+ "epoch": 29.964192139737992,
+ "grad_norm": 0.016580764204263687,
+ "learning_rate": 0.0006,
+ "loss": 4.290809631347656,
+ "step": 2157
+ },
+ {
+ "epoch": 29.978165938864628,
+ "grad_norm": 0.01623128168284893,
+ "learning_rate": 0.0006,
+ "loss": 4.205451488494873,
+ "step": 2158
+ },
+ {
+ "epoch": 29.992139737991266,
+ "grad_norm": 0.014904248528182507,
+ "learning_rate": 0.0006,
+ "loss": 4.24924373626709,
+ "step": 2159
+ },
+ {
+ "epoch": 30.0,
+ "grad_norm": 0.017089299857616425,
+ "learning_rate": 0.0006,
+ "loss": 4.222072601318359,
+ "step": 2160
+ },
+ {
+ "epoch": 30.0,
+ "eval_loss": 4.585411071777344,
+ "eval_runtime": 57.1164,
+ "eval_samples_per_second": 42.755,
+ "eval_steps_per_second": 1.348,
+ "step": 2160
+ },
+ {
+ "epoch": 30.01397379912664,
+ "grad_norm": 0.017226383090019226,
+ "learning_rate": 0.0006,
+ "loss": 4.19467830657959,
+ "step": 2161
+ },
+ {
+ "epoch": 30.027947598253274,
+ "grad_norm": 0.01635066606104374,
+ "learning_rate": 0.0006,
+ "loss": 4.176875591278076,
+ "step": 2162
+ },
+ {
+ "epoch": 30.041921397379912,
+ "grad_norm": 0.01735362410545349,
+ "learning_rate": 0.0006,
+ "loss": 4.13667106628418,
+ "step": 2163
+ },
+ {
+ "epoch": 30.05589519650655,
+ "grad_norm": 0.017047109082341194,
+ "learning_rate": 0.0006,
+ "loss": 4.136329650878906,
+ "step": 2164
+ },
+ {
+ "epoch": 30.069868995633186,
+ "grad_norm": 0.018116053193807602,
+ "learning_rate": 0.0006,
+ "loss": 4.228706359863281,
+ "step": 2165
+ },
+ {
+ "epoch": 30.083842794759825,
+ "grad_norm": 0.01791389286518097,
+ "learning_rate": 0.0006,
+ "loss": 4.1597747802734375,
+ "step": 2166
+ },
+ {
+ "epoch": 30.097816593886463,
+ "grad_norm": 0.017793502658605576,
+ "learning_rate": 0.0006,
+ "loss": 4.0807976722717285,
+ "step": 2167
+ },
+ {
+ "epoch": 30.111790393013102,
+ "grad_norm": 0.01827503927052021,
+ "learning_rate": 0.0006,
+ "loss": 4.170646667480469,
+ "step": 2168
+ },
+ {
+ "epoch": 30.125764192139737,
+ "grad_norm": 0.017228564247488976,
+ "learning_rate": 0.0006,
+ "loss": 4.146934509277344,
+ "step": 2169
+ },
+ {
+ "epoch": 30.139737991266376,
+ "grad_norm": 0.01757437363266945,
+ "learning_rate": 0.0006,
+ "loss": 4.1859130859375,
+ "step": 2170
+ },
+ {
+ "epoch": 30.153711790393015,
+ "grad_norm": 0.01701093092560768,
+ "learning_rate": 0.0006,
+ "loss": 4.126981258392334,
+ "step": 2171
+ },
+ {
+ "epoch": 30.16768558951965,
+ "grad_norm": 0.015220489352941513,
+ "learning_rate": 0.0006,
+ "loss": 4.164018630981445,
+ "step": 2172
+ },
+ {
+ "epoch": 30.18165938864629,
+ "grad_norm": 0.016254238784313202,
+ "learning_rate": 0.0006,
+ "loss": 4.258275508880615,
+ "step": 2173
+ },
+ {
+ "epoch": 30.195633187772927,
+ "grad_norm": 0.016609614714980125,
+ "learning_rate": 0.0006,
+ "loss": 4.222233772277832,
+ "step": 2174
+ },
+ {
+ "epoch": 30.209606986899562,
+ "grad_norm": 0.016531318426132202,
+ "learning_rate": 0.0006,
+ "loss": 4.144489288330078,
+ "step": 2175
+ },
+ {
+ "epoch": 30.2235807860262,
+ "grad_norm": 0.01567668654024601,
+ "learning_rate": 0.0006,
+ "loss": 4.181916236877441,
+ "step": 2176
+ },
+ {
+ "epoch": 30.23755458515284,
+ "grad_norm": 0.017195681110024452,
+ "learning_rate": 0.0006,
+ "loss": 4.189693927764893,
+ "step": 2177
+ },
+ {
+ "epoch": 30.251528384279474,
+ "grad_norm": 0.016882948577404022,
+ "learning_rate": 0.0006,
+ "loss": 4.21604061126709,
+ "step": 2178
+ },
+ {
+ "epoch": 30.265502183406113,
+ "grad_norm": 0.016261622309684753,
+ "learning_rate": 0.0006,
+ "loss": 4.227639198303223,
+ "step": 2179
+ },
+ {
+ "epoch": 30.27947598253275,
+ "grad_norm": 0.016411812976002693,
+ "learning_rate": 0.0006,
+ "loss": 4.130362510681152,
+ "step": 2180
+ },
+ {
+ "epoch": 30.29344978165939,
+ "grad_norm": 0.01732019893825054,
+ "learning_rate": 0.0006,
+ "loss": 4.230443000793457,
+ "step": 2181
+ },
+ {
+ "epoch": 30.307423580786025,
+ "grad_norm": 0.01825905591249466,
+ "learning_rate": 0.0006,
+ "loss": 4.175817489624023,
+ "step": 2182
+ },
+ {
+ "epoch": 30.321397379912664,
+ "grad_norm": 0.01887713558971882,
+ "learning_rate": 0.0006,
+ "loss": 4.169020652770996,
+ "step": 2183
+ },
+ {
+ "epoch": 30.335371179039303,
+ "grad_norm": 0.019391661509871483,
+ "learning_rate": 0.0006,
+ "loss": 4.176318645477295,
+ "step": 2184
+ },
+ {
+ "epoch": 30.349344978165938,
+ "grad_norm": 0.020840002223849297,
+ "learning_rate": 0.0006,
+ "loss": 4.174228668212891,
+ "step": 2185
+ },
+ {
+ "epoch": 30.363318777292577,
+ "grad_norm": 0.020193178206682205,
+ "learning_rate": 0.0006,
+ "loss": 4.127919673919678,
+ "step": 2186
+ },
+ {
+ "epoch": 30.377292576419215,
+ "grad_norm": 0.018449561670422554,
+ "learning_rate": 0.0006,
+ "loss": 4.229487895965576,
+ "step": 2187
+ },
+ {
+ "epoch": 30.39126637554585,
+ "grad_norm": 0.01711983233690262,
+ "learning_rate": 0.0006,
+ "loss": 4.201651573181152,
+ "step": 2188
+ },
+ {
+ "epoch": 30.40524017467249,
+ "grad_norm": 0.017293395474553108,
+ "learning_rate": 0.0006,
+ "loss": 4.168704032897949,
+ "step": 2189
+ },
+ {
+ "epoch": 30.419213973799128,
+ "grad_norm": 0.018526358529925346,
+ "learning_rate": 0.0006,
+ "loss": 4.171864032745361,
+ "step": 2190
+ },
+ {
+ "epoch": 30.433187772925763,
+ "grad_norm": 0.019287291914224625,
+ "learning_rate": 0.0006,
+ "loss": 4.039534568786621,
+ "step": 2191
+ },
+ {
+ "epoch": 30.4471615720524,
+ "grad_norm": 0.016902444884181023,
+ "learning_rate": 0.0006,
+ "loss": 4.282787799835205,
+ "step": 2192
+ },
+ {
+ "epoch": 30.46113537117904,
+ "grad_norm": 0.01656750962138176,
+ "learning_rate": 0.0006,
+ "loss": 4.213098526000977,
+ "step": 2193
+ },
+ {
+ "epoch": 30.475109170305675,
+ "grad_norm": 0.016805065795779228,
+ "learning_rate": 0.0006,
+ "loss": 4.1583476066589355,
+ "step": 2194
+ },
+ {
+ "epoch": 30.489082969432314,
+ "grad_norm": 0.017166638746857643,
+ "learning_rate": 0.0006,
+ "loss": 4.188634395599365,
+ "step": 2195
+ },
+ {
+ "epoch": 30.503056768558952,
+ "grad_norm": 0.01592106744647026,
+ "learning_rate": 0.0006,
+ "loss": 4.234043121337891,
+ "step": 2196
+ },
+ {
+ "epoch": 30.51703056768559,
+ "grad_norm": 0.015270989388227463,
+ "learning_rate": 0.0006,
+ "loss": 4.301420211791992,
+ "step": 2197
+ },
+ {
+ "epoch": 30.531004366812226,
+ "grad_norm": 0.015194724313914776,
+ "learning_rate": 0.0006,
+ "loss": 4.18703031539917,
+ "step": 2198
+ },
+ {
+ "epoch": 30.544978165938865,
+ "grad_norm": 0.015587746165692806,
+ "learning_rate": 0.0006,
+ "loss": 4.138928413391113,
+ "step": 2199
+ },
+ {
+ "epoch": 30.558951965065503,
+ "grad_norm": 0.016029570251703262,
+ "learning_rate": 0.0006,
+ "loss": 4.290863990783691,
+ "step": 2200
+ },
+ {
+ "epoch": 30.57292576419214,
+ "grad_norm": 0.015651309862732887,
+ "learning_rate": 0.0006,
+ "loss": 4.272062301635742,
+ "step": 2201
+ },
+ {
+ "epoch": 30.586899563318777,
+ "grad_norm": 0.015992306172847748,
+ "learning_rate": 0.0006,
+ "loss": 4.2487406730651855,
+ "step": 2202
+ },
+ {
+ "epoch": 30.600873362445416,
+ "grad_norm": 0.014969157986342907,
+ "learning_rate": 0.0006,
+ "loss": 4.115347862243652,
+ "step": 2203
+ },
+ {
+ "epoch": 30.61484716157205,
+ "grad_norm": 0.014639027416706085,
+ "learning_rate": 0.0006,
+ "loss": 4.2207489013671875,
+ "step": 2204
+ },
+ {
+ "epoch": 30.62882096069869,
+ "grad_norm": 0.015059413388371468,
+ "learning_rate": 0.0006,
+ "loss": 4.274044990539551,
+ "step": 2205
+ },
+ {
+ "epoch": 30.64279475982533,
+ "grad_norm": 0.0166641678661108,
+ "learning_rate": 0.0006,
+ "loss": 4.191373348236084,
+ "step": 2206
+ },
+ {
+ "epoch": 30.656768558951963,
+ "grad_norm": 0.017806345596909523,
+ "learning_rate": 0.0006,
+ "loss": 4.139800548553467,
+ "step": 2207
+ },
+ {
+ "epoch": 30.670742358078602,
+ "grad_norm": 0.018712421879172325,
+ "learning_rate": 0.0006,
+ "loss": 4.035965442657471,
+ "step": 2208
+ },
+ {
+ "epoch": 30.68471615720524,
+ "grad_norm": 0.018457243219017982,
+ "learning_rate": 0.0006,
+ "loss": 4.2579240798950195,
+ "step": 2209
+ },
+ {
+ "epoch": 30.69868995633188,
+ "grad_norm": 0.01920999586582184,
+ "learning_rate": 0.0006,
+ "loss": 4.131112098693848,
+ "step": 2210
+ },
+ {
+ "epoch": 30.712663755458514,
+ "grad_norm": 0.01729944534599781,
+ "learning_rate": 0.0006,
+ "loss": 4.254059314727783,
+ "step": 2211
+ },
+ {
+ "epoch": 30.726637554585153,
+ "grad_norm": 0.016914231702685356,
+ "learning_rate": 0.0006,
+ "loss": 4.233428955078125,
+ "step": 2212
+ },
+ {
+ "epoch": 30.74061135371179,
+ "grad_norm": 0.01913098618388176,
+ "learning_rate": 0.0006,
+ "loss": 4.23525333404541,
+ "step": 2213
+ },
+ {
+ "epoch": 30.754585152838427,
+ "grad_norm": 0.019150281324982643,
+ "learning_rate": 0.0006,
+ "loss": 4.213695049285889,
+ "step": 2214
+ },
+ {
+ "epoch": 30.768558951965066,
+ "grad_norm": 0.018947452306747437,
+ "learning_rate": 0.0006,
+ "loss": 4.151697158813477,
+ "step": 2215
+ },
+ {
+ "epoch": 30.782532751091704,
+ "grad_norm": 0.016058508306741714,
+ "learning_rate": 0.0006,
+ "loss": 4.220280170440674,
+ "step": 2216
+ },
+ {
+ "epoch": 30.79650655021834,
+ "grad_norm": 0.017488619312644005,
+ "learning_rate": 0.0006,
+ "loss": 4.154428482055664,
+ "step": 2217
+ },
+ {
+ "epoch": 30.810480349344978,
+ "grad_norm": 0.017355090007185936,
+ "learning_rate": 0.0006,
+ "loss": 4.256086349487305,
+ "step": 2218
+ },
+ {
+ "epoch": 30.824454148471617,
+ "grad_norm": 0.016636032611131668,
+ "learning_rate": 0.0006,
+ "loss": 4.227564334869385,
+ "step": 2219
+ },
+ {
+ "epoch": 30.83842794759825,
+ "grad_norm": 0.016273891553282738,
+ "learning_rate": 0.0006,
+ "loss": 4.200466156005859,
+ "step": 2220
+ },
+ {
+ "epoch": 30.85240174672489,
+ "grad_norm": 0.016939911991357803,
+ "learning_rate": 0.0006,
+ "loss": 4.1571221351623535,
+ "step": 2221
+ },
+ {
+ "epoch": 30.86637554585153,
+ "grad_norm": 0.018332522362470627,
+ "learning_rate": 0.0006,
+ "loss": 4.228398323059082,
+ "step": 2222
+ },
+ {
+ "epoch": 30.880349344978168,
+ "grad_norm": 0.019508114084601402,
+ "learning_rate": 0.0006,
+ "loss": 4.222960472106934,
+ "step": 2223
+ },
+ {
+ "epoch": 30.894323144104803,
+ "grad_norm": 0.018440278246998787,
+ "learning_rate": 0.0006,
+ "loss": 4.261051177978516,
+ "step": 2224
+ },
+ {
+ "epoch": 30.90829694323144,
+ "grad_norm": 0.018113229423761368,
+ "learning_rate": 0.0006,
+ "loss": 4.109235763549805,
+ "step": 2225
+ },
+ {
+ "epoch": 30.92227074235808,
+ "grad_norm": 0.015392083674669266,
+ "learning_rate": 0.0006,
+ "loss": 4.21950101852417,
+ "step": 2226
+ },
+ {
+ "epoch": 30.936244541484715,
+ "grad_norm": 0.01633290946483612,
+ "learning_rate": 0.0006,
+ "loss": 4.176053047180176,
+ "step": 2227
+ },
+ {
+ "epoch": 30.950218340611354,
+ "grad_norm": 0.016236383467912674,
+ "learning_rate": 0.0006,
+ "loss": 4.314309120178223,
+ "step": 2228
+ },
+ {
+ "epoch": 30.964192139737992,
+ "grad_norm": 0.015284373424947262,
+ "learning_rate": 0.0006,
+ "loss": 4.303133487701416,
+ "step": 2229
+ },
+ {
+ "epoch": 30.978165938864628,
+ "grad_norm": 0.015982897952198982,
+ "learning_rate": 0.0006,
+ "loss": 4.25739860534668,
+ "step": 2230
+ },
+ {
+ "epoch": 30.992139737991266,
+ "grad_norm": 0.016682198271155357,
+ "learning_rate": 0.0006,
+ "loss": 4.292210578918457,
+ "step": 2231
+ },
+ {
+ "epoch": 31.0,
+ "grad_norm": 0.017521467059850693,
+ "learning_rate": 0.0006,
+ "loss": 4.170637130737305,
+ "step": 2232
+ },
+ {
+ "epoch": 31.0,
+ "eval_loss": 4.642786979675293,
+ "eval_runtime": 57.029,
+ "eval_samples_per_second": 42.82,
+ "eval_steps_per_second": 1.35,
+ "step": 2232
+ },
+ {
+ "epoch": 31.01397379912664,
+ "grad_norm": 0.016607852652668953,
+ "learning_rate": 0.0006,
+ "loss": 4.106224536895752,
+ "step": 2233
+ },
+ {
+ "epoch": 31.027947598253274,
+ "grad_norm": 0.018448904156684875,
+ "learning_rate": 0.0006,
+ "loss": 4.224459648132324,
+ "step": 2234
+ },
+ {
+ "epoch": 31.041921397379912,
+ "grad_norm": 0.020481228828430176,
+ "learning_rate": 0.0006,
+ "loss": 4.305548667907715,
+ "step": 2235
+ },
+ {
+ "epoch": 31.05589519650655,
+ "grad_norm": 0.01999707892537117,
+ "learning_rate": 0.0006,
+ "loss": 4.253663539886475,
+ "step": 2236
+ },
+ {
+ "epoch": 31.069868995633186,
+ "grad_norm": 0.020428303629159927,
+ "learning_rate": 0.0006,
+ "loss": 4.166954040527344,
+ "step": 2237
+ },
+ {
+ "epoch": 31.083842794759825,
+ "grad_norm": 0.022608213126659393,
+ "learning_rate": 0.0006,
+ "loss": 4.172819137573242,
+ "step": 2238
+ },
+ {
+ "epoch": 31.097816593886463,
+ "grad_norm": 0.024669520556926727,
+ "learning_rate": 0.0006,
+ "loss": 4.106586456298828,
+ "step": 2239
+ },
+ {
+ "epoch": 31.111790393013102,
+ "grad_norm": 0.02247351035475731,
+ "learning_rate": 0.0006,
+ "loss": 4.160271644592285,
+ "step": 2240
+ },
+ {
+ "epoch": 31.125764192139737,
+ "grad_norm": 0.0207810141146183,
+ "learning_rate": 0.0006,
+ "loss": 4.272491931915283,
+ "step": 2241
+ },
+ {
+ "epoch": 31.139737991266376,
+ "grad_norm": 0.021576499566435814,
+ "learning_rate": 0.0006,
+ "loss": 4.171474456787109,
+ "step": 2242
+ },
+ {
+ "epoch": 31.153711790393015,
+ "grad_norm": 0.02337106689810753,
+ "learning_rate": 0.0006,
+ "loss": 4.094712734222412,
+ "step": 2243
+ },
+ {
+ "epoch": 31.16768558951965,
+ "grad_norm": 0.02219875156879425,
+ "learning_rate": 0.0006,
+ "loss": 4.1800994873046875,
+ "step": 2244
+ },
+ {
+ "epoch": 31.18165938864629,
+ "grad_norm": 0.022680504247546196,
+ "learning_rate": 0.0006,
+ "loss": 4.2054362297058105,
+ "step": 2245
+ },
+ {
+ "epoch": 31.195633187772927,
+ "grad_norm": 0.0208734143525362,
+ "learning_rate": 0.0006,
+ "loss": 4.06068229675293,
+ "step": 2246
+ },
+ {
+ "epoch": 31.209606986899562,
+ "grad_norm": 0.01811066083610058,
+ "learning_rate": 0.0006,
+ "loss": 4.2465128898620605,
+ "step": 2247
+ },
+ {
+ "epoch": 31.2235807860262,
+ "grad_norm": 0.018571637570858,
+ "learning_rate": 0.0006,
+ "loss": 4.191955089569092,
+ "step": 2248
+ },
+ {
+ "epoch": 31.23755458515284,
+ "grad_norm": 0.018598807975649834,
+ "learning_rate": 0.0006,
+ "loss": 4.200872421264648,
+ "step": 2249
+ },
+ {
+ "epoch": 31.251528384279474,
+ "grad_norm": 0.017577625811100006,
+ "learning_rate": 0.0006,
+ "loss": 4.148700714111328,
+ "step": 2250
+ },
+ {
+ "epoch": 31.265502183406113,
+ "grad_norm": 0.01817980408668518,
+ "learning_rate": 0.0006,
+ "loss": 4.256978988647461,
+ "step": 2251
+ },
+ {
+ "epoch": 31.27947598253275,
+ "grad_norm": 0.017983486875891685,
+ "learning_rate": 0.0006,
+ "loss": 4.175005912780762,
+ "step": 2252
+ },
+ {
+ "epoch": 31.29344978165939,
+ "grad_norm": 0.016751961782574654,
+ "learning_rate": 0.0006,
+ "loss": 4.237518310546875,
+ "step": 2253
+ },
+ {
+ "epoch": 31.307423580786025,
+ "grad_norm": 0.016407202929258347,
+ "learning_rate": 0.0006,
+ "loss": 4.197269439697266,
+ "step": 2254
+ },
+ {
+ "epoch": 31.321397379912664,
+ "grad_norm": 0.01578645594418049,
+ "learning_rate": 0.0006,
+ "loss": 4.179751396179199,
+ "step": 2255
+ },
+ {
+ "epoch": 31.335371179039303,
+ "grad_norm": 0.016001226380467415,
+ "learning_rate": 0.0006,
+ "loss": 4.08821964263916,
+ "step": 2256
+ },
+ {
+ "epoch": 31.349344978165938,
+ "grad_norm": 0.015122811309993267,
+ "learning_rate": 0.0006,
+ "loss": 4.047828674316406,
+ "step": 2257
+ },
+ {
+ "epoch": 31.363318777292577,
+ "grad_norm": 0.015399965457618237,
+ "learning_rate": 0.0006,
+ "loss": 4.1506524085998535,
+ "step": 2258
+ },
+ {
+ "epoch": 31.377292576419215,
+ "grad_norm": 0.01534605398774147,
+ "learning_rate": 0.0006,
+ "loss": 4.27952766418457,
+ "step": 2259
+ },
+ {
+ "epoch": 31.39126637554585,
+ "grad_norm": 0.015461381524801254,
+ "learning_rate": 0.0006,
+ "loss": 4.27102518081665,
+ "step": 2260
+ },
+ {
+ "epoch": 31.40524017467249,
+ "grad_norm": 0.016769153997302055,
+ "learning_rate": 0.0006,
+ "loss": 4.191592216491699,
+ "step": 2261
+ },
+ {
+ "epoch": 31.419213973799128,
+ "grad_norm": 0.016557959839701653,
+ "learning_rate": 0.0006,
+ "loss": 4.159705638885498,
+ "step": 2262
+ },
+ {
+ "epoch": 31.433187772925763,
+ "grad_norm": 0.01659569889307022,
+ "learning_rate": 0.0006,
+ "loss": 4.2136664390563965,
+ "step": 2263
+ },
+ {
+ "epoch": 31.4471615720524,
+ "grad_norm": 0.017101937904953957,
+ "learning_rate": 0.0006,
+ "loss": 4.2716498374938965,
+ "step": 2264
+ },
+ {
+ "epoch": 31.46113537117904,
+ "grad_norm": 0.01722985878586769,
+ "learning_rate": 0.0006,
+ "loss": 4.300534248352051,
+ "step": 2265
+ },
+ {
+ "epoch": 31.475109170305675,
+ "grad_norm": 0.016800185665488243,
+ "learning_rate": 0.0006,
+ "loss": 4.113240718841553,
+ "step": 2266
+ },
+ {
+ "epoch": 31.489082969432314,
+ "grad_norm": 0.01603875495493412,
+ "learning_rate": 0.0006,
+ "loss": 4.157105445861816,
+ "step": 2267
+ },
+ {
+ "epoch": 31.503056768558952,
+ "grad_norm": 0.015903932973742485,
+ "learning_rate": 0.0006,
+ "loss": 4.233669757843018,
+ "step": 2268
+ },
+ {
+ "epoch": 31.51703056768559,
+ "grad_norm": 0.0160983856767416,
+ "learning_rate": 0.0006,
+ "loss": 4.147704124450684,
+ "step": 2269
+ },
+ {
+ "epoch": 31.531004366812226,
+ "grad_norm": 0.01604730263352394,
+ "learning_rate": 0.0006,
+ "loss": 4.1465911865234375,
+ "step": 2270
+ },
+ {
+ "epoch": 31.544978165938865,
+ "grad_norm": 0.016205785796046257,
+ "learning_rate": 0.0006,
+ "loss": 4.106704235076904,
+ "step": 2271
+ },
+ {
+ "epoch": 31.558951965065503,
+ "grad_norm": 0.017008036375045776,
+ "learning_rate": 0.0006,
+ "loss": 4.191000938415527,
+ "step": 2272
+ },
+ {
+ "epoch": 31.57292576419214,
+ "grad_norm": 0.01699635200202465,
+ "learning_rate": 0.0006,
+ "loss": 4.187042236328125,
+ "step": 2273
+ },
+ {
+ "epoch": 31.586899563318777,
+ "grad_norm": 0.01692967116832733,
+ "learning_rate": 0.0006,
+ "loss": 4.208805561065674,
+ "step": 2274
+ },
+ {
+ "epoch": 31.600873362445416,
+ "grad_norm": 0.017608756199479103,
+ "learning_rate": 0.0006,
+ "loss": 4.200512886047363,
+ "step": 2275
+ },
+ {
+ "epoch": 31.61484716157205,
+ "grad_norm": 0.016590990126132965,
+ "learning_rate": 0.0006,
+ "loss": 4.194705009460449,
+ "step": 2276
+ },
+ {
+ "epoch": 31.62882096069869,
+ "grad_norm": 0.0170235987752676,
+ "learning_rate": 0.0006,
+ "loss": 4.210326194763184,
+ "step": 2277
+ },
+ {
+ "epoch": 31.64279475982533,
+ "grad_norm": 0.01604801043868065,
+ "learning_rate": 0.0006,
+ "loss": 4.196024417877197,
+ "step": 2278
+ },
+ {
+ "epoch": 31.656768558951963,
+ "grad_norm": 0.015694202855229378,
+ "learning_rate": 0.0006,
+ "loss": 4.178236484527588,
+ "step": 2279
+ },
+ {
+ "epoch": 31.670742358078602,
+ "grad_norm": 0.015473959036171436,
+ "learning_rate": 0.0006,
+ "loss": 4.259339809417725,
+ "step": 2280
+ },
+ {
+ "epoch": 31.68471615720524,
+ "grad_norm": 0.014689362607896328,
+ "learning_rate": 0.0006,
+ "loss": 3.9702324867248535,
+ "step": 2281
+ },
+ {
+ "epoch": 31.69868995633188,
+ "grad_norm": 0.015245123766362667,
+ "learning_rate": 0.0006,
+ "loss": 4.197025775909424,
+ "step": 2282
+ },
+ {
+ "epoch": 31.712663755458514,
+ "grad_norm": 0.016148347407579422,
+ "learning_rate": 0.0006,
+ "loss": 4.231856346130371,
+ "step": 2283
+ },
+ {
+ "epoch": 31.726637554585153,
+ "grad_norm": 0.015169818885624409,
+ "learning_rate": 0.0006,
+ "loss": 4.230169296264648,
+ "step": 2284
+ },
+ {
+ "epoch": 31.74061135371179,
+ "grad_norm": 0.014843013137578964,
+ "learning_rate": 0.0006,
+ "loss": 4.195633411407471,
+ "step": 2285
+ },
+ {
+ "epoch": 31.754585152838427,
+ "grad_norm": 0.015229340642690659,
+ "learning_rate": 0.0006,
+ "loss": 4.231280326843262,
+ "step": 2286
+ },
+ {
+ "epoch": 31.768558951965066,
+ "grad_norm": 0.01551708485931158,
+ "learning_rate": 0.0006,
+ "loss": 4.17064094543457,
+ "step": 2287
+ },
+ {
+ "epoch": 31.782532751091704,
+ "grad_norm": 0.015863575041294098,
+ "learning_rate": 0.0006,
+ "loss": 4.1579484939575195,
+ "step": 2288
+ },
+ {
+ "epoch": 31.79650655021834,
+ "grad_norm": 0.015118198469281197,
+ "learning_rate": 0.0006,
+ "loss": 4.1626787185668945,
+ "step": 2289
+ },
+ {
+ "epoch": 31.810480349344978,
+ "grad_norm": 0.015220776200294495,
+ "learning_rate": 0.0006,
+ "loss": 4.135746955871582,
+ "step": 2290
+ },
+ {
+ "epoch": 31.824454148471617,
+ "grad_norm": 0.014529784210026264,
+ "learning_rate": 0.0006,
+ "loss": 4.164931297302246,
+ "step": 2291
+ },
+ {
+ "epoch": 31.83842794759825,
+ "grad_norm": 0.015533296391367912,
+ "learning_rate": 0.0006,
+ "loss": 4.112464904785156,
+ "step": 2292
+ },
+ {
+ "epoch": 31.85240174672489,
+ "grad_norm": 0.016562335193157196,
+ "learning_rate": 0.0006,
+ "loss": 4.164898872375488,
+ "step": 2293
+ },
+ {
+ "epoch": 31.86637554585153,
+ "grad_norm": 0.016258113086223602,
+ "learning_rate": 0.0006,
+ "loss": 4.1979265213012695,
+ "step": 2294
+ },
+ {
+ "epoch": 31.880349344978168,
+ "grad_norm": 0.01558110024780035,
+ "learning_rate": 0.0006,
+ "loss": 4.174675941467285,
+ "step": 2295
+ },
+ {
+ "epoch": 31.894323144104803,
+ "grad_norm": 0.016412515193223953,
+ "learning_rate": 0.0006,
+ "loss": 4.217909336090088,
+ "step": 2296
+ },
+ {
+ "epoch": 31.90829694323144,
+ "grad_norm": 0.015118442475795746,
+ "learning_rate": 0.0006,
+ "loss": 4.137143611907959,
+ "step": 2297
+ },
+ {
+ "epoch": 31.92227074235808,
+ "grad_norm": 0.01506161317229271,
+ "learning_rate": 0.0006,
+ "loss": 4.196122169494629,
+ "step": 2298
+ },
+ {
+ "epoch": 31.936244541484715,
+ "grad_norm": 0.017165271565318108,
+ "learning_rate": 0.0006,
+ "loss": 4.14834451675415,
+ "step": 2299
+ },
+ {
+ "epoch": 31.950218340611354,
+ "grad_norm": 0.01696440577507019,
+ "learning_rate": 0.0006,
+ "loss": 4.218868732452393,
+ "step": 2300
+ },
+ {
+ "epoch": 31.964192139737992,
+ "grad_norm": 0.017349394038319588,
+ "learning_rate": 0.0006,
+ "loss": 4.2754011154174805,
+ "step": 2301
+ },
+ {
+ "epoch": 31.978165938864628,
+ "grad_norm": 0.018387921154499054,
+ "learning_rate": 0.0006,
+ "loss": 4.282623291015625,
+ "step": 2302
+ },
+ {
+ "epoch": 31.992139737991266,
+ "grad_norm": 0.018837926909327507,
+ "learning_rate": 0.0006,
+ "loss": 4.185425281524658,
+ "step": 2303
+ },
+ {
+ "epoch": 32.0,
+ "grad_norm": 0.018998464569449425,
+ "learning_rate": 0.0006,
+ "loss": 4.151291847229004,
+ "step": 2304
+ },
+ {
+ "epoch": 32.0,
+ "eval_loss": 4.522614002227783,
+ "eval_runtime": 57.0106,
+ "eval_samples_per_second": 42.834,
+ "eval_steps_per_second": 1.351,
+ "step": 2304
+ },
+ {
+ "epoch": 32.01397379912664,
+ "grad_norm": 0.019036833196878433,
+ "learning_rate": 0.0006,
+ "loss": 4.183202266693115,
+ "step": 2305
+ },
+ {
+ "epoch": 32.02794759825328,
+ "grad_norm": 0.02207397110760212,
+ "learning_rate": 0.0006,
+ "loss": 4.17875862121582,
+ "step": 2306
+ },
+ {
+ "epoch": 32.041921397379916,
+ "grad_norm": 0.022008279338479042,
+ "learning_rate": 0.0006,
+ "loss": 4.08042049407959,
+ "step": 2307
+ },
+ {
+ "epoch": 32.05589519650655,
+ "grad_norm": 0.022407017648220062,
+ "learning_rate": 0.0006,
+ "loss": 4.112189769744873,
+ "step": 2308
+ },
+ {
+ "epoch": 32.069868995633186,
+ "grad_norm": 0.021599747240543365,
+ "learning_rate": 0.0006,
+ "loss": 4.126455307006836,
+ "step": 2309
+ },
+ {
+ "epoch": 32.083842794759825,
+ "grad_norm": 0.020340966060757637,
+ "learning_rate": 0.0006,
+ "loss": 4.080538272857666,
+ "step": 2310
+ },
+ {
+ "epoch": 32.09781659388646,
+ "grad_norm": 0.020956892520189285,
+ "learning_rate": 0.0006,
+ "loss": 4.171943664550781,
+ "step": 2311
+ },
+ {
+ "epoch": 32.1117903930131,
+ "grad_norm": 0.019966667518019676,
+ "learning_rate": 0.0006,
+ "loss": 4.144007682800293,
+ "step": 2312
+ },
+ {
+ "epoch": 32.12576419213974,
+ "grad_norm": 0.018256952986121178,
+ "learning_rate": 0.0006,
+ "loss": 4.208107948303223,
+ "step": 2313
+ },
+ {
+ "epoch": 32.13973799126637,
+ "grad_norm": 0.017289170995354652,
+ "learning_rate": 0.0006,
+ "loss": 4.182024955749512,
+ "step": 2314
+ },
+ {
+ "epoch": 32.15371179039301,
+ "grad_norm": 0.017656700685620308,
+ "learning_rate": 0.0006,
+ "loss": 4.068373680114746,
+ "step": 2315
+ },
+ {
+ "epoch": 32.16768558951965,
+ "grad_norm": 0.017903361469507217,
+ "learning_rate": 0.0006,
+ "loss": 4.2345428466796875,
+ "step": 2316
+ },
+ {
+ "epoch": 32.18165938864629,
+ "grad_norm": 0.01785109005868435,
+ "learning_rate": 0.0006,
+ "loss": 4.219776153564453,
+ "step": 2317
+ },
+ {
+ "epoch": 32.19563318777293,
+ "grad_norm": 0.01887257769703865,
+ "learning_rate": 0.0006,
+ "loss": 4.139206409454346,
+ "step": 2318
+ },
+ {
+ "epoch": 32.209606986899566,
+ "grad_norm": 0.01924685761332512,
+ "learning_rate": 0.0006,
+ "loss": 4.121052265167236,
+ "step": 2319
+ },
+ {
+ "epoch": 32.223580786026204,
+ "grad_norm": 0.020574072375893593,
+ "learning_rate": 0.0006,
+ "loss": 4.13262414932251,
+ "step": 2320
+ },
+ {
+ "epoch": 32.237554585152836,
+ "grad_norm": 0.023849064484238625,
+ "learning_rate": 0.0006,
+ "loss": 4.239119529724121,
+ "step": 2321
+ },
+ {
+ "epoch": 32.251528384279474,
+ "grad_norm": 0.026239361613988876,
+ "learning_rate": 0.0006,
+ "loss": 4.248993873596191,
+ "step": 2322
+ },
+ {
+ "epoch": 32.26550218340611,
+ "grad_norm": 0.023980243131518364,
+ "learning_rate": 0.0006,
+ "loss": 4.159852504730225,
+ "step": 2323
+ },
+ {
+ "epoch": 32.27947598253275,
+ "grad_norm": 0.020442405715584755,
+ "learning_rate": 0.0006,
+ "loss": 4.133546829223633,
+ "step": 2324
+ },
+ {
+ "epoch": 32.29344978165939,
+ "grad_norm": 0.020661186426877975,
+ "learning_rate": 0.0006,
+ "loss": 4.078243732452393,
+ "step": 2325
+ },
+ {
+ "epoch": 32.30742358078603,
+ "grad_norm": 0.02046387456357479,
+ "learning_rate": 0.0006,
+ "loss": 4.172012805938721,
+ "step": 2326
+ },
+ {
+ "epoch": 32.32139737991266,
+ "grad_norm": 0.02089597098529339,
+ "learning_rate": 0.0006,
+ "loss": 4.0604705810546875,
+ "step": 2327
+ },
+ {
+ "epoch": 32.3353711790393,
+ "grad_norm": 0.01875591278076172,
+ "learning_rate": 0.0006,
+ "loss": 4.125911235809326,
+ "step": 2328
+ },
+ {
+ "epoch": 32.34934497816594,
+ "grad_norm": 0.018606796860694885,
+ "learning_rate": 0.0006,
+ "loss": 4.097362518310547,
+ "step": 2329
+ },
+ {
+ "epoch": 32.36331877729258,
+ "grad_norm": 0.01766626164317131,
+ "learning_rate": 0.0006,
+ "loss": 4.118616580963135,
+ "step": 2330
+ },
+ {
+ "epoch": 32.377292576419215,
+ "grad_norm": 0.017692960798740387,
+ "learning_rate": 0.0006,
+ "loss": 4.1457672119140625,
+ "step": 2331
+ },
+ {
+ "epoch": 32.391266375545854,
+ "grad_norm": 0.017619704827666283,
+ "learning_rate": 0.0006,
+ "loss": 4.175961494445801,
+ "step": 2332
+ },
+ {
+ "epoch": 32.40524017467249,
+ "grad_norm": 0.01784413494169712,
+ "learning_rate": 0.0006,
+ "loss": 4.051291465759277,
+ "step": 2333
+ },
+ {
+ "epoch": 32.419213973799124,
+ "grad_norm": 0.018508171662688255,
+ "learning_rate": 0.0006,
+ "loss": 4.0638556480407715,
+ "step": 2334
+ },
+ {
+ "epoch": 32.43318777292576,
+ "grad_norm": 0.01731436885893345,
+ "learning_rate": 0.0006,
+ "loss": 4.1805419921875,
+ "step": 2335
+ },
+ {
+ "epoch": 32.4471615720524,
+ "grad_norm": 0.015886498615145683,
+ "learning_rate": 0.0006,
+ "loss": 4.151026725769043,
+ "step": 2336
+ },
+ {
+ "epoch": 32.46113537117904,
+ "grad_norm": 0.016583820804953575,
+ "learning_rate": 0.0006,
+ "loss": 4.203604698181152,
+ "step": 2337
+ },
+ {
+ "epoch": 32.47510917030568,
+ "grad_norm": 0.0176097322255373,
+ "learning_rate": 0.0006,
+ "loss": 4.135159015655518,
+ "step": 2338
+ },
+ {
+ "epoch": 32.48908296943232,
+ "grad_norm": 0.016657903790473938,
+ "learning_rate": 0.0006,
+ "loss": 4.126681804656982,
+ "step": 2339
+ },
+ {
+ "epoch": 32.50305676855895,
+ "grad_norm": 0.016866130754351616,
+ "learning_rate": 0.0006,
+ "loss": 4.162341117858887,
+ "step": 2340
+ },
+ {
+ "epoch": 32.51703056768559,
+ "grad_norm": 0.016286680474877357,
+ "learning_rate": 0.0006,
+ "loss": 4.261801242828369,
+ "step": 2341
+ },
+ {
+ "epoch": 32.531004366812226,
+ "grad_norm": 0.01481365505605936,
+ "learning_rate": 0.0006,
+ "loss": 4.110016822814941,
+ "step": 2342
+ },
+ {
+ "epoch": 32.544978165938865,
+ "grad_norm": 0.016605013981461525,
+ "learning_rate": 0.0006,
+ "loss": 4.230024814605713,
+ "step": 2343
+ },
+ {
+ "epoch": 32.5589519650655,
+ "grad_norm": 0.01645175740122795,
+ "learning_rate": 0.0006,
+ "loss": 4.174612045288086,
+ "step": 2344
+ },
+ {
+ "epoch": 32.57292576419214,
+ "grad_norm": 0.014725026674568653,
+ "learning_rate": 0.0006,
+ "loss": 4.066371440887451,
+ "step": 2345
+ },
+ {
+ "epoch": 32.58689956331878,
+ "grad_norm": 0.015003837645053864,
+ "learning_rate": 0.0006,
+ "loss": 4.193343639373779,
+ "step": 2346
+ },
+ {
+ "epoch": 32.60087336244541,
+ "grad_norm": 0.015346302650868893,
+ "learning_rate": 0.0006,
+ "loss": 4.210659980773926,
+ "step": 2347
+ },
+ {
+ "epoch": 32.61484716157205,
+ "grad_norm": 0.015135962516069412,
+ "learning_rate": 0.0006,
+ "loss": 4.204618453979492,
+ "step": 2348
+ },
+ {
+ "epoch": 32.62882096069869,
+ "grad_norm": 0.01601956970989704,
+ "learning_rate": 0.0006,
+ "loss": 4.145564556121826,
+ "step": 2349
+ },
+ {
+ "epoch": 32.64279475982533,
+ "grad_norm": 0.01517851185053587,
+ "learning_rate": 0.0006,
+ "loss": 4.051673889160156,
+ "step": 2350
+ },
+ {
+ "epoch": 32.65676855895197,
+ "grad_norm": 0.014894292689859867,
+ "learning_rate": 0.0006,
+ "loss": 4.184818267822266,
+ "step": 2351
+ },
+ {
+ "epoch": 32.670742358078606,
+ "grad_norm": 0.015720834955573082,
+ "learning_rate": 0.0006,
+ "loss": 4.220036506652832,
+ "step": 2352
+ },
+ {
+ "epoch": 32.68471615720524,
+ "grad_norm": 0.01573433168232441,
+ "learning_rate": 0.0006,
+ "loss": 4.076773166656494,
+ "step": 2353
+ },
+ {
+ "epoch": 32.698689956331876,
+ "grad_norm": 0.015828052535653114,
+ "learning_rate": 0.0006,
+ "loss": 4.116329669952393,
+ "step": 2354
+ },
+ {
+ "epoch": 32.712663755458514,
+ "grad_norm": 0.015462259761989117,
+ "learning_rate": 0.0006,
+ "loss": 4.1155242919921875,
+ "step": 2355
+ },
+ {
+ "epoch": 32.72663755458515,
+ "grad_norm": 0.015763849020004272,
+ "learning_rate": 0.0006,
+ "loss": 4.169687747955322,
+ "step": 2356
+ },
+ {
+ "epoch": 32.74061135371179,
+ "grad_norm": 0.014935145154595375,
+ "learning_rate": 0.0006,
+ "loss": 4.2090911865234375,
+ "step": 2357
+ },
+ {
+ "epoch": 32.75458515283843,
+ "grad_norm": 0.015119184739887714,
+ "learning_rate": 0.0006,
+ "loss": 4.131838798522949,
+ "step": 2358
+ },
+ {
+ "epoch": 32.76855895196506,
+ "grad_norm": 0.014833525754511356,
+ "learning_rate": 0.0006,
+ "loss": 4.17618465423584,
+ "step": 2359
+ },
+ {
+ "epoch": 32.7825327510917,
+ "grad_norm": 0.01588359661400318,
+ "learning_rate": 0.0006,
+ "loss": 4.163723945617676,
+ "step": 2360
+ },
+ {
+ "epoch": 32.79650655021834,
+ "grad_norm": 0.0168188214302063,
+ "learning_rate": 0.0006,
+ "loss": 4.178189277648926,
+ "step": 2361
+ },
+ {
+ "epoch": 32.81048034934498,
+ "grad_norm": 0.01726095750927925,
+ "learning_rate": 0.0006,
+ "loss": 4.144223690032959,
+ "step": 2362
+ },
+ {
+ "epoch": 32.82445414847162,
+ "grad_norm": 0.01701410301029682,
+ "learning_rate": 0.0006,
+ "loss": 4.209896087646484,
+ "step": 2363
+ },
+ {
+ "epoch": 32.838427947598255,
+ "grad_norm": 0.01682264916598797,
+ "learning_rate": 0.0006,
+ "loss": 4.26223087310791,
+ "step": 2364
+ },
+ {
+ "epoch": 32.852401746724894,
+ "grad_norm": 0.015597456134855747,
+ "learning_rate": 0.0006,
+ "loss": 4.292664051055908,
+ "step": 2365
+ },
+ {
+ "epoch": 32.866375545851525,
+ "grad_norm": 0.015258579514920712,
+ "learning_rate": 0.0006,
+ "loss": 4.165006637573242,
+ "step": 2366
+ },
+ {
+ "epoch": 32.880349344978164,
+ "grad_norm": 0.01662200503051281,
+ "learning_rate": 0.0006,
+ "loss": 4.309786319732666,
+ "step": 2367
+ },
+ {
+ "epoch": 32.8943231441048,
+ "grad_norm": 0.015691161155700684,
+ "learning_rate": 0.0006,
+ "loss": 4.167202949523926,
+ "step": 2368
+ },
+ {
+ "epoch": 32.90829694323144,
+ "grad_norm": 0.014870637096464634,
+ "learning_rate": 0.0006,
+ "loss": 4.142415523529053,
+ "step": 2369
+ },
+ {
+ "epoch": 32.92227074235808,
+ "grad_norm": 0.01567252166569233,
+ "learning_rate": 0.0006,
+ "loss": 4.2312092781066895,
+ "step": 2370
+ },
+ {
+ "epoch": 32.93624454148472,
+ "grad_norm": 0.014948616735637188,
+ "learning_rate": 0.0006,
+ "loss": 4.1514129638671875,
+ "step": 2371
+ },
+ {
+ "epoch": 32.95021834061135,
+ "grad_norm": 0.014654500409960747,
+ "learning_rate": 0.0006,
+ "loss": 4.205010890960693,
+ "step": 2372
+ },
+ {
+ "epoch": 32.96419213973799,
+ "grad_norm": 0.015059148892760277,
+ "learning_rate": 0.0006,
+ "loss": 4.148405075073242,
+ "step": 2373
+ },
+ {
+ "epoch": 32.97816593886463,
+ "grad_norm": 0.015026576817035675,
+ "learning_rate": 0.0006,
+ "loss": 4.182535171508789,
+ "step": 2374
+ },
+ {
+ "epoch": 32.992139737991266,
+ "grad_norm": 0.016394320875406265,
+ "learning_rate": 0.0006,
+ "loss": 4.129051208496094,
+ "step": 2375
+ },
+ {
+ "epoch": 33.0,
+ "grad_norm": 0.01867043413221836,
+ "learning_rate": 0.0006,
+ "loss": 4.185556411743164,
+ "step": 2376
+ },
+ {
+ "epoch": 33.0,
+ "eval_loss": 4.640405654907227,
+ "eval_runtime": 56.5861,
+ "eval_samples_per_second": 43.155,
+ "eval_steps_per_second": 1.361,
+ "step": 2376
+ },
+ {
+ "epoch": 33.01397379912664,
+ "grad_norm": 0.018626727163791656,
+ "learning_rate": 0.0006,
+ "loss": 4.143031597137451,
+ "step": 2377
+ },
+ {
+ "epoch": 33.02794759825328,
+ "grad_norm": 0.019283363595604897,
+ "learning_rate": 0.0006,
+ "loss": 4.117845058441162,
+ "step": 2378
+ },
+ {
+ "epoch": 33.041921397379916,
+ "grad_norm": 0.018185822293162346,
+ "learning_rate": 0.0006,
+ "loss": 4.076862812042236,
+ "step": 2379
+ },
+ {
+ "epoch": 33.05589519650655,
+ "grad_norm": 0.019411196932196617,
+ "learning_rate": 0.0006,
+ "loss": 4.124156475067139,
+ "step": 2380
+ },
+ {
+ "epoch": 33.069868995633186,
+ "grad_norm": 0.019248466938734055,
+ "learning_rate": 0.0006,
+ "loss": 4.201735019683838,
+ "step": 2381
+ },
+ {
+ "epoch": 33.083842794759825,
+ "grad_norm": 0.018856720998883247,
+ "learning_rate": 0.0006,
+ "loss": 4.00296688079834,
+ "step": 2382
+ },
+ {
+ "epoch": 33.09781659388646,
+ "grad_norm": 0.018554294481873512,
+ "learning_rate": 0.0006,
+ "loss": 4.270577430725098,
+ "step": 2383
+ },
+ {
+ "epoch": 33.1117903930131,
+ "grad_norm": 0.019946224987506866,
+ "learning_rate": 0.0006,
+ "loss": 4.119589328765869,
+ "step": 2384
+ },
+ {
+ "epoch": 33.12576419213974,
+ "grad_norm": 0.021395670250058174,
+ "learning_rate": 0.0006,
+ "loss": 4.128847122192383,
+ "step": 2385
+ },
+ {
+ "epoch": 33.13973799126637,
+ "grad_norm": 0.02009674906730652,
+ "learning_rate": 0.0006,
+ "loss": 4.183228969573975,
+ "step": 2386
+ },
+ {
+ "epoch": 33.15371179039301,
+ "grad_norm": 0.01820223033428192,
+ "learning_rate": 0.0006,
+ "loss": 4.131601810455322,
+ "step": 2387
+ },
+ {
+ "epoch": 33.16768558951965,
+ "grad_norm": 0.018653156235814095,
+ "learning_rate": 0.0006,
+ "loss": 4.116430282592773,
+ "step": 2388
+ },
+ {
+ "epoch": 33.18165938864629,
+ "grad_norm": 0.016828063875436783,
+ "learning_rate": 0.0006,
+ "loss": 4.054663181304932,
+ "step": 2389
+ },
+ {
+ "epoch": 33.19563318777293,
+ "grad_norm": 0.017025409266352654,
+ "learning_rate": 0.0006,
+ "loss": 4.110044956207275,
+ "step": 2390
+ },
+ {
+ "epoch": 33.209606986899566,
+ "grad_norm": 0.016608452424407005,
+ "learning_rate": 0.0006,
+ "loss": 4.102668285369873,
+ "step": 2391
+ },
+ {
+ "epoch": 33.223580786026204,
+ "grad_norm": 0.01743577979505062,
+ "learning_rate": 0.0006,
+ "loss": 4.166791915893555,
+ "step": 2392
+ },
+ {
+ "epoch": 33.237554585152836,
+ "grad_norm": 0.017253734171390533,
+ "learning_rate": 0.0006,
+ "loss": 4.217832088470459,
+ "step": 2393
+ },
+ {
+ "epoch": 33.251528384279474,
+ "grad_norm": 0.017937595024704933,
+ "learning_rate": 0.0006,
+ "loss": 4.2173967361450195,
+ "step": 2394
+ },
+ {
+ "epoch": 33.26550218340611,
+ "grad_norm": 0.01985993981361389,
+ "learning_rate": 0.0006,
+ "loss": 4.17849588394165,
+ "step": 2395
+ },
+ {
+ "epoch": 33.27947598253275,
+ "grad_norm": 0.019243355840444565,
+ "learning_rate": 0.0006,
+ "loss": 4.086523056030273,
+ "step": 2396
+ },
+ {
+ "epoch": 33.29344978165939,
+ "grad_norm": 0.01924416422843933,
+ "learning_rate": 0.0006,
+ "loss": 4.128574371337891,
+ "step": 2397
+ },
+ {
+ "epoch": 33.30742358078603,
+ "grad_norm": 0.01983758620917797,
+ "learning_rate": 0.0006,
+ "loss": 4.061581611633301,
+ "step": 2398
+ },
+ {
+ "epoch": 33.32139737991266,
+ "grad_norm": 0.01930462196469307,
+ "learning_rate": 0.0006,
+ "loss": 4.091731071472168,
+ "step": 2399
+ },
+ {
+ "epoch": 33.3353711790393,
+ "grad_norm": 0.017988424748182297,
+ "learning_rate": 0.0006,
+ "loss": 4.155599117279053,
+ "step": 2400
+ },
+ {
+ "epoch": 33.34934497816594,
+ "grad_norm": 0.015194767154753208,
+ "learning_rate": 0.0006,
+ "loss": 4.144134044647217,
+ "step": 2401
+ },
+ {
+ "epoch": 33.36331877729258,
+ "grad_norm": 0.01536081824451685,
+ "learning_rate": 0.0006,
+ "loss": 4.127272129058838,
+ "step": 2402
+ },
+ {
+ "epoch": 33.377292576419215,
+ "grad_norm": 0.015431297942996025,
+ "learning_rate": 0.0006,
+ "loss": 4.3213019371032715,
+ "step": 2403
+ },
+ {
+ "epoch": 33.391266375545854,
+ "grad_norm": 0.015740280970931053,
+ "learning_rate": 0.0006,
+ "loss": 4.121467590332031,
+ "step": 2404
+ },
+ {
+ "epoch": 33.40524017467249,
+ "grad_norm": 0.015379384160041809,
+ "learning_rate": 0.0006,
+ "loss": 4.095216751098633,
+ "step": 2405
+ },
+ {
+ "epoch": 33.419213973799124,
+ "grad_norm": 0.01528073474764824,
+ "learning_rate": 0.0006,
+ "loss": 4.13047981262207,
+ "step": 2406
+ },
+ {
+ "epoch": 33.43318777292576,
+ "grad_norm": 0.015739979222416878,
+ "learning_rate": 0.0006,
+ "loss": 4.2533135414123535,
+ "step": 2407
+ },
+ {
+ "epoch": 33.4471615720524,
+ "grad_norm": 0.01519764307886362,
+ "learning_rate": 0.0006,
+ "loss": 4.131782054901123,
+ "step": 2408
+ },
+ {
+ "epoch": 33.46113537117904,
+ "grad_norm": 0.015425996854901314,
+ "learning_rate": 0.0006,
+ "loss": 4.02811336517334,
+ "step": 2409
+ },
+ {
+ "epoch": 33.47510917030568,
+ "grad_norm": 0.015378830023109913,
+ "learning_rate": 0.0006,
+ "loss": 4.08652925491333,
+ "step": 2410
+ },
+ {
+ "epoch": 33.48908296943232,
+ "grad_norm": 0.01473075058311224,
+ "learning_rate": 0.0006,
+ "loss": 4.214364528656006,
+ "step": 2411
+ },
+ {
+ "epoch": 33.50305676855895,
+ "grad_norm": 0.015141637064516544,
+ "learning_rate": 0.0006,
+ "loss": 4.155712127685547,
+ "step": 2412
+ },
+ {
+ "epoch": 33.51703056768559,
+ "grad_norm": 0.01641593500971794,
+ "learning_rate": 0.0006,
+ "loss": 4.097907066345215,
+ "step": 2413
+ },
+ {
+ "epoch": 33.531004366812226,
+ "grad_norm": 0.015182029455900192,
+ "learning_rate": 0.0006,
+ "loss": 4.238499641418457,
+ "step": 2414
+ },
+ {
+ "epoch": 33.544978165938865,
+ "grad_norm": 0.01643732376396656,
+ "learning_rate": 0.0006,
+ "loss": 4.0661492347717285,
+ "step": 2415
+ },
+ {
+ "epoch": 33.5589519650655,
+ "grad_norm": 0.015614228323101997,
+ "learning_rate": 0.0006,
+ "loss": 4.120795249938965,
+ "step": 2416
+ },
+ {
+ "epoch": 33.57292576419214,
+ "grad_norm": 0.016066187992691994,
+ "learning_rate": 0.0006,
+ "loss": 4.096465110778809,
+ "step": 2417
+ },
+ {
+ "epoch": 33.58689956331878,
+ "grad_norm": 0.017214342951774597,
+ "learning_rate": 0.0006,
+ "loss": 4.242628574371338,
+ "step": 2418
+ },
+ {
+ "epoch": 33.60087336244541,
+ "grad_norm": 0.01787586510181427,
+ "learning_rate": 0.0006,
+ "loss": 4.149589538574219,
+ "step": 2419
+ },
+ {
+ "epoch": 33.61484716157205,
+ "grad_norm": 0.019455241039395332,
+ "learning_rate": 0.0006,
+ "loss": 4.267796039581299,
+ "step": 2420
+ },
+ {
+ "epoch": 33.62882096069869,
+ "grad_norm": 0.01984674669802189,
+ "learning_rate": 0.0006,
+ "loss": 4.175220966339111,
+ "step": 2421
+ },
+ {
+ "epoch": 33.64279475982533,
+ "grad_norm": 0.021933717653155327,
+ "learning_rate": 0.0006,
+ "loss": 4.106647491455078,
+ "step": 2422
+ },
+ {
+ "epoch": 33.65676855895197,
+ "grad_norm": 0.02123589813709259,
+ "learning_rate": 0.0006,
+ "loss": 4.204996109008789,
+ "step": 2423
+ },
+ {
+ "epoch": 33.670742358078606,
+ "grad_norm": 0.019814589992165565,
+ "learning_rate": 0.0006,
+ "loss": 4.186171531677246,
+ "step": 2424
+ },
+ {
+ "epoch": 33.68471615720524,
+ "grad_norm": 0.02059728465974331,
+ "learning_rate": 0.0006,
+ "loss": 4.170509338378906,
+ "step": 2425
+ },
+ {
+ "epoch": 33.698689956331876,
+ "grad_norm": 0.018483608961105347,
+ "learning_rate": 0.0006,
+ "loss": 4.085585594177246,
+ "step": 2426
+ },
+ {
+ "epoch": 33.712663755458514,
+ "grad_norm": 0.018403852358460426,
+ "learning_rate": 0.0006,
+ "loss": 4.2551774978637695,
+ "step": 2427
+ },
+ {
+ "epoch": 33.72663755458515,
+ "grad_norm": 0.018653830513358116,
+ "learning_rate": 0.0006,
+ "loss": 4.091766357421875,
+ "step": 2428
+ },
+ {
+ "epoch": 33.74061135371179,
+ "grad_norm": 0.017692334949970245,
+ "learning_rate": 0.0006,
+ "loss": 4.100838661193848,
+ "step": 2429
+ },
+ {
+ "epoch": 33.75458515283843,
+ "grad_norm": 0.016029126942157745,
+ "learning_rate": 0.0006,
+ "loss": 4.1530680656433105,
+ "step": 2430
+ },
+ {
+ "epoch": 33.76855895196506,
+ "grad_norm": 0.016095684841275215,
+ "learning_rate": 0.0006,
+ "loss": 4.20544958114624,
+ "step": 2431
+ },
+ {
+ "epoch": 33.7825327510917,
+ "grad_norm": 0.016375798732042313,
+ "learning_rate": 0.0006,
+ "loss": 4.199177265167236,
+ "step": 2432
+ },
+ {
+ "epoch": 33.79650655021834,
+ "grad_norm": 0.01590786688029766,
+ "learning_rate": 0.0006,
+ "loss": 4.200799465179443,
+ "step": 2433
+ },
+ {
+ "epoch": 33.81048034934498,
+ "grad_norm": 0.014896097593009472,
+ "learning_rate": 0.0006,
+ "loss": 4.185850143432617,
+ "step": 2434
+ },
+ {
+ "epoch": 33.82445414847162,
+ "grad_norm": 0.01438513956964016,
+ "learning_rate": 0.0006,
+ "loss": 4.214962005615234,
+ "step": 2435
+ },
+ {
+ "epoch": 33.838427947598255,
+ "grad_norm": 0.014459339901804924,
+ "learning_rate": 0.0006,
+ "loss": 4.196972370147705,
+ "step": 2436
+ },
+ {
+ "epoch": 33.852401746724894,
+ "grad_norm": 0.014782169833779335,
+ "learning_rate": 0.0006,
+ "loss": 4.0885910987854,
+ "step": 2437
+ },
+ {
+ "epoch": 33.866375545851525,
+ "grad_norm": 0.01520370040088892,
+ "learning_rate": 0.0006,
+ "loss": 4.0323262214660645,
+ "step": 2438
+ },
+ {
+ "epoch": 33.880349344978164,
+ "grad_norm": 0.014056282117962837,
+ "learning_rate": 0.0006,
+ "loss": 4.1386518478393555,
+ "step": 2439
+ },
+ {
+ "epoch": 33.8943231441048,
+ "grad_norm": 0.013567990623414516,
+ "learning_rate": 0.0006,
+ "loss": 4.095754623413086,
+ "step": 2440
+ },
+ {
+ "epoch": 33.90829694323144,
+ "grad_norm": 0.013306716457009315,
+ "learning_rate": 0.0006,
+ "loss": 4.1623077392578125,
+ "step": 2441
+ },
+ {
+ "epoch": 33.92227074235808,
+ "grad_norm": 0.014047150500118732,
+ "learning_rate": 0.0006,
+ "loss": 4.142541885375977,
+ "step": 2442
+ },
+ {
+ "epoch": 33.93624454148472,
+ "grad_norm": 0.01377617847174406,
+ "learning_rate": 0.0006,
+ "loss": 4.268551826477051,
+ "step": 2443
+ },
+ {
+ "epoch": 33.95021834061135,
+ "grad_norm": 0.014671621844172478,
+ "learning_rate": 0.0006,
+ "loss": 4.198196887969971,
+ "step": 2444
+ },
+ {
+ "epoch": 33.96419213973799,
+ "grad_norm": 0.014806744642555714,
+ "learning_rate": 0.0006,
+ "loss": 4.057899475097656,
+ "step": 2445
+ },
+ {
+ "epoch": 33.97816593886463,
+ "grad_norm": 0.015734272077679634,
+ "learning_rate": 0.0006,
+ "loss": 4.163505554199219,
+ "step": 2446
+ },
+ {
+ "epoch": 33.992139737991266,
+ "grad_norm": 0.015250146389007568,
+ "learning_rate": 0.0006,
+ "loss": 4.076294898986816,
+ "step": 2447
+ },
+ {
+ "epoch": 34.0,
+ "grad_norm": 0.016733702272176743,
+ "learning_rate": 0.0006,
+ "loss": 4.138420104980469,
+ "step": 2448
+ },
+ {
+ "epoch": 34.0,
+ "eval_loss": 4.542503356933594,
+ "eval_runtime": 56.983,
+ "eval_samples_per_second": 42.855,
+ "eval_steps_per_second": 1.351,
+ "step": 2448
+ },
+ {
+ "epoch": 34.01397379912664,
+ "grad_norm": 0.015866613015532494,
+ "learning_rate": 0.0006,
+ "loss": 4.032081127166748,
+ "step": 2449
+ },
+ {
+ "epoch": 34.02794759825328,
+ "grad_norm": 0.01946570910513401,
+ "learning_rate": 0.0006,
+ "loss": 4.122320175170898,
+ "step": 2450
+ },
+ {
+ "epoch": 34.041921397379916,
+ "grad_norm": 0.020265938714146614,
+ "learning_rate": 0.0006,
+ "loss": 4.06460428237915,
+ "step": 2451
+ },
+ {
+ "epoch": 34.05589519650655,
+ "grad_norm": 0.018887171521782875,
+ "learning_rate": 0.0006,
+ "loss": 4.096906661987305,
+ "step": 2452
+ },
+ {
+ "epoch": 34.069868995633186,
+ "grad_norm": 0.017304692417383194,
+ "learning_rate": 0.0006,
+ "loss": 4.137261390686035,
+ "step": 2453
+ },
+ {
+ "epoch": 34.083842794759825,
+ "grad_norm": 0.017029082402586937,
+ "learning_rate": 0.0006,
+ "loss": 4.081112861633301,
+ "step": 2454
+ },
+ {
+ "epoch": 34.09781659388646,
+ "grad_norm": 0.016414813697338104,
+ "learning_rate": 0.0006,
+ "loss": 4.140783309936523,
+ "step": 2455
+ },
+ {
+ "epoch": 34.1117903930131,
+ "grad_norm": 0.018238281831145287,
+ "learning_rate": 0.0006,
+ "loss": 4.161116600036621,
+ "step": 2456
+ },
+ {
+ "epoch": 34.12576419213974,
+ "grad_norm": 0.019929232075810432,
+ "learning_rate": 0.0006,
+ "loss": 4.243722438812256,
+ "step": 2457
+ },
+ {
+ "epoch": 34.13973799126637,
+ "grad_norm": 0.019375786185264587,
+ "learning_rate": 0.0006,
+ "loss": 4.134873390197754,
+ "step": 2458
+ },
+ {
+ "epoch": 34.15371179039301,
+ "grad_norm": 0.018151137977838516,
+ "learning_rate": 0.0006,
+ "loss": 4.098745346069336,
+ "step": 2459
+ },
+ {
+ "epoch": 34.16768558951965,
+ "grad_norm": 0.017054934054613113,
+ "learning_rate": 0.0006,
+ "loss": 4.194662570953369,
+ "step": 2460
+ },
+ {
+ "epoch": 34.18165938864629,
+ "grad_norm": 0.018220555037260056,
+ "learning_rate": 0.0006,
+ "loss": 4.128866195678711,
+ "step": 2461
+ },
+ {
+ "epoch": 34.19563318777293,
+ "grad_norm": 0.01692879945039749,
+ "learning_rate": 0.0006,
+ "loss": 4.062477111816406,
+ "step": 2462
+ },
+ {
+ "epoch": 34.209606986899566,
+ "grad_norm": 0.01674199476838112,
+ "learning_rate": 0.0006,
+ "loss": 4.186149597167969,
+ "step": 2463
+ },
+ {
+ "epoch": 34.223580786026204,
+ "grad_norm": 0.016527455300092697,
+ "learning_rate": 0.0006,
+ "loss": 4.0734028816223145,
+ "step": 2464
+ },
+ {
+ "epoch": 34.237554585152836,
+ "grad_norm": 0.016448386013507843,
+ "learning_rate": 0.0006,
+ "loss": 4.098896026611328,
+ "step": 2465
+ },
+ {
+ "epoch": 34.251528384279474,
+ "grad_norm": 0.014644953422248363,
+ "learning_rate": 0.0006,
+ "loss": 4.1031904220581055,
+ "step": 2466
+ },
+ {
+ "epoch": 34.26550218340611,
+ "grad_norm": 0.016649341210722923,
+ "learning_rate": 0.0006,
+ "loss": 4.148791313171387,
+ "step": 2467
+ },
+ {
+ "epoch": 34.27947598253275,
+ "grad_norm": 0.017442714422941208,
+ "learning_rate": 0.0006,
+ "loss": 4.070799827575684,
+ "step": 2468
+ },
+ {
+ "epoch": 34.29344978165939,
+ "grad_norm": 0.017852013930678368,
+ "learning_rate": 0.0006,
+ "loss": 4.031924247741699,
+ "step": 2469
+ },
+ {
+ "epoch": 34.30742358078603,
+ "grad_norm": 0.016371270641684532,
+ "learning_rate": 0.0006,
+ "loss": 4.149420738220215,
+ "step": 2470
+ },
+ {
+ "epoch": 34.32139737991266,
+ "grad_norm": 0.015846814960241318,
+ "learning_rate": 0.0006,
+ "loss": 4.15427303314209,
+ "step": 2471
+ },
+ {
+ "epoch": 34.3353711790393,
+ "grad_norm": 0.017411163076758385,
+ "learning_rate": 0.0006,
+ "loss": 4.1304192543029785,
+ "step": 2472
+ },
+ {
+ "epoch": 34.34934497816594,
+ "grad_norm": 0.01780046708881855,
+ "learning_rate": 0.0006,
+ "loss": 4.191732406616211,
+ "step": 2473
+ },
+ {
+ "epoch": 34.36331877729258,
+ "grad_norm": 0.01958383060991764,
+ "learning_rate": 0.0006,
+ "loss": 4.068355083465576,
+ "step": 2474
+ },
+ {
+ "epoch": 34.377292576419215,
+ "grad_norm": 0.022173317149281502,
+ "learning_rate": 0.0006,
+ "loss": 4.182929039001465,
+ "step": 2475
+ },
+ {
+ "epoch": 34.391266375545854,
+ "grad_norm": 0.022389156743884087,
+ "learning_rate": 0.0006,
+ "loss": 4.110218048095703,
+ "step": 2476
+ },
+ {
+ "epoch": 34.40524017467249,
+ "grad_norm": 0.022478511556982994,
+ "learning_rate": 0.0006,
+ "loss": 4.095479488372803,
+ "step": 2477
+ },
+ {
+ "epoch": 34.419213973799124,
+ "grad_norm": 0.019504647701978683,
+ "learning_rate": 0.0006,
+ "loss": 4.101964950561523,
+ "step": 2478
+ },
+ {
+ "epoch": 34.43318777292576,
+ "grad_norm": 0.019378576427698135,
+ "learning_rate": 0.0006,
+ "loss": 4.14620304107666,
+ "step": 2479
+ },
+ {
+ "epoch": 34.4471615720524,
+ "grad_norm": 0.017993204295635223,
+ "learning_rate": 0.0006,
+ "loss": 4.183492660522461,
+ "step": 2480
+ },
+ {
+ "epoch": 34.46113537117904,
+ "grad_norm": 0.01814805157482624,
+ "learning_rate": 0.0006,
+ "loss": 4.151247978210449,
+ "step": 2481
+ },
+ {
+ "epoch": 34.47510917030568,
+ "grad_norm": 0.018622448667883873,
+ "learning_rate": 0.0006,
+ "loss": 4.020938396453857,
+ "step": 2482
+ },
+ {
+ "epoch": 34.48908296943232,
+ "grad_norm": 0.018360132351517677,
+ "learning_rate": 0.0006,
+ "loss": 4.064734935760498,
+ "step": 2483
+ },
+ {
+ "epoch": 34.50305676855895,
+ "grad_norm": 0.017715711146593094,
+ "learning_rate": 0.0006,
+ "loss": 4.185687065124512,
+ "step": 2484
+ },
+ {
+ "epoch": 34.51703056768559,
+ "grad_norm": 0.01714295893907547,
+ "learning_rate": 0.0006,
+ "loss": 4.1044721603393555,
+ "step": 2485
+ },
+ {
+ "epoch": 34.531004366812226,
+ "grad_norm": 0.016943803057074547,
+ "learning_rate": 0.0006,
+ "loss": 4.106837272644043,
+ "step": 2486
+ },
+ {
+ "epoch": 34.544978165938865,
+ "grad_norm": 0.0182423684746027,
+ "learning_rate": 0.0006,
+ "loss": 4.1041340827941895,
+ "step": 2487
+ },
+ {
+ "epoch": 34.5589519650655,
+ "grad_norm": 0.01636587455868721,
+ "learning_rate": 0.0006,
+ "loss": 3.9735546112060547,
+ "step": 2488
+ },
+ {
+ "epoch": 34.57292576419214,
+ "grad_norm": 0.014593546278774738,
+ "learning_rate": 0.0006,
+ "loss": 4.041980266571045,
+ "step": 2489
+ },
+ {
+ "epoch": 34.58689956331878,
+ "grad_norm": 0.015436965972185135,
+ "learning_rate": 0.0006,
+ "loss": 4.301148414611816,
+ "step": 2490
+ },
+ {
+ "epoch": 34.60087336244541,
+ "grad_norm": 0.015075269155204296,
+ "learning_rate": 0.0006,
+ "loss": 4.118527412414551,
+ "step": 2491
+ },
+ {
+ "epoch": 34.61484716157205,
+ "grad_norm": 0.01701514609158039,
+ "learning_rate": 0.0006,
+ "loss": 4.245968818664551,
+ "step": 2492
+ },
+ {
+ "epoch": 34.62882096069869,
+ "grad_norm": 0.018047887831926346,
+ "learning_rate": 0.0006,
+ "loss": 4.227261066436768,
+ "step": 2493
+ },
+ {
+ "epoch": 34.64279475982533,
+ "grad_norm": 0.018499085679650307,
+ "learning_rate": 0.0006,
+ "loss": 4.121360778808594,
+ "step": 2494
+ },
+ {
+ "epoch": 34.65676855895197,
+ "grad_norm": 0.018987711519002914,
+ "learning_rate": 0.0006,
+ "loss": 4.131564617156982,
+ "step": 2495
+ },
+ {
+ "epoch": 34.670742358078606,
+ "grad_norm": 0.01834476925432682,
+ "learning_rate": 0.0006,
+ "loss": 4.224941253662109,
+ "step": 2496
+ },
+ {
+ "epoch": 34.68471615720524,
+ "grad_norm": 0.017507942393422127,
+ "learning_rate": 0.0006,
+ "loss": 4.1394548416137695,
+ "step": 2497
+ },
+ {
+ "epoch": 34.698689956331876,
+ "grad_norm": 0.01680580899119377,
+ "learning_rate": 0.0006,
+ "loss": 4.1073431968688965,
+ "step": 2498
+ },
+ {
+ "epoch": 34.712663755458514,
+ "grad_norm": 0.015233514830470085,
+ "learning_rate": 0.0006,
+ "loss": 4.136994361877441,
+ "step": 2499
+ },
+ {
+ "epoch": 34.72663755458515,
+ "grad_norm": 0.015482917428016663,
+ "learning_rate": 0.0006,
+ "loss": 4.110719680786133,
+ "step": 2500
+ },
+ {
+ "epoch": 34.74061135371179,
+ "grad_norm": 0.015940461307764053,
+ "learning_rate": 0.0006,
+ "loss": 4.207851886749268,
+ "step": 2501
+ },
+ {
+ "epoch": 34.75458515283843,
+ "grad_norm": 0.015355597250163555,
+ "learning_rate": 0.0006,
+ "loss": 4.169469833374023,
+ "step": 2502
+ },
+ {
+ "epoch": 34.76855895196506,
+ "grad_norm": 0.015974000096321106,
+ "learning_rate": 0.0006,
+ "loss": 4.110061168670654,
+ "step": 2503
+ },
+ {
+ "epoch": 34.7825327510917,
+ "grad_norm": 0.016088927164673805,
+ "learning_rate": 0.0006,
+ "loss": 4.155452728271484,
+ "step": 2504
+ },
+ {
+ "epoch": 34.79650655021834,
+ "grad_norm": 0.016156120225787163,
+ "learning_rate": 0.0006,
+ "loss": 4.178092956542969,
+ "step": 2505
+ },
+ {
+ "epoch": 34.81048034934498,
+ "grad_norm": 0.014928647316992283,
+ "learning_rate": 0.0006,
+ "loss": 4.05457878112793,
+ "step": 2506
+ },
+ {
+ "epoch": 34.82445414847162,
+ "grad_norm": 0.015289328061044216,
+ "learning_rate": 0.0006,
+ "loss": 4.18869686126709,
+ "step": 2507
+ },
+ {
+ "epoch": 34.838427947598255,
+ "grad_norm": 0.015799039974808693,
+ "learning_rate": 0.0006,
+ "loss": 4.106073379516602,
+ "step": 2508
+ },
+ {
+ "epoch": 34.852401746724894,
+ "grad_norm": 0.015824418514966965,
+ "learning_rate": 0.0006,
+ "loss": 4.0634589195251465,
+ "step": 2509
+ },
+ {
+ "epoch": 34.866375545851525,
+ "grad_norm": 0.015321656130254269,
+ "learning_rate": 0.0006,
+ "loss": 4.093090057373047,
+ "step": 2510
+ },
+ {
+ "epoch": 34.880349344978164,
+ "grad_norm": 0.014380166307091713,
+ "learning_rate": 0.0006,
+ "loss": 4.1750640869140625,
+ "step": 2511
+ },
+ {
+ "epoch": 34.8943231441048,
+ "grad_norm": 0.014962567947804928,
+ "learning_rate": 0.0006,
+ "loss": 4.185277938842773,
+ "step": 2512
+ },
+ {
+ "epoch": 34.90829694323144,
+ "grad_norm": 0.014467543922364712,
+ "learning_rate": 0.0006,
+ "loss": 4.1600799560546875,
+ "step": 2513
+ },
+ {
+ "epoch": 34.92227074235808,
+ "grad_norm": 0.01545181218534708,
+ "learning_rate": 0.0006,
+ "loss": 4.096299648284912,
+ "step": 2514
+ },
+ {
+ "epoch": 34.93624454148472,
+ "grad_norm": 0.01605340465903282,
+ "learning_rate": 0.0006,
+ "loss": 4.203323841094971,
+ "step": 2515
+ },
+ {
+ "epoch": 34.95021834061135,
+ "grad_norm": 0.015203815884888172,
+ "learning_rate": 0.0006,
+ "loss": 4.201385021209717,
+ "step": 2516
+ },
+ {
+ "epoch": 34.96419213973799,
+ "grad_norm": 0.015149409882724285,
+ "learning_rate": 0.0006,
+ "loss": 4.088164329528809,
+ "step": 2517
+ },
+ {
+ "epoch": 34.97816593886463,
+ "grad_norm": 0.014885558746755123,
+ "learning_rate": 0.0006,
+ "loss": 4.073911666870117,
+ "step": 2518
+ },
+ {
+ "epoch": 34.992139737991266,
+ "grad_norm": 0.015003419481217861,
+ "learning_rate": 0.0006,
+ "loss": 4.157113552093506,
+ "step": 2519
+ },
+ {
+ "epoch": 35.0,
+ "grad_norm": 0.01641916297376156,
+ "learning_rate": 0.0006,
+ "loss": 3.970052719116211,
+ "step": 2520
+ },
+ {
+ "epoch": 35.0,
+ "eval_loss": 4.641692161560059,
+ "eval_runtime": 56.8836,
+ "eval_samples_per_second": 42.93,
+ "eval_steps_per_second": 1.354,
+ "step": 2520
+ },
+ {
+ "epoch": 35.01397379912664,
+ "grad_norm": 0.015894673764705658,
+ "learning_rate": 0.0006,
+ "loss": 4.104437351226807,
+ "step": 2521
+ },
+ {
+ "epoch": 35.02794759825328,
+ "grad_norm": 0.0173867866396904,
+ "learning_rate": 0.0006,
+ "loss": 4.091726303100586,
+ "step": 2522
+ },
+ {
+ "epoch": 35.041921397379916,
+ "grad_norm": 0.019272832199931145,
+ "learning_rate": 0.0006,
+ "loss": 4.217905044555664,
+ "step": 2523
+ },
+ {
+ "epoch": 35.05589519650655,
+ "grad_norm": 0.021820371970534325,
+ "learning_rate": 0.0006,
+ "loss": 4.111428260803223,
+ "step": 2524
+ },
+ {
+ "epoch": 35.069868995633186,
+ "grad_norm": 0.02340741828083992,
+ "learning_rate": 0.0006,
+ "loss": 4.115771293640137,
+ "step": 2525
+ },
+ {
+ "epoch": 35.083842794759825,
+ "grad_norm": 0.023760691285133362,
+ "learning_rate": 0.0006,
+ "loss": 4.132383346557617,
+ "step": 2526
+ },
+ {
+ "epoch": 35.09781659388646,
+ "grad_norm": 0.02073468267917633,
+ "learning_rate": 0.0006,
+ "loss": 4.272691249847412,
+ "step": 2527
+ },
+ {
+ "epoch": 35.1117903930131,
+ "grad_norm": 0.02197718247771263,
+ "learning_rate": 0.0006,
+ "loss": 4.105586528778076,
+ "step": 2528
+ },
+ {
+ "epoch": 35.12576419213974,
+ "grad_norm": 0.0200254675000906,
+ "learning_rate": 0.0006,
+ "loss": 3.968820095062256,
+ "step": 2529
+ },
+ {
+ "epoch": 35.13973799126637,
+ "grad_norm": 0.01929587684571743,
+ "learning_rate": 0.0006,
+ "loss": 4.045468807220459,
+ "step": 2530
+ },
+ {
+ "epoch": 35.15371179039301,
+ "grad_norm": 0.01886162906885147,
+ "learning_rate": 0.0006,
+ "loss": 4.185571670532227,
+ "step": 2531
+ },
+ {
+ "epoch": 35.16768558951965,
+ "grad_norm": 0.019789839163422585,
+ "learning_rate": 0.0006,
+ "loss": 3.9734551906585693,
+ "step": 2532
+ },
+ {
+ "epoch": 35.18165938864629,
+ "grad_norm": 0.0198142658919096,
+ "learning_rate": 0.0006,
+ "loss": 4.2662506103515625,
+ "step": 2533
+ },
+ {
+ "epoch": 35.19563318777293,
+ "grad_norm": 0.02087012305855751,
+ "learning_rate": 0.0006,
+ "loss": 4.175931930541992,
+ "step": 2534
+ },
+ {
+ "epoch": 35.209606986899566,
+ "grad_norm": 0.020294535905122757,
+ "learning_rate": 0.0006,
+ "loss": 4.060324668884277,
+ "step": 2535
+ },
+ {
+ "epoch": 35.223580786026204,
+ "grad_norm": 0.01908983662724495,
+ "learning_rate": 0.0006,
+ "loss": 4.217939853668213,
+ "step": 2536
+ },
+ {
+ "epoch": 35.237554585152836,
+ "grad_norm": 0.0178892333060503,
+ "learning_rate": 0.0006,
+ "loss": 4.190216541290283,
+ "step": 2537
+ },
+ {
+ "epoch": 35.251528384279474,
+ "grad_norm": 0.017892315983772278,
+ "learning_rate": 0.0006,
+ "loss": 4.067419052124023,
+ "step": 2538
+ },
+ {
+ "epoch": 35.26550218340611,
+ "grad_norm": 0.018556250259280205,
+ "learning_rate": 0.0006,
+ "loss": 4.105958461761475,
+ "step": 2539
+ },
+ {
+ "epoch": 35.27947598253275,
+ "grad_norm": 0.01944422349333763,
+ "learning_rate": 0.0006,
+ "loss": 4.155775547027588,
+ "step": 2540
+ },
+ {
+ "epoch": 35.29344978165939,
+ "grad_norm": 0.017333146184682846,
+ "learning_rate": 0.0006,
+ "loss": 4.244513511657715,
+ "step": 2541
+ },
+ {
+ "epoch": 35.30742358078603,
+ "grad_norm": 0.01639244332909584,
+ "learning_rate": 0.0006,
+ "loss": 4.167815685272217,
+ "step": 2542
+ },
+ {
+ "epoch": 35.32139737991266,
+ "grad_norm": 0.01610143668949604,
+ "learning_rate": 0.0006,
+ "loss": 4.121486663818359,
+ "step": 2543
+ },
+ {
+ "epoch": 35.3353711790393,
+ "grad_norm": 0.01656411588191986,
+ "learning_rate": 0.0006,
+ "loss": 4.1755499839782715,
+ "step": 2544
+ },
+ {
+ "epoch": 35.34934497816594,
+ "grad_norm": 0.015075297094881535,
+ "learning_rate": 0.0006,
+ "loss": 4.102992534637451,
+ "step": 2545
+ },
+ {
+ "epoch": 35.36331877729258,
+ "grad_norm": 0.01538170501589775,
+ "learning_rate": 0.0006,
+ "loss": 4.161603927612305,
+ "step": 2546
+ },
+ {
+ "epoch": 35.377292576419215,
+ "grad_norm": 0.01483312901109457,
+ "learning_rate": 0.0006,
+ "loss": 4.085355758666992,
+ "step": 2547
+ },
+ {
+ "epoch": 35.391266375545854,
+ "grad_norm": 0.014894738793373108,
+ "learning_rate": 0.0006,
+ "loss": 4.083585262298584,
+ "step": 2548
+ },
+ {
+ "epoch": 35.40524017467249,
+ "grad_norm": 0.01475780550390482,
+ "learning_rate": 0.0006,
+ "loss": 4.012645721435547,
+ "step": 2549
+ },
+ {
+ "epoch": 35.419213973799124,
+ "grad_norm": 0.014531653374433517,
+ "learning_rate": 0.0006,
+ "loss": 4.197113990783691,
+ "step": 2550
+ },
+ {
+ "epoch": 35.43318777292576,
+ "grad_norm": 0.015316493809223175,
+ "learning_rate": 0.0006,
+ "loss": 4.227970123291016,
+ "step": 2551
+ },
+ {
+ "epoch": 35.4471615720524,
+ "grad_norm": 0.014685571193695068,
+ "learning_rate": 0.0006,
+ "loss": 4.089415550231934,
+ "step": 2552
+ },
+ {
+ "epoch": 35.46113537117904,
+ "grad_norm": 0.015337035991251469,
+ "learning_rate": 0.0006,
+ "loss": 4.080583572387695,
+ "step": 2553
+ },
+ {
+ "epoch": 35.47510917030568,
+ "grad_norm": 0.016754208132624626,
+ "learning_rate": 0.0006,
+ "loss": 4.163935661315918,
+ "step": 2554
+ },
+ {
+ "epoch": 35.48908296943232,
+ "grad_norm": 0.017535798251628876,
+ "learning_rate": 0.0006,
+ "loss": 4.093027591705322,
+ "step": 2555
+ },
+ {
+ "epoch": 35.50305676855895,
+ "grad_norm": 0.017336975783109665,
+ "learning_rate": 0.0006,
+ "loss": 4.127961158752441,
+ "step": 2556
+ },
+ {
+ "epoch": 35.51703056768559,
+ "grad_norm": 0.016242152079939842,
+ "learning_rate": 0.0006,
+ "loss": 4.155593395233154,
+ "step": 2557
+ },
+ {
+ "epoch": 35.531004366812226,
+ "grad_norm": 0.015121322125196457,
+ "learning_rate": 0.0006,
+ "loss": 3.984004020690918,
+ "step": 2558
+ },
+ {
+ "epoch": 35.544978165938865,
+ "grad_norm": 0.015178884379565716,
+ "learning_rate": 0.0006,
+ "loss": 4.111564636230469,
+ "step": 2559
+ },
+ {
+ "epoch": 35.5589519650655,
+ "grad_norm": 0.014777131378650665,
+ "learning_rate": 0.0006,
+ "loss": 4.041402816772461,
+ "step": 2560
+ },
+ {
+ "epoch": 35.57292576419214,
+ "grad_norm": 0.015197676606476307,
+ "learning_rate": 0.0006,
+ "loss": 4.133706092834473,
+ "step": 2561
+ },
+ {
+ "epoch": 35.58689956331878,
+ "grad_norm": 0.015905972570180893,
+ "learning_rate": 0.0006,
+ "loss": 4.1069135665893555,
+ "step": 2562
+ },
+ {
+ "epoch": 35.60087336244541,
+ "grad_norm": 0.015942877158522606,
+ "learning_rate": 0.0006,
+ "loss": 4.0720624923706055,
+ "step": 2563
+ },
+ {
+ "epoch": 35.61484716157205,
+ "grad_norm": 0.016697583720088005,
+ "learning_rate": 0.0006,
+ "loss": 4.167448043823242,
+ "step": 2564
+ },
+ {
+ "epoch": 35.62882096069869,
+ "grad_norm": 0.015603577718138695,
+ "learning_rate": 0.0006,
+ "loss": 4.127327919006348,
+ "step": 2565
+ },
+ {
+ "epoch": 35.64279475982533,
+ "grad_norm": 0.017153460532426834,
+ "learning_rate": 0.0006,
+ "loss": 4.056543350219727,
+ "step": 2566
+ },
+ {
+ "epoch": 35.65676855895197,
+ "grad_norm": 0.016731958836317062,
+ "learning_rate": 0.0006,
+ "loss": 4.0681843757629395,
+ "step": 2567
+ },
+ {
+ "epoch": 35.670742358078606,
+ "grad_norm": 0.016167359426617622,
+ "learning_rate": 0.0006,
+ "loss": 4.174992084503174,
+ "step": 2568
+ },
+ {
+ "epoch": 35.68471615720524,
+ "grad_norm": 0.01615108922123909,
+ "learning_rate": 0.0006,
+ "loss": 4.1383161544799805,
+ "step": 2569
+ },
+ {
+ "epoch": 35.698689956331876,
+ "grad_norm": 0.015750816091895103,
+ "learning_rate": 0.0006,
+ "loss": 4.031595706939697,
+ "step": 2570
+ },
+ {
+ "epoch": 35.712663755458514,
+ "grad_norm": 0.015416189096868038,
+ "learning_rate": 0.0006,
+ "loss": 4.067498207092285,
+ "step": 2571
+ },
+ {
+ "epoch": 35.72663755458515,
+ "grad_norm": 0.01570792682468891,
+ "learning_rate": 0.0006,
+ "loss": 4.1338300704956055,
+ "step": 2572
+ },
+ {
+ "epoch": 35.74061135371179,
+ "grad_norm": 0.016367752104997635,
+ "learning_rate": 0.0006,
+ "loss": 4.152523040771484,
+ "step": 2573
+ },
+ {
+ "epoch": 35.75458515283843,
+ "grad_norm": 0.016349803656339645,
+ "learning_rate": 0.0006,
+ "loss": 4.101985454559326,
+ "step": 2574
+ },
+ {
+ "epoch": 35.76855895196506,
+ "grad_norm": 0.01663660630583763,
+ "learning_rate": 0.0006,
+ "loss": 4.190130710601807,
+ "step": 2575
+ },
+ {
+ "epoch": 35.7825327510917,
+ "grad_norm": 0.015599170699715614,
+ "learning_rate": 0.0006,
+ "loss": 4.0792436599731445,
+ "step": 2576
+ },
+ {
+ "epoch": 35.79650655021834,
+ "grad_norm": 0.015472427010536194,
+ "learning_rate": 0.0006,
+ "loss": 4.186446666717529,
+ "step": 2577
+ },
+ {
+ "epoch": 35.81048034934498,
+ "grad_norm": 0.015520439483225346,
+ "learning_rate": 0.0006,
+ "loss": 4.246635437011719,
+ "step": 2578
+ },
+ {
+ "epoch": 35.82445414847162,
+ "grad_norm": 0.016998281702399254,
+ "learning_rate": 0.0006,
+ "loss": 4.13435173034668,
+ "step": 2579
+ },
+ {
+ "epoch": 35.838427947598255,
+ "grad_norm": 0.016290361061692238,
+ "learning_rate": 0.0006,
+ "loss": 4.107428073883057,
+ "step": 2580
+ },
+ {
+ "epoch": 35.852401746724894,
+ "grad_norm": 0.0165516659617424,
+ "learning_rate": 0.0006,
+ "loss": 4.1612138748168945,
+ "step": 2581
+ },
+ {
+ "epoch": 35.866375545851525,
+ "grad_norm": 0.01736091636121273,
+ "learning_rate": 0.0006,
+ "loss": 4.1072869300842285,
+ "step": 2582
+ },
+ {
+ "epoch": 35.880349344978164,
+ "grad_norm": 0.01787031814455986,
+ "learning_rate": 0.0006,
+ "loss": 4.1076812744140625,
+ "step": 2583
+ },
+ {
+ "epoch": 35.8943231441048,
+ "grad_norm": 0.015437985770404339,
+ "learning_rate": 0.0006,
+ "loss": 4.097488880157471,
+ "step": 2584
+ },
+ {
+ "epoch": 35.90829694323144,
+ "grad_norm": 0.016306741163134575,
+ "learning_rate": 0.0006,
+ "loss": 4.105017185211182,
+ "step": 2585
+ },
+ {
+ "epoch": 35.92227074235808,
+ "grad_norm": 0.015723245218396187,
+ "learning_rate": 0.0006,
+ "loss": 4.205283164978027,
+ "step": 2586
+ },
+ {
+ "epoch": 35.93624454148472,
+ "grad_norm": 0.014626456424593925,
+ "learning_rate": 0.0006,
+ "loss": 4.086552619934082,
+ "step": 2587
+ },
+ {
+ "epoch": 35.95021834061135,
+ "grad_norm": 0.014348454773426056,
+ "learning_rate": 0.0006,
+ "loss": 4.171329975128174,
+ "step": 2588
+ },
+ {
+ "epoch": 35.96419213973799,
+ "grad_norm": 0.015598480589687824,
+ "learning_rate": 0.0006,
+ "loss": 4.140815258026123,
+ "step": 2589
+ },
+ {
+ "epoch": 35.97816593886463,
+ "grad_norm": 0.017126062884926796,
+ "learning_rate": 0.0006,
+ "loss": 4.157176971435547,
+ "step": 2590
+ },
+ {
+ "epoch": 35.992139737991266,
+ "grad_norm": 0.01860199309885502,
+ "learning_rate": 0.0006,
+ "loss": 4.252880096435547,
+ "step": 2591
+ },
+ {
+ "epoch": 36.0,
+ "grad_norm": 0.021780220791697502,
+ "learning_rate": 0.0006,
+ "loss": 4.249861717224121,
+ "step": 2592
+ },
+ {
+ "epoch": 36.0,
+ "eval_loss": 4.606466293334961,
+ "eval_runtime": 56.6217,
+ "eval_samples_per_second": 43.128,
+ "eval_steps_per_second": 1.36,
+ "step": 2592
+ },
+ {
+ "epoch": 36.01397379912664,
+ "grad_norm": 0.020892828702926636,
+ "learning_rate": 0.0006,
+ "loss": 4.102055549621582,
+ "step": 2593
+ },
+ {
+ "epoch": 36.02794759825328,
+ "grad_norm": 0.020306255668401718,
+ "learning_rate": 0.0006,
+ "loss": 4.184308052062988,
+ "step": 2594
+ },
+ {
+ "epoch": 36.041921397379916,
+ "grad_norm": 0.02025097794830799,
+ "learning_rate": 0.0006,
+ "loss": 3.9691858291625977,
+ "step": 2595
+ },
+ {
+ "epoch": 36.05589519650655,
+ "grad_norm": 0.020254382863640785,
+ "learning_rate": 0.0006,
+ "loss": 4.02957820892334,
+ "step": 2596
+ },
+ {
+ "epoch": 36.069868995633186,
+ "grad_norm": 0.019307196140289307,
+ "learning_rate": 0.0006,
+ "loss": 4.033181667327881,
+ "step": 2597
+ },
+ {
+ "epoch": 36.083842794759825,
+ "grad_norm": 0.0198476854711771,
+ "learning_rate": 0.0006,
+ "loss": 4.0816874504089355,
+ "step": 2598
+ },
+ {
+ "epoch": 36.09781659388646,
+ "grad_norm": 0.019916515797376633,
+ "learning_rate": 0.0006,
+ "loss": 4.164889812469482,
+ "step": 2599
+ },
+ {
+ "epoch": 36.1117903930131,
+ "grad_norm": 0.0189144778996706,
+ "learning_rate": 0.0006,
+ "loss": 4.038427352905273,
+ "step": 2600
+ },
+ {
+ "epoch": 36.12576419213974,
+ "grad_norm": 0.01754768192768097,
+ "learning_rate": 0.0006,
+ "loss": 4.010427474975586,
+ "step": 2601
+ },
+ {
+ "epoch": 36.13973799126637,
+ "grad_norm": 0.016862990334630013,
+ "learning_rate": 0.0006,
+ "loss": 4.035970211029053,
+ "step": 2602
+ },
+ {
+ "epoch": 36.15371179039301,
+ "grad_norm": 0.019616752862930298,
+ "learning_rate": 0.0006,
+ "loss": 4.079898834228516,
+ "step": 2603
+ },
+ {
+ "epoch": 36.16768558951965,
+ "grad_norm": 0.020131465047597885,
+ "learning_rate": 0.0006,
+ "loss": 4.144684791564941,
+ "step": 2604
+ },
+ {
+ "epoch": 36.18165938864629,
+ "grad_norm": 0.02072521299123764,
+ "learning_rate": 0.0006,
+ "loss": 4.013660430908203,
+ "step": 2605
+ },
+ {
+ "epoch": 36.19563318777293,
+ "grad_norm": 0.01924615167081356,
+ "learning_rate": 0.0006,
+ "loss": 4.071398735046387,
+ "step": 2606
+ },
+ {
+ "epoch": 36.209606986899566,
+ "grad_norm": 0.02013954520225525,
+ "learning_rate": 0.0006,
+ "loss": 4.127094268798828,
+ "step": 2607
+ },
+ {
+ "epoch": 36.223580786026204,
+ "grad_norm": 0.02042008936405182,
+ "learning_rate": 0.0006,
+ "loss": 4.154261589050293,
+ "step": 2608
+ },
+ {
+ "epoch": 36.237554585152836,
+ "grad_norm": 0.021998876705765724,
+ "learning_rate": 0.0006,
+ "loss": 4.204150199890137,
+ "step": 2609
+ },
+ {
+ "epoch": 36.251528384279474,
+ "grad_norm": 0.022946683689951897,
+ "learning_rate": 0.0006,
+ "loss": 4.013632774353027,
+ "step": 2610
+ },
+ {
+ "epoch": 36.26550218340611,
+ "grad_norm": 0.02061830274760723,
+ "learning_rate": 0.0006,
+ "loss": 4.171084403991699,
+ "step": 2611
+ },
+ {
+ "epoch": 36.27947598253275,
+ "grad_norm": 0.01883152686059475,
+ "learning_rate": 0.0006,
+ "loss": 4.149848461151123,
+ "step": 2612
+ },
+ {
+ "epoch": 36.29344978165939,
+ "grad_norm": 0.019945282489061356,
+ "learning_rate": 0.0006,
+ "loss": 4.174656867980957,
+ "step": 2613
+ },
+ {
+ "epoch": 36.30742358078603,
+ "grad_norm": 0.020751051604747772,
+ "learning_rate": 0.0006,
+ "loss": 4.080747127532959,
+ "step": 2614
+ },
+ {
+ "epoch": 36.32139737991266,
+ "grad_norm": 0.01879691146314144,
+ "learning_rate": 0.0006,
+ "loss": 4.057356357574463,
+ "step": 2615
+ },
+ {
+ "epoch": 36.3353711790393,
+ "grad_norm": 0.019037071615457535,
+ "learning_rate": 0.0006,
+ "loss": 4.204607009887695,
+ "step": 2616
+ },
+ {
+ "epoch": 36.34934497816594,
+ "grad_norm": 0.01872078701853752,
+ "learning_rate": 0.0006,
+ "loss": 4.050586700439453,
+ "step": 2617
+ },
+ {
+ "epoch": 36.36331877729258,
+ "grad_norm": 0.018092811107635498,
+ "learning_rate": 0.0006,
+ "loss": 4.102832794189453,
+ "step": 2618
+ },
+ {
+ "epoch": 36.377292576419215,
+ "grad_norm": 0.017010487616062164,
+ "learning_rate": 0.0006,
+ "loss": 4.191788673400879,
+ "step": 2619
+ },
+ {
+ "epoch": 36.391266375545854,
+ "grad_norm": 0.01691143773496151,
+ "learning_rate": 0.0006,
+ "loss": 4.160456657409668,
+ "step": 2620
+ },
+ {
+ "epoch": 36.40524017467249,
+ "grad_norm": 0.017046816647052765,
+ "learning_rate": 0.0006,
+ "loss": 4.073828220367432,
+ "step": 2621
+ },
+ {
+ "epoch": 36.419213973799124,
+ "grad_norm": 0.01707981713116169,
+ "learning_rate": 0.0006,
+ "loss": 4.0980095863342285,
+ "step": 2622
+ },
+ {
+ "epoch": 36.43318777292576,
+ "grad_norm": 0.01630452275276184,
+ "learning_rate": 0.0006,
+ "loss": 4.101321220397949,
+ "step": 2623
+ },
+ {
+ "epoch": 36.4471615720524,
+ "grad_norm": 0.01576915942132473,
+ "learning_rate": 0.0006,
+ "loss": 4.158249855041504,
+ "step": 2624
+ },
+ {
+ "epoch": 36.46113537117904,
+ "grad_norm": 0.016179712489247322,
+ "learning_rate": 0.0006,
+ "loss": 4.077852249145508,
+ "step": 2625
+ },
+ {
+ "epoch": 36.47510917030568,
+ "grad_norm": 0.01609169878065586,
+ "learning_rate": 0.0006,
+ "loss": 4.187317848205566,
+ "step": 2626
+ },
+ {
+ "epoch": 36.48908296943232,
+ "grad_norm": 0.017148254439234734,
+ "learning_rate": 0.0006,
+ "loss": 4.109071731567383,
+ "step": 2627
+ },
+ {
+ "epoch": 36.50305676855895,
+ "grad_norm": 0.015842624008655548,
+ "learning_rate": 0.0006,
+ "loss": 4.145096302032471,
+ "step": 2628
+ },
+ {
+ "epoch": 36.51703056768559,
+ "grad_norm": 0.014906194061040878,
+ "learning_rate": 0.0006,
+ "loss": 4.028440475463867,
+ "step": 2629
+ },
+ {
+ "epoch": 36.531004366812226,
+ "grad_norm": 0.01485038734972477,
+ "learning_rate": 0.0006,
+ "loss": 4.175201416015625,
+ "step": 2630
+ },
+ {
+ "epoch": 36.544978165938865,
+ "grad_norm": 0.015834150835871696,
+ "learning_rate": 0.0006,
+ "loss": 4.076434135437012,
+ "step": 2631
+ },
+ {
+ "epoch": 36.5589519650655,
+ "grad_norm": 0.015411358326673508,
+ "learning_rate": 0.0006,
+ "loss": 4.03912878036499,
+ "step": 2632
+ },
+ {
+ "epoch": 36.57292576419214,
+ "grad_norm": 0.014762846753001213,
+ "learning_rate": 0.0006,
+ "loss": 4.019827365875244,
+ "step": 2633
+ },
+ {
+ "epoch": 36.58689956331878,
+ "grad_norm": 0.015269497409462929,
+ "learning_rate": 0.0006,
+ "loss": 4.116957187652588,
+ "step": 2634
+ },
+ {
+ "epoch": 36.60087336244541,
+ "grad_norm": 0.015171617269515991,
+ "learning_rate": 0.0006,
+ "loss": 4.2069926261901855,
+ "step": 2635
+ },
+ {
+ "epoch": 36.61484716157205,
+ "grad_norm": 0.014770902693271637,
+ "learning_rate": 0.0006,
+ "loss": 4.092021942138672,
+ "step": 2636
+ },
+ {
+ "epoch": 36.62882096069869,
+ "grad_norm": 0.014957531355321407,
+ "learning_rate": 0.0006,
+ "loss": 4.16819953918457,
+ "step": 2637
+ },
+ {
+ "epoch": 36.64279475982533,
+ "grad_norm": 0.014774451963603497,
+ "learning_rate": 0.0006,
+ "loss": 4.046073913574219,
+ "step": 2638
+ },
+ {
+ "epoch": 36.65676855895197,
+ "grad_norm": 0.014320445246994495,
+ "learning_rate": 0.0006,
+ "loss": 3.977792501449585,
+ "step": 2639
+ },
+ {
+ "epoch": 36.670742358078606,
+ "grad_norm": 0.01413130946457386,
+ "learning_rate": 0.0006,
+ "loss": 4.130649566650391,
+ "step": 2640
+ },
+ {
+ "epoch": 36.68471615720524,
+ "grad_norm": 0.014336307533085346,
+ "learning_rate": 0.0006,
+ "loss": 4.161357402801514,
+ "step": 2641
+ },
+ {
+ "epoch": 36.698689956331876,
+ "grad_norm": 0.014516931027173996,
+ "learning_rate": 0.0006,
+ "loss": 4.14248514175415,
+ "step": 2642
+ },
+ {
+ "epoch": 36.712663755458514,
+ "grad_norm": 0.014377126470208168,
+ "learning_rate": 0.0006,
+ "loss": 4.062711715698242,
+ "step": 2643
+ },
+ {
+ "epoch": 36.72663755458515,
+ "grad_norm": 0.014332668855786324,
+ "learning_rate": 0.0006,
+ "loss": 4.06423807144165,
+ "step": 2644
+ },
+ {
+ "epoch": 36.74061135371179,
+ "grad_norm": 0.014030571095645428,
+ "learning_rate": 0.0006,
+ "loss": 4.178613662719727,
+ "step": 2645
+ },
+ {
+ "epoch": 36.75458515283843,
+ "grad_norm": 0.014376776292920113,
+ "learning_rate": 0.0006,
+ "loss": 4.073130130767822,
+ "step": 2646
+ },
+ {
+ "epoch": 36.76855895196506,
+ "grad_norm": 0.013827716931700706,
+ "learning_rate": 0.0006,
+ "loss": 4.043869972229004,
+ "step": 2647
+ },
+ {
+ "epoch": 36.7825327510917,
+ "grad_norm": 0.015548762865364552,
+ "learning_rate": 0.0006,
+ "loss": 4.134049415588379,
+ "step": 2648
+ },
+ {
+ "epoch": 36.79650655021834,
+ "grad_norm": 0.017041970044374466,
+ "learning_rate": 0.0006,
+ "loss": 4.205024242401123,
+ "step": 2649
+ },
+ {
+ "epoch": 36.81048034934498,
+ "grad_norm": 0.017638275399804115,
+ "learning_rate": 0.0006,
+ "loss": 4.199924468994141,
+ "step": 2650
+ },
+ {
+ "epoch": 36.82445414847162,
+ "grad_norm": 0.017771627753973007,
+ "learning_rate": 0.0006,
+ "loss": 4.15706729888916,
+ "step": 2651
+ },
+ {
+ "epoch": 36.838427947598255,
+ "grad_norm": 0.015933562070131302,
+ "learning_rate": 0.0006,
+ "loss": 4.220742225646973,
+ "step": 2652
+ },
+ {
+ "epoch": 36.852401746724894,
+ "grad_norm": 0.01668144389986992,
+ "learning_rate": 0.0006,
+ "loss": 4.161426544189453,
+ "step": 2653
+ },
+ {
+ "epoch": 36.866375545851525,
+ "grad_norm": 0.018014680594205856,
+ "learning_rate": 0.0006,
+ "loss": 4.183046817779541,
+ "step": 2654
+ },
+ {
+ "epoch": 36.880349344978164,
+ "grad_norm": 0.018808670341968536,
+ "learning_rate": 0.0006,
+ "loss": 4.105293273925781,
+ "step": 2655
+ },
+ {
+ "epoch": 36.8943231441048,
+ "grad_norm": 0.018003476783633232,
+ "learning_rate": 0.0006,
+ "loss": 4.103171348571777,
+ "step": 2656
+ },
+ {
+ "epoch": 36.90829694323144,
+ "grad_norm": 0.016865914687514305,
+ "learning_rate": 0.0006,
+ "loss": 4.125634670257568,
+ "step": 2657
+ },
+ {
+ "epoch": 36.92227074235808,
+ "grad_norm": 0.018371153622865677,
+ "learning_rate": 0.0006,
+ "loss": 4.192461013793945,
+ "step": 2658
+ },
+ {
+ "epoch": 36.93624454148472,
+ "grad_norm": 0.019045377150177956,
+ "learning_rate": 0.0006,
+ "loss": 4.198105812072754,
+ "step": 2659
+ },
+ {
+ "epoch": 36.95021834061135,
+ "grad_norm": 0.01594168320298195,
+ "learning_rate": 0.0006,
+ "loss": 4.169788360595703,
+ "step": 2660
+ },
+ {
+ "epoch": 36.96419213973799,
+ "grad_norm": 0.015645820647478104,
+ "learning_rate": 0.0006,
+ "loss": 4.047277450561523,
+ "step": 2661
+ },
+ {
+ "epoch": 36.97816593886463,
+ "grad_norm": 0.016313156113028526,
+ "learning_rate": 0.0006,
+ "loss": 4.087392807006836,
+ "step": 2662
+ },
+ {
+ "epoch": 36.992139737991266,
+ "grad_norm": 0.017050625756382942,
+ "learning_rate": 0.0006,
+ "loss": 4.0908002853393555,
+ "step": 2663
+ },
+ {
+ "epoch": 37.0,
+ "grad_norm": 0.017730968073010445,
+ "learning_rate": 0.0006,
+ "loss": 4.110957622528076,
+ "step": 2664
+ },
+ {
+ "epoch": 37.0,
+ "eval_loss": 4.563793182373047,
+ "eval_runtime": 56.5922,
+ "eval_samples_per_second": 43.151,
+ "eval_steps_per_second": 1.361,
+ "step": 2664
+ },
+ {
+ "epoch": 37.01397379912664,
+ "grad_norm": 0.015687910839915276,
+ "learning_rate": 0.0006,
+ "loss": 4.143589019775391,
+ "step": 2665
+ },
+ {
+ "epoch": 37.02794759825328,
+ "grad_norm": 0.017451366409659386,
+ "learning_rate": 0.0006,
+ "loss": 4.0326666831970215,
+ "step": 2666
+ },
+ {
+ "epoch": 37.041921397379916,
+ "grad_norm": 0.01909538544714451,
+ "learning_rate": 0.0006,
+ "loss": 3.9706249237060547,
+ "step": 2667
+ },
+ {
+ "epoch": 37.05589519650655,
+ "grad_norm": 0.01988379657268524,
+ "learning_rate": 0.0006,
+ "loss": 3.9531009197235107,
+ "step": 2668
+ },
+ {
+ "epoch": 37.069868995633186,
+ "grad_norm": 0.02035115472972393,
+ "learning_rate": 0.0006,
+ "loss": 3.962214469909668,
+ "step": 2669
+ },
+ {
+ "epoch": 37.083842794759825,
+ "grad_norm": 0.01952001079916954,
+ "learning_rate": 0.0006,
+ "loss": 4.182262897491455,
+ "step": 2670
+ },
+ {
+ "epoch": 37.09781659388646,
+ "grad_norm": 0.01854325644671917,
+ "learning_rate": 0.0006,
+ "loss": 4.127663612365723,
+ "step": 2671
+ },
+ {
+ "epoch": 37.1117903930131,
+ "grad_norm": 0.017757616937160492,
+ "learning_rate": 0.0006,
+ "loss": 4.100804805755615,
+ "step": 2672
+ },
+ {
+ "epoch": 37.12576419213974,
+ "grad_norm": 0.016712479293346405,
+ "learning_rate": 0.0006,
+ "loss": 4.026284217834473,
+ "step": 2673
+ },
+ {
+ "epoch": 37.13973799126637,
+ "grad_norm": 0.0168139711022377,
+ "learning_rate": 0.0006,
+ "loss": 4.183882713317871,
+ "step": 2674
+ },
+ {
+ "epoch": 37.15371179039301,
+ "grad_norm": 0.017060423269867897,
+ "learning_rate": 0.0006,
+ "loss": 4.131832599639893,
+ "step": 2675
+ },
+ {
+ "epoch": 37.16768558951965,
+ "grad_norm": 0.01726631447672844,
+ "learning_rate": 0.0006,
+ "loss": 3.9156057834625244,
+ "step": 2676
+ },
+ {
+ "epoch": 37.18165938864629,
+ "grad_norm": 0.016348622739315033,
+ "learning_rate": 0.0006,
+ "loss": 4.1731367111206055,
+ "step": 2677
+ },
+ {
+ "epoch": 37.19563318777293,
+ "grad_norm": 0.016640059649944305,
+ "learning_rate": 0.0006,
+ "loss": 4.10997200012207,
+ "step": 2678
+ },
+ {
+ "epoch": 37.209606986899566,
+ "grad_norm": 0.01601342298090458,
+ "learning_rate": 0.0006,
+ "loss": 4.081145286560059,
+ "step": 2679
+ },
+ {
+ "epoch": 37.223580786026204,
+ "grad_norm": 0.01735631190240383,
+ "learning_rate": 0.0006,
+ "loss": 4.202108860015869,
+ "step": 2680
+ },
+ {
+ "epoch": 37.237554585152836,
+ "grad_norm": 0.019164685159921646,
+ "learning_rate": 0.0006,
+ "loss": 4.104109764099121,
+ "step": 2681
+ },
+ {
+ "epoch": 37.251528384279474,
+ "grad_norm": 0.01881759613752365,
+ "learning_rate": 0.0006,
+ "loss": 4.145532131195068,
+ "step": 2682
+ },
+ {
+ "epoch": 37.26550218340611,
+ "grad_norm": 0.017656555399298668,
+ "learning_rate": 0.0006,
+ "loss": 4.012782096862793,
+ "step": 2683
+ },
+ {
+ "epoch": 37.27947598253275,
+ "grad_norm": 0.01624823734164238,
+ "learning_rate": 0.0006,
+ "loss": 4.001727104187012,
+ "step": 2684
+ },
+ {
+ "epoch": 37.29344978165939,
+ "grad_norm": 0.016973139718174934,
+ "learning_rate": 0.0006,
+ "loss": 4.095399856567383,
+ "step": 2685
+ },
+ {
+ "epoch": 37.30742358078603,
+ "grad_norm": 0.015503671951591969,
+ "learning_rate": 0.0006,
+ "loss": 4.034431457519531,
+ "step": 2686
+ },
+ {
+ "epoch": 37.32139737991266,
+ "grad_norm": 0.015370250679552555,
+ "learning_rate": 0.0006,
+ "loss": 4.075836181640625,
+ "step": 2687
+ },
+ {
+ "epoch": 37.3353711790393,
+ "grad_norm": 0.01564021222293377,
+ "learning_rate": 0.0006,
+ "loss": 4.125202178955078,
+ "step": 2688
+ },
+ {
+ "epoch": 37.34934497816594,
+ "grad_norm": 0.016717975959181786,
+ "learning_rate": 0.0006,
+ "loss": 4.153017044067383,
+ "step": 2689
+ },
+ {
+ "epoch": 37.36331877729258,
+ "grad_norm": 0.015263617038726807,
+ "learning_rate": 0.0006,
+ "loss": 4.021814823150635,
+ "step": 2690
+ },
+ {
+ "epoch": 37.377292576419215,
+ "grad_norm": 0.016497142612934113,
+ "learning_rate": 0.0006,
+ "loss": 4.069095611572266,
+ "step": 2691
+ },
+ {
+ "epoch": 37.391266375545854,
+ "grad_norm": 0.01669355481863022,
+ "learning_rate": 0.0006,
+ "loss": 4.174587726593018,
+ "step": 2692
+ },
+ {
+ "epoch": 37.40524017467249,
+ "grad_norm": 0.019695773720741272,
+ "learning_rate": 0.0006,
+ "loss": 4.0073041915893555,
+ "step": 2693
+ },
+ {
+ "epoch": 37.419213973799124,
+ "grad_norm": 0.023720644414424896,
+ "learning_rate": 0.0006,
+ "loss": 4.111948013305664,
+ "step": 2694
+ },
+ {
+ "epoch": 37.43318777292576,
+ "grad_norm": 0.02399350143969059,
+ "learning_rate": 0.0006,
+ "loss": 4.081164360046387,
+ "step": 2695
+ },
+ {
+ "epoch": 37.4471615720524,
+ "grad_norm": 0.019630271941423416,
+ "learning_rate": 0.0006,
+ "loss": 4.065731048583984,
+ "step": 2696
+ },
+ {
+ "epoch": 37.46113537117904,
+ "grad_norm": 0.019679008051753044,
+ "learning_rate": 0.0006,
+ "loss": 3.993852376937866,
+ "step": 2697
+ },
+ {
+ "epoch": 37.47510917030568,
+ "grad_norm": 0.020597299560904503,
+ "learning_rate": 0.0006,
+ "loss": 4.275555610656738,
+ "step": 2698
+ },
+ {
+ "epoch": 37.48908296943232,
+ "grad_norm": 0.01983889564871788,
+ "learning_rate": 0.0006,
+ "loss": 4.254708290100098,
+ "step": 2699
+ },
+ {
+ "epoch": 37.50305676855895,
+ "grad_norm": 0.019150668755173683,
+ "learning_rate": 0.0006,
+ "loss": 4.097885608673096,
+ "step": 2700
+ },
+ {
+ "epoch": 37.51703056768559,
+ "grad_norm": 0.018847715109586716,
+ "learning_rate": 0.0006,
+ "loss": 4.054081439971924,
+ "step": 2701
+ },
+ {
+ "epoch": 37.531004366812226,
+ "grad_norm": 0.019607288762927055,
+ "learning_rate": 0.0006,
+ "loss": 3.9891440868377686,
+ "step": 2702
+ },
+ {
+ "epoch": 37.544978165938865,
+ "grad_norm": 0.018863150849938393,
+ "learning_rate": 0.0006,
+ "loss": 4.084877014160156,
+ "step": 2703
+ },
+ {
+ "epoch": 37.5589519650655,
+ "grad_norm": 0.017340239137411118,
+ "learning_rate": 0.0006,
+ "loss": 4.113290786743164,
+ "step": 2704
+ },
+ {
+ "epoch": 37.57292576419214,
+ "grad_norm": 0.015628693625330925,
+ "learning_rate": 0.0006,
+ "loss": 4.1483001708984375,
+ "step": 2705
+ },
+ {
+ "epoch": 37.58689956331878,
+ "grad_norm": 0.016493283212184906,
+ "learning_rate": 0.0006,
+ "loss": 4.1040143966674805,
+ "step": 2706
+ },
+ {
+ "epoch": 37.60087336244541,
+ "grad_norm": 0.016732443124055862,
+ "learning_rate": 0.0006,
+ "loss": 4.07205057144165,
+ "step": 2707
+ },
+ {
+ "epoch": 37.61484716157205,
+ "grad_norm": 0.017176156863570213,
+ "learning_rate": 0.0006,
+ "loss": 4.1279754638671875,
+ "step": 2708
+ },
+ {
+ "epoch": 37.62882096069869,
+ "grad_norm": 0.015730608254671097,
+ "learning_rate": 0.0006,
+ "loss": 4.086658000946045,
+ "step": 2709
+ },
+ {
+ "epoch": 37.64279475982533,
+ "grad_norm": 0.016111090779304504,
+ "learning_rate": 0.0006,
+ "loss": 4.064335823059082,
+ "step": 2710
+ },
+ {
+ "epoch": 37.65676855895197,
+ "grad_norm": 0.01769379712641239,
+ "learning_rate": 0.0006,
+ "loss": 4.116047382354736,
+ "step": 2711
+ },
+ {
+ "epoch": 37.670742358078606,
+ "grad_norm": 0.01683102548122406,
+ "learning_rate": 0.0006,
+ "loss": 4.095986843109131,
+ "step": 2712
+ },
+ {
+ "epoch": 37.68471615720524,
+ "grad_norm": 0.015726450830698013,
+ "learning_rate": 0.0006,
+ "loss": 4.10775089263916,
+ "step": 2713
+ },
+ {
+ "epoch": 37.698689956331876,
+ "grad_norm": 0.014760235324501991,
+ "learning_rate": 0.0006,
+ "loss": 4.1234893798828125,
+ "step": 2714
+ },
+ {
+ "epoch": 37.712663755458514,
+ "grad_norm": 0.014933438040316105,
+ "learning_rate": 0.0006,
+ "loss": 4.079200744628906,
+ "step": 2715
+ },
+ {
+ "epoch": 37.72663755458515,
+ "grad_norm": 0.015723997727036476,
+ "learning_rate": 0.0006,
+ "loss": 4.07865047454834,
+ "step": 2716
+ },
+ {
+ "epoch": 37.74061135371179,
+ "grad_norm": 0.014539001509547234,
+ "learning_rate": 0.0006,
+ "loss": 4.036242485046387,
+ "step": 2717
+ },
+ {
+ "epoch": 37.75458515283843,
+ "grad_norm": 0.014844979159533978,
+ "learning_rate": 0.0006,
+ "loss": 4.1051483154296875,
+ "step": 2718
+ },
+ {
+ "epoch": 37.76855895196506,
+ "grad_norm": 0.015939844772219658,
+ "learning_rate": 0.0006,
+ "loss": 4.017253875732422,
+ "step": 2719
+ },
+ {
+ "epoch": 37.7825327510917,
+ "grad_norm": 0.014771555550396442,
+ "learning_rate": 0.0006,
+ "loss": 4.0694661140441895,
+ "step": 2720
+ },
+ {
+ "epoch": 37.79650655021834,
+ "grad_norm": 0.015777921304106712,
+ "learning_rate": 0.0006,
+ "loss": 3.947587490081787,
+ "step": 2721
+ },
+ {
+ "epoch": 37.81048034934498,
+ "grad_norm": 0.017451291903853416,
+ "learning_rate": 0.0006,
+ "loss": 4.093625068664551,
+ "step": 2722
+ },
+ {
+ "epoch": 37.82445414847162,
+ "grad_norm": 0.016829246655106544,
+ "learning_rate": 0.0006,
+ "loss": 4.09030818939209,
+ "step": 2723
+ },
+ {
+ "epoch": 37.838427947598255,
+ "grad_norm": 0.016984397545456886,
+ "learning_rate": 0.0006,
+ "loss": 4.140626907348633,
+ "step": 2724
+ },
+ {
+ "epoch": 37.852401746724894,
+ "grad_norm": 0.01732603646814823,
+ "learning_rate": 0.0006,
+ "loss": 4.128276824951172,
+ "step": 2725
+ },
+ {
+ "epoch": 37.866375545851525,
+ "grad_norm": 0.01744619384407997,
+ "learning_rate": 0.0006,
+ "loss": 4.103130340576172,
+ "step": 2726
+ },
+ {
+ "epoch": 37.880349344978164,
+ "grad_norm": 0.016331320628523827,
+ "learning_rate": 0.0006,
+ "loss": 4.037951469421387,
+ "step": 2727
+ },
+ {
+ "epoch": 37.8943231441048,
+ "grad_norm": 0.015919430181384087,
+ "learning_rate": 0.0006,
+ "loss": 4.121513366699219,
+ "step": 2728
+ },
+ {
+ "epoch": 37.90829694323144,
+ "grad_norm": 0.016153695061802864,
+ "learning_rate": 0.0006,
+ "loss": 4.099384307861328,
+ "step": 2729
+ },
+ {
+ "epoch": 37.92227074235808,
+ "grad_norm": 0.01628800667822361,
+ "learning_rate": 0.0006,
+ "loss": 4.092325687408447,
+ "step": 2730
+ },
+ {
+ "epoch": 37.93624454148472,
+ "grad_norm": 0.017285684123635292,
+ "learning_rate": 0.0006,
+ "loss": 3.9828906059265137,
+ "step": 2731
+ },
+ {
+ "epoch": 37.95021834061135,
+ "grad_norm": 0.016047554090619087,
+ "learning_rate": 0.0006,
+ "loss": 4.100419044494629,
+ "step": 2732
+ },
+ {
+ "epoch": 37.96419213973799,
+ "grad_norm": 0.01562986709177494,
+ "learning_rate": 0.0006,
+ "loss": 4.065832614898682,
+ "step": 2733
+ },
+ {
+ "epoch": 37.97816593886463,
+ "grad_norm": 0.01677163504064083,
+ "learning_rate": 0.0006,
+ "loss": 4.077518463134766,
+ "step": 2734
+ },
+ {
+ "epoch": 37.992139737991266,
+ "grad_norm": 0.01755492016673088,
+ "learning_rate": 0.0006,
+ "loss": 4.1076154708862305,
+ "step": 2735
+ },
+ {
+ "epoch": 38.0,
+ "grad_norm": 0.01929762214422226,
+ "learning_rate": 0.0006,
+ "loss": 4.09603214263916,
+ "step": 2736
+ },
+ {
+ "epoch": 38.0,
+ "eval_loss": 4.582442283630371,
+ "eval_runtime": 62.8649,
+ "eval_samples_per_second": 38.845,
+ "eval_steps_per_second": 1.225,
+ "step": 2736
+ },
+ {
+ "epoch": 38.01397379912664,
+ "grad_norm": 0.019129136577248573,
+ "learning_rate": 0.0006,
+ "loss": 4.095005989074707,
+ "step": 2737
+ },
+ {
+ "epoch": 38.02794759825328,
+ "grad_norm": 0.019167408347129822,
+ "learning_rate": 0.0006,
+ "loss": 4.054426670074463,
+ "step": 2738
+ },
+ {
+ "epoch": 38.041921397379916,
+ "grad_norm": 0.01789460889995098,
+ "learning_rate": 0.0006,
+ "loss": 4.0434417724609375,
+ "step": 2739
+ },
+ {
+ "epoch": 38.05589519650655,
+ "grad_norm": 0.015947027131915092,
+ "learning_rate": 0.0006,
+ "loss": 4.074516296386719,
+ "step": 2740
+ },
+ {
+ "epoch": 38.069868995633186,
+ "grad_norm": 0.016667252406477928,
+ "learning_rate": 0.0006,
+ "loss": 4.108545303344727,
+ "step": 2741
+ },
+ {
+ "epoch": 38.083842794759825,
+ "grad_norm": 0.016014399006962776,
+ "learning_rate": 0.0006,
+ "loss": 4.026925086975098,
+ "step": 2742
+ },
+ {
+ "epoch": 38.09781659388646,
+ "grad_norm": 0.015397695824503899,
+ "learning_rate": 0.0006,
+ "loss": 4.163970470428467,
+ "step": 2743
+ },
+ {
+ "epoch": 38.1117903930131,
+ "grad_norm": 0.01563289947807789,
+ "learning_rate": 0.0006,
+ "loss": 4.116578578948975,
+ "step": 2744
+ },
+ {
+ "epoch": 38.12576419213974,
+ "grad_norm": 0.015512318350374699,
+ "learning_rate": 0.0006,
+ "loss": 4.096311569213867,
+ "step": 2745
+ },
+ {
+ "epoch": 38.13973799126637,
+ "grad_norm": 0.015426642261445522,
+ "learning_rate": 0.0006,
+ "loss": 4.017065048217773,
+ "step": 2746
+ },
+ {
+ "epoch": 38.15371179039301,
+ "grad_norm": 0.015115435235202312,
+ "learning_rate": 0.0006,
+ "loss": 4.005434989929199,
+ "step": 2747
+ },
+ {
+ "epoch": 38.16768558951965,
+ "grad_norm": 0.015235894359648228,
+ "learning_rate": 0.0006,
+ "loss": 3.982743263244629,
+ "step": 2748
+ },
+ {
+ "epoch": 38.18165938864629,
+ "grad_norm": 0.016095101833343506,
+ "learning_rate": 0.0006,
+ "loss": 4.0849385261535645,
+ "step": 2749
+ },
+ {
+ "epoch": 38.19563318777293,
+ "grad_norm": 0.015201197937130928,
+ "learning_rate": 0.0006,
+ "loss": 4.103863716125488,
+ "step": 2750
+ },
+ {
+ "epoch": 38.209606986899566,
+ "grad_norm": 0.015687478706240654,
+ "learning_rate": 0.0006,
+ "loss": 4.12074089050293,
+ "step": 2751
+ },
+ {
+ "epoch": 38.223580786026204,
+ "grad_norm": 0.015201589092612267,
+ "learning_rate": 0.0006,
+ "loss": 4.056872367858887,
+ "step": 2752
+ },
+ {
+ "epoch": 38.237554585152836,
+ "grad_norm": 0.015154894441366196,
+ "learning_rate": 0.0006,
+ "loss": 4.089670181274414,
+ "step": 2753
+ },
+ {
+ "epoch": 38.251528384279474,
+ "grad_norm": 0.01586608588695526,
+ "learning_rate": 0.0006,
+ "loss": 4.155317306518555,
+ "step": 2754
+ },
+ {
+ "epoch": 38.26550218340611,
+ "grad_norm": 0.01609654724597931,
+ "learning_rate": 0.0006,
+ "loss": 3.995800495147705,
+ "step": 2755
+ },
+ {
+ "epoch": 38.27947598253275,
+ "grad_norm": 0.015495185740292072,
+ "learning_rate": 0.0006,
+ "loss": 3.995769739151001,
+ "step": 2756
+ },
+ {
+ "epoch": 38.29344978165939,
+ "grad_norm": 0.017042381688952446,
+ "learning_rate": 0.0006,
+ "loss": 4.132832050323486,
+ "step": 2757
+ },
+ {
+ "epoch": 38.30742358078603,
+ "grad_norm": 0.018706614151597023,
+ "learning_rate": 0.0006,
+ "loss": 4.038820743560791,
+ "step": 2758
+ },
+ {
+ "epoch": 38.32139737991266,
+ "grad_norm": 0.018316026777029037,
+ "learning_rate": 0.0006,
+ "loss": 3.972052574157715,
+ "step": 2759
+ },
+ {
+ "epoch": 38.3353711790393,
+ "grad_norm": 0.018375389277935028,
+ "learning_rate": 0.0006,
+ "loss": 4.152097702026367,
+ "step": 2760
+ },
+ {
+ "epoch": 38.34934497816594,
+ "grad_norm": 0.016879886388778687,
+ "learning_rate": 0.0006,
+ "loss": 4.020869731903076,
+ "step": 2761
+ },
+ {
+ "epoch": 38.36331877729258,
+ "grad_norm": 0.015728183090686798,
+ "learning_rate": 0.0006,
+ "loss": 4.008016586303711,
+ "step": 2762
+ },
+ {
+ "epoch": 38.377292576419215,
+ "grad_norm": 0.016147758811712265,
+ "learning_rate": 0.0006,
+ "loss": 4.035854339599609,
+ "step": 2763
+ },
+ {
+ "epoch": 38.391266375545854,
+ "grad_norm": 0.017774797976017,
+ "learning_rate": 0.0006,
+ "loss": 4.122506141662598,
+ "step": 2764
+ },
+ {
+ "epoch": 38.40524017467249,
+ "grad_norm": 0.016590069979429245,
+ "learning_rate": 0.0006,
+ "loss": 4.082847595214844,
+ "step": 2765
+ },
+ {
+ "epoch": 38.419213973799124,
+ "grad_norm": 0.016544528305530548,
+ "learning_rate": 0.0006,
+ "loss": 4.184270858764648,
+ "step": 2766
+ },
+ {
+ "epoch": 38.43318777292576,
+ "grad_norm": 0.017972925677895546,
+ "learning_rate": 0.0006,
+ "loss": 3.987558603286743,
+ "step": 2767
+ },
+ {
+ "epoch": 38.4471615720524,
+ "grad_norm": 0.018562227487564087,
+ "learning_rate": 0.0006,
+ "loss": 4.1193461418151855,
+ "step": 2768
+ },
+ {
+ "epoch": 38.46113537117904,
+ "grad_norm": 0.018896685913205147,
+ "learning_rate": 0.0006,
+ "loss": 4.057808876037598,
+ "step": 2769
+ },
+ {
+ "epoch": 38.47510917030568,
+ "grad_norm": 0.016793809831142426,
+ "learning_rate": 0.0006,
+ "loss": 4.069523334503174,
+ "step": 2770
+ },
+ {
+ "epoch": 38.48908296943232,
+ "grad_norm": 0.01698676310479641,
+ "learning_rate": 0.0006,
+ "loss": 3.997624158859253,
+ "step": 2771
+ },
+ {
+ "epoch": 38.50305676855895,
+ "grad_norm": 0.016784202307462692,
+ "learning_rate": 0.0006,
+ "loss": 4.174795627593994,
+ "step": 2772
+ },
+ {
+ "epoch": 38.51703056768559,
+ "grad_norm": 0.016785014420747757,
+ "learning_rate": 0.0006,
+ "loss": 3.999321222305298,
+ "step": 2773
+ },
+ {
+ "epoch": 38.531004366812226,
+ "grad_norm": 0.018615184351801872,
+ "learning_rate": 0.0006,
+ "loss": 4.191237926483154,
+ "step": 2774
+ },
+ {
+ "epoch": 38.544978165938865,
+ "grad_norm": 0.019090790301561356,
+ "learning_rate": 0.0006,
+ "loss": 4.106966018676758,
+ "step": 2775
+ },
+ {
+ "epoch": 38.5589519650655,
+ "grad_norm": 0.018298320472240448,
+ "learning_rate": 0.0006,
+ "loss": 4.063363552093506,
+ "step": 2776
+ },
+ {
+ "epoch": 38.57292576419214,
+ "grad_norm": 0.018541090190410614,
+ "learning_rate": 0.0006,
+ "loss": 4.101193428039551,
+ "step": 2777
+ },
+ {
+ "epoch": 38.58689956331878,
+ "grad_norm": 0.01849970407783985,
+ "learning_rate": 0.0006,
+ "loss": 4.071335792541504,
+ "step": 2778
+ },
+ {
+ "epoch": 38.60087336244541,
+ "grad_norm": 0.018309077247977257,
+ "learning_rate": 0.0006,
+ "loss": 4.245305061340332,
+ "step": 2779
+ },
+ {
+ "epoch": 38.61484716157205,
+ "grad_norm": 0.017681840807199478,
+ "learning_rate": 0.0006,
+ "loss": 4.081173419952393,
+ "step": 2780
+ },
+ {
+ "epoch": 38.62882096069869,
+ "grad_norm": 0.016286224126815796,
+ "learning_rate": 0.0006,
+ "loss": 4.024351596832275,
+ "step": 2781
+ },
+ {
+ "epoch": 38.64279475982533,
+ "grad_norm": 0.015837235376238823,
+ "learning_rate": 0.0006,
+ "loss": 4.028205394744873,
+ "step": 2782
+ },
+ {
+ "epoch": 38.65676855895197,
+ "grad_norm": 0.016062596812844276,
+ "learning_rate": 0.0006,
+ "loss": 4.131122589111328,
+ "step": 2783
+ },
+ {
+ "epoch": 38.670742358078606,
+ "grad_norm": 0.01711239665746689,
+ "learning_rate": 0.0006,
+ "loss": 4.124416351318359,
+ "step": 2784
+ },
+ {
+ "epoch": 38.68471615720524,
+ "grad_norm": 0.015304913744330406,
+ "learning_rate": 0.0006,
+ "loss": 4.056122779846191,
+ "step": 2785
+ },
+ {
+ "epoch": 38.698689956331876,
+ "grad_norm": 0.015971887856721878,
+ "learning_rate": 0.0006,
+ "loss": 4.132555961608887,
+ "step": 2786
+ },
+ {
+ "epoch": 38.712663755458514,
+ "grad_norm": 0.015787221491336823,
+ "learning_rate": 0.0006,
+ "loss": 4.051596164703369,
+ "step": 2787
+ },
+ {
+ "epoch": 38.72663755458515,
+ "grad_norm": 0.01605444774031639,
+ "learning_rate": 0.0006,
+ "loss": 4.191808700561523,
+ "step": 2788
+ },
+ {
+ "epoch": 38.74061135371179,
+ "grad_norm": 0.016383009031414986,
+ "learning_rate": 0.0006,
+ "loss": 4.0648417472839355,
+ "step": 2789
+ },
+ {
+ "epoch": 38.75458515283843,
+ "grad_norm": 0.016135554760694504,
+ "learning_rate": 0.0006,
+ "loss": 3.988563060760498,
+ "step": 2790
+ },
+ {
+ "epoch": 38.76855895196506,
+ "grad_norm": 0.016634158790111542,
+ "learning_rate": 0.0006,
+ "loss": 4.133493423461914,
+ "step": 2791
+ },
+ {
+ "epoch": 38.7825327510917,
+ "grad_norm": 0.01688210666179657,
+ "learning_rate": 0.0006,
+ "loss": 4.091277122497559,
+ "step": 2792
+ },
+ {
+ "epoch": 38.79650655021834,
+ "grad_norm": 0.016102122142910957,
+ "learning_rate": 0.0006,
+ "loss": 4.051219940185547,
+ "step": 2793
+ },
+ {
+ "epoch": 38.81048034934498,
+ "grad_norm": 0.016585038974881172,
+ "learning_rate": 0.0006,
+ "loss": 4.058587074279785,
+ "step": 2794
+ },
+ {
+ "epoch": 38.82445414847162,
+ "grad_norm": 0.017700951546430588,
+ "learning_rate": 0.0006,
+ "loss": 4.016030311584473,
+ "step": 2795
+ },
+ {
+ "epoch": 38.838427947598255,
+ "grad_norm": 0.017922203987836838,
+ "learning_rate": 0.0006,
+ "loss": 4.079977989196777,
+ "step": 2796
+ },
+ {
+ "epoch": 38.852401746724894,
+ "grad_norm": 0.017299890518188477,
+ "learning_rate": 0.0006,
+ "loss": 4.062984943389893,
+ "step": 2797
+ },
+ {
+ "epoch": 38.866375545851525,
+ "grad_norm": 0.018648581579327583,
+ "learning_rate": 0.0006,
+ "loss": 4.065717697143555,
+ "step": 2798
+ },
+ {
+ "epoch": 38.880349344978164,
+ "grad_norm": 0.019057920202612877,
+ "learning_rate": 0.0006,
+ "loss": 4.227892875671387,
+ "step": 2799
+ },
+ {
+ "epoch": 38.8943231441048,
+ "grad_norm": 0.018824422731995583,
+ "learning_rate": 0.0006,
+ "loss": 4.116058349609375,
+ "step": 2800
+ },
+ {
+ "epoch": 38.90829694323144,
+ "grad_norm": 0.019084978848695755,
+ "learning_rate": 0.0006,
+ "loss": 4.061677932739258,
+ "step": 2801
+ },
+ {
+ "epoch": 38.92227074235808,
+ "grad_norm": 0.017842981964349747,
+ "learning_rate": 0.0006,
+ "loss": 4.083813667297363,
+ "step": 2802
+ },
+ {
+ "epoch": 38.93624454148472,
+ "grad_norm": 0.01880858838558197,
+ "learning_rate": 0.0006,
+ "loss": 4.152561187744141,
+ "step": 2803
+ },
+ {
+ "epoch": 38.95021834061135,
+ "grad_norm": 0.017400087788701057,
+ "learning_rate": 0.0006,
+ "loss": 4.154635429382324,
+ "step": 2804
+ },
+ {
+ "epoch": 38.96419213973799,
+ "grad_norm": 0.017682507634162903,
+ "learning_rate": 0.0006,
+ "loss": 4.151128768920898,
+ "step": 2805
+ },
+ {
+ "epoch": 38.97816593886463,
+ "grad_norm": 0.015803052112460136,
+ "learning_rate": 0.0006,
+ "loss": 4.10498571395874,
+ "step": 2806
+ },
+ {
+ "epoch": 38.992139737991266,
+ "grad_norm": 0.01582511141896248,
+ "learning_rate": 0.0006,
+ "loss": 4.140159606933594,
+ "step": 2807
+ },
+ {
+ "epoch": 39.0,
+ "grad_norm": 0.01925666816532612,
+ "learning_rate": 0.0006,
+ "loss": 4.126394271850586,
+ "step": 2808
+ },
+ {
+ "epoch": 39.0,
+ "eval_loss": 4.615617275238037,
+ "eval_runtime": 56.8895,
+ "eval_samples_per_second": 42.925,
+ "eval_steps_per_second": 1.354,
+ "step": 2808
+ },
+ {
+ "epoch": 39.01397379912664,
+ "grad_norm": 0.01677853614091873,
+ "learning_rate": 0.0006,
+ "loss": 4.102395057678223,
+ "step": 2809
+ },
+ {
+ "epoch": 39.02794759825328,
+ "grad_norm": 0.017901137471199036,
+ "learning_rate": 0.0006,
+ "loss": 3.976609230041504,
+ "step": 2810
+ },
+ {
+ "epoch": 39.041921397379916,
+ "grad_norm": 0.017978668212890625,
+ "learning_rate": 0.0006,
+ "loss": 3.944416046142578,
+ "step": 2811
+ },
+ {
+ "epoch": 39.05589519650655,
+ "grad_norm": 0.019967492669820786,
+ "learning_rate": 0.0006,
+ "loss": 3.992234468460083,
+ "step": 2812
+ },
+ {
+ "epoch": 39.069868995633186,
+ "grad_norm": 0.018937544897198677,
+ "learning_rate": 0.0006,
+ "loss": 4.028573036193848,
+ "step": 2813
+ },
+ {
+ "epoch": 39.083842794759825,
+ "grad_norm": 0.016852952539920807,
+ "learning_rate": 0.0006,
+ "loss": 4.062442779541016,
+ "step": 2814
+ },
+ {
+ "epoch": 39.09781659388646,
+ "grad_norm": 0.016643647104501724,
+ "learning_rate": 0.0006,
+ "loss": 4.08871603012085,
+ "step": 2815
+ },
+ {
+ "epoch": 39.1117903930131,
+ "grad_norm": 0.016872866079211235,
+ "learning_rate": 0.0006,
+ "loss": 3.9514245986938477,
+ "step": 2816
+ },
+ {
+ "epoch": 39.12576419213974,
+ "grad_norm": 0.018045669421553612,
+ "learning_rate": 0.0006,
+ "loss": 4.038842678070068,
+ "step": 2817
+ },
+ {
+ "epoch": 39.13973799126637,
+ "grad_norm": 0.017853403463959694,
+ "learning_rate": 0.0006,
+ "loss": 3.93709135055542,
+ "step": 2818
+ },
+ {
+ "epoch": 39.15371179039301,
+ "grad_norm": 0.017507443204522133,
+ "learning_rate": 0.0006,
+ "loss": 3.98665714263916,
+ "step": 2819
+ },
+ {
+ "epoch": 39.16768558951965,
+ "grad_norm": 0.01766853965818882,
+ "learning_rate": 0.0006,
+ "loss": 3.937126636505127,
+ "step": 2820
+ },
+ {
+ "epoch": 39.18165938864629,
+ "grad_norm": 0.016660021618008614,
+ "learning_rate": 0.0006,
+ "loss": 3.8397014141082764,
+ "step": 2821
+ },
+ {
+ "epoch": 39.19563318777293,
+ "grad_norm": 0.017719309777021408,
+ "learning_rate": 0.0006,
+ "loss": 4.003579616546631,
+ "step": 2822
+ },
+ {
+ "epoch": 39.209606986899566,
+ "grad_norm": 0.018234414979815483,
+ "learning_rate": 0.0006,
+ "loss": 4.053859710693359,
+ "step": 2823
+ },
+ {
+ "epoch": 39.223580786026204,
+ "grad_norm": 0.018777024000883102,
+ "learning_rate": 0.0006,
+ "loss": 4.056182861328125,
+ "step": 2824
+ },
+ {
+ "epoch": 39.237554585152836,
+ "grad_norm": 0.017443502321839333,
+ "learning_rate": 0.0006,
+ "loss": 4.025068283081055,
+ "step": 2825
+ },
+ {
+ "epoch": 39.251528384279474,
+ "grad_norm": 0.016500331461429596,
+ "learning_rate": 0.0006,
+ "loss": 4.101631164550781,
+ "step": 2826
+ },
+ {
+ "epoch": 39.26550218340611,
+ "grad_norm": 0.016101544722914696,
+ "learning_rate": 0.0006,
+ "loss": 4.027082443237305,
+ "step": 2827
+ },
+ {
+ "epoch": 39.27947598253275,
+ "grad_norm": 0.015471171587705612,
+ "learning_rate": 0.0006,
+ "loss": 4.022909641265869,
+ "step": 2828
+ },
+ {
+ "epoch": 39.29344978165939,
+ "grad_norm": 0.015779748558998108,
+ "learning_rate": 0.0006,
+ "loss": 4.18796968460083,
+ "step": 2829
+ },
+ {
+ "epoch": 39.30742358078603,
+ "grad_norm": 0.015522679314017296,
+ "learning_rate": 0.0006,
+ "loss": 4.005777359008789,
+ "step": 2830
+ },
+ {
+ "epoch": 39.32139737991266,
+ "grad_norm": 0.015226504765450954,
+ "learning_rate": 0.0006,
+ "loss": 4.108416557312012,
+ "step": 2831
+ },
+ {
+ "epoch": 39.3353711790393,
+ "grad_norm": 0.015095778740942478,
+ "learning_rate": 0.0006,
+ "loss": 4.08029842376709,
+ "step": 2832
+ },
+ {
+ "epoch": 39.34934497816594,
+ "grad_norm": 0.015601031482219696,
+ "learning_rate": 0.0006,
+ "loss": 3.989492416381836,
+ "step": 2833
+ },
+ {
+ "epoch": 39.36331877729258,
+ "grad_norm": 0.014825052581727505,
+ "learning_rate": 0.0006,
+ "loss": 3.998520612716675,
+ "step": 2834
+ },
+ {
+ "epoch": 39.377292576419215,
+ "grad_norm": 0.014821100980043411,
+ "learning_rate": 0.0006,
+ "loss": 4.001777172088623,
+ "step": 2835
+ },
+ {
+ "epoch": 39.391266375545854,
+ "grad_norm": 0.015333754941821098,
+ "learning_rate": 0.0006,
+ "loss": 4.11985969543457,
+ "step": 2836
+ },
+ {
+ "epoch": 39.40524017467249,
+ "grad_norm": 0.015940699726343155,
+ "learning_rate": 0.0006,
+ "loss": 4.025373935699463,
+ "step": 2837
+ },
+ {
+ "epoch": 39.419213973799124,
+ "grad_norm": 0.015096287243068218,
+ "learning_rate": 0.0006,
+ "loss": 4.068551063537598,
+ "step": 2838
+ },
+ {
+ "epoch": 39.43318777292576,
+ "grad_norm": 0.01533548254519701,
+ "learning_rate": 0.0006,
+ "loss": 3.9800801277160645,
+ "step": 2839
+ },
+ {
+ "epoch": 39.4471615720524,
+ "grad_norm": 0.014648732729256153,
+ "learning_rate": 0.0006,
+ "loss": 3.9670190811157227,
+ "step": 2840
+ },
+ {
+ "epoch": 39.46113537117904,
+ "grad_norm": 0.015759671106934547,
+ "learning_rate": 0.0006,
+ "loss": 4.050443649291992,
+ "step": 2841
+ },
+ {
+ "epoch": 39.47510917030568,
+ "grad_norm": 0.016054073348641396,
+ "learning_rate": 0.0006,
+ "loss": 4.22734260559082,
+ "step": 2842
+ },
+ {
+ "epoch": 39.48908296943232,
+ "grad_norm": 0.0160264540463686,
+ "learning_rate": 0.0006,
+ "loss": 4.116174221038818,
+ "step": 2843
+ },
+ {
+ "epoch": 39.50305676855895,
+ "grad_norm": 0.015426610596477985,
+ "learning_rate": 0.0006,
+ "loss": 4.000178337097168,
+ "step": 2844
+ },
+ {
+ "epoch": 39.51703056768559,
+ "grad_norm": 0.01520642638206482,
+ "learning_rate": 0.0006,
+ "loss": 4.218410968780518,
+ "step": 2845
+ },
+ {
+ "epoch": 39.531004366812226,
+ "grad_norm": 0.015305949375033379,
+ "learning_rate": 0.0006,
+ "loss": 4.025891304016113,
+ "step": 2846
+ },
+ {
+ "epoch": 39.544978165938865,
+ "grad_norm": 0.015461564064025879,
+ "learning_rate": 0.0006,
+ "loss": 4.100950717926025,
+ "step": 2847
+ },
+ {
+ "epoch": 39.5589519650655,
+ "grad_norm": 0.01712625101208687,
+ "learning_rate": 0.0006,
+ "loss": 4.0620856285095215,
+ "step": 2848
+ },
+ {
+ "epoch": 39.57292576419214,
+ "grad_norm": 0.017133301123976707,
+ "learning_rate": 0.0006,
+ "loss": 4.097138404846191,
+ "step": 2849
+ },
+ {
+ "epoch": 39.58689956331878,
+ "grad_norm": 0.01683625392615795,
+ "learning_rate": 0.0006,
+ "loss": 4.084364891052246,
+ "step": 2850
+ },
+ {
+ "epoch": 39.60087336244541,
+ "grad_norm": 0.0165710486471653,
+ "learning_rate": 0.0006,
+ "loss": 4.042361259460449,
+ "step": 2851
+ },
+ {
+ "epoch": 39.61484716157205,
+ "grad_norm": 0.01660393364727497,
+ "learning_rate": 0.0006,
+ "loss": 4.1395487785339355,
+ "step": 2852
+ },
+ {
+ "epoch": 39.62882096069869,
+ "grad_norm": 0.017025845125317574,
+ "learning_rate": 0.0006,
+ "loss": 4.1207075119018555,
+ "step": 2853
+ },
+ {
+ "epoch": 39.64279475982533,
+ "grad_norm": 0.01660497486591339,
+ "learning_rate": 0.0006,
+ "loss": 4.103846073150635,
+ "step": 2854
+ },
+ {
+ "epoch": 39.65676855895197,
+ "grad_norm": 0.01607946865260601,
+ "learning_rate": 0.0006,
+ "loss": 4.189016342163086,
+ "step": 2855
+ },
+ {
+ "epoch": 39.670742358078606,
+ "grad_norm": 0.01614505983889103,
+ "learning_rate": 0.0006,
+ "loss": 4.079860687255859,
+ "step": 2856
+ },
+ {
+ "epoch": 39.68471615720524,
+ "grad_norm": 0.017085086554288864,
+ "learning_rate": 0.0006,
+ "loss": 3.9774293899536133,
+ "step": 2857
+ },
+ {
+ "epoch": 39.698689956331876,
+ "grad_norm": 0.01816520094871521,
+ "learning_rate": 0.0006,
+ "loss": 4.113666534423828,
+ "step": 2858
+ },
+ {
+ "epoch": 39.712663755458514,
+ "grad_norm": 0.016597526147961617,
+ "learning_rate": 0.0006,
+ "loss": 4.068242073059082,
+ "step": 2859
+ },
+ {
+ "epoch": 39.72663755458515,
+ "grad_norm": 0.01672235131263733,
+ "learning_rate": 0.0006,
+ "loss": 4.024444580078125,
+ "step": 2860
+ },
+ {
+ "epoch": 39.74061135371179,
+ "grad_norm": 0.020093556493520737,
+ "learning_rate": 0.0006,
+ "loss": 4.064124584197998,
+ "step": 2861
+ },
+ {
+ "epoch": 39.75458515283843,
+ "grad_norm": 0.023141304031014442,
+ "learning_rate": 0.0006,
+ "loss": 4.130439758300781,
+ "step": 2862
+ },
+ {
+ "epoch": 39.76855895196506,
+ "grad_norm": 0.02289372682571411,
+ "learning_rate": 0.0006,
+ "loss": 4.14794921875,
+ "step": 2863
+ },
+ {
+ "epoch": 39.7825327510917,
+ "grad_norm": 0.01919020712375641,
+ "learning_rate": 0.0006,
+ "loss": 4.123961925506592,
+ "step": 2864
+ },
+ {
+ "epoch": 39.79650655021834,
+ "grad_norm": 0.016932329162955284,
+ "learning_rate": 0.0006,
+ "loss": 4.0944132804870605,
+ "step": 2865
+ },
+ {
+ "epoch": 39.81048034934498,
+ "grad_norm": 0.016644535586237907,
+ "learning_rate": 0.0006,
+ "loss": 4.124269962310791,
+ "step": 2866
+ },
+ {
+ "epoch": 39.82445414847162,
+ "grad_norm": 0.016302216798067093,
+ "learning_rate": 0.0006,
+ "loss": 4.0267205238342285,
+ "step": 2867
+ },
+ {
+ "epoch": 39.838427947598255,
+ "grad_norm": 0.01673300378024578,
+ "learning_rate": 0.0006,
+ "loss": 4.10650634765625,
+ "step": 2868
+ },
+ {
+ "epoch": 39.852401746724894,
+ "grad_norm": 0.016581356525421143,
+ "learning_rate": 0.0006,
+ "loss": 4.1547956466674805,
+ "step": 2869
+ },
+ {
+ "epoch": 39.866375545851525,
+ "grad_norm": 0.01691797934472561,
+ "learning_rate": 0.0006,
+ "loss": 4.056936740875244,
+ "step": 2870
+ },
+ {
+ "epoch": 39.880349344978164,
+ "grad_norm": 0.01596726104617119,
+ "learning_rate": 0.0006,
+ "loss": 3.9891414642333984,
+ "step": 2871
+ },
+ {
+ "epoch": 39.8943231441048,
+ "grad_norm": 0.015615091659128666,
+ "learning_rate": 0.0006,
+ "loss": 4.137948036193848,
+ "step": 2872
+ },
+ {
+ "epoch": 39.90829694323144,
+ "grad_norm": 0.015703413635492325,
+ "learning_rate": 0.0006,
+ "loss": 3.9844930171966553,
+ "step": 2873
+ },
+ {
+ "epoch": 39.92227074235808,
+ "grad_norm": 0.015453094616532326,
+ "learning_rate": 0.0006,
+ "loss": 4.098443508148193,
+ "step": 2874
+ },
+ {
+ "epoch": 39.93624454148472,
+ "grad_norm": 0.01449158601462841,
+ "learning_rate": 0.0006,
+ "loss": 4.090852737426758,
+ "step": 2875
+ },
+ {
+ "epoch": 39.95021834061135,
+ "grad_norm": 0.01537515688687563,
+ "learning_rate": 0.0006,
+ "loss": 4.239612579345703,
+ "step": 2876
+ },
+ {
+ "epoch": 39.96419213973799,
+ "grad_norm": 0.015989921987056732,
+ "learning_rate": 0.0006,
+ "loss": 4.142073154449463,
+ "step": 2877
+ },
+ {
+ "epoch": 39.97816593886463,
+ "grad_norm": 0.015985611826181412,
+ "learning_rate": 0.0006,
+ "loss": 4.100076675415039,
+ "step": 2878
+ },
+ {
+ "epoch": 39.992139737991266,
+ "grad_norm": 0.015260833315551281,
+ "learning_rate": 0.0006,
+ "loss": 4.068576335906982,
+ "step": 2879
+ },
+ {
+ "epoch": 40.0,
+ "grad_norm": 0.018047941848635674,
+ "learning_rate": 0.0006,
+ "loss": 4.005878925323486,
+ "step": 2880
+ },
+ {
+ "epoch": 40.0,
+ "eval_loss": 4.622848987579346,
+ "eval_runtime": 56.9618,
+ "eval_samples_per_second": 42.871,
+ "eval_steps_per_second": 1.352,
+ "step": 2880
+ },
+ {
+ "epoch": 40.01397379912664,
+ "grad_norm": 0.015976838767528534,
+ "learning_rate": 0.0006,
+ "loss": 4.035689353942871,
+ "step": 2881
+ },
+ {
+ "epoch": 40.02794759825328,
+ "grad_norm": 0.015176220797002316,
+ "learning_rate": 0.0006,
+ "loss": 3.932197332382202,
+ "step": 2882
+ },
+ {
+ "epoch": 40.041921397379916,
+ "grad_norm": 0.016676228493452072,
+ "learning_rate": 0.0006,
+ "loss": 4.014171123504639,
+ "step": 2883
+ },
+ {
+ "epoch": 40.05589519650655,
+ "grad_norm": 0.01698281429708004,
+ "learning_rate": 0.0006,
+ "loss": 4.117893218994141,
+ "step": 2884
+ },
+ {
+ "epoch": 40.069868995633186,
+ "grad_norm": 0.017872486263513565,
+ "learning_rate": 0.0006,
+ "loss": 4.132119178771973,
+ "step": 2885
+ },
+ {
+ "epoch": 40.083842794759825,
+ "grad_norm": 0.017615554854273796,
+ "learning_rate": 0.0006,
+ "loss": 4.072000980377197,
+ "step": 2886
+ },
+ {
+ "epoch": 40.09781659388646,
+ "grad_norm": 0.017469394952058792,
+ "learning_rate": 0.0006,
+ "loss": 4.017726421356201,
+ "step": 2887
+ },
+ {
+ "epoch": 40.1117903930131,
+ "grad_norm": 0.01954762637615204,
+ "learning_rate": 0.0006,
+ "loss": 4.014443874359131,
+ "step": 2888
+ },
+ {
+ "epoch": 40.12576419213974,
+ "grad_norm": 0.018935853615403175,
+ "learning_rate": 0.0006,
+ "loss": 4.0042948722839355,
+ "step": 2889
+ },
+ {
+ "epoch": 40.13973799126637,
+ "grad_norm": 0.019382692873477936,
+ "learning_rate": 0.0006,
+ "loss": 4.06160831451416,
+ "step": 2890
+ },
+ {
+ "epoch": 40.15371179039301,
+ "grad_norm": 0.019677115604281425,
+ "learning_rate": 0.0006,
+ "loss": 4.13466215133667,
+ "step": 2891
+ },
+ {
+ "epoch": 40.16768558951965,
+ "grad_norm": 0.020365122705698013,
+ "learning_rate": 0.0006,
+ "loss": 4.095000267028809,
+ "step": 2892
+ },
+ {
+ "epoch": 40.18165938864629,
+ "grad_norm": 0.021842138841748238,
+ "learning_rate": 0.0006,
+ "loss": 3.9877853393554688,
+ "step": 2893
+ },
+ {
+ "epoch": 40.19563318777293,
+ "grad_norm": 0.02354881353676319,
+ "learning_rate": 0.0006,
+ "loss": 3.9430112838745117,
+ "step": 2894
+ },
+ {
+ "epoch": 40.209606986899566,
+ "grad_norm": 0.026602735742926598,
+ "learning_rate": 0.0006,
+ "loss": 4.012076377868652,
+ "step": 2895
+ },
+ {
+ "epoch": 40.223580786026204,
+ "grad_norm": 0.023976484313607216,
+ "learning_rate": 0.0006,
+ "loss": 3.9768595695495605,
+ "step": 2896
+ },
+ {
+ "epoch": 40.237554585152836,
+ "grad_norm": 0.02336616814136505,
+ "learning_rate": 0.0006,
+ "loss": 4.1096038818359375,
+ "step": 2897
+ },
+ {
+ "epoch": 40.251528384279474,
+ "grad_norm": 0.02426023595035076,
+ "learning_rate": 0.0006,
+ "loss": 4.157189846038818,
+ "step": 2898
+ },
+ {
+ "epoch": 40.26550218340611,
+ "grad_norm": 0.021572696045041084,
+ "learning_rate": 0.0006,
+ "loss": 3.979175090789795,
+ "step": 2899
+ },
+ {
+ "epoch": 40.27947598253275,
+ "grad_norm": 0.08300432562828064,
+ "learning_rate": 0.0006,
+ "loss": 4.092336654663086,
+ "step": 2900
+ },
+ {
+ "epoch": 40.29344978165939,
+ "grad_norm": 0.7323967814445496,
+ "learning_rate": 0.0006,
+ "loss": 5.337990760803223,
+ "step": 2901
+ },
+ {
+ "epoch": 40.30742358078603,
+ "grad_norm": 0.3363337814807892,
+ "learning_rate": 0.0006,
+ "loss": 6.40043830871582,
+ "step": 2902
+ },
+ {
+ "epoch": 40.32139737991266,
+ "grad_norm": 0.598686933517456,
+ "learning_rate": 0.0006,
+ "loss": 7.004765510559082,
+ "step": 2903
+ },
+ {
+ "epoch": 40.3353711790393,
+ "grad_norm": 0.5998364686965942,
+ "learning_rate": 0.0006,
+ "loss": 8.184747695922852,
+ "step": 2904
+ },
+ {
+ "epoch": 40.34934497816594,
+ "grad_norm": 1.408208966255188,
+ "learning_rate": 0.0006,
+ "loss": 8.88978385925293,
+ "step": 2905
+ },
+ {
+ "epoch": 40.36331877729258,
+ "grad_norm": 0.7829785943031311,
+ "learning_rate": 0.0006,
+ "loss": 11.13601303100586,
+ "step": 2906
+ },
+ {
+ "epoch": 40.377292576419215,
+ "grad_norm": 0.7769215106964111,
+ "learning_rate": 0.0006,
+ "loss": 9.90416145324707,
+ "step": 2907
+ },
+ {
+ "epoch": 40.391266375545854,
+ "grad_norm": 0.46508079767227173,
+ "learning_rate": 0.0006,
+ "loss": 9.647209167480469,
+ "step": 2908
+ },
+ {
+ "epoch": 40.40524017467249,
+ "grad_norm": 1.0228968858718872,
+ "learning_rate": 0.0006,
+ "loss": 11.685434341430664,
+ "step": 2909
+ },
+ {
+ "epoch": 40.419213973799124,
+ "grad_norm": 0.20553778111934662,
+ "learning_rate": 0.0006,
+ "loss": 8.621708869934082,
+ "step": 2910
+ },
+ {
+ "epoch": 40.43318777292576,
+ "grad_norm": 0.18466424942016602,
+ "learning_rate": 0.0006,
+ "loss": 8.590787887573242,
+ "step": 2911
+ },
+ {
+ "epoch": 40.4471615720524,
+ "grad_norm": 0.10458594560623169,
+ "learning_rate": 0.0006,
+ "loss": 8.324458122253418,
+ "step": 2912
+ },
+ {
+ "epoch": 40.46113537117904,
+ "grad_norm": 0.1096842959523201,
+ "learning_rate": 0.0006,
+ "loss": 8.300851821899414,
+ "step": 2913
+ },
+ {
+ "epoch": 40.47510917030568,
+ "grad_norm": 0.10501226782798767,
+ "learning_rate": 0.0006,
+ "loss": 8.21009635925293,
+ "step": 2914
+ },
+ {
+ "epoch": 40.48908296943232,
+ "grad_norm": 0.13704165816307068,
+ "learning_rate": 0.0006,
+ "loss": 8.141355514526367,
+ "step": 2915
+ },
+ {
+ "epoch": 40.50305676855895,
+ "grad_norm": 0.12129108607769012,
+ "learning_rate": 0.0006,
+ "loss": 8.074612617492676,
+ "step": 2916
+ },
+ {
+ "epoch": 40.51703056768559,
+ "grad_norm": 0.10028897225856781,
+ "learning_rate": 0.0006,
+ "loss": 8.00384521484375,
+ "step": 2917
+ },
+ {
+ "epoch": 40.531004366812226,
+ "grad_norm": 0.11764246970415115,
+ "learning_rate": 0.0006,
+ "loss": 8.078598022460938,
+ "step": 2918
+ },
+ {
+ "epoch": 40.544978165938865,
+ "grad_norm": 0.10081959515810013,
+ "learning_rate": 0.0006,
+ "loss": 8.008938789367676,
+ "step": 2919
+ },
+ {
+ "epoch": 40.5589519650655,
+ "grad_norm": 0.1003170907497406,
+ "learning_rate": 0.0006,
+ "loss": 7.962940216064453,
+ "step": 2920
+ },
+ {
+ "epoch": 40.57292576419214,
+ "grad_norm": 0.08763492107391357,
+ "learning_rate": 0.0006,
+ "loss": 7.909419536590576,
+ "step": 2921
+ },
+ {
+ "epoch": 40.58689956331878,
+ "grad_norm": 0.0636807233095169,
+ "learning_rate": 0.0006,
+ "loss": 7.878905773162842,
+ "step": 2922
+ },
+ {
+ "epoch": 40.60087336244541,
+ "grad_norm": 0.06005936488509178,
+ "learning_rate": 0.0006,
+ "loss": 7.850980758666992,
+ "step": 2923
+ },
+ {
+ "epoch": 40.61484716157205,
+ "grad_norm": 0.06415296345949173,
+ "learning_rate": 0.0006,
+ "loss": 7.830353260040283,
+ "step": 2924
+ },
+ {
+ "epoch": 40.62882096069869,
+ "grad_norm": 0.0732007846236229,
+ "learning_rate": 0.0006,
+ "loss": 7.830746650695801,
+ "step": 2925
+ },
+ {
+ "epoch": 40.64279475982533,
+ "grad_norm": 0.06628110259771347,
+ "learning_rate": 0.0006,
+ "loss": 7.82661771774292,
+ "step": 2926
+ },
+ {
+ "epoch": 40.65676855895197,
+ "grad_norm": 0.05544218048453331,
+ "learning_rate": 0.0006,
+ "loss": 7.814305782318115,
+ "step": 2927
+ },
+ {
+ "epoch": 40.670742358078606,
+ "grad_norm": 0.050327103585004807,
+ "learning_rate": 0.0006,
+ "loss": 7.764055252075195,
+ "step": 2928
+ },
+ {
+ "epoch": 40.68471615720524,
+ "grad_norm": 0.04918425530195236,
+ "learning_rate": 0.0006,
+ "loss": 7.744625091552734,
+ "step": 2929
+ },
+ {
+ "epoch": 40.698689956331876,
+ "grad_norm": 0.042862266302108765,
+ "learning_rate": 0.0006,
+ "loss": 7.764426231384277,
+ "step": 2930
+ },
+ {
+ "epoch": 40.712663755458514,
+ "grad_norm": 0.04338089004158974,
+ "learning_rate": 0.0006,
+ "loss": 7.712530136108398,
+ "step": 2931
+ },
+ {
+ "epoch": 40.72663755458515,
+ "grad_norm": 0.04127737879753113,
+ "learning_rate": 0.0006,
+ "loss": 7.742180347442627,
+ "step": 2932
+ },
+ {
+ "epoch": 40.74061135371179,
+ "grad_norm": 0.03358631581068039,
+ "learning_rate": 0.0006,
+ "loss": 7.714834213256836,
+ "step": 2933
+ },
+ {
+ "epoch": 40.75458515283843,
+ "grad_norm": 0.02845790982246399,
+ "learning_rate": 0.0006,
+ "loss": 7.7021026611328125,
+ "step": 2934
+ },
+ {
+ "epoch": 40.76855895196506,
+ "grad_norm": 0.03012419492006302,
+ "learning_rate": 0.0006,
+ "loss": 7.712377548217773,
+ "step": 2935
+ },
+ {
+ "epoch": 40.7825327510917,
+ "grad_norm": 0.032981179654598236,
+ "learning_rate": 0.0006,
+ "loss": 7.666385650634766,
+ "step": 2936
+ },
+ {
+ "epoch": 40.79650655021834,
+ "grad_norm": 0.03357074037194252,
+ "learning_rate": 0.0006,
+ "loss": 7.649097919464111,
+ "step": 2937
+ },
+ {
+ "epoch": 40.81048034934498,
+ "grad_norm": 0.030914708971977234,
+ "learning_rate": 0.0006,
+ "loss": 7.649196624755859,
+ "step": 2938
+ },
+ {
+ "epoch": 40.82445414847162,
+ "grad_norm": 0.026296131312847137,
+ "learning_rate": 0.0006,
+ "loss": 7.629964828491211,
+ "step": 2939
+ },
+ {
+ "epoch": 40.838427947598255,
+ "grad_norm": 0.026677245274186134,
+ "learning_rate": 0.0006,
+ "loss": 7.599973678588867,
+ "step": 2940
+ },
+ {
+ "epoch": 40.852401746724894,
+ "grad_norm": 0.023208219558000565,
+ "learning_rate": 0.0006,
+ "loss": 7.601624488830566,
+ "step": 2941
+ },
+ {
+ "epoch": 40.866375545851525,
+ "grad_norm": 0.020741475746035576,
+ "learning_rate": 0.0006,
+ "loss": 7.603902339935303,
+ "step": 2942
+ },
+ {
+ "epoch": 40.880349344978164,
+ "grad_norm": 0.019221751019358635,
+ "learning_rate": 0.0006,
+ "loss": 7.5535888671875,
+ "step": 2943
+ },
+ {
+ "epoch": 40.8943231441048,
+ "grad_norm": 0.019491951912641525,
+ "learning_rate": 0.0006,
+ "loss": 7.54597282409668,
+ "step": 2944
+ },
+ {
+ "epoch": 40.90829694323144,
+ "grad_norm": 0.019702572375535965,
+ "learning_rate": 0.0006,
+ "loss": 7.537567138671875,
+ "step": 2945
+ },
+ {
+ "epoch": 40.92227074235808,
+ "grad_norm": 0.01768800988793373,
+ "learning_rate": 0.0006,
+ "loss": 7.512800216674805,
+ "step": 2946
+ },
+ {
+ "epoch": 40.93624454148472,
+ "grad_norm": 0.01926138624548912,
+ "learning_rate": 0.0006,
+ "loss": 7.509932518005371,
+ "step": 2947
+ },
+ {
+ "epoch": 40.95021834061135,
+ "grad_norm": 0.019030006602406502,
+ "learning_rate": 0.0006,
+ "loss": 7.497983932495117,
+ "step": 2948
+ },
+ {
+ "epoch": 40.96419213973799,
+ "grad_norm": 0.017313841730356216,
+ "learning_rate": 0.0006,
+ "loss": 7.463518142700195,
+ "step": 2949
+ },
+ {
+ "epoch": 40.97816593886463,
+ "grad_norm": 0.01310762483626604,
+ "learning_rate": 0.0006,
+ "loss": 7.474668502807617,
+ "step": 2950
+ },
+ {
+ "epoch": 40.992139737991266,
+ "grad_norm": 0.013008282519876957,
+ "learning_rate": 0.0006,
+ "loss": 7.45982551574707,
+ "step": 2951
+ },
+ {
+ "epoch": 41.0,
+ "grad_norm": 0.01441363524645567,
+ "learning_rate": 0.0006,
+ "loss": 7.446529388427734,
+ "step": 2952
+ },
+ {
+ "epoch": 41.0,
+ "eval_loss": 7.486120223999023,
+ "eval_runtime": 70.4986,
+ "eval_samples_per_second": 34.639,
+ "eval_steps_per_second": 1.092,
+ "step": 2952
+ },
+ {
+ "epoch": 41.01397379912664,
+ "grad_norm": 0.014061855152249336,
+ "learning_rate": 0.0006,
+ "loss": 7.4692182540893555,
+ "step": 2953
+ },
+ {
+ "epoch": 41.02794759825328,
+ "grad_norm": 0.01587500050663948,
+ "learning_rate": 0.0006,
+ "loss": 7.4182024002075195,
+ "step": 2954
+ },
+ {
+ "epoch": 41.041921397379916,
+ "grad_norm": 0.014783984050154686,
+ "learning_rate": 0.0006,
+ "loss": 7.466778755187988,
+ "step": 2955
+ },
+ {
+ "epoch": 41.05589519650655,
+ "grad_norm": 0.011560854502022266,
+ "learning_rate": 0.0006,
+ "loss": 7.438002586364746,
+ "step": 2956
+ },
+ {
+ "epoch": 41.069868995633186,
+ "grad_norm": 0.00970957800745964,
+ "learning_rate": 0.0006,
+ "loss": 7.442694187164307,
+ "step": 2957
+ },
+ {
+ "epoch": 41.083842794759825,
+ "grad_norm": 0.009633459150791168,
+ "learning_rate": 0.0006,
+ "loss": 7.458884239196777,
+ "step": 2958
+ },
+ {
+ "epoch": 41.09781659388646,
+ "grad_norm": 0.010664350353181362,
+ "learning_rate": 0.0006,
+ "loss": 7.48378324508667,
+ "step": 2959
+ },
+ {
+ "epoch": 41.1117903930131,
+ "grad_norm": 0.011440591886639595,
+ "learning_rate": 0.0006,
+ "loss": 7.474484920501709,
+ "step": 2960
+ },
+ {
+ "epoch": 41.12576419213974,
+ "grad_norm": 0.01150908600538969,
+ "learning_rate": 0.0006,
+ "loss": 7.417902946472168,
+ "step": 2961
+ },
+ {
+ "epoch": 41.13973799126637,
+ "grad_norm": 0.010375697165727615,
+ "learning_rate": 0.0006,
+ "loss": 7.403279781341553,
+ "step": 2962
+ },
+ {
+ "epoch": 41.15371179039301,
+ "grad_norm": 0.009075703099370003,
+ "learning_rate": 0.0006,
+ "loss": 7.447347164154053,
+ "step": 2963
+ },
+ {
+ "epoch": 41.16768558951965,
+ "grad_norm": 0.008892177604138851,
+ "learning_rate": 0.0006,
+ "loss": 7.402681350708008,
+ "step": 2964
+ },
+ {
+ "epoch": 41.18165938864629,
+ "grad_norm": 0.009492084383964539,
+ "learning_rate": 0.0006,
+ "loss": 7.3871355056762695,
+ "step": 2965
+ },
+ {
+ "epoch": 41.19563318777293,
+ "grad_norm": 0.009139460511505604,
+ "learning_rate": 0.0006,
+ "loss": 7.419617176055908,
+ "step": 2966
+ },
+ {
+ "epoch": 41.209606986899566,
+ "grad_norm": 0.007959138602018356,
+ "learning_rate": 0.0006,
+ "loss": 7.374949932098389,
+ "step": 2967
+ },
+ {
+ "epoch": 41.223580786026204,
+ "grad_norm": 0.007571548223495483,
+ "learning_rate": 0.0006,
+ "loss": 7.4205241203308105,
+ "step": 2968
+ },
+ {
+ "epoch": 41.237554585152836,
+ "grad_norm": 0.007885400205850601,
+ "learning_rate": 0.0006,
+ "loss": 7.416543006896973,
+ "step": 2969
+ },
+ {
+ "epoch": 41.251528384279474,
+ "grad_norm": 0.008650158531963825,
+ "learning_rate": 0.0006,
+ "loss": 7.357326984405518,
+ "step": 2970
+ },
+ {
+ "epoch": 41.26550218340611,
+ "grad_norm": 0.007483262103050947,
+ "learning_rate": 0.0006,
+ "loss": 7.410891056060791,
+ "step": 2971
+ },
+ {
+ "epoch": 41.27947598253275,
+ "grad_norm": 0.00869277399033308,
+ "learning_rate": 0.0006,
+ "loss": 7.392127990722656,
+ "step": 2972
+ },
+ {
+ "epoch": 41.29344978165939,
+ "grad_norm": 0.00885078776627779,
+ "learning_rate": 0.0006,
+ "loss": 7.405449867248535,
+ "step": 2973
+ },
+ {
+ "epoch": 41.30742358078603,
+ "grad_norm": 0.00958226341754198,
+ "learning_rate": 0.0006,
+ "loss": 7.399868488311768,
+ "step": 2974
+ },
+ {
+ "epoch": 41.32139737991266,
+ "grad_norm": 0.006991599686443806,
+ "learning_rate": 0.0006,
+ "loss": 7.35369873046875,
+ "step": 2975
+ },
+ {
+ "epoch": 41.3353711790393,
+ "grad_norm": 0.007415142375975847,
+ "learning_rate": 0.0006,
+ "loss": 7.358353137969971,
+ "step": 2976
+ },
+ {
+ "epoch": 41.34934497816594,
+ "grad_norm": 0.007024371065199375,
+ "learning_rate": 0.0006,
+ "loss": 7.4070892333984375,
+ "step": 2977
+ },
+ {
+ "epoch": 41.36331877729258,
+ "grad_norm": 0.007234332151710987,
+ "learning_rate": 0.0006,
+ "loss": 7.379120349884033,
+ "step": 2978
+ },
+ {
+ "epoch": 41.377292576419215,
+ "grad_norm": 0.00798892229795456,
+ "learning_rate": 0.0006,
+ "loss": 7.365455627441406,
+ "step": 2979
+ },
+ {
+ "epoch": 41.391266375545854,
+ "grad_norm": 0.007303398102521896,
+ "learning_rate": 0.0006,
+ "loss": 7.341798782348633,
+ "step": 2980
+ },
+ {
+ "epoch": 41.40524017467249,
+ "grad_norm": 0.007119669578969479,
+ "learning_rate": 0.0006,
+ "loss": 7.363280296325684,
+ "step": 2981
+ },
+ {
+ "epoch": 41.419213973799124,
+ "grad_norm": 0.007457619532942772,
+ "learning_rate": 0.0006,
+ "loss": 7.39388370513916,
+ "step": 2982
+ },
+ {
+ "epoch": 41.43318777292576,
+ "grad_norm": 0.007905688136816025,
+ "learning_rate": 0.0006,
+ "loss": 7.380652904510498,
+ "step": 2983
+ },
+ {
+ "epoch": 41.4471615720524,
+ "grad_norm": 0.006212855689227581,
+ "learning_rate": 0.0006,
+ "loss": 7.398769378662109,
+ "step": 2984
+ },
+ {
+ "epoch": 41.46113537117904,
+ "grad_norm": 0.0070641543716192245,
+ "learning_rate": 0.0006,
+ "loss": 7.3503546714782715,
+ "step": 2985
+ },
+ {
+ "epoch": 41.47510917030568,
+ "grad_norm": 0.00787805300205946,
+ "learning_rate": 0.0006,
+ "loss": 7.347678184509277,
+ "step": 2986
+ },
+ {
+ "epoch": 41.48908296943232,
+ "grad_norm": 0.006760426796972752,
+ "learning_rate": 0.0006,
+ "loss": 7.380006790161133,
+ "step": 2987
+ },
+ {
+ "epoch": 41.50305676855895,
+ "grad_norm": 0.00772385997697711,
+ "learning_rate": 0.0006,
+ "loss": 7.363680362701416,
+ "step": 2988
+ },
+ {
+ "epoch": 41.51703056768559,
+ "grad_norm": 0.007373564876616001,
+ "learning_rate": 0.0006,
+ "loss": 7.333014488220215,
+ "step": 2989
+ },
+ {
+ "epoch": 41.531004366812226,
+ "grad_norm": 0.006717653013765812,
+ "learning_rate": 0.0006,
+ "loss": 7.360076427459717,
+ "step": 2990
+ },
+ {
+ "epoch": 41.544978165938865,
+ "grad_norm": 0.007191413082182407,
+ "learning_rate": 0.0006,
+ "loss": 7.397030830383301,
+ "step": 2991
+ },
+ {
+ "epoch": 41.5589519650655,
+ "grad_norm": 0.007831960916519165,
+ "learning_rate": 0.0006,
+ "loss": 7.369959831237793,
+ "step": 2992
+ },
+ {
+ "epoch": 41.57292576419214,
+ "grad_norm": 0.006947481073439121,
+ "learning_rate": 0.0006,
+ "loss": 7.333254337310791,
+ "step": 2993
+ },
+ {
+ "epoch": 41.58689956331878,
+ "grad_norm": 0.006243606563657522,
+ "learning_rate": 0.0006,
+ "loss": 7.327378273010254,
+ "step": 2994
+ },
+ {
+ "epoch": 41.60087336244541,
+ "grad_norm": 0.006298339460045099,
+ "learning_rate": 0.0006,
+ "loss": 7.323482036590576,
+ "step": 2995
+ },
+ {
+ "epoch": 41.61484716157205,
+ "grad_norm": 0.007587412837892771,
+ "learning_rate": 0.0006,
+ "loss": 7.313789367675781,
+ "step": 2996
+ },
+ {
+ "epoch": 41.62882096069869,
+ "grad_norm": 0.007140746805816889,
+ "learning_rate": 0.0006,
+ "loss": 7.356067657470703,
+ "step": 2997
+ },
+ {
+ "epoch": 41.64279475982533,
+ "grad_norm": 0.008315625600516796,
+ "learning_rate": 0.0006,
+ "loss": 7.3339643478393555,
+ "step": 2998
+ },
+ {
+ "epoch": 41.65676855895197,
+ "grad_norm": 0.006547102238982916,
+ "learning_rate": 0.0006,
+ "loss": 7.342191696166992,
+ "step": 2999
+ },
+ {
+ "epoch": 41.670742358078606,
+ "grad_norm": 0.006068122107535601,
+ "learning_rate": 0.0006,
+ "loss": 7.321015357971191,
+ "step": 3000
+ },
+ {
+ "epoch": 41.68471615720524,
+ "grad_norm": 0.007316991221159697,
+ "learning_rate": 0.0006,
+ "loss": 7.346490859985352,
+ "step": 3001
+ },
+ {
+ "epoch": 41.698689956331876,
+ "grad_norm": 0.0067388564348220825,
+ "learning_rate": 0.0006,
+ "loss": 7.340723991394043,
+ "step": 3002
+ },
+ {
+ "epoch": 41.712663755458514,
+ "grad_norm": 0.006839963141828775,
+ "learning_rate": 0.0006,
+ "loss": 7.329365253448486,
+ "step": 3003
+ },
+ {
+ "epoch": 41.72663755458515,
+ "grad_norm": 0.00805041566491127,
+ "learning_rate": 0.0006,
+ "loss": 7.298934459686279,
+ "step": 3004
+ },
+ {
+ "epoch": 41.74061135371179,
+ "grad_norm": 0.007197503466159105,
+ "learning_rate": 0.0006,
+ "loss": 7.331496238708496,
+ "step": 3005
+ },
+ {
+ "epoch": 41.75458515283843,
+ "grad_norm": 0.007220805156975985,
+ "learning_rate": 0.0006,
+ "loss": 7.317695140838623,
+ "step": 3006
+ },
+ {
+ "epoch": 41.76855895196506,
+ "grad_norm": 0.00874276738613844,
+ "learning_rate": 0.0006,
+ "loss": 7.272614479064941,
+ "step": 3007
+ },
+ {
+ "epoch": 41.7825327510917,
+ "grad_norm": 0.008386565372347832,
+ "learning_rate": 0.0006,
+ "loss": 7.295244216918945,
+ "step": 3008
+ },
+ {
+ "epoch": 41.79650655021834,
+ "grad_norm": 0.0077730282209813595,
+ "learning_rate": 0.0006,
+ "loss": 7.3395490646362305,
+ "step": 3009
+ },
+ {
+ "epoch": 41.81048034934498,
+ "grad_norm": 0.008441480807960033,
+ "learning_rate": 0.0006,
+ "loss": 7.319921493530273,
+ "step": 3010
+ },
+ {
+ "epoch": 41.82445414847162,
+ "grad_norm": 0.007398269604891539,
+ "learning_rate": 0.0006,
+ "loss": 7.3029465675354,
+ "step": 3011
+ },
+ {
+ "epoch": 41.838427947598255,
+ "grad_norm": 0.006878677289932966,
+ "learning_rate": 0.0006,
+ "loss": 7.318467140197754,
+ "step": 3012
+ },
+ {
+ "epoch": 41.852401746724894,
+ "grad_norm": 0.007616598624736071,
+ "learning_rate": 0.0006,
+ "loss": 7.3112077713012695,
+ "step": 3013
+ },
+ {
+ "epoch": 41.866375545851525,
+ "grad_norm": 0.011629699729382992,
+ "learning_rate": 0.0006,
+ "loss": 7.32523250579834,
+ "step": 3014
+ },
+ {
+ "epoch": 41.880349344978164,
+ "grad_norm": 0.011011249385774136,
+ "learning_rate": 0.0006,
+ "loss": 7.287247657775879,
+ "step": 3015
+ },
+ {
+ "epoch": 41.8943231441048,
+ "grad_norm": 0.010208295658230782,
+ "learning_rate": 0.0006,
+ "loss": 7.292117118835449,
+ "step": 3016
+ },
+ {
+ "epoch": 41.90829694323144,
+ "grad_norm": 0.008651095442473888,
+ "learning_rate": 0.0006,
+ "loss": 7.302201271057129,
+ "step": 3017
+ },
+ {
+ "epoch": 41.92227074235808,
+ "grad_norm": 0.006692920345813036,
+ "learning_rate": 0.0006,
+ "loss": 7.319568634033203,
+ "step": 3018
+ },
+ {
+ "epoch": 41.93624454148472,
+ "grad_norm": 0.010033348575234413,
+ "learning_rate": 0.0006,
+ "loss": 7.2738566398620605,
+ "step": 3019
+ },
+ {
+ "epoch": 41.95021834061135,
+ "grad_norm": 0.011790832504630089,
+ "learning_rate": 0.0006,
+ "loss": 7.298540115356445,
+ "step": 3020
+ },
+ {
+ "epoch": 41.96419213973799,
+ "grad_norm": 0.018622662872076035,
+ "learning_rate": 0.0006,
+ "loss": 7.267182350158691,
+ "step": 3021
+ },
+ {
+ "epoch": 41.97816593886463,
+ "grad_norm": 0.02520257607102394,
+ "learning_rate": 0.0006,
+ "loss": 7.261566162109375,
+ "step": 3022
+ },
+ {
+ "epoch": 41.992139737991266,
+ "grad_norm": 0.036332644522190094,
+ "learning_rate": 0.0006,
+ "loss": 7.270719051361084,
+ "step": 3023
+ },
+ {
+ "epoch": 42.0,
+ "grad_norm": 0.04717305675148964,
+ "learning_rate": 0.0006,
+ "loss": 7.293332099914551,
+ "step": 3024
+ },
+ {
+ "epoch": 42.0,
+ "eval_loss": 7.354280471801758,
+ "eval_runtime": 56.4353,
+ "eval_samples_per_second": 43.271,
+ "eval_steps_per_second": 1.364,
+ "step": 3024
+ },
+ {
+ "epoch": 42.01397379912664,
+ "grad_norm": 0.06661626696586609,
+ "learning_rate": 0.0006,
+ "loss": 7.338189601898193,
+ "step": 3025
+ },
+ {
+ "epoch": 42.02794759825328,
+ "grad_norm": 0.05452054366469383,
+ "learning_rate": 0.0006,
+ "loss": 7.316764831542969,
+ "step": 3026
+ },
+ {
+ "epoch": 42.041921397379916,
+ "grad_norm": 0.025638775900006294,
+ "learning_rate": 0.0006,
+ "loss": 7.34204626083374,
+ "step": 3027
+ },
+ {
+ "epoch": 42.05589519650655,
+ "grad_norm": 0.022398799657821655,
+ "learning_rate": 0.0006,
+ "loss": 7.262149333953857,
+ "step": 3028
+ },
+ {
+ "epoch": 42.069868995633186,
+ "grad_norm": 0.03713499382138252,
+ "learning_rate": 0.0006,
+ "loss": 7.27695369720459,
+ "step": 3029
+ },
+ {
+ "epoch": 42.083842794759825,
+ "grad_norm": 0.014949247241020203,
+ "learning_rate": 0.0006,
+ "loss": 7.296508312225342,
+ "step": 3030
+ },
+ {
+ "epoch": 42.09781659388646,
+ "grad_norm": 0.02703995071351528,
+ "learning_rate": 0.0006,
+ "loss": 7.258018493652344,
+ "step": 3031
+ },
+ {
+ "epoch": 42.1117903930131,
+ "grad_norm": 0.020352816209197044,
+ "learning_rate": 0.0006,
+ "loss": 7.241450309753418,
+ "step": 3032
+ },
+ {
+ "epoch": 42.12576419213974,
+ "grad_norm": 0.014578479342162609,
+ "learning_rate": 0.0006,
+ "loss": 7.257323265075684,
+ "step": 3033
+ },
+ {
+ "epoch": 42.13973799126637,
+ "grad_norm": 0.026053044945001602,
+ "learning_rate": 0.0006,
+ "loss": 7.254680633544922,
+ "step": 3034
+ },
+ {
+ "epoch": 42.15371179039301,
+ "grad_norm": 0.015120275318622589,
+ "learning_rate": 0.0006,
+ "loss": 7.261074066162109,
+ "step": 3035
+ },
+ {
+ "epoch": 42.16768558951965,
+ "grad_norm": 0.016822177916765213,
+ "learning_rate": 0.0006,
+ "loss": 7.279064178466797,
+ "step": 3036
+ },
+ {
+ "epoch": 42.18165938864629,
+ "grad_norm": 0.016367588192224503,
+ "learning_rate": 0.0006,
+ "loss": 7.251429557800293,
+ "step": 3037
+ },
+ {
+ "epoch": 42.19563318777293,
+ "grad_norm": 0.014350418001413345,
+ "learning_rate": 0.0006,
+ "loss": 7.259264945983887,
+ "step": 3038
+ },
+ {
+ "epoch": 42.209606986899566,
+ "grad_norm": 0.017078228294849396,
+ "learning_rate": 0.0006,
+ "loss": 7.227105140686035,
+ "step": 3039
+ },
+ {
+ "epoch": 42.223580786026204,
+ "grad_norm": 0.011759317480027676,
+ "learning_rate": 0.0006,
+ "loss": 7.275888442993164,
+ "step": 3040
+ },
+ {
+ "epoch": 42.237554585152836,
+ "grad_norm": 0.013019340112805367,
+ "learning_rate": 0.0006,
+ "loss": 7.2517008781433105,
+ "step": 3041
+ },
+ {
+ "epoch": 42.251528384279474,
+ "grad_norm": 0.01178441196680069,
+ "learning_rate": 0.0006,
+ "loss": 7.23576545715332,
+ "step": 3042
+ },
+ {
+ "epoch": 42.26550218340611,
+ "grad_norm": 0.014518891461193562,
+ "learning_rate": 0.0006,
+ "loss": 7.253895282745361,
+ "step": 3043
+ },
+ {
+ "epoch": 42.27947598253275,
+ "grad_norm": 0.018159950152039528,
+ "learning_rate": 0.0006,
+ "loss": 7.278058052062988,
+ "step": 3044
+ },
+ {
+ "epoch": 42.29344978165939,
+ "grad_norm": 0.007587796077132225,
+ "learning_rate": 0.0006,
+ "loss": 7.278578758239746,
+ "step": 3045
+ },
+ {
+ "epoch": 42.30742358078603,
+ "grad_norm": 0.015994979068636894,
+ "learning_rate": 0.0006,
+ "loss": 7.239841461181641,
+ "step": 3046
+ },
+ {
+ "epoch": 42.32139737991266,
+ "grad_norm": 0.017830168828368187,
+ "learning_rate": 0.0006,
+ "loss": 7.2443976402282715,
+ "step": 3047
+ },
+ {
+ "epoch": 42.3353711790393,
+ "grad_norm": 0.008738759905099869,
+ "learning_rate": 0.0006,
+ "loss": 7.232264518737793,
+ "step": 3048
+ },
+ {
+ "epoch": 42.34934497816594,
+ "grad_norm": 0.0175449438393116,
+ "learning_rate": 0.0006,
+ "loss": 7.228000640869141,
+ "step": 3049
+ },
+ {
+ "epoch": 42.36331877729258,
+ "grad_norm": 0.019611820578575134,
+ "learning_rate": 0.0006,
+ "loss": 7.234999179840088,
+ "step": 3050
+ },
+ {
+ "epoch": 42.377292576419215,
+ "grad_norm": 0.015302863903343678,
+ "learning_rate": 0.0006,
+ "loss": 7.242547988891602,
+ "step": 3051
+ },
+ {
+ "epoch": 42.391266375545854,
+ "grad_norm": 0.025919731706380844,
+ "learning_rate": 0.0006,
+ "loss": 7.216652870178223,
+ "step": 3052
+ },
+ {
+ "epoch": 42.40524017467249,
+ "grad_norm": 0.034587468951940536,
+ "learning_rate": 0.0006,
+ "loss": 7.245336532592773,
+ "step": 3053
+ },
+ {
+ "epoch": 42.419213973799124,
+ "grad_norm": 0.019198840484023094,
+ "learning_rate": 0.0006,
+ "loss": 7.256758689880371,
+ "step": 3054
+ },
+ {
+ "epoch": 42.43318777292576,
+ "grad_norm": 0.012316389009356499,
+ "learning_rate": 0.0006,
+ "loss": 7.209102630615234,
+ "step": 3055
+ },
+ {
+ "epoch": 42.4471615720524,
+ "grad_norm": 0.023512162268161774,
+ "learning_rate": 0.0006,
+ "loss": 7.250742435455322,
+ "step": 3056
+ },
+ {
+ "epoch": 42.46113537117904,
+ "grad_norm": 0.025457851588726044,
+ "learning_rate": 0.0006,
+ "loss": 7.263757705688477,
+ "step": 3057
+ },
+ {
+ "epoch": 42.47510917030568,
+ "grad_norm": 0.02192617766559124,
+ "learning_rate": 0.0006,
+ "loss": 7.246977806091309,
+ "step": 3058
+ },
+ {
+ "epoch": 42.48908296943232,
+ "grad_norm": 0.02541281282901764,
+ "learning_rate": 0.0006,
+ "loss": 7.232361316680908,
+ "step": 3059
+ },
+ {
+ "epoch": 42.50305676855895,
+ "grad_norm": 0.036353763192892075,
+ "learning_rate": 0.0006,
+ "loss": 7.242556571960449,
+ "step": 3060
+ },
+ {
+ "epoch": 42.51703056768559,
+ "grad_norm": 0.04555872827768326,
+ "learning_rate": 0.0006,
+ "loss": 7.228940963745117,
+ "step": 3061
+ },
+ {
+ "epoch": 42.531004366812226,
+ "grad_norm": 0.04563378915190697,
+ "learning_rate": 0.0006,
+ "loss": 7.235276222229004,
+ "step": 3062
+ },
+ {
+ "epoch": 42.544978165938865,
+ "grad_norm": 0.04275760427117348,
+ "learning_rate": 0.0006,
+ "loss": 7.238895416259766,
+ "step": 3063
+ },
+ {
+ "epoch": 42.5589519650655,
+ "grad_norm": 0.04535885155200958,
+ "learning_rate": 0.0006,
+ "loss": 7.205336570739746,
+ "step": 3064
+ },
+ {
+ "epoch": 42.57292576419214,
+ "grad_norm": 0.015534617006778717,
+ "learning_rate": 0.0006,
+ "loss": 7.238393783569336,
+ "step": 3065
+ },
+ {
+ "epoch": 42.58689956331878,
+ "grad_norm": 0.03286299481987953,
+ "learning_rate": 0.0006,
+ "loss": 7.216453552246094,
+ "step": 3066
+ },
+ {
+ "epoch": 42.60087336244541,
+ "grad_norm": 0.025453168898820877,
+ "learning_rate": 0.0006,
+ "loss": 7.203750133514404,
+ "step": 3067
+ },
+ {
+ "epoch": 42.61484716157205,
+ "grad_norm": 0.021501649171113968,
+ "learning_rate": 0.0006,
+ "loss": 7.241621971130371,
+ "step": 3068
+ },
+ {
+ "epoch": 42.62882096069869,
+ "grad_norm": 0.02743634767830372,
+ "learning_rate": 0.0006,
+ "loss": 7.22076940536499,
+ "step": 3069
+ },
+ {
+ "epoch": 42.64279475982533,
+ "grad_norm": 0.020890483632683754,
+ "learning_rate": 0.0006,
+ "loss": 7.218412399291992,
+ "step": 3070
+ },
+ {
+ "epoch": 42.65676855895197,
+ "grad_norm": 0.022499850019812584,
+ "learning_rate": 0.0006,
+ "loss": 7.217696189880371,
+ "step": 3071
+ },
+ {
+ "epoch": 42.670742358078606,
+ "grad_norm": 0.01653696969151497,
+ "learning_rate": 0.0006,
+ "loss": 7.156417369842529,
+ "step": 3072
+ },
+ {
+ "epoch": 42.68471615720524,
+ "grad_norm": 0.01834225095808506,
+ "learning_rate": 0.0006,
+ "loss": 7.211247444152832,
+ "step": 3073
+ },
+ {
+ "epoch": 42.698689956331876,
+ "grad_norm": 0.020137229934334755,
+ "learning_rate": 0.0006,
+ "loss": 7.232499599456787,
+ "step": 3074
+ },
+ {
+ "epoch": 42.712663755458514,
+ "grad_norm": 0.012671459466218948,
+ "learning_rate": 0.0006,
+ "loss": 7.172441482543945,
+ "step": 3075
+ },
+ {
+ "epoch": 42.72663755458515,
+ "grad_norm": 0.016803130507469177,
+ "learning_rate": 0.0006,
+ "loss": 7.152103424072266,
+ "step": 3076
+ },
+ {
+ "epoch": 42.74061135371179,
+ "grad_norm": 0.01623712107539177,
+ "learning_rate": 0.0006,
+ "loss": 7.187032699584961,
+ "step": 3077
+ },
+ {
+ "epoch": 42.75458515283843,
+ "grad_norm": 0.015938356518745422,
+ "learning_rate": 0.0006,
+ "loss": 7.171140670776367,
+ "step": 3078
+ },
+ {
+ "epoch": 42.76855895196506,
+ "grad_norm": 0.02580736204981804,
+ "learning_rate": 0.0006,
+ "loss": 7.204472541809082,
+ "step": 3079
+ },
+ {
+ "epoch": 42.7825327510917,
+ "grad_norm": 0.023372117429971695,
+ "learning_rate": 0.0006,
+ "loss": 7.165554523468018,
+ "step": 3080
+ },
+ {
+ "epoch": 42.79650655021834,
+ "grad_norm": 0.04440586268901825,
+ "learning_rate": 0.0006,
+ "loss": 7.169431209564209,
+ "step": 3081
+ },
+ {
+ "epoch": 42.81048034934498,
+ "grad_norm": 0.06797406077384949,
+ "learning_rate": 0.0006,
+ "loss": 7.168193340301514,
+ "step": 3082
+ },
+ {
+ "epoch": 42.82445414847162,
+ "grad_norm": 0.0914321169257164,
+ "learning_rate": 0.0006,
+ "loss": 7.204671859741211,
+ "step": 3083
+ },
+ {
+ "epoch": 42.838427947598255,
+ "grad_norm": 0.0506940595805645,
+ "learning_rate": 0.0006,
+ "loss": 7.15949821472168,
+ "step": 3084
+ },
+ {
+ "epoch": 42.852401746724894,
+ "grad_norm": 0.04819890484213829,
+ "learning_rate": 0.0006,
+ "loss": 7.192837238311768,
+ "step": 3085
+ },
+ {
+ "epoch": 42.866375545851525,
+ "grad_norm": 0.040044233202934265,
+ "learning_rate": 0.0006,
+ "loss": 7.1600165367126465,
+ "step": 3086
+ },
+ {
+ "epoch": 42.880349344978164,
+ "grad_norm": 0.02573299966752529,
+ "learning_rate": 0.0006,
+ "loss": 7.153427600860596,
+ "step": 3087
+ },
+ {
+ "epoch": 42.8943231441048,
+ "grad_norm": 0.046942535787820816,
+ "learning_rate": 0.0006,
+ "loss": 7.16767692565918,
+ "step": 3088
+ },
+ {
+ "epoch": 42.90829694323144,
+ "grad_norm": 0.02683161571621895,
+ "learning_rate": 0.0006,
+ "loss": 7.17144250869751,
+ "step": 3089
+ },
+ {
+ "epoch": 42.92227074235808,
+ "grad_norm": 0.05363674834370613,
+ "learning_rate": 0.0006,
+ "loss": 7.120128631591797,
+ "step": 3090
+ },
+ {
+ "epoch": 42.93624454148472,
+ "grad_norm": 0.023756247013807297,
+ "learning_rate": 0.0006,
+ "loss": 7.169703006744385,
+ "step": 3091
+ },
+ {
+ "epoch": 42.95021834061135,
+ "grad_norm": 0.04216783866286278,
+ "learning_rate": 0.0006,
+ "loss": 7.1206865310668945,
+ "step": 3092
+ },
+ {
+ "epoch": 42.96419213973799,
+ "grad_norm": 0.02219565585255623,
+ "learning_rate": 0.0006,
+ "loss": 7.157215118408203,
+ "step": 3093
+ },
+ {
+ "epoch": 42.97816593886463,
+ "grad_norm": 0.03527127206325531,
+ "learning_rate": 0.0006,
+ "loss": 7.166672706604004,
+ "step": 3094
+ },
+ {
+ "epoch": 42.992139737991266,
+ "grad_norm": 0.017748817801475525,
+ "learning_rate": 0.0006,
+ "loss": 7.154747009277344,
+ "step": 3095
+ },
+ {
+ "epoch": 43.0,
+ "grad_norm": 0.033165838569402695,
+ "learning_rate": 0.0006,
+ "loss": 7.150866508483887,
+ "step": 3096
+ },
+ {
+ "epoch": 43.0,
+ "eval_loss": 7.144403457641602,
+ "eval_runtime": 56.8853,
+ "eval_samples_per_second": 42.928,
+ "eval_steps_per_second": 1.354,
+ "step": 3096
+ },
+ {
+ "epoch": 43.01397379912664,
+ "grad_norm": 0.015199829824268818,
+ "learning_rate": 0.0006,
+ "loss": 7.122735023498535,
+ "step": 3097
+ },
+ {
+ "epoch": 43.02794759825328,
+ "grad_norm": 0.03220256045460701,
+ "learning_rate": 0.0006,
+ "loss": 7.143553733825684,
+ "step": 3098
+ },
+ {
+ "epoch": 43.041921397379916,
+ "grad_norm": 0.014687920920550823,
+ "learning_rate": 0.0006,
+ "loss": 7.080960750579834,
+ "step": 3099
+ },
+ {
+ "epoch": 43.05589519650655,
+ "grad_norm": 0.023522932082414627,
+ "learning_rate": 0.0006,
+ "loss": 7.137356758117676,
+ "step": 3100
+ },
+ {
+ "epoch": 43.069868995633186,
+ "grad_norm": 0.014808216132223606,
+ "learning_rate": 0.0006,
+ "loss": 7.09143590927124,
+ "step": 3101
+ },
+ {
+ "epoch": 43.083842794759825,
+ "grad_norm": 0.029339246451854706,
+ "learning_rate": 0.0006,
+ "loss": 7.122198104858398,
+ "step": 3102
+ },
+ {
+ "epoch": 43.09781659388646,
+ "grad_norm": 0.022953078150749207,
+ "learning_rate": 0.0006,
+ "loss": 7.092505931854248,
+ "step": 3103
+ },
+ {
+ "epoch": 43.1117903930131,
+ "grad_norm": 0.04590052738785744,
+ "learning_rate": 0.0006,
+ "loss": 7.113100051879883,
+ "step": 3104
+ },
+ {
+ "epoch": 43.12576419213974,
+ "grad_norm": 0.0524442233145237,
+ "learning_rate": 0.0006,
+ "loss": 7.147022247314453,
+ "step": 3105
+ },
+ {
+ "epoch": 43.13973799126637,
+ "grad_norm": 0.0315794013440609,
+ "learning_rate": 0.0006,
+ "loss": 7.094045162200928,
+ "step": 3106
+ },
+ {
+ "epoch": 43.15371179039301,
+ "grad_norm": 0.05112722888588905,
+ "learning_rate": 0.0006,
+ "loss": 7.054388523101807,
+ "step": 3107
+ },
+ {
+ "epoch": 43.16768558951965,
+ "grad_norm": 0.03146776929497719,
+ "learning_rate": 0.0006,
+ "loss": 7.116674423217773,
+ "step": 3108
+ },
+ {
+ "epoch": 43.18165938864629,
+ "grad_norm": 0.03391246497631073,
+ "learning_rate": 0.0006,
+ "loss": 7.150345802307129,
+ "step": 3109
+ },
+ {
+ "epoch": 43.19563318777293,
+ "grad_norm": 0.026957903057336807,
+ "learning_rate": 0.0006,
+ "loss": 7.058531284332275,
+ "step": 3110
+ },
+ {
+ "epoch": 43.209606986899566,
+ "grad_norm": 0.041532568633556366,
+ "learning_rate": 0.0006,
+ "loss": 7.085498809814453,
+ "step": 3111
+ },
+ {
+ "epoch": 43.223580786026204,
+ "grad_norm": 0.020001588389277458,
+ "learning_rate": 0.0006,
+ "loss": 7.131487846374512,
+ "step": 3112
+ },
+ {
+ "epoch": 43.237554585152836,
+ "grad_norm": 0.02268712781369686,
+ "learning_rate": 0.0006,
+ "loss": 7.0790863037109375,
+ "step": 3113
+ },
+ {
+ "epoch": 43.251528384279474,
+ "grad_norm": 0.023363925516605377,
+ "learning_rate": 0.0006,
+ "loss": 7.067816734313965,
+ "step": 3114
+ },
+ {
+ "epoch": 43.26550218340611,
+ "grad_norm": 0.019367508590221405,
+ "learning_rate": 0.0006,
+ "loss": 7.042041778564453,
+ "step": 3115
+ },
+ {
+ "epoch": 43.27947598253275,
+ "grad_norm": 0.02745090238749981,
+ "learning_rate": 0.0006,
+ "loss": 7.064733028411865,
+ "step": 3116
+ },
+ {
+ "epoch": 43.29344978165939,
+ "grad_norm": 0.02543802745640278,
+ "learning_rate": 0.0006,
+ "loss": 7.058948040008545,
+ "step": 3117
+ },
+ {
+ "epoch": 43.30742358078603,
+ "grad_norm": 0.018874092027544975,
+ "learning_rate": 0.0006,
+ "loss": 7.009946346282959,
+ "step": 3118
+ },
+ {
+ "epoch": 43.32139737991266,
+ "grad_norm": 0.024662822484970093,
+ "learning_rate": 0.0006,
+ "loss": 7.015913963317871,
+ "step": 3119
+ },
+ {
+ "epoch": 43.3353711790393,
+ "grad_norm": 0.02441416308283806,
+ "learning_rate": 0.0006,
+ "loss": 7.061622142791748,
+ "step": 3120
+ },
+ {
+ "epoch": 43.34934497816594,
+ "grad_norm": 0.031887322664260864,
+ "learning_rate": 0.0006,
+ "loss": 7.046883583068848,
+ "step": 3121
+ },
+ {
+ "epoch": 43.36331877729258,
+ "grad_norm": 0.021930739283561707,
+ "learning_rate": 0.0006,
+ "loss": 7.0309624671936035,
+ "step": 3122
+ },
+ {
+ "epoch": 43.377292576419215,
+ "grad_norm": 0.02722783200442791,
+ "learning_rate": 0.0006,
+ "loss": 7.072924613952637,
+ "step": 3123
+ },
+ {
+ "epoch": 43.391266375545854,
+ "grad_norm": 0.01483179721981287,
+ "learning_rate": 0.0006,
+ "loss": 7.028624534606934,
+ "step": 3124
+ },
+ {
+ "epoch": 43.40524017467249,
+ "grad_norm": 0.024793945252895355,
+ "learning_rate": 0.0006,
+ "loss": 7.001623153686523,
+ "step": 3125
+ },
+ {
+ "epoch": 43.419213973799124,
+ "grad_norm": 0.020148713141679764,
+ "learning_rate": 0.0006,
+ "loss": 6.9938812255859375,
+ "step": 3126
+ },
+ {
+ "epoch": 43.43318777292576,
+ "grad_norm": 0.02413850650191307,
+ "learning_rate": 0.0006,
+ "loss": 7.001243591308594,
+ "step": 3127
+ },
+ {
+ "epoch": 43.4471615720524,
+ "grad_norm": 0.028549203649163246,
+ "learning_rate": 0.0006,
+ "loss": 6.987671375274658,
+ "step": 3128
+ },
+ {
+ "epoch": 43.46113537117904,
+ "grad_norm": 0.025988785549998283,
+ "learning_rate": 0.0006,
+ "loss": 6.9551167488098145,
+ "step": 3129
+ },
+ {
+ "epoch": 43.47510917030568,
+ "grad_norm": 0.027140729129314423,
+ "learning_rate": 0.0006,
+ "loss": 6.9547810554504395,
+ "step": 3130
+ },
+ {
+ "epoch": 43.48908296943232,
+ "grad_norm": 0.030130082741379738,
+ "learning_rate": 0.0006,
+ "loss": 6.993186950683594,
+ "step": 3131
+ },
+ {
+ "epoch": 43.50305676855895,
+ "grad_norm": 0.025996994227170944,
+ "learning_rate": 0.0006,
+ "loss": 6.984624862670898,
+ "step": 3132
+ },
+ {
+ "epoch": 43.51703056768559,
+ "grad_norm": 0.02423008903861046,
+ "learning_rate": 0.0006,
+ "loss": 6.98813009262085,
+ "step": 3133
+ },
+ {
+ "epoch": 43.531004366812226,
+ "grad_norm": 0.029665837064385414,
+ "learning_rate": 0.0006,
+ "loss": 6.986610412597656,
+ "step": 3134
+ },
+ {
+ "epoch": 43.544978165938865,
+ "grad_norm": 0.028768595308065414,
+ "learning_rate": 0.0006,
+ "loss": 6.898797035217285,
+ "step": 3135
+ },
+ {
+ "epoch": 43.5589519650655,
+ "grad_norm": 0.03111208975315094,
+ "learning_rate": 0.0006,
+ "loss": 6.923741817474365,
+ "step": 3136
+ },
+ {
+ "epoch": 43.57292576419214,
+ "grad_norm": 0.0368187390267849,
+ "learning_rate": 0.0006,
+ "loss": 6.994343280792236,
+ "step": 3137
+ },
+ {
+ "epoch": 43.58689956331878,
+ "grad_norm": 0.04076763987541199,
+ "learning_rate": 0.0006,
+ "loss": 6.909458160400391,
+ "step": 3138
+ },
+ {
+ "epoch": 43.60087336244541,
+ "grad_norm": 0.03281061723828316,
+ "learning_rate": 0.0006,
+ "loss": 6.924851417541504,
+ "step": 3139
+ },
+ {
+ "epoch": 43.61484716157205,
+ "grad_norm": 0.025486988946795464,
+ "learning_rate": 0.0006,
+ "loss": 6.962926864624023,
+ "step": 3140
+ },
+ {
+ "epoch": 43.62882096069869,
+ "grad_norm": 0.04583485797047615,
+ "learning_rate": 0.0006,
+ "loss": 6.876555442810059,
+ "step": 3141
+ },
+ {
+ "epoch": 43.64279475982533,
+ "grad_norm": 0.05460785701870918,
+ "learning_rate": 0.0006,
+ "loss": 6.952932357788086,
+ "step": 3142
+ },
+ {
+ "epoch": 43.65676855895197,
+ "grad_norm": 0.040917906910181046,
+ "learning_rate": 0.0006,
+ "loss": 6.9624924659729,
+ "step": 3143
+ },
+ {
+ "epoch": 43.670742358078606,
+ "grad_norm": 0.04009668901562691,
+ "learning_rate": 0.0006,
+ "loss": 6.94033145904541,
+ "step": 3144
+ },
+ {
+ "epoch": 43.68471615720524,
+ "grad_norm": 0.04500625655055046,
+ "learning_rate": 0.0006,
+ "loss": 6.961839199066162,
+ "step": 3145
+ },
+ {
+ "epoch": 43.698689956331876,
+ "grad_norm": 0.03182258829474449,
+ "learning_rate": 0.0006,
+ "loss": 6.930580139160156,
+ "step": 3146
+ },
+ {
+ "epoch": 43.712663755458514,
+ "grad_norm": 0.053621504455804825,
+ "learning_rate": 0.0006,
+ "loss": 6.950235366821289,
+ "step": 3147
+ },
+ {
+ "epoch": 43.72663755458515,
+ "grad_norm": 0.031919289380311966,
+ "learning_rate": 0.0006,
+ "loss": 6.895111560821533,
+ "step": 3148
+ },
+ {
+ "epoch": 43.74061135371179,
+ "grad_norm": 0.042713891714811325,
+ "learning_rate": 0.0006,
+ "loss": 6.940885543823242,
+ "step": 3149
+ },
+ {
+ "epoch": 43.75458515283843,
+ "grad_norm": 0.02385653182864189,
+ "learning_rate": 0.0006,
+ "loss": 6.948801517486572,
+ "step": 3150
+ },
+ {
+ "epoch": 43.76855895196506,
+ "grad_norm": 0.04100663587450981,
+ "learning_rate": 0.0006,
+ "loss": 6.970411777496338,
+ "step": 3151
+ },
+ {
+ "epoch": 43.7825327510917,
+ "grad_norm": 0.02729583904147148,
+ "learning_rate": 0.0006,
+ "loss": 6.899379253387451,
+ "step": 3152
+ },
+ {
+ "epoch": 43.79650655021834,
+ "grad_norm": 0.03128620237112045,
+ "learning_rate": 0.0006,
+ "loss": 6.931319236755371,
+ "step": 3153
+ },
+ {
+ "epoch": 43.81048034934498,
+ "grad_norm": 0.025488659739494324,
+ "learning_rate": 0.0006,
+ "loss": 6.8486409187316895,
+ "step": 3154
+ },
+ {
+ "epoch": 43.82445414847162,
+ "grad_norm": 0.03679376840591431,
+ "learning_rate": 0.0006,
+ "loss": 6.840811252593994,
+ "step": 3155
+ },
+ {
+ "epoch": 43.838427947598255,
+ "grad_norm": 0.020578034222126007,
+ "learning_rate": 0.0006,
+ "loss": 6.9088616371154785,
+ "step": 3156
+ },
+ {
+ "epoch": 43.852401746724894,
+ "grad_norm": 0.02935841865837574,
+ "learning_rate": 0.0006,
+ "loss": 6.806676864624023,
+ "step": 3157
+ },
+ {
+ "epoch": 43.866375545851525,
+ "grad_norm": 0.023319244384765625,
+ "learning_rate": 0.0006,
+ "loss": 6.878457546234131,
+ "step": 3158
+ },
+ {
+ "epoch": 43.880349344978164,
+ "grad_norm": 0.02779589593410492,
+ "learning_rate": 0.0006,
+ "loss": 6.822447776794434,
+ "step": 3159
+ },
+ {
+ "epoch": 43.8943231441048,
+ "grad_norm": 0.01879825070500374,
+ "learning_rate": 0.0006,
+ "loss": 6.854121208190918,
+ "step": 3160
+ },
+ {
+ "epoch": 43.90829694323144,
+ "grad_norm": 0.023286139592528343,
+ "learning_rate": 0.0006,
+ "loss": 6.836827278137207,
+ "step": 3161
+ },
+ {
+ "epoch": 43.92227074235808,
+ "grad_norm": 0.030358297750353813,
+ "learning_rate": 0.0006,
+ "loss": 6.860037803649902,
+ "step": 3162
+ },
+ {
+ "epoch": 43.93624454148472,
+ "grad_norm": 0.029621105641126633,
+ "learning_rate": 0.0006,
+ "loss": 6.884025573730469,
+ "step": 3163
+ },
+ {
+ "epoch": 43.95021834061135,
+ "grad_norm": 0.029829824343323708,
+ "learning_rate": 0.0006,
+ "loss": 6.8702392578125,
+ "step": 3164
+ },
+ {
+ "epoch": 43.96419213973799,
+ "grad_norm": 0.036892473697662354,
+ "learning_rate": 0.0006,
+ "loss": 6.806652069091797,
+ "step": 3165
+ },
+ {
+ "epoch": 43.97816593886463,
+ "grad_norm": 0.03206709027290344,
+ "learning_rate": 0.0006,
+ "loss": 6.828495979309082,
+ "step": 3166
+ },
+ {
+ "epoch": 43.992139737991266,
+ "grad_norm": 0.02744000405073166,
+ "learning_rate": 0.0006,
+ "loss": 6.832002639770508,
+ "step": 3167
+ },
+ {
+ "epoch": 44.0,
+ "grad_norm": 0.02205660194158554,
+ "learning_rate": 0.0006,
+ "loss": 6.898828506469727,
+ "step": 3168
+ },
+ {
+ "epoch": 44.0,
+ "eval_loss": 6.8150553703308105,
+ "eval_runtime": 56.784,
+ "eval_samples_per_second": 43.005,
+ "eval_steps_per_second": 1.356,
+ "step": 3168
+ },
+ {
+ "epoch": 44.01397379912664,
+ "grad_norm": 0.030421260744333267,
+ "learning_rate": 0.0006,
+ "loss": 6.775057792663574,
+ "step": 3169
+ },
+ {
+ "epoch": 44.02794759825328,
+ "grad_norm": 0.0465611033141613,
+ "learning_rate": 0.0006,
+ "loss": 6.849390506744385,
+ "step": 3170
+ },
+ {
+ "epoch": 44.041921397379916,
+ "grad_norm": 0.05977972596883774,
+ "learning_rate": 0.0006,
+ "loss": 6.725576400756836,
+ "step": 3171
+ },
+ {
+ "epoch": 44.05589519650655,
+ "grad_norm": 0.05432697385549545,
+ "learning_rate": 0.0006,
+ "loss": 6.768024444580078,
+ "step": 3172
+ },
+ {
+ "epoch": 44.069868995633186,
+ "grad_norm": 0.04850967228412628,
+ "learning_rate": 0.0006,
+ "loss": 6.799457550048828,
+ "step": 3173
+ },
+ {
+ "epoch": 44.083842794759825,
+ "grad_norm": 0.06201786920428276,
+ "learning_rate": 0.0006,
+ "loss": 6.897234916687012,
+ "step": 3174
+ },
+ {
+ "epoch": 44.09781659388646,
+ "grad_norm": 0.04695591703057289,
+ "learning_rate": 0.0006,
+ "loss": 6.826470375061035,
+ "step": 3175
+ },
+ {
+ "epoch": 44.1117903930131,
+ "grad_norm": 0.03850164636969566,
+ "learning_rate": 0.0006,
+ "loss": 6.835901260375977,
+ "step": 3176
+ },
+ {
+ "epoch": 44.12576419213974,
+ "grad_norm": 0.04250155761837959,
+ "learning_rate": 0.0006,
+ "loss": 6.782310485839844,
+ "step": 3177
+ },
+ {
+ "epoch": 44.13973799126637,
+ "grad_norm": 0.041465479880571365,
+ "learning_rate": 0.0006,
+ "loss": 6.785732746124268,
+ "step": 3178
+ },
+ {
+ "epoch": 44.15371179039301,
+ "grad_norm": 0.02527807280421257,
+ "learning_rate": 0.0006,
+ "loss": 6.78061580657959,
+ "step": 3179
+ },
+ {
+ "epoch": 44.16768558951965,
+ "grad_norm": 0.03660374879837036,
+ "learning_rate": 0.0006,
+ "loss": 6.783992767333984,
+ "step": 3180
+ },
+ {
+ "epoch": 44.18165938864629,
+ "grad_norm": 0.027567964047193527,
+ "learning_rate": 0.0006,
+ "loss": 6.765413284301758,
+ "step": 3181
+ },
+ {
+ "epoch": 44.19563318777293,
+ "grad_norm": 0.024301424622535706,
+ "learning_rate": 0.0006,
+ "loss": 6.760368347167969,
+ "step": 3182
+ },
+ {
+ "epoch": 44.209606986899566,
+ "grad_norm": 0.0266821701079607,
+ "learning_rate": 0.0006,
+ "loss": 6.71417236328125,
+ "step": 3183
+ },
+ {
+ "epoch": 44.223580786026204,
+ "grad_norm": 0.028125090524554253,
+ "learning_rate": 0.0006,
+ "loss": 6.646126747131348,
+ "step": 3184
+ },
+ {
+ "epoch": 44.237554585152836,
+ "grad_norm": 0.0185737032443285,
+ "learning_rate": 0.0006,
+ "loss": 6.712944030761719,
+ "step": 3185
+ },
+ {
+ "epoch": 44.251528384279474,
+ "grad_norm": 0.025836674496531487,
+ "learning_rate": 0.0006,
+ "loss": 6.707030296325684,
+ "step": 3186
+ },
+ {
+ "epoch": 44.26550218340611,
+ "grad_norm": 0.029744870960712433,
+ "learning_rate": 0.0006,
+ "loss": 6.6271071434021,
+ "step": 3187
+ },
+ {
+ "epoch": 44.27947598253275,
+ "grad_norm": 0.03021426685154438,
+ "learning_rate": 0.0006,
+ "loss": 6.677826404571533,
+ "step": 3188
+ },
+ {
+ "epoch": 44.29344978165939,
+ "grad_norm": 0.032727014273405075,
+ "learning_rate": 0.0006,
+ "loss": 6.657764434814453,
+ "step": 3189
+ },
+ {
+ "epoch": 44.30742358078603,
+ "grad_norm": 0.03696778789162636,
+ "learning_rate": 0.0006,
+ "loss": 6.632883548736572,
+ "step": 3190
+ },
+ {
+ "epoch": 44.32139737991266,
+ "grad_norm": 0.034630827605724335,
+ "learning_rate": 0.0006,
+ "loss": 6.759583473205566,
+ "step": 3191
+ },
+ {
+ "epoch": 44.3353711790393,
+ "grad_norm": 0.030517758801579475,
+ "learning_rate": 0.0006,
+ "loss": 6.702268600463867,
+ "step": 3192
+ },
+ {
+ "epoch": 44.34934497816594,
+ "grad_norm": 0.04105190932750702,
+ "learning_rate": 0.0006,
+ "loss": 6.705792427062988,
+ "step": 3193
+ },
+ {
+ "epoch": 44.36331877729258,
+ "grad_norm": 0.06659174710512161,
+ "learning_rate": 0.0006,
+ "loss": 6.73862361907959,
+ "step": 3194
+ },
+ {
+ "epoch": 44.377292576419215,
+ "grad_norm": 0.08013619482517242,
+ "learning_rate": 0.0006,
+ "loss": 6.710282802581787,
+ "step": 3195
+ },
+ {
+ "epoch": 44.391266375545854,
+ "grad_norm": 0.049850963056087494,
+ "learning_rate": 0.0006,
+ "loss": 6.684974670410156,
+ "step": 3196
+ },
+ {
+ "epoch": 44.40524017467249,
+ "grad_norm": 0.061247408390045166,
+ "learning_rate": 0.0006,
+ "loss": 6.716034412384033,
+ "step": 3197
+ },
+ {
+ "epoch": 44.419213973799124,
+ "grad_norm": 0.05868857726454735,
+ "learning_rate": 0.0006,
+ "loss": 6.589107513427734,
+ "step": 3198
+ },
+ {
+ "epoch": 44.43318777292576,
+ "grad_norm": 0.048878561705350876,
+ "learning_rate": 0.0006,
+ "loss": 6.694855690002441,
+ "step": 3199
+ },
+ {
+ "epoch": 44.4471615720524,
+ "grad_norm": 0.028748324140906334,
+ "learning_rate": 0.0006,
+ "loss": 6.629969596862793,
+ "step": 3200
+ },
+ {
+ "epoch": 44.46113537117904,
+ "grad_norm": 0.04060306400060654,
+ "learning_rate": 0.0006,
+ "loss": 6.721431255340576,
+ "step": 3201
+ },
+ {
+ "epoch": 44.47510917030568,
+ "grad_norm": 0.03833824768662453,
+ "learning_rate": 0.0006,
+ "loss": 6.625029563903809,
+ "step": 3202
+ },
+ {
+ "epoch": 44.48908296943232,
+ "grad_norm": 0.03791997581720352,
+ "learning_rate": 0.0006,
+ "loss": 6.669187545776367,
+ "step": 3203
+ },
+ {
+ "epoch": 44.50305676855895,
+ "grad_norm": 0.03259684890508652,
+ "learning_rate": 0.0006,
+ "loss": 6.665643692016602,
+ "step": 3204
+ },
+ {
+ "epoch": 44.51703056768559,
+ "grad_norm": 0.040658626705408096,
+ "learning_rate": 0.0006,
+ "loss": 6.639796257019043,
+ "step": 3205
+ },
+ {
+ "epoch": 44.531004366812226,
+ "grad_norm": 0.03691193461418152,
+ "learning_rate": 0.0006,
+ "loss": 6.692244529724121,
+ "step": 3206
+ },
+ {
+ "epoch": 44.544978165938865,
+ "grad_norm": 0.02814812771975994,
+ "learning_rate": 0.0006,
+ "loss": 6.61978006362915,
+ "step": 3207
+ },
+ {
+ "epoch": 44.5589519650655,
+ "grad_norm": 0.022155074402689934,
+ "learning_rate": 0.0006,
+ "loss": 6.625905990600586,
+ "step": 3208
+ },
+ {
+ "epoch": 44.57292576419214,
+ "grad_norm": 0.030794885009527206,
+ "learning_rate": 0.0006,
+ "loss": 6.678500652313232,
+ "step": 3209
+ },
+ {
+ "epoch": 44.58689956331878,
+ "grad_norm": 0.03201768547296524,
+ "learning_rate": 0.0006,
+ "loss": 6.6015729904174805,
+ "step": 3210
+ },
+ {
+ "epoch": 44.60087336244541,
+ "grad_norm": 0.025138361379504204,
+ "learning_rate": 0.0006,
+ "loss": 6.568234443664551,
+ "step": 3211
+ },
+ {
+ "epoch": 44.61484716157205,
+ "grad_norm": 0.02969353087246418,
+ "learning_rate": 0.0006,
+ "loss": 6.588366508483887,
+ "step": 3212
+ },
+ {
+ "epoch": 44.62882096069869,
+ "grad_norm": 0.03136962279677391,
+ "learning_rate": 0.0006,
+ "loss": 6.479961395263672,
+ "step": 3213
+ },
+ {
+ "epoch": 44.64279475982533,
+ "grad_norm": 0.030178170651197433,
+ "learning_rate": 0.0006,
+ "loss": 6.622910976409912,
+ "step": 3214
+ },
+ {
+ "epoch": 44.65676855895197,
+ "grad_norm": 0.02682245895266533,
+ "learning_rate": 0.0006,
+ "loss": 6.597707748413086,
+ "step": 3215
+ },
+ {
+ "epoch": 44.670742358078606,
+ "grad_norm": 0.025154557079076767,
+ "learning_rate": 0.0006,
+ "loss": 6.630541801452637,
+ "step": 3216
+ },
+ {
+ "epoch": 44.68471615720524,
+ "grad_norm": 0.03564463183283806,
+ "learning_rate": 0.0006,
+ "loss": 6.542823791503906,
+ "step": 3217
+ },
+ {
+ "epoch": 44.698689956331876,
+ "grad_norm": 0.049477651715278625,
+ "learning_rate": 0.0006,
+ "loss": 6.646030426025391,
+ "step": 3218
+ },
+ {
+ "epoch": 44.712663755458514,
+ "grad_norm": 0.06536535918712616,
+ "learning_rate": 0.0006,
+ "loss": 6.595463275909424,
+ "step": 3219
+ },
+ {
+ "epoch": 44.72663755458515,
+ "grad_norm": 0.07710316777229309,
+ "learning_rate": 0.0006,
+ "loss": 6.55524206161499,
+ "step": 3220
+ },
+ {
+ "epoch": 44.74061135371179,
+ "grad_norm": 0.09571163356304169,
+ "learning_rate": 0.0006,
+ "loss": 6.659764289855957,
+ "step": 3221
+ },
+ {
+ "epoch": 44.75458515283843,
+ "grad_norm": 0.055076830089092255,
+ "learning_rate": 0.0006,
+ "loss": 6.658882141113281,
+ "step": 3222
+ },
+ {
+ "epoch": 44.76855895196506,
+ "grad_norm": 0.07671099156141281,
+ "learning_rate": 0.0006,
+ "loss": 6.605809211730957,
+ "step": 3223
+ },
+ {
+ "epoch": 44.7825327510917,
+ "grad_norm": 0.05295584350824356,
+ "learning_rate": 0.0006,
+ "loss": 6.641455173492432,
+ "step": 3224
+ },
+ {
+ "epoch": 44.79650655021834,
+ "grad_norm": 0.09128042310476303,
+ "learning_rate": 0.0006,
+ "loss": 6.691066741943359,
+ "step": 3225
+ },
+ {
+ "epoch": 44.81048034934498,
+ "grad_norm": 0.05362958461046219,
+ "learning_rate": 0.0006,
+ "loss": 6.573406219482422,
+ "step": 3226
+ },
+ {
+ "epoch": 44.82445414847162,
+ "grad_norm": 0.046910062432289124,
+ "learning_rate": 0.0006,
+ "loss": 6.550876617431641,
+ "step": 3227
+ },
+ {
+ "epoch": 44.838427947598255,
+ "grad_norm": 0.03494197130203247,
+ "learning_rate": 0.0006,
+ "loss": 6.540179252624512,
+ "step": 3228
+ },
+ {
+ "epoch": 44.852401746724894,
+ "grad_norm": 0.04716387391090393,
+ "learning_rate": 0.0006,
+ "loss": 6.490459442138672,
+ "step": 3229
+ },
+ {
+ "epoch": 44.866375545851525,
+ "grad_norm": 0.04190506786108017,
+ "learning_rate": 0.0006,
+ "loss": 6.509925365447998,
+ "step": 3230
+ },
+ {
+ "epoch": 44.880349344978164,
+ "grad_norm": 0.04749064892530441,
+ "learning_rate": 0.0006,
+ "loss": 6.478779315948486,
+ "step": 3231
+ },
+ {
+ "epoch": 44.8943231441048,
+ "grad_norm": 0.028002185747027397,
+ "learning_rate": 0.0006,
+ "loss": 6.528904914855957,
+ "step": 3232
+ },
+ {
+ "epoch": 44.90829694323144,
+ "grad_norm": 0.036473166197538376,
+ "learning_rate": 0.0006,
+ "loss": 6.549042701721191,
+ "step": 3233
+ },
+ {
+ "epoch": 44.92227074235808,
+ "grad_norm": 0.02543315477669239,
+ "learning_rate": 0.0006,
+ "loss": 6.498875141143799,
+ "step": 3234
+ },
+ {
+ "epoch": 44.93624454148472,
+ "grad_norm": 0.029906807467341423,
+ "learning_rate": 0.0006,
+ "loss": 6.453019142150879,
+ "step": 3235
+ },
+ {
+ "epoch": 44.95021834061135,
+ "grad_norm": 0.027231857180595398,
+ "learning_rate": 0.0006,
+ "loss": 6.428829193115234,
+ "step": 3236
+ },
+ {
+ "epoch": 44.96419213973799,
+ "grad_norm": 0.02506154775619507,
+ "learning_rate": 0.0006,
+ "loss": 6.420858383178711,
+ "step": 3237
+ },
+ {
+ "epoch": 44.97816593886463,
+ "grad_norm": 0.0269472673535347,
+ "learning_rate": 0.0006,
+ "loss": 6.484989643096924,
+ "step": 3238
+ },
+ {
+ "epoch": 44.992139737991266,
+ "grad_norm": 0.023410487920045853,
+ "learning_rate": 0.0006,
+ "loss": 6.454830169677734,
+ "step": 3239
+ },
+ {
+ "epoch": 45.0,
+ "grad_norm": 0.034175023436546326,
+ "learning_rate": 0.0006,
+ "loss": 6.430975437164307,
+ "step": 3240
+ },
+ {
+ "epoch": 45.0,
+ "eval_loss": 6.458896160125732,
+ "eval_runtime": 56.6934,
+ "eval_samples_per_second": 43.074,
+ "eval_steps_per_second": 1.358,
+ "step": 3240
+ },
+ {
+ "epoch": 45.01397379912664,
+ "grad_norm": 0.045031629502773285,
+ "learning_rate": 0.0006,
+ "loss": 6.4243011474609375,
+ "step": 3241
+ },
+ {
+ "epoch": 45.02794759825328,
+ "grad_norm": 0.04968734085559845,
+ "learning_rate": 0.0006,
+ "loss": 6.380091667175293,
+ "step": 3242
+ },
+ {
+ "epoch": 45.041921397379916,
+ "grad_norm": 0.047145187854766846,
+ "learning_rate": 0.0006,
+ "loss": 6.477178573608398,
+ "step": 3243
+ },
+ {
+ "epoch": 45.05589519650655,
+ "grad_norm": 0.04304948449134827,
+ "learning_rate": 0.0006,
+ "loss": 6.399726867675781,
+ "step": 3244
+ },
+ {
+ "epoch": 45.069868995633186,
+ "grad_norm": 0.04466010630130768,
+ "learning_rate": 0.0006,
+ "loss": 6.3415679931640625,
+ "step": 3245
+ },
+ {
+ "epoch": 45.083842794759825,
+ "grad_norm": 0.04305306077003479,
+ "learning_rate": 0.0006,
+ "loss": 6.3413310050964355,
+ "step": 3246
+ },
+ {
+ "epoch": 45.09781659388646,
+ "grad_norm": 0.043064575642347336,
+ "learning_rate": 0.0006,
+ "loss": 6.4838714599609375,
+ "step": 3247
+ },
+ {
+ "epoch": 45.1117903930131,
+ "grad_norm": 0.049503277987241745,
+ "learning_rate": 0.0006,
+ "loss": 6.38473653793335,
+ "step": 3248
+ },
+ {
+ "epoch": 45.12576419213974,
+ "grad_norm": 0.04613238573074341,
+ "learning_rate": 0.0006,
+ "loss": 6.370769500732422,
+ "step": 3249
+ },
+ {
+ "epoch": 45.13973799126637,
+ "grad_norm": 0.052119478583335876,
+ "learning_rate": 0.0006,
+ "loss": 6.362321853637695,
+ "step": 3250
+ },
+ {
+ "epoch": 45.15371179039301,
+ "grad_norm": 0.045110076665878296,
+ "learning_rate": 0.0006,
+ "loss": 6.369844913482666,
+ "step": 3251
+ },
+ {
+ "epoch": 45.16768558951965,
+ "grad_norm": 0.03798583149909973,
+ "learning_rate": 0.0006,
+ "loss": 6.324633598327637,
+ "step": 3252
+ },
+ {
+ "epoch": 45.18165938864629,
+ "grad_norm": 0.03744659945368767,
+ "learning_rate": 0.0006,
+ "loss": 6.349421501159668,
+ "step": 3253
+ },
+ {
+ "epoch": 45.19563318777293,
+ "grad_norm": 0.03767292574048042,
+ "learning_rate": 0.0006,
+ "loss": 6.334238052368164,
+ "step": 3254
+ },
+ {
+ "epoch": 45.209606986899566,
+ "grad_norm": 0.03577509894967079,
+ "learning_rate": 0.0006,
+ "loss": 6.3453826904296875,
+ "step": 3255
+ },
+ {
+ "epoch": 45.223580786026204,
+ "grad_norm": 0.03403240442276001,
+ "learning_rate": 0.0006,
+ "loss": 6.2836174964904785,
+ "step": 3256
+ },
+ {
+ "epoch": 45.237554585152836,
+ "grad_norm": 0.03517331928014755,
+ "learning_rate": 0.0006,
+ "loss": 6.281718730926514,
+ "step": 3257
+ },
+ {
+ "epoch": 45.251528384279474,
+ "grad_norm": 0.04409139230847359,
+ "learning_rate": 0.0006,
+ "loss": 6.320815086364746,
+ "step": 3258
+ },
+ {
+ "epoch": 45.26550218340611,
+ "grad_norm": 0.054750364273786545,
+ "learning_rate": 0.0006,
+ "loss": 6.343618392944336,
+ "step": 3259
+ },
+ {
+ "epoch": 45.27947598253275,
+ "grad_norm": 0.07612067461013794,
+ "learning_rate": 0.0006,
+ "loss": 6.310614585876465,
+ "step": 3260
+ },
+ {
+ "epoch": 45.29344978165939,
+ "grad_norm": 0.08438985049724579,
+ "learning_rate": 0.0006,
+ "loss": 6.2369184494018555,
+ "step": 3261
+ },
+ {
+ "epoch": 45.30742358078603,
+ "grad_norm": 0.0782201811671257,
+ "learning_rate": 0.0006,
+ "loss": 6.254452705383301,
+ "step": 3262
+ },
+ {
+ "epoch": 45.32139737991266,
+ "grad_norm": 0.09928355365991592,
+ "learning_rate": 0.0006,
+ "loss": 6.3112077713012695,
+ "step": 3263
+ },
+ {
+ "epoch": 45.3353711790393,
+ "grad_norm": 0.07233133912086487,
+ "learning_rate": 0.0006,
+ "loss": 6.408087253570557,
+ "step": 3264
+ },
+ {
+ "epoch": 45.34934497816594,
+ "grad_norm": 0.07478684186935425,
+ "learning_rate": 0.0006,
+ "loss": 6.3770036697387695,
+ "step": 3265
+ },
+ {
+ "epoch": 45.36331877729258,
+ "grad_norm": 0.06036846712231636,
+ "learning_rate": 0.0006,
+ "loss": 6.33865213394165,
+ "step": 3266
+ },
+ {
+ "epoch": 45.377292576419215,
+ "grad_norm": 0.05783458054065704,
+ "learning_rate": 0.0006,
+ "loss": 6.363203048706055,
+ "step": 3267
+ },
+ {
+ "epoch": 45.391266375545854,
+ "grad_norm": 0.05443909019231796,
+ "learning_rate": 0.0006,
+ "loss": 6.287069797515869,
+ "step": 3268
+ },
+ {
+ "epoch": 45.40524017467249,
+ "grad_norm": 0.0557909794151783,
+ "learning_rate": 0.0006,
+ "loss": 6.297784805297852,
+ "step": 3269
+ },
+ {
+ "epoch": 45.419213973799124,
+ "grad_norm": 0.03483372926712036,
+ "learning_rate": 0.0006,
+ "loss": 6.280734062194824,
+ "step": 3270
+ },
+ {
+ "epoch": 45.43318777292576,
+ "grad_norm": 0.0425543375313282,
+ "learning_rate": 0.0006,
+ "loss": 6.33565616607666,
+ "step": 3271
+ },
+ {
+ "epoch": 45.4471615720524,
+ "grad_norm": 0.03230159357190132,
+ "learning_rate": 0.0006,
+ "loss": 6.197698593139648,
+ "step": 3272
+ },
+ {
+ "epoch": 45.46113537117904,
+ "grad_norm": 0.036897290498018265,
+ "learning_rate": 0.0006,
+ "loss": 6.241239547729492,
+ "step": 3273
+ },
+ {
+ "epoch": 45.47510917030568,
+ "grad_norm": 0.031103838235139847,
+ "learning_rate": 0.0006,
+ "loss": 6.159435749053955,
+ "step": 3274
+ },
+ {
+ "epoch": 45.48908296943232,
+ "grad_norm": 0.027116449549794197,
+ "learning_rate": 0.0006,
+ "loss": 6.212225914001465,
+ "step": 3275
+ },
+ {
+ "epoch": 45.50305676855895,
+ "grad_norm": 0.029923981055617332,
+ "learning_rate": 0.0006,
+ "loss": 6.157269477844238,
+ "step": 3276
+ },
+ {
+ "epoch": 45.51703056768559,
+ "grad_norm": 0.02521538734436035,
+ "learning_rate": 0.0006,
+ "loss": 6.146918773651123,
+ "step": 3277
+ },
+ {
+ "epoch": 45.531004366812226,
+ "grad_norm": 0.02758920192718506,
+ "learning_rate": 0.0006,
+ "loss": 6.0446977615356445,
+ "step": 3278
+ },
+ {
+ "epoch": 45.544978165938865,
+ "grad_norm": 0.024328220635652542,
+ "learning_rate": 0.0006,
+ "loss": 6.099508285522461,
+ "step": 3279
+ },
+ {
+ "epoch": 45.5589519650655,
+ "grad_norm": 0.027425771579146385,
+ "learning_rate": 0.0006,
+ "loss": 6.084299087524414,
+ "step": 3280
+ },
+ {
+ "epoch": 45.57292576419214,
+ "grad_norm": 0.026927651837468147,
+ "learning_rate": 0.0006,
+ "loss": 6.146334171295166,
+ "step": 3281
+ },
+ {
+ "epoch": 45.58689956331878,
+ "grad_norm": 0.04054379090666771,
+ "learning_rate": 0.0006,
+ "loss": 6.1301493644714355,
+ "step": 3282
+ },
+ {
+ "epoch": 45.60087336244541,
+ "grad_norm": 0.04456290975213051,
+ "learning_rate": 0.0006,
+ "loss": 6.152511119842529,
+ "step": 3283
+ },
+ {
+ "epoch": 45.61484716157205,
+ "grad_norm": 0.052821505814790726,
+ "learning_rate": 0.0006,
+ "loss": 6.113544940948486,
+ "step": 3284
+ },
+ {
+ "epoch": 45.62882096069869,
+ "grad_norm": 0.07251739501953125,
+ "learning_rate": 0.0006,
+ "loss": 6.118298530578613,
+ "step": 3285
+ },
+ {
+ "epoch": 45.64279475982533,
+ "grad_norm": 0.09433556348085403,
+ "learning_rate": 0.0006,
+ "loss": 6.170897483825684,
+ "step": 3286
+ },
+ {
+ "epoch": 45.65676855895197,
+ "grad_norm": 0.057683493942022324,
+ "learning_rate": 0.0006,
+ "loss": 6.1747822761535645,
+ "step": 3287
+ },
+ {
+ "epoch": 45.670742358078606,
+ "grad_norm": 0.06363270431756973,
+ "learning_rate": 0.0006,
+ "loss": 6.1617889404296875,
+ "step": 3288
+ },
+ {
+ "epoch": 45.68471615720524,
+ "grad_norm": 0.07528169453144073,
+ "learning_rate": 0.0006,
+ "loss": 6.230060577392578,
+ "step": 3289
+ },
+ {
+ "epoch": 45.698689956331876,
+ "grad_norm": 0.07816170156002045,
+ "learning_rate": 0.0006,
+ "loss": 6.171257019042969,
+ "step": 3290
+ },
+ {
+ "epoch": 45.712663755458514,
+ "grad_norm": 0.05860164389014244,
+ "learning_rate": 0.0006,
+ "loss": 6.095484256744385,
+ "step": 3291
+ },
+ {
+ "epoch": 45.72663755458515,
+ "grad_norm": 0.04747429117560387,
+ "learning_rate": 0.0006,
+ "loss": 6.060067176818848,
+ "step": 3292
+ },
+ {
+ "epoch": 45.74061135371179,
+ "grad_norm": 0.04514474794268608,
+ "learning_rate": 0.0006,
+ "loss": 6.100218772888184,
+ "step": 3293
+ },
+ {
+ "epoch": 45.75458515283843,
+ "grad_norm": 0.05168681964278221,
+ "learning_rate": 0.0006,
+ "loss": 6.0621867179870605,
+ "step": 3294
+ },
+ {
+ "epoch": 45.76855895196506,
+ "grad_norm": 0.04665876924991608,
+ "learning_rate": 0.0006,
+ "loss": 6.118292808532715,
+ "step": 3295
+ },
+ {
+ "epoch": 45.7825327510917,
+ "grad_norm": 0.049374762922525406,
+ "learning_rate": 0.0006,
+ "loss": 6.098394393920898,
+ "step": 3296
+ },
+ {
+ "epoch": 45.79650655021834,
+ "grad_norm": 0.03793549910187721,
+ "learning_rate": 0.0006,
+ "loss": 6.069541931152344,
+ "step": 3297
+ },
+ {
+ "epoch": 45.81048034934498,
+ "grad_norm": 0.036362383514642715,
+ "learning_rate": 0.0006,
+ "loss": 6.1188435554504395,
+ "step": 3298
+ },
+ {
+ "epoch": 45.82445414847162,
+ "grad_norm": 0.03397097811102867,
+ "learning_rate": 0.0006,
+ "loss": 6.117589950561523,
+ "step": 3299
+ },
+ {
+ "epoch": 45.838427947598255,
+ "grad_norm": 0.030977753922343254,
+ "learning_rate": 0.0006,
+ "loss": 6.0236711502075195,
+ "step": 3300
+ },
+ {
+ "epoch": 45.852401746724894,
+ "grad_norm": 0.02800217643380165,
+ "learning_rate": 0.0006,
+ "loss": 6.034825325012207,
+ "step": 3301
+ },
+ {
+ "epoch": 45.866375545851525,
+ "grad_norm": 0.03129778802394867,
+ "learning_rate": 0.0006,
+ "loss": 5.980271339416504,
+ "step": 3302
+ },
+ {
+ "epoch": 45.880349344978164,
+ "grad_norm": 0.025886094197630882,
+ "learning_rate": 0.0006,
+ "loss": 5.966419219970703,
+ "step": 3303
+ },
+ {
+ "epoch": 45.8943231441048,
+ "grad_norm": 0.025712328031659126,
+ "learning_rate": 0.0006,
+ "loss": 5.938652515411377,
+ "step": 3304
+ },
+ {
+ "epoch": 45.90829694323144,
+ "grad_norm": 0.022508298978209496,
+ "learning_rate": 0.0006,
+ "loss": 5.982312202453613,
+ "step": 3305
+ },
+ {
+ "epoch": 45.92227074235808,
+ "grad_norm": 0.024880951270461082,
+ "learning_rate": 0.0006,
+ "loss": 5.931150436401367,
+ "step": 3306
+ },
+ {
+ "epoch": 45.93624454148472,
+ "grad_norm": 0.030007576569914818,
+ "learning_rate": 0.0006,
+ "loss": 5.936639308929443,
+ "step": 3307
+ },
+ {
+ "epoch": 45.95021834061135,
+ "grad_norm": 0.033314041793346405,
+ "learning_rate": 0.0006,
+ "loss": 5.944248676300049,
+ "step": 3308
+ },
+ {
+ "epoch": 45.96419213973799,
+ "grad_norm": 0.043078381568193436,
+ "learning_rate": 0.0006,
+ "loss": 5.971286773681641,
+ "step": 3309
+ },
+ {
+ "epoch": 45.97816593886463,
+ "grad_norm": 0.059296440333127975,
+ "learning_rate": 0.0006,
+ "loss": 5.910497665405273,
+ "step": 3310
+ },
+ {
+ "epoch": 45.992139737991266,
+ "grad_norm": 0.06232666224241257,
+ "learning_rate": 0.0006,
+ "loss": 5.929817199707031,
+ "step": 3311
+ },
+ {
+ "epoch": 46.0,
+ "grad_norm": 0.05303940549492836,
+ "learning_rate": 0.0006,
+ "loss": 5.910822868347168,
+ "step": 3312
+ },
+ {
+ "epoch": 46.0,
+ "eval_loss": 5.977851390838623,
+ "eval_runtime": 56.2435,
+ "eval_samples_per_second": 43.418,
+ "eval_steps_per_second": 1.369,
+ "step": 3312
+ },
+ {
+ "epoch": 46.01397379912664,
+ "grad_norm": 0.06943798810243607,
+ "learning_rate": 0.0006,
+ "loss": 5.854922294616699,
+ "step": 3313
+ },
+ {
+ "epoch": 46.02794759825328,
+ "grad_norm": 0.05236278846859932,
+ "learning_rate": 0.0006,
+ "loss": 5.876611232757568,
+ "step": 3314
+ },
+ {
+ "epoch": 46.041921397379916,
+ "grad_norm": 0.06240001320838928,
+ "learning_rate": 0.0006,
+ "loss": 5.878912448883057,
+ "step": 3315
+ },
+ {
+ "epoch": 46.05589519650655,
+ "grad_norm": 0.07401192933320999,
+ "learning_rate": 0.0006,
+ "loss": 5.985136032104492,
+ "step": 3316
+ },
+ {
+ "epoch": 46.069868995633186,
+ "grad_norm": 0.0826529711484909,
+ "learning_rate": 0.0006,
+ "loss": 5.939439296722412,
+ "step": 3317
+ },
+ {
+ "epoch": 46.083842794759825,
+ "grad_norm": 0.07999005168676376,
+ "learning_rate": 0.0006,
+ "loss": 5.982336044311523,
+ "step": 3318
+ },
+ {
+ "epoch": 46.09781659388646,
+ "grad_norm": 0.07014869153499603,
+ "learning_rate": 0.0006,
+ "loss": 5.896534442901611,
+ "step": 3319
+ },
+ {
+ "epoch": 46.1117903930131,
+ "grad_norm": 0.05795128643512726,
+ "learning_rate": 0.0006,
+ "loss": 5.843111991882324,
+ "step": 3320
+ },
+ {
+ "epoch": 46.12576419213974,
+ "grad_norm": 0.06374258548021317,
+ "learning_rate": 0.0006,
+ "loss": 5.78752326965332,
+ "step": 3321
+ },
+ {
+ "epoch": 46.13973799126637,
+ "grad_norm": 0.045524805784225464,
+ "learning_rate": 0.0006,
+ "loss": 5.86755895614624,
+ "step": 3322
+ },
+ {
+ "epoch": 46.15371179039301,
+ "grad_norm": 0.03773504123091698,
+ "learning_rate": 0.0006,
+ "loss": 5.83627462387085,
+ "step": 3323
+ },
+ {
+ "epoch": 46.16768558951965,
+ "grad_norm": 0.035815201699733734,
+ "learning_rate": 0.0006,
+ "loss": 5.798238277435303,
+ "step": 3324
+ },
+ {
+ "epoch": 46.18165938864629,
+ "grad_norm": 0.03756783902645111,
+ "learning_rate": 0.0006,
+ "loss": 5.845197677612305,
+ "step": 3325
+ },
+ {
+ "epoch": 46.19563318777293,
+ "grad_norm": 0.03569113463163376,
+ "learning_rate": 0.0006,
+ "loss": 5.888932704925537,
+ "step": 3326
+ },
+ {
+ "epoch": 46.209606986899566,
+ "grad_norm": 0.029467813670635223,
+ "learning_rate": 0.0006,
+ "loss": 5.792047500610352,
+ "step": 3327
+ },
+ {
+ "epoch": 46.223580786026204,
+ "grad_norm": 0.025698406621813774,
+ "learning_rate": 0.0006,
+ "loss": 5.783734321594238,
+ "step": 3328
+ },
+ {
+ "epoch": 46.237554585152836,
+ "grad_norm": 0.028426671400666237,
+ "learning_rate": 0.0006,
+ "loss": 5.762084007263184,
+ "step": 3329
+ },
+ {
+ "epoch": 46.251528384279474,
+ "grad_norm": 0.03099622018635273,
+ "learning_rate": 0.0006,
+ "loss": 5.719109058380127,
+ "step": 3330
+ },
+ {
+ "epoch": 46.26550218340611,
+ "grad_norm": 0.0310219656676054,
+ "learning_rate": 0.0006,
+ "loss": 5.737057209014893,
+ "step": 3331
+ },
+ {
+ "epoch": 46.27947598253275,
+ "grad_norm": 0.02713446132838726,
+ "learning_rate": 0.0006,
+ "loss": 5.758657455444336,
+ "step": 3332
+ },
+ {
+ "epoch": 46.29344978165939,
+ "grad_norm": 0.03099844604730606,
+ "learning_rate": 0.0006,
+ "loss": 5.742086410522461,
+ "step": 3333
+ },
+ {
+ "epoch": 46.30742358078603,
+ "grad_norm": 0.030258730053901672,
+ "learning_rate": 0.0006,
+ "loss": 5.767426490783691,
+ "step": 3334
+ },
+ {
+ "epoch": 46.32139737991266,
+ "grad_norm": 0.029034852981567383,
+ "learning_rate": 0.0006,
+ "loss": 5.651792049407959,
+ "step": 3335
+ },
+ {
+ "epoch": 46.3353711790393,
+ "grad_norm": 0.03138665854930878,
+ "learning_rate": 0.0006,
+ "loss": 5.664980888366699,
+ "step": 3336
+ },
+ {
+ "epoch": 46.34934497816594,
+ "grad_norm": 0.040546901524066925,
+ "learning_rate": 0.0006,
+ "loss": 5.665165901184082,
+ "step": 3337
+ },
+ {
+ "epoch": 46.36331877729258,
+ "grad_norm": 0.06395485252141953,
+ "learning_rate": 0.0006,
+ "loss": 5.7776713371276855,
+ "step": 3338
+ },
+ {
+ "epoch": 46.377292576419215,
+ "grad_norm": 0.10088056325912476,
+ "learning_rate": 0.0006,
+ "loss": 5.755645751953125,
+ "step": 3339
+ },
+ {
+ "epoch": 46.391266375545854,
+ "grad_norm": 0.09527938067913055,
+ "learning_rate": 0.0006,
+ "loss": 5.838809490203857,
+ "step": 3340
+ },
+ {
+ "epoch": 46.40524017467249,
+ "grad_norm": 0.0849648267030716,
+ "learning_rate": 0.0006,
+ "loss": 5.781932830810547,
+ "step": 3341
+ },
+ {
+ "epoch": 46.419213973799124,
+ "grad_norm": 0.10142666101455688,
+ "learning_rate": 0.0006,
+ "loss": 5.838970184326172,
+ "step": 3342
+ },
+ {
+ "epoch": 46.43318777292576,
+ "grad_norm": 0.42731454968452454,
+ "learning_rate": 0.0006,
+ "loss": 6.071902275085449,
+ "step": 3343
+ },
+ {
+ "epoch": 46.4471615720524,
+ "grad_norm": 0.13102488219738007,
+ "learning_rate": 0.0006,
+ "loss": 6.8635993003845215,
+ "step": 3344
+ },
+ {
+ "epoch": 46.46113537117904,
+ "grad_norm": 0.23691938817501068,
+ "learning_rate": 0.0006,
+ "loss": 7.313158988952637,
+ "step": 3345
+ },
+ {
+ "epoch": 46.47510917030568,
+ "grad_norm": 0.11670669168233871,
+ "learning_rate": 0.0006,
+ "loss": 7.240363121032715,
+ "step": 3346
+ },
+ {
+ "epoch": 46.48908296943232,
+ "grad_norm": 0.09985820949077606,
+ "learning_rate": 0.0006,
+ "loss": 7.113332271575928,
+ "step": 3347
+ },
+ {
+ "epoch": 46.50305676855895,
+ "grad_norm": 0.11406408250331879,
+ "learning_rate": 0.0006,
+ "loss": 7.051285743713379,
+ "step": 3348
+ },
+ {
+ "epoch": 46.51703056768559,
+ "grad_norm": 0.07900336384773254,
+ "learning_rate": 0.0006,
+ "loss": 7.014365196228027,
+ "step": 3349
+ },
+ {
+ "epoch": 46.531004366812226,
+ "grad_norm": 0.05380333960056305,
+ "learning_rate": 0.0006,
+ "loss": 6.8868818283081055,
+ "step": 3350
+ },
+ {
+ "epoch": 46.544978165938865,
+ "grad_norm": 0.05868449807167053,
+ "learning_rate": 0.0006,
+ "loss": 6.783599853515625,
+ "step": 3351
+ },
+ {
+ "epoch": 46.5589519650655,
+ "grad_norm": 0.04428785294294357,
+ "learning_rate": 0.0006,
+ "loss": 6.798511505126953,
+ "step": 3352
+ },
+ {
+ "epoch": 46.57292576419214,
+ "grad_norm": 0.055527277290821075,
+ "learning_rate": 0.0006,
+ "loss": 6.694507598876953,
+ "step": 3353
+ },
+ {
+ "epoch": 46.58689956331878,
+ "grad_norm": 0.03872432932257652,
+ "learning_rate": 0.0006,
+ "loss": 6.736631393432617,
+ "step": 3354
+ },
+ {
+ "epoch": 46.60087336244541,
+ "grad_norm": 0.04330779239535332,
+ "learning_rate": 0.0006,
+ "loss": 6.576806545257568,
+ "step": 3355
+ },
+ {
+ "epoch": 46.61484716157205,
+ "grad_norm": 0.036767736077308655,
+ "learning_rate": 0.0006,
+ "loss": 6.559871196746826,
+ "step": 3356
+ },
+ {
+ "epoch": 46.62882096069869,
+ "grad_norm": 0.039857108145952225,
+ "learning_rate": 0.0006,
+ "loss": 6.5749831199646,
+ "step": 3357
+ },
+ {
+ "epoch": 46.64279475982533,
+ "grad_norm": 0.048719972372055054,
+ "learning_rate": 0.0006,
+ "loss": 6.4915618896484375,
+ "step": 3358
+ },
+ {
+ "epoch": 46.65676855895197,
+ "grad_norm": 0.042849015444517136,
+ "learning_rate": 0.0006,
+ "loss": 6.413516044616699,
+ "step": 3359
+ },
+ {
+ "epoch": 46.670742358078606,
+ "grad_norm": 0.02871009334921837,
+ "learning_rate": 0.0006,
+ "loss": 6.320952415466309,
+ "step": 3360
+ },
+ {
+ "epoch": 46.68471615720524,
+ "grad_norm": 0.03193620592355728,
+ "learning_rate": 0.0006,
+ "loss": 6.4176435470581055,
+ "step": 3361
+ },
+ {
+ "epoch": 46.698689956331876,
+ "grad_norm": 0.027417024597525597,
+ "learning_rate": 0.0006,
+ "loss": 6.342319488525391,
+ "step": 3362
+ },
+ {
+ "epoch": 46.712663755458514,
+ "grad_norm": 0.03305937349796295,
+ "learning_rate": 0.0006,
+ "loss": 6.246112823486328,
+ "step": 3363
+ },
+ {
+ "epoch": 46.72663755458515,
+ "grad_norm": 0.031403183937072754,
+ "learning_rate": 0.0006,
+ "loss": 6.149503231048584,
+ "step": 3364
+ },
+ {
+ "epoch": 46.74061135371179,
+ "grad_norm": 0.031563788652420044,
+ "learning_rate": 0.0006,
+ "loss": 6.278808116912842,
+ "step": 3365
+ },
+ {
+ "epoch": 46.75458515283843,
+ "grad_norm": 0.028366589918732643,
+ "learning_rate": 0.0006,
+ "loss": 6.057036876678467,
+ "step": 3366
+ },
+ {
+ "epoch": 46.76855895196506,
+ "grad_norm": 0.023715803399682045,
+ "learning_rate": 0.0006,
+ "loss": 6.111324310302734,
+ "step": 3367
+ },
+ {
+ "epoch": 46.7825327510917,
+ "grad_norm": 0.026546811684966087,
+ "learning_rate": 0.0006,
+ "loss": 6.105669975280762,
+ "step": 3368
+ },
+ {
+ "epoch": 46.79650655021834,
+ "grad_norm": 0.023531019687652588,
+ "learning_rate": 0.0006,
+ "loss": 6.068072319030762,
+ "step": 3369
+ },
+ {
+ "epoch": 46.81048034934498,
+ "grad_norm": 0.024480128660798073,
+ "learning_rate": 0.0006,
+ "loss": 5.947896957397461,
+ "step": 3370
+ },
+ {
+ "epoch": 46.82445414847162,
+ "grad_norm": 0.02517310529947281,
+ "learning_rate": 0.0006,
+ "loss": 6.030037879943848,
+ "step": 3371
+ },
+ {
+ "epoch": 46.838427947598255,
+ "grad_norm": 0.026723403483629227,
+ "learning_rate": 0.0006,
+ "loss": 5.963776111602783,
+ "step": 3372
+ },
+ {
+ "epoch": 46.852401746724894,
+ "grad_norm": 0.03163504600524902,
+ "learning_rate": 0.0006,
+ "loss": 5.893466949462891,
+ "step": 3373
+ },
+ {
+ "epoch": 46.866375545851525,
+ "grad_norm": 0.04058683291077614,
+ "learning_rate": 0.0006,
+ "loss": 5.9202799797058105,
+ "step": 3374
+ },
+ {
+ "epoch": 46.880349344978164,
+ "grad_norm": 0.04964117333292961,
+ "learning_rate": 0.0006,
+ "loss": 5.9366559982299805,
+ "step": 3375
+ },
+ {
+ "epoch": 46.8943231441048,
+ "grad_norm": 0.055406033992767334,
+ "learning_rate": 0.0006,
+ "loss": 5.912230968475342,
+ "step": 3376
+ },
+ {
+ "epoch": 46.90829694323144,
+ "grad_norm": 0.04643625393509865,
+ "learning_rate": 0.0006,
+ "loss": 5.87260103225708,
+ "step": 3377
+ },
+ {
+ "epoch": 46.92227074235808,
+ "grad_norm": 0.029325876384973526,
+ "learning_rate": 0.0006,
+ "loss": 5.854763984680176,
+ "step": 3378
+ },
+ {
+ "epoch": 46.93624454148472,
+ "grad_norm": 0.03159382566809654,
+ "learning_rate": 0.0006,
+ "loss": 5.880882263183594,
+ "step": 3379
+ },
+ {
+ "epoch": 46.95021834061135,
+ "grad_norm": 0.030224012210965157,
+ "learning_rate": 0.0006,
+ "loss": 5.787360191345215,
+ "step": 3380
+ },
+ {
+ "epoch": 46.96419213973799,
+ "grad_norm": 0.028368130326271057,
+ "learning_rate": 0.0006,
+ "loss": 5.817195892333984,
+ "step": 3381
+ },
+ {
+ "epoch": 46.97816593886463,
+ "grad_norm": 0.029411399737000465,
+ "learning_rate": 0.0006,
+ "loss": 5.783403396606445,
+ "step": 3382
+ },
+ {
+ "epoch": 46.992139737991266,
+ "grad_norm": 0.029927462339401245,
+ "learning_rate": 0.0006,
+ "loss": 5.662444591522217,
+ "step": 3383
+ },
+ {
+ "epoch": 47.0,
+ "grad_norm": 0.02872961387038231,
+ "learning_rate": 0.0006,
+ "loss": 5.742781639099121,
+ "step": 3384
+ },
+ {
+ "epoch": 47.0,
+ "eval_loss": 5.870260238647461,
+ "eval_runtime": 57.3734,
+ "eval_samples_per_second": 42.563,
+ "eval_steps_per_second": 1.342,
+ "step": 3384
+ },
+ {
+ "epoch": 47.01397379912664,
+ "grad_norm": 0.028542131185531616,
+ "learning_rate": 0.0006,
+ "loss": 5.753256797790527,
+ "step": 3385
+ },
+ {
+ "epoch": 47.02794759825328,
+ "grad_norm": 0.03227119892835617,
+ "learning_rate": 0.0006,
+ "loss": 5.722246170043945,
+ "step": 3386
+ },
+ {
+ "epoch": 47.041921397379916,
+ "grad_norm": 0.031461723148822784,
+ "learning_rate": 0.0006,
+ "loss": 5.740520000457764,
+ "step": 3387
+ },
+ {
+ "epoch": 47.05589519650655,
+ "grad_norm": 0.03184220567345619,
+ "learning_rate": 0.0006,
+ "loss": 5.720890045166016,
+ "step": 3388
+ },
+ {
+ "epoch": 47.069868995633186,
+ "grad_norm": 0.030015580356121063,
+ "learning_rate": 0.0006,
+ "loss": 5.597814559936523,
+ "step": 3389
+ },
+ {
+ "epoch": 47.083842794759825,
+ "grad_norm": 0.029931779950857162,
+ "learning_rate": 0.0006,
+ "loss": 5.5412468910217285,
+ "step": 3390
+ },
+ {
+ "epoch": 47.09781659388646,
+ "grad_norm": 0.03206505626440048,
+ "learning_rate": 0.0006,
+ "loss": 5.6967597007751465,
+ "step": 3391
+ },
+ {
+ "epoch": 47.1117903930131,
+ "grad_norm": 0.03978161886334419,
+ "learning_rate": 0.0006,
+ "loss": 5.59661340713501,
+ "step": 3392
+ },
+ {
+ "epoch": 47.12576419213974,
+ "grad_norm": 0.04619080200791359,
+ "learning_rate": 0.0006,
+ "loss": 5.795788764953613,
+ "step": 3393
+ },
+ {
+ "epoch": 47.13973799126637,
+ "grad_norm": 0.049085382372140884,
+ "learning_rate": 0.0006,
+ "loss": 5.667335510253906,
+ "step": 3394
+ },
+ {
+ "epoch": 47.15371179039301,
+ "grad_norm": 0.04830858111381531,
+ "learning_rate": 0.0006,
+ "loss": 5.668806552886963,
+ "step": 3395
+ },
+ {
+ "epoch": 47.16768558951965,
+ "grad_norm": 0.0583178736269474,
+ "learning_rate": 0.0006,
+ "loss": 5.692285537719727,
+ "step": 3396
+ },
+ {
+ "epoch": 47.18165938864629,
+ "grad_norm": 0.05182253196835518,
+ "learning_rate": 0.0006,
+ "loss": 5.678645133972168,
+ "step": 3397
+ },
+ {
+ "epoch": 47.19563318777293,
+ "grad_norm": 0.04011905938386917,
+ "learning_rate": 0.0006,
+ "loss": 5.661550998687744,
+ "step": 3398
+ },
+ {
+ "epoch": 47.209606986899566,
+ "grad_norm": 0.03888486698269844,
+ "learning_rate": 0.0006,
+ "loss": 5.643001556396484,
+ "step": 3399
+ },
+ {
+ "epoch": 47.223580786026204,
+ "grad_norm": 0.030658265575766563,
+ "learning_rate": 0.0006,
+ "loss": 5.659090995788574,
+ "step": 3400
+ },
+ {
+ "epoch": 47.237554585152836,
+ "grad_norm": 0.03609256446361542,
+ "learning_rate": 0.0006,
+ "loss": 5.544501781463623,
+ "step": 3401
+ },
+ {
+ "epoch": 47.251528384279474,
+ "grad_norm": 0.033589769154787064,
+ "learning_rate": 0.0006,
+ "loss": 5.463878631591797,
+ "step": 3402
+ },
+ {
+ "epoch": 47.26550218340611,
+ "grad_norm": 0.035128287971019745,
+ "learning_rate": 0.0006,
+ "loss": 5.5680060386657715,
+ "step": 3403
+ },
+ {
+ "epoch": 47.27947598253275,
+ "grad_norm": 0.038684550672769547,
+ "learning_rate": 0.0006,
+ "loss": 5.587898254394531,
+ "step": 3404
+ },
+ {
+ "epoch": 47.29344978165939,
+ "grad_norm": 0.043566226959228516,
+ "learning_rate": 0.0006,
+ "loss": 5.580896854400635,
+ "step": 3405
+ },
+ {
+ "epoch": 47.30742358078603,
+ "grad_norm": 0.04428374394774437,
+ "learning_rate": 0.0006,
+ "loss": 5.603460311889648,
+ "step": 3406
+ },
+ {
+ "epoch": 47.32139737991266,
+ "grad_norm": 0.037368934601545334,
+ "learning_rate": 0.0006,
+ "loss": 5.564740180969238,
+ "step": 3407
+ },
+ {
+ "epoch": 47.3353711790393,
+ "grad_norm": 0.04092350974678993,
+ "learning_rate": 0.0006,
+ "loss": 5.584153652191162,
+ "step": 3408
+ },
+ {
+ "epoch": 47.34934497816594,
+ "grad_norm": 0.045031506568193436,
+ "learning_rate": 0.0006,
+ "loss": 5.601186275482178,
+ "step": 3409
+ },
+ {
+ "epoch": 47.36331877729258,
+ "grad_norm": 0.041830409318208694,
+ "learning_rate": 0.0006,
+ "loss": 5.512788772583008,
+ "step": 3410
+ },
+ {
+ "epoch": 47.377292576419215,
+ "grad_norm": 0.037633974105119705,
+ "learning_rate": 0.0006,
+ "loss": 5.503828525543213,
+ "step": 3411
+ },
+ {
+ "epoch": 47.391266375545854,
+ "grad_norm": 0.03473588079214096,
+ "learning_rate": 0.0006,
+ "loss": 5.570557594299316,
+ "step": 3412
+ },
+ {
+ "epoch": 47.40524017467249,
+ "grad_norm": 0.039325181394815445,
+ "learning_rate": 0.0006,
+ "loss": 5.427783966064453,
+ "step": 3413
+ },
+ {
+ "epoch": 47.419213973799124,
+ "grad_norm": 0.04034624621272087,
+ "learning_rate": 0.0006,
+ "loss": 5.572591304779053,
+ "step": 3414
+ },
+ {
+ "epoch": 47.43318777292576,
+ "grad_norm": 0.05245184153318405,
+ "learning_rate": 0.0006,
+ "loss": 5.505208969116211,
+ "step": 3415
+ },
+ {
+ "epoch": 47.4471615720524,
+ "grad_norm": 0.0568675734102726,
+ "learning_rate": 0.0006,
+ "loss": 5.589373588562012,
+ "step": 3416
+ },
+ {
+ "epoch": 47.46113537117904,
+ "grad_norm": 0.048336587846279144,
+ "learning_rate": 0.0006,
+ "loss": 5.522005081176758,
+ "step": 3417
+ },
+ {
+ "epoch": 47.47510917030568,
+ "grad_norm": 0.04194975271821022,
+ "learning_rate": 0.0006,
+ "loss": 5.387576103210449,
+ "step": 3418
+ },
+ {
+ "epoch": 47.48908296943232,
+ "grad_norm": 0.04019433259963989,
+ "learning_rate": 0.0006,
+ "loss": 5.522615432739258,
+ "step": 3419
+ },
+ {
+ "epoch": 47.50305676855895,
+ "grad_norm": 0.05709933117032051,
+ "learning_rate": 0.0006,
+ "loss": 5.420286178588867,
+ "step": 3420
+ },
+ {
+ "epoch": 47.51703056768559,
+ "grad_norm": 0.09130338579416275,
+ "learning_rate": 0.0006,
+ "loss": 5.484622478485107,
+ "step": 3421
+ },
+ {
+ "epoch": 47.531004366812226,
+ "grad_norm": 0.2757541835308075,
+ "learning_rate": 0.0006,
+ "loss": 5.821832656860352,
+ "step": 3422
+ },
+ {
+ "epoch": 47.544978165938865,
+ "grad_norm": 0.28292080760002136,
+ "learning_rate": 0.0006,
+ "loss": 6.579934120178223,
+ "step": 3423
+ },
+ {
+ "epoch": 47.5589519650655,
+ "grad_norm": 0.15293604135513306,
+ "learning_rate": 0.0006,
+ "loss": 6.478213310241699,
+ "step": 3424
+ },
+ {
+ "epoch": 47.57292576419214,
+ "grad_norm": 0.11284393817186356,
+ "learning_rate": 0.0006,
+ "loss": 6.4066386222839355,
+ "step": 3425
+ },
+ {
+ "epoch": 47.58689956331878,
+ "grad_norm": 0.10933073610067368,
+ "learning_rate": 0.0006,
+ "loss": 6.3306169509887695,
+ "step": 3426
+ },
+ {
+ "epoch": 47.60087336244541,
+ "grad_norm": 0.05837641656398773,
+ "learning_rate": 0.0006,
+ "loss": 6.250000953674316,
+ "step": 3427
+ },
+ {
+ "epoch": 47.61484716157205,
+ "grad_norm": 0.05987229570746422,
+ "learning_rate": 0.0006,
+ "loss": 6.152463436126709,
+ "step": 3428
+ },
+ {
+ "epoch": 47.62882096069869,
+ "grad_norm": 0.04064938426017761,
+ "learning_rate": 0.0006,
+ "loss": 6.117897033691406,
+ "step": 3429
+ },
+ {
+ "epoch": 47.64279475982533,
+ "grad_norm": 0.04565057530999184,
+ "learning_rate": 0.0006,
+ "loss": 6.042291641235352,
+ "step": 3430
+ },
+ {
+ "epoch": 47.65676855895197,
+ "grad_norm": 0.03475063294172287,
+ "learning_rate": 0.0006,
+ "loss": 6.0071868896484375,
+ "step": 3431
+ },
+ {
+ "epoch": 47.670742358078606,
+ "grad_norm": 0.03809220343828201,
+ "learning_rate": 0.0006,
+ "loss": 6.0128021240234375,
+ "step": 3432
+ },
+ {
+ "epoch": 47.68471615720524,
+ "grad_norm": 0.03307219594717026,
+ "learning_rate": 0.0006,
+ "loss": 5.894664287567139,
+ "step": 3433
+ },
+ {
+ "epoch": 47.698689956331876,
+ "grad_norm": 0.028824346140027046,
+ "learning_rate": 0.0006,
+ "loss": 5.914865493774414,
+ "step": 3434
+ },
+ {
+ "epoch": 47.712663755458514,
+ "grad_norm": 0.027923690155148506,
+ "learning_rate": 0.0006,
+ "loss": 5.851759910583496,
+ "step": 3435
+ },
+ {
+ "epoch": 47.72663755458515,
+ "grad_norm": 0.025492684915661812,
+ "learning_rate": 0.0006,
+ "loss": 5.749186038970947,
+ "step": 3436
+ },
+ {
+ "epoch": 47.74061135371179,
+ "grad_norm": 0.025105174630880356,
+ "learning_rate": 0.0006,
+ "loss": 5.803566932678223,
+ "step": 3437
+ },
+ {
+ "epoch": 47.75458515283843,
+ "grad_norm": 0.023298880085349083,
+ "learning_rate": 0.0006,
+ "loss": 5.76276159286499,
+ "step": 3438
+ },
+ {
+ "epoch": 47.76855895196506,
+ "grad_norm": 0.02468576282262802,
+ "learning_rate": 0.0006,
+ "loss": 5.833852767944336,
+ "step": 3439
+ },
+ {
+ "epoch": 47.7825327510917,
+ "grad_norm": 0.022730689495801926,
+ "learning_rate": 0.0006,
+ "loss": 5.736724376678467,
+ "step": 3440
+ },
+ {
+ "epoch": 47.79650655021834,
+ "grad_norm": 0.021335354074835777,
+ "learning_rate": 0.0006,
+ "loss": 5.644248008728027,
+ "step": 3441
+ },
+ {
+ "epoch": 47.81048034934498,
+ "grad_norm": 0.02288101427257061,
+ "learning_rate": 0.0006,
+ "loss": 5.7011942863464355,
+ "step": 3442
+ },
+ {
+ "epoch": 47.82445414847162,
+ "grad_norm": 0.01961585320532322,
+ "learning_rate": 0.0006,
+ "loss": 5.63215446472168,
+ "step": 3443
+ },
+ {
+ "epoch": 47.838427947598255,
+ "grad_norm": 0.017659205943346024,
+ "learning_rate": 0.0006,
+ "loss": 5.744845390319824,
+ "step": 3444
+ },
+ {
+ "epoch": 47.852401746724894,
+ "grad_norm": 0.02281784638762474,
+ "learning_rate": 0.0006,
+ "loss": 5.696993350982666,
+ "step": 3445
+ },
+ {
+ "epoch": 47.866375545851525,
+ "grad_norm": 0.016394242644309998,
+ "learning_rate": 0.0006,
+ "loss": 5.659870147705078,
+ "step": 3446
+ },
+ {
+ "epoch": 47.880349344978164,
+ "grad_norm": 0.01967368647456169,
+ "learning_rate": 0.0006,
+ "loss": 5.638072490692139,
+ "step": 3447
+ },
+ {
+ "epoch": 47.8943231441048,
+ "grad_norm": 0.017326267436146736,
+ "learning_rate": 0.0006,
+ "loss": 5.638836860656738,
+ "step": 3448
+ },
+ {
+ "epoch": 47.90829694323144,
+ "grad_norm": 0.017778463661670685,
+ "learning_rate": 0.0006,
+ "loss": 5.634466648101807,
+ "step": 3449
+ },
+ {
+ "epoch": 47.92227074235808,
+ "grad_norm": 0.017269758507609367,
+ "learning_rate": 0.0006,
+ "loss": 5.506277561187744,
+ "step": 3450
+ },
+ {
+ "epoch": 47.93624454148472,
+ "grad_norm": 0.01733289286494255,
+ "learning_rate": 0.0006,
+ "loss": 5.544455528259277,
+ "step": 3451
+ },
+ {
+ "epoch": 47.95021834061135,
+ "grad_norm": 0.017787795513868332,
+ "learning_rate": 0.0006,
+ "loss": 5.498943328857422,
+ "step": 3452
+ },
+ {
+ "epoch": 47.96419213973799,
+ "grad_norm": 0.015770932659506798,
+ "learning_rate": 0.0006,
+ "loss": 5.537227630615234,
+ "step": 3453
+ },
+ {
+ "epoch": 47.97816593886463,
+ "grad_norm": 0.016132809221744537,
+ "learning_rate": 0.0006,
+ "loss": 5.489081382751465,
+ "step": 3454
+ },
+ {
+ "epoch": 47.992139737991266,
+ "grad_norm": 0.01749664731323719,
+ "learning_rate": 0.0006,
+ "loss": 5.454263687133789,
+ "step": 3455
+ },
+ {
+ "epoch": 48.0,
+ "grad_norm": 0.016711296513676643,
+ "learning_rate": 0.0006,
+ "loss": 5.530783653259277,
+ "step": 3456
+ },
+ {
+ "epoch": 48.0,
+ "eval_loss": 5.596925258636475,
+ "eval_runtime": 56.9612,
+ "eval_samples_per_second": 42.871,
+ "eval_steps_per_second": 1.352,
+ "step": 3456
+ },
+ {
+ "epoch": 48.01397379912664,
+ "grad_norm": 0.01867641694843769,
+ "learning_rate": 0.0006,
+ "loss": 5.506853103637695,
+ "step": 3457
+ },
+ {
+ "epoch": 48.02794759825328,
+ "grad_norm": 0.023250792175531387,
+ "learning_rate": 0.0006,
+ "loss": 5.4074177742004395,
+ "step": 3458
+ },
+ {
+ "epoch": 48.041921397379916,
+ "grad_norm": 0.01970936730504036,
+ "learning_rate": 0.0006,
+ "loss": 5.508777618408203,
+ "step": 3459
+ },
+ {
+ "epoch": 48.05589519650655,
+ "grad_norm": 0.01879260502755642,
+ "learning_rate": 0.0006,
+ "loss": 5.412424087524414,
+ "step": 3460
+ },
+ {
+ "epoch": 48.069868995633186,
+ "grad_norm": 0.017420092597603798,
+ "learning_rate": 0.0006,
+ "loss": 5.43361759185791,
+ "step": 3461
+ },
+ {
+ "epoch": 48.083842794759825,
+ "grad_norm": 0.018707862123847008,
+ "learning_rate": 0.0006,
+ "loss": 5.476800918579102,
+ "step": 3462
+ },
+ {
+ "epoch": 48.09781659388646,
+ "grad_norm": 0.023185882717370987,
+ "learning_rate": 0.0006,
+ "loss": 5.465272426605225,
+ "step": 3463
+ },
+ {
+ "epoch": 48.1117903930131,
+ "grad_norm": 0.023570368066430092,
+ "learning_rate": 0.0006,
+ "loss": 5.391471862792969,
+ "step": 3464
+ },
+ {
+ "epoch": 48.12576419213974,
+ "grad_norm": 0.02651144750416279,
+ "learning_rate": 0.0006,
+ "loss": 5.422799110412598,
+ "step": 3465
+ },
+ {
+ "epoch": 48.13973799126637,
+ "grad_norm": 0.03316329047083855,
+ "learning_rate": 0.0006,
+ "loss": 5.3952202796936035,
+ "step": 3466
+ },
+ {
+ "epoch": 48.15371179039301,
+ "grad_norm": 0.04012278467416763,
+ "learning_rate": 0.0006,
+ "loss": 5.540853023529053,
+ "step": 3467
+ },
+ {
+ "epoch": 48.16768558951965,
+ "grad_norm": 0.04763203114271164,
+ "learning_rate": 0.0006,
+ "loss": 5.387890338897705,
+ "step": 3468
+ },
+ {
+ "epoch": 48.18165938864629,
+ "grad_norm": 0.05129165202379227,
+ "learning_rate": 0.0006,
+ "loss": 5.414722442626953,
+ "step": 3469
+ },
+ {
+ "epoch": 48.19563318777293,
+ "grad_norm": 0.04438071325421333,
+ "learning_rate": 0.0006,
+ "loss": 5.386244773864746,
+ "step": 3470
+ },
+ {
+ "epoch": 48.209606986899566,
+ "grad_norm": 0.040234919637441635,
+ "learning_rate": 0.0006,
+ "loss": 5.431617736816406,
+ "step": 3471
+ },
+ {
+ "epoch": 48.223580786026204,
+ "grad_norm": 0.03217008337378502,
+ "learning_rate": 0.0006,
+ "loss": 5.30168342590332,
+ "step": 3472
+ },
+ {
+ "epoch": 48.237554585152836,
+ "grad_norm": 0.03087974339723587,
+ "learning_rate": 0.0006,
+ "loss": 5.316389083862305,
+ "step": 3473
+ },
+ {
+ "epoch": 48.251528384279474,
+ "grad_norm": 0.02789347991347313,
+ "learning_rate": 0.0006,
+ "loss": 5.385970592498779,
+ "step": 3474
+ },
+ {
+ "epoch": 48.26550218340611,
+ "grad_norm": 0.024792097508907318,
+ "learning_rate": 0.0006,
+ "loss": 5.362329959869385,
+ "step": 3475
+ },
+ {
+ "epoch": 48.27947598253275,
+ "grad_norm": 0.025869522243738174,
+ "learning_rate": 0.0006,
+ "loss": 5.28400182723999,
+ "step": 3476
+ },
+ {
+ "epoch": 48.29344978165939,
+ "grad_norm": 0.021339882165193558,
+ "learning_rate": 0.0006,
+ "loss": 5.36669921875,
+ "step": 3477
+ },
+ {
+ "epoch": 48.30742358078603,
+ "grad_norm": 0.021462520584464073,
+ "learning_rate": 0.0006,
+ "loss": 5.311222076416016,
+ "step": 3478
+ },
+ {
+ "epoch": 48.32139737991266,
+ "grad_norm": 0.02193879708647728,
+ "learning_rate": 0.0006,
+ "loss": 5.291441917419434,
+ "step": 3479
+ },
+ {
+ "epoch": 48.3353711790393,
+ "grad_norm": 0.01963093690574169,
+ "learning_rate": 0.0006,
+ "loss": 5.342780113220215,
+ "step": 3480
+ },
+ {
+ "epoch": 48.34934497816594,
+ "grad_norm": 0.020634736865758896,
+ "learning_rate": 0.0006,
+ "loss": 5.144141674041748,
+ "step": 3481
+ },
+ {
+ "epoch": 48.36331877729258,
+ "grad_norm": 0.02126968279480934,
+ "learning_rate": 0.0006,
+ "loss": 5.232624530792236,
+ "step": 3482
+ },
+ {
+ "epoch": 48.377292576419215,
+ "grad_norm": 0.021203994750976562,
+ "learning_rate": 0.0006,
+ "loss": 5.290219783782959,
+ "step": 3483
+ },
+ {
+ "epoch": 48.391266375545854,
+ "grad_norm": 0.02193225920200348,
+ "learning_rate": 0.0006,
+ "loss": 5.29197359085083,
+ "step": 3484
+ },
+ {
+ "epoch": 48.40524017467249,
+ "grad_norm": 0.023174211382865906,
+ "learning_rate": 0.0006,
+ "loss": 5.204349517822266,
+ "step": 3485
+ },
+ {
+ "epoch": 48.419213973799124,
+ "grad_norm": 0.024161914363503456,
+ "learning_rate": 0.0006,
+ "loss": 5.339351654052734,
+ "step": 3486
+ },
+ {
+ "epoch": 48.43318777292576,
+ "grad_norm": 0.02221524342894554,
+ "learning_rate": 0.0006,
+ "loss": 5.263001441955566,
+ "step": 3487
+ },
+ {
+ "epoch": 48.4471615720524,
+ "grad_norm": 0.026632830500602722,
+ "learning_rate": 0.0006,
+ "loss": 5.221281051635742,
+ "step": 3488
+ },
+ {
+ "epoch": 48.46113537117904,
+ "grad_norm": 0.034923870116472244,
+ "learning_rate": 0.0006,
+ "loss": 5.296039581298828,
+ "step": 3489
+ },
+ {
+ "epoch": 48.47510917030568,
+ "grad_norm": 0.046184979379177094,
+ "learning_rate": 0.0006,
+ "loss": 5.212769985198975,
+ "step": 3490
+ },
+ {
+ "epoch": 48.48908296943232,
+ "grad_norm": 0.05154098570346832,
+ "learning_rate": 0.0006,
+ "loss": 5.337864875793457,
+ "step": 3491
+ },
+ {
+ "epoch": 48.50305676855895,
+ "grad_norm": 0.042967259883880615,
+ "learning_rate": 0.0006,
+ "loss": 5.267958641052246,
+ "step": 3492
+ },
+ {
+ "epoch": 48.51703056768559,
+ "grad_norm": 0.05356583371758461,
+ "learning_rate": 0.0006,
+ "loss": 5.1404218673706055,
+ "step": 3493
+ },
+ {
+ "epoch": 48.531004366812226,
+ "grad_norm": 0.08219032734632492,
+ "learning_rate": 0.0006,
+ "loss": 5.300687789916992,
+ "step": 3494
+ },
+ {
+ "epoch": 48.544978165938865,
+ "grad_norm": 0.19204089045524597,
+ "learning_rate": 0.0006,
+ "loss": 5.456025123596191,
+ "step": 3495
+ },
+ {
+ "epoch": 48.5589519650655,
+ "grad_norm": 0.25917696952819824,
+ "learning_rate": 0.0006,
+ "loss": 6.156761169433594,
+ "step": 3496
+ },
+ {
+ "epoch": 48.57292576419214,
+ "grad_norm": 0.1566283404827118,
+ "learning_rate": 0.0006,
+ "loss": 5.862346649169922,
+ "step": 3497
+ },
+ {
+ "epoch": 48.58689956331878,
+ "grad_norm": 0.1761486828327179,
+ "learning_rate": 0.0006,
+ "loss": 6.012795448303223,
+ "step": 3498
+ },
+ {
+ "epoch": 48.60087336244541,
+ "grad_norm": 0.09857621043920517,
+ "learning_rate": 0.0006,
+ "loss": 5.9248151779174805,
+ "step": 3499
+ },
+ {
+ "epoch": 48.61484716157205,
+ "grad_norm": 0.05968055501580238,
+ "learning_rate": 0.0006,
+ "loss": 5.780706882476807,
+ "step": 3500
+ },
+ {
+ "epoch": 48.62882096069869,
+ "grad_norm": 0.057433564215898514,
+ "learning_rate": 0.0006,
+ "loss": 5.645388603210449,
+ "step": 3501
+ },
+ {
+ "epoch": 48.64279475982533,
+ "grad_norm": 0.039659615606069565,
+ "learning_rate": 0.0006,
+ "loss": 5.695138931274414,
+ "step": 3502
+ },
+ {
+ "epoch": 48.65676855895197,
+ "grad_norm": 0.04019695147871971,
+ "learning_rate": 0.0006,
+ "loss": 5.672435283660889,
+ "step": 3503
+ },
+ {
+ "epoch": 48.670742358078606,
+ "grad_norm": 0.03033018670976162,
+ "learning_rate": 0.0006,
+ "loss": 5.682260513305664,
+ "step": 3504
+ },
+ {
+ "epoch": 48.68471615720524,
+ "grad_norm": 0.03215787187218666,
+ "learning_rate": 0.0006,
+ "loss": 5.604182243347168,
+ "step": 3505
+ },
+ {
+ "epoch": 48.698689956331876,
+ "grad_norm": 0.026865126565098763,
+ "learning_rate": 0.0006,
+ "loss": 5.64040470123291,
+ "step": 3506
+ },
+ {
+ "epoch": 48.712663755458514,
+ "grad_norm": 0.02831609919667244,
+ "learning_rate": 0.0006,
+ "loss": 5.552720069885254,
+ "step": 3507
+ },
+ {
+ "epoch": 48.72663755458515,
+ "grad_norm": 0.024226613342761993,
+ "learning_rate": 0.0006,
+ "loss": 5.556946277618408,
+ "step": 3508
+ },
+ {
+ "epoch": 48.74061135371179,
+ "grad_norm": 0.025393934920430183,
+ "learning_rate": 0.0006,
+ "loss": 5.476178169250488,
+ "step": 3509
+ },
+ {
+ "epoch": 48.75458515283843,
+ "grad_norm": 0.01984412968158722,
+ "learning_rate": 0.0006,
+ "loss": 5.471088409423828,
+ "step": 3510
+ },
+ {
+ "epoch": 48.76855895196506,
+ "grad_norm": 0.02035129815340042,
+ "learning_rate": 0.0006,
+ "loss": 5.438812732696533,
+ "step": 3511
+ },
+ {
+ "epoch": 48.7825327510917,
+ "grad_norm": 0.0208350270986557,
+ "learning_rate": 0.0006,
+ "loss": 5.381746292114258,
+ "step": 3512
+ },
+ {
+ "epoch": 48.79650655021834,
+ "grad_norm": 0.017783882096409798,
+ "learning_rate": 0.0006,
+ "loss": 5.351675510406494,
+ "step": 3513
+ },
+ {
+ "epoch": 48.81048034934498,
+ "grad_norm": 0.01671770215034485,
+ "learning_rate": 0.0006,
+ "loss": 5.329428672790527,
+ "step": 3514
+ },
+ {
+ "epoch": 48.82445414847162,
+ "grad_norm": 0.017891215160489082,
+ "learning_rate": 0.0006,
+ "loss": 5.212074279785156,
+ "step": 3515
+ },
+ {
+ "epoch": 48.838427947598255,
+ "grad_norm": 0.016847791150212288,
+ "learning_rate": 0.0006,
+ "loss": 5.299534797668457,
+ "step": 3516
+ },
+ {
+ "epoch": 48.852401746724894,
+ "grad_norm": 0.01593637838959694,
+ "learning_rate": 0.0006,
+ "loss": 5.30466890335083,
+ "step": 3517
+ },
+ {
+ "epoch": 48.866375545851525,
+ "grad_norm": 0.014713102020323277,
+ "learning_rate": 0.0006,
+ "loss": 5.294345378875732,
+ "step": 3518
+ },
+ {
+ "epoch": 48.880349344978164,
+ "grad_norm": 0.01433857623487711,
+ "learning_rate": 0.0006,
+ "loss": 5.2148542404174805,
+ "step": 3519
+ },
+ {
+ "epoch": 48.8943231441048,
+ "grad_norm": 0.014979379251599312,
+ "learning_rate": 0.0006,
+ "loss": 5.298582077026367,
+ "step": 3520
+ },
+ {
+ "epoch": 48.90829694323144,
+ "grad_norm": 0.013494192622601986,
+ "learning_rate": 0.0006,
+ "loss": 5.309783935546875,
+ "step": 3521
+ },
+ {
+ "epoch": 48.92227074235808,
+ "grad_norm": 0.014020012691617012,
+ "learning_rate": 0.0006,
+ "loss": 5.284879684448242,
+ "step": 3522
+ },
+ {
+ "epoch": 48.93624454148472,
+ "grad_norm": 0.01348058134317398,
+ "learning_rate": 0.0006,
+ "loss": 5.249116897583008,
+ "step": 3523
+ },
+ {
+ "epoch": 48.95021834061135,
+ "grad_norm": 0.013280218467116356,
+ "learning_rate": 0.0006,
+ "loss": 5.273489952087402,
+ "step": 3524
+ },
+ {
+ "epoch": 48.96419213973799,
+ "grad_norm": 0.012647110037505627,
+ "learning_rate": 0.0006,
+ "loss": 5.306652069091797,
+ "step": 3525
+ },
+ {
+ "epoch": 48.97816593886463,
+ "grad_norm": 0.012847724370658398,
+ "learning_rate": 0.0006,
+ "loss": 5.235745429992676,
+ "step": 3526
+ },
+ {
+ "epoch": 48.992139737991266,
+ "grad_norm": 0.011820166371762753,
+ "learning_rate": 0.0006,
+ "loss": 5.238091468811035,
+ "step": 3527
+ },
+ {
+ "epoch": 49.0,
+ "grad_norm": 0.01255585066974163,
+ "learning_rate": 0.0006,
+ "loss": 5.231295585632324,
+ "step": 3528
+ },
+ {
+ "epoch": 49.0,
+ "eval_loss": 5.394587993621826,
+ "eval_runtime": 56.3734,
+ "eval_samples_per_second": 43.318,
+ "eval_steps_per_second": 1.366,
+ "step": 3528
+ },
+ {
+ "epoch": 49.01397379912664,
+ "grad_norm": 0.011031972244381905,
+ "learning_rate": 0.0006,
+ "loss": 5.1354570388793945,
+ "step": 3529
+ },
+ {
+ "epoch": 49.02794759825328,
+ "grad_norm": 0.010631263256072998,
+ "learning_rate": 0.0006,
+ "loss": 5.177639961242676,
+ "step": 3530
+ },
+ {
+ "epoch": 49.041921397379916,
+ "grad_norm": 0.01133775431662798,
+ "learning_rate": 0.0006,
+ "loss": 5.283428192138672,
+ "step": 3531
+ },
+ {
+ "epoch": 49.05589519650655,
+ "grad_norm": 0.01154172234237194,
+ "learning_rate": 0.0006,
+ "loss": 5.206364154815674,
+ "step": 3532
+ },
+ {
+ "epoch": 49.069868995633186,
+ "grad_norm": 0.010192356072366238,
+ "learning_rate": 0.0006,
+ "loss": 5.2385735511779785,
+ "step": 3533
+ },
+ {
+ "epoch": 49.083842794759825,
+ "grad_norm": 0.012789200060069561,
+ "learning_rate": 0.0006,
+ "loss": 5.071800231933594,
+ "step": 3534
+ },
+ {
+ "epoch": 49.09781659388646,
+ "grad_norm": 0.011282820254564285,
+ "learning_rate": 0.0006,
+ "loss": 5.218020439147949,
+ "step": 3535
+ },
+ {
+ "epoch": 49.1117903930131,
+ "grad_norm": 0.012153448536992073,
+ "learning_rate": 0.0006,
+ "loss": 5.134698867797852,
+ "step": 3536
+ },
+ {
+ "epoch": 49.12576419213974,
+ "grad_norm": 0.011920543387532234,
+ "learning_rate": 0.0006,
+ "loss": 5.141620635986328,
+ "step": 3537
+ },
+ {
+ "epoch": 49.13973799126637,
+ "grad_norm": 0.010949329473078251,
+ "learning_rate": 0.0006,
+ "loss": 5.116259574890137,
+ "step": 3538
+ },
+ {
+ "epoch": 49.15371179039301,
+ "grad_norm": 0.012362448498606682,
+ "learning_rate": 0.0006,
+ "loss": 5.090177059173584,
+ "step": 3539
+ },
+ {
+ "epoch": 49.16768558951965,
+ "grad_norm": 0.011575071141123772,
+ "learning_rate": 0.0006,
+ "loss": 5.093441009521484,
+ "step": 3540
+ },
+ {
+ "epoch": 49.18165938864629,
+ "grad_norm": 0.011126236990094185,
+ "learning_rate": 0.0006,
+ "loss": 5.060235023498535,
+ "step": 3541
+ },
+ {
+ "epoch": 49.19563318777293,
+ "grad_norm": 0.0107401292771101,
+ "learning_rate": 0.0006,
+ "loss": 5.140704154968262,
+ "step": 3542
+ },
+ {
+ "epoch": 49.209606986899566,
+ "grad_norm": 0.011202906258404255,
+ "learning_rate": 0.0006,
+ "loss": 5.10855770111084,
+ "step": 3543
+ },
+ {
+ "epoch": 49.223580786026204,
+ "grad_norm": 0.011083454824984074,
+ "learning_rate": 0.0006,
+ "loss": 5.092751502990723,
+ "step": 3544
+ },
+ {
+ "epoch": 49.237554585152836,
+ "grad_norm": 0.009523800574243069,
+ "learning_rate": 0.0006,
+ "loss": 5.13733434677124,
+ "step": 3545
+ },
+ {
+ "epoch": 49.251528384279474,
+ "grad_norm": 0.011298353783786297,
+ "learning_rate": 0.0006,
+ "loss": 5.036048889160156,
+ "step": 3546
+ },
+ {
+ "epoch": 49.26550218340611,
+ "grad_norm": 0.010676483623683453,
+ "learning_rate": 0.0006,
+ "loss": 5.051901340484619,
+ "step": 3547
+ },
+ {
+ "epoch": 49.27947598253275,
+ "grad_norm": 0.011463399045169353,
+ "learning_rate": 0.0006,
+ "loss": 4.983170986175537,
+ "step": 3548
+ },
+ {
+ "epoch": 49.29344978165939,
+ "grad_norm": 0.010603402741253376,
+ "learning_rate": 0.0006,
+ "loss": 5.112311363220215,
+ "step": 3549
+ },
+ {
+ "epoch": 49.30742358078603,
+ "grad_norm": 0.010535670444369316,
+ "learning_rate": 0.0006,
+ "loss": 5.085227966308594,
+ "step": 3550
+ },
+ {
+ "epoch": 49.32139737991266,
+ "grad_norm": 0.011843928135931492,
+ "learning_rate": 0.0006,
+ "loss": 5.07575798034668,
+ "step": 3551
+ },
+ {
+ "epoch": 49.3353711790393,
+ "grad_norm": 0.010587952099740505,
+ "learning_rate": 0.0006,
+ "loss": 5.019237041473389,
+ "step": 3552
+ },
+ {
+ "epoch": 49.34934497816594,
+ "grad_norm": 0.012285659089684486,
+ "learning_rate": 0.0006,
+ "loss": 5.144603729248047,
+ "step": 3553
+ },
+ {
+ "epoch": 49.36331877729258,
+ "grad_norm": 0.013119431212544441,
+ "learning_rate": 0.0006,
+ "loss": 5.035256862640381,
+ "step": 3554
+ },
+ {
+ "epoch": 49.377292576419215,
+ "grad_norm": 0.012138537131249905,
+ "learning_rate": 0.0006,
+ "loss": 4.996667861938477,
+ "step": 3555
+ },
+ {
+ "epoch": 49.391266375545854,
+ "grad_norm": 0.012142804451286793,
+ "learning_rate": 0.0006,
+ "loss": 4.95058536529541,
+ "step": 3556
+ },
+ {
+ "epoch": 49.40524017467249,
+ "grad_norm": 0.012205676175653934,
+ "learning_rate": 0.0006,
+ "loss": 5.096931457519531,
+ "step": 3557
+ },
+ {
+ "epoch": 49.419213973799124,
+ "grad_norm": 0.012086744420230389,
+ "learning_rate": 0.0006,
+ "loss": 4.99384880065918,
+ "step": 3558
+ },
+ {
+ "epoch": 49.43318777292576,
+ "grad_norm": 0.012918103486299515,
+ "learning_rate": 0.0006,
+ "loss": 5.031294822692871,
+ "step": 3559
+ },
+ {
+ "epoch": 49.4471615720524,
+ "grad_norm": 0.012006713077425957,
+ "learning_rate": 0.0006,
+ "loss": 5.0547943115234375,
+ "step": 3560
+ },
+ {
+ "epoch": 49.46113537117904,
+ "grad_norm": 0.0120399734005332,
+ "learning_rate": 0.0006,
+ "loss": 5.014238357543945,
+ "step": 3561
+ },
+ {
+ "epoch": 49.47510917030568,
+ "grad_norm": 0.013306990265846252,
+ "learning_rate": 0.0006,
+ "loss": 4.972925186157227,
+ "step": 3562
+ },
+ {
+ "epoch": 49.48908296943232,
+ "grad_norm": 0.013878636993467808,
+ "learning_rate": 0.0006,
+ "loss": 5.079033851623535,
+ "step": 3563
+ },
+ {
+ "epoch": 49.50305676855895,
+ "grad_norm": 0.012454566545784473,
+ "learning_rate": 0.0006,
+ "loss": 4.93708610534668,
+ "step": 3564
+ },
+ {
+ "epoch": 49.51703056768559,
+ "grad_norm": 0.014075061306357384,
+ "learning_rate": 0.0006,
+ "loss": 4.965582370758057,
+ "step": 3565
+ },
+ {
+ "epoch": 49.531004366812226,
+ "grad_norm": 0.012525617145001888,
+ "learning_rate": 0.0006,
+ "loss": 4.968148708343506,
+ "step": 3566
+ },
+ {
+ "epoch": 49.544978165938865,
+ "grad_norm": 0.012418225407600403,
+ "learning_rate": 0.0006,
+ "loss": 5.06901741027832,
+ "step": 3567
+ },
+ {
+ "epoch": 49.5589519650655,
+ "grad_norm": 0.015034239739179611,
+ "learning_rate": 0.0006,
+ "loss": 4.984277725219727,
+ "step": 3568
+ },
+ {
+ "epoch": 49.57292576419214,
+ "grad_norm": 0.014329628087580204,
+ "learning_rate": 0.0006,
+ "loss": 4.951558589935303,
+ "step": 3569
+ },
+ {
+ "epoch": 49.58689956331878,
+ "grad_norm": 0.017769070342183113,
+ "learning_rate": 0.0006,
+ "loss": 4.966532230377197,
+ "step": 3570
+ },
+ {
+ "epoch": 49.60087336244541,
+ "grad_norm": 0.01909453794360161,
+ "learning_rate": 0.0006,
+ "loss": 4.919890880584717,
+ "step": 3571
+ },
+ {
+ "epoch": 49.61484716157205,
+ "grad_norm": 0.017519688233733177,
+ "learning_rate": 0.0006,
+ "loss": 5.031286716461182,
+ "step": 3572
+ },
+ {
+ "epoch": 49.62882096069869,
+ "grad_norm": 0.017143897712230682,
+ "learning_rate": 0.0006,
+ "loss": 4.969764709472656,
+ "step": 3573
+ },
+ {
+ "epoch": 49.64279475982533,
+ "grad_norm": 0.01688285917043686,
+ "learning_rate": 0.0006,
+ "loss": 5.0733442306518555,
+ "step": 3574
+ },
+ {
+ "epoch": 49.65676855895197,
+ "grad_norm": 0.018547311425209045,
+ "learning_rate": 0.0006,
+ "loss": 4.980863571166992,
+ "step": 3575
+ },
+ {
+ "epoch": 49.670742358078606,
+ "grad_norm": 0.018877137452363968,
+ "learning_rate": 0.0006,
+ "loss": 4.954646110534668,
+ "step": 3576
+ },
+ {
+ "epoch": 49.68471615720524,
+ "grad_norm": 0.018681600689888,
+ "learning_rate": 0.0006,
+ "loss": 4.959702014923096,
+ "step": 3577
+ },
+ {
+ "epoch": 49.698689956331876,
+ "grad_norm": 0.018947336822748184,
+ "learning_rate": 0.0006,
+ "loss": 5.070502758026123,
+ "step": 3578
+ },
+ {
+ "epoch": 49.712663755458514,
+ "grad_norm": 0.018817484378814697,
+ "learning_rate": 0.0006,
+ "loss": 5.056508541107178,
+ "step": 3579
+ },
+ {
+ "epoch": 49.72663755458515,
+ "grad_norm": 0.01931724324822426,
+ "learning_rate": 0.0006,
+ "loss": 4.922451972961426,
+ "step": 3580
+ },
+ {
+ "epoch": 49.74061135371179,
+ "grad_norm": 0.022047819569706917,
+ "learning_rate": 0.0006,
+ "loss": 5.053170204162598,
+ "step": 3581
+ },
+ {
+ "epoch": 49.75458515283843,
+ "grad_norm": 0.02279801294207573,
+ "learning_rate": 0.0006,
+ "loss": 4.8768439292907715,
+ "step": 3582
+ },
+ {
+ "epoch": 49.76855895196506,
+ "grad_norm": 0.020661601796746254,
+ "learning_rate": 0.0006,
+ "loss": 4.907928466796875,
+ "step": 3583
+ },
+ {
+ "epoch": 49.7825327510917,
+ "grad_norm": 0.020479509606957436,
+ "learning_rate": 0.0006,
+ "loss": 5.040694236755371,
+ "step": 3584
+ },
+ {
+ "epoch": 49.79650655021834,
+ "grad_norm": 0.021654162555933,
+ "learning_rate": 0.0006,
+ "loss": 5.018434524536133,
+ "step": 3585
+ },
+ {
+ "epoch": 49.81048034934498,
+ "grad_norm": 0.022486824542284012,
+ "learning_rate": 0.0006,
+ "loss": 5.086538314819336,
+ "step": 3586
+ },
+ {
+ "epoch": 49.82445414847162,
+ "grad_norm": 0.02141369879245758,
+ "learning_rate": 0.0006,
+ "loss": 5.035694122314453,
+ "step": 3587
+ },
+ {
+ "epoch": 49.838427947598255,
+ "grad_norm": 0.015926677733659744,
+ "learning_rate": 0.0006,
+ "loss": 4.905038833618164,
+ "step": 3588
+ },
+ {
+ "epoch": 49.852401746724894,
+ "grad_norm": 0.015775298699736595,
+ "learning_rate": 0.0006,
+ "loss": 5.003902435302734,
+ "step": 3589
+ },
+ {
+ "epoch": 49.866375545851525,
+ "grad_norm": 0.01872118003666401,
+ "learning_rate": 0.0006,
+ "loss": 4.891060829162598,
+ "step": 3590
+ },
+ {
+ "epoch": 49.880349344978164,
+ "grad_norm": 0.02066192403435707,
+ "learning_rate": 0.0006,
+ "loss": 5.002897262573242,
+ "step": 3591
+ },
+ {
+ "epoch": 49.8943231441048,
+ "grad_norm": 0.02348714880645275,
+ "learning_rate": 0.0006,
+ "loss": 4.9189958572387695,
+ "step": 3592
+ },
+ {
+ "epoch": 49.90829694323144,
+ "grad_norm": 0.020402103662490845,
+ "learning_rate": 0.0006,
+ "loss": 4.992661476135254,
+ "step": 3593
+ },
+ {
+ "epoch": 49.92227074235808,
+ "grad_norm": 0.020466072484850883,
+ "learning_rate": 0.0006,
+ "loss": 4.962968826293945,
+ "step": 3594
+ },
+ {
+ "epoch": 49.93624454148472,
+ "grad_norm": 0.01922314241528511,
+ "learning_rate": 0.0006,
+ "loss": 4.897730827331543,
+ "step": 3595
+ },
+ {
+ "epoch": 49.95021834061135,
+ "grad_norm": 0.017228754237294197,
+ "learning_rate": 0.0006,
+ "loss": 4.964842796325684,
+ "step": 3596
+ },
+ {
+ "epoch": 49.96419213973799,
+ "grad_norm": 0.016021743416786194,
+ "learning_rate": 0.0006,
+ "loss": 4.910758972167969,
+ "step": 3597
+ },
+ {
+ "epoch": 49.97816593886463,
+ "grad_norm": 0.017985103651881218,
+ "learning_rate": 0.0006,
+ "loss": 4.916072368621826,
+ "step": 3598
+ },
+ {
+ "epoch": 49.992139737991266,
+ "grad_norm": 0.018780240789055824,
+ "learning_rate": 0.0006,
+ "loss": 4.94658088684082,
+ "step": 3599
+ },
+ {
+ "epoch": 50.0,
+ "grad_norm": 0.01776314526796341,
+ "learning_rate": 0.0006,
+ "loss": 4.969357967376709,
+ "step": 3600
+ },
+ {
+ "epoch": 50.0,
+ "eval_loss": 5.141158580780029,
+ "eval_runtime": 56.2332,
+ "eval_samples_per_second": 43.426,
+ "eval_steps_per_second": 1.369,
+ "step": 3600
+ },
+ {
+ "epoch": 50.01397379912664,
+ "grad_norm": 0.016308696940541267,
+ "learning_rate": 0.0006,
+ "loss": 4.924664497375488,
+ "step": 3601
+ },
+ {
+ "epoch": 50.02794759825328,
+ "grad_norm": 0.018810199573636055,
+ "learning_rate": 0.0006,
+ "loss": 4.82293701171875,
+ "step": 3602
+ },
+ {
+ "epoch": 50.041921397379916,
+ "grad_norm": 0.02227160707116127,
+ "learning_rate": 0.0006,
+ "loss": 4.9612298011779785,
+ "step": 3603
+ },
+ {
+ "epoch": 50.05589519650655,
+ "grad_norm": 0.025266462936997414,
+ "learning_rate": 0.0006,
+ "loss": 4.963835716247559,
+ "step": 3604
+ },
+ {
+ "epoch": 50.069868995633186,
+ "grad_norm": 0.028588181361556053,
+ "learning_rate": 0.0006,
+ "loss": 4.992118835449219,
+ "step": 3605
+ },
+ {
+ "epoch": 50.083842794759825,
+ "grad_norm": 0.02739487960934639,
+ "learning_rate": 0.0006,
+ "loss": 4.915547847747803,
+ "step": 3606
+ },
+ {
+ "epoch": 50.09781659388646,
+ "grad_norm": 0.027000300586223602,
+ "learning_rate": 0.0006,
+ "loss": 4.883801460266113,
+ "step": 3607
+ },
+ {
+ "epoch": 50.1117903930131,
+ "grad_norm": 0.023468483239412308,
+ "learning_rate": 0.0006,
+ "loss": 5.006168365478516,
+ "step": 3608
+ },
+ {
+ "epoch": 50.12576419213974,
+ "grad_norm": 0.026753634214401245,
+ "learning_rate": 0.0006,
+ "loss": 4.92526912689209,
+ "step": 3609
+ },
+ {
+ "epoch": 50.13973799126637,
+ "grad_norm": 0.026816733181476593,
+ "learning_rate": 0.0006,
+ "loss": 4.847546100616455,
+ "step": 3610
+ },
+ {
+ "epoch": 50.15371179039301,
+ "grad_norm": 0.027362391352653503,
+ "learning_rate": 0.0006,
+ "loss": 4.862548351287842,
+ "step": 3611
+ },
+ {
+ "epoch": 50.16768558951965,
+ "grad_norm": 0.03450845181941986,
+ "learning_rate": 0.0006,
+ "loss": 4.963202953338623,
+ "step": 3612
+ },
+ {
+ "epoch": 50.18165938864629,
+ "grad_norm": 0.03801408037543297,
+ "learning_rate": 0.0006,
+ "loss": 4.770240306854248,
+ "step": 3613
+ },
+ {
+ "epoch": 50.19563318777293,
+ "grad_norm": 0.03579140082001686,
+ "learning_rate": 0.0006,
+ "loss": 4.797981262207031,
+ "step": 3614
+ },
+ {
+ "epoch": 50.209606986899566,
+ "grad_norm": 0.029393760487437248,
+ "learning_rate": 0.0006,
+ "loss": 4.798581600189209,
+ "step": 3615
+ },
+ {
+ "epoch": 50.223580786026204,
+ "grad_norm": 0.026111142709851265,
+ "learning_rate": 0.0006,
+ "loss": 4.990712642669678,
+ "step": 3616
+ },
+ {
+ "epoch": 50.237554585152836,
+ "grad_norm": 0.02213387005031109,
+ "learning_rate": 0.0006,
+ "loss": 4.865842342376709,
+ "step": 3617
+ },
+ {
+ "epoch": 50.251528384279474,
+ "grad_norm": 0.025450684130191803,
+ "learning_rate": 0.0006,
+ "loss": 4.903746128082275,
+ "step": 3618
+ },
+ {
+ "epoch": 50.26550218340611,
+ "grad_norm": 0.020915117114782333,
+ "learning_rate": 0.0006,
+ "loss": 4.951366424560547,
+ "step": 3619
+ },
+ {
+ "epoch": 50.27947598253275,
+ "grad_norm": 0.020972849801182747,
+ "learning_rate": 0.0006,
+ "loss": 4.783311367034912,
+ "step": 3620
+ },
+ {
+ "epoch": 50.29344978165939,
+ "grad_norm": 0.016956273466348648,
+ "learning_rate": 0.0006,
+ "loss": 4.955767631530762,
+ "step": 3621
+ },
+ {
+ "epoch": 50.30742358078603,
+ "grad_norm": 0.018045252189040184,
+ "learning_rate": 0.0006,
+ "loss": 4.869322776794434,
+ "step": 3622
+ },
+ {
+ "epoch": 50.32139737991266,
+ "grad_norm": 0.018199188634753227,
+ "learning_rate": 0.0006,
+ "loss": 4.95680046081543,
+ "step": 3623
+ },
+ {
+ "epoch": 50.3353711790393,
+ "grad_norm": 0.016688082367181778,
+ "learning_rate": 0.0006,
+ "loss": 4.783535957336426,
+ "step": 3624
+ },
+ {
+ "epoch": 50.34934497816594,
+ "grad_norm": 0.017540352419018745,
+ "learning_rate": 0.0006,
+ "loss": 4.8236083984375,
+ "step": 3625
+ },
+ {
+ "epoch": 50.36331877729258,
+ "grad_norm": 0.018054187297821045,
+ "learning_rate": 0.0006,
+ "loss": 4.764655113220215,
+ "step": 3626
+ },
+ {
+ "epoch": 50.377292576419215,
+ "grad_norm": 0.01741131953895092,
+ "learning_rate": 0.0006,
+ "loss": 4.875613689422607,
+ "step": 3627
+ },
+ {
+ "epoch": 50.391266375545854,
+ "grad_norm": 0.01682269014418125,
+ "learning_rate": 0.0006,
+ "loss": 4.818509101867676,
+ "step": 3628
+ },
+ {
+ "epoch": 50.40524017467249,
+ "grad_norm": 0.016652341932058334,
+ "learning_rate": 0.0006,
+ "loss": 4.974020481109619,
+ "step": 3629
+ },
+ {
+ "epoch": 50.419213973799124,
+ "grad_norm": 0.02035282365977764,
+ "learning_rate": 0.0006,
+ "loss": 4.893037796020508,
+ "step": 3630
+ },
+ {
+ "epoch": 50.43318777292576,
+ "grad_norm": 0.019907061010599136,
+ "learning_rate": 0.0006,
+ "loss": 4.848119735717773,
+ "step": 3631
+ },
+ {
+ "epoch": 50.4471615720524,
+ "grad_norm": 0.023861493915319443,
+ "learning_rate": 0.0006,
+ "loss": 4.819353103637695,
+ "step": 3632
+ },
+ {
+ "epoch": 50.46113537117904,
+ "grad_norm": 0.025254026055336,
+ "learning_rate": 0.0006,
+ "loss": 4.769312858581543,
+ "step": 3633
+ },
+ {
+ "epoch": 50.47510917030568,
+ "grad_norm": 0.026154937222599983,
+ "learning_rate": 0.0006,
+ "loss": 4.7586774826049805,
+ "step": 3634
+ },
+ {
+ "epoch": 50.48908296943232,
+ "grad_norm": 0.026342090219259262,
+ "learning_rate": 0.0006,
+ "loss": 4.861218452453613,
+ "step": 3635
+ },
+ {
+ "epoch": 50.50305676855895,
+ "grad_norm": 0.028169559314846992,
+ "learning_rate": 0.0006,
+ "loss": 4.818380355834961,
+ "step": 3636
+ },
+ {
+ "epoch": 50.51703056768559,
+ "grad_norm": 0.02820264920592308,
+ "learning_rate": 0.0006,
+ "loss": 4.892871856689453,
+ "step": 3637
+ },
+ {
+ "epoch": 50.531004366812226,
+ "grad_norm": 0.03345080465078354,
+ "learning_rate": 0.0006,
+ "loss": 4.811084747314453,
+ "step": 3638
+ },
+ {
+ "epoch": 50.544978165938865,
+ "grad_norm": 0.04005736857652664,
+ "learning_rate": 0.0006,
+ "loss": 4.765410900115967,
+ "step": 3639
+ },
+ {
+ "epoch": 50.5589519650655,
+ "grad_norm": 0.04333870857954025,
+ "learning_rate": 0.0006,
+ "loss": 4.956520080566406,
+ "step": 3640
+ },
+ {
+ "epoch": 50.57292576419214,
+ "grad_norm": 0.04216952994465828,
+ "learning_rate": 0.0006,
+ "loss": 4.793441295623779,
+ "step": 3641
+ },
+ {
+ "epoch": 50.58689956331878,
+ "grad_norm": 0.02936951443552971,
+ "learning_rate": 0.0006,
+ "loss": 4.8447957038879395,
+ "step": 3642
+ },
+ {
+ "epoch": 50.60087336244541,
+ "grad_norm": 0.026937026530504227,
+ "learning_rate": 0.0006,
+ "loss": 4.962215423583984,
+ "step": 3643
+ },
+ {
+ "epoch": 50.61484716157205,
+ "grad_norm": 0.02628890983760357,
+ "learning_rate": 0.0006,
+ "loss": 4.883061408996582,
+ "step": 3644
+ },
+ {
+ "epoch": 50.62882096069869,
+ "grad_norm": 0.025984834879636765,
+ "learning_rate": 0.0006,
+ "loss": 4.768662452697754,
+ "step": 3645
+ },
+ {
+ "epoch": 50.64279475982533,
+ "grad_norm": 0.023998012766242027,
+ "learning_rate": 0.0006,
+ "loss": 4.851325035095215,
+ "step": 3646
+ },
+ {
+ "epoch": 50.65676855895197,
+ "grad_norm": 0.02143184095621109,
+ "learning_rate": 0.0006,
+ "loss": 4.8274688720703125,
+ "step": 3647
+ },
+ {
+ "epoch": 50.670742358078606,
+ "grad_norm": 0.02488900162279606,
+ "learning_rate": 0.0006,
+ "loss": 4.831140518188477,
+ "step": 3648
+ },
+ {
+ "epoch": 50.68471615720524,
+ "grad_norm": 0.02330533228814602,
+ "learning_rate": 0.0006,
+ "loss": 4.850462913513184,
+ "step": 3649
+ },
+ {
+ "epoch": 50.698689956331876,
+ "grad_norm": 0.02121943049132824,
+ "learning_rate": 0.0006,
+ "loss": 4.870253086090088,
+ "step": 3650
+ },
+ {
+ "epoch": 50.712663755458514,
+ "grad_norm": 0.02080935798585415,
+ "learning_rate": 0.0006,
+ "loss": 4.798770904541016,
+ "step": 3651
+ },
+ {
+ "epoch": 50.72663755458515,
+ "grad_norm": 0.019047029316425323,
+ "learning_rate": 0.0006,
+ "loss": 4.8427300453186035,
+ "step": 3652
+ },
+ {
+ "epoch": 50.74061135371179,
+ "grad_norm": 0.01824103854596615,
+ "learning_rate": 0.0006,
+ "loss": 4.820980548858643,
+ "step": 3653
+ },
+ {
+ "epoch": 50.75458515283843,
+ "grad_norm": 0.018129967153072357,
+ "learning_rate": 0.0006,
+ "loss": 4.946732997894287,
+ "step": 3654
+ },
+ {
+ "epoch": 50.76855895196506,
+ "grad_norm": 0.01784815452992916,
+ "learning_rate": 0.0006,
+ "loss": 4.829516410827637,
+ "step": 3655
+ },
+ {
+ "epoch": 50.7825327510917,
+ "grad_norm": 0.020021555945277214,
+ "learning_rate": 0.0006,
+ "loss": 4.780428409576416,
+ "step": 3656
+ },
+ {
+ "epoch": 50.79650655021834,
+ "grad_norm": 0.01863253302872181,
+ "learning_rate": 0.0006,
+ "loss": 4.791567802429199,
+ "step": 3657
+ },
+ {
+ "epoch": 50.81048034934498,
+ "grad_norm": 0.019151683896780014,
+ "learning_rate": 0.0006,
+ "loss": 4.824892997741699,
+ "step": 3658
+ },
+ {
+ "epoch": 50.82445414847162,
+ "grad_norm": 0.01891648955643177,
+ "learning_rate": 0.0006,
+ "loss": 4.789605140686035,
+ "step": 3659
+ },
+ {
+ "epoch": 50.838427947598255,
+ "grad_norm": 0.016019510105252266,
+ "learning_rate": 0.0006,
+ "loss": 4.779555320739746,
+ "step": 3660
+ },
+ {
+ "epoch": 50.852401746724894,
+ "grad_norm": 0.016096491366624832,
+ "learning_rate": 0.0006,
+ "loss": 4.753842353820801,
+ "step": 3661
+ },
+ {
+ "epoch": 50.866375545851525,
+ "grad_norm": 0.01693476364016533,
+ "learning_rate": 0.0006,
+ "loss": 4.768477439880371,
+ "step": 3662
+ },
+ {
+ "epoch": 50.880349344978164,
+ "grad_norm": 0.01749575324356556,
+ "learning_rate": 0.0006,
+ "loss": 4.779143333435059,
+ "step": 3663
+ },
+ {
+ "epoch": 50.8943231441048,
+ "grad_norm": 0.016851043328642845,
+ "learning_rate": 0.0006,
+ "loss": 4.780387878417969,
+ "step": 3664
+ },
+ {
+ "epoch": 50.90829694323144,
+ "grad_norm": 0.019045202061533928,
+ "learning_rate": 0.0006,
+ "loss": 4.769913673400879,
+ "step": 3665
+ },
+ {
+ "epoch": 50.92227074235808,
+ "grad_norm": 0.01954123005270958,
+ "learning_rate": 0.0006,
+ "loss": 4.769009590148926,
+ "step": 3666
+ },
+ {
+ "epoch": 50.93624454148472,
+ "grad_norm": 0.022527096793055534,
+ "learning_rate": 0.0006,
+ "loss": 4.789215087890625,
+ "step": 3667
+ },
+ {
+ "epoch": 50.95021834061135,
+ "grad_norm": 0.02871634066104889,
+ "learning_rate": 0.0006,
+ "loss": 4.749727725982666,
+ "step": 3668
+ },
+ {
+ "epoch": 50.96419213973799,
+ "grad_norm": 0.04045053571462631,
+ "learning_rate": 0.0006,
+ "loss": 4.909038543701172,
+ "step": 3669
+ },
+ {
+ "epoch": 50.97816593886463,
+ "grad_norm": 0.04156879335641861,
+ "learning_rate": 0.0006,
+ "loss": 4.876831531524658,
+ "step": 3670
+ },
+ {
+ "epoch": 50.992139737991266,
+ "grad_norm": 0.02621397376060486,
+ "learning_rate": 0.0006,
+ "loss": 4.797173500061035,
+ "step": 3671
+ },
+ {
+ "epoch": 51.0,
+ "grad_norm": 0.03389597684144974,
+ "learning_rate": 0.0006,
+ "loss": 4.8428850173950195,
+ "step": 3672
+ },
+ {
+ "epoch": 51.0,
+ "eval_loss": 4.989261627197266,
+ "eval_runtime": 56.358,
+ "eval_samples_per_second": 43.33,
+ "eval_steps_per_second": 1.366,
+ "step": 3672
+ },
+ {
+ "epoch": 51.01397379912664,
+ "grad_norm": 0.03216972202062607,
+ "learning_rate": 0.0006,
+ "loss": 4.755102634429932,
+ "step": 3673
+ },
+ {
+ "epoch": 51.02794759825328,
+ "grad_norm": 0.03190374746918678,
+ "learning_rate": 0.0006,
+ "loss": 4.720424652099609,
+ "step": 3674
+ },
+ {
+ "epoch": 51.041921397379916,
+ "grad_norm": 0.031727299094200134,
+ "learning_rate": 0.0006,
+ "loss": 4.803277969360352,
+ "step": 3675
+ },
+ {
+ "epoch": 51.05589519650655,
+ "grad_norm": 0.02473079040646553,
+ "learning_rate": 0.0006,
+ "loss": 4.761087417602539,
+ "step": 3676
+ },
+ {
+ "epoch": 51.069868995633186,
+ "grad_norm": 0.020360209047794342,
+ "learning_rate": 0.0006,
+ "loss": 4.753884792327881,
+ "step": 3677
+ },
+ {
+ "epoch": 51.083842794759825,
+ "grad_norm": 0.02078923024237156,
+ "learning_rate": 0.0006,
+ "loss": 4.647130489349365,
+ "step": 3678
+ },
+ {
+ "epoch": 51.09781659388646,
+ "grad_norm": 0.019387200474739075,
+ "learning_rate": 0.0006,
+ "loss": 4.74135160446167,
+ "step": 3679
+ },
+ {
+ "epoch": 51.1117903930131,
+ "grad_norm": 0.01874949410557747,
+ "learning_rate": 0.0006,
+ "loss": 4.795329570770264,
+ "step": 3680
+ },
+ {
+ "epoch": 51.12576419213974,
+ "grad_norm": 0.019331837072968483,
+ "learning_rate": 0.0006,
+ "loss": 4.897967338562012,
+ "step": 3681
+ },
+ {
+ "epoch": 51.13973799126637,
+ "grad_norm": 0.017066050320863724,
+ "learning_rate": 0.0006,
+ "loss": 4.825351715087891,
+ "step": 3682
+ },
+ {
+ "epoch": 51.15371179039301,
+ "grad_norm": 0.01757810264825821,
+ "learning_rate": 0.0006,
+ "loss": 4.719335079193115,
+ "step": 3683
+ },
+ {
+ "epoch": 51.16768558951965,
+ "grad_norm": 0.018174266442656517,
+ "learning_rate": 0.0006,
+ "loss": 4.732680320739746,
+ "step": 3684
+ },
+ {
+ "epoch": 51.18165938864629,
+ "grad_norm": 0.018497727811336517,
+ "learning_rate": 0.0006,
+ "loss": 4.731050491333008,
+ "step": 3685
+ },
+ {
+ "epoch": 51.19563318777293,
+ "grad_norm": 0.0186338871717453,
+ "learning_rate": 0.0006,
+ "loss": 4.748035430908203,
+ "step": 3686
+ },
+ {
+ "epoch": 51.209606986899566,
+ "grad_norm": 0.018531711772084236,
+ "learning_rate": 0.0006,
+ "loss": 4.9007463455200195,
+ "step": 3687
+ },
+ {
+ "epoch": 51.223580786026204,
+ "grad_norm": 0.018221354112029076,
+ "learning_rate": 0.0006,
+ "loss": 4.84077262878418,
+ "step": 3688
+ },
+ {
+ "epoch": 51.237554585152836,
+ "grad_norm": 0.0203656405210495,
+ "learning_rate": 0.0006,
+ "loss": 4.73009729385376,
+ "step": 3689
+ },
+ {
+ "epoch": 51.251528384279474,
+ "grad_norm": 0.01964893378317356,
+ "learning_rate": 0.0006,
+ "loss": 4.772583961486816,
+ "step": 3690
+ },
+ {
+ "epoch": 51.26550218340611,
+ "grad_norm": 0.019606510177254677,
+ "learning_rate": 0.0006,
+ "loss": 4.599812030792236,
+ "step": 3691
+ },
+ {
+ "epoch": 51.27947598253275,
+ "grad_norm": 0.020501255989074707,
+ "learning_rate": 0.0006,
+ "loss": 4.809898853302002,
+ "step": 3692
+ },
+ {
+ "epoch": 51.29344978165939,
+ "grad_norm": 0.01870838552713394,
+ "learning_rate": 0.0006,
+ "loss": 4.741898059844971,
+ "step": 3693
+ },
+ {
+ "epoch": 51.30742358078603,
+ "grad_norm": 0.01802559196949005,
+ "learning_rate": 0.0006,
+ "loss": 4.746562480926514,
+ "step": 3694
+ },
+ {
+ "epoch": 51.32139737991266,
+ "grad_norm": 0.02053048275411129,
+ "learning_rate": 0.0006,
+ "loss": 4.69932746887207,
+ "step": 3695
+ },
+ {
+ "epoch": 51.3353711790393,
+ "grad_norm": 0.02661825157701969,
+ "learning_rate": 0.0006,
+ "loss": 4.68748664855957,
+ "step": 3696
+ },
+ {
+ "epoch": 51.34934497816594,
+ "grad_norm": 0.028269147500395775,
+ "learning_rate": 0.0006,
+ "loss": 4.731565475463867,
+ "step": 3697
+ },
+ {
+ "epoch": 51.36331877729258,
+ "grad_norm": 0.02650964818894863,
+ "learning_rate": 0.0006,
+ "loss": 4.733857154846191,
+ "step": 3698
+ },
+ {
+ "epoch": 51.377292576419215,
+ "grad_norm": 0.023696066811680794,
+ "learning_rate": 0.0006,
+ "loss": 4.812362194061279,
+ "step": 3699
+ },
+ {
+ "epoch": 51.391266375545854,
+ "grad_norm": 0.02365880273282528,
+ "learning_rate": 0.0006,
+ "loss": 4.733426094055176,
+ "step": 3700
+ },
+ {
+ "epoch": 51.40524017467249,
+ "grad_norm": 0.026249447837471962,
+ "learning_rate": 0.0006,
+ "loss": 4.715622901916504,
+ "step": 3701
+ },
+ {
+ "epoch": 51.419213973799124,
+ "grad_norm": 0.03301899880170822,
+ "learning_rate": 0.0006,
+ "loss": 4.7223663330078125,
+ "step": 3702
+ },
+ {
+ "epoch": 51.43318777292576,
+ "grad_norm": 0.044686660170555115,
+ "learning_rate": 0.0006,
+ "loss": 4.8121161460876465,
+ "step": 3703
+ },
+ {
+ "epoch": 51.4471615720524,
+ "grad_norm": 0.04733522981405258,
+ "learning_rate": 0.0006,
+ "loss": 4.841394424438477,
+ "step": 3704
+ },
+ {
+ "epoch": 51.46113537117904,
+ "grad_norm": 0.0395209975540638,
+ "learning_rate": 0.0006,
+ "loss": 4.773837089538574,
+ "step": 3705
+ },
+ {
+ "epoch": 51.47510917030568,
+ "grad_norm": 0.032790958881378174,
+ "learning_rate": 0.0006,
+ "loss": 4.833066463470459,
+ "step": 3706
+ },
+ {
+ "epoch": 51.48908296943232,
+ "grad_norm": 0.030578427016735077,
+ "learning_rate": 0.0006,
+ "loss": 4.685190200805664,
+ "step": 3707
+ },
+ {
+ "epoch": 51.50305676855895,
+ "grad_norm": 0.031619083136320114,
+ "learning_rate": 0.0006,
+ "loss": 4.740378379821777,
+ "step": 3708
+ },
+ {
+ "epoch": 51.51703056768559,
+ "grad_norm": 0.02385825850069523,
+ "learning_rate": 0.0006,
+ "loss": 4.812911510467529,
+ "step": 3709
+ },
+ {
+ "epoch": 51.531004366812226,
+ "grad_norm": 0.023523425683379173,
+ "learning_rate": 0.0006,
+ "loss": 4.669580459594727,
+ "step": 3710
+ },
+ {
+ "epoch": 51.544978165938865,
+ "grad_norm": 0.021316280588507652,
+ "learning_rate": 0.0006,
+ "loss": 4.76888370513916,
+ "step": 3711
+ },
+ {
+ "epoch": 51.5589519650655,
+ "grad_norm": 0.024663880467414856,
+ "learning_rate": 0.0006,
+ "loss": 4.702592372894287,
+ "step": 3712
+ },
+ {
+ "epoch": 51.57292576419214,
+ "grad_norm": 0.024173280224204063,
+ "learning_rate": 0.0006,
+ "loss": 4.710604667663574,
+ "step": 3713
+ },
+ {
+ "epoch": 51.58689956331878,
+ "grad_norm": 0.02754114381968975,
+ "learning_rate": 0.0006,
+ "loss": 4.746356964111328,
+ "step": 3714
+ },
+ {
+ "epoch": 51.60087336244541,
+ "grad_norm": 0.027903622016310692,
+ "learning_rate": 0.0006,
+ "loss": 4.7546491622924805,
+ "step": 3715
+ },
+ {
+ "epoch": 51.61484716157205,
+ "grad_norm": 0.024871617555618286,
+ "learning_rate": 0.0006,
+ "loss": 4.605746269226074,
+ "step": 3716
+ },
+ {
+ "epoch": 51.62882096069869,
+ "grad_norm": 0.021239031106233597,
+ "learning_rate": 0.0006,
+ "loss": 4.6942458152771,
+ "step": 3717
+ },
+ {
+ "epoch": 51.64279475982533,
+ "grad_norm": 0.01925622671842575,
+ "learning_rate": 0.0006,
+ "loss": 4.593558311462402,
+ "step": 3718
+ },
+ {
+ "epoch": 51.65676855895197,
+ "grad_norm": 0.01956283673644066,
+ "learning_rate": 0.0006,
+ "loss": 4.631172180175781,
+ "step": 3719
+ },
+ {
+ "epoch": 51.670742358078606,
+ "grad_norm": 0.02269871160387993,
+ "learning_rate": 0.0006,
+ "loss": 4.729236602783203,
+ "step": 3720
+ },
+ {
+ "epoch": 51.68471615720524,
+ "grad_norm": 0.02253916673362255,
+ "learning_rate": 0.0006,
+ "loss": 4.7191033363342285,
+ "step": 3721
+ },
+ {
+ "epoch": 51.698689956331876,
+ "grad_norm": 0.026231050491333008,
+ "learning_rate": 0.0006,
+ "loss": 4.811850547790527,
+ "step": 3722
+ },
+ {
+ "epoch": 51.712663755458514,
+ "grad_norm": 0.02558821812272072,
+ "learning_rate": 0.0006,
+ "loss": 4.681466102600098,
+ "step": 3723
+ },
+ {
+ "epoch": 51.72663755458515,
+ "grad_norm": 0.02124209702014923,
+ "learning_rate": 0.0006,
+ "loss": 4.693664073944092,
+ "step": 3724
+ },
+ {
+ "epoch": 51.74061135371179,
+ "grad_norm": 0.016638319939374924,
+ "learning_rate": 0.0006,
+ "loss": 4.771440029144287,
+ "step": 3725
+ },
+ {
+ "epoch": 51.75458515283843,
+ "grad_norm": 0.018048347905278206,
+ "learning_rate": 0.0006,
+ "loss": 4.643252372741699,
+ "step": 3726
+ },
+ {
+ "epoch": 51.76855895196506,
+ "grad_norm": 0.019904717803001404,
+ "learning_rate": 0.0006,
+ "loss": 4.66530704498291,
+ "step": 3727
+ },
+ {
+ "epoch": 51.7825327510917,
+ "grad_norm": 0.0194614976644516,
+ "learning_rate": 0.0006,
+ "loss": 4.711943626403809,
+ "step": 3728
+ },
+ {
+ "epoch": 51.79650655021834,
+ "grad_norm": 0.019175555557012558,
+ "learning_rate": 0.0006,
+ "loss": 4.721931457519531,
+ "step": 3729
+ },
+ {
+ "epoch": 51.81048034934498,
+ "grad_norm": 0.02093447372317314,
+ "learning_rate": 0.0006,
+ "loss": 4.633059024810791,
+ "step": 3730
+ },
+ {
+ "epoch": 51.82445414847162,
+ "grad_norm": 0.01930283196270466,
+ "learning_rate": 0.0006,
+ "loss": 4.659373760223389,
+ "step": 3731
+ },
+ {
+ "epoch": 51.838427947598255,
+ "grad_norm": 0.015350676141679287,
+ "learning_rate": 0.0006,
+ "loss": 4.763708114624023,
+ "step": 3732
+ },
+ {
+ "epoch": 51.852401746724894,
+ "grad_norm": 0.014139448292553425,
+ "learning_rate": 0.0006,
+ "loss": 4.753891468048096,
+ "step": 3733
+ },
+ {
+ "epoch": 51.866375545851525,
+ "grad_norm": 0.016215872019529343,
+ "learning_rate": 0.0006,
+ "loss": 4.658967018127441,
+ "step": 3734
+ },
+ {
+ "epoch": 51.880349344978164,
+ "grad_norm": 0.017328619956970215,
+ "learning_rate": 0.0006,
+ "loss": 4.696579933166504,
+ "step": 3735
+ },
+ {
+ "epoch": 51.8943231441048,
+ "grad_norm": 0.018103396520018578,
+ "learning_rate": 0.0006,
+ "loss": 4.627464771270752,
+ "step": 3736
+ },
+ {
+ "epoch": 51.90829694323144,
+ "grad_norm": 0.01715235412120819,
+ "learning_rate": 0.0006,
+ "loss": 4.7368011474609375,
+ "step": 3737
+ },
+ {
+ "epoch": 51.92227074235808,
+ "grad_norm": 0.014790094457566738,
+ "learning_rate": 0.0006,
+ "loss": 4.629441738128662,
+ "step": 3738
+ },
+ {
+ "epoch": 51.93624454148472,
+ "grad_norm": 0.017114296555519104,
+ "learning_rate": 0.0006,
+ "loss": 4.761256694793701,
+ "step": 3739
+ },
+ {
+ "epoch": 51.95021834061135,
+ "grad_norm": 0.019009845331311226,
+ "learning_rate": 0.0006,
+ "loss": 4.789155006408691,
+ "step": 3740
+ },
+ {
+ "epoch": 51.96419213973799,
+ "grad_norm": 0.019696692004799843,
+ "learning_rate": 0.0006,
+ "loss": 4.721269607543945,
+ "step": 3741
+ },
+ {
+ "epoch": 51.97816593886463,
+ "grad_norm": 0.020542822778224945,
+ "learning_rate": 0.0006,
+ "loss": 4.687038898468018,
+ "step": 3742
+ },
+ {
+ "epoch": 51.992139737991266,
+ "grad_norm": 0.021267160773277283,
+ "learning_rate": 0.0006,
+ "loss": 4.747860908508301,
+ "step": 3743
+ },
+ {
+ "epoch": 52.0,
+ "grad_norm": 0.025049567222595215,
+ "learning_rate": 0.0006,
+ "loss": 4.8473100662231445,
+ "step": 3744
+ },
+ {
+ "epoch": 52.0,
+ "eval_loss": 4.943043231964111,
+ "eval_runtime": 56.4929,
+ "eval_samples_per_second": 43.227,
+ "eval_steps_per_second": 1.363,
+ "step": 3744
+ },
+ {
+ "epoch": 52.01397379912664,
+ "grad_norm": 0.03054986707866192,
+ "learning_rate": 0.0006,
+ "loss": 4.856931686401367,
+ "step": 3745
+ },
+ {
+ "epoch": 52.02794759825328,
+ "grad_norm": 0.03838375583291054,
+ "learning_rate": 0.0006,
+ "loss": 4.699807167053223,
+ "step": 3746
+ },
+ {
+ "epoch": 52.041921397379916,
+ "grad_norm": 0.03383488208055496,
+ "learning_rate": 0.0006,
+ "loss": 4.6626787185668945,
+ "step": 3747
+ },
+ {
+ "epoch": 52.05589519650655,
+ "grad_norm": 0.0201561376452446,
+ "learning_rate": 0.0006,
+ "loss": 4.621831893920898,
+ "step": 3748
+ },
+ {
+ "epoch": 52.069868995633186,
+ "grad_norm": 0.028749438002705574,
+ "learning_rate": 0.0006,
+ "loss": 4.682540416717529,
+ "step": 3749
+ },
+ {
+ "epoch": 52.083842794759825,
+ "grad_norm": 0.02681174874305725,
+ "learning_rate": 0.0006,
+ "loss": 4.629997253417969,
+ "step": 3750
+ },
+ {
+ "epoch": 52.09781659388646,
+ "grad_norm": 0.024063318967819214,
+ "learning_rate": 0.0006,
+ "loss": 4.573021411895752,
+ "step": 3751
+ },
+ {
+ "epoch": 52.1117903930131,
+ "grad_norm": 0.02324330434203148,
+ "learning_rate": 0.0006,
+ "loss": 4.739465236663818,
+ "step": 3752
+ },
+ {
+ "epoch": 52.12576419213974,
+ "grad_norm": 0.0225426834076643,
+ "learning_rate": 0.0006,
+ "loss": 4.7335333824157715,
+ "step": 3753
+ },
+ {
+ "epoch": 52.13973799126637,
+ "grad_norm": 0.019593428820371628,
+ "learning_rate": 0.0006,
+ "loss": 4.704958438873291,
+ "step": 3754
+ },
+ {
+ "epoch": 52.15371179039301,
+ "grad_norm": 0.020777082070708275,
+ "learning_rate": 0.0006,
+ "loss": 4.638696670532227,
+ "step": 3755
+ },
+ {
+ "epoch": 52.16768558951965,
+ "grad_norm": 0.021525604650378227,
+ "learning_rate": 0.0006,
+ "loss": 4.643706798553467,
+ "step": 3756
+ },
+ {
+ "epoch": 52.18165938864629,
+ "grad_norm": 0.018943654373288155,
+ "learning_rate": 0.0006,
+ "loss": 4.5173211097717285,
+ "step": 3757
+ },
+ {
+ "epoch": 52.19563318777293,
+ "grad_norm": 0.01856580562889576,
+ "learning_rate": 0.0006,
+ "loss": 4.678788661956787,
+ "step": 3758
+ },
+ {
+ "epoch": 52.209606986899566,
+ "grad_norm": 0.019417259842157364,
+ "learning_rate": 0.0006,
+ "loss": 4.587244987487793,
+ "step": 3759
+ },
+ {
+ "epoch": 52.223580786026204,
+ "grad_norm": 0.01911911740899086,
+ "learning_rate": 0.0006,
+ "loss": 4.688503265380859,
+ "step": 3760
+ },
+ {
+ "epoch": 52.237554585152836,
+ "grad_norm": 0.021017879247665405,
+ "learning_rate": 0.0006,
+ "loss": 4.65829610824585,
+ "step": 3761
+ },
+ {
+ "epoch": 52.251528384279474,
+ "grad_norm": 0.02048875205218792,
+ "learning_rate": 0.0006,
+ "loss": 4.624055862426758,
+ "step": 3762
+ },
+ {
+ "epoch": 52.26550218340611,
+ "grad_norm": 0.01932772807776928,
+ "learning_rate": 0.0006,
+ "loss": 4.556562900543213,
+ "step": 3763
+ },
+ {
+ "epoch": 52.27947598253275,
+ "grad_norm": 0.01871614344418049,
+ "learning_rate": 0.0006,
+ "loss": 4.581201076507568,
+ "step": 3764
+ },
+ {
+ "epoch": 52.29344978165939,
+ "grad_norm": 0.017253106459975243,
+ "learning_rate": 0.0006,
+ "loss": 4.605219841003418,
+ "step": 3765
+ },
+ {
+ "epoch": 52.30742358078603,
+ "grad_norm": 0.01831238530576229,
+ "learning_rate": 0.0006,
+ "loss": 4.759900093078613,
+ "step": 3766
+ },
+ {
+ "epoch": 52.32139737991266,
+ "grad_norm": 0.017946207895874977,
+ "learning_rate": 0.0006,
+ "loss": 4.5598344802856445,
+ "step": 3767
+ },
+ {
+ "epoch": 52.3353711790393,
+ "grad_norm": 0.01680106669664383,
+ "learning_rate": 0.0006,
+ "loss": 4.605034828186035,
+ "step": 3768
+ },
+ {
+ "epoch": 52.34934497816594,
+ "grad_norm": 0.021482540294528008,
+ "learning_rate": 0.0006,
+ "loss": 4.620411396026611,
+ "step": 3769
+ },
+ {
+ "epoch": 52.36331877729258,
+ "grad_norm": 0.02210252732038498,
+ "learning_rate": 0.0006,
+ "loss": 4.667577743530273,
+ "step": 3770
+ },
+ {
+ "epoch": 52.377292576419215,
+ "grad_norm": 0.02237485907971859,
+ "learning_rate": 0.0006,
+ "loss": 4.628812789916992,
+ "step": 3771
+ },
+ {
+ "epoch": 52.391266375545854,
+ "grad_norm": 0.02196541801095009,
+ "learning_rate": 0.0006,
+ "loss": 4.572988510131836,
+ "step": 3772
+ },
+ {
+ "epoch": 52.40524017467249,
+ "grad_norm": 0.01930229924619198,
+ "learning_rate": 0.0006,
+ "loss": 4.588339805603027,
+ "step": 3773
+ },
+ {
+ "epoch": 52.419213973799124,
+ "grad_norm": 0.02222434990108013,
+ "learning_rate": 0.0006,
+ "loss": 4.766482353210449,
+ "step": 3774
+ },
+ {
+ "epoch": 52.43318777292576,
+ "grad_norm": 0.028806306421756744,
+ "learning_rate": 0.0006,
+ "loss": 4.617892742156982,
+ "step": 3775
+ },
+ {
+ "epoch": 52.4471615720524,
+ "grad_norm": 0.03306042402982712,
+ "learning_rate": 0.0006,
+ "loss": 4.665334701538086,
+ "step": 3776
+ },
+ {
+ "epoch": 52.46113537117904,
+ "grad_norm": 0.0358956903219223,
+ "learning_rate": 0.0006,
+ "loss": 4.566668510437012,
+ "step": 3777
+ },
+ {
+ "epoch": 52.47510917030568,
+ "grad_norm": 0.030406497418880463,
+ "learning_rate": 0.0006,
+ "loss": 4.513612270355225,
+ "step": 3778
+ },
+ {
+ "epoch": 52.48908296943232,
+ "grad_norm": 0.030295712873339653,
+ "learning_rate": 0.0006,
+ "loss": 4.640189170837402,
+ "step": 3779
+ },
+ {
+ "epoch": 52.50305676855895,
+ "grad_norm": 0.024981003254652023,
+ "learning_rate": 0.0006,
+ "loss": 4.57558012008667,
+ "step": 3780
+ },
+ {
+ "epoch": 52.51703056768559,
+ "grad_norm": 0.024990713223814964,
+ "learning_rate": 0.0006,
+ "loss": 4.708589553833008,
+ "step": 3781
+ },
+ {
+ "epoch": 52.531004366812226,
+ "grad_norm": 0.025034775957465172,
+ "learning_rate": 0.0006,
+ "loss": 4.70468282699585,
+ "step": 3782
+ },
+ {
+ "epoch": 52.544978165938865,
+ "grad_norm": 0.028243759647011757,
+ "learning_rate": 0.0006,
+ "loss": 4.712060451507568,
+ "step": 3783
+ },
+ {
+ "epoch": 52.5589519650655,
+ "grad_norm": 0.026984544470906258,
+ "learning_rate": 0.0006,
+ "loss": 4.505004405975342,
+ "step": 3784
+ },
+ {
+ "epoch": 52.57292576419214,
+ "grad_norm": 0.050843529403209686,
+ "learning_rate": 0.0006,
+ "loss": 4.593271255493164,
+ "step": 3785
+ },
+ {
+ "epoch": 52.58689956331878,
+ "grad_norm": 0.25422802567481995,
+ "learning_rate": 0.0006,
+ "loss": 4.7178802490234375,
+ "step": 3786
+ },
+ {
+ "epoch": 52.60087336244541,
+ "grad_norm": 0.6078406572341919,
+ "learning_rate": 0.0006,
+ "loss": 6.144372940063477,
+ "step": 3787
+ },
+ {
+ "epoch": 52.61484716157205,
+ "grad_norm": 0.25933101773262024,
+ "learning_rate": 0.0006,
+ "loss": 5.971083641052246,
+ "step": 3788
+ },
+ {
+ "epoch": 52.62882096069869,
+ "grad_norm": 0.294657826423645,
+ "learning_rate": 0.0006,
+ "loss": 6.204090118408203,
+ "step": 3789
+ },
+ {
+ "epoch": 52.64279475982533,
+ "grad_norm": 0.23641522228717804,
+ "learning_rate": 0.0006,
+ "loss": 6.869574546813965,
+ "step": 3790
+ },
+ {
+ "epoch": 52.65676855895197,
+ "grad_norm": 0.23717311024665833,
+ "learning_rate": 0.0006,
+ "loss": 6.739034175872803,
+ "step": 3791
+ },
+ {
+ "epoch": 52.670742358078606,
+ "grad_norm": 0.10300362855195999,
+ "learning_rate": 0.0006,
+ "loss": 6.49191951751709,
+ "step": 3792
+ },
+ {
+ "epoch": 52.68471615720524,
+ "grad_norm": 0.1484968513250351,
+ "learning_rate": 0.0006,
+ "loss": 6.6021270751953125,
+ "step": 3793
+ },
+ {
+ "epoch": 52.698689956331876,
+ "grad_norm": 0.08052115142345428,
+ "learning_rate": 0.0006,
+ "loss": 6.432796478271484,
+ "step": 3794
+ },
+ {
+ "epoch": 52.712663755458514,
+ "grad_norm": 0.06869253516197205,
+ "learning_rate": 0.0006,
+ "loss": 6.212784767150879,
+ "step": 3795
+ },
+ {
+ "epoch": 52.72663755458515,
+ "grad_norm": 0.05589582398533821,
+ "learning_rate": 0.0006,
+ "loss": 6.168288707733154,
+ "step": 3796
+ },
+ {
+ "epoch": 52.74061135371179,
+ "grad_norm": 0.04840940609574318,
+ "learning_rate": 0.0006,
+ "loss": 6.163174629211426,
+ "step": 3797
+ },
+ {
+ "epoch": 52.75458515283843,
+ "grad_norm": 0.04578601196408272,
+ "learning_rate": 0.0006,
+ "loss": 6.038390636444092,
+ "step": 3798
+ },
+ {
+ "epoch": 52.76855895196506,
+ "grad_norm": 0.04102053865790367,
+ "learning_rate": 0.0006,
+ "loss": 5.929539680480957,
+ "step": 3799
+ },
+ {
+ "epoch": 52.7825327510917,
+ "grad_norm": 0.038398485630750656,
+ "learning_rate": 0.0006,
+ "loss": 6.007378101348877,
+ "step": 3800
+ },
+ {
+ "epoch": 52.79650655021834,
+ "grad_norm": 0.03466642275452614,
+ "learning_rate": 0.0006,
+ "loss": 5.767291069030762,
+ "step": 3801
+ },
+ {
+ "epoch": 52.81048034934498,
+ "grad_norm": 0.03328905999660492,
+ "learning_rate": 0.0006,
+ "loss": 5.785032272338867,
+ "step": 3802
+ },
+ {
+ "epoch": 52.82445414847162,
+ "grad_norm": 0.03186681121587753,
+ "learning_rate": 0.0006,
+ "loss": 5.6942057609558105,
+ "step": 3803
+ },
+ {
+ "epoch": 52.838427947598255,
+ "grad_norm": 0.030553551390767097,
+ "learning_rate": 0.0006,
+ "loss": 5.680774688720703,
+ "step": 3804
+ },
+ {
+ "epoch": 52.852401746724894,
+ "grad_norm": 0.02578839473426342,
+ "learning_rate": 0.0006,
+ "loss": 5.703461647033691,
+ "step": 3805
+ },
+ {
+ "epoch": 52.866375545851525,
+ "grad_norm": 0.026261312887072563,
+ "learning_rate": 0.0006,
+ "loss": 5.642549514770508,
+ "step": 3806
+ },
+ {
+ "epoch": 52.880349344978164,
+ "grad_norm": 0.02782430872321129,
+ "learning_rate": 0.0006,
+ "loss": 5.563597679138184,
+ "step": 3807
+ },
+ {
+ "epoch": 52.8943231441048,
+ "grad_norm": 0.024467799812555313,
+ "learning_rate": 0.0006,
+ "loss": 5.583395957946777,
+ "step": 3808
+ },
+ {
+ "epoch": 52.90829694323144,
+ "grad_norm": 0.022602897137403488,
+ "learning_rate": 0.0006,
+ "loss": 5.50742769241333,
+ "step": 3809
+ },
+ {
+ "epoch": 52.92227074235808,
+ "grad_norm": 0.024292029440402985,
+ "learning_rate": 0.0006,
+ "loss": 5.367804527282715,
+ "step": 3810
+ },
+ {
+ "epoch": 52.93624454148472,
+ "grad_norm": 0.021062295883893967,
+ "learning_rate": 0.0006,
+ "loss": 5.423705101013184,
+ "step": 3811
+ },
+ {
+ "epoch": 52.95021834061135,
+ "grad_norm": 0.022226469591259956,
+ "learning_rate": 0.0006,
+ "loss": 5.344039440155029,
+ "step": 3812
+ },
+ {
+ "epoch": 52.96419213973799,
+ "grad_norm": 0.023691752925515175,
+ "learning_rate": 0.0006,
+ "loss": 5.2581892013549805,
+ "step": 3813
+ },
+ {
+ "epoch": 52.97816593886463,
+ "grad_norm": 0.02181999757885933,
+ "learning_rate": 0.0006,
+ "loss": 5.2710113525390625,
+ "step": 3814
+ },
+ {
+ "epoch": 52.992139737991266,
+ "grad_norm": 0.025295857340097427,
+ "learning_rate": 0.0006,
+ "loss": 5.278355598449707,
+ "step": 3815
+ },
+ {
+ "epoch": 53.0,
+ "grad_norm": 0.02216828241944313,
+ "learning_rate": 0.0006,
+ "loss": 5.127945899963379,
+ "step": 3816
+ },
+ {
+ "epoch": 53.0,
+ "eval_loss": 5.3411736488342285,
+ "eval_runtime": 56.2837,
+ "eval_samples_per_second": 43.387,
+ "eval_steps_per_second": 1.368,
+ "step": 3816
+ },
+ {
+ "epoch": 53.01397379912664,
+ "grad_norm": 0.022496026009321213,
+ "learning_rate": 0.0006,
+ "loss": 5.19655704498291,
+ "step": 3817
+ },
+ {
+ "epoch": 53.02794759825328,
+ "grad_norm": 0.03201168775558472,
+ "learning_rate": 0.0006,
+ "loss": 5.094621181488037,
+ "step": 3818
+ },
+ {
+ "epoch": 53.041921397379916,
+ "grad_norm": 0.02683178335428238,
+ "learning_rate": 0.0006,
+ "loss": 5.19899845123291,
+ "step": 3819
+ },
+ {
+ "epoch": 53.05589519650655,
+ "grad_norm": 0.022659169510006905,
+ "learning_rate": 0.0006,
+ "loss": 5.184103012084961,
+ "step": 3820
+ },
+ {
+ "epoch": 53.069868995633186,
+ "grad_norm": 0.024784443899989128,
+ "learning_rate": 0.0006,
+ "loss": 5.124286651611328,
+ "step": 3821
+ },
+ {
+ "epoch": 53.083842794759825,
+ "grad_norm": 0.028710370883345604,
+ "learning_rate": 0.0006,
+ "loss": 5.130426406860352,
+ "step": 3822
+ },
+ {
+ "epoch": 53.09781659388646,
+ "grad_norm": 0.02366088517010212,
+ "learning_rate": 0.0006,
+ "loss": 5.1453399658203125,
+ "step": 3823
+ },
+ {
+ "epoch": 53.1117903930131,
+ "grad_norm": 0.01813938096165657,
+ "learning_rate": 0.0006,
+ "loss": 5.05271053314209,
+ "step": 3824
+ },
+ {
+ "epoch": 53.12576419213974,
+ "grad_norm": 0.02051435597240925,
+ "learning_rate": 0.0006,
+ "loss": 5.023038864135742,
+ "step": 3825
+ },
+ {
+ "epoch": 53.13973799126637,
+ "grad_norm": 0.02185754105448723,
+ "learning_rate": 0.0006,
+ "loss": 4.8950018882751465,
+ "step": 3826
+ },
+ {
+ "epoch": 53.15371179039301,
+ "grad_norm": 0.017223743721842766,
+ "learning_rate": 0.0006,
+ "loss": 5.078621864318848,
+ "step": 3827
+ },
+ {
+ "epoch": 53.16768558951965,
+ "grad_norm": 0.01832217164337635,
+ "learning_rate": 0.0006,
+ "loss": 4.921341419219971,
+ "step": 3828
+ },
+ {
+ "epoch": 53.18165938864629,
+ "grad_norm": 0.02114005573093891,
+ "learning_rate": 0.0006,
+ "loss": 4.997865676879883,
+ "step": 3829
+ },
+ {
+ "epoch": 53.19563318777293,
+ "grad_norm": 0.021426431834697723,
+ "learning_rate": 0.0006,
+ "loss": 4.877785682678223,
+ "step": 3830
+ },
+ {
+ "epoch": 53.209606986899566,
+ "grad_norm": 0.021327560767531395,
+ "learning_rate": 0.0006,
+ "loss": 4.847512245178223,
+ "step": 3831
+ },
+ {
+ "epoch": 53.223580786026204,
+ "grad_norm": 0.027825530618429184,
+ "learning_rate": 0.0006,
+ "loss": 5.004823207855225,
+ "step": 3832
+ },
+ {
+ "epoch": 53.237554585152836,
+ "grad_norm": 0.03756481409072876,
+ "learning_rate": 0.0006,
+ "loss": 4.817290306091309,
+ "step": 3833
+ },
+ {
+ "epoch": 53.251528384279474,
+ "grad_norm": 0.04933721572160721,
+ "learning_rate": 0.0006,
+ "loss": 4.863317489624023,
+ "step": 3834
+ },
+ {
+ "epoch": 53.26550218340611,
+ "grad_norm": 0.04588694870471954,
+ "learning_rate": 0.0006,
+ "loss": 4.864742755889893,
+ "step": 3835
+ },
+ {
+ "epoch": 53.27947598253275,
+ "grad_norm": 0.02728264033794403,
+ "learning_rate": 0.0006,
+ "loss": 4.9068756103515625,
+ "step": 3836
+ },
+ {
+ "epoch": 53.29344978165939,
+ "grad_norm": 0.036240726709365845,
+ "learning_rate": 0.0006,
+ "loss": 4.902045249938965,
+ "step": 3837
+ },
+ {
+ "epoch": 53.30742358078603,
+ "grad_norm": 0.02703195810317993,
+ "learning_rate": 0.0006,
+ "loss": 4.908980369567871,
+ "step": 3838
+ },
+ {
+ "epoch": 53.32139737991266,
+ "grad_norm": 0.024894049391150475,
+ "learning_rate": 0.0006,
+ "loss": 4.794848442077637,
+ "step": 3839
+ },
+ {
+ "epoch": 53.3353711790393,
+ "grad_norm": 0.02389993891119957,
+ "learning_rate": 0.0006,
+ "loss": 4.8234734535217285,
+ "step": 3840
+ },
+ {
+ "epoch": 53.34934497816594,
+ "grad_norm": 0.02486400306224823,
+ "learning_rate": 0.0006,
+ "loss": 4.851463794708252,
+ "step": 3841
+ },
+ {
+ "epoch": 53.36331877729258,
+ "grad_norm": 0.02613813802599907,
+ "learning_rate": 0.0006,
+ "loss": 4.860444068908691,
+ "step": 3842
+ },
+ {
+ "epoch": 53.377292576419215,
+ "grad_norm": 0.0229744054377079,
+ "learning_rate": 0.0006,
+ "loss": 4.709606170654297,
+ "step": 3843
+ },
+ {
+ "epoch": 53.391266375545854,
+ "grad_norm": 0.02123042568564415,
+ "learning_rate": 0.0006,
+ "loss": 4.654613494873047,
+ "step": 3844
+ },
+ {
+ "epoch": 53.40524017467249,
+ "grad_norm": 0.017352338880300522,
+ "learning_rate": 0.0006,
+ "loss": 4.769761085510254,
+ "step": 3845
+ },
+ {
+ "epoch": 53.419213973799124,
+ "grad_norm": 0.019808724522590637,
+ "learning_rate": 0.0006,
+ "loss": 4.98514461517334,
+ "step": 3846
+ },
+ {
+ "epoch": 53.43318777292576,
+ "grad_norm": 0.018776152282953262,
+ "learning_rate": 0.0006,
+ "loss": 4.749990463256836,
+ "step": 3847
+ },
+ {
+ "epoch": 53.4471615720524,
+ "grad_norm": 0.02017521858215332,
+ "learning_rate": 0.0006,
+ "loss": 4.757381439208984,
+ "step": 3848
+ },
+ {
+ "epoch": 53.46113537117904,
+ "grad_norm": 0.019509416073560715,
+ "learning_rate": 0.0006,
+ "loss": 4.8111572265625,
+ "step": 3849
+ },
+ {
+ "epoch": 53.47510917030568,
+ "grad_norm": 0.015543580986559391,
+ "learning_rate": 0.0006,
+ "loss": 4.7501678466796875,
+ "step": 3850
+ },
+ {
+ "epoch": 53.48908296943232,
+ "grad_norm": 0.017941651865839958,
+ "learning_rate": 0.0006,
+ "loss": 4.701838493347168,
+ "step": 3851
+ },
+ {
+ "epoch": 53.50305676855895,
+ "grad_norm": 0.01678212359547615,
+ "learning_rate": 0.0006,
+ "loss": 4.787798881530762,
+ "step": 3852
+ },
+ {
+ "epoch": 53.51703056768559,
+ "grad_norm": 0.019050654023885727,
+ "learning_rate": 0.0006,
+ "loss": 4.754859924316406,
+ "step": 3853
+ },
+ {
+ "epoch": 53.531004366812226,
+ "grad_norm": 0.017406078055500984,
+ "learning_rate": 0.0006,
+ "loss": 4.752772808074951,
+ "step": 3854
+ },
+ {
+ "epoch": 53.544978165938865,
+ "grad_norm": 0.01773344725370407,
+ "learning_rate": 0.0006,
+ "loss": 4.651812553405762,
+ "step": 3855
+ },
+ {
+ "epoch": 53.5589519650655,
+ "grad_norm": 0.017221346497535706,
+ "learning_rate": 0.0006,
+ "loss": 4.792244911193848,
+ "step": 3856
+ },
+ {
+ "epoch": 53.57292576419214,
+ "grad_norm": 0.01722758449614048,
+ "learning_rate": 0.0006,
+ "loss": 4.7902960777282715,
+ "step": 3857
+ },
+ {
+ "epoch": 53.58689956331878,
+ "grad_norm": 0.020445192232728004,
+ "learning_rate": 0.0006,
+ "loss": 4.67392110824585,
+ "step": 3858
+ },
+ {
+ "epoch": 53.60087336244541,
+ "grad_norm": 0.02267114631831646,
+ "learning_rate": 0.0006,
+ "loss": 4.796089172363281,
+ "step": 3859
+ },
+ {
+ "epoch": 53.61484716157205,
+ "grad_norm": 0.027129942551255226,
+ "learning_rate": 0.0006,
+ "loss": 4.774303913116455,
+ "step": 3860
+ },
+ {
+ "epoch": 53.62882096069869,
+ "grad_norm": 0.025454839691519737,
+ "learning_rate": 0.0006,
+ "loss": 4.708813190460205,
+ "step": 3861
+ },
+ {
+ "epoch": 53.64279475982533,
+ "grad_norm": 0.019248757511377335,
+ "learning_rate": 0.0006,
+ "loss": 4.760467529296875,
+ "step": 3862
+ },
+ {
+ "epoch": 53.65676855895197,
+ "grad_norm": 0.01791176199913025,
+ "learning_rate": 0.0006,
+ "loss": 4.789203643798828,
+ "step": 3863
+ },
+ {
+ "epoch": 53.670742358078606,
+ "grad_norm": 0.016306612640619278,
+ "learning_rate": 0.0006,
+ "loss": 4.732605934143066,
+ "step": 3864
+ },
+ {
+ "epoch": 53.68471615720524,
+ "grad_norm": 0.017836976796388626,
+ "learning_rate": 0.0006,
+ "loss": 4.691871643066406,
+ "step": 3865
+ },
+ {
+ "epoch": 53.698689956331876,
+ "grad_norm": 0.015102861449122429,
+ "learning_rate": 0.0006,
+ "loss": 4.724156856536865,
+ "step": 3866
+ },
+ {
+ "epoch": 53.712663755458514,
+ "grad_norm": 0.017458142712712288,
+ "learning_rate": 0.0006,
+ "loss": 4.73524808883667,
+ "step": 3867
+ },
+ {
+ "epoch": 53.72663755458515,
+ "grad_norm": 0.01591932773590088,
+ "learning_rate": 0.0006,
+ "loss": 4.705888748168945,
+ "step": 3868
+ },
+ {
+ "epoch": 53.74061135371179,
+ "grad_norm": 0.01713593676686287,
+ "learning_rate": 0.0006,
+ "loss": 4.644516944885254,
+ "step": 3869
+ },
+ {
+ "epoch": 53.75458515283843,
+ "grad_norm": 0.01905885338783264,
+ "learning_rate": 0.0006,
+ "loss": 4.698416233062744,
+ "step": 3870
+ },
+ {
+ "epoch": 53.76855895196506,
+ "grad_norm": 0.01766437292098999,
+ "learning_rate": 0.0006,
+ "loss": 4.6796555519104,
+ "step": 3871
+ },
+ {
+ "epoch": 53.7825327510917,
+ "grad_norm": 0.01552243810147047,
+ "learning_rate": 0.0006,
+ "loss": 4.78138542175293,
+ "step": 3872
+ },
+ {
+ "epoch": 53.79650655021834,
+ "grad_norm": 0.014191331341862679,
+ "learning_rate": 0.0006,
+ "loss": 4.703588485717773,
+ "step": 3873
+ },
+ {
+ "epoch": 53.81048034934498,
+ "grad_norm": 0.015872279182076454,
+ "learning_rate": 0.0006,
+ "loss": 4.705611228942871,
+ "step": 3874
+ },
+ {
+ "epoch": 53.82445414847162,
+ "grad_norm": 0.01787560060620308,
+ "learning_rate": 0.0006,
+ "loss": 4.8227949142456055,
+ "step": 3875
+ },
+ {
+ "epoch": 53.838427947598255,
+ "grad_norm": 0.0158209390938282,
+ "learning_rate": 0.0006,
+ "loss": 4.645227432250977,
+ "step": 3876
+ },
+ {
+ "epoch": 53.852401746724894,
+ "grad_norm": 0.014362678863108158,
+ "learning_rate": 0.0006,
+ "loss": 4.7225422859191895,
+ "step": 3877
+ },
+ {
+ "epoch": 53.866375545851525,
+ "grad_norm": 0.014899161644279957,
+ "learning_rate": 0.0006,
+ "loss": 4.685914039611816,
+ "step": 3878
+ },
+ {
+ "epoch": 53.880349344978164,
+ "grad_norm": 0.01628505438566208,
+ "learning_rate": 0.0006,
+ "loss": 4.674065589904785,
+ "step": 3879
+ },
+ {
+ "epoch": 53.8943231441048,
+ "grad_norm": 0.01873331144452095,
+ "learning_rate": 0.0006,
+ "loss": 4.811744213104248,
+ "step": 3880
+ },
+ {
+ "epoch": 53.90829694323144,
+ "grad_norm": 0.02229069173336029,
+ "learning_rate": 0.0006,
+ "loss": 4.721514701843262,
+ "step": 3881
+ },
+ {
+ "epoch": 53.92227074235808,
+ "grad_norm": 0.028012916445732117,
+ "learning_rate": 0.0006,
+ "loss": 4.773603439331055,
+ "step": 3882
+ },
+ {
+ "epoch": 53.93624454148472,
+ "grad_norm": 0.036322783678770065,
+ "learning_rate": 0.0006,
+ "loss": 4.787134170532227,
+ "step": 3883
+ },
+ {
+ "epoch": 53.95021834061135,
+ "grad_norm": 0.048149142414331436,
+ "learning_rate": 0.0006,
+ "loss": 4.736196517944336,
+ "step": 3884
+ },
+ {
+ "epoch": 53.96419213973799,
+ "grad_norm": 0.051918916404247284,
+ "learning_rate": 0.0006,
+ "loss": 4.686728000640869,
+ "step": 3885
+ },
+ {
+ "epoch": 53.97816593886463,
+ "grad_norm": 0.0481342189013958,
+ "learning_rate": 0.0006,
+ "loss": 4.733730792999268,
+ "step": 3886
+ },
+ {
+ "epoch": 53.992139737991266,
+ "grad_norm": 0.045424818992614746,
+ "learning_rate": 0.0006,
+ "loss": 4.748723983764648,
+ "step": 3887
+ },
+ {
+ "epoch": 54.0,
+ "grad_norm": 0.04153316840529442,
+ "learning_rate": 0.0006,
+ "loss": 4.715220928192139,
+ "step": 3888
+ },
+ {
+ "epoch": 54.0,
+ "eval_loss": 4.964215278625488,
+ "eval_runtime": 56.7757,
+ "eval_samples_per_second": 43.011,
+ "eval_steps_per_second": 1.356,
+ "step": 3888
+ },
+ {
+ "epoch": 54.01397379912664,
+ "grad_norm": 0.03459750860929489,
+ "learning_rate": 0.0006,
+ "loss": 4.661557674407959,
+ "step": 3889
+ },
+ {
+ "epoch": 54.02794759825328,
+ "grad_norm": 0.025409208610653877,
+ "learning_rate": 0.0006,
+ "loss": 4.7065582275390625,
+ "step": 3890
+ },
+ {
+ "epoch": 54.041921397379916,
+ "grad_norm": 0.03202550485730171,
+ "learning_rate": 0.0006,
+ "loss": 4.751070976257324,
+ "step": 3891
+ },
+ {
+ "epoch": 54.05589519650655,
+ "grad_norm": 0.02968018501996994,
+ "learning_rate": 0.0006,
+ "loss": 4.635388374328613,
+ "step": 3892
+ },
+ {
+ "epoch": 54.069868995633186,
+ "grad_norm": 0.02757406234741211,
+ "learning_rate": 0.0006,
+ "loss": 4.706214427947998,
+ "step": 3893
+ },
+ {
+ "epoch": 54.083842794759825,
+ "grad_norm": 0.025223620235919952,
+ "learning_rate": 0.0006,
+ "loss": 4.703792572021484,
+ "step": 3894
+ },
+ {
+ "epoch": 54.09781659388646,
+ "grad_norm": 0.02431778982281685,
+ "learning_rate": 0.0006,
+ "loss": 4.658116340637207,
+ "step": 3895
+ },
+ {
+ "epoch": 54.1117903930131,
+ "grad_norm": 0.02792172320187092,
+ "learning_rate": 0.0006,
+ "loss": 4.638895034790039,
+ "step": 3896
+ },
+ {
+ "epoch": 54.12576419213974,
+ "grad_norm": 0.032253511250019073,
+ "learning_rate": 0.0006,
+ "loss": 4.693902492523193,
+ "step": 3897
+ },
+ {
+ "epoch": 54.13973799126637,
+ "grad_norm": 0.027044931426644325,
+ "learning_rate": 0.0006,
+ "loss": 4.575805187225342,
+ "step": 3898
+ },
+ {
+ "epoch": 54.15371179039301,
+ "grad_norm": 0.023270361125469208,
+ "learning_rate": 0.0006,
+ "loss": 4.67645263671875,
+ "step": 3899
+ },
+ {
+ "epoch": 54.16768558951965,
+ "grad_norm": 0.01929079182446003,
+ "learning_rate": 0.0006,
+ "loss": 4.68062686920166,
+ "step": 3900
+ },
+ {
+ "epoch": 54.18165938864629,
+ "grad_norm": 0.020343473181128502,
+ "learning_rate": 0.0006,
+ "loss": 4.708086967468262,
+ "step": 3901
+ },
+ {
+ "epoch": 54.19563318777293,
+ "grad_norm": 0.0211793202906847,
+ "learning_rate": 0.0006,
+ "loss": 4.677323341369629,
+ "step": 3902
+ },
+ {
+ "epoch": 54.209606986899566,
+ "grad_norm": 0.020341217517852783,
+ "learning_rate": 0.0006,
+ "loss": 4.760937690734863,
+ "step": 3903
+ },
+ {
+ "epoch": 54.223580786026204,
+ "grad_norm": 0.019421054050326347,
+ "learning_rate": 0.0006,
+ "loss": 4.508834362030029,
+ "step": 3904
+ },
+ {
+ "epoch": 54.237554585152836,
+ "grad_norm": 0.016701679676771164,
+ "learning_rate": 0.0006,
+ "loss": 4.664399147033691,
+ "step": 3905
+ },
+ {
+ "epoch": 54.251528384279474,
+ "grad_norm": 0.016561470925807953,
+ "learning_rate": 0.0006,
+ "loss": 4.742990493774414,
+ "step": 3906
+ },
+ {
+ "epoch": 54.26550218340611,
+ "grad_norm": 0.017446085810661316,
+ "learning_rate": 0.0006,
+ "loss": 4.63038444519043,
+ "step": 3907
+ },
+ {
+ "epoch": 54.27947598253275,
+ "grad_norm": 0.014340086840093136,
+ "learning_rate": 0.0006,
+ "loss": 4.688791275024414,
+ "step": 3908
+ },
+ {
+ "epoch": 54.29344978165939,
+ "grad_norm": 0.013580698519945145,
+ "learning_rate": 0.0006,
+ "loss": 4.625302314758301,
+ "step": 3909
+ },
+ {
+ "epoch": 54.30742358078603,
+ "grad_norm": 0.014931539073586464,
+ "learning_rate": 0.0006,
+ "loss": 4.6167497634887695,
+ "step": 3910
+ },
+ {
+ "epoch": 54.32139737991266,
+ "grad_norm": 0.013492150232195854,
+ "learning_rate": 0.0006,
+ "loss": 4.657521724700928,
+ "step": 3911
+ },
+ {
+ "epoch": 54.3353711790393,
+ "grad_norm": 0.015151413157582283,
+ "learning_rate": 0.0006,
+ "loss": 4.634521484375,
+ "step": 3912
+ },
+ {
+ "epoch": 54.34934497816594,
+ "grad_norm": 0.01302040833979845,
+ "learning_rate": 0.0006,
+ "loss": 4.718928337097168,
+ "step": 3913
+ },
+ {
+ "epoch": 54.36331877729258,
+ "grad_norm": 0.013433529995381832,
+ "learning_rate": 0.0006,
+ "loss": 4.584803581237793,
+ "step": 3914
+ },
+ {
+ "epoch": 54.377292576419215,
+ "grad_norm": 0.012987039051949978,
+ "learning_rate": 0.0006,
+ "loss": 4.6393585205078125,
+ "step": 3915
+ },
+ {
+ "epoch": 54.391266375545854,
+ "grad_norm": 0.013695075176656246,
+ "learning_rate": 0.0006,
+ "loss": 4.556313514709473,
+ "step": 3916
+ },
+ {
+ "epoch": 54.40524017467249,
+ "grad_norm": 0.013718255795538425,
+ "learning_rate": 0.0006,
+ "loss": 4.576223850250244,
+ "step": 3917
+ },
+ {
+ "epoch": 54.419213973799124,
+ "grad_norm": 0.013699112460017204,
+ "learning_rate": 0.0006,
+ "loss": 4.63889217376709,
+ "step": 3918
+ },
+ {
+ "epoch": 54.43318777292576,
+ "grad_norm": 0.013516264036297798,
+ "learning_rate": 0.0006,
+ "loss": 4.669532299041748,
+ "step": 3919
+ },
+ {
+ "epoch": 54.4471615720524,
+ "grad_norm": 0.013597294688224792,
+ "learning_rate": 0.0006,
+ "loss": 4.6644287109375,
+ "step": 3920
+ },
+ {
+ "epoch": 54.46113537117904,
+ "grad_norm": 0.013185849413275719,
+ "learning_rate": 0.0006,
+ "loss": 4.636900901794434,
+ "step": 3921
+ },
+ {
+ "epoch": 54.47510917030568,
+ "grad_norm": 0.014311990700662136,
+ "learning_rate": 0.0006,
+ "loss": 4.6047587394714355,
+ "step": 3922
+ },
+ {
+ "epoch": 54.48908296943232,
+ "grad_norm": 0.012426027096807957,
+ "learning_rate": 0.0006,
+ "loss": 4.6161394119262695,
+ "step": 3923
+ },
+ {
+ "epoch": 54.50305676855895,
+ "grad_norm": 0.01283493172377348,
+ "learning_rate": 0.0006,
+ "loss": 4.584190368652344,
+ "step": 3924
+ },
+ {
+ "epoch": 54.51703056768559,
+ "grad_norm": 0.01339943427592516,
+ "learning_rate": 0.0006,
+ "loss": 4.624184608459473,
+ "step": 3925
+ },
+ {
+ "epoch": 54.531004366812226,
+ "grad_norm": 0.012859724462032318,
+ "learning_rate": 0.0006,
+ "loss": 4.6174635887146,
+ "step": 3926
+ },
+ {
+ "epoch": 54.544978165938865,
+ "grad_norm": 0.014420023187994957,
+ "learning_rate": 0.0006,
+ "loss": 4.536544322967529,
+ "step": 3927
+ },
+ {
+ "epoch": 54.5589519650655,
+ "grad_norm": 0.014237185008823872,
+ "learning_rate": 0.0006,
+ "loss": 4.681514739990234,
+ "step": 3928
+ },
+ {
+ "epoch": 54.57292576419214,
+ "grad_norm": 0.013340328820049763,
+ "learning_rate": 0.0006,
+ "loss": 4.579689979553223,
+ "step": 3929
+ },
+ {
+ "epoch": 54.58689956331878,
+ "grad_norm": 0.013894530944526196,
+ "learning_rate": 0.0006,
+ "loss": 4.638958930969238,
+ "step": 3930
+ },
+ {
+ "epoch": 54.60087336244541,
+ "grad_norm": 0.016181141138076782,
+ "learning_rate": 0.0006,
+ "loss": 4.573508262634277,
+ "step": 3931
+ },
+ {
+ "epoch": 54.61484716157205,
+ "grad_norm": 0.013221992179751396,
+ "learning_rate": 0.0006,
+ "loss": 4.6229047775268555,
+ "step": 3932
+ },
+ {
+ "epoch": 54.62882096069869,
+ "grad_norm": 0.013241561129689217,
+ "learning_rate": 0.0006,
+ "loss": 4.641152858734131,
+ "step": 3933
+ },
+ {
+ "epoch": 54.64279475982533,
+ "grad_norm": 0.016237538307905197,
+ "learning_rate": 0.0006,
+ "loss": 4.6421217918396,
+ "step": 3934
+ },
+ {
+ "epoch": 54.65676855895197,
+ "grad_norm": 0.020866507664322853,
+ "learning_rate": 0.0006,
+ "loss": 4.646058082580566,
+ "step": 3935
+ },
+ {
+ "epoch": 54.670742358078606,
+ "grad_norm": 0.022182893007993698,
+ "learning_rate": 0.0006,
+ "loss": 4.554501533508301,
+ "step": 3936
+ },
+ {
+ "epoch": 54.68471615720524,
+ "grad_norm": 0.02254868485033512,
+ "learning_rate": 0.0006,
+ "loss": 4.563869953155518,
+ "step": 3937
+ },
+ {
+ "epoch": 54.698689956331876,
+ "grad_norm": 0.019730940461158752,
+ "learning_rate": 0.0006,
+ "loss": 4.592165946960449,
+ "step": 3938
+ },
+ {
+ "epoch": 54.712663755458514,
+ "grad_norm": 0.015146561898291111,
+ "learning_rate": 0.0006,
+ "loss": 4.6228742599487305,
+ "step": 3939
+ },
+ {
+ "epoch": 54.72663755458515,
+ "grad_norm": 0.01824716106057167,
+ "learning_rate": 0.0006,
+ "loss": 4.716236114501953,
+ "step": 3940
+ },
+ {
+ "epoch": 54.74061135371179,
+ "grad_norm": 0.017730550840497017,
+ "learning_rate": 0.0006,
+ "loss": 4.650232315063477,
+ "step": 3941
+ },
+ {
+ "epoch": 54.75458515283843,
+ "grad_norm": 0.014875595457851887,
+ "learning_rate": 0.0006,
+ "loss": 4.6377458572387695,
+ "step": 3942
+ },
+ {
+ "epoch": 54.76855895196506,
+ "grad_norm": 0.01613209955394268,
+ "learning_rate": 0.0006,
+ "loss": 4.5768537521362305,
+ "step": 3943
+ },
+ {
+ "epoch": 54.7825327510917,
+ "grad_norm": 0.017311280593276024,
+ "learning_rate": 0.0006,
+ "loss": 4.625744342803955,
+ "step": 3944
+ },
+ {
+ "epoch": 54.79650655021834,
+ "grad_norm": 0.018770085647702217,
+ "learning_rate": 0.0006,
+ "loss": 4.613787651062012,
+ "step": 3945
+ },
+ {
+ "epoch": 54.81048034934498,
+ "grad_norm": 0.0165257528424263,
+ "learning_rate": 0.0006,
+ "loss": 4.5991315841674805,
+ "step": 3946
+ },
+ {
+ "epoch": 54.82445414847162,
+ "grad_norm": 0.017433857545256615,
+ "learning_rate": 0.0006,
+ "loss": 4.513872146606445,
+ "step": 3947
+ },
+ {
+ "epoch": 54.838427947598255,
+ "grad_norm": 0.020126905292272568,
+ "learning_rate": 0.0006,
+ "loss": 4.616185188293457,
+ "step": 3948
+ },
+ {
+ "epoch": 54.852401746724894,
+ "grad_norm": 0.01819358579814434,
+ "learning_rate": 0.0006,
+ "loss": 4.535691261291504,
+ "step": 3949
+ },
+ {
+ "epoch": 54.866375545851525,
+ "grad_norm": 0.01899927295744419,
+ "learning_rate": 0.0006,
+ "loss": 4.595199108123779,
+ "step": 3950
+ },
+ {
+ "epoch": 54.880349344978164,
+ "grad_norm": 0.02005566656589508,
+ "learning_rate": 0.0006,
+ "loss": 4.633289337158203,
+ "step": 3951
+ },
+ {
+ "epoch": 54.8943231441048,
+ "grad_norm": 0.021570220589637756,
+ "learning_rate": 0.0006,
+ "loss": 4.592503547668457,
+ "step": 3952
+ },
+ {
+ "epoch": 54.90829694323144,
+ "grad_norm": 0.01883506216108799,
+ "learning_rate": 0.0006,
+ "loss": 4.6597795486450195,
+ "step": 3953
+ },
+ {
+ "epoch": 54.92227074235808,
+ "grad_norm": 0.018070021644234657,
+ "learning_rate": 0.0006,
+ "loss": 4.601474285125732,
+ "step": 3954
+ },
+ {
+ "epoch": 54.93624454148472,
+ "grad_norm": 0.018090378493070602,
+ "learning_rate": 0.0006,
+ "loss": 4.569846153259277,
+ "step": 3955
+ },
+ {
+ "epoch": 54.95021834061135,
+ "grad_norm": 0.018751563504338264,
+ "learning_rate": 0.0006,
+ "loss": 4.646064281463623,
+ "step": 3956
+ },
+ {
+ "epoch": 54.96419213973799,
+ "grad_norm": 0.017512254416942596,
+ "learning_rate": 0.0006,
+ "loss": 4.606670379638672,
+ "step": 3957
+ },
+ {
+ "epoch": 54.97816593886463,
+ "grad_norm": 0.018771110102534294,
+ "learning_rate": 0.0006,
+ "loss": 4.604315757751465,
+ "step": 3958
+ },
+ {
+ "epoch": 54.992139737991266,
+ "grad_norm": 0.01611141301691532,
+ "learning_rate": 0.0006,
+ "loss": 4.643245220184326,
+ "step": 3959
+ },
+ {
+ "epoch": 55.0,
+ "grad_norm": 0.0156728383153677,
+ "learning_rate": 0.0006,
+ "loss": 4.6739044189453125,
+ "step": 3960
+ },
+ {
+ "epoch": 55.0,
+ "eval_loss": 4.851032257080078,
+ "eval_runtime": 56.8699,
+ "eval_samples_per_second": 42.94,
+ "eval_steps_per_second": 1.354,
+ "step": 3960
+ },
+ {
+ "epoch": 55.01397379912664,
+ "grad_norm": 0.016214672476053238,
+ "learning_rate": 0.0006,
+ "loss": 4.571749687194824,
+ "step": 3961
+ },
+ {
+ "epoch": 55.02794759825328,
+ "grad_norm": 0.01649155467748642,
+ "learning_rate": 0.0006,
+ "loss": 4.553592205047607,
+ "step": 3962
+ },
+ {
+ "epoch": 55.041921397379916,
+ "grad_norm": 0.016793379560112953,
+ "learning_rate": 0.0006,
+ "loss": 4.650628566741943,
+ "step": 3963
+ },
+ {
+ "epoch": 55.05589519650655,
+ "grad_norm": 0.0191937405616045,
+ "learning_rate": 0.0006,
+ "loss": 4.620532989501953,
+ "step": 3964
+ },
+ {
+ "epoch": 55.069868995633186,
+ "grad_norm": 0.020976556465029716,
+ "learning_rate": 0.0006,
+ "loss": 4.6329545974731445,
+ "step": 3965
+ },
+ {
+ "epoch": 55.083842794759825,
+ "grad_norm": 0.023450477048754692,
+ "learning_rate": 0.0006,
+ "loss": 4.60828971862793,
+ "step": 3966
+ },
+ {
+ "epoch": 55.09781659388646,
+ "grad_norm": 0.026778312399983406,
+ "learning_rate": 0.0006,
+ "loss": 4.617481708526611,
+ "step": 3967
+ },
+ {
+ "epoch": 55.1117903930131,
+ "grad_norm": 0.02333568036556244,
+ "learning_rate": 0.0006,
+ "loss": 4.580722332000732,
+ "step": 3968
+ },
+ {
+ "epoch": 55.12576419213974,
+ "grad_norm": 0.018491310998797417,
+ "learning_rate": 0.0006,
+ "loss": 4.590953350067139,
+ "step": 3969
+ },
+ {
+ "epoch": 55.13973799126637,
+ "grad_norm": 0.020339803770184517,
+ "learning_rate": 0.0006,
+ "loss": 4.518374443054199,
+ "step": 3970
+ },
+ {
+ "epoch": 55.15371179039301,
+ "grad_norm": 0.01759706623852253,
+ "learning_rate": 0.0006,
+ "loss": 4.560522079467773,
+ "step": 3971
+ },
+ {
+ "epoch": 55.16768558951965,
+ "grad_norm": 0.015875980257987976,
+ "learning_rate": 0.0006,
+ "loss": 4.4795708656311035,
+ "step": 3972
+ },
+ {
+ "epoch": 55.18165938864629,
+ "grad_norm": 0.01650134287774563,
+ "learning_rate": 0.0006,
+ "loss": 4.57435941696167,
+ "step": 3973
+ },
+ {
+ "epoch": 55.19563318777293,
+ "grad_norm": 0.01734331250190735,
+ "learning_rate": 0.0006,
+ "loss": 4.661007881164551,
+ "step": 3974
+ },
+ {
+ "epoch": 55.209606986899566,
+ "grad_norm": 0.01651303470134735,
+ "learning_rate": 0.0006,
+ "loss": 4.511148452758789,
+ "step": 3975
+ },
+ {
+ "epoch": 55.223580786026204,
+ "grad_norm": 0.016189761459827423,
+ "learning_rate": 0.0006,
+ "loss": 4.592350006103516,
+ "step": 3976
+ },
+ {
+ "epoch": 55.237554585152836,
+ "grad_norm": 0.01833222061395645,
+ "learning_rate": 0.0006,
+ "loss": 4.506649971008301,
+ "step": 3977
+ },
+ {
+ "epoch": 55.251528384279474,
+ "grad_norm": 0.02433410845696926,
+ "learning_rate": 0.0006,
+ "loss": 4.49500036239624,
+ "step": 3978
+ },
+ {
+ "epoch": 55.26550218340611,
+ "grad_norm": 0.0314289927482605,
+ "learning_rate": 0.0006,
+ "loss": 4.507624626159668,
+ "step": 3979
+ },
+ {
+ "epoch": 55.27947598253275,
+ "grad_norm": 0.03013679012656212,
+ "learning_rate": 0.0006,
+ "loss": 4.533641815185547,
+ "step": 3980
+ },
+ {
+ "epoch": 55.29344978165939,
+ "grad_norm": 0.024524061009287834,
+ "learning_rate": 0.0006,
+ "loss": 4.5721116065979,
+ "step": 3981
+ },
+ {
+ "epoch": 55.30742358078603,
+ "grad_norm": 0.02640336938202381,
+ "learning_rate": 0.0006,
+ "loss": 4.652521133422852,
+ "step": 3982
+ },
+ {
+ "epoch": 55.32139737991266,
+ "grad_norm": 0.021309101954102516,
+ "learning_rate": 0.0006,
+ "loss": 4.600375175476074,
+ "step": 3983
+ },
+ {
+ "epoch": 55.3353711790393,
+ "grad_norm": 0.023008549585938454,
+ "learning_rate": 0.0006,
+ "loss": 4.53914737701416,
+ "step": 3984
+ },
+ {
+ "epoch": 55.34934497816594,
+ "grad_norm": 0.022491468116641045,
+ "learning_rate": 0.0006,
+ "loss": 4.661163330078125,
+ "step": 3985
+ },
+ {
+ "epoch": 55.36331877729258,
+ "grad_norm": 0.01904081553220749,
+ "learning_rate": 0.0006,
+ "loss": 4.5426177978515625,
+ "step": 3986
+ },
+ {
+ "epoch": 55.377292576419215,
+ "grad_norm": 0.019229518249630928,
+ "learning_rate": 0.0006,
+ "loss": 4.606066703796387,
+ "step": 3987
+ },
+ {
+ "epoch": 55.391266375545854,
+ "grad_norm": 0.01781383901834488,
+ "learning_rate": 0.0006,
+ "loss": 4.54642391204834,
+ "step": 3988
+ },
+ {
+ "epoch": 55.40524017467249,
+ "grad_norm": 0.0174600500613451,
+ "learning_rate": 0.0006,
+ "loss": 4.711891174316406,
+ "step": 3989
+ },
+ {
+ "epoch": 55.419213973799124,
+ "grad_norm": 0.018154339864850044,
+ "learning_rate": 0.0006,
+ "loss": 4.527122974395752,
+ "step": 3990
+ },
+ {
+ "epoch": 55.43318777292576,
+ "grad_norm": 0.01914648339152336,
+ "learning_rate": 0.0006,
+ "loss": 4.641975402832031,
+ "step": 3991
+ },
+ {
+ "epoch": 55.4471615720524,
+ "grad_norm": 0.017118625342845917,
+ "learning_rate": 0.0006,
+ "loss": 4.511425971984863,
+ "step": 3992
+ },
+ {
+ "epoch": 55.46113537117904,
+ "grad_norm": 0.01551489345729351,
+ "learning_rate": 0.0006,
+ "loss": 4.551141738891602,
+ "step": 3993
+ },
+ {
+ "epoch": 55.47510917030568,
+ "grad_norm": 0.015072065405547619,
+ "learning_rate": 0.0006,
+ "loss": 4.651115894317627,
+ "step": 3994
+ },
+ {
+ "epoch": 55.48908296943232,
+ "grad_norm": 0.017010986804962158,
+ "learning_rate": 0.0006,
+ "loss": 4.583282470703125,
+ "step": 3995
+ },
+ {
+ "epoch": 55.50305676855895,
+ "grad_norm": 0.01764831133186817,
+ "learning_rate": 0.0006,
+ "loss": 4.593364238739014,
+ "step": 3996
+ },
+ {
+ "epoch": 55.51703056768559,
+ "grad_norm": 0.01762884296476841,
+ "learning_rate": 0.0006,
+ "loss": 4.45671272277832,
+ "step": 3997
+ },
+ {
+ "epoch": 55.531004366812226,
+ "grad_norm": 0.01600457727909088,
+ "learning_rate": 0.0006,
+ "loss": 4.511260032653809,
+ "step": 3998
+ },
+ {
+ "epoch": 55.544978165938865,
+ "grad_norm": 0.015706021338701248,
+ "learning_rate": 0.0006,
+ "loss": 4.47593355178833,
+ "step": 3999
+ },
+ {
+ "epoch": 55.5589519650655,
+ "grad_norm": 0.017718996852636337,
+ "learning_rate": 0.0006,
+ "loss": 4.610543251037598,
+ "step": 4000
+ },
+ {
+ "epoch": 55.57292576419214,
+ "grad_norm": 0.015073757618665695,
+ "learning_rate": 0.0006,
+ "loss": 4.466620445251465,
+ "step": 4001
+ },
+ {
+ "epoch": 55.58689956331878,
+ "grad_norm": 0.01789247989654541,
+ "learning_rate": 0.0006,
+ "loss": 4.503409385681152,
+ "step": 4002
+ },
+ {
+ "epoch": 55.60087336244541,
+ "grad_norm": 0.022329598665237427,
+ "learning_rate": 0.0006,
+ "loss": 4.62099552154541,
+ "step": 4003
+ },
+ {
+ "epoch": 55.61484716157205,
+ "grad_norm": 0.02343829534947872,
+ "learning_rate": 0.0006,
+ "loss": 4.620813369750977,
+ "step": 4004
+ },
+ {
+ "epoch": 55.62882096069869,
+ "grad_norm": 0.023279687389731407,
+ "learning_rate": 0.0006,
+ "loss": 4.545765399932861,
+ "step": 4005
+ },
+ {
+ "epoch": 55.64279475982533,
+ "grad_norm": 0.020817464217543602,
+ "learning_rate": 0.0006,
+ "loss": 4.561470031738281,
+ "step": 4006
+ },
+ {
+ "epoch": 55.65676855895197,
+ "grad_norm": 0.020270058885216713,
+ "learning_rate": 0.0006,
+ "loss": 4.5875020027160645,
+ "step": 4007
+ },
+ {
+ "epoch": 55.670742358078606,
+ "grad_norm": 0.021085843443870544,
+ "learning_rate": 0.0006,
+ "loss": 4.504560470581055,
+ "step": 4008
+ },
+ {
+ "epoch": 55.68471615720524,
+ "grad_norm": 0.018507730215787888,
+ "learning_rate": 0.0006,
+ "loss": 4.520644664764404,
+ "step": 4009
+ },
+ {
+ "epoch": 55.698689956331876,
+ "grad_norm": 0.020571516826748848,
+ "learning_rate": 0.0006,
+ "loss": 4.539643287658691,
+ "step": 4010
+ },
+ {
+ "epoch": 55.712663755458514,
+ "grad_norm": 0.021484510973095894,
+ "learning_rate": 0.0006,
+ "loss": 4.628833293914795,
+ "step": 4011
+ },
+ {
+ "epoch": 55.72663755458515,
+ "grad_norm": 0.017693819478154182,
+ "learning_rate": 0.0006,
+ "loss": 4.6023783683776855,
+ "step": 4012
+ },
+ {
+ "epoch": 55.74061135371179,
+ "grad_norm": 0.020186059176921844,
+ "learning_rate": 0.0006,
+ "loss": 4.634884834289551,
+ "step": 4013
+ },
+ {
+ "epoch": 55.75458515283843,
+ "grad_norm": 0.01835780404508114,
+ "learning_rate": 0.0006,
+ "loss": 4.577791213989258,
+ "step": 4014
+ },
+ {
+ "epoch": 55.76855895196506,
+ "grad_norm": 0.018162988126277924,
+ "learning_rate": 0.0006,
+ "loss": 4.641597270965576,
+ "step": 4015
+ },
+ {
+ "epoch": 55.7825327510917,
+ "grad_norm": 0.018909504637122154,
+ "learning_rate": 0.0006,
+ "loss": 4.487447738647461,
+ "step": 4016
+ },
+ {
+ "epoch": 55.79650655021834,
+ "grad_norm": 0.018972251564264297,
+ "learning_rate": 0.0006,
+ "loss": 4.590317726135254,
+ "step": 4017
+ },
+ {
+ "epoch": 55.81048034934498,
+ "grad_norm": 0.01793607883155346,
+ "learning_rate": 0.0006,
+ "loss": 4.4994611740112305,
+ "step": 4018
+ },
+ {
+ "epoch": 55.82445414847162,
+ "grad_norm": 0.02123514749109745,
+ "learning_rate": 0.0006,
+ "loss": 4.479006767272949,
+ "step": 4019
+ },
+ {
+ "epoch": 55.838427947598255,
+ "grad_norm": 0.021246599033474922,
+ "learning_rate": 0.0006,
+ "loss": 4.4738640785217285,
+ "step": 4020
+ },
+ {
+ "epoch": 55.852401746724894,
+ "grad_norm": 0.020131602883338928,
+ "learning_rate": 0.0006,
+ "loss": 4.600308418273926,
+ "step": 4021
+ },
+ {
+ "epoch": 55.866375545851525,
+ "grad_norm": 0.019159460440278053,
+ "learning_rate": 0.0006,
+ "loss": 4.540908336639404,
+ "step": 4022
+ },
+ {
+ "epoch": 55.880349344978164,
+ "grad_norm": 0.023114705458283424,
+ "learning_rate": 0.0006,
+ "loss": 4.641247749328613,
+ "step": 4023
+ },
+ {
+ "epoch": 55.8943231441048,
+ "grad_norm": 0.02296135015785694,
+ "learning_rate": 0.0006,
+ "loss": 4.444624423980713,
+ "step": 4024
+ },
+ {
+ "epoch": 55.90829694323144,
+ "grad_norm": 0.02204655110836029,
+ "learning_rate": 0.0006,
+ "loss": 4.506349563598633,
+ "step": 4025
+ },
+ {
+ "epoch": 55.92227074235808,
+ "grad_norm": 0.019338225945830345,
+ "learning_rate": 0.0006,
+ "loss": 4.616047382354736,
+ "step": 4026
+ },
+ {
+ "epoch": 55.93624454148472,
+ "grad_norm": 0.020357884466648102,
+ "learning_rate": 0.0006,
+ "loss": 4.48170280456543,
+ "step": 4027
+ },
+ {
+ "epoch": 55.95021834061135,
+ "grad_norm": 0.02426772564649582,
+ "learning_rate": 0.0006,
+ "loss": 4.576877117156982,
+ "step": 4028
+ },
+ {
+ "epoch": 55.96419213973799,
+ "grad_norm": 0.023351816460490227,
+ "learning_rate": 0.0006,
+ "loss": 4.458885192871094,
+ "step": 4029
+ },
+ {
+ "epoch": 55.97816593886463,
+ "grad_norm": 0.019892243668437004,
+ "learning_rate": 0.0006,
+ "loss": 4.520229816436768,
+ "step": 4030
+ },
+ {
+ "epoch": 55.992139737991266,
+ "grad_norm": 0.018331089988350868,
+ "learning_rate": 0.0006,
+ "loss": 4.432201862335205,
+ "step": 4031
+ },
+ {
+ "epoch": 56.0,
+ "grad_norm": 0.021300828084349632,
+ "learning_rate": 0.0006,
+ "loss": 4.617072582244873,
+ "step": 4032
+ },
+ {
+ "epoch": 56.0,
+ "eval_loss": 4.811229705810547,
+ "eval_runtime": 56.4846,
+ "eval_samples_per_second": 43.233,
+ "eval_steps_per_second": 1.363,
+ "step": 4032
+ },
+ {
+ "epoch": 56.01397379912664,
+ "grad_norm": 0.02199401892721653,
+ "learning_rate": 0.0006,
+ "loss": 4.513965606689453,
+ "step": 4033
+ },
+ {
+ "epoch": 56.02794759825328,
+ "grad_norm": 0.022109858691692352,
+ "learning_rate": 0.0006,
+ "loss": 4.557568550109863,
+ "step": 4034
+ },
+ {
+ "epoch": 56.041921397379916,
+ "grad_norm": 0.020103694871068,
+ "learning_rate": 0.0006,
+ "loss": 4.569512844085693,
+ "step": 4035
+ },
+ {
+ "epoch": 56.05589519650655,
+ "grad_norm": 0.01898941583931446,
+ "learning_rate": 0.0006,
+ "loss": 4.623138904571533,
+ "step": 4036
+ },
+ {
+ "epoch": 56.069868995633186,
+ "grad_norm": 0.02227010577917099,
+ "learning_rate": 0.0006,
+ "loss": 4.433316230773926,
+ "step": 4037
+ },
+ {
+ "epoch": 56.083842794759825,
+ "grad_norm": 0.024683045223355293,
+ "learning_rate": 0.0006,
+ "loss": 4.458296775817871,
+ "step": 4038
+ },
+ {
+ "epoch": 56.09781659388646,
+ "grad_norm": 0.02900184690952301,
+ "learning_rate": 0.0006,
+ "loss": 4.531060218811035,
+ "step": 4039
+ },
+ {
+ "epoch": 56.1117903930131,
+ "grad_norm": 0.025557566434144974,
+ "learning_rate": 0.0006,
+ "loss": 4.56685733795166,
+ "step": 4040
+ },
+ {
+ "epoch": 56.12576419213974,
+ "grad_norm": 0.021414656192064285,
+ "learning_rate": 0.0006,
+ "loss": 4.469037055969238,
+ "step": 4041
+ },
+ {
+ "epoch": 56.13973799126637,
+ "grad_norm": 0.023262616246938705,
+ "learning_rate": 0.0006,
+ "loss": 4.493876934051514,
+ "step": 4042
+ },
+ {
+ "epoch": 56.15371179039301,
+ "grad_norm": 0.023011326789855957,
+ "learning_rate": 0.0006,
+ "loss": 4.4572906494140625,
+ "step": 4043
+ },
+ {
+ "epoch": 56.16768558951965,
+ "grad_norm": 0.02177615463733673,
+ "learning_rate": 0.0006,
+ "loss": 4.615389823913574,
+ "step": 4044
+ },
+ {
+ "epoch": 56.18165938864629,
+ "grad_norm": 0.021527713164687157,
+ "learning_rate": 0.0006,
+ "loss": 4.613762855529785,
+ "step": 4045
+ },
+ {
+ "epoch": 56.19563318777293,
+ "grad_norm": 0.01982325315475464,
+ "learning_rate": 0.0006,
+ "loss": 4.53670597076416,
+ "step": 4046
+ },
+ {
+ "epoch": 56.209606986899566,
+ "grad_norm": 0.022484654560685158,
+ "learning_rate": 0.0006,
+ "loss": 4.605565071105957,
+ "step": 4047
+ },
+ {
+ "epoch": 56.223580786026204,
+ "grad_norm": 0.02346140146255493,
+ "learning_rate": 0.0006,
+ "loss": 4.6265950202941895,
+ "step": 4048
+ },
+ {
+ "epoch": 56.237554585152836,
+ "grad_norm": 0.01876714453101158,
+ "learning_rate": 0.0006,
+ "loss": 4.481525421142578,
+ "step": 4049
+ },
+ {
+ "epoch": 56.251528384279474,
+ "grad_norm": 0.01926851086318493,
+ "learning_rate": 0.0006,
+ "loss": 4.630258083343506,
+ "step": 4050
+ },
+ {
+ "epoch": 56.26550218340611,
+ "grad_norm": 0.020858589559793472,
+ "learning_rate": 0.0006,
+ "loss": 4.5265913009643555,
+ "step": 4051
+ },
+ {
+ "epoch": 56.27947598253275,
+ "grad_norm": 0.01889280416071415,
+ "learning_rate": 0.0006,
+ "loss": 4.581275463104248,
+ "step": 4052
+ },
+ {
+ "epoch": 56.29344978165939,
+ "grad_norm": 0.01801002025604248,
+ "learning_rate": 0.0006,
+ "loss": 4.498557090759277,
+ "step": 4053
+ },
+ {
+ "epoch": 56.30742358078603,
+ "grad_norm": 0.02252180315554142,
+ "learning_rate": 0.0006,
+ "loss": 4.565267086029053,
+ "step": 4054
+ },
+ {
+ "epoch": 56.32139737991266,
+ "grad_norm": 0.02306070365011692,
+ "learning_rate": 0.0006,
+ "loss": 4.3591389656066895,
+ "step": 4055
+ },
+ {
+ "epoch": 56.3353711790393,
+ "grad_norm": 0.020292513072490692,
+ "learning_rate": 0.0006,
+ "loss": 4.407517433166504,
+ "step": 4056
+ },
+ {
+ "epoch": 56.34934497816594,
+ "grad_norm": 0.019154123961925507,
+ "learning_rate": 0.0006,
+ "loss": 4.482261657714844,
+ "step": 4057
+ },
+ {
+ "epoch": 56.36331877729258,
+ "grad_norm": 0.017338937148451805,
+ "learning_rate": 0.0006,
+ "loss": 4.4463582038879395,
+ "step": 4058
+ },
+ {
+ "epoch": 56.377292576419215,
+ "grad_norm": 0.01730559952557087,
+ "learning_rate": 0.0006,
+ "loss": 4.542474746704102,
+ "step": 4059
+ },
+ {
+ "epoch": 56.391266375545854,
+ "grad_norm": 0.017029426991939545,
+ "learning_rate": 0.0006,
+ "loss": 4.552488327026367,
+ "step": 4060
+ },
+ {
+ "epoch": 56.40524017467249,
+ "grad_norm": 0.01745498552918434,
+ "learning_rate": 0.0006,
+ "loss": 4.52928352355957,
+ "step": 4061
+ },
+ {
+ "epoch": 56.419213973799124,
+ "grad_norm": 0.0168515145778656,
+ "learning_rate": 0.0006,
+ "loss": 4.4516801834106445,
+ "step": 4062
+ },
+ {
+ "epoch": 56.43318777292576,
+ "grad_norm": 0.018545417115092278,
+ "learning_rate": 0.0006,
+ "loss": 4.569811820983887,
+ "step": 4063
+ },
+ {
+ "epoch": 56.4471615720524,
+ "grad_norm": 0.02227838523685932,
+ "learning_rate": 0.0006,
+ "loss": 4.412224769592285,
+ "step": 4064
+ },
+ {
+ "epoch": 56.46113537117904,
+ "grad_norm": 0.023338504135608673,
+ "learning_rate": 0.0006,
+ "loss": 4.490184307098389,
+ "step": 4065
+ },
+ {
+ "epoch": 56.47510917030568,
+ "grad_norm": 0.02257535606622696,
+ "learning_rate": 0.0006,
+ "loss": 4.479152679443359,
+ "step": 4066
+ },
+ {
+ "epoch": 56.48908296943232,
+ "grad_norm": 0.022737598046660423,
+ "learning_rate": 0.0006,
+ "loss": 4.418235778808594,
+ "step": 4067
+ },
+ {
+ "epoch": 56.50305676855895,
+ "grad_norm": 0.024603573605418205,
+ "learning_rate": 0.0006,
+ "loss": 4.45781135559082,
+ "step": 4068
+ },
+ {
+ "epoch": 56.51703056768559,
+ "grad_norm": 0.024953778833150864,
+ "learning_rate": 0.0006,
+ "loss": 4.593907833099365,
+ "step": 4069
+ },
+ {
+ "epoch": 56.531004366812226,
+ "grad_norm": 0.02195645309984684,
+ "learning_rate": 0.0006,
+ "loss": 4.467643737792969,
+ "step": 4070
+ },
+ {
+ "epoch": 56.544978165938865,
+ "grad_norm": 0.017126578837633133,
+ "learning_rate": 0.0006,
+ "loss": 4.555484771728516,
+ "step": 4071
+ },
+ {
+ "epoch": 56.5589519650655,
+ "grad_norm": 0.019482605159282684,
+ "learning_rate": 0.0006,
+ "loss": 4.430915832519531,
+ "step": 4072
+ },
+ {
+ "epoch": 56.57292576419214,
+ "grad_norm": 0.0219392292201519,
+ "learning_rate": 0.0006,
+ "loss": 4.593365669250488,
+ "step": 4073
+ },
+ {
+ "epoch": 56.58689956331878,
+ "grad_norm": 0.02420145645737648,
+ "learning_rate": 0.0006,
+ "loss": 4.491467475891113,
+ "step": 4074
+ },
+ {
+ "epoch": 56.60087336244541,
+ "grad_norm": 0.021523723378777504,
+ "learning_rate": 0.0006,
+ "loss": 4.463634490966797,
+ "step": 4075
+ },
+ {
+ "epoch": 56.61484716157205,
+ "grad_norm": 0.020242048427462578,
+ "learning_rate": 0.0006,
+ "loss": 4.577968597412109,
+ "step": 4076
+ },
+ {
+ "epoch": 56.62882096069869,
+ "grad_norm": 0.017771173268556595,
+ "learning_rate": 0.0006,
+ "loss": 4.568269729614258,
+ "step": 4077
+ },
+ {
+ "epoch": 56.64279475982533,
+ "grad_norm": 0.01598544418811798,
+ "learning_rate": 0.0006,
+ "loss": 4.476841926574707,
+ "step": 4078
+ },
+ {
+ "epoch": 56.65676855895197,
+ "grad_norm": 0.0168446097522974,
+ "learning_rate": 0.0006,
+ "loss": 4.480965614318848,
+ "step": 4079
+ },
+ {
+ "epoch": 56.670742358078606,
+ "grad_norm": 0.014948660507798195,
+ "learning_rate": 0.0006,
+ "loss": 4.535400390625,
+ "step": 4080
+ },
+ {
+ "epoch": 56.68471615720524,
+ "grad_norm": 0.014899153262376785,
+ "learning_rate": 0.0006,
+ "loss": 4.476048946380615,
+ "step": 4081
+ },
+ {
+ "epoch": 56.698689956331876,
+ "grad_norm": 0.014051870442926884,
+ "learning_rate": 0.0006,
+ "loss": 4.533884048461914,
+ "step": 4082
+ },
+ {
+ "epoch": 56.712663755458514,
+ "grad_norm": 0.014708572067320347,
+ "learning_rate": 0.0006,
+ "loss": 4.586690902709961,
+ "step": 4083
+ },
+ {
+ "epoch": 56.72663755458515,
+ "grad_norm": 0.014797299169003963,
+ "learning_rate": 0.0006,
+ "loss": 4.446291446685791,
+ "step": 4084
+ },
+ {
+ "epoch": 56.74061135371179,
+ "grad_norm": 0.015256541781127453,
+ "learning_rate": 0.0006,
+ "loss": 4.553572654724121,
+ "step": 4085
+ },
+ {
+ "epoch": 56.75458515283843,
+ "grad_norm": 0.016042491421103477,
+ "learning_rate": 0.0006,
+ "loss": 4.507801055908203,
+ "step": 4086
+ },
+ {
+ "epoch": 56.76855895196506,
+ "grad_norm": 0.018835173919796944,
+ "learning_rate": 0.0006,
+ "loss": 4.393911838531494,
+ "step": 4087
+ },
+ {
+ "epoch": 56.7825327510917,
+ "grad_norm": 0.021994033828377724,
+ "learning_rate": 0.0006,
+ "loss": 4.543368339538574,
+ "step": 4088
+ },
+ {
+ "epoch": 56.79650655021834,
+ "grad_norm": 0.0230990182608366,
+ "learning_rate": 0.0006,
+ "loss": 4.507420539855957,
+ "step": 4089
+ },
+ {
+ "epoch": 56.81048034934498,
+ "grad_norm": 0.018818842247128487,
+ "learning_rate": 0.0006,
+ "loss": 4.499941349029541,
+ "step": 4090
+ },
+ {
+ "epoch": 56.82445414847162,
+ "grad_norm": 0.018257278949022293,
+ "learning_rate": 0.0006,
+ "loss": 4.380057334899902,
+ "step": 4091
+ },
+ {
+ "epoch": 56.838427947598255,
+ "grad_norm": 0.020073018968105316,
+ "learning_rate": 0.0006,
+ "loss": 4.572811603546143,
+ "step": 4092
+ },
+ {
+ "epoch": 56.852401746724894,
+ "grad_norm": 0.02078658528625965,
+ "learning_rate": 0.0006,
+ "loss": 4.446835517883301,
+ "step": 4093
+ },
+ {
+ "epoch": 56.866375545851525,
+ "grad_norm": 0.01793118566274643,
+ "learning_rate": 0.0006,
+ "loss": 4.517854690551758,
+ "step": 4094
+ },
+ {
+ "epoch": 56.880349344978164,
+ "grad_norm": 0.01608281210064888,
+ "learning_rate": 0.0006,
+ "loss": 4.506482124328613,
+ "step": 4095
+ },
+ {
+ "epoch": 56.8943231441048,
+ "grad_norm": 0.017405368387699127,
+ "learning_rate": 0.0006,
+ "loss": 4.439718246459961,
+ "step": 4096
+ },
+ {
+ "epoch": 56.90829694323144,
+ "grad_norm": 0.017930863425135612,
+ "learning_rate": 0.0006,
+ "loss": 4.409365653991699,
+ "step": 4097
+ },
+ {
+ "epoch": 56.92227074235808,
+ "grad_norm": 0.015757670626044273,
+ "learning_rate": 0.0006,
+ "loss": 4.505547523498535,
+ "step": 4098
+ },
+ {
+ "epoch": 56.93624454148472,
+ "grad_norm": 0.01580345816910267,
+ "learning_rate": 0.0006,
+ "loss": 4.462774753570557,
+ "step": 4099
+ },
+ {
+ "epoch": 56.95021834061135,
+ "grad_norm": 0.01659180410206318,
+ "learning_rate": 0.0006,
+ "loss": 4.49345588684082,
+ "step": 4100
+ },
+ {
+ "epoch": 56.96419213973799,
+ "grad_norm": 0.01877259835600853,
+ "learning_rate": 0.0006,
+ "loss": 4.574443340301514,
+ "step": 4101
+ },
+ {
+ "epoch": 56.97816593886463,
+ "grad_norm": 0.01977839693427086,
+ "learning_rate": 0.0006,
+ "loss": 4.574034690856934,
+ "step": 4102
+ },
+ {
+ "epoch": 56.992139737991266,
+ "grad_norm": 0.020627597346901894,
+ "learning_rate": 0.0006,
+ "loss": 4.50969123840332,
+ "step": 4103
+ },
+ {
+ "epoch": 57.0,
+ "grad_norm": 0.025128807872533798,
+ "learning_rate": 0.0006,
+ "loss": 4.401662826538086,
+ "step": 4104
+ },
+ {
+ "epoch": 57.0,
+ "eval_loss": 4.779252052307129,
+ "eval_runtime": 56.8104,
+ "eval_samples_per_second": 42.985,
+ "eval_steps_per_second": 1.355,
+ "step": 4104
+ },
+ {
+ "epoch": 57.01397379912664,
+ "grad_norm": 0.027712570503354073,
+ "learning_rate": 0.0006,
+ "loss": 4.472955703735352,
+ "step": 4105
+ },
+ {
+ "epoch": 57.02794759825328,
+ "grad_norm": 0.02164197526872158,
+ "learning_rate": 0.0006,
+ "loss": 4.387596130371094,
+ "step": 4106
+ },
+ {
+ "epoch": 57.041921397379916,
+ "grad_norm": 0.017972752451896667,
+ "learning_rate": 0.0006,
+ "loss": 4.4326276779174805,
+ "step": 4107
+ },
+ {
+ "epoch": 57.05589519650655,
+ "grad_norm": 0.020811019465327263,
+ "learning_rate": 0.0006,
+ "loss": 4.452951431274414,
+ "step": 4108
+ },
+ {
+ "epoch": 57.069868995633186,
+ "grad_norm": 0.019913259893655777,
+ "learning_rate": 0.0006,
+ "loss": 4.448301315307617,
+ "step": 4109
+ },
+ {
+ "epoch": 57.083842794759825,
+ "grad_norm": 0.017409102991223335,
+ "learning_rate": 0.0006,
+ "loss": 4.4108357429504395,
+ "step": 4110
+ },
+ {
+ "epoch": 57.09781659388646,
+ "grad_norm": 0.019374269992113113,
+ "learning_rate": 0.0006,
+ "loss": 4.398412227630615,
+ "step": 4111
+ },
+ {
+ "epoch": 57.1117903930131,
+ "grad_norm": 0.015521776862442493,
+ "learning_rate": 0.0006,
+ "loss": 4.508594989776611,
+ "step": 4112
+ },
+ {
+ "epoch": 57.12576419213974,
+ "grad_norm": 0.017201920971274376,
+ "learning_rate": 0.0006,
+ "loss": 4.449558734893799,
+ "step": 4113
+ },
+ {
+ "epoch": 57.13973799126637,
+ "grad_norm": 0.018062541261315346,
+ "learning_rate": 0.0006,
+ "loss": 4.517874240875244,
+ "step": 4114
+ },
+ {
+ "epoch": 57.15371179039301,
+ "grad_norm": 0.015846073627471924,
+ "learning_rate": 0.0006,
+ "loss": 4.427501678466797,
+ "step": 4115
+ },
+ {
+ "epoch": 57.16768558951965,
+ "grad_norm": 0.01849362626671791,
+ "learning_rate": 0.0006,
+ "loss": 4.433181285858154,
+ "step": 4116
+ },
+ {
+ "epoch": 57.18165938864629,
+ "grad_norm": 0.01948804222047329,
+ "learning_rate": 0.0006,
+ "loss": 4.446122646331787,
+ "step": 4117
+ },
+ {
+ "epoch": 57.19563318777293,
+ "grad_norm": 0.017354389652609825,
+ "learning_rate": 0.0006,
+ "loss": 4.431153774261475,
+ "step": 4118
+ },
+ {
+ "epoch": 57.209606986899566,
+ "grad_norm": 0.015817373991012573,
+ "learning_rate": 0.0006,
+ "loss": 4.448566913604736,
+ "step": 4119
+ },
+ {
+ "epoch": 57.223580786026204,
+ "grad_norm": 0.01635785959661007,
+ "learning_rate": 0.0006,
+ "loss": 4.408031463623047,
+ "step": 4120
+ },
+ {
+ "epoch": 57.237554585152836,
+ "grad_norm": 0.014528338797390461,
+ "learning_rate": 0.0006,
+ "loss": 4.541668891906738,
+ "step": 4121
+ },
+ {
+ "epoch": 57.251528384279474,
+ "grad_norm": 0.01589174196124077,
+ "learning_rate": 0.0006,
+ "loss": 4.513199329376221,
+ "step": 4122
+ },
+ {
+ "epoch": 57.26550218340611,
+ "grad_norm": 0.016093695536255836,
+ "learning_rate": 0.0006,
+ "loss": 4.516178131103516,
+ "step": 4123
+ },
+ {
+ "epoch": 57.27947598253275,
+ "grad_norm": 0.01788361556828022,
+ "learning_rate": 0.0006,
+ "loss": 4.524880409240723,
+ "step": 4124
+ },
+ {
+ "epoch": 57.29344978165939,
+ "grad_norm": 0.01591576263308525,
+ "learning_rate": 0.0006,
+ "loss": 4.44764518737793,
+ "step": 4125
+ },
+ {
+ "epoch": 57.30742358078603,
+ "grad_norm": 0.015452570281922817,
+ "learning_rate": 0.0006,
+ "loss": 4.46844482421875,
+ "step": 4126
+ },
+ {
+ "epoch": 57.32139737991266,
+ "grad_norm": 0.016404492780566216,
+ "learning_rate": 0.0006,
+ "loss": 4.443678855895996,
+ "step": 4127
+ },
+ {
+ "epoch": 57.3353711790393,
+ "grad_norm": 0.019967148080468178,
+ "learning_rate": 0.0006,
+ "loss": 4.430230140686035,
+ "step": 4128
+ },
+ {
+ "epoch": 57.34934497816594,
+ "grad_norm": 0.02594015561044216,
+ "learning_rate": 0.0006,
+ "loss": 4.47061824798584,
+ "step": 4129
+ },
+ {
+ "epoch": 57.36331877729258,
+ "grad_norm": 0.031096890568733215,
+ "learning_rate": 0.0006,
+ "loss": 4.462275981903076,
+ "step": 4130
+ },
+ {
+ "epoch": 57.377292576419215,
+ "grad_norm": 0.02834898792207241,
+ "learning_rate": 0.0006,
+ "loss": 4.41143798828125,
+ "step": 4131
+ },
+ {
+ "epoch": 57.391266375545854,
+ "grad_norm": 0.023621153086423874,
+ "learning_rate": 0.0006,
+ "loss": 4.562548637390137,
+ "step": 4132
+ },
+ {
+ "epoch": 57.40524017467249,
+ "grad_norm": 0.02277296595275402,
+ "learning_rate": 0.0006,
+ "loss": 4.456247329711914,
+ "step": 4133
+ },
+ {
+ "epoch": 57.419213973799124,
+ "grad_norm": 0.019868768751621246,
+ "learning_rate": 0.0006,
+ "loss": 4.524300575256348,
+ "step": 4134
+ },
+ {
+ "epoch": 57.43318777292576,
+ "grad_norm": 0.01887078955769539,
+ "learning_rate": 0.0006,
+ "loss": 4.540097236633301,
+ "step": 4135
+ },
+ {
+ "epoch": 57.4471615720524,
+ "grad_norm": 0.024268802255392075,
+ "learning_rate": 0.0006,
+ "loss": 4.397233963012695,
+ "step": 4136
+ },
+ {
+ "epoch": 57.46113537117904,
+ "grad_norm": 0.025739721953868866,
+ "learning_rate": 0.0006,
+ "loss": 4.354194641113281,
+ "step": 4137
+ },
+ {
+ "epoch": 57.47510917030568,
+ "grad_norm": 0.02442290261387825,
+ "learning_rate": 0.0006,
+ "loss": 4.303822994232178,
+ "step": 4138
+ },
+ {
+ "epoch": 57.48908296943232,
+ "grad_norm": 0.022215209901332855,
+ "learning_rate": 0.0006,
+ "loss": 4.4870452880859375,
+ "step": 4139
+ },
+ {
+ "epoch": 57.50305676855895,
+ "grad_norm": 0.01789158768951893,
+ "learning_rate": 0.0006,
+ "loss": 4.343809127807617,
+ "step": 4140
+ },
+ {
+ "epoch": 57.51703056768559,
+ "grad_norm": 0.020381739363074303,
+ "learning_rate": 0.0006,
+ "loss": 4.53312873840332,
+ "step": 4141
+ },
+ {
+ "epoch": 57.531004366812226,
+ "grad_norm": 0.019585080444812775,
+ "learning_rate": 0.0006,
+ "loss": 4.510162353515625,
+ "step": 4142
+ },
+ {
+ "epoch": 57.544978165938865,
+ "grad_norm": 0.01985127478837967,
+ "learning_rate": 0.0006,
+ "loss": 4.3654022216796875,
+ "step": 4143
+ },
+ {
+ "epoch": 57.5589519650655,
+ "grad_norm": 0.01962905190885067,
+ "learning_rate": 0.0006,
+ "loss": 4.560830116271973,
+ "step": 4144
+ },
+ {
+ "epoch": 57.57292576419214,
+ "grad_norm": 0.02679629437625408,
+ "learning_rate": 0.0006,
+ "loss": 4.300804138183594,
+ "step": 4145
+ },
+ {
+ "epoch": 57.58689956331878,
+ "grad_norm": 0.039830856025218964,
+ "learning_rate": 0.0006,
+ "loss": 4.54150915145874,
+ "step": 4146
+ },
+ {
+ "epoch": 57.60087336244541,
+ "grad_norm": 0.049527477473020554,
+ "learning_rate": 0.0006,
+ "loss": 4.405792236328125,
+ "step": 4147
+ },
+ {
+ "epoch": 57.61484716157205,
+ "grad_norm": 0.042966946959495544,
+ "learning_rate": 0.0006,
+ "loss": 4.500643730163574,
+ "step": 4148
+ },
+ {
+ "epoch": 57.62882096069869,
+ "grad_norm": 0.08999177813529968,
+ "learning_rate": 0.0006,
+ "loss": 4.504790782928467,
+ "step": 4149
+ },
+ {
+ "epoch": 57.64279475982533,
+ "grad_norm": 0.44693124294281006,
+ "learning_rate": 0.0006,
+ "loss": 4.845344066619873,
+ "step": 4150
+ },
+ {
+ "epoch": 57.65676855895197,
+ "grad_norm": 0.7143144011497498,
+ "learning_rate": 0.0006,
+ "loss": 6.907845497131348,
+ "step": 4151
+ },
+ {
+ "epoch": 57.670742358078606,
+ "grad_norm": 0.25239595770835876,
+ "learning_rate": 0.0006,
+ "loss": 7.082579612731934,
+ "step": 4152
+ },
+ {
+ "epoch": 57.68471615720524,
+ "grad_norm": 0.26550790667533875,
+ "learning_rate": 0.0006,
+ "loss": 7.50180196762085,
+ "step": 4153
+ },
+ {
+ "epoch": 57.698689956331876,
+ "grad_norm": 0.14735378324985504,
+ "learning_rate": 0.0006,
+ "loss": 7.214590072631836,
+ "step": 4154
+ },
+ {
+ "epoch": 57.712663755458514,
+ "grad_norm": 0.17685334384441376,
+ "learning_rate": 0.0006,
+ "loss": 6.861320495605469,
+ "step": 4155
+ },
+ {
+ "epoch": 57.72663755458515,
+ "grad_norm": 0.18231339752674103,
+ "learning_rate": 0.0006,
+ "loss": 6.791839599609375,
+ "step": 4156
+ },
+ {
+ "epoch": 57.74061135371179,
+ "grad_norm": 0.09230359643697739,
+ "learning_rate": 0.0006,
+ "loss": 6.760223388671875,
+ "step": 4157
+ },
+ {
+ "epoch": 57.75458515283843,
+ "grad_norm": 0.06918249279260635,
+ "learning_rate": 0.0006,
+ "loss": 6.532999515533447,
+ "step": 4158
+ },
+ {
+ "epoch": 57.76855895196506,
+ "grad_norm": 0.0901007279753685,
+ "learning_rate": 0.0006,
+ "loss": 6.52907133102417,
+ "step": 4159
+ },
+ {
+ "epoch": 57.7825327510917,
+ "grad_norm": 0.05678229779005051,
+ "learning_rate": 0.0006,
+ "loss": 6.295290470123291,
+ "step": 4160
+ },
+ {
+ "epoch": 57.79650655021834,
+ "grad_norm": 0.04450210556387901,
+ "learning_rate": 0.0006,
+ "loss": 6.14900541305542,
+ "step": 4161
+ },
+ {
+ "epoch": 57.81048034934498,
+ "grad_norm": 0.049191512167453766,
+ "learning_rate": 0.0006,
+ "loss": 6.19650411605835,
+ "step": 4162
+ },
+ {
+ "epoch": 57.82445414847162,
+ "grad_norm": 0.03908967226743698,
+ "learning_rate": 0.0006,
+ "loss": 6.054948806762695,
+ "step": 4163
+ },
+ {
+ "epoch": 57.838427947598255,
+ "grad_norm": 0.04023387283086777,
+ "learning_rate": 0.0006,
+ "loss": 6.034873962402344,
+ "step": 4164
+ },
+ {
+ "epoch": 57.852401746724894,
+ "grad_norm": 0.03304268419742584,
+ "learning_rate": 0.0006,
+ "loss": 5.769852161407471,
+ "step": 4165
+ },
+ {
+ "epoch": 57.866375545851525,
+ "grad_norm": 0.03432450816035271,
+ "learning_rate": 0.0006,
+ "loss": 5.788785457611084,
+ "step": 4166
+ },
+ {
+ "epoch": 57.880349344978164,
+ "grad_norm": 0.029965840280056,
+ "learning_rate": 0.0006,
+ "loss": 5.815243721008301,
+ "step": 4167
+ },
+ {
+ "epoch": 57.8943231441048,
+ "grad_norm": 0.03637698292732239,
+ "learning_rate": 0.0006,
+ "loss": 5.7204060554504395,
+ "step": 4168
+ },
+ {
+ "epoch": 57.90829694323144,
+ "grad_norm": 0.03477517142891884,
+ "learning_rate": 0.0006,
+ "loss": 5.605930328369141,
+ "step": 4169
+ },
+ {
+ "epoch": 57.92227074235808,
+ "grad_norm": 0.05061859264969826,
+ "learning_rate": 0.0006,
+ "loss": 5.547054767608643,
+ "step": 4170
+ },
+ {
+ "epoch": 57.93624454148472,
+ "grad_norm": 0.06299655884504318,
+ "learning_rate": 0.0006,
+ "loss": 5.552166938781738,
+ "step": 4171
+ },
+ {
+ "epoch": 57.95021834061135,
+ "grad_norm": 0.04215948283672333,
+ "learning_rate": 0.0006,
+ "loss": 5.375457763671875,
+ "step": 4172
+ },
+ {
+ "epoch": 57.96419213973799,
+ "grad_norm": 0.0348566472530365,
+ "learning_rate": 0.0006,
+ "loss": 5.418606281280518,
+ "step": 4173
+ },
+ {
+ "epoch": 57.97816593886463,
+ "grad_norm": 0.03807530924677849,
+ "learning_rate": 0.0006,
+ "loss": 5.403156280517578,
+ "step": 4174
+ },
+ {
+ "epoch": 57.992139737991266,
+ "grad_norm": 0.02989993803203106,
+ "learning_rate": 0.0006,
+ "loss": 5.355951309204102,
+ "step": 4175
+ },
+ {
+ "epoch": 58.0,
+ "grad_norm": 0.034300558269023895,
+ "learning_rate": 0.0006,
+ "loss": 5.44920539855957,
+ "step": 4176
+ },
+ {
+ "epoch": 58.0,
+ "eval_loss": 5.416731834411621,
+ "eval_runtime": 56.7934,
+ "eval_samples_per_second": 42.998,
+ "eval_steps_per_second": 1.356,
+ "step": 4176
+ },
+ {
+ "epoch": 58.01397379912664,
+ "grad_norm": 0.033961448818445206,
+ "learning_rate": 0.0006,
+ "loss": 5.250707626342773,
+ "step": 4177
+ },
+ {
+ "epoch": 58.02794759825328,
+ "grad_norm": 0.027343502268195152,
+ "learning_rate": 0.0006,
+ "loss": 5.149238109588623,
+ "step": 4178
+ },
+ {
+ "epoch": 58.041921397379916,
+ "grad_norm": 0.03033650480210781,
+ "learning_rate": 0.0006,
+ "loss": 5.072463512420654,
+ "step": 4179
+ },
+ {
+ "epoch": 58.05589519650655,
+ "grad_norm": 0.028262868523597717,
+ "learning_rate": 0.0006,
+ "loss": 5.109417915344238,
+ "step": 4180
+ },
+ {
+ "epoch": 58.069868995633186,
+ "grad_norm": 0.02373034693300724,
+ "learning_rate": 0.0006,
+ "loss": 5.082464218139648,
+ "step": 4181
+ },
+ {
+ "epoch": 58.083842794759825,
+ "grad_norm": 0.025879688560962677,
+ "learning_rate": 0.0006,
+ "loss": 5.056792259216309,
+ "step": 4182
+ },
+ {
+ "epoch": 58.09781659388646,
+ "grad_norm": 0.026252275332808495,
+ "learning_rate": 0.0006,
+ "loss": 5.064239978790283,
+ "step": 4183
+ },
+ {
+ "epoch": 58.1117903930131,
+ "grad_norm": 0.028177792206406593,
+ "learning_rate": 0.0006,
+ "loss": 4.922310829162598,
+ "step": 4184
+ },
+ {
+ "epoch": 58.12576419213974,
+ "grad_norm": 0.02630820870399475,
+ "learning_rate": 0.0006,
+ "loss": 4.992884635925293,
+ "step": 4185
+ },
+ {
+ "epoch": 58.13973799126637,
+ "grad_norm": 0.024890903383493423,
+ "learning_rate": 0.0006,
+ "loss": 4.910036563873291,
+ "step": 4186
+ },
+ {
+ "epoch": 58.15371179039301,
+ "grad_norm": 0.02822992391884327,
+ "learning_rate": 0.0006,
+ "loss": 4.866640090942383,
+ "step": 4187
+ },
+ {
+ "epoch": 58.16768558951965,
+ "grad_norm": 0.03253607079386711,
+ "learning_rate": 0.0006,
+ "loss": 4.9112958908081055,
+ "step": 4188
+ },
+ {
+ "epoch": 58.18165938864629,
+ "grad_norm": 0.03765944764018059,
+ "learning_rate": 0.0006,
+ "loss": 4.774468421936035,
+ "step": 4189
+ },
+ {
+ "epoch": 58.19563318777293,
+ "grad_norm": 0.039874881505966187,
+ "learning_rate": 0.0006,
+ "loss": 4.831678867340088,
+ "step": 4190
+ },
+ {
+ "epoch": 58.209606986899566,
+ "grad_norm": 0.034595925360918045,
+ "learning_rate": 0.0006,
+ "loss": 4.8243794441223145,
+ "step": 4191
+ },
+ {
+ "epoch": 58.223580786026204,
+ "grad_norm": 0.04840189591050148,
+ "learning_rate": 0.0006,
+ "loss": 4.854156494140625,
+ "step": 4192
+ },
+ {
+ "epoch": 58.237554585152836,
+ "grad_norm": 0.06358753889799118,
+ "learning_rate": 0.0006,
+ "loss": 4.836430549621582,
+ "step": 4193
+ },
+ {
+ "epoch": 58.251528384279474,
+ "grad_norm": 0.0507555715739727,
+ "learning_rate": 0.0006,
+ "loss": 4.88357400894165,
+ "step": 4194
+ },
+ {
+ "epoch": 58.26550218340611,
+ "grad_norm": 0.04384690150618553,
+ "learning_rate": 0.0006,
+ "loss": 4.774049282073975,
+ "step": 4195
+ },
+ {
+ "epoch": 58.27947598253275,
+ "grad_norm": 0.03246928006410599,
+ "learning_rate": 0.0006,
+ "loss": 4.686605930328369,
+ "step": 4196
+ },
+ {
+ "epoch": 58.29344978165939,
+ "grad_norm": 0.028878789395093918,
+ "learning_rate": 0.0006,
+ "loss": 4.796070575714111,
+ "step": 4197
+ },
+ {
+ "epoch": 58.30742358078603,
+ "grad_norm": 0.028175361454486847,
+ "learning_rate": 0.0006,
+ "loss": 4.728490829467773,
+ "step": 4198
+ },
+ {
+ "epoch": 58.32139737991266,
+ "grad_norm": 0.029428014531731606,
+ "learning_rate": 0.0006,
+ "loss": 4.7159624099731445,
+ "step": 4199
+ },
+ {
+ "epoch": 58.3353711790393,
+ "grad_norm": 0.031227827072143555,
+ "learning_rate": 0.0006,
+ "loss": 4.702625274658203,
+ "step": 4200
+ },
+ {
+ "epoch": 58.34934497816594,
+ "grad_norm": 0.028843428939580917,
+ "learning_rate": 0.0006,
+ "loss": 4.811799049377441,
+ "step": 4201
+ },
+ {
+ "epoch": 58.36331877729258,
+ "grad_norm": 0.026780391111969948,
+ "learning_rate": 0.0006,
+ "loss": 4.81662654876709,
+ "step": 4202
+ },
+ {
+ "epoch": 58.377292576419215,
+ "grad_norm": 0.019699757918715477,
+ "learning_rate": 0.0006,
+ "loss": 4.713685989379883,
+ "step": 4203
+ },
+ {
+ "epoch": 58.391266375545854,
+ "grad_norm": 0.024389250203967094,
+ "learning_rate": 0.0006,
+ "loss": 4.72354793548584,
+ "step": 4204
+ },
+ {
+ "epoch": 58.40524017467249,
+ "grad_norm": 0.02218274027109146,
+ "learning_rate": 0.0006,
+ "loss": 4.674125671386719,
+ "step": 4205
+ },
+ {
+ "epoch": 58.419213973799124,
+ "grad_norm": 0.02041488140821457,
+ "learning_rate": 0.0006,
+ "loss": 4.7909698486328125,
+ "step": 4206
+ },
+ {
+ "epoch": 58.43318777292576,
+ "grad_norm": 0.018561935052275658,
+ "learning_rate": 0.0006,
+ "loss": 4.808770179748535,
+ "step": 4207
+ },
+ {
+ "epoch": 58.4471615720524,
+ "grad_norm": 0.01708846725523472,
+ "learning_rate": 0.0006,
+ "loss": 4.701626777648926,
+ "step": 4208
+ },
+ {
+ "epoch": 58.46113537117904,
+ "grad_norm": 0.0222734697163105,
+ "learning_rate": 0.0006,
+ "loss": 4.505410194396973,
+ "step": 4209
+ },
+ {
+ "epoch": 58.47510917030568,
+ "grad_norm": 0.03197057917714119,
+ "learning_rate": 0.0006,
+ "loss": 4.610360145568848,
+ "step": 4210
+ },
+ {
+ "epoch": 58.48908296943232,
+ "grad_norm": 0.06854520738124847,
+ "learning_rate": 0.0006,
+ "loss": 4.625649452209473,
+ "step": 4211
+ },
+ {
+ "epoch": 58.50305676855895,
+ "grad_norm": 0.09981521219015121,
+ "learning_rate": 0.0006,
+ "loss": 4.832150459289551,
+ "step": 4212
+ },
+ {
+ "epoch": 58.51703056768559,
+ "grad_norm": 0.046356040984392166,
+ "learning_rate": 0.0006,
+ "loss": 4.5641632080078125,
+ "step": 4213
+ },
+ {
+ "epoch": 58.531004366812226,
+ "grad_norm": 0.03882667422294617,
+ "learning_rate": 0.0006,
+ "loss": 4.666974067687988,
+ "step": 4214
+ },
+ {
+ "epoch": 58.544978165938865,
+ "grad_norm": 0.033203233033418655,
+ "learning_rate": 0.0006,
+ "loss": 4.629413604736328,
+ "step": 4215
+ },
+ {
+ "epoch": 58.5589519650655,
+ "grad_norm": 0.02806735597550869,
+ "learning_rate": 0.0006,
+ "loss": 4.763245582580566,
+ "step": 4216
+ },
+ {
+ "epoch": 58.57292576419214,
+ "grad_norm": 0.026728278025984764,
+ "learning_rate": 0.0006,
+ "loss": 4.629597187042236,
+ "step": 4217
+ },
+ {
+ "epoch": 58.58689956331878,
+ "grad_norm": 0.024380959570407867,
+ "learning_rate": 0.0006,
+ "loss": 4.633614540100098,
+ "step": 4218
+ },
+ {
+ "epoch": 58.60087336244541,
+ "grad_norm": 0.022312544286251068,
+ "learning_rate": 0.0006,
+ "loss": 4.671379566192627,
+ "step": 4219
+ },
+ {
+ "epoch": 58.61484716157205,
+ "grad_norm": 0.022049568593502045,
+ "learning_rate": 0.0006,
+ "loss": 4.661557197570801,
+ "step": 4220
+ },
+ {
+ "epoch": 58.62882096069869,
+ "grad_norm": 0.02119818702340126,
+ "learning_rate": 0.0006,
+ "loss": 4.688562393188477,
+ "step": 4221
+ },
+ {
+ "epoch": 58.64279475982533,
+ "grad_norm": 0.019769301638007164,
+ "learning_rate": 0.0006,
+ "loss": 4.607748031616211,
+ "step": 4222
+ },
+ {
+ "epoch": 58.65676855895197,
+ "grad_norm": 0.02124079130589962,
+ "learning_rate": 0.0006,
+ "loss": 4.570640563964844,
+ "step": 4223
+ },
+ {
+ "epoch": 58.670742358078606,
+ "grad_norm": 0.016344040632247925,
+ "learning_rate": 0.0006,
+ "loss": 4.676113128662109,
+ "step": 4224
+ },
+ {
+ "epoch": 58.68471615720524,
+ "grad_norm": 0.01735256239771843,
+ "learning_rate": 0.0006,
+ "loss": 4.62088680267334,
+ "step": 4225
+ },
+ {
+ "epoch": 58.698689956331876,
+ "grad_norm": 0.019323458895087242,
+ "learning_rate": 0.0006,
+ "loss": 4.579405307769775,
+ "step": 4226
+ },
+ {
+ "epoch": 58.712663755458514,
+ "grad_norm": 0.01760544814169407,
+ "learning_rate": 0.0006,
+ "loss": 4.625407695770264,
+ "step": 4227
+ },
+ {
+ "epoch": 58.72663755458515,
+ "grad_norm": 0.015593250282108784,
+ "learning_rate": 0.0006,
+ "loss": 4.6326117515563965,
+ "step": 4228
+ },
+ {
+ "epoch": 58.74061135371179,
+ "grad_norm": 0.014357523061335087,
+ "learning_rate": 0.0006,
+ "loss": 4.6869401931762695,
+ "step": 4229
+ },
+ {
+ "epoch": 58.75458515283843,
+ "grad_norm": 0.01604357920587063,
+ "learning_rate": 0.0006,
+ "loss": 4.690784931182861,
+ "step": 4230
+ },
+ {
+ "epoch": 58.76855895196506,
+ "grad_norm": 0.01673845760524273,
+ "learning_rate": 0.0006,
+ "loss": 4.513410568237305,
+ "step": 4231
+ },
+ {
+ "epoch": 58.7825327510917,
+ "grad_norm": 0.013657779432833195,
+ "learning_rate": 0.0006,
+ "loss": 4.637057781219482,
+ "step": 4232
+ },
+ {
+ "epoch": 58.79650655021834,
+ "grad_norm": 0.013373114168643951,
+ "learning_rate": 0.0006,
+ "loss": 4.719967842102051,
+ "step": 4233
+ },
+ {
+ "epoch": 58.81048034934498,
+ "grad_norm": 0.016345568001270294,
+ "learning_rate": 0.0006,
+ "loss": 4.604109764099121,
+ "step": 4234
+ },
+ {
+ "epoch": 58.82445414847162,
+ "grad_norm": 0.01573033444583416,
+ "learning_rate": 0.0006,
+ "loss": 4.6471428871154785,
+ "step": 4235
+ },
+ {
+ "epoch": 58.838427947598255,
+ "grad_norm": 0.01517449039965868,
+ "learning_rate": 0.0006,
+ "loss": 4.722684860229492,
+ "step": 4236
+ },
+ {
+ "epoch": 58.852401746724894,
+ "grad_norm": 0.0145102022215724,
+ "learning_rate": 0.0006,
+ "loss": 4.662420272827148,
+ "step": 4237
+ },
+ {
+ "epoch": 58.866375545851525,
+ "grad_norm": 0.01457217987626791,
+ "learning_rate": 0.0006,
+ "loss": 4.585169792175293,
+ "step": 4238
+ },
+ {
+ "epoch": 58.880349344978164,
+ "grad_norm": 0.01406773366034031,
+ "learning_rate": 0.0006,
+ "loss": 4.5413665771484375,
+ "step": 4239
+ },
+ {
+ "epoch": 58.8943231441048,
+ "grad_norm": 0.012653871439397335,
+ "learning_rate": 0.0006,
+ "loss": 4.553636074066162,
+ "step": 4240
+ },
+ {
+ "epoch": 58.90829694323144,
+ "grad_norm": 0.01261200476437807,
+ "learning_rate": 0.0006,
+ "loss": 4.534193992614746,
+ "step": 4241
+ },
+ {
+ "epoch": 58.92227074235808,
+ "grad_norm": 0.012955864891409874,
+ "learning_rate": 0.0006,
+ "loss": 4.697248458862305,
+ "step": 4242
+ },
+ {
+ "epoch": 58.93624454148472,
+ "grad_norm": 0.012134749442338943,
+ "learning_rate": 0.0006,
+ "loss": 4.557095527648926,
+ "step": 4243
+ },
+ {
+ "epoch": 58.95021834061135,
+ "grad_norm": 0.012164677493274212,
+ "learning_rate": 0.0006,
+ "loss": 4.623579025268555,
+ "step": 4244
+ },
+ {
+ "epoch": 58.96419213973799,
+ "grad_norm": 0.013802947476506233,
+ "learning_rate": 0.0006,
+ "loss": 4.562838077545166,
+ "step": 4245
+ },
+ {
+ "epoch": 58.97816593886463,
+ "grad_norm": 0.011420476250350475,
+ "learning_rate": 0.0006,
+ "loss": 4.62265682220459,
+ "step": 4246
+ },
+ {
+ "epoch": 58.992139737991266,
+ "grad_norm": 0.01169576682150364,
+ "learning_rate": 0.0006,
+ "loss": 4.727449417114258,
+ "step": 4247
+ },
+ {
+ "epoch": 59.0,
+ "grad_norm": 0.013577710837125778,
+ "learning_rate": 0.0006,
+ "loss": 4.536325454711914,
+ "step": 4248
+ },
+ {
+ "epoch": 59.0,
+ "eval_loss": 4.86527681350708,
+ "eval_runtime": 60.4463,
+ "eval_samples_per_second": 40.399,
+ "eval_steps_per_second": 1.274,
+ "step": 4248
+ },
+ {
+ "epoch": 59.01397379912664,
+ "grad_norm": 0.015251655131578445,
+ "learning_rate": 0.0006,
+ "loss": 4.634598731994629,
+ "step": 4249
+ },
+ {
+ "epoch": 59.02794759825328,
+ "grad_norm": 0.017414938658475876,
+ "learning_rate": 0.0006,
+ "loss": 4.521329879760742,
+ "step": 4250
+ },
+ {
+ "epoch": 59.041921397379916,
+ "grad_norm": 0.01656588353216648,
+ "learning_rate": 0.0006,
+ "loss": 4.446192264556885,
+ "step": 4251
+ },
+ {
+ "epoch": 59.05589519650655,
+ "grad_norm": 0.012038925662636757,
+ "learning_rate": 0.0006,
+ "loss": 4.523855209350586,
+ "step": 4252
+ },
+ {
+ "epoch": 59.069868995633186,
+ "grad_norm": 0.014403033070266247,
+ "learning_rate": 0.0006,
+ "loss": 4.561079978942871,
+ "step": 4253
+ },
+ {
+ "epoch": 59.083842794759825,
+ "grad_norm": 0.016243597492575645,
+ "learning_rate": 0.0006,
+ "loss": 4.673602104187012,
+ "step": 4254
+ },
+ {
+ "epoch": 59.09781659388646,
+ "grad_norm": 0.016492877155542374,
+ "learning_rate": 0.0006,
+ "loss": 4.564824104309082,
+ "step": 4255
+ },
+ {
+ "epoch": 59.1117903930131,
+ "grad_norm": 0.015980370342731476,
+ "learning_rate": 0.0006,
+ "loss": 4.339249134063721,
+ "step": 4256
+ },
+ {
+ "epoch": 59.12576419213974,
+ "grad_norm": 0.019687356427311897,
+ "learning_rate": 0.0006,
+ "loss": 4.481834888458252,
+ "step": 4257
+ },
+ {
+ "epoch": 59.13973799126637,
+ "grad_norm": 0.02983798086643219,
+ "learning_rate": 0.0006,
+ "loss": 4.4495697021484375,
+ "step": 4258
+ },
+ {
+ "epoch": 59.15371179039301,
+ "grad_norm": 0.04600981995463371,
+ "learning_rate": 0.0006,
+ "loss": 4.567838668823242,
+ "step": 4259
+ },
+ {
+ "epoch": 59.16768558951965,
+ "grad_norm": 0.04852001741528511,
+ "learning_rate": 0.0006,
+ "loss": 4.594701766967773,
+ "step": 4260
+ },
+ {
+ "epoch": 59.18165938864629,
+ "grad_norm": 0.03429504856467247,
+ "learning_rate": 0.0006,
+ "loss": 4.573735237121582,
+ "step": 4261
+ },
+ {
+ "epoch": 59.19563318777293,
+ "grad_norm": 0.029345160350203514,
+ "learning_rate": 0.0006,
+ "loss": 4.488258361816406,
+ "step": 4262
+ },
+ {
+ "epoch": 59.209606986899566,
+ "grad_norm": 0.01961623504757881,
+ "learning_rate": 0.0006,
+ "loss": 4.545222282409668,
+ "step": 4263
+ },
+ {
+ "epoch": 59.223580786026204,
+ "grad_norm": 0.020406130701303482,
+ "learning_rate": 0.0006,
+ "loss": 4.471898078918457,
+ "step": 4264
+ },
+ {
+ "epoch": 59.237554585152836,
+ "grad_norm": 0.020500272512435913,
+ "learning_rate": 0.0006,
+ "loss": 4.538337707519531,
+ "step": 4265
+ },
+ {
+ "epoch": 59.251528384279474,
+ "grad_norm": 0.020727017894387245,
+ "learning_rate": 0.0006,
+ "loss": 4.437284469604492,
+ "step": 4266
+ },
+ {
+ "epoch": 59.26550218340611,
+ "grad_norm": 0.03395998477935791,
+ "learning_rate": 0.0006,
+ "loss": 4.459124565124512,
+ "step": 4267
+ },
+ {
+ "epoch": 59.27947598253275,
+ "grad_norm": 0.04635605961084366,
+ "learning_rate": 0.0006,
+ "loss": 4.520228385925293,
+ "step": 4268
+ },
+ {
+ "epoch": 59.29344978165939,
+ "grad_norm": 0.04253006726503372,
+ "learning_rate": 0.0006,
+ "loss": 4.578673362731934,
+ "step": 4269
+ },
+ {
+ "epoch": 59.30742358078603,
+ "grad_norm": 0.02403509058058262,
+ "learning_rate": 0.0006,
+ "loss": 4.484639644622803,
+ "step": 4270
+ },
+ {
+ "epoch": 59.32139737991266,
+ "grad_norm": 0.02254190482199192,
+ "learning_rate": 0.0006,
+ "loss": 4.511314868927002,
+ "step": 4271
+ },
+ {
+ "epoch": 59.3353711790393,
+ "grad_norm": 0.021585190668702126,
+ "learning_rate": 0.0006,
+ "loss": 4.530239105224609,
+ "step": 4272
+ },
+ {
+ "epoch": 59.34934497816594,
+ "grad_norm": 0.021210316568613052,
+ "learning_rate": 0.0006,
+ "loss": 4.593514442443848,
+ "step": 4273
+ },
+ {
+ "epoch": 59.36331877729258,
+ "grad_norm": 0.019317107275128365,
+ "learning_rate": 0.0006,
+ "loss": 4.5591535568237305,
+ "step": 4274
+ },
+ {
+ "epoch": 59.377292576419215,
+ "grad_norm": 0.018059708178043365,
+ "learning_rate": 0.0006,
+ "loss": 4.4499592781066895,
+ "step": 4275
+ },
+ {
+ "epoch": 59.391266375545854,
+ "grad_norm": 0.017847446724772453,
+ "learning_rate": 0.0006,
+ "loss": 4.473999500274658,
+ "step": 4276
+ },
+ {
+ "epoch": 59.40524017467249,
+ "grad_norm": 0.014944756403565407,
+ "learning_rate": 0.0006,
+ "loss": 4.5934529304504395,
+ "step": 4277
+ },
+ {
+ "epoch": 59.419213973799124,
+ "grad_norm": 0.01663832925260067,
+ "learning_rate": 0.0006,
+ "loss": 4.504262924194336,
+ "step": 4278
+ },
+ {
+ "epoch": 59.43318777292576,
+ "grad_norm": 0.015571796335279942,
+ "learning_rate": 0.0006,
+ "loss": 4.512310981750488,
+ "step": 4279
+ },
+ {
+ "epoch": 59.4471615720524,
+ "grad_norm": 0.015896400436758995,
+ "learning_rate": 0.0006,
+ "loss": 4.452714920043945,
+ "step": 4280
+ },
+ {
+ "epoch": 59.46113537117904,
+ "grad_norm": 0.016010284423828125,
+ "learning_rate": 0.0006,
+ "loss": 4.6188063621521,
+ "step": 4281
+ },
+ {
+ "epoch": 59.47510917030568,
+ "grad_norm": 0.015075593255460262,
+ "learning_rate": 0.0006,
+ "loss": 4.477978229522705,
+ "step": 4282
+ },
+ {
+ "epoch": 59.48908296943232,
+ "grad_norm": 0.014698950573801994,
+ "learning_rate": 0.0006,
+ "loss": 4.544955730438232,
+ "step": 4283
+ },
+ {
+ "epoch": 59.50305676855895,
+ "grad_norm": 0.0136245833709836,
+ "learning_rate": 0.0006,
+ "loss": 4.530460357666016,
+ "step": 4284
+ },
+ {
+ "epoch": 59.51703056768559,
+ "grad_norm": 0.013118310831487179,
+ "learning_rate": 0.0006,
+ "loss": 4.614217758178711,
+ "step": 4285
+ },
+ {
+ "epoch": 59.531004366812226,
+ "grad_norm": 0.01295961532741785,
+ "learning_rate": 0.0006,
+ "loss": 4.43156099319458,
+ "step": 4286
+ },
+ {
+ "epoch": 59.544978165938865,
+ "grad_norm": 0.012429811991751194,
+ "learning_rate": 0.0006,
+ "loss": 4.517333030700684,
+ "step": 4287
+ },
+ {
+ "epoch": 59.5589519650655,
+ "grad_norm": 0.01261830423027277,
+ "learning_rate": 0.0006,
+ "loss": 4.569780349731445,
+ "step": 4288
+ },
+ {
+ "epoch": 59.57292576419214,
+ "grad_norm": 0.012394499965012074,
+ "learning_rate": 0.0006,
+ "loss": 4.5970869064331055,
+ "step": 4289
+ },
+ {
+ "epoch": 59.58689956331878,
+ "grad_norm": 0.01290759164839983,
+ "learning_rate": 0.0006,
+ "loss": 4.437100410461426,
+ "step": 4290
+ },
+ {
+ "epoch": 59.60087336244541,
+ "grad_norm": 0.012042169459164143,
+ "learning_rate": 0.0006,
+ "loss": 4.567404747009277,
+ "step": 4291
+ },
+ {
+ "epoch": 59.61484716157205,
+ "grad_norm": 0.012338520959019661,
+ "learning_rate": 0.0006,
+ "loss": 4.476193428039551,
+ "step": 4292
+ },
+ {
+ "epoch": 59.62882096069869,
+ "grad_norm": 0.01321258582174778,
+ "learning_rate": 0.0006,
+ "loss": 4.566680431365967,
+ "step": 4293
+ },
+ {
+ "epoch": 59.64279475982533,
+ "grad_norm": 0.013478122651576996,
+ "learning_rate": 0.0006,
+ "loss": 4.487398147583008,
+ "step": 4294
+ },
+ {
+ "epoch": 59.65676855895197,
+ "grad_norm": 0.01537603884935379,
+ "learning_rate": 0.0006,
+ "loss": 4.449337005615234,
+ "step": 4295
+ },
+ {
+ "epoch": 59.670742358078606,
+ "grad_norm": 0.017140565440058708,
+ "learning_rate": 0.0006,
+ "loss": 4.5174360275268555,
+ "step": 4296
+ },
+ {
+ "epoch": 59.68471615720524,
+ "grad_norm": 0.017931461334228516,
+ "learning_rate": 0.0006,
+ "loss": 4.522104740142822,
+ "step": 4297
+ },
+ {
+ "epoch": 59.698689956331876,
+ "grad_norm": 0.018037918955087662,
+ "learning_rate": 0.0006,
+ "loss": 4.496928691864014,
+ "step": 4298
+ },
+ {
+ "epoch": 59.712663755458514,
+ "grad_norm": 0.01684504561126232,
+ "learning_rate": 0.0006,
+ "loss": 4.5332417488098145,
+ "step": 4299
+ },
+ {
+ "epoch": 59.72663755458515,
+ "grad_norm": 0.016268383711576462,
+ "learning_rate": 0.0006,
+ "loss": 4.468893051147461,
+ "step": 4300
+ },
+ {
+ "epoch": 59.74061135371179,
+ "grad_norm": 0.014859385788440704,
+ "learning_rate": 0.0006,
+ "loss": 4.551092147827148,
+ "step": 4301
+ },
+ {
+ "epoch": 59.75458515283843,
+ "grad_norm": 0.013589047826826572,
+ "learning_rate": 0.0006,
+ "loss": 4.470579147338867,
+ "step": 4302
+ },
+ {
+ "epoch": 59.76855895196506,
+ "grad_norm": 0.015704551711678505,
+ "learning_rate": 0.0006,
+ "loss": 4.330068111419678,
+ "step": 4303
+ },
+ {
+ "epoch": 59.7825327510917,
+ "grad_norm": 0.017869921401143074,
+ "learning_rate": 0.0006,
+ "loss": 4.428255081176758,
+ "step": 4304
+ },
+ {
+ "epoch": 59.79650655021834,
+ "grad_norm": 0.019828537479043007,
+ "learning_rate": 0.0006,
+ "loss": 4.496023178100586,
+ "step": 4305
+ },
+ {
+ "epoch": 59.81048034934498,
+ "grad_norm": 0.015856236219406128,
+ "learning_rate": 0.0006,
+ "loss": 4.427152156829834,
+ "step": 4306
+ },
+ {
+ "epoch": 59.82445414847162,
+ "grad_norm": 0.013349304907023907,
+ "learning_rate": 0.0006,
+ "loss": 4.417333602905273,
+ "step": 4307
+ },
+ {
+ "epoch": 59.838427947598255,
+ "grad_norm": 0.014111637137830257,
+ "learning_rate": 0.0006,
+ "loss": 4.4507575035095215,
+ "step": 4308
+ },
+ {
+ "epoch": 59.852401746724894,
+ "grad_norm": 0.015186597593128681,
+ "learning_rate": 0.0006,
+ "loss": 4.375116348266602,
+ "step": 4309
+ },
+ {
+ "epoch": 59.866375545851525,
+ "grad_norm": 0.01604730449616909,
+ "learning_rate": 0.0006,
+ "loss": 4.604518890380859,
+ "step": 4310
+ },
+ {
+ "epoch": 59.880349344978164,
+ "grad_norm": 0.015308464877307415,
+ "learning_rate": 0.0006,
+ "loss": 4.507232189178467,
+ "step": 4311
+ },
+ {
+ "epoch": 59.8943231441048,
+ "grad_norm": 0.014017206616699696,
+ "learning_rate": 0.0006,
+ "loss": 4.398487091064453,
+ "step": 4312
+ },
+ {
+ "epoch": 59.90829694323144,
+ "grad_norm": 0.012911394238471985,
+ "learning_rate": 0.0006,
+ "loss": 4.567872047424316,
+ "step": 4313
+ },
+ {
+ "epoch": 59.92227074235808,
+ "grad_norm": 0.012775142677128315,
+ "learning_rate": 0.0006,
+ "loss": 4.562105655670166,
+ "step": 4314
+ },
+ {
+ "epoch": 59.93624454148472,
+ "grad_norm": 0.014750408008694649,
+ "learning_rate": 0.0006,
+ "loss": 4.456263542175293,
+ "step": 4315
+ },
+ {
+ "epoch": 59.95021834061135,
+ "grad_norm": 0.016726011410355568,
+ "learning_rate": 0.0006,
+ "loss": 4.360722064971924,
+ "step": 4316
+ },
+ {
+ "epoch": 59.96419213973799,
+ "grad_norm": 0.01734776981174946,
+ "learning_rate": 0.0006,
+ "loss": 4.572544574737549,
+ "step": 4317
+ },
+ {
+ "epoch": 59.97816593886463,
+ "grad_norm": 0.01822415366768837,
+ "learning_rate": 0.0006,
+ "loss": 4.490501880645752,
+ "step": 4318
+ },
+ {
+ "epoch": 59.992139737991266,
+ "grad_norm": 0.014638577587902546,
+ "learning_rate": 0.0006,
+ "loss": 4.440951824188232,
+ "step": 4319
+ },
+ {
+ "epoch": 60.0,
+ "grad_norm": 0.01255202479660511,
+ "learning_rate": 0.0006,
+ "loss": 4.559123516082764,
+ "step": 4320
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 28800,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 400,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": true,
+ "should_log": false,
+ "should_save": false,
+ "should_training_stop": false
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 1.8417680167993344e+18,
+ "train_batch_size": 8,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4320/training_args.bin b/runs/l2r50-i2-fulle-lm/checkpoint-4320/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..cad232831c222cc7f8b938b8bbe42f61e7d89c3a
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-4320/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:f7c40e3bfa2f5ac1450ca1243e3f2a895d44a06ed5995e6aefd97f78d229d8c8
+size 4792
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4608/chat_template.jinja b/runs/l2r50-i2-fulle-lm/checkpoint-4608/chat_template.jinja
new file mode 100644
index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-4608/chat_template.jinja
@@ -0,0 +1,85 @@
+{%- if tools %}
+ {{- '<|im_start|>system\n' }}
+ {%- if messages[0].role == 'system' %}
+ {{- messages[0].content + '\n\n' }}
+ {%- endif %}
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }}
+ {%- for tool in tools %}
+ {{- "\n" }}
+ {{- tool | tojson }}
+ {%- endfor %}
+ {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }}
+{%- else %}
+ {%- if messages[0].role == 'system' %}
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
+ {%- endif %}
+{%- endif %}
+{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
+{%- for message in messages[::-1] %}
+ {%- set index = (messages|length - 1) - loop.index0 %}
+ {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %}
+ {%- set ns.multi_step_tool = false %}
+ {%- set ns.last_query_index = index %}
+ {%- endif %}
+{%- endfor %}
+{%- for message in messages %}
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
+ {%- elif message.role == "assistant" %}
+ {%- set content = message.content %}
+ {%- set reasoning_content = '' %}
+ {%- if message.reasoning_content is defined and message.reasoning_content is not none %}
+ {%- set reasoning_content = message.reasoning_content %}
+ {%- else %}
+ {%- if '' in message.content %}
+ {%- set content = message.content.split('')[-1].lstrip('\n') %}
+ {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %}
+ {%- endif %}
+ {%- endif %}
+ {%- if loop.index0 > ns.last_query_index %}
+ {%- if loop.last or (not loop.last and reasoning_content) %}
+ {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- if message.tool_calls %}
+ {%- for tool_call in message.tool_calls %}
+ {%- if (loop.first and content) or (not loop.first) %}
+ {{- '\n' }}
+ {%- endif %}
+ {%- if tool_call.function %}
+ {%- set tool_call = tool_call.function %}
+ {%- endif %}
+ {{- '\n{"name": "' }}
+ {{- tool_call.name }}
+ {{- '", "arguments": ' }}
+ {%- if tool_call.arguments is string %}
+ {{- tool_call.arguments }}
+ {%- else %}
+ {{- tool_call.arguments | tojson }}
+ {%- endif %}
+ {{- '}\n' }}
+ {%- endfor %}
+ {%- endif %}
+ {{- '<|im_end|>\n' }}
+ {%- elif message.role == "tool" %}
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
+ {{- '<|im_start|>user' }}
+ {%- endif %}
+ {{- '\n\n' }}
+ {{- message.content }}
+ {{- '\n' }}
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
+ {{- '<|im_end|>\n' }}
+ {%- endif %}
+ {%- endif %}
+{%- endfor %}
+{%- if add_generation_prompt %}
+ {{- '<|im_start|>assistant\n' }}
+ {%- if enable_thinking is defined and enable_thinking is false %}
+ {{- '\n\n\n\n' }}
+ {%- endif %}
+{%- endif %}
\ No newline at end of file
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4608/config.json b/runs/l2r50-i2-fulle-lm/checkpoint-4608/config.json
new file mode 100644
index 0000000000000000000000000000000000000000..22f250953c6ce8fc91ee4411856dd73041a415e2
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-4608/config.json
@@ -0,0 +1,32 @@
+{
+ "architectures": [
+ "LlamaForCausalLM"
+ ],
+ "attention_bias": false,
+ "attention_dropout": 0.0,
+ "bos_token_id": null,
+ "dtype": "float32",
+ "eos_token_id": 151645,
+ "head_dim": 128,
+ "hidden_act": "silu",
+ "hidden_size": 512,
+ "initializer_range": 0.02,
+ "intermediate_size": 1536,
+ "max_position_embeddings": 2048,
+ "mlp_bias": false,
+ "model_type": "llama",
+ "num_attention_heads": 4,
+ "num_hidden_layers": 20,
+ "num_key_value_heads": 4,
+ "pad_token_id": 151645,
+ "pretraining_tp": 1,
+ "rms_norm_eps": 1e-06,
+ "rope_parameters": {
+ "rope_theta": 10000.0,
+ "rope_type": "default"
+ },
+ "tie_word_embeddings": true,
+ "transformers_version": "5.5.0",
+ "use_cache": false,
+ "vocab_size": 151673
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4608/generation_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-4608/generation_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-4608/generation_config.json
@@ -0,0 +1,11 @@
+{
+ "_from_model_config": true,
+ "eos_token_id": [
+ 151645
+ ],
+ "output_attentions": false,
+ "output_hidden_states": false,
+ "pad_token_id": 151645,
+ "transformers_version": "5.5.0",
+ "use_cache": true
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4608/model.safetensors b/runs/l2r50-i2-fulle-lm/checkpoint-4608/model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..a4e8790d4cd6c8f19a0256a5e946ed4d1a94109c
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-4608/model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:b64b7e8382655d170b13d1a6577e69b6f8470911a15d5110093e386bff274fab
+size 583360328
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4608/optimizer.pt b/runs/l2r50-i2-fulle-lm/checkpoint-4608/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..bf6e229277fa7ecf83e54904a5d52a400fac0f4b
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-4608/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:257a71904ee506eea0c7c4afbf5510ffc96cd430a8f34d09c3552bb3aaa7c6ea
+size 1166833530
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4608/rng_state_0.pth b/runs/l2r50-i2-fulle-lm/checkpoint-4608/rng_state_0.pth
new file mode 100644
index 0000000000000000000000000000000000000000..f91567317ab848c2dff078899eb662dd51945bae
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-4608/rng_state_0.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:82bad169f67148d86bef2a656e092bfc17814ad929da77d19575a4e4d385d43c
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4608/rng_state_1.pth b/runs/l2r50-i2-fulle-lm/checkpoint-4608/rng_state_1.pth
new file mode 100644
index 0000000000000000000000000000000000000000..ed61f380a617838b1ac7d8ff8c039ec4d4cbe5c6
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-4608/rng_state_1.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:9d116daf8954ab7a831f7098c4b5c2ad6a178ca829b12cf6ed99988af6804c5b
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4608/rng_state_2.pth b/runs/l2r50-i2-fulle-lm/checkpoint-4608/rng_state_2.pth
new file mode 100644
index 0000000000000000000000000000000000000000..d68e42e69f3b5f81be555c974a3b3df909c2ea1d
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-4608/rng_state_2.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:d320b2fa71fd92a501be41c393f5297bdc9d88e31c0a36dfc68fb94a01290b14
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4608/rng_state_3.pth b/runs/l2r50-i2-fulle-lm/checkpoint-4608/rng_state_3.pth
new file mode 100644
index 0000000000000000000000000000000000000000..ed2916c2a4e214438982b7fcbcb279f2806549e2
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-4608/rng_state_3.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:56f75bbaa62722ad1b7a78bdf0b5790f29b39f7cbeaf0c2d6fc2de4d46c7b867
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4608/scheduler.pt b/runs/l2r50-i2-fulle-lm/checkpoint-4608/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..e577101181b24dfcc9b18d0a62144d06cb4b37c2
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-4608/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:8d97e4c65a4c8c2376a5585ec88bf61c1243526bb687aa9ffb5e9d7512df2daa
+size 1064
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4608/tokenizer.json b/runs/l2r50-i2-fulle-lm/checkpoint-4608/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..53786016f070e11184813f3c83101dc5e93e392e
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-4608/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:c3c63fda1646a555448e8f03662542e04b11d103721319e4d9b267d7da03ceb8
+size 11424463
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4608/tokenizer_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-4608/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..98e2fb9b2cc54785dc9d871b8e503f8b8f2defe7
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-4608/tokenizer_config.json
@@ -0,0 +1,20 @@
+{
+ "add_prefix_space": false,
+ "backend": "tokenizers",
+ "bos_token": null,
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "<|im_end|>",
+ "errors": "replace",
+ "extra_special_tokens": [
+ "<|l2r_pred|>",
+ "<|r2l_pred|>",
+ "<|next_1_pred|>",
+ "<|next_2_pred|>"
+ ],
+ "is_local": false,
+ "model_max_length": 131072,
+ "pad_token": "<|im_end|>",
+ "split_special_tokens": false,
+ "tokenizer_class": "Qwen2Tokenizer",
+ "unk_token": null
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4608/trainer_state.json b/runs/l2r50-i2-fulle-lm/checkpoint-4608/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..46f1e702d3ce871749b48bd1c1214a16fa5388a1
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-4608/trainer_state.json
@@ -0,0 +1,32794 @@
+{
+ "best_global_step": null,
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 64.0,
+ "eval_steps": 500,
+ "global_step": 4608,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.013973799126637555,
+ "grad_norm": 0.1739506870508194,
+ "learning_rate": 0.0,
+ "loss": 11.99299430847168,
+ "step": 1
+ },
+ {
+ "epoch": 0.02794759825327511,
+ "grad_norm": 0.17216597497463226,
+ "learning_rate": 5.999999999999999e-06,
+ "loss": 11.988704681396484,
+ "step": 2
+ },
+ {
+ "epoch": 0.04192139737991266,
+ "grad_norm": 0.17459715902805328,
+ "learning_rate": 1.1999999999999999e-05,
+ "loss": 11.945581436157227,
+ "step": 3
+ },
+ {
+ "epoch": 0.05589519650655022,
+ "grad_norm": 0.17577414214611053,
+ "learning_rate": 1.7999999999999997e-05,
+ "loss": 11.859772682189941,
+ "step": 4
+ },
+ {
+ "epoch": 0.06986899563318777,
+ "grad_norm": 0.16163651645183563,
+ "learning_rate": 2.3999999999999997e-05,
+ "loss": 11.744415283203125,
+ "step": 5
+ },
+ {
+ "epoch": 0.08384279475982533,
+ "grad_norm": 0.14409932494163513,
+ "learning_rate": 2.9999999999999997e-05,
+ "loss": 11.618099212646484,
+ "step": 6
+ },
+ {
+ "epoch": 0.09781659388646288,
+ "grad_norm": 0.13079848885536194,
+ "learning_rate": 3.5999999999999994e-05,
+ "loss": 11.505935668945312,
+ "step": 7
+ },
+ {
+ "epoch": 0.11179039301310044,
+ "grad_norm": 0.11964613944292068,
+ "learning_rate": 4.2e-05,
+ "loss": 11.416532516479492,
+ "step": 8
+ },
+ {
+ "epoch": 0.125764192139738,
+ "grad_norm": 0.11345142871141434,
+ "learning_rate": 4.7999999999999994e-05,
+ "loss": 11.33206558227539,
+ "step": 9
+ },
+ {
+ "epoch": 0.13973799126637554,
+ "grad_norm": 0.11033952981233597,
+ "learning_rate": 5.399999999999999e-05,
+ "loss": 11.259052276611328,
+ "step": 10
+ },
+ {
+ "epoch": 0.1537117903930131,
+ "grad_norm": 0.10776489973068237,
+ "learning_rate": 5.9999999999999995e-05,
+ "loss": 11.203384399414062,
+ "step": 11
+ },
+ {
+ "epoch": 0.16768558951965065,
+ "grad_norm": 0.10721946507692337,
+ "learning_rate": 6.599999999999999e-05,
+ "loss": 11.153295516967773,
+ "step": 12
+ },
+ {
+ "epoch": 0.18165938864628822,
+ "grad_norm": 0.10673430562019348,
+ "learning_rate": 7.199999999999999e-05,
+ "loss": 11.110200881958008,
+ "step": 13
+ },
+ {
+ "epoch": 0.19563318777292577,
+ "grad_norm": 0.10630254447460175,
+ "learning_rate": 7.8e-05,
+ "loss": 11.067477226257324,
+ "step": 14
+ },
+ {
+ "epoch": 0.2096069868995633,
+ "grad_norm": 0.10631615668535233,
+ "learning_rate": 8.4e-05,
+ "loss": 11.022941589355469,
+ "step": 15
+ },
+ {
+ "epoch": 0.22358078602620088,
+ "grad_norm": 0.10609443485736847,
+ "learning_rate": 8.999999999999999e-05,
+ "loss": 10.970746040344238,
+ "step": 16
+ },
+ {
+ "epoch": 0.23755458515283842,
+ "grad_norm": 0.10648232698440552,
+ "learning_rate": 9.599999999999999e-05,
+ "loss": 10.91267204284668,
+ "step": 17
+ },
+ {
+ "epoch": 0.251528384279476,
+ "grad_norm": 0.10584589838981628,
+ "learning_rate": 0.000102,
+ "loss": 10.854547500610352,
+ "step": 18
+ },
+ {
+ "epoch": 0.26550218340611353,
+ "grad_norm": 0.10591720044612885,
+ "learning_rate": 0.00010799999999999998,
+ "loss": 10.788917541503906,
+ "step": 19
+ },
+ {
+ "epoch": 0.2794759825327511,
+ "grad_norm": 0.10611672699451447,
+ "learning_rate": 0.00011399999999999999,
+ "loss": 10.717556953430176,
+ "step": 20
+ },
+ {
+ "epoch": 0.2934497816593886,
+ "grad_norm": 0.10455359518527985,
+ "learning_rate": 0.00011999999999999999,
+ "loss": 10.650186538696289,
+ "step": 21
+ },
+ {
+ "epoch": 0.3074235807860262,
+ "grad_norm": 0.1047525629401207,
+ "learning_rate": 0.00012599999999999997,
+ "loss": 10.56808090209961,
+ "step": 22
+ },
+ {
+ "epoch": 0.32139737991266376,
+ "grad_norm": 0.10511605441570282,
+ "learning_rate": 0.00013199999999999998,
+ "loss": 10.484028816223145,
+ "step": 23
+ },
+ {
+ "epoch": 0.3353711790393013,
+ "grad_norm": 0.10467177629470825,
+ "learning_rate": 0.000138,
+ "loss": 10.399947166442871,
+ "step": 24
+ },
+ {
+ "epoch": 0.34934497816593885,
+ "grad_norm": 0.10432136058807373,
+ "learning_rate": 0.00014399999999999998,
+ "loss": 10.312714576721191,
+ "step": 25
+ },
+ {
+ "epoch": 0.36331877729257644,
+ "grad_norm": 0.10527262091636658,
+ "learning_rate": 0.00015,
+ "loss": 10.212261199951172,
+ "step": 26
+ },
+ {
+ "epoch": 0.377292576419214,
+ "grad_norm": 0.103610560297966,
+ "learning_rate": 0.000156,
+ "loss": 10.132034301757812,
+ "step": 27
+ },
+ {
+ "epoch": 0.39126637554585153,
+ "grad_norm": 0.10553059726953506,
+ "learning_rate": 0.000162,
+ "loss": 10.020434379577637,
+ "step": 28
+ },
+ {
+ "epoch": 0.4052401746724891,
+ "grad_norm": 0.1037299782037735,
+ "learning_rate": 0.000168,
+ "loss": 9.934466361999512,
+ "step": 29
+ },
+ {
+ "epoch": 0.4192139737991266,
+ "grad_norm": 0.1042739674448967,
+ "learning_rate": 0.00017399999999999997,
+ "loss": 9.822539329528809,
+ "step": 30
+ },
+ {
+ "epoch": 0.4331877729257642,
+ "grad_norm": 0.10300865024328232,
+ "learning_rate": 0.00017999999999999998,
+ "loss": 9.734994888305664,
+ "step": 31
+ },
+ {
+ "epoch": 0.44716157205240176,
+ "grad_norm": 0.10291654616594315,
+ "learning_rate": 0.000186,
+ "loss": 9.624505996704102,
+ "step": 32
+ },
+ {
+ "epoch": 0.4611353711790393,
+ "grad_norm": 0.10210851579904556,
+ "learning_rate": 0.00019199999999999998,
+ "loss": 9.518205642700195,
+ "step": 33
+ },
+ {
+ "epoch": 0.47510917030567684,
+ "grad_norm": 0.10222529619932175,
+ "learning_rate": 0.000198,
+ "loss": 9.412222862243652,
+ "step": 34
+ },
+ {
+ "epoch": 0.4890829694323144,
+ "grad_norm": 0.09992019087076187,
+ "learning_rate": 0.000204,
+ "loss": 9.326566696166992,
+ "step": 35
+ },
+ {
+ "epoch": 0.503056768558952,
+ "grad_norm": 0.09931084513664246,
+ "learning_rate": 0.00020999999999999998,
+ "loss": 9.210111618041992,
+ "step": 36
+ },
+ {
+ "epoch": 0.5170305676855895,
+ "grad_norm": 0.09928256273269653,
+ "learning_rate": 0.00021599999999999996,
+ "loss": 9.100493431091309,
+ "step": 37
+ },
+ {
+ "epoch": 0.5310043668122271,
+ "grad_norm": 0.09843816608190536,
+ "learning_rate": 0.00022199999999999998,
+ "loss": 8.996543884277344,
+ "step": 38
+ },
+ {
+ "epoch": 0.5449781659388646,
+ "grad_norm": 0.09825374186038971,
+ "learning_rate": 0.00022799999999999999,
+ "loss": 8.888991355895996,
+ "step": 39
+ },
+ {
+ "epoch": 0.5589519650655022,
+ "grad_norm": 0.09602385759353638,
+ "learning_rate": 0.000234,
+ "loss": 8.802907943725586,
+ "step": 40
+ },
+ {
+ "epoch": 0.5729257641921397,
+ "grad_norm": 0.09414064139127731,
+ "learning_rate": 0.00023999999999999998,
+ "loss": 8.715423583984375,
+ "step": 41
+ },
+ {
+ "epoch": 0.5868995633187772,
+ "grad_norm": 0.09273893386125565,
+ "learning_rate": 0.00024599999999999996,
+ "loss": 8.61374282836914,
+ "step": 42
+ },
+ {
+ "epoch": 0.6008733624454149,
+ "grad_norm": 0.09079599380493164,
+ "learning_rate": 0.00025199999999999995,
+ "loss": 8.502904891967773,
+ "step": 43
+ },
+ {
+ "epoch": 0.6148471615720524,
+ "grad_norm": 0.08605825901031494,
+ "learning_rate": 0.000258,
+ "loss": 8.453008651733398,
+ "step": 44
+ },
+ {
+ "epoch": 0.62882096069869,
+ "grad_norm": 0.08338996022939682,
+ "learning_rate": 0.00026399999999999997,
+ "loss": 8.365163803100586,
+ "step": 45
+ },
+ {
+ "epoch": 0.6427947598253275,
+ "grad_norm": 0.08150867372751236,
+ "learning_rate": 0.00027,
+ "loss": 8.285372734069824,
+ "step": 46
+ },
+ {
+ "epoch": 0.6567685589519651,
+ "grad_norm": 0.07794807851314545,
+ "learning_rate": 0.000276,
+ "loss": 8.185811996459961,
+ "step": 47
+ },
+ {
+ "epoch": 0.6707423580786026,
+ "grad_norm": 0.07592720538377762,
+ "learning_rate": 0.00028199999999999997,
+ "loss": 8.098482131958008,
+ "step": 48
+ },
+ {
+ "epoch": 0.6847161572052402,
+ "grad_norm": 0.0709897130727768,
+ "learning_rate": 0.00028799999999999995,
+ "loss": 8.042091369628906,
+ "step": 49
+ },
+ {
+ "epoch": 0.6986899563318777,
+ "grad_norm": 0.06582305580377579,
+ "learning_rate": 0.000294,
+ "loss": 7.972844123840332,
+ "step": 50
+ },
+ {
+ "epoch": 0.7126637554585152,
+ "grad_norm": 0.059452351182699203,
+ "learning_rate": 0.0003,
+ "loss": 7.924579620361328,
+ "step": 51
+ },
+ {
+ "epoch": 0.7266375545851529,
+ "grad_norm": 0.052426718175411224,
+ "learning_rate": 0.00030599999999999996,
+ "loss": 7.876336097717285,
+ "step": 52
+ },
+ {
+ "epoch": 0.7406113537117904,
+ "grad_norm": 0.04768150672316551,
+ "learning_rate": 0.000312,
+ "loss": 7.8102707862854,
+ "step": 53
+ },
+ {
+ "epoch": 0.754585152838428,
+ "grad_norm": 0.03970075398683548,
+ "learning_rate": 0.000318,
+ "loss": 7.791913986206055,
+ "step": 54
+ },
+ {
+ "epoch": 0.7685589519650655,
+ "grad_norm": 0.03559752553701401,
+ "learning_rate": 0.000324,
+ "loss": 7.756960868835449,
+ "step": 55
+ },
+ {
+ "epoch": 0.7825327510917031,
+ "grad_norm": 0.029321353882551193,
+ "learning_rate": 0.00033,
+ "loss": 7.724878787994385,
+ "step": 56
+ },
+ {
+ "epoch": 0.7965065502183406,
+ "grad_norm": 0.023485427722334862,
+ "learning_rate": 0.000336,
+ "loss": 7.72553014755249,
+ "step": 57
+ },
+ {
+ "epoch": 0.8104803493449781,
+ "grad_norm": 0.019319474697113037,
+ "learning_rate": 0.00034199999999999996,
+ "loss": 7.709493160247803,
+ "step": 58
+ },
+ {
+ "epoch": 0.8244541484716157,
+ "grad_norm": 0.016705691814422607,
+ "learning_rate": 0.00034799999999999995,
+ "loss": 7.6873297691345215,
+ "step": 59
+ },
+ {
+ "epoch": 0.8384279475982532,
+ "grad_norm": 0.017667734995484352,
+ "learning_rate": 0.00035399999999999993,
+ "loss": 7.677059173583984,
+ "step": 60
+ },
+ {
+ "epoch": 0.8524017467248908,
+ "grad_norm": 0.01671992614865303,
+ "learning_rate": 0.00035999999999999997,
+ "loss": 7.643512725830078,
+ "step": 61
+ },
+ {
+ "epoch": 0.8663755458515284,
+ "grad_norm": 0.012288172729313374,
+ "learning_rate": 0.00036599999999999995,
+ "loss": 7.658343315124512,
+ "step": 62
+ },
+ {
+ "epoch": 0.880349344978166,
+ "grad_norm": 0.012004776857793331,
+ "learning_rate": 0.000372,
+ "loss": 7.649759292602539,
+ "step": 63
+ },
+ {
+ "epoch": 0.8943231441048035,
+ "grad_norm": 0.012487445957958698,
+ "learning_rate": 0.00037799999999999997,
+ "loss": 7.631503105163574,
+ "step": 64
+ },
+ {
+ "epoch": 0.9082969432314411,
+ "grad_norm": 0.01489241048693657,
+ "learning_rate": 0.00038399999999999996,
+ "loss": 7.647110462188721,
+ "step": 65
+ },
+ {
+ "epoch": 0.9222707423580786,
+ "grad_norm": 0.01428980939090252,
+ "learning_rate": 0.00039,
+ "loss": 7.636983871459961,
+ "step": 66
+ },
+ {
+ "epoch": 0.9362445414847161,
+ "grad_norm": 0.010711144655942917,
+ "learning_rate": 0.000396,
+ "loss": 7.621130466461182,
+ "step": 67
+ },
+ {
+ "epoch": 0.9502183406113537,
+ "grad_norm": 0.010178886353969574,
+ "learning_rate": 0.000402,
+ "loss": 7.593924522399902,
+ "step": 68
+ },
+ {
+ "epoch": 0.9641921397379912,
+ "grad_norm": 0.010604874230921268,
+ "learning_rate": 0.000408,
+ "loss": 7.6015496253967285,
+ "step": 69
+ },
+ {
+ "epoch": 0.9781659388646288,
+ "grad_norm": 0.009751426056027412,
+ "learning_rate": 0.0004139999999999999,
+ "loss": 7.571964263916016,
+ "step": 70
+ },
+ {
+ "epoch": 0.9921397379912664,
+ "grad_norm": 0.009816481731832027,
+ "learning_rate": 0.00041999999999999996,
+ "loss": 7.57344913482666,
+ "step": 71
+ },
+ {
+ "epoch": 1.0,
+ "grad_norm": 0.008665637113153934,
+ "learning_rate": 0.00042599999999999995,
+ "loss": 7.579050064086914,
+ "step": 72
+ },
+ {
+ "epoch": 1.0,
+ "eval_loss": 7.581204891204834,
+ "eval_runtime": 59.0091,
+ "eval_samples_per_second": 41.383,
+ "eval_steps_per_second": 1.305,
+ "step": 72
+ },
+ {
+ "epoch": 1.0139737991266375,
+ "grad_norm": 0.009684201329946518,
+ "learning_rate": 0.00043199999999999993,
+ "loss": 7.568594932556152,
+ "step": 73
+ },
+ {
+ "epoch": 1.027947598253275,
+ "grad_norm": 0.025783710181713104,
+ "learning_rate": 0.00043799999999999997,
+ "loss": 7.578948497772217,
+ "step": 74
+ },
+ {
+ "epoch": 1.0419213973799126,
+ "grad_norm": 0.0327819399535656,
+ "learning_rate": 0.00044399999999999995,
+ "loss": 7.548173904418945,
+ "step": 75
+ },
+ {
+ "epoch": 1.0558951965065502,
+ "grad_norm": 0.021604498848319054,
+ "learning_rate": 0.00045,
+ "loss": 7.552953720092773,
+ "step": 76
+ },
+ {
+ "epoch": 1.0698689956331877,
+ "grad_norm": 0.010230105370283127,
+ "learning_rate": 0.00045599999999999997,
+ "loss": 7.5213212966918945,
+ "step": 77
+ },
+ {
+ "epoch": 1.0838427947598253,
+ "grad_norm": 0.023315589874982834,
+ "learning_rate": 0.00046199999999999995,
+ "loss": 7.501640319824219,
+ "step": 78
+ },
+ {
+ "epoch": 1.0978165938864628,
+ "grad_norm": 0.01109944935888052,
+ "learning_rate": 0.000468,
+ "loss": 7.539519786834717,
+ "step": 79
+ },
+ {
+ "epoch": 1.1117903930131003,
+ "grad_norm": 0.026093758642673492,
+ "learning_rate": 0.000474,
+ "loss": 7.525928974151611,
+ "step": 80
+ },
+ {
+ "epoch": 1.125764192139738,
+ "grad_norm": 0.03274895250797272,
+ "learning_rate": 0.00047999999999999996,
+ "loss": 7.535602569580078,
+ "step": 81
+ },
+ {
+ "epoch": 1.1397379912663754,
+ "grad_norm": 0.01543602254241705,
+ "learning_rate": 0.000486,
+ "loss": 7.513683795928955,
+ "step": 82
+ },
+ {
+ "epoch": 1.1537117903930132,
+ "grad_norm": 0.028377249836921692,
+ "learning_rate": 0.0004919999999999999,
+ "loss": 7.49384880065918,
+ "step": 83
+ },
+ {
+ "epoch": 1.1676855895196507,
+ "grad_norm": 0.04121481627225876,
+ "learning_rate": 0.000498,
+ "loss": 7.488447189331055,
+ "step": 84
+ },
+ {
+ "epoch": 1.1816593886462883,
+ "grad_norm": 0.013282302767038345,
+ "learning_rate": 0.0005039999999999999,
+ "loss": 7.463010787963867,
+ "step": 85
+ },
+ {
+ "epoch": 1.1956331877729258,
+ "grad_norm": 0.03200364485383034,
+ "learning_rate": 0.0005099999999999999,
+ "loss": 7.473515033721924,
+ "step": 86
+ },
+ {
+ "epoch": 1.2096069868995634,
+ "grad_norm": 0.020768029615283012,
+ "learning_rate": 0.000516,
+ "loss": 7.464688777923584,
+ "step": 87
+ },
+ {
+ "epoch": 1.223580786026201,
+ "grad_norm": 0.021715788170695305,
+ "learning_rate": 0.000522,
+ "loss": 7.4835028648376465,
+ "step": 88
+ },
+ {
+ "epoch": 1.2375545851528384,
+ "grad_norm": 0.026255754753947258,
+ "learning_rate": 0.0005279999999999999,
+ "loss": 7.443392753601074,
+ "step": 89
+ },
+ {
+ "epoch": 1.251528384279476,
+ "grad_norm": 0.022325951606035233,
+ "learning_rate": 0.000534,
+ "loss": 7.457373142242432,
+ "step": 90
+ },
+ {
+ "epoch": 1.2655021834061135,
+ "grad_norm": 0.009270387701690197,
+ "learning_rate": 0.00054,
+ "loss": 7.455986976623535,
+ "step": 91
+ },
+ {
+ "epoch": 1.279475982532751,
+ "grad_norm": 0.02369544468820095,
+ "learning_rate": 0.0005459999999999999,
+ "loss": 7.442216873168945,
+ "step": 92
+ },
+ {
+ "epoch": 1.2934497816593886,
+ "grad_norm": 0.008474506437778473,
+ "learning_rate": 0.000552,
+ "loss": 7.4294915199279785,
+ "step": 93
+ },
+ {
+ "epoch": 1.3074235807860262,
+ "grad_norm": 0.017043249681591988,
+ "learning_rate": 0.000558,
+ "loss": 7.406932830810547,
+ "step": 94
+ },
+ {
+ "epoch": 1.3213973799126637,
+ "grad_norm": 0.011813412420451641,
+ "learning_rate": 0.0005639999999999999,
+ "loss": 7.415900230407715,
+ "step": 95
+ },
+ {
+ "epoch": 1.3353711790393012,
+ "grad_norm": 0.01086883433163166,
+ "learning_rate": 0.00057,
+ "loss": 7.452998638153076,
+ "step": 96
+ },
+ {
+ "epoch": 1.3493449781659388,
+ "grad_norm": 0.014700951054692268,
+ "learning_rate": 0.0005759999999999999,
+ "loss": 7.410249710083008,
+ "step": 97
+ },
+ {
+ "epoch": 1.3633187772925766,
+ "grad_norm": 0.014512976631522179,
+ "learning_rate": 0.0005819999999999999,
+ "loss": 7.429132461547852,
+ "step": 98
+ },
+ {
+ "epoch": 1.3772925764192139,
+ "grad_norm": 0.012519920244812965,
+ "learning_rate": 0.000588,
+ "loss": 7.430817604064941,
+ "step": 99
+ },
+ {
+ "epoch": 1.3912663755458516,
+ "grad_norm": 0.007132664322853088,
+ "learning_rate": 0.0005939999999999999,
+ "loss": 7.405234336853027,
+ "step": 100
+ },
+ {
+ "epoch": 1.405240174672489,
+ "grad_norm": 0.019489388912916183,
+ "learning_rate": 0.0006,
+ "loss": 7.427791595458984,
+ "step": 101
+ },
+ {
+ "epoch": 1.4192139737991267,
+ "grad_norm": 0.027252664789557457,
+ "learning_rate": 0.0006,
+ "loss": 7.41761589050293,
+ "step": 102
+ },
+ {
+ "epoch": 1.4331877729257643,
+ "grad_norm": 0.02501637488603592,
+ "learning_rate": 0.0006,
+ "loss": 7.416735649108887,
+ "step": 103
+ },
+ {
+ "epoch": 1.4471615720524018,
+ "grad_norm": 0.00989772193133831,
+ "learning_rate": 0.0006,
+ "loss": 7.3756632804870605,
+ "step": 104
+ },
+ {
+ "epoch": 1.4611353711790394,
+ "grad_norm": 0.01622949168086052,
+ "learning_rate": 0.0006,
+ "loss": 7.400811672210693,
+ "step": 105
+ },
+ {
+ "epoch": 1.475109170305677,
+ "grad_norm": 0.012407763861119747,
+ "learning_rate": 0.0006,
+ "loss": 7.37675666809082,
+ "step": 106
+ },
+ {
+ "epoch": 1.4890829694323144,
+ "grad_norm": 0.010353565216064453,
+ "learning_rate": 0.0006,
+ "loss": 7.4013142585754395,
+ "step": 107
+ },
+ {
+ "epoch": 1.503056768558952,
+ "grad_norm": 0.0120553532615304,
+ "learning_rate": 0.0006,
+ "loss": 7.379873275756836,
+ "step": 108
+ },
+ {
+ "epoch": 1.5170305676855895,
+ "grad_norm": 0.01610225811600685,
+ "learning_rate": 0.0006,
+ "loss": 7.371631145477295,
+ "step": 109
+ },
+ {
+ "epoch": 1.531004366812227,
+ "grad_norm": 0.01159879844635725,
+ "learning_rate": 0.0006,
+ "loss": 7.389007568359375,
+ "step": 110
+ },
+ {
+ "epoch": 1.5449781659388646,
+ "grad_norm": 0.008013816550374031,
+ "learning_rate": 0.0006,
+ "loss": 7.395920753479004,
+ "step": 111
+ },
+ {
+ "epoch": 1.5589519650655022,
+ "grad_norm": 0.016334809362888336,
+ "learning_rate": 0.0006,
+ "loss": 7.393121719360352,
+ "step": 112
+ },
+ {
+ "epoch": 1.5729257641921397,
+ "grad_norm": 0.018821721896529198,
+ "learning_rate": 0.0006,
+ "loss": 7.367543697357178,
+ "step": 113
+ },
+ {
+ "epoch": 1.5868995633187772,
+ "grad_norm": 0.014088994823396206,
+ "learning_rate": 0.0006,
+ "loss": 7.385537624359131,
+ "step": 114
+ },
+ {
+ "epoch": 1.600873362445415,
+ "grad_norm": 0.005864050704985857,
+ "learning_rate": 0.0006,
+ "loss": 7.365346908569336,
+ "step": 115
+ },
+ {
+ "epoch": 1.6148471615720523,
+ "grad_norm": 0.011745430529117584,
+ "learning_rate": 0.0006,
+ "loss": 7.390227317810059,
+ "step": 116
+ },
+ {
+ "epoch": 1.62882096069869,
+ "grad_norm": 0.017205573618412018,
+ "learning_rate": 0.0006,
+ "loss": 7.3558549880981445,
+ "step": 117
+ },
+ {
+ "epoch": 1.6427947598253274,
+ "grad_norm": 0.021256346255540848,
+ "learning_rate": 0.0006,
+ "loss": 7.377346992492676,
+ "step": 118
+ },
+ {
+ "epoch": 1.6567685589519652,
+ "grad_norm": 0.019335204735398293,
+ "learning_rate": 0.0006,
+ "loss": 7.346194267272949,
+ "step": 119
+ },
+ {
+ "epoch": 1.6707423580786025,
+ "grad_norm": 0.008134003728628159,
+ "learning_rate": 0.0006,
+ "loss": 7.3063578605651855,
+ "step": 120
+ },
+ {
+ "epoch": 1.6847161572052403,
+ "grad_norm": 0.017613831907510757,
+ "learning_rate": 0.0006,
+ "loss": 7.338171005249023,
+ "step": 121
+ },
+ {
+ "epoch": 1.6986899563318776,
+ "grad_norm": 0.00890734139829874,
+ "learning_rate": 0.0006,
+ "loss": 7.355384826660156,
+ "step": 122
+ },
+ {
+ "epoch": 1.7126637554585153,
+ "grad_norm": 0.013448155485093594,
+ "learning_rate": 0.0006,
+ "loss": 7.344296455383301,
+ "step": 123
+ },
+ {
+ "epoch": 1.726637554585153,
+ "grad_norm": 0.009241633117198944,
+ "learning_rate": 0.0006,
+ "loss": 7.3454060554504395,
+ "step": 124
+ },
+ {
+ "epoch": 1.7406113537117904,
+ "grad_norm": 0.018561311066150665,
+ "learning_rate": 0.0006,
+ "loss": 7.32348108291626,
+ "step": 125
+ },
+ {
+ "epoch": 1.754585152838428,
+ "grad_norm": 0.025667704641819,
+ "learning_rate": 0.0006,
+ "loss": 7.356142520904541,
+ "step": 126
+ },
+ {
+ "epoch": 1.7685589519650655,
+ "grad_norm": 0.036232151091098785,
+ "learning_rate": 0.0006,
+ "loss": 7.363539695739746,
+ "step": 127
+ },
+ {
+ "epoch": 1.782532751091703,
+ "grad_norm": 0.059483714401721954,
+ "learning_rate": 0.0006,
+ "loss": 7.402331352233887,
+ "step": 128
+ },
+ {
+ "epoch": 1.7965065502183406,
+ "grad_norm": 0.05365009233355522,
+ "learning_rate": 0.0006,
+ "loss": 7.40457820892334,
+ "step": 129
+ },
+ {
+ "epoch": 1.8104803493449781,
+ "grad_norm": 0.02738415263593197,
+ "learning_rate": 0.0006,
+ "loss": 7.3756208419799805,
+ "step": 130
+ },
+ {
+ "epoch": 1.8244541484716157,
+ "grad_norm": 0.035017091780900955,
+ "learning_rate": 0.0006,
+ "loss": 7.379263877868652,
+ "step": 131
+ },
+ {
+ "epoch": 1.8384279475982532,
+ "grad_norm": 0.02819863148033619,
+ "learning_rate": 0.0006,
+ "loss": 7.327673435211182,
+ "step": 132
+ },
+ {
+ "epoch": 1.8524017467248908,
+ "grad_norm": 0.017754238098859787,
+ "learning_rate": 0.0006,
+ "loss": 7.333420276641846,
+ "step": 133
+ },
+ {
+ "epoch": 1.8663755458515285,
+ "grad_norm": 0.024685733020305634,
+ "learning_rate": 0.0006,
+ "loss": 7.34493350982666,
+ "step": 134
+ },
+ {
+ "epoch": 1.8803493449781659,
+ "grad_norm": 0.019426532089710236,
+ "learning_rate": 0.0006,
+ "loss": 7.328448295593262,
+ "step": 135
+ },
+ {
+ "epoch": 1.8943231441048036,
+ "grad_norm": 0.02106391079723835,
+ "learning_rate": 0.0006,
+ "loss": 7.327917098999023,
+ "step": 136
+ },
+ {
+ "epoch": 1.908296943231441,
+ "grad_norm": 0.02056879736483097,
+ "learning_rate": 0.0006,
+ "loss": 7.3311991691589355,
+ "step": 137
+ },
+ {
+ "epoch": 1.9222707423580787,
+ "grad_norm": 0.016865141689777374,
+ "learning_rate": 0.0006,
+ "loss": 7.312766075134277,
+ "step": 138
+ },
+ {
+ "epoch": 1.936244541484716,
+ "grad_norm": 0.017549855634570122,
+ "learning_rate": 0.0006,
+ "loss": 7.31191349029541,
+ "step": 139
+ },
+ {
+ "epoch": 1.9502183406113538,
+ "grad_norm": 0.015930157154798508,
+ "learning_rate": 0.0006,
+ "loss": 7.330901622772217,
+ "step": 140
+ },
+ {
+ "epoch": 1.9641921397379911,
+ "grad_norm": 0.013283558189868927,
+ "learning_rate": 0.0006,
+ "loss": 7.298318386077881,
+ "step": 141
+ },
+ {
+ "epoch": 1.9781659388646289,
+ "grad_norm": 0.012023469433188438,
+ "learning_rate": 0.0006,
+ "loss": 7.282337188720703,
+ "step": 142
+ },
+ {
+ "epoch": 1.9921397379912664,
+ "grad_norm": 0.01459511648863554,
+ "learning_rate": 0.0006,
+ "loss": 7.282924175262451,
+ "step": 143
+ },
+ {
+ "epoch": 2.0,
+ "grad_norm": 0.017474235966801643,
+ "learning_rate": 0.0006,
+ "loss": 7.254333972930908,
+ "step": 144
+ },
+ {
+ "epoch": 2.0,
+ "eval_loss": 7.318708896636963,
+ "eval_runtime": 59.6555,
+ "eval_samples_per_second": 40.935,
+ "eval_steps_per_second": 1.291,
+ "step": 144
+ },
+ {
+ "epoch": 2.0139737991266378,
+ "grad_norm": 0.019757507368922234,
+ "learning_rate": 0.0006,
+ "loss": 7.30288553237915,
+ "step": 145
+ },
+ {
+ "epoch": 2.027947598253275,
+ "grad_norm": 0.014571248553693295,
+ "learning_rate": 0.0006,
+ "loss": 7.274681568145752,
+ "step": 146
+ },
+ {
+ "epoch": 2.041921397379913,
+ "grad_norm": 0.015744198113679886,
+ "learning_rate": 0.0006,
+ "loss": 7.278632640838623,
+ "step": 147
+ },
+ {
+ "epoch": 2.05589519650655,
+ "grad_norm": 0.028048858046531677,
+ "learning_rate": 0.0006,
+ "loss": 7.26558780670166,
+ "step": 148
+ },
+ {
+ "epoch": 2.069868995633188,
+ "grad_norm": 0.06547307223081589,
+ "learning_rate": 0.0006,
+ "loss": 7.32567834854126,
+ "step": 149
+ },
+ {
+ "epoch": 2.0838427947598253,
+ "grad_norm": 0.05969763547182083,
+ "learning_rate": 0.0006,
+ "loss": 7.395836353302002,
+ "step": 150
+ },
+ {
+ "epoch": 2.097816593886463,
+ "grad_norm": 0.032574281096458435,
+ "learning_rate": 0.0006,
+ "loss": 7.318150997161865,
+ "step": 151
+ },
+ {
+ "epoch": 2.1117903930131003,
+ "grad_norm": 0.029452962800860405,
+ "learning_rate": 0.0006,
+ "loss": 7.308066368103027,
+ "step": 152
+ },
+ {
+ "epoch": 2.125764192139738,
+ "grad_norm": 0.03161991387605667,
+ "learning_rate": 0.0006,
+ "loss": 7.306290626525879,
+ "step": 153
+ },
+ {
+ "epoch": 2.1397379912663754,
+ "grad_norm": 0.027096295729279518,
+ "learning_rate": 0.0006,
+ "loss": 7.297796249389648,
+ "step": 154
+ },
+ {
+ "epoch": 2.153711790393013,
+ "grad_norm": 0.028657056391239166,
+ "learning_rate": 0.0006,
+ "loss": 7.317424774169922,
+ "step": 155
+ },
+ {
+ "epoch": 2.1676855895196505,
+ "grad_norm": 0.02046995982527733,
+ "learning_rate": 0.0006,
+ "loss": 7.297986030578613,
+ "step": 156
+ },
+ {
+ "epoch": 2.1816593886462883,
+ "grad_norm": 0.02220369316637516,
+ "learning_rate": 0.0006,
+ "loss": 7.268237590789795,
+ "step": 157
+ },
+ {
+ "epoch": 2.1956331877729256,
+ "grad_norm": 0.018381357192993164,
+ "learning_rate": 0.0006,
+ "loss": 7.259415626525879,
+ "step": 158
+ },
+ {
+ "epoch": 2.2096069868995634,
+ "grad_norm": 0.0204413291066885,
+ "learning_rate": 0.0006,
+ "loss": 7.258173942565918,
+ "step": 159
+ },
+ {
+ "epoch": 2.2235807860262007,
+ "grad_norm": 0.015874288976192474,
+ "learning_rate": 0.0006,
+ "loss": 7.281834602355957,
+ "step": 160
+ },
+ {
+ "epoch": 2.2375545851528384,
+ "grad_norm": 0.01634068600833416,
+ "learning_rate": 0.0006,
+ "loss": 7.2469868659973145,
+ "step": 161
+ },
+ {
+ "epoch": 2.251528384279476,
+ "grad_norm": 0.01728747971355915,
+ "learning_rate": 0.0006,
+ "loss": 7.280995845794678,
+ "step": 162
+ },
+ {
+ "epoch": 2.2655021834061135,
+ "grad_norm": 0.016341188922524452,
+ "learning_rate": 0.0006,
+ "loss": 7.273357391357422,
+ "step": 163
+ },
+ {
+ "epoch": 2.279475982532751,
+ "grad_norm": 0.016125807538628578,
+ "learning_rate": 0.0006,
+ "loss": 7.273510932922363,
+ "step": 164
+ },
+ {
+ "epoch": 2.2934497816593886,
+ "grad_norm": 0.01629287749528885,
+ "learning_rate": 0.0006,
+ "loss": 7.267029762268066,
+ "step": 165
+ },
+ {
+ "epoch": 2.3074235807860264,
+ "grad_norm": 0.015296131372451782,
+ "learning_rate": 0.0006,
+ "loss": 7.211989402770996,
+ "step": 166
+ },
+ {
+ "epoch": 2.3213973799126637,
+ "grad_norm": 0.01709570363163948,
+ "learning_rate": 0.0006,
+ "loss": 7.242851734161377,
+ "step": 167
+ },
+ {
+ "epoch": 2.3353711790393015,
+ "grad_norm": 0.012552388943731785,
+ "learning_rate": 0.0006,
+ "loss": 7.241696834564209,
+ "step": 168
+ },
+ {
+ "epoch": 2.349344978165939,
+ "grad_norm": 0.014571291394531727,
+ "learning_rate": 0.0006,
+ "loss": 7.270681381225586,
+ "step": 169
+ },
+ {
+ "epoch": 2.3633187772925766,
+ "grad_norm": 0.013389437459409237,
+ "learning_rate": 0.0006,
+ "loss": 7.191600799560547,
+ "step": 170
+ },
+ {
+ "epoch": 2.377292576419214,
+ "grad_norm": 0.011895314790308475,
+ "learning_rate": 0.0006,
+ "loss": 7.213610649108887,
+ "step": 171
+ },
+ {
+ "epoch": 2.3912663755458516,
+ "grad_norm": 0.01345642190426588,
+ "learning_rate": 0.0006,
+ "loss": 7.193787574768066,
+ "step": 172
+ },
+ {
+ "epoch": 2.405240174672489,
+ "grad_norm": 0.0091794328764081,
+ "learning_rate": 0.0006,
+ "loss": 7.233364105224609,
+ "step": 173
+ },
+ {
+ "epoch": 2.4192139737991267,
+ "grad_norm": 0.013767086900770664,
+ "learning_rate": 0.0006,
+ "loss": 7.217272758483887,
+ "step": 174
+ },
+ {
+ "epoch": 2.433187772925764,
+ "grad_norm": 0.015134960412979126,
+ "learning_rate": 0.0006,
+ "loss": 7.192791938781738,
+ "step": 175
+ },
+ {
+ "epoch": 2.447161572052402,
+ "grad_norm": 0.01790648326277733,
+ "learning_rate": 0.0006,
+ "loss": 7.229467391967773,
+ "step": 176
+ },
+ {
+ "epoch": 2.461135371179039,
+ "grad_norm": 0.015362377278506756,
+ "learning_rate": 0.0006,
+ "loss": 7.180346488952637,
+ "step": 177
+ },
+ {
+ "epoch": 2.475109170305677,
+ "grad_norm": 0.010505498386919498,
+ "learning_rate": 0.0006,
+ "loss": 7.188833236694336,
+ "step": 178
+ },
+ {
+ "epoch": 2.489082969432314,
+ "grad_norm": 0.009129747748374939,
+ "learning_rate": 0.0006,
+ "loss": 7.203153133392334,
+ "step": 179
+ },
+ {
+ "epoch": 2.503056768558952,
+ "grad_norm": 0.01147883478552103,
+ "learning_rate": 0.0006,
+ "loss": 7.163824558258057,
+ "step": 180
+ },
+ {
+ "epoch": 2.5170305676855893,
+ "grad_norm": 0.011338942684233189,
+ "learning_rate": 0.0006,
+ "loss": 7.16287899017334,
+ "step": 181
+ },
+ {
+ "epoch": 2.531004366812227,
+ "grad_norm": 0.013277675025165081,
+ "learning_rate": 0.0006,
+ "loss": 7.155905723571777,
+ "step": 182
+ },
+ {
+ "epoch": 2.544978165938865,
+ "grad_norm": 0.015269347466528416,
+ "learning_rate": 0.0006,
+ "loss": 7.189371109008789,
+ "step": 183
+ },
+ {
+ "epoch": 2.558951965065502,
+ "grad_norm": 0.017905596643686295,
+ "learning_rate": 0.0006,
+ "loss": 7.177547931671143,
+ "step": 184
+ },
+ {
+ "epoch": 2.5729257641921395,
+ "grad_norm": 0.024431610479950905,
+ "learning_rate": 0.0006,
+ "loss": 7.192169189453125,
+ "step": 185
+ },
+ {
+ "epoch": 2.5868995633187772,
+ "grad_norm": 0.03656961768865585,
+ "learning_rate": 0.0006,
+ "loss": 7.124387741088867,
+ "step": 186
+ },
+ {
+ "epoch": 2.600873362445415,
+ "grad_norm": 0.04337048903107643,
+ "learning_rate": 0.0006,
+ "loss": 7.140231609344482,
+ "step": 187
+ },
+ {
+ "epoch": 2.6148471615720523,
+ "grad_norm": 0.04047228768467903,
+ "learning_rate": 0.0006,
+ "loss": 7.171927452087402,
+ "step": 188
+ },
+ {
+ "epoch": 2.62882096069869,
+ "grad_norm": 0.07825770974159241,
+ "learning_rate": 0.0006,
+ "loss": 7.207631587982178,
+ "step": 189
+ },
+ {
+ "epoch": 2.6427947598253274,
+ "grad_norm": 0.07118063420057297,
+ "learning_rate": 0.0006,
+ "loss": 7.209488391876221,
+ "step": 190
+ },
+ {
+ "epoch": 2.656768558951965,
+ "grad_norm": 0.05153293535113335,
+ "learning_rate": 0.0006,
+ "loss": 7.14793062210083,
+ "step": 191
+ },
+ {
+ "epoch": 2.6707423580786025,
+ "grad_norm": 0.03318839520215988,
+ "learning_rate": 0.0006,
+ "loss": 7.163010597229004,
+ "step": 192
+ },
+ {
+ "epoch": 2.6847161572052403,
+ "grad_norm": 0.029333539307117462,
+ "learning_rate": 0.0006,
+ "loss": 7.159955978393555,
+ "step": 193
+ },
+ {
+ "epoch": 2.6986899563318776,
+ "grad_norm": 0.029405880719423294,
+ "learning_rate": 0.0006,
+ "loss": 7.136178493499756,
+ "step": 194
+ },
+ {
+ "epoch": 2.7126637554585153,
+ "grad_norm": 0.037815988063812256,
+ "learning_rate": 0.0006,
+ "loss": 7.183186054229736,
+ "step": 195
+ },
+ {
+ "epoch": 2.726637554585153,
+ "grad_norm": 0.020307250320911407,
+ "learning_rate": 0.0006,
+ "loss": 7.133164882659912,
+ "step": 196
+ },
+ {
+ "epoch": 2.7406113537117904,
+ "grad_norm": 0.026048069819808006,
+ "learning_rate": 0.0006,
+ "loss": 7.124977111816406,
+ "step": 197
+ },
+ {
+ "epoch": 2.7545851528384278,
+ "grad_norm": 0.027316724881529808,
+ "learning_rate": 0.0006,
+ "loss": 7.07265567779541,
+ "step": 198
+ },
+ {
+ "epoch": 2.7685589519650655,
+ "grad_norm": 0.021629009395837784,
+ "learning_rate": 0.0006,
+ "loss": 7.150020599365234,
+ "step": 199
+ },
+ {
+ "epoch": 2.7825327510917033,
+ "grad_norm": 0.01605929434299469,
+ "learning_rate": 0.0006,
+ "loss": 7.10919189453125,
+ "step": 200
+ },
+ {
+ "epoch": 2.7965065502183406,
+ "grad_norm": 0.020800089463591576,
+ "learning_rate": 0.0006,
+ "loss": 7.132045269012451,
+ "step": 201
+ },
+ {
+ "epoch": 2.810480349344978,
+ "grad_norm": 0.019225774332880974,
+ "learning_rate": 0.0006,
+ "loss": 7.128364562988281,
+ "step": 202
+ },
+ {
+ "epoch": 2.8244541484716157,
+ "grad_norm": 0.01561372634023428,
+ "learning_rate": 0.0006,
+ "loss": 7.083669662475586,
+ "step": 203
+ },
+ {
+ "epoch": 2.8384279475982535,
+ "grad_norm": 0.021969519555568695,
+ "learning_rate": 0.0006,
+ "loss": 7.050841331481934,
+ "step": 204
+ },
+ {
+ "epoch": 2.8524017467248908,
+ "grad_norm": 0.021453695371747017,
+ "learning_rate": 0.0006,
+ "loss": 7.097085475921631,
+ "step": 205
+ },
+ {
+ "epoch": 2.8663755458515285,
+ "grad_norm": 0.022836022078990936,
+ "learning_rate": 0.0006,
+ "loss": 7.10405969619751,
+ "step": 206
+ },
+ {
+ "epoch": 2.880349344978166,
+ "grad_norm": 0.02430851384997368,
+ "learning_rate": 0.0006,
+ "loss": 7.045181751251221,
+ "step": 207
+ },
+ {
+ "epoch": 2.8943231441048036,
+ "grad_norm": 0.020166993141174316,
+ "learning_rate": 0.0006,
+ "loss": 7.068183898925781,
+ "step": 208
+ },
+ {
+ "epoch": 2.908296943231441,
+ "grad_norm": 0.0118505684658885,
+ "learning_rate": 0.0006,
+ "loss": 7.07132625579834,
+ "step": 209
+ },
+ {
+ "epoch": 2.9222707423580787,
+ "grad_norm": 0.023772427812218666,
+ "learning_rate": 0.0006,
+ "loss": 7.055238723754883,
+ "step": 210
+ },
+ {
+ "epoch": 2.936244541484716,
+ "grad_norm": 0.02350712940096855,
+ "learning_rate": 0.0006,
+ "loss": 7.055898666381836,
+ "step": 211
+ },
+ {
+ "epoch": 2.950218340611354,
+ "grad_norm": 0.017244907096028328,
+ "learning_rate": 0.0006,
+ "loss": 7.0099711418151855,
+ "step": 212
+ },
+ {
+ "epoch": 2.964192139737991,
+ "grad_norm": 0.021565575152635574,
+ "learning_rate": 0.0006,
+ "loss": 6.991751670837402,
+ "step": 213
+ },
+ {
+ "epoch": 2.978165938864629,
+ "grad_norm": 0.03241860866546631,
+ "learning_rate": 0.0006,
+ "loss": 7.005980491638184,
+ "step": 214
+ },
+ {
+ "epoch": 2.992139737991266,
+ "grad_norm": 0.04613726586103439,
+ "learning_rate": 0.0006,
+ "loss": 7.027859687805176,
+ "step": 215
+ },
+ {
+ "epoch": 3.0,
+ "grad_norm": 0.06404894590377808,
+ "learning_rate": 0.0006,
+ "loss": 7.123725891113281,
+ "step": 216
+ },
+ {
+ "epoch": 3.0,
+ "eval_loss": 7.070647716522217,
+ "eval_runtime": 58.5801,
+ "eval_samples_per_second": 41.687,
+ "eval_steps_per_second": 1.314,
+ "step": 216
+ },
+ {
+ "epoch": 3.0139737991266378,
+ "grad_norm": 0.05542432889342308,
+ "learning_rate": 0.0006,
+ "loss": 7.078299522399902,
+ "step": 217
+ },
+ {
+ "epoch": 3.027947598253275,
+ "grad_norm": 0.026484526693820953,
+ "learning_rate": 0.0006,
+ "loss": 6.961159706115723,
+ "step": 218
+ },
+ {
+ "epoch": 3.041921397379913,
+ "grad_norm": 0.028276391327381134,
+ "learning_rate": 0.0006,
+ "loss": 7.019961833953857,
+ "step": 219
+ },
+ {
+ "epoch": 3.05589519650655,
+ "grad_norm": 0.024486595764756203,
+ "learning_rate": 0.0006,
+ "loss": 7.039677619934082,
+ "step": 220
+ },
+ {
+ "epoch": 3.069868995633188,
+ "grad_norm": 0.017620893195271492,
+ "learning_rate": 0.0006,
+ "loss": 6.9270782470703125,
+ "step": 221
+ },
+ {
+ "epoch": 3.0838427947598253,
+ "grad_norm": 0.02547302283346653,
+ "learning_rate": 0.0006,
+ "loss": 6.924233436584473,
+ "step": 222
+ },
+ {
+ "epoch": 3.097816593886463,
+ "grad_norm": 0.02539009042084217,
+ "learning_rate": 0.0006,
+ "loss": 6.973979949951172,
+ "step": 223
+ },
+ {
+ "epoch": 3.1117903930131003,
+ "grad_norm": 0.03256227448582649,
+ "learning_rate": 0.0006,
+ "loss": 6.933725357055664,
+ "step": 224
+ },
+ {
+ "epoch": 3.125764192139738,
+ "grad_norm": 0.04651800915598869,
+ "learning_rate": 0.0006,
+ "loss": 6.951044082641602,
+ "step": 225
+ },
+ {
+ "epoch": 3.1397379912663754,
+ "grad_norm": 0.04822975769639015,
+ "learning_rate": 0.0006,
+ "loss": 6.9614410400390625,
+ "step": 226
+ },
+ {
+ "epoch": 3.153711790393013,
+ "grad_norm": 0.031165020540356636,
+ "learning_rate": 0.0006,
+ "loss": 6.91193151473999,
+ "step": 227
+ },
+ {
+ "epoch": 3.1676855895196505,
+ "grad_norm": 0.029754292219877243,
+ "learning_rate": 0.0006,
+ "loss": 6.91802978515625,
+ "step": 228
+ },
+ {
+ "epoch": 3.1816593886462883,
+ "grad_norm": 0.02193152718245983,
+ "learning_rate": 0.0006,
+ "loss": 6.928090572357178,
+ "step": 229
+ },
+ {
+ "epoch": 3.1956331877729256,
+ "grad_norm": 0.02428470179438591,
+ "learning_rate": 0.0006,
+ "loss": 6.90396785736084,
+ "step": 230
+ },
+ {
+ "epoch": 3.2096069868995634,
+ "grad_norm": 0.023374345153570175,
+ "learning_rate": 0.0006,
+ "loss": 6.885858535766602,
+ "step": 231
+ },
+ {
+ "epoch": 3.2235807860262007,
+ "grad_norm": 0.023405000567436218,
+ "learning_rate": 0.0006,
+ "loss": 6.8855767250061035,
+ "step": 232
+ },
+ {
+ "epoch": 3.2375545851528384,
+ "grad_norm": 0.017277134582400322,
+ "learning_rate": 0.0006,
+ "loss": 6.901449203491211,
+ "step": 233
+ },
+ {
+ "epoch": 3.251528384279476,
+ "grad_norm": 0.019496750086545944,
+ "learning_rate": 0.0006,
+ "loss": 6.885687828063965,
+ "step": 234
+ },
+ {
+ "epoch": 3.2655021834061135,
+ "grad_norm": 0.016893276944756508,
+ "learning_rate": 0.0006,
+ "loss": 6.872369766235352,
+ "step": 235
+ },
+ {
+ "epoch": 3.279475982532751,
+ "grad_norm": 0.02058715932071209,
+ "learning_rate": 0.0006,
+ "loss": 6.823280334472656,
+ "step": 236
+ },
+ {
+ "epoch": 3.2934497816593886,
+ "grad_norm": 0.015530755743384361,
+ "learning_rate": 0.0006,
+ "loss": 6.860499858856201,
+ "step": 237
+ },
+ {
+ "epoch": 3.3074235807860264,
+ "grad_norm": 0.01934926211833954,
+ "learning_rate": 0.0006,
+ "loss": 6.817448616027832,
+ "step": 238
+ },
+ {
+ "epoch": 3.3213973799126637,
+ "grad_norm": 0.017413552850484848,
+ "learning_rate": 0.0006,
+ "loss": 6.881955146789551,
+ "step": 239
+ },
+ {
+ "epoch": 3.3353711790393015,
+ "grad_norm": 0.026478875428438187,
+ "learning_rate": 0.0006,
+ "loss": 6.827404499053955,
+ "step": 240
+ },
+ {
+ "epoch": 3.349344978165939,
+ "grad_norm": 0.032611701637506485,
+ "learning_rate": 0.0006,
+ "loss": 6.78285551071167,
+ "step": 241
+ },
+ {
+ "epoch": 3.3633187772925766,
+ "grad_norm": 0.041900306940078735,
+ "learning_rate": 0.0006,
+ "loss": 6.783053874969482,
+ "step": 242
+ },
+ {
+ "epoch": 3.377292576419214,
+ "grad_norm": 0.04149497672915459,
+ "learning_rate": 0.0006,
+ "loss": 6.745220184326172,
+ "step": 243
+ },
+ {
+ "epoch": 3.3912663755458516,
+ "grad_norm": 0.023949826136231422,
+ "learning_rate": 0.0006,
+ "loss": 6.841533184051514,
+ "step": 244
+ },
+ {
+ "epoch": 3.405240174672489,
+ "grad_norm": 0.028601842001080513,
+ "learning_rate": 0.0006,
+ "loss": 6.852852821350098,
+ "step": 245
+ },
+ {
+ "epoch": 3.4192139737991267,
+ "grad_norm": 0.027365155518054962,
+ "learning_rate": 0.0006,
+ "loss": 6.74376106262207,
+ "step": 246
+ },
+ {
+ "epoch": 3.433187772925764,
+ "grad_norm": 0.02856568619608879,
+ "learning_rate": 0.0006,
+ "loss": 6.74906063079834,
+ "step": 247
+ },
+ {
+ "epoch": 3.447161572052402,
+ "grad_norm": 0.019172416999936104,
+ "learning_rate": 0.0006,
+ "loss": 6.809206008911133,
+ "step": 248
+ },
+ {
+ "epoch": 3.461135371179039,
+ "grad_norm": 0.02161746844649315,
+ "learning_rate": 0.0006,
+ "loss": 6.789888381958008,
+ "step": 249
+ },
+ {
+ "epoch": 3.475109170305677,
+ "grad_norm": 0.020273666828870773,
+ "learning_rate": 0.0006,
+ "loss": 6.73512077331543,
+ "step": 250
+ },
+ {
+ "epoch": 3.489082969432314,
+ "grad_norm": 0.022329283878207207,
+ "learning_rate": 0.0006,
+ "loss": 6.826148509979248,
+ "step": 251
+ },
+ {
+ "epoch": 3.503056768558952,
+ "grad_norm": 0.02284284122288227,
+ "learning_rate": 0.0006,
+ "loss": 6.748521327972412,
+ "step": 252
+ },
+ {
+ "epoch": 3.5170305676855893,
+ "grad_norm": 0.026368247345089912,
+ "learning_rate": 0.0006,
+ "loss": 6.645879745483398,
+ "step": 253
+ },
+ {
+ "epoch": 3.531004366812227,
+ "grad_norm": 0.03627244383096695,
+ "learning_rate": 0.0006,
+ "loss": 6.666064739227295,
+ "step": 254
+ },
+ {
+ "epoch": 3.544978165938865,
+ "grad_norm": 0.04140935093164444,
+ "learning_rate": 0.0006,
+ "loss": 6.712275981903076,
+ "step": 255
+ },
+ {
+ "epoch": 3.558951965065502,
+ "grad_norm": 0.0390462800860405,
+ "learning_rate": 0.0006,
+ "loss": 6.665900230407715,
+ "step": 256
+ },
+ {
+ "epoch": 3.5729257641921395,
+ "grad_norm": 0.05178583785891533,
+ "learning_rate": 0.0006,
+ "loss": 6.753298759460449,
+ "step": 257
+ },
+ {
+ "epoch": 3.5868995633187772,
+ "grad_norm": 0.054328061640262604,
+ "learning_rate": 0.0006,
+ "loss": 6.732633113861084,
+ "step": 258
+ },
+ {
+ "epoch": 3.600873362445415,
+ "grad_norm": 0.04590460658073425,
+ "learning_rate": 0.0006,
+ "loss": 6.65250301361084,
+ "step": 259
+ },
+ {
+ "epoch": 3.6148471615720523,
+ "grad_norm": 0.044331666082143784,
+ "learning_rate": 0.0006,
+ "loss": 6.68654203414917,
+ "step": 260
+ },
+ {
+ "epoch": 3.62882096069869,
+ "grad_norm": 0.05308730527758598,
+ "learning_rate": 0.0006,
+ "loss": 6.642482757568359,
+ "step": 261
+ },
+ {
+ "epoch": 3.6427947598253274,
+ "grad_norm": 0.03999016433954239,
+ "learning_rate": 0.0006,
+ "loss": 6.625227451324463,
+ "step": 262
+ },
+ {
+ "epoch": 3.656768558951965,
+ "grad_norm": 0.04088086634874344,
+ "learning_rate": 0.0006,
+ "loss": 6.67623233795166,
+ "step": 263
+ },
+ {
+ "epoch": 3.6707423580786025,
+ "grad_norm": 0.03615636005997658,
+ "learning_rate": 0.0006,
+ "loss": 6.6869001388549805,
+ "step": 264
+ },
+ {
+ "epoch": 3.6847161572052403,
+ "grad_norm": 0.02817360684275627,
+ "learning_rate": 0.0006,
+ "loss": 6.674601078033447,
+ "step": 265
+ },
+ {
+ "epoch": 3.6986899563318776,
+ "grad_norm": 0.032586123794317245,
+ "learning_rate": 0.0006,
+ "loss": 6.612994194030762,
+ "step": 266
+ },
+ {
+ "epoch": 3.7126637554585153,
+ "grad_norm": 0.031117867678403854,
+ "learning_rate": 0.0006,
+ "loss": 6.6244354248046875,
+ "step": 267
+ },
+ {
+ "epoch": 3.726637554585153,
+ "grad_norm": 0.028330687433481216,
+ "learning_rate": 0.0006,
+ "loss": 6.669557094573975,
+ "step": 268
+ },
+ {
+ "epoch": 3.7406113537117904,
+ "grad_norm": 0.021489202976226807,
+ "learning_rate": 0.0006,
+ "loss": 6.632035732269287,
+ "step": 269
+ },
+ {
+ "epoch": 3.7545851528384278,
+ "grad_norm": 0.024269424378871918,
+ "learning_rate": 0.0006,
+ "loss": 6.660867691040039,
+ "step": 270
+ },
+ {
+ "epoch": 3.7685589519650655,
+ "grad_norm": 0.02237873338162899,
+ "learning_rate": 0.0006,
+ "loss": 6.602587699890137,
+ "step": 271
+ },
+ {
+ "epoch": 3.7825327510917033,
+ "grad_norm": 0.026503929868340492,
+ "learning_rate": 0.0006,
+ "loss": 6.599200248718262,
+ "step": 272
+ },
+ {
+ "epoch": 3.7965065502183406,
+ "grad_norm": 0.03687124326825142,
+ "learning_rate": 0.0006,
+ "loss": 6.567296504974365,
+ "step": 273
+ },
+ {
+ "epoch": 3.810480349344978,
+ "grad_norm": 0.04708952456712723,
+ "learning_rate": 0.0006,
+ "loss": 6.621265411376953,
+ "step": 274
+ },
+ {
+ "epoch": 3.8244541484716157,
+ "grad_norm": 0.041846614331007004,
+ "learning_rate": 0.0006,
+ "loss": 6.550536155700684,
+ "step": 275
+ },
+ {
+ "epoch": 3.8384279475982535,
+ "grad_norm": 0.031249675899744034,
+ "learning_rate": 0.0006,
+ "loss": 6.54278564453125,
+ "step": 276
+ },
+ {
+ "epoch": 3.8524017467248908,
+ "grad_norm": 0.023082256317138672,
+ "learning_rate": 0.0006,
+ "loss": 6.544787406921387,
+ "step": 277
+ },
+ {
+ "epoch": 3.8663755458515285,
+ "grad_norm": 0.020995570346713066,
+ "learning_rate": 0.0006,
+ "loss": 6.518143653869629,
+ "step": 278
+ },
+ {
+ "epoch": 3.880349344978166,
+ "grad_norm": 0.020839324221014977,
+ "learning_rate": 0.0006,
+ "loss": 6.546667098999023,
+ "step": 279
+ },
+ {
+ "epoch": 3.8943231441048036,
+ "grad_norm": 0.018377432599663734,
+ "learning_rate": 0.0006,
+ "loss": 6.478307723999023,
+ "step": 280
+ },
+ {
+ "epoch": 3.908296943231441,
+ "grad_norm": 0.021713724359869957,
+ "learning_rate": 0.0006,
+ "loss": 6.485574245452881,
+ "step": 281
+ },
+ {
+ "epoch": 3.9222707423580787,
+ "grad_norm": 0.01359301246702671,
+ "learning_rate": 0.0006,
+ "loss": 6.570268630981445,
+ "step": 282
+ },
+ {
+ "epoch": 3.936244541484716,
+ "grad_norm": 0.016233332455158234,
+ "learning_rate": 0.0006,
+ "loss": 6.57150936126709,
+ "step": 283
+ },
+ {
+ "epoch": 3.950218340611354,
+ "grad_norm": 0.020687608048319817,
+ "learning_rate": 0.0006,
+ "loss": 6.527956008911133,
+ "step": 284
+ },
+ {
+ "epoch": 3.964192139737991,
+ "grad_norm": 0.02032964862883091,
+ "learning_rate": 0.0006,
+ "loss": 6.468169689178467,
+ "step": 285
+ },
+ {
+ "epoch": 3.978165938864629,
+ "grad_norm": 0.016676288098096848,
+ "learning_rate": 0.0006,
+ "loss": 6.448418617248535,
+ "step": 286
+ },
+ {
+ "epoch": 3.992139737991266,
+ "grad_norm": 0.0168539360165596,
+ "learning_rate": 0.0006,
+ "loss": 6.461180686950684,
+ "step": 287
+ },
+ {
+ "epoch": 4.0,
+ "grad_norm": 0.02666233666241169,
+ "learning_rate": 0.0006,
+ "loss": 6.344893932342529,
+ "step": 288
+ },
+ {
+ "epoch": 4.0,
+ "eval_loss": 6.5384979248046875,
+ "eval_runtime": 58.814,
+ "eval_samples_per_second": 41.521,
+ "eval_steps_per_second": 1.309,
+ "step": 288
+ },
+ {
+ "epoch": 4.013973799126638,
+ "grad_norm": 0.04154708981513977,
+ "learning_rate": 0.0006,
+ "loss": 6.41855525970459,
+ "step": 289
+ },
+ {
+ "epoch": 4.0279475982532755,
+ "grad_norm": 0.04895668104290962,
+ "learning_rate": 0.0006,
+ "loss": 6.499302864074707,
+ "step": 290
+ },
+ {
+ "epoch": 4.041921397379912,
+ "grad_norm": 0.034464482218027115,
+ "learning_rate": 0.0006,
+ "loss": 6.438611030578613,
+ "step": 291
+ },
+ {
+ "epoch": 4.05589519650655,
+ "grad_norm": 0.06845773756504059,
+ "learning_rate": 0.0006,
+ "loss": 6.538361072540283,
+ "step": 292
+ },
+ {
+ "epoch": 4.069868995633188,
+ "grad_norm": 0.053768858313560486,
+ "learning_rate": 0.0006,
+ "loss": 6.527524948120117,
+ "step": 293
+ },
+ {
+ "epoch": 4.083842794759826,
+ "grad_norm": 0.039037931710481644,
+ "learning_rate": 0.0006,
+ "loss": 6.503847122192383,
+ "step": 294
+ },
+ {
+ "epoch": 4.097816593886463,
+ "grad_norm": 0.03962196782231331,
+ "learning_rate": 0.0006,
+ "loss": 6.475178241729736,
+ "step": 295
+ },
+ {
+ "epoch": 4.1117903930131,
+ "grad_norm": 0.049101535230875015,
+ "learning_rate": 0.0006,
+ "loss": 6.474587917327881,
+ "step": 296
+ },
+ {
+ "epoch": 4.125764192139738,
+ "grad_norm": 0.02830282226204872,
+ "learning_rate": 0.0006,
+ "loss": 6.45065975189209,
+ "step": 297
+ },
+ {
+ "epoch": 4.139737991266376,
+ "grad_norm": 0.023153482005000114,
+ "learning_rate": 0.0006,
+ "loss": 6.5218706130981445,
+ "step": 298
+ },
+ {
+ "epoch": 4.153711790393013,
+ "grad_norm": 0.025064002722501755,
+ "learning_rate": 0.0006,
+ "loss": 6.456475257873535,
+ "step": 299
+ },
+ {
+ "epoch": 4.1676855895196505,
+ "grad_norm": 0.02357092685997486,
+ "learning_rate": 0.0006,
+ "loss": 6.485952377319336,
+ "step": 300
+ },
+ {
+ "epoch": 4.181659388646288,
+ "grad_norm": 0.021832305938005447,
+ "learning_rate": 0.0006,
+ "loss": 6.425506591796875,
+ "step": 301
+ },
+ {
+ "epoch": 4.195633187772926,
+ "grad_norm": 0.0227900929749012,
+ "learning_rate": 0.0006,
+ "loss": 6.439155578613281,
+ "step": 302
+ },
+ {
+ "epoch": 4.209606986899563,
+ "grad_norm": 0.015961607918143272,
+ "learning_rate": 0.0006,
+ "loss": 6.460862636566162,
+ "step": 303
+ },
+ {
+ "epoch": 4.223580786026201,
+ "grad_norm": 0.017876390367746353,
+ "learning_rate": 0.0006,
+ "loss": 6.44564151763916,
+ "step": 304
+ },
+ {
+ "epoch": 4.2375545851528384,
+ "grad_norm": 0.013703204691410065,
+ "learning_rate": 0.0006,
+ "loss": 6.418414115905762,
+ "step": 305
+ },
+ {
+ "epoch": 4.251528384279476,
+ "grad_norm": 0.015240306034684181,
+ "learning_rate": 0.0006,
+ "loss": 6.411238670349121,
+ "step": 306
+ },
+ {
+ "epoch": 4.265502183406113,
+ "grad_norm": 0.01587662845849991,
+ "learning_rate": 0.0006,
+ "loss": 6.395642280578613,
+ "step": 307
+ },
+ {
+ "epoch": 4.279475982532751,
+ "grad_norm": 0.014650700613856316,
+ "learning_rate": 0.0006,
+ "loss": 6.302793025970459,
+ "step": 308
+ },
+ {
+ "epoch": 4.293449781659389,
+ "grad_norm": 0.016533225774765015,
+ "learning_rate": 0.0006,
+ "loss": 6.380588054656982,
+ "step": 309
+ },
+ {
+ "epoch": 4.307423580786026,
+ "grad_norm": 0.019963741302490234,
+ "learning_rate": 0.0006,
+ "loss": 6.345336437225342,
+ "step": 310
+ },
+ {
+ "epoch": 4.321397379912664,
+ "grad_norm": 0.02237936295568943,
+ "learning_rate": 0.0006,
+ "loss": 6.387770652770996,
+ "step": 311
+ },
+ {
+ "epoch": 4.335371179039301,
+ "grad_norm": 0.026658328250050545,
+ "learning_rate": 0.0006,
+ "loss": 6.323662757873535,
+ "step": 312
+ },
+ {
+ "epoch": 4.349344978165939,
+ "grad_norm": 0.02852284163236618,
+ "learning_rate": 0.0006,
+ "loss": 6.382204055786133,
+ "step": 313
+ },
+ {
+ "epoch": 4.3633187772925766,
+ "grad_norm": 0.028815090656280518,
+ "learning_rate": 0.0006,
+ "loss": 6.402390480041504,
+ "step": 314
+ },
+ {
+ "epoch": 4.377292576419214,
+ "grad_norm": 0.029393529519438744,
+ "learning_rate": 0.0006,
+ "loss": 6.280278205871582,
+ "step": 315
+ },
+ {
+ "epoch": 4.391266375545851,
+ "grad_norm": 0.02186041697859764,
+ "learning_rate": 0.0006,
+ "loss": 6.302707672119141,
+ "step": 316
+ },
+ {
+ "epoch": 4.405240174672489,
+ "grad_norm": 0.02071218006312847,
+ "learning_rate": 0.0006,
+ "loss": 6.3365020751953125,
+ "step": 317
+ },
+ {
+ "epoch": 4.419213973799127,
+ "grad_norm": 0.02779117226600647,
+ "learning_rate": 0.0006,
+ "loss": 6.252157211303711,
+ "step": 318
+ },
+ {
+ "epoch": 4.4331877729257645,
+ "grad_norm": 0.03470654785633087,
+ "learning_rate": 0.0006,
+ "loss": 6.319070339202881,
+ "step": 319
+ },
+ {
+ "epoch": 4.447161572052401,
+ "grad_norm": 0.037067804485559464,
+ "learning_rate": 0.0006,
+ "loss": 6.314116477966309,
+ "step": 320
+ },
+ {
+ "epoch": 4.461135371179039,
+ "grad_norm": 0.032182756811380386,
+ "learning_rate": 0.0006,
+ "loss": 6.267298698425293,
+ "step": 321
+ },
+ {
+ "epoch": 4.475109170305677,
+ "grad_norm": 0.026305217295885086,
+ "learning_rate": 0.0006,
+ "loss": 6.331688404083252,
+ "step": 322
+ },
+ {
+ "epoch": 4.489082969432315,
+ "grad_norm": 0.027228357270359993,
+ "learning_rate": 0.0006,
+ "loss": 6.31139612197876,
+ "step": 323
+ },
+ {
+ "epoch": 4.503056768558952,
+ "grad_norm": 0.025781169533729553,
+ "learning_rate": 0.0006,
+ "loss": 6.194684982299805,
+ "step": 324
+ },
+ {
+ "epoch": 4.517030567685589,
+ "grad_norm": 0.029186977073550224,
+ "learning_rate": 0.0006,
+ "loss": 6.31306266784668,
+ "step": 325
+ },
+ {
+ "epoch": 4.531004366812227,
+ "grad_norm": 0.023608777672052383,
+ "learning_rate": 0.0006,
+ "loss": 6.285243034362793,
+ "step": 326
+ },
+ {
+ "epoch": 4.544978165938865,
+ "grad_norm": 0.02267594076693058,
+ "learning_rate": 0.0006,
+ "loss": 6.30342960357666,
+ "step": 327
+ },
+ {
+ "epoch": 4.558951965065502,
+ "grad_norm": 0.01926310732960701,
+ "learning_rate": 0.0006,
+ "loss": 6.280606269836426,
+ "step": 328
+ },
+ {
+ "epoch": 4.5729257641921395,
+ "grad_norm": 0.026484837755560875,
+ "learning_rate": 0.0006,
+ "loss": 6.237980365753174,
+ "step": 329
+ },
+ {
+ "epoch": 4.586899563318777,
+ "grad_norm": 0.026228854432702065,
+ "learning_rate": 0.0006,
+ "loss": 6.3018293380737305,
+ "step": 330
+ },
+ {
+ "epoch": 4.600873362445415,
+ "grad_norm": 0.022096967324614525,
+ "learning_rate": 0.0006,
+ "loss": 6.253863334655762,
+ "step": 331
+ },
+ {
+ "epoch": 4.614847161572053,
+ "grad_norm": 0.027223890647292137,
+ "learning_rate": 0.0006,
+ "loss": 6.166882038116455,
+ "step": 332
+ },
+ {
+ "epoch": 4.62882096069869,
+ "grad_norm": 0.03160124272108078,
+ "learning_rate": 0.0006,
+ "loss": 6.230309963226318,
+ "step": 333
+ },
+ {
+ "epoch": 4.642794759825327,
+ "grad_norm": 0.04110539332032204,
+ "learning_rate": 0.0006,
+ "loss": 6.2334818840026855,
+ "step": 334
+ },
+ {
+ "epoch": 4.656768558951965,
+ "grad_norm": 0.04355933889746666,
+ "learning_rate": 0.0006,
+ "loss": 6.160956382751465,
+ "step": 335
+ },
+ {
+ "epoch": 4.670742358078603,
+ "grad_norm": 0.03611086308956146,
+ "learning_rate": 0.0006,
+ "loss": 6.23399543762207,
+ "step": 336
+ },
+ {
+ "epoch": 4.68471615720524,
+ "grad_norm": 0.051811400800943375,
+ "learning_rate": 0.0006,
+ "loss": 6.241855144500732,
+ "step": 337
+ },
+ {
+ "epoch": 4.698689956331878,
+ "grad_norm": 0.057434193789958954,
+ "learning_rate": 0.0006,
+ "loss": 6.220993995666504,
+ "step": 338
+ },
+ {
+ "epoch": 4.712663755458515,
+ "grad_norm": 0.04420953616499901,
+ "learning_rate": 0.0006,
+ "loss": 6.300461292266846,
+ "step": 339
+ },
+ {
+ "epoch": 4.726637554585153,
+ "grad_norm": 0.03515457361936569,
+ "learning_rate": 0.0006,
+ "loss": 6.269349098205566,
+ "step": 340
+ },
+ {
+ "epoch": 4.74061135371179,
+ "grad_norm": 0.03222600743174553,
+ "learning_rate": 0.0006,
+ "loss": 6.248613357543945,
+ "step": 341
+ },
+ {
+ "epoch": 4.754585152838428,
+ "grad_norm": 0.0384046845138073,
+ "learning_rate": 0.0006,
+ "loss": 6.266979217529297,
+ "step": 342
+ },
+ {
+ "epoch": 4.7685589519650655,
+ "grad_norm": 0.0514867827296257,
+ "learning_rate": 0.0006,
+ "loss": 6.234103202819824,
+ "step": 343
+ },
+ {
+ "epoch": 4.782532751091703,
+ "grad_norm": 0.050747793167829514,
+ "learning_rate": 0.0006,
+ "loss": 6.322582244873047,
+ "step": 344
+ },
+ {
+ "epoch": 4.796506550218341,
+ "grad_norm": 0.040878988802433014,
+ "learning_rate": 0.0006,
+ "loss": 6.212509632110596,
+ "step": 345
+ },
+ {
+ "epoch": 4.810480349344978,
+ "grad_norm": 0.034655820578336716,
+ "learning_rate": 0.0006,
+ "loss": 6.265879154205322,
+ "step": 346
+ },
+ {
+ "epoch": 4.824454148471616,
+ "grad_norm": 0.04202224686741829,
+ "learning_rate": 0.0006,
+ "loss": 6.2896409034729,
+ "step": 347
+ },
+ {
+ "epoch": 4.8384279475982535,
+ "grad_norm": 0.03632277995347977,
+ "learning_rate": 0.0006,
+ "loss": 6.253917217254639,
+ "step": 348
+ },
+ {
+ "epoch": 4.85240174672489,
+ "grad_norm": 0.028645765036344528,
+ "learning_rate": 0.0006,
+ "loss": 6.207965850830078,
+ "step": 349
+ },
+ {
+ "epoch": 4.866375545851528,
+ "grad_norm": 0.025305170565843582,
+ "learning_rate": 0.0006,
+ "loss": 6.168068885803223,
+ "step": 350
+ },
+ {
+ "epoch": 4.880349344978166,
+ "grad_norm": 0.022137803956866264,
+ "learning_rate": 0.0006,
+ "loss": 6.235283374786377,
+ "step": 351
+ },
+ {
+ "epoch": 4.894323144104804,
+ "grad_norm": 0.01814538985490799,
+ "learning_rate": 0.0006,
+ "loss": 6.259641647338867,
+ "step": 352
+ },
+ {
+ "epoch": 4.908296943231441,
+ "grad_norm": 0.018465086817741394,
+ "learning_rate": 0.0006,
+ "loss": 6.111738681793213,
+ "step": 353
+ },
+ {
+ "epoch": 4.922270742358078,
+ "grad_norm": 0.018638933077454567,
+ "learning_rate": 0.0006,
+ "loss": 6.225642204284668,
+ "step": 354
+ },
+ {
+ "epoch": 4.936244541484716,
+ "grad_norm": 0.016461260616779327,
+ "learning_rate": 0.0006,
+ "loss": 6.140647888183594,
+ "step": 355
+ },
+ {
+ "epoch": 4.950218340611354,
+ "grad_norm": 0.015458385460078716,
+ "learning_rate": 0.0006,
+ "loss": 6.155224800109863,
+ "step": 356
+ },
+ {
+ "epoch": 4.964192139737992,
+ "grad_norm": 0.014934048056602478,
+ "learning_rate": 0.0006,
+ "loss": 6.101555824279785,
+ "step": 357
+ },
+ {
+ "epoch": 4.978165938864628,
+ "grad_norm": 0.013979545794427395,
+ "learning_rate": 0.0006,
+ "loss": 6.161718845367432,
+ "step": 358
+ },
+ {
+ "epoch": 4.992139737991266,
+ "grad_norm": 0.01640394888818264,
+ "learning_rate": 0.0006,
+ "loss": 6.055768966674805,
+ "step": 359
+ },
+ {
+ "epoch": 5.0,
+ "grad_norm": 0.015135680325329304,
+ "learning_rate": 0.0006,
+ "loss": 6.169695854187012,
+ "step": 360
+ },
+ {
+ "epoch": 5.0,
+ "eval_loss": 6.16408109664917,
+ "eval_runtime": 59.3649,
+ "eval_samples_per_second": 41.135,
+ "eval_steps_per_second": 1.297,
+ "step": 360
+ },
+ {
+ "epoch": 5.013973799126638,
+ "grad_norm": 0.012849048711359501,
+ "learning_rate": 0.0006,
+ "loss": 6.055830001831055,
+ "step": 361
+ },
+ {
+ "epoch": 5.0279475982532755,
+ "grad_norm": 0.018169576302170753,
+ "learning_rate": 0.0006,
+ "loss": 6.082655429840088,
+ "step": 362
+ },
+ {
+ "epoch": 5.041921397379912,
+ "grad_norm": 0.024293430149555206,
+ "learning_rate": 0.0006,
+ "loss": 6.109032154083252,
+ "step": 363
+ },
+ {
+ "epoch": 5.05589519650655,
+ "grad_norm": 0.03266787901520729,
+ "learning_rate": 0.0006,
+ "loss": 6.030613899230957,
+ "step": 364
+ },
+ {
+ "epoch": 5.069868995633188,
+ "grad_norm": 0.03572266176342964,
+ "learning_rate": 0.0006,
+ "loss": 6.050390720367432,
+ "step": 365
+ },
+ {
+ "epoch": 5.083842794759826,
+ "grad_norm": 0.032625213265419006,
+ "learning_rate": 0.0006,
+ "loss": 6.169313430786133,
+ "step": 366
+ },
+ {
+ "epoch": 5.097816593886463,
+ "grad_norm": 0.04051872715353966,
+ "learning_rate": 0.0006,
+ "loss": 6.10398006439209,
+ "step": 367
+ },
+ {
+ "epoch": 5.1117903930131,
+ "grad_norm": 0.0351213738322258,
+ "learning_rate": 0.0006,
+ "loss": 6.121973514556885,
+ "step": 368
+ },
+ {
+ "epoch": 5.125764192139738,
+ "grad_norm": 0.03523759916424751,
+ "learning_rate": 0.0006,
+ "loss": 6.0626654624938965,
+ "step": 369
+ },
+ {
+ "epoch": 5.139737991266376,
+ "grad_norm": 0.02706182189285755,
+ "learning_rate": 0.0006,
+ "loss": 6.049354076385498,
+ "step": 370
+ },
+ {
+ "epoch": 5.153711790393013,
+ "grad_norm": 0.024212589487433434,
+ "learning_rate": 0.0006,
+ "loss": 6.034826278686523,
+ "step": 371
+ },
+ {
+ "epoch": 5.1676855895196505,
+ "grad_norm": 0.024798082187771797,
+ "learning_rate": 0.0006,
+ "loss": 6.004058837890625,
+ "step": 372
+ },
+ {
+ "epoch": 5.181659388646288,
+ "grad_norm": 0.02325567416846752,
+ "learning_rate": 0.0006,
+ "loss": 5.986461639404297,
+ "step": 373
+ },
+ {
+ "epoch": 5.195633187772926,
+ "grad_norm": 0.019060570746660233,
+ "learning_rate": 0.0006,
+ "loss": 6.059736251831055,
+ "step": 374
+ },
+ {
+ "epoch": 5.209606986899563,
+ "grad_norm": 0.016822976991534233,
+ "learning_rate": 0.0006,
+ "loss": 6.072957515716553,
+ "step": 375
+ },
+ {
+ "epoch": 5.223580786026201,
+ "grad_norm": 0.017218658700585365,
+ "learning_rate": 0.0006,
+ "loss": 6.031739234924316,
+ "step": 376
+ },
+ {
+ "epoch": 5.2375545851528384,
+ "grad_norm": 0.014889383688569069,
+ "learning_rate": 0.0006,
+ "loss": 6.025674343109131,
+ "step": 377
+ },
+ {
+ "epoch": 5.251528384279476,
+ "grad_norm": 0.01708107627928257,
+ "learning_rate": 0.0006,
+ "loss": 6.090451240539551,
+ "step": 378
+ },
+ {
+ "epoch": 5.265502183406113,
+ "grad_norm": 0.020260317251086235,
+ "learning_rate": 0.0006,
+ "loss": 6.094257354736328,
+ "step": 379
+ },
+ {
+ "epoch": 5.279475982532751,
+ "grad_norm": 0.020110400393605232,
+ "learning_rate": 0.0006,
+ "loss": 6.021188735961914,
+ "step": 380
+ },
+ {
+ "epoch": 5.293449781659389,
+ "grad_norm": 0.02027830481529236,
+ "learning_rate": 0.0006,
+ "loss": 6.022156238555908,
+ "step": 381
+ },
+ {
+ "epoch": 5.307423580786026,
+ "grad_norm": 0.022747062146663666,
+ "learning_rate": 0.0006,
+ "loss": 6.011836051940918,
+ "step": 382
+ },
+ {
+ "epoch": 5.321397379912664,
+ "grad_norm": 0.01990230567753315,
+ "learning_rate": 0.0006,
+ "loss": 6.003054618835449,
+ "step": 383
+ },
+ {
+ "epoch": 5.335371179039301,
+ "grad_norm": 0.01596238650381565,
+ "learning_rate": 0.0006,
+ "loss": 5.991410732269287,
+ "step": 384
+ },
+ {
+ "epoch": 5.349344978165939,
+ "grad_norm": 0.015847105532884598,
+ "learning_rate": 0.0006,
+ "loss": 6.100622177124023,
+ "step": 385
+ },
+ {
+ "epoch": 5.3633187772925766,
+ "grad_norm": 0.016179397702217102,
+ "learning_rate": 0.0006,
+ "loss": 6.032659530639648,
+ "step": 386
+ },
+ {
+ "epoch": 5.377292576419214,
+ "grad_norm": 0.018256602808833122,
+ "learning_rate": 0.0006,
+ "loss": 5.961983680725098,
+ "step": 387
+ },
+ {
+ "epoch": 5.391266375545851,
+ "grad_norm": 0.02805912122130394,
+ "learning_rate": 0.0006,
+ "loss": 5.983541965484619,
+ "step": 388
+ },
+ {
+ "epoch": 5.405240174672489,
+ "grad_norm": 0.039082255214452744,
+ "learning_rate": 0.0006,
+ "loss": 6.021474361419678,
+ "step": 389
+ },
+ {
+ "epoch": 5.419213973799127,
+ "grad_norm": 0.036757659167051315,
+ "learning_rate": 0.0006,
+ "loss": 6.005046367645264,
+ "step": 390
+ },
+ {
+ "epoch": 5.4331877729257645,
+ "grad_norm": 0.035277366638183594,
+ "learning_rate": 0.0006,
+ "loss": 6.036296844482422,
+ "step": 391
+ },
+ {
+ "epoch": 5.447161572052401,
+ "grad_norm": 0.034404635429382324,
+ "learning_rate": 0.0006,
+ "loss": 5.994539260864258,
+ "step": 392
+ },
+ {
+ "epoch": 5.461135371179039,
+ "grad_norm": 0.0377206988632679,
+ "learning_rate": 0.0006,
+ "loss": 5.941534519195557,
+ "step": 393
+ },
+ {
+ "epoch": 5.475109170305677,
+ "grad_norm": 0.0389922633767128,
+ "learning_rate": 0.0006,
+ "loss": 5.983644485473633,
+ "step": 394
+ },
+ {
+ "epoch": 5.489082969432315,
+ "grad_norm": 0.04176778718829155,
+ "learning_rate": 0.0006,
+ "loss": 6.030922889709473,
+ "step": 395
+ },
+ {
+ "epoch": 5.503056768558952,
+ "grad_norm": 0.0337153784930706,
+ "learning_rate": 0.0006,
+ "loss": 5.961367130279541,
+ "step": 396
+ },
+ {
+ "epoch": 5.517030567685589,
+ "grad_norm": 0.036116816103458405,
+ "learning_rate": 0.0006,
+ "loss": 5.819280624389648,
+ "step": 397
+ },
+ {
+ "epoch": 5.531004366812227,
+ "grad_norm": 0.030724041163921356,
+ "learning_rate": 0.0006,
+ "loss": 5.906380653381348,
+ "step": 398
+ },
+ {
+ "epoch": 5.544978165938865,
+ "grad_norm": 0.030264202505350113,
+ "learning_rate": 0.0006,
+ "loss": 5.975833415985107,
+ "step": 399
+ },
+ {
+ "epoch": 5.558951965065502,
+ "grad_norm": 0.032096318900585175,
+ "learning_rate": 0.0006,
+ "loss": 6.001348972320557,
+ "step": 400
+ },
+ {
+ "epoch": 5.5729257641921395,
+ "grad_norm": 0.030579356476664543,
+ "learning_rate": 0.0006,
+ "loss": 5.928768634796143,
+ "step": 401
+ },
+ {
+ "epoch": 5.586899563318777,
+ "grad_norm": 0.028241973370313644,
+ "learning_rate": 0.0006,
+ "loss": 5.92582893371582,
+ "step": 402
+ },
+ {
+ "epoch": 5.600873362445415,
+ "grad_norm": 0.023086953908205032,
+ "learning_rate": 0.0006,
+ "loss": 5.892926216125488,
+ "step": 403
+ },
+ {
+ "epoch": 5.614847161572053,
+ "grad_norm": 0.02547992393374443,
+ "learning_rate": 0.0006,
+ "loss": 5.903195381164551,
+ "step": 404
+ },
+ {
+ "epoch": 5.62882096069869,
+ "grad_norm": 0.023089831694960594,
+ "learning_rate": 0.0006,
+ "loss": 5.912033557891846,
+ "step": 405
+ },
+ {
+ "epoch": 5.642794759825327,
+ "grad_norm": 0.025446368381381035,
+ "learning_rate": 0.0006,
+ "loss": 5.926138401031494,
+ "step": 406
+ },
+ {
+ "epoch": 5.656768558951965,
+ "grad_norm": 0.03748747706413269,
+ "learning_rate": 0.0006,
+ "loss": 5.913451194763184,
+ "step": 407
+ },
+ {
+ "epoch": 5.670742358078603,
+ "grad_norm": 0.03493810072541237,
+ "learning_rate": 0.0006,
+ "loss": 5.97633695602417,
+ "step": 408
+ },
+ {
+ "epoch": 5.68471615720524,
+ "grad_norm": 0.023131275549530983,
+ "learning_rate": 0.0006,
+ "loss": 5.899571418762207,
+ "step": 409
+ },
+ {
+ "epoch": 5.698689956331878,
+ "grad_norm": 0.02986014075577259,
+ "learning_rate": 0.0006,
+ "loss": 5.894903182983398,
+ "step": 410
+ },
+ {
+ "epoch": 5.712663755458515,
+ "grad_norm": 0.030171144753694534,
+ "learning_rate": 0.0006,
+ "loss": 5.876595497131348,
+ "step": 411
+ },
+ {
+ "epoch": 5.726637554585153,
+ "grad_norm": 0.029377546161413193,
+ "learning_rate": 0.0006,
+ "loss": 5.911637306213379,
+ "step": 412
+ },
+ {
+ "epoch": 5.74061135371179,
+ "grad_norm": 0.029829490929841995,
+ "learning_rate": 0.0006,
+ "loss": 5.819629669189453,
+ "step": 413
+ },
+ {
+ "epoch": 5.754585152838428,
+ "grad_norm": 0.024139299988746643,
+ "learning_rate": 0.0006,
+ "loss": 5.81728458404541,
+ "step": 414
+ },
+ {
+ "epoch": 5.7685589519650655,
+ "grad_norm": 0.021453432738780975,
+ "learning_rate": 0.0006,
+ "loss": 5.987326145172119,
+ "step": 415
+ },
+ {
+ "epoch": 5.782532751091703,
+ "grad_norm": 0.017942246049642563,
+ "learning_rate": 0.0006,
+ "loss": 5.893008232116699,
+ "step": 416
+ },
+ {
+ "epoch": 5.796506550218341,
+ "grad_norm": 0.019724637269973755,
+ "learning_rate": 0.0006,
+ "loss": 5.912441253662109,
+ "step": 417
+ },
+ {
+ "epoch": 5.810480349344978,
+ "grad_norm": 0.018597912043333054,
+ "learning_rate": 0.0006,
+ "loss": 5.860394477844238,
+ "step": 418
+ },
+ {
+ "epoch": 5.824454148471616,
+ "grad_norm": 0.016232412308454514,
+ "learning_rate": 0.0006,
+ "loss": 5.94991397857666,
+ "step": 419
+ },
+ {
+ "epoch": 5.8384279475982535,
+ "grad_norm": 0.01667204685509205,
+ "learning_rate": 0.0006,
+ "loss": 5.8506364822387695,
+ "step": 420
+ },
+ {
+ "epoch": 5.85240174672489,
+ "grad_norm": 0.014232151210308075,
+ "learning_rate": 0.0006,
+ "loss": 5.928424835205078,
+ "step": 421
+ },
+ {
+ "epoch": 5.866375545851528,
+ "grad_norm": 0.01570476032793522,
+ "learning_rate": 0.0006,
+ "loss": 5.953692436218262,
+ "step": 422
+ },
+ {
+ "epoch": 5.880349344978166,
+ "grad_norm": 0.01586179807782173,
+ "learning_rate": 0.0006,
+ "loss": 5.820014476776123,
+ "step": 423
+ },
+ {
+ "epoch": 5.894323144104804,
+ "grad_norm": 0.014729145914316177,
+ "learning_rate": 0.0006,
+ "loss": 5.860154628753662,
+ "step": 424
+ },
+ {
+ "epoch": 5.908296943231441,
+ "grad_norm": 0.013111459091305733,
+ "learning_rate": 0.0006,
+ "loss": 5.85988187789917,
+ "step": 425
+ },
+ {
+ "epoch": 5.922270742358078,
+ "grad_norm": 0.012937922962009907,
+ "learning_rate": 0.0006,
+ "loss": 5.756265640258789,
+ "step": 426
+ },
+ {
+ "epoch": 5.936244541484716,
+ "grad_norm": 0.01377453189343214,
+ "learning_rate": 0.0006,
+ "loss": 5.8333048820495605,
+ "step": 427
+ },
+ {
+ "epoch": 5.950218340611354,
+ "grad_norm": 0.014054795727133751,
+ "learning_rate": 0.0006,
+ "loss": 5.803333759307861,
+ "step": 428
+ },
+ {
+ "epoch": 5.964192139737992,
+ "grad_norm": 0.01674959622323513,
+ "learning_rate": 0.0006,
+ "loss": 5.8718485832214355,
+ "step": 429
+ },
+ {
+ "epoch": 5.978165938864628,
+ "grad_norm": 0.018733017146587372,
+ "learning_rate": 0.0006,
+ "loss": 5.875979900360107,
+ "step": 430
+ },
+ {
+ "epoch": 5.992139737991266,
+ "grad_norm": 0.02088983915746212,
+ "learning_rate": 0.0006,
+ "loss": 5.763634204864502,
+ "step": 431
+ },
+ {
+ "epoch": 6.0,
+ "grad_norm": 0.026705985888838768,
+ "learning_rate": 0.0006,
+ "loss": 5.6922478675842285,
+ "step": 432
+ },
+ {
+ "epoch": 6.0,
+ "eval_loss": 5.847958087921143,
+ "eval_runtime": 58.5451,
+ "eval_samples_per_second": 41.711,
+ "eval_steps_per_second": 1.315,
+ "step": 432
+ },
+ {
+ "epoch": 6.013973799126638,
+ "grad_norm": 0.03181544691324234,
+ "learning_rate": 0.0006,
+ "loss": 5.786368370056152,
+ "step": 433
+ },
+ {
+ "epoch": 6.0279475982532755,
+ "grad_norm": 0.03238112851977348,
+ "learning_rate": 0.0006,
+ "loss": 5.810310363769531,
+ "step": 434
+ },
+ {
+ "epoch": 6.041921397379912,
+ "grad_norm": 0.03916756808757782,
+ "learning_rate": 0.0006,
+ "loss": 5.7761993408203125,
+ "step": 435
+ },
+ {
+ "epoch": 6.05589519650655,
+ "grad_norm": 0.044009730219841,
+ "learning_rate": 0.0006,
+ "loss": 5.898112773895264,
+ "step": 436
+ },
+ {
+ "epoch": 6.069868995633188,
+ "grad_norm": 0.045235246419906616,
+ "learning_rate": 0.0006,
+ "loss": 5.876371383666992,
+ "step": 437
+ },
+ {
+ "epoch": 6.083842794759826,
+ "grad_norm": 0.050218384712934494,
+ "learning_rate": 0.0006,
+ "loss": 5.805103302001953,
+ "step": 438
+ },
+ {
+ "epoch": 6.097816593886463,
+ "grad_norm": 0.05444490164518356,
+ "learning_rate": 0.0006,
+ "loss": 5.826424598693848,
+ "step": 439
+ },
+ {
+ "epoch": 6.1117903930131,
+ "grad_norm": 0.04890982061624527,
+ "learning_rate": 0.0006,
+ "loss": 5.7735395431518555,
+ "step": 440
+ },
+ {
+ "epoch": 6.125764192139738,
+ "grad_norm": 0.05478639155626297,
+ "learning_rate": 0.0006,
+ "loss": 5.823677062988281,
+ "step": 441
+ },
+ {
+ "epoch": 6.139737991266376,
+ "grad_norm": 0.057562313973903656,
+ "learning_rate": 0.0006,
+ "loss": 5.889334678649902,
+ "step": 442
+ },
+ {
+ "epoch": 6.153711790393013,
+ "grad_norm": 0.06447703391313553,
+ "learning_rate": 0.0006,
+ "loss": 5.84621524810791,
+ "step": 443
+ },
+ {
+ "epoch": 6.1676855895196505,
+ "grad_norm": 0.048861872404813766,
+ "learning_rate": 0.0006,
+ "loss": 5.931595802307129,
+ "step": 444
+ },
+ {
+ "epoch": 6.181659388646288,
+ "grad_norm": 0.05521339178085327,
+ "learning_rate": 0.0006,
+ "loss": 5.8229780197143555,
+ "step": 445
+ },
+ {
+ "epoch": 6.195633187772926,
+ "grad_norm": 0.053666990250349045,
+ "learning_rate": 0.0006,
+ "loss": 5.879191875457764,
+ "step": 446
+ },
+ {
+ "epoch": 6.209606986899563,
+ "grad_norm": 0.0451025515794754,
+ "learning_rate": 0.0006,
+ "loss": 5.801628112792969,
+ "step": 447
+ },
+ {
+ "epoch": 6.223580786026201,
+ "grad_norm": 0.04291655868291855,
+ "learning_rate": 0.0006,
+ "loss": 5.914680480957031,
+ "step": 448
+ },
+ {
+ "epoch": 6.2375545851528384,
+ "grad_norm": 0.035102490335702896,
+ "learning_rate": 0.0006,
+ "loss": 5.843682765960693,
+ "step": 449
+ },
+ {
+ "epoch": 6.251528384279476,
+ "grad_norm": 0.03403995931148529,
+ "learning_rate": 0.0006,
+ "loss": 5.827154159545898,
+ "step": 450
+ },
+ {
+ "epoch": 6.265502183406113,
+ "grad_norm": 0.03444375470280647,
+ "learning_rate": 0.0006,
+ "loss": 5.800136566162109,
+ "step": 451
+ },
+ {
+ "epoch": 6.279475982532751,
+ "grad_norm": 0.03165159001946449,
+ "learning_rate": 0.0006,
+ "loss": 5.906252384185791,
+ "step": 452
+ },
+ {
+ "epoch": 6.293449781659389,
+ "grad_norm": 0.026153093203902245,
+ "learning_rate": 0.0006,
+ "loss": 5.7423200607299805,
+ "step": 453
+ },
+ {
+ "epoch": 6.307423580786026,
+ "grad_norm": 0.024357657879590988,
+ "learning_rate": 0.0006,
+ "loss": 5.701364517211914,
+ "step": 454
+ },
+ {
+ "epoch": 6.321397379912664,
+ "grad_norm": 0.021508049219846725,
+ "learning_rate": 0.0006,
+ "loss": 5.8328094482421875,
+ "step": 455
+ },
+ {
+ "epoch": 6.335371179039301,
+ "grad_norm": 0.017313921824097633,
+ "learning_rate": 0.0006,
+ "loss": 5.792145252227783,
+ "step": 456
+ },
+ {
+ "epoch": 6.349344978165939,
+ "grad_norm": 0.018563397228717804,
+ "learning_rate": 0.0006,
+ "loss": 5.732672214508057,
+ "step": 457
+ },
+ {
+ "epoch": 6.3633187772925766,
+ "grad_norm": 0.016568679362535477,
+ "learning_rate": 0.0006,
+ "loss": 5.81948184967041,
+ "step": 458
+ },
+ {
+ "epoch": 6.377292576419214,
+ "grad_norm": 0.014133770950138569,
+ "learning_rate": 0.0006,
+ "loss": 5.734982490539551,
+ "step": 459
+ },
+ {
+ "epoch": 6.391266375545851,
+ "grad_norm": 0.014114447869360447,
+ "learning_rate": 0.0006,
+ "loss": 5.70073127746582,
+ "step": 460
+ },
+ {
+ "epoch": 6.405240174672489,
+ "grad_norm": 0.012907393276691437,
+ "learning_rate": 0.0006,
+ "loss": 5.758626937866211,
+ "step": 461
+ },
+ {
+ "epoch": 6.419213973799127,
+ "grad_norm": 0.012673444114625454,
+ "learning_rate": 0.0006,
+ "loss": 5.775138854980469,
+ "step": 462
+ },
+ {
+ "epoch": 6.4331877729257645,
+ "grad_norm": 0.01311410591006279,
+ "learning_rate": 0.0006,
+ "loss": 5.70920991897583,
+ "step": 463
+ },
+ {
+ "epoch": 6.447161572052401,
+ "grad_norm": 0.012935544364154339,
+ "learning_rate": 0.0006,
+ "loss": 5.781479835510254,
+ "step": 464
+ },
+ {
+ "epoch": 6.461135371179039,
+ "grad_norm": 0.01439738366752863,
+ "learning_rate": 0.0006,
+ "loss": 5.691315650939941,
+ "step": 465
+ },
+ {
+ "epoch": 6.475109170305677,
+ "grad_norm": 0.01205186452716589,
+ "learning_rate": 0.0006,
+ "loss": 5.662499904632568,
+ "step": 466
+ },
+ {
+ "epoch": 6.489082969432315,
+ "grad_norm": 0.011575652286410332,
+ "learning_rate": 0.0006,
+ "loss": 5.6271772384643555,
+ "step": 467
+ },
+ {
+ "epoch": 6.503056768558952,
+ "grad_norm": 0.011672502383589745,
+ "learning_rate": 0.0006,
+ "loss": 5.761662006378174,
+ "step": 468
+ },
+ {
+ "epoch": 6.517030567685589,
+ "grad_norm": 0.011541751213371754,
+ "learning_rate": 0.0006,
+ "loss": 5.762078285217285,
+ "step": 469
+ },
+ {
+ "epoch": 6.531004366812227,
+ "grad_norm": 0.011396365240216255,
+ "learning_rate": 0.0006,
+ "loss": 5.67873477935791,
+ "step": 470
+ },
+ {
+ "epoch": 6.544978165938865,
+ "grad_norm": 0.010688499547541142,
+ "learning_rate": 0.0006,
+ "loss": 5.717051982879639,
+ "step": 471
+ },
+ {
+ "epoch": 6.558951965065502,
+ "grad_norm": 0.012224317528307438,
+ "learning_rate": 0.0006,
+ "loss": 5.707948684692383,
+ "step": 472
+ },
+ {
+ "epoch": 6.5729257641921395,
+ "grad_norm": 0.011856825090944767,
+ "learning_rate": 0.0006,
+ "loss": 5.70878791809082,
+ "step": 473
+ },
+ {
+ "epoch": 6.586899563318777,
+ "grad_norm": 0.01199064590036869,
+ "learning_rate": 0.0006,
+ "loss": 5.708697319030762,
+ "step": 474
+ },
+ {
+ "epoch": 6.600873362445415,
+ "grad_norm": 0.01219446212053299,
+ "learning_rate": 0.0006,
+ "loss": 5.6061577796936035,
+ "step": 475
+ },
+ {
+ "epoch": 6.614847161572053,
+ "grad_norm": 0.013397484086453915,
+ "learning_rate": 0.0006,
+ "loss": 5.624789714813232,
+ "step": 476
+ },
+ {
+ "epoch": 6.62882096069869,
+ "grad_norm": 0.01483136136084795,
+ "learning_rate": 0.0006,
+ "loss": 5.721141338348389,
+ "step": 477
+ },
+ {
+ "epoch": 6.642794759825327,
+ "grad_norm": 0.0194488987326622,
+ "learning_rate": 0.0006,
+ "loss": 5.604279518127441,
+ "step": 478
+ },
+ {
+ "epoch": 6.656768558951965,
+ "grad_norm": 0.02079896256327629,
+ "learning_rate": 0.0006,
+ "loss": 5.586322784423828,
+ "step": 479
+ },
+ {
+ "epoch": 6.670742358078603,
+ "grad_norm": 0.018996229395270348,
+ "learning_rate": 0.0006,
+ "loss": 5.5987772941589355,
+ "step": 480
+ },
+ {
+ "epoch": 6.68471615720524,
+ "grad_norm": 0.015779603272676468,
+ "learning_rate": 0.0006,
+ "loss": 5.588602066040039,
+ "step": 481
+ },
+ {
+ "epoch": 6.698689956331878,
+ "grad_norm": 0.015322973020374775,
+ "learning_rate": 0.0006,
+ "loss": 5.686467170715332,
+ "step": 482
+ },
+ {
+ "epoch": 6.712663755458515,
+ "grad_norm": 0.015282213687896729,
+ "learning_rate": 0.0006,
+ "loss": 5.664676666259766,
+ "step": 483
+ },
+ {
+ "epoch": 6.726637554585153,
+ "grad_norm": 0.016589272767305374,
+ "learning_rate": 0.0006,
+ "loss": 5.666739463806152,
+ "step": 484
+ },
+ {
+ "epoch": 6.74061135371179,
+ "grad_norm": 0.01811421848833561,
+ "learning_rate": 0.0006,
+ "loss": 5.647593021392822,
+ "step": 485
+ },
+ {
+ "epoch": 6.754585152838428,
+ "grad_norm": 0.019652029499411583,
+ "learning_rate": 0.0006,
+ "loss": 5.628726959228516,
+ "step": 486
+ },
+ {
+ "epoch": 6.7685589519650655,
+ "grad_norm": 0.02118665538728237,
+ "learning_rate": 0.0006,
+ "loss": 5.6580705642700195,
+ "step": 487
+ },
+ {
+ "epoch": 6.782532751091703,
+ "grad_norm": 0.02237832546234131,
+ "learning_rate": 0.0006,
+ "loss": 5.635931968688965,
+ "step": 488
+ },
+ {
+ "epoch": 6.796506550218341,
+ "grad_norm": 0.023897631093859673,
+ "learning_rate": 0.0006,
+ "loss": 5.517949104309082,
+ "step": 489
+ },
+ {
+ "epoch": 6.810480349344978,
+ "grad_norm": 0.02442227490246296,
+ "learning_rate": 0.0006,
+ "loss": 5.600021839141846,
+ "step": 490
+ },
+ {
+ "epoch": 6.824454148471616,
+ "grad_norm": 0.024675287306308746,
+ "learning_rate": 0.0006,
+ "loss": 5.6639909744262695,
+ "step": 491
+ },
+ {
+ "epoch": 6.8384279475982535,
+ "grad_norm": 0.030372392386198044,
+ "learning_rate": 0.0006,
+ "loss": 5.582888603210449,
+ "step": 492
+ },
+ {
+ "epoch": 6.85240174672489,
+ "grad_norm": 0.03237690031528473,
+ "learning_rate": 0.0006,
+ "loss": 5.573896884918213,
+ "step": 493
+ },
+ {
+ "epoch": 6.866375545851528,
+ "grad_norm": 0.03511589393019676,
+ "learning_rate": 0.0006,
+ "loss": 5.565878868103027,
+ "step": 494
+ },
+ {
+ "epoch": 6.880349344978166,
+ "grad_norm": 0.03883938118815422,
+ "learning_rate": 0.0006,
+ "loss": 5.58540678024292,
+ "step": 495
+ },
+ {
+ "epoch": 6.894323144104804,
+ "grad_norm": 0.04175502806901932,
+ "learning_rate": 0.0006,
+ "loss": 5.6095194816589355,
+ "step": 496
+ },
+ {
+ "epoch": 6.908296943231441,
+ "grad_norm": 0.04013441503047943,
+ "learning_rate": 0.0006,
+ "loss": 5.654401779174805,
+ "step": 497
+ },
+ {
+ "epoch": 6.922270742358078,
+ "grad_norm": 0.04618770629167557,
+ "learning_rate": 0.0006,
+ "loss": 5.6483473777771,
+ "step": 498
+ },
+ {
+ "epoch": 6.936244541484716,
+ "grad_norm": 0.04507308453321457,
+ "learning_rate": 0.0006,
+ "loss": 5.5856218338012695,
+ "step": 499
+ },
+ {
+ "epoch": 6.950218340611354,
+ "grad_norm": 0.03151383996009827,
+ "learning_rate": 0.0006,
+ "loss": 5.594086647033691,
+ "step": 500
+ },
+ {
+ "epoch": 6.964192139737992,
+ "grad_norm": 0.026883797720074654,
+ "learning_rate": 0.0006,
+ "loss": 5.607676029205322,
+ "step": 501
+ },
+ {
+ "epoch": 6.978165938864628,
+ "grad_norm": 0.02981836162507534,
+ "learning_rate": 0.0006,
+ "loss": 5.563666820526123,
+ "step": 502
+ },
+ {
+ "epoch": 6.992139737991266,
+ "grad_norm": 0.03233477845788002,
+ "learning_rate": 0.0006,
+ "loss": 5.667543411254883,
+ "step": 503
+ },
+ {
+ "epoch": 7.0,
+ "grad_norm": 0.029927456751465797,
+ "learning_rate": 0.0006,
+ "loss": 5.733482837677002,
+ "step": 504
+ },
+ {
+ "epoch": 7.0,
+ "eval_loss": 5.679076671600342,
+ "eval_runtime": 59.5965,
+ "eval_samples_per_second": 40.976,
+ "eval_steps_per_second": 1.292,
+ "step": 504
+ },
+ {
+ "epoch": 7.013973799126638,
+ "grad_norm": 0.027081597596406937,
+ "learning_rate": 0.0006,
+ "loss": 5.634487152099609,
+ "step": 505
+ },
+ {
+ "epoch": 7.0279475982532755,
+ "grad_norm": 0.031959421932697296,
+ "learning_rate": 0.0006,
+ "loss": 5.643294811248779,
+ "step": 506
+ },
+ {
+ "epoch": 7.041921397379912,
+ "grad_norm": 0.02802063524723053,
+ "learning_rate": 0.0006,
+ "loss": 5.573239326477051,
+ "step": 507
+ },
+ {
+ "epoch": 7.05589519650655,
+ "grad_norm": 0.032579705119132996,
+ "learning_rate": 0.0006,
+ "loss": 5.627150058746338,
+ "step": 508
+ },
+ {
+ "epoch": 7.069868995633188,
+ "grad_norm": 0.02503928542137146,
+ "learning_rate": 0.0006,
+ "loss": 5.6045308113098145,
+ "step": 509
+ },
+ {
+ "epoch": 7.083842794759826,
+ "grad_norm": 0.02421317622065544,
+ "learning_rate": 0.0006,
+ "loss": 5.522153377532959,
+ "step": 510
+ },
+ {
+ "epoch": 7.097816593886463,
+ "grad_norm": 0.021983426064252853,
+ "learning_rate": 0.0006,
+ "loss": 5.632939338684082,
+ "step": 511
+ },
+ {
+ "epoch": 7.1117903930131,
+ "grad_norm": 0.021933183073997498,
+ "learning_rate": 0.0006,
+ "loss": 5.592185974121094,
+ "step": 512
+ },
+ {
+ "epoch": 7.125764192139738,
+ "grad_norm": 0.02287031151354313,
+ "learning_rate": 0.0006,
+ "loss": 5.613700866699219,
+ "step": 513
+ },
+ {
+ "epoch": 7.139737991266376,
+ "grad_norm": 0.02281602844595909,
+ "learning_rate": 0.0006,
+ "loss": 5.551383018493652,
+ "step": 514
+ },
+ {
+ "epoch": 7.153711790393013,
+ "grad_norm": 0.021050360053777695,
+ "learning_rate": 0.0006,
+ "loss": 5.55275821685791,
+ "step": 515
+ },
+ {
+ "epoch": 7.1676855895196505,
+ "grad_norm": 0.018299926072359085,
+ "learning_rate": 0.0006,
+ "loss": 5.528225421905518,
+ "step": 516
+ },
+ {
+ "epoch": 7.181659388646288,
+ "grad_norm": 0.02024853602051735,
+ "learning_rate": 0.0006,
+ "loss": 5.522453784942627,
+ "step": 517
+ },
+ {
+ "epoch": 7.195633187772926,
+ "grad_norm": 0.02135239914059639,
+ "learning_rate": 0.0006,
+ "loss": 5.527522563934326,
+ "step": 518
+ },
+ {
+ "epoch": 7.209606986899563,
+ "grad_norm": 0.023394184187054634,
+ "learning_rate": 0.0006,
+ "loss": 5.4535017013549805,
+ "step": 519
+ },
+ {
+ "epoch": 7.223580786026201,
+ "grad_norm": 0.025952080264687538,
+ "learning_rate": 0.0006,
+ "loss": 5.5374674797058105,
+ "step": 520
+ },
+ {
+ "epoch": 7.2375545851528384,
+ "grad_norm": 0.022287718951702118,
+ "learning_rate": 0.0006,
+ "loss": 5.493753433227539,
+ "step": 521
+ },
+ {
+ "epoch": 7.251528384279476,
+ "grad_norm": 0.02018447406589985,
+ "learning_rate": 0.0006,
+ "loss": 5.510575294494629,
+ "step": 522
+ },
+ {
+ "epoch": 7.265502183406113,
+ "grad_norm": 0.02050507813692093,
+ "learning_rate": 0.0006,
+ "loss": 5.466026782989502,
+ "step": 523
+ },
+ {
+ "epoch": 7.279475982532751,
+ "grad_norm": 0.023688802495598793,
+ "learning_rate": 0.0006,
+ "loss": 5.485815525054932,
+ "step": 524
+ },
+ {
+ "epoch": 7.293449781659389,
+ "grad_norm": 0.021078843623399734,
+ "learning_rate": 0.0006,
+ "loss": 5.415581703186035,
+ "step": 525
+ },
+ {
+ "epoch": 7.307423580786026,
+ "grad_norm": 0.018266011029481888,
+ "learning_rate": 0.0006,
+ "loss": 5.402815818786621,
+ "step": 526
+ },
+ {
+ "epoch": 7.321397379912664,
+ "grad_norm": 0.019693493843078613,
+ "learning_rate": 0.0006,
+ "loss": 5.504674911499023,
+ "step": 527
+ },
+ {
+ "epoch": 7.335371179039301,
+ "grad_norm": 0.02031373605132103,
+ "learning_rate": 0.0006,
+ "loss": 5.503837585449219,
+ "step": 528
+ },
+ {
+ "epoch": 7.349344978165939,
+ "grad_norm": 0.017636626958847046,
+ "learning_rate": 0.0006,
+ "loss": 5.530580997467041,
+ "step": 529
+ },
+ {
+ "epoch": 7.3633187772925766,
+ "grad_norm": 0.01787244901061058,
+ "learning_rate": 0.0006,
+ "loss": 5.450218677520752,
+ "step": 530
+ },
+ {
+ "epoch": 7.377292576419214,
+ "grad_norm": 0.0170669574290514,
+ "learning_rate": 0.0006,
+ "loss": 5.481570243835449,
+ "step": 531
+ },
+ {
+ "epoch": 7.391266375545851,
+ "grad_norm": 0.01802165061235428,
+ "learning_rate": 0.0006,
+ "loss": 5.386394500732422,
+ "step": 532
+ },
+ {
+ "epoch": 7.405240174672489,
+ "grad_norm": 0.022949472069740295,
+ "learning_rate": 0.0006,
+ "loss": 5.424929618835449,
+ "step": 533
+ },
+ {
+ "epoch": 7.419213973799127,
+ "grad_norm": 0.034213390201330185,
+ "learning_rate": 0.0006,
+ "loss": 5.346663475036621,
+ "step": 534
+ },
+ {
+ "epoch": 7.4331877729257645,
+ "grad_norm": 0.03847593814134598,
+ "learning_rate": 0.0006,
+ "loss": 5.418482780456543,
+ "step": 535
+ },
+ {
+ "epoch": 7.447161572052401,
+ "grad_norm": 0.029393676668405533,
+ "learning_rate": 0.0006,
+ "loss": 5.456090927124023,
+ "step": 536
+ },
+ {
+ "epoch": 7.461135371179039,
+ "grad_norm": 0.03388667106628418,
+ "learning_rate": 0.0006,
+ "loss": 5.572983741760254,
+ "step": 537
+ },
+ {
+ "epoch": 7.475109170305677,
+ "grad_norm": 0.039690613746643066,
+ "learning_rate": 0.0006,
+ "loss": 5.414067268371582,
+ "step": 538
+ },
+ {
+ "epoch": 7.489082969432315,
+ "grad_norm": 0.03630776330828667,
+ "learning_rate": 0.0006,
+ "loss": 5.523739814758301,
+ "step": 539
+ },
+ {
+ "epoch": 7.503056768558952,
+ "grad_norm": 0.03356870263814926,
+ "learning_rate": 0.0006,
+ "loss": 5.444767951965332,
+ "step": 540
+ },
+ {
+ "epoch": 7.517030567685589,
+ "grad_norm": 0.030812304466962814,
+ "learning_rate": 0.0006,
+ "loss": 5.461244106292725,
+ "step": 541
+ },
+ {
+ "epoch": 7.531004366812227,
+ "grad_norm": 0.03319217637181282,
+ "learning_rate": 0.0006,
+ "loss": 5.502161026000977,
+ "step": 542
+ },
+ {
+ "epoch": 7.544978165938865,
+ "grad_norm": 0.032997481524944305,
+ "learning_rate": 0.0006,
+ "loss": 5.4162278175354,
+ "step": 543
+ },
+ {
+ "epoch": 7.558951965065502,
+ "grad_norm": 0.03364962339401245,
+ "learning_rate": 0.0006,
+ "loss": 5.415736198425293,
+ "step": 544
+ },
+ {
+ "epoch": 7.5729257641921395,
+ "grad_norm": 0.0344221405684948,
+ "learning_rate": 0.0006,
+ "loss": 5.541967868804932,
+ "step": 545
+ },
+ {
+ "epoch": 7.586899563318777,
+ "grad_norm": 0.029609503224492073,
+ "learning_rate": 0.0006,
+ "loss": 5.440614223480225,
+ "step": 546
+ },
+ {
+ "epoch": 7.600873362445415,
+ "grad_norm": 0.029055926948785782,
+ "learning_rate": 0.0006,
+ "loss": 5.462865352630615,
+ "step": 547
+ },
+ {
+ "epoch": 7.614847161572053,
+ "grad_norm": 0.02658848837018013,
+ "learning_rate": 0.0006,
+ "loss": 5.444467544555664,
+ "step": 548
+ },
+ {
+ "epoch": 7.62882096069869,
+ "grad_norm": 0.026276404038071632,
+ "learning_rate": 0.0006,
+ "loss": 5.462278842926025,
+ "step": 549
+ },
+ {
+ "epoch": 7.642794759825327,
+ "grad_norm": 0.0286889486014843,
+ "learning_rate": 0.0006,
+ "loss": 5.4849534034729,
+ "step": 550
+ },
+ {
+ "epoch": 7.656768558951965,
+ "grad_norm": 0.026849817484617233,
+ "learning_rate": 0.0006,
+ "loss": 5.431473731994629,
+ "step": 551
+ },
+ {
+ "epoch": 7.670742358078603,
+ "grad_norm": 0.02312396466732025,
+ "learning_rate": 0.0006,
+ "loss": 5.45462703704834,
+ "step": 552
+ },
+ {
+ "epoch": 7.68471615720524,
+ "grad_norm": 0.026197485625743866,
+ "learning_rate": 0.0006,
+ "loss": 5.452552318572998,
+ "step": 553
+ },
+ {
+ "epoch": 7.698689956331878,
+ "grad_norm": 0.02747255191206932,
+ "learning_rate": 0.0006,
+ "loss": 5.447169780731201,
+ "step": 554
+ },
+ {
+ "epoch": 7.712663755458515,
+ "grad_norm": 0.028123432770371437,
+ "learning_rate": 0.0006,
+ "loss": 5.394400596618652,
+ "step": 555
+ },
+ {
+ "epoch": 7.726637554585153,
+ "grad_norm": 0.02599870041012764,
+ "learning_rate": 0.0006,
+ "loss": 5.398341178894043,
+ "step": 556
+ },
+ {
+ "epoch": 7.74061135371179,
+ "grad_norm": 0.022171657532453537,
+ "learning_rate": 0.0006,
+ "loss": 5.368820667266846,
+ "step": 557
+ },
+ {
+ "epoch": 7.754585152838428,
+ "grad_norm": 0.022309480234980583,
+ "learning_rate": 0.0006,
+ "loss": 5.416772365570068,
+ "step": 558
+ },
+ {
+ "epoch": 7.7685589519650655,
+ "grad_norm": 0.024074165150523186,
+ "learning_rate": 0.0006,
+ "loss": 5.359002113342285,
+ "step": 559
+ },
+ {
+ "epoch": 7.782532751091703,
+ "grad_norm": 0.02636653557419777,
+ "learning_rate": 0.0006,
+ "loss": 5.484368324279785,
+ "step": 560
+ },
+ {
+ "epoch": 7.796506550218341,
+ "grad_norm": 0.02203752100467682,
+ "learning_rate": 0.0006,
+ "loss": 5.375498294830322,
+ "step": 561
+ },
+ {
+ "epoch": 7.810480349344978,
+ "grad_norm": 0.019924819469451904,
+ "learning_rate": 0.0006,
+ "loss": 5.334723472595215,
+ "step": 562
+ },
+ {
+ "epoch": 7.824454148471616,
+ "grad_norm": 0.018755966797471046,
+ "learning_rate": 0.0006,
+ "loss": 5.348781108856201,
+ "step": 563
+ },
+ {
+ "epoch": 7.8384279475982535,
+ "grad_norm": 0.016747845336794853,
+ "learning_rate": 0.0006,
+ "loss": 5.433079719543457,
+ "step": 564
+ },
+ {
+ "epoch": 7.85240174672489,
+ "grad_norm": 0.0170395877212286,
+ "learning_rate": 0.0006,
+ "loss": 5.346794128417969,
+ "step": 565
+ },
+ {
+ "epoch": 7.866375545851528,
+ "grad_norm": 0.016074176877737045,
+ "learning_rate": 0.0006,
+ "loss": 5.3277740478515625,
+ "step": 566
+ },
+ {
+ "epoch": 7.880349344978166,
+ "grad_norm": 0.014119806699454784,
+ "learning_rate": 0.0006,
+ "loss": 5.483551979064941,
+ "step": 567
+ },
+ {
+ "epoch": 7.894323144104804,
+ "grad_norm": 0.015270394273102283,
+ "learning_rate": 0.0006,
+ "loss": 5.378519535064697,
+ "step": 568
+ },
+ {
+ "epoch": 7.908296943231441,
+ "grad_norm": 0.015677539631724358,
+ "learning_rate": 0.0006,
+ "loss": 5.290210247039795,
+ "step": 569
+ },
+ {
+ "epoch": 7.922270742358078,
+ "grad_norm": 0.015930423513054848,
+ "learning_rate": 0.0006,
+ "loss": 5.427360534667969,
+ "step": 570
+ },
+ {
+ "epoch": 7.936244541484716,
+ "grad_norm": 0.016641564667224884,
+ "learning_rate": 0.0006,
+ "loss": 5.301599979400635,
+ "step": 571
+ },
+ {
+ "epoch": 7.950218340611354,
+ "grad_norm": 0.018293552100658417,
+ "learning_rate": 0.0006,
+ "loss": 5.277889251708984,
+ "step": 572
+ },
+ {
+ "epoch": 7.964192139737992,
+ "grad_norm": 0.017618799582123756,
+ "learning_rate": 0.0006,
+ "loss": 5.297194480895996,
+ "step": 573
+ },
+ {
+ "epoch": 7.978165938864628,
+ "grad_norm": 0.015034242533147335,
+ "learning_rate": 0.0006,
+ "loss": 5.42428731918335,
+ "step": 574
+ },
+ {
+ "epoch": 7.992139737991266,
+ "grad_norm": 0.01617511734366417,
+ "learning_rate": 0.0006,
+ "loss": 5.420044898986816,
+ "step": 575
+ },
+ {
+ "epoch": 8.0,
+ "grad_norm": 0.01837257295846939,
+ "learning_rate": 0.0006,
+ "loss": 5.240387916564941,
+ "step": 576
+ },
+ {
+ "epoch": 8.0,
+ "eval_loss": 5.423932075500488,
+ "eval_runtime": 58.0841,
+ "eval_samples_per_second": 42.042,
+ "eval_steps_per_second": 1.326,
+ "step": 576
+ },
+ {
+ "epoch": 8.013973799126637,
+ "grad_norm": 0.020953809842467308,
+ "learning_rate": 0.0006,
+ "loss": 5.4320783615112305,
+ "step": 577
+ },
+ {
+ "epoch": 8.027947598253276,
+ "grad_norm": 0.0227745920419693,
+ "learning_rate": 0.0006,
+ "loss": 5.310197830200195,
+ "step": 578
+ },
+ {
+ "epoch": 8.041921397379912,
+ "grad_norm": 0.021372603252530098,
+ "learning_rate": 0.0006,
+ "loss": 5.339468002319336,
+ "step": 579
+ },
+ {
+ "epoch": 8.055895196506551,
+ "grad_norm": 0.022401731461286545,
+ "learning_rate": 0.0006,
+ "loss": 5.279829978942871,
+ "step": 580
+ },
+ {
+ "epoch": 8.069868995633188,
+ "grad_norm": 0.023942379280924797,
+ "learning_rate": 0.0006,
+ "loss": 5.372323989868164,
+ "step": 581
+ },
+ {
+ "epoch": 8.083842794759825,
+ "grad_norm": 0.024380534887313843,
+ "learning_rate": 0.0006,
+ "loss": 5.253486633300781,
+ "step": 582
+ },
+ {
+ "epoch": 8.097816593886463,
+ "grad_norm": 0.03246625140309334,
+ "learning_rate": 0.0006,
+ "loss": 5.305774211883545,
+ "step": 583
+ },
+ {
+ "epoch": 8.1117903930131,
+ "grad_norm": 0.030399736016988754,
+ "learning_rate": 0.0006,
+ "loss": 5.370976448059082,
+ "step": 584
+ },
+ {
+ "epoch": 8.125764192139737,
+ "grad_norm": 0.027148913592100143,
+ "learning_rate": 0.0006,
+ "loss": 5.368010520935059,
+ "step": 585
+ },
+ {
+ "epoch": 8.139737991266376,
+ "grad_norm": 0.02925540879368782,
+ "learning_rate": 0.0006,
+ "loss": 5.303013801574707,
+ "step": 586
+ },
+ {
+ "epoch": 8.153711790393013,
+ "grad_norm": 0.02915453538298607,
+ "learning_rate": 0.0006,
+ "loss": 5.329667091369629,
+ "step": 587
+ },
+ {
+ "epoch": 8.167685589519651,
+ "grad_norm": 0.031732626259326935,
+ "learning_rate": 0.0006,
+ "loss": 5.3634934425354,
+ "step": 588
+ },
+ {
+ "epoch": 8.181659388646288,
+ "grad_norm": 0.036166731268167496,
+ "learning_rate": 0.0006,
+ "loss": 5.223980903625488,
+ "step": 589
+ },
+ {
+ "epoch": 8.195633187772925,
+ "grad_norm": 0.035654108971357346,
+ "learning_rate": 0.0006,
+ "loss": 5.361059188842773,
+ "step": 590
+ },
+ {
+ "epoch": 8.209606986899564,
+ "grad_norm": 0.0365324430167675,
+ "learning_rate": 0.0006,
+ "loss": 5.336151599884033,
+ "step": 591
+ },
+ {
+ "epoch": 8.2235807860262,
+ "grad_norm": 0.03578880429267883,
+ "learning_rate": 0.0006,
+ "loss": 5.434076309204102,
+ "step": 592
+ },
+ {
+ "epoch": 8.237554585152838,
+ "grad_norm": 0.035734012722969055,
+ "learning_rate": 0.0006,
+ "loss": 5.330893516540527,
+ "step": 593
+ },
+ {
+ "epoch": 8.251528384279476,
+ "grad_norm": 0.03196857124567032,
+ "learning_rate": 0.0006,
+ "loss": 5.306009769439697,
+ "step": 594
+ },
+ {
+ "epoch": 8.265502183406113,
+ "grad_norm": 0.032042086124420166,
+ "learning_rate": 0.0006,
+ "loss": 5.343267917633057,
+ "step": 595
+ },
+ {
+ "epoch": 8.279475982532752,
+ "grad_norm": 0.03160746395587921,
+ "learning_rate": 0.0006,
+ "loss": 5.2353715896606445,
+ "step": 596
+ },
+ {
+ "epoch": 8.293449781659389,
+ "grad_norm": 0.026689818128943443,
+ "learning_rate": 0.0006,
+ "loss": 5.254042625427246,
+ "step": 597
+ },
+ {
+ "epoch": 8.307423580786025,
+ "grad_norm": 0.02880188450217247,
+ "learning_rate": 0.0006,
+ "loss": 5.326833248138428,
+ "step": 598
+ },
+ {
+ "epoch": 8.321397379912664,
+ "grad_norm": 0.027516381815075874,
+ "learning_rate": 0.0006,
+ "loss": 5.342336177825928,
+ "step": 599
+ },
+ {
+ "epoch": 8.335371179039301,
+ "grad_norm": 0.027875030413269997,
+ "learning_rate": 0.0006,
+ "loss": 5.200719833374023,
+ "step": 600
+ },
+ {
+ "epoch": 8.34934497816594,
+ "grad_norm": 0.0268265288323164,
+ "learning_rate": 0.0006,
+ "loss": 5.249449729919434,
+ "step": 601
+ },
+ {
+ "epoch": 8.363318777292577,
+ "grad_norm": 0.024824608117341995,
+ "learning_rate": 0.0006,
+ "loss": 5.318952560424805,
+ "step": 602
+ },
+ {
+ "epoch": 8.377292576419213,
+ "grad_norm": 0.022990627214312553,
+ "learning_rate": 0.0006,
+ "loss": 5.244993209838867,
+ "step": 603
+ },
+ {
+ "epoch": 8.391266375545852,
+ "grad_norm": 0.022804604843258858,
+ "learning_rate": 0.0006,
+ "loss": 5.3236846923828125,
+ "step": 604
+ },
+ {
+ "epoch": 8.405240174672489,
+ "grad_norm": 0.02419872209429741,
+ "learning_rate": 0.0006,
+ "loss": 5.255486488342285,
+ "step": 605
+ },
+ {
+ "epoch": 8.419213973799126,
+ "grad_norm": 0.021952059119939804,
+ "learning_rate": 0.0006,
+ "loss": 5.3511552810668945,
+ "step": 606
+ },
+ {
+ "epoch": 8.433187772925764,
+ "grad_norm": 0.022375497967004776,
+ "learning_rate": 0.0006,
+ "loss": 5.294790267944336,
+ "step": 607
+ },
+ {
+ "epoch": 8.447161572052401,
+ "grad_norm": 0.019242815673351288,
+ "learning_rate": 0.0006,
+ "loss": 5.252618789672852,
+ "step": 608
+ },
+ {
+ "epoch": 8.46113537117904,
+ "grad_norm": 0.0173486340790987,
+ "learning_rate": 0.0006,
+ "loss": 5.30147647857666,
+ "step": 609
+ },
+ {
+ "epoch": 8.475109170305677,
+ "grad_norm": 0.01612604409456253,
+ "learning_rate": 0.0006,
+ "loss": 5.23960542678833,
+ "step": 610
+ },
+ {
+ "epoch": 8.489082969432314,
+ "grad_norm": 0.014293976128101349,
+ "learning_rate": 0.0006,
+ "loss": 5.119932651519775,
+ "step": 611
+ },
+ {
+ "epoch": 8.503056768558952,
+ "grad_norm": 0.014453536830842495,
+ "learning_rate": 0.0006,
+ "loss": 5.166794776916504,
+ "step": 612
+ },
+ {
+ "epoch": 8.51703056768559,
+ "grad_norm": 0.014685769565403461,
+ "learning_rate": 0.0006,
+ "loss": 5.264822959899902,
+ "step": 613
+ },
+ {
+ "epoch": 8.531004366812226,
+ "grad_norm": 0.01427740603685379,
+ "learning_rate": 0.0006,
+ "loss": 5.278433799743652,
+ "step": 614
+ },
+ {
+ "epoch": 8.544978165938865,
+ "grad_norm": 0.014864951372146606,
+ "learning_rate": 0.0006,
+ "loss": 5.214512348175049,
+ "step": 615
+ },
+ {
+ "epoch": 8.558951965065502,
+ "grad_norm": 0.01565164513885975,
+ "learning_rate": 0.0006,
+ "loss": 5.183066368103027,
+ "step": 616
+ },
+ {
+ "epoch": 8.57292576419214,
+ "grad_norm": 0.01779816672205925,
+ "learning_rate": 0.0006,
+ "loss": 5.163185119628906,
+ "step": 617
+ },
+ {
+ "epoch": 8.586899563318777,
+ "grad_norm": 0.017254827544093132,
+ "learning_rate": 0.0006,
+ "loss": 5.243555068969727,
+ "step": 618
+ },
+ {
+ "epoch": 8.600873362445414,
+ "grad_norm": 0.01650584116578102,
+ "learning_rate": 0.0006,
+ "loss": 5.234884738922119,
+ "step": 619
+ },
+ {
+ "epoch": 8.614847161572053,
+ "grad_norm": 0.017021115869283676,
+ "learning_rate": 0.0006,
+ "loss": 5.12271785736084,
+ "step": 620
+ },
+ {
+ "epoch": 8.62882096069869,
+ "grad_norm": 0.01773759163916111,
+ "learning_rate": 0.0006,
+ "loss": 5.242153167724609,
+ "step": 621
+ },
+ {
+ "epoch": 8.642794759825328,
+ "grad_norm": 0.015679042786359787,
+ "learning_rate": 0.0006,
+ "loss": 5.170779228210449,
+ "step": 622
+ },
+ {
+ "epoch": 8.656768558951965,
+ "grad_norm": 0.013760549947619438,
+ "learning_rate": 0.0006,
+ "loss": 5.238644599914551,
+ "step": 623
+ },
+ {
+ "epoch": 8.670742358078602,
+ "grad_norm": 0.014571522362530231,
+ "learning_rate": 0.0006,
+ "loss": 5.177056312561035,
+ "step": 624
+ },
+ {
+ "epoch": 8.68471615720524,
+ "grad_norm": 0.016209015622735023,
+ "learning_rate": 0.0006,
+ "loss": 5.163750648498535,
+ "step": 625
+ },
+ {
+ "epoch": 8.698689956331878,
+ "grad_norm": 0.016813941299915314,
+ "learning_rate": 0.0006,
+ "loss": 5.183428764343262,
+ "step": 626
+ },
+ {
+ "epoch": 8.712663755458514,
+ "grad_norm": 0.020985357463359833,
+ "learning_rate": 0.0006,
+ "loss": 5.229465007781982,
+ "step": 627
+ },
+ {
+ "epoch": 8.726637554585153,
+ "grad_norm": 0.02679632417857647,
+ "learning_rate": 0.0006,
+ "loss": 5.257368087768555,
+ "step": 628
+ },
+ {
+ "epoch": 8.74061135371179,
+ "grad_norm": 0.025756070390343666,
+ "learning_rate": 0.0006,
+ "loss": 5.253736972808838,
+ "step": 629
+ },
+ {
+ "epoch": 8.754585152838429,
+ "grad_norm": 0.02643490768969059,
+ "learning_rate": 0.0006,
+ "loss": 5.228633403778076,
+ "step": 630
+ },
+ {
+ "epoch": 8.768558951965066,
+ "grad_norm": 0.03303210437297821,
+ "learning_rate": 0.0006,
+ "loss": 5.163540363311768,
+ "step": 631
+ },
+ {
+ "epoch": 8.782532751091702,
+ "grad_norm": 0.030432431027293205,
+ "learning_rate": 0.0006,
+ "loss": 5.243169784545898,
+ "step": 632
+ },
+ {
+ "epoch": 8.796506550218341,
+ "grad_norm": 0.028631288558244705,
+ "learning_rate": 0.0006,
+ "loss": 5.166398048400879,
+ "step": 633
+ },
+ {
+ "epoch": 8.810480349344978,
+ "grad_norm": 0.026188310235738754,
+ "learning_rate": 0.0006,
+ "loss": 5.128122329711914,
+ "step": 634
+ },
+ {
+ "epoch": 8.824454148471617,
+ "grad_norm": 0.025308528915047646,
+ "learning_rate": 0.0006,
+ "loss": 5.211284637451172,
+ "step": 635
+ },
+ {
+ "epoch": 8.838427947598253,
+ "grad_norm": 0.020657729357481003,
+ "learning_rate": 0.0006,
+ "loss": 5.161575794219971,
+ "step": 636
+ },
+ {
+ "epoch": 8.85240174672489,
+ "grad_norm": 0.021901234984397888,
+ "learning_rate": 0.0006,
+ "loss": 5.20050048828125,
+ "step": 637
+ },
+ {
+ "epoch": 8.866375545851529,
+ "grad_norm": 0.021219318732619286,
+ "learning_rate": 0.0006,
+ "loss": 5.211699485778809,
+ "step": 638
+ },
+ {
+ "epoch": 8.880349344978166,
+ "grad_norm": 0.0200974028557539,
+ "learning_rate": 0.0006,
+ "loss": 5.092503547668457,
+ "step": 639
+ },
+ {
+ "epoch": 8.894323144104803,
+ "grad_norm": 0.023804882541298866,
+ "learning_rate": 0.0006,
+ "loss": 5.216068267822266,
+ "step": 640
+ },
+ {
+ "epoch": 8.908296943231441,
+ "grad_norm": 0.026088010519742966,
+ "learning_rate": 0.0006,
+ "loss": 5.155592918395996,
+ "step": 641
+ },
+ {
+ "epoch": 8.922270742358078,
+ "grad_norm": 0.02601276896893978,
+ "learning_rate": 0.0006,
+ "loss": 5.197238922119141,
+ "step": 642
+ },
+ {
+ "epoch": 8.936244541484717,
+ "grad_norm": 0.020387211814522743,
+ "learning_rate": 0.0006,
+ "loss": 5.1955437660217285,
+ "step": 643
+ },
+ {
+ "epoch": 8.950218340611354,
+ "grad_norm": 0.019214622676372528,
+ "learning_rate": 0.0006,
+ "loss": 5.206346035003662,
+ "step": 644
+ },
+ {
+ "epoch": 8.96419213973799,
+ "grad_norm": 0.019674314185976982,
+ "learning_rate": 0.0006,
+ "loss": 5.219857215881348,
+ "step": 645
+ },
+ {
+ "epoch": 8.97816593886463,
+ "grad_norm": 0.020318234339356422,
+ "learning_rate": 0.0006,
+ "loss": 5.191132545471191,
+ "step": 646
+ },
+ {
+ "epoch": 8.992139737991266,
+ "grad_norm": 0.021428676322102547,
+ "learning_rate": 0.0006,
+ "loss": 5.214695930480957,
+ "step": 647
+ },
+ {
+ "epoch": 9.0,
+ "grad_norm": 0.023235054686665535,
+ "learning_rate": 0.0006,
+ "loss": 5.229091644287109,
+ "step": 648
+ },
+ {
+ "epoch": 9.0,
+ "eval_loss": 5.289593696594238,
+ "eval_runtime": 56.9256,
+ "eval_samples_per_second": 42.898,
+ "eval_steps_per_second": 1.353,
+ "step": 648
+ },
+ {
+ "epoch": 9.013973799126637,
+ "grad_norm": 0.02320289984345436,
+ "learning_rate": 0.0006,
+ "loss": 4.983669281005859,
+ "step": 649
+ },
+ {
+ "epoch": 9.027947598253276,
+ "grad_norm": 0.018521282821893692,
+ "learning_rate": 0.0006,
+ "loss": 5.200575828552246,
+ "step": 650
+ },
+ {
+ "epoch": 9.041921397379912,
+ "grad_norm": 0.0180169939994812,
+ "learning_rate": 0.0006,
+ "loss": 5.148033618927002,
+ "step": 651
+ },
+ {
+ "epoch": 9.055895196506551,
+ "grad_norm": 0.019969860091805458,
+ "learning_rate": 0.0006,
+ "loss": 5.099125385284424,
+ "step": 652
+ },
+ {
+ "epoch": 9.069868995633188,
+ "grad_norm": 0.01643305830657482,
+ "learning_rate": 0.0006,
+ "loss": 5.092347145080566,
+ "step": 653
+ },
+ {
+ "epoch": 9.083842794759825,
+ "grad_norm": 0.016022363677620888,
+ "learning_rate": 0.0006,
+ "loss": 5.129279613494873,
+ "step": 654
+ },
+ {
+ "epoch": 9.097816593886463,
+ "grad_norm": 0.01657041162252426,
+ "learning_rate": 0.0006,
+ "loss": 5.159562587738037,
+ "step": 655
+ },
+ {
+ "epoch": 9.1117903930131,
+ "grad_norm": 0.017924228683114052,
+ "learning_rate": 0.0006,
+ "loss": 5.165497779846191,
+ "step": 656
+ },
+ {
+ "epoch": 9.125764192139737,
+ "grad_norm": 0.021592361852526665,
+ "learning_rate": 0.0006,
+ "loss": 5.204775810241699,
+ "step": 657
+ },
+ {
+ "epoch": 9.139737991266376,
+ "grad_norm": 0.026924120262265205,
+ "learning_rate": 0.0006,
+ "loss": 5.128296852111816,
+ "step": 658
+ },
+ {
+ "epoch": 9.153711790393013,
+ "grad_norm": 0.023156899958848953,
+ "learning_rate": 0.0006,
+ "loss": 5.1255340576171875,
+ "step": 659
+ },
+ {
+ "epoch": 9.167685589519651,
+ "grad_norm": 0.020670367404818535,
+ "learning_rate": 0.0006,
+ "loss": 5.128479957580566,
+ "step": 660
+ },
+ {
+ "epoch": 9.181659388646288,
+ "grad_norm": 0.025447173044085503,
+ "learning_rate": 0.0006,
+ "loss": 5.154821872711182,
+ "step": 661
+ },
+ {
+ "epoch": 9.195633187772925,
+ "grad_norm": 0.027940068393945694,
+ "learning_rate": 0.0006,
+ "loss": 5.143466472625732,
+ "step": 662
+ },
+ {
+ "epoch": 9.209606986899564,
+ "grad_norm": 0.026847844943404198,
+ "learning_rate": 0.0006,
+ "loss": 5.045665740966797,
+ "step": 663
+ },
+ {
+ "epoch": 9.2235807860262,
+ "grad_norm": 0.02340601570904255,
+ "learning_rate": 0.0006,
+ "loss": 5.21933126449585,
+ "step": 664
+ },
+ {
+ "epoch": 9.237554585152838,
+ "grad_norm": 0.02340371161699295,
+ "learning_rate": 0.0006,
+ "loss": 4.942949295043945,
+ "step": 665
+ },
+ {
+ "epoch": 9.251528384279476,
+ "grad_norm": 0.02221992425620556,
+ "learning_rate": 0.0006,
+ "loss": 5.197099208831787,
+ "step": 666
+ },
+ {
+ "epoch": 9.265502183406113,
+ "grad_norm": 0.023508677259087563,
+ "learning_rate": 0.0006,
+ "loss": 5.022897720336914,
+ "step": 667
+ },
+ {
+ "epoch": 9.279475982532752,
+ "grad_norm": 0.026120394468307495,
+ "learning_rate": 0.0006,
+ "loss": 5.087725639343262,
+ "step": 668
+ },
+ {
+ "epoch": 9.293449781659389,
+ "grad_norm": 0.027273228392004967,
+ "learning_rate": 0.0006,
+ "loss": 5.090963840484619,
+ "step": 669
+ },
+ {
+ "epoch": 9.307423580786025,
+ "grad_norm": 0.03241586685180664,
+ "learning_rate": 0.0006,
+ "loss": 5.047842979431152,
+ "step": 670
+ },
+ {
+ "epoch": 9.321397379912664,
+ "grad_norm": 0.030245667323470116,
+ "learning_rate": 0.0006,
+ "loss": 5.104803085327148,
+ "step": 671
+ },
+ {
+ "epoch": 9.335371179039301,
+ "grad_norm": 0.027698364108800888,
+ "learning_rate": 0.0006,
+ "loss": 5.0644097328186035,
+ "step": 672
+ },
+ {
+ "epoch": 9.34934497816594,
+ "grad_norm": 0.029692554846405983,
+ "learning_rate": 0.0006,
+ "loss": 5.111942291259766,
+ "step": 673
+ },
+ {
+ "epoch": 9.363318777292577,
+ "grad_norm": 0.03328656405210495,
+ "learning_rate": 0.0006,
+ "loss": 5.1574554443359375,
+ "step": 674
+ },
+ {
+ "epoch": 9.377292576419213,
+ "grad_norm": 0.031938180327415466,
+ "learning_rate": 0.0006,
+ "loss": 5.190371990203857,
+ "step": 675
+ },
+ {
+ "epoch": 9.391266375545852,
+ "grad_norm": 0.024858945980668068,
+ "learning_rate": 0.0006,
+ "loss": 5.121420860290527,
+ "step": 676
+ },
+ {
+ "epoch": 9.405240174672489,
+ "grad_norm": 0.023033203557133675,
+ "learning_rate": 0.0006,
+ "loss": 5.0974345207214355,
+ "step": 677
+ },
+ {
+ "epoch": 9.419213973799126,
+ "grad_norm": 0.022083545103669167,
+ "learning_rate": 0.0006,
+ "loss": 5.046319961547852,
+ "step": 678
+ },
+ {
+ "epoch": 9.433187772925764,
+ "grad_norm": 0.0228437427431345,
+ "learning_rate": 0.0006,
+ "loss": 5.119963645935059,
+ "step": 679
+ },
+ {
+ "epoch": 9.447161572052401,
+ "grad_norm": 0.0242351982742548,
+ "learning_rate": 0.0006,
+ "loss": 5.0444560050964355,
+ "step": 680
+ },
+ {
+ "epoch": 9.46113537117904,
+ "grad_norm": 0.02401185780763626,
+ "learning_rate": 0.0006,
+ "loss": 5.063238620758057,
+ "step": 681
+ },
+ {
+ "epoch": 9.475109170305677,
+ "grad_norm": 0.02560959942638874,
+ "learning_rate": 0.0006,
+ "loss": 5.035224914550781,
+ "step": 682
+ },
+ {
+ "epoch": 9.489082969432314,
+ "grad_norm": 0.024985626339912415,
+ "learning_rate": 0.0006,
+ "loss": 5.049284934997559,
+ "step": 683
+ },
+ {
+ "epoch": 9.503056768558952,
+ "grad_norm": 0.023178115487098694,
+ "learning_rate": 0.0006,
+ "loss": 5.2266621589660645,
+ "step": 684
+ },
+ {
+ "epoch": 9.51703056768559,
+ "grad_norm": 0.0215512253344059,
+ "learning_rate": 0.0006,
+ "loss": 5.1875715255737305,
+ "step": 685
+ },
+ {
+ "epoch": 9.531004366812226,
+ "grad_norm": 0.01662837713956833,
+ "learning_rate": 0.0006,
+ "loss": 5.137012481689453,
+ "step": 686
+ },
+ {
+ "epoch": 9.544978165938865,
+ "grad_norm": 0.01952764391899109,
+ "learning_rate": 0.0006,
+ "loss": 5.072500228881836,
+ "step": 687
+ },
+ {
+ "epoch": 9.558951965065502,
+ "grad_norm": 0.017521077767014503,
+ "learning_rate": 0.0006,
+ "loss": 5.045711994171143,
+ "step": 688
+ },
+ {
+ "epoch": 9.57292576419214,
+ "grad_norm": 0.017355265095829964,
+ "learning_rate": 0.0006,
+ "loss": 5.02881383895874,
+ "step": 689
+ },
+ {
+ "epoch": 9.586899563318777,
+ "grad_norm": 0.015587719157338142,
+ "learning_rate": 0.0006,
+ "loss": 5.132787704467773,
+ "step": 690
+ },
+ {
+ "epoch": 9.600873362445414,
+ "grad_norm": 0.017469845712184906,
+ "learning_rate": 0.0006,
+ "loss": 5.105447292327881,
+ "step": 691
+ },
+ {
+ "epoch": 9.614847161572053,
+ "grad_norm": 0.019630100578069687,
+ "learning_rate": 0.0006,
+ "loss": 5.032464981079102,
+ "step": 692
+ },
+ {
+ "epoch": 9.62882096069869,
+ "grad_norm": 0.01784994639456272,
+ "learning_rate": 0.0006,
+ "loss": 5.081345558166504,
+ "step": 693
+ },
+ {
+ "epoch": 9.642794759825328,
+ "grad_norm": 0.01872754842042923,
+ "learning_rate": 0.0006,
+ "loss": 4.9772114753723145,
+ "step": 694
+ },
+ {
+ "epoch": 9.656768558951965,
+ "grad_norm": 0.01999032311141491,
+ "learning_rate": 0.0006,
+ "loss": 5.089740753173828,
+ "step": 695
+ },
+ {
+ "epoch": 9.670742358078602,
+ "grad_norm": 0.01710767112672329,
+ "learning_rate": 0.0006,
+ "loss": 4.9869890213012695,
+ "step": 696
+ },
+ {
+ "epoch": 9.68471615720524,
+ "grad_norm": 0.015144161880016327,
+ "learning_rate": 0.0006,
+ "loss": 5.022564888000488,
+ "step": 697
+ },
+ {
+ "epoch": 9.698689956331878,
+ "grad_norm": 0.01618543453514576,
+ "learning_rate": 0.0006,
+ "loss": 5.035921573638916,
+ "step": 698
+ },
+ {
+ "epoch": 9.712663755458514,
+ "grad_norm": 0.014653387479484081,
+ "learning_rate": 0.0006,
+ "loss": 4.914456367492676,
+ "step": 699
+ },
+ {
+ "epoch": 9.726637554585153,
+ "grad_norm": 0.013434977270662785,
+ "learning_rate": 0.0006,
+ "loss": 5.063764572143555,
+ "step": 700
+ },
+ {
+ "epoch": 9.74061135371179,
+ "grad_norm": 0.015128864906728268,
+ "learning_rate": 0.0006,
+ "loss": 5.052367687225342,
+ "step": 701
+ },
+ {
+ "epoch": 9.754585152838429,
+ "grad_norm": 0.014210768043994904,
+ "learning_rate": 0.0006,
+ "loss": 5.016489028930664,
+ "step": 702
+ },
+ {
+ "epoch": 9.768558951965066,
+ "grad_norm": 0.015191680751740932,
+ "learning_rate": 0.0006,
+ "loss": 4.990176200866699,
+ "step": 703
+ },
+ {
+ "epoch": 9.782532751091702,
+ "grad_norm": 0.018285127356648445,
+ "learning_rate": 0.0006,
+ "loss": 5.015164375305176,
+ "step": 704
+ },
+ {
+ "epoch": 9.796506550218341,
+ "grad_norm": 0.0204079058021307,
+ "learning_rate": 0.0006,
+ "loss": 5.091142654418945,
+ "step": 705
+ },
+ {
+ "epoch": 9.810480349344978,
+ "grad_norm": 0.01980586163699627,
+ "learning_rate": 0.0006,
+ "loss": 5.03350830078125,
+ "step": 706
+ },
+ {
+ "epoch": 9.824454148471617,
+ "grad_norm": 0.02048729732632637,
+ "learning_rate": 0.0006,
+ "loss": 5.03611421585083,
+ "step": 707
+ },
+ {
+ "epoch": 9.838427947598253,
+ "grad_norm": 0.02058163844048977,
+ "learning_rate": 0.0006,
+ "loss": 4.9613447189331055,
+ "step": 708
+ },
+ {
+ "epoch": 9.85240174672489,
+ "grad_norm": 0.018004924058914185,
+ "learning_rate": 0.0006,
+ "loss": 4.966541767120361,
+ "step": 709
+ },
+ {
+ "epoch": 9.866375545851529,
+ "grad_norm": 0.0198152307420969,
+ "learning_rate": 0.0006,
+ "loss": 5.01750373840332,
+ "step": 710
+ },
+ {
+ "epoch": 9.880349344978166,
+ "grad_norm": 0.018532969057559967,
+ "learning_rate": 0.0006,
+ "loss": 5.097424507141113,
+ "step": 711
+ },
+ {
+ "epoch": 9.894323144104803,
+ "grad_norm": 0.018939971923828125,
+ "learning_rate": 0.0006,
+ "loss": 5.177223205566406,
+ "step": 712
+ },
+ {
+ "epoch": 9.908296943231441,
+ "grad_norm": 0.017969170585274696,
+ "learning_rate": 0.0006,
+ "loss": 5.118013858795166,
+ "step": 713
+ },
+ {
+ "epoch": 9.922270742358078,
+ "grad_norm": 0.018021270632743835,
+ "learning_rate": 0.0006,
+ "loss": 5.034787178039551,
+ "step": 714
+ },
+ {
+ "epoch": 9.936244541484717,
+ "grad_norm": 0.017009710893034935,
+ "learning_rate": 0.0006,
+ "loss": 5.003920555114746,
+ "step": 715
+ },
+ {
+ "epoch": 9.950218340611354,
+ "grad_norm": 0.017722049728035927,
+ "learning_rate": 0.0006,
+ "loss": 5.031939506530762,
+ "step": 716
+ },
+ {
+ "epoch": 9.96419213973799,
+ "grad_norm": 0.017134087160229683,
+ "learning_rate": 0.0006,
+ "loss": 5.093474388122559,
+ "step": 717
+ },
+ {
+ "epoch": 9.97816593886463,
+ "grad_norm": 0.018511613830924034,
+ "learning_rate": 0.0006,
+ "loss": 5.078604698181152,
+ "step": 718
+ },
+ {
+ "epoch": 9.992139737991266,
+ "grad_norm": 0.021206334233283997,
+ "learning_rate": 0.0006,
+ "loss": 5.052572727203369,
+ "step": 719
+ },
+ {
+ "epoch": 10.0,
+ "grad_norm": 0.02239062264561653,
+ "learning_rate": 0.0006,
+ "loss": 5.063547134399414,
+ "step": 720
+ },
+ {
+ "epoch": 10.0,
+ "eval_loss": 5.173044681549072,
+ "eval_runtime": 56.7287,
+ "eval_samples_per_second": 43.047,
+ "eval_steps_per_second": 1.357,
+ "step": 720
+ },
+ {
+ "epoch": 10.013973799126637,
+ "grad_norm": 0.020757969468832016,
+ "learning_rate": 0.0006,
+ "loss": 5.045648574829102,
+ "step": 721
+ },
+ {
+ "epoch": 10.027947598253276,
+ "grad_norm": 0.022697702050209045,
+ "learning_rate": 0.0006,
+ "loss": 4.993833065032959,
+ "step": 722
+ },
+ {
+ "epoch": 10.041921397379912,
+ "grad_norm": 0.02434331737458706,
+ "learning_rate": 0.0006,
+ "loss": 5.073219299316406,
+ "step": 723
+ },
+ {
+ "epoch": 10.055895196506551,
+ "grad_norm": 0.025499660521745682,
+ "learning_rate": 0.0006,
+ "loss": 5.019072532653809,
+ "step": 724
+ },
+ {
+ "epoch": 10.069868995633188,
+ "grad_norm": 0.029281053692102432,
+ "learning_rate": 0.0006,
+ "loss": 4.959851264953613,
+ "step": 725
+ },
+ {
+ "epoch": 10.083842794759825,
+ "grad_norm": 0.033878110349178314,
+ "learning_rate": 0.0006,
+ "loss": 5.132717609405518,
+ "step": 726
+ },
+ {
+ "epoch": 10.097816593886463,
+ "grad_norm": 0.03385764732956886,
+ "learning_rate": 0.0006,
+ "loss": 5.0482892990112305,
+ "step": 727
+ },
+ {
+ "epoch": 10.1117903930131,
+ "grad_norm": 0.03466648980975151,
+ "learning_rate": 0.0006,
+ "loss": 4.946855545043945,
+ "step": 728
+ },
+ {
+ "epoch": 10.125764192139737,
+ "grad_norm": 0.028546305373311043,
+ "learning_rate": 0.0006,
+ "loss": 5.027774333953857,
+ "step": 729
+ },
+ {
+ "epoch": 10.139737991266376,
+ "grad_norm": 0.030575869604945183,
+ "learning_rate": 0.0006,
+ "loss": 4.9768476486206055,
+ "step": 730
+ },
+ {
+ "epoch": 10.153711790393013,
+ "grad_norm": 0.03614303469657898,
+ "learning_rate": 0.0006,
+ "loss": 5.078113555908203,
+ "step": 731
+ },
+ {
+ "epoch": 10.167685589519651,
+ "grad_norm": 0.040991660207509995,
+ "learning_rate": 0.0006,
+ "loss": 5.033185958862305,
+ "step": 732
+ },
+ {
+ "epoch": 10.181659388646288,
+ "grad_norm": 0.04551699757575989,
+ "learning_rate": 0.0006,
+ "loss": 5.0820746421813965,
+ "step": 733
+ },
+ {
+ "epoch": 10.195633187772925,
+ "grad_norm": 0.040193989872932434,
+ "learning_rate": 0.0006,
+ "loss": 5.059700965881348,
+ "step": 734
+ },
+ {
+ "epoch": 10.209606986899564,
+ "grad_norm": 0.035851605236530304,
+ "learning_rate": 0.0006,
+ "loss": 5.142467498779297,
+ "step": 735
+ },
+ {
+ "epoch": 10.2235807860262,
+ "grad_norm": 0.034200169146060944,
+ "learning_rate": 0.0006,
+ "loss": 5.072312355041504,
+ "step": 736
+ },
+ {
+ "epoch": 10.237554585152838,
+ "grad_norm": 0.033621110022068024,
+ "learning_rate": 0.0006,
+ "loss": 5.08037805557251,
+ "step": 737
+ },
+ {
+ "epoch": 10.251528384279476,
+ "grad_norm": 0.037372369319200516,
+ "learning_rate": 0.0006,
+ "loss": 4.998425483703613,
+ "step": 738
+ },
+ {
+ "epoch": 10.265502183406113,
+ "grad_norm": 0.037044707685709,
+ "learning_rate": 0.0006,
+ "loss": 4.923993110656738,
+ "step": 739
+ },
+ {
+ "epoch": 10.279475982532752,
+ "grad_norm": 0.030454808846116066,
+ "learning_rate": 0.0006,
+ "loss": 5.0889997482299805,
+ "step": 740
+ },
+ {
+ "epoch": 10.293449781659389,
+ "grad_norm": 0.030969638377428055,
+ "learning_rate": 0.0006,
+ "loss": 4.984112739562988,
+ "step": 741
+ },
+ {
+ "epoch": 10.307423580786025,
+ "grad_norm": 0.0276983380317688,
+ "learning_rate": 0.0006,
+ "loss": 4.955412864685059,
+ "step": 742
+ },
+ {
+ "epoch": 10.321397379912664,
+ "grad_norm": 0.02716052532196045,
+ "learning_rate": 0.0006,
+ "loss": 5.0098876953125,
+ "step": 743
+ },
+ {
+ "epoch": 10.335371179039301,
+ "grad_norm": 0.02403552085161209,
+ "learning_rate": 0.0006,
+ "loss": 5.165160655975342,
+ "step": 744
+ },
+ {
+ "epoch": 10.34934497816594,
+ "grad_norm": 0.0257862601429224,
+ "learning_rate": 0.0006,
+ "loss": 5.001547813415527,
+ "step": 745
+ },
+ {
+ "epoch": 10.363318777292577,
+ "grad_norm": 0.02300378680229187,
+ "learning_rate": 0.0006,
+ "loss": 5.038888931274414,
+ "step": 746
+ },
+ {
+ "epoch": 10.377292576419213,
+ "grad_norm": 0.019766854122281075,
+ "learning_rate": 0.0006,
+ "loss": 4.924291610717773,
+ "step": 747
+ },
+ {
+ "epoch": 10.391266375545852,
+ "grad_norm": 0.01888395845890045,
+ "learning_rate": 0.0006,
+ "loss": 5.059971332550049,
+ "step": 748
+ },
+ {
+ "epoch": 10.405240174672489,
+ "grad_norm": 0.015577499754726887,
+ "learning_rate": 0.0006,
+ "loss": 4.987686634063721,
+ "step": 749
+ },
+ {
+ "epoch": 10.419213973799126,
+ "grad_norm": 0.016012346372008324,
+ "learning_rate": 0.0006,
+ "loss": 5.015157699584961,
+ "step": 750
+ },
+ {
+ "epoch": 10.433187772925764,
+ "grad_norm": 0.016277998685836792,
+ "learning_rate": 0.0006,
+ "loss": 5.004931449890137,
+ "step": 751
+ },
+ {
+ "epoch": 10.447161572052401,
+ "grad_norm": 0.017504561692476273,
+ "learning_rate": 0.0006,
+ "loss": 4.902827739715576,
+ "step": 752
+ },
+ {
+ "epoch": 10.46113537117904,
+ "grad_norm": 0.01569022797048092,
+ "learning_rate": 0.0006,
+ "loss": 4.954863548278809,
+ "step": 753
+ },
+ {
+ "epoch": 10.475109170305677,
+ "grad_norm": 0.014114422723650932,
+ "learning_rate": 0.0006,
+ "loss": 5.00656795501709,
+ "step": 754
+ },
+ {
+ "epoch": 10.489082969432314,
+ "grad_norm": 0.01573832333087921,
+ "learning_rate": 0.0006,
+ "loss": 4.978353023529053,
+ "step": 755
+ },
+ {
+ "epoch": 10.503056768558952,
+ "grad_norm": 0.016480036079883575,
+ "learning_rate": 0.0006,
+ "loss": 4.953327655792236,
+ "step": 756
+ },
+ {
+ "epoch": 10.51703056768559,
+ "grad_norm": 0.014008025638759136,
+ "learning_rate": 0.0006,
+ "loss": 5.050840377807617,
+ "step": 757
+ },
+ {
+ "epoch": 10.531004366812226,
+ "grad_norm": 0.013147998601198196,
+ "learning_rate": 0.0006,
+ "loss": 4.974964141845703,
+ "step": 758
+ },
+ {
+ "epoch": 10.544978165938865,
+ "grad_norm": 0.013807603158056736,
+ "learning_rate": 0.0006,
+ "loss": 4.927907943725586,
+ "step": 759
+ },
+ {
+ "epoch": 10.558951965065502,
+ "grad_norm": 0.013555224984884262,
+ "learning_rate": 0.0006,
+ "loss": 4.980690956115723,
+ "step": 760
+ },
+ {
+ "epoch": 10.57292576419214,
+ "grad_norm": 0.013045408762991428,
+ "learning_rate": 0.0006,
+ "loss": 4.9436116218566895,
+ "step": 761
+ },
+ {
+ "epoch": 10.586899563318777,
+ "grad_norm": 0.012858862057328224,
+ "learning_rate": 0.0006,
+ "loss": 4.9595770835876465,
+ "step": 762
+ },
+ {
+ "epoch": 10.600873362445414,
+ "grad_norm": 0.01468253880739212,
+ "learning_rate": 0.0006,
+ "loss": 4.90725040435791,
+ "step": 763
+ },
+ {
+ "epoch": 10.614847161572053,
+ "grad_norm": 0.013224679045379162,
+ "learning_rate": 0.0006,
+ "loss": 4.877615928649902,
+ "step": 764
+ },
+ {
+ "epoch": 10.62882096069869,
+ "grad_norm": 0.013555348850786686,
+ "learning_rate": 0.0006,
+ "loss": 4.9456610679626465,
+ "step": 765
+ },
+ {
+ "epoch": 10.642794759825328,
+ "grad_norm": 0.014387160539627075,
+ "learning_rate": 0.0006,
+ "loss": 4.9233717918396,
+ "step": 766
+ },
+ {
+ "epoch": 10.656768558951965,
+ "grad_norm": 0.014354088343679905,
+ "learning_rate": 0.0006,
+ "loss": 4.983511924743652,
+ "step": 767
+ },
+ {
+ "epoch": 10.670742358078602,
+ "grad_norm": 0.01332700252532959,
+ "learning_rate": 0.0006,
+ "loss": 4.934416770935059,
+ "step": 768
+ },
+ {
+ "epoch": 10.68471615720524,
+ "grad_norm": 0.011543313041329384,
+ "learning_rate": 0.0006,
+ "loss": 4.96388053894043,
+ "step": 769
+ },
+ {
+ "epoch": 10.698689956331878,
+ "grad_norm": 0.012330878525972366,
+ "learning_rate": 0.0006,
+ "loss": 4.946396350860596,
+ "step": 770
+ },
+ {
+ "epoch": 10.712663755458514,
+ "grad_norm": 0.01379779726266861,
+ "learning_rate": 0.0006,
+ "loss": 4.981019020080566,
+ "step": 771
+ },
+ {
+ "epoch": 10.726637554585153,
+ "grad_norm": 0.014612431637942791,
+ "learning_rate": 0.0006,
+ "loss": 4.894567966461182,
+ "step": 772
+ },
+ {
+ "epoch": 10.74061135371179,
+ "grad_norm": 0.013062899932265282,
+ "learning_rate": 0.0006,
+ "loss": 5.033038139343262,
+ "step": 773
+ },
+ {
+ "epoch": 10.754585152838429,
+ "grad_norm": 0.013898049481213093,
+ "learning_rate": 0.0006,
+ "loss": 4.823636054992676,
+ "step": 774
+ },
+ {
+ "epoch": 10.768558951965066,
+ "grad_norm": 0.01502019353210926,
+ "learning_rate": 0.0006,
+ "loss": 5.020156383514404,
+ "step": 775
+ },
+ {
+ "epoch": 10.782532751091702,
+ "grad_norm": 0.013810204342007637,
+ "learning_rate": 0.0006,
+ "loss": 4.935647010803223,
+ "step": 776
+ },
+ {
+ "epoch": 10.796506550218341,
+ "grad_norm": 0.013616513460874557,
+ "learning_rate": 0.0006,
+ "loss": 4.901486396789551,
+ "step": 777
+ },
+ {
+ "epoch": 10.810480349344978,
+ "grad_norm": 0.013601024635136127,
+ "learning_rate": 0.0006,
+ "loss": 4.931873321533203,
+ "step": 778
+ },
+ {
+ "epoch": 10.824454148471617,
+ "grad_norm": 0.013668350875377655,
+ "learning_rate": 0.0006,
+ "loss": 4.937256813049316,
+ "step": 779
+ },
+ {
+ "epoch": 10.838427947598253,
+ "grad_norm": 0.01581823080778122,
+ "learning_rate": 0.0006,
+ "loss": 4.9611406326293945,
+ "step": 780
+ },
+ {
+ "epoch": 10.85240174672489,
+ "grad_norm": 0.019478484988212585,
+ "learning_rate": 0.0006,
+ "loss": 4.9212236404418945,
+ "step": 781
+ },
+ {
+ "epoch": 10.866375545851529,
+ "grad_norm": 0.026921333745121956,
+ "learning_rate": 0.0006,
+ "loss": 4.88501501083374,
+ "step": 782
+ },
+ {
+ "epoch": 10.880349344978166,
+ "grad_norm": 0.03155914694070816,
+ "learning_rate": 0.0006,
+ "loss": 4.989967346191406,
+ "step": 783
+ },
+ {
+ "epoch": 10.894323144104803,
+ "grad_norm": 0.029689429327845573,
+ "learning_rate": 0.0006,
+ "loss": 4.936000823974609,
+ "step": 784
+ },
+ {
+ "epoch": 10.908296943231441,
+ "grad_norm": 0.026626063510775566,
+ "learning_rate": 0.0006,
+ "loss": 4.984785079956055,
+ "step": 785
+ },
+ {
+ "epoch": 10.922270742358078,
+ "grad_norm": 0.026168212294578552,
+ "learning_rate": 0.0006,
+ "loss": 4.908282279968262,
+ "step": 786
+ },
+ {
+ "epoch": 10.936244541484717,
+ "grad_norm": 0.02828158251941204,
+ "learning_rate": 0.0006,
+ "loss": 4.926390647888184,
+ "step": 787
+ },
+ {
+ "epoch": 10.950218340611354,
+ "grad_norm": 0.02649078331887722,
+ "learning_rate": 0.0006,
+ "loss": 5.00400972366333,
+ "step": 788
+ },
+ {
+ "epoch": 10.96419213973799,
+ "grad_norm": 0.0256856270134449,
+ "learning_rate": 0.0006,
+ "loss": 5.0317487716674805,
+ "step": 789
+ },
+ {
+ "epoch": 10.97816593886463,
+ "grad_norm": 0.024728883057832718,
+ "learning_rate": 0.0006,
+ "loss": 4.9299635887146,
+ "step": 790
+ },
+ {
+ "epoch": 10.992139737991266,
+ "grad_norm": 0.026241116225719452,
+ "learning_rate": 0.0006,
+ "loss": 4.947319984436035,
+ "step": 791
+ },
+ {
+ "epoch": 11.0,
+ "grad_norm": 0.023526743054389954,
+ "learning_rate": 0.0006,
+ "loss": 4.94429349899292,
+ "step": 792
+ },
+ {
+ "epoch": 11.0,
+ "eval_loss": 5.104345798492432,
+ "eval_runtime": 57.0955,
+ "eval_samples_per_second": 42.77,
+ "eval_steps_per_second": 1.349,
+ "step": 792
+ },
+ {
+ "epoch": 11.013973799126637,
+ "grad_norm": 0.0216788612306118,
+ "learning_rate": 0.0006,
+ "loss": 4.881950855255127,
+ "step": 793
+ },
+ {
+ "epoch": 11.027947598253276,
+ "grad_norm": 0.01830463856458664,
+ "learning_rate": 0.0006,
+ "loss": 4.971016883850098,
+ "step": 794
+ },
+ {
+ "epoch": 11.041921397379912,
+ "grad_norm": 0.02204521745443344,
+ "learning_rate": 0.0006,
+ "loss": 4.894623756408691,
+ "step": 795
+ },
+ {
+ "epoch": 11.055895196506551,
+ "grad_norm": 0.0210255216807127,
+ "learning_rate": 0.0006,
+ "loss": 4.834009170532227,
+ "step": 796
+ },
+ {
+ "epoch": 11.069868995633188,
+ "grad_norm": 0.019797448068857193,
+ "learning_rate": 0.0006,
+ "loss": 5.086714744567871,
+ "step": 797
+ },
+ {
+ "epoch": 11.083842794759825,
+ "grad_norm": 0.02036641724407673,
+ "learning_rate": 0.0006,
+ "loss": 4.920056343078613,
+ "step": 798
+ },
+ {
+ "epoch": 11.097816593886463,
+ "grad_norm": 0.01966066285967827,
+ "learning_rate": 0.0006,
+ "loss": 4.879859447479248,
+ "step": 799
+ },
+ {
+ "epoch": 11.1117903930131,
+ "grad_norm": 0.0202149897813797,
+ "learning_rate": 0.0006,
+ "loss": 4.933640480041504,
+ "step": 800
+ },
+ {
+ "epoch": 11.125764192139737,
+ "grad_norm": 0.01893039606511593,
+ "learning_rate": 0.0006,
+ "loss": 4.901970863342285,
+ "step": 801
+ },
+ {
+ "epoch": 11.139737991266376,
+ "grad_norm": 0.016040155664086342,
+ "learning_rate": 0.0006,
+ "loss": 4.905491828918457,
+ "step": 802
+ },
+ {
+ "epoch": 11.153711790393013,
+ "grad_norm": 0.01720350608229637,
+ "learning_rate": 0.0006,
+ "loss": 4.798816680908203,
+ "step": 803
+ },
+ {
+ "epoch": 11.167685589519651,
+ "grad_norm": 0.017002852633595467,
+ "learning_rate": 0.0006,
+ "loss": 4.851859092712402,
+ "step": 804
+ },
+ {
+ "epoch": 11.181659388646288,
+ "grad_norm": 0.015594680793583393,
+ "learning_rate": 0.0006,
+ "loss": 4.90328311920166,
+ "step": 805
+ },
+ {
+ "epoch": 11.195633187772925,
+ "grad_norm": 0.014904462732374668,
+ "learning_rate": 0.0006,
+ "loss": 4.904215335845947,
+ "step": 806
+ },
+ {
+ "epoch": 11.209606986899564,
+ "grad_norm": 0.01366228237748146,
+ "learning_rate": 0.0006,
+ "loss": 4.9813055992126465,
+ "step": 807
+ },
+ {
+ "epoch": 11.2235807860262,
+ "grad_norm": 0.014719638973474503,
+ "learning_rate": 0.0006,
+ "loss": 4.927306175231934,
+ "step": 808
+ },
+ {
+ "epoch": 11.237554585152838,
+ "grad_norm": 0.015016036108136177,
+ "learning_rate": 0.0006,
+ "loss": 4.874215126037598,
+ "step": 809
+ },
+ {
+ "epoch": 11.251528384279476,
+ "grad_norm": 0.014925424009561539,
+ "learning_rate": 0.0006,
+ "loss": 4.838257789611816,
+ "step": 810
+ },
+ {
+ "epoch": 11.265502183406113,
+ "grad_norm": 0.014335823245346546,
+ "learning_rate": 0.0006,
+ "loss": 4.940608024597168,
+ "step": 811
+ },
+ {
+ "epoch": 11.279475982532752,
+ "grad_norm": 0.014987512491643429,
+ "learning_rate": 0.0006,
+ "loss": 4.831531524658203,
+ "step": 812
+ },
+ {
+ "epoch": 11.293449781659389,
+ "grad_norm": 0.016087457537651062,
+ "learning_rate": 0.0006,
+ "loss": 4.863227367401123,
+ "step": 813
+ },
+ {
+ "epoch": 11.307423580786025,
+ "grad_norm": 0.01671300269663334,
+ "learning_rate": 0.0006,
+ "loss": 4.8535661697387695,
+ "step": 814
+ },
+ {
+ "epoch": 11.321397379912664,
+ "grad_norm": 0.015229142270982265,
+ "learning_rate": 0.0006,
+ "loss": 4.948635101318359,
+ "step": 815
+ },
+ {
+ "epoch": 11.335371179039301,
+ "grad_norm": 0.01532587967813015,
+ "learning_rate": 0.0006,
+ "loss": 4.9035797119140625,
+ "step": 816
+ },
+ {
+ "epoch": 11.34934497816594,
+ "grad_norm": 0.015715476125478745,
+ "learning_rate": 0.0006,
+ "loss": 4.894709587097168,
+ "step": 817
+ },
+ {
+ "epoch": 11.363318777292577,
+ "grad_norm": 0.014156977646052837,
+ "learning_rate": 0.0006,
+ "loss": 4.960206985473633,
+ "step": 818
+ },
+ {
+ "epoch": 11.377292576419213,
+ "grad_norm": 0.017076566815376282,
+ "learning_rate": 0.0006,
+ "loss": 5.005850791931152,
+ "step": 819
+ },
+ {
+ "epoch": 11.391266375545852,
+ "grad_norm": 0.0172084029763937,
+ "learning_rate": 0.0006,
+ "loss": 4.903355598449707,
+ "step": 820
+ },
+ {
+ "epoch": 11.405240174672489,
+ "grad_norm": 0.0180258359760046,
+ "learning_rate": 0.0006,
+ "loss": 4.842667579650879,
+ "step": 821
+ },
+ {
+ "epoch": 11.419213973799126,
+ "grad_norm": 0.022130120545625687,
+ "learning_rate": 0.0006,
+ "loss": 4.799970626831055,
+ "step": 822
+ },
+ {
+ "epoch": 11.433187772925764,
+ "grad_norm": 0.02384897693991661,
+ "learning_rate": 0.0006,
+ "loss": 4.860746383666992,
+ "step": 823
+ },
+ {
+ "epoch": 11.447161572052401,
+ "grad_norm": 0.021881572902202606,
+ "learning_rate": 0.0006,
+ "loss": 4.918107032775879,
+ "step": 824
+ },
+ {
+ "epoch": 11.46113537117904,
+ "grad_norm": 0.023085080087184906,
+ "learning_rate": 0.0006,
+ "loss": 4.878946304321289,
+ "step": 825
+ },
+ {
+ "epoch": 11.475109170305677,
+ "grad_norm": 0.02291012555360794,
+ "learning_rate": 0.0006,
+ "loss": 4.958887100219727,
+ "step": 826
+ },
+ {
+ "epoch": 11.489082969432314,
+ "grad_norm": 0.0199968870729208,
+ "learning_rate": 0.0006,
+ "loss": 4.80955696105957,
+ "step": 827
+ },
+ {
+ "epoch": 11.503056768558952,
+ "grad_norm": 0.018702229484915733,
+ "learning_rate": 0.0006,
+ "loss": 5.0276408195495605,
+ "step": 828
+ },
+ {
+ "epoch": 11.51703056768559,
+ "grad_norm": 0.017787311226129532,
+ "learning_rate": 0.0006,
+ "loss": 4.96018123626709,
+ "step": 829
+ },
+ {
+ "epoch": 11.531004366812226,
+ "grad_norm": 0.01704501546919346,
+ "learning_rate": 0.0006,
+ "loss": 4.914553165435791,
+ "step": 830
+ },
+ {
+ "epoch": 11.544978165938865,
+ "grad_norm": 0.017519677057862282,
+ "learning_rate": 0.0006,
+ "loss": 4.798098087310791,
+ "step": 831
+ },
+ {
+ "epoch": 11.558951965065502,
+ "grad_norm": 0.018511991947889328,
+ "learning_rate": 0.0006,
+ "loss": 4.817785263061523,
+ "step": 832
+ },
+ {
+ "epoch": 11.57292576419214,
+ "grad_norm": 0.01876644790172577,
+ "learning_rate": 0.0006,
+ "loss": 4.818995475769043,
+ "step": 833
+ },
+ {
+ "epoch": 11.586899563318777,
+ "grad_norm": 0.01753568835556507,
+ "learning_rate": 0.0006,
+ "loss": 4.811938285827637,
+ "step": 834
+ },
+ {
+ "epoch": 11.600873362445414,
+ "grad_norm": 0.017046531662344933,
+ "learning_rate": 0.0006,
+ "loss": 4.914674282073975,
+ "step": 835
+ },
+ {
+ "epoch": 11.614847161572053,
+ "grad_norm": 0.01802094094455242,
+ "learning_rate": 0.0006,
+ "loss": 4.993284225463867,
+ "step": 836
+ },
+ {
+ "epoch": 11.62882096069869,
+ "grad_norm": 0.019493764266371727,
+ "learning_rate": 0.0006,
+ "loss": 4.895628452301025,
+ "step": 837
+ },
+ {
+ "epoch": 11.642794759825328,
+ "grad_norm": 0.019461210817098618,
+ "learning_rate": 0.0006,
+ "loss": 4.869269371032715,
+ "step": 838
+ },
+ {
+ "epoch": 11.656768558951965,
+ "grad_norm": 0.01981567218899727,
+ "learning_rate": 0.0006,
+ "loss": 4.863052845001221,
+ "step": 839
+ },
+ {
+ "epoch": 11.670742358078602,
+ "grad_norm": 0.019894491881132126,
+ "learning_rate": 0.0006,
+ "loss": 4.8775787353515625,
+ "step": 840
+ },
+ {
+ "epoch": 11.68471615720524,
+ "grad_norm": 0.01875786855816841,
+ "learning_rate": 0.0006,
+ "loss": 4.817763328552246,
+ "step": 841
+ },
+ {
+ "epoch": 11.698689956331878,
+ "grad_norm": 0.020926695317029953,
+ "learning_rate": 0.0006,
+ "loss": 4.7470526695251465,
+ "step": 842
+ },
+ {
+ "epoch": 11.712663755458514,
+ "grad_norm": 0.021477103233337402,
+ "learning_rate": 0.0006,
+ "loss": 4.762682914733887,
+ "step": 843
+ },
+ {
+ "epoch": 11.726637554585153,
+ "grad_norm": 0.020628711208701134,
+ "learning_rate": 0.0006,
+ "loss": 4.807985782623291,
+ "step": 844
+ },
+ {
+ "epoch": 11.74061135371179,
+ "grad_norm": 0.021489612758159637,
+ "learning_rate": 0.0006,
+ "loss": 4.843091011047363,
+ "step": 845
+ },
+ {
+ "epoch": 11.754585152838429,
+ "grad_norm": 0.021194491535425186,
+ "learning_rate": 0.0006,
+ "loss": 4.8300251960754395,
+ "step": 846
+ },
+ {
+ "epoch": 11.768558951965066,
+ "grad_norm": 0.01833350397646427,
+ "learning_rate": 0.0006,
+ "loss": 5.000202178955078,
+ "step": 847
+ },
+ {
+ "epoch": 11.782532751091702,
+ "grad_norm": 0.018096931278705597,
+ "learning_rate": 0.0006,
+ "loss": 4.880648136138916,
+ "step": 848
+ },
+ {
+ "epoch": 11.796506550218341,
+ "grad_norm": 0.0197161752730608,
+ "learning_rate": 0.0006,
+ "loss": 4.788316249847412,
+ "step": 849
+ },
+ {
+ "epoch": 11.810480349344978,
+ "grad_norm": 0.0203181654214859,
+ "learning_rate": 0.0006,
+ "loss": 4.870944976806641,
+ "step": 850
+ },
+ {
+ "epoch": 11.824454148471617,
+ "grad_norm": 0.01908455230295658,
+ "learning_rate": 0.0006,
+ "loss": 4.890144348144531,
+ "step": 851
+ },
+ {
+ "epoch": 11.838427947598253,
+ "grad_norm": 0.021281801164150238,
+ "learning_rate": 0.0006,
+ "loss": 4.89578914642334,
+ "step": 852
+ },
+ {
+ "epoch": 11.85240174672489,
+ "grad_norm": 0.02124273031949997,
+ "learning_rate": 0.0006,
+ "loss": 4.883749008178711,
+ "step": 853
+ },
+ {
+ "epoch": 11.866375545851529,
+ "grad_norm": 0.020969398319721222,
+ "learning_rate": 0.0006,
+ "loss": 4.809142589569092,
+ "step": 854
+ },
+ {
+ "epoch": 11.880349344978166,
+ "grad_norm": 0.01967657171189785,
+ "learning_rate": 0.0006,
+ "loss": 4.821864604949951,
+ "step": 855
+ },
+ {
+ "epoch": 11.894323144104803,
+ "grad_norm": 0.020887333899736404,
+ "learning_rate": 0.0006,
+ "loss": 4.917985916137695,
+ "step": 856
+ },
+ {
+ "epoch": 11.908296943231441,
+ "grad_norm": 0.019203083589673042,
+ "learning_rate": 0.0006,
+ "loss": 4.824960708618164,
+ "step": 857
+ },
+ {
+ "epoch": 11.922270742358078,
+ "grad_norm": 0.020162401720881462,
+ "learning_rate": 0.0006,
+ "loss": 4.772617816925049,
+ "step": 858
+ },
+ {
+ "epoch": 11.936244541484717,
+ "grad_norm": 0.02251487225294113,
+ "learning_rate": 0.0006,
+ "loss": 4.735418319702148,
+ "step": 859
+ },
+ {
+ "epoch": 11.950218340611354,
+ "grad_norm": 0.020313527435064316,
+ "learning_rate": 0.0006,
+ "loss": 4.939233779907227,
+ "step": 860
+ },
+ {
+ "epoch": 11.96419213973799,
+ "grad_norm": 0.021635599434375763,
+ "learning_rate": 0.0006,
+ "loss": 4.861807346343994,
+ "step": 861
+ },
+ {
+ "epoch": 11.97816593886463,
+ "grad_norm": 0.02285311557352543,
+ "learning_rate": 0.0006,
+ "loss": 4.932732582092285,
+ "step": 862
+ },
+ {
+ "epoch": 11.992139737991266,
+ "grad_norm": 0.02072535641491413,
+ "learning_rate": 0.0006,
+ "loss": 4.796767711639404,
+ "step": 863
+ },
+ {
+ "epoch": 12.0,
+ "grad_norm": 0.020588304847478867,
+ "learning_rate": 0.0006,
+ "loss": 4.997537612915039,
+ "step": 864
+ },
+ {
+ "epoch": 12.0,
+ "eval_loss": 5.044651985168457,
+ "eval_runtime": 56.9145,
+ "eval_samples_per_second": 42.906,
+ "eval_steps_per_second": 1.353,
+ "step": 864
+ },
+ {
+ "epoch": 12.013973799126637,
+ "grad_norm": 0.017597166821360588,
+ "learning_rate": 0.0006,
+ "loss": 4.830524444580078,
+ "step": 865
+ },
+ {
+ "epoch": 12.027947598253276,
+ "grad_norm": 0.01716788113117218,
+ "learning_rate": 0.0006,
+ "loss": 4.894325256347656,
+ "step": 866
+ },
+ {
+ "epoch": 12.041921397379912,
+ "grad_norm": 0.017712457105517387,
+ "learning_rate": 0.0006,
+ "loss": 4.868297576904297,
+ "step": 867
+ },
+ {
+ "epoch": 12.055895196506551,
+ "grad_norm": 0.021486390382051468,
+ "learning_rate": 0.0006,
+ "loss": 4.922049522399902,
+ "step": 868
+ },
+ {
+ "epoch": 12.069868995633188,
+ "grad_norm": 0.024560654535889626,
+ "learning_rate": 0.0006,
+ "loss": 4.734241008758545,
+ "step": 869
+ },
+ {
+ "epoch": 12.083842794759825,
+ "grad_norm": 0.025870252400636673,
+ "learning_rate": 0.0006,
+ "loss": 4.868566989898682,
+ "step": 870
+ },
+ {
+ "epoch": 12.097816593886463,
+ "grad_norm": 0.0277389045804739,
+ "learning_rate": 0.0006,
+ "loss": 4.755006790161133,
+ "step": 871
+ },
+ {
+ "epoch": 12.1117903930131,
+ "grad_norm": 0.029100844636559486,
+ "learning_rate": 0.0006,
+ "loss": 4.909667015075684,
+ "step": 872
+ },
+ {
+ "epoch": 12.125764192139737,
+ "grad_norm": 0.031833067536354065,
+ "learning_rate": 0.0006,
+ "loss": 4.858674049377441,
+ "step": 873
+ },
+ {
+ "epoch": 12.139737991266376,
+ "grad_norm": 0.032709237188100815,
+ "learning_rate": 0.0006,
+ "loss": 4.823696136474609,
+ "step": 874
+ },
+ {
+ "epoch": 12.153711790393013,
+ "grad_norm": 0.02995380386710167,
+ "learning_rate": 0.0006,
+ "loss": 4.850235462188721,
+ "step": 875
+ },
+ {
+ "epoch": 12.167685589519651,
+ "grad_norm": 0.034098610281944275,
+ "learning_rate": 0.0006,
+ "loss": 4.8169989585876465,
+ "step": 876
+ },
+ {
+ "epoch": 12.181659388646288,
+ "grad_norm": 0.03377080336213112,
+ "learning_rate": 0.0006,
+ "loss": 4.90738582611084,
+ "step": 877
+ },
+ {
+ "epoch": 12.195633187772925,
+ "grad_norm": 0.029780825600028038,
+ "learning_rate": 0.0006,
+ "loss": 4.811163902282715,
+ "step": 878
+ },
+ {
+ "epoch": 12.209606986899564,
+ "grad_norm": 0.03118024580180645,
+ "learning_rate": 0.0006,
+ "loss": 4.861286640167236,
+ "step": 879
+ },
+ {
+ "epoch": 12.2235807860262,
+ "grad_norm": 0.033052004873752594,
+ "learning_rate": 0.0006,
+ "loss": 4.865021228790283,
+ "step": 880
+ },
+ {
+ "epoch": 12.237554585152838,
+ "grad_norm": 0.029070032760500908,
+ "learning_rate": 0.0006,
+ "loss": 4.947303771972656,
+ "step": 881
+ },
+ {
+ "epoch": 12.251528384279476,
+ "grad_norm": 0.028355715796351433,
+ "learning_rate": 0.0006,
+ "loss": 4.804648399353027,
+ "step": 882
+ },
+ {
+ "epoch": 12.265502183406113,
+ "grad_norm": 0.031257227063179016,
+ "learning_rate": 0.0006,
+ "loss": 4.757718086242676,
+ "step": 883
+ },
+ {
+ "epoch": 12.279475982532752,
+ "grad_norm": 0.031438373029232025,
+ "learning_rate": 0.0006,
+ "loss": 4.851268768310547,
+ "step": 884
+ },
+ {
+ "epoch": 12.293449781659389,
+ "grad_norm": 0.030492138117551804,
+ "learning_rate": 0.0006,
+ "loss": 4.7891035079956055,
+ "step": 885
+ },
+ {
+ "epoch": 12.307423580786025,
+ "grad_norm": 0.026115991175174713,
+ "learning_rate": 0.0006,
+ "loss": 4.86474609375,
+ "step": 886
+ },
+ {
+ "epoch": 12.321397379912664,
+ "grad_norm": 0.024184465408325195,
+ "learning_rate": 0.0006,
+ "loss": 4.860456943511963,
+ "step": 887
+ },
+ {
+ "epoch": 12.335371179039301,
+ "grad_norm": 0.025437982752919197,
+ "learning_rate": 0.0006,
+ "loss": 4.806251049041748,
+ "step": 888
+ },
+ {
+ "epoch": 12.34934497816594,
+ "grad_norm": 0.022390231490135193,
+ "learning_rate": 0.0006,
+ "loss": 4.807086944580078,
+ "step": 889
+ },
+ {
+ "epoch": 12.363318777292577,
+ "grad_norm": 0.020174294710159302,
+ "learning_rate": 0.0006,
+ "loss": 4.78292179107666,
+ "step": 890
+ },
+ {
+ "epoch": 12.377292576419213,
+ "grad_norm": 0.019734324887394905,
+ "learning_rate": 0.0006,
+ "loss": 4.784509658813477,
+ "step": 891
+ },
+ {
+ "epoch": 12.391266375545852,
+ "grad_norm": 0.016733523458242416,
+ "learning_rate": 0.0006,
+ "loss": 4.811755180358887,
+ "step": 892
+ },
+ {
+ "epoch": 12.405240174672489,
+ "grad_norm": 0.01570146717131138,
+ "learning_rate": 0.0006,
+ "loss": 4.720088005065918,
+ "step": 893
+ },
+ {
+ "epoch": 12.419213973799126,
+ "grad_norm": 0.01601393334567547,
+ "learning_rate": 0.0006,
+ "loss": 4.783543109893799,
+ "step": 894
+ },
+ {
+ "epoch": 12.433187772925764,
+ "grad_norm": 0.014035487547516823,
+ "learning_rate": 0.0006,
+ "loss": 4.848340034484863,
+ "step": 895
+ },
+ {
+ "epoch": 12.447161572052401,
+ "grad_norm": 0.015248487703502178,
+ "learning_rate": 0.0006,
+ "loss": 4.773287773132324,
+ "step": 896
+ },
+ {
+ "epoch": 12.46113537117904,
+ "grad_norm": 0.01532268151640892,
+ "learning_rate": 0.0006,
+ "loss": 4.799646854400635,
+ "step": 897
+ },
+ {
+ "epoch": 12.475109170305677,
+ "grad_norm": 0.013982534408569336,
+ "learning_rate": 0.0006,
+ "loss": 4.8090500831604,
+ "step": 898
+ },
+ {
+ "epoch": 12.489082969432314,
+ "grad_norm": 0.014081164263188839,
+ "learning_rate": 0.0006,
+ "loss": 4.751888751983643,
+ "step": 899
+ },
+ {
+ "epoch": 12.503056768558952,
+ "grad_norm": 0.01380133256316185,
+ "learning_rate": 0.0006,
+ "loss": 4.78144645690918,
+ "step": 900
+ },
+ {
+ "epoch": 12.51703056768559,
+ "grad_norm": 0.013555269688367844,
+ "learning_rate": 0.0006,
+ "loss": 4.779257774353027,
+ "step": 901
+ },
+ {
+ "epoch": 12.531004366812226,
+ "grad_norm": 0.013075701892375946,
+ "learning_rate": 0.0006,
+ "loss": 4.820705413818359,
+ "step": 902
+ },
+ {
+ "epoch": 12.544978165938865,
+ "grad_norm": 0.01276316400617361,
+ "learning_rate": 0.0006,
+ "loss": 4.836733818054199,
+ "step": 903
+ },
+ {
+ "epoch": 12.558951965065502,
+ "grad_norm": 0.013033892959356308,
+ "learning_rate": 0.0006,
+ "loss": 4.764178276062012,
+ "step": 904
+ },
+ {
+ "epoch": 12.57292576419214,
+ "grad_norm": 0.013996980153024197,
+ "learning_rate": 0.0006,
+ "loss": 4.767983436584473,
+ "step": 905
+ },
+ {
+ "epoch": 12.586899563318777,
+ "grad_norm": 0.015222184360027313,
+ "learning_rate": 0.0006,
+ "loss": 4.759721755981445,
+ "step": 906
+ },
+ {
+ "epoch": 12.600873362445414,
+ "grad_norm": 0.01623941957950592,
+ "learning_rate": 0.0006,
+ "loss": 4.849811553955078,
+ "step": 907
+ },
+ {
+ "epoch": 12.614847161572053,
+ "grad_norm": 0.015711182728409767,
+ "learning_rate": 0.0006,
+ "loss": 4.7829084396362305,
+ "step": 908
+ },
+ {
+ "epoch": 12.62882096069869,
+ "grad_norm": 0.014137467369437218,
+ "learning_rate": 0.0006,
+ "loss": 4.73321533203125,
+ "step": 909
+ },
+ {
+ "epoch": 12.642794759825328,
+ "grad_norm": 0.012738462537527084,
+ "learning_rate": 0.0006,
+ "loss": 4.870052337646484,
+ "step": 910
+ },
+ {
+ "epoch": 12.656768558951965,
+ "grad_norm": 0.013683130033314228,
+ "learning_rate": 0.0006,
+ "loss": 4.82658052444458,
+ "step": 911
+ },
+ {
+ "epoch": 12.670742358078602,
+ "grad_norm": 0.014409830793738365,
+ "learning_rate": 0.0006,
+ "loss": 4.707494735717773,
+ "step": 912
+ },
+ {
+ "epoch": 12.68471615720524,
+ "grad_norm": 0.013806809671223164,
+ "learning_rate": 0.0006,
+ "loss": 4.778928279876709,
+ "step": 913
+ },
+ {
+ "epoch": 12.698689956331878,
+ "grad_norm": 0.012857912108302116,
+ "learning_rate": 0.0006,
+ "loss": 4.890717506408691,
+ "step": 914
+ },
+ {
+ "epoch": 12.712663755458514,
+ "grad_norm": 0.013828758150339127,
+ "learning_rate": 0.0006,
+ "loss": 4.8357439041137695,
+ "step": 915
+ },
+ {
+ "epoch": 12.726637554585153,
+ "grad_norm": 0.01478351280093193,
+ "learning_rate": 0.0006,
+ "loss": 4.869986534118652,
+ "step": 916
+ },
+ {
+ "epoch": 12.74061135371179,
+ "grad_norm": 0.01326421182602644,
+ "learning_rate": 0.0006,
+ "loss": 4.830646514892578,
+ "step": 917
+ },
+ {
+ "epoch": 12.754585152838429,
+ "grad_norm": 0.013335433788597584,
+ "learning_rate": 0.0006,
+ "loss": 4.7786149978637695,
+ "step": 918
+ },
+ {
+ "epoch": 12.768558951965066,
+ "grad_norm": 0.014261147007346153,
+ "learning_rate": 0.0006,
+ "loss": 4.788693428039551,
+ "step": 919
+ },
+ {
+ "epoch": 12.782532751091702,
+ "grad_norm": 0.014274783432483673,
+ "learning_rate": 0.0006,
+ "loss": 4.77295446395874,
+ "step": 920
+ },
+ {
+ "epoch": 12.796506550218341,
+ "grad_norm": 0.015074674971401691,
+ "learning_rate": 0.0006,
+ "loss": 4.830754280090332,
+ "step": 921
+ },
+ {
+ "epoch": 12.810480349344978,
+ "grad_norm": 0.015599515289068222,
+ "learning_rate": 0.0006,
+ "loss": 4.837759017944336,
+ "step": 922
+ },
+ {
+ "epoch": 12.824454148471617,
+ "grad_norm": 0.014595242217183113,
+ "learning_rate": 0.0006,
+ "loss": 4.882965564727783,
+ "step": 923
+ },
+ {
+ "epoch": 12.838427947598253,
+ "grad_norm": 0.014951467514038086,
+ "learning_rate": 0.0006,
+ "loss": 4.709748268127441,
+ "step": 924
+ },
+ {
+ "epoch": 12.85240174672489,
+ "grad_norm": 0.01539295818656683,
+ "learning_rate": 0.0006,
+ "loss": 4.791160583496094,
+ "step": 925
+ },
+ {
+ "epoch": 12.866375545851529,
+ "grad_norm": 0.015987906605005264,
+ "learning_rate": 0.0006,
+ "loss": 4.661876678466797,
+ "step": 926
+ },
+ {
+ "epoch": 12.880349344978166,
+ "grad_norm": 0.016527341678738594,
+ "learning_rate": 0.0006,
+ "loss": 4.721952438354492,
+ "step": 927
+ },
+ {
+ "epoch": 12.894323144104803,
+ "grad_norm": 0.01637161523103714,
+ "learning_rate": 0.0006,
+ "loss": 4.704446792602539,
+ "step": 928
+ },
+ {
+ "epoch": 12.908296943231441,
+ "grad_norm": 0.015221050009131432,
+ "learning_rate": 0.0006,
+ "loss": 4.771907806396484,
+ "step": 929
+ },
+ {
+ "epoch": 12.922270742358078,
+ "grad_norm": 0.016242166981101036,
+ "learning_rate": 0.0006,
+ "loss": 4.806267738342285,
+ "step": 930
+ },
+ {
+ "epoch": 12.936244541484717,
+ "grad_norm": 0.01628357358276844,
+ "learning_rate": 0.0006,
+ "loss": 4.7252326011657715,
+ "step": 931
+ },
+ {
+ "epoch": 12.950218340611354,
+ "grad_norm": 0.017428133636713028,
+ "learning_rate": 0.0006,
+ "loss": 4.765689849853516,
+ "step": 932
+ },
+ {
+ "epoch": 12.96419213973799,
+ "grad_norm": 0.01870061084628105,
+ "learning_rate": 0.0006,
+ "loss": 4.814488887786865,
+ "step": 933
+ },
+ {
+ "epoch": 12.97816593886463,
+ "grad_norm": 0.019796060398221016,
+ "learning_rate": 0.0006,
+ "loss": 4.739185810089111,
+ "step": 934
+ },
+ {
+ "epoch": 12.992139737991266,
+ "grad_norm": 0.01985820196568966,
+ "learning_rate": 0.0006,
+ "loss": 4.810608863830566,
+ "step": 935
+ },
+ {
+ "epoch": 13.0,
+ "grad_norm": 0.020834477618336678,
+ "learning_rate": 0.0006,
+ "loss": 4.727858066558838,
+ "step": 936
+ },
+ {
+ "epoch": 13.0,
+ "eval_loss": 4.979116916656494,
+ "eval_runtime": 57.2877,
+ "eval_samples_per_second": 42.627,
+ "eval_steps_per_second": 1.344,
+ "step": 936
+ },
+ {
+ "epoch": 13.013973799126637,
+ "grad_norm": 0.02025519125163555,
+ "learning_rate": 0.0006,
+ "loss": 4.628875732421875,
+ "step": 937
+ },
+ {
+ "epoch": 13.027947598253276,
+ "grad_norm": 0.019596368074417114,
+ "learning_rate": 0.0006,
+ "loss": 4.694699287414551,
+ "step": 938
+ },
+ {
+ "epoch": 13.041921397379912,
+ "grad_norm": 0.020494934171438217,
+ "learning_rate": 0.0006,
+ "loss": 4.691821098327637,
+ "step": 939
+ },
+ {
+ "epoch": 13.055895196506551,
+ "grad_norm": 0.020121760666370392,
+ "learning_rate": 0.0006,
+ "loss": 4.71478271484375,
+ "step": 940
+ },
+ {
+ "epoch": 13.069868995633188,
+ "grad_norm": 0.021091405302286148,
+ "learning_rate": 0.0006,
+ "loss": 4.695915222167969,
+ "step": 941
+ },
+ {
+ "epoch": 13.083842794759825,
+ "grad_norm": 0.019787423312664032,
+ "learning_rate": 0.0006,
+ "loss": 4.8652448654174805,
+ "step": 942
+ },
+ {
+ "epoch": 13.097816593886463,
+ "grad_norm": 0.0185833927243948,
+ "learning_rate": 0.0006,
+ "loss": 4.7065887451171875,
+ "step": 943
+ },
+ {
+ "epoch": 13.1117903930131,
+ "grad_norm": 0.021220475435256958,
+ "learning_rate": 0.0006,
+ "loss": 4.741232395172119,
+ "step": 944
+ },
+ {
+ "epoch": 13.125764192139737,
+ "grad_norm": 0.02136622555553913,
+ "learning_rate": 0.0006,
+ "loss": 4.705967903137207,
+ "step": 945
+ },
+ {
+ "epoch": 13.139737991266376,
+ "grad_norm": 0.0191587433218956,
+ "learning_rate": 0.0006,
+ "loss": 4.758667945861816,
+ "step": 946
+ },
+ {
+ "epoch": 13.153711790393013,
+ "grad_norm": 0.018290022388100624,
+ "learning_rate": 0.0006,
+ "loss": 4.760666847229004,
+ "step": 947
+ },
+ {
+ "epoch": 13.167685589519651,
+ "grad_norm": 0.021704409271478653,
+ "learning_rate": 0.0006,
+ "loss": 4.750555992126465,
+ "step": 948
+ },
+ {
+ "epoch": 13.181659388646288,
+ "grad_norm": 0.021668918430805206,
+ "learning_rate": 0.0006,
+ "loss": 4.840903282165527,
+ "step": 949
+ },
+ {
+ "epoch": 13.195633187772925,
+ "grad_norm": 0.019497141242027283,
+ "learning_rate": 0.0006,
+ "loss": 4.738969326019287,
+ "step": 950
+ },
+ {
+ "epoch": 13.209606986899564,
+ "grad_norm": 0.018500596284866333,
+ "learning_rate": 0.0006,
+ "loss": 4.75154972076416,
+ "step": 951
+ },
+ {
+ "epoch": 13.2235807860262,
+ "grad_norm": 0.01872239261865616,
+ "learning_rate": 0.0006,
+ "loss": 4.811738014221191,
+ "step": 952
+ },
+ {
+ "epoch": 13.237554585152838,
+ "grad_norm": 0.020321547985076904,
+ "learning_rate": 0.0006,
+ "loss": 4.815384387969971,
+ "step": 953
+ },
+ {
+ "epoch": 13.251528384279476,
+ "grad_norm": 0.02283575013279915,
+ "learning_rate": 0.0006,
+ "loss": 4.74582052230835,
+ "step": 954
+ },
+ {
+ "epoch": 13.265502183406113,
+ "grad_norm": 0.02474398910999298,
+ "learning_rate": 0.0006,
+ "loss": 4.738375186920166,
+ "step": 955
+ },
+ {
+ "epoch": 13.279475982532752,
+ "grad_norm": 0.02065764181315899,
+ "learning_rate": 0.0006,
+ "loss": 4.746875286102295,
+ "step": 956
+ },
+ {
+ "epoch": 13.293449781659389,
+ "grad_norm": 0.020534520968794823,
+ "learning_rate": 0.0006,
+ "loss": 4.656914710998535,
+ "step": 957
+ },
+ {
+ "epoch": 13.307423580786025,
+ "grad_norm": 0.0206378772854805,
+ "learning_rate": 0.0006,
+ "loss": 4.842615604400635,
+ "step": 958
+ },
+ {
+ "epoch": 13.321397379912664,
+ "grad_norm": 0.017935654148459435,
+ "learning_rate": 0.0006,
+ "loss": 4.717827796936035,
+ "step": 959
+ },
+ {
+ "epoch": 13.335371179039301,
+ "grad_norm": 0.016958363354206085,
+ "learning_rate": 0.0006,
+ "loss": 4.681070327758789,
+ "step": 960
+ },
+ {
+ "epoch": 13.34934497816594,
+ "grad_norm": 0.015735412016510963,
+ "learning_rate": 0.0006,
+ "loss": 4.7957048416137695,
+ "step": 961
+ },
+ {
+ "epoch": 13.363318777292577,
+ "grad_norm": 0.01563699170947075,
+ "learning_rate": 0.0006,
+ "loss": 4.701192378997803,
+ "step": 962
+ },
+ {
+ "epoch": 13.377292576419213,
+ "grad_norm": 0.015207760035991669,
+ "learning_rate": 0.0006,
+ "loss": 4.73256254196167,
+ "step": 963
+ },
+ {
+ "epoch": 13.391266375545852,
+ "grad_norm": 0.015396883711218834,
+ "learning_rate": 0.0006,
+ "loss": 4.76412296295166,
+ "step": 964
+ },
+ {
+ "epoch": 13.405240174672489,
+ "grad_norm": 0.015690583735704422,
+ "learning_rate": 0.0006,
+ "loss": 4.791157245635986,
+ "step": 965
+ },
+ {
+ "epoch": 13.419213973799126,
+ "grad_norm": 0.01755724661052227,
+ "learning_rate": 0.0006,
+ "loss": 4.845666885375977,
+ "step": 966
+ },
+ {
+ "epoch": 13.433187772925764,
+ "grad_norm": 0.019443422555923462,
+ "learning_rate": 0.0006,
+ "loss": 4.5902299880981445,
+ "step": 967
+ },
+ {
+ "epoch": 13.447161572052401,
+ "grad_norm": 0.02118338644504547,
+ "learning_rate": 0.0006,
+ "loss": 4.740627288818359,
+ "step": 968
+ },
+ {
+ "epoch": 13.46113537117904,
+ "grad_norm": 0.02174839936196804,
+ "learning_rate": 0.0006,
+ "loss": 4.814560890197754,
+ "step": 969
+ },
+ {
+ "epoch": 13.475109170305677,
+ "grad_norm": 0.022860374301671982,
+ "learning_rate": 0.0006,
+ "loss": 4.66440486907959,
+ "step": 970
+ },
+ {
+ "epoch": 13.489082969432314,
+ "grad_norm": 0.024742677807807922,
+ "learning_rate": 0.0006,
+ "loss": 4.7451653480529785,
+ "step": 971
+ },
+ {
+ "epoch": 13.503056768558952,
+ "grad_norm": 0.027068907395005226,
+ "learning_rate": 0.0006,
+ "loss": 4.802584171295166,
+ "step": 972
+ },
+ {
+ "epoch": 13.51703056768559,
+ "grad_norm": 0.026316625997424126,
+ "learning_rate": 0.0006,
+ "loss": 4.842397689819336,
+ "step": 973
+ },
+ {
+ "epoch": 13.531004366812226,
+ "grad_norm": 0.024610962718725204,
+ "learning_rate": 0.0006,
+ "loss": 4.650721073150635,
+ "step": 974
+ },
+ {
+ "epoch": 13.544978165938865,
+ "grad_norm": 0.024122603237628937,
+ "learning_rate": 0.0006,
+ "loss": 4.6984357833862305,
+ "step": 975
+ },
+ {
+ "epoch": 13.558951965065502,
+ "grad_norm": 0.022699708119034767,
+ "learning_rate": 0.0006,
+ "loss": 4.793007850646973,
+ "step": 976
+ },
+ {
+ "epoch": 13.57292576419214,
+ "grad_norm": 0.021706534549593925,
+ "learning_rate": 0.0006,
+ "loss": 4.691362380981445,
+ "step": 977
+ },
+ {
+ "epoch": 13.586899563318777,
+ "grad_norm": 0.021363291889429092,
+ "learning_rate": 0.0006,
+ "loss": 4.76762580871582,
+ "step": 978
+ },
+ {
+ "epoch": 13.600873362445414,
+ "grad_norm": 0.023476677015423775,
+ "learning_rate": 0.0006,
+ "loss": 4.777202606201172,
+ "step": 979
+ },
+ {
+ "epoch": 13.614847161572053,
+ "grad_norm": 0.02457100711762905,
+ "learning_rate": 0.0006,
+ "loss": 4.590703010559082,
+ "step": 980
+ },
+ {
+ "epoch": 13.62882096069869,
+ "grad_norm": 0.02469099499285221,
+ "learning_rate": 0.0006,
+ "loss": 4.640185356140137,
+ "step": 981
+ },
+ {
+ "epoch": 13.642794759825328,
+ "grad_norm": 0.024914603680372238,
+ "learning_rate": 0.0006,
+ "loss": 4.772201061248779,
+ "step": 982
+ },
+ {
+ "epoch": 13.656768558951965,
+ "grad_norm": 0.025133173912763596,
+ "learning_rate": 0.0006,
+ "loss": 4.743337154388428,
+ "step": 983
+ },
+ {
+ "epoch": 13.670742358078602,
+ "grad_norm": 0.02034137211740017,
+ "learning_rate": 0.0006,
+ "loss": 4.752127647399902,
+ "step": 984
+ },
+ {
+ "epoch": 13.68471615720524,
+ "grad_norm": 0.019645661115646362,
+ "learning_rate": 0.0006,
+ "loss": 4.838453769683838,
+ "step": 985
+ },
+ {
+ "epoch": 13.698689956331878,
+ "grad_norm": 0.01900297962129116,
+ "learning_rate": 0.0006,
+ "loss": 4.804868221282959,
+ "step": 986
+ },
+ {
+ "epoch": 13.712663755458514,
+ "grad_norm": 0.020292341709136963,
+ "learning_rate": 0.0006,
+ "loss": 4.609743118286133,
+ "step": 987
+ },
+ {
+ "epoch": 13.726637554585153,
+ "grad_norm": 0.017014775425195694,
+ "learning_rate": 0.0006,
+ "loss": 4.740612983703613,
+ "step": 988
+ },
+ {
+ "epoch": 13.74061135371179,
+ "grad_norm": 0.015199844725430012,
+ "learning_rate": 0.0006,
+ "loss": 4.719701766967773,
+ "step": 989
+ },
+ {
+ "epoch": 13.754585152838429,
+ "grad_norm": 0.015125414356589317,
+ "learning_rate": 0.0006,
+ "loss": 4.686114311218262,
+ "step": 990
+ },
+ {
+ "epoch": 13.768558951965066,
+ "grad_norm": 0.014779540710151196,
+ "learning_rate": 0.0006,
+ "loss": 4.674347877502441,
+ "step": 991
+ },
+ {
+ "epoch": 13.782532751091702,
+ "grad_norm": 0.015255163423717022,
+ "learning_rate": 0.0006,
+ "loss": 4.6309309005737305,
+ "step": 992
+ },
+ {
+ "epoch": 13.796506550218341,
+ "grad_norm": 0.013692197389900684,
+ "learning_rate": 0.0006,
+ "loss": 4.698497295379639,
+ "step": 993
+ },
+ {
+ "epoch": 13.810480349344978,
+ "grad_norm": 0.014529623091220856,
+ "learning_rate": 0.0006,
+ "loss": 4.628443717956543,
+ "step": 994
+ },
+ {
+ "epoch": 13.824454148471617,
+ "grad_norm": 0.01576659083366394,
+ "learning_rate": 0.0006,
+ "loss": 4.788015842437744,
+ "step": 995
+ },
+ {
+ "epoch": 13.838427947598253,
+ "grad_norm": 0.015834277495741844,
+ "learning_rate": 0.0006,
+ "loss": 4.672703742980957,
+ "step": 996
+ },
+ {
+ "epoch": 13.85240174672489,
+ "grad_norm": 0.015539503656327724,
+ "learning_rate": 0.0006,
+ "loss": 4.731503486633301,
+ "step": 997
+ },
+ {
+ "epoch": 13.866375545851529,
+ "grad_norm": 0.013463746756315231,
+ "learning_rate": 0.0006,
+ "loss": 4.7006330490112305,
+ "step": 998
+ },
+ {
+ "epoch": 13.880349344978166,
+ "grad_norm": 0.016377726569771767,
+ "learning_rate": 0.0006,
+ "loss": 4.636074066162109,
+ "step": 999
+ },
+ {
+ "epoch": 13.894323144104803,
+ "grad_norm": 0.016660034656524658,
+ "learning_rate": 0.0006,
+ "loss": 4.742693901062012,
+ "step": 1000
+ },
+ {
+ "epoch": 13.908296943231441,
+ "grad_norm": 0.014902938157320023,
+ "learning_rate": 0.0006,
+ "loss": 4.769811630249023,
+ "step": 1001
+ },
+ {
+ "epoch": 13.922270742358078,
+ "grad_norm": 0.014883476309478283,
+ "learning_rate": 0.0006,
+ "loss": 4.731159210205078,
+ "step": 1002
+ },
+ {
+ "epoch": 13.936244541484717,
+ "grad_norm": 0.014494217932224274,
+ "learning_rate": 0.0006,
+ "loss": 4.659485816955566,
+ "step": 1003
+ },
+ {
+ "epoch": 13.950218340611354,
+ "grad_norm": 0.015426991507411003,
+ "learning_rate": 0.0006,
+ "loss": 4.755229949951172,
+ "step": 1004
+ },
+ {
+ "epoch": 13.96419213973799,
+ "grad_norm": 0.016648586839437485,
+ "learning_rate": 0.0006,
+ "loss": 4.74111270904541,
+ "step": 1005
+ },
+ {
+ "epoch": 13.97816593886463,
+ "grad_norm": 0.016980541869997978,
+ "learning_rate": 0.0006,
+ "loss": 4.74072265625,
+ "step": 1006
+ },
+ {
+ "epoch": 13.992139737991266,
+ "grad_norm": 0.018070152029395103,
+ "learning_rate": 0.0006,
+ "loss": 4.732975006103516,
+ "step": 1007
+ },
+ {
+ "epoch": 14.0,
+ "grad_norm": 0.019910665228962898,
+ "learning_rate": 0.0006,
+ "loss": 4.634697914123535,
+ "step": 1008
+ },
+ {
+ "epoch": 14.0,
+ "eval_loss": 4.850640296936035,
+ "eval_runtime": 57.3039,
+ "eval_samples_per_second": 42.615,
+ "eval_steps_per_second": 1.344,
+ "step": 1008
+ },
+ {
+ "epoch": 14.013973799126637,
+ "grad_norm": 0.01980053074657917,
+ "learning_rate": 0.0006,
+ "loss": 4.628744602203369,
+ "step": 1009
+ },
+ {
+ "epoch": 14.027947598253276,
+ "grad_norm": 0.02497623674571514,
+ "learning_rate": 0.0006,
+ "loss": 4.75930118560791,
+ "step": 1010
+ },
+ {
+ "epoch": 14.041921397379912,
+ "grad_norm": 0.02794097363948822,
+ "learning_rate": 0.0006,
+ "loss": 4.691340446472168,
+ "step": 1011
+ },
+ {
+ "epoch": 14.055895196506551,
+ "grad_norm": 0.023725535720586777,
+ "learning_rate": 0.0006,
+ "loss": 4.733664512634277,
+ "step": 1012
+ },
+ {
+ "epoch": 14.069868995633188,
+ "grad_norm": 0.022966474294662476,
+ "learning_rate": 0.0006,
+ "loss": 4.664808750152588,
+ "step": 1013
+ },
+ {
+ "epoch": 14.083842794759825,
+ "grad_norm": 0.02530563250184059,
+ "learning_rate": 0.0006,
+ "loss": 4.8047285079956055,
+ "step": 1014
+ },
+ {
+ "epoch": 14.097816593886463,
+ "grad_norm": 0.02491082064807415,
+ "learning_rate": 0.0006,
+ "loss": 4.667956352233887,
+ "step": 1015
+ },
+ {
+ "epoch": 14.1117903930131,
+ "grad_norm": 0.022110294550657272,
+ "learning_rate": 0.0006,
+ "loss": 4.666140556335449,
+ "step": 1016
+ },
+ {
+ "epoch": 14.125764192139737,
+ "grad_norm": 0.02065849117934704,
+ "learning_rate": 0.0006,
+ "loss": 4.581960678100586,
+ "step": 1017
+ },
+ {
+ "epoch": 14.139737991266376,
+ "grad_norm": 0.01941298507153988,
+ "learning_rate": 0.0006,
+ "loss": 4.614226341247559,
+ "step": 1018
+ },
+ {
+ "epoch": 14.153711790393013,
+ "grad_norm": 0.022441856563091278,
+ "learning_rate": 0.0006,
+ "loss": 4.638352394104004,
+ "step": 1019
+ },
+ {
+ "epoch": 14.167685589519651,
+ "grad_norm": 0.021519694477319717,
+ "learning_rate": 0.0006,
+ "loss": 4.700160980224609,
+ "step": 1020
+ },
+ {
+ "epoch": 14.181659388646288,
+ "grad_norm": 0.018956486135721207,
+ "learning_rate": 0.0006,
+ "loss": 4.649593353271484,
+ "step": 1021
+ },
+ {
+ "epoch": 14.195633187772925,
+ "grad_norm": 0.016681140288710594,
+ "learning_rate": 0.0006,
+ "loss": 4.603952407836914,
+ "step": 1022
+ },
+ {
+ "epoch": 14.209606986899564,
+ "grad_norm": 0.01788809522986412,
+ "learning_rate": 0.0006,
+ "loss": 4.669160842895508,
+ "step": 1023
+ },
+ {
+ "epoch": 14.2235807860262,
+ "grad_norm": 0.016047228127717972,
+ "learning_rate": 0.0006,
+ "loss": 4.6367950439453125,
+ "step": 1024
+ },
+ {
+ "epoch": 14.237554585152838,
+ "grad_norm": 0.01614239625632763,
+ "learning_rate": 0.0006,
+ "loss": 4.664529323577881,
+ "step": 1025
+ },
+ {
+ "epoch": 14.251528384279476,
+ "grad_norm": 0.016362549737095833,
+ "learning_rate": 0.0006,
+ "loss": 4.702373504638672,
+ "step": 1026
+ },
+ {
+ "epoch": 14.265502183406113,
+ "grad_norm": 0.015800409018993378,
+ "learning_rate": 0.0006,
+ "loss": 4.7594428062438965,
+ "step": 1027
+ },
+ {
+ "epoch": 14.279475982532752,
+ "grad_norm": 0.016730139032006264,
+ "learning_rate": 0.0006,
+ "loss": 4.677244186401367,
+ "step": 1028
+ },
+ {
+ "epoch": 14.293449781659389,
+ "grad_norm": 0.01661759801208973,
+ "learning_rate": 0.0006,
+ "loss": 4.633679389953613,
+ "step": 1029
+ },
+ {
+ "epoch": 14.307423580786025,
+ "grad_norm": 0.017673347145318985,
+ "learning_rate": 0.0006,
+ "loss": 4.649107456207275,
+ "step": 1030
+ },
+ {
+ "epoch": 14.321397379912664,
+ "grad_norm": 0.019243352115154266,
+ "learning_rate": 0.0006,
+ "loss": 4.611257553100586,
+ "step": 1031
+ },
+ {
+ "epoch": 14.335371179039301,
+ "grad_norm": 0.019817164167761803,
+ "learning_rate": 0.0006,
+ "loss": 4.588086128234863,
+ "step": 1032
+ },
+ {
+ "epoch": 14.34934497816594,
+ "grad_norm": 0.022292211651802063,
+ "learning_rate": 0.0006,
+ "loss": 4.783066749572754,
+ "step": 1033
+ },
+ {
+ "epoch": 14.363318777292577,
+ "grad_norm": 0.02140096202492714,
+ "learning_rate": 0.0006,
+ "loss": 4.604718208312988,
+ "step": 1034
+ },
+ {
+ "epoch": 14.377292576419213,
+ "grad_norm": 0.0187817495316267,
+ "learning_rate": 0.0006,
+ "loss": 4.721773147583008,
+ "step": 1035
+ },
+ {
+ "epoch": 14.391266375545852,
+ "grad_norm": 0.019879210740327835,
+ "learning_rate": 0.0006,
+ "loss": 4.789286136627197,
+ "step": 1036
+ },
+ {
+ "epoch": 14.405240174672489,
+ "grad_norm": 0.01885785162448883,
+ "learning_rate": 0.0006,
+ "loss": 4.669743537902832,
+ "step": 1037
+ },
+ {
+ "epoch": 14.419213973799126,
+ "grad_norm": 0.018324270844459534,
+ "learning_rate": 0.0006,
+ "loss": 4.65901517868042,
+ "step": 1038
+ },
+ {
+ "epoch": 14.433187772925764,
+ "grad_norm": 0.017664821818470955,
+ "learning_rate": 0.0006,
+ "loss": 4.667839527130127,
+ "step": 1039
+ },
+ {
+ "epoch": 14.447161572052401,
+ "grad_norm": 0.015702905133366585,
+ "learning_rate": 0.0006,
+ "loss": 4.64700984954834,
+ "step": 1040
+ },
+ {
+ "epoch": 14.46113537117904,
+ "grad_norm": 0.016674194484949112,
+ "learning_rate": 0.0006,
+ "loss": 4.694093227386475,
+ "step": 1041
+ },
+ {
+ "epoch": 14.475109170305677,
+ "grad_norm": 0.016869477927684784,
+ "learning_rate": 0.0006,
+ "loss": 4.619544982910156,
+ "step": 1042
+ },
+ {
+ "epoch": 14.489082969432314,
+ "grad_norm": 0.016557540744543076,
+ "learning_rate": 0.0006,
+ "loss": 4.568667411804199,
+ "step": 1043
+ },
+ {
+ "epoch": 14.503056768558952,
+ "grad_norm": 0.014500945806503296,
+ "learning_rate": 0.0006,
+ "loss": 4.561741828918457,
+ "step": 1044
+ },
+ {
+ "epoch": 14.51703056768559,
+ "grad_norm": 0.014228739775717258,
+ "learning_rate": 0.0006,
+ "loss": 4.654277324676514,
+ "step": 1045
+ },
+ {
+ "epoch": 14.531004366812226,
+ "grad_norm": 0.013894238509237766,
+ "learning_rate": 0.0006,
+ "loss": 4.5867719650268555,
+ "step": 1046
+ },
+ {
+ "epoch": 14.544978165938865,
+ "grad_norm": 0.01550068985670805,
+ "learning_rate": 0.0006,
+ "loss": 4.61168098449707,
+ "step": 1047
+ },
+ {
+ "epoch": 14.558951965065502,
+ "grad_norm": 0.015940772369503975,
+ "learning_rate": 0.0006,
+ "loss": 4.691247940063477,
+ "step": 1048
+ },
+ {
+ "epoch": 14.57292576419214,
+ "grad_norm": 0.01714535616338253,
+ "learning_rate": 0.0006,
+ "loss": 4.61183500289917,
+ "step": 1049
+ },
+ {
+ "epoch": 14.586899563318777,
+ "grad_norm": 0.017925038933753967,
+ "learning_rate": 0.0006,
+ "loss": 4.719022750854492,
+ "step": 1050
+ },
+ {
+ "epoch": 14.600873362445414,
+ "grad_norm": 0.018141552805900574,
+ "learning_rate": 0.0006,
+ "loss": 4.760281562805176,
+ "step": 1051
+ },
+ {
+ "epoch": 14.614847161572053,
+ "grad_norm": 0.017131328582763672,
+ "learning_rate": 0.0006,
+ "loss": 4.709680080413818,
+ "step": 1052
+ },
+ {
+ "epoch": 14.62882096069869,
+ "grad_norm": 0.018730048090219498,
+ "learning_rate": 0.0006,
+ "loss": 4.5994720458984375,
+ "step": 1053
+ },
+ {
+ "epoch": 14.642794759825328,
+ "grad_norm": 0.018894298002123833,
+ "learning_rate": 0.0006,
+ "loss": 4.669070720672607,
+ "step": 1054
+ },
+ {
+ "epoch": 14.656768558951965,
+ "grad_norm": 0.017643198370933533,
+ "learning_rate": 0.0006,
+ "loss": 4.641790390014648,
+ "step": 1055
+ },
+ {
+ "epoch": 14.670742358078602,
+ "grad_norm": 0.017647244036197662,
+ "learning_rate": 0.0006,
+ "loss": 4.653899192810059,
+ "step": 1056
+ },
+ {
+ "epoch": 14.68471615720524,
+ "grad_norm": 0.017304198816418648,
+ "learning_rate": 0.0006,
+ "loss": 4.7487473487854,
+ "step": 1057
+ },
+ {
+ "epoch": 14.698689956331878,
+ "grad_norm": 0.01638447679579258,
+ "learning_rate": 0.0006,
+ "loss": 4.71945858001709,
+ "step": 1058
+ },
+ {
+ "epoch": 14.712663755458514,
+ "grad_norm": 0.015892071649432182,
+ "learning_rate": 0.0006,
+ "loss": 4.744903564453125,
+ "step": 1059
+ },
+ {
+ "epoch": 14.726637554585153,
+ "grad_norm": 0.017420614138245583,
+ "learning_rate": 0.0006,
+ "loss": 4.781563758850098,
+ "step": 1060
+ },
+ {
+ "epoch": 14.74061135371179,
+ "grad_norm": 0.016498520970344543,
+ "learning_rate": 0.0006,
+ "loss": 4.622779846191406,
+ "step": 1061
+ },
+ {
+ "epoch": 14.754585152838429,
+ "grad_norm": 0.017482127994298935,
+ "learning_rate": 0.0006,
+ "loss": 4.62825870513916,
+ "step": 1062
+ },
+ {
+ "epoch": 14.768558951965066,
+ "grad_norm": 0.019018379971385002,
+ "learning_rate": 0.0006,
+ "loss": 4.693194389343262,
+ "step": 1063
+ },
+ {
+ "epoch": 14.782532751091702,
+ "grad_norm": 0.01993914321064949,
+ "learning_rate": 0.0006,
+ "loss": 4.742903709411621,
+ "step": 1064
+ },
+ {
+ "epoch": 14.796506550218341,
+ "grad_norm": 0.02301335707306862,
+ "learning_rate": 0.0006,
+ "loss": 4.613251686096191,
+ "step": 1065
+ },
+ {
+ "epoch": 14.810480349344978,
+ "grad_norm": 0.024088222533464432,
+ "learning_rate": 0.0006,
+ "loss": 4.594305992126465,
+ "step": 1066
+ },
+ {
+ "epoch": 14.824454148471617,
+ "grad_norm": 0.02493320405483246,
+ "learning_rate": 0.0006,
+ "loss": 4.700588226318359,
+ "step": 1067
+ },
+ {
+ "epoch": 14.838427947598253,
+ "grad_norm": 0.02501937560737133,
+ "learning_rate": 0.0006,
+ "loss": 4.601742744445801,
+ "step": 1068
+ },
+ {
+ "epoch": 14.85240174672489,
+ "grad_norm": 0.023490751162171364,
+ "learning_rate": 0.0006,
+ "loss": 4.589211463928223,
+ "step": 1069
+ },
+ {
+ "epoch": 14.866375545851529,
+ "grad_norm": 0.019695665687322617,
+ "learning_rate": 0.0006,
+ "loss": 4.635081768035889,
+ "step": 1070
+ },
+ {
+ "epoch": 14.880349344978166,
+ "grad_norm": 0.019276980310678482,
+ "learning_rate": 0.0006,
+ "loss": 4.737002372741699,
+ "step": 1071
+ },
+ {
+ "epoch": 14.894323144104803,
+ "grad_norm": 0.022430511191487312,
+ "learning_rate": 0.0006,
+ "loss": 4.808295726776123,
+ "step": 1072
+ },
+ {
+ "epoch": 14.908296943231441,
+ "grad_norm": 0.021543916314840317,
+ "learning_rate": 0.0006,
+ "loss": 4.559475421905518,
+ "step": 1073
+ },
+ {
+ "epoch": 14.922270742358078,
+ "grad_norm": 0.019885241985321045,
+ "learning_rate": 0.0006,
+ "loss": 4.647948741912842,
+ "step": 1074
+ },
+ {
+ "epoch": 14.936244541484717,
+ "grad_norm": 0.02115175500512123,
+ "learning_rate": 0.0006,
+ "loss": 4.680505752563477,
+ "step": 1075
+ },
+ {
+ "epoch": 14.950218340611354,
+ "grad_norm": 0.02156151458621025,
+ "learning_rate": 0.0006,
+ "loss": 4.673477649688721,
+ "step": 1076
+ },
+ {
+ "epoch": 14.96419213973799,
+ "grad_norm": 0.02184601128101349,
+ "learning_rate": 0.0006,
+ "loss": 4.644402027130127,
+ "step": 1077
+ },
+ {
+ "epoch": 14.97816593886463,
+ "grad_norm": 0.0211983360350132,
+ "learning_rate": 0.0006,
+ "loss": 4.821558475494385,
+ "step": 1078
+ },
+ {
+ "epoch": 14.992139737991266,
+ "grad_norm": 0.019636567682027817,
+ "learning_rate": 0.0006,
+ "loss": 4.578868865966797,
+ "step": 1079
+ },
+ {
+ "epoch": 15.0,
+ "grad_norm": 0.019997350871562958,
+ "learning_rate": 0.0006,
+ "loss": 4.691307544708252,
+ "step": 1080
+ },
+ {
+ "epoch": 15.0,
+ "eval_loss": 4.872478008270264,
+ "eval_runtime": 56.8513,
+ "eval_samples_per_second": 42.954,
+ "eval_steps_per_second": 1.354,
+ "step": 1080
+ },
+ {
+ "epoch": 15.013973799126637,
+ "grad_norm": 0.018830662593245506,
+ "learning_rate": 0.0006,
+ "loss": 4.761779308319092,
+ "step": 1081
+ },
+ {
+ "epoch": 15.027947598253276,
+ "grad_norm": 0.016876481473445892,
+ "learning_rate": 0.0006,
+ "loss": 4.61148738861084,
+ "step": 1082
+ },
+ {
+ "epoch": 15.041921397379912,
+ "grad_norm": 0.015550863929092884,
+ "learning_rate": 0.0006,
+ "loss": 4.676124095916748,
+ "step": 1083
+ },
+ {
+ "epoch": 15.055895196506551,
+ "grad_norm": 0.017178896814584732,
+ "learning_rate": 0.0006,
+ "loss": 4.588935852050781,
+ "step": 1084
+ },
+ {
+ "epoch": 15.069868995633188,
+ "grad_norm": 0.015465447679162025,
+ "learning_rate": 0.0006,
+ "loss": 4.605499267578125,
+ "step": 1085
+ },
+ {
+ "epoch": 15.083842794759825,
+ "grad_norm": 0.01606195978820324,
+ "learning_rate": 0.0006,
+ "loss": 4.723373889923096,
+ "step": 1086
+ },
+ {
+ "epoch": 15.097816593886463,
+ "grad_norm": 0.016521936282515526,
+ "learning_rate": 0.0006,
+ "loss": 4.646608352661133,
+ "step": 1087
+ },
+ {
+ "epoch": 15.1117903930131,
+ "grad_norm": 0.01589205674827099,
+ "learning_rate": 0.0006,
+ "loss": 4.498138427734375,
+ "step": 1088
+ },
+ {
+ "epoch": 15.125764192139737,
+ "grad_norm": 0.01524856686592102,
+ "learning_rate": 0.0006,
+ "loss": 4.594562530517578,
+ "step": 1089
+ },
+ {
+ "epoch": 15.139737991266376,
+ "grad_norm": 0.014714641496539116,
+ "learning_rate": 0.0006,
+ "loss": 4.630744934082031,
+ "step": 1090
+ },
+ {
+ "epoch": 15.153711790393013,
+ "grad_norm": 0.015333331190049648,
+ "learning_rate": 0.0006,
+ "loss": 4.682388782501221,
+ "step": 1091
+ },
+ {
+ "epoch": 15.167685589519651,
+ "grad_norm": 0.01550073828548193,
+ "learning_rate": 0.0006,
+ "loss": 4.692652702331543,
+ "step": 1092
+ },
+ {
+ "epoch": 15.181659388646288,
+ "grad_norm": 0.014783318154513836,
+ "learning_rate": 0.0006,
+ "loss": 4.6223835945129395,
+ "step": 1093
+ },
+ {
+ "epoch": 15.195633187772925,
+ "grad_norm": 0.014372751116752625,
+ "learning_rate": 0.0006,
+ "loss": 4.52687406539917,
+ "step": 1094
+ },
+ {
+ "epoch": 15.209606986899564,
+ "grad_norm": 0.014779130928218365,
+ "learning_rate": 0.0006,
+ "loss": 4.637009620666504,
+ "step": 1095
+ },
+ {
+ "epoch": 15.2235807860262,
+ "grad_norm": 0.015697073191404343,
+ "learning_rate": 0.0006,
+ "loss": 4.682981014251709,
+ "step": 1096
+ },
+ {
+ "epoch": 15.237554585152838,
+ "grad_norm": 0.016523664817214012,
+ "learning_rate": 0.0006,
+ "loss": 4.705722808837891,
+ "step": 1097
+ },
+ {
+ "epoch": 15.251528384279476,
+ "grad_norm": 0.017633996903896332,
+ "learning_rate": 0.0006,
+ "loss": 4.574686050415039,
+ "step": 1098
+ },
+ {
+ "epoch": 15.265502183406113,
+ "grad_norm": 0.019955076277256012,
+ "learning_rate": 0.0006,
+ "loss": 4.634450912475586,
+ "step": 1099
+ },
+ {
+ "epoch": 15.279475982532752,
+ "grad_norm": 0.02132618986070156,
+ "learning_rate": 0.0006,
+ "loss": 4.56110143661499,
+ "step": 1100
+ },
+ {
+ "epoch": 15.293449781659389,
+ "grad_norm": 0.0207593385130167,
+ "learning_rate": 0.0006,
+ "loss": 4.629396438598633,
+ "step": 1101
+ },
+ {
+ "epoch": 15.307423580786025,
+ "grad_norm": 0.017882540822029114,
+ "learning_rate": 0.0006,
+ "loss": 4.730963230133057,
+ "step": 1102
+ },
+ {
+ "epoch": 15.321397379912664,
+ "grad_norm": 0.01917876861989498,
+ "learning_rate": 0.0006,
+ "loss": 4.64585542678833,
+ "step": 1103
+ },
+ {
+ "epoch": 15.335371179039301,
+ "grad_norm": 0.01943974196910858,
+ "learning_rate": 0.0006,
+ "loss": 4.656181335449219,
+ "step": 1104
+ },
+ {
+ "epoch": 15.34934497816594,
+ "grad_norm": 0.019904715940356255,
+ "learning_rate": 0.0006,
+ "loss": 4.619329452514648,
+ "step": 1105
+ },
+ {
+ "epoch": 15.363318777292577,
+ "grad_norm": 0.019174346700310707,
+ "learning_rate": 0.0006,
+ "loss": 4.647892475128174,
+ "step": 1106
+ },
+ {
+ "epoch": 15.377292576419213,
+ "grad_norm": 0.018244758248329163,
+ "learning_rate": 0.0006,
+ "loss": 4.56166934967041,
+ "step": 1107
+ },
+ {
+ "epoch": 15.391266375545852,
+ "grad_norm": 0.018914205953478813,
+ "learning_rate": 0.0006,
+ "loss": 4.648063659667969,
+ "step": 1108
+ },
+ {
+ "epoch": 15.405240174672489,
+ "grad_norm": 0.01961071416735649,
+ "learning_rate": 0.0006,
+ "loss": 4.665525436401367,
+ "step": 1109
+ },
+ {
+ "epoch": 15.419213973799126,
+ "grad_norm": 0.01841311901807785,
+ "learning_rate": 0.0006,
+ "loss": 4.611342430114746,
+ "step": 1110
+ },
+ {
+ "epoch": 15.433187772925764,
+ "grad_norm": 0.01863427273929119,
+ "learning_rate": 0.0006,
+ "loss": 4.664256572723389,
+ "step": 1111
+ },
+ {
+ "epoch": 15.447161572052401,
+ "grad_norm": 0.020506877452135086,
+ "learning_rate": 0.0006,
+ "loss": 4.631237506866455,
+ "step": 1112
+ },
+ {
+ "epoch": 15.46113537117904,
+ "grad_norm": 0.02223833277821541,
+ "learning_rate": 0.0006,
+ "loss": 4.668654441833496,
+ "step": 1113
+ },
+ {
+ "epoch": 15.475109170305677,
+ "grad_norm": 0.023336702957749367,
+ "learning_rate": 0.0006,
+ "loss": 4.562833309173584,
+ "step": 1114
+ },
+ {
+ "epoch": 15.489082969432314,
+ "grad_norm": 0.023052413016557693,
+ "learning_rate": 0.0006,
+ "loss": 4.621532917022705,
+ "step": 1115
+ },
+ {
+ "epoch": 15.503056768558952,
+ "grad_norm": 0.023420924320816994,
+ "learning_rate": 0.0006,
+ "loss": 4.665465354919434,
+ "step": 1116
+ },
+ {
+ "epoch": 15.51703056768559,
+ "grad_norm": 0.024071281775832176,
+ "learning_rate": 0.0006,
+ "loss": 4.595980644226074,
+ "step": 1117
+ },
+ {
+ "epoch": 15.531004366812226,
+ "grad_norm": 0.023406682536005974,
+ "learning_rate": 0.0006,
+ "loss": 4.642635345458984,
+ "step": 1118
+ },
+ {
+ "epoch": 15.544978165938865,
+ "grad_norm": 0.02138647809624672,
+ "learning_rate": 0.0006,
+ "loss": 4.704513072967529,
+ "step": 1119
+ },
+ {
+ "epoch": 15.558951965065502,
+ "grad_norm": 0.0221384447067976,
+ "learning_rate": 0.0006,
+ "loss": 4.567503452301025,
+ "step": 1120
+ },
+ {
+ "epoch": 15.57292576419214,
+ "grad_norm": 0.020879851654171944,
+ "learning_rate": 0.0006,
+ "loss": 4.664057731628418,
+ "step": 1121
+ },
+ {
+ "epoch": 15.586899563318777,
+ "grad_norm": 0.021478967741131783,
+ "learning_rate": 0.0006,
+ "loss": 4.539627552032471,
+ "step": 1122
+ },
+ {
+ "epoch": 15.600873362445414,
+ "grad_norm": 0.022748656570911407,
+ "learning_rate": 0.0006,
+ "loss": 4.5654497146606445,
+ "step": 1123
+ },
+ {
+ "epoch": 15.614847161572053,
+ "grad_norm": 0.021870248019695282,
+ "learning_rate": 0.0006,
+ "loss": 4.56899356842041,
+ "step": 1124
+ },
+ {
+ "epoch": 15.62882096069869,
+ "grad_norm": 0.021711355075240135,
+ "learning_rate": 0.0006,
+ "loss": 4.657808303833008,
+ "step": 1125
+ },
+ {
+ "epoch": 15.642794759825328,
+ "grad_norm": 0.022810909897089005,
+ "learning_rate": 0.0006,
+ "loss": 4.6526689529418945,
+ "step": 1126
+ },
+ {
+ "epoch": 15.656768558951965,
+ "grad_norm": 0.022561343386769295,
+ "learning_rate": 0.0006,
+ "loss": 4.534887790679932,
+ "step": 1127
+ },
+ {
+ "epoch": 15.670742358078602,
+ "grad_norm": 0.019909802824258804,
+ "learning_rate": 0.0006,
+ "loss": 4.600790023803711,
+ "step": 1128
+ },
+ {
+ "epoch": 15.68471615720524,
+ "grad_norm": 0.020092325285077095,
+ "learning_rate": 0.0006,
+ "loss": 4.773830413818359,
+ "step": 1129
+ },
+ {
+ "epoch": 15.698689956331878,
+ "grad_norm": 0.018834445625543594,
+ "learning_rate": 0.0006,
+ "loss": 4.628868103027344,
+ "step": 1130
+ },
+ {
+ "epoch": 15.712663755458514,
+ "grad_norm": 0.017552798613905907,
+ "learning_rate": 0.0006,
+ "loss": 4.647500038146973,
+ "step": 1131
+ },
+ {
+ "epoch": 15.726637554585153,
+ "grad_norm": 0.01751251146197319,
+ "learning_rate": 0.0006,
+ "loss": 4.590545654296875,
+ "step": 1132
+ },
+ {
+ "epoch": 15.74061135371179,
+ "grad_norm": 0.01699099875986576,
+ "learning_rate": 0.0006,
+ "loss": 4.509699821472168,
+ "step": 1133
+ },
+ {
+ "epoch": 15.754585152838429,
+ "grad_norm": 0.015808649361133575,
+ "learning_rate": 0.0006,
+ "loss": 4.606992244720459,
+ "step": 1134
+ },
+ {
+ "epoch": 15.768558951965066,
+ "grad_norm": 0.01655443385243416,
+ "learning_rate": 0.0006,
+ "loss": 4.585219860076904,
+ "step": 1135
+ },
+ {
+ "epoch": 15.782532751091702,
+ "grad_norm": 0.01708163693547249,
+ "learning_rate": 0.0006,
+ "loss": 4.668147087097168,
+ "step": 1136
+ },
+ {
+ "epoch": 15.796506550218341,
+ "grad_norm": 0.017161913216114044,
+ "learning_rate": 0.0006,
+ "loss": 4.7170634269714355,
+ "step": 1137
+ },
+ {
+ "epoch": 15.810480349344978,
+ "grad_norm": 0.017262674868106842,
+ "learning_rate": 0.0006,
+ "loss": 4.572132110595703,
+ "step": 1138
+ },
+ {
+ "epoch": 15.824454148471617,
+ "grad_norm": 0.017713576555252075,
+ "learning_rate": 0.0006,
+ "loss": 4.702538967132568,
+ "step": 1139
+ },
+ {
+ "epoch": 15.838427947598253,
+ "grad_norm": 0.018730733543634415,
+ "learning_rate": 0.0006,
+ "loss": 4.629721641540527,
+ "step": 1140
+ },
+ {
+ "epoch": 15.85240174672489,
+ "grad_norm": 0.017960187047719955,
+ "learning_rate": 0.0006,
+ "loss": 4.661039352416992,
+ "step": 1141
+ },
+ {
+ "epoch": 15.866375545851529,
+ "grad_norm": 0.01750079356133938,
+ "learning_rate": 0.0006,
+ "loss": 4.503710746765137,
+ "step": 1142
+ },
+ {
+ "epoch": 15.880349344978166,
+ "grad_norm": 0.016630422323942184,
+ "learning_rate": 0.0006,
+ "loss": 4.6582136154174805,
+ "step": 1143
+ },
+ {
+ "epoch": 15.894323144104803,
+ "grad_norm": 0.015896832570433617,
+ "learning_rate": 0.0006,
+ "loss": 4.596693992614746,
+ "step": 1144
+ },
+ {
+ "epoch": 15.908296943231441,
+ "grad_norm": 0.01566407084465027,
+ "learning_rate": 0.0006,
+ "loss": 4.6337199211120605,
+ "step": 1145
+ },
+ {
+ "epoch": 15.922270742358078,
+ "grad_norm": 0.016823330894112587,
+ "learning_rate": 0.0006,
+ "loss": 4.686994552612305,
+ "step": 1146
+ },
+ {
+ "epoch": 15.936244541484717,
+ "grad_norm": 0.016208255663514137,
+ "learning_rate": 0.0006,
+ "loss": 4.6205549240112305,
+ "step": 1147
+ },
+ {
+ "epoch": 15.950218340611354,
+ "grad_norm": 0.0171707421541214,
+ "learning_rate": 0.0006,
+ "loss": 4.582968711853027,
+ "step": 1148
+ },
+ {
+ "epoch": 15.96419213973799,
+ "grad_norm": 0.017009446397423744,
+ "learning_rate": 0.0006,
+ "loss": 4.510274887084961,
+ "step": 1149
+ },
+ {
+ "epoch": 15.97816593886463,
+ "grad_norm": 0.01565811224281788,
+ "learning_rate": 0.0006,
+ "loss": 4.55834436416626,
+ "step": 1150
+ },
+ {
+ "epoch": 15.992139737991266,
+ "grad_norm": 0.015276037156581879,
+ "learning_rate": 0.0006,
+ "loss": 4.537870407104492,
+ "step": 1151
+ },
+ {
+ "epoch": 16.0,
+ "grad_norm": 0.017305882647633553,
+ "learning_rate": 0.0006,
+ "loss": 4.611305236816406,
+ "step": 1152
+ },
+ {
+ "epoch": 16.0,
+ "eval_loss": 4.775513172149658,
+ "eval_runtime": 56.7341,
+ "eval_samples_per_second": 43.043,
+ "eval_steps_per_second": 1.357,
+ "step": 1152
+ },
+ {
+ "epoch": 16.01397379912664,
+ "grad_norm": 0.016752436757087708,
+ "learning_rate": 0.0006,
+ "loss": 4.541312217712402,
+ "step": 1153
+ },
+ {
+ "epoch": 16.027947598253274,
+ "grad_norm": 0.01643134094774723,
+ "learning_rate": 0.0006,
+ "loss": 4.585402965545654,
+ "step": 1154
+ },
+ {
+ "epoch": 16.041921397379912,
+ "grad_norm": 0.01888720504939556,
+ "learning_rate": 0.0006,
+ "loss": 4.597825050354004,
+ "step": 1155
+ },
+ {
+ "epoch": 16.05589519650655,
+ "grad_norm": 0.022044632583856583,
+ "learning_rate": 0.0006,
+ "loss": 4.532772064208984,
+ "step": 1156
+ },
+ {
+ "epoch": 16.069868995633186,
+ "grad_norm": 0.022986987605690956,
+ "learning_rate": 0.0006,
+ "loss": 4.551351547241211,
+ "step": 1157
+ },
+ {
+ "epoch": 16.083842794759825,
+ "grad_norm": 0.020116791129112244,
+ "learning_rate": 0.0006,
+ "loss": 4.484149932861328,
+ "step": 1158
+ },
+ {
+ "epoch": 16.097816593886463,
+ "grad_norm": 0.021442247554659843,
+ "learning_rate": 0.0006,
+ "loss": 4.515316486358643,
+ "step": 1159
+ },
+ {
+ "epoch": 16.111790393013102,
+ "grad_norm": 0.020974772050976753,
+ "learning_rate": 0.0006,
+ "loss": 4.51606559753418,
+ "step": 1160
+ },
+ {
+ "epoch": 16.125764192139737,
+ "grad_norm": 0.021274864673614502,
+ "learning_rate": 0.0006,
+ "loss": 4.587329864501953,
+ "step": 1161
+ },
+ {
+ "epoch": 16.139737991266376,
+ "grad_norm": 0.022799784317612648,
+ "learning_rate": 0.0006,
+ "loss": 4.568149089813232,
+ "step": 1162
+ },
+ {
+ "epoch": 16.153711790393015,
+ "grad_norm": 0.021784933283925056,
+ "learning_rate": 0.0006,
+ "loss": 4.515336990356445,
+ "step": 1163
+ },
+ {
+ "epoch": 16.16768558951965,
+ "grad_norm": 0.024261271581053734,
+ "learning_rate": 0.0006,
+ "loss": 4.5127081871032715,
+ "step": 1164
+ },
+ {
+ "epoch": 16.18165938864629,
+ "grad_norm": 0.02248375490307808,
+ "learning_rate": 0.0006,
+ "loss": 4.754251003265381,
+ "step": 1165
+ },
+ {
+ "epoch": 16.195633187772927,
+ "grad_norm": 0.02136492356657982,
+ "learning_rate": 0.0006,
+ "loss": 4.58193826675415,
+ "step": 1166
+ },
+ {
+ "epoch": 16.209606986899562,
+ "grad_norm": 0.020727066323161125,
+ "learning_rate": 0.0006,
+ "loss": 4.564209938049316,
+ "step": 1167
+ },
+ {
+ "epoch": 16.2235807860262,
+ "grad_norm": 0.017789119854569435,
+ "learning_rate": 0.0006,
+ "loss": 4.698680877685547,
+ "step": 1168
+ },
+ {
+ "epoch": 16.23755458515284,
+ "grad_norm": 0.017603034153580666,
+ "learning_rate": 0.0006,
+ "loss": 4.501171588897705,
+ "step": 1169
+ },
+ {
+ "epoch": 16.251528384279474,
+ "grad_norm": 0.017345862463116646,
+ "learning_rate": 0.0006,
+ "loss": 4.593590259552002,
+ "step": 1170
+ },
+ {
+ "epoch": 16.265502183406113,
+ "grad_norm": 0.018047425895929337,
+ "learning_rate": 0.0006,
+ "loss": 4.597467422485352,
+ "step": 1171
+ },
+ {
+ "epoch": 16.27947598253275,
+ "grad_norm": 0.01947222836315632,
+ "learning_rate": 0.0006,
+ "loss": 4.5072221755981445,
+ "step": 1172
+ },
+ {
+ "epoch": 16.29344978165939,
+ "grad_norm": 0.019598500803112984,
+ "learning_rate": 0.0006,
+ "loss": 4.601351737976074,
+ "step": 1173
+ },
+ {
+ "epoch": 16.307423580786025,
+ "grad_norm": 0.020002514123916626,
+ "learning_rate": 0.0006,
+ "loss": 4.580431938171387,
+ "step": 1174
+ },
+ {
+ "epoch": 16.321397379912664,
+ "grad_norm": 0.0216530729085207,
+ "learning_rate": 0.0006,
+ "loss": 4.424681663513184,
+ "step": 1175
+ },
+ {
+ "epoch": 16.335371179039303,
+ "grad_norm": 0.01949235238134861,
+ "learning_rate": 0.0006,
+ "loss": 4.751883029937744,
+ "step": 1176
+ },
+ {
+ "epoch": 16.349344978165938,
+ "grad_norm": 0.017779016867280006,
+ "learning_rate": 0.0006,
+ "loss": 4.541490077972412,
+ "step": 1177
+ },
+ {
+ "epoch": 16.363318777292577,
+ "grad_norm": 0.01909121684730053,
+ "learning_rate": 0.0006,
+ "loss": 4.594141006469727,
+ "step": 1178
+ },
+ {
+ "epoch": 16.377292576419215,
+ "grad_norm": 0.019894849509000778,
+ "learning_rate": 0.0006,
+ "loss": 4.555070400238037,
+ "step": 1179
+ },
+ {
+ "epoch": 16.39126637554585,
+ "grad_norm": 0.01993032731115818,
+ "learning_rate": 0.0006,
+ "loss": 4.521401405334473,
+ "step": 1180
+ },
+ {
+ "epoch": 16.40524017467249,
+ "grad_norm": 0.0210788045078516,
+ "learning_rate": 0.0006,
+ "loss": 4.672842979431152,
+ "step": 1181
+ },
+ {
+ "epoch": 16.419213973799128,
+ "grad_norm": 0.02274598926305771,
+ "learning_rate": 0.0006,
+ "loss": 4.629504203796387,
+ "step": 1182
+ },
+ {
+ "epoch": 16.433187772925763,
+ "grad_norm": 0.020448120310902596,
+ "learning_rate": 0.0006,
+ "loss": 4.548209190368652,
+ "step": 1183
+ },
+ {
+ "epoch": 16.4471615720524,
+ "grad_norm": 0.017565643414855003,
+ "learning_rate": 0.0006,
+ "loss": 4.529054641723633,
+ "step": 1184
+ },
+ {
+ "epoch": 16.46113537117904,
+ "grad_norm": 0.019504237920045853,
+ "learning_rate": 0.0006,
+ "loss": 4.48182487487793,
+ "step": 1185
+ },
+ {
+ "epoch": 16.475109170305675,
+ "grad_norm": 0.016823742538690567,
+ "learning_rate": 0.0006,
+ "loss": 4.608696937561035,
+ "step": 1186
+ },
+ {
+ "epoch": 16.489082969432314,
+ "grad_norm": 0.01602894812822342,
+ "learning_rate": 0.0006,
+ "loss": 4.620062351226807,
+ "step": 1187
+ },
+ {
+ "epoch": 16.503056768558952,
+ "grad_norm": 0.016615642234683037,
+ "learning_rate": 0.0006,
+ "loss": 4.69378662109375,
+ "step": 1188
+ },
+ {
+ "epoch": 16.51703056768559,
+ "grad_norm": 0.018576730042696,
+ "learning_rate": 0.0006,
+ "loss": 4.537041664123535,
+ "step": 1189
+ },
+ {
+ "epoch": 16.531004366812226,
+ "grad_norm": 0.018611151725053787,
+ "learning_rate": 0.0006,
+ "loss": 4.496169090270996,
+ "step": 1190
+ },
+ {
+ "epoch": 16.544978165938865,
+ "grad_norm": 0.018622413277626038,
+ "learning_rate": 0.0006,
+ "loss": 4.557358264923096,
+ "step": 1191
+ },
+ {
+ "epoch": 16.558951965065503,
+ "grad_norm": 0.01823544315993786,
+ "learning_rate": 0.0006,
+ "loss": 4.730792999267578,
+ "step": 1192
+ },
+ {
+ "epoch": 16.57292576419214,
+ "grad_norm": 0.018796846270561218,
+ "learning_rate": 0.0006,
+ "loss": 4.781791687011719,
+ "step": 1193
+ },
+ {
+ "epoch": 16.586899563318777,
+ "grad_norm": 0.01728292554616928,
+ "learning_rate": 0.0006,
+ "loss": 4.515257835388184,
+ "step": 1194
+ },
+ {
+ "epoch": 16.600873362445416,
+ "grad_norm": 0.01699589006602764,
+ "learning_rate": 0.0006,
+ "loss": 4.5996904373168945,
+ "step": 1195
+ },
+ {
+ "epoch": 16.61484716157205,
+ "grad_norm": 0.015905970707535744,
+ "learning_rate": 0.0006,
+ "loss": 4.601301193237305,
+ "step": 1196
+ },
+ {
+ "epoch": 16.62882096069869,
+ "grad_norm": 0.01338571310043335,
+ "learning_rate": 0.0006,
+ "loss": 4.557708740234375,
+ "step": 1197
+ },
+ {
+ "epoch": 16.64279475982533,
+ "grad_norm": 0.015298855490982533,
+ "learning_rate": 0.0006,
+ "loss": 4.542287349700928,
+ "step": 1198
+ },
+ {
+ "epoch": 16.656768558951963,
+ "grad_norm": 0.014753367751836777,
+ "learning_rate": 0.0006,
+ "loss": 4.595787048339844,
+ "step": 1199
+ },
+ {
+ "epoch": 16.670742358078602,
+ "grad_norm": 0.015527483075857162,
+ "learning_rate": 0.0006,
+ "loss": 4.607529163360596,
+ "step": 1200
+ },
+ {
+ "epoch": 16.68471615720524,
+ "grad_norm": 0.015042162500321865,
+ "learning_rate": 0.0006,
+ "loss": 4.686626434326172,
+ "step": 1201
+ },
+ {
+ "epoch": 16.69868995633188,
+ "grad_norm": 0.0159393772482872,
+ "learning_rate": 0.0006,
+ "loss": 4.47656774520874,
+ "step": 1202
+ },
+ {
+ "epoch": 16.712663755458514,
+ "grad_norm": 0.01532787922769785,
+ "learning_rate": 0.0006,
+ "loss": 4.55678653717041,
+ "step": 1203
+ },
+ {
+ "epoch": 16.726637554585153,
+ "grad_norm": 0.016332164406776428,
+ "learning_rate": 0.0006,
+ "loss": 4.51441764831543,
+ "step": 1204
+ },
+ {
+ "epoch": 16.74061135371179,
+ "grad_norm": 0.018901726230978966,
+ "learning_rate": 0.0006,
+ "loss": 4.57227897644043,
+ "step": 1205
+ },
+ {
+ "epoch": 16.754585152838427,
+ "grad_norm": 0.018589433282613754,
+ "learning_rate": 0.0006,
+ "loss": 4.550070762634277,
+ "step": 1206
+ },
+ {
+ "epoch": 16.768558951965066,
+ "grad_norm": 0.01634034886956215,
+ "learning_rate": 0.0006,
+ "loss": 4.586910724639893,
+ "step": 1207
+ },
+ {
+ "epoch": 16.782532751091704,
+ "grad_norm": 0.017599433660507202,
+ "learning_rate": 0.0006,
+ "loss": 4.5891923904418945,
+ "step": 1208
+ },
+ {
+ "epoch": 16.79650655021834,
+ "grad_norm": 0.01878446340560913,
+ "learning_rate": 0.0006,
+ "loss": 4.57227087020874,
+ "step": 1209
+ },
+ {
+ "epoch": 16.810480349344978,
+ "grad_norm": 0.01803469844162464,
+ "learning_rate": 0.0006,
+ "loss": 4.583845138549805,
+ "step": 1210
+ },
+ {
+ "epoch": 16.824454148471617,
+ "grad_norm": 0.018505144864320755,
+ "learning_rate": 0.0006,
+ "loss": 4.561151504516602,
+ "step": 1211
+ },
+ {
+ "epoch": 16.83842794759825,
+ "grad_norm": 0.01920081488788128,
+ "learning_rate": 0.0006,
+ "loss": 4.506834030151367,
+ "step": 1212
+ },
+ {
+ "epoch": 16.85240174672489,
+ "grad_norm": 0.0177712831646204,
+ "learning_rate": 0.0006,
+ "loss": 4.557195663452148,
+ "step": 1213
+ },
+ {
+ "epoch": 16.86637554585153,
+ "grad_norm": 0.019113918766379356,
+ "learning_rate": 0.0006,
+ "loss": 4.651227951049805,
+ "step": 1214
+ },
+ {
+ "epoch": 16.880349344978168,
+ "grad_norm": 0.02012539468705654,
+ "learning_rate": 0.0006,
+ "loss": 4.4952239990234375,
+ "step": 1215
+ },
+ {
+ "epoch": 16.894323144104803,
+ "grad_norm": 0.019993646070361137,
+ "learning_rate": 0.0006,
+ "loss": 4.517433166503906,
+ "step": 1216
+ },
+ {
+ "epoch": 16.90829694323144,
+ "grad_norm": 0.020041441544890404,
+ "learning_rate": 0.0006,
+ "loss": 4.489407539367676,
+ "step": 1217
+ },
+ {
+ "epoch": 16.92227074235808,
+ "grad_norm": 0.0183611661195755,
+ "learning_rate": 0.0006,
+ "loss": 4.5435638427734375,
+ "step": 1218
+ },
+ {
+ "epoch": 16.936244541484715,
+ "grad_norm": 0.019138429313898087,
+ "learning_rate": 0.0006,
+ "loss": 4.545949459075928,
+ "step": 1219
+ },
+ {
+ "epoch": 16.950218340611354,
+ "grad_norm": 0.018554266542196274,
+ "learning_rate": 0.0006,
+ "loss": 4.579620361328125,
+ "step": 1220
+ },
+ {
+ "epoch": 16.964192139737992,
+ "grad_norm": 0.0193545650690794,
+ "learning_rate": 0.0006,
+ "loss": 4.588255405426025,
+ "step": 1221
+ },
+ {
+ "epoch": 16.978165938864628,
+ "grad_norm": 0.019389165565371513,
+ "learning_rate": 0.0006,
+ "loss": 4.520698547363281,
+ "step": 1222
+ },
+ {
+ "epoch": 16.992139737991266,
+ "grad_norm": 0.016601379960775375,
+ "learning_rate": 0.0006,
+ "loss": 4.4809064865112305,
+ "step": 1223
+ },
+ {
+ "epoch": 17.0,
+ "grad_norm": 0.01771303080022335,
+ "learning_rate": 0.0006,
+ "loss": 4.721122741699219,
+ "step": 1224
+ },
+ {
+ "epoch": 17.0,
+ "eval_loss": 4.7903666496276855,
+ "eval_runtime": 56.9826,
+ "eval_samples_per_second": 42.855,
+ "eval_steps_per_second": 1.351,
+ "step": 1224
+ },
+ {
+ "epoch": 17.01397379912664,
+ "grad_norm": 0.017720289528369904,
+ "learning_rate": 0.0006,
+ "loss": 4.523834705352783,
+ "step": 1225
+ },
+ {
+ "epoch": 17.027947598253274,
+ "grad_norm": 0.020137161016464233,
+ "learning_rate": 0.0006,
+ "loss": 4.574061870574951,
+ "step": 1226
+ },
+ {
+ "epoch": 17.041921397379912,
+ "grad_norm": 0.02002314105629921,
+ "learning_rate": 0.0006,
+ "loss": 4.512407302856445,
+ "step": 1227
+ },
+ {
+ "epoch": 17.05589519650655,
+ "grad_norm": 0.019628183916211128,
+ "learning_rate": 0.0006,
+ "loss": 4.502533435821533,
+ "step": 1228
+ },
+ {
+ "epoch": 17.069868995633186,
+ "grad_norm": 0.019808458164334297,
+ "learning_rate": 0.0006,
+ "loss": 4.463156700134277,
+ "step": 1229
+ },
+ {
+ "epoch": 17.083842794759825,
+ "grad_norm": 0.021250786259770393,
+ "learning_rate": 0.0006,
+ "loss": 4.622682094573975,
+ "step": 1230
+ },
+ {
+ "epoch": 17.097816593886463,
+ "grad_norm": 0.022031694650650024,
+ "learning_rate": 0.0006,
+ "loss": 4.451882839202881,
+ "step": 1231
+ },
+ {
+ "epoch": 17.111790393013102,
+ "grad_norm": 0.01898987777531147,
+ "learning_rate": 0.0006,
+ "loss": 4.529331684112549,
+ "step": 1232
+ },
+ {
+ "epoch": 17.125764192139737,
+ "grad_norm": 0.018749834969639778,
+ "learning_rate": 0.0006,
+ "loss": 4.435815334320068,
+ "step": 1233
+ },
+ {
+ "epoch": 17.139737991266376,
+ "grad_norm": 0.01896674931049347,
+ "learning_rate": 0.0006,
+ "loss": 4.484029293060303,
+ "step": 1234
+ },
+ {
+ "epoch": 17.153711790393015,
+ "grad_norm": 0.018286997452378273,
+ "learning_rate": 0.0006,
+ "loss": 4.523510932922363,
+ "step": 1235
+ },
+ {
+ "epoch": 17.16768558951965,
+ "grad_norm": 0.017273059114813805,
+ "learning_rate": 0.0006,
+ "loss": 4.451767444610596,
+ "step": 1236
+ },
+ {
+ "epoch": 17.18165938864629,
+ "grad_norm": 0.016499150544404984,
+ "learning_rate": 0.0006,
+ "loss": 4.5568647384643555,
+ "step": 1237
+ },
+ {
+ "epoch": 17.195633187772927,
+ "grad_norm": 0.015934469178318977,
+ "learning_rate": 0.0006,
+ "loss": 4.4639997482299805,
+ "step": 1238
+ },
+ {
+ "epoch": 17.209606986899562,
+ "grad_norm": 0.01714209094643593,
+ "learning_rate": 0.0006,
+ "loss": 4.547614097595215,
+ "step": 1239
+ },
+ {
+ "epoch": 17.2235807860262,
+ "grad_norm": 0.019167495891451836,
+ "learning_rate": 0.0006,
+ "loss": 4.571097373962402,
+ "step": 1240
+ },
+ {
+ "epoch": 17.23755458515284,
+ "grad_norm": 0.018276996910572052,
+ "learning_rate": 0.0006,
+ "loss": 4.489764213562012,
+ "step": 1241
+ },
+ {
+ "epoch": 17.251528384279474,
+ "grad_norm": 0.021296866238117218,
+ "learning_rate": 0.0006,
+ "loss": 4.526913642883301,
+ "step": 1242
+ },
+ {
+ "epoch": 17.265502183406113,
+ "grad_norm": 0.022606629878282547,
+ "learning_rate": 0.0006,
+ "loss": 4.521624565124512,
+ "step": 1243
+ },
+ {
+ "epoch": 17.27947598253275,
+ "grad_norm": 0.021204711869359016,
+ "learning_rate": 0.0006,
+ "loss": 4.500864028930664,
+ "step": 1244
+ },
+ {
+ "epoch": 17.29344978165939,
+ "grad_norm": 0.01988849975168705,
+ "learning_rate": 0.0006,
+ "loss": 4.45162296295166,
+ "step": 1245
+ },
+ {
+ "epoch": 17.307423580786025,
+ "grad_norm": 0.020268384367227554,
+ "learning_rate": 0.0006,
+ "loss": 4.536198616027832,
+ "step": 1246
+ },
+ {
+ "epoch": 17.321397379912664,
+ "grad_norm": 0.020185677334666252,
+ "learning_rate": 0.0006,
+ "loss": 4.440262794494629,
+ "step": 1247
+ },
+ {
+ "epoch": 17.335371179039303,
+ "grad_norm": 0.020106853917241096,
+ "learning_rate": 0.0006,
+ "loss": 4.475907325744629,
+ "step": 1248
+ },
+ {
+ "epoch": 17.349344978165938,
+ "grad_norm": 0.018883297219872475,
+ "learning_rate": 0.0006,
+ "loss": 4.640661716461182,
+ "step": 1249
+ },
+ {
+ "epoch": 17.363318777292577,
+ "grad_norm": 0.01852620206773281,
+ "learning_rate": 0.0006,
+ "loss": 4.578243255615234,
+ "step": 1250
+ },
+ {
+ "epoch": 17.377292576419215,
+ "grad_norm": 0.01886451430618763,
+ "learning_rate": 0.0006,
+ "loss": 4.546026229858398,
+ "step": 1251
+ },
+ {
+ "epoch": 17.39126637554585,
+ "grad_norm": 0.01777452416718006,
+ "learning_rate": 0.0006,
+ "loss": 4.521542549133301,
+ "step": 1252
+ },
+ {
+ "epoch": 17.40524017467249,
+ "grad_norm": 0.016626447439193726,
+ "learning_rate": 0.0006,
+ "loss": 4.52629280090332,
+ "step": 1253
+ },
+ {
+ "epoch": 17.419213973799128,
+ "grad_norm": 0.01599477231502533,
+ "learning_rate": 0.0006,
+ "loss": 4.4796037673950195,
+ "step": 1254
+ },
+ {
+ "epoch": 17.433187772925763,
+ "grad_norm": 0.017021458595991135,
+ "learning_rate": 0.0006,
+ "loss": 4.468813419342041,
+ "step": 1255
+ },
+ {
+ "epoch": 17.4471615720524,
+ "grad_norm": 0.016925333067774773,
+ "learning_rate": 0.0006,
+ "loss": 4.584898948669434,
+ "step": 1256
+ },
+ {
+ "epoch": 17.46113537117904,
+ "grad_norm": 0.01714998669922352,
+ "learning_rate": 0.0006,
+ "loss": 4.523544788360596,
+ "step": 1257
+ },
+ {
+ "epoch": 17.475109170305675,
+ "grad_norm": 0.01802288182079792,
+ "learning_rate": 0.0006,
+ "loss": 4.595581531524658,
+ "step": 1258
+ },
+ {
+ "epoch": 17.489082969432314,
+ "grad_norm": 0.0180289875715971,
+ "learning_rate": 0.0006,
+ "loss": 4.394440650939941,
+ "step": 1259
+ },
+ {
+ "epoch": 17.503056768558952,
+ "grad_norm": 0.018504245206713676,
+ "learning_rate": 0.0006,
+ "loss": 4.4794416427612305,
+ "step": 1260
+ },
+ {
+ "epoch": 17.51703056768559,
+ "grad_norm": 0.019973233342170715,
+ "learning_rate": 0.0006,
+ "loss": 4.4241437911987305,
+ "step": 1261
+ },
+ {
+ "epoch": 17.531004366812226,
+ "grad_norm": 0.020063083618879318,
+ "learning_rate": 0.0006,
+ "loss": 4.48068380355835,
+ "step": 1262
+ },
+ {
+ "epoch": 17.544978165938865,
+ "grad_norm": 0.02066640742123127,
+ "learning_rate": 0.0006,
+ "loss": 4.628993034362793,
+ "step": 1263
+ },
+ {
+ "epoch": 17.558951965065503,
+ "grad_norm": 0.02206416428089142,
+ "learning_rate": 0.0006,
+ "loss": 4.564790725708008,
+ "step": 1264
+ },
+ {
+ "epoch": 17.57292576419214,
+ "grad_norm": 0.020640335977077484,
+ "learning_rate": 0.0006,
+ "loss": 4.578705787658691,
+ "step": 1265
+ },
+ {
+ "epoch": 17.586899563318777,
+ "grad_norm": 0.019093405455350876,
+ "learning_rate": 0.0006,
+ "loss": 4.5023603439331055,
+ "step": 1266
+ },
+ {
+ "epoch": 17.600873362445416,
+ "grad_norm": 0.019459193572402,
+ "learning_rate": 0.0006,
+ "loss": 4.594531059265137,
+ "step": 1267
+ },
+ {
+ "epoch": 17.61484716157205,
+ "grad_norm": 0.02021171525120735,
+ "learning_rate": 0.0006,
+ "loss": 4.770299911499023,
+ "step": 1268
+ },
+ {
+ "epoch": 17.62882096069869,
+ "grad_norm": 0.021161191165447235,
+ "learning_rate": 0.0006,
+ "loss": 4.5380401611328125,
+ "step": 1269
+ },
+ {
+ "epoch": 17.64279475982533,
+ "grad_norm": 0.021013403311371803,
+ "learning_rate": 0.0006,
+ "loss": 4.455524921417236,
+ "step": 1270
+ },
+ {
+ "epoch": 17.656768558951963,
+ "grad_norm": 0.020336592569947243,
+ "learning_rate": 0.0006,
+ "loss": 4.533672332763672,
+ "step": 1271
+ },
+ {
+ "epoch": 17.670742358078602,
+ "grad_norm": 0.020159434527158737,
+ "learning_rate": 0.0006,
+ "loss": 4.596151351928711,
+ "step": 1272
+ },
+ {
+ "epoch": 17.68471615720524,
+ "grad_norm": 0.020432960242033005,
+ "learning_rate": 0.0006,
+ "loss": 4.55531644821167,
+ "step": 1273
+ },
+ {
+ "epoch": 17.69868995633188,
+ "grad_norm": 0.025466229766607285,
+ "learning_rate": 0.0006,
+ "loss": 4.566912651062012,
+ "step": 1274
+ },
+ {
+ "epoch": 17.712663755458514,
+ "grad_norm": 0.026537558063864708,
+ "learning_rate": 0.0006,
+ "loss": 4.526003837585449,
+ "step": 1275
+ },
+ {
+ "epoch": 17.726637554585153,
+ "grad_norm": 0.024397607892751694,
+ "learning_rate": 0.0006,
+ "loss": 4.541755676269531,
+ "step": 1276
+ },
+ {
+ "epoch": 17.74061135371179,
+ "grad_norm": 0.022062312811613083,
+ "learning_rate": 0.0006,
+ "loss": 4.6532793045043945,
+ "step": 1277
+ },
+ {
+ "epoch": 17.754585152838427,
+ "grad_norm": 0.02204282023012638,
+ "learning_rate": 0.0006,
+ "loss": 4.582825660705566,
+ "step": 1278
+ },
+ {
+ "epoch": 17.768558951965066,
+ "grad_norm": 0.020844416692852974,
+ "learning_rate": 0.0006,
+ "loss": 4.522336483001709,
+ "step": 1279
+ },
+ {
+ "epoch": 17.782532751091704,
+ "grad_norm": 0.019815703853964806,
+ "learning_rate": 0.0006,
+ "loss": 4.607398509979248,
+ "step": 1280
+ },
+ {
+ "epoch": 17.79650655021834,
+ "grad_norm": 0.016676954925060272,
+ "learning_rate": 0.0006,
+ "loss": 4.446206092834473,
+ "step": 1281
+ },
+ {
+ "epoch": 17.810480349344978,
+ "grad_norm": 0.0166452769190073,
+ "learning_rate": 0.0006,
+ "loss": 4.505931854248047,
+ "step": 1282
+ },
+ {
+ "epoch": 17.824454148471617,
+ "grad_norm": 0.015928907319903374,
+ "learning_rate": 0.0006,
+ "loss": 4.455267429351807,
+ "step": 1283
+ },
+ {
+ "epoch": 17.83842794759825,
+ "grad_norm": 0.015219560824334621,
+ "learning_rate": 0.0006,
+ "loss": 4.5623674392700195,
+ "step": 1284
+ },
+ {
+ "epoch": 17.85240174672489,
+ "grad_norm": 0.015125355683267117,
+ "learning_rate": 0.0006,
+ "loss": 4.554104804992676,
+ "step": 1285
+ },
+ {
+ "epoch": 17.86637554585153,
+ "grad_norm": 0.015612718649208546,
+ "learning_rate": 0.0006,
+ "loss": 4.484847068786621,
+ "step": 1286
+ },
+ {
+ "epoch": 17.880349344978168,
+ "grad_norm": 0.015215662308037281,
+ "learning_rate": 0.0006,
+ "loss": 4.4802093505859375,
+ "step": 1287
+ },
+ {
+ "epoch": 17.894323144104803,
+ "grad_norm": 0.01511684525758028,
+ "learning_rate": 0.0006,
+ "loss": 4.537298679351807,
+ "step": 1288
+ },
+ {
+ "epoch": 17.90829694323144,
+ "grad_norm": 0.013841322623193264,
+ "learning_rate": 0.0006,
+ "loss": 4.427488327026367,
+ "step": 1289
+ },
+ {
+ "epoch": 17.92227074235808,
+ "grad_norm": 0.015115964226424694,
+ "learning_rate": 0.0006,
+ "loss": 4.4547119140625,
+ "step": 1290
+ },
+ {
+ "epoch": 17.936244541484715,
+ "grad_norm": 0.016411006450653076,
+ "learning_rate": 0.0006,
+ "loss": 4.52994966506958,
+ "step": 1291
+ },
+ {
+ "epoch": 17.950218340611354,
+ "grad_norm": 0.017313152551651,
+ "learning_rate": 0.0006,
+ "loss": 4.5337653160095215,
+ "step": 1292
+ },
+ {
+ "epoch": 17.964192139737992,
+ "grad_norm": 0.016559796407818794,
+ "learning_rate": 0.0006,
+ "loss": 4.5277910232543945,
+ "step": 1293
+ },
+ {
+ "epoch": 17.978165938864628,
+ "grad_norm": 0.015854761004447937,
+ "learning_rate": 0.0006,
+ "loss": 4.6334309577941895,
+ "step": 1294
+ },
+ {
+ "epoch": 17.992139737991266,
+ "grad_norm": 0.015658238902688026,
+ "learning_rate": 0.0006,
+ "loss": 4.613372802734375,
+ "step": 1295
+ },
+ {
+ "epoch": 18.0,
+ "grad_norm": 0.0169155802577734,
+ "learning_rate": 0.0006,
+ "loss": 4.527356147766113,
+ "step": 1296
+ },
+ {
+ "epoch": 18.0,
+ "eval_loss": 4.764543056488037,
+ "eval_runtime": 56.5998,
+ "eval_samples_per_second": 43.145,
+ "eval_steps_per_second": 1.36,
+ "step": 1296
+ },
+ {
+ "epoch": 18.01397379912664,
+ "grad_norm": 0.015348963439464569,
+ "learning_rate": 0.0006,
+ "loss": 4.466423988342285,
+ "step": 1297
+ },
+ {
+ "epoch": 18.027947598253274,
+ "grad_norm": 0.01768588088452816,
+ "learning_rate": 0.0006,
+ "loss": 4.545933246612549,
+ "step": 1298
+ },
+ {
+ "epoch": 18.041921397379912,
+ "grad_norm": 0.017381226643919945,
+ "learning_rate": 0.0006,
+ "loss": 4.494060039520264,
+ "step": 1299
+ },
+ {
+ "epoch": 18.05589519650655,
+ "grad_norm": 0.019002335146069527,
+ "learning_rate": 0.0006,
+ "loss": 4.395200729370117,
+ "step": 1300
+ },
+ {
+ "epoch": 18.069868995633186,
+ "grad_norm": 0.01827932707965374,
+ "learning_rate": 0.0006,
+ "loss": 4.424848556518555,
+ "step": 1301
+ },
+ {
+ "epoch": 18.083842794759825,
+ "grad_norm": 0.016383031383156776,
+ "learning_rate": 0.0006,
+ "loss": 4.494722366333008,
+ "step": 1302
+ },
+ {
+ "epoch": 18.097816593886463,
+ "grad_norm": 0.016265157610177994,
+ "learning_rate": 0.0006,
+ "loss": 4.416522979736328,
+ "step": 1303
+ },
+ {
+ "epoch": 18.111790393013102,
+ "grad_norm": 0.01702386513352394,
+ "learning_rate": 0.0006,
+ "loss": 4.600103378295898,
+ "step": 1304
+ },
+ {
+ "epoch": 18.125764192139737,
+ "grad_norm": 0.018495498225092888,
+ "learning_rate": 0.0006,
+ "loss": 4.588401794433594,
+ "step": 1305
+ },
+ {
+ "epoch": 18.139737991266376,
+ "grad_norm": 0.017921684309840202,
+ "learning_rate": 0.0006,
+ "loss": 4.478075981140137,
+ "step": 1306
+ },
+ {
+ "epoch": 18.153711790393015,
+ "grad_norm": 0.01801278069615364,
+ "learning_rate": 0.0006,
+ "loss": 4.413186550140381,
+ "step": 1307
+ },
+ {
+ "epoch": 18.16768558951965,
+ "grad_norm": 0.017211275175213814,
+ "learning_rate": 0.0006,
+ "loss": 4.461341857910156,
+ "step": 1308
+ },
+ {
+ "epoch": 18.18165938864629,
+ "grad_norm": 0.017589328810572624,
+ "learning_rate": 0.0006,
+ "loss": 4.546969413757324,
+ "step": 1309
+ },
+ {
+ "epoch": 18.195633187772927,
+ "grad_norm": 0.01797698810696602,
+ "learning_rate": 0.0006,
+ "loss": 4.530797004699707,
+ "step": 1310
+ },
+ {
+ "epoch": 18.209606986899562,
+ "grad_norm": 0.020408619195222855,
+ "learning_rate": 0.0006,
+ "loss": 4.62999153137207,
+ "step": 1311
+ },
+ {
+ "epoch": 18.2235807860262,
+ "grad_norm": 0.01994038000702858,
+ "learning_rate": 0.0006,
+ "loss": 4.586811065673828,
+ "step": 1312
+ },
+ {
+ "epoch": 18.23755458515284,
+ "grad_norm": 0.019381340593099594,
+ "learning_rate": 0.0006,
+ "loss": 4.4769368171691895,
+ "step": 1313
+ },
+ {
+ "epoch": 18.251528384279474,
+ "grad_norm": 0.020484555512666702,
+ "learning_rate": 0.0006,
+ "loss": 4.524302959442139,
+ "step": 1314
+ },
+ {
+ "epoch": 18.265502183406113,
+ "grad_norm": 0.02426416426897049,
+ "learning_rate": 0.0006,
+ "loss": 4.561705589294434,
+ "step": 1315
+ },
+ {
+ "epoch": 18.27947598253275,
+ "grad_norm": 0.024748101830482483,
+ "learning_rate": 0.0006,
+ "loss": 4.4654316902160645,
+ "step": 1316
+ },
+ {
+ "epoch": 18.29344978165939,
+ "grad_norm": 0.020703328773379326,
+ "learning_rate": 0.0006,
+ "loss": 4.53033447265625,
+ "step": 1317
+ },
+ {
+ "epoch": 18.307423580786025,
+ "grad_norm": 0.01765458658337593,
+ "learning_rate": 0.0006,
+ "loss": 4.5154266357421875,
+ "step": 1318
+ },
+ {
+ "epoch": 18.321397379912664,
+ "grad_norm": 0.018646420910954475,
+ "learning_rate": 0.0006,
+ "loss": 4.5224714279174805,
+ "step": 1319
+ },
+ {
+ "epoch": 18.335371179039303,
+ "grad_norm": 0.01907474733889103,
+ "learning_rate": 0.0006,
+ "loss": 4.5055437088012695,
+ "step": 1320
+ },
+ {
+ "epoch": 18.349344978165938,
+ "grad_norm": 0.018179846927523613,
+ "learning_rate": 0.0006,
+ "loss": 4.462161064147949,
+ "step": 1321
+ },
+ {
+ "epoch": 18.363318777292577,
+ "grad_norm": 0.016697878018021584,
+ "learning_rate": 0.0006,
+ "loss": 4.473718643188477,
+ "step": 1322
+ },
+ {
+ "epoch": 18.377292576419215,
+ "grad_norm": 0.017504019662737846,
+ "learning_rate": 0.0006,
+ "loss": 4.474348068237305,
+ "step": 1323
+ },
+ {
+ "epoch": 18.39126637554585,
+ "grad_norm": 0.018889913335442543,
+ "learning_rate": 0.0006,
+ "loss": 4.4531636238098145,
+ "step": 1324
+ },
+ {
+ "epoch": 18.40524017467249,
+ "grad_norm": 0.019167251884937286,
+ "learning_rate": 0.0006,
+ "loss": 4.4278364181518555,
+ "step": 1325
+ },
+ {
+ "epoch": 18.419213973799128,
+ "grad_norm": 0.015895741060376167,
+ "learning_rate": 0.0006,
+ "loss": 4.474592208862305,
+ "step": 1326
+ },
+ {
+ "epoch": 18.433187772925763,
+ "grad_norm": 0.01531882956624031,
+ "learning_rate": 0.0006,
+ "loss": 4.403757572174072,
+ "step": 1327
+ },
+ {
+ "epoch": 18.4471615720524,
+ "grad_norm": 0.01565849967300892,
+ "learning_rate": 0.0006,
+ "loss": 4.507664680480957,
+ "step": 1328
+ },
+ {
+ "epoch": 18.46113537117904,
+ "grad_norm": 0.01602315716445446,
+ "learning_rate": 0.0006,
+ "loss": 4.5648908615112305,
+ "step": 1329
+ },
+ {
+ "epoch": 18.475109170305675,
+ "grad_norm": 0.016467366367578506,
+ "learning_rate": 0.0006,
+ "loss": 4.428762912750244,
+ "step": 1330
+ },
+ {
+ "epoch": 18.489082969432314,
+ "grad_norm": 0.01604127697646618,
+ "learning_rate": 0.0006,
+ "loss": 4.527771472930908,
+ "step": 1331
+ },
+ {
+ "epoch": 18.503056768558952,
+ "grad_norm": 0.018825586885213852,
+ "learning_rate": 0.0006,
+ "loss": 4.485386848449707,
+ "step": 1332
+ },
+ {
+ "epoch": 18.51703056768559,
+ "grad_norm": 0.019258007407188416,
+ "learning_rate": 0.0006,
+ "loss": 4.414972305297852,
+ "step": 1333
+ },
+ {
+ "epoch": 18.531004366812226,
+ "grad_norm": 0.018192021176218987,
+ "learning_rate": 0.0006,
+ "loss": 4.412166595458984,
+ "step": 1334
+ },
+ {
+ "epoch": 18.544978165938865,
+ "grad_norm": 0.016120698302984238,
+ "learning_rate": 0.0006,
+ "loss": 4.484317779541016,
+ "step": 1335
+ },
+ {
+ "epoch": 18.558951965065503,
+ "grad_norm": 0.016496622934937477,
+ "learning_rate": 0.0006,
+ "loss": 4.514986515045166,
+ "step": 1336
+ },
+ {
+ "epoch": 18.57292576419214,
+ "grad_norm": 0.01795116998255253,
+ "learning_rate": 0.0006,
+ "loss": 4.41473388671875,
+ "step": 1337
+ },
+ {
+ "epoch": 18.586899563318777,
+ "grad_norm": 0.019062276929616928,
+ "learning_rate": 0.0006,
+ "loss": 4.601996898651123,
+ "step": 1338
+ },
+ {
+ "epoch": 18.600873362445416,
+ "grad_norm": 0.017174631357192993,
+ "learning_rate": 0.0006,
+ "loss": 4.571218490600586,
+ "step": 1339
+ },
+ {
+ "epoch": 18.61484716157205,
+ "grad_norm": 0.014015606604516506,
+ "learning_rate": 0.0006,
+ "loss": 4.526606559753418,
+ "step": 1340
+ },
+ {
+ "epoch": 18.62882096069869,
+ "grad_norm": 0.015358510427176952,
+ "learning_rate": 0.0006,
+ "loss": 4.604286193847656,
+ "step": 1341
+ },
+ {
+ "epoch": 18.64279475982533,
+ "grad_norm": 0.01619391329586506,
+ "learning_rate": 0.0006,
+ "loss": 4.560883522033691,
+ "step": 1342
+ },
+ {
+ "epoch": 18.656768558951963,
+ "grad_norm": 0.015972964465618134,
+ "learning_rate": 0.0006,
+ "loss": 4.479755401611328,
+ "step": 1343
+ },
+ {
+ "epoch": 18.670742358078602,
+ "grad_norm": 0.016861025243997574,
+ "learning_rate": 0.0006,
+ "loss": 4.474471092224121,
+ "step": 1344
+ },
+ {
+ "epoch": 18.68471615720524,
+ "grad_norm": 0.01822078227996826,
+ "learning_rate": 0.0006,
+ "loss": 4.509489059448242,
+ "step": 1345
+ },
+ {
+ "epoch": 18.69868995633188,
+ "grad_norm": 0.019415007904171944,
+ "learning_rate": 0.0006,
+ "loss": 4.508941173553467,
+ "step": 1346
+ },
+ {
+ "epoch": 18.712663755458514,
+ "grad_norm": 0.019856126978993416,
+ "learning_rate": 0.0006,
+ "loss": 4.446670055389404,
+ "step": 1347
+ },
+ {
+ "epoch": 18.726637554585153,
+ "grad_norm": 0.01842520199716091,
+ "learning_rate": 0.0006,
+ "loss": 4.524422645568848,
+ "step": 1348
+ },
+ {
+ "epoch": 18.74061135371179,
+ "grad_norm": 0.017196480184793472,
+ "learning_rate": 0.0006,
+ "loss": 4.498466968536377,
+ "step": 1349
+ },
+ {
+ "epoch": 18.754585152838427,
+ "grad_norm": 0.01823010854423046,
+ "learning_rate": 0.0006,
+ "loss": 4.534684181213379,
+ "step": 1350
+ },
+ {
+ "epoch": 18.768558951965066,
+ "grad_norm": 0.020398065447807312,
+ "learning_rate": 0.0006,
+ "loss": 4.456494331359863,
+ "step": 1351
+ },
+ {
+ "epoch": 18.782532751091704,
+ "grad_norm": 0.022992262616753578,
+ "learning_rate": 0.0006,
+ "loss": 4.476374626159668,
+ "step": 1352
+ },
+ {
+ "epoch": 18.79650655021834,
+ "grad_norm": 0.02235420234501362,
+ "learning_rate": 0.0006,
+ "loss": 4.587776184082031,
+ "step": 1353
+ },
+ {
+ "epoch": 18.810480349344978,
+ "grad_norm": 0.02093626745045185,
+ "learning_rate": 0.0006,
+ "loss": 4.424172401428223,
+ "step": 1354
+ },
+ {
+ "epoch": 18.824454148471617,
+ "grad_norm": 0.01997302658855915,
+ "learning_rate": 0.0006,
+ "loss": 4.448066711425781,
+ "step": 1355
+ },
+ {
+ "epoch": 18.83842794759825,
+ "grad_norm": 0.018265459686517715,
+ "learning_rate": 0.0006,
+ "loss": 4.436610698699951,
+ "step": 1356
+ },
+ {
+ "epoch": 18.85240174672489,
+ "grad_norm": 0.01673833839595318,
+ "learning_rate": 0.0006,
+ "loss": 4.5502214431762695,
+ "step": 1357
+ },
+ {
+ "epoch": 18.86637554585153,
+ "grad_norm": 0.017908833920955658,
+ "learning_rate": 0.0006,
+ "loss": 4.550530433654785,
+ "step": 1358
+ },
+ {
+ "epoch": 18.880349344978168,
+ "grad_norm": 0.01796654611825943,
+ "learning_rate": 0.0006,
+ "loss": 4.531463146209717,
+ "step": 1359
+ },
+ {
+ "epoch": 18.894323144104803,
+ "grad_norm": 0.01646183803677559,
+ "learning_rate": 0.0006,
+ "loss": 4.498582363128662,
+ "step": 1360
+ },
+ {
+ "epoch": 18.90829694323144,
+ "grad_norm": 0.015430327504873276,
+ "learning_rate": 0.0006,
+ "loss": 4.52571439743042,
+ "step": 1361
+ },
+ {
+ "epoch": 18.92227074235808,
+ "grad_norm": 0.016527952626347542,
+ "learning_rate": 0.0006,
+ "loss": 4.547013759613037,
+ "step": 1362
+ },
+ {
+ "epoch": 18.936244541484715,
+ "grad_norm": 0.016263660043478012,
+ "learning_rate": 0.0006,
+ "loss": 4.468754291534424,
+ "step": 1363
+ },
+ {
+ "epoch": 18.950218340611354,
+ "grad_norm": 0.016347210854291916,
+ "learning_rate": 0.0006,
+ "loss": 4.450741767883301,
+ "step": 1364
+ },
+ {
+ "epoch": 18.964192139737992,
+ "grad_norm": 0.01701398566365242,
+ "learning_rate": 0.0006,
+ "loss": 4.534899711608887,
+ "step": 1365
+ },
+ {
+ "epoch": 18.978165938864628,
+ "grad_norm": 0.017693081870675087,
+ "learning_rate": 0.0006,
+ "loss": 4.460553169250488,
+ "step": 1366
+ },
+ {
+ "epoch": 18.992139737991266,
+ "grad_norm": 0.018746308982372284,
+ "learning_rate": 0.0006,
+ "loss": 4.538792610168457,
+ "step": 1367
+ },
+ {
+ "epoch": 19.0,
+ "grad_norm": 0.020099416375160217,
+ "learning_rate": 0.0006,
+ "loss": 4.449617385864258,
+ "step": 1368
+ },
+ {
+ "epoch": 19.0,
+ "eval_loss": 4.7756571769714355,
+ "eval_runtime": 57.1804,
+ "eval_samples_per_second": 42.707,
+ "eval_steps_per_second": 1.347,
+ "step": 1368
+ },
+ {
+ "epoch": 19.01397379912664,
+ "grad_norm": 0.018726933747529984,
+ "learning_rate": 0.0006,
+ "loss": 4.446059703826904,
+ "step": 1369
+ },
+ {
+ "epoch": 19.027947598253274,
+ "grad_norm": 0.01956087350845337,
+ "learning_rate": 0.0006,
+ "loss": 4.4520039558410645,
+ "step": 1370
+ },
+ {
+ "epoch": 19.041921397379912,
+ "grad_norm": 0.020092811435461044,
+ "learning_rate": 0.0006,
+ "loss": 4.543343544006348,
+ "step": 1371
+ },
+ {
+ "epoch": 19.05589519650655,
+ "grad_norm": 0.01843622885644436,
+ "learning_rate": 0.0006,
+ "loss": 4.535579204559326,
+ "step": 1372
+ },
+ {
+ "epoch": 19.069868995633186,
+ "grad_norm": 0.016989488154649734,
+ "learning_rate": 0.0006,
+ "loss": 4.52316951751709,
+ "step": 1373
+ },
+ {
+ "epoch": 19.083842794759825,
+ "grad_norm": 0.01805422455072403,
+ "learning_rate": 0.0006,
+ "loss": 4.5393266677856445,
+ "step": 1374
+ },
+ {
+ "epoch": 19.097816593886463,
+ "grad_norm": 0.01874629408121109,
+ "learning_rate": 0.0006,
+ "loss": 4.422059059143066,
+ "step": 1375
+ },
+ {
+ "epoch": 19.111790393013102,
+ "grad_norm": 0.01804221048951149,
+ "learning_rate": 0.0006,
+ "loss": 4.50508451461792,
+ "step": 1376
+ },
+ {
+ "epoch": 19.125764192139737,
+ "grad_norm": 0.01785474829375744,
+ "learning_rate": 0.0006,
+ "loss": 4.412599563598633,
+ "step": 1377
+ },
+ {
+ "epoch": 19.139737991266376,
+ "grad_norm": 0.01971791312098503,
+ "learning_rate": 0.0006,
+ "loss": 4.518318176269531,
+ "step": 1378
+ },
+ {
+ "epoch": 19.153711790393015,
+ "grad_norm": 0.02061633951961994,
+ "learning_rate": 0.0006,
+ "loss": 4.332393169403076,
+ "step": 1379
+ },
+ {
+ "epoch": 19.16768558951965,
+ "grad_norm": 0.020563429221510887,
+ "learning_rate": 0.0006,
+ "loss": 4.445863723754883,
+ "step": 1380
+ },
+ {
+ "epoch": 19.18165938864629,
+ "grad_norm": 0.019995713606476784,
+ "learning_rate": 0.0006,
+ "loss": 4.497312068939209,
+ "step": 1381
+ },
+ {
+ "epoch": 19.195633187772927,
+ "grad_norm": 0.019823070615530014,
+ "learning_rate": 0.0006,
+ "loss": 4.475085258483887,
+ "step": 1382
+ },
+ {
+ "epoch": 19.209606986899562,
+ "grad_norm": 0.019039355218410492,
+ "learning_rate": 0.0006,
+ "loss": 4.50432014465332,
+ "step": 1383
+ },
+ {
+ "epoch": 19.2235807860262,
+ "grad_norm": 0.017876973375678062,
+ "learning_rate": 0.0006,
+ "loss": 4.384788513183594,
+ "step": 1384
+ },
+ {
+ "epoch": 19.23755458515284,
+ "grad_norm": 0.018697699531912804,
+ "learning_rate": 0.0006,
+ "loss": 4.410778045654297,
+ "step": 1385
+ },
+ {
+ "epoch": 19.251528384279474,
+ "grad_norm": 0.018826806917786598,
+ "learning_rate": 0.0006,
+ "loss": 4.524034023284912,
+ "step": 1386
+ },
+ {
+ "epoch": 19.265502183406113,
+ "grad_norm": 0.01824226602911949,
+ "learning_rate": 0.0006,
+ "loss": 4.526393890380859,
+ "step": 1387
+ },
+ {
+ "epoch": 19.27947598253275,
+ "grad_norm": 0.01753208599984646,
+ "learning_rate": 0.0006,
+ "loss": 4.493703365325928,
+ "step": 1388
+ },
+ {
+ "epoch": 19.29344978165939,
+ "grad_norm": 0.01742432825267315,
+ "learning_rate": 0.0006,
+ "loss": 4.392831802368164,
+ "step": 1389
+ },
+ {
+ "epoch": 19.307423580786025,
+ "grad_norm": 0.017099356278777122,
+ "learning_rate": 0.0006,
+ "loss": 4.578668594360352,
+ "step": 1390
+ },
+ {
+ "epoch": 19.321397379912664,
+ "grad_norm": 0.017946023494005203,
+ "learning_rate": 0.0006,
+ "loss": 4.456718444824219,
+ "step": 1391
+ },
+ {
+ "epoch": 19.335371179039303,
+ "grad_norm": 0.017705943435430527,
+ "learning_rate": 0.0006,
+ "loss": 4.553682327270508,
+ "step": 1392
+ },
+ {
+ "epoch": 19.349344978165938,
+ "grad_norm": 0.016647523269057274,
+ "learning_rate": 0.0006,
+ "loss": 4.525016784667969,
+ "step": 1393
+ },
+ {
+ "epoch": 19.363318777292577,
+ "grad_norm": 0.015407416969537735,
+ "learning_rate": 0.0006,
+ "loss": 4.500231742858887,
+ "step": 1394
+ },
+ {
+ "epoch": 19.377292576419215,
+ "grad_norm": 0.017264019697904587,
+ "learning_rate": 0.0006,
+ "loss": 4.4615349769592285,
+ "step": 1395
+ },
+ {
+ "epoch": 19.39126637554585,
+ "grad_norm": 0.01871555484831333,
+ "learning_rate": 0.0006,
+ "loss": 4.281026840209961,
+ "step": 1396
+ },
+ {
+ "epoch": 19.40524017467249,
+ "grad_norm": 0.017361650243401527,
+ "learning_rate": 0.0006,
+ "loss": 4.479382514953613,
+ "step": 1397
+ },
+ {
+ "epoch": 19.419213973799128,
+ "grad_norm": 0.019781362265348434,
+ "learning_rate": 0.0006,
+ "loss": 4.402320861816406,
+ "step": 1398
+ },
+ {
+ "epoch": 19.433187772925763,
+ "grad_norm": 0.019043298438191414,
+ "learning_rate": 0.0006,
+ "loss": 4.476983070373535,
+ "step": 1399
+ },
+ {
+ "epoch": 19.4471615720524,
+ "grad_norm": 0.017137985676527023,
+ "learning_rate": 0.0006,
+ "loss": 4.418689250946045,
+ "step": 1400
+ },
+ {
+ "epoch": 19.46113537117904,
+ "grad_norm": 0.016481833532452583,
+ "learning_rate": 0.0006,
+ "loss": 4.579145431518555,
+ "step": 1401
+ },
+ {
+ "epoch": 19.475109170305675,
+ "grad_norm": 0.016507714986801147,
+ "learning_rate": 0.0006,
+ "loss": 4.519158363342285,
+ "step": 1402
+ },
+ {
+ "epoch": 19.489082969432314,
+ "grad_norm": 0.018092872574925423,
+ "learning_rate": 0.0006,
+ "loss": 4.513748645782471,
+ "step": 1403
+ },
+ {
+ "epoch": 19.503056768558952,
+ "grad_norm": 0.018153440207242966,
+ "learning_rate": 0.0006,
+ "loss": 4.645384311676025,
+ "step": 1404
+ },
+ {
+ "epoch": 19.51703056768559,
+ "grad_norm": 0.01657119207084179,
+ "learning_rate": 0.0006,
+ "loss": 4.399757385253906,
+ "step": 1405
+ },
+ {
+ "epoch": 19.531004366812226,
+ "grad_norm": 0.0156280305236578,
+ "learning_rate": 0.0006,
+ "loss": 4.467836380004883,
+ "step": 1406
+ },
+ {
+ "epoch": 19.544978165938865,
+ "grad_norm": 0.016838403418660164,
+ "learning_rate": 0.0006,
+ "loss": 4.529176235198975,
+ "step": 1407
+ },
+ {
+ "epoch": 19.558951965065503,
+ "grad_norm": 0.01743236929178238,
+ "learning_rate": 0.0006,
+ "loss": 4.440032482147217,
+ "step": 1408
+ },
+ {
+ "epoch": 19.57292576419214,
+ "grad_norm": 0.01553610060364008,
+ "learning_rate": 0.0006,
+ "loss": 4.425588607788086,
+ "step": 1409
+ },
+ {
+ "epoch": 19.586899563318777,
+ "grad_norm": 0.017259271815419197,
+ "learning_rate": 0.0006,
+ "loss": 4.405452728271484,
+ "step": 1410
+ },
+ {
+ "epoch": 19.600873362445416,
+ "grad_norm": 0.015568481758236885,
+ "learning_rate": 0.0006,
+ "loss": 4.444279193878174,
+ "step": 1411
+ },
+ {
+ "epoch": 19.61484716157205,
+ "grad_norm": 0.015331248752772808,
+ "learning_rate": 0.0006,
+ "loss": 4.451846122741699,
+ "step": 1412
+ },
+ {
+ "epoch": 19.62882096069869,
+ "grad_norm": 0.015667378902435303,
+ "learning_rate": 0.0006,
+ "loss": 4.533931255340576,
+ "step": 1413
+ },
+ {
+ "epoch": 19.64279475982533,
+ "grad_norm": 0.016177602112293243,
+ "learning_rate": 0.0006,
+ "loss": 4.467313289642334,
+ "step": 1414
+ },
+ {
+ "epoch": 19.656768558951963,
+ "grad_norm": 0.01733129471540451,
+ "learning_rate": 0.0006,
+ "loss": 4.496893882751465,
+ "step": 1415
+ },
+ {
+ "epoch": 19.670742358078602,
+ "grad_norm": 0.016539594158530235,
+ "learning_rate": 0.0006,
+ "loss": 4.401961326599121,
+ "step": 1416
+ },
+ {
+ "epoch": 19.68471615720524,
+ "grad_norm": 0.016058551147580147,
+ "learning_rate": 0.0006,
+ "loss": 4.503375053405762,
+ "step": 1417
+ },
+ {
+ "epoch": 19.69868995633188,
+ "grad_norm": 0.01715777814388275,
+ "learning_rate": 0.0006,
+ "loss": 4.3808369636535645,
+ "step": 1418
+ },
+ {
+ "epoch": 19.712663755458514,
+ "grad_norm": 0.01681119203567505,
+ "learning_rate": 0.0006,
+ "loss": 4.5469560623168945,
+ "step": 1419
+ },
+ {
+ "epoch": 19.726637554585153,
+ "grad_norm": 0.01634550653398037,
+ "learning_rate": 0.0006,
+ "loss": 4.328197956085205,
+ "step": 1420
+ },
+ {
+ "epoch": 19.74061135371179,
+ "grad_norm": 0.01549347210675478,
+ "learning_rate": 0.0006,
+ "loss": 4.534621238708496,
+ "step": 1421
+ },
+ {
+ "epoch": 19.754585152838427,
+ "grad_norm": 0.01654665172100067,
+ "learning_rate": 0.0006,
+ "loss": 4.4724225997924805,
+ "step": 1422
+ },
+ {
+ "epoch": 19.768558951965066,
+ "grad_norm": 0.017119025811553,
+ "learning_rate": 0.0006,
+ "loss": 4.3897247314453125,
+ "step": 1423
+ },
+ {
+ "epoch": 19.782532751091704,
+ "grad_norm": 0.01603223942220211,
+ "learning_rate": 0.0006,
+ "loss": 4.412731170654297,
+ "step": 1424
+ },
+ {
+ "epoch": 19.79650655021834,
+ "grad_norm": 0.016774067655205727,
+ "learning_rate": 0.0006,
+ "loss": 4.54707145690918,
+ "step": 1425
+ },
+ {
+ "epoch": 19.810480349344978,
+ "grad_norm": 0.016453025862574577,
+ "learning_rate": 0.0006,
+ "loss": 4.462501049041748,
+ "step": 1426
+ },
+ {
+ "epoch": 19.824454148471617,
+ "grad_norm": 0.01597629487514496,
+ "learning_rate": 0.0006,
+ "loss": 4.229758262634277,
+ "step": 1427
+ },
+ {
+ "epoch": 19.83842794759825,
+ "grad_norm": 0.017999034374952316,
+ "learning_rate": 0.0006,
+ "loss": 4.389263153076172,
+ "step": 1428
+ },
+ {
+ "epoch": 19.85240174672489,
+ "grad_norm": 0.02165898121893406,
+ "learning_rate": 0.0006,
+ "loss": 4.342130661010742,
+ "step": 1429
+ },
+ {
+ "epoch": 19.86637554585153,
+ "grad_norm": 0.02222822792828083,
+ "learning_rate": 0.0006,
+ "loss": 4.492661476135254,
+ "step": 1430
+ },
+ {
+ "epoch": 19.880349344978168,
+ "grad_norm": 0.019378066062927246,
+ "learning_rate": 0.0006,
+ "loss": 4.4614105224609375,
+ "step": 1431
+ },
+ {
+ "epoch": 19.894323144104803,
+ "grad_norm": 0.019908083602786064,
+ "learning_rate": 0.0006,
+ "loss": 4.370929718017578,
+ "step": 1432
+ },
+ {
+ "epoch": 19.90829694323144,
+ "grad_norm": 0.022296471521258354,
+ "learning_rate": 0.0006,
+ "loss": 4.438320159912109,
+ "step": 1433
+ },
+ {
+ "epoch": 19.92227074235808,
+ "grad_norm": 0.021810343489050865,
+ "learning_rate": 0.0006,
+ "loss": 4.462067127227783,
+ "step": 1434
+ },
+ {
+ "epoch": 19.936244541484715,
+ "grad_norm": 0.02129209227859974,
+ "learning_rate": 0.0006,
+ "loss": 4.541261196136475,
+ "step": 1435
+ },
+ {
+ "epoch": 19.950218340611354,
+ "grad_norm": 0.02079629711806774,
+ "learning_rate": 0.0006,
+ "loss": 4.36712646484375,
+ "step": 1436
+ },
+ {
+ "epoch": 19.964192139737992,
+ "grad_norm": 0.022501740604639053,
+ "learning_rate": 0.0006,
+ "loss": 4.5279083251953125,
+ "step": 1437
+ },
+ {
+ "epoch": 19.978165938864628,
+ "grad_norm": 0.019121676683425903,
+ "learning_rate": 0.0006,
+ "loss": 4.477789878845215,
+ "step": 1438
+ },
+ {
+ "epoch": 19.992139737991266,
+ "grad_norm": 0.017429588362574577,
+ "learning_rate": 0.0006,
+ "loss": 4.4023847579956055,
+ "step": 1439
+ },
+ {
+ "epoch": 20.0,
+ "grad_norm": 0.018857665359973907,
+ "learning_rate": 0.0006,
+ "loss": 4.494743347167969,
+ "step": 1440
+ },
+ {
+ "epoch": 20.0,
+ "eval_loss": 4.694555759429932,
+ "eval_runtime": 56.4146,
+ "eval_samples_per_second": 43.287,
+ "eval_steps_per_second": 1.365,
+ "step": 1440
+ },
+ {
+ "epoch": 20.01397379912664,
+ "grad_norm": 0.016505861654877663,
+ "learning_rate": 0.0006,
+ "loss": 4.506229877471924,
+ "step": 1441
+ },
+ {
+ "epoch": 20.027947598253274,
+ "grad_norm": 0.017571842297911644,
+ "learning_rate": 0.0006,
+ "loss": 4.384073257446289,
+ "step": 1442
+ },
+ {
+ "epoch": 20.041921397379912,
+ "grad_norm": 0.019398365169763565,
+ "learning_rate": 0.0006,
+ "loss": 4.482177734375,
+ "step": 1443
+ },
+ {
+ "epoch": 20.05589519650655,
+ "grad_norm": 0.02011062018573284,
+ "learning_rate": 0.0006,
+ "loss": 4.440773963928223,
+ "step": 1444
+ },
+ {
+ "epoch": 20.069868995633186,
+ "grad_norm": 0.019026953727006912,
+ "learning_rate": 0.0006,
+ "loss": 4.507168769836426,
+ "step": 1445
+ },
+ {
+ "epoch": 20.083842794759825,
+ "grad_norm": 0.020506031811237335,
+ "learning_rate": 0.0006,
+ "loss": 4.469079971313477,
+ "step": 1446
+ },
+ {
+ "epoch": 20.097816593886463,
+ "grad_norm": 0.019608162343502045,
+ "learning_rate": 0.0006,
+ "loss": 4.395927429199219,
+ "step": 1447
+ },
+ {
+ "epoch": 20.111790393013102,
+ "grad_norm": 0.019419124349951744,
+ "learning_rate": 0.0006,
+ "loss": 4.484235763549805,
+ "step": 1448
+ },
+ {
+ "epoch": 20.125764192139737,
+ "grad_norm": 0.021775128319859505,
+ "learning_rate": 0.0006,
+ "loss": 4.333173751831055,
+ "step": 1449
+ },
+ {
+ "epoch": 20.139737991266376,
+ "grad_norm": 0.025512272492051125,
+ "learning_rate": 0.0006,
+ "loss": 4.460362434387207,
+ "step": 1450
+ },
+ {
+ "epoch": 20.153711790393015,
+ "grad_norm": 0.025157401338219643,
+ "learning_rate": 0.0006,
+ "loss": 4.4361395835876465,
+ "step": 1451
+ },
+ {
+ "epoch": 20.16768558951965,
+ "grad_norm": 0.019862616434693336,
+ "learning_rate": 0.0006,
+ "loss": 4.426807403564453,
+ "step": 1452
+ },
+ {
+ "epoch": 20.18165938864629,
+ "grad_norm": 0.01932152360677719,
+ "learning_rate": 0.0006,
+ "loss": 4.375677108764648,
+ "step": 1453
+ },
+ {
+ "epoch": 20.195633187772927,
+ "grad_norm": 0.01760999858379364,
+ "learning_rate": 0.0006,
+ "loss": 4.42259407043457,
+ "step": 1454
+ },
+ {
+ "epoch": 20.209606986899562,
+ "grad_norm": 0.016998112201690674,
+ "learning_rate": 0.0006,
+ "loss": 4.481280326843262,
+ "step": 1455
+ },
+ {
+ "epoch": 20.2235807860262,
+ "grad_norm": 0.017199190333485603,
+ "learning_rate": 0.0006,
+ "loss": 4.496712684631348,
+ "step": 1456
+ },
+ {
+ "epoch": 20.23755458515284,
+ "grad_norm": 0.016833679750561714,
+ "learning_rate": 0.0006,
+ "loss": 4.40736722946167,
+ "step": 1457
+ },
+ {
+ "epoch": 20.251528384279474,
+ "grad_norm": 0.017546730116009712,
+ "learning_rate": 0.0006,
+ "loss": 4.4915266036987305,
+ "step": 1458
+ },
+ {
+ "epoch": 20.265502183406113,
+ "grad_norm": 0.016595078632235527,
+ "learning_rate": 0.0006,
+ "loss": 4.460815906524658,
+ "step": 1459
+ },
+ {
+ "epoch": 20.27947598253275,
+ "grad_norm": 0.017728568986058235,
+ "learning_rate": 0.0006,
+ "loss": 4.346048831939697,
+ "step": 1460
+ },
+ {
+ "epoch": 20.29344978165939,
+ "grad_norm": 0.01849197782576084,
+ "learning_rate": 0.0006,
+ "loss": 4.399681091308594,
+ "step": 1461
+ },
+ {
+ "epoch": 20.307423580786025,
+ "grad_norm": 0.018336588516831398,
+ "learning_rate": 0.0006,
+ "loss": 4.414183616638184,
+ "step": 1462
+ },
+ {
+ "epoch": 20.321397379912664,
+ "grad_norm": 0.018086519092321396,
+ "learning_rate": 0.0006,
+ "loss": 4.346586227416992,
+ "step": 1463
+ },
+ {
+ "epoch": 20.335371179039303,
+ "grad_norm": 0.019697299227118492,
+ "learning_rate": 0.0006,
+ "loss": 4.4420952796936035,
+ "step": 1464
+ },
+ {
+ "epoch": 20.349344978165938,
+ "grad_norm": 0.02198074571788311,
+ "learning_rate": 0.0006,
+ "loss": 4.588729381561279,
+ "step": 1465
+ },
+ {
+ "epoch": 20.363318777292577,
+ "grad_norm": 0.022813329473137856,
+ "learning_rate": 0.0006,
+ "loss": 4.49729585647583,
+ "step": 1466
+ },
+ {
+ "epoch": 20.377292576419215,
+ "grad_norm": 0.022006649523973465,
+ "learning_rate": 0.0006,
+ "loss": 4.475641250610352,
+ "step": 1467
+ },
+ {
+ "epoch": 20.39126637554585,
+ "grad_norm": 0.023188438266515732,
+ "learning_rate": 0.0006,
+ "loss": 4.404322624206543,
+ "step": 1468
+ },
+ {
+ "epoch": 20.40524017467249,
+ "grad_norm": 0.022226877510547638,
+ "learning_rate": 0.0006,
+ "loss": 4.50904655456543,
+ "step": 1469
+ },
+ {
+ "epoch": 20.419213973799128,
+ "grad_norm": 0.021082637831568718,
+ "learning_rate": 0.0006,
+ "loss": 4.339011192321777,
+ "step": 1470
+ },
+ {
+ "epoch": 20.433187772925763,
+ "grad_norm": 0.02013542130589485,
+ "learning_rate": 0.0006,
+ "loss": 4.516942024230957,
+ "step": 1471
+ },
+ {
+ "epoch": 20.4471615720524,
+ "grad_norm": 0.01898968778550625,
+ "learning_rate": 0.0006,
+ "loss": 4.315804481506348,
+ "step": 1472
+ },
+ {
+ "epoch": 20.46113537117904,
+ "grad_norm": 0.019450880587100983,
+ "learning_rate": 0.0006,
+ "loss": 4.4794921875,
+ "step": 1473
+ },
+ {
+ "epoch": 20.475109170305675,
+ "grad_norm": 0.018577802926301956,
+ "learning_rate": 0.0006,
+ "loss": 4.422323703765869,
+ "step": 1474
+ },
+ {
+ "epoch": 20.489082969432314,
+ "grad_norm": 0.01778412237763405,
+ "learning_rate": 0.0006,
+ "loss": 4.429789066314697,
+ "step": 1475
+ },
+ {
+ "epoch": 20.503056768558952,
+ "grad_norm": 0.017134476453065872,
+ "learning_rate": 0.0006,
+ "loss": 4.416459083557129,
+ "step": 1476
+ },
+ {
+ "epoch": 20.51703056768559,
+ "grad_norm": 0.017153557389974594,
+ "learning_rate": 0.0006,
+ "loss": 4.446170806884766,
+ "step": 1477
+ },
+ {
+ "epoch": 20.531004366812226,
+ "grad_norm": 0.014998532831668854,
+ "learning_rate": 0.0006,
+ "loss": 4.419447898864746,
+ "step": 1478
+ },
+ {
+ "epoch": 20.544978165938865,
+ "grad_norm": 0.015877850353717804,
+ "learning_rate": 0.0006,
+ "loss": 4.312338829040527,
+ "step": 1479
+ },
+ {
+ "epoch": 20.558951965065503,
+ "grad_norm": 0.01569489948451519,
+ "learning_rate": 0.0006,
+ "loss": 4.572905540466309,
+ "step": 1480
+ },
+ {
+ "epoch": 20.57292576419214,
+ "grad_norm": 0.01591474376618862,
+ "learning_rate": 0.0006,
+ "loss": 4.414907455444336,
+ "step": 1481
+ },
+ {
+ "epoch": 20.586899563318777,
+ "grad_norm": 0.015395903028547764,
+ "learning_rate": 0.0006,
+ "loss": 4.549131393432617,
+ "step": 1482
+ },
+ {
+ "epoch": 20.600873362445416,
+ "grad_norm": 0.015715358778834343,
+ "learning_rate": 0.0006,
+ "loss": 4.530194282531738,
+ "step": 1483
+ },
+ {
+ "epoch": 20.61484716157205,
+ "grad_norm": 0.016042975708842278,
+ "learning_rate": 0.0006,
+ "loss": 4.322149276733398,
+ "step": 1484
+ },
+ {
+ "epoch": 20.62882096069869,
+ "grad_norm": 0.01664186641573906,
+ "learning_rate": 0.0006,
+ "loss": 4.424866676330566,
+ "step": 1485
+ },
+ {
+ "epoch": 20.64279475982533,
+ "grad_norm": 0.01631534844636917,
+ "learning_rate": 0.0006,
+ "loss": 4.441752910614014,
+ "step": 1486
+ },
+ {
+ "epoch": 20.656768558951963,
+ "grad_norm": 0.014057058840990067,
+ "learning_rate": 0.0006,
+ "loss": 4.399982452392578,
+ "step": 1487
+ },
+ {
+ "epoch": 20.670742358078602,
+ "grad_norm": 0.015076273120939732,
+ "learning_rate": 0.0006,
+ "loss": 4.441324234008789,
+ "step": 1488
+ },
+ {
+ "epoch": 20.68471615720524,
+ "grad_norm": 0.014138277620077133,
+ "learning_rate": 0.0006,
+ "loss": 4.4312543869018555,
+ "step": 1489
+ },
+ {
+ "epoch": 20.69868995633188,
+ "grad_norm": 0.013762440532445908,
+ "learning_rate": 0.0006,
+ "loss": 4.387807846069336,
+ "step": 1490
+ },
+ {
+ "epoch": 20.712663755458514,
+ "grad_norm": 0.015058411285281181,
+ "learning_rate": 0.0006,
+ "loss": 4.594747543334961,
+ "step": 1491
+ },
+ {
+ "epoch": 20.726637554585153,
+ "grad_norm": 0.013827123679220676,
+ "learning_rate": 0.0006,
+ "loss": 4.434604167938232,
+ "step": 1492
+ },
+ {
+ "epoch": 20.74061135371179,
+ "grad_norm": 0.015315295197069645,
+ "learning_rate": 0.0006,
+ "loss": 4.588066101074219,
+ "step": 1493
+ },
+ {
+ "epoch": 20.754585152838427,
+ "grad_norm": 0.018678732216358185,
+ "learning_rate": 0.0006,
+ "loss": 4.3514204025268555,
+ "step": 1494
+ },
+ {
+ "epoch": 20.768558951965066,
+ "grad_norm": 0.019521350041031837,
+ "learning_rate": 0.0006,
+ "loss": 4.437658786773682,
+ "step": 1495
+ },
+ {
+ "epoch": 20.782532751091704,
+ "grad_norm": 0.018363403156399727,
+ "learning_rate": 0.0006,
+ "loss": 4.495002269744873,
+ "step": 1496
+ },
+ {
+ "epoch": 20.79650655021834,
+ "grad_norm": 0.016830725595355034,
+ "learning_rate": 0.0006,
+ "loss": 4.402471542358398,
+ "step": 1497
+ },
+ {
+ "epoch": 20.810480349344978,
+ "grad_norm": 0.01722072623670101,
+ "learning_rate": 0.0006,
+ "loss": 4.437653541564941,
+ "step": 1498
+ },
+ {
+ "epoch": 20.824454148471617,
+ "grad_norm": 0.017459599301218987,
+ "learning_rate": 0.0006,
+ "loss": 4.366603851318359,
+ "step": 1499
+ },
+ {
+ "epoch": 20.83842794759825,
+ "grad_norm": 0.018902219831943512,
+ "learning_rate": 0.0006,
+ "loss": 4.501561164855957,
+ "step": 1500
+ },
+ {
+ "epoch": 20.85240174672489,
+ "grad_norm": 0.020814096555113792,
+ "learning_rate": 0.0006,
+ "loss": 4.535295486450195,
+ "step": 1501
+ },
+ {
+ "epoch": 20.86637554585153,
+ "grad_norm": 0.018990976735949516,
+ "learning_rate": 0.0006,
+ "loss": 4.333580017089844,
+ "step": 1502
+ },
+ {
+ "epoch": 20.880349344978168,
+ "grad_norm": 0.01837880350649357,
+ "learning_rate": 0.0006,
+ "loss": 4.313388824462891,
+ "step": 1503
+ },
+ {
+ "epoch": 20.894323144104803,
+ "grad_norm": 0.019820360466837883,
+ "learning_rate": 0.0006,
+ "loss": 4.388157844543457,
+ "step": 1504
+ },
+ {
+ "epoch": 20.90829694323144,
+ "grad_norm": 0.020181143656373024,
+ "learning_rate": 0.0006,
+ "loss": 4.572224140167236,
+ "step": 1505
+ },
+ {
+ "epoch": 20.92227074235808,
+ "grad_norm": 0.021206077188253403,
+ "learning_rate": 0.0006,
+ "loss": 4.3127970695495605,
+ "step": 1506
+ },
+ {
+ "epoch": 20.936244541484715,
+ "grad_norm": 0.017962060868740082,
+ "learning_rate": 0.0006,
+ "loss": 4.384734630584717,
+ "step": 1507
+ },
+ {
+ "epoch": 20.950218340611354,
+ "grad_norm": 0.01737220585346222,
+ "learning_rate": 0.0006,
+ "loss": 4.4075751304626465,
+ "step": 1508
+ },
+ {
+ "epoch": 20.964192139737992,
+ "grad_norm": 0.01839268207550049,
+ "learning_rate": 0.0006,
+ "loss": 4.553779125213623,
+ "step": 1509
+ },
+ {
+ "epoch": 20.978165938864628,
+ "grad_norm": 0.020102640613913536,
+ "learning_rate": 0.0006,
+ "loss": 4.399398326873779,
+ "step": 1510
+ },
+ {
+ "epoch": 20.992139737991266,
+ "grad_norm": 0.018831463530659676,
+ "learning_rate": 0.0006,
+ "loss": 4.358499526977539,
+ "step": 1511
+ },
+ {
+ "epoch": 21.0,
+ "grad_norm": 0.018254734575748444,
+ "learning_rate": 0.0006,
+ "loss": 4.384119510650635,
+ "step": 1512
+ },
+ {
+ "epoch": 21.0,
+ "eval_loss": 4.7524237632751465,
+ "eval_runtime": 57.8812,
+ "eval_samples_per_second": 42.19,
+ "eval_steps_per_second": 1.33,
+ "step": 1512
+ },
+ {
+ "epoch": 21.01397379912664,
+ "grad_norm": 0.019728587940335274,
+ "learning_rate": 0.0006,
+ "loss": 4.32363224029541,
+ "step": 1513
+ },
+ {
+ "epoch": 21.027947598253274,
+ "grad_norm": 0.02216625027358532,
+ "learning_rate": 0.0006,
+ "loss": 4.436988353729248,
+ "step": 1514
+ },
+ {
+ "epoch": 21.041921397379912,
+ "grad_norm": 0.02227606810629368,
+ "learning_rate": 0.0006,
+ "loss": 4.406063556671143,
+ "step": 1515
+ },
+ {
+ "epoch": 21.05589519650655,
+ "grad_norm": 0.020187009125947952,
+ "learning_rate": 0.0006,
+ "loss": 4.324402332305908,
+ "step": 1516
+ },
+ {
+ "epoch": 21.069868995633186,
+ "grad_norm": 0.02095157466828823,
+ "learning_rate": 0.0006,
+ "loss": 4.456544876098633,
+ "step": 1517
+ },
+ {
+ "epoch": 21.083842794759825,
+ "grad_norm": 0.02146013453602791,
+ "learning_rate": 0.0006,
+ "loss": 4.368387222290039,
+ "step": 1518
+ },
+ {
+ "epoch": 21.097816593886463,
+ "grad_norm": 0.02162035182118416,
+ "learning_rate": 0.0006,
+ "loss": 4.48838996887207,
+ "step": 1519
+ },
+ {
+ "epoch": 21.111790393013102,
+ "grad_norm": 0.021745676174759865,
+ "learning_rate": 0.0006,
+ "loss": 4.30778694152832,
+ "step": 1520
+ },
+ {
+ "epoch": 21.125764192139737,
+ "grad_norm": 0.02173588052392006,
+ "learning_rate": 0.0006,
+ "loss": 4.433149814605713,
+ "step": 1521
+ },
+ {
+ "epoch": 21.139737991266376,
+ "grad_norm": 0.020950648933649063,
+ "learning_rate": 0.0006,
+ "loss": 4.520089149475098,
+ "step": 1522
+ },
+ {
+ "epoch": 21.153711790393015,
+ "grad_norm": 0.020684752613306046,
+ "learning_rate": 0.0006,
+ "loss": 4.372437477111816,
+ "step": 1523
+ },
+ {
+ "epoch": 21.16768558951965,
+ "grad_norm": 0.019472869113087654,
+ "learning_rate": 0.0006,
+ "loss": 4.437015533447266,
+ "step": 1524
+ },
+ {
+ "epoch": 21.18165938864629,
+ "grad_norm": 0.01846962980926037,
+ "learning_rate": 0.0006,
+ "loss": 4.449531555175781,
+ "step": 1525
+ },
+ {
+ "epoch": 21.195633187772927,
+ "grad_norm": 0.017377803102135658,
+ "learning_rate": 0.0006,
+ "loss": 4.458279132843018,
+ "step": 1526
+ },
+ {
+ "epoch": 21.209606986899562,
+ "grad_norm": 0.016002200543880463,
+ "learning_rate": 0.0006,
+ "loss": 4.266297817230225,
+ "step": 1527
+ },
+ {
+ "epoch": 21.2235807860262,
+ "grad_norm": 0.015575903467833996,
+ "learning_rate": 0.0006,
+ "loss": 4.342632293701172,
+ "step": 1528
+ },
+ {
+ "epoch": 21.23755458515284,
+ "grad_norm": 0.015634698793292046,
+ "learning_rate": 0.0006,
+ "loss": 4.338623523712158,
+ "step": 1529
+ },
+ {
+ "epoch": 21.251528384279474,
+ "grad_norm": 0.01449581328779459,
+ "learning_rate": 0.0006,
+ "loss": 4.354330062866211,
+ "step": 1530
+ },
+ {
+ "epoch": 21.265502183406113,
+ "grad_norm": 0.01487616915255785,
+ "learning_rate": 0.0006,
+ "loss": 4.486835479736328,
+ "step": 1531
+ },
+ {
+ "epoch": 21.27947598253275,
+ "grad_norm": 0.016346555203199387,
+ "learning_rate": 0.0006,
+ "loss": 4.407098770141602,
+ "step": 1532
+ },
+ {
+ "epoch": 21.29344978165939,
+ "grad_norm": 0.01682182401418686,
+ "learning_rate": 0.0006,
+ "loss": 4.436809539794922,
+ "step": 1533
+ },
+ {
+ "epoch": 21.307423580786025,
+ "grad_norm": 0.017704004421830177,
+ "learning_rate": 0.0006,
+ "loss": 4.3854780197143555,
+ "step": 1534
+ },
+ {
+ "epoch": 21.321397379912664,
+ "grad_norm": 0.018437420949339867,
+ "learning_rate": 0.0006,
+ "loss": 4.413283824920654,
+ "step": 1535
+ },
+ {
+ "epoch": 21.335371179039303,
+ "grad_norm": 0.016911081969738007,
+ "learning_rate": 0.0006,
+ "loss": 4.357306003570557,
+ "step": 1536
+ },
+ {
+ "epoch": 21.349344978165938,
+ "grad_norm": 0.01642223633825779,
+ "learning_rate": 0.0006,
+ "loss": 4.422322750091553,
+ "step": 1537
+ },
+ {
+ "epoch": 21.363318777292577,
+ "grad_norm": 0.01780877821147442,
+ "learning_rate": 0.0006,
+ "loss": 4.387620449066162,
+ "step": 1538
+ },
+ {
+ "epoch": 21.377292576419215,
+ "grad_norm": 0.01653975248336792,
+ "learning_rate": 0.0006,
+ "loss": 4.327267646789551,
+ "step": 1539
+ },
+ {
+ "epoch": 21.39126637554585,
+ "grad_norm": 0.01572156324982643,
+ "learning_rate": 0.0006,
+ "loss": 4.475979804992676,
+ "step": 1540
+ },
+ {
+ "epoch": 21.40524017467249,
+ "grad_norm": 0.017156066372990608,
+ "learning_rate": 0.0006,
+ "loss": 4.4011125564575195,
+ "step": 1541
+ },
+ {
+ "epoch": 21.419213973799128,
+ "grad_norm": 0.01788829080760479,
+ "learning_rate": 0.0006,
+ "loss": 4.4174089431762695,
+ "step": 1542
+ },
+ {
+ "epoch": 21.433187772925763,
+ "grad_norm": 0.01728985086083412,
+ "learning_rate": 0.0006,
+ "loss": 4.429244041442871,
+ "step": 1543
+ },
+ {
+ "epoch": 21.4471615720524,
+ "grad_norm": 0.016534466296434402,
+ "learning_rate": 0.0006,
+ "loss": 4.326068878173828,
+ "step": 1544
+ },
+ {
+ "epoch": 21.46113537117904,
+ "grad_norm": 0.016582584008574486,
+ "learning_rate": 0.0006,
+ "loss": 4.2975969314575195,
+ "step": 1545
+ },
+ {
+ "epoch": 21.475109170305675,
+ "grad_norm": 0.018034178763628006,
+ "learning_rate": 0.0006,
+ "loss": 4.501319408416748,
+ "step": 1546
+ },
+ {
+ "epoch": 21.489082969432314,
+ "grad_norm": 0.017827702686190605,
+ "learning_rate": 0.0006,
+ "loss": 4.502284526824951,
+ "step": 1547
+ },
+ {
+ "epoch": 21.503056768558952,
+ "grad_norm": 0.017182299867272377,
+ "learning_rate": 0.0006,
+ "loss": 4.380828857421875,
+ "step": 1548
+ },
+ {
+ "epoch": 21.51703056768559,
+ "grad_norm": 0.019101588055491447,
+ "learning_rate": 0.0006,
+ "loss": 4.31564998626709,
+ "step": 1549
+ },
+ {
+ "epoch": 21.531004366812226,
+ "grad_norm": 0.0204677302390337,
+ "learning_rate": 0.0006,
+ "loss": 4.3840131759643555,
+ "step": 1550
+ },
+ {
+ "epoch": 21.544978165938865,
+ "grad_norm": 0.0188993439078331,
+ "learning_rate": 0.0006,
+ "loss": 4.4257731437683105,
+ "step": 1551
+ },
+ {
+ "epoch": 21.558951965065503,
+ "grad_norm": 0.01730230636894703,
+ "learning_rate": 0.0006,
+ "loss": 4.445132255554199,
+ "step": 1552
+ },
+ {
+ "epoch": 21.57292576419214,
+ "grad_norm": 0.01801574043929577,
+ "learning_rate": 0.0006,
+ "loss": 4.427064895629883,
+ "step": 1553
+ },
+ {
+ "epoch": 21.586899563318777,
+ "grad_norm": 0.018417565152049065,
+ "learning_rate": 0.0006,
+ "loss": 4.319595813751221,
+ "step": 1554
+ },
+ {
+ "epoch": 21.600873362445416,
+ "grad_norm": 0.01683160290122032,
+ "learning_rate": 0.0006,
+ "loss": 4.349218368530273,
+ "step": 1555
+ },
+ {
+ "epoch": 21.61484716157205,
+ "grad_norm": 0.016973106190562248,
+ "learning_rate": 0.0006,
+ "loss": 4.375990390777588,
+ "step": 1556
+ },
+ {
+ "epoch": 21.62882096069869,
+ "grad_norm": 0.015688583254814148,
+ "learning_rate": 0.0006,
+ "loss": 4.408170223236084,
+ "step": 1557
+ },
+ {
+ "epoch": 21.64279475982533,
+ "grad_norm": 0.01573043316602707,
+ "learning_rate": 0.0006,
+ "loss": 4.416255950927734,
+ "step": 1558
+ },
+ {
+ "epoch": 21.656768558951963,
+ "grad_norm": 0.01744293048977852,
+ "learning_rate": 0.0006,
+ "loss": 4.477141380310059,
+ "step": 1559
+ },
+ {
+ "epoch": 21.670742358078602,
+ "grad_norm": 0.017503513023257256,
+ "learning_rate": 0.0006,
+ "loss": 4.3367486000061035,
+ "step": 1560
+ },
+ {
+ "epoch": 21.68471615720524,
+ "grad_norm": 0.016592150554060936,
+ "learning_rate": 0.0006,
+ "loss": 4.314785480499268,
+ "step": 1561
+ },
+ {
+ "epoch": 21.69868995633188,
+ "grad_norm": 0.015774084255099297,
+ "learning_rate": 0.0006,
+ "loss": 4.448209285736084,
+ "step": 1562
+ },
+ {
+ "epoch": 21.712663755458514,
+ "grad_norm": 0.015779150649905205,
+ "learning_rate": 0.0006,
+ "loss": 4.392690658569336,
+ "step": 1563
+ },
+ {
+ "epoch": 21.726637554585153,
+ "grad_norm": 0.015392648056149483,
+ "learning_rate": 0.0006,
+ "loss": 4.437741756439209,
+ "step": 1564
+ },
+ {
+ "epoch": 21.74061135371179,
+ "grad_norm": 0.01530836895108223,
+ "learning_rate": 0.0006,
+ "loss": 4.424595832824707,
+ "step": 1565
+ },
+ {
+ "epoch": 21.754585152838427,
+ "grad_norm": 0.015600779093801975,
+ "learning_rate": 0.0006,
+ "loss": 4.335417747497559,
+ "step": 1566
+ },
+ {
+ "epoch": 21.768558951965066,
+ "grad_norm": 0.014646518044173717,
+ "learning_rate": 0.0006,
+ "loss": 4.279081344604492,
+ "step": 1567
+ },
+ {
+ "epoch": 21.782532751091704,
+ "grad_norm": 0.014132864773273468,
+ "learning_rate": 0.0006,
+ "loss": 4.345950126647949,
+ "step": 1568
+ },
+ {
+ "epoch": 21.79650655021834,
+ "grad_norm": 0.01425724383443594,
+ "learning_rate": 0.0006,
+ "loss": 4.277588844299316,
+ "step": 1569
+ },
+ {
+ "epoch": 21.810480349344978,
+ "grad_norm": 0.015396572649478912,
+ "learning_rate": 0.0006,
+ "loss": 4.326196670532227,
+ "step": 1570
+ },
+ {
+ "epoch": 21.824454148471617,
+ "grad_norm": 0.01683669537305832,
+ "learning_rate": 0.0006,
+ "loss": 4.3294172286987305,
+ "step": 1571
+ },
+ {
+ "epoch": 21.83842794759825,
+ "grad_norm": 0.016365647315979004,
+ "learning_rate": 0.0006,
+ "loss": 4.523321628570557,
+ "step": 1572
+ },
+ {
+ "epoch": 21.85240174672489,
+ "grad_norm": 0.017142774537205696,
+ "learning_rate": 0.0006,
+ "loss": 4.475842475891113,
+ "step": 1573
+ },
+ {
+ "epoch": 21.86637554585153,
+ "grad_norm": 0.015950839966535568,
+ "learning_rate": 0.0006,
+ "loss": 4.397187232971191,
+ "step": 1574
+ },
+ {
+ "epoch": 21.880349344978168,
+ "grad_norm": 0.016180362552404404,
+ "learning_rate": 0.0006,
+ "loss": 4.414778709411621,
+ "step": 1575
+ },
+ {
+ "epoch": 21.894323144104803,
+ "grad_norm": 0.017615556716918945,
+ "learning_rate": 0.0006,
+ "loss": 4.348567962646484,
+ "step": 1576
+ },
+ {
+ "epoch": 21.90829694323144,
+ "grad_norm": 0.01896277628839016,
+ "learning_rate": 0.0006,
+ "loss": 4.431762218475342,
+ "step": 1577
+ },
+ {
+ "epoch": 21.92227074235808,
+ "grad_norm": 0.01990971714258194,
+ "learning_rate": 0.0006,
+ "loss": 4.431285858154297,
+ "step": 1578
+ },
+ {
+ "epoch": 21.936244541484715,
+ "grad_norm": 0.0201258547604084,
+ "learning_rate": 0.0006,
+ "loss": 4.309872627258301,
+ "step": 1579
+ },
+ {
+ "epoch": 21.950218340611354,
+ "grad_norm": 0.019620198756456375,
+ "learning_rate": 0.0006,
+ "loss": 4.484044551849365,
+ "step": 1580
+ },
+ {
+ "epoch": 21.964192139737992,
+ "grad_norm": 0.01771123707294464,
+ "learning_rate": 0.0006,
+ "loss": 4.368093013763428,
+ "step": 1581
+ },
+ {
+ "epoch": 21.978165938864628,
+ "grad_norm": 0.017501654103398323,
+ "learning_rate": 0.0006,
+ "loss": 4.372651100158691,
+ "step": 1582
+ },
+ {
+ "epoch": 21.992139737991266,
+ "grad_norm": 0.01693608984351158,
+ "learning_rate": 0.0006,
+ "loss": 4.379167556762695,
+ "step": 1583
+ },
+ {
+ "epoch": 22.0,
+ "grad_norm": 0.02002849616110325,
+ "learning_rate": 0.0006,
+ "loss": 4.329627990722656,
+ "step": 1584
+ },
+ {
+ "epoch": 22.0,
+ "eval_loss": 4.723696231842041,
+ "eval_runtime": 56.1758,
+ "eval_samples_per_second": 43.471,
+ "eval_steps_per_second": 1.371,
+ "step": 1584
+ },
+ {
+ "epoch": 22.01397379912664,
+ "grad_norm": 0.02030305378139019,
+ "learning_rate": 0.0006,
+ "loss": 4.355566024780273,
+ "step": 1585
+ },
+ {
+ "epoch": 22.027947598253274,
+ "grad_norm": 0.018638934940099716,
+ "learning_rate": 0.0006,
+ "loss": 4.314009189605713,
+ "step": 1586
+ },
+ {
+ "epoch": 22.041921397379912,
+ "grad_norm": 0.019917072728276253,
+ "learning_rate": 0.0006,
+ "loss": 4.367525577545166,
+ "step": 1587
+ },
+ {
+ "epoch": 22.05589519650655,
+ "grad_norm": 0.020785167813301086,
+ "learning_rate": 0.0006,
+ "loss": 4.375561714172363,
+ "step": 1588
+ },
+ {
+ "epoch": 22.069868995633186,
+ "grad_norm": 0.021214453503489494,
+ "learning_rate": 0.0006,
+ "loss": 4.418305397033691,
+ "step": 1589
+ },
+ {
+ "epoch": 22.083842794759825,
+ "grad_norm": 0.019051678478717804,
+ "learning_rate": 0.0006,
+ "loss": 4.215126991271973,
+ "step": 1590
+ },
+ {
+ "epoch": 22.097816593886463,
+ "grad_norm": 0.02323366142809391,
+ "learning_rate": 0.0006,
+ "loss": 4.306807994842529,
+ "step": 1591
+ },
+ {
+ "epoch": 22.111790393013102,
+ "grad_norm": 0.026186738163232803,
+ "learning_rate": 0.0006,
+ "loss": 4.408508777618408,
+ "step": 1592
+ },
+ {
+ "epoch": 22.125764192139737,
+ "grad_norm": 0.027361227199435234,
+ "learning_rate": 0.0006,
+ "loss": 4.329322814941406,
+ "step": 1593
+ },
+ {
+ "epoch": 22.139737991266376,
+ "grad_norm": 0.026564646512269974,
+ "learning_rate": 0.0006,
+ "loss": 4.346642017364502,
+ "step": 1594
+ },
+ {
+ "epoch": 22.153711790393015,
+ "grad_norm": 0.025073667988181114,
+ "learning_rate": 0.0006,
+ "loss": 4.3984198570251465,
+ "step": 1595
+ },
+ {
+ "epoch": 22.16768558951965,
+ "grad_norm": 0.02347794733941555,
+ "learning_rate": 0.0006,
+ "loss": 4.352538108825684,
+ "step": 1596
+ },
+ {
+ "epoch": 22.18165938864629,
+ "grad_norm": 0.02203468047082424,
+ "learning_rate": 0.0006,
+ "loss": 4.298148155212402,
+ "step": 1597
+ },
+ {
+ "epoch": 22.195633187772927,
+ "grad_norm": 0.020621957257390022,
+ "learning_rate": 0.0006,
+ "loss": 4.480309963226318,
+ "step": 1598
+ },
+ {
+ "epoch": 22.209606986899562,
+ "grad_norm": 0.023710055276751518,
+ "learning_rate": 0.0006,
+ "loss": 4.339317321777344,
+ "step": 1599
+ },
+ {
+ "epoch": 22.2235807860262,
+ "grad_norm": 0.0233482476323843,
+ "learning_rate": 0.0006,
+ "loss": 4.436896324157715,
+ "step": 1600
+ },
+ {
+ "epoch": 22.23755458515284,
+ "grad_norm": 0.020122403278946877,
+ "learning_rate": 0.0006,
+ "loss": 4.236002445220947,
+ "step": 1601
+ },
+ {
+ "epoch": 22.251528384279474,
+ "grad_norm": 0.01957266591489315,
+ "learning_rate": 0.0006,
+ "loss": 4.497328281402588,
+ "step": 1602
+ },
+ {
+ "epoch": 22.265502183406113,
+ "grad_norm": 0.018964895978569984,
+ "learning_rate": 0.0006,
+ "loss": 4.36360502243042,
+ "step": 1603
+ },
+ {
+ "epoch": 22.27947598253275,
+ "grad_norm": 0.018902769312262535,
+ "learning_rate": 0.0006,
+ "loss": 4.427093505859375,
+ "step": 1604
+ },
+ {
+ "epoch": 22.29344978165939,
+ "grad_norm": 0.018918119370937347,
+ "learning_rate": 0.0006,
+ "loss": 4.294950485229492,
+ "step": 1605
+ },
+ {
+ "epoch": 22.307423580786025,
+ "grad_norm": 0.019315775483846664,
+ "learning_rate": 0.0006,
+ "loss": 4.427952289581299,
+ "step": 1606
+ },
+ {
+ "epoch": 22.321397379912664,
+ "grad_norm": 0.018350522965192795,
+ "learning_rate": 0.0006,
+ "loss": 4.272945404052734,
+ "step": 1607
+ },
+ {
+ "epoch": 22.335371179039303,
+ "grad_norm": 0.01759534701704979,
+ "learning_rate": 0.0006,
+ "loss": 4.348023891448975,
+ "step": 1608
+ },
+ {
+ "epoch": 22.349344978165938,
+ "grad_norm": 0.01646057888865471,
+ "learning_rate": 0.0006,
+ "loss": 4.278518199920654,
+ "step": 1609
+ },
+ {
+ "epoch": 22.363318777292577,
+ "grad_norm": 0.015875060111284256,
+ "learning_rate": 0.0006,
+ "loss": 4.423527240753174,
+ "step": 1610
+ },
+ {
+ "epoch": 22.377292576419215,
+ "grad_norm": 0.01638529635965824,
+ "learning_rate": 0.0006,
+ "loss": 4.321723937988281,
+ "step": 1611
+ },
+ {
+ "epoch": 22.39126637554585,
+ "grad_norm": 0.016942061483860016,
+ "learning_rate": 0.0006,
+ "loss": 4.346164703369141,
+ "step": 1612
+ },
+ {
+ "epoch": 22.40524017467249,
+ "grad_norm": 0.015209621749818325,
+ "learning_rate": 0.0006,
+ "loss": 4.434340000152588,
+ "step": 1613
+ },
+ {
+ "epoch": 22.419213973799128,
+ "grad_norm": 0.014423009008169174,
+ "learning_rate": 0.0006,
+ "loss": 4.342560768127441,
+ "step": 1614
+ },
+ {
+ "epoch": 22.433187772925763,
+ "grad_norm": 0.014369525946676731,
+ "learning_rate": 0.0006,
+ "loss": 4.296586990356445,
+ "step": 1615
+ },
+ {
+ "epoch": 22.4471615720524,
+ "grad_norm": 0.01392540242522955,
+ "learning_rate": 0.0006,
+ "loss": 4.433176040649414,
+ "step": 1616
+ },
+ {
+ "epoch": 22.46113537117904,
+ "grad_norm": 0.013206899166107178,
+ "learning_rate": 0.0006,
+ "loss": 4.32962703704834,
+ "step": 1617
+ },
+ {
+ "epoch": 22.475109170305675,
+ "grad_norm": 0.013427077792584896,
+ "learning_rate": 0.0006,
+ "loss": 4.420437335968018,
+ "step": 1618
+ },
+ {
+ "epoch": 22.489082969432314,
+ "grad_norm": 0.014194196090102196,
+ "learning_rate": 0.0006,
+ "loss": 4.410478591918945,
+ "step": 1619
+ },
+ {
+ "epoch": 22.503056768558952,
+ "grad_norm": 0.01362364087253809,
+ "learning_rate": 0.0006,
+ "loss": 4.338932037353516,
+ "step": 1620
+ },
+ {
+ "epoch": 22.51703056768559,
+ "grad_norm": 0.014299589209258556,
+ "learning_rate": 0.0006,
+ "loss": 4.34089994430542,
+ "step": 1621
+ },
+ {
+ "epoch": 22.531004366812226,
+ "grad_norm": 0.014061295427381992,
+ "learning_rate": 0.0006,
+ "loss": 4.29201602935791,
+ "step": 1622
+ },
+ {
+ "epoch": 22.544978165938865,
+ "grad_norm": 0.013945229351520538,
+ "learning_rate": 0.0006,
+ "loss": 4.326052188873291,
+ "step": 1623
+ },
+ {
+ "epoch": 22.558951965065503,
+ "grad_norm": 0.01387068536132574,
+ "learning_rate": 0.0006,
+ "loss": 4.393521308898926,
+ "step": 1624
+ },
+ {
+ "epoch": 22.57292576419214,
+ "grad_norm": 0.013931898400187492,
+ "learning_rate": 0.0006,
+ "loss": 4.3739752769470215,
+ "step": 1625
+ },
+ {
+ "epoch": 22.586899563318777,
+ "grad_norm": 0.01497740764170885,
+ "learning_rate": 0.0006,
+ "loss": 4.3991618156433105,
+ "step": 1626
+ },
+ {
+ "epoch": 22.600873362445416,
+ "grad_norm": 0.015842905268073082,
+ "learning_rate": 0.0006,
+ "loss": 4.405340671539307,
+ "step": 1627
+ },
+ {
+ "epoch": 22.61484716157205,
+ "grad_norm": 0.01580633595585823,
+ "learning_rate": 0.0006,
+ "loss": 4.280102729797363,
+ "step": 1628
+ },
+ {
+ "epoch": 22.62882096069869,
+ "grad_norm": 0.016567382961511612,
+ "learning_rate": 0.0006,
+ "loss": 4.407657146453857,
+ "step": 1629
+ },
+ {
+ "epoch": 22.64279475982533,
+ "grad_norm": 0.01648622378706932,
+ "learning_rate": 0.0006,
+ "loss": 4.392590522766113,
+ "step": 1630
+ },
+ {
+ "epoch": 22.656768558951963,
+ "grad_norm": 0.017066307365894318,
+ "learning_rate": 0.0006,
+ "loss": 4.324862003326416,
+ "step": 1631
+ },
+ {
+ "epoch": 22.670742358078602,
+ "grad_norm": 0.01637062057852745,
+ "learning_rate": 0.0006,
+ "loss": 4.34598445892334,
+ "step": 1632
+ },
+ {
+ "epoch": 22.68471615720524,
+ "grad_norm": 0.018038960173726082,
+ "learning_rate": 0.0006,
+ "loss": 4.320711612701416,
+ "step": 1633
+ },
+ {
+ "epoch": 22.69868995633188,
+ "grad_norm": 0.018096428364515305,
+ "learning_rate": 0.0006,
+ "loss": 4.536527156829834,
+ "step": 1634
+ },
+ {
+ "epoch": 22.712663755458514,
+ "grad_norm": 0.017889291048049927,
+ "learning_rate": 0.0006,
+ "loss": 4.388131141662598,
+ "step": 1635
+ },
+ {
+ "epoch": 22.726637554585153,
+ "grad_norm": 0.018771521747112274,
+ "learning_rate": 0.0006,
+ "loss": 4.357668876647949,
+ "step": 1636
+ },
+ {
+ "epoch": 22.74061135371179,
+ "grad_norm": 0.02100873365998268,
+ "learning_rate": 0.0006,
+ "loss": 4.293972015380859,
+ "step": 1637
+ },
+ {
+ "epoch": 22.754585152838427,
+ "grad_norm": 0.017715785652399063,
+ "learning_rate": 0.0006,
+ "loss": 4.349849224090576,
+ "step": 1638
+ },
+ {
+ "epoch": 22.768558951965066,
+ "grad_norm": 0.015546333976089954,
+ "learning_rate": 0.0006,
+ "loss": 4.311728000640869,
+ "step": 1639
+ },
+ {
+ "epoch": 22.782532751091704,
+ "grad_norm": 0.01605990156531334,
+ "learning_rate": 0.0006,
+ "loss": 4.499123573303223,
+ "step": 1640
+ },
+ {
+ "epoch": 22.79650655021834,
+ "grad_norm": 0.016642142087221146,
+ "learning_rate": 0.0006,
+ "loss": 4.4319257736206055,
+ "step": 1641
+ },
+ {
+ "epoch": 22.810480349344978,
+ "grad_norm": 0.016120387241244316,
+ "learning_rate": 0.0006,
+ "loss": 4.4373650550842285,
+ "step": 1642
+ },
+ {
+ "epoch": 22.824454148471617,
+ "grad_norm": 0.015174890868365765,
+ "learning_rate": 0.0006,
+ "loss": 4.314082145690918,
+ "step": 1643
+ },
+ {
+ "epoch": 22.83842794759825,
+ "grad_norm": 0.01551714539527893,
+ "learning_rate": 0.0006,
+ "loss": 4.360363960266113,
+ "step": 1644
+ },
+ {
+ "epoch": 22.85240174672489,
+ "grad_norm": 0.014966742135584354,
+ "learning_rate": 0.0006,
+ "loss": 4.346253395080566,
+ "step": 1645
+ },
+ {
+ "epoch": 22.86637554585153,
+ "grad_norm": 0.014288028702139854,
+ "learning_rate": 0.0006,
+ "loss": 4.464366436004639,
+ "step": 1646
+ },
+ {
+ "epoch": 22.880349344978168,
+ "grad_norm": 0.015291682444512844,
+ "learning_rate": 0.0006,
+ "loss": 4.504101753234863,
+ "step": 1647
+ },
+ {
+ "epoch": 22.894323144104803,
+ "grad_norm": 0.014534426853060722,
+ "learning_rate": 0.0006,
+ "loss": 4.462172031402588,
+ "step": 1648
+ },
+ {
+ "epoch": 22.90829694323144,
+ "grad_norm": 0.014129819348454475,
+ "learning_rate": 0.0006,
+ "loss": 4.350126266479492,
+ "step": 1649
+ },
+ {
+ "epoch": 22.92227074235808,
+ "grad_norm": 0.014932668767869473,
+ "learning_rate": 0.0006,
+ "loss": 4.374392032623291,
+ "step": 1650
+ },
+ {
+ "epoch": 22.936244541484715,
+ "grad_norm": 0.014647853560745716,
+ "learning_rate": 0.0006,
+ "loss": 4.252264499664307,
+ "step": 1651
+ },
+ {
+ "epoch": 22.950218340611354,
+ "grad_norm": 0.013553360477089882,
+ "learning_rate": 0.0006,
+ "loss": 4.379729270935059,
+ "step": 1652
+ },
+ {
+ "epoch": 22.964192139737992,
+ "grad_norm": 0.014606194570660591,
+ "learning_rate": 0.0006,
+ "loss": 4.363372802734375,
+ "step": 1653
+ },
+ {
+ "epoch": 22.978165938864628,
+ "grad_norm": 0.015000631101429462,
+ "learning_rate": 0.0006,
+ "loss": 4.408048629760742,
+ "step": 1654
+ },
+ {
+ "epoch": 22.992139737991266,
+ "grad_norm": 0.015370816923677921,
+ "learning_rate": 0.0006,
+ "loss": 4.297451019287109,
+ "step": 1655
+ },
+ {
+ "epoch": 23.0,
+ "grad_norm": 0.01709144562482834,
+ "learning_rate": 0.0006,
+ "loss": 4.294776916503906,
+ "step": 1656
+ },
+ {
+ "epoch": 23.0,
+ "eval_loss": 4.678618907928467,
+ "eval_runtime": 56.1668,
+ "eval_samples_per_second": 43.478,
+ "eval_steps_per_second": 1.371,
+ "step": 1656
+ },
+ {
+ "epoch": 23.01397379912664,
+ "grad_norm": 0.01727290451526642,
+ "learning_rate": 0.0006,
+ "loss": 4.311701774597168,
+ "step": 1657
+ },
+ {
+ "epoch": 23.027947598253274,
+ "grad_norm": 0.020645759999752045,
+ "learning_rate": 0.0006,
+ "loss": 4.429897308349609,
+ "step": 1658
+ },
+ {
+ "epoch": 23.041921397379912,
+ "grad_norm": 0.02204226516187191,
+ "learning_rate": 0.0006,
+ "loss": 4.365151882171631,
+ "step": 1659
+ },
+ {
+ "epoch": 23.05589519650655,
+ "grad_norm": 0.023069269955158234,
+ "learning_rate": 0.0006,
+ "loss": 4.3710832595825195,
+ "step": 1660
+ },
+ {
+ "epoch": 23.069868995633186,
+ "grad_norm": 0.024151179939508438,
+ "learning_rate": 0.0006,
+ "loss": 4.403108596801758,
+ "step": 1661
+ },
+ {
+ "epoch": 23.083842794759825,
+ "grad_norm": 0.02511233650147915,
+ "learning_rate": 0.0006,
+ "loss": 4.414128303527832,
+ "step": 1662
+ },
+ {
+ "epoch": 23.097816593886463,
+ "grad_norm": 0.025802113115787506,
+ "learning_rate": 0.0006,
+ "loss": 4.388265609741211,
+ "step": 1663
+ },
+ {
+ "epoch": 23.111790393013102,
+ "grad_norm": 0.02377348765730858,
+ "learning_rate": 0.0006,
+ "loss": 4.357505798339844,
+ "step": 1664
+ },
+ {
+ "epoch": 23.125764192139737,
+ "grad_norm": 0.020876651629805565,
+ "learning_rate": 0.0006,
+ "loss": 4.23043966293335,
+ "step": 1665
+ },
+ {
+ "epoch": 23.139737991266376,
+ "grad_norm": 0.01959727331995964,
+ "learning_rate": 0.0006,
+ "loss": 4.280970096588135,
+ "step": 1666
+ },
+ {
+ "epoch": 23.153711790393015,
+ "grad_norm": 0.018678616732358932,
+ "learning_rate": 0.0006,
+ "loss": 4.346716403961182,
+ "step": 1667
+ },
+ {
+ "epoch": 23.16768558951965,
+ "grad_norm": 0.02108944021165371,
+ "learning_rate": 0.0006,
+ "loss": 4.277904510498047,
+ "step": 1668
+ },
+ {
+ "epoch": 23.18165938864629,
+ "grad_norm": 0.020221566781401634,
+ "learning_rate": 0.0006,
+ "loss": 4.367693901062012,
+ "step": 1669
+ },
+ {
+ "epoch": 23.195633187772927,
+ "grad_norm": 0.019599711522459984,
+ "learning_rate": 0.0006,
+ "loss": 4.257092475891113,
+ "step": 1670
+ },
+ {
+ "epoch": 23.209606986899562,
+ "grad_norm": 0.019833361729979515,
+ "learning_rate": 0.0006,
+ "loss": 4.540737628936768,
+ "step": 1671
+ },
+ {
+ "epoch": 23.2235807860262,
+ "grad_norm": 0.020352095365524292,
+ "learning_rate": 0.0006,
+ "loss": 4.360923767089844,
+ "step": 1672
+ },
+ {
+ "epoch": 23.23755458515284,
+ "grad_norm": 0.01939748041331768,
+ "learning_rate": 0.0006,
+ "loss": 4.393110275268555,
+ "step": 1673
+ },
+ {
+ "epoch": 23.251528384279474,
+ "grad_norm": 0.01884354278445244,
+ "learning_rate": 0.0006,
+ "loss": 4.383938789367676,
+ "step": 1674
+ },
+ {
+ "epoch": 23.265502183406113,
+ "grad_norm": 0.02008470520377159,
+ "learning_rate": 0.0006,
+ "loss": 4.346344947814941,
+ "step": 1675
+ },
+ {
+ "epoch": 23.27947598253275,
+ "grad_norm": 0.019065195694565773,
+ "learning_rate": 0.0006,
+ "loss": 4.36611795425415,
+ "step": 1676
+ },
+ {
+ "epoch": 23.29344978165939,
+ "grad_norm": 0.01938541606068611,
+ "learning_rate": 0.0006,
+ "loss": 4.4009599685668945,
+ "step": 1677
+ },
+ {
+ "epoch": 23.307423580786025,
+ "grad_norm": 0.019044723361730576,
+ "learning_rate": 0.0006,
+ "loss": 4.340169906616211,
+ "step": 1678
+ },
+ {
+ "epoch": 23.321397379912664,
+ "grad_norm": 0.018850378692150116,
+ "learning_rate": 0.0006,
+ "loss": 4.37674617767334,
+ "step": 1679
+ },
+ {
+ "epoch": 23.335371179039303,
+ "grad_norm": 0.019076095893979073,
+ "learning_rate": 0.0006,
+ "loss": 4.226874351501465,
+ "step": 1680
+ },
+ {
+ "epoch": 23.349344978165938,
+ "grad_norm": 0.017543038353323936,
+ "learning_rate": 0.0006,
+ "loss": 4.337751388549805,
+ "step": 1681
+ },
+ {
+ "epoch": 23.363318777292577,
+ "grad_norm": 0.017170218750834465,
+ "learning_rate": 0.0006,
+ "loss": 4.282512664794922,
+ "step": 1682
+ },
+ {
+ "epoch": 23.377292576419215,
+ "grad_norm": 0.01574796624481678,
+ "learning_rate": 0.0006,
+ "loss": 4.30638313293457,
+ "step": 1683
+ },
+ {
+ "epoch": 23.39126637554585,
+ "grad_norm": 0.015160846523940563,
+ "learning_rate": 0.0006,
+ "loss": 4.336716175079346,
+ "step": 1684
+ },
+ {
+ "epoch": 23.40524017467249,
+ "grad_norm": 0.01534526702016592,
+ "learning_rate": 0.0006,
+ "loss": 4.2174787521362305,
+ "step": 1685
+ },
+ {
+ "epoch": 23.419213973799128,
+ "grad_norm": 0.015289999544620514,
+ "learning_rate": 0.0006,
+ "loss": 4.272546291351318,
+ "step": 1686
+ },
+ {
+ "epoch": 23.433187772925763,
+ "grad_norm": 0.014894185587763786,
+ "learning_rate": 0.0006,
+ "loss": 4.296045303344727,
+ "step": 1687
+ },
+ {
+ "epoch": 23.4471615720524,
+ "grad_norm": 0.014869502745568752,
+ "learning_rate": 0.0006,
+ "loss": 4.425461769104004,
+ "step": 1688
+ },
+ {
+ "epoch": 23.46113537117904,
+ "grad_norm": 0.01474483497440815,
+ "learning_rate": 0.0006,
+ "loss": 4.361867427825928,
+ "step": 1689
+ },
+ {
+ "epoch": 23.475109170305675,
+ "grad_norm": 0.015299245715141296,
+ "learning_rate": 0.0006,
+ "loss": 4.364523410797119,
+ "step": 1690
+ },
+ {
+ "epoch": 23.489082969432314,
+ "grad_norm": 0.015201340429484844,
+ "learning_rate": 0.0006,
+ "loss": 4.285182952880859,
+ "step": 1691
+ },
+ {
+ "epoch": 23.503056768558952,
+ "grad_norm": 0.016057293862104416,
+ "learning_rate": 0.0006,
+ "loss": 4.284873962402344,
+ "step": 1692
+ },
+ {
+ "epoch": 23.51703056768559,
+ "grad_norm": 0.015871062874794006,
+ "learning_rate": 0.0006,
+ "loss": 4.326470851898193,
+ "step": 1693
+ },
+ {
+ "epoch": 23.531004366812226,
+ "grad_norm": 0.01586288772523403,
+ "learning_rate": 0.0006,
+ "loss": 4.324883460998535,
+ "step": 1694
+ },
+ {
+ "epoch": 23.544978165938865,
+ "grad_norm": 0.016092827543616295,
+ "learning_rate": 0.0006,
+ "loss": 4.3258867263793945,
+ "step": 1695
+ },
+ {
+ "epoch": 23.558951965065503,
+ "grad_norm": 0.01582016795873642,
+ "learning_rate": 0.0006,
+ "loss": 4.358892440795898,
+ "step": 1696
+ },
+ {
+ "epoch": 23.57292576419214,
+ "grad_norm": 0.01688452623784542,
+ "learning_rate": 0.0006,
+ "loss": 4.280842304229736,
+ "step": 1697
+ },
+ {
+ "epoch": 23.586899563318777,
+ "grad_norm": 0.015865886583924294,
+ "learning_rate": 0.0006,
+ "loss": 4.245567321777344,
+ "step": 1698
+ },
+ {
+ "epoch": 23.600873362445416,
+ "grad_norm": 0.015303662046790123,
+ "learning_rate": 0.0006,
+ "loss": 4.383112907409668,
+ "step": 1699
+ },
+ {
+ "epoch": 23.61484716157205,
+ "grad_norm": 0.015649516135454178,
+ "learning_rate": 0.0006,
+ "loss": 4.519082069396973,
+ "step": 1700
+ },
+ {
+ "epoch": 23.62882096069869,
+ "grad_norm": 0.0167120099067688,
+ "learning_rate": 0.0006,
+ "loss": 4.447847366333008,
+ "step": 1701
+ },
+ {
+ "epoch": 23.64279475982533,
+ "grad_norm": 0.016368716955184937,
+ "learning_rate": 0.0006,
+ "loss": 4.434186935424805,
+ "step": 1702
+ },
+ {
+ "epoch": 23.656768558951963,
+ "grad_norm": 0.017379503697156906,
+ "learning_rate": 0.0006,
+ "loss": 4.317594528198242,
+ "step": 1703
+ },
+ {
+ "epoch": 23.670742358078602,
+ "grad_norm": 0.018573811277747154,
+ "learning_rate": 0.0006,
+ "loss": 4.205793857574463,
+ "step": 1704
+ },
+ {
+ "epoch": 23.68471615720524,
+ "grad_norm": 0.01749090477824211,
+ "learning_rate": 0.0006,
+ "loss": 4.3985700607299805,
+ "step": 1705
+ },
+ {
+ "epoch": 23.69868995633188,
+ "grad_norm": 0.0165668074041605,
+ "learning_rate": 0.0006,
+ "loss": 4.329226970672607,
+ "step": 1706
+ },
+ {
+ "epoch": 23.712663755458514,
+ "grad_norm": 0.016913846135139465,
+ "learning_rate": 0.0006,
+ "loss": 4.388485908508301,
+ "step": 1707
+ },
+ {
+ "epoch": 23.726637554585153,
+ "grad_norm": 0.016813859343528748,
+ "learning_rate": 0.0006,
+ "loss": 4.3395915031433105,
+ "step": 1708
+ },
+ {
+ "epoch": 23.74061135371179,
+ "grad_norm": 0.01676975190639496,
+ "learning_rate": 0.0006,
+ "loss": 4.254745960235596,
+ "step": 1709
+ },
+ {
+ "epoch": 23.754585152838427,
+ "grad_norm": 0.016370350494980812,
+ "learning_rate": 0.0006,
+ "loss": 4.2612223625183105,
+ "step": 1710
+ },
+ {
+ "epoch": 23.768558951965066,
+ "grad_norm": 0.01696198247373104,
+ "learning_rate": 0.0006,
+ "loss": 4.327524662017822,
+ "step": 1711
+ },
+ {
+ "epoch": 23.782532751091704,
+ "grad_norm": 0.0189108457416296,
+ "learning_rate": 0.0006,
+ "loss": 4.325839519500732,
+ "step": 1712
+ },
+ {
+ "epoch": 23.79650655021834,
+ "grad_norm": 0.01862351782619953,
+ "learning_rate": 0.0006,
+ "loss": 4.333662509918213,
+ "step": 1713
+ },
+ {
+ "epoch": 23.810480349344978,
+ "grad_norm": 0.01655014418065548,
+ "learning_rate": 0.0006,
+ "loss": 4.39305305480957,
+ "step": 1714
+ },
+ {
+ "epoch": 23.824454148471617,
+ "grad_norm": 0.015899376943707466,
+ "learning_rate": 0.0006,
+ "loss": 4.390053749084473,
+ "step": 1715
+ },
+ {
+ "epoch": 23.83842794759825,
+ "grad_norm": 0.016519056633114815,
+ "learning_rate": 0.0006,
+ "loss": 4.255903244018555,
+ "step": 1716
+ },
+ {
+ "epoch": 23.85240174672489,
+ "grad_norm": 0.0167539119720459,
+ "learning_rate": 0.0006,
+ "loss": 4.255727767944336,
+ "step": 1717
+ },
+ {
+ "epoch": 23.86637554585153,
+ "grad_norm": 0.0171353816986084,
+ "learning_rate": 0.0006,
+ "loss": 4.376740455627441,
+ "step": 1718
+ },
+ {
+ "epoch": 23.880349344978168,
+ "grad_norm": 0.016719117760658264,
+ "learning_rate": 0.0006,
+ "loss": 4.3376688957214355,
+ "step": 1719
+ },
+ {
+ "epoch": 23.894323144104803,
+ "grad_norm": 0.01561494916677475,
+ "learning_rate": 0.0006,
+ "loss": 4.3345746994018555,
+ "step": 1720
+ },
+ {
+ "epoch": 23.90829694323144,
+ "grad_norm": 0.016303174197673798,
+ "learning_rate": 0.0006,
+ "loss": 4.332371234893799,
+ "step": 1721
+ },
+ {
+ "epoch": 23.92227074235808,
+ "grad_norm": 0.016722865402698517,
+ "learning_rate": 0.0006,
+ "loss": 4.267125129699707,
+ "step": 1722
+ },
+ {
+ "epoch": 23.936244541484715,
+ "grad_norm": 0.017072124406695366,
+ "learning_rate": 0.0006,
+ "loss": 4.322751998901367,
+ "step": 1723
+ },
+ {
+ "epoch": 23.950218340611354,
+ "grad_norm": 0.01699932850897312,
+ "learning_rate": 0.0006,
+ "loss": 4.378422737121582,
+ "step": 1724
+ },
+ {
+ "epoch": 23.964192139737992,
+ "grad_norm": 0.01742720976471901,
+ "learning_rate": 0.0006,
+ "loss": 4.345083236694336,
+ "step": 1725
+ },
+ {
+ "epoch": 23.978165938864628,
+ "grad_norm": 0.018797501921653748,
+ "learning_rate": 0.0006,
+ "loss": 4.290443420410156,
+ "step": 1726
+ },
+ {
+ "epoch": 23.992139737991266,
+ "grad_norm": 0.018090683966875076,
+ "learning_rate": 0.0006,
+ "loss": 4.377580642700195,
+ "step": 1727
+ },
+ {
+ "epoch": 24.0,
+ "grad_norm": 0.017305459827184677,
+ "learning_rate": 0.0006,
+ "loss": 4.364426612854004,
+ "step": 1728
+ },
+ {
+ "epoch": 24.0,
+ "eval_loss": 4.647762298583984,
+ "eval_runtime": 56.5402,
+ "eval_samples_per_second": 43.191,
+ "eval_steps_per_second": 1.362,
+ "step": 1728
+ },
+ {
+ "epoch": 24.01397379912664,
+ "grad_norm": 0.016493607312440872,
+ "learning_rate": 0.0006,
+ "loss": 4.252255916595459,
+ "step": 1729
+ },
+ {
+ "epoch": 24.027947598253274,
+ "grad_norm": 0.016529450193047523,
+ "learning_rate": 0.0006,
+ "loss": 4.352746963500977,
+ "step": 1730
+ },
+ {
+ "epoch": 24.041921397379912,
+ "grad_norm": 0.016940191388130188,
+ "learning_rate": 0.0006,
+ "loss": 4.301870822906494,
+ "step": 1731
+ },
+ {
+ "epoch": 24.05589519650655,
+ "grad_norm": 0.01752558909356594,
+ "learning_rate": 0.0006,
+ "loss": 4.2762451171875,
+ "step": 1732
+ },
+ {
+ "epoch": 24.069868995633186,
+ "grad_norm": 0.017112884670495987,
+ "learning_rate": 0.0006,
+ "loss": 4.2659149169921875,
+ "step": 1733
+ },
+ {
+ "epoch": 24.083842794759825,
+ "grad_norm": 0.01900586113333702,
+ "learning_rate": 0.0006,
+ "loss": 4.306058883666992,
+ "step": 1734
+ },
+ {
+ "epoch": 24.097816593886463,
+ "grad_norm": 0.019447680562734604,
+ "learning_rate": 0.0006,
+ "loss": 4.450128555297852,
+ "step": 1735
+ },
+ {
+ "epoch": 24.111790393013102,
+ "grad_norm": 0.01948186755180359,
+ "learning_rate": 0.0006,
+ "loss": 4.202869415283203,
+ "step": 1736
+ },
+ {
+ "epoch": 24.125764192139737,
+ "grad_norm": 0.01716383546590805,
+ "learning_rate": 0.0006,
+ "loss": 4.318876266479492,
+ "step": 1737
+ },
+ {
+ "epoch": 24.139737991266376,
+ "grad_norm": 0.017049988731741905,
+ "learning_rate": 0.0006,
+ "loss": 4.237238883972168,
+ "step": 1738
+ },
+ {
+ "epoch": 24.153711790393015,
+ "grad_norm": 0.015169923193752766,
+ "learning_rate": 0.0006,
+ "loss": 4.289426803588867,
+ "step": 1739
+ },
+ {
+ "epoch": 24.16768558951965,
+ "grad_norm": 0.014760667458176613,
+ "learning_rate": 0.0006,
+ "loss": 4.356924057006836,
+ "step": 1740
+ },
+ {
+ "epoch": 24.18165938864629,
+ "grad_norm": 0.015127750113606453,
+ "learning_rate": 0.0006,
+ "loss": 4.142916679382324,
+ "step": 1741
+ },
+ {
+ "epoch": 24.195633187772927,
+ "grad_norm": 0.01549555640667677,
+ "learning_rate": 0.0006,
+ "loss": 4.345563888549805,
+ "step": 1742
+ },
+ {
+ "epoch": 24.209606986899562,
+ "grad_norm": 0.016393091529607773,
+ "learning_rate": 0.0006,
+ "loss": 4.417495250701904,
+ "step": 1743
+ },
+ {
+ "epoch": 24.2235807860262,
+ "grad_norm": 0.01785266026854515,
+ "learning_rate": 0.0006,
+ "loss": 4.3553547859191895,
+ "step": 1744
+ },
+ {
+ "epoch": 24.23755458515284,
+ "grad_norm": 0.017574617639183998,
+ "learning_rate": 0.0006,
+ "loss": 4.241024494171143,
+ "step": 1745
+ },
+ {
+ "epoch": 24.251528384279474,
+ "grad_norm": 0.019060306251049042,
+ "learning_rate": 0.0006,
+ "loss": 4.3685150146484375,
+ "step": 1746
+ },
+ {
+ "epoch": 24.265502183406113,
+ "grad_norm": 0.0214633010327816,
+ "learning_rate": 0.0006,
+ "loss": 4.323866844177246,
+ "step": 1747
+ },
+ {
+ "epoch": 24.27947598253275,
+ "grad_norm": 0.021055003628134727,
+ "learning_rate": 0.0006,
+ "loss": 4.278757095336914,
+ "step": 1748
+ },
+ {
+ "epoch": 24.29344978165939,
+ "grad_norm": 0.018618572503328323,
+ "learning_rate": 0.0006,
+ "loss": 4.28849983215332,
+ "step": 1749
+ },
+ {
+ "epoch": 24.307423580786025,
+ "grad_norm": 0.018666435033082962,
+ "learning_rate": 0.0006,
+ "loss": 4.368590354919434,
+ "step": 1750
+ },
+ {
+ "epoch": 24.321397379912664,
+ "grad_norm": 0.020626146346330643,
+ "learning_rate": 0.0006,
+ "loss": 4.2951836585998535,
+ "step": 1751
+ },
+ {
+ "epoch": 24.335371179039303,
+ "grad_norm": 0.01929759792983532,
+ "learning_rate": 0.0006,
+ "loss": 4.3664703369140625,
+ "step": 1752
+ },
+ {
+ "epoch": 24.349344978165938,
+ "grad_norm": 0.020956002175807953,
+ "learning_rate": 0.0006,
+ "loss": 4.345308303833008,
+ "step": 1753
+ },
+ {
+ "epoch": 24.363318777292577,
+ "grad_norm": 0.019227957352995872,
+ "learning_rate": 0.0006,
+ "loss": 4.349849224090576,
+ "step": 1754
+ },
+ {
+ "epoch": 24.377292576419215,
+ "grad_norm": 0.015970097854733467,
+ "learning_rate": 0.0006,
+ "loss": 4.223012447357178,
+ "step": 1755
+ },
+ {
+ "epoch": 24.39126637554585,
+ "grad_norm": 0.015951646491885185,
+ "learning_rate": 0.0006,
+ "loss": 4.236420631408691,
+ "step": 1756
+ },
+ {
+ "epoch": 24.40524017467249,
+ "grad_norm": 0.016032515093684196,
+ "learning_rate": 0.0006,
+ "loss": 4.226899147033691,
+ "step": 1757
+ },
+ {
+ "epoch": 24.419213973799128,
+ "grad_norm": 0.01573132909834385,
+ "learning_rate": 0.0006,
+ "loss": 4.315916061401367,
+ "step": 1758
+ },
+ {
+ "epoch": 24.433187772925763,
+ "grad_norm": 0.015367005951702595,
+ "learning_rate": 0.0006,
+ "loss": 4.358939170837402,
+ "step": 1759
+ },
+ {
+ "epoch": 24.4471615720524,
+ "grad_norm": 0.01669907011091709,
+ "learning_rate": 0.0006,
+ "loss": 4.232147216796875,
+ "step": 1760
+ },
+ {
+ "epoch": 24.46113537117904,
+ "grad_norm": 0.018708152696490288,
+ "learning_rate": 0.0006,
+ "loss": 4.329852104187012,
+ "step": 1761
+ },
+ {
+ "epoch": 24.475109170305675,
+ "grad_norm": 0.019035987555980682,
+ "learning_rate": 0.0006,
+ "loss": 4.260340213775635,
+ "step": 1762
+ },
+ {
+ "epoch": 24.489082969432314,
+ "grad_norm": 0.01832144521176815,
+ "learning_rate": 0.0006,
+ "loss": 4.359543800354004,
+ "step": 1763
+ },
+ {
+ "epoch": 24.503056768558952,
+ "grad_norm": 0.015313294716179371,
+ "learning_rate": 0.0006,
+ "loss": 4.279088020324707,
+ "step": 1764
+ },
+ {
+ "epoch": 24.51703056768559,
+ "grad_norm": 0.015803923830389977,
+ "learning_rate": 0.0006,
+ "loss": 4.230682849884033,
+ "step": 1765
+ },
+ {
+ "epoch": 24.531004366812226,
+ "grad_norm": 0.016748374328017235,
+ "learning_rate": 0.0006,
+ "loss": 4.349661827087402,
+ "step": 1766
+ },
+ {
+ "epoch": 24.544978165938865,
+ "grad_norm": 0.015732428058981895,
+ "learning_rate": 0.0006,
+ "loss": 4.372441291809082,
+ "step": 1767
+ },
+ {
+ "epoch": 24.558951965065503,
+ "grad_norm": 0.016564829275012016,
+ "learning_rate": 0.0006,
+ "loss": 4.323090553283691,
+ "step": 1768
+ },
+ {
+ "epoch": 24.57292576419214,
+ "grad_norm": 0.01603010483086109,
+ "learning_rate": 0.0006,
+ "loss": 4.347898006439209,
+ "step": 1769
+ },
+ {
+ "epoch": 24.586899563318777,
+ "grad_norm": 0.01674809865653515,
+ "learning_rate": 0.0006,
+ "loss": 4.269165515899658,
+ "step": 1770
+ },
+ {
+ "epoch": 24.600873362445416,
+ "grad_norm": 0.017333930358290672,
+ "learning_rate": 0.0006,
+ "loss": 4.353298187255859,
+ "step": 1771
+ },
+ {
+ "epoch": 24.61484716157205,
+ "grad_norm": 0.017519580200314522,
+ "learning_rate": 0.0006,
+ "loss": 4.265199661254883,
+ "step": 1772
+ },
+ {
+ "epoch": 24.62882096069869,
+ "grad_norm": 0.017675306648015976,
+ "learning_rate": 0.0006,
+ "loss": 4.322596549987793,
+ "step": 1773
+ },
+ {
+ "epoch": 24.64279475982533,
+ "grad_norm": 0.017166534438729286,
+ "learning_rate": 0.0006,
+ "loss": 4.413900375366211,
+ "step": 1774
+ },
+ {
+ "epoch": 24.656768558951963,
+ "grad_norm": 0.016362346708774567,
+ "learning_rate": 0.0006,
+ "loss": 4.307236671447754,
+ "step": 1775
+ },
+ {
+ "epoch": 24.670742358078602,
+ "grad_norm": 0.016086872667074203,
+ "learning_rate": 0.0006,
+ "loss": 4.235316276550293,
+ "step": 1776
+ },
+ {
+ "epoch": 24.68471615720524,
+ "grad_norm": 0.01689537614583969,
+ "learning_rate": 0.0006,
+ "loss": 4.390591621398926,
+ "step": 1777
+ },
+ {
+ "epoch": 24.69868995633188,
+ "grad_norm": 0.016836725175380707,
+ "learning_rate": 0.0006,
+ "loss": 4.2990031242370605,
+ "step": 1778
+ },
+ {
+ "epoch": 24.712663755458514,
+ "grad_norm": 0.015627335757017136,
+ "learning_rate": 0.0006,
+ "loss": 4.166745185852051,
+ "step": 1779
+ },
+ {
+ "epoch": 24.726637554585153,
+ "grad_norm": 0.015007016249001026,
+ "learning_rate": 0.0006,
+ "loss": 4.295273303985596,
+ "step": 1780
+ },
+ {
+ "epoch": 24.74061135371179,
+ "grad_norm": 0.016084209084510803,
+ "learning_rate": 0.0006,
+ "loss": 4.342424392700195,
+ "step": 1781
+ },
+ {
+ "epoch": 24.754585152838427,
+ "grad_norm": 0.01595073565840721,
+ "learning_rate": 0.0006,
+ "loss": 4.3060102462768555,
+ "step": 1782
+ },
+ {
+ "epoch": 24.768558951965066,
+ "grad_norm": 0.016113461926579475,
+ "learning_rate": 0.0006,
+ "loss": 4.267908096313477,
+ "step": 1783
+ },
+ {
+ "epoch": 24.782532751091704,
+ "grad_norm": 0.01564556173980236,
+ "learning_rate": 0.0006,
+ "loss": 4.324642658233643,
+ "step": 1784
+ },
+ {
+ "epoch": 24.79650655021834,
+ "grad_norm": 0.01575268618762493,
+ "learning_rate": 0.0006,
+ "loss": 4.328181743621826,
+ "step": 1785
+ },
+ {
+ "epoch": 24.810480349344978,
+ "grad_norm": 0.016830824315547943,
+ "learning_rate": 0.0006,
+ "loss": 4.247906684875488,
+ "step": 1786
+ },
+ {
+ "epoch": 24.824454148471617,
+ "grad_norm": 0.017989547923207283,
+ "learning_rate": 0.0006,
+ "loss": 4.228281021118164,
+ "step": 1787
+ },
+ {
+ "epoch": 24.83842794759825,
+ "grad_norm": 0.016783086583018303,
+ "learning_rate": 0.0006,
+ "loss": 4.2880964279174805,
+ "step": 1788
+ },
+ {
+ "epoch": 24.85240174672489,
+ "grad_norm": 0.016467561945319176,
+ "learning_rate": 0.0006,
+ "loss": 4.36118745803833,
+ "step": 1789
+ },
+ {
+ "epoch": 24.86637554585153,
+ "grad_norm": 0.01780523732304573,
+ "learning_rate": 0.0006,
+ "loss": 4.172321319580078,
+ "step": 1790
+ },
+ {
+ "epoch": 24.880349344978168,
+ "grad_norm": 0.017102031037211418,
+ "learning_rate": 0.0006,
+ "loss": 4.352826118469238,
+ "step": 1791
+ },
+ {
+ "epoch": 24.894323144104803,
+ "grad_norm": 0.016409218311309814,
+ "learning_rate": 0.0006,
+ "loss": 4.212162017822266,
+ "step": 1792
+ },
+ {
+ "epoch": 24.90829694323144,
+ "grad_norm": 0.018660584464669228,
+ "learning_rate": 0.0006,
+ "loss": 4.346684455871582,
+ "step": 1793
+ },
+ {
+ "epoch": 24.92227074235808,
+ "grad_norm": 0.017389440909028053,
+ "learning_rate": 0.0006,
+ "loss": 4.197913646697998,
+ "step": 1794
+ },
+ {
+ "epoch": 24.936244541484715,
+ "grad_norm": 0.01596485823392868,
+ "learning_rate": 0.0006,
+ "loss": 4.370617866516113,
+ "step": 1795
+ },
+ {
+ "epoch": 24.950218340611354,
+ "grad_norm": 0.01709776371717453,
+ "learning_rate": 0.0006,
+ "loss": 4.313847064971924,
+ "step": 1796
+ },
+ {
+ "epoch": 24.964192139737992,
+ "grad_norm": 0.016571134328842163,
+ "learning_rate": 0.0006,
+ "loss": 4.1895856857299805,
+ "step": 1797
+ },
+ {
+ "epoch": 24.978165938864628,
+ "grad_norm": 0.016851048916578293,
+ "learning_rate": 0.0006,
+ "loss": 4.350772857666016,
+ "step": 1798
+ },
+ {
+ "epoch": 24.992139737991266,
+ "grad_norm": 0.017007440328598022,
+ "learning_rate": 0.0006,
+ "loss": 4.399008750915527,
+ "step": 1799
+ },
+ {
+ "epoch": 25.0,
+ "grad_norm": 0.01791212148964405,
+ "learning_rate": 0.0006,
+ "loss": 4.306121826171875,
+ "step": 1800
+ },
+ {
+ "epoch": 25.0,
+ "eval_loss": 4.636417865753174,
+ "eval_runtime": 56.8296,
+ "eval_samples_per_second": 42.971,
+ "eval_steps_per_second": 1.355,
+ "step": 1800
+ },
+ {
+ "epoch": 25.01397379912664,
+ "grad_norm": 0.01729617640376091,
+ "learning_rate": 0.0006,
+ "loss": 4.257018089294434,
+ "step": 1801
+ },
+ {
+ "epoch": 25.027947598253274,
+ "grad_norm": 0.017466643825173378,
+ "learning_rate": 0.0006,
+ "loss": 4.307605266571045,
+ "step": 1802
+ },
+ {
+ "epoch": 25.041921397379912,
+ "grad_norm": 0.018614279106259346,
+ "learning_rate": 0.0006,
+ "loss": 4.360141277313232,
+ "step": 1803
+ },
+ {
+ "epoch": 25.05589519650655,
+ "grad_norm": 0.02010771445930004,
+ "learning_rate": 0.0006,
+ "loss": 4.196428298950195,
+ "step": 1804
+ },
+ {
+ "epoch": 25.069868995633186,
+ "grad_norm": 0.02090228535234928,
+ "learning_rate": 0.0006,
+ "loss": 4.288947105407715,
+ "step": 1805
+ },
+ {
+ "epoch": 25.083842794759825,
+ "grad_norm": 0.02050255797803402,
+ "learning_rate": 0.0006,
+ "loss": 4.259582042694092,
+ "step": 1806
+ },
+ {
+ "epoch": 25.097816593886463,
+ "grad_norm": 0.021844089031219482,
+ "learning_rate": 0.0006,
+ "loss": 4.405138969421387,
+ "step": 1807
+ },
+ {
+ "epoch": 25.111790393013102,
+ "grad_norm": 0.022287018597126007,
+ "learning_rate": 0.0006,
+ "loss": 4.333793640136719,
+ "step": 1808
+ },
+ {
+ "epoch": 25.125764192139737,
+ "grad_norm": 0.02288329415023327,
+ "learning_rate": 0.0006,
+ "loss": 4.3910980224609375,
+ "step": 1809
+ },
+ {
+ "epoch": 25.139737991266376,
+ "grad_norm": 0.02373199723660946,
+ "learning_rate": 0.0006,
+ "loss": 4.32921028137207,
+ "step": 1810
+ },
+ {
+ "epoch": 25.153711790393015,
+ "grad_norm": 0.024960234761238098,
+ "learning_rate": 0.0006,
+ "loss": 4.295609474182129,
+ "step": 1811
+ },
+ {
+ "epoch": 25.16768558951965,
+ "grad_norm": 0.026743892580270767,
+ "learning_rate": 0.0006,
+ "loss": 4.315308094024658,
+ "step": 1812
+ },
+ {
+ "epoch": 25.18165938864629,
+ "grad_norm": 0.022138185799121857,
+ "learning_rate": 0.0006,
+ "loss": 4.360208511352539,
+ "step": 1813
+ },
+ {
+ "epoch": 25.195633187772927,
+ "grad_norm": 0.022120574489235878,
+ "learning_rate": 0.0006,
+ "loss": 4.1973772048950195,
+ "step": 1814
+ },
+ {
+ "epoch": 25.209606986899562,
+ "grad_norm": 0.02230502851307392,
+ "learning_rate": 0.0006,
+ "loss": 4.285558700561523,
+ "step": 1815
+ },
+ {
+ "epoch": 25.2235807860262,
+ "grad_norm": 0.022301986813545227,
+ "learning_rate": 0.0006,
+ "loss": 4.238343238830566,
+ "step": 1816
+ },
+ {
+ "epoch": 25.23755458515284,
+ "grad_norm": 0.020050447434186935,
+ "learning_rate": 0.0006,
+ "loss": 4.298235893249512,
+ "step": 1817
+ },
+ {
+ "epoch": 25.251528384279474,
+ "grad_norm": 0.01987724006175995,
+ "learning_rate": 0.0006,
+ "loss": 4.301384449005127,
+ "step": 1818
+ },
+ {
+ "epoch": 25.265502183406113,
+ "grad_norm": 0.020440855994820595,
+ "learning_rate": 0.0006,
+ "loss": 4.269218444824219,
+ "step": 1819
+ },
+ {
+ "epoch": 25.27947598253275,
+ "grad_norm": 0.01891893707215786,
+ "learning_rate": 0.0006,
+ "loss": 4.2401957511901855,
+ "step": 1820
+ },
+ {
+ "epoch": 25.29344978165939,
+ "grad_norm": 0.01903688907623291,
+ "learning_rate": 0.0006,
+ "loss": 4.139670372009277,
+ "step": 1821
+ },
+ {
+ "epoch": 25.307423580786025,
+ "grad_norm": 0.018856938928365707,
+ "learning_rate": 0.0006,
+ "loss": 4.233214378356934,
+ "step": 1822
+ },
+ {
+ "epoch": 25.321397379912664,
+ "grad_norm": 0.017791323363780975,
+ "learning_rate": 0.0006,
+ "loss": 4.2342023849487305,
+ "step": 1823
+ },
+ {
+ "epoch": 25.335371179039303,
+ "grad_norm": 0.017304565757513046,
+ "learning_rate": 0.0006,
+ "loss": 4.355198860168457,
+ "step": 1824
+ },
+ {
+ "epoch": 25.349344978165938,
+ "grad_norm": 0.01576206088066101,
+ "learning_rate": 0.0006,
+ "loss": 4.114147186279297,
+ "step": 1825
+ },
+ {
+ "epoch": 25.363318777292577,
+ "grad_norm": 0.015105154365301132,
+ "learning_rate": 0.0006,
+ "loss": 4.313795566558838,
+ "step": 1826
+ },
+ {
+ "epoch": 25.377292576419215,
+ "grad_norm": 0.01563587784767151,
+ "learning_rate": 0.0006,
+ "loss": 4.308282852172852,
+ "step": 1827
+ },
+ {
+ "epoch": 25.39126637554585,
+ "grad_norm": 0.015980906784534454,
+ "learning_rate": 0.0006,
+ "loss": 4.345149040222168,
+ "step": 1828
+ },
+ {
+ "epoch": 25.40524017467249,
+ "grad_norm": 0.01669284701347351,
+ "learning_rate": 0.0006,
+ "loss": 4.237611293792725,
+ "step": 1829
+ },
+ {
+ "epoch": 25.419213973799128,
+ "grad_norm": 0.016202064231038094,
+ "learning_rate": 0.0006,
+ "loss": 4.3072309494018555,
+ "step": 1830
+ },
+ {
+ "epoch": 25.433187772925763,
+ "grad_norm": 0.01478662807494402,
+ "learning_rate": 0.0006,
+ "loss": 4.276215076446533,
+ "step": 1831
+ },
+ {
+ "epoch": 25.4471615720524,
+ "grad_norm": 0.015168731100857258,
+ "learning_rate": 0.0006,
+ "loss": 4.37509822845459,
+ "step": 1832
+ },
+ {
+ "epoch": 25.46113537117904,
+ "grad_norm": 0.01648980937898159,
+ "learning_rate": 0.0006,
+ "loss": 4.349167346954346,
+ "step": 1833
+ },
+ {
+ "epoch": 25.475109170305675,
+ "grad_norm": 0.01708398200571537,
+ "learning_rate": 0.0006,
+ "loss": 4.346580505371094,
+ "step": 1834
+ },
+ {
+ "epoch": 25.489082969432314,
+ "grad_norm": 0.017370784655213356,
+ "learning_rate": 0.0006,
+ "loss": 4.299466133117676,
+ "step": 1835
+ },
+ {
+ "epoch": 25.503056768558952,
+ "grad_norm": 0.016248228028416634,
+ "learning_rate": 0.0006,
+ "loss": 4.184610366821289,
+ "step": 1836
+ },
+ {
+ "epoch": 25.51703056768559,
+ "grad_norm": 0.017106913030147552,
+ "learning_rate": 0.0006,
+ "loss": 4.366856098175049,
+ "step": 1837
+ },
+ {
+ "epoch": 25.531004366812226,
+ "grad_norm": 0.016071965917944908,
+ "learning_rate": 0.0006,
+ "loss": 4.297691345214844,
+ "step": 1838
+ },
+ {
+ "epoch": 25.544978165938865,
+ "grad_norm": 0.016451986506581306,
+ "learning_rate": 0.0006,
+ "loss": 4.229983329772949,
+ "step": 1839
+ },
+ {
+ "epoch": 25.558951965065503,
+ "grad_norm": 0.016195081174373627,
+ "learning_rate": 0.0006,
+ "loss": 4.259893417358398,
+ "step": 1840
+ },
+ {
+ "epoch": 25.57292576419214,
+ "grad_norm": 0.0165175162255764,
+ "learning_rate": 0.0006,
+ "loss": 4.3348236083984375,
+ "step": 1841
+ },
+ {
+ "epoch": 25.586899563318777,
+ "grad_norm": 0.01651517115533352,
+ "learning_rate": 0.0006,
+ "loss": 4.319541931152344,
+ "step": 1842
+ },
+ {
+ "epoch": 25.600873362445416,
+ "grad_norm": 0.016143690794706345,
+ "learning_rate": 0.0006,
+ "loss": 4.345652103424072,
+ "step": 1843
+ },
+ {
+ "epoch": 25.61484716157205,
+ "grad_norm": 0.015140696428716183,
+ "learning_rate": 0.0006,
+ "loss": 4.212047576904297,
+ "step": 1844
+ },
+ {
+ "epoch": 25.62882096069869,
+ "grad_norm": 0.014637443237006664,
+ "learning_rate": 0.0006,
+ "loss": 4.187453269958496,
+ "step": 1845
+ },
+ {
+ "epoch": 25.64279475982533,
+ "grad_norm": 0.014585614204406738,
+ "learning_rate": 0.0006,
+ "loss": 4.253936767578125,
+ "step": 1846
+ },
+ {
+ "epoch": 25.656768558951963,
+ "grad_norm": 0.015158289112150669,
+ "learning_rate": 0.0006,
+ "loss": 4.34612512588501,
+ "step": 1847
+ },
+ {
+ "epoch": 25.670742358078602,
+ "grad_norm": 0.015202849172055721,
+ "learning_rate": 0.0006,
+ "loss": 4.270634651184082,
+ "step": 1848
+ },
+ {
+ "epoch": 25.68471615720524,
+ "grad_norm": 0.015363575890660286,
+ "learning_rate": 0.0006,
+ "loss": 4.271080493927002,
+ "step": 1849
+ },
+ {
+ "epoch": 25.69868995633188,
+ "grad_norm": 0.014360226690769196,
+ "learning_rate": 0.0006,
+ "loss": 4.375033855438232,
+ "step": 1850
+ },
+ {
+ "epoch": 25.712663755458514,
+ "grad_norm": 0.014263181947171688,
+ "learning_rate": 0.0006,
+ "loss": 4.424138069152832,
+ "step": 1851
+ },
+ {
+ "epoch": 25.726637554585153,
+ "grad_norm": 0.014398688450455666,
+ "learning_rate": 0.0006,
+ "loss": 4.410154819488525,
+ "step": 1852
+ },
+ {
+ "epoch": 25.74061135371179,
+ "grad_norm": 0.016131488606333733,
+ "learning_rate": 0.0006,
+ "loss": 4.324014663696289,
+ "step": 1853
+ },
+ {
+ "epoch": 25.754585152838427,
+ "grad_norm": 0.017568735405802727,
+ "learning_rate": 0.0006,
+ "loss": 4.367861747741699,
+ "step": 1854
+ },
+ {
+ "epoch": 25.768558951965066,
+ "grad_norm": 0.017907511442899704,
+ "learning_rate": 0.0006,
+ "loss": 4.352965354919434,
+ "step": 1855
+ },
+ {
+ "epoch": 25.782532751091704,
+ "grad_norm": 0.016918016597628593,
+ "learning_rate": 0.0006,
+ "loss": 4.252397537231445,
+ "step": 1856
+ },
+ {
+ "epoch": 25.79650655021834,
+ "grad_norm": 0.017274610698223114,
+ "learning_rate": 0.0006,
+ "loss": 4.342309951782227,
+ "step": 1857
+ },
+ {
+ "epoch": 25.810480349344978,
+ "grad_norm": 0.017879825085401535,
+ "learning_rate": 0.0006,
+ "loss": 4.330987453460693,
+ "step": 1858
+ },
+ {
+ "epoch": 25.824454148471617,
+ "grad_norm": 0.01736099272966385,
+ "learning_rate": 0.0006,
+ "loss": 4.307255744934082,
+ "step": 1859
+ },
+ {
+ "epoch": 25.83842794759825,
+ "grad_norm": 0.016575824469327927,
+ "learning_rate": 0.0006,
+ "loss": 4.262016773223877,
+ "step": 1860
+ },
+ {
+ "epoch": 25.85240174672489,
+ "grad_norm": 0.015181140042841434,
+ "learning_rate": 0.0006,
+ "loss": 4.291863441467285,
+ "step": 1861
+ },
+ {
+ "epoch": 25.86637554585153,
+ "grad_norm": 0.015965687111020088,
+ "learning_rate": 0.0006,
+ "loss": 4.424836158752441,
+ "step": 1862
+ },
+ {
+ "epoch": 25.880349344978168,
+ "grad_norm": 0.01687219925224781,
+ "learning_rate": 0.0006,
+ "loss": 4.343531131744385,
+ "step": 1863
+ },
+ {
+ "epoch": 25.894323144104803,
+ "grad_norm": 0.015115504153072834,
+ "learning_rate": 0.0006,
+ "loss": 4.357808589935303,
+ "step": 1864
+ },
+ {
+ "epoch": 25.90829694323144,
+ "grad_norm": 0.014150998555123806,
+ "learning_rate": 0.0006,
+ "loss": 4.324550151824951,
+ "step": 1865
+ },
+ {
+ "epoch": 25.92227074235808,
+ "grad_norm": 0.014989510178565979,
+ "learning_rate": 0.0006,
+ "loss": 4.380428314208984,
+ "step": 1866
+ },
+ {
+ "epoch": 25.936244541484715,
+ "grad_norm": 0.01527960691601038,
+ "learning_rate": 0.0006,
+ "loss": 4.399374961853027,
+ "step": 1867
+ },
+ {
+ "epoch": 25.950218340611354,
+ "grad_norm": 0.015434633940458298,
+ "learning_rate": 0.0006,
+ "loss": 4.260984420776367,
+ "step": 1868
+ },
+ {
+ "epoch": 25.964192139737992,
+ "grad_norm": 0.015370228327810764,
+ "learning_rate": 0.0006,
+ "loss": 4.3340253829956055,
+ "step": 1869
+ },
+ {
+ "epoch": 25.978165938864628,
+ "grad_norm": 0.015348542481660843,
+ "learning_rate": 0.0006,
+ "loss": 4.369132995605469,
+ "step": 1870
+ },
+ {
+ "epoch": 25.992139737991266,
+ "grad_norm": 0.015354936942458153,
+ "learning_rate": 0.0006,
+ "loss": 4.381937026977539,
+ "step": 1871
+ },
+ {
+ "epoch": 26.0,
+ "grad_norm": 0.01796272210776806,
+ "learning_rate": 0.0006,
+ "loss": 4.286768436431885,
+ "step": 1872
+ },
+ {
+ "epoch": 26.0,
+ "eval_loss": 4.576984405517578,
+ "eval_runtime": 57.2091,
+ "eval_samples_per_second": 42.685,
+ "eval_steps_per_second": 1.346,
+ "step": 1872
+ },
+ {
+ "epoch": 26.01397379912664,
+ "grad_norm": 0.01748845726251602,
+ "learning_rate": 0.0006,
+ "loss": 4.337504863739014,
+ "step": 1873
+ },
+ {
+ "epoch": 26.027947598253274,
+ "grad_norm": 0.019244296476244926,
+ "learning_rate": 0.0006,
+ "loss": 4.264280796051025,
+ "step": 1874
+ },
+ {
+ "epoch": 26.041921397379912,
+ "grad_norm": 0.0215692650526762,
+ "learning_rate": 0.0006,
+ "loss": 4.350830078125,
+ "step": 1875
+ },
+ {
+ "epoch": 26.05589519650655,
+ "grad_norm": 0.022189252078533173,
+ "learning_rate": 0.0006,
+ "loss": 4.299098968505859,
+ "step": 1876
+ },
+ {
+ "epoch": 26.069868995633186,
+ "grad_norm": 0.022384043782949448,
+ "learning_rate": 0.0006,
+ "loss": 4.1185173988342285,
+ "step": 1877
+ },
+ {
+ "epoch": 26.083842794759825,
+ "grad_norm": 0.02076014317572117,
+ "learning_rate": 0.0006,
+ "loss": 4.297612190246582,
+ "step": 1878
+ },
+ {
+ "epoch": 26.097816593886463,
+ "grad_norm": 0.020843051373958588,
+ "learning_rate": 0.0006,
+ "loss": 4.294098854064941,
+ "step": 1879
+ },
+ {
+ "epoch": 26.111790393013102,
+ "grad_norm": 0.020654456689953804,
+ "learning_rate": 0.0006,
+ "loss": 4.267856597900391,
+ "step": 1880
+ },
+ {
+ "epoch": 26.125764192139737,
+ "grad_norm": 0.01860472559928894,
+ "learning_rate": 0.0006,
+ "loss": 4.284232139587402,
+ "step": 1881
+ },
+ {
+ "epoch": 26.139737991266376,
+ "grad_norm": 0.018719421699643135,
+ "learning_rate": 0.0006,
+ "loss": 4.277214527130127,
+ "step": 1882
+ },
+ {
+ "epoch": 26.153711790393015,
+ "grad_norm": 0.020903829485177994,
+ "learning_rate": 0.0006,
+ "loss": 4.254947662353516,
+ "step": 1883
+ },
+ {
+ "epoch": 26.16768558951965,
+ "grad_norm": 0.02020149491727352,
+ "learning_rate": 0.0006,
+ "loss": 4.2526936531066895,
+ "step": 1884
+ },
+ {
+ "epoch": 26.18165938864629,
+ "grad_norm": 0.0180392786860466,
+ "learning_rate": 0.0006,
+ "loss": 4.309922695159912,
+ "step": 1885
+ },
+ {
+ "epoch": 26.195633187772927,
+ "grad_norm": 0.018163125962018967,
+ "learning_rate": 0.0006,
+ "loss": 4.13860559463501,
+ "step": 1886
+ },
+ {
+ "epoch": 26.209606986899562,
+ "grad_norm": 0.01853291131556034,
+ "learning_rate": 0.0006,
+ "loss": 4.190926551818848,
+ "step": 1887
+ },
+ {
+ "epoch": 26.2235807860262,
+ "grad_norm": 0.01757677271962166,
+ "learning_rate": 0.0006,
+ "loss": 4.163999557495117,
+ "step": 1888
+ },
+ {
+ "epoch": 26.23755458515284,
+ "grad_norm": 0.01767992228269577,
+ "learning_rate": 0.0006,
+ "loss": 4.184430122375488,
+ "step": 1889
+ },
+ {
+ "epoch": 26.251528384279474,
+ "grad_norm": 0.018139973282814026,
+ "learning_rate": 0.0006,
+ "loss": 4.300766944885254,
+ "step": 1890
+ },
+ {
+ "epoch": 26.265502183406113,
+ "grad_norm": 0.020184461027383804,
+ "learning_rate": 0.0006,
+ "loss": 4.232537269592285,
+ "step": 1891
+ },
+ {
+ "epoch": 26.27947598253275,
+ "grad_norm": 0.02028856985270977,
+ "learning_rate": 0.0006,
+ "loss": 4.324883460998535,
+ "step": 1892
+ },
+ {
+ "epoch": 26.29344978165939,
+ "grad_norm": 0.017120616510510445,
+ "learning_rate": 0.0006,
+ "loss": 4.23493766784668,
+ "step": 1893
+ },
+ {
+ "epoch": 26.307423580786025,
+ "grad_norm": 0.016202978789806366,
+ "learning_rate": 0.0006,
+ "loss": 4.197602272033691,
+ "step": 1894
+ },
+ {
+ "epoch": 26.321397379912664,
+ "grad_norm": 0.017231421545147896,
+ "learning_rate": 0.0006,
+ "loss": 4.302001953125,
+ "step": 1895
+ },
+ {
+ "epoch": 26.335371179039303,
+ "grad_norm": 0.01725655607879162,
+ "learning_rate": 0.0006,
+ "loss": 4.356800079345703,
+ "step": 1896
+ },
+ {
+ "epoch": 26.349344978165938,
+ "grad_norm": 0.016657443717122078,
+ "learning_rate": 0.0006,
+ "loss": 4.266307830810547,
+ "step": 1897
+ },
+ {
+ "epoch": 26.363318777292577,
+ "grad_norm": 0.01616556942462921,
+ "learning_rate": 0.0006,
+ "loss": 4.307868003845215,
+ "step": 1898
+ },
+ {
+ "epoch": 26.377292576419215,
+ "grad_norm": 0.016621306538581848,
+ "learning_rate": 0.0006,
+ "loss": 4.172905921936035,
+ "step": 1899
+ },
+ {
+ "epoch": 26.39126637554585,
+ "grad_norm": 0.016243206337094307,
+ "learning_rate": 0.0006,
+ "loss": 4.162896156311035,
+ "step": 1900
+ },
+ {
+ "epoch": 26.40524017467249,
+ "grad_norm": 0.016663808375597,
+ "learning_rate": 0.0006,
+ "loss": 4.258232116699219,
+ "step": 1901
+ },
+ {
+ "epoch": 26.419213973799128,
+ "grad_norm": 0.01677478663623333,
+ "learning_rate": 0.0006,
+ "loss": 4.336090087890625,
+ "step": 1902
+ },
+ {
+ "epoch": 26.433187772925763,
+ "grad_norm": 0.01641303487122059,
+ "learning_rate": 0.0006,
+ "loss": 4.191476821899414,
+ "step": 1903
+ },
+ {
+ "epoch": 26.4471615720524,
+ "grad_norm": 0.016700323671102524,
+ "learning_rate": 0.0006,
+ "loss": 4.299427509307861,
+ "step": 1904
+ },
+ {
+ "epoch": 26.46113537117904,
+ "grad_norm": 0.017595035955309868,
+ "learning_rate": 0.0006,
+ "loss": 4.312828063964844,
+ "step": 1905
+ },
+ {
+ "epoch": 26.475109170305675,
+ "grad_norm": 0.020300284028053284,
+ "learning_rate": 0.0006,
+ "loss": 4.28331184387207,
+ "step": 1906
+ },
+ {
+ "epoch": 26.489082969432314,
+ "grad_norm": 0.02198321744799614,
+ "learning_rate": 0.0006,
+ "loss": 4.3280839920043945,
+ "step": 1907
+ },
+ {
+ "epoch": 26.503056768558952,
+ "grad_norm": 0.02003200724720955,
+ "learning_rate": 0.0006,
+ "loss": 4.325623989105225,
+ "step": 1908
+ },
+ {
+ "epoch": 26.51703056768559,
+ "grad_norm": 0.01870287023484707,
+ "learning_rate": 0.0006,
+ "loss": 4.251203536987305,
+ "step": 1909
+ },
+ {
+ "epoch": 26.531004366812226,
+ "grad_norm": 0.01835448481142521,
+ "learning_rate": 0.0006,
+ "loss": 4.335144996643066,
+ "step": 1910
+ },
+ {
+ "epoch": 26.544978165938865,
+ "grad_norm": 0.02046327292919159,
+ "learning_rate": 0.0006,
+ "loss": 4.238023281097412,
+ "step": 1911
+ },
+ {
+ "epoch": 26.558951965065503,
+ "grad_norm": 0.019771311432123184,
+ "learning_rate": 0.0006,
+ "loss": 4.353594779968262,
+ "step": 1912
+ },
+ {
+ "epoch": 26.57292576419214,
+ "grad_norm": 0.018090544268488884,
+ "learning_rate": 0.0006,
+ "loss": 4.304343223571777,
+ "step": 1913
+ },
+ {
+ "epoch": 26.586899563318777,
+ "grad_norm": 0.017489777877926826,
+ "learning_rate": 0.0006,
+ "loss": 4.204375267028809,
+ "step": 1914
+ },
+ {
+ "epoch": 26.600873362445416,
+ "grad_norm": 0.016785891726613045,
+ "learning_rate": 0.0006,
+ "loss": 4.286669731140137,
+ "step": 1915
+ },
+ {
+ "epoch": 26.61484716157205,
+ "grad_norm": 0.016382789239287376,
+ "learning_rate": 0.0006,
+ "loss": 4.347662448883057,
+ "step": 1916
+ },
+ {
+ "epoch": 26.62882096069869,
+ "grad_norm": 0.016581876203417778,
+ "learning_rate": 0.0006,
+ "loss": 4.2716193199157715,
+ "step": 1917
+ },
+ {
+ "epoch": 26.64279475982533,
+ "grad_norm": 0.015627184882760048,
+ "learning_rate": 0.0006,
+ "loss": 4.311489105224609,
+ "step": 1918
+ },
+ {
+ "epoch": 26.656768558951963,
+ "grad_norm": 0.015908483415842056,
+ "learning_rate": 0.0006,
+ "loss": 4.262099742889404,
+ "step": 1919
+ },
+ {
+ "epoch": 26.670742358078602,
+ "grad_norm": 0.015576236881315708,
+ "learning_rate": 0.0006,
+ "loss": 4.275871276855469,
+ "step": 1920
+ },
+ {
+ "epoch": 26.68471615720524,
+ "grad_norm": 0.014621064998209476,
+ "learning_rate": 0.0006,
+ "loss": 4.184260845184326,
+ "step": 1921
+ },
+ {
+ "epoch": 26.69868995633188,
+ "grad_norm": 0.014797811396420002,
+ "learning_rate": 0.0006,
+ "loss": 4.3412628173828125,
+ "step": 1922
+ },
+ {
+ "epoch": 26.712663755458514,
+ "grad_norm": 0.015610570088028908,
+ "learning_rate": 0.0006,
+ "loss": 4.181094646453857,
+ "step": 1923
+ },
+ {
+ "epoch": 26.726637554585153,
+ "grad_norm": 0.016048265621066093,
+ "learning_rate": 0.0006,
+ "loss": 4.365290641784668,
+ "step": 1924
+ },
+ {
+ "epoch": 26.74061135371179,
+ "grad_norm": 0.016604071483016014,
+ "learning_rate": 0.0006,
+ "loss": 4.2628912925720215,
+ "step": 1925
+ },
+ {
+ "epoch": 26.754585152838427,
+ "grad_norm": 0.017044221982359886,
+ "learning_rate": 0.0006,
+ "loss": 4.378512382507324,
+ "step": 1926
+ },
+ {
+ "epoch": 26.768558951965066,
+ "grad_norm": 0.01746384985744953,
+ "learning_rate": 0.0006,
+ "loss": 4.3888773918151855,
+ "step": 1927
+ },
+ {
+ "epoch": 26.782532751091704,
+ "grad_norm": 0.01572711206972599,
+ "learning_rate": 0.0006,
+ "loss": 4.189935684204102,
+ "step": 1928
+ },
+ {
+ "epoch": 26.79650655021834,
+ "grad_norm": 0.014976629987359047,
+ "learning_rate": 0.0006,
+ "loss": 4.280525207519531,
+ "step": 1929
+ },
+ {
+ "epoch": 26.810480349344978,
+ "grad_norm": 0.014804664999246597,
+ "learning_rate": 0.0006,
+ "loss": 4.270630836486816,
+ "step": 1930
+ },
+ {
+ "epoch": 26.824454148471617,
+ "grad_norm": 0.014641920104622841,
+ "learning_rate": 0.0006,
+ "loss": 4.319505214691162,
+ "step": 1931
+ },
+ {
+ "epoch": 26.83842794759825,
+ "grad_norm": 0.014928505755960941,
+ "learning_rate": 0.0006,
+ "loss": 4.180147647857666,
+ "step": 1932
+ },
+ {
+ "epoch": 26.85240174672489,
+ "grad_norm": 0.016130365431308746,
+ "learning_rate": 0.0006,
+ "loss": 4.260660648345947,
+ "step": 1933
+ },
+ {
+ "epoch": 26.86637554585153,
+ "grad_norm": 0.015642493963241577,
+ "learning_rate": 0.0006,
+ "loss": 4.304064750671387,
+ "step": 1934
+ },
+ {
+ "epoch": 26.880349344978168,
+ "grad_norm": 0.015988217666745186,
+ "learning_rate": 0.0006,
+ "loss": 4.2548017501831055,
+ "step": 1935
+ },
+ {
+ "epoch": 26.894323144104803,
+ "grad_norm": 0.016501398757100105,
+ "learning_rate": 0.0006,
+ "loss": 4.260058879852295,
+ "step": 1936
+ },
+ {
+ "epoch": 26.90829694323144,
+ "grad_norm": 0.01645040698349476,
+ "learning_rate": 0.0006,
+ "loss": 4.483081340789795,
+ "step": 1937
+ },
+ {
+ "epoch": 26.92227074235808,
+ "grad_norm": 0.016140863299369812,
+ "learning_rate": 0.0006,
+ "loss": 4.1776838302612305,
+ "step": 1938
+ },
+ {
+ "epoch": 26.936244541484715,
+ "grad_norm": 0.016681019216775894,
+ "learning_rate": 0.0006,
+ "loss": 4.224747657775879,
+ "step": 1939
+ },
+ {
+ "epoch": 26.950218340611354,
+ "grad_norm": 0.01650378853082657,
+ "learning_rate": 0.0006,
+ "loss": 4.305703163146973,
+ "step": 1940
+ },
+ {
+ "epoch": 26.964192139737992,
+ "grad_norm": 0.017051683738827705,
+ "learning_rate": 0.0006,
+ "loss": 4.285835266113281,
+ "step": 1941
+ },
+ {
+ "epoch": 26.978165938864628,
+ "grad_norm": 0.016894327476620674,
+ "learning_rate": 0.0006,
+ "loss": 4.318901538848877,
+ "step": 1942
+ },
+ {
+ "epoch": 26.992139737991266,
+ "grad_norm": 0.017313100397586823,
+ "learning_rate": 0.0006,
+ "loss": 4.281434535980225,
+ "step": 1943
+ },
+ {
+ "epoch": 27.0,
+ "grad_norm": 0.018450884148478508,
+ "learning_rate": 0.0006,
+ "loss": 4.335104942321777,
+ "step": 1944
+ },
+ {
+ "epoch": 27.0,
+ "eval_loss": 4.640687465667725,
+ "eval_runtime": 56.9624,
+ "eval_samples_per_second": 42.87,
+ "eval_steps_per_second": 1.352,
+ "step": 1944
+ },
+ {
+ "epoch": 27.01397379912664,
+ "grad_norm": 0.017755698412656784,
+ "learning_rate": 0.0006,
+ "loss": 4.278824329376221,
+ "step": 1945
+ },
+ {
+ "epoch": 27.027947598253274,
+ "grad_norm": 0.01724310778081417,
+ "learning_rate": 0.0006,
+ "loss": 4.311049461364746,
+ "step": 1946
+ },
+ {
+ "epoch": 27.041921397379912,
+ "grad_norm": 0.01690123789012432,
+ "learning_rate": 0.0006,
+ "loss": 4.261994361877441,
+ "step": 1947
+ },
+ {
+ "epoch": 27.05589519650655,
+ "grad_norm": 0.018406696617603302,
+ "learning_rate": 0.0006,
+ "loss": 4.1925153732299805,
+ "step": 1948
+ },
+ {
+ "epoch": 27.069868995633186,
+ "grad_norm": 0.019945867359638214,
+ "learning_rate": 0.0006,
+ "loss": 4.108537197113037,
+ "step": 1949
+ },
+ {
+ "epoch": 27.083842794759825,
+ "grad_norm": 0.020316628739237785,
+ "learning_rate": 0.0006,
+ "loss": 4.209043025970459,
+ "step": 1950
+ },
+ {
+ "epoch": 27.097816593886463,
+ "grad_norm": 0.019690370187163353,
+ "learning_rate": 0.0006,
+ "loss": 4.195789337158203,
+ "step": 1951
+ },
+ {
+ "epoch": 27.111790393013102,
+ "grad_norm": 0.01785232499241829,
+ "learning_rate": 0.0006,
+ "loss": 4.21466064453125,
+ "step": 1952
+ },
+ {
+ "epoch": 27.125764192139737,
+ "grad_norm": 0.0164767038077116,
+ "learning_rate": 0.0006,
+ "loss": 4.270247459411621,
+ "step": 1953
+ },
+ {
+ "epoch": 27.139737991266376,
+ "grad_norm": 0.018008455634117126,
+ "learning_rate": 0.0006,
+ "loss": 4.201042175292969,
+ "step": 1954
+ },
+ {
+ "epoch": 27.153711790393015,
+ "grad_norm": 0.01807340793311596,
+ "learning_rate": 0.0006,
+ "loss": 4.25289249420166,
+ "step": 1955
+ },
+ {
+ "epoch": 27.16768558951965,
+ "grad_norm": 0.016246909275650978,
+ "learning_rate": 0.0006,
+ "loss": 4.116283416748047,
+ "step": 1956
+ },
+ {
+ "epoch": 27.18165938864629,
+ "grad_norm": 0.017545776441693306,
+ "learning_rate": 0.0006,
+ "loss": 4.316999435424805,
+ "step": 1957
+ },
+ {
+ "epoch": 27.195633187772927,
+ "grad_norm": 0.017915375530719757,
+ "learning_rate": 0.0006,
+ "loss": 4.253859519958496,
+ "step": 1958
+ },
+ {
+ "epoch": 27.209606986899562,
+ "grad_norm": 0.018112223595380783,
+ "learning_rate": 0.0006,
+ "loss": 4.2541022300720215,
+ "step": 1959
+ },
+ {
+ "epoch": 27.2235807860262,
+ "grad_norm": 0.01846139505505562,
+ "learning_rate": 0.0006,
+ "loss": 4.20438289642334,
+ "step": 1960
+ },
+ {
+ "epoch": 27.23755458515284,
+ "grad_norm": 0.01943897269666195,
+ "learning_rate": 0.0006,
+ "loss": 4.2799577713012695,
+ "step": 1961
+ },
+ {
+ "epoch": 27.251528384279474,
+ "grad_norm": 0.018253348767757416,
+ "learning_rate": 0.0006,
+ "loss": 4.1720499992370605,
+ "step": 1962
+ },
+ {
+ "epoch": 27.265502183406113,
+ "grad_norm": 0.018292246386408806,
+ "learning_rate": 0.0006,
+ "loss": 4.28231143951416,
+ "step": 1963
+ },
+ {
+ "epoch": 27.27947598253275,
+ "grad_norm": 0.01943155936896801,
+ "learning_rate": 0.0006,
+ "loss": 4.314976692199707,
+ "step": 1964
+ },
+ {
+ "epoch": 27.29344978165939,
+ "grad_norm": 0.019064994528889656,
+ "learning_rate": 0.0006,
+ "loss": 4.25328254699707,
+ "step": 1965
+ },
+ {
+ "epoch": 27.307423580786025,
+ "grad_norm": 0.017423273995518684,
+ "learning_rate": 0.0006,
+ "loss": 4.195992469787598,
+ "step": 1966
+ },
+ {
+ "epoch": 27.321397379912664,
+ "grad_norm": 0.01677924580872059,
+ "learning_rate": 0.0006,
+ "loss": 4.260043144226074,
+ "step": 1967
+ },
+ {
+ "epoch": 27.335371179039303,
+ "grad_norm": 0.01720673404633999,
+ "learning_rate": 0.0006,
+ "loss": 4.222780704498291,
+ "step": 1968
+ },
+ {
+ "epoch": 27.349344978165938,
+ "grad_norm": 0.017300743609666824,
+ "learning_rate": 0.0006,
+ "loss": 4.109524250030518,
+ "step": 1969
+ },
+ {
+ "epoch": 27.363318777292577,
+ "grad_norm": 0.016065871343016624,
+ "learning_rate": 0.0006,
+ "loss": 4.36108922958374,
+ "step": 1970
+ },
+ {
+ "epoch": 27.377292576419215,
+ "grad_norm": 0.016626616939902306,
+ "learning_rate": 0.0006,
+ "loss": 4.220630645751953,
+ "step": 1971
+ },
+ {
+ "epoch": 27.39126637554585,
+ "grad_norm": 0.016482602804899216,
+ "learning_rate": 0.0006,
+ "loss": 4.344797134399414,
+ "step": 1972
+ },
+ {
+ "epoch": 27.40524017467249,
+ "grad_norm": 0.01727953553199768,
+ "learning_rate": 0.0006,
+ "loss": 4.296645164489746,
+ "step": 1973
+ },
+ {
+ "epoch": 27.419213973799128,
+ "grad_norm": 0.01522731315344572,
+ "learning_rate": 0.0006,
+ "loss": 4.223751068115234,
+ "step": 1974
+ },
+ {
+ "epoch": 27.433187772925763,
+ "grad_norm": 0.01581776700913906,
+ "learning_rate": 0.0006,
+ "loss": 4.351379871368408,
+ "step": 1975
+ },
+ {
+ "epoch": 27.4471615720524,
+ "grad_norm": 0.016394605860114098,
+ "learning_rate": 0.0006,
+ "loss": 4.2412919998168945,
+ "step": 1976
+ },
+ {
+ "epoch": 27.46113537117904,
+ "grad_norm": 0.01751169003546238,
+ "learning_rate": 0.0006,
+ "loss": 4.257606506347656,
+ "step": 1977
+ },
+ {
+ "epoch": 27.475109170305675,
+ "grad_norm": 0.017452310770750046,
+ "learning_rate": 0.0006,
+ "loss": 4.380134582519531,
+ "step": 1978
+ },
+ {
+ "epoch": 27.489082969432314,
+ "grad_norm": 0.017741898074746132,
+ "learning_rate": 0.0006,
+ "loss": 4.279632568359375,
+ "step": 1979
+ },
+ {
+ "epoch": 27.503056768558952,
+ "grad_norm": 0.017680590972304344,
+ "learning_rate": 0.0006,
+ "loss": 4.273406028747559,
+ "step": 1980
+ },
+ {
+ "epoch": 27.51703056768559,
+ "grad_norm": 0.01864718459546566,
+ "learning_rate": 0.0006,
+ "loss": 4.278583526611328,
+ "step": 1981
+ },
+ {
+ "epoch": 27.531004366812226,
+ "grad_norm": 0.019558541476726532,
+ "learning_rate": 0.0006,
+ "loss": 4.211919784545898,
+ "step": 1982
+ },
+ {
+ "epoch": 27.544978165938865,
+ "grad_norm": 0.019467400386929512,
+ "learning_rate": 0.0006,
+ "loss": 4.276822566986084,
+ "step": 1983
+ },
+ {
+ "epoch": 27.558951965065503,
+ "grad_norm": 0.01756172813475132,
+ "learning_rate": 0.0006,
+ "loss": 4.165379047393799,
+ "step": 1984
+ },
+ {
+ "epoch": 27.57292576419214,
+ "grad_norm": 0.017458263784646988,
+ "learning_rate": 0.0006,
+ "loss": 4.2992401123046875,
+ "step": 1985
+ },
+ {
+ "epoch": 27.586899563318777,
+ "grad_norm": 0.018089665099978447,
+ "learning_rate": 0.0006,
+ "loss": 4.243378639221191,
+ "step": 1986
+ },
+ {
+ "epoch": 27.600873362445416,
+ "grad_norm": 0.019535524770617485,
+ "learning_rate": 0.0006,
+ "loss": 4.321477890014648,
+ "step": 1987
+ },
+ {
+ "epoch": 27.61484716157205,
+ "grad_norm": 0.01997970975935459,
+ "learning_rate": 0.0006,
+ "loss": 4.123082637786865,
+ "step": 1988
+ },
+ {
+ "epoch": 27.62882096069869,
+ "grad_norm": 0.01913067139685154,
+ "learning_rate": 0.0006,
+ "loss": 4.289896488189697,
+ "step": 1989
+ },
+ {
+ "epoch": 27.64279475982533,
+ "grad_norm": 0.0169806070625782,
+ "learning_rate": 0.0006,
+ "loss": 4.247004985809326,
+ "step": 1990
+ },
+ {
+ "epoch": 27.656768558951963,
+ "grad_norm": 0.017315007746219635,
+ "learning_rate": 0.0006,
+ "loss": 4.2095441818237305,
+ "step": 1991
+ },
+ {
+ "epoch": 27.670742358078602,
+ "grad_norm": 0.015098759904503822,
+ "learning_rate": 0.0006,
+ "loss": 4.145560264587402,
+ "step": 1992
+ },
+ {
+ "epoch": 27.68471615720524,
+ "grad_norm": 0.016240514814853668,
+ "learning_rate": 0.0006,
+ "loss": 4.254927635192871,
+ "step": 1993
+ },
+ {
+ "epoch": 27.69868995633188,
+ "grad_norm": 0.01621108129620552,
+ "learning_rate": 0.0006,
+ "loss": 4.219725608825684,
+ "step": 1994
+ },
+ {
+ "epoch": 27.712663755458514,
+ "grad_norm": 0.015191009268164635,
+ "learning_rate": 0.0006,
+ "loss": 4.194345474243164,
+ "step": 1995
+ },
+ {
+ "epoch": 27.726637554585153,
+ "grad_norm": 0.014387983828783035,
+ "learning_rate": 0.0006,
+ "loss": 4.211467742919922,
+ "step": 1996
+ },
+ {
+ "epoch": 27.74061135371179,
+ "grad_norm": 0.015460561029613018,
+ "learning_rate": 0.0006,
+ "loss": 4.358700752258301,
+ "step": 1997
+ },
+ {
+ "epoch": 27.754585152838427,
+ "grad_norm": 0.015100076794624329,
+ "learning_rate": 0.0006,
+ "loss": 4.171223163604736,
+ "step": 1998
+ },
+ {
+ "epoch": 27.768558951965066,
+ "grad_norm": 0.015607891604304314,
+ "learning_rate": 0.0006,
+ "loss": 4.217874526977539,
+ "step": 1999
+ },
+ {
+ "epoch": 27.782532751091704,
+ "grad_norm": 0.015650689601898193,
+ "learning_rate": 0.0006,
+ "loss": 4.280780792236328,
+ "step": 2000
+ },
+ {
+ "epoch": 27.79650655021834,
+ "grad_norm": 0.01580994389951229,
+ "learning_rate": 0.0006,
+ "loss": 4.159212112426758,
+ "step": 2001
+ },
+ {
+ "epoch": 27.810480349344978,
+ "grad_norm": 0.016607413068413734,
+ "learning_rate": 0.0006,
+ "loss": 4.238853454589844,
+ "step": 2002
+ },
+ {
+ "epoch": 27.824454148471617,
+ "grad_norm": 0.017383860424160957,
+ "learning_rate": 0.0006,
+ "loss": 4.233944892883301,
+ "step": 2003
+ },
+ {
+ "epoch": 27.83842794759825,
+ "grad_norm": 0.014966659247875214,
+ "learning_rate": 0.0006,
+ "loss": 4.167392730712891,
+ "step": 2004
+ },
+ {
+ "epoch": 27.85240174672489,
+ "grad_norm": 0.014589878730475903,
+ "learning_rate": 0.0006,
+ "loss": 4.2390899658203125,
+ "step": 2005
+ },
+ {
+ "epoch": 27.86637554585153,
+ "grad_norm": 0.017179450020194054,
+ "learning_rate": 0.0006,
+ "loss": 4.365785598754883,
+ "step": 2006
+ },
+ {
+ "epoch": 27.880349344978168,
+ "grad_norm": 0.01893901452422142,
+ "learning_rate": 0.0006,
+ "loss": 4.289430141448975,
+ "step": 2007
+ },
+ {
+ "epoch": 27.894323144104803,
+ "grad_norm": 0.018619712442159653,
+ "learning_rate": 0.0006,
+ "loss": 4.310665130615234,
+ "step": 2008
+ },
+ {
+ "epoch": 27.90829694323144,
+ "grad_norm": 0.019180385395884514,
+ "learning_rate": 0.0006,
+ "loss": 4.27454137802124,
+ "step": 2009
+ },
+ {
+ "epoch": 27.92227074235808,
+ "grad_norm": 0.017811523750424385,
+ "learning_rate": 0.0006,
+ "loss": 4.312562465667725,
+ "step": 2010
+ },
+ {
+ "epoch": 27.936244541484715,
+ "grad_norm": 0.016583573073148727,
+ "learning_rate": 0.0006,
+ "loss": 4.202253341674805,
+ "step": 2011
+ },
+ {
+ "epoch": 27.950218340611354,
+ "grad_norm": 0.01658615842461586,
+ "learning_rate": 0.0006,
+ "loss": 4.183433532714844,
+ "step": 2012
+ },
+ {
+ "epoch": 27.964192139737992,
+ "grad_norm": 0.015327401459217072,
+ "learning_rate": 0.0006,
+ "loss": 4.2125139236450195,
+ "step": 2013
+ },
+ {
+ "epoch": 27.978165938864628,
+ "grad_norm": 0.015004601329565048,
+ "learning_rate": 0.0006,
+ "loss": 4.257392883300781,
+ "step": 2014
+ },
+ {
+ "epoch": 27.992139737991266,
+ "grad_norm": 0.015259161591529846,
+ "learning_rate": 0.0006,
+ "loss": 4.241818428039551,
+ "step": 2015
+ },
+ {
+ "epoch": 28.0,
+ "grad_norm": 0.016808776184916496,
+ "learning_rate": 0.0006,
+ "loss": 4.146241188049316,
+ "step": 2016
+ },
+ {
+ "epoch": 28.0,
+ "eval_loss": 4.663090705871582,
+ "eval_runtime": 56.4559,
+ "eval_samples_per_second": 43.255,
+ "eval_steps_per_second": 1.364,
+ "step": 2016
+ },
+ {
+ "epoch": 28.01397379912664,
+ "grad_norm": 0.015784382820129395,
+ "learning_rate": 0.0006,
+ "loss": 4.044191360473633,
+ "step": 2017
+ },
+ {
+ "epoch": 28.027947598253274,
+ "grad_norm": 0.015636710450053215,
+ "learning_rate": 0.0006,
+ "loss": 4.128754615783691,
+ "step": 2018
+ },
+ {
+ "epoch": 28.041921397379912,
+ "grad_norm": 0.015364975668489933,
+ "learning_rate": 0.0006,
+ "loss": 4.191835403442383,
+ "step": 2019
+ },
+ {
+ "epoch": 28.05589519650655,
+ "grad_norm": 0.015015700832009315,
+ "learning_rate": 0.0006,
+ "loss": 4.120022773742676,
+ "step": 2020
+ },
+ {
+ "epoch": 28.069868995633186,
+ "grad_norm": 0.016666799783706665,
+ "learning_rate": 0.0006,
+ "loss": 4.33906364440918,
+ "step": 2021
+ },
+ {
+ "epoch": 28.083842794759825,
+ "grad_norm": 0.01780104823410511,
+ "learning_rate": 0.0006,
+ "loss": 4.279458999633789,
+ "step": 2022
+ },
+ {
+ "epoch": 28.097816593886463,
+ "grad_norm": 0.01870882883667946,
+ "learning_rate": 0.0006,
+ "loss": 4.34728479385376,
+ "step": 2023
+ },
+ {
+ "epoch": 28.111790393013102,
+ "grad_norm": 0.018406381830573082,
+ "learning_rate": 0.0006,
+ "loss": 4.211311340332031,
+ "step": 2024
+ },
+ {
+ "epoch": 28.125764192139737,
+ "grad_norm": 0.017833448946475983,
+ "learning_rate": 0.0006,
+ "loss": 4.223254203796387,
+ "step": 2025
+ },
+ {
+ "epoch": 28.139737991266376,
+ "grad_norm": 0.017552688717842102,
+ "learning_rate": 0.0006,
+ "loss": 4.208823204040527,
+ "step": 2026
+ },
+ {
+ "epoch": 28.153711790393015,
+ "grad_norm": 0.018500229343771935,
+ "learning_rate": 0.0006,
+ "loss": 4.198025226593018,
+ "step": 2027
+ },
+ {
+ "epoch": 28.16768558951965,
+ "grad_norm": 0.01782156340777874,
+ "learning_rate": 0.0006,
+ "loss": 4.179978370666504,
+ "step": 2028
+ },
+ {
+ "epoch": 28.18165938864629,
+ "grad_norm": 0.019580967724323273,
+ "learning_rate": 0.0006,
+ "loss": 4.291114807128906,
+ "step": 2029
+ },
+ {
+ "epoch": 28.195633187772927,
+ "grad_norm": 0.020804036408662796,
+ "learning_rate": 0.0006,
+ "loss": 4.220607280731201,
+ "step": 2030
+ },
+ {
+ "epoch": 28.209606986899562,
+ "grad_norm": 0.019595693796873093,
+ "learning_rate": 0.0006,
+ "loss": 4.2905731201171875,
+ "step": 2031
+ },
+ {
+ "epoch": 28.2235807860262,
+ "grad_norm": 0.019247086718678474,
+ "learning_rate": 0.0006,
+ "loss": 4.2467360496521,
+ "step": 2032
+ },
+ {
+ "epoch": 28.23755458515284,
+ "grad_norm": 0.01958036608994007,
+ "learning_rate": 0.0006,
+ "loss": 4.286137580871582,
+ "step": 2033
+ },
+ {
+ "epoch": 28.251528384279474,
+ "grad_norm": 0.021354753524065018,
+ "learning_rate": 0.0006,
+ "loss": 4.234755039215088,
+ "step": 2034
+ },
+ {
+ "epoch": 28.265502183406113,
+ "grad_norm": 0.02206375077366829,
+ "learning_rate": 0.0006,
+ "loss": 4.149641036987305,
+ "step": 2035
+ },
+ {
+ "epoch": 28.27947598253275,
+ "grad_norm": 0.02201310358941555,
+ "learning_rate": 0.0006,
+ "loss": 4.1928629875183105,
+ "step": 2036
+ },
+ {
+ "epoch": 28.29344978165939,
+ "grad_norm": 0.01961950585246086,
+ "learning_rate": 0.0006,
+ "loss": 4.211404323577881,
+ "step": 2037
+ },
+ {
+ "epoch": 28.307423580786025,
+ "grad_norm": 0.019211795181035995,
+ "learning_rate": 0.0006,
+ "loss": 4.200976848602295,
+ "step": 2038
+ },
+ {
+ "epoch": 28.321397379912664,
+ "grad_norm": 0.01779370754957199,
+ "learning_rate": 0.0006,
+ "loss": 4.249148845672607,
+ "step": 2039
+ },
+ {
+ "epoch": 28.335371179039303,
+ "grad_norm": 0.018635908141732216,
+ "learning_rate": 0.0006,
+ "loss": 4.11806058883667,
+ "step": 2040
+ },
+ {
+ "epoch": 28.349344978165938,
+ "grad_norm": 0.019792694598436356,
+ "learning_rate": 0.0006,
+ "loss": 4.335224151611328,
+ "step": 2041
+ },
+ {
+ "epoch": 28.363318777292577,
+ "grad_norm": 0.021731717512011528,
+ "learning_rate": 0.0006,
+ "loss": 4.302778244018555,
+ "step": 2042
+ },
+ {
+ "epoch": 28.377292576419215,
+ "grad_norm": 0.020823447033762932,
+ "learning_rate": 0.0006,
+ "loss": 4.260610580444336,
+ "step": 2043
+ },
+ {
+ "epoch": 28.39126637554585,
+ "grad_norm": 0.019385412335395813,
+ "learning_rate": 0.0006,
+ "loss": 4.2247724533081055,
+ "step": 2044
+ },
+ {
+ "epoch": 28.40524017467249,
+ "grad_norm": 0.019539108499884605,
+ "learning_rate": 0.0006,
+ "loss": 4.236655235290527,
+ "step": 2045
+ },
+ {
+ "epoch": 28.419213973799128,
+ "grad_norm": 0.01942690648138523,
+ "learning_rate": 0.0006,
+ "loss": 4.251850605010986,
+ "step": 2046
+ },
+ {
+ "epoch": 28.433187772925763,
+ "grad_norm": 0.019667886197566986,
+ "learning_rate": 0.0006,
+ "loss": 4.222312927246094,
+ "step": 2047
+ },
+ {
+ "epoch": 28.4471615720524,
+ "grad_norm": 0.02029012329876423,
+ "learning_rate": 0.0006,
+ "loss": 4.246252059936523,
+ "step": 2048
+ },
+ {
+ "epoch": 28.46113537117904,
+ "grad_norm": 0.01784469373524189,
+ "learning_rate": 0.0006,
+ "loss": 4.192128658294678,
+ "step": 2049
+ },
+ {
+ "epoch": 28.475109170305675,
+ "grad_norm": 0.0160287544131279,
+ "learning_rate": 0.0006,
+ "loss": 4.208607196807861,
+ "step": 2050
+ },
+ {
+ "epoch": 28.489082969432314,
+ "grad_norm": 0.015348346903920174,
+ "learning_rate": 0.0006,
+ "loss": 4.236598014831543,
+ "step": 2051
+ },
+ {
+ "epoch": 28.503056768558952,
+ "grad_norm": 0.015198206529021263,
+ "learning_rate": 0.0006,
+ "loss": 4.235552787780762,
+ "step": 2052
+ },
+ {
+ "epoch": 28.51703056768559,
+ "grad_norm": 0.01523754745721817,
+ "learning_rate": 0.0006,
+ "loss": 4.1941728591918945,
+ "step": 2053
+ },
+ {
+ "epoch": 28.531004366812226,
+ "grad_norm": 0.015559614636003971,
+ "learning_rate": 0.0006,
+ "loss": 4.294757843017578,
+ "step": 2054
+ },
+ {
+ "epoch": 28.544978165938865,
+ "grad_norm": 0.01568358950316906,
+ "learning_rate": 0.0006,
+ "loss": 4.213065147399902,
+ "step": 2055
+ },
+ {
+ "epoch": 28.558951965065503,
+ "grad_norm": 0.015336516313254833,
+ "learning_rate": 0.0006,
+ "loss": 4.240293979644775,
+ "step": 2056
+ },
+ {
+ "epoch": 28.57292576419214,
+ "grad_norm": 0.015041496604681015,
+ "learning_rate": 0.0006,
+ "loss": 4.273397445678711,
+ "step": 2057
+ },
+ {
+ "epoch": 28.586899563318777,
+ "grad_norm": 0.01610128954052925,
+ "learning_rate": 0.0006,
+ "loss": 4.125369071960449,
+ "step": 2058
+ },
+ {
+ "epoch": 28.600873362445416,
+ "grad_norm": 0.015974976122379303,
+ "learning_rate": 0.0006,
+ "loss": 4.133459091186523,
+ "step": 2059
+ },
+ {
+ "epoch": 28.61484716157205,
+ "grad_norm": 0.015936799347400665,
+ "learning_rate": 0.0006,
+ "loss": 4.306667327880859,
+ "step": 2060
+ },
+ {
+ "epoch": 28.62882096069869,
+ "grad_norm": 0.01795247197151184,
+ "learning_rate": 0.0006,
+ "loss": 4.205156326293945,
+ "step": 2061
+ },
+ {
+ "epoch": 28.64279475982533,
+ "grad_norm": 0.019639814272522926,
+ "learning_rate": 0.0006,
+ "loss": 4.277614593505859,
+ "step": 2062
+ },
+ {
+ "epoch": 28.656768558951963,
+ "grad_norm": 0.01722300611436367,
+ "learning_rate": 0.0006,
+ "loss": 4.191032886505127,
+ "step": 2063
+ },
+ {
+ "epoch": 28.670742358078602,
+ "grad_norm": 0.01612633652985096,
+ "learning_rate": 0.0006,
+ "loss": 4.20717191696167,
+ "step": 2064
+ },
+ {
+ "epoch": 28.68471615720524,
+ "grad_norm": 0.01603279449045658,
+ "learning_rate": 0.0006,
+ "loss": 4.228884220123291,
+ "step": 2065
+ },
+ {
+ "epoch": 28.69868995633188,
+ "grad_norm": 0.016405927017331123,
+ "learning_rate": 0.0006,
+ "loss": 4.201292991638184,
+ "step": 2066
+ },
+ {
+ "epoch": 28.712663755458514,
+ "grad_norm": 0.016644183546304703,
+ "learning_rate": 0.0006,
+ "loss": 4.239282131195068,
+ "step": 2067
+ },
+ {
+ "epoch": 28.726637554585153,
+ "grad_norm": 0.0175587497651577,
+ "learning_rate": 0.0006,
+ "loss": 4.289038181304932,
+ "step": 2068
+ },
+ {
+ "epoch": 28.74061135371179,
+ "grad_norm": 0.018223397433757782,
+ "learning_rate": 0.0006,
+ "loss": 4.2266998291015625,
+ "step": 2069
+ },
+ {
+ "epoch": 28.754585152838427,
+ "grad_norm": 0.018680868670344353,
+ "learning_rate": 0.0006,
+ "loss": 4.348849296569824,
+ "step": 2070
+ },
+ {
+ "epoch": 28.768558951965066,
+ "grad_norm": 0.016726728528738022,
+ "learning_rate": 0.0006,
+ "loss": 4.292080879211426,
+ "step": 2071
+ },
+ {
+ "epoch": 28.782532751091704,
+ "grad_norm": 0.017386795952916145,
+ "learning_rate": 0.0006,
+ "loss": 4.208870887756348,
+ "step": 2072
+ },
+ {
+ "epoch": 28.79650655021834,
+ "grad_norm": 0.01865178905427456,
+ "learning_rate": 0.0006,
+ "loss": 4.054150581359863,
+ "step": 2073
+ },
+ {
+ "epoch": 28.810480349344978,
+ "grad_norm": 0.01776815392076969,
+ "learning_rate": 0.0006,
+ "loss": 4.202404499053955,
+ "step": 2074
+ },
+ {
+ "epoch": 28.824454148471617,
+ "grad_norm": 0.01643497310578823,
+ "learning_rate": 0.0006,
+ "loss": 4.131855010986328,
+ "step": 2075
+ },
+ {
+ "epoch": 28.83842794759825,
+ "grad_norm": 0.01574273779988289,
+ "learning_rate": 0.0006,
+ "loss": 4.296788692474365,
+ "step": 2076
+ },
+ {
+ "epoch": 28.85240174672489,
+ "grad_norm": 0.015613908879458904,
+ "learning_rate": 0.0006,
+ "loss": 4.210949897766113,
+ "step": 2077
+ },
+ {
+ "epoch": 28.86637554585153,
+ "grad_norm": 0.015543212182819843,
+ "learning_rate": 0.0006,
+ "loss": 4.218136310577393,
+ "step": 2078
+ },
+ {
+ "epoch": 28.880349344978168,
+ "grad_norm": 0.014681251719594002,
+ "learning_rate": 0.0006,
+ "loss": 4.19401741027832,
+ "step": 2079
+ },
+ {
+ "epoch": 28.894323144104803,
+ "grad_norm": 0.014393828809261322,
+ "learning_rate": 0.0006,
+ "loss": 4.198468208312988,
+ "step": 2080
+ },
+ {
+ "epoch": 28.90829694323144,
+ "grad_norm": 0.01580170728266239,
+ "learning_rate": 0.0006,
+ "loss": 4.224969863891602,
+ "step": 2081
+ },
+ {
+ "epoch": 28.92227074235808,
+ "grad_norm": 0.01484165620058775,
+ "learning_rate": 0.0006,
+ "loss": 4.12236213684082,
+ "step": 2082
+ },
+ {
+ "epoch": 28.936244541484715,
+ "grad_norm": 0.01497623696923256,
+ "learning_rate": 0.0006,
+ "loss": 4.24141788482666,
+ "step": 2083
+ },
+ {
+ "epoch": 28.950218340611354,
+ "grad_norm": 0.015080046840012074,
+ "learning_rate": 0.0006,
+ "loss": 4.284151077270508,
+ "step": 2084
+ },
+ {
+ "epoch": 28.964192139737992,
+ "grad_norm": 0.016207212582230568,
+ "learning_rate": 0.0006,
+ "loss": 4.298693656921387,
+ "step": 2085
+ },
+ {
+ "epoch": 28.978165938864628,
+ "grad_norm": 0.015474417246878147,
+ "learning_rate": 0.0006,
+ "loss": 4.2412261962890625,
+ "step": 2086
+ },
+ {
+ "epoch": 28.992139737991266,
+ "grad_norm": 0.01515351701527834,
+ "learning_rate": 0.0006,
+ "loss": 4.3054423332214355,
+ "step": 2087
+ },
+ {
+ "epoch": 29.0,
+ "grad_norm": 0.016411345452070236,
+ "learning_rate": 0.0006,
+ "loss": 4.104830741882324,
+ "step": 2088
+ },
+ {
+ "epoch": 29.0,
+ "eval_loss": 4.661567687988281,
+ "eval_runtime": 57.1244,
+ "eval_samples_per_second": 42.749,
+ "eval_steps_per_second": 1.348,
+ "step": 2088
+ },
+ {
+ "epoch": 29.01397379912664,
+ "grad_norm": 0.01665342226624489,
+ "learning_rate": 0.0006,
+ "loss": 4.311391830444336,
+ "step": 2089
+ },
+ {
+ "epoch": 29.027947598253274,
+ "grad_norm": 0.018053214997053146,
+ "learning_rate": 0.0006,
+ "loss": 4.135597229003906,
+ "step": 2090
+ },
+ {
+ "epoch": 29.041921397379912,
+ "grad_norm": 0.018213748931884766,
+ "learning_rate": 0.0006,
+ "loss": 4.303299903869629,
+ "step": 2091
+ },
+ {
+ "epoch": 29.05589519650655,
+ "grad_norm": 0.018460242077708244,
+ "learning_rate": 0.0006,
+ "loss": 4.20850944519043,
+ "step": 2092
+ },
+ {
+ "epoch": 29.069868995633186,
+ "grad_norm": 0.018365809693932533,
+ "learning_rate": 0.0006,
+ "loss": 4.275847434997559,
+ "step": 2093
+ },
+ {
+ "epoch": 29.083842794759825,
+ "grad_norm": 0.018346186727285385,
+ "learning_rate": 0.0006,
+ "loss": 4.202091217041016,
+ "step": 2094
+ },
+ {
+ "epoch": 29.097816593886463,
+ "grad_norm": 0.017920587211847305,
+ "learning_rate": 0.0006,
+ "loss": 4.182092666625977,
+ "step": 2095
+ },
+ {
+ "epoch": 29.111790393013102,
+ "grad_norm": 0.01812003180384636,
+ "learning_rate": 0.0006,
+ "loss": 4.259721755981445,
+ "step": 2096
+ },
+ {
+ "epoch": 29.125764192139737,
+ "grad_norm": 0.01700986735522747,
+ "learning_rate": 0.0006,
+ "loss": 4.067573070526123,
+ "step": 2097
+ },
+ {
+ "epoch": 29.139737991266376,
+ "grad_norm": 0.018069760873913765,
+ "learning_rate": 0.0006,
+ "loss": 4.234400749206543,
+ "step": 2098
+ },
+ {
+ "epoch": 29.153711790393015,
+ "grad_norm": 0.01905428245663643,
+ "learning_rate": 0.0006,
+ "loss": 4.16609001159668,
+ "step": 2099
+ },
+ {
+ "epoch": 29.16768558951965,
+ "grad_norm": 0.018711242824792862,
+ "learning_rate": 0.0006,
+ "loss": 4.249088764190674,
+ "step": 2100
+ },
+ {
+ "epoch": 29.18165938864629,
+ "grad_norm": 0.018721306696534157,
+ "learning_rate": 0.0006,
+ "loss": 4.14540958404541,
+ "step": 2101
+ },
+ {
+ "epoch": 29.195633187772927,
+ "grad_norm": 0.020139874890446663,
+ "learning_rate": 0.0006,
+ "loss": 4.049248695373535,
+ "step": 2102
+ },
+ {
+ "epoch": 29.209606986899562,
+ "grad_norm": 0.020849574357271194,
+ "learning_rate": 0.0006,
+ "loss": 4.235224723815918,
+ "step": 2103
+ },
+ {
+ "epoch": 29.2235807860262,
+ "grad_norm": 0.018693549558520317,
+ "learning_rate": 0.0006,
+ "loss": 4.232961177825928,
+ "step": 2104
+ },
+ {
+ "epoch": 29.23755458515284,
+ "grad_norm": 0.017889857292175293,
+ "learning_rate": 0.0006,
+ "loss": 4.171422004699707,
+ "step": 2105
+ },
+ {
+ "epoch": 29.251528384279474,
+ "grad_norm": 0.016690224409103394,
+ "learning_rate": 0.0006,
+ "loss": 4.244717597961426,
+ "step": 2106
+ },
+ {
+ "epoch": 29.265502183406113,
+ "grad_norm": 0.01708616502583027,
+ "learning_rate": 0.0006,
+ "loss": 4.125068664550781,
+ "step": 2107
+ },
+ {
+ "epoch": 29.27947598253275,
+ "grad_norm": 0.017928237095475197,
+ "learning_rate": 0.0006,
+ "loss": 4.208292007446289,
+ "step": 2108
+ },
+ {
+ "epoch": 29.29344978165939,
+ "grad_norm": 0.018250394612550735,
+ "learning_rate": 0.0006,
+ "loss": 4.118630886077881,
+ "step": 2109
+ },
+ {
+ "epoch": 29.307423580786025,
+ "grad_norm": 0.020160246640443802,
+ "learning_rate": 0.0006,
+ "loss": 4.206025123596191,
+ "step": 2110
+ },
+ {
+ "epoch": 29.321397379912664,
+ "grad_norm": 0.020045453682541847,
+ "learning_rate": 0.0006,
+ "loss": 4.2758684158325195,
+ "step": 2111
+ },
+ {
+ "epoch": 29.335371179039303,
+ "grad_norm": 0.019556907936930656,
+ "learning_rate": 0.0006,
+ "loss": 4.188272953033447,
+ "step": 2112
+ },
+ {
+ "epoch": 29.349344978165938,
+ "grad_norm": 0.019185200333595276,
+ "learning_rate": 0.0006,
+ "loss": 4.17451286315918,
+ "step": 2113
+ },
+ {
+ "epoch": 29.363318777292577,
+ "grad_norm": 0.01816629432141781,
+ "learning_rate": 0.0006,
+ "loss": 4.165246963500977,
+ "step": 2114
+ },
+ {
+ "epoch": 29.377292576419215,
+ "grad_norm": 0.01865146867930889,
+ "learning_rate": 0.0006,
+ "loss": 4.148184776306152,
+ "step": 2115
+ },
+ {
+ "epoch": 29.39126637554585,
+ "grad_norm": 0.018338464200496674,
+ "learning_rate": 0.0006,
+ "loss": 4.288936614990234,
+ "step": 2116
+ },
+ {
+ "epoch": 29.40524017467249,
+ "grad_norm": 0.018986187875270844,
+ "learning_rate": 0.0006,
+ "loss": 4.183379650115967,
+ "step": 2117
+ },
+ {
+ "epoch": 29.419213973799128,
+ "grad_norm": 0.018409637734293938,
+ "learning_rate": 0.0006,
+ "loss": 4.057574272155762,
+ "step": 2118
+ },
+ {
+ "epoch": 29.433187772925763,
+ "grad_norm": 0.016664542257785797,
+ "learning_rate": 0.0006,
+ "loss": 4.18517541885376,
+ "step": 2119
+ },
+ {
+ "epoch": 29.4471615720524,
+ "grad_norm": 0.016486089676618576,
+ "learning_rate": 0.0006,
+ "loss": 4.179367542266846,
+ "step": 2120
+ },
+ {
+ "epoch": 29.46113537117904,
+ "grad_norm": 0.01787523552775383,
+ "learning_rate": 0.0006,
+ "loss": 4.184126377105713,
+ "step": 2121
+ },
+ {
+ "epoch": 29.475109170305675,
+ "grad_norm": 0.016574613749980927,
+ "learning_rate": 0.0006,
+ "loss": 4.290170192718506,
+ "step": 2122
+ },
+ {
+ "epoch": 29.489082969432314,
+ "grad_norm": 0.01643376611173153,
+ "learning_rate": 0.0006,
+ "loss": 4.052913665771484,
+ "step": 2123
+ },
+ {
+ "epoch": 29.503056768558952,
+ "grad_norm": 0.016503969207406044,
+ "learning_rate": 0.0006,
+ "loss": 4.335184097290039,
+ "step": 2124
+ },
+ {
+ "epoch": 29.51703056768559,
+ "grad_norm": 0.017596479505300522,
+ "learning_rate": 0.0006,
+ "loss": 4.165539741516113,
+ "step": 2125
+ },
+ {
+ "epoch": 29.531004366812226,
+ "grad_norm": 0.017468582838773727,
+ "learning_rate": 0.0006,
+ "loss": 4.315964698791504,
+ "step": 2126
+ },
+ {
+ "epoch": 29.544978165938865,
+ "grad_norm": 0.01724618673324585,
+ "learning_rate": 0.0006,
+ "loss": 4.243681907653809,
+ "step": 2127
+ },
+ {
+ "epoch": 29.558951965065503,
+ "grad_norm": 0.018384616822004318,
+ "learning_rate": 0.0006,
+ "loss": 4.112447261810303,
+ "step": 2128
+ },
+ {
+ "epoch": 29.57292576419214,
+ "grad_norm": 0.01824074238538742,
+ "learning_rate": 0.0006,
+ "loss": 4.236065864562988,
+ "step": 2129
+ },
+ {
+ "epoch": 29.586899563318777,
+ "grad_norm": 0.01625971868634224,
+ "learning_rate": 0.0006,
+ "loss": 4.18143892288208,
+ "step": 2130
+ },
+ {
+ "epoch": 29.600873362445416,
+ "grad_norm": 0.015537494793534279,
+ "learning_rate": 0.0006,
+ "loss": 4.225147247314453,
+ "step": 2131
+ },
+ {
+ "epoch": 29.61484716157205,
+ "grad_norm": 0.015818974003195763,
+ "learning_rate": 0.0006,
+ "loss": 4.3100175857543945,
+ "step": 2132
+ },
+ {
+ "epoch": 29.62882096069869,
+ "grad_norm": 0.015909671783447266,
+ "learning_rate": 0.0006,
+ "loss": 4.169775009155273,
+ "step": 2133
+ },
+ {
+ "epoch": 29.64279475982533,
+ "grad_norm": 0.01769658550620079,
+ "learning_rate": 0.0006,
+ "loss": 4.278904914855957,
+ "step": 2134
+ },
+ {
+ "epoch": 29.656768558951963,
+ "grad_norm": 0.015641704201698303,
+ "learning_rate": 0.0006,
+ "loss": 4.2426228523254395,
+ "step": 2135
+ },
+ {
+ "epoch": 29.670742358078602,
+ "grad_norm": 0.01565658301115036,
+ "learning_rate": 0.0006,
+ "loss": 4.232687950134277,
+ "step": 2136
+ },
+ {
+ "epoch": 29.68471615720524,
+ "grad_norm": 0.016485940665006638,
+ "learning_rate": 0.0006,
+ "loss": 4.209839820861816,
+ "step": 2137
+ },
+ {
+ "epoch": 29.69868995633188,
+ "grad_norm": 0.017622729763388634,
+ "learning_rate": 0.0006,
+ "loss": 4.196017265319824,
+ "step": 2138
+ },
+ {
+ "epoch": 29.712663755458514,
+ "grad_norm": 0.018320564180612564,
+ "learning_rate": 0.0006,
+ "loss": 4.225683689117432,
+ "step": 2139
+ },
+ {
+ "epoch": 29.726637554585153,
+ "grad_norm": 0.018140794709324837,
+ "learning_rate": 0.0006,
+ "loss": 4.311944961547852,
+ "step": 2140
+ },
+ {
+ "epoch": 29.74061135371179,
+ "grad_norm": 0.016227100044488907,
+ "learning_rate": 0.0006,
+ "loss": 4.230257511138916,
+ "step": 2141
+ },
+ {
+ "epoch": 29.754585152838427,
+ "grad_norm": 0.015759488567709923,
+ "learning_rate": 0.0006,
+ "loss": 4.119174003601074,
+ "step": 2142
+ },
+ {
+ "epoch": 29.768558951965066,
+ "grad_norm": 0.01757766678929329,
+ "learning_rate": 0.0006,
+ "loss": 4.165826797485352,
+ "step": 2143
+ },
+ {
+ "epoch": 29.782532751091704,
+ "grad_norm": 0.019661923870444298,
+ "learning_rate": 0.0006,
+ "loss": 4.392203330993652,
+ "step": 2144
+ },
+ {
+ "epoch": 29.79650655021834,
+ "grad_norm": 0.019478755071759224,
+ "learning_rate": 0.0006,
+ "loss": 4.348371982574463,
+ "step": 2145
+ },
+ {
+ "epoch": 29.810480349344978,
+ "grad_norm": 0.019149569794535637,
+ "learning_rate": 0.0006,
+ "loss": 4.34104061126709,
+ "step": 2146
+ },
+ {
+ "epoch": 29.824454148471617,
+ "grad_norm": 0.017707521095871925,
+ "learning_rate": 0.0006,
+ "loss": 4.228226661682129,
+ "step": 2147
+ },
+ {
+ "epoch": 29.83842794759825,
+ "grad_norm": 0.015615378506481647,
+ "learning_rate": 0.0006,
+ "loss": 4.175541877746582,
+ "step": 2148
+ },
+ {
+ "epoch": 29.85240174672489,
+ "grad_norm": 0.016111129894852638,
+ "learning_rate": 0.0006,
+ "loss": 4.267936706542969,
+ "step": 2149
+ },
+ {
+ "epoch": 29.86637554585153,
+ "grad_norm": 0.01528779324144125,
+ "learning_rate": 0.0006,
+ "loss": 4.186014175415039,
+ "step": 2150
+ },
+ {
+ "epoch": 29.880349344978168,
+ "grad_norm": 0.014925646595656872,
+ "learning_rate": 0.0006,
+ "loss": 4.174860000610352,
+ "step": 2151
+ },
+ {
+ "epoch": 29.894323144104803,
+ "grad_norm": 0.016917673870921135,
+ "learning_rate": 0.0006,
+ "loss": 4.33908748626709,
+ "step": 2152
+ },
+ {
+ "epoch": 29.90829694323144,
+ "grad_norm": 0.016007075086236,
+ "learning_rate": 0.0006,
+ "loss": 4.266141414642334,
+ "step": 2153
+ },
+ {
+ "epoch": 29.92227074235808,
+ "grad_norm": 0.016178956255316734,
+ "learning_rate": 0.0006,
+ "loss": 4.271215915679932,
+ "step": 2154
+ },
+ {
+ "epoch": 29.936244541484715,
+ "grad_norm": 0.016180871054530144,
+ "learning_rate": 0.0006,
+ "loss": 4.278801441192627,
+ "step": 2155
+ },
+ {
+ "epoch": 29.950218340611354,
+ "grad_norm": 0.015335503034293652,
+ "learning_rate": 0.0006,
+ "loss": 4.194978713989258,
+ "step": 2156
+ },
+ {
+ "epoch": 29.964192139737992,
+ "grad_norm": 0.016580764204263687,
+ "learning_rate": 0.0006,
+ "loss": 4.290809631347656,
+ "step": 2157
+ },
+ {
+ "epoch": 29.978165938864628,
+ "grad_norm": 0.01623128168284893,
+ "learning_rate": 0.0006,
+ "loss": 4.205451488494873,
+ "step": 2158
+ },
+ {
+ "epoch": 29.992139737991266,
+ "grad_norm": 0.014904248528182507,
+ "learning_rate": 0.0006,
+ "loss": 4.24924373626709,
+ "step": 2159
+ },
+ {
+ "epoch": 30.0,
+ "grad_norm": 0.017089299857616425,
+ "learning_rate": 0.0006,
+ "loss": 4.222072601318359,
+ "step": 2160
+ },
+ {
+ "epoch": 30.0,
+ "eval_loss": 4.585411071777344,
+ "eval_runtime": 57.1164,
+ "eval_samples_per_second": 42.755,
+ "eval_steps_per_second": 1.348,
+ "step": 2160
+ },
+ {
+ "epoch": 30.01397379912664,
+ "grad_norm": 0.017226383090019226,
+ "learning_rate": 0.0006,
+ "loss": 4.19467830657959,
+ "step": 2161
+ },
+ {
+ "epoch": 30.027947598253274,
+ "grad_norm": 0.01635066606104374,
+ "learning_rate": 0.0006,
+ "loss": 4.176875591278076,
+ "step": 2162
+ },
+ {
+ "epoch": 30.041921397379912,
+ "grad_norm": 0.01735362410545349,
+ "learning_rate": 0.0006,
+ "loss": 4.13667106628418,
+ "step": 2163
+ },
+ {
+ "epoch": 30.05589519650655,
+ "grad_norm": 0.017047109082341194,
+ "learning_rate": 0.0006,
+ "loss": 4.136329650878906,
+ "step": 2164
+ },
+ {
+ "epoch": 30.069868995633186,
+ "grad_norm": 0.018116053193807602,
+ "learning_rate": 0.0006,
+ "loss": 4.228706359863281,
+ "step": 2165
+ },
+ {
+ "epoch": 30.083842794759825,
+ "grad_norm": 0.01791389286518097,
+ "learning_rate": 0.0006,
+ "loss": 4.1597747802734375,
+ "step": 2166
+ },
+ {
+ "epoch": 30.097816593886463,
+ "grad_norm": 0.017793502658605576,
+ "learning_rate": 0.0006,
+ "loss": 4.0807976722717285,
+ "step": 2167
+ },
+ {
+ "epoch": 30.111790393013102,
+ "grad_norm": 0.01827503927052021,
+ "learning_rate": 0.0006,
+ "loss": 4.170646667480469,
+ "step": 2168
+ },
+ {
+ "epoch": 30.125764192139737,
+ "grad_norm": 0.017228564247488976,
+ "learning_rate": 0.0006,
+ "loss": 4.146934509277344,
+ "step": 2169
+ },
+ {
+ "epoch": 30.139737991266376,
+ "grad_norm": 0.01757437363266945,
+ "learning_rate": 0.0006,
+ "loss": 4.1859130859375,
+ "step": 2170
+ },
+ {
+ "epoch": 30.153711790393015,
+ "grad_norm": 0.01701093092560768,
+ "learning_rate": 0.0006,
+ "loss": 4.126981258392334,
+ "step": 2171
+ },
+ {
+ "epoch": 30.16768558951965,
+ "grad_norm": 0.015220489352941513,
+ "learning_rate": 0.0006,
+ "loss": 4.164018630981445,
+ "step": 2172
+ },
+ {
+ "epoch": 30.18165938864629,
+ "grad_norm": 0.016254238784313202,
+ "learning_rate": 0.0006,
+ "loss": 4.258275508880615,
+ "step": 2173
+ },
+ {
+ "epoch": 30.195633187772927,
+ "grad_norm": 0.016609614714980125,
+ "learning_rate": 0.0006,
+ "loss": 4.222233772277832,
+ "step": 2174
+ },
+ {
+ "epoch": 30.209606986899562,
+ "grad_norm": 0.016531318426132202,
+ "learning_rate": 0.0006,
+ "loss": 4.144489288330078,
+ "step": 2175
+ },
+ {
+ "epoch": 30.2235807860262,
+ "grad_norm": 0.01567668654024601,
+ "learning_rate": 0.0006,
+ "loss": 4.181916236877441,
+ "step": 2176
+ },
+ {
+ "epoch": 30.23755458515284,
+ "grad_norm": 0.017195681110024452,
+ "learning_rate": 0.0006,
+ "loss": 4.189693927764893,
+ "step": 2177
+ },
+ {
+ "epoch": 30.251528384279474,
+ "grad_norm": 0.016882948577404022,
+ "learning_rate": 0.0006,
+ "loss": 4.21604061126709,
+ "step": 2178
+ },
+ {
+ "epoch": 30.265502183406113,
+ "grad_norm": 0.016261622309684753,
+ "learning_rate": 0.0006,
+ "loss": 4.227639198303223,
+ "step": 2179
+ },
+ {
+ "epoch": 30.27947598253275,
+ "grad_norm": 0.016411812976002693,
+ "learning_rate": 0.0006,
+ "loss": 4.130362510681152,
+ "step": 2180
+ },
+ {
+ "epoch": 30.29344978165939,
+ "grad_norm": 0.01732019893825054,
+ "learning_rate": 0.0006,
+ "loss": 4.230443000793457,
+ "step": 2181
+ },
+ {
+ "epoch": 30.307423580786025,
+ "grad_norm": 0.01825905591249466,
+ "learning_rate": 0.0006,
+ "loss": 4.175817489624023,
+ "step": 2182
+ },
+ {
+ "epoch": 30.321397379912664,
+ "grad_norm": 0.01887713558971882,
+ "learning_rate": 0.0006,
+ "loss": 4.169020652770996,
+ "step": 2183
+ },
+ {
+ "epoch": 30.335371179039303,
+ "grad_norm": 0.019391661509871483,
+ "learning_rate": 0.0006,
+ "loss": 4.176318645477295,
+ "step": 2184
+ },
+ {
+ "epoch": 30.349344978165938,
+ "grad_norm": 0.020840002223849297,
+ "learning_rate": 0.0006,
+ "loss": 4.174228668212891,
+ "step": 2185
+ },
+ {
+ "epoch": 30.363318777292577,
+ "grad_norm": 0.020193178206682205,
+ "learning_rate": 0.0006,
+ "loss": 4.127919673919678,
+ "step": 2186
+ },
+ {
+ "epoch": 30.377292576419215,
+ "grad_norm": 0.018449561670422554,
+ "learning_rate": 0.0006,
+ "loss": 4.229487895965576,
+ "step": 2187
+ },
+ {
+ "epoch": 30.39126637554585,
+ "grad_norm": 0.01711983233690262,
+ "learning_rate": 0.0006,
+ "loss": 4.201651573181152,
+ "step": 2188
+ },
+ {
+ "epoch": 30.40524017467249,
+ "grad_norm": 0.017293395474553108,
+ "learning_rate": 0.0006,
+ "loss": 4.168704032897949,
+ "step": 2189
+ },
+ {
+ "epoch": 30.419213973799128,
+ "grad_norm": 0.018526358529925346,
+ "learning_rate": 0.0006,
+ "loss": 4.171864032745361,
+ "step": 2190
+ },
+ {
+ "epoch": 30.433187772925763,
+ "grad_norm": 0.019287291914224625,
+ "learning_rate": 0.0006,
+ "loss": 4.039534568786621,
+ "step": 2191
+ },
+ {
+ "epoch": 30.4471615720524,
+ "grad_norm": 0.016902444884181023,
+ "learning_rate": 0.0006,
+ "loss": 4.282787799835205,
+ "step": 2192
+ },
+ {
+ "epoch": 30.46113537117904,
+ "grad_norm": 0.01656750962138176,
+ "learning_rate": 0.0006,
+ "loss": 4.213098526000977,
+ "step": 2193
+ },
+ {
+ "epoch": 30.475109170305675,
+ "grad_norm": 0.016805065795779228,
+ "learning_rate": 0.0006,
+ "loss": 4.1583476066589355,
+ "step": 2194
+ },
+ {
+ "epoch": 30.489082969432314,
+ "grad_norm": 0.017166638746857643,
+ "learning_rate": 0.0006,
+ "loss": 4.188634395599365,
+ "step": 2195
+ },
+ {
+ "epoch": 30.503056768558952,
+ "grad_norm": 0.01592106744647026,
+ "learning_rate": 0.0006,
+ "loss": 4.234043121337891,
+ "step": 2196
+ },
+ {
+ "epoch": 30.51703056768559,
+ "grad_norm": 0.015270989388227463,
+ "learning_rate": 0.0006,
+ "loss": 4.301420211791992,
+ "step": 2197
+ },
+ {
+ "epoch": 30.531004366812226,
+ "grad_norm": 0.015194724313914776,
+ "learning_rate": 0.0006,
+ "loss": 4.18703031539917,
+ "step": 2198
+ },
+ {
+ "epoch": 30.544978165938865,
+ "grad_norm": 0.015587746165692806,
+ "learning_rate": 0.0006,
+ "loss": 4.138928413391113,
+ "step": 2199
+ },
+ {
+ "epoch": 30.558951965065503,
+ "grad_norm": 0.016029570251703262,
+ "learning_rate": 0.0006,
+ "loss": 4.290863990783691,
+ "step": 2200
+ },
+ {
+ "epoch": 30.57292576419214,
+ "grad_norm": 0.015651309862732887,
+ "learning_rate": 0.0006,
+ "loss": 4.272062301635742,
+ "step": 2201
+ },
+ {
+ "epoch": 30.586899563318777,
+ "grad_norm": 0.015992306172847748,
+ "learning_rate": 0.0006,
+ "loss": 4.2487406730651855,
+ "step": 2202
+ },
+ {
+ "epoch": 30.600873362445416,
+ "grad_norm": 0.014969157986342907,
+ "learning_rate": 0.0006,
+ "loss": 4.115347862243652,
+ "step": 2203
+ },
+ {
+ "epoch": 30.61484716157205,
+ "grad_norm": 0.014639027416706085,
+ "learning_rate": 0.0006,
+ "loss": 4.2207489013671875,
+ "step": 2204
+ },
+ {
+ "epoch": 30.62882096069869,
+ "grad_norm": 0.015059413388371468,
+ "learning_rate": 0.0006,
+ "loss": 4.274044990539551,
+ "step": 2205
+ },
+ {
+ "epoch": 30.64279475982533,
+ "grad_norm": 0.0166641678661108,
+ "learning_rate": 0.0006,
+ "loss": 4.191373348236084,
+ "step": 2206
+ },
+ {
+ "epoch": 30.656768558951963,
+ "grad_norm": 0.017806345596909523,
+ "learning_rate": 0.0006,
+ "loss": 4.139800548553467,
+ "step": 2207
+ },
+ {
+ "epoch": 30.670742358078602,
+ "grad_norm": 0.018712421879172325,
+ "learning_rate": 0.0006,
+ "loss": 4.035965442657471,
+ "step": 2208
+ },
+ {
+ "epoch": 30.68471615720524,
+ "grad_norm": 0.018457243219017982,
+ "learning_rate": 0.0006,
+ "loss": 4.2579240798950195,
+ "step": 2209
+ },
+ {
+ "epoch": 30.69868995633188,
+ "grad_norm": 0.01920999586582184,
+ "learning_rate": 0.0006,
+ "loss": 4.131112098693848,
+ "step": 2210
+ },
+ {
+ "epoch": 30.712663755458514,
+ "grad_norm": 0.01729944534599781,
+ "learning_rate": 0.0006,
+ "loss": 4.254059314727783,
+ "step": 2211
+ },
+ {
+ "epoch": 30.726637554585153,
+ "grad_norm": 0.016914231702685356,
+ "learning_rate": 0.0006,
+ "loss": 4.233428955078125,
+ "step": 2212
+ },
+ {
+ "epoch": 30.74061135371179,
+ "grad_norm": 0.01913098618388176,
+ "learning_rate": 0.0006,
+ "loss": 4.23525333404541,
+ "step": 2213
+ },
+ {
+ "epoch": 30.754585152838427,
+ "grad_norm": 0.019150281324982643,
+ "learning_rate": 0.0006,
+ "loss": 4.213695049285889,
+ "step": 2214
+ },
+ {
+ "epoch": 30.768558951965066,
+ "grad_norm": 0.018947452306747437,
+ "learning_rate": 0.0006,
+ "loss": 4.151697158813477,
+ "step": 2215
+ },
+ {
+ "epoch": 30.782532751091704,
+ "grad_norm": 0.016058508306741714,
+ "learning_rate": 0.0006,
+ "loss": 4.220280170440674,
+ "step": 2216
+ },
+ {
+ "epoch": 30.79650655021834,
+ "grad_norm": 0.017488619312644005,
+ "learning_rate": 0.0006,
+ "loss": 4.154428482055664,
+ "step": 2217
+ },
+ {
+ "epoch": 30.810480349344978,
+ "grad_norm": 0.017355090007185936,
+ "learning_rate": 0.0006,
+ "loss": 4.256086349487305,
+ "step": 2218
+ },
+ {
+ "epoch": 30.824454148471617,
+ "grad_norm": 0.016636032611131668,
+ "learning_rate": 0.0006,
+ "loss": 4.227564334869385,
+ "step": 2219
+ },
+ {
+ "epoch": 30.83842794759825,
+ "grad_norm": 0.016273891553282738,
+ "learning_rate": 0.0006,
+ "loss": 4.200466156005859,
+ "step": 2220
+ },
+ {
+ "epoch": 30.85240174672489,
+ "grad_norm": 0.016939911991357803,
+ "learning_rate": 0.0006,
+ "loss": 4.1571221351623535,
+ "step": 2221
+ },
+ {
+ "epoch": 30.86637554585153,
+ "grad_norm": 0.018332522362470627,
+ "learning_rate": 0.0006,
+ "loss": 4.228398323059082,
+ "step": 2222
+ },
+ {
+ "epoch": 30.880349344978168,
+ "grad_norm": 0.019508114084601402,
+ "learning_rate": 0.0006,
+ "loss": 4.222960472106934,
+ "step": 2223
+ },
+ {
+ "epoch": 30.894323144104803,
+ "grad_norm": 0.018440278246998787,
+ "learning_rate": 0.0006,
+ "loss": 4.261051177978516,
+ "step": 2224
+ },
+ {
+ "epoch": 30.90829694323144,
+ "grad_norm": 0.018113229423761368,
+ "learning_rate": 0.0006,
+ "loss": 4.109235763549805,
+ "step": 2225
+ },
+ {
+ "epoch": 30.92227074235808,
+ "grad_norm": 0.015392083674669266,
+ "learning_rate": 0.0006,
+ "loss": 4.21950101852417,
+ "step": 2226
+ },
+ {
+ "epoch": 30.936244541484715,
+ "grad_norm": 0.01633290946483612,
+ "learning_rate": 0.0006,
+ "loss": 4.176053047180176,
+ "step": 2227
+ },
+ {
+ "epoch": 30.950218340611354,
+ "grad_norm": 0.016236383467912674,
+ "learning_rate": 0.0006,
+ "loss": 4.314309120178223,
+ "step": 2228
+ },
+ {
+ "epoch": 30.964192139737992,
+ "grad_norm": 0.015284373424947262,
+ "learning_rate": 0.0006,
+ "loss": 4.303133487701416,
+ "step": 2229
+ },
+ {
+ "epoch": 30.978165938864628,
+ "grad_norm": 0.015982897952198982,
+ "learning_rate": 0.0006,
+ "loss": 4.25739860534668,
+ "step": 2230
+ },
+ {
+ "epoch": 30.992139737991266,
+ "grad_norm": 0.016682198271155357,
+ "learning_rate": 0.0006,
+ "loss": 4.292210578918457,
+ "step": 2231
+ },
+ {
+ "epoch": 31.0,
+ "grad_norm": 0.017521467059850693,
+ "learning_rate": 0.0006,
+ "loss": 4.170637130737305,
+ "step": 2232
+ },
+ {
+ "epoch": 31.0,
+ "eval_loss": 4.642786979675293,
+ "eval_runtime": 57.029,
+ "eval_samples_per_second": 42.82,
+ "eval_steps_per_second": 1.35,
+ "step": 2232
+ },
+ {
+ "epoch": 31.01397379912664,
+ "grad_norm": 0.016607852652668953,
+ "learning_rate": 0.0006,
+ "loss": 4.106224536895752,
+ "step": 2233
+ },
+ {
+ "epoch": 31.027947598253274,
+ "grad_norm": 0.018448904156684875,
+ "learning_rate": 0.0006,
+ "loss": 4.224459648132324,
+ "step": 2234
+ },
+ {
+ "epoch": 31.041921397379912,
+ "grad_norm": 0.020481228828430176,
+ "learning_rate": 0.0006,
+ "loss": 4.305548667907715,
+ "step": 2235
+ },
+ {
+ "epoch": 31.05589519650655,
+ "grad_norm": 0.01999707892537117,
+ "learning_rate": 0.0006,
+ "loss": 4.253663539886475,
+ "step": 2236
+ },
+ {
+ "epoch": 31.069868995633186,
+ "grad_norm": 0.020428303629159927,
+ "learning_rate": 0.0006,
+ "loss": 4.166954040527344,
+ "step": 2237
+ },
+ {
+ "epoch": 31.083842794759825,
+ "grad_norm": 0.022608213126659393,
+ "learning_rate": 0.0006,
+ "loss": 4.172819137573242,
+ "step": 2238
+ },
+ {
+ "epoch": 31.097816593886463,
+ "grad_norm": 0.024669520556926727,
+ "learning_rate": 0.0006,
+ "loss": 4.106586456298828,
+ "step": 2239
+ },
+ {
+ "epoch": 31.111790393013102,
+ "grad_norm": 0.02247351035475731,
+ "learning_rate": 0.0006,
+ "loss": 4.160271644592285,
+ "step": 2240
+ },
+ {
+ "epoch": 31.125764192139737,
+ "grad_norm": 0.0207810141146183,
+ "learning_rate": 0.0006,
+ "loss": 4.272491931915283,
+ "step": 2241
+ },
+ {
+ "epoch": 31.139737991266376,
+ "grad_norm": 0.021576499566435814,
+ "learning_rate": 0.0006,
+ "loss": 4.171474456787109,
+ "step": 2242
+ },
+ {
+ "epoch": 31.153711790393015,
+ "grad_norm": 0.02337106689810753,
+ "learning_rate": 0.0006,
+ "loss": 4.094712734222412,
+ "step": 2243
+ },
+ {
+ "epoch": 31.16768558951965,
+ "grad_norm": 0.02219875156879425,
+ "learning_rate": 0.0006,
+ "loss": 4.1800994873046875,
+ "step": 2244
+ },
+ {
+ "epoch": 31.18165938864629,
+ "grad_norm": 0.022680504247546196,
+ "learning_rate": 0.0006,
+ "loss": 4.2054362297058105,
+ "step": 2245
+ },
+ {
+ "epoch": 31.195633187772927,
+ "grad_norm": 0.0208734143525362,
+ "learning_rate": 0.0006,
+ "loss": 4.06068229675293,
+ "step": 2246
+ },
+ {
+ "epoch": 31.209606986899562,
+ "grad_norm": 0.01811066083610058,
+ "learning_rate": 0.0006,
+ "loss": 4.2465128898620605,
+ "step": 2247
+ },
+ {
+ "epoch": 31.2235807860262,
+ "grad_norm": 0.018571637570858,
+ "learning_rate": 0.0006,
+ "loss": 4.191955089569092,
+ "step": 2248
+ },
+ {
+ "epoch": 31.23755458515284,
+ "grad_norm": 0.018598807975649834,
+ "learning_rate": 0.0006,
+ "loss": 4.200872421264648,
+ "step": 2249
+ },
+ {
+ "epoch": 31.251528384279474,
+ "grad_norm": 0.017577625811100006,
+ "learning_rate": 0.0006,
+ "loss": 4.148700714111328,
+ "step": 2250
+ },
+ {
+ "epoch": 31.265502183406113,
+ "grad_norm": 0.01817980408668518,
+ "learning_rate": 0.0006,
+ "loss": 4.256978988647461,
+ "step": 2251
+ },
+ {
+ "epoch": 31.27947598253275,
+ "grad_norm": 0.017983486875891685,
+ "learning_rate": 0.0006,
+ "loss": 4.175005912780762,
+ "step": 2252
+ },
+ {
+ "epoch": 31.29344978165939,
+ "grad_norm": 0.016751961782574654,
+ "learning_rate": 0.0006,
+ "loss": 4.237518310546875,
+ "step": 2253
+ },
+ {
+ "epoch": 31.307423580786025,
+ "grad_norm": 0.016407202929258347,
+ "learning_rate": 0.0006,
+ "loss": 4.197269439697266,
+ "step": 2254
+ },
+ {
+ "epoch": 31.321397379912664,
+ "grad_norm": 0.01578645594418049,
+ "learning_rate": 0.0006,
+ "loss": 4.179751396179199,
+ "step": 2255
+ },
+ {
+ "epoch": 31.335371179039303,
+ "grad_norm": 0.016001226380467415,
+ "learning_rate": 0.0006,
+ "loss": 4.08821964263916,
+ "step": 2256
+ },
+ {
+ "epoch": 31.349344978165938,
+ "grad_norm": 0.015122811309993267,
+ "learning_rate": 0.0006,
+ "loss": 4.047828674316406,
+ "step": 2257
+ },
+ {
+ "epoch": 31.363318777292577,
+ "grad_norm": 0.015399965457618237,
+ "learning_rate": 0.0006,
+ "loss": 4.1506524085998535,
+ "step": 2258
+ },
+ {
+ "epoch": 31.377292576419215,
+ "grad_norm": 0.01534605398774147,
+ "learning_rate": 0.0006,
+ "loss": 4.27952766418457,
+ "step": 2259
+ },
+ {
+ "epoch": 31.39126637554585,
+ "grad_norm": 0.015461381524801254,
+ "learning_rate": 0.0006,
+ "loss": 4.27102518081665,
+ "step": 2260
+ },
+ {
+ "epoch": 31.40524017467249,
+ "grad_norm": 0.016769153997302055,
+ "learning_rate": 0.0006,
+ "loss": 4.191592216491699,
+ "step": 2261
+ },
+ {
+ "epoch": 31.419213973799128,
+ "grad_norm": 0.016557959839701653,
+ "learning_rate": 0.0006,
+ "loss": 4.159705638885498,
+ "step": 2262
+ },
+ {
+ "epoch": 31.433187772925763,
+ "grad_norm": 0.01659569889307022,
+ "learning_rate": 0.0006,
+ "loss": 4.2136664390563965,
+ "step": 2263
+ },
+ {
+ "epoch": 31.4471615720524,
+ "grad_norm": 0.017101937904953957,
+ "learning_rate": 0.0006,
+ "loss": 4.2716498374938965,
+ "step": 2264
+ },
+ {
+ "epoch": 31.46113537117904,
+ "grad_norm": 0.01722985878586769,
+ "learning_rate": 0.0006,
+ "loss": 4.300534248352051,
+ "step": 2265
+ },
+ {
+ "epoch": 31.475109170305675,
+ "grad_norm": 0.016800185665488243,
+ "learning_rate": 0.0006,
+ "loss": 4.113240718841553,
+ "step": 2266
+ },
+ {
+ "epoch": 31.489082969432314,
+ "grad_norm": 0.01603875495493412,
+ "learning_rate": 0.0006,
+ "loss": 4.157105445861816,
+ "step": 2267
+ },
+ {
+ "epoch": 31.503056768558952,
+ "grad_norm": 0.015903932973742485,
+ "learning_rate": 0.0006,
+ "loss": 4.233669757843018,
+ "step": 2268
+ },
+ {
+ "epoch": 31.51703056768559,
+ "grad_norm": 0.0160983856767416,
+ "learning_rate": 0.0006,
+ "loss": 4.147704124450684,
+ "step": 2269
+ },
+ {
+ "epoch": 31.531004366812226,
+ "grad_norm": 0.01604730263352394,
+ "learning_rate": 0.0006,
+ "loss": 4.1465911865234375,
+ "step": 2270
+ },
+ {
+ "epoch": 31.544978165938865,
+ "grad_norm": 0.016205785796046257,
+ "learning_rate": 0.0006,
+ "loss": 4.106704235076904,
+ "step": 2271
+ },
+ {
+ "epoch": 31.558951965065503,
+ "grad_norm": 0.017008036375045776,
+ "learning_rate": 0.0006,
+ "loss": 4.191000938415527,
+ "step": 2272
+ },
+ {
+ "epoch": 31.57292576419214,
+ "grad_norm": 0.01699635200202465,
+ "learning_rate": 0.0006,
+ "loss": 4.187042236328125,
+ "step": 2273
+ },
+ {
+ "epoch": 31.586899563318777,
+ "grad_norm": 0.01692967116832733,
+ "learning_rate": 0.0006,
+ "loss": 4.208805561065674,
+ "step": 2274
+ },
+ {
+ "epoch": 31.600873362445416,
+ "grad_norm": 0.017608756199479103,
+ "learning_rate": 0.0006,
+ "loss": 4.200512886047363,
+ "step": 2275
+ },
+ {
+ "epoch": 31.61484716157205,
+ "grad_norm": 0.016590990126132965,
+ "learning_rate": 0.0006,
+ "loss": 4.194705009460449,
+ "step": 2276
+ },
+ {
+ "epoch": 31.62882096069869,
+ "grad_norm": 0.0170235987752676,
+ "learning_rate": 0.0006,
+ "loss": 4.210326194763184,
+ "step": 2277
+ },
+ {
+ "epoch": 31.64279475982533,
+ "grad_norm": 0.01604801043868065,
+ "learning_rate": 0.0006,
+ "loss": 4.196024417877197,
+ "step": 2278
+ },
+ {
+ "epoch": 31.656768558951963,
+ "grad_norm": 0.015694202855229378,
+ "learning_rate": 0.0006,
+ "loss": 4.178236484527588,
+ "step": 2279
+ },
+ {
+ "epoch": 31.670742358078602,
+ "grad_norm": 0.015473959036171436,
+ "learning_rate": 0.0006,
+ "loss": 4.259339809417725,
+ "step": 2280
+ },
+ {
+ "epoch": 31.68471615720524,
+ "grad_norm": 0.014689362607896328,
+ "learning_rate": 0.0006,
+ "loss": 3.9702324867248535,
+ "step": 2281
+ },
+ {
+ "epoch": 31.69868995633188,
+ "grad_norm": 0.015245123766362667,
+ "learning_rate": 0.0006,
+ "loss": 4.197025775909424,
+ "step": 2282
+ },
+ {
+ "epoch": 31.712663755458514,
+ "grad_norm": 0.016148347407579422,
+ "learning_rate": 0.0006,
+ "loss": 4.231856346130371,
+ "step": 2283
+ },
+ {
+ "epoch": 31.726637554585153,
+ "grad_norm": 0.015169818885624409,
+ "learning_rate": 0.0006,
+ "loss": 4.230169296264648,
+ "step": 2284
+ },
+ {
+ "epoch": 31.74061135371179,
+ "grad_norm": 0.014843013137578964,
+ "learning_rate": 0.0006,
+ "loss": 4.195633411407471,
+ "step": 2285
+ },
+ {
+ "epoch": 31.754585152838427,
+ "grad_norm": 0.015229340642690659,
+ "learning_rate": 0.0006,
+ "loss": 4.231280326843262,
+ "step": 2286
+ },
+ {
+ "epoch": 31.768558951965066,
+ "grad_norm": 0.01551708485931158,
+ "learning_rate": 0.0006,
+ "loss": 4.17064094543457,
+ "step": 2287
+ },
+ {
+ "epoch": 31.782532751091704,
+ "grad_norm": 0.015863575041294098,
+ "learning_rate": 0.0006,
+ "loss": 4.1579484939575195,
+ "step": 2288
+ },
+ {
+ "epoch": 31.79650655021834,
+ "grad_norm": 0.015118198469281197,
+ "learning_rate": 0.0006,
+ "loss": 4.1626787185668945,
+ "step": 2289
+ },
+ {
+ "epoch": 31.810480349344978,
+ "grad_norm": 0.015220776200294495,
+ "learning_rate": 0.0006,
+ "loss": 4.135746955871582,
+ "step": 2290
+ },
+ {
+ "epoch": 31.824454148471617,
+ "grad_norm": 0.014529784210026264,
+ "learning_rate": 0.0006,
+ "loss": 4.164931297302246,
+ "step": 2291
+ },
+ {
+ "epoch": 31.83842794759825,
+ "grad_norm": 0.015533296391367912,
+ "learning_rate": 0.0006,
+ "loss": 4.112464904785156,
+ "step": 2292
+ },
+ {
+ "epoch": 31.85240174672489,
+ "grad_norm": 0.016562335193157196,
+ "learning_rate": 0.0006,
+ "loss": 4.164898872375488,
+ "step": 2293
+ },
+ {
+ "epoch": 31.86637554585153,
+ "grad_norm": 0.016258113086223602,
+ "learning_rate": 0.0006,
+ "loss": 4.1979265213012695,
+ "step": 2294
+ },
+ {
+ "epoch": 31.880349344978168,
+ "grad_norm": 0.01558110024780035,
+ "learning_rate": 0.0006,
+ "loss": 4.174675941467285,
+ "step": 2295
+ },
+ {
+ "epoch": 31.894323144104803,
+ "grad_norm": 0.016412515193223953,
+ "learning_rate": 0.0006,
+ "loss": 4.217909336090088,
+ "step": 2296
+ },
+ {
+ "epoch": 31.90829694323144,
+ "grad_norm": 0.015118442475795746,
+ "learning_rate": 0.0006,
+ "loss": 4.137143611907959,
+ "step": 2297
+ },
+ {
+ "epoch": 31.92227074235808,
+ "grad_norm": 0.01506161317229271,
+ "learning_rate": 0.0006,
+ "loss": 4.196122169494629,
+ "step": 2298
+ },
+ {
+ "epoch": 31.936244541484715,
+ "grad_norm": 0.017165271565318108,
+ "learning_rate": 0.0006,
+ "loss": 4.14834451675415,
+ "step": 2299
+ },
+ {
+ "epoch": 31.950218340611354,
+ "grad_norm": 0.01696440577507019,
+ "learning_rate": 0.0006,
+ "loss": 4.218868732452393,
+ "step": 2300
+ },
+ {
+ "epoch": 31.964192139737992,
+ "grad_norm": 0.017349394038319588,
+ "learning_rate": 0.0006,
+ "loss": 4.2754011154174805,
+ "step": 2301
+ },
+ {
+ "epoch": 31.978165938864628,
+ "grad_norm": 0.018387921154499054,
+ "learning_rate": 0.0006,
+ "loss": 4.282623291015625,
+ "step": 2302
+ },
+ {
+ "epoch": 31.992139737991266,
+ "grad_norm": 0.018837926909327507,
+ "learning_rate": 0.0006,
+ "loss": 4.185425281524658,
+ "step": 2303
+ },
+ {
+ "epoch": 32.0,
+ "grad_norm": 0.018998464569449425,
+ "learning_rate": 0.0006,
+ "loss": 4.151291847229004,
+ "step": 2304
+ },
+ {
+ "epoch": 32.0,
+ "eval_loss": 4.522614002227783,
+ "eval_runtime": 57.0106,
+ "eval_samples_per_second": 42.834,
+ "eval_steps_per_second": 1.351,
+ "step": 2304
+ },
+ {
+ "epoch": 32.01397379912664,
+ "grad_norm": 0.019036833196878433,
+ "learning_rate": 0.0006,
+ "loss": 4.183202266693115,
+ "step": 2305
+ },
+ {
+ "epoch": 32.02794759825328,
+ "grad_norm": 0.02207397110760212,
+ "learning_rate": 0.0006,
+ "loss": 4.17875862121582,
+ "step": 2306
+ },
+ {
+ "epoch": 32.041921397379916,
+ "grad_norm": 0.022008279338479042,
+ "learning_rate": 0.0006,
+ "loss": 4.08042049407959,
+ "step": 2307
+ },
+ {
+ "epoch": 32.05589519650655,
+ "grad_norm": 0.022407017648220062,
+ "learning_rate": 0.0006,
+ "loss": 4.112189769744873,
+ "step": 2308
+ },
+ {
+ "epoch": 32.069868995633186,
+ "grad_norm": 0.021599747240543365,
+ "learning_rate": 0.0006,
+ "loss": 4.126455307006836,
+ "step": 2309
+ },
+ {
+ "epoch": 32.083842794759825,
+ "grad_norm": 0.020340966060757637,
+ "learning_rate": 0.0006,
+ "loss": 4.080538272857666,
+ "step": 2310
+ },
+ {
+ "epoch": 32.09781659388646,
+ "grad_norm": 0.020956892520189285,
+ "learning_rate": 0.0006,
+ "loss": 4.171943664550781,
+ "step": 2311
+ },
+ {
+ "epoch": 32.1117903930131,
+ "grad_norm": 0.019966667518019676,
+ "learning_rate": 0.0006,
+ "loss": 4.144007682800293,
+ "step": 2312
+ },
+ {
+ "epoch": 32.12576419213974,
+ "grad_norm": 0.018256952986121178,
+ "learning_rate": 0.0006,
+ "loss": 4.208107948303223,
+ "step": 2313
+ },
+ {
+ "epoch": 32.13973799126637,
+ "grad_norm": 0.017289170995354652,
+ "learning_rate": 0.0006,
+ "loss": 4.182024955749512,
+ "step": 2314
+ },
+ {
+ "epoch": 32.15371179039301,
+ "grad_norm": 0.017656700685620308,
+ "learning_rate": 0.0006,
+ "loss": 4.068373680114746,
+ "step": 2315
+ },
+ {
+ "epoch": 32.16768558951965,
+ "grad_norm": 0.017903361469507217,
+ "learning_rate": 0.0006,
+ "loss": 4.2345428466796875,
+ "step": 2316
+ },
+ {
+ "epoch": 32.18165938864629,
+ "grad_norm": 0.01785109005868435,
+ "learning_rate": 0.0006,
+ "loss": 4.219776153564453,
+ "step": 2317
+ },
+ {
+ "epoch": 32.19563318777293,
+ "grad_norm": 0.01887257769703865,
+ "learning_rate": 0.0006,
+ "loss": 4.139206409454346,
+ "step": 2318
+ },
+ {
+ "epoch": 32.209606986899566,
+ "grad_norm": 0.01924685761332512,
+ "learning_rate": 0.0006,
+ "loss": 4.121052265167236,
+ "step": 2319
+ },
+ {
+ "epoch": 32.223580786026204,
+ "grad_norm": 0.020574072375893593,
+ "learning_rate": 0.0006,
+ "loss": 4.13262414932251,
+ "step": 2320
+ },
+ {
+ "epoch": 32.237554585152836,
+ "grad_norm": 0.023849064484238625,
+ "learning_rate": 0.0006,
+ "loss": 4.239119529724121,
+ "step": 2321
+ },
+ {
+ "epoch": 32.251528384279474,
+ "grad_norm": 0.026239361613988876,
+ "learning_rate": 0.0006,
+ "loss": 4.248993873596191,
+ "step": 2322
+ },
+ {
+ "epoch": 32.26550218340611,
+ "grad_norm": 0.023980243131518364,
+ "learning_rate": 0.0006,
+ "loss": 4.159852504730225,
+ "step": 2323
+ },
+ {
+ "epoch": 32.27947598253275,
+ "grad_norm": 0.020442405715584755,
+ "learning_rate": 0.0006,
+ "loss": 4.133546829223633,
+ "step": 2324
+ },
+ {
+ "epoch": 32.29344978165939,
+ "grad_norm": 0.020661186426877975,
+ "learning_rate": 0.0006,
+ "loss": 4.078243732452393,
+ "step": 2325
+ },
+ {
+ "epoch": 32.30742358078603,
+ "grad_norm": 0.02046387456357479,
+ "learning_rate": 0.0006,
+ "loss": 4.172012805938721,
+ "step": 2326
+ },
+ {
+ "epoch": 32.32139737991266,
+ "grad_norm": 0.02089597098529339,
+ "learning_rate": 0.0006,
+ "loss": 4.0604705810546875,
+ "step": 2327
+ },
+ {
+ "epoch": 32.3353711790393,
+ "grad_norm": 0.01875591278076172,
+ "learning_rate": 0.0006,
+ "loss": 4.125911235809326,
+ "step": 2328
+ },
+ {
+ "epoch": 32.34934497816594,
+ "grad_norm": 0.018606796860694885,
+ "learning_rate": 0.0006,
+ "loss": 4.097362518310547,
+ "step": 2329
+ },
+ {
+ "epoch": 32.36331877729258,
+ "grad_norm": 0.01766626164317131,
+ "learning_rate": 0.0006,
+ "loss": 4.118616580963135,
+ "step": 2330
+ },
+ {
+ "epoch": 32.377292576419215,
+ "grad_norm": 0.017692960798740387,
+ "learning_rate": 0.0006,
+ "loss": 4.1457672119140625,
+ "step": 2331
+ },
+ {
+ "epoch": 32.391266375545854,
+ "grad_norm": 0.017619704827666283,
+ "learning_rate": 0.0006,
+ "loss": 4.175961494445801,
+ "step": 2332
+ },
+ {
+ "epoch": 32.40524017467249,
+ "grad_norm": 0.01784413494169712,
+ "learning_rate": 0.0006,
+ "loss": 4.051291465759277,
+ "step": 2333
+ },
+ {
+ "epoch": 32.419213973799124,
+ "grad_norm": 0.018508171662688255,
+ "learning_rate": 0.0006,
+ "loss": 4.0638556480407715,
+ "step": 2334
+ },
+ {
+ "epoch": 32.43318777292576,
+ "grad_norm": 0.01731436885893345,
+ "learning_rate": 0.0006,
+ "loss": 4.1805419921875,
+ "step": 2335
+ },
+ {
+ "epoch": 32.4471615720524,
+ "grad_norm": 0.015886498615145683,
+ "learning_rate": 0.0006,
+ "loss": 4.151026725769043,
+ "step": 2336
+ },
+ {
+ "epoch": 32.46113537117904,
+ "grad_norm": 0.016583820804953575,
+ "learning_rate": 0.0006,
+ "loss": 4.203604698181152,
+ "step": 2337
+ },
+ {
+ "epoch": 32.47510917030568,
+ "grad_norm": 0.0176097322255373,
+ "learning_rate": 0.0006,
+ "loss": 4.135159015655518,
+ "step": 2338
+ },
+ {
+ "epoch": 32.48908296943232,
+ "grad_norm": 0.016657903790473938,
+ "learning_rate": 0.0006,
+ "loss": 4.126681804656982,
+ "step": 2339
+ },
+ {
+ "epoch": 32.50305676855895,
+ "grad_norm": 0.016866130754351616,
+ "learning_rate": 0.0006,
+ "loss": 4.162341117858887,
+ "step": 2340
+ },
+ {
+ "epoch": 32.51703056768559,
+ "grad_norm": 0.016286680474877357,
+ "learning_rate": 0.0006,
+ "loss": 4.261801242828369,
+ "step": 2341
+ },
+ {
+ "epoch": 32.531004366812226,
+ "grad_norm": 0.01481365505605936,
+ "learning_rate": 0.0006,
+ "loss": 4.110016822814941,
+ "step": 2342
+ },
+ {
+ "epoch": 32.544978165938865,
+ "grad_norm": 0.016605013981461525,
+ "learning_rate": 0.0006,
+ "loss": 4.230024814605713,
+ "step": 2343
+ },
+ {
+ "epoch": 32.5589519650655,
+ "grad_norm": 0.01645175740122795,
+ "learning_rate": 0.0006,
+ "loss": 4.174612045288086,
+ "step": 2344
+ },
+ {
+ "epoch": 32.57292576419214,
+ "grad_norm": 0.014725026674568653,
+ "learning_rate": 0.0006,
+ "loss": 4.066371440887451,
+ "step": 2345
+ },
+ {
+ "epoch": 32.58689956331878,
+ "grad_norm": 0.015003837645053864,
+ "learning_rate": 0.0006,
+ "loss": 4.193343639373779,
+ "step": 2346
+ },
+ {
+ "epoch": 32.60087336244541,
+ "grad_norm": 0.015346302650868893,
+ "learning_rate": 0.0006,
+ "loss": 4.210659980773926,
+ "step": 2347
+ },
+ {
+ "epoch": 32.61484716157205,
+ "grad_norm": 0.015135962516069412,
+ "learning_rate": 0.0006,
+ "loss": 4.204618453979492,
+ "step": 2348
+ },
+ {
+ "epoch": 32.62882096069869,
+ "grad_norm": 0.01601956970989704,
+ "learning_rate": 0.0006,
+ "loss": 4.145564556121826,
+ "step": 2349
+ },
+ {
+ "epoch": 32.64279475982533,
+ "grad_norm": 0.01517851185053587,
+ "learning_rate": 0.0006,
+ "loss": 4.051673889160156,
+ "step": 2350
+ },
+ {
+ "epoch": 32.65676855895197,
+ "grad_norm": 0.014894292689859867,
+ "learning_rate": 0.0006,
+ "loss": 4.184818267822266,
+ "step": 2351
+ },
+ {
+ "epoch": 32.670742358078606,
+ "grad_norm": 0.015720834955573082,
+ "learning_rate": 0.0006,
+ "loss": 4.220036506652832,
+ "step": 2352
+ },
+ {
+ "epoch": 32.68471615720524,
+ "grad_norm": 0.01573433168232441,
+ "learning_rate": 0.0006,
+ "loss": 4.076773166656494,
+ "step": 2353
+ },
+ {
+ "epoch": 32.698689956331876,
+ "grad_norm": 0.015828052535653114,
+ "learning_rate": 0.0006,
+ "loss": 4.116329669952393,
+ "step": 2354
+ },
+ {
+ "epoch": 32.712663755458514,
+ "grad_norm": 0.015462259761989117,
+ "learning_rate": 0.0006,
+ "loss": 4.1155242919921875,
+ "step": 2355
+ },
+ {
+ "epoch": 32.72663755458515,
+ "grad_norm": 0.015763849020004272,
+ "learning_rate": 0.0006,
+ "loss": 4.169687747955322,
+ "step": 2356
+ },
+ {
+ "epoch": 32.74061135371179,
+ "grad_norm": 0.014935145154595375,
+ "learning_rate": 0.0006,
+ "loss": 4.2090911865234375,
+ "step": 2357
+ },
+ {
+ "epoch": 32.75458515283843,
+ "grad_norm": 0.015119184739887714,
+ "learning_rate": 0.0006,
+ "loss": 4.131838798522949,
+ "step": 2358
+ },
+ {
+ "epoch": 32.76855895196506,
+ "grad_norm": 0.014833525754511356,
+ "learning_rate": 0.0006,
+ "loss": 4.17618465423584,
+ "step": 2359
+ },
+ {
+ "epoch": 32.7825327510917,
+ "grad_norm": 0.01588359661400318,
+ "learning_rate": 0.0006,
+ "loss": 4.163723945617676,
+ "step": 2360
+ },
+ {
+ "epoch": 32.79650655021834,
+ "grad_norm": 0.0168188214302063,
+ "learning_rate": 0.0006,
+ "loss": 4.178189277648926,
+ "step": 2361
+ },
+ {
+ "epoch": 32.81048034934498,
+ "grad_norm": 0.01726095750927925,
+ "learning_rate": 0.0006,
+ "loss": 4.144223690032959,
+ "step": 2362
+ },
+ {
+ "epoch": 32.82445414847162,
+ "grad_norm": 0.01701410301029682,
+ "learning_rate": 0.0006,
+ "loss": 4.209896087646484,
+ "step": 2363
+ },
+ {
+ "epoch": 32.838427947598255,
+ "grad_norm": 0.01682264916598797,
+ "learning_rate": 0.0006,
+ "loss": 4.26223087310791,
+ "step": 2364
+ },
+ {
+ "epoch": 32.852401746724894,
+ "grad_norm": 0.015597456134855747,
+ "learning_rate": 0.0006,
+ "loss": 4.292664051055908,
+ "step": 2365
+ },
+ {
+ "epoch": 32.866375545851525,
+ "grad_norm": 0.015258579514920712,
+ "learning_rate": 0.0006,
+ "loss": 4.165006637573242,
+ "step": 2366
+ },
+ {
+ "epoch": 32.880349344978164,
+ "grad_norm": 0.01662200503051281,
+ "learning_rate": 0.0006,
+ "loss": 4.309786319732666,
+ "step": 2367
+ },
+ {
+ "epoch": 32.8943231441048,
+ "grad_norm": 0.015691161155700684,
+ "learning_rate": 0.0006,
+ "loss": 4.167202949523926,
+ "step": 2368
+ },
+ {
+ "epoch": 32.90829694323144,
+ "grad_norm": 0.014870637096464634,
+ "learning_rate": 0.0006,
+ "loss": 4.142415523529053,
+ "step": 2369
+ },
+ {
+ "epoch": 32.92227074235808,
+ "grad_norm": 0.01567252166569233,
+ "learning_rate": 0.0006,
+ "loss": 4.2312092781066895,
+ "step": 2370
+ },
+ {
+ "epoch": 32.93624454148472,
+ "grad_norm": 0.014948616735637188,
+ "learning_rate": 0.0006,
+ "loss": 4.1514129638671875,
+ "step": 2371
+ },
+ {
+ "epoch": 32.95021834061135,
+ "grad_norm": 0.014654500409960747,
+ "learning_rate": 0.0006,
+ "loss": 4.205010890960693,
+ "step": 2372
+ },
+ {
+ "epoch": 32.96419213973799,
+ "grad_norm": 0.015059148892760277,
+ "learning_rate": 0.0006,
+ "loss": 4.148405075073242,
+ "step": 2373
+ },
+ {
+ "epoch": 32.97816593886463,
+ "grad_norm": 0.015026576817035675,
+ "learning_rate": 0.0006,
+ "loss": 4.182535171508789,
+ "step": 2374
+ },
+ {
+ "epoch": 32.992139737991266,
+ "grad_norm": 0.016394320875406265,
+ "learning_rate": 0.0006,
+ "loss": 4.129051208496094,
+ "step": 2375
+ },
+ {
+ "epoch": 33.0,
+ "grad_norm": 0.01867043413221836,
+ "learning_rate": 0.0006,
+ "loss": 4.185556411743164,
+ "step": 2376
+ },
+ {
+ "epoch": 33.0,
+ "eval_loss": 4.640405654907227,
+ "eval_runtime": 56.5861,
+ "eval_samples_per_second": 43.155,
+ "eval_steps_per_second": 1.361,
+ "step": 2376
+ },
+ {
+ "epoch": 33.01397379912664,
+ "grad_norm": 0.018626727163791656,
+ "learning_rate": 0.0006,
+ "loss": 4.143031597137451,
+ "step": 2377
+ },
+ {
+ "epoch": 33.02794759825328,
+ "grad_norm": 0.019283363595604897,
+ "learning_rate": 0.0006,
+ "loss": 4.117845058441162,
+ "step": 2378
+ },
+ {
+ "epoch": 33.041921397379916,
+ "grad_norm": 0.018185822293162346,
+ "learning_rate": 0.0006,
+ "loss": 4.076862812042236,
+ "step": 2379
+ },
+ {
+ "epoch": 33.05589519650655,
+ "grad_norm": 0.019411196932196617,
+ "learning_rate": 0.0006,
+ "loss": 4.124156475067139,
+ "step": 2380
+ },
+ {
+ "epoch": 33.069868995633186,
+ "grad_norm": 0.019248466938734055,
+ "learning_rate": 0.0006,
+ "loss": 4.201735019683838,
+ "step": 2381
+ },
+ {
+ "epoch": 33.083842794759825,
+ "grad_norm": 0.018856720998883247,
+ "learning_rate": 0.0006,
+ "loss": 4.00296688079834,
+ "step": 2382
+ },
+ {
+ "epoch": 33.09781659388646,
+ "grad_norm": 0.018554294481873512,
+ "learning_rate": 0.0006,
+ "loss": 4.270577430725098,
+ "step": 2383
+ },
+ {
+ "epoch": 33.1117903930131,
+ "grad_norm": 0.019946224987506866,
+ "learning_rate": 0.0006,
+ "loss": 4.119589328765869,
+ "step": 2384
+ },
+ {
+ "epoch": 33.12576419213974,
+ "grad_norm": 0.021395670250058174,
+ "learning_rate": 0.0006,
+ "loss": 4.128847122192383,
+ "step": 2385
+ },
+ {
+ "epoch": 33.13973799126637,
+ "grad_norm": 0.02009674906730652,
+ "learning_rate": 0.0006,
+ "loss": 4.183228969573975,
+ "step": 2386
+ },
+ {
+ "epoch": 33.15371179039301,
+ "grad_norm": 0.01820223033428192,
+ "learning_rate": 0.0006,
+ "loss": 4.131601810455322,
+ "step": 2387
+ },
+ {
+ "epoch": 33.16768558951965,
+ "grad_norm": 0.018653156235814095,
+ "learning_rate": 0.0006,
+ "loss": 4.116430282592773,
+ "step": 2388
+ },
+ {
+ "epoch": 33.18165938864629,
+ "grad_norm": 0.016828063875436783,
+ "learning_rate": 0.0006,
+ "loss": 4.054663181304932,
+ "step": 2389
+ },
+ {
+ "epoch": 33.19563318777293,
+ "grad_norm": 0.017025409266352654,
+ "learning_rate": 0.0006,
+ "loss": 4.110044956207275,
+ "step": 2390
+ },
+ {
+ "epoch": 33.209606986899566,
+ "grad_norm": 0.016608452424407005,
+ "learning_rate": 0.0006,
+ "loss": 4.102668285369873,
+ "step": 2391
+ },
+ {
+ "epoch": 33.223580786026204,
+ "grad_norm": 0.01743577979505062,
+ "learning_rate": 0.0006,
+ "loss": 4.166791915893555,
+ "step": 2392
+ },
+ {
+ "epoch": 33.237554585152836,
+ "grad_norm": 0.017253734171390533,
+ "learning_rate": 0.0006,
+ "loss": 4.217832088470459,
+ "step": 2393
+ },
+ {
+ "epoch": 33.251528384279474,
+ "grad_norm": 0.017937595024704933,
+ "learning_rate": 0.0006,
+ "loss": 4.2173967361450195,
+ "step": 2394
+ },
+ {
+ "epoch": 33.26550218340611,
+ "grad_norm": 0.01985993981361389,
+ "learning_rate": 0.0006,
+ "loss": 4.17849588394165,
+ "step": 2395
+ },
+ {
+ "epoch": 33.27947598253275,
+ "grad_norm": 0.019243355840444565,
+ "learning_rate": 0.0006,
+ "loss": 4.086523056030273,
+ "step": 2396
+ },
+ {
+ "epoch": 33.29344978165939,
+ "grad_norm": 0.01924416422843933,
+ "learning_rate": 0.0006,
+ "loss": 4.128574371337891,
+ "step": 2397
+ },
+ {
+ "epoch": 33.30742358078603,
+ "grad_norm": 0.01983758620917797,
+ "learning_rate": 0.0006,
+ "loss": 4.061581611633301,
+ "step": 2398
+ },
+ {
+ "epoch": 33.32139737991266,
+ "grad_norm": 0.01930462196469307,
+ "learning_rate": 0.0006,
+ "loss": 4.091731071472168,
+ "step": 2399
+ },
+ {
+ "epoch": 33.3353711790393,
+ "grad_norm": 0.017988424748182297,
+ "learning_rate": 0.0006,
+ "loss": 4.155599117279053,
+ "step": 2400
+ },
+ {
+ "epoch": 33.34934497816594,
+ "grad_norm": 0.015194767154753208,
+ "learning_rate": 0.0006,
+ "loss": 4.144134044647217,
+ "step": 2401
+ },
+ {
+ "epoch": 33.36331877729258,
+ "grad_norm": 0.01536081824451685,
+ "learning_rate": 0.0006,
+ "loss": 4.127272129058838,
+ "step": 2402
+ },
+ {
+ "epoch": 33.377292576419215,
+ "grad_norm": 0.015431297942996025,
+ "learning_rate": 0.0006,
+ "loss": 4.3213019371032715,
+ "step": 2403
+ },
+ {
+ "epoch": 33.391266375545854,
+ "grad_norm": 0.015740280970931053,
+ "learning_rate": 0.0006,
+ "loss": 4.121467590332031,
+ "step": 2404
+ },
+ {
+ "epoch": 33.40524017467249,
+ "grad_norm": 0.015379384160041809,
+ "learning_rate": 0.0006,
+ "loss": 4.095216751098633,
+ "step": 2405
+ },
+ {
+ "epoch": 33.419213973799124,
+ "grad_norm": 0.01528073474764824,
+ "learning_rate": 0.0006,
+ "loss": 4.13047981262207,
+ "step": 2406
+ },
+ {
+ "epoch": 33.43318777292576,
+ "grad_norm": 0.015739979222416878,
+ "learning_rate": 0.0006,
+ "loss": 4.2533135414123535,
+ "step": 2407
+ },
+ {
+ "epoch": 33.4471615720524,
+ "grad_norm": 0.01519764307886362,
+ "learning_rate": 0.0006,
+ "loss": 4.131782054901123,
+ "step": 2408
+ },
+ {
+ "epoch": 33.46113537117904,
+ "grad_norm": 0.015425996854901314,
+ "learning_rate": 0.0006,
+ "loss": 4.02811336517334,
+ "step": 2409
+ },
+ {
+ "epoch": 33.47510917030568,
+ "grad_norm": 0.015378830023109913,
+ "learning_rate": 0.0006,
+ "loss": 4.08652925491333,
+ "step": 2410
+ },
+ {
+ "epoch": 33.48908296943232,
+ "grad_norm": 0.01473075058311224,
+ "learning_rate": 0.0006,
+ "loss": 4.214364528656006,
+ "step": 2411
+ },
+ {
+ "epoch": 33.50305676855895,
+ "grad_norm": 0.015141637064516544,
+ "learning_rate": 0.0006,
+ "loss": 4.155712127685547,
+ "step": 2412
+ },
+ {
+ "epoch": 33.51703056768559,
+ "grad_norm": 0.01641593500971794,
+ "learning_rate": 0.0006,
+ "loss": 4.097907066345215,
+ "step": 2413
+ },
+ {
+ "epoch": 33.531004366812226,
+ "grad_norm": 0.015182029455900192,
+ "learning_rate": 0.0006,
+ "loss": 4.238499641418457,
+ "step": 2414
+ },
+ {
+ "epoch": 33.544978165938865,
+ "grad_norm": 0.01643732376396656,
+ "learning_rate": 0.0006,
+ "loss": 4.0661492347717285,
+ "step": 2415
+ },
+ {
+ "epoch": 33.5589519650655,
+ "grad_norm": 0.015614228323101997,
+ "learning_rate": 0.0006,
+ "loss": 4.120795249938965,
+ "step": 2416
+ },
+ {
+ "epoch": 33.57292576419214,
+ "grad_norm": 0.016066187992691994,
+ "learning_rate": 0.0006,
+ "loss": 4.096465110778809,
+ "step": 2417
+ },
+ {
+ "epoch": 33.58689956331878,
+ "grad_norm": 0.017214342951774597,
+ "learning_rate": 0.0006,
+ "loss": 4.242628574371338,
+ "step": 2418
+ },
+ {
+ "epoch": 33.60087336244541,
+ "grad_norm": 0.01787586510181427,
+ "learning_rate": 0.0006,
+ "loss": 4.149589538574219,
+ "step": 2419
+ },
+ {
+ "epoch": 33.61484716157205,
+ "grad_norm": 0.019455241039395332,
+ "learning_rate": 0.0006,
+ "loss": 4.267796039581299,
+ "step": 2420
+ },
+ {
+ "epoch": 33.62882096069869,
+ "grad_norm": 0.01984674669802189,
+ "learning_rate": 0.0006,
+ "loss": 4.175220966339111,
+ "step": 2421
+ },
+ {
+ "epoch": 33.64279475982533,
+ "grad_norm": 0.021933717653155327,
+ "learning_rate": 0.0006,
+ "loss": 4.106647491455078,
+ "step": 2422
+ },
+ {
+ "epoch": 33.65676855895197,
+ "grad_norm": 0.02123589813709259,
+ "learning_rate": 0.0006,
+ "loss": 4.204996109008789,
+ "step": 2423
+ },
+ {
+ "epoch": 33.670742358078606,
+ "grad_norm": 0.019814589992165565,
+ "learning_rate": 0.0006,
+ "loss": 4.186171531677246,
+ "step": 2424
+ },
+ {
+ "epoch": 33.68471615720524,
+ "grad_norm": 0.02059728465974331,
+ "learning_rate": 0.0006,
+ "loss": 4.170509338378906,
+ "step": 2425
+ },
+ {
+ "epoch": 33.698689956331876,
+ "grad_norm": 0.018483608961105347,
+ "learning_rate": 0.0006,
+ "loss": 4.085585594177246,
+ "step": 2426
+ },
+ {
+ "epoch": 33.712663755458514,
+ "grad_norm": 0.018403852358460426,
+ "learning_rate": 0.0006,
+ "loss": 4.2551774978637695,
+ "step": 2427
+ },
+ {
+ "epoch": 33.72663755458515,
+ "grad_norm": 0.018653830513358116,
+ "learning_rate": 0.0006,
+ "loss": 4.091766357421875,
+ "step": 2428
+ },
+ {
+ "epoch": 33.74061135371179,
+ "grad_norm": 0.017692334949970245,
+ "learning_rate": 0.0006,
+ "loss": 4.100838661193848,
+ "step": 2429
+ },
+ {
+ "epoch": 33.75458515283843,
+ "grad_norm": 0.016029126942157745,
+ "learning_rate": 0.0006,
+ "loss": 4.1530680656433105,
+ "step": 2430
+ },
+ {
+ "epoch": 33.76855895196506,
+ "grad_norm": 0.016095684841275215,
+ "learning_rate": 0.0006,
+ "loss": 4.20544958114624,
+ "step": 2431
+ },
+ {
+ "epoch": 33.7825327510917,
+ "grad_norm": 0.016375798732042313,
+ "learning_rate": 0.0006,
+ "loss": 4.199177265167236,
+ "step": 2432
+ },
+ {
+ "epoch": 33.79650655021834,
+ "grad_norm": 0.01590786688029766,
+ "learning_rate": 0.0006,
+ "loss": 4.200799465179443,
+ "step": 2433
+ },
+ {
+ "epoch": 33.81048034934498,
+ "grad_norm": 0.014896097593009472,
+ "learning_rate": 0.0006,
+ "loss": 4.185850143432617,
+ "step": 2434
+ },
+ {
+ "epoch": 33.82445414847162,
+ "grad_norm": 0.01438513956964016,
+ "learning_rate": 0.0006,
+ "loss": 4.214962005615234,
+ "step": 2435
+ },
+ {
+ "epoch": 33.838427947598255,
+ "grad_norm": 0.014459339901804924,
+ "learning_rate": 0.0006,
+ "loss": 4.196972370147705,
+ "step": 2436
+ },
+ {
+ "epoch": 33.852401746724894,
+ "grad_norm": 0.014782169833779335,
+ "learning_rate": 0.0006,
+ "loss": 4.0885910987854,
+ "step": 2437
+ },
+ {
+ "epoch": 33.866375545851525,
+ "grad_norm": 0.01520370040088892,
+ "learning_rate": 0.0006,
+ "loss": 4.0323262214660645,
+ "step": 2438
+ },
+ {
+ "epoch": 33.880349344978164,
+ "grad_norm": 0.014056282117962837,
+ "learning_rate": 0.0006,
+ "loss": 4.1386518478393555,
+ "step": 2439
+ },
+ {
+ "epoch": 33.8943231441048,
+ "grad_norm": 0.013567990623414516,
+ "learning_rate": 0.0006,
+ "loss": 4.095754623413086,
+ "step": 2440
+ },
+ {
+ "epoch": 33.90829694323144,
+ "grad_norm": 0.013306716457009315,
+ "learning_rate": 0.0006,
+ "loss": 4.1623077392578125,
+ "step": 2441
+ },
+ {
+ "epoch": 33.92227074235808,
+ "grad_norm": 0.014047150500118732,
+ "learning_rate": 0.0006,
+ "loss": 4.142541885375977,
+ "step": 2442
+ },
+ {
+ "epoch": 33.93624454148472,
+ "grad_norm": 0.01377617847174406,
+ "learning_rate": 0.0006,
+ "loss": 4.268551826477051,
+ "step": 2443
+ },
+ {
+ "epoch": 33.95021834061135,
+ "grad_norm": 0.014671621844172478,
+ "learning_rate": 0.0006,
+ "loss": 4.198196887969971,
+ "step": 2444
+ },
+ {
+ "epoch": 33.96419213973799,
+ "grad_norm": 0.014806744642555714,
+ "learning_rate": 0.0006,
+ "loss": 4.057899475097656,
+ "step": 2445
+ },
+ {
+ "epoch": 33.97816593886463,
+ "grad_norm": 0.015734272077679634,
+ "learning_rate": 0.0006,
+ "loss": 4.163505554199219,
+ "step": 2446
+ },
+ {
+ "epoch": 33.992139737991266,
+ "grad_norm": 0.015250146389007568,
+ "learning_rate": 0.0006,
+ "loss": 4.076294898986816,
+ "step": 2447
+ },
+ {
+ "epoch": 34.0,
+ "grad_norm": 0.016733702272176743,
+ "learning_rate": 0.0006,
+ "loss": 4.138420104980469,
+ "step": 2448
+ },
+ {
+ "epoch": 34.0,
+ "eval_loss": 4.542503356933594,
+ "eval_runtime": 56.983,
+ "eval_samples_per_second": 42.855,
+ "eval_steps_per_second": 1.351,
+ "step": 2448
+ },
+ {
+ "epoch": 34.01397379912664,
+ "grad_norm": 0.015866613015532494,
+ "learning_rate": 0.0006,
+ "loss": 4.032081127166748,
+ "step": 2449
+ },
+ {
+ "epoch": 34.02794759825328,
+ "grad_norm": 0.01946570910513401,
+ "learning_rate": 0.0006,
+ "loss": 4.122320175170898,
+ "step": 2450
+ },
+ {
+ "epoch": 34.041921397379916,
+ "grad_norm": 0.020265938714146614,
+ "learning_rate": 0.0006,
+ "loss": 4.06460428237915,
+ "step": 2451
+ },
+ {
+ "epoch": 34.05589519650655,
+ "grad_norm": 0.018887171521782875,
+ "learning_rate": 0.0006,
+ "loss": 4.096906661987305,
+ "step": 2452
+ },
+ {
+ "epoch": 34.069868995633186,
+ "grad_norm": 0.017304692417383194,
+ "learning_rate": 0.0006,
+ "loss": 4.137261390686035,
+ "step": 2453
+ },
+ {
+ "epoch": 34.083842794759825,
+ "grad_norm": 0.017029082402586937,
+ "learning_rate": 0.0006,
+ "loss": 4.081112861633301,
+ "step": 2454
+ },
+ {
+ "epoch": 34.09781659388646,
+ "grad_norm": 0.016414813697338104,
+ "learning_rate": 0.0006,
+ "loss": 4.140783309936523,
+ "step": 2455
+ },
+ {
+ "epoch": 34.1117903930131,
+ "grad_norm": 0.018238281831145287,
+ "learning_rate": 0.0006,
+ "loss": 4.161116600036621,
+ "step": 2456
+ },
+ {
+ "epoch": 34.12576419213974,
+ "grad_norm": 0.019929232075810432,
+ "learning_rate": 0.0006,
+ "loss": 4.243722438812256,
+ "step": 2457
+ },
+ {
+ "epoch": 34.13973799126637,
+ "grad_norm": 0.019375786185264587,
+ "learning_rate": 0.0006,
+ "loss": 4.134873390197754,
+ "step": 2458
+ },
+ {
+ "epoch": 34.15371179039301,
+ "grad_norm": 0.018151137977838516,
+ "learning_rate": 0.0006,
+ "loss": 4.098745346069336,
+ "step": 2459
+ },
+ {
+ "epoch": 34.16768558951965,
+ "grad_norm": 0.017054934054613113,
+ "learning_rate": 0.0006,
+ "loss": 4.194662570953369,
+ "step": 2460
+ },
+ {
+ "epoch": 34.18165938864629,
+ "grad_norm": 0.018220555037260056,
+ "learning_rate": 0.0006,
+ "loss": 4.128866195678711,
+ "step": 2461
+ },
+ {
+ "epoch": 34.19563318777293,
+ "grad_norm": 0.01692879945039749,
+ "learning_rate": 0.0006,
+ "loss": 4.062477111816406,
+ "step": 2462
+ },
+ {
+ "epoch": 34.209606986899566,
+ "grad_norm": 0.01674199476838112,
+ "learning_rate": 0.0006,
+ "loss": 4.186149597167969,
+ "step": 2463
+ },
+ {
+ "epoch": 34.223580786026204,
+ "grad_norm": 0.016527455300092697,
+ "learning_rate": 0.0006,
+ "loss": 4.0734028816223145,
+ "step": 2464
+ },
+ {
+ "epoch": 34.237554585152836,
+ "grad_norm": 0.016448386013507843,
+ "learning_rate": 0.0006,
+ "loss": 4.098896026611328,
+ "step": 2465
+ },
+ {
+ "epoch": 34.251528384279474,
+ "grad_norm": 0.014644953422248363,
+ "learning_rate": 0.0006,
+ "loss": 4.1031904220581055,
+ "step": 2466
+ },
+ {
+ "epoch": 34.26550218340611,
+ "grad_norm": 0.016649341210722923,
+ "learning_rate": 0.0006,
+ "loss": 4.148791313171387,
+ "step": 2467
+ },
+ {
+ "epoch": 34.27947598253275,
+ "grad_norm": 0.017442714422941208,
+ "learning_rate": 0.0006,
+ "loss": 4.070799827575684,
+ "step": 2468
+ },
+ {
+ "epoch": 34.29344978165939,
+ "grad_norm": 0.017852013930678368,
+ "learning_rate": 0.0006,
+ "loss": 4.031924247741699,
+ "step": 2469
+ },
+ {
+ "epoch": 34.30742358078603,
+ "grad_norm": 0.016371270641684532,
+ "learning_rate": 0.0006,
+ "loss": 4.149420738220215,
+ "step": 2470
+ },
+ {
+ "epoch": 34.32139737991266,
+ "grad_norm": 0.015846814960241318,
+ "learning_rate": 0.0006,
+ "loss": 4.15427303314209,
+ "step": 2471
+ },
+ {
+ "epoch": 34.3353711790393,
+ "grad_norm": 0.017411163076758385,
+ "learning_rate": 0.0006,
+ "loss": 4.1304192543029785,
+ "step": 2472
+ },
+ {
+ "epoch": 34.34934497816594,
+ "grad_norm": 0.01780046708881855,
+ "learning_rate": 0.0006,
+ "loss": 4.191732406616211,
+ "step": 2473
+ },
+ {
+ "epoch": 34.36331877729258,
+ "grad_norm": 0.01958383060991764,
+ "learning_rate": 0.0006,
+ "loss": 4.068355083465576,
+ "step": 2474
+ },
+ {
+ "epoch": 34.377292576419215,
+ "grad_norm": 0.022173317149281502,
+ "learning_rate": 0.0006,
+ "loss": 4.182929039001465,
+ "step": 2475
+ },
+ {
+ "epoch": 34.391266375545854,
+ "grad_norm": 0.022389156743884087,
+ "learning_rate": 0.0006,
+ "loss": 4.110218048095703,
+ "step": 2476
+ },
+ {
+ "epoch": 34.40524017467249,
+ "grad_norm": 0.022478511556982994,
+ "learning_rate": 0.0006,
+ "loss": 4.095479488372803,
+ "step": 2477
+ },
+ {
+ "epoch": 34.419213973799124,
+ "grad_norm": 0.019504647701978683,
+ "learning_rate": 0.0006,
+ "loss": 4.101964950561523,
+ "step": 2478
+ },
+ {
+ "epoch": 34.43318777292576,
+ "grad_norm": 0.019378576427698135,
+ "learning_rate": 0.0006,
+ "loss": 4.14620304107666,
+ "step": 2479
+ },
+ {
+ "epoch": 34.4471615720524,
+ "grad_norm": 0.017993204295635223,
+ "learning_rate": 0.0006,
+ "loss": 4.183492660522461,
+ "step": 2480
+ },
+ {
+ "epoch": 34.46113537117904,
+ "grad_norm": 0.01814805157482624,
+ "learning_rate": 0.0006,
+ "loss": 4.151247978210449,
+ "step": 2481
+ },
+ {
+ "epoch": 34.47510917030568,
+ "grad_norm": 0.018622448667883873,
+ "learning_rate": 0.0006,
+ "loss": 4.020938396453857,
+ "step": 2482
+ },
+ {
+ "epoch": 34.48908296943232,
+ "grad_norm": 0.018360132351517677,
+ "learning_rate": 0.0006,
+ "loss": 4.064734935760498,
+ "step": 2483
+ },
+ {
+ "epoch": 34.50305676855895,
+ "grad_norm": 0.017715711146593094,
+ "learning_rate": 0.0006,
+ "loss": 4.185687065124512,
+ "step": 2484
+ },
+ {
+ "epoch": 34.51703056768559,
+ "grad_norm": 0.01714295893907547,
+ "learning_rate": 0.0006,
+ "loss": 4.1044721603393555,
+ "step": 2485
+ },
+ {
+ "epoch": 34.531004366812226,
+ "grad_norm": 0.016943803057074547,
+ "learning_rate": 0.0006,
+ "loss": 4.106837272644043,
+ "step": 2486
+ },
+ {
+ "epoch": 34.544978165938865,
+ "grad_norm": 0.0182423684746027,
+ "learning_rate": 0.0006,
+ "loss": 4.1041340827941895,
+ "step": 2487
+ },
+ {
+ "epoch": 34.5589519650655,
+ "grad_norm": 0.01636587455868721,
+ "learning_rate": 0.0006,
+ "loss": 3.9735546112060547,
+ "step": 2488
+ },
+ {
+ "epoch": 34.57292576419214,
+ "grad_norm": 0.014593546278774738,
+ "learning_rate": 0.0006,
+ "loss": 4.041980266571045,
+ "step": 2489
+ },
+ {
+ "epoch": 34.58689956331878,
+ "grad_norm": 0.015436965972185135,
+ "learning_rate": 0.0006,
+ "loss": 4.301148414611816,
+ "step": 2490
+ },
+ {
+ "epoch": 34.60087336244541,
+ "grad_norm": 0.015075269155204296,
+ "learning_rate": 0.0006,
+ "loss": 4.118527412414551,
+ "step": 2491
+ },
+ {
+ "epoch": 34.61484716157205,
+ "grad_norm": 0.01701514609158039,
+ "learning_rate": 0.0006,
+ "loss": 4.245968818664551,
+ "step": 2492
+ },
+ {
+ "epoch": 34.62882096069869,
+ "grad_norm": 0.018047887831926346,
+ "learning_rate": 0.0006,
+ "loss": 4.227261066436768,
+ "step": 2493
+ },
+ {
+ "epoch": 34.64279475982533,
+ "grad_norm": 0.018499085679650307,
+ "learning_rate": 0.0006,
+ "loss": 4.121360778808594,
+ "step": 2494
+ },
+ {
+ "epoch": 34.65676855895197,
+ "grad_norm": 0.018987711519002914,
+ "learning_rate": 0.0006,
+ "loss": 4.131564617156982,
+ "step": 2495
+ },
+ {
+ "epoch": 34.670742358078606,
+ "grad_norm": 0.01834476925432682,
+ "learning_rate": 0.0006,
+ "loss": 4.224941253662109,
+ "step": 2496
+ },
+ {
+ "epoch": 34.68471615720524,
+ "grad_norm": 0.017507942393422127,
+ "learning_rate": 0.0006,
+ "loss": 4.1394548416137695,
+ "step": 2497
+ },
+ {
+ "epoch": 34.698689956331876,
+ "grad_norm": 0.01680580899119377,
+ "learning_rate": 0.0006,
+ "loss": 4.1073431968688965,
+ "step": 2498
+ },
+ {
+ "epoch": 34.712663755458514,
+ "grad_norm": 0.015233514830470085,
+ "learning_rate": 0.0006,
+ "loss": 4.136994361877441,
+ "step": 2499
+ },
+ {
+ "epoch": 34.72663755458515,
+ "grad_norm": 0.015482917428016663,
+ "learning_rate": 0.0006,
+ "loss": 4.110719680786133,
+ "step": 2500
+ },
+ {
+ "epoch": 34.74061135371179,
+ "grad_norm": 0.015940461307764053,
+ "learning_rate": 0.0006,
+ "loss": 4.207851886749268,
+ "step": 2501
+ },
+ {
+ "epoch": 34.75458515283843,
+ "grad_norm": 0.015355597250163555,
+ "learning_rate": 0.0006,
+ "loss": 4.169469833374023,
+ "step": 2502
+ },
+ {
+ "epoch": 34.76855895196506,
+ "grad_norm": 0.015974000096321106,
+ "learning_rate": 0.0006,
+ "loss": 4.110061168670654,
+ "step": 2503
+ },
+ {
+ "epoch": 34.7825327510917,
+ "grad_norm": 0.016088927164673805,
+ "learning_rate": 0.0006,
+ "loss": 4.155452728271484,
+ "step": 2504
+ },
+ {
+ "epoch": 34.79650655021834,
+ "grad_norm": 0.016156120225787163,
+ "learning_rate": 0.0006,
+ "loss": 4.178092956542969,
+ "step": 2505
+ },
+ {
+ "epoch": 34.81048034934498,
+ "grad_norm": 0.014928647316992283,
+ "learning_rate": 0.0006,
+ "loss": 4.05457878112793,
+ "step": 2506
+ },
+ {
+ "epoch": 34.82445414847162,
+ "grad_norm": 0.015289328061044216,
+ "learning_rate": 0.0006,
+ "loss": 4.18869686126709,
+ "step": 2507
+ },
+ {
+ "epoch": 34.838427947598255,
+ "grad_norm": 0.015799039974808693,
+ "learning_rate": 0.0006,
+ "loss": 4.106073379516602,
+ "step": 2508
+ },
+ {
+ "epoch": 34.852401746724894,
+ "grad_norm": 0.015824418514966965,
+ "learning_rate": 0.0006,
+ "loss": 4.0634589195251465,
+ "step": 2509
+ },
+ {
+ "epoch": 34.866375545851525,
+ "grad_norm": 0.015321656130254269,
+ "learning_rate": 0.0006,
+ "loss": 4.093090057373047,
+ "step": 2510
+ },
+ {
+ "epoch": 34.880349344978164,
+ "grad_norm": 0.014380166307091713,
+ "learning_rate": 0.0006,
+ "loss": 4.1750640869140625,
+ "step": 2511
+ },
+ {
+ "epoch": 34.8943231441048,
+ "grad_norm": 0.014962567947804928,
+ "learning_rate": 0.0006,
+ "loss": 4.185277938842773,
+ "step": 2512
+ },
+ {
+ "epoch": 34.90829694323144,
+ "grad_norm": 0.014467543922364712,
+ "learning_rate": 0.0006,
+ "loss": 4.1600799560546875,
+ "step": 2513
+ },
+ {
+ "epoch": 34.92227074235808,
+ "grad_norm": 0.01545181218534708,
+ "learning_rate": 0.0006,
+ "loss": 4.096299648284912,
+ "step": 2514
+ },
+ {
+ "epoch": 34.93624454148472,
+ "grad_norm": 0.01605340465903282,
+ "learning_rate": 0.0006,
+ "loss": 4.203323841094971,
+ "step": 2515
+ },
+ {
+ "epoch": 34.95021834061135,
+ "grad_norm": 0.015203815884888172,
+ "learning_rate": 0.0006,
+ "loss": 4.201385021209717,
+ "step": 2516
+ },
+ {
+ "epoch": 34.96419213973799,
+ "grad_norm": 0.015149409882724285,
+ "learning_rate": 0.0006,
+ "loss": 4.088164329528809,
+ "step": 2517
+ },
+ {
+ "epoch": 34.97816593886463,
+ "grad_norm": 0.014885558746755123,
+ "learning_rate": 0.0006,
+ "loss": 4.073911666870117,
+ "step": 2518
+ },
+ {
+ "epoch": 34.992139737991266,
+ "grad_norm": 0.015003419481217861,
+ "learning_rate": 0.0006,
+ "loss": 4.157113552093506,
+ "step": 2519
+ },
+ {
+ "epoch": 35.0,
+ "grad_norm": 0.01641916297376156,
+ "learning_rate": 0.0006,
+ "loss": 3.970052719116211,
+ "step": 2520
+ },
+ {
+ "epoch": 35.0,
+ "eval_loss": 4.641692161560059,
+ "eval_runtime": 56.8836,
+ "eval_samples_per_second": 42.93,
+ "eval_steps_per_second": 1.354,
+ "step": 2520
+ },
+ {
+ "epoch": 35.01397379912664,
+ "grad_norm": 0.015894673764705658,
+ "learning_rate": 0.0006,
+ "loss": 4.104437351226807,
+ "step": 2521
+ },
+ {
+ "epoch": 35.02794759825328,
+ "grad_norm": 0.0173867866396904,
+ "learning_rate": 0.0006,
+ "loss": 4.091726303100586,
+ "step": 2522
+ },
+ {
+ "epoch": 35.041921397379916,
+ "grad_norm": 0.019272832199931145,
+ "learning_rate": 0.0006,
+ "loss": 4.217905044555664,
+ "step": 2523
+ },
+ {
+ "epoch": 35.05589519650655,
+ "grad_norm": 0.021820371970534325,
+ "learning_rate": 0.0006,
+ "loss": 4.111428260803223,
+ "step": 2524
+ },
+ {
+ "epoch": 35.069868995633186,
+ "grad_norm": 0.02340741828083992,
+ "learning_rate": 0.0006,
+ "loss": 4.115771293640137,
+ "step": 2525
+ },
+ {
+ "epoch": 35.083842794759825,
+ "grad_norm": 0.023760691285133362,
+ "learning_rate": 0.0006,
+ "loss": 4.132383346557617,
+ "step": 2526
+ },
+ {
+ "epoch": 35.09781659388646,
+ "grad_norm": 0.02073468267917633,
+ "learning_rate": 0.0006,
+ "loss": 4.272691249847412,
+ "step": 2527
+ },
+ {
+ "epoch": 35.1117903930131,
+ "grad_norm": 0.02197718247771263,
+ "learning_rate": 0.0006,
+ "loss": 4.105586528778076,
+ "step": 2528
+ },
+ {
+ "epoch": 35.12576419213974,
+ "grad_norm": 0.0200254675000906,
+ "learning_rate": 0.0006,
+ "loss": 3.968820095062256,
+ "step": 2529
+ },
+ {
+ "epoch": 35.13973799126637,
+ "grad_norm": 0.01929587684571743,
+ "learning_rate": 0.0006,
+ "loss": 4.045468807220459,
+ "step": 2530
+ },
+ {
+ "epoch": 35.15371179039301,
+ "grad_norm": 0.01886162906885147,
+ "learning_rate": 0.0006,
+ "loss": 4.185571670532227,
+ "step": 2531
+ },
+ {
+ "epoch": 35.16768558951965,
+ "grad_norm": 0.019789839163422585,
+ "learning_rate": 0.0006,
+ "loss": 3.9734551906585693,
+ "step": 2532
+ },
+ {
+ "epoch": 35.18165938864629,
+ "grad_norm": 0.0198142658919096,
+ "learning_rate": 0.0006,
+ "loss": 4.2662506103515625,
+ "step": 2533
+ },
+ {
+ "epoch": 35.19563318777293,
+ "grad_norm": 0.02087012305855751,
+ "learning_rate": 0.0006,
+ "loss": 4.175931930541992,
+ "step": 2534
+ },
+ {
+ "epoch": 35.209606986899566,
+ "grad_norm": 0.020294535905122757,
+ "learning_rate": 0.0006,
+ "loss": 4.060324668884277,
+ "step": 2535
+ },
+ {
+ "epoch": 35.223580786026204,
+ "grad_norm": 0.01908983662724495,
+ "learning_rate": 0.0006,
+ "loss": 4.217939853668213,
+ "step": 2536
+ },
+ {
+ "epoch": 35.237554585152836,
+ "grad_norm": 0.0178892333060503,
+ "learning_rate": 0.0006,
+ "loss": 4.190216541290283,
+ "step": 2537
+ },
+ {
+ "epoch": 35.251528384279474,
+ "grad_norm": 0.017892315983772278,
+ "learning_rate": 0.0006,
+ "loss": 4.067419052124023,
+ "step": 2538
+ },
+ {
+ "epoch": 35.26550218340611,
+ "grad_norm": 0.018556250259280205,
+ "learning_rate": 0.0006,
+ "loss": 4.105958461761475,
+ "step": 2539
+ },
+ {
+ "epoch": 35.27947598253275,
+ "grad_norm": 0.01944422349333763,
+ "learning_rate": 0.0006,
+ "loss": 4.155775547027588,
+ "step": 2540
+ },
+ {
+ "epoch": 35.29344978165939,
+ "grad_norm": 0.017333146184682846,
+ "learning_rate": 0.0006,
+ "loss": 4.244513511657715,
+ "step": 2541
+ },
+ {
+ "epoch": 35.30742358078603,
+ "grad_norm": 0.01639244332909584,
+ "learning_rate": 0.0006,
+ "loss": 4.167815685272217,
+ "step": 2542
+ },
+ {
+ "epoch": 35.32139737991266,
+ "grad_norm": 0.01610143668949604,
+ "learning_rate": 0.0006,
+ "loss": 4.121486663818359,
+ "step": 2543
+ },
+ {
+ "epoch": 35.3353711790393,
+ "grad_norm": 0.01656411588191986,
+ "learning_rate": 0.0006,
+ "loss": 4.1755499839782715,
+ "step": 2544
+ },
+ {
+ "epoch": 35.34934497816594,
+ "grad_norm": 0.015075297094881535,
+ "learning_rate": 0.0006,
+ "loss": 4.102992534637451,
+ "step": 2545
+ },
+ {
+ "epoch": 35.36331877729258,
+ "grad_norm": 0.01538170501589775,
+ "learning_rate": 0.0006,
+ "loss": 4.161603927612305,
+ "step": 2546
+ },
+ {
+ "epoch": 35.377292576419215,
+ "grad_norm": 0.01483312901109457,
+ "learning_rate": 0.0006,
+ "loss": 4.085355758666992,
+ "step": 2547
+ },
+ {
+ "epoch": 35.391266375545854,
+ "grad_norm": 0.014894738793373108,
+ "learning_rate": 0.0006,
+ "loss": 4.083585262298584,
+ "step": 2548
+ },
+ {
+ "epoch": 35.40524017467249,
+ "grad_norm": 0.01475780550390482,
+ "learning_rate": 0.0006,
+ "loss": 4.012645721435547,
+ "step": 2549
+ },
+ {
+ "epoch": 35.419213973799124,
+ "grad_norm": 0.014531653374433517,
+ "learning_rate": 0.0006,
+ "loss": 4.197113990783691,
+ "step": 2550
+ },
+ {
+ "epoch": 35.43318777292576,
+ "grad_norm": 0.015316493809223175,
+ "learning_rate": 0.0006,
+ "loss": 4.227970123291016,
+ "step": 2551
+ },
+ {
+ "epoch": 35.4471615720524,
+ "grad_norm": 0.014685571193695068,
+ "learning_rate": 0.0006,
+ "loss": 4.089415550231934,
+ "step": 2552
+ },
+ {
+ "epoch": 35.46113537117904,
+ "grad_norm": 0.015337035991251469,
+ "learning_rate": 0.0006,
+ "loss": 4.080583572387695,
+ "step": 2553
+ },
+ {
+ "epoch": 35.47510917030568,
+ "grad_norm": 0.016754208132624626,
+ "learning_rate": 0.0006,
+ "loss": 4.163935661315918,
+ "step": 2554
+ },
+ {
+ "epoch": 35.48908296943232,
+ "grad_norm": 0.017535798251628876,
+ "learning_rate": 0.0006,
+ "loss": 4.093027591705322,
+ "step": 2555
+ },
+ {
+ "epoch": 35.50305676855895,
+ "grad_norm": 0.017336975783109665,
+ "learning_rate": 0.0006,
+ "loss": 4.127961158752441,
+ "step": 2556
+ },
+ {
+ "epoch": 35.51703056768559,
+ "grad_norm": 0.016242152079939842,
+ "learning_rate": 0.0006,
+ "loss": 4.155593395233154,
+ "step": 2557
+ },
+ {
+ "epoch": 35.531004366812226,
+ "grad_norm": 0.015121322125196457,
+ "learning_rate": 0.0006,
+ "loss": 3.984004020690918,
+ "step": 2558
+ },
+ {
+ "epoch": 35.544978165938865,
+ "grad_norm": 0.015178884379565716,
+ "learning_rate": 0.0006,
+ "loss": 4.111564636230469,
+ "step": 2559
+ },
+ {
+ "epoch": 35.5589519650655,
+ "grad_norm": 0.014777131378650665,
+ "learning_rate": 0.0006,
+ "loss": 4.041402816772461,
+ "step": 2560
+ },
+ {
+ "epoch": 35.57292576419214,
+ "grad_norm": 0.015197676606476307,
+ "learning_rate": 0.0006,
+ "loss": 4.133706092834473,
+ "step": 2561
+ },
+ {
+ "epoch": 35.58689956331878,
+ "grad_norm": 0.015905972570180893,
+ "learning_rate": 0.0006,
+ "loss": 4.1069135665893555,
+ "step": 2562
+ },
+ {
+ "epoch": 35.60087336244541,
+ "grad_norm": 0.015942877158522606,
+ "learning_rate": 0.0006,
+ "loss": 4.0720624923706055,
+ "step": 2563
+ },
+ {
+ "epoch": 35.61484716157205,
+ "grad_norm": 0.016697583720088005,
+ "learning_rate": 0.0006,
+ "loss": 4.167448043823242,
+ "step": 2564
+ },
+ {
+ "epoch": 35.62882096069869,
+ "grad_norm": 0.015603577718138695,
+ "learning_rate": 0.0006,
+ "loss": 4.127327919006348,
+ "step": 2565
+ },
+ {
+ "epoch": 35.64279475982533,
+ "grad_norm": 0.017153460532426834,
+ "learning_rate": 0.0006,
+ "loss": 4.056543350219727,
+ "step": 2566
+ },
+ {
+ "epoch": 35.65676855895197,
+ "grad_norm": 0.016731958836317062,
+ "learning_rate": 0.0006,
+ "loss": 4.0681843757629395,
+ "step": 2567
+ },
+ {
+ "epoch": 35.670742358078606,
+ "grad_norm": 0.016167359426617622,
+ "learning_rate": 0.0006,
+ "loss": 4.174992084503174,
+ "step": 2568
+ },
+ {
+ "epoch": 35.68471615720524,
+ "grad_norm": 0.01615108922123909,
+ "learning_rate": 0.0006,
+ "loss": 4.1383161544799805,
+ "step": 2569
+ },
+ {
+ "epoch": 35.698689956331876,
+ "grad_norm": 0.015750816091895103,
+ "learning_rate": 0.0006,
+ "loss": 4.031595706939697,
+ "step": 2570
+ },
+ {
+ "epoch": 35.712663755458514,
+ "grad_norm": 0.015416189096868038,
+ "learning_rate": 0.0006,
+ "loss": 4.067498207092285,
+ "step": 2571
+ },
+ {
+ "epoch": 35.72663755458515,
+ "grad_norm": 0.01570792682468891,
+ "learning_rate": 0.0006,
+ "loss": 4.1338300704956055,
+ "step": 2572
+ },
+ {
+ "epoch": 35.74061135371179,
+ "grad_norm": 0.016367752104997635,
+ "learning_rate": 0.0006,
+ "loss": 4.152523040771484,
+ "step": 2573
+ },
+ {
+ "epoch": 35.75458515283843,
+ "grad_norm": 0.016349803656339645,
+ "learning_rate": 0.0006,
+ "loss": 4.101985454559326,
+ "step": 2574
+ },
+ {
+ "epoch": 35.76855895196506,
+ "grad_norm": 0.01663660630583763,
+ "learning_rate": 0.0006,
+ "loss": 4.190130710601807,
+ "step": 2575
+ },
+ {
+ "epoch": 35.7825327510917,
+ "grad_norm": 0.015599170699715614,
+ "learning_rate": 0.0006,
+ "loss": 4.0792436599731445,
+ "step": 2576
+ },
+ {
+ "epoch": 35.79650655021834,
+ "grad_norm": 0.015472427010536194,
+ "learning_rate": 0.0006,
+ "loss": 4.186446666717529,
+ "step": 2577
+ },
+ {
+ "epoch": 35.81048034934498,
+ "grad_norm": 0.015520439483225346,
+ "learning_rate": 0.0006,
+ "loss": 4.246635437011719,
+ "step": 2578
+ },
+ {
+ "epoch": 35.82445414847162,
+ "grad_norm": 0.016998281702399254,
+ "learning_rate": 0.0006,
+ "loss": 4.13435173034668,
+ "step": 2579
+ },
+ {
+ "epoch": 35.838427947598255,
+ "grad_norm": 0.016290361061692238,
+ "learning_rate": 0.0006,
+ "loss": 4.107428073883057,
+ "step": 2580
+ },
+ {
+ "epoch": 35.852401746724894,
+ "grad_norm": 0.0165516659617424,
+ "learning_rate": 0.0006,
+ "loss": 4.1612138748168945,
+ "step": 2581
+ },
+ {
+ "epoch": 35.866375545851525,
+ "grad_norm": 0.01736091636121273,
+ "learning_rate": 0.0006,
+ "loss": 4.1072869300842285,
+ "step": 2582
+ },
+ {
+ "epoch": 35.880349344978164,
+ "grad_norm": 0.01787031814455986,
+ "learning_rate": 0.0006,
+ "loss": 4.1076812744140625,
+ "step": 2583
+ },
+ {
+ "epoch": 35.8943231441048,
+ "grad_norm": 0.015437985770404339,
+ "learning_rate": 0.0006,
+ "loss": 4.097488880157471,
+ "step": 2584
+ },
+ {
+ "epoch": 35.90829694323144,
+ "grad_norm": 0.016306741163134575,
+ "learning_rate": 0.0006,
+ "loss": 4.105017185211182,
+ "step": 2585
+ },
+ {
+ "epoch": 35.92227074235808,
+ "grad_norm": 0.015723245218396187,
+ "learning_rate": 0.0006,
+ "loss": 4.205283164978027,
+ "step": 2586
+ },
+ {
+ "epoch": 35.93624454148472,
+ "grad_norm": 0.014626456424593925,
+ "learning_rate": 0.0006,
+ "loss": 4.086552619934082,
+ "step": 2587
+ },
+ {
+ "epoch": 35.95021834061135,
+ "grad_norm": 0.014348454773426056,
+ "learning_rate": 0.0006,
+ "loss": 4.171329975128174,
+ "step": 2588
+ },
+ {
+ "epoch": 35.96419213973799,
+ "grad_norm": 0.015598480589687824,
+ "learning_rate": 0.0006,
+ "loss": 4.140815258026123,
+ "step": 2589
+ },
+ {
+ "epoch": 35.97816593886463,
+ "grad_norm": 0.017126062884926796,
+ "learning_rate": 0.0006,
+ "loss": 4.157176971435547,
+ "step": 2590
+ },
+ {
+ "epoch": 35.992139737991266,
+ "grad_norm": 0.01860199309885502,
+ "learning_rate": 0.0006,
+ "loss": 4.252880096435547,
+ "step": 2591
+ },
+ {
+ "epoch": 36.0,
+ "grad_norm": 0.021780220791697502,
+ "learning_rate": 0.0006,
+ "loss": 4.249861717224121,
+ "step": 2592
+ },
+ {
+ "epoch": 36.0,
+ "eval_loss": 4.606466293334961,
+ "eval_runtime": 56.6217,
+ "eval_samples_per_second": 43.128,
+ "eval_steps_per_second": 1.36,
+ "step": 2592
+ },
+ {
+ "epoch": 36.01397379912664,
+ "grad_norm": 0.020892828702926636,
+ "learning_rate": 0.0006,
+ "loss": 4.102055549621582,
+ "step": 2593
+ },
+ {
+ "epoch": 36.02794759825328,
+ "grad_norm": 0.020306255668401718,
+ "learning_rate": 0.0006,
+ "loss": 4.184308052062988,
+ "step": 2594
+ },
+ {
+ "epoch": 36.041921397379916,
+ "grad_norm": 0.02025097794830799,
+ "learning_rate": 0.0006,
+ "loss": 3.9691858291625977,
+ "step": 2595
+ },
+ {
+ "epoch": 36.05589519650655,
+ "grad_norm": 0.020254382863640785,
+ "learning_rate": 0.0006,
+ "loss": 4.02957820892334,
+ "step": 2596
+ },
+ {
+ "epoch": 36.069868995633186,
+ "grad_norm": 0.019307196140289307,
+ "learning_rate": 0.0006,
+ "loss": 4.033181667327881,
+ "step": 2597
+ },
+ {
+ "epoch": 36.083842794759825,
+ "grad_norm": 0.0198476854711771,
+ "learning_rate": 0.0006,
+ "loss": 4.0816874504089355,
+ "step": 2598
+ },
+ {
+ "epoch": 36.09781659388646,
+ "grad_norm": 0.019916515797376633,
+ "learning_rate": 0.0006,
+ "loss": 4.164889812469482,
+ "step": 2599
+ },
+ {
+ "epoch": 36.1117903930131,
+ "grad_norm": 0.0189144778996706,
+ "learning_rate": 0.0006,
+ "loss": 4.038427352905273,
+ "step": 2600
+ },
+ {
+ "epoch": 36.12576419213974,
+ "grad_norm": 0.01754768192768097,
+ "learning_rate": 0.0006,
+ "loss": 4.010427474975586,
+ "step": 2601
+ },
+ {
+ "epoch": 36.13973799126637,
+ "grad_norm": 0.016862990334630013,
+ "learning_rate": 0.0006,
+ "loss": 4.035970211029053,
+ "step": 2602
+ },
+ {
+ "epoch": 36.15371179039301,
+ "grad_norm": 0.019616752862930298,
+ "learning_rate": 0.0006,
+ "loss": 4.079898834228516,
+ "step": 2603
+ },
+ {
+ "epoch": 36.16768558951965,
+ "grad_norm": 0.020131465047597885,
+ "learning_rate": 0.0006,
+ "loss": 4.144684791564941,
+ "step": 2604
+ },
+ {
+ "epoch": 36.18165938864629,
+ "grad_norm": 0.02072521299123764,
+ "learning_rate": 0.0006,
+ "loss": 4.013660430908203,
+ "step": 2605
+ },
+ {
+ "epoch": 36.19563318777293,
+ "grad_norm": 0.01924615167081356,
+ "learning_rate": 0.0006,
+ "loss": 4.071398735046387,
+ "step": 2606
+ },
+ {
+ "epoch": 36.209606986899566,
+ "grad_norm": 0.02013954520225525,
+ "learning_rate": 0.0006,
+ "loss": 4.127094268798828,
+ "step": 2607
+ },
+ {
+ "epoch": 36.223580786026204,
+ "grad_norm": 0.02042008936405182,
+ "learning_rate": 0.0006,
+ "loss": 4.154261589050293,
+ "step": 2608
+ },
+ {
+ "epoch": 36.237554585152836,
+ "grad_norm": 0.021998876705765724,
+ "learning_rate": 0.0006,
+ "loss": 4.204150199890137,
+ "step": 2609
+ },
+ {
+ "epoch": 36.251528384279474,
+ "grad_norm": 0.022946683689951897,
+ "learning_rate": 0.0006,
+ "loss": 4.013632774353027,
+ "step": 2610
+ },
+ {
+ "epoch": 36.26550218340611,
+ "grad_norm": 0.02061830274760723,
+ "learning_rate": 0.0006,
+ "loss": 4.171084403991699,
+ "step": 2611
+ },
+ {
+ "epoch": 36.27947598253275,
+ "grad_norm": 0.01883152686059475,
+ "learning_rate": 0.0006,
+ "loss": 4.149848461151123,
+ "step": 2612
+ },
+ {
+ "epoch": 36.29344978165939,
+ "grad_norm": 0.019945282489061356,
+ "learning_rate": 0.0006,
+ "loss": 4.174656867980957,
+ "step": 2613
+ },
+ {
+ "epoch": 36.30742358078603,
+ "grad_norm": 0.020751051604747772,
+ "learning_rate": 0.0006,
+ "loss": 4.080747127532959,
+ "step": 2614
+ },
+ {
+ "epoch": 36.32139737991266,
+ "grad_norm": 0.01879691146314144,
+ "learning_rate": 0.0006,
+ "loss": 4.057356357574463,
+ "step": 2615
+ },
+ {
+ "epoch": 36.3353711790393,
+ "grad_norm": 0.019037071615457535,
+ "learning_rate": 0.0006,
+ "loss": 4.204607009887695,
+ "step": 2616
+ },
+ {
+ "epoch": 36.34934497816594,
+ "grad_norm": 0.01872078701853752,
+ "learning_rate": 0.0006,
+ "loss": 4.050586700439453,
+ "step": 2617
+ },
+ {
+ "epoch": 36.36331877729258,
+ "grad_norm": 0.018092811107635498,
+ "learning_rate": 0.0006,
+ "loss": 4.102832794189453,
+ "step": 2618
+ },
+ {
+ "epoch": 36.377292576419215,
+ "grad_norm": 0.017010487616062164,
+ "learning_rate": 0.0006,
+ "loss": 4.191788673400879,
+ "step": 2619
+ },
+ {
+ "epoch": 36.391266375545854,
+ "grad_norm": 0.01691143773496151,
+ "learning_rate": 0.0006,
+ "loss": 4.160456657409668,
+ "step": 2620
+ },
+ {
+ "epoch": 36.40524017467249,
+ "grad_norm": 0.017046816647052765,
+ "learning_rate": 0.0006,
+ "loss": 4.073828220367432,
+ "step": 2621
+ },
+ {
+ "epoch": 36.419213973799124,
+ "grad_norm": 0.01707981713116169,
+ "learning_rate": 0.0006,
+ "loss": 4.0980095863342285,
+ "step": 2622
+ },
+ {
+ "epoch": 36.43318777292576,
+ "grad_norm": 0.01630452275276184,
+ "learning_rate": 0.0006,
+ "loss": 4.101321220397949,
+ "step": 2623
+ },
+ {
+ "epoch": 36.4471615720524,
+ "grad_norm": 0.01576915942132473,
+ "learning_rate": 0.0006,
+ "loss": 4.158249855041504,
+ "step": 2624
+ },
+ {
+ "epoch": 36.46113537117904,
+ "grad_norm": 0.016179712489247322,
+ "learning_rate": 0.0006,
+ "loss": 4.077852249145508,
+ "step": 2625
+ },
+ {
+ "epoch": 36.47510917030568,
+ "grad_norm": 0.01609169878065586,
+ "learning_rate": 0.0006,
+ "loss": 4.187317848205566,
+ "step": 2626
+ },
+ {
+ "epoch": 36.48908296943232,
+ "grad_norm": 0.017148254439234734,
+ "learning_rate": 0.0006,
+ "loss": 4.109071731567383,
+ "step": 2627
+ },
+ {
+ "epoch": 36.50305676855895,
+ "grad_norm": 0.015842624008655548,
+ "learning_rate": 0.0006,
+ "loss": 4.145096302032471,
+ "step": 2628
+ },
+ {
+ "epoch": 36.51703056768559,
+ "grad_norm": 0.014906194061040878,
+ "learning_rate": 0.0006,
+ "loss": 4.028440475463867,
+ "step": 2629
+ },
+ {
+ "epoch": 36.531004366812226,
+ "grad_norm": 0.01485038734972477,
+ "learning_rate": 0.0006,
+ "loss": 4.175201416015625,
+ "step": 2630
+ },
+ {
+ "epoch": 36.544978165938865,
+ "grad_norm": 0.015834150835871696,
+ "learning_rate": 0.0006,
+ "loss": 4.076434135437012,
+ "step": 2631
+ },
+ {
+ "epoch": 36.5589519650655,
+ "grad_norm": 0.015411358326673508,
+ "learning_rate": 0.0006,
+ "loss": 4.03912878036499,
+ "step": 2632
+ },
+ {
+ "epoch": 36.57292576419214,
+ "grad_norm": 0.014762846753001213,
+ "learning_rate": 0.0006,
+ "loss": 4.019827365875244,
+ "step": 2633
+ },
+ {
+ "epoch": 36.58689956331878,
+ "grad_norm": 0.015269497409462929,
+ "learning_rate": 0.0006,
+ "loss": 4.116957187652588,
+ "step": 2634
+ },
+ {
+ "epoch": 36.60087336244541,
+ "grad_norm": 0.015171617269515991,
+ "learning_rate": 0.0006,
+ "loss": 4.2069926261901855,
+ "step": 2635
+ },
+ {
+ "epoch": 36.61484716157205,
+ "grad_norm": 0.014770902693271637,
+ "learning_rate": 0.0006,
+ "loss": 4.092021942138672,
+ "step": 2636
+ },
+ {
+ "epoch": 36.62882096069869,
+ "grad_norm": 0.014957531355321407,
+ "learning_rate": 0.0006,
+ "loss": 4.16819953918457,
+ "step": 2637
+ },
+ {
+ "epoch": 36.64279475982533,
+ "grad_norm": 0.014774451963603497,
+ "learning_rate": 0.0006,
+ "loss": 4.046073913574219,
+ "step": 2638
+ },
+ {
+ "epoch": 36.65676855895197,
+ "grad_norm": 0.014320445246994495,
+ "learning_rate": 0.0006,
+ "loss": 3.977792501449585,
+ "step": 2639
+ },
+ {
+ "epoch": 36.670742358078606,
+ "grad_norm": 0.01413130946457386,
+ "learning_rate": 0.0006,
+ "loss": 4.130649566650391,
+ "step": 2640
+ },
+ {
+ "epoch": 36.68471615720524,
+ "grad_norm": 0.014336307533085346,
+ "learning_rate": 0.0006,
+ "loss": 4.161357402801514,
+ "step": 2641
+ },
+ {
+ "epoch": 36.698689956331876,
+ "grad_norm": 0.014516931027173996,
+ "learning_rate": 0.0006,
+ "loss": 4.14248514175415,
+ "step": 2642
+ },
+ {
+ "epoch": 36.712663755458514,
+ "grad_norm": 0.014377126470208168,
+ "learning_rate": 0.0006,
+ "loss": 4.062711715698242,
+ "step": 2643
+ },
+ {
+ "epoch": 36.72663755458515,
+ "grad_norm": 0.014332668855786324,
+ "learning_rate": 0.0006,
+ "loss": 4.06423807144165,
+ "step": 2644
+ },
+ {
+ "epoch": 36.74061135371179,
+ "grad_norm": 0.014030571095645428,
+ "learning_rate": 0.0006,
+ "loss": 4.178613662719727,
+ "step": 2645
+ },
+ {
+ "epoch": 36.75458515283843,
+ "grad_norm": 0.014376776292920113,
+ "learning_rate": 0.0006,
+ "loss": 4.073130130767822,
+ "step": 2646
+ },
+ {
+ "epoch": 36.76855895196506,
+ "grad_norm": 0.013827716931700706,
+ "learning_rate": 0.0006,
+ "loss": 4.043869972229004,
+ "step": 2647
+ },
+ {
+ "epoch": 36.7825327510917,
+ "grad_norm": 0.015548762865364552,
+ "learning_rate": 0.0006,
+ "loss": 4.134049415588379,
+ "step": 2648
+ },
+ {
+ "epoch": 36.79650655021834,
+ "grad_norm": 0.017041970044374466,
+ "learning_rate": 0.0006,
+ "loss": 4.205024242401123,
+ "step": 2649
+ },
+ {
+ "epoch": 36.81048034934498,
+ "grad_norm": 0.017638275399804115,
+ "learning_rate": 0.0006,
+ "loss": 4.199924468994141,
+ "step": 2650
+ },
+ {
+ "epoch": 36.82445414847162,
+ "grad_norm": 0.017771627753973007,
+ "learning_rate": 0.0006,
+ "loss": 4.15706729888916,
+ "step": 2651
+ },
+ {
+ "epoch": 36.838427947598255,
+ "grad_norm": 0.015933562070131302,
+ "learning_rate": 0.0006,
+ "loss": 4.220742225646973,
+ "step": 2652
+ },
+ {
+ "epoch": 36.852401746724894,
+ "grad_norm": 0.01668144389986992,
+ "learning_rate": 0.0006,
+ "loss": 4.161426544189453,
+ "step": 2653
+ },
+ {
+ "epoch": 36.866375545851525,
+ "grad_norm": 0.018014680594205856,
+ "learning_rate": 0.0006,
+ "loss": 4.183046817779541,
+ "step": 2654
+ },
+ {
+ "epoch": 36.880349344978164,
+ "grad_norm": 0.018808670341968536,
+ "learning_rate": 0.0006,
+ "loss": 4.105293273925781,
+ "step": 2655
+ },
+ {
+ "epoch": 36.8943231441048,
+ "grad_norm": 0.018003476783633232,
+ "learning_rate": 0.0006,
+ "loss": 4.103171348571777,
+ "step": 2656
+ },
+ {
+ "epoch": 36.90829694323144,
+ "grad_norm": 0.016865914687514305,
+ "learning_rate": 0.0006,
+ "loss": 4.125634670257568,
+ "step": 2657
+ },
+ {
+ "epoch": 36.92227074235808,
+ "grad_norm": 0.018371153622865677,
+ "learning_rate": 0.0006,
+ "loss": 4.192461013793945,
+ "step": 2658
+ },
+ {
+ "epoch": 36.93624454148472,
+ "grad_norm": 0.019045377150177956,
+ "learning_rate": 0.0006,
+ "loss": 4.198105812072754,
+ "step": 2659
+ },
+ {
+ "epoch": 36.95021834061135,
+ "grad_norm": 0.01594168320298195,
+ "learning_rate": 0.0006,
+ "loss": 4.169788360595703,
+ "step": 2660
+ },
+ {
+ "epoch": 36.96419213973799,
+ "grad_norm": 0.015645820647478104,
+ "learning_rate": 0.0006,
+ "loss": 4.047277450561523,
+ "step": 2661
+ },
+ {
+ "epoch": 36.97816593886463,
+ "grad_norm": 0.016313156113028526,
+ "learning_rate": 0.0006,
+ "loss": 4.087392807006836,
+ "step": 2662
+ },
+ {
+ "epoch": 36.992139737991266,
+ "grad_norm": 0.017050625756382942,
+ "learning_rate": 0.0006,
+ "loss": 4.0908002853393555,
+ "step": 2663
+ },
+ {
+ "epoch": 37.0,
+ "grad_norm": 0.017730968073010445,
+ "learning_rate": 0.0006,
+ "loss": 4.110957622528076,
+ "step": 2664
+ },
+ {
+ "epoch": 37.0,
+ "eval_loss": 4.563793182373047,
+ "eval_runtime": 56.5922,
+ "eval_samples_per_second": 43.151,
+ "eval_steps_per_second": 1.361,
+ "step": 2664
+ },
+ {
+ "epoch": 37.01397379912664,
+ "grad_norm": 0.015687910839915276,
+ "learning_rate": 0.0006,
+ "loss": 4.143589019775391,
+ "step": 2665
+ },
+ {
+ "epoch": 37.02794759825328,
+ "grad_norm": 0.017451366409659386,
+ "learning_rate": 0.0006,
+ "loss": 4.0326666831970215,
+ "step": 2666
+ },
+ {
+ "epoch": 37.041921397379916,
+ "grad_norm": 0.01909538544714451,
+ "learning_rate": 0.0006,
+ "loss": 3.9706249237060547,
+ "step": 2667
+ },
+ {
+ "epoch": 37.05589519650655,
+ "grad_norm": 0.01988379657268524,
+ "learning_rate": 0.0006,
+ "loss": 3.9531009197235107,
+ "step": 2668
+ },
+ {
+ "epoch": 37.069868995633186,
+ "grad_norm": 0.02035115472972393,
+ "learning_rate": 0.0006,
+ "loss": 3.962214469909668,
+ "step": 2669
+ },
+ {
+ "epoch": 37.083842794759825,
+ "grad_norm": 0.01952001079916954,
+ "learning_rate": 0.0006,
+ "loss": 4.182262897491455,
+ "step": 2670
+ },
+ {
+ "epoch": 37.09781659388646,
+ "grad_norm": 0.01854325644671917,
+ "learning_rate": 0.0006,
+ "loss": 4.127663612365723,
+ "step": 2671
+ },
+ {
+ "epoch": 37.1117903930131,
+ "grad_norm": 0.017757616937160492,
+ "learning_rate": 0.0006,
+ "loss": 4.100804805755615,
+ "step": 2672
+ },
+ {
+ "epoch": 37.12576419213974,
+ "grad_norm": 0.016712479293346405,
+ "learning_rate": 0.0006,
+ "loss": 4.026284217834473,
+ "step": 2673
+ },
+ {
+ "epoch": 37.13973799126637,
+ "grad_norm": 0.0168139711022377,
+ "learning_rate": 0.0006,
+ "loss": 4.183882713317871,
+ "step": 2674
+ },
+ {
+ "epoch": 37.15371179039301,
+ "grad_norm": 0.017060423269867897,
+ "learning_rate": 0.0006,
+ "loss": 4.131832599639893,
+ "step": 2675
+ },
+ {
+ "epoch": 37.16768558951965,
+ "grad_norm": 0.01726631447672844,
+ "learning_rate": 0.0006,
+ "loss": 3.9156057834625244,
+ "step": 2676
+ },
+ {
+ "epoch": 37.18165938864629,
+ "grad_norm": 0.016348622739315033,
+ "learning_rate": 0.0006,
+ "loss": 4.1731367111206055,
+ "step": 2677
+ },
+ {
+ "epoch": 37.19563318777293,
+ "grad_norm": 0.016640059649944305,
+ "learning_rate": 0.0006,
+ "loss": 4.10997200012207,
+ "step": 2678
+ },
+ {
+ "epoch": 37.209606986899566,
+ "grad_norm": 0.01601342298090458,
+ "learning_rate": 0.0006,
+ "loss": 4.081145286560059,
+ "step": 2679
+ },
+ {
+ "epoch": 37.223580786026204,
+ "grad_norm": 0.01735631190240383,
+ "learning_rate": 0.0006,
+ "loss": 4.202108860015869,
+ "step": 2680
+ },
+ {
+ "epoch": 37.237554585152836,
+ "grad_norm": 0.019164685159921646,
+ "learning_rate": 0.0006,
+ "loss": 4.104109764099121,
+ "step": 2681
+ },
+ {
+ "epoch": 37.251528384279474,
+ "grad_norm": 0.01881759613752365,
+ "learning_rate": 0.0006,
+ "loss": 4.145532131195068,
+ "step": 2682
+ },
+ {
+ "epoch": 37.26550218340611,
+ "grad_norm": 0.017656555399298668,
+ "learning_rate": 0.0006,
+ "loss": 4.012782096862793,
+ "step": 2683
+ },
+ {
+ "epoch": 37.27947598253275,
+ "grad_norm": 0.01624823734164238,
+ "learning_rate": 0.0006,
+ "loss": 4.001727104187012,
+ "step": 2684
+ },
+ {
+ "epoch": 37.29344978165939,
+ "grad_norm": 0.016973139718174934,
+ "learning_rate": 0.0006,
+ "loss": 4.095399856567383,
+ "step": 2685
+ },
+ {
+ "epoch": 37.30742358078603,
+ "grad_norm": 0.015503671951591969,
+ "learning_rate": 0.0006,
+ "loss": 4.034431457519531,
+ "step": 2686
+ },
+ {
+ "epoch": 37.32139737991266,
+ "grad_norm": 0.015370250679552555,
+ "learning_rate": 0.0006,
+ "loss": 4.075836181640625,
+ "step": 2687
+ },
+ {
+ "epoch": 37.3353711790393,
+ "grad_norm": 0.01564021222293377,
+ "learning_rate": 0.0006,
+ "loss": 4.125202178955078,
+ "step": 2688
+ },
+ {
+ "epoch": 37.34934497816594,
+ "grad_norm": 0.016717975959181786,
+ "learning_rate": 0.0006,
+ "loss": 4.153017044067383,
+ "step": 2689
+ },
+ {
+ "epoch": 37.36331877729258,
+ "grad_norm": 0.015263617038726807,
+ "learning_rate": 0.0006,
+ "loss": 4.021814823150635,
+ "step": 2690
+ },
+ {
+ "epoch": 37.377292576419215,
+ "grad_norm": 0.016497142612934113,
+ "learning_rate": 0.0006,
+ "loss": 4.069095611572266,
+ "step": 2691
+ },
+ {
+ "epoch": 37.391266375545854,
+ "grad_norm": 0.01669355481863022,
+ "learning_rate": 0.0006,
+ "loss": 4.174587726593018,
+ "step": 2692
+ },
+ {
+ "epoch": 37.40524017467249,
+ "grad_norm": 0.019695773720741272,
+ "learning_rate": 0.0006,
+ "loss": 4.0073041915893555,
+ "step": 2693
+ },
+ {
+ "epoch": 37.419213973799124,
+ "grad_norm": 0.023720644414424896,
+ "learning_rate": 0.0006,
+ "loss": 4.111948013305664,
+ "step": 2694
+ },
+ {
+ "epoch": 37.43318777292576,
+ "grad_norm": 0.02399350143969059,
+ "learning_rate": 0.0006,
+ "loss": 4.081164360046387,
+ "step": 2695
+ },
+ {
+ "epoch": 37.4471615720524,
+ "grad_norm": 0.019630271941423416,
+ "learning_rate": 0.0006,
+ "loss": 4.065731048583984,
+ "step": 2696
+ },
+ {
+ "epoch": 37.46113537117904,
+ "grad_norm": 0.019679008051753044,
+ "learning_rate": 0.0006,
+ "loss": 3.993852376937866,
+ "step": 2697
+ },
+ {
+ "epoch": 37.47510917030568,
+ "grad_norm": 0.020597299560904503,
+ "learning_rate": 0.0006,
+ "loss": 4.275555610656738,
+ "step": 2698
+ },
+ {
+ "epoch": 37.48908296943232,
+ "grad_norm": 0.01983889564871788,
+ "learning_rate": 0.0006,
+ "loss": 4.254708290100098,
+ "step": 2699
+ },
+ {
+ "epoch": 37.50305676855895,
+ "grad_norm": 0.019150668755173683,
+ "learning_rate": 0.0006,
+ "loss": 4.097885608673096,
+ "step": 2700
+ },
+ {
+ "epoch": 37.51703056768559,
+ "grad_norm": 0.018847715109586716,
+ "learning_rate": 0.0006,
+ "loss": 4.054081439971924,
+ "step": 2701
+ },
+ {
+ "epoch": 37.531004366812226,
+ "grad_norm": 0.019607288762927055,
+ "learning_rate": 0.0006,
+ "loss": 3.9891440868377686,
+ "step": 2702
+ },
+ {
+ "epoch": 37.544978165938865,
+ "grad_norm": 0.018863150849938393,
+ "learning_rate": 0.0006,
+ "loss": 4.084877014160156,
+ "step": 2703
+ },
+ {
+ "epoch": 37.5589519650655,
+ "grad_norm": 0.017340239137411118,
+ "learning_rate": 0.0006,
+ "loss": 4.113290786743164,
+ "step": 2704
+ },
+ {
+ "epoch": 37.57292576419214,
+ "grad_norm": 0.015628693625330925,
+ "learning_rate": 0.0006,
+ "loss": 4.1483001708984375,
+ "step": 2705
+ },
+ {
+ "epoch": 37.58689956331878,
+ "grad_norm": 0.016493283212184906,
+ "learning_rate": 0.0006,
+ "loss": 4.1040143966674805,
+ "step": 2706
+ },
+ {
+ "epoch": 37.60087336244541,
+ "grad_norm": 0.016732443124055862,
+ "learning_rate": 0.0006,
+ "loss": 4.07205057144165,
+ "step": 2707
+ },
+ {
+ "epoch": 37.61484716157205,
+ "grad_norm": 0.017176156863570213,
+ "learning_rate": 0.0006,
+ "loss": 4.1279754638671875,
+ "step": 2708
+ },
+ {
+ "epoch": 37.62882096069869,
+ "grad_norm": 0.015730608254671097,
+ "learning_rate": 0.0006,
+ "loss": 4.086658000946045,
+ "step": 2709
+ },
+ {
+ "epoch": 37.64279475982533,
+ "grad_norm": 0.016111090779304504,
+ "learning_rate": 0.0006,
+ "loss": 4.064335823059082,
+ "step": 2710
+ },
+ {
+ "epoch": 37.65676855895197,
+ "grad_norm": 0.01769379712641239,
+ "learning_rate": 0.0006,
+ "loss": 4.116047382354736,
+ "step": 2711
+ },
+ {
+ "epoch": 37.670742358078606,
+ "grad_norm": 0.01683102548122406,
+ "learning_rate": 0.0006,
+ "loss": 4.095986843109131,
+ "step": 2712
+ },
+ {
+ "epoch": 37.68471615720524,
+ "grad_norm": 0.015726450830698013,
+ "learning_rate": 0.0006,
+ "loss": 4.10775089263916,
+ "step": 2713
+ },
+ {
+ "epoch": 37.698689956331876,
+ "grad_norm": 0.014760235324501991,
+ "learning_rate": 0.0006,
+ "loss": 4.1234893798828125,
+ "step": 2714
+ },
+ {
+ "epoch": 37.712663755458514,
+ "grad_norm": 0.014933438040316105,
+ "learning_rate": 0.0006,
+ "loss": 4.079200744628906,
+ "step": 2715
+ },
+ {
+ "epoch": 37.72663755458515,
+ "grad_norm": 0.015723997727036476,
+ "learning_rate": 0.0006,
+ "loss": 4.07865047454834,
+ "step": 2716
+ },
+ {
+ "epoch": 37.74061135371179,
+ "grad_norm": 0.014539001509547234,
+ "learning_rate": 0.0006,
+ "loss": 4.036242485046387,
+ "step": 2717
+ },
+ {
+ "epoch": 37.75458515283843,
+ "grad_norm": 0.014844979159533978,
+ "learning_rate": 0.0006,
+ "loss": 4.1051483154296875,
+ "step": 2718
+ },
+ {
+ "epoch": 37.76855895196506,
+ "grad_norm": 0.015939844772219658,
+ "learning_rate": 0.0006,
+ "loss": 4.017253875732422,
+ "step": 2719
+ },
+ {
+ "epoch": 37.7825327510917,
+ "grad_norm": 0.014771555550396442,
+ "learning_rate": 0.0006,
+ "loss": 4.0694661140441895,
+ "step": 2720
+ },
+ {
+ "epoch": 37.79650655021834,
+ "grad_norm": 0.015777921304106712,
+ "learning_rate": 0.0006,
+ "loss": 3.947587490081787,
+ "step": 2721
+ },
+ {
+ "epoch": 37.81048034934498,
+ "grad_norm": 0.017451291903853416,
+ "learning_rate": 0.0006,
+ "loss": 4.093625068664551,
+ "step": 2722
+ },
+ {
+ "epoch": 37.82445414847162,
+ "grad_norm": 0.016829246655106544,
+ "learning_rate": 0.0006,
+ "loss": 4.09030818939209,
+ "step": 2723
+ },
+ {
+ "epoch": 37.838427947598255,
+ "grad_norm": 0.016984397545456886,
+ "learning_rate": 0.0006,
+ "loss": 4.140626907348633,
+ "step": 2724
+ },
+ {
+ "epoch": 37.852401746724894,
+ "grad_norm": 0.01732603646814823,
+ "learning_rate": 0.0006,
+ "loss": 4.128276824951172,
+ "step": 2725
+ },
+ {
+ "epoch": 37.866375545851525,
+ "grad_norm": 0.01744619384407997,
+ "learning_rate": 0.0006,
+ "loss": 4.103130340576172,
+ "step": 2726
+ },
+ {
+ "epoch": 37.880349344978164,
+ "grad_norm": 0.016331320628523827,
+ "learning_rate": 0.0006,
+ "loss": 4.037951469421387,
+ "step": 2727
+ },
+ {
+ "epoch": 37.8943231441048,
+ "grad_norm": 0.015919430181384087,
+ "learning_rate": 0.0006,
+ "loss": 4.121513366699219,
+ "step": 2728
+ },
+ {
+ "epoch": 37.90829694323144,
+ "grad_norm": 0.016153695061802864,
+ "learning_rate": 0.0006,
+ "loss": 4.099384307861328,
+ "step": 2729
+ },
+ {
+ "epoch": 37.92227074235808,
+ "grad_norm": 0.01628800667822361,
+ "learning_rate": 0.0006,
+ "loss": 4.092325687408447,
+ "step": 2730
+ },
+ {
+ "epoch": 37.93624454148472,
+ "grad_norm": 0.017285684123635292,
+ "learning_rate": 0.0006,
+ "loss": 3.9828906059265137,
+ "step": 2731
+ },
+ {
+ "epoch": 37.95021834061135,
+ "grad_norm": 0.016047554090619087,
+ "learning_rate": 0.0006,
+ "loss": 4.100419044494629,
+ "step": 2732
+ },
+ {
+ "epoch": 37.96419213973799,
+ "grad_norm": 0.01562986709177494,
+ "learning_rate": 0.0006,
+ "loss": 4.065832614898682,
+ "step": 2733
+ },
+ {
+ "epoch": 37.97816593886463,
+ "grad_norm": 0.01677163504064083,
+ "learning_rate": 0.0006,
+ "loss": 4.077518463134766,
+ "step": 2734
+ },
+ {
+ "epoch": 37.992139737991266,
+ "grad_norm": 0.01755492016673088,
+ "learning_rate": 0.0006,
+ "loss": 4.1076154708862305,
+ "step": 2735
+ },
+ {
+ "epoch": 38.0,
+ "grad_norm": 0.01929762214422226,
+ "learning_rate": 0.0006,
+ "loss": 4.09603214263916,
+ "step": 2736
+ },
+ {
+ "epoch": 38.0,
+ "eval_loss": 4.582442283630371,
+ "eval_runtime": 62.8649,
+ "eval_samples_per_second": 38.845,
+ "eval_steps_per_second": 1.225,
+ "step": 2736
+ },
+ {
+ "epoch": 38.01397379912664,
+ "grad_norm": 0.019129136577248573,
+ "learning_rate": 0.0006,
+ "loss": 4.095005989074707,
+ "step": 2737
+ },
+ {
+ "epoch": 38.02794759825328,
+ "grad_norm": 0.019167408347129822,
+ "learning_rate": 0.0006,
+ "loss": 4.054426670074463,
+ "step": 2738
+ },
+ {
+ "epoch": 38.041921397379916,
+ "grad_norm": 0.01789460889995098,
+ "learning_rate": 0.0006,
+ "loss": 4.0434417724609375,
+ "step": 2739
+ },
+ {
+ "epoch": 38.05589519650655,
+ "grad_norm": 0.015947027131915092,
+ "learning_rate": 0.0006,
+ "loss": 4.074516296386719,
+ "step": 2740
+ },
+ {
+ "epoch": 38.069868995633186,
+ "grad_norm": 0.016667252406477928,
+ "learning_rate": 0.0006,
+ "loss": 4.108545303344727,
+ "step": 2741
+ },
+ {
+ "epoch": 38.083842794759825,
+ "grad_norm": 0.016014399006962776,
+ "learning_rate": 0.0006,
+ "loss": 4.026925086975098,
+ "step": 2742
+ },
+ {
+ "epoch": 38.09781659388646,
+ "grad_norm": 0.015397695824503899,
+ "learning_rate": 0.0006,
+ "loss": 4.163970470428467,
+ "step": 2743
+ },
+ {
+ "epoch": 38.1117903930131,
+ "grad_norm": 0.01563289947807789,
+ "learning_rate": 0.0006,
+ "loss": 4.116578578948975,
+ "step": 2744
+ },
+ {
+ "epoch": 38.12576419213974,
+ "grad_norm": 0.015512318350374699,
+ "learning_rate": 0.0006,
+ "loss": 4.096311569213867,
+ "step": 2745
+ },
+ {
+ "epoch": 38.13973799126637,
+ "grad_norm": 0.015426642261445522,
+ "learning_rate": 0.0006,
+ "loss": 4.017065048217773,
+ "step": 2746
+ },
+ {
+ "epoch": 38.15371179039301,
+ "grad_norm": 0.015115435235202312,
+ "learning_rate": 0.0006,
+ "loss": 4.005434989929199,
+ "step": 2747
+ },
+ {
+ "epoch": 38.16768558951965,
+ "grad_norm": 0.015235894359648228,
+ "learning_rate": 0.0006,
+ "loss": 3.982743263244629,
+ "step": 2748
+ },
+ {
+ "epoch": 38.18165938864629,
+ "grad_norm": 0.016095101833343506,
+ "learning_rate": 0.0006,
+ "loss": 4.0849385261535645,
+ "step": 2749
+ },
+ {
+ "epoch": 38.19563318777293,
+ "grad_norm": 0.015201197937130928,
+ "learning_rate": 0.0006,
+ "loss": 4.103863716125488,
+ "step": 2750
+ },
+ {
+ "epoch": 38.209606986899566,
+ "grad_norm": 0.015687478706240654,
+ "learning_rate": 0.0006,
+ "loss": 4.12074089050293,
+ "step": 2751
+ },
+ {
+ "epoch": 38.223580786026204,
+ "grad_norm": 0.015201589092612267,
+ "learning_rate": 0.0006,
+ "loss": 4.056872367858887,
+ "step": 2752
+ },
+ {
+ "epoch": 38.237554585152836,
+ "grad_norm": 0.015154894441366196,
+ "learning_rate": 0.0006,
+ "loss": 4.089670181274414,
+ "step": 2753
+ },
+ {
+ "epoch": 38.251528384279474,
+ "grad_norm": 0.01586608588695526,
+ "learning_rate": 0.0006,
+ "loss": 4.155317306518555,
+ "step": 2754
+ },
+ {
+ "epoch": 38.26550218340611,
+ "grad_norm": 0.01609654724597931,
+ "learning_rate": 0.0006,
+ "loss": 3.995800495147705,
+ "step": 2755
+ },
+ {
+ "epoch": 38.27947598253275,
+ "grad_norm": 0.015495185740292072,
+ "learning_rate": 0.0006,
+ "loss": 3.995769739151001,
+ "step": 2756
+ },
+ {
+ "epoch": 38.29344978165939,
+ "grad_norm": 0.017042381688952446,
+ "learning_rate": 0.0006,
+ "loss": 4.132832050323486,
+ "step": 2757
+ },
+ {
+ "epoch": 38.30742358078603,
+ "grad_norm": 0.018706614151597023,
+ "learning_rate": 0.0006,
+ "loss": 4.038820743560791,
+ "step": 2758
+ },
+ {
+ "epoch": 38.32139737991266,
+ "grad_norm": 0.018316026777029037,
+ "learning_rate": 0.0006,
+ "loss": 3.972052574157715,
+ "step": 2759
+ },
+ {
+ "epoch": 38.3353711790393,
+ "grad_norm": 0.018375389277935028,
+ "learning_rate": 0.0006,
+ "loss": 4.152097702026367,
+ "step": 2760
+ },
+ {
+ "epoch": 38.34934497816594,
+ "grad_norm": 0.016879886388778687,
+ "learning_rate": 0.0006,
+ "loss": 4.020869731903076,
+ "step": 2761
+ },
+ {
+ "epoch": 38.36331877729258,
+ "grad_norm": 0.015728183090686798,
+ "learning_rate": 0.0006,
+ "loss": 4.008016586303711,
+ "step": 2762
+ },
+ {
+ "epoch": 38.377292576419215,
+ "grad_norm": 0.016147758811712265,
+ "learning_rate": 0.0006,
+ "loss": 4.035854339599609,
+ "step": 2763
+ },
+ {
+ "epoch": 38.391266375545854,
+ "grad_norm": 0.017774797976017,
+ "learning_rate": 0.0006,
+ "loss": 4.122506141662598,
+ "step": 2764
+ },
+ {
+ "epoch": 38.40524017467249,
+ "grad_norm": 0.016590069979429245,
+ "learning_rate": 0.0006,
+ "loss": 4.082847595214844,
+ "step": 2765
+ },
+ {
+ "epoch": 38.419213973799124,
+ "grad_norm": 0.016544528305530548,
+ "learning_rate": 0.0006,
+ "loss": 4.184270858764648,
+ "step": 2766
+ },
+ {
+ "epoch": 38.43318777292576,
+ "grad_norm": 0.017972925677895546,
+ "learning_rate": 0.0006,
+ "loss": 3.987558603286743,
+ "step": 2767
+ },
+ {
+ "epoch": 38.4471615720524,
+ "grad_norm": 0.018562227487564087,
+ "learning_rate": 0.0006,
+ "loss": 4.1193461418151855,
+ "step": 2768
+ },
+ {
+ "epoch": 38.46113537117904,
+ "grad_norm": 0.018896685913205147,
+ "learning_rate": 0.0006,
+ "loss": 4.057808876037598,
+ "step": 2769
+ },
+ {
+ "epoch": 38.47510917030568,
+ "grad_norm": 0.016793809831142426,
+ "learning_rate": 0.0006,
+ "loss": 4.069523334503174,
+ "step": 2770
+ },
+ {
+ "epoch": 38.48908296943232,
+ "grad_norm": 0.01698676310479641,
+ "learning_rate": 0.0006,
+ "loss": 3.997624158859253,
+ "step": 2771
+ },
+ {
+ "epoch": 38.50305676855895,
+ "grad_norm": 0.016784202307462692,
+ "learning_rate": 0.0006,
+ "loss": 4.174795627593994,
+ "step": 2772
+ },
+ {
+ "epoch": 38.51703056768559,
+ "grad_norm": 0.016785014420747757,
+ "learning_rate": 0.0006,
+ "loss": 3.999321222305298,
+ "step": 2773
+ },
+ {
+ "epoch": 38.531004366812226,
+ "grad_norm": 0.018615184351801872,
+ "learning_rate": 0.0006,
+ "loss": 4.191237926483154,
+ "step": 2774
+ },
+ {
+ "epoch": 38.544978165938865,
+ "grad_norm": 0.019090790301561356,
+ "learning_rate": 0.0006,
+ "loss": 4.106966018676758,
+ "step": 2775
+ },
+ {
+ "epoch": 38.5589519650655,
+ "grad_norm": 0.018298320472240448,
+ "learning_rate": 0.0006,
+ "loss": 4.063363552093506,
+ "step": 2776
+ },
+ {
+ "epoch": 38.57292576419214,
+ "grad_norm": 0.018541090190410614,
+ "learning_rate": 0.0006,
+ "loss": 4.101193428039551,
+ "step": 2777
+ },
+ {
+ "epoch": 38.58689956331878,
+ "grad_norm": 0.01849970407783985,
+ "learning_rate": 0.0006,
+ "loss": 4.071335792541504,
+ "step": 2778
+ },
+ {
+ "epoch": 38.60087336244541,
+ "grad_norm": 0.018309077247977257,
+ "learning_rate": 0.0006,
+ "loss": 4.245305061340332,
+ "step": 2779
+ },
+ {
+ "epoch": 38.61484716157205,
+ "grad_norm": 0.017681840807199478,
+ "learning_rate": 0.0006,
+ "loss": 4.081173419952393,
+ "step": 2780
+ },
+ {
+ "epoch": 38.62882096069869,
+ "grad_norm": 0.016286224126815796,
+ "learning_rate": 0.0006,
+ "loss": 4.024351596832275,
+ "step": 2781
+ },
+ {
+ "epoch": 38.64279475982533,
+ "grad_norm": 0.015837235376238823,
+ "learning_rate": 0.0006,
+ "loss": 4.028205394744873,
+ "step": 2782
+ },
+ {
+ "epoch": 38.65676855895197,
+ "grad_norm": 0.016062596812844276,
+ "learning_rate": 0.0006,
+ "loss": 4.131122589111328,
+ "step": 2783
+ },
+ {
+ "epoch": 38.670742358078606,
+ "grad_norm": 0.01711239665746689,
+ "learning_rate": 0.0006,
+ "loss": 4.124416351318359,
+ "step": 2784
+ },
+ {
+ "epoch": 38.68471615720524,
+ "grad_norm": 0.015304913744330406,
+ "learning_rate": 0.0006,
+ "loss": 4.056122779846191,
+ "step": 2785
+ },
+ {
+ "epoch": 38.698689956331876,
+ "grad_norm": 0.015971887856721878,
+ "learning_rate": 0.0006,
+ "loss": 4.132555961608887,
+ "step": 2786
+ },
+ {
+ "epoch": 38.712663755458514,
+ "grad_norm": 0.015787221491336823,
+ "learning_rate": 0.0006,
+ "loss": 4.051596164703369,
+ "step": 2787
+ },
+ {
+ "epoch": 38.72663755458515,
+ "grad_norm": 0.01605444774031639,
+ "learning_rate": 0.0006,
+ "loss": 4.191808700561523,
+ "step": 2788
+ },
+ {
+ "epoch": 38.74061135371179,
+ "grad_norm": 0.016383009031414986,
+ "learning_rate": 0.0006,
+ "loss": 4.0648417472839355,
+ "step": 2789
+ },
+ {
+ "epoch": 38.75458515283843,
+ "grad_norm": 0.016135554760694504,
+ "learning_rate": 0.0006,
+ "loss": 3.988563060760498,
+ "step": 2790
+ },
+ {
+ "epoch": 38.76855895196506,
+ "grad_norm": 0.016634158790111542,
+ "learning_rate": 0.0006,
+ "loss": 4.133493423461914,
+ "step": 2791
+ },
+ {
+ "epoch": 38.7825327510917,
+ "grad_norm": 0.01688210666179657,
+ "learning_rate": 0.0006,
+ "loss": 4.091277122497559,
+ "step": 2792
+ },
+ {
+ "epoch": 38.79650655021834,
+ "grad_norm": 0.016102122142910957,
+ "learning_rate": 0.0006,
+ "loss": 4.051219940185547,
+ "step": 2793
+ },
+ {
+ "epoch": 38.81048034934498,
+ "grad_norm": 0.016585038974881172,
+ "learning_rate": 0.0006,
+ "loss": 4.058587074279785,
+ "step": 2794
+ },
+ {
+ "epoch": 38.82445414847162,
+ "grad_norm": 0.017700951546430588,
+ "learning_rate": 0.0006,
+ "loss": 4.016030311584473,
+ "step": 2795
+ },
+ {
+ "epoch": 38.838427947598255,
+ "grad_norm": 0.017922203987836838,
+ "learning_rate": 0.0006,
+ "loss": 4.079977989196777,
+ "step": 2796
+ },
+ {
+ "epoch": 38.852401746724894,
+ "grad_norm": 0.017299890518188477,
+ "learning_rate": 0.0006,
+ "loss": 4.062984943389893,
+ "step": 2797
+ },
+ {
+ "epoch": 38.866375545851525,
+ "grad_norm": 0.018648581579327583,
+ "learning_rate": 0.0006,
+ "loss": 4.065717697143555,
+ "step": 2798
+ },
+ {
+ "epoch": 38.880349344978164,
+ "grad_norm": 0.019057920202612877,
+ "learning_rate": 0.0006,
+ "loss": 4.227892875671387,
+ "step": 2799
+ },
+ {
+ "epoch": 38.8943231441048,
+ "grad_norm": 0.018824422731995583,
+ "learning_rate": 0.0006,
+ "loss": 4.116058349609375,
+ "step": 2800
+ },
+ {
+ "epoch": 38.90829694323144,
+ "grad_norm": 0.019084978848695755,
+ "learning_rate": 0.0006,
+ "loss": 4.061677932739258,
+ "step": 2801
+ },
+ {
+ "epoch": 38.92227074235808,
+ "grad_norm": 0.017842981964349747,
+ "learning_rate": 0.0006,
+ "loss": 4.083813667297363,
+ "step": 2802
+ },
+ {
+ "epoch": 38.93624454148472,
+ "grad_norm": 0.01880858838558197,
+ "learning_rate": 0.0006,
+ "loss": 4.152561187744141,
+ "step": 2803
+ },
+ {
+ "epoch": 38.95021834061135,
+ "grad_norm": 0.017400087788701057,
+ "learning_rate": 0.0006,
+ "loss": 4.154635429382324,
+ "step": 2804
+ },
+ {
+ "epoch": 38.96419213973799,
+ "grad_norm": 0.017682507634162903,
+ "learning_rate": 0.0006,
+ "loss": 4.151128768920898,
+ "step": 2805
+ },
+ {
+ "epoch": 38.97816593886463,
+ "grad_norm": 0.015803052112460136,
+ "learning_rate": 0.0006,
+ "loss": 4.10498571395874,
+ "step": 2806
+ },
+ {
+ "epoch": 38.992139737991266,
+ "grad_norm": 0.01582511141896248,
+ "learning_rate": 0.0006,
+ "loss": 4.140159606933594,
+ "step": 2807
+ },
+ {
+ "epoch": 39.0,
+ "grad_norm": 0.01925666816532612,
+ "learning_rate": 0.0006,
+ "loss": 4.126394271850586,
+ "step": 2808
+ },
+ {
+ "epoch": 39.0,
+ "eval_loss": 4.615617275238037,
+ "eval_runtime": 56.8895,
+ "eval_samples_per_second": 42.925,
+ "eval_steps_per_second": 1.354,
+ "step": 2808
+ },
+ {
+ "epoch": 39.01397379912664,
+ "grad_norm": 0.01677853614091873,
+ "learning_rate": 0.0006,
+ "loss": 4.102395057678223,
+ "step": 2809
+ },
+ {
+ "epoch": 39.02794759825328,
+ "grad_norm": 0.017901137471199036,
+ "learning_rate": 0.0006,
+ "loss": 3.976609230041504,
+ "step": 2810
+ },
+ {
+ "epoch": 39.041921397379916,
+ "grad_norm": 0.017978668212890625,
+ "learning_rate": 0.0006,
+ "loss": 3.944416046142578,
+ "step": 2811
+ },
+ {
+ "epoch": 39.05589519650655,
+ "grad_norm": 0.019967492669820786,
+ "learning_rate": 0.0006,
+ "loss": 3.992234468460083,
+ "step": 2812
+ },
+ {
+ "epoch": 39.069868995633186,
+ "grad_norm": 0.018937544897198677,
+ "learning_rate": 0.0006,
+ "loss": 4.028573036193848,
+ "step": 2813
+ },
+ {
+ "epoch": 39.083842794759825,
+ "grad_norm": 0.016852952539920807,
+ "learning_rate": 0.0006,
+ "loss": 4.062442779541016,
+ "step": 2814
+ },
+ {
+ "epoch": 39.09781659388646,
+ "grad_norm": 0.016643647104501724,
+ "learning_rate": 0.0006,
+ "loss": 4.08871603012085,
+ "step": 2815
+ },
+ {
+ "epoch": 39.1117903930131,
+ "grad_norm": 0.016872866079211235,
+ "learning_rate": 0.0006,
+ "loss": 3.9514245986938477,
+ "step": 2816
+ },
+ {
+ "epoch": 39.12576419213974,
+ "grad_norm": 0.018045669421553612,
+ "learning_rate": 0.0006,
+ "loss": 4.038842678070068,
+ "step": 2817
+ },
+ {
+ "epoch": 39.13973799126637,
+ "grad_norm": 0.017853403463959694,
+ "learning_rate": 0.0006,
+ "loss": 3.93709135055542,
+ "step": 2818
+ },
+ {
+ "epoch": 39.15371179039301,
+ "grad_norm": 0.017507443204522133,
+ "learning_rate": 0.0006,
+ "loss": 3.98665714263916,
+ "step": 2819
+ },
+ {
+ "epoch": 39.16768558951965,
+ "grad_norm": 0.01766853965818882,
+ "learning_rate": 0.0006,
+ "loss": 3.937126636505127,
+ "step": 2820
+ },
+ {
+ "epoch": 39.18165938864629,
+ "grad_norm": 0.016660021618008614,
+ "learning_rate": 0.0006,
+ "loss": 3.8397014141082764,
+ "step": 2821
+ },
+ {
+ "epoch": 39.19563318777293,
+ "grad_norm": 0.017719309777021408,
+ "learning_rate": 0.0006,
+ "loss": 4.003579616546631,
+ "step": 2822
+ },
+ {
+ "epoch": 39.209606986899566,
+ "grad_norm": 0.018234414979815483,
+ "learning_rate": 0.0006,
+ "loss": 4.053859710693359,
+ "step": 2823
+ },
+ {
+ "epoch": 39.223580786026204,
+ "grad_norm": 0.018777024000883102,
+ "learning_rate": 0.0006,
+ "loss": 4.056182861328125,
+ "step": 2824
+ },
+ {
+ "epoch": 39.237554585152836,
+ "grad_norm": 0.017443502321839333,
+ "learning_rate": 0.0006,
+ "loss": 4.025068283081055,
+ "step": 2825
+ },
+ {
+ "epoch": 39.251528384279474,
+ "grad_norm": 0.016500331461429596,
+ "learning_rate": 0.0006,
+ "loss": 4.101631164550781,
+ "step": 2826
+ },
+ {
+ "epoch": 39.26550218340611,
+ "grad_norm": 0.016101544722914696,
+ "learning_rate": 0.0006,
+ "loss": 4.027082443237305,
+ "step": 2827
+ },
+ {
+ "epoch": 39.27947598253275,
+ "grad_norm": 0.015471171587705612,
+ "learning_rate": 0.0006,
+ "loss": 4.022909641265869,
+ "step": 2828
+ },
+ {
+ "epoch": 39.29344978165939,
+ "grad_norm": 0.015779748558998108,
+ "learning_rate": 0.0006,
+ "loss": 4.18796968460083,
+ "step": 2829
+ },
+ {
+ "epoch": 39.30742358078603,
+ "grad_norm": 0.015522679314017296,
+ "learning_rate": 0.0006,
+ "loss": 4.005777359008789,
+ "step": 2830
+ },
+ {
+ "epoch": 39.32139737991266,
+ "grad_norm": 0.015226504765450954,
+ "learning_rate": 0.0006,
+ "loss": 4.108416557312012,
+ "step": 2831
+ },
+ {
+ "epoch": 39.3353711790393,
+ "grad_norm": 0.015095778740942478,
+ "learning_rate": 0.0006,
+ "loss": 4.08029842376709,
+ "step": 2832
+ },
+ {
+ "epoch": 39.34934497816594,
+ "grad_norm": 0.015601031482219696,
+ "learning_rate": 0.0006,
+ "loss": 3.989492416381836,
+ "step": 2833
+ },
+ {
+ "epoch": 39.36331877729258,
+ "grad_norm": 0.014825052581727505,
+ "learning_rate": 0.0006,
+ "loss": 3.998520612716675,
+ "step": 2834
+ },
+ {
+ "epoch": 39.377292576419215,
+ "grad_norm": 0.014821100980043411,
+ "learning_rate": 0.0006,
+ "loss": 4.001777172088623,
+ "step": 2835
+ },
+ {
+ "epoch": 39.391266375545854,
+ "grad_norm": 0.015333754941821098,
+ "learning_rate": 0.0006,
+ "loss": 4.11985969543457,
+ "step": 2836
+ },
+ {
+ "epoch": 39.40524017467249,
+ "grad_norm": 0.015940699726343155,
+ "learning_rate": 0.0006,
+ "loss": 4.025373935699463,
+ "step": 2837
+ },
+ {
+ "epoch": 39.419213973799124,
+ "grad_norm": 0.015096287243068218,
+ "learning_rate": 0.0006,
+ "loss": 4.068551063537598,
+ "step": 2838
+ },
+ {
+ "epoch": 39.43318777292576,
+ "grad_norm": 0.01533548254519701,
+ "learning_rate": 0.0006,
+ "loss": 3.9800801277160645,
+ "step": 2839
+ },
+ {
+ "epoch": 39.4471615720524,
+ "grad_norm": 0.014648732729256153,
+ "learning_rate": 0.0006,
+ "loss": 3.9670190811157227,
+ "step": 2840
+ },
+ {
+ "epoch": 39.46113537117904,
+ "grad_norm": 0.015759671106934547,
+ "learning_rate": 0.0006,
+ "loss": 4.050443649291992,
+ "step": 2841
+ },
+ {
+ "epoch": 39.47510917030568,
+ "grad_norm": 0.016054073348641396,
+ "learning_rate": 0.0006,
+ "loss": 4.22734260559082,
+ "step": 2842
+ },
+ {
+ "epoch": 39.48908296943232,
+ "grad_norm": 0.0160264540463686,
+ "learning_rate": 0.0006,
+ "loss": 4.116174221038818,
+ "step": 2843
+ },
+ {
+ "epoch": 39.50305676855895,
+ "grad_norm": 0.015426610596477985,
+ "learning_rate": 0.0006,
+ "loss": 4.000178337097168,
+ "step": 2844
+ },
+ {
+ "epoch": 39.51703056768559,
+ "grad_norm": 0.01520642638206482,
+ "learning_rate": 0.0006,
+ "loss": 4.218410968780518,
+ "step": 2845
+ },
+ {
+ "epoch": 39.531004366812226,
+ "grad_norm": 0.015305949375033379,
+ "learning_rate": 0.0006,
+ "loss": 4.025891304016113,
+ "step": 2846
+ },
+ {
+ "epoch": 39.544978165938865,
+ "grad_norm": 0.015461564064025879,
+ "learning_rate": 0.0006,
+ "loss": 4.100950717926025,
+ "step": 2847
+ },
+ {
+ "epoch": 39.5589519650655,
+ "grad_norm": 0.01712625101208687,
+ "learning_rate": 0.0006,
+ "loss": 4.0620856285095215,
+ "step": 2848
+ },
+ {
+ "epoch": 39.57292576419214,
+ "grad_norm": 0.017133301123976707,
+ "learning_rate": 0.0006,
+ "loss": 4.097138404846191,
+ "step": 2849
+ },
+ {
+ "epoch": 39.58689956331878,
+ "grad_norm": 0.01683625392615795,
+ "learning_rate": 0.0006,
+ "loss": 4.084364891052246,
+ "step": 2850
+ },
+ {
+ "epoch": 39.60087336244541,
+ "grad_norm": 0.0165710486471653,
+ "learning_rate": 0.0006,
+ "loss": 4.042361259460449,
+ "step": 2851
+ },
+ {
+ "epoch": 39.61484716157205,
+ "grad_norm": 0.01660393364727497,
+ "learning_rate": 0.0006,
+ "loss": 4.1395487785339355,
+ "step": 2852
+ },
+ {
+ "epoch": 39.62882096069869,
+ "grad_norm": 0.017025845125317574,
+ "learning_rate": 0.0006,
+ "loss": 4.1207075119018555,
+ "step": 2853
+ },
+ {
+ "epoch": 39.64279475982533,
+ "grad_norm": 0.01660497486591339,
+ "learning_rate": 0.0006,
+ "loss": 4.103846073150635,
+ "step": 2854
+ },
+ {
+ "epoch": 39.65676855895197,
+ "grad_norm": 0.01607946865260601,
+ "learning_rate": 0.0006,
+ "loss": 4.189016342163086,
+ "step": 2855
+ },
+ {
+ "epoch": 39.670742358078606,
+ "grad_norm": 0.01614505983889103,
+ "learning_rate": 0.0006,
+ "loss": 4.079860687255859,
+ "step": 2856
+ },
+ {
+ "epoch": 39.68471615720524,
+ "grad_norm": 0.017085086554288864,
+ "learning_rate": 0.0006,
+ "loss": 3.9774293899536133,
+ "step": 2857
+ },
+ {
+ "epoch": 39.698689956331876,
+ "grad_norm": 0.01816520094871521,
+ "learning_rate": 0.0006,
+ "loss": 4.113666534423828,
+ "step": 2858
+ },
+ {
+ "epoch": 39.712663755458514,
+ "grad_norm": 0.016597526147961617,
+ "learning_rate": 0.0006,
+ "loss": 4.068242073059082,
+ "step": 2859
+ },
+ {
+ "epoch": 39.72663755458515,
+ "grad_norm": 0.01672235131263733,
+ "learning_rate": 0.0006,
+ "loss": 4.024444580078125,
+ "step": 2860
+ },
+ {
+ "epoch": 39.74061135371179,
+ "grad_norm": 0.020093556493520737,
+ "learning_rate": 0.0006,
+ "loss": 4.064124584197998,
+ "step": 2861
+ },
+ {
+ "epoch": 39.75458515283843,
+ "grad_norm": 0.023141304031014442,
+ "learning_rate": 0.0006,
+ "loss": 4.130439758300781,
+ "step": 2862
+ },
+ {
+ "epoch": 39.76855895196506,
+ "grad_norm": 0.02289372682571411,
+ "learning_rate": 0.0006,
+ "loss": 4.14794921875,
+ "step": 2863
+ },
+ {
+ "epoch": 39.7825327510917,
+ "grad_norm": 0.01919020712375641,
+ "learning_rate": 0.0006,
+ "loss": 4.123961925506592,
+ "step": 2864
+ },
+ {
+ "epoch": 39.79650655021834,
+ "grad_norm": 0.016932329162955284,
+ "learning_rate": 0.0006,
+ "loss": 4.0944132804870605,
+ "step": 2865
+ },
+ {
+ "epoch": 39.81048034934498,
+ "grad_norm": 0.016644535586237907,
+ "learning_rate": 0.0006,
+ "loss": 4.124269962310791,
+ "step": 2866
+ },
+ {
+ "epoch": 39.82445414847162,
+ "grad_norm": 0.016302216798067093,
+ "learning_rate": 0.0006,
+ "loss": 4.0267205238342285,
+ "step": 2867
+ },
+ {
+ "epoch": 39.838427947598255,
+ "grad_norm": 0.01673300378024578,
+ "learning_rate": 0.0006,
+ "loss": 4.10650634765625,
+ "step": 2868
+ },
+ {
+ "epoch": 39.852401746724894,
+ "grad_norm": 0.016581356525421143,
+ "learning_rate": 0.0006,
+ "loss": 4.1547956466674805,
+ "step": 2869
+ },
+ {
+ "epoch": 39.866375545851525,
+ "grad_norm": 0.01691797934472561,
+ "learning_rate": 0.0006,
+ "loss": 4.056936740875244,
+ "step": 2870
+ },
+ {
+ "epoch": 39.880349344978164,
+ "grad_norm": 0.01596726104617119,
+ "learning_rate": 0.0006,
+ "loss": 3.9891414642333984,
+ "step": 2871
+ },
+ {
+ "epoch": 39.8943231441048,
+ "grad_norm": 0.015615091659128666,
+ "learning_rate": 0.0006,
+ "loss": 4.137948036193848,
+ "step": 2872
+ },
+ {
+ "epoch": 39.90829694323144,
+ "grad_norm": 0.015703413635492325,
+ "learning_rate": 0.0006,
+ "loss": 3.9844930171966553,
+ "step": 2873
+ },
+ {
+ "epoch": 39.92227074235808,
+ "grad_norm": 0.015453094616532326,
+ "learning_rate": 0.0006,
+ "loss": 4.098443508148193,
+ "step": 2874
+ },
+ {
+ "epoch": 39.93624454148472,
+ "grad_norm": 0.01449158601462841,
+ "learning_rate": 0.0006,
+ "loss": 4.090852737426758,
+ "step": 2875
+ },
+ {
+ "epoch": 39.95021834061135,
+ "grad_norm": 0.01537515688687563,
+ "learning_rate": 0.0006,
+ "loss": 4.239612579345703,
+ "step": 2876
+ },
+ {
+ "epoch": 39.96419213973799,
+ "grad_norm": 0.015989921987056732,
+ "learning_rate": 0.0006,
+ "loss": 4.142073154449463,
+ "step": 2877
+ },
+ {
+ "epoch": 39.97816593886463,
+ "grad_norm": 0.015985611826181412,
+ "learning_rate": 0.0006,
+ "loss": 4.100076675415039,
+ "step": 2878
+ },
+ {
+ "epoch": 39.992139737991266,
+ "grad_norm": 0.015260833315551281,
+ "learning_rate": 0.0006,
+ "loss": 4.068576335906982,
+ "step": 2879
+ },
+ {
+ "epoch": 40.0,
+ "grad_norm": 0.018047941848635674,
+ "learning_rate": 0.0006,
+ "loss": 4.005878925323486,
+ "step": 2880
+ },
+ {
+ "epoch": 40.0,
+ "eval_loss": 4.622848987579346,
+ "eval_runtime": 56.9618,
+ "eval_samples_per_second": 42.871,
+ "eval_steps_per_second": 1.352,
+ "step": 2880
+ },
+ {
+ "epoch": 40.01397379912664,
+ "grad_norm": 0.015976838767528534,
+ "learning_rate": 0.0006,
+ "loss": 4.035689353942871,
+ "step": 2881
+ },
+ {
+ "epoch": 40.02794759825328,
+ "grad_norm": 0.015176220797002316,
+ "learning_rate": 0.0006,
+ "loss": 3.932197332382202,
+ "step": 2882
+ },
+ {
+ "epoch": 40.041921397379916,
+ "grad_norm": 0.016676228493452072,
+ "learning_rate": 0.0006,
+ "loss": 4.014171123504639,
+ "step": 2883
+ },
+ {
+ "epoch": 40.05589519650655,
+ "grad_norm": 0.01698281429708004,
+ "learning_rate": 0.0006,
+ "loss": 4.117893218994141,
+ "step": 2884
+ },
+ {
+ "epoch": 40.069868995633186,
+ "grad_norm": 0.017872486263513565,
+ "learning_rate": 0.0006,
+ "loss": 4.132119178771973,
+ "step": 2885
+ },
+ {
+ "epoch": 40.083842794759825,
+ "grad_norm": 0.017615554854273796,
+ "learning_rate": 0.0006,
+ "loss": 4.072000980377197,
+ "step": 2886
+ },
+ {
+ "epoch": 40.09781659388646,
+ "grad_norm": 0.017469394952058792,
+ "learning_rate": 0.0006,
+ "loss": 4.017726421356201,
+ "step": 2887
+ },
+ {
+ "epoch": 40.1117903930131,
+ "grad_norm": 0.01954762637615204,
+ "learning_rate": 0.0006,
+ "loss": 4.014443874359131,
+ "step": 2888
+ },
+ {
+ "epoch": 40.12576419213974,
+ "grad_norm": 0.018935853615403175,
+ "learning_rate": 0.0006,
+ "loss": 4.0042948722839355,
+ "step": 2889
+ },
+ {
+ "epoch": 40.13973799126637,
+ "grad_norm": 0.019382692873477936,
+ "learning_rate": 0.0006,
+ "loss": 4.06160831451416,
+ "step": 2890
+ },
+ {
+ "epoch": 40.15371179039301,
+ "grad_norm": 0.019677115604281425,
+ "learning_rate": 0.0006,
+ "loss": 4.13466215133667,
+ "step": 2891
+ },
+ {
+ "epoch": 40.16768558951965,
+ "grad_norm": 0.020365122705698013,
+ "learning_rate": 0.0006,
+ "loss": 4.095000267028809,
+ "step": 2892
+ },
+ {
+ "epoch": 40.18165938864629,
+ "grad_norm": 0.021842138841748238,
+ "learning_rate": 0.0006,
+ "loss": 3.9877853393554688,
+ "step": 2893
+ },
+ {
+ "epoch": 40.19563318777293,
+ "grad_norm": 0.02354881353676319,
+ "learning_rate": 0.0006,
+ "loss": 3.9430112838745117,
+ "step": 2894
+ },
+ {
+ "epoch": 40.209606986899566,
+ "grad_norm": 0.026602735742926598,
+ "learning_rate": 0.0006,
+ "loss": 4.012076377868652,
+ "step": 2895
+ },
+ {
+ "epoch": 40.223580786026204,
+ "grad_norm": 0.023976484313607216,
+ "learning_rate": 0.0006,
+ "loss": 3.9768595695495605,
+ "step": 2896
+ },
+ {
+ "epoch": 40.237554585152836,
+ "grad_norm": 0.02336616814136505,
+ "learning_rate": 0.0006,
+ "loss": 4.1096038818359375,
+ "step": 2897
+ },
+ {
+ "epoch": 40.251528384279474,
+ "grad_norm": 0.02426023595035076,
+ "learning_rate": 0.0006,
+ "loss": 4.157189846038818,
+ "step": 2898
+ },
+ {
+ "epoch": 40.26550218340611,
+ "grad_norm": 0.021572696045041084,
+ "learning_rate": 0.0006,
+ "loss": 3.979175090789795,
+ "step": 2899
+ },
+ {
+ "epoch": 40.27947598253275,
+ "grad_norm": 0.08300432562828064,
+ "learning_rate": 0.0006,
+ "loss": 4.092336654663086,
+ "step": 2900
+ },
+ {
+ "epoch": 40.29344978165939,
+ "grad_norm": 0.7323967814445496,
+ "learning_rate": 0.0006,
+ "loss": 5.337990760803223,
+ "step": 2901
+ },
+ {
+ "epoch": 40.30742358078603,
+ "grad_norm": 0.3363337814807892,
+ "learning_rate": 0.0006,
+ "loss": 6.40043830871582,
+ "step": 2902
+ },
+ {
+ "epoch": 40.32139737991266,
+ "grad_norm": 0.598686933517456,
+ "learning_rate": 0.0006,
+ "loss": 7.004765510559082,
+ "step": 2903
+ },
+ {
+ "epoch": 40.3353711790393,
+ "grad_norm": 0.5998364686965942,
+ "learning_rate": 0.0006,
+ "loss": 8.184747695922852,
+ "step": 2904
+ },
+ {
+ "epoch": 40.34934497816594,
+ "grad_norm": 1.408208966255188,
+ "learning_rate": 0.0006,
+ "loss": 8.88978385925293,
+ "step": 2905
+ },
+ {
+ "epoch": 40.36331877729258,
+ "grad_norm": 0.7829785943031311,
+ "learning_rate": 0.0006,
+ "loss": 11.13601303100586,
+ "step": 2906
+ },
+ {
+ "epoch": 40.377292576419215,
+ "grad_norm": 0.7769215106964111,
+ "learning_rate": 0.0006,
+ "loss": 9.90416145324707,
+ "step": 2907
+ },
+ {
+ "epoch": 40.391266375545854,
+ "grad_norm": 0.46508079767227173,
+ "learning_rate": 0.0006,
+ "loss": 9.647209167480469,
+ "step": 2908
+ },
+ {
+ "epoch": 40.40524017467249,
+ "grad_norm": 1.0228968858718872,
+ "learning_rate": 0.0006,
+ "loss": 11.685434341430664,
+ "step": 2909
+ },
+ {
+ "epoch": 40.419213973799124,
+ "grad_norm": 0.20553778111934662,
+ "learning_rate": 0.0006,
+ "loss": 8.621708869934082,
+ "step": 2910
+ },
+ {
+ "epoch": 40.43318777292576,
+ "grad_norm": 0.18466424942016602,
+ "learning_rate": 0.0006,
+ "loss": 8.590787887573242,
+ "step": 2911
+ },
+ {
+ "epoch": 40.4471615720524,
+ "grad_norm": 0.10458594560623169,
+ "learning_rate": 0.0006,
+ "loss": 8.324458122253418,
+ "step": 2912
+ },
+ {
+ "epoch": 40.46113537117904,
+ "grad_norm": 0.1096842959523201,
+ "learning_rate": 0.0006,
+ "loss": 8.300851821899414,
+ "step": 2913
+ },
+ {
+ "epoch": 40.47510917030568,
+ "grad_norm": 0.10501226782798767,
+ "learning_rate": 0.0006,
+ "loss": 8.21009635925293,
+ "step": 2914
+ },
+ {
+ "epoch": 40.48908296943232,
+ "grad_norm": 0.13704165816307068,
+ "learning_rate": 0.0006,
+ "loss": 8.141355514526367,
+ "step": 2915
+ },
+ {
+ "epoch": 40.50305676855895,
+ "grad_norm": 0.12129108607769012,
+ "learning_rate": 0.0006,
+ "loss": 8.074612617492676,
+ "step": 2916
+ },
+ {
+ "epoch": 40.51703056768559,
+ "grad_norm": 0.10028897225856781,
+ "learning_rate": 0.0006,
+ "loss": 8.00384521484375,
+ "step": 2917
+ },
+ {
+ "epoch": 40.531004366812226,
+ "grad_norm": 0.11764246970415115,
+ "learning_rate": 0.0006,
+ "loss": 8.078598022460938,
+ "step": 2918
+ },
+ {
+ "epoch": 40.544978165938865,
+ "grad_norm": 0.10081959515810013,
+ "learning_rate": 0.0006,
+ "loss": 8.008938789367676,
+ "step": 2919
+ },
+ {
+ "epoch": 40.5589519650655,
+ "grad_norm": 0.1003170907497406,
+ "learning_rate": 0.0006,
+ "loss": 7.962940216064453,
+ "step": 2920
+ },
+ {
+ "epoch": 40.57292576419214,
+ "grad_norm": 0.08763492107391357,
+ "learning_rate": 0.0006,
+ "loss": 7.909419536590576,
+ "step": 2921
+ },
+ {
+ "epoch": 40.58689956331878,
+ "grad_norm": 0.0636807233095169,
+ "learning_rate": 0.0006,
+ "loss": 7.878905773162842,
+ "step": 2922
+ },
+ {
+ "epoch": 40.60087336244541,
+ "grad_norm": 0.06005936488509178,
+ "learning_rate": 0.0006,
+ "loss": 7.850980758666992,
+ "step": 2923
+ },
+ {
+ "epoch": 40.61484716157205,
+ "grad_norm": 0.06415296345949173,
+ "learning_rate": 0.0006,
+ "loss": 7.830353260040283,
+ "step": 2924
+ },
+ {
+ "epoch": 40.62882096069869,
+ "grad_norm": 0.0732007846236229,
+ "learning_rate": 0.0006,
+ "loss": 7.830746650695801,
+ "step": 2925
+ },
+ {
+ "epoch": 40.64279475982533,
+ "grad_norm": 0.06628110259771347,
+ "learning_rate": 0.0006,
+ "loss": 7.82661771774292,
+ "step": 2926
+ },
+ {
+ "epoch": 40.65676855895197,
+ "grad_norm": 0.05544218048453331,
+ "learning_rate": 0.0006,
+ "loss": 7.814305782318115,
+ "step": 2927
+ },
+ {
+ "epoch": 40.670742358078606,
+ "grad_norm": 0.050327103585004807,
+ "learning_rate": 0.0006,
+ "loss": 7.764055252075195,
+ "step": 2928
+ },
+ {
+ "epoch": 40.68471615720524,
+ "grad_norm": 0.04918425530195236,
+ "learning_rate": 0.0006,
+ "loss": 7.744625091552734,
+ "step": 2929
+ },
+ {
+ "epoch": 40.698689956331876,
+ "grad_norm": 0.042862266302108765,
+ "learning_rate": 0.0006,
+ "loss": 7.764426231384277,
+ "step": 2930
+ },
+ {
+ "epoch": 40.712663755458514,
+ "grad_norm": 0.04338089004158974,
+ "learning_rate": 0.0006,
+ "loss": 7.712530136108398,
+ "step": 2931
+ },
+ {
+ "epoch": 40.72663755458515,
+ "grad_norm": 0.04127737879753113,
+ "learning_rate": 0.0006,
+ "loss": 7.742180347442627,
+ "step": 2932
+ },
+ {
+ "epoch": 40.74061135371179,
+ "grad_norm": 0.03358631581068039,
+ "learning_rate": 0.0006,
+ "loss": 7.714834213256836,
+ "step": 2933
+ },
+ {
+ "epoch": 40.75458515283843,
+ "grad_norm": 0.02845790982246399,
+ "learning_rate": 0.0006,
+ "loss": 7.7021026611328125,
+ "step": 2934
+ },
+ {
+ "epoch": 40.76855895196506,
+ "grad_norm": 0.03012419492006302,
+ "learning_rate": 0.0006,
+ "loss": 7.712377548217773,
+ "step": 2935
+ },
+ {
+ "epoch": 40.7825327510917,
+ "grad_norm": 0.032981179654598236,
+ "learning_rate": 0.0006,
+ "loss": 7.666385650634766,
+ "step": 2936
+ },
+ {
+ "epoch": 40.79650655021834,
+ "grad_norm": 0.03357074037194252,
+ "learning_rate": 0.0006,
+ "loss": 7.649097919464111,
+ "step": 2937
+ },
+ {
+ "epoch": 40.81048034934498,
+ "grad_norm": 0.030914708971977234,
+ "learning_rate": 0.0006,
+ "loss": 7.649196624755859,
+ "step": 2938
+ },
+ {
+ "epoch": 40.82445414847162,
+ "grad_norm": 0.026296131312847137,
+ "learning_rate": 0.0006,
+ "loss": 7.629964828491211,
+ "step": 2939
+ },
+ {
+ "epoch": 40.838427947598255,
+ "grad_norm": 0.026677245274186134,
+ "learning_rate": 0.0006,
+ "loss": 7.599973678588867,
+ "step": 2940
+ },
+ {
+ "epoch": 40.852401746724894,
+ "grad_norm": 0.023208219558000565,
+ "learning_rate": 0.0006,
+ "loss": 7.601624488830566,
+ "step": 2941
+ },
+ {
+ "epoch": 40.866375545851525,
+ "grad_norm": 0.020741475746035576,
+ "learning_rate": 0.0006,
+ "loss": 7.603902339935303,
+ "step": 2942
+ },
+ {
+ "epoch": 40.880349344978164,
+ "grad_norm": 0.019221751019358635,
+ "learning_rate": 0.0006,
+ "loss": 7.5535888671875,
+ "step": 2943
+ },
+ {
+ "epoch": 40.8943231441048,
+ "grad_norm": 0.019491951912641525,
+ "learning_rate": 0.0006,
+ "loss": 7.54597282409668,
+ "step": 2944
+ },
+ {
+ "epoch": 40.90829694323144,
+ "grad_norm": 0.019702572375535965,
+ "learning_rate": 0.0006,
+ "loss": 7.537567138671875,
+ "step": 2945
+ },
+ {
+ "epoch": 40.92227074235808,
+ "grad_norm": 0.01768800988793373,
+ "learning_rate": 0.0006,
+ "loss": 7.512800216674805,
+ "step": 2946
+ },
+ {
+ "epoch": 40.93624454148472,
+ "grad_norm": 0.01926138624548912,
+ "learning_rate": 0.0006,
+ "loss": 7.509932518005371,
+ "step": 2947
+ },
+ {
+ "epoch": 40.95021834061135,
+ "grad_norm": 0.019030006602406502,
+ "learning_rate": 0.0006,
+ "loss": 7.497983932495117,
+ "step": 2948
+ },
+ {
+ "epoch": 40.96419213973799,
+ "grad_norm": 0.017313841730356216,
+ "learning_rate": 0.0006,
+ "loss": 7.463518142700195,
+ "step": 2949
+ },
+ {
+ "epoch": 40.97816593886463,
+ "grad_norm": 0.01310762483626604,
+ "learning_rate": 0.0006,
+ "loss": 7.474668502807617,
+ "step": 2950
+ },
+ {
+ "epoch": 40.992139737991266,
+ "grad_norm": 0.013008282519876957,
+ "learning_rate": 0.0006,
+ "loss": 7.45982551574707,
+ "step": 2951
+ },
+ {
+ "epoch": 41.0,
+ "grad_norm": 0.01441363524645567,
+ "learning_rate": 0.0006,
+ "loss": 7.446529388427734,
+ "step": 2952
+ },
+ {
+ "epoch": 41.0,
+ "eval_loss": 7.486120223999023,
+ "eval_runtime": 70.4986,
+ "eval_samples_per_second": 34.639,
+ "eval_steps_per_second": 1.092,
+ "step": 2952
+ },
+ {
+ "epoch": 41.01397379912664,
+ "grad_norm": 0.014061855152249336,
+ "learning_rate": 0.0006,
+ "loss": 7.4692182540893555,
+ "step": 2953
+ },
+ {
+ "epoch": 41.02794759825328,
+ "grad_norm": 0.01587500050663948,
+ "learning_rate": 0.0006,
+ "loss": 7.4182024002075195,
+ "step": 2954
+ },
+ {
+ "epoch": 41.041921397379916,
+ "grad_norm": 0.014783984050154686,
+ "learning_rate": 0.0006,
+ "loss": 7.466778755187988,
+ "step": 2955
+ },
+ {
+ "epoch": 41.05589519650655,
+ "grad_norm": 0.011560854502022266,
+ "learning_rate": 0.0006,
+ "loss": 7.438002586364746,
+ "step": 2956
+ },
+ {
+ "epoch": 41.069868995633186,
+ "grad_norm": 0.00970957800745964,
+ "learning_rate": 0.0006,
+ "loss": 7.442694187164307,
+ "step": 2957
+ },
+ {
+ "epoch": 41.083842794759825,
+ "grad_norm": 0.009633459150791168,
+ "learning_rate": 0.0006,
+ "loss": 7.458884239196777,
+ "step": 2958
+ },
+ {
+ "epoch": 41.09781659388646,
+ "grad_norm": 0.010664350353181362,
+ "learning_rate": 0.0006,
+ "loss": 7.48378324508667,
+ "step": 2959
+ },
+ {
+ "epoch": 41.1117903930131,
+ "grad_norm": 0.011440591886639595,
+ "learning_rate": 0.0006,
+ "loss": 7.474484920501709,
+ "step": 2960
+ },
+ {
+ "epoch": 41.12576419213974,
+ "grad_norm": 0.01150908600538969,
+ "learning_rate": 0.0006,
+ "loss": 7.417902946472168,
+ "step": 2961
+ },
+ {
+ "epoch": 41.13973799126637,
+ "grad_norm": 0.010375697165727615,
+ "learning_rate": 0.0006,
+ "loss": 7.403279781341553,
+ "step": 2962
+ },
+ {
+ "epoch": 41.15371179039301,
+ "grad_norm": 0.009075703099370003,
+ "learning_rate": 0.0006,
+ "loss": 7.447347164154053,
+ "step": 2963
+ },
+ {
+ "epoch": 41.16768558951965,
+ "grad_norm": 0.008892177604138851,
+ "learning_rate": 0.0006,
+ "loss": 7.402681350708008,
+ "step": 2964
+ },
+ {
+ "epoch": 41.18165938864629,
+ "grad_norm": 0.009492084383964539,
+ "learning_rate": 0.0006,
+ "loss": 7.3871355056762695,
+ "step": 2965
+ },
+ {
+ "epoch": 41.19563318777293,
+ "grad_norm": 0.009139460511505604,
+ "learning_rate": 0.0006,
+ "loss": 7.419617176055908,
+ "step": 2966
+ },
+ {
+ "epoch": 41.209606986899566,
+ "grad_norm": 0.007959138602018356,
+ "learning_rate": 0.0006,
+ "loss": 7.374949932098389,
+ "step": 2967
+ },
+ {
+ "epoch": 41.223580786026204,
+ "grad_norm": 0.007571548223495483,
+ "learning_rate": 0.0006,
+ "loss": 7.4205241203308105,
+ "step": 2968
+ },
+ {
+ "epoch": 41.237554585152836,
+ "grad_norm": 0.007885400205850601,
+ "learning_rate": 0.0006,
+ "loss": 7.416543006896973,
+ "step": 2969
+ },
+ {
+ "epoch": 41.251528384279474,
+ "grad_norm": 0.008650158531963825,
+ "learning_rate": 0.0006,
+ "loss": 7.357326984405518,
+ "step": 2970
+ },
+ {
+ "epoch": 41.26550218340611,
+ "grad_norm": 0.007483262103050947,
+ "learning_rate": 0.0006,
+ "loss": 7.410891056060791,
+ "step": 2971
+ },
+ {
+ "epoch": 41.27947598253275,
+ "grad_norm": 0.00869277399033308,
+ "learning_rate": 0.0006,
+ "loss": 7.392127990722656,
+ "step": 2972
+ },
+ {
+ "epoch": 41.29344978165939,
+ "grad_norm": 0.00885078776627779,
+ "learning_rate": 0.0006,
+ "loss": 7.405449867248535,
+ "step": 2973
+ },
+ {
+ "epoch": 41.30742358078603,
+ "grad_norm": 0.00958226341754198,
+ "learning_rate": 0.0006,
+ "loss": 7.399868488311768,
+ "step": 2974
+ },
+ {
+ "epoch": 41.32139737991266,
+ "grad_norm": 0.006991599686443806,
+ "learning_rate": 0.0006,
+ "loss": 7.35369873046875,
+ "step": 2975
+ },
+ {
+ "epoch": 41.3353711790393,
+ "grad_norm": 0.007415142375975847,
+ "learning_rate": 0.0006,
+ "loss": 7.358353137969971,
+ "step": 2976
+ },
+ {
+ "epoch": 41.34934497816594,
+ "grad_norm": 0.007024371065199375,
+ "learning_rate": 0.0006,
+ "loss": 7.4070892333984375,
+ "step": 2977
+ },
+ {
+ "epoch": 41.36331877729258,
+ "grad_norm": 0.007234332151710987,
+ "learning_rate": 0.0006,
+ "loss": 7.379120349884033,
+ "step": 2978
+ },
+ {
+ "epoch": 41.377292576419215,
+ "grad_norm": 0.00798892229795456,
+ "learning_rate": 0.0006,
+ "loss": 7.365455627441406,
+ "step": 2979
+ },
+ {
+ "epoch": 41.391266375545854,
+ "grad_norm": 0.007303398102521896,
+ "learning_rate": 0.0006,
+ "loss": 7.341798782348633,
+ "step": 2980
+ },
+ {
+ "epoch": 41.40524017467249,
+ "grad_norm": 0.007119669578969479,
+ "learning_rate": 0.0006,
+ "loss": 7.363280296325684,
+ "step": 2981
+ },
+ {
+ "epoch": 41.419213973799124,
+ "grad_norm": 0.007457619532942772,
+ "learning_rate": 0.0006,
+ "loss": 7.39388370513916,
+ "step": 2982
+ },
+ {
+ "epoch": 41.43318777292576,
+ "grad_norm": 0.007905688136816025,
+ "learning_rate": 0.0006,
+ "loss": 7.380652904510498,
+ "step": 2983
+ },
+ {
+ "epoch": 41.4471615720524,
+ "grad_norm": 0.006212855689227581,
+ "learning_rate": 0.0006,
+ "loss": 7.398769378662109,
+ "step": 2984
+ },
+ {
+ "epoch": 41.46113537117904,
+ "grad_norm": 0.0070641543716192245,
+ "learning_rate": 0.0006,
+ "loss": 7.3503546714782715,
+ "step": 2985
+ },
+ {
+ "epoch": 41.47510917030568,
+ "grad_norm": 0.00787805300205946,
+ "learning_rate": 0.0006,
+ "loss": 7.347678184509277,
+ "step": 2986
+ },
+ {
+ "epoch": 41.48908296943232,
+ "grad_norm": 0.006760426796972752,
+ "learning_rate": 0.0006,
+ "loss": 7.380006790161133,
+ "step": 2987
+ },
+ {
+ "epoch": 41.50305676855895,
+ "grad_norm": 0.00772385997697711,
+ "learning_rate": 0.0006,
+ "loss": 7.363680362701416,
+ "step": 2988
+ },
+ {
+ "epoch": 41.51703056768559,
+ "grad_norm": 0.007373564876616001,
+ "learning_rate": 0.0006,
+ "loss": 7.333014488220215,
+ "step": 2989
+ },
+ {
+ "epoch": 41.531004366812226,
+ "grad_norm": 0.006717653013765812,
+ "learning_rate": 0.0006,
+ "loss": 7.360076427459717,
+ "step": 2990
+ },
+ {
+ "epoch": 41.544978165938865,
+ "grad_norm": 0.007191413082182407,
+ "learning_rate": 0.0006,
+ "loss": 7.397030830383301,
+ "step": 2991
+ },
+ {
+ "epoch": 41.5589519650655,
+ "grad_norm": 0.007831960916519165,
+ "learning_rate": 0.0006,
+ "loss": 7.369959831237793,
+ "step": 2992
+ },
+ {
+ "epoch": 41.57292576419214,
+ "grad_norm": 0.006947481073439121,
+ "learning_rate": 0.0006,
+ "loss": 7.333254337310791,
+ "step": 2993
+ },
+ {
+ "epoch": 41.58689956331878,
+ "grad_norm": 0.006243606563657522,
+ "learning_rate": 0.0006,
+ "loss": 7.327378273010254,
+ "step": 2994
+ },
+ {
+ "epoch": 41.60087336244541,
+ "grad_norm": 0.006298339460045099,
+ "learning_rate": 0.0006,
+ "loss": 7.323482036590576,
+ "step": 2995
+ },
+ {
+ "epoch": 41.61484716157205,
+ "grad_norm": 0.007587412837892771,
+ "learning_rate": 0.0006,
+ "loss": 7.313789367675781,
+ "step": 2996
+ },
+ {
+ "epoch": 41.62882096069869,
+ "grad_norm": 0.007140746805816889,
+ "learning_rate": 0.0006,
+ "loss": 7.356067657470703,
+ "step": 2997
+ },
+ {
+ "epoch": 41.64279475982533,
+ "grad_norm": 0.008315625600516796,
+ "learning_rate": 0.0006,
+ "loss": 7.3339643478393555,
+ "step": 2998
+ },
+ {
+ "epoch": 41.65676855895197,
+ "grad_norm": 0.006547102238982916,
+ "learning_rate": 0.0006,
+ "loss": 7.342191696166992,
+ "step": 2999
+ },
+ {
+ "epoch": 41.670742358078606,
+ "grad_norm": 0.006068122107535601,
+ "learning_rate": 0.0006,
+ "loss": 7.321015357971191,
+ "step": 3000
+ },
+ {
+ "epoch": 41.68471615720524,
+ "grad_norm": 0.007316991221159697,
+ "learning_rate": 0.0006,
+ "loss": 7.346490859985352,
+ "step": 3001
+ },
+ {
+ "epoch": 41.698689956331876,
+ "grad_norm": 0.0067388564348220825,
+ "learning_rate": 0.0006,
+ "loss": 7.340723991394043,
+ "step": 3002
+ },
+ {
+ "epoch": 41.712663755458514,
+ "grad_norm": 0.006839963141828775,
+ "learning_rate": 0.0006,
+ "loss": 7.329365253448486,
+ "step": 3003
+ },
+ {
+ "epoch": 41.72663755458515,
+ "grad_norm": 0.00805041566491127,
+ "learning_rate": 0.0006,
+ "loss": 7.298934459686279,
+ "step": 3004
+ },
+ {
+ "epoch": 41.74061135371179,
+ "grad_norm": 0.007197503466159105,
+ "learning_rate": 0.0006,
+ "loss": 7.331496238708496,
+ "step": 3005
+ },
+ {
+ "epoch": 41.75458515283843,
+ "grad_norm": 0.007220805156975985,
+ "learning_rate": 0.0006,
+ "loss": 7.317695140838623,
+ "step": 3006
+ },
+ {
+ "epoch": 41.76855895196506,
+ "grad_norm": 0.00874276738613844,
+ "learning_rate": 0.0006,
+ "loss": 7.272614479064941,
+ "step": 3007
+ },
+ {
+ "epoch": 41.7825327510917,
+ "grad_norm": 0.008386565372347832,
+ "learning_rate": 0.0006,
+ "loss": 7.295244216918945,
+ "step": 3008
+ },
+ {
+ "epoch": 41.79650655021834,
+ "grad_norm": 0.0077730282209813595,
+ "learning_rate": 0.0006,
+ "loss": 7.3395490646362305,
+ "step": 3009
+ },
+ {
+ "epoch": 41.81048034934498,
+ "grad_norm": 0.008441480807960033,
+ "learning_rate": 0.0006,
+ "loss": 7.319921493530273,
+ "step": 3010
+ },
+ {
+ "epoch": 41.82445414847162,
+ "grad_norm": 0.007398269604891539,
+ "learning_rate": 0.0006,
+ "loss": 7.3029465675354,
+ "step": 3011
+ },
+ {
+ "epoch": 41.838427947598255,
+ "grad_norm": 0.006878677289932966,
+ "learning_rate": 0.0006,
+ "loss": 7.318467140197754,
+ "step": 3012
+ },
+ {
+ "epoch": 41.852401746724894,
+ "grad_norm": 0.007616598624736071,
+ "learning_rate": 0.0006,
+ "loss": 7.3112077713012695,
+ "step": 3013
+ },
+ {
+ "epoch": 41.866375545851525,
+ "grad_norm": 0.011629699729382992,
+ "learning_rate": 0.0006,
+ "loss": 7.32523250579834,
+ "step": 3014
+ },
+ {
+ "epoch": 41.880349344978164,
+ "grad_norm": 0.011011249385774136,
+ "learning_rate": 0.0006,
+ "loss": 7.287247657775879,
+ "step": 3015
+ },
+ {
+ "epoch": 41.8943231441048,
+ "grad_norm": 0.010208295658230782,
+ "learning_rate": 0.0006,
+ "loss": 7.292117118835449,
+ "step": 3016
+ },
+ {
+ "epoch": 41.90829694323144,
+ "grad_norm": 0.008651095442473888,
+ "learning_rate": 0.0006,
+ "loss": 7.302201271057129,
+ "step": 3017
+ },
+ {
+ "epoch": 41.92227074235808,
+ "grad_norm": 0.006692920345813036,
+ "learning_rate": 0.0006,
+ "loss": 7.319568634033203,
+ "step": 3018
+ },
+ {
+ "epoch": 41.93624454148472,
+ "grad_norm": 0.010033348575234413,
+ "learning_rate": 0.0006,
+ "loss": 7.2738566398620605,
+ "step": 3019
+ },
+ {
+ "epoch": 41.95021834061135,
+ "grad_norm": 0.011790832504630089,
+ "learning_rate": 0.0006,
+ "loss": 7.298540115356445,
+ "step": 3020
+ },
+ {
+ "epoch": 41.96419213973799,
+ "grad_norm": 0.018622662872076035,
+ "learning_rate": 0.0006,
+ "loss": 7.267182350158691,
+ "step": 3021
+ },
+ {
+ "epoch": 41.97816593886463,
+ "grad_norm": 0.02520257607102394,
+ "learning_rate": 0.0006,
+ "loss": 7.261566162109375,
+ "step": 3022
+ },
+ {
+ "epoch": 41.992139737991266,
+ "grad_norm": 0.036332644522190094,
+ "learning_rate": 0.0006,
+ "loss": 7.270719051361084,
+ "step": 3023
+ },
+ {
+ "epoch": 42.0,
+ "grad_norm": 0.04717305675148964,
+ "learning_rate": 0.0006,
+ "loss": 7.293332099914551,
+ "step": 3024
+ },
+ {
+ "epoch": 42.0,
+ "eval_loss": 7.354280471801758,
+ "eval_runtime": 56.4353,
+ "eval_samples_per_second": 43.271,
+ "eval_steps_per_second": 1.364,
+ "step": 3024
+ },
+ {
+ "epoch": 42.01397379912664,
+ "grad_norm": 0.06661626696586609,
+ "learning_rate": 0.0006,
+ "loss": 7.338189601898193,
+ "step": 3025
+ },
+ {
+ "epoch": 42.02794759825328,
+ "grad_norm": 0.05452054366469383,
+ "learning_rate": 0.0006,
+ "loss": 7.316764831542969,
+ "step": 3026
+ },
+ {
+ "epoch": 42.041921397379916,
+ "grad_norm": 0.025638775900006294,
+ "learning_rate": 0.0006,
+ "loss": 7.34204626083374,
+ "step": 3027
+ },
+ {
+ "epoch": 42.05589519650655,
+ "grad_norm": 0.022398799657821655,
+ "learning_rate": 0.0006,
+ "loss": 7.262149333953857,
+ "step": 3028
+ },
+ {
+ "epoch": 42.069868995633186,
+ "grad_norm": 0.03713499382138252,
+ "learning_rate": 0.0006,
+ "loss": 7.27695369720459,
+ "step": 3029
+ },
+ {
+ "epoch": 42.083842794759825,
+ "grad_norm": 0.014949247241020203,
+ "learning_rate": 0.0006,
+ "loss": 7.296508312225342,
+ "step": 3030
+ },
+ {
+ "epoch": 42.09781659388646,
+ "grad_norm": 0.02703995071351528,
+ "learning_rate": 0.0006,
+ "loss": 7.258018493652344,
+ "step": 3031
+ },
+ {
+ "epoch": 42.1117903930131,
+ "grad_norm": 0.020352816209197044,
+ "learning_rate": 0.0006,
+ "loss": 7.241450309753418,
+ "step": 3032
+ },
+ {
+ "epoch": 42.12576419213974,
+ "grad_norm": 0.014578479342162609,
+ "learning_rate": 0.0006,
+ "loss": 7.257323265075684,
+ "step": 3033
+ },
+ {
+ "epoch": 42.13973799126637,
+ "grad_norm": 0.026053044945001602,
+ "learning_rate": 0.0006,
+ "loss": 7.254680633544922,
+ "step": 3034
+ },
+ {
+ "epoch": 42.15371179039301,
+ "grad_norm": 0.015120275318622589,
+ "learning_rate": 0.0006,
+ "loss": 7.261074066162109,
+ "step": 3035
+ },
+ {
+ "epoch": 42.16768558951965,
+ "grad_norm": 0.016822177916765213,
+ "learning_rate": 0.0006,
+ "loss": 7.279064178466797,
+ "step": 3036
+ },
+ {
+ "epoch": 42.18165938864629,
+ "grad_norm": 0.016367588192224503,
+ "learning_rate": 0.0006,
+ "loss": 7.251429557800293,
+ "step": 3037
+ },
+ {
+ "epoch": 42.19563318777293,
+ "grad_norm": 0.014350418001413345,
+ "learning_rate": 0.0006,
+ "loss": 7.259264945983887,
+ "step": 3038
+ },
+ {
+ "epoch": 42.209606986899566,
+ "grad_norm": 0.017078228294849396,
+ "learning_rate": 0.0006,
+ "loss": 7.227105140686035,
+ "step": 3039
+ },
+ {
+ "epoch": 42.223580786026204,
+ "grad_norm": 0.011759317480027676,
+ "learning_rate": 0.0006,
+ "loss": 7.275888442993164,
+ "step": 3040
+ },
+ {
+ "epoch": 42.237554585152836,
+ "grad_norm": 0.013019340112805367,
+ "learning_rate": 0.0006,
+ "loss": 7.2517008781433105,
+ "step": 3041
+ },
+ {
+ "epoch": 42.251528384279474,
+ "grad_norm": 0.01178441196680069,
+ "learning_rate": 0.0006,
+ "loss": 7.23576545715332,
+ "step": 3042
+ },
+ {
+ "epoch": 42.26550218340611,
+ "grad_norm": 0.014518891461193562,
+ "learning_rate": 0.0006,
+ "loss": 7.253895282745361,
+ "step": 3043
+ },
+ {
+ "epoch": 42.27947598253275,
+ "grad_norm": 0.018159950152039528,
+ "learning_rate": 0.0006,
+ "loss": 7.278058052062988,
+ "step": 3044
+ },
+ {
+ "epoch": 42.29344978165939,
+ "grad_norm": 0.007587796077132225,
+ "learning_rate": 0.0006,
+ "loss": 7.278578758239746,
+ "step": 3045
+ },
+ {
+ "epoch": 42.30742358078603,
+ "grad_norm": 0.015994979068636894,
+ "learning_rate": 0.0006,
+ "loss": 7.239841461181641,
+ "step": 3046
+ },
+ {
+ "epoch": 42.32139737991266,
+ "grad_norm": 0.017830168828368187,
+ "learning_rate": 0.0006,
+ "loss": 7.2443976402282715,
+ "step": 3047
+ },
+ {
+ "epoch": 42.3353711790393,
+ "grad_norm": 0.008738759905099869,
+ "learning_rate": 0.0006,
+ "loss": 7.232264518737793,
+ "step": 3048
+ },
+ {
+ "epoch": 42.34934497816594,
+ "grad_norm": 0.0175449438393116,
+ "learning_rate": 0.0006,
+ "loss": 7.228000640869141,
+ "step": 3049
+ },
+ {
+ "epoch": 42.36331877729258,
+ "grad_norm": 0.019611820578575134,
+ "learning_rate": 0.0006,
+ "loss": 7.234999179840088,
+ "step": 3050
+ },
+ {
+ "epoch": 42.377292576419215,
+ "grad_norm": 0.015302863903343678,
+ "learning_rate": 0.0006,
+ "loss": 7.242547988891602,
+ "step": 3051
+ },
+ {
+ "epoch": 42.391266375545854,
+ "grad_norm": 0.025919731706380844,
+ "learning_rate": 0.0006,
+ "loss": 7.216652870178223,
+ "step": 3052
+ },
+ {
+ "epoch": 42.40524017467249,
+ "grad_norm": 0.034587468951940536,
+ "learning_rate": 0.0006,
+ "loss": 7.245336532592773,
+ "step": 3053
+ },
+ {
+ "epoch": 42.419213973799124,
+ "grad_norm": 0.019198840484023094,
+ "learning_rate": 0.0006,
+ "loss": 7.256758689880371,
+ "step": 3054
+ },
+ {
+ "epoch": 42.43318777292576,
+ "grad_norm": 0.012316389009356499,
+ "learning_rate": 0.0006,
+ "loss": 7.209102630615234,
+ "step": 3055
+ },
+ {
+ "epoch": 42.4471615720524,
+ "grad_norm": 0.023512162268161774,
+ "learning_rate": 0.0006,
+ "loss": 7.250742435455322,
+ "step": 3056
+ },
+ {
+ "epoch": 42.46113537117904,
+ "grad_norm": 0.025457851588726044,
+ "learning_rate": 0.0006,
+ "loss": 7.263757705688477,
+ "step": 3057
+ },
+ {
+ "epoch": 42.47510917030568,
+ "grad_norm": 0.02192617766559124,
+ "learning_rate": 0.0006,
+ "loss": 7.246977806091309,
+ "step": 3058
+ },
+ {
+ "epoch": 42.48908296943232,
+ "grad_norm": 0.02541281282901764,
+ "learning_rate": 0.0006,
+ "loss": 7.232361316680908,
+ "step": 3059
+ },
+ {
+ "epoch": 42.50305676855895,
+ "grad_norm": 0.036353763192892075,
+ "learning_rate": 0.0006,
+ "loss": 7.242556571960449,
+ "step": 3060
+ },
+ {
+ "epoch": 42.51703056768559,
+ "grad_norm": 0.04555872827768326,
+ "learning_rate": 0.0006,
+ "loss": 7.228940963745117,
+ "step": 3061
+ },
+ {
+ "epoch": 42.531004366812226,
+ "grad_norm": 0.04563378915190697,
+ "learning_rate": 0.0006,
+ "loss": 7.235276222229004,
+ "step": 3062
+ },
+ {
+ "epoch": 42.544978165938865,
+ "grad_norm": 0.04275760427117348,
+ "learning_rate": 0.0006,
+ "loss": 7.238895416259766,
+ "step": 3063
+ },
+ {
+ "epoch": 42.5589519650655,
+ "grad_norm": 0.04535885155200958,
+ "learning_rate": 0.0006,
+ "loss": 7.205336570739746,
+ "step": 3064
+ },
+ {
+ "epoch": 42.57292576419214,
+ "grad_norm": 0.015534617006778717,
+ "learning_rate": 0.0006,
+ "loss": 7.238393783569336,
+ "step": 3065
+ },
+ {
+ "epoch": 42.58689956331878,
+ "grad_norm": 0.03286299481987953,
+ "learning_rate": 0.0006,
+ "loss": 7.216453552246094,
+ "step": 3066
+ },
+ {
+ "epoch": 42.60087336244541,
+ "grad_norm": 0.025453168898820877,
+ "learning_rate": 0.0006,
+ "loss": 7.203750133514404,
+ "step": 3067
+ },
+ {
+ "epoch": 42.61484716157205,
+ "grad_norm": 0.021501649171113968,
+ "learning_rate": 0.0006,
+ "loss": 7.241621971130371,
+ "step": 3068
+ },
+ {
+ "epoch": 42.62882096069869,
+ "grad_norm": 0.02743634767830372,
+ "learning_rate": 0.0006,
+ "loss": 7.22076940536499,
+ "step": 3069
+ },
+ {
+ "epoch": 42.64279475982533,
+ "grad_norm": 0.020890483632683754,
+ "learning_rate": 0.0006,
+ "loss": 7.218412399291992,
+ "step": 3070
+ },
+ {
+ "epoch": 42.65676855895197,
+ "grad_norm": 0.022499850019812584,
+ "learning_rate": 0.0006,
+ "loss": 7.217696189880371,
+ "step": 3071
+ },
+ {
+ "epoch": 42.670742358078606,
+ "grad_norm": 0.01653696969151497,
+ "learning_rate": 0.0006,
+ "loss": 7.156417369842529,
+ "step": 3072
+ },
+ {
+ "epoch": 42.68471615720524,
+ "grad_norm": 0.01834225095808506,
+ "learning_rate": 0.0006,
+ "loss": 7.211247444152832,
+ "step": 3073
+ },
+ {
+ "epoch": 42.698689956331876,
+ "grad_norm": 0.020137229934334755,
+ "learning_rate": 0.0006,
+ "loss": 7.232499599456787,
+ "step": 3074
+ },
+ {
+ "epoch": 42.712663755458514,
+ "grad_norm": 0.012671459466218948,
+ "learning_rate": 0.0006,
+ "loss": 7.172441482543945,
+ "step": 3075
+ },
+ {
+ "epoch": 42.72663755458515,
+ "grad_norm": 0.016803130507469177,
+ "learning_rate": 0.0006,
+ "loss": 7.152103424072266,
+ "step": 3076
+ },
+ {
+ "epoch": 42.74061135371179,
+ "grad_norm": 0.01623712107539177,
+ "learning_rate": 0.0006,
+ "loss": 7.187032699584961,
+ "step": 3077
+ },
+ {
+ "epoch": 42.75458515283843,
+ "grad_norm": 0.015938356518745422,
+ "learning_rate": 0.0006,
+ "loss": 7.171140670776367,
+ "step": 3078
+ },
+ {
+ "epoch": 42.76855895196506,
+ "grad_norm": 0.02580736204981804,
+ "learning_rate": 0.0006,
+ "loss": 7.204472541809082,
+ "step": 3079
+ },
+ {
+ "epoch": 42.7825327510917,
+ "grad_norm": 0.023372117429971695,
+ "learning_rate": 0.0006,
+ "loss": 7.165554523468018,
+ "step": 3080
+ },
+ {
+ "epoch": 42.79650655021834,
+ "grad_norm": 0.04440586268901825,
+ "learning_rate": 0.0006,
+ "loss": 7.169431209564209,
+ "step": 3081
+ },
+ {
+ "epoch": 42.81048034934498,
+ "grad_norm": 0.06797406077384949,
+ "learning_rate": 0.0006,
+ "loss": 7.168193340301514,
+ "step": 3082
+ },
+ {
+ "epoch": 42.82445414847162,
+ "grad_norm": 0.0914321169257164,
+ "learning_rate": 0.0006,
+ "loss": 7.204671859741211,
+ "step": 3083
+ },
+ {
+ "epoch": 42.838427947598255,
+ "grad_norm": 0.0506940595805645,
+ "learning_rate": 0.0006,
+ "loss": 7.15949821472168,
+ "step": 3084
+ },
+ {
+ "epoch": 42.852401746724894,
+ "grad_norm": 0.04819890484213829,
+ "learning_rate": 0.0006,
+ "loss": 7.192837238311768,
+ "step": 3085
+ },
+ {
+ "epoch": 42.866375545851525,
+ "grad_norm": 0.040044233202934265,
+ "learning_rate": 0.0006,
+ "loss": 7.1600165367126465,
+ "step": 3086
+ },
+ {
+ "epoch": 42.880349344978164,
+ "grad_norm": 0.02573299966752529,
+ "learning_rate": 0.0006,
+ "loss": 7.153427600860596,
+ "step": 3087
+ },
+ {
+ "epoch": 42.8943231441048,
+ "grad_norm": 0.046942535787820816,
+ "learning_rate": 0.0006,
+ "loss": 7.16767692565918,
+ "step": 3088
+ },
+ {
+ "epoch": 42.90829694323144,
+ "grad_norm": 0.02683161571621895,
+ "learning_rate": 0.0006,
+ "loss": 7.17144250869751,
+ "step": 3089
+ },
+ {
+ "epoch": 42.92227074235808,
+ "grad_norm": 0.05363674834370613,
+ "learning_rate": 0.0006,
+ "loss": 7.120128631591797,
+ "step": 3090
+ },
+ {
+ "epoch": 42.93624454148472,
+ "grad_norm": 0.023756247013807297,
+ "learning_rate": 0.0006,
+ "loss": 7.169703006744385,
+ "step": 3091
+ },
+ {
+ "epoch": 42.95021834061135,
+ "grad_norm": 0.04216783866286278,
+ "learning_rate": 0.0006,
+ "loss": 7.1206865310668945,
+ "step": 3092
+ },
+ {
+ "epoch": 42.96419213973799,
+ "grad_norm": 0.02219565585255623,
+ "learning_rate": 0.0006,
+ "loss": 7.157215118408203,
+ "step": 3093
+ },
+ {
+ "epoch": 42.97816593886463,
+ "grad_norm": 0.03527127206325531,
+ "learning_rate": 0.0006,
+ "loss": 7.166672706604004,
+ "step": 3094
+ },
+ {
+ "epoch": 42.992139737991266,
+ "grad_norm": 0.017748817801475525,
+ "learning_rate": 0.0006,
+ "loss": 7.154747009277344,
+ "step": 3095
+ },
+ {
+ "epoch": 43.0,
+ "grad_norm": 0.033165838569402695,
+ "learning_rate": 0.0006,
+ "loss": 7.150866508483887,
+ "step": 3096
+ },
+ {
+ "epoch": 43.0,
+ "eval_loss": 7.144403457641602,
+ "eval_runtime": 56.8853,
+ "eval_samples_per_second": 42.928,
+ "eval_steps_per_second": 1.354,
+ "step": 3096
+ },
+ {
+ "epoch": 43.01397379912664,
+ "grad_norm": 0.015199829824268818,
+ "learning_rate": 0.0006,
+ "loss": 7.122735023498535,
+ "step": 3097
+ },
+ {
+ "epoch": 43.02794759825328,
+ "grad_norm": 0.03220256045460701,
+ "learning_rate": 0.0006,
+ "loss": 7.143553733825684,
+ "step": 3098
+ },
+ {
+ "epoch": 43.041921397379916,
+ "grad_norm": 0.014687920920550823,
+ "learning_rate": 0.0006,
+ "loss": 7.080960750579834,
+ "step": 3099
+ },
+ {
+ "epoch": 43.05589519650655,
+ "grad_norm": 0.023522932082414627,
+ "learning_rate": 0.0006,
+ "loss": 7.137356758117676,
+ "step": 3100
+ },
+ {
+ "epoch": 43.069868995633186,
+ "grad_norm": 0.014808216132223606,
+ "learning_rate": 0.0006,
+ "loss": 7.09143590927124,
+ "step": 3101
+ },
+ {
+ "epoch": 43.083842794759825,
+ "grad_norm": 0.029339246451854706,
+ "learning_rate": 0.0006,
+ "loss": 7.122198104858398,
+ "step": 3102
+ },
+ {
+ "epoch": 43.09781659388646,
+ "grad_norm": 0.022953078150749207,
+ "learning_rate": 0.0006,
+ "loss": 7.092505931854248,
+ "step": 3103
+ },
+ {
+ "epoch": 43.1117903930131,
+ "grad_norm": 0.04590052738785744,
+ "learning_rate": 0.0006,
+ "loss": 7.113100051879883,
+ "step": 3104
+ },
+ {
+ "epoch": 43.12576419213974,
+ "grad_norm": 0.0524442233145237,
+ "learning_rate": 0.0006,
+ "loss": 7.147022247314453,
+ "step": 3105
+ },
+ {
+ "epoch": 43.13973799126637,
+ "grad_norm": 0.0315794013440609,
+ "learning_rate": 0.0006,
+ "loss": 7.094045162200928,
+ "step": 3106
+ },
+ {
+ "epoch": 43.15371179039301,
+ "grad_norm": 0.05112722888588905,
+ "learning_rate": 0.0006,
+ "loss": 7.054388523101807,
+ "step": 3107
+ },
+ {
+ "epoch": 43.16768558951965,
+ "grad_norm": 0.03146776929497719,
+ "learning_rate": 0.0006,
+ "loss": 7.116674423217773,
+ "step": 3108
+ },
+ {
+ "epoch": 43.18165938864629,
+ "grad_norm": 0.03391246497631073,
+ "learning_rate": 0.0006,
+ "loss": 7.150345802307129,
+ "step": 3109
+ },
+ {
+ "epoch": 43.19563318777293,
+ "grad_norm": 0.026957903057336807,
+ "learning_rate": 0.0006,
+ "loss": 7.058531284332275,
+ "step": 3110
+ },
+ {
+ "epoch": 43.209606986899566,
+ "grad_norm": 0.041532568633556366,
+ "learning_rate": 0.0006,
+ "loss": 7.085498809814453,
+ "step": 3111
+ },
+ {
+ "epoch": 43.223580786026204,
+ "grad_norm": 0.020001588389277458,
+ "learning_rate": 0.0006,
+ "loss": 7.131487846374512,
+ "step": 3112
+ },
+ {
+ "epoch": 43.237554585152836,
+ "grad_norm": 0.02268712781369686,
+ "learning_rate": 0.0006,
+ "loss": 7.0790863037109375,
+ "step": 3113
+ },
+ {
+ "epoch": 43.251528384279474,
+ "grad_norm": 0.023363925516605377,
+ "learning_rate": 0.0006,
+ "loss": 7.067816734313965,
+ "step": 3114
+ },
+ {
+ "epoch": 43.26550218340611,
+ "grad_norm": 0.019367508590221405,
+ "learning_rate": 0.0006,
+ "loss": 7.042041778564453,
+ "step": 3115
+ },
+ {
+ "epoch": 43.27947598253275,
+ "grad_norm": 0.02745090238749981,
+ "learning_rate": 0.0006,
+ "loss": 7.064733028411865,
+ "step": 3116
+ },
+ {
+ "epoch": 43.29344978165939,
+ "grad_norm": 0.02543802745640278,
+ "learning_rate": 0.0006,
+ "loss": 7.058948040008545,
+ "step": 3117
+ },
+ {
+ "epoch": 43.30742358078603,
+ "grad_norm": 0.018874092027544975,
+ "learning_rate": 0.0006,
+ "loss": 7.009946346282959,
+ "step": 3118
+ },
+ {
+ "epoch": 43.32139737991266,
+ "grad_norm": 0.024662822484970093,
+ "learning_rate": 0.0006,
+ "loss": 7.015913963317871,
+ "step": 3119
+ },
+ {
+ "epoch": 43.3353711790393,
+ "grad_norm": 0.02441416308283806,
+ "learning_rate": 0.0006,
+ "loss": 7.061622142791748,
+ "step": 3120
+ },
+ {
+ "epoch": 43.34934497816594,
+ "grad_norm": 0.031887322664260864,
+ "learning_rate": 0.0006,
+ "loss": 7.046883583068848,
+ "step": 3121
+ },
+ {
+ "epoch": 43.36331877729258,
+ "grad_norm": 0.021930739283561707,
+ "learning_rate": 0.0006,
+ "loss": 7.0309624671936035,
+ "step": 3122
+ },
+ {
+ "epoch": 43.377292576419215,
+ "grad_norm": 0.02722783200442791,
+ "learning_rate": 0.0006,
+ "loss": 7.072924613952637,
+ "step": 3123
+ },
+ {
+ "epoch": 43.391266375545854,
+ "grad_norm": 0.01483179721981287,
+ "learning_rate": 0.0006,
+ "loss": 7.028624534606934,
+ "step": 3124
+ },
+ {
+ "epoch": 43.40524017467249,
+ "grad_norm": 0.024793945252895355,
+ "learning_rate": 0.0006,
+ "loss": 7.001623153686523,
+ "step": 3125
+ },
+ {
+ "epoch": 43.419213973799124,
+ "grad_norm": 0.020148713141679764,
+ "learning_rate": 0.0006,
+ "loss": 6.9938812255859375,
+ "step": 3126
+ },
+ {
+ "epoch": 43.43318777292576,
+ "grad_norm": 0.02413850650191307,
+ "learning_rate": 0.0006,
+ "loss": 7.001243591308594,
+ "step": 3127
+ },
+ {
+ "epoch": 43.4471615720524,
+ "grad_norm": 0.028549203649163246,
+ "learning_rate": 0.0006,
+ "loss": 6.987671375274658,
+ "step": 3128
+ },
+ {
+ "epoch": 43.46113537117904,
+ "grad_norm": 0.025988785549998283,
+ "learning_rate": 0.0006,
+ "loss": 6.9551167488098145,
+ "step": 3129
+ },
+ {
+ "epoch": 43.47510917030568,
+ "grad_norm": 0.027140729129314423,
+ "learning_rate": 0.0006,
+ "loss": 6.9547810554504395,
+ "step": 3130
+ },
+ {
+ "epoch": 43.48908296943232,
+ "grad_norm": 0.030130082741379738,
+ "learning_rate": 0.0006,
+ "loss": 6.993186950683594,
+ "step": 3131
+ },
+ {
+ "epoch": 43.50305676855895,
+ "grad_norm": 0.025996994227170944,
+ "learning_rate": 0.0006,
+ "loss": 6.984624862670898,
+ "step": 3132
+ },
+ {
+ "epoch": 43.51703056768559,
+ "grad_norm": 0.02423008903861046,
+ "learning_rate": 0.0006,
+ "loss": 6.98813009262085,
+ "step": 3133
+ },
+ {
+ "epoch": 43.531004366812226,
+ "grad_norm": 0.029665837064385414,
+ "learning_rate": 0.0006,
+ "loss": 6.986610412597656,
+ "step": 3134
+ },
+ {
+ "epoch": 43.544978165938865,
+ "grad_norm": 0.028768595308065414,
+ "learning_rate": 0.0006,
+ "loss": 6.898797035217285,
+ "step": 3135
+ },
+ {
+ "epoch": 43.5589519650655,
+ "grad_norm": 0.03111208975315094,
+ "learning_rate": 0.0006,
+ "loss": 6.923741817474365,
+ "step": 3136
+ },
+ {
+ "epoch": 43.57292576419214,
+ "grad_norm": 0.0368187390267849,
+ "learning_rate": 0.0006,
+ "loss": 6.994343280792236,
+ "step": 3137
+ },
+ {
+ "epoch": 43.58689956331878,
+ "grad_norm": 0.04076763987541199,
+ "learning_rate": 0.0006,
+ "loss": 6.909458160400391,
+ "step": 3138
+ },
+ {
+ "epoch": 43.60087336244541,
+ "grad_norm": 0.03281061723828316,
+ "learning_rate": 0.0006,
+ "loss": 6.924851417541504,
+ "step": 3139
+ },
+ {
+ "epoch": 43.61484716157205,
+ "grad_norm": 0.025486988946795464,
+ "learning_rate": 0.0006,
+ "loss": 6.962926864624023,
+ "step": 3140
+ },
+ {
+ "epoch": 43.62882096069869,
+ "grad_norm": 0.04583485797047615,
+ "learning_rate": 0.0006,
+ "loss": 6.876555442810059,
+ "step": 3141
+ },
+ {
+ "epoch": 43.64279475982533,
+ "grad_norm": 0.05460785701870918,
+ "learning_rate": 0.0006,
+ "loss": 6.952932357788086,
+ "step": 3142
+ },
+ {
+ "epoch": 43.65676855895197,
+ "grad_norm": 0.040917906910181046,
+ "learning_rate": 0.0006,
+ "loss": 6.9624924659729,
+ "step": 3143
+ },
+ {
+ "epoch": 43.670742358078606,
+ "grad_norm": 0.04009668901562691,
+ "learning_rate": 0.0006,
+ "loss": 6.94033145904541,
+ "step": 3144
+ },
+ {
+ "epoch": 43.68471615720524,
+ "grad_norm": 0.04500625655055046,
+ "learning_rate": 0.0006,
+ "loss": 6.961839199066162,
+ "step": 3145
+ },
+ {
+ "epoch": 43.698689956331876,
+ "grad_norm": 0.03182258829474449,
+ "learning_rate": 0.0006,
+ "loss": 6.930580139160156,
+ "step": 3146
+ },
+ {
+ "epoch": 43.712663755458514,
+ "grad_norm": 0.053621504455804825,
+ "learning_rate": 0.0006,
+ "loss": 6.950235366821289,
+ "step": 3147
+ },
+ {
+ "epoch": 43.72663755458515,
+ "grad_norm": 0.031919289380311966,
+ "learning_rate": 0.0006,
+ "loss": 6.895111560821533,
+ "step": 3148
+ },
+ {
+ "epoch": 43.74061135371179,
+ "grad_norm": 0.042713891714811325,
+ "learning_rate": 0.0006,
+ "loss": 6.940885543823242,
+ "step": 3149
+ },
+ {
+ "epoch": 43.75458515283843,
+ "grad_norm": 0.02385653182864189,
+ "learning_rate": 0.0006,
+ "loss": 6.948801517486572,
+ "step": 3150
+ },
+ {
+ "epoch": 43.76855895196506,
+ "grad_norm": 0.04100663587450981,
+ "learning_rate": 0.0006,
+ "loss": 6.970411777496338,
+ "step": 3151
+ },
+ {
+ "epoch": 43.7825327510917,
+ "grad_norm": 0.02729583904147148,
+ "learning_rate": 0.0006,
+ "loss": 6.899379253387451,
+ "step": 3152
+ },
+ {
+ "epoch": 43.79650655021834,
+ "grad_norm": 0.03128620237112045,
+ "learning_rate": 0.0006,
+ "loss": 6.931319236755371,
+ "step": 3153
+ },
+ {
+ "epoch": 43.81048034934498,
+ "grad_norm": 0.025488659739494324,
+ "learning_rate": 0.0006,
+ "loss": 6.8486409187316895,
+ "step": 3154
+ },
+ {
+ "epoch": 43.82445414847162,
+ "grad_norm": 0.03679376840591431,
+ "learning_rate": 0.0006,
+ "loss": 6.840811252593994,
+ "step": 3155
+ },
+ {
+ "epoch": 43.838427947598255,
+ "grad_norm": 0.020578034222126007,
+ "learning_rate": 0.0006,
+ "loss": 6.9088616371154785,
+ "step": 3156
+ },
+ {
+ "epoch": 43.852401746724894,
+ "grad_norm": 0.02935841865837574,
+ "learning_rate": 0.0006,
+ "loss": 6.806676864624023,
+ "step": 3157
+ },
+ {
+ "epoch": 43.866375545851525,
+ "grad_norm": 0.023319244384765625,
+ "learning_rate": 0.0006,
+ "loss": 6.878457546234131,
+ "step": 3158
+ },
+ {
+ "epoch": 43.880349344978164,
+ "grad_norm": 0.02779589593410492,
+ "learning_rate": 0.0006,
+ "loss": 6.822447776794434,
+ "step": 3159
+ },
+ {
+ "epoch": 43.8943231441048,
+ "grad_norm": 0.01879825070500374,
+ "learning_rate": 0.0006,
+ "loss": 6.854121208190918,
+ "step": 3160
+ },
+ {
+ "epoch": 43.90829694323144,
+ "grad_norm": 0.023286139592528343,
+ "learning_rate": 0.0006,
+ "loss": 6.836827278137207,
+ "step": 3161
+ },
+ {
+ "epoch": 43.92227074235808,
+ "grad_norm": 0.030358297750353813,
+ "learning_rate": 0.0006,
+ "loss": 6.860037803649902,
+ "step": 3162
+ },
+ {
+ "epoch": 43.93624454148472,
+ "grad_norm": 0.029621105641126633,
+ "learning_rate": 0.0006,
+ "loss": 6.884025573730469,
+ "step": 3163
+ },
+ {
+ "epoch": 43.95021834061135,
+ "grad_norm": 0.029829824343323708,
+ "learning_rate": 0.0006,
+ "loss": 6.8702392578125,
+ "step": 3164
+ },
+ {
+ "epoch": 43.96419213973799,
+ "grad_norm": 0.036892473697662354,
+ "learning_rate": 0.0006,
+ "loss": 6.806652069091797,
+ "step": 3165
+ },
+ {
+ "epoch": 43.97816593886463,
+ "grad_norm": 0.03206709027290344,
+ "learning_rate": 0.0006,
+ "loss": 6.828495979309082,
+ "step": 3166
+ },
+ {
+ "epoch": 43.992139737991266,
+ "grad_norm": 0.02744000405073166,
+ "learning_rate": 0.0006,
+ "loss": 6.832002639770508,
+ "step": 3167
+ },
+ {
+ "epoch": 44.0,
+ "grad_norm": 0.02205660194158554,
+ "learning_rate": 0.0006,
+ "loss": 6.898828506469727,
+ "step": 3168
+ },
+ {
+ "epoch": 44.0,
+ "eval_loss": 6.8150553703308105,
+ "eval_runtime": 56.784,
+ "eval_samples_per_second": 43.005,
+ "eval_steps_per_second": 1.356,
+ "step": 3168
+ },
+ {
+ "epoch": 44.01397379912664,
+ "grad_norm": 0.030421260744333267,
+ "learning_rate": 0.0006,
+ "loss": 6.775057792663574,
+ "step": 3169
+ },
+ {
+ "epoch": 44.02794759825328,
+ "grad_norm": 0.0465611033141613,
+ "learning_rate": 0.0006,
+ "loss": 6.849390506744385,
+ "step": 3170
+ },
+ {
+ "epoch": 44.041921397379916,
+ "grad_norm": 0.05977972596883774,
+ "learning_rate": 0.0006,
+ "loss": 6.725576400756836,
+ "step": 3171
+ },
+ {
+ "epoch": 44.05589519650655,
+ "grad_norm": 0.05432697385549545,
+ "learning_rate": 0.0006,
+ "loss": 6.768024444580078,
+ "step": 3172
+ },
+ {
+ "epoch": 44.069868995633186,
+ "grad_norm": 0.04850967228412628,
+ "learning_rate": 0.0006,
+ "loss": 6.799457550048828,
+ "step": 3173
+ },
+ {
+ "epoch": 44.083842794759825,
+ "grad_norm": 0.06201786920428276,
+ "learning_rate": 0.0006,
+ "loss": 6.897234916687012,
+ "step": 3174
+ },
+ {
+ "epoch": 44.09781659388646,
+ "grad_norm": 0.04695591703057289,
+ "learning_rate": 0.0006,
+ "loss": 6.826470375061035,
+ "step": 3175
+ },
+ {
+ "epoch": 44.1117903930131,
+ "grad_norm": 0.03850164636969566,
+ "learning_rate": 0.0006,
+ "loss": 6.835901260375977,
+ "step": 3176
+ },
+ {
+ "epoch": 44.12576419213974,
+ "grad_norm": 0.04250155761837959,
+ "learning_rate": 0.0006,
+ "loss": 6.782310485839844,
+ "step": 3177
+ },
+ {
+ "epoch": 44.13973799126637,
+ "grad_norm": 0.041465479880571365,
+ "learning_rate": 0.0006,
+ "loss": 6.785732746124268,
+ "step": 3178
+ },
+ {
+ "epoch": 44.15371179039301,
+ "grad_norm": 0.02527807280421257,
+ "learning_rate": 0.0006,
+ "loss": 6.78061580657959,
+ "step": 3179
+ },
+ {
+ "epoch": 44.16768558951965,
+ "grad_norm": 0.03660374879837036,
+ "learning_rate": 0.0006,
+ "loss": 6.783992767333984,
+ "step": 3180
+ },
+ {
+ "epoch": 44.18165938864629,
+ "grad_norm": 0.027567964047193527,
+ "learning_rate": 0.0006,
+ "loss": 6.765413284301758,
+ "step": 3181
+ },
+ {
+ "epoch": 44.19563318777293,
+ "grad_norm": 0.024301424622535706,
+ "learning_rate": 0.0006,
+ "loss": 6.760368347167969,
+ "step": 3182
+ },
+ {
+ "epoch": 44.209606986899566,
+ "grad_norm": 0.0266821701079607,
+ "learning_rate": 0.0006,
+ "loss": 6.71417236328125,
+ "step": 3183
+ },
+ {
+ "epoch": 44.223580786026204,
+ "grad_norm": 0.028125090524554253,
+ "learning_rate": 0.0006,
+ "loss": 6.646126747131348,
+ "step": 3184
+ },
+ {
+ "epoch": 44.237554585152836,
+ "grad_norm": 0.0185737032443285,
+ "learning_rate": 0.0006,
+ "loss": 6.712944030761719,
+ "step": 3185
+ },
+ {
+ "epoch": 44.251528384279474,
+ "grad_norm": 0.025836674496531487,
+ "learning_rate": 0.0006,
+ "loss": 6.707030296325684,
+ "step": 3186
+ },
+ {
+ "epoch": 44.26550218340611,
+ "grad_norm": 0.029744870960712433,
+ "learning_rate": 0.0006,
+ "loss": 6.6271071434021,
+ "step": 3187
+ },
+ {
+ "epoch": 44.27947598253275,
+ "grad_norm": 0.03021426685154438,
+ "learning_rate": 0.0006,
+ "loss": 6.677826404571533,
+ "step": 3188
+ },
+ {
+ "epoch": 44.29344978165939,
+ "grad_norm": 0.032727014273405075,
+ "learning_rate": 0.0006,
+ "loss": 6.657764434814453,
+ "step": 3189
+ },
+ {
+ "epoch": 44.30742358078603,
+ "grad_norm": 0.03696778789162636,
+ "learning_rate": 0.0006,
+ "loss": 6.632883548736572,
+ "step": 3190
+ },
+ {
+ "epoch": 44.32139737991266,
+ "grad_norm": 0.034630827605724335,
+ "learning_rate": 0.0006,
+ "loss": 6.759583473205566,
+ "step": 3191
+ },
+ {
+ "epoch": 44.3353711790393,
+ "grad_norm": 0.030517758801579475,
+ "learning_rate": 0.0006,
+ "loss": 6.702268600463867,
+ "step": 3192
+ },
+ {
+ "epoch": 44.34934497816594,
+ "grad_norm": 0.04105190932750702,
+ "learning_rate": 0.0006,
+ "loss": 6.705792427062988,
+ "step": 3193
+ },
+ {
+ "epoch": 44.36331877729258,
+ "grad_norm": 0.06659174710512161,
+ "learning_rate": 0.0006,
+ "loss": 6.73862361907959,
+ "step": 3194
+ },
+ {
+ "epoch": 44.377292576419215,
+ "grad_norm": 0.08013619482517242,
+ "learning_rate": 0.0006,
+ "loss": 6.710282802581787,
+ "step": 3195
+ },
+ {
+ "epoch": 44.391266375545854,
+ "grad_norm": 0.049850963056087494,
+ "learning_rate": 0.0006,
+ "loss": 6.684974670410156,
+ "step": 3196
+ },
+ {
+ "epoch": 44.40524017467249,
+ "grad_norm": 0.061247408390045166,
+ "learning_rate": 0.0006,
+ "loss": 6.716034412384033,
+ "step": 3197
+ },
+ {
+ "epoch": 44.419213973799124,
+ "grad_norm": 0.05868857726454735,
+ "learning_rate": 0.0006,
+ "loss": 6.589107513427734,
+ "step": 3198
+ },
+ {
+ "epoch": 44.43318777292576,
+ "grad_norm": 0.048878561705350876,
+ "learning_rate": 0.0006,
+ "loss": 6.694855690002441,
+ "step": 3199
+ },
+ {
+ "epoch": 44.4471615720524,
+ "grad_norm": 0.028748324140906334,
+ "learning_rate": 0.0006,
+ "loss": 6.629969596862793,
+ "step": 3200
+ },
+ {
+ "epoch": 44.46113537117904,
+ "grad_norm": 0.04060306400060654,
+ "learning_rate": 0.0006,
+ "loss": 6.721431255340576,
+ "step": 3201
+ },
+ {
+ "epoch": 44.47510917030568,
+ "grad_norm": 0.03833824768662453,
+ "learning_rate": 0.0006,
+ "loss": 6.625029563903809,
+ "step": 3202
+ },
+ {
+ "epoch": 44.48908296943232,
+ "grad_norm": 0.03791997581720352,
+ "learning_rate": 0.0006,
+ "loss": 6.669187545776367,
+ "step": 3203
+ },
+ {
+ "epoch": 44.50305676855895,
+ "grad_norm": 0.03259684890508652,
+ "learning_rate": 0.0006,
+ "loss": 6.665643692016602,
+ "step": 3204
+ },
+ {
+ "epoch": 44.51703056768559,
+ "grad_norm": 0.040658626705408096,
+ "learning_rate": 0.0006,
+ "loss": 6.639796257019043,
+ "step": 3205
+ },
+ {
+ "epoch": 44.531004366812226,
+ "grad_norm": 0.03691193461418152,
+ "learning_rate": 0.0006,
+ "loss": 6.692244529724121,
+ "step": 3206
+ },
+ {
+ "epoch": 44.544978165938865,
+ "grad_norm": 0.02814812771975994,
+ "learning_rate": 0.0006,
+ "loss": 6.61978006362915,
+ "step": 3207
+ },
+ {
+ "epoch": 44.5589519650655,
+ "grad_norm": 0.022155074402689934,
+ "learning_rate": 0.0006,
+ "loss": 6.625905990600586,
+ "step": 3208
+ },
+ {
+ "epoch": 44.57292576419214,
+ "grad_norm": 0.030794885009527206,
+ "learning_rate": 0.0006,
+ "loss": 6.678500652313232,
+ "step": 3209
+ },
+ {
+ "epoch": 44.58689956331878,
+ "grad_norm": 0.03201768547296524,
+ "learning_rate": 0.0006,
+ "loss": 6.6015729904174805,
+ "step": 3210
+ },
+ {
+ "epoch": 44.60087336244541,
+ "grad_norm": 0.025138361379504204,
+ "learning_rate": 0.0006,
+ "loss": 6.568234443664551,
+ "step": 3211
+ },
+ {
+ "epoch": 44.61484716157205,
+ "grad_norm": 0.02969353087246418,
+ "learning_rate": 0.0006,
+ "loss": 6.588366508483887,
+ "step": 3212
+ },
+ {
+ "epoch": 44.62882096069869,
+ "grad_norm": 0.03136962279677391,
+ "learning_rate": 0.0006,
+ "loss": 6.479961395263672,
+ "step": 3213
+ },
+ {
+ "epoch": 44.64279475982533,
+ "grad_norm": 0.030178170651197433,
+ "learning_rate": 0.0006,
+ "loss": 6.622910976409912,
+ "step": 3214
+ },
+ {
+ "epoch": 44.65676855895197,
+ "grad_norm": 0.02682245895266533,
+ "learning_rate": 0.0006,
+ "loss": 6.597707748413086,
+ "step": 3215
+ },
+ {
+ "epoch": 44.670742358078606,
+ "grad_norm": 0.025154557079076767,
+ "learning_rate": 0.0006,
+ "loss": 6.630541801452637,
+ "step": 3216
+ },
+ {
+ "epoch": 44.68471615720524,
+ "grad_norm": 0.03564463183283806,
+ "learning_rate": 0.0006,
+ "loss": 6.542823791503906,
+ "step": 3217
+ },
+ {
+ "epoch": 44.698689956331876,
+ "grad_norm": 0.049477651715278625,
+ "learning_rate": 0.0006,
+ "loss": 6.646030426025391,
+ "step": 3218
+ },
+ {
+ "epoch": 44.712663755458514,
+ "grad_norm": 0.06536535918712616,
+ "learning_rate": 0.0006,
+ "loss": 6.595463275909424,
+ "step": 3219
+ },
+ {
+ "epoch": 44.72663755458515,
+ "grad_norm": 0.07710316777229309,
+ "learning_rate": 0.0006,
+ "loss": 6.55524206161499,
+ "step": 3220
+ },
+ {
+ "epoch": 44.74061135371179,
+ "grad_norm": 0.09571163356304169,
+ "learning_rate": 0.0006,
+ "loss": 6.659764289855957,
+ "step": 3221
+ },
+ {
+ "epoch": 44.75458515283843,
+ "grad_norm": 0.055076830089092255,
+ "learning_rate": 0.0006,
+ "loss": 6.658882141113281,
+ "step": 3222
+ },
+ {
+ "epoch": 44.76855895196506,
+ "grad_norm": 0.07671099156141281,
+ "learning_rate": 0.0006,
+ "loss": 6.605809211730957,
+ "step": 3223
+ },
+ {
+ "epoch": 44.7825327510917,
+ "grad_norm": 0.05295584350824356,
+ "learning_rate": 0.0006,
+ "loss": 6.641455173492432,
+ "step": 3224
+ },
+ {
+ "epoch": 44.79650655021834,
+ "grad_norm": 0.09128042310476303,
+ "learning_rate": 0.0006,
+ "loss": 6.691066741943359,
+ "step": 3225
+ },
+ {
+ "epoch": 44.81048034934498,
+ "grad_norm": 0.05362958461046219,
+ "learning_rate": 0.0006,
+ "loss": 6.573406219482422,
+ "step": 3226
+ },
+ {
+ "epoch": 44.82445414847162,
+ "grad_norm": 0.046910062432289124,
+ "learning_rate": 0.0006,
+ "loss": 6.550876617431641,
+ "step": 3227
+ },
+ {
+ "epoch": 44.838427947598255,
+ "grad_norm": 0.03494197130203247,
+ "learning_rate": 0.0006,
+ "loss": 6.540179252624512,
+ "step": 3228
+ },
+ {
+ "epoch": 44.852401746724894,
+ "grad_norm": 0.04716387391090393,
+ "learning_rate": 0.0006,
+ "loss": 6.490459442138672,
+ "step": 3229
+ },
+ {
+ "epoch": 44.866375545851525,
+ "grad_norm": 0.04190506786108017,
+ "learning_rate": 0.0006,
+ "loss": 6.509925365447998,
+ "step": 3230
+ },
+ {
+ "epoch": 44.880349344978164,
+ "grad_norm": 0.04749064892530441,
+ "learning_rate": 0.0006,
+ "loss": 6.478779315948486,
+ "step": 3231
+ },
+ {
+ "epoch": 44.8943231441048,
+ "grad_norm": 0.028002185747027397,
+ "learning_rate": 0.0006,
+ "loss": 6.528904914855957,
+ "step": 3232
+ },
+ {
+ "epoch": 44.90829694323144,
+ "grad_norm": 0.036473166197538376,
+ "learning_rate": 0.0006,
+ "loss": 6.549042701721191,
+ "step": 3233
+ },
+ {
+ "epoch": 44.92227074235808,
+ "grad_norm": 0.02543315477669239,
+ "learning_rate": 0.0006,
+ "loss": 6.498875141143799,
+ "step": 3234
+ },
+ {
+ "epoch": 44.93624454148472,
+ "grad_norm": 0.029906807467341423,
+ "learning_rate": 0.0006,
+ "loss": 6.453019142150879,
+ "step": 3235
+ },
+ {
+ "epoch": 44.95021834061135,
+ "grad_norm": 0.027231857180595398,
+ "learning_rate": 0.0006,
+ "loss": 6.428829193115234,
+ "step": 3236
+ },
+ {
+ "epoch": 44.96419213973799,
+ "grad_norm": 0.02506154775619507,
+ "learning_rate": 0.0006,
+ "loss": 6.420858383178711,
+ "step": 3237
+ },
+ {
+ "epoch": 44.97816593886463,
+ "grad_norm": 0.0269472673535347,
+ "learning_rate": 0.0006,
+ "loss": 6.484989643096924,
+ "step": 3238
+ },
+ {
+ "epoch": 44.992139737991266,
+ "grad_norm": 0.023410487920045853,
+ "learning_rate": 0.0006,
+ "loss": 6.454830169677734,
+ "step": 3239
+ },
+ {
+ "epoch": 45.0,
+ "grad_norm": 0.034175023436546326,
+ "learning_rate": 0.0006,
+ "loss": 6.430975437164307,
+ "step": 3240
+ },
+ {
+ "epoch": 45.0,
+ "eval_loss": 6.458896160125732,
+ "eval_runtime": 56.6934,
+ "eval_samples_per_second": 43.074,
+ "eval_steps_per_second": 1.358,
+ "step": 3240
+ },
+ {
+ "epoch": 45.01397379912664,
+ "grad_norm": 0.045031629502773285,
+ "learning_rate": 0.0006,
+ "loss": 6.4243011474609375,
+ "step": 3241
+ },
+ {
+ "epoch": 45.02794759825328,
+ "grad_norm": 0.04968734085559845,
+ "learning_rate": 0.0006,
+ "loss": 6.380091667175293,
+ "step": 3242
+ },
+ {
+ "epoch": 45.041921397379916,
+ "grad_norm": 0.047145187854766846,
+ "learning_rate": 0.0006,
+ "loss": 6.477178573608398,
+ "step": 3243
+ },
+ {
+ "epoch": 45.05589519650655,
+ "grad_norm": 0.04304948449134827,
+ "learning_rate": 0.0006,
+ "loss": 6.399726867675781,
+ "step": 3244
+ },
+ {
+ "epoch": 45.069868995633186,
+ "grad_norm": 0.04466010630130768,
+ "learning_rate": 0.0006,
+ "loss": 6.3415679931640625,
+ "step": 3245
+ },
+ {
+ "epoch": 45.083842794759825,
+ "grad_norm": 0.04305306077003479,
+ "learning_rate": 0.0006,
+ "loss": 6.3413310050964355,
+ "step": 3246
+ },
+ {
+ "epoch": 45.09781659388646,
+ "grad_norm": 0.043064575642347336,
+ "learning_rate": 0.0006,
+ "loss": 6.4838714599609375,
+ "step": 3247
+ },
+ {
+ "epoch": 45.1117903930131,
+ "grad_norm": 0.049503277987241745,
+ "learning_rate": 0.0006,
+ "loss": 6.38473653793335,
+ "step": 3248
+ },
+ {
+ "epoch": 45.12576419213974,
+ "grad_norm": 0.04613238573074341,
+ "learning_rate": 0.0006,
+ "loss": 6.370769500732422,
+ "step": 3249
+ },
+ {
+ "epoch": 45.13973799126637,
+ "grad_norm": 0.052119478583335876,
+ "learning_rate": 0.0006,
+ "loss": 6.362321853637695,
+ "step": 3250
+ },
+ {
+ "epoch": 45.15371179039301,
+ "grad_norm": 0.045110076665878296,
+ "learning_rate": 0.0006,
+ "loss": 6.369844913482666,
+ "step": 3251
+ },
+ {
+ "epoch": 45.16768558951965,
+ "grad_norm": 0.03798583149909973,
+ "learning_rate": 0.0006,
+ "loss": 6.324633598327637,
+ "step": 3252
+ },
+ {
+ "epoch": 45.18165938864629,
+ "grad_norm": 0.03744659945368767,
+ "learning_rate": 0.0006,
+ "loss": 6.349421501159668,
+ "step": 3253
+ },
+ {
+ "epoch": 45.19563318777293,
+ "grad_norm": 0.03767292574048042,
+ "learning_rate": 0.0006,
+ "loss": 6.334238052368164,
+ "step": 3254
+ },
+ {
+ "epoch": 45.209606986899566,
+ "grad_norm": 0.03577509894967079,
+ "learning_rate": 0.0006,
+ "loss": 6.3453826904296875,
+ "step": 3255
+ },
+ {
+ "epoch": 45.223580786026204,
+ "grad_norm": 0.03403240442276001,
+ "learning_rate": 0.0006,
+ "loss": 6.2836174964904785,
+ "step": 3256
+ },
+ {
+ "epoch": 45.237554585152836,
+ "grad_norm": 0.03517331928014755,
+ "learning_rate": 0.0006,
+ "loss": 6.281718730926514,
+ "step": 3257
+ },
+ {
+ "epoch": 45.251528384279474,
+ "grad_norm": 0.04409139230847359,
+ "learning_rate": 0.0006,
+ "loss": 6.320815086364746,
+ "step": 3258
+ },
+ {
+ "epoch": 45.26550218340611,
+ "grad_norm": 0.054750364273786545,
+ "learning_rate": 0.0006,
+ "loss": 6.343618392944336,
+ "step": 3259
+ },
+ {
+ "epoch": 45.27947598253275,
+ "grad_norm": 0.07612067461013794,
+ "learning_rate": 0.0006,
+ "loss": 6.310614585876465,
+ "step": 3260
+ },
+ {
+ "epoch": 45.29344978165939,
+ "grad_norm": 0.08438985049724579,
+ "learning_rate": 0.0006,
+ "loss": 6.2369184494018555,
+ "step": 3261
+ },
+ {
+ "epoch": 45.30742358078603,
+ "grad_norm": 0.0782201811671257,
+ "learning_rate": 0.0006,
+ "loss": 6.254452705383301,
+ "step": 3262
+ },
+ {
+ "epoch": 45.32139737991266,
+ "grad_norm": 0.09928355365991592,
+ "learning_rate": 0.0006,
+ "loss": 6.3112077713012695,
+ "step": 3263
+ },
+ {
+ "epoch": 45.3353711790393,
+ "grad_norm": 0.07233133912086487,
+ "learning_rate": 0.0006,
+ "loss": 6.408087253570557,
+ "step": 3264
+ },
+ {
+ "epoch": 45.34934497816594,
+ "grad_norm": 0.07478684186935425,
+ "learning_rate": 0.0006,
+ "loss": 6.3770036697387695,
+ "step": 3265
+ },
+ {
+ "epoch": 45.36331877729258,
+ "grad_norm": 0.06036846712231636,
+ "learning_rate": 0.0006,
+ "loss": 6.33865213394165,
+ "step": 3266
+ },
+ {
+ "epoch": 45.377292576419215,
+ "grad_norm": 0.05783458054065704,
+ "learning_rate": 0.0006,
+ "loss": 6.363203048706055,
+ "step": 3267
+ },
+ {
+ "epoch": 45.391266375545854,
+ "grad_norm": 0.05443909019231796,
+ "learning_rate": 0.0006,
+ "loss": 6.287069797515869,
+ "step": 3268
+ },
+ {
+ "epoch": 45.40524017467249,
+ "grad_norm": 0.0557909794151783,
+ "learning_rate": 0.0006,
+ "loss": 6.297784805297852,
+ "step": 3269
+ },
+ {
+ "epoch": 45.419213973799124,
+ "grad_norm": 0.03483372926712036,
+ "learning_rate": 0.0006,
+ "loss": 6.280734062194824,
+ "step": 3270
+ },
+ {
+ "epoch": 45.43318777292576,
+ "grad_norm": 0.0425543375313282,
+ "learning_rate": 0.0006,
+ "loss": 6.33565616607666,
+ "step": 3271
+ },
+ {
+ "epoch": 45.4471615720524,
+ "grad_norm": 0.03230159357190132,
+ "learning_rate": 0.0006,
+ "loss": 6.197698593139648,
+ "step": 3272
+ },
+ {
+ "epoch": 45.46113537117904,
+ "grad_norm": 0.036897290498018265,
+ "learning_rate": 0.0006,
+ "loss": 6.241239547729492,
+ "step": 3273
+ },
+ {
+ "epoch": 45.47510917030568,
+ "grad_norm": 0.031103838235139847,
+ "learning_rate": 0.0006,
+ "loss": 6.159435749053955,
+ "step": 3274
+ },
+ {
+ "epoch": 45.48908296943232,
+ "grad_norm": 0.027116449549794197,
+ "learning_rate": 0.0006,
+ "loss": 6.212225914001465,
+ "step": 3275
+ },
+ {
+ "epoch": 45.50305676855895,
+ "grad_norm": 0.029923981055617332,
+ "learning_rate": 0.0006,
+ "loss": 6.157269477844238,
+ "step": 3276
+ },
+ {
+ "epoch": 45.51703056768559,
+ "grad_norm": 0.02521538734436035,
+ "learning_rate": 0.0006,
+ "loss": 6.146918773651123,
+ "step": 3277
+ },
+ {
+ "epoch": 45.531004366812226,
+ "grad_norm": 0.02758920192718506,
+ "learning_rate": 0.0006,
+ "loss": 6.0446977615356445,
+ "step": 3278
+ },
+ {
+ "epoch": 45.544978165938865,
+ "grad_norm": 0.024328220635652542,
+ "learning_rate": 0.0006,
+ "loss": 6.099508285522461,
+ "step": 3279
+ },
+ {
+ "epoch": 45.5589519650655,
+ "grad_norm": 0.027425771579146385,
+ "learning_rate": 0.0006,
+ "loss": 6.084299087524414,
+ "step": 3280
+ },
+ {
+ "epoch": 45.57292576419214,
+ "grad_norm": 0.026927651837468147,
+ "learning_rate": 0.0006,
+ "loss": 6.146334171295166,
+ "step": 3281
+ },
+ {
+ "epoch": 45.58689956331878,
+ "grad_norm": 0.04054379090666771,
+ "learning_rate": 0.0006,
+ "loss": 6.1301493644714355,
+ "step": 3282
+ },
+ {
+ "epoch": 45.60087336244541,
+ "grad_norm": 0.04456290975213051,
+ "learning_rate": 0.0006,
+ "loss": 6.152511119842529,
+ "step": 3283
+ },
+ {
+ "epoch": 45.61484716157205,
+ "grad_norm": 0.052821505814790726,
+ "learning_rate": 0.0006,
+ "loss": 6.113544940948486,
+ "step": 3284
+ },
+ {
+ "epoch": 45.62882096069869,
+ "grad_norm": 0.07251739501953125,
+ "learning_rate": 0.0006,
+ "loss": 6.118298530578613,
+ "step": 3285
+ },
+ {
+ "epoch": 45.64279475982533,
+ "grad_norm": 0.09433556348085403,
+ "learning_rate": 0.0006,
+ "loss": 6.170897483825684,
+ "step": 3286
+ },
+ {
+ "epoch": 45.65676855895197,
+ "grad_norm": 0.057683493942022324,
+ "learning_rate": 0.0006,
+ "loss": 6.1747822761535645,
+ "step": 3287
+ },
+ {
+ "epoch": 45.670742358078606,
+ "grad_norm": 0.06363270431756973,
+ "learning_rate": 0.0006,
+ "loss": 6.1617889404296875,
+ "step": 3288
+ },
+ {
+ "epoch": 45.68471615720524,
+ "grad_norm": 0.07528169453144073,
+ "learning_rate": 0.0006,
+ "loss": 6.230060577392578,
+ "step": 3289
+ },
+ {
+ "epoch": 45.698689956331876,
+ "grad_norm": 0.07816170156002045,
+ "learning_rate": 0.0006,
+ "loss": 6.171257019042969,
+ "step": 3290
+ },
+ {
+ "epoch": 45.712663755458514,
+ "grad_norm": 0.05860164389014244,
+ "learning_rate": 0.0006,
+ "loss": 6.095484256744385,
+ "step": 3291
+ },
+ {
+ "epoch": 45.72663755458515,
+ "grad_norm": 0.04747429117560387,
+ "learning_rate": 0.0006,
+ "loss": 6.060067176818848,
+ "step": 3292
+ },
+ {
+ "epoch": 45.74061135371179,
+ "grad_norm": 0.04514474794268608,
+ "learning_rate": 0.0006,
+ "loss": 6.100218772888184,
+ "step": 3293
+ },
+ {
+ "epoch": 45.75458515283843,
+ "grad_norm": 0.05168681964278221,
+ "learning_rate": 0.0006,
+ "loss": 6.0621867179870605,
+ "step": 3294
+ },
+ {
+ "epoch": 45.76855895196506,
+ "grad_norm": 0.04665876924991608,
+ "learning_rate": 0.0006,
+ "loss": 6.118292808532715,
+ "step": 3295
+ },
+ {
+ "epoch": 45.7825327510917,
+ "grad_norm": 0.049374762922525406,
+ "learning_rate": 0.0006,
+ "loss": 6.098394393920898,
+ "step": 3296
+ },
+ {
+ "epoch": 45.79650655021834,
+ "grad_norm": 0.03793549910187721,
+ "learning_rate": 0.0006,
+ "loss": 6.069541931152344,
+ "step": 3297
+ },
+ {
+ "epoch": 45.81048034934498,
+ "grad_norm": 0.036362383514642715,
+ "learning_rate": 0.0006,
+ "loss": 6.1188435554504395,
+ "step": 3298
+ },
+ {
+ "epoch": 45.82445414847162,
+ "grad_norm": 0.03397097811102867,
+ "learning_rate": 0.0006,
+ "loss": 6.117589950561523,
+ "step": 3299
+ },
+ {
+ "epoch": 45.838427947598255,
+ "grad_norm": 0.030977753922343254,
+ "learning_rate": 0.0006,
+ "loss": 6.0236711502075195,
+ "step": 3300
+ },
+ {
+ "epoch": 45.852401746724894,
+ "grad_norm": 0.02800217643380165,
+ "learning_rate": 0.0006,
+ "loss": 6.034825325012207,
+ "step": 3301
+ },
+ {
+ "epoch": 45.866375545851525,
+ "grad_norm": 0.03129778802394867,
+ "learning_rate": 0.0006,
+ "loss": 5.980271339416504,
+ "step": 3302
+ },
+ {
+ "epoch": 45.880349344978164,
+ "grad_norm": 0.025886094197630882,
+ "learning_rate": 0.0006,
+ "loss": 5.966419219970703,
+ "step": 3303
+ },
+ {
+ "epoch": 45.8943231441048,
+ "grad_norm": 0.025712328031659126,
+ "learning_rate": 0.0006,
+ "loss": 5.938652515411377,
+ "step": 3304
+ },
+ {
+ "epoch": 45.90829694323144,
+ "grad_norm": 0.022508298978209496,
+ "learning_rate": 0.0006,
+ "loss": 5.982312202453613,
+ "step": 3305
+ },
+ {
+ "epoch": 45.92227074235808,
+ "grad_norm": 0.024880951270461082,
+ "learning_rate": 0.0006,
+ "loss": 5.931150436401367,
+ "step": 3306
+ },
+ {
+ "epoch": 45.93624454148472,
+ "grad_norm": 0.030007576569914818,
+ "learning_rate": 0.0006,
+ "loss": 5.936639308929443,
+ "step": 3307
+ },
+ {
+ "epoch": 45.95021834061135,
+ "grad_norm": 0.033314041793346405,
+ "learning_rate": 0.0006,
+ "loss": 5.944248676300049,
+ "step": 3308
+ },
+ {
+ "epoch": 45.96419213973799,
+ "grad_norm": 0.043078381568193436,
+ "learning_rate": 0.0006,
+ "loss": 5.971286773681641,
+ "step": 3309
+ },
+ {
+ "epoch": 45.97816593886463,
+ "grad_norm": 0.059296440333127975,
+ "learning_rate": 0.0006,
+ "loss": 5.910497665405273,
+ "step": 3310
+ },
+ {
+ "epoch": 45.992139737991266,
+ "grad_norm": 0.06232666224241257,
+ "learning_rate": 0.0006,
+ "loss": 5.929817199707031,
+ "step": 3311
+ },
+ {
+ "epoch": 46.0,
+ "grad_norm": 0.05303940549492836,
+ "learning_rate": 0.0006,
+ "loss": 5.910822868347168,
+ "step": 3312
+ },
+ {
+ "epoch": 46.0,
+ "eval_loss": 5.977851390838623,
+ "eval_runtime": 56.2435,
+ "eval_samples_per_second": 43.418,
+ "eval_steps_per_second": 1.369,
+ "step": 3312
+ },
+ {
+ "epoch": 46.01397379912664,
+ "grad_norm": 0.06943798810243607,
+ "learning_rate": 0.0006,
+ "loss": 5.854922294616699,
+ "step": 3313
+ },
+ {
+ "epoch": 46.02794759825328,
+ "grad_norm": 0.05236278846859932,
+ "learning_rate": 0.0006,
+ "loss": 5.876611232757568,
+ "step": 3314
+ },
+ {
+ "epoch": 46.041921397379916,
+ "grad_norm": 0.06240001320838928,
+ "learning_rate": 0.0006,
+ "loss": 5.878912448883057,
+ "step": 3315
+ },
+ {
+ "epoch": 46.05589519650655,
+ "grad_norm": 0.07401192933320999,
+ "learning_rate": 0.0006,
+ "loss": 5.985136032104492,
+ "step": 3316
+ },
+ {
+ "epoch": 46.069868995633186,
+ "grad_norm": 0.0826529711484909,
+ "learning_rate": 0.0006,
+ "loss": 5.939439296722412,
+ "step": 3317
+ },
+ {
+ "epoch": 46.083842794759825,
+ "grad_norm": 0.07999005168676376,
+ "learning_rate": 0.0006,
+ "loss": 5.982336044311523,
+ "step": 3318
+ },
+ {
+ "epoch": 46.09781659388646,
+ "grad_norm": 0.07014869153499603,
+ "learning_rate": 0.0006,
+ "loss": 5.896534442901611,
+ "step": 3319
+ },
+ {
+ "epoch": 46.1117903930131,
+ "grad_norm": 0.05795128643512726,
+ "learning_rate": 0.0006,
+ "loss": 5.843111991882324,
+ "step": 3320
+ },
+ {
+ "epoch": 46.12576419213974,
+ "grad_norm": 0.06374258548021317,
+ "learning_rate": 0.0006,
+ "loss": 5.78752326965332,
+ "step": 3321
+ },
+ {
+ "epoch": 46.13973799126637,
+ "grad_norm": 0.045524805784225464,
+ "learning_rate": 0.0006,
+ "loss": 5.86755895614624,
+ "step": 3322
+ },
+ {
+ "epoch": 46.15371179039301,
+ "grad_norm": 0.03773504123091698,
+ "learning_rate": 0.0006,
+ "loss": 5.83627462387085,
+ "step": 3323
+ },
+ {
+ "epoch": 46.16768558951965,
+ "grad_norm": 0.035815201699733734,
+ "learning_rate": 0.0006,
+ "loss": 5.798238277435303,
+ "step": 3324
+ },
+ {
+ "epoch": 46.18165938864629,
+ "grad_norm": 0.03756783902645111,
+ "learning_rate": 0.0006,
+ "loss": 5.845197677612305,
+ "step": 3325
+ },
+ {
+ "epoch": 46.19563318777293,
+ "grad_norm": 0.03569113463163376,
+ "learning_rate": 0.0006,
+ "loss": 5.888932704925537,
+ "step": 3326
+ },
+ {
+ "epoch": 46.209606986899566,
+ "grad_norm": 0.029467813670635223,
+ "learning_rate": 0.0006,
+ "loss": 5.792047500610352,
+ "step": 3327
+ },
+ {
+ "epoch": 46.223580786026204,
+ "grad_norm": 0.025698406621813774,
+ "learning_rate": 0.0006,
+ "loss": 5.783734321594238,
+ "step": 3328
+ },
+ {
+ "epoch": 46.237554585152836,
+ "grad_norm": 0.028426671400666237,
+ "learning_rate": 0.0006,
+ "loss": 5.762084007263184,
+ "step": 3329
+ },
+ {
+ "epoch": 46.251528384279474,
+ "grad_norm": 0.03099622018635273,
+ "learning_rate": 0.0006,
+ "loss": 5.719109058380127,
+ "step": 3330
+ },
+ {
+ "epoch": 46.26550218340611,
+ "grad_norm": 0.0310219656676054,
+ "learning_rate": 0.0006,
+ "loss": 5.737057209014893,
+ "step": 3331
+ },
+ {
+ "epoch": 46.27947598253275,
+ "grad_norm": 0.02713446132838726,
+ "learning_rate": 0.0006,
+ "loss": 5.758657455444336,
+ "step": 3332
+ },
+ {
+ "epoch": 46.29344978165939,
+ "grad_norm": 0.03099844604730606,
+ "learning_rate": 0.0006,
+ "loss": 5.742086410522461,
+ "step": 3333
+ },
+ {
+ "epoch": 46.30742358078603,
+ "grad_norm": 0.030258730053901672,
+ "learning_rate": 0.0006,
+ "loss": 5.767426490783691,
+ "step": 3334
+ },
+ {
+ "epoch": 46.32139737991266,
+ "grad_norm": 0.029034852981567383,
+ "learning_rate": 0.0006,
+ "loss": 5.651792049407959,
+ "step": 3335
+ },
+ {
+ "epoch": 46.3353711790393,
+ "grad_norm": 0.03138665854930878,
+ "learning_rate": 0.0006,
+ "loss": 5.664980888366699,
+ "step": 3336
+ },
+ {
+ "epoch": 46.34934497816594,
+ "grad_norm": 0.040546901524066925,
+ "learning_rate": 0.0006,
+ "loss": 5.665165901184082,
+ "step": 3337
+ },
+ {
+ "epoch": 46.36331877729258,
+ "grad_norm": 0.06395485252141953,
+ "learning_rate": 0.0006,
+ "loss": 5.7776713371276855,
+ "step": 3338
+ },
+ {
+ "epoch": 46.377292576419215,
+ "grad_norm": 0.10088056325912476,
+ "learning_rate": 0.0006,
+ "loss": 5.755645751953125,
+ "step": 3339
+ },
+ {
+ "epoch": 46.391266375545854,
+ "grad_norm": 0.09527938067913055,
+ "learning_rate": 0.0006,
+ "loss": 5.838809490203857,
+ "step": 3340
+ },
+ {
+ "epoch": 46.40524017467249,
+ "grad_norm": 0.0849648267030716,
+ "learning_rate": 0.0006,
+ "loss": 5.781932830810547,
+ "step": 3341
+ },
+ {
+ "epoch": 46.419213973799124,
+ "grad_norm": 0.10142666101455688,
+ "learning_rate": 0.0006,
+ "loss": 5.838970184326172,
+ "step": 3342
+ },
+ {
+ "epoch": 46.43318777292576,
+ "grad_norm": 0.42731454968452454,
+ "learning_rate": 0.0006,
+ "loss": 6.071902275085449,
+ "step": 3343
+ },
+ {
+ "epoch": 46.4471615720524,
+ "grad_norm": 0.13102488219738007,
+ "learning_rate": 0.0006,
+ "loss": 6.8635993003845215,
+ "step": 3344
+ },
+ {
+ "epoch": 46.46113537117904,
+ "grad_norm": 0.23691938817501068,
+ "learning_rate": 0.0006,
+ "loss": 7.313158988952637,
+ "step": 3345
+ },
+ {
+ "epoch": 46.47510917030568,
+ "grad_norm": 0.11670669168233871,
+ "learning_rate": 0.0006,
+ "loss": 7.240363121032715,
+ "step": 3346
+ },
+ {
+ "epoch": 46.48908296943232,
+ "grad_norm": 0.09985820949077606,
+ "learning_rate": 0.0006,
+ "loss": 7.113332271575928,
+ "step": 3347
+ },
+ {
+ "epoch": 46.50305676855895,
+ "grad_norm": 0.11406408250331879,
+ "learning_rate": 0.0006,
+ "loss": 7.051285743713379,
+ "step": 3348
+ },
+ {
+ "epoch": 46.51703056768559,
+ "grad_norm": 0.07900336384773254,
+ "learning_rate": 0.0006,
+ "loss": 7.014365196228027,
+ "step": 3349
+ },
+ {
+ "epoch": 46.531004366812226,
+ "grad_norm": 0.05380333960056305,
+ "learning_rate": 0.0006,
+ "loss": 6.8868818283081055,
+ "step": 3350
+ },
+ {
+ "epoch": 46.544978165938865,
+ "grad_norm": 0.05868449807167053,
+ "learning_rate": 0.0006,
+ "loss": 6.783599853515625,
+ "step": 3351
+ },
+ {
+ "epoch": 46.5589519650655,
+ "grad_norm": 0.04428785294294357,
+ "learning_rate": 0.0006,
+ "loss": 6.798511505126953,
+ "step": 3352
+ },
+ {
+ "epoch": 46.57292576419214,
+ "grad_norm": 0.055527277290821075,
+ "learning_rate": 0.0006,
+ "loss": 6.694507598876953,
+ "step": 3353
+ },
+ {
+ "epoch": 46.58689956331878,
+ "grad_norm": 0.03872432932257652,
+ "learning_rate": 0.0006,
+ "loss": 6.736631393432617,
+ "step": 3354
+ },
+ {
+ "epoch": 46.60087336244541,
+ "grad_norm": 0.04330779239535332,
+ "learning_rate": 0.0006,
+ "loss": 6.576806545257568,
+ "step": 3355
+ },
+ {
+ "epoch": 46.61484716157205,
+ "grad_norm": 0.036767736077308655,
+ "learning_rate": 0.0006,
+ "loss": 6.559871196746826,
+ "step": 3356
+ },
+ {
+ "epoch": 46.62882096069869,
+ "grad_norm": 0.039857108145952225,
+ "learning_rate": 0.0006,
+ "loss": 6.5749831199646,
+ "step": 3357
+ },
+ {
+ "epoch": 46.64279475982533,
+ "grad_norm": 0.048719972372055054,
+ "learning_rate": 0.0006,
+ "loss": 6.4915618896484375,
+ "step": 3358
+ },
+ {
+ "epoch": 46.65676855895197,
+ "grad_norm": 0.042849015444517136,
+ "learning_rate": 0.0006,
+ "loss": 6.413516044616699,
+ "step": 3359
+ },
+ {
+ "epoch": 46.670742358078606,
+ "grad_norm": 0.02871009334921837,
+ "learning_rate": 0.0006,
+ "loss": 6.320952415466309,
+ "step": 3360
+ },
+ {
+ "epoch": 46.68471615720524,
+ "grad_norm": 0.03193620592355728,
+ "learning_rate": 0.0006,
+ "loss": 6.4176435470581055,
+ "step": 3361
+ },
+ {
+ "epoch": 46.698689956331876,
+ "grad_norm": 0.027417024597525597,
+ "learning_rate": 0.0006,
+ "loss": 6.342319488525391,
+ "step": 3362
+ },
+ {
+ "epoch": 46.712663755458514,
+ "grad_norm": 0.03305937349796295,
+ "learning_rate": 0.0006,
+ "loss": 6.246112823486328,
+ "step": 3363
+ },
+ {
+ "epoch": 46.72663755458515,
+ "grad_norm": 0.031403183937072754,
+ "learning_rate": 0.0006,
+ "loss": 6.149503231048584,
+ "step": 3364
+ },
+ {
+ "epoch": 46.74061135371179,
+ "grad_norm": 0.031563788652420044,
+ "learning_rate": 0.0006,
+ "loss": 6.278808116912842,
+ "step": 3365
+ },
+ {
+ "epoch": 46.75458515283843,
+ "grad_norm": 0.028366589918732643,
+ "learning_rate": 0.0006,
+ "loss": 6.057036876678467,
+ "step": 3366
+ },
+ {
+ "epoch": 46.76855895196506,
+ "grad_norm": 0.023715803399682045,
+ "learning_rate": 0.0006,
+ "loss": 6.111324310302734,
+ "step": 3367
+ },
+ {
+ "epoch": 46.7825327510917,
+ "grad_norm": 0.026546811684966087,
+ "learning_rate": 0.0006,
+ "loss": 6.105669975280762,
+ "step": 3368
+ },
+ {
+ "epoch": 46.79650655021834,
+ "grad_norm": 0.023531019687652588,
+ "learning_rate": 0.0006,
+ "loss": 6.068072319030762,
+ "step": 3369
+ },
+ {
+ "epoch": 46.81048034934498,
+ "grad_norm": 0.024480128660798073,
+ "learning_rate": 0.0006,
+ "loss": 5.947896957397461,
+ "step": 3370
+ },
+ {
+ "epoch": 46.82445414847162,
+ "grad_norm": 0.02517310529947281,
+ "learning_rate": 0.0006,
+ "loss": 6.030037879943848,
+ "step": 3371
+ },
+ {
+ "epoch": 46.838427947598255,
+ "grad_norm": 0.026723403483629227,
+ "learning_rate": 0.0006,
+ "loss": 5.963776111602783,
+ "step": 3372
+ },
+ {
+ "epoch": 46.852401746724894,
+ "grad_norm": 0.03163504600524902,
+ "learning_rate": 0.0006,
+ "loss": 5.893466949462891,
+ "step": 3373
+ },
+ {
+ "epoch": 46.866375545851525,
+ "grad_norm": 0.04058683291077614,
+ "learning_rate": 0.0006,
+ "loss": 5.9202799797058105,
+ "step": 3374
+ },
+ {
+ "epoch": 46.880349344978164,
+ "grad_norm": 0.04964117333292961,
+ "learning_rate": 0.0006,
+ "loss": 5.9366559982299805,
+ "step": 3375
+ },
+ {
+ "epoch": 46.8943231441048,
+ "grad_norm": 0.055406033992767334,
+ "learning_rate": 0.0006,
+ "loss": 5.912230968475342,
+ "step": 3376
+ },
+ {
+ "epoch": 46.90829694323144,
+ "grad_norm": 0.04643625393509865,
+ "learning_rate": 0.0006,
+ "loss": 5.87260103225708,
+ "step": 3377
+ },
+ {
+ "epoch": 46.92227074235808,
+ "grad_norm": 0.029325876384973526,
+ "learning_rate": 0.0006,
+ "loss": 5.854763984680176,
+ "step": 3378
+ },
+ {
+ "epoch": 46.93624454148472,
+ "grad_norm": 0.03159382566809654,
+ "learning_rate": 0.0006,
+ "loss": 5.880882263183594,
+ "step": 3379
+ },
+ {
+ "epoch": 46.95021834061135,
+ "grad_norm": 0.030224012210965157,
+ "learning_rate": 0.0006,
+ "loss": 5.787360191345215,
+ "step": 3380
+ },
+ {
+ "epoch": 46.96419213973799,
+ "grad_norm": 0.028368130326271057,
+ "learning_rate": 0.0006,
+ "loss": 5.817195892333984,
+ "step": 3381
+ },
+ {
+ "epoch": 46.97816593886463,
+ "grad_norm": 0.029411399737000465,
+ "learning_rate": 0.0006,
+ "loss": 5.783403396606445,
+ "step": 3382
+ },
+ {
+ "epoch": 46.992139737991266,
+ "grad_norm": 0.029927462339401245,
+ "learning_rate": 0.0006,
+ "loss": 5.662444591522217,
+ "step": 3383
+ },
+ {
+ "epoch": 47.0,
+ "grad_norm": 0.02872961387038231,
+ "learning_rate": 0.0006,
+ "loss": 5.742781639099121,
+ "step": 3384
+ },
+ {
+ "epoch": 47.0,
+ "eval_loss": 5.870260238647461,
+ "eval_runtime": 57.3734,
+ "eval_samples_per_second": 42.563,
+ "eval_steps_per_second": 1.342,
+ "step": 3384
+ },
+ {
+ "epoch": 47.01397379912664,
+ "grad_norm": 0.028542131185531616,
+ "learning_rate": 0.0006,
+ "loss": 5.753256797790527,
+ "step": 3385
+ },
+ {
+ "epoch": 47.02794759825328,
+ "grad_norm": 0.03227119892835617,
+ "learning_rate": 0.0006,
+ "loss": 5.722246170043945,
+ "step": 3386
+ },
+ {
+ "epoch": 47.041921397379916,
+ "grad_norm": 0.031461723148822784,
+ "learning_rate": 0.0006,
+ "loss": 5.740520000457764,
+ "step": 3387
+ },
+ {
+ "epoch": 47.05589519650655,
+ "grad_norm": 0.03184220567345619,
+ "learning_rate": 0.0006,
+ "loss": 5.720890045166016,
+ "step": 3388
+ },
+ {
+ "epoch": 47.069868995633186,
+ "grad_norm": 0.030015580356121063,
+ "learning_rate": 0.0006,
+ "loss": 5.597814559936523,
+ "step": 3389
+ },
+ {
+ "epoch": 47.083842794759825,
+ "grad_norm": 0.029931779950857162,
+ "learning_rate": 0.0006,
+ "loss": 5.5412468910217285,
+ "step": 3390
+ },
+ {
+ "epoch": 47.09781659388646,
+ "grad_norm": 0.03206505626440048,
+ "learning_rate": 0.0006,
+ "loss": 5.6967597007751465,
+ "step": 3391
+ },
+ {
+ "epoch": 47.1117903930131,
+ "grad_norm": 0.03978161886334419,
+ "learning_rate": 0.0006,
+ "loss": 5.59661340713501,
+ "step": 3392
+ },
+ {
+ "epoch": 47.12576419213974,
+ "grad_norm": 0.04619080200791359,
+ "learning_rate": 0.0006,
+ "loss": 5.795788764953613,
+ "step": 3393
+ },
+ {
+ "epoch": 47.13973799126637,
+ "grad_norm": 0.049085382372140884,
+ "learning_rate": 0.0006,
+ "loss": 5.667335510253906,
+ "step": 3394
+ },
+ {
+ "epoch": 47.15371179039301,
+ "grad_norm": 0.04830858111381531,
+ "learning_rate": 0.0006,
+ "loss": 5.668806552886963,
+ "step": 3395
+ },
+ {
+ "epoch": 47.16768558951965,
+ "grad_norm": 0.0583178736269474,
+ "learning_rate": 0.0006,
+ "loss": 5.692285537719727,
+ "step": 3396
+ },
+ {
+ "epoch": 47.18165938864629,
+ "grad_norm": 0.05182253196835518,
+ "learning_rate": 0.0006,
+ "loss": 5.678645133972168,
+ "step": 3397
+ },
+ {
+ "epoch": 47.19563318777293,
+ "grad_norm": 0.04011905938386917,
+ "learning_rate": 0.0006,
+ "loss": 5.661550998687744,
+ "step": 3398
+ },
+ {
+ "epoch": 47.209606986899566,
+ "grad_norm": 0.03888486698269844,
+ "learning_rate": 0.0006,
+ "loss": 5.643001556396484,
+ "step": 3399
+ },
+ {
+ "epoch": 47.223580786026204,
+ "grad_norm": 0.030658265575766563,
+ "learning_rate": 0.0006,
+ "loss": 5.659090995788574,
+ "step": 3400
+ },
+ {
+ "epoch": 47.237554585152836,
+ "grad_norm": 0.03609256446361542,
+ "learning_rate": 0.0006,
+ "loss": 5.544501781463623,
+ "step": 3401
+ },
+ {
+ "epoch": 47.251528384279474,
+ "grad_norm": 0.033589769154787064,
+ "learning_rate": 0.0006,
+ "loss": 5.463878631591797,
+ "step": 3402
+ },
+ {
+ "epoch": 47.26550218340611,
+ "grad_norm": 0.035128287971019745,
+ "learning_rate": 0.0006,
+ "loss": 5.5680060386657715,
+ "step": 3403
+ },
+ {
+ "epoch": 47.27947598253275,
+ "grad_norm": 0.038684550672769547,
+ "learning_rate": 0.0006,
+ "loss": 5.587898254394531,
+ "step": 3404
+ },
+ {
+ "epoch": 47.29344978165939,
+ "grad_norm": 0.043566226959228516,
+ "learning_rate": 0.0006,
+ "loss": 5.580896854400635,
+ "step": 3405
+ },
+ {
+ "epoch": 47.30742358078603,
+ "grad_norm": 0.04428374394774437,
+ "learning_rate": 0.0006,
+ "loss": 5.603460311889648,
+ "step": 3406
+ },
+ {
+ "epoch": 47.32139737991266,
+ "grad_norm": 0.037368934601545334,
+ "learning_rate": 0.0006,
+ "loss": 5.564740180969238,
+ "step": 3407
+ },
+ {
+ "epoch": 47.3353711790393,
+ "grad_norm": 0.04092350974678993,
+ "learning_rate": 0.0006,
+ "loss": 5.584153652191162,
+ "step": 3408
+ },
+ {
+ "epoch": 47.34934497816594,
+ "grad_norm": 0.045031506568193436,
+ "learning_rate": 0.0006,
+ "loss": 5.601186275482178,
+ "step": 3409
+ },
+ {
+ "epoch": 47.36331877729258,
+ "grad_norm": 0.041830409318208694,
+ "learning_rate": 0.0006,
+ "loss": 5.512788772583008,
+ "step": 3410
+ },
+ {
+ "epoch": 47.377292576419215,
+ "grad_norm": 0.037633974105119705,
+ "learning_rate": 0.0006,
+ "loss": 5.503828525543213,
+ "step": 3411
+ },
+ {
+ "epoch": 47.391266375545854,
+ "grad_norm": 0.03473588079214096,
+ "learning_rate": 0.0006,
+ "loss": 5.570557594299316,
+ "step": 3412
+ },
+ {
+ "epoch": 47.40524017467249,
+ "grad_norm": 0.039325181394815445,
+ "learning_rate": 0.0006,
+ "loss": 5.427783966064453,
+ "step": 3413
+ },
+ {
+ "epoch": 47.419213973799124,
+ "grad_norm": 0.04034624621272087,
+ "learning_rate": 0.0006,
+ "loss": 5.572591304779053,
+ "step": 3414
+ },
+ {
+ "epoch": 47.43318777292576,
+ "grad_norm": 0.05245184153318405,
+ "learning_rate": 0.0006,
+ "loss": 5.505208969116211,
+ "step": 3415
+ },
+ {
+ "epoch": 47.4471615720524,
+ "grad_norm": 0.0568675734102726,
+ "learning_rate": 0.0006,
+ "loss": 5.589373588562012,
+ "step": 3416
+ },
+ {
+ "epoch": 47.46113537117904,
+ "grad_norm": 0.048336587846279144,
+ "learning_rate": 0.0006,
+ "loss": 5.522005081176758,
+ "step": 3417
+ },
+ {
+ "epoch": 47.47510917030568,
+ "grad_norm": 0.04194975271821022,
+ "learning_rate": 0.0006,
+ "loss": 5.387576103210449,
+ "step": 3418
+ },
+ {
+ "epoch": 47.48908296943232,
+ "grad_norm": 0.04019433259963989,
+ "learning_rate": 0.0006,
+ "loss": 5.522615432739258,
+ "step": 3419
+ },
+ {
+ "epoch": 47.50305676855895,
+ "grad_norm": 0.05709933117032051,
+ "learning_rate": 0.0006,
+ "loss": 5.420286178588867,
+ "step": 3420
+ },
+ {
+ "epoch": 47.51703056768559,
+ "grad_norm": 0.09130338579416275,
+ "learning_rate": 0.0006,
+ "loss": 5.484622478485107,
+ "step": 3421
+ },
+ {
+ "epoch": 47.531004366812226,
+ "grad_norm": 0.2757541835308075,
+ "learning_rate": 0.0006,
+ "loss": 5.821832656860352,
+ "step": 3422
+ },
+ {
+ "epoch": 47.544978165938865,
+ "grad_norm": 0.28292080760002136,
+ "learning_rate": 0.0006,
+ "loss": 6.579934120178223,
+ "step": 3423
+ },
+ {
+ "epoch": 47.5589519650655,
+ "grad_norm": 0.15293604135513306,
+ "learning_rate": 0.0006,
+ "loss": 6.478213310241699,
+ "step": 3424
+ },
+ {
+ "epoch": 47.57292576419214,
+ "grad_norm": 0.11284393817186356,
+ "learning_rate": 0.0006,
+ "loss": 6.4066386222839355,
+ "step": 3425
+ },
+ {
+ "epoch": 47.58689956331878,
+ "grad_norm": 0.10933073610067368,
+ "learning_rate": 0.0006,
+ "loss": 6.3306169509887695,
+ "step": 3426
+ },
+ {
+ "epoch": 47.60087336244541,
+ "grad_norm": 0.05837641656398773,
+ "learning_rate": 0.0006,
+ "loss": 6.250000953674316,
+ "step": 3427
+ },
+ {
+ "epoch": 47.61484716157205,
+ "grad_norm": 0.05987229570746422,
+ "learning_rate": 0.0006,
+ "loss": 6.152463436126709,
+ "step": 3428
+ },
+ {
+ "epoch": 47.62882096069869,
+ "grad_norm": 0.04064938426017761,
+ "learning_rate": 0.0006,
+ "loss": 6.117897033691406,
+ "step": 3429
+ },
+ {
+ "epoch": 47.64279475982533,
+ "grad_norm": 0.04565057530999184,
+ "learning_rate": 0.0006,
+ "loss": 6.042291641235352,
+ "step": 3430
+ },
+ {
+ "epoch": 47.65676855895197,
+ "grad_norm": 0.03475063294172287,
+ "learning_rate": 0.0006,
+ "loss": 6.0071868896484375,
+ "step": 3431
+ },
+ {
+ "epoch": 47.670742358078606,
+ "grad_norm": 0.03809220343828201,
+ "learning_rate": 0.0006,
+ "loss": 6.0128021240234375,
+ "step": 3432
+ },
+ {
+ "epoch": 47.68471615720524,
+ "grad_norm": 0.03307219594717026,
+ "learning_rate": 0.0006,
+ "loss": 5.894664287567139,
+ "step": 3433
+ },
+ {
+ "epoch": 47.698689956331876,
+ "grad_norm": 0.028824346140027046,
+ "learning_rate": 0.0006,
+ "loss": 5.914865493774414,
+ "step": 3434
+ },
+ {
+ "epoch": 47.712663755458514,
+ "grad_norm": 0.027923690155148506,
+ "learning_rate": 0.0006,
+ "loss": 5.851759910583496,
+ "step": 3435
+ },
+ {
+ "epoch": 47.72663755458515,
+ "grad_norm": 0.025492684915661812,
+ "learning_rate": 0.0006,
+ "loss": 5.749186038970947,
+ "step": 3436
+ },
+ {
+ "epoch": 47.74061135371179,
+ "grad_norm": 0.025105174630880356,
+ "learning_rate": 0.0006,
+ "loss": 5.803566932678223,
+ "step": 3437
+ },
+ {
+ "epoch": 47.75458515283843,
+ "grad_norm": 0.023298880085349083,
+ "learning_rate": 0.0006,
+ "loss": 5.76276159286499,
+ "step": 3438
+ },
+ {
+ "epoch": 47.76855895196506,
+ "grad_norm": 0.02468576282262802,
+ "learning_rate": 0.0006,
+ "loss": 5.833852767944336,
+ "step": 3439
+ },
+ {
+ "epoch": 47.7825327510917,
+ "grad_norm": 0.022730689495801926,
+ "learning_rate": 0.0006,
+ "loss": 5.736724376678467,
+ "step": 3440
+ },
+ {
+ "epoch": 47.79650655021834,
+ "grad_norm": 0.021335354074835777,
+ "learning_rate": 0.0006,
+ "loss": 5.644248008728027,
+ "step": 3441
+ },
+ {
+ "epoch": 47.81048034934498,
+ "grad_norm": 0.02288101427257061,
+ "learning_rate": 0.0006,
+ "loss": 5.7011942863464355,
+ "step": 3442
+ },
+ {
+ "epoch": 47.82445414847162,
+ "grad_norm": 0.01961585320532322,
+ "learning_rate": 0.0006,
+ "loss": 5.63215446472168,
+ "step": 3443
+ },
+ {
+ "epoch": 47.838427947598255,
+ "grad_norm": 0.017659205943346024,
+ "learning_rate": 0.0006,
+ "loss": 5.744845390319824,
+ "step": 3444
+ },
+ {
+ "epoch": 47.852401746724894,
+ "grad_norm": 0.02281784638762474,
+ "learning_rate": 0.0006,
+ "loss": 5.696993350982666,
+ "step": 3445
+ },
+ {
+ "epoch": 47.866375545851525,
+ "grad_norm": 0.016394242644309998,
+ "learning_rate": 0.0006,
+ "loss": 5.659870147705078,
+ "step": 3446
+ },
+ {
+ "epoch": 47.880349344978164,
+ "grad_norm": 0.01967368647456169,
+ "learning_rate": 0.0006,
+ "loss": 5.638072490692139,
+ "step": 3447
+ },
+ {
+ "epoch": 47.8943231441048,
+ "grad_norm": 0.017326267436146736,
+ "learning_rate": 0.0006,
+ "loss": 5.638836860656738,
+ "step": 3448
+ },
+ {
+ "epoch": 47.90829694323144,
+ "grad_norm": 0.017778463661670685,
+ "learning_rate": 0.0006,
+ "loss": 5.634466648101807,
+ "step": 3449
+ },
+ {
+ "epoch": 47.92227074235808,
+ "grad_norm": 0.017269758507609367,
+ "learning_rate": 0.0006,
+ "loss": 5.506277561187744,
+ "step": 3450
+ },
+ {
+ "epoch": 47.93624454148472,
+ "grad_norm": 0.01733289286494255,
+ "learning_rate": 0.0006,
+ "loss": 5.544455528259277,
+ "step": 3451
+ },
+ {
+ "epoch": 47.95021834061135,
+ "grad_norm": 0.017787795513868332,
+ "learning_rate": 0.0006,
+ "loss": 5.498943328857422,
+ "step": 3452
+ },
+ {
+ "epoch": 47.96419213973799,
+ "grad_norm": 0.015770932659506798,
+ "learning_rate": 0.0006,
+ "loss": 5.537227630615234,
+ "step": 3453
+ },
+ {
+ "epoch": 47.97816593886463,
+ "grad_norm": 0.016132809221744537,
+ "learning_rate": 0.0006,
+ "loss": 5.489081382751465,
+ "step": 3454
+ },
+ {
+ "epoch": 47.992139737991266,
+ "grad_norm": 0.01749664731323719,
+ "learning_rate": 0.0006,
+ "loss": 5.454263687133789,
+ "step": 3455
+ },
+ {
+ "epoch": 48.0,
+ "grad_norm": 0.016711296513676643,
+ "learning_rate": 0.0006,
+ "loss": 5.530783653259277,
+ "step": 3456
+ },
+ {
+ "epoch": 48.0,
+ "eval_loss": 5.596925258636475,
+ "eval_runtime": 56.9612,
+ "eval_samples_per_second": 42.871,
+ "eval_steps_per_second": 1.352,
+ "step": 3456
+ },
+ {
+ "epoch": 48.01397379912664,
+ "grad_norm": 0.01867641694843769,
+ "learning_rate": 0.0006,
+ "loss": 5.506853103637695,
+ "step": 3457
+ },
+ {
+ "epoch": 48.02794759825328,
+ "grad_norm": 0.023250792175531387,
+ "learning_rate": 0.0006,
+ "loss": 5.4074177742004395,
+ "step": 3458
+ },
+ {
+ "epoch": 48.041921397379916,
+ "grad_norm": 0.01970936730504036,
+ "learning_rate": 0.0006,
+ "loss": 5.508777618408203,
+ "step": 3459
+ },
+ {
+ "epoch": 48.05589519650655,
+ "grad_norm": 0.01879260502755642,
+ "learning_rate": 0.0006,
+ "loss": 5.412424087524414,
+ "step": 3460
+ },
+ {
+ "epoch": 48.069868995633186,
+ "grad_norm": 0.017420092597603798,
+ "learning_rate": 0.0006,
+ "loss": 5.43361759185791,
+ "step": 3461
+ },
+ {
+ "epoch": 48.083842794759825,
+ "grad_norm": 0.018707862123847008,
+ "learning_rate": 0.0006,
+ "loss": 5.476800918579102,
+ "step": 3462
+ },
+ {
+ "epoch": 48.09781659388646,
+ "grad_norm": 0.023185882717370987,
+ "learning_rate": 0.0006,
+ "loss": 5.465272426605225,
+ "step": 3463
+ },
+ {
+ "epoch": 48.1117903930131,
+ "grad_norm": 0.023570368066430092,
+ "learning_rate": 0.0006,
+ "loss": 5.391471862792969,
+ "step": 3464
+ },
+ {
+ "epoch": 48.12576419213974,
+ "grad_norm": 0.02651144750416279,
+ "learning_rate": 0.0006,
+ "loss": 5.422799110412598,
+ "step": 3465
+ },
+ {
+ "epoch": 48.13973799126637,
+ "grad_norm": 0.03316329047083855,
+ "learning_rate": 0.0006,
+ "loss": 5.3952202796936035,
+ "step": 3466
+ },
+ {
+ "epoch": 48.15371179039301,
+ "grad_norm": 0.04012278467416763,
+ "learning_rate": 0.0006,
+ "loss": 5.540853023529053,
+ "step": 3467
+ },
+ {
+ "epoch": 48.16768558951965,
+ "grad_norm": 0.04763203114271164,
+ "learning_rate": 0.0006,
+ "loss": 5.387890338897705,
+ "step": 3468
+ },
+ {
+ "epoch": 48.18165938864629,
+ "grad_norm": 0.05129165202379227,
+ "learning_rate": 0.0006,
+ "loss": 5.414722442626953,
+ "step": 3469
+ },
+ {
+ "epoch": 48.19563318777293,
+ "grad_norm": 0.04438071325421333,
+ "learning_rate": 0.0006,
+ "loss": 5.386244773864746,
+ "step": 3470
+ },
+ {
+ "epoch": 48.209606986899566,
+ "grad_norm": 0.040234919637441635,
+ "learning_rate": 0.0006,
+ "loss": 5.431617736816406,
+ "step": 3471
+ },
+ {
+ "epoch": 48.223580786026204,
+ "grad_norm": 0.03217008337378502,
+ "learning_rate": 0.0006,
+ "loss": 5.30168342590332,
+ "step": 3472
+ },
+ {
+ "epoch": 48.237554585152836,
+ "grad_norm": 0.03087974339723587,
+ "learning_rate": 0.0006,
+ "loss": 5.316389083862305,
+ "step": 3473
+ },
+ {
+ "epoch": 48.251528384279474,
+ "grad_norm": 0.02789347991347313,
+ "learning_rate": 0.0006,
+ "loss": 5.385970592498779,
+ "step": 3474
+ },
+ {
+ "epoch": 48.26550218340611,
+ "grad_norm": 0.024792097508907318,
+ "learning_rate": 0.0006,
+ "loss": 5.362329959869385,
+ "step": 3475
+ },
+ {
+ "epoch": 48.27947598253275,
+ "grad_norm": 0.025869522243738174,
+ "learning_rate": 0.0006,
+ "loss": 5.28400182723999,
+ "step": 3476
+ },
+ {
+ "epoch": 48.29344978165939,
+ "grad_norm": 0.021339882165193558,
+ "learning_rate": 0.0006,
+ "loss": 5.36669921875,
+ "step": 3477
+ },
+ {
+ "epoch": 48.30742358078603,
+ "grad_norm": 0.021462520584464073,
+ "learning_rate": 0.0006,
+ "loss": 5.311222076416016,
+ "step": 3478
+ },
+ {
+ "epoch": 48.32139737991266,
+ "grad_norm": 0.02193879708647728,
+ "learning_rate": 0.0006,
+ "loss": 5.291441917419434,
+ "step": 3479
+ },
+ {
+ "epoch": 48.3353711790393,
+ "grad_norm": 0.01963093690574169,
+ "learning_rate": 0.0006,
+ "loss": 5.342780113220215,
+ "step": 3480
+ },
+ {
+ "epoch": 48.34934497816594,
+ "grad_norm": 0.020634736865758896,
+ "learning_rate": 0.0006,
+ "loss": 5.144141674041748,
+ "step": 3481
+ },
+ {
+ "epoch": 48.36331877729258,
+ "grad_norm": 0.02126968279480934,
+ "learning_rate": 0.0006,
+ "loss": 5.232624530792236,
+ "step": 3482
+ },
+ {
+ "epoch": 48.377292576419215,
+ "grad_norm": 0.021203994750976562,
+ "learning_rate": 0.0006,
+ "loss": 5.290219783782959,
+ "step": 3483
+ },
+ {
+ "epoch": 48.391266375545854,
+ "grad_norm": 0.02193225920200348,
+ "learning_rate": 0.0006,
+ "loss": 5.29197359085083,
+ "step": 3484
+ },
+ {
+ "epoch": 48.40524017467249,
+ "grad_norm": 0.023174211382865906,
+ "learning_rate": 0.0006,
+ "loss": 5.204349517822266,
+ "step": 3485
+ },
+ {
+ "epoch": 48.419213973799124,
+ "grad_norm": 0.024161914363503456,
+ "learning_rate": 0.0006,
+ "loss": 5.339351654052734,
+ "step": 3486
+ },
+ {
+ "epoch": 48.43318777292576,
+ "grad_norm": 0.02221524342894554,
+ "learning_rate": 0.0006,
+ "loss": 5.263001441955566,
+ "step": 3487
+ },
+ {
+ "epoch": 48.4471615720524,
+ "grad_norm": 0.026632830500602722,
+ "learning_rate": 0.0006,
+ "loss": 5.221281051635742,
+ "step": 3488
+ },
+ {
+ "epoch": 48.46113537117904,
+ "grad_norm": 0.034923870116472244,
+ "learning_rate": 0.0006,
+ "loss": 5.296039581298828,
+ "step": 3489
+ },
+ {
+ "epoch": 48.47510917030568,
+ "grad_norm": 0.046184979379177094,
+ "learning_rate": 0.0006,
+ "loss": 5.212769985198975,
+ "step": 3490
+ },
+ {
+ "epoch": 48.48908296943232,
+ "grad_norm": 0.05154098570346832,
+ "learning_rate": 0.0006,
+ "loss": 5.337864875793457,
+ "step": 3491
+ },
+ {
+ "epoch": 48.50305676855895,
+ "grad_norm": 0.042967259883880615,
+ "learning_rate": 0.0006,
+ "loss": 5.267958641052246,
+ "step": 3492
+ },
+ {
+ "epoch": 48.51703056768559,
+ "grad_norm": 0.05356583371758461,
+ "learning_rate": 0.0006,
+ "loss": 5.1404218673706055,
+ "step": 3493
+ },
+ {
+ "epoch": 48.531004366812226,
+ "grad_norm": 0.08219032734632492,
+ "learning_rate": 0.0006,
+ "loss": 5.300687789916992,
+ "step": 3494
+ },
+ {
+ "epoch": 48.544978165938865,
+ "grad_norm": 0.19204089045524597,
+ "learning_rate": 0.0006,
+ "loss": 5.456025123596191,
+ "step": 3495
+ },
+ {
+ "epoch": 48.5589519650655,
+ "grad_norm": 0.25917696952819824,
+ "learning_rate": 0.0006,
+ "loss": 6.156761169433594,
+ "step": 3496
+ },
+ {
+ "epoch": 48.57292576419214,
+ "grad_norm": 0.1566283404827118,
+ "learning_rate": 0.0006,
+ "loss": 5.862346649169922,
+ "step": 3497
+ },
+ {
+ "epoch": 48.58689956331878,
+ "grad_norm": 0.1761486828327179,
+ "learning_rate": 0.0006,
+ "loss": 6.012795448303223,
+ "step": 3498
+ },
+ {
+ "epoch": 48.60087336244541,
+ "grad_norm": 0.09857621043920517,
+ "learning_rate": 0.0006,
+ "loss": 5.9248151779174805,
+ "step": 3499
+ },
+ {
+ "epoch": 48.61484716157205,
+ "grad_norm": 0.05968055501580238,
+ "learning_rate": 0.0006,
+ "loss": 5.780706882476807,
+ "step": 3500
+ },
+ {
+ "epoch": 48.62882096069869,
+ "grad_norm": 0.057433564215898514,
+ "learning_rate": 0.0006,
+ "loss": 5.645388603210449,
+ "step": 3501
+ },
+ {
+ "epoch": 48.64279475982533,
+ "grad_norm": 0.039659615606069565,
+ "learning_rate": 0.0006,
+ "loss": 5.695138931274414,
+ "step": 3502
+ },
+ {
+ "epoch": 48.65676855895197,
+ "grad_norm": 0.04019695147871971,
+ "learning_rate": 0.0006,
+ "loss": 5.672435283660889,
+ "step": 3503
+ },
+ {
+ "epoch": 48.670742358078606,
+ "grad_norm": 0.03033018670976162,
+ "learning_rate": 0.0006,
+ "loss": 5.682260513305664,
+ "step": 3504
+ },
+ {
+ "epoch": 48.68471615720524,
+ "grad_norm": 0.03215787187218666,
+ "learning_rate": 0.0006,
+ "loss": 5.604182243347168,
+ "step": 3505
+ },
+ {
+ "epoch": 48.698689956331876,
+ "grad_norm": 0.026865126565098763,
+ "learning_rate": 0.0006,
+ "loss": 5.64040470123291,
+ "step": 3506
+ },
+ {
+ "epoch": 48.712663755458514,
+ "grad_norm": 0.02831609919667244,
+ "learning_rate": 0.0006,
+ "loss": 5.552720069885254,
+ "step": 3507
+ },
+ {
+ "epoch": 48.72663755458515,
+ "grad_norm": 0.024226613342761993,
+ "learning_rate": 0.0006,
+ "loss": 5.556946277618408,
+ "step": 3508
+ },
+ {
+ "epoch": 48.74061135371179,
+ "grad_norm": 0.025393934920430183,
+ "learning_rate": 0.0006,
+ "loss": 5.476178169250488,
+ "step": 3509
+ },
+ {
+ "epoch": 48.75458515283843,
+ "grad_norm": 0.01984412968158722,
+ "learning_rate": 0.0006,
+ "loss": 5.471088409423828,
+ "step": 3510
+ },
+ {
+ "epoch": 48.76855895196506,
+ "grad_norm": 0.02035129815340042,
+ "learning_rate": 0.0006,
+ "loss": 5.438812732696533,
+ "step": 3511
+ },
+ {
+ "epoch": 48.7825327510917,
+ "grad_norm": 0.0208350270986557,
+ "learning_rate": 0.0006,
+ "loss": 5.381746292114258,
+ "step": 3512
+ },
+ {
+ "epoch": 48.79650655021834,
+ "grad_norm": 0.017783882096409798,
+ "learning_rate": 0.0006,
+ "loss": 5.351675510406494,
+ "step": 3513
+ },
+ {
+ "epoch": 48.81048034934498,
+ "grad_norm": 0.01671770215034485,
+ "learning_rate": 0.0006,
+ "loss": 5.329428672790527,
+ "step": 3514
+ },
+ {
+ "epoch": 48.82445414847162,
+ "grad_norm": 0.017891215160489082,
+ "learning_rate": 0.0006,
+ "loss": 5.212074279785156,
+ "step": 3515
+ },
+ {
+ "epoch": 48.838427947598255,
+ "grad_norm": 0.016847791150212288,
+ "learning_rate": 0.0006,
+ "loss": 5.299534797668457,
+ "step": 3516
+ },
+ {
+ "epoch": 48.852401746724894,
+ "grad_norm": 0.01593637838959694,
+ "learning_rate": 0.0006,
+ "loss": 5.30466890335083,
+ "step": 3517
+ },
+ {
+ "epoch": 48.866375545851525,
+ "grad_norm": 0.014713102020323277,
+ "learning_rate": 0.0006,
+ "loss": 5.294345378875732,
+ "step": 3518
+ },
+ {
+ "epoch": 48.880349344978164,
+ "grad_norm": 0.01433857623487711,
+ "learning_rate": 0.0006,
+ "loss": 5.2148542404174805,
+ "step": 3519
+ },
+ {
+ "epoch": 48.8943231441048,
+ "grad_norm": 0.014979379251599312,
+ "learning_rate": 0.0006,
+ "loss": 5.298582077026367,
+ "step": 3520
+ },
+ {
+ "epoch": 48.90829694323144,
+ "grad_norm": 0.013494192622601986,
+ "learning_rate": 0.0006,
+ "loss": 5.309783935546875,
+ "step": 3521
+ },
+ {
+ "epoch": 48.92227074235808,
+ "grad_norm": 0.014020012691617012,
+ "learning_rate": 0.0006,
+ "loss": 5.284879684448242,
+ "step": 3522
+ },
+ {
+ "epoch": 48.93624454148472,
+ "grad_norm": 0.01348058134317398,
+ "learning_rate": 0.0006,
+ "loss": 5.249116897583008,
+ "step": 3523
+ },
+ {
+ "epoch": 48.95021834061135,
+ "grad_norm": 0.013280218467116356,
+ "learning_rate": 0.0006,
+ "loss": 5.273489952087402,
+ "step": 3524
+ },
+ {
+ "epoch": 48.96419213973799,
+ "grad_norm": 0.012647110037505627,
+ "learning_rate": 0.0006,
+ "loss": 5.306652069091797,
+ "step": 3525
+ },
+ {
+ "epoch": 48.97816593886463,
+ "grad_norm": 0.012847724370658398,
+ "learning_rate": 0.0006,
+ "loss": 5.235745429992676,
+ "step": 3526
+ },
+ {
+ "epoch": 48.992139737991266,
+ "grad_norm": 0.011820166371762753,
+ "learning_rate": 0.0006,
+ "loss": 5.238091468811035,
+ "step": 3527
+ },
+ {
+ "epoch": 49.0,
+ "grad_norm": 0.01255585066974163,
+ "learning_rate": 0.0006,
+ "loss": 5.231295585632324,
+ "step": 3528
+ },
+ {
+ "epoch": 49.0,
+ "eval_loss": 5.394587993621826,
+ "eval_runtime": 56.3734,
+ "eval_samples_per_second": 43.318,
+ "eval_steps_per_second": 1.366,
+ "step": 3528
+ },
+ {
+ "epoch": 49.01397379912664,
+ "grad_norm": 0.011031972244381905,
+ "learning_rate": 0.0006,
+ "loss": 5.1354570388793945,
+ "step": 3529
+ },
+ {
+ "epoch": 49.02794759825328,
+ "grad_norm": 0.010631263256072998,
+ "learning_rate": 0.0006,
+ "loss": 5.177639961242676,
+ "step": 3530
+ },
+ {
+ "epoch": 49.041921397379916,
+ "grad_norm": 0.01133775431662798,
+ "learning_rate": 0.0006,
+ "loss": 5.283428192138672,
+ "step": 3531
+ },
+ {
+ "epoch": 49.05589519650655,
+ "grad_norm": 0.01154172234237194,
+ "learning_rate": 0.0006,
+ "loss": 5.206364154815674,
+ "step": 3532
+ },
+ {
+ "epoch": 49.069868995633186,
+ "grad_norm": 0.010192356072366238,
+ "learning_rate": 0.0006,
+ "loss": 5.2385735511779785,
+ "step": 3533
+ },
+ {
+ "epoch": 49.083842794759825,
+ "grad_norm": 0.012789200060069561,
+ "learning_rate": 0.0006,
+ "loss": 5.071800231933594,
+ "step": 3534
+ },
+ {
+ "epoch": 49.09781659388646,
+ "grad_norm": 0.011282820254564285,
+ "learning_rate": 0.0006,
+ "loss": 5.218020439147949,
+ "step": 3535
+ },
+ {
+ "epoch": 49.1117903930131,
+ "grad_norm": 0.012153448536992073,
+ "learning_rate": 0.0006,
+ "loss": 5.134698867797852,
+ "step": 3536
+ },
+ {
+ "epoch": 49.12576419213974,
+ "grad_norm": 0.011920543387532234,
+ "learning_rate": 0.0006,
+ "loss": 5.141620635986328,
+ "step": 3537
+ },
+ {
+ "epoch": 49.13973799126637,
+ "grad_norm": 0.010949329473078251,
+ "learning_rate": 0.0006,
+ "loss": 5.116259574890137,
+ "step": 3538
+ },
+ {
+ "epoch": 49.15371179039301,
+ "grad_norm": 0.012362448498606682,
+ "learning_rate": 0.0006,
+ "loss": 5.090177059173584,
+ "step": 3539
+ },
+ {
+ "epoch": 49.16768558951965,
+ "grad_norm": 0.011575071141123772,
+ "learning_rate": 0.0006,
+ "loss": 5.093441009521484,
+ "step": 3540
+ },
+ {
+ "epoch": 49.18165938864629,
+ "grad_norm": 0.011126236990094185,
+ "learning_rate": 0.0006,
+ "loss": 5.060235023498535,
+ "step": 3541
+ },
+ {
+ "epoch": 49.19563318777293,
+ "grad_norm": 0.0107401292771101,
+ "learning_rate": 0.0006,
+ "loss": 5.140704154968262,
+ "step": 3542
+ },
+ {
+ "epoch": 49.209606986899566,
+ "grad_norm": 0.011202906258404255,
+ "learning_rate": 0.0006,
+ "loss": 5.10855770111084,
+ "step": 3543
+ },
+ {
+ "epoch": 49.223580786026204,
+ "grad_norm": 0.011083454824984074,
+ "learning_rate": 0.0006,
+ "loss": 5.092751502990723,
+ "step": 3544
+ },
+ {
+ "epoch": 49.237554585152836,
+ "grad_norm": 0.009523800574243069,
+ "learning_rate": 0.0006,
+ "loss": 5.13733434677124,
+ "step": 3545
+ },
+ {
+ "epoch": 49.251528384279474,
+ "grad_norm": 0.011298353783786297,
+ "learning_rate": 0.0006,
+ "loss": 5.036048889160156,
+ "step": 3546
+ },
+ {
+ "epoch": 49.26550218340611,
+ "grad_norm": 0.010676483623683453,
+ "learning_rate": 0.0006,
+ "loss": 5.051901340484619,
+ "step": 3547
+ },
+ {
+ "epoch": 49.27947598253275,
+ "grad_norm": 0.011463399045169353,
+ "learning_rate": 0.0006,
+ "loss": 4.983170986175537,
+ "step": 3548
+ },
+ {
+ "epoch": 49.29344978165939,
+ "grad_norm": 0.010603402741253376,
+ "learning_rate": 0.0006,
+ "loss": 5.112311363220215,
+ "step": 3549
+ },
+ {
+ "epoch": 49.30742358078603,
+ "grad_norm": 0.010535670444369316,
+ "learning_rate": 0.0006,
+ "loss": 5.085227966308594,
+ "step": 3550
+ },
+ {
+ "epoch": 49.32139737991266,
+ "grad_norm": 0.011843928135931492,
+ "learning_rate": 0.0006,
+ "loss": 5.07575798034668,
+ "step": 3551
+ },
+ {
+ "epoch": 49.3353711790393,
+ "grad_norm": 0.010587952099740505,
+ "learning_rate": 0.0006,
+ "loss": 5.019237041473389,
+ "step": 3552
+ },
+ {
+ "epoch": 49.34934497816594,
+ "grad_norm": 0.012285659089684486,
+ "learning_rate": 0.0006,
+ "loss": 5.144603729248047,
+ "step": 3553
+ },
+ {
+ "epoch": 49.36331877729258,
+ "grad_norm": 0.013119431212544441,
+ "learning_rate": 0.0006,
+ "loss": 5.035256862640381,
+ "step": 3554
+ },
+ {
+ "epoch": 49.377292576419215,
+ "grad_norm": 0.012138537131249905,
+ "learning_rate": 0.0006,
+ "loss": 4.996667861938477,
+ "step": 3555
+ },
+ {
+ "epoch": 49.391266375545854,
+ "grad_norm": 0.012142804451286793,
+ "learning_rate": 0.0006,
+ "loss": 4.95058536529541,
+ "step": 3556
+ },
+ {
+ "epoch": 49.40524017467249,
+ "grad_norm": 0.012205676175653934,
+ "learning_rate": 0.0006,
+ "loss": 5.096931457519531,
+ "step": 3557
+ },
+ {
+ "epoch": 49.419213973799124,
+ "grad_norm": 0.012086744420230389,
+ "learning_rate": 0.0006,
+ "loss": 4.99384880065918,
+ "step": 3558
+ },
+ {
+ "epoch": 49.43318777292576,
+ "grad_norm": 0.012918103486299515,
+ "learning_rate": 0.0006,
+ "loss": 5.031294822692871,
+ "step": 3559
+ },
+ {
+ "epoch": 49.4471615720524,
+ "grad_norm": 0.012006713077425957,
+ "learning_rate": 0.0006,
+ "loss": 5.0547943115234375,
+ "step": 3560
+ },
+ {
+ "epoch": 49.46113537117904,
+ "grad_norm": 0.0120399734005332,
+ "learning_rate": 0.0006,
+ "loss": 5.014238357543945,
+ "step": 3561
+ },
+ {
+ "epoch": 49.47510917030568,
+ "grad_norm": 0.013306990265846252,
+ "learning_rate": 0.0006,
+ "loss": 4.972925186157227,
+ "step": 3562
+ },
+ {
+ "epoch": 49.48908296943232,
+ "grad_norm": 0.013878636993467808,
+ "learning_rate": 0.0006,
+ "loss": 5.079033851623535,
+ "step": 3563
+ },
+ {
+ "epoch": 49.50305676855895,
+ "grad_norm": 0.012454566545784473,
+ "learning_rate": 0.0006,
+ "loss": 4.93708610534668,
+ "step": 3564
+ },
+ {
+ "epoch": 49.51703056768559,
+ "grad_norm": 0.014075061306357384,
+ "learning_rate": 0.0006,
+ "loss": 4.965582370758057,
+ "step": 3565
+ },
+ {
+ "epoch": 49.531004366812226,
+ "grad_norm": 0.012525617145001888,
+ "learning_rate": 0.0006,
+ "loss": 4.968148708343506,
+ "step": 3566
+ },
+ {
+ "epoch": 49.544978165938865,
+ "grad_norm": 0.012418225407600403,
+ "learning_rate": 0.0006,
+ "loss": 5.06901741027832,
+ "step": 3567
+ },
+ {
+ "epoch": 49.5589519650655,
+ "grad_norm": 0.015034239739179611,
+ "learning_rate": 0.0006,
+ "loss": 4.984277725219727,
+ "step": 3568
+ },
+ {
+ "epoch": 49.57292576419214,
+ "grad_norm": 0.014329628087580204,
+ "learning_rate": 0.0006,
+ "loss": 4.951558589935303,
+ "step": 3569
+ },
+ {
+ "epoch": 49.58689956331878,
+ "grad_norm": 0.017769070342183113,
+ "learning_rate": 0.0006,
+ "loss": 4.966532230377197,
+ "step": 3570
+ },
+ {
+ "epoch": 49.60087336244541,
+ "grad_norm": 0.01909453794360161,
+ "learning_rate": 0.0006,
+ "loss": 4.919890880584717,
+ "step": 3571
+ },
+ {
+ "epoch": 49.61484716157205,
+ "grad_norm": 0.017519688233733177,
+ "learning_rate": 0.0006,
+ "loss": 5.031286716461182,
+ "step": 3572
+ },
+ {
+ "epoch": 49.62882096069869,
+ "grad_norm": 0.017143897712230682,
+ "learning_rate": 0.0006,
+ "loss": 4.969764709472656,
+ "step": 3573
+ },
+ {
+ "epoch": 49.64279475982533,
+ "grad_norm": 0.01688285917043686,
+ "learning_rate": 0.0006,
+ "loss": 5.0733442306518555,
+ "step": 3574
+ },
+ {
+ "epoch": 49.65676855895197,
+ "grad_norm": 0.018547311425209045,
+ "learning_rate": 0.0006,
+ "loss": 4.980863571166992,
+ "step": 3575
+ },
+ {
+ "epoch": 49.670742358078606,
+ "grad_norm": 0.018877137452363968,
+ "learning_rate": 0.0006,
+ "loss": 4.954646110534668,
+ "step": 3576
+ },
+ {
+ "epoch": 49.68471615720524,
+ "grad_norm": 0.018681600689888,
+ "learning_rate": 0.0006,
+ "loss": 4.959702014923096,
+ "step": 3577
+ },
+ {
+ "epoch": 49.698689956331876,
+ "grad_norm": 0.018947336822748184,
+ "learning_rate": 0.0006,
+ "loss": 5.070502758026123,
+ "step": 3578
+ },
+ {
+ "epoch": 49.712663755458514,
+ "grad_norm": 0.018817484378814697,
+ "learning_rate": 0.0006,
+ "loss": 5.056508541107178,
+ "step": 3579
+ },
+ {
+ "epoch": 49.72663755458515,
+ "grad_norm": 0.01931724324822426,
+ "learning_rate": 0.0006,
+ "loss": 4.922451972961426,
+ "step": 3580
+ },
+ {
+ "epoch": 49.74061135371179,
+ "grad_norm": 0.022047819569706917,
+ "learning_rate": 0.0006,
+ "loss": 5.053170204162598,
+ "step": 3581
+ },
+ {
+ "epoch": 49.75458515283843,
+ "grad_norm": 0.02279801294207573,
+ "learning_rate": 0.0006,
+ "loss": 4.8768439292907715,
+ "step": 3582
+ },
+ {
+ "epoch": 49.76855895196506,
+ "grad_norm": 0.020661601796746254,
+ "learning_rate": 0.0006,
+ "loss": 4.907928466796875,
+ "step": 3583
+ },
+ {
+ "epoch": 49.7825327510917,
+ "grad_norm": 0.020479509606957436,
+ "learning_rate": 0.0006,
+ "loss": 5.040694236755371,
+ "step": 3584
+ },
+ {
+ "epoch": 49.79650655021834,
+ "grad_norm": 0.021654162555933,
+ "learning_rate": 0.0006,
+ "loss": 5.018434524536133,
+ "step": 3585
+ },
+ {
+ "epoch": 49.81048034934498,
+ "grad_norm": 0.022486824542284012,
+ "learning_rate": 0.0006,
+ "loss": 5.086538314819336,
+ "step": 3586
+ },
+ {
+ "epoch": 49.82445414847162,
+ "grad_norm": 0.02141369879245758,
+ "learning_rate": 0.0006,
+ "loss": 5.035694122314453,
+ "step": 3587
+ },
+ {
+ "epoch": 49.838427947598255,
+ "grad_norm": 0.015926677733659744,
+ "learning_rate": 0.0006,
+ "loss": 4.905038833618164,
+ "step": 3588
+ },
+ {
+ "epoch": 49.852401746724894,
+ "grad_norm": 0.015775298699736595,
+ "learning_rate": 0.0006,
+ "loss": 5.003902435302734,
+ "step": 3589
+ },
+ {
+ "epoch": 49.866375545851525,
+ "grad_norm": 0.01872118003666401,
+ "learning_rate": 0.0006,
+ "loss": 4.891060829162598,
+ "step": 3590
+ },
+ {
+ "epoch": 49.880349344978164,
+ "grad_norm": 0.02066192403435707,
+ "learning_rate": 0.0006,
+ "loss": 5.002897262573242,
+ "step": 3591
+ },
+ {
+ "epoch": 49.8943231441048,
+ "grad_norm": 0.02348714880645275,
+ "learning_rate": 0.0006,
+ "loss": 4.9189958572387695,
+ "step": 3592
+ },
+ {
+ "epoch": 49.90829694323144,
+ "grad_norm": 0.020402103662490845,
+ "learning_rate": 0.0006,
+ "loss": 4.992661476135254,
+ "step": 3593
+ },
+ {
+ "epoch": 49.92227074235808,
+ "grad_norm": 0.020466072484850883,
+ "learning_rate": 0.0006,
+ "loss": 4.962968826293945,
+ "step": 3594
+ },
+ {
+ "epoch": 49.93624454148472,
+ "grad_norm": 0.01922314241528511,
+ "learning_rate": 0.0006,
+ "loss": 4.897730827331543,
+ "step": 3595
+ },
+ {
+ "epoch": 49.95021834061135,
+ "grad_norm": 0.017228754237294197,
+ "learning_rate": 0.0006,
+ "loss": 4.964842796325684,
+ "step": 3596
+ },
+ {
+ "epoch": 49.96419213973799,
+ "grad_norm": 0.016021743416786194,
+ "learning_rate": 0.0006,
+ "loss": 4.910758972167969,
+ "step": 3597
+ },
+ {
+ "epoch": 49.97816593886463,
+ "grad_norm": 0.017985103651881218,
+ "learning_rate": 0.0006,
+ "loss": 4.916072368621826,
+ "step": 3598
+ },
+ {
+ "epoch": 49.992139737991266,
+ "grad_norm": 0.018780240789055824,
+ "learning_rate": 0.0006,
+ "loss": 4.94658088684082,
+ "step": 3599
+ },
+ {
+ "epoch": 50.0,
+ "grad_norm": 0.01776314526796341,
+ "learning_rate": 0.0006,
+ "loss": 4.969357967376709,
+ "step": 3600
+ },
+ {
+ "epoch": 50.0,
+ "eval_loss": 5.141158580780029,
+ "eval_runtime": 56.2332,
+ "eval_samples_per_second": 43.426,
+ "eval_steps_per_second": 1.369,
+ "step": 3600
+ },
+ {
+ "epoch": 50.01397379912664,
+ "grad_norm": 0.016308696940541267,
+ "learning_rate": 0.0006,
+ "loss": 4.924664497375488,
+ "step": 3601
+ },
+ {
+ "epoch": 50.02794759825328,
+ "grad_norm": 0.018810199573636055,
+ "learning_rate": 0.0006,
+ "loss": 4.82293701171875,
+ "step": 3602
+ },
+ {
+ "epoch": 50.041921397379916,
+ "grad_norm": 0.02227160707116127,
+ "learning_rate": 0.0006,
+ "loss": 4.9612298011779785,
+ "step": 3603
+ },
+ {
+ "epoch": 50.05589519650655,
+ "grad_norm": 0.025266462936997414,
+ "learning_rate": 0.0006,
+ "loss": 4.963835716247559,
+ "step": 3604
+ },
+ {
+ "epoch": 50.069868995633186,
+ "grad_norm": 0.028588181361556053,
+ "learning_rate": 0.0006,
+ "loss": 4.992118835449219,
+ "step": 3605
+ },
+ {
+ "epoch": 50.083842794759825,
+ "grad_norm": 0.02739487960934639,
+ "learning_rate": 0.0006,
+ "loss": 4.915547847747803,
+ "step": 3606
+ },
+ {
+ "epoch": 50.09781659388646,
+ "grad_norm": 0.027000300586223602,
+ "learning_rate": 0.0006,
+ "loss": 4.883801460266113,
+ "step": 3607
+ },
+ {
+ "epoch": 50.1117903930131,
+ "grad_norm": 0.023468483239412308,
+ "learning_rate": 0.0006,
+ "loss": 5.006168365478516,
+ "step": 3608
+ },
+ {
+ "epoch": 50.12576419213974,
+ "grad_norm": 0.026753634214401245,
+ "learning_rate": 0.0006,
+ "loss": 4.92526912689209,
+ "step": 3609
+ },
+ {
+ "epoch": 50.13973799126637,
+ "grad_norm": 0.026816733181476593,
+ "learning_rate": 0.0006,
+ "loss": 4.847546100616455,
+ "step": 3610
+ },
+ {
+ "epoch": 50.15371179039301,
+ "grad_norm": 0.027362391352653503,
+ "learning_rate": 0.0006,
+ "loss": 4.862548351287842,
+ "step": 3611
+ },
+ {
+ "epoch": 50.16768558951965,
+ "grad_norm": 0.03450845181941986,
+ "learning_rate": 0.0006,
+ "loss": 4.963202953338623,
+ "step": 3612
+ },
+ {
+ "epoch": 50.18165938864629,
+ "grad_norm": 0.03801408037543297,
+ "learning_rate": 0.0006,
+ "loss": 4.770240306854248,
+ "step": 3613
+ },
+ {
+ "epoch": 50.19563318777293,
+ "grad_norm": 0.03579140082001686,
+ "learning_rate": 0.0006,
+ "loss": 4.797981262207031,
+ "step": 3614
+ },
+ {
+ "epoch": 50.209606986899566,
+ "grad_norm": 0.029393760487437248,
+ "learning_rate": 0.0006,
+ "loss": 4.798581600189209,
+ "step": 3615
+ },
+ {
+ "epoch": 50.223580786026204,
+ "grad_norm": 0.026111142709851265,
+ "learning_rate": 0.0006,
+ "loss": 4.990712642669678,
+ "step": 3616
+ },
+ {
+ "epoch": 50.237554585152836,
+ "grad_norm": 0.02213387005031109,
+ "learning_rate": 0.0006,
+ "loss": 4.865842342376709,
+ "step": 3617
+ },
+ {
+ "epoch": 50.251528384279474,
+ "grad_norm": 0.025450684130191803,
+ "learning_rate": 0.0006,
+ "loss": 4.903746128082275,
+ "step": 3618
+ },
+ {
+ "epoch": 50.26550218340611,
+ "grad_norm": 0.020915117114782333,
+ "learning_rate": 0.0006,
+ "loss": 4.951366424560547,
+ "step": 3619
+ },
+ {
+ "epoch": 50.27947598253275,
+ "grad_norm": 0.020972849801182747,
+ "learning_rate": 0.0006,
+ "loss": 4.783311367034912,
+ "step": 3620
+ },
+ {
+ "epoch": 50.29344978165939,
+ "grad_norm": 0.016956273466348648,
+ "learning_rate": 0.0006,
+ "loss": 4.955767631530762,
+ "step": 3621
+ },
+ {
+ "epoch": 50.30742358078603,
+ "grad_norm": 0.018045252189040184,
+ "learning_rate": 0.0006,
+ "loss": 4.869322776794434,
+ "step": 3622
+ },
+ {
+ "epoch": 50.32139737991266,
+ "grad_norm": 0.018199188634753227,
+ "learning_rate": 0.0006,
+ "loss": 4.95680046081543,
+ "step": 3623
+ },
+ {
+ "epoch": 50.3353711790393,
+ "grad_norm": 0.016688082367181778,
+ "learning_rate": 0.0006,
+ "loss": 4.783535957336426,
+ "step": 3624
+ },
+ {
+ "epoch": 50.34934497816594,
+ "grad_norm": 0.017540352419018745,
+ "learning_rate": 0.0006,
+ "loss": 4.8236083984375,
+ "step": 3625
+ },
+ {
+ "epoch": 50.36331877729258,
+ "grad_norm": 0.018054187297821045,
+ "learning_rate": 0.0006,
+ "loss": 4.764655113220215,
+ "step": 3626
+ },
+ {
+ "epoch": 50.377292576419215,
+ "grad_norm": 0.01741131953895092,
+ "learning_rate": 0.0006,
+ "loss": 4.875613689422607,
+ "step": 3627
+ },
+ {
+ "epoch": 50.391266375545854,
+ "grad_norm": 0.01682269014418125,
+ "learning_rate": 0.0006,
+ "loss": 4.818509101867676,
+ "step": 3628
+ },
+ {
+ "epoch": 50.40524017467249,
+ "grad_norm": 0.016652341932058334,
+ "learning_rate": 0.0006,
+ "loss": 4.974020481109619,
+ "step": 3629
+ },
+ {
+ "epoch": 50.419213973799124,
+ "grad_norm": 0.02035282365977764,
+ "learning_rate": 0.0006,
+ "loss": 4.893037796020508,
+ "step": 3630
+ },
+ {
+ "epoch": 50.43318777292576,
+ "grad_norm": 0.019907061010599136,
+ "learning_rate": 0.0006,
+ "loss": 4.848119735717773,
+ "step": 3631
+ },
+ {
+ "epoch": 50.4471615720524,
+ "grad_norm": 0.023861493915319443,
+ "learning_rate": 0.0006,
+ "loss": 4.819353103637695,
+ "step": 3632
+ },
+ {
+ "epoch": 50.46113537117904,
+ "grad_norm": 0.025254026055336,
+ "learning_rate": 0.0006,
+ "loss": 4.769312858581543,
+ "step": 3633
+ },
+ {
+ "epoch": 50.47510917030568,
+ "grad_norm": 0.026154937222599983,
+ "learning_rate": 0.0006,
+ "loss": 4.7586774826049805,
+ "step": 3634
+ },
+ {
+ "epoch": 50.48908296943232,
+ "grad_norm": 0.026342090219259262,
+ "learning_rate": 0.0006,
+ "loss": 4.861218452453613,
+ "step": 3635
+ },
+ {
+ "epoch": 50.50305676855895,
+ "grad_norm": 0.028169559314846992,
+ "learning_rate": 0.0006,
+ "loss": 4.818380355834961,
+ "step": 3636
+ },
+ {
+ "epoch": 50.51703056768559,
+ "grad_norm": 0.02820264920592308,
+ "learning_rate": 0.0006,
+ "loss": 4.892871856689453,
+ "step": 3637
+ },
+ {
+ "epoch": 50.531004366812226,
+ "grad_norm": 0.03345080465078354,
+ "learning_rate": 0.0006,
+ "loss": 4.811084747314453,
+ "step": 3638
+ },
+ {
+ "epoch": 50.544978165938865,
+ "grad_norm": 0.04005736857652664,
+ "learning_rate": 0.0006,
+ "loss": 4.765410900115967,
+ "step": 3639
+ },
+ {
+ "epoch": 50.5589519650655,
+ "grad_norm": 0.04333870857954025,
+ "learning_rate": 0.0006,
+ "loss": 4.956520080566406,
+ "step": 3640
+ },
+ {
+ "epoch": 50.57292576419214,
+ "grad_norm": 0.04216952994465828,
+ "learning_rate": 0.0006,
+ "loss": 4.793441295623779,
+ "step": 3641
+ },
+ {
+ "epoch": 50.58689956331878,
+ "grad_norm": 0.02936951443552971,
+ "learning_rate": 0.0006,
+ "loss": 4.8447957038879395,
+ "step": 3642
+ },
+ {
+ "epoch": 50.60087336244541,
+ "grad_norm": 0.026937026530504227,
+ "learning_rate": 0.0006,
+ "loss": 4.962215423583984,
+ "step": 3643
+ },
+ {
+ "epoch": 50.61484716157205,
+ "grad_norm": 0.02628890983760357,
+ "learning_rate": 0.0006,
+ "loss": 4.883061408996582,
+ "step": 3644
+ },
+ {
+ "epoch": 50.62882096069869,
+ "grad_norm": 0.025984834879636765,
+ "learning_rate": 0.0006,
+ "loss": 4.768662452697754,
+ "step": 3645
+ },
+ {
+ "epoch": 50.64279475982533,
+ "grad_norm": 0.023998012766242027,
+ "learning_rate": 0.0006,
+ "loss": 4.851325035095215,
+ "step": 3646
+ },
+ {
+ "epoch": 50.65676855895197,
+ "grad_norm": 0.02143184095621109,
+ "learning_rate": 0.0006,
+ "loss": 4.8274688720703125,
+ "step": 3647
+ },
+ {
+ "epoch": 50.670742358078606,
+ "grad_norm": 0.02488900162279606,
+ "learning_rate": 0.0006,
+ "loss": 4.831140518188477,
+ "step": 3648
+ },
+ {
+ "epoch": 50.68471615720524,
+ "grad_norm": 0.02330533228814602,
+ "learning_rate": 0.0006,
+ "loss": 4.850462913513184,
+ "step": 3649
+ },
+ {
+ "epoch": 50.698689956331876,
+ "grad_norm": 0.02121943049132824,
+ "learning_rate": 0.0006,
+ "loss": 4.870253086090088,
+ "step": 3650
+ },
+ {
+ "epoch": 50.712663755458514,
+ "grad_norm": 0.02080935798585415,
+ "learning_rate": 0.0006,
+ "loss": 4.798770904541016,
+ "step": 3651
+ },
+ {
+ "epoch": 50.72663755458515,
+ "grad_norm": 0.019047029316425323,
+ "learning_rate": 0.0006,
+ "loss": 4.8427300453186035,
+ "step": 3652
+ },
+ {
+ "epoch": 50.74061135371179,
+ "grad_norm": 0.01824103854596615,
+ "learning_rate": 0.0006,
+ "loss": 4.820980548858643,
+ "step": 3653
+ },
+ {
+ "epoch": 50.75458515283843,
+ "grad_norm": 0.018129967153072357,
+ "learning_rate": 0.0006,
+ "loss": 4.946732997894287,
+ "step": 3654
+ },
+ {
+ "epoch": 50.76855895196506,
+ "grad_norm": 0.01784815452992916,
+ "learning_rate": 0.0006,
+ "loss": 4.829516410827637,
+ "step": 3655
+ },
+ {
+ "epoch": 50.7825327510917,
+ "grad_norm": 0.020021555945277214,
+ "learning_rate": 0.0006,
+ "loss": 4.780428409576416,
+ "step": 3656
+ },
+ {
+ "epoch": 50.79650655021834,
+ "grad_norm": 0.01863253302872181,
+ "learning_rate": 0.0006,
+ "loss": 4.791567802429199,
+ "step": 3657
+ },
+ {
+ "epoch": 50.81048034934498,
+ "grad_norm": 0.019151683896780014,
+ "learning_rate": 0.0006,
+ "loss": 4.824892997741699,
+ "step": 3658
+ },
+ {
+ "epoch": 50.82445414847162,
+ "grad_norm": 0.01891648955643177,
+ "learning_rate": 0.0006,
+ "loss": 4.789605140686035,
+ "step": 3659
+ },
+ {
+ "epoch": 50.838427947598255,
+ "grad_norm": 0.016019510105252266,
+ "learning_rate": 0.0006,
+ "loss": 4.779555320739746,
+ "step": 3660
+ },
+ {
+ "epoch": 50.852401746724894,
+ "grad_norm": 0.016096491366624832,
+ "learning_rate": 0.0006,
+ "loss": 4.753842353820801,
+ "step": 3661
+ },
+ {
+ "epoch": 50.866375545851525,
+ "grad_norm": 0.01693476364016533,
+ "learning_rate": 0.0006,
+ "loss": 4.768477439880371,
+ "step": 3662
+ },
+ {
+ "epoch": 50.880349344978164,
+ "grad_norm": 0.01749575324356556,
+ "learning_rate": 0.0006,
+ "loss": 4.779143333435059,
+ "step": 3663
+ },
+ {
+ "epoch": 50.8943231441048,
+ "grad_norm": 0.016851043328642845,
+ "learning_rate": 0.0006,
+ "loss": 4.780387878417969,
+ "step": 3664
+ },
+ {
+ "epoch": 50.90829694323144,
+ "grad_norm": 0.019045202061533928,
+ "learning_rate": 0.0006,
+ "loss": 4.769913673400879,
+ "step": 3665
+ },
+ {
+ "epoch": 50.92227074235808,
+ "grad_norm": 0.01954123005270958,
+ "learning_rate": 0.0006,
+ "loss": 4.769009590148926,
+ "step": 3666
+ },
+ {
+ "epoch": 50.93624454148472,
+ "grad_norm": 0.022527096793055534,
+ "learning_rate": 0.0006,
+ "loss": 4.789215087890625,
+ "step": 3667
+ },
+ {
+ "epoch": 50.95021834061135,
+ "grad_norm": 0.02871634066104889,
+ "learning_rate": 0.0006,
+ "loss": 4.749727725982666,
+ "step": 3668
+ },
+ {
+ "epoch": 50.96419213973799,
+ "grad_norm": 0.04045053571462631,
+ "learning_rate": 0.0006,
+ "loss": 4.909038543701172,
+ "step": 3669
+ },
+ {
+ "epoch": 50.97816593886463,
+ "grad_norm": 0.04156879335641861,
+ "learning_rate": 0.0006,
+ "loss": 4.876831531524658,
+ "step": 3670
+ },
+ {
+ "epoch": 50.992139737991266,
+ "grad_norm": 0.02621397376060486,
+ "learning_rate": 0.0006,
+ "loss": 4.797173500061035,
+ "step": 3671
+ },
+ {
+ "epoch": 51.0,
+ "grad_norm": 0.03389597684144974,
+ "learning_rate": 0.0006,
+ "loss": 4.8428850173950195,
+ "step": 3672
+ },
+ {
+ "epoch": 51.0,
+ "eval_loss": 4.989261627197266,
+ "eval_runtime": 56.358,
+ "eval_samples_per_second": 43.33,
+ "eval_steps_per_second": 1.366,
+ "step": 3672
+ },
+ {
+ "epoch": 51.01397379912664,
+ "grad_norm": 0.03216972202062607,
+ "learning_rate": 0.0006,
+ "loss": 4.755102634429932,
+ "step": 3673
+ },
+ {
+ "epoch": 51.02794759825328,
+ "grad_norm": 0.03190374746918678,
+ "learning_rate": 0.0006,
+ "loss": 4.720424652099609,
+ "step": 3674
+ },
+ {
+ "epoch": 51.041921397379916,
+ "grad_norm": 0.031727299094200134,
+ "learning_rate": 0.0006,
+ "loss": 4.803277969360352,
+ "step": 3675
+ },
+ {
+ "epoch": 51.05589519650655,
+ "grad_norm": 0.02473079040646553,
+ "learning_rate": 0.0006,
+ "loss": 4.761087417602539,
+ "step": 3676
+ },
+ {
+ "epoch": 51.069868995633186,
+ "grad_norm": 0.020360209047794342,
+ "learning_rate": 0.0006,
+ "loss": 4.753884792327881,
+ "step": 3677
+ },
+ {
+ "epoch": 51.083842794759825,
+ "grad_norm": 0.02078923024237156,
+ "learning_rate": 0.0006,
+ "loss": 4.647130489349365,
+ "step": 3678
+ },
+ {
+ "epoch": 51.09781659388646,
+ "grad_norm": 0.019387200474739075,
+ "learning_rate": 0.0006,
+ "loss": 4.74135160446167,
+ "step": 3679
+ },
+ {
+ "epoch": 51.1117903930131,
+ "grad_norm": 0.01874949410557747,
+ "learning_rate": 0.0006,
+ "loss": 4.795329570770264,
+ "step": 3680
+ },
+ {
+ "epoch": 51.12576419213974,
+ "grad_norm": 0.019331837072968483,
+ "learning_rate": 0.0006,
+ "loss": 4.897967338562012,
+ "step": 3681
+ },
+ {
+ "epoch": 51.13973799126637,
+ "grad_norm": 0.017066050320863724,
+ "learning_rate": 0.0006,
+ "loss": 4.825351715087891,
+ "step": 3682
+ },
+ {
+ "epoch": 51.15371179039301,
+ "grad_norm": 0.01757810264825821,
+ "learning_rate": 0.0006,
+ "loss": 4.719335079193115,
+ "step": 3683
+ },
+ {
+ "epoch": 51.16768558951965,
+ "grad_norm": 0.018174266442656517,
+ "learning_rate": 0.0006,
+ "loss": 4.732680320739746,
+ "step": 3684
+ },
+ {
+ "epoch": 51.18165938864629,
+ "grad_norm": 0.018497727811336517,
+ "learning_rate": 0.0006,
+ "loss": 4.731050491333008,
+ "step": 3685
+ },
+ {
+ "epoch": 51.19563318777293,
+ "grad_norm": 0.0186338871717453,
+ "learning_rate": 0.0006,
+ "loss": 4.748035430908203,
+ "step": 3686
+ },
+ {
+ "epoch": 51.209606986899566,
+ "grad_norm": 0.018531711772084236,
+ "learning_rate": 0.0006,
+ "loss": 4.9007463455200195,
+ "step": 3687
+ },
+ {
+ "epoch": 51.223580786026204,
+ "grad_norm": 0.018221354112029076,
+ "learning_rate": 0.0006,
+ "loss": 4.84077262878418,
+ "step": 3688
+ },
+ {
+ "epoch": 51.237554585152836,
+ "grad_norm": 0.0203656405210495,
+ "learning_rate": 0.0006,
+ "loss": 4.73009729385376,
+ "step": 3689
+ },
+ {
+ "epoch": 51.251528384279474,
+ "grad_norm": 0.01964893378317356,
+ "learning_rate": 0.0006,
+ "loss": 4.772583961486816,
+ "step": 3690
+ },
+ {
+ "epoch": 51.26550218340611,
+ "grad_norm": 0.019606510177254677,
+ "learning_rate": 0.0006,
+ "loss": 4.599812030792236,
+ "step": 3691
+ },
+ {
+ "epoch": 51.27947598253275,
+ "grad_norm": 0.020501255989074707,
+ "learning_rate": 0.0006,
+ "loss": 4.809898853302002,
+ "step": 3692
+ },
+ {
+ "epoch": 51.29344978165939,
+ "grad_norm": 0.01870838552713394,
+ "learning_rate": 0.0006,
+ "loss": 4.741898059844971,
+ "step": 3693
+ },
+ {
+ "epoch": 51.30742358078603,
+ "grad_norm": 0.01802559196949005,
+ "learning_rate": 0.0006,
+ "loss": 4.746562480926514,
+ "step": 3694
+ },
+ {
+ "epoch": 51.32139737991266,
+ "grad_norm": 0.02053048275411129,
+ "learning_rate": 0.0006,
+ "loss": 4.69932746887207,
+ "step": 3695
+ },
+ {
+ "epoch": 51.3353711790393,
+ "grad_norm": 0.02661825157701969,
+ "learning_rate": 0.0006,
+ "loss": 4.68748664855957,
+ "step": 3696
+ },
+ {
+ "epoch": 51.34934497816594,
+ "grad_norm": 0.028269147500395775,
+ "learning_rate": 0.0006,
+ "loss": 4.731565475463867,
+ "step": 3697
+ },
+ {
+ "epoch": 51.36331877729258,
+ "grad_norm": 0.02650964818894863,
+ "learning_rate": 0.0006,
+ "loss": 4.733857154846191,
+ "step": 3698
+ },
+ {
+ "epoch": 51.377292576419215,
+ "grad_norm": 0.023696066811680794,
+ "learning_rate": 0.0006,
+ "loss": 4.812362194061279,
+ "step": 3699
+ },
+ {
+ "epoch": 51.391266375545854,
+ "grad_norm": 0.02365880273282528,
+ "learning_rate": 0.0006,
+ "loss": 4.733426094055176,
+ "step": 3700
+ },
+ {
+ "epoch": 51.40524017467249,
+ "grad_norm": 0.026249447837471962,
+ "learning_rate": 0.0006,
+ "loss": 4.715622901916504,
+ "step": 3701
+ },
+ {
+ "epoch": 51.419213973799124,
+ "grad_norm": 0.03301899880170822,
+ "learning_rate": 0.0006,
+ "loss": 4.7223663330078125,
+ "step": 3702
+ },
+ {
+ "epoch": 51.43318777292576,
+ "grad_norm": 0.044686660170555115,
+ "learning_rate": 0.0006,
+ "loss": 4.8121161460876465,
+ "step": 3703
+ },
+ {
+ "epoch": 51.4471615720524,
+ "grad_norm": 0.04733522981405258,
+ "learning_rate": 0.0006,
+ "loss": 4.841394424438477,
+ "step": 3704
+ },
+ {
+ "epoch": 51.46113537117904,
+ "grad_norm": 0.0395209975540638,
+ "learning_rate": 0.0006,
+ "loss": 4.773837089538574,
+ "step": 3705
+ },
+ {
+ "epoch": 51.47510917030568,
+ "grad_norm": 0.032790958881378174,
+ "learning_rate": 0.0006,
+ "loss": 4.833066463470459,
+ "step": 3706
+ },
+ {
+ "epoch": 51.48908296943232,
+ "grad_norm": 0.030578427016735077,
+ "learning_rate": 0.0006,
+ "loss": 4.685190200805664,
+ "step": 3707
+ },
+ {
+ "epoch": 51.50305676855895,
+ "grad_norm": 0.031619083136320114,
+ "learning_rate": 0.0006,
+ "loss": 4.740378379821777,
+ "step": 3708
+ },
+ {
+ "epoch": 51.51703056768559,
+ "grad_norm": 0.02385825850069523,
+ "learning_rate": 0.0006,
+ "loss": 4.812911510467529,
+ "step": 3709
+ },
+ {
+ "epoch": 51.531004366812226,
+ "grad_norm": 0.023523425683379173,
+ "learning_rate": 0.0006,
+ "loss": 4.669580459594727,
+ "step": 3710
+ },
+ {
+ "epoch": 51.544978165938865,
+ "grad_norm": 0.021316280588507652,
+ "learning_rate": 0.0006,
+ "loss": 4.76888370513916,
+ "step": 3711
+ },
+ {
+ "epoch": 51.5589519650655,
+ "grad_norm": 0.024663880467414856,
+ "learning_rate": 0.0006,
+ "loss": 4.702592372894287,
+ "step": 3712
+ },
+ {
+ "epoch": 51.57292576419214,
+ "grad_norm": 0.024173280224204063,
+ "learning_rate": 0.0006,
+ "loss": 4.710604667663574,
+ "step": 3713
+ },
+ {
+ "epoch": 51.58689956331878,
+ "grad_norm": 0.02754114381968975,
+ "learning_rate": 0.0006,
+ "loss": 4.746356964111328,
+ "step": 3714
+ },
+ {
+ "epoch": 51.60087336244541,
+ "grad_norm": 0.027903622016310692,
+ "learning_rate": 0.0006,
+ "loss": 4.7546491622924805,
+ "step": 3715
+ },
+ {
+ "epoch": 51.61484716157205,
+ "grad_norm": 0.024871617555618286,
+ "learning_rate": 0.0006,
+ "loss": 4.605746269226074,
+ "step": 3716
+ },
+ {
+ "epoch": 51.62882096069869,
+ "grad_norm": 0.021239031106233597,
+ "learning_rate": 0.0006,
+ "loss": 4.6942458152771,
+ "step": 3717
+ },
+ {
+ "epoch": 51.64279475982533,
+ "grad_norm": 0.01925622671842575,
+ "learning_rate": 0.0006,
+ "loss": 4.593558311462402,
+ "step": 3718
+ },
+ {
+ "epoch": 51.65676855895197,
+ "grad_norm": 0.01956283673644066,
+ "learning_rate": 0.0006,
+ "loss": 4.631172180175781,
+ "step": 3719
+ },
+ {
+ "epoch": 51.670742358078606,
+ "grad_norm": 0.02269871160387993,
+ "learning_rate": 0.0006,
+ "loss": 4.729236602783203,
+ "step": 3720
+ },
+ {
+ "epoch": 51.68471615720524,
+ "grad_norm": 0.02253916673362255,
+ "learning_rate": 0.0006,
+ "loss": 4.7191033363342285,
+ "step": 3721
+ },
+ {
+ "epoch": 51.698689956331876,
+ "grad_norm": 0.026231050491333008,
+ "learning_rate": 0.0006,
+ "loss": 4.811850547790527,
+ "step": 3722
+ },
+ {
+ "epoch": 51.712663755458514,
+ "grad_norm": 0.02558821812272072,
+ "learning_rate": 0.0006,
+ "loss": 4.681466102600098,
+ "step": 3723
+ },
+ {
+ "epoch": 51.72663755458515,
+ "grad_norm": 0.02124209702014923,
+ "learning_rate": 0.0006,
+ "loss": 4.693664073944092,
+ "step": 3724
+ },
+ {
+ "epoch": 51.74061135371179,
+ "grad_norm": 0.016638319939374924,
+ "learning_rate": 0.0006,
+ "loss": 4.771440029144287,
+ "step": 3725
+ },
+ {
+ "epoch": 51.75458515283843,
+ "grad_norm": 0.018048347905278206,
+ "learning_rate": 0.0006,
+ "loss": 4.643252372741699,
+ "step": 3726
+ },
+ {
+ "epoch": 51.76855895196506,
+ "grad_norm": 0.019904717803001404,
+ "learning_rate": 0.0006,
+ "loss": 4.66530704498291,
+ "step": 3727
+ },
+ {
+ "epoch": 51.7825327510917,
+ "grad_norm": 0.0194614976644516,
+ "learning_rate": 0.0006,
+ "loss": 4.711943626403809,
+ "step": 3728
+ },
+ {
+ "epoch": 51.79650655021834,
+ "grad_norm": 0.019175555557012558,
+ "learning_rate": 0.0006,
+ "loss": 4.721931457519531,
+ "step": 3729
+ },
+ {
+ "epoch": 51.81048034934498,
+ "grad_norm": 0.02093447372317314,
+ "learning_rate": 0.0006,
+ "loss": 4.633059024810791,
+ "step": 3730
+ },
+ {
+ "epoch": 51.82445414847162,
+ "grad_norm": 0.01930283196270466,
+ "learning_rate": 0.0006,
+ "loss": 4.659373760223389,
+ "step": 3731
+ },
+ {
+ "epoch": 51.838427947598255,
+ "grad_norm": 0.015350676141679287,
+ "learning_rate": 0.0006,
+ "loss": 4.763708114624023,
+ "step": 3732
+ },
+ {
+ "epoch": 51.852401746724894,
+ "grad_norm": 0.014139448292553425,
+ "learning_rate": 0.0006,
+ "loss": 4.753891468048096,
+ "step": 3733
+ },
+ {
+ "epoch": 51.866375545851525,
+ "grad_norm": 0.016215872019529343,
+ "learning_rate": 0.0006,
+ "loss": 4.658967018127441,
+ "step": 3734
+ },
+ {
+ "epoch": 51.880349344978164,
+ "grad_norm": 0.017328619956970215,
+ "learning_rate": 0.0006,
+ "loss": 4.696579933166504,
+ "step": 3735
+ },
+ {
+ "epoch": 51.8943231441048,
+ "grad_norm": 0.018103396520018578,
+ "learning_rate": 0.0006,
+ "loss": 4.627464771270752,
+ "step": 3736
+ },
+ {
+ "epoch": 51.90829694323144,
+ "grad_norm": 0.01715235412120819,
+ "learning_rate": 0.0006,
+ "loss": 4.7368011474609375,
+ "step": 3737
+ },
+ {
+ "epoch": 51.92227074235808,
+ "grad_norm": 0.014790094457566738,
+ "learning_rate": 0.0006,
+ "loss": 4.629441738128662,
+ "step": 3738
+ },
+ {
+ "epoch": 51.93624454148472,
+ "grad_norm": 0.017114296555519104,
+ "learning_rate": 0.0006,
+ "loss": 4.761256694793701,
+ "step": 3739
+ },
+ {
+ "epoch": 51.95021834061135,
+ "grad_norm": 0.019009845331311226,
+ "learning_rate": 0.0006,
+ "loss": 4.789155006408691,
+ "step": 3740
+ },
+ {
+ "epoch": 51.96419213973799,
+ "grad_norm": 0.019696692004799843,
+ "learning_rate": 0.0006,
+ "loss": 4.721269607543945,
+ "step": 3741
+ },
+ {
+ "epoch": 51.97816593886463,
+ "grad_norm": 0.020542822778224945,
+ "learning_rate": 0.0006,
+ "loss": 4.687038898468018,
+ "step": 3742
+ },
+ {
+ "epoch": 51.992139737991266,
+ "grad_norm": 0.021267160773277283,
+ "learning_rate": 0.0006,
+ "loss": 4.747860908508301,
+ "step": 3743
+ },
+ {
+ "epoch": 52.0,
+ "grad_norm": 0.025049567222595215,
+ "learning_rate": 0.0006,
+ "loss": 4.8473100662231445,
+ "step": 3744
+ },
+ {
+ "epoch": 52.0,
+ "eval_loss": 4.943043231964111,
+ "eval_runtime": 56.4929,
+ "eval_samples_per_second": 43.227,
+ "eval_steps_per_second": 1.363,
+ "step": 3744
+ },
+ {
+ "epoch": 52.01397379912664,
+ "grad_norm": 0.03054986707866192,
+ "learning_rate": 0.0006,
+ "loss": 4.856931686401367,
+ "step": 3745
+ },
+ {
+ "epoch": 52.02794759825328,
+ "grad_norm": 0.03838375583291054,
+ "learning_rate": 0.0006,
+ "loss": 4.699807167053223,
+ "step": 3746
+ },
+ {
+ "epoch": 52.041921397379916,
+ "grad_norm": 0.03383488208055496,
+ "learning_rate": 0.0006,
+ "loss": 4.6626787185668945,
+ "step": 3747
+ },
+ {
+ "epoch": 52.05589519650655,
+ "grad_norm": 0.0201561376452446,
+ "learning_rate": 0.0006,
+ "loss": 4.621831893920898,
+ "step": 3748
+ },
+ {
+ "epoch": 52.069868995633186,
+ "grad_norm": 0.028749438002705574,
+ "learning_rate": 0.0006,
+ "loss": 4.682540416717529,
+ "step": 3749
+ },
+ {
+ "epoch": 52.083842794759825,
+ "grad_norm": 0.02681174874305725,
+ "learning_rate": 0.0006,
+ "loss": 4.629997253417969,
+ "step": 3750
+ },
+ {
+ "epoch": 52.09781659388646,
+ "grad_norm": 0.024063318967819214,
+ "learning_rate": 0.0006,
+ "loss": 4.573021411895752,
+ "step": 3751
+ },
+ {
+ "epoch": 52.1117903930131,
+ "grad_norm": 0.02324330434203148,
+ "learning_rate": 0.0006,
+ "loss": 4.739465236663818,
+ "step": 3752
+ },
+ {
+ "epoch": 52.12576419213974,
+ "grad_norm": 0.0225426834076643,
+ "learning_rate": 0.0006,
+ "loss": 4.7335333824157715,
+ "step": 3753
+ },
+ {
+ "epoch": 52.13973799126637,
+ "grad_norm": 0.019593428820371628,
+ "learning_rate": 0.0006,
+ "loss": 4.704958438873291,
+ "step": 3754
+ },
+ {
+ "epoch": 52.15371179039301,
+ "grad_norm": 0.020777082070708275,
+ "learning_rate": 0.0006,
+ "loss": 4.638696670532227,
+ "step": 3755
+ },
+ {
+ "epoch": 52.16768558951965,
+ "grad_norm": 0.021525604650378227,
+ "learning_rate": 0.0006,
+ "loss": 4.643706798553467,
+ "step": 3756
+ },
+ {
+ "epoch": 52.18165938864629,
+ "grad_norm": 0.018943654373288155,
+ "learning_rate": 0.0006,
+ "loss": 4.5173211097717285,
+ "step": 3757
+ },
+ {
+ "epoch": 52.19563318777293,
+ "grad_norm": 0.01856580562889576,
+ "learning_rate": 0.0006,
+ "loss": 4.678788661956787,
+ "step": 3758
+ },
+ {
+ "epoch": 52.209606986899566,
+ "grad_norm": 0.019417259842157364,
+ "learning_rate": 0.0006,
+ "loss": 4.587244987487793,
+ "step": 3759
+ },
+ {
+ "epoch": 52.223580786026204,
+ "grad_norm": 0.01911911740899086,
+ "learning_rate": 0.0006,
+ "loss": 4.688503265380859,
+ "step": 3760
+ },
+ {
+ "epoch": 52.237554585152836,
+ "grad_norm": 0.021017879247665405,
+ "learning_rate": 0.0006,
+ "loss": 4.65829610824585,
+ "step": 3761
+ },
+ {
+ "epoch": 52.251528384279474,
+ "grad_norm": 0.02048875205218792,
+ "learning_rate": 0.0006,
+ "loss": 4.624055862426758,
+ "step": 3762
+ },
+ {
+ "epoch": 52.26550218340611,
+ "grad_norm": 0.01932772807776928,
+ "learning_rate": 0.0006,
+ "loss": 4.556562900543213,
+ "step": 3763
+ },
+ {
+ "epoch": 52.27947598253275,
+ "grad_norm": 0.01871614344418049,
+ "learning_rate": 0.0006,
+ "loss": 4.581201076507568,
+ "step": 3764
+ },
+ {
+ "epoch": 52.29344978165939,
+ "grad_norm": 0.017253106459975243,
+ "learning_rate": 0.0006,
+ "loss": 4.605219841003418,
+ "step": 3765
+ },
+ {
+ "epoch": 52.30742358078603,
+ "grad_norm": 0.01831238530576229,
+ "learning_rate": 0.0006,
+ "loss": 4.759900093078613,
+ "step": 3766
+ },
+ {
+ "epoch": 52.32139737991266,
+ "grad_norm": 0.017946207895874977,
+ "learning_rate": 0.0006,
+ "loss": 4.5598344802856445,
+ "step": 3767
+ },
+ {
+ "epoch": 52.3353711790393,
+ "grad_norm": 0.01680106669664383,
+ "learning_rate": 0.0006,
+ "loss": 4.605034828186035,
+ "step": 3768
+ },
+ {
+ "epoch": 52.34934497816594,
+ "grad_norm": 0.021482540294528008,
+ "learning_rate": 0.0006,
+ "loss": 4.620411396026611,
+ "step": 3769
+ },
+ {
+ "epoch": 52.36331877729258,
+ "grad_norm": 0.02210252732038498,
+ "learning_rate": 0.0006,
+ "loss": 4.667577743530273,
+ "step": 3770
+ },
+ {
+ "epoch": 52.377292576419215,
+ "grad_norm": 0.02237485907971859,
+ "learning_rate": 0.0006,
+ "loss": 4.628812789916992,
+ "step": 3771
+ },
+ {
+ "epoch": 52.391266375545854,
+ "grad_norm": 0.02196541801095009,
+ "learning_rate": 0.0006,
+ "loss": 4.572988510131836,
+ "step": 3772
+ },
+ {
+ "epoch": 52.40524017467249,
+ "grad_norm": 0.01930229924619198,
+ "learning_rate": 0.0006,
+ "loss": 4.588339805603027,
+ "step": 3773
+ },
+ {
+ "epoch": 52.419213973799124,
+ "grad_norm": 0.02222434990108013,
+ "learning_rate": 0.0006,
+ "loss": 4.766482353210449,
+ "step": 3774
+ },
+ {
+ "epoch": 52.43318777292576,
+ "grad_norm": 0.028806306421756744,
+ "learning_rate": 0.0006,
+ "loss": 4.617892742156982,
+ "step": 3775
+ },
+ {
+ "epoch": 52.4471615720524,
+ "grad_norm": 0.03306042402982712,
+ "learning_rate": 0.0006,
+ "loss": 4.665334701538086,
+ "step": 3776
+ },
+ {
+ "epoch": 52.46113537117904,
+ "grad_norm": 0.0358956903219223,
+ "learning_rate": 0.0006,
+ "loss": 4.566668510437012,
+ "step": 3777
+ },
+ {
+ "epoch": 52.47510917030568,
+ "grad_norm": 0.030406497418880463,
+ "learning_rate": 0.0006,
+ "loss": 4.513612270355225,
+ "step": 3778
+ },
+ {
+ "epoch": 52.48908296943232,
+ "grad_norm": 0.030295712873339653,
+ "learning_rate": 0.0006,
+ "loss": 4.640189170837402,
+ "step": 3779
+ },
+ {
+ "epoch": 52.50305676855895,
+ "grad_norm": 0.024981003254652023,
+ "learning_rate": 0.0006,
+ "loss": 4.57558012008667,
+ "step": 3780
+ },
+ {
+ "epoch": 52.51703056768559,
+ "grad_norm": 0.024990713223814964,
+ "learning_rate": 0.0006,
+ "loss": 4.708589553833008,
+ "step": 3781
+ },
+ {
+ "epoch": 52.531004366812226,
+ "grad_norm": 0.025034775957465172,
+ "learning_rate": 0.0006,
+ "loss": 4.70468282699585,
+ "step": 3782
+ },
+ {
+ "epoch": 52.544978165938865,
+ "grad_norm": 0.028243759647011757,
+ "learning_rate": 0.0006,
+ "loss": 4.712060451507568,
+ "step": 3783
+ },
+ {
+ "epoch": 52.5589519650655,
+ "grad_norm": 0.026984544470906258,
+ "learning_rate": 0.0006,
+ "loss": 4.505004405975342,
+ "step": 3784
+ },
+ {
+ "epoch": 52.57292576419214,
+ "grad_norm": 0.050843529403209686,
+ "learning_rate": 0.0006,
+ "loss": 4.593271255493164,
+ "step": 3785
+ },
+ {
+ "epoch": 52.58689956331878,
+ "grad_norm": 0.25422802567481995,
+ "learning_rate": 0.0006,
+ "loss": 4.7178802490234375,
+ "step": 3786
+ },
+ {
+ "epoch": 52.60087336244541,
+ "grad_norm": 0.6078406572341919,
+ "learning_rate": 0.0006,
+ "loss": 6.144372940063477,
+ "step": 3787
+ },
+ {
+ "epoch": 52.61484716157205,
+ "grad_norm": 0.25933101773262024,
+ "learning_rate": 0.0006,
+ "loss": 5.971083641052246,
+ "step": 3788
+ },
+ {
+ "epoch": 52.62882096069869,
+ "grad_norm": 0.294657826423645,
+ "learning_rate": 0.0006,
+ "loss": 6.204090118408203,
+ "step": 3789
+ },
+ {
+ "epoch": 52.64279475982533,
+ "grad_norm": 0.23641522228717804,
+ "learning_rate": 0.0006,
+ "loss": 6.869574546813965,
+ "step": 3790
+ },
+ {
+ "epoch": 52.65676855895197,
+ "grad_norm": 0.23717311024665833,
+ "learning_rate": 0.0006,
+ "loss": 6.739034175872803,
+ "step": 3791
+ },
+ {
+ "epoch": 52.670742358078606,
+ "grad_norm": 0.10300362855195999,
+ "learning_rate": 0.0006,
+ "loss": 6.49191951751709,
+ "step": 3792
+ },
+ {
+ "epoch": 52.68471615720524,
+ "grad_norm": 0.1484968513250351,
+ "learning_rate": 0.0006,
+ "loss": 6.6021270751953125,
+ "step": 3793
+ },
+ {
+ "epoch": 52.698689956331876,
+ "grad_norm": 0.08052115142345428,
+ "learning_rate": 0.0006,
+ "loss": 6.432796478271484,
+ "step": 3794
+ },
+ {
+ "epoch": 52.712663755458514,
+ "grad_norm": 0.06869253516197205,
+ "learning_rate": 0.0006,
+ "loss": 6.212784767150879,
+ "step": 3795
+ },
+ {
+ "epoch": 52.72663755458515,
+ "grad_norm": 0.05589582398533821,
+ "learning_rate": 0.0006,
+ "loss": 6.168288707733154,
+ "step": 3796
+ },
+ {
+ "epoch": 52.74061135371179,
+ "grad_norm": 0.04840940609574318,
+ "learning_rate": 0.0006,
+ "loss": 6.163174629211426,
+ "step": 3797
+ },
+ {
+ "epoch": 52.75458515283843,
+ "grad_norm": 0.04578601196408272,
+ "learning_rate": 0.0006,
+ "loss": 6.038390636444092,
+ "step": 3798
+ },
+ {
+ "epoch": 52.76855895196506,
+ "grad_norm": 0.04102053865790367,
+ "learning_rate": 0.0006,
+ "loss": 5.929539680480957,
+ "step": 3799
+ },
+ {
+ "epoch": 52.7825327510917,
+ "grad_norm": 0.038398485630750656,
+ "learning_rate": 0.0006,
+ "loss": 6.007378101348877,
+ "step": 3800
+ },
+ {
+ "epoch": 52.79650655021834,
+ "grad_norm": 0.03466642275452614,
+ "learning_rate": 0.0006,
+ "loss": 5.767291069030762,
+ "step": 3801
+ },
+ {
+ "epoch": 52.81048034934498,
+ "grad_norm": 0.03328905999660492,
+ "learning_rate": 0.0006,
+ "loss": 5.785032272338867,
+ "step": 3802
+ },
+ {
+ "epoch": 52.82445414847162,
+ "grad_norm": 0.03186681121587753,
+ "learning_rate": 0.0006,
+ "loss": 5.6942057609558105,
+ "step": 3803
+ },
+ {
+ "epoch": 52.838427947598255,
+ "grad_norm": 0.030553551390767097,
+ "learning_rate": 0.0006,
+ "loss": 5.680774688720703,
+ "step": 3804
+ },
+ {
+ "epoch": 52.852401746724894,
+ "grad_norm": 0.02578839473426342,
+ "learning_rate": 0.0006,
+ "loss": 5.703461647033691,
+ "step": 3805
+ },
+ {
+ "epoch": 52.866375545851525,
+ "grad_norm": 0.026261312887072563,
+ "learning_rate": 0.0006,
+ "loss": 5.642549514770508,
+ "step": 3806
+ },
+ {
+ "epoch": 52.880349344978164,
+ "grad_norm": 0.02782430872321129,
+ "learning_rate": 0.0006,
+ "loss": 5.563597679138184,
+ "step": 3807
+ },
+ {
+ "epoch": 52.8943231441048,
+ "grad_norm": 0.024467799812555313,
+ "learning_rate": 0.0006,
+ "loss": 5.583395957946777,
+ "step": 3808
+ },
+ {
+ "epoch": 52.90829694323144,
+ "grad_norm": 0.022602897137403488,
+ "learning_rate": 0.0006,
+ "loss": 5.50742769241333,
+ "step": 3809
+ },
+ {
+ "epoch": 52.92227074235808,
+ "grad_norm": 0.024292029440402985,
+ "learning_rate": 0.0006,
+ "loss": 5.367804527282715,
+ "step": 3810
+ },
+ {
+ "epoch": 52.93624454148472,
+ "grad_norm": 0.021062295883893967,
+ "learning_rate": 0.0006,
+ "loss": 5.423705101013184,
+ "step": 3811
+ },
+ {
+ "epoch": 52.95021834061135,
+ "grad_norm": 0.022226469591259956,
+ "learning_rate": 0.0006,
+ "loss": 5.344039440155029,
+ "step": 3812
+ },
+ {
+ "epoch": 52.96419213973799,
+ "grad_norm": 0.023691752925515175,
+ "learning_rate": 0.0006,
+ "loss": 5.2581892013549805,
+ "step": 3813
+ },
+ {
+ "epoch": 52.97816593886463,
+ "grad_norm": 0.02181999757885933,
+ "learning_rate": 0.0006,
+ "loss": 5.2710113525390625,
+ "step": 3814
+ },
+ {
+ "epoch": 52.992139737991266,
+ "grad_norm": 0.025295857340097427,
+ "learning_rate": 0.0006,
+ "loss": 5.278355598449707,
+ "step": 3815
+ },
+ {
+ "epoch": 53.0,
+ "grad_norm": 0.02216828241944313,
+ "learning_rate": 0.0006,
+ "loss": 5.127945899963379,
+ "step": 3816
+ },
+ {
+ "epoch": 53.0,
+ "eval_loss": 5.3411736488342285,
+ "eval_runtime": 56.2837,
+ "eval_samples_per_second": 43.387,
+ "eval_steps_per_second": 1.368,
+ "step": 3816
+ },
+ {
+ "epoch": 53.01397379912664,
+ "grad_norm": 0.022496026009321213,
+ "learning_rate": 0.0006,
+ "loss": 5.19655704498291,
+ "step": 3817
+ },
+ {
+ "epoch": 53.02794759825328,
+ "grad_norm": 0.03201168775558472,
+ "learning_rate": 0.0006,
+ "loss": 5.094621181488037,
+ "step": 3818
+ },
+ {
+ "epoch": 53.041921397379916,
+ "grad_norm": 0.02683178335428238,
+ "learning_rate": 0.0006,
+ "loss": 5.19899845123291,
+ "step": 3819
+ },
+ {
+ "epoch": 53.05589519650655,
+ "grad_norm": 0.022659169510006905,
+ "learning_rate": 0.0006,
+ "loss": 5.184103012084961,
+ "step": 3820
+ },
+ {
+ "epoch": 53.069868995633186,
+ "grad_norm": 0.024784443899989128,
+ "learning_rate": 0.0006,
+ "loss": 5.124286651611328,
+ "step": 3821
+ },
+ {
+ "epoch": 53.083842794759825,
+ "grad_norm": 0.028710370883345604,
+ "learning_rate": 0.0006,
+ "loss": 5.130426406860352,
+ "step": 3822
+ },
+ {
+ "epoch": 53.09781659388646,
+ "grad_norm": 0.02366088517010212,
+ "learning_rate": 0.0006,
+ "loss": 5.1453399658203125,
+ "step": 3823
+ },
+ {
+ "epoch": 53.1117903930131,
+ "grad_norm": 0.01813938096165657,
+ "learning_rate": 0.0006,
+ "loss": 5.05271053314209,
+ "step": 3824
+ },
+ {
+ "epoch": 53.12576419213974,
+ "grad_norm": 0.02051435597240925,
+ "learning_rate": 0.0006,
+ "loss": 5.023038864135742,
+ "step": 3825
+ },
+ {
+ "epoch": 53.13973799126637,
+ "grad_norm": 0.02185754105448723,
+ "learning_rate": 0.0006,
+ "loss": 4.8950018882751465,
+ "step": 3826
+ },
+ {
+ "epoch": 53.15371179039301,
+ "grad_norm": 0.017223743721842766,
+ "learning_rate": 0.0006,
+ "loss": 5.078621864318848,
+ "step": 3827
+ },
+ {
+ "epoch": 53.16768558951965,
+ "grad_norm": 0.01832217164337635,
+ "learning_rate": 0.0006,
+ "loss": 4.921341419219971,
+ "step": 3828
+ },
+ {
+ "epoch": 53.18165938864629,
+ "grad_norm": 0.02114005573093891,
+ "learning_rate": 0.0006,
+ "loss": 4.997865676879883,
+ "step": 3829
+ },
+ {
+ "epoch": 53.19563318777293,
+ "grad_norm": 0.021426431834697723,
+ "learning_rate": 0.0006,
+ "loss": 4.877785682678223,
+ "step": 3830
+ },
+ {
+ "epoch": 53.209606986899566,
+ "grad_norm": 0.021327560767531395,
+ "learning_rate": 0.0006,
+ "loss": 4.847512245178223,
+ "step": 3831
+ },
+ {
+ "epoch": 53.223580786026204,
+ "grad_norm": 0.027825530618429184,
+ "learning_rate": 0.0006,
+ "loss": 5.004823207855225,
+ "step": 3832
+ },
+ {
+ "epoch": 53.237554585152836,
+ "grad_norm": 0.03756481409072876,
+ "learning_rate": 0.0006,
+ "loss": 4.817290306091309,
+ "step": 3833
+ },
+ {
+ "epoch": 53.251528384279474,
+ "grad_norm": 0.04933721572160721,
+ "learning_rate": 0.0006,
+ "loss": 4.863317489624023,
+ "step": 3834
+ },
+ {
+ "epoch": 53.26550218340611,
+ "grad_norm": 0.04588694870471954,
+ "learning_rate": 0.0006,
+ "loss": 4.864742755889893,
+ "step": 3835
+ },
+ {
+ "epoch": 53.27947598253275,
+ "grad_norm": 0.02728264033794403,
+ "learning_rate": 0.0006,
+ "loss": 4.9068756103515625,
+ "step": 3836
+ },
+ {
+ "epoch": 53.29344978165939,
+ "grad_norm": 0.036240726709365845,
+ "learning_rate": 0.0006,
+ "loss": 4.902045249938965,
+ "step": 3837
+ },
+ {
+ "epoch": 53.30742358078603,
+ "grad_norm": 0.02703195810317993,
+ "learning_rate": 0.0006,
+ "loss": 4.908980369567871,
+ "step": 3838
+ },
+ {
+ "epoch": 53.32139737991266,
+ "grad_norm": 0.024894049391150475,
+ "learning_rate": 0.0006,
+ "loss": 4.794848442077637,
+ "step": 3839
+ },
+ {
+ "epoch": 53.3353711790393,
+ "grad_norm": 0.02389993891119957,
+ "learning_rate": 0.0006,
+ "loss": 4.8234734535217285,
+ "step": 3840
+ },
+ {
+ "epoch": 53.34934497816594,
+ "grad_norm": 0.02486400306224823,
+ "learning_rate": 0.0006,
+ "loss": 4.851463794708252,
+ "step": 3841
+ },
+ {
+ "epoch": 53.36331877729258,
+ "grad_norm": 0.02613813802599907,
+ "learning_rate": 0.0006,
+ "loss": 4.860444068908691,
+ "step": 3842
+ },
+ {
+ "epoch": 53.377292576419215,
+ "grad_norm": 0.0229744054377079,
+ "learning_rate": 0.0006,
+ "loss": 4.709606170654297,
+ "step": 3843
+ },
+ {
+ "epoch": 53.391266375545854,
+ "grad_norm": 0.02123042568564415,
+ "learning_rate": 0.0006,
+ "loss": 4.654613494873047,
+ "step": 3844
+ },
+ {
+ "epoch": 53.40524017467249,
+ "grad_norm": 0.017352338880300522,
+ "learning_rate": 0.0006,
+ "loss": 4.769761085510254,
+ "step": 3845
+ },
+ {
+ "epoch": 53.419213973799124,
+ "grad_norm": 0.019808724522590637,
+ "learning_rate": 0.0006,
+ "loss": 4.98514461517334,
+ "step": 3846
+ },
+ {
+ "epoch": 53.43318777292576,
+ "grad_norm": 0.018776152282953262,
+ "learning_rate": 0.0006,
+ "loss": 4.749990463256836,
+ "step": 3847
+ },
+ {
+ "epoch": 53.4471615720524,
+ "grad_norm": 0.02017521858215332,
+ "learning_rate": 0.0006,
+ "loss": 4.757381439208984,
+ "step": 3848
+ },
+ {
+ "epoch": 53.46113537117904,
+ "grad_norm": 0.019509416073560715,
+ "learning_rate": 0.0006,
+ "loss": 4.8111572265625,
+ "step": 3849
+ },
+ {
+ "epoch": 53.47510917030568,
+ "grad_norm": 0.015543580986559391,
+ "learning_rate": 0.0006,
+ "loss": 4.7501678466796875,
+ "step": 3850
+ },
+ {
+ "epoch": 53.48908296943232,
+ "grad_norm": 0.017941651865839958,
+ "learning_rate": 0.0006,
+ "loss": 4.701838493347168,
+ "step": 3851
+ },
+ {
+ "epoch": 53.50305676855895,
+ "grad_norm": 0.01678212359547615,
+ "learning_rate": 0.0006,
+ "loss": 4.787798881530762,
+ "step": 3852
+ },
+ {
+ "epoch": 53.51703056768559,
+ "grad_norm": 0.019050654023885727,
+ "learning_rate": 0.0006,
+ "loss": 4.754859924316406,
+ "step": 3853
+ },
+ {
+ "epoch": 53.531004366812226,
+ "grad_norm": 0.017406078055500984,
+ "learning_rate": 0.0006,
+ "loss": 4.752772808074951,
+ "step": 3854
+ },
+ {
+ "epoch": 53.544978165938865,
+ "grad_norm": 0.01773344725370407,
+ "learning_rate": 0.0006,
+ "loss": 4.651812553405762,
+ "step": 3855
+ },
+ {
+ "epoch": 53.5589519650655,
+ "grad_norm": 0.017221346497535706,
+ "learning_rate": 0.0006,
+ "loss": 4.792244911193848,
+ "step": 3856
+ },
+ {
+ "epoch": 53.57292576419214,
+ "grad_norm": 0.01722758449614048,
+ "learning_rate": 0.0006,
+ "loss": 4.7902960777282715,
+ "step": 3857
+ },
+ {
+ "epoch": 53.58689956331878,
+ "grad_norm": 0.020445192232728004,
+ "learning_rate": 0.0006,
+ "loss": 4.67392110824585,
+ "step": 3858
+ },
+ {
+ "epoch": 53.60087336244541,
+ "grad_norm": 0.02267114631831646,
+ "learning_rate": 0.0006,
+ "loss": 4.796089172363281,
+ "step": 3859
+ },
+ {
+ "epoch": 53.61484716157205,
+ "grad_norm": 0.027129942551255226,
+ "learning_rate": 0.0006,
+ "loss": 4.774303913116455,
+ "step": 3860
+ },
+ {
+ "epoch": 53.62882096069869,
+ "grad_norm": 0.025454839691519737,
+ "learning_rate": 0.0006,
+ "loss": 4.708813190460205,
+ "step": 3861
+ },
+ {
+ "epoch": 53.64279475982533,
+ "grad_norm": 0.019248757511377335,
+ "learning_rate": 0.0006,
+ "loss": 4.760467529296875,
+ "step": 3862
+ },
+ {
+ "epoch": 53.65676855895197,
+ "grad_norm": 0.01791176199913025,
+ "learning_rate": 0.0006,
+ "loss": 4.789203643798828,
+ "step": 3863
+ },
+ {
+ "epoch": 53.670742358078606,
+ "grad_norm": 0.016306612640619278,
+ "learning_rate": 0.0006,
+ "loss": 4.732605934143066,
+ "step": 3864
+ },
+ {
+ "epoch": 53.68471615720524,
+ "grad_norm": 0.017836976796388626,
+ "learning_rate": 0.0006,
+ "loss": 4.691871643066406,
+ "step": 3865
+ },
+ {
+ "epoch": 53.698689956331876,
+ "grad_norm": 0.015102861449122429,
+ "learning_rate": 0.0006,
+ "loss": 4.724156856536865,
+ "step": 3866
+ },
+ {
+ "epoch": 53.712663755458514,
+ "grad_norm": 0.017458142712712288,
+ "learning_rate": 0.0006,
+ "loss": 4.73524808883667,
+ "step": 3867
+ },
+ {
+ "epoch": 53.72663755458515,
+ "grad_norm": 0.01591932773590088,
+ "learning_rate": 0.0006,
+ "loss": 4.705888748168945,
+ "step": 3868
+ },
+ {
+ "epoch": 53.74061135371179,
+ "grad_norm": 0.01713593676686287,
+ "learning_rate": 0.0006,
+ "loss": 4.644516944885254,
+ "step": 3869
+ },
+ {
+ "epoch": 53.75458515283843,
+ "grad_norm": 0.01905885338783264,
+ "learning_rate": 0.0006,
+ "loss": 4.698416233062744,
+ "step": 3870
+ },
+ {
+ "epoch": 53.76855895196506,
+ "grad_norm": 0.01766437292098999,
+ "learning_rate": 0.0006,
+ "loss": 4.6796555519104,
+ "step": 3871
+ },
+ {
+ "epoch": 53.7825327510917,
+ "grad_norm": 0.01552243810147047,
+ "learning_rate": 0.0006,
+ "loss": 4.78138542175293,
+ "step": 3872
+ },
+ {
+ "epoch": 53.79650655021834,
+ "grad_norm": 0.014191331341862679,
+ "learning_rate": 0.0006,
+ "loss": 4.703588485717773,
+ "step": 3873
+ },
+ {
+ "epoch": 53.81048034934498,
+ "grad_norm": 0.015872279182076454,
+ "learning_rate": 0.0006,
+ "loss": 4.705611228942871,
+ "step": 3874
+ },
+ {
+ "epoch": 53.82445414847162,
+ "grad_norm": 0.01787560060620308,
+ "learning_rate": 0.0006,
+ "loss": 4.8227949142456055,
+ "step": 3875
+ },
+ {
+ "epoch": 53.838427947598255,
+ "grad_norm": 0.0158209390938282,
+ "learning_rate": 0.0006,
+ "loss": 4.645227432250977,
+ "step": 3876
+ },
+ {
+ "epoch": 53.852401746724894,
+ "grad_norm": 0.014362678863108158,
+ "learning_rate": 0.0006,
+ "loss": 4.7225422859191895,
+ "step": 3877
+ },
+ {
+ "epoch": 53.866375545851525,
+ "grad_norm": 0.014899161644279957,
+ "learning_rate": 0.0006,
+ "loss": 4.685914039611816,
+ "step": 3878
+ },
+ {
+ "epoch": 53.880349344978164,
+ "grad_norm": 0.01628505438566208,
+ "learning_rate": 0.0006,
+ "loss": 4.674065589904785,
+ "step": 3879
+ },
+ {
+ "epoch": 53.8943231441048,
+ "grad_norm": 0.01873331144452095,
+ "learning_rate": 0.0006,
+ "loss": 4.811744213104248,
+ "step": 3880
+ },
+ {
+ "epoch": 53.90829694323144,
+ "grad_norm": 0.02229069173336029,
+ "learning_rate": 0.0006,
+ "loss": 4.721514701843262,
+ "step": 3881
+ },
+ {
+ "epoch": 53.92227074235808,
+ "grad_norm": 0.028012916445732117,
+ "learning_rate": 0.0006,
+ "loss": 4.773603439331055,
+ "step": 3882
+ },
+ {
+ "epoch": 53.93624454148472,
+ "grad_norm": 0.036322783678770065,
+ "learning_rate": 0.0006,
+ "loss": 4.787134170532227,
+ "step": 3883
+ },
+ {
+ "epoch": 53.95021834061135,
+ "grad_norm": 0.048149142414331436,
+ "learning_rate": 0.0006,
+ "loss": 4.736196517944336,
+ "step": 3884
+ },
+ {
+ "epoch": 53.96419213973799,
+ "grad_norm": 0.051918916404247284,
+ "learning_rate": 0.0006,
+ "loss": 4.686728000640869,
+ "step": 3885
+ },
+ {
+ "epoch": 53.97816593886463,
+ "grad_norm": 0.0481342189013958,
+ "learning_rate": 0.0006,
+ "loss": 4.733730792999268,
+ "step": 3886
+ },
+ {
+ "epoch": 53.992139737991266,
+ "grad_norm": 0.045424818992614746,
+ "learning_rate": 0.0006,
+ "loss": 4.748723983764648,
+ "step": 3887
+ },
+ {
+ "epoch": 54.0,
+ "grad_norm": 0.04153316840529442,
+ "learning_rate": 0.0006,
+ "loss": 4.715220928192139,
+ "step": 3888
+ },
+ {
+ "epoch": 54.0,
+ "eval_loss": 4.964215278625488,
+ "eval_runtime": 56.7757,
+ "eval_samples_per_second": 43.011,
+ "eval_steps_per_second": 1.356,
+ "step": 3888
+ },
+ {
+ "epoch": 54.01397379912664,
+ "grad_norm": 0.03459750860929489,
+ "learning_rate": 0.0006,
+ "loss": 4.661557674407959,
+ "step": 3889
+ },
+ {
+ "epoch": 54.02794759825328,
+ "grad_norm": 0.025409208610653877,
+ "learning_rate": 0.0006,
+ "loss": 4.7065582275390625,
+ "step": 3890
+ },
+ {
+ "epoch": 54.041921397379916,
+ "grad_norm": 0.03202550485730171,
+ "learning_rate": 0.0006,
+ "loss": 4.751070976257324,
+ "step": 3891
+ },
+ {
+ "epoch": 54.05589519650655,
+ "grad_norm": 0.02968018501996994,
+ "learning_rate": 0.0006,
+ "loss": 4.635388374328613,
+ "step": 3892
+ },
+ {
+ "epoch": 54.069868995633186,
+ "grad_norm": 0.02757406234741211,
+ "learning_rate": 0.0006,
+ "loss": 4.706214427947998,
+ "step": 3893
+ },
+ {
+ "epoch": 54.083842794759825,
+ "grad_norm": 0.025223620235919952,
+ "learning_rate": 0.0006,
+ "loss": 4.703792572021484,
+ "step": 3894
+ },
+ {
+ "epoch": 54.09781659388646,
+ "grad_norm": 0.02431778982281685,
+ "learning_rate": 0.0006,
+ "loss": 4.658116340637207,
+ "step": 3895
+ },
+ {
+ "epoch": 54.1117903930131,
+ "grad_norm": 0.02792172320187092,
+ "learning_rate": 0.0006,
+ "loss": 4.638895034790039,
+ "step": 3896
+ },
+ {
+ "epoch": 54.12576419213974,
+ "grad_norm": 0.032253511250019073,
+ "learning_rate": 0.0006,
+ "loss": 4.693902492523193,
+ "step": 3897
+ },
+ {
+ "epoch": 54.13973799126637,
+ "grad_norm": 0.027044931426644325,
+ "learning_rate": 0.0006,
+ "loss": 4.575805187225342,
+ "step": 3898
+ },
+ {
+ "epoch": 54.15371179039301,
+ "grad_norm": 0.023270361125469208,
+ "learning_rate": 0.0006,
+ "loss": 4.67645263671875,
+ "step": 3899
+ },
+ {
+ "epoch": 54.16768558951965,
+ "grad_norm": 0.01929079182446003,
+ "learning_rate": 0.0006,
+ "loss": 4.68062686920166,
+ "step": 3900
+ },
+ {
+ "epoch": 54.18165938864629,
+ "grad_norm": 0.020343473181128502,
+ "learning_rate": 0.0006,
+ "loss": 4.708086967468262,
+ "step": 3901
+ },
+ {
+ "epoch": 54.19563318777293,
+ "grad_norm": 0.0211793202906847,
+ "learning_rate": 0.0006,
+ "loss": 4.677323341369629,
+ "step": 3902
+ },
+ {
+ "epoch": 54.209606986899566,
+ "grad_norm": 0.020341217517852783,
+ "learning_rate": 0.0006,
+ "loss": 4.760937690734863,
+ "step": 3903
+ },
+ {
+ "epoch": 54.223580786026204,
+ "grad_norm": 0.019421054050326347,
+ "learning_rate": 0.0006,
+ "loss": 4.508834362030029,
+ "step": 3904
+ },
+ {
+ "epoch": 54.237554585152836,
+ "grad_norm": 0.016701679676771164,
+ "learning_rate": 0.0006,
+ "loss": 4.664399147033691,
+ "step": 3905
+ },
+ {
+ "epoch": 54.251528384279474,
+ "grad_norm": 0.016561470925807953,
+ "learning_rate": 0.0006,
+ "loss": 4.742990493774414,
+ "step": 3906
+ },
+ {
+ "epoch": 54.26550218340611,
+ "grad_norm": 0.017446085810661316,
+ "learning_rate": 0.0006,
+ "loss": 4.63038444519043,
+ "step": 3907
+ },
+ {
+ "epoch": 54.27947598253275,
+ "grad_norm": 0.014340086840093136,
+ "learning_rate": 0.0006,
+ "loss": 4.688791275024414,
+ "step": 3908
+ },
+ {
+ "epoch": 54.29344978165939,
+ "grad_norm": 0.013580698519945145,
+ "learning_rate": 0.0006,
+ "loss": 4.625302314758301,
+ "step": 3909
+ },
+ {
+ "epoch": 54.30742358078603,
+ "grad_norm": 0.014931539073586464,
+ "learning_rate": 0.0006,
+ "loss": 4.6167497634887695,
+ "step": 3910
+ },
+ {
+ "epoch": 54.32139737991266,
+ "grad_norm": 0.013492150232195854,
+ "learning_rate": 0.0006,
+ "loss": 4.657521724700928,
+ "step": 3911
+ },
+ {
+ "epoch": 54.3353711790393,
+ "grad_norm": 0.015151413157582283,
+ "learning_rate": 0.0006,
+ "loss": 4.634521484375,
+ "step": 3912
+ },
+ {
+ "epoch": 54.34934497816594,
+ "grad_norm": 0.01302040833979845,
+ "learning_rate": 0.0006,
+ "loss": 4.718928337097168,
+ "step": 3913
+ },
+ {
+ "epoch": 54.36331877729258,
+ "grad_norm": 0.013433529995381832,
+ "learning_rate": 0.0006,
+ "loss": 4.584803581237793,
+ "step": 3914
+ },
+ {
+ "epoch": 54.377292576419215,
+ "grad_norm": 0.012987039051949978,
+ "learning_rate": 0.0006,
+ "loss": 4.6393585205078125,
+ "step": 3915
+ },
+ {
+ "epoch": 54.391266375545854,
+ "grad_norm": 0.013695075176656246,
+ "learning_rate": 0.0006,
+ "loss": 4.556313514709473,
+ "step": 3916
+ },
+ {
+ "epoch": 54.40524017467249,
+ "grad_norm": 0.013718255795538425,
+ "learning_rate": 0.0006,
+ "loss": 4.576223850250244,
+ "step": 3917
+ },
+ {
+ "epoch": 54.419213973799124,
+ "grad_norm": 0.013699112460017204,
+ "learning_rate": 0.0006,
+ "loss": 4.63889217376709,
+ "step": 3918
+ },
+ {
+ "epoch": 54.43318777292576,
+ "grad_norm": 0.013516264036297798,
+ "learning_rate": 0.0006,
+ "loss": 4.669532299041748,
+ "step": 3919
+ },
+ {
+ "epoch": 54.4471615720524,
+ "grad_norm": 0.013597294688224792,
+ "learning_rate": 0.0006,
+ "loss": 4.6644287109375,
+ "step": 3920
+ },
+ {
+ "epoch": 54.46113537117904,
+ "grad_norm": 0.013185849413275719,
+ "learning_rate": 0.0006,
+ "loss": 4.636900901794434,
+ "step": 3921
+ },
+ {
+ "epoch": 54.47510917030568,
+ "grad_norm": 0.014311990700662136,
+ "learning_rate": 0.0006,
+ "loss": 4.6047587394714355,
+ "step": 3922
+ },
+ {
+ "epoch": 54.48908296943232,
+ "grad_norm": 0.012426027096807957,
+ "learning_rate": 0.0006,
+ "loss": 4.6161394119262695,
+ "step": 3923
+ },
+ {
+ "epoch": 54.50305676855895,
+ "grad_norm": 0.01283493172377348,
+ "learning_rate": 0.0006,
+ "loss": 4.584190368652344,
+ "step": 3924
+ },
+ {
+ "epoch": 54.51703056768559,
+ "grad_norm": 0.01339943427592516,
+ "learning_rate": 0.0006,
+ "loss": 4.624184608459473,
+ "step": 3925
+ },
+ {
+ "epoch": 54.531004366812226,
+ "grad_norm": 0.012859724462032318,
+ "learning_rate": 0.0006,
+ "loss": 4.6174635887146,
+ "step": 3926
+ },
+ {
+ "epoch": 54.544978165938865,
+ "grad_norm": 0.014420023187994957,
+ "learning_rate": 0.0006,
+ "loss": 4.536544322967529,
+ "step": 3927
+ },
+ {
+ "epoch": 54.5589519650655,
+ "grad_norm": 0.014237185008823872,
+ "learning_rate": 0.0006,
+ "loss": 4.681514739990234,
+ "step": 3928
+ },
+ {
+ "epoch": 54.57292576419214,
+ "grad_norm": 0.013340328820049763,
+ "learning_rate": 0.0006,
+ "loss": 4.579689979553223,
+ "step": 3929
+ },
+ {
+ "epoch": 54.58689956331878,
+ "grad_norm": 0.013894530944526196,
+ "learning_rate": 0.0006,
+ "loss": 4.638958930969238,
+ "step": 3930
+ },
+ {
+ "epoch": 54.60087336244541,
+ "grad_norm": 0.016181141138076782,
+ "learning_rate": 0.0006,
+ "loss": 4.573508262634277,
+ "step": 3931
+ },
+ {
+ "epoch": 54.61484716157205,
+ "grad_norm": 0.013221992179751396,
+ "learning_rate": 0.0006,
+ "loss": 4.6229047775268555,
+ "step": 3932
+ },
+ {
+ "epoch": 54.62882096069869,
+ "grad_norm": 0.013241561129689217,
+ "learning_rate": 0.0006,
+ "loss": 4.641152858734131,
+ "step": 3933
+ },
+ {
+ "epoch": 54.64279475982533,
+ "grad_norm": 0.016237538307905197,
+ "learning_rate": 0.0006,
+ "loss": 4.6421217918396,
+ "step": 3934
+ },
+ {
+ "epoch": 54.65676855895197,
+ "grad_norm": 0.020866507664322853,
+ "learning_rate": 0.0006,
+ "loss": 4.646058082580566,
+ "step": 3935
+ },
+ {
+ "epoch": 54.670742358078606,
+ "grad_norm": 0.022182893007993698,
+ "learning_rate": 0.0006,
+ "loss": 4.554501533508301,
+ "step": 3936
+ },
+ {
+ "epoch": 54.68471615720524,
+ "grad_norm": 0.02254868485033512,
+ "learning_rate": 0.0006,
+ "loss": 4.563869953155518,
+ "step": 3937
+ },
+ {
+ "epoch": 54.698689956331876,
+ "grad_norm": 0.019730940461158752,
+ "learning_rate": 0.0006,
+ "loss": 4.592165946960449,
+ "step": 3938
+ },
+ {
+ "epoch": 54.712663755458514,
+ "grad_norm": 0.015146561898291111,
+ "learning_rate": 0.0006,
+ "loss": 4.6228742599487305,
+ "step": 3939
+ },
+ {
+ "epoch": 54.72663755458515,
+ "grad_norm": 0.01824716106057167,
+ "learning_rate": 0.0006,
+ "loss": 4.716236114501953,
+ "step": 3940
+ },
+ {
+ "epoch": 54.74061135371179,
+ "grad_norm": 0.017730550840497017,
+ "learning_rate": 0.0006,
+ "loss": 4.650232315063477,
+ "step": 3941
+ },
+ {
+ "epoch": 54.75458515283843,
+ "grad_norm": 0.014875595457851887,
+ "learning_rate": 0.0006,
+ "loss": 4.6377458572387695,
+ "step": 3942
+ },
+ {
+ "epoch": 54.76855895196506,
+ "grad_norm": 0.01613209955394268,
+ "learning_rate": 0.0006,
+ "loss": 4.5768537521362305,
+ "step": 3943
+ },
+ {
+ "epoch": 54.7825327510917,
+ "grad_norm": 0.017311280593276024,
+ "learning_rate": 0.0006,
+ "loss": 4.625744342803955,
+ "step": 3944
+ },
+ {
+ "epoch": 54.79650655021834,
+ "grad_norm": 0.018770085647702217,
+ "learning_rate": 0.0006,
+ "loss": 4.613787651062012,
+ "step": 3945
+ },
+ {
+ "epoch": 54.81048034934498,
+ "grad_norm": 0.0165257528424263,
+ "learning_rate": 0.0006,
+ "loss": 4.5991315841674805,
+ "step": 3946
+ },
+ {
+ "epoch": 54.82445414847162,
+ "grad_norm": 0.017433857545256615,
+ "learning_rate": 0.0006,
+ "loss": 4.513872146606445,
+ "step": 3947
+ },
+ {
+ "epoch": 54.838427947598255,
+ "grad_norm": 0.020126905292272568,
+ "learning_rate": 0.0006,
+ "loss": 4.616185188293457,
+ "step": 3948
+ },
+ {
+ "epoch": 54.852401746724894,
+ "grad_norm": 0.01819358579814434,
+ "learning_rate": 0.0006,
+ "loss": 4.535691261291504,
+ "step": 3949
+ },
+ {
+ "epoch": 54.866375545851525,
+ "grad_norm": 0.01899927295744419,
+ "learning_rate": 0.0006,
+ "loss": 4.595199108123779,
+ "step": 3950
+ },
+ {
+ "epoch": 54.880349344978164,
+ "grad_norm": 0.02005566656589508,
+ "learning_rate": 0.0006,
+ "loss": 4.633289337158203,
+ "step": 3951
+ },
+ {
+ "epoch": 54.8943231441048,
+ "grad_norm": 0.021570220589637756,
+ "learning_rate": 0.0006,
+ "loss": 4.592503547668457,
+ "step": 3952
+ },
+ {
+ "epoch": 54.90829694323144,
+ "grad_norm": 0.01883506216108799,
+ "learning_rate": 0.0006,
+ "loss": 4.6597795486450195,
+ "step": 3953
+ },
+ {
+ "epoch": 54.92227074235808,
+ "grad_norm": 0.018070021644234657,
+ "learning_rate": 0.0006,
+ "loss": 4.601474285125732,
+ "step": 3954
+ },
+ {
+ "epoch": 54.93624454148472,
+ "grad_norm": 0.018090378493070602,
+ "learning_rate": 0.0006,
+ "loss": 4.569846153259277,
+ "step": 3955
+ },
+ {
+ "epoch": 54.95021834061135,
+ "grad_norm": 0.018751563504338264,
+ "learning_rate": 0.0006,
+ "loss": 4.646064281463623,
+ "step": 3956
+ },
+ {
+ "epoch": 54.96419213973799,
+ "grad_norm": 0.017512254416942596,
+ "learning_rate": 0.0006,
+ "loss": 4.606670379638672,
+ "step": 3957
+ },
+ {
+ "epoch": 54.97816593886463,
+ "grad_norm": 0.018771110102534294,
+ "learning_rate": 0.0006,
+ "loss": 4.604315757751465,
+ "step": 3958
+ },
+ {
+ "epoch": 54.992139737991266,
+ "grad_norm": 0.01611141301691532,
+ "learning_rate": 0.0006,
+ "loss": 4.643245220184326,
+ "step": 3959
+ },
+ {
+ "epoch": 55.0,
+ "grad_norm": 0.0156728383153677,
+ "learning_rate": 0.0006,
+ "loss": 4.6739044189453125,
+ "step": 3960
+ },
+ {
+ "epoch": 55.0,
+ "eval_loss": 4.851032257080078,
+ "eval_runtime": 56.8699,
+ "eval_samples_per_second": 42.94,
+ "eval_steps_per_second": 1.354,
+ "step": 3960
+ },
+ {
+ "epoch": 55.01397379912664,
+ "grad_norm": 0.016214672476053238,
+ "learning_rate": 0.0006,
+ "loss": 4.571749687194824,
+ "step": 3961
+ },
+ {
+ "epoch": 55.02794759825328,
+ "grad_norm": 0.01649155467748642,
+ "learning_rate": 0.0006,
+ "loss": 4.553592205047607,
+ "step": 3962
+ },
+ {
+ "epoch": 55.041921397379916,
+ "grad_norm": 0.016793379560112953,
+ "learning_rate": 0.0006,
+ "loss": 4.650628566741943,
+ "step": 3963
+ },
+ {
+ "epoch": 55.05589519650655,
+ "grad_norm": 0.0191937405616045,
+ "learning_rate": 0.0006,
+ "loss": 4.620532989501953,
+ "step": 3964
+ },
+ {
+ "epoch": 55.069868995633186,
+ "grad_norm": 0.020976556465029716,
+ "learning_rate": 0.0006,
+ "loss": 4.6329545974731445,
+ "step": 3965
+ },
+ {
+ "epoch": 55.083842794759825,
+ "grad_norm": 0.023450477048754692,
+ "learning_rate": 0.0006,
+ "loss": 4.60828971862793,
+ "step": 3966
+ },
+ {
+ "epoch": 55.09781659388646,
+ "grad_norm": 0.026778312399983406,
+ "learning_rate": 0.0006,
+ "loss": 4.617481708526611,
+ "step": 3967
+ },
+ {
+ "epoch": 55.1117903930131,
+ "grad_norm": 0.02333568036556244,
+ "learning_rate": 0.0006,
+ "loss": 4.580722332000732,
+ "step": 3968
+ },
+ {
+ "epoch": 55.12576419213974,
+ "grad_norm": 0.018491310998797417,
+ "learning_rate": 0.0006,
+ "loss": 4.590953350067139,
+ "step": 3969
+ },
+ {
+ "epoch": 55.13973799126637,
+ "grad_norm": 0.020339803770184517,
+ "learning_rate": 0.0006,
+ "loss": 4.518374443054199,
+ "step": 3970
+ },
+ {
+ "epoch": 55.15371179039301,
+ "grad_norm": 0.01759706623852253,
+ "learning_rate": 0.0006,
+ "loss": 4.560522079467773,
+ "step": 3971
+ },
+ {
+ "epoch": 55.16768558951965,
+ "grad_norm": 0.015875980257987976,
+ "learning_rate": 0.0006,
+ "loss": 4.4795708656311035,
+ "step": 3972
+ },
+ {
+ "epoch": 55.18165938864629,
+ "grad_norm": 0.01650134287774563,
+ "learning_rate": 0.0006,
+ "loss": 4.57435941696167,
+ "step": 3973
+ },
+ {
+ "epoch": 55.19563318777293,
+ "grad_norm": 0.01734331250190735,
+ "learning_rate": 0.0006,
+ "loss": 4.661007881164551,
+ "step": 3974
+ },
+ {
+ "epoch": 55.209606986899566,
+ "grad_norm": 0.01651303470134735,
+ "learning_rate": 0.0006,
+ "loss": 4.511148452758789,
+ "step": 3975
+ },
+ {
+ "epoch": 55.223580786026204,
+ "grad_norm": 0.016189761459827423,
+ "learning_rate": 0.0006,
+ "loss": 4.592350006103516,
+ "step": 3976
+ },
+ {
+ "epoch": 55.237554585152836,
+ "grad_norm": 0.01833222061395645,
+ "learning_rate": 0.0006,
+ "loss": 4.506649971008301,
+ "step": 3977
+ },
+ {
+ "epoch": 55.251528384279474,
+ "grad_norm": 0.02433410845696926,
+ "learning_rate": 0.0006,
+ "loss": 4.49500036239624,
+ "step": 3978
+ },
+ {
+ "epoch": 55.26550218340611,
+ "grad_norm": 0.0314289927482605,
+ "learning_rate": 0.0006,
+ "loss": 4.507624626159668,
+ "step": 3979
+ },
+ {
+ "epoch": 55.27947598253275,
+ "grad_norm": 0.03013679012656212,
+ "learning_rate": 0.0006,
+ "loss": 4.533641815185547,
+ "step": 3980
+ },
+ {
+ "epoch": 55.29344978165939,
+ "grad_norm": 0.024524061009287834,
+ "learning_rate": 0.0006,
+ "loss": 4.5721116065979,
+ "step": 3981
+ },
+ {
+ "epoch": 55.30742358078603,
+ "grad_norm": 0.02640336938202381,
+ "learning_rate": 0.0006,
+ "loss": 4.652521133422852,
+ "step": 3982
+ },
+ {
+ "epoch": 55.32139737991266,
+ "grad_norm": 0.021309101954102516,
+ "learning_rate": 0.0006,
+ "loss": 4.600375175476074,
+ "step": 3983
+ },
+ {
+ "epoch": 55.3353711790393,
+ "grad_norm": 0.023008549585938454,
+ "learning_rate": 0.0006,
+ "loss": 4.53914737701416,
+ "step": 3984
+ },
+ {
+ "epoch": 55.34934497816594,
+ "grad_norm": 0.022491468116641045,
+ "learning_rate": 0.0006,
+ "loss": 4.661163330078125,
+ "step": 3985
+ },
+ {
+ "epoch": 55.36331877729258,
+ "grad_norm": 0.01904081553220749,
+ "learning_rate": 0.0006,
+ "loss": 4.5426177978515625,
+ "step": 3986
+ },
+ {
+ "epoch": 55.377292576419215,
+ "grad_norm": 0.019229518249630928,
+ "learning_rate": 0.0006,
+ "loss": 4.606066703796387,
+ "step": 3987
+ },
+ {
+ "epoch": 55.391266375545854,
+ "grad_norm": 0.01781383901834488,
+ "learning_rate": 0.0006,
+ "loss": 4.54642391204834,
+ "step": 3988
+ },
+ {
+ "epoch": 55.40524017467249,
+ "grad_norm": 0.0174600500613451,
+ "learning_rate": 0.0006,
+ "loss": 4.711891174316406,
+ "step": 3989
+ },
+ {
+ "epoch": 55.419213973799124,
+ "grad_norm": 0.018154339864850044,
+ "learning_rate": 0.0006,
+ "loss": 4.527122974395752,
+ "step": 3990
+ },
+ {
+ "epoch": 55.43318777292576,
+ "grad_norm": 0.01914648339152336,
+ "learning_rate": 0.0006,
+ "loss": 4.641975402832031,
+ "step": 3991
+ },
+ {
+ "epoch": 55.4471615720524,
+ "grad_norm": 0.017118625342845917,
+ "learning_rate": 0.0006,
+ "loss": 4.511425971984863,
+ "step": 3992
+ },
+ {
+ "epoch": 55.46113537117904,
+ "grad_norm": 0.01551489345729351,
+ "learning_rate": 0.0006,
+ "loss": 4.551141738891602,
+ "step": 3993
+ },
+ {
+ "epoch": 55.47510917030568,
+ "grad_norm": 0.015072065405547619,
+ "learning_rate": 0.0006,
+ "loss": 4.651115894317627,
+ "step": 3994
+ },
+ {
+ "epoch": 55.48908296943232,
+ "grad_norm": 0.017010986804962158,
+ "learning_rate": 0.0006,
+ "loss": 4.583282470703125,
+ "step": 3995
+ },
+ {
+ "epoch": 55.50305676855895,
+ "grad_norm": 0.01764831133186817,
+ "learning_rate": 0.0006,
+ "loss": 4.593364238739014,
+ "step": 3996
+ },
+ {
+ "epoch": 55.51703056768559,
+ "grad_norm": 0.01762884296476841,
+ "learning_rate": 0.0006,
+ "loss": 4.45671272277832,
+ "step": 3997
+ },
+ {
+ "epoch": 55.531004366812226,
+ "grad_norm": 0.01600457727909088,
+ "learning_rate": 0.0006,
+ "loss": 4.511260032653809,
+ "step": 3998
+ },
+ {
+ "epoch": 55.544978165938865,
+ "grad_norm": 0.015706021338701248,
+ "learning_rate": 0.0006,
+ "loss": 4.47593355178833,
+ "step": 3999
+ },
+ {
+ "epoch": 55.5589519650655,
+ "grad_norm": 0.017718996852636337,
+ "learning_rate": 0.0006,
+ "loss": 4.610543251037598,
+ "step": 4000
+ },
+ {
+ "epoch": 55.57292576419214,
+ "grad_norm": 0.015073757618665695,
+ "learning_rate": 0.0006,
+ "loss": 4.466620445251465,
+ "step": 4001
+ },
+ {
+ "epoch": 55.58689956331878,
+ "grad_norm": 0.01789247989654541,
+ "learning_rate": 0.0006,
+ "loss": 4.503409385681152,
+ "step": 4002
+ },
+ {
+ "epoch": 55.60087336244541,
+ "grad_norm": 0.022329598665237427,
+ "learning_rate": 0.0006,
+ "loss": 4.62099552154541,
+ "step": 4003
+ },
+ {
+ "epoch": 55.61484716157205,
+ "grad_norm": 0.02343829534947872,
+ "learning_rate": 0.0006,
+ "loss": 4.620813369750977,
+ "step": 4004
+ },
+ {
+ "epoch": 55.62882096069869,
+ "grad_norm": 0.023279687389731407,
+ "learning_rate": 0.0006,
+ "loss": 4.545765399932861,
+ "step": 4005
+ },
+ {
+ "epoch": 55.64279475982533,
+ "grad_norm": 0.020817464217543602,
+ "learning_rate": 0.0006,
+ "loss": 4.561470031738281,
+ "step": 4006
+ },
+ {
+ "epoch": 55.65676855895197,
+ "grad_norm": 0.020270058885216713,
+ "learning_rate": 0.0006,
+ "loss": 4.5875020027160645,
+ "step": 4007
+ },
+ {
+ "epoch": 55.670742358078606,
+ "grad_norm": 0.021085843443870544,
+ "learning_rate": 0.0006,
+ "loss": 4.504560470581055,
+ "step": 4008
+ },
+ {
+ "epoch": 55.68471615720524,
+ "grad_norm": 0.018507730215787888,
+ "learning_rate": 0.0006,
+ "loss": 4.520644664764404,
+ "step": 4009
+ },
+ {
+ "epoch": 55.698689956331876,
+ "grad_norm": 0.020571516826748848,
+ "learning_rate": 0.0006,
+ "loss": 4.539643287658691,
+ "step": 4010
+ },
+ {
+ "epoch": 55.712663755458514,
+ "grad_norm": 0.021484510973095894,
+ "learning_rate": 0.0006,
+ "loss": 4.628833293914795,
+ "step": 4011
+ },
+ {
+ "epoch": 55.72663755458515,
+ "grad_norm": 0.017693819478154182,
+ "learning_rate": 0.0006,
+ "loss": 4.6023783683776855,
+ "step": 4012
+ },
+ {
+ "epoch": 55.74061135371179,
+ "grad_norm": 0.020186059176921844,
+ "learning_rate": 0.0006,
+ "loss": 4.634884834289551,
+ "step": 4013
+ },
+ {
+ "epoch": 55.75458515283843,
+ "grad_norm": 0.01835780404508114,
+ "learning_rate": 0.0006,
+ "loss": 4.577791213989258,
+ "step": 4014
+ },
+ {
+ "epoch": 55.76855895196506,
+ "grad_norm": 0.018162988126277924,
+ "learning_rate": 0.0006,
+ "loss": 4.641597270965576,
+ "step": 4015
+ },
+ {
+ "epoch": 55.7825327510917,
+ "grad_norm": 0.018909504637122154,
+ "learning_rate": 0.0006,
+ "loss": 4.487447738647461,
+ "step": 4016
+ },
+ {
+ "epoch": 55.79650655021834,
+ "grad_norm": 0.018972251564264297,
+ "learning_rate": 0.0006,
+ "loss": 4.590317726135254,
+ "step": 4017
+ },
+ {
+ "epoch": 55.81048034934498,
+ "grad_norm": 0.01793607883155346,
+ "learning_rate": 0.0006,
+ "loss": 4.4994611740112305,
+ "step": 4018
+ },
+ {
+ "epoch": 55.82445414847162,
+ "grad_norm": 0.02123514749109745,
+ "learning_rate": 0.0006,
+ "loss": 4.479006767272949,
+ "step": 4019
+ },
+ {
+ "epoch": 55.838427947598255,
+ "grad_norm": 0.021246599033474922,
+ "learning_rate": 0.0006,
+ "loss": 4.4738640785217285,
+ "step": 4020
+ },
+ {
+ "epoch": 55.852401746724894,
+ "grad_norm": 0.020131602883338928,
+ "learning_rate": 0.0006,
+ "loss": 4.600308418273926,
+ "step": 4021
+ },
+ {
+ "epoch": 55.866375545851525,
+ "grad_norm": 0.019159460440278053,
+ "learning_rate": 0.0006,
+ "loss": 4.540908336639404,
+ "step": 4022
+ },
+ {
+ "epoch": 55.880349344978164,
+ "grad_norm": 0.023114705458283424,
+ "learning_rate": 0.0006,
+ "loss": 4.641247749328613,
+ "step": 4023
+ },
+ {
+ "epoch": 55.8943231441048,
+ "grad_norm": 0.02296135015785694,
+ "learning_rate": 0.0006,
+ "loss": 4.444624423980713,
+ "step": 4024
+ },
+ {
+ "epoch": 55.90829694323144,
+ "grad_norm": 0.02204655110836029,
+ "learning_rate": 0.0006,
+ "loss": 4.506349563598633,
+ "step": 4025
+ },
+ {
+ "epoch": 55.92227074235808,
+ "grad_norm": 0.019338225945830345,
+ "learning_rate": 0.0006,
+ "loss": 4.616047382354736,
+ "step": 4026
+ },
+ {
+ "epoch": 55.93624454148472,
+ "grad_norm": 0.020357884466648102,
+ "learning_rate": 0.0006,
+ "loss": 4.48170280456543,
+ "step": 4027
+ },
+ {
+ "epoch": 55.95021834061135,
+ "grad_norm": 0.02426772564649582,
+ "learning_rate": 0.0006,
+ "loss": 4.576877117156982,
+ "step": 4028
+ },
+ {
+ "epoch": 55.96419213973799,
+ "grad_norm": 0.023351816460490227,
+ "learning_rate": 0.0006,
+ "loss": 4.458885192871094,
+ "step": 4029
+ },
+ {
+ "epoch": 55.97816593886463,
+ "grad_norm": 0.019892243668437004,
+ "learning_rate": 0.0006,
+ "loss": 4.520229816436768,
+ "step": 4030
+ },
+ {
+ "epoch": 55.992139737991266,
+ "grad_norm": 0.018331089988350868,
+ "learning_rate": 0.0006,
+ "loss": 4.432201862335205,
+ "step": 4031
+ },
+ {
+ "epoch": 56.0,
+ "grad_norm": 0.021300828084349632,
+ "learning_rate": 0.0006,
+ "loss": 4.617072582244873,
+ "step": 4032
+ },
+ {
+ "epoch": 56.0,
+ "eval_loss": 4.811229705810547,
+ "eval_runtime": 56.4846,
+ "eval_samples_per_second": 43.233,
+ "eval_steps_per_second": 1.363,
+ "step": 4032
+ },
+ {
+ "epoch": 56.01397379912664,
+ "grad_norm": 0.02199401892721653,
+ "learning_rate": 0.0006,
+ "loss": 4.513965606689453,
+ "step": 4033
+ },
+ {
+ "epoch": 56.02794759825328,
+ "grad_norm": 0.022109858691692352,
+ "learning_rate": 0.0006,
+ "loss": 4.557568550109863,
+ "step": 4034
+ },
+ {
+ "epoch": 56.041921397379916,
+ "grad_norm": 0.020103694871068,
+ "learning_rate": 0.0006,
+ "loss": 4.569512844085693,
+ "step": 4035
+ },
+ {
+ "epoch": 56.05589519650655,
+ "grad_norm": 0.01898941583931446,
+ "learning_rate": 0.0006,
+ "loss": 4.623138904571533,
+ "step": 4036
+ },
+ {
+ "epoch": 56.069868995633186,
+ "grad_norm": 0.02227010577917099,
+ "learning_rate": 0.0006,
+ "loss": 4.433316230773926,
+ "step": 4037
+ },
+ {
+ "epoch": 56.083842794759825,
+ "grad_norm": 0.024683045223355293,
+ "learning_rate": 0.0006,
+ "loss": 4.458296775817871,
+ "step": 4038
+ },
+ {
+ "epoch": 56.09781659388646,
+ "grad_norm": 0.02900184690952301,
+ "learning_rate": 0.0006,
+ "loss": 4.531060218811035,
+ "step": 4039
+ },
+ {
+ "epoch": 56.1117903930131,
+ "grad_norm": 0.025557566434144974,
+ "learning_rate": 0.0006,
+ "loss": 4.56685733795166,
+ "step": 4040
+ },
+ {
+ "epoch": 56.12576419213974,
+ "grad_norm": 0.021414656192064285,
+ "learning_rate": 0.0006,
+ "loss": 4.469037055969238,
+ "step": 4041
+ },
+ {
+ "epoch": 56.13973799126637,
+ "grad_norm": 0.023262616246938705,
+ "learning_rate": 0.0006,
+ "loss": 4.493876934051514,
+ "step": 4042
+ },
+ {
+ "epoch": 56.15371179039301,
+ "grad_norm": 0.023011326789855957,
+ "learning_rate": 0.0006,
+ "loss": 4.4572906494140625,
+ "step": 4043
+ },
+ {
+ "epoch": 56.16768558951965,
+ "grad_norm": 0.02177615463733673,
+ "learning_rate": 0.0006,
+ "loss": 4.615389823913574,
+ "step": 4044
+ },
+ {
+ "epoch": 56.18165938864629,
+ "grad_norm": 0.021527713164687157,
+ "learning_rate": 0.0006,
+ "loss": 4.613762855529785,
+ "step": 4045
+ },
+ {
+ "epoch": 56.19563318777293,
+ "grad_norm": 0.01982325315475464,
+ "learning_rate": 0.0006,
+ "loss": 4.53670597076416,
+ "step": 4046
+ },
+ {
+ "epoch": 56.209606986899566,
+ "grad_norm": 0.022484654560685158,
+ "learning_rate": 0.0006,
+ "loss": 4.605565071105957,
+ "step": 4047
+ },
+ {
+ "epoch": 56.223580786026204,
+ "grad_norm": 0.02346140146255493,
+ "learning_rate": 0.0006,
+ "loss": 4.6265950202941895,
+ "step": 4048
+ },
+ {
+ "epoch": 56.237554585152836,
+ "grad_norm": 0.01876714453101158,
+ "learning_rate": 0.0006,
+ "loss": 4.481525421142578,
+ "step": 4049
+ },
+ {
+ "epoch": 56.251528384279474,
+ "grad_norm": 0.01926851086318493,
+ "learning_rate": 0.0006,
+ "loss": 4.630258083343506,
+ "step": 4050
+ },
+ {
+ "epoch": 56.26550218340611,
+ "grad_norm": 0.020858589559793472,
+ "learning_rate": 0.0006,
+ "loss": 4.5265913009643555,
+ "step": 4051
+ },
+ {
+ "epoch": 56.27947598253275,
+ "grad_norm": 0.01889280416071415,
+ "learning_rate": 0.0006,
+ "loss": 4.581275463104248,
+ "step": 4052
+ },
+ {
+ "epoch": 56.29344978165939,
+ "grad_norm": 0.01801002025604248,
+ "learning_rate": 0.0006,
+ "loss": 4.498557090759277,
+ "step": 4053
+ },
+ {
+ "epoch": 56.30742358078603,
+ "grad_norm": 0.02252180315554142,
+ "learning_rate": 0.0006,
+ "loss": 4.565267086029053,
+ "step": 4054
+ },
+ {
+ "epoch": 56.32139737991266,
+ "grad_norm": 0.02306070365011692,
+ "learning_rate": 0.0006,
+ "loss": 4.3591389656066895,
+ "step": 4055
+ },
+ {
+ "epoch": 56.3353711790393,
+ "grad_norm": 0.020292513072490692,
+ "learning_rate": 0.0006,
+ "loss": 4.407517433166504,
+ "step": 4056
+ },
+ {
+ "epoch": 56.34934497816594,
+ "grad_norm": 0.019154123961925507,
+ "learning_rate": 0.0006,
+ "loss": 4.482261657714844,
+ "step": 4057
+ },
+ {
+ "epoch": 56.36331877729258,
+ "grad_norm": 0.017338937148451805,
+ "learning_rate": 0.0006,
+ "loss": 4.4463582038879395,
+ "step": 4058
+ },
+ {
+ "epoch": 56.377292576419215,
+ "grad_norm": 0.01730559952557087,
+ "learning_rate": 0.0006,
+ "loss": 4.542474746704102,
+ "step": 4059
+ },
+ {
+ "epoch": 56.391266375545854,
+ "grad_norm": 0.017029426991939545,
+ "learning_rate": 0.0006,
+ "loss": 4.552488327026367,
+ "step": 4060
+ },
+ {
+ "epoch": 56.40524017467249,
+ "grad_norm": 0.01745498552918434,
+ "learning_rate": 0.0006,
+ "loss": 4.52928352355957,
+ "step": 4061
+ },
+ {
+ "epoch": 56.419213973799124,
+ "grad_norm": 0.0168515145778656,
+ "learning_rate": 0.0006,
+ "loss": 4.4516801834106445,
+ "step": 4062
+ },
+ {
+ "epoch": 56.43318777292576,
+ "grad_norm": 0.018545417115092278,
+ "learning_rate": 0.0006,
+ "loss": 4.569811820983887,
+ "step": 4063
+ },
+ {
+ "epoch": 56.4471615720524,
+ "grad_norm": 0.02227838523685932,
+ "learning_rate": 0.0006,
+ "loss": 4.412224769592285,
+ "step": 4064
+ },
+ {
+ "epoch": 56.46113537117904,
+ "grad_norm": 0.023338504135608673,
+ "learning_rate": 0.0006,
+ "loss": 4.490184307098389,
+ "step": 4065
+ },
+ {
+ "epoch": 56.47510917030568,
+ "grad_norm": 0.02257535606622696,
+ "learning_rate": 0.0006,
+ "loss": 4.479152679443359,
+ "step": 4066
+ },
+ {
+ "epoch": 56.48908296943232,
+ "grad_norm": 0.022737598046660423,
+ "learning_rate": 0.0006,
+ "loss": 4.418235778808594,
+ "step": 4067
+ },
+ {
+ "epoch": 56.50305676855895,
+ "grad_norm": 0.024603573605418205,
+ "learning_rate": 0.0006,
+ "loss": 4.45781135559082,
+ "step": 4068
+ },
+ {
+ "epoch": 56.51703056768559,
+ "grad_norm": 0.024953778833150864,
+ "learning_rate": 0.0006,
+ "loss": 4.593907833099365,
+ "step": 4069
+ },
+ {
+ "epoch": 56.531004366812226,
+ "grad_norm": 0.02195645309984684,
+ "learning_rate": 0.0006,
+ "loss": 4.467643737792969,
+ "step": 4070
+ },
+ {
+ "epoch": 56.544978165938865,
+ "grad_norm": 0.017126578837633133,
+ "learning_rate": 0.0006,
+ "loss": 4.555484771728516,
+ "step": 4071
+ },
+ {
+ "epoch": 56.5589519650655,
+ "grad_norm": 0.019482605159282684,
+ "learning_rate": 0.0006,
+ "loss": 4.430915832519531,
+ "step": 4072
+ },
+ {
+ "epoch": 56.57292576419214,
+ "grad_norm": 0.0219392292201519,
+ "learning_rate": 0.0006,
+ "loss": 4.593365669250488,
+ "step": 4073
+ },
+ {
+ "epoch": 56.58689956331878,
+ "grad_norm": 0.02420145645737648,
+ "learning_rate": 0.0006,
+ "loss": 4.491467475891113,
+ "step": 4074
+ },
+ {
+ "epoch": 56.60087336244541,
+ "grad_norm": 0.021523723378777504,
+ "learning_rate": 0.0006,
+ "loss": 4.463634490966797,
+ "step": 4075
+ },
+ {
+ "epoch": 56.61484716157205,
+ "grad_norm": 0.020242048427462578,
+ "learning_rate": 0.0006,
+ "loss": 4.577968597412109,
+ "step": 4076
+ },
+ {
+ "epoch": 56.62882096069869,
+ "grad_norm": 0.017771173268556595,
+ "learning_rate": 0.0006,
+ "loss": 4.568269729614258,
+ "step": 4077
+ },
+ {
+ "epoch": 56.64279475982533,
+ "grad_norm": 0.01598544418811798,
+ "learning_rate": 0.0006,
+ "loss": 4.476841926574707,
+ "step": 4078
+ },
+ {
+ "epoch": 56.65676855895197,
+ "grad_norm": 0.0168446097522974,
+ "learning_rate": 0.0006,
+ "loss": 4.480965614318848,
+ "step": 4079
+ },
+ {
+ "epoch": 56.670742358078606,
+ "grad_norm": 0.014948660507798195,
+ "learning_rate": 0.0006,
+ "loss": 4.535400390625,
+ "step": 4080
+ },
+ {
+ "epoch": 56.68471615720524,
+ "grad_norm": 0.014899153262376785,
+ "learning_rate": 0.0006,
+ "loss": 4.476048946380615,
+ "step": 4081
+ },
+ {
+ "epoch": 56.698689956331876,
+ "grad_norm": 0.014051870442926884,
+ "learning_rate": 0.0006,
+ "loss": 4.533884048461914,
+ "step": 4082
+ },
+ {
+ "epoch": 56.712663755458514,
+ "grad_norm": 0.014708572067320347,
+ "learning_rate": 0.0006,
+ "loss": 4.586690902709961,
+ "step": 4083
+ },
+ {
+ "epoch": 56.72663755458515,
+ "grad_norm": 0.014797299169003963,
+ "learning_rate": 0.0006,
+ "loss": 4.446291446685791,
+ "step": 4084
+ },
+ {
+ "epoch": 56.74061135371179,
+ "grad_norm": 0.015256541781127453,
+ "learning_rate": 0.0006,
+ "loss": 4.553572654724121,
+ "step": 4085
+ },
+ {
+ "epoch": 56.75458515283843,
+ "grad_norm": 0.016042491421103477,
+ "learning_rate": 0.0006,
+ "loss": 4.507801055908203,
+ "step": 4086
+ },
+ {
+ "epoch": 56.76855895196506,
+ "grad_norm": 0.018835173919796944,
+ "learning_rate": 0.0006,
+ "loss": 4.393911838531494,
+ "step": 4087
+ },
+ {
+ "epoch": 56.7825327510917,
+ "grad_norm": 0.021994033828377724,
+ "learning_rate": 0.0006,
+ "loss": 4.543368339538574,
+ "step": 4088
+ },
+ {
+ "epoch": 56.79650655021834,
+ "grad_norm": 0.0230990182608366,
+ "learning_rate": 0.0006,
+ "loss": 4.507420539855957,
+ "step": 4089
+ },
+ {
+ "epoch": 56.81048034934498,
+ "grad_norm": 0.018818842247128487,
+ "learning_rate": 0.0006,
+ "loss": 4.499941349029541,
+ "step": 4090
+ },
+ {
+ "epoch": 56.82445414847162,
+ "grad_norm": 0.018257278949022293,
+ "learning_rate": 0.0006,
+ "loss": 4.380057334899902,
+ "step": 4091
+ },
+ {
+ "epoch": 56.838427947598255,
+ "grad_norm": 0.020073018968105316,
+ "learning_rate": 0.0006,
+ "loss": 4.572811603546143,
+ "step": 4092
+ },
+ {
+ "epoch": 56.852401746724894,
+ "grad_norm": 0.02078658528625965,
+ "learning_rate": 0.0006,
+ "loss": 4.446835517883301,
+ "step": 4093
+ },
+ {
+ "epoch": 56.866375545851525,
+ "grad_norm": 0.01793118566274643,
+ "learning_rate": 0.0006,
+ "loss": 4.517854690551758,
+ "step": 4094
+ },
+ {
+ "epoch": 56.880349344978164,
+ "grad_norm": 0.01608281210064888,
+ "learning_rate": 0.0006,
+ "loss": 4.506482124328613,
+ "step": 4095
+ },
+ {
+ "epoch": 56.8943231441048,
+ "grad_norm": 0.017405368387699127,
+ "learning_rate": 0.0006,
+ "loss": 4.439718246459961,
+ "step": 4096
+ },
+ {
+ "epoch": 56.90829694323144,
+ "grad_norm": 0.017930863425135612,
+ "learning_rate": 0.0006,
+ "loss": 4.409365653991699,
+ "step": 4097
+ },
+ {
+ "epoch": 56.92227074235808,
+ "grad_norm": 0.015757670626044273,
+ "learning_rate": 0.0006,
+ "loss": 4.505547523498535,
+ "step": 4098
+ },
+ {
+ "epoch": 56.93624454148472,
+ "grad_norm": 0.01580345816910267,
+ "learning_rate": 0.0006,
+ "loss": 4.462774753570557,
+ "step": 4099
+ },
+ {
+ "epoch": 56.95021834061135,
+ "grad_norm": 0.01659180410206318,
+ "learning_rate": 0.0006,
+ "loss": 4.49345588684082,
+ "step": 4100
+ },
+ {
+ "epoch": 56.96419213973799,
+ "grad_norm": 0.01877259835600853,
+ "learning_rate": 0.0006,
+ "loss": 4.574443340301514,
+ "step": 4101
+ },
+ {
+ "epoch": 56.97816593886463,
+ "grad_norm": 0.01977839693427086,
+ "learning_rate": 0.0006,
+ "loss": 4.574034690856934,
+ "step": 4102
+ },
+ {
+ "epoch": 56.992139737991266,
+ "grad_norm": 0.020627597346901894,
+ "learning_rate": 0.0006,
+ "loss": 4.50969123840332,
+ "step": 4103
+ },
+ {
+ "epoch": 57.0,
+ "grad_norm": 0.025128807872533798,
+ "learning_rate": 0.0006,
+ "loss": 4.401662826538086,
+ "step": 4104
+ },
+ {
+ "epoch": 57.0,
+ "eval_loss": 4.779252052307129,
+ "eval_runtime": 56.8104,
+ "eval_samples_per_second": 42.985,
+ "eval_steps_per_second": 1.355,
+ "step": 4104
+ },
+ {
+ "epoch": 57.01397379912664,
+ "grad_norm": 0.027712570503354073,
+ "learning_rate": 0.0006,
+ "loss": 4.472955703735352,
+ "step": 4105
+ },
+ {
+ "epoch": 57.02794759825328,
+ "grad_norm": 0.02164197526872158,
+ "learning_rate": 0.0006,
+ "loss": 4.387596130371094,
+ "step": 4106
+ },
+ {
+ "epoch": 57.041921397379916,
+ "grad_norm": 0.017972752451896667,
+ "learning_rate": 0.0006,
+ "loss": 4.4326276779174805,
+ "step": 4107
+ },
+ {
+ "epoch": 57.05589519650655,
+ "grad_norm": 0.020811019465327263,
+ "learning_rate": 0.0006,
+ "loss": 4.452951431274414,
+ "step": 4108
+ },
+ {
+ "epoch": 57.069868995633186,
+ "grad_norm": 0.019913259893655777,
+ "learning_rate": 0.0006,
+ "loss": 4.448301315307617,
+ "step": 4109
+ },
+ {
+ "epoch": 57.083842794759825,
+ "grad_norm": 0.017409102991223335,
+ "learning_rate": 0.0006,
+ "loss": 4.4108357429504395,
+ "step": 4110
+ },
+ {
+ "epoch": 57.09781659388646,
+ "grad_norm": 0.019374269992113113,
+ "learning_rate": 0.0006,
+ "loss": 4.398412227630615,
+ "step": 4111
+ },
+ {
+ "epoch": 57.1117903930131,
+ "grad_norm": 0.015521776862442493,
+ "learning_rate": 0.0006,
+ "loss": 4.508594989776611,
+ "step": 4112
+ },
+ {
+ "epoch": 57.12576419213974,
+ "grad_norm": 0.017201920971274376,
+ "learning_rate": 0.0006,
+ "loss": 4.449558734893799,
+ "step": 4113
+ },
+ {
+ "epoch": 57.13973799126637,
+ "grad_norm": 0.018062541261315346,
+ "learning_rate": 0.0006,
+ "loss": 4.517874240875244,
+ "step": 4114
+ },
+ {
+ "epoch": 57.15371179039301,
+ "grad_norm": 0.015846073627471924,
+ "learning_rate": 0.0006,
+ "loss": 4.427501678466797,
+ "step": 4115
+ },
+ {
+ "epoch": 57.16768558951965,
+ "grad_norm": 0.01849362626671791,
+ "learning_rate": 0.0006,
+ "loss": 4.433181285858154,
+ "step": 4116
+ },
+ {
+ "epoch": 57.18165938864629,
+ "grad_norm": 0.01948804222047329,
+ "learning_rate": 0.0006,
+ "loss": 4.446122646331787,
+ "step": 4117
+ },
+ {
+ "epoch": 57.19563318777293,
+ "grad_norm": 0.017354389652609825,
+ "learning_rate": 0.0006,
+ "loss": 4.431153774261475,
+ "step": 4118
+ },
+ {
+ "epoch": 57.209606986899566,
+ "grad_norm": 0.015817373991012573,
+ "learning_rate": 0.0006,
+ "loss": 4.448566913604736,
+ "step": 4119
+ },
+ {
+ "epoch": 57.223580786026204,
+ "grad_norm": 0.01635785959661007,
+ "learning_rate": 0.0006,
+ "loss": 4.408031463623047,
+ "step": 4120
+ },
+ {
+ "epoch": 57.237554585152836,
+ "grad_norm": 0.014528338797390461,
+ "learning_rate": 0.0006,
+ "loss": 4.541668891906738,
+ "step": 4121
+ },
+ {
+ "epoch": 57.251528384279474,
+ "grad_norm": 0.01589174196124077,
+ "learning_rate": 0.0006,
+ "loss": 4.513199329376221,
+ "step": 4122
+ },
+ {
+ "epoch": 57.26550218340611,
+ "grad_norm": 0.016093695536255836,
+ "learning_rate": 0.0006,
+ "loss": 4.516178131103516,
+ "step": 4123
+ },
+ {
+ "epoch": 57.27947598253275,
+ "grad_norm": 0.01788361556828022,
+ "learning_rate": 0.0006,
+ "loss": 4.524880409240723,
+ "step": 4124
+ },
+ {
+ "epoch": 57.29344978165939,
+ "grad_norm": 0.01591576263308525,
+ "learning_rate": 0.0006,
+ "loss": 4.44764518737793,
+ "step": 4125
+ },
+ {
+ "epoch": 57.30742358078603,
+ "grad_norm": 0.015452570281922817,
+ "learning_rate": 0.0006,
+ "loss": 4.46844482421875,
+ "step": 4126
+ },
+ {
+ "epoch": 57.32139737991266,
+ "grad_norm": 0.016404492780566216,
+ "learning_rate": 0.0006,
+ "loss": 4.443678855895996,
+ "step": 4127
+ },
+ {
+ "epoch": 57.3353711790393,
+ "grad_norm": 0.019967148080468178,
+ "learning_rate": 0.0006,
+ "loss": 4.430230140686035,
+ "step": 4128
+ },
+ {
+ "epoch": 57.34934497816594,
+ "grad_norm": 0.02594015561044216,
+ "learning_rate": 0.0006,
+ "loss": 4.47061824798584,
+ "step": 4129
+ },
+ {
+ "epoch": 57.36331877729258,
+ "grad_norm": 0.031096890568733215,
+ "learning_rate": 0.0006,
+ "loss": 4.462275981903076,
+ "step": 4130
+ },
+ {
+ "epoch": 57.377292576419215,
+ "grad_norm": 0.02834898792207241,
+ "learning_rate": 0.0006,
+ "loss": 4.41143798828125,
+ "step": 4131
+ },
+ {
+ "epoch": 57.391266375545854,
+ "grad_norm": 0.023621153086423874,
+ "learning_rate": 0.0006,
+ "loss": 4.562548637390137,
+ "step": 4132
+ },
+ {
+ "epoch": 57.40524017467249,
+ "grad_norm": 0.02277296595275402,
+ "learning_rate": 0.0006,
+ "loss": 4.456247329711914,
+ "step": 4133
+ },
+ {
+ "epoch": 57.419213973799124,
+ "grad_norm": 0.019868768751621246,
+ "learning_rate": 0.0006,
+ "loss": 4.524300575256348,
+ "step": 4134
+ },
+ {
+ "epoch": 57.43318777292576,
+ "grad_norm": 0.01887078955769539,
+ "learning_rate": 0.0006,
+ "loss": 4.540097236633301,
+ "step": 4135
+ },
+ {
+ "epoch": 57.4471615720524,
+ "grad_norm": 0.024268802255392075,
+ "learning_rate": 0.0006,
+ "loss": 4.397233963012695,
+ "step": 4136
+ },
+ {
+ "epoch": 57.46113537117904,
+ "grad_norm": 0.025739721953868866,
+ "learning_rate": 0.0006,
+ "loss": 4.354194641113281,
+ "step": 4137
+ },
+ {
+ "epoch": 57.47510917030568,
+ "grad_norm": 0.02442290261387825,
+ "learning_rate": 0.0006,
+ "loss": 4.303822994232178,
+ "step": 4138
+ },
+ {
+ "epoch": 57.48908296943232,
+ "grad_norm": 0.022215209901332855,
+ "learning_rate": 0.0006,
+ "loss": 4.4870452880859375,
+ "step": 4139
+ },
+ {
+ "epoch": 57.50305676855895,
+ "grad_norm": 0.01789158768951893,
+ "learning_rate": 0.0006,
+ "loss": 4.343809127807617,
+ "step": 4140
+ },
+ {
+ "epoch": 57.51703056768559,
+ "grad_norm": 0.020381739363074303,
+ "learning_rate": 0.0006,
+ "loss": 4.53312873840332,
+ "step": 4141
+ },
+ {
+ "epoch": 57.531004366812226,
+ "grad_norm": 0.019585080444812775,
+ "learning_rate": 0.0006,
+ "loss": 4.510162353515625,
+ "step": 4142
+ },
+ {
+ "epoch": 57.544978165938865,
+ "grad_norm": 0.01985127478837967,
+ "learning_rate": 0.0006,
+ "loss": 4.3654022216796875,
+ "step": 4143
+ },
+ {
+ "epoch": 57.5589519650655,
+ "grad_norm": 0.01962905190885067,
+ "learning_rate": 0.0006,
+ "loss": 4.560830116271973,
+ "step": 4144
+ },
+ {
+ "epoch": 57.57292576419214,
+ "grad_norm": 0.02679629437625408,
+ "learning_rate": 0.0006,
+ "loss": 4.300804138183594,
+ "step": 4145
+ },
+ {
+ "epoch": 57.58689956331878,
+ "grad_norm": 0.039830856025218964,
+ "learning_rate": 0.0006,
+ "loss": 4.54150915145874,
+ "step": 4146
+ },
+ {
+ "epoch": 57.60087336244541,
+ "grad_norm": 0.049527477473020554,
+ "learning_rate": 0.0006,
+ "loss": 4.405792236328125,
+ "step": 4147
+ },
+ {
+ "epoch": 57.61484716157205,
+ "grad_norm": 0.042966946959495544,
+ "learning_rate": 0.0006,
+ "loss": 4.500643730163574,
+ "step": 4148
+ },
+ {
+ "epoch": 57.62882096069869,
+ "grad_norm": 0.08999177813529968,
+ "learning_rate": 0.0006,
+ "loss": 4.504790782928467,
+ "step": 4149
+ },
+ {
+ "epoch": 57.64279475982533,
+ "grad_norm": 0.44693124294281006,
+ "learning_rate": 0.0006,
+ "loss": 4.845344066619873,
+ "step": 4150
+ },
+ {
+ "epoch": 57.65676855895197,
+ "grad_norm": 0.7143144011497498,
+ "learning_rate": 0.0006,
+ "loss": 6.907845497131348,
+ "step": 4151
+ },
+ {
+ "epoch": 57.670742358078606,
+ "grad_norm": 0.25239595770835876,
+ "learning_rate": 0.0006,
+ "loss": 7.082579612731934,
+ "step": 4152
+ },
+ {
+ "epoch": 57.68471615720524,
+ "grad_norm": 0.26550790667533875,
+ "learning_rate": 0.0006,
+ "loss": 7.50180196762085,
+ "step": 4153
+ },
+ {
+ "epoch": 57.698689956331876,
+ "grad_norm": 0.14735378324985504,
+ "learning_rate": 0.0006,
+ "loss": 7.214590072631836,
+ "step": 4154
+ },
+ {
+ "epoch": 57.712663755458514,
+ "grad_norm": 0.17685334384441376,
+ "learning_rate": 0.0006,
+ "loss": 6.861320495605469,
+ "step": 4155
+ },
+ {
+ "epoch": 57.72663755458515,
+ "grad_norm": 0.18231339752674103,
+ "learning_rate": 0.0006,
+ "loss": 6.791839599609375,
+ "step": 4156
+ },
+ {
+ "epoch": 57.74061135371179,
+ "grad_norm": 0.09230359643697739,
+ "learning_rate": 0.0006,
+ "loss": 6.760223388671875,
+ "step": 4157
+ },
+ {
+ "epoch": 57.75458515283843,
+ "grad_norm": 0.06918249279260635,
+ "learning_rate": 0.0006,
+ "loss": 6.532999515533447,
+ "step": 4158
+ },
+ {
+ "epoch": 57.76855895196506,
+ "grad_norm": 0.0901007279753685,
+ "learning_rate": 0.0006,
+ "loss": 6.52907133102417,
+ "step": 4159
+ },
+ {
+ "epoch": 57.7825327510917,
+ "grad_norm": 0.05678229779005051,
+ "learning_rate": 0.0006,
+ "loss": 6.295290470123291,
+ "step": 4160
+ },
+ {
+ "epoch": 57.79650655021834,
+ "grad_norm": 0.04450210556387901,
+ "learning_rate": 0.0006,
+ "loss": 6.14900541305542,
+ "step": 4161
+ },
+ {
+ "epoch": 57.81048034934498,
+ "grad_norm": 0.049191512167453766,
+ "learning_rate": 0.0006,
+ "loss": 6.19650411605835,
+ "step": 4162
+ },
+ {
+ "epoch": 57.82445414847162,
+ "grad_norm": 0.03908967226743698,
+ "learning_rate": 0.0006,
+ "loss": 6.054948806762695,
+ "step": 4163
+ },
+ {
+ "epoch": 57.838427947598255,
+ "grad_norm": 0.04023387283086777,
+ "learning_rate": 0.0006,
+ "loss": 6.034873962402344,
+ "step": 4164
+ },
+ {
+ "epoch": 57.852401746724894,
+ "grad_norm": 0.03304268419742584,
+ "learning_rate": 0.0006,
+ "loss": 5.769852161407471,
+ "step": 4165
+ },
+ {
+ "epoch": 57.866375545851525,
+ "grad_norm": 0.03432450816035271,
+ "learning_rate": 0.0006,
+ "loss": 5.788785457611084,
+ "step": 4166
+ },
+ {
+ "epoch": 57.880349344978164,
+ "grad_norm": 0.029965840280056,
+ "learning_rate": 0.0006,
+ "loss": 5.815243721008301,
+ "step": 4167
+ },
+ {
+ "epoch": 57.8943231441048,
+ "grad_norm": 0.03637698292732239,
+ "learning_rate": 0.0006,
+ "loss": 5.7204060554504395,
+ "step": 4168
+ },
+ {
+ "epoch": 57.90829694323144,
+ "grad_norm": 0.03477517142891884,
+ "learning_rate": 0.0006,
+ "loss": 5.605930328369141,
+ "step": 4169
+ },
+ {
+ "epoch": 57.92227074235808,
+ "grad_norm": 0.05061859264969826,
+ "learning_rate": 0.0006,
+ "loss": 5.547054767608643,
+ "step": 4170
+ },
+ {
+ "epoch": 57.93624454148472,
+ "grad_norm": 0.06299655884504318,
+ "learning_rate": 0.0006,
+ "loss": 5.552166938781738,
+ "step": 4171
+ },
+ {
+ "epoch": 57.95021834061135,
+ "grad_norm": 0.04215948283672333,
+ "learning_rate": 0.0006,
+ "loss": 5.375457763671875,
+ "step": 4172
+ },
+ {
+ "epoch": 57.96419213973799,
+ "grad_norm": 0.0348566472530365,
+ "learning_rate": 0.0006,
+ "loss": 5.418606281280518,
+ "step": 4173
+ },
+ {
+ "epoch": 57.97816593886463,
+ "grad_norm": 0.03807530924677849,
+ "learning_rate": 0.0006,
+ "loss": 5.403156280517578,
+ "step": 4174
+ },
+ {
+ "epoch": 57.992139737991266,
+ "grad_norm": 0.02989993803203106,
+ "learning_rate": 0.0006,
+ "loss": 5.355951309204102,
+ "step": 4175
+ },
+ {
+ "epoch": 58.0,
+ "grad_norm": 0.034300558269023895,
+ "learning_rate": 0.0006,
+ "loss": 5.44920539855957,
+ "step": 4176
+ },
+ {
+ "epoch": 58.0,
+ "eval_loss": 5.416731834411621,
+ "eval_runtime": 56.7934,
+ "eval_samples_per_second": 42.998,
+ "eval_steps_per_second": 1.356,
+ "step": 4176
+ },
+ {
+ "epoch": 58.01397379912664,
+ "grad_norm": 0.033961448818445206,
+ "learning_rate": 0.0006,
+ "loss": 5.250707626342773,
+ "step": 4177
+ },
+ {
+ "epoch": 58.02794759825328,
+ "grad_norm": 0.027343502268195152,
+ "learning_rate": 0.0006,
+ "loss": 5.149238109588623,
+ "step": 4178
+ },
+ {
+ "epoch": 58.041921397379916,
+ "grad_norm": 0.03033650480210781,
+ "learning_rate": 0.0006,
+ "loss": 5.072463512420654,
+ "step": 4179
+ },
+ {
+ "epoch": 58.05589519650655,
+ "grad_norm": 0.028262868523597717,
+ "learning_rate": 0.0006,
+ "loss": 5.109417915344238,
+ "step": 4180
+ },
+ {
+ "epoch": 58.069868995633186,
+ "grad_norm": 0.02373034693300724,
+ "learning_rate": 0.0006,
+ "loss": 5.082464218139648,
+ "step": 4181
+ },
+ {
+ "epoch": 58.083842794759825,
+ "grad_norm": 0.025879688560962677,
+ "learning_rate": 0.0006,
+ "loss": 5.056792259216309,
+ "step": 4182
+ },
+ {
+ "epoch": 58.09781659388646,
+ "grad_norm": 0.026252275332808495,
+ "learning_rate": 0.0006,
+ "loss": 5.064239978790283,
+ "step": 4183
+ },
+ {
+ "epoch": 58.1117903930131,
+ "grad_norm": 0.028177792206406593,
+ "learning_rate": 0.0006,
+ "loss": 4.922310829162598,
+ "step": 4184
+ },
+ {
+ "epoch": 58.12576419213974,
+ "grad_norm": 0.02630820870399475,
+ "learning_rate": 0.0006,
+ "loss": 4.992884635925293,
+ "step": 4185
+ },
+ {
+ "epoch": 58.13973799126637,
+ "grad_norm": 0.024890903383493423,
+ "learning_rate": 0.0006,
+ "loss": 4.910036563873291,
+ "step": 4186
+ },
+ {
+ "epoch": 58.15371179039301,
+ "grad_norm": 0.02822992391884327,
+ "learning_rate": 0.0006,
+ "loss": 4.866640090942383,
+ "step": 4187
+ },
+ {
+ "epoch": 58.16768558951965,
+ "grad_norm": 0.03253607079386711,
+ "learning_rate": 0.0006,
+ "loss": 4.9112958908081055,
+ "step": 4188
+ },
+ {
+ "epoch": 58.18165938864629,
+ "grad_norm": 0.03765944764018059,
+ "learning_rate": 0.0006,
+ "loss": 4.774468421936035,
+ "step": 4189
+ },
+ {
+ "epoch": 58.19563318777293,
+ "grad_norm": 0.039874881505966187,
+ "learning_rate": 0.0006,
+ "loss": 4.831678867340088,
+ "step": 4190
+ },
+ {
+ "epoch": 58.209606986899566,
+ "grad_norm": 0.034595925360918045,
+ "learning_rate": 0.0006,
+ "loss": 4.8243794441223145,
+ "step": 4191
+ },
+ {
+ "epoch": 58.223580786026204,
+ "grad_norm": 0.04840189591050148,
+ "learning_rate": 0.0006,
+ "loss": 4.854156494140625,
+ "step": 4192
+ },
+ {
+ "epoch": 58.237554585152836,
+ "grad_norm": 0.06358753889799118,
+ "learning_rate": 0.0006,
+ "loss": 4.836430549621582,
+ "step": 4193
+ },
+ {
+ "epoch": 58.251528384279474,
+ "grad_norm": 0.0507555715739727,
+ "learning_rate": 0.0006,
+ "loss": 4.88357400894165,
+ "step": 4194
+ },
+ {
+ "epoch": 58.26550218340611,
+ "grad_norm": 0.04384690150618553,
+ "learning_rate": 0.0006,
+ "loss": 4.774049282073975,
+ "step": 4195
+ },
+ {
+ "epoch": 58.27947598253275,
+ "grad_norm": 0.03246928006410599,
+ "learning_rate": 0.0006,
+ "loss": 4.686605930328369,
+ "step": 4196
+ },
+ {
+ "epoch": 58.29344978165939,
+ "grad_norm": 0.028878789395093918,
+ "learning_rate": 0.0006,
+ "loss": 4.796070575714111,
+ "step": 4197
+ },
+ {
+ "epoch": 58.30742358078603,
+ "grad_norm": 0.028175361454486847,
+ "learning_rate": 0.0006,
+ "loss": 4.728490829467773,
+ "step": 4198
+ },
+ {
+ "epoch": 58.32139737991266,
+ "grad_norm": 0.029428014531731606,
+ "learning_rate": 0.0006,
+ "loss": 4.7159624099731445,
+ "step": 4199
+ },
+ {
+ "epoch": 58.3353711790393,
+ "grad_norm": 0.031227827072143555,
+ "learning_rate": 0.0006,
+ "loss": 4.702625274658203,
+ "step": 4200
+ },
+ {
+ "epoch": 58.34934497816594,
+ "grad_norm": 0.028843428939580917,
+ "learning_rate": 0.0006,
+ "loss": 4.811799049377441,
+ "step": 4201
+ },
+ {
+ "epoch": 58.36331877729258,
+ "grad_norm": 0.026780391111969948,
+ "learning_rate": 0.0006,
+ "loss": 4.81662654876709,
+ "step": 4202
+ },
+ {
+ "epoch": 58.377292576419215,
+ "grad_norm": 0.019699757918715477,
+ "learning_rate": 0.0006,
+ "loss": 4.713685989379883,
+ "step": 4203
+ },
+ {
+ "epoch": 58.391266375545854,
+ "grad_norm": 0.024389250203967094,
+ "learning_rate": 0.0006,
+ "loss": 4.72354793548584,
+ "step": 4204
+ },
+ {
+ "epoch": 58.40524017467249,
+ "grad_norm": 0.02218274027109146,
+ "learning_rate": 0.0006,
+ "loss": 4.674125671386719,
+ "step": 4205
+ },
+ {
+ "epoch": 58.419213973799124,
+ "grad_norm": 0.02041488140821457,
+ "learning_rate": 0.0006,
+ "loss": 4.7909698486328125,
+ "step": 4206
+ },
+ {
+ "epoch": 58.43318777292576,
+ "grad_norm": 0.018561935052275658,
+ "learning_rate": 0.0006,
+ "loss": 4.808770179748535,
+ "step": 4207
+ },
+ {
+ "epoch": 58.4471615720524,
+ "grad_norm": 0.01708846725523472,
+ "learning_rate": 0.0006,
+ "loss": 4.701626777648926,
+ "step": 4208
+ },
+ {
+ "epoch": 58.46113537117904,
+ "grad_norm": 0.0222734697163105,
+ "learning_rate": 0.0006,
+ "loss": 4.505410194396973,
+ "step": 4209
+ },
+ {
+ "epoch": 58.47510917030568,
+ "grad_norm": 0.03197057917714119,
+ "learning_rate": 0.0006,
+ "loss": 4.610360145568848,
+ "step": 4210
+ },
+ {
+ "epoch": 58.48908296943232,
+ "grad_norm": 0.06854520738124847,
+ "learning_rate": 0.0006,
+ "loss": 4.625649452209473,
+ "step": 4211
+ },
+ {
+ "epoch": 58.50305676855895,
+ "grad_norm": 0.09981521219015121,
+ "learning_rate": 0.0006,
+ "loss": 4.832150459289551,
+ "step": 4212
+ },
+ {
+ "epoch": 58.51703056768559,
+ "grad_norm": 0.046356040984392166,
+ "learning_rate": 0.0006,
+ "loss": 4.5641632080078125,
+ "step": 4213
+ },
+ {
+ "epoch": 58.531004366812226,
+ "grad_norm": 0.03882667422294617,
+ "learning_rate": 0.0006,
+ "loss": 4.666974067687988,
+ "step": 4214
+ },
+ {
+ "epoch": 58.544978165938865,
+ "grad_norm": 0.033203233033418655,
+ "learning_rate": 0.0006,
+ "loss": 4.629413604736328,
+ "step": 4215
+ },
+ {
+ "epoch": 58.5589519650655,
+ "grad_norm": 0.02806735597550869,
+ "learning_rate": 0.0006,
+ "loss": 4.763245582580566,
+ "step": 4216
+ },
+ {
+ "epoch": 58.57292576419214,
+ "grad_norm": 0.026728278025984764,
+ "learning_rate": 0.0006,
+ "loss": 4.629597187042236,
+ "step": 4217
+ },
+ {
+ "epoch": 58.58689956331878,
+ "grad_norm": 0.024380959570407867,
+ "learning_rate": 0.0006,
+ "loss": 4.633614540100098,
+ "step": 4218
+ },
+ {
+ "epoch": 58.60087336244541,
+ "grad_norm": 0.022312544286251068,
+ "learning_rate": 0.0006,
+ "loss": 4.671379566192627,
+ "step": 4219
+ },
+ {
+ "epoch": 58.61484716157205,
+ "grad_norm": 0.022049568593502045,
+ "learning_rate": 0.0006,
+ "loss": 4.661557197570801,
+ "step": 4220
+ },
+ {
+ "epoch": 58.62882096069869,
+ "grad_norm": 0.02119818702340126,
+ "learning_rate": 0.0006,
+ "loss": 4.688562393188477,
+ "step": 4221
+ },
+ {
+ "epoch": 58.64279475982533,
+ "grad_norm": 0.019769301638007164,
+ "learning_rate": 0.0006,
+ "loss": 4.607748031616211,
+ "step": 4222
+ },
+ {
+ "epoch": 58.65676855895197,
+ "grad_norm": 0.02124079130589962,
+ "learning_rate": 0.0006,
+ "loss": 4.570640563964844,
+ "step": 4223
+ },
+ {
+ "epoch": 58.670742358078606,
+ "grad_norm": 0.016344040632247925,
+ "learning_rate": 0.0006,
+ "loss": 4.676113128662109,
+ "step": 4224
+ },
+ {
+ "epoch": 58.68471615720524,
+ "grad_norm": 0.01735256239771843,
+ "learning_rate": 0.0006,
+ "loss": 4.62088680267334,
+ "step": 4225
+ },
+ {
+ "epoch": 58.698689956331876,
+ "grad_norm": 0.019323458895087242,
+ "learning_rate": 0.0006,
+ "loss": 4.579405307769775,
+ "step": 4226
+ },
+ {
+ "epoch": 58.712663755458514,
+ "grad_norm": 0.01760544814169407,
+ "learning_rate": 0.0006,
+ "loss": 4.625407695770264,
+ "step": 4227
+ },
+ {
+ "epoch": 58.72663755458515,
+ "grad_norm": 0.015593250282108784,
+ "learning_rate": 0.0006,
+ "loss": 4.6326117515563965,
+ "step": 4228
+ },
+ {
+ "epoch": 58.74061135371179,
+ "grad_norm": 0.014357523061335087,
+ "learning_rate": 0.0006,
+ "loss": 4.6869401931762695,
+ "step": 4229
+ },
+ {
+ "epoch": 58.75458515283843,
+ "grad_norm": 0.01604357920587063,
+ "learning_rate": 0.0006,
+ "loss": 4.690784931182861,
+ "step": 4230
+ },
+ {
+ "epoch": 58.76855895196506,
+ "grad_norm": 0.01673845760524273,
+ "learning_rate": 0.0006,
+ "loss": 4.513410568237305,
+ "step": 4231
+ },
+ {
+ "epoch": 58.7825327510917,
+ "grad_norm": 0.013657779432833195,
+ "learning_rate": 0.0006,
+ "loss": 4.637057781219482,
+ "step": 4232
+ },
+ {
+ "epoch": 58.79650655021834,
+ "grad_norm": 0.013373114168643951,
+ "learning_rate": 0.0006,
+ "loss": 4.719967842102051,
+ "step": 4233
+ },
+ {
+ "epoch": 58.81048034934498,
+ "grad_norm": 0.016345568001270294,
+ "learning_rate": 0.0006,
+ "loss": 4.604109764099121,
+ "step": 4234
+ },
+ {
+ "epoch": 58.82445414847162,
+ "grad_norm": 0.01573033444583416,
+ "learning_rate": 0.0006,
+ "loss": 4.6471428871154785,
+ "step": 4235
+ },
+ {
+ "epoch": 58.838427947598255,
+ "grad_norm": 0.01517449039965868,
+ "learning_rate": 0.0006,
+ "loss": 4.722684860229492,
+ "step": 4236
+ },
+ {
+ "epoch": 58.852401746724894,
+ "grad_norm": 0.0145102022215724,
+ "learning_rate": 0.0006,
+ "loss": 4.662420272827148,
+ "step": 4237
+ },
+ {
+ "epoch": 58.866375545851525,
+ "grad_norm": 0.01457217987626791,
+ "learning_rate": 0.0006,
+ "loss": 4.585169792175293,
+ "step": 4238
+ },
+ {
+ "epoch": 58.880349344978164,
+ "grad_norm": 0.01406773366034031,
+ "learning_rate": 0.0006,
+ "loss": 4.5413665771484375,
+ "step": 4239
+ },
+ {
+ "epoch": 58.8943231441048,
+ "grad_norm": 0.012653871439397335,
+ "learning_rate": 0.0006,
+ "loss": 4.553636074066162,
+ "step": 4240
+ },
+ {
+ "epoch": 58.90829694323144,
+ "grad_norm": 0.01261200476437807,
+ "learning_rate": 0.0006,
+ "loss": 4.534193992614746,
+ "step": 4241
+ },
+ {
+ "epoch": 58.92227074235808,
+ "grad_norm": 0.012955864891409874,
+ "learning_rate": 0.0006,
+ "loss": 4.697248458862305,
+ "step": 4242
+ },
+ {
+ "epoch": 58.93624454148472,
+ "grad_norm": 0.012134749442338943,
+ "learning_rate": 0.0006,
+ "loss": 4.557095527648926,
+ "step": 4243
+ },
+ {
+ "epoch": 58.95021834061135,
+ "grad_norm": 0.012164677493274212,
+ "learning_rate": 0.0006,
+ "loss": 4.623579025268555,
+ "step": 4244
+ },
+ {
+ "epoch": 58.96419213973799,
+ "grad_norm": 0.013802947476506233,
+ "learning_rate": 0.0006,
+ "loss": 4.562838077545166,
+ "step": 4245
+ },
+ {
+ "epoch": 58.97816593886463,
+ "grad_norm": 0.011420476250350475,
+ "learning_rate": 0.0006,
+ "loss": 4.62265682220459,
+ "step": 4246
+ },
+ {
+ "epoch": 58.992139737991266,
+ "grad_norm": 0.01169576682150364,
+ "learning_rate": 0.0006,
+ "loss": 4.727449417114258,
+ "step": 4247
+ },
+ {
+ "epoch": 59.0,
+ "grad_norm": 0.013577710837125778,
+ "learning_rate": 0.0006,
+ "loss": 4.536325454711914,
+ "step": 4248
+ },
+ {
+ "epoch": 59.0,
+ "eval_loss": 4.86527681350708,
+ "eval_runtime": 60.4463,
+ "eval_samples_per_second": 40.399,
+ "eval_steps_per_second": 1.274,
+ "step": 4248
+ },
+ {
+ "epoch": 59.01397379912664,
+ "grad_norm": 0.015251655131578445,
+ "learning_rate": 0.0006,
+ "loss": 4.634598731994629,
+ "step": 4249
+ },
+ {
+ "epoch": 59.02794759825328,
+ "grad_norm": 0.017414938658475876,
+ "learning_rate": 0.0006,
+ "loss": 4.521329879760742,
+ "step": 4250
+ },
+ {
+ "epoch": 59.041921397379916,
+ "grad_norm": 0.01656588353216648,
+ "learning_rate": 0.0006,
+ "loss": 4.446192264556885,
+ "step": 4251
+ },
+ {
+ "epoch": 59.05589519650655,
+ "grad_norm": 0.012038925662636757,
+ "learning_rate": 0.0006,
+ "loss": 4.523855209350586,
+ "step": 4252
+ },
+ {
+ "epoch": 59.069868995633186,
+ "grad_norm": 0.014403033070266247,
+ "learning_rate": 0.0006,
+ "loss": 4.561079978942871,
+ "step": 4253
+ },
+ {
+ "epoch": 59.083842794759825,
+ "grad_norm": 0.016243597492575645,
+ "learning_rate": 0.0006,
+ "loss": 4.673602104187012,
+ "step": 4254
+ },
+ {
+ "epoch": 59.09781659388646,
+ "grad_norm": 0.016492877155542374,
+ "learning_rate": 0.0006,
+ "loss": 4.564824104309082,
+ "step": 4255
+ },
+ {
+ "epoch": 59.1117903930131,
+ "grad_norm": 0.015980370342731476,
+ "learning_rate": 0.0006,
+ "loss": 4.339249134063721,
+ "step": 4256
+ },
+ {
+ "epoch": 59.12576419213974,
+ "grad_norm": 0.019687356427311897,
+ "learning_rate": 0.0006,
+ "loss": 4.481834888458252,
+ "step": 4257
+ },
+ {
+ "epoch": 59.13973799126637,
+ "grad_norm": 0.02983798086643219,
+ "learning_rate": 0.0006,
+ "loss": 4.4495697021484375,
+ "step": 4258
+ },
+ {
+ "epoch": 59.15371179039301,
+ "grad_norm": 0.04600981995463371,
+ "learning_rate": 0.0006,
+ "loss": 4.567838668823242,
+ "step": 4259
+ },
+ {
+ "epoch": 59.16768558951965,
+ "grad_norm": 0.04852001741528511,
+ "learning_rate": 0.0006,
+ "loss": 4.594701766967773,
+ "step": 4260
+ },
+ {
+ "epoch": 59.18165938864629,
+ "grad_norm": 0.03429504856467247,
+ "learning_rate": 0.0006,
+ "loss": 4.573735237121582,
+ "step": 4261
+ },
+ {
+ "epoch": 59.19563318777293,
+ "grad_norm": 0.029345160350203514,
+ "learning_rate": 0.0006,
+ "loss": 4.488258361816406,
+ "step": 4262
+ },
+ {
+ "epoch": 59.209606986899566,
+ "grad_norm": 0.01961623504757881,
+ "learning_rate": 0.0006,
+ "loss": 4.545222282409668,
+ "step": 4263
+ },
+ {
+ "epoch": 59.223580786026204,
+ "grad_norm": 0.020406130701303482,
+ "learning_rate": 0.0006,
+ "loss": 4.471898078918457,
+ "step": 4264
+ },
+ {
+ "epoch": 59.237554585152836,
+ "grad_norm": 0.020500272512435913,
+ "learning_rate": 0.0006,
+ "loss": 4.538337707519531,
+ "step": 4265
+ },
+ {
+ "epoch": 59.251528384279474,
+ "grad_norm": 0.020727017894387245,
+ "learning_rate": 0.0006,
+ "loss": 4.437284469604492,
+ "step": 4266
+ },
+ {
+ "epoch": 59.26550218340611,
+ "grad_norm": 0.03395998477935791,
+ "learning_rate": 0.0006,
+ "loss": 4.459124565124512,
+ "step": 4267
+ },
+ {
+ "epoch": 59.27947598253275,
+ "grad_norm": 0.04635605961084366,
+ "learning_rate": 0.0006,
+ "loss": 4.520228385925293,
+ "step": 4268
+ },
+ {
+ "epoch": 59.29344978165939,
+ "grad_norm": 0.04253006726503372,
+ "learning_rate": 0.0006,
+ "loss": 4.578673362731934,
+ "step": 4269
+ },
+ {
+ "epoch": 59.30742358078603,
+ "grad_norm": 0.02403509058058262,
+ "learning_rate": 0.0006,
+ "loss": 4.484639644622803,
+ "step": 4270
+ },
+ {
+ "epoch": 59.32139737991266,
+ "grad_norm": 0.02254190482199192,
+ "learning_rate": 0.0006,
+ "loss": 4.511314868927002,
+ "step": 4271
+ },
+ {
+ "epoch": 59.3353711790393,
+ "grad_norm": 0.021585190668702126,
+ "learning_rate": 0.0006,
+ "loss": 4.530239105224609,
+ "step": 4272
+ },
+ {
+ "epoch": 59.34934497816594,
+ "grad_norm": 0.021210316568613052,
+ "learning_rate": 0.0006,
+ "loss": 4.593514442443848,
+ "step": 4273
+ },
+ {
+ "epoch": 59.36331877729258,
+ "grad_norm": 0.019317107275128365,
+ "learning_rate": 0.0006,
+ "loss": 4.5591535568237305,
+ "step": 4274
+ },
+ {
+ "epoch": 59.377292576419215,
+ "grad_norm": 0.018059708178043365,
+ "learning_rate": 0.0006,
+ "loss": 4.4499592781066895,
+ "step": 4275
+ },
+ {
+ "epoch": 59.391266375545854,
+ "grad_norm": 0.017847446724772453,
+ "learning_rate": 0.0006,
+ "loss": 4.473999500274658,
+ "step": 4276
+ },
+ {
+ "epoch": 59.40524017467249,
+ "grad_norm": 0.014944756403565407,
+ "learning_rate": 0.0006,
+ "loss": 4.5934529304504395,
+ "step": 4277
+ },
+ {
+ "epoch": 59.419213973799124,
+ "grad_norm": 0.01663832925260067,
+ "learning_rate": 0.0006,
+ "loss": 4.504262924194336,
+ "step": 4278
+ },
+ {
+ "epoch": 59.43318777292576,
+ "grad_norm": 0.015571796335279942,
+ "learning_rate": 0.0006,
+ "loss": 4.512310981750488,
+ "step": 4279
+ },
+ {
+ "epoch": 59.4471615720524,
+ "grad_norm": 0.015896400436758995,
+ "learning_rate": 0.0006,
+ "loss": 4.452714920043945,
+ "step": 4280
+ },
+ {
+ "epoch": 59.46113537117904,
+ "grad_norm": 0.016010284423828125,
+ "learning_rate": 0.0006,
+ "loss": 4.6188063621521,
+ "step": 4281
+ },
+ {
+ "epoch": 59.47510917030568,
+ "grad_norm": 0.015075593255460262,
+ "learning_rate": 0.0006,
+ "loss": 4.477978229522705,
+ "step": 4282
+ },
+ {
+ "epoch": 59.48908296943232,
+ "grad_norm": 0.014698950573801994,
+ "learning_rate": 0.0006,
+ "loss": 4.544955730438232,
+ "step": 4283
+ },
+ {
+ "epoch": 59.50305676855895,
+ "grad_norm": 0.0136245833709836,
+ "learning_rate": 0.0006,
+ "loss": 4.530460357666016,
+ "step": 4284
+ },
+ {
+ "epoch": 59.51703056768559,
+ "grad_norm": 0.013118310831487179,
+ "learning_rate": 0.0006,
+ "loss": 4.614217758178711,
+ "step": 4285
+ },
+ {
+ "epoch": 59.531004366812226,
+ "grad_norm": 0.01295961532741785,
+ "learning_rate": 0.0006,
+ "loss": 4.43156099319458,
+ "step": 4286
+ },
+ {
+ "epoch": 59.544978165938865,
+ "grad_norm": 0.012429811991751194,
+ "learning_rate": 0.0006,
+ "loss": 4.517333030700684,
+ "step": 4287
+ },
+ {
+ "epoch": 59.5589519650655,
+ "grad_norm": 0.01261830423027277,
+ "learning_rate": 0.0006,
+ "loss": 4.569780349731445,
+ "step": 4288
+ },
+ {
+ "epoch": 59.57292576419214,
+ "grad_norm": 0.012394499965012074,
+ "learning_rate": 0.0006,
+ "loss": 4.5970869064331055,
+ "step": 4289
+ },
+ {
+ "epoch": 59.58689956331878,
+ "grad_norm": 0.01290759164839983,
+ "learning_rate": 0.0006,
+ "loss": 4.437100410461426,
+ "step": 4290
+ },
+ {
+ "epoch": 59.60087336244541,
+ "grad_norm": 0.012042169459164143,
+ "learning_rate": 0.0006,
+ "loss": 4.567404747009277,
+ "step": 4291
+ },
+ {
+ "epoch": 59.61484716157205,
+ "grad_norm": 0.012338520959019661,
+ "learning_rate": 0.0006,
+ "loss": 4.476193428039551,
+ "step": 4292
+ },
+ {
+ "epoch": 59.62882096069869,
+ "grad_norm": 0.01321258582174778,
+ "learning_rate": 0.0006,
+ "loss": 4.566680431365967,
+ "step": 4293
+ },
+ {
+ "epoch": 59.64279475982533,
+ "grad_norm": 0.013478122651576996,
+ "learning_rate": 0.0006,
+ "loss": 4.487398147583008,
+ "step": 4294
+ },
+ {
+ "epoch": 59.65676855895197,
+ "grad_norm": 0.01537603884935379,
+ "learning_rate": 0.0006,
+ "loss": 4.449337005615234,
+ "step": 4295
+ },
+ {
+ "epoch": 59.670742358078606,
+ "grad_norm": 0.017140565440058708,
+ "learning_rate": 0.0006,
+ "loss": 4.5174360275268555,
+ "step": 4296
+ },
+ {
+ "epoch": 59.68471615720524,
+ "grad_norm": 0.017931461334228516,
+ "learning_rate": 0.0006,
+ "loss": 4.522104740142822,
+ "step": 4297
+ },
+ {
+ "epoch": 59.698689956331876,
+ "grad_norm": 0.018037918955087662,
+ "learning_rate": 0.0006,
+ "loss": 4.496928691864014,
+ "step": 4298
+ },
+ {
+ "epoch": 59.712663755458514,
+ "grad_norm": 0.01684504561126232,
+ "learning_rate": 0.0006,
+ "loss": 4.5332417488098145,
+ "step": 4299
+ },
+ {
+ "epoch": 59.72663755458515,
+ "grad_norm": 0.016268383711576462,
+ "learning_rate": 0.0006,
+ "loss": 4.468893051147461,
+ "step": 4300
+ },
+ {
+ "epoch": 59.74061135371179,
+ "grad_norm": 0.014859385788440704,
+ "learning_rate": 0.0006,
+ "loss": 4.551092147827148,
+ "step": 4301
+ },
+ {
+ "epoch": 59.75458515283843,
+ "grad_norm": 0.013589047826826572,
+ "learning_rate": 0.0006,
+ "loss": 4.470579147338867,
+ "step": 4302
+ },
+ {
+ "epoch": 59.76855895196506,
+ "grad_norm": 0.015704551711678505,
+ "learning_rate": 0.0006,
+ "loss": 4.330068111419678,
+ "step": 4303
+ },
+ {
+ "epoch": 59.7825327510917,
+ "grad_norm": 0.017869921401143074,
+ "learning_rate": 0.0006,
+ "loss": 4.428255081176758,
+ "step": 4304
+ },
+ {
+ "epoch": 59.79650655021834,
+ "grad_norm": 0.019828537479043007,
+ "learning_rate": 0.0006,
+ "loss": 4.496023178100586,
+ "step": 4305
+ },
+ {
+ "epoch": 59.81048034934498,
+ "grad_norm": 0.015856236219406128,
+ "learning_rate": 0.0006,
+ "loss": 4.427152156829834,
+ "step": 4306
+ },
+ {
+ "epoch": 59.82445414847162,
+ "grad_norm": 0.013349304907023907,
+ "learning_rate": 0.0006,
+ "loss": 4.417333602905273,
+ "step": 4307
+ },
+ {
+ "epoch": 59.838427947598255,
+ "grad_norm": 0.014111637137830257,
+ "learning_rate": 0.0006,
+ "loss": 4.4507575035095215,
+ "step": 4308
+ },
+ {
+ "epoch": 59.852401746724894,
+ "grad_norm": 0.015186597593128681,
+ "learning_rate": 0.0006,
+ "loss": 4.375116348266602,
+ "step": 4309
+ },
+ {
+ "epoch": 59.866375545851525,
+ "grad_norm": 0.01604730449616909,
+ "learning_rate": 0.0006,
+ "loss": 4.604518890380859,
+ "step": 4310
+ },
+ {
+ "epoch": 59.880349344978164,
+ "grad_norm": 0.015308464877307415,
+ "learning_rate": 0.0006,
+ "loss": 4.507232189178467,
+ "step": 4311
+ },
+ {
+ "epoch": 59.8943231441048,
+ "grad_norm": 0.014017206616699696,
+ "learning_rate": 0.0006,
+ "loss": 4.398487091064453,
+ "step": 4312
+ },
+ {
+ "epoch": 59.90829694323144,
+ "grad_norm": 0.012911394238471985,
+ "learning_rate": 0.0006,
+ "loss": 4.567872047424316,
+ "step": 4313
+ },
+ {
+ "epoch": 59.92227074235808,
+ "grad_norm": 0.012775142677128315,
+ "learning_rate": 0.0006,
+ "loss": 4.562105655670166,
+ "step": 4314
+ },
+ {
+ "epoch": 59.93624454148472,
+ "grad_norm": 0.014750408008694649,
+ "learning_rate": 0.0006,
+ "loss": 4.456263542175293,
+ "step": 4315
+ },
+ {
+ "epoch": 59.95021834061135,
+ "grad_norm": 0.016726011410355568,
+ "learning_rate": 0.0006,
+ "loss": 4.360722064971924,
+ "step": 4316
+ },
+ {
+ "epoch": 59.96419213973799,
+ "grad_norm": 0.01734776981174946,
+ "learning_rate": 0.0006,
+ "loss": 4.572544574737549,
+ "step": 4317
+ },
+ {
+ "epoch": 59.97816593886463,
+ "grad_norm": 0.01822415366768837,
+ "learning_rate": 0.0006,
+ "loss": 4.490501880645752,
+ "step": 4318
+ },
+ {
+ "epoch": 59.992139737991266,
+ "grad_norm": 0.014638577587902546,
+ "learning_rate": 0.0006,
+ "loss": 4.440951824188232,
+ "step": 4319
+ },
+ {
+ "epoch": 60.0,
+ "grad_norm": 0.01255202479660511,
+ "learning_rate": 0.0006,
+ "loss": 4.559123516082764,
+ "step": 4320
+ },
+ {
+ "epoch": 60.0,
+ "eval_loss": 4.741089344024658,
+ "eval_runtime": 56.8997,
+ "eval_samples_per_second": 42.918,
+ "eval_steps_per_second": 1.353,
+ "step": 4320
+ },
+ {
+ "epoch": 60.01397379912664,
+ "grad_norm": 0.016548393294215202,
+ "learning_rate": 0.0006,
+ "loss": 4.442024230957031,
+ "step": 4321
+ },
+ {
+ "epoch": 60.02794759825328,
+ "grad_norm": 0.019759127870202065,
+ "learning_rate": 0.0006,
+ "loss": 4.394334316253662,
+ "step": 4322
+ },
+ {
+ "epoch": 60.041921397379916,
+ "grad_norm": 0.018325170502066612,
+ "learning_rate": 0.0006,
+ "loss": 4.318270683288574,
+ "step": 4323
+ },
+ {
+ "epoch": 60.05589519650655,
+ "grad_norm": 0.015289463102817535,
+ "learning_rate": 0.0006,
+ "loss": 4.318228721618652,
+ "step": 4324
+ },
+ {
+ "epoch": 60.069868995633186,
+ "grad_norm": 0.015424872748553753,
+ "learning_rate": 0.0006,
+ "loss": 4.3605499267578125,
+ "step": 4325
+ },
+ {
+ "epoch": 60.083842794759825,
+ "grad_norm": 0.014085669070482254,
+ "learning_rate": 0.0006,
+ "loss": 4.359585762023926,
+ "step": 4326
+ },
+ {
+ "epoch": 60.09781659388646,
+ "grad_norm": 0.014004210010170937,
+ "learning_rate": 0.0006,
+ "loss": 4.447783470153809,
+ "step": 4327
+ },
+ {
+ "epoch": 60.1117903930131,
+ "grad_norm": 0.013514582999050617,
+ "learning_rate": 0.0006,
+ "loss": 4.4850239753723145,
+ "step": 4328
+ },
+ {
+ "epoch": 60.12576419213974,
+ "grad_norm": 0.013049394823610783,
+ "learning_rate": 0.0006,
+ "loss": 4.528675556182861,
+ "step": 4329
+ },
+ {
+ "epoch": 60.13973799126637,
+ "grad_norm": 0.01378058921545744,
+ "learning_rate": 0.0006,
+ "loss": 4.444823741912842,
+ "step": 4330
+ },
+ {
+ "epoch": 60.15371179039301,
+ "grad_norm": 0.013752233237028122,
+ "learning_rate": 0.0006,
+ "loss": 4.5205206871032715,
+ "step": 4331
+ },
+ {
+ "epoch": 60.16768558951965,
+ "grad_norm": 0.013431803323328495,
+ "learning_rate": 0.0006,
+ "loss": 4.547320365905762,
+ "step": 4332
+ },
+ {
+ "epoch": 60.18165938864629,
+ "grad_norm": 0.014544196426868439,
+ "learning_rate": 0.0006,
+ "loss": 4.517143249511719,
+ "step": 4333
+ },
+ {
+ "epoch": 60.19563318777293,
+ "grad_norm": 0.01612176187336445,
+ "learning_rate": 0.0006,
+ "loss": 4.517232894897461,
+ "step": 4334
+ },
+ {
+ "epoch": 60.209606986899566,
+ "grad_norm": 0.01803239807486534,
+ "learning_rate": 0.0006,
+ "loss": 4.377574920654297,
+ "step": 4335
+ },
+ {
+ "epoch": 60.223580786026204,
+ "grad_norm": 0.014871489256620407,
+ "learning_rate": 0.0006,
+ "loss": 4.538285255432129,
+ "step": 4336
+ },
+ {
+ "epoch": 60.237554585152836,
+ "grad_norm": 0.013927377760410309,
+ "learning_rate": 0.0006,
+ "loss": 4.386175632476807,
+ "step": 4337
+ },
+ {
+ "epoch": 60.251528384279474,
+ "grad_norm": 0.014757219702005386,
+ "learning_rate": 0.0006,
+ "loss": 4.528553009033203,
+ "step": 4338
+ },
+ {
+ "epoch": 60.26550218340611,
+ "grad_norm": 0.01433873176574707,
+ "learning_rate": 0.0006,
+ "loss": 4.476929187774658,
+ "step": 4339
+ },
+ {
+ "epoch": 60.27947598253275,
+ "grad_norm": 0.015768803656101227,
+ "learning_rate": 0.0006,
+ "loss": 4.50970458984375,
+ "step": 4340
+ },
+ {
+ "epoch": 60.29344978165939,
+ "grad_norm": 0.015651429072022438,
+ "learning_rate": 0.0006,
+ "loss": 4.427133083343506,
+ "step": 4341
+ },
+ {
+ "epoch": 60.30742358078603,
+ "grad_norm": 0.01673000678420067,
+ "learning_rate": 0.0006,
+ "loss": 4.447723865509033,
+ "step": 4342
+ },
+ {
+ "epoch": 60.32139737991266,
+ "grad_norm": 0.018481194972991943,
+ "learning_rate": 0.0006,
+ "loss": 4.559832572937012,
+ "step": 4343
+ },
+ {
+ "epoch": 60.3353711790393,
+ "grad_norm": 0.016731027513742447,
+ "learning_rate": 0.0006,
+ "loss": 4.470893859863281,
+ "step": 4344
+ },
+ {
+ "epoch": 60.34934497816594,
+ "grad_norm": 0.01504011545330286,
+ "learning_rate": 0.0006,
+ "loss": 4.483014106750488,
+ "step": 4345
+ },
+ {
+ "epoch": 60.36331877729258,
+ "grad_norm": 0.01602456159889698,
+ "learning_rate": 0.0006,
+ "loss": 4.548428058624268,
+ "step": 4346
+ },
+ {
+ "epoch": 60.377292576419215,
+ "grad_norm": 0.016027364879846573,
+ "learning_rate": 0.0006,
+ "loss": 4.426779747009277,
+ "step": 4347
+ },
+ {
+ "epoch": 60.391266375545854,
+ "grad_norm": 0.01873827911913395,
+ "learning_rate": 0.0006,
+ "loss": 4.353686332702637,
+ "step": 4348
+ },
+ {
+ "epoch": 60.40524017467249,
+ "grad_norm": 0.01907406374812126,
+ "learning_rate": 0.0006,
+ "loss": 4.502945423126221,
+ "step": 4349
+ },
+ {
+ "epoch": 60.419213973799124,
+ "grad_norm": 0.016083354130387306,
+ "learning_rate": 0.0006,
+ "loss": 4.431731700897217,
+ "step": 4350
+ },
+ {
+ "epoch": 60.43318777292576,
+ "grad_norm": 0.015293709933757782,
+ "learning_rate": 0.0006,
+ "loss": 4.480112552642822,
+ "step": 4351
+ },
+ {
+ "epoch": 60.4471615720524,
+ "grad_norm": 0.01688101328909397,
+ "learning_rate": 0.0006,
+ "loss": 4.50853157043457,
+ "step": 4352
+ },
+ {
+ "epoch": 60.46113537117904,
+ "grad_norm": 0.02094118855893612,
+ "learning_rate": 0.0006,
+ "loss": 4.359930038452148,
+ "step": 4353
+ },
+ {
+ "epoch": 60.47510917030568,
+ "grad_norm": 0.018188107758760452,
+ "learning_rate": 0.0006,
+ "loss": 4.549592018127441,
+ "step": 4354
+ },
+ {
+ "epoch": 60.48908296943232,
+ "grad_norm": 0.01462319865822792,
+ "learning_rate": 0.0006,
+ "loss": 4.49467658996582,
+ "step": 4355
+ },
+ {
+ "epoch": 60.50305676855895,
+ "grad_norm": 0.017508579418063164,
+ "learning_rate": 0.0006,
+ "loss": 4.398318767547607,
+ "step": 4356
+ },
+ {
+ "epoch": 60.51703056768559,
+ "grad_norm": 0.018200315535068512,
+ "learning_rate": 0.0006,
+ "loss": 4.44169282913208,
+ "step": 4357
+ },
+ {
+ "epoch": 60.531004366812226,
+ "grad_norm": 0.01717468351125717,
+ "learning_rate": 0.0006,
+ "loss": 4.481400489807129,
+ "step": 4358
+ },
+ {
+ "epoch": 60.544978165938865,
+ "grad_norm": 0.019329151138663292,
+ "learning_rate": 0.0006,
+ "loss": 4.578839302062988,
+ "step": 4359
+ },
+ {
+ "epoch": 60.5589519650655,
+ "grad_norm": 0.016109555959701538,
+ "learning_rate": 0.0006,
+ "loss": 4.461043357849121,
+ "step": 4360
+ },
+ {
+ "epoch": 60.57292576419214,
+ "grad_norm": 0.01494457945227623,
+ "learning_rate": 0.0006,
+ "loss": 4.505096912384033,
+ "step": 4361
+ },
+ {
+ "epoch": 60.58689956331878,
+ "grad_norm": 0.01729399710893631,
+ "learning_rate": 0.0006,
+ "loss": 4.366747856140137,
+ "step": 4362
+ },
+ {
+ "epoch": 60.60087336244541,
+ "grad_norm": 0.01705821603536606,
+ "learning_rate": 0.0006,
+ "loss": 4.42534065246582,
+ "step": 4363
+ },
+ {
+ "epoch": 60.61484716157205,
+ "grad_norm": 0.018391354009509087,
+ "learning_rate": 0.0006,
+ "loss": 4.350294589996338,
+ "step": 4364
+ },
+ {
+ "epoch": 60.62882096069869,
+ "grad_norm": 0.016989829018712044,
+ "learning_rate": 0.0006,
+ "loss": 4.45905876159668,
+ "step": 4365
+ },
+ {
+ "epoch": 60.64279475982533,
+ "grad_norm": 0.017192212864756584,
+ "learning_rate": 0.0006,
+ "loss": 4.526487350463867,
+ "step": 4366
+ },
+ {
+ "epoch": 60.65676855895197,
+ "grad_norm": 0.018091315403580666,
+ "learning_rate": 0.0006,
+ "loss": 4.309333324432373,
+ "step": 4367
+ },
+ {
+ "epoch": 60.670742358078606,
+ "grad_norm": 0.014051459729671478,
+ "learning_rate": 0.0006,
+ "loss": 4.338171005249023,
+ "step": 4368
+ },
+ {
+ "epoch": 60.68471615720524,
+ "grad_norm": 0.013166418299078941,
+ "learning_rate": 0.0006,
+ "loss": 4.409758567810059,
+ "step": 4369
+ },
+ {
+ "epoch": 60.698689956331876,
+ "grad_norm": 0.014260428957641125,
+ "learning_rate": 0.0006,
+ "loss": 4.524338245391846,
+ "step": 4370
+ },
+ {
+ "epoch": 60.712663755458514,
+ "grad_norm": 0.017019255086779594,
+ "learning_rate": 0.0006,
+ "loss": 4.387436866760254,
+ "step": 4371
+ },
+ {
+ "epoch": 60.72663755458515,
+ "grad_norm": 0.017070675268769264,
+ "learning_rate": 0.0006,
+ "loss": 4.437321662902832,
+ "step": 4372
+ },
+ {
+ "epoch": 60.74061135371179,
+ "grad_norm": 0.015197164379060268,
+ "learning_rate": 0.0006,
+ "loss": 4.371205806732178,
+ "step": 4373
+ },
+ {
+ "epoch": 60.75458515283843,
+ "grad_norm": 0.014791185967624187,
+ "learning_rate": 0.0006,
+ "loss": 4.44439697265625,
+ "step": 4374
+ },
+ {
+ "epoch": 60.76855895196506,
+ "grad_norm": 0.016458261758089066,
+ "learning_rate": 0.0006,
+ "loss": 4.445584297180176,
+ "step": 4375
+ },
+ {
+ "epoch": 60.7825327510917,
+ "grad_norm": 0.017706342041492462,
+ "learning_rate": 0.0006,
+ "loss": 4.348708152770996,
+ "step": 4376
+ },
+ {
+ "epoch": 60.79650655021834,
+ "grad_norm": 0.017561476677656174,
+ "learning_rate": 0.0006,
+ "loss": 4.521929740905762,
+ "step": 4377
+ },
+ {
+ "epoch": 60.81048034934498,
+ "grad_norm": 0.016557445749640465,
+ "learning_rate": 0.0006,
+ "loss": 4.545806884765625,
+ "step": 4378
+ },
+ {
+ "epoch": 60.82445414847162,
+ "grad_norm": 0.016953222453594208,
+ "learning_rate": 0.0006,
+ "loss": 4.4230875968933105,
+ "step": 4379
+ },
+ {
+ "epoch": 60.838427947598255,
+ "grad_norm": 0.017094967886805534,
+ "learning_rate": 0.0006,
+ "loss": 4.423985481262207,
+ "step": 4380
+ },
+ {
+ "epoch": 60.852401746724894,
+ "grad_norm": 0.014787515625357628,
+ "learning_rate": 0.0006,
+ "loss": 4.39473819732666,
+ "step": 4381
+ },
+ {
+ "epoch": 60.866375545851525,
+ "grad_norm": 0.016208263114094734,
+ "learning_rate": 0.0006,
+ "loss": 4.3810038566589355,
+ "step": 4382
+ },
+ {
+ "epoch": 60.880349344978164,
+ "grad_norm": 0.017844321206212044,
+ "learning_rate": 0.0006,
+ "loss": 4.35086727142334,
+ "step": 4383
+ },
+ {
+ "epoch": 60.8943231441048,
+ "grad_norm": 0.017539294436573982,
+ "learning_rate": 0.0006,
+ "loss": 4.460260391235352,
+ "step": 4384
+ },
+ {
+ "epoch": 60.90829694323144,
+ "grad_norm": 0.01612042263150215,
+ "learning_rate": 0.0006,
+ "loss": 4.370500564575195,
+ "step": 4385
+ },
+ {
+ "epoch": 60.92227074235808,
+ "grad_norm": 0.015381601639091969,
+ "learning_rate": 0.0006,
+ "loss": 4.435868263244629,
+ "step": 4386
+ },
+ {
+ "epoch": 60.93624454148472,
+ "grad_norm": 0.01603585295379162,
+ "learning_rate": 0.0006,
+ "loss": 4.434070587158203,
+ "step": 4387
+ },
+ {
+ "epoch": 60.95021834061135,
+ "grad_norm": 0.016268130391836166,
+ "learning_rate": 0.0006,
+ "loss": 4.404343128204346,
+ "step": 4388
+ },
+ {
+ "epoch": 60.96419213973799,
+ "grad_norm": 0.014980032108724117,
+ "learning_rate": 0.0006,
+ "loss": 4.399213790893555,
+ "step": 4389
+ },
+ {
+ "epoch": 60.97816593886463,
+ "grad_norm": 0.016899267211556435,
+ "learning_rate": 0.0006,
+ "loss": 4.508069038391113,
+ "step": 4390
+ },
+ {
+ "epoch": 60.992139737991266,
+ "grad_norm": 0.013756908476352692,
+ "learning_rate": 0.0006,
+ "loss": 4.541858673095703,
+ "step": 4391
+ },
+ {
+ "epoch": 61.0,
+ "grad_norm": 0.01578413136303425,
+ "learning_rate": 0.0006,
+ "loss": 4.543513298034668,
+ "step": 4392
+ },
+ {
+ "epoch": 61.0,
+ "eval_loss": 4.756770610809326,
+ "eval_runtime": 56.4572,
+ "eval_samples_per_second": 43.254,
+ "eval_steps_per_second": 1.364,
+ "step": 4392
+ },
+ {
+ "epoch": 61.01397379912664,
+ "grad_norm": 0.013387506827712059,
+ "learning_rate": 0.0006,
+ "loss": 4.475942611694336,
+ "step": 4393
+ },
+ {
+ "epoch": 61.02794759825328,
+ "grad_norm": 0.014278904534876347,
+ "learning_rate": 0.0006,
+ "loss": 4.442403793334961,
+ "step": 4394
+ },
+ {
+ "epoch": 61.041921397379916,
+ "grad_norm": 0.014359553344547749,
+ "learning_rate": 0.0006,
+ "loss": 4.509463310241699,
+ "step": 4395
+ },
+ {
+ "epoch": 61.05589519650655,
+ "grad_norm": 0.01718050241470337,
+ "learning_rate": 0.0006,
+ "loss": 4.354458808898926,
+ "step": 4396
+ },
+ {
+ "epoch": 61.069868995633186,
+ "grad_norm": 0.015988312661647797,
+ "learning_rate": 0.0006,
+ "loss": 4.484508514404297,
+ "step": 4397
+ },
+ {
+ "epoch": 61.083842794759825,
+ "grad_norm": 0.015466735698282719,
+ "learning_rate": 0.0006,
+ "loss": 4.326908111572266,
+ "step": 4398
+ },
+ {
+ "epoch": 61.09781659388646,
+ "grad_norm": 0.014541332609951496,
+ "learning_rate": 0.0006,
+ "loss": 4.405823230743408,
+ "step": 4399
+ },
+ {
+ "epoch": 61.1117903930131,
+ "grad_norm": 0.014803987927734852,
+ "learning_rate": 0.0006,
+ "loss": 4.375386714935303,
+ "step": 4400
+ },
+ {
+ "epoch": 61.12576419213974,
+ "grad_norm": 0.013962093740701675,
+ "learning_rate": 0.0006,
+ "loss": 4.300838470458984,
+ "step": 4401
+ },
+ {
+ "epoch": 61.13973799126637,
+ "grad_norm": 0.015213954262435436,
+ "learning_rate": 0.0006,
+ "loss": 4.388306617736816,
+ "step": 4402
+ },
+ {
+ "epoch": 61.15371179039301,
+ "grad_norm": 0.018965506926178932,
+ "learning_rate": 0.0006,
+ "loss": 4.388340950012207,
+ "step": 4403
+ },
+ {
+ "epoch": 61.16768558951965,
+ "grad_norm": 0.01958410255610943,
+ "learning_rate": 0.0006,
+ "loss": 4.35012149810791,
+ "step": 4404
+ },
+ {
+ "epoch": 61.18165938864629,
+ "grad_norm": 0.01938101276755333,
+ "learning_rate": 0.0006,
+ "loss": 4.368372917175293,
+ "step": 4405
+ },
+ {
+ "epoch": 61.19563318777293,
+ "grad_norm": 0.016556516289711,
+ "learning_rate": 0.0006,
+ "loss": 4.486198425292969,
+ "step": 4406
+ },
+ {
+ "epoch": 61.209606986899566,
+ "grad_norm": 0.017031289637088776,
+ "learning_rate": 0.0006,
+ "loss": 4.421676158905029,
+ "step": 4407
+ },
+ {
+ "epoch": 61.223580786026204,
+ "grad_norm": 0.018023479729890823,
+ "learning_rate": 0.0006,
+ "loss": 4.38751220703125,
+ "step": 4408
+ },
+ {
+ "epoch": 61.237554585152836,
+ "grad_norm": 0.0178135447204113,
+ "learning_rate": 0.0006,
+ "loss": 4.336528301239014,
+ "step": 4409
+ },
+ {
+ "epoch": 61.251528384279474,
+ "grad_norm": 0.016782190650701523,
+ "learning_rate": 0.0006,
+ "loss": 4.386693954467773,
+ "step": 4410
+ },
+ {
+ "epoch": 61.26550218340611,
+ "grad_norm": 0.015494848601520061,
+ "learning_rate": 0.0006,
+ "loss": 4.308948040008545,
+ "step": 4411
+ },
+ {
+ "epoch": 61.27947598253275,
+ "grad_norm": 0.015002378262579441,
+ "learning_rate": 0.0006,
+ "loss": 4.314206123352051,
+ "step": 4412
+ },
+ {
+ "epoch": 61.29344978165939,
+ "grad_norm": 0.015798920765519142,
+ "learning_rate": 0.0006,
+ "loss": 4.502900123596191,
+ "step": 4413
+ },
+ {
+ "epoch": 61.30742358078603,
+ "grad_norm": 0.015444336459040642,
+ "learning_rate": 0.0006,
+ "loss": 4.473160743713379,
+ "step": 4414
+ },
+ {
+ "epoch": 61.32139737991266,
+ "grad_norm": 0.017025096341967583,
+ "learning_rate": 0.0006,
+ "loss": 4.460621356964111,
+ "step": 4415
+ },
+ {
+ "epoch": 61.3353711790393,
+ "grad_norm": 0.019502537325024605,
+ "learning_rate": 0.0006,
+ "loss": 4.385551452636719,
+ "step": 4416
+ },
+ {
+ "epoch": 61.34934497816594,
+ "grad_norm": 0.018529541790485382,
+ "learning_rate": 0.0006,
+ "loss": 4.425390720367432,
+ "step": 4417
+ },
+ {
+ "epoch": 61.36331877729258,
+ "grad_norm": 0.013155706226825714,
+ "learning_rate": 0.0006,
+ "loss": 4.467020034790039,
+ "step": 4418
+ },
+ {
+ "epoch": 61.377292576419215,
+ "grad_norm": 0.01518090907484293,
+ "learning_rate": 0.0006,
+ "loss": 4.369664192199707,
+ "step": 4419
+ },
+ {
+ "epoch": 61.391266375545854,
+ "grad_norm": 0.01626511849462986,
+ "learning_rate": 0.0006,
+ "loss": 4.463231086730957,
+ "step": 4420
+ },
+ {
+ "epoch": 61.40524017467249,
+ "grad_norm": 0.01701558195054531,
+ "learning_rate": 0.0006,
+ "loss": 4.43708610534668,
+ "step": 4421
+ },
+ {
+ "epoch": 61.419213973799124,
+ "grad_norm": 0.01761862449347973,
+ "learning_rate": 0.0006,
+ "loss": 4.3314208984375,
+ "step": 4422
+ },
+ {
+ "epoch": 61.43318777292576,
+ "grad_norm": 0.017084648832678795,
+ "learning_rate": 0.0006,
+ "loss": 4.39676570892334,
+ "step": 4423
+ },
+ {
+ "epoch": 61.4471615720524,
+ "grad_norm": 0.01667078025639057,
+ "learning_rate": 0.0006,
+ "loss": 4.3789873123168945,
+ "step": 4424
+ },
+ {
+ "epoch": 61.46113537117904,
+ "grad_norm": 0.01877436414361,
+ "learning_rate": 0.0006,
+ "loss": 4.415275573730469,
+ "step": 4425
+ },
+ {
+ "epoch": 61.47510917030568,
+ "grad_norm": 0.020250486209988594,
+ "learning_rate": 0.0006,
+ "loss": 4.478635787963867,
+ "step": 4426
+ },
+ {
+ "epoch": 61.48908296943232,
+ "grad_norm": 0.020244868472218513,
+ "learning_rate": 0.0006,
+ "loss": 4.296854019165039,
+ "step": 4427
+ },
+ {
+ "epoch": 61.50305676855895,
+ "grad_norm": 0.018723875284194946,
+ "learning_rate": 0.0006,
+ "loss": 4.3760881423950195,
+ "step": 4428
+ },
+ {
+ "epoch": 61.51703056768559,
+ "grad_norm": 0.015138164162635803,
+ "learning_rate": 0.0006,
+ "loss": 4.39974308013916,
+ "step": 4429
+ },
+ {
+ "epoch": 61.531004366812226,
+ "grad_norm": 0.015111138112843037,
+ "learning_rate": 0.0006,
+ "loss": 4.34659481048584,
+ "step": 4430
+ },
+ {
+ "epoch": 61.544978165938865,
+ "grad_norm": 0.0164666585624218,
+ "learning_rate": 0.0006,
+ "loss": 4.415737152099609,
+ "step": 4431
+ },
+ {
+ "epoch": 61.5589519650655,
+ "grad_norm": 0.020891312509775162,
+ "learning_rate": 0.0006,
+ "loss": 4.300319671630859,
+ "step": 4432
+ },
+ {
+ "epoch": 61.57292576419214,
+ "grad_norm": 0.02089156024158001,
+ "learning_rate": 0.0006,
+ "loss": 4.373029708862305,
+ "step": 4433
+ },
+ {
+ "epoch": 61.58689956331878,
+ "grad_norm": 0.020114559680223465,
+ "learning_rate": 0.0006,
+ "loss": 4.434499740600586,
+ "step": 4434
+ },
+ {
+ "epoch": 61.60087336244541,
+ "grad_norm": 0.01791365258395672,
+ "learning_rate": 0.0006,
+ "loss": 4.466872215270996,
+ "step": 4435
+ },
+ {
+ "epoch": 61.61484716157205,
+ "grad_norm": 0.01879458874464035,
+ "learning_rate": 0.0006,
+ "loss": 4.5176520347595215,
+ "step": 4436
+ },
+ {
+ "epoch": 61.62882096069869,
+ "grad_norm": 0.016371937468647957,
+ "learning_rate": 0.0006,
+ "loss": 4.3507585525512695,
+ "step": 4437
+ },
+ {
+ "epoch": 61.64279475982533,
+ "grad_norm": 0.01648143120110035,
+ "learning_rate": 0.0006,
+ "loss": 4.398778438568115,
+ "step": 4438
+ },
+ {
+ "epoch": 61.65676855895197,
+ "grad_norm": 0.017240576446056366,
+ "learning_rate": 0.0006,
+ "loss": 4.513933181762695,
+ "step": 4439
+ },
+ {
+ "epoch": 61.670742358078606,
+ "grad_norm": 0.01734284684062004,
+ "learning_rate": 0.0006,
+ "loss": 4.315343856811523,
+ "step": 4440
+ },
+ {
+ "epoch": 61.68471615720524,
+ "grad_norm": 0.01639772206544876,
+ "learning_rate": 0.0006,
+ "loss": 4.4641218185424805,
+ "step": 4441
+ },
+ {
+ "epoch": 61.698689956331876,
+ "grad_norm": 0.0164708960801363,
+ "learning_rate": 0.0006,
+ "loss": 4.439968585968018,
+ "step": 4442
+ },
+ {
+ "epoch": 61.712663755458514,
+ "grad_norm": 0.01757933758199215,
+ "learning_rate": 0.0006,
+ "loss": 4.408327102661133,
+ "step": 4443
+ },
+ {
+ "epoch": 61.72663755458515,
+ "grad_norm": 0.01789589412510395,
+ "learning_rate": 0.0006,
+ "loss": 4.478257179260254,
+ "step": 4444
+ },
+ {
+ "epoch": 61.74061135371179,
+ "grad_norm": 0.01822184957563877,
+ "learning_rate": 0.0006,
+ "loss": 4.412046909332275,
+ "step": 4445
+ },
+ {
+ "epoch": 61.75458515283843,
+ "grad_norm": 0.018624618649482727,
+ "learning_rate": 0.0006,
+ "loss": 4.427122116088867,
+ "step": 4446
+ },
+ {
+ "epoch": 61.76855895196506,
+ "grad_norm": 0.018674887716770172,
+ "learning_rate": 0.0006,
+ "loss": 4.429147720336914,
+ "step": 4447
+ },
+ {
+ "epoch": 61.7825327510917,
+ "grad_norm": 0.015317119657993317,
+ "learning_rate": 0.0006,
+ "loss": 4.2808451652526855,
+ "step": 4448
+ },
+ {
+ "epoch": 61.79650655021834,
+ "grad_norm": 0.015400845557451248,
+ "learning_rate": 0.0006,
+ "loss": 4.420193672180176,
+ "step": 4449
+ },
+ {
+ "epoch": 61.81048034934498,
+ "grad_norm": 0.016743650659918785,
+ "learning_rate": 0.0006,
+ "loss": 4.498254776000977,
+ "step": 4450
+ },
+ {
+ "epoch": 61.82445414847162,
+ "grad_norm": 0.014330781064927578,
+ "learning_rate": 0.0006,
+ "loss": 4.445384979248047,
+ "step": 4451
+ },
+ {
+ "epoch": 61.838427947598255,
+ "grad_norm": 0.015046479180455208,
+ "learning_rate": 0.0006,
+ "loss": 4.368041038513184,
+ "step": 4452
+ },
+ {
+ "epoch": 61.852401746724894,
+ "grad_norm": 0.01630168780684471,
+ "learning_rate": 0.0006,
+ "loss": 4.460142135620117,
+ "step": 4453
+ },
+ {
+ "epoch": 61.866375545851525,
+ "grad_norm": 0.01637885719537735,
+ "learning_rate": 0.0006,
+ "loss": 4.416606426239014,
+ "step": 4454
+ },
+ {
+ "epoch": 61.880349344978164,
+ "grad_norm": 0.016785888001322746,
+ "learning_rate": 0.0006,
+ "loss": 4.452545642852783,
+ "step": 4455
+ },
+ {
+ "epoch": 61.8943231441048,
+ "grad_norm": 0.015044555068016052,
+ "learning_rate": 0.0006,
+ "loss": 4.427334785461426,
+ "step": 4456
+ },
+ {
+ "epoch": 61.90829694323144,
+ "grad_norm": 0.013699211180210114,
+ "learning_rate": 0.0006,
+ "loss": 4.408761024475098,
+ "step": 4457
+ },
+ {
+ "epoch": 61.92227074235808,
+ "grad_norm": 0.014021803624927998,
+ "learning_rate": 0.0006,
+ "loss": 4.543607711791992,
+ "step": 4458
+ },
+ {
+ "epoch": 61.93624454148472,
+ "grad_norm": 0.015438460744917393,
+ "learning_rate": 0.0006,
+ "loss": 4.313363075256348,
+ "step": 4459
+ },
+ {
+ "epoch": 61.95021834061135,
+ "grad_norm": 0.017515428364276886,
+ "learning_rate": 0.0006,
+ "loss": 4.310412406921387,
+ "step": 4460
+ },
+ {
+ "epoch": 61.96419213973799,
+ "grad_norm": 0.016683880239725113,
+ "learning_rate": 0.0006,
+ "loss": 4.3378448486328125,
+ "step": 4461
+ },
+ {
+ "epoch": 61.97816593886463,
+ "grad_norm": 0.019384048879146576,
+ "learning_rate": 0.0006,
+ "loss": 4.450568675994873,
+ "step": 4462
+ },
+ {
+ "epoch": 61.992139737991266,
+ "grad_norm": 0.020651493221521378,
+ "learning_rate": 0.0006,
+ "loss": 4.453197479248047,
+ "step": 4463
+ },
+ {
+ "epoch": 62.0,
+ "grad_norm": 0.020991764962673187,
+ "learning_rate": 0.0006,
+ "loss": 4.441527366638184,
+ "step": 4464
+ },
+ {
+ "epoch": 62.0,
+ "eval_loss": 4.700193881988525,
+ "eval_runtime": 56.6098,
+ "eval_samples_per_second": 43.137,
+ "eval_steps_per_second": 1.36,
+ "step": 4464
+ },
+ {
+ "epoch": 62.01397379912664,
+ "grad_norm": 0.01811373233795166,
+ "learning_rate": 0.0006,
+ "loss": 4.37346076965332,
+ "step": 4465
+ },
+ {
+ "epoch": 62.02794759825328,
+ "grad_norm": 0.01667127199470997,
+ "learning_rate": 0.0006,
+ "loss": 4.237151145935059,
+ "step": 4466
+ },
+ {
+ "epoch": 62.041921397379916,
+ "grad_norm": 0.014971534721553326,
+ "learning_rate": 0.0006,
+ "loss": 4.367053031921387,
+ "step": 4467
+ },
+ {
+ "epoch": 62.05589519650655,
+ "grad_norm": 0.015655893832445145,
+ "learning_rate": 0.0006,
+ "loss": 4.437426567077637,
+ "step": 4468
+ },
+ {
+ "epoch": 62.069868995633186,
+ "grad_norm": 0.014935447834432125,
+ "learning_rate": 0.0006,
+ "loss": 4.3635125160217285,
+ "step": 4469
+ },
+ {
+ "epoch": 62.083842794759825,
+ "grad_norm": 0.015363802202045918,
+ "learning_rate": 0.0006,
+ "loss": 4.40492057800293,
+ "step": 4470
+ },
+ {
+ "epoch": 62.09781659388646,
+ "grad_norm": 0.0159201230853796,
+ "learning_rate": 0.0006,
+ "loss": 4.392058849334717,
+ "step": 4471
+ },
+ {
+ "epoch": 62.1117903930131,
+ "grad_norm": 0.014598241075873375,
+ "learning_rate": 0.0006,
+ "loss": 4.369944095611572,
+ "step": 4472
+ },
+ {
+ "epoch": 62.12576419213974,
+ "grad_norm": 0.015394841320812702,
+ "learning_rate": 0.0006,
+ "loss": 4.425097942352295,
+ "step": 4473
+ },
+ {
+ "epoch": 62.13973799126637,
+ "grad_norm": 0.015913628041744232,
+ "learning_rate": 0.0006,
+ "loss": 4.420596122741699,
+ "step": 4474
+ },
+ {
+ "epoch": 62.15371179039301,
+ "grad_norm": 0.01575258933007717,
+ "learning_rate": 0.0006,
+ "loss": 4.32290506362915,
+ "step": 4475
+ },
+ {
+ "epoch": 62.16768558951965,
+ "grad_norm": 0.0155344782397151,
+ "learning_rate": 0.0006,
+ "loss": 4.273068428039551,
+ "step": 4476
+ },
+ {
+ "epoch": 62.18165938864629,
+ "grad_norm": 0.014947040006518364,
+ "learning_rate": 0.0006,
+ "loss": 4.384868621826172,
+ "step": 4477
+ },
+ {
+ "epoch": 62.19563318777293,
+ "grad_norm": 0.01557591650635004,
+ "learning_rate": 0.0006,
+ "loss": 4.458410263061523,
+ "step": 4478
+ },
+ {
+ "epoch": 62.209606986899566,
+ "grad_norm": 0.015182198956608772,
+ "learning_rate": 0.0006,
+ "loss": 4.300738334655762,
+ "step": 4479
+ },
+ {
+ "epoch": 62.223580786026204,
+ "grad_norm": 0.01634717360138893,
+ "learning_rate": 0.0006,
+ "loss": 4.436875820159912,
+ "step": 4480
+ },
+ {
+ "epoch": 62.237554585152836,
+ "grad_norm": 0.018813718110322952,
+ "learning_rate": 0.0006,
+ "loss": 4.454665184020996,
+ "step": 4481
+ },
+ {
+ "epoch": 62.251528384279474,
+ "grad_norm": 0.020965630188584328,
+ "learning_rate": 0.0006,
+ "loss": 4.357341766357422,
+ "step": 4482
+ },
+ {
+ "epoch": 62.26550218340611,
+ "grad_norm": 0.020738353952765465,
+ "learning_rate": 0.0006,
+ "loss": 4.456644058227539,
+ "step": 4483
+ },
+ {
+ "epoch": 62.27947598253275,
+ "grad_norm": 0.018813546746969223,
+ "learning_rate": 0.0006,
+ "loss": 4.381941795349121,
+ "step": 4484
+ },
+ {
+ "epoch": 62.29344978165939,
+ "grad_norm": 0.015431761741638184,
+ "learning_rate": 0.0006,
+ "loss": 4.41126823425293,
+ "step": 4485
+ },
+ {
+ "epoch": 62.30742358078603,
+ "grad_norm": 0.016335977241396904,
+ "learning_rate": 0.0006,
+ "loss": 4.419488906860352,
+ "step": 4486
+ },
+ {
+ "epoch": 62.32139737991266,
+ "grad_norm": 0.01832406409084797,
+ "learning_rate": 0.0006,
+ "loss": 4.579026222229004,
+ "step": 4487
+ },
+ {
+ "epoch": 62.3353711790393,
+ "grad_norm": 0.01790742017328739,
+ "learning_rate": 0.0006,
+ "loss": 4.289964199066162,
+ "step": 4488
+ },
+ {
+ "epoch": 62.34934497816594,
+ "grad_norm": 0.017266202718019485,
+ "learning_rate": 0.0006,
+ "loss": 4.415959358215332,
+ "step": 4489
+ },
+ {
+ "epoch": 62.36331877729258,
+ "grad_norm": 0.015976132825016975,
+ "learning_rate": 0.0006,
+ "loss": 4.358587265014648,
+ "step": 4490
+ },
+ {
+ "epoch": 62.377292576419215,
+ "grad_norm": 0.016873881220817566,
+ "learning_rate": 0.0006,
+ "loss": 4.354607105255127,
+ "step": 4491
+ },
+ {
+ "epoch": 62.391266375545854,
+ "grad_norm": 0.01476745493710041,
+ "learning_rate": 0.0006,
+ "loss": 4.384904384613037,
+ "step": 4492
+ },
+ {
+ "epoch": 62.40524017467249,
+ "grad_norm": 0.013810782693326473,
+ "learning_rate": 0.0006,
+ "loss": 4.459230899810791,
+ "step": 4493
+ },
+ {
+ "epoch": 62.419213973799124,
+ "grad_norm": 0.015150421299040318,
+ "learning_rate": 0.0006,
+ "loss": 4.369471549987793,
+ "step": 4494
+ },
+ {
+ "epoch": 62.43318777292576,
+ "grad_norm": 0.014539610594511032,
+ "learning_rate": 0.0006,
+ "loss": 4.3564982414245605,
+ "step": 4495
+ },
+ {
+ "epoch": 62.4471615720524,
+ "grad_norm": 0.015554017387330532,
+ "learning_rate": 0.0006,
+ "loss": 4.414379119873047,
+ "step": 4496
+ },
+ {
+ "epoch": 62.46113537117904,
+ "grad_norm": 0.016608357429504395,
+ "learning_rate": 0.0006,
+ "loss": 4.311585426330566,
+ "step": 4497
+ },
+ {
+ "epoch": 62.47510917030568,
+ "grad_norm": 0.01885881833732128,
+ "learning_rate": 0.0006,
+ "loss": 4.306629180908203,
+ "step": 4498
+ },
+ {
+ "epoch": 62.48908296943232,
+ "grad_norm": 0.022128663957118988,
+ "learning_rate": 0.0006,
+ "loss": 4.345483779907227,
+ "step": 4499
+ },
+ {
+ "epoch": 62.50305676855895,
+ "grad_norm": 0.023401018232107162,
+ "learning_rate": 0.0006,
+ "loss": 4.398204326629639,
+ "step": 4500
+ },
+ {
+ "epoch": 62.51703056768559,
+ "grad_norm": 0.022080009803175926,
+ "learning_rate": 0.0006,
+ "loss": 4.375027179718018,
+ "step": 4501
+ },
+ {
+ "epoch": 62.531004366812226,
+ "grad_norm": 0.020004864782094955,
+ "learning_rate": 0.0006,
+ "loss": 4.33494758605957,
+ "step": 4502
+ },
+ {
+ "epoch": 62.544978165938865,
+ "grad_norm": 0.01786700263619423,
+ "learning_rate": 0.0006,
+ "loss": 4.4428300857543945,
+ "step": 4503
+ },
+ {
+ "epoch": 62.5589519650655,
+ "grad_norm": 0.01885388419032097,
+ "learning_rate": 0.0006,
+ "loss": 4.4138946533203125,
+ "step": 4504
+ },
+ {
+ "epoch": 62.57292576419214,
+ "grad_norm": 0.019890939816832542,
+ "learning_rate": 0.0006,
+ "loss": 4.278632164001465,
+ "step": 4505
+ },
+ {
+ "epoch": 62.58689956331878,
+ "grad_norm": 0.02109031192958355,
+ "learning_rate": 0.0006,
+ "loss": 4.3756327629089355,
+ "step": 4506
+ },
+ {
+ "epoch": 62.60087336244541,
+ "grad_norm": 0.020813899114727974,
+ "learning_rate": 0.0006,
+ "loss": 4.410213470458984,
+ "step": 4507
+ },
+ {
+ "epoch": 62.61484716157205,
+ "grad_norm": 0.021655689924955368,
+ "learning_rate": 0.0006,
+ "loss": 4.30633020401001,
+ "step": 4508
+ },
+ {
+ "epoch": 62.62882096069869,
+ "grad_norm": 0.018155649304389954,
+ "learning_rate": 0.0006,
+ "loss": 4.407284736633301,
+ "step": 4509
+ },
+ {
+ "epoch": 62.64279475982533,
+ "grad_norm": 0.017765069380402565,
+ "learning_rate": 0.0006,
+ "loss": 4.321715354919434,
+ "step": 4510
+ },
+ {
+ "epoch": 62.65676855895197,
+ "grad_norm": 0.017724554985761642,
+ "learning_rate": 0.0006,
+ "loss": 4.493371963500977,
+ "step": 4511
+ },
+ {
+ "epoch": 62.670742358078606,
+ "grad_norm": 0.01492092851549387,
+ "learning_rate": 0.0006,
+ "loss": 4.443025588989258,
+ "step": 4512
+ },
+ {
+ "epoch": 62.68471615720524,
+ "grad_norm": 0.01615324057638645,
+ "learning_rate": 0.0006,
+ "loss": 4.3289594650268555,
+ "step": 4513
+ },
+ {
+ "epoch": 62.698689956331876,
+ "grad_norm": 0.016569366678595543,
+ "learning_rate": 0.0006,
+ "loss": 4.369720458984375,
+ "step": 4514
+ },
+ {
+ "epoch": 62.712663755458514,
+ "grad_norm": 0.016509901732206345,
+ "learning_rate": 0.0006,
+ "loss": 4.292316436767578,
+ "step": 4515
+ },
+ {
+ "epoch": 62.72663755458515,
+ "grad_norm": 0.014270003885030746,
+ "learning_rate": 0.0006,
+ "loss": 4.349854469299316,
+ "step": 4516
+ },
+ {
+ "epoch": 62.74061135371179,
+ "grad_norm": 0.014235340990126133,
+ "learning_rate": 0.0006,
+ "loss": 4.473727226257324,
+ "step": 4517
+ },
+ {
+ "epoch": 62.75458515283843,
+ "grad_norm": 0.015163225121796131,
+ "learning_rate": 0.0006,
+ "loss": 4.452878475189209,
+ "step": 4518
+ },
+ {
+ "epoch": 62.76855895196506,
+ "grad_norm": 0.015748055651783943,
+ "learning_rate": 0.0006,
+ "loss": 4.521087169647217,
+ "step": 4519
+ },
+ {
+ "epoch": 62.7825327510917,
+ "grad_norm": 0.01661108434200287,
+ "learning_rate": 0.0006,
+ "loss": 4.443885326385498,
+ "step": 4520
+ },
+ {
+ "epoch": 62.79650655021834,
+ "grad_norm": 0.01610301062464714,
+ "learning_rate": 0.0006,
+ "loss": 4.312990188598633,
+ "step": 4521
+ },
+ {
+ "epoch": 62.81048034934498,
+ "grad_norm": 0.017007583752274513,
+ "learning_rate": 0.0006,
+ "loss": 4.408937454223633,
+ "step": 4522
+ },
+ {
+ "epoch": 62.82445414847162,
+ "grad_norm": 0.01749461516737938,
+ "learning_rate": 0.0006,
+ "loss": 4.346066951751709,
+ "step": 4523
+ },
+ {
+ "epoch": 62.838427947598255,
+ "grad_norm": 0.018243493512272835,
+ "learning_rate": 0.0006,
+ "loss": 4.357658386230469,
+ "step": 4524
+ },
+ {
+ "epoch": 62.852401746724894,
+ "grad_norm": 0.017499985173344612,
+ "learning_rate": 0.0006,
+ "loss": 4.374061584472656,
+ "step": 4525
+ },
+ {
+ "epoch": 62.866375545851525,
+ "grad_norm": 0.014536325819790363,
+ "learning_rate": 0.0006,
+ "loss": 4.189321994781494,
+ "step": 4526
+ },
+ {
+ "epoch": 62.880349344978164,
+ "grad_norm": 0.014728891663253307,
+ "learning_rate": 0.0006,
+ "loss": 4.425910949707031,
+ "step": 4527
+ },
+ {
+ "epoch": 62.8943231441048,
+ "grad_norm": 0.015013772062957287,
+ "learning_rate": 0.0006,
+ "loss": 4.337319850921631,
+ "step": 4528
+ },
+ {
+ "epoch": 62.90829694323144,
+ "grad_norm": 0.01626027375459671,
+ "learning_rate": 0.0006,
+ "loss": 4.330339431762695,
+ "step": 4529
+ },
+ {
+ "epoch": 62.92227074235808,
+ "grad_norm": 0.014773444272577763,
+ "learning_rate": 0.0006,
+ "loss": 4.4188737869262695,
+ "step": 4530
+ },
+ {
+ "epoch": 62.93624454148472,
+ "grad_norm": 0.01422932744026184,
+ "learning_rate": 0.0006,
+ "loss": 4.355986595153809,
+ "step": 4531
+ },
+ {
+ "epoch": 62.95021834061135,
+ "grad_norm": 0.012170110829174519,
+ "learning_rate": 0.0006,
+ "loss": 4.405722618103027,
+ "step": 4532
+ },
+ {
+ "epoch": 62.96419213973799,
+ "grad_norm": 0.013233253732323647,
+ "learning_rate": 0.0006,
+ "loss": 4.377286911010742,
+ "step": 4533
+ },
+ {
+ "epoch": 62.97816593886463,
+ "grad_norm": 0.014497299678623676,
+ "learning_rate": 0.0006,
+ "loss": 4.508796691894531,
+ "step": 4534
+ },
+ {
+ "epoch": 62.992139737991266,
+ "grad_norm": 0.014718741178512573,
+ "learning_rate": 0.0006,
+ "loss": 4.370919704437256,
+ "step": 4535
+ },
+ {
+ "epoch": 63.0,
+ "grad_norm": 0.015833502635359764,
+ "learning_rate": 0.0006,
+ "loss": 4.258753776550293,
+ "step": 4536
+ },
+ {
+ "epoch": 63.0,
+ "eval_loss": 4.714163303375244,
+ "eval_runtime": 57.0269,
+ "eval_samples_per_second": 42.822,
+ "eval_steps_per_second": 1.35,
+ "step": 4536
+ },
+ {
+ "epoch": 63.01397379912664,
+ "grad_norm": 0.014277510344982147,
+ "learning_rate": 0.0006,
+ "loss": 4.3112640380859375,
+ "step": 4537
+ },
+ {
+ "epoch": 63.02794759825328,
+ "grad_norm": 0.015377935022115707,
+ "learning_rate": 0.0006,
+ "loss": 4.342782974243164,
+ "step": 4538
+ },
+ {
+ "epoch": 63.041921397379916,
+ "grad_norm": 0.015406244434416294,
+ "learning_rate": 0.0006,
+ "loss": 4.474056720733643,
+ "step": 4539
+ },
+ {
+ "epoch": 63.05589519650655,
+ "grad_norm": 0.018121499568223953,
+ "learning_rate": 0.0006,
+ "loss": 4.397963523864746,
+ "step": 4540
+ },
+ {
+ "epoch": 63.069868995633186,
+ "grad_norm": 0.022634636610746384,
+ "learning_rate": 0.0006,
+ "loss": 4.306631088256836,
+ "step": 4541
+ },
+ {
+ "epoch": 63.083842794759825,
+ "grad_norm": 0.026448970660567284,
+ "learning_rate": 0.0006,
+ "loss": 4.283175468444824,
+ "step": 4542
+ },
+ {
+ "epoch": 63.09781659388646,
+ "grad_norm": 0.024104537442326546,
+ "learning_rate": 0.0006,
+ "loss": 4.315826416015625,
+ "step": 4543
+ },
+ {
+ "epoch": 63.1117903930131,
+ "grad_norm": 0.019040480256080627,
+ "learning_rate": 0.0006,
+ "loss": 4.44057035446167,
+ "step": 4544
+ },
+ {
+ "epoch": 63.12576419213974,
+ "grad_norm": 0.02006000280380249,
+ "learning_rate": 0.0006,
+ "loss": 4.405603408813477,
+ "step": 4545
+ },
+ {
+ "epoch": 63.13973799126637,
+ "grad_norm": 0.019980071112513542,
+ "learning_rate": 0.0006,
+ "loss": 4.393726348876953,
+ "step": 4546
+ },
+ {
+ "epoch": 63.15371179039301,
+ "grad_norm": 0.01641865260899067,
+ "learning_rate": 0.0006,
+ "loss": 4.272546768188477,
+ "step": 4547
+ },
+ {
+ "epoch": 63.16768558951965,
+ "grad_norm": 0.01890353485941887,
+ "learning_rate": 0.0006,
+ "loss": 4.382566928863525,
+ "step": 4548
+ },
+ {
+ "epoch": 63.18165938864629,
+ "grad_norm": 0.018370937556028366,
+ "learning_rate": 0.0006,
+ "loss": 4.423953533172607,
+ "step": 4549
+ },
+ {
+ "epoch": 63.19563318777293,
+ "grad_norm": 0.014985294081270695,
+ "learning_rate": 0.0006,
+ "loss": 4.308627128601074,
+ "step": 4550
+ },
+ {
+ "epoch": 63.209606986899566,
+ "grad_norm": 0.017513327300548553,
+ "learning_rate": 0.0006,
+ "loss": 4.248819351196289,
+ "step": 4551
+ },
+ {
+ "epoch": 63.223580786026204,
+ "grad_norm": 0.019464161247015,
+ "learning_rate": 0.0006,
+ "loss": 4.359729766845703,
+ "step": 4552
+ },
+ {
+ "epoch": 63.237554585152836,
+ "grad_norm": 0.018678035587072372,
+ "learning_rate": 0.0006,
+ "loss": 4.3299665451049805,
+ "step": 4553
+ },
+ {
+ "epoch": 63.251528384279474,
+ "grad_norm": 0.017909523099660873,
+ "learning_rate": 0.0006,
+ "loss": 4.246147155761719,
+ "step": 4554
+ },
+ {
+ "epoch": 63.26550218340611,
+ "grad_norm": 0.01747293211519718,
+ "learning_rate": 0.0006,
+ "loss": 4.41165018081665,
+ "step": 4555
+ },
+ {
+ "epoch": 63.27947598253275,
+ "grad_norm": 0.018670853227376938,
+ "learning_rate": 0.0006,
+ "loss": 4.409883499145508,
+ "step": 4556
+ },
+ {
+ "epoch": 63.29344978165939,
+ "grad_norm": 0.0189529862254858,
+ "learning_rate": 0.0006,
+ "loss": 4.418654918670654,
+ "step": 4557
+ },
+ {
+ "epoch": 63.30742358078603,
+ "grad_norm": 0.018394414335489273,
+ "learning_rate": 0.0006,
+ "loss": 4.38355827331543,
+ "step": 4558
+ },
+ {
+ "epoch": 63.32139737991266,
+ "grad_norm": 0.018317820504307747,
+ "learning_rate": 0.0006,
+ "loss": 4.456843376159668,
+ "step": 4559
+ },
+ {
+ "epoch": 63.3353711790393,
+ "grad_norm": 0.019225597381591797,
+ "learning_rate": 0.0006,
+ "loss": 4.232264518737793,
+ "step": 4560
+ },
+ {
+ "epoch": 63.34934497816594,
+ "grad_norm": 0.016663681715726852,
+ "learning_rate": 0.0006,
+ "loss": 4.393517017364502,
+ "step": 4561
+ },
+ {
+ "epoch": 63.36331877729258,
+ "grad_norm": 0.014975365251302719,
+ "learning_rate": 0.0006,
+ "loss": 4.375183582305908,
+ "step": 4562
+ },
+ {
+ "epoch": 63.377292576419215,
+ "grad_norm": 0.01771732047200203,
+ "learning_rate": 0.0006,
+ "loss": 4.3073410987854,
+ "step": 4563
+ },
+ {
+ "epoch": 63.391266375545854,
+ "grad_norm": 0.01663549244403839,
+ "learning_rate": 0.0006,
+ "loss": 4.311003684997559,
+ "step": 4564
+ },
+ {
+ "epoch": 63.40524017467249,
+ "grad_norm": 0.016301177442073822,
+ "learning_rate": 0.0006,
+ "loss": 4.326807022094727,
+ "step": 4565
+ },
+ {
+ "epoch": 63.419213973799124,
+ "grad_norm": 0.016781188547611237,
+ "learning_rate": 0.0006,
+ "loss": 4.306562423706055,
+ "step": 4566
+ },
+ {
+ "epoch": 63.43318777292576,
+ "grad_norm": 0.01767745055258274,
+ "learning_rate": 0.0006,
+ "loss": 4.367082595825195,
+ "step": 4567
+ },
+ {
+ "epoch": 63.4471615720524,
+ "grad_norm": 0.016728565096855164,
+ "learning_rate": 0.0006,
+ "loss": 4.327400207519531,
+ "step": 4568
+ },
+ {
+ "epoch": 63.46113537117904,
+ "grad_norm": 0.016289403662085533,
+ "learning_rate": 0.0006,
+ "loss": 4.392712116241455,
+ "step": 4569
+ },
+ {
+ "epoch": 63.47510917030568,
+ "grad_norm": 0.016233716160058975,
+ "learning_rate": 0.0006,
+ "loss": 4.354372024536133,
+ "step": 4570
+ },
+ {
+ "epoch": 63.48908296943232,
+ "grad_norm": 0.016547974199056625,
+ "learning_rate": 0.0006,
+ "loss": 4.309050559997559,
+ "step": 4571
+ },
+ {
+ "epoch": 63.50305676855895,
+ "grad_norm": 0.018466509878635406,
+ "learning_rate": 0.0006,
+ "loss": 4.318345069885254,
+ "step": 4572
+ },
+ {
+ "epoch": 63.51703056768559,
+ "grad_norm": 0.017685720697045326,
+ "learning_rate": 0.0006,
+ "loss": 4.399726867675781,
+ "step": 4573
+ },
+ {
+ "epoch": 63.531004366812226,
+ "grad_norm": 0.015801234170794487,
+ "learning_rate": 0.0006,
+ "loss": 4.328495502471924,
+ "step": 4574
+ },
+ {
+ "epoch": 63.544978165938865,
+ "grad_norm": 0.0188873540610075,
+ "learning_rate": 0.0006,
+ "loss": 4.253244400024414,
+ "step": 4575
+ },
+ {
+ "epoch": 63.5589519650655,
+ "grad_norm": 0.01868271268904209,
+ "learning_rate": 0.0006,
+ "loss": 4.463746070861816,
+ "step": 4576
+ },
+ {
+ "epoch": 63.57292576419214,
+ "grad_norm": 0.01844949834048748,
+ "learning_rate": 0.0006,
+ "loss": 4.400300025939941,
+ "step": 4577
+ },
+ {
+ "epoch": 63.58689956331878,
+ "grad_norm": 0.015256480313837528,
+ "learning_rate": 0.0006,
+ "loss": 4.445844650268555,
+ "step": 4578
+ },
+ {
+ "epoch": 63.60087336244541,
+ "grad_norm": 0.015305760316550732,
+ "learning_rate": 0.0006,
+ "loss": 4.317327499389648,
+ "step": 4579
+ },
+ {
+ "epoch": 63.61484716157205,
+ "grad_norm": 0.015200769528746605,
+ "learning_rate": 0.0006,
+ "loss": 4.25112247467041,
+ "step": 4580
+ },
+ {
+ "epoch": 63.62882096069869,
+ "grad_norm": 0.015919538214802742,
+ "learning_rate": 0.0006,
+ "loss": 4.277937889099121,
+ "step": 4581
+ },
+ {
+ "epoch": 63.64279475982533,
+ "grad_norm": 0.01572665013372898,
+ "learning_rate": 0.0006,
+ "loss": 4.302237510681152,
+ "step": 4582
+ },
+ {
+ "epoch": 63.65676855895197,
+ "grad_norm": 0.016428731381893158,
+ "learning_rate": 0.0006,
+ "loss": 4.36762809753418,
+ "step": 4583
+ },
+ {
+ "epoch": 63.670742358078606,
+ "grad_norm": 0.01706082746386528,
+ "learning_rate": 0.0006,
+ "loss": 4.236526966094971,
+ "step": 4584
+ },
+ {
+ "epoch": 63.68471615720524,
+ "grad_norm": 0.017850136384367943,
+ "learning_rate": 0.0006,
+ "loss": 4.264129638671875,
+ "step": 4585
+ },
+ {
+ "epoch": 63.698689956331876,
+ "grad_norm": 0.016079438850283623,
+ "learning_rate": 0.0006,
+ "loss": 4.449799060821533,
+ "step": 4586
+ },
+ {
+ "epoch": 63.712663755458514,
+ "grad_norm": 0.014996383339166641,
+ "learning_rate": 0.0006,
+ "loss": 4.404574394226074,
+ "step": 4587
+ },
+ {
+ "epoch": 63.72663755458515,
+ "grad_norm": 0.015910431742668152,
+ "learning_rate": 0.0006,
+ "loss": 4.348101615905762,
+ "step": 4588
+ },
+ {
+ "epoch": 63.74061135371179,
+ "grad_norm": 0.01630021072924137,
+ "learning_rate": 0.0006,
+ "loss": 4.3193817138671875,
+ "step": 4589
+ },
+ {
+ "epoch": 63.75458515283843,
+ "grad_norm": 0.014571773819625378,
+ "learning_rate": 0.0006,
+ "loss": 4.374402046203613,
+ "step": 4590
+ },
+ {
+ "epoch": 63.76855895196506,
+ "grad_norm": 0.01662345975637436,
+ "learning_rate": 0.0006,
+ "loss": 4.329075813293457,
+ "step": 4591
+ },
+ {
+ "epoch": 63.7825327510917,
+ "grad_norm": 0.016604198142886162,
+ "learning_rate": 0.0006,
+ "loss": 4.436126232147217,
+ "step": 4592
+ },
+ {
+ "epoch": 63.79650655021834,
+ "grad_norm": 0.016578884795308113,
+ "learning_rate": 0.0006,
+ "loss": 4.361468315124512,
+ "step": 4593
+ },
+ {
+ "epoch": 63.81048034934498,
+ "grad_norm": 0.016339825466275215,
+ "learning_rate": 0.0006,
+ "loss": 4.449892044067383,
+ "step": 4594
+ },
+ {
+ "epoch": 63.82445414847162,
+ "grad_norm": 0.016805455088615417,
+ "learning_rate": 0.0006,
+ "loss": 4.415866851806641,
+ "step": 4595
+ },
+ {
+ "epoch": 63.838427947598255,
+ "grad_norm": 0.017366213724017143,
+ "learning_rate": 0.0006,
+ "loss": 4.361730575561523,
+ "step": 4596
+ },
+ {
+ "epoch": 63.852401746724894,
+ "grad_norm": 0.015599003992974758,
+ "learning_rate": 0.0006,
+ "loss": 4.363637924194336,
+ "step": 4597
+ },
+ {
+ "epoch": 63.866375545851525,
+ "grad_norm": 0.015674632042646408,
+ "learning_rate": 0.0006,
+ "loss": 4.332045078277588,
+ "step": 4598
+ },
+ {
+ "epoch": 63.880349344978164,
+ "grad_norm": 0.015350443311035633,
+ "learning_rate": 0.0006,
+ "loss": 4.320063591003418,
+ "step": 4599
+ },
+ {
+ "epoch": 63.8943231441048,
+ "grad_norm": 0.015388146974146366,
+ "learning_rate": 0.0006,
+ "loss": 4.398199081420898,
+ "step": 4600
+ },
+ {
+ "epoch": 63.90829694323144,
+ "grad_norm": 0.01450140681117773,
+ "learning_rate": 0.0006,
+ "loss": 4.244076728820801,
+ "step": 4601
+ },
+ {
+ "epoch": 63.92227074235808,
+ "grad_norm": 0.01637491025030613,
+ "learning_rate": 0.0006,
+ "loss": 4.213549613952637,
+ "step": 4602
+ },
+ {
+ "epoch": 63.93624454148472,
+ "grad_norm": 0.0159814041107893,
+ "learning_rate": 0.0006,
+ "loss": 4.421775817871094,
+ "step": 4603
+ },
+ {
+ "epoch": 63.95021834061135,
+ "grad_norm": 0.017386827617883682,
+ "learning_rate": 0.0006,
+ "loss": 4.340023994445801,
+ "step": 4604
+ },
+ {
+ "epoch": 63.96419213973799,
+ "grad_norm": 0.019679656252264977,
+ "learning_rate": 0.0006,
+ "loss": 4.352534294128418,
+ "step": 4605
+ },
+ {
+ "epoch": 63.97816593886463,
+ "grad_norm": 0.022391516715288162,
+ "learning_rate": 0.0006,
+ "loss": 4.3538312911987305,
+ "step": 4606
+ },
+ {
+ "epoch": 63.992139737991266,
+ "grad_norm": 0.023281006142497063,
+ "learning_rate": 0.0006,
+ "loss": 4.432089328765869,
+ "step": 4607
+ },
+ {
+ "epoch": 64.0,
+ "grad_norm": 0.019916288554668427,
+ "learning_rate": 0.0006,
+ "loss": 4.466937065124512,
+ "step": 4608
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 28800,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 400,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": true,
+ "should_log": false,
+ "should_save": false,
+ "should_training_stop": false
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 1.9645525512526234e+18,
+ "train_batch_size": 8,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4608/training_args.bin b/runs/l2r50-i2-fulle-lm/checkpoint-4608/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..cad232831c222cc7f8b938b8bbe42f61e7d89c3a
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-4608/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:f7c40e3bfa2f5ac1450ca1243e3f2a895d44a06ed5995e6aefd97f78d229d8c8
+size 4792
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4896/chat_template.jinja b/runs/l2r50-i2-fulle-lm/checkpoint-4896/chat_template.jinja
new file mode 100644
index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-4896/chat_template.jinja
@@ -0,0 +1,85 @@
+{%- if tools %}
+ {{- '<|im_start|>system\n' }}
+ {%- if messages[0].role == 'system' %}
+ {{- messages[0].content + '\n\n' }}
+ {%- endif %}
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }}
+ {%- for tool in tools %}
+ {{- "\n" }}
+ {{- tool | tojson }}
+ {%- endfor %}
+ {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }}
+{%- else %}
+ {%- if messages[0].role == 'system' %}
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
+ {%- endif %}
+{%- endif %}
+{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
+{%- for message in messages[::-1] %}
+ {%- set index = (messages|length - 1) - loop.index0 %}
+ {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %}
+ {%- set ns.multi_step_tool = false %}
+ {%- set ns.last_query_index = index %}
+ {%- endif %}
+{%- endfor %}
+{%- for message in messages %}
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
+ {%- elif message.role == "assistant" %}
+ {%- set content = message.content %}
+ {%- set reasoning_content = '' %}
+ {%- if message.reasoning_content is defined and message.reasoning_content is not none %}
+ {%- set reasoning_content = message.reasoning_content %}
+ {%- else %}
+ {%- if '' in message.content %}
+ {%- set content = message.content.split('')[-1].lstrip('\n') %}
+ {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %}
+ {%- endif %}
+ {%- endif %}
+ {%- if loop.index0 > ns.last_query_index %}
+ {%- if loop.last or (not loop.last and reasoning_content) %}
+ {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- if message.tool_calls %}
+ {%- for tool_call in message.tool_calls %}
+ {%- if (loop.first and content) or (not loop.first) %}
+ {{- '\n' }}
+ {%- endif %}
+ {%- if tool_call.function %}
+ {%- set tool_call = tool_call.function %}
+ {%- endif %}
+ {{- '\n{"name": "' }}
+ {{- tool_call.name }}
+ {{- '", "arguments": ' }}
+ {%- if tool_call.arguments is string %}
+ {{- tool_call.arguments }}
+ {%- else %}
+ {{- tool_call.arguments | tojson }}
+ {%- endif %}
+ {{- '}\n' }}
+ {%- endfor %}
+ {%- endif %}
+ {{- '<|im_end|>\n' }}
+ {%- elif message.role == "tool" %}
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
+ {{- '<|im_start|>user' }}
+ {%- endif %}
+ {{- '\n\n' }}
+ {{- message.content }}
+ {{- '\n' }}
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
+ {{- '<|im_end|>\n' }}
+ {%- endif %}
+ {%- endif %}
+{%- endfor %}
+{%- if add_generation_prompt %}
+ {{- '<|im_start|>assistant\n' }}
+ {%- if enable_thinking is defined and enable_thinking is false %}
+ {{- '\n\n\n\n' }}
+ {%- endif %}
+{%- endif %}
\ No newline at end of file
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4896/config.json b/runs/l2r50-i2-fulle-lm/checkpoint-4896/config.json
new file mode 100644
index 0000000000000000000000000000000000000000..22f250953c6ce8fc91ee4411856dd73041a415e2
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-4896/config.json
@@ -0,0 +1,32 @@
+{
+ "architectures": [
+ "LlamaForCausalLM"
+ ],
+ "attention_bias": false,
+ "attention_dropout": 0.0,
+ "bos_token_id": null,
+ "dtype": "float32",
+ "eos_token_id": 151645,
+ "head_dim": 128,
+ "hidden_act": "silu",
+ "hidden_size": 512,
+ "initializer_range": 0.02,
+ "intermediate_size": 1536,
+ "max_position_embeddings": 2048,
+ "mlp_bias": false,
+ "model_type": "llama",
+ "num_attention_heads": 4,
+ "num_hidden_layers": 20,
+ "num_key_value_heads": 4,
+ "pad_token_id": 151645,
+ "pretraining_tp": 1,
+ "rms_norm_eps": 1e-06,
+ "rope_parameters": {
+ "rope_theta": 10000.0,
+ "rope_type": "default"
+ },
+ "tie_word_embeddings": true,
+ "transformers_version": "5.5.0",
+ "use_cache": false,
+ "vocab_size": 151673
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4896/generation_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-4896/generation_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-4896/generation_config.json
@@ -0,0 +1,11 @@
+{
+ "_from_model_config": true,
+ "eos_token_id": [
+ 151645
+ ],
+ "output_attentions": false,
+ "output_hidden_states": false,
+ "pad_token_id": 151645,
+ "transformers_version": "5.5.0",
+ "use_cache": true
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4896/model.safetensors b/runs/l2r50-i2-fulle-lm/checkpoint-4896/model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..6eda745eca2b3b50d1ce4441fcc02cb99bf30786
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-4896/model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:5df3108be16e0dcc8f841715dc42ed32fde017827e7e67080a1ea4d3447ca624
+size 583360328
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4896/optimizer.pt b/runs/l2r50-i2-fulle-lm/checkpoint-4896/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..13d3c7cfe6712d55c65b7a4a71190481285bca97
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-4896/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:159e7c5f95d1fd59d1a34afda01a4d02812101e4b25682aa81ddaa510d1e77ce
+size 1166833530
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4896/rng_state_0.pth b/runs/l2r50-i2-fulle-lm/checkpoint-4896/rng_state_0.pth
new file mode 100644
index 0000000000000000000000000000000000000000..750ff4ede83419f049eeb660a694b7e3ecc7c8d6
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-4896/rng_state_0.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:5fa62fb1de6e4743a12d17318debd141345b5a642afd38db7758d0634c7a7f13
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4896/rng_state_1.pth b/runs/l2r50-i2-fulle-lm/checkpoint-4896/rng_state_1.pth
new file mode 100644
index 0000000000000000000000000000000000000000..9ff82b1335a99fc6b3002849e9ff7716b929c804
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-4896/rng_state_1.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:4f69ce94754005caa11960e9c7c89ce4524851a59ae27cb35de69f711d23b218
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4896/rng_state_2.pth b/runs/l2r50-i2-fulle-lm/checkpoint-4896/rng_state_2.pth
new file mode 100644
index 0000000000000000000000000000000000000000..088a0e343cacf37055aca354fb38e71ffd425531
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-4896/rng_state_2.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:993b3848c8a8140c118f377b7499a5492beb127bdfe31a0d905e551e1c466859
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4896/rng_state_3.pth b/runs/l2r50-i2-fulle-lm/checkpoint-4896/rng_state_3.pth
new file mode 100644
index 0000000000000000000000000000000000000000..0610ab758a90414848478fab4b81e86746f261c3
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-4896/rng_state_3.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:2bd27efae4d1ce0c2c890f064a5e19e1258acc7c189fd707628a2140a1a99d14
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4896/scheduler.pt b/runs/l2r50-i2-fulle-lm/checkpoint-4896/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..f446f829d82a1f86c912714abb53d7fba343010b
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-4896/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:94642e081368d7b75b2ef306469e253305fe85f7bdeb714618031927a58d2a80
+size 1064
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4896/tokenizer.json b/runs/l2r50-i2-fulle-lm/checkpoint-4896/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..53786016f070e11184813f3c83101dc5e93e392e
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-4896/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:c3c63fda1646a555448e8f03662542e04b11d103721319e4d9b267d7da03ceb8
+size 11424463
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4896/tokenizer_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-4896/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..98e2fb9b2cc54785dc9d871b8e503f8b8f2defe7
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-4896/tokenizer_config.json
@@ -0,0 +1,20 @@
+{
+ "add_prefix_space": false,
+ "backend": "tokenizers",
+ "bos_token": null,
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "<|im_end|>",
+ "errors": "replace",
+ "extra_special_tokens": [
+ "<|l2r_pred|>",
+ "<|r2l_pred|>",
+ "<|next_1_pred|>",
+ "<|next_2_pred|>"
+ ],
+ "is_local": false,
+ "model_max_length": 131072,
+ "pad_token": "<|im_end|>",
+ "split_special_tokens": false,
+ "tokenizer_class": "Qwen2Tokenizer",
+ "unk_token": null
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4896/trainer_state.json b/runs/l2r50-i2-fulle-lm/checkpoint-4896/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..e7d06748e87af64234d992b80f80e04c74e9f768
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-4896/trainer_state.json
@@ -0,0 +1,34842 @@
+{
+ "best_global_step": null,
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 68.0,
+ "eval_steps": 500,
+ "global_step": 4896,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.013973799126637555,
+ "grad_norm": 0.1739506870508194,
+ "learning_rate": 0.0,
+ "loss": 11.99299430847168,
+ "step": 1
+ },
+ {
+ "epoch": 0.02794759825327511,
+ "grad_norm": 0.17216597497463226,
+ "learning_rate": 5.999999999999999e-06,
+ "loss": 11.988704681396484,
+ "step": 2
+ },
+ {
+ "epoch": 0.04192139737991266,
+ "grad_norm": 0.17459715902805328,
+ "learning_rate": 1.1999999999999999e-05,
+ "loss": 11.945581436157227,
+ "step": 3
+ },
+ {
+ "epoch": 0.05589519650655022,
+ "grad_norm": 0.17577414214611053,
+ "learning_rate": 1.7999999999999997e-05,
+ "loss": 11.859772682189941,
+ "step": 4
+ },
+ {
+ "epoch": 0.06986899563318777,
+ "grad_norm": 0.16163651645183563,
+ "learning_rate": 2.3999999999999997e-05,
+ "loss": 11.744415283203125,
+ "step": 5
+ },
+ {
+ "epoch": 0.08384279475982533,
+ "grad_norm": 0.14409932494163513,
+ "learning_rate": 2.9999999999999997e-05,
+ "loss": 11.618099212646484,
+ "step": 6
+ },
+ {
+ "epoch": 0.09781659388646288,
+ "grad_norm": 0.13079848885536194,
+ "learning_rate": 3.5999999999999994e-05,
+ "loss": 11.505935668945312,
+ "step": 7
+ },
+ {
+ "epoch": 0.11179039301310044,
+ "grad_norm": 0.11964613944292068,
+ "learning_rate": 4.2e-05,
+ "loss": 11.416532516479492,
+ "step": 8
+ },
+ {
+ "epoch": 0.125764192139738,
+ "grad_norm": 0.11345142871141434,
+ "learning_rate": 4.7999999999999994e-05,
+ "loss": 11.33206558227539,
+ "step": 9
+ },
+ {
+ "epoch": 0.13973799126637554,
+ "grad_norm": 0.11033952981233597,
+ "learning_rate": 5.399999999999999e-05,
+ "loss": 11.259052276611328,
+ "step": 10
+ },
+ {
+ "epoch": 0.1537117903930131,
+ "grad_norm": 0.10776489973068237,
+ "learning_rate": 5.9999999999999995e-05,
+ "loss": 11.203384399414062,
+ "step": 11
+ },
+ {
+ "epoch": 0.16768558951965065,
+ "grad_norm": 0.10721946507692337,
+ "learning_rate": 6.599999999999999e-05,
+ "loss": 11.153295516967773,
+ "step": 12
+ },
+ {
+ "epoch": 0.18165938864628822,
+ "grad_norm": 0.10673430562019348,
+ "learning_rate": 7.199999999999999e-05,
+ "loss": 11.110200881958008,
+ "step": 13
+ },
+ {
+ "epoch": 0.19563318777292577,
+ "grad_norm": 0.10630254447460175,
+ "learning_rate": 7.8e-05,
+ "loss": 11.067477226257324,
+ "step": 14
+ },
+ {
+ "epoch": 0.2096069868995633,
+ "grad_norm": 0.10631615668535233,
+ "learning_rate": 8.4e-05,
+ "loss": 11.022941589355469,
+ "step": 15
+ },
+ {
+ "epoch": 0.22358078602620088,
+ "grad_norm": 0.10609443485736847,
+ "learning_rate": 8.999999999999999e-05,
+ "loss": 10.970746040344238,
+ "step": 16
+ },
+ {
+ "epoch": 0.23755458515283842,
+ "grad_norm": 0.10648232698440552,
+ "learning_rate": 9.599999999999999e-05,
+ "loss": 10.91267204284668,
+ "step": 17
+ },
+ {
+ "epoch": 0.251528384279476,
+ "grad_norm": 0.10584589838981628,
+ "learning_rate": 0.000102,
+ "loss": 10.854547500610352,
+ "step": 18
+ },
+ {
+ "epoch": 0.26550218340611353,
+ "grad_norm": 0.10591720044612885,
+ "learning_rate": 0.00010799999999999998,
+ "loss": 10.788917541503906,
+ "step": 19
+ },
+ {
+ "epoch": 0.2794759825327511,
+ "grad_norm": 0.10611672699451447,
+ "learning_rate": 0.00011399999999999999,
+ "loss": 10.717556953430176,
+ "step": 20
+ },
+ {
+ "epoch": 0.2934497816593886,
+ "grad_norm": 0.10455359518527985,
+ "learning_rate": 0.00011999999999999999,
+ "loss": 10.650186538696289,
+ "step": 21
+ },
+ {
+ "epoch": 0.3074235807860262,
+ "grad_norm": 0.1047525629401207,
+ "learning_rate": 0.00012599999999999997,
+ "loss": 10.56808090209961,
+ "step": 22
+ },
+ {
+ "epoch": 0.32139737991266376,
+ "grad_norm": 0.10511605441570282,
+ "learning_rate": 0.00013199999999999998,
+ "loss": 10.484028816223145,
+ "step": 23
+ },
+ {
+ "epoch": 0.3353711790393013,
+ "grad_norm": 0.10467177629470825,
+ "learning_rate": 0.000138,
+ "loss": 10.399947166442871,
+ "step": 24
+ },
+ {
+ "epoch": 0.34934497816593885,
+ "grad_norm": 0.10432136058807373,
+ "learning_rate": 0.00014399999999999998,
+ "loss": 10.312714576721191,
+ "step": 25
+ },
+ {
+ "epoch": 0.36331877729257644,
+ "grad_norm": 0.10527262091636658,
+ "learning_rate": 0.00015,
+ "loss": 10.212261199951172,
+ "step": 26
+ },
+ {
+ "epoch": 0.377292576419214,
+ "grad_norm": 0.103610560297966,
+ "learning_rate": 0.000156,
+ "loss": 10.132034301757812,
+ "step": 27
+ },
+ {
+ "epoch": 0.39126637554585153,
+ "grad_norm": 0.10553059726953506,
+ "learning_rate": 0.000162,
+ "loss": 10.020434379577637,
+ "step": 28
+ },
+ {
+ "epoch": 0.4052401746724891,
+ "grad_norm": 0.1037299782037735,
+ "learning_rate": 0.000168,
+ "loss": 9.934466361999512,
+ "step": 29
+ },
+ {
+ "epoch": 0.4192139737991266,
+ "grad_norm": 0.1042739674448967,
+ "learning_rate": 0.00017399999999999997,
+ "loss": 9.822539329528809,
+ "step": 30
+ },
+ {
+ "epoch": 0.4331877729257642,
+ "grad_norm": 0.10300865024328232,
+ "learning_rate": 0.00017999999999999998,
+ "loss": 9.734994888305664,
+ "step": 31
+ },
+ {
+ "epoch": 0.44716157205240176,
+ "grad_norm": 0.10291654616594315,
+ "learning_rate": 0.000186,
+ "loss": 9.624505996704102,
+ "step": 32
+ },
+ {
+ "epoch": 0.4611353711790393,
+ "grad_norm": 0.10210851579904556,
+ "learning_rate": 0.00019199999999999998,
+ "loss": 9.518205642700195,
+ "step": 33
+ },
+ {
+ "epoch": 0.47510917030567684,
+ "grad_norm": 0.10222529619932175,
+ "learning_rate": 0.000198,
+ "loss": 9.412222862243652,
+ "step": 34
+ },
+ {
+ "epoch": 0.4890829694323144,
+ "grad_norm": 0.09992019087076187,
+ "learning_rate": 0.000204,
+ "loss": 9.326566696166992,
+ "step": 35
+ },
+ {
+ "epoch": 0.503056768558952,
+ "grad_norm": 0.09931084513664246,
+ "learning_rate": 0.00020999999999999998,
+ "loss": 9.210111618041992,
+ "step": 36
+ },
+ {
+ "epoch": 0.5170305676855895,
+ "grad_norm": 0.09928256273269653,
+ "learning_rate": 0.00021599999999999996,
+ "loss": 9.100493431091309,
+ "step": 37
+ },
+ {
+ "epoch": 0.5310043668122271,
+ "grad_norm": 0.09843816608190536,
+ "learning_rate": 0.00022199999999999998,
+ "loss": 8.996543884277344,
+ "step": 38
+ },
+ {
+ "epoch": 0.5449781659388646,
+ "grad_norm": 0.09825374186038971,
+ "learning_rate": 0.00022799999999999999,
+ "loss": 8.888991355895996,
+ "step": 39
+ },
+ {
+ "epoch": 0.5589519650655022,
+ "grad_norm": 0.09602385759353638,
+ "learning_rate": 0.000234,
+ "loss": 8.802907943725586,
+ "step": 40
+ },
+ {
+ "epoch": 0.5729257641921397,
+ "grad_norm": 0.09414064139127731,
+ "learning_rate": 0.00023999999999999998,
+ "loss": 8.715423583984375,
+ "step": 41
+ },
+ {
+ "epoch": 0.5868995633187772,
+ "grad_norm": 0.09273893386125565,
+ "learning_rate": 0.00024599999999999996,
+ "loss": 8.61374282836914,
+ "step": 42
+ },
+ {
+ "epoch": 0.6008733624454149,
+ "grad_norm": 0.09079599380493164,
+ "learning_rate": 0.00025199999999999995,
+ "loss": 8.502904891967773,
+ "step": 43
+ },
+ {
+ "epoch": 0.6148471615720524,
+ "grad_norm": 0.08605825901031494,
+ "learning_rate": 0.000258,
+ "loss": 8.453008651733398,
+ "step": 44
+ },
+ {
+ "epoch": 0.62882096069869,
+ "grad_norm": 0.08338996022939682,
+ "learning_rate": 0.00026399999999999997,
+ "loss": 8.365163803100586,
+ "step": 45
+ },
+ {
+ "epoch": 0.6427947598253275,
+ "grad_norm": 0.08150867372751236,
+ "learning_rate": 0.00027,
+ "loss": 8.285372734069824,
+ "step": 46
+ },
+ {
+ "epoch": 0.6567685589519651,
+ "grad_norm": 0.07794807851314545,
+ "learning_rate": 0.000276,
+ "loss": 8.185811996459961,
+ "step": 47
+ },
+ {
+ "epoch": 0.6707423580786026,
+ "grad_norm": 0.07592720538377762,
+ "learning_rate": 0.00028199999999999997,
+ "loss": 8.098482131958008,
+ "step": 48
+ },
+ {
+ "epoch": 0.6847161572052402,
+ "grad_norm": 0.0709897130727768,
+ "learning_rate": 0.00028799999999999995,
+ "loss": 8.042091369628906,
+ "step": 49
+ },
+ {
+ "epoch": 0.6986899563318777,
+ "grad_norm": 0.06582305580377579,
+ "learning_rate": 0.000294,
+ "loss": 7.972844123840332,
+ "step": 50
+ },
+ {
+ "epoch": 0.7126637554585152,
+ "grad_norm": 0.059452351182699203,
+ "learning_rate": 0.0003,
+ "loss": 7.924579620361328,
+ "step": 51
+ },
+ {
+ "epoch": 0.7266375545851529,
+ "grad_norm": 0.052426718175411224,
+ "learning_rate": 0.00030599999999999996,
+ "loss": 7.876336097717285,
+ "step": 52
+ },
+ {
+ "epoch": 0.7406113537117904,
+ "grad_norm": 0.04768150672316551,
+ "learning_rate": 0.000312,
+ "loss": 7.8102707862854,
+ "step": 53
+ },
+ {
+ "epoch": 0.754585152838428,
+ "grad_norm": 0.03970075398683548,
+ "learning_rate": 0.000318,
+ "loss": 7.791913986206055,
+ "step": 54
+ },
+ {
+ "epoch": 0.7685589519650655,
+ "grad_norm": 0.03559752553701401,
+ "learning_rate": 0.000324,
+ "loss": 7.756960868835449,
+ "step": 55
+ },
+ {
+ "epoch": 0.7825327510917031,
+ "grad_norm": 0.029321353882551193,
+ "learning_rate": 0.00033,
+ "loss": 7.724878787994385,
+ "step": 56
+ },
+ {
+ "epoch": 0.7965065502183406,
+ "grad_norm": 0.023485427722334862,
+ "learning_rate": 0.000336,
+ "loss": 7.72553014755249,
+ "step": 57
+ },
+ {
+ "epoch": 0.8104803493449781,
+ "grad_norm": 0.019319474697113037,
+ "learning_rate": 0.00034199999999999996,
+ "loss": 7.709493160247803,
+ "step": 58
+ },
+ {
+ "epoch": 0.8244541484716157,
+ "grad_norm": 0.016705691814422607,
+ "learning_rate": 0.00034799999999999995,
+ "loss": 7.6873297691345215,
+ "step": 59
+ },
+ {
+ "epoch": 0.8384279475982532,
+ "grad_norm": 0.017667734995484352,
+ "learning_rate": 0.00035399999999999993,
+ "loss": 7.677059173583984,
+ "step": 60
+ },
+ {
+ "epoch": 0.8524017467248908,
+ "grad_norm": 0.01671992614865303,
+ "learning_rate": 0.00035999999999999997,
+ "loss": 7.643512725830078,
+ "step": 61
+ },
+ {
+ "epoch": 0.8663755458515284,
+ "grad_norm": 0.012288172729313374,
+ "learning_rate": 0.00036599999999999995,
+ "loss": 7.658343315124512,
+ "step": 62
+ },
+ {
+ "epoch": 0.880349344978166,
+ "grad_norm": 0.012004776857793331,
+ "learning_rate": 0.000372,
+ "loss": 7.649759292602539,
+ "step": 63
+ },
+ {
+ "epoch": 0.8943231441048035,
+ "grad_norm": 0.012487445957958698,
+ "learning_rate": 0.00037799999999999997,
+ "loss": 7.631503105163574,
+ "step": 64
+ },
+ {
+ "epoch": 0.9082969432314411,
+ "grad_norm": 0.01489241048693657,
+ "learning_rate": 0.00038399999999999996,
+ "loss": 7.647110462188721,
+ "step": 65
+ },
+ {
+ "epoch": 0.9222707423580786,
+ "grad_norm": 0.01428980939090252,
+ "learning_rate": 0.00039,
+ "loss": 7.636983871459961,
+ "step": 66
+ },
+ {
+ "epoch": 0.9362445414847161,
+ "grad_norm": 0.010711144655942917,
+ "learning_rate": 0.000396,
+ "loss": 7.621130466461182,
+ "step": 67
+ },
+ {
+ "epoch": 0.9502183406113537,
+ "grad_norm": 0.010178886353969574,
+ "learning_rate": 0.000402,
+ "loss": 7.593924522399902,
+ "step": 68
+ },
+ {
+ "epoch": 0.9641921397379912,
+ "grad_norm": 0.010604874230921268,
+ "learning_rate": 0.000408,
+ "loss": 7.6015496253967285,
+ "step": 69
+ },
+ {
+ "epoch": 0.9781659388646288,
+ "grad_norm": 0.009751426056027412,
+ "learning_rate": 0.0004139999999999999,
+ "loss": 7.571964263916016,
+ "step": 70
+ },
+ {
+ "epoch": 0.9921397379912664,
+ "grad_norm": 0.009816481731832027,
+ "learning_rate": 0.00041999999999999996,
+ "loss": 7.57344913482666,
+ "step": 71
+ },
+ {
+ "epoch": 1.0,
+ "grad_norm": 0.008665637113153934,
+ "learning_rate": 0.00042599999999999995,
+ "loss": 7.579050064086914,
+ "step": 72
+ },
+ {
+ "epoch": 1.0,
+ "eval_loss": 7.581204891204834,
+ "eval_runtime": 59.0091,
+ "eval_samples_per_second": 41.383,
+ "eval_steps_per_second": 1.305,
+ "step": 72
+ },
+ {
+ "epoch": 1.0139737991266375,
+ "grad_norm": 0.009684201329946518,
+ "learning_rate": 0.00043199999999999993,
+ "loss": 7.568594932556152,
+ "step": 73
+ },
+ {
+ "epoch": 1.027947598253275,
+ "grad_norm": 0.025783710181713104,
+ "learning_rate": 0.00043799999999999997,
+ "loss": 7.578948497772217,
+ "step": 74
+ },
+ {
+ "epoch": 1.0419213973799126,
+ "grad_norm": 0.0327819399535656,
+ "learning_rate": 0.00044399999999999995,
+ "loss": 7.548173904418945,
+ "step": 75
+ },
+ {
+ "epoch": 1.0558951965065502,
+ "grad_norm": 0.021604498848319054,
+ "learning_rate": 0.00045,
+ "loss": 7.552953720092773,
+ "step": 76
+ },
+ {
+ "epoch": 1.0698689956331877,
+ "grad_norm": 0.010230105370283127,
+ "learning_rate": 0.00045599999999999997,
+ "loss": 7.5213212966918945,
+ "step": 77
+ },
+ {
+ "epoch": 1.0838427947598253,
+ "grad_norm": 0.023315589874982834,
+ "learning_rate": 0.00046199999999999995,
+ "loss": 7.501640319824219,
+ "step": 78
+ },
+ {
+ "epoch": 1.0978165938864628,
+ "grad_norm": 0.01109944935888052,
+ "learning_rate": 0.000468,
+ "loss": 7.539519786834717,
+ "step": 79
+ },
+ {
+ "epoch": 1.1117903930131003,
+ "grad_norm": 0.026093758642673492,
+ "learning_rate": 0.000474,
+ "loss": 7.525928974151611,
+ "step": 80
+ },
+ {
+ "epoch": 1.125764192139738,
+ "grad_norm": 0.03274895250797272,
+ "learning_rate": 0.00047999999999999996,
+ "loss": 7.535602569580078,
+ "step": 81
+ },
+ {
+ "epoch": 1.1397379912663754,
+ "grad_norm": 0.01543602254241705,
+ "learning_rate": 0.000486,
+ "loss": 7.513683795928955,
+ "step": 82
+ },
+ {
+ "epoch": 1.1537117903930132,
+ "grad_norm": 0.028377249836921692,
+ "learning_rate": 0.0004919999999999999,
+ "loss": 7.49384880065918,
+ "step": 83
+ },
+ {
+ "epoch": 1.1676855895196507,
+ "grad_norm": 0.04121481627225876,
+ "learning_rate": 0.000498,
+ "loss": 7.488447189331055,
+ "step": 84
+ },
+ {
+ "epoch": 1.1816593886462883,
+ "grad_norm": 0.013282302767038345,
+ "learning_rate": 0.0005039999999999999,
+ "loss": 7.463010787963867,
+ "step": 85
+ },
+ {
+ "epoch": 1.1956331877729258,
+ "grad_norm": 0.03200364485383034,
+ "learning_rate": 0.0005099999999999999,
+ "loss": 7.473515033721924,
+ "step": 86
+ },
+ {
+ "epoch": 1.2096069868995634,
+ "grad_norm": 0.020768029615283012,
+ "learning_rate": 0.000516,
+ "loss": 7.464688777923584,
+ "step": 87
+ },
+ {
+ "epoch": 1.223580786026201,
+ "grad_norm": 0.021715788170695305,
+ "learning_rate": 0.000522,
+ "loss": 7.4835028648376465,
+ "step": 88
+ },
+ {
+ "epoch": 1.2375545851528384,
+ "grad_norm": 0.026255754753947258,
+ "learning_rate": 0.0005279999999999999,
+ "loss": 7.443392753601074,
+ "step": 89
+ },
+ {
+ "epoch": 1.251528384279476,
+ "grad_norm": 0.022325951606035233,
+ "learning_rate": 0.000534,
+ "loss": 7.457373142242432,
+ "step": 90
+ },
+ {
+ "epoch": 1.2655021834061135,
+ "grad_norm": 0.009270387701690197,
+ "learning_rate": 0.00054,
+ "loss": 7.455986976623535,
+ "step": 91
+ },
+ {
+ "epoch": 1.279475982532751,
+ "grad_norm": 0.02369544468820095,
+ "learning_rate": 0.0005459999999999999,
+ "loss": 7.442216873168945,
+ "step": 92
+ },
+ {
+ "epoch": 1.2934497816593886,
+ "grad_norm": 0.008474506437778473,
+ "learning_rate": 0.000552,
+ "loss": 7.4294915199279785,
+ "step": 93
+ },
+ {
+ "epoch": 1.3074235807860262,
+ "grad_norm": 0.017043249681591988,
+ "learning_rate": 0.000558,
+ "loss": 7.406932830810547,
+ "step": 94
+ },
+ {
+ "epoch": 1.3213973799126637,
+ "grad_norm": 0.011813412420451641,
+ "learning_rate": 0.0005639999999999999,
+ "loss": 7.415900230407715,
+ "step": 95
+ },
+ {
+ "epoch": 1.3353711790393012,
+ "grad_norm": 0.01086883433163166,
+ "learning_rate": 0.00057,
+ "loss": 7.452998638153076,
+ "step": 96
+ },
+ {
+ "epoch": 1.3493449781659388,
+ "grad_norm": 0.014700951054692268,
+ "learning_rate": 0.0005759999999999999,
+ "loss": 7.410249710083008,
+ "step": 97
+ },
+ {
+ "epoch": 1.3633187772925766,
+ "grad_norm": 0.014512976631522179,
+ "learning_rate": 0.0005819999999999999,
+ "loss": 7.429132461547852,
+ "step": 98
+ },
+ {
+ "epoch": 1.3772925764192139,
+ "grad_norm": 0.012519920244812965,
+ "learning_rate": 0.000588,
+ "loss": 7.430817604064941,
+ "step": 99
+ },
+ {
+ "epoch": 1.3912663755458516,
+ "grad_norm": 0.007132664322853088,
+ "learning_rate": 0.0005939999999999999,
+ "loss": 7.405234336853027,
+ "step": 100
+ },
+ {
+ "epoch": 1.405240174672489,
+ "grad_norm": 0.019489388912916183,
+ "learning_rate": 0.0006,
+ "loss": 7.427791595458984,
+ "step": 101
+ },
+ {
+ "epoch": 1.4192139737991267,
+ "grad_norm": 0.027252664789557457,
+ "learning_rate": 0.0006,
+ "loss": 7.41761589050293,
+ "step": 102
+ },
+ {
+ "epoch": 1.4331877729257643,
+ "grad_norm": 0.02501637488603592,
+ "learning_rate": 0.0006,
+ "loss": 7.416735649108887,
+ "step": 103
+ },
+ {
+ "epoch": 1.4471615720524018,
+ "grad_norm": 0.00989772193133831,
+ "learning_rate": 0.0006,
+ "loss": 7.3756632804870605,
+ "step": 104
+ },
+ {
+ "epoch": 1.4611353711790394,
+ "grad_norm": 0.01622949168086052,
+ "learning_rate": 0.0006,
+ "loss": 7.400811672210693,
+ "step": 105
+ },
+ {
+ "epoch": 1.475109170305677,
+ "grad_norm": 0.012407763861119747,
+ "learning_rate": 0.0006,
+ "loss": 7.37675666809082,
+ "step": 106
+ },
+ {
+ "epoch": 1.4890829694323144,
+ "grad_norm": 0.010353565216064453,
+ "learning_rate": 0.0006,
+ "loss": 7.4013142585754395,
+ "step": 107
+ },
+ {
+ "epoch": 1.503056768558952,
+ "grad_norm": 0.0120553532615304,
+ "learning_rate": 0.0006,
+ "loss": 7.379873275756836,
+ "step": 108
+ },
+ {
+ "epoch": 1.5170305676855895,
+ "grad_norm": 0.01610225811600685,
+ "learning_rate": 0.0006,
+ "loss": 7.371631145477295,
+ "step": 109
+ },
+ {
+ "epoch": 1.531004366812227,
+ "grad_norm": 0.01159879844635725,
+ "learning_rate": 0.0006,
+ "loss": 7.389007568359375,
+ "step": 110
+ },
+ {
+ "epoch": 1.5449781659388646,
+ "grad_norm": 0.008013816550374031,
+ "learning_rate": 0.0006,
+ "loss": 7.395920753479004,
+ "step": 111
+ },
+ {
+ "epoch": 1.5589519650655022,
+ "grad_norm": 0.016334809362888336,
+ "learning_rate": 0.0006,
+ "loss": 7.393121719360352,
+ "step": 112
+ },
+ {
+ "epoch": 1.5729257641921397,
+ "grad_norm": 0.018821721896529198,
+ "learning_rate": 0.0006,
+ "loss": 7.367543697357178,
+ "step": 113
+ },
+ {
+ "epoch": 1.5868995633187772,
+ "grad_norm": 0.014088994823396206,
+ "learning_rate": 0.0006,
+ "loss": 7.385537624359131,
+ "step": 114
+ },
+ {
+ "epoch": 1.600873362445415,
+ "grad_norm": 0.005864050704985857,
+ "learning_rate": 0.0006,
+ "loss": 7.365346908569336,
+ "step": 115
+ },
+ {
+ "epoch": 1.6148471615720523,
+ "grad_norm": 0.011745430529117584,
+ "learning_rate": 0.0006,
+ "loss": 7.390227317810059,
+ "step": 116
+ },
+ {
+ "epoch": 1.62882096069869,
+ "grad_norm": 0.017205573618412018,
+ "learning_rate": 0.0006,
+ "loss": 7.3558549880981445,
+ "step": 117
+ },
+ {
+ "epoch": 1.6427947598253274,
+ "grad_norm": 0.021256346255540848,
+ "learning_rate": 0.0006,
+ "loss": 7.377346992492676,
+ "step": 118
+ },
+ {
+ "epoch": 1.6567685589519652,
+ "grad_norm": 0.019335204735398293,
+ "learning_rate": 0.0006,
+ "loss": 7.346194267272949,
+ "step": 119
+ },
+ {
+ "epoch": 1.6707423580786025,
+ "grad_norm": 0.008134003728628159,
+ "learning_rate": 0.0006,
+ "loss": 7.3063578605651855,
+ "step": 120
+ },
+ {
+ "epoch": 1.6847161572052403,
+ "grad_norm": 0.017613831907510757,
+ "learning_rate": 0.0006,
+ "loss": 7.338171005249023,
+ "step": 121
+ },
+ {
+ "epoch": 1.6986899563318776,
+ "grad_norm": 0.00890734139829874,
+ "learning_rate": 0.0006,
+ "loss": 7.355384826660156,
+ "step": 122
+ },
+ {
+ "epoch": 1.7126637554585153,
+ "grad_norm": 0.013448155485093594,
+ "learning_rate": 0.0006,
+ "loss": 7.344296455383301,
+ "step": 123
+ },
+ {
+ "epoch": 1.726637554585153,
+ "grad_norm": 0.009241633117198944,
+ "learning_rate": 0.0006,
+ "loss": 7.3454060554504395,
+ "step": 124
+ },
+ {
+ "epoch": 1.7406113537117904,
+ "grad_norm": 0.018561311066150665,
+ "learning_rate": 0.0006,
+ "loss": 7.32348108291626,
+ "step": 125
+ },
+ {
+ "epoch": 1.754585152838428,
+ "grad_norm": 0.025667704641819,
+ "learning_rate": 0.0006,
+ "loss": 7.356142520904541,
+ "step": 126
+ },
+ {
+ "epoch": 1.7685589519650655,
+ "grad_norm": 0.036232151091098785,
+ "learning_rate": 0.0006,
+ "loss": 7.363539695739746,
+ "step": 127
+ },
+ {
+ "epoch": 1.782532751091703,
+ "grad_norm": 0.059483714401721954,
+ "learning_rate": 0.0006,
+ "loss": 7.402331352233887,
+ "step": 128
+ },
+ {
+ "epoch": 1.7965065502183406,
+ "grad_norm": 0.05365009233355522,
+ "learning_rate": 0.0006,
+ "loss": 7.40457820892334,
+ "step": 129
+ },
+ {
+ "epoch": 1.8104803493449781,
+ "grad_norm": 0.02738415263593197,
+ "learning_rate": 0.0006,
+ "loss": 7.3756208419799805,
+ "step": 130
+ },
+ {
+ "epoch": 1.8244541484716157,
+ "grad_norm": 0.035017091780900955,
+ "learning_rate": 0.0006,
+ "loss": 7.379263877868652,
+ "step": 131
+ },
+ {
+ "epoch": 1.8384279475982532,
+ "grad_norm": 0.02819863148033619,
+ "learning_rate": 0.0006,
+ "loss": 7.327673435211182,
+ "step": 132
+ },
+ {
+ "epoch": 1.8524017467248908,
+ "grad_norm": 0.017754238098859787,
+ "learning_rate": 0.0006,
+ "loss": 7.333420276641846,
+ "step": 133
+ },
+ {
+ "epoch": 1.8663755458515285,
+ "grad_norm": 0.024685733020305634,
+ "learning_rate": 0.0006,
+ "loss": 7.34493350982666,
+ "step": 134
+ },
+ {
+ "epoch": 1.8803493449781659,
+ "grad_norm": 0.019426532089710236,
+ "learning_rate": 0.0006,
+ "loss": 7.328448295593262,
+ "step": 135
+ },
+ {
+ "epoch": 1.8943231441048036,
+ "grad_norm": 0.02106391079723835,
+ "learning_rate": 0.0006,
+ "loss": 7.327917098999023,
+ "step": 136
+ },
+ {
+ "epoch": 1.908296943231441,
+ "grad_norm": 0.02056879736483097,
+ "learning_rate": 0.0006,
+ "loss": 7.3311991691589355,
+ "step": 137
+ },
+ {
+ "epoch": 1.9222707423580787,
+ "grad_norm": 0.016865141689777374,
+ "learning_rate": 0.0006,
+ "loss": 7.312766075134277,
+ "step": 138
+ },
+ {
+ "epoch": 1.936244541484716,
+ "grad_norm": 0.017549855634570122,
+ "learning_rate": 0.0006,
+ "loss": 7.31191349029541,
+ "step": 139
+ },
+ {
+ "epoch": 1.9502183406113538,
+ "grad_norm": 0.015930157154798508,
+ "learning_rate": 0.0006,
+ "loss": 7.330901622772217,
+ "step": 140
+ },
+ {
+ "epoch": 1.9641921397379911,
+ "grad_norm": 0.013283558189868927,
+ "learning_rate": 0.0006,
+ "loss": 7.298318386077881,
+ "step": 141
+ },
+ {
+ "epoch": 1.9781659388646289,
+ "grad_norm": 0.012023469433188438,
+ "learning_rate": 0.0006,
+ "loss": 7.282337188720703,
+ "step": 142
+ },
+ {
+ "epoch": 1.9921397379912664,
+ "grad_norm": 0.01459511648863554,
+ "learning_rate": 0.0006,
+ "loss": 7.282924175262451,
+ "step": 143
+ },
+ {
+ "epoch": 2.0,
+ "grad_norm": 0.017474235966801643,
+ "learning_rate": 0.0006,
+ "loss": 7.254333972930908,
+ "step": 144
+ },
+ {
+ "epoch": 2.0,
+ "eval_loss": 7.318708896636963,
+ "eval_runtime": 59.6555,
+ "eval_samples_per_second": 40.935,
+ "eval_steps_per_second": 1.291,
+ "step": 144
+ },
+ {
+ "epoch": 2.0139737991266378,
+ "grad_norm": 0.019757507368922234,
+ "learning_rate": 0.0006,
+ "loss": 7.30288553237915,
+ "step": 145
+ },
+ {
+ "epoch": 2.027947598253275,
+ "grad_norm": 0.014571248553693295,
+ "learning_rate": 0.0006,
+ "loss": 7.274681568145752,
+ "step": 146
+ },
+ {
+ "epoch": 2.041921397379913,
+ "grad_norm": 0.015744198113679886,
+ "learning_rate": 0.0006,
+ "loss": 7.278632640838623,
+ "step": 147
+ },
+ {
+ "epoch": 2.05589519650655,
+ "grad_norm": 0.028048858046531677,
+ "learning_rate": 0.0006,
+ "loss": 7.26558780670166,
+ "step": 148
+ },
+ {
+ "epoch": 2.069868995633188,
+ "grad_norm": 0.06547307223081589,
+ "learning_rate": 0.0006,
+ "loss": 7.32567834854126,
+ "step": 149
+ },
+ {
+ "epoch": 2.0838427947598253,
+ "grad_norm": 0.05969763547182083,
+ "learning_rate": 0.0006,
+ "loss": 7.395836353302002,
+ "step": 150
+ },
+ {
+ "epoch": 2.097816593886463,
+ "grad_norm": 0.032574281096458435,
+ "learning_rate": 0.0006,
+ "loss": 7.318150997161865,
+ "step": 151
+ },
+ {
+ "epoch": 2.1117903930131003,
+ "grad_norm": 0.029452962800860405,
+ "learning_rate": 0.0006,
+ "loss": 7.308066368103027,
+ "step": 152
+ },
+ {
+ "epoch": 2.125764192139738,
+ "grad_norm": 0.03161991387605667,
+ "learning_rate": 0.0006,
+ "loss": 7.306290626525879,
+ "step": 153
+ },
+ {
+ "epoch": 2.1397379912663754,
+ "grad_norm": 0.027096295729279518,
+ "learning_rate": 0.0006,
+ "loss": 7.297796249389648,
+ "step": 154
+ },
+ {
+ "epoch": 2.153711790393013,
+ "grad_norm": 0.028657056391239166,
+ "learning_rate": 0.0006,
+ "loss": 7.317424774169922,
+ "step": 155
+ },
+ {
+ "epoch": 2.1676855895196505,
+ "grad_norm": 0.02046995982527733,
+ "learning_rate": 0.0006,
+ "loss": 7.297986030578613,
+ "step": 156
+ },
+ {
+ "epoch": 2.1816593886462883,
+ "grad_norm": 0.02220369316637516,
+ "learning_rate": 0.0006,
+ "loss": 7.268237590789795,
+ "step": 157
+ },
+ {
+ "epoch": 2.1956331877729256,
+ "grad_norm": 0.018381357192993164,
+ "learning_rate": 0.0006,
+ "loss": 7.259415626525879,
+ "step": 158
+ },
+ {
+ "epoch": 2.2096069868995634,
+ "grad_norm": 0.0204413291066885,
+ "learning_rate": 0.0006,
+ "loss": 7.258173942565918,
+ "step": 159
+ },
+ {
+ "epoch": 2.2235807860262007,
+ "grad_norm": 0.015874288976192474,
+ "learning_rate": 0.0006,
+ "loss": 7.281834602355957,
+ "step": 160
+ },
+ {
+ "epoch": 2.2375545851528384,
+ "grad_norm": 0.01634068600833416,
+ "learning_rate": 0.0006,
+ "loss": 7.2469868659973145,
+ "step": 161
+ },
+ {
+ "epoch": 2.251528384279476,
+ "grad_norm": 0.01728747971355915,
+ "learning_rate": 0.0006,
+ "loss": 7.280995845794678,
+ "step": 162
+ },
+ {
+ "epoch": 2.2655021834061135,
+ "grad_norm": 0.016341188922524452,
+ "learning_rate": 0.0006,
+ "loss": 7.273357391357422,
+ "step": 163
+ },
+ {
+ "epoch": 2.279475982532751,
+ "grad_norm": 0.016125807538628578,
+ "learning_rate": 0.0006,
+ "loss": 7.273510932922363,
+ "step": 164
+ },
+ {
+ "epoch": 2.2934497816593886,
+ "grad_norm": 0.01629287749528885,
+ "learning_rate": 0.0006,
+ "loss": 7.267029762268066,
+ "step": 165
+ },
+ {
+ "epoch": 2.3074235807860264,
+ "grad_norm": 0.015296131372451782,
+ "learning_rate": 0.0006,
+ "loss": 7.211989402770996,
+ "step": 166
+ },
+ {
+ "epoch": 2.3213973799126637,
+ "grad_norm": 0.01709570363163948,
+ "learning_rate": 0.0006,
+ "loss": 7.242851734161377,
+ "step": 167
+ },
+ {
+ "epoch": 2.3353711790393015,
+ "grad_norm": 0.012552388943731785,
+ "learning_rate": 0.0006,
+ "loss": 7.241696834564209,
+ "step": 168
+ },
+ {
+ "epoch": 2.349344978165939,
+ "grad_norm": 0.014571291394531727,
+ "learning_rate": 0.0006,
+ "loss": 7.270681381225586,
+ "step": 169
+ },
+ {
+ "epoch": 2.3633187772925766,
+ "grad_norm": 0.013389437459409237,
+ "learning_rate": 0.0006,
+ "loss": 7.191600799560547,
+ "step": 170
+ },
+ {
+ "epoch": 2.377292576419214,
+ "grad_norm": 0.011895314790308475,
+ "learning_rate": 0.0006,
+ "loss": 7.213610649108887,
+ "step": 171
+ },
+ {
+ "epoch": 2.3912663755458516,
+ "grad_norm": 0.01345642190426588,
+ "learning_rate": 0.0006,
+ "loss": 7.193787574768066,
+ "step": 172
+ },
+ {
+ "epoch": 2.405240174672489,
+ "grad_norm": 0.0091794328764081,
+ "learning_rate": 0.0006,
+ "loss": 7.233364105224609,
+ "step": 173
+ },
+ {
+ "epoch": 2.4192139737991267,
+ "grad_norm": 0.013767086900770664,
+ "learning_rate": 0.0006,
+ "loss": 7.217272758483887,
+ "step": 174
+ },
+ {
+ "epoch": 2.433187772925764,
+ "grad_norm": 0.015134960412979126,
+ "learning_rate": 0.0006,
+ "loss": 7.192791938781738,
+ "step": 175
+ },
+ {
+ "epoch": 2.447161572052402,
+ "grad_norm": 0.01790648326277733,
+ "learning_rate": 0.0006,
+ "loss": 7.229467391967773,
+ "step": 176
+ },
+ {
+ "epoch": 2.461135371179039,
+ "grad_norm": 0.015362377278506756,
+ "learning_rate": 0.0006,
+ "loss": 7.180346488952637,
+ "step": 177
+ },
+ {
+ "epoch": 2.475109170305677,
+ "grad_norm": 0.010505498386919498,
+ "learning_rate": 0.0006,
+ "loss": 7.188833236694336,
+ "step": 178
+ },
+ {
+ "epoch": 2.489082969432314,
+ "grad_norm": 0.009129747748374939,
+ "learning_rate": 0.0006,
+ "loss": 7.203153133392334,
+ "step": 179
+ },
+ {
+ "epoch": 2.503056768558952,
+ "grad_norm": 0.01147883478552103,
+ "learning_rate": 0.0006,
+ "loss": 7.163824558258057,
+ "step": 180
+ },
+ {
+ "epoch": 2.5170305676855893,
+ "grad_norm": 0.011338942684233189,
+ "learning_rate": 0.0006,
+ "loss": 7.16287899017334,
+ "step": 181
+ },
+ {
+ "epoch": 2.531004366812227,
+ "grad_norm": 0.013277675025165081,
+ "learning_rate": 0.0006,
+ "loss": 7.155905723571777,
+ "step": 182
+ },
+ {
+ "epoch": 2.544978165938865,
+ "grad_norm": 0.015269347466528416,
+ "learning_rate": 0.0006,
+ "loss": 7.189371109008789,
+ "step": 183
+ },
+ {
+ "epoch": 2.558951965065502,
+ "grad_norm": 0.017905596643686295,
+ "learning_rate": 0.0006,
+ "loss": 7.177547931671143,
+ "step": 184
+ },
+ {
+ "epoch": 2.5729257641921395,
+ "grad_norm": 0.024431610479950905,
+ "learning_rate": 0.0006,
+ "loss": 7.192169189453125,
+ "step": 185
+ },
+ {
+ "epoch": 2.5868995633187772,
+ "grad_norm": 0.03656961768865585,
+ "learning_rate": 0.0006,
+ "loss": 7.124387741088867,
+ "step": 186
+ },
+ {
+ "epoch": 2.600873362445415,
+ "grad_norm": 0.04337048903107643,
+ "learning_rate": 0.0006,
+ "loss": 7.140231609344482,
+ "step": 187
+ },
+ {
+ "epoch": 2.6148471615720523,
+ "grad_norm": 0.04047228768467903,
+ "learning_rate": 0.0006,
+ "loss": 7.171927452087402,
+ "step": 188
+ },
+ {
+ "epoch": 2.62882096069869,
+ "grad_norm": 0.07825770974159241,
+ "learning_rate": 0.0006,
+ "loss": 7.207631587982178,
+ "step": 189
+ },
+ {
+ "epoch": 2.6427947598253274,
+ "grad_norm": 0.07118063420057297,
+ "learning_rate": 0.0006,
+ "loss": 7.209488391876221,
+ "step": 190
+ },
+ {
+ "epoch": 2.656768558951965,
+ "grad_norm": 0.05153293535113335,
+ "learning_rate": 0.0006,
+ "loss": 7.14793062210083,
+ "step": 191
+ },
+ {
+ "epoch": 2.6707423580786025,
+ "grad_norm": 0.03318839520215988,
+ "learning_rate": 0.0006,
+ "loss": 7.163010597229004,
+ "step": 192
+ },
+ {
+ "epoch": 2.6847161572052403,
+ "grad_norm": 0.029333539307117462,
+ "learning_rate": 0.0006,
+ "loss": 7.159955978393555,
+ "step": 193
+ },
+ {
+ "epoch": 2.6986899563318776,
+ "grad_norm": 0.029405880719423294,
+ "learning_rate": 0.0006,
+ "loss": 7.136178493499756,
+ "step": 194
+ },
+ {
+ "epoch": 2.7126637554585153,
+ "grad_norm": 0.037815988063812256,
+ "learning_rate": 0.0006,
+ "loss": 7.183186054229736,
+ "step": 195
+ },
+ {
+ "epoch": 2.726637554585153,
+ "grad_norm": 0.020307250320911407,
+ "learning_rate": 0.0006,
+ "loss": 7.133164882659912,
+ "step": 196
+ },
+ {
+ "epoch": 2.7406113537117904,
+ "grad_norm": 0.026048069819808006,
+ "learning_rate": 0.0006,
+ "loss": 7.124977111816406,
+ "step": 197
+ },
+ {
+ "epoch": 2.7545851528384278,
+ "grad_norm": 0.027316724881529808,
+ "learning_rate": 0.0006,
+ "loss": 7.07265567779541,
+ "step": 198
+ },
+ {
+ "epoch": 2.7685589519650655,
+ "grad_norm": 0.021629009395837784,
+ "learning_rate": 0.0006,
+ "loss": 7.150020599365234,
+ "step": 199
+ },
+ {
+ "epoch": 2.7825327510917033,
+ "grad_norm": 0.01605929434299469,
+ "learning_rate": 0.0006,
+ "loss": 7.10919189453125,
+ "step": 200
+ },
+ {
+ "epoch": 2.7965065502183406,
+ "grad_norm": 0.020800089463591576,
+ "learning_rate": 0.0006,
+ "loss": 7.132045269012451,
+ "step": 201
+ },
+ {
+ "epoch": 2.810480349344978,
+ "grad_norm": 0.019225774332880974,
+ "learning_rate": 0.0006,
+ "loss": 7.128364562988281,
+ "step": 202
+ },
+ {
+ "epoch": 2.8244541484716157,
+ "grad_norm": 0.01561372634023428,
+ "learning_rate": 0.0006,
+ "loss": 7.083669662475586,
+ "step": 203
+ },
+ {
+ "epoch": 2.8384279475982535,
+ "grad_norm": 0.021969519555568695,
+ "learning_rate": 0.0006,
+ "loss": 7.050841331481934,
+ "step": 204
+ },
+ {
+ "epoch": 2.8524017467248908,
+ "grad_norm": 0.021453695371747017,
+ "learning_rate": 0.0006,
+ "loss": 7.097085475921631,
+ "step": 205
+ },
+ {
+ "epoch": 2.8663755458515285,
+ "grad_norm": 0.022836022078990936,
+ "learning_rate": 0.0006,
+ "loss": 7.10405969619751,
+ "step": 206
+ },
+ {
+ "epoch": 2.880349344978166,
+ "grad_norm": 0.02430851384997368,
+ "learning_rate": 0.0006,
+ "loss": 7.045181751251221,
+ "step": 207
+ },
+ {
+ "epoch": 2.8943231441048036,
+ "grad_norm": 0.020166993141174316,
+ "learning_rate": 0.0006,
+ "loss": 7.068183898925781,
+ "step": 208
+ },
+ {
+ "epoch": 2.908296943231441,
+ "grad_norm": 0.0118505684658885,
+ "learning_rate": 0.0006,
+ "loss": 7.07132625579834,
+ "step": 209
+ },
+ {
+ "epoch": 2.9222707423580787,
+ "grad_norm": 0.023772427812218666,
+ "learning_rate": 0.0006,
+ "loss": 7.055238723754883,
+ "step": 210
+ },
+ {
+ "epoch": 2.936244541484716,
+ "grad_norm": 0.02350712940096855,
+ "learning_rate": 0.0006,
+ "loss": 7.055898666381836,
+ "step": 211
+ },
+ {
+ "epoch": 2.950218340611354,
+ "grad_norm": 0.017244907096028328,
+ "learning_rate": 0.0006,
+ "loss": 7.0099711418151855,
+ "step": 212
+ },
+ {
+ "epoch": 2.964192139737991,
+ "grad_norm": 0.021565575152635574,
+ "learning_rate": 0.0006,
+ "loss": 6.991751670837402,
+ "step": 213
+ },
+ {
+ "epoch": 2.978165938864629,
+ "grad_norm": 0.03241860866546631,
+ "learning_rate": 0.0006,
+ "loss": 7.005980491638184,
+ "step": 214
+ },
+ {
+ "epoch": 2.992139737991266,
+ "grad_norm": 0.04613726586103439,
+ "learning_rate": 0.0006,
+ "loss": 7.027859687805176,
+ "step": 215
+ },
+ {
+ "epoch": 3.0,
+ "grad_norm": 0.06404894590377808,
+ "learning_rate": 0.0006,
+ "loss": 7.123725891113281,
+ "step": 216
+ },
+ {
+ "epoch": 3.0,
+ "eval_loss": 7.070647716522217,
+ "eval_runtime": 58.5801,
+ "eval_samples_per_second": 41.687,
+ "eval_steps_per_second": 1.314,
+ "step": 216
+ },
+ {
+ "epoch": 3.0139737991266378,
+ "grad_norm": 0.05542432889342308,
+ "learning_rate": 0.0006,
+ "loss": 7.078299522399902,
+ "step": 217
+ },
+ {
+ "epoch": 3.027947598253275,
+ "grad_norm": 0.026484526693820953,
+ "learning_rate": 0.0006,
+ "loss": 6.961159706115723,
+ "step": 218
+ },
+ {
+ "epoch": 3.041921397379913,
+ "grad_norm": 0.028276391327381134,
+ "learning_rate": 0.0006,
+ "loss": 7.019961833953857,
+ "step": 219
+ },
+ {
+ "epoch": 3.05589519650655,
+ "grad_norm": 0.024486595764756203,
+ "learning_rate": 0.0006,
+ "loss": 7.039677619934082,
+ "step": 220
+ },
+ {
+ "epoch": 3.069868995633188,
+ "grad_norm": 0.017620893195271492,
+ "learning_rate": 0.0006,
+ "loss": 6.9270782470703125,
+ "step": 221
+ },
+ {
+ "epoch": 3.0838427947598253,
+ "grad_norm": 0.02547302283346653,
+ "learning_rate": 0.0006,
+ "loss": 6.924233436584473,
+ "step": 222
+ },
+ {
+ "epoch": 3.097816593886463,
+ "grad_norm": 0.02539009042084217,
+ "learning_rate": 0.0006,
+ "loss": 6.973979949951172,
+ "step": 223
+ },
+ {
+ "epoch": 3.1117903930131003,
+ "grad_norm": 0.03256227448582649,
+ "learning_rate": 0.0006,
+ "loss": 6.933725357055664,
+ "step": 224
+ },
+ {
+ "epoch": 3.125764192139738,
+ "grad_norm": 0.04651800915598869,
+ "learning_rate": 0.0006,
+ "loss": 6.951044082641602,
+ "step": 225
+ },
+ {
+ "epoch": 3.1397379912663754,
+ "grad_norm": 0.04822975769639015,
+ "learning_rate": 0.0006,
+ "loss": 6.9614410400390625,
+ "step": 226
+ },
+ {
+ "epoch": 3.153711790393013,
+ "grad_norm": 0.031165020540356636,
+ "learning_rate": 0.0006,
+ "loss": 6.91193151473999,
+ "step": 227
+ },
+ {
+ "epoch": 3.1676855895196505,
+ "grad_norm": 0.029754292219877243,
+ "learning_rate": 0.0006,
+ "loss": 6.91802978515625,
+ "step": 228
+ },
+ {
+ "epoch": 3.1816593886462883,
+ "grad_norm": 0.02193152718245983,
+ "learning_rate": 0.0006,
+ "loss": 6.928090572357178,
+ "step": 229
+ },
+ {
+ "epoch": 3.1956331877729256,
+ "grad_norm": 0.02428470179438591,
+ "learning_rate": 0.0006,
+ "loss": 6.90396785736084,
+ "step": 230
+ },
+ {
+ "epoch": 3.2096069868995634,
+ "grad_norm": 0.023374345153570175,
+ "learning_rate": 0.0006,
+ "loss": 6.885858535766602,
+ "step": 231
+ },
+ {
+ "epoch": 3.2235807860262007,
+ "grad_norm": 0.023405000567436218,
+ "learning_rate": 0.0006,
+ "loss": 6.8855767250061035,
+ "step": 232
+ },
+ {
+ "epoch": 3.2375545851528384,
+ "grad_norm": 0.017277134582400322,
+ "learning_rate": 0.0006,
+ "loss": 6.901449203491211,
+ "step": 233
+ },
+ {
+ "epoch": 3.251528384279476,
+ "grad_norm": 0.019496750086545944,
+ "learning_rate": 0.0006,
+ "loss": 6.885687828063965,
+ "step": 234
+ },
+ {
+ "epoch": 3.2655021834061135,
+ "grad_norm": 0.016893276944756508,
+ "learning_rate": 0.0006,
+ "loss": 6.872369766235352,
+ "step": 235
+ },
+ {
+ "epoch": 3.279475982532751,
+ "grad_norm": 0.02058715932071209,
+ "learning_rate": 0.0006,
+ "loss": 6.823280334472656,
+ "step": 236
+ },
+ {
+ "epoch": 3.2934497816593886,
+ "grad_norm": 0.015530755743384361,
+ "learning_rate": 0.0006,
+ "loss": 6.860499858856201,
+ "step": 237
+ },
+ {
+ "epoch": 3.3074235807860264,
+ "grad_norm": 0.01934926211833954,
+ "learning_rate": 0.0006,
+ "loss": 6.817448616027832,
+ "step": 238
+ },
+ {
+ "epoch": 3.3213973799126637,
+ "grad_norm": 0.017413552850484848,
+ "learning_rate": 0.0006,
+ "loss": 6.881955146789551,
+ "step": 239
+ },
+ {
+ "epoch": 3.3353711790393015,
+ "grad_norm": 0.026478875428438187,
+ "learning_rate": 0.0006,
+ "loss": 6.827404499053955,
+ "step": 240
+ },
+ {
+ "epoch": 3.349344978165939,
+ "grad_norm": 0.032611701637506485,
+ "learning_rate": 0.0006,
+ "loss": 6.78285551071167,
+ "step": 241
+ },
+ {
+ "epoch": 3.3633187772925766,
+ "grad_norm": 0.041900306940078735,
+ "learning_rate": 0.0006,
+ "loss": 6.783053874969482,
+ "step": 242
+ },
+ {
+ "epoch": 3.377292576419214,
+ "grad_norm": 0.04149497672915459,
+ "learning_rate": 0.0006,
+ "loss": 6.745220184326172,
+ "step": 243
+ },
+ {
+ "epoch": 3.3912663755458516,
+ "grad_norm": 0.023949826136231422,
+ "learning_rate": 0.0006,
+ "loss": 6.841533184051514,
+ "step": 244
+ },
+ {
+ "epoch": 3.405240174672489,
+ "grad_norm": 0.028601842001080513,
+ "learning_rate": 0.0006,
+ "loss": 6.852852821350098,
+ "step": 245
+ },
+ {
+ "epoch": 3.4192139737991267,
+ "grad_norm": 0.027365155518054962,
+ "learning_rate": 0.0006,
+ "loss": 6.74376106262207,
+ "step": 246
+ },
+ {
+ "epoch": 3.433187772925764,
+ "grad_norm": 0.02856568619608879,
+ "learning_rate": 0.0006,
+ "loss": 6.74906063079834,
+ "step": 247
+ },
+ {
+ "epoch": 3.447161572052402,
+ "grad_norm": 0.019172416999936104,
+ "learning_rate": 0.0006,
+ "loss": 6.809206008911133,
+ "step": 248
+ },
+ {
+ "epoch": 3.461135371179039,
+ "grad_norm": 0.02161746844649315,
+ "learning_rate": 0.0006,
+ "loss": 6.789888381958008,
+ "step": 249
+ },
+ {
+ "epoch": 3.475109170305677,
+ "grad_norm": 0.020273666828870773,
+ "learning_rate": 0.0006,
+ "loss": 6.73512077331543,
+ "step": 250
+ },
+ {
+ "epoch": 3.489082969432314,
+ "grad_norm": 0.022329283878207207,
+ "learning_rate": 0.0006,
+ "loss": 6.826148509979248,
+ "step": 251
+ },
+ {
+ "epoch": 3.503056768558952,
+ "grad_norm": 0.02284284122288227,
+ "learning_rate": 0.0006,
+ "loss": 6.748521327972412,
+ "step": 252
+ },
+ {
+ "epoch": 3.5170305676855893,
+ "grad_norm": 0.026368247345089912,
+ "learning_rate": 0.0006,
+ "loss": 6.645879745483398,
+ "step": 253
+ },
+ {
+ "epoch": 3.531004366812227,
+ "grad_norm": 0.03627244383096695,
+ "learning_rate": 0.0006,
+ "loss": 6.666064739227295,
+ "step": 254
+ },
+ {
+ "epoch": 3.544978165938865,
+ "grad_norm": 0.04140935093164444,
+ "learning_rate": 0.0006,
+ "loss": 6.712275981903076,
+ "step": 255
+ },
+ {
+ "epoch": 3.558951965065502,
+ "grad_norm": 0.0390462800860405,
+ "learning_rate": 0.0006,
+ "loss": 6.665900230407715,
+ "step": 256
+ },
+ {
+ "epoch": 3.5729257641921395,
+ "grad_norm": 0.05178583785891533,
+ "learning_rate": 0.0006,
+ "loss": 6.753298759460449,
+ "step": 257
+ },
+ {
+ "epoch": 3.5868995633187772,
+ "grad_norm": 0.054328061640262604,
+ "learning_rate": 0.0006,
+ "loss": 6.732633113861084,
+ "step": 258
+ },
+ {
+ "epoch": 3.600873362445415,
+ "grad_norm": 0.04590460658073425,
+ "learning_rate": 0.0006,
+ "loss": 6.65250301361084,
+ "step": 259
+ },
+ {
+ "epoch": 3.6148471615720523,
+ "grad_norm": 0.044331666082143784,
+ "learning_rate": 0.0006,
+ "loss": 6.68654203414917,
+ "step": 260
+ },
+ {
+ "epoch": 3.62882096069869,
+ "grad_norm": 0.05308730527758598,
+ "learning_rate": 0.0006,
+ "loss": 6.642482757568359,
+ "step": 261
+ },
+ {
+ "epoch": 3.6427947598253274,
+ "grad_norm": 0.03999016433954239,
+ "learning_rate": 0.0006,
+ "loss": 6.625227451324463,
+ "step": 262
+ },
+ {
+ "epoch": 3.656768558951965,
+ "grad_norm": 0.04088086634874344,
+ "learning_rate": 0.0006,
+ "loss": 6.67623233795166,
+ "step": 263
+ },
+ {
+ "epoch": 3.6707423580786025,
+ "grad_norm": 0.03615636005997658,
+ "learning_rate": 0.0006,
+ "loss": 6.6869001388549805,
+ "step": 264
+ },
+ {
+ "epoch": 3.6847161572052403,
+ "grad_norm": 0.02817360684275627,
+ "learning_rate": 0.0006,
+ "loss": 6.674601078033447,
+ "step": 265
+ },
+ {
+ "epoch": 3.6986899563318776,
+ "grad_norm": 0.032586123794317245,
+ "learning_rate": 0.0006,
+ "loss": 6.612994194030762,
+ "step": 266
+ },
+ {
+ "epoch": 3.7126637554585153,
+ "grad_norm": 0.031117867678403854,
+ "learning_rate": 0.0006,
+ "loss": 6.6244354248046875,
+ "step": 267
+ },
+ {
+ "epoch": 3.726637554585153,
+ "grad_norm": 0.028330687433481216,
+ "learning_rate": 0.0006,
+ "loss": 6.669557094573975,
+ "step": 268
+ },
+ {
+ "epoch": 3.7406113537117904,
+ "grad_norm": 0.021489202976226807,
+ "learning_rate": 0.0006,
+ "loss": 6.632035732269287,
+ "step": 269
+ },
+ {
+ "epoch": 3.7545851528384278,
+ "grad_norm": 0.024269424378871918,
+ "learning_rate": 0.0006,
+ "loss": 6.660867691040039,
+ "step": 270
+ },
+ {
+ "epoch": 3.7685589519650655,
+ "grad_norm": 0.02237873338162899,
+ "learning_rate": 0.0006,
+ "loss": 6.602587699890137,
+ "step": 271
+ },
+ {
+ "epoch": 3.7825327510917033,
+ "grad_norm": 0.026503929868340492,
+ "learning_rate": 0.0006,
+ "loss": 6.599200248718262,
+ "step": 272
+ },
+ {
+ "epoch": 3.7965065502183406,
+ "grad_norm": 0.03687124326825142,
+ "learning_rate": 0.0006,
+ "loss": 6.567296504974365,
+ "step": 273
+ },
+ {
+ "epoch": 3.810480349344978,
+ "grad_norm": 0.04708952456712723,
+ "learning_rate": 0.0006,
+ "loss": 6.621265411376953,
+ "step": 274
+ },
+ {
+ "epoch": 3.8244541484716157,
+ "grad_norm": 0.041846614331007004,
+ "learning_rate": 0.0006,
+ "loss": 6.550536155700684,
+ "step": 275
+ },
+ {
+ "epoch": 3.8384279475982535,
+ "grad_norm": 0.031249675899744034,
+ "learning_rate": 0.0006,
+ "loss": 6.54278564453125,
+ "step": 276
+ },
+ {
+ "epoch": 3.8524017467248908,
+ "grad_norm": 0.023082256317138672,
+ "learning_rate": 0.0006,
+ "loss": 6.544787406921387,
+ "step": 277
+ },
+ {
+ "epoch": 3.8663755458515285,
+ "grad_norm": 0.020995570346713066,
+ "learning_rate": 0.0006,
+ "loss": 6.518143653869629,
+ "step": 278
+ },
+ {
+ "epoch": 3.880349344978166,
+ "grad_norm": 0.020839324221014977,
+ "learning_rate": 0.0006,
+ "loss": 6.546667098999023,
+ "step": 279
+ },
+ {
+ "epoch": 3.8943231441048036,
+ "grad_norm": 0.018377432599663734,
+ "learning_rate": 0.0006,
+ "loss": 6.478307723999023,
+ "step": 280
+ },
+ {
+ "epoch": 3.908296943231441,
+ "grad_norm": 0.021713724359869957,
+ "learning_rate": 0.0006,
+ "loss": 6.485574245452881,
+ "step": 281
+ },
+ {
+ "epoch": 3.9222707423580787,
+ "grad_norm": 0.01359301246702671,
+ "learning_rate": 0.0006,
+ "loss": 6.570268630981445,
+ "step": 282
+ },
+ {
+ "epoch": 3.936244541484716,
+ "grad_norm": 0.016233332455158234,
+ "learning_rate": 0.0006,
+ "loss": 6.57150936126709,
+ "step": 283
+ },
+ {
+ "epoch": 3.950218340611354,
+ "grad_norm": 0.020687608048319817,
+ "learning_rate": 0.0006,
+ "loss": 6.527956008911133,
+ "step": 284
+ },
+ {
+ "epoch": 3.964192139737991,
+ "grad_norm": 0.02032964862883091,
+ "learning_rate": 0.0006,
+ "loss": 6.468169689178467,
+ "step": 285
+ },
+ {
+ "epoch": 3.978165938864629,
+ "grad_norm": 0.016676288098096848,
+ "learning_rate": 0.0006,
+ "loss": 6.448418617248535,
+ "step": 286
+ },
+ {
+ "epoch": 3.992139737991266,
+ "grad_norm": 0.0168539360165596,
+ "learning_rate": 0.0006,
+ "loss": 6.461180686950684,
+ "step": 287
+ },
+ {
+ "epoch": 4.0,
+ "grad_norm": 0.02666233666241169,
+ "learning_rate": 0.0006,
+ "loss": 6.344893932342529,
+ "step": 288
+ },
+ {
+ "epoch": 4.0,
+ "eval_loss": 6.5384979248046875,
+ "eval_runtime": 58.814,
+ "eval_samples_per_second": 41.521,
+ "eval_steps_per_second": 1.309,
+ "step": 288
+ },
+ {
+ "epoch": 4.013973799126638,
+ "grad_norm": 0.04154708981513977,
+ "learning_rate": 0.0006,
+ "loss": 6.41855525970459,
+ "step": 289
+ },
+ {
+ "epoch": 4.0279475982532755,
+ "grad_norm": 0.04895668104290962,
+ "learning_rate": 0.0006,
+ "loss": 6.499302864074707,
+ "step": 290
+ },
+ {
+ "epoch": 4.041921397379912,
+ "grad_norm": 0.034464482218027115,
+ "learning_rate": 0.0006,
+ "loss": 6.438611030578613,
+ "step": 291
+ },
+ {
+ "epoch": 4.05589519650655,
+ "grad_norm": 0.06845773756504059,
+ "learning_rate": 0.0006,
+ "loss": 6.538361072540283,
+ "step": 292
+ },
+ {
+ "epoch": 4.069868995633188,
+ "grad_norm": 0.053768858313560486,
+ "learning_rate": 0.0006,
+ "loss": 6.527524948120117,
+ "step": 293
+ },
+ {
+ "epoch": 4.083842794759826,
+ "grad_norm": 0.039037931710481644,
+ "learning_rate": 0.0006,
+ "loss": 6.503847122192383,
+ "step": 294
+ },
+ {
+ "epoch": 4.097816593886463,
+ "grad_norm": 0.03962196782231331,
+ "learning_rate": 0.0006,
+ "loss": 6.475178241729736,
+ "step": 295
+ },
+ {
+ "epoch": 4.1117903930131,
+ "grad_norm": 0.049101535230875015,
+ "learning_rate": 0.0006,
+ "loss": 6.474587917327881,
+ "step": 296
+ },
+ {
+ "epoch": 4.125764192139738,
+ "grad_norm": 0.02830282226204872,
+ "learning_rate": 0.0006,
+ "loss": 6.45065975189209,
+ "step": 297
+ },
+ {
+ "epoch": 4.139737991266376,
+ "grad_norm": 0.023153482005000114,
+ "learning_rate": 0.0006,
+ "loss": 6.5218706130981445,
+ "step": 298
+ },
+ {
+ "epoch": 4.153711790393013,
+ "grad_norm": 0.025064002722501755,
+ "learning_rate": 0.0006,
+ "loss": 6.456475257873535,
+ "step": 299
+ },
+ {
+ "epoch": 4.1676855895196505,
+ "grad_norm": 0.02357092685997486,
+ "learning_rate": 0.0006,
+ "loss": 6.485952377319336,
+ "step": 300
+ },
+ {
+ "epoch": 4.181659388646288,
+ "grad_norm": 0.021832305938005447,
+ "learning_rate": 0.0006,
+ "loss": 6.425506591796875,
+ "step": 301
+ },
+ {
+ "epoch": 4.195633187772926,
+ "grad_norm": 0.0227900929749012,
+ "learning_rate": 0.0006,
+ "loss": 6.439155578613281,
+ "step": 302
+ },
+ {
+ "epoch": 4.209606986899563,
+ "grad_norm": 0.015961607918143272,
+ "learning_rate": 0.0006,
+ "loss": 6.460862636566162,
+ "step": 303
+ },
+ {
+ "epoch": 4.223580786026201,
+ "grad_norm": 0.017876390367746353,
+ "learning_rate": 0.0006,
+ "loss": 6.44564151763916,
+ "step": 304
+ },
+ {
+ "epoch": 4.2375545851528384,
+ "grad_norm": 0.013703204691410065,
+ "learning_rate": 0.0006,
+ "loss": 6.418414115905762,
+ "step": 305
+ },
+ {
+ "epoch": 4.251528384279476,
+ "grad_norm": 0.015240306034684181,
+ "learning_rate": 0.0006,
+ "loss": 6.411238670349121,
+ "step": 306
+ },
+ {
+ "epoch": 4.265502183406113,
+ "grad_norm": 0.01587662845849991,
+ "learning_rate": 0.0006,
+ "loss": 6.395642280578613,
+ "step": 307
+ },
+ {
+ "epoch": 4.279475982532751,
+ "grad_norm": 0.014650700613856316,
+ "learning_rate": 0.0006,
+ "loss": 6.302793025970459,
+ "step": 308
+ },
+ {
+ "epoch": 4.293449781659389,
+ "grad_norm": 0.016533225774765015,
+ "learning_rate": 0.0006,
+ "loss": 6.380588054656982,
+ "step": 309
+ },
+ {
+ "epoch": 4.307423580786026,
+ "grad_norm": 0.019963741302490234,
+ "learning_rate": 0.0006,
+ "loss": 6.345336437225342,
+ "step": 310
+ },
+ {
+ "epoch": 4.321397379912664,
+ "grad_norm": 0.02237936295568943,
+ "learning_rate": 0.0006,
+ "loss": 6.387770652770996,
+ "step": 311
+ },
+ {
+ "epoch": 4.335371179039301,
+ "grad_norm": 0.026658328250050545,
+ "learning_rate": 0.0006,
+ "loss": 6.323662757873535,
+ "step": 312
+ },
+ {
+ "epoch": 4.349344978165939,
+ "grad_norm": 0.02852284163236618,
+ "learning_rate": 0.0006,
+ "loss": 6.382204055786133,
+ "step": 313
+ },
+ {
+ "epoch": 4.3633187772925766,
+ "grad_norm": 0.028815090656280518,
+ "learning_rate": 0.0006,
+ "loss": 6.402390480041504,
+ "step": 314
+ },
+ {
+ "epoch": 4.377292576419214,
+ "grad_norm": 0.029393529519438744,
+ "learning_rate": 0.0006,
+ "loss": 6.280278205871582,
+ "step": 315
+ },
+ {
+ "epoch": 4.391266375545851,
+ "grad_norm": 0.02186041697859764,
+ "learning_rate": 0.0006,
+ "loss": 6.302707672119141,
+ "step": 316
+ },
+ {
+ "epoch": 4.405240174672489,
+ "grad_norm": 0.02071218006312847,
+ "learning_rate": 0.0006,
+ "loss": 6.3365020751953125,
+ "step": 317
+ },
+ {
+ "epoch": 4.419213973799127,
+ "grad_norm": 0.02779117226600647,
+ "learning_rate": 0.0006,
+ "loss": 6.252157211303711,
+ "step": 318
+ },
+ {
+ "epoch": 4.4331877729257645,
+ "grad_norm": 0.03470654785633087,
+ "learning_rate": 0.0006,
+ "loss": 6.319070339202881,
+ "step": 319
+ },
+ {
+ "epoch": 4.447161572052401,
+ "grad_norm": 0.037067804485559464,
+ "learning_rate": 0.0006,
+ "loss": 6.314116477966309,
+ "step": 320
+ },
+ {
+ "epoch": 4.461135371179039,
+ "grad_norm": 0.032182756811380386,
+ "learning_rate": 0.0006,
+ "loss": 6.267298698425293,
+ "step": 321
+ },
+ {
+ "epoch": 4.475109170305677,
+ "grad_norm": 0.026305217295885086,
+ "learning_rate": 0.0006,
+ "loss": 6.331688404083252,
+ "step": 322
+ },
+ {
+ "epoch": 4.489082969432315,
+ "grad_norm": 0.027228357270359993,
+ "learning_rate": 0.0006,
+ "loss": 6.31139612197876,
+ "step": 323
+ },
+ {
+ "epoch": 4.503056768558952,
+ "grad_norm": 0.025781169533729553,
+ "learning_rate": 0.0006,
+ "loss": 6.194684982299805,
+ "step": 324
+ },
+ {
+ "epoch": 4.517030567685589,
+ "grad_norm": 0.029186977073550224,
+ "learning_rate": 0.0006,
+ "loss": 6.31306266784668,
+ "step": 325
+ },
+ {
+ "epoch": 4.531004366812227,
+ "grad_norm": 0.023608777672052383,
+ "learning_rate": 0.0006,
+ "loss": 6.285243034362793,
+ "step": 326
+ },
+ {
+ "epoch": 4.544978165938865,
+ "grad_norm": 0.02267594076693058,
+ "learning_rate": 0.0006,
+ "loss": 6.30342960357666,
+ "step": 327
+ },
+ {
+ "epoch": 4.558951965065502,
+ "grad_norm": 0.01926310732960701,
+ "learning_rate": 0.0006,
+ "loss": 6.280606269836426,
+ "step": 328
+ },
+ {
+ "epoch": 4.5729257641921395,
+ "grad_norm": 0.026484837755560875,
+ "learning_rate": 0.0006,
+ "loss": 6.237980365753174,
+ "step": 329
+ },
+ {
+ "epoch": 4.586899563318777,
+ "grad_norm": 0.026228854432702065,
+ "learning_rate": 0.0006,
+ "loss": 6.3018293380737305,
+ "step": 330
+ },
+ {
+ "epoch": 4.600873362445415,
+ "grad_norm": 0.022096967324614525,
+ "learning_rate": 0.0006,
+ "loss": 6.253863334655762,
+ "step": 331
+ },
+ {
+ "epoch": 4.614847161572053,
+ "grad_norm": 0.027223890647292137,
+ "learning_rate": 0.0006,
+ "loss": 6.166882038116455,
+ "step": 332
+ },
+ {
+ "epoch": 4.62882096069869,
+ "grad_norm": 0.03160124272108078,
+ "learning_rate": 0.0006,
+ "loss": 6.230309963226318,
+ "step": 333
+ },
+ {
+ "epoch": 4.642794759825327,
+ "grad_norm": 0.04110539332032204,
+ "learning_rate": 0.0006,
+ "loss": 6.2334818840026855,
+ "step": 334
+ },
+ {
+ "epoch": 4.656768558951965,
+ "grad_norm": 0.04355933889746666,
+ "learning_rate": 0.0006,
+ "loss": 6.160956382751465,
+ "step": 335
+ },
+ {
+ "epoch": 4.670742358078603,
+ "grad_norm": 0.03611086308956146,
+ "learning_rate": 0.0006,
+ "loss": 6.23399543762207,
+ "step": 336
+ },
+ {
+ "epoch": 4.68471615720524,
+ "grad_norm": 0.051811400800943375,
+ "learning_rate": 0.0006,
+ "loss": 6.241855144500732,
+ "step": 337
+ },
+ {
+ "epoch": 4.698689956331878,
+ "grad_norm": 0.057434193789958954,
+ "learning_rate": 0.0006,
+ "loss": 6.220993995666504,
+ "step": 338
+ },
+ {
+ "epoch": 4.712663755458515,
+ "grad_norm": 0.04420953616499901,
+ "learning_rate": 0.0006,
+ "loss": 6.300461292266846,
+ "step": 339
+ },
+ {
+ "epoch": 4.726637554585153,
+ "grad_norm": 0.03515457361936569,
+ "learning_rate": 0.0006,
+ "loss": 6.269349098205566,
+ "step": 340
+ },
+ {
+ "epoch": 4.74061135371179,
+ "grad_norm": 0.03222600743174553,
+ "learning_rate": 0.0006,
+ "loss": 6.248613357543945,
+ "step": 341
+ },
+ {
+ "epoch": 4.754585152838428,
+ "grad_norm": 0.0384046845138073,
+ "learning_rate": 0.0006,
+ "loss": 6.266979217529297,
+ "step": 342
+ },
+ {
+ "epoch": 4.7685589519650655,
+ "grad_norm": 0.0514867827296257,
+ "learning_rate": 0.0006,
+ "loss": 6.234103202819824,
+ "step": 343
+ },
+ {
+ "epoch": 4.782532751091703,
+ "grad_norm": 0.050747793167829514,
+ "learning_rate": 0.0006,
+ "loss": 6.322582244873047,
+ "step": 344
+ },
+ {
+ "epoch": 4.796506550218341,
+ "grad_norm": 0.040878988802433014,
+ "learning_rate": 0.0006,
+ "loss": 6.212509632110596,
+ "step": 345
+ },
+ {
+ "epoch": 4.810480349344978,
+ "grad_norm": 0.034655820578336716,
+ "learning_rate": 0.0006,
+ "loss": 6.265879154205322,
+ "step": 346
+ },
+ {
+ "epoch": 4.824454148471616,
+ "grad_norm": 0.04202224686741829,
+ "learning_rate": 0.0006,
+ "loss": 6.2896409034729,
+ "step": 347
+ },
+ {
+ "epoch": 4.8384279475982535,
+ "grad_norm": 0.03632277995347977,
+ "learning_rate": 0.0006,
+ "loss": 6.253917217254639,
+ "step": 348
+ },
+ {
+ "epoch": 4.85240174672489,
+ "grad_norm": 0.028645765036344528,
+ "learning_rate": 0.0006,
+ "loss": 6.207965850830078,
+ "step": 349
+ },
+ {
+ "epoch": 4.866375545851528,
+ "grad_norm": 0.025305170565843582,
+ "learning_rate": 0.0006,
+ "loss": 6.168068885803223,
+ "step": 350
+ },
+ {
+ "epoch": 4.880349344978166,
+ "grad_norm": 0.022137803956866264,
+ "learning_rate": 0.0006,
+ "loss": 6.235283374786377,
+ "step": 351
+ },
+ {
+ "epoch": 4.894323144104804,
+ "grad_norm": 0.01814538985490799,
+ "learning_rate": 0.0006,
+ "loss": 6.259641647338867,
+ "step": 352
+ },
+ {
+ "epoch": 4.908296943231441,
+ "grad_norm": 0.018465086817741394,
+ "learning_rate": 0.0006,
+ "loss": 6.111738681793213,
+ "step": 353
+ },
+ {
+ "epoch": 4.922270742358078,
+ "grad_norm": 0.018638933077454567,
+ "learning_rate": 0.0006,
+ "loss": 6.225642204284668,
+ "step": 354
+ },
+ {
+ "epoch": 4.936244541484716,
+ "grad_norm": 0.016461260616779327,
+ "learning_rate": 0.0006,
+ "loss": 6.140647888183594,
+ "step": 355
+ },
+ {
+ "epoch": 4.950218340611354,
+ "grad_norm": 0.015458385460078716,
+ "learning_rate": 0.0006,
+ "loss": 6.155224800109863,
+ "step": 356
+ },
+ {
+ "epoch": 4.964192139737992,
+ "grad_norm": 0.014934048056602478,
+ "learning_rate": 0.0006,
+ "loss": 6.101555824279785,
+ "step": 357
+ },
+ {
+ "epoch": 4.978165938864628,
+ "grad_norm": 0.013979545794427395,
+ "learning_rate": 0.0006,
+ "loss": 6.161718845367432,
+ "step": 358
+ },
+ {
+ "epoch": 4.992139737991266,
+ "grad_norm": 0.01640394888818264,
+ "learning_rate": 0.0006,
+ "loss": 6.055768966674805,
+ "step": 359
+ },
+ {
+ "epoch": 5.0,
+ "grad_norm": 0.015135680325329304,
+ "learning_rate": 0.0006,
+ "loss": 6.169695854187012,
+ "step": 360
+ },
+ {
+ "epoch": 5.0,
+ "eval_loss": 6.16408109664917,
+ "eval_runtime": 59.3649,
+ "eval_samples_per_second": 41.135,
+ "eval_steps_per_second": 1.297,
+ "step": 360
+ },
+ {
+ "epoch": 5.013973799126638,
+ "grad_norm": 0.012849048711359501,
+ "learning_rate": 0.0006,
+ "loss": 6.055830001831055,
+ "step": 361
+ },
+ {
+ "epoch": 5.0279475982532755,
+ "grad_norm": 0.018169576302170753,
+ "learning_rate": 0.0006,
+ "loss": 6.082655429840088,
+ "step": 362
+ },
+ {
+ "epoch": 5.041921397379912,
+ "grad_norm": 0.024293430149555206,
+ "learning_rate": 0.0006,
+ "loss": 6.109032154083252,
+ "step": 363
+ },
+ {
+ "epoch": 5.05589519650655,
+ "grad_norm": 0.03266787901520729,
+ "learning_rate": 0.0006,
+ "loss": 6.030613899230957,
+ "step": 364
+ },
+ {
+ "epoch": 5.069868995633188,
+ "grad_norm": 0.03572266176342964,
+ "learning_rate": 0.0006,
+ "loss": 6.050390720367432,
+ "step": 365
+ },
+ {
+ "epoch": 5.083842794759826,
+ "grad_norm": 0.032625213265419006,
+ "learning_rate": 0.0006,
+ "loss": 6.169313430786133,
+ "step": 366
+ },
+ {
+ "epoch": 5.097816593886463,
+ "grad_norm": 0.04051872715353966,
+ "learning_rate": 0.0006,
+ "loss": 6.10398006439209,
+ "step": 367
+ },
+ {
+ "epoch": 5.1117903930131,
+ "grad_norm": 0.0351213738322258,
+ "learning_rate": 0.0006,
+ "loss": 6.121973514556885,
+ "step": 368
+ },
+ {
+ "epoch": 5.125764192139738,
+ "grad_norm": 0.03523759916424751,
+ "learning_rate": 0.0006,
+ "loss": 6.0626654624938965,
+ "step": 369
+ },
+ {
+ "epoch": 5.139737991266376,
+ "grad_norm": 0.02706182189285755,
+ "learning_rate": 0.0006,
+ "loss": 6.049354076385498,
+ "step": 370
+ },
+ {
+ "epoch": 5.153711790393013,
+ "grad_norm": 0.024212589487433434,
+ "learning_rate": 0.0006,
+ "loss": 6.034826278686523,
+ "step": 371
+ },
+ {
+ "epoch": 5.1676855895196505,
+ "grad_norm": 0.024798082187771797,
+ "learning_rate": 0.0006,
+ "loss": 6.004058837890625,
+ "step": 372
+ },
+ {
+ "epoch": 5.181659388646288,
+ "grad_norm": 0.02325567416846752,
+ "learning_rate": 0.0006,
+ "loss": 5.986461639404297,
+ "step": 373
+ },
+ {
+ "epoch": 5.195633187772926,
+ "grad_norm": 0.019060570746660233,
+ "learning_rate": 0.0006,
+ "loss": 6.059736251831055,
+ "step": 374
+ },
+ {
+ "epoch": 5.209606986899563,
+ "grad_norm": 0.016822976991534233,
+ "learning_rate": 0.0006,
+ "loss": 6.072957515716553,
+ "step": 375
+ },
+ {
+ "epoch": 5.223580786026201,
+ "grad_norm": 0.017218658700585365,
+ "learning_rate": 0.0006,
+ "loss": 6.031739234924316,
+ "step": 376
+ },
+ {
+ "epoch": 5.2375545851528384,
+ "grad_norm": 0.014889383688569069,
+ "learning_rate": 0.0006,
+ "loss": 6.025674343109131,
+ "step": 377
+ },
+ {
+ "epoch": 5.251528384279476,
+ "grad_norm": 0.01708107627928257,
+ "learning_rate": 0.0006,
+ "loss": 6.090451240539551,
+ "step": 378
+ },
+ {
+ "epoch": 5.265502183406113,
+ "grad_norm": 0.020260317251086235,
+ "learning_rate": 0.0006,
+ "loss": 6.094257354736328,
+ "step": 379
+ },
+ {
+ "epoch": 5.279475982532751,
+ "grad_norm": 0.020110400393605232,
+ "learning_rate": 0.0006,
+ "loss": 6.021188735961914,
+ "step": 380
+ },
+ {
+ "epoch": 5.293449781659389,
+ "grad_norm": 0.02027830481529236,
+ "learning_rate": 0.0006,
+ "loss": 6.022156238555908,
+ "step": 381
+ },
+ {
+ "epoch": 5.307423580786026,
+ "grad_norm": 0.022747062146663666,
+ "learning_rate": 0.0006,
+ "loss": 6.011836051940918,
+ "step": 382
+ },
+ {
+ "epoch": 5.321397379912664,
+ "grad_norm": 0.01990230567753315,
+ "learning_rate": 0.0006,
+ "loss": 6.003054618835449,
+ "step": 383
+ },
+ {
+ "epoch": 5.335371179039301,
+ "grad_norm": 0.01596238650381565,
+ "learning_rate": 0.0006,
+ "loss": 5.991410732269287,
+ "step": 384
+ },
+ {
+ "epoch": 5.349344978165939,
+ "grad_norm": 0.015847105532884598,
+ "learning_rate": 0.0006,
+ "loss": 6.100622177124023,
+ "step": 385
+ },
+ {
+ "epoch": 5.3633187772925766,
+ "grad_norm": 0.016179397702217102,
+ "learning_rate": 0.0006,
+ "loss": 6.032659530639648,
+ "step": 386
+ },
+ {
+ "epoch": 5.377292576419214,
+ "grad_norm": 0.018256602808833122,
+ "learning_rate": 0.0006,
+ "loss": 5.961983680725098,
+ "step": 387
+ },
+ {
+ "epoch": 5.391266375545851,
+ "grad_norm": 0.02805912122130394,
+ "learning_rate": 0.0006,
+ "loss": 5.983541965484619,
+ "step": 388
+ },
+ {
+ "epoch": 5.405240174672489,
+ "grad_norm": 0.039082255214452744,
+ "learning_rate": 0.0006,
+ "loss": 6.021474361419678,
+ "step": 389
+ },
+ {
+ "epoch": 5.419213973799127,
+ "grad_norm": 0.036757659167051315,
+ "learning_rate": 0.0006,
+ "loss": 6.005046367645264,
+ "step": 390
+ },
+ {
+ "epoch": 5.4331877729257645,
+ "grad_norm": 0.035277366638183594,
+ "learning_rate": 0.0006,
+ "loss": 6.036296844482422,
+ "step": 391
+ },
+ {
+ "epoch": 5.447161572052401,
+ "grad_norm": 0.034404635429382324,
+ "learning_rate": 0.0006,
+ "loss": 5.994539260864258,
+ "step": 392
+ },
+ {
+ "epoch": 5.461135371179039,
+ "grad_norm": 0.0377206988632679,
+ "learning_rate": 0.0006,
+ "loss": 5.941534519195557,
+ "step": 393
+ },
+ {
+ "epoch": 5.475109170305677,
+ "grad_norm": 0.0389922633767128,
+ "learning_rate": 0.0006,
+ "loss": 5.983644485473633,
+ "step": 394
+ },
+ {
+ "epoch": 5.489082969432315,
+ "grad_norm": 0.04176778718829155,
+ "learning_rate": 0.0006,
+ "loss": 6.030922889709473,
+ "step": 395
+ },
+ {
+ "epoch": 5.503056768558952,
+ "grad_norm": 0.0337153784930706,
+ "learning_rate": 0.0006,
+ "loss": 5.961367130279541,
+ "step": 396
+ },
+ {
+ "epoch": 5.517030567685589,
+ "grad_norm": 0.036116816103458405,
+ "learning_rate": 0.0006,
+ "loss": 5.819280624389648,
+ "step": 397
+ },
+ {
+ "epoch": 5.531004366812227,
+ "grad_norm": 0.030724041163921356,
+ "learning_rate": 0.0006,
+ "loss": 5.906380653381348,
+ "step": 398
+ },
+ {
+ "epoch": 5.544978165938865,
+ "grad_norm": 0.030264202505350113,
+ "learning_rate": 0.0006,
+ "loss": 5.975833415985107,
+ "step": 399
+ },
+ {
+ "epoch": 5.558951965065502,
+ "grad_norm": 0.032096318900585175,
+ "learning_rate": 0.0006,
+ "loss": 6.001348972320557,
+ "step": 400
+ },
+ {
+ "epoch": 5.5729257641921395,
+ "grad_norm": 0.030579356476664543,
+ "learning_rate": 0.0006,
+ "loss": 5.928768634796143,
+ "step": 401
+ },
+ {
+ "epoch": 5.586899563318777,
+ "grad_norm": 0.028241973370313644,
+ "learning_rate": 0.0006,
+ "loss": 5.92582893371582,
+ "step": 402
+ },
+ {
+ "epoch": 5.600873362445415,
+ "grad_norm": 0.023086953908205032,
+ "learning_rate": 0.0006,
+ "loss": 5.892926216125488,
+ "step": 403
+ },
+ {
+ "epoch": 5.614847161572053,
+ "grad_norm": 0.02547992393374443,
+ "learning_rate": 0.0006,
+ "loss": 5.903195381164551,
+ "step": 404
+ },
+ {
+ "epoch": 5.62882096069869,
+ "grad_norm": 0.023089831694960594,
+ "learning_rate": 0.0006,
+ "loss": 5.912033557891846,
+ "step": 405
+ },
+ {
+ "epoch": 5.642794759825327,
+ "grad_norm": 0.025446368381381035,
+ "learning_rate": 0.0006,
+ "loss": 5.926138401031494,
+ "step": 406
+ },
+ {
+ "epoch": 5.656768558951965,
+ "grad_norm": 0.03748747706413269,
+ "learning_rate": 0.0006,
+ "loss": 5.913451194763184,
+ "step": 407
+ },
+ {
+ "epoch": 5.670742358078603,
+ "grad_norm": 0.03493810072541237,
+ "learning_rate": 0.0006,
+ "loss": 5.97633695602417,
+ "step": 408
+ },
+ {
+ "epoch": 5.68471615720524,
+ "grad_norm": 0.023131275549530983,
+ "learning_rate": 0.0006,
+ "loss": 5.899571418762207,
+ "step": 409
+ },
+ {
+ "epoch": 5.698689956331878,
+ "grad_norm": 0.02986014075577259,
+ "learning_rate": 0.0006,
+ "loss": 5.894903182983398,
+ "step": 410
+ },
+ {
+ "epoch": 5.712663755458515,
+ "grad_norm": 0.030171144753694534,
+ "learning_rate": 0.0006,
+ "loss": 5.876595497131348,
+ "step": 411
+ },
+ {
+ "epoch": 5.726637554585153,
+ "grad_norm": 0.029377546161413193,
+ "learning_rate": 0.0006,
+ "loss": 5.911637306213379,
+ "step": 412
+ },
+ {
+ "epoch": 5.74061135371179,
+ "grad_norm": 0.029829490929841995,
+ "learning_rate": 0.0006,
+ "loss": 5.819629669189453,
+ "step": 413
+ },
+ {
+ "epoch": 5.754585152838428,
+ "grad_norm": 0.024139299988746643,
+ "learning_rate": 0.0006,
+ "loss": 5.81728458404541,
+ "step": 414
+ },
+ {
+ "epoch": 5.7685589519650655,
+ "grad_norm": 0.021453432738780975,
+ "learning_rate": 0.0006,
+ "loss": 5.987326145172119,
+ "step": 415
+ },
+ {
+ "epoch": 5.782532751091703,
+ "grad_norm": 0.017942246049642563,
+ "learning_rate": 0.0006,
+ "loss": 5.893008232116699,
+ "step": 416
+ },
+ {
+ "epoch": 5.796506550218341,
+ "grad_norm": 0.019724637269973755,
+ "learning_rate": 0.0006,
+ "loss": 5.912441253662109,
+ "step": 417
+ },
+ {
+ "epoch": 5.810480349344978,
+ "grad_norm": 0.018597912043333054,
+ "learning_rate": 0.0006,
+ "loss": 5.860394477844238,
+ "step": 418
+ },
+ {
+ "epoch": 5.824454148471616,
+ "grad_norm": 0.016232412308454514,
+ "learning_rate": 0.0006,
+ "loss": 5.94991397857666,
+ "step": 419
+ },
+ {
+ "epoch": 5.8384279475982535,
+ "grad_norm": 0.01667204685509205,
+ "learning_rate": 0.0006,
+ "loss": 5.8506364822387695,
+ "step": 420
+ },
+ {
+ "epoch": 5.85240174672489,
+ "grad_norm": 0.014232151210308075,
+ "learning_rate": 0.0006,
+ "loss": 5.928424835205078,
+ "step": 421
+ },
+ {
+ "epoch": 5.866375545851528,
+ "grad_norm": 0.01570476032793522,
+ "learning_rate": 0.0006,
+ "loss": 5.953692436218262,
+ "step": 422
+ },
+ {
+ "epoch": 5.880349344978166,
+ "grad_norm": 0.01586179807782173,
+ "learning_rate": 0.0006,
+ "loss": 5.820014476776123,
+ "step": 423
+ },
+ {
+ "epoch": 5.894323144104804,
+ "grad_norm": 0.014729145914316177,
+ "learning_rate": 0.0006,
+ "loss": 5.860154628753662,
+ "step": 424
+ },
+ {
+ "epoch": 5.908296943231441,
+ "grad_norm": 0.013111459091305733,
+ "learning_rate": 0.0006,
+ "loss": 5.85988187789917,
+ "step": 425
+ },
+ {
+ "epoch": 5.922270742358078,
+ "grad_norm": 0.012937922962009907,
+ "learning_rate": 0.0006,
+ "loss": 5.756265640258789,
+ "step": 426
+ },
+ {
+ "epoch": 5.936244541484716,
+ "grad_norm": 0.01377453189343214,
+ "learning_rate": 0.0006,
+ "loss": 5.8333048820495605,
+ "step": 427
+ },
+ {
+ "epoch": 5.950218340611354,
+ "grad_norm": 0.014054795727133751,
+ "learning_rate": 0.0006,
+ "loss": 5.803333759307861,
+ "step": 428
+ },
+ {
+ "epoch": 5.964192139737992,
+ "grad_norm": 0.01674959622323513,
+ "learning_rate": 0.0006,
+ "loss": 5.8718485832214355,
+ "step": 429
+ },
+ {
+ "epoch": 5.978165938864628,
+ "grad_norm": 0.018733017146587372,
+ "learning_rate": 0.0006,
+ "loss": 5.875979900360107,
+ "step": 430
+ },
+ {
+ "epoch": 5.992139737991266,
+ "grad_norm": 0.02088983915746212,
+ "learning_rate": 0.0006,
+ "loss": 5.763634204864502,
+ "step": 431
+ },
+ {
+ "epoch": 6.0,
+ "grad_norm": 0.026705985888838768,
+ "learning_rate": 0.0006,
+ "loss": 5.6922478675842285,
+ "step": 432
+ },
+ {
+ "epoch": 6.0,
+ "eval_loss": 5.847958087921143,
+ "eval_runtime": 58.5451,
+ "eval_samples_per_second": 41.711,
+ "eval_steps_per_second": 1.315,
+ "step": 432
+ },
+ {
+ "epoch": 6.013973799126638,
+ "grad_norm": 0.03181544691324234,
+ "learning_rate": 0.0006,
+ "loss": 5.786368370056152,
+ "step": 433
+ },
+ {
+ "epoch": 6.0279475982532755,
+ "grad_norm": 0.03238112851977348,
+ "learning_rate": 0.0006,
+ "loss": 5.810310363769531,
+ "step": 434
+ },
+ {
+ "epoch": 6.041921397379912,
+ "grad_norm": 0.03916756808757782,
+ "learning_rate": 0.0006,
+ "loss": 5.7761993408203125,
+ "step": 435
+ },
+ {
+ "epoch": 6.05589519650655,
+ "grad_norm": 0.044009730219841,
+ "learning_rate": 0.0006,
+ "loss": 5.898112773895264,
+ "step": 436
+ },
+ {
+ "epoch": 6.069868995633188,
+ "grad_norm": 0.045235246419906616,
+ "learning_rate": 0.0006,
+ "loss": 5.876371383666992,
+ "step": 437
+ },
+ {
+ "epoch": 6.083842794759826,
+ "grad_norm": 0.050218384712934494,
+ "learning_rate": 0.0006,
+ "loss": 5.805103302001953,
+ "step": 438
+ },
+ {
+ "epoch": 6.097816593886463,
+ "grad_norm": 0.05444490164518356,
+ "learning_rate": 0.0006,
+ "loss": 5.826424598693848,
+ "step": 439
+ },
+ {
+ "epoch": 6.1117903930131,
+ "grad_norm": 0.04890982061624527,
+ "learning_rate": 0.0006,
+ "loss": 5.7735395431518555,
+ "step": 440
+ },
+ {
+ "epoch": 6.125764192139738,
+ "grad_norm": 0.05478639155626297,
+ "learning_rate": 0.0006,
+ "loss": 5.823677062988281,
+ "step": 441
+ },
+ {
+ "epoch": 6.139737991266376,
+ "grad_norm": 0.057562313973903656,
+ "learning_rate": 0.0006,
+ "loss": 5.889334678649902,
+ "step": 442
+ },
+ {
+ "epoch": 6.153711790393013,
+ "grad_norm": 0.06447703391313553,
+ "learning_rate": 0.0006,
+ "loss": 5.84621524810791,
+ "step": 443
+ },
+ {
+ "epoch": 6.1676855895196505,
+ "grad_norm": 0.048861872404813766,
+ "learning_rate": 0.0006,
+ "loss": 5.931595802307129,
+ "step": 444
+ },
+ {
+ "epoch": 6.181659388646288,
+ "grad_norm": 0.05521339178085327,
+ "learning_rate": 0.0006,
+ "loss": 5.8229780197143555,
+ "step": 445
+ },
+ {
+ "epoch": 6.195633187772926,
+ "grad_norm": 0.053666990250349045,
+ "learning_rate": 0.0006,
+ "loss": 5.879191875457764,
+ "step": 446
+ },
+ {
+ "epoch": 6.209606986899563,
+ "grad_norm": 0.0451025515794754,
+ "learning_rate": 0.0006,
+ "loss": 5.801628112792969,
+ "step": 447
+ },
+ {
+ "epoch": 6.223580786026201,
+ "grad_norm": 0.04291655868291855,
+ "learning_rate": 0.0006,
+ "loss": 5.914680480957031,
+ "step": 448
+ },
+ {
+ "epoch": 6.2375545851528384,
+ "grad_norm": 0.035102490335702896,
+ "learning_rate": 0.0006,
+ "loss": 5.843682765960693,
+ "step": 449
+ },
+ {
+ "epoch": 6.251528384279476,
+ "grad_norm": 0.03403995931148529,
+ "learning_rate": 0.0006,
+ "loss": 5.827154159545898,
+ "step": 450
+ },
+ {
+ "epoch": 6.265502183406113,
+ "grad_norm": 0.03444375470280647,
+ "learning_rate": 0.0006,
+ "loss": 5.800136566162109,
+ "step": 451
+ },
+ {
+ "epoch": 6.279475982532751,
+ "grad_norm": 0.03165159001946449,
+ "learning_rate": 0.0006,
+ "loss": 5.906252384185791,
+ "step": 452
+ },
+ {
+ "epoch": 6.293449781659389,
+ "grad_norm": 0.026153093203902245,
+ "learning_rate": 0.0006,
+ "loss": 5.7423200607299805,
+ "step": 453
+ },
+ {
+ "epoch": 6.307423580786026,
+ "grad_norm": 0.024357657879590988,
+ "learning_rate": 0.0006,
+ "loss": 5.701364517211914,
+ "step": 454
+ },
+ {
+ "epoch": 6.321397379912664,
+ "grad_norm": 0.021508049219846725,
+ "learning_rate": 0.0006,
+ "loss": 5.8328094482421875,
+ "step": 455
+ },
+ {
+ "epoch": 6.335371179039301,
+ "grad_norm": 0.017313921824097633,
+ "learning_rate": 0.0006,
+ "loss": 5.792145252227783,
+ "step": 456
+ },
+ {
+ "epoch": 6.349344978165939,
+ "grad_norm": 0.018563397228717804,
+ "learning_rate": 0.0006,
+ "loss": 5.732672214508057,
+ "step": 457
+ },
+ {
+ "epoch": 6.3633187772925766,
+ "grad_norm": 0.016568679362535477,
+ "learning_rate": 0.0006,
+ "loss": 5.81948184967041,
+ "step": 458
+ },
+ {
+ "epoch": 6.377292576419214,
+ "grad_norm": 0.014133770950138569,
+ "learning_rate": 0.0006,
+ "loss": 5.734982490539551,
+ "step": 459
+ },
+ {
+ "epoch": 6.391266375545851,
+ "grad_norm": 0.014114447869360447,
+ "learning_rate": 0.0006,
+ "loss": 5.70073127746582,
+ "step": 460
+ },
+ {
+ "epoch": 6.405240174672489,
+ "grad_norm": 0.012907393276691437,
+ "learning_rate": 0.0006,
+ "loss": 5.758626937866211,
+ "step": 461
+ },
+ {
+ "epoch": 6.419213973799127,
+ "grad_norm": 0.012673444114625454,
+ "learning_rate": 0.0006,
+ "loss": 5.775138854980469,
+ "step": 462
+ },
+ {
+ "epoch": 6.4331877729257645,
+ "grad_norm": 0.01311410591006279,
+ "learning_rate": 0.0006,
+ "loss": 5.70920991897583,
+ "step": 463
+ },
+ {
+ "epoch": 6.447161572052401,
+ "grad_norm": 0.012935544364154339,
+ "learning_rate": 0.0006,
+ "loss": 5.781479835510254,
+ "step": 464
+ },
+ {
+ "epoch": 6.461135371179039,
+ "grad_norm": 0.01439738366752863,
+ "learning_rate": 0.0006,
+ "loss": 5.691315650939941,
+ "step": 465
+ },
+ {
+ "epoch": 6.475109170305677,
+ "grad_norm": 0.01205186452716589,
+ "learning_rate": 0.0006,
+ "loss": 5.662499904632568,
+ "step": 466
+ },
+ {
+ "epoch": 6.489082969432315,
+ "grad_norm": 0.011575652286410332,
+ "learning_rate": 0.0006,
+ "loss": 5.6271772384643555,
+ "step": 467
+ },
+ {
+ "epoch": 6.503056768558952,
+ "grad_norm": 0.011672502383589745,
+ "learning_rate": 0.0006,
+ "loss": 5.761662006378174,
+ "step": 468
+ },
+ {
+ "epoch": 6.517030567685589,
+ "grad_norm": 0.011541751213371754,
+ "learning_rate": 0.0006,
+ "loss": 5.762078285217285,
+ "step": 469
+ },
+ {
+ "epoch": 6.531004366812227,
+ "grad_norm": 0.011396365240216255,
+ "learning_rate": 0.0006,
+ "loss": 5.67873477935791,
+ "step": 470
+ },
+ {
+ "epoch": 6.544978165938865,
+ "grad_norm": 0.010688499547541142,
+ "learning_rate": 0.0006,
+ "loss": 5.717051982879639,
+ "step": 471
+ },
+ {
+ "epoch": 6.558951965065502,
+ "grad_norm": 0.012224317528307438,
+ "learning_rate": 0.0006,
+ "loss": 5.707948684692383,
+ "step": 472
+ },
+ {
+ "epoch": 6.5729257641921395,
+ "grad_norm": 0.011856825090944767,
+ "learning_rate": 0.0006,
+ "loss": 5.70878791809082,
+ "step": 473
+ },
+ {
+ "epoch": 6.586899563318777,
+ "grad_norm": 0.01199064590036869,
+ "learning_rate": 0.0006,
+ "loss": 5.708697319030762,
+ "step": 474
+ },
+ {
+ "epoch": 6.600873362445415,
+ "grad_norm": 0.01219446212053299,
+ "learning_rate": 0.0006,
+ "loss": 5.6061577796936035,
+ "step": 475
+ },
+ {
+ "epoch": 6.614847161572053,
+ "grad_norm": 0.013397484086453915,
+ "learning_rate": 0.0006,
+ "loss": 5.624789714813232,
+ "step": 476
+ },
+ {
+ "epoch": 6.62882096069869,
+ "grad_norm": 0.01483136136084795,
+ "learning_rate": 0.0006,
+ "loss": 5.721141338348389,
+ "step": 477
+ },
+ {
+ "epoch": 6.642794759825327,
+ "grad_norm": 0.0194488987326622,
+ "learning_rate": 0.0006,
+ "loss": 5.604279518127441,
+ "step": 478
+ },
+ {
+ "epoch": 6.656768558951965,
+ "grad_norm": 0.02079896256327629,
+ "learning_rate": 0.0006,
+ "loss": 5.586322784423828,
+ "step": 479
+ },
+ {
+ "epoch": 6.670742358078603,
+ "grad_norm": 0.018996229395270348,
+ "learning_rate": 0.0006,
+ "loss": 5.5987772941589355,
+ "step": 480
+ },
+ {
+ "epoch": 6.68471615720524,
+ "grad_norm": 0.015779603272676468,
+ "learning_rate": 0.0006,
+ "loss": 5.588602066040039,
+ "step": 481
+ },
+ {
+ "epoch": 6.698689956331878,
+ "grad_norm": 0.015322973020374775,
+ "learning_rate": 0.0006,
+ "loss": 5.686467170715332,
+ "step": 482
+ },
+ {
+ "epoch": 6.712663755458515,
+ "grad_norm": 0.015282213687896729,
+ "learning_rate": 0.0006,
+ "loss": 5.664676666259766,
+ "step": 483
+ },
+ {
+ "epoch": 6.726637554585153,
+ "grad_norm": 0.016589272767305374,
+ "learning_rate": 0.0006,
+ "loss": 5.666739463806152,
+ "step": 484
+ },
+ {
+ "epoch": 6.74061135371179,
+ "grad_norm": 0.01811421848833561,
+ "learning_rate": 0.0006,
+ "loss": 5.647593021392822,
+ "step": 485
+ },
+ {
+ "epoch": 6.754585152838428,
+ "grad_norm": 0.019652029499411583,
+ "learning_rate": 0.0006,
+ "loss": 5.628726959228516,
+ "step": 486
+ },
+ {
+ "epoch": 6.7685589519650655,
+ "grad_norm": 0.02118665538728237,
+ "learning_rate": 0.0006,
+ "loss": 5.6580705642700195,
+ "step": 487
+ },
+ {
+ "epoch": 6.782532751091703,
+ "grad_norm": 0.02237832546234131,
+ "learning_rate": 0.0006,
+ "loss": 5.635931968688965,
+ "step": 488
+ },
+ {
+ "epoch": 6.796506550218341,
+ "grad_norm": 0.023897631093859673,
+ "learning_rate": 0.0006,
+ "loss": 5.517949104309082,
+ "step": 489
+ },
+ {
+ "epoch": 6.810480349344978,
+ "grad_norm": 0.02442227490246296,
+ "learning_rate": 0.0006,
+ "loss": 5.600021839141846,
+ "step": 490
+ },
+ {
+ "epoch": 6.824454148471616,
+ "grad_norm": 0.024675287306308746,
+ "learning_rate": 0.0006,
+ "loss": 5.6639909744262695,
+ "step": 491
+ },
+ {
+ "epoch": 6.8384279475982535,
+ "grad_norm": 0.030372392386198044,
+ "learning_rate": 0.0006,
+ "loss": 5.582888603210449,
+ "step": 492
+ },
+ {
+ "epoch": 6.85240174672489,
+ "grad_norm": 0.03237690031528473,
+ "learning_rate": 0.0006,
+ "loss": 5.573896884918213,
+ "step": 493
+ },
+ {
+ "epoch": 6.866375545851528,
+ "grad_norm": 0.03511589393019676,
+ "learning_rate": 0.0006,
+ "loss": 5.565878868103027,
+ "step": 494
+ },
+ {
+ "epoch": 6.880349344978166,
+ "grad_norm": 0.03883938118815422,
+ "learning_rate": 0.0006,
+ "loss": 5.58540678024292,
+ "step": 495
+ },
+ {
+ "epoch": 6.894323144104804,
+ "grad_norm": 0.04175502806901932,
+ "learning_rate": 0.0006,
+ "loss": 5.6095194816589355,
+ "step": 496
+ },
+ {
+ "epoch": 6.908296943231441,
+ "grad_norm": 0.04013441503047943,
+ "learning_rate": 0.0006,
+ "loss": 5.654401779174805,
+ "step": 497
+ },
+ {
+ "epoch": 6.922270742358078,
+ "grad_norm": 0.04618770629167557,
+ "learning_rate": 0.0006,
+ "loss": 5.6483473777771,
+ "step": 498
+ },
+ {
+ "epoch": 6.936244541484716,
+ "grad_norm": 0.04507308453321457,
+ "learning_rate": 0.0006,
+ "loss": 5.5856218338012695,
+ "step": 499
+ },
+ {
+ "epoch": 6.950218340611354,
+ "grad_norm": 0.03151383996009827,
+ "learning_rate": 0.0006,
+ "loss": 5.594086647033691,
+ "step": 500
+ },
+ {
+ "epoch": 6.964192139737992,
+ "grad_norm": 0.026883797720074654,
+ "learning_rate": 0.0006,
+ "loss": 5.607676029205322,
+ "step": 501
+ },
+ {
+ "epoch": 6.978165938864628,
+ "grad_norm": 0.02981836162507534,
+ "learning_rate": 0.0006,
+ "loss": 5.563666820526123,
+ "step": 502
+ },
+ {
+ "epoch": 6.992139737991266,
+ "grad_norm": 0.03233477845788002,
+ "learning_rate": 0.0006,
+ "loss": 5.667543411254883,
+ "step": 503
+ },
+ {
+ "epoch": 7.0,
+ "grad_norm": 0.029927456751465797,
+ "learning_rate": 0.0006,
+ "loss": 5.733482837677002,
+ "step": 504
+ },
+ {
+ "epoch": 7.0,
+ "eval_loss": 5.679076671600342,
+ "eval_runtime": 59.5965,
+ "eval_samples_per_second": 40.976,
+ "eval_steps_per_second": 1.292,
+ "step": 504
+ },
+ {
+ "epoch": 7.013973799126638,
+ "grad_norm": 0.027081597596406937,
+ "learning_rate": 0.0006,
+ "loss": 5.634487152099609,
+ "step": 505
+ },
+ {
+ "epoch": 7.0279475982532755,
+ "grad_norm": 0.031959421932697296,
+ "learning_rate": 0.0006,
+ "loss": 5.643294811248779,
+ "step": 506
+ },
+ {
+ "epoch": 7.041921397379912,
+ "grad_norm": 0.02802063524723053,
+ "learning_rate": 0.0006,
+ "loss": 5.573239326477051,
+ "step": 507
+ },
+ {
+ "epoch": 7.05589519650655,
+ "grad_norm": 0.032579705119132996,
+ "learning_rate": 0.0006,
+ "loss": 5.627150058746338,
+ "step": 508
+ },
+ {
+ "epoch": 7.069868995633188,
+ "grad_norm": 0.02503928542137146,
+ "learning_rate": 0.0006,
+ "loss": 5.6045308113098145,
+ "step": 509
+ },
+ {
+ "epoch": 7.083842794759826,
+ "grad_norm": 0.02421317622065544,
+ "learning_rate": 0.0006,
+ "loss": 5.522153377532959,
+ "step": 510
+ },
+ {
+ "epoch": 7.097816593886463,
+ "grad_norm": 0.021983426064252853,
+ "learning_rate": 0.0006,
+ "loss": 5.632939338684082,
+ "step": 511
+ },
+ {
+ "epoch": 7.1117903930131,
+ "grad_norm": 0.021933183073997498,
+ "learning_rate": 0.0006,
+ "loss": 5.592185974121094,
+ "step": 512
+ },
+ {
+ "epoch": 7.125764192139738,
+ "grad_norm": 0.02287031151354313,
+ "learning_rate": 0.0006,
+ "loss": 5.613700866699219,
+ "step": 513
+ },
+ {
+ "epoch": 7.139737991266376,
+ "grad_norm": 0.02281602844595909,
+ "learning_rate": 0.0006,
+ "loss": 5.551383018493652,
+ "step": 514
+ },
+ {
+ "epoch": 7.153711790393013,
+ "grad_norm": 0.021050360053777695,
+ "learning_rate": 0.0006,
+ "loss": 5.55275821685791,
+ "step": 515
+ },
+ {
+ "epoch": 7.1676855895196505,
+ "grad_norm": 0.018299926072359085,
+ "learning_rate": 0.0006,
+ "loss": 5.528225421905518,
+ "step": 516
+ },
+ {
+ "epoch": 7.181659388646288,
+ "grad_norm": 0.02024853602051735,
+ "learning_rate": 0.0006,
+ "loss": 5.522453784942627,
+ "step": 517
+ },
+ {
+ "epoch": 7.195633187772926,
+ "grad_norm": 0.02135239914059639,
+ "learning_rate": 0.0006,
+ "loss": 5.527522563934326,
+ "step": 518
+ },
+ {
+ "epoch": 7.209606986899563,
+ "grad_norm": 0.023394184187054634,
+ "learning_rate": 0.0006,
+ "loss": 5.4535017013549805,
+ "step": 519
+ },
+ {
+ "epoch": 7.223580786026201,
+ "grad_norm": 0.025952080264687538,
+ "learning_rate": 0.0006,
+ "loss": 5.5374674797058105,
+ "step": 520
+ },
+ {
+ "epoch": 7.2375545851528384,
+ "grad_norm": 0.022287718951702118,
+ "learning_rate": 0.0006,
+ "loss": 5.493753433227539,
+ "step": 521
+ },
+ {
+ "epoch": 7.251528384279476,
+ "grad_norm": 0.02018447406589985,
+ "learning_rate": 0.0006,
+ "loss": 5.510575294494629,
+ "step": 522
+ },
+ {
+ "epoch": 7.265502183406113,
+ "grad_norm": 0.02050507813692093,
+ "learning_rate": 0.0006,
+ "loss": 5.466026782989502,
+ "step": 523
+ },
+ {
+ "epoch": 7.279475982532751,
+ "grad_norm": 0.023688802495598793,
+ "learning_rate": 0.0006,
+ "loss": 5.485815525054932,
+ "step": 524
+ },
+ {
+ "epoch": 7.293449781659389,
+ "grad_norm": 0.021078843623399734,
+ "learning_rate": 0.0006,
+ "loss": 5.415581703186035,
+ "step": 525
+ },
+ {
+ "epoch": 7.307423580786026,
+ "grad_norm": 0.018266011029481888,
+ "learning_rate": 0.0006,
+ "loss": 5.402815818786621,
+ "step": 526
+ },
+ {
+ "epoch": 7.321397379912664,
+ "grad_norm": 0.019693493843078613,
+ "learning_rate": 0.0006,
+ "loss": 5.504674911499023,
+ "step": 527
+ },
+ {
+ "epoch": 7.335371179039301,
+ "grad_norm": 0.02031373605132103,
+ "learning_rate": 0.0006,
+ "loss": 5.503837585449219,
+ "step": 528
+ },
+ {
+ "epoch": 7.349344978165939,
+ "grad_norm": 0.017636626958847046,
+ "learning_rate": 0.0006,
+ "loss": 5.530580997467041,
+ "step": 529
+ },
+ {
+ "epoch": 7.3633187772925766,
+ "grad_norm": 0.01787244901061058,
+ "learning_rate": 0.0006,
+ "loss": 5.450218677520752,
+ "step": 530
+ },
+ {
+ "epoch": 7.377292576419214,
+ "grad_norm": 0.0170669574290514,
+ "learning_rate": 0.0006,
+ "loss": 5.481570243835449,
+ "step": 531
+ },
+ {
+ "epoch": 7.391266375545851,
+ "grad_norm": 0.01802165061235428,
+ "learning_rate": 0.0006,
+ "loss": 5.386394500732422,
+ "step": 532
+ },
+ {
+ "epoch": 7.405240174672489,
+ "grad_norm": 0.022949472069740295,
+ "learning_rate": 0.0006,
+ "loss": 5.424929618835449,
+ "step": 533
+ },
+ {
+ "epoch": 7.419213973799127,
+ "grad_norm": 0.034213390201330185,
+ "learning_rate": 0.0006,
+ "loss": 5.346663475036621,
+ "step": 534
+ },
+ {
+ "epoch": 7.4331877729257645,
+ "grad_norm": 0.03847593814134598,
+ "learning_rate": 0.0006,
+ "loss": 5.418482780456543,
+ "step": 535
+ },
+ {
+ "epoch": 7.447161572052401,
+ "grad_norm": 0.029393676668405533,
+ "learning_rate": 0.0006,
+ "loss": 5.456090927124023,
+ "step": 536
+ },
+ {
+ "epoch": 7.461135371179039,
+ "grad_norm": 0.03388667106628418,
+ "learning_rate": 0.0006,
+ "loss": 5.572983741760254,
+ "step": 537
+ },
+ {
+ "epoch": 7.475109170305677,
+ "grad_norm": 0.039690613746643066,
+ "learning_rate": 0.0006,
+ "loss": 5.414067268371582,
+ "step": 538
+ },
+ {
+ "epoch": 7.489082969432315,
+ "grad_norm": 0.03630776330828667,
+ "learning_rate": 0.0006,
+ "loss": 5.523739814758301,
+ "step": 539
+ },
+ {
+ "epoch": 7.503056768558952,
+ "grad_norm": 0.03356870263814926,
+ "learning_rate": 0.0006,
+ "loss": 5.444767951965332,
+ "step": 540
+ },
+ {
+ "epoch": 7.517030567685589,
+ "grad_norm": 0.030812304466962814,
+ "learning_rate": 0.0006,
+ "loss": 5.461244106292725,
+ "step": 541
+ },
+ {
+ "epoch": 7.531004366812227,
+ "grad_norm": 0.03319217637181282,
+ "learning_rate": 0.0006,
+ "loss": 5.502161026000977,
+ "step": 542
+ },
+ {
+ "epoch": 7.544978165938865,
+ "grad_norm": 0.032997481524944305,
+ "learning_rate": 0.0006,
+ "loss": 5.4162278175354,
+ "step": 543
+ },
+ {
+ "epoch": 7.558951965065502,
+ "grad_norm": 0.03364962339401245,
+ "learning_rate": 0.0006,
+ "loss": 5.415736198425293,
+ "step": 544
+ },
+ {
+ "epoch": 7.5729257641921395,
+ "grad_norm": 0.0344221405684948,
+ "learning_rate": 0.0006,
+ "loss": 5.541967868804932,
+ "step": 545
+ },
+ {
+ "epoch": 7.586899563318777,
+ "grad_norm": 0.029609503224492073,
+ "learning_rate": 0.0006,
+ "loss": 5.440614223480225,
+ "step": 546
+ },
+ {
+ "epoch": 7.600873362445415,
+ "grad_norm": 0.029055926948785782,
+ "learning_rate": 0.0006,
+ "loss": 5.462865352630615,
+ "step": 547
+ },
+ {
+ "epoch": 7.614847161572053,
+ "grad_norm": 0.02658848837018013,
+ "learning_rate": 0.0006,
+ "loss": 5.444467544555664,
+ "step": 548
+ },
+ {
+ "epoch": 7.62882096069869,
+ "grad_norm": 0.026276404038071632,
+ "learning_rate": 0.0006,
+ "loss": 5.462278842926025,
+ "step": 549
+ },
+ {
+ "epoch": 7.642794759825327,
+ "grad_norm": 0.0286889486014843,
+ "learning_rate": 0.0006,
+ "loss": 5.4849534034729,
+ "step": 550
+ },
+ {
+ "epoch": 7.656768558951965,
+ "grad_norm": 0.026849817484617233,
+ "learning_rate": 0.0006,
+ "loss": 5.431473731994629,
+ "step": 551
+ },
+ {
+ "epoch": 7.670742358078603,
+ "grad_norm": 0.02312396466732025,
+ "learning_rate": 0.0006,
+ "loss": 5.45462703704834,
+ "step": 552
+ },
+ {
+ "epoch": 7.68471615720524,
+ "grad_norm": 0.026197485625743866,
+ "learning_rate": 0.0006,
+ "loss": 5.452552318572998,
+ "step": 553
+ },
+ {
+ "epoch": 7.698689956331878,
+ "grad_norm": 0.02747255191206932,
+ "learning_rate": 0.0006,
+ "loss": 5.447169780731201,
+ "step": 554
+ },
+ {
+ "epoch": 7.712663755458515,
+ "grad_norm": 0.028123432770371437,
+ "learning_rate": 0.0006,
+ "loss": 5.394400596618652,
+ "step": 555
+ },
+ {
+ "epoch": 7.726637554585153,
+ "grad_norm": 0.02599870041012764,
+ "learning_rate": 0.0006,
+ "loss": 5.398341178894043,
+ "step": 556
+ },
+ {
+ "epoch": 7.74061135371179,
+ "grad_norm": 0.022171657532453537,
+ "learning_rate": 0.0006,
+ "loss": 5.368820667266846,
+ "step": 557
+ },
+ {
+ "epoch": 7.754585152838428,
+ "grad_norm": 0.022309480234980583,
+ "learning_rate": 0.0006,
+ "loss": 5.416772365570068,
+ "step": 558
+ },
+ {
+ "epoch": 7.7685589519650655,
+ "grad_norm": 0.024074165150523186,
+ "learning_rate": 0.0006,
+ "loss": 5.359002113342285,
+ "step": 559
+ },
+ {
+ "epoch": 7.782532751091703,
+ "grad_norm": 0.02636653557419777,
+ "learning_rate": 0.0006,
+ "loss": 5.484368324279785,
+ "step": 560
+ },
+ {
+ "epoch": 7.796506550218341,
+ "grad_norm": 0.02203752100467682,
+ "learning_rate": 0.0006,
+ "loss": 5.375498294830322,
+ "step": 561
+ },
+ {
+ "epoch": 7.810480349344978,
+ "grad_norm": 0.019924819469451904,
+ "learning_rate": 0.0006,
+ "loss": 5.334723472595215,
+ "step": 562
+ },
+ {
+ "epoch": 7.824454148471616,
+ "grad_norm": 0.018755966797471046,
+ "learning_rate": 0.0006,
+ "loss": 5.348781108856201,
+ "step": 563
+ },
+ {
+ "epoch": 7.8384279475982535,
+ "grad_norm": 0.016747845336794853,
+ "learning_rate": 0.0006,
+ "loss": 5.433079719543457,
+ "step": 564
+ },
+ {
+ "epoch": 7.85240174672489,
+ "grad_norm": 0.0170395877212286,
+ "learning_rate": 0.0006,
+ "loss": 5.346794128417969,
+ "step": 565
+ },
+ {
+ "epoch": 7.866375545851528,
+ "grad_norm": 0.016074176877737045,
+ "learning_rate": 0.0006,
+ "loss": 5.3277740478515625,
+ "step": 566
+ },
+ {
+ "epoch": 7.880349344978166,
+ "grad_norm": 0.014119806699454784,
+ "learning_rate": 0.0006,
+ "loss": 5.483551979064941,
+ "step": 567
+ },
+ {
+ "epoch": 7.894323144104804,
+ "grad_norm": 0.015270394273102283,
+ "learning_rate": 0.0006,
+ "loss": 5.378519535064697,
+ "step": 568
+ },
+ {
+ "epoch": 7.908296943231441,
+ "grad_norm": 0.015677539631724358,
+ "learning_rate": 0.0006,
+ "loss": 5.290210247039795,
+ "step": 569
+ },
+ {
+ "epoch": 7.922270742358078,
+ "grad_norm": 0.015930423513054848,
+ "learning_rate": 0.0006,
+ "loss": 5.427360534667969,
+ "step": 570
+ },
+ {
+ "epoch": 7.936244541484716,
+ "grad_norm": 0.016641564667224884,
+ "learning_rate": 0.0006,
+ "loss": 5.301599979400635,
+ "step": 571
+ },
+ {
+ "epoch": 7.950218340611354,
+ "grad_norm": 0.018293552100658417,
+ "learning_rate": 0.0006,
+ "loss": 5.277889251708984,
+ "step": 572
+ },
+ {
+ "epoch": 7.964192139737992,
+ "grad_norm": 0.017618799582123756,
+ "learning_rate": 0.0006,
+ "loss": 5.297194480895996,
+ "step": 573
+ },
+ {
+ "epoch": 7.978165938864628,
+ "grad_norm": 0.015034242533147335,
+ "learning_rate": 0.0006,
+ "loss": 5.42428731918335,
+ "step": 574
+ },
+ {
+ "epoch": 7.992139737991266,
+ "grad_norm": 0.01617511734366417,
+ "learning_rate": 0.0006,
+ "loss": 5.420044898986816,
+ "step": 575
+ },
+ {
+ "epoch": 8.0,
+ "grad_norm": 0.01837257295846939,
+ "learning_rate": 0.0006,
+ "loss": 5.240387916564941,
+ "step": 576
+ },
+ {
+ "epoch": 8.0,
+ "eval_loss": 5.423932075500488,
+ "eval_runtime": 58.0841,
+ "eval_samples_per_second": 42.042,
+ "eval_steps_per_second": 1.326,
+ "step": 576
+ },
+ {
+ "epoch": 8.013973799126637,
+ "grad_norm": 0.020953809842467308,
+ "learning_rate": 0.0006,
+ "loss": 5.4320783615112305,
+ "step": 577
+ },
+ {
+ "epoch": 8.027947598253276,
+ "grad_norm": 0.0227745920419693,
+ "learning_rate": 0.0006,
+ "loss": 5.310197830200195,
+ "step": 578
+ },
+ {
+ "epoch": 8.041921397379912,
+ "grad_norm": 0.021372603252530098,
+ "learning_rate": 0.0006,
+ "loss": 5.339468002319336,
+ "step": 579
+ },
+ {
+ "epoch": 8.055895196506551,
+ "grad_norm": 0.022401731461286545,
+ "learning_rate": 0.0006,
+ "loss": 5.279829978942871,
+ "step": 580
+ },
+ {
+ "epoch": 8.069868995633188,
+ "grad_norm": 0.023942379280924797,
+ "learning_rate": 0.0006,
+ "loss": 5.372323989868164,
+ "step": 581
+ },
+ {
+ "epoch": 8.083842794759825,
+ "grad_norm": 0.024380534887313843,
+ "learning_rate": 0.0006,
+ "loss": 5.253486633300781,
+ "step": 582
+ },
+ {
+ "epoch": 8.097816593886463,
+ "grad_norm": 0.03246625140309334,
+ "learning_rate": 0.0006,
+ "loss": 5.305774211883545,
+ "step": 583
+ },
+ {
+ "epoch": 8.1117903930131,
+ "grad_norm": 0.030399736016988754,
+ "learning_rate": 0.0006,
+ "loss": 5.370976448059082,
+ "step": 584
+ },
+ {
+ "epoch": 8.125764192139737,
+ "grad_norm": 0.027148913592100143,
+ "learning_rate": 0.0006,
+ "loss": 5.368010520935059,
+ "step": 585
+ },
+ {
+ "epoch": 8.139737991266376,
+ "grad_norm": 0.02925540879368782,
+ "learning_rate": 0.0006,
+ "loss": 5.303013801574707,
+ "step": 586
+ },
+ {
+ "epoch": 8.153711790393013,
+ "grad_norm": 0.02915453538298607,
+ "learning_rate": 0.0006,
+ "loss": 5.329667091369629,
+ "step": 587
+ },
+ {
+ "epoch": 8.167685589519651,
+ "grad_norm": 0.031732626259326935,
+ "learning_rate": 0.0006,
+ "loss": 5.3634934425354,
+ "step": 588
+ },
+ {
+ "epoch": 8.181659388646288,
+ "grad_norm": 0.036166731268167496,
+ "learning_rate": 0.0006,
+ "loss": 5.223980903625488,
+ "step": 589
+ },
+ {
+ "epoch": 8.195633187772925,
+ "grad_norm": 0.035654108971357346,
+ "learning_rate": 0.0006,
+ "loss": 5.361059188842773,
+ "step": 590
+ },
+ {
+ "epoch": 8.209606986899564,
+ "grad_norm": 0.0365324430167675,
+ "learning_rate": 0.0006,
+ "loss": 5.336151599884033,
+ "step": 591
+ },
+ {
+ "epoch": 8.2235807860262,
+ "grad_norm": 0.03578880429267883,
+ "learning_rate": 0.0006,
+ "loss": 5.434076309204102,
+ "step": 592
+ },
+ {
+ "epoch": 8.237554585152838,
+ "grad_norm": 0.035734012722969055,
+ "learning_rate": 0.0006,
+ "loss": 5.330893516540527,
+ "step": 593
+ },
+ {
+ "epoch": 8.251528384279476,
+ "grad_norm": 0.03196857124567032,
+ "learning_rate": 0.0006,
+ "loss": 5.306009769439697,
+ "step": 594
+ },
+ {
+ "epoch": 8.265502183406113,
+ "grad_norm": 0.032042086124420166,
+ "learning_rate": 0.0006,
+ "loss": 5.343267917633057,
+ "step": 595
+ },
+ {
+ "epoch": 8.279475982532752,
+ "grad_norm": 0.03160746395587921,
+ "learning_rate": 0.0006,
+ "loss": 5.2353715896606445,
+ "step": 596
+ },
+ {
+ "epoch": 8.293449781659389,
+ "grad_norm": 0.026689818128943443,
+ "learning_rate": 0.0006,
+ "loss": 5.254042625427246,
+ "step": 597
+ },
+ {
+ "epoch": 8.307423580786025,
+ "grad_norm": 0.02880188450217247,
+ "learning_rate": 0.0006,
+ "loss": 5.326833248138428,
+ "step": 598
+ },
+ {
+ "epoch": 8.321397379912664,
+ "grad_norm": 0.027516381815075874,
+ "learning_rate": 0.0006,
+ "loss": 5.342336177825928,
+ "step": 599
+ },
+ {
+ "epoch": 8.335371179039301,
+ "grad_norm": 0.027875030413269997,
+ "learning_rate": 0.0006,
+ "loss": 5.200719833374023,
+ "step": 600
+ },
+ {
+ "epoch": 8.34934497816594,
+ "grad_norm": 0.0268265288323164,
+ "learning_rate": 0.0006,
+ "loss": 5.249449729919434,
+ "step": 601
+ },
+ {
+ "epoch": 8.363318777292577,
+ "grad_norm": 0.024824608117341995,
+ "learning_rate": 0.0006,
+ "loss": 5.318952560424805,
+ "step": 602
+ },
+ {
+ "epoch": 8.377292576419213,
+ "grad_norm": 0.022990627214312553,
+ "learning_rate": 0.0006,
+ "loss": 5.244993209838867,
+ "step": 603
+ },
+ {
+ "epoch": 8.391266375545852,
+ "grad_norm": 0.022804604843258858,
+ "learning_rate": 0.0006,
+ "loss": 5.3236846923828125,
+ "step": 604
+ },
+ {
+ "epoch": 8.405240174672489,
+ "grad_norm": 0.02419872209429741,
+ "learning_rate": 0.0006,
+ "loss": 5.255486488342285,
+ "step": 605
+ },
+ {
+ "epoch": 8.419213973799126,
+ "grad_norm": 0.021952059119939804,
+ "learning_rate": 0.0006,
+ "loss": 5.3511552810668945,
+ "step": 606
+ },
+ {
+ "epoch": 8.433187772925764,
+ "grad_norm": 0.022375497967004776,
+ "learning_rate": 0.0006,
+ "loss": 5.294790267944336,
+ "step": 607
+ },
+ {
+ "epoch": 8.447161572052401,
+ "grad_norm": 0.019242815673351288,
+ "learning_rate": 0.0006,
+ "loss": 5.252618789672852,
+ "step": 608
+ },
+ {
+ "epoch": 8.46113537117904,
+ "grad_norm": 0.0173486340790987,
+ "learning_rate": 0.0006,
+ "loss": 5.30147647857666,
+ "step": 609
+ },
+ {
+ "epoch": 8.475109170305677,
+ "grad_norm": 0.01612604409456253,
+ "learning_rate": 0.0006,
+ "loss": 5.23960542678833,
+ "step": 610
+ },
+ {
+ "epoch": 8.489082969432314,
+ "grad_norm": 0.014293976128101349,
+ "learning_rate": 0.0006,
+ "loss": 5.119932651519775,
+ "step": 611
+ },
+ {
+ "epoch": 8.503056768558952,
+ "grad_norm": 0.014453536830842495,
+ "learning_rate": 0.0006,
+ "loss": 5.166794776916504,
+ "step": 612
+ },
+ {
+ "epoch": 8.51703056768559,
+ "grad_norm": 0.014685769565403461,
+ "learning_rate": 0.0006,
+ "loss": 5.264822959899902,
+ "step": 613
+ },
+ {
+ "epoch": 8.531004366812226,
+ "grad_norm": 0.01427740603685379,
+ "learning_rate": 0.0006,
+ "loss": 5.278433799743652,
+ "step": 614
+ },
+ {
+ "epoch": 8.544978165938865,
+ "grad_norm": 0.014864951372146606,
+ "learning_rate": 0.0006,
+ "loss": 5.214512348175049,
+ "step": 615
+ },
+ {
+ "epoch": 8.558951965065502,
+ "grad_norm": 0.01565164513885975,
+ "learning_rate": 0.0006,
+ "loss": 5.183066368103027,
+ "step": 616
+ },
+ {
+ "epoch": 8.57292576419214,
+ "grad_norm": 0.01779816672205925,
+ "learning_rate": 0.0006,
+ "loss": 5.163185119628906,
+ "step": 617
+ },
+ {
+ "epoch": 8.586899563318777,
+ "grad_norm": 0.017254827544093132,
+ "learning_rate": 0.0006,
+ "loss": 5.243555068969727,
+ "step": 618
+ },
+ {
+ "epoch": 8.600873362445414,
+ "grad_norm": 0.01650584116578102,
+ "learning_rate": 0.0006,
+ "loss": 5.234884738922119,
+ "step": 619
+ },
+ {
+ "epoch": 8.614847161572053,
+ "grad_norm": 0.017021115869283676,
+ "learning_rate": 0.0006,
+ "loss": 5.12271785736084,
+ "step": 620
+ },
+ {
+ "epoch": 8.62882096069869,
+ "grad_norm": 0.01773759163916111,
+ "learning_rate": 0.0006,
+ "loss": 5.242153167724609,
+ "step": 621
+ },
+ {
+ "epoch": 8.642794759825328,
+ "grad_norm": 0.015679042786359787,
+ "learning_rate": 0.0006,
+ "loss": 5.170779228210449,
+ "step": 622
+ },
+ {
+ "epoch": 8.656768558951965,
+ "grad_norm": 0.013760549947619438,
+ "learning_rate": 0.0006,
+ "loss": 5.238644599914551,
+ "step": 623
+ },
+ {
+ "epoch": 8.670742358078602,
+ "grad_norm": 0.014571522362530231,
+ "learning_rate": 0.0006,
+ "loss": 5.177056312561035,
+ "step": 624
+ },
+ {
+ "epoch": 8.68471615720524,
+ "grad_norm": 0.016209015622735023,
+ "learning_rate": 0.0006,
+ "loss": 5.163750648498535,
+ "step": 625
+ },
+ {
+ "epoch": 8.698689956331878,
+ "grad_norm": 0.016813941299915314,
+ "learning_rate": 0.0006,
+ "loss": 5.183428764343262,
+ "step": 626
+ },
+ {
+ "epoch": 8.712663755458514,
+ "grad_norm": 0.020985357463359833,
+ "learning_rate": 0.0006,
+ "loss": 5.229465007781982,
+ "step": 627
+ },
+ {
+ "epoch": 8.726637554585153,
+ "grad_norm": 0.02679632417857647,
+ "learning_rate": 0.0006,
+ "loss": 5.257368087768555,
+ "step": 628
+ },
+ {
+ "epoch": 8.74061135371179,
+ "grad_norm": 0.025756070390343666,
+ "learning_rate": 0.0006,
+ "loss": 5.253736972808838,
+ "step": 629
+ },
+ {
+ "epoch": 8.754585152838429,
+ "grad_norm": 0.02643490768969059,
+ "learning_rate": 0.0006,
+ "loss": 5.228633403778076,
+ "step": 630
+ },
+ {
+ "epoch": 8.768558951965066,
+ "grad_norm": 0.03303210437297821,
+ "learning_rate": 0.0006,
+ "loss": 5.163540363311768,
+ "step": 631
+ },
+ {
+ "epoch": 8.782532751091702,
+ "grad_norm": 0.030432431027293205,
+ "learning_rate": 0.0006,
+ "loss": 5.243169784545898,
+ "step": 632
+ },
+ {
+ "epoch": 8.796506550218341,
+ "grad_norm": 0.028631288558244705,
+ "learning_rate": 0.0006,
+ "loss": 5.166398048400879,
+ "step": 633
+ },
+ {
+ "epoch": 8.810480349344978,
+ "grad_norm": 0.026188310235738754,
+ "learning_rate": 0.0006,
+ "loss": 5.128122329711914,
+ "step": 634
+ },
+ {
+ "epoch": 8.824454148471617,
+ "grad_norm": 0.025308528915047646,
+ "learning_rate": 0.0006,
+ "loss": 5.211284637451172,
+ "step": 635
+ },
+ {
+ "epoch": 8.838427947598253,
+ "grad_norm": 0.020657729357481003,
+ "learning_rate": 0.0006,
+ "loss": 5.161575794219971,
+ "step": 636
+ },
+ {
+ "epoch": 8.85240174672489,
+ "grad_norm": 0.021901234984397888,
+ "learning_rate": 0.0006,
+ "loss": 5.20050048828125,
+ "step": 637
+ },
+ {
+ "epoch": 8.866375545851529,
+ "grad_norm": 0.021219318732619286,
+ "learning_rate": 0.0006,
+ "loss": 5.211699485778809,
+ "step": 638
+ },
+ {
+ "epoch": 8.880349344978166,
+ "grad_norm": 0.0200974028557539,
+ "learning_rate": 0.0006,
+ "loss": 5.092503547668457,
+ "step": 639
+ },
+ {
+ "epoch": 8.894323144104803,
+ "grad_norm": 0.023804882541298866,
+ "learning_rate": 0.0006,
+ "loss": 5.216068267822266,
+ "step": 640
+ },
+ {
+ "epoch": 8.908296943231441,
+ "grad_norm": 0.026088010519742966,
+ "learning_rate": 0.0006,
+ "loss": 5.155592918395996,
+ "step": 641
+ },
+ {
+ "epoch": 8.922270742358078,
+ "grad_norm": 0.02601276896893978,
+ "learning_rate": 0.0006,
+ "loss": 5.197238922119141,
+ "step": 642
+ },
+ {
+ "epoch": 8.936244541484717,
+ "grad_norm": 0.020387211814522743,
+ "learning_rate": 0.0006,
+ "loss": 5.1955437660217285,
+ "step": 643
+ },
+ {
+ "epoch": 8.950218340611354,
+ "grad_norm": 0.019214622676372528,
+ "learning_rate": 0.0006,
+ "loss": 5.206346035003662,
+ "step": 644
+ },
+ {
+ "epoch": 8.96419213973799,
+ "grad_norm": 0.019674314185976982,
+ "learning_rate": 0.0006,
+ "loss": 5.219857215881348,
+ "step": 645
+ },
+ {
+ "epoch": 8.97816593886463,
+ "grad_norm": 0.020318234339356422,
+ "learning_rate": 0.0006,
+ "loss": 5.191132545471191,
+ "step": 646
+ },
+ {
+ "epoch": 8.992139737991266,
+ "grad_norm": 0.021428676322102547,
+ "learning_rate": 0.0006,
+ "loss": 5.214695930480957,
+ "step": 647
+ },
+ {
+ "epoch": 9.0,
+ "grad_norm": 0.023235054686665535,
+ "learning_rate": 0.0006,
+ "loss": 5.229091644287109,
+ "step": 648
+ },
+ {
+ "epoch": 9.0,
+ "eval_loss": 5.289593696594238,
+ "eval_runtime": 56.9256,
+ "eval_samples_per_second": 42.898,
+ "eval_steps_per_second": 1.353,
+ "step": 648
+ },
+ {
+ "epoch": 9.013973799126637,
+ "grad_norm": 0.02320289984345436,
+ "learning_rate": 0.0006,
+ "loss": 4.983669281005859,
+ "step": 649
+ },
+ {
+ "epoch": 9.027947598253276,
+ "grad_norm": 0.018521282821893692,
+ "learning_rate": 0.0006,
+ "loss": 5.200575828552246,
+ "step": 650
+ },
+ {
+ "epoch": 9.041921397379912,
+ "grad_norm": 0.0180169939994812,
+ "learning_rate": 0.0006,
+ "loss": 5.148033618927002,
+ "step": 651
+ },
+ {
+ "epoch": 9.055895196506551,
+ "grad_norm": 0.019969860091805458,
+ "learning_rate": 0.0006,
+ "loss": 5.099125385284424,
+ "step": 652
+ },
+ {
+ "epoch": 9.069868995633188,
+ "grad_norm": 0.01643305830657482,
+ "learning_rate": 0.0006,
+ "loss": 5.092347145080566,
+ "step": 653
+ },
+ {
+ "epoch": 9.083842794759825,
+ "grad_norm": 0.016022363677620888,
+ "learning_rate": 0.0006,
+ "loss": 5.129279613494873,
+ "step": 654
+ },
+ {
+ "epoch": 9.097816593886463,
+ "grad_norm": 0.01657041162252426,
+ "learning_rate": 0.0006,
+ "loss": 5.159562587738037,
+ "step": 655
+ },
+ {
+ "epoch": 9.1117903930131,
+ "grad_norm": 0.017924228683114052,
+ "learning_rate": 0.0006,
+ "loss": 5.165497779846191,
+ "step": 656
+ },
+ {
+ "epoch": 9.125764192139737,
+ "grad_norm": 0.021592361852526665,
+ "learning_rate": 0.0006,
+ "loss": 5.204775810241699,
+ "step": 657
+ },
+ {
+ "epoch": 9.139737991266376,
+ "grad_norm": 0.026924120262265205,
+ "learning_rate": 0.0006,
+ "loss": 5.128296852111816,
+ "step": 658
+ },
+ {
+ "epoch": 9.153711790393013,
+ "grad_norm": 0.023156899958848953,
+ "learning_rate": 0.0006,
+ "loss": 5.1255340576171875,
+ "step": 659
+ },
+ {
+ "epoch": 9.167685589519651,
+ "grad_norm": 0.020670367404818535,
+ "learning_rate": 0.0006,
+ "loss": 5.128479957580566,
+ "step": 660
+ },
+ {
+ "epoch": 9.181659388646288,
+ "grad_norm": 0.025447173044085503,
+ "learning_rate": 0.0006,
+ "loss": 5.154821872711182,
+ "step": 661
+ },
+ {
+ "epoch": 9.195633187772925,
+ "grad_norm": 0.027940068393945694,
+ "learning_rate": 0.0006,
+ "loss": 5.143466472625732,
+ "step": 662
+ },
+ {
+ "epoch": 9.209606986899564,
+ "grad_norm": 0.026847844943404198,
+ "learning_rate": 0.0006,
+ "loss": 5.045665740966797,
+ "step": 663
+ },
+ {
+ "epoch": 9.2235807860262,
+ "grad_norm": 0.02340601570904255,
+ "learning_rate": 0.0006,
+ "loss": 5.21933126449585,
+ "step": 664
+ },
+ {
+ "epoch": 9.237554585152838,
+ "grad_norm": 0.02340371161699295,
+ "learning_rate": 0.0006,
+ "loss": 4.942949295043945,
+ "step": 665
+ },
+ {
+ "epoch": 9.251528384279476,
+ "grad_norm": 0.02221992425620556,
+ "learning_rate": 0.0006,
+ "loss": 5.197099208831787,
+ "step": 666
+ },
+ {
+ "epoch": 9.265502183406113,
+ "grad_norm": 0.023508677259087563,
+ "learning_rate": 0.0006,
+ "loss": 5.022897720336914,
+ "step": 667
+ },
+ {
+ "epoch": 9.279475982532752,
+ "grad_norm": 0.026120394468307495,
+ "learning_rate": 0.0006,
+ "loss": 5.087725639343262,
+ "step": 668
+ },
+ {
+ "epoch": 9.293449781659389,
+ "grad_norm": 0.027273228392004967,
+ "learning_rate": 0.0006,
+ "loss": 5.090963840484619,
+ "step": 669
+ },
+ {
+ "epoch": 9.307423580786025,
+ "grad_norm": 0.03241586685180664,
+ "learning_rate": 0.0006,
+ "loss": 5.047842979431152,
+ "step": 670
+ },
+ {
+ "epoch": 9.321397379912664,
+ "grad_norm": 0.030245667323470116,
+ "learning_rate": 0.0006,
+ "loss": 5.104803085327148,
+ "step": 671
+ },
+ {
+ "epoch": 9.335371179039301,
+ "grad_norm": 0.027698364108800888,
+ "learning_rate": 0.0006,
+ "loss": 5.0644097328186035,
+ "step": 672
+ },
+ {
+ "epoch": 9.34934497816594,
+ "grad_norm": 0.029692554846405983,
+ "learning_rate": 0.0006,
+ "loss": 5.111942291259766,
+ "step": 673
+ },
+ {
+ "epoch": 9.363318777292577,
+ "grad_norm": 0.03328656405210495,
+ "learning_rate": 0.0006,
+ "loss": 5.1574554443359375,
+ "step": 674
+ },
+ {
+ "epoch": 9.377292576419213,
+ "grad_norm": 0.031938180327415466,
+ "learning_rate": 0.0006,
+ "loss": 5.190371990203857,
+ "step": 675
+ },
+ {
+ "epoch": 9.391266375545852,
+ "grad_norm": 0.024858945980668068,
+ "learning_rate": 0.0006,
+ "loss": 5.121420860290527,
+ "step": 676
+ },
+ {
+ "epoch": 9.405240174672489,
+ "grad_norm": 0.023033203557133675,
+ "learning_rate": 0.0006,
+ "loss": 5.0974345207214355,
+ "step": 677
+ },
+ {
+ "epoch": 9.419213973799126,
+ "grad_norm": 0.022083545103669167,
+ "learning_rate": 0.0006,
+ "loss": 5.046319961547852,
+ "step": 678
+ },
+ {
+ "epoch": 9.433187772925764,
+ "grad_norm": 0.0228437427431345,
+ "learning_rate": 0.0006,
+ "loss": 5.119963645935059,
+ "step": 679
+ },
+ {
+ "epoch": 9.447161572052401,
+ "grad_norm": 0.0242351982742548,
+ "learning_rate": 0.0006,
+ "loss": 5.0444560050964355,
+ "step": 680
+ },
+ {
+ "epoch": 9.46113537117904,
+ "grad_norm": 0.02401185780763626,
+ "learning_rate": 0.0006,
+ "loss": 5.063238620758057,
+ "step": 681
+ },
+ {
+ "epoch": 9.475109170305677,
+ "grad_norm": 0.02560959942638874,
+ "learning_rate": 0.0006,
+ "loss": 5.035224914550781,
+ "step": 682
+ },
+ {
+ "epoch": 9.489082969432314,
+ "grad_norm": 0.024985626339912415,
+ "learning_rate": 0.0006,
+ "loss": 5.049284934997559,
+ "step": 683
+ },
+ {
+ "epoch": 9.503056768558952,
+ "grad_norm": 0.023178115487098694,
+ "learning_rate": 0.0006,
+ "loss": 5.2266621589660645,
+ "step": 684
+ },
+ {
+ "epoch": 9.51703056768559,
+ "grad_norm": 0.0215512253344059,
+ "learning_rate": 0.0006,
+ "loss": 5.1875715255737305,
+ "step": 685
+ },
+ {
+ "epoch": 9.531004366812226,
+ "grad_norm": 0.01662837713956833,
+ "learning_rate": 0.0006,
+ "loss": 5.137012481689453,
+ "step": 686
+ },
+ {
+ "epoch": 9.544978165938865,
+ "grad_norm": 0.01952764391899109,
+ "learning_rate": 0.0006,
+ "loss": 5.072500228881836,
+ "step": 687
+ },
+ {
+ "epoch": 9.558951965065502,
+ "grad_norm": 0.017521077767014503,
+ "learning_rate": 0.0006,
+ "loss": 5.045711994171143,
+ "step": 688
+ },
+ {
+ "epoch": 9.57292576419214,
+ "grad_norm": 0.017355265095829964,
+ "learning_rate": 0.0006,
+ "loss": 5.02881383895874,
+ "step": 689
+ },
+ {
+ "epoch": 9.586899563318777,
+ "grad_norm": 0.015587719157338142,
+ "learning_rate": 0.0006,
+ "loss": 5.132787704467773,
+ "step": 690
+ },
+ {
+ "epoch": 9.600873362445414,
+ "grad_norm": 0.017469845712184906,
+ "learning_rate": 0.0006,
+ "loss": 5.105447292327881,
+ "step": 691
+ },
+ {
+ "epoch": 9.614847161572053,
+ "grad_norm": 0.019630100578069687,
+ "learning_rate": 0.0006,
+ "loss": 5.032464981079102,
+ "step": 692
+ },
+ {
+ "epoch": 9.62882096069869,
+ "grad_norm": 0.01784994639456272,
+ "learning_rate": 0.0006,
+ "loss": 5.081345558166504,
+ "step": 693
+ },
+ {
+ "epoch": 9.642794759825328,
+ "grad_norm": 0.01872754842042923,
+ "learning_rate": 0.0006,
+ "loss": 4.9772114753723145,
+ "step": 694
+ },
+ {
+ "epoch": 9.656768558951965,
+ "grad_norm": 0.01999032311141491,
+ "learning_rate": 0.0006,
+ "loss": 5.089740753173828,
+ "step": 695
+ },
+ {
+ "epoch": 9.670742358078602,
+ "grad_norm": 0.01710767112672329,
+ "learning_rate": 0.0006,
+ "loss": 4.9869890213012695,
+ "step": 696
+ },
+ {
+ "epoch": 9.68471615720524,
+ "grad_norm": 0.015144161880016327,
+ "learning_rate": 0.0006,
+ "loss": 5.022564888000488,
+ "step": 697
+ },
+ {
+ "epoch": 9.698689956331878,
+ "grad_norm": 0.01618543453514576,
+ "learning_rate": 0.0006,
+ "loss": 5.035921573638916,
+ "step": 698
+ },
+ {
+ "epoch": 9.712663755458514,
+ "grad_norm": 0.014653387479484081,
+ "learning_rate": 0.0006,
+ "loss": 4.914456367492676,
+ "step": 699
+ },
+ {
+ "epoch": 9.726637554585153,
+ "grad_norm": 0.013434977270662785,
+ "learning_rate": 0.0006,
+ "loss": 5.063764572143555,
+ "step": 700
+ },
+ {
+ "epoch": 9.74061135371179,
+ "grad_norm": 0.015128864906728268,
+ "learning_rate": 0.0006,
+ "loss": 5.052367687225342,
+ "step": 701
+ },
+ {
+ "epoch": 9.754585152838429,
+ "grad_norm": 0.014210768043994904,
+ "learning_rate": 0.0006,
+ "loss": 5.016489028930664,
+ "step": 702
+ },
+ {
+ "epoch": 9.768558951965066,
+ "grad_norm": 0.015191680751740932,
+ "learning_rate": 0.0006,
+ "loss": 4.990176200866699,
+ "step": 703
+ },
+ {
+ "epoch": 9.782532751091702,
+ "grad_norm": 0.018285127356648445,
+ "learning_rate": 0.0006,
+ "loss": 5.015164375305176,
+ "step": 704
+ },
+ {
+ "epoch": 9.796506550218341,
+ "grad_norm": 0.0204079058021307,
+ "learning_rate": 0.0006,
+ "loss": 5.091142654418945,
+ "step": 705
+ },
+ {
+ "epoch": 9.810480349344978,
+ "grad_norm": 0.01980586163699627,
+ "learning_rate": 0.0006,
+ "loss": 5.03350830078125,
+ "step": 706
+ },
+ {
+ "epoch": 9.824454148471617,
+ "grad_norm": 0.02048729732632637,
+ "learning_rate": 0.0006,
+ "loss": 5.03611421585083,
+ "step": 707
+ },
+ {
+ "epoch": 9.838427947598253,
+ "grad_norm": 0.02058163844048977,
+ "learning_rate": 0.0006,
+ "loss": 4.9613447189331055,
+ "step": 708
+ },
+ {
+ "epoch": 9.85240174672489,
+ "grad_norm": 0.018004924058914185,
+ "learning_rate": 0.0006,
+ "loss": 4.966541767120361,
+ "step": 709
+ },
+ {
+ "epoch": 9.866375545851529,
+ "grad_norm": 0.0198152307420969,
+ "learning_rate": 0.0006,
+ "loss": 5.01750373840332,
+ "step": 710
+ },
+ {
+ "epoch": 9.880349344978166,
+ "grad_norm": 0.018532969057559967,
+ "learning_rate": 0.0006,
+ "loss": 5.097424507141113,
+ "step": 711
+ },
+ {
+ "epoch": 9.894323144104803,
+ "grad_norm": 0.018939971923828125,
+ "learning_rate": 0.0006,
+ "loss": 5.177223205566406,
+ "step": 712
+ },
+ {
+ "epoch": 9.908296943231441,
+ "grad_norm": 0.017969170585274696,
+ "learning_rate": 0.0006,
+ "loss": 5.118013858795166,
+ "step": 713
+ },
+ {
+ "epoch": 9.922270742358078,
+ "grad_norm": 0.018021270632743835,
+ "learning_rate": 0.0006,
+ "loss": 5.034787178039551,
+ "step": 714
+ },
+ {
+ "epoch": 9.936244541484717,
+ "grad_norm": 0.017009710893034935,
+ "learning_rate": 0.0006,
+ "loss": 5.003920555114746,
+ "step": 715
+ },
+ {
+ "epoch": 9.950218340611354,
+ "grad_norm": 0.017722049728035927,
+ "learning_rate": 0.0006,
+ "loss": 5.031939506530762,
+ "step": 716
+ },
+ {
+ "epoch": 9.96419213973799,
+ "grad_norm": 0.017134087160229683,
+ "learning_rate": 0.0006,
+ "loss": 5.093474388122559,
+ "step": 717
+ },
+ {
+ "epoch": 9.97816593886463,
+ "grad_norm": 0.018511613830924034,
+ "learning_rate": 0.0006,
+ "loss": 5.078604698181152,
+ "step": 718
+ },
+ {
+ "epoch": 9.992139737991266,
+ "grad_norm": 0.021206334233283997,
+ "learning_rate": 0.0006,
+ "loss": 5.052572727203369,
+ "step": 719
+ },
+ {
+ "epoch": 10.0,
+ "grad_norm": 0.02239062264561653,
+ "learning_rate": 0.0006,
+ "loss": 5.063547134399414,
+ "step": 720
+ },
+ {
+ "epoch": 10.0,
+ "eval_loss": 5.173044681549072,
+ "eval_runtime": 56.7287,
+ "eval_samples_per_second": 43.047,
+ "eval_steps_per_second": 1.357,
+ "step": 720
+ },
+ {
+ "epoch": 10.013973799126637,
+ "grad_norm": 0.020757969468832016,
+ "learning_rate": 0.0006,
+ "loss": 5.045648574829102,
+ "step": 721
+ },
+ {
+ "epoch": 10.027947598253276,
+ "grad_norm": 0.022697702050209045,
+ "learning_rate": 0.0006,
+ "loss": 4.993833065032959,
+ "step": 722
+ },
+ {
+ "epoch": 10.041921397379912,
+ "grad_norm": 0.02434331737458706,
+ "learning_rate": 0.0006,
+ "loss": 5.073219299316406,
+ "step": 723
+ },
+ {
+ "epoch": 10.055895196506551,
+ "grad_norm": 0.025499660521745682,
+ "learning_rate": 0.0006,
+ "loss": 5.019072532653809,
+ "step": 724
+ },
+ {
+ "epoch": 10.069868995633188,
+ "grad_norm": 0.029281053692102432,
+ "learning_rate": 0.0006,
+ "loss": 4.959851264953613,
+ "step": 725
+ },
+ {
+ "epoch": 10.083842794759825,
+ "grad_norm": 0.033878110349178314,
+ "learning_rate": 0.0006,
+ "loss": 5.132717609405518,
+ "step": 726
+ },
+ {
+ "epoch": 10.097816593886463,
+ "grad_norm": 0.03385764732956886,
+ "learning_rate": 0.0006,
+ "loss": 5.0482892990112305,
+ "step": 727
+ },
+ {
+ "epoch": 10.1117903930131,
+ "grad_norm": 0.03466648980975151,
+ "learning_rate": 0.0006,
+ "loss": 4.946855545043945,
+ "step": 728
+ },
+ {
+ "epoch": 10.125764192139737,
+ "grad_norm": 0.028546305373311043,
+ "learning_rate": 0.0006,
+ "loss": 5.027774333953857,
+ "step": 729
+ },
+ {
+ "epoch": 10.139737991266376,
+ "grad_norm": 0.030575869604945183,
+ "learning_rate": 0.0006,
+ "loss": 4.9768476486206055,
+ "step": 730
+ },
+ {
+ "epoch": 10.153711790393013,
+ "grad_norm": 0.03614303469657898,
+ "learning_rate": 0.0006,
+ "loss": 5.078113555908203,
+ "step": 731
+ },
+ {
+ "epoch": 10.167685589519651,
+ "grad_norm": 0.040991660207509995,
+ "learning_rate": 0.0006,
+ "loss": 5.033185958862305,
+ "step": 732
+ },
+ {
+ "epoch": 10.181659388646288,
+ "grad_norm": 0.04551699757575989,
+ "learning_rate": 0.0006,
+ "loss": 5.0820746421813965,
+ "step": 733
+ },
+ {
+ "epoch": 10.195633187772925,
+ "grad_norm": 0.040193989872932434,
+ "learning_rate": 0.0006,
+ "loss": 5.059700965881348,
+ "step": 734
+ },
+ {
+ "epoch": 10.209606986899564,
+ "grad_norm": 0.035851605236530304,
+ "learning_rate": 0.0006,
+ "loss": 5.142467498779297,
+ "step": 735
+ },
+ {
+ "epoch": 10.2235807860262,
+ "grad_norm": 0.034200169146060944,
+ "learning_rate": 0.0006,
+ "loss": 5.072312355041504,
+ "step": 736
+ },
+ {
+ "epoch": 10.237554585152838,
+ "grad_norm": 0.033621110022068024,
+ "learning_rate": 0.0006,
+ "loss": 5.08037805557251,
+ "step": 737
+ },
+ {
+ "epoch": 10.251528384279476,
+ "grad_norm": 0.037372369319200516,
+ "learning_rate": 0.0006,
+ "loss": 4.998425483703613,
+ "step": 738
+ },
+ {
+ "epoch": 10.265502183406113,
+ "grad_norm": 0.037044707685709,
+ "learning_rate": 0.0006,
+ "loss": 4.923993110656738,
+ "step": 739
+ },
+ {
+ "epoch": 10.279475982532752,
+ "grad_norm": 0.030454808846116066,
+ "learning_rate": 0.0006,
+ "loss": 5.0889997482299805,
+ "step": 740
+ },
+ {
+ "epoch": 10.293449781659389,
+ "grad_norm": 0.030969638377428055,
+ "learning_rate": 0.0006,
+ "loss": 4.984112739562988,
+ "step": 741
+ },
+ {
+ "epoch": 10.307423580786025,
+ "grad_norm": 0.0276983380317688,
+ "learning_rate": 0.0006,
+ "loss": 4.955412864685059,
+ "step": 742
+ },
+ {
+ "epoch": 10.321397379912664,
+ "grad_norm": 0.02716052532196045,
+ "learning_rate": 0.0006,
+ "loss": 5.0098876953125,
+ "step": 743
+ },
+ {
+ "epoch": 10.335371179039301,
+ "grad_norm": 0.02403552085161209,
+ "learning_rate": 0.0006,
+ "loss": 5.165160655975342,
+ "step": 744
+ },
+ {
+ "epoch": 10.34934497816594,
+ "grad_norm": 0.0257862601429224,
+ "learning_rate": 0.0006,
+ "loss": 5.001547813415527,
+ "step": 745
+ },
+ {
+ "epoch": 10.363318777292577,
+ "grad_norm": 0.02300378680229187,
+ "learning_rate": 0.0006,
+ "loss": 5.038888931274414,
+ "step": 746
+ },
+ {
+ "epoch": 10.377292576419213,
+ "grad_norm": 0.019766854122281075,
+ "learning_rate": 0.0006,
+ "loss": 4.924291610717773,
+ "step": 747
+ },
+ {
+ "epoch": 10.391266375545852,
+ "grad_norm": 0.01888395845890045,
+ "learning_rate": 0.0006,
+ "loss": 5.059971332550049,
+ "step": 748
+ },
+ {
+ "epoch": 10.405240174672489,
+ "grad_norm": 0.015577499754726887,
+ "learning_rate": 0.0006,
+ "loss": 4.987686634063721,
+ "step": 749
+ },
+ {
+ "epoch": 10.419213973799126,
+ "grad_norm": 0.016012346372008324,
+ "learning_rate": 0.0006,
+ "loss": 5.015157699584961,
+ "step": 750
+ },
+ {
+ "epoch": 10.433187772925764,
+ "grad_norm": 0.016277998685836792,
+ "learning_rate": 0.0006,
+ "loss": 5.004931449890137,
+ "step": 751
+ },
+ {
+ "epoch": 10.447161572052401,
+ "grad_norm": 0.017504561692476273,
+ "learning_rate": 0.0006,
+ "loss": 4.902827739715576,
+ "step": 752
+ },
+ {
+ "epoch": 10.46113537117904,
+ "grad_norm": 0.01569022797048092,
+ "learning_rate": 0.0006,
+ "loss": 4.954863548278809,
+ "step": 753
+ },
+ {
+ "epoch": 10.475109170305677,
+ "grad_norm": 0.014114422723650932,
+ "learning_rate": 0.0006,
+ "loss": 5.00656795501709,
+ "step": 754
+ },
+ {
+ "epoch": 10.489082969432314,
+ "grad_norm": 0.01573832333087921,
+ "learning_rate": 0.0006,
+ "loss": 4.978353023529053,
+ "step": 755
+ },
+ {
+ "epoch": 10.503056768558952,
+ "grad_norm": 0.016480036079883575,
+ "learning_rate": 0.0006,
+ "loss": 4.953327655792236,
+ "step": 756
+ },
+ {
+ "epoch": 10.51703056768559,
+ "grad_norm": 0.014008025638759136,
+ "learning_rate": 0.0006,
+ "loss": 5.050840377807617,
+ "step": 757
+ },
+ {
+ "epoch": 10.531004366812226,
+ "grad_norm": 0.013147998601198196,
+ "learning_rate": 0.0006,
+ "loss": 4.974964141845703,
+ "step": 758
+ },
+ {
+ "epoch": 10.544978165938865,
+ "grad_norm": 0.013807603158056736,
+ "learning_rate": 0.0006,
+ "loss": 4.927907943725586,
+ "step": 759
+ },
+ {
+ "epoch": 10.558951965065502,
+ "grad_norm": 0.013555224984884262,
+ "learning_rate": 0.0006,
+ "loss": 4.980690956115723,
+ "step": 760
+ },
+ {
+ "epoch": 10.57292576419214,
+ "grad_norm": 0.013045408762991428,
+ "learning_rate": 0.0006,
+ "loss": 4.9436116218566895,
+ "step": 761
+ },
+ {
+ "epoch": 10.586899563318777,
+ "grad_norm": 0.012858862057328224,
+ "learning_rate": 0.0006,
+ "loss": 4.9595770835876465,
+ "step": 762
+ },
+ {
+ "epoch": 10.600873362445414,
+ "grad_norm": 0.01468253880739212,
+ "learning_rate": 0.0006,
+ "loss": 4.90725040435791,
+ "step": 763
+ },
+ {
+ "epoch": 10.614847161572053,
+ "grad_norm": 0.013224679045379162,
+ "learning_rate": 0.0006,
+ "loss": 4.877615928649902,
+ "step": 764
+ },
+ {
+ "epoch": 10.62882096069869,
+ "grad_norm": 0.013555348850786686,
+ "learning_rate": 0.0006,
+ "loss": 4.9456610679626465,
+ "step": 765
+ },
+ {
+ "epoch": 10.642794759825328,
+ "grad_norm": 0.014387160539627075,
+ "learning_rate": 0.0006,
+ "loss": 4.9233717918396,
+ "step": 766
+ },
+ {
+ "epoch": 10.656768558951965,
+ "grad_norm": 0.014354088343679905,
+ "learning_rate": 0.0006,
+ "loss": 4.983511924743652,
+ "step": 767
+ },
+ {
+ "epoch": 10.670742358078602,
+ "grad_norm": 0.01332700252532959,
+ "learning_rate": 0.0006,
+ "loss": 4.934416770935059,
+ "step": 768
+ },
+ {
+ "epoch": 10.68471615720524,
+ "grad_norm": 0.011543313041329384,
+ "learning_rate": 0.0006,
+ "loss": 4.96388053894043,
+ "step": 769
+ },
+ {
+ "epoch": 10.698689956331878,
+ "grad_norm": 0.012330878525972366,
+ "learning_rate": 0.0006,
+ "loss": 4.946396350860596,
+ "step": 770
+ },
+ {
+ "epoch": 10.712663755458514,
+ "grad_norm": 0.01379779726266861,
+ "learning_rate": 0.0006,
+ "loss": 4.981019020080566,
+ "step": 771
+ },
+ {
+ "epoch": 10.726637554585153,
+ "grad_norm": 0.014612431637942791,
+ "learning_rate": 0.0006,
+ "loss": 4.894567966461182,
+ "step": 772
+ },
+ {
+ "epoch": 10.74061135371179,
+ "grad_norm": 0.013062899932265282,
+ "learning_rate": 0.0006,
+ "loss": 5.033038139343262,
+ "step": 773
+ },
+ {
+ "epoch": 10.754585152838429,
+ "grad_norm": 0.013898049481213093,
+ "learning_rate": 0.0006,
+ "loss": 4.823636054992676,
+ "step": 774
+ },
+ {
+ "epoch": 10.768558951965066,
+ "grad_norm": 0.01502019353210926,
+ "learning_rate": 0.0006,
+ "loss": 5.020156383514404,
+ "step": 775
+ },
+ {
+ "epoch": 10.782532751091702,
+ "grad_norm": 0.013810204342007637,
+ "learning_rate": 0.0006,
+ "loss": 4.935647010803223,
+ "step": 776
+ },
+ {
+ "epoch": 10.796506550218341,
+ "grad_norm": 0.013616513460874557,
+ "learning_rate": 0.0006,
+ "loss": 4.901486396789551,
+ "step": 777
+ },
+ {
+ "epoch": 10.810480349344978,
+ "grad_norm": 0.013601024635136127,
+ "learning_rate": 0.0006,
+ "loss": 4.931873321533203,
+ "step": 778
+ },
+ {
+ "epoch": 10.824454148471617,
+ "grad_norm": 0.013668350875377655,
+ "learning_rate": 0.0006,
+ "loss": 4.937256813049316,
+ "step": 779
+ },
+ {
+ "epoch": 10.838427947598253,
+ "grad_norm": 0.01581823080778122,
+ "learning_rate": 0.0006,
+ "loss": 4.9611406326293945,
+ "step": 780
+ },
+ {
+ "epoch": 10.85240174672489,
+ "grad_norm": 0.019478484988212585,
+ "learning_rate": 0.0006,
+ "loss": 4.9212236404418945,
+ "step": 781
+ },
+ {
+ "epoch": 10.866375545851529,
+ "grad_norm": 0.026921333745121956,
+ "learning_rate": 0.0006,
+ "loss": 4.88501501083374,
+ "step": 782
+ },
+ {
+ "epoch": 10.880349344978166,
+ "grad_norm": 0.03155914694070816,
+ "learning_rate": 0.0006,
+ "loss": 4.989967346191406,
+ "step": 783
+ },
+ {
+ "epoch": 10.894323144104803,
+ "grad_norm": 0.029689429327845573,
+ "learning_rate": 0.0006,
+ "loss": 4.936000823974609,
+ "step": 784
+ },
+ {
+ "epoch": 10.908296943231441,
+ "grad_norm": 0.026626063510775566,
+ "learning_rate": 0.0006,
+ "loss": 4.984785079956055,
+ "step": 785
+ },
+ {
+ "epoch": 10.922270742358078,
+ "grad_norm": 0.026168212294578552,
+ "learning_rate": 0.0006,
+ "loss": 4.908282279968262,
+ "step": 786
+ },
+ {
+ "epoch": 10.936244541484717,
+ "grad_norm": 0.02828158251941204,
+ "learning_rate": 0.0006,
+ "loss": 4.926390647888184,
+ "step": 787
+ },
+ {
+ "epoch": 10.950218340611354,
+ "grad_norm": 0.02649078331887722,
+ "learning_rate": 0.0006,
+ "loss": 5.00400972366333,
+ "step": 788
+ },
+ {
+ "epoch": 10.96419213973799,
+ "grad_norm": 0.0256856270134449,
+ "learning_rate": 0.0006,
+ "loss": 5.0317487716674805,
+ "step": 789
+ },
+ {
+ "epoch": 10.97816593886463,
+ "grad_norm": 0.024728883057832718,
+ "learning_rate": 0.0006,
+ "loss": 4.9299635887146,
+ "step": 790
+ },
+ {
+ "epoch": 10.992139737991266,
+ "grad_norm": 0.026241116225719452,
+ "learning_rate": 0.0006,
+ "loss": 4.947319984436035,
+ "step": 791
+ },
+ {
+ "epoch": 11.0,
+ "grad_norm": 0.023526743054389954,
+ "learning_rate": 0.0006,
+ "loss": 4.94429349899292,
+ "step": 792
+ },
+ {
+ "epoch": 11.0,
+ "eval_loss": 5.104345798492432,
+ "eval_runtime": 57.0955,
+ "eval_samples_per_second": 42.77,
+ "eval_steps_per_second": 1.349,
+ "step": 792
+ },
+ {
+ "epoch": 11.013973799126637,
+ "grad_norm": 0.0216788612306118,
+ "learning_rate": 0.0006,
+ "loss": 4.881950855255127,
+ "step": 793
+ },
+ {
+ "epoch": 11.027947598253276,
+ "grad_norm": 0.01830463856458664,
+ "learning_rate": 0.0006,
+ "loss": 4.971016883850098,
+ "step": 794
+ },
+ {
+ "epoch": 11.041921397379912,
+ "grad_norm": 0.02204521745443344,
+ "learning_rate": 0.0006,
+ "loss": 4.894623756408691,
+ "step": 795
+ },
+ {
+ "epoch": 11.055895196506551,
+ "grad_norm": 0.0210255216807127,
+ "learning_rate": 0.0006,
+ "loss": 4.834009170532227,
+ "step": 796
+ },
+ {
+ "epoch": 11.069868995633188,
+ "grad_norm": 0.019797448068857193,
+ "learning_rate": 0.0006,
+ "loss": 5.086714744567871,
+ "step": 797
+ },
+ {
+ "epoch": 11.083842794759825,
+ "grad_norm": 0.02036641724407673,
+ "learning_rate": 0.0006,
+ "loss": 4.920056343078613,
+ "step": 798
+ },
+ {
+ "epoch": 11.097816593886463,
+ "grad_norm": 0.01966066285967827,
+ "learning_rate": 0.0006,
+ "loss": 4.879859447479248,
+ "step": 799
+ },
+ {
+ "epoch": 11.1117903930131,
+ "grad_norm": 0.0202149897813797,
+ "learning_rate": 0.0006,
+ "loss": 4.933640480041504,
+ "step": 800
+ },
+ {
+ "epoch": 11.125764192139737,
+ "grad_norm": 0.01893039606511593,
+ "learning_rate": 0.0006,
+ "loss": 4.901970863342285,
+ "step": 801
+ },
+ {
+ "epoch": 11.139737991266376,
+ "grad_norm": 0.016040155664086342,
+ "learning_rate": 0.0006,
+ "loss": 4.905491828918457,
+ "step": 802
+ },
+ {
+ "epoch": 11.153711790393013,
+ "grad_norm": 0.01720350608229637,
+ "learning_rate": 0.0006,
+ "loss": 4.798816680908203,
+ "step": 803
+ },
+ {
+ "epoch": 11.167685589519651,
+ "grad_norm": 0.017002852633595467,
+ "learning_rate": 0.0006,
+ "loss": 4.851859092712402,
+ "step": 804
+ },
+ {
+ "epoch": 11.181659388646288,
+ "grad_norm": 0.015594680793583393,
+ "learning_rate": 0.0006,
+ "loss": 4.90328311920166,
+ "step": 805
+ },
+ {
+ "epoch": 11.195633187772925,
+ "grad_norm": 0.014904462732374668,
+ "learning_rate": 0.0006,
+ "loss": 4.904215335845947,
+ "step": 806
+ },
+ {
+ "epoch": 11.209606986899564,
+ "grad_norm": 0.01366228237748146,
+ "learning_rate": 0.0006,
+ "loss": 4.9813055992126465,
+ "step": 807
+ },
+ {
+ "epoch": 11.2235807860262,
+ "grad_norm": 0.014719638973474503,
+ "learning_rate": 0.0006,
+ "loss": 4.927306175231934,
+ "step": 808
+ },
+ {
+ "epoch": 11.237554585152838,
+ "grad_norm": 0.015016036108136177,
+ "learning_rate": 0.0006,
+ "loss": 4.874215126037598,
+ "step": 809
+ },
+ {
+ "epoch": 11.251528384279476,
+ "grad_norm": 0.014925424009561539,
+ "learning_rate": 0.0006,
+ "loss": 4.838257789611816,
+ "step": 810
+ },
+ {
+ "epoch": 11.265502183406113,
+ "grad_norm": 0.014335823245346546,
+ "learning_rate": 0.0006,
+ "loss": 4.940608024597168,
+ "step": 811
+ },
+ {
+ "epoch": 11.279475982532752,
+ "grad_norm": 0.014987512491643429,
+ "learning_rate": 0.0006,
+ "loss": 4.831531524658203,
+ "step": 812
+ },
+ {
+ "epoch": 11.293449781659389,
+ "grad_norm": 0.016087457537651062,
+ "learning_rate": 0.0006,
+ "loss": 4.863227367401123,
+ "step": 813
+ },
+ {
+ "epoch": 11.307423580786025,
+ "grad_norm": 0.01671300269663334,
+ "learning_rate": 0.0006,
+ "loss": 4.8535661697387695,
+ "step": 814
+ },
+ {
+ "epoch": 11.321397379912664,
+ "grad_norm": 0.015229142270982265,
+ "learning_rate": 0.0006,
+ "loss": 4.948635101318359,
+ "step": 815
+ },
+ {
+ "epoch": 11.335371179039301,
+ "grad_norm": 0.01532587967813015,
+ "learning_rate": 0.0006,
+ "loss": 4.9035797119140625,
+ "step": 816
+ },
+ {
+ "epoch": 11.34934497816594,
+ "grad_norm": 0.015715476125478745,
+ "learning_rate": 0.0006,
+ "loss": 4.894709587097168,
+ "step": 817
+ },
+ {
+ "epoch": 11.363318777292577,
+ "grad_norm": 0.014156977646052837,
+ "learning_rate": 0.0006,
+ "loss": 4.960206985473633,
+ "step": 818
+ },
+ {
+ "epoch": 11.377292576419213,
+ "grad_norm": 0.017076566815376282,
+ "learning_rate": 0.0006,
+ "loss": 5.005850791931152,
+ "step": 819
+ },
+ {
+ "epoch": 11.391266375545852,
+ "grad_norm": 0.0172084029763937,
+ "learning_rate": 0.0006,
+ "loss": 4.903355598449707,
+ "step": 820
+ },
+ {
+ "epoch": 11.405240174672489,
+ "grad_norm": 0.0180258359760046,
+ "learning_rate": 0.0006,
+ "loss": 4.842667579650879,
+ "step": 821
+ },
+ {
+ "epoch": 11.419213973799126,
+ "grad_norm": 0.022130120545625687,
+ "learning_rate": 0.0006,
+ "loss": 4.799970626831055,
+ "step": 822
+ },
+ {
+ "epoch": 11.433187772925764,
+ "grad_norm": 0.02384897693991661,
+ "learning_rate": 0.0006,
+ "loss": 4.860746383666992,
+ "step": 823
+ },
+ {
+ "epoch": 11.447161572052401,
+ "grad_norm": 0.021881572902202606,
+ "learning_rate": 0.0006,
+ "loss": 4.918107032775879,
+ "step": 824
+ },
+ {
+ "epoch": 11.46113537117904,
+ "grad_norm": 0.023085080087184906,
+ "learning_rate": 0.0006,
+ "loss": 4.878946304321289,
+ "step": 825
+ },
+ {
+ "epoch": 11.475109170305677,
+ "grad_norm": 0.02291012555360794,
+ "learning_rate": 0.0006,
+ "loss": 4.958887100219727,
+ "step": 826
+ },
+ {
+ "epoch": 11.489082969432314,
+ "grad_norm": 0.0199968870729208,
+ "learning_rate": 0.0006,
+ "loss": 4.80955696105957,
+ "step": 827
+ },
+ {
+ "epoch": 11.503056768558952,
+ "grad_norm": 0.018702229484915733,
+ "learning_rate": 0.0006,
+ "loss": 5.0276408195495605,
+ "step": 828
+ },
+ {
+ "epoch": 11.51703056768559,
+ "grad_norm": 0.017787311226129532,
+ "learning_rate": 0.0006,
+ "loss": 4.96018123626709,
+ "step": 829
+ },
+ {
+ "epoch": 11.531004366812226,
+ "grad_norm": 0.01704501546919346,
+ "learning_rate": 0.0006,
+ "loss": 4.914553165435791,
+ "step": 830
+ },
+ {
+ "epoch": 11.544978165938865,
+ "grad_norm": 0.017519677057862282,
+ "learning_rate": 0.0006,
+ "loss": 4.798098087310791,
+ "step": 831
+ },
+ {
+ "epoch": 11.558951965065502,
+ "grad_norm": 0.018511991947889328,
+ "learning_rate": 0.0006,
+ "loss": 4.817785263061523,
+ "step": 832
+ },
+ {
+ "epoch": 11.57292576419214,
+ "grad_norm": 0.01876644790172577,
+ "learning_rate": 0.0006,
+ "loss": 4.818995475769043,
+ "step": 833
+ },
+ {
+ "epoch": 11.586899563318777,
+ "grad_norm": 0.01753568835556507,
+ "learning_rate": 0.0006,
+ "loss": 4.811938285827637,
+ "step": 834
+ },
+ {
+ "epoch": 11.600873362445414,
+ "grad_norm": 0.017046531662344933,
+ "learning_rate": 0.0006,
+ "loss": 4.914674282073975,
+ "step": 835
+ },
+ {
+ "epoch": 11.614847161572053,
+ "grad_norm": 0.01802094094455242,
+ "learning_rate": 0.0006,
+ "loss": 4.993284225463867,
+ "step": 836
+ },
+ {
+ "epoch": 11.62882096069869,
+ "grad_norm": 0.019493764266371727,
+ "learning_rate": 0.0006,
+ "loss": 4.895628452301025,
+ "step": 837
+ },
+ {
+ "epoch": 11.642794759825328,
+ "grad_norm": 0.019461210817098618,
+ "learning_rate": 0.0006,
+ "loss": 4.869269371032715,
+ "step": 838
+ },
+ {
+ "epoch": 11.656768558951965,
+ "grad_norm": 0.01981567218899727,
+ "learning_rate": 0.0006,
+ "loss": 4.863052845001221,
+ "step": 839
+ },
+ {
+ "epoch": 11.670742358078602,
+ "grad_norm": 0.019894491881132126,
+ "learning_rate": 0.0006,
+ "loss": 4.8775787353515625,
+ "step": 840
+ },
+ {
+ "epoch": 11.68471615720524,
+ "grad_norm": 0.01875786855816841,
+ "learning_rate": 0.0006,
+ "loss": 4.817763328552246,
+ "step": 841
+ },
+ {
+ "epoch": 11.698689956331878,
+ "grad_norm": 0.020926695317029953,
+ "learning_rate": 0.0006,
+ "loss": 4.7470526695251465,
+ "step": 842
+ },
+ {
+ "epoch": 11.712663755458514,
+ "grad_norm": 0.021477103233337402,
+ "learning_rate": 0.0006,
+ "loss": 4.762682914733887,
+ "step": 843
+ },
+ {
+ "epoch": 11.726637554585153,
+ "grad_norm": 0.020628711208701134,
+ "learning_rate": 0.0006,
+ "loss": 4.807985782623291,
+ "step": 844
+ },
+ {
+ "epoch": 11.74061135371179,
+ "grad_norm": 0.021489612758159637,
+ "learning_rate": 0.0006,
+ "loss": 4.843091011047363,
+ "step": 845
+ },
+ {
+ "epoch": 11.754585152838429,
+ "grad_norm": 0.021194491535425186,
+ "learning_rate": 0.0006,
+ "loss": 4.8300251960754395,
+ "step": 846
+ },
+ {
+ "epoch": 11.768558951965066,
+ "grad_norm": 0.01833350397646427,
+ "learning_rate": 0.0006,
+ "loss": 5.000202178955078,
+ "step": 847
+ },
+ {
+ "epoch": 11.782532751091702,
+ "grad_norm": 0.018096931278705597,
+ "learning_rate": 0.0006,
+ "loss": 4.880648136138916,
+ "step": 848
+ },
+ {
+ "epoch": 11.796506550218341,
+ "grad_norm": 0.0197161752730608,
+ "learning_rate": 0.0006,
+ "loss": 4.788316249847412,
+ "step": 849
+ },
+ {
+ "epoch": 11.810480349344978,
+ "grad_norm": 0.0203181654214859,
+ "learning_rate": 0.0006,
+ "loss": 4.870944976806641,
+ "step": 850
+ },
+ {
+ "epoch": 11.824454148471617,
+ "grad_norm": 0.01908455230295658,
+ "learning_rate": 0.0006,
+ "loss": 4.890144348144531,
+ "step": 851
+ },
+ {
+ "epoch": 11.838427947598253,
+ "grad_norm": 0.021281801164150238,
+ "learning_rate": 0.0006,
+ "loss": 4.89578914642334,
+ "step": 852
+ },
+ {
+ "epoch": 11.85240174672489,
+ "grad_norm": 0.02124273031949997,
+ "learning_rate": 0.0006,
+ "loss": 4.883749008178711,
+ "step": 853
+ },
+ {
+ "epoch": 11.866375545851529,
+ "grad_norm": 0.020969398319721222,
+ "learning_rate": 0.0006,
+ "loss": 4.809142589569092,
+ "step": 854
+ },
+ {
+ "epoch": 11.880349344978166,
+ "grad_norm": 0.01967657171189785,
+ "learning_rate": 0.0006,
+ "loss": 4.821864604949951,
+ "step": 855
+ },
+ {
+ "epoch": 11.894323144104803,
+ "grad_norm": 0.020887333899736404,
+ "learning_rate": 0.0006,
+ "loss": 4.917985916137695,
+ "step": 856
+ },
+ {
+ "epoch": 11.908296943231441,
+ "grad_norm": 0.019203083589673042,
+ "learning_rate": 0.0006,
+ "loss": 4.824960708618164,
+ "step": 857
+ },
+ {
+ "epoch": 11.922270742358078,
+ "grad_norm": 0.020162401720881462,
+ "learning_rate": 0.0006,
+ "loss": 4.772617816925049,
+ "step": 858
+ },
+ {
+ "epoch": 11.936244541484717,
+ "grad_norm": 0.02251487225294113,
+ "learning_rate": 0.0006,
+ "loss": 4.735418319702148,
+ "step": 859
+ },
+ {
+ "epoch": 11.950218340611354,
+ "grad_norm": 0.020313527435064316,
+ "learning_rate": 0.0006,
+ "loss": 4.939233779907227,
+ "step": 860
+ },
+ {
+ "epoch": 11.96419213973799,
+ "grad_norm": 0.021635599434375763,
+ "learning_rate": 0.0006,
+ "loss": 4.861807346343994,
+ "step": 861
+ },
+ {
+ "epoch": 11.97816593886463,
+ "grad_norm": 0.02285311557352543,
+ "learning_rate": 0.0006,
+ "loss": 4.932732582092285,
+ "step": 862
+ },
+ {
+ "epoch": 11.992139737991266,
+ "grad_norm": 0.02072535641491413,
+ "learning_rate": 0.0006,
+ "loss": 4.796767711639404,
+ "step": 863
+ },
+ {
+ "epoch": 12.0,
+ "grad_norm": 0.020588304847478867,
+ "learning_rate": 0.0006,
+ "loss": 4.997537612915039,
+ "step": 864
+ },
+ {
+ "epoch": 12.0,
+ "eval_loss": 5.044651985168457,
+ "eval_runtime": 56.9145,
+ "eval_samples_per_second": 42.906,
+ "eval_steps_per_second": 1.353,
+ "step": 864
+ },
+ {
+ "epoch": 12.013973799126637,
+ "grad_norm": 0.017597166821360588,
+ "learning_rate": 0.0006,
+ "loss": 4.830524444580078,
+ "step": 865
+ },
+ {
+ "epoch": 12.027947598253276,
+ "grad_norm": 0.01716788113117218,
+ "learning_rate": 0.0006,
+ "loss": 4.894325256347656,
+ "step": 866
+ },
+ {
+ "epoch": 12.041921397379912,
+ "grad_norm": 0.017712457105517387,
+ "learning_rate": 0.0006,
+ "loss": 4.868297576904297,
+ "step": 867
+ },
+ {
+ "epoch": 12.055895196506551,
+ "grad_norm": 0.021486390382051468,
+ "learning_rate": 0.0006,
+ "loss": 4.922049522399902,
+ "step": 868
+ },
+ {
+ "epoch": 12.069868995633188,
+ "grad_norm": 0.024560654535889626,
+ "learning_rate": 0.0006,
+ "loss": 4.734241008758545,
+ "step": 869
+ },
+ {
+ "epoch": 12.083842794759825,
+ "grad_norm": 0.025870252400636673,
+ "learning_rate": 0.0006,
+ "loss": 4.868566989898682,
+ "step": 870
+ },
+ {
+ "epoch": 12.097816593886463,
+ "grad_norm": 0.0277389045804739,
+ "learning_rate": 0.0006,
+ "loss": 4.755006790161133,
+ "step": 871
+ },
+ {
+ "epoch": 12.1117903930131,
+ "grad_norm": 0.029100844636559486,
+ "learning_rate": 0.0006,
+ "loss": 4.909667015075684,
+ "step": 872
+ },
+ {
+ "epoch": 12.125764192139737,
+ "grad_norm": 0.031833067536354065,
+ "learning_rate": 0.0006,
+ "loss": 4.858674049377441,
+ "step": 873
+ },
+ {
+ "epoch": 12.139737991266376,
+ "grad_norm": 0.032709237188100815,
+ "learning_rate": 0.0006,
+ "loss": 4.823696136474609,
+ "step": 874
+ },
+ {
+ "epoch": 12.153711790393013,
+ "grad_norm": 0.02995380386710167,
+ "learning_rate": 0.0006,
+ "loss": 4.850235462188721,
+ "step": 875
+ },
+ {
+ "epoch": 12.167685589519651,
+ "grad_norm": 0.034098610281944275,
+ "learning_rate": 0.0006,
+ "loss": 4.8169989585876465,
+ "step": 876
+ },
+ {
+ "epoch": 12.181659388646288,
+ "grad_norm": 0.03377080336213112,
+ "learning_rate": 0.0006,
+ "loss": 4.90738582611084,
+ "step": 877
+ },
+ {
+ "epoch": 12.195633187772925,
+ "grad_norm": 0.029780825600028038,
+ "learning_rate": 0.0006,
+ "loss": 4.811163902282715,
+ "step": 878
+ },
+ {
+ "epoch": 12.209606986899564,
+ "grad_norm": 0.03118024580180645,
+ "learning_rate": 0.0006,
+ "loss": 4.861286640167236,
+ "step": 879
+ },
+ {
+ "epoch": 12.2235807860262,
+ "grad_norm": 0.033052004873752594,
+ "learning_rate": 0.0006,
+ "loss": 4.865021228790283,
+ "step": 880
+ },
+ {
+ "epoch": 12.237554585152838,
+ "grad_norm": 0.029070032760500908,
+ "learning_rate": 0.0006,
+ "loss": 4.947303771972656,
+ "step": 881
+ },
+ {
+ "epoch": 12.251528384279476,
+ "grad_norm": 0.028355715796351433,
+ "learning_rate": 0.0006,
+ "loss": 4.804648399353027,
+ "step": 882
+ },
+ {
+ "epoch": 12.265502183406113,
+ "grad_norm": 0.031257227063179016,
+ "learning_rate": 0.0006,
+ "loss": 4.757718086242676,
+ "step": 883
+ },
+ {
+ "epoch": 12.279475982532752,
+ "grad_norm": 0.031438373029232025,
+ "learning_rate": 0.0006,
+ "loss": 4.851268768310547,
+ "step": 884
+ },
+ {
+ "epoch": 12.293449781659389,
+ "grad_norm": 0.030492138117551804,
+ "learning_rate": 0.0006,
+ "loss": 4.7891035079956055,
+ "step": 885
+ },
+ {
+ "epoch": 12.307423580786025,
+ "grad_norm": 0.026115991175174713,
+ "learning_rate": 0.0006,
+ "loss": 4.86474609375,
+ "step": 886
+ },
+ {
+ "epoch": 12.321397379912664,
+ "grad_norm": 0.024184465408325195,
+ "learning_rate": 0.0006,
+ "loss": 4.860456943511963,
+ "step": 887
+ },
+ {
+ "epoch": 12.335371179039301,
+ "grad_norm": 0.025437982752919197,
+ "learning_rate": 0.0006,
+ "loss": 4.806251049041748,
+ "step": 888
+ },
+ {
+ "epoch": 12.34934497816594,
+ "grad_norm": 0.022390231490135193,
+ "learning_rate": 0.0006,
+ "loss": 4.807086944580078,
+ "step": 889
+ },
+ {
+ "epoch": 12.363318777292577,
+ "grad_norm": 0.020174294710159302,
+ "learning_rate": 0.0006,
+ "loss": 4.78292179107666,
+ "step": 890
+ },
+ {
+ "epoch": 12.377292576419213,
+ "grad_norm": 0.019734324887394905,
+ "learning_rate": 0.0006,
+ "loss": 4.784509658813477,
+ "step": 891
+ },
+ {
+ "epoch": 12.391266375545852,
+ "grad_norm": 0.016733523458242416,
+ "learning_rate": 0.0006,
+ "loss": 4.811755180358887,
+ "step": 892
+ },
+ {
+ "epoch": 12.405240174672489,
+ "grad_norm": 0.01570146717131138,
+ "learning_rate": 0.0006,
+ "loss": 4.720088005065918,
+ "step": 893
+ },
+ {
+ "epoch": 12.419213973799126,
+ "grad_norm": 0.01601393334567547,
+ "learning_rate": 0.0006,
+ "loss": 4.783543109893799,
+ "step": 894
+ },
+ {
+ "epoch": 12.433187772925764,
+ "grad_norm": 0.014035487547516823,
+ "learning_rate": 0.0006,
+ "loss": 4.848340034484863,
+ "step": 895
+ },
+ {
+ "epoch": 12.447161572052401,
+ "grad_norm": 0.015248487703502178,
+ "learning_rate": 0.0006,
+ "loss": 4.773287773132324,
+ "step": 896
+ },
+ {
+ "epoch": 12.46113537117904,
+ "grad_norm": 0.01532268151640892,
+ "learning_rate": 0.0006,
+ "loss": 4.799646854400635,
+ "step": 897
+ },
+ {
+ "epoch": 12.475109170305677,
+ "grad_norm": 0.013982534408569336,
+ "learning_rate": 0.0006,
+ "loss": 4.8090500831604,
+ "step": 898
+ },
+ {
+ "epoch": 12.489082969432314,
+ "grad_norm": 0.014081164263188839,
+ "learning_rate": 0.0006,
+ "loss": 4.751888751983643,
+ "step": 899
+ },
+ {
+ "epoch": 12.503056768558952,
+ "grad_norm": 0.01380133256316185,
+ "learning_rate": 0.0006,
+ "loss": 4.78144645690918,
+ "step": 900
+ },
+ {
+ "epoch": 12.51703056768559,
+ "grad_norm": 0.013555269688367844,
+ "learning_rate": 0.0006,
+ "loss": 4.779257774353027,
+ "step": 901
+ },
+ {
+ "epoch": 12.531004366812226,
+ "grad_norm": 0.013075701892375946,
+ "learning_rate": 0.0006,
+ "loss": 4.820705413818359,
+ "step": 902
+ },
+ {
+ "epoch": 12.544978165938865,
+ "grad_norm": 0.01276316400617361,
+ "learning_rate": 0.0006,
+ "loss": 4.836733818054199,
+ "step": 903
+ },
+ {
+ "epoch": 12.558951965065502,
+ "grad_norm": 0.013033892959356308,
+ "learning_rate": 0.0006,
+ "loss": 4.764178276062012,
+ "step": 904
+ },
+ {
+ "epoch": 12.57292576419214,
+ "grad_norm": 0.013996980153024197,
+ "learning_rate": 0.0006,
+ "loss": 4.767983436584473,
+ "step": 905
+ },
+ {
+ "epoch": 12.586899563318777,
+ "grad_norm": 0.015222184360027313,
+ "learning_rate": 0.0006,
+ "loss": 4.759721755981445,
+ "step": 906
+ },
+ {
+ "epoch": 12.600873362445414,
+ "grad_norm": 0.01623941957950592,
+ "learning_rate": 0.0006,
+ "loss": 4.849811553955078,
+ "step": 907
+ },
+ {
+ "epoch": 12.614847161572053,
+ "grad_norm": 0.015711182728409767,
+ "learning_rate": 0.0006,
+ "loss": 4.7829084396362305,
+ "step": 908
+ },
+ {
+ "epoch": 12.62882096069869,
+ "grad_norm": 0.014137467369437218,
+ "learning_rate": 0.0006,
+ "loss": 4.73321533203125,
+ "step": 909
+ },
+ {
+ "epoch": 12.642794759825328,
+ "grad_norm": 0.012738462537527084,
+ "learning_rate": 0.0006,
+ "loss": 4.870052337646484,
+ "step": 910
+ },
+ {
+ "epoch": 12.656768558951965,
+ "grad_norm": 0.013683130033314228,
+ "learning_rate": 0.0006,
+ "loss": 4.82658052444458,
+ "step": 911
+ },
+ {
+ "epoch": 12.670742358078602,
+ "grad_norm": 0.014409830793738365,
+ "learning_rate": 0.0006,
+ "loss": 4.707494735717773,
+ "step": 912
+ },
+ {
+ "epoch": 12.68471615720524,
+ "grad_norm": 0.013806809671223164,
+ "learning_rate": 0.0006,
+ "loss": 4.778928279876709,
+ "step": 913
+ },
+ {
+ "epoch": 12.698689956331878,
+ "grad_norm": 0.012857912108302116,
+ "learning_rate": 0.0006,
+ "loss": 4.890717506408691,
+ "step": 914
+ },
+ {
+ "epoch": 12.712663755458514,
+ "grad_norm": 0.013828758150339127,
+ "learning_rate": 0.0006,
+ "loss": 4.8357439041137695,
+ "step": 915
+ },
+ {
+ "epoch": 12.726637554585153,
+ "grad_norm": 0.01478351280093193,
+ "learning_rate": 0.0006,
+ "loss": 4.869986534118652,
+ "step": 916
+ },
+ {
+ "epoch": 12.74061135371179,
+ "grad_norm": 0.01326421182602644,
+ "learning_rate": 0.0006,
+ "loss": 4.830646514892578,
+ "step": 917
+ },
+ {
+ "epoch": 12.754585152838429,
+ "grad_norm": 0.013335433788597584,
+ "learning_rate": 0.0006,
+ "loss": 4.7786149978637695,
+ "step": 918
+ },
+ {
+ "epoch": 12.768558951965066,
+ "grad_norm": 0.014261147007346153,
+ "learning_rate": 0.0006,
+ "loss": 4.788693428039551,
+ "step": 919
+ },
+ {
+ "epoch": 12.782532751091702,
+ "grad_norm": 0.014274783432483673,
+ "learning_rate": 0.0006,
+ "loss": 4.77295446395874,
+ "step": 920
+ },
+ {
+ "epoch": 12.796506550218341,
+ "grad_norm": 0.015074674971401691,
+ "learning_rate": 0.0006,
+ "loss": 4.830754280090332,
+ "step": 921
+ },
+ {
+ "epoch": 12.810480349344978,
+ "grad_norm": 0.015599515289068222,
+ "learning_rate": 0.0006,
+ "loss": 4.837759017944336,
+ "step": 922
+ },
+ {
+ "epoch": 12.824454148471617,
+ "grad_norm": 0.014595242217183113,
+ "learning_rate": 0.0006,
+ "loss": 4.882965564727783,
+ "step": 923
+ },
+ {
+ "epoch": 12.838427947598253,
+ "grad_norm": 0.014951467514038086,
+ "learning_rate": 0.0006,
+ "loss": 4.709748268127441,
+ "step": 924
+ },
+ {
+ "epoch": 12.85240174672489,
+ "grad_norm": 0.01539295818656683,
+ "learning_rate": 0.0006,
+ "loss": 4.791160583496094,
+ "step": 925
+ },
+ {
+ "epoch": 12.866375545851529,
+ "grad_norm": 0.015987906605005264,
+ "learning_rate": 0.0006,
+ "loss": 4.661876678466797,
+ "step": 926
+ },
+ {
+ "epoch": 12.880349344978166,
+ "grad_norm": 0.016527341678738594,
+ "learning_rate": 0.0006,
+ "loss": 4.721952438354492,
+ "step": 927
+ },
+ {
+ "epoch": 12.894323144104803,
+ "grad_norm": 0.01637161523103714,
+ "learning_rate": 0.0006,
+ "loss": 4.704446792602539,
+ "step": 928
+ },
+ {
+ "epoch": 12.908296943231441,
+ "grad_norm": 0.015221050009131432,
+ "learning_rate": 0.0006,
+ "loss": 4.771907806396484,
+ "step": 929
+ },
+ {
+ "epoch": 12.922270742358078,
+ "grad_norm": 0.016242166981101036,
+ "learning_rate": 0.0006,
+ "loss": 4.806267738342285,
+ "step": 930
+ },
+ {
+ "epoch": 12.936244541484717,
+ "grad_norm": 0.01628357358276844,
+ "learning_rate": 0.0006,
+ "loss": 4.7252326011657715,
+ "step": 931
+ },
+ {
+ "epoch": 12.950218340611354,
+ "grad_norm": 0.017428133636713028,
+ "learning_rate": 0.0006,
+ "loss": 4.765689849853516,
+ "step": 932
+ },
+ {
+ "epoch": 12.96419213973799,
+ "grad_norm": 0.01870061084628105,
+ "learning_rate": 0.0006,
+ "loss": 4.814488887786865,
+ "step": 933
+ },
+ {
+ "epoch": 12.97816593886463,
+ "grad_norm": 0.019796060398221016,
+ "learning_rate": 0.0006,
+ "loss": 4.739185810089111,
+ "step": 934
+ },
+ {
+ "epoch": 12.992139737991266,
+ "grad_norm": 0.01985820196568966,
+ "learning_rate": 0.0006,
+ "loss": 4.810608863830566,
+ "step": 935
+ },
+ {
+ "epoch": 13.0,
+ "grad_norm": 0.020834477618336678,
+ "learning_rate": 0.0006,
+ "loss": 4.727858066558838,
+ "step": 936
+ },
+ {
+ "epoch": 13.0,
+ "eval_loss": 4.979116916656494,
+ "eval_runtime": 57.2877,
+ "eval_samples_per_second": 42.627,
+ "eval_steps_per_second": 1.344,
+ "step": 936
+ },
+ {
+ "epoch": 13.013973799126637,
+ "grad_norm": 0.02025519125163555,
+ "learning_rate": 0.0006,
+ "loss": 4.628875732421875,
+ "step": 937
+ },
+ {
+ "epoch": 13.027947598253276,
+ "grad_norm": 0.019596368074417114,
+ "learning_rate": 0.0006,
+ "loss": 4.694699287414551,
+ "step": 938
+ },
+ {
+ "epoch": 13.041921397379912,
+ "grad_norm": 0.020494934171438217,
+ "learning_rate": 0.0006,
+ "loss": 4.691821098327637,
+ "step": 939
+ },
+ {
+ "epoch": 13.055895196506551,
+ "grad_norm": 0.020121760666370392,
+ "learning_rate": 0.0006,
+ "loss": 4.71478271484375,
+ "step": 940
+ },
+ {
+ "epoch": 13.069868995633188,
+ "grad_norm": 0.021091405302286148,
+ "learning_rate": 0.0006,
+ "loss": 4.695915222167969,
+ "step": 941
+ },
+ {
+ "epoch": 13.083842794759825,
+ "grad_norm": 0.019787423312664032,
+ "learning_rate": 0.0006,
+ "loss": 4.8652448654174805,
+ "step": 942
+ },
+ {
+ "epoch": 13.097816593886463,
+ "grad_norm": 0.0185833927243948,
+ "learning_rate": 0.0006,
+ "loss": 4.7065887451171875,
+ "step": 943
+ },
+ {
+ "epoch": 13.1117903930131,
+ "grad_norm": 0.021220475435256958,
+ "learning_rate": 0.0006,
+ "loss": 4.741232395172119,
+ "step": 944
+ },
+ {
+ "epoch": 13.125764192139737,
+ "grad_norm": 0.02136622555553913,
+ "learning_rate": 0.0006,
+ "loss": 4.705967903137207,
+ "step": 945
+ },
+ {
+ "epoch": 13.139737991266376,
+ "grad_norm": 0.0191587433218956,
+ "learning_rate": 0.0006,
+ "loss": 4.758667945861816,
+ "step": 946
+ },
+ {
+ "epoch": 13.153711790393013,
+ "grad_norm": 0.018290022388100624,
+ "learning_rate": 0.0006,
+ "loss": 4.760666847229004,
+ "step": 947
+ },
+ {
+ "epoch": 13.167685589519651,
+ "grad_norm": 0.021704409271478653,
+ "learning_rate": 0.0006,
+ "loss": 4.750555992126465,
+ "step": 948
+ },
+ {
+ "epoch": 13.181659388646288,
+ "grad_norm": 0.021668918430805206,
+ "learning_rate": 0.0006,
+ "loss": 4.840903282165527,
+ "step": 949
+ },
+ {
+ "epoch": 13.195633187772925,
+ "grad_norm": 0.019497141242027283,
+ "learning_rate": 0.0006,
+ "loss": 4.738969326019287,
+ "step": 950
+ },
+ {
+ "epoch": 13.209606986899564,
+ "grad_norm": 0.018500596284866333,
+ "learning_rate": 0.0006,
+ "loss": 4.75154972076416,
+ "step": 951
+ },
+ {
+ "epoch": 13.2235807860262,
+ "grad_norm": 0.01872239261865616,
+ "learning_rate": 0.0006,
+ "loss": 4.811738014221191,
+ "step": 952
+ },
+ {
+ "epoch": 13.237554585152838,
+ "grad_norm": 0.020321547985076904,
+ "learning_rate": 0.0006,
+ "loss": 4.815384387969971,
+ "step": 953
+ },
+ {
+ "epoch": 13.251528384279476,
+ "grad_norm": 0.02283575013279915,
+ "learning_rate": 0.0006,
+ "loss": 4.74582052230835,
+ "step": 954
+ },
+ {
+ "epoch": 13.265502183406113,
+ "grad_norm": 0.02474398910999298,
+ "learning_rate": 0.0006,
+ "loss": 4.738375186920166,
+ "step": 955
+ },
+ {
+ "epoch": 13.279475982532752,
+ "grad_norm": 0.02065764181315899,
+ "learning_rate": 0.0006,
+ "loss": 4.746875286102295,
+ "step": 956
+ },
+ {
+ "epoch": 13.293449781659389,
+ "grad_norm": 0.020534520968794823,
+ "learning_rate": 0.0006,
+ "loss": 4.656914710998535,
+ "step": 957
+ },
+ {
+ "epoch": 13.307423580786025,
+ "grad_norm": 0.0206378772854805,
+ "learning_rate": 0.0006,
+ "loss": 4.842615604400635,
+ "step": 958
+ },
+ {
+ "epoch": 13.321397379912664,
+ "grad_norm": 0.017935654148459435,
+ "learning_rate": 0.0006,
+ "loss": 4.717827796936035,
+ "step": 959
+ },
+ {
+ "epoch": 13.335371179039301,
+ "grad_norm": 0.016958363354206085,
+ "learning_rate": 0.0006,
+ "loss": 4.681070327758789,
+ "step": 960
+ },
+ {
+ "epoch": 13.34934497816594,
+ "grad_norm": 0.015735412016510963,
+ "learning_rate": 0.0006,
+ "loss": 4.7957048416137695,
+ "step": 961
+ },
+ {
+ "epoch": 13.363318777292577,
+ "grad_norm": 0.01563699170947075,
+ "learning_rate": 0.0006,
+ "loss": 4.701192378997803,
+ "step": 962
+ },
+ {
+ "epoch": 13.377292576419213,
+ "grad_norm": 0.015207760035991669,
+ "learning_rate": 0.0006,
+ "loss": 4.73256254196167,
+ "step": 963
+ },
+ {
+ "epoch": 13.391266375545852,
+ "grad_norm": 0.015396883711218834,
+ "learning_rate": 0.0006,
+ "loss": 4.76412296295166,
+ "step": 964
+ },
+ {
+ "epoch": 13.405240174672489,
+ "grad_norm": 0.015690583735704422,
+ "learning_rate": 0.0006,
+ "loss": 4.791157245635986,
+ "step": 965
+ },
+ {
+ "epoch": 13.419213973799126,
+ "grad_norm": 0.01755724661052227,
+ "learning_rate": 0.0006,
+ "loss": 4.845666885375977,
+ "step": 966
+ },
+ {
+ "epoch": 13.433187772925764,
+ "grad_norm": 0.019443422555923462,
+ "learning_rate": 0.0006,
+ "loss": 4.5902299880981445,
+ "step": 967
+ },
+ {
+ "epoch": 13.447161572052401,
+ "grad_norm": 0.02118338644504547,
+ "learning_rate": 0.0006,
+ "loss": 4.740627288818359,
+ "step": 968
+ },
+ {
+ "epoch": 13.46113537117904,
+ "grad_norm": 0.02174839936196804,
+ "learning_rate": 0.0006,
+ "loss": 4.814560890197754,
+ "step": 969
+ },
+ {
+ "epoch": 13.475109170305677,
+ "grad_norm": 0.022860374301671982,
+ "learning_rate": 0.0006,
+ "loss": 4.66440486907959,
+ "step": 970
+ },
+ {
+ "epoch": 13.489082969432314,
+ "grad_norm": 0.024742677807807922,
+ "learning_rate": 0.0006,
+ "loss": 4.7451653480529785,
+ "step": 971
+ },
+ {
+ "epoch": 13.503056768558952,
+ "grad_norm": 0.027068907395005226,
+ "learning_rate": 0.0006,
+ "loss": 4.802584171295166,
+ "step": 972
+ },
+ {
+ "epoch": 13.51703056768559,
+ "grad_norm": 0.026316625997424126,
+ "learning_rate": 0.0006,
+ "loss": 4.842397689819336,
+ "step": 973
+ },
+ {
+ "epoch": 13.531004366812226,
+ "grad_norm": 0.024610962718725204,
+ "learning_rate": 0.0006,
+ "loss": 4.650721073150635,
+ "step": 974
+ },
+ {
+ "epoch": 13.544978165938865,
+ "grad_norm": 0.024122603237628937,
+ "learning_rate": 0.0006,
+ "loss": 4.6984357833862305,
+ "step": 975
+ },
+ {
+ "epoch": 13.558951965065502,
+ "grad_norm": 0.022699708119034767,
+ "learning_rate": 0.0006,
+ "loss": 4.793007850646973,
+ "step": 976
+ },
+ {
+ "epoch": 13.57292576419214,
+ "grad_norm": 0.021706534549593925,
+ "learning_rate": 0.0006,
+ "loss": 4.691362380981445,
+ "step": 977
+ },
+ {
+ "epoch": 13.586899563318777,
+ "grad_norm": 0.021363291889429092,
+ "learning_rate": 0.0006,
+ "loss": 4.76762580871582,
+ "step": 978
+ },
+ {
+ "epoch": 13.600873362445414,
+ "grad_norm": 0.023476677015423775,
+ "learning_rate": 0.0006,
+ "loss": 4.777202606201172,
+ "step": 979
+ },
+ {
+ "epoch": 13.614847161572053,
+ "grad_norm": 0.02457100711762905,
+ "learning_rate": 0.0006,
+ "loss": 4.590703010559082,
+ "step": 980
+ },
+ {
+ "epoch": 13.62882096069869,
+ "grad_norm": 0.02469099499285221,
+ "learning_rate": 0.0006,
+ "loss": 4.640185356140137,
+ "step": 981
+ },
+ {
+ "epoch": 13.642794759825328,
+ "grad_norm": 0.024914603680372238,
+ "learning_rate": 0.0006,
+ "loss": 4.772201061248779,
+ "step": 982
+ },
+ {
+ "epoch": 13.656768558951965,
+ "grad_norm": 0.025133173912763596,
+ "learning_rate": 0.0006,
+ "loss": 4.743337154388428,
+ "step": 983
+ },
+ {
+ "epoch": 13.670742358078602,
+ "grad_norm": 0.02034137211740017,
+ "learning_rate": 0.0006,
+ "loss": 4.752127647399902,
+ "step": 984
+ },
+ {
+ "epoch": 13.68471615720524,
+ "grad_norm": 0.019645661115646362,
+ "learning_rate": 0.0006,
+ "loss": 4.838453769683838,
+ "step": 985
+ },
+ {
+ "epoch": 13.698689956331878,
+ "grad_norm": 0.01900297962129116,
+ "learning_rate": 0.0006,
+ "loss": 4.804868221282959,
+ "step": 986
+ },
+ {
+ "epoch": 13.712663755458514,
+ "grad_norm": 0.020292341709136963,
+ "learning_rate": 0.0006,
+ "loss": 4.609743118286133,
+ "step": 987
+ },
+ {
+ "epoch": 13.726637554585153,
+ "grad_norm": 0.017014775425195694,
+ "learning_rate": 0.0006,
+ "loss": 4.740612983703613,
+ "step": 988
+ },
+ {
+ "epoch": 13.74061135371179,
+ "grad_norm": 0.015199844725430012,
+ "learning_rate": 0.0006,
+ "loss": 4.719701766967773,
+ "step": 989
+ },
+ {
+ "epoch": 13.754585152838429,
+ "grad_norm": 0.015125414356589317,
+ "learning_rate": 0.0006,
+ "loss": 4.686114311218262,
+ "step": 990
+ },
+ {
+ "epoch": 13.768558951965066,
+ "grad_norm": 0.014779540710151196,
+ "learning_rate": 0.0006,
+ "loss": 4.674347877502441,
+ "step": 991
+ },
+ {
+ "epoch": 13.782532751091702,
+ "grad_norm": 0.015255163423717022,
+ "learning_rate": 0.0006,
+ "loss": 4.6309309005737305,
+ "step": 992
+ },
+ {
+ "epoch": 13.796506550218341,
+ "grad_norm": 0.013692197389900684,
+ "learning_rate": 0.0006,
+ "loss": 4.698497295379639,
+ "step": 993
+ },
+ {
+ "epoch": 13.810480349344978,
+ "grad_norm": 0.014529623091220856,
+ "learning_rate": 0.0006,
+ "loss": 4.628443717956543,
+ "step": 994
+ },
+ {
+ "epoch": 13.824454148471617,
+ "grad_norm": 0.01576659083366394,
+ "learning_rate": 0.0006,
+ "loss": 4.788015842437744,
+ "step": 995
+ },
+ {
+ "epoch": 13.838427947598253,
+ "grad_norm": 0.015834277495741844,
+ "learning_rate": 0.0006,
+ "loss": 4.672703742980957,
+ "step": 996
+ },
+ {
+ "epoch": 13.85240174672489,
+ "grad_norm": 0.015539503656327724,
+ "learning_rate": 0.0006,
+ "loss": 4.731503486633301,
+ "step": 997
+ },
+ {
+ "epoch": 13.866375545851529,
+ "grad_norm": 0.013463746756315231,
+ "learning_rate": 0.0006,
+ "loss": 4.7006330490112305,
+ "step": 998
+ },
+ {
+ "epoch": 13.880349344978166,
+ "grad_norm": 0.016377726569771767,
+ "learning_rate": 0.0006,
+ "loss": 4.636074066162109,
+ "step": 999
+ },
+ {
+ "epoch": 13.894323144104803,
+ "grad_norm": 0.016660034656524658,
+ "learning_rate": 0.0006,
+ "loss": 4.742693901062012,
+ "step": 1000
+ },
+ {
+ "epoch": 13.908296943231441,
+ "grad_norm": 0.014902938157320023,
+ "learning_rate": 0.0006,
+ "loss": 4.769811630249023,
+ "step": 1001
+ },
+ {
+ "epoch": 13.922270742358078,
+ "grad_norm": 0.014883476309478283,
+ "learning_rate": 0.0006,
+ "loss": 4.731159210205078,
+ "step": 1002
+ },
+ {
+ "epoch": 13.936244541484717,
+ "grad_norm": 0.014494217932224274,
+ "learning_rate": 0.0006,
+ "loss": 4.659485816955566,
+ "step": 1003
+ },
+ {
+ "epoch": 13.950218340611354,
+ "grad_norm": 0.015426991507411003,
+ "learning_rate": 0.0006,
+ "loss": 4.755229949951172,
+ "step": 1004
+ },
+ {
+ "epoch": 13.96419213973799,
+ "grad_norm": 0.016648586839437485,
+ "learning_rate": 0.0006,
+ "loss": 4.74111270904541,
+ "step": 1005
+ },
+ {
+ "epoch": 13.97816593886463,
+ "grad_norm": 0.016980541869997978,
+ "learning_rate": 0.0006,
+ "loss": 4.74072265625,
+ "step": 1006
+ },
+ {
+ "epoch": 13.992139737991266,
+ "grad_norm": 0.018070152029395103,
+ "learning_rate": 0.0006,
+ "loss": 4.732975006103516,
+ "step": 1007
+ },
+ {
+ "epoch": 14.0,
+ "grad_norm": 0.019910665228962898,
+ "learning_rate": 0.0006,
+ "loss": 4.634697914123535,
+ "step": 1008
+ },
+ {
+ "epoch": 14.0,
+ "eval_loss": 4.850640296936035,
+ "eval_runtime": 57.3039,
+ "eval_samples_per_second": 42.615,
+ "eval_steps_per_second": 1.344,
+ "step": 1008
+ },
+ {
+ "epoch": 14.013973799126637,
+ "grad_norm": 0.01980053074657917,
+ "learning_rate": 0.0006,
+ "loss": 4.628744602203369,
+ "step": 1009
+ },
+ {
+ "epoch": 14.027947598253276,
+ "grad_norm": 0.02497623674571514,
+ "learning_rate": 0.0006,
+ "loss": 4.75930118560791,
+ "step": 1010
+ },
+ {
+ "epoch": 14.041921397379912,
+ "grad_norm": 0.02794097363948822,
+ "learning_rate": 0.0006,
+ "loss": 4.691340446472168,
+ "step": 1011
+ },
+ {
+ "epoch": 14.055895196506551,
+ "grad_norm": 0.023725535720586777,
+ "learning_rate": 0.0006,
+ "loss": 4.733664512634277,
+ "step": 1012
+ },
+ {
+ "epoch": 14.069868995633188,
+ "grad_norm": 0.022966474294662476,
+ "learning_rate": 0.0006,
+ "loss": 4.664808750152588,
+ "step": 1013
+ },
+ {
+ "epoch": 14.083842794759825,
+ "grad_norm": 0.02530563250184059,
+ "learning_rate": 0.0006,
+ "loss": 4.8047285079956055,
+ "step": 1014
+ },
+ {
+ "epoch": 14.097816593886463,
+ "grad_norm": 0.02491082064807415,
+ "learning_rate": 0.0006,
+ "loss": 4.667956352233887,
+ "step": 1015
+ },
+ {
+ "epoch": 14.1117903930131,
+ "grad_norm": 0.022110294550657272,
+ "learning_rate": 0.0006,
+ "loss": 4.666140556335449,
+ "step": 1016
+ },
+ {
+ "epoch": 14.125764192139737,
+ "grad_norm": 0.02065849117934704,
+ "learning_rate": 0.0006,
+ "loss": 4.581960678100586,
+ "step": 1017
+ },
+ {
+ "epoch": 14.139737991266376,
+ "grad_norm": 0.01941298507153988,
+ "learning_rate": 0.0006,
+ "loss": 4.614226341247559,
+ "step": 1018
+ },
+ {
+ "epoch": 14.153711790393013,
+ "grad_norm": 0.022441856563091278,
+ "learning_rate": 0.0006,
+ "loss": 4.638352394104004,
+ "step": 1019
+ },
+ {
+ "epoch": 14.167685589519651,
+ "grad_norm": 0.021519694477319717,
+ "learning_rate": 0.0006,
+ "loss": 4.700160980224609,
+ "step": 1020
+ },
+ {
+ "epoch": 14.181659388646288,
+ "grad_norm": 0.018956486135721207,
+ "learning_rate": 0.0006,
+ "loss": 4.649593353271484,
+ "step": 1021
+ },
+ {
+ "epoch": 14.195633187772925,
+ "grad_norm": 0.016681140288710594,
+ "learning_rate": 0.0006,
+ "loss": 4.603952407836914,
+ "step": 1022
+ },
+ {
+ "epoch": 14.209606986899564,
+ "grad_norm": 0.01788809522986412,
+ "learning_rate": 0.0006,
+ "loss": 4.669160842895508,
+ "step": 1023
+ },
+ {
+ "epoch": 14.2235807860262,
+ "grad_norm": 0.016047228127717972,
+ "learning_rate": 0.0006,
+ "loss": 4.6367950439453125,
+ "step": 1024
+ },
+ {
+ "epoch": 14.237554585152838,
+ "grad_norm": 0.01614239625632763,
+ "learning_rate": 0.0006,
+ "loss": 4.664529323577881,
+ "step": 1025
+ },
+ {
+ "epoch": 14.251528384279476,
+ "grad_norm": 0.016362549737095833,
+ "learning_rate": 0.0006,
+ "loss": 4.702373504638672,
+ "step": 1026
+ },
+ {
+ "epoch": 14.265502183406113,
+ "grad_norm": 0.015800409018993378,
+ "learning_rate": 0.0006,
+ "loss": 4.7594428062438965,
+ "step": 1027
+ },
+ {
+ "epoch": 14.279475982532752,
+ "grad_norm": 0.016730139032006264,
+ "learning_rate": 0.0006,
+ "loss": 4.677244186401367,
+ "step": 1028
+ },
+ {
+ "epoch": 14.293449781659389,
+ "grad_norm": 0.01661759801208973,
+ "learning_rate": 0.0006,
+ "loss": 4.633679389953613,
+ "step": 1029
+ },
+ {
+ "epoch": 14.307423580786025,
+ "grad_norm": 0.017673347145318985,
+ "learning_rate": 0.0006,
+ "loss": 4.649107456207275,
+ "step": 1030
+ },
+ {
+ "epoch": 14.321397379912664,
+ "grad_norm": 0.019243352115154266,
+ "learning_rate": 0.0006,
+ "loss": 4.611257553100586,
+ "step": 1031
+ },
+ {
+ "epoch": 14.335371179039301,
+ "grad_norm": 0.019817164167761803,
+ "learning_rate": 0.0006,
+ "loss": 4.588086128234863,
+ "step": 1032
+ },
+ {
+ "epoch": 14.34934497816594,
+ "grad_norm": 0.022292211651802063,
+ "learning_rate": 0.0006,
+ "loss": 4.783066749572754,
+ "step": 1033
+ },
+ {
+ "epoch": 14.363318777292577,
+ "grad_norm": 0.02140096202492714,
+ "learning_rate": 0.0006,
+ "loss": 4.604718208312988,
+ "step": 1034
+ },
+ {
+ "epoch": 14.377292576419213,
+ "grad_norm": 0.0187817495316267,
+ "learning_rate": 0.0006,
+ "loss": 4.721773147583008,
+ "step": 1035
+ },
+ {
+ "epoch": 14.391266375545852,
+ "grad_norm": 0.019879210740327835,
+ "learning_rate": 0.0006,
+ "loss": 4.789286136627197,
+ "step": 1036
+ },
+ {
+ "epoch": 14.405240174672489,
+ "grad_norm": 0.01885785162448883,
+ "learning_rate": 0.0006,
+ "loss": 4.669743537902832,
+ "step": 1037
+ },
+ {
+ "epoch": 14.419213973799126,
+ "grad_norm": 0.018324270844459534,
+ "learning_rate": 0.0006,
+ "loss": 4.65901517868042,
+ "step": 1038
+ },
+ {
+ "epoch": 14.433187772925764,
+ "grad_norm": 0.017664821818470955,
+ "learning_rate": 0.0006,
+ "loss": 4.667839527130127,
+ "step": 1039
+ },
+ {
+ "epoch": 14.447161572052401,
+ "grad_norm": 0.015702905133366585,
+ "learning_rate": 0.0006,
+ "loss": 4.64700984954834,
+ "step": 1040
+ },
+ {
+ "epoch": 14.46113537117904,
+ "grad_norm": 0.016674194484949112,
+ "learning_rate": 0.0006,
+ "loss": 4.694093227386475,
+ "step": 1041
+ },
+ {
+ "epoch": 14.475109170305677,
+ "grad_norm": 0.016869477927684784,
+ "learning_rate": 0.0006,
+ "loss": 4.619544982910156,
+ "step": 1042
+ },
+ {
+ "epoch": 14.489082969432314,
+ "grad_norm": 0.016557540744543076,
+ "learning_rate": 0.0006,
+ "loss": 4.568667411804199,
+ "step": 1043
+ },
+ {
+ "epoch": 14.503056768558952,
+ "grad_norm": 0.014500945806503296,
+ "learning_rate": 0.0006,
+ "loss": 4.561741828918457,
+ "step": 1044
+ },
+ {
+ "epoch": 14.51703056768559,
+ "grad_norm": 0.014228739775717258,
+ "learning_rate": 0.0006,
+ "loss": 4.654277324676514,
+ "step": 1045
+ },
+ {
+ "epoch": 14.531004366812226,
+ "grad_norm": 0.013894238509237766,
+ "learning_rate": 0.0006,
+ "loss": 4.5867719650268555,
+ "step": 1046
+ },
+ {
+ "epoch": 14.544978165938865,
+ "grad_norm": 0.01550068985670805,
+ "learning_rate": 0.0006,
+ "loss": 4.61168098449707,
+ "step": 1047
+ },
+ {
+ "epoch": 14.558951965065502,
+ "grad_norm": 0.015940772369503975,
+ "learning_rate": 0.0006,
+ "loss": 4.691247940063477,
+ "step": 1048
+ },
+ {
+ "epoch": 14.57292576419214,
+ "grad_norm": 0.01714535616338253,
+ "learning_rate": 0.0006,
+ "loss": 4.61183500289917,
+ "step": 1049
+ },
+ {
+ "epoch": 14.586899563318777,
+ "grad_norm": 0.017925038933753967,
+ "learning_rate": 0.0006,
+ "loss": 4.719022750854492,
+ "step": 1050
+ },
+ {
+ "epoch": 14.600873362445414,
+ "grad_norm": 0.018141552805900574,
+ "learning_rate": 0.0006,
+ "loss": 4.760281562805176,
+ "step": 1051
+ },
+ {
+ "epoch": 14.614847161572053,
+ "grad_norm": 0.017131328582763672,
+ "learning_rate": 0.0006,
+ "loss": 4.709680080413818,
+ "step": 1052
+ },
+ {
+ "epoch": 14.62882096069869,
+ "grad_norm": 0.018730048090219498,
+ "learning_rate": 0.0006,
+ "loss": 4.5994720458984375,
+ "step": 1053
+ },
+ {
+ "epoch": 14.642794759825328,
+ "grad_norm": 0.018894298002123833,
+ "learning_rate": 0.0006,
+ "loss": 4.669070720672607,
+ "step": 1054
+ },
+ {
+ "epoch": 14.656768558951965,
+ "grad_norm": 0.017643198370933533,
+ "learning_rate": 0.0006,
+ "loss": 4.641790390014648,
+ "step": 1055
+ },
+ {
+ "epoch": 14.670742358078602,
+ "grad_norm": 0.017647244036197662,
+ "learning_rate": 0.0006,
+ "loss": 4.653899192810059,
+ "step": 1056
+ },
+ {
+ "epoch": 14.68471615720524,
+ "grad_norm": 0.017304198816418648,
+ "learning_rate": 0.0006,
+ "loss": 4.7487473487854,
+ "step": 1057
+ },
+ {
+ "epoch": 14.698689956331878,
+ "grad_norm": 0.01638447679579258,
+ "learning_rate": 0.0006,
+ "loss": 4.71945858001709,
+ "step": 1058
+ },
+ {
+ "epoch": 14.712663755458514,
+ "grad_norm": 0.015892071649432182,
+ "learning_rate": 0.0006,
+ "loss": 4.744903564453125,
+ "step": 1059
+ },
+ {
+ "epoch": 14.726637554585153,
+ "grad_norm": 0.017420614138245583,
+ "learning_rate": 0.0006,
+ "loss": 4.781563758850098,
+ "step": 1060
+ },
+ {
+ "epoch": 14.74061135371179,
+ "grad_norm": 0.016498520970344543,
+ "learning_rate": 0.0006,
+ "loss": 4.622779846191406,
+ "step": 1061
+ },
+ {
+ "epoch": 14.754585152838429,
+ "grad_norm": 0.017482127994298935,
+ "learning_rate": 0.0006,
+ "loss": 4.62825870513916,
+ "step": 1062
+ },
+ {
+ "epoch": 14.768558951965066,
+ "grad_norm": 0.019018379971385002,
+ "learning_rate": 0.0006,
+ "loss": 4.693194389343262,
+ "step": 1063
+ },
+ {
+ "epoch": 14.782532751091702,
+ "grad_norm": 0.01993914321064949,
+ "learning_rate": 0.0006,
+ "loss": 4.742903709411621,
+ "step": 1064
+ },
+ {
+ "epoch": 14.796506550218341,
+ "grad_norm": 0.02301335707306862,
+ "learning_rate": 0.0006,
+ "loss": 4.613251686096191,
+ "step": 1065
+ },
+ {
+ "epoch": 14.810480349344978,
+ "grad_norm": 0.024088222533464432,
+ "learning_rate": 0.0006,
+ "loss": 4.594305992126465,
+ "step": 1066
+ },
+ {
+ "epoch": 14.824454148471617,
+ "grad_norm": 0.02493320405483246,
+ "learning_rate": 0.0006,
+ "loss": 4.700588226318359,
+ "step": 1067
+ },
+ {
+ "epoch": 14.838427947598253,
+ "grad_norm": 0.02501937560737133,
+ "learning_rate": 0.0006,
+ "loss": 4.601742744445801,
+ "step": 1068
+ },
+ {
+ "epoch": 14.85240174672489,
+ "grad_norm": 0.023490751162171364,
+ "learning_rate": 0.0006,
+ "loss": 4.589211463928223,
+ "step": 1069
+ },
+ {
+ "epoch": 14.866375545851529,
+ "grad_norm": 0.019695665687322617,
+ "learning_rate": 0.0006,
+ "loss": 4.635081768035889,
+ "step": 1070
+ },
+ {
+ "epoch": 14.880349344978166,
+ "grad_norm": 0.019276980310678482,
+ "learning_rate": 0.0006,
+ "loss": 4.737002372741699,
+ "step": 1071
+ },
+ {
+ "epoch": 14.894323144104803,
+ "grad_norm": 0.022430511191487312,
+ "learning_rate": 0.0006,
+ "loss": 4.808295726776123,
+ "step": 1072
+ },
+ {
+ "epoch": 14.908296943231441,
+ "grad_norm": 0.021543916314840317,
+ "learning_rate": 0.0006,
+ "loss": 4.559475421905518,
+ "step": 1073
+ },
+ {
+ "epoch": 14.922270742358078,
+ "grad_norm": 0.019885241985321045,
+ "learning_rate": 0.0006,
+ "loss": 4.647948741912842,
+ "step": 1074
+ },
+ {
+ "epoch": 14.936244541484717,
+ "grad_norm": 0.02115175500512123,
+ "learning_rate": 0.0006,
+ "loss": 4.680505752563477,
+ "step": 1075
+ },
+ {
+ "epoch": 14.950218340611354,
+ "grad_norm": 0.02156151458621025,
+ "learning_rate": 0.0006,
+ "loss": 4.673477649688721,
+ "step": 1076
+ },
+ {
+ "epoch": 14.96419213973799,
+ "grad_norm": 0.02184601128101349,
+ "learning_rate": 0.0006,
+ "loss": 4.644402027130127,
+ "step": 1077
+ },
+ {
+ "epoch": 14.97816593886463,
+ "grad_norm": 0.0211983360350132,
+ "learning_rate": 0.0006,
+ "loss": 4.821558475494385,
+ "step": 1078
+ },
+ {
+ "epoch": 14.992139737991266,
+ "grad_norm": 0.019636567682027817,
+ "learning_rate": 0.0006,
+ "loss": 4.578868865966797,
+ "step": 1079
+ },
+ {
+ "epoch": 15.0,
+ "grad_norm": 0.019997350871562958,
+ "learning_rate": 0.0006,
+ "loss": 4.691307544708252,
+ "step": 1080
+ },
+ {
+ "epoch": 15.0,
+ "eval_loss": 4.872478008270264,
+ "eval_runtime": 56.8513,
+ "eval_samples_per_second": 42.954,
+ "eval_steps_per_second": 1.354,
+ "step": 1080
+ },
+ {
+ "epoch": 15.013973799126637,
+ "grad_norm": 0.018830662593245506,
+ "learning_rate": 0.0006,
+ "loss": 4.761779308319092,
+ "step": 1081
+ },
+ {
+ "epoch": 15.027947598253276,
+ "grad_norm": 0.016876481473445892,
+ "learning_rate": 0.0006,
+ "loss": 4.61148738861084,
+ "step": 1082
+ },
+ {
+ "epoch": 15.041921397379912,
+ "grad_norm": 0.015550863929092884,
+ "learning_rate": 0.0006,
+ "loss": 4.676124095916748,
+ "step": 1083
+ },
+ {
+ "epoch": 15.055895196506551,
+ "grad_norm": 0.017178896814584732,
+ "learning_rate": 0.0006,
+ "loss": 4.588935852050781,
+ "step": 1084
+ },
+ {
+ "epoch": 15.069868995633188,
+ "grad_norm": 0.015465447679162025,
+ "learning_rate": 0.0006,
+ "loss": 4.605499267578125,
+ "step": 1085
+ },
+ {
+ "epoch": 15.083842794759825,
+ "grad_norm": 0.01606195978820324,
+ "learning_rate": 0.0006,
+ "loss": 4.723373889923096,
+ "step": 1086
+ },
+ {
+ "epoch": 15.097816593886463,
+ "grad_norm": 0.016521936282515526,
+ "learning_rate": 0.0006,
+ "loss": 4.646608352661133,
+ "step": 1087
+ },
+ {
+ "epoch": 15.1117903930131,
+ "grad_norm": 0.01589205674827099,
+ "learning_rate": 0.0006,
+ "loss": 4.498138427734375,
+ "step": 1088
+ },
+ {
+ "epoch": 15.125764192139737,
+ "grad_norm": 0.01524856686592102,
+ "learning_rate": 0.0006,
+ "loss": 4.594562530517578,
+ "step": 1089
+ },
+ {
+ "epoch": 15.139737991266376,
+ "grad_norm": 0.014714641496539116,
+ "learning_rate": 0.0006,
+ "loss": 4.630744934082031,
+ "step": 1090
+ },
+ {
+ "epoch": 15.153711790393013,
+ "grad_norm": 0.015333331190049648,
+ "learning_rate": 0.0006,
+ "loss": 4.682388782501221,
+ "step": 1091
+ },
+ {
+ "epoch": 15.167685589519651,
+ "grad_norm": 0.01550073828548193,
+ "learning_rate": 0.0006,
+ "loss": 4.692652702331543,
+ "step": 1092
+ },
+ {
+ "epoch": 15.181659388646288,
+ "grad_norm": 0.014783318154513836,
+ "learning_rate": 0.0006,
+ "loss": 4.6223835945129395,
+ "step": 1093
+ },
+ {
+ "epoch": 15.195633187772925,
+ "grad_norm": 0.014372751116752625,
+ "learning_rate": 0.0006,
+ "loss": 4.52687406539917,
+ "step": 1094
+ },
+ {
+ "epoch": 15.209606986899564,
+ "grad_norm": 0.014779130928218365,
+ "learning_rate": 0.0006,
+ "loss": 4.637009620666504,
+ "step": 1095
+ },
+ {
+ "epoch": 15.2235807860262,
+ "grad_norm": 0.015697073191404343,
+ "learning_rate": 0.0006,
+ "loss": 4.682981014251709,
+ "step": 1096
+ },
+ {
+ "epoch": 15.237554585152838,
+ "grad_norm": 0.016523664817214012,
+ "learning_rate": 0.0006,
+ "loss": 4.705722808837891,
+ "step": 1097
+ },
+ {
+ "epoch": 15.251528384279476,
+ "grad_norm": 0.017633996903896332,
+ "learning_rate": 0.0006,
+ "loss": 4.574686050415039,
+ "step": 1098
+ },
+ {
+ "epoch": 15.265502183406113,
+ "grad_norm": 0.019955076277256012,
+ "learning_rate": 0.0006,
+ "loss": 4.634450912475586,
+ "step": 1099
+ },
+ {
+ "epoch": 15.279475982532752,
+ "grad_norm": 0.02132618986070156,
+ "learning_rate": 0.0006,
+ "loss": 4.56110143661499,
+ "step": 1100
+ },
+ {
+ "epoch": 15.293449781659389,
+ "grad_norm": 0.0207593385130167,
+ "learning_rate": 0.0006,
+ "loss": 4.629396438598633,
+ "step": 1101
+ },
+ {
+ "epoch": 15.307423580786025,
+ "grad_norm": 0.017882540822029114,
+ "learning_rate": 0.0006,
+ "loss": 4.730963230133057,
+ "step": 1102
+ },
+ {
+ "epoch": 15.321397379912664,
+ "grad_norm": 0.01917876861989498,
+ "learning_rate": 0.0006,
+ "loss": 4.64585542678833,
+ "step": 1103
+ },
+ {
+ "epoch": 15.335371179039301,
+ "grad_norm": 0.01943974196910858,
+ "learning_rate": 0.0006,
+ "loss": 4.656181335449219,
+ "step": 1104
+ },
+ {
+ "epoch": 15.34934497816594,
+ "grad_norm": 0.019904715940356255,
+ "learning_rate": 0.0006,
+ "loss": 4.619329452514648,
+ "step": 1105
+ },
+ {
+ "epoch": 15.363318777292577,
+ "grad_norm": 0.019174346700310707,
+ "learning_rate": 0.0006,
+ "loss": 4.647892475128174,
+ "step": 1106
+ },
+ {
+ "epoch": 15.377292576419213,
+ "grad_norm": 0.018244758248329163,
+ "learning_rate": 0.0006,
+ "loss": 4.56166934967041,
+ "step": 1107
+ },
+ {
+ "epoch": 15.391266375545852,
+ "grad_norm": 0.018914205953478813,
+ "learning_rate": 0.0006,
+ "loss": 4.648063659667969,
+ "step": 1108
+ },
+ {
+ "epoch": 15.405240174672489,
+ "grad_norm": 0.01961071416735649,
+ "learning_rate": 0.0006,
+ "loss": 4.665525436401367,
+ "step": 1109
+ },
+ {
+ "epoch": 15.419213973799126,
+ "grad_norm": 0.01841311901807785,
+ "learning_rate": 0.0006,
+ "loss": 4.611342430114746,
+ "step": 1110
+ },
+ {
+ "epoch": 15.433187772925764,
+ "grad_norm": 0.01863427273929119,
+ "learning_rate": 0.0006,
+ "loss": 4.664256572723389,
+ "step": 1111
+ },
+ {
+ "epoch": 15.447161572052401,
+ "grad_norm": 0.020506877452135086,
+ "learning_rate": 0.0006,
+ "loss": 4.631237506866455,
+ "step": 1112
+ },
+ {
+ "epoch": 15.46113537117904,
+ "grad_norm": 0.02223833277821541,
+ "learning_rate": 0.0006,
+ "loss": 4.668654441833496,
+ "step": 1113
+ },
+ {
+ "epoch": 15.475109170305677,
+ "grad_norm": 0.023336702957749367,
+ "learning_rate": 0.0006,
+ "loss": 4.562833309173584,
+ "step": 1114
+ },
+ {
+ "epoch": 15.489082969432314,
+ "grad_norm": 0.023052413016557693,
+ "learning_rate": 0.0006,
+ "loss": 4.621532917022705,
+ "step": 1115
+ },
+ {
+ "epoch": 15.503056768558952,
+ "grad_norm": 0.023420924320816994,
+ "learning_rate": 0.0006,
+ "loss": 4.665465354919434,
+ "step": 1116
+ },
+ {
+ "epoch": 15.51703056768559,
+ "grad_norm": 0.024071281775832176,
+ "learning_rate": 0.0006,
+ "loss": 4.595980644226074,
+ "step": 1117
+ },
+ {
+ "epoch": 15.531004366812226,
+ "grad_norm": 0.023406682536005974,
+ "learning_rate": 0.0006,
+ "loss": 4.642635345458984,
+ "step": 1118
+ },
+ {
+ "epoch": 15.544978165938865,
+ "grad_norm": 0.02138647809624672,
+ "learning_rate": 0.0006,
+ "loss": 4.704513072967529,
+ "step": 1119
+ },
+ {
+ "epoch": 15.558951965065502,
+ "grad_norm": 0.0221384447067976,
+ "learning_rate": 0.0006,
+ "loss": 4.567503452301025,
+ "step": 1120
+ },
+ {
+ "epoch": 15.57292576419214,
+ "grad_norm": 0.020879851654171944,
+ "learning_rate": 0.0006,
+ "loss": 4.664057731628418,
+ "step": 1121
+ },
+ {
+ "epoch": 15.586899563318777,
+ "grad_norm": 0.021478967741131783,
+ "learning_rate": 0.0006,
+ "loss": 4.539627552032471,
+ "step": 1122
+ },
+ {
+ "epoch": 15.600873362445414,
+ "grad_norm": 0.022748656570911407,
+ "learning_rate": 0.0006,
+ "loss": 4.5654497146606445,
+ "step": 1123
+ },
+ {
+ "epoch": 15.614847161572053,
+ "grad_norm": 0.021870248019695282,
+ "learning_rate": 0.0006,
+ "loss": 4.56899356842041,
+ "step": 1124
+ },
+ {
+ "epoch": 15.62882096069869,
+ "grad_norm": 0.021711355075240135,
+ "learning_rate": 0.0006,
+ "loss": 4.657808303833008,
+ "step": 1125
+ },
+ {
+ "epoch": 15.642794759825328,
+ "grad_norm": 0.022810909897089005,
+ "learning_rate": 0.0006,
+ "loss": 4.6526689529418945,
+ "step": 1126
+ },
+ {
+ "epoch": 15.656768558951965,
+ "grad_norm": 0.022561343386769295,
+ "learning_rate": 0.0006,
+ "loss": 4.534887790679932,
+ "step": 1127
+ },
+ {
+ "epoch": 15.670742358078602,
+ "grad_norm": 0.019909802824258804,
+ "learning_rate": 0.0006,
+ "loss": 4.600790023803711,
+ "step": 1128
+ },
+ {
+ "epoch": 15.68471615720524,
+ "grad_norm": 0.020092325285077095,
+ "learning_rate": 0.0006,
+ "loss": 4.773830413818359,
+ "step": 1129
+ },
+ {
+ "epoch": 15.698689956331878,
+ "grad_norm": 0.018834445625543594,
+ "learning_rate": 0.0006,
+ "loss": 4.628868103027344,
+ "step": 1130
+ },
+ {
+ "epoch": 15.712663755458514,
+ "grad_norm": 0.017552798613905907,
+ "learning_rate": 0.0006,
+ "loss": 4.647500038146973,
+ "step": 1131
+ },
+ {
+ "epoch": 15.726637554585153,
+ "grad_norm": 0.01751251146197319,
+ "learning_rate": 0.0006,
+ "loss": 4.590545654296875,
+ "step": 1132
+ },
+ {
+ "epoch": 15.74061135371179,
+ "grad_norm": 0.01699099875986576,
+ "learning_rate": 0.0006,
+ "loss": 4.509699821472168,
+ "step": 1133
+ },
+ {
+ "epoch": 15.754585152838429,
+ "grad_norm": 0.015808649361133575,
+ "learning_rate": 0.0006,
+ "loss": 4.606992244720459,
+ "step": 1134
+ },
+ {
+ "epoch": 15.768558951965066,
+ "grad_norm": 0.01655443385243416,
+ "learning_rate": 0.0006,
+ "loss": 4.585219860076904,
+ "step": 1135
+ },
+ {
+ "epoch": 15.782532751091702,
+ "grad_norm": 0.01708163693547249,
+ "learning_rate": 0.0006,
+ "loss": 4.668147087097168,
+ "step": 1136
+ },
+ {
+ "epoch": 15.796506550218341,
+ "grad_norm": 0.017161913216114044,
+ "learning_rate": 0.0006,
+ "loss": 4.7170634269714355,
+ "step": 1137
+ },
+ {
+ "epoch": 15.810480349344978,
+ "grad_norm": 0.017262674868106842,
+ "learning_rate": 0.0006,
+ "loss": 4.572132110595703,
+ "step": 1138
+ },
+ {
+ "epoch": 15.824454148471617,
+ "grad_norm": 0.017713576555252075,
+ "learning_rate": 0.0006,
+ "loss": 4.702538967132568,
+ "step": 1139
+ },
+ {
+ "epoch": 15.838427947598253,
+ "grad_norm": 0.018730733543634415,
+ "learning_rate": 0.0006,
+ "loss": 4.629721641540527,
+ "step": 1140
+ },
+ {
+ "epoch": 15.85240174672489,
+ "grad_norm": 0.017960187047719955,
+ "learning_rate": 0.0006,
+ "loss": 4.661039352416992,
+ "step": 1141
+ },
+ {
+ "epoch": 15.866375545851529,
+ "grad_norm": 0.01750079356133938,
+ "learning_rate": 0.0006,
+ "loss": 4.503710746765137,
+ "step": 1142
+ },
+ {
+ "epoch": 15.880349344978166,
+ "grad_norm": 0.016630422323942184,
+ "learning_rate": 0.0006,
+ "loss": 4.6582136154174805,
+ "step": 1143
+ },
+ {
+ "epoch": 15.894323144104803,
+ "grad_norm": 0.015896832570433617,
+ "learning_rate": 0.0006,
+ "loss": 4.596693992614746,
+ "step": 1144
+ },
+ {
+ "epoch": 15.908296943231441,
+ "grad_norm": 0.01566407084465027,
+ "learning_rate": 0.0006,
+ "loss": 4.6337199211120605,
+ "step": 1145
+ },
+ {
+ "epoch": 15.922270742358078,
+ "grad_norm": 0.016823330894112587,
+ "learning_rate": 0.0006,
+ "loss": 4.686994552612305,
+ "step": 1146
+ },
+ {
+ "epoch": 15.936244541484717,
+ "grad_norm": 0.016208255663514137,
+ "learning_rate": 0.0006,
+ "loss": 4.6205549240112305,
+ "step": 1147
+ },
+ {
+ "epoch": 15.950218340611354,
+ "grad_norm": 0.0171707421541214,
+ "learning_rate": 0.0006,
+ "loss": 4.582968711853027,
+ "step": 1148
+ },
+ {
+ "epoch": 15.96419213973799,
+ "grad_norm": 0.017009446397423744,
+ "learning_rate": 0.0006,
+ "loss": 4.510274887084961,
+ "step": 1149
+ },
+ {
+ "epoch": 15.97816593886463,
+ "grad_norm": 0.01565811224281788,
+ "learning_rate": 0.0006,
+ "loss": 4.55834436416626,
+ "step": 1150
+ },
+ {
+ "epoch": 15.992139737991266,
+ "grad_norm": 0.015276037156581879,
+ "learning_rate": 0.0006,
+ "loss": 4.537870407104492,
+ "step": 1151
+ },
+ {
+ "epoch": 16.0,
+ "grad_norm": 0.017305882647633553,
+ "learning_rate": 0.0006,
+ "loss": 4.611305236816406,
+ "step": 1152
+ },
+ {
+ "epoch": 16.0,
+ "eval_loss": 4.775513172149658,
+ "eval_runtime": 56.7341,
+ "eval_samples_per_second": 43.043,
+ "eval_steps_per_second": 1.357,
+ "step": 1152
+ },
+ {
+ "epoch": 16.01397379912664,
+ "grad_norm": 0.016752436757087708,
+ "learning_rate": 0.0006,
+ "loss": 4.541312217712402,
+ "step": 1153
+ },
+ {
+ "epoch": 16.027947598253274,
+ "grad_norm": 0.01643134094774723,
+ "learning_rate": 0.0006,
+ "loss": 4.585402965545654,
+ "step": 1154
+ },
+ {
+ "epoch": 16.041921397379912,
+ "grad_norm": 0.01888720504939556,
+ "learning_rate": 0.0006,
+ "loss": 4.597825050354004,
+ "step": 1155
+ },
+ {
+ "epoch": 16.05589519650655,
+ "grad_norm": 0.022044632583856583,
+ "learning_rate": 0.0006,
+ "loss": 4.532772064208984,
+ "step": 1156
+ },
+ {
+ "epoch": 16.069868995633186,
+ "grad_norm": 0.022986987605690956,
+ "learning_rate": 0.0006,
+ "loss": 4.551351547241211,
+ "step": 1157
+ },
+ {
+ "epoch": 16.083842794759825,
+ "grad_norm": 0.020116791129112244,
+ "learning_rate": 0.0006,
+ "loss": 4.484149932861328,
+ "step": 1158
+ },
+ {
+ "epoch": 16.097816593886463,
+ "grad_norm": 0.021442247554659843,
+ "learning_rate": 0.0006,
+ "loss": 4.515316486358643,
+ "step": 1159
+ },
+ {
+ "epoch": 16.111790393013102,
+ "grad_norm": 0.020974772050976753,
+ "learning_rate": 0.0006,
+ "loss": 4.51606559753418,
+ "step": 1160
+ },
+ {
+ "epoch": 16.125764192139737,
+ "grad_norm": 0.021274864673614502,
+ "learning_rate": 0.0006,
+ "loss": 4.587329864501953,
+ "step": 1161
+ },
+ {
+ "epoch": 16.139737991266376,
+ "grad_norm": 0.022799784317612648,
+ "learning_rate": 0.0006,
+ "loss": 4.568149089813232,
+ "step": 1162
+ },
+ {
+ "epoch": 16.153711790393015,
+ "grad_norm": 0.021784933283925056,
+ "learning_rate": 0.0006,
+ "loss": 4.515336990356445,
+ "step": 1163
+ },
+ {
+ "epoch": 16.16768558951965,
+ "grad_norm": 0.024261271581053734,
+ "learning_rate": 0.0006,
+ "loss": 4.5127081871032715,
+ "step": 1164
+ },
+ {
+ "epoch": 16.18165938864629,
+ "grad_norm": 0.02248375490307808,
+ "learning_rate": 0.0006,
+ "loss": 4.754251003265381,
+ "step": 1165
+ },
+ {
+ "epoch": 16.195633187772927,
+ "grad_norm": 0.02136492356657982,
+ "learning_rate": 0.0006,
+ "loss": 4.58193826675415,
+ "step": 1166
+ },
+ {
+ "epoch": 16.209606986899562,
+ "grad_norm": 0.020727066323161125,
+ "learning_rate": 0.0006,
+ "loss": 4.564209938049316,
+ "step": 1167
+ },
+ {
+ "epoch": 16.2235807860262,
+ "grad_norm": 0.017789119854569435,
+ "learning_rate": 0.0006,
+ "loss": 4.698680877685547,
+ "step": 1168
+ },
+ {
+ "epoch": 16.23755458515284,
+ "grad_norm": 0.017603034153580666,
+ "learning_rate": 0.0006,
+ "loss": 4.501171588897705,
+ "step": 1169
+ },
+ {
+ "epoch": 16.251528384279474,
+ "grad_norm": 0.017345862463116646,
+ "learning_rate": 0.0006,
+ "loss": 4.593590259552002,
+ "step": 1170
+ },
+ {
+ "epoch": 16.265502183406113,
+ "grad_norm": 0.018047425895929337,
+ "learning_rate": 0.0006,
+ "loss": 4.597467422485352,
+ "step": 1171
+ },
+ {
+ "epoch": 16.27947598253275,
+ "grad_norm": 0.01947222836315632,
+ "learning_rate": 0.0006,
+ "loss": 4.5072221755981445,
+ "step": 1172
+ },
+ {
+ "epoch": 16.29344978165939,
+ "grad_norm": 0.019598500803112984,
+ "learning_rate": 0.0006,
+ "loss": 4.601351737976074,
+ "step": 1173
+ },
+ {
+ "epoch": 16.307423580786025,
+ "grad_norm": 0.020002514123916626,
+ "learning_rate": 0.0006,
+ "loss": 4.580431938171387,
+ "step": 1174
+ },
+ {
+ "epoch": 16.321397379912664,
+ "grad_norm": 0.0216530729085207,
+ "learning_rate": 0.0006,
+ "loss": 4.424681663513184,
+ "step": 1175
+ },
+ {
+ "epoch": 16.335371179039303,
+ "grad_norm": 0.01949235238134861,
+ "learning_rate": 0.0006,
+ "loss": 4.751883029937744,
+ "step": 1176
+ },
+ {
+ "epoch": 16.349344978165938,
+ "grad_norm": 0.017779016867280006,
+ "learning_rate": 0.0006,
+ "loss": 4.541490077972412,
+ "step": 1177
+ },
+ {
+ "epoch": 16.363318777292577,
+ "grad_norm": 0.01909121684730053,
+ "learning_rate": 0.0006,
+ "loss": 4.594141006469727,
+ "step": 1178
+ },
+ {
+ "epoch": 16.377292576419215,
+ "grad_norm": 0.019894849509000778,
+ "learning_rate": 0.0006,
+ "loss": 4.555070400238037,
+ "step": 1179
+ },
+ {
+ "epoch": 16.39126637554585,
+ "grad_norm": 0.01993032731115818,
+ "learning_rate": 0.0006,
+ "loss": 4.521401405334473,
+ "step": 1180
+ },
+ {
+ "epoch": 16.40524017467249,
+ "grad_norm": 0.0210788045078516,
+ "learning_rate": 0.0006,
+ "loss": 4.672842979431152,
+ "step": 1181
+ },
+ {
+ "epoch": 16.419213973799128,
+ "grad_norm": 0.02274598926305771,
+ "learning_rate": 0.0006,
+ "loss": 4.629504203796387,
+ "step": 1182
+ },
+ {
+ "epoch": 16.433187772925763,
+ "grad_norm": 0.020448120310902596,
+ "learning_rate": 0.0006,
+ "loss": 4.548209190368652,
+ "step": 1183
+ },
+ {
+ "epoch": 16.4471615720524,
+ "grad_norm": 0.017565643414855003,
+ "learning_rate": 0.0006,
+ "loss": 4.529054641723633,
+ "step": 1184
+ },
+ {
+ "epoch": 16.46113537117904,
+ "grad_norm": 0.019504237920045853,
+ "learning_rate": 0.0006,
+ "loss": 4.48182487487793,
+ "step": 1185
+ },
+ {
+ "epoch": 16.475109170305675,
+ "grad_norm": 0.016823742538690567,
+ "learning_rate": 0.0006,
+ "loss": 4.608696937561035,
+ "step": 1186
+ },
+ {
+ "epoch": 16.489082969432314,
+ "grad_norm": 0.01602894812822342,
+ "learning_rate": 0.0006,
+ "loss": 4.620062351226807,
+ "step": 1187
+ },
+ {
+ "epoch": 16.503056768558952,
+ "grad_norm": 0.016615642234683037,
+ "learning_rate": 0.0006,
+ "loss": 4.69378662109375,
+ "step": 1188
+ },
+ {
+ "epoch": 16.51703056768559,
+ "grad_norm": 0.018576730042696,
+ "learning_rate": 0.0006,
+ "loss": 4.537041664123535,
+ "step": 1189
+ },
+ {
+ "epoch": 16.531004366812226,
+ "grad_norm": 0.018611151725053787,
+ "learning_rate": 0.0006,
+ "loss": 4.496169090270996,
+ "step": 1190
+ },
+ {
+ "epoch": 16.544978165938865,
+ "grad_norm": 0.018622413277626038,
+ "learning_rate": 0.0006,
+ "loss": 4.557358264923096,
+ "step": 1191
+ },
+ {
+ "epoch": 16.558951965065503,
+ "grad_norm": 0.01823544315993786,
+ "learning_rate": 0.0006,
+ "loss": 4.730792999267578,
+ "step": 1192
+ },
+ {
+ "epoch": 16.57292576419214,
+ "grad_norm": 0.018796846270561218,
+ "learning_rate": 0.0006,
+ "loss": 4.781791687011719,
+ "step": 1193
+ },
+ {
+ "epoch": 16.586899563318777,
+ "grad_norm": 0.01728292554616928,
+ "learning_rate": 0.0006,
+ "loss": 4.515257835388184,
+ "step": 1194
+ },
+ {
+ "epoch": 16.600873362445416,
+ "grad_norm": 0.01699589006602764,
+ "learning_rate": 0.0006,
+ "loss": 4.5996904373168945,
+ "step": 1195
+ },
+ {
+ "epoch": 16.61484716157205,
+ "grad_norm": 0.015905970707535744,
+ "learning_rate": 0.0006,
+ "loss": 4.601301193237305,
+ "step": 1196
+ },
+ {
+ "epoch": 16.62882096069869,
+ "grad_norm": 0.01338571310043335,
+ "learning_rate": 0.0006,
+ "loss": 4.557708740234375,
+ "step": 1197
+ },
+ {
+ "epoch": 16.64279475982533,
+ "grad_norm": 0.015298855490982533,
+ "learning_rate": 0.0006,
+ "loss": 4.542287349700928,
+ "step": 1198
+ },
+ {
+ "epoch": 16.656768558951963,
+ "grad_norm": 0.014753367751836777,
+ "learning_rate": 0.0006,
+ "loss": 4.595787048339844,
+ "step": 1199
+ },
+ {
+ "epoch": 16.670742358078602,
+ "grad_norm": 0.015527483075857162,
+ "learning_rate": 0.0006,
+ "loss": 4.607529163360596,
+ "step": 1200
+ },
+ {
+ "epoch": 16.68471615720524,
+ "grad_norm": 0.015042162500321865,
+ "learning_rate": 0.0006,
+ "loss": 4.686626434326172,
+ "step": 1201
+ },
+ {
+ "epoch": 16.69868995633188,
+ "grad_norm": 0.0159393772482872,
+ "learning_rate": 0.0006,
+ "loss": 4.47656774520874,
+ "step": 1202
+ },
+ {
+ "epoch": 16.712663755458514,
+ "grad_norm": 0.01532787922769785,
+ "learning_rate": 0.0006,
+ "loss": 4.55678653717041,
+ "step": 1203
+ },
+ {
+ "epoch": 16.726637554585153,
+ "grad_norm": 0.016332164406776428,
+ "learning_rate": 0.0006,
+ "loss": 4.51441764831543,
+ "step": 1204
+ },
+ {
+ "epoch": 16.74061135371179,
+ "grad_norm": 0.018901726230978966,
+ "learning_rate": 0.0006,
+ "loss": 4.57227897644043,
+ "step": 1205
+ },
+ {
+ "epoch": 16.754585152838427,
+ "grad_norm": 0.018589433282613754,
+ "learning_rate": 0.0006,
+ "loss": 4.550070762634277,
+ "step": 1206
+ },
+ {
+ "epoch": 16.768558951965066,
+ "grad_norm": 0.01634034886956215,
+ "learning_rate": 0.0006,
+ "loss": 4.586910724639893,
+ "step": 1207
+ },
+ {
+ "epoch": 16.782532751091704,
+ "grad_norm": 0.017599433660507202,
+ "learning_rate": 0.0006,
+ "loss": 4.5891923904418945,
+ "step": 1208
+ },
+ {
+ "epoch": 16.79650655021834,
+ "grad_norm": 0.01878446340560913,
+ "learning_rate": 0.0006,
+ "loss": 4.57227087020874,
+ "step": 1209
+ },
+ {
+ "epoch": 16.810480349344978,
+ "grad_norm": 0.01803469844162464,
+ "learning_rate": 0.0006,
+ "loss": 4.583845138549805,
+ "step": 1210
+ },
+ {
+ "epoch": 16.824454148471617,
+ "grad_norm": 0.018505144864320755,
+ "learning_rate": 0.0006,
+ "loss": 4.561151504516602,
+ "step": 1211
+ },
+ {
+ "epoch": 16.83842794759825,
+ "grad_norm": 0.01920081488788128,
+ "learning_rate": 0.0006,
+ "loss": 4.506834030151367,
+ "step": 1212
+ },
+ {
+ "epoch": 16.85240174672489,
+ "grad_norm": 0.0177712831646204,
+ "learning_rate": 0.0006,
+ "loss": 4.557195663452148,
+ "step": 1213
+ },
+ {
+ "epoch": 16.86637554585153,
+ "grad_norm": 0.019113918766379356,
+ "learning_rate": 0.0006,
+ "loss": 4.651227951049805,
+ "step": 1214
+ },
+ {
+ "epoch": 16.880349344978168,
+ "grad_norm": 0.02012539468705654,
+ "learning_rate": 0.0006,
+ "loss": 4.4952239990234375,
+ "step": 1215
+ },
+ {
+ "epoch": 16.894323144104803,
+ "grad_norm": 0.019993646070361137,
+ "learning_rate": 0.0006,
+ "loss": 4.517433166503906,
+ "step": 1216
+ },
+ {
+ "epoch": 16.90829694323144,
+ "grad_norm": 0.020041441544890404,
+ "learning_rate": 0.0006,
+ "loss": 4.489407539367676,
+ "step": 1217
+ },
+ {
+ "epoch": 16.92227074235808,
+ "grad_norm": 0.0183611661195755,
+ "learning_rate": 0.0006,
+ "loss": 4.5435638427734375,
+ "step": 1218
+ },
+ {
+ "epoch": 16.936244541484715,
+ "grad_norm": 0.019138429313898087,
+ "learning_rate": 0.0006,
+ "loss": 4.545949459075928,
+ "step": 1219
+ },
+ {
+ "epoch": 16.950218340611354,
+ "grad_norm": 0.018554266542196274,
+ "learning_rate": 0.0006,
+ "loss": 4.579620361328125,
+ "step": 1220
+ },
+ {
+ "epoch": 16.964192139737992,
+ "grad_norm": 0.0193545650690794,
+ "learning_rate": 0.0006,
+ "loss": 4.588255405426025,
+ "step": 1221
+ },
+ {
+ "epoch": 16.978165938864628,
+ "grad_norm": 0.019389165565371513,
+ "learning_rate": 0.0006,
+ "loss": 4.520698547363281,
+ "step": 1222
+ },
+ {
+ "epoch": 16.992139737991266,
+ "grad_norm": 0.016601379960775375,
+ "learning_rate": 0.0006,
+ "loss": 4.4809064865112305,
+ "step": 1223
+ },
+ {
+ "epoch": 17.0,
+ "grad_norm": 0.01771303080022335,
+ "learning_rate": 0.0006,
+ "loss": 4.721122741699219,
+ "step": 1224
+ },
+ {
+ "epoch": 17.0,
+ "eval_loss": 4.7903666496276855,
+ "eval_runtime": 56.9826,
+ "eval_samples_per_second": 42.855,
+ "eval_steps_per_second": 1.351,
+ "step": 1224
+ },
+ {
+ "epoch": 17.01397379912664,
+ "grad_norm": 0.017720289528369904,
+ "learning_rate": 0.0006,
+ "loss": 4.523834705352783,
+ "step": 1225
+ },
+ {
+ "epoch": 17.027947598253274,
+ "grad_norm": 0.020137161016464233,
+ "learning_rate": 0.0006,
+ "loss": 4.574061870574951,
+ "step": 1226
+ },
+ {
+ "epoch": 17.041921397379912,
+ "grad_norm": 0.02002314105629921,
+ "learning_rate": 0.0006,
+ "loss": 4.512407302856445,
+ "step": 1227
+ },
+ {
+ "epoch": 17.05589519650655,
+ "grad_norm": 0.019628183916211128,
+ "learning_rate": 0.0006,
+ "loss": 4.502533435821533,
+ "step": 1228
+ },
+ {
+ "epoch": 17.069868995633186,
+ "grad_norm": 0.019808458164334297,
+ "learning_rate": 0.0006,
+ "loss": 4.463156700134277,
+ "step": 1229
+ },
+ {
+ "epoch": 17.083842794759825,
+ "grad_norm": 0.021250786259770393,
+ "learning_rate": 0.0006,
+ "loss": 4.622682094573975,
+ "step": 1230
+ },
+ {
+ "epoch": 17.097816593886463,
+ "grad_norm": 0.022031694650650024,
+ "learning_rate": 0.0006,
+ "loss": 4.451882839202881,
+ "step": 1231
+ },
+ {
+ "epoch": 17.111790393013102,
+ "grad_norm": 0.01898987777531147,
+ "learning_rate": 0.0006,
+ "loss": 4.529331684112549,
+ "step": 1232
+ },
+ {
+ "epoch": 17.125764192139737,
+ "grad_norm": 0.018749834969639778,
+ "learning_rate": 0.0006,
+ "loss": 4.435815334320068,
+ "step": 1233
+ },
+ {
+ "epoch": 17.139737991266376,
+ "grad_norm": 0.01896674931049347,
+ "learning_rate": 0.0006,
+ "loss": 4.484029293060303,
+ "step": 1234
+ },
+ {
+ "epoch": 17.153711790393015,
+ "grad_norm": 0.018286997452378273,
+ "learning_rate": 0.0006,
+ "loss": 4.523510932922363,
+ "step": 1235
+ },
+ {
+ "epoch": 17.16768558951965,
+ "grad_norm": 0.017273059114813805,
+ "learning_rate": 0.0006,
+ "loss": 4.451767444610596,
+ "step": 1236
+ },
+ {
+ "epoch": 17.18165938864629,
+ "grad_norm": 0.016499150544404984,
+ "learning_rate": 0.0006,
+ "loss": 4.5568647384643555,
+ "step": 1237
+ },
+ {
+ "epoch": 17.195633187772927,
+ "grad_norm": 0.015934469178318977,
+ "learning_rate": 0.0006,
+ "loss": 4.4639997482299805,
+ "step": 1238
+ },
+ {
+ "epoch": 17.209606986899562,
+ "grad_norm": 0.01714209094643593,
+ "learning_rate": 0.0006,
+ "loss": 4.547614097595215,
+ "step": 1239
+ },
+ {
+ "epoch": 17.2235807860262,
+ "grad_norm": 0.019167495891451836,
+ "learning_rate": 0.0006,
+ "loss": 4.571097373962402,
+ "step": 1240
+ },
+ {
+ "epoch": 17.23755458515284,
+ "grad_norm": 0.018276996910572052,
+ "learning_rate": 0.0006,
+ "loss": 4.489764213562012,
+ "step": 1241
+ },
+ {
+ "epoch": 17.251528384279474,
+ "grad_norm": 0.021296866238117218,
+ "learning_rate": 0.0006,
+ "loss": 4.526913642883301,
+ "step": 1242
+ },
+ {
+ "epoch": 17.265502183406113,
+ "grad_norm": 0.022606629878282547,
+ "learning_rate": 0.0006,
+ "loss": 4.521624565124512,
+ "step": 1243
+ },
+ {
+ "epoch": 17.27947598253275,
+ "grad_norm": 0.021204711869359016,
+ "learning_rate": 0.0006,
+ "loss": 4.500864028930664,
+ "step": 1244
+ },
+ {
+ "epoch": 17.29344978165939,
+ "grad_norm": 0.01988849975168705,
+ "learning_rate": 0.0006,
+ "loss": 4.45162296295166,
+ "step": 1245
+ },
+ {
+ "epoch": 17.307423580786025,
+ "grad_norm": 0.020268384367227554,
+ "learning_rate": 0.0006,
+ "loss": 4.536198616027832,
+ "step": 1246
+ },
+ {
+ "epoch": 17.321397379912664,
+ "grad_norm": 0.020185677334666252,
+ "learning_rate": 0.0006,
+ "loss": 4.440262794494629,
+ "step": 1247
+ },
+ {
+ "epoch": 17.335371179039303,
+ "grad_norm": 0.020106853917241096,
+ "learning_rate": 0.0006,
+ "loss": 4.475907325744629,
+ "step": 1248
+ },
+ {
+ "epoch": 17.349344978165938,
+ "grad_norm": 0.018883297219872475,
+ "learning_rate": 0.0006,
+ "loss": 4.640661716461182,
+ "step": 1249
+ },
+ {
+ "epoch": 17.363318777292577,
+ "grad_norm": 0.01852620206773281,
+ "learning_rate": 0.0006,
+ "loss": 4.578243255615234,
+ "step": 1250
+ },
+ {
+ "epoch": 17.377292576419215,
+ "grad_norm": 0.01886451430618763,
+ "learning_rate": 0.0006,
+ "loss": 4.546026229858398,
+ "step": 1251
+ },
+ {
+ "epoch": 17.39126637554585,
+ "grad_norm": 0.01777452416718006,
+ "learning_rate": 0.0006,
+ "loss": 4.521542549133301,
+ "step": 1252
+ },
+ {
+ "epoch": 17.40524017467249,
+ "grad_norm": 0.016626447439193726,
+ "learning_rate": 0.0006,
+ "loss": 4.52629280090332,
+ "step": 1253
+ },
+ {
+ "epoch": 17.419213973799128,
+ "grad_norm": 0.01599477231502533,
+ "learning_rate": 0.0006,
+ "loss": 4.4796037673950195,
+ "step": 1254
+ },
+ {
+ "epoch": 17.433187772925763,
+ "grad_norm": 0.017021458595991135,
+ "learning_rate": 0.0006,
+ "loss": 4.468813419342041,
+ "step": 1255
+ },
+ {
+ "epoch": 17.4471615720524,
+ "grad_norm": 0.016925333067774773,
+ "learning_rate": 0.0006,
+ "loss": 4.584898948669434,
+ "step": 1256
+ },
+ {
+ "epoch": 17.46113537117904,
+ "grad_norm": 0.01714998669922352,
+ "learning_rate": 0.0006,
+ "loss": 4.523544788360596,
+ "step": 1257
+ },
+ {
+ "epoch": 17.475109170305675,
+ "grad_norm": 0.01802288182079792,
+ "learning_rate": 0.0006,
+ "loss": 4.595581531524658,
+ "step": 1258
+ },
+ {
+ "epoch": 17.489082969432314,
+ "grad_norm": 0.0180289875715971,
+ "learning_rate": 0.0006,
+ "loss": 4.394440650939941,
+ "step": 1259
+ },
+ {
+ "epoch": 17.503056768558952,
+ "grad_norm": 0.018504245206713676,
+ "learning_rate": 0.0006,
+ "loss": 4.4794416427612305,
+ "step": 1260
+ },
+ {
+ "epoch": 17.51703056768559,
+ "grad_norm": 0.019973233342170715,
+ "learning_rate": 0.0006,
+ "loss": 4.4241437911987305,
+ "step": 1261
+ },
+ {
+ "epoch": 17.531004366812226,
+ "grad_norm": 0.020063083618879318,
+ "learning_rate": 0.0006,
+ "loss": 4.48068380355835,
+ "step": 1262
+ },
+ {
+ "epoch": 17.544978165938865,
+ "grad_norm": 0.02066640742123127,
+ "learning_rate": 0.0006,
+ "loss": 4.628993034362793,
+ "step": 1263
+ },
+ {
+ "epoch": 17.558951965065503,
+ "grad_norm": 0.02206416428089142,
+ "learning_rate": 0.0006,
+ "loss": 4.564790725708008,
+ "step": 1264
+ },
+ {
+ "epoch": 17.57292576419214,
+ "grad_norm": 0.020640335977077484,
+ "learning_rate": 0.0006,
+ "loss": 4.578705787658691,
+ "step": 1265
+ },
+ {
+ "epoch": 17.586899563318777,
+ "grad_norm": 0.019093405455350876,
+ "learning_rate": 0.0006,
+ "loss": 4.5023603439331055,
+ "step": 1266
+ },
+ {
+ "epoch": 17.600873362445416,
+ "grad_norm": 0.019459193572402,
+ "learning_rate": 0.0006,
+ "loss": 4.594531059265137,
+ "step": 1267
+ },
+ {
+ "epoch": 17.61484716157205,
+ "grad_norm": 0.02021171525120735,
+ "learning_rate": 0.0006,
+ "loss": 4.770299911499023,
+ "step": 1268
+ },
+ {
+ "epoch": 17.62882096069869,
+ "grad_norm": 0.021161191165447235,
+ "learning_rate": 0.0006,
+ "loss": 4.5380401611328125,
+ "step": 1269
+ },
+ {
+ "epoch": 17.64279475982533,
+ "grad_norm": 0.021013403311371803,
+ "learning_rate": 0.0006,
+ "loss": 4.455524921417236,
+ "step": 1270
+ },
+ {
+ "epoch": 17.656768558951963,
+ "grad_norm": 0.020336592569947243,
+ "learning_rate": 0.0006,
+ "loss": 4.533672332763672,
+ "step": 1271
+ },
+ {
+ "epoch": 17.670742358078602,
+ "grad_norm": 0.020159434527158737,
+ "learning_rate": 0.0006,
+ "loss": 4.596151351928711,
+ "step": 1272
+ },
+ {
+ "epoch": 17.68471615720524,
+ "grad_norm": 0.020432960242033005,
+ "learning_rate": 0.0006,
+ "loss": 4.55531644821167,
+ "step": 1273
+ },
+ {
+ "epoch": 17.69868995633188,
+ "grad_norm": 0.025466229766607285,
+ "learning_rate": 0.0006,
+ "loss": 4.566912651062012,
+ "step": 1274
+ },
+ {
+ "epoch": 17.712663755458514,
+ "grad_norm": 0.026537558063864708,
+ "learning_rate": 0.0006,
+ "loss": 4.526003837585449,
+ "step": 1275
+ },
+ {
+ "epoch": 17.726637554585153,
+ "grad_norm": 0.024397607892751694,
+ "learning_rate": 0.0006,
+ "loss": 4.541755676269531,
+ "step": 1276
+ },
+ {
+ "epoch": 17.74061135371179,
+ "grad_norm": 0.022062312811613083,
+ "learning_rate": 0.0006,
+ "loss": 4.6532793045043945,
+ "step": 1277
+ },
+ {
+ "epoch": 17.754585152838427,
+ "grad_norm": 0.02204282023012638,
+ "learning_rate": 0.0006,
+ "loss": 4.582825660705566,
+ "step": 1278
+ },
+ {
+ "epoch": 17.768558951965066,
+ "grad_norm": 0.020844416692852974,
+ "learning_rate": 0.0006,
+ "loss": 4.522336483001709,
+ "step": 1279
+ },
+ {
+ "epoch": 17.782532751091704,
+ "grad_norm": 0.019815703853964806,
+ "learning_rate": 0.0006,
+ "loss": 4.607398509979248,
+ "step": 1280
+ },
+ {
+ "epoch": 17.79650655021834,
+ "grad_norm": 0.016676954925060272,
+ "learning_rate": 0.0006,
+ "loss": 4.446206092834473,
+ "step": 1281
+ },
+ {
+ "epoch": 17.810480349344978,
+ "grad_norm": 0.0166452769190073,
+ "learning_rate": 0.0006,
+ "loss": 4.505931854248047,
+ "step": 1282
+ },
+ {
+ "epoch": 17.824454148471617,
+ "grad_norm": 0.015928907319903374,
+ "learning_rate": 0.0006,
+ "loss": 4.455267429351807,
+ "step": 1283
+ },
+ {
+ "epoch": 17.83842794759825,
+ "grad_norm": 0.015219560824334621,
+ "learning_rate": 0.0006,
+ "loss": 4.5623674392700195,
+ "step": 1284
+ },
+ {
+ "epoch": 17.85240174672489,
+ "grad_norm": 0.015125355683267117,
+ "learning_rate": 0.0006,
+ "loss": 4.554104804992676,
+ "step": 1285
+ },
+ {
+ "epoch": 17.86637554585153,
+ "grad_norm": 0.015612718649208546,
+ "learning_rate": 0.0006,
+ "loss": 4.484847068786621,
+ "step": 1286
+ },
+ {
+ "epoch": 17.880349344978168,
+ "grad_norm": 0.015215662308037281,
+ "learning_rate": 0.0006,
+ "loss": 4.4802093505859375,
+ "step": 1287
+ },
+ {
+ "epoch": 17.894323144104803,
+ "grad_norm": 0.01511684525758028,
+ "learning_rate": 0.0006,
+ "loss": 4.537298679351807,
+ "step": 1288
+ },
+ {
+ "epoch": 17.90829694323144,
+ "grad_norm": 0.013841322623193264,
+ "learning_rate": 0.0006,
+ "loss": 4.427488327026367,
+ "step": 1289
+ },
+ {
+ "epoch": 17.92227074235808,
+ "grad_norm": 0.015115964226424694,
+ "learning_rate": 0.0006,
+ "loss": 4.4547119140625,
+ "step": 1290
+ },
+ {
+ "epoch": 17.936244541484715,
+ "grad_norm": 0.016411006450653076,
+ "learning_rate": 0.0006,
+ "loss": 4.52994966506958,
+ "step": 1291
+ },
+ {
+ "epoch": 17.950218340611354,
+ "grad_norm": 0.017313152551651,
+ "learning_rate": 0.0006,
+ "loss": 4.5337653160095215,
+ "step": 1292
+ },
+ {
+ "epoch": 17.964192139737992,
+ "grad_norm": 0.016559796407818794,
+ "learning_rate": 0.0006,
+ "loss": 4.5277910232543945,
+ "step": 1293
+ },
+ {
+ "epoch": 17.978165938864628,
+ "grad_norm": 0.015854761004447937,
+ "learning_rate": 0.0006,
+ "loss": 4.6334309577941895,
+ "step": 1294
+ },
+ {
+ "epoch": 17.992139737991266,
+ "grad_norm": 0.015658238902688026,
+ "learning_rate": 0.0006,
+ "loss": 4.613372802734375,
+ "step": 1295
+ },
+ {
+ "epoch": 18.0,
+ "grad_norm": 0.0169155802577734,
+ "learning_rate": 0.0006,
+ "loss": 4.527356147766113,
+ "step": 1296
+ },
+ {
+ "epoch": 18.0,
+ "eval_loss": 4.764543056488037,
+ "eval_runtime": 56.5998,
+ "eval_samples_per_second": 43.145,
+ "eval_steps_per_second": 1.36,
+ "step": 1296
+ },
+ {
+ "epoch": 18.01397379912664,
+ "grad_norm": 0.015348963439464569,
+ "learning_rate": 0.0006,
+ "loss": 4.466423988342285,
+ "step": 1297
+ },
+ {
+ "epoch": 18.027947598253274,
+ "grad_norm": 0.01768588088452816,
+ "learning_rate": 0.0006,
+ "loss": 4.545933246612549,
+ "step": 1298
+ },
+ {
+ "epoch": 18.041921397379912,
+ "grad_norm": 0.017381226643919945,
+ "learning_rate": 0.0006,
+ "loss": 4.494060039520264,
+ "step": 1299
+ },
+ {
+ "epoch": 18.05589519650655,
+ "grad_norm": 0.019002335146069527,
+ "learning_rate": 0.0006,
+ "loss": 4.395200729370117,
+ "step": 1300
+ },
+ {
+ "epoch": 18.069868995633186,
+ "grad_norm": 0.01827932707965374,
+ "learning_rate": 0.0006,
+ "loss": 4.424848556518555,
+ "step": 1301
+ },
+ {
+ "epoch": 18.083842794759825,
+ "grad_norm": 0.016383031383156776,
+ "learning_rate": 0.0006,
+ "loss": 4.494722366333008,
+ "step": 1302
+ },
+ {
+ "epoch": 18.097816593886463,
+ "grad_norm": 0.016265157610177994,
+ "learning_rate": 0.0006,
+ "loss": 4.416522979736328,
+ "step": 1303
+ },
+ {
+ "epoch": 18.111790393013102,
+ "grad_norm": 0.01702386513352394,
+ "learning_rate": 0.0006,
+ "loss": 4.600103378295898,
+ "step": 1304
+ },
+ {
+ "epoch": 18.125764192139737,
+ "grad_norm": 0.018495498225092888,
+ "learning_rate": 0.0006,
+ "loss": 4.588401794433594,
+ "step": 1305
+ },
+ {
+ "epoch": 18.139737991266376,
+ "grad_norm": 0.017921684309840202,
+ "learning_rate": 0.0006,
+ "loss": 4.478075981140137,
+ "step": 1306
+ },
+ {
+ "epoch": 18.153711790393015,
+ "grad_norm": 0.01801278069615364,
+ "learning_rate": 0.0006,
+ "loss": 4.413186550140381,
+ "step": 1307
+ },
+ {
+ "epoch": 18.16768558951965,
+ "grad_norm": 0.017211275175213814,
+ "learning_rate": 0.0006,
+ "loss": 4.461341857910156,
+ "step": 1308
+ },
+ {
+ "epoch": 18.18165938864629,
+ "grad_norm": 0.017589328810572624,
+ "learning_rate": 0.0006,
+ "loss": 4.546969413757324,
+ "step": 1309
+ },
+ {
+ "epoch": 18.195633187772927,
+ "grad_norm": 0.01797698810696602,
+ "learning_rate": 0.0006,
+ "loss": 4.530797004699707,
+ "step": 1310
+ },
+ {
+ "epoch": 18.209606986899562,
+ "grad_norm": 0.020408619195222855,
+ "learning_rate": 0.0006,
+ "loss": 4.62999153137207,
+ "step": 1311
+ },
+ {
+ "epoch": 18.2235807860262,
+ "grad_norm": 0.01994038000702858,
+ "learning_rate": 0.0006,
+ "loss": 4.586811065673828,
+ "step": 1312
+ },
+ {
+ "epoch": 18.23755458515284,
+ "grad_norm": 0.019381340593099594,
+ "learning_rate": 0.0006,
+ "loss": 4.4769368171691895,
+ "step": 1313
+ },
+ {
+ "epoch": 18.251528384279474,
+ "grad_norm": 0.020484555512666702,
+ "learning_rate": 0.0006,
+ "loss": 4.524302959442139,
+ "step": 1314
+ },
+ {
+ "epoch": 18.265502183406113,
+ "grad_norm": 0.02426416426897049,
+ "learning_rate": 0.0006,
+ "loss": 4.561705589294434,
+ "step": 1315
+ },
+ {
+ "epoch": 18.27947598253275,
+ "grad_norm": 0.024748101830482483,
+ "learning_rate": 0.0006,
+ "loss": 4.4654316902160645,
+ "step": 1316
+ },
+ {
+ "epoch": 18.29344978165939,
+ "grad_norm": 0.020703328773379326,
+ "learning_rate": 0.0006,
+ "loss": 4.53033447265625,
+ "step": 1317
+ },
+ {
+ "epoch": 18.307423580786025,
+ "grad_norm": 0.01765458658337593,
+ "learning_rate": 0.0006,
+ "loss": 4.5154266357421875,
+ "step": 1318
+ },
+ {
+ "epoch": 18.321397379912664,
+ "grad_norm": 0.018646420910954475,
+ "learning_rate": 0.0006,
+ "loss": 4.5224714279174805,
+ "step": 1319
+ },
+ {
+ "epoch": 18.335371179039303,
+ "grad_norm": 0.01907474733889103,
+ "learning_rate": 0.0006,
+ "loss": 4.5055437088012695,
+ "step": 1320
+ },
+ {
+ "epoch": 18.349344978165938,
+ "grad_norm": 0.018179846927523613,
+ "learning_rate": 0.0006,
+ "loss": 4.462161064147949,
+ "step": 1321
+ },
+ {
+ "epoch": 18.363318777292577,
+ "grad_norm": 0.016697878018021584,
+ "learning_rate": 0.0006,
+ "loss": 4.473718643188477,
+ "step": 1322
+ },
+ {
+ "epoch": 18.377292576419215,
+ "grad_norm": 0.017504019662737846,
+ "learning_rate": 0.0006,
+ "loss": 4.474348068237305,
+ "step": 1323
+ },
+ {
+ "epoch": 18.39126637554585,
+ "grad_norm": 0.018889913335442543,
+ "learning_rate": 0.0006,
+ "loss": 4.4531636238098145,
+ "step": 1324
+ },
+ {
+ "epoch": 18.40524017467249,
+ "grad_norm": 0.019167251884937286,
+ "learning_rate": 0.0006,
+ "loss": 4.4278364181518555,
+ "step": 1325
+ },
+ {
+ "epoch": 18.419213973799128,
+ "grad_norm": 0.015895741060376167,
+ "learning_rate": 0.0006,
+ "loss": 4.474592208862305,
+ "step": 1326
+ },
+ {
+ "epoch": 18.433187772925763,
+ "grad_norm": 0.01531882956624031,
+ "learning_rate": 0.0006,
+ "loss": 4.403757572174072,
+ "step": 1327
+ },
+ {
+ "epoch": 18.4471615720524,
+ "grad_norm": 0.01565849967300892,
+ "learning_rate": 0.0006,
+ "loss": 4.507664680480957,
+ "step": 1328
+ },
+ {
+ "epoch": 18.46113537117904,
+ "grad_norm": 0.01602315716445446,
+ "learning_rate": 0.0006,
+ "loss": 4.5648908615112305,
+ "step": 1329
+ },
+ {
+ "epoch": 18.475109170305675,
+ "grad_norm": 0.016467366367578506,
+ "learning_rate": 0.0006,
+ "loss": 4.428762912750244,
+ "step": 1330
+ },
+ {
+ "epoch": 18.489082969432314,
+ "grad_norm": 0.01604127697646618,
+ "learning_rate": 0.0006,
+ "loss": 4.527771472930908,
+ "step": 1331
+ },
+ {
+ "epoch": 18.503056768558952,
+ "grad_norm": 0.018825586885213852,
+ "learning_rate": 0.0006,
+ "loss": 4.485386848449707,
+ "step": 1332
+ },
+ {
+ "epoch": 18.51703056768559,
+ "grad_norm": 0.019258007407188416,
+ "learning_rate": 0.0006,
+ "loss": 4.414972305297852,
+ "step": 1333
+ },
+ {
+ "epoch": 18.531004366812226,
+ "grad_norm": 0.018192021176218987,
+ "learning_rate": 0.0006,
+ "loss": 4.412166595458984,
+ "step": 1334
+ },
+ {
+ "epoch": 18.544978165938865,
+ "grad_norm": 0.016120698302984238,
+ "learning_rate": 0.0006,
+ "loss": 4.484317779541016,
+ "step": 1335
+ },
+ {
+ "epoch": 18.558951965065503,
+ "grad_norm": 0.016496622934937477,
+ "learning_rate": 0.0006,
+ "loss": 4.514986515045166,
+ "step": 1336
+ },
+ {
+ "epoch": 18.57292576419214,
+ "grad_norm": 0.01795116998255253,
+ "learning_rate": 0.0006,
+ "loss": 4.41473388671875,
+ "step": 1337
+ },
+ {
+ "epoch": 18.586899563318777,
+ "grad_norm": 0.019062276929616928,
+ "learning_rate": 0.0006,
+ "loss": 4.601996898651123,
+ "step": 1338
+ },
+ {
+ "epoch": 18.600873362445416,
+ "grad_norm": 0.017174631357192993,
+ "learning_rate": 0.0006,
+ "loss": 4.571218490600586,
+ "step": 1339
+ },
+ {
+ "epoch": 18.61484716157205,
+ "grad_norm": 0.014015606604516506,
+ "learning_rate": 0.0006,
+ "loss": 4.526606559753418,
+ "step": 1340
+ },
+ {
+ "epoch": 18.62882096069869,
+ "grad_norm": 0.015358510427176952,
+ "learning_rate": 0.0006,
+ "loss": 4.604286193847656,
+ "step": 1341
+ },
+ {
+ "epoch": 18.64279475982533,
+ "grad_norm": 0.01619391329586506,
+ "learning_rate": 0.0006,
+ "loss": 4.560883522033691,
+ "step": 1342
+ },
+ {
+ "epoch": 18.656768558951963,
+ "grad_norm": 0.015972964465618134,
+ "learning_rate": 0.0006,
+ "loss": 4.479755401611328,
+ "step": 1343
+ },
+ {
+ "epoch": 18.670742358078602,
+ "grad_norm": 0.016861025243997574,
+ "learning_rate": 0.0006,
+ "loss": 4.474471092224121,
+ "step": 1344
+ },
+ {
+ "epoch": 18.68471615720524,
+ "grad_norm": 0.01822078227996826,
+ "learning_rate": 0.0006,
+ "loss": 4.509489059448242,
+ "step": 1345
+ },
+ {
+ "epoch": 18.69868995633188,
+ "grad_norm": 0.019415007904171944,
+ "learning_rate": 0.0006,
+ "loss": 4.508941173553467,
+ "step": 1346
+ },
+ {
+ "epoch": 18.712663755458514,
+ "grad_norm": 0.019856126978993416,
+ "learning_rate": 0.0006,
+ "loss": 4.446670055389404,
+ "step": 1347
+ },
+ {
+ "epoch": 18.726637554585153,
+ "grad_norm": 0.01842520199716091,
+ "learning_rate": 0.0006,
+ "loss": 4.524422645568848,
+ "step": 1348
+ },
+ {
+ "epoch": 18.74061135371179,
+ "grad_norm": 0.017196480184793472,
+ "learning_rate": 0.0006,
+ "loss": 4.498466968536377,
+ "step": 1349
+ },
+ {
+ "epoch": 18.754585152838427,
+ "grad_norm": 0.01823010854423046,
+ "learning_rate": 0.0006,
+ "loss": 4.534684181213379,
+ "step": 1350
+ },
+ {
+ "epoch": 18.768558951965066,
+ "grad_norm": 0.020398065447807312,
+ "learning_rate": 0.0006,
+ "loss": 4.456494331359863,
+ "step": 1351
+ },
+ {
+ "epoch": 18.782532751091704,
+ "grad_norm": 0.022992262616753578,
+ "learning_rate": 0.0006,
+ "loss": 4.476374626159668,
+ "step": 1352
+ },
+ {
+ "epoch": 18.79650655021834,
+ "grad_norm": 0.02235420234501362,
+ "learning_rate": 0.0006,
+ "loss": 4.587776184082031,
+ "step": 1353
+ },
+ {
+ "epoch": 18.810480349344978,
+ "grad_norm": 0.02093626745045185,
+ "learning_rate": 0.0006,
+ "loss": 4.424172401428223,
+ "step": 1354
+ },
+ {
+ "epoch": 18.824454148471617,
+ "grad_norm": 0.01997302658855915,
+ "learning_rate": 0.0006,
+ "loss": 4.448066711425781,
+ "step": 1355
+ },
+ {
+ "epoch": 18.83842794759825,
+ "grad_norm": 0.018265459686517715,
+ "learning_rate": 0.0006,
+ "loss": 4.436610698699951,
+ "step": 1356
+ },
+ {
+ "epoch": 18.85240174672489,
+ "grad_norm": 0.01673833839595318,
+ "learning_rate": 0.0006,
+ "loss": 4.5502214431762695,
+ "step": 1357
+ },
+ {
+ "epoch": 18.86637554585153,
+ "grad_norm": 0.017908833920955658,
+ "learning_rate": 0.0006,
+ "loss": 4.550530433654785,
+ "step": 1358
+ },
+ {
+ "epoch": 18.880349344978168,
+ "grad_norm": 0.01796654611825943,
+ "learning_rate": 0.0006,
+ "loss": 4.531463146209717,
+ "step": 1359
+ },
+ {
+ "epoch": 18.894323144104803,
+ "grad_norm": 0.01646183803677559,
+ "learning_rate": 0.0006,
+ "loss": 4.498582363128662,
+ "step": 1360
+ },
+ {
+ "epoch": 18.90829694323144,
+ "grad_norm": 0.015430327504873276,
+ "learning_rate": 0.0006,
+ "loss": 4.52571439743042,
+ "step": 1361
+ },
+ {
+ "epoch": 18.92227074235808,
+ "grad_norm": 0.016527952626347542,
+ "learning_rate": 0.0006,
+ "loss": 4.547013759613037,
+ "step": 1362
+ },
+ {
+ "epoch": 18.936244541484715,
+ "grad_norm": 0.016263660043478012,
+ "learning_rate": 0.0006,
+ "loss": 4.468754291534424,
+ "step": 1363
+ },
+ {
+ "epoch": 18.950218340611354,
+ "grad_norm": 0.016347210854291916,
+ "learning_rate": 0.0006,
+ "loss": 4.450741767883301,
+ "step": 1364
+ },
+ {
+ "epoch": 18.964192139737992,
+ "grad_norm": 0.01701398566365242,
+ "learning_rate": 0.0006,
+ "loss": 4.534899711608887,
+ "step": 1365
+ },
+ {
+ "epoch": 18.978165938864628,
+ "grad_norm": 0.017693081870675087,
+ "learning_rate": 0.0006,
+ "loss": 4.460553169250488,
+ "step": 1366
+ },
+ {
+ "epoch": 18.992139737991266,
+ "grad_norm": 0.018746308982372284,
+ "learning_rate": 0.0006,
+ "loss": 4.538792610168457,
+ "step": 1367
+ },
+ {
+ "epoch": 19.0,
+ "grad_norm": 0.020099416375160217,
+ "learning_rate": 0.0006,
+ "loss": 4.449617385864258,
+ "step": 1368
+ },
+ {
+ "epoch": 19.0,
+ "eval_loss": 4.7756571769714355,
+ "eval_runtime": 57.1804,
+ "eval_samples_per_second": 42.707,
+ "eval_steps_per_second": 1.347,
+ "step": 1368
+ },
+ {
+ "epoch": 19.01397379912664,
+ "grad_norm": 0.018726933747529984,
+ "learning_rate": 0.0006,
+ "loss": 4.446059703826904,
+ "step": 1369
+ },
+ {
+ "epoch": 19.027947598253274,
+ "grad_norm": 0.01956087350845337,
+ "learning_rate": 0.0006,
+ "loss": 4.4520039558410645,
+ "step": 1370
+ },
+ {
+ "epoch": 19.041921397379912,
+ "grad_norm": 0.020092811435461044,
+ "learning_rate": 0.0006,
+ "loss": 4.543343544006348,
+ "step": 1371
+ },
+ {
+ "epoch": 19.05589519650655,
+ "grad_norm": 0.01843622885644436,
+ "learning_rate": 0.0006,
+ "loss": 4.535579204559326,
+ "step": 1372
+ },
+ {
+ "epoch": 19.069868995633186,
+ "grad_norm": 0.016989488154649734,
+ "learning_rate": 0.0006,
+ "loss": 4.52316951751709,
+ "step": 1373
+ },
+ {
+ "epoch": 19.083842794759825,
+ "grad_norm": 0.01805422455072403,
+ "learning_rate": 0.0006,
+ "loss": 4.5393266677856445,
+ "step": 1374
+ },
+ {
+ "epoch": 19.097816593886463,
+ "grad_norm": 0.01874629408121109,
+ "learning_rate": 0.0006,
+ "loss": 4.422059059143066,
+ "step": 1375
+ },
+ {
+ "epoch": 19.111790393013102,
+ "grad_norm": 0.01804221048951149,
+ "learning_rate": 0.0006,
+ "loss": 4.50508451461792,
+ "step": 1376
+ },
+ {
+ "epoch": 19.125764192139737,
+ "grad_norm": 0.01785474829375744,
+ "learning_rate": 0.0006,
+ "loss": 4.412599563598633,
+ "step": 1377
+ },
+ {
+ "epoch": 19.139737991266376,
+ "grad_norm": 0.01971791312098503,
+ "learning_rate": 0.0006,
+ "loss": 4.518318176269531,
+ "step": 1378
+ },
+ {
+ "epoch": 19.153711790393015,
+ "grad_norm": 0.02061633951961994,
+ "learning_rate": 0.0006,
+ "loss": 4.332393169403076,
+ "step": 1379
+ },
+ {
+ "epoch": 19.16768558951965,
+ "grad_norm": 0.020563429221510887,
+ "learning_rate": 0.0006,
+ "loss": 4.445863723754883,
+ "step": 1380
+ },
+ {
+ "epoch": 19.18165938864629,
+ "grad_norm": 0.019995713606476784,
+ "learning_rate": 0.0006,
+ "loss": 4.497312068939209,
+ "step": 1381
+ },
+ {
+ "epoch": 19.195633187772927,
+ "grad_norm": 0.019823070615530014,
+ "learning_rate": 0.0006,
+ "loss": 4.475085258483887,
+ "step": 1382
+ },
+ {
+ "epoch": 19.209606986899562,
+ "grad_norm": 0.019039355218410492,
+ "learning_rate": 0.0006,
+ "loss": 4.50432014465332,
+ "step": 1383
+ },
+ {
+ "epoch": 19.2235807860262,
+ "grad_norm": 0.017876973375678062,
+ "learning_rate": 0.0006,
+ "loss": 4.384788513183594,
+ "step": 1384
+ },
+ {
+ "epoch": 19.23755458515284,
+ "grad_norm": 0.018697699531912804,
+ "learning_rate": 0.0006,
+ "loss": 4.410778045654297,
+ "step": 1385
+ },
+ {
+ "epoch": 19.251528384279474,
+ "grad_norm": 0.018826806917786598,
+ "learning_rate": 0.0006,
+ "loss": 4.524034023284912,
+ "step": 1386
+ },
+ {
+ "epoch": 19.265502183406113,
+ "grad_norm": 0.01824226602911949,
+ "learning_rate": 0.0006,
+ "loss": 4.526393890380859,
+ "step": 1387
+ },
+ {
+ "epoch": 19.27947598253275,
+ "grad_norm": 0.01753208599984646,
+ "learning_rate": 0.0006,
+ "loss": 4.493703365325928,
+ "step": 1388
+ },
+ {
+ "epoch": 19.29344978165939,
+ "grad_norm": 0.01742432825267315,
+ "learning_rate": 0.0006,
+ "loss": 4.392831802368164,
+ "step": 1389
+ },
+ {
+ "epoch": 19.307423580786025,
+ "grad_norm": 0.017099356278777122,
+ "learning_rate": 0.0006,
+ "loss": 4.578668594360352,
+ "step": 1390
+ },
+ {
+ "epoch": 19.321397379912664,
+ "grad_norm": 0.017946023494005203,
+ "learning_rate": 0.0006,
+ "loss": 4.456718444824219,
+ "step": 1391
+ },
+ {
+ "epoch": 19.335371179039303,
+ "grad_norm": 0.017705943435430527,
+ "learning_rate": 0.0006,
+ "loss": 4.553682327270508,
+ "step": 1392
+ },
+ {
+ "epoch": 19.349344978165938,
+ "grad_norm": 0.016647523269057274,
+ "learning_rate": 0.0006,
+ "loss": 4.525016784667969,
+ "step": 1393
+ },
+ {
+ "epoch": 19.363318777292577,
+ "grad_norm": 0.015407416969537735,
+ "learning_rate": 0.0006,
+ "loss": 4.500231742858887,
+ "step": 1394
+ },
+ {
+ "epoch": 19.377292576419215,
+ "grad_norm": 0.017264019697904587,
+ "learning_rate": 0.0006,
+ "loss": 4.4615349769592285,
+ "step": 1395
+ },
+ {
+ "epoch": 19.39126637554585,
+ "grad_norm": 0.01871555484831333,
+ "learning_rate": 0.0006,
+ "loss": 4.281026840209961,
+ "step": 1396
+ },
+ {
+ "epoch": 19.40524017467249,
+ "grad_norm": 0.017361650243401527,
+ "learning_rate": 0.0006,
+ "loss": 4.479382514953613,
+ "step": 1397
+ },
+ {
+ "epoch": 19.419213973799128,
+ "grad_norm": 0.019781362265348434,
+ "learning_rate": 0.0006,
+ "loss": 4.402320861816406,
+ "step": 1398
+ },
+ {
+ "epoch": 19.433187772925763,
+ "grad_norm": 0.019043298438191414,
+ "learning_rate": 0.0006,
+ "loss": 4.476983070373535,
+ "step": 1399
+ },
+ {
+ "epoch": 19.4471615720524,
+ "grad_norm": 0.017137985676527023,
+ "learning_rate": 0.0006,
+ "loss": 4.418689250946045,
+ "step": 1400
+ },
+ {
+ "epoch": 19.46113537117904,
+ "grad_norm": 0.016481833532452583,
+ "learning_rate": 0.0006,
+ "loss": 4.579145431518555,
+ "step": 1401
+ },
+ {
+ "epoch": 19.475109170305675,
+ "grad_norm": 0.016507714986801147,
+ "learning_rate": 0.0006,
+ "loss": 4.519158363342285,
+ "step": 1402
+ },
+ {
+ "epoch": 19.489082969432314,
+ "grad_norm": 0.018092872574925423,
+ "learning_rate": 0.0006,
+ "loss": 4.513748645782471,
+ "step": 1403
+ },
+ {
+ "epoch": 19.503056768558952,
+ "grad_norm": 0.018153440207242966,
+ "learning_rate": 0.0006,
+ "loss": 4.645384311676025,
+ "step": 1404
+ },
+ {
+ "epoch": 19.51703056768559,
+ "grad_norm": 0.01657119207084179,
+ "learning_rate": 0.0006,
+ "loss": 4.399757385253906,
+ "step": 1405
+ },
+ {
+ "epoch": 19.531004366812226,
+ "grad_norm": 0.0156280305236578,
+ "learning_rate": 0.0006,
+ "loss": 4.467836380004883,
+ "step": 1406
+ },
+ {
+ "epoch": 19.544978165938865,
+ "grad_norm": 0.016838403418660164,
+ "learning_rate": 0.0006,
+ "loss": 4.529176235198975,
+ "step": 1407
+ },
+ {
+ "epoch": 19.558951965065503,
+ "grad_norm": 0.01743236929178238,
+ "learning_rate": 0.0006,
+ "loss": 4.440032482147217,
+ "step": 1408
+ },
+ {
+ "epoch": 19.57292576419214,
+ "grad_norm": 0.01553610060364008,
+ "learning_rate": 0.0006,
+ "loss": 4.425588607788086,
+ "step": 1409
+ },
+ {
+ "epoch": 19.586899563318777,
+ "grad_norm": 0.017259271815419197,
+ "learning_rate": 0.0006,
+ "loss": 4.405452728271484,
+ "step": 1410
+ },
+ {
+ "epoch": 19.600873362445416,
+ "grad_norm": 0.015568481758236885,
+ "learning_rate": 0.0006,
+ "loss": 4.444279193878174,
+ "step": 1411
+ },
+ {
+ "epoch": 19.61484716157205,
+ "grad_norm": 0.015331248752772808,
+ "learning_rate": 0.0006,
+ "loss": 4.451846122741699,
+ "step": 1412
+ },
+ {
+ "epoch": 19.62882096069869,
+ "grad_norm": 0.015667378902435303,
+ "learning_rate": 0.0006,
+ "loss": 4.533931255340576,
+ "step": 1413
+ },
+ {
+ "epoch": 19.64279475982533,
+ "grad_norm": 0.016177602112293243,
+ "learning_rate": 0.0006,
+ "loss": 4.467313289642334,
+ "step": 1414
+ },
+ {
+ "epoch": 19.656768558951963,
+ "grad_norm": 0.01733129471540451,
+ "learning_rate": 0.0006,
+ "loss": 4.496893882751465,
+ "step": 1415
+ },
+ {
+ "epoch": 19.670742358078602,
+ "grad_norm": 0.016539594158530235,
+ "learning_rate": 0.0006,
+ "loss": 4.401961326599121,
+ "step": 1416
+ },
+ {
+ "epoch": 19.68471615720524,
+ "grad_norm": 0.016058551147580147,
+ "learning_rate": 0.0006,
+ "loss": 4.503375053405762,
+ "step": 1417
+ },
+ {
+ "epoch": 19.69868995633188,
+ "grad_norm": 0.01715777814388275,
+ "learning_rate": 0.0006,
+ "loss": 4.3808369636535645,
+ "step": 1418
+ },
+ {
+ "epoch": 19.712663755458514,
+ "grad_norm": 0.01681119203567505,
+ "learning_rate": 0.0006,
+ "loss": 4.5469560623168945,
+ "step": 1419
+ },
+ {
+ "epoch": 19.726637554585153,
+ "grad_norm": 0.01634550653398037,
+ "learning_rate": 0.0006,
+ "loss": 4.328197956085205,
+ "step": 1420
+ },
+ {
+ "epoch": 19.74061135371179,
+ "grad_norm": 0.01549347210675478,
+ "learning_rate": 0.0006,
+ "loss": 4.534621238708496,
+ "step": 1421
+ },
+ {
+ "epoch": 19.754585152838427,
+ "grad_norm": 0.01654665172100067,
+ "learning_rate": 0.0006,
+ "loss": 4.4724225997924805,
+ "step": 1422
+ },
+ {
+ "epoch": 19.768558951965066,
+ "grad_norm": 0.017119025811553,
+ "learning_rate": 0.0006,
+ "loss": 4.3897247314453125,
+ "step": 1423
+ },
+ {
+ "epoch": 19.782532751091704,
+ "grad_norm": 0.01603223942220211,
+ "learning_rate": 0.0006,
+ "loss": 4.412731170654297,
+ "step": 1424
+ },
+ {
+ "epoch": 19.79650655021834,
+ "grad_norm": 0.016774067655205727,
+ "learning_rate": 0.0006,
+ "loss": 4.54707145690918,
+ "step": 1425
+ },
+ {
+ "epoch": 19.810480349344978,
+ "grad_norm": 0.016453025862574577,
+ "learning_rate": 0.0006,
+ "loss": 4.462501049041748,
+ "step": 1426
+ },
+ {
+ "epoch": 19.824454148471617,
+ "grad_norm": 0.01597629487514496,
+ "learning_rate": 0.0006,
+ "loss": 4.229758262634277,
+ "step": 1427
+ },
+ {
+ "epoch": 19.83842794759825,
+ "grad_norm": 0.017999034374952316,
+ "learning_rate": 0.0006,
+ "loss": 4.389263153076172,
+ "step": 1428
+ },
+ {
+ "epoch": 19.85240174672489,
+ "grad_norm": 0.02165898121893406,
+ "learning_rate": 0.0006,
+ "loss": 4.342130661010742,
+ "step": 1429
+ },
+ {
+ "epoch": 19.86637554585153,
+ "grad_norm": 0.02222822792828083,
+ "learning_rate": 0.0006,
+ "loss": 4.492661476135254,
+ "step": 1430
+ },
+ {
+ "epoch": 19.880349344978168,
+ "grad_norm": 0.019378066062927246,
+ "learning_rate": 0.0006,
+ "loss": 4.4614105224609375,
+ "step": 1431
+ },
+ {
+ "epoch": 19.894323144104803,
+ "grad_norm": 0.019908083602786064,
+ "learning_rate": 0.0006,
+ "loss": 4.370929718017578,
+ "step": 1432
+ },
+ {
+ "epoch": 19.90829694323144,
+ "grad_norm": 0.022296471521258354,
+ "learning_rate": 0.0006,
+ "loss": 4.438320159912109,
+ "step": 1433
+ },
+ {
+ "epoch": 19.92227074235808,
+ "grad_norm": 0.021810343489050865,
+ "learning_rate": 0.0006,
+ "loss": 4.462067127227783,
+ "step": 1434
+ },
+ {
+ "epoch": 19.936244541484715,
+ "grad_norm": 0.02129209227859974,
+ "learning_rate": 0.0006,
+ "loss": 4.541261196136475,
+ "step": 1435
+ },
+ {
+ "epoch": 19.950218340611354,
+ "grad_norm": 0.02079629711806774,
+ "learning_rate": 0.0006,
+ "loss": 4.36712646484375,
+ "step": 1436
+ },
+ {
+ "epoch": 19.964192139737992,
+ "grad_norm": 0.022501740604639053,
+ "learning_rate": 0.0006,
+ "loss": 4.5279083251953125,
+ "step": 1437
+ },
+ {
+ "epoch": 19.978165938864628,
+ "grad_norm": 0.019121676683425903,
+ "learning_rate": 0.0006,
+ "loss": 4.477789878845215,
+ "step": 1438
+ },
+ {
+ "epoch": 19.992139737991266,
+ "grad_norm": 0.017429588362574577,
+ "learning_rate": 0.0006,
+ "loss": 4.4023847579956055,
+ "step": 1439
+ },
+ {
+ "epoch": 20.0,
+ "grad_norm": 0.018857665359973907,
+ "learning_rate": 0.0006,
+ "loss": 4.494743347167969,
+ "step": 1440
+ },
+ {
+ "epoch": 20.0,
+ "eval_loss": 4.694555759429932,
+ "eval_runtime": 56.4146,
+ "eval_samples_per_second": 43.287,
+ "eval_steps_per_second": 1.365,
+ "step": 1440
+ },
+ {
+ "epoch": 20.01397379912664,
+ "grad_norm": 0.016505861654877663,
+ "learning_rate": 0.0006,
+ "loss": 4.506229877471924,
+ "step": 1441
+ },
+ {
+ "epoch": 20.027947598253274,
+ "grad_norm": 0.017571842297911644,
+ "learning_rate": 0.0006,
+ "loss": 4.384073257446289,
+ "step": 1442
+ },
+ {
+ "epoch": 20.041921397379912,
+ "grad_norm": 0.019398365169763565,
+ "learning_rate": 0.0006,
+ "loss": 4.482177734375,
+ "step": 1443
+ },
+ {
+ "epoch": 20.05589519650655,
+ "grad_norm": 0.02011062018573284,
+ "learning_rate": 0.0006,
+ "loss": 4.440773963928223,
+ "step": 1444
+ },
+ {
+ "epoch": 20.069868995633186,
+ "grad_norm": 0.019026953727006912,
+ "learning_rate": 0.0006,
+ "loss": 4.507168769836426,
+ "step": 1445
+ },
+ {
+ "epoch": 20.083842794759825,
+ "grad_norm": 0.020506031811237335,
+ "learning_rate": 0.0006,
+ "loss": 4.469079971313477,
+ "step": 1446
+ },
+ {
+ "epoch": 20.097816593886463,
+ "grad_norm": 0.019608162343502045,
+ "learning_rate": 0.0006,
+ "loss": 4.395927429199219,
+ "step": 1447
+ },
+ {
+ "epoch": 20.111790393013102,
+ "grad_norm": 0.019419124349951744,
+ "learning_rate": 0.0006,
+ "loss": 4.484235763549805,
+ "step": 1448
+ },
+ {
+ "epoch": 20.125764192139737,
+ "grad_norm": 0.021775128319859505,
+ "learning_rate": 0.0006,
+ "loss": 4.333173751831055,
+ "step": 1449
+ },
+ {
+ "epoch": 20.139737991266376,
+ "grad_norm": 0.025512272492051125,
+ "learning_rate": 0.0006,
+ "loss": 4.460362434387207,
+ "step": 1450
+ },
+ {
+ "epoch": 20.153711790393015,
+ "grad_norm": 0.025157401338219643,
+ "learning_rate": 0.0006,
+ "loss": 4.4361395835876465,
+ "step": 1451
+ },
+ {
+ "epoch": 20.16768558951965,
+ "grad_norm": 0.019862616434693336,
+ "learning_rate": 0.0006,
+ "loss": 4.426807403564453,
+ "step": 1452
+ },
+ {
+ "epoch": 20.18165938864629,
+ "grad_norm": 0.01932152360677719,
+ "learning_rate": 0.0006,
+ "loss": 4.375677108764648,
+ "step": 1453
+ },
+ {
+ "epoch": 20.195633187772927,
+ "grad_norm": 0.01760999858379364,
+ "learning_rate": 0.0006,
+ "loss": 4.42259407043457,
+ "step": 1454
+ },
+ {
+ "epoch": 20.209606986899562,
+ "grad_norm": 0.016998112201690674,
+ "learning_rate": 0.0006,
+ "loss": 4.481280326843262,
+ "step": 1455
+ },
+ {
+ "epoch": 20.2235807860262,
+ "grad_norm": 0.017199190333485603,
+ "learning_rate": 0.0006,
+ "loss": 4.496712684631348,
+ "step": 1456
+ },
+ {
+ "epoch": 20.23755458515284,
+ "grad_norm": 0.016833679750561714,
+ "learning_rate": 0.0006,
+ "loss": 4.40736722946167,
+ "step": 1457
+ },
+ {
+ "epoch": 20.251528384279474,
+ "grad_norm": 0.017546730116009712,
+ "learning_rate": 0.0006,
+ "loss": 4.4915266036987305,
+ "step": 1458
+ },
+ {
+ "epoch": 20.265502183406113,
+ "grad_norm": 0.016595078632235527,
+ "learning_rate": 0.0006,
+ "loss": 4.460815906524658,
+ "step": 1459
+ },
+ {
+ "epoch": 20.27947598253275,
+ "grad_norm": 0.017728568986058235,
+ "learning_rate": 0.0006,
+ "loss": 4.346048831939697,
+ "step": 1460
+ },
+ {
+ "epoch": 20.29344978165939,
+ "grad_norm": 0.01849197782576084,
+ "learning_rate": 0.0006,
+ "loss": 4.399681091308594,
+ "step": 1461
+ },
+ {
+ "epoch": 20.307423580786025,
+ "grad_norm": 0.018336588516831398,
+ "learning_rate": 0.0006,
+ "loss": 4.414183616638184,
+ "step": 1462
+ },
+ {
+ "epoch": 20.321397379912664,
+ "grad_norm": 0.018086519092321396,
+ "learning_rate": 0.0006,
+ "loss": 4.346586227416992,
+ "step": 1463
+ },
+ {
+ "epoch": 20.335371179039303,
+ "grad_norm": 0.019697299227118492,
+ "learning_rate": 0.0006,
+ "loss": 4.4420952796936035,
+ "step": 1464
+ },
+ {
+ "epoch": 20.349344978165938,
+ "grad_norm": 0.02198074571788311,
+ "learning_rate": 0.0006,
+ "loss": 4.588729381561279,
+ "step": 1465
+ },
+ {
+ "epoch": 20.363318777292577,
+ "grad_norm": 0.022813329473137856,
+ "learning_rate": 0.0006,
+ "loss": 4.49729585647583,
+ "step": 1466
+ },
+ {
+ "epoch": 20.377292576419215,
+ "grad_norm": 0.022006649523973465,
+ "learning_rate": 0.0006,
+ "loss": 4.475641250610352,
+ "step": 1467
+ },
+ {
+ "epoch": 20.39126637554585,
+ "grad_norm": 0.023188438266515732,
+ "learning_rate": 0.0006,
+ "loss": 4.404322624206543,
+ "step": 1468
+ },
+ {
+ "epoch": 20.40524017467249,
+ "grad_norm": 0.022226877510547638,
+ "learning_rate": 0.0006,
+ "loss": 4.50904655456543,
+ "step": 1469
+ },
+ {
+ "epoch": 20.419213973799128,
+ "grad_norm": 0.021082637831568718,
+ "learning_rate": 0.0006,
+ "loss": 4.339011192321777,
+ "step": 1470
+ },
+ {
+ "epoch": 20.433187772925763,
+ "grad_norm": 0.02013542130589485,
+ "learning_rate": 0.0006,
+ "loss": 4.516942024230957,
+ "step": 1471
+ },
+ {
+ "epoch": 20.4471615720524,
+ "grad_norm": 0.01898968778550625,
+ "learning_rate": 0.0006,
+ "loss": 4.315804481506348,
+ "step": 1472
+ },
+ {
+ "epoch": 20.46113537117904,
+ "grad_norm": 0.019450880587100983,
+ "learning_rate": 0.0006,
+ "loss": 4.4794921875,
+ "step": 1473
+ },
+ {
+ "epoch": 20.475109170305675,
+ "grad_norm": 0.018577802926301956,
+ "learning_rate": 0.0006,
+ "loss": 4.422323703765869,
+ "step": 1474
+ },
+ {
+ "epoch": 20.489082969432314,
+ "grad_norm": 0.01778412237763405,
+ "learning_rate": 0.0006,
+ "loss": 4.429789066314697,
+ "step": 1475
+ },
+ {
+ "epoch": 20.503056768558952,
+ "grad_norm": 0.017134476453065872,
+ "learning_rate": 0.0006,
+ "loss": 4.416459083557129,
+ "step": 1476
+ },
+ {
+ "epoch": 20.51703056768559,
+ "grad_norm": 0.017153557389974594,
+ "learning_rate": 0.0006,
+ "loss": 4.446170806884766,
+ "step": 1477
+ },
+ {
+ "epoch": 20.531004366812226,
+ "grad_norm": 0.014998532831668854,
+ "learning_rate": 0.0006,
+ "loss": 4.419447898864746,
+ "step": 1478
+ },
+ {
+ "epoch": 20.544978165938865,
+ "grad_norm": 0.015877850353717804,
+ "learning_rate": 0.0006,
+ "loss": 4.312338829040527,
+ "step": 1479
+ },
+ {
+ "epoch": 20.558951965065503,
+ "grad_norm": 0.01569489948451519,
+ "learning_rate": 0.0006,
+ "loss": 4.572905540466309,
+ "step": 1480
+ },
+ {
+ "epoch": 20.57292576419214,
+ "grad_norm": 0.01591474376618862,
+ "learning_rate": 0.0006,
+ "loss": 4.414907455444336,
+ "step": 1481
+ },
+ {
+ "epoch": 20.586899563318777,
+ "grad_norm": 0.015395903028547764,
+ "learning_rate": 0.0006,
+ "loss": 4.549131393432617,
+ "step": 1482
+ },
+ {
+ "epoch": 20.600873362445416,
+ "grad_norm": 0.015715358778834343,
+ "learning_rate": 0.0006,
+ "loss": 4.530194282531738,
+ "step": 1483
+ },
+ {
+ "epoch": 20.61484716157205,
+ "grad_norm": 0.016042975708842278,
+ "learning_rate": 0.0006,
+ "loss": 4.322149276733398,
+ "step": 1484
+ },
+ {
+ "epoch": 20.62882096069869,
+ "grad_norm": 0.01664186641573906,
+ "learning_rate": 0.0006,
+ "loss": 4.424866676330566,
+ "step": 1485
+ },
+ {
+ "epoch": 20.64279475982533,
+ "grad_norm": 0.01631534844636917,
+ "learning_rate": 0.0006,
+ "loss": 4.441752910614014,
+ "step": 1486
+ },
+ {
+ "epoch": 20.656768558951963,
+ "grad_norm": 0.014057058840990067,
+ "learning_rate": 0.0006,
+ "loss": 4.399982452392578,
+ "step": 1487
+ },
+ {
+ "epoch": 20.670742358078602,
+ "grad_norm": 0.015076273120939732,
+ "learning_rate": 0.0006,
+ "loss": 4.441324234008789,
+ "step": 1488
+ },
+ {
+ "epoch": 20.68471615720524,
+ "grad_norm": 0.014138277620077133,
+ "learning_rate": 0.0006,
+ "loss": 4.4312543869018555,
+ "step": 1489
+ },
+ {
+ "epoch": 20.69868995633188,
+ "grad_norm": 0.013762440532445908,
+ "learning_rate": 0.0006,
+ "loss": 4.387807846069336,
+ "step": 1490
+ },
+ {
+ "epoch": 20.712663755458514,
+ "grad_norm": 0.015058411285281181,
+ "learning_rate": 0.0006,
+ "loss": 4.594747543334961,
+ "step": 1491
+ },
+ {
+ "epoch": 20.726637554585153,
+ "grad_norm": 0.013827123679220676,
+ "learning_rate": 0.0006,
+ "loss": 4.434604167938232,
+ "step": 1492
+ },
+ {
+ "epoch": 20.74061135371179,
+ "grad_norm": 0.015315295197069645,
+ "learning_rate": 0.0006,
+ "loss": 4.588066101074219,
+ "step": 1493
+ },
+ {
+ "epoch": 20.754585152838427,
+ "grad_norm": 0.018678732216358185,
+ "learning_rate": 0.0006,
+ "loss": 4.3514204025268555,
+ "step": 1494
+ },
+ {
+ "epoch": 20.768558951965066,
+ "grad_norm": 0.019521350041031837,
+ "learning_rate": 0.0006,
+ "loss": 4.437658786773682,
+ "step": 1495
+ },
+ {
+ "epoch": 20.782532751091704,
+ "grad_norm": 0.018363403156399727,
+ "learning_rate": 0.0006,
+ "loss": 4.495002269744873,
+ "step": 1496
+ },
+ {
+ "epoch": 20.79650655021834,
+ "grad_norm": 0.016830725595355034,
+ "learning_rate": 0.0006,
+ "loss": 4.402471542358398,
+ "step": 1497
+ },
+ {
+ "epoch": 20.810480349344978,
+ "grad_norm": 0.01722072623670101,
+ "learning_rate": 0.0006,
+ "loss": 4.437653541564941,
+ "step": 1498
+ },
+ {
+ "epoch": 20.824454148471617,
+ "grad_norm": 0.017459599301218987,
+ "learning_rate": 0.0006,
+ "loss": 4.366603851318359,
+ "step": 1499
+ },
+ {
+ "epoch": 20.83842794759825,
+ "grad_norm": 0.018902219831943512,
+ "learning_rate": 0.0006,
+ "loss": 4.501561164855957,
+ "step": 1500
+ },
+ {
+ "epoch": 20.85240174672489,
+ "grad_norm": 0.020814096555113792,
+ "learning_rate": 0.0006,
+ "loss": 4.535295486450195,
+ "step": 1501
+ },
+ {
+ "epoch": 20.86637554585153,
+ "grad_norm": 0.018990976735949516,
+ "learning_rate": 0.0006,
+ "loss": 4.333580017089844,
+ "step": 1502
+ },
+ {
+ "epoch": 20.880349344978168,
+ "grad_norm": 0.01837880350649357,
+ "learning_rate": 0.0006,
+ "loss": 4.313388824462891,
+ "step": 1503
+ },
+ {
+ "epoch": 20.894323144104803,
+ "grad_norm": 0.019820360466837883,
+ "learning_rate": 0.0006,
+ "loss": 4.388157844543457,
+ "step": 1504
+ },
+ {
+ "epoch": 20.90829694323144,
+ "grad_norm": 0.020181143656373024,
+ "learning_rate": 0.0006,
+ "loss": 4.572224140167236,
+ "step": 1505
+ },
+ {
+ "epoch": 20.92227074235808,
+ "grad_norm": 0.021206077188253403,
+ "learning_rate": 0.0006,
+ "loss": 4.3127970695495605,
+ "step": 1506
+ },
+ {
+ "epoch": 20.936244541484715,
+ "grad_norm": 0.017962060868740082,
+ "learning_rate": 0.0006,
+ "loss": 4.384734630584717,
+ "step": 1507
+ },
+ {
+ "epoch": 20.950218340611354,
+ "grad_norm": 0.01737220585346222,
+ "learning_rate": 0.0006,
+ "loss": 4.4075751304626465,
+ "step": 1508
+ },
+ {
+ "epoch": 20.964192139737992,
+ "grad_norm": 0.01839268207550049,
+ "learning_rate": 0.0006,
+ "loss": 4.553779125213623,
+ "step": 1509
+ },
+ {
+ "epoch": 20.978165938864628,
+ "grad_norm": 0.020102640613913536,
+ "learning_rate": 0.0006,
+ "loss": 4.399398326873779,
+ "step": 1510
+ },
+ {
+ "epoch": 20.992139737991266,
+ "grad_norm": 0.018831463530659676,
+ "learning_rate": 0.0006,
+ "loss": 4.358499526977539,
+ "step": 1511
+ },
+ {
+ "epoch": 21.0,
+ "grad_norm": 0.018254734575748444,
+ "learning_rate": 0.0006,
+ "loss": 4.384119510650635,
+ "step": 1512
+ },
+ {
+ "epoch": 21.0,
+ "eval_loss": 4.7524237632751465,
+ "eval_runtime": 57.8812,
+ "eval_samples_per_second": 42.19,
+ "eval_steps_per_second": 1.33,
+ "step": 1512
+ },
+ {
+ "epoch": 21.01397379912664,
+ "grad_norm": 0.019728587940335274,
+ "learning_rate": 0.0006,
+ "loss": 4.32363224029541,
+ "step": 1513
+ },
+ {
+ "epoch": 21.027947598253274,
+ "grad_norm": 0.02216625027358532,
+ "learning_rate": 0.0006,
+ "loss": 4.436988353729248,
+ "step": 1514
+ },
+ {
+ "epoch": 21.041921397379912,
+ "grad_norm": 0.02227606810629368,
+ "learning_rate": 0.0006,
+ "loss": 4.406063556671143,
+ "step": 1515
+ },
+ {
+ "epoch": 21.05589519650655,
+ "grad_norm": 0.020187009125947952,
+ "learning_rate": 0.0006,
+ "loss": 4.324402332305908,
+ "step": 1516
+ },
+ {
+ "epoch": 21.069868995633186,
+ "grad_norm": 0.02095157466828823,
+ "learning_rate": 0.0006,
+ "loss": 4.456544876098633,
+ "step": 1517
+ },
+ {
+ "epoch": 21.083842794759825,
+ "grad_norm": 0.02146013453602791,
+ "learning_rate": 0.0006,
+ "loss": 4.368387222290039,
+ "step": 1518
+ },
+ {
+ "epoch": 21.097816593886463,
+ "grad_norm": 0.02162035182118416,
+ "learning_rate": 0.0006,
+ "loss": 4.48838996887207,
+ "step": 1519
+ },
+ {
+ "epoch": 21.111790393013102,
+ "grad_norm": 0.021745676174759865,
+ "learning_rate": 0.0006,
+ "loss": 4.30778694152832,
+ "step": 1520
+ },
+ {
+ "epoch": 21.125764192139737,
+ "grad_norm": 0.02173588052392006,
+ "learning_rate": 0.0006,
+ "loss": 4.433149814605713,
+ "step": 1521
+ },
+ {
+ "epoch": 21.139737991266376,
+ "grad_norm": 0.020950648933649063,
+ "learning_rate": 0.0006,
+ "loss": 4.520089149475098,
+ "step": 1522
+ },
+ {
+ "epoch": 21.153711790393015,
+ "grad_norm": 0.020684752613306046,
+ "learning_rate": 0.0006,
+ "loss": 4.372437477111816,
+ "step": 1523
+ },
+ {
+ "epoch": 21.16768558951965,
+ "grad_norm": 0.019472869113087654,
+ "learning_rate": 0.0006,
+ "loss": 4.437015533447266,
+ "step": 1524
+ },
+ {
+ "epoch": 21.18165938864629,
+ "grad_norm": 0.01846962980926037,
+ "learning_rate": 0.0006,
+ "loss": 4.449531555175781,
+ "step": 1525
+ },
+ {
+ "epoch": 21.195633187772927,
+ "grad_norm": 0.017377803102135658,
+ "learning_rate": 0.0006,
+ "loss": 4.458279132843018,
+ "step": 1526
+ },
+ {
+ "epoch": 21.209606986899562,
+ "grad_norm": 0.016002200543880463,
+ "learning_rate": 0.0006,
+ "loss": 4.266297817230225,
+ "step": 1527
+ },
+ {
+ "epoch": 21.2235807860262,
+ "grad_norm": 0.015575903467833996,
+ "learning_rate": 0.0006,
+ "loss": 4.342632293701172,
+ "step": 1528
+ },
+ {
+ "epoch": 21.23755458515284,
+ "grad_norm": 0.015634698793292046,
+ "learning_rate": 0.0006,
+ "loss": 4.338623523712158,
+ "step": 1529
+ },
+ {
+ "epoch": 21.251528384279474,
+ "grad_norm": 0.01449581328779459,
+ "learning_rate": 0.0006,
+ "loss": 4.354330062866211,
+ "step": 1530
+ },
+ {
+ "epoch": 21.265502183406113,
+ "grad_norm": 0.01487616915255785,
+ "learning_rate": 0.0006,
+ "loss": 4.486835479736328,
+ "step": 1531
+ },
+ {
+ "epoch": 21.27947598253275,
+ "grad_norm": 0.016346555203199387,
+ "learning_rate": 0.0006,
+ "loss": 4.407098770141602,
+ "step": 1532
+ },
+ {
+ "epoch": 21.29344978165939,
+ "grad_norm": 0.01682182401418686,
+ "learning_rate": 0.0006,
+ "loss": 4.436809539794922,
+ "step": 1533
+ },
+ {
+ "epoch": 21.307423580786025,
+ "grad_norm": 0.017704004421830177,
+ "learning_rate": 0.0006,
+ "loss": 4.3854780197143555,
+ "step": 1534
+ },
+ {
+ "epoch": 21.321397379912664,
+ "grad_norm": 0.018437420949339867,
+ "learning_rate": 0.0006,
+ "loss": 4.413283824920654,
+ "step": 1535
+ },
+ {
+ "epoch": 21.335371179039303,
+ "grad_norm": 0.016911081969738007,
+ "learning_rate": 0.0006,
+ "loss": 4.357306003570557,
+ "step": 1536
+ },
+ {
+ "epoch": 21.349344978165938,
+ "grad_norm": 0.01642223633825779,
+ "learning_rate": 0.0006,
+ "loss": 4.422322750091553,
+ "step": 1537
+ },
+ {
+ "epoch": 21.363318777292577,
+ "grad_norm": 0.01780877821147442,
+ "learning_rate": 0.0006,
+ "loss": 4.387620449066162,
+ "step": 1538
+ },
+ {
+ "epoch": 21.377292576419215,
+ "grad_norm": 0.01653975248336792,
+ "learning_rate": 0.0006,
+ "loss": 4.327267646789551,
+ "step": 1539
+ },
+ {
+ "epoch": 21.39126637554585,
+ "grad_norm": 0.01572156324982643,
+ "learning_rate": 0.0006,
+ "loss": 4.475979804992676,
+ "step": 1540
+ },
+ {
+ "epoch": 21.40524017467249,
+ "grad_norm": 0.017156066372990608,
+ "learning_rate": 0.0006,
+ "loss": 4.4011125564575195,
+ "step": 1541
+ },
+ {
+ "epoch": 21.419213973799128,
+ "grad_norm": 0.01788829080760479,
+ "learning_rate": 0.0006,
+ "loss": 4.4174089431762695,
+ "step": 1542
+ },
+ {
+ "epoch": 21.433187772925763,
+ "grad_norm": 0.01728985086083412,
+ "learning_rate": 0.0006,
+ "loss": 4.429244041442871,
+ "step": 1543
+ },
+ {
+ "epoch": 21.4471615720524,
+ "grad_norm": 0.016534466296434402,
+ "learning_rate": 0.0006,
+ "loss": 4.326068878173828,
+ "step": 1544
+ },
+ {
+ "epoch": 21.46113537117904,
+ "grad_norm": 0.016582584008574486,
+ "learning_rate": 0.0006,
+ "loss": 4.2975969314575195,
+ "step": 1545
+ },
+ {
+ "epoch": 21.475109170305675,
+ "grad_norm": 0.018034178763628006,
+ "learning_rate": 0.0006,
+ "loss": 4.501319408416748,
+ "step": 1546
+ },
+ {
+ "epoch": 21.489082969432314,
+ "grad_norm": 0.017827702686190605,
+ "learning_rate": 0.0006,
+ "loss": 4.502284526824951,
+ "step": 1547
+ },
+ {
+ "epoch": 21.503056768558952,
+ "grad_norm": 0.017182299867272377,
+ "learning_rate": 0.0006,
+ "loss": 4.380828857421875,
+ "step": 1548
+ },
+ {
+ "epoch": 21.51703056768559,
+ "grad_norm": 0.019101588055491447,
+ "learning_rate": 0.0006,
+ "loss": 4.31564998626709,
+ "step": 1549
+ },
+ {
+ "epoch": 21.531004366812226,
+ "grad_norm": 0.0204677302390337,
+ "learning_rate": 0.0006,
+ "loss": 4.3840131759643555,
+ "step": 1550
+ },
+ {
+ "epoch": 21.544978165938865,
+ "grad_norm": 0.0188993439078331,
+ "learning_rate": 0.0006,
+ "loss": 4.4257731437683105,
+ "step": 1551
+ },
+ {
+ "epoch": 21.558951965065503,
+ "grad_norm": 0.01730230636894703,
+ "learning_rate": 0.0006,
+ "loss": 4.445132255554199,
+ "step": 1552
+ },
+ {
+ "epoch": 21.57292576419214,
+ "grad_norm": 0.01801574043929577,
+ "learning_rate": 0.0006,
+ "loss": 4.427064895629883,
+ "step": 1553
+ },
+ {
+ "epoch": 21.586899563318777,
+ "grad_norm": 0.018417565152049065,
+ "learning_rate": 0.0006,
+ "loss": 4.319595813751221,
+ "step": 1554
+ },
+ {
+ "epoch": 21.600873362445416,
+ "grad_norm": 0.01683160290122032,
+ "learning_rate": 0.0006,
+ "loss": 4.349218368530273,
+ "step": 1555
+ },
+ {
+ "epoch": 21.61484716157205,
+ "grad_norm": 0.016973106190562248,
+ "learning_rate": 0.0006,
+ "loss": 4.375990390777588,
+ "step": 1556
+ },
+ {
+ "epoch": 21.62882096069869,
+ "grad_norm": 0.015688583254814148,
+ "learning_rate": 0.0006,
+ "loss": 4.408170223236084,
+ "step": 1557
+ },
+ {
+ "epoch": 21.64279475982533,
+ "grad_norm": 0.01573043316602707,
+ "learning_rate": 0.0006,
+ "loss": 4.416255950927734,
+ "step": 1558
+ },
+ {
+ "epoch": 21.656768558951963,
+ "grad_norm": 0.01744293048977852,
+ "learning_rate": 0.0006,
+ "loss": 4.477141380310059,
+ "step": 1559
+ },
+ {
+ "epoch": 21.670742358078602,
+ "grad_norm": 0.017503513023257256,
+ "learning_rate": 0.0006,
+ "loss": 4.3367486000061035,
+ "step": 1560
+ },
+ {
+ "epoch": 21.68471615720524,
+ "grad_norm": 0.016592150554060936,
+ "learning_rate": 0.0006,
+ "loss": 4.314785480499268,
+ "step": 1561
+ },
+ {
+ "epoch": 21.69868995633188,
+ "grad_norm": 0.015774084255099297,
+ "learning_rate": 0.0006,
+ "loss": 4.448209285736084,
+ "step": 1562
+ },
+ {
+ "epoch": 21.712663755458514,
+ "grad_norm": 0.015779150649905205,
+ "learning_rate": 0.0006,
+ "loss": 4.392690658569336,
+ "step": 1563
+ },
+ {
+ "epoch": 21.726637554585153,
+ "grad_norm": 0.015392648056149483,
+ "learning_rate": 0.0006,
+ "loss": 4.437741756439209,
+ "step": 1564
+ },
+ {
+ "epoch": 21.74061135371179,
+ "grad_norm": 0.01530836895108223,
+ "learning_rate": 0.0006,
+ "loss": 4.424595832824707,
+ "step": 1565
+ },
+ {
+ "epoch": 21.754585152838427,
+ "grad_norm": 0.015600779093801975,
+ "learning_rate": 0.0006,
+ "loss": 4.335417747497559,
+ "step": 1566
+ },
+ {
+ "epoch": 21.768558951965066,
+ "grad_norm": 0.014646518044173717,
+ "learning_rate": 0.0006,
+ "loss": 4.279081344604492,
+ "step": 1567
+ },
+ {
+ "epoch": 21.782532751091704,
+ "grad_norm": 0.014132864773273468,
+ "learning_rate": 0.0006,
+ "loss": 4.345950126647949,
+ "step": 1568
+ },
+ {
+ "epoch": 21.79650655021834,
+ "grad_norm": 0.01425724383443594,
+ "learning_rate": 0.0006,
+ "loss": 4.277588844299316,
+ "step": 1569
+ },
+ {
+ "epoch": 21.810480349344978,
+ "grad_norm": 0.015396572649478912,
+ "learning_rate": 0.0006,
+ "loss": 4.326196670532227,
+ "step": 1570
+ },
+ {
+ "epoch": 21.824454148471617,
+ "grad_norm": 0.01683669537305832,
+ "learning_rate": 0.0006,
+ "loss": 4.3294172286987305,
+ "step": 1571
+ },
+ {
+ "epoch": 21.83842794759825,
+ "grad_norm": 0.016365647315979004,
+ "learning_rate": 0.0006,
+ "loss": 4.523321628570557,
+ "step": 1572
+ },
+ {
+ "epoch": 21.85240174672489,
+ "grad_norm": 0.017142774537205696,
+ "learning_rate": 0.0006,
+ "loss": 4.475842475891113,
+ "step": 1573
+ },
+ {
+ "epoch": 21.86637554585153,
+ "grad_norm": 0.015950839966535568,
+ "learning_rate": 0.0006,
+ "loss": 4.397187232971191,
+ "step": 1574
+ },
+ {
+ "epoch": 21.880349344978168,
+ "grad_norm": 0.016180362552404404,
+ "learning_rate": 0.0006,
+ "loss": 4.414778709411621,
+ "step": 1575
+ },
+ {
+ "epoch": 21.894323144104803,
+ "grad_norm": 0.017615556716918945,
+ "learning_rate": 0.0006,
+ "loss": 4.348567962646484,
+ "step": 1576
+ },
+ {
+ "epoch": 21.90829694323144,
+ "grad_norm": 0.01896277628839016,
+ "learning_rate": 0.0006,
+ "loss": 4.431762218475342,
+ "step": 1577
+ },
+ {
+ "epoch": 21.92227074235808,
+ "grad_norm": 0.01990971714258194,
+ "learning_rate": 0.0006,
+ "loss": 4.431285858154297,
+ "step": 1578
+ },
+ {
+ "epoch": 21.936244541484715,
+ "grad_norm": 0.0201258547604084,
+ "learning_rate": 0.0006,
+ "loss": 4.309872627258301,
+ "step": 1579
+ },
+ {
+ "epoch": 21.950218340611354,
+ "grad_norm": 0.019620198756456375,
+ "learning_rate": 0.0006,
+ "loss": 4.484044551849365,
+ "step": 1580
+ },
+ {
+ "epoch": 21.964192139737992,
+ "grad_norm": 0.01771123707294464,
+ "learning_rate": 0.0006,
+ "loss": 4.368093013763428,
+ "step": 1581
+ },
+ {
+ "epoch": 21.978165938864628,
+ "grad_norm": 0.017501654103398323,
+ "learning_rate": 0.0006,
+ "loss": 4.372651100158691,
+ "step": 1582
+ },
+ {
+ "epoch": 21.992139737991266,
+ "grad_norm": 0.01693608984351158,
+ "learning_rate": 0.0006,
+ "loss": 4.379167556762695,
+ "step": 1583
+ },
+ {
+ "epoch": 22.0,
+ "grad_norm": 0.02002849616110325,
+ "learning_rate": 0.0006,
+ "loss": 4.329627990722656,
+ "step": 1584
+ },
+ {
+ "epoch": 22.0,
+ "eval_loss": 4.723696231842041,
+ "eval_runtime": 56.1758,
+ "eval_samples_per_second": 43.471,
+ "eval_steps_per_second": 1.371,
+ "step": 1584
+ },
+ {
+ "epoch": 22.01397379912664,
+ "grad_norm": 0.02030305378139019,
+ "learning_rate": 0.0006,
+ "loss": 4.355566024780273,
+ "step": 1585
+ },
+ {
+ "epoch": 22.027947598253274,
+ "grad_norm": 0.018638934940099716,
+ "learning_rate": 0.0006,
+ "loss": 4.314009189605713,
+ "step": 1586
+ },
+ {
+ "epoch": 22.041921397379912,
+ "grad_norm": 0.019917072728276253,
+ "learning_rate": 0.0006,
+ "loss": 4.367525577545166,
+ "step": 1587
+ },
+ {
+ "epoch": 22.05589519650655,
+ "grad_norm": 0.020785167813301086,
+ "learning_rate": 0.0006,
+ "loss": 4.375561714172363,
+ "step": 1588
+ },
+ {
+ "epoch": 22.069868995633186,
+ "grad_norm": 0.021214453503489494,
+ "learning_rate": 0.0006,
+ "loss": 4.418305397033691,
+ "step": 1589
+ },
+ {
+ "epoch": 22.083842794759825,
+ "grad_norm": 0.019051678478717804,
+ "learning_rate": 0.0006,
+ "loss": 4.215126991271973,
+ "step": 1590
+ },
+ {
+ "epoch": 22.097816593886463,
+ "grad_norm": 0.02323366142809391,
+ "learning_rate": 0.0006,
+ "loss": 4.306807994842529,
+ "step": 1591
+ },
+ {
+ "epoch": 22.111790393013102,
+ "grad_norm": 0.026186738163232803,
+ "learning_rate": 0.0006,
+ "loss": 4.408508777618408,
+ "step": 1592
+ },
+ {
+ "epoch": 22.125764192139737,
+ "grad_norm": 0.027361227199435234,
+ "learning_rate": 0.0006,
+ "loss": 4.329322814941406,
+ "step": 1593
+ },
+ {
+ "epoch": 22.139737991266376,
+ "grad_norm": 0.026564646512269974,
+ "learning_rate": 0.0006,
+ "loss": 4.346642017364502,
+ "step": 1594
+ },
+ {
+ "epoch": 22.153711790393015,
+ "grad_norm": 0.025073667988181114,
+ "learning_rate": 0.0006,
+ "loss": 4.3984198570251465,
+ "step": 1595
+ },
+ {
+ "epoch": 22.16768558951965,
+ "grad_norm": 0.02347794733941555,
+ "learning_rate": 0.0006,
+ "loss": 4.352538108825684,
+ "step": 1596
+ },
+ {
+ "epoch": 22.18165938864629,
+ "grad_norm": 0.02203468047082424,
+ "learning_rate": 0.0006,
+ "loss": 4.298148155212402,
+ "step": 1597
+ },
+ {
+ "epoch": 22.195633187772927,
+ "grad_norm": 0.020621957257390022,
+ "learning_rate": 0.0006,
+ "loss": 4.480309963226318,
+ "step": 1598
+ },
+ {
+ "epoch": 22.209606986899562,
+ "grad_norm": 0.023710055276751518,
+ "learning_rate": 0.0006,
+ "loss": 4.339317321777344,
+ "step": 1599
+ },
+ {
+ "epoch": 22.2235807860262,
+ "grad_norm": 0.0233482476323843,
+ "learning_rate": 0.0006,
+ "loss": 4.436896324157715,
+ "step": 1600
+ },
+ {
+ "epoch": 22.23755458515284,
+ "grad_norm": 0.020122403278946877,
+ "learning_rate": 0.0006,
+ "loss": 4.236002445220947,
+ "step": 1601
+ },
+ {
+ "epoch": 22.251528384279474,
+ "grad_norm": 0.01957266591489315,
+ "learning_rate": 0.0006,
+ "loss": 4.497328281402588,
+ "step": 1602
+ },
+ {
+ "epoch": 22.265502183406113,
+ "grad_norm": 0.018964895978569984,
+ "learning_rate": 0.0006,
+ "loss": 4.36360502243042,
+ "step": 1603
+ },
+ {
+ "epoch": 22.27947598253275,
+ "grad_norm": 0.018902769312262535,
+ "learning_rate": 0.0006,
+ "loss": 4.427093505859375,
+ "step": 1604
+ },
+ {
+ "epoch": 22.29344978165939,
+ "grad_norm": 0.018918119370937347,
+ "learning_rate": 0.0006,
+ "loss": 4.294950485229492,
+ "step": 1605
+ },
+ {
+ "epoch": 22.307423580786025,
+ "grad_norm": 0.019315775483846664,
+ "learning_rate": 0.0006,
+ "loss": 4.427952289581299,
+ "step": 1606
+ },
+ {
+ "epoch": 22.321397379912664,
+ "grad_norm": 0.018350522965192795,
+ "learning_rate": 0.0006,
+ "loss": 4.272945404052734,
+ "step": 1607
+ },
+ {
+ "epoch": 22.335371179039303,
+ "grad_norm": 0.01759534701704979,
+ "learning_rate": 0.0006,
+ "loss": 4.348023891448975,
+ "step": 1608
+ },
+ {
+ "epoch": 22.349344978165938,
+ "grad_norm": 0.01646057888865471,
+ "learning_rate": 0.0006,
+ "loss": 4.278518199920654,
+ "step": 1609
+ },
+ {
+ "epoch": 22.363318777292577,
+ "grad_norm": 0.015875060111284256,
+ "learning_rate": 0.0006,
+ "loss": 4.423527240753174,
+ "step": 1610
+ },
+ {
+ "epoch": 22.377292576419215,
+ "grad_norm": 0.01638529635965824,
+ "learning_rate": 0.0006,
+ "loss": 4.321723937988281,
+ "step": 1611
+ },
+ {
+ "epoch": 22.39126637554585,
+ "grad_norm": 0.016942061483860016,
+ "learning_rate": 0.0006,
+ "loss": 4.346164703369141,
+ "step": 1612
+ },
+ {
+ "epoch": 22.40524017467249,
+ "grad_norm": 0.015209621749818325,
+ "learning_rate": 0.0006,
+ "loss": 4.434340000152588,
+ "step": 1613
+ },
+ {
+ "epoch": 22.419213973799128,
+ "grad_norm": 0.014423009008169174,
+ "learning_rate": 0.0006,
+ "loss": 4.342560768127441,
+ "step": 1614
+ },
+ {
+ "epoch": 22.433187772925763,
+ "grad_norm": 0.014369525946676731,
+ "learning_rate": 0.0006,
+ "loss": 4.296586990356445,
+ "step": 1615
+ },
+ {
+ "epoch": 22.4471615720524,
+ "grad_norm": 0.01392540242522955,
+ "learning_rate": 0.0006,
+ "loss": 4.433176040649414,
+ "step": 1616
+ },
+ {
+ "epoch": 22.46113537117904,
+ "grad_norm": 0.013206899166107178,
+ "learning_rate": 0.0006,
+ "loss": 4.32962703704834,
+ "step": 1617
+ },
+ {
+ "epoch": 22.475109170305675,
+ "grad_norm": 0.013427077792584896,
+ "learning_rate": 0.0006,
+ "loss": 4.420437335968018,
+ "step": 1618
+ },
+ {
+ "epoch": 22.489082969432314,
+ "grad_norm": 0.014194196090102196,
+ "learning_rate": 0.0006,
+ "loss": 4.410478591918945,
+ "step": 1619
+ },
+ {
+ "epoch": 22.503056768558952,
+ "grad_norm": 0.01362364087253809,
+ "learning_rate": 0.0006,
+ "loss": 4.338932037353516,
+ "step": 1620
+ },
+ {
+ "epoch": 22.51703056768559,
+ "grad_norm": 0.014299589209258556,
+ "learning_rate": 0.0006,
+ "loss": 4.34089994430542,
+ "step": 1621
+ },
+ {
+ "epoch": 22.531004366812226,
+ "grad_norm": 0.014061295427381992,
+ "learning_rate": 0.0006,
+ "loss": 4.29201602935791,
+ "step": 1622
+ },
+ {
+ "epoch": 22.544978165938865,
+ "grad_norm": 0.013945229351520538,
+ "learning_rate": 0.0006,
+ "loss": 4.326052188873291,
+ "step": 1623
+ },
+ {
+ "epoch": 22.558951965065503,
+ "grad_norm": 0.01387068536132574,
+ "learning_rate": 0.0006,
+ "loss": 4.393521308898926,
+ "step": 1624
+ },
+ {
+ "epoch": 22.57292576419214,
+ "grad_norm": 0.013931898400187492,
+ "learning_rate": 0.0006,
+ "loss": 4.3739752769470215,
+ "step": 1625
+ },
+ {
+ "epoch": 22.586899563318777,
+ "grad_norm": 0.01497740764170885,
+ "learning_rate": 0.0006,
+ "loss": 4.3991618156433105,
+ "step": 1626
+ },
+ {
+ "epoch": 22.600873362445416,
+ "grad_norm": 0.015842905268073082,
+ "learning_rate": 0.0006,
+ "loss": 4.405340671539307,
+ "step": 1627
+ },
+ {
+ "epoch": 22.61484716157205,
+ "grad_norm": 0.01580633595585823,
+ "learning_rate": 0.0006,
+ "loss": 4.280102729797363,
+ "step": 1628
+ },
+ {
+ "epoch": 22.62882096069869,
+ "grad_norm": 0.016567382961511612,
+ "learning_rate": 0.0006,
+ "loss": 4.407657146453857,
+ "step": 1629
+ },
+ {
+ "epoch": 22.64279475982533,
+ "grad_norm": 0.01648622378706932,
+ "learning_rate": 0.0006,
+ "loss": 4.392590522766113,
+ "step": 1630
+ },
+ {
+ "epoch": 22.656768558951963,
+ "grad_norm": 0.017066307365894318,
+ "learning_rate": 0.0006,
+ "loss": 4.324862003326416,
+ "step": 1631
+ },
+ {
+ "epoch": 22.670742358078602,
+ "grad_norm": 0.01637062057852745,
+ "learning_rate": 0.0006,
+ "loss": 4.34598445892334,
+ "step": 1632
+ },
+ {
+ "epoch": 22.68471615720524,
+ "grad_norm": 0.018038960173726082,
+ "learning_rate": 0.0006,
+ "loss": 4.320711612701416,
+ "step": 1633
+ },
+ {
+ "epoch": 22.69868995633188,
+ "grad_norm": 0.018096428364515305,
+ "learning_rate": 0.0006,
+ "loss": 4.536527156829834,
+ "step": 1634
+ },
+ {
+ "epoch": 22.712663755458514,
+ "grad_norm": 0.017889291048049927,
+ "learning_rate": 0.0006,
+ "loss": 4.388131141662598,
+ "step": 1635
+ },
+ {
+ "epoch": 22.726637554585153,
+ "grad_norm": 0.018771521747112274,
+ "learning_rate": 0.0006,
+ "loss": 4.357668876647949,
+ "step": 1636
+ },
+ {
+ "epoch": 22.74061135371179,
+ "grad_norm": 0.02100873365998268,
+ "learning_rate": 0.0006,
+ "loss": 4.293972015380859,
+ "step": 1637
+ },
+ {
+ "epoch": 22.754585152838427,
+ "grad_norm": 0.017715785652399063,
+ "learning_rate": 0.0006,
+ "loss": 4.349849224090576,
+ "step": 1638
+ },
+ {
+ "epoch": 22.768558951965066,
+ "grad_norm": 0.015546333976089954,
+ "learning_rate": 0.0006,
+ "loss": 4.311728000640869,
+ "step": 1639
+ },
+ {
+ "epoch": 22.782532751091704,
+ "grad_norm": 0.01605990156531334,
+ "learning_rate": 0.0006,
+ "loss": 4.499123573303223,
+ "step": 1640
+ },
+ {
+ "epoch": 22.79650655021834,
+ "grad_norm": 0.016642142087221146,
+ "learning_rate": 0.0006,
+ "loss": 4.4319257736206055,
+ "step": 1641
+ },
+ {
+ "epoch": 22.810480349344978,
+ "grad_norm": 0.016120387241244316,
+ "learning_rate": 0.0006,
+ "loss": 4.4373650550842285,
+ "step": 1642
+ },
+ {
+ "epoch": 22.824454148471617,
+ "grad_norm": 0.015174890868365765,
+ "learning_rate": 0.0006,
+ "loss": 4.314082145690918,
+ "step": 1643
+ },
+ {
+ "epoch": 22.83842794759825,
+ "grad_norm": 0.01551714539527893,
+ "learning_rate": 0.0006,
+ "loss": 4.360363960266113,
+ "step": 1644
+ },
+ {
+ "epoch": 22.85240174672489,
+ "grad_norm": 0.014966742135584354,
+ "learning_rate": 0.0006,
+ "loss": 4.346253395080566,
+ "step": 1645
+ },
+ {
+ "epoch": 22.86637554585153,
+ "grad_norm": 0.014288028702139854,
+ "learning_rate": 0.0006,
+ "loss": 4.464366436004639,
+ "step": 1646
+ },
+ {
+ "epoch": 22.880349344978168,
+ "grad_norm": 0.015291682444512844,
+ "learning_rate": 0.0006,
+ "loss": 4.504101753234863,
+ "step": 1647
+ },
+ {
+ "epoch": 22.894323144104803,
+ "grad_norm": 0.014534426853060722,
+ "learning_rate": 0.0006,
+ "loss": 4.462172031402588,
+ "step": 1648
+ },
+ {
+ "epoch": 22.90829694323144,
+ "grad_norm": 0.014129819348454475,
+ "learning_rate": 0.0006,
+ "loss": 4.350126266479492,
+ "step": 1649
+ },
+ {
+ "epoch": 22.92227074235808,
+ "grad_norm": 0.014932668767869473,
+ "learning_rate": 0.0006,
+ "loss": 4.374392032623291,
+ "step": 1650
+ },
+ {
+ "epoch": 22.936244541484715,
+ "grad_norm": 0.014647853560745716,
+ "learning_rate": 0.0006,
+ "loss": 4.252264499664307,
+ "step": 1651
+ },
+ {
+ "epoch": 22.950218340611354,
+ "grad_norm": 0.013553360477089882,
+ "learning_rate": 0.0006,
+ "loss": 4.379729270935059,
+ "step": 1652
+ },
+ {
+ "epoch": 22.964192139737992,
+ "grad_norm": 0.014606194570660591,
+ "learning_rate": 0.0006,
+ "loss": 4.363372802734375,
+ "step": 1653
+ },
+ {
+ "epoch": 22.978165938864628,
+ "grad_norm": 0.015000631101429462,
+ "learning_rate": 0.0006,
+ "loss": 4.408048629760742,
+ "step": 1654
+ },
+ {
+ "epoch": 22.992139737991266,
+ "grad_norm": 0.015370816923677921,
+ "learning_rate": 0.0006,
+ "loss": 4.297451019287109,
+ "step": 1655
+ },
+ {
+ "epoch": 23.0,
+ "grad_norm": 0.01709144562482834,
+ "learning_rate": 0.0006,
+ "loss": 4.294776916503906,
+ "step": 1656
+ },
+ {
+ "epoch": 23.0,
+ "eval_loss": 4.678618907928467,
+ "eval_runtime": 56.1668,
+ "eval_samples_per_second": 43.478,
+ "eval_steps_per_second": 1.371,
+ "step": 1656
+ },
+ {
+ "epoch": 23.01397379912664,
+ "grad_norm": 0.01727290451526642,
+ "learning_rate": 0.0006,
+ "loss": 4.311701774597168,
+ "step": 1657
+ },
+ {
+ "epoch": 23.027947598253274,
+ "grad_norm": 0.020645759999752045,
+ "learning_rate": 0.0006,
+ "loss": 4.429897308349609,
+ "step": 1658
+ },
+ {
+ "epoch": 23.041921397379912,
+ "grad_norm": 0.02204226516187191,
+ "learning_rate": 0.0006,
+ "loss": 4.365151882171631,
+ "step": 1659
+ },
+ {
+ "epoch": 23.05589519650655,
+ "grad_norm": 0.023069269955158234,
+ "learning_rate": 0.0006,
+ "loss": 4.3710832595825195,
+ "step": 1660
+ },
+ {
+ "epoch": 23.069868995633186,
+ "grad_norm": 0.024151179939508438,
+ "learning_rate": 0.0006,
+ "loss": 4.403108596801758,
+ "step": 1661
+ },
+ {
+ "epoch": 23.083842794759825,
+ "grad_norm": 0.02511233650147915,
+ "learning_rate": 0.0006,
+ "loss": 4.414128303527832,
+ "step": 1662
+ },
+ {
+ "epoch": 23.097816593886463,
+ "grad_norm": 0.025802113115787506,
+ "learning_rate": 0.0006,
+ "loss": 4.388265609741211,
+ "step": 1663
+ },
+ {
+ "epoch": 23.111790393013102,
+ "grad_norm": 0.02377348765730858,
+ "learning_rate": 0.0006,
+ "loss": 4.357505798339844,
+ "step": 1664
+ },
+ {
+ "epoch": 23.125764192139737,
+ "grad_norm": 0.020876651629805565,
+ "learning_rate": 0.0006,
+ "loss": 4.23043966293335,
+ "step": 1665
+ },
+ {
+ "epoch": 23.139737991266376,
+ "grad_norm": 0.01959727331995964,
+ "learning_rate": 0.0006,
+ "loss": 4.280970096588135,
+ "step": 1666
+ },
+ {
+ "epoch": 23.153711790393015,
+ "grad_norm": 0.018678616732358932,
+ "learning_rate": 0.0006,
+ "loss": 4.346716403961182,
+ "step": 1667
+ },
+ {
+ "epoch": 23.16768558951965,
+ "grad_norm": 0.02108944021165371,
+ "learning_rate": 0.0006,
+ "loss": 4.277904510498047,
+ "step": 1668
+ },
+ {
+ "epoch": 23.18165938864629,
+ "grad_norm": 0.020221566781401634,
+ "learning_rate": 0.0006,
+ "loss": 4.367693901062012,
+ "step": 1669
+ },
+ {
+ "epoch": 23.195633187772927,
+ "grad_norm": 0.019599711522459984,
+ "learning_rate": 0.0006,
+ "loss": 4.257092475891113,
+ "step": 1670
+ },
+ {
+ "epoch": 23.209606986899562,
+ "grad_norm": 0.019833361729979515,
+ "learning_rate": 0.0006,
+ "loss": 4.540737628936768,
+ "step": 1671
+ },
+ {
+ "epoch": 23.2235807860262,
+ "grad_norm": 0.020352095365524292,
+ "learning_rate": 0.0006,
+ "loss": 4.360923767089844,
+ "step": 1672
+ },
+ {
+ "epoch": 23.23755458515284,
+ "grad_norm": 0.01939748041331768,
+ "learning_rate": 0.0006,
+ "loss": 4.393110275268555,
+ "step": 1673
+ },
+ {
+ "epoch": 23.251528384279474,
+ "grad_norm": 0.01884354278445244,
+ "learning_rate": 0.0006,
+ "loss": 4.383938789367676,
+ "step": 1674
+ },
+ {
+ "epoch": 23.265502183406113,
+ "grad_norm": 0.02008470520377159,
+ "learning_rate": 0.0006,
+ "loss": 4.346344947814941,
+ "step": 1675
+ },
+ {
+ "epoch": 23.27947598253275,
+ "grad_norm": 0.019065195694565773,
+ "learning_rate": 0.0006,
+ "loss": 4.36611795425415,
+ "step": 1676
+ },
+ {
+ "epoch": 23.29344978165939,
+ "grad_norm": 0.01938541606068611,
+ "learning_rate": 0.0006,
+ "loss": 4.4009599685668945,
+ "step": 1677
+ },
+ {
+ "epoch": 23.307423580786025,
+ "grad_norm": 0.019044723361730576,
+ "learning_rate": 0.0006,
+ "loss": 4.340169906616211,
+ "step": 1678
+ },
+ {
+ "epoch": 23.321397379912664,
+ "grad_norm": 0.018850378692150116,
+ "learning_rate": 0.0006,
+ "loss": 4.37674617767334,
+ "step": 1679
+ },
+ {
+ "epoch": 23.335371179039303,
+ "grad_norm": 0.019076095893979073,
+ "learning_rate": 0.0006,
+ "loss": 4.226874351501465,
+ "step": 1680
+ },
+ {
+ "epoch": 23.349344978165938,
+ "grad_norm": 0.017543038353323936,
+ "learning_rate": 0.0006,
+ "loss": 4.337751388549805,
+ "step": 1681
+ },
+ {
+ "epoch": 23.363318777292577,
+ "grad_norm": 0.017170218750834465,
+ "learning_rate": 0.0006,
+ "loss": 4.282512664794922,
+ "step": 1682
+ },
+ {
+ "epoch": 23.377292576419215,
+ "grad_norm": 0.01574796624481678,
+ "learning_rate": 0.0006,
+ "loss": 4.30638313293457,
+ "step": 1683
+ },
+ {
+ "epoch": 23.39126637554585,
+ "grad_norm": 0.015160846523940563,
+ "learning_rate": 0.0006,
+ "loss": 4.336716175079346,
+ "step": 1684
+ },
+ {
+ "epoch": 23.40524017467249,
+ "grad_norm": 0.01534526702016592,
+ "learning_rate": 0.0006,
+ "loss": 4.2174787521362305,
+ "step": 1685
+ },
+ {
+ "epoch": 23.419213973799128,
+ "grad_norm": 0.015289999544620514,
+ "learning_rate": 0.0006,
+ "loss": 4.272546291351318,
+ "step": 1686
+ },
+ {
+ "epoch": 23.433187772925763,
+ "grad_norm": 0.014894185587763786,
+ "learning_rate": 0.0006,
+ "loss": 4.296045303344727,
+ "step": 1687
+ },
+ {
+ "epoch": 23.4471615720524,
+ "grad_norm": 0.014869502745568752,
+ "learning_rate": 0.0006,
+ "loss": 4.425461769104004,
+ "step": 1688
+ },
+ {
+ "epoch": 23.46113537117904,
+ "grad_norm": 0.01474483497440815,
+ "learning_rate": 0.0006,
+ "loss": 4.361867427825928,
+ "step": 1689
+ },
+ {
+ "epoch": 23.475109170305675,
+ "grad_norm": 0.015299245715141296,
+ "learning_rate": 0.0006,
+ "loss": 4.364523410797119,
+ "step": 1690
+ },
+ {
+ "epoch": 23.489082969432314,
+ "grad_norm": 0.015201340429484844,
+ "learning_rate": 0.0006,
+ "loss": 4.285182952880859,
+ "step": 1691
+ },
+ {
+ "epoch": 23.503056768558952,
+ "grad_norm": 0.016057293862104416,
+ "learning_rate": 0.0006,
+ "loss": 4.284873962402344,
+ "step": 1692
+ },
+ {
+ "epoch": 23.51703056768559,
+ "grad_norm": 0.015871062874794006,
+ "learning_rate": 0.0006,
+ "loss": 4.326470851898193,
+ "step": 1693
+ },
+ {
+ "epoch": 23.531004366812226,
+ "grad_norm": 0.01586288772523403,
+ "learning_rate": 0.0006,
+ "loss": 4.324883460998535,
+ "step": 1694
+ },
+ {
+ "epoch": 23.544978165938865,
+ "grad_norm": 0.016092827543616295,
+ "learning_rate": 0.0006,
+ "loss": 4.3258867263793945,
+ "step": 1695
+ },
+ {
+ "epoch": 23.558951965065503,
+ "grad_norm": 0.01582016795873642,
+ "learning_rate": 0.0006,
+ "loss": 4.358892440795898,
+ "step": 1696
+ },
+ {
+ "epoch": 23.57292576419214,
+ "grad_norm": 0.01688452623784542,
+ "learning_rate": 0.0006,
+ "loss": 4.280842304229736,
+ "step": 1697
+ },
+ {
+ "epoch": 23.586899563318777,
+ "grad_norm": 0.015865886583924294,
+ "learning_rate": 0.0006,
+ "loss": 4.245567321777344,
+ "step": 1698
+ },
+ {
+ "epoch": 23.600873362445416,
+ "grad_norm": 0.015303662046790123,
+ "learning_rate": 0.0006,
+ "loss": 4.383112907409668,
+ "step": 1699
+ },
+ {
+ "epoch": 23.61484716157205,
+ "grad_norm": 0.015649516135454178,
+ "learning_rate": 0.0006,
+ "loss": 4.519082069396973,
+ "step": 1700
+ },
+ {
+ "epoch": 23.62882096069869,
+ "grad_norm": 0.0167120099067688,
+ "learning_rate": 0.0006,
+ "loss": 4.447847366333008,
+ "step": 1701
+ },
+ {
+ "epoch": 23.64279475982533,
+ "grad_norm": 0.016368716955184937,
+ "learning_rate": 0.0006,
+ "loss": 4.434186935424805,
+ "step": 1702
+ },
+ {
+ "epoch": 23.656768558951963,
+ "grad_norm": 0.017379503697156906,
+ "learning_rate": 0.0006,
+ "loss": 4.317594528198242,
+ "step": 1703
+ },
+ {
+ "epoch": 23.670742358078602,
+ "grad_norm": 0.018573811277747154,
+ "learning_rate": 0.0006,
+ "loss": 4.205793857574463,
+ "step": 1704
+ },
+ {
+ "epoch": 23.68471615720524,
+ "grad_norm": 0.01749090477824211,
+ "learning_rate": 0.0006,
+ "loss": 4.3985700607299805,
+ "step": 1705
+ },
+ {
+ "epoch": 23.69868995633188,
+ "grad_norm": 0.0165668074041605,
+ "learning_rate": 0.0006,
+ "loss": 4.329226970672607,
+ "step": 1706
+ },
+ {
+ "epoch": 23.712663755458514,
+ "grad_norm": 0.016913846135139465,
+ "learning_rate": 0.0006,
+ "loss": 4.388485908508301,
+ "step": 1707
+ },
+ {
+ "epoch": 23.726637554585153,
+ "grad_norm": 0.016813859343528748,
+ "learning_rate": 0.0006,
+ "loss": 4.3395915031433105,
+ "step": 1708
+ },
+ {
+ "epoch": 23.74061135371179,
+ "grad_norm": 0.01676975190639496,
+ "learning_rate": 0.0006,
+ "loss": 4.254745960235596,
+ "step": 1709
+ },
+ {
+ "epoch": 23.754585152838427,
+ "grad_norm": 0.016370350494980812,
+ "learning_rate": 0.0006,
+ "loss": 4.2612223625183105,
+ "step": 1710
+ },
+ {
+ "epoch": 23.768558951965066,
+ "grad_norm": 0.01696198247373104,
+ "learning_rate": 0.0006,
+ "loss": 4.327524662017822,
+ "step": 1711
+ },
+ {
+ "epoch": 23.782532751091704,
+ "grad_norm": 0.0189108457416296,
+ "learning_rate": 0.0006,
+ "loss": 4.325839519500732,
+ "step": 1712
+ },
+ {
+ "epoch": 23.79650655021834,
+ "grad_norm": 0.01862351782619953,
+ "learning_rate": 0.0006,
+ "loss": 4.333662509918213,
+ "step": 1713
+ },
+ {
+ "epoch": 23.810480349344978,
+ "grad_norm": 0.01655014418065548,
+ "learning_rate": 0.0006,
+ "loss": 4.39305305480957,
+ "step": 1714
+ },
+ {
+ "epoch": 23.824454148471617,
+ "grad_norm": 0.015899376943707466,
+ "learning_rate": 0.0006,
+ "loss": 4.390053749084473,
+ "step": 1715
+ },
+ {
+ "epoch": 23.83842794759825,
+ "grad_norm": 0.016519056633114815,
+ "learning_rate": 0.0006,
+ "loss": 4.255903244018555,
+ "step": 1716
+ },
+ {
+ "epoch": 23.85240174672489,
+ "grad_norm": 0.0167539119720459,
+ "learning_rate": 0.0006,
+ "loss": 4.255727767944336,
+ "step": 1717
+ },
+ {
+ "epoch": 23.86637554585153,
+ "grad_norm": 0.0171353816986084,
+ "learning_rate": 0.0006,
+ "loss": 4.376740455627441,
+ "step": 1718
+ },
+ {
+ "epoch": 23.880349344978168,
+ "grad_norm": 0.016719117760658264,
+ "learning_rate": 0.0006,
+ "loss": 4.3376688957214355,
+ "step": 1719
+ },
+ {
+ "epoch": 23.894323144104803,
+ "grad_norm": 0.01561494916677475,
+ "learning_rate": 0.0006,
+ "loss": 4.3345746994018555,
+ "step": 1720
+ },
+ {
+ "epoch": 23.90829694323144,
+ "grad_norm": 0.016303174197673798,
+ "learning_rate": 0.0006,
+ "loss": 4.332371234893799,
+ "step": 1721
+ },
+ {
+ "epoch": 23.92227074235808,
+ "grad_norm": 0.016722865402698517,
+ "learning_rate": 0.0006,
+ "loss": 4.267125129699707,
+ "step": 1722
+ },
+ {
+ "epoch": 23.936244541484715,
+ "grad_norm": 0.017072124406695366,
+ "learning_rate": 0.0006,
+ "loss": 4.322751998901367,
+ "step": 1723
+ },
+ {
+ "epoch": 23.950218340611354,
+ "grad_norm": 0.01699932850897312,
+ "learning_rate": 0.0006,
+ "loss": 4.378422737121582,
+ "step": 1724
+ },
+ {
+ "epoch": 23.964192139737992,
+ "grad_norm": 0.01742720976471901,
+ "learning_rate": 0.0006,
+ "loss": 4.345083236694336,
+ "step": 1725
+ },
+ {
+ "epoch": 23.978165938864628,
+ "grad_norm": 0.018797501921653748,
+ "learning_rate": 0.0006,
+ "loss": 4.290443420410156,
+ "step": 1726
+ },
+ {
+ "epoch": 23.992139737991266,
+ "grad_norm": 0.018090683966875076,
+ "learning_rate": 0.0006,
+ "loss": 4.377580642700195,
+ "step": 1727
+ },
+ {
+ "epoch": 24.0,
+ "grad_norm": 0.017305459827184677,
+ "learning_rate": 0.0006,
+ "loss": 4.364426612854004,
+ "step": 1728
+ },
+ {
+ "epoch": 24.0,
+ "eval_loss": 4.647762298583984,
+ "eval_runtime": 56.5402,
+ "eval_samples_per_second": 43.191,
+ "eval_steps_per_second": 1.362,
+ "step": 1728
+ },
+ {
+ "epoch": 24.01397379912664,
+ "grad_norm": 0.016493607312440872,
+ "learning_rate": 0.0006,
+ "loss": 4.252255916595459,
+ "step": 1729
+ },
+ {
+ "epoch": 24.027947598253274,
+ "grad_norm": 0.016529450193047523,
+ "learning_rate": 0.0006,
+ "loss": 4.352746963500977,
+ "step": 1730
+ },
+ {
+ "epoch": 24.041921397379912,
+ "grad_norm": 0.016940191388130188,
+ "learning_rate": 0.0006,
+ "loss": 4.301870822906494,
+ "step": 1731
+ },
+ {
+ "epoch": 24.05589519650655,
+ "grad_norm": 0.01752558909356594,
+ "learning_rate": 0.0006,
+ "loss": 4.2762451171875,
+ "step": 1732
+ },
+ {
+ "epoch": 24.069868995633186,
+ "grad_norm": 0.017112884670495987,
+ "learning_rate": 0.0006,
+ "loss": 4.2659149169921875,
+ "step": 1733
+ },
+ {
+ "epoch": 24.083842794759825,
+ "grad_norm": 0.01900586113333702,
+ "learning_rate": 0.0006,
+ "loss": 4.306058883666992,
+ "step": 1734
+ },
+ {
+ "epoch": 24.097816593886463,
+ "grad_norm": 0.019447680562734604,
+ "learning_rate": 0.0006,
+ "loss": 4.450128555297852,
+ "step": 1735
+ },
+ {
+ "epoch": 24.111790393013102,
+ "grad_norm": 0.01948186755180359,
+ "learning_rate": 0.0006,
+ "loss": 4.202869415283203,
+ "step": 1736
+ },
+ {
+ "epoch": 24.125764192139737,
+ "grad_norm": 0.01716383546590805,
+ "learning_rate": 0.0006,
+ "loss": 4.318876266479492,
+ "step": 1737
+ },
+ {
+ "epoch": 24.139737991266376,
+ "grad_norm": 0.017049988731741905,
+ "learning_rate": 0.0006,
+ "loss": 4.237238883972168,
+ "step": 1738
+ },
+ {
+ "epoch": 24.153711790393015,
+ "grad_norm": 0.015169923193752766,
+ "learning_rate": 0.0006,
+ "loss": 4.289426803588867,
+ "step": 1739
+ },
+ {
+ "epoch": 24.16768558951965,
+ "grad_norm": 0.014760667458176613,
+ "learning_rate": 0.0006,
+ "loss": 4.356924057006836,
+ "step": 1740
+ },
+ {
+ "epoch": 24.18165938864629,
+ "grad_norm": 0.015127750113606453,
+ "learning_rate": 0.0006,
+ "loss": 4.142916679382324,
+ "step": 1741
+ },
+ {
+ "epoch": 24.195633187772927,
+ "grad_norm": 0.01549555640667677,
+ "learning_rate": 0.0006,
+ "loss": 4.345563888549805,
+ "step": 1742
+ },
+ {
+ "epoch": 24.209606986899562,
+ "grad_norm": 0.016393091529607773,
+ "learning_rate": 0.0006,
+ "loss": 4.417495250701904,
+ "step": 1743
+ },
+ {
+ "epoch": 24.2235807860262,
+ "grad_norm": 0.01785266026854515,
+ "learning_rate": 0.0006,
+ "loss": 4.3553547859191895,
+ "step": 1744
+ },
+ {
+ "epoch": 24.23755458515284,
+ "grad_norm": 0.017574617639183998,
+ "learning_rate": 0.0006,
+ "loss": 4.241024494171143,
+ "step": 1745
+ },
+ {
+ "epoch": 24.251528384279474,
+ "grad_norm": 0.019060306251049042,
+ "learning_rate": 0.0006,
+ "loss": 4.3685150146484375,
+ "step": 1746
+ },
+ {
+ "epoch": 24.265502183406113,
+ "grad_norm": 0.0214633010327816,
+ "learning_rate": 0.0006,
+ "loss": 4.323866844177246,
+ "step": 1747
+ },
+ {
+ "epoch": 24.27947598253275,
+ "grad_norm": 0.021055003628134727,
+ "learning_rate": 0.0006,
+ "loss": 4.278757095336914,
+ "step": 1748
+ },
+ {
+ "epoch": 24.29344978165939,
+ "grad_norm": 0.018618572503328323,
+ "learning_rate": 0.0006,
+ "loss": 4.28849983215332,
+ "step": 1749
+ },
+ {
+ "epoch": 24.307423580786025,
+ "grad_norm": 0.018666435033082962,
+ "learning_rate": 0.0006,
+ "loss": 4.368590354919434,
+ "step": 1750
+ },
+ {
+ "epoch": 24.321397379912664,
+ "grad_norm": 0.020626146346330643,
+ "learning_rate": 0.0006,
+ "loss": 4.2951836585998535,
+ "step": 1751
+ },
+ {
+ "epoch": 24.335371179039303,
+ "grad_norm": 0.01929759792983532,
+ "learning_rate": 0.0006,
+ "loss": 4.3664703369140625,
+ "step": 1752
+ },
+ {
+ "epoch": 24.349344978165938,
+ "grad_norm": 0.020956002175807953,
+ "learning_rate": 0.0006,
+ "loss": 4.345308303833008,
+ "step": 1753
+ },
+ {
+ "epoch": 24.363318777292577,
+ "grad_norm": 0.019227957352995872,
+ "learning_rate": 0.0006,
+ "loss": 4.349849224090576,
+ "step": 1754
+ },
+ {
+ "epoch": 24.377292576419215,
+ "grad_norm": 0.015970097854733467,
+ "learning_rate": 0.0006,
+ "loss": 4.223012447357178,
+ "step": 1755
+ },
+ {
+ "epoch": 24.39126637554585,
+ "grad_norm": 0.015951646491885185,
+ "learning_rate": 0.0006,
+ "loss": 4.236420631408691,
+ "step": 1756
+ },
+ {
+ "epoch": 24.40524017467249,
+ "grad_norm": 0.016032515093684196,
+ "learning_rate": 0.0006,
+ "loss": 4.226899147033691,
+ "step": 1757
+ },
+ {
+ "epoch": 24.419213973799128,
+ "grad_norm": 0.01573132909834385,
+ "learning_rate": 0.0006,
+ "loss": 4.315916061401367,
+ "step": 1758
+ },
+ {
+ "epoch": 24.433187772925763,
+ "grad_norm": 0.015367005951702595,
+ "learning_rate": 0.0006,
+ "loss": 4.358939170837402,
+ "step": 1759
+ },
+ {
+ "epoch": 24.4471615720524,
+ "grad_norm": 0.01669907011091709,
+ "learning_rate": 0.0006,
+ "loss": 4.232147216796875,
+ "step": 1760
+ },
+ {
+ "epoch": 24.46113537117904,
+ "grad_norm": 0.018708152696490288,
+ "learning_rate": 0.0006,
+ "loss": 4.329852104187012,
+ "step": 1761
+ },
+ {
+ "epoch": 24.475109170305675,
+ "grad_norm": 0.019035987555980682,
+ "learning_rate": 0.0006,
+ "loss": 4.260340213775635,
+ "step": 1762
+ },
+ {
+ "epoch": 24.489082969432314,
+ "grad_norm": 0.01832144521176815,
+ "learning_rate": 0.0006,
+ "loss": 4.359543800354004,
+ "step": 1763
+ },
+ {
+ "epoch": 24.503056768558952,
+ "grad_norm": 0.015313294716179371,
+ "learning_rate": 0.0006,
+ "loss": 4.279088020324707,
+ "step": 1764
+ },
+ {
+ "epoch": 24.51703056768559,
+ "grad_norm": 0.015803923830389977,
+ "learning_rate": 0.0006,
+ "loss": 4.230682849884033,
+ "step": 1765
+ },
+ {
+ "epoch": 24.531004366812226,
+ "grad_norm": 0.016748374328017235,
+ "learning_rate": 0.0006,
+ "loss": 4.349661827087402,
+ "step": 1766
+ },
+ {
+ "epoch": 24.544978165938865,
+ "grad_norm": 0.015732428058981895,
+ "learning_rate": 0.0006,
+ "loss": 4.372441291809082,
+ "step": 1767
+ },
+ {
+ "epoch": 24.558951965065503,
+ "grad_norm": 0.016564829275012016,
+ "learning_rate": 0.0006,
+ "loss": 4.323090553283691,
+ "step": 1768
+ },
+ {
+ "epoch": 24.57292576419214,
+ "grad_norm": 0.01603010483086109,
+ "learning_rate": 0.0006,
+ "loss": 4.347898006439209,
+ "step": 1769
+ },
+ {
+ "epoch": 24.586899563318777,
+ "grad_norm": 0.01674809865653515,
+ "learning_rate": 0.0006,
+ "loss": 4.269165515899658,
+ "step": 1770
+ },
+ {
+ "epoch": 24.600873362445416,
+ "grad_norm": 0.017333930358290672,
+ "learning_rate": 0.0006,
+ "loss": 4.353298187255859,
+ "step": 1771
+ },
+ {
+ "epoch": 24.61484716157205,
+ "grad_norm": 0.017519580200314522,
+ "learning_rate": 0.0006,
+ "loss": 4.265199661254883,
+ "step": 1772
+ },
+ {
+ "epoch": 24.62882096069869,
+ "grad_norm": 0.017675306648015976,
+ "learning_rate": 0.0006,
+ "loss": 4.322596549987793,
+ "step": 1773
+ },
+ {
+ "epoch": 24.64279475982533,
+ "grad_norm": 0.017166534438729286,
+ "learning_rate": 0.0006,
+ "loss": 4.413900375366211,
+ "step": 1774
+ },
+ {
+ "epoch": 24.656768558951963,
+ "grad_norm": 0.016362346708774567,
+ "learning_rate": 0.0006,
+ "loss": 4.307236671447754,
+ "step": 1775
+ },
+ {
+ "epoch": 24.670742358078602,
+ "grad_norm": 0.016086872667074203,
+ "learning_rate": 0.0006,
+ "loss": 4.235316276550293,
+ "step": 1776
+ },
+ {
+ "epoch": 24.68471615720524,
+ "grad_norm": 0.01689537614583969,
+ "learning_rate": 0.0006,
+ "loss": 4.390591621398926,
+ "step": 1777
+ },
+ {
+ "epoch": 24.69868995633188,
+ "grad_norm": 0.016836725175380707,
+ "learning_rate": 0.0006,
+ "loss": 4.2990031242370605,
+ "step": 1778
+ },
+ {
+ "epoch": 24.712663755458514,
+ "grad_norm": 0.015627335757017136,
+ "learning_rate": 0.0006,
+ "loss": 4.166745185852051,
+ "step": 1779
+ },
+ {
+ "epoch": 24.726637554585153,
+ "grad_norm": 0.015007016249001026,
+ "learning_rate": 0.0006,
+ "loss": 4.295273303985596,
+ "step": 1780
+ },
+ {
+ "epoch": 24.74061135371179,
+ "grad_norm": 0.016084209084510803,
+ "learning_rate": 0.0006,
+ "loss": 4.342424392700195,
+ "step": 1781
+ },
+ {
+ "epoch": 24.754585152838427,
+ "grad_norm": 0.01595073565840721,
+ "learning_rate": 0.0006,
+ "loss": 4.3060102462768555,
+ "step": 1782
+ },
+ {
+ "epoch": 24.768558951965066,
+ "grad_norm": 0.016113461926579475,
+ "learning_rate": 0.0006,
+ "loss": 4.267908096313477,
+ "step": 1783
+ },
+ {
+ "epoch": 24.782532751091704,
+ "grad_norm": 0.01564556173980236,
+ "learning_rate": 0.0006,
+ "loss": 4.324642658233643,
+ "step": 1784
+ },
+ {
+ "epoch": 24.79650655021834,
+ "grad_norm": 0.01575268618762493,
+ "learning_rate": 0.0006,
+ "loss": 4.328181743621826,
+ "step": 1785
+ },
+ {
+ "epoch": 24.810480349344978,
+ "grad_norm": 0.016830824315547943,
+ "learning_rate": 0.0006,
+ "loss": 4.247906684875488,
+ "step": 1786
+ },
+ {
+ "epoch": 24.824454148471617,
+ "grad_norm": 0.017989547923207283,
+ "learning_rate": 0.0006,
+ "loss": 4.228281021118164,
+ "step": 1787
+ },
+ {
+ "epoch": 24.83842794759825,
+ "grad_norm": 0.016783086583018303,
+ "learning_rate": 0.0006,
+ "loss": 4.2880964279174805,
+ "step": 1788
+ },
+ {
+ "epoch": 24.85240174672489,
+ "grad_norm": 0.016467561945319176,
+ "learning_rate": 0.0006,
+ "loss": 4.36118745803833,
+ "step": 1789
+ },
+ {
+ "epoch": 24.86637554585153,
+ "grad_norm": 0.01780523732304573,
+ "learning_rate": 0.0006,
+ "loss": 4.172321319580078,
+ "step": 1790
+ },
+ {
+ "epoch": 24.880349344978168,
+ "grad_norm": 0.017102031037211418,
+ "learning_rate": 0.0006,
+ "loss": 4.352826118469238,
+ "step": 1791
+ },
+ {
+ "epoch": 24.894323144104803,
+ "grad_norm": 0.016409218311309814,
+ "learning_rate": 0.0006,
+ "loss": 4.212162017822266,
+ "step": 1792
+ },
+ {
+ "epoch": 24.90829694323144,
+ "grad_norm": 0.018660584464669228,
+ "learning_rate": 0.0006,
+ "loss": 4.346684455871582,
+ "step": 1793
+ },
+ {
+ "epoch": 24.92227074235808,
+ "grad_norm": 0.017389440909028053,
+ "learning_rate": 0.0006,
+ "loss": 4.197913646697998,
+ "step": 1794
+ },
+ {
+ "epoch": 24.936244541484715,
+ "grad_norm": 0.01596485823392868,
+ "learning_rate": 0.0006,
+ "loss": 4.370617866516113,
+ "step": 1795
+ },
+ {
+ "epoch": 24.950218340611354,
+ "grad_norm": 0.01709776371717453,
+ "learning_rate": 0.0006,
+ "loss": 4.313847064971924,
+ "step": 1796
+ },
+ {
+ "epoch": 24.964192139737992,
+ "grad_norm": 0.016571134328842163,
+ "learning_rate": 0.0006,
+ "loss": 4.1895856857299805,
+ "step": 1797
+ },
+ {
+ "epoch": 24.978165938864628,
+ "grad_norm": 0.016851048916578293,
+ "learning_rate": 0.0006,
+ "loss": 4.350772857666016,
+ "step": 1798
+ },
+ {
+ "epoch": 24.992139737991266,
+ "grad_norm": 0.017007440328598022,
+ "learning_rate": 0.0006,
+ "loss": 4.399008750915527,
+ "step": 1799
+ },
+ {
+ "epoch": 25.0,
+ "grad_norm": 0.01791212148964405,
+ "learning_rate": 0.0006,
+ "loss": 4.306121826171875,
+ "step": 1800
+ },
+ {
+ "epoch": 25.0,
+ "eval_loss": 4.636417865753174,
+ "eval_runtime": 56.8296,
+ "eval_samples_per_second": 42.971,
+ "eval_steps_per_second": 1.355,
+ "step": 1800
+ },
+ {
+ "epoch": 25.01397379912664,
+ "grad_norm": 0.01729617640376091,
+ "learning_rate": 0.0006,
+ "loss": 4.257018089294434,
+ "step": 1801
+ },
+ {
+ "epoch": 25.027947598253274,
+ "grad_norm": 0.017466643825173378,
+ "learning_rate": 0.0006,
+ "loss": 4.307605266571045,
+ "step": 1802
+ },
+ {
+ "epoch": 25.041921397379912,
+ "grad_norm": 0.018614279106259346,
+ "learning_rate": 0.0006,
+ "loss": 4.360141277313232,
+ "step": 1803
+ },
+ {
+ "epoch": 25.05589519650655,
+ "grad_norm": 0.02010771445930004,
+ "learning_rate": 0.0006,
+ "loss": 4.196428298950195,
+ "step": 1804
+ },
+ {
+ "epoch": 25.069868995633186,
+ "grad_norm": 0.02090228535234928,
+ "learning_rate": 0.0006,
+ "loss": 4.288947105407715,
+ "step": 1805
+ },
+ {
+ "epoch": 25.083842794759825,
+ "grad_norm": 0.02050255797803402,
+ "learning_rate": 0.0006,
+ "loss": 4.259582042694092,
+ "step": 1806
+ },
+ {
+ "epoch": 25.097816593886463,
+ "grad_norm": 0.021844089031219482,
+ "learning_rate": 0.0006,
+ "loss": 4.405138969421387,
+ "step": 1807
+ },
+ {
+ "epoch": 25.111790393013102,
+ "grad_norm": 0.022287018597126007,
+ "learning_rate": 0.0006,
+ "loss": 4.333793640136719,
+ "step": 1808
+ },
+ {
+ "epoch": 25.125764192139737,
+ "grad_norm": 0.02288329415023327,
+ "learning_rate": 0.0006,
+ "loss": 4.3910980224609375,
+ "step": 1809
+ },
+ {
+ "epoch": 25.139737991266376,
+ "grad_norm": 0.02373199723660946,
+ "learning_rate": 0.0006,
+ "loss": 4.32921028137207,
+ "step": 1810
+ },
+ {
+ "epoch": 25.153711790393015,
+ "grad_norm": 0.024960234761238098,
+ "learning_rate": 0.0006,
+ "loss": 4.295609474182129,
+ "step": 1811
+ },
+ {
+ "epoch": 25.16768558951965,
+ "grad_norm": 0.026743892580270767,
+ "learning_rate": 0.0006,
+ "loss": 4.315308094024658,
+ "step": 1812
+ },
+ {
+ "epoch": 25.18165938864629,
+ "grad_norm": 0.022138185799121857,
+ "learning_rate": 0.0006,
+ "loss": 4.360208511352539,
+ "step": 1813
+ },
+ {
+ "epoch": 25.195633187772927,
+ "grad_norm": 0.022120574489235878,
+ "learning_rate": 0.0006,
+ "loss": 4.1973772048950195,
+ "step": 1814
+ },
+ {
+ "epoch": 25.209606986899562,
+ "grad_norm": 0.02230502851307392,
+ "learning_rate": 0.0006,
+ "loss": 4.285558700561523,
+ "step": 1815
+ },
+ {
+ "epoch": 25.2235807860262,
+ "grad_norm": 0.022301986813545227,
+ "learning_rate": 0.0006,
+ "loss": 4.238343238830566,
+ "step": 1816
+ },
+ {
+ "epoch": 25.23755458515284,
+ "grad_norm": 0.020050447434186935,
+ "learning_rate": 0.0006,
+ "loss": 4.298235893249512,
+ "step": 1817
+ },
+ {
+ "epoch": 25.251528384279474,
+ "grad_norm": 0.01987724006175995,
+ "learning_rate": 0.0006,
+ "loss": 4.301384449005127,
+ "step": 1818
+ },
+ {
+ "epoch": 25.265502183406113,
+ "grad_norm": 0.020440855994820595,
+ "learning_rate": 0.0006,
+ "loss": 4.269218444824219,
+ "step": 1819
+ },
+ {
+ "epoch": 25.27947598253275,
+ "grad_norm": 0.01891893707215786,
+ "learning_rate": 0.0006,
+ "loss": 4.2401957511901855,
+ "step": 1820
+ },
+ {
+ "epoch": 25.29344978165939,
+ "grad_norm": 0.01903688907623291,
+ "learning_rate": 0.0006,
+ "loss": 4.139670372009277,
+ "step": 1821
+ },
+ {
+ "epoch": 25.307423580786025,
+ "grad_norm": 0.018856938928365707,
+ "learning_rate": 0.0006,
+ "loss": 4.233214378356934,
+ "step": 1822
+ },
+ {
+ "epoch": 25.321397379912664,
+ "grad_norm": 0.017791323363780975,
+ "learning_rate": 0.0006,
+ "loss": 4.2342023849487305,
+ "step": 1823
+ },
+ {
+ "epoch": 25.335371179039303,
+ "grad_norm": 0.017304565757513046,
+ "learning_rate": 0.0006,
+ "loss": 4.355198860168457,
+ "step": 1824
+ },
+ {
+ "epoch": 25.349344978165938,
+ "grad_norm": 0.01576206088066101,
+ "learning_rate": 0.0006,
+ "loss": 4.114147186279297,
+ "step": 1825
+ },
+ {
+ "epoch": 25.363318777292577,
+ "grad_norm": 0.015105154365301132,
+ "learning_rate": 0.0006,
+ "loss": 4.313795566558838,
+ "step": 1826
+ },
+ {
+ "epoch": 25.377292576419215,
+ "grad_norm": 0.01563587784767151,
+ "learning_rate": 0.0006,
+ "loss": 4.308282852172852,
+ "step": 1827
+ },
+ {
+ "epoch": 25.39126637554585,
+ "grad_norm": 0.015980906784534454,
+ "learning_rate": 0.0006,
+ "loss": 4.345149040222168,
+ "step": 1828
+ },
+ {
+ "epoch": 25.40524017467249,
+ "grad_norm": 0.01669284701347351,
+ "learning_rate": 0.0006,
+ "loss": 4.237611293792725,
+ "step": 1829
+ },
+ {
+ "epoch": 25.419213973799128,
+ "grad_norm": 0.016202064231038094,
+ "learning_rate": 0.0006,
+ "loss": 4.3072309494018555,
+ "step": 1830
+ },
+ {
+ "epoch": 25.433187772925763,
+ "grad_norm": 0.01478662807494402,
+ "learning_rate": 0.0006,
+ "loss": 4.276215076446533,
+ "step": 1831
+ },
+ {
+ "epoch": 25.4471615720524,
+ "grad_norm": 0.015168731100857258,
+ "learning_rate": 0.0006,
+ "loss": 4.37509822845459,
+ "step": 1832
+ },
+ {
+ "epoch": 25.46113537117904,
+ "grad_norm": 0.01648980937898159,
+ "learning_rate": 0.0006,
+ "loss": 4.349167346954346,
+ "step": 1833
+ },
+ {
+ "epoch": 25.475109170305675,
+ "grad_norm": 0.01708398200571537,
+ "learning_rate": 0.0006,
+ "loss": 4.346580505371094,
+ "step": 1834
+ },
+ {
+ "epoch": 25.489082969432314,
+ "grad_norm": 0.017370784655213356,
+ "learning_rate": 0.0006,
+ "loss": 4.299466133117676,
+ "step": 1835
+ },
+ {
+ "epoch": 25.503056768558952,
+ "grad_norm": 0.016248228028416634,
+ "learning_rate": 0.0006,
+ "loss": 4.184610366821289,
+ "step": 1836
+ },
+ {
+ "epoch": 25.51703056768559,
+ "grad_norm": 0.017106913030147552,
+ "learning_rate": 0.0006,
+ "loss": 4.366856098175049,
+ "step": 1837
+ },
+ {
+ "epoch": 25.531004366812226,
+ "grad_norm": 0.016071965917944908,
+ "learning_rate": 0.0006,
+ "loss": 4.297691345214844,
+ "step": 1838
+ },
+ {
+ "epoch": 25.544978165938865,
+ "grad_norm": 0.016451986506581306,
+ "learning_rate": 0.0006,
+ "loss": 4.229983329772949,
+ "step": 1839
+ },
+ {
+ "epoch": 25.558951965065503,
+ "grad_norm": 0.016195081174373627,
+ "learning_rate": 0.0006,
+ "loss": 4.259893417358398,
+ "step": 1840
+ },
+ {
+ "epoch": 25.57292576419214,
+ "grad_norm": 0.0165175162255764,
+ "learning_rate": 0.0006,
+ "loss": 4.3348236083984375,
+ "step": 1841
+ },
+ {
+ "epoch": 25.586899563318777,
+ "grad_norm": 0.01651517115533352,
+ "learning_rate": 0.0006,
+ "loss": 4.319541931152344,
+ "step": 1842
+ },
+ {
+ "epoch": 25.600873362445416,
+ "grad_norm": 0.016143690794706345,
+ "learning_rate": 0.0006,
+ "loss": 4.345652103424072,
+ "step": 1843
+ },
+ {
+ "epoch": 25.61484716157205,
+ "grad_norm": 0.015140696428716183,
+ "learning_rate": 0.0006,
+ "loss": 4.212047576904297,
+ "step": 1844
+ },
+ {
+ "epoch": 25.62882096069869,
+ "grad_norm": 0.014637443237006664,
+ "learning_rate": 0.0006,
+ "loss": 4.187453269958496,
+ "step": 1845
+ },
+ {
+ "epoch": 25.64279475982533,
+ "grad_norm": 0.014585614204406738,
+ "learning_rate": 0.0006,
+ "loss": 4.253936767578125,
+ "step": 1846
+ },
+ {
+ "epoch": 25.656768558951963,
+ "grad_norm": 0.015158289112150669,
+ "learning_rate": 0.0006,
+ "loss": 4.34612512588501,
+ "step": 1847
+ },
+ {
+ "epoch": 25.670742358078602,
+ "grad_norm": 0.015202849172055721,
+ "learning_rate": 0.0006,
+ "loss": 4.270634651184082,
+ "step": 1848
+ },
+ {
+ "epoch": 25.68471615720524,
+ "grad_norm": 0.015363575890660286,
+ "learning_rate": 0.0006,
+ "loss": 4.271080493927002,
+ "step": 1849
+ },
+ {
+ "epoch": 25.69868995633188,
+ "grad_norm": 0.014360226690769196,
+ "learning_rate": 0.0006,
+ "loss": 4.375033855438232,
+ "step": 1850
+ },
+ {
+ "epoch": 25.712663755458514,
+ "grad_norm": 0.014263181947171688,
+ "learning_rate": 0.0006,
+ "loss": 4.424138069152832,
+ "step": 1851
+ },
+ {
+ "epoch": 25.726637554585153,
+ "grad_norm": 0.014398688450455666,
+ "learning_rate": 0.0006,
+ "loss": 4.410154819488525,
+ "step": 1852
+ },
+ {
+ "epoch": 25.74061135371179,
+ "grad_norm": 0.016131488606333733,
+ "learning_rate": 0.0006,
+ "loss": 4.324014663696289,
+ "step": 1853
+ },
+ {
+ "epoch": 25.754585152838427,
+ "grad_norm": 0.017568735405802727,
+ "learning_rate": 0.0006,
+ "loss": 4.367861747741699,
+ "step": 1854
+ },
+ {
+ "epoch": 25.768558951965066,
+ "grad_norm": 0.017907511442899704,
+ "learning_rate": 0.0006,
+ "loss": 4.352965354919434,
+ "step": 1855
+ },
+ {
+ "epoch": 25.782532751091704,
+ "grad_norm": 0.016918016597628593,
+ "learning_rate": 0.0006,
+ "loss": 4.252397537231445,
+ "step": 1856
+ },
+ {
+ "epoch": 25.79650655021834,
+ "grad_norm": 0.017274610698223114,
+ "learning_rate": 0.0006,
+ "loss": 4.342309951782227,
+ "step": 1857
+ },
+ {
+ "epoch": 25.810480349344978,
+ "grad_norm": 0.017879825085401535,
+ "learning_rate": 0.0006,
+ "loss": 4.330987453460693,
+ "step": 1858
+ },
+ {
+ "epoch": 25.824454148471617,
+ "grad_norm": 0.01736099272966385,
+ "learning_rate": 0.0006,
+ "loss": 4.307255744934082,
+ "step": 1859
+ },
+ {
+ "epoch": 25.83842794759825,
+ "grad_norm": 0.016575824469327927,
+ "learning_rate": 0.0006,
+ "loss": 4.262016773223877,
+ "step": 1860
+ },
+ {
+ "epoch": 25.85240174672489,
+ "grad_norm": 0.015181140042841434,
+ "learning_rate": 0.0006,
+ "loss": 4.291863441467285,
+ "step": 1861
+ },
+ {
+ "epoch": 25.86637554585153,
+ "grad_norm": 0.015965687111020088,
+ "learning_rate": 0.0006,
+ "loss": 4.424836158752441,
+ "step": 1862
+ },
+ {
+ "epoch": 25.880349344978168,
+ "grad_norm": 0.01687219925224781,
+ "learning_rate": 0.0006,
+ "loss": 4.343531131744385,
+ "step": 1863
+ },
+ {
+ "epoch": 25.894323144104803,
+ "grad_norm": 0.015115504153072834,
+ "learning_rate": 0.0006,
+ "loss": 4.357808589935303,
+ "step": 1864
+ },
+ {
+ "epoch": 25.90829694323144,
+ "grad_norm": 0.014150998555123806,
+ "learning_rate": 0.0006,
+ "loss": 4.324550151824951,
+ "step": 1865
+ },
+ {
+ "epoch": 25.92227074235808,
+ "grad_norm": 0.014989510178565979,
+ "learning_rate": 0.0006,
+ "loss": 4.380428314208984,
+ "step": 1866
+ },
+ {
+ "epoch": 25.936244541484715,
+ "grad_norm": 0.01527960691601038,
+ "learning_rate": 0.0006,
+ "loss": 4.399374961853027,
+ "step": 1867
+ },
+ {
+ "epoch": 25.950218340611354,
+ "grad_norm": 0.015434633940458298,
+ "learning_rate": 0.0006,
+ "loss": 4.260984420776367,
+ "step": 1868
+ },
+ {
+ "epoch": 25.964192139737992,
+ "grad_norm": 0.015370228327810764,
+ "learning_rate": 0.0006,
+ "loss": 4.3340253829956055,
+ "step": 1869
+ },
+ {
+ "epoch": 25.978165938864628,
+ "grad_norm": 0.015348542481660843,
+ "learning_rate": 0.0006,
+ "loss": 4.369132995605469,
+ "step": 1870
+ },
+ {
+ "epoch": 25.992139737991266,
+ "grad_norm": 0.015354936942458153,
+ "learning_rate": 0.0006,
+ "loss": 4.381937026977539,
+ "step": 1871
+ },
+ {
+ "epoch": 26.0,
+ "grad_norm": 0.01796272210776806,
+ "learning_rate": 0.0006,
+ "loss": 4.286768436431885,
+ "step": 1872
+ },
+ {
+ "epoch": 26.0,
+ "eval_loss": 4.576984405517578,
+ "eval_runtime": 57.2091,
+ "eval_samples_per_second": 42.685,
+ "eval_steps_per_second": 1.346,
+ "step": 1872
+ },
+ {
+ "epoch": 26.01397379912664,
+ "grad_norm": 0.01748845726251602,
+ "learning_rate": 0.0006,
+ "loss": 4.337504863739014,
+ "step": 1873
+ },
+ {
+ "epoch": 26.027947598253274,
+ "grad_norm": 0.019244296476244926,
+ "learning_rate": 0.0006,
+ "loss": 4.264280796051025,
+ "step": 1874
+ },
+ {
+ "epoch": 26.041921397379912,
+ "grad_norm": 0.0215692650526762,
+ "learning_rate": 0.0006,
+ "loss": 4.350830078125,
+ "step": 1875
+ },
+ {
+ "epoch": 26.05589519650655,
+ "grad_norm": 0.022189252078533173,
+ "learning_rate": 0.0006,
+ "loss": 4.299098968505859,
+ "step": 1876
+ },
+ {
+ "epoch": 26.069868995633186,
+ "grad_norm": 0.022384043782949448,
+ "learning_rate": 0.0006,
+ "loss": 4.1185173988342285,
+ "step": 1877
+ },
+ {
+ "epoch": 26.083842794759825,
+ "grad_norm": 0.02076014317572117,
+ "learning_rate": 0.0006,
+ "loss": 4.297612190246582,
+ "step": 1878
+ },
+ {
+ "epoch": 26.097816593886463,
+ "grad_norm": 0.020843051373958588,
+ "learning_rate": 0.0006,
+ "loss": 4.294098854064941,
+ "step": 1879
+ },
+ {
+ "epoch": 26.111790393013102,
+ "grad_norm": 0.020654456689953804,
+ "learning_rate": 0.0006,
+ "loss": 4.267856597900391,
+ "step": 1880
+ },
+ {
+ "epoch": 26.125764192139737,
+ "grad_norm": 0.01860472559928894,
+ "learning_rate": 0.0006,
+ "loss": 4.284232139587402,
+ "step": 1881
+ },
+ {
+ "epoch": 26.139737991266376,
+ "grad_norm": 0.018719421699643135,
+ "learning_rate": 0.0006,
+ "loss": 4.277214527130127,
+ "step": 1882
+ },
+ {
+ "epoch": 26.153711790393015,
+ "grad_norm": 0.020903829485177994,
+ "learning_rate": 0.0006,
+ "loss": 4.254947662353516,
+ "step": 1883
+ },
+ {
+ "epoch": 26.16768558951965,
+ "grad_norm": 0.02020149491727352,
+ "learning_rate": 0.0006,
+ "loss": 4.2526936531066895,
+ "step": 1884
+ },
+ {
+ "epoch": 26.18165938864629,
+ "grad_norm": 0.0180392786860466,
+ "learning_rate": 0.0006,
+ "loss": 4.309922695159912,
+ "step": 1885
+ },
+ {
+ "epoch": 26.195633187772927,
+ "grad_norm": 0.018163125962018967,
+ "learning_rate": 0.0006,
+ "loss": 4.13860559463501,
+ "step": 1886
+ },
+ {
+ "epoch": 26.209606986899562,
+ "grad_norm": 0.01853291131556034,
+ "learning_rate": 0.0006,
+ "loss": 4.190926551818848,
+ "step": 1887
+ },
+ {
+ "epoch": 26.2235807860262,
+ "grad_norm": 0.01757677271962166,
+ "learning_rate": 0.0006,
+ "loss": 4.163999557495117,
+ "step": 1888
+ },
+ {
+ "epoch": 26.23755458515284,
+ "grad_norm": 0.01767992228269577,
+ "learning_rate": 0.0006,
+ "loss": 4.184430122375488,
+ "step": 1889
+ },
+ {
+ "epoch": 26.251528384279474,
+ "grad_norm": 0.018139973282814026,
+ "learning_rate": 0.0006,
+ "loss": 4.300766944885254,
+ "step": 1890
+ },
+ {
+ "epoch": 26.265502183406113,
+ "grad_norm": 0.020184461027383804,
+ "learning_rate": 0.0006,
+ "loss": 4.232537269592285,
+ "step": 1891
+ },
+ {
+ "epoch": 26.27947598253275,
+ "grad_norm": 0.02028856985270977,
+ "learning_rate": 0.0006,
+ "loss": 4.324883460998535,
+ "step": 1892
+ },
+ {
+ "epoch": 26.29344978165939,
+ "grad_norm": 0.017120616510510445,
+ "learning_rate": 0.0006,
+ "loss": 4.23493766784668,
+ "step": 1893
+ },
+ {
+ "epoch": 26.307423580786025,
+ "grad_norm": 0.016202978789806366,
+ "learning_rate": 0.0006,
+ "loss": 4.197602272033691,
+ "step": 1894
+ },
+ {
+ "epoch": 26.321397379912664,
+ "grad_norm": 0.017231421545147896,
+ "learning_rate": 0.0006,
+ "loss": 4.302001953125,
+ "step": 1895
+ },
+ {
+ "epoch": 26.335371179039303,
+ "grad_norm": 0.01725655607879162,
+ "learning_rate": 0.0006,
+ "loss": 4.356800079345703,
+ "step": 1896
+ },
+ {
+ "epoch": 26.349344978165938,
+ "grad_norm": 0.016657443717122078,
+ "learning_rate": 0.0006,
+ "loss": 4.266307830810547,
+ "step": 1897
+ },
+ {
+ "epoch": 26.363318777292577,
+ "grad_norm": 0.01616556942462921,
+ "learning_rate": 0.0006,
+ "loss": 4.307868003845215,
+ "step": 1898
+ },
+ {
+ "epoch": 26.377292576419215,
+ "grad_norm": 0.016621306538581848,
+ "learning_rate": 0.0006,
+ "loss": 4.172905921936035,
+ "step": 1899
+ },
+ {
+ "epoch": 26.39126637554585,
+ "grad_norm": 0.016243206337094307,
+ "learning_rate": 0.0006,
+ "loss": 4.162896156311035,
+ "step": 1900
+ },
+ {
+ "epoch": 26.40524017467249,
+ "grad_norm": 0.016663808375597,
+ "learning_rate": 0.0006,
+ "loss": 4.258232116699219,
+ "step": 1901
+ },
+ {
+ "epoch": 26.419213973799128,
+ "grad_norm": 0.01677478663623333,
+ "learning_rate": 0.0006,
+ "loss": 4.336090087890625,
+ "step": 1902
+ },
+ {
+ "epoch": 26.433187772925763,
+ "grad_norm": 0.01641303487122059,
+ "learning_rate": 0.0006,
+ "loss": 4.191476821899414,
+ "step": 1903
+ },
+ {
+ "epoch": 26.4471615720524,
+ "grad_norm": 0.016700323671102524,
+ "learning_rate": 0.0006,
+ "loss": 4.299427509307861,
+ "step": 1904
+ },
+ {
+ "epoch": 26.46113537117904,
+ "grad_norm": 0.017595035955309868,
+ "learning_rate": 0.0006,
+ "loss": 4.312828063964844,
+ "step": 1905
+ },
+ {
+ "epoch": 26.475109170305675,
+ "grad_norm": 0.020300284028053284,
+ "learning_rate": 0.0006,
+ "loss": 4.28331184387207,
+ "step": 1906
+ },
+ {
+ "epoch": 26.489082969432314,
+ "grad_norm": 0.02198321744799614,
+ "learning_rate": 0.0006,
+ "loss": 4.3280839920043945,
+ "step": 1907
+ },
+ {
+ "epoch": 26.503056768558952,
+ "grad_norm": 0.02003200724720955,
+ "learning_rate": 0.0006,
+ "loss": 4.325623989105225,
+ "step": 1908
+ },
+ {
+ "epoch": 26.51703056768559,
+ "grad_norm": 0.01870287023484707,
+ "learning_rate": 0.0006,
+ "loss": 4.251203536987305,
+ "step": 1909
+ },
+ {
+ "epoch": 26.531004366812226,
+ "grad_norm": 0.01835448481142521,
+ "learning_rate": 0.0006,
+ "loss": 4.335144996643066,
+ "step": 1910
+ },
+ {
+ "epoch": 26.544978165938865,
+ "grad_norm": 0.02046327292919159,
+ "learning_rate": 0.0006,
+ "loss": 4.238023281097412,
+ "step": 1911
+ },
+ {
+ "epoch": 26.558951965065503,
+ "grad_norm": 0.019771311432123184,
+ "learning_rate": 0.0006,
+ "loss": 4.353594779968262,
+ "step": 1912
+ },
+ {
+ "epoch": 26.57292576419214,
+ "grad_norm": 0.018090544268488884,
+ "learning_rate": 0.0006,
+ "loss": 4.304343223571777,
+ "step": 1913
+ },
+ {
+ "epoch": 26.586899563318777,
+ "grad_norm": 0.017489777877926826,
+ "learning_rate": 0.0006,
+ "loss": 4.204375267028809,
+ "step": 1914
+ },
+ {
+ "epoch": 26.600873362445416,
+ "grad_norm": 0.016785891726613045,
+ "learning_rate": 0.0006,
+ "loss": 4.286669731140137,
+ "step": 1915
+ },
+ {
+ "epoch": 26.61484716157205,
+ "grad_norm": 0.016382789239287376,
+ "learning_rate": 0.0006,
+ "loss": 4.347662448883057,
+ "step": 1916
+ },
+ {
+ "epoch": 26.62882096069869,
+ "grad_norm": 0.016581876203417778,
+ "learning_rate": 0.0006,
+ "loss": 4.2716193199157715,
+ "step": 1917
+ },
+ {
+ "epoch": 26.64279475982533,
+ "grad_norm": 0.015627184882760048,
+ "learning_rate": 0.0006,
+ "loss": 4.311489105224609,
+ "step": 1918
+ },
+ {
+ "epoch": 26.656768558951963,
+ "grad_norm": 0.015908483415842056,
+ "learning_rate": 0.0006,
+ "loss": 4.262099742889404,
+ "step": 1919
+ },
+ {
+ "epoch": 26.670742358078602,
+ "grad_norm": 0.015576236881315708,
+ "learning_rate": 0.0006,
+ "loss": 4.275871276855469,
+ "step": 1920
+ },
+ {
+ "epoch": 26.68471615720524,
+ "grad_norm": 0.014621064998209476,
+ "learning_rate": 0.0006,
+ "loss": 4.184260845184326,
+ "step": 1921
+ },
+ {
+ "epoch": 26.69868995633188,
+ "grad_norm": 0.014797811396420002,
+ "learning_rate": 0.0006,
+ "loss": 4.3412628173828125,
+ "step": 1922
+ },
+ {
+ "epoch": 26.712663755458514,
+ "grad_norm": 0.015610570088028908,
+ "learning_rate": 0.0006,
+ "loss": 4.181094646453857,
+ "step": 1923
+ },
+ {
+ "epoch": 26.726637554585153,
+ "grad_norm": 0.016048265621066093,
+ "learning_rate": 0.0006,
+ "loss": 4.365290641784668,
+ "step": 1924
+ },
+ {
+ "epoch": 26.74061135371179,
+ "grad_norm": 0.016604071483016014,
+ "learning_rate": 0.0006,
+ "loss": 4.2628912925720215,
+ "step": 1925
+ },
+ {
+ "epoch": 26.754585152838427,
+ "grad_norm": 0.017044221982359886,
+ "learning_rate": 0.0006,
+ "loss": 4.378512382507324,
+ "step": 1926
+ },
+ {
+ "epoch": 26.768558951965066,
+ "grad_norm": 0.01746384985744953,
+ "learning_rate": 0.0006,
+ "loss": 4.3888773918151855,
+ "step": 1927
+ },
+ {
+ "epoch": 26.782532751091704,
+ "grad_norm": 0.01572711206972599,
+ "learning_rate": 0.0006,
+ "loss": 4.189935684204102,
+ "step": 1928
+ },
+ {
+ "epoch": 26.79650655021834,
+ "grad_norm": 0.014976629987359047,
+ "learning_rate": 0.0006,
+ "loss": 4.280525207519531,
+ "step": 1929
+ },
+ {
+ "epoch": 26.810480349344978,
+ "grad_norm": 0.014804664999246597,
+ "learning_rate": 0.0006,
+ "loss": 4.270630836486816,
+ "step": 1930
+ },
+ {
+ "epoch": 26.824454148471617,
+ "grad_norm": 0.014641920104622841,
+ "learning_rate": 0.0006,
+ "loss": 4.319505214691162,
+ "step": 1931
+ },
+ {
+ "epoch": 26.83842794759825,
+ "grad_norm": 0.014928505755960941,
+ "learning_rate": 0.0006,
+ "loss": 4.180147647857666,
+ "step": 1932
+ },
+ {
+ "epoch": 26.85240174672489,
+ "grad_norm": 0.016130365431308746,
+ "learning_rate": 0.0006,
+ "loss": 4.260660648345947,
+ "step": 1933
+ },
+ {
+ "epoch": 26.86637554585153,
+ "grad_norm": 0.015642493963241577,
+ "learning_rate": 0.0006,
+ "loss": 4.304064750671387,
+ "step": 1934
+ },
+ {
+ "epoch": 26.880349344978168,
+ "grad_norm": 0.015988217666745186,
+ "learning_rate": 0.0006,
+ "loss": 4.2548017501831055,
+ "step": 1935
+ },
+ {
+ "epoch": 26.894323144104803,
+ "grad_norm": 0.016501398757100105,
+ "learning_rate": 0.0006,
+ "loss": 4.260058879852295,
+ "step": 1936
+ },
+ {
+ "epoch": 26.90829694323144,
+ "grad_norm": 0.01645040698349476,
+ "learning_rate": 0.0006,
+ "loss": 4.483081340789795,
+ "step": 1937
+ },
+ {
+ "epoch": 26.92227074235808,
+ "grad_norm": 0.016140863299369812,
+ "learning_rate": 0.0006,
+ "loss": 4.1776838302612305,
+ "step": 1938
+ },
+ {
+ "epoch": 26.936244541484715,
+ "grad_norm": 0.016681019216775894,
+ "learning_rate": 0.0006,
+ "loss": 4.224747657775879,
+ "step": 1939
+ },
+ {
+ "epoch": 26.950218340611354,
+ "grad_norm": 0.01650378853082657,
+ "learning_rate": 0.0006,
+ "loss": 4.305703163146973,
+ "step": 1940
+ },
+ {
+ "epoch": 26.964192139737992,
+ "grad_norm": 0.017051683738827705,
+ "learning_rate": 0.0006,
+ "loss": 4.285835266113281,
+ "step": 1941
+ },
+ {
+ "epoch": 26.978165938864628,
+ "grad_norm": 0.016894327476620674,
+ "learning_rate": 0.0006,
+ "loss": 4.318901538848877,
+ "step": 1942
+ },
+ {
+ "epoch": 26.992139737991266,
+ "grad_norm": 0.017313100397586823,
+ "learning_rate": 0.0006,
+ "loss": 4.281434535980225,
+ "step": 1943
+ },
+ {
+ "epoch": 27.0,
+ "grad_norm": 0.018450884148478508,
+ "learning_rate": 0.0006,
+ "loss": 4.335104942321777,
+ "step": 1944
+ },
+ {
+ "epoch": 27.0,
+ "eval_loss": 4.640687465667725,
+ "eval_runtime": 56.9624,
+ "eval_samples_per_second": 42.87,
+ "eval_steps_per_second": 1.352,
+ "step": 1944
+ },
+ {
+ "epoch": 27.01397379912664,
+ "grad_norm": 0.017755698412656784,
+ "learning_rate": 0.0006,
+ "loss": 4.278824329376221,
+ "step": 1945
+ },
+ {
+ "epoch": 27.027947598253274,
+ "grad_norm": 0.01724310778081417,
+ "learning_rate": 0.0006,
+ "loss": 4.311049461364746,
+ "step": 1946
+ },
+ {
+ "epoch": 27.041921397379912,
+ "grad_norm": 0.01690123789012432,
+ "learning_rate": 0.0006,
+ "loss": 4.261994361877441,
+ "step": 1947
+ },
+ {
+ "epoch": 27.05589519650655,
+ "grad_norm": 0.018406696617603302,
+ "learning_rate": 0.0006,
+ "loss": 4.1925153732299805,
+ "step": 1948
+ },
+ {
+ "epoch": 27.069868995633186,
+ "grad_norm": 0.019945867359638214,
+ "learning_rate": 0.0006,
+ "loss": 4.108537197113037,
+ "step": 1949
+ },
+ {
+ "epoch": 27.083842794759825,
+ "grad_norm": 0.020316628739237785,
+ "learning_rate": 0.0006,
+ "loss": 4.209043025970459,
+ "step": 1950
+ },
+ {
+ "epoch": 27.097816593886463,
+ "grad_norm": 0.019690370187163353,
+ "learning_rate": 0.0006,
+ "loss": 4.195789337158203,
+ "step": 1951
+ },
+ {
+ "epoch": 27.111790393013102,
+ "grad_norm": 0.01785232499241829,
+ "learning_rate": 0.0006,
+ "loss": 4.21466064453125,
+ "step": 1952
+ },
+ {
+ "epoch": 27.125764192139737,
+ "grad_norm": 0.0164767038077116,
+ "learning_rate": 0.0006,
+ "loss": 4.270247459411621,
+ "step": 1953
+ },
+ {
+ "epoch": 27.139737991266376,
+ "grad_norm": 0.018008455634117126,
+ "learning_rate": 0.0006,
+ "loss": 4.201042175292969,
+ "step": 1954
+ },
+ {
+ "epoch": 27.153711790393015,
+ "grad_norm": 0.01807340793311596,
+ "learning_rate": 0.0006,
+ "loss": 4.25289249420166,
+ "step": 1955
+ },
+ {
+ "epoch": 27.16768558951965,
+ "grad_norm": 0.016246909275650978,
+ "learning_rate": 0.0006,
+ "loss": 4.116283416748047,
+ "step": 1956
+ },
+ {
+ "epoch": 27.18165938864629,
+ "grad_norm": 0.017545776441693306,
+ "learning_rate": 0.0006,
+ "loss": 4.316999435424805,
+ "step": 1957
+ },
+ {
+ "epoch": 27.195633187772927,
+ "grad_norm": 0.017915375530719757,
+ "learning_rate": 0.0006,
+ "loss": 4.253859519958496,
+ "step": 1958
+ },
+ {
+ "epoch": 27.209606986899562,
+ "grad_norm": 0.018112223595380783,
+ "learning_rate": 0.0006,
+ "loss": 4.2541022300720215,
+ "step": 1959
+ },
+ {
+ "epoch": 27.2235807860262,
+ "grad_norm": 0.01846139505505562,
+ "learning_rate": 0.0006,
+ "loss": 4.20438289642334,
+ "step": 1960
+ },
+ {
+ "epoch": 27.23755458515284,
+ "grad_norm": 0.01943897269666195,
+ "learning_rate": 0.0006,
+ "loss": 4.2799577713012695,
+ "step": 1961
+ },
+ {
+ "epoch": 27.251528384279474,
+ "grad_norm": 0.018253348767757416,
+ "learning_rate": 0.0006,
+ "loss": 4.1720499992370605,
+ "step": 1962
+ },
+ {
+ "epoch": 27.265502183406113,
+ "grad_norm": 0.018292246386408806,
+ "learning_rate": 0.0006,
+ "loss": 4.28231143951416,
+ "step": 1963
+ },
+ {
+ "epoch": 27.27947598253275,
+ "grad_norm": 0.01943155936896801,
+ "learning_rate": 0.0006,
+ "loss": 4.314976692199707,
+ "step": 1964
+ },
+ {
+ "epoch": 27.29344978165939,
+ "grad_norm": 0.019064994528889656,
+ "learning_rate": 0.0006,
+ "loss": 4.25328254699707,
+ "step": 1965
+ },
+ {
+ "epoch": 27.307423580786025,
+ "grad_norm": 0.017423273995518684,
+ "learning_rate": 0.0006,
+ "loss": 4.195992469787598,
+ "step": 1966
+ },
+ {
+ "epoch": 27.321397379912664,
+ "grad_norm": 0.01677924580872059,
+ "learning_rate": 0.0006,
+ "loss": 4.260043144226074,
+ "step": 1967
+ },
+ {
+ "epoch": 27.335371179039303,
+ "grad_norm": 0.01720673404633999,
+ "learning_rate": 0.0006,
+ "loss": 4.222780704498291,
+ "step": 1968
+ },
+ {
+ "epoch": 27.349344978165938,
+ "grad_norm": 0.017300743609666824,
+ "learning_rate": 0.0006,
+ "loss": 4.109524250030518,
+ "step": 1969
+ },
+ {
+ "epoch": 27.363318777292577,
+ "grad_norm": 0.016065871343016624,
+ "learning_rate": 0.0006,
+ "loss": 4.36108922958374,
+ "step": 1970
+ },
+ {
+ "epoch": 27.377292576419215,
+ "grad_norm": 0.016626616939902306,
+ "learning_rate": 0.0006,
+ "loss": 4.220630645751953,
+ "step": 1971
+ },
+ {
+ "epoch": 27.39126637554585,
+ "grad_norm": 0.016482602804899216,
+ "learning_rate": 0.0006,
+ "loss": 4.344797134399414,
+ "step": 1972
+ },
+ {
+ "epoch": 27.40524017467249,
+ "grad_norm": 0.01727953553199768,
+ "learning_rate": 0.0006,
+ "loss": 4.296645164489746,
+ "step": 1973
+ },
+ {
+ "epoch": 27.419213973799128,
+ "grad_norm": 0.01522731315344572,
+ "learning_rate": 0.0006,
+ "loss": 4.223751068115234,
+ "step": 1974
+ },
+ {
+ "epoch": 27.433187772925763,
+ "grad_norm": 0.01581776700913906,
+ "learning_rate": 0.0006,
+ "loss": 4.351379871368408,
+ "step": 1975
+ },
+ {
+ "epoch": 27.4471615720524,
+ "grad_norm": 0.016394605860114098,
+ "learning_rate": 0.0006,
+ "loss": 4.2412919998168945,
+ "step": 1976
+ },
+ {
+ "epoch": 27.46113537117904,
+ "grad_norm": 0.01751169003546238,
+ "learning_rate": 0.0006,
+ "loss": 4.257606506347656,
+ "step": 1977
+ },
+ {
+ "epoch": 27.475109170305675,
+ "grad_norm": 0.017452310770750046,
+ "learning_rate": 0.0006,
+ "loss": 4.380134582519531,
+ "step": 1978
+ },
+ {
+ "epoch": 27.489082969432314,
+ "grad_norm": 0.017741898074746132,
+ "learning_rate": 0.0006,
+ "loss": 4.279632568359375,
+ "step": 1979
+ },
+ {
+ "epoch": 27.503056768558952,
+ "grad_norm": 0.017680590972304344,
+ "learning_rate": 0.0006,
+ "loss": 4.273406028747559,
+ "step": 1980
+ },
+ {
+ "epoch": 27.51703056768559,
+ "grad_norm": 0.01864718459546566,
+ "learning_rate": 0.0006,
+ "loss": 4.278583526611328,
+ "step": 1981
+ },
+ {
+ "epoch": 27.531004366812226,
+ "grad_norm": 0.019558541476726532,
+ "learning_rate": 0.0006,
+ "loss": 4.211919784545898,
+ "step": 1982
+ },
+ {
+ "epoch": 27.544978165938865,
+ "grad_norm": 0.019467400386929512,
+ "learning_rate": 0.0006,
+ "loss": 4.276822566986084,
+ "step": 1983
+ },
+ {
+ "epoch": 27.558951965065503,
+ "grad_norm": 0.01756172813475132,
+ "learning_rate": 0.0006,
+ "loss": 4.165379047393799,
+ "step": 1984
+ },
+ {
+ "epoch": 27.57292576419214,
+ "grad_norm": 0.017458263784646988,
+ "learning_rate": 0.0006,
+ "loss": 4.2992401123046875,
+ "step": 1985
+ },
+ {
+ "epoch": 27.586899563318777,
+ "grad_norm": 0.018089665099978447,
+ "learning_rate": 0.0006,
+ "loss": 4.243378639221191,
+ "step": 1986
+ },
+ {
+ "epoch": 27.600873362445416,
+ "grad_norm": 0.019535524770617485,
+ "learning_rate": 0.0006,
+ "loss": 4.321477890014648,
+ "step": 1987
+ },
+ {
+ "epoch": 27.61484716157205,
+ "grad_norm": 0.01997970975935459,
+ "learning_rate": 0.0006,
+ "loss": 4.123082637786865,
+ "step": 1988
+ },
+ {
+ "epoch": 27.62882096069869,
+ "grad_norm": 0.01913067139685154,
+ "learning_rate": 0.0006,
+ "loss": 4.289896488189697,
+ "step": 1989
+ },
+ {
+ "epoch": 27.64279475982533,
+ "grad_norm": 0.0169806070625782,
+ "learning_rate": 0.0006,
+ "loss": 4.247004985809326,
+ "step": 1990
+ },
+ {
+ "epoch": 27.656768558951963,
+ "grad_norm": 0.017315007746219635,
+ "learning_rate": 0.0006,
+ "loss": 4.2095441818237305,
+ "step": 1991
+ },
+ {
+ "epoch": 27.670742358078602,
+ "grad_norm": 0.015098759904503822,
+ "learning_rate": 0.0006,
+ "loss": 4.145560264587402,
+ "step": 1992
+ },
+ {
+ "epoch": 27.68471615720524,
+ "grad_norm": 0.016240514814853668,
+ "learning_rate": 0.0006,
+ "loss": 4.254927635192871,
+ "step": 1993
+ },
+ {
+ "epoch": 27.69868995633188,
+ "grad_norm": 0.01621108129620552,
+ "learning_rate": 0.0006,
+ "loss": 4.219725608825684,
+ "step": 1994
+ },
+ {
+ "epoch": 27.712663755458514,
+ "grad_norm": 0.015191009268164635,
+ "learning_rate": 0.0006,
+ "loss": 4.194345474243164,
+ "step": 1995
+ },
+ {
+ "epoch": 27.726637554585153,
+ "grad_norm": 0.014387983828783035,
+ "learning_rate": 0.0006,
+ "loss": 4.211467742919922,
+ "step": 1996
+ },
+ {
+ "epoch": 27.74061135371179,
+ "grad_norm": 0.015460561029613018,
+ "learning_rate": 0.0006,
+ "loss": 4.358700752258301,
+ "step": 1997
+ },
+ {
+ "epoch": 27.754585152838427,
+ "grad_norm": 0.015100076794624329,
+ "learning_rate": 0.0006,
+ "loss": 4.171223163604736,
+ "step": 1998
+ },
+ {
+ "epoch": 27.768558951965066,
+ "grad_norm": 0.015607891604304314,
+ "learning_rate": 0.0006,
+ "loss": 4.217874526977539,
+ "step": 1999
+ },
+ {
+ "epoch": 27.782532751091704,
+ "grad_norm": 0.015650689601898193,
+ "learning_rate": 0.0006,
+ "loss": 4.280780792236328,
+ "step": 2000
+ },
+ {
+ "epoch": 27.79650655021834,
+ "grad_norm": 0.01580994389951229,
+ "learning_rate": 0.0006,
+ "loss": 4.159212112426758,
+ "step": 2001
+ },
+ {
+ "epoch": 27.810480349344978,
+ "grad_norm": 0.016607413068413734,
+ "learning_rate": 0.0006,
+ "loss": 4.238853454589844,
+ "step": 2002
+ },
+ {
+ "epoch": 27.824454148471617,
+ "grad_norm": 0.017383860424160957,
+ "learning_rate": 0.0006,
+ "loss": 4.233944892883301,
+ "step": 2003
+ },
+ {
+ "epoch": 27.83842794759825,
+ "grad_norm": 0.014966659247875214,
+ "learning_rate": 0.0006,
+ "loss": 4.167392730712891,
+ "step": 2004
+ },
+ {
+ "epoch": 27.85240174672489,
+ "grad_norm": 0.014589878730475903,
+ "learning_rate": 0.0006,
+ "loss": 4.2390899658203125,
+ "step": 2005
+ },
+ {
+ "epoch": 27.86637554585153,
+ "grad_norm": 0.017179450020194054,
+ "learning_rate": 0.0006,
+ "loss": 4.365785598754883,
+ "step": 2006
+ },
+ {
+ "epoch": 27.880349344978168,
+ "grad_norm": 0.01893901452422142,
+ "learning_rate": 0.0006,
+ "loss": 4.289430141448975,
+ "step": 2007
+ },
+ {
+ "epoch": 27.894323144104803,
+ "grad_norm": 0.018619712442159653,
+ "learning_rate": 0.0006,
+ "loss": 4.310665130615234,
+ "step": 2008
+ },
+ {
+ "epoch": 27.90829694323144,
+ "grad_norm": 0.019180385395884514,
+ "learning_rate": 0.0006,
+ "loss": 4.27454137802124,
+ "step": 2009
+ },
+ {
+ "epoch": 27.92227074235808,
+ "grad_norm": 0.017811523750424385,
+ "learning_rate": 0.0006,
+ "loss": 4.312562465667725,
+ "step": 2010
+ },
+ {
+ "epoch": 27.936244541484715,
+ "grad_norm": 0.016583573073148727,
+ "learning_rate": 0.0006,
+ "loss": 4.202253341674805,
+ "step": 2011
+ },
+ {
+ "epoch": 27.950218340611354,
+ "grad_norm": 0.01658615842461586,
+ "learning_rate": 0.0006,
+ "loss": 4.183433532714844,
+ "step": 2012
+ },
+ {
+ "epoch": 27.964192139737992,
+ "grad_norm": 0.015327401459217072,
+ "learning_rate": 0.0006,
+ "loss": 4.2125139236450195,
+ "step": 2013
+ },
+ {
+ "epoch": 27.978165938864628,
+ "grad_norm": 0.015004601329565048,
+ "learning_rate": 0.0006,
+ "loss": 4.257392883300781,
+ "step": 2014
+ },
+ {
+ "epoch": 27.992139737991266,
+ "grad_norm": 0.015259161591529846,
+ "learning_rate": 0.0006,
+ "loss": 4.241818428039551,
+ "step": 2015
+ },
+ {
+ "epoch": 28.0,
+ "grad_norm": 0.016808776184916496,
+ "learning_rate": 0.0006,
+ "loss": 4.146241188049316,
+ "step": 2016
+ },
+ {
+ "epoch": 28.0,
+ "eval_loss": 4.663090705871582,
+ "eval_runtime": 56.4559,
+ "eval_samples_per_second": 43.255,
+ "eval_steps_per_second": 1.364,
+ "step": 2016
+ },
+ {
+ "epoch": 28.01397379912664,
+ "grad_norm": 0.015784382820129395,
+ "learning_rate": 0.0006,
+ "loss": 4.044191360473633,
+ "step": 2017
+ },
+ {
+ "epoch": 28.027947598253274,
+ "grad_norm": 0.015636710450053215,
+ "learning_rate": 0.0006,
+ "loss": 4.128754615783691,
+ "step": 2018
+ },
+ {
+ "epoch": 28.041921397379912,
+ "grad_norm": 0.015364975668489933,
+ "learning_rate": 0.0006,
+ "loss": 4.191835403442383,
+ "step": 2019
+ },
+ {
+ "epoch": 28.05589519650655,
+ "grad_norm": 0.015015700832009315,
+ "learning_rate": 0.0006,
+ "loss": 4.120022773742676,
+ "step": 2020
+ },
+ {
+ "epoch": 28.069868995633186,
+ "grad_norm": 0.016666799783706665,
+ "learning_rate": 0.0006,
+ "loss": 4.33906364440918,
+ "step": 2021
+ },
+ {
+ "epoch": 28.083842794759825,
+ "grad_norm": 0.01780104823410511,
+ "learning_rate": 0.0006,
+ "loss": 4.279458999633789,
+ "step": 2022
+ },
+ {
+ "epoch": 28.097816593886463,
+ "grad_norm": 0.01870882883667946,
+ "learning_rate": 0.0006,
+ "loss": 4.34728479385376,
+ "step": 2023
+ },
+ {
+ "epoch": 28.111790393013102,
+ "grad_norm": 0.018406381830573082,
+ "learning_rate": 0.0006,
+ "loss": 4.211311340332031,
+ "step": 2024
+ },
+ {
+ "epoch": 28.125764192139737,
+ "grad_norm": 0.017833448946475983,
+ "learning_rate": 0.0006,
+ "loss": 4.223254203796387,
+ "step": 2025
+ },
+ {
+ "epoch": 28.139737991266376,
+ "grad_norm": 0.017552688717842102,
+ "learning_rate": 0.0006,
+ "loss": 4.208823204040527,
+ "step": 2026
+ },
+ {
+ "epoch": 28.153711790393015,
+ "grad_norm": 0.018500229343771935,
+ "learning_rate": 0.0006,
+ "loss": 4.198025226593018,
+ "step": 2027
+ },
+ {
+ "epoch": 28.16768558951965,
+ "grad_norm": 0.01782156340777874,
+ "learning_rate": 0.0006,
+ "loss": 4.179978370666504,
+ "step": 2028
+ },
+ {
+ "epoch": 28.18165938864629,
+ "grad_norm": 0.019580967724323273,
+ "learning_rate": 0.0006,
+ "loss": 4.291114807128906,
+ "step": 2029
+ },
+ {
+ "epoch": 28.195633187772927,
+ "grad_norm": 0.020804036408662796,
+ "learning_rate": 0.0006,
+ "loss": 4.220607280731201,
+ "step": 2030
+ },
+ {
+ "epoch": 28.209606986899562,
+ "grad_norm": 0.019595693796873093,
+ "learning_rate": 0.0006,
+ "loss": 4.2905731201171875,
+ "step": 2031
+ },
+ {
+ "epoch": 28.2235807860262,
+ "grad_norm": 0.019247086718678474,
+ "learning_rate": 0.0006,
+ "loss": 4.2467360496521,
+ "step": 2032
+ },
+ {
+ "epoch": 28.23755458515284,
+ "grad_norm": 0.01958036608994007,
+ "learning_rate": 0.0006,
+ "loss": 4.286137580871582,
+ "step": 2033
+ },
+ {
+ "epoch": 28.251528384279474,
+ "grad_norm": 0.021354753524065018,
+ "learning_rate": 0.0006,
+ "loss": 4.234755039215088,
+ "step": 2034
+ },
+ {
+ "epoch": 28.265502183406113,
+ "grad_norm": 0.02206375077366829,
+ "learning_rate": 0.0006,
+ "loss": 4.149641036987305,
+ "step": 2035
+ },
+ {
+ "epoch": 28.27947598253275,
+ "grad_norm": 0.02201310358941555,
+ "learning_rate": 0.0006,
+ "loss": 4.1928629875183105,
+ "step": 2036
+ },
+ {
+ "epoch": 28.29344978165939,
+ "grad_norm": 0.01961950585246086,
+ "learning_rate": 0.0006,
+ "loss": 4.211404323577881,
+ "step": 2037
+ },
+ {
+ "epoch": 28.307423580786025,
+ "grad_norm": 0.019211795181035995,
+ "learning_rate": 0.0006,
+ "loss": 4.200976848602295,
+ "step": 2038
+ },
+ {
+ "epoch": 28.321397379912664,
+ "grad_norm": 0.01779370754957199,
+ "learning_rate": 0.0006,
+ "loss": 4.249148845672607,
+ "step": 2039
+ },
+ {
+ "epoch": 28.335371179039303,
+ "grad_norm": 0.018635908141732216,
+ "learning_rate": 0.0006,
+ "loss": 4.11806058883667,
+ "step": 2040
+ },
+ {
+ "epoch": 28.349344978165938,
+ "grad_norm": 0.019792694598436356,
+ "learning_rate": 0.0006,
+ "loss": 4.335224151611328,
+ "step": 2041
+ },
+ {
+ "epoch": 28.363318777292577,
+ "grad_norm": 0.021731717512011528,
+ "learning_rate": 0.0006,
+ "loss": 4.302778244018555,
+ "step": 2042
+ },
+ {
+ "epoch": 28.377292576419215,
+ "grad_norm": 0.020823447033762932,
+ "learning_rate": 0.0006,
+ "loss": 4.260610580444336,
+ "step": 2043
+ },
+ {
+ "epoch": 28.39126637554585,
+ "grad_norm": 0.019385412335395813,
+ "learning_rate": 0.0006,
+ "loss": 4.2247724533081055,
+ "step": 2044
+ },
+ {
+ "epoch": 28.40524017467249,
+ "grad_norm": 0.019539108499884605,
+ "learning_rate": 0.0006,
+ "loss": 4.236655235290527,
+ "step": 2045
+ },
+ {
+ "epoch": 28.419213973799128,
+ "grad_norm": 0.01942690648138523,
+ "learning_rate": 0.0006,
+ "loss": 4.251850605010986,
+ "step": 2046
+ },
+ {
+ "epoch": 28.433187772925763,
+ "grad_norm": 0.019667886197566986,
+ "learning_rate": 0.0006,
+ "loss": 4.222312927246094,
+ "step": 2047
+ },
+ {
+ "epoch": 28.4471615720524,
+ "grad_norm": 0.02029012329876423,
+ "learning_rate": 0.0006,
+ "loss": 4.246252059936523,
+ "step": 2048
+ },
+ {
+ "epoch": 28.46113537117904,
+ "grad_norm": 0.01784469373524189,
+ "learning_rate": 0.0006,
+ "loss": 4.192128658294678,
+ "step": 2049
+ },
+ {
+ "epoch": 28.475109170305675,
+ "grad_norm": 0.0160287544131279,
+ "learning_rate": 0.0006,
+ "loss": 4.208607196807861,
+ "step": 2050
+ },
+ {
+ "epoch": 28.489082969432314,
+ "grad_norm": 0.015348346903920174,
+ "learning_rate": 0.0006,
+ "loss": 4.236598014831543,
+ "step": 2051
+ },
+ {
+ "epoch": 28.503056768558952,
+ "grad_norm": 0.015198206529021263,
+ "learning_rate": 0.0006,
+ "loss": 4.235552787780762,
+ "step": 2052
+ },
+ {
+ "epoch": 28.51703056768559,
+ "grad_norm": 0.01523754745721817,
+ "learning_rate": 0.0006,
+ "loss": 4.1941728591918945,
+ "step": 2053
+ },
+ {
+ "epoch": 28.531004366812226,
+ "grad_norm": 0.015559614636003971,
+ "learning_rate": 0.0006,
+ "loss": 4.294757843017578,
+ "step": 2054
+ },
+ {
+ "epoch": 28.544978165938865,
+ "grad_norm": 0.01568358950316906,
+ "learning_rate": 0.0006,
+ "loss": 4.213065147399902,
+ "step": 2055
+ },
+ {
+ "epoch": 28.558951965065503,
+ "grad_norm": 0.015336516313254833,
+ "learning_rate": 0.0006,
+ "loss": 4.240293979644775,
+ "step": 2056
+ },
+ {
+ "epoch": 28.57292576419214,
+ "grad_norm": 0.015041496604681015,
+ "learning_rate": 0.0006,
+ "loss": 4.273397445678711,
+ "step": 2057
+ },
+ {
+ "epoch": 28.586899563318777,
+ "grad_norm": 0.01610128954052925,
+ "learning_rate": 0.0006,
+ "loss": 4.125369071960449,
+ "step": 2058
+ },
+ {
+ "epoch": 28.600873362445416,
+ "grad_norm": 0.015974976122379303,
+ "learning_rate": 0.0006,
+ "loss": 4.133459091186523,
+ "step": 2059
+ },
+ {
+ "epoch": 28.61484716157205,
+ "grad_norm": 0.015936799347400665,
+ "learning_rate": 0.0006,
+ "loss": 4.306667327880859,
+ "step": 2060
+ },
+ {
+ "epoch": 28.62882096069869,
+ "grad_norm": 0.01795247197151184,
+ "learning_rate": 0.0006,
+ "loss": 4.205156326293945,
+ "step": 2061
+ },
+ {
+ "epoch": 28.64279475982533,
+ "grad_norm": 0.019639814272522926,
+ "learning_rate": 0.0006,
+ "loss": 4.277614593505859,
+ "step": 2062
+ },
+ {
+ "epoch": 28.656768558951963,
+ "grad_norm": 0.01722300611436367,
+ "learning_rate": 0.0006,
+ "loss": 4.191032886505127,
+ "step": 2063
+ },
+ {
+ "epoch": 28.670742358078602,
+ "grad_norm": 0.01612633652985096,
+ "learning_rate": 0.0006,
+ "loss": 4.20717191696167,
+ "step": 2064
+ },
+ {
+ "epoch": 28.68471615720524,
+ "grad_norm": 0.01603279449045658,
+ "learning_rate": 0.0006,
+ "loss": 4.228884220123291,
+ "step": 2065
+ },
+ {
+ "epoch": 28.69868995633188,
+ "grad_norm": 0.016405927017331123,
+ "learning_rate": 0.0006,
+ "loss": 4.201292991638184,
+ "step": 2066
+ },
+ {
+ "epoch": 28.712663755458514,
+ "grad_norm": 0.016644183546304703,
+ "learning_rate": 0.0006,
+ "loss": 4.239282131195068,
+ "step": 2067
+ },
+ {
+ "epoch": 28.726637554585153,
+ "grad_norm": 0.0175587497651577,
+ "learning_rate": 0.0006,
+ "loss": 4.289038181304932,
+ "step": 2068
+ },
+ {
+ "epoch": 28.74061135371179,
+ "grad_norm": 0.018223397433757782,
+ "learning_rate": 0.0006,
+ "loss": 4.2266998291015625,
+ "step": 2069
+ },
+ {
+ "epoch": 28.754585152838427,
+ "grad_norm": 0.018680868670344353,
+ "learning_rate": 0.0006,
+ "loss": 4.348849296569824,
+ "step": 2070
+ },
+ {
+ "epoch": 28.768558951965066,
+ "grad_norm": 0.016726728528738022,
+ "learning_rate": 0.0006,
+ "loss": 4.292080879211426,
+ "step": 2071
+ },
+ {
+ "epoch": 28.782532751091704,
+ "grad_norm": 0.017386795952916145,
+ "learning_rate": 0.0006,
+ "loss": 4.208870887756348,
+ "step": 2072
+ },
+ {
+ "epoch": 28.79650655021834,
+ "grad_norm": 0.01865178905427456,
+ "learning_rate": 0.0006,
+ "loss": 4.054150581359863,
+ "step": 2073
+ },
+ {
+ "epoch": 28.810480349344978,
+ "grad_norm": 0.01776815392076969,
+ "learning_rate": 0.0006,
+ "loss": 4.202404499053955,
+ "step": 2074
+ },
+ {
+ "epoch": 28.824454148471617,
+ "grad_norm": 0.01643497310578823,
+ "learning_rate": 0.0006,
+ "loss": 4.131855010986328,
+ "step": 2075
+ },
+ {
+ "epoch": 28.83842794759825,
+ "grad_norm": 0.01574273779988289,
+ "learning_rate": 0.0006,
+ "loss": 4.296788692474365,
+ "step": 2076
+ },
+ {
+ "epoch": 28.85240174672489,
+ "grad_norm": 0.015613908879458904,
+ "learning_rate": 0.0006,
+ "loss": 4.210949897766113,
+ "step": 2077
+ },
+ {
+ "epoch": 28.86637554585153,
+ "grad_norm": 0.015543212182819843,
+ "learning_rate": 0.0006,
+ "loss": 4.218136310577393,
+ "step": 2078
+ },
+ {
+ "epoch": 28.880349344978168,
+ "grad_norm": 0.014681251719594002,
+ "learning_rate": 0.0006,
+ "loss": 4.19401741027832,
+ "step": 2079
+ },
+ {
+ "epoch": 28.894323144104803,
+ "grad_norm": 0.014393828809261322,
+ "learning_rate": 0.0006,
+ "loss": 4.198468208312988,
+ "step": 2080
+ },
+ {
+ "epoch": 28.90829694323144,
+ "grad_norm": 0.01580170728266239,
+ "learning_rate": 0.0006,
+ "loss": 4.224969863891602,
+ "step": 2081
+ },
+ {
+ "epoch": 28.92227074235808,
+ "grad_norm": 0.01484165620058775,
+ "learning_rate": 0.0006,
+ "loss": 4.12236213684082,
+ "step": 2082
+ },
+ {
+ "epoch": 28.936244541484715,
+ "grad_norm": 0.01497623696923256,
+ "learning_rate": 0.0006,
+ "loss": 4.24141788482666,
+ "step": 2083
+ },
+ {
+ "epoch": 28.950218340611354,
+ "grad_norm": 0.015080046840012074,
+ "learning_rate": 0.0006,
+ "loss": 4.284151077270508,
+ "step": 2084
+ },
+ {
+ "epoch": 28.964192139737992,
+ "grad_norm": 0.016207212582230568,
+ "learning_rate": 0.0006,
+ "loss": 4.298693656921387,
+ "step": 2085
+ },
+ {
+ "epoch": 28.978165938864628,
+ "grad_norm": 0.015474417246878147,
+ "learning_rate": 0.0006,
+ "loss": 4.2412261962890625,
+ "step": 2086
+ },
+ {
+ "epoch": 28.992139737991266,
+ "grad_norm": 0.01515351701527834,
+ "learning_rate": 0.0006,
+ "loss": 4.3054423332214355,
+ "step": 2087
+ },
+ {
+ "epoch": 29.0,
+ "grad_norm": 0.016411345452070236,
+ "learning_rate": 0.0006,
+ "loss": 4.104830741882324,
+ "step": 2088
+ },
+ {
+ "epoch": 29.0,
+ "eval_loss": 4.661567687988281,
+ "eval_runtime": 57.1244,
+ "eval_samples_per_second": 42.749,
+ "eval_steps_per_second": 1.348,
+ "step": 2088
+ },
+ {
+ "epoch": 29.01397379912664,
+ "grad_norm": 0.01665342226624489,
+ "learning_rate": 0.0006,
+ "loss": 4.311391830444336,
+ "step": 2089
+ },
+ {
+ "epoch": 29.027947598253274,
+ "grad_norm": 0.018053214997053146,
+ "learning_rate": 0.0006,
+ "loss": 4.135597229003906,
+ "step": 2090
+ },
+ {
+ "epoch": 29.041921397379912,
+ "grad_norm": 0.018213748931884766,
+ "learning_rate": 0.0006,
+ "loss": 4.303299903869629,
+ "step": 2091
+ },
+ {
+ "epoch": 29.05589519650655,
+ "grad_norm": 0.018460242077708244,
+ "learning_rate": 0.0006,
+ "loss": 4.20850944519043,
+ "step": 2092
+ },
+ {
+ "epoch": 29.069868995633186,
+ "grad_norm": 0.018365809693932533,
+ "learning_rate": 0.0006,
+ "loss": 4.275847434997559,
+ "step": 2093
+ },
+ {
+ "epoch": 29.083842794759825,
+ "grad_norm": 0.018346186727285385,
+ "learning_rate": 0.0006,
+ "loss": 4.202091217041016,
+ "step": 2094
+ },
+ {
+ "epoch": 29.097816593886463,
+ "grad_norm": 0.017920587211847305,
+ "learning_rate": 0.0006,
+ "loss": 4.182092666625977,
+ "step": 2095
+ },
+ {
+ "epoch": 29.111790393013102,
+ "grad_norm": 0.01812003180384636,
+ "learning_rate": 0.0006,
+ "loss": 4.259721755981445,
+ "step": 2096
+ },
+ {
+ "epoch": 29.125764192139737,
+ "grad_norm": 0.01700986735522747,
+ "learning_rate": 0.0006,
+ "loss": 4.067573070526123,
+ "step": 2097
+ },
+ {
+ "epoch": 29.139737991266376,
+ "grad_norm": 0.018069760873913765,
+ "learning_rate": 0.0006,
+ "loss": 4.234400749206543,
+ "step": 2098
+ },
+ {
+ "epoch": 29.153711790393015,
+ "grad_norm": 0.01905428245663643,
+ "learning_rate": 0.0006,
+ "loss": 4.16609001159668,
+ "step": 2099
+ },
+ {
+ "epoch": 29.16768558951965,
+ "grad_norm": 0.018711242824792862,
+ "learning_rate": 0.0006,
+ "loss": 4.249088764190674,
+ "step": 2100
+ },
+ {
+ "epoch": 29.18165938864629,
+ "grad_norm": 0.018721306696534157,
+ "learning_rate": 0.0006,
+ "loss": 4.14540958404541,
+ "step": 2101
+ },
+ {
+ "epoch": 29.195633187772927,
+ "grad_norm": 0.020139874890446663,
+ "learning_rate": 0.0006,
+ "loss": 4.049248695373535,
+ "step": 2102
+ },
+ {
+ "epoch": 29.209606986899562,
+ "grad_norm": 0.020849574357271194,
+ "learning_rate": 0.0006,
+ "loss": 4.235224723815918,
+ "step": 2103
+ },
+ {
+ "epoch": 29.2235807860262,
+ "grad_norm": 0.018693549558520317,
+ "learning_rate": 0.0006,
+ "loss": 4.232961177825928,
+ "step": 2104
+ },
+ {
+ "epoch": 29.23755458515284,
+ "grad_norm": 0.017889857292175293,
+ "learning_rate": 0.0006,
+ "loss": 4.171422004699707,
+ "step": 2105
+ },
+ {
+ "epoch": 29.251528384279474,
+ "grad_norm": 0.016690224409103394,
+ "learning_rate": 0.0006,
+ "loss": 4.244717597961426,
+ "step": 2106
+ },
+ {
+ "epoch": 29.265502183406113,
+ "grad_norm": 0.01708616502583027,
+ "learning_rate": 0.0006,
+ "loss": 4.125068664550781,
+ "step": 2107
+ },
+ {
+ "epoch": 29.27947598253275,
+ "grad_norm": 0.017928237095475197,
+ "learning_rate": 0.0006,
+ "loss": 4.208292007446289,
+ "step": 2108
+ },
+ {
+ "epoch": 29.29344978165939,
+ "grad_norm": 0.018250394612550735,
+ "learning_rate": 0.0006,
+ "loss": 4.118630886077881,
+ "step": 2109
+ },
+ {
+ "epoch": 29.307423580786025,
+ "grad_norm": 0.020160246640443802,
+ "learning_rate": 0.0006,
+ "loss": 4.206025123596191,
+ "step": 2110
+ },
+ {
+ "epoch": 29.321397379912664,
+ "grad_norm": 0.020045453682541847,
+ "learning_rate": 0.0006,
+ "loss": 4.2758684158325195,
+ "step": 2111
+ },
+ {
+ "epoch": 29.335371179039303,
+ "grad_norm": 0.019556907936930656,
+ "learning_rate": 0.0006,
+ "loss": 4.188272953033447,
+ "step": 2112
+ },
+ {
+ "epoch": 29.349344978165938,
+ "grad_norm": 0.019185200333595276,
+ "learning_rate": 0.0006,
+ "loss": 4.17451286315918,
+ "step": 2113
+ },
+ {
+ "epoch": 29.363318777292577,
+ "grad_norm": 0.01816629432141781,
+ "learning_rate": 0.0006,
+ "loss": 4.165246963500977,
+ "step": 2114
+ },
+ {
+ "epoch": 29.377292576419215,
+ "grad_norm": 0.01865146867930889,
+ "learning_rate": 0.0006,
+ "loss": 4.148184776306152,
+ "step": 2115
+ },
+ {
+ "epoch": 29.39126637554585,
+ "grad_norm": 0.018338464200496674,
+ "learning_rate": 0.0006,
+ "loss": 4.288936614990234,
+ "step": 2116
+ },
+ {
+ "epoch": 29.40524017467249,
+ "grad_norm": 0.018986187875270844,
+ "learning_rate": 0.0006,
+ "loss": 4.183379650115967,
+ "step": 2117
+ },
+ {
+ "epoch": 29.419213973799128,
+ "grad_norm": 0.018409637734293938,
+ "learning_rate": 0.0006,
+ "loss": 4.057574272155762,
+ "step": 2118
+ },
+ {
+ "epoch": 29.433187772925763,
+ "grad_norm": 0.016664542257785797,
+ "learning_rate": 0.0006,
+ "loss": 4.18517541885376,
+ "step": 2119
+ },
+ {
+ "epoch": 29.4471615720524,
+ "grad_norm": 0.016486089676618576,
+ "learning_rate": 0.0006,
+ "loss": 4.179367542266846,
+ "step": 2120
+ },
+ {
+ "epoch": 29.46113537117904,
+ "grad_norm": 0.01787523552775383,
+ "learning_rate": 0.0006,
+ "loss": 4.184126377105713,
+ "step": 2121
+ },
+ {
+ "epoch": 29.475109170305675,
+ "grad_norm": 0.016574613749980927,
+ "learning_rate": 0.0006,
+ "loss": 4.290170192718506,
+ "step": 2122
+ },
+ {
+ "epoch": 29.489082969432314,
+ "grad_norm": 0.01643376611173153,
+ "learning_rate": 0.0006,
+ "loss": 4.052913665771484,
+ "step": 2123
+ },
+ {
+ "epoch": 29.503056768558952,
+ "grad_norm": 0.016503969207406044,
+ "learning_rate": 0.0006,
+ "loss": 4.335184097290039,
+ "step": 2124
+ },
+ {
+ "epoch": 29.51703056768559,
+ "grad_norm": 0.017596479505300522,
+ "learning_rate": 0.0006,
+ "loss": 4.165539741516113,
+ "step": 2125
+ },
+ {
+ "epoch": 29.531004366812226,
+ "grad_norm": 0.017468582838773727,
+ "learning_rate": 0.0006,
+ "loss": 4.315964698791504,
+ "step": 2126
+ },
+ {
+ "epoch": 29.544978165938865,
+ "grad_norm": 0.01724618673324585,
+ "learning_rate": 0.0006,
+ "loss": 4.243681907653809,
+ "step": 2127
+ },
+ {
+ "epoch": 29.558951965065503,
+ "grad_norm": 0.018384616822004318,
+ "learning_rate": 0.0006,
+ "loss": 4.112447261810303,
+ "step": 2128
+ },
+ {
+ "epoch": 29.57292576419214,
+ "grad_norm": 0.01824074238538742,
+ "learning_rate": 0.0006,
+ "loss": 4.236065864562988,
+ "step": 2129
+ },
+ {
+ "epoch": 29.586899563318777,
+ "grad_norm": 0.01625971868634224,
+ "learning_rate": 0.0006,
+ "loss": 4.18143892288208,
+ "step": 2130
+ },
+ {
+ "epoch": 29.600873362445416,
+ "grad_norm": 0.015537494793534279,
+ "learning_rate": 0.0006,
+ "loss": 4.225147247314453,
+ "step": 2131
+ },
+ {
+ "epoch": 29.61484716157205,
+ "grad_norm": 0.015818974003195763,
+ "learning_rate": 0.0006,
+ "loss": 4.3100175857543945,
+ "step": 2132
+ },
+ {
+ "epoch": 29.62882096069869,
+ "grad_norm": 0.015909671783447266,
+ "learning_rate": 0.0006,
+ "loss": 4.169775009155273,
+ "step": 2133
+ },
+ {
+ "epoch": 29.64279475982533,
+ "grad_norm": 0.01769658550620079,
+ "learning_rate": 0.0006,
+ "loss": 4.278904914855957,
+ "step": 2134
+ },
+ {
+ "epoch": 29.656768558951963,
+ "grad_norm": 0.015641704201698303,
+ "learning_rate": 0.0006,
+ "loss": 4.2426228523254395,
+ "step": 2135
+ },
+ {
+ "epoch": 29.670742358078602,
+ "grad_norm": 0.01565658301115036,
+ "learning_rate": 0.0006,
+ "loss": 4.232687950134277,
+ "step": 2136
+ },
+ {
+ "epoch": 29.68471615720524,
+ "grad_norm": 0.016485940665006638,
+ "learning_rate": 0.0006,
+ "loss": 4.209839820861816,
+ "step": 2137
+ },
+ {
+ "epoch": 29.69868995633188,
+ "grad_norm": 0.017622729763388634,
+ "learning_rate": 0.0006,
+ "loss": 4.196017265319824,
+ "step": 2138
+ },
+ {
+ "epoch": 29.712663755458514,
+ "grad_norm": 0.018320564180612564,
+ "learning_rate": 0.0006,
+ "loss": 4.225683689117432,
+ "step": 2139
+ },
+ {
+ "epoch": 29.726637554585153,
+ "grad_norm": 0.018140794709324837,
+ "learning_rate": 0.0006,
+ "loss": 4.311944961547852,
+ "step": 2140
+ },
+ {
+ "epoch": 29.74061135371179,
+ "grad_norm": 0.016227100044488907,
+ "learning_rate": 0.0006,
+ "loss": 4.230257511138916,
+ "step": 2141
+ },
+ {
+ "epoch": 29.754585152838427,
+ "grad_norm": 0.015759488567709923,
+ "learning_rate": 0.0006,
+ "loss": 4.119174003601074,
+ "step": 2142
+ },
+ {
+ "epoch": 29.768558951965066,
+ "grad_norm": 0.01757766678929329,
+ "learning_rate": 0.0006,
+ "loss": 4.165826797485352,
+ "step": 2143
+ },
+ {
+ "epoch": 29.782532751091704,
+ "grad_norm": 0.019661923870444298,
+ "learning_rate": 0.0006,
+ "loss": 4.392203330993652,
+ "step": 2144
+ },
+ {
+ "epoch": 29.79650655021834,
+ "grad_norm": 0.019478755071759224,
+ "learning_rate": 0.0006,
+ "loss": 4.348371982574463,
+ "step": 2145
+ },
+ {
+ "epoch": 29.810480349344978,
+ "grad_norm": 0.019149569794535637,
+ "learning_rate": 0.0006,
+ "loss": 4.34104061126709,
+ "step": 2146
+ },
+ {
+ "epoch": 29.824454148471617,
+ "grad_norm": 0.017707521095871925,
+ "learning_rate": 0.0006,
+ "loss": 4.228226661682129,
+ "step": 2147
+ },
+ {
+ "epoch": 29.83842794759825,
+ "grad_norm": 0.015615378506481647,
+ "learning_rate": 0.0006,
+ "loss": 4.175541877746582,
+ "step": 2148
+ },
+ {
+ "epoch": 29.85240174672489,
+ "grad_norm": 0.016111129894852638,
+ "learning_rate": 0.0006,
+ "loss": 4.267936706542969,
+ "step": 2149
+ },
+ {
+ "epoch": 29.86637554585153,
+ "grad_norm": 0.01528779324144125,
+ "learning_rate": 0.0006,
+ "loss": 4.186014175415039,
+ "step": 2150
+ },
+ {
+ "epoch": 29.880349344978168,
+ "grad_norm": 0.014925646595656872,
+ "learning_rate": 0.0006,
+ "loss": 4.174860000610352,
+ "step": 2151
+ },
+ {
+ "epoch": 29.894323144104803,
+ "grad_norm": 0.016917673870921135,
+ "learning_rate": 0.0006,
+ "loss": 4.33908748626709,
+ "step": 2152
+ },
+ {
+ "epoch": 29.90829694323144,
+ "grad_norm": 0.016007075086236,
+ "learning_rate": 0.0006,
+ "loss": 4.266141414642334,
+ "step": 2153
+ },
+ {
+ "epoch": 29.92227074235808,
+ "grad_norm": 0.016178956255316734,
+ "learning_rate": 0.0006,
+ "loss": 4.271215915679932,
+ "step": 2154
+ },
+ {
+ "epoch": 29.936244541484715,
+ "grad_norm": 0.016180871054530144,
+ "learning_rate": 0.0006,
+ "loss": 4.278801441192627,
+ "step": 2155
+ },
+ {
+ "epoch": 29.950218340611354,
+ "grad_norm": 0.015335503034293652,
+ "learning_rate": 0.0006,
+ "loss": 4.194978713989258,
+ "step": 2156
+ },
+ {
+ "epoch": 29.964192139737992,
+ "grad_norm": 0.016580764204263687,
+ "learning_rate": 0.0006,
+ "loss": 4.290809631347656,
+ "step": 2157
+ },
+ {
+ "epoch": 29.978165938864628,
+ "grad_norm": 0.01623128168284893,
+ "learning_rate": 0.0006,
+ "loss": 4.205451488494873,
+ "step": 2158
+ },
+ {
+ "epoch": 29.992139737991266,
+ "grad_norm": 0.014904248528182507,
+ "learning_rate": 0.0006,
+ "loss": 4.24924373626709,
+ "step": 2159
+ },
+ {
+ "epoch": 30.0,
+ "grad_norm": 0.017089299857616425,
+ "learning_rate": 0.0006,
+ "loss": 4.222072601318359,
+ "step": 2160
+ },
+ {
+ "epoch": 30.0,
+ "eval_loss": 4.585411071777344,
+ "eval_runtime": 57.1164,
+ "eval_samples_per_second": 42.755,
+ "eval_steps_per_second": 1.348,
+ "step": 2160
+ },
+ {
+ "epoch": 30.01397379912664,
+ "grad_norm": 0.017226383090019226,
+ "learning_rate": 0.0006,
+ "loss": 4.19467830657959,
+ "step": 2161
+ },
+ {
+ "epoch": 30.027947598253274,
+ "grad_norm": 0.01635066606104374,
+ "learning_rate": 0.0006,
+ "loss": 4.176875591278076,
+ "step": 2162
+ },
+ {
+ "epoch": 30.041921397379912,
+ "grad_norm": 0.01735362410545349,
+ "learning_rate": 0.0006,
+ "loss": 4.13667106628418,
+ "step": 2163
+ },
+ {
+ "epoch": 30.05589519650655,
+ "grad_norm": 0.017047109082341194,
+ "learning_rate": 0.0006,
+ "loss": 4.136329650878906,
+ "step": 2164
+ },
+ {
+ "epoch": 30.069868995633186,
+ "grad_norm": 0.018116053193807602,
+ "learning_rate": 0.0006,
+ "loss": 4.228706359863281,
+ "step": 2165
+ },
+ {
+ "epoch": 30.083842794759825,
+ "grad_norm": 0.01791389286518097,
+ "learning_rate": 0.0006,
+ "loss": 4.1597747802734375,
+ "step": 2166
+ },
+ {
+ "epoch": 30.097816593886463,
+ "grad_norm": 0.017793502658605576,
+ "learning_rate": 0.0006,
+ "loss": 4.0807976722717285,
+ "step": 2167
+ },
+ {
+ "epoch": 30.111790393013102,
+ "grad_norm": 0.01827503927052021,
+ "learning_rate": 0.0006,
+ "loss": 4.170646667480469,
+ "step": 2168
+ },
+ {
+ "epoch": 30.125764192139737,
+ "grad_norm": 0.017228564247488976,
+ "learning_rate": 0.0006,
+ "loss": 4.146934509277344,
+ "step": 2169
+ },
+ {
+ "epoch": 30.139737991266376,
+ "grad_norm": 0.01757437363266945,
+ "learning_rate": 0.0006,
+ "loss": 4.1859130859375,
+ "step": 2170
+ },
+ {
+ "epoch": 30.153711790393015,
+ "grad_norm": 0.01701093092560768,
+ "learning_rate": 0.0006,
+ "loss": 4.126981258392334,
+ "step": 2171
+ },
+ {
+ "epoch": 30.16768558951965,
+ "grad_norm": 0.015220489352941513,
+ "learning_rate": 0.0006,
+ "loss": 4.164018630981445,
+ "step": 2172
+ },
+ {
+ "epoch": 30.18165938864629,
+ "grad_norm": 0.016254238784313202,
+ "learning_rate": 0.0006,
+ "loss": 4.258275508880615,
+ "step": 2173
+ },
+ {
+ "epoch": 30.195633187772927,
+ "grad_norm": 0.016609614714980125,
+ "learning_rate": 0.0006,
+ "loss": 4.222233772277832,
+ "step": 2174
+ },
+ {
+ "epoch": 30.209606986899562,
+ "grad_norm": 0.016531318426132202,
+ "learning_rate": 0.0006,
+ "loss": 4.144489288330078,
+ "step": 2175
+ },
+ {
+ "epoch": 30.2235807860262,
+ "grad_norm": 0.01567668654024601,
+ "learning_rate": 0.0006,
+ "loss": 4.181916236877441,
+ "step": 2176
+ },
+ {
+ "epoch": 30.23755458515284,
+ "grad_norm": 0.017195681110024452,
+ "learning_rate": 0.0006,
+ "loss": 4.189693927764893,
+ "step": 2177
+ },
+ {
+ "epoch": 30.251528384279474,
+ "grad_norm": 0.016882948577404022,
+ "learning_rate": 0.0006,
+ "loss": 4.21604061126709,
+ "step": 2178
+ },
+ {
+ "epoch": 30.265502183406113,
+ "grad_norm": 0.016261622309684753,
+ "learning_rate": 0.0006,
+ "loss": 4.227639198303223,
+ "step": 2179
+ },
+ {
+ "epoch": 30.27947598253275,
+ "grad_norm": 0.016411812976002693,
+ "learning_rate": 0.0006,
+ "loss": 4.130362510681152,
+ "step": 2180
+ },
+ {
+ "epoch": 30.29344978165939,
+ "grad_norm": 0.01732019893825054,
+ "learning_rate": 0.0006,
+ "loss": 4.230443000793457,
+ "step": 2181
+ },
+ {
+ "epoch": 30.307423580786025,
+ "grad_norm": 0.01825905591249466,
+ "learning_rate": 0.0006,
+ "loss": 4.175817489624023,
+ "step": 2182
+ },
+ {
+ "epoch": 30.321397379912664,
+ "grad_norm": 0.01887713558971882,
+ "learning_rate": 0.0006,
+ "loss": 4.169020652770996,
+ "step": 2183
+ },
+ {
+ "epoch": 30.335371179039303,
+ "grad_norm": 0.019391661509871483,
+ "learning_rate": 0.0006,
+ "loss": 4.176318645477295,
+ "step": 2184
+ },
+ {
+ "epoch": 30.349344978165938,
+ "grad_norm": 0.020840002223849297,
+ "learning_rate": 0.0006,
+ "loss": 4.174228668212891,
+ "step": 2185
+ },
+ {
+ "epoch": 30.363318777292577,
+ "grad_norm": 0.020193178206682205,
+ "learning_rate": 0.0006,
+ "loss": 4.127919673919678,
+ "step": 2186
+ },
+ {
+ "epoch": 30.377292576419215,
+ "grad_norm": 0.018449561670422554,
+ "learning_rate": 0.0006,
+ "loss": 4.229487895965576,
+ "step": 2187
+ },
+ {
+ "epoch": 30.39126637554585,
+ "grad_norm": 0.01711983233690262,
+ "learning_rate": 0.0006,
+ "loss": 4.201651573181152,
+ "step": 2188
+ },
+ {
+ "epoch": 30.40524017467249,
+ "grad_norm": 0.017293395474553108,
+ "learning_rate": 0.0006,
+ "loss": 4.168704032897949,
+ "step": 2189
+ },
+ {
+ "epoch": 30.419213973799128,
+ "grad_norm": 0.018526358529925346,
+ "learning_rate": 0.0006,
+ "loss": 4.171864032745361,
+ "step": 2190
+ },
+ {
+ "epoch": 30.433187772925763,
+ "grad_norm": 0.019287291914224625,
+ "learning_rate": 0.0006,
+ "loss": 4.039534568786621,
+ "step": 2191
+ },
+ {
+ "epoch": 30.4471615720524,
+ "grad_norm": 0.016902444884181023,
+ "learning_rate": 0.0006,
+ "loss": 4.282787799835205,
+ "step": 2192
+ },
+ {
+ "epoch": 30.46113537117904,
+ "grad_norm": 0.01656750962138176,
+ "learning_rate": 0.0006,
+ "loss": 4.213098526000977,
+ "step": 2193
+ },
+ {
+ "epoch": 30.475109170305675,
+ "grad_norm": 0.016805065795779228,
+ "learning_rate": 0.0006,
+ "loss": 4.1583476066589355,
+ "step": 2194
+ },
+ {
+ "epoch": 30.489082969432314,
+ "grad_norm": 0.017166638746857643,
+ "learning_rate": 0.0006,
+ "loss": 4.188634395599365,
+ "step": 2195
+ },
+ {
+ "epoch": 30.503056768558952,
+ "grad_norm": 0.01592106744647026,
+ "learning_rate": 0.0006,
+ "loss": 4.234043121337891,
+ "step": 2196
+ },
+ {
+ "epoch": 30.51703056768559,
+ "grad_norm": 0.015270989388227463,
+ "learning_rate": 0.0006,
+ "loss": 4.301420211791992,
+ "step": 2197
+ },
+ {
+ "epoch": 30.531004366812226,
+ "grad_norm": 0.015194724313914776,
+ "learning_rate": 0.0006,
+ "loss": 4.18703031539917,
+ "step": 2198
+ },
+ {
+ "epoch": 30.544978165938865,
+ "grad_norm": 0.015587746165692806,
+ "learning_rate": 0.0006,
+ "loss": 4.138928413391113,
+ "step": 2199
+ },
+ {
+ "epoch": 30.558951965065503,
+ "grad_norm": 0.016029570251703262,
+ "learning_rate": 0.0006,
+ "loss": 4.290863990783691,
+ "step": 2200
+ },
+ {
+ "epoch": 30.57292576419214,
+ "grad_norm": 0.015651309862732887,
+ "learning_rate": 0.0006,
+ "loss": 4.272062301635742,
+ "step": 2201
+ },
+ {
+ "epoch": 30.586899563318777,
+ "grad_norm": 0.015992306172847748,
+ "learning_rate": 0.0006,
+ "loss": 4.2487406730651855,
+ "step": 2202
+ },
+ {
+ "epoch": 30.600873362445416,
+ "grad_norm": 0.014969157986342907,
+ "learning_rate": 0.0006,
+ "loss": 4.115347862243652,
+ "step": 2203
+ },
+ {
+ "epoch": 30.61484716157205,
+ "grad_norm": 0.014639027416706085,
+ "learning_rate": 0.0006,
+ "loss": 4.2207489013671875,
+ "step": 2204
+ },
+ {
+ "epoch": 30.62882096069869,
+ "grad_norm": 0.015059413388371468,
+ "learning_rate": 0.0006,
+ "loss": 4.274044990539551,
+ "step": 2205
+ },
+ {
+ "epoch": 30.64279475982533,
+ "grad_norm": 0.0166641678661108,
+ "learning_rate": 0.0006,
+ "loss": 4.191373348236084,
+ "step": 2206
+ },
+ {
+ "epoch": 30.656768558951963,
+ "grad_norm": 0.017806345596909523,
+ "learning_rate": 0.0006,
+ "loss": 4.139800548553467,
+ "step": 2207
+ },
+ {
+ "epoch": 30.670742358078602,
+ "grad_norm": 0.018712421879172325,
+ "learning_rate": 0.0006,
+ "loss": 4.035965442657471,
+ "step": 2208
+ },
+ {
+ "epoch": 30.68471615720524,
+ "grad_norm": 0.018457243219017982,
+ "learning_rate": 0.0006,
+ "loss": 4.2579240798950195,
+ "step": 2209
+ },
+ {
+ "epoch": 30.69868995633188,
+ "grad_norm": 0.01920999586582184,
+ "learning_rate": 0.0006,
+ "loss": 4.131112098693848,
+ "step": 2210
+ },
+ {
+ "epoch": 30.712663755458514,
+ "grad_norm": 0.01729944534599781,
+ "learning_rate": 0.0006,
+ "loss": 4.254059314727783,
+ "step": 2211
+ },
+ {
+ "epoch": 30.726637554585153,
+ "grad_norm": 0.016914231702685356,
+ "learning_rate": 0.0006,
+ "loss": 4.233428955078125,
+ "step": 2212
+ },
+ {
+ "epoch": 30.74061135371179,
+ "grad_norm": 0.01913098618388176,
+ "learning_rate": 0.0006,
+ "loss": 4.23525333404541,
+ "step": 2213
+ },
+ {
+ "epoch": 30.754585152838427,
+ "grad_norm": 0.019150281324982643,
+ "learning_rate": 0.0006,
+ "loss": 4.213695049285889,
+ "step": 2214
+ },
+ {
+ "epoch": 30.768558951965066,
+ "grad_norm": 0.018947452306747437,
+ "learning_rate": 0.0006,
+ "loss": 4.151697158813477,
+ "step": 2215
+ },
+ {
+ "epoch": 30.782532751091704,
+ "grad_norm": 0.016058508306741714,
+ "learning_rate": 0.0006,
+ "loss": 4.220280170440674,
+ "step": 2216
+ },
+ {
+ "epoch": 30.79650655021834,
+ "grad_norm": 0.017488619312644005,
+ "learning_rate": 0.0006,
+ "loss": 4.154428482055664,
+ "step": 2217
+ },
+ {
+ "epoch": 30.810480349344978,
+ "grad_norm": 0.017355090007185936,
+ "learning_rate": 0.0006,
+ "loss": 4.256086349487305,
+ "step": 2218
+ },
+ {
+ "epoch": 30.824454148471617,
+ "grad_norm": 0.016636032611131668,
+ "learning_rate": 0.0006,
+ "loss": 4.227564334869385,
+ "step": 2219
+ },
+ {
+ "epoch": 30.83842794759825,
+ "grad_norm": 0.016273891553282738,
+ "learning_rate": 0.0006,
+ "loss": 4.200466156005859,
+ "step": 2220
+ },
+ {
+ "epoch": 30.85240174672489,
+ "grad_norm": 0.016939911991357803,
+ "learning_rate": 0.0006,
+ "loss": 4.1571221351623535,
+ "step": 2221
+ },
+ {
+ "epoch": 30.86637554585153,
+ "grad_norm": 0.018332522362470627,
+ "learning_rate": 0.0006,
+ "loss": 4.228398323059082,
+ "step": 2222
+ },
+ {
+ "epoch": 30.880349344978168,
+ "grad_norm": 0.019508114084601402,
+ "learning_rate": 0.0006,
+ "loss": 4.222960472106934,
+ "step": 2223
+ },
+ {
+ "epoch": 30.894323144104803,
+ "grad_norm": 0.018440278246998787,
+ "learning_rate": 0.0006,
+ "loss": 4.261051177978516,
+ "step": 2224
+ },
+ {
+ "epoch": 30.90829694323144,
+ "grad_norm": 0.018113229423761368,
+ "learning_rate": 0.0006,
+ "loss": 4.109235763549805,
+ "step": 2225
+ },
+ {
+ "epoch": 30.92227074235808,
+ "grad_norm": 0.015392083674669266,
+ "learning_rate": 0.0006,
+ "loss": 4.21950101852417,
+ "step": 2226
+ },
+ {
+ "epoch": 30.936244541484715,
+ "grad_norm": 0.01633290946483612,
+ "learning_rate": 0.0006,
+ "loss": 4.176053047180176,
+ "step": 2227
+ },
+ {
+ "epoch": 30.950218340611354,
+ "grad_norm": 0.016236383467912674,
+ "learning_rate": 0.0006,
+ "loss": 4.314309120178223,
+ "step": 2228
+ },
+ {
+ "epoch": 30.964192139737992,
+ "grad_norm": 0.015284373424947262,
+ "learning_rate": 0.0006,
+ "loss": 4.303133487701416,
+ "step": 2229
+ },
+ {
+ "epoch": 30.978165938864628,
+ "grad_norm": 0.015982897952198982,
+ "learning_rate": 0.0006,
+ "loss": 4.25739860534668,
+ "step": 2230
+ },
+ {
+ "epoch": 30.992139737991266,
+ "grad_norm": 0.016682198271155357,
+ "learning_rate": 0.0006,
+ "loss": 4.292210578918457,
+ "step": 2231
+ },
+ {
+ "epoch": 31.0,
+ "grad_norm": 0.017521467059850693,
+ "learning_rate": 0.0006,
+ "loss": 4.170637130737305,
+ "step": 2232
+ },
+ {
+ "epoch": 31.0,
+ "eval_loss": 4.642786979675293,
+ "eval_runtime": 57.029,
+ "eval_samples_per_second": 42.82,
+ "eval_steps_per_second": 1.35,
+ "step": 2232
+ },
+ {
+ "epoch": 31.01397379912664,
+ "grad_norm": 0.016607852652668953,
+ "learning_rate": 0.0006,
+ "loss": 4.106224536895752,
+ "step": 2233
+ },
+ {
+ "epoch": 31.027947598253274,
+ "grad_norm": 0.018448904156684875,
+ "learning_rate": 0.0006,
+ "loss": 4.224459648132324,
+ "step": 2234
+ },
+ {
+ "epoch": 31.041921397379912,
+ "grad_norm": 0.020481228828430176,
+ "learning_rate": 0.0006,
+ "loss": 4.305548667907715,
+ "step": 2235
+ },
+ {
+ "epoch": 31.05589519650655,
+ "grad_norm": 0.01999707892537117,
+ "learning_rate": 0.0006,
+ "loss": 4.253663539886475,
+ "step": 2236
+ },
+ {
+ "epoch": 31.069868995633186,
+ "grad_norm": 0.020428303629159927,
+ "learning_rate": 0.0006,
+ "loss": 4.166954040527344,
+ "step": 2237
+ },
+ {
+ "epoch": 31.083842794759825,
+ "grad_norm": 0.022608213126659393,
+ "learning_rate": 0.0006,
+ "loss": 4.172819137573242,
+ "step": 2238
+ },
+ {
+ "epoch": 31.097816593886463,
+ "grad_norm": 0.024669520556926727,
+ "learning_rate": 0.0006,
+ "loss": 4.106586456298828,
+ "step": 2239
+ },
+ {
+ "epoch": 31.111790393013102,
+ "grad_norm": 0.02247351035475731,
+ "learning_rate": 0.0006,
+ "loss": 4.160271644592285,
+ "step": 2240
+ },
+ {
+ "epoch": 31.125764192139737,
+ "grad_norm": 0.0207810141146183,
+ "learning_rate": 0.0006,
+ "loss": 4.272491931915283,
+ "step": 2241
+ },
+ {
+ "epoch": 31.139737991266376,
+ "grad_norm": 0.021576499566435814,
+ "learning_rate": 0.0006,
+ "loss": 4.171474456787109,
+ "step": 2242
+ },
+ {
+ "epoch": 31.153711790393015,
+ "grad_norm": 0.02337106689810753,
+ "learning_rate": 0.0006,
+ "loss": 4.094712734222412,
+ "step": 2243
+ },
+ {
+ "epoch": 31.16768558951965,
+ "grad_norm": 0.02219875156879425,
+ "learning_rate": 0.0006,
+ "loss": 4.1800994873046875,
+ "step": 2244
+ },
+ {
+ "epoch": 31.18165938864629,
+ "grad_norm": 0.022680504247546196,
+ "learning_rate": 0.0006,
+ "loss": 4.2054362297058105,
+ "step": 2245
+ },
+ {
+ "epoch": 31.195633187772927,
+ "grad_norm": 0.0208734143525362,
+ "learning_rate": 0.0006,
+ "loss": 4.06068229675293,
+ "step": 2246
+ },
+ {
+ "epoch": 31.209606986899562,
+ "grad_norm": 0.01811066083610058,
+ "learning_rate": 0.0006,
+ "loss": 4.2465128898620605,
+ "step": 2247
+ },
+ {
+ "epoch": 31.2235807860262,
+ "grad_norm": 0.018571637570858,
+ "learning_rate": 0.0006,
+ "loss": 4.191955089569092,
+ "step": 2248
+ },
+ {
+ "epoch": 31.23755458515284,
+ "grad_norm": 0.018598807975649834,
+ "learning_rate": 0.0006,
+ "loss": 4.200872421264648,
+ "step": 2249
+ },
+ {
+ "epoch": 31.251528384279474,
+ "grad_norm": 0.017577625811100006,
+ "learning_rate": 0.0006,
+ "loss": 4.148700714111328,
+ "step": 2250
+ },
+ {
+ "epoch": 31.265502183406113,
+ "grad_norm": 0.01817980408668518,
+ "learning_rate": 0.0006,
+ "loss": 4.256978988647461,
+ "step": 2251
+ },
+ {
+ "epoch": 31.27947598253275,
+ "grad_norm": 0.017983486875891685,
+ "learning_rate": 0.0006,
+ "loss": 4.175005912780762,
+ "step": 2252
+ },
+ {
+ "epoch": 31.29344978165939,
+ "grad_norm": 0.016751961782574654,
+ "learning_rate": 0.0006,
+ "loss": 4.237518310546875,
+ "step": 2253
+ },
+ {
+ "epoch": 31.307423580786025,
+ "grad_norm": 0.016407202929258347,
+ "learning_rate": 0.0006,
+ "loss": 4.197269439697266,
+ "step": 2254
+ },
+ {
+ "epoch": 31.321397379912664,
+ "grad_norm": 0.01578645594418049,
+ "learning_rate": 0.0006,
+ "loss": 4.179751396179199,
+ "step": 2255
+ },
+ {
+ "epoch": 31.335371179039303,
+ "grad_norm": 0.016001226380467415,
+ "learning_rate": 0.0006,
+ "loss": 4.08821964263916,
+ "step": 2256
+ },
+ {
+ "epoch": 31.349344978165938,
+ "grad_norm": 0.015122811309993267,
+ "learning_rate": 0.0006,
+ "loss": 4.047828674316406,
+ "step": 2257
+ },
+ {
+ "epoch": 31.363318777292577,
+ "grad_norm": 0.015399965457618237,
+ "learning_rate": 0.0006,
+ "loss": 4.1506524085998535,
+ "step": 2258
+ },
+ {
+ "epoch": 31.377292576419215,
+ "grad_norm": 0.01534605398774147,
+ "learning_rate": 0.0006,
+ "loss": 4.27952766418457,
+ "step": 2259
+ },
+ {
+ "epoch": 31.39126637554585,
+ "grad_norm": 0.015461381524801254,
+ "learning_rate": 0.0006,
+ "loss": 4.27102518081665,
+ "step": 2260
+ },
+ {
+ "epoch": 31.40524017467249,
+ "grad_norm": 0.016769153997302055,
+ "learning_rate": 0.0006,
+ "loss": 4.191592216491699,
+ "step": 2261
+ },
+ {
+ "epoch": 31.419213973799128,
+ "grad_norm": 0.016557959839701653,
+ "learning_rate": 0.0006,
+ "loss": 4.159705638885498,
+ "step": 2262
+ },
+ {
+ "epoch": 31.433187772925763,
+ "grad_norm": 0.01659569889307022,
+ "learning_rate": 0.0006,
+ "loss": 4.2136664390563965,
+ "step": 2263
+ },
+ {
+ "epoch": 31.4471615720524,
+ "grad_norm": 0.017101937904953957,
+ "learning_rate": 0.0006,
+ "loss": 4.2716498374938965,
+ "step": 2264
+ },
+ {
+ "epoch": 31.46113537117904,
+ "grad_norm": 0.01722985878586769,
+ "learning_rate": 0.0006,
+ "loss": 4.300534248352051,
+ "step": 2265
+ },
+ {
+ "epoch": 31.475109170305675,
+ "grad_norm": 0.016800185665488243,
+ "learning_rate": 0.0006,
+ "loss": 4.113240718841553,
+ "step": 2266
+ },
+ {
+ "epoch": 31.489082969432314,
+ "grad_norm": 0.01603875495493412,
+ "learning_rate": 0.0006,
+ "loss": 4.157105445861816,
+ "step": 2267
+ },
+ {
+ "epoch": 31.503056768558952,
+ "grad_norm": 0.015903932973742485,
+ "learning_rate": 0.0006,
+ "loss": 4.233669757843018,
+ "step": 2268
+ },
+ {
+ "epoch": 31.51703056768559,
+ "grad_norm": 0.0160983856767416,
+ "learning_rate": 0.0006,
+ "loss": 4.147704124450684,
+ "step": 2269
+ },
+ {
+ "epoch": 31.531004366812226,
+ "grad_norm": 0.01604730263352394,
+ "learning_rate": 0.0006,
+ "loss": 4.1465911865234375,
+ "step": 2270
+ },
+ {
+ "epoch": 31.544978165938865,
+ "grad_norm": 0.016205785796046257,
+ "learning_rate": 0.0006,
+ "loss": 4.106704235076904,
+ "step": 2271
+ },
+ {
+ "epoch": 31.558951965065503,
+ "grad_norm": 0.017008036375045776,
+ "learning_rate": 0.0006,
+ "loss": 4.191000938415527,
+ "step": 2272
+ },
+ {
+ "epoch": 31.57292576419214,
+ "grad_norm": 0.01699635200202465,
+ "learning_rate": 0.0006,
+ "loss": 4.187042236328125,
+ "step": 2273
+ },
+ {
+ "epoch": 31.586899563318777,
+ "grad_norm": 0.01692967116832733,
+ "learning_rate": 0.0006,
+ "loss": 4.208805561065674,
+ "step": 2274
+ },
+ {
+ "epoch": 31.600873362445416,
+ "grad_norm": 0.017608756199479103,
+ "learning_rate": 0.0006,
+ "loss": 4.200512886047363,
+ "step": 2275
+ },
+ {
+ "epoch": 31.61484716157205,
+ "grad_norm": 0.016590990126132965,
+ "learning_rate": 0.0006,
+ "loss": 4.194705009460449,
+ "step": 2276
+ },
+ {
+ "epoch": 31.62882096069869,
+ "grad_norm": 0.0170235987752676,
+ "learning_rate": 0.0006,
+ "loss": 4.210326194763184,
+ "step": 2277
+ },
+ {
+ "epoch": 31.64279475982533,
+ "grad_norm": 0.01604801043868065,
+ "learning_rate": 0.0006,
+ "loss": 4.196024417877197,
+ "step": 2278
+ },
+ {
+ "epoch": 31.656768558951963,
+ "grad_norm": 0.015694202855229378,
+ "learning_rate": 0.0006,
+ "loss": 4.178236484527588,
+ "step": 2279
+ },
+ {
+ "epoch": 31.670742358078602,
+ "grad_norm": 0.015473959036171436,
+ "learning_rate": 0.0006,
+ "loss": 4.259339809417725,
+ "step": 2280
+ },
+ {
+ "epoch": 31.68471615720524,
+ "grad_norm": 0.014689362607896328,
+ "learning_rate": 0.0006,
+ "loss": 3.9702324867248535,
+ "step": 2281
+ },
+ {
+ "epoch": 31.69868995633188,
+ "grad_norm": 0.015245123766362667,
+ "learning_rate": 0.0006,
+ "loss": 4.197025775909424,
+ "step": 2282
+ },
+ {
+ "epoch": 31.712663755458514,
+ "grad_norm": 0.016148347407579422,
+ "learning_rate": 0.0006,
+ "loss": 4.231856346130371,
+ "step": 2283
+ },
+ {
+ "epoch": 31.726637554585153,
+ "grad_norm": 0.015169818885624409,
+ "learning_rate": 0.0006,
+ "loss": 4.230169296264648,
+ "step": 2284
+ },
+ {
+ "epoch": 31.74061135371179,
+ "grad_norm": 0.014843013137578964,
+ "learning_rate": 0.0006,
+ "loss": 4.195633411407471,
+ "step": 2285
+ },
+ {
+ "epoch": 31.754585152838427,
+ "grad_norm": 0.015229340642690659,
+ "learning_rate": 0.0006,
+ "loss": 4.231280326843262,
+ "step": 2286
+ },
+ {
+ "epoch": 31.768558951965066,
+ "grad_norm": 0.01551708485931158,
+ "learning_rate": 0.0006,
+ "loss": 4.17064094543457,
+ "step": 2287
+ },
+ {
+ "epoch": 31.782532751091704,
+ "grad_norm": 0.015863575041294098,
+ "learning_rate": 0.0006,
+ "loss": 4.1579484939575195,
+ "step": 2288
+ },
+ {
+ "epoch": 31.79650655021834,
+ "grad_norm": 0.015118198469281197,
+ "learning_rate": 0.0006,
+ "loss": 4.1626787185668945,
+ "step": 2289
+ },
+ {
+ "epoch": 31.810480349344978,
+ "grad_norm": 0.015220776200294495,
+ "learning_rate": 0.0006,
+ "loss": 4.135746955871582,
+ "step": 2290
+ },
+ {
+ "epoch": 31.824454148471617,
+ "grad_norm": 0.014529784210026264,
+ "learning_rate": 0.0006,
+ "loss": 4.164931297302246,
+ "step": 2291
+ },
+ {
+ "epoch": 31.83842794759825,
+ "grad_norm": 0.015533296391367912,
+ "learning_rate": 0.0006,
+ "loss": 4.112464904785156,
+ "step": 2292
+ },
+ {
+ "epoch": 31.85240174672489,
+ "grad_norm": 0.016562335193157196,
+ "learning_rate": 0.0006,
+ "loss": 4.164898872375488,
+ "step": 2293
+ },
+ {
+ "epoch": 31.86637554585153,
+ "grad_norm": 0.016258113086223602,
+ "learning_rate": 0.0006,
+ "loss": 4.1979265213012695,
+ "step": 2294
+ },
+ {
+ "epoch": 31.880349344978168,
+ "grad_norm": 0.01558110024780035,
+ "learning_rate": 0.0006,
+ "loss": 4.174675941467285,
+ "step": 2295
+ },
+ {
+ "epoch": 31.894323144104803,
+ "grad_norm": 0.016412515193223953,
+ "learning_rate": 0.0006,
+ "loss": 4.217909336090088,
+ "step": 2296
+ },
+ {
+ "epoch": 31.90829694323144,
+ "grad_norm": 0.015118442475795746,
+ "learning_rate": 0.0006,
+ "loss": 4.137143611907959,
+ "step": 2297
+ },
+ {
+ "epoch": 31.92227074235808,
+ "grad_norm": 0.01506161317229271,
+ "learning_rate": 0.0006,
+ "loss": 4.196122169494629,
+ "step": 2298
+ },
+ {
+ "epoch": 31.936244541484715,
+ "grad_norm": 0.017165271565318108,
+ "learning_rate": 0.0006,
+ "loss": 4.14834451675415,
+ "step": 2299
+ },
+ {
+ "epoch": 31.950218340611354,
+ "grad_norm": 0.01696440577507019,
+ "learning_rate": 0.0006,
+ "loss": 4.218868732452393,
+ "step": 2300
+ },
+ {
+ "epoch": 31.964192139737992,
+ "grad_norm": 0.017349394038319588,
+ "learning_rate": 0.0006,
+ "loss": 4.2754011154174805,
+ "step": 2301
+ },
+ {
+ "epoch": 31.978165938864628,
+ "grad_norm": 0.018387921154499054,
+ "learning_rate": 0.0006,
+ "loss": 4.282623291015625,
+ "step": 2302
+ },
+ {
+ "epoch": 31.992139737991266,
+ "grad_norm": 0.018837926909327507,
+ "learning_rate": 0.0006,
+ "loss": 4.185425281524658,
+ "step": 2303
+ },
+ {
+ "epoch": 32.0,
+ "grad_norm": 0.018998464569449425,
+ "learning_rate": 0.0006,
+ "loss": 4.151291847229004,
+ "step": 2304
+ },
+ {
+ "epoch": 32.0,
+ "eval_loss": 4.522614002227783,
+ "eval_runtime": 57.0106,
+ "eval_samples_per_second": 42.834,
+ "eval_steps_per_second": 1.351,
+ "step": 2304
+ },
+ {
+ "epoch": 32.01397379912664,
+ "grad_norm": 0.019036833196878433,
+ "learning_rate": 0.0006,
+ "loss": 4.183202266693115,
+ "step": 2305
+ },
+ {
+ "epoch": 32.02794759825328,
+ "grad_norm": 0.02207397110760212,
+ "learning_rate": 0.0006,
+ "loss": 4.17875862121582,
+ "step": 2306
+ },
+ {
+ "epoch": 32.041921397379916,
+ "grad_norm": 0.022008279338479042,
+ "learning_rate": 0.0006,
+ "loss": 4.08042049407959,
+ "step": 2307
+ },
+ {
+ "epoch": 32.05589519650655,
+ "grad_norm": 0.022407017648220062,
+ "learning_rate": 0.0006,
+ "loss": 4.112189769744873,
+ "step": 2308
+ },
+ {
+ "epoch": 32.069868995633186,
+ "grad_norm": 0.021599747240543365,
+ "learning_rate": 0.0006,
+ "loss": 4.126455307006836,
+ "step": 2309
+ },
+ {
+ "epoch": 32.083842794759825,
+ "grad_norm": 0.020340966060757637,
+ "learning_rate": 0.0006,
+ "loss": 4.080538272857666,
+ "step": 2310
+ },
+ {
+ "epoch": 32.09781659388646,
+ "grad_norm": 0.020956892520189285,
+ "learning_rate": 0.0006,
+ "loss": 4.171943664550781,
+ "step": 2311
+ },
+ {
+ "epoch": 32.1117903930131,
+ "grad_norm": 0.019966667518019676,
+ "learning_rate": 0.0006,
+ "loss": 4.144007682800293,
+ "step": 2312
+ },
+ {
+ "epoch": 32.12576419213974,
+ "grad_norm": 0.018256952986121178,
+ "learning_rate": 0.0006,
+ "loss": 4.208107948303223,
+ "step": 2313
+ },
+ {
+ "epoch": 32.13973799126637,
+ "grad_norm": 0.017289170995354652,
+ "learning_rate": 0.0006,
+ "loss": 4.182024955749512,
+ "step": 2314
+ },
+ {
+ "epoch": 32.15371179039301,
+ "grad_norm": 0.017656700685620308,
+ "learning_rate": 0.0006,
+ "loss": 4.068373680114746,
+ "step": 2315
+ },
+ {
+ "epoch": 32.16768558951965,
+ "grad_norm": 0.017903361469507217,
+ "learning_rate": 0.0006,
+ "loss": 4.2345428466796875,
+ "step": 2316
+ },
+ {
+ "epoch": 32.18165938864629,
+ "grad_norm": 0.01785109005868435,
+ "learning_rate": 0.0006,
+ "loss": 4.219776153564453,
+ "step": 2317
+ },
+ {
+ "epoch": 32.19563318777293,
+ "grad_norm": 0.01887257769703865,
+ "learning_rate": 0.0006,
+ "loss": 4.139206409454346,
+ "step": 2318
+ },
+ {
+ "epoch": 32.209606986899566,
+ "grad_norm": 0.01924685761332512,
+ "learning_rate": 0.0006,
+ "loss": 4.121052265167236,
+ "step": 2319
+ },
+ {
+ "epoch": 32.223580786026204,
+ "grad_norm": 0.020574072375893593,
+ "learning_rate": 0.0006,
+ "loss": 4.13262414932251,
+ "step": 2320
+ },
+ {
+ "epoch": 32.237554585152836,
+ "grad_norm": 0.023849064484238625,
+ "learning_rate": 0.0006,
+ "loss": 4.239119529724121,
+ "step": 2321
+ },
+ {
+ "epoch": 32.251528384279474,
+ "grad_norm": 0.026239361613988876,
+ "learning_rate": 0.0006,
+ "loss": 4.248993873596191,
+ "step": 2322
+ },
+ {
+ "epoch": 32.26550218340611,
+ "grad_norm": 0.023980243131518364,
+ "learning_rate": 0.0006,
+ "loss": 4.159852504730225,
+ "step": 2323
+ },
+ {
+ "epoch": 32.27947598253275,
+ "grad_norm": 0.020442405715584755,
+ "learning_rate": 0.0006,
+ "loss": 4.133546829223633,
+ "step": 2324
+ },
+ {
+ "epoch": 32.29344978165939,
+ "grad_norm": 0.020661186426877975,
+ "learning_rate": 0.0006,
+ "loss": 4.078243732452393,
+ "step": 2325
+ },
+ {
+ "epoch": 32.30742358078603,
+ "grad_norm": 0.02046387456357479,
+ "learning_rate": 0.0006,
+ "loss": 4.172012805938721,
+ "step": 2326
+ },
+ {
+ "epoch": 32.32139737991266,
+ "grad_norm": 0.02089597098529339,
+ "learning_rate": 0.0006,
+ "loss": 4.0604705810546875,
+ "step": 2327
+ },
+ {
+ "epoch": 32.3353711790393,
+ "grad_norm": 0.01875591278076172,
+ "learning_rate": 0.0006,
+ "loss": 4.125911235809326,
+ "step": 2328
+ },
+ {
+ "epoch": 32.34934497816594,
+ "grad_norm": 0.018606796860694885,
+ "learning_rate": 0.0006,
+ "loss": 4.097362518310547,
+ "step": 2329
+ },
+ {
+ "epoch": 32.36331877729258,
+ "grad_norm": 0.01766626164317131,
+ "learning_rate": 0.0006,
+ "loss": 4.118616580963135,
+ "step": 2330
+ },
+ {
+ "epoch": 32.377292576419215,
+ "grad_norm": 0.017692960798740387,
+ "learning_rate": 0.0006,
+ "loss": 4.1457672119140625,
+ "step": 2331
+ },
+ {
+ "epoch": 32.391266375545854,
+ "grad_norm": 0.017619704827666283,
+ "learning_rate": 0.0006,
+ "loss": 4.175961494445801,
+ "step": 2332
+ },
+ {
+ "epoch": 32.40524017467249,
+ "grad_norm": 0.01784413494169712,
+ "learning_rate": 0.0006,
+ "loss": 4.051291465759277,
+ "step": 2333
+ },
+ {
+ "epoch": 32.419213973799124,
+ "grad_norm": 0.018508171662688255,
+ "learning_rate": 0.0006,
+ "loss": 4.0638556480407715,
+ "step": 2334
+ },
+ {
+ "epoch": 32.43318777292576,
+ "grad_norm": 0.01731436885893345,
+ "learning_rate": 0.0006,
+ "loss": 4.1805419921875,
+ "step": 2335
+ },
+ {
+ "epoch": 32.4471615720524,
+ "grad_norm": 0.015886498615145683,
+ "learning_rate": 0.0006,
+ "loss": 4.151026725769043,
+ "step": 2336
+ },
+ {
+ "epoch": 32.46113537117904,
+ "grad_norm": 0.016583820804953575,
+ "learning_rate": 0.0006,
+ "loss": 4.203604698181152,
+ "step": 2337
+ },
+ {
+ "epoch": 32.47510917030568,
+ "grad_norm": 0.0176097322255373,
+ "learning_rate": 0.0006,
+ "loss": 4.135159015655518,
+ "step": 2338
+ },
+ {
+ "epoch": 32.48908296943232,
+ "grad_norm": 0.016657903790473938,
+ "learning_rate": 0.0006,
+ "loss": 4.126681804656982,
+ "step": 2339
+ },
+ {
+ "epoch": 32.50305676855895,
+ "grad_norm": 0.016866130754351616,
+ "learning_rate": 0.0006,
+ "loss": 4.162341117858887,
+ "step": 2340
+ },
+ {
+ "epoch": 32.51703056768559,
+ "grad_norm": 0.016286680474877357,
+ "learning_rate": 0.0006,
+ "loss": 4.261801242828369,
+ "step": 2341
+ },
+ {
+ "epoch": 32.531004366812226,
+ "grad_norm": 0.01481365505605936,
+ "learning_rate": 0.0006,
+ "loss": 4.110016822814941,
+ "step": 2342
+ },
+ {
+ "epoch": 32.544978165938865,
+ "grad_norm": 0.016605013981461525,
+ "learning_rate": 0.0006,
+ "loss": 4.230024814605713,
+ "step": 2343
+ },
+ {
+ "epoch": 32.5589519650655,
+ "grad_norm": 0.01645175740122795,
+ "learning_rate": 0.0006,
+ "loss": 4.174612045288086,
+ "step": 2344
+ },
+ {
+ "epoch": 32.57292576419214,
+ "grad_norm": 0.014725026674568653,
+ "learning_rate": 0.0006,
+ "loss": 4.066371440887451,
+ "step": 2345
+ },
+ {
+ "epoch": 32.58689956331878,
+ "grad_norm": 0.015003837645053864,
+ "learning_rate": 0.0006,
+ "loss": 4.193343639373779,
+ "step": 2346
+ },
+ {
+ "epoch": 32.60087336244541,
+ "grad_norm": 0.015346302650868893,
+ "learning_rate": 0.0006,
+ "loss": 4.210659980773926,
+ "step": 2347
+ },
+ {
+ "epoch": 32.61484716157205,
+ "grad_norm": 0.015135962516069412,
+ "learning_rate": 0.0006,
+ "loss": 4.204618453979492,
+ "step": 2348
+ },
+ {
+ "epoch": 32.62882096069869,
+ "grad_norm": 0.01601956970989704,
+ "learning_rate": 0.0006,
+ "loss": 4.145564556121826,
+ "step": 2349
+ },
+ {
+ "epoch": 32.64279475982533,
+ "grad_norm": 0.01517851185053587,
+ "learning_rate": 0.0006,
+ "loss": 4.051673889160156,
+ "step": 2350
+ },
+ {
+ "epoch": 32.65676855895197,
+ "grad_norm": 0.014894292689859867,
+ "learning_rate": 0.0006,
+ "loss": 4.184818267822266,
+ "step": 2351
+ },
+ {
+ "epoch": 32.670742358078606,
+ "grad_norm": 0.015720834955573082,
+ "learning_rate": 0.0006,
+ "loss": 4.220036506652832,
+ "step": 2352
+ },
+ {
+ "epoch": 32.68471615720524,
+ "grad_norm": 0.01573433168232441,
+ "learning_rate": 0.0006,
+ "loss": 4.076773166656494,
+ "step": 2353
+ },
+ {
+ "epoch": 32.698689956331876,
+ "grad_norm": 0.015828052535653114,
+ "learning_rate": 0.0006,
+ "loss": 4.116329669952393,
+ "step": 2354
+ },
+ {
+ "epoch": 32.712663755458514,
+ "grad_norm": 0.015462259761989117,
+ "learning_rate": 0.0006,
+ "loss": 4.1155242919921875,
+ "step": 2355
+ },
+ {
+ "epoch": 32.72663755458515,
+ "grad_norm": 0.015763849020004272,
+ "learning_rate": 0.0006,
+ "loss": 4.169687747955322,
+ "step": 2356
+ },
+ {
+ "epoch": 32.74061135371179,
+ "grad_norm": 0.014935145154595375,
+ "learning_rate": 0.0006,
+ "loss": 4.2090911865234375,
+ "step": 2357
+ },
+ {
+ "epoch": 32.75458515283843,
+ "grad_norm": 0.015119184739887714,
+ "learning_rate": 0.0006,
+ "loss": 4.131838798522949,
+ "step": 2358
+ },
+ {
+ "epoch": 32.76855895196506,
+ "grad_norm": 0.014833525754511356,
+ "learning_rate": 0.0006,
+ "loss": 4.17618465423584,
+ "step": 2359
+ },
+ {
+ "epoch": 32.7825327510917,
+ "grad_norm": 0.01588359661400318,
+ "learning_rate": 0.0006,
+ "loss": 4.163723945617676,
+ "step": 2360
+ },
+ {
+ "epoch": 32.79650655021834,
+ "grad_norm": 0.0168188214302063,
+ "learning_rate": 0.0006,
+ "loss": 4.178189277648926,
+ "step": 2361
+ },
+ {
+ "epoch": 32.81048034934498,
+ "grad_norm": 0.01726095750927925,
+ "learning_rate": 0.0006,
+ "loss": 4.144223690032959,
+ "step": 2362
+ },
+ {
+ "epoch": 32.82445414847162,
+ "grad_norm": 0.01701410301029682,
+ "learning_rate": 0.0006,
+ "loss": 4.209896087646484,
+ "step": 2363
+ },
+ {
+ "epoch": 32.838427947598255,
+ "grad_norm": 0.01682264916598797,
+ "learning_rate": 0.0006,
+ "loss": 4.26223087310791,
+ "step": 2364
+ },
+ {
+ "epoch": 32.852401746724894,
+ "grad_norm": 0.015597456134855747,
+ "learning_rate": 0.0006,
+ "loss": 4.292664051055908,
+ "step": 2365
+ },
+ {
+ "epoch": 32.866375545851525,
+ "grad_norm": 0.015258579514920712,
+ "learning_rate": 0.0006,
+ "loss": 4.165006637573242,
+ "step": 2366
+ },
+ {
+ "epoch": 32.880349344978164,
+ "grad_norm": 0.01662200503051281,
+ "learning_rate": 0.0006,
+ "loss": 4.309786319732666,
+ "step": 2367
+ },
+ {
+ "epoch": 32.8943231441048,
+ "grad_norm": 0.015691161155700684,
+ "learning_rate": 0.0006,
+ "loss": 4.167202949523926,
+ "step": 2368
+ },
+ {
+ "epoch": 32.90829694323144,
+ "grad_norm": 0.014870637096464634,
+ "learning_rate": 0.0006,
+ "loss": 4.142415523529053,
+ "step": 2369
+ },
+ {
+ "epoch": 32.92227074235808,
+ "grad_norm": 0.01567252166569233,
+ "learning_rate": 0.0006,
+ "loss": 4.2312092781066895,
+ "step": 2370
+ },
+ {
+ "epoch": 32.93624454148472,
+ "grad_norm": 0.014948616735637188,
+ "learning_rate": 0.0006,
+ "loss": 4.1514129638671875,
+ "step": 2371
+ },
+ {
+ "epoch": 32.95021834061135,
+ "grad_norm": 0.014654500409960747,
+ "learning_rate": 0.0006,
+ "loss": 4.205010890960693,
+ "step": 2372
+ },
+ {
+ "epoch": 32.96419213973799,
+ "grad_norm": 0.015059148892760277,
+ "learning_rate": 0.0006,
+ "loss": 4.148405075073242,
+ "step": 2373
+ },
+ {
+ "epoch": 32.97816593886463,
+ "grad_norm": 0.015026576817035675,
+ "learning_rate": 0.0006,
+ "loss": 4.182535171508789,
+ "step": 2374
+ },
+ {
+ "epoch": 32.992139737991266,
+ "grad_norm": 0.016394320875406265,
+ "learning_rate": 0.0006,
+ "loss": 4.129051208496094,
+ "step": 2375
+ },
+ {
+ "epoch": 33.0,
+ "grad_norm": 0.01867043413221836,
+ "learning_rate": 0.0006,
+ "loss": 4.185556411743164,
+ "step": 2376
+ },
+ {
+ "epoch": 33.0,
+ "eval_loss": 4.640405654907227,
+ "eval_runtime": 56.5861,
+ "eval_samples_per_second": 43.155,
+ "eval_steps_per_second": 1.361,
+ "step": 2376
+ },
+ {
+ "epoch": 33.01397379912664,
+ "grad_norm": 0.018626727163791656,
+ "learning_rate": 0.0006,
+ "loss": 4.143031597137451,
+ "step": 2377
+ },
+ {
+ "epoch": 33.02794759825328,
+ "grad_norm": 0.019283363595604897,
+ "learning_rate": 0.0006,
+ "loss": 4.117845058441162,
+ "step": 2378
+ },
+ {
+ "epoch": 33.041921397379916,
+ "grad_norm": 0.018185822293162346,
+ "learning_rate": 0.0006,
+ "loss": 4.076862812042236,
+ "step": 2379
+ },
+ {
+ "epoch": 33.05589519650655,
+ "grad_norm": 0.019411196932196617,
+ "learning_rate": 0.0006,
+ "loss": 4.124156475067139,
+ "step": 2380
+ },
+ {
+ "epoch": 33.069868995633186,
+ "grad_norm": 0.019248466938734055,
+ "learning_rate": 0.0006,
+ "loss": 4.201735019683838,
+ "step": 2381
+ },
+ {
+ "epoch": 33.083842794759825,
+ "grad_norm": 0.018856720998883247,
+ "learning_rate": 0.0006,
+ "loss": 4.00296688079834,
+ "step": 2382
+ },
+ {
+ "epoch": 33.09781659388646,
+ "grad_norm": 0.018554294481873512,
+ "learning_rate": 0.0006,
+ "loss": 4.270577430725098,
+ "step": 2383
+ },
+ {
+ "epoch": 33.1117903930131,
+ "grad_norm": 0.019946224987506866,
+ "learning_rate": 0.0006,
+ "loss": 4.119589328765869,
+ "step": 2384
+ },
+ {
+ "epoch": 33.12576419213974,
+ "grad_norm": 0.021395670250058174,
+ "learning_rate": 0.0006,
+ "loss": 4.128847122192383,
+ "step": 2385
+ },
+ {
+ "epoch": 33.13973799126637,
+ "grad_norm": 0.02009674906730652,
+ "learning_rate": 0.0006,
+ "loss": 4.183228969573975,
+ "step": 2386
+ },
+ {
+ "epoch": 33.15371179039301,
+ "grad_norm": 0.01820223033428192,
+ "learning_rate": 0.0006,
+ "loss": 4.131601810455322,
+ "step": 2387
+ },
+ {
+ "epoch": 33.16768558951965,
+ "grad_norm": 0.018653156235814095,
+ "learning_rate": 0.0006,
+ "loss": 4.116430282592773,
+ "step": 2388
+ },
+ {
+ "epoch": 33.18165938864629,
+ "grad_norm": 0.016828063875436783,
+ "learning_rate": 0.0006,
+ "loss": 4.054663181304932,
+ "step": 2389
+ },
+ {
+ "epoch": 33.19563318777293,
+ "grad_norm": 0.017025409266352654,
+ "learning_rate": 0.0006,
+ "loss": 4.110044956207275,
+ "step": 2390
+ },
+ {
+ "epoch": 33.209606986899566,
+ "grad_norm": 0.016608452424407005,
+ "learning_rate": 0.0006,
+ "loss": 4.102668285369873,
+ "step": 2391
+ },
+ {
+ "epoch": 33.223580786026204,
+ "grad_norm": 0.01743577979505062,
+ "learning_rate": 0.0006,
+ "loss": 4.166791915893555,
+ "step": 2392
+ },
+ {
+ "epoch": 33.237554585152836,
+ "grad_norm": 0.017253734171390533,
+ "learning_rate": 0.0006,
+ "loss": 4.217832088470459,
+ "step": 2393
+ },
+ {
+ "epoch": 33.251528384279474,
+ "grad_norm": 0.017937595024704933,
+ "learning_rate": 0.0006,
+ "loss": 4.2173967361450195,
+ "step": 2394
+ },
+ {
+ "epoch": 33.26550218340611,
+ "grad_norm": 0.01985993981361389,
+ "learning_rate": 0.0006,
+ "loss": 4.17849588394165,
+ "step": 2395
+ },
+ {
+ "epoch": 33.27947598253275,
+ "grad_norm": 0.019243355840444565,
+ "learning_rate": 0.0006,
+ "loss": 4.086523056030273,
+ "step": 2396
+ },
+ {
+ "epoch": 33.29344978165939,
+ "grad_norm": 0.01924416422843933,
+ "learning_rate": 0.0006,
+ "loss": 4.128574371337891,
+ "step": 2397
+ },
+ {
+ "epoch": 33.30742358078603,
+ "grad_norm": 0.01983758620917797,
+ "learning_rate": 0.0006,
+ "loss": 4.061581611633301,
+ "step": 2398
+ },
+ {
+ "epoch": 33.32139737991266,
+ "grad_norm": 0.01930462196469307,
+ "learning_rate": 0.0006,
+ "loss": 4.091731071472168,
+ "step": 2399
+ },
+ {
+ "epoch": 33.3353711790393,
+ "grad_norm": 0.017988424748182297,
+ "learning_rate": 0.0006,
+ "loss": 4.155599117279053,
+ "step": 2400
+ },
+ {
+ "epoch": 33.34934497816594,
+ "grad_norm": 0.015194767154753208,
+ "learning_rate": 0.0006,
+ "loss": 4.144134044647217,
+ "step": 2401
+ },
+ {
+ "epoch": 33.36331877729258,
+ "grad_norm": 0.01536081824451685,
+ "learning_rate": 0.0006,
+ "loss": 4.127272129058838,
+ "step": 2402
+ },
+ {
+ "epoch": 33.377292576419215,
+ "grad_norm": 0.015431297942996025,
+ "learning_rate": 0.0006,
+ "loss": 4.3213019371032715,
+ "step": 2403
+ },
+ {
+ "epoch": 33.391266375545854,
+ "grad_norm": 0.015740280970931053,
+ "learning_rate": 0.0006,
+ "loss": 4.121467590332031,
+ "step": 2404
+ },
+ {
+ "epoch": 33.40524017467249,
+ "grad_norm": 0.015379384160041809,
+ "learning_rate": 0.0006,
+ "loss": 4.095216751098633,
+ "step": 2405
+ },
+ {
+ "epoch": 33.419213973799124,
+ "grad_norm": 0.01528073474764824,
+ "learning_rate": 0.0006,
+ "loss": 4.13047981262207,
+ "step": 2406
+ },
+ {
+ "epoch": 33.43318777292576,
+ "grad_norm": 0.015739979222416878,
+ "learning_rate": 0.0006,
+ "loss": 4.2533135414123535,
+ "step": 2407
+ },
+ {
+ "epoch": 33.4471615720524,
+ "grad_norm": 0.01519764307886362,
+ "learning_rate": 0.0006,
+ "loss": 4.131782054901123,
+ "step": 2408
+ },
+ {
+ "epoch": 33.46113537117904,
+ "grad_norm": 0.015425996854901314,
+ "learning_rate": 0.0006,
+ "loss": 4.02811336517334,
+ "step": 2409
+ },
+ {
+ "epoch": 33.47510917030568,
+ "grad_norm": 0.015378830023109913,
+ "learning_rate": 0.0006,
+ "loss": 4.08652925491333,
+ "step": 2410
+ },
+ {
+ "epoch": 33.48908296943232,
+ "grad_norm": 0.01473075058311224,
+ "learning_rate": 0.0006,
+ "loss": 4.214364528656006,
+ "step": 2411
+ },
+ {
+ "epoch": 33.50305676855895,
+ "grad_norm": 0.015141637064516544,
+ "learning_rate": 0.0006,
+ "loss": 4.155712127685547,
+ "step": 2412
+ },
+ {
+ "epoch": 33.51703056768559,
+ "grad_norm": 0.01641593500971794,
+ "learning_rate": 0.0006,
+ "loss": 4.097907066345215,
+ "step": 2413
+ },
+ {
+ "epoch": 33.531004366812226,
+ "grad_norm": 0.015182029455900192,
+ "learning_rate": 0.0006,
+ "loss": 4.238499641418457,
+ "step": 2414
+ },
+ {
+ "epoch": 33.544978165938865,
+ "grad_norm": 0.01643732376396656,
+ "learning_rate": 0.0006,
+ "loss": 4.0661492347717285,
+ "step": 2415
+ },
+ {
+ "epoch": 33.5589519650655,
+ "grad_norm": 0.015614228323101997,
+ "learning_rate": 0.0006,
+ "loss": 4.120795249938965,
+ "step": 2416
+ },
+ {
+ "epoch": 33.57292576419214,
+ "grad_norm": 0.016066187992691994,
+ "learning_rate": 0.0006,
+ "loss": 4.096465110778809,
+ "step": 2417
+ },
+ {
+ "epoch": 33.58689956331878,
+ "grad_norm": 0.017214342951774597,
+ "learning_rate": 0.0006,
+ "loss": 4.242628574371338,
+ "step": 2418
+ },
+ {
+ "epoch": 33.60087336244541,
+ "grad_norm": 0.01787586510181427,
+ "learning_rate": 0.0006,
+ "loss": 4.149589538574219,
+ "step": 2419
+ },
+ {
+ "epoch": 33.61484716157205,
+ "grad_norm": 0.019455241039395332,
+ "learning_rate": 0.0006,
+ "loss": 4.267796039581299,
+ "step": 2420
+ },
+ {
+ "epoch": 33.62882096069869,
+ "grad_norm": 0.01984674669802189,
+ "learning_rate": 0.0006,
+ "loss": 4.175220966339111,
+ "step": 2421
+ },
+ {
+ "epoch": 33.64279475982533,
+ "grad_norm": 0.021933717653155327,
+ "learning_rate": 0.0006,
+ "loss": 4.106647491455078,
+ "step": 2422
+ },
+ {
+ "epoch": 33.65676855895197,
+ "grad_norm": 0.02123589813709259,
+ "learning_rate": 0.0006,
+ "loss": 4.204996109008789,
+ "step": 2423
+ },
+ {
+ "epoch": 33.670742358078606,
+ "grad_norm": 0.019814589992165565,
+ "learning_rate": 0.0006,
+ "loss": 4.186171531677246,
+ "step": 2424
+ },
+ {
+ "epoch": 33.68471615720524,
+ "grad_norm": 0.02059728465974331,
+ "learning_rate": 0.0006,
+ "loss": 4.170509338378906,
+ "step": 2425
+ },
+ {
+ "epoch": 33.698689956331876,
+ "grad_norm": 0.018483608961105347,
+ "learning_rate": 0.0006,
+ "loss": 4.085585594177246,
+ "step": 2426
+ },
+ {
+ "epoch": 33.712663755458514,
+ "grad_norm": 0.018403852358460426,
+ "learning_rate": 0.0006,
+ "loss": 4.2551774978637695,
+ "step": 2427
+ },
+ {
+ "epoch": 33.72663755458515,
+ "grad_norm": 0.018653830513358116,
+ "learning_rate": 0.0006,
+ "loss": 4.091766357421875,
+ "step": 2428
+ },
+ {
+ "epoch": 33.74061135371179,
+ "grad_norm": 0.017692334949970245,
+ "learning_rate": 0.0006,
+ "loss": 4.100838661193848,
+ "step": 2429
+ },
+ {
+ "epoch": 33.75458515283843,
+ "grad_norm": 0.016029126942157745,
+ "learning_rate": 0.0006,
+ "loss": 4.1530680656433105,
+ "step": 2430
+ },
+ {
+ "epoch": 33.76855895196506,
+ "grad_norm": 0.016095684841275215,
+ "learning_rate": 0.0006,
+ "loss": 4.20544958114624,
+ "step": 2431
+ },
+ {
+ "epoch": 33.7825327510917,
+ "grad_norm": 0.016375798732042313,
+ "learning_rate": 0.0006,
+ "loss": 4.199177265167236,
+ "step": 2432
+ },
+ {
+ "epoch": 33.79650655021834,
+ "grad_norm": 0.01590786688029766,
+ "learning_rate": 0.0006,
+ "loss": 4.200799465179443,
+ "step": 2433
+ },
+ {
+ "epoch": 33.81048034934498,
+ "grad_norm": 0.014896097593009472,
+ "learning_rate": 0.0006,
+ "loss": 4.185850143432617,
+ "step": 2434
+ },
+ {
+ "epoch": 33.82445414847162,
+ "grad_norm": 0.01438513956964016,
+ "learning_rate": 0.0006,
+ "loss": 4.214962005615234,
+ "step": 2435
+ },
+ {
+ "epoch": 33.838427947598255,
+ "grad_norm": 0.014459339901804924,
+ "learning_rate": 0.0006,
+ "loss": 4.196972370147705,
+ "step": 2436
+ },
+ {
+ "epoch": 33.852401746724894,
+ "grad_norm": 0.014782169833779335,
+ "learning_rate": 0.0006,
+ "loss": 4.0885910987854,
+ "step": 2437
+ },
+ {
+ "epoch": 33.866375545851525,
+ "grad_norm": 0.01520370040088892,
+ "learning_rate": 0.0006,
+ "loss": 4.0323262214660645,
+ "step": 2438
+ },
+ {
+ "epoch": 33.880349344978164,
+ "grad_norm": 0.014056282117962837,
+ "learning_rate": 0.0006,
+ "loss": 4.1386518478393555,
+ "step": 2439
+ },
+ {
+ "epoch": 33.8943231441048,
+ "grad_norm": 0.013567990623414516,
+ "learning_rate": 0.0006,
+ "loss": 4.095754623413086,
+ "step": 2440
+ },
+ {
+ "epoch": 33.90829694323144,
+ "grad_norm": 0.013306716457009315,
+ "learning_rate": 0.0006,
+ "loss": 4.1623077392578125,
+ "step": 2441
+ },
+ {
+ "epoch": 33.92227074235808,
+ "grad_norm": 0.014047150500118732,
+ "learning_rate": 0.0006,
+ "loss": 4.142541885375977,
+ "step": 2442
+ },
+ {
+ "epoch": 33.93624454148472,
+ "grad_norm": 0.01377617847174406,
+ "learning_rate": 0.0006,
+ "loss": 4.268551826477051,
+ "step": 2443
+ },
+ {
+ "epoch": 33.95021834061135,
+ "grad_norm": 0.014671621844172478,
+ "learning_rate": 0.0006,
+ "loss": 4.198196887969971,
+ "step": 2444
+ },
+ {
+ "epoch": 33.96419213973799,
+ "grad_norm": 0.014806744642555714,
+ "learning_rate": 0.0006,
+ "loss": 4.057899475097656,
+ "step": 2445
+ },
+ {
+ "epoch": 33.97816593886463,
+ "grad_norm": 0.015734272077679634,
+ "learning_rate": 0.0006,
+ "loss": 4.163505554199219,
+ "step": 2446
+ },
+ {
+ "epoch": 33.992139737991266,
+ "grad_norm": 0.015250146389007568,
+ "learning_rate": 0.0006,
+ "loss": 4.076294898986816,
+ "step": 2447
+ },
+ {
+ "epoch": 34.0,
+ "grad_norm": 0.016733702272176743,
+ "learning_rate": 0.0006,
+ "loss": 4.138420104980469,
+ "step": 2448
+ },
+ {
+ "epoch": 34.0,
+ "eval_loss": 4.542503356933594,
+ "eval_runtime": 56.983,
+ "eval_samples_per_second": 42.855,
+ "eval_steps_per_second": 1.351,
+ "step": 2448
+ },
+ {
+ "epoch": 34.01397379912664,
+ "grad_norm": 0.015866613015532494,
+ "learning_rate": 0.0006,
+ "loss": 4.032081127166748,
+ "step": 2449
+ },
+ {
+ "epoch": 34.02794759825328,
+ "grad_norm": 0.01946570910513401,
+ "learning_rate": 0.0006,
+ "loss": 4.122320175170898,
+ "step": 2450
+ },
+ {
+ "epoch": 34.041921397379916,
+ "grad_norm": 0.020265938714146614,
+ "learning_rate": 0.0006,
+ "loss": 4.06460428237915,
+ "step": 2451
+ },
+ {
+ "epoch": 34.05589519650655,
+ "grad_norm": 0.018887171521782875,
+ "learning_rate": 0.0006,
+ "loss": 4.096906661987305,
+ "step": 2452
+ },
+ {
+ "epoch": 34.069868995633186,
+ "grad_norm": 0.017304692417383194,
+ "learning_rate": 0.0006,
+ "loss": 4.137261390686035,
+ "step": 2453
+ },
+ {
+ "epoch": 34.083842794759825,
+ "grad_norm": 0.017029082402586937,
+ "learning_rate": 0.0006,
+ "loss": 4.081112861633301,
+ "step": 2454
+ },
+ {
+ "epoch": 34.09781659388646,
+ "grad_norm": 0.016414813697338104,
+ "learning_rate": 0.0006,
+ "loss": 4.140783309936523,
+ "step": 2455
+ },
+ {
+ "epoch": 34.1117903930131,
+ "grad_norm": 0.018238281831145287,
+ "learning_rate": 0.0006,
+ "loss": 4.161116600036621,
+ "step": 2456
+ },
+ {
+ "epoch": 34.12576419213974,
+ "grad_norm": 0.019929232075810432,
+ "learning_rate": 0.0006,
+ "loss": 4.243722438812256,
+ "step": 2457
+ },
+ {
+ "epoch": 34.13973799126637,
+ "grad_norm": 0.019375786185264587,
+ "learning_rate": 0.0006,
+ "loss": 4.134873390197754,
+ "step": 2458
+ },
+ {
+ "epoch": 34.15371179039301,
+ "grad_norm": 0.018151137977838516,
+ "learning_rate": 0.0006,
+ "loss": 4.098745346069336,
+ "step": 2459
+ },
+ {
+ "epoch": 34.16768558951965,
+ "grad_norm": 0.017054934054613113,
+ "learning_rate": 0.0006,
+ "loss": 4.194662570953369,
+ "step": 2460
+ },
+ {
+ "epoch": 34.18165938864629,
+ "grad_norm": 0.018220555037260056,
+ "learning_rate": 0.0006,
+ "loss": 4.128866195678711,
+ "step": 2461
+ },
+ {
+ "epoch": 34.19563318777293,
+ "grad_norm": 0.01692879945039749,
+ "learning_rate": 0.0006,
+ "loss": 4.062477111816406,
+ "step": 2462
+ },
+ {
+ "epoch": 34.209606986899566,
+ "grad_norm": 0.01674199476838112,
+ "learning_rate": 0.0006,
+ "loss": 4.186149597167969,
+ "step": 2463
+ },
+ {
+ "epoch": 34.223580786026204,
+ "grad_norm": 0.016527455300092697,
+ "learning_rate": 0.0006,
+ "loss": 4.0734028816223145,
+ "step": 2464
+ },
+ {
+ "epoch": 34.237554585152836,
+ "grad_norm": 0.016448386013507843,
+ "learning_rate": 0.0006,
+ "loss": 4.098896026611328,
+ "step": 2465
+ },
+ {
+ "epoch": 34.251528384279474,
+ "grad_norm": 0.014644953422248363,
+ "learning_rate": 0.0006,
+ "loss": 4.1031904220581055,
+ "step": 2466
+ },
+ {
+ "epoch": 34.26550218340611,
+ "grad_norm": 0.016649341210722923,
+ "learning_rate": 0.0006,
+ "loss": 4.148791313171387,
+ "step": 2467
+ },
+ {
+ "epoch": 34.27947598253275,
+ "grad_norm": 0.017442714422941208,
+ "learning_rate": 0.0006,
+ "loss": 4.070799827575684,
+ "step": 2468
+ },
+ {
+ "epoch": 34.29344978165939,
+ "grad_norm": 0.017852013930678368,
+ "learning_rate": 0.0006,
+ "loss": 4.031924247741699,
+ "step": 2469
+ },
+ {
+ "epoch": 34.30742358078603,
+ "grad_norm": 0.016371270641684532,
+ "learning_rate": 0.0006,
+ "loss": 4.149420738220215,
+ "step": 2470
+ },
+ {
+ "epoch": 34.32139737991266,
+ "grad_norm": 0.015846814960241318,
+ "learning_rate": 0.0006,
+ "loss": 4.15427303314209,
+ "step": 2471
+ },
+ {
+ "epoch": 34.3353711790393,
+ "grad_norm": 0.017411163076758385,
+ "learning_rate": 0.0006,
+ "loss": 4.1304192543029785,
+ "step": 2472
+ },
+ {
+ "epoch": 34.34934497816594,
+ "grad_norm": 0.01780046708881855,
+ "learning_rate": 0.0006,
+ "loss": 4.191732406616211,
+ "step": 2473
+ },
+ {
+ "epoch": 34.36331877729258,
+ "grad_norm": 0.01958383060991764,
+ "learning_rate": 0.0006,
+ "loss": 4.068355083465576,
+ "step": 2474
+ },
+ {
+ "epoch": 34.377292576419215,
+ "grad_norm": 0.022173317149281502,
+ "learning_rate": 0.0006,
+ "loss": 4.182929039001465,
+ "step": 2475
+ },
+ {
+ "epoch": 34.391266375545854,
+ "grad_norm": 0.022389156743884087,
+ "learning_rate": 0.0006,
+ "loss": 4.110218048095703,
+ "step": 2476
+ },
+ {
+ "epoch": 34.40524017467249,
+ "grad_norm": 0.022478511556982994,
+ "learning_rate": 0.0006,
+ "loss": 4.095479488372803,
+ "step": 2477
+ },
+ {
+ "epoch": 34.419213973799124,
+ "grad_norm": 0.019504647701978683,
+ "learning_rate": 0.0006,
+ "loss": 4.101964950561523,
+ "step": 2478
+ },
+ {
+ "epoch": 34.43318777292576,
+ "grad_norm": 0.019378576427698135,
+ "learning_rate": 0.0006,
+ "loss": 4.14620304107666,
+ "step": 2479
+ },
+ {
+ "epoch": 34.4471615720524,
+ "grad_norm": 0.017993204295635223,
+ "learning_rate": 0.0006,
+ "loss": 4.183492660522461,
+ "step": 2480
+ },
+ {
+ "epoch": 34.46113537117904,
+ "grad_norm": 0.01814805157482624,
+ "learning_rate": 0.0006,
+ "loss": 4.151247978210449,
+ "step": 2481
+ },
+ {
+ "epoch": 34.47510917030568,
+ "grad_norm": 0.018622448667883873,
+ "learning_rate": 0.0006,
+ "loss": 4.020938396453857,
+ "step": 2482
+ },
+ {
+ "epoch": 34.48908296943232,
+ "grad_norm": 0.018360132351517677,
+ "learning_rate": 0.0006,
+ "loss": 4.064734935760498,
+ "step": 2483
+ },
+ {
+ "epoch": 34.50305676855895,
+ "grad_norm": 0.017715711146593094,
+ "learning_rate": 0.0006,
+ "loss": 4.185687065124512,
+ "step": 2484
+ },
+ {
+ "epoch": 34.51703056768559,
+ "grad_norm": 0.01714295893907547,
+ "learning_rate": 0.0006,
+ "loss": 4.1044721603393555,
+ "step": 2485
+ },
+ {
+ "epoch": 34.531004366812226,
+ "grad_norm": 0.016943803057074547,
+ "learning_rate": 0.0006,
+ "loss": 4.106837272644043,
+ "step": 2486
+ },
+ {
+ "epoch": 34.544978165938865,
+ "grad_norm": 0.0182423684746027,
+ "learning_rate": 0.0006,
+ "loss": 4.1041340827941895,
+ "step": 2487
+ },
+ {
+ "epoch": 34.5589519650655,
+ "grad_norm": 0.01636587455868721,
+ "learning_rate": 0.0006,
+ "loss": 3.9735546112060547,
+ "step": 2488
+ },
+ {
+ "epoch": 34.57292576419214,
+ "grad_norm": 0.014593546278774738,
+ "learning_rate": 0.0006,
+ "loss": 4.041980266571045,
+ "step": 2489
+ },
+ {
+ "epoch": 34.58689956331878,
+ "grad_norm": 0.015436965972185135,
+ "learning_rate": 0.0006,
+ "loss": 4.301148414611816,
+ "step": 2490
+ },
+ {
+ "epoch": 34.60087336244541,
+ "grad_norm": 0.015075269155204296,
+ "learning_rate": 0.0006,
+ "loss": 4.118527412414551,
+ "step": 2491
+ },
+ {
+ "epoch": 34.61484716157205,
+ "grad_norm": 0.01701514609158039,
+ "learning_rate": 0.0006,
+ "loss": 4.245968818664551,
+ "step": 2492
+ },
+ {
+ "epoch": 34.62882096069869,
+ "grad_norm": 0.018047887831926346,
+ "learning_rate": 0.0006,
+ "loss": 4.227261066436768,
+ "step": 2493
+ },
+ {
+ "epoch": 34.64279475982533,
+ "grad_norm": 0.018499085679650307,
+ "learning_rate": 0.0006,
+ "loss": 4.121360778808594,
+ "step": 2494
+ },
+ {
+ "epoch": 34.65676855895197,
+ "grad_norm": 0.018987711519002914,
+ "learning_rate": 0.0006,
+ "loss": 4.131564617156982,
+ "step": 2495
+ },
+ {
+ "epoch": 34.670742358078606,
+ "grad_norm": 0.01834476925432682,
+ "learning_rate": 0.0006,
+ "loss": 4.224941253662109,
+ "step": 2496
+ },
+ {
+ "epoch": 34.68471615720524,
+ "grad_norm": 0.017507942393422127,
+ "learning_rate": 0.0006,
+ "loss": 4.1394548416137695,
+ "step": 2497
+ },
+ {
+ "epoch": 34.698689956331876,
+ "grad_norm": 0.01680580899119377,
+ "learning_rate": 0.0006,
+ "loss": 4.1073431968688965,
+ "step": 2498
+ },
+ {
+ "epoch": 34.712663755458514,
+ "grad_norm": 0.015233514830470085,
+ "learning_rate": 0.0006,
+ "loss": 4.136994361877441,
+ "step": 2499
+ },
+ {
+ "epoch": 34.72663755458515,
+ "grad_norm": 0.015482917428016663,
+ "learning_rate": 0.0006,
+ "loss": 4.110719680786133,
+ "step": 2500
+ },
+ {
+ "epoch": 34.74061135371179,
+ "grad_norm": 0.015940461307764053,
+ "learning_rate": 0.0006,
+ "loss": 4.207851886749268,
+ "step": 2501
+ },
+ {
+ "epoch": 34.75458515283843,
+ "grad_norm": 0.015355597250163555,
+ "learning_rate": 0.0006,
+ "loss": 4.169469833374023,
+ "step": 2502
+ },
+ {
+ "epoch": 34.76855895196506,
+ "grad_norm": 0.015974000096321106,
+ "learning_rate": 0.0006,
+ "loss": 4.110061168670654,
+ "step": 2503
+ },
+ {
+ "epoch": 34.7825327510917,
+ "grad_norm": 0.016088927164673805,
+ "learning_rate": 0.0006,
+ "loss": 4.155452728271484,
+ "step": 2504
+ },
+ {
+ "epoch": 34.79650655021834,
+ "grad_norm": 0.016156120225787163,
+ "learning_rate": 0.0006,
+ "loss": 4.178092956542969,
+ "step": 2505
+ },
+ {
+ "epoch": 34.81048034934498,
+ "grad_norm": 0.014928647316992283,
+ "learning_rate": 0.0006,
+ "loss": 4.05457878112793,
+ "step": 2506
+ },
+ {
+ "epoch": 34.82445414847162,
+ "grad_norm": 0.015289328061044216,
+ "learning_rate": 0.0006,
+ "loss": 4.18869686126709,
+ "step": 2507
+ },
+ {
+ "epoch": 34.838427947598255,
+ "grad_norm": 0.015799039974808693,
+ "learning_rate": 0.0006,
+ "loss": 4.106073379516602,
+ "step": 2508
+ },
+ {
+ "epoch": 34.852401746724894,
+ "grad_norm": 0.015824418514966965,
+ "learning_rate": 0.0006,
+ "loss": 4.0634589195251465,
+ "step": 2509
+ },
+ {
+ "epoch": 34.866375545851525,
+ "grad_norm": 0.015321656130254269,
+ "learning_rate": 0.0006,
+ "loss": 4.093090057373047,
+ "step": 2510
+ },
+ {
+ "epoch": 34.880349344978164,
+ "grad_norm": 0.014380166307091713,
+ "learning_rate": 0.0006,
+ "loss": 4.1750640869140625,
+ "step": 2511
+ },
+ {
+ "epoch": 34.8943231441048,
+ "grad_norm": 0.014962567947804928,
+ "learning_rate": 0.0006,
+ "loss": 4.185277938842773,
+ "step": 2512
+ },
+ {
+ "epoch": 34.90829694323144,
+ "grad_norm": 0.014467543922364712,
+ "learning_rate": 0.0006,
+ "loss": 4.1600799560546875,
+ "step": 2513
+ },
+ {
+ "epoch": 34.92227074235808,
+ "grad_norm": 0.01545181218534708,
+ "learning_rate": 0.0006,
+ "loss": 4.096299648284912,
+ "step": 2514
+ },
+ {
+ "epoch": 34.93624454148472,
+ "grad_norm": 0.01605340465903282,
+ "learning_rate": 0.0006,
+ "loss": 4.203323841094971,
+ "step": 2515
+ },
+ {
+ "epoch": 34.95021834061135,
+ "grad_norm": 0.015203815884888172,
+ "learning_rate": 0.0006,
+ "loss": 4.201385021209717,
+ "step": 2516
+ },
+ {
+ "epoch": 34.96419213973799,
+ "grad_norm": 0.015149409882724285,
+ "learning_rate": 0.0006,
+ "loss": 4.088164329528809,
+ "step": 2517
+ },
+ {
+ "epoch": 34.97816593886463,
+ "grad_norm": 0.014885558746755123,
+ "learning_rate": 0.0006,
+ "loss": 4.073911666870117,
+ "step": 2518
+ },
+ {
+ "epoch": 34.992139737991266,
+ "grad_norm": 0.015003419481217861,
+ "learning_rate": 0.0006,
+ "loss": 4.157113552093506,
+ "step": 2519
+ },
+ {
+ "epoch": 35.0,
+ "grad_norm": 0.01641916297376156,
+ "learning_rate": 0.0006,
+ "loss": 3.970052719116211,
+ "step": 2520
+ },
+ {
+ "epoch": 35.0,
+ "eval_loss": 4.641692161560059,
+ "eval_runtime": 56.8836,
+ "eval_samples_per_second": 42.93,
+ "eval_steps_per_second": 1.354,
+ "step": 2520
+ },
+ {
+ "epoch": 35.01397379912664,
+ "grad_norm": 0.015894673764705658,
+ "learning_rate": 0.0006,
+ "loss": 4.104437351226807,
+ "step": 2521
+ },
+ {
+ "epoch": 35.02794759825328,
+ "grad_norm": 0.0173867866396904,
+ "learning_rate": 0.0006,
+ "loss": 4.091726303100586,
+ "step": 2522
+ },
+ {
+ "epoch": 35.041921397379916,
+ "grad_norm": 0.019272832199931145,
+ "learning_rate": 0.0006,
+ "loss": 4.217905044555664,
+ "step": 2523
+ },
+ {
+ "epoch": 35.05589519650655,
+ "grad_norm": 0.021820371970534325,
+ "learning_rate": 0.0006,
+ "loss": 4.111428260803223,
+ "step": 2524
+ },
+ {
+ "epoch": 35.069868995633186,
+ "grad_norm": 0.02340741828083992,
+ "learning_rate": 0.0006,
+ "loss": 4.115771293640137,
+ "step": 2525
+ },
+ {
+ "epoch": 35.083842794759825,
+ "grad_norm": 0.023760691285133362,
+ "learning_rate": 0.0006,
+ "loss": 4.132383346557617,
+ "step": 2526
+ },
+ {
+ "epoch": 35.09781659388646,
+ "grad_norm": 0.02073468267917633,
+ "learning_rate": 0.0006,
+ "loss": 4.272691249847412,
+ "step": 2527
+ },
+ {
+ "epoch": 35.1117903930131,
+ "grad_norm": 0.02197718247771263,
+ "learning_rate": 0.0006,
+ "loss": 4.105586528778076,
+ "step": 2528
+ },
+ {
+ "epoch": 35.12576419213974,
+ "grad_norm": 0.0200254675000906,
+ "learning_rate": 0.0006,
+ "loss": 3.968820095062256,
+ "step": 2529
+ },
+ {
+ "epoch": 35.13973799126637,
+ "grad_norm": 0.01929587684571743,
+ "learning_rate": 0.0006,
+ "loss": 4.045468807220459,
+ "step": 2530
+ },
+ {
+ "epoch": 35.15371179039301,
+ "grad_norm": 0.01886162906885147,
+ "learning_rate": 0.0006,
+ "loss": 4.185571670532227,
+ "step": 2531
+ },
+ {
+ "epoch": 35.16768558951965,
+ "grad_norm": 0.019789839163422585,
+ "learning_rate": 0.0006,
+ "loss": 3.9734551906585693,
+ "step": 2532
+ },
+ {
+ "epoch": 35.18165938864629,
+ "grad_norm": 0.0198142658919096,
+ "learning_rate": 0.0006,
+ "loss": 4.2662506103515625,
+ "step": 2533
+ },
+ {
+ "epoch": 35.19563318777293,
+ "grad_norm": 0.02087012305855751,
+ "learning_rate": 0.0006,
+ "loss": 4.175931930541992,
+ "step": 2534
+ },
+ {
+ "epoch": 35.209606986899566,
+ "grad_norm": 0.020294535905122757,
+ "learning_rate": 0.0006,
+ "loss": 4.060324668884277,
+ "step": 2535
+ },
+ {
+ "epoch": 35.223580786026204,
+ "grad_norm": 0.01908983662724495,
+ "learning_rate": 0.0006,
+ "loss": 4.217939853668213,
+ "step": 2536
+ },
+ {
+ "epoch": 35.237554585152836,
+ "grad_norm": 0.0178892333060503,
+ "learning_rate": 0.0006,
+ "loss": 4.190216541290283,
+ "step": 2537
+ },
+ {
+ "epoch": 35.251528384279474,
+ "grad_norm": 0.017892315983772278,
+ "learning_rate": 0.0006,
+ "loss": 4.067419052124023,
+ "step": 2538
+ },
+ {
+ "epoch": 35.26550218340611,
+ "grad_norm": 0.018556250259280205,
+ "learning_rate": 0.0006,
+ "loss": 4.105958461761475,
+ "step": 2539
+ },
+ {
+ "epoch": 35.27947598253275,
+ "grad_norm": 0.01944422349333763,
+ "learning_rate": 0.0006,
+ "loss": 4.155775547027588,
+ "step": 2540
+ },
+ {
+ "epoch": 35.29344978165939,
+ "grad_norm": 0.017333146184682846,
+ "learning_rate": 0.0006,
+ "loss": 4.244513511657715,
+ "step": 2541
+ },
+ {
+ "epoch": 35.30742358078603,
+ "grad_norm": 0.01639244332909584,
+ "learning_rate": 0.0006,
+ "loss": 4.167815685272217,
+ "step": 2542
+ },
+ {
+ "epoch": 35.32139737991266,
+ "grad_norm": 0.01610143668949604,
+ "learning_rate": 0.0006,
+ "loss": 4.121486663818359,
+ "step": 2543
+ },
+ {
+ "epoch": 35.3353711790393,
+ "grad_norm": 0.01656411588191986,
+ "learning_rate": 0.0006,
+ "loss": 4.1755499839782715,
+ "step": 2544
+ },
+ {
+ "epoch": 35.34934497816594,
+ "grad_norm": 0.015075297094881535,
+ "learning_rate": 0.0006,
+ "loss": 4.102992534637451,
+ "step": 2545
+ },
+ {
+ "epoch": 35.36331877729258,
+ "grad_norm": 0.01538170501589775,
+ "learning_rate": 0.0006,
+ "loss": 4.161603927612305,
+ "step": 2546
+ },
+ {
+ "epoch": 35.377292576419215,
+ "grad_norm": 0.01483312901109457,
+ "learning_rate": 0.0006,
+ "loss": 4.085355758666992,
+ "step": 2547
+ },
+ {
+ "epoch": 35.391266375545854,
+ "grad_norm": 0.014894738793373108,
+ "learning_rate": 0.0006,
+ "loss": 4.083585262298584,
+ "step": 2548
+ },
+ {
+ "epoch": 35.40524017467249,
+ "grad_norm": 0.01475780550390482,
+ "learning_rate": 0.0006,
+ "loss": 4.012645721435547,
+ "step": 2549
+ },
+ {
+ "epoch": 35.419213973799124,
+ "grad_norm": 0.014531653374433517,
+ "learning_rate": 0.0006,
+ "loss": 4.197113990783691,
+ "step": 2550
+ },
+ {
+ "epoch": 35.43318777292576,
+ "grad_norm": 0.015316493809223175,
+ "learning_rate": 0.0006,
+ "loss": 4.227970123291016,
+ "step": 2551
+ },
+ {
+ "epoch": 35.4471615720524,
+ "grad_norm": 0.014685571193695068,
+ "learning_rate": 0.0006,
+ "loss": 4.089415550231934,
+ "step": 2552
+ },
+ {
+ "epoch": 35.46113537117904,
+ "grad_norm": 0.015337035991251469,
+ "learning_rate": 0.0006,
+ "loss": 4.080583572387695,
+ "step": 2553
+ },
+ {
+ "epoch": 35.47510917030568,
+ "grad_norm": 0.016754208132624626,
+ "learning_rate": 0.0006,
+ "loss": 4.163935661315918,
+ "step": 2554
+ },
+ {
+ "epoch": 35.48908296943232,
+ "grad_norm": 0.017535798251628876,
+ "learning_rate": 0.0006,
+ "loss": 4.093027591705322,
+ "step": 2555
+ },
+ {
+ "epoch": 35.50305676855895,
+ "grad_norm": 0.017336975783109665,
+ "learning_rate": 0.0006,
+ "loss": 4.127961158752441,
+ "step": 2556
+ },
+ {
+ "epoch": 35.51703056768559,
+ "grad_norm": 0.016242152079939842,
+ "learning_rate": 0.0006,
+ "loss": 4.155593395233154,
+ "step": 2557
+ },
+ {
+ "epoch": 35.531004366812226,
+ "grad_norm": 0.015121322125196457,
+ "learning_rate": 0.0006,
+ "loss": 3.984004020690918,
+ "step": 2558
+ },
+ {
+ "epoch": 35.544978165938865,
+ "grad_norm": 0.015178884379565716,
+ "learning_rate": 0.0006,
+ "loss": 4.111564636230469,
+ "step": 2559
+ },
+ {
+ "epoch": 35.5589519650655,
+ "grad_norm": 0.014777131378650665,
+ "learning_rate": 0.0006,
+ "loss": 4.041402816772461,
+ "step": 2560
+ },
+ {
+ "epoch": 35.57292576419214,
+ "grad_norm": 0.015197676606476307,
+ "learning_rate": 0.0006,
+ "loss": 4.133706092834473,
+ "step": 2561
+ },
+ {
+ "epoch": 35.58689956331878,
+ "grad_norm": 0.015905972570180893,
+ "learning_rate": 0.0006,
+ "loss": 4.1069135665893555,
+ "step": 2562
+ },
+ {
+ "epoch": 35.60087336244541,
+ "grad_norm": 0.015942877158522606,
+ "learning_rate": 0.0006,
+ "loss": 4.0720624923706055,
+ "step": 2563
+ },
+ {
+ "epoch": 35.61484716157205,
+ "grad_norm": 0.016697583720088005,
+ "learning_rate": 0.0006,
+ "loss": 4.167448043823242,
+ "step": 2564
+ },
+ {
+ "epoch": 35.62882096069869,
+ "grad_norm": 0.015603577718138695,
+ "learning_rate": 0.0006,
+ "loss": 4.127327919006348,
+ "step": 2565
+ },
+ {
+ "epoch": 35.64279475982533,
+ "grad_norm": 0.017153460532426834,
+ "learning_rate": 0.0006,
+ "loss": 4.056543350219727,
+ "step": 2566
+ },
+ {
+ "epoch": 35.65676855895197,
+ "grad_norm": 0.016731958836317062,
+ "learning_rate": 0.0006,
+ "loss": 4.0681843757629395,
+ "step": 2567
+ },
+ {
+ "epoch": 35.670742358078606,
+ "grad_norm": 0.016167359426617622,
+ "learning_rate": 0.0006,
+ "loss": 4.174992084503174,
+ "step": 2568
+ },
+ {
+ "epoch": 35.68471615720524,
+ "grad_norm": 0.01615108922123909,
+ "learning_rate": 0.0006,
+ "loss": 4.1383161544799805,
+ "step": 2569
+ },
+ {
+ "epoch": 35.698689956331876,
+ "grad_norm": 0.015750816091895103,
+ "learning_rate": 0.0006,
+ "loss": 4.031595706939697,
+ "step": 2570
+ },
+ {
+ "epoch": 35.712663755458514,
+ "grad_norm": 0.015416189096868038,
+ "learning_rate": 0.0006,
+ "loss": 4.067498207092285,
+ "step": 2571
+ },
+ {
+ "epoch": 35.72663755458515,
+ "grad_norm": 0.01570792682468891,
+ "learning_rate": 0.0006,
+ "loss": 4.1338300704956055,
+ "step": 2572
+ },
+ {
+ "epoch": 35.74061135371179,
+ "grad_norm": 0.016367752104997635,
+ "learning_rate": 0.0006,
+ "loss": 4.152523040771484,
+ "step": 2573
+ },
+ {
+ "epoch": 35.75458515283843,
+ "grad_norm": 0.016349803656339645,
+ "learning_rate": 0.0006,
+ "loss": 4.101985454559326,
+ "step": 2574
+ },
+ {
+ "epoch": 35.76855895196506,
+ "grad_norm": 0.01663660630583763,
+ "learning_rate": 0.0006,
+ "loss": 4.190130710601807,
+ "step": 2575
+ },
+ {
+ "epoch": 35.7825327510917,
+ "grad_norm": 0.015599170699715614,
+ "learning_rate": 0.0006,
+ "loss": 4.0792436599731445,
+ "step": 2576
+ },
+ {
+ "epoch": 35.79650655021834,
+ "grad_norm": 0.015472427010536194,
+ "learning_rate": 0.0006,
+ "loss": 4.186446666717529,
+ "step": 2577
+ },
+ {
+ "epoch": 35.81048034934498,
+ "grad_norm": 0.015520439483225346,
+ "learning_rate": 0.0006,
+ "loss": 4.246635437011719,
+ "step": 2578
+ },
+ {
+ "epoch": 35.82445414847162,
+ "grad_norm": 0.016998281702399254,
+ "learning_rate": 0.0006,
+ "loss": 4.13435173034668,
+ "step": 2579
+ },
+ {
+ "epoch": 35.838427947598255,
+ "grad_norm": 0.016290361061692238,
+ "learning_rate": 0.0006,
+ "loss": 4.107428073883057,
+ "step": 2580
+ },
+ {
+ "epoch": 35.852401746724894,
+ "grad_norm": 0.0165516659617424,
+ "learning_rate": 0.0006,
+ "loss": 4.1612138748168945,
+ "step": 2581
+ },
+ {
+ "epoch": 35.866375545851525,
+ "grad_norm": 0.01736091636121273,
+ "learning_rate": 0.0006,
+ "loss": 4.1072869300842285,
+ "step": 2582
+ },
+ {
+ "epoch": 35.880349344978164,
+ "grad_norm": 0.01787031814455986,
+ "learning_rate": 0.0006,
+ "loss": 4.1076812744140625,
+ "step": 2583
+ },
+ {
+ "epoch": 35.8943231441048,
+ "grad_norm": 0.015437985770404339,
+ "learning_rate": 0.0006,
+ "loss": 4.097488880157471,
+ "step": 2584
+ },
+ {
+ "epoch": 35.90829694323144,
+ "grad_norm": 0.016306741163134575,
+ "learning_rate": 0.0006,
+ "loss": 4.105017185211182,
+ "step": 2585
+ },
+ {
+ "epoch": 35.92227074235808,
+ "grad_norm": 0.015723245218396187,
+ "learning_rate": 0.0006,
+ "loss": 4.205283164978027,
+ "step": 2586
+ },
+ {
+ "epoch": 35.93624454148472,
+ "grad_norm": 0.014626456424593925,
+ "learning_rate": 0.0006,
+ "loss": 4.086552619934082,
+ "step": 2587
+ },
+ {
+ "epoch": 35.95021834061135,
+ "grad_norm": 0.014348454773426056,
+ "learning_rate": 0.0006,
+ "loss": 4.171329975128174,
+ "step": 2588
+ },
+ {
+ "epoch": 35.96419213973799,
+ "grad_norm": 0.015598480589687824,
+ "learning_rate": 0.0006,
+ "loss": 4.140815258026123,
+ "step": 2589
+ },
+ {
+ "epoch": 35.97816593886463,
+ "grad_norm": 0.017126062884926796,
+ "learning_rate": 0.0006,
+ "loss": 4.157176971435547,
+ "step": 2590
+ },
+ {
+ "epoch": 35.992139737991266,
+ "grad_norm": 0.01860199309885502,
+ "learning_rate": 0.0006,
+ "loss": 4.252880096435547,
+ "step": 2591
+ },
+ {
+ "epoch": 36.0,
+ "grad_norm": 0.021780220791697502,
+ "learning_rate": 0.0006,
+ "loss": 4.249861717224121,
+ "step": 2592
+ },
+ {
+ "epoch": 36.0,
+ "eval_loss": 4.606466293334961,
+ "eval_runtime": 56.6217,
+ "eval_samples_per_second": 43.128,
+ "eval_steps_per_second": 1.36,
+ "step": 2592
+ },
+ {
+ "epoch": 36.01397379912664,
+ "grad_norm": 0.020892828702926636,
+ "learning_rate": 0.0006,
+ "loss": 4.102055549621582,
+ "step": 2593
+ },
+ {
+ "epoch": 36.02794759825328,
+ "grad_norm": 0.020306255668401718,
+ "learning_rate": 0.0006,
+ "loss": 4.184308052062988,
+ "step": 2594
+ },
+ {
+ "epoch": 36.041921397379916,
+ "grad_norm": 0.02025097794830799,
+ "learning_rate": 0.0006,
+ "loss": 3.9691858291625977,
+ "step": 2595
+ },
+ {
+ "epoch": 36.05589519650655,
+ "grad_norm": 0.020254382863640785,
+ "learning_rate": 0.0006,
+ "loss": 4.02957820892334,
+ "step": 2596
+ },
+ {
+ "epoch": 36.069868995633186,
+ "grad_norm": 0.019307196140289307,
+ "learning_rate": 0.0006,
+ "loss": 4.033181667327881,
+ "step": 2597
+ },
+ {
+ "epoch": 36.083842794759825,
+ "grad_norm": 0.0198476854711771,
+ "learning_rate": 0.0006,
+ "loss": 4.0816874504089355,
+ "step": 2598
+ },
+ {
+ "epoch": 36.09781659388646,
+ "grad_norm": 0.019916515797376633,
+ "learning_rate": 0.0006,
+ "loss": 4.164889812469482,
+ "step": 2599
+ },
+ {
+ "epoch": 36.1117903930131,
+ "grad_norm": 0.0189144778996706,
+ "learning_rate": 0.0006,
+ "loss": 4.038427352905273,
+ "step": 2600
+ },
+ {
+ "epoch": 36.12576419213974,
+ "grad_norm": 0.01754768192768097,
+ "learning_rate": 0.0006,
+ "loss": 4.010427474975586,
+ "step": 2601
+ },
+ {
+ "epoch": 36.13973799126637,
+ "grad_norm": 0.016862990334630013,
+ "learning_rate": 0.0006,
+ "loss": 4.035970211029053,
+ "step": 2602
+ },
+ {
+ "epoch": 36.15371179039301,
+ "grad_norm": 0.019616752862930298,
+ "learning_rate": 0.0006,
+ "loss": 4.079898834228516,
+ "step": 2603
+ },
+ {
+ "epoch": 36.16768558951965,
+ "grad_norm": 0.020131465047597885,
+ "learning_rate": 0.0006,
+ "loss": 4.144684791564941,
+ "step": 2604
+ },
+ {
+ "epoch": 36.18165938864629,
+ "grad_norm": 0.02072521299123764,
+ "learning_rate": 0.0006,
+ "loss": 4.013660430908203,
+ "step": 2605
+ },
+ {
+ "epoch": 36.19563318777293,
+ "grad_norm": 0.01924615167081356,
+ "learning_rate": 0.0006,
+ "loss": 4.071398735046387,
+ "step": 2606
+ },
+ {
+ "epoch": 36.209606986899566,
+ "grad_norm": 0.02013954520225525,
+ "learning_rate": 0.0006,
+ "loss": 4.127094268798828,
+ "step": 2607
+ },
+ {
+ "epoch": 36.223580786026204,
+ "grad_norm": 0.02042008936405182,
+ "learning_rate": 0.0006,
+ "loss": 4.154261589050293,
+ "step": 2608
+ },
+ {
+ "epoch": 36.237554585152836,
+ "grad_norm": 0.021998876705765724,
+ "learning_rate": 0.0006,
+ "loss": 4.204150199890137,
+ "step": 2609
+ },
+ {
+ "epoch": 36.251528384279474,
+ "grad_norm": 0.022946683689951897,
+ "learning_rate": 0.0006,
+ "loss": 4.013632774353027,
+ "step": 2610
+ },
+ {
+ "epoch": 36.26550218340611,
+ "grad_norm": 0.02061830274760723,
+ "learning_rate": 0.0006,
+ "loss": 4.171084403991699,
+ "step": 2611
+ },
+ {
+ "epoch": 36.27947598253275,
+ "grad_norm": 0.01883152686059475,
+ "learning_rate": 0.0006,
+ "loss": 4.149848461151123,
+ "step": 2612
+ },
+ {
+ "epoch": 36.29344978165939,
+ "grad_norm": 0.019945282489061356,
+ "learning_rate": 0.0006,
+ "loss": 4.174656867980957,
+ "step": 2613
+ },
+ {
+ "epoch": 36.30742358078603,
+ "grad_norm": 0.020751051604747772,
+ "learning_rate": 0.0006,
+ "loss": 4.080747127532959,
+ "step": 2614
+ },
+ {
+ "epoch": 36.32139737991266,
+ "grad_norm": 0.01879691146314144,
+ "learning_rate": 0.0006,
+ "loss": 4.057356357574463,
+ "step": 2615
+ },
+ {
+ "epoch": 36.3353711790393,
+ "grad_norm": 0.019037071615457535,
+ "learning_rate": 0.0006,
+ "loss": 4.204607009887695,
+ "step": 2616
+ },
+ {
+ "epoch": 36.34934497816594,
+ "grad_norm": 0.01872078701853752,
+ "learning_rate": 0.0006,
+ "loss": 4.050586700439453,
+ "step": 2617
+ },
+ {
+ "epoch": 36.36331877729258,
+ "grad_norm": 0.018092811107635498,
+ "learning_rate": 0.0006,
+ "loss": 4.102832794189453,
+ "step": 2618
+ },
+ {
+ "epoch": 36.377292576419215,
+ "grad_norm": 0.017010487616062164,
+ "learning_rate": 0.0006,
+ "loss": 4.191788673400879,
+ "step": 2619
+ },
+ {
+ "epoch": 36.391266375545854,
+ "grad_norm": 0.01691143773496151,
+ "learning_rate": 0.0006,
+ "loss": 4.160456657409668,
+ "step": 2620
+ },
+ {
+ "epoch": 36.40524017467249,
+ "grad_norm": 0.017046816647052765,
+ "learning_rate": 0.0006,
+ "loss": 4.073828220367432,
+ "step": 2621
+ },
+ {
+ "epoch": 36.419213973799124,
+ "grad_norm": 0.01707981713116169,
+ "learning_rate": 0.0006,
+ "loss": 4.0980095863342285,
+ "step": 2622
+ },
+ {
+ "epoch": 36.43318777292576,
+ "grad_norm": 0.01630452275276184,
+ "learning_rate": 0.0006,
+ "loss": 4.101321220397949,
+ "step": 2623
+ },
+ {
+ "epoch": 36.4471615720524,
+ "grad_norm": 0.01576915942132473,
+ "learning_rate": 0.0006,
+ "loss": 4.158249855041504,
+ "step": 2624
+ },
+ {
+ "epoch": 36.46113537117904,
+ "grad_norm": 0.016179712489247322,
+ "learning_rate": 0.0006,
+ "loss": 4.077852249145508,
+ "step": 2625
+ },
+ {
+ "epoch": 36.47510917030568,
+ "grad_norm": 0.01609169878065586,
+ "learning_rate": 0.0006,
+ "loss": 4.187317848205566,
+ "step": 2626
+ },
+ {
+ "epoch": 36.48908296943232,
+ "grad_norm": 0.017148254439234734,
+ "learning_rate": 0.0006,
+ "loss": 4.109071731567383,
+ "step": 2627
+ },
+ {
+ "epoch": 36.50305676855895,
+ "grad_norm": 0.015842624008655548,
+ "learning_rate": 0.0006,
+ "loss": 4.145096302032471,
+ "step": 2628
+ },
+ {
+ "epoch": 36.51703056768559,
+ "grad_norm": 0.014906194061040878,
+ "learning_rate": 0.0006,
+ "loss": 4.028440475463867,
+ "step": 2629
+ },
+ {
+ "epoch": 36.531004366812226,
+ "grad_norm": 0.01485038734972477,
+ "learning_rate": 0.0006,
+ "loss": 4.175201416015625,
+ "step": 2630
+ },
+ {
+ "epoch": 36.544978165938865,
+ "grad_norm": 0.015834150835871696,
+ "learning_rate": 0.0006,
+ "loss": 4.076434135437012,
+ "step": 2631
+ },
+ {
+ "epoch": 36.5589519650655,
+ "grad_norm": 0.015411358326673508,
+ "learning_rate": 0.0006,
+ "loss": 4.03912878036499,
+ "step": 2632
+ },
+ {
+ "epoch": 36.57292576419214,
+ "grad_norm": 0.014762846753001213,
+ "learning_rate": 0.0006,
+ "loss": 4.019827365875244,
+ "step": 2633
+ },
+ {
+ "epoch": 36.58689956331878,
+ "grad_norm": 0.015269497409462929,
+ "learning_rate": 0.0006,
+ "loss": 4.116957187652588,
+ "step": 2634
+ },
+ {
+ "epoch": 36.60087336244541,
+ "grad_norm": 0.015171617269515991,
+ "learning_rate": 0.0006,
+ "loss": 4.2069926261901855,
+ "step": 2635
+ },
+ {
+ "epoch": 36.61484716157205,
+ "grad_norm": 0.014770902693271637,
+ "learning_rate": 0.0006,
+ "loss": 4.092021942138672,
+ "step": 2636
+ },
+ {
+ "epoch": 36.62882096069869,
+ "grad_norm": 0.014957531355321407,
+ "learning_rate": 0.0006,
+ "loss": 4.16819953918457,
+ "step": 2637
+ },
+ {
+ "epoch": 36.64279475982533,
+ "grad_norm": 0.014774451963603497,
+ "learning_rate": 0.0006,
+ "loss": 4.046073913574219,
+ "step": 2638
+ },
+ {
+ "epoch": 36.65676855895197,
+ "grad_norm": 0.014320445246994495,
+ "learning_rate": 0.0006,
+ "loss": 3.977792501449585,
+ "step": 2639
+ },
+ {
+ "epoch": 36.670742358078606,
+ "grad_norm": 0.01413130946457386,
+ "learning_rate": 0.0006,
+ "loss": 4.130649566650391,
+ "step": 2640
+ },
+ {
+ "epoch": 36.68471615720524,
+ "grad_norm": 0.014336307533085346,
+ "learning_rate": 0.0006,
+ "loss": 4.161357402801514,
+ "step": 2641
+ },
+ {
+ "epoch": 36.698689956331876,
+ "grad_norm": 0.014516931027173996,
+ "learning_rate": 0.0006,
+ "loss": 4.14248514175415,
+ "step": 2642
+ },
+ {
+ "epoch": 36.712663755458514,
+ "grad_norm": 0.014377126470208168,
+ "learning_rate": 0.0006,
+ "loss": 4.062711715698242,
+ "step": 2643
+ },
+ {
+ "epoch": 36.72663755458515,
+ "grad_norm": 0.014332668855786324,
+ "learning_rate": 0.0006,
+ "loss": 4.06423807144165,
+ "step": 2644
+ },
+ {
+ "epoch": 36.74061135371179,
+ "grad_norm": 0.014030571095645428,
+ "learning_rate": 0.0006,
+ "loss": 4.178613662719727,
+ "step": 2645
+ },
+ {
+ "epoch": 36.75458515283843,
+ "grad_norm": 0.014376776292920113,
+ "learning_rate": 0.0006,
+ "loss": 4.073130130767822,
+ "step": 2646
+ },
+ {
+ "epoch": 36.76855895196506,
+ "grad_norm": 0.013827716931700706,
+ "learning_rate": 0.0006,
+ "loss": 4.043869972229004,
+ "step": 2647
+ },
+ {
+ "epoch": 36.7825327510917,
+ "grad_norm": 0.015548762865364552,
+ "learning_rate": 0.0006,
+ "loss": 4.134049415588379,
+ "step": 2648
+ },
+ {
+ "epoch": 36.79650655021834,
+ "grad_norm": 0.017041970044374466,
+ "learning_rate": 0.0006,
+ "loss": 4.205024242401123,
+ "step": 2649
+ },
+ {
+ "epoch": 36.81048034934498,
+ "grad_norm": 0.017638275399804115,
+ "learning_rate": 0.0006,
+ "loss": 4.199924468994141,
+ "step": 2650
+ },
+ {
+ "epoch": 36.82445414847162,
+ "grad_norm": 0.017771627753973007,
+ "learning_rate": 0.0006,
+ "loss": 4.15706729888916,
+ "step": 2651
+ },
+ {
+ "epoch": 36.838427947598255,
+ "grad_norm": 0.015933562070131302,
+ "learning_rate": 0.0006,
+ "loss": 4.220742225646973,
+ "step": 2652
+ },
+ {
+ "epoch": 36.852401746724894,
+ "grad_norm": 0.01668144389986992,
+ "learning_rate": 0.0006,
+ "loss": 4.161426544189453,
+ "step": 2653
+ },
+ {
+ "epoch": 36.866375545851525,
+ "grad_norm": 0.018014680594205856,
+ "learning_rate": 0.0006,
+ "loss": 4.183046817779541,
+ "step": 2654
+ },
+ {
+ "epoch": 36.880349344978164,
+ "grad_norm": 0.018808670341968536,
+ "learning_rate": 0.0006,
+ "loss": 4.105293273925781,
+ "step": 2655
+ },
+ {
+ "epoch": 36.8943231441048,
+ "grad_norm": 0.018003476783633232,
+ "learning_rate": 0.0006,
+ "loss": 4.103171348571777,
+ "step": 2656
+ },
+ {
+ "epoch": 36.90829694323144,
+ "grad_norm": 0.016865914687514305,
+ "learning_rate": 0.0006,
+ "loss": 4.125634670257568,
+ "step": 2657
+ },
+ {
+ "epoch": 36.92227074235808,
+ "grad_norm": 0.018371153622865677,
+ "learning_rate": 0.0006,
+ "loss": 4.192461013793945,
+ "step": 2658
+ },
+ {
+ "epoch": 36.93624454148472,
+ "grad_norm": 0.019045377150177956,
+ "learning_rate": 0.0006,
+ "loss": 4.198105812072754,
+ "step": 2659
+ },
+ {
+ "epoch": 36.95021834061135,
+ "grad_norm": 0.01594168320298195,
+ "learning_rate": 0.0006,
+ "loss": 4.169788360595703,
+ "step": 2660
+ },
+ {
+ "epoch": 36.96419213973799,
+ "grad_norm": 0.015645820647478104,
+ "learning_rate": 0.0006,
+ "loss": 4.047277450561523,
+ "step": 2661
+ },
+ {
+ "epoch": 36.97816593886463,
+ "grad_norm": 0.016313156113028526,
+ "learning_rate": 0.0006,
+ "loss": 4.087392807006836,
+ "step": 2662
+ },
+ {
+ "epoch": 36.992139737991266,
+ "grad_norm": 0.017050625756382942,
+ "learning_rate": 0.0006,
+ "loss": 4.0908002853393555,
+ "step": 2663
+ },
+ {
+ "epoch": 37.0,
+ "grad_norm": 0.017730968073010445,
+ "learning_rate": 0.0006,
+ "loss": 4.110957622528076,
+ "step": 2664
+ },
+ {
+ "epoch": 37.0,
+ "eval_loss": 4.563793182373047,
+ "eval_runtime": 56.5922,
+ "eval_samples_per_second": 43.151,
+ "eval_steps_per_second": 1.361,
+ "step": 2664
+ },
+ {
+ "epoch": 37.01397379912664,
+ "grad_norm": 0.015687910839915276,
+ "learning_rate": 0.0006,
+ "loss": 4.143589019775391,
+ "step": 2665
+ },
+ {
+ "epoch": 37.02794759825328,
+ "grad_norm": 0.017451366409659386,
+ "learning_rate": 0.0006,
+ "loss": 4.0326666831970215,
+ "step": 2666
+ },
+ {
+ "epoch": 37.041921397379916,
+ "grad_norm": 0.01909538544714451,
+ "learning_rate": 0.0006,
+ "loss": 3.9706249237060547,
+ "step": 2667
+ },
+ {
+ "epoch": 37.05589519650655,
+ "grad_norm": 0.01988379657268524,
+ "learning_rate": 0.0006,
+ "loss": 3.9531009197235107,
+ "step": 2668
+ },
+ {
+ "epoch": 37.069868995633186,
+ "grad_norm": 0.02035115472972393,
+ "learning_rate": 0.0006,
+ "loss": 3.962214469909668,
+ "step": 2669
+ },
+ {
+ "epoch": 37.083842794759825,
+ "grad_norm": 0.01952001079916954,
+ "learning_rate": 0.0006,
+ "loss": 4.182262897491455,
+ "step": 2670
+ },
+ {
+ "epoch": 37.09781659388646,
+ "grad_norm": 0.01854325644671917,
+ "learning_rate": 0.0006,
+ "loss": 4.127663612365723,
+ "step": 2671
+ },
+ {
+ "epoch": 37.1117903930131,
+ "grad_norm": 0.017757616937160492,
+ "learning_rate": 0.0006,
+ "loss": 4.100804805755615,
+ "step": 2672
+ },
+ {
+ "epoch": 37.12576419213974,
+ "grad_norm": 0.016712479293346405,
+ "learning_rate": 0.0006,
+ "loss": 4.026284217834473,
+ "step": 2673
+ },
+ {
+ "epoch": 37.13973799126637,
+ "grad_norm": 0.0168139711022377,
+ "learning_rate": 0.0006,
+ "loss": 4.183882713317871,
+ "step": 2674
+ },
+ {
+ "epoch": 37.15371179039301,
+ "grad_norm": 0.017060423269867897,
+ "learning_rate": 0.0006,
+ "loss": 4.131832599639893,
+ "step": 2675
+ },
+ {
+ "epoch": 37.16768558951965,
+ "grad_norm": 0.01726631447672844,
+ "learning_rate": 0.0006,
+ "loss": 3.9156057834625244,
+ "step": 2676
+ },
+ {
+ "epoch": 37.18165938864629,
+ "grad_norm": 0.016348622739315033,
+ "learning_rate": 0.0006,
+ "loss": 4.1731367111206055,
+ "step": 2677
+ },
+ {
+ "epoch": 37.19563318777293,
+ "grad_norm": 0.016640059649944305,
+ "learning_rate": 0.0006,
+ "loss": 4.10997200012207,
+ "step": 2678
+ },
+ {
+ "epoch": 37.209606986899566,
+ "grad_norm": 0.01601342298090458,
+ "learning_rate": 0.0006,
+ "loss": 4.081145286560059,
+ "step": 2679
+ },
+ {
+ "epoch": 37.223580786026204,
+ "grad_norm": 0.01735631190240383,
+ "learning_rate": 0.0006,
+ "loss": 4.202108860015869,
+ "step": 2680
+ },
+ {
+ "epoch": 37.237554585152836,
+ "grad_norm": 0.019164685159921646,
+ "learning_rate": 0.0006,
+ "loss": 4.104109764099121,
+ "step": 2681
+ },
+ {
+ "epoch": 37.251528384279474,
+ "grad_norm": 0.01881759613752365,
+ "learning_rate": 0.0006,
+ "loss": 4.145532131195068,
+ "step": 2682
+ },
+ {
+ "epoch": 37.26550218340611,
+ "grad_norm": 0.017656555399298668,
+ "learning_rate": 0.0006,
+ "loss": 4.012782096862793,
+ "step": 2683
+ },
+ {
+ "epoch": 37.27947598253275,
+ "grad_norm": 0.01624823734164238,
+ "learning_rate": 0.0006,
+ "loss": 4.001727104187012,
+ "step": 2684
+ },
+ {
+ "epoch": 37.29344978165939,
+ "grad_norm": 0.016973139718174934,
+ "learning_rate": 0.0006,
+ "loss": 4.095399856567383,
+ "step": 2685
+ },
+ {
+ "epoch": 37.30742358078603,
+ "grad_norm": 0.015503671951591969,
+ "learning_rate": 0.0006,
+ "loss": 4.034431457519531,
+ "step": 2686
+ },
+ {
+ "epoch": 37.32139737991266,
+ "grad_norm": 0.015370250679552555,
+ "learning_rate": 0.0006,
+ "loss": 4.075836181640625,
+ "step": 2687
+ },
+ {
+ "epoch": 37.3353711790393,
+ "grad_norm": 0.01564021222293377,
+ "learning_rate": 0.0006,
+ "loss": 4.125202178955078,
+ "step": 2688
+ },
+ {
+ "epoch": 37.34934497816594,
+ "grad_norm": 0.016717975959181786,
+ "learning_rate": 0.0006,
+ "loss": 4.153017044067383,
+ "step": 2689
+ },
+ {
+ "epoch": 37.36331877729258,
+ "grad_norm": 0.015263617038726807,
+ "learning_rate": 0.0006,
+ "loss": 4.021814823150635,
+ "step": 2690
+ },
+ {
+ "epoch": 37.377292576419215,
+ "grad_norm": 0.016497142612934113,
+ "learning_rate": 0.0006,
+ "loss": 4.069095611572266,
+ "step": 2691
+ },
+ {
+ "epoch": 37.391266375545854,
+ "grad_norm": 0.01669355481863022,
+ "learning_rate": 0.0006,
+ "loss": 4.174587726593018,
+ "step": 2692
+ },
+ {
+ "epoch": 37.40524017467249,
+ "grad_norm": 0.019695773720741272,
+ "learning_rate": 0.0006,
+ "loss": 4.0073041915893555,
+ "step": 2693
+ },
+ {
+ "epoch": 37.419213973799124,
+ "grad_norm": 0.023720644414424896,
+ "learning_rate": 0.0006,
+ "loss": 4.111948013305664,
+ "step": 2694
+ },
+ {
+ "epoch": 37.43318777292576,
+ "grad_norm": 0.02399350143969059,
+ "learning_rate": 0.0006,
+ "loss": 4.081164360046387,
+ "step": 2695
+ },
+ {
+ "epoch": 37.4471615720524,
+ "grad_norm": 0.019630271941423416,
+ "learning_rate": 0.0006,
+ "loss": 4.065731048583984,
+ "step": 2696
+ },
+ {
+ "epoch": 37.46113537117904,
+ "grad_norm": 0.019679008051753044,
+ "learning_rate": 0.0006,
+ "loss": 3.993852376937866,
+ "step": 2697
+ },
+ {
+ "epoch": 37.47510917030568,
+ "grad_norm": 0.020597299560904503,
+ "learning_rate": 0.0006,
+ "loss": 4.275555610656738,
+ "step": 2698
+ },
+ {
+ "epoch": 37.48908296943232,
+ "grad_norm": 0.01983889564871788,
+ "learning_rate": 0.0006,
+ "loss": 4.254708290100098,
+ "step": 2699
+ },
+ {
+ "epoch": 37.50305676855895,
+ "grad_norm": 0.019150668755173683,
+ "learning_rate": 0.0006,
+ "loss": 4.097885608673096,
+ "step": 2700
+ },
+ {
+ "epoch": 37.51703056768559,
+ "grad_norm": 0.018847715109586716,
+ "learning_rate": 0.0006,
+ "loss": 4.054081439971924,
+ "step": 2701
+ },
+ {
+ "epoch": 37.531004366812226,
+ "grad_norm": 0.019607288762927055,
+ "learning_rate": 0.0006,
+ "loss": 3.9891440868377686,
+ "step": 2702
+ },
+ {
+ "epoch": 37.544978165938865,
+ "grad_norm": 0.018863150849938393,
+ "learning_rate": 0.0006,
+ "loss": 4.084877014160156,
+ "step": 2703
+ },
+ {
+ "epoch": 37.5589519650655,
+ "grad_norm": 0.017340239137411118,
+ "learning_rate": 0.0006,
+ "loss": 4.113290786743164,
+ "step": 2704
+ },
+ {
+ "epoch": 37.57292576419214,
+ "grad_norm": 0.015628693625330925,
+ "learning_rate": 0.0006,
+ "loss": 4.1483001708984375,
+ "step": 2705
+ },
+ {
+ "epoch": 37.58689956331878,
+ "grad_norm": 0.016493283212184906,
+ "learning_rate": 0.0006,
+ "loss": 4.1040143966674805,
+ "step": 2706
+ },
+ {
+ "epoch": 37.60087336244541,
+ "grad_norm": 0.016732443124055862,
+ "learning_rate": 0.0006,
+ "loss": 4.07205057144165,
+ "step": 2707
+ },
+ {
+ "epoch": 37.61484716157205,
+ "grad_norm": 0.017176156863570213,
+ "learning_rate": 0.0006,
+ "loss": 4.1279754638671875,
+ "step": 2708
+ },
+ {
+ "epoch": 37.62882096069869,
+ "grad_norm": 0.015730608254671097,
+ "learning_rate": 0.0006,
+ "loss": 4.086658000946045,
+ "step": 2709
+ },
+ {
+ "epoch": 37.64279475982533,
+ "grad_norm": 0.016111090779304504,
+ "learning_rate": 0.0006,
+ "loss": 4.064335823059082,
+ "step": 2710
+ },
+ {
+ "epoch": 37.65676855895197,
+ "grad_norm": 0.01769379712641239,
+ "learning_rate": 0.0006,
+ "loss": 4.116047382354736,
+ "step": 2711
+ },
+ {
+ "epoch": 37.670742358078606,
+ "grad_norm": 0.01683102548122406,
+ "learning_rate": 0.0006,
+ "loss": 4.095986843109131,
+ "step": 2712
+ },
+ {
+ "epoch": 37.68471615720524,
+ "grad_norm": 0.015726450830698013,
+ "learning_rate": 0.0006,
+ "loss": 4.10775089263916,
+ "step": 2713
+ },
+ {
+ "epoch": 37.698689956331876,
+ "grad_norm": 0.014760235324501991,
+ "learning_rate": 0.0006,
+ "loss": 4.1234893798828125,
+ "step": 2714
+ },
+ {
+ "epoch": 37.712663755458514,
+ "grad_norm": 0.014933438040316105,
+ "learning_rate": 0.0006,
+ "loss": 4.079200744628906,
+ "step": 2715
+ },
+ {
+ "epoch": 37.72663755458515,
+ "grad_norm": 0.015723997727036476,
+ "learning_rate": 0.0006,
+ "loss": 4.07865047454834,
+ "step": 2716
+ },
+ {
+ "epoch": 37.74061135371179,
+ "grad_norm": 0.014539001509547234,
+ "learning_rate": 0.0006,
+ "loss": 4.036242485046387,
+ "step": 2717
+ },
+ {
+ "epoch": 37.75458515283843,
+ "grad_norm": 0.014844979159533978,
+ "learning_rate": 0.0006,
+ "loss": 4.1051483154296875,
+ "step": 2718
+ },
+ {
+ "epoch": 37.76855895196506,
+ "grad_norm": 0.015939844772219658,
+ "learning_rate": 0.0006,
+ "loss": 4.017253875732422,
+ "step": 2719
+ },
+ {
+ "epoch": 37.7825327510917,
+ "grad_norm": 0.014771555550396442,
+ "learning_rate": 0.0006,
+ "loss": 4.0694661140441895,
+ "step": 2720
+ },
+ {
+ "epoch": 37.79650655021834,
+ "grad_norm": 0.015777921304106712,
+ "learning_rate": 0.0006,
+ "loss": 3.947587490081787,
+ "step": 2721
+ },
+ {
+ "epoch": 37.81048034934498,
+ "grad_norm": 0.017451291903853416,
+ "learning_rate": 0.0006,
+ "loss": 4.093625068664551,
+ "step": 2722
+ },
+ {
+ "epoch": 37.82445414847162,
+ "grad_norm": 0.016829246655106544,
+ "learning_rate": 0.0006,
+ "loss": 4.09030818939209,
+ "step": 2723
+ },
+ {
+ "epoch": 37.838427947598255,
+ "grad_norm": 0.016984397545456886,
+ "learning_rate": 0.0006,
+ "loss": 4.140626907348633,
+ "step": 2724
+ },
+ {
+ "epoch": 37.852401746724894,
+ "grad_norm": 0.01732603646814823,
+ "learning_rate": 0.0006,
+ "loss": 4.128276824951172,
+ "step": 2725
+ },
+ {
+ "epoch": 37.866375545851525,
+ "grad_norm": 0.01744619384407997,
+ "learning_rate": 0.0006,
+ "loss": 4.103130340576172,
+ "step": 2726
+ },
+ {
+ "epoch": 37.880349344978164,
+ "grad_norm": 0.016331320628523827,
+ "learning_rate": 0.0006,
+ "loss": 4.037951469421387,
+ "step": 2727
+ },
+ {
+ "epoch": 37.8943231441048,
+ "grad_norm": 0.015919430181384087,
+ "learning_rate": 0.0006,
+ "loss": 4.121513366699219,
+ "step": 2728
+ },
+ {
+ "epoch": 37.90829694323144,
+ "grad_norm": 0.016153695061802864,
+ "learning_rate": 0.0006,
+ "loss": 4.099384307861328,
+ "step": 2729
+ },
+ {
+ "epoch": 37.92227074235808,
+ "grad_norm": 0.01628800667822361,
+ "learning_rate": 0.0006,
+ "loss": 4.092325687408447,
+ "step": 2730
+ },
+ {
+ "epoch": 37.93624454148472,
+ "grad_norm": 0.017285684123635292,
+ "learning_rate": 0.0006,
+ "loss": 3.9828906059265137,
+ "step": 2731
+ },
+ {
+ "epoch": 37.95021834061135,
+ "grad_norm": 0.016047554090619087,
+ "learning_rate": 0.0006,
+ "loss": 4.100419044494629,
+ "step": 2732
+ },
+ {
+ "epoch": 37.96419213973799,
+ "grad_norm": 0.01562986709177494,
+ "learning_rate": 0.0006,
+ "loss": 4.065832614898682,
+ "step": 2733
+ },
+ {
+ "epoch": 37.97816593886463,
+ "grad_norm": 0.01677163504064083,
+ "learning_rate": 0.0006,
+ "loss": 4.077518463134766,
+ "step": 2734
+ },
+ {
+ "epoch": 37.992139737991266,
+ "grad_norm": 0.01755492016673088,
+ "learning_rate": 0.0006,
+ "loss": 4.1076154708862305,
+ "step": 2735
+ },
+ {
+ "epoch": 38.0,
+ "grad_norm": 0.01929762214422226,
+ "learning_rate": 0.0006,
+ "loss": 4.09603214263916,
+ "step": 2736
+ },
+ {
+ "epoch": 38.0,
+ "eval_loss": 4.582442283630371,
+ "eval_runtime": 62.8649,
+ "eval_samples_per_second": 38.845,
+ "eval_steps_per_second": 1.225,
+ "step": 2736
+ },
+ {
+ "epoch": 38.01397379912664,
+ "grad_norm": 0.019129136577248573,
+ "learning_rate": 0.0006,
+ "loss": 4.095005989074707,
+ "step": 2737
+ },
+ {
+ "epoch": 38.02794759825328,
+ "grad_norm": 0.019167408347129822,
+ "learning_rate": 0.0006,
+ "loss": 4.054426670074463,
+ "step": 2738
+ },
+ {
+ "epoch": 38.041921397379916,
+ "grad_norm": 0.01789460889995098,
+ "learning_rate": 0.0006,
+ "loss": 4.0434417724609375,
+ "step": 2739
+ },
+ {
+ "epoch": 38.05589519650655,
+ "grad_norm": 0.015947027131915092,
+ "learning_rate": 0.0006,
+ "loss": 4.074516296386719,
+ "step": 2740
+ },
+ {
+ "epoch": 38.069868995633186,
+ "grad_norm": 0.016667252406477928,
+ "learning_rate": 0.0006,
+ "loss": 4.108545303344727,
+ "step": 2741
+ },
+ {
+ "epoch": 38.083842794759825,
+ "grad_norm": 0.016014399006962776,
+ "learning_rate": 0.0006,
+ "loss": 4.026925086975098,
+ "step": 2742
+ },
+ {
+ "epoch": 38.09781659388646,
+ "grad_norm": 0.015397695824503899,
+ "learning_rate": 0.0006,
+ "loss": 4.163970470428467,
+ "step": 2743
+ },
+ {
+ "epoch": 38.1117903930131,
+ "grad_norm": 0.01563289947807789,
+ "learning_rate": 0.0006,
+ "loss": 4.116578578948975,
+ "step": 2744
+ },
+ {
+ "epoch": 38.12576419213974,
+ "grad_norm": 0.015512318350374699,
+ "learning_rate": 0.0006,
+ "loss": 4.096311569213867,
+ "step": 2745
+ },
+ {
+ "epoch": 38.13973799126637,
+ "grad_norm": 0.015426642261445522,
+ "learning_rate": 0.0006,
+ "loss": 4.017065048217773,
+ "step": 2746
+ },
+ {
+ "epoch": 38.15371179039301,
+ "grad_norm": 0.015115435235202312,
+ "learning_rate": 0.0006,
+ "loss": 4.005434989929199,
+ "step": 2747
+ },
+ {
+ "epoch": 38.16768558951965,
+ "grad_norm": 0.015235894359648228,
+ "learning_rate": 0.0006,
+ "loss": 3.982743263244629,
+ "step": 2748
+ },
+ {
+ "epoch": 38.18165938864629,
+ "grad_norm": 0.016095101833343506,
+ "learning_rate": 0.0006,
+ "loss": 4.0849385261535645,
+ "step": 2749
+ },
+ {
+ "epoch": 38.19563318777293,
+ "grad_norm": 0.015201197937130928,
+ "learning_rate": 0.0006,
+ "loss": 4.103863716125488,
+ "step": 2750
+ },
+ {
+ "epoch": 38.209606986899566,
+ "grad_norm": 0.015687478706240654,
+ "learning_rate": 0.0006,
+ "loss": 4.12074089050293,
+ "step": 2751
+ },
+ {
+ "epoch": 38.223580786026204,
+ "grad_norm": 0.015201589092612267,
+ "learning_rate": 0.0006,
+ "loss": 4.056872367858887,
+ "step": 2752
+ },
+ {
+ "epoch": 38.237554585152836,
+ "grad_norm": 0.015154894441366196,
+ "learning_rate": 0.0006,
+ "loss": 4.089670181274414,
+ "step": 2753
+ },
+ {
+ "epoch": 38.251528384279474,
+ "grad_norm": 0.01586608588695526,
+ "learning_rate": 0.0006,
+ "loss": 4.155317306518555,
+ "step": 2754
+ },
+ {
+ "epoch": 38.26550218340611,
+ "grad_norm": 0.01609654724597931,
+ "learning_rate": 0.0006,
+ "loss": 3.995800495147705,
+ "step": 2755
+ },
+ {
+ "epoch": 38.27947598253275,
+ "grad_norm": 0.015495185740292072,
+ "learning_rate": 0.0006,
+ "loss": 3.995769739151001,
+ "step": 2756
+ },
+ {
+ "epoch": 38.29344978165939,
+ "grad_norm": 0.017042381688952446,
+ "learning_rate": 0.0006,
+ "loss": 4.132832050323486,
+ "step": 2757
+ },
+ {
+ "epoch": 38.30742358078603,
+ "grad_norm": 0.018706614151597023,
+ "learning_rate": 0.0006,
+ "loss": 4.038820743560791,
+ "step": 2758
+ },
+ {
+ "epoch": 38.32139737991266,
+ "grad_norm": 0.018316026777029037,
+ "learning_rate": 0.0006,
+ "loss": 3.972052574157715,
+ "step": 2759
+ },
+ {
+ "epoch": 38.3353711790393,
+ "grad_norm": 0.018375389277935028,
+ "learning_rate": 0.0006,
+ "loss": 4.152097702026367,
+ "step": 2760
+ },
+ {
+ "epoch": 38.34934497816594,
+ "grad_norm": 0.016879886388778687,
+ "learning_rate": 0.0006,
+ "loss": 4.020869731903076,
+ "step": 2761
+ },
+ {
+ "epoch": 38.36331877729258,
+ "grad_norm": 0.015728183090686798,
+ "learning_rate": 0.0006,
+ "loss": 4.008016586303711,
+ "step": 2762
+ },
+ {
+ "epoch": 38.377292576419215,
+ "grad_norm": 0.016147758811712265,
+ "learning_rate": 0.0006,
+ "loss": 4.035854339599609,
+ "step": 2763
+ },
+ {
+ "epoch": 38.391266375545854,
+ "grad_norm": 0.017774797976017,
+ "learning_rate": 0.0006,
+ "loss": 4.122506141662598,
+ "step": 2764
+ },
+ {
+ "epoch": 38.40524017467249,
+ "grad_norm": 0.016590069979429245,
+ "learning_rate": 0.0006,
+ "loss": 4.082847595214844,
+ "step": 2765
+ },
+ {
+ "epoch": 38.419213973799124,
+ "grad_norm": 0.016544528305530548,
+ "learning_rate": 0.0006,
+ "loss": 4.184270858764648,
+ "step": 2766
+ },
+ {
+ "epoch": 38.43318777292576,
+ "grad_norm": 0.017972925677895546,
+ "learning_rate": 0.0006,
+ "loss": 3.987558603286743,
+ "step": 2767
+ },
+ {
+ "epoch": 38.4471615720524,
+ "grad_norm": 0.018562227487564087,
+ "learning_rate": 0.0006,
+ "loss": 4.1193461418151855,
+ "step": 2768
+ },
+ {
+ "epoch": 38.46113537117904,
+ "grad_norm": 0.018896685913205147,
+ "learning_rate": 0.0006,
+ "loss": 4.057808876037598,
+ "step": 2769
+ },
+ {
+ "epoch": 38.47510917030568,
+ "grad_norm": 0.016793809831142426,
+ "learning_rate": 0.0006,
+ "loss": 4.069523334503174,
+ "step": 2770
+ },
+ {
+ "epoch": 38.48908296943232,
+ "grad_norm": 0.01698676310479641,
+ "learning_rate": 0.0006,
+ "loss": 3.997624158859253,
+ "step": 2771
+ },
+ {
+ "epoch": 38.50305676855895,
+ "grad_norm": 0.016784202307462692,
+ "learning_rate": 0.0006,
+ "loss": 4.174795627593994,
+ "step": 2772
+ },
+ {
+ "epoch": 38.51703056768559,
+ "grad_norm": 0.016785014420747757,
+ "learning_rate": 0.0006,
+ "loss": 3.999321222305298,
+ "step": 2773
+ },
+ {
+ "epoch": 38.531004366812226,
+ "grad_norm": 0.018615184351801872,
+ "learning_rate": 0.0006,
+ "loss": 4.191237926483154,
+ "step": 2774
+ },
+ {
+ "epoch": 38.544978165938865,
+ "grad_norm": 0.019090790301561356,
+ "learning_rate": 0.0006,
+ "loss": 4.106966018676758,
+ "step": 2775
+ },
+ {
+ "epoch": 38.5589519650655,
+ "grad_norm": 0.018298320472240448,
+ "learning_rate": 0.0006,
+ "loss": 4.063363552093506,
+ "step": 2776
+ },
+ {
+ "epoch": 38.57292576419214,
+ "grad_norm": 0.018541090190410614,
+ "learning_rate": 0.0006,
+ "loss": 4.101193428039551,
+ "step": 2777
+ },
+ {
+ "epoch": 38.58689956331878,
+ "grad_norm": 0.01849970407783985,
+ "learning_rate": 0.0006,
+ "loss": 4.071335792541504,
+ "step": 2778
+ },
+ {
+ "epoch": 38.60087336244541,
+ "grad_norm": 0.018309077247977257,
+ "learning_rate": 0.0006,
+ "loss": 4.245305061340332,
+ "step": 2779
+ },
+ {
+ "epoch": 38.61484716157205,
+ "grad_norm": 0.017681840807199478,
+ "learning_rate": 0.0006,
+ "loss": 4.081173419952393,
+ "step": 2780
+ },
+ {
+ "epoch": 38.62882096069869,
+ "grad_norm": 0.016286224126815796,
+ "learning_rate": 0.0006,
+ "loss": 4.024351596832275,
+ "step": 2781
+ },
+ {
+ "epoch": 38.64279475982533,
+ "grad_norm": 0.015837235376238823,
+ "learning_rate": 0.0006,
+ "loss": 4.028205394744873,
+ "step": 2782
+ },
+ {
+ "epoch": 38.65676855895197,
+ "grad_norm": 0.016062596812844276,
+ "learning_rate": 0.0006,
+ "loss": 4.131122589111328,
+ "step": 2783
+ },
+ {
+ "epoch": 38.670742358078606,
+ "grad_norm": 0.01711239665746689,
+ "learning_rate": 0.0006,
+ "loss": 4.124416351318359,
+ "step": 2784
+ },
+ {
+ "epoch": 38.68471615720524,
+ "grad_norm": 0.015304913744330406,
+ "learning_rate": 0.0006,
+ "loss": 4.056122779846191,
+ "step": 2785
+ },
+ {
+ "epoch": 38.698689956331876,
+ "grad_norm": 0.015971887856721878,
+ "learning_rate": 0.0006,
+ "loss": 4.132555961608887,
+ "step": 2786
+ },
+ {
+ "epoch": 38.712663755458514,
+ "grad_norm": 0.015787221491336823,
+ "learning_rate": 0.0006,
+ "loss": 4.051596164703369,
+ "step": 2787
+ },
+ {
+ "epoch": 38.72663755458515,
+ "grad_norm": 0.01605444774031639,
+ "learning_rate": 0.0006,
+ "loss": 4.191808700561523,
+ "step": 2788
+ },
+ {
+ "epoch": 38.74061135371179,
+ "grad_norm": 0.016383009031414986,
+ "learning_rate": 0.0006,
+ "loss": 4.0648417472839355,
+ "step": 2789
+ },
+ {
+ "epoch": 38.75458515283843,
+ "grad_norm": 0.016135554760694504,
+ "learning_rate": 0.0006,
+ "loss": 3.988563060760498,
+ "step": 2790
+ },
+ {
+ "epoch": 38.76855895196506,
+ "grad_norm": 0.016634158790111542,
+ "learning_rate": 0.0006,
+ "loss": 4.133493423461914,
+ "step": 2791
+ },
+ {
+ "epoch": 38.7825327510917,
+ "grad_norm": 0.01688210666179657,
+ "learning_rate": 0.0006,
+ "loss": 4.091277122497559,
+ "step": 2792
+ },
+ {
+ "epoch": 38.79650655021834,
+ "grad_norm": 0.016102122142910957,
+ "learning_rate": 0.0006,
+ "loss": 4.051219940185547,
+ "step": 2793
+ },
+ {
+ "epoch": 38.81048034934498,
+ "grad_norm": 0.016585038974881172,
+ "learning_rate": 0.0006,
+ "loss": 4.058587074279785,
+ "step": 2794
+ },
+ {
+ "epoch": 38.82445414847162,
+ "grad_norm": 0.017700951546430588,
+ "learning_rate": 0.0006,
+ "loss": 4.016030311584473,
+ "step": 2795
+ },
+ {
+ "epoch": 38.838427947598255,
+ "grad_norm": 0.017922203987836838,
+ "learning_rate": 0.0006,
+ "loss": 4.079977989196777,
+ "step": 2796
+ },
+ {
+ "epoch": 38.852401746724894,
+ "grad_norm": 0.017299890518188477,
+ "learning_rate": 0.0006,
+ "loss": 4.062984943389893,
+ "step": 2797
+ },
+ {
+ "epoch": 38.866375545851525,
+ "grad_norm": 0.018648581579327583,
+ "learning_rate": 0.0006,
+ "loss": 4.065717697143555,
+ "step": 2798
+ },
+ {
+ "epoch": 38.880349344978164,
+ "grad_norm": 0.019057920202612877,
+ "learning_rate": 0.0006,
+ "loss": 4.227892875671387,
+ "step": 2799
+ },
+ {
+ "epoch": 38.8943231441048,
+ "grad_norm": 0.018824422731995583,
+ "learning_rate": 0.0006,
+ "loss": 4.116058349609375,
+ "step": 2800
+ },
+ {
+ "epoch": 38.90829694323144,
+ "grad_norm": 0.019084978848695755,
+ "learning_rate": 0.0006,
+ "loss": 4.061677932739258,
+ "step": 2801
+ },
+ {
+ "epoch": 38.92227074235808,
+ "grad_norm": 0.017842981964349747,
+ "learning_rate": 0.0006,
+ "loss": 4.083813667297363,
+ "step": 2802
+ },
+ {
+ "epoch": 38.93624454148472,
+ "grad_norm": 0.01880858838558197,
+ "learning_rate": 0.0006,
+ "loss": 4.152561187744141,
+ "step": 2803
+ },
+ {
+ "epoch": 38.95021834061135,
+ "grad_norm": 0.017400087788701057,
+ "learning_rate": 0.0006,
+ "loss": 4.154635429382324,
+ "step": 2804
+ },
+ {
+ "epoch": 38.96419213973799,
+ "grad_norm": 0.017682507634162903,
+ "learning_rate": 0.0006,
+ "loss": 4.151128768920898,
+ "step": 2805
+ },
+ {
+ "epoch": 38.97816593886463,
+ "grad_norm": 0.015803052112460136,
+ "learning_rate": 0.0006,
+ "loss": 4.10498571395874,
+ "step": 2806
+ },
+ {
+ "epoch": 38.992139737991266,
+ "grad_norm": 0.01582511141896248,
+ "learning_rate": 0.0006,
+ "loss": 4.140159606933594,
+ "step": 2807
+ },
+ {
+ "epoch": 39.0,
+ "grad_norm": 0.01925666816532612,
+ "learning_rate": 0.0006,
+ "loss": 4.126394271850586,
+ "step": 2808
+ },
+ {
+ "epoch": 39.0,
+ "eval_loss": 4.615617275238037,
+ "eval_runtime": 56.8895,
+ "eval_samples_per_second": 42.925,
+ "eval_steps_per_second": 1.354,
+ "step": 2808
+ },
+ {
+ "epoch": 39.01397379912664,
+ "grad_norm": 0.01677853614091873,
+ "learning_rate": 0.0006,
+ "loss": 4.102395057678223,
+ "step": 2809
+ },
+ {
+ "epoch": 39.02794759825328,
+ "grad_norm": 0.017901137471199036,
+ "learning_rate": 0.0006,
+ "loss": 3.976609230041504,
+ "step": 2810
+ },
+ {
+ "epoch": 39.041921397379916,
+ "grad_norm": 0.017978668212890625,
+ "learning_rate": 0.0006,
+ "loss": 3.944416046142578,
+ "step": 2811
+ },
+ {
+ "epoch": 39.05589519650655,
+ "grad_norm": 0.019967492669820786,
+ "learning_rate": 0.0006,
+ "loss": 3.992234468460083,
+ "step": 2812
+ },
+ {
+ "epoch": 39.069868995633186,
+ "grad_norm": 0.018937544897198677,
+ "learning_rate": 0.0006,
+ "loss": 4.028573036193848,
+ "step": 2813
+ },
+ {
+ "epoch": 39.083842794759825,
+ "grad_norm": 0.016852952539920807,
+ "learning_rate": 0.0006,
+ "loss": 4.062442779541016,
+ "step": 2814
+ },
+ {
+ "epoch": 39.09781659388646,
+ "grad_norm": 0.016643647104501724,
+ "learning_rate": 0.0006,
+ "loss": 4.08871603012085,
+ "step": 2815
+ },
+ {
+ "epoch": 39.1117903930131,
+ "grad_norm": 0.016872866079211235,
+ "learning_rate": 0.0006,
+ "loss": 3.9514245986938477,
+ "step": 2816
+ },
+ {
+ "epoch": 39.12576419213974,
+ "grad_norm": 0.018045669421553612,
+ "learning_rate": 0.0006,
+ "loss": 4.038842678070068,
+ "step": 2817
+ },
+ {
+ "epoch": 39.13973799126637,
+ "grad_norm": 0.017853403463959694,
+ "learning_rate": 0.0006,
+ "loss": 3.93709135055542,
+ "step": 2818
+ },
+ {
+ "epoch": 39.15371179039301,
+ "grad_norm": 0.017507443204522133,
+ "learning_rate": 0.0006,
+ "loss": 3.98665714263916,
+ "step": 2819
+ },
+ {
+ "epoch": 39.16768558951965,
+ "grad_norm": 0.01766853965818882,
+ "learning_rate": 0.0006,
+ "loss": 3.937126636505127,
+ "step": 2820
+ },
+ {
+ "epoch": 39.18165938864629,
+ "grad_norm": 0.016660021618008614,
+ "learning_rate": 0.0006,
+ "loss": 3.8397014141082764,
+ "step": 2821
+ },
+ {
+ "epoch": 39.19563318777293,
+ "grad_norm": 0.017719309777021408,
+ "learning_rate": 0.0006,
+ "loss": 4.003579616546631,
+ "step": 2822
+ },
+ {
+ "epoch": 39.209606986899566,
+ "grad_norm": 0.018234414979815483,
+ "learning_rate": 0.0006,
+ "loss": 4.053859710693359,
+ "step": 2823
+ },
+ {
+ "epoch": 39.223580786026204,
+ "grad_norm": 0.018777024000883102,
+ "learning_rate": 0.0006,
+ "loss": 4.056182861328125,
+ "step": 2824
+ },
+ {
+ "epoch": 39.237554585152836,
+ "grad_norm": 0.017443502321839333,
+ "learning_rate": 0.0006,
+ "loss": 4.025068283081055,
+ "step": 2825
+ },
+ {
+ "epoch": 39.251528384279474,
+ "grad_norm": 0.016500331461429596,
+ "learning_rate": 0.0006,
+ "loss": 4.101631164550781,
+ "step": 2826
+ },
+ {
+ "epoch": 39.26550218340611,
+ "grad_norm": 0.016101544722914696,
+ "learning_rate": 0.0006,
+ "loss": 4.027082443237305,
+ "step": 2827
+ },
+ {
+ "epoch": 39.27947598253275,
+ "grad_norm": 0.015471171587705612,
+ "learning_rate": 0.0006,
+ "loss": 4.022909641265869,
+ "step": 2828
+ },
+ {
+ "epoch": 39.29344978165939,
+ "grad_norm": 0.015779748558998108,
+ "learning_rate": 0.0006,
+ "loss": 4.18796968460083,
+ "step": 2829
+ },
+ {
+ "epoch": 39.30742358078603,
+ "grad_norm": 0.015522679314017296,
+ "learning_rate": 0.0006,
+ "loss": 4.005777359008789,
+ "step": 2830
+ },
+ {
+ "epoch": 39.32139737991266,
+ "grad_norm": 0.015226504765450954,
+ "learning_rate": 0.0006,
+ "loss": 4.108416557312012,
+ "step": 2831
+ },
+ {
+ "epoch": 39.3353711790393,
+ "grad_norm": 0.015095778740942478,
+ "learning_rate": 0.0006,
+ "loss": 4.08029842376709,
+ "step": 2832
+ },
+ {
+ "epoch": 39.34934497816594,
+ "grad_norm": 0.015601031482219696,
+ "learning_rate": 0.0006,
+ "loss": 3.989492416381836,
+ "step": 2833
+ },
+ {
+ "epoch": 39.36331877729258,
+ "grad_norm": 0.014825052581727505,
+ "learning_rate": 0.0006,
+ "loss": 3.998520612716675,
+ "step": 2834
+ },
+ {
+ "epoch": 39.377292576419215,
+ "grad_norm": 0.014821100980043411,
+ "learning_rate": 0.0006,
+ "loss": 4.001777172088623,
+ "step": 2835
+ },
+ {
+ "epoch": 39.391266375545854,
+ "grad_norm": 0.015333754941821098,
+ "learning_rate": 0.0006,
+ "loss": 4.11985969543457,
+ "step": 2836
+ },
+ {
+ "epoch": 39.40524017467249,
+ "grad_norm": 0.015940699726343155,
+ "learning_rate": 0.0006,
+ "loss": 4.025373935699463,
+ "step": 2837
+ },
+ {
+ "epoch": 39.419213973799124,
+ "grad_norm": 0.015096287243068218,
+ "learning_rate": 0.0006,
+ "loss": 4.068551063537598,
+ "step": 2838
+ },
+ {
+ "epoch": 39.43318777292576,
+ "grad_norm": 0.01533548254519701,
+ "learning_rate": 0.0006,
+ "loss": 3.9800801277160645,
+ "step": 2839
+ },
+ {
+ "epoch": 39.4471615720524,
+ "grad_norm": 0.014648732729256153,
+ "learning_rate": 0.0006,
+ "loss": 3.9670190811157227,
+ "step": 2840
+ },
+ {
+ "epoch": 39.46113537117904,
+ "grad_norm": 0.015759671106934547,
+ "learning_rate": 0.0006,
+ "loss": 4.050443649291992,
+ "step": 2841
+ },
+ {
+ "epoch": 39.47510917030568,
+ "grad_norm": 0.016054073348641396,
+ "learning_rate": 0.0006,
+ "loss": 4.22734260559082,
+ "step": 2842
+ },
+ {
+ "epoch": 39.48908296943232,
+ "grad_norm": 0.0160264540463686,
+ "learning_rate": 0.0006,
+ "loss": 4.116174221038818,
+ "step": 2843
+ },
+ {
+ "epoch": 39.50305676855895,
+ "grad_norm": 0.015426610596477985,
+ "learning_rate": 0.0006,
+ "loss": 4.000178337097168,
+ "step": 2844
+ },
+ {
+ "epoch": 39.51703056768559,
+ "grad_norm": 0.01520642638206482,
+ "learning_rate": 0.0006,
+ "loss": 4.218410968780518,
+ "step": 2845
+ },
+ {
+ "epoch": 39.531004366812226,
+ "grad_norm": 0.015305949375033379,
+ "learning_rate": 0.0006,
+ "loss": 4.025891304016113,
+ "step": 2846
+ },
+ {
+ "epoch": 39.544978165938865,
+ "grad_norm": 0.015461564064025879,
+ "learning_rate": 0.0006,
+ "loss": 4.100950717926025,
+ "step": 2847
+ },
+ {
+ "epoch": 39.5589519650655,
+ "grad_norm": 0.01712625101208687,
+ "learning_rate": 0.0006,
+ "loss": 4.0620856285095215,
+ "step": 2848
+ },
+ {
+ "epoch": 39.57292576419214,
+ "grad_norm": 0.017133301123976707,
+ "learning_rate": 0.0006,
+ "loss": 4.097138404846191,
+ "step": 2849
+ },
+ {
+ "epoch": 39.58689956331878,
+ "grad_norm": 0.01683625392615795,
+ "learning_rate": 0.0006,
+ "loss": 4.084364891052246,
+ "step": 2850
+ },
+ {
+ "epoch": 39.60087336244541,
+ "grad_norm": 0.0165710486471653,
+ "learning_rate": 0.0006,
+ "loss": 4.042361259460449,
+ "step": 2851
+ },
+ {
+ "epoch": 39.61484716157205,
+ "grad_norm": 0.01660393364727497,
+ "learning_rate": 0.0006,
+ "loss": 4.1395487785339355,
+ "step": 2852
+ },
+ {
+ "epoch": 39.62882096069869,
+ "grad_norm": 0.017025845125317574,
+ "learning_rate": 0.0006,
+ "loss": 4.1207075119018555,
+ "step": 2853
+ },
+ {
+ "epoch": 39.64279475982533,
+ "grad_norm": 0.01660497486591339,
+ "learning_rate": 0.0006,
+ "loss": 4.103846073150635,
+ "step": 2854
+ },
+ {
+ "epoch": 39.65676855895197,
+ "grad_norm": 0.01607946865260601,
+ "learning_rate": 0.0006,
+ "loss": 4.189016342163086,
+ "step": 2855
+ },
+ {
+ "epoch": 39.670742358078606,
+ "grad_norm": 0.01614505983889103,
+ "learning_rate": 0.0006,
+ "loss": 4.079860687255859,
+ "step": 2856
+ },
+ {
+ "epoch": 39.68471615720524,
+ "grad_norm": 0.017085086554288864,
+ "learning_rate": 0.0006,
+ "loss": 3.9774293899536133,
+ "step": 2857
+ },
+ {
+ "epoch": 39.698689956331876,
+ "grad_norm": 0.01816520094871521,
+ "learning_rate": 0.0006,
+ "loss": 4.113666534423828,
+ "step": 2858
+ },
+ {
+ "epoch": 39.712663755458514,
+ "grad_norm": 0.016597526147961617,
+ "learning_rate": 0.0006,
+ "loss": 4.068242073059082,
+ "step": 2859
+ },
+ {
+ "epoch": 39.72663755458515,
+ "grad_norm": 0.01672235131263733,
+ "learning_rate": 0.0006,
+ "loss": 4.024444580078125,
+ "step": 2860
+ },
+ {
+ "epoch": 39.74061135371179,
+ "grad_norm": 0.020093556493520737,
+ "learning_rate": 0.0006,
+ "loss": 4.064124584197998,
+ "step": 2861
+ },
+ {
+ "epoch": 39.75458515283843,
+ "grad_norm": 0.023141304031014442,
+ "learning_rate": 0.0006,
+ "loss": 4.130439758300781,
+ "step": 2862
+ },
+ {
+ "epoch": 39.76855895196506,
+ "grad_norm": 0.02289372682571411,
+ "learning_rate": 0.0006,
+ "loss": 4.14794921875,
+ "step": 2863
+ },
+ {
+ "epoch": 39.7825327510917,
+ "grad_norm": 0.01919020712375641,
+ "learning_rate": 0.0006,
+ "loss": 4.123961925506592,
+ "step": 2864
+ },
+ {
+ "epoch": 39.79650655021834,
+ "grad_norm": 0.016932329162955284,
+ "learning_rate": 0.0006,
+ "loss": 4.0944132804870605,
+ "step": 2865
+ },
+ {
+ "epoch": 39.81048034934498,
+ "grad_norm": 0.016644535586237907,
+ "learning_rate": 0.0006,
+ "loss": 4.124269962310791,
+ "step": 2866
+ },
+ {
+ "epoch": 39.82445414847162,
+ "grad_norm": 0.016302216798067093,
+ "learning_rate": 0.0006,
+ "loss": 4.0267205238342285,
+ "step": 2867
+ },
+ {
+ "epoch": 39.838427947598255,
+ "grad_norm": 0.01673300378024578,
+ "learning_rate": 0.0006,
+ "loss": 4.10650634765625,
+ "step": 2868
+ },
+ {
+ "epoch": 39.852401746724894,
+ "grad_norm": 0.016581356525421143,
+ "learning_rate": 0.0006,
+ "loss": 4.1547956466674805,
+ "step": 2869
+ },
+ {
+ "epoch": 39.866375545851525,
+ "grad_norm": 0.01691797934472561,
+ "learning_rate": 0.0006,
+ "loss": 4.056936740875244,
+ "step": 2870
+ },
+ {
+ "epoch": 39.880349344978164,
+ "grad_norm": 0.01596726104617119,
+ "learning_rate": 0.0006,
+ "loss": 3.9891414642333984,
+ "step": 2871
+ },
+ {
+ "epoch": 39.8943231441048,
+ "grad_norm": 0.015615091659128666,
+ "learning_rate": 0.0006,
+ "loss": 4.137948036193848,
+ "step": 2872
+ },
+ {
+ "epoch": 39.90829694323144,
+ "grad_norm": 0.015703413635492325,
+ "learning_rate": 0.0006,
+ "loss": 3.9844930171966553,
+ "step": 2873
+ },
+ {
+ "epoch": 39.92227074235808,
+ "grad_norm": 0.015453094616532326,
+ "learning_rate": 0.0006,
+ "loss": 4.098443508148193,
+ "step": 2874
+ },
+ {
+ "epoch": 39.93624454148472,
+ "grad_norm": 0.01449158601462841,
+ "learning_rate": 0.0006,
+ "loss": 4.090852737426758,
+ "step": 2875
+ },
+ {
+ "epoch": 39.95021834061135,
+ "grad_norm": 0.01537515688687563,
+ "learning_rate": 0.0006,
+ "loss": 4.239612579345703,
+ "step": 2876
+ },
+ {
+ "epoch": 39.96419213973799,
+ "grad_norm": 0.015989921987056732,
+ "learning_rate": 0.0006,
+ "loss": 4.142073154449463,
+ "step": 2877
+ },
+ {
+ "epoch": 39.97816593886463,
+ "grad_norm": 0.015985611826181412,
+ "learning_rate": 0.0006,
+ "loss": 4.100076675415039,
+ "step": 2878
+ },
+ {
+ "epoch": 39.992139737991266,
+ "grad_norm": 0.015260833315551281,
+ "learning_rate": 0.0006,
+ "loss": 4.068576335906982,
+ "step": 2879
+ },
+ {
+ "epoch": 40.0,
+ "grad_norm": 0.018047941848635674,
+ "learning_rate": 0.0006,
+ "loss": 4.005878925323486,
+ "step": 2880
+ },
+ {
+ "epoch": 40.0,
+ "eval_loss": 4.622848987579346,
+ "eval_runtime": 56.9618,
+ "eval_samples_per_second": 42.871,
+ "eval_steps_per_second": 1.352,
+ "step": 2880
+ },
+ {
+ "epoch": 40.01397379912664,
+ "grad_norm": 0.015976838767528534,
+ "learning_rate": 0.0006,
+ "loss": 4.035689353942871,
+ "step": 2881
+ },
+ {
+ "epoch": 40.02794759825328,
+ "grad_norm": 0.015176220797002316,
+ "learning_rate": 0.0006,
+ "loss": 3.932197332382202,
+ "step": 2882
+ },
+ {
+ "epoch": 40.041921397379916,
+ "grad_norm": 0.016676228493452072,
+ "learning_rate": 0.0006,
+ "loss": 4.014171123504639,
+ "step": 2883
+ },
+ {
+ "epoch": 40.05589519650655,
+ "grad_norm": 0.01698281429708004,
+ "learning_rate": 0.0006,
+ "loss": 4.117893218994141,
+ "step": 2884
+ },
+ {
+ "epoch": 40.069868995633186,
+ "grad_norm": 0.017872486263513565,
+ "learning_rate": 0.0006,
+ "loss": 4.132119178771973,
+ "step": 2885
+ },
+ {
+ "epoch": 40.083842794759825,
+ "grad_norm": 0.017615554854273796,
+ "learning_rate": 0.0006,
+ "loss": 4.072000980377197,
+ "step": 2886
+ },
+ {
+ "epoch": 40.09781659388646,
+ "grad_norm": 0.017469394952058792,
+ "learning_rate": 0.0006,
+ "loss": 4.017726421356201,
+ "step": 2887
+ },
+ {
+ "epoch": 40.1117903930131,
+ "grad_norm": 0.01954762637615204,
+ "learning_rate": 0.0006,
+ "loss": 4.014443874359131,
+ "step": 2888
+ },
+ {
+ "epoch": 40.12576419213974,
+ "grad_norm": 0.018935853615403175,
+ "learning_rate": 0.0006,
+ "loss": 4.0042948722839355,
+ "step": 2889
+ },
+ {
+ "epoch": 40.13973799126637,
+ "grad_norm": 0.019382692873477936,
+ "learning_rate": 0.0006,
+ "loss": 4.06160831451416,
+ "step": 2890
+ },
+ {
+ "epoch": 40.15371179039301,
+ "grad_norm": 0.019677115604281425,
+ "learning_rate": 0.0006,
+ "loss": 4.13466215133667,
+ "step": 2891
+ },
+ {
+ "epoch": 40.16768558951965,
+ "grad_norm": 0.020365122705698013,
+ "learning_rate": 0.0006,
+ "loss": 4.095000267028809,
+ "step": 2892
+ },
+ {
+ "epoch": 40.18165938864629,
+ "grad_norm": 0.021842138841748238,
+ "learning_rate": 0.0006,
+ "loss": 3.9877853393554688,
+ "step": 2893
+ },
+ {
+ "epoch": 40.19563318777293,
+ "grad_norm": 0.02354881353676319,
+ "learning_rate": 0.0006,
+ "loss": 3.9430112838745117,
+ "step": 2894
+ },
+ {
+ "epoch": 40.209606986899566,
+ "grad_norm": 0.026602735742926598,
+ "learning_rate": 0.0006,
+ "loss": 4.012076377868652,
+ "step": 2895
+ },
+ {
+ "epoch": 40.223580786026204,
+ "grad_norm": 0.023976484313607216,
+ "learning_rate": 0.0006,
+ "loss": 3.9768595695495605,
+ "step": 2896
+ },
+ {
+ "epoch": 40.237554585152836,
+ "grad_norm": 0.02336616814136505,
+ "learning_rate": 0.0006,
+ "loss": 4.1096038818359375,
+ "step": 2897
+ },
+ {
+ "epoch": 40.251528384279474,
+ "grad_norm": 0.02426023595035076,
+ "learning_rate": 0.0006,
+ "loss": 4.157189846038818,
+ "step": 2898
+ },
+ {
+ "epoch": 40.26550218340611,
+ "grad_norm": 0.021572696045041084,
+ "learning_rate": 0.0006,
+ "loss": 3.979175090789795,
+ "step": 2899
+ },
+ {
+ "epoch": 40.27947598253275,
+ "grad_norm": 0.08300432562828064,
+ "learning_rate": 0.0006,
+ "loss": 4.092336654663086,
+ "step": 2900
+ },
+ {
+ "epoch": 40.29344978165939,
+ "grad_norm": 0.7323967814445496,
+ "learning_rate": 0.0006,
+ "loss": 5.337990760803223,
+ "step": 2901
+ },
+ {
+ "epoch": 40.30742358078603,
+ "grad_norm": 0.3363337814807892,
+ "learning_rate": 0.0006,
+ "loss": 6.40043830871582,
+ "step": 2902
+ },
+ {
+ "epoch": 40.32139737991266,
+ "grad_norm": 0.598686933517456,
+ "learning_rate": 0.0006,
+ "loss": 7.004765510559082,
+ "step": 2903
+ },
+ {
+ "epoch": 40.3353711790393,
+ "grad_norm": 0.5998364686965942,
+ "learning_rate": 0.0006,
+ "loss": 8.184747695922852,
+ "step": 2904
+ },
+ {
+ "epoch": 40.34934497816594,
+ "grad_norm": 1.408208966255188,
+ "learning_rate": 0.0006,
+ "loss": 8.88978385925293,
+ "step": 2905
+ },
+ {
+ "epoch": 40.36331877729258,
+ "grad_norm": 0.7829785943031311,
+ "learning_rate": 0.0006,
+ "loss": 11.13601303100586,
+ "step": 2906
+ },
+ {
+ "epoch": 40.377292576419215,
+ "grad_norm": 0.7769215106964111,
+ "learning_rate": 0.0006,
+ "loss": 9.90416145324707,
+ "step": 2907
+ },
+ {
+ "epoch": 40.391266375545854,
+ "grad_norm": 0.46508079767227173,
+ "learning_rate": 0.0006,
+ "loss": 9.647209167480469,
+ "step": 2908
+ },
+ {
+ "epoch": 40.40524017467249,
+ "grad_norm": 1.0228968858718872,
+ "learning_rate": 0.0006,
+ "loss": 11.685434341430664,
+ "step": 2909
+ },
+ {
+ "epoch": 40.419213973799124,
+ "grad_norm": 0.20553778111934662,
+ "learning_rate": 0.0006,
+ "loss": 8.621708869934082,
+ "step": 2910
+ },
+ {
+ "epoch": 40.43318777292576,
+ "grad_norm": 0.18466424942016602,
+ "learning_rate": 0.0006,
+ "loss": 8.590787887573242,
+ "step": 2911
+ },
+ {
+ "epoch": 40.4471615720524,
+ "grad_norm": 0.10458594560623169,
+ "learning_rate": 0.0006,
+ "loss": 8.324458122253418,
+ "step": 2912
+ },
+ {
+ "epoch": 40.46113537117904,
+ "grad_norm": 0.1096842959523201,
+ "learning_rate": 0.0006,
+ "loss": 8.300851821899414,
+ "step": 2913
+ },
+ {
+ "epoch": 40.47510917030568,
+ "grad_norm": 0.10501226782798767,
+ "learning_rate": 0.0006,
+ "loss": 8.21009635925293,
+ "step": 2914
+ },
+ {
+ "epoch": 40.48908296943232,
+ "grad_norm": 0.13704165816307068,
+ "learning_rate": 0.0006,
+ "loss": 8.141355514526367,
+ "step": 2915
+ },
+ {
+ "epoch": 40.50305676855895,
+ "grad_norm": 0.12129108607769012,
+ "learning_rate": 0.0006,
+ "loss": 8.074612617492676,
+ "step": 2916
+ },
+ {
+ "epoch": 40.51703056768559,
+ "grad_norm": 0.10028897225856781,
+ "learning_rate": 0.0006,
+ "loss": 8.00384521484375,
+ "step": 2917
+ },
+ {
+ "epoch": 40.531004366812226,
+ "grad_norm": 0.11764246970415115,
+ "learning_rate": 0.0006,
+ "loss": 8.078598022460938,
+ "step": 2918
+ },
+ {
+ "epoch": 40.544978165938865,
+ "grad_norm": 0.10081959515810013,
+ "learning_rate": 0.0006,
+ "loss": 8.008938789367676,
+ "step": 2919
+ },
+ {
+ "epoch": 40.5589519650655,
+ "grad_norm": 0.1003170907497406,
+ "learning_rate": 0.0006,
+ "loss": 7.962940216064453,
+ "step": 2920
+ },
+ {
+ "epoch": 40.57292576419214,
+ "grad_norm": 0.08763492107391357,
+ "learning_rate": 0.0006,
+ "loss": 7.909419536590576,
+ "step": 2921
+ },
+ {
+ "epoch": 40.58689956331878,
+ "grad_norm": 0.0636807233095169,
+ "learning_rate": 0.0006,
+ "loss": 7.878905773162842,
+ "step": 2922
+ },
+ {
+ "epoch": 40.60087336244541,
+ "grad_norm": 0.06005936488509178,
+ "learning_rate": 0.0006,
+ "loss": 7.850980758666992,
+ "step": 2923
+ },
+ {
+ "epoch": 40.61484716157205,
+ "grad_norm": 0.06415296345949173,
+ "learning_rate": 0.0006,
+ "loss": 7.830353260040283,
+ "step": 2924
+ },
+ {
+ "epoch": 40.62882096069869,
+ "grad_norm": 0.0732007846236229,
+ "learning_rate": 0.0006,
+ "loss": 7.830746650695801,
+ "step": 2925
+ },
+ {
+ "epoch": 40.64279475982533,
+ "grad_norm": 0.06628110259771347,
+ "learning_rate": 0.0006,
+ "loss": 7.82661771774292,
+ "step": 2926
+ },
+ {
+ "epoch": 40.65676855895197,
+ "grad_norm": 0.05544218048453331,
+ "learning_rate": 0.0006,
+ "loss": 7.814305782318115,
+ "step": 2927
+ },
+ {
+ "epoch": 40.670742358078606,
+ "grad_norm": 0.050327103585004807,
+ "learning_rate": 0.0006,
+ "loss": 7.764055252075195,
+ "step": 2928
+ },
+ {
+ "epoch": 40.68471615720524,
+ "grad_norm": 0.04918425530195236,
+ "learning_rate": 0.0006,
+ "loss": 7.744625091552734,
+ "step": 2929
+ },
+ {
+ "epoch": 40.698689956331876,
+ "grad_norm": 0.042862266302108765,
+ "learning_rate": 0.0006,
+ "loss": 7.764426231384277,
+ "step": 2930
+ },
+ {
+ "epoch": 40.712663755458514,
+ "grad_norm": 0.04338089004158974,
+ "learning_rate": 0.0006,
+ "loss": 7.712530136108398,
+ "step": 2931
+ },
+ {
+ "epoch": 40.72663755458515,
+ "grad_norm": 0.04127737879753113,
+ "learning_rate": 0.0006,
+ "loss": 7.742180347442627,
+ "step": 2932
+ },
+ {
+ "epoch": 40.74061135371179,
+ "grad_norm": 0.03358631581068039,
+ "learning_rate": 0.0006,
+ "loss": 7.714834213256836,
+ "step": 2933
+ },
+ {
+ "epoch": 40.75458515283843,
+ "grad_norm": 0.02845790982246399,
+ "learning_rate": 0.0006,
+ "loss": 7.7021026611328125,
+ "step": 2934
+ },
+ {
+ "epoch": 40.76855895196506,
+ "grad_norm": 0.03012419492006302,
+ "learning_rate": 0.0006,
+ "loss": 7.712377548217773,
+ "step": 2935
+ },
+ {
+ "epoch": 40.7825327510917,
+ "grad_norm": 0.032981179654598236,
+ "learning_rate": 0.0006,
+ "loss": 7.666385650634766,
+ "step": 2936
+ },
+ {
+ "epoch": 40.79650655021834,
+ "grad_norm": 0.03357074037194252,
+ "learning_rate": 0.0006,
+ "loss": 7.649097919464111,
+ "step": 2937
+ },
+ {
+ "epoch": 40.81048034934498,
+ "grad_norm": 0.030914708971977234,
+ "learning_rate": 0.0006,
+ "loss": 7.649196624755859,
+ "step": 2938
+ },
+ {
+ "epoch": 40.82445414847162,
+ "grad_norm": 0.026296131312847137,
+ "learning_rate": 0.0006,
+ "loss": 7.629964828491211,
+ "step": 2939
+ },
+ {
+ "epoch": 40.838427947598255,
+ "grad_norm": 0.026677245274186134,
+ "learning_rate": 0.0006,
+ "loss": 7.599973678588867,
+ "step": 2940
+ },
+ {
+ "epoch": 40.852401746724894,
+ "grad_norm": 0.023208219558000565,
+ "learning_rate": 0.0006,
+ "loss": 7.601624488830566,
+ "step": 2941
+ },
+ {
+ "epoch": 40.866375545851525,
+ "grad_norm": 0.020741475746035576,
+ "learning_rate": 0.0006,
+ "loss": 7.603902339935303,
+ "step": 2942
+ },
+ {
+ "epoch": 40.880349344978164,
+ "grad_norm": 0.019221751019358635,
+ "learning_rate": 0.0006,
+ "loss": 7.5535888671875,
+ "step": 2943
+ },
+ {
+ "epoch": 40.8943231441048,
+ "grad_norm": 0.019491951912641525,
+ "learning_rate": 0.0006,
+ "loss": 7.54597282409668,
+ "step": 2944
+ },
+ {
+ "epoch": 40.90829694323144,
+ "grad_norm": 0.019702572375535965,
+ "learning_rate": 0.0006,
+ "loss": 7.537567138671875,
+ "step": 2945
+ },
+ {
+ "epoch": 40.92227074235808,
+ "grad_norm": 0.01768800988793373,
+ "learning_rate": 0.0006,
+ "loss": 7.512800216674805,
+ "step": 2946
+ },
+ {
+ "epoch": 40.93624454148472,
+ "grad_norm": 0.01926138624548912,
+ "learning_rate": 0.0006,
+ "loss": 7.509932518005371,
+ "step": 2947
+ },
+ {
+ "epoch": 40.95021834061135,
+ "grad_norm": 0.019030006602406502,
+ "learning_rate": 0.0006,
+ "loss": 7.497983932495117,
+ "step": 2948
+ },
+ {
+ "epoch": 40.96419213973799,
+ "grad_norm": 0.017313841730356216,
+ "learning_rate": 0.0006,
+ "loss": 7.463518142700195,
+ "step": 2949
+ },
+ {
+ "epoch": 40.97816593886463,
+ "grad_norm": 0.01310762483626604,
+ "learning_rate": 0.0006,
+ "loss": 7.474668502807617,
+ "step": 2950
+ },
+ {
+ "epoch": 40.992139737991266,
+ "grad_norm": 0.013008282519876957,
+ "learning_rate": 0.0006,
+ "loss": 7.45982551574707,
+ "step": 2951
+ },
+ {
+ "epoch": 41.0,
+ "grad_norm": 0.01441363524645567,
+ "learning_rate": 0.0006,
+ "loss": 7.446529388427734,
+ "step": 2952
+ },
+ {
+ "epoch": 41.0,
+ "eval_loss": 7.486120223999023,
+ "eval_runtime": 70.4986,
+ "eval_samples_per_second": 34.639,
+ "eval_steps_per_second": 1.092,
+ "step": 2952
+ },
+ {
+ "epoch": 41.01397379912664,
+ "grad_norm": 0.014061855152249336,
+ "learning_rate": 0.0006,
+ "loss": 7.4692182540893555,
+ "step": 2953
+ },
+ {
+ "epoch": 41.02794759825328,
+ "grad_norm": 0.01587500050663948,
+ "learning_rate": 0.0006,
+ "loss": 7.4182024002075195,
+ "step": 2954
+ },
+ {
+ "epoch": 41.041921397379916,
+ "grad_norm": 0.014783984050154686,
+ "learning_rate": 0.0006,
+ "loss": 7.466778755187988,
+ "step": 2955
+ },
+ {
+ "epoch": 41.05589519650655,
+ "grad_norm": 0.011560854502022266,
+ "learning_rate": 0.0006,
+ "loss": 7.438002586364746,
+ "step": 2956
+ },
+ {
+ "epoch": 41.069868995633186,
+ "grad_norm": 0.00970957800745964,
+ "learning_rate": 0.0006,
+ "loss": 7.442694187164307,
+ "step": 2957
+ },
+ {
+ "epoch": 41.083842794759825,
+ "grad_norm": 0.009633459150791168,
+ "learning_rate": 0.0006,
+ "loss": 7.458884239196777,
+ "step": 2958
+ },
+ {
+ "epoch": 41.09781659388646,
+ "grad_norm": 0.010664350353181362,
+ "learning_rate": 0.0006,
+ "loss": 7.48378324508667,
+ "step": 2959
+ },
+ {
+ "epoch": 41.1117903930131,
+ "grad_norm": 0.011440591886639595,
+ "learning_rate": 0.0006,
+ "loss": 7.474484920501709,
+ "step": 2960
+ },
+ {
+ "epoch": 41.12576419213974,
+ "grad_norm": 0.01150908600538969,
+ "learning_rate": 0.0006,
+ "loss": 7.417902946472168,
+ "step": 2961
+ },
+ {
+ "epoch": 41.13973799126637,
+ "grad_norm": 0.010375697165727615,
+ "learning_rate": 0.0006,
+ "loss": 7.403279781341553,
+ "step": 2962
+ },
+ {
+ "epoch": 41.15371179039301,
+ "grad_norm": 0.009075703099370003,
+ "learning_rate": 0.0006,
+ "loss": 7.447347164154053,
+ "step": 2963
+ },
+ {
+ "epoch": 41.16768558951965,
+ "grad_norm": 0.008892177604138851,
+ "learning_rate": 0.0006,
+ "loss": 7.402681350708008,
+ "step": 2964
+ },
+ {
+ "epoch": 41.18165938864629,
+ "grad_norm": 0.009492084383964539,
+ "learning_rate": 0.0006,
+ "loss": 7.3871355056762695,
+ "step": 2965
+ },
+ {
+ "epoch": 41.19563318777293,
+ "grad_norm": 0.009139460511505604,
+ "learning_rate": 0.0006,
+ "loss": 7.419617176055908,
+ "step": 2966
+ },
+ {
+ "epoch": 41.209606986899566,
+ "grad_norm": 0.007959138602018356,
+ "learning_rate": 0.0006,
+ "loss": 7.374949932098389,
+ "step": 2967
+ },
+ {
+ "epoch": 41.223580786026204,
+ "grad_norm": 0.007571548223495483,
+ "learning_rate": 0.0006,
+ "loss": 7.4205241203308105,
+ "step": 2968
+ },
+ {
+ "epoch": 41.237554585152836,
+ "grad_norm": 0.007885400205850601,
+ "learning_rate": 0.0006,
+ "loss": 7.416543006896973,
+ "step": 2969
+ },
+ {
+ "epoch": 41.251528384279474,
+ "grad_norm": 0.008650158531963825,
+ "learning_rate": 0.0006,
+ "loss": 7.357326984405518,
+ "step": 2970
+ },
+ {
+ "epoch": 41.26550218340611,
+ "grad_norm": 0.007483262103050947,
+ "learning_rate": 0.0006,
+ "loss": 7.410891056060791,
+ "step": 2971
+ },
+ {
+ "epoch": 41.27947598253275,
+ "grad_norm": 0.00869277399033308,
+ "learning_rate": 0.0006,
+ "loss": 7.392127990722656,
+ "step": 2972
+ },
+ {
+ "epoch": 41.29344978165939,
+ "grad_norm": 0.00885078776627779,
+ "learning_rate": 0.0006,
+ "loss": 7.405449867248535,
+ "step": 2973
+ },
+ {
+ "epoch": 41.30742358078603,
+ "grad_norm": 0.00958226341754198,
+ "learning_rate": 0.0006,
+ "loss": 7.399868488311768,
+ "step": 2974
+ },
+ {
+ "epoch": 41.32139737991266,
+ "grad_norm": 0.006991599686443806,
+ "learning_rate": 0.0006,
+ "loss": 7.35369873046875,
+ "step": 2975
+ },
+ {
+ "epoch": 41.3353711790393,
+ "grad_norm": 0.007415142375975847,
+ "learning_rate": 0.0006,
+ "loss": 7.358353137969971,
+ "step": 2976
+ },
+ {
+ "epoch": 41.34934497816594,
+ "grad_norm": 0.007024371065199375,
+ "learning_rate": 0.0006,
+ "loss": 7.4070892333984375,
+ "step": 2977
+ },
+ {
+ "epoch": 41.36331877729258,
+ "grad_norm": 0.007234332151710987,
+ "learning_rate": 0.0006,
+ "loss": 7.379120349884033,
+ "step": 2978
+ },
+ {
+ "epoch": 41.377292576419215,
+ "grad_norm": 0.00798892229795456,
+ "learning_rate": 0.0006,
+ "loss": 7.365455627441406,
+ "step": 2979
+ },
+ {
+ "epoch": 41.391266375545854,
+ "grad_norm": 0.007303398102521896,
+ "learning_rate": 0.0006,
+ "loss": 7.341798782348633,
+ "step": 2980
+ },
+ {
+ "epoch": 41.40524017467249,
+ "grad_norm": 0.007119669578969479,
+ "learning_rate": 0.0006,
+ "loss": 7.363280296325684,
+ "step": 2981
+ },
+ {
+ "epoch": 41.419213973799124,
+ "grad_norm": 0.007457619532942772,
+ "learning_rate": 0.0006,
+ "loss": 7.39388370513916,
+ "step": 2982
+ },
+ {
+ "epoch": 41.43318777292576,
+ "grad_norm": 0.007905688136816025,
+ "learning_rate": 0.0006,
+ "loss": 7.380652904510498,
+ "step": 2983
+ },
+ {
+ "epoch": 41.4471615720524,
+ "grad_norm": 0.006212855689227581,
+ "learning_rate": 0.0006,
+ "loss": 7.398769378662109,
+ "step": 2984
+ },
+ {
+ "epoch": 41.46113537117904,
+ "grad_norm": 0.0070641543716192245,
+ "learning_rate": 0.0006,
+ "loss": 7.3503546714782715,
+ "step": 2985
+ },
+ {
+ "epoch": 41.47510917030568,
+ "grad_norm": 0.00787805300205946,
+ "learning_rate": 0.0006,
+ "loss": 7.347678184509277,
+ "step": 2986
+ },
+ {
+ "epoch": 41.48908296943232,
+ "grad_norm": 0.006760426796972752,
+ "learning_rate": 0.0006,
+ "loss": 7.380006790161133,
+ "step": 2987
+ },
+ {
+ "epoch": 41.50305676855895,
+ "grad_norm": 0.00772385997697711,
+ "learning_rate": 0.0006,
+ "loss": 7.363680362701416,
+ "step": 2988
+ },
+ {
+ "epoch": 41.51703056768559,
+ "grad_norm": 0.007373564876616001,
+ "learning_rate": 0.0006,
+ "loss": 7.333014488220215,
+ "step": 2989
+ },
+ {
+ "epoch": 41.531004366812226,
+ "grad_norm": 0.006717653013765812,
+ "learning_rate": 0.0006,
+ "loss": 7.360076427459717,
+ "step": 2990
+ },
+ {
+ "epoch": 41.544978165938865,
+ "grad_norm": 0.007191413082182407,
+ "learning_rate": 0.0006,
+ "loss": 7.397030830383301,
+ "step": 2991
+ },
+ {
+ "epoch": 41.5589519650655,
+ "grad_norm": 0.007831960916519165,
+ "learning_rate": 0.0006,
+ "loss": 7.369959831237793,
+ "step": 2992
+ },
+ {
+ "epoch": 41.57292576419214,
+ "grad_norm": 0.006947481073439121,
+ "learning_rate": 0.0006,
+ "loss": 7.333254337310791,
+ "step": 2993
+ },
+ {
+ "epoch": 41.58689956331878,
+ "grad_norm": 0.006243606563657522,
+ "learning_rate": 0.0006,
+ "loss": 7.327378273010254,
+ "step": 2994
+ },
+ {
+ "epoch": 41.60087336244541,
+ "grad_norm": 0.006298339460045099,
+ "learning_rate": 0.0006,
+ "loss": 7.323482036590576,
+ "step": 2995
+ },
+ {
+ "epoch": 41.61484716157205,
+ "grad_norm": 0.007587412837892771,
+ "learning_rate": 0.0006,
+ "loss": 7.313789367675781,
+ "step": 2996
+ },
+ {
+ "epoch": 41.62882096069869,
+ "grad_norm": 0.007140746805816889,
+ "learning_rate": 0.0006,
+ "loss": 7.356067657470703,
+ "step": 2997
+ },
+ {
+ "epoch": 41.64279475982533,
+ "grad_norm": 0.008315625600516796,
+ "learning_rate": 0.0006,
+ "loss": 7.3339643478393555,
+ "step": 2998
+ },
+ {
+ "epoch": 41.65676855895197,
+ "grad_norm": 0.006547102238982916,
+ "learning_rate": 0.0006,
+ "loss": 7.342191696166992,
+ "step": 2999
+ },
+ {
+ "epoch": 41.670742358078606,
+ "grad_norm": 0.006068122107535601,
+ "learning_rate": 0.0006,
+ "loss": 7.321015357971191,
+ "step": 3000
+ },
+ {
+ "epoch": 41.68471615720524,
+ "grad_norm": 0.007316991221159697,
+ "learning_rate": 0.0006,
+ "loss": 7.346490859985352,
+ "step": 3001
+ },
+ {
+ "epoch": 41.698689956331876,
+ "grad_norm": 0.0067388564348220825,
+ "learning_rate": 0.0006,
+ "loss": 7.340723991394043,
+ "step": 3002
+ },
+ {
+ "epoch": 41.712663755458514,
+ "grad_norm": 0.006839963141828775,
+ "learning_rate": 0.0006,
+ "loss": 7.329365253448486,
+ "step": 3003
+ },
+ {
+ "epoch": 41.72663755458515,
+ "grad_norm": 0.00805041566491127,
+ "learning_rate": 0.0006,
+ "loss": 7.298934459686279,
+ "step": 3004
+ },
+ {
+ "epoch": 41.74061135371179,
+ "grad_norm": 0.007197503466159105,
+ "learning_rate": 0.0006,
+ "loss": 7.331496238708496,
+ "step": 3005
+ },
+ {
+ "epoch": 41.75458515283843,
+ "grad_norm": 0.007220805156975985,
+ "learning_rate": 0.0006,
+ "loss": 7.317695140838623,
+ "step": 3006
+ },
+ {
+ "epoch": 41.76855895196506,
+ "grad_norm": 0.00874276738613844,
+ "learning_rate": 0.0006,
+ "loss": 7.272614479064941,
+ "step": 3007
+ },
+ {
+ "epoch": 41.7825327510917,
+ "grad_norm": 0.008386565372347832,
+ "learning_rate": 0.0006,
+ "loss": 7.295244216918945,
+ "step": 3008
+ },
+ {
+ "epoch": 41.79650655021834,
+ "grad_norm": 0.0077730282209813595,
+ "learning_rate": 0.0006,
+ "loss": 7.3395490646362305,
+ "step": 3009
+ },
+ {
+ "epoch": 41.81048034934498,
+ "grad_norm": 0.008441480807960033,
+ "learning_rate": 0.0006,
+ "loss": 7.319921493530273,
+ "step": 3010
+ },
+ {
+ "epoch": 41.82445414847162,
+ "grad_norm": 0.007398269604891539,
+ "learning_rate": 0.0006,
+ "loss": 7.3029465675354,
+ "step": 3011
+ },
+ {
+ "epoch": 41.838427947598255,
+ "grad_norm": 0.006878677289932966,
+ "learning_rate": 0.0006,
+ "loss": 7.318467140197754,
+ "step": 3012
+ },
+ {
+ "epoch": 41.852401746724894,
+ "grad_norm": 0.007616598624736071,
+ "learning_rate": 0.0006,
+ "loss": 7.3112077713012695,
+ "step": 3013
+ },
+ {
+ "epoch": 41.866375545851525,
+ "grad_norm": 0.011629699729382992,
+ "learning_rate": 0.0006,
+ "loss": 7.32523250579834,
+ "step": 3014
+ },
+ {
+ "epoch": 41.880349344978164,
+ "grad_norm": 0.011011249385774136,
+ "learning_rate": 0.0006,
+ "loss": 7.287247657775879,
+ "step": 3015
+ },
+ {
+ "epoch": 41.8943231441048,
+ "grad_norm": 0.010208295658230782,
+ "learning_rate": 0.0006,
+ "loss": 7.292117118835449,
+ "step": 3016
+ },
+ {
+ "epoch": 41.90829694323144,
+ "grad_norm": 0.008651095442473888,
+ "learning_rate": 0.0006,
+ "loss": 7.302201271057129,
+ "step": 3017
+ },
+ {
+ "epoch": 41.92227074235808,
+ "grad_norm": 0.006692920345813036,
+ "learning_rate": 0.0006,
+ "loss": 7.319568634033203,
+ "step": 3018
+ },
+ {
+ "epoch": 41.93624454148472,
+ "grad_norm": 0.010033348575234413,
+ "learning_rate": 0.0006,
+ "loss": 7.2738566398620605,
+ "step": 3019
+ },
+ {
+ "epoch": 41.95021834061135,
+ "grad_norm": 0.011790832504630089,
+ "learning_rate": 0.0006,
+ "loss": 7.298540115356445,
+ "step": 3020
+ },
+ {
+ "epoch": 41.96419213973799,
+ "grad_norm": 0.018622662872076035,
+ "learning_rate": 0.0006,
+ "loss": 7.267182350158691,
+ "step": 3021
+ },
+ {
+ "epoch": 41.97816593886463,
+ "grad_norm": 0.02520257607102394,
+ "learning_rate": 0.0006,
+ "loss": 7.261566162109375,
+ "step": 3022
+ },
+ {
+ "epoch": 41.992139737991266,
+ "grad_norm": 0.036332644522190094,
+ "learning_rate": 0.0006,
+ "loss": 7.270719051361084,
+ "step": 3023
+ },
+ {
+ "epoch": 42.0,
+ "grad_norm": 0.04717305675148964,
+ "learning_rate": 0.0006,
+ "loss": 7.293332099914551,
+ "step": 3024
+ },
+ {
+ "epoch": 42.0,
+ "eval_loss": 7.354280471801758,
+ "eval_runtime": 56.4353,
+ "eval_samples_per_second": 43.271,
+ "eval_steps_per_second": 1.364,
+ "step": 3024
+ },
+ {
+ "epoch": 42.01397379912664,
+ "grad_norm": 0.06661626696586609,
+ "learning_rate": 0.0006,
+ "loss": 7.338189601898193,
+ "step": 3025
+ },
+ {
+ "epoch": 42.02794759825328,
+ "grad_norm": 0.05452054366469383,
+ "learning_rate": 0.0006,
+ "loss": 7.316764831542969,
+ "step": 3026
+ },
+ {
+ "epoch": 42.041921397379916,
+ "grad_norm": 0.025638775900006294,
+ "learning_rate": 0.0006,
+ "loss": 7.34204626083374,
+ "step": 3027
+ },
+ {
+ "epoch": 42.05589519650655,
+ "grad_norm": 0.022398799657821655,
+ "learning_rate": 0.0006,
+ "loss": 7.262149333953857,
+ "step": 3028
+ },
+ {
+ "epoch": 42.069868995633186,
+ "grad_norm": 0.03713499382138252,
+ "learning_rate": 0.0006,
+ "loss": 7.27695369720459,
+ "step": 3029
+ },
+ {
+ "epoch": 42.083842794759825,
+ "grad_norm": 0.014949247241020203,
+ "learning_rate": 0.0006,
+ "loss": 7.296508312225342,
+ "step": 3030
+ },
+ {
+ "epoch": 42.09781659388646,
+ "grad_norm": 0.02703995071351528,
+ "learning_rate": 0.0006,
+ "loss": 7.258018493652344,
+ "step": 3031
+ },
+ {
+ "epoch": 42.1117903930131,
+ "grad_norm": 0.020352816209197044,
+ "learning_rate": 0.0006,
+ "loss": 7.241450309753418,
+ "step": 3032
+ },
+ {
+ "epoch": 42.12576419213974,
+ "grad_norm": 0.014578479342162609,
+ "learning_rate": 0.0006,
+ "loss": 7.257323265075684,
+ "step": 3033
+ },
+ {
+ "epoch": 42.13973799126637,
+ "grad_norm": 0.026053044945001602,
+ "learning_rate": 0.0006,
+ "loss": 7.254680633544922,
+ "step": 3034
+ },
+ {
+ "epoch": 42.15371179039301,
+ "grad_norm": 0.015120275318622589,
+ "learning_rate": 0.0006,
+ "loss": 7.261074066162109,
+ "step": 3035
+ },
+ {
+ "epoch": 42.16768558951965,
+ "grad_norm": 0.016822177916765213,
+ "learning_rate": 0.0006,
+ "loss": 7.279064178466797,
+ "step": 3036
+ },
+ {
+ "epoch": 42.18165938864629,
+ "grad_norm": 0.016367588192224503,
+ "learning_rate": 0.0006,
+ "loss": 7.251429557800293,
+ "step": 3037
+ },
+ {
+ "epoch": 42.19563318777293,
+ "grad_norm": 0.014350418001413345,
+ "learning_rate": 0.0006,
+ "loss": 7.259264945983887,
+ "step": 3038
+ },
+ {
+ "epoch": 42.209606986899566,
+ "grad_norm": 0.017078228294849396,
+ "learning_rate": 0.0006,
+ "loss": 7.227105140686035,
+ "step": 3039
+ },
+ {
+ "epoch": 42.223580786026204,
+ "grad_norm": 0.011759317480027676,
+ "learning_rate": 0.0006,
+ "loss": 7.275888442993164,
+ "step": 3040
+ },
+ {
+ "epoch": 42.237554585152836,
+ "grad_norm": 0.013019340112805367,
+ "learning_rate": 0.0006,
+ "loss": 7.2517008781433105,
+ "step": 3041
+ },
+ {
+ "epoch": 42.251528384279474,
+ "grad_norm": 0.01178441196680069,
+ "learning_rate": 0.0006,
+ "loss": 7.23576545715332,
+ "step": 3042
+ },
+ {
+ "epoch": 42.26550218340611,
+ "grad_norm": 0.014518891461193562,
+ "learning_rate": 0.0006,
+ "loss": 7.253895282745361,
+ "step": 3043
+ },
+ {
+ "epoch": 42.27947598253275,
+ "grad_norm": 0.018159950152039528,
+ "learning_rate": 0.0006,
+ "loss": 7.278058052062988,
+ "step": 3044
+ },
+ {
+ "epoch": 42.29344978165939,
+ "grad_norm": 0.007587796077132225,
+ "learning_rate": 0.0006,
+ "loss": 7.278578758239746,
+ "step": 3045
+ },
+ {
+ "epoch": 42.30742358078603,
+ "grad_norm": 0.015994979068636894,
+ "learning_rate": 0.0006,
+ "loss": 7.239841461181641,
+ "step": 3046
+ },
+ {
+ "epoch": 42.32139737991266,
+ "grad_norm": 0.017830168828368187,
+ "learning_rate": 0.0006,
+ "loss": 7.2443976402282715,
+ "step": 3047
+ },
+ {
+ "epoch": 42.3353711790393,
+ "grad_norm": 0.008738759905099869,
+ "learning_rate": 0.0006,
+ "loss": 7.232264518737793,
+ "step": 3048
+ },
+ {
+ "epoch": 42.34934497816594,
+ "grad_norm": 0.0175449438393116,
+ "learning_rate": 0.0006,
+ "loss": 7.228000640869141,
+ "step": 3049
+ },
+ {
+ "epoch": 42.36331877729258,
+ "grad_norm": 0.019611820578575134,
+ "learning_rate": 0.0006,
+ "loss": 7.234999179840088,
+ "step": 3050
+ },
+ {
+ "epoch": 42.377292576419215,
+ "grad_norm": 0.015302863903343678,
+ "learning_rate": 0.0006,
+ "loss": 7.242547988891602,
+ "step": 3051
+ },
+ {
+ "epoch": 42.391266375545854,
+ "grad_norm": 0.025919731706380844,
+ "learning_rate": 0.0006,
+ "loss": 7.216652870178223,
+ "step": 3052
+ },
+ {
+ "epoch": 42.40524017467249,
+ "grad_norm": 0.034587468951940536,
+ "learning_rate": 0.0006,
+ "loss": 7.245336532592773,
+ "step": 3053
+ },
+ {
+ "epoch": 42.419213973799124,
+ "grad_norm": 0.019198840484023094,
+ "learning_rate": 0.0006,
+ "loss": 7.256758689880371,
+ "step": 3054
+ },
+ {
+ "epoch": 42.43318777292576,
+ "grad_norm": 0.012316389009356499,
+ "learning_rate": 0.0006,
+ "loss": 7.209102630615234,
+ "step": 3055
+ },
+ {
+ "epoch": 42.4471615720524,
+ "grad_norm": 0.023512162268161774,
+ "learning_rate": 0.0006,
+ "loss": 7.250742435455322,
+ "step": 3056
+ },
+ {
+ "epoch": 42.46113537117904,
+ "grad_norm": 0.025457851588726044,
+ "learning_rate": 0.0006,
+ "loss": 7.263757705688477,
+ "step": 3057
+ },
+ {
+ "epoch": 42.47510917030568,
+ "grad_norm": 0.02192617766559124,
+ "learning_rate": 0.0006,
+ "loss": 7.246977806091309,
+ "step": 3058
+ },
+ {
+ "epoch": 42.48908296943232,
+ "grad_norm": 0.02541281282901764,
+ "learning_rate": 0.0006,
+ "loss": 7.232361316680908,
+ "step": 3059
+ },
+ {
+ "epoch": 42.50305676855895,
+ "grad_norm": 0.036353763192892075,
+ "learning_rate": 0.0006,
+ "loss": 7.242556571960449,
+ "step": 3060
+ },
+ {
+ "epoch": 42.51703056768559,
+ "grad_norm": 0.04555872827768326,
+ "learning_rate": 0.0006,
+ "loss": 7.228940963745117,
+ "step": 3061
+ },
+ {
+ "epoch": 42.531004366812226,
+ "grad_norm": 0.04563378915190697,
+ "learning_rate": 0.0006,
+ "loss": 7.235276222229004,
+ "step": 3062
+ },
+ {
+ "epoch": 42.544978165938865,
+ "grad_norm": 0.04275760427117348,
+ "learning_rate": 0.0006,
+ "loss": 7.238895416259766,
+ "step": 3063
+ },
+ {
+ "epoch": 42.5589519650655,
+ "grad_norm": 0.04535885155200958,
+ "learning_rate": 0.0006,
+ "loss": 7.205336570739746,
+ "step": 3064
+ },
+ {
+ "epoch": 42.57292576419214,
+ "grad_norm": 0.015534617006778717,
+ "learning_rate": 0.0006,
+ "loss": 7.238393783569336,
+ "step": 3065
+ },
+ {
+ "epoch": 42.58689956331878,
+ "grad_norm": 0.03286299481987953,
+ "learning_rate": 0.0006,
+ "loss": 7.216453552246094,
+ "step": 3066
+ },
+ {
+ "epoch": 42.60087336244541,
+ "grad_norm": 0.025453168898820877,
+ "learning_rate": 0.0006,
+ "loss": 7.203750133514404,
+ "step": 3067
+ },
+ {
+ "epoch": 42.61484716157205,
+ "grad_norm": 0.021501649171113968,
+ "learning_rate": 0.0006,
+ "loss": 7.241621971130371,
+ "step": 3068
+ },
+ {
+ "epoch": 42.62882096069869,
+ "grad_norm": 0.02743634767830372,
+ "learning_rate": 0.0006,
+ "loss": 7.22076940536499,
+ "step": 3069
+ },
+ {
+ "epoch": 42.64279475982533,
+ "grad_norm": 0.020890483632683754,
+ "learning_rate": 0.0006,
+ "loss": 7.218412399291992,
+ "step": 3070
+ },
+ {
+ "epoch": 42.65676855895197,
+ "grad_norm": 0.022499850019812584,
+ "learning_rate": 0.0006,
+ "loss": 7.217696189880371,
+ "step": 3071
+ },
+ {
+ "epoch": 42.670742358078606,
+ "grad_norm": 0.01653696969151497,
+ "learning_rate": 0.0006,
+ "loss": 7.156417369842529,
+ "step": 3072
+ },
+ {
+ "epoch": 42.68471615720524,
+ "grad_norm": 0.01834225095808506,
+ "learning_rate": 0.0006,
+ "loss": 7.211247444152832,
+ "step": 3073
+ },
+ {
+ "epoch": 42.698689956331876,
+ "grad_norm": 0.020137229934334755,
+ "learning_rate": 0.0006,
+ "loss": 7.232499599456787,
+ "step": 3074
+ },
+ {
+ "epoch": 42.712663755458514,
+ "grad_norm": 0.012671459466218948,
+ "learning_rate": 0.0006,
+ "loss": 7.172441482543945,
+ "step": 3075
+ },
+ {
+ "epoch": 42.72663755458515,
+ "grad_norm": 0.016803130507469177,
+ "learning_rate": 0.0006,
+ "loss": 7.152103424072266,
+ "step": 3076
+ },
+ {
+ "epoch": 42.74061135371179,
+ "grad_norm": 0.01623712107539177,
+ "learning_rate": 0.0006,
+ "loss": 7.187032699584961,
+ "step": 3077
+ },
+ {
+ "epoch": 42.75458515283843,
+ "grad_norm": 0.015938356518745422,
+ "learning_rate": 0.0006,
+ "loss": 7.171140670776367,
+ "step": 3078
+ },
+ {
+ "epoch": 42.76855895196506,
+ "grad_norm": 0.02580736204981804,
+ "learning_rate": 0.0006,
+ "loss": 7.204472541809082,
+ "step": 3079
+ },
+ {
+ "epoch": 42.7825327510917,
+ "grad_norm": 0.023372117429971695,
+ "learning_rate": 0.0006,
+ "loss": 7.165554523468018,
+ "step": 3080
+ },
+ {
+ "epoch": 42.79650655021834,
+ "grad_norm": 0.04440586268901825,
+ "learning_rate": 0.0006,
+ "loss": 7.169431209564209,
+ "step": 3081
+ },
+ {
+ "epoch": 42.81048034934498,
+ "grad_norm": 0.06797406077384949,
+ "learning_rate": 0.0006,
+ "loss": 7.168193340301514,
+ "step": 3082
+ },
+ {
+ "epoch": 42.82445414847162,
+ "grad_norm": 0.0914321169257164,
+ "learning_rate": 0.0006,
+ "loss": 7.204671859741211,
+ "step": 3083
+ },
+ {
+ "epoch": 42.838427947598255,
+ "grad_norm": 0.0506940595805645,
+ "learning_rate": 0.0006,
+ "loss": 7.15949821472168,
+ "step": 3084
+ },
+ {
+ "epoch": 42.852401746724894,
+ "grad_norm": 0.04819890484213829,
+ "learning_rate": 0.0006,
+ "loss": 7.192837238311768,
+ "step": 3085
+ },
+ {
+ "epoch": 42.866375545851525,
+ "grad_norm": 0.040044233202934265,
+ "learning_rate": 0.0006,
+ "loss": 7.1600165367126465,
+ "step": 3086
+ },
+ {
+ "epoch": 42.880349344978164,
+ "grad_norm": 0.02573299966752529,
+ "learning_rate": 0.0006,
+ "loss": 7.153427600860596,
+ "step": 3087
+ },
+ {
+ "epoch": 42.8943231441048,
+ "grad_norm": 0.046942535787820816,
+ "learning_rate": 0.0006,
+ "loss": 7.16767692565918,
+ "step": 3088
+ },
+ {
+ "epoch": 42.90829694323144,
+ "grad_norm": 0.02683161571621895,
+ "learning_rate": 0.0006,
+ "loss": 7.17144250869751,
+ "step": 3089
+ },
+ {
+ "epoch": 42.92227074235808,
+ "grad_norm": 0.05363674834370613,
+ "learning_rate": 0.0006,
+ "loss": 7.120128631591797,
+ "step": 3090
+ },
+ {
+ "epoch": 42.93624454148472,
+ "grad_norm": 0.023756247013807297,
+ "learning_rate": 0.0006,
+ "loss": 7.169703006744385,
+ "step": 3091
+ },
+ {
+ "epoch": 42.95021834061135,
+ "grad_norm": 0.04216783866286278,
+ "learning_rate": 0.0006,
+ "loss": 7.1206865310668945,
+ "step": 3092
+ },
+ {
+ "epoch": 42.96419213973799,
+ "grad_norm": 0.02219565585255623,
+ "learning_rate": 0.0006,
+ "loss": 7.157215118408203,
+ "step": 3093
+ },
+ {
+ "epoch": 42.97816593886463,
+ "grad_norm": 0.03527127206325531,
+ "learning_rate": 0.0006,
+ "loss": 7.166672706604004,
+ "step": 3094
+ },
+ {
+ "epoch": 42.992139737991266,
+ "grad_norm": 0.017748817801475525,
+ "learning_rate": 0.0006,
+ "loss": 7.154747009277344,
+ "step": 3095
+ },
+ {
+ "epoch": 43.0,
+ "grad_norm": 0.033165838569402695,
+ "learning_rate": 0.0006,
+ "loss": 7.150866508483887,
+ "step": 3096
+ },
+ {
+ "epoch": 43.0,
+ "eval_loss": 7.144403457641602,
+ "eval_runtime": 56.8853,
+ "eval_samples_per_second": 42.928,
+ "eval_steps_per_second": 1.354,
+ "step": 3096
+ },
+ {
+ "epoch": 43.01397379912664,
+ "grad_norm": 0.015199829824268818,
+ "learning_rate": 0.0006,
+ "loss": 7.122735023498535,
+ "step": 3097
+ },
+ {
+ "epoch": 43.02794759825328,
+ "grad_norm": 0.03220256045460701,
+ "learning_rate": 0.0006,
+ "loss": 7.143553733825684,
+ "step": 3098
+ },
+ {
+ "epoch": 43.041921397379916,
+ "grad_norm": 0.014687920920550823,
+ "learning_rate": 0.0006,
+ "loss": 7.080960750579834,
+ "step": 3099
+ },
+ {
+ "epoch": 43.05589519650655,
+ "grad_norm": 0.023522932082414627,
+ "learning_rate": 0.0006,
+ "loss": 7.137356758117676,
+ "step": 3100
+ },
+ {
+ "epoch": 43.069868995633186,
+ "grad_norm": 0.014808216132223606,
+ "learning_rate": 0.0006,
+ "loss": 7.09143590927124,
+ "step": 3101
+ },
+ {
+ "epoch": 43.083842794759825,
+ "grad_norm": 0.029339246451854706,
+ "learning_rate": 0.0006,
+ "loss": 7.122198104858398,
+ "step": 3102
+ },
+ {
+ "epoch": 43.09781659388646,
+ "grad_norm": 0.022953078150749207,
+ "learning_rate": 0.0006,
+ "loss": 7.092505931854248,
+ "step": 3103
+ },
+ {
+ "epoch": 43.1117903930131,
+ "grad_norm": 0.04590052738785744,
+ "learning_rate": 0.0006,
+ "loss": 7.113100051879883,
+ "step": 3104
+ },
+ {
+ "epoch": 43.12576419213974,
+ "grad_norm": 0.0524442233145237,
+ "learning_rate": 0.0006,
+ "loss": 7.147022247314453,
+ "step": 3105
+ },
+ {
+ "epoch": 43.13973799126637,
+ "grad_norm": 0.0315794013440609,
+ "learning_rate": 0.0006,
+ "loss": 7.094045162200928,
+ "step": 3106
+ },
+ {
+ "epoch": 43.15371179039301,
+ "grad_norm": 0.05112722888588905,
+ "learning_rate": 0.0006,
+ "loss": 7.054388523101807,
+ "step": 3107
+ },
+ {
+ "epoch": 43.16768558951965,
+ "grad_norm": 0.03146776929497719,
+ "learning_rate": 0.0006,
+ "loss": 7.116674423217773,
+ "step": 3108
+ },
+ {
+ "epoch": 43.18165938864629,
+ "grad_norm": 0.03391246497631073,
+ "learning_rate": 0.0006,
+ "loss": 7.150345802307129,
+ "step": 3109
+ },
+ {
+ "epoch": 43.19563318777293,
+ "grad_norm": 0.026957903057336807,
+ "learning_rate": 0.0006,
+ "loss": 7.058531284332275,
+ "step": 3110
+ },
+ {
+ "epoch": 43.209606986899566,
+ "grad_norm": 0.041532568633556366,
+ "learning_rate": 0.0006,
+ "loss": 7.085498809814453,
+ "step": 3111
+ },
+ {
+ "epoch": 43.223580786026204,
+ "grad_norm": 0.020001588389277458,
+ "learning_rate": 0.0006,
+ "loss": 7.131487846374512,
+ "step": 3112
+ },
+ {
+ "epoch": 43.237554585152836,
+ "grad_norm": 0.02268712781369686,
+ "learning_rate": 0.0006,
+ "loss": 7.0790863037109375,
+ "step": 3113
+ },
+ {
+ "epoch": 43.251528384279474,
+ "grad_norm": 0.023363925516605377,
+ "learning_rate": 0.0006,
+ "loss": 7.067816734313965,
+ "step": 3114
+ },
+ {
+ "epoch": 43.26550218340611,
+ "grad_norm": 0.019367508590221405,
+ "learning_rate": 0.0006,
+ "loss": 7.042041778564453,
+ "step": 3115
+ },
+ {
+ "epoch": 43.27947598253275,
+ "grad_norm": 0.02745090238749981,
+ "learning_rate": 0.0006,
+ "loss": 7.064733028411865,
+ "step": 3116
+ },
+ {
+ "epoch": 43.29344978165939,
+ "grad_norm": 0.02543802745640278,
+ "learning_rate": 0.0006,
+ "loss": 7.058948040008545,
+ "step": 3117
+ },
+ {
+ "epoch": 43.30742358078603,
+ "grad_norm": 0.018874092027544975,
+ "learning_rate": 0.0006,
+ "loss": 7.009946346282959,
+ "step": 3118
+ },
+ {
+ "epoch": 43.32139737991266,
+ "grad_norm": 0.024662822484970093,
+ "learning_rate": 0.0006,
+ "loss": 7.015913963317871,
+ "step": 3119
+ },
+ {
+ "epoch": 43.3353711790393,
+ "grad_norm": 0.02441416308283806,
+ "learning_rate": 0.0006,
+ "loss": 7.061622142791748,
+ "step": 3120
+ },
+ {
+ "epoch": 43.34934497816594,
+ "grad_norm": 0.031887322664260864,
+ "learning_rate": 0.0006,
+ "loss": 7.046883583068848,
+ "step": 3121
+ },
+ {
+ "epoch": 43.36331877729258,
+ "grad_norm": 0.021930739283561707,
+ "learning_rate": 0.0006,
+ "loss": 7.0309624671936035,
+ "step": 3122
+ },
+ {
+ "epoch": 43.377292576419215,
+ "grad_norm": 0.02722783200442791,
+ "learning_rate": 0.0006,
+ "loss": 7.072924613952637,
+ "step": 3123
+ },
+ {
+ "epoch": 43.391266375545854,
+ "grad_norm": 0.01483179721981287,
+ "learning_rate": 0.0006,
+ "loss": 7.028624534606934,
+ "step": 3124
+ },
+ {
+ "epoch": 43.40524017467249,
+ "grad_norm": 0.024793945252895355,
+ "learning_rate": 0.0006,
+ "loss": 7.001623153686523,
+ "step": 3125
+ },
+ {
+ "epoch": 43.419213973799124,
+ "grad_norm": 0.020148713141679764,
+ "learning_rate": 0.0006,
+ "loss": 6.9938812255859375,
+ "step": 3126
+ },
+ {
+ "epoch": 43.43318777292576,
+ "grad_norm": 0.02413850650191307,
+ "learning_rate": 0.0006,
+ "loss": 7.001243591308594,
+ "step": 3127
+ },
+ {
+ "epoch": 43.4471615720524,
+ "grad_norm": 0.028549203649163246,
+ "learning_rate": 0.0006,
+ "loss": 6.987671375274658,
+ "step": 3128
+ },
+ {
+ "epoch": 43.46113537117904,
+ "grad_norm": 0.025988785549998283,
+ "learning_rate": 0.0006,
+ "loss": 6.9551167488098145,
+ "step": 3129
+ },
+ {
+ "epoch": 43.47510917030568,
+ "grad_norm": 0.027140729129314423,
+ "learning_rate": 0.0006,
+ "loss": 6.9547810554504395,
+ "step": 3130
+ },
+ {
+ "epoch": 43.48908296943232,
+ "grad_norm": 0.030130082741379738,
+ "learning_rate": 0.0006,
+ "loss": 6.993186950683594,
+ "step": 3131
+ },
+ {
+ "epoch": 43.50305676855895,
+ "grad_norm": 0.025996994227170944,
+ "learning_rate": 0.0006,
+ "loss": 6.984624862670898,
+ "step": 3132
+ },
+ {
+ "epoch": 43.51703056768559,
+ "grad_norm": 0.02423008903861046,
+ "learning_rate": 0.0006,
+ "loss": 6.98813009262085,
+ "step": 3133
+ },
+ {
+ "epoch": 43.531004366812226,
+ "grad_norm": 0.029665837064385414,
+ "learning_rate": 0.0006,
+ "loss": 6.986610412597656,
+ "step": 3134
+ },
+ {
+ "epoch": 43.544978165938865,
+ "grad_norm": 0.028768595308065414,
+ "learning_rate": 0.0006,
+ "loss": 6.898797035217285,
+ "step": 3135
+ },
+ {
+ "epoch": 43.5589519650655,
+ "grad_norm": 0.03111208975315094,
+ "learning_rate": 0.0006,
+ "loss": 6.923741817474365,
+ "step": 3136
+ },
+ {
+ "epoch": 43.57292576419214,
+ "grad_norm": 0.0368187390267849,
+ "learning_rate": 0.0006,
+ "loss": 6.994343280792236,
+ "step": 3137
+ },
+ {
+ "epoch": 43.58689956331878,
+ "grad_norm": 0.04076763987541199,
+ "learning_rate": 0.0006,
+ "loss": 6.909458160400391,
+ "step": 3138
+ },
+ {
+ "epoch": 43.60087336244541,
+ "grad_norm": 0.03281061723828316,
+ "learning_rate": 0.0006,
+ "loss": 6.924851417541504,
+ "step": 3139
+ },
+ {
+ "epoch": 43.61484716157205,
+ "grad_norm": 0.025486988946795464,
+ "learning_rate": 0.0006,
+ "loss": 6.962926864624023,
+ "step": 3140
+ },
+ {
+ "epoch": 43.62882096069869,
+ "grad_norm": 0.04583485797047615,
+ "learning_rate": 0.0006,
+ "loss": 6.876555442810059,
+ "step": 3141
+ },
+ {
+ "epoch": 43.64279475982533,
+ "grad_norm": 0.05460785701870918,
+ "learning_rate": 0.0006,
+ "loss": 6.952932357788086,
+ "step": 3142
+ },
+ {
+ "epoch": 43.65676855895197,
+ "grad_norm": 0.040917906910181046,
+ "learning_rate": 0.0006,
+ "loss": 6.9624924659729,
+ "step": 3143
+ },
+ {
+ "epoch": 43.670742358078606,
+ "grad_norm": 0.04009668901562691,
+ "learning_rate": 0.0006,
+ "loss": 6.94033145904541,
+ "step": 3144
+ },
+ {
+ "epoch": 43.68471615720524,
+ "grad_norm": 0.04500625655055046,
+ "learning_rate": 0.0006,
+ "loss": 6.961839199066162,
+ "step": 3145
+ },
+ {
+ "epoch": 43.698689956331876,
+ "grad_norm": 0.03182258829474449,
+ "learning_rate": 0.0006,
+ "loss": 6.930580139160156,
+ "step": 3146
+ },
+ {
+ "epoch": 43.712663755458514,
+ "grad_norm": 0.053621504455804825,
+ "learning_rate": 0.0006,
+ "loss": 6.950235366821289,
+ "step": 3147
+ },
+ {
+ "epoch": 43.72663755458515,
+ "grad_norm": 0.031919289380311966,
+ "learning_rate": 0.0006,
+ "loss": 6.895111560821533,
+ "step": 3148
+ },
+ {
+ "epoch": 43.74061135371179,
+ "grad_norm": 0.042713891714811325,
+ "learning_rate": 0.0006,
+ "loss": 6.940885543823242,
+ "step": 3149
+ },
+ {
+ "epoch": 43.75458515283843,
+ "grad_norm": 0.02385653182864189,
+ "learning_rate": 0.0006,
+ "loss": 6.948801517486572,
+ "step": 3150
+ },
+ {
+ "epoch": 43.76855895196506,
+ "grad_norm": 0.04100663587450981,
+ "learning_rate": 0.0006,
+ "loss": 6.970411777496338,
+ "step": 3151
+ },
+ {
+ "epoch": 43.7825327510917,
+ "grad_norm": 0.02729583904147148,
+ "learning_rate": 0.0006,
+ "loss": 6.899379253387451,
+ "step": 3152
+ },
+ {
+ "epoch": 43.79650655021834,
+ "grad_norm": 0.03128620237112045,
+ "learning_rate": 0.0006,
+ "loss": 6.931319236755371,
+ "step": 3153
+ },
+ {
+ "epoch": 43.81048034934498,
+ "grad_norm": 0.025488659739494324,
+ "learning_rate": 0.0006,
+ "loss": 6.8486409187316895,
+ "step": 3154
+ },
+ {
+ "epoch": 43.82445414847162,
+ "grad_norm": 0.03679376840591431,
+ "learning_rate": 0.0006,
+ "loss": 6.840811252593994,
+ "step": 3155
+ },
+ {
+ "epoch": 43.838427947598255,
+ "grad_norm": 0.020578034222126007,
+ "learning_rate": 0.0006,
+ "loss": 6.9088616371154785,
+ "step": 3156
+ },
+ {
+ "epoch": 43.852401746724894,
+ "grad_norm": 0.02935841865837574,
+ "learning_rate": 0.0006,
+ "loss": 6.806676864624023,
+ "step": 3157
+ },
+ {
+ "epoch": 43.866375545851525,
+ "grad_norm": 0.023319244384765625,
+ "learning_rate": 0.0006,
+ "loss": 6.878457546234131,
+ "step": 3158
+ },
+ {
+ "epoch": 43.880349344978164,
+ "grad_norm": 0.02779589593410492,
+ "learning_rate": 0.0006,
+ "loss": 6.822447776794434,
+ "step": 3159
+ },
+ {
+ "epoch": 43.8943231441048,
+ "grad_norm": 0.01879825070500374,
+ "learning_rate": 0.0006,
+ "loss": 6.854121208190918,
+ "step": 3160
+ },
+ {
+ "epoch": 43.90829694323144,
+ "grad_norm": 0.023286139592528343,
+ "learning_rate": 0.0006,
+ "loss": 6.836827278137207,
+ "step": 3161
+ },
+ {
+ "epoch": 43.92227074235808,
+ "grad_norm": 0.030358297750353813,
+ "learning_rate": 0.0006,
+ "loss": 6.860037803649902,
+ "step": 3162
+ },
+ {
+ "epoch": 43.93624454148472,
+ "grad_norm": 0.029621105641126633,
+ "learning_rate": 0.0006,
+ "loss": 6.884025573730469,
+ "step": 3163
+ },
+ {
+ "epoch": 43.95021834061135,
+ "grad_norm": 0.029829824343323708,
+ "learning_rate": 0.0006,
+ "loss": 6.8702392578125,
+ "step": 3164
+ },
+ {
+ "epoch": 43.96419213973799,
+ "grad_norm": 0.036892473697662354,
+ "learning_rate": 0.0006,
+ "loss": 6.806652069091797,
+ "step": 3165
+ },
+ {
+ "epoch": 43.97816593886463,
+ "grad_norm": 0.03206709027290344,
+ "learning_rate": 0.0006,
+ "loss": 6.828495979309082,
+ "step": 3166
+ },
+ {
+ "epoch": 43.992139737991266,
+ "grad_norm": 0.02744000405073166,
+ "learning_rate": 0.0006,
+ "loss": 6.832002639770508,
+ "step": 3167
+ },
+ {
+ "epoch": 44.0,
+ "grad_norm": 0.02205660194158554,
+ "learning_rate": 0.0006,
+ "loss": 6.898828506469727,
+ "step": 3168
+ },
+ {
+ "epoch": 44.0,
+ "eval_loss": 6.8150553703308105,
+ "eval_runtime": 56.784,
+ "eval_samples_per_second": 43.005,
+ "eval_steps_per_second": 1.356,
+ "step": 3168
+ },
+ {
+ "epoch": 44.01397379912664,
+ "grad_norm": 0.030421260744333267,
+ "learning_rate": 0.0006,
+ "loss": 6.775057792663574,
+ "step": 3169
+ },
+ {
+ "epoch": 44.02794759825328,
+ "grad_norm": 0.0465611033141613,
+ "learning_rate": 0.0006,
+ "loss": 6.849390506744385,
+ "step": 3170
+ },
+ {
+ "epoch": 44.041921397379916,
+ "grad_norm": 0.05977972596883774,
+ "learning_rate": 0.0006,
+ "loss": 6.725576400756836,
+ "step": 3171
+ },
+ {
+ "epoch": 44.05589519650655,
+ "grad_norm": 0.05432697385549545,
+ "learning_rate": 0.0006,
+ "loss": 6.768024444580078,
+ "step": 3172
+ },
+ {
+ "epoch": 44.069868995633186,
+ "grad_norm": 0.04850967228412628,
+ "learning_rate": 0.0006,
+ "loss": 6.799457550048828,
+ "step": 3173
+ },
+ {
+ "epoch": 44.083842794759825,
+ "grad_norm": 0.06201786920428276,
+ "learning_rate": 0.0006,
+ "loss": 6.897234916687012,
+ "step": 3174
+ },
+ {
+ "epoch": 44.09781659388646,
+ "grad_norm": 0.04695591703057289,
+ "learning_rate": 0.0006,
+ "loss": 6.826470375061035,
+ "step": 3175
+ },
+ {
+ "epoch": 44.1117903930131,
+ "grad_norm": 0.03850164636969566,
+ "learning_rate": 0.0006,
+ "loss": 6.835901260375977,
+ "step": 3176
+ },
+ {
+ "epoch": 44.12576419213974,
+ "grad_norm": 0.04250155761837959,
+ "learning_rate": 0.0006,
+ "loss": 6.782310485839844,
+ "step": 3177
+ },
+ {
+ "epoch": 44.13973799126637,
+ "grad_norm": 0.041465479880571365,
+ "learning_rate": 0.0006,
+ "loss": 6.785732746124268,
+ "step": 3178
+ },
+ {
+ "epoch": 44.15371179039301,
+ "grad_norm": 0.02527807280421257,
+ "learning_rate": 0.0006,
+ "loss": 6.78061580657959,
+ "step": 3179
+ },
+ {
+ "epoch": 44.16768558951965,
+ "grad_norm": 0.03660374879837036,
+ "learning_rate": 0.0006,
+ "loss": 6.783992767333984,
+ "step": 3180
+ },
+ {
+ "epoch": 44.18165938864629,
+ "grad_norm": 0.027567964047193527,
+ "learning_rate": 0.0006,
+ "loss": 6.765413284301758,
+ "step": 3181
+ },
+ {
+ "epoch": 44.19563318777293,
+ "grad_norm": 0.024301424622535706,
+ "learning_rate": 0.0006,
+ "loss": 6.760368347167969,
+ "step": 3182
+ },
+ {
+ "epoch": 44.209606986899566,
+ "grad_norm": 0.0266821701079607,
+ "learning_rate": 0.0006,
+ "loss": 6.71417236328125,
+ "step": 3183
+ },
+ {
+ "epoch": 44.223580786026204,
+ "grad_norm": 0.028125090524554253,
+ "learning_rate": 0.0006,
+ "loss": 6.646126747131348,
+ "step": 3184
+ },
+ {
+ "epoch": 44.237554585152836,
+ "grad_norm": 0.0185737032443285,
+ "learning_rate": 0.0006,
+ "loss": 6.712944030761719,
+ "step": 3185
+ },
+ {
+ "epoch": 44.251528384279474,
+ "grad_norm": 0.025836674496531487,
+ "learning_rate": 0.0006,
+ "loss": 6.707030296325684,
+ "step": 3186
+ },
+ {
+ "epoch": 44.26550218340611,
+ "grad_norm": 0.029744870960712433,
+ "learning_rate": 0.0006,
+ "loss": 6.6271071434021,
+ "step": 3187
+ },
+ {
+ "epoch": 44.27947598253275,
+ "grad_norm": 0.03021426685154438,
+ "learning_rate": 0.0006,
+ "loss": 6.677826404571533,
+ "step": 3188
+ },
+ {
+ "epoch": 44.29344978165939,
+ "grad_norm": 0.032727014273405075,
+ "learning_rate": 0.0006,
+ "loss": 6.657764434814453,
+ "step": 3189
+ },
+ {
+ "epoch": 44.30742358078603,
+ "grad_norm": 0.03696778789162636,
+ "learning_rate": 0.0006,
+ "loss": 6.632883548736572,
+ "step": 3190
+ },
+ {
+ "epoch": 44.32139737991266,
+ "grad_norm": 0.034630827605724335,
+ "learning_rate": 0.0006,
+ "loss": 6.759583473205566,
+ "step": 3191
+ },
+ {
+ "epoch": 44.3353711790393,
+ "grad_norm": 0.030517758801579475,
+ "learning_rate": 0.0006,
+ "loss": 6.702268600463867,
+ "step": 3192
+ },
+ {
+ "epoch": 44.34934497816594,
+ "grad_norm": 0.04105190932750702,
+ "learning_rate": 0.0006,
+ "loss": 6.705792427062988,
+ "step": 3193
+ },
+ {
+ "epoch": 44.36331877729258,
+ "grad_norm": 0.06659174710512161,
+ "learning_rate": 0.0006,
+ "loss": 6.73862361907959,
+ "step": 3194
+ },
+ {
+ "epoch": 44.377292576419215,
+ "grad_norm": 0.08013619482517242,
+ "learning_rate": 0.0006,
+ "loss": 6.710282802581787,
+ "step": 3195
+ },
+ {
+ "epoch": 44.391266375545854,
+ "grad_norm": 0.049850963056087494,
+ "learning_rate": 0.0006,
+ "loss": 6.684974670410156,
+ "step": 3196
+ },
+ {
+ "epoch": 44.40524017467249,
+ "grad_norm": 0.061247408390045166,
+ "learning_rate": 0.0006,
+ "loss": 6.716034412384033,
+ "step": 3197
+ },
+ {
+ "epoch": 44.419213973799124,
+ "grad_norm": 0.05868857726454735,
+ "learning_rate": 0.0006,
+ "loss": 6.589107513427734,
+ "step": 3198
+ },
+ {
+ "epoch": 44.43318777292576,
+ "grad_norm": 0.048878561705350876,
+ "learning_rate": 0.0006,
+ "loss": 6.694855690002441,
+ "step": 3199
+ },
+ {
+ "epoch": 44.4471615720524,
+ "grad_norm": 0.028748324140906334,
+ "learning_rate": 0.0006,
+ "loss": 6.629969596862793,
+ "step": 3200
+ },
+ {
+ "epoch": 44.46113537117904,
+ "grad_norm": 0.04060306400060654,
+ "learning_rate": 0.0006,
+ "loss": 6.721431255340576,
+ "step": 3201
+ },
+ {
+ "epoch": 44.47510917030568,
+ "grad_norm": 0.03833824768662453,
+ "learning_rate": 0.0006,
+ "loss": 6.625029563903809,
+ "step": 3202
+ },
+ {
+ "epoch": 44.48908296943232,
+ "grad_norm": 0.03791997581720352,
+ "learning_rate": 0.0006,
+ "loss": 6.669187545776367,
+ "step": 3203
+ },
+ {
+ "epoch": 44.50305676855895,
+ "grad_norm": 0.03259684890508652,
+ "learning_rate": 0.0006,
+ "loss": 6.665643692016602,
+ "step": 3204
+ },
+ {
+ "epoch": 44.51703056768559,
+ "grad_norm": 0.040658626705408096,
+ "learning_rate": 0.0006,
+ "loss": 6.639796257019043,
+ "step": 3205
+ },
+ {
+ "epoch": 44.531004366812226,
+ "grad_norm": 0.03691193461418152,
+ "learning_rate": 0.0006,
+ "loss": 6.692244529724121,
+ "step": 3206
+ },
+ {
+ "epoch": 44.544978165938865,
+ "grad_norm": 0.02814812771975994,
+ "learning_rate": 0.0006,
+ "loss": 6.61978006362915,
+ "step": 3207
+ },
+ {
+ "epoch": 44.5589519650655,
+ "grad_norm": 0.022155074402689934,
+ "learning_rate": 0.0006,
+ "loss": 6.625905990600586,
+ "step": 3208
+ },
+ {
+ "epoch": 44.57292576419214,
+ "grad_norm": 0.030794885009527206,
+ "learning_rate": 0.0006,
+ "loss": 6.678500652313232,
+ "step": 3209
+ },
+ {
+ "epoch": 44.58689956331878,
+ "grad_norm": 0.03201768547296524,
+ "learning_rate": 0.0006,
+ "loss": 6.6015729904174805,
+ "step": 3210
+ },
+ {
+ "epoch": 44.60087336244541,
+ "grad_norm": 0.025138361379504204,
+ "learning_rate": 0.0006,
+ "loss": 6.568234443664551,
+ "step": 3211
+ },
+ {
+ "epoch": 44.61484716157205,
+ "grad_norm": 0.02969353087246418,
+ "learning_rate": 0.0006,
+ "loss": 6.588366508483887,
+ "step": 3212
+ },
+ {
+ "epoch": 44.62882096069869,
+ "grad_norm": 0.03136962279677391,
+ "learning_rate": 0.0006,
+ "loss": 6.479961395263672,
+ "step": 3213
+ },
+ {
+ "epoch": 44.64279475982533,
+ "grad_norm": 0.030178170651197433,
+ "learning_rate": 0.0006,
+ "loss": 6.622910976409912,
+ "step": 3214
+ },
+ {
+ "epoch": 44.65676855895197,
+ "grad_norm": 0.02682245895266533,
+ "learning_rate": 0.0006,
+ "loss": 6.597707748413086,
+ "step": 3215
+ },
+ {
+ "epoch": 44.670742358078606,
+ "grad_norm": 0.025154557079076767,
+ "learning_rate": 0.0006,
+ "loss": 6.630541801452637,
+ "step": 3216
+ },
+ {
+ "epoch": 44.68471615720524,
+ "grad_norm": 0.03564463183283806,
+ "learning_rate": 0.0006,
+ "loss": 6.542823791503906,
+ "step": 3217
+ },
+ {
+ "epoch": 44.698689956331876,
+ "grad_norm": 0.049477651715278625,
+ "learning_rate": 0.0006,
+ "loss": 6.646030426025391,
+ "step": 3218
+ },
+ {
+ "epoch": 44.712663755458514,
+ "grad_norm": 0.06536535918712616,
+ "learning_rate": 0.0006,
+ "loss": 6.595463275909424,
+ "step": 3219
+ },
+ {
+ "epoch": 44.72663755458515,
+ "grad_norm": 0.07710316777229309,
+ "learning_rate": 0.0006,
+ "loss": 6.55524206161499,
+ "step": 3220
+ },
+ {
+ "epoch": 44.74061135371179,
+ "grad_norm": 0.09571163356304169,
+ "learning_rate": 0.0006,
+ "loss": 6.659764289855957,
+ "step": 3221
+ },
+ {
+ "epoch": 44.75458515283843,
+ "grad_norm": 0.055076830089092255,
+ "learning_rate": 0.0006,
+ "loss": 6.658882141113281,
+ "step": 3222
+ },
+ {
+ "epoch": 44.76855895196506,
+ "grad_norm": 0.07671099156141281,
+ "learning_rate": 0.0006,
+ "loss": 6.605809211730957,
+ "step": 3223
+ },
+ {
+ "epoch": 44.7825327510917,
+ "grad_norm": 0.05295584350824356,
+ "learning_rate": 0.0006,
+ "loss": 6.641455173492432,
+ "step": 3224
+ },
+ {
+ "epoch": 44.79650655021834,
+ "grad_norm": 0.09128042310476303,
+ "learning_rate": 0.0006,
+ "loss": 6.691066741943359,
+ "step": 3225
+ },
+ {
+ "epoch": 44.81048034934498,
+ "grad_norm": 0.05362958461046219,
+ "learning_rate": 0.0006,
+ "loss": 6.573406219482422,
+ "step": 3226
+ },
+ {
+ "epoch": 44.82445414847162,
+ "grad_norm": 0.046910062432289124,
+ "learning_rate": 0.0006,
+ "loss": 6.550876617431641,
+ "step": 3227
+ },
+ {
+ "epoch": 44.838427947598255,
+ "grad_norm": 0.03494197130203247,
+ "learning_rate": 0.0006,
+ "loss": 6.540179252624512,
+ "step": 3228
+ },
+ {
+ "epoch": 44.852401746724894,
+ "grad_norm": 0.04716387391090393,
+ "learning_rate": 0.0006,
+ "loss": 6.490459442138672,
+ "step": 3229
+ },
+ {
+ "epoch": 44.866375545851525,
+ "grad_norm": 0.04190506786108017,
+ "learning_rate": 0.0006,
+ "loss": 6.509925365447998,
+ "step": 3230
+ },
+ {
+ "epoch": 44.880349344978164,
+ "grad_norm": 0.04749064892530441,
+ "learning_rate": 0.0006,
+ "loss": 6.478779315948486,
+ "step": 3231
+ },
+ {
+ "epoch": 44.8943231441048,
+ "grad_norm": 0.028002185747027397,
+ "learning_rate": 0.0006,
+ "loss": 6.528904914855957,
+ "step": 3232
+ },
+ {
+ "epoch": 44.90829694323144,
+ "grad_norm": 0.036473166197538376,
+ "learning_rate": 0.0006,
+ "loss": 6.549042701721191,
+ "step": 3233
+ },
+ {
+ "epoch": 44.92227074235808,
+ "grad_norm": 0.02543315477669239,
+ "learning_rate": 0.0006,
+ "loss": 6.498875141143799,
+ "step": 3234
+ },
+ {
+ "epoch": 44.93624454148472,
+ "grad_norm": 0.029906807467341423,
+ "learning_rate": 0.0006,
+ "loss": 6.453019142150879,
+ "step": 3235
+ },
+ {
+ "epoch": 44.95021834061135,
+ "grad_norm": 0.027231857180595398,
+ "learning_rate": 0.0006,
+ "loss": 6.428829193115234,
+ "step": 3236
+ },
+ {
+ "epoch": 44.96419213973799,
+ "grad_norm": 0.02506154775619507,
+ "learning_rate": 0.0006,
+ "loss": 6.420858383178711,
+ "step": 3237
+ },
+ {
+ "epoch": 44.97816593886463,
+ "grad_norm": 0.0269472673535347,
+ "learning_rate": 0.0006,
+ "loss": 6.484989643096924,
+ "step": 3238
+ },
+ {
+ "epoch": 44.992139737991266,
+ "grad_norm": 0.023410487920045853,
+ "learning_rate": 0.0006,
+ "loss": 6.454830169677734,
+ "step": 3239
+ },
+ {
+ "epoch": 45.0,
+ "grad_norm": 0.034175023436546326,
+ "learning_rate": 0.0006,
+ "loss": 6.430975437164307,
+ "step": 3240
+ },
+ {
+ "epoch": 45.0,
+ "eval_loss": 6.458896160125732,
+ "eval_runtime": 56.6934,
+ "eval_samples_per_second": 43.074,
+ "eval_steps_per_second": 1.358,
+ "step": 3240
+ },
+ {
+ "epoch": 45.01397379912664,
+ "grad_norm": 0.045031629502773285,
+ "learning_rate": 0.0006,
+ "loss": 6.4243011474609375,
+ "step": 3241
+ },
+ {
+ "epoch": 45.02794759825328,
+ "grad_norm": 0.04968734085559845,
+ "learning_rate": 0.0006,
+ "loss": 6.380091667175293,
+ "step": 3242
+ },
+ {
+ "epoch": 45.041921397379916,
+ "grad_norm": 0.047145187854766846,
+ "learning_rate": 0.0006,
+ "loss": 6.477178573608398,
+ "step": 3243
+ },
+ {
+ "epoch": 45.05589519650655,
+ "grad_norm": 0.04304948449134827,
+ "learning_rate": 0.0006,
+ "loss": 6.399726867675781,
+ "step": 3244
+ },
+ {
+ "epoch": 45.069868995633186,
+ "grad_norm": 0.04466010630130768,
+ "learning_rate": 0.0006,
+ "loss": 6.3415679931640625,
+ "step": 3245
+ },
+ {
+ "epoch": 45.083842794759825,
+ "grad_norm": 0.04305306077003479,
+ "learning_rate": 0.0006,
+ "loss": 6.3413310050964355,
+ "step": 3246
+ },
+ {
+ "epoch": 45.09781659388646,
+ "grad_norm": 0.043064575642347336,
+ "learning_rate": 0.0006,
+ "loss": 6.4838714599609375,
+ "step": 3247
+ },
+ {
+ "epoch": 45.1117903930131,
+ "grad_norm": 0.049503277987241745,
+ "learning_rate": 0.0006,
+ "loss": 6.38473653793335,
+ "step": 3248
+ },
+ {
+ "epoch": 45.12576419213974,
+ "grad_norm": 0.04613238573074341,
+ "learning_rate": 0.0006,
+ "loss": 6.370769500732422,
+ "step": 3249
+ },
+ {
+ "epoch": 45.13973799126637,
+ "grad_norm": 0.052119478583335876,
+ "learning_rate": 0.0006,
+ "loss": 6.362321853637695,
+ "step": 3250
+ },
+ {
+ "epoch": 45.15371179039301,
+ "grad_norm": 0.045110076665878296,
+ "learning_rate": 0.0006,
+ "loss": 6.369844913482666,
+ "step": 3251
+ },
+ {
+ "epoch": 45.16768558951965,
+ "grad_norm": 0.03798583149909973,
+ "learning_rate": 0.0006,
+ "loss": 6.324633598327637,
+ "step": 3252
+ },
+ {
+ "epoch": 45.18165938864629,
+ "grad_norm": 0.03744659945368767,
+ "learning_rate": 0.0006,
+ "loss": 6.349421501159668,
+ "step": 3253
+ },
+ {
+ "epoch": 45.19563318777293,
+ "grad_norm": 0.03767292574048042,
+ "learning_rate": 0.0006,
+ "loss": 6.334238052368164,
+ "step": 3254
+ },
+ {
+ "epoch": 45.209606986899566,
+ "grad_norm": 0.03577509894967079,
+ "learning_rate": 0.0006,
+ "loss": 6.3453826904296875,
+ "step": 3255
+ },
+ {
+ "epoch": 45.223580786026204,
+ "grad_norm": 0.03403240442276001,
+ "learning_rate": 0.0006,
+ "loss": 6.2836174964904785,
+ "step": 3256
+ },
+ {
+ "epoch": 45.237554585152836,
+ "grad_norm": 0.03517331928014755,
+ "learning_rate": 0.0006,
+ "loss": 6.281718730926514,
+ "step": 3257
+ },
+ {
+ "epoch": 45.251528384279474,
+ "grad_norm": 0.04409139230847359,
+ "learning_rate": 0.0006,
+ "loss": 6.320815086364746,
+ "step": 3258
+ },
+ {
+ "epoch": 45.26550218340611,
+ "grad_norm": 0.054750364273786545,
+ "learning_rate": 0.0006,
+ "loss": 6.343618392944336,
+ "step": 3259
+ },
+ {
+ "epoch": 45.27947598253275,
+ "grad_norm": 0.07612067461013794,
+ "learning_rate": 0.0006,
+ "loss": 6.310614585876465,
+ "step": 3260
+ },
+ {
+ "epoch": 45.29344978165939,
+ "grad_norm": 0.08438985049724579,
+ "learning_rate": 0.0006,
+ "loss": 6.2369184494018555,
+ "step": 3261
+ },
+ {
+ "epoch": 45.30742358078603,
+ "grad_norm": 0.0782201811671257,
+ "learning_rate": 0.0006,
+ "loss": 6.254452705383301,
+ "step": 3262
+ },
+ {
+ "epoch": 45.32139737991266,
+ "grad_norm": 0.09928355365991592,
+ "learning_rate": 0.0006,
+ "loss": 6.3112077713012695,
+ "step": 3263
+ },
+ {
+ "epoch": 45.3353711790393,
+ "grad_norm": 0.07233133912086487,
+ "learning_rate": 0.0006,
+ "loss": 6.408087253570557,
+ "step": 3264
+ },
+ {
+ "epoch": 45.34934497816594,
+ "grad_norm": 0.07478684186935425,
+ "learning_rate": 0.0006,
+ "loss": 6.3770036697387695,
+ "step": 3265
+ },
+ {
+ "epoch": 45.36331877729258,
+ "grad_norm": 0.06036846712231636,
+ "learning_rate": 0.0006,
+ "loss": 6.33865213394165,
+ "step": 3266
+ },
+ {
+ "epoch": 45.377292576419215,
+ "grad_norm": 0.05783458054065704,
+ "learning_rate": 0.0006,
+ "loss": 6.363203048706055,
+ "step": 3267
+ },
+ {
+ "epoch": 45.391266375545854,
+ "grad_norm": 0.05443909019231796,
+ "learning_rate": 0.0006,
+ "loss": 6.287069797515869,
+ "step": 3268
+ },
+ {
+ "epoch": 45.40524017467249,
+ "grad_norm": 0.0557909794151783,
+ "learning_rate": 0.0006,
+ "loss": 6.297784805297852,
+ "step": 3269
+ },
+ {
+ "epoch": 45.419213973799124,
+ "grad_norm": 0.03483372926712036,
+ "learning_rate": 0.0006,
+ "loss": 6.280734062194824,
+ "step": 3270
+ },
+ {
+ "epoch": 45.43318777292576,
+ "grad_norm": 0.0425543375313282,
+ "learning_rate": 0.0006,
+ "loss": 6.33565616607666,
+ "step": 3271
+ },
+ {
+ "epoch": 45.4471615720524,
+ "grad_norm": 0.03230159357190132,
+ "learning_rate": 0.0006,
+ "loss": 6.197698593139648,
+ "step": 3272
+ },
+ {
+ "epoch": 45.46113537117904,
+ "grad_norm": 0.036897290498018265,
+ "learning_rate": 0.0006,
+ "loss": 6.241239547729492,
+ "step": 3273
+ },
+ {
+ "epoch": 45.47510917030568,
+ "grad_norm": 0.031103838235139847,
+ "learning_rate": 0.0006,
+ "loss": 6.159435749053955,
+ "step": 3274
+ },
+ {
+ "epoch": 45.48908296943232,
+ "grad_norm": 0.027116449549794197,
+ "learning_rate": 0.0006,
+ "loss": 6.212225914001465,
+ "step": 3275
+ },
+ {
+ "epoch": 45.50305676855895,
+ "grad_norm": 0.029923981055617332,
+ "learning_rate": 0.0006,
+ "loss": 6.157269477844238,
+ "step": 3276
+ },
+ {
+ "epoch": 45.51703056768559,
+ "grad_norm": 0.02521538734436035,
+ "learning_rate": 0.0006,
+ "loss": 6.146918773651123,
+ "step": 3277
+ },
+ {
+ "epoch": 45.531004366812226,
+ "grad_norm": 0.02758920192718506,
+ "learning_rate": 0.0006,
+ "loss": 6.0446977615356445,
+ "step": 3278
+ },
+ {
+ "epoch": 45.544978165938865,
+ "grad_norm": 0.024328220635652542,
+ "learning_rate": 0.0006,
+ "loss": 6.099508285522461,
+ "step": 3279
+ },
+ {
+ "epoch": 45.5589519650655,
+ "grad_norm": 0.027425771579146385,
+ "learning_rate": 0.0006,
+ "loss": 6.084299087524414,
+ "step": 3280
+ },
+ {
+ "epoch": 45.57292576419214,
+ "grad_norm": 0.026927651837468147,
+ "learning_rate": 0.0006,
+ "loss": 6.146334171295166,
+ "step": 3281
+ },
+ {
+ "epoch": 45.58689956331878,
+ "grad_norm": 0.04054379090666771,
+ "learning_rate": 0.0006,
+ "loss": 6.1301493644714355,
+ "step": 3282
+ },
+ {
+ "epoch": 45.60087336244541,
+ "grad_norm": 0.04456290975213051,
+ "learning_rate": 0.0006,
+ "loss": 6.152511119842529,
+ "step": 3283
+ },
+ {
+ "epoch": 45.61484716157205,
+ "grad_norm": 0.052821505814790726,
+ "learning_rate": 0.0006,
+ "loss": 6.113544940948486,
+ "step": 3284
+ },
+ {
+ "epoch": 45.62882096069869,
+ "grad_norm": 0.07251739501953125,
+ "learning_rate": 0.0006,
+ "loss": 6.118298530578613,
+ "step": 3285
+ },
+ {
+ "epoch": 45.64279475982533,
+ "grad_norm": 0.09433556348085403,
+ "learning_rate": 0.0006,
+ "loss": 6.170897483825684,
+ "step": 3286
+ },
+ {
+ "epoch": 45.65676855895197,
+ "grad_norm": 0.057683493942022324,
+ "learning_rate": 0.0006,
+ "loss": 6.1747822761535645,
+ "step": 3287
+ },
+ {
+ "epoch": 45.670742358078606,
+ "grad_norm": 0.06363270431756973,
+ "learning_rate": 0.0006,
+ "loss": 6.1617889404296875,
+ "step": 3288
+ },
+ {
+ "epoch": 45.68471615720524,
+ "grad_norm": 0.07528169453144073,
+ "learning_rate": 0.0006,
+ "loss": 6.230060577392578,
+ "step": 3289
+ },
+ {
+ "epoch": 45.698689956331876,
+ "grad_norm": 0.07816170156002045,
+ "learning_rate": 0.0006,
+ "loss": 6.171257019042969,
+ "step": 3290
+ },
+ {
+ "epoch": 45.712663755458514,
+ "grad_norm": 0.05860164389014244,
+ "learning_rate": 0.0006,
+ "loss": 6.095484256744385,
+ "step": 3291
+ },
+ {
+ "epoch": 45.72663755458515,
+ "grad_norm": 0.04747429117560387,
+ "learning_rate": 0.0006,
+ "loss": 6.060067176818848,
+ "step": 3292
+ },
+ {
+ "epoch": 45.74061135371179,
+ "grad_norm": 0.04514474794268608,
+ "learning_rate": 0.0006,
+ "loss": 6.100218772888184,
+ "step": 3293
+ },
+ {
+ "epoch": 45.75458515283843,
+ "grad_norm": 0.05168681964278221,
+ "learning_rate": 0.0006,
+ "loss": 6.0621867179870605,
+ "step": 3294
+ },
+ {
+ "epoch": 45.76855895196506,
+ "grad_norm": 0.04665876924991608,
+ "learning_rate": 0.0006,
+ "loss": 6.118292808532715,
+ "step": 3295
+ },
+ {
+ "epoch": 45.7825327510917,
+ "grad_norm": 0.049374762922525406,
+ "learning_rate": 0.0006,
+ "loss": 6.098394393920898,
+ "step": 3296
+ },
+ {
+ "epoch": 45.79650655021834,
+ "grad_norm": 0.03793549910187721,
+ "learning_rate": 0.0006,
+ "loss": 6.069541931152344,
+ "step": 3297
+ },
+ {
+ "epoch": 45.81048034934498,
+ "grad_norm": 0.036362383514642715,
+ "learning_rate": 0.0006,
+ "loss": 6.1188435554504395,
+ "step": 3298
+ },
+ {
+ "epoch": 45.82445414847162,
+ "grad_norm": 0.03397097811102867,
+ "learning_rate": 0.0006,
+ "loss": 6.117589950561523,
+ "step": 3299
+ },
+ {
+ "epoch": 45.838427947598255,
+ "grad_norm": 0.030977753922343254,
+ "learning_rate": 0.0006,
+ "loss": 6.0236711502075195,
+ "step": 3300
+ },
+ {
+ "epoch": 45.852401746724894,
+ "grad_norm": 0.02800217643380165,
+ "learning_rate": 0.0006,
+ "loss": 6.034825325012207,
+ "step": 3301
+ },
+ {
+ "epoch": 45.866375545851525,
+ "grad_norm": 0.03129778802394867,
+ "learning_rate": 0.0006,
+ "loss": 5.980271339416504,
+ "step": 3302
+ },
+ {
+ "epoch": 45.880349344978164,
+ "grad_norm": 0.025886094197630882,
+ "learning_rate": 0.0006,
+ "loss": 5.966419219970703,
+ "step": 3303
+ },
+ {
+ "epoch": 45.8943231441048,
+ "grad_norm": 0.025712328031659126,
+ "learning_rate": 0.0006,
+ "loss": 5.938652515411377,
+ "step": 3304
+ },
+ {
+ "epoch": 45.90829694323144,
+ "grad_norm": 0.022508298978209496,
+ "learning_rate": 0.0006,
+ "loss": 5.982312202453613,
+ "step": 3305
+ },
+ {
+ "epoch": 45.92227074235808,
+ "grad_norm": 0.024880951270461082,
+ "learning_rate": 0.0006,
+ "loss": 5.931150436401367,
+ "step": 3306
+ },
+ {
+ "epoch": 45.93624454148472,
+ "grad_norm": 0.030007576569914818,
+ "learning_rate": 0.0006,
+ "loss": 5.936639308929443,
+ "step": 3307
+ },
+ {
+ "epoch": 45.95021834061135,
+ "grad_norm": 0.033314041793346405,
+ "learning_rate": 0.0006,
+ "loss": 5.944248676300049,
+ "step": 3308
+ },
+ {
+ "epoch": 45.96419213973799,
+ "grad_norm": 0.043078381568193436,
+ "learning_rate": 0.0006,
+ "loss": 5.971286773681641,
+ "step": 3309
+ },
+ {
+ "epoch": 45.97816593886463,
+ "grad_norm": 0.059296440333127975,
+ "learning_rate": 0.0006,
+ "loss": 5.910497665405273,
+ "step": 3310
+ },
+ {
+ "epoch": 45.992139737991266,
+ "grad_norm": 0.06232666224241257,
+ "learning_rate": 0.0006,
+ "loss": 5.929817199707031,
+ "step": 3311
+ },
+ {
+ "epoch": 46.0,
+ "grad_norm": 0.05303940549492836,
+ "learning_rate": 0.0006,
+ "loss": 5.910822868347168,
+ "step": 3312
+ },
+ {
+ "epoch": 46.0,
+ "eval_loss": 5.977851390838623,
+ "eval_runtime": 56.2435,
+ "eval_samples_per_second": 43.418,
+ "eval_steps_per_second": 1.369,
+ "step": 3312
+ },
+ {
+ "epoch": 46.01397379912664,
+ "grad_norm": 0.06943798810243607,
+ "learning_rate": 0.0006,
+ "loss": 5.854922294616699,
+ "step": 3313
+ },
+ {
+ "epoch": 46.02794759825328,
+ "grad_norm": 0.05236278846859932,
+ "learning_rate": 0.0006,
+ "loss": 5.876611232757568,
+ "step": 3314
+ },
+ {
+ "epoch": 46.041921397379916,
+ "grad_norm": 0.06240001320838928,
+ "learning_rate": 0.0006,
+ "loss": 5.878912448883057,
+ "step": 3315
+ },
+ {
+ "epoch": 46.05589519650655,
+ "grad_norm": 0.07401192933320999,
+ "learning_rate": 0.0006,
+ "loss": 5.985136032104492,
+ "step": 3316
+ },
+ {
+ "epoch": 46.069868995633186,
+ "grad_norm": 0.0826529711484909,
+ "learning_rate": 0.0006,
+ "loss": 5.939439296722412,
+ "step": 3317
+ },
+ {
+ "epoch": 46.083842794759825,
+ "grad_norm": 0.07999005168676376,
+ "learning_rate": 0.0006,
+ "loss": 5.982336044311523,
+ "step": 3318
+ },
+ {
+ "epoch": 46.09781659388646,
+ "grad_norm": 0.07014869153499603,
+ "learning_rate": 0.0006,
+ "loss": 5.896534442901611,
+ "step": 3319
+ },
+ {
+ "epoch": 46.1117903930131,
+ "grad_norm": 0.05795128643512726,
+ "learning_rate": 0.0006,
+ "loss": 5.843111991882324,
+ "step": 3320
+ },
+ {
+ "epoch": 46.12576419213974,
+ "grad_norm": 0.06374258548021317,
+ "learning_rate": 0.0006,
+ "loss": 5.78752326965332,
+ "step": 3321
+ },
+ {
+ "epoch": 46.13973799126637,
+ "grad_norm": 0.045524805784225464,
+ "learning_rate": 0.0006,
+ "loss": 5.86755895614624,
+ "step": 3322
+ },
+ {
+ "epoch": 46.15371179039301,
+ "grad_norm": 0.03773504123091698,
+ "learning_rate": 0.0006,
+ "loss": 5.83627462387085,
+ "step": 3323
+ },
+ {
+ "epoch": 46.16768558951965,
+ "grad_norm": 0.035815201699733734,
+ "learning_rate": 0.0006,
+ "loss": 5.798238277435303,
+ "step": 3324
+ },
+ {
+ "epoch": 46.18165938864629,
+ "grad_norm": 0.03756783902645111,
+ "learning_rate": 0.0006,
+ "loss": 5.845197677612305,
+ "step": 3325
+ },
+ {
+ "epoch": 46.19563318777293,
+ "grad_norm": 0.03569113463163376,
+ "learning_rate": 0.0006,
+ "loss": 5.888932704925537,
+ "step": 3326
+ },
+ {
+ "epoch": 46.209606986899566,
+ "grad_norm": 0.029467813670635223,
+ "learning_rate": 0.0006,
+ "loss": 5.792047500610352,
+ "step": 3327
+ },
+ {
+ "epoch": 46.223580786026204,
+ "grad_norm": 0.025698406621813774,
+ "learning_rate": 0.0006,
+ "loss": 5.783734321594238,
+ "step": 3328
+ },
+ {
+ "epoch": 46.237554585152836,
+ "grad_norm": 0.028426671400666237,
+ "learning_rate": 0.0006,
+ "loss": 5.762084007263184,
+ "step": 3329
+ },
+ {
+ "epoch": 46.251528384279474,
+ "grad_norm": 0.03099622018635273,
+ "learning_rate": 0.0006,
+ "loss": 5.719109058380127,
+ "step": 3330
+ },
+ {
+ "epoch": 46.26550218340611,
+ "grad_norm": 0.0310219656676054,
+ "learning_rate": 0.0006,
+ "loss": 5.737057209014893,
+ "step": 3331
+ },
+ {
+ "epoch": 46.27947598253275,
+ "grad_norm": 0.02713446132838726,
+ "learning_rate": 0.0006,
+ "loss": 5.758657455444336,
+ "step": 3332
+ },
+ {
+ "epoch": 46.29344978165939,
+ "grad_norm": 0.03099844604730606,
+ "learning_rate": 0.0006,
+ "loss": 5.742086410522461,
+ "step": 3333
+ },
+ {
+ "epoch": 46.30742358078603,
+ "grad_norm": 0.030258730053901672,
+ "learning_rate": 0.0006,
+ "loss": 5.767426490783691,
+ "step": 3334
+ },
+ {
+ "epoch": 46.32139737991266,
+ "grad_norm": 0.029034852981567383,
+ "learning_rate": 0.0006,
+ "loss": 5.651792049407959,
+ "step": 3335
+ },
+ {
+ "epoch": 46.3353711790393,
+ "grad_norm": 0.03138665854930878,
+ "learning_rate": 0.0006,
+ "loss": 5.664980888366699,
+ "step": 3336
+ },
+ {
+ "epoch": 46.34934497816594,
+ "grad_norm": 0.040546901524066925,
+ "learning_rate": 0.0006,
+ "loss": 5.665165901184082,
+ "step": 3337
+ },
+ {
+ "epoch": 46.36331877729258,
+ "grad_norm": 0.06395485252141953,
+ "learning_rate": 0.0006,
+ "loss": 5.7776713371276855,
+ "step": 3338
+ },
+ {
+ "epoch": 46.377292576419215,
+ "grad_norm": 0.10088056325912476,
+ "learning_rate": 0.0006,
+ "loss": 5.755645751953125,
+ "step": 3339
+ },
+ {
+ "epoch": 46.391266375545854,
+ "grad_norm": 0.09527938067913055,
+ "learning_rate": 0.0006,
+ "loss": 5.838809490203857,
+ "step": 3340
+ },
+ {
+ "epoch": 46.40524017467249,
+ "grad_norm": 0.0849648267030716,
+ "learning_rate": 0.0006,
+ "loss": 5.781932830810547,
+ "step": 3341
+ },
+ {
+ "epoch": 46.419213973799124,
+ "grad_norm": 0.10142666101455688,
+ "learning_rate": 0.0006,
+ "loss": 5.838970184326172,
+ "step": 3342
+ },
+ {
+ "epoch": 46.43318777292576,
+ "grad_norm": 0.42731454968452454,
+ "learning_rate": 0.0006,
+ "loss": 6.071902275085449,
+ "step": 3343
+ },
+ {
+ "epoch": 46.4471615720524,
+ "grad_norm": 0.13102488219738007,
+ "learning_rate": 0.0006,
+ "loss": 6.8635993003845215,
+ "step": 3344
+ },
+ {
+ "epoch": 46.46113537117904,
+ "grad_norm": 0.23691938817501068,
+ "learning_rate": 0.0006,
+ "loss": 7.313158988952637,
+ "step": 3345
+ },
+ {
+ "epoch": 46.47510917030568,
+ "grad_norm": 0.11670669168233871,
+ "learning_rate": 0.0006,
+ "loss": 7.240363121032715,
+ "step": 3346
+ },
+ {
+ "epoch": 46.48908296943232,
+ "grad_norm": 0.09985820949077606,
+ "learning_rate": 0.0006,
+ "loss": 7.113332271575928,
+ "step": 3347
+ },
+ {
+ "epoch": 46.50305676855895,
+ "grad_norm": 0.11406408250331879,
+ "learning_rate": 0.0006,
+ "loss": 7.051285743713379,
+ "step": 3348
+ },
+ {
+ "epoch": 46.51703056768559,
+ "grad_norm": 0.07900336384773254,
+ "learning_rate": 0.0006,
+ "loss": 7.014365196228027,
+ "step": 3349
+ },
+ {
+ "epoch": 46.531004366812226,
+ "grad_norm": 0.05380333960056305,
+ "learning_rate": 0.0006,
+ "loss": 6.8868818283081055,
+ "step": 3350
+ },
+ {
+ "epoch": 46.544978165938865,
+ "grad_norm": 0.05868449807167053,
+ "learning_rate": 0.0006,
+ "loss": 6.783599853515625,
+ "step": 3351
+ },
+ {
+ "epoch": 46.5589519650655,
+ "grad_norm": 0.04428785294294357,
+ "learning_rate": 0.0006,
+ "loss": 6.798511505126953,
+ "step": 3352
+ },
+ {
+ "epoch": 46.57292576419214,
+ "grad_norm": 0.055527277290821075,
+ "learning_rate": 0.0006,
+ "loss": 6.694507598876953,
+ "step": 3353
+ },
+ {
+ "epoch": 46.58689956331878,
+ "grad_norm": 0.03872432932257652,
+ "learning_rate": 0.0006,
+ "loss": 6.736631393432617,
+ "step": 3354
+ },
+ {
+ "epoch": 46.60087336244541,
+ "grad_norm": 0.04330779239535332,
+ "learning_rate": 0.0006,
+ "loss": 6.576806545257568,
+ "step": 3355
+ },
+ {
+ "epoch": 46.61484716157205,
+ "grad_norm": 0.036767736077308655,
+ "learning_rate": 0.0006,
+ "loss": 6.559871196746826,
+ "step": 3356
+ },
+ {
+ "epoch": 46.62882096069869,
+ "grad_norm": 0.039857108145952225,
+ "learning_rate": 0.0006,
+ "loss": 6.5749831199646,
+ "step": 3357
+ },
+ {
+ "epoch": 46.64279475982533,
+ "grad_norm": 0.048719972372055054,
+ "learning_rate": 0.0006,
+ "loss": 6.4915618896484375,
+ "step": 3358
+ },
+ {
+ "epoch": 46.65676855895197,
+ "grad_norm": 0.042849015444517136,
+ "learning_rate": 0.0006,
+ "loss": 6.413516044616699,
+ "step": 3359
+ },
+ {
+ "epoch": 46.670742358078606,
+ "grad_norm": 0.02871009334921837,
+ "learning_rate": 0.0006,
+ "loss": 6.320952415466309,
+ "step": 3360
+ },
+ {
+ "epoch": 46.68471615720524,
+ "grad_norm": 0.03193620592355728,
+ "learning_rate": 0.0006,
+ "loss": 6.4176435470581055,
+ "step": 3361
+ },
+ {
+ "epoch": 46.698689956331876,
+ "grad_norm": 0.027417024597525597,
+ "learning_rate": 0.0006,
+ "loss": 6.342319488525391,
+ "step": 3362
+ },
+ {
+ "epoch": 46.712663755458514,
+ "grad_norm": 0.03305937349796295,
+ "learning_rate": 0.0006,
+ "loss": 6.246112823486328,
+ "step": 3363
+ },
+ {
+ "epoch": 46.72663755458515,
+ "grad_norm": 0.031403183937072754,
+ "learning_rate": 0.0006,
+ "loss": 6.149503231048584,
+ "step": 3364
+ },
+ {
+ "epoch": 46.74061135371179,
+ "grad_norm": 0.031563788652420044,
+ "learning_rate": 0.0006,
+ "loss": 6.278808116912842,
+ "step": 3365
+ },
+ {
+ "epoch": 46.75458515283843,
+ "grad_norm": 0.028366589918732643,
+ "learning_rate": 0.0006,
+ "loss": 6.057036876678467,
+ "step": 3366
+ },
+ {
+ "epoch": 46.76855895196506,
+ "grad_norm": 0.023715803399682045,
+ "learning_rate": 0.0006,
+ "loss": 6.111324310302734,
+ "step": 3367
+ },
+ {
+ "epoch": 46.7825327510917,
+ "grad_norm": 0.026546811684966087,
+ "learning_rate": 0.0006,
+ "loss": 6.105669975280762,
+ "step": 3368
+ },
+ {
+ "epoch": 46.79650655021834,
+ "grad_norm": 0.023531019687652588,
+ "learning_rate": 0.0006,
+ "loss": 6.068072319030762,
+ "step": 3369
+ },
+ {
+ "epoch": 46.81048034934498,
+ "grad_norm": 0.024480128660798073,
+ "learning_rate": 0.0006,
+ "loss": 5.947896957397461,
+ "step": 3370
+ },
+ {
+ "epoch": 46.82445414847162,
+ "grad_norm": 0.02517310529947281,
+ "learning_rate": 0.0006,
+ "loss": 6.030037879943848,
+ "step": 3371
+ },
+ {
+ "epoch": 46.838427947598255,
+ "grad_norm": 0.026723403483629227,
+ "learning_rate": 0.0006,
+ "loss": 5.963776111602783,
+ "step": 3372
+ },
+ {
+ "epoch": 46.852401746724894,
+ "grad_norm": 0.03163504600524902,
+ "learning_rate": 0.0006,
+ "loss": 5.893466949462891,
+ "step": 3373
+ },
+ {
+ "epoch": 46.866375545851525,
+ "grad_norm": 0.04058683291077614,
+ "learning_rate": 0.0006,
+ "loss": 5.9202799797058105,
+ "step": 3374
+ },
+ {
+ "epoch": 46.880349344978164,
+ "grad_norm": 0.04964117333292961,
+ "learning_rate": 0.0006,
+ "loss": 5.9366559982299805,
+ "step": 3375
+ },
+ {
+ "epoch": 46.8943231441048,
+ "grad_norm": 0.055406033992767334,
+ "learning_rate": 0.0006,
+ "loss": 5.912230968475342,
+ "step": 3376
+ },
+ {
+ "epoch": 46.90829694323144,
+ "grad_norm": 0.04643625393509865,
+ "learning_rate": 0.0006,
+ "loss": 5.87260103225708,
+ "step": 3377
+ },
+ {
+ "epoch": 46.92227074235808,
+ "grad_norm": 0.029325876384973526,
+ "learning_rate": 0.0006,
+ "loss": 5.854763984680176,
+ "step": 3378
+ },
+ {
+ "epoch": 46.93624454148472,
+ "grad_norm": 0.03159382566809654,
+ "learning_rate": 0.0006,
+ "loss": 5.880882263183594,
+ "step": 3379
+ },
+ {
+ "epoch": 46.95021834061135,
+ "grad_norm": 0.030224012210965157,
+ "learning_rate": 0.0006,
+ "loss": 5.787360191345215,
+ "step": 3380
+ },
+ {
+ "epoch": 46.96419213973799,
+ "grad_norm": 0.028368130326271057,
+ "learning_rate": 0.0006,
+ "loss": 5.817195892333984,
+ "step": 3381
+ },
+ {
+ "epoch": 46.97816593886463,
+ "grad_norm": 0.029411399737000465,
+ "learning_rate": 0.0006,
+ "loss": 5.783403396606445,
+ "step": 3382
+ },
+ {
+ "epoch": 46.992139737991266,
+ "grad_norm": 0.029927462339401245,
+ "learning_rate": 0.0006,
+ "loss": 5.662444591522217,
+ "step": 3383
+ },
+ {
+ "epoch": 47.0,
+ "grad_norm": 0.02872961387038231,
+ "learning_rate": 0.0006,
+ "loss": 5.742781639099121,
+ "step": 3384
+ },
+ {
+ "epoch": 47.0,
+ "eval_loss": 5.870260238647461,
+ "eval_runtime": 57.3734,
+ "eval_samples_per_second": 42.563,
+ "eval_steps_per_second": 1.342,
+ "step": 3384
+ },
+ {
+ "epoch": 47.01397379912664,
+ "grad_norm": 0.028542131185531616,
+ "learning_rate": 0.0006,
+ "loss": 5.753256797790527,
+ "step": 3385
+ },
+ {
+ "epoch": 47.02794759825328,
+ "grad_norm": 0.03227119892835617,
+ "learning_rate": 0.0006,
+ "loss": 5.722246170043945,
+ "step": 3386
+ },
+ {
+ "epoch": 47.041921397379916,
+ "grad_norm": 0.031461723148822784,
+ "learning_rate": 0.0006,
+ "loss": 5.740520000457764,
+ "step": 3387
+ },
+ {
+ "epoch": 47.05589519650655,
+ "grad_norm": 0.03184220567345619,
+ "learning_rate": 0.0006,
+ "loss": 5.720890045166016,
+ "step": 3388
+ },
+ {
+ "epoch": 47.069868995633186,
+ "grad_norm": 0.030015580356121063,
+ "learning_rate": 0.0006,
+ "loss": 5.597814559936523,
+ "step": 3389
+ },
+ {
+ "epoch": 47.083842794759825,
+ "grad_norm": 0.029931779950857162,
+ "learning_rate": 0.0006,
+ "loss": 5.5412468910217285,
+ "step": 3390
+ },
+ {
+ "epoch": 47.09781659388646,
+ "grad_norm": 0.03206505626440048,
+ "learning_rate": 0.0006,
+ "loss": 5.6967597007751465,
+ "step": 3391
+ },
+ {
+ "epoch": 47.1117903930131,
+ "grad_norm": 0.03978161886334419,
+ "learning_rate": 0.0006,
+ "loss": 5.59661340713501,
+ "step": 3392
+ },
+ {
+ "epoch": 47.12576419213974,
+ "grad_norm": 0.04619080200791359,
+ "learning_rate": 0.0006,
+ "loss": 5.795788764953613,
+ "step": 3393
+ },
+ {
+ "epoch": 47.13973799126637,
+ "grad_norm": 0.049085382372140884,
+ "learning_rate": 0.0006,
+ "loss": 5.667335510253906,
+ "step": 3394
+ },
+ {
+ "epoch": 47.15371179039301,
+ "grad_norm": 0.04830858111381531,
+ "learning_rate": 0.0006,
+ "loss": 5.668806552886963,
+ "step": 3395
+ },
+ {
+ "epoch": 47.16768558951965,
+ "grad_norm": 0.0583178736269474,
+ "learning_rate": 0.0006,
+ "loss": 5.692285537719727,
+ "step": 3396
+ },
+ {
+ "epoch": 47.18165938864629,
+ "grad_norm": 0.05182253196835518,
+ "learning_rate": 0.0006,
+ "loss": 5.678645133972168,
+ "step": 3397
+ },
+ {
+ "epoch": 47.19563318777293,
+ "grad_norm": 0.04011905938386917,
+ "learning_rate": 0.0006,
+ "loss": 5.661550998687744,
+ "step": 3398
+ },
+ {
+ "epoch": 47.209606986899566,
+ "grad_norm": 0.03888486698269844,
+ "learning_rate": 0.0006,
+ "loss": 5.643001556396484,
+ "step": 3399
+ },
+ {
+ "epoch": 47.223580786026204,
+ "grad_norm": 0.030658265575766563,
+ "learning_rate": 0.0006,
+ "loss": 5.659090995788574,
+ "step": 3400
+ },
+ {
+ "epoch": 47.237554585152836,
+ "grad_norm": 0.03609256446361542,
+ "learning_rate": 0.0006,
+ "loss": 5.544501781463623,
+ "step": 3401
+ },
+ {
+ "epoch": 47.251528384279474,
+ "grad_norm": 0.033589769154787064,
+ "learning_rate": 0.0006,
+ "loss": 5.463878631591797,
+ "step": 3402
+ },
+ {
+ "epoch": 47.26550218340611,
+ "grad_norm": 0.035128287971019745,
+ "learning_rate": 0.0006,
+ "loss": 5.5680060386657715,
+ "step": 3403
+ },
+ {
+ "epoch": 47.27947598253275,
+ "grad_norm": 0.038684550672769547,
+ "learning_rate": 0.0006,
+ "loss": 5.587898254394531,
+ "step": 3404
+ },
+ {
+ "epoch": 47.29344978165939,
+ "grad_norm": 0.043566226959228516,
+ "learning_rate": 0.0006,
+ "loss": 5.580896854400635,
+ "step": 3405
+ },
+ {
+ "epoch": 47.30742358078603,
+ "grad_norm": 0.04428374394774437,
+ "learning_rate": 0.0006,
+ "loss": 5.603460311889648,
+ "step": 3406
+ },
+ {
+ "epoch": 47.32139737991266,
+ "grad_norm": 0.037368934601545334,
+ "learning_rate": 0.0006,
+ "loss": 5.564740180969238,
+ "step": 3407
+ },
+ {
+ "epoch": 47.3353711790393,
+ "grad_norm": 0.04092350974678993,
+ "learning_rate": 0.0006,
+ "loss": 5.584153652191162,
+ "step": 3408
+ },
+ {
+ "epoch": 47.34934497816594,
+ "grad_norm": 0.045031506568193436,
+ "learning_rate": 0.0006,
+ "loss": 5.601186275482178,
+ "step": 3409
+ },
+ {
+ "epoch": 47.36331877729258,
+ "grad_norm": 0.041830409318208694,
+ "learning_rate": 0.0006,
+ "loss": 5.512788772583008,
+ "step": 3410
+ },
+ {
+ "epoch": 47.377292576419215,
+ "grad_norm": 0.037633974105119705,
+ "learning_rate": 0.0006,
+ "loss": 5.503828525543213,
+ "step": 3411
+ },
+ {
+ "epoch": 47.391266375545854,
+ "grad_norm": 0.03473588079214096,
+ "learning_rate": 0.0006,
+ "loss": 5.570557594299316,
+ "step": 3412
+ },
+ {
+ "epoch": 47.40524017467249,
+ "grad_norm": 0.039325181394815445,
+ "learning_rate": 0.0006,
+ "loss": 5.427783966064453,
+ "step": 3413
+ },
+ {
+ "epoch": 47.419213973799124,
+ "grad_norm": 0.04034624621272087,
+ "learning_rate": 0.0006,
+ "loss": 5.572591304779053,
+ "step": 3414
+ },
+ {
+ "epoch": 47.43318777292576,
+ "grad_norm": 0.05245184153318405,
+ "learning_rate": 0.0006,
+ "loss": 5.505208969116211,
+ "step": 3415
+ },
+ {
+ "epoch": 47.4471615720524,
+ "grad_norm": 0.0568675734102726,
+ "learning_rate": 0.0006,
+ "loss": 5.589373588562012,
+ "step": 3416
+ },
+ {
+ "epoch": 47.46113537117904,
+ "grad_norm": 0.048336587846279144,
+ "learning_rate": 0.0006,
+ "loss": 5.522005081176758,
+ "step": 3417
+ },
+ {
+ "epoch": 47.47510917030568,
+ "grad_norm": 0.04194975271821022,
+ "learning_rate": 0.0006,
+ "loss": 5.387576103210449,
+ "step": 3418
+ },
+ {
+ "epoch": 47.48908296943232,
+ "grad_norm": 0.04019433259963989,
+ "learning_rate": 0.0006,
+ "loss": 5.522615432739258,
+ "step": 3419
+ },
+ {
+ "epoch": 47.50305676855895,
+ "grad_norm": 0.05709933117032051,
+ "learning_rate": 0.0006,
+ "loss": 5.420286178588867,
+ "step": 3420
+ },
+ {
+ "epoch": 47.51703056768559,
+ "grad_norm": 0.09130338579416275,
+ "learning_rate": 0.0006,
+ "loss": 5.484622478485107,
+ "step": 3421
+ },
+ {
+ "epoch": 47.531004366812226,
+ "grad_norm": 0.2757541835308075,
+ "learning_rate": 0.0006,
+ "loss": 5.821832656860352,
+ "step": 3422
+ },
+ {
+ "epoch": 47.544978165938865,
+ "grad_norm": 0.28292080760002136,
+ "learning_rate": 0.0006,
+ "loss": 6.579934120178223,
+ "step": 3423
+ },
+ {
+ "epoch": 47.5589519650655,
+ "grad_norm": 0.15293604135513306,
+ "learning_rate": 0.0006,
+ "loss": 6.478213310241699,
+ "step": 3424
+ },
+ {
+ "epoch": 47.57292576419214,
+ "grad_norm": 0.11284393817186356,
+ "learning_rate": 0.0006,
+ "loss": 6.4066386222839355,
+ "step": 3425
+ },
+ {
+ "epoch": 47.58689956331878,
+ "grad_norm": 0.10933073610067368,
+ "learning_rate": 0.0006,
+ "loss": 6.3306169509887695,
+ "step": 3426
+ },
+ {
+ "epoch": 47.60087336244541,
+ "grad_norm": 0.05837641656398773,
+ "learning_rate": 0.0006,
+ "loss": 6.250000953674316,
+ "step": 3427
+ },
+ {
+ "epoch": 47.61484716157205,
+ "grad_norm": 0.05987229570746422,
+ "learning_rate": 0.0006,
+ "loss": 6.152463436126709,
+ "step": 3428
+ },
+ {
+ "epoch": 47.62882096069869,
+ "grad_norm": 0.04064938426017761,
+ "learning_rate": 0.0006,
+ "loss": 6.117897033691406,
+ "step": 3429
+ },
+ {
+ "epoch": 47.64279475982533,
+ "grad_norm": 0.04565057530999184,
+ "learning_rate": 0.0006,
+ "loss": 6.042291641235352,
+ "step": 3430
+ },
+ {
+ "epoch": 47.65676855895197,
+ "grad_norm": 0.03475063294172287,
+ "learning_rate": 0.0006,
+ "loss": 6.0071868896484375,
+ "step": 3431
+ },
+ {
+ "epoch": 47.670742358078606,
+ "grad_norm": 0.03809220343828201,
+ "learning_rate": 0.0006,
+ "loss": 6.0128021240234375,
+ "step": 3432
+ },
+ {
+ "epoch": 47.68471615720524,
+ "grad_norm": 0.03307219594717026,
+ "learning_rate": 0.0006,
+ "loss": 5.894664287567139,
+ "step": 3433
+ },
+ {
+ "epoch": 47.698689956331876,
+ "grad_norm": 0.028824346140027046,
+ "learning_rate": 0.0006,
+ "loss": 5.914865493774414,
+ "step": 3434
+ },
+ {
+ "epoch": 47.712663755458514,
+ "grad_norm": 0.027923690155148506,
+ "learning_rate": 0.0006,
+ "loss": 5.851759910583496,
+ "step": 3435
+ },
+ {
+ "epoch": 47.72663755458515,
+ "grad_norm": 0.025492684915661812,
+ "learning_rate": 0.0006,
+ "loss": 5.749186038970947,
+ "step": 3436
+ },
+ {
+ "epoch": 47.74061135371179,
+ "grad_norm": 0.025105174630880356,
+ "learning_rate": 0.0006,
+ "loss": 5.803566932678223,
+ "step": 3437
+ },
+ {
+ "epoch": 47.75458515283843,
+ "grad_norm": 0.023298880085349083,
+ "learning_rate": 0.0006,
+ "loss": 5.76276159286499,
+ "step": 3438
+ },
+ {
+ "epoch": 47.76855895196506,
+ "grad_norm": 0.02468576282262802,
+ "learning_rate": 0.0006,
+ "loss": 5.833852767944336,
+ "step": 3439
+ },
+ {
+ "epoch": 47.7825327510917,
+ "grad_norm": 0.022730689495801926,
+ "learning_rate": 0.0006,
+ "loss": 5.736724376678467,
+ "step": 3440
+ },
+ {
+ "epoch": 47.79650655021834,
+ "grad_norm": 0.021335354074835777,
+ "learning_rate": 0.0006,
+ "loss": 5.644248008728027,
+ "step": 3441
+ },
+ {
+ "epoch": 47.81048034934498,
+ "grad_norm": 0.02288101427257061,
+ "learning_rate": 0.0006,
+ "loss": 5.7011942863464355,
+ "step": 3442
+ },
+ {
+ "epoch": 47.82445414847162,
+ "grad_norm": 0.01961585320532322,
+ "learning_rate": 0.0006,
+ "loss": 5.63215446472168,
+ "step": 3443
+ },
+ {
+ "epoch": 47.838427947598255,
+ "grad_norm": 0.017659205943346024,
+ "learning_rate": 0.0006,
+ "loss": 5.744845390319824,
+ "step": 3444
+ },
+ {
+ "epoch": 47.852401746724894,
+ "grad_norm": 0.02281784638762474,
+ "learning_rate": 0.0006,
+ "loss": 5.696993350982666,
+ "step": 3445
+ },
+ {
+ "epoch": 47.866375545851525,
+ "grad_norm": 0.016394242644309998,
+ "learning_rate": 0.0006,
+ "loss": 5.659870147705078,
+ "step": 3446
+ },
+ {
+ "epoch": 47.880349344978164,
+ "grad_norm": 0.01967368647456169,
+ "learning_rate": 0.0006,
+ "loss": 5.638072490692139,
+ "step": 3447
+ },
+ {
+ "epoch": 47.8943231441048,
+ "grad_norm": 0.017326267436146736,
+ "learning_rate": 0.0006,
+ "loss": 5.638836860656738,
+ "step": 3448
+ },
+ {
+ "epoch": 47.90829694323144,
+ "grad_norm": 0.017778463661670685,
+ "learning_rate": 0.0006,
+ "loss": 5.634466648101807,
+ "step": 3449
+ },
+ {
+ "epoch": 47.92227074235808,
+ "grad_norm": 0.017269758507609367,
+ "learning_rate": 0.0006,
+ "loss": 5.506277561187744,
+ "step": 3450
+ },
+ {
+ "epoch": 47.93624454148472,
+ "grad_norm": 0.01733289286494255,
+ "learning_rate": 0.0006,
+ "loss": 5.544455528259277,
+ "step": 3451
+ },
+ {
+ "epoch": 47.95021834061135,
+ "grad_norm": 0.017787795513868332,
+ "learning_rate": 0.0006,
+ "loss": 5.498943328857422,
+ "step": 3452
+ },
+ {
+ "epoch": 47.96419213973799,
+ "grad_norm": 0.015770932659506798,
+ "learning_rate": 0.0006,
+ "loss": 5.537227630615234,
+ "step": 3453
+ },
+ {
+ "epoch": 47.97816593886463,
+ "grad_norm": 0.016132809221744537,
+ "learning_rate": 0.0006,
+ "loss": 5.489081382751465,
+ "step": 3454
+ },
+ {
+ "epoch": 47.992139737991266,
+ "grad_norm": 0.01749664731323719,
+ "learning_rate": 0.0006,
+ "loss": 5.454263687133789,
+ "step": 3455
+ },
+ {
+ "epoch": 48.0,
+ "grad_norm": 0.016711296513676643,
+ "learning_rate": 0.0006,
+ "loss": 5.530783653259277,
+ "step": 3456
+ },
+ {
+ "epoch": 48.0,
+ "eval_loss": 5.596925258636475,
+ "eval_runtime": 56.9612,
+ "eval_samples_per_second": 42.871,
+ "eval_steps_per_second": 1.352,
+ "step": 3456
+ },
+ {
+ "epoch": 48.01397379912664,
+ "grad_norm": 0.01867641694843769,
+ "learning_rate": 0.0006,
+ "loss": 5.506853103637695,
+ "step": 3457
+ },
+ {
+ "epoch": 48.02794759825328,
+ "grad_norm": 0.023250792175531387,
+ "learning_rate": 0.0006,
+ "loss": 5.4074177742004395,
+ "step": 3458
+ },
+ {
+ "epoch": 48.041921397379916,
+ "grad_norm": 0.01970936730504036,
+ "learning_rate": 0.0006,
+ "loss": 5.508777618408203,
+ "step": 3459
+ },
+ {
+ "epoch": 48.05589519650655,
+ "grad_norm": 0.01879260502755642,
+ "learning_rate": 0.0006,
+ "loss": 5.412424087524414,
+ "step": 3460
+ },
+ {
+ "epoch": 48.069868995633186,
+ "grad_norm": 0.017420092597603798,
+ "learning_rate": 0.0006,
+ "loss": 5.43361759185791,
+ "step": 3461
+ },
+ {
+ "epoch": 48.083842794759825,
+ "grad_norm": 0.018707862123847008,
+ "learning_rate": 0.0006,
+ "loss": 5.476800918579102,
+ "step": 3462
+ },
+ {
+ "epoch": 48.09781659388646,
+ "grad_norm": 0.023185882717370987,
+ "learning_rate": 0.0006,
+ "loss": 5.465272426605225,
+ "step": 3463
+ },
+ {
+ "epoch": 48.1117903930131,
+ "grad_norm": 0.023570368066430092,
+ "learning_rate": 0.0006,
+ "loss": 5.391471862792969,
+ "step": 3464
+ },
+ {
+ "epoch": 48.12576419213974,
+ "grad_norm": 0.02651144750416279,
+ "learning_rate": 0.0006,
+ "loss": 5.422799110412598,
+ "step": 3465
+ },
+ {
+ "epoch": 48.13973799126637,
+ "grad_norm": 0.03316329047083855,
+ "learning_rate": 0.0006,
+ "loss": 5.3952202796936035,
+ "step": 3466
+ },
+ {
+ "epoch": 48.15371179039301,
+ "grad_norm": 0.04012278467416763,
+ "learning_rate": 0.0006,
+ "loss": 5.540853023529053,
+ "step": 3467
+ },
+ {
+ "epoch": 48.16768558951965,
+ "grad_norm": 0.04763203114271164,
+ "learning_rate": 0.0006,
+ "loss": 5.387890338897705,
+ "step": 3468
+ },
+ {
+ "epoch": 48.18165938864629,
+ "grad_norm": 0.05129165202379227,
+ "learning_rate": 0.0006,
+ "loss": 5.414722442626953,
+ "step": 3469
+ },
+ {
+ "epoch": 48.19563318777293,
+ "grad_norm": 0.04438071325421333,
+ "learning_rate": 0.0006,
+ "loss": 5.386244773864746,
+ "step": 3470
+ },
+ {
+ "epoch": 48.209606986899566,
+ "grad_norm": 0.040234919637441635,
+ "learning_rate": 0.0006,
+ "loss": 5.431617736816406,
+ "step": 3471
+ },
+ {
+ "epoch": 48.223580786026204,
+ "grad_norm": 0.03217008337378502,
+ "learning_rate": 0.0006,
+ "loss": 5.30168342590332,
+ "step": 3472
+ },
+ {
+ "epoch": 48.237554585152836,
+ "grad_norm": 0.03087974339723587,
+ "learning_rate": 0.0006,
+ "loss": 5.316389083862305,
+ "step": 3473
+ },
+ {
+ "epoch": 48.251528384279474,
+ "grad_norm": 0.02789347991347313,
+ "learning_rate": 0.0006,
+ "loss": 5.385970592498779,
+ "step": 3474
+ },
+ {
+ "epoch": 48.26550218340611,
+ "grad_norm": 0.024792097508907318,
+ "learning_rate": 0.0006,
+ "loss": 5.362329959869385,
+ "step": 3475
+ },
+ {
+ "epoch": 48.27947598253275,
+ "grad_norm": 0.025869522243738174,
+ "learning_rate": 0.0006,
+ "loss": 5.28400182723999,
+ "step": 3476
+ },
+ {
+ "epoch": 48.29344978165939,
+ "grad_norm": 0.021339882165193558,
+ "learning_rate": 0.0006,
+ "loss": 5.36669921875,
+ "step": 3477
+ },
+ {
+ "epoch": 48.30742358078603,
+ "grad_norm": 0.021462520584464073,
+ "learning_rate": 0.0006,
+ "loss": 5.311222076416016,
+ "step": 3478
+ },
+ {
+ "epoch": 48.32139737991266,
+ "grad_norm": 0.02193879708647728,
+ "learning_rate": 0.0006,
+ "loss": 5.291441917419434,
+ "step": 3479
+ },
+ {
+ "epoch": 48.3353711790393,
+ "grad_norm": 0.01963093690574169,
+ "learning_rate": 0.0006,
+ "loss": 5.342780113220215,
+ "step": 3480
+ },
+ {
+ "epoch": 48.34934497816594,
+ "grad_norm": 0.020634736865758896,
+ "learning_rate": 0.0006,
+ "loss": 5.144141674041748,
+ "step": 3481
+ },
+ {
+ "epoch": 48.36331877729258,
+ "grad_norm": 0.02126968279480934,
+ "learning_rate": 0.0006,
+ "loss": 5.232624530792236,
+ "step": 3482
+ },
+ {
+ "epoch": 48.377292576419215,
+ "grad_norm": 0.021203994750976562,
+ "learning_rate": 0.0006,
+ "loss": 5.290219783782959,
+ "step": 3483
+ },
+ {
+ "epoch": 48.391266375545854,
+ "grad_norm": 0.02193225920200348,
+ "learning_rate": 0.0006,
+ "loss": 5.29197359085083,
+ "step": 3484
+ },
+ {
+ "epoch": 48.40524017467249,
+ "grad_norm": 0.023174211382865906,
+ "learning_rate": 0.0006,
+ "loss": 5.204349517822266,
+ "step": 3485
+ },
+ {
+ "epoch": 48.419213973799124,
+ "grad_norm": 0.024161914363503456,
+ "learning_rate": 0.0006,
+ "loss": 5.339351654052734,
+ "step": 3486
+ },
+ {
+ "epoch": 48.43318777292576,
+ "grad_norm": 0.02221524342894554,
+ "learning_rate": 0.0006,
+ "loss": 5.263001441955566,
+ "step": 3487
+ },
+ {
+ "epoch": 48.4471615720524,
+ "grad_norm": 0.026632830500602722,
+ "learning_rate": 0.0006,
+ "loss": 5.221281051635742,
+ "step": 3488
+ },
+ {
+ "epoch": 48.46113537117904,
+ "grad_norm": 0.034923870116472244,
+ "learning_rate": 0.0006,
+ "loss": 5.296039581298828,
+ "step": 3489
+ },
+ {
+ "epoch": 48.47510917030568,
+ "grad_norm": 0.046184979379177094,
+ "learning_rate": 0.0006,
+ "loss": 5.212769985198975,
+ "step": 3490
+ },
+ {
+ "epoch": 48.48908296943232,
+ "grad_norm": 0.05154098570346832,
+ "learning_rate": 0.0006,
+ "loss": 5.337864875793457,
+ "step": 3491
+ },
+ {
+ "epoch": 48.50305676855895,
+ "grad_norm": 0.042967259883880615,
+ "learning_rate": 0.0006,
+ "loss": 5.267958641052246,
+ "step": 3492
+ },
+ {
+ "epoch": 48.51703056768559,
+ "grad_norm": 0.05356583371758461,
+ "learning_rate": 0.0006,
+ "loss": 5.1404218673706055,
+ "step": 3493
+ },
+ {
+ "epoch": 48.531004366812226,
+ "grad_norm": 0.08219032734632492,
+ "learning_rate": 0.0006,
+ "loss": 5.300687789916992,
+ "step": 3494
+ },
+ {
+ "epoch": 48.544978165938865,
+ "grad_norm": 0.19204089045524597,
+ "learning_rate": 0.0006,
+ "loss": 5.456025123596191,
+ "step": 3495
+ },
+ {
+ "epoch": 48.5589519650655,
+ "grad_norm": 0.25917696952819824,
+ "learning_rate": 0.0006,
+ "loss": 6.156761169433594,
+ "step": 3496
+ },
+ {
+ "epoch": 48.57292576419214,
+ "grad_norm": 0.1566283404827118,
+ "learning_rate": 0.0006,
+ "loss": 5.862346649169922,
+ "step": 3497
+ },
+ {
+ "epoch": 48.58689956331878,
+ "grad_norm": 0.1761486828327179,
+ "learning_rate": 0.0006,
+ "loss": 6.012795448303223,
+ "step": 3498
+ },
+ {
+ "epoch": 48.60087336244541,
+ "grad_norm": 0.09857621043920517,
+ "learning_rate": 0.0006,
+ "loss": 5.9248151779174805,
+ "step": 3499
+ },
+ {
+ "epoch": 48.61484716157205,
+ "grad_norm": 0.05968055501580238,
+ "learning_rate": 0.0006,
+ "loss": 5.780706882476807,
+ "step": 3500
+ },
+ {
+ "epoch": 48.62882096069869,
+ "grad_norm": 0.057433564215898514,
+ "learning_rate": 0.0006,
+ "loss": 5.645388603210449,
+ "step": 3501
+ },
+ {
+ "epoch": 48.64279475982533,
+ "grad_norm": 0.039659615606069565,
+ "learning_rate": 0.0006,
+ "loss": 5.695138931274414,
+ "step": 3502
+ },
+ {
+ "epoch": 48.65676855895197,
+ "grad_norm": 0.04019695147871971,
+ "learning_rate": 0.0006,
+ "loss": 5.672435283660889,
+ "step": 3503
+ },
+ {
+ "epoch": 48.670742358078606,
+ "grad_norm": 0.03033018670976162,
+ "learning_rate": 0.0006,
+ "loss": 5.682260513305664,
+ "step": 3504
+ },
+ {
+ "epoch": 48.68471615720524,
+ "grad_norm": 0.03215787187218666,
+ "learning_rate": 0.0006,
+ "loss": 5.604182243347168,
+ "step": 3505
+ },
+ {
+ "epoch": 48.698689956331876,
+ "grad_norm": 0.026865126565098763,
+ "learning_rate": 0.0006,
+ "loss": 5.64040470123291,
+ "step": 3506
+ },
+ {
+ "epoch": 48.712663755458514,
+ "grad_norm": 0.02831609919667244,
+ "learning_rate": 0.0006,
+ "loss": 5.552720069885254,
+ "step": 3507
+ },
+ {
+ "epoch": 48.72663755458515,
+ "grad_norm": 0.024226613342761993,
+ "learning_rate": 0.0006,
+ "loss": 5.556946277618408,
+ "step": 3508
+ },
+ {
+ "epoch": 48.74061135371179,
+ "grad_norm": 0.025393934920430183,
+ "learning_rate": 0.0006,
+ "loss": 5.476178169250488,
+ "step": 3509
+ },
+ {
+ "epoch": 48.75458515283843,
+ "grad_norm": 0.01984412968158722,
+ "learning_rate": 0.0006,
+ "loss": 5.471088409423828,
+ "step": 3510
+ },
+ {
+ "epoch": 48.76855895196506,
+ "grad_norm": 0.02035129815340042,
+ "learning_rate": 0.0006,
+ "loss": 5.438812732696533,
+ "step": 3511
+ },
+ {
+ "epoch": 48.7825327510917,
+ "grad_norm": 0.0208350270986557,
+ "learning_rate": 0.0006,
+ "loss": 5.381746292114258,
+ "step": 3512
+ },
+ {
+ "epoch": 48.79650655021834,
+ "grad_norm": 0.017783882096409798,
+ "learning_rate": 0.0006,
+ "loss": 5.351675510406494,
+ "step": 3513
+ },
+ {
+ "epoch": 48.81048034934498,
+ "grad_norm": 0.01671770215034485,
+ "learning_rate": 0.0006,
+ "loss": 5.329428672790527,
+ "step": 3514
+ },
+ {
+ "epoch": 48.82445414847162,
+ "grad_norm": 0.017891215160489082,
+ "learning_rate": 0.0006,
+ "loss": 5.212074279785156,
+ "step": 3515
+ },
+ {
+ "epoch": 48.838427947598255,
+ "grad_norm": 0.016847791150212288,
+ "learning_rate": 0.0006,
+ "loss": 5.299534797668457,
+ "step": 3516
+ },
+ {
+ "epoch": 48.852401746724894,
+ "grad_norm": 0.01593637838959694,
+ "learning_rate": 0.0006,
+ "loss": 5.30466890335083,
+ "step": 3517
+ },
+ {
+ "epoch": 48.866375545851525,
+ "grad_norm": 0.014713102020323277,
+ "learning_rate": 0.0006,
+ "loss": 5.294345378875732,
+ "step": 3518
+ },
+ {
+ "epoch": 48.880349344978164,
+ "grad_norm": 0.01433857623487711,
+ "learning_rate": 0.0006,
+ "loss": 5.2148542404174805,
+ "step": 3519
+ },
+ {
+ "epoch": 48.8943231441048,
+ "grad_norm": 0.014979379251599312,
+ "learning_rate": 0.0006,
+ "loss": 5.298582077026367,
+ "step": 3520
+ },
+ {
+ "epoch": 48.90829694323144,
+ "grad_norm": 0.013494192622601986,
+ "learning_rate": 0.0006,
+ "loss": 5.309783935546875,
+ "step": 3521
+ },
+ {
+ "epoch": 48.92227074235808,
+ "grad_norm": 0.014020012691617012,
+ "learning_rate": 0.0006,
+ "loss": 5.284879684448242,
+ "step": 3522
+ },
+ {
+ "epoch": 48.93624454148472,
+ "grad_norm": 0.01348058134317398,
+ "learning_rate": 0.0006,
+ "loss": 5.249116897583008,
+ "step": 3523
+ },
+ {
+ "epoch": 48.95021834061135,
+ "grad_norm": 0.013280218467116356,
+ "learning_rate": 0.0006,
+ "loss": 5.273489952087402,
+ "step": 3524
+ },
+ {
+ "epoch": 48.96419213973799,
+ "grad_norm": 0.012647110037505627,
+ "learning_rate": 0.0006,
+ "loss": 5.306652069091797,
+ "step": 3525
+ },
+ {
+ "epoch": 48.97816593886463,
+ "grad_norm": 0.012847724370658398,
+ "learning_rate": 0.0006,
+ "loss": 5.235745429992676,
+ "step": 3526
+ },
+ {
+ "epoch": 48.992139737991266,
+ "grad_norm": 0.011820166371762753,
+ "learning_rate": 0.0006,
+ "loss": 5.238091468811035,
+ "step": 3527
+ },
+ {
+ "epoch": 49.0,
+ "grad_norm": 0.01255585066974163,
+ "learning_rate": 0.0006,
+ "loss": 5.231295585632324,
+ "step": 3528
+ },
+ {
+ "epoch": 49.0,
+ "eval_loss": 5.394587993621826,
+ "eval_runtime": 56.3734,
+ "eval_samples_per_second": 43.318,
+ "eval_steps_per_second": 1.366,
+ "step": 3528
+ },
+ {
+ "epoch": 49.01397379912664,
+ "grad_norm": 0.011031972244381905,
+ "learning_rate": 0.0006,
+ "loss": 5.1354570388793945,
+ "step": 3529
+ },
+ {
+ "epoch": 49.02794759825328,
+ "grad_norm": 0.010631263256072998,
+ "learning_rate": 0.0006,
+ "loss": 5.177639961242676,
+ "step": 3530
+ },
+ {
+ "epoch": 49.041921397379916,
+ "grad_norm": 0.01133775431662798,
+ "learning_rate": 0.0006,
+ "loss": 5.283428192138672,
+ "step": 3531
+ },
+ {
+ "epoch": 49.05589519650655,
+ "grad_norm": 0.01154172234237194,
+ "learning_rate": 0.0006,
+ "loss": 5.206364154815674,
+ "step": 3532
+ },
+ {
+ "epoch": 49.069868995633186,
+ "grad_norm": 0.010192356072366238,
+ "learning_rate": 0.0006,
+ "loss": 5.2385735511779785,
+ "step": 3533
+ },
+ {
+ "epoch": 49.083842794759825,
+ "grad_norm": 0.012789200060069561,
+ "learning_rate": 0.0006,
+ "loss": 5.071800231933594,
+ "step": 3534
+ },
+ {
+ "epoch": 49.09781659388646,
+ "grad_norm": 0.011282820254564285,
+ "learning_rate": 0.0006,
+ "loss": 5.218020439147949,
+ "step": 3535
+ },
+ {
+ "epoch": 49.1117903930131,
+ "grad_norm": 0.012153448536992073,
+ "learning_rate": 0.0006,
+ "loss": 5.134698867797852,
+ "step": 3536
+ },
+ {
+ "epoch": 49.12576419213974,
+ "grad_norm": 0.011920543387532234,
+ "learning_rate": 0.0006,
+ "loss": 5.141620635986328,
+ "step": 3537
+ },
+ {
+ "epoch": 49.13973799126637,
+ "grad_norm": 0.010949329473078251,
+ "learning_rate": 0.0006,
+ "loss": 5.116259574890137,
+ "step": 3538
+ },
+ {
+ "epoch": 49.15371179039301,
+ "grad_norm": 0.012362448498606682,
+ "learning_rate": 0.0006,
+ "loss": 5.090177059173584,
+ "step": 3539
+ },
+ {
+ "epoch": 49.16768558951965,
+ "grad_norm": 0.011575071141123772,
+ "learning_rate": 0.0006,
+ "loss": 5.093441009521484,
+ "step": 3540
+ },
+ {
+ "epoch": 49.18165938864629,
+ "grad_norm": 0.011126236990094185,
+ "learning_rate": 0.0006,
+ "loss": 5.060235023498535,
+ "step": 3541
+ },
+ {
+ "epoch": 49.19563318777293,
+ "grad_norm": 0.0107401292771101,
+ "learning_rate": 0.0006,
+ "loss": 5.140704154968262,
+ "step": 3542
+ },
+ {
+ "epoch": 49.209606986899566,
+ "grad_norm": 0.011202906258404255,
+ "learning_rate": 0.0006,
+ "loss": 5.10855770111084,
+ "step": 3543
+ },
+ {
+ "epoch": 49.223580786026204,
+ "grad_norm": 0.011083454824984074,
+ "learning_rate": 0.0006,
+ "loss": 5.092751502990723,
+ "step": 3544
+ },
+ {
+ "epoch": 49.237554585152836,
+ "grad_norm": 0.009523800574243069,
+ "learning_rate": 0.0006,
+ "loss": 5.13733434677124,
+ "step": 3545
+ },
+ {
+ "epoch": 49.251528384279474,
+ "grad_norm": 0.011298353783786297,
+ "learning_rate": 0.0006,
+ "loss": 5.036048889160156,
+ "step": 3546
+ },
+ {
+ "epoch": 49.26550218340611,
+ "grad_norm": 0.010676483623683453,
+ "learning_rate": 0.0006,
+ "loss": 5.051901340484619,
+ "step": 3547
+ },
+ {
+ "epoch": 49.27947598253275,
+ "grad_norm": 0.011463399045169353,
+ "learning_rate": 0.0006,
+ "loss": 4.983170986175537,
+ "step": 3548
+ },
+ {
+ "epoch": 49.29344978165939,
+ "grad_norm": 0.010603402741253376,
+ "learning_rate": 0.0006,
+ "loss": 5.112311363220215,
+ "step": 3549
+ },
+ {
+ "epoch": 49.30742358078603,
+ "grad_norm": 0.010535670444369316,
+ "learning_rate": 0.0006,
+ "loss": 5.085227966308594,
+ "step": 3550
+ },
+ {
+ "epoch": 49.32139737991266,
+ "grad_norm": 0.011843928135931492,
+ "learning_rate": 0.0006,
+ "loss": 5.07575798034668,
+ "step": 3551
+ },
+ {
+ "epoch": 49.3353711790393,
+ "grad_norm": 0.010587952099740505,
+ "learning_rate": 0.0006,
+ "loss": 5.019237041473389,
+ "step": 3552
+ },
+ {
+ "epoch": 49.34934497816594,
+ "grad_norm": 0.012285659089684486,
+ "learning_rate": 0.0006,
+ "loss": 5.144603729248047,
+ "step": 3553
+ },
+ {
+ "epoch": 49.36331877729258,
+ "grad_norm": 0.013119431212544441,
+ "learning_rate": 0.0006,
+ "loss": 5.035256862640381,
+ "step": 3554
+ },
+ {
+ "epoch": 49.377292576419215,
+ "grad_norm": 0.012138537131249905,
+ "learning_rate": 0.0006,
+ "loss": 4.996667861938477,
+ "step": 3555
+ },
+ {
+ "epoch": 49.391266375545854,
+ "grad_norm": 0.012142804451286793,
+ "learning_rate": 0.0006,
+ "loss": 4.95058536529541,
+ "step": 3556
+ },
+ {
+ "epoch": 49.40524017467249,
+ "grad_norm": 0.012205676175653934,
+ "learning_rate": 0.0006,
+ "loss": 5.096931457519531,
+ "step": 3557
+ },
+ {
+ "epoch": 49.419213973799124,
+ "grad_norm": 0.012086744420230389,
+ "learning_rate": 0.0006,
+ "loss": 4.99384880065918,
+ "step": 3558
+ },
+ {
+ "epoch": 49.43318777292576,
+ "grad_norm": 0.012918103486299515,
+ "learning_rate": 0.0006,
+ "loss": 5.031294822692871,
+ "step": 3559
+ },
+ {
+ "epoch": 49.4471615720524,
+ "grad_norm": 0.012006713077425957,
+ "learning_rate": 0.0006,
+ "loss": 5.0547943115234375,
+ "step": 3560
+ },
+ {
+ "epoch": 49.46113537117904,
+ "grad_norm": 0.0120399734005332,
+ "learning_rate": 0.0006,
+ "loss": 5.014238357543945,
+ "step": 3561
+ },
+ {
+ "epoch": 49.47510917030568,
+ "grad_norm": 0.013306990265846252,
+ "learning_rate": 0.0006,
+ "loss": 4.972925186157227,
+ "step": 3562
+ },
+ {
+ "epoch": 49.48908296943232,
+ "grad_norm": 0.013878636993467808,
+ "learning_rate": 0.0006,
+ "loss": 5.079033851623535,
+ "step": 3563
+ },
+ {
+ "epoch": 49.50305676855895,
+ "grad_norm": 0.012454566545784473,
+ "learning_rate": 0.0006,
+ "loss": 4.93708610534668,
+ "step": 3564
+ },
+ {
+ "epoch": 49.51703056768559,
+ "grad_norm": 0.014075061306357384,
+ "learning_rate": 0.0006,
+ "loss": 4.965582370758057,
+ "step": 3565
+ },
+ {
+ "epoch": 49.531004366812226,
+ "grad_norm": 0.012525617145001888,
+ "learning_rate": 0.0006,
+ "loss": 4.968148708343506,
+ "step": 3566
+ },
+ {
+ "epoch": 49.544978165938865,
+ "grad_norm": 0.012418225407600403,
+ "learning_rate": 0.0006,
+ "loss": 5.06901741027832,
+ "step": 3567
+ },
+ {
+ "epoch": 49.5589519650655,
+ "grad_norm": 0.015034239739179611,
+ "learning_rate": 0.0006,
+ "loss": 4.984277725219727,
+ "step": 3568
+ },
+ {
+ "epoch": 49.57292576419214,
+ "grad_norm": 0.014329628087580204,
+ "learning_rate": 0.0006,
+ "loss": 4.951558589935303,
+ "step": 3569
+ },
+ {
+ "epoch": 49.58689956331878,
+ "grad_norm": 0.017769070342183113,
+ "learning_rate": 0.0006,
+ "loss": 4.966532230377197,
+ "step": 3570
+ },
+ {
+ "epoch": 49.60087336244541,
+ "grad_norm": 0.01909453794360161,
+ "learning_rate": 0.0006,
+ "loss": 4.919890880584717,
+ "step": 3571
+ },
+ {
+ "epoch": 49.61484716157205,
+ "grad_norm": 0.017519688233733177,
+ "learning_rate": 0.0006,
+ "loss": 5.031286716461182,
+ "step": 3572
+ },
+ {
+ "epoch": 49.62882096069869,
+ "grad_norm": 0.017143897712230682,
+ "learning_rate": 0.0006,
+ "loss": 4.969764709472656,
+ "step": 3573
+ },
+ {
+ "epoch": 49.64279475982533,
+ "grad_norm": 0.01688285917043686,
+ "learning_rate": 0.0006,
+ "loss": 5.0733442306518555,
+ "step": 3574
+ },
+ {
+ "epoch": 49.65676855895197,
+ "grad_norm": 0.018547311425209045,
+ "learning_rate": 0.0006,
+ "loss": 4.980863571166992,
+ "step": 3575
+ },
+ {
+ "epoch": 49.670742358078606,
+ "grad_norm": 0.018877137452363968,
+ "learning_rate": 0.0006,
+ "loss": 4.954646110534668,
+ "step": 3576
+ },
+ {
+ "epoch": 49.68471615720524,
+ "grad_norm": 0.018681600689888,
+ "learning_rate": 0.0006,
+ "loss": 4.959702014923096,
+ "step": 3577
+ },
+ {
+ "epoch": 49.698689956331876,
+ "grad_norm": 0.018947336822748184,
+ "learning_rate": 0.0006,
+ "loss": 5.070502758026123,
+ "step": 3578
+ },
+ {
+ "epoch": 49.712663755458514,
+ "grad_norm": 0.018817484378814697,
+ "learning_rate": 0.0006,
+ "loss": 5.056508541107178,
+ "step": 3579
+ },
+ {
+ "epoch": 49.72663755458515,
+ "grad_norm": 0.01931724324822426,
+ "learning_rate": 0.0006,
+ "loss": 4.922451972961426,
+ "step": 3580
+ },
+ {
+ "epoch": 49.74061135371179,
+ "grad_norm": 0.022047819569706917,
+ "learning_rate": 0.0006,
+ "loss": 5.053170204162598,
+ "step": 3581
+ },
+ {
+ "epoch": 49.75458515283843,
+ "grad_norm": 0.02279801294207573,
+ "learning_rate": 0.0006,
+ "loss": 4.8768439292907715,
+ "step": 3582
+ },
+ {
+ "epoch": 49.76855895196506,
+ "grad_norm": 0.020661601796746254,
+ "learning_rate": 0.0006,
+ "loss": 4.907928466796875,
+ "step": 3583
+ },
+ {
+ "epoch": 49.7825327510917,
+ "grad_norm": 0.020479509606957436,
+ "learning_rate": 0.0006,
+ "loss": 5.040694236755371,
+ "step": 3584
+ },
+ {
+ "epoch": 49.79650655021834,
+ "grad_norm": 0.021654162555933,
+ "learning_rate": 0.0006,
+ "loss": 5.018434524536133,
+ "step": 3585
+ },
+ {
+ "epoch": 49.81048034934498,
+ "grad_norm": 0.022486824542284012,
+ "learning_rate": 0.0006,
+ "loss": 5.086538314819336,
+ "step": 3586
+ },
+ {
+ "epoch": 49.82445414847162,
+ "grad_norm": 0.02141369879245758,
+ "learning_rate": 0.0006,
+ "loss": 5.035694122314453,
+ "step": 3587
+ },
+ {
+ "epoch": 49.838427947598255,
+ "grad_norm": 0.015926677733659744,
+ "learning_rate": 0.0006,
+ "loss": 4.905038833618164,
+ "step": 3588
+ },
+ {
+ "epoch": 49.852401746724894,
+ "grad_norm": 0.015775298699736595,
+ "learning_rate": 0.0006,
+ "loss": 5.003902435302734,
+ "step": 3589
+ },
+ {
+ "epoch": 49.866375545851525,
+ "grad_norm": 0.01872118003666401,
+ "learning_rate": 0.0006,
+ "loss": 4.891060829162598,
+ "step": 3590
+ },
+ {
+ "epoch": 49.880349344978164,
+ "grad_norm": 0.02066192403435707,
+ "learning_rate": 0.0006,
+ "loss": 5.002897262573242,
+ "step": 3591
+ },
+ {
+ "epoch": 49.8943231441048,
+ "grad_norm": 0.02348714880645275,
+ "learning_rate": 0.0006,
+ "loss": 4.9189958572387695,
+ "step": 3592
+ },
+ {
+ "epoch": 49.90829694323144,
+ "grad_norm": 0.020402103662490845,
+ "learning_rate": 0.0006,
+ "loss": 4.992661476135254,
+ "step": 3593
+ },
+ {
+ "epoch": 49.92227074235808,
+ "grad_norm": 0.020466072484850883,
+ "learning_rate": 0.0006,
+ "loss": 4.962968826293945,
+ "step": 3594
+ },
+ {
+ "epoch": 49.93624454148472,
+ "grad_norm": 0.01922314241528511,
+ "learning_rate": 0.0006,
+ "loss": 4.897730827331543,
+ "step": 3595
+ },
+ {
+ "epoch": 49.95021834061135,
+ "grad_norm": 0.017228754237294197,
+ "learning_rate": 0.0006,
+ "loss": 4.964842796325684,
+ "step": 3596
+ },
+ {
+ "epoch": 49.96419213973799,
+ "grad_norm": 0.016021743416786194,
+ "learning_rate": 0.0006,
+ "loss": 4.910758972167969,
+ "step": 3597
+ },
+ {
+ "epoch": 49.97816593886463,
+ "grad_norm": 0.017985103651881218,
+ "learning_rate": 0.0006,
+ "loss": 4.916072368621826,
+ "step": 3598
+ },
+ {
+ "epoch": 49.992139737991266,
+ "grad_norm": 0.018780240789055824,
+ "learning_rate": 0.0006,
+ "loss": 4.94658088684082,
+ "step": 3599
+ },
+ {
+ "epoch": 50.0,
+ "grad_norm": 0.01776314526796341,
+ "learning_rate": 0.0006,
+ "loss": 4.969357967376709,
+ "step": 3600
+ },
+ {
+ "epoch": 50.0,
+ "eval_loss": 5.141158580780029,
+ "eval_runtime": 56.2332,
+ "eval_samples_per_second": 43.426,
+ "eval_steps_per_second": 1.369,
+ "step": 3600
+ },
+ {
+ "epoch": 50.01397379912664,
+ "grad_norm": 0.016308696940541267,
+ "learning_rate": 0.0006,
+ "loss": 4.924664497375488,
+ "step": 3601
+ },
+ {
+ "epoch": 50.02794759825328,
+ "grad_norm": 0.018810199573636055,
+ "learning_rate": 0.0006,
+ "loss": 4.82293701171875,
+ "step": 3602
+ },
+ {
+ "epoch": 50.041921397379916,
+ "grad_norm": 0.02227160707116127,
+ "learning_rate": 0.0006,
+ "loss": 4.9612298011779785,
+ "step": 3603
+ },
+ {
+ "epoch": 50.05589519650655,
+ "grad_norm": 0.025266462936997414,
+ "learning_rate": 0.0006,
+ "loss": 4.963835716247559,
+ "step": 3604
+ },
+ {
+ "epoch": 50.069868995633186,
+ "grad_norm": 0.028588181361556053,
+ "learning_rate": 0.0006,
+ "loss": 4.992118835449219,
+ "step": 3605
+ },
+ {
+ "epoch": 50.083842794759825,
+ "grad_norm": 0.02739487960934639,
+ "learning_rate": 0.0006,
+ "loss": 4.915547847747803,
+ "step": 3606
+ },
+ {
+ "epoch": 50.09781659388646,
+ "grad_norm": 0.027000300586223602,
+ "learning_rate": 0.0006,
+ "loss": 4.883801460266113,
+ "step": 3607
+ },
+ {
+ "epoch": 50.1117903930131,
+ "grad_norm": 0.023468483239412308,
+ "learning_rate": 0.0006,
+ "loss": 5.006168365478516,
+ "step": 3608
+ },
+ {
+ "epoch": 50.12576419213974,
+ "grad_norm": 0.026753634214401245,
+ "learning_rate": 0.0006,
+ "loss": 4.92526912689209,
+ "step": 3609
+ },
+ {
+ "epoch": 50.13973799126637,
+ "grad_norm": 0.026816733181476593,
+ "learning_rate": 0.0006,
+ "loss": 4.847546100616455,
+ "step": 3610
+ },
+ {
+ "epoch": 50.15371179039301,
+ "grad_norm": 0.027362391352653503,
+ "learning_rate": 0.0006,
+ "loss": 4.862548351287842,
+ "step": 3611
+ },
+ {
+ "epoch": 50.16768558951965,
+ "grad_norm": 0.03450845181941986,
+ "learning_rate": 0.0006,
+ "loss": 4.963202953338623,
+ "step": 3612
+ },
+ {
+ "epoch": 50.18165938864629,
+ "grad_norm": 0.03801408037543297,
+ "learning_rate": 0.0006,
+ "loss": 4.770240306854248,
+ "step": 3613
+ },
+ {
+ "epoch": 50.19563318777293,
+ "grad_norm": 0.03579140082001686,
+ "learning_rate": 0.0006,
+ "loss": 4.797981262207031,
+ "step": 3614
+ },
+ {
+ "epoch": 50.209606986899566,
+ "grad_norm": 0.029393760487437248,
+ "learning_rate": 0.0006,
+ "loss": 4.798581600189209,
+ "step": 3615
+ },
+ {
+ "epoch": 50.223580786026204,
+ "grad_norm": 0.026111142709851265,
+ "learning_rate": 0.0006,
+ "loss": 4.990712642669678,
+ "step": 3616
+ },
+ {
+ "epoch": 50.237554585152836,
+ "grad_norm": 0.02213387005031109,
+ "learning_rate": 0.0006,
+ "loss": 4.865842342376709,
+ "step": 3617
+ },
+ {
+ "epoch": 50.251528384279474,
+ "grad_norm": 0.025450684130191803,
+ "learning_rate": 0.0006,
+ "loss": 4.903746128082275,
+ "step": 3618
+ },
+ {
+ "epoch": 50.26550218340611,
+ "grad_norm": 0.020915117114782333,
+ "learning_rate": 0.0006,
+ "loss": 4.951366424560547,
+ "step": 3619
+ },
+ {
+ "epoch": 50.27947598253275,
+ "grad_norm": 0.020972849801182747,
+ "learning_rate": 0.0006,
+ "loss": 4.783311367034912,
+ "step": 3620
+ },
+ {
+ "epoch": 50.29344978165939,
+ "grad_norm": 0.016956273466348648,
+ "learning_rate": 0.0006,
+ "loss": 4.955767631530762,
+ "step": 3621
+ },
+ {
+ "epoch": 50.30742358078603,
+ "grad_norm": 0.018045252189040184,
+ "learning_rate": 0.0006,
+ "loss": 4.869322776794434,
+ "step": 3622
+ },
+ {
+ "epoch": 50.32139737991266,
+ "grad_norm": 0.018199188634753227,
+ "learning_rate": 0.0006,
+ "loss": 4.95680046081543,
+ "step": 3623
+ },
+ {
+ "epoch": 50.3353711790393,
+ "grad_norm": 0.016688082367181778,
+ "learning_rate": 0.0006,
+ "loss": 4.783535957336426,
+ "step": 3624
+ },
+ {
+ "epoch": 50.34934497816594,
+ "grad_norm": 0.017540352419018745,
+ "learning_rate": 0.0006,
+ "loss": 4.8236083984375,
+ "step": 3625
+ },
+ {
+ "epoch": 50.36331877729258,
+ "grad_norm": 0.018054187297821045,
+ "learning_rate": 0.0006,
+ "loss": 4.764655113220215,
+ "step": 3626
+ },
+ {
+ "epoch": 50.377292576419215,
+ "grad_norm": 0.01741131953895092,
+ "learning_rate": 0.0006,
+ "loss": 4.875613689422607,
+ "step": 3627
+ },
+ {
+ "epoch": 50.391266375545854,
+ "grad_norm": 0.01682269014418125,
+ "learning_rate": 0.0006,
+ "loss": 4.818509101867676,
+ "step": 3628
+ },
+ {
+ "epoch": 50.40524017467249,
+ "grad_norm": 0.016652341932058334,
+ "learning_rate": 0.0006,
+ "loss": 4.974020481109619,
+ "step": 3629
+ },
+ {
+ "epoch": 50.419213973799124,
+ "grad_norm": 0.02035282365977764,
+ "learning_rate": 0.0006,
+ "loss": 4.893037796020508,
+ "step": 3630
+ },
+ {
+ "epoch": 50.43318777292576,
+ "grad_norm": 0.019907061010599136,
+ "learning_rate": 0.0006,
+ "loss": 4.848119735717773,
+ "step": 3631
+ },
+ {
+ "epoch": 50.4471615720524,
+ "grad_norm": 0.023861493915319443,
+ "learning_rate": 0.0006,
+ "loss": 4.819353103637695,
+ "step": 3632
+ },
+ {
+ "epoch": 50.46113537117904,
+ "grad_norm": 0.025254026055336,
+ "learning_rate": 0.0006,
+ "loss": 4.769312858581543,
+ "step": 3633
+ },
+ {
+ "epoch": 50.47510917030568,
+ "grad_norm": 0.026154937222599983,
+ "learning_rate": 0.0006,
+ "loss": 4.7586774826049805,
+ "step": 3634
+ },
+ {
+ "epoch": 50.48908296943232,
+ "grad_norm": 0.026342090219259262,
+ "learning_rate": 0.0006,
+ "loss": 4.861218452453613,
+ "step": 3635
+ },
+ {
+ "epoch": 50.50305676855895,
+ "grad_norm": 0.028169559314846992,
+ "learning_rate": 0.0006,
+ "loss": 4.818380355834961,
+ "step": 3636
+ },
+ {
+ "epoch": 50.51703056768559,
+ "grad_norm": 0.02820264920592308,
+ "learning_rate": 0.0006,
+ "loss": 4.892871856689453,
+ "step": 3637
+ },
+ {
+ "epoch": 50.531004366812226,
+ "grad_norm": 0.03345080465078354,
+ "learning_rate": 0.0006,
+ "loss": 4.811084747314453,
+ "step": 3638
+ },
+ {
+ "epoch": 50.544978165938865,
+ "grad_norm": 0.04005736857652664,
+ "learning_rate": 0.0006,
+ "loss": 4.765410900115967,
+ "step": 3639
+ },
+ {
+ "epoch": 50.5589519650655,
+ "grad_norm": 0.04333870857954025,
+ "learning_rate": 0.0006,
+ "loss": 4.956520080566406,
+ "step": 3640
+ },
+ {
+ "epoch": 50.57292576419214,
+ "grad_norm": 0.04216952994465828,
+ "learning_rate": 0.0006,
+ "loss": 4.793441295623779,
+ "step": 3641
+ },
+ {
+ "epoch": 50.58689956331878,
+ "grad_norm": 0.02936951443552971,
+ "learning_rate": 0.0006,
+ "loss": 4.8447957038879395,
+ "step": 3642
+ },
+ {
+ "epoch": 50.60087336244541,
+ "grad_norm": 0.026937026530504227,
+ "learning_rate": 0.0006,
+ "loss": 4.962215423583984,
+ "step": 3643
+ },
+ {
+ "epoch": 50.61484716157205,
+ "grad_norm": 0.02628890983760357,
+ "learning_rate": 0.0006,
+ "loss": 4.883061408996582,
+ "step": 3644
+ },
+ {
+ "epoch": 50.62882096069869,
+ "grad_norm": 0.025984834879636765,
+ "learning_rate": 0.0006,
+ "loss": 4.768662452697754,
+ "step": 3645
+ },
+ {
+ "epoch": 50.64279475982533,
+ "grad_norm": 0.023998012766242027,
+ "learning_rate": 0.0006,
+ "loss": 4.851325035095215,
+ "step": 3646
+ },
+ {
+ "epoch": 50.65676855895197,
+ "grad_norm": 0.02143184095621109,
+ "learning_rate": 0.0006,
+ "loss": 4.8274688720703125,
+ "step": 3647
+ },
+ {
+ "epoch": 50.670742358078606,
+ "grad_norm": 0.02488900162279606,
+ "learning_rate": 0.0006,
+ "loss": 4.831140518188477,
+ "step": 3648
+ },
+ {
+ "epoch": 50.68471615720524,
+ "grad_norm": 0.02330533228814602,
+ "learning_rate": 0.0006,
+ "loss": 4.850462913513184,
+ "step": 3649
+ },
+ {
+ "epoch": 50.698689956331876,
+ "grad_norm": 0.02121943049132824,
+ "learning_rate": 0.0006,
+ "loss": 4.870253086090088,
+ "step": 3650
+ },
+ {
+ "epoch": 50.712663755458514,
+ "grad_norm": 0.02080935798585415,
+ "learning_rate": 0.0006,
+ "loss": 4.798770904541016,
+ "step": 3651
+ },
+ {
+ "epoch": 50.72663755458515,
+ "grad_norm": 0.019047029316425323,
+ "learning_rate": 0.0006,
+ "loss": 4.8427300453186035,
+ "step": 3652
+ },
+ {
+ "epoch": 50.74061135371179,
+ "grad_norm": 0.01824103854596615,
+ "learning_rate": 0.0006,
+ "loss": 4.820980548858643,
+ "step": 3653
+ },
+ {
+ "epoch": 50.75458515283843,
+ "grad_norm": 0.018129967153072357,
+ "learning_rate": 0.0006,
+ "loss": 4.946732997894287,
+ "step": 3654
+ },
+ {
+ "epoch": 50.76855895196506,
+ "grad_norm": 0.01784815452992916,
+ "learning_rate": 0.0006,
+ "loss": 4.829516410827637,
+ "step": 3655
+ },
+ {
+ "epoch": 50.7825327510917,
+ "grad_norm": 0.020021555945277214,
+ "learning_rate": 0.0006,
+ "loss": 4.780428409576416,
+ "step": 3656
+ },
+ {
+ "epoch": 50.79650655021834,
+ "grad_norm": 0.01863253302872181,
+ "learning_rate": 0.0006,
+ "loss": 4.791567802429199,
+ "step": 3657
+ },
+ {
+ "epoch": 50.81048034934498,
+ "grad_norm": 0.019151683896780014,
+ "learning_rate": 0.0006,
+ "loss": 4.824892997741699,
+ "step": 3658
+ },
+ {
+ "epoch": 50.82445414847162,
+ "grad_norm": 0.01891648955643177,
+ "learning_rate": 0.0006,
+ "loss": 4.789605140686035,
+ "step": 3659
+ },
+ {
+ "epoch": 50.838427947598255,
+ "grad_norm": 0.016019510105252266,
+ "learning_rate": 0.0006,
+ "loss": 4.779555320739746,
+ "step": 3660
+ },
+ {
+ "epoch": 50.852401746724894,
+ "grad_norm": 0.016096491366624832,
+ "learning_rate": 0.0006,
+ "loss": 4.753842353820801,
+ "step": 3661
+ },
+ {
+ "epoch": 50.866375545851525,
+ "grad_norm": 0.01693476364016533,
+ "learning_rate": 0.0006,
+ "loss": 4.768477439880371,
+ "step": 3662
+ },
+ {
+ "epoch": 50.880349344978164,
+ "grad_norm": 0.01749575324356556,
+ "learning_rate": 0.0006,
+ "loss": 4.779143333435059,
+ "step": 3663
+ },
+ {
+ "epoch": 50.8943231441048,
+ "grad_norm": 0.016851043328642845,
+ "learning_rate": 0.0006,
+ "loss": 4.780387878417969,
+ "step": 3664
+ },
+ {
+ "epoch": 50.90829694323144,
+ "grad_norm": 0.019045202061533928,
+ "learning_rate": 0.0006,
+ "loss": 4.769913673400879,
+ "step": 3665
+ },
+ {
+ "epoch": 50.92227074235808,
+ "grad_norm": 0.01954123005270958,
+ "learning_rate": 0.0006,
+ "loss": 4.769009590148926,
+ "step": 3666
+ },
+ {
+ "epoch": 50.93624454148472,
+ "grad_norm": 0.022527096793055534,
+ "learning_rate": 0.0006,
+ "loss": 4.789215087890625,
+ "step": 3667
+ },
+ {
+ "epoch": 50.95021834061135,
+ "grad_norm": 0.02871634066104889,
+ "learning_rate": 0.0006,
+ "loss": 4.749727725982666,
+ "step": 3668
+ },
+ {
+ "epoch": 50.96419213973799,
+ "grad_norm": 0.04045053571462631,
+ "learning_rate": 0.0006,
+ "loss": 4.909038543701172,
+ "step": 3669
+ },
+ {
+ "epoch": 50.97816593886463,
+ "grad_norm": 0.04156879335641861,
+ "learning_rate": 0.0006,
+ "loss": 4.876831531524658,
+ "step": 3670
+ },
+ {
+ "epoch": 50.992139737991266,
+ "grad_norm": 0.02621397376060486,
+ "learning_rate": 0.0006,
+ "loss": 4.797173500061035,
+ "step": 3671
+ },
+ {
+ "epoch": 51.0,
+ "grad_norm": 0.03389597684144974,
+ "learning_rate": 0.0006,
+ "loss": 4.8428850173950195,
+ "step": 3672
+ },
+ {
+ "epoch": 51.0,
+ "eval_loss": 4.989261627197266,
+ "eval_runtime": 56.358,
+ "eval_samples_per_second": 43.33,
+ "eval_steps_per_second": 1.366,
+ "step": 3672
+ },
+ {
+ "epoch": 51.01397379912664,
+ "grad_norm": 0.03216972202062607,
+ "learning_rate": 0.0006,
+ "loss": 4.755102634429932,
+ "step": 3673
+ },
+ {
+ "epoch": 51.02794759825328,
+ "grad_norm": 0.03190374746918678,
+ "learning_rate": 0.0006,
+ "loss": 4.720424652099609,
+ "step": 3674
+ },
+ {
+ "epoch": 51.041921397379916,
+ "grad_norm": 0.031727299094200134,
+ "learning_rate": 0.0006,
+ "loss": 4.803277969360352,
+ "step": 3675
+ },
+ {
+ "epoch": 51.05589519650655,
+ "grad_norm": 0.02473079040646553,
+ "learning_rate": 0.0006,
+ "loss": 4.761087417602539,
+ "step": 3676
+ },
+ {
+ "epoch": 51.069868995633186,
+ "grad_norm": 0.020360209047794342,
+ "learning_rate": 0.0006,
+ "loss": 4.753884792327881,
+ "step": 3677
+ },
+ {
+ "epoch": 51.083842794759825,
+ "grad_norm": 0.02078923024237156,
+ "learning_rate": 0.0006,
+ "loss": 4.647130489349365,
+ "step": 3678
+ },
+ {
+ "epoch": 51.09781659388646,
+ "grad_norm": 0.019387200474739075,
+ "learning_rate": 0.0006,
+ "loss": 4.74135160446167,
+ "step": 3679
+ },
+ {
+ "epoch": 51.1117903930131,
+ "grad_norm": 0.01874949410557747,
+ "learning_rate": 0.0006,
+ "loss": 4.795329570770264,
+ "step": 3680
+ },
+ {
+ "epoch": 51.12576419213974,
+ "grad_norm": 0.019331837072968483,
+ "learning_rate": 0.0006,
+ "loss": 4.897967338562012,
+ "step": 3681
+ },
+ {
+ "epoch": 51.13973799126637,
+ "grad_norm": 0.017066050320863724,
+ "learning_rate": 0.0006,
+ "loss": 4.825351715087891,
+ "step": 3682
+ },
+ {
+ "epoch": 51.15371179039301,
+ "grad_norm": 0.01757810264825821,
+ "learning_rate": 0.0006,
+ "loss": 4.719335079193115,
+ "step": 3683
+ },
+ {
+ "epoch": 51.16768558951965,
+ "grad_norm": 0.018174266442656517,
+ "learning_rate": 0.0006,
+ "loss": 4.732680320739746,
+ "step": 3684
+ },
+ {
+ "epoch": 51.18165938864629,
+ "grad_norm": 0.018497727811336517,
+ "learning_rate": 0.0006,
+ "loss": 4.731050491333008,
+ "step": 3685
+ },
+ {
+ "epoch": 51.19563318777293,
+ "grad_norm": 0.0186338871717453,
+ "learning_rate": 0.0006,
+ "loss": 4.748035430908203,
+ "step": 3686
+ },
+ {
+ "epoch": 51.209606986899566,
+ "grad_norm": 0.018531711772084236,
+ "learning_rate": 0.0006,
+ "loss": 4.9007463455200195,
+ "step": 3687
+ },
+ {
+ "epoch": 51.223580786026204,
+ "grad_norm": 0.018221354112029076,
+ "learning_rate": 0.0006,
+ "loss": 4.84077262878418,
+ "step": 3688
+ },
+ {
+ "epoch": 51.237554585152836,
+ "grad_norm": 0.0203656405210495,
+ "learning_rate": 0.0006,
+ "loss": 4.73009729385376,
+ "step": 3689
+ },
+ {
+ "epoch": 51.251528384279474,
+ "grad_norm": 0.01964893378317356,
+ "learning_rate": 0.0006,
+ "loss": 4.772583961486816,
+ "step": 3690
+ },
+ {
+ "epoch": 51.26550218340611,
+ "grad_norm": 0.019606510177254677,
+ "learning_rate": 0.0006,
+ "loss": 4.599812030792236,
+ "step": 3691
+ },
+ {
+ "epoch": 51.27947598253275,
+ "grad_norm": 0.020501255989074707,
+ "learning_rate": 0.0006,
+ "loss": 4.809898853302002,
+ "step": 3692
+ },
+ {
+ "epoch": 51.29344978165939,
+ "grad_norm": 0.01870838552713394,
+ "learning_rate": 0.0006,
+ "loss": 4.741898059844971,
+ "step": 3693
+ },
+ {
+ "epoch": 51.30742358078603,
+ "grad_norm": 0.01802559196949005,
+ "learning_rate": 0.0006,
+ "loss": 4.746562480926514,
+ "step": 3694
+ },
+ {
+ "epoch": 51.32139737991266,
+ "grad_norm": 0.02053048275411129,
+ "learning_rate": 0.0006,
+ "loss": 4.69932746887207,
+ "step": 3695
+ },
+ {
+ "epoch": 51.3353711790393,
+ "grad_norm": 0.02661825157701969,
+ "learning_rate": 0.0006,
+ "loss": 4.68748664855957,
+ "step": 3696
+ },
+ {
+ "epoch": 51.34934497816594,
+ "grad_norm": 0.028269147500395775,
+ "learning_rate": 0.0006,
+ "loss": 4.731565475463867,
+ "step": 3697
+ },
+ {
+ "epoch": 51.36331877729258,
+ "grad_norm": 0.02650964818894863,
+ "learning_rate": 0.0006,
+ "loss": 4.733857154846191,
+ "step": 3698
+ },
+ {
+ "epoch": 51.377292576419215,
+ "grad_norm": 0.023696066811680794,
+ "learning_rate": 0.0006,
+ "loss": 4.812362194061279,
+ "step": 3699
+ },
+ {
+ "epoch": 51.391266375545854,
+ "grad_norm": 0.02365880273282528,
+ "learning_rate": 0.0006,
+ "loss": 4.733426094055176,
+ "step": 3700
+ },
+ {
+ "epoch": 51.40524017467249,
+ "grad_norm": 0.026249447837471962,
+ "learning_rate": 0.0006,
+ "loss": 4.715622901916504,
+ "step": 3701
+ },
+ {
+ "epoch": 51.419213973799124,
+ "grad_norm": 0.03301899880170822,
+ "learning_rate": 0.0006,
+ "loss": 4.7223663330078125,
+ "step": 3702
+ },
+ {
+ "epoch": 51.43318777292576,
+ "grad_norm": 0.044686660170555115,
+ "learning_rate": 0.0006,
+ "loss": 4.8121161460876465,
+ "step": 3703
+ },
+ {
+ "epoch": 51.4471615720524,
+ "grad_norm": 0.04733522981405258,
+ "learning_rate": 0.0006,
+ "loss": 4.841394424438477,
+ "step": 3704
+ },
+ {
+ "epoch": 51.46113537117904,
+ "grad_norm": 0.0395209975540638,
+ "learning_rate": 0.0006,
+ "loss": 4.773837089538574,
+ "step": 3705
+ },
+ {
+ "epoch": 51.47510917030568,
+ "grad_norm": 0.032790958881378174,
+ "learning_rate": 0.0006,
+ "loss": 4.833066463470459,
+ "step": 3706
+ },
+ {
+ "epoch": 51.48908296943232,
+ "grad_norm": 0.030578427016735077,
+ "learning_rate": 0.0006,
+ "loss": 4.685190200805664,
+ "step": 3707
+ },
+ {
+ "epoch": 51.50305676855895,
+ "grad_norm": 0.031619083136320114,
+ "learning_rate": 0.0006,
+ "loss": 4.740378379821777,
+ "step": 3708
+ },
+ {
+ "epoch": 51.51703056768559,
+ "grad_norm": 0.02385825850069523,
+ "learning_rate": 0.0006,
+ "loss": 4.812911510467529,
+ "step": 3709
+ },
+ {
+ "epoch": 51.531004366812226,
+ "grad_norm": 0.023523425683379173,
+ "learning_rate": 0.0006,
+ "loss": 4.669580459594727,
+ "step": 3710
+ },
+ {
+ "epoch": 51.544978165938865,
+ "grad_norm": 0.021316280588507652,
+ "learning_rate": 0.0006,
+ "loss": 4.76888370513916,
+ "step": 3711
+ },
+ {
+ "epoch": 51.5589519650655,
+ "grad_norm": 0.024663880467414856,
+ "learning_rate": 0.0006,
+ "loss": 4.702592372894287,
+ "step": 3712
+ },
+ {
+ "epoch": 51.57292576419214,
+ "grad_norm": 0.024173280224204063,
+ "learning_rate": 0.0006,
+ "loss": 4.710604667663574,
+ "step": 3713
+ },
+ {
+ "epoch": 51.58689956331878,
+ "grad_norm": 0.02754114381968975,
+ "learning_rate": 0.0006,
+ "loss": 4.746356964111328,
+ "step": 3714
+ },
+ {
+ "epoch": 51.60087336244541,
+ "grad_norm": 0.027903622016310692,
+ "learning_rate": 0.0006,
+ "loss": 4.7546491622924805,
+ "step": 3715
+ },
+ {
+ "epoch": 51.61484716157205,
+ "grad_norm": 0.024871617555618286,
+ "learning_rate": 0.0006,
+ "loss": 4.605746269226074,
+ "step": 3716
+ },
+ {
+ "epoch": 51.62882096069869,
+ "grad_norm": 0.021239031106233597,
+ "learning_rate": 0.0006,
+ "loss": 4.6942458152771,
+ "step": 3717
+ },
+ {
+ "epoch": 51.64279475982533,
+ "grad_norm": 0.01925622671842575,
+ "learning_rate": 0.0006,
+ "loss": 4.593558311462402,
+ "step": 3718
+ },
+ {
+ "epoch": 51.65676855895197,
+ "grad_norm": 0.01956283673644066,
+ "learning_rate": 0.0006,
+ "loss": 4.631172180175781,
+ "step": 3719
+ },
+ {
+ "epoch": 51.670742358078606,
+ "grad_norm": 0.02269871160387993,
+ "learning_rate": 0.0006,
+ "loss": 4.729236602783203,
+ "step": 3720
+ },
+ {
+ "epoch": 51.68471615720524,
+ "grad_norm": 0.02253916673362255,
+ "learning_rate": 0.0006,
+ "loss": 4.7191033363342285,
+ "step": 3721
+ },
+ {
+ "epoch": 51.698689956331876,
+ "grad_norm": 0.026231050491333008,
+ "learning_rate": 0.0006,
+ "loss": 4.811850547790527,
+ "step": 3722
+ },
+ {
+ "epoch": 51.712663755458514,
+ "grad_norm": 0.02558821812272072,
+ "learning_rate": 0.0006,
+ "loss": 4.681466102600098,
+ "step": 3723
+ },
+ {
+ "epoch": 51.72663755458515,
+ "grad_norm": 0.02124209702014923,
+ "learning_rate": 0.0006,
+ "loss": 4.693664073944092,
+ "step": 3724
+ },
+ {
+ "epoch": 51.74061135371179,
+ "grad_norm": 0.016638319939374924,
+ "learning_rate": 0.0006,
+ "loss": 4.771440029144287,
+ "step": 3725
+ },
+ {
+ "epoch": 51.75458515283843,
+ "grad_norm": 0.018048347905278206,
+ "learning_rate": 0.0006,
+ "loss": 4.643252372741699,
+ "step": 3726
+ },
+ {
+ "epoch": 51.76855895196506,
+ "grad_norm": 0.019904717803001404,
+ "learning_rate": 0.0006,
+ "loss": 4.66530704498291,
+ "step": 3727
+ },
+ {
+ "epoch": 51.7825327510917,
+ "grad_norm": 0.0194614976644516,
+ "learning_rate": 0.0006,
+ "loss": 4.711943626403809,
+ "step": 3728
+ },
+ {
+ "epoch": 51.79650655021834,
+ "grad_norm": 0.019175555557012558,
+ "learning_rate": 0.0006,
+ "loss": 4.721931457519531,
+ "step": 3729
+ },
+ {
+ "epoch": 51.81048034934498,
+ "grad_norm": 0.02093447372317314,
+ "learning_rate": 0.0006,
+ "loss": 4.633059024810791,
+ "step": 3730
+ },
+ {
+ "epoch": 51.82445414847162,
+ "grad_norm": 0.01930283196270466,
+ "learning_rate": 0.0006,
+ "loss": 4.659373760223389,
+ "step": 3731
+ },
+ {
+ "epoch": 51.838427947598255,
+ "grad_norm": 0.015350676141679287,
+ "learning_rate": 0.0006,
+ "loss": 4.763708114624023,
+ "step": 3732
+ },
+ {
+ "epoch": 51.852401746724894,
+ "grad_norm": 0.014139448292553425,
+ "learning_rate": 0.0006,
+ "loss": 4.753891468048096,
+ "step": 3733
+ },
+ {
+ "epoch": 51.866375545851525,
+ "grad_norm": 0.016215872019529343,
+ "learning_rate": 0.0006,
+ "loss": 4.658967018127441,
+ "step": 3734
+ },
+ {
+ "epoch": 51.880349344978164,
+ "grad_norm": 0.017328619956970215,
+ "learning_rate": 0.0006,
+ "loss": 4.696579933166504,
+ "step": 3735
+ },
+ {
+ "epoch": 51.8943231441048,
+ "grad_norm": 0.018103396520018578,
+ "learning_rate": 0.0006,
+ "loss": 4.627464771270752,
+ "step": 3736
+ },
+ {
+ "epoch": 51.90829694323144,
+ "grad_norm": 0.01715235412120819,
+ "learning_rate": 0.0006,
+ "loss": 4.7368011474609375,
+ "step": 3737
+ },
+ {
+ "epoch": 51.92227074235808,
+ "grad_norm": 0.014790094457566738,
+ "learning_rate": 0.0006,
+ "loss": 4.629441738128662,
+ "step": 3738
+ },
+ {
+ "epoch": 51.93624454148472,
+ "grad_norm": 0.017114296555519104,
+ "learning_rate": 0.0006,
+ "loss": 4.761256694793701,
+ "step": 3739
+ },
+ {
+ "epoch": 51.95021834061135,
+ "grad_norm": 0.019009845331311226,
+ "learning_rate": 0.0006,
+ "loss": 4.789155006408691,
+ "step": 3740
+ },
+ {
+ "epoch": 51.96419213973799,
+ "grad_norm": 0.019696692004799843,
+ "learning_rate": 0.0006,
+ "loss": 4.721269607543945,
+ "step": 3741
+ },
+ {
+ "epoch": 51.97816593886463,
+ "grad_norm": 0.020542822778224945,
+ "learning_rate": 0.0006,
+ "loss": 4.687038898468018,
+ "step": 3742
+ },
+ {
+ "epoch": 51.992139737991266,
+ "grad_norm": 0.021267160773277283,
+ "learning_rate": 0.0006,
+ "loss": 4.747860908508301,
+ "step": 3743
+ },
+ {
+ "epoch": 52.0,
+ "grad_norm": 0.025049567222595215,
+ "learning_rate": 0.0006,
+ "loss": 4.8473100662231445,
+ "step": 3744
+ },
+ {
+ "epoch": 52.0,
+ "eval_loss": 4.943043231964111,
+ "eval_runtime": 56.4929,
+ "eval_samples_per_second": 43.227,
+ "eval_steps_per_second": 1.363,
+ "step": 3744
+ },
+ {
+ "epoch": 52.01397379912664,
+ "grad_norm": 0.03054986707866192,
+ "learning_rate": 0.0006,
+ "loss": 4.856931686401367,
+ "step": 3745
+ },
+ {
+ "epoch": 52.02794759825328,
+ "grad_norm": 0.03838375583291054,
+ "learning_rate": 0.0006,
+ "loss": 4.699807167053223,
+ "step": 3746
+ },
+ {
+ "epoch": 52.041921397379916,
+ "grad_norm": 0.03383488208055496,
+ "learning_rate": 0.0006,
+ "loss": 4.6626787185668945,
+ "step": 3747
+ },
+ {
+ "epoch": 52.05589519650655,
+ "grad_norm": 0.0201561376452446,
+ "learning_rate": 0.0006,
+ "loss": 4.621831893920898,
+ "step": 3748
+ },
+ {
+ "epoch": 52.069868995633186,
+ "grad_norm": 0.028749438002705574,
+ "learning_rate": 0.0006,
+ "loss": 4.682540416717529,
+ "step": 3749
+ },
+ {
+ "epoch": 52.083842794759825,
+ "grad_norm": 0.02681174874305725,
+ "learning_rate": 0.0006,
+ "loss": 4.629997253417969,
+ "step": 3750
+ },
+ {
+ "epoch": 52.09781659388646,
+ "grad_norm": 0.024063318967819214,
+ "learning_rate": 0.0006,
+ "loss": 4.573021411895752,
+ "step": 3751
+ },
+ {
+ "epoch": 52.1117903930131,
+ "grad_norm": 0.02324330434203148,
+ "learning_rate": 0.0006,
+ "loss": 4.739465236663818,
+ "step": 3752
+ },
+ {
+ "epoch": 52.12576419213974,
+ "grad_norm": 0.0225426834076643,
+ "learning_rate": 0.0006,
+ "loss": 4.7335333824157715,
+ "step": 3753
+ },
+ {
+ "epoch": 52.13973799126637,
+ "grad_norm": 0.019593428820371628,
+ "learning_rate": 0.0006,
+ "loss": 4.704958438873291,
+ "step": 3754
+ },
+ {
+ "epoch": 52.15371179039301,
+ "grad_norm": 0.020777082070708275,
+ "learning_rate": 0.0006,
+ "loss": 4.638696670532227,
+ "step": 3755
+ },
+ {
+ "epoch": 52.16768558951965,
+ "grad_norm": 0.021525604650378227,
+ "learning_rate": 0.0006,
+ "loss": 4.643706798553467,
+ "step": 3756
+ },
+ {
+ "epoch": 52.18165938864629,
+ "grad_norm": 0.018943654373288155,
+ "learning_rate": 0.0006,
+ "loss": 4.5173211097717285,
+ "step": 3757
+ },
+ {
+ "epoch": 52.19563318777293,
+ "grad_norm": 0.01856580562889576,
+ "learning_rate": 0.0006,
+ "loss": 4.678788661956787,
+ "step": 3758
+ },
+ {
+ "epoch": 52.209606986899566,
+ "grad_norm": 0.019417259842157364,
+ "learning_rate": 0.0006,
+ "loss": 4.587244987487793,
+ "step": 3759
+ },
+ {
+ "epoch": 52.223580786026204,
+ "grad_norm": 0.01911911740899086,
+ "learning_rate": 0.0006,
+ "loss": 4.688503265380859,
+ "step": 3760
+ },
+ {
+ "epoch": 52.237554585152836,
+ "grad_norm": 0.021017879247665405,
+ "learning_rate": 0.0006,
+ "loss": 4.65829610824585,
+ "step": 3761
+ },
+ {
+ "epoch": 52.251528384279474,
+ "grad_norm": 0.02048875205218792,
+ "learning_rate": 0.0006,
+ "loss": 4.624055862426758,
+ "step": 3762
+ },
+ {
+ "epoch": 52.26550218340611,
+ "grad_norm": 0.01932772807776928,
+ "learning_rate": 0.0006,
+ "loss": 4.556562900543213,
+ "step": 3763
+ },
+ {
+ "epoch": 52.27947598253275,
+ "grad_norm": 0.01871614344418049,
+ "learning_rate": 0.0006,
+ "loss": 4.581201076507568,
+ "step": 3764
+ },
+ {
+ "epoch": 52.29344978165939,
+ "grad_norm": 0.017253106459975243,
+ "learning_rate": 0.0006,
+ "loss": 4.605219841003418,
+ "step": 3765
+ },
+ {
+ "epoch": 52.30742358078603,
+ "grad_norm": 0.01831238530576229,
+ "learning_rate": 0.0006,
+ "loss": 4.759900093078613,
+ "step": 3766
+ },
+ {
+ "epoch": 52.32139737991266,
+ "grad_norm": 0.017946207895874977,
+ "learning_rate": 0.0006,
+ "loss": 4.5598344802856445,
+ "step": 3767
+ },
+ {
+ "epoch": 52.3353711790393,
+ "grad_norm": 0.01680106669664383,
+ "learning_rate": 0.0006,
+ "loss": 4.605034828186035,
+ "step": 3768
+ },
+ {
+ "epoch": 52.34934497816594,
+ "grad_norm": 0.021482540294528008,
+ "learning_rate": 0.0006,
+ "loss": 4.620411396026611,
+ "step": 3769
+ },
+ {
+ "epoch": 52.36331877729258,
+ "grad_norm": 0.02210252732038498,
+ "learning_rate": 0.0006,
+ "loss": 4.667577743530273,
+ "step": 3770
+ },
+ {
+ "epoch": 52.377292576419215,
+ "grad_norm": 0.02237485907971859,
+ "learning_rate": 0.0006,
+ "loss": 4.628812789916992,
+ "step": 3771
+ },
+ {
+ "epoch": 52.391266375545854,
+ "grad_norm": 0.02196541801095009,
+ "learning_rate": 0.0006,
+ "loss": 4.572988510131836,
+ "step": 3772
+ },
+ {
+ "epoch": 52.40524017467249,
+ "grad_norm": 0.01930229924619198,
+ "learning_rate": 0.0006,
+ "loss": 4.588339805603027,
+ "step": 3773
+ },
+ {
+ "epoch": 52.419213973799124,
+ "grad_norm": 0.02222434990108013,
+ "learning_rate": 0.0006,
+ "loss": 4.766482353210449,
+ "step": 3774
+ },
+ {
+ "epoch": 52.43318777292576,
+ "grad_norm": 0.028806306421756744,
+ "learning_rate": 0.0006,
+ "loss": 4.617892742156982,
+ "step": 3775
+ },
+ {
+ "epoch": 52.4471615720524,
+ "grad_norm": 0.03306042402982712,
+ "learning_rate": 0.0006,
+ "loss": 4.665334701538086,
+ "step": 3776
+ },
+ {
+ "epoch": 52.46113537117904,
+ "grad_norm": 0.0358956903219223,
+ "learning_rate": 0.0006,
+ "loss": 4.566668510437012,
+ "step": 3777
+ },
+ {
+ "epoch": 52.47510917030568,
+ "grad_norm": 0.030406497418880463,
+ "learning_rate": 0.0006,
+ "loss": 4.513612270355225,
+ "step": 3778
+ },
+ {
+ "epoch": 52.48908296943232,
+ "grad_norm": 0.030295712873339653,
+ "learning_rate": 0.0006,
+ "loss": 4.640189170837402,
+ "step": 3779
+ },
+ {
+ "epoch": 52.50305676855895,
+ "grad_norm": 0.024981003254652023,
+ "learning_rate": 0.0006,
+ "loss": 4.57558012008667,
+ "step": 3780
+ },
+ {
+ "epoch": 52.51703056768559,
+ "grad_norm": 0.024990713223814964,
+ "learning_rate": 0.0006,
+ "loss": 4.708589553833008,
+ "step": 3781
+ },
+ {
+ "epoch": 52.531004366812226,
+ "grad_norm": 0.025034775957465172,
+ "learning_rate": 0.0006,
+ "loss": 4.70468282699585,
+ "step": 3782
+ },
+ {
+ "epoch": 52.544978165938865,
+ "grad_norm": 0.028243759647011757,
+ "learning_rate": 0.0006,
+ "loss": 4.712060451507568,
+ "step": 3783
+ },
+ {
+ "epoch": 52.5589519650655,
+ "grad_norm": 0.026984544470906258,
+ "learning_rate": 0.0006,
+ "loss": 4.505004405975342,
+ "step": 3784
+ },
+ {
+ "epoch": 52.57292576419214,
+ "grad_norm": 0.050843529403209686,
+ "learning_rate": 0.0006,
+ "loss": 4.593271255493164,
+ "step": 3785
+ },
+ {
+ "epoch": 52.58689956331878,
+ "grad_norm": 0.25422802567481995,
+ "learning_rate": 0.0006,
+ "loss": 4.7178802490234375,
+ "step": 3786
+ },
+ {
+ "epoch": 52.60087336244541,
+ "grad_norm": 0.6078406572341919,
+ "learning_rate": 0.0006,
+ "loss": 6.144372940063477,
+ "step": 3787
+ },
+ {
+ "epoch": 52.61484716157205,
+ "grad_norm": 0.25933101773262024,
+ "learning_rate": 0.0006,
+ "loss": 5.971083641052246,
+ "step": 3788
+ },
+ {
+ "epoch": 52.62882096069869,
+ "grad_norm": 0.294657826423645,
+ "learning_rate": 0.0006,
+ "loss": 6.204090118408203,
+ "step": 3789
+ },
+ {
+ "epoch": 52.64279475982533,
+ "grad_norm": 0.23641522228717804,
+ "learning_rate": 0.0006,
+ "loss": 6.869574546813965,
+ "step": 3790
+ },
+ {
+ "epoch": 52.65676855895197,
+ "grad_norm": 0.23717311024665833,
+ "learning_rate": 0.0006,
+ "loss": 6.739034175872803,
+ "step": 3791
+ },
+ {
+ "epoch": 52.670742358078606,
+ "grad_norm": 0.10300362855195999,
+ "learning_rate": 0.0006,
+ "loss": 6.49191951751709,
+ "step": 3792
+ },
+ {
+ "epoch": 52.68471615720524,
+ "grad_norm": 0.1484968513250351,
+ "learning_rate": 0.0006,
+ "loss": 6.6021270751953125,
+ "step": 3793
+ },
+ {
+ "epoch": 52.698689956331876,
+ "grad_norm": 0.08052115142345428,
+ "learning_rate": 0.0006,
+ "loss": 6.432796478271484,
+ "step": 3794
+ },
+ {
+ "epoch": 52.712663755458514,
+ "grad_norm": 0.06869253516197205,
+ "learning_rate": 0.0006,
+ "loss": 6.212784767150879,
+ "step": 3795
+ },
+ {
+ "epoch": 52.72663755458515,
+ "grad_norm": 0.05589582398533821,
+ "learning_rate": 0.0006,
+ "loss": 6.168288707733154,
+ "step": 3796
+ },
+ {
+ "epoch": 52.74061135371179,
+ "grad_norm": 0.04840940609574318,
+ "learning_rate": 0.0006,
+ "loss": 6.163174629211426,
+ "step": 3797
+ },
+ {
+ "epoch": 52.75458515283843,
+ "grad_norm": 0.04578601196408272,
+ "learning_rate": 0.0006,
+ "loss": 6.038390636444092,
+ "step": 3798
+ },
+ {
+ "epoch": 52.76855895196506,
+ "grad_norm": 0.04102053865790367,
+ "learning_rate": 0.0006,
+ "loss": 5.929539680480957,
+ "step": 3799
+ },
+ {
+ "epoch": 52.7825327510917,
+ "grad_norm": 0.038398485630750656,
+ "learning_rate": 0.0006,
+ "loss": 6.007378101348877,
+ "step": 3800
+ },
+ {
+ "epoch": 52.79650655021834,
+ "grad_norm": 0.03466642275452614,
+ "learning_rate": 0.0006,
+ "loss": 5.767291069030762,
+ "step": 3801
+ },
+ {
+ "epoch": 52.81048034934498,
+ "grad_norm": 0.03328905999660492,
+ "learning_rate": 0.0006,
+ "loss": 5.785032272338867,
+ "step": 3802
+ },
+ {
+ "epoch": 52.82445414847162,
+ "grad_norm": 0.03186681121587753,
+ "learning_rate": 0.0006,
+ "loss": 5.6942057609558105,
+ "step": 3803
+ },
+ {
+ "epoch": 52.838427947598255,
+ "grad_norm": 0.030553551390767097,
+ "learning_rate": 0.0006,
+ "loss": 5.680774688720703,
+ "step": 3804
+ },
+ {
+ "epoch": 52.852401746724894,
+ "grad_norm": 0.02578839473426342,
+ "learning_rate": 0.0006,
+ "loss": 5.703461647033691,
+ "step": 3805
+ },
+ {
+ "epoch": 52.866375545851525,
+ "grad_norm": 0.026261312887072563,
+ "learning_rate": 0.0006,
+ "loss": 5.642549514770508,
+ "step": 3806
+ },
+ {
+ "epoch": 52.880349344978164,
+ "grad_norm": 0.02782430872321129,
+ "learning_rate": 0.0006,
+ "loss": 5.563597679138184,
+ "step": 3807
+ },
+ {
+ "epoch": 52.8943231441048,
+ "grad_norm": 0.024467799812555313,
+ "learning_rate": 0.0006,
+ "loss": 5.583395957946777,
+ "step": 3808
+ },
+ {
+ "epoch": 52.90829694323144,
+ "grad_norm": 0.022602897137403488,
+ "learning_rate": 0.0006,
+ "loss": 5.50742769241333,
+ "step": 3809
+ },
+ {
+ "epoch": 52.92227074235808,
+ "grad_norm": 0.024292029440402985,
+ "learning_rate": 0.0006,
+ "loss": 5.367804527282715,
+ "step": 3810
+ },
+ {
+ "epoch": 52.93624454148472,
+ "grad_norm": 0.021062295883893967,
+ "learning_rate": 0.0006,
+ "loss": 5.423705101013184,
+ "step": 3811
+ },
+ {
+ "epoch": 52.95021834061135,
+ "grad_norm": 0.022226469591259956,
+ "learning_rate": 0.0006,
+ "loss": 5.344039440155029,
+ "step": 3812
+ },
+ {
+ "epoch": 52.96419213973799,
+ "grad_norm": 0.023691752925515175,
+ "learning_rate": 0.0006,
+ "loss": 5.2581892013549805,
+ "step": 3813
+ },
+ {
+ "epoch": 52.97816593886463,
+ "grad_norm": 0.02181999757885933,
+ "learning_rate": 0.0006,
+ "loss": 5.2710113525390625,
+ "step": 3814
+ },
+ {
+ "epoch": 52.992139737991266,
+ "grad_norm": 0.025295857340097427,
+ "learning_rate": 0.0006,
+ "loss": 5.278355598449707,
+ "step": 3815
+ },
+ {
+ "epoch": 53.0,
+ "grad_norm": 0.02216828241944313,
+ "learning_rate": 0.0006,
+ "loss": 5.127945899963379,
+ "step": 3816
+ },
+ {
+ "epoch": 53.0,
+ "eval_loss": 5.3411736488342285,
+ "eval_runtime": 56.2837,
+ "eval_samples_per_second": 43.387,
+ "eval_steps_per_second": 1.368,
+ "step": 3816
+ },
+ {
+ "epoch": 53.01397379912664,
+ "grad_norm": 0.022496026009321213,
+ "learning_rate": 0.0006,
+ "loss": 5.19655704498291,
+ "step": 3817
+ },
+ {
+ "epoch": 53.02794759825328,
+ "grad_norm": 0.03201168775558472,
+ "learning_rate": 0.0006,
+ "loss": 5.094621181488037,
+ "step": 3818
+ },
+ {
+ "epoch": 53.041921397379916,
+ "grad_norm": 0.02683178335428238,
+ "learning_rate": 0.0006,
+ "loss": 5.19899845123291,
+ "step": 3819
+ },
+ {
+ "epoch": 53.05589519650655,
+ "grad_norm": 0.022659169510006905,
+ "learning_rate": 0.0006,
+ "loss": 5.184103012084961,
+ "step": 3820
+ },
+ {
+ "epoch": 53.069868995633186,
+ "grad_norm": 0.024784443899989128,
+ "learning_rate": 0.0006,
+ "loss": 5.124286651611328,
+ "step": 3821
+ },
+ {
+ "epoch": 53.083842794759825,
+ "grad_norm": 0.028710370883345604,
+ "learning_rate": 0.0006,
+ "loss": 5.130426406860352,
+ "step": 3822
+ },
+ {
+ "epoch": 53.09781659388646,
+ "grad_norm": 0.02366088517010212,
+ "learning_rate": 0.0006,
+ "loss": 5.1453399658203125,
+ "step": 3823
+ },
+ {
+ "epoch": 53.1117903930131,
+ "grad_norm": 0.01813938096165657,
+ "learning_rate": 0.0006,
+ "loss": 5.05271053314209,
+ "step": 3824
+ },
+ {
+ "epoch": 53.12576419213974,
+ "grad_norm": 0.02051435597240925,
+ "learning_rate": 0.0006,
+ "loss": 5.023038864135742,
+ "step": 3825
+ },
+ {
+ "epoch": 53.13973799126637,
+ "grad_norm": 0.02185754105448723,
+ "learning_rate": 0.0006,
+ "loss": 4.8950018882751465,
+ "step": 3826
+ },
+ {
+ "epoch": 53.15371179039301,
+ "grad_norm": 0.017223743721842766,
+ "learning_rate": 0.0006,
+ "loss": 5.078621864318848,
+ "step": 3827
+ },
+ {
+ "epoch": 53.16768558951965,
+ "grad_norm": 0.01832217164337635,
+ "learning_rate": 0.0006,
+ "loss": 4.921341419219971,
+ "step": 3828
+ },
+ {
+ "epoch": 53.18165938864629,
+ "grad_norm": 0.02114005573093891,
+ "learning_rate": 0.0006,
+ "loss": 4.997865676879883,
+ "step": 3829
+ },
+ {
+ "epoch": 53.19563318777293,
+ "grad_norm": 0.021426431834697723,
+ "learning_rate": 0.0006,
+ "loss": 4.877785682678223,
+ "step": 3830
+ },
+ {
+ "epoch": 53.209606986899566,
+ "grad_norm": 0.021327560767531395,
+ "learning_rate": 0.0006,
+ "loss": 4.847512245178223,
+ "step": 3831
+ },
+ {
+ "epoch": 53.223580786026204,
+ "grad_norm": 0.027825530618429184,
+ "learning_rate": 0.0006,
+ "loss": 5.004823207855225,
+ "step": 3832
+ },
+ {
+ "epoch": 53.237554585152836,
+ "grad_norm": 0.03756481409072876,
+ "learning_rate": 0.0006,
+ "loss": 4.817290306091309,
+ "step": 3833
+ },
+ {
+ "epoch": 53.251528384279474,
+ "grad_norm": 0.04933721572160721,
+ "learning_rate": 0.0006,
+ "loss": 4.863317489624023,
+ "step": 3834
+ },
+ {
+ "epoch": 53.26550218340611,
+ "grad_norm": 0.04588694870471954,
+ "learning_rate": 0.0006,
+ "loss": 4.864742755889893,
+ "step": 3835
+ },
+ {
+ "epoch": 53.27947598253275,
+ "grad_norm": 0.02728264033794403,
+ "learning_rate": 0.0006,
+ "loss": 4.9068756103515625,
+ "step": 3836
+ },
+ {
+ "epoch": 53.29344978165939,
+ "grad_norm": 0.036240726709365845,
+ "learning_rate": 0.0006,
+ "loss": 4.902045249938965,
+ "step": 3837
+ },
+ {
+ "epoch": 53.30742358078603,
+ "grad_norm": 0.02703195810317993,
+ "learning_rate": 0.0006,
+ "loss": 4.908980369567871,
+ "step": 3838
+ },
+ {
+ "epoch": 53.32139737991266,
+ "grad_norm": 0.024894049391150475,
+ "learning_rate": 0.0006,
+ "loss": 4.794848442077637,
+ "step": 3839
+ },
+ {
+ "epoch": 53.3353711790393,
+ "grad_norm": 0.02389993891119957,
+ "learning_rate": 0.0006,
+ "loss": 4.8234734535217285,
+ "step": 3840
+ },
+ {
+ "epoch": 53.34934497816594,
+ "grad_norm": 0.02486400306224823,
+ "learning_rate": 0.0006,
+ "loss": 4.851463794708252,
+ "step": 3841
+ },
+ {
+ "epoch": 53.36331877729258,
+ "grad_norm": 0.02613813802599907,
+ "learning_rate": 0.0006,
+ "loss": 4.860444068908691,
+ "step": 3842
+ },
+ {
+ "epoch": 53.377292576419215,
+ "grad_norm": 0.0229744054377079,
+ "learning_rate": 0.0006,
+ "loss": 4.709606170654297,
+ "step": 3843
+ },
+ {
+ "epoch": 53.391266375545854,
+ "grad_norm": 0.02123042568564415,
+ "learning_rate": 0.0006,
+ "loss": 4.654613494873047,
+ "step": 3844
+ },
+ {
+ "epoch": 53.40524017467249,
+ "grad_norm": 0.017352338880300522,
+ "learning_rate": 0.0006,
+ "loss": 4.769761085510254,
+ "step": 3845
+ },
+ {
+ "epoch": 53.419213973799124,
+ "grad_norm": 0.019808724522590637,
+ "learning_rate": 0.0006,
+ "loss": 4.98514461517334,
+ "step": 3846
+ },
+ {
+ "epoch": 53.43318777292576,
+ "grad_norm": 0.018776152282953262,
+ "learning_rate": 0.0006,
+ "loss": 4.749990463256836,
+ "step": 3847
+ },
+ {
+ "epoch": 53.4471615720524,
+ "grad_norm": 0.02017521858215332,
+ "learning_rate": 0.0006,
+ "loss": 4.757381439208984,
+ "step": 3848
+ },
+ {
+ "epoch": 53.46113537117904,
+ "grad_norm": 0.019509416073560715,
+ "learning_rate": 0.0006,
+ "loss": 4.8111572265625,
+ "step": 3849
+ },
+ {
+ "epoch": 53.47510917030568,
+ "grad_norm": 0.015543580986559391,
+ "learning_rate": 0.0006,
+ "loss": 4.7501678466796875,
+ "step": 3850
+ },
+ {
+ "epoch": 53.48908296943232,
+ "grad_norm": 0.017941651865839958,
+ "learning_rate": 0.0006,
+ "loss": 4.701838493347168,
+ "step": 3851
+ },
+ {
+ "epoch": 53.50305676855895,
+ "grad_norm": 0.01678212359547615,
+ "learning_rate": 0.0006,
+ "loss": 4.787798881530762,
+ "step": 3852
+ },
+ {
+ "epoch": 53.51703056768559,
+ "grad_norm": 0.019050654023885727,
+ "learning_rate": 0.0006,
+ "loss": 4.754859924316406,
+ "step": 3853
+ },
+ {
+ "epoch": 53.531004366812226,
+ "grad_norm": 0.017406078055500984,
+ "learning_rate": 0.0006,
+ "loss": 4.752772808074951,
+ "step": 3854
+ },
+ {
+ "epoch": 53.544978165938865,
+ "grad_norm": 0.01773344725370407,
+ "learning_rate": 0.0006,
+ "loss": 4.651812553405762,
+ "step": 3855
+ },
+ {
+ "epoch": 53.5589519650655,
+ "grad_norm": 0.017221346497535706,
+ "learning_rate": 0.0006,
+ "loss": 4.792244911193848,
+ "step": 3856
+ },
+ {
+ "epoch": 53.57292576419214,
+ "grad_norm": 0.01722758449614048,
+ "learning_rate": 0.0006,
+ "loss": 4.7902960777282715,
+ "step": 3857
+ },
+ {
+ "epoch": 53.58689956331878,
+ "grad_norm": 0.020445192232728004,
+ "learning_rate": 0.0006,
+ "loss": 4.67392110824585,
+ "step": 3858
+ },
+ {
+ "epoch": 53.60087336244541,
+ "grad_norm": 0.02267114631831646,
+ "learning_rate": 0.0006,
+ "loss": 4.796089172363281,
+ "step": 3859
+ },
+ {
+ "epoch": 53.61484716157205,
+ "grad_norm": 0.027129942551255226,
+ "learning_rate": 0.0006,
+ "loss": 4.774303913116455,
+ "step": 3860
+ },
+ {
+ "epoch": 53.62882096069869,
+ "grad_norm": 0.025454839691519737,
+ "learning_rate": 0.0006,
+ "loss": 4.708813190460205,
+ "step": 3861
+ },
+ {
+ "epoch": 53.64279475982533,
+ "grad_norm": 0.019248757511377335,
+ "learning_rate": 0.0006,
+ "loss": 4.760467529296875,
+ "step": 3862
+ },
+ {
+ "epoch": 53.65676855895197,
+ "grad_norm": 0.01791176199913025,
+ "learning_rate": 0.0006,
+ "loss": 4.789203643798828,
+ "step": 3863
+ },
+ {
+ "epoch": 53.670742358078606,
+ "grad_norm": 0.016306612640619278,
+ "learning_rate": 0.0006,
+ "loss": 4.732605934143066,
+ "step": 3864
+ },
+ {
+ "epoch": 53.68471615720524,
+ "grad_norm": 0.017836976796388626,
+ "learning_rate": 0.0006,
+ "loss": 4.691871643066406,
+ "step": 3865
+ },
+ {
+ "epoch": 53.698689956331876,
+ "grad_norm": 0.015102861449122429,
+ "learning_rate": 0.0006,
+ "loss": 4.724156856536865,
+ "step": 3866
+ },
+ {
+ "epoch": 53.712663755458514,
+ "grad_norm": 0.017458142712712288,
+ "learning_rate": 0.0006,
+ "loss": 4.73524808883667,
+ "step": 3867
+ },
+ {
+ "epoch": 53.72663755458515,
+ "grad_norm": 0.01591932773590088,
+ "learning_rate": 0.0006,
+ "loss": 4.705888748168945,
+ "step": 3868
+ },
+ {
+ "epoch": 53.74061135371179,
+ "grad_norm": 0.01713593676686287,
+ "learning_rate": 0.0006,
+ "loss": 4.644516944885254,
+ "step": 3869
+ },
+ {
+ "epoch": 53.75458515283843,
+ "grad_norm": 0.01905885338783264,
+ "learning_rate": 0.0006,
+ "loss": 4.698416233062744,
+ "step": 3870
+ },
+ {
+ "epoch": 53.76855895196506,
+ "grad_norm": 0.01766437292098999,
+ "learning_rate": 0.0006,
+ "loss": 4.6796555519104,
+ "step": 3871
+ },
+ {
+ "epoch": 53.7825327510917,
+ "grad_norm": 0.01552243810147047,
+ "learning_rate": 0.0006,
+ "loss": 4.78138542175293,
+ "step": 3872
+ },
+ {
+ "epoch": 53.79650655021834,
+ "grad_norm": 0.014191331341862679,
+ "learning_rate": 0.0006,
+ "loss": 4.703588485717773,
+ "step": 3873
+ },
+ {
+ "epoch": 53.81048034934498,
+ "grad_norm": 0.015872279182076454,
+ "learning_rate": 0.0006,
+ "loss": 4.705611228942871,
+ "step": 3874
+ },
+ {
+ "epoch": 53.82445414847162,
+ "grad_norm": 0.01787560060620308,
+ "learning_rate": 0.0006,
+ "loss": 4.8227949142456055,
+ "step": 3875
+ },
+ {
+ "epoch": 53.838427947598255,
+ "grad_norm": 0.0158209390938282,
+ "learning_rate": 0.0006,
+ "loss": 4.645227432250977,
+ "step": 3876
+ },
+ {
+ "epoch": 53.852401746724894,
+ "grad_norm": 0.014362678863108158,
+ "learning_rate": 0.0006,
+ "loss": 4.7225422859191895,
+ "step": 3877
+ },
+ {
+ "epoch": 53.866375545851525,
+ "grad_norm": 0.014899161644279957,
+ "learning_rate": 0.0006,
+ "loss": 4.685914039611816,
+ "step": 3878
+ },
+ {
+ "epoch": 53.880349344978164,
+ "grad_norm": 0.01628505438566208,
+ "learning_rate": 0.0006,
+ "loss": 4.674065589904785,
+ "step": 3879
+ },
+ {
+ "epoch": 53.8943231441048,
+ "grad_norm": 0.01873331144452095,
+ "learning_rate": 0.0006,
+ "loss": 4.811744213104248,
+ "step": 3880
+ },
+ {
+ "epoch": 53.90829694323144,
+ "grad_norm": 0.02229069173336029,
+ "learning_rate": 0.0006,
+ "loss": 4.721514701843262,
+ "step": 3881
+ },
+ {
+ "epoch": 53.92227074235808,
+ "grad_norm": 0.028012916445732117,
+ "learning_rate": 0.0006,
+ "loss": 4.773603439331055,
+ "step": 3882
+ },
+ {
+ "epoch": 53.93624454148472,
+ "grad_norm": 0.036322783678770065,
+ "learning_rate": 0.0006,
+ "loss": 4.787134170532227,
+ "step": 3883
+ },
+ {
+ "epoch": 53.95021834061135,
+ "grad_norm": 0.048149142414331436,
+ "learning_rate": 0.0006,
+ "loss": 4.736196517944336,
+ "step": 3884
+ },
+ {
+ "epoch": 53.96419213973799,
+ "grad_norm": 0.051918916404247284,
+ "learning_rate": 0.0006,
+ "loss": 4.686728000640869,
+ "step": 3885
+ },
+ {
+ "epoch": 53.97816593886463,
+ "grad_norm": 0.0481342189013958,
+ "learning_rate": 0.0006,
+ "loss": 4.733730792999268,
+ "step": 3886
+ },
+ {
+ "epoch": 53.992139737991266,
+ "grad_norm": 0.045424818992614746,
+ "learning_rate": 0.0006,
+ "loss": 4.748723983764648,
+ "step": 3887
+ },
+ {
+ "epoch": 54.0,
+ "grad_norm": 0.04153316840529442,
+ "learning_rate": 0.0006,
+ "loss": 4.715220928192139,
+ "step": 3888
+ },
+ {
+ "epoch": 54.0,
+ "eval_loss": 4.964215278625488,
+ "eval_runtime": 56.7757,
+ "eval_samples_per_second": 43.011,
+ "eval_steps_per_second": 1.356,
+ "step": 3888
+ },
+ {
+ "epoch": 54.01397379912664,
+ "grad_norm": 0.03459750860929489,
+ "learning_rate": 0.0006,
+ "loss": 4.661557674407959,
+ "step": 3889
+ },
+ {
+ "epoch": 54.02794759825328,
+ "grad_norm": 0.025409208610653877,
+ "learning_rate": 0.0006,
+ "loss": 4.7065582275390625,
+ "step": 3890
+ },
+ {
+ "epoch": 54.041921397379916,
+ "grad_norm": 0.03202550485730171,
+ "learning_rate": 0.0006,
+ "loss": 4.751070976257324,
+ "step": 3891
+ },
+ {
+ "epoch": 54.05589519650655,
+ "grad_norm": 0.02968018501996994,
+ "learning_rate": 0.0006,
+ "loss": 4.635388374328613,
+ "step": 3892
+ },
+ {
+ "epoch": 54.069868995633186,
+ "grad_norm": 0.02757406234741211,
+ "learning_rate": 0.0006,
+ "loss": 4.706214427947998,
+ "step": 3893
+ },
+ {
+ "epoch": 54.083842794759825,
+ "grad_norm": 0.025223620235919952,
+ "learning_rate": 0.0006,
+ "loss": 4.703792572021484,
+ "step": 3894
+ },
+ {
+ "epoch": 54.09781659388646,
+ "grad_norm": 0.02431778982281685,
+ "learning_rate": 0.0006,
+ "loss": 4.658116340637207,
+ "step": 3895
+ },
+ {
+ "epoch": 54.1117903930131,
+ "grad_norm": 0.02792172320187092,
+ "learning_rate": 0.0006,
+ "loss": 4.638895034790039,
+ "step": 3896
+ },
+ {
+ "epoch": 54.12576419213974,
+ "grad_norm": 0.032253511250019073,
+ "learning_rate": 0.0006,
+ "loss": 4.693902492523193,
+ "step": 3897
+ },
+ {
+ "epoch": 54.13973799126637,
+ "grad_norm": 0.027044931426644325,
+ "learning_rate": 0.0006,
+ "loss": 4.575805187225342,
+ "step": 3898
+ },
+ {
+ "epoch": 54.15371179039301,
+ "grad_norm": 0.023270361125469208,
+ "learning_rate": 0.0006,
+ "loss": 4.67645263671875,
+ "step": 3899
+ },
+ {
+ "epoch": 54.16768558951965,
+ "grad_norm": 0.01929079182446003,
+ "learning_rate": 0.0006,
+ "loss": 4.68062686920166,
+ "step": 3900
+ },
+ {
+ "epoch": 54.18165938864629,
+ "grad_norm": 0.020343473181128502,
+ "learning_rate": 0.0006,
+ "loss": 4.708086967468262,
+ "step": 3901
+ },
+ {
+ "epoch": 54.19563318777293,
+ "grad_norm": 0.0211793202906847,
+ "learning_rate": 0.0006,
+ "loss": 4.677323341369629,
+ "step": 3902
+ },
+ {
+ "epoch": 54.209606986899566,
+ "grad_norm": 0.020341217517852783,
+ "learning_rate": 0.0006,
+ "loss": 4.760937690734863,
+ "step": 3903
+ },
+ {
+ "epoch": 54.223580786026204,
+ "grad_norm": 0.019421054050326347,
+ "learning_rate": 0.0006,
+ "loss": 4.508834362030029,
+ "step": 3904
+ },
+ {
+ "epoch": 54.237554585152836,
+ "grad_norm": 0.016701679676771164,
+ "learning_rate": 0.0006,
+ "loss": 4.664399147033691,
+ "step": 3905
+ },
+ {
+ "epoch": 54.251528384279474,
+ "grad_norm": 0.016561470925807953,
+ "learning_rate": 0.0006,
+ "loss": 4.742990493774414,
+ "step": 3906
+ },
+ {
+ "epoch": 54.26550218340611,
+ "grad_norm": 0.017446085810661316,
+ "learning_rate": 0.0006,
+ "loss": 4.63038444519043,
+ "step": 3907
+ },
+ {
+ "epoch": 54.27947598253275,
+ "grad_norm": 0.014340086840093136,
+ "learning_rate": 0.0006,
+ "loss": 4.688791275024414,
+ "step": 3908
+ },
+ {
+ "epoch": 54.29344978165939,
+ "grad_norm": 0.013580698519945145,
+ "learning_rate": 0.0006,
+ "loss": 4.625302314758301,
+ "step": 3909
+ },
+ {
+ "epoch": 54.30742358078603,
+ "grad_norm": 0.014931539073586464,
+ "learning_rate": 0.0006,
+ "loss": 4.6167497634887695,
+ "step": 3910
+ },
+ {
+ "epoch": 54.32139737991266,
+ "grad_norm": 0.013492150232195854,
+ "learning_rate": 0.0006,
+ "loss": 4.657521724700928,
+ "step": 3911
+ },
+ {
+ "epoch": 54.3353711790393,
+ "grad_norm": 0.015151413157582283,
+ "learning_rate": 0.0006,
+ "loss": 4.634521484375,
+ "step": 3912
+ },
+ {
+ "epoch": 54.34934497816594,
+ "grad_norm": 0.01302040833979845,
+ "learning_rate": 0.0006,
+ "loss": 4.718928337097168,
+ "step": 3913
+ },
+ {
+ "epoch": 54.36331877729258,
+ "grad_norm": 0.013433529995381832,
+ "learning_rate": 0.0006,
+ "loss": 4.584803581237793,
+ "step": 3914
+ },
+ {
+ "epoch": 54.377292576419215,
+ "grad_norm": 0.012987039051949978,
+ "learning_rate": 0.0006,
+ "loss": 4.6393585205078125,
+ "step": 3915
+ },
+ {
+ "epoch": 54.391266375545854,
+ "grad_norm": 0.013695075176656246,
+ "learning_rate": 0.0006,
+ "loss": 4.556313514709473,
+ "step": 3916
+ },
+ {
+ "epoch": 54.40524017467249,
+ "grad_norm": 0.013718255795538425,
+ "learning_rate": 0.0006,
+ "loss": 4.576223850250244,
+ "step": 3917
+ },
+ {
+ "epoch": 54.419213973799124,
+ "grad_norm": 0.013699112460017204,
+ "learning_rate": 0.0006,
+ "loss": 4.63889217376709,
+ "step": 3918
+ },
+ {
+ "epoch": 54.43318777292576,
+ "grad_norm": 0.013516264036297798,
+ "learning_rate": 0.0006,
+ "loss": 4.669532299041748,
+ "step": 3919
+ },
+ {
+ "epoch": 54.4471615720524,
+ "grad_norm": 0.013597294688224792,
+ "learning_rate": 0.0006,
+ "loss": 4.6644287109375,
+ "step": 3920
+ },
+ {
+ "epoch": 54.46113537117904,
+ "grad_norm": 0.013185849413275719,
+ "learning_rate": 0.0006,
+ "loss": 4.636900901794434,
+ "step": 3921
+ },
+ {
+ "epoch": 54.47510917030568,
+ "grad_norm": 0.014311990700662136,
+ "learning_rate": 0.0006,
+ "loss": 4.6047587394714355,
+ "step": 3922
+ },
+ {
+ "epoch": 54.48908296943232,
+ "grad_norm": 0.012426027096807957,
+ "learning_rate": 0.0006,
+ "loss": 4.6161394119262695,
+ "step": 3923
+ },
+ {
+ "epoch": 54.50305676855895,
+ "grad_norm": 0.01283493172377348,
+ "learning_rate": 0.0006,
+ "loss": 4.584190368652344,
+ "step": 3924
+ },
+ {
+ "epoch": 54.51703056768559,
+ "grad_norm": 0.01339943427592516,
+ "learning_rate": 0.0006,
+ "loss": 4.624184608459473,
+ "step": 3925
+ },
+ {
+ "epoch": 54.531004366812226,
+ "grad_norm": 0.012859724462032318,
+ "learning_rate": 0.0006,
+ "loss": 4.6174635887146,
+ "step": 3926
+ },
+ {
+ "epoch": 54.544978165938865,
+ "grad_norm": 0.014420023187994957,
+ "learning_rate": 0.0006,
+ "loss": 4.536544322967529,
+ "step": 3927
+ },
+ {
+ "epoch": 54.5589519650655,
+ "grad_norm": 0.014237185008823872,
+ "learning_rate": 0.0006,
+ "loss": 4.681514739990234,
+ "step": 3928
+ },
+ {
+ "epoch": 54.57292576419214,
+ "grad_norm": 0.013340328820049763,
+ "learning_rate": 0.0006,
+ "loss": 4.579689979553223,
+ "step": 3929
+ },
+ {
+ "epoch": 54.58689956331878,
+ "grad_norm": 0.013894530944526196,
+ "learning_rate": 0.0006,
+ "loss": 4.638958930969238,
+ "step": 3930
+ },
+ {
+ "epoch": 54.60087336244541,
+ "grad_norm": 0.016181141138076782,
+ "learning_rate": 0.0006,
+ "loss": 4.573508262634277,
+ "step": 3931
+ },
+ {
+ "epoch": 54.61484716157205,
+ "grad_norm": 0.013221992179751396,
+ "learning_rate": 0.0006,
+ "loss": 4.6229047775268555,
+ "step": 3932
+ },
+ {
+ "epoch": 54.62882096069869,
+ "grad_norm": 0.013241561129689217,
+ "learning_rate": 0.0006,
+ "loss": 4.641152858734131,
+ "step": 3933
+ },
+ {
+ "epoch": 54.64279475982533,
+ "grad_norm": 0.016237538307905197,
+ "learning_rate": 0.0006,
+ "loss": 4.6421217918396,
+ "step": 3934
+ },
+ {
+ "epoch": 54.65676855895197,
+ "grad_norm": 0.020866507664322853,
+ "learning_rate": 0.0006,
+ "loss": 4.646058082580566,
+ "step": 3935
+ },
+ {
+ "epoch": 54.670742358078606,
+ "grad_norm": 0.022182893007993698,
+ "learning_rate": 0.0006,
+ "loss": 4.554501533508301,
+ "step": 3936
+ },
+ {
+ "epoch": 54.68471615720524,
+ "grad_norm": 0.02254868485033512,
+ "learning_rate": 0.0006,
+ "loss": 4.563869953155518,
+ "step": 3937
+ },
+ {
+ "epoch": 54.698689956331876,
+ "grad_norm": 0.019730940461158752,
+ "learning_rate": 0.0006,
+ "loss": 4.592165946960449,
+ "step": 3938
+ },
+ {
+ "epoch": 54.712663755458514,
+ "grad_norm": 0.015146561898291111,
+ "learning_rate": 0.0006,
+ "loss": 4.6228742599487305,
+ "step": 3939
+ },
+ {
+ "epoch": 54.72663755458515,
+ "grad_norm": 0.01824716106057167,
+ "learning_rate": 0.0006,
+ "loss": 4.716236114501953,
+ "step": 3940
+ },
+ {
+ "epoch": 54.74061135371179,
+ "grad_norm": 0.017730550840497017,
+ "learning_rate": 0.0006,
+ "loss": 4.650232315063477,
+ "step": 3941
+ },
+ {
+ "epoch": 54.75458515283843,
+ "grad_norm": 0.014875595457851887,
+ "learning_rate": 0.0006,
+ "loss": 4.6377458572387695,
+ "step": 3942
+ },
+ {
+ "epoch": 54.76855895196506,
+ "grad_norm": 0.01613209955394268,
+ "learning_rate": 0.0006,
+ "loss": 4.5768537521362305,
+ "step": 3943
+ },
+ {
+ "epoch": 54.7825327510917,
+ "grad_norm": 0.017311280593276024,
+ "learning_rate": 0.0006,
+ "loss": 4.625744342803955,
+ "step": 3944
+ },
+ {
+ "epoch": 54.79650655021834,
+ "grad_norm": 0.018770085647702217,
+ "learning_rate": 0.0006,
+ "loss": 4.613787651062012,
+ "step": 3945
+ },
+ {
+ "epoch": 54.81048034934498,
+ "grad_norm": 0.0165257528424263,
+ "learning_rate": 0.0006,
+ "loss": 4.5991315841674805,
+ "step": 3946
+ },
+ {
+ "epoch": 54.82445414847162,
+ "grad_norm": 0.017433857545256615,
+ "learning_rate": 0.0006,
+ "loss": 4.513872146606445,
+ "step": 3947
+ },
+ {
+ "epoch": 54.838427947598255,
+ "grad_norm": 0.020126905292272568,
+ "learning_rate": 0.0006,
+ "loss": 4.616185188293457,
+ "step": 3948
+ },
+ {
+ "epoch": 54.852401746724894,
+ "grad_norm": 0.01819358579814434,
+ "learning_rate": 0.0006,
+ "loss": 4.535691261291504,
+ "step": 3949
+ },
+ {
+ "epoch": 54.866375545851525,
+ "grad_norm": 0.01899927295744419,
+ "learning_rate": 0.0006,
+ "loss": 4.595199108123779,
+ "step": 3950
+ },
+ {
+ "epoch": 54.880349344978164,
+ "grad_norm": 0.02005566656589508,
+ "learning_rate": 0.0006,
+ "loss": 4.633289337158203,
+ "step": 3951
+ },
+ {
+ "epoch": 54.8943231441048,
+ "grad_norm": 0.021570220589637756,
+ "learning_rate": 0.0006,
+ "loss": 4.592503547668457,
+ "step": 3952
+ },
+ {
+ "epoch": 54.90829694323144,
+ "grad_norm": 0.01883506216108799,
+ "learning_rate": 0.0006,
+ "loss": 4.6597795486450195,
+ "step": 3953
+ },
+ {
+ "epoch": 54.92227074235808,
+ "grad_norm": 0.018070021644234657,
+ "learning_rate": 0.0006,
+ "loss": 4.601474285125732,
+ "step": 3954
+ },
+ {
+ "epoch": 54.93624454148472,
+ "grad_norm": 0.018090378493070602,
+ "learning_rate": 0.0006,
+ "loss": 4.569846153259277,
+ "step": 3955
+ },
+ {
+ "epoch": 54.95021834061135,
+ "grad_norm": 0.018751563504338264,
+ "learning_rate": 0.0006,
+ "loss": 4.646064281463623,
+ "step": 3956
+ },
+ {
+ "epoch": 54.96419213973799,
+ "grad_norm": 0.017512254416942596,
+ "learning_rate": 0.0006,
+ "loss": 4.606670379638672,
+ "step": 3957
+ },
+ {
+ "epoch": 54.97816593886463,
+ "grad_norm": 0.018771110102534294,
+ "learning_rate": 0.0006,
+ "loss": 4.604315757751465,
+ "step": 3958
+ },
+ {
+ "epoch": 54.992139737991266,
+ "grad_norm": 0.01611141301691532,
+ "learning_rate": 0.0006,
+ "loss": 4.643245220184326,
+ "step": 3959
+ },
+ {
+ "epoch": 55.0,
+ "grad_norm": 0.0156728383153677,
+ "learning_rate": 0.0006,
+ "loss": 4.6739044189453125,
+ "step": 3960
+ },
+ {
+ "epoch": 55.0,
+ "eval_loss": 4.851032257080078,
+ "eval_runtime": 56.8699,
+ "eval_samples_per_second": 42.94,
+ "eval_steps_per_second": 1.354,
+ "step": 3960
+ },
+ {
+ "epoch": 55.01397379912664,
+ "grad_norm": 0.016214672476053238,
+ "learning_rate": 0.0006,
+ "loss": 4.571749687194824,
+ "step": 3961
+ },
+ {
+ "epoch": 55.02794759825328,
+ "grad_norm": 0.01649155467748642,
+ "learning_rate": 0.0006,
+ "loss": 4.553592205047607,
+ "step": 3962
+ },
+ {
+ "epoch": 55.041921397379916,
+ "grad_norm": 0.016793379560112953,
+ "learning_rate": 0.0006,
+ "loss": 4.650628566741943,
+ "step": 3963
+ },
+ {
+ "epoch": 55.05589519650655,
+ "grad_norm": 0.0191937405616045,
+ "learning_rate": 0.0006,
+ "loss": 4.620532989501953,
+ "step": 3964
+ },
+ {
+ "epoch": 55.069868995633186,
+ "grad_norm": 0.020976556465029716,
+ "learning_rate": 0.0006,
+ "loss": 4.6329545974731445,
+ "step": 3965
+ },
+ {
+ "epoch": 55.083842794759825,
+ "grad_norm": 0.023450477048754692,
+ "learning_rate": 0.0006,
+ "loss": 4.60828971862793,
+ "step": 3966
+ },
+ {
+ "epoch": 55.09781659388646,
+ "grad_norm": 0.026778312399983406,
+ "learning_rate": 0.0006,
+ "loss": 4.617481708526611,
+ "step": 3967
+ },
+ {
+ "epoch": 55.1117903930131,
+ "grad_norm": 0.02333568036556244,
+ "learning_rate": 0.0006,
+ "loss": 4.580722332000732,
+ "step": 3968
+ },
+ {
+ "epoch": 55.12576419213974,
+ "grad_norm": 0.018491310998797417,
+ "learning_rate": 0.0006,
+ "loss": 4.590953350067139,
+ "step": 3969
+ },
+ {
+ "epoch": 55.13973799126637,
+ "grad_norm": 0.020339803770184517,
+ "learning_rate": 0.0006,
+ "loss": 4.518374443054199,
+ "step": 3970
+ },
+ {
+ "epoch": 55.15371179039301,
+ "grad_norm": 0.01759706623852253,
+ "learning_rate": 0.0006,
+ "loss": 4.560522079467773,
+ "step": 3971
+ },
+ {
+ "epoch": 55.16768558951965,
+ "grad_norm": 0.015875980257987976,
+ "learning_rate": 0.0006,
+ "loss": 4.4795708656311035,
+ "step": 3972
+ },
+ {
+ "epoch": 55.18165938864629,
+ "grad_norm": 0.01650134287774563,
+ "learning_rate": 0.0006,
+ "loss": 4.57435941696167,
+ "step": 3973
+ },
+ {
+ "epoch": 55.19563318777293,
+ "grad_norm": 0.01734331250190735,
+ "learning_rate": 0.0006,
+ "loss": 4.661007881164551,
+ "step": 3974
+ },
+ {
+ "epoch": 55.209606986899566,
+ "grad_norm": 0.01651303470134735,
+ "learning_rate": 0.0006,
+ "loss": 4.511148452758789,
+ "step": 3975
+ },
+ {
+ "epoch": 55.223580786026204,
+ "grad_norm": 0.016189761459827423,
+ "learning_rate": 0.0006,
+ "loss": 4.592350006103516,
+ "step": 3976
+ },
+ {
+ "epoch": 55.237554585152836,
+ "grad_norm": 0.01833222061395645,
+ "learning_rate": 0.0006,
+ "loss": 4.506649971008301,
+ "step": 3977
+ },
+ {
+ "epoch": 55.251528384279474,
+ "grad_norm": 0.02433410845696926,
+ "learning_rate": 0.0006,
+ "loss": 4.49500036239624,
+ "step": 3978
+ },
+ {
+ "epoch": 55.26550218340611,
+ "grad_norm": 0.0314289927482605,
+ "learning_rate": 0.0006,
+ "loss": 4.507624626159668,
+ "step": 3979
+ },
+ {
+ "epoch": 55.27947598253275,
+ "grad_norm": 0.03013679012656212,
+ "learning_rate": 0.0006,
+ "loss": 4.533641815185547,
+ "step": 3980
+ },
+ {
+ "epoch": 55.29344978165939,
+ "grad_norm": 0.024524061009287834,
+ "learning_rate": 0.0006,
+ "loss": 4.5721116065979,
+ "step": 3981
+ },
+ {
+ "epoch": 55.30742358078603,
+ "grad_norm": 0.02640336938202381,
+ "learning_rate": 0.0006,
+ "loss": 4.652521133422852,
+ "step": 3982
+ },
+ {
+ "epoch": 55.32139737991266,
+ "grad_norm": 0.021309101954102516,
+ "learning_rate": 0.0006,
+ "loss": 4.600375175476074,
+ "step": 3983
+ },
+ {
+ "epoch": 55.3353711790393,
+ "grad_norm": 0.023008549585938454,
+ "learning_rate": 0.0006,
+ "loss": 4.53914737701416,
+ "step": 3984
+ },
+ {
+ "epoch": 55.34934497816594,
+ "grad_norm": 0.022491468116641045,
+ "learning_rate": 0.0006,
+ "loss": 4.661163330078125,
+ "step": 3985
+ },
+ {
+ "epoch": 55.36331877729258,
+ "grad_norm": 0.01904081553220749,
+ "learning_rate": 0.0006,
+ "loss": 4.5426177978515625,
+ "step": 3986
+ },
+ {
+ "epoch": 55.377292576419215,
+ "grad_norm": 0.019229518249630928,
+ "learning_rate": 0.0006,
+ "loss": 4.606066703796387,
+ "step": 3987
+ },
+ {
+ "epoch": 55.391266375545854,
+ "grad_norm": 0.01781383901834488,
+ "learning_rate": 0.0006,
+ "loss": 4.54642391204834,
+ "step": 3988
+ },
+ {
+ "epoch": 55.40524017467249,
+ "grad_norm": 0.0174600500613451,
+ "learning_rate": 0.0006,
+ "loss": 4.711891174316406,
+ "step": 3989
+ },
+ {
+ "epoch": 55.419213973799124,
+ "grad_norm": 0.018154339864850044,
+ "learning_rate": 0.0006,
+ "loss": 4.527122974395752,
+ "step": 3990
+ },
+ {
+ "epoch": 55.43318777292576,
+ "grad_norm": 0.01914648339152336,
+ "learning_rate": 0.0006,
+ "loss": 4.641975402832031,
+ "step": 3991
+ },
+ {
+ "epoch": 55.4471615720524,
+ "grad_norm": 0.017118625342845917,
+ "learning_rate": 0.0006,
+ "loss": 4.511425971984863,
+ "step": 3992
+ },
+ {
+ "epoch": 55.46113537117904,
+ "grad_norm": 0.01551489345729351,
+ "learning_rate": 0.0006,
+ "loss": 4.551141738891602,
+ "step": 3993
+ },
+ {
+ "epoch": 55.47510917030568,
+ "grad_norm": 0.015072065405547619,
+ "learning_rate": 0.0006,
+ "loss": 4.651115894317627,
+ "step": 3994
+ },
+ {
+ "epoch": 55.48908296943232,
+ "grad_norm": 0.017010986804962158,
+ "learning_rate": 0.0006,
+ "loss": 4.583282470703125,
+ "step": 3995
+ },
+ {
+ "epoch": 55.50305676855895,
+ "grad_norm": 0.01764831133186817,
+ "learning_rate": 0.0006,
+ "loss": 4.593364238739014,
+ "step": 3996
+ },
+ {
+ "epoch": 55.51703056768559,
+ "grad_norm": 0.01762884296476841,
+ "learning_rate": 0.0006,
+ "loss": 4.45671272277832,
+ "step": 3997
+ },
+ {
+ "epoch": 55.531004366812226,
+ "grad_norm": 0.01600457727909088,
+ "learning_rate": 0.0006,
+ "loss": 4.511260032653809,
+ "step": 3998
+ },
+ {
+ "epoch": 55.544978165938865,
+ "grad_norm": 0.015706021338701248,
+ "learning_rate": 0.0006,
+ "loss": 4.47593355178833,
+ "step": 3999
+ },
+ {
+ "epoch": 55.5589519650655,
+ "grad_norm": 0.017718996852636337,
+ "learning_rate": 0.0006,
+ "loss": 4.610543251037598,
+ "step": 4000
+ },
+ {
+ "epoch": 55.57292576419214,
+ "grad_norm": 0.015073757618665695,
+ "learning_rate": 0.0006,
+ "loss": 4.466620445251465,
+ "step": 4001
+ },
+ {
+ "epoch": 55.58689956331878,
+ "grad_norm": 0.01789247989654541,
+ "learning_rate": 0.0006,
+ "loss": 4.503409385681152,
+ "step": 4002
+ },
+ {
+ "epoch": 55.60087336244541,
+ "grad_norm": 0.022329598665237427,
+ "learning_rate": 0.0006,
+ "loss": 4.62099552154541,
+ "step": 4003
+ },
+ {
+ "epoch": 55.61484716157205,
+ "grad_norm": 0.02343829534947872,
+ "learning_rate": 0.0006,
+ "loss": 4.620813369750977,
+ "step": 4004
+ },
+ {
+ "epoch": 55.62882096069869,
+ "grad_norm": 0.023279687389731407,
+ "learning_rate": 0.0006,
+ "loss": 4.545765399932861,
+ "step": 4005
+ },
+ {
+ "epoch": 55.64279475982533,
+ "grad_norm": 0.020817464217543602,
+ "learning_rate": 0.0006,
+ "loss": 4.561470031738281,
+ "step": 4006
+ },
+ {
+ "epoch": 55.65676855895197,
+ "grad_norm": 0.020270058885216713,
+ "learning_rate": 0.0006,
+ "loss": 4.5875020027160645,
+ "step": 4007
+ },
+ {
+ "epoch": 55.670742358078606,
+ "grad_norm": 0.021085843443870544,
+ "learning_rate": 0.0006,
+ "loss": 4.504560470581055,
+ "step": 4008
+ },
+ {
+ "epoch": 55.68471615720524,
+ "grad_norm": 0.018507730215787888,
+ "learning_rate": 0.0006,
+ "loss": 4.520644664764404,
+ "step": 4009
+ },
+ {
+ "epoch": 55.698689956331876,
+ "grad_norm": 0.020571516826748848,
+ "learning_rate": 0.0006,
+ "loss": 4.539643287658691,
+ "step": 4010
+ },
+ {
+ "epoch": 55.712663755458514,
+ "grad_norm": 0.021484510973095894,
+ "learning_rate": 0.0006,
+ "loss": 4.628833293914795,
+ "step": 4011
+ },
+ {
+ "epoch": 55.72663755458515,
+ "grad_norm": 0.017693819478154182,
+ "learning_rate": 0.0006,
+ "loss": 4.6023783683776855,
+ "step": 4012
+ },
+ {
+ "epoch": 55.74061135371179,
+ "grad_norm": 0.020186059176921844,
+ "learning_rate": 0.0006,
+ "loss": 4.634884834289551,
+ "step": 4013
+ },
+ {
+ "epoch": 55.75458515283843,
+ "grad_norm": 0.01835780404508114,
+ "learning_rate": 0.0006,
+ "loss": 4.577791213989258,
+ "step": 4014
+ },
+ {
+ "epoch": 55.76855895196506,
+ "grad_norm": 0.018162988126277924,
+ "learning_rate": 0.0006,
+ "loss": 4.641597270965576,
+ "step": 4015
+ },
+ {
+ "epoch": 55.7825327510917,
+ "grad_norm": 0.018909504637122154,
+ "learning_rate": 0.0006,
+ "loss": 4.487447738647461,
+ "step": 4016
+ },
+ {
+ "epoch": 55.79650655021834,
+ "grad_norm": 0.018972251564264297,
+ "learning_rate": 0.0006,
+ "loss": 4.590317726135254,
+ "step": 4017
+ },
+ {
+ "epoch": 55.81048034934498,
+ "grad_norm": 0.01793607883155346,
+ "learning_rate": 0.0006,
+ "loss": 4.4994611740112305,
+ "step": 4018
+ },
+ {
+ "epoch": 55.82445414847162,
+ "grad_norm": 0.02123514749109745,
+ "learning_rate": 0.0006,
+ "loss": 4.479006767272949,
+ "step": 4019
+ },
+ {
+ "epoch": 55.838427947598255,
+ "grad_norm": 0.021246599033474922,
+ "learning_rate": 0.0006,
+ "loss": 4.4738640785217285,
+ "step": 4020
+ },
+ {
+ "epoch": 55.852401746724894,
+ "grad_norm": 0.020131602883338928,
+ "learning_rate": 0.0006,
+ "loss": 4.600308418273926,
+ "step": 4021
+ },
+ {
+ "epoch": 55.866375545851525,
+ "grad_norm": 0.019159460440278053,
+ "learning_rate": 0.0006,
+ "loss": 4.540908336639404,
+ "step": 4022
+ },
+ {
+ "epoch": 55.880349344978164,
+ "grad_norm": 0.023114705458283424,
+ "learning_rate": 0.0006,
+ "loss": 4.641247749328613,
+ "step": 4023
+ },
+ {
+ "epoch": 55.8943231441048,
+ "grad_norm": 0.02296135015785694,
+ "learning_rate": 0.0006,
+ "loss": 4.444624423980713,
+ "step": 4024
+ },
+ {
+ "epoch": 55.90829694323144,
+ "grad_norm": 0.02204655110836029,
+ "learning_rate": 0.0006,
+ "loss": 4.506349563598633,
+ "step": 4025
+ },
+ {
+ "epoch": 55.92227074235808,
+ "grad_norm": 0.019338225945830345,
+ "learning_rate": 0.0006,
+ "loss": 4.616047382354736,
+ "step": 4026
+ },
+ {
+ "epoch": 55.93624454148472,
+ "grad_norm": 0.020357884466648102,
+ "learning_rate": 0.0006,
+ "loss": 4.48170280456543,
+ "step": 4027
+ },
+ {
+ "epoch": 55.95021834061135,
+ "grad_norm": 0.02426772564649582,
+ "learning_rate": 0.0006,
+ "loss": 4.576877117156982,
+ "step": 4028
+ },
+ {
+ "epoch": 55.96419213973799,
+ "grad_norm": 0.023351816460490227,
+ "learning_rate": 0.0006,
+ "loss": 4.458885192871094,
+ "step": 4029
+ },
+ {
+ "epoch": 55.97816593886463,
+ "grad_norm": 0.019892243668437004,
+ "learning_rate": 0.0006,
+ "loss": 4.520229816436768,
+ "step": 4030
+ },
+ {
+ "epoch": 55.992139737991266,
+ "grad_norm": 0.018331089988350868,
+ "learning_rate": 0.0006,
+ "loss": 4.432201862335205,
+ "step": 4031
+ },
+ {
+ "epoch": 56.0,
+ "grad_norm": 0.021300828084349632,
+ "learning_rate": 0.0006,
+ "loss": 4.617072582244873,
+ "step": 4032
+ },
+ {
+ "epoch": 56.0,
+ "eval_loss": 4.811229705810547,
+ "eval_runtime": 56.4846,
+ "eval_samples_per_second": 43.233,
+ "eval_steps_per_second": 1.363,
+ "step": 4032
+ },
+ {
+ "epoch": 56.01397379912664,
+ "grad_norm": 0.02199401892721653,
+ "learning_rate": 0.0006,
+ "loss": 4.513965606689453,
+ "step": 4033
+ },
+ {
+ "epoch": 56.02794759825328,
+ "grad_norm": 0.022109858691692352,
+ "learning_rate": 0.0006,
+ "loss": 4.557568550109863,
+ "step": 4034
+ },
+ {
+ "epoch": 56.041921397379916,
+ "grad_norm": 0.020103694871068,
+ "learning_rate": 0.0006,
+ "loss": 4.569512844085693,
+ "step": 4035
+ },
+ {
+ "epoch": 56.05589519650655,
+ "grad_norm": 0.01898941583931446,
+ "learning_rate": 0.0006,
+ "loss": 4.623138904571533,
+ "step": 4036
+ },
+ {
+ "epoch": 56.069868995633186,
+ "grad_norm": 0.02227010577917099,
+ "learning_rate": 0.0006,
+ "loss": 4.433316230773926,
+ "step": 4037
+ },
+ {
+ "epoch": 56.083842794759825,
+ "grad_norm": 0.024683045223355293,
+ "learning_rate": 0.0006,
+ "loss": 4.458296775817871,
+ "step": 4038
+ },
+ {
+ "epoch": 56.09781659388646,
+ "grad_norm": 0.02900184690952301,
+ "learning_rate": 0.0006,
+ "loss": 4.531060218811035,
+ "step": 4039
+ },
+ {
+ "epoch": 56.1117903930131,
+ "grad_norm": 0.025557566434144974,
+ "learning_rate": 0.0006,
+ "loss": 4.56685733795166,
+ "step": 4040
+ },
+ {
+ "epoch": 56.12576419213974,
+ "grad_norm": 0.021414656192064285,
+ "learning_rate": 0.0006,
+ "loss": 4.469037055969238,
+ "step": 4041
+ },
+ {
+ "epoch": 56.13973799126637,
+ "grad_norm": 0.023262616246938705,
+ "learning_rate": 0.0006,
+ "loss": 4.493876934051514,
+ "step": 4042
+ },
+ {
+ "epoch": 56.15371179039301,
+ "grad_norm": 0.023011326789855957,
+ "learning_rate": 0.0006,
+ "loss": 4.4572906494140625,
+ "step": 4043
+ },
+ {
+ "epoch": 56.16768558951965,
+ "grad_norm": 0.02177615463733673,
+ "learning_rate": 0.0006,
+ "loss": 4.615389823913574,
+ "step": 4044
+ },
+ {
+ "epoch": 56.18165938864629,
+ "grad_norm": 0.021527713164687157,
+ "learning_rate": 0.0006,
+ "loss": 4.613762855529785,
+ "step": 4045
+ },
+ {
+ "epoch": 56.19563318777293,
+ "grad_norm": 0.01982325315475464,
+ "learning_rate": 0.0006,
+ "loss": 4.53670597076416,
+ "step": 4046
+ },
+ {
+ "epoch": 56.209606986899566,
+ "grad_norm": 0.022484654560685158,
+ "learning_rate": 0.0006,
+ "loss": 4.605565071105957,
+ "step": 4047
+ },
+ {
+ "epoch": 56.223580786026204,
+ "grad_norm": 0.02346140146255493,
+ "learning_rate": 0.0006,
+ "loss": 4.6265950202941895,
+ "step": 4048
+ },
+ {
+ "epoch": 56.237554585152836,
+ "grad_norm": 0.01876714453101158,
+ "learning_rate": 0.0006,
+ "loss": 4.481525421142578,
+ "step": 4049
+ },
+ {
+ "epoch": 56.251528384279474,
+ "grad_norm": 0.01926851086318493,
+ "learning_rate": 0.0006,
+ "loss": 4.630258083343506,
+ "step": 4050
+ },
+ {
+ "epoch": 56.26550218340611,
+ "grad_norm": 0.020858589559793472,
+ "learning_rate": 0.0006,
+ "loss": 4.5265913009643555,
+ "step": 4051
+ },
+ {
+ "epoch": 56.27947598253275,
+ "grad_norm": 0.01889280416071415,
+ "learning_rate": 0.0006,
+ "loss": 4.581275463104248,
+ "step": 4052
+ },
+ {
+ "epoch": 56.29344978165939,
+ "grad_norm": 0.01801002025604248,
+ "learning_rate": 0.0006,
+ "loss": 4.498557090759277,
+ "step": 4053
+ },
+ {
+ "epoch": 56.30742358078603,
+ "grad_norm": 0.02252180315554142,
+ "learning_rate": 0.0006,
+ "loss": 4.565267086029053,
+ "step": 4054
+ },
+ {
+ "epoch": 56.32139737991266,
+ "grad_norm": 0.02306070365011692,
+ "learning_rate": 0.0006,
+ "loss": 4.3591389656066895,
+ "step": 4055
+ },
+ {
+ "epoch": 56.3353711790393,
+ "grad_norm": 0.020292513072490692,
+ "learning_rate": 0.0006,
+ "loss": 4.407517433166504,
+ "step": 4056
+ },
+ {
+ "epoch": 56.34934497816594,
+ "grad_norm": 0.019154123961925507,
+ "learning_rate": 0.0006,
+ "loss": 4.482261657714844,
+ "step": 4057
+ },
+ {
+ "epoch": 56.36331877729258,
+ "grad_norm": 0.017338937148451805,
+ "learning_rate": 0.0006,
+ "loss": 4.4463582038879395,
+ "step": 4058
+ },
+ {
+ "epoch": 56.377292576419215,
+ "grad_norm": 0.01730559952557087,
+ "learning_rate": 0.0006,
+ "loss": 4.542474746704102,
+ "step": 4059
+ },
+ {
+ "epoch": 56.391266375545854,
+ "grad_norm": 0.017029426991939545,
+ "learning_rate": 0.0006,
+ "loss": 4.552488327026367,
+ "step": 4060
+ },
+ {
+ "epoch": 56.40524017467249,
+ "grad_norm": 0.01745498552918434,
+ "learning_rate": 0.0006,
+ "loss": 4.52928352355957,
+ "step": 4061
+ },
+ {
+ "epoch": 56.419213973799124,
+ "grad_norm": 0.0168515145778656,
+ "learning_rate": 0.0006,
+ "loss": 4.4516801834106445,
+ "step": 4062
+ },
+ {
+ "epoch": 56.43318777292576,
+ "grad_norm": 0.018545417115092278,
+ "learning_rate": 0.0006,
+ "loss": 4.569811820983887,
+ "step": 4063
+ },
+ {
+ "epoch": 56.4471615720524,
+ "grad_norm": 0.02227838523685932,
+ "learning_rate": 0.0006,
+ "loss": 4.412224769592285,
+ "step": 4064
+ },
+ {
+ "epoch": 56.46113537117904,
+ "grad_norm": 0.023338504135608673,
+ "learning_rate": 0.0006,
+ "loss": 4.490184307098389,
+ "step": 4065
+ },
+ {
+ "epoch": 56.47510917030568,
+ "grad_norm": 0.02257535606622696,
+ "learning_rate": 0.0006,
+ "loss": 4.479152679443359,
+ "step": 4066
+ },
+ {
+ "epoch": 56.48908296943232,
+ "grad_norm": 0.022737598046660423,
+ "learning_rate": 0.0006,
+ "loss": 4.418235778808594,
+ "step": 4067
+ },
+ {
+ "epoch": 56.50305676855895,
+ "grad_norm": 0.024603573605418205,
+ "learning_rate": 0.0006,
+ "loss": 4.45781135559082,
+ "step": 4068
+ },
+ {
+ "epoch": 56.51703056768559,
+ "grad_norm": 0.024953778833150864,
+ "learning_rate": 0.0006,
+ "loss": 4.593907833099365,
+ "step": 4069
+ },
+ {
+ "epoch": 56.531004366812226,
+ "grad_norm": 0.02195645309984684,
+ "learning_rate": 0.0006,
+ "loss": 4.467643737792969,
+ "step": 4070
+ },
+ {
+ "epoch": 56.544978165938865,
+ "grad_norm": 0.017126578837633133,
+ "learning_rate": 0.0006,
+ "loss": 4.555484771728516,
+ "step": 4071
+ },
+ {
+ "epoch": 56.5589519650655,
+ "grad_norm": 0.019482605159282684,
+ "learning_rate": 0.0006,
+ "loss": 4.430915832519531,
+ "step": 4072
+ },
+ {
+ "epoch": 56.57292576419214,
+ "grad_norm": 0.0219392292201519,
+ "learning_rate": 0.0006,
+ "loss": 4.593365669250488,
+ "step": 4073
+ },
+ {
+ "epoch": 56.58689956331878,
+ "grad_norm": 0.02420145645737648,
+ "learning_rate": 0.0006,
+ "loss": 4.491467475891113,
+ "step": 4074
+ },
+ {
+ "epoch": 56.60087336244541,
+ "grad_norm": 0.021523723378777504,
+ "learning_rate": 0.0006,
+ "loss": 4.463634490966797,
+ "step": 4075
+ },
+ {
+ "epoch": 56.61484716157205,
+ "grad_norm": 0.020242048427462578,
+ "learning_rate": 0.0006,
+ "loss": 4.577968597412109,
+ "step": 4076
+ },
+ {
+ "epoch": 56.62882096069869,
+ "grad_norm": 0.017771173268556595,
+ "learning_rate": 0.0006,
+ "loss": 4.568269729614258,
+ "step": 4077
+ },
+ {
+ "epoch": 56.64279475982533,
+ "grad_norm": 0.01598544418811798,
+ "learning_rate": 0.0006,
+ "loss": 4.476841926574707,
+ "step": 4078
+ },
+ {
+ "epoch": 56.65676855895197,
+ "grad_norm": 0.0168446097522974,
+ "learning_rate": 0.0006,
+ "loss": 4.480965614318848,
+ "step": 4079
+ },
+ {
+ "epoch": 56.670742358078606,
+ "grad_norm": 0.014948660507798195,
+ "learning_rate": 0.0006,
+ "loss": 4.535400390625,
+ "step": 4080
+ },
+ {
+ "epoch": 56.68471615720524,
+ "grad_norm": 0.014899153262376785,
+ "learning_rate": 0.0006,
+ "loss": 4.476048946380615,
+ "step": 4081
+ },
+ {
+ "epoch": 56.698689956331876,
+ "grad_norm": 0.014051870442926884,
+ "learning_rate": 0.0006,
+ "loss": 4.533884048461914,
+ "step": 4082
+ },
+ {
+ "epoch": 56.712663755458514,
+ "grad_norm": 0.014708572067320347,
+ "learning_rate": 0.0006,
+ "loss": 4.586690902709961,
+ "step": 4083
+ },
+ {
+ "epoch": 56.72663755458515,
+ "grad_norm": 0.014797299169003963,
+ "learning_rate": 0.0006,
+ "loss": 4.446291446685791,
+ "step": 4084
+ },
+ {
+ "epoch": 56.74061135371179,
+ "grad_norm": 0.015256541781127453,
+ "learning_rate": 0.0006,
+ "loss": 4.553572654724121,
+ "step": 4085
+ },
+ {
+ "epoch": 56.75458515283843,
+ "grad_norm": 0.016042491421103477,
+ "learning_rate": 0.0006,
+ "loss": 4.507801055908203,
+ "step": 4086
+ },
+ {
+ "epoch": 56.76855895196506,
+ "grad_norm": 0.018835173919796944,
+ "learning_rate": 0.0006,
+ "loss": 4.393911838531494,
+ "step": 4087
+ },
+ {
+ "epoch": 56.7825327510917,
+ "grad_norm": 0.021994033828377724,
+ "learning_rate": 0.0006,
+ "loss": 4.543368339538574,
+ "step": 4088
+ },
+ {
+ "epoch": 56.79650655021834,
+ "grad_norm": 0.0230990182608366,
+ "learning_rate": 0.0006,
+ "loss": 4.507420539855957,
+ "step": 4089
+ },
+ {
+ "epoch": 56.81048034934498,
+ "grad_norm": 0.018818842247128487,
+ "learning_rate": 0.0006,
+ "loss": 4.499941349029541,
+ "step": 4090
+ },
+ {
+ "epoch": 56.82445414847162,
+ "grad_norm": 0.018257278949022293,
+ "learning_rate": 0.0006,
+ "loss": 4.380057334899902,
+ "step": 4091
+ },
+ {
+ "epoch": 56.838427947598255,
+ "grad_norm": 0.020073018968105316,
+ "learning_rate": 0.0006,
+ "loss": 4.572811603546143,
+ "step": 4092
+ },
+ {
+ "epoch": 56.852401746724894,
+ "grad_norm": 0.02078658528625965,
+ "learning_rate": 0.0006,
+ "loss": 4.446835517883301,
+ "step": 4093
+ },
+ {
+ "epoch": 56.866375545851525,
+ "grad_norm": 0.01793118566274643,
+ "learning_rate": 0.0006,
+ "loss": 4.517854690551758,
+ "step": 4094
+ },
+ {
+ "epoch": 56.880349344978164,
+ "grad_norm": 0.01608281210064888,
+ "learning_rate": 0.0006,
+ "loss": 4.506482124328613,
+ "step": 4095
+ },
+ {
+ "epoch": 56.8943231441048,
+ "grad_norm": 0.017405368387699127,
+ "learning_rate": 0.0006,
+ "loss": 4.439718246459961,
+ "step": 4096
+ },
+ {
+ "epoch": 56.90829694323144,
+ "grad_norm": 0.017930863425135612,
+ "learning_rate": 0.0006,
+ "loss": 4.409365653991699,
+ "step": 4097
+ },
+ {
+ "epoch": 56.92227074235808,
+ "grad_norm": 0.015757670626044273,
+ "learning_rate": 0.0006,
+ "loss": 4.505547523498535,
+ "step": 4098
+ },
+ {
+ "epoch": 56.93624454148472,
+ "grad_norm": 0.01580345816910267,
+ "learning_rate": 0.0006,
+ "loss": 4.462774753570557,
+ "step": 4099
+ },
+ {
+ "epoch": 56.95021834061135,
+ "grad_norm": 0.01659180410206318,
+ "learning_rate": 0.0006,
+ "loss": 4.49345588684082,
+ "step": 4100
+ },
+ {
+ "epoch": 56.96419213973799,
+ "grad_norm": 0.01877259835600853,
+ "learning_rate": 0.0006,
+ "loss": 4.574443340301514,
+ "step": 4101
+ },
+ {
+ "epoch": 56.97816593886463,
+ "grad_norm": 0.01977839693427086,
+ "learning_rate": 0.0006,
+ "loss": 4.574034690856934,
+ "step": 4102
+ },
+ {
+ "epoch": 56.992139737991266,
+ "grad_norm": 0.020627597346901894,
+ "learning_rate": 0.0006,
+ "loss": 4.50969123840332,
+ "step": 4103
+ },
+ {
+ "epoch": 57.0,
+ "grad_norm": 0.025128807872533798,
+ "learning_rate": 0.0006,
+ "loss": 4.401662826538086,
+ "step": 4104
+ },
+ {
+ "epoch": 57.0,
+ "eval_loss": 4.779252052307129,
+ "eval_runtime": 56.8104,
+ "eval_samples_per_second": 42.985,
+ "eval_steps_per_second": 1.355,
+ "step": 4104
+ },
+ {
+ "epoch": 57.01397379912664,
+ "grad_norm": 0.027712570503354073,
+ "learning_rate": 0.0006,
+ "loss": 4.472955703735352,
+ "step": 4105
+ },
+ {
+ "epoch": 57.02794759825328,
+ "grad_norm": 0.02164197526872158,
+ "learning_rate": 0.0006,
+ "loss": 4.387596130371094,
+ "step": 4106
+ },
+ {
+ "epoch": 57.041921397379916,
+ "grad_norm": 0.017972752451896667,
+ "learning_rate": 0.0006,
+ "loss": 4.4326276779174805,
+ "step": 4107
+ },
+ {
+ "epoch": 57.05589519650655,
+ "grad_norm": 0.020811019465327263,
+ "learning_rate": 0.0006,
+ "loss": 4.452951431274414,
+ "step": 4108
+ },
+ {
+ "epoch": 57.069868995633186,
+ "grad_norm": 0.019913259893655777,
+ "learning_rate": 0.0006,
+ "loss": 4.448301315307617,
+ "step": 4109
+ },
+ {
+ "epoch": 57.083842794759825,
+ "grad_norm": 0.017409102991223335,
+ "learning_rate": 0.0006,
+ "loss": 4.4108357429504395,
+ "step": 4110
+ },
+ {
+ "epoch": 57.09781659388646,
+ "grad_norm": 0.019374269992113113,
+ "learning_rate": 0.0006,
+ "loss": 4.398412227630615,
+ "step": 4111
+ },
+ {
+ "epoch": 57.1117903930131,
+ "grad_norm": 0.015521776862442493,
+ "learning_rate": 0.0006,
+ "loss": 4.508594989776611,
+ "step": 4112
+ },
+ {
+ "epoch": 57.12576419213974,
+ "grad_norm": 0.017201920971274376,
+ "learning_rate": 0.0006,
+ "loss": 4.449558734893799,
+ "step": 4113
+ },
+ {
+ "epoch": 57.13973799126637,
+ "grad_norm": 0.018062541261315346,
+ "learning_rate": 0.0006,
+ "loss": 4.517874240875244,
+ "step": 4114
+ },
+ {
+ "epoch": 57.15371179039301,
+ "grad_norm": 0.015846073627471924,
+ "learning_rate": 0.0006,
+ "loss": 4.427501678466797,
+ "step": 4115
+ },
+ {
+ "epoch": 57.16768558951965,
+ "grad_norm": 0.01849362626671791,
+ "learning_rate": 0.0006,
+ "loss": 4.433181285858154,
+ "step": 4116
+ },
+ {
+ "epoch": 57.18165938864629,
+ "grad_norm": 0.01948804222047329,
+ "learning_rate": 0.0006,
+ "loss": 4.446122646331787,
+ "step": 4117
+ },
+ {
+ "epoch": 57.19563318777293,
+ "grad_norm": 0.017354389652609825,
+ "learning_rate": 0.0006,
+ "loss": 4.431153774261475,
+ "step": 4118
+ },
+ {
+ "epoch": 57.209606986899566,
+ "grad_norm": 0.015817373991012573,
+ "learning_rate": 0.0006,
+ "loss": 4.448566913604736,
+ "step": 4119
+ },
+ {
+ "epoch": 57.223580786026204,
+ "grad_norm": 0.01635785959661007,
+ "learning_rate": 0.0006,
+ "loss": 4.408031463623047,
+ "step": 4120
+ },
+ {
+ "epoch": 57.237554585152836,
+ "grad_norm": 0.014528338797390461,
+ "learning_rate": 0.0006,
+ "loss": 4.541668891906738,
+ "step": 4121
+ },
+ {
+ "epoch": 57.251528384279474,
+ "grad_norm": 0.01589174196124077,
+ "learning_rate": 0.0006,
+ "loss": 4.513199329376221,
+ "step": 4122
+ },
+ {
+ "epoch": 57.26550218340611,
+ "grad_norm": 0.016093695536255836,
+ "learning_rate": 0.0006,
+ "loss": 4.516178131103516,
+ "step": 4123
+ },
+ {
+ "epoch": 57.27947598253275,
+ "grad_norm": 0.01788361556828022,
+ "learning_rate": 0.0006,
+ "loss": 4.524880409240723,
+ "step": 4124
+ },
+ {
+ "epoch": 57.29344978165939,
+ "grad_norm": 0.01591576263308525,
+ "learning_rate": 0.0006,
+ "loss": 4.44764518737793,
+ "step": 4125
+ },
+ {
+ "epoch": 57.30742358078603,
+ "grad_norm": 0.015452570281922817,
+ "learning_rate": 0.0006,
+ "loss": 4.46844482421875,
+ "step": 4126
+ },
+ {
+ "epoch": 57.32139737991266,
+ "grad_norm": 0.016404492780566216,
+ "learning_rate": 0.0006,
+ "loss": 4.443678855895996,
+ "step": 4127
+ },
+ {
+ "epoch": 57.3353711790393,
+ "grad_norm": 0.019967148080468178,
+ "learning_rate": 0.0006,
+ "loss": 4.430230140686035,
+ "step": 4128
+ },
+ {
+ "epoch": 57.34934497816594,
+ "grad_norm": 0.02594015561044216,
+ "learning_rate": 0.0006,
+ "loss": 4.47061824798584,
+ "step": 4129
+ },
+ {
+ "epoch": 57.36331877729258,
+ "grad_norm": 0.031096890568733215,
+ "learning_rate": 0.0006,
+ "loss": 4.462275981903076,
+ "step": 4130
+ },
+ {
+ "epoch": 57.377292576419215,
+ "grad_norm": 0.02834898792207241,
+ "learning_rate": 0.0006,
+ "loss": 4.41143798828125,
+ "step": 4131
+ },
+ {
+ "epoch": 57.391266375545854,
+ "grad_norm": 0.023621153086423874,
+ "learning_rate": 0.0006,
+ "loss": 4.562548637390137,
+ "step": 4132
+ },
+ {
+ "epoch": 57.40524017467249,
+ "grad_norm": 0.02277296595275402,
+ "learning_rate": 0.0006,
+ "loss": 4.456247329711914,
+ "step": 4133
+ },
+ {
+ "epoch": 57.419213973799124,
+ "grad_norm": 0.019868768751621246,
+ "learning_rate": 0.0006,
+ "loss": 4.524300575256348,
+ "step": 4134
+ },
+ {
+ "epoch": 57.43318777292576,
+ "grad_norm": 0.01887078955769539,
+ "learning_rate": 0.0006,
+ "loss": 4.540097236633301,
+ "step": 4135
+ },
+ {
+ "epoch": 57.4471615720524,
+ "grad_norm": 0.024268802255392075,
+ "learning_rate": 0.0006,
+ "loss": 4.397233963012695,
+ "step": 4136
+ },
+ {
+ "epoch": 57.46113537117904,
+ "grad_norm": 0.025739721953868866,
+ "learning_rate": 0.0006,
+ "loss": 4.354194641113281,
+ "step": 4137
+ },
+ {
+ "epoch": 57.47510917030568,
+ "grad_norm": 0.02442290261387825,
+ "learning_rate": 0.0006,
+ "loss": 4.303822994232178,
+ "step": 4138
+ },
+ {
+ "epoch": 57.48908296943232,
+ "grad_norm": 0.022215209901332855,
+ "learning_rate": 0.0006,
+ "loss": 4.4870452880859375,
+ "step": 4139
+ },
+ {
+ "epoch": 57.50305676855895,
+ "grad_norm": 0.01789158768951893,
+ "learning_rate": 0.0006,
+ "loss": 4.343809127807617,
+ "step": 4140
+ },
+ {
+ "epoch": 57.51703056768559,
+ "grad_norm": 0.020381739363074303,
+ "learning_rate": 0.0006,
+ "loss": 4.53312873840332,
+ "step": 4141
+ },
+ {
+ "epoch": 57.531004366812226,
+ "grad_norm": 0.019585080444812775,
+ "learning_rate": 0.0006,
+ "loss": 4.510162353515625,
+ "step": 4142
+ },
+ {
+ "epoch": 57.544978165938865,
+ "grad_norm": 0.01985127478837967,
+ "learning_rate": 0.0006,
+ "loss": 4.3654022216796875,
+ "step": 4143
+ },
+ {
+ "epoch": 57.5589519650655,
+ "grad_norm": 0.01962905190885067,
+ "learning_rate": 0.0006,
+ "loss": 4.560830116271973,
+ "step": 4144
+ },
+ {
+ "epoch": 57.57292576419214,
+ "grad_norm": 0.02679629437625408,
+ "learning_rate": 0.0006,
+ "loss": 4.300804138183594,
+ "step": 4145
+ },
+ {
+ "epoch": 57.58689956331878,
+ "grad_norm": 0.039830856025218964,
+ "learning_rate": 0.0006,
+ "loss": 4.54150915145874,
+ "step": 4146
+ },
+ {
+ "epoch": 57.60087336244541,
+ "grad_norm": 0.049527477473020554,
+ "learning_rate": 0.0006,
+ "loss": 4.405792236328125,
+ "step": 4147
+ },
+ {
+ "epoch": 57.61484716157205,
+ "grad_norm": 0.042966946959495544,
+ "learning_rate": 0.0006,
+ "loss": 4.500643730163574,
+ "step": 4148
+ },
+ {
+ "epoch": 57.62882096069869,
+ "grad_norm": 0.08999177813529968,
+ "learning_rate": 0.0006,
+ "loss": 4.504790782928467,
+ "step": 4149
+ },
+ {
+ "epoch": 57.64279475982533,
+ "grad_norm": 0.44693124294281006,
+ "learning_rate": 0.0006,
+ "loss": 4.845344066619873,
+ "step": 4150
+ },
+ {
+ "epoch": 57.65676855895197,
+ "grad_norm": 0.7143144011497498,
+ "learning_rate": 0.0006,
+ "loss": 6.907845497131348,
+ "step": 4151
+ },
+ {
+ "epoch": 57.670742358078606,
+ "grad_norm": 0.25239595770835876,
+ "learning_rate": 0.0006,
+ "loss": 7.082579612731934,
+ "step": 4152
+ },
+ {
+ "epoch": 57.68471615720524,
+ "grad_norm": 0.26550790667533875,
+ "learning_rate": 0.0006,
+ "loss": 7.50180196762085,
+ "step": 4153
+ },
+ {
+ "epoch": 57.698689956331876,
+ "grad_norm": 0.14735378324985504,
+ "learning_rate": 0.0006,
+ "loss": 7.214590072631836,
+ "step": 4154
+ },
+ {
+ "epoch": 57.712663755458514,
+ "grad_norm": 0.17685334384441376,
+ "learning_rate": 0.0006,
+ "loss": 6.861320495605469,
+ "step": 4155
+ },
+ {
+ "epoch": 57.72663755458515,
+ "grad_norm": 0.18231339752674103,
+ "learning_rate": 0.0006,
+ "loss": 6.791839599609375,
+ "step": 4156
+ },
+ {
+ "epoch": 57.74061135371179,
+ "grad_norm": 0.09230359643697739,
+ "learning_rate": 0.0006,
+ "loss": 6.760223388671875,
+ "step": 4157
+ },
+ {
+ "epoch": 57.75458515283843,
+ "grad_norm": 0.06918249279260635,
+ "learning_rate": 0.0006,
+ "loss": 6.532999515533447,
+ "step": 4158
+ },
+ {
+ "epoch": 57.76855895196506,
+ "grad_norm": 0.0901007279753685,
+ "learning_rate": 0.0006,
+ "loss": 6.52907133102417,
+ "step": 4159
+ },
+ {
+ "epoch": 57.7825327510917,
+ "grad_norm": 0.05678229779005051,
+ "learning_rate": 0.0006,
+ "loss": 6.295290470123291,
+ "step": 4160
+ },
+ {
+ "epoch": 57.79650655021834,
+ "grad_norm": 0.04450210556387901,
+ "learning_rate": 0.0006,
+ "loss": 6.14900541305542,
+ "step": 4161
+ },
+ {
+ "epoch": 57.81048034934498,
+ "grad_norm": 0.049191512167453766,
+ "learning_rate": 0.0006,
+ "loss": 6.19650411605835,
+ "step": 4162
+ },
+ {
+ "epoch": 57.82445414847162,
+ "grad_norm": 0.03908967226743698,
+ "learning_rate": 0.0006,
+ "loss": 6.054948806762695,
+ "step": 4163
+ },
+ {
+ "epoch": 57.838427947598255,
+ "grad_norm": 0.04023387283086777,
+ "learning_rate": 0.0006,
+ "loss": 6.034873962402344,
+ "step": 4164
+ },
+ {
+ "epoch": 57.852401746724894,
+ "grad_norm": 0.03304268419742584,
+ "learning_rate": 0.0006,
+ "loss": 5.769852161407471,
+ "step": 4165
+ },
+ {
+ "epoch": 57.866375545851525,
+ "grad_norm": 0.03432450816035271,
+ "learning_rate": 0.0006,
+ "loss": 5.788785457611084,
+ "step": 4166
+ },
+ {
+ "epoch": 57.880349344978164,
+ "grad_norm": 0.029965840280056,
+ "learning_rate": 0.0006,
+ "loss": 5.815243721008301,
+ "step": 4167
+ },
+ {
+ "epoch": 57.8943231441048,
+ "grad_norm": 0.03637698292732239,
+ "learning_rate": 0.0006,
+ "loss": 5.7204060554504395,
+ "step": 4168
+ },
+ {
+ "epoch": 57.90829694323144,
+ "grad_norm": 0.03477517142891884,
+ "learning_rate": 0.0006,
+ "loss": 5.605930328369141,
+ "step": 4169
+ },
+ {
+ "epoch": 57.92227074235808,
+ "grad_norm": 0.05061859264969826,
+ "learning_rate": 0.0006,
+ "loss": 5.547054767608643,
+ "step": 4170
+ },
+ {
+ "epoch": 57.93624454148472,
+ "grad_norm": 0.06299655884504318,
+ "learning_rate": 0.0006,
+ "loss": 5.552166938781738,
+ "step": 4171
+ },
+ {
+ "epoch": 57.95021834061135,
+ "grad_norm": 0.04215948283672333,
+ "learning_rate": 0.0006,
+ "loss": 5.375457763671875,
+ "step": 4172
+ },
+ {
+ "epoch": 57.96419213973799,
+ "grad_norm": 0.0348566472530365,
+ "learning_rate": 0.0006,
+ "loss": 5.418606281280518,
+ "step": 4173
+ },
+ {
+ "epoch": 57.97816593886463,
+ "grad_norm": 0.03807530924677849,
+ "learning_rate": 0.0006,
+ "loss": 5.403156280517578,
+ "step": 4174
+ },
+ {
+ "epoch": 57.992139737991266,
+ "grad_norm": 0.02989993803203106,
+ "learning_rate": 0.0006,
+ "loss": 5.355951309204102,
+ "step": 4175
+ },
+ {
+ "epoch": 58.0,
+ "grad_norm": 0.034300558269023895,
+ "learning_rate": 0.0006,
+ "loss": 5.44920539855957,
+ "step": 4176
+ },
+ {
+ "epoch": 58.0,
+ "eval_loss": 5.416731834411621,
+ "eval_runtime": 56.7934,
+ "eval_samples_per_second": 42.998,
+ "eval_steps_per_second": 1.356,
+ "step": 4176
+ },
+ {
+ "epoch": 58.01397379912664,
+ "grad_norm": 0.033961448818445206,
+ "learning_rate": 0.0006,
+ "loss": 5.250707626342773,
+ "step": 4177
+ },
+ {
+ "epoch": 58.02794759825328,
+ "grad_norm": 0.027343502268195152,
+ "learning_rate": 0.0006,
+ "loss": 5.149238109588623,
+ "step": 4178
+ },
+ {
+ "epoch": 58.041921397379916,
+ "grad_norm": 0.03033650480210781,
+ "learning_rate": 0.0006,
+ "loss": 5.072463512420654,
+ "step": 4179
+ },
+ {
+ "epoch": 58.05589519650655,
+ "grad_norm": 0.028262868523597717,
+ "learning_rate": 0.0006,
+ "loss": 5.109417915344238,
+ "step": 4180
+ },
+ {
+ "epoch": 58.069868995633186,
+ "grad_norm": 0.02373034693300724,
+ "learning_rate": 0.0006,
+ "loss": 5.082464218139648,
+ "step": 4181
+ },
+ {
+ "epoch": 58.083842794759825,
+ "grad_norm": 0.025879688560962677,
+ "learning_rate": 0.0006,
+ "loss": 5.056792259216309,
+ "step": 4182
+ },
+ {
+ "epoch": 58.09781659388646,
+ "grad_norm": 0.026252275332808495,
+ "learning_rate": 0.0006,
+ "loss": 5.064239978790283,
+ "step": 4183
+ },
+ {
+ "epoch": 58.1117903930131,
+ "grad_norm": 0.028177792206406593,
+ "learning_rate": 0.0006,
+ "loss": 4.922310829162598,
+ "step": 4184
+ },
+ {
+ "epoch": 58.12576419213974,
+ "grad_norm": 0.02630820870399475,
+ "learning_rate": 0.0006,
+ "loss": 4.992884635925293,
+ "step": 4185
+ },
+ {
+ "epoch": 58.13973799126637,
+ "grad_norm": 0.024890903383493423,
+ "learning_rate": 0.0006,
+ "loss": 4.910036563873291,
+ "step": 4186
+ },
+ {
+ "epoch": 58.15371179039301,
+ "grad_norm": 0.02822992391884327,
+ "learning_rate": 0.0006,
+ "loss": 4.866640090942383,
+ "step": 4187
+ },
+ {
+ "epoch": 58.16768558951965,
+ "grad_norm": 0.03253607079386711,
+ "learning_rate": 0.0006,
+ "loss": 4.9112958908081055,
+ "step": 4188
+ },
+ {
+ "epoch": 58.18165938864629,
+ "grad_norm": 0.03765944764018059,
+ "learning_rate": 0.0006,
+ "loss": 4.774468421936035,
+ "step": 4189
+ },
+ {
+ "epoch": 58.19563318777293,
+ "grad_norm": 0.039874881505966187,
+ "learning_rate": 0.0006,
+ "loss": 4.831678867340088,
+ "step": 4190
+ },
+ {
+ "epoch": 58.209606986899566,
+ "grad_norm": 0.034595925360918045,
+ "learning_rate": 0.0006,
+ "loss": 4.8243794441223145,
+ "step": 4191
+ },
+ {
+ "epoch": 58.223580786026204,
+ "grad_norm": 0.04840189591050148,
+ "learning_rate": 0.0006,
+ "loss": 4.854156494140625,
+ "step": 4192
+ },
+ {
+ "epoch": 58.237554585152836,
+ "grad_norm": 0.06358753889799118,
+ "learning_rate": 0.0006,
+ "loss": 4.836430549621582,
+ "step": 4193
+ },
+ {
+ "epoch": 58.251528384279474,
+ "grad_norm": 0.0507555715739727,
+ "learning_rate": 0.0006,
+ "loss": 4.88357400894165,
+ "step": 4194
+ },
+ {
+ "epoch": 58.26550218340611,
+ "grad_norm": 0.04384690150618553,
+ "learning_rate": 0.0006,
+ "loss": 4.774049282073975,
+ "step": 4195
+ },
+ {
+ "epoch": 58.27947598253275,
+ "grad_norm": 0.03246928006410599,
+ "learning_rate": 0.0006,
+ "loss": 4.686605930328369,
+ "step": 4196
+ },
+ {
+ "epoch": 58.29344978165939,
+ "grad_norm": 0.028878789395093918,
+ "learning_rate": 0.0006,
+ "loss": 4.796070575714111,
+ "step": 4197
+ },
+ {
+ "epoch": 58.30742358078603,
+ "grad_norm": 0.028175361454486847,
+ "learning_rate": 0.0006,
+ "loss": 4.728490829467773,
+ "step": 4198
+ },
+ {
+ "epoch": 58.32139737991266,
+ "grad_norm": 0.029428014531731606,
+ "learning_rate": 0.0006,
+ "loss": 4.7159624099731445,
+ "step": 4199
+ },
+ {
+ "epoch": 58.3353711790393,
+ "grad_norm": 0.031227827072143555,
+ "learning_rate": 0.0006,
+ "loss": 4.702625274658203,
+ "step": 4200
+ },
+ {
+ "epoch": 58.34934497816594,
+ "grad_norm": 0.028843428939580917,
+ "learning_rate": 0.0006,
+ "loss": 4.811799049377441,
+ "step": 4201
+ },
+ {
+ "epoch": 58.36331877729258,
+ "grad_norm": 0.026780391111969948,
+ "learning_rate": 0.0006,
+ "loss": 4.81662654876709,
+ "step": 4202
+ },
+ {
+ "epoch": 58.377292576419215,
+ "grad_norm": 0.019699757918715477,
+ "learning_rate": 0.0006,
+ "loss": 4.713685989379883,
+ "step": 4203
+ },
+ {
+ "epoch": 58.391266375545854,
+ "grad_norm": 0.024389250203967094,
+ "learning_rate": 0.0006,
+ "loss": 4.72354793548584,
+ "step": 4204
+ },
+ {
+ "epoch": 58.40524017467249,
+ "grad_norm": 0.02218274027109146,
+ "learning_rate": 0.0006,
+ "loss": 4.674125671386719,
+ "step": 4205
+ },
+ {
+ "epoch": 58.419213973799124,
+ "grad_norm": 0.02041488140821457,
+ "learning_rate": 0.0006,
+ "loss": 4.7909698486328125,
+ "step": 4206
+ },
+ {
+ "epoch": 58.43318777292576,
+ "grad_norm": 0.018561935052275658,
+ "learning_rate": 0.0006,
+ "loss": 4.808770179748535,
+ "step": 4207
+ },
+ {
+ "epoch": 58.4471615720524,
+ "grad_norm": 0.01708846725523472,
+ "learning_rate": 0.0006,
+ "loss": 4.701626777648926,
+ "step": 4208
+ },
+ {
+ "epoch": 58.46113537117904,
+ "grad_norm": 0.0222734697163105,
+ "learning_rate": 0.0006,
+ "loss": 4.505410194396973,
+ "step": 4209
+ },
+ {
+ "epoch": 58.47510917030568,
+ "grad_norm": 0.03197057917714119,
+ "learning_rate": 0.0006,
+ "loss": 4.610360145568848,
+ "step": 4210
+ },
+ {
+ "epoch": 58.48908296943232,
+ "grad_norm": 0.06854520738124847,
+ "learning_rate": 0.0006,
+ "loss": 4.625649452209473,
+ "step": 4211
+ },
+ {
+ "epoch": 58.50305676855895,
+ "grad_norm": 0.09981521219015121,
+ "learning_rate": 0.0006,
+ "loss": 4.832150459289551,
+ "step": 4212
+ },
+ {
+ "epoch": 58.51703056768559,
+ "grad_norm": 0.046356040984392166,
+ "learning_rate": 0.0006,
+ "loss": 4.5641632080078125,
+ "step": 4213
+ },
+ {
+ "epoch": 58.531004366812226,
+ "grad_norm": 0.03882667422294617,
+ "learning_rate": 0.0006,
+ "loss": 4.666974067687988,
+ "step": 4214
+ },
+ {
+ "epoch": 58.544978165938865,
+ "grad_norm": 0.033203233033418655,
+ "learning_rate": 0.0006,
+ "loss": 4.629413604736328,
+ "step": 4215
+ },
+ {
+ "epoch": 58.5589519650655,
+ "grad_norm": 0.02806735597550869,
+ "learning_rate": 0.0006,
+ "loss": 4.763245582580566,
+ "step": 4216
+ },
+ {
+ "epoch": 58.57292576419214,
+ "grad_norm": 0.026728278025984764,
+ "learning_rate": 0.0006,
+ "loss": 4.629597187042236,
+ "step": 4217
+ },
+ {
+ "epoch": 58.58689956331878,
+ "grad_norm": 0.024380959570407867,
+ "learning_rate": 0.0006,
+ "loss": 4.633614540100098,
+ "step": 4218
+ },
+ {
+ "epoch": 58.60087336244541,
+ "grad_norm": 0.022312544286251068,
+ "learning_rate": 0.0006,
+ "loss": 4.671379566192627,
+ "step": 4219
+ },
+ {
+ "epoch": 58.61484716157205,
+ "grad_norm": 0.022049568593502045,
+ "learning_rate": 0.0006,
+ "loss": 4.661557197570801,
+ "step": 4220
+ },
+ {
+ "epoch": 58.62882096069869,
+ "grad_norm": 0.02119818702340126,
+ "learning_rate": 0.0006,
+ "loss": 4.688562393188477,
+ "step": 4221
+ },
+ {
+ "epoch": 58.64279475982533,
+ "grad_norm": 0.019769301638007164,
+ "learning_rate": 0.0006,
+ "loss": 4.607748031616211,
+ "step": 4222
+ },
+ {
+ "epoch": 58.65676855895197,
+ "grad_norm": 0.02124079130589962,
+ "learning_rate": 0.0006,
+ "loss": 4.570640563964844,
+ "step": 4223
+ },
+ {
+ "epoch": 58.670742358078606,
+ "grad_norm": 0.016344040632247925,
+ "learning_rate": 0.0006,
+ "loss": 4.676113128662109,
+ "step": 4224
+ },
+ {
+ "epoch": 58.68471615720524,
+ "grad_norm": 0.01735256239771843,
+ "learning_rate": 0.0006,
+ "loss": 4.62088680267334,
+ "step": 4225
+ },
+ {
+ "epoch": 58.698689956331876,
+ "grad_norm": 0.019323458895087242,
+ "learning_rate": 0.0006,
+ "loss": 4.579405307769775,
+ "step": 4226
+ },
+ {
+ "epoch": 58.712663755458514,
+ "grad_norm": 0.01760544814169407,
+ "learning_rate": 0.0006,
+ "loss": 4.625407695770264,
+ "step": 4227
+ },
+ {
+ "epoch": 58.72663755458515,
+ "grad_norm": 0.015593250282108784,
+ "learning_rate": 0.0006,
+ "loss": 4.6326117515563965,
+ "step": 4228
+ },
+ {
+ "epoch": 58.74061135371179,
+ "grad_norm": 0.014357523061335087,
+ "learning_rate": 0.0006,
+ "loss": 4.6869401931762695,
+ "step": 4229
+ },
+ {
+ "epoch": 58.75458515283843,
+ "grad_norm": 0.01604357920587063,
+ "learning_rate": 0.0006,
+ "loss": 4.690784931182861,
+ "step": 4230
+ },
+ {
+ "epoch": 58.76855895196506,
+ "grad_norm": 0.01673845760524273,
+ "learning_rate": 0.0006,
+ "loss": 4.513410568237305,
+ "step": 4231
+ },
+ {
+ "epoch": 58.7825327510917,
+ "grad_norm": 0.013657779432833195,
+ "learning_rate": 0.0006,
+ "loss": 4.637057781219482,
+ "step": 4232
+ },
+ {
+ "epoch": 58.79650655021834,
+ "grad_norm": 0.013373114168643951,
+ "learning_rate": 0.0006,
+ "loss": 4.719967842102051,
+ "step": 4233
+ },
+ {
+ "epoch": 58.81048034934498,
+ "grad_norm": 0.016345568001270294,
+ "learning_rate": 0.0006,
+ "loss": 4.604109764099121,
+ "step": 4234
+ },
+ {
+ "epoch": 58.82445414847162,
+ "grad_norm": 0.01573033444583416,
+ "learning_rate": 0.0006,
+ "loss": 4.6471428871154785,
+ "step": 4235
+ },
+ {
+ "epoch": 58.838427947598255,
+ "grad_norm": 0.01517449039965868,
+ "learning_rate": 0.0006,
+ "loss": 4.722684860229492,
+ "step": 4236
+ },
+ {
+ "epoch": 58.852401746724894,
+ "grad_norm": 0.0145102022215724,
+ "learning_rate": 0.0006,
+ "loss": 4.662420272827148,
+ "step": 4237
+ },
+ {
+ "epoch": 58.866375545851525,
+ "grad_norm": 0.01457217987626791,
+ "learning_rate": 0.0006,
+ "loss": 4.585169792175293,
+ "step": 4238
+ },
+ {
+ "epoch": 58.880349344978164,
+ "grad_norm": 0.01406773366034031,
+ "learning_rate": 0.0006,
+ "loss": 4.5413665771484375,
+ "step": 4239
+ },
+ {
+ "epoch": 58.8943231441048,
+ "grad_norm": 0.012653871439397335,
+ "learning_rate": 0.0006,
+ "loss": 4.553636074066162,
+ "step": 4240
+ },
+ {
+ "epoch": 58.90829694323144,
+ "grad_norm": 0.01261200476437807,
+ "learning_rate": 0.0006,
+ "loss": 4.534193992614746,
+ "step": 4241
+ },
+ {
+ "epoch": 58.92227074235808,
+ "grad_norm": 0.012955864891409874,
+ "learning_rate": 0.0006,
+ "loss": 4.697248458862305,
+ "step": 4242
+ },
+ {
+ "epoch": 58.93624454148472,
+ "grad_norm": 0.012134749442338943,
+ "learning_rate": 0.0006,
+ "loss": 4.557095527648926,
+ "step": 4243
+ },
+ {
+ "epoch": 58.95021834061135,
+ "grad_norm": 0.012164677493274212,
+ "learning_rate": 0.0006,
+ "loss": 4.623579025268555,
+ "step": 4244
+ },
+ {
+ "epoch": 58.96419213973799,
+ "grad_norm": 0.013802947476506233,
+ "learning_rate": 0.0006,
+ "loss": 4.562838077545166,
+ "step": 4245
+ },
+ {
+ "epoch": 58.97816593886463,
+ "grad_norm": 0.011420476250350475,
+ "learning_rate": 0.0006,
+ "loss": 4.62265682220459,
+ "step": 4246
+ },
+ {
+ "epoch": 58.992139737991266,
+ "grad_norm": 0.01169576682150364,
+ "learning_rate": 0.0006,
+ "loss": 4.727449417114258,
+ "step": 4247
+ },
+ {
+ "epoch": 59.0,
+ "grad_norm": 0.013577710837125778,
+ "learning_rate": 0.0006,
+ "loss": 4.536325454711914,
+ "step": 4248
+ },
+ {
+ "epoch": 59.0,
+ "eval_loss": 4.86527681350708,
+ "eval_runtime": 60.4463,
+ "eval_samples_per_second": 40.399,
+ "eval_steps_per_second": 1.274,
+ "step": 4248
+ },
+ {
+ "epoch": 59.01397379912664,
+ "grad_norm": 0.015251655131578445,
+ "learning_rate": 0.0006,
+ "loss": 4.634598731994629,
+ "step": 4249
+ },
+ {
+ "epoch": 59.02794759825328,
+ "grad_norm": 0.017414938658475876,
+ "learning_rate": 0.0006,
+ "loss": 4.521329879760742,
+ "step": 4250
+ },
+ {
+ "epoch": 59.041921397379916,
+ "grad_norm": 0.01656588353216648,
+ "learning_rate": 0.0006,
+ "loss": 4.446192264556885,
+ "step": 4251
+ },
+ {
+ "epoch": 59.05589519650655,
+ "grad_norm": 0.012038925662636757,
+ "learning_rate": 0.0006,
+ "loss": 4.523855209350586,
+ "step": 4252
+ },
+ {
+ "epoch": 59.069868995633186,
+ "grad_norm": 0.014403033070266247,
+ "learning_rate": 0.0006,
+ "loss": 4.561079978942871,
+ "step": 4253
+ },
+ {
+ "epoch": 59.083842794759825,
+ "grad_norm": 0.016243597492575645,
+ "learning_rate": 0.0006,
+ "loss": 4.673602104187012,
+ "step": 4254
+ },
+ {
+ "epoch": 59.09781659388646,
+ "grad_norm": 0.016492877155542374,
+ "learning_rate": 0.0006,
+ "loss": 4.564824104309082,
+ "step": 4255
+ },
+ {
+ "epoch": 59.1117903930131,
+ "grad_norm": 0.015980370342731476,
+ "learning_rate": 0.0006,
+ "loss": 4.339249134063721,
+ "step": 4256
+ },
+ {
+ "epoch": 59.12576419213974,
+ "grad_norm": 0.019687356427311897,
+ "learning_rate": 0.0006,
+ "loss": 4.481834888458252,
+ "step": 4257
+ },
+ {
+ "epoch": 59.13973799126637,
+ "grad_norm": 0.02983798086643219,
+ "learning_rate": 0.0006,
+ "loss": 4.4495697021484375,
+ "step": 4258
+ },
+ {
+ "epoch": 59.15371179039301,
+ "grad_norm": 0.04600981995463371,
+ "learning_rate": 0.0006,
+ "loss": 4.567838668823242,
+ "step": 4259
+ },
+ {
+ "epoch": 59.16768558951965,
+ "grad_norm": 0.04852001741528511,
+ "learning_rate": 0.0006,
+ "loss": 4.594701766967773,
+ "step": 4260
+ },
+ {
+ "epoch": 59.18165938864629,
+ "grad_norm": 0.03429504856467247,
+ "learning_rate": 0.0006,
+ "loss": 4.573735237121582,
+ "step": 4261
+ },
+ {
+ "epoch": 59.19563318777293,
+ "grad_norm": 0.029345160350203514,
+ "learning_rate": 0.0006,
+ "loss": 4.488258361816406,
+ "step": 4262
+ },
+ {
+ "epoch": 59.209606986899566,
+ "grad_norm": 0.01961623504757881,
+ "learning_rate": 0.0006,
+ "loss": 4.545222282409668,
+ "step": 4263
+ },
+ {
+ "epoch": 59.223580786026204,
+ "grad_norm": 0.020406130701303482,
+ "learning_rate": 0.0006,
+ "loss": 4.471898078918457,
+ "step": 4264
+ },
+ {
+ "epoch": 59.237554585152836,
+ "grad_norm": 0.020500272512435913,
+ "learning_rate": 0.0006,
+ "loss": 4.538337707519531,
+ "step": 4265
+ },
+ {
+ "epoch": 59.251528384279474,
+ "grad_norm": 0.020727017894387245,
+ "learning_rate": 0.0006,
+ "loss": 4.437284469604492,
+ "step": 4266
+ },
+ {
+ "epoch": 59.26550218340611,
+ "grad_norm": 0.03395998477935791,
+ "learning_rate": 0.0006,
+ "loss": 4.459124565124512,
+ "step": 4267
+ },
+ {
+ "epoch": 59.27947598253275,
+ "grad_norm": 0.04635605961084366,
+ "learning_rate": 0.0006,
+ "loss": 4.520228385925293,
+ "step": 4268
+ },
+ {
+ "epoch": 59.29344978165939,
+ "grad_norm": 0.04253006726503372,
+ "learning_rate": 0.0006,
+ "loss": 4.578673362731934,
+ "step": 4269
+ },
+ {
+ "epoch": 59.30742358078603,
+ "grad_norm": 0.02403509058058262,
+ "learning_rate": 0.0006,
+ "loss": 4.484639644622803,
+ "step": 4270
+ },
+ {
+ "epoch": 59.32139737991266,
+ "grad_norm": 0.02254190482199192,
+ "learning_rate": 0.0006,
+ "loss": 4.511314868927002,
+ "step": 4271
+ },
+ {
+ "epoch": 59.3353711790393,
+ "grad_norm": 0.021585190668702126,
+ "learning_rate": 0.0006,
+ "loss": 4.530239105224609,
+ "step": 4272
+ },
+ {
+ "epoch": 59.34934497816594,
+ "grad_norm": 0.021210316568613052,
+ "learning_rate": 0.0006,
+ "loss": 4.593514442443848,
+ "step": 4273
+ },
+ {
+ "epoch": 59.36331877729258,
+ "grad_norm": 0.019317107275128365,
+ "learning_rate": 0.0006,
+ "loss": 4.5591535568237305,
+ "step": 4274
+ },
+ {
+ "epoch": 59.377292576419215,
+ "grad_norm": 0.018059708178043365,
+ "learning_rate": 0.0006,
+ "loss": 4.4499592781066895,
+ "step": 4275
+ },
+ {
+ "epoch": 59.391266375545854,
+ "grad_norm": 0.017847446724772453,
+ "learning_rate": 0.0006,
+ "loss": 4.473999500274658,
+ "step": 4276
+ },
+ {
+ "epoch": 59.40524017467249,
+ "grad_norm": 0.014944756403565407,
+ "learning_rate": 0.0006,
+ "loss": 4.5934529304504395,
+ "step": 4277
+ },
+ {
+ "epoch": 59.419213973799124,
+ "grad_norm": 0.01663832925260067,
+ "learning_rate": 0.0006,
+ "loss": 4.504262924194336,
+ "step": 4278
+ },
+ {
+ "epoch": 59.43318777292576,
+ "grad_norm": 0.015571796335279942,
+ "learning_rate": 0.0006,
+ "loss": 4.512310981750488,
+ "step": 4279
+ },
+ {
+ "epoch": 59.4471615720524,
+ "grad_norm": 0.015896400436758995,
+ "learning_rate": 0.0006,
+ "loss": 4.452714920043945,
+ "step": 4280
+ },
+ {
+ "epoch": 59.46113537117904,
+ "grad_norm": 0.016010284423828125,
+ "learning_rate": 0.0006,
+ "loss": 4.6188063621521,
+ "step": 4281
+ },
+ {
+ "epoch": 59.47510917030568,
+ "grad_norm": 0.015075593255460262,
+ "learning_rate": 0.0006,
+ "loss": 4.477978229522705,
+ "step": 4282
+ },
+ {
+ "epoch": 59.48908296943232,
+ "grad_norm": 0.014698950573801994,
+ "learning_rate": 0.0006,
+ "loss": 4.544955730438232,
+ "step": 4283
+ },
+ {
+ "epoch": 59.50305676855895,
+ "grad_norm": 0.0136245833709836,
+ "learning_rate": 0.0006,
+ "loss": 4.530460357666016,
+ "step": 4284
+ },
+ {
+ "epoch": 59.51703056768559,
+ "grad_norm": 0.013118310831487179,
+ "learning_rate": 0.0006,
+ "loss": 4.614217758178711,
+ "step": 4285
+ },
+ {
+ "epoch": 59.531004366812226,
+ "grad_norm": 0.01295961532741785,
+ "learning_rate": 0.0006,
+ "loss": 4.43156099319458,
+ "step": 4286
+ },
+ {
+ "epoch": 59.544978165938865,
+ "grad_norm": 0.012429811991751194,
+ "learning_rate": 0.0006,
+ "loss": 4.517333030700684,
+ "step": 4287
+ },
+ {
+ "epoch": 59.5589519650655,
+ "grad_norm": 0.01261830423027277,
+ "learning_rate": 0.0006,
+ "loss": 4.569780349731445,
+ "step": 4288
+ },
+ {
+ "epoch": 59.57292576419214,
+ "grad_norm": 0.012394499965012074,
+ "learning_rate": 0.0006,
+ "loss": 4.5970869064331055,
+ "step": 4289
+ },
+ {
+ "epoch": 59.58689956331878,
+ "grad_norm": 0.01290759164839983,
+ "learning_rate": 0.0006,
+ "loss": 4.437100410461426,
+ "step": 4290
+ },
+ {
+ "epoch": 59.60087336244541,
+ "grad_norm": 0.012042169459164143,
+ "learning_rate": 0.0006,
+ "loss": 4.567404747009277,
+ "step": 4291
+ },
+ {
+ "epoch": 59.61484716157205,
+ "grad_norm": 0.012338520959019661,
+ "learning_rate": 0.0006,
+ "loss": 4.476193428039551,
+ "step": 4292
+ },
+ {
+ "epoch": 59.62882096069869,
+ "grad_norm": 0.01321258582174778,
+ "learning_rate": 0.0006,
+ "loss": 4.566680431365967,
+ "step": 4293
+ },
+ {
+ "epoch": 59.64279475982533,
+ "grad_norm": 0.013478122651576996,
+ "learning_rate": 0.0006,
+ "loss": 4.487398147583008,
+ "step": 4294
+ },
+ {
+ "epoch": 59.65676855895197,
+ "grad_norm": 0.01537603884935379,
+ "learning_rate": 0.0006,
+ "loss": 4.449337005615234,
+ "step": 4295
+ },
+ {
+ "epoch": 59.670742358078606,
+ "grad_norm": 0.017140565440058708,
+ "learning_rate": 0.0006,
+ "loss": 4.5174360275268555,
+ "step": 4296
+ },
+ {
+ "epoch": 59.68471615720524,
+ "grad_norm": 0.017931461334228516,
+ "learning_rate": 0.0006,
+ "loss": 4.522104740142822,
+ "step": 4297
+ },
+ {
+ "epoch": 59.698689956331876,
+ "grad_norm": 0.018037918955087662,
+ "learning_rate": 0.0006,
+ "loss": 4.496928691864014,
+ "step": 4298
+ },
+ {
+ "epoch": 59.712663755458514,
+ "grad_norm": 0.01684504561126232,
+ "learning_rate": 0.0006,
+ "loss": 4.5332417488098145,
+ "step": 4299
+ },
+ {
+ "epoch": 59.72663755458515,
+ "grad_norm": 0.016268383711576462,
+ "learning_rate": 0.0006,
+ "loss": 4.468893051147461,
+ "step": 4300
+ },
+ {
+ "epoch": 59.74061135371179,
+ "grad_norm": 0.014859385788440704,
+ "learning_rate": 0.0006,
+ "loss": 4.551092147827148,
+ "step": 4301
+ },
+ {
+ "epoch": 59.75458515283843,
+ "grad_norm": 0.013589047826826572,
+ "learning_rate": 0.0006,
+ "loss": 4.470579147338867,
+ "step": 4302
+ },
+ {
+ "epoch": 59.76855895196506,
+ "grad_norm": 0.015704551711678505,
+ "learning_rate": 0.0006,
+ "loss": 4.330068111419678,
+ "step": 4303
+ },
+ {
+ "epoch": 59.7825327510917,
+ "grad_norm": 0.017869921401143074,
+ "learning_rate": 0.0006,
+ "loss": 4.428255081176758,
+ "step": 4304
+ },
+ {
+ "epoch": 59.79650655021834,
+ "grad_norm": 0.019828537479043007,
+ "learning_rate": 0.0006,
+ "loss": 4.496023178100586,
+ "step": 4305
+ },
+ {
+ "epoch": 59.81048034934498,
+ "grad_norm": 0.015856236219406128,
+ "learning_rate": 0.0006,
+ "loss": 4.427152156829834,
+ "step": 4306
+ },
+ {
+ "epoch": 59.82445414847162,
+ "grad_norm": 0.013349304907023907,
+ "learning_rate": 0.0006,
+ "loss": 4.417333602905273,
+ "step": 4307
+ },
+ {
+ "epoch": 59.838427947598255,
+ "grad_norm": 0.014111637137830257,
+ "learning_rate": 0.0006,
+ "loss": 4.4507575035095215,
+ "step": 4308
+ },
+ {
+ "epoch": 59.852401746724894,
+ "grad_norm": 0.015186597593128681,
+ "learning_rate": 0.0006,
+ "loss": 4.375116348266602,
+ "step": 4309
+ },
+ {
+ "epoch": 59.866375545851525,
+ "grad_norm": 0.01604730449616909,
+ "learning_rate": 0.0006,
+ "loss": 4.604518890380859,
+ "step": 4310
+ },
+ {
+ "epoch": 59.880349344978164,
+ "grad_norm": 0.015308464877307415,
+ "learning_rate": 0.0006,
+ "loss": 4.507232189178467,
+ "step": 4311
+ },
+ {
+ "epoch": 59.8943231441048,
+ "grad_norm": 0.014017206616699696,
+ "learning_rate": 0.0006,
+ "loss": 4.398487091064453,
+ "step": 4312
+ },
+ {
+ "epoch": 59.90829694323144,
+ "grad_norm": 0.012911394238471985,
+ "learning_rate": 0.0006,
+ "loss": 4.567872047424316,
+ "step": 4313
+ },
+ {
+ "epoch": 59.92227074235808,
+ "grad_norm": 0.012775142677128315,
+ "learning_rate": 0.0006,
+ "loss": 4.562105655670166,
+ "step": 4314
+ },
+ {
+ "epoch": 59.93624454148472,
+ "grad_norm": 0.014750408008694649,
+ "learning_rate": 0.0006,
+ "loss": 4.456263542175293,
+ "step": 4315
+ },
+ {
+ "epoch": 59.95021834061135,
+ "grad_norm": 0.016726011410355568,
+ "learning_rate": 0.0006,
+ "loss": 4.360722064971924,
+ "step": 4316
+ },
+ {
+ "epoch": 59.96419213973799,
+ "grad_norm": 0.01734776981174946,
+ "learning_rate": 0.0006,
+ "loss": 4.572544574737549,
+ "step": 4317
+ },
+ {
+ "epoch": 59.97816593886463,
+ "grad_norm": 0.01822415366768837,
+ "learning_rate": 0.0006,
+ "loss": 4.490501880645752,
+ "step": 4318
+ },
+ {
+ "epoch": 59.992139737991266,
+ "grad_norm": 0.014638577587902546,
+ "learning_rate": 0.0006,
+ "loss": 4.440951824188232,
+ "step": 4319
+ },
+ {
+ "epoch": 60.0,
+ "grad_norm": 0.01255202479660511,
+ "learning_rate": 0.0006,
+ "loss": 4.559123516082764,
+ "step": 4320
+ },
+ {
+ "epoch": 60.0,
+ "eval_loss": 4.741089344024658,
+ "eval_runtime": 56.8997,
+ "eval_samples_per_second": 42.918,
+ "eval_steps_per_second": 1.353,
+ "step": 4320
+ },
+ {
+ "epoch": 60.01397379912664,
+ "grad_norm": 0.016548393294215202,
+ "learning_rate": 0.0006,
+ "loss": 4.442024230957031,
+ "step": 4321
+ },
+ {
+ "epoch": 60.02794759825328,
+ "grad_norm": 0.019759127870202065,
+ "learning_rate": 0.0006,
+ "loss": 4.394334316253662,
+ "step": 4322
+ },
+ {
+ "epoch": 60.041921397379916,
+ "grad_norm": 0.018325170502066612,
+ "learning_rate": 0.0006,
+ "loss": 4.318270683288574,
+ "step": 4323
+ },
+ {
+ "epoch": 60.05589519650655,
+ "grad_norm": 0.015289463102817535,
+ "learning_rate": 0.0006,
+ "loss": 4.318228721618652,
+ "step": 4324
+ },
+ {
+ "epoch": 60.069868995633186,
+ "grad_norm": 0.015424872748553753,
+ "learning_rate": 0.0006,
+ "loss": 4.3605499267578125,
+ "step": 4325
+ },
+ {
+ "epoch": 60.083842794759825,
+ "grad_norm": 0.014085669070482254,
+ "learning_rate": 0.0006,
+ "loss": 4.359585762023926,
+ "step": 4326
+ },
+ {
+ "epoch": 60.09781659388646,
+ "grad_norm": 0.014004210010170937,
+ "learning_rate": 0.0006,
+ "loss": 4.447783470153809,
+ "step": 4327
+ },
+ {
+ "epoch": 60.1117903930131,
+ "grad_norm": 0.013514582999050617,
+ "learning_rate": 0.0006,
+ "loss": 4.4850239753723145,
+ "step": 4328
+ },
+ {
+ "epoch": 60.12576419213974,
+ "grad_norm": 0.013049394823610783,
+ "learning_rate": 0.0006,
+ "loss": 4.528675556182861,
+ "step": 4329
+ },
+ {
+ "epoch": 60.13973799126637,
+ "grad_norm": 0.01378058921545744,
+ "learning_rate": 0.0006,
+ "loss": 4.444823741912842,
+ "step": 4330
+ },
+ {
+ "epoch": 60.15371179039301,
+ "grad_norm": 0.013752233237028122,
+ "learning_rate": 0.0006,
+ "loss": 4.5205206871032715,
+ "step": 4331
+ },
+ {
+ "epoch": 60.16768558951965,
+ "grad_norm": 0.013431803323328495,
+ "learning_rate": 0.0006,
+ "loss": 4.547320365905762,
+ "step": 4332
+ },
+ {
+ "epoch": 60.18165938864629,
+ "grad_norm": 0.014544196426868439,
+ "learning_rate": 0.0006,
+ "loss": 4.517143249511719,
+ "step": 4333
+ },
+ {
+ "epoch": 60.19563318777293,
+ "grad_norm": 0.01612176187336445,
+ "learning_rate": 0.0006,
+ "loss": 4.517232894897461,
+ "step": 4334
+ },
+ {
+ "epoch": 60.209606986899566,
+ "grad_norm": 0.01803239807486534,
+ "learning_rate": 0.0006,
+ "loss": 4.377574920654297,
+ "step": 4335
+ },
+ {
+ "epoch": 60.223580786026204,
+ "grad_norm": 0.014871489256620407,
+ "learning_rate": 0.0006,
+ "loss": 4.538285255432129,
+ "step": 4336
+ },
+ {
+ "epoch": 60.237554585152836,
+ "grad_norm": 0.013927377760410309,
+ "learning_rate": 0.0006,
+ "loss": 4.386175632476807,
+ "step": 4337
+ },
+ {
+ "epoch": 60.251528384279474,
+ "grad_norm": 0.014757219702005386,
+ "learning_rate": 0.0006,
+ "loss": 4.528553009033203,
+ "step": 4338
+ },
+ {
+ "epoch": 60.26550218340611,
+ "grad_norm": 0.01433873176574707,
+ "learning_rate": 0.0006,
+ "loss": 4.476929187774658,
+ "step": 4339
+ },
+ {
+ "epoch": 60.27947598253275,
+ "grad_norm": 0.015768803656101227,
+ "learning_rate": 0.0006,
+ "loss": 4.50970458984375,
+ "step": 4340
+ },
+ {
+ "epoch": 60.29344978165939,
+ "grad_norm": 0.015651429072022438,
+ "learning_rate": 0.0006,
+ "loss": 4.427133083343506,
+ "step": 4341
+ },
+ {
+ "epoch": 60.30742358078603,
+ "grad_norm": 0.01673000678420067,
+ "learning_rate": 0.0006,
+ "loss": 4.447723865509033,
+ "step": 4342
+ },
+ {
+ "epoch": 60.32139737991266,
+ "grad_norm": 0.018481194972991943,
+ "learning_rate": 0.0006,
+ "loss": 4.559832572937012,
+ "step": 4343
+ },
+ {
+ "epoch": 60.3353711790393,
+ "grad_norm": 0.016731027513742447,
+ "learning_rate": 0.0006,
+ "loss": 4.470893859863281,
+ "step": 4344
+ },
+ {
+ "epoch": 60.34934497816594,
+ "grad_norm": 0.01504011545330286,
+ "learning_rate": 0.0006,
+ "loss": 4.483014106750488,
+ "step": 4345
+ },
+ {
+ "epoch": 60.36331877729258,
+ "grad_norm": 0.01602456159889698,
+ "learning_rate": 0.0006,
+ "loss": 4.548428058624268,
+ "step": 4346
+ },
+ {
+ "epoch": 60.377292576419215,
+ "grad_norm": 0.016027364879846573,
+ "learning_rate": 0.0006,
+ "loss": 4.426779747009277,
+ "step": 4347
+ },
+ {
+ "epoch": 60.391266375545854,
+ "grad_norm": 0.01873827911913395,
+ "learning_rate": 0.0006,
+ "loss": 4.353686332702637,
+ "step": 4348
+ },
+ {
+ "epoch": 60.40524017467249,
+ "grad_norm": 0.01907406374812126,
+ "learning_rate": 0.0006,
+ "loss": 4.502945423126221,
+ "step": 4349
+ },
+ {
+ "epoch": 60.419213973799124,
+ "grad_norm": 0.016083354130387306,
+ "learning_rate": 0.0006,
+ "loss": 4.431731700897217,
+ "step": 4350
+ },
+ {
+ "epoch": 60.43318777292576,
+ "grad_norm": 0.015293709933757782,
+ "learning_rate": 0.0006,
+ "loss": 4.480112552642822,
+ "step": 4351
+ },
+ {
+ "epoch": 60.4471615720524,
+ "grad_norm": 0.01688101328909397,
+ "learning_rate": 0.0006,
+ "loss": 4.50853157043457,
+ "step": 4352
+ },
+ {
+ "epoch": 60.46113537117904,
+ "grad_norm": 0.02094118855893612,
+ "learning_rate": 0.0006,
+ "loss": 4.359930038452148,
+ "step": 4353
+ },
+ {
+ "epoch": 60.47510917030568,
+ "grad_norm": 0.018188107758760452,
+ "learning_rate": 0.0006,
+ "loss": 4.549592018127441,
+ "step": 4354
+ },
+ {
+ "epoch": 60.48908296943232,
+ "grad_norm": 0.01462319865822792,
+ "learning_rate": 0.0006,
+ "loss": 4.49467658996582,
+ "step": 4355
+ },
+ {
+ "epoch": 60.50305676855895,
+ "grad_norm": 0.017508579418063164,
+ "learning_rate": 0.0006,
+ "loss": 4.398318767547607,
+ "step": 4356
+ },
+ {
+ "epoch": 60.51703056768559,
+ "grad_norm": 0.018200315535068512,
+ "learning_rate": 0.0006,
+ "loss": 4.44169282913208,
+ "step": 4357
+ },
+ {
+ "epoch": 60.531004366812226,
+ "grad_norm": 0.01717468351125717,
+ "learning_rate": 0.0006,
+ "loss": 4.481400489807129,
+ "step": 4358
+ },
+ {
+ "epoch": 60.544978165938865,
+ "grad_norm": 0.019329151138663292,
+ "learning_rate": 0.0006,
+ "loss": 4.578839302062988,
+ "step": 4359
+ },
+ {
+ "epoch": 60.5589519650655,
+ "grad_norm": 0.016109555959701538,
+ "learning_rate": 0.0006,
+ "loss": 4.461043357849121,
+ "step": 4360
+ },
+ {
+ "epoch": 60.57292576419214,
+ "grad_norm": 0.01494457945227623,
+ "learning_rate": 0.0006,
+ "loss": 4.505096912384033,
+ "step": 4361
+ },
+ {
+ "epoch": 60.58689956331878,
+ "grad_norm": 0.01729399710893631,
+ "learning_rate": 0.0006,
+ "loss": 4.366747856140137,
+ "step": 4362
+ },
+ {
+ "epoch": 60.60087336244541,
+ "grad_norm": 0.01705821603536606,
+ "learning_rate": 0.0006,
+ "loss": 4.42534065246582,
+ "step": 4363
+ },
+ {
+ "epoch": 60.61484716157205,
+ "grad_norm": 0.018391354009509087,
+ "learning_rate": 0.0006,
+ "loss": 4.350294589996338,
+ "step": 4364
+ },
+ {
+ "epoch": 60.62882096069869,
+ "grad_norm": 0.016989829018712044,
+ "learning_rate": 0.0006,
+ "loss": 4.45905876159668,
+ "step": 4365
+ },
+ {
+ "epoch": 60.64279475982533,
+ "grad_norm": 0.017192212864756584,
+ "learning_rate": 0.0006,
+ "loss": 4.526487350463867,
+ "step": 4366
+ },
+ {
+ "epoch": 60.65676855895197,
+ "grad_norm": 0.018091315403580666,
+ "learning_rate": 0.0006,
+ "loss": 4.309333324432373,
+ "step": 4367
+ },
+ {
+ "epoch": 60.670742358078606,
+ "grad_norm": 0.014051459729671478,
+ "learning_rate": 0.0006,
+ "loss": 4.338171005249023,
+ "step": 4368
+ },
+ {
+ "epoch": 60.68471615720524,
+ "grad_norm": 0.013166418299078941,
+ "learning_rate": 0.0006,
+ "loss": 4.409758567810059,
+ "step": 4369
+ },
+ {
+ "epoch": 60.698689956331876,
+ "grad_norm": 0.014260428957641125,
+ "learning_rate": 0.0006,
+ "loss": 4.524338245391846,
+ "step": 4370
+ },
+ {
+ "epoch": 60.712663755458514,
+ "grad_norm": 0.017019255086779594,
+ "learning_rate": 0.0006,
+ "loss": 4.387436866760254,
+ "step": 4371
+ },
+ {
+ "epoch": 60.72663755458515,
+ "grad_norm": 0.017070675268769264,
+ "learning_rate": 0.0006,
+ "loss": 4.437321662902832,
+ "step": 4372
+ },
+ {
+ "epoch": 60.74061135371179,
+ "grad_norm": 0.015197164379060268,
+ "learning_rate": 0.0006,
+ "loss": 4.371205806732178,
+ "step": 4373
+ },
+ {
+ "epoch": 60.75458515283843,
+ "grad_norm": 0.014791185967624187,
+ "learning_rate": 0.0006,
+ "loss": 4.44439697265625,
+ "step": 4374
+ },
+ {
+ "epoch": 60.76855895196506,
+ "grad_norm": 0.016458261758089066,
+ "learning_rate": 0.0006,
+ "loss": 4.445584297180176,
+ "step": 4375
+ },
+ {
+ "epoch": 60.7825327510917,
+ "grad_norm": 0.017706342041492462,
+ "learning_rate": 0.0006,
+ "loss": 4.348708152770996,
+ "step": 4376
+ },
+ {
+ "epoch": 60.79650655021834,
+ "grad_norm": 0.017561476677656174,
+ "learning_rate": 0.0006,
+ "loss": 4.521929740905762,
+ "step": 4377
+ },
+ {
+ "epoch": 60.81048034934498,
+ "grad_norm": 0.016557445749640465,
+ "learning_rate": 0.0006,
+ "loss": 4.545806884765625,
+ "step": 4378
+ },
+ {
+ "epoch": 60.82445414847162,
+ "grad_norm": 0.016953222453594208,
+ "learning_rate": 0.0006,
+ "loss": 4.4230875968933105,
+ "step": 4379
+ },
+ {
+ "epoch": 60.838427947598255,
+ "grad_norm": 0.017094967886805534,
+ "learning_rate": 0.0006,
+ "loss": 4.423985481262207,
+ "step": 4380
+ },
+ {
+ "epoch": 60.852401746724894,
+ "grad_norm": 0.014787515625357628,
+ "learning_rate": 0.0006,
+ "loss": 4.39473819732666,
+ "step": 4381
+ },
+ {
+ "epoch": 60.866375545851525,
+ "grad_norm": 0.016208263114094734,
+ "learning_rate": 0.0006,
+ "loss": 4.3810038566589355,
+ "step": 4382
+ },
+ {
+ "epoch": 60.880349344978164,
+ "grad_norm": 0.017844321206212044,
+ "learning_rate": 0.0006,
+ "loss": 4.35086727142334,
+ "step": 4383
+ },
+ {
+ "epoch": 60.8943231441048,
+ "grad_norm": 0.017539294436573982,
+ "learning_rate": 0.0006,
+ "loss": 4.460260391235352,
+ "step": 4384
+ },
+ {
+ "epoch": 60.90829694323144,
+ "grad_norm": 0.01612042263150215,
+ "learning_rate": 0.0006,
+ "loss": 4.370500564575195,
+ "step": 4385
+ },
+ {
+ "epoch": 60.92227074235808,
+ "grad_norm": 0.015381601639091969,
+ "learning_rate": 0.0006,
+ "loss": 4.435868263244629,
+ "step": 4386
+ },
+ {
+ "epoch": 60.93624454148472,
+ "grad_norm": 0.01603585295379162,
+ "learning_rate": 0.0006,
+ "loss": 4.434070587158203,
+ "step": 4387
+ },
+ {
+ "epoch": 60.95021834061135,
+ "grad_norm": 0.016268130391836166,
+ "learning_rate": 0.0006,
+ "loss": 4.404343128204346,
+ "step": 4388
+ },
+ {
+ "epoch": 60.96419213973799,
+ "grad_norm": 0.014980032108724117,
+ "learning_rate": 0.0006,
+ "loss": 4.399213790893555,
+ "step": 4389
+ },
+ {
+ "epoch": 60.97816593886463,
+ "grad_norm": 0.016899267211556435,
+ "learning_rate": 0.0006,
+ "loss": 4.508069038391113,
+ "step": 4390
+ },
+ {
+ "epoch": 60.992139737991266,
+ "grad_norm": 0.013756908476352692,
+ "learning_rate": 0.0006,
+ "loss": 4.541858673095703,
+ "step": 4391
+ },
+ {
+ "epoch": 61.0,
+ "grad_norm": 0.01578413136303425,
+ "learning_rate": 0.0006,
+ "loss": 4.543513298034668,
+ "step": 4392
+ },
+ {
+ "epoch": 61.0,
+ "eval_loss": 4.756770610809326,
+ "eval_runtime": 56.4572,
+ "eval_samples_per_second": 43.254,
+ "eval_steps_per_second": 1.364,
+ "step": 4392
+ },
+ {
+ "epoch": 61.01397379912664,
+ "grad_norm": 0.013387506827712059,
+ "learning_rate": 0.0006,
+ "loss": 4.475942611694336,
+ "step": 4393
+ },
+ {
+ "epoch": 61.02794759825328,
+ "grad_norm": 0.014278904534876347,
+ "learning_rate": 0.0006,
+ "loss": 4.442403793334961,
+ "step": 4394
+ },
+ {
+ "epoch": 61.041921397379916,
+ "grad_norm": 0.014359553344547749,
+ "learning_rate": 0.0006,
+ "loss": 4.509463310241699,
+ "step": 4395
+ },
+ {
+ "epoch": 61.05589519650655,
+ "grad_norm": 0.01718050241470337,
+ "learning_rate": 0.0006,
+ "loss": 4.354458808898926,
+ "step": 4396
+ },
+ {
+ "epoch": 61.069868995633186,
+ "grad_norm": 0.015988312661647797,
+ "learning_rate": 0.0006,
+ "loss": 4.484508514404297,
+ "step": 4397
+ },
+ {
+ "epoch": 61.083842794759825,
+ "grad_norm": 0.015466735698282719,
+ "learning_rate": 0.0006,
+ "loss": 4.326908111572266,
+ "step": 4398
+ },
+ {
+ "epoch": 61.09781659388646,
+ "grad_norm": 0.014541332609951496,
+ "learning_rate": 0.0006,
+ "loss": 4.405823230743408,
+ "step": 4399
+ },
+ {
+ "epoch": 61.1117903930131,
+ "grad_norm": 0.014803987927734852,
+ "learning_rate": 0.0006,
+ "loss": 4.375386714935303,
+ "step": 4400
+ },
+ {
+ "epoch": 61.12576419213974,
+ "grad_norm": 0.013962093740701675,
+ "learning_rate": 0.0006,
+ "loss": 4.300838470458984,
+ "step": 4401
+ },
+ {
+ "epoch": 61.13973799126637,
+ "grad_norm": 0.015213954262435436,
+ "learning_rate": 0.0006,
+ "loss": 4.388306617736816,
+ "step": 4402
+ },
+ {
+ "epoch": 61.15371179039301,
+ "grad_norm": 0.018965506926178932,
+ "learning_rate": 0.0006,
+ "loss": 4.388340950012207,
+ "step": 4403
+ },
+ {
+ "epoch": 61.16768558951965,
+ "grad_norm": 0.01958410255610943,
+ "learning_rate": 0.0006,
+ "loss": 4.35012149810791,
+ "step": 4404
+ },
+ {
+ "epoch": 61.18165938864629,
+ "grad_norm": 0.01938101276755333,
+ "learning_rate": 0.0006,
+ "loss": 4.368372917175293,
+ "step": 4405
+ },
+ {
+ "epoch": 61.19563318777293,
+ "grad_norm": 0.016556516289711,
+ "learning_rate": 0.0006,
+ "loss": 4.486198425292969,
+ "step": 4406
+ },
+ {
+ "epoch": 61.209606986899566,
+ "grad_norm": 0.017031289637088776,
+ "learning_rate": 0.0006,
+ "loss": 4.421676158905029,
+ "step": 4407
+ },
+ {
+ "epoch": 61.223580786026204,
+ "grad_norm": 0.018023479729890823,
+ "learning_rate": 0.0006,
+ "loss": 4.38751220703125,
+ "step": 4408
+ },
+ {
+ "epoch": 61.237554585152836,
+ "grad_norm": 0.0178135447204113,
+ "learning_rate": 0.0006,
+ "loss": 4.336528301239014,
+ "step": 4409
+ },
+ {
+ "epoch": 61.251528384279474,
+ "grad_norm": 0.016782190650701523,
+ "learning_rate": 0.0006,
+ "loss": 4.386693954467773,
+ "step": 4410
+ },
+ {
+ "epoch": 61.26550218340611,
+ "grad_norm": 0.015494848601520061,
+ "learning_rate": 0.0006,
+ "loss": 4.308948040008545,
+ "step": 4411
+ },
+ {
+ "epoch": 61.27947598253275,
+ "grad_norm": 0.015002378262579441,
+ "learning_rate": 0.0006,
+ "loss": 4.314206123352051,
+ "step": 4412
+ },
+ {
+ "epoch": 61.29344978165939,
+ "grad_norm": 0.015798920765519142,
+ "learning_rate": 0.0006,
+ "loss": 4.502900123596191,
+ "step": 4413
+ },
+ {
+ "epoch": 61.30742358078603,
+ "grad_norm": 0.015444336459040642,
+ "learning_rate": 0.0006,
+ "loss": 4.473160743713379,
+ "step": 4414
+ },
+ {
+ "epoch": 61.32139737991266,
+ "grad_norm": 0.017025096341967583,
+ "learning_rate": 0.0006,
+ "loss": 4.460621356964111,
+ "step": 4415
+ },
+ {
+ "epoch": 61.3353711790393,
+ "grad_norm": 0.019502537325024605,
+ "learning_rate": 0.0006,
+ "loss": 4.385551452636719,
+ "step": 4416
+ },
+ {
+ "epoch": 61.34934497816594,
+ "grad_norm": 0.018529541790485382,
+ "learning_rate": 0.0006,
+ "loss": 4.425390720367432,
+ "step": 4417
+ },
+ {
+ "epoch": 61.36331877729258,
+ "grad_norm": 0.013155706226825714,
+ "learning_rate": 0.0006,
+ "loss": 4.467020034790039,
+ "step": 4418
+ },
+ {
+ "epoch": 61.377292576419215,
+ "grad_norm": 0.01518090907484293,
+ "learning_rate": 0.0006,
+ "loss": 4.369664192199707,
+ "step": 4419
+ },
+ {
+ "epoch": 61.391266375545854,
+ "grad_norm": 0.01626511849462986,
+ "learning_rate": 0.0006,
+ "loss": 4.463231086730957,
+ "step": 4420
+ },
+ {
+ "epoch": 61.40524017467249,
+ "grad_norm": 0.01701558195054531,
+ "learning_rate": 0.0006,
+ "loss": 4.43708610534668,
+ "step": 4421
+ },
+ {
+ "epoch": 61.419213973799124,
+ "grad_norm": 0.01761862449347973,
+ "learning_rate": 0.0006,
+ "loss": 4.3314208984375,
+ "step": 4422
+ },
+ {
+ "epoch": 61.43318777292576,
+ "grad_norm": 0.017084648832678795,
+ "learning_rate": 0.0006,
+ "loss": 4.39676570892334,
+ "step": 4423
+ },
+ {
+ "epoch": 61.4471615720524,
+ "grad_norm": 0.01667078025639057,
+ "learning_rate": 0.0006,
+ "loss": 4.3789873123168945,
+ "step": 4424
+ },
+ {
+ "epoch": 61.46113537117904,
+ "grad_norm": 0.01877436414361,
+ "learning_rate": 0.0006,
+ "loss": 4.415275573730469,
+ "step": 4425
+ },
+ {
+ "epoch": 61.47510917030568,
+ "grad_norm": 0.020250486209988594,
+ "learning_rate": 0.0006,
+ "loss": 4.478635787963867,
+ "step": 4426
+ },
+ {
+ "epoch": 61.48908296943232,
+ "grad_norm": 0.020244868472218513,
+ "learning_rate": 0.0006,
+ "loss": 4.296854019165039,
+ "step": 4427
+ },
+ {
+ "epoch": 61.50305676855895,
+ "grad_norm": 0.018723875284194946,
+ "learning_rate": 0.0006,
+ "loss": 4.3760881423950195,
+ "step": 4428
+ },
+ {
+ "epoch": 61.51703056768559,
+ "grad_norm": 0.015138164162635803,
+ "learning_rate": 0.0006,
+ "loss": 4.39974308013916,
+ "step": 4429
+ },
+ {
+ "epoch": 61.531004366812226,
+ "grad_norm": 0.015111138112843037,
+ "learning_rate": 0.0006,
+ "loss": 4.34659481048584,
+ "step": 4430
+ },
+ {
+ "epoch": 61.544978165938865,
+ "grad_norm": 0.0164666585624218,
+ "learning_rate": 0.0006,
+ "loss": 4.415737152099609,
+ "step": 4431
+ },
+ {
+ "epoch": 61.5589519650655,
+ "grad_norm": 0.020891312509775162,
+ "learning_rate": 0.0006,
+ "loss": 4.300319671630859,
+ "step": 4432
+ },
+ {
+ "epoch": 61.57292576419214,
+ "grad_norm": 0.02089156024158001,
+ "learning_rate": 0.0006,
+ "loss": 4.373029708862305,
+ "step": 4433
+ },
+ {
+ "epoch": 61.58689956331878,
+ "grad_norm": 0.020114559680223465,
+ "learning_rate": 0.0006,
+ "loss": 4.434499740600586,
+ "step": 4434
+ },
+ {
+ "epoch": 61.60087336244541,
+ "grad_norm": 0.01791365258395672,
+ "learning_rate": 0.0006,
+ "loss": 4.466872215270996,
+ "step": 4435
+ },
+ {
+ "epoch": 61.61484716157205,
+ "grad_norm": 0.01879458874464035,
+ "learning_rate": 0.0006,
+ "loss": 4.5176520347595215,
+ "step": 4436
+ },
+ {
+ "epoch": 61.62882096069869,
+ "grad_norm": 0.016371937468647957,
+ "learning_rate": 0.0006,
+ "loss": 4.3507585525512695,
+ "step": 4437
+ },
+ {
+ "epoch": 61.64279475982533,
+ "grad_norm": 0.01648143120110035,
+ "learning_rate": 0.0006,
+ "loss": 4.398778438568115,
+ "step": 4438
+ },
+ {
+ "epoch": 61.65676855895197,
+ "grad_norm": 0.017240576446056366,
+ "learning_rate": 0.0006,
+ "loss": 4.513933181762695,
+ "step": 4439
+ },
+ {
+ "epoch": 61.670742358078606,
+ "grad_norm": 0.01734284684062004,
+ "learning_rate": 0.0006,
+ "loss": 4.315343856811523,
+ "step": 4440
+ },
+ {
+ "epoch": 61.68471615720524,
+ "grad_norm": 0.01639772206544876,
+ "learning_rate": 0.0006,
+ "loss": 4.4641218185424805,
+ "step": 4441
+ },
+ {
+ "epoch": 61.698689956331876,
+ "grad_norm": 0.0164708960801363,
+ "learning_rate": 0.0006,
+ "loss": 4.439968585968018,
+ "step": 4442
+ },
+ {
+ "epoch": 61.712663755458514,
+ "grad_norm": 0.01757933758199215,
+ "learning_rate": 0.0006,
+ "loss": 4.408327102661133,
+ "step": 4443
+ },
+ {
+ "epoch": 61.72663755458515,
+ "grad_norm": 0.01789589412510395,
+ "learning_rate": 0.0006,
+ "loss": 4.478257179260254,
+ "step": 4444
+ },
+ {
+ "epoch": 61.74061135371179,
+ "grad_norm": 0.01822184957563877,
+ "learning_rate": 0.0006,
+ "loss": 4.412046909332275,
+ "step": 4445
+ },
+ {
+ "epoch": 61.75458515283843,
+ "grad_norm": 0.018624618649482727,
+ "learning_rate": 0.0006,
+ "loss": 4.427122116088867,
+ "step": 4446
+ },
+ {
+ "epoch": 61.76855895196506,
+ "grad_norm": 0.018674887716770172,
+ "learning_rate": 0.0006,
+ "loss": 4.429147720336914,
+ "step": 4447
+ },
+ {
+ "epoch": 61.7825327510917,
+ "grad_norm": 0.015317119657993317,
+ "learning_rate": 0.0006,
+ "loss": 4.2808451652526855,
+ "step": 4448
+ },
+ {
+ "epoch": 61.79650655021834,
+ "grad_norm": 0.015400845557451248,
+ "learning_rate": 0.0006,
+ "loss": 4.420193672180176,
+ "step": 4449
+ },
+ {
+ "epoch": 61.81048034934498,
+ "grad_norm": 0.016743650659918785,
+ "learning_rate": 0.0006,
+ "loss": 4.498254776000977,
+ "step": 4450
+ },
+ {
+ "epoch": 61.82445414847162,
+ "grad_norm": 0.014330781064927578,
+ "learning_rate": 0.0006,
+ "loss": 4.445384979248047,
+ "step": 4451
+ },
+ {
+ "epoch": 61.838427947598255,
+ "grad_norm": 0.015046479180455208,
+ "learning_rate": 0.0006,
+ "loss": 4.368041038513184,
+ "step": 4452
+ },
+ {
+ "epoch": 61.852401746724894,
+ "grad_norm": 0.01630168780684471,
+ "learning_rate": 0.0006,
+ "loss": 4.460142135620117,
+ "step": 4453
+ },
+ {
+ "epoch": 61.866375545851525,
+ "grad_norm": 0.01637885719537735,
+ "learning_rate": 0.0006,
+ "loss": 4.416606426239014,
+ "step": 4454
+ },
+ {
+ "epoch": 61.880349344978164,
+ "grad_norm": 0.016785888001322746,
+ "learning_rate": 0.0006,
+ "loss": 4.452545642852783,
+ "step": 4455
+ },
+ {
+ "epoch": 61.8943231441048,
+ "grad_norm": 0.015044555068016052,
+ "learning_rate": 0.0006,
+ "loss": 4.427334785461426,
+ "step": 4456
+ },
+ {
+ "epoch": 61.90829694323144,
+ "grad_norm": 0.013699211180210114,
+ "learning_rate": 0.0006,
+ "loss": 4.408761024475098,
+ "step": 4457
+ },
+ {
+ "epoch": 61.92227074235808,
+ "grad_norm": 0.014021803624927998,
+ "learning_rate": 0.0006,
+ "loss": 4.543607711791992,
+ "step": 4458
+ },
+ {
+ "epoch": 61.93624454148472,
+ "grad_norm": 0.015438460744917393,
+ "learning_rate": 0.0006,
+ "loss": 4.313363075256348,
+ "step": 4459
+ },
+ {
+ "epoch": 61.95021834061135,
+ "grad_norm": 0.017515428364276886,
+ "learning_rate": 0.0006,
+ "loss": 4.310412406921387,
+ "step": 4460
+ },
+ {
+ "epoch": 61.96419213973799,
+ "grad_norm": 0.016683880239725113,
+ "learning_rate": 0.0006,
+ "loss": 4.3378448486328125,
+ "step": 4461
+ },
+ {
+ "epoch": 61.97816593886463,
+ "grad_norm": 0.019384048879146576,
+ "learning_rate": 0.0006,
+ "loss": 4.450568675994873,
+ "step": 4462
+ },
+ {
+ "epoch": 61.992139737991266,
+ "grad_norm": 0.020651493221521378,
+ "learning_rate": 0.0006,
+ "loss": 4.453197479248047,
+ "step": 4463
+ },
+ {
+ "epoch": 62.0,
+ "grad_norm": 0.020991764962673187,
+ "learning_rate": 0.0006,
+ "loss": 4.441527366638184,
+ "step": 4464
+ },
+ {
+ "epoch": 62.0,
+ "eval_loss": 4.700193881988525,
+ "eval_runtime": 56.6098,
+ "eval_samples_per_second": 43.137,
+ "eval_steps_per_second": 1.36,
+ "step": 4464
+ },
+ {
+ "epoch": 62.01397379912664,
+ "grad_norm": 0.01811373233795166,
+ "learning_rate": 0.0006,
+ "loss": 4.37346076965332,
+ "step": 4465
+ },
+ {
+ "epoch": 62.02794759825328,
+ "grad_norm": 0.01667127199470997,
+ "learning_rate": 0.0006,
+ "loss": 4.237151145935059,
+ "step": 4466
+ },
+ {
+ "epoch": 62.041921397379916,
+ "grad_norm": 0.014971534721553326,
+ "learning_rate": 0.0006,
+ "loss": 4.367053031921387,
+ "step": 4467
+ },
+ {
+ "epoch": 62.05589519650655,
+ "grad_norm": 0.015655893832445145,
+ "learning_rate": 0.0006,
+ "loss": 4.437426567077637,
+ "step": 4468
+ },
+ {
+ "epoch": 62.069868995633186,
+ "grad_norm": 0.014935447834432125,
+ "learning_rate": 0.0006,
+ "loss": 4.3635125160217285,
+ "step": 4469
+ },
+ {
+ "epoch": 62.083842794759825,
+ "grad_norm": 0.015363802202045918,
+ "learning_rate": 0.0006,
+ "loss": 4.40492057800293,
+ "step": 4470
+ },
+ {
+ "epoch": 62.09781659388646,
+ "grad_norm": 0.0159201230853796,
+ "learning_rate": 0.0006,
+ "loss": 4.392058849334717,
+ "step": 4471
+ },
+ {
+ "epoch": 62.1117903930131,
+ "grad_norm": 0.014598241075873375,
+ "learning_rate": 0.0006,
+ "loss": 4.369944095611572,
+ "step": 4472
+ },
+ {
+ "epoch": 62.12576419213974,
+ "grad_norm": 0.015394841320812702,
+ "learning_rate": 0.0006,
+ "loss": 4.425097942352295,
+ "step": 4473
+ },
+ {
+ "epoch": 62.13973799126637,
+ "grad_norm": 0.015913628041744232,
+ "learning_rate": 0.0006,
+ "loss": 4.420596122741699,
+ "step": 4474
+ },
+ {
+ "epoch": 62.15371179039301,
+ "grad_norm": 0.01575258933007717,
+ "learning_rate": 0.0006,
+ "loss": 4.32290506362915,
+ "step": 4475
+ },
+ {
+ "epoch": 62.16768558951965,
+ "grad_norm": 0.0155344782397151,
+ "learning_rate": 0.0006,
+ "loss": 4.273068428039551,
+ "step": 4476
+ },
+ {
+ "epoch": 62.18165938864629,
+ "grad_norm": 0.014947040006518364,
+ "learning_rate": 0.0006,
+ "loss": 4.384868621826172,
+ "step": 4477
+ },
+ {
+ "epoch": 62.19563318777293,
+ "grad_norm": 0.01557591650635004,
+ "learning_rate": 0.0006,
+ "loss": 4.458410263061523,
+ "step": 4478
+ },
+ {
+ "epoch": 62.209606986899566,
+ "grad_norm": 0.015182198956608772,
+ "learning_rate": 0.0006,
+ "loss": 4.300738334655762,
+ "step": 4479
+ },
+ {
+ "epoch": 62.223580786026204,
+ "grad_norm": 0.01634717360138893,
+ "learning_rate": 0.0006,
+ "loss": 4.436875820159912,
+ "step": 4480
+ },
+ {
+ "epoch": 62.237554585152836,
+ "grad_norm": 0.018813718110322952,
+ "learning_rate": 0.0006,
+ "loss": 4.454665184020996,
+ "step": 4481
+ },
+ {
+ "epoch": 62.251528384279474,
+ "grad_norm": 0.020965630188584328,
+ "learning_rate": 0.0006,
+ "loss": 4.357341766357422,
+ "step": 4482
+ },
+ {
+ "epoch": 62.26550218340611,
+ "grad_norm": 0.020738353952765465,
+ "learning_rate": 0.0006,
+ "loss": 4.456644058227539,
+ "step": 4483
+ },
+ {
+ "epoch": 62.27947598253275,
+ "grad_norm": 0.018813546746969223,
+ "learning_rate": 0.0006,
+ "loss": 4.381941795349121,
+ "step": 4484
+ },
+ {
+ "epoch": 62.29344978165939,
+ "grad_norm": 0.015431761741638184,
+ "learning_rate": 0.0006,
+ "loss": 4.41126823425293,
+ "step": 4485
+ },
+ {
+ "epoch": 62.30742358078603,
+ "grad_norm": 0.016335977241396904,
+ "learning_rate": 0.0006,
+ "loss": 4.419488906860352,
+ "step": 4486
+ },
+ {
+ "epoch": 62.32139737991266,
+ "grad_norm": 0.01832406409084797,
+ "learning_rate": 0.0006,
+ "loss": 4.579026222229004,
+ "step": 4487
+ },
+ {
+ "epoch": 62.3353711790393,
+ "grad_norm": 0.01790742017328739,
+ "learning_rate": 0.0006,
+ "loss": 4.289964199066162,
+ "step": 4488
+ },
+ {
+ "epoch": 62.34934497816594,
+ "grad_norm": 0.017266202718019485,
+ "learning_rate": 0.0006,
+ "loss": 4.415959358215332,
+ "step": 4489
+ },
+ {
+ "epoch": 62.36331877729258,
+ "grad_norm": 0.015976132825016975,
+ "learning_rate": 0.0006,
+ "loss": 4.358587265014648,
+ "step": 4490
+ },
+ {
+ "epoch": 62.377292576419215,
+ "grad_norm": 0.016873881220817566,
+ "learning_rate": 0.0006,
+ "loss": 4.354607105255127,
+ "step": 4491
+ },
+ {
+ "epoch": 62.391266375545854,
+ "grad_norm": 0.01476745493710041,
+ "learning_rate": 0.0006,
+ "loss": 4.384904384613037,
+ "step": 4492
+ },
+ {
+ "epoch": 62.40524017467249,
+ "grad_norm": 0.013810782693326473,
+ "learning_rate": 0.0006,
+ "loss": 4.459230899810791,
+ "step": 4493
+ },
+ {
+ "epoch": 62.419213973799124,
+ "grad_norm": 0.015150421299040318,
+ "learning_rate": 0.0006,
+ "loss": 4.369471549987793,
+ "step": 4494
+ },
+ {
+ "epoch": 62.43318777292576,
+ "grad_norm": 0.014539610594511032,
+ "learning_rate": 0.0006,
+ "loss": 4.3564982414245605,
+ "step": 4495
+ },
+ {
+ "epoch": 62.4471615720524,
+ "grad_norm": 0.015554017387330532,
+ "learning_rate": 0.0006,
+ "loss": 4.414379119873047,
+ "step": 4496
+ },
+ {
+ "epoch": 62.46113537117904,
+ "grad_norm": 0.016608357429504395,
+ "learning_rate": 0.0006,
+ "loss": 4.311585426330566,
+ "step": 4497
+ },
+ {
+ "epoch": 62.47510917030568,
+ "grad_norm": 0.01885881833732128,
+ "learning_rate": 0.0006,
+ "loss": 4.306629180908203,
+ "step": 4498
+ },
+ {
+ "epoch": 62.48908296943232,
+ "grad_norm": 0.022128663957118988,
+ "learning_rate": 0.0006,
+ "loss": 4.345483779907227,
+ "step": 4499
+ },
+ {
+ "epoch": 62.50305676855895,
+ "grad_norm": 0.023401018232107162,
+ "learning_rate": 0.0006,
+ "loss": 4.398204326629639,
+ "step": 4500
+ },
+ {
+ "epoch": 62.51703056768559,
+ "grad_norm": 0.022080009803175926,
+ "learning_rate": 0.0006,
+ "loss": 4.375027179718018,
+ "step": 4501
+ },
+ {
+ "epoch": 62.531004366812226,
+ "grad_norm": 0.020004864782094955,
+ "learning_rate": 0.0006,
+ "loss": 4.33494758605957,
+ "step": 4502
+ },
+ {
+ "epoch": 62.544978165938865,
+ "grad_norm": 0.01786700263619423,
+ "learning_rate": 0.0006,
+ "loss": 4.4428300857543945,
+ "step": 4503
+ },
+ {
+ "epoch": 62.5589519650655,
+ "grad_norm": 0.01885388419032097,
+ "learning_rate": 0.0006,
+ "loss": 4.4138946533203125,
+ "step": 4504
+ },
+ {
+ "epoch": 62.57292576419214,
+ "grad_norm": 0.019890939816832542,
+ "learning_rate": 0.0006,
+ "loss": 4.278632164001465,
+ "step": 4505
+ },
+ {
+ "epoch": 62.58689956331878,
+ "grad_norm": 0.02109031192958355,
+ "learning_rate": 0.0006,
+ "loss": 4.3756327629089355,
+ "step": 4506
+ },
+ {
+ "epoch": 62.60087336244541,
+ "grad_norm": 0.020813899114727974,
+ "learning_rate": 0.0006,
+ "loss": 4.410213470458984,
+ "step": 4507
+ },
+ {
+ "epoch": 62.61484716157205,
+ "grad_norm": 0.021655689924955368,
+ "learning_rate": 0.0006,
+ "loss": 4.30633020401001,
+ "step": 4508
+ },
+ {
+ "epoch": 62.62882096069869,
+ "grad_norm": 0.018155649304389954,
+ "learning_rate": 0.0006,
+ "loss": 4.407284736633301,
+ "step": 4509
+ },
+ {
+ "epoch": 62.64279475982533,
+ "grad_norm": 0.017765069380402565,
+ "learning_rate": 0.0006,
+ "loss": 4.321715354919434,
+ "step": 4510
+ },
+ {
+ "epoch": 62.65676855895197,
+ "grad_norm": 0.017724554985761642,
+ "learning_rate": 0.0006,
+ "loss": 4.493371963500977,
+ "step": 4511
+ },
+ {
+ "epoch": 62.670742358078606,
+ "grad_norm": 0.01492092851549387,
+ "learning_rate": 0.0006,
+ "loss": 4.443025588989258,
+ "step": 4512
+ },
+ {
+ "epoch": 62.68471615720524,
+ "grad_norm": 0.01615324057638645,
+ "learning_rate": 0.0006,
+ "loss": 4.3289594650268555,
+ "step": 4513
+ },
+ {
+ "epoch": 62.698689956331876,
+ "grad_norm": 0.016569366678595543,
+ "learning_rate": 0.0006,
+ "loss": 4.369720458984375,
+ "step": 4514
+ },
+ {
+ "epoch": 62.712663755458514,
+ "grad_norm": 0.016509901732206345,
+ "learning_rate": 0.0006,
+ "loss": 4.292316436767578,
+ "step": 4515
+ },
+ {
+ "epoch": 62.72663755458515,
+ "grad_norm": 0.014270003885030746,
+ "learning_rate": 0.0006,
+ "loss": 4.349854469299316,
+ "step": 4516
+ },
+ {
+ "epoch": 62.74061135371179,
+ "grad_norm": 0.014235340990126133,
+ "learning_rate": 0.0006,
+ "loss": 4.473727226257324,
+ "step": 4517
+ },
+ {
+ "epoch": 62.75458515283843,
+ "grad_norm": 0.015163225121796131,
+ "learning_rate": 0.0006,
+ "loss": 4.452878475189209,
+ "step": 4518
+ },
+ {
+ "epoch": 62.76855895196506,
+ "grad_norm": 0.015748055651783943,
+ "learning_rate": 0.0006,
+ "loss": 4.521087169647217,
+ "step": 4519
+ },
+ {
+ "epoch": 62.7825327510917,
+ "grad_norm": 0.01661108434200287,
+ "learning_rate": 0.0006,
+ "loss": 4.443885326385498,
+ "step": 4520
+ },
+ {
+ "epoch": 62.79650655021834,
+ "grad_norm": 0.01610301062464714,
+ "learning_rate": 0.0006,
+ "loss": 4.312990188598633,
+ "step": 4521
+ },
+ {
+ "epoch": 62.81048034934498,
+ "grad_norm": 0.017007583752274513,
+ "learning_rate": 0.0006,
+ "loss": 4.408937454223633,
+ "step": 4522
+ },
+ {
+ "epoch": 62.82445414847162,
+ "grad_norm": 0.01749461516737938,
+ "learning_rate": 0.0006,
+ "loss": 4.346066951751709,
+ "step": 4523
+ },
+ {
+ "epoch": 62.838427947598255,
+ "grad_norm": 0.018243493512272835,
+ "learning_rate": 0.0006,
+ "loss": 4.357658386230469,
+ "step": 4524
+ },
+ {
+ "epoch": 62.852401746724894,
+ "grad_norm": 0.017499985173344612,
+ "learning_rate": 0.0006,
+ "loss": 4.374061584472656,
+ "step": 4525
+ },
+ {
+ "epoch": 62.866375545851525,
+ "grad_norm": 0.014536325819790363,
+ "learning_rate": 0.0006,
+ "loss": 4.189321994781494,
+ "step": 4526
+ },
+ {
+ "epoch": 62.880349344978164,
+ "grad_norm": 0.014728891663253307,
+ "learning_rate": 0.0006,
+ "loss": 4.425910949707031,
+ "step": 4527
+ },
+ {
+ "epoch": 62.8943231441048,
+ "grad_norm": 0.015013772062957287,
+ "learning_rate": 0.0006,
+ "loss": 4.337319850921631,
+ "step": 4528
+ },
+ {
+ "epoch": 62.90829694323144,
+ "grad_norm": 0.01626027375459671,
+ "learning_rate": 0.0006,
+ "loss": 4.330339431762695,
+ "step": 4529
+ },
+ {
+ "epoch": 62.92227074235808,
+ "grad_norm": 0.014773444272577763,
+ "learning_rate": 0.0006,
+ "loss": 4.4188737869262695,
+ "step": 4530
+ },
+ {
+ "epoch": 62.93624454148472,
+ "grad_norm": 0.01422932744026184,
+ "learning_rate": 0.0006,
+ "loss": 4.355986595153809,
+ "step": 4531
+ },
+ {
+ "epoch": 62.95021834061135,
+ "grad_norm": 0.012170110829174519,
+ "learning_rate": 0.0006,
+ "loss": 4.405722618103027,
+ "step": 4532
+ },
+ {
+ "epoch": 62.96419213973799,
+ "grad_norm": 0.013233253732323647,
+ "learning_rate": 0.0006,
+ "loss": 4.377286911010742,
+ "step": 4533
+ },
+ {
+ "epoch": 62.97816593886463,
+ "grad_norm": 0.014497299678623676,
+ "learning_rate": 0.0006,
+ "loss": 4.508796691894531,
+ "step": 4534
+ },
+ {
+ "epoch": 62.992139737991266,
+ "grad_norm": 0.014718741178512573,
+ "learning_rate": 0.0006,
+ "loss": 4.370919704437256,
+ "step": 4535
+ },
+ {
+ "epoch": 63.0,
+ "grad_norm": 0.015833502635359764,
+ "learning_rate": 0.0006,
+ "loss": 4.258753776550293,
+ "step": 4536
+ },
+ {
+ "epoch": 63.0,
+ "eval_loss": 4.714163303375244,
+ "eval_runtime": 57.0269,
+ "eval_samples_per_second": 42.822,
+ "eval_steps_per_second": 1.35,
+ "step": 4536
+ },
+ {
+ "epoch": 63.01397379912664,
+ "grad_norm": 0.014277510344982147,
+ "learning_rate": 0.0006,
+ "loss": 4.3112640380859375,
+ "step": 4537
+ },
+ {
+ "epoch": 63.02794759825328,
+ "grad_norm": 0.015377935022115707,
+ "learning_rate": 0.0006,
+ "loss": 4.342782974243164,
+ "step": 4538
+ },
+ {
+ "epoch": 63.041921397379916,
+ "grad_norm": 0.015406244434416294,
+ "learning_rate": 0.0006,
+ "loss": 4.474056720733643,
+ "step": 4539
+ },
+ {
+ "epoch": 63.05589519650655,
+ "grad_norm": 0.018121499568223953,
+ "learning_rate": 0.0006,
+ "loss": 4.397963523864746,
+ "step": 4540
+ },
+ {
+ "epoch": 63.069868995633186,
+ "grad_norm": 0.022634636610746384,
+ "learning_rate": 0.0006,
+ "loss": 4.306631088256836,
+ "step": 4541
+ },
+ {
+ "epoch": 63.083842794759825,
+ "grad_norm": 0.026448970660567284,
+ "learning_rate": 0.0006,
+ "loss": 4.283175468444824,
+ "step": 4542
+ },
+ {
+ "epoch": 63.09781659388646,
+ "grad_norm": 0.024104537442326546,
+ "learning_rate": 0.0006,
+ "loss": 4.315826416015625,
+ "step": 4543
+ },
+ {
+ "epoch": 63.1117903930131,
+ "grad_norm": 0.019040480256080627,
+ "learning_rate": 0.0006,
+ "loss": 4.44057035446167,
+ "step": 4544
+ },
+ {
+ "epoch": 63.12576419213974,
+ "grad_norm": 0.02006000280380249,
+ "learning_rate": 0.0006,
+ "loss": 4.405603408813477,
+ "step": 4545
+ },
+ {
+ "epoch": 63.13973799126637,
+ "grad_norm": 0.019980071112513542,
+ "learning_rate": 0.0006,
+ "loss": 4.393726348876953,
+ "step": 4546
+ },
+ {
+ "epoch": 63.15371179039301,
+ "grad_norm": 0.01641865260899067,
+ "learning_rate": 0.0006,
+ "loss": 4.272546768188477,
+ "step": 4547
+ },
+ {
+ "epoch": 63.16768558951965,
+ "grad_norm": 0.01890353485941887,
+ "learning_rate": 0.0006,
+ "loss": 4.382566928863525,
+ "step": 4548
+ },
+ {
+ "epoch": 63.18165938864629,
+ "grad_norm": 0.018370937556028366,
+ "learning_rate": 0.0006,
+ "loss": 4.423953533172607,
+ "step": 4549
+ },
+ {
+ "epoch": 63.19563318777293,
+ "grad_norm": 0.014985294081270695,
+ "learning_rate": 0.0006,
+ "loss": 4.308627128601074,
+ "step": 4550
+ },
+ {
+ "epoch": 63.209606986899566,
+ "grad_norm": 0.017513327300548553,
+ "learning_rate": 0.0006,
+ "loss": 4.248819351196289,
+ "step": 4551
+ },
+ {
+ "epoch": 63.223580786026204,
+ "grad_norm": 0.019464161247015,
+ "learning_rate": 0.0006,
+ "loss": 4.359729766845703,
+ "step": 4552
+ },
+ {
+ "epoch": 63.237554585152836,
+ "grad_norm": 0.018678035587072372,
+ "learning_rate": 0.0006,
+ "loss": 4.3299665451049805,
+ "step": 4553
+ },
+ {
+ "epoch": 63.251528384279474,
+ "grad_norm": 0.017909523099660873,
+ "learning_rate": 0.0006,
+ "loss": 4.246147155761719,
+ "step": 4554
+ },
+ {
+ "epoch": 63.26550218340611,
+ "grad_norm": 0.01747293211519718,
+ "learning_rate": 0.0006,
+ "loss": 4.41165018081665,
+ "step": 4555
+ },
+ {
+ "epoch": 63.27947598253275,
+ "grad_norm": 0.018670853227376938,
+ "learning_rate": 0.0006,
+ "loss": 4.409883499145508,
+ "step": 4556
+ },
+ {
+ "epoch": 63.29344978165939,
+ "grad_norm": 0.0189529862254858,
+ "learning_rate": 0.0006,
+ "loss": 4.418654918670654,
+ "step": 4557
+ },
+ {
+ "epoch": 63.30742358078603,
+ "grad_norm": 0.018394414335489273,
+ "learning_rate": 0.0006,
+ "loss": 4.38355827331543,
+ "step": 4558
+ },
+ {
+ "epoch": 63.32139737991266,
+ "grad_norm": 0.018317820504307747,
+ "learning_rate": 0.0006,
+ "loss": 4.456843376159668,
+ "step": 4559
+ },
+ {
+ "epoch": 63.3353711790393,
+ "grad_norm": 0.019225597381591797,
+ "learning_rate": 0.0006,
+ "loss": 4.232264518737793,
+ "step": 4560
+ },
+ {
+ "epoch": 63.34934497816594,
+ "grad_norm": 0.016663681715726852,
+ "learning_rate": 0.0006,
+ "loss": 4.393517017364502,
+ "step": 4561
+ },
+ {
+ "epoch": 63.36331877729258,
+ "grad_norm": 0.014975365251302719,
+ "learning_rate": 0.0006,
+ "loss": 4.375183582305908,
+ "step": 4562
+ },
+ {
+ "epoch": 63.377292576419215,
+ "grad_norm": 0.01771732047200203,
+ "learning_rate": 0.0006,
+ "loss": 4.3073410987854,
+ "step": 4563
+ },
+ {
+ "epoch": 63.391266375545854,
+ "grad_norm": 0.01663549244403839,
+ "learning_rate": 0.0006,
+ "loss": 4.311003684997559,
+ "step": 4564
+ },
+ {
+ "epoch": 63.40524017467249,
+ "grad_norm": 0.016301177442073822,
+ "learning_rate": 0.0006,
+ "loss": 4.326807022094727,
+ "step": 4565
+ },
+ {
+ "epoch": 63.419213973799124,
+ "grad_norm": 0.016781188547611237,
+ "learning_rate": 0.0006,
+ "loss": 4.306562423706055,
+ "step": 4566
+ },
+ {
+ "epoch": 63.43318777292576,
+ "grad_norm": 0.01767745055258274,
+ "learning_rate": 0.0006,
+ "loss": 4.367082595825195,
+ "step": 4567
+ },
+ {
+ "epoch": 63.4471615720524,
+ "grad_norm": 0.016728565096855164,
+ "learning_rate": 0.0006,
+ "loss": 4.327400207519531,
+ "step": 4568
+ },
+ {
+ "epoch": 63.46113537117904,
+ "grad_norm": 0.016289403662085533,
+ "learning_rate": 0.0006,
+ "loss": 4.392712116241455,
+ "step": 4569
+ },
+ {
+ "epoch": 63.47510917030568,
+ "grad_norm": 0.016233716160058975,
+ "learning_rate": 0.0006,
+ "loss": 4.354372024536133,
+ "step": 4570
+ },
+ {
+ "epoch": 63.48908296943232,
+ "grad_norm": 0.016547974199056625,
+ "learning_rate": 0.0006,
+ "loss": 4.309050559997559,
+ "step": 4571
+ },
+ {
+ "epoch": 63.50305676855895,
+ "grad_norm": 0.018466509878635406,
+ "learning_rate": 0.0006,
+ "loss": 4.318345069885254,
+ "step": 4572
+ },
+ {
+ "epoch": 63.51703056768559,
+ "grad_norm": 0.017685720697045326,
+ "learning_rate": 0.0006,
+ "loss": 4.399726867675781,
+ "step": 4573
+ },
+ {
+ "epoch": 63.531004366812226,
+ "grad_norm": 0.015801234170794487,
+ "learning_rate": 0.0006,
+ "loss": 4.328495502471924,
+ "step": 4574
+ },
+ {
+ "epoch": 63.544978165938865,
+ "grad_norm": 0.0188873540610075,
+ "learning_rate": 0.0006,
+ "loss": 4.253244400024414,
+ "step": 4575
+ },
+ {
+ "epoch": 63.5589519650655,
+ "grad_norm": 0.01868271268904209,
+ "learning_rate": 0.0006,
+ "loss": 4.463746070861816,
+ "step": 4576
+ },
+ {
+ "epoch": 63.57292576419214,
+ "grad_norm": 0.01844949834048748,
+ "learning_rate": 0.0006,
+ "loss": 4.400300025939941,
+ "step": 4577
+ },
+ {
+ "epoch": 63.58689956331878,
+ "grad_norm": 0.015256480313837528,
+ "learning_rate": 0.0006,
+ "loss": 4.445844650268555,
+ "step": 4578
+ },
+ {
+ "epoch": 63.60087336244541,
+ "grad_norm": 0.015305760316550732,
+ "learning_rate": 0.0006,
+ "loss": 4.317327499389648,
+ "step": 4579
+ },
+ {
+ "epoch": 63.61484716157205,
+ "grad_norm": 0.015200769528746605,
+ "learning_rate": 0.0006,
+ "loss": 4.25112247467041,
+ "step": 4580
+ },
+ {
+ "epoch": 63.62882096069869,
+ "grad_norm": 0.015919538214802742,
+ "learning_rate": 0.0006,
+ "loss": 4.277937889099121,
+ "step": 4581
+ },
+ {
+ "epoch": 63.64279475982533,
+ "grad_norm": 0.01572665013372898,
+ "learning_rate": 0.0006,
+ "loss": 4.302237510681152,
+ "step": 4582
+ },
+ {
+ "epoch": 63.65676855895197,
+ "grad_norm": 0.016428731381893158,
+ "learning_rate": 0.0006,
+ "loss": 4.36762809753418,
+ "step": 4583
+ },
+ {
+ "epoch": 63.670742358078606,
+ "grad_norm": 0.01706082746386528,
+ "learning_rate": 0.0006,
+ "loss": 4.236526966094971,
+ "step": 4584
+ },
+ {
+ "epoch": 63.68471615720524,
+ "grad_norm": 0.017850136384367943,
+ "learning_rate": 0.0006,
+ "loss": 4.264129638671875,
+ "step": 4585
+ },
+ {
+ "epoch": 63.698689956331876,
+ "grad_norm": 0.016079438850283623,
+ "learning_rate": 0.0006,
+ "loss": 4.449799060821533,
+ "step": 4586
+ },
+ {
+ "epoch": 63.712663755458514,
+ "grad_norm": 0.014996383339166641,
+ "learning_rate": 0.0006,
+ "loss": 4.404574394226074,
+ "step": 4587
+ },
+ {
+ "epoch": 63.72663755458515,
+ "grad_norm": 0.015910431742668152,
+ "learning_rate": 0.0006,
+ "loss": 4.348101615905762,
+ "step": 4588
+ },
+ {
+ "epoch": 63.74061135371179,
+ "grad_norm": 0.01630021072924137,
+ "learning_rate": 0.0006,
+ "loss": 4.3193817138671875,
+ "step": 4589
+ },
+ {
+ "epoch": 63.75458515283843,
+ "grad_norm": 0.014571773819625378,
+ "learning_rate": 0.0006,
+ "loss": 4.374402046203613,
+ "step": 4590
+ },
+ {
+ "epoch": 63.76855895196506,
+ "grad_norm": 0.01662345975637436,
+ "learning_rate": 0.0006,
+ "loss": 4.329075813293457,
+ "step": 4591
+ },
+ {
+ "epoch": 63.7825327510917,
+ "grad_norm": 0.016604198142886162,
+ "learning_rate": 0.0006,
+ "loss": 4.436126232147217,
+ "step": 4592
+ },
+ {
+ "epoch": 63.79650655021834,
+ "grad_norm": 0.016578884795308113,
+ "learning_rate": 0.0006,
+ "loss": 4.361468315124512,
+ "step": 4593
+ },
+ {
+ "epoch": 63.81048034934498,
+ "grad_norm": 0.016339825466275215,
+ "learning_rate": 0.0006,
+ "loss": 4.449892044067383,
+ "step": 4594
+ },
+ {
+ "epoch": 63.82445414847162,
+ "grad_norm": 0.016805455088615417,
+ "learning_rate": 0.0006,
+ "loss": 4.415866851806641,
+ "step": 4595
+ },
+ {
+ "epoch": 63.838427947598255,
+ "grad_norm": 0.017366213724017143,
+ "learning_rate": 0.0006,
+ "loss": 4.361730575561523,
+ "step": 4596
+ },
+ {
+ "epoch": 63.852401746724894,
+ "grad_norm": 0.015599003992974758,
+ "learning_rate": 0.0006,
+ "loss": 4.363637924194336,
+ "step": 4597
+ },
+ {
+ "epoch": 63.866375545851525,
+ "grad_norm": 0.015674632042646408,
+ "learning_rate": 0.0006,
+ "loss": 4.332045078277588,
+ "step": 4598
+ },
+ {
+ "epoch": 63.880349344978164,
+ "grad_norm": 0.015350443311035633,
+ "learning_rate": 0.0006,
+ "loss": 4.320063591003418,
+ "step": 4599
+ },
+ {
+ "epoch": 63.8943231441048,
+ "grad_norm": 0.015388146974146366,
+ "learning_rate": 0.0006,
+ "loss": 4.398199081420898,
+ "step": 4600
+ },
+ {
+ "epoch": 63.90829694323144,
+ "grad_norm": 0.01450140681117773,
+ "learning_rate": 0.0006,
+ "loss": 4.244076728820801,
+ "step": 4601
+ },
+ {
+ "epoch": 63.92227074235808,
+ "grad_norm": 0.01637491025030613,
+ "learning_rate": 0.0006,
+ "loss": 4.213549613952637,
+ "step": 4602
+ },
+ {
+ "epoch": 63.93624454148472,
+ "grad_norm": 0.0159814041107893,
+ "learning_rate": 0.0006,
+ "loss": 4.421775817871094,
+ "step": 4603
+ },
+ {
+ "epoch": 63.95021834061135,
+ "grad_norm": 0.017386827617883682,
+ "learning_rate": 0.0006,
+ "loss": 4.340023994445801,
+ "step": 4604
+ },
+ {
+ "epoch": 63.96419213973799,
+ "grad_norm": 0.019679656252264977,
+ "learning_rate": 0.0006,
+ "loss": 4.352534294128418,
+ "step": 4605
+ },
+ {
+ "epoch": 63.97816593886463,
+ "grad_norm": 0.022391516715288162,
+ "learning_rate": 0.0006,
+ "loss": 4.3538312911987305,
+ "step": 4606
+ },
+ {
+ "epoch": 63.992139737991266,
+ "grad_norm": 0.023281006142497063,
+ "learning_rate": 0.0006,
+ "loss": 4.432089328765869,
+ "step": 4607
+ },
+ {
+ "epoch": 64.0,
+ "grad_norm": 0.019916288554668427,
+ "learning_rate": 0.0006,
+ "loss": 4.466937065124512,
+ "step": 4608
+ },
+ {
+ "epoch": 64.0,
+ "eval_loss": 4.659008502960205,
+ "eval_runtime": 56.0019,
+ "eval_samples_per_second": 43.606,
+ "eval_steps_per_second": 1.375,
+ "step": 4608
+ },
+ {
+ "epoch": 64.01397379912663,
+ "grad_norm": 0.017351187765598297,
+ "learning_rate": 0.0006,
+ "loss": 4.294129371643066,
+ "step": 4609
+ },
+ {
+ "epoch": 64.02794759825328,
+ "grad_norm": 0.02405865117907524,
+ "learning_rate": 0.0006,
+ "loss": 4.266717910766602,
+ "step": 4610
+ },
+ {
+ "epoch": 64.04192139737991,
+ "grad_norm": 0.022158386185765266,
+ "learning_rate": 0.0006,
+ "loss": 4.395932197570801,
+ "step": 4611
+ },
+ {
+ "epoch": 64.05589519650655,
+ "grad_norm": 0.019238505512475967,
+ "learning_rate": 0.0006,
+ "loss": 4.266704082489014,
+ "step": 4612
+ },
+ {
+ "epoch": 64.06986899563319,
+ "grad_norm": 0.021670425310730934,
+ "learning_rate": 0.0006,
+ "loss": 4.3160552978515625,
+ "step": 4613
+ },
+ {
+ "epoch": 64.08384279475983,
+ "grad_norm": 0.02076641097664833,
+ "learning_rate": 0.0006,
+ "loss": 4.415626525878906,
+ "step": 4614
+ },
+ {
+ "epoch": 64.09781659388646,
+ "grad_norm": 0.018223078921437263,
+ "learning_rate": 0.0006,
+ "loss": 4.292387008666992,
+ "step": 4615
+ },
+ {
+ "epoch": 64.1117903930131,
+ "grad_norm": 0.017401937395334244,
+ "learning_rate": 0.0006,
+ "loss": 4.424524307250977,
+ "step": 4616
+ },
+ {
+ "epoch": 64.12576419213974,
+ "grad_norm": 0.01663743332028389,
+ "learning_rate": 0.0006,
+ "loss": 4.350032806396484,
+ "step": 4617
+ },
+ {
+ "epoch": 64.13973799126637,
+ "grad_norm": 0.018716629594564438,
+ "learning_rate": 0.0006,
+ "loss": 4.309157371520996,
+ "step": 4618
+ },
+ {
+ "epoch": 64.15371179039302,
+ "grad_norm": 0.019569825381040573,
+ "learning_rate": 0.0006,
+ "loss": 4.358292579650879,
+ "step": 4619
+ },
+ {
+ "epoch": 64.16768558951965,
+ "grad_norm": 0.018994422629475594,
+ "learning_rate": 0.0006,
+ "loss": 4.3338823318481445,
+ "step": 4620
+ },
+ {
+ "epoch": 64.18165938864628,
+ "grad_norm": 0.017690688371658325,
+ "learning_rate": 0.0006,
+ "loss": 4.414468765258789,
+ "step": 4621
+ },
+ {
+ "epoch": 64.19563318777293,
+ "grad_norm": 0.0170609001070261,
+ "learning_rate": 0.0006,
+ "loss": 4.4337568283081055,
+ "step": 4622
+ },
+ {
+ "epoch": 64.20960698689956,
+ "grad_norm": 0.016984131187200546,
+ "learning_rate": 0.0006,
+ "loss": 4.304839134216309,
+ "step": 4623
+ },
+ {
+ "epoch": 64.2235807860262,
+ "grad_norm": 0.01619867794215679,
+ "learning_rate": 0.0006,
+ "loss": 4.389623165130615,
+ "step": 4624
+ },
+ {
+ "epoch": 64.23755458515284,
+ "grad_norm": 0.015428110957145691,
+ "learning_rate": 0.0006,
+ "loss": 4.277515888214111,
+ "step": 4625
+ },
+ {
+ "epoch": 64.25152838427948,
+ "grad_norm": 0.015909532085061073,
+ "learning_rate": 0.0006,
+ "loss": 4.266029357910156,
+ "step": 4626
+ },
+ {
+ "epoch": 64.26550218340611,
+ "grad_norm": 0.014296560548245907,
+ "learning_rate": 0.0006,
+ "loss": 4.286067485809326,
+ "step": 4627
+ },
+ {
+ "epoch": 64.27947598253274,
+ "grad_norm": 0.014065971598029137,
+ "learning_rate": 0.0006,
+ "loss": 4.353528022766113,
+ "step": 4628
+ },
+ {
+ "epoch": 64.29344978165939,
+ "grad_norm": 0.014386293478310108,
+ "learning_rate": 0.0006,
+ "loss": 4.318715572357178,
+ "step": 4629
+ },
+ {
+ "epoch": 64.30742358078602,
+ "grad_norm": 0.014001123607158661,
+ "learning_rate": 0.0006,
+ "loss": 4.362497329711914,
+ "step": 4630
+ },
+ {
+ "epoch": 64.32139737991267,
+ "grad_norm": 0.013408979400992393,
+ "learning_rate": 0.0006,
+ "loss": 4.417397499084473,
+ "step": 4631
+ },
+ {
+ "epoch": 64.3353711790393,
+ "grad_norm": 0.014960093423724174,
+ "learning_rate": 0.0006,
+ "loss": 4.359710693359375,
+ "step": 4632
+ },
+ {
+ "epoch": 64.34934497816595,
+ "grad_norm": 0.014658777043223381,
+ "learning_rate": 0.0006,
+ "loss": 4.318822383880615,
+ "step": 4633
+ },
+ {
+ "epoch": 64.36331877729258,
+ "grad_norm": 0.013766744174063206,
+ "learning_rate": 0.0006,
+ "loss": 4.265679836273193,
+ "step": 4634
+ },
+ {
+ "epoch": 64.37729257641921,
+ "grad_norm": 0.013821255415678024,
+ "learning_rate": 0.0006,
+ "loss": 4.3452911376953125,
+ "step": 4635
+ },
+ {
+ "epoch": 64.39126637554585,
+ "grad_norm": 0.013541066087782383,
+ "learning_rate": 0.0006,
+ "loss": 4.429254531860352,
+ "step": 4636
+ },
+ {
+ "epoch": 64.40524017467249,
+ "grad_norm": 0.013985645025968552,
+ "learning_rate": 0.0006,
+ "loss": 4.304452896118164,
+ "step": 4637
+ },
+ {
+ "epoch": 64.41921397379913,
+ "grad_norm": 0.015079808421432972,
+ "learning_rate": 0.0006,
+ "loss": 4.296828746795654,
+ "step": 4638
+ },
+ {
+ "epoch": 64.43318777292576,
+ "grad_norm": 0.014896944165229797,
+ "learning_rate": 0.0006,
+ "loss": 4.299156665802002,
+ "step": 4639
+ },
+ {
+ "epoch": 64.44716157205241,
+ "grad_norm": 0.013770041987299919,
+ "learning_rate": 0.0006,
+ "loss": 4.351439952850342,
+ "step": 4640
+ },
+ {
+ "epoch": 64.46113537117904,
+ "grad_norm": 0.01531178504228592,
+ "learning_rate": 0.0006,
+ "loss": 4.373715877532959,
+ "step": 4641
+ },
+ {
+ "epoch": 64.47510917030567,
+ "grad_norm": 0.014024450443685055,
+ "learning_rate": 0.0006,
+ "loss": 4.368053436279297,
+ "step": 4642
+ },
+ {
+ "epoch": 64.48908296943232,
+ "grad_norm": 0.013864024542272091,
+ "learning_rate": 0.0006,
+ "loss": 4.378564834594727,
+ "step": 4643
+ },
+ {
+ "epoch": 64.50305676855895,
+ "grad_norm": 0.013925445266067982,
+ "learning_rate": 0.0006,
+ "loss": 4.261663436889648,
+ "step": 4644
+ },
+ {
+ "epoch": 64.5170305676856,
+ "grad_norm": 0.015011338517069817,
+ "learning_rate": 0.0006,
+ "loss": 4.337285995483398,
+ "step": 4645
+ },
+ {
+ "epoch": 64.53100436681223,
+ "grad_norm": 0.016941986978054047,
+ "learning_rate": 0.0006,
+ "loss": 4.336877822875977,
+ "step": 4646
+ },
+ {
+ "epoch": 64.54497816593886,
+ "grad_norm": 0.01719909906387329,
+ "learning_rate": 0.0006,
+ "loss": 4.149813175201416,
+ "step": 4647
+ },
+ {
+ "epoch": 64.5589519650655,
+ "grad_norm": 0.01608402095735073,
+ "learning_rate": 0.0006,
+ "loss": 4.191280364990234,
+ "step": 4648
+ },
+ {
+ "epoch": 64.57292576419214,
+ "grad_norm": 0.015284820459783077,
+ "learning_rate": 0.0006,
+ "loss": 4.35953426361084,
+ "step": 4649
+ },
+ {
+ "epoch": 64.58689956331878,
+ "grad_norm": 0.014855682849884033,
+ "learning_rate": 0.0006,
+ "loss": 4.308401107788086,
+ "step": 4650
+ },
+ {
+ "epoch": 64.60087336244541,
+ "grad_norm": 0.017553111538290977,
+ "learning_rate": 0.0006,
+ "loss": 4.279720306396484,
+ "step": 4651
+ },
+ {
+ "epoch": 64.61484716157206,
+ "grad_norm": 0.019506998360157013,
+ "learning_rate": 0.0006,
+ "loss": 4.319524765014648,
+ "step": 4652
+ },
+ {
+ "epoch": 64.62882096069869,
+ "grad_norm": 0.020293425768613815,
+ "learning_rate": 0.0006,
+ "loss": 4.340523719787598,
+ "step": 4653
+ },
+ {
+ "epoch": 64.64279475982532,
+ "grad_norm": 0.01664556749165058,
+ "learning_rate": 0.0006,
+ "loss": 4.296785831451416,
+ "step": 4654
+ },
+ {
+ "epoch": 64.65676855895197,
+ "grad_norm": 0.012592585757374763,
+ "learning_rate": 0.0006,
+ "loss": 4.464595317840576,
+ "step": 4655
+ },
+ {
+ "epoch": 64.6707423580786,
+ "grad_norm": 0.01549589168280363,
+ "learning_rate": 0.0006,
+ "loss": 4.275053024291992,
+ "step": 4656
+ },
+ {
+ "epoch": 64.68471615720524,
+ "grad_norm": 0.01705172471702099,
+ "learning_rate": 0.0006,
+ "loss": 4.388431549072266,
+ "step": 4657
+ },
+ {
+ "epoch": 64.69868995633188,
+ "grad_norm": 0.015438582748174667,
+ "learning_rate": 0.0006,
+ "loss": 4.312070846557617,
+ "step": 4658
+ },
+ {
+ "epoch": 64.71266375545852,
+ "grad_norm": 0.014448083005845547,
+ "learning_rate": 0.0006,
+ "loss": 4.237599849700928,
+ "step": 4659
+ },
+ {
+ "epoch": 64.72663755458515,
+ "grad_norm": 0.015267071314156055,
+ "learning_rate": 0.0006,
+ "loss": 4.29893684387207,
+ "step": 4660
+ },
+ {
+ "epoch": 64.74061135371178,
+ "grad_norm": 0.016615847125649452,
+ "learning_rate": 0.0006,
+ "loss": 4.432273864746094,
+ "step": 4661
+ },
+ {
+ "epoch": 64.75458515283843,
+ "grad_norm": 0.016584362834692,
+ "learning_rate": 0.0006,
+ "loss": 4.202266216278076,
+ "step": 4662
+ },
+ {
+ "epoch": 64.76855895196506,
+ "grad_norm": 0.018313994631171227,
+ "learning_rate": 0.0006,
+ "loss": 4.367966651916504,
+ "step": 4663
+ },
+ {
+ "epoch": 64.78253275109171,
+ "grad_norm": 0.017946481704711914,
+ "learning_rate": 0.0006,
+ "loss": 4.205172538757324,
+ "step": 4664
+ },
+ {
+ "epoch": 64.79650655021834,
+ "grad_norm": 0.017744354903697968,
+ "learning_rate": 0.0006,
+ "loss": 4.295661926269531,
+ "step": 4665
+ },
+ {
+ "epoch": 64.81048034934499,
+ "grad_norm": 0.016103824600577354,
+ "learning_rate": 0.0006,
+ "loss": 4.276543617248535,
+ "step": 4666
+ },
+ {
+ "epoch": 64.82445414847162,
+ "grad_norm": 0.014590012840926647,
+ "learning_rate": 0.0006,
+ "loss": 4.414247035980225,
+ "step": 4667
+ },
+ {
+ "epoch": 64.83842794759825,
+ "grad_norm": 0.01782386749982834,
+ "learning_rate": 0.0006,
+ "loss": 4.203220367431641,
+ "step": 4668
+ },
+ {
+ "epoch": 64.8524017467249,
+ "grad_norm": 0.022122733294963837,
+ "learning_rate": 0.0006,
+ "loss": 4.38078498840332,
+ "step": 4669
+ },
+ {
+ "epoch": 64.86637554585153,
+ "grad_norm": 0.0250293780118227,
+ "learning_rate": 0.0006,
+ "loss": 4.3310136795043945,
+ "step": 4670
+ },
+ {
+ "epoch": 64.88034934497817,
+ "grad_norm": 0.022438062354922295,
+ "learning_rate": 0.0006,
+ "loss": 4.428631782531738,
+ "step": 4671
+ },
+ {
+ "epoch": 64.8943231441048,
+ "grad_norm": 0.018518192693591118,
+ "learning_rate": 0.0006,
+ "loss": 4.428250312805176,
+ "step": 4672
+ },
+ {
+ "epoch": 64.90829694323143,
+ "grad_norm": 0.018105225637555122,
+ "learning_rate": 0.0006,
+ "loss": 4.336209297180176,
+ "step": 4673
+ },
+ {
+ "epoch": 64.92227074235808,
+ "grad_norm": 0.017564399167895317,
+ "learning_rate": 0.0006,
+ "loss": 4.416011810302734,
+ "step": 4674
+ },
+ {
+ "epoch": 64.93624454148471,
+ "grad_norm": 0.018747031688690186,
+ "learning_rate": 0.0006,
+ "loss": 4.419938087463379,
+ "step": 4675
+ },
+ {
+ "epoch": 64.95021834061136,
+ "grad_norm": 0.018686484545469284,
+ "learning_rate": 0.0006,
+ "loss": 4.347185134887695,
+ "step": 4676
+ },
+ {
+ "epoch": 64.96419213973799,
+ "grad_norm": 0.019329898059368134,
+ "learning_rate": 0.0006,
+ "loss": 4.343749046325684,
+ "step": 4677
+ },
+ {
+ "epoch": 64.97816593886463,
+ "grad_norm": 0.018517782911658287,
+ "learning_rate": 0.0006,
+ "loss": 4.264604568481445,
+ "step": 4678
+ },
+ {
+ "epoch": 64.99213973799127,
+ "grad_norm": 0.016579125076532364,
+ "learning_rate": 0.0006,
+ "loss": 4.303423881530762,
+ "step": 4679
+ },
+ {
+ "epoch": 65.0,
+ "grad_norm": 0.01617567613720894,
+ "learning_rate": 0.0006,
+ "loss": 4.326556205749512,
+ "step": 4680
+ },
+ {
+ "epoch": 65.0,
+ "eval_loss": 4.655026912689209,
+ "eval_runtime": 57.1299,
+ "eval_samples_per_second": 42.745,
+ "eval_steps_per_second": 1.348,
+ "step": 4680
+ },
+ {
+ "epoch": 65.01397379912663,
+ "grad_norm": 0.01607249490916729,
+ "learning_rate": 0.0006,
+ "loss": 4.276607036590576,
+ "step": 4681
+ },
+ {
+ "epoch": 65.02794759825328,
+ "grad_norm": 0.016779497265815735,
+ "learning_rate": 0.0006,
+ "loss": 4.412464618682861,
+ "step": 4682
+ },
+ {
+ "epoch": 65.04192139737991,
+ "grad_norm": 0.018161555752158165,
+ "learning_rate": 0.0006,
+ "loss": 4.199329376220703,
+ "step": 4683
+ },
+ {
+ "epoch": 65.05589519650655,
+ "grad_norm": 0.017508309334516525,
+ "learning_rate": 0.0006,
+ "loss": 4.217567443847656,
+ "step": 4684
+ },
+ {
+ "epoch": 65.06986899563319,
+ "grad_norm": 0.015716979280114174,
+ "learning_rate": 0.0006,
+ "loss": 4.322023391723633,
+ "step": 4685
+ },
+ {
+ "epoch": 65.08384279475983,
+ "grad_norm": 0.015113635919988155,
+ "learning_rate": 0.0006,
+ "loss": 4.333515167236328,
+ "step": 4686
+ },
+ {
+ "epoch": 65.09781659388646,
+ "grad_norm": 0.01437560748308897,
+ "learning_rate": 0.0006,
+ "loss": 4.295609474182129,
+ "step": 4687
+ },
+ {
+ "epoch": 65.1117903930131,
+ "grad_norm": 0.013074862770736217,
+ "learning_rate": 0.0006,
+ "loss": 4.314120769500732,
+ "step": 4688
+ },
+ {
+ "epoch": 65.12576419213974,
+ "grad_norm": 0.012908284552395344,
+ "learning_rate": 0.0006,
+ "loss": 4.216890811920166,
+ "step": 4689
+ },
+ {
+ "epoch": 65.13973799126637,
+ "grad_norm": 0.013001754879951477,
+ "learning_rate": 0.0006,
+ "loss": 4.287717342376709,
+ "step": 4690
+ },
+ {
+ "epoch": 65.15371179039302,
+ "grad_norm": 0.013694263994693756,
+ "learning_rate": 0.0006,
+ "loss": 4.333837985992432,
+ "step": 4691
+ },
+ {
+ "epoch": 65.16768558951965,
+ "grad_norm": 0.014461833983659744,
+ "learning_rate": 0.0006,
+ "loss": 4.323545455932617,
+ "step": 4692
+ },
+ {
+ "epoch": 65.18165938864628,
+ "grad_norm": 0.013670675456523895,
+ "learning_rate": 0.0006,
+ "loss": 4.3313493728637695,
+ "step": 4693
+ },
+ {
+ "epoch": 65.19563318777293,
+ "grad_norm": 0.013094350695610046,
+ "learning_rate": 0.0006,
+ "loss": 4.23858118057251,
+ "step": 4694
+ },
+ {
+ "epoch": 65.20960698689956,
+ "grad_norm": 0.013996944762766361,
+ "learning_rate": 0.0006,
+ "loss": 4.428705215454102,
+ "step": 4695
+ },
+ {
+ "epoch": 65.2235807860262,
+ "grad_norm": 0.013295048847794533,
+ "learning_rate": 0.0006,
+ "loss": 4.238950252532959,
+ "step": 4696
+ },
+ {
+ "epoch": 65.23755458515284,
+ "grad_norm": 0.013380616903305054,
+ "learning_rate": 0.0006,
+ "loss": 4.273185729980469,
+ "step": 4697
+ },
+ {
+ "epoch": 65.25152838427948,
+ "grad_norm": 0.015062890015542507,
+ "learning_rate": 0.0006,
+ "loss": 4.29472017288208,
+ "step": 4698
+ },
+ {
+ "epoch": 65.26550218340611,
+ "grad_norm": 0.017317263409495354,
+ "learning_rate": 0.0006,
+ "loss": 4.337954998016357,
+ "step": 4699
+ },
+ {
+ "epoch": 65.27947598253274,
+ "grad_norm": 0.017698992043733597,
+ "learning_rate": 0.0006,
+ "loss": 4.336437225341797,
+ "step": 4700
+ },
+ {
+ "epoch": 65.29344978165939,
+ "grad_norm": 0.01628575474023819,
+ "learning_rate": 0.0006,
+ "loss": 4.302372455596924,
+ "step": 4701
+ },
+ {
+ "epoch": 65.30742358078602,
+ "grad_norm": 0.014531854540109634,
+ "learning_rate": 0.0006,
+ "loss": 4.256450653076172,
+ "step": 4702
+ },
+ {
+ "epoch": 65.32139737991267,
+ "grad_norm": 0.014549105428159237,
+ "learning_rate": 0.0006,
+ "loss": 4.302240371704102,
+ "step": 4703
+ },
+ {
+ "epoch": 65.3353711790393,
+ "grad_norm": 0.014881663955748081,
+ "learning_rate": 0.0006,
+ "loss": 4.3316121101379395,
+ "step": 4704
+ },
+ {
+ "epoch": 65.34934497816595,
+ "grad_norm": 0.014513437636196613,
+ "learning_rate": 0.0006,
+ "loss": 4.441212177276611,
+ "step": 4705
+ },
+ {
+ "epoch": 65.36331877729258,
+ "grad_norm": 0.014122087508440018,
+ "learning_rate": 0.0006,
+ "loss": 4.343296527862549,
+ "step": 4706
+ },
+ {
+ "epoch": 65.37729257641921,
+ "grad_norm": 0.014817348681390285,
+ "learning_rate": 0.0006,
+ "loss": 4.195674896240234,
+ "step": 4707
+ },
+ {
+ "epoch": 65.39126637554585,
+ "grad_norm": 0.01787056028842926,
+ "learning_rate": 0.0006,
+ "loss": 4.3507914543151855,
+ "step": 4708
+ },
+ {
+ "epoch": 65.40524017467249,
+ "grad_norm": 0.015295136719942093,
+ "learning_rate": 0.0006,
+ "loss": 4.234372138977051,
+ "step": 4709
+ },
+ {
+ "epoch": 65.41921397379913,
+ "grad_norm": 0.014324234798550606,
+ "learning_rate": 0.0006,
+ "loss": 4.2907867431640625,
+ "step": 4710
+ },
+ {
+ "epoch": 65.43318777292576,
+ "grad_norm": 0.016090670600533485,
+ "learning_rate": 0.0006,
+ "loss": 4.177244186401367,
+ "step": 4711
+ },
+ {
+ "epoch": 65.44716157205241,
+ "grad_norm": 0.0174604132771492,
+ "learning_rate": 0.0006,
+ "loss": 4.335577964782715,
+ "step": 4712
+ },
+ {
+ "epoch": 65.46113537117904,
+ "grad_norm": 0.016731899231672287,
+ "learning_rate": 0.0006,
+ "loss": 4.30197811126709,
+ "step": 4713
+ },
+ {
+ "epoch": 65.47510917030567,
+ "grad_norm": 0.017423735931515694,
+ "learning_rate": 0.0006,
+ "loss": 4.3528852462768555,
+ "step": 4714
+ },
+ {
+ "epoch": 65.48908296943232,
+ "grad_norm": 0.01780976541340351,
+ "learning_rate": 0.0006,
+ "loss": 4.185036659240723,
+ "step": 4715
+ },
+ {
+ "epoch": 65.50305676855895,
+ "grad_norm": 0.019977200776338577,
+ "learning_rate": 0.0006,
+ "loss": 4.328766345977783,
+ "step": 4716
+ },
+ {
+ "epoch": 65.5170305676856,
+ "grad_norm": 0.020284287631511688,
+ "learning_rate": 0.0006,
+ "loss": 4.285494804382324,
+ "step": 4717
+ },
+ {
+ "epoch": 65.53100436681223,
+ "grad_norm": 0.019168002530932426,
+ "learning_rate": 0.0006,
+ "loss": 4.328560829162598,
+ "step": 4718
+ },
+ {
+ "epoch": 65.54497816593886,
+ "grad_norm": 0.019217776134610176,
+ "learning_rate": 0.0006,
+ "loss": 4.189998626708984,
+ "step": 4719
+ },
+ {
+ "epoch": 65.5589519650655,
+ "grad_norm": 0.018960488960146904,
+ "learning_rate": 0.0006,
+ "loss": 4.4995012283325195,
+ "step": 4720
+ },
+ {
+ "epoch": 65.57292576419214,
+ "grad_norm": 0.01941288262605667,
+ "learning_rate": 0.0006,
+ "loss": 4.26570987701416,
+ "step": 4721
+ },
+ {
+ "epoch": 65.58689956331878,
+ "grad_norm": 0.020263448357582092,
+ "learning_rate": 0.0006,
+ "loss": 4.411645889282227,
+ "step": 4722
+ },
+ {
+ "epoch": 65.60087336244541,
+ "grad_norm": 0.02144799567759037,
+ "learning_rate": 0.0006,
+ "loss": 4.163675308227539,
+ "step": 4723
+ },
+ {
+ "epoch": 65.61484716157206,
+ "grad_norm": 0.019790446385741234,
+ "learning_rate": 0.0006,
+ "loss": 4.2361063957214355,
+ "step": 4724
+ },
+ {
+ "epoch": 65.62882096069869,
+ "grad_norm": 0.01829805225133896,
+ "learning_rate": 0.0006,
+ "loss": 4.449810028076172,
+ "step": 4725
+ },
+ {
+ "epoch": 65.64279475982532,
+ "grad_norm": 0.01571105420589447,
+ "learning_rate": 0.0006,
+ "loss": 4.281406879425049,
+ "step": 4726
+ },
+ {
+ "epoch": 65.65676855895197,
+ "grad_norm": 0.016477147117257118,
+ "learning_rate": 0.0006,
+ "loss": 4.293408393859863,
+ "step": 4727
+ },
+ {
+ "epoch": 65.6707423580786,
+ "grad_norm": 0.016766032204031944,
+ "learning_rate": 0.0006,
+ "loss": 4.378252983093262,
+ "step": 4728
+ },
+ {
+ "epoch": 65.68471615720524,
+ "grad_norm": 0.0159933939576149,
+ "learning_rate": 0.0006,
+ "loss": 4.3065667152404785,
+ "step": 4729
+ },
+ {
+ "epoch": 65.69868995633188,
+ "grad_norm": 0.01658209227025509,
+ "learning_rate": 0.0006,
+ "loss": 4.174026966094971,
+ "step": 4730
+ },
+ {
+ "epoch": 65.71266375545852,
+ "grad_norm": 0.015081380493938923,
+ "learning_rate": 0.0006,
+ "loss": 4.360222339630127,
+ "step": 4731
+ },
+ {
+ "epoch": 65.72663755458515,
+ "grad_norm": 0.015110890381038189,
+ "learning_rate": 0.0006,
+ "loss": 4.218810558319092,
+ "step": 4732
+ },
+ {
+ "epoch": 65.74061135371178,
+ "grad_norm": 0.015300953760743141,
+ "learning_rate": 0.0006,
+ "loss": 4.331979274749756,
+ "step": 4733
+ },
+ {
+ "epoch": 65.75458515283843,
+ "grad_norm": 0.01715504191815853,
+ "learning_rate": 0.0006,
+ "loss": 4.349112033843994,
+ "step": 4734
+ },
+ {
+ "epoch": 65.76855895196506,
+ "grad_norm": 0.015870366245508194,
+ "learning_rate": 0.0006,
+ "loss": 4.329850196838379,
+ "step": 4735
+ },
+ {
+ "epoch": 65.78253275109171,
+ "grad_norm": 0.01541937980800867,
+ "learning_rate": 0.0006,
+ "loss": 4.3400163650512695,
+ "step": 4736
+ },
+ {
+ "epoch": 65.79650655021834,
+ "grad_norm": 0.01738942787051201,
+ "learning_rate": 0.0006,
+ "loss": 4.220798492431641,
+ "step": 4737
+ },
+ {
+ "epoch": 65.81048034934499,
+ "grad_norm": 0.01685990020632744,
+ "learning_rate": 0.0006,
+ "loss": 4.177238464355469,
+ "step": 4738
+ },
+ {
+ "epoch": 65.82445414847162,
+ "grad_norm": 0.016183465719223022,
+ "learning_rate": 0.0006,
+ "loss": 4.327216625213623,
+ "step": 4739
+ },
+ {
+ "epoch": 65.83842794759825,
+ "grad_norm": 0.01677694357931614,
+ "learning_rate": 0.0006,
+ "loss": 4.27766227722168,
+ "step": 4740
+ },
+ {
+ "epoch": 65.8524017467249,
+ "grad_norm": 0.01645066775381565,
+ "learning_rate": 0.0006,
+ "loss": 4.288347244262695,
+ "step": 4741
+ },
+ {
+ "epoch": 65.86637554585153,
+ "grad_norm": 0.018034635111689568,
+ "learning_rate": 0.0006,
+ "loss": 4.292458534240723,
+ "step": 4742
+ },
+ {
+ "epoch": 65.88034934497817,
+ "grad_norm": 0.017973346635699272,
+ "learning_rate": 0.0006,
+ "loss": 4.3117265701293945,
+ "step": 4743
+ },
+ {
+ "epoch": 65.8943231441048,
+ "grad_norm": 0.016605669632554054,
+ "learning_rate": 0.0006,
+ "loss": 4.280615329742432,
+ "step": 4744
+ },
+ {
+ "epoch": 65.90829694323143,
+ "grad_norm": 0.016298074275255203,
+ "learning_rate": 0.0006,
+ "loss": 4.301933288574219,
+ "step": 4745
+ },
+ {
+ "epoch": 65.92227074235808,
+ "grad_norm": 0.015383216552436352,
+ "learning_rate": 0.0006,
+ "loss": 4.352383136749268,
+ "step": 4746
+ },
+ {
+ "epoch": 65.93624454148471,
+ "grad_norm": 0.016222182661294937,
+ "learning_rate": 0.0006,
+ "loss": 4.305725574493408,
+ "step": 4747
+ },
+ {
+ "epoch": 65.95021834061136,
+ "grad_norm": 0.01691329851746559,
+ "learning_rate": 0.0006,
+ "loss": 4.29502010345459,
+ "step": 4748
+ },
+ {
+ "epoch": 65.96419213973799,
+ "grad_norm": 0.017637677490711212,
+ "learning_rate": 0.0006,
+ "loss": 4.3133039474487305,
+ "step": 4749
+ },
+ {
+ "epoch": 65.97816593886463,
+ "grad_norm": 0.021362047642469406,
+ "learning_rate": 0.0006,
+ "loss": 4.458834648132324,
+ "step": 4750
+ },
+ {
+ "epoch": 65.99213973799127,
+ "grad_norm": 0.023553011938929558,
+ "learning_rate": 0.0006,
+ "loss": 4.298727989196777,
+ "step": 4751
+ },
+ {
+ "epoch": 66.0,
+ "grad_norm": 0.02154950052499771,
+ "learning_rate": 0.0006,
+ "loss": 4.403392791748047,
+ "step": 4752
+ },
+ {
+ "epoch": 66.0,
+ "eval_loss": 4.724340438842773,
+ "eval_runtime": 56.9555,
+ "eval_samples_per_second": 42.876,
+ "eval_steps_per_second": 1.352,
+ "step": 4752
+ },
+ {
+ "epoch": 66.01397379912663,
+ "grad_norm": 0.01916784606873989,
+ "learning_rate": 0.0006,
+ "loss": 4.22931432723999,
+ "step": 4753
+ },
+ {
+ "epoch": 66.02794759825328,
+ "grad_norm": 0.017492054030299187,
+ "learning_rate": 0.0006,
+ "loss": 4.271522521972656,
+ "step": 4754
+ },
+ {
+ "epoch": 66.04192139737991,
+ "grad_norm": 0.018408095464110374,
+ "learning_rate": 0.0006,
+ "loss": 4.337705135345459,
+ "step": 4755
+ },
+ {
+ "epoch": 66.05589519650655,
+ "grad_norm": 0.019220108166337013,
+ "learning_rate": 0.0006,
+ "loss": 4.280340194702148,
+ "step": 4756
+ },
+ {
+ "epoch": 66.06986899563319,
+ "grad_norm": 0.01773735322058201,
+ "learning_rate": 0.0006,
+ "loss": 4.390201568603516,
+ "step": 4757
+ },
+ {
+ "epoch": 66.08384279475983,
+ "grad_norm": 0.01679125428199768,
+ "learning_rate": 0.0006,
+ "loss": 4.299282073974609,
+ "step": 4758
+ },
+ {
+ "epoch": 66.09781659388646,
+ "grad_norm": 0.01479059923440218,
+ "learning_rate": 0.0006,
+ "loss": 4.472630500793457,
+ "step": 4759
+ },
+ {
+ "epoch": 66.1117903930131,
+ "grad_norm": 0.015130104497075081,
+ "learning_rate": 0.0006,
+ "loss": 4.301194190979004,
+ "step": 4760
+ },
+ {
+ "epoch": 66.12576419213974,
+ "grad_norm": 0.0158979594707489,
+ "learning_rate": 0.0006,
+ "loss": 4.375751972198486,
+ "step": 4761
+ },
+ {
+ "epoch": 66.13973799126637,
+ "grad_norm": 0.014432408846914768,
+ "learning_rate": 0.0006,
+ "loss": 4.279788017272949,
+ "step": 4762
+ },
+ {
+ "epoch": 66.15371179039302,
+ "grad_norm": 0.013026784174144268,
+ "learning_rate": 0.0006,
+ "loss": 4.255035400390625,
+ "step": 4763
+ },
+ {
+ "epoch": 66.16768558951965,
+ "grad_norm": 0.014422653242945671,
+ "learning_rate": 0.0006,
+ "loss": 4.261832237243652,
+ "step": 4764
+ },
+ {
+ "epoch": 66.18165938864628,
+ "grad_norm": 0.014613019302487373,
+ "learning_rate": 0.0006,
+ "loss": 4.382184028625488,
+ "step": 4765
+ },
+ {
+ "epoch": 66.19563318777293,
+ "grad_norm": 0.014989894814789295,
+ "learning_rate": 0.0006,
+ "loss": 4.40528678894043,
+ "step": 4766
+ },
+ {
+ "epoch": 66.20960698689956,
+ "grad_norm": 0.018190350383520126,
+ "learning_rate": 0.0006,
+ "loss": 4.302414894104004,
+ "step": 4767
+ },
+ {
+ "epoch": 66.2235807860262,
+ "grad_norm": 0.020887890830636024,
+ "learning_rate": 0.0006,
+ "loss": 4.289920806884766,
+ "step": 4768
+ },
+ {
+ "epoch": 66.23755458515284,
+ "grad_norm": 0.02376876398921013,
+ "learning_rate": 0.0006,
+ "loss": 4.297855377197266,
+ "step": 4769
+ },
+ {
+ "epoch": 66.25152838427948,
+ "grad_norm": 0.02152269519865513,
+ "learning_rate": 0.0006,
+ "loss": 4.1972832679748535,
+ "step": 4770
+ },
+ {
+ "epoch": 66.26550218340611,
+ "grad_norm": 0.015572802163660526,
+ "learning_rate": 0.0006,
+ "loss": 4.307487487792969,
+ "step": 4771
+ },
+ {
+ "epoch": 66.27947598253274,
+ "grad_norm": 0.015228749252855778,
+ "learning_rate": 0.0006,
+ "loss": 4.103250503540039,
+ "step": 4772
+ },
+ {
+ "epoch": 66.29344978165939,
+ "grad_norm": 0.016925504431128502,
+ "learning_rate": 0.0006,
+ "loss": 4.247461318969727,
+ "step": 4773
+ },
+ {
+ "epoch": 66.30742358078602,
+ "grad_norm": 0.01613711006939411,
+ "learning_rate": 0.0006,
+ "loss": 4.4699273109436035,
+ "step": 4774
+ },
+ {
+ "epoch": 66.32139737991267,
+ "grad_norm": 0.018332140520215034,
+ "learning_rate": 0.0006,
+ "loss": 4.22969388961792,
+ "step": 4775
+ },
+ {
+ "epoch": 66.3353711790393,
+ "grad_norm": 0.019575074315071106,
+ "learning_rate": 0.0006,
+ "loss": 4.29397439956665,
+ "step": 4776
+ },
+ {
+ "epoch": 66.34934497816595,
+ "grad_norm": 0.018177419900894165,
+ "learning_rate": 0.0006,
+ "loss": 4.221899509429932,
+ "step": 4777
+ },
+ {
+ "epoch": 66.36331877729258,
+ "grad_norm": 0.014292889274656773,
+ "learning_rate": 0.0006,
+ "loss": 4.325457572937012,
+ "step": 4778
+ },
+ {
+ "epoch": 66.37729257641921,
+ "grad_norm": 0.015098796226084232,
+ "learning_rate": 0.0006,
+ "loss": 4.364837169647217,
+ "step": 4779
+ },
+ {
+ "epoch": 66.39126637554585,
+ "grad_norm": 0.013992635533213615,
+ "learning_rate": 0.0006,
+ "loss": 4.299082279205322,
+ "step": 4780
+ },
+ {
+ "epoch": 66.40524017467249,
+ "grad_norm": 0.01504511572420597,
+ "learning_rate": 0.0006,
+ "loss": 4.230877876281738,
+ "step": 4781
+ },
+ {
+ "epoch": 66.41921397379913,
+ "grad_norm": 0.014875639230012894,
+ "learning_rate": 0.0006,
+ "loss": 4.368405818939209,
+ "step": 4782
+ },
+ {
+ "epoch": 66.43318777292576,
+ "grad_norm": 0.014306887984275818,
+ "learning_rate": 0.0006,
+ "loss": 4.389184474945068,
+ "step": 4783
+ },
+ {
+ "epoch": 66.44716157205241,
+ "grad_norm": 0.012959563173353672,
+ "learning_rate": 0.0006,
+ "loss": 4.37934684753418,
+ "step": 4784
+ },
+ {
+ "epoch": 66.46113537117904,
+ "grad_norm": 0.013206703588366508,
+ "learning_rate": 0.0006,
+ "loss": 4.442799091339111,
+ "step": 4785
+ },
+ {
+ "epoch": 66.47510917030567,
+ "grad_norm": 0.014848909340798855,
+ "learning_rate": 0.0006,
+ "loss": 4.203481674194336,
+ "step": 4786
+ },
+ {
+ "epoch": 66.48908296943232,
+ "grad_norm": 0.015637682750821114,
+ "learning_rate": 0.0006,
+ "loss": 4.27552604675293,
+ "step": 4787
+ },
+ {
+ "epoch": 66.50305676855895,
+ "grad_norm": 0.016017816960811615,
+ "learning_rate": 0.0006,
+ "loss": 4.296402931213379,
+ "step": 4788
+ },
+ {
+ "epoch": 66.5170305676856,
+ "grad_norm": 0.01577814109623432,
+ "learning_rate": 0.0006,
+ "loss": 4.301733016967773,
+ "step": 4789
+ },
+ {
+ "epoch": 66.53100436681223,
+ "grad_norm": 0.015293175354599953,
+ "learning_rate": 0.0006,
+ "loss": 4.369877815246582,
+ "step": 4790
+ },
+ {
+ "epoch": 66.54497816593886,
+ "grad_norm": 0.014262278564274311,
+ "learning_rate": 0.0006,
+ "loss": 4.3314738273620605,
+ "step": 4791
+ },
+ {
+ "epoch": 66.5589519650655,
+ "grad_norm": 0.015123482793569565,
+ "learning_rate": 0.0006,
+ "loss": 4.313591957092285,
+ "step": 4792
+ },
+ {
+ "epoch": 66.57292576419214,
+ "grad_norm": 0.015458766371011734,
+ "learning_rate": 0.0006,
+ "loss": 4.3045854568481445,
+ "step": 4793
+ },
+ {
+ "epoch": 66.58689956331878,
+ "grad_norm": 0.017217233777046204,
+ "learning_rate": 0.0006,
+ "loss": 4.269944190979004,
+ "step": 4794
+ },
+ {
+ "epoch": 66.60087336244541,
+ "grad_norm": 0.01962282694876194,
+ "learning_rate": 0.0006,
+ "loss": 4.271735191345215,
+ "step": 4795
+ },
+ {
+ "epoch": 66.61484716157206,
+ "grad_norm": 0.01917066052556038,
+ "learning_rate": 0.0006,
+ "loss": 4.288168430328369,
+ "step": 4796
+ },
+ {
+ "epoch": 66.62882096069869,
+ "grad_norm": 0.018211524933576584,
+ "learning_rate": 0.0006,
+ "loss": 4.280660629272461,
+ "step": 4797
+ },
+ {
+ "epoch": 66.64279475982532,
+ "grad_norm": 0.01936524175107479,
+ "learning_rate": 0.0006,
+ "loss": 4.3710551261901855,
+ "step": 4798
+ },
+ {
+ "epoch": 66.65676855895197,
+ "grad_norm": 0.01882351003587246,
+ "learning_rate": 0.0006,
+ "loss": 4.342750549316406,
+ "step": 4799
+ },
+ {
+ "epoch": 66.6707423580786,
+ "grad_norm": 0.017149122431874275,
+ "learning_rate": 0.0006,
+ "loss": 4.32134485244751,
+ "step": 4800
+ },
+ {
+ "epoch": 66.68471615720524,
+ "grad_norm": 0.01699327677488327,
+ "learning_rate": 0.0006,
+ "loss": 4.248893737792969,
+ "step": 4801
+ },
+ {
+ "epoch": 66.69868995633188,
+ "grad_norm": 0.016873255372047424,
+ "learning_rate": 0.0006,
+ "loss": 4.277350425720215,
+ "step": 4802
+ },
+ {
+ "epoch": 66.71266375545852,
+ "grad_norm": 0.016592837870121002,
+ "learning_rate": 0.0006,
+ "loss": 4.221665859222412,
+ "step": 4803
+ },
+ {
+ "epoch": 66.72663755458515,
+ "grad_norm": 0.015887131914496422,
+ "learning_rate": 0.0006,
+ "loss": 4.328040599822998,
+ "step": 4804
+ },
+ {
+ "epoch": 66.74061135371178,
+ "grad_norm": 0.014057286083698273,
+ "learning_rate": 0.0006,
+ "loss": 4.3294548988342285,
+ "step": 4805
+ },
+ {
+ "epoch": 66.75458515283843,
+ "grad_norm": 0.013135905377566814,
+ "learning_rate": 0.0006,
+ "loss": 4.254383563995361,
+ "step": 4806
+ },
+ {
+ "epoch": 66.76855895196506,
+ "grad_norm": 0.013632557354867458,
+ "learning_rate": 0.0006,
+ "loss": 4.342815399169922,
+ "step": 4807
+ },
+ {
+ "epoch": 66.78253275109171,
+ "grad_norm": 0.015512688085436821,
+ "learning_rate": 0.0006,
+ "loss": 4.297835350036621,
+ "step": 4808
+ },
+ {
+ "epoch": 66.79650655021834,
+ "grad_norm": 0.01621832884848118,
+ "learning_rate": 0.0006,
+ "loss": 4.314773082733154,
+ "step": 4809
+ },
+ {
+ "epoch": 66.81048034934499,
+ "grad_norm": 0.016196032986044884,
+ "learning_rate": 0.0006,
+ "loss": 4.29209566116333,
+ "step": 4810
+ },
+ {
+ "epoch": 66.82445414847162,
+ "grad_norm": 0.015463353134691715,
+ "learning_rate": 0.0006,
+ "loss": 4.262795448303223,
+ "step": 4811
+ },
+ {
+ "epoch": 66.83842794759825,
+ "grad_norm": 0.016143787652254105,
+ "learning_rate": 0.0006,
+ "loss": 4.274284362792969,
+ "step": 4812
+ },
+ {
+ "epoch": 66.8524017467249,
+ "grad_norm": 0.01610061153769493,
+ "learning_rate": 0.0006,
+ "loss": 4.337160587310791,
+ "step": 4813
+ },
+ {
+ "epoch": 66.86637554585153,
+ "grad_norm": 0.01582838036119938,
+ "learning_rate": 0.0006,
+ "loss": 4.192723274230957,
+ "step": 4814
+ },
+ {
+ "epoch": 66.88034934497817,
+ "grad_norm": 0.015081540681421757,
+ "learning_rate": 0.0006,
+ "loss": 4.370708465576172,
+ "step": 4815
+ },
+ {
+ "epoch": 66.8943231441048,
+ "grad_norm": 0.015929723158478737,
+ "learning_rate": 0.0006,
+ "loss": 4.340244293212891,
+ "step": 4816
+ },
+ {
+ "epoch": 66.90829694323143,
+ "grad_norm": 0.01784711331129074,
+ "learning_rate": 0.0006,
+ "loss": 4.260406970977783,
+ "step": 4817
+ },
+ {
+ "epoch": 66.92227074235808,
+ "grad_norm": 0.017434703186154366,
+ "learning_rate": 0.0006,
+ "loss": 4.286900043487549,
+ "step": 4818
+ },
+ {
+ "epoch": 66.93624454148471,
+ "grad_norm": 0.016262540593743324,
+ "learning_rate": 0.0006,
+ "loss": 4.292142391204834,
+ "step": 4819
+ },
+ {
+ "epoch": 66.95021834061136,
+ "grad_norm": 0.016122829169034958,
+ "learning_rate": 0.0006,
+ "loss": 4.291485786437988,
+ "step": 4820
+ },
+ {
+ "epoch": 66.96419213973799,
+ "grad_norm": 0.014361178502440453,
+ "learning_rate": 0.0006,
+ "loss": 4.275425910949707,
+ "step": 4821
+ },
+ {
+ "epoch": 66.97816593886463,
+ "grad_norm": 0.01373987551778555,
+ "learning_rate": 0.0006,
+ "loss": 4.223120212554932,
+ "step": 4822
+ },
+ {
+ "epoch": 66.99213973799127,
+ "grad_norm": 0.015687376260757446,
+ "learning_rate": 0.0006,
+ "loss": 4.355381011962891,
+ "step": 4823
+ },
+ {
+ "epoch": 67.0,
+ "grad_norm": 0.01998465321958065,
+ "learning_rate": 0.0006,
+ "loss": 4.264053821563721,
+ "step": 4824
+ },
+ {
+ "epoch": 67.0,
+ "eval_loss": 4.662398338317871,
+ "eval_runtime": 56.7928,
+ "eval_samples_per_second": 42.998,
+ "eval_steps_per_second": 1.356,
+ "step": 4824
+ },
+ {
+ "epoch": 67.01397379912663,
+ "grad_norm": 0.021736836060881615,
+ "learning_rate": 0.0006,
+ "loss": 4.243354797363281,
+ "step": 4825
+ },
+ {
+ "epoch": 67.02794759825328,
+ "grad_norm": 0.021476037800312042,
+ "learning_rate": 0.0006,
+ "loss": 4.218207836151123,
+ "step": 4826
+ },
+ {
+ "epoch": 67.04192139737991,
+ "grad_norm": 0.017188388854265213,
+ "learning_rate": 0.0006,
+ "loss": 4.18048095703125,
+ "step": 4827
+ },
+ {
+ "epoch": 67.05589519650655,
+ "grad_norm": 0.017293719574809074,
+ "learning_rate": 0.0006,
+ "loss": 4.239145278930664,
+ "step": 4828
+ },
+ {
+ "epoch": 67.06986899563319,
+ "grad_norm": 0.017383819445967674,
+ "learning_rate": 0.0006,
+ "loss": 4.114765167236328,
+ "step": 4829
+ },
+ {
+ "epoch": 67.08384279475983,
+ "grad_norm": 0.01651529036462307,
+ "learning_rate": 0.0006,
+ "loss": 4.351023197174072,
+ "step": 4830
+ },
+ {
+ "epoch": 67.09781659388646,
+ "grad_norm": 0.015893027186393738,
+ "learning_rate": 0.0006,
+ "loss": 4.261432647705078,
+ "step": 4831
+ },
+ {
+ "epoch": 67.1117903930131,
+ "grad_norm": 0.013381517492234707,
+ "learning_rate": 0.0006,
+ "loss": 4.172064781188965,
+ "step": 4832
+ },
+ {
+ "epoch": 67.12576419213974,
+ "grad_norm": 0.014642654918134212,
+ "learning_rate": 0.0006,
+ "loss": 4.356171131134033,
+ "step": 4833
+ },
+ {
+ "epoch": 67.13973799126637,
+ "grad_norm": 0.01425282284617424,
+ "learning_rate": 0.0006,
+ "loss": 4.303495407104492,
+ "step": 4834
+ },
+ {
+ "epoch": 67.15371179039302,
+ "grad_norm": 0.014574599452316761,
+ "learning_rate": 0.0006,
+ "loss": 4.2850847244262695,
+ "step": 4835
+ },
+ {
+ "epoch": 67.16768558951965,
+ "grad_norm": 0.012676282785832882,
+ "learning_rate": 0.0006,
+ "loss": 4.297845840454102,
+ "step": 4836
+ },
+ {
+ "epoch": 67.18165938864628,
+ "grad_norm": 0.013948136940598488,
+ "learning_rate": 0.0006,
+ "loss": 4.324530601501465,
+ "step": 4837
+ },
+ {
+ "epoch": 67.19563318777293,
+ "grad_norm": 0.014195773750543594,
+ "learning_rate": 0.0006,
+ "loss": 4.31236457824707,
+ "step": 4838
+ },
+ {
+ "epoch": 67.20960698689956,
+ "grad_norm": 0.013631382025778294,
+ "learning_rate": 0.0006,
+ "loss": 4.086362838745117,
+ "step": 4839
+ },
+ {
+ "epoch": 67.2235807860262,
+ "grad_norm": 0.012798587791621685,
+ "learning_rate": 0.0006,
+ "loss": 4.328471660614014,
+ "step": 4840
+ },
+ {
+ "epoch": 67.23755458515284,
+ "grad_norm": 0.014443321153521538,
+ "learning_rate": 0.0006,
+ "loss": 4.278019905090332,
+ "step": 4841
+ },
+ {
+ "epoch": 67.25152838427948,
+ "grad_norm": 0.015355420298874378,
+ "learning_rate": 0.0006,
+ "loss": 4.333611011505127,
+ "step": 4842
+ },
+ {
+ "epoch": 67.26550218340611,
+ "grad_norm": 0.016487296670675278,
+ "learning_rate": 0.0006,
+ "loss": 4.351336479187012,
+ "step": 4843
+ },
+ {
+ "epoch": 67.27947598253274,
+ "grad_norm": 0.017422541975975037,
+ "learning_rate": 0.0006,
+ "loss": 4.386096477508545,
+ "step": 4844
+ },
+ {
+ "epoch": 67.29344978165939,
+ "grad_norm": 0.015790918841958046,
+ "learning_rate": 0.0006,
+ "loss": 4.301377773284912,
+ "step": 4845
+ },
+ {
+ "epoch": 67.30742358078602,
+ "grad_norm": 0.016427364200353622,
+ "learning_rate": 0.0006,
+ "loss": 4.2838945388793945,
+ "step": 4846
+ },
+ {
+ "epoch": 67.32139737991267,
+ "grad_norm": 0.017050761729478836,
+ "learning_rate": 0.0006,
+ "loss": 4.282417297363281,
+ "step": 4847
+ },
+ {
+ "epoch": 67.3353711790393,
+ "grad_norm": 0.015006253495812416,
+ "learning_rate": 0.0006,
+ "loss": 4.352236270904541,
+ "step": 4848
+ },
+ {
+ "epoch": 67.34934497816595,
+ "grad_norm": 0.012996627949178219,
+ "learning_rate": 0.0006,
+ "loss": 4.309242248535156,
+ "step": 4849
+ },
+ {
+ "epoch": 67.36331877729258,
+ "grad_norm": 0.01367615070194006,
+ "learning_rate": 0.0006,
+ "loss": 4.211801528930664,
+ "step": 4850
+ },
+ {
+ "epoch": 67.37729257641921,
+ "grad_norm": 0.014775963500142097,
+ "learning_rate": 0.0006,
+ "loss": 4.360859394073486,
+ "step": 4851
+ },
+ {
+ "epoch": 67.39126637554585,
+ "grad_norm": 0.015854304656386375,
+ "learning_rate": 0.0006,
+ "loss": 4.211830139160156,
+ "step": 4852
+ },
+ {
+ "epoch": 67.40524017467249,
+ "grad_norm": 0.016888894140720367,
+ "learning_rate": 0.0006,
+ "loss": 4.216549873352051,
+ "step": 4853
+ },
+ {
+ "epoch": 67.41921397379913,
+ "grad_norm": 0.0170894768089056,
+ "learning_rate": 0.0006,
+ "loss": 4.263486385345459,
+ "step": 4854
+ },
+ {
+ "epoch": 67.43318777292576,
+ "grad_norm": 0.015829868614673615,
+ "learning_rate": 0.0006,
+ "loss": 4.360196113586426,
+ "step": 4855
+ },
+ {
+ "epoch": 67.44716157205241,
+ "grad_norm": 0.015220202505588531,
+ "learning_rate": 0.0006,
+ "loss": 4.243794918060303,
+ "step": 4856
+ },
+ {
+ "epoch": 67.46113537117904,
+ "grad_norm": 0.01413388829678297,
+ "learning_rate": 0.0006,
+ "loss": 4.286624908447266,
+ "step": 4857
+ },
+ {
+ "epoch": 67.47510917030567,
+ "grad_norm": 0.01426626369357109,
+ "learning_rate": 0.0006,
+ "loss": 4.257316589355469,
+ "step": 4858
+ },
+ {
+ "epoch": 67.48908296943232,
+ "grad_norm": 0.015957845374941826,
+ "learning_rate": 0.0006,
+ "loss": 4.300443649291992,
+ "step": 4859
+ },
+ {
+ "epoch": 67.50305676855895,
+ "grad_norm": 0.015499051660299301,
+ "learning_rate": 0.0006,
+ "loss": 4.278162002563477,
+ "step": 4860
+ },
+ {
+ "epoch": 67.5170305676856,
+ "grad_norm": 0.015145624987781048,
+ "learning_rate": 0.0006,
+ "loss": 4.267950534820557,
+ "step": 4861
+ },
+ {
+ "epoch": 67.53100436681223,
+ "grad_norm": 0.014610803686082363,
+ "learning_rate": 0.0006,
+ "loss": 4.247499942779541,
+ "step": 4862
+ },
+ {
+ "epoch": 67.54497816593886,
+ "grad_norm": 0.014860909432172775,
+ "learning_rate": 0.0006,
+ "loss": 4.263200759887695,
+ "step": 4863
+ },
+ {
+ "epoch": 67.5589519650655,
+ "grad_norm": 0.01718984916806221,
+ "learning_rate": 0.0006,
+ "loss": 4.290019989013672,
+ "step": 4864
+ },
+ {
+ "epoch": 67.57292576419214,
+ "grad_norm": 0.01558963768184185,
+ "learning_rate": 0.0006,
+ "loss": 4.223816871643066,
+ "step": 4865
+ },
+ {
+ "epoch": 67.58689956331878,
+ "grad_norm": 0.014927188865840435,
+ "learning_rate": 0.0006,
+ "loss": 4.363586902618408,
+ "step": 4866
+ },
+ {
+ "epoch": 67.60087336244541,
+ "grad_norm": 0.01573595404624939,
+ "learning_rate": 0.0006,
+ "loss": 4.254765510559082,
+ "step": 4867
+ },
+ {
+ "epoch": 67.61484716157206,
+ "grad_norm": 0.015618929639458656,
+ "learning_rate": 0.0006,
+ "loss": 4.25072717666626,
+ "step": 4868
+ },
+ {
+ "epoch": 67.62882096069869,
+ "grad_norm": 0.01605590060353279,
+ "learning_rate": 0.0006,
+ "loss": 4.25208854675293,
+ "step": 4869
+ },
+ {
+ "epoch": 67.64279475982532,
+ "grad_norm": 0.01804601214826107,
+ "learning_rate": 0.0006,
+ "loss": 4.351354122161865,
+ "step": 4870
+ },
+ {
+ "epoch": 67.65676855895197,
+ "grad_norm": 0.020595047622919083,
+ "learning_rate": 0.0006,
+ "loss": 4.193732738494873,
+ "step": 4871
+ },
+ {
+ "epoch": 67.6707423580786,
+ "grad_norm": 0.023037966340780258,
+ "learning_rate": 0.0006,
+ "loss": 4.27754020690918,
+ "step": 4872
+ },
+ {
+ "epoch": 67.68471615720524,
+ "grad_norm": 0.020680895075201988,
+ "learning_rate": 0.0006,
+ "loss": 4.314730167388916,
+ "step": 4873
+ },
+ {
+ "epoch": 67.69868995633188,
+ "grad_norm": 0.01887434534728527,
+ "learning_rate": 0.0006,
+ "loss": 4.252305030822754,
+ "step": 4874
+ },
+ {
+ "epoch": 67.71266375545852,
+ "grad_norm": 0.016333596780896187,
+ "learning_rate": 0.0006,
+ "loss": 4.146123886108398,
+ "step": 4875
+ },
+ {
+ "epoch": 67.72663755458515,
+ "grad_norm": 0.015688810497522354,
+ "learning_rate": 0.0006,
+ "loss": 4.2814154624938965,
+ "step": 4876
+ },
+ {
+ "epoch": 67.74061135371178,
+ "grad_norm": 0.016270726919174194,
+ "learning_rate": 0.0006,
+ "loss": 4.323200225830078,
+ "step": 4877
+ },
+ {
+ "epoch": 67.75458515283843,
+ "grad_norm": 0.016824988648295403,
+ "learning_rate": 0.0006,
+ "loss": 4.269501686096191,
+ "step": 4878
+ },
+ {
+ "epoch": 67.76855895196506,
+ "grad_norm": 0.01730562001466751,
+ "learning_rate": 0.0006,
+ "loss": 4.187993049621582,
+ "step": 4879
+ },
+ {
+ "epoch": 67.78253275109171,
+ "grad_norm": 0.01546655036509037,
+ "learning_rate": 0.0006,
+ "loss": 4.340415000915527,
+ "step": 4880
+ },
+ {
+ "epoch": 67.79650655021834,
+ "grad_norm": 0.014173958450555801,
+ "learning_rate": 0.0006,
+ "loss": 4.217419624328613,
+ "step": 4881
+ },
+ {
+ "epoch": 67.81048034934499,
+ "grad_norm": 0.016810474917292595,
+ "learning_rate": 0.0006,
+ "loss": 4.352302074432373,
+ "step": 4882
+ },
+ {
+ "epoch": 67.82445414847162,
+ "grad_norm": 0.0175373163074255,
+ "learning_rate": 0.0006,
+ "loss": 4.20377254486084,
+ "step": 4883
+ },
+ {
+ "epoch": 67.83842794759825,
+ "grad_norm": 0.015734676271677017,
+ "learning_rate": 0.0006,
+ "loss": 4.147495746612549,
+ "step": 4884
+ },
+ {
+ "epoch": 67.8524017467249,
+ "grad_norm": 0.015091817826032639,
+ "learning_rate": 0.0006,
+ "loss": 4.272252082824707,
+ "step": 4885
+ },
+ {
+ "epoch": 67.86637554585153,
+ "grad_norm": 0.015000219456851482,
+ "learning_rate": 0.0006,
+ "loss": 4.302773952484131,
+ "step": 4886
+ },
+ {
+ "epoch": 67.88034934497817,
+ "grad_norm": 0.013827184215188026,
+ "learning_rate": 0.0006,
+ "loss": 4.337242126464844,
+ "step": 4887
+ },
+ {
+ "epoch": 67.8943231441048,
+ "grad_norm": 0.015308346599340439,
+ "learning_rate": 0.0006,
+ "loss": 4.140895366668701,
+ "step": 4888
+ },
+ {
+ "epoch": 67.90829694323143,
+ "grad_norm": 0.015356178395450115,
+ "learning_rate": 0.0006,
+ "loss": 4.312541484832764,
+ "step": 4889
+ },
+ {
+ "epoch": 67.92227074235808,
+ "grad_norm": 0.0137353939935565,
+ "learning_rate": 0.0006,
+ "loss": 4.173184394836426,
+ "step": 4890
+ },
+ {
+ "epoch": 67.93624454148471,
+ "grad_norm": 0.01332368515431881,
+ "learning_rate": 0.0006,
+ "loss": 4.191199779510498,
+ "step": 4891
+ },
+ {
+ "epoch": 67.95021834061136,
+ "grad_norm": 0.01488130260258913,
+ "learning_rate": 0.0006,
+ "loss": 4.331752300262451,
+ "step": 4892
+ },
+ {
+ "epoch": 67.96419213973799,
+ "grad_norm": 0.015449733473360538,
+ "learning_rate": 0.0006,
+ "loss": 4.2385454177856445,
+ "step": 4893
+ },
+ {
+ "epoch": 67.97816593886463,
+ "grad_norm": 0.01556103490293026,
+ "learning_rate": 0.0006,
+ "loss": 4.262353897094727,
+ "step": 4894
+ },
+ {
+ "epoch": 67.99213973799127,
+ "grad_norm": 0.014613849110901356,
+ "learning_rate": 0.0006,
+ "loss": 4.223945617675781,
+ "step": 4895
+ },
+ {
+ "epoch": 68.0,
+ "grad_norm": 0.015323462896049023,
+ "learning_rate": 0.0006,
+ "loss": 4.317168712615967,
+ "step": 4896
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 28800,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 400,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": true,
+ "should_log": false,
+ "should_save": false,
+ "should_training_stop": false
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 2.0873370857059123e+18,
+ "train_batch_size": 8,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4896/training_args.bin b/runs/l2r50-i2-fulle-lm/checkpoint-4896/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..cad232831c222cc7f8b938b8bbe42f61e7d89c3a
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-4896/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:f7c40e3bfa2f5ac1450ca1243e3f2a895d44a06ed5995e6aefd97f78d229d8c8
+size 4792
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-5184/chat_template.jinja b/runs/l2r50-i2-fulle-lm/checkpoint-5184/chat_template.jinja
new file mode 100644
index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-5184/chat_template.jinja
@@ -0,0 +1,85 @@
+{%- if tools %}
+ {{- '<|im_start|>system\n' }}
+ {%- if messages[0].role == 'system' %}
+ {{- messages[0].content + '\n\n' }}
+ {%- endif %}
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }}
+ {%- for tool in tools %}
+ {{- "\n" }}
+ {{- tool | tojson }}
+ {%- endfor %}
+ {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }}
+{%- else %}
+ {%- if messages[0].role == 'system' %}
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
+ {%- endif %}
+{%- endif %}
+{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
+{%- for message in messages[::-1] %}
+ {%- set index = (messages|length - 1) - loop.index0 %}
+ {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %}
+ {%- set ns.multi_step_tool = false %}
+ {%- set ns.last_query_index = index %}
+ {%- endif %}
+{%- endfor %}
+{%- for message in messages %}
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
+ {%- elif message.role == "assistant" %}
+ {%- set content = message.content %}
+ {%- set reasoning_content = '' %}
+ {%- if message.reasoning_content is defined and message.reasoning_content is not none %}
+ {%- set reasoning_content = message.reasoning_content %}
+ {%- else %}
+ {%- if '' in message.content %}
+ {%- set content = message.content.split('')[-1].lstrip('\n') %}
+ {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %}
+ {%- endif %}
+ {%- endif %}
+ {%- if loop.index0 > ns.last_query_index %}
+ {%- if loop.last or (not loop.last and reasoning_content) %}
+ {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- if message.tool_calls %}
+ {%- for tool_call in message.tool_calls %}
+ {%- if (loop.first and content) or (not loop.first) %}
+ {{- '\n' }}
+ {%- endif %}
+ {%- if tool_call.function %}
+ {%- set tool_call = tool_call.function %}
+ {%- endif %}
+ {{- '\n{"name": "' }}
+ {{- tool_call.name }}
+ {{- '", "arguments": ' }}
+ {%- if tool_call.arguments is string %}
+ {{- tool_call.arguments }}
+ {%- else %}
+ {{- tool_call.arguments | tojson }}
+ {%- endif %}
+ {{- '}\n' }}
+ {%- endfor %}
+ {%- endif %}
+ {{- '<|im_end|>\n' }}
+ {%- elif message.role == "tool" %}
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
+ {{- '<|im_start|>user' }}
+ {%- endif %}
+ {{- '\n\n' }}
+ {{- message.content }}
+ {{- '\n' }}
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
+ {{- '<|im_end|>\n' }}
+ {%- endif %}
+ {%- endif %}
+{%- endfor %}
+{%- if add_generation_prompt %}
+ {{- '<|im_start|>assistant\n' }}
+ {%- if enable_thinking is defined and enable_thinking is false %}
+ {{- '\n\n\n\n' }}
+ {%- endif %}
+{%- endif %}
\ No newline at end of file
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-5184/config.json b/runs/l2r50-i2-fulle-lm/checkpoint-5184/config.json
new file mode 100644
index 0000000000000000000000000000000000000000..22f250953c6ce8fc91ee4411856dd73041a415e2
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-5184/config.json
@@ -0,0 +1,32 @@
+{
+ "architectures": [
+ "LlamaForCausalLM"
+ ],
+ "attention_bias": false,
+ "attention_dropout": 0.0,
+ "bos_token_id": null,
+ "dtype": "float32",
+ "eos_token_id": 151645,
+ "head_dim": 128,
+ "hidden_act": "silu",
+ "hidden_size": 512,
+ "initializer_range": 0.02,
+ "intermediate_size": 1536,
+ "max_position_embeddings": 2048,
+ "mlp_bias": false,
+ "model_type": "llama",
+ "num_attention_heads": 4,
+ "num_hidden_layers": 20,
+ "num_key_value_heads": 4,
+ "pad_token_id": 151645,
+ "pretraining_tp": 1,
+ "rms_norm_eps": 1e-06,
+ "rope_parameters": {
+ "rope_theta": 10000.0,
+ "rope_type": "default"
+ },
+ "tie_word_embeddings": true,
+ "transformers_version": "5.5.0",
+ "use_cache": false,
+ "vocab_size": 151673
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-5184/generation_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-5184/generation_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-5184/generation_config.json
@@ -0,0 +1,11 @@
+{
+ "_from_model_config": true,
+ "eos_token_id": [
+ 151645
+ ],
+ "output_attentions": false,
+ "output_hidden_states": false,
+ "pad_token_id": 151645,
+ "transformers_version": "5.5.0",
+ "use_cache": true
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-5184/model.safetensors b/runs/l2r50-i2-fulle-lm/checkpoint-5184/model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..3d6964b90321ecf947102216d4e9befa10b7992d
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-5184/model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:bdd0baa43cae174cd7ab844b8496cdaad7bac65f887d1fd0c1428e0137c373f7
+size 583360328
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-5184/optimizer.pt b/runs/l2r50-i2-fulle-lm/checkpoint-5184/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..dc2838fe5fc58605fa80a338d826ad594ac2feae
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-5184/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:679c528f3a002006294bb8deab8424c8ccf33503816d9605c2ae50708cb95ea4
+size 1166833530
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-5184/rng_state_0.pth b/runs/l2r50-i2-fulle-lm/checkpoint-5184/rng_state_0.pth
new file mode 100644
index 0000000000000000000000000000000000000000..8330111548d3ceb55eec2ce0d100fb6911e3f119
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-5184/rng_state_0.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:22a799ebf18d5949b0d687c2fca394611640bd54f2edeab8c785f7bfd8ca1210
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-5184/rng_state_1.pth b/runs/l2r50-i2-fulle-lm/checkpoint-5184/rng_state_1.pth
new file mode 100644
index 0000000000000000000000000000000000000000..94c164b966d4b94a3d97959474303b95d0335ab3
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-5184/rng_state_1.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:b86e8ab98f075cb83d5d0c3f8da246a7181ec97b387d2d0b3b2596f3a4026908
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-5184/rng_state_2.pth b/runs/l2r50-i2-fulle-lm/checkpoint-5184/rng_state_2.pth
new file mode 100644
index 0000000000000000000000000000000000000000..f938712cd32e74263ab38e7e4e77dc3a52a35a56
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-5184/rng_state_2.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:7305ed0dd577328473d6a79b96b2c5846f83995e5a3b25f37615248aa85284b5
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-5184/rng_state_3.pth b/runs/l2r50-i2-fulle-lm/checkpoint-5184/rng_state_3.pth
new file mode 100644
index 0000000000000000000000000000000000000000..b29fdbf1cbfdea70acf84092a9920a959ea9ee42
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-5184/rng_state_3.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:f01527c462cf991ac25914ec71ab83c5feaff9e269428a30e5c3f3dcc48b907e
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-5184/scheduler.pt b/runs/l2r50-i2-fulle-lm/checkpoint-5184/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..87c25f34fb9c3e31517de12eec62aa55b15b7f35
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-5184/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:f27f1d0b18f28ca972b7026932ac402d35af3b241ee72040352067f9ee1e6988
+size 1064
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-5184/tokenizer.json b/runs/l2r50-i2-fulle-lm/checkpoint-5184/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..53786016f070e11184813f3c83101dc5e93e392e
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-5184/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:c3c63fda1646a555448e8f03662542e04b11d103721319e4d9b267d7da03ceb8
+size 11424463
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-5184/tokenizer_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-5184/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..98e2fb9b2cc54785dc9d871b8e503f8b8f2defe7
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-5184/tokenizer_config.json
@@ -0,0 +1,20 @@
+{
+ "add_prefix_space": false,
+ "backend": "tokenizers",
+ "bos_token": null,
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "<|im_end|>",
+ "errors": "replace",
+ "extra_special_tokens": [
+ "<|l2r_pred|>",
+ "<|r2l_pred|>",
+ "<|next_1_pred|>",
+ "<|next_2_pred|>"
+ ],
+ "is_local": false,
+ "model_max_length": 131072,
+ "pad_token": "<|im_end|>",
+ "split_special_tokens": false,
+ "tokenizer_class": "Qwen2Tokenizer",
+ "unk_token": null
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-5184/trainer_state.json b/runs/l2r50-i2-fulle-lm/checkpoint-5184/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..12afb136dca88a2b936bd3342d80e61693cebeb9
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-5184/trainer_state.json
@@ -0,0 +1,36890 @@
+{
+ "best_global_step": null,
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 72.0,
+ "eval_steps": 500,
+ "global_step": 5184,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.013973799126637555,
+ "grad_norm": 0.1739506870508194,
+ "learning_rate": 0.0,
+ "loss": 11.99299430847168,
+ "step": 1
+ },
+ {
+ "epoch": 0.02794759825327511,
+ "grad_norm": 0.17216597497463226,
+ "learning_rate": 5.999999999999999e-06,
+ "loss": 11.988704681396484,
+ "step": 2
+ },
+ {
+ "epoch": 0.04192139737991266,
+ "grad_norm": 0.17459715902805328,
+ "learning_rate": 1.1999999999999999e-05,
+ "loss": 11.945581436157227,
+ "step": 3
+ },
+ {
+ "epoch": 0.05589519650655022,
+ "grad_norm": 0.17577414214611053,
+ "learning_rate": 1.7999999999999997e-05,
+ "loss": 11.859772682189941,
+ "step": 4
+ },
+ {
+ "epoch": 0.06986899563318777,
+ "grad_norm": 0.16163651645183563,
+ "learning_rate": 2.3999999999999997e-05,
+ "loss": 11.744415283203125,
+ "step": 5
+ },
+ {
+ "epoch": 0.08384279475982533,
+ "grad_norm": 0.14409932494163513,
+ "learning_rate": 2.9999999999999997e-05,
+ "loss": 11.618099212646484,
+ "step": 6
+ },
+ {
+ "epoch": 0.09781659388646288,
+ "grad_norm": 0.13079848885536194,
+ "learning_rate": 3.5999999999999994e-05,
+ "loss": 11.505935668945312,
+ "step": 7
+ },
+ {
+ "epoch": 0.11179039301310044,
+ "grad_norm": 0.11964613944292068,
+ "learning_rate": 4.2e-05,
+ "loss": 11.416532516479492,
+ "step": 8
+ },
+ {
+ "epoch": 0.125764192139738,
+ "grad_norm": 0.11345142871141434,
+ "learning_rate": 4.7999999999999994e-05,
+ "loss": 11.33206558227539,
+ "step": 9
+ },
+ {
+ "epoch": 0.13973799126637554,
+ "grad_norm": 0.11033952981233597,
+ "learning_rate": 5.399999999999999e-05,
+ "loss": 11.259052276611328,
+ "step": 10
+ },
+ {
+ "epoch": 0.1537117903930131,
+ "grad_norm": 0.10776489973068237,
+ "learning_rate": 5.9999999999999995e-05,
+ "loss": 11.203384399414062,
+ "step": 11
+ },
+ {
+ "epoch": 0.16768558951965065,
+ "grad_norm": 0.10721946507692337,
+ "learning_rate": 6.599999999999999e-05,
+ "loss": 11.153295516967773,
+ "step": 12
+ },
+ {
+ "epoch": 0.18165938864628822,
+ "grad_norm": 0.10673430562019348,
+ "learning_rate": 7.199999999999999e-05,
+ "loss": 11.110200881958008,
+ "step": 13
+ },
+ {
+ "epoch": 0.19563318777292577,
+ "grad_norm": 0.10630254447460175,
+ "learning_rate": 7.8e-05,
+ "loss": 11.067477226257324,
+ "step": 14
+ },
+ {
+ "epoch": 0.2096069868995633,
+ "grad_norm": 0.10631615668535233,
+ "learning_rate": 8.4e-05,
+ "loss": 11.022941589355469,
+ "step": 15
+ },
+ {
+ "epoch": 0.22358078602620088,
+ "grad_norm": 0.10609443485736847,
+ "learning_rate": 8.999999999999999e-05,
+ "loss": 10.970746040344238,
+ "step": 16
+ },
+ {
+ "epoch": 0.23755458515283842,
+ "grad_norm": 0.10648232698440552,
+ "learning_rate": 9.599999999999999e-05,
+ "loss": 10.91267204284668,
+ "step": 17
+ },
+ {
+ "epoch": 0.251528384279476,
+ "grad_norm": 0.10584589838981628,
+ "learning_rate": 0.000102,
+ "loss": 10.854547500610352,
+ "step": 18
+ },
+ {
+ "epoch": 0.26550218340611353,
+ "grad_norm": 0.10591720044612885,
+ "learning_rate": 0.00010799999999999998,
+ "loss": 10.788917541503906,
+ "step": 19
+ },
+ {
+ "epoch": 0.2794759825327511,
+ "grad_norm": 0.10611672699451447,
+ "learning_rate": 0.00011399999999999999,
+ "loss": 10.717556953430176,
+ "step": 20
+ },
+ {
+ "epoch": 0.2934497816593886,
+ "grad_norm": 0.10455359518527985,
+ "learning_rate": 0.00011999999999999999,
+ "loss": 10.650186538696289,
+ "step": 21
+ },
+ {
+ "epoch": 0.3074235807860262,
+ "grad_norm": 0.1047525629401207,
+ "learning_rate": 0.00012599999999999997,
+ "loss": 10.56808090209961,
+ "step": 22
+ },
+ {
+ "epoch": 0.32139737991266376,
+ "grad_norm": 0.10511605441570282,
+ "learning_rate": 0.00013199999999999998,
+ "loss": 10.484028816223145,
+ "step": 23
+ },
+ {
+ "epoch": 0.3353711790393013,
+ "grad_norm": 0.10467177629470825,
+ "learning_rate": 0.000138,
+ "loss": 10.399947166442871,
+ "step": 24
+ },
+ {
+ "epoch": 0.34934497816593885,
+ "grad_norm": 0.10432136058807373,
+ "learning_rate": 0.00014399999999999998,
+ "loss": 10.312714576721191,
+ "step": 25
+ },
+ {
+ "epoch": 0.36331877729257644,
+ "grad_norm": 0.10527262091636658,
+ "learning_rate": 0.00015,
+ "loss": 10.212261199951172,
+ "step": 26
+ },
+ {
+ "epoch": 0.377292576419214,
+ "grad_norm": 0.103610560297966,
+ "learning_rate": 0.000156,
+ "loss": 10.132034301757812,
+ "step": 27
+ },
+ {
+ "epoch": 0.39126637554585153,
+ "grad_norm": 0.10553059726953506,
+ "learning_rate": 0.000162,
+ "loss": 10.020434379577637,
+ "step": 28
+ },
+ {
+ "epoch": 0.4052401746724891,
+ "grad_norm": 0.1037299782037735,
+ "learning_rate": 0.000168,
+ "loss": 9.934466361999512,
+ "step": 29
+ },
+ {
+ "epoch": 0.4192139737991266,
+ "grad_norm": 0.1042739674448967,
+ "learning_rate": 0.00017399999999999997,
+ "loss": 9.822539329528809,
+ "step": 30
+ },
+ {
+ "epoch": 0.4331877729257642,
+ "grad_norm": 0.10300865024328232,
+ "learning_rate": 0.00017999999999999998,
+ "loss": 9.734994888305664,
+ "step": 31
+ },
+ {
+ "epoch": 0.44716157205240176,
+ "grad_norm": 0.10291654616594315,
+ "learning_rate": 0.000186,
+ "loss": 9.624505996704102,
+ "step": 32
+ },
+ {
+ "epoch": 0.4611353711790393,
+ "grad_norm": 0.10210851579904556,
+ "learning_rate": 0.00019199999999999998,
+ "loss": 9.518205642700195,
+ "step": 33
+ },
+ {
+ "epoch": 0.47510917030567684,
+ "grad_norm": 0.10222529619932175,
+ "learning_rate": 0.000198,
+ "loss": 9.412222862243652,
+ "step": 34
+ },
+ {
+ "epoch": 0.4890829694323144,
+ "grad_norm": 0.09992019087076187,
+ "learning_rate": 0.000204,
+ "loss": 9.326566696166992,
+ "step": 35
+ },
+ {
+ "epoch": 0.503056768558952,
+ "grad_norm": 0.09931084513664246,
+ "learning_rate": 0.00020999999999999998,
+ "loss": 9.210111618041992,
+ "step": 36
+ },
+ {
+ "epoch": 0.5170305676855895,
+ "grad_norm": 0.09928256273269653,
+ "learning_rate": 0.00021599999999999996,
+ "loss": 9.100493431091309,
+ "step": 37
+ },
+ {
+ "epoch": 0.5310043668122271,
+ "grad_norm": 0.09843816608190536,
+ "learning_rate": 0.00022199999999999998,
+ "loss": 8.996543884277344,
+ "step": 38
+ },
+ {
+ "epoch": 0.5449781659388646,
+ "grad_norm": 0.09825374186038971,
+ "learning_rate": 0.00022799999999999999,
+ "loss": 8.888991355895996,
+ "step": 39
+ },
+ {
+ "epoch": 0.5589519650655022,
+ "grad_norm": 0.09602385759353638,
+ "learning_rate": 0.000234,
+ "loss": 8.802907943725586,
+ "step": 40
+ },
+ {
+ "epoch": 0.5729257641921397,
+ "grad_norm": 0.09414064139127731,
+ "learning_rate": 0.00023999999999999998,
+ "loss": 8.715423583984375,
+ "step": 41
+ },
+ {
+ "epoch": 0.5868995633187772,
+ "grad_norm": 0.09273893386125565,
+ "learning_rate": 0.00024599999999999996,
+ "loss": 8.61374282836914,
+ "step": 42
+ },
+ {
+ "epoch": 0.6008733624454149,
+ "grad_norm": 0.09079599380493164,
+ "learning_rate": 0.00025199999999999995,
+ "loss": 8.502904891967773,
+ "step": 43
+ },
+ {
+ "epoch": 0.6148471615720524,
+ "grad_norm": 0.08605825901031494,
+ "learning_rate": 0.000258,
+ "loss": 8.453008651733398,
+ "step": 44
+ },
+ {
+ "epoch": 0.62882096069869,
+ "grad_norm": 0.08338996022939682,
+ "learning_rate": 0.00026399999999999997,
+ "loss": 8.365163803100586,
+ "step": 45
+ },
+ {
+ "epoch": 0.6427947598253275,
+ "grad_norm": 0.08150867372751236,
+ "learning_rate": 0.00027,
+ "loss": 8.285372734069824,
+ "step": 46
+ },
+ {
+ "epoch": 0.6567685589519651,
+ "grad_norm": 0.07794807851314545,
+ "learning_rate": 0.000276,
+ "loss": 8.185811996459961,
+ "step": 47
+ },
+ {
+ "epoch": 0.6707423580786026,
+ "grad_norm": 0.07592720538377762,
+ "learning_rate": 0.00028199999999999997,
+ "loss": 8.098482131958008,
+ "step": 48
+ },
+ {
+ "epoch": 0.6847161572052402,
+ "grad_norm": 0.0709897130727768,
+ "learning_rate": 0.00028799999999999995,
+ "loss": 8.042091369628906,
+ "step": 49
+ },
+ {
+ "epoch": 0.6986899563318777,
+ "grad_norm": 0.06582305580377579,
+ "learning_rate": 0.000294,
+ "loss": 7.972844123840332,
+ "step": 50
+ },
+ {
+ "epoch": 0.7126637554585152,
+ "grad_norm": 0.059452351182699203,
+ "learning_rate": 0.0003,
+ "loss": 7.924579620361328,
+ "step": 51
+ },
+ {
+ "epoch": 0.7266375545851529,
+ "grad_norm": 0.052426718175411224,
+ "learning_rate": 0.00030599999999999996,
+ "loss": 7.876336097717285,
+ "step": 52
+ },
+ {
+ "epoch": 0.7406113537117904,
+ "grad_norm": 0.04768150672316551,
+ "learning_rate": 0.000312,
+ "loss": 7.8102707862854,
+ "step": 53
+ },
+ {
+ "epoch": 0.754585152838428,
+ "grad_norm": 0.03970075398683548,
+ "learning_rate": 0.000318,
+ "loss": 7.791913986206055,
+ "step": 54
+ },
+ {
+ "epoch": 0.7685589519650655,
+ "grad_norm": 0.03559752553701401,
+ "learning_rate": 0.000324,
+ "loss": 7.756960868835449,
+ "step": 55
+ },
+ {
+ "epoch": 0.7825327510917031,
+ "grad_norm": 0.029321353882551193,
+ "learning_rate": 0.00033,
+ "loss": 7.724878787994385,
+ "step": 56
+ },
+ {
+ "epoch": 0.7965065502183406,
+ "grad_norm": 0.023485427722334862,
+ "learning_rate": 0.000336,
+ "loss": 7.72553014755249,
+ "step": 57
+ },
+ {
+ "epoch": 0.8104803493449781,
+ "grad_norm": 0.019319474697113037,
+ "learning_rate": 0.00034199999999999996,
+ "loss": 7.709493160247803,
+ "step": 58
+ },
+ {
+ "epoch": 0.8244541484716157,
+ "grad_norm": 0.016705691814422607,
+ "learning_rate": 0.00034799999999999995,
+ "loss": 7.6873297691345215,
+ "step": 59
+ },
+ {
+ "epoch": 0.8384279475982532,
+ "grad_norm": 0.017667734995484352,
+ "learning_rate": 0.00035399999999999993,
+ "loss": 7.677059173583984,
+ "step": 60
+ },
+ {
+ "epoch": 0.8524017467248908,
+ "grad_norm": 0.01671992614865303,
+ "learning_rate": 0.00035999999999999997,
+ "loss": 7.643512725830078,
+ "step": 61
+ },
+ {
+ "epoch": 0.8663755458515284,
+ "grad_norm": 0.012288172729313374,
+ "learning_rate": 0.00036599999999999995,
+ "loss": 7.658343315124512,
+ "step": 62
+ },
+ {
+ "epoch": 0.880349344978166,
+ "grad_norm": 0.012004776857793331,
+ "learning_rate": 0.000372,
+ "loss": 7.649759292602539,
+ "step": 63
+ },
+ {
+ "epoch": 0.8943231441048035,
+ "grad_norm": 0.012487445957958698,
+ "learning_rate": 0.00037799999999999997,
+ "loss": 7.631503105163574,
+ "step": 64
+ },
+ {
+ "epoch": 0.9082969432314411,
+ "grad_norm": 0.01489241048693657,
+ "learning_rate": 0.00038399999999999996,
+ "loss": 7.647110462188721,
+ "step": 65
+ },
+ {
+ "epoch": 0.9222707423580786,
+ "grad_norm": 0.01428980939090252,
+ "learning_rate": 0.00039,
+ "loss": 7.636983871459961,
+ "step": 66
+ },
+ {
+ "epoch": 0.9362445414847161,
+ "grad_norm": 0.010711144655942917,
+ "learning_rate": 0.000396,
+ "loss": 7.621130466461182,
+ "step": 67
+ },
+ {
+ "epoch": 0.9502183406113537,
+ "grad_norm": 0.010178886353969574,
+ "learning_rate": 0.000402,
+ "loss": 7.593924522399902,
+ "step": 68
+ },
+ {
+ "epoch": 0.9641921397379912,
+ "grad_norm": 0.010604874230921268,
+ "learning_rate": 0.000408,
+ "loss": 7.6015496253967285,
+ "step": 69
+ },
+ {
+ "epoch": 0.9781659388646288,
+ "grad_norm": 0.009751426056027412,
+ "learning_rate": 0.0004139999999999999,
+ "loss": 7.571964263916016,
+ "step": 70
+ },
+ {
+ "epoch": 0.9921397379912664,
+ "grad_norm": 0.009816481731832027,
+ "learning_rate": 0.00041999999999999996,
+ "loss": 7.57344913482666,
+ "step": 71
+ },
+ {
+ "epoch": 1.0,
+ "grad_norm": 0.008665637113153934,
+ "learning_rate": 0.00042599999999999995,
+ "loss": 7.579050064086914,
+ "step": 72
+ },
+ {
+ "epoch": 1.0,
+ "eval_loss": 7.581204891204834,
+ "eval_runtime": 59.0091,
+ "eval_samples_per_second": 41.383,
+ "eval_steps_per_second": 1.305,
+ "step": 72
+ },
+ {
+ "epoch": 1.0139737991266375,
+ "grad_norm": 0.009684201329946518,
+ "learning_rate": 0.00043199999999999993,
+ "loss": 7.568594932556152,
+ "step": 73
+ },
+ {
+ "epoch": 1.027947598253275,
+ "grad_norm": 0.025783710181713104,
+ "learning_rate": 0.00043799999999999997,
+ "loss": 7.578948497772217,
+ "step": 74
+ },
+ {
+ "epoch": 1.0419213973799126,
+ "grad_norm": 0.0327819399535656,
+ "learning_rate": 0.00044399999999999995,
+ "loss": 7.548173904418945,
+ "step": 75
+ },
+ {
+ "epoch": 1.0558951965065502,
+ "grad_norm": 0.021604498848319054,
+ "learning_rate": 0.00045,
+ "loss": 7.552953720092773,
+ "step": 76
+ },
+ {
+ "epoch": 1.0698689956331877,
+ "grad_norm": 0.010230105370283127,
+ "learning_rate": 0.00045599999999999997,
+ "loss": 7.5213212966918945,
+ "step": 77
+ },
+ {
+ "epoch": 1.0838427947598253,
+ "grad_norm": 0.023315589874982834,
+ "learning_rate": 0.00046199999999999995,
+ "loss": 7.501640319824219,
+ "step": 78
+ },
+ {
+ "epoch": 1.0978165938864628,
+ "grad_norm": 0.01109944935888052,
+ "learning_rate": 0.000468,
+ "loss": 7.539519786834717,
+ "step": 79
+ },
+ {
+ "epoch": 1.1117903930131003,
+ "grad_norm": 0.026093758642673492,
+ "learning_rate": 0.000474,
+ "loss": 7.525928974151611,
+ "step": 80
+ },
+ {
+ "epoch": 1.125764192139738,
+ "grad_norm": 0.03274895250797272,
+ "learning_rate": 0.00047999999999999996,
+ "loss": 7.535602569580078,
+ "step": 81
+ },
+ {
+ "epoch": 1.1397379912663754,
+ "grad_norm": 0.01543602254241705,
+ "learning_rate": 0.000486,
+ "loss": 7.513683795928955,
+ "step": 82
+ },
+ {
+ "epoch": 1.1537117903930132,
+ "grad_norm": 0.028377249836921692,
+ "learning_rate": 0.0004919999999999999,
+ "loss": 7.49384880065918,
+ "step": 83
+ },
+ {
+ "epoch": 1.1676855895196507,
+ "grad_norm": 0.04121481627225876,
+ "learning_rate": 0.000498,
+ "loss": 7.488447189331055,
+ "step": 84
+ },
+ {
+ "epoch": 1.1816593886462883,
+ "grad_norm": 0.013282302767038345,
+ "learning_rate": 0.0005039999999999999,
+ "loss": 7.463010787963867,
+ "step": 85
+ },
+ {
+ "epoch": 1.1956331877729258,
+ "grad_norm": 0.03200364485383034,
+ "learning_rate": 0.0005099999999999999,
+ "loss": 7.473515033721924,
+ "step": 86
+ },
+ {
+ "epoch": 1.2096069868995634,
+ "grad_norm": 0.020768029615283012,
+ "learning_rate": 0.000516,
+ "loss": 7.464688777923584,
+ "step": 87
+ },
+ {
+ "epoch": 1.223580786026201,
+ "grad_norm": 0.021715788170695305,
+ "learning_rate": 0.000522,
+ "loss": 7.4835028648376465,
+ "step": 88
+ },
+ {
+ "epoch": 1.2375545851528384,
+ "grad_norm": 0.026255754753947258,
+ "learning_rate": 0.0005279999999999999,
+ "loss": 7.443392753601074,
+ "step": 89
+ },
+ {
+ "epoch": 1.251528384279476,
+ "grad_norm": 0.022325951606035233,
+ "learning_rate": 0.000534,
+ "loss": 7.457373142242432,
+ "step": 90
+ },
+ {
+ "epoch": 1.2655021834061135,
+ "grad_norm": 0.009270387701690197,
+ "learning_rate": 0.00054,
+ "loss": 7.455986976623535,
+ "step": 91
+ },
+ {
+ "epoch": 1.279475982532751,
+ "grad_norm": 0.02369544468820095,
+ "learning_rate": 0.0005459999999999999,
+ "loss": 7.442216873168945,
+ "step": 92
+ },
+ {
+ "epoch": 1.2934497816593886,
+ "grad_norm": 0.008474506437778473,
+ "learning_rate": 0.000552,
+ "loss": 7.4294915199279785,
+ "step": 93
+ },
+ {
+ "epoch": 1.3074235807860262,
+ "grad_norm": 0.017043249681591988,
+ "learning_rate": 0.000558,
+ "loss": 7.406932830810547,
+ "step": 94
+ },
+ {
+ "epoch": 1.3213973799126637,
+ "grad_norm": 0.011813412420451641,
+ "learning_rate": 0.0005639999999999999,
+ "loss": 7.415900230407715,
+ "step": 95
+ },
+ {
+ "epoch": 1.3353711790393012,
+ "grad_norm": 0.01086883433163166,
+ "learning_rate": 0.00057,
+ "loss": 7.452998638153076,
+ "step": 96
+ },
+ {
+ "epoch": 1.3493449781659388,
+ "grad_norm": 0.014700951054692268,
+ "learning_rate": 0.0005759999999999999,
+ "loss": 7.410249710083008,
+ "step": 97
+ },
+ {
+ "epoch": 1.3633187772925766,
+ "grad_norm": 0.014512976631522179,
+ "learning_rate": 0.0005819999999999999,
+ "loss": 7.429132461547852,
+ "step": 98
+ },
+ {
+ "epoch": 1.3772925764192139,
+ "grad_norm": 0.012519920244812965,
+ "learning_rate": 0.000588,
+ "loss": 7.430817604064941,
+ "step": 99
+ },
+ {
+ "epoch": 1.3912663755458516,
+ "grad_norm": 0.007132664322853088,
+ "learning_rate": 0.0005939999999999999,
+ "loss": 7.405234336853027,
+ "step": 100
+ },
+ {
+ "epoch": 1.405240174672489,
+ "grad_norm": 0.019489388912916183,
+ "learning_rate": 0.0006,
+ "loss": 7.427791595458984,
+ "step": 101
+ },
+ {
+ "epoch": 1.4192139737991267,
+ "grad_norm": 0.027252664789557457,
+ "learning_rate": 0.0006,
+ "loss": 7.41761589050293,
+ "step": 102
+ },
+ {
+ "epoch": 1.4331877729257643,
+ "grad_norm": 0.02501637488603592,
+ "learning_rate": 0.0006,
+ "loss": 7.416735649108887,
+ "step": 103
+ },
+ {
+ "epoch": 1.4471615720524018,
+ "grad_norm": 0.00989772193133831,
+ "learning_rate": 0.0006,
+ "loss": 7.3756632804870605,
+ "step": 104
+ },
+ {
+ "epoch": 1.4611353711790394,
+ "grad_norm": 0.01622949168086052,
+ "learning_rate": 0.0006,
+ "loss": 7.400811672210693,
+ "step": 105
+ },
+ {
+ "epoch": 1.475109170305677,
+ "grad_norm": 0.012407763861119747,
+ "learning_rate": 0.0006,
+ "loss": 7.37675666809082,
+ "step": 106
+ },
+ {
+ "epoch": 1.4890829694323144,
+ "grad_norm": 0.010353565216064453,
+ "learning_rate": 0.0006,
+ "loss": 7.4013142585754395,
+ "step": 107
+ },
+ {
+ "epoch": 1.503056768558952,
+ "grad_norm": 0.0120553532615304,
+ "learning_rate": 0.0006,
+ "loss": 7.379873275756836,
+ "step": 108
+ },
+ {
+ "epoch": 1.5170305676855895,
+ "grad_norm": 0.01610225811600685,
+ "learning_rate": 0.0006,
+ "loss": 7.371631145477295,
+ "step": 109
+ },
+ {
+ "epoch": 1.531004366812227,
+ "grad_norm": 0.01159879844635725,
+ "learning_rate": 0.0006,
+ "loss": 7.389007568359375,
+ "step": 110
+ },
+ {
+ "epoch": 1.5449781659388646,
+ "grad_norm": 0.008013816550374031,
+ "learning_rate": 0.0006,
+ "loss": 7.395920753479004,
+ "step": 111
+ },
+ {
+ "epoch": 1.5589519650655022,
+ "grad_norm": 0.016334809362888336,
+ "learning_rate": 0.0006,
+ "loss": 7.393121719360352,
+ "step": 112
+ },
+ {
+ "epoch": 1.5729257641921397,
+ "grad_norm": 0.018821721896529198,
+ "learning_rate": 0.0006,
+ "loss": 7.367543697357178,
+ "step": 113
+ },
+ {
+ "epoch": 1.5868995633187772,
+ "grad_norm": 0.014088994823396206,
+ "learning_rate": 0.0006,
+ "loss": 7.385537624359131,
+ "step": 114
+ },
+ {
+ "epoch": 1.600873362445415,
+ "grad_norm": 0.005864050704985857,
+ "learning_rate": 0.0006,
+ "loss": 7.365346908569336,
+ "step": 115
+ },
+ {
+ "epoch": 1.6148471615720523,
+ "grad_norm": 0.011745430529117584,
+ "learning_rate": 0.0006,
+ "loss": 7.390227317810059,
+ "step": 116
+ },
+ {
+ "epoch": 1.62882096069869,
+ "grad_norm": 0.017205573618412018,
+ "learning_rate": 0.0006,
+ "loss": 7.3558549880981445,
+ "step": 117
+ },
+ {
+ "epoch": 1.6427947598253274,
+ "grad_norm": 0.021256346255540848,
+ "learning_rate": 0.0006,
+ "loss": 7.377346992492676,
+ "step": 118
+ },
+ {
+ "epoch": 1.6567685589519652,
+ "grad_norm": 0.019335204735398293,
+ "learning_rate": 0.0006,
+ "loss": 7.346194267272949,
+ "step": 119
+ },
+ {
+ "epoch": 1.6707423580786025,
+ "grad_norm": 0.008134003728628159,
+ "learning_rate": 0.0006,
+ "loss": 7.3063578605651855,
+ "step": 120
+ },
+ {
+ "epoch": 1.6847161572052403,
+ "grad_norm": 0.017613831907510757,
+ "learning_rate": 0.0006,
+ "loss": 7.338171005249023,
+ "step": 121
+ },
+ {
+ "epoch": 1.6986899563318776,
+ "grad_norm": 0.00890734139829874,
+ "learning_rate": 0.0006,
+ "loss": 7.355384826660156,
+ "step": 122
+ },
+ {
+ "epoch": 1.7126637554585153,
+ "grad_norm": 0.013448155485093594,
+ "learning_rate": 0.0006,
+ "loss": 7.344296455383301,
+ "step": 123
+ },
+ {
+ "epoch": 1.726637554585153,
+ "grad_norm": 0.009241633117198944,
+ "learning_rate": 0.0006,
+ "loss": 7.3454060554504395,
+ "step": 124
+ },
+ {
+ "epoch": 1.7406113537117904,
+ "grad_norm": 0.018561311066150665,
+ "learning_rate": 0.0006,
+ "loss": 7.32348108291626,
+ "step": 125
+ },
+ {
+ "epoch": 1.754585152838428,
+ "grad_norm": 0.025667704641819,
+ "learning_rate": 0.0006,
+ "loss": 7.356142520904541,
+ "step": 126
+ },
+ {
+ "epoch": 1.7685589519650655,
+ "grad_norm": 0.036232151091098785,
+ "learning_rate": 0.0006,
+ "loss": 7.363539695739746,
+ "step": 127
+ },
+ {
+ "epoch": 1.782532751091703,
+ "grad_norm": 0.059483714401721954,
+ "learning_rate": 0.0006,
+ "loss": 7.402331352233887,
+ "step": 128
+ },
+ {
+ "epoch": 1.7965065502183406,
+ "grad_norm": 0.05365009233355522,
+ "learning_rate": 0.0006,
+ "loss": 7.40457820892334,
+ "step": 129
+ },
+ {
+ "epoch": 1.8104803493449781,
+ "grad_norm": 0.02738415263593197,
+ "learning_rate": 0.0006,
+ "loss": 7.3756208419799805,
+ "step": 130
+ },
+ {
+ "epoch": 1.8244541484716157,
+ "grad_norm": 0.035017091780900955,
+ "learning_rate": 0.0006,
+ "loss": 7.379263877868652,
+ "step": 131
+ },
+ {
+ "epoch": 1.8384279475982532,
+ "grad_norm": 0.02819863148033619,
+ "learning_rate": 0.0006,
+ "loss": 7.327673435211182,
+ "step": 132
+ },
+ {
+ "epoch": 1.8524017467248908,
+ "grad_norm": 0.017754238098859787,
+ "learning_rate": 0.0006,
+ "loss": 7.333420276641846,
+ "step": 133
+ },
+ {
+ "epoch": 1.8663755458515285,
+ "grad_norm": 0.024685733020305634,
+ "learning_rate": 0.0006,
+ "loss": 7.34493350982666,
+ "step": 134
+ },
+ {
+ "epoch": 1.8803493449781659,
+ "grad_norm": 0.019426532089710236,
+ "learning_rate": 0.0006,
+ "loss": 7.328448295593262,
+ "step": 135
+ },
+ {
+ "epoch": 1.8943231441048036,
+ "grad_norm": 0.02106391079723835,
+ "learning_rate": 0.0006,
+ "loss": 7.327917098999023,
+ "step": 136
+ },
+ {
+ "epoch": 1.908296943231441,
+ "grad_norm": 0.02056879736483097,
+ "learning_rate": 0.0006,
+ "loss": 7.3311991691589355,
+ "step": 137
+ },
+ {
+ "epoch": 1.9222707423580787,
+ "grad_norm": 0.016865141689777374,
+ "learning_rate": 0.0006,
+ "loss": 7.312766075134277,
+ "step": 138
+ },
+ {
+ "epoch": 1.936244541484716,
+ "grad_norm": 0.017549855634570122,
+ "learning_rate": 0.0006,
+ "loss": 7.31191349029541,
+ "step": 139
+ },
+ {
+ "epoch": 1.9502183406113538,
+ "grad_norm": 0.015930157154798508,
+ "learning_rate": 0.0006,
+ "loss": 7.330901622772217,
+ "step": 140
+ },
+ {
+ "epoch": 1.9641921397379911,
+ "grad_norm": 0.013283558189868927,
+ "learning_rate": 0.0006,
+ "loss": 7.298318386077881,
+ "step": 141
+ },
+ {
+ "epoch": 1.9781659388646289,
+ "grad_norm": 0.012023469433188438,
+ "learning_rate": 0.0006,
+ "loss": 7.282337188720703,
+ "step": 142
+ },
+ {
+ "epoch": 1.9921397379912664,
+ "grad_norm": 0.01459511648863554,
+ "learning_rate": 0.0006,
+ "loss": 7.282924175262451,
+ "step": 143
+ },
+ {
+ "epoch": 2.0,
+ "grad_norm": 0.017474235966801643,
+ "learning_rate": 0.0006,
+ "loss": 7.254333972930908,
+ "step": 144
+ },
+ {
+ "epoch": 2.0,
+ "eval_loss": 7.318708896636963,
+ "eval_runtime": 59.6555,
+ "eval_samples_per_second": 40.935,
+ "eval_steps_per_second": 1.291,
+ "step": 144
+ },
+ {
+ "epoch": 2.0139737991266378,
+ "grad_norm": 0.019757507368922234,
+ "learning_rate": 0.0006,
+ "loss": 7.30288553237915,
+ "step": 145
+ },
+ {
+ "epoch": 2.027947598253275,
+ "grad_norm": 0.014571248553693295,
+ "learning_rate": 0.0006,
+ "loss": 7.274681568145752,
+ "step": 146
+ },
+ {
+ "epoch": 2.041921397379913,
+ "grad_norm": 0.015744198113679886,
+ "learning_rate": 0.0006,
+ "loss": 7.278632640838623,
+ "step": 147
+ },
+ {
+ "epoch": 2.05589519650655,
+ "grad_norm": 0.028048858046531677,
+ "learning_rate": 0.0006,
+ "loss": 7.26558780670166,
+ "step": 148
+ },
+ {
+ "epoch": 2.069868995633188,
+ "grad_norm": 0.06547307223081589,
+ "learning_rate": 0.0006,
+ "loss": 7.32567834854126,
+ "step": 149
+ },
+ {
+ "epoch": 2.0838427947598253,
+ "grad_norm": 0.05969763547182083,
+ "learning_rate": 0.0006,
+ "loss": 7.395836353302002,
+ "step": 150
+ },
+ {
+ "epoch": 2.097816593886463,
+ "grad_norm": 0.032574281096458435,
+ "learning_rate": 0.0006,
+ "loss": 7.318150997161865,
+ "step": 151
+ },
+ {
+ "epoch": 2.1117903930131003,
+ "grad_norm": 0.029452962800860405,
+ "learning_rate": 0.0006,
+ "loss": 7.308066368103027,
+ "step": 152
+ },
+ {
+ "epoch": 2.125764192139738,
+ "grad_norm": 0.03161991387605667,
+ "learning_rate": 0.0006,
+ "loss": 7.306290626525879,
+ "step": 153
+ },
+ {
+ "epoch": 2.1397379912663754,
+ "grad_norm": 0.027096295729279518,
+ "learning_rate": 0.0006,
+ "loss": 7.297796249389648,
+ "step": 154
+ },
+ {
+ "epoch": 2.153711790393013,
+ "grad_norm": 0.028657056391239166,
+ "learning_rate": 0.0006,
+ "loss": 7.317424774169922,
+ "step": 155
+ },
+ {
+ "epoch": 2.1676855895196505,
+ "grad_norm": 0.02046995982527733,
+ "learning_rate": 0.0006,
+ "loss": 7.297986030578613,
+ "step": 156
+ },
+ {
+ "epoch": 2.1816593886462883,
+ "grad_norm": 0.02220369316637516,
+ "learning_rate": 0.0006,
+ "loss": 7.268237590789795,
+ "step": 157
+ },
+ {
+ "epoch": 2.1956331877729256,
+ "grad_norm": 0.018381357192993164,
+ "learning_rate": 0.0006,
+ "loss": 7.259415626525879,
+ "step": 158
+ },
+ {
+ "epoch": 2.2096069868995634,
+ "grad_norm": 0.0204413291066885,
+ "learning_rate": 0.0006,
+ "loss": 7.258173942565918,
+ "step": 159
+ },
+ {
+ "epoch": 2.2235807860262007,
+ "grad_norm": 0.015874288976192474,
+ "learning_rate": 0.0006,
+ "loss": 7.281834602355957,
+ "step": 160
+ },
+ {
+ "epoch": 2.2375545851528384,
+ "grad_norm": 0.01634068600833416,
+ "learning_rate": 0.0006,
+ "loss": 7.2469868659973145,
+ "step": 161
+ },
+ {
+ "epoch": 2.251528384279476,
+ "grad_norm": 0.01728747971355915,
+ "learning_rate": 0.0006,
+ "loss": 7.280995845794678,
+ "step": 162
+ },
+ {
+ "epoch": 2.2655021834061135,
+ "grad_norm": 0.016341188922524452,
+ "learning_rate": 0.0006,
+ "loss": 7.273357391357422,
+ "step": 163
+ },
+ {
+ "epoch": 2.279475982532751,
+ "grad_norm": 0.016125807538628578,
+ "learning_rate": 0.0006,
+ "loss": 7.273510932922363,
+ "step": 164
+ },
+ {
+ "epoch": 2.2934497816593886,
+ "grad_norm": 0.01629287749528885,
+ "learning_rate": 0.0006,
+ "loss": 7.267029762268066,
+ "step": 165
+ },
+ {
+ "epoch": 2.3074235807860264,
+ "grad_norm": 0.015296131372451782,
+ "learning_rate": 0.0006,
+ "loss": 7.211989402770996,
+ "step": 166
+ },
+ {
+ "epoch": 2.3213973799126637,
+ "grad_norm": 0.01709570363163948,
+ "learning_rate": 0.0006,
+ "loss": 7.242851734161377,
+ "step": 167
+ },
+ {
+ "epoch": 2.3353711790393015,
+ "grad_norm": 0.012552388943731785,
+ "learning_rate": 0.0006,
+ "loss": 7.241696834564209,
+ "step": 168
+ },
+ {
+ "epoch": 2.349344978165939,
+ "grad_norm": 0.014571291394531727,
+ "learning_rate": 0.0006,
+ "loss": 7.270681381225586,
+ "step": 169
+ },
+ {
+ "epoch": 2.3633187772925766,
+ "grad_norm": 0.013389437459409237,
+ "learning_rate": 0.0006,
+ "loss": 7.191600799560547,
+ "step": 170
+ },
+ {
+ "epoch": 2.377292576419214,
+ "grad_norm": 0.011895314790308475,
+ "learning_rate": 0.0006,
+ "loss": 7.213610649108887,
+ "step": 171
+ },
+ {
+ "epoch": 2.3912663755458516,
+ "grad_norm": 0.01345642190426588,
+ "learning_rate": 0.0006,
+ "loss": 7.193787574768066,
+ "step": 172
+ },
+ {
+ "epoch": 2.405240174672489,
+ "grad_norm": 0.0091794328764081,
+ "learning_rate": 0.0006,
+ "loss": 7.233364105224609,
+ "step": 173
+ },
+ {
+ "epoch": 2.4192139737991267,
+ "grad_norm": 0.013767086900770664,
+ "learning_rate": 0.0006,
+ "loss": 7.217272758483887,
+ "step": 174
+ },
+ {
+ "epoch": 2.433187772925764,
+ "grad_norm": 0.015134960412979126,
+ "learning_rate": 0.0006,
+ "loss": 7.192791938781738,
+ "step": 175
+ },
+ {
+ "epoch": 2.447161572052402,
+ "grad_norm": 0.01790648326277733,
+ "learning_rate": 0.0006,
+ "loss": 7.229467391967773,
+ "step": 176
+ },
+ {
+ "epoch": 2.461135371179039,
+ "grad_norm": 0.015362377278506756,
+ "learning_rate": 0.0006,
+ "loss": 7.180346488952637,
+ "step": 177
+ },
+ {
+ "epoch": 2.475109170305677,
+ "grad_norm": 0.010505498386919498,
+ "learning_rate": 0.0006,
+ "loss": 7.188833236694336,
+ "step": 178
+ },
+ {
+ "epoch": 2.489082969432314,
+ "grad_norm": 0.009129747748374939,
+ "learning_rate": 0.0006,
+ "loss": 7.203153133392334,
+ "step": 179
+ },
+ {
+ "epoch": 2.503056768558952,
+ "grad_norm": 0.01147883478552103,
+ "learning_rate": 0.0006,
+ "loss": 7.163824558258057,
+ "step": 180
+ },
+ {
+ "epoch": 2.5170305676855893,
+ "grad_norm": 0.011338942684233189,
+ "learning_rate": 0.0006,
+ "loss": 7.16287899017334,
+ "step": 181
+ },
+ {
+ "epoch": 2.531004366812227,
+ "grad_norm": 0.013277675025165081,
+ "learning_rate": 0.0006,
+ "loss": 7.155905723571777,
+ "step": 182
+ },
+ {
+ "epoch": 2.544978165938865,
+ "grad_norm": 0.015269347466528416,
+ "learning_rate": 0.0006,
+ "loss": 7.189371109008789,
+ "step": 183
+ },
+ {
+ "epoch": 2.558951965065502,
+ "grad_norm": 0.017905596643686295,
+ "learning_rate": 0.0006,
+ "loss": 7.177547931671143,
+ "step": 184
+ },
+ {
+ "epoch": 2.5729257641921395,
+ "grad_norm": 0.024431610479950905,
+ "learning_rate": 0.0006,
+ "loss": 7.192169189453125,
+ "step": 185
+ },
+ {
+ "epoch": 2.5868995633187772,
+ "grad_norm": 0.03656961768865585,
+ "learning_rate": 0.0006,
+ "loss": 7.124387741088867,
+ "step": 186
+ },
+ {
+ "epoch": 2.600873362445415,
+ "grad_norm": 0.04337048903107643,
+ "learning_rate": 0.0006,
+ "loss": 7.140231609344482,
+ "step": 187
+ },
+ {
+ "epoch": 2.6148471615720523,
+ "grad_norm": 0.04047228768467903,
+ "learning_rate": 0.0006,
+ "loss": 7.171927452087402,
+ "step": 188
+ },
+ {
+ "epoch": 2.62882096069869,
+ "grad_norm": 0.07825770974159241,
+ "learning_rate": 0.0006,
+ "loss": 7.207631587982178,
+ "step": 189
+ },
+ {
+ "epoch": 2.6427947598253274,
+ "grad_norm": 0.07118063420057297,
+ "learning_rate": 0.0006,
+ "loss": 7.209488391876221,
+ "step": 190
+ },
+ {
+ "epoch": 2.656768558951965,
+ "grad_norm": 0.05153293535113335,
+ "learning_rate": 0.0006,
+ "loss": 7.14793062210083,
+ "step": 191
+ },
+ {
+ "epoch": 2.6707423580786025,
+ "grad_norm": 0.03318839520215988,
+ "learning_rate": 0.0006,
+ "loss": 7.163010597229004,
+ "step": 192
+ },
+ {
+ "epoch": 2.6847161572052403,
+ "grad_norm": 0.029333539307117462,
+ "learning_rate": 0.0006,
+ "loss": 7.159955978393555,
+ "step": 193
+ },
+ {
+ "epoch": 2.6986899563318776,
+ "grad_norm": 0.029405880719423294,
+ "learning_rate": 0.0006,
+ "loss": 7.136178493499756,
+ "step": 194
+ },
+ {
+ "epoch": 2.7126637554585153,
+ "grad_norm": 0.037815988063812256,
+ "learning_rate": 0.0006,
+ "loss": 7.183186054229736,
+ "step": 195
+ },
+ {
+ "epoch": 2.726637554585153,
+ "grad_norm": 0.020307250320911407,
+ "learning_rate": 0.0006,
+ "loss": 7.133164882659912,
+ "step": 196
+ },
+ {
+ "epoch": 2.7406113537117904,
+ "grad_norm": 0.026048069819808006,
+ "learning_rate": 0.0006,
+ "loss": 7.124977111816406,
+ "step": 197
+ },
+ {
+ "epoch": 2.7545851528384278,
+ "grad_norm": 0.027316724881529808,
+ "learning_rate": 0.0006,
+ "loss": 7.07265567779541,
+ "step": 198
+ },
+ {
+ "epoch": 2.7685589519650655,
+ "grad_norm": 0.021629009395837784,
+ "learning_rate": 0.0006,
+ "loss": 7.150020599365234,
+ "step": 199
+ },
+ {
+ "epoch": 2.7825327510917033,
+ "grad_norm": 0.01605929434299469,
+ "learning_rate": 0.0006,
+ "loss": 7.10919189453125,
+ "step": 200
+ },
+ {
+ "epoch": 2.7965065502183406,
+ "grad_norm": 0.020800089463591576,
+ "learning_rate": 0.0006,
+ "loss": 7.132045269012451,
+ "step": 201
+ },
+ {
+ "epoch": 2.810480349344978,
+ "grad_norm": 0.019225774332880974,
+ "learning_rate": 0.0006,
+ "loss": 7.128364562988281,
+ "step": 202
+ },
+ {
+ "epoch": 2.8244541484716157,
+ "grad_norm": 0.01561372634023428,
+ "learning_rate": 0.0006,
+ "loss": 7.083669662475586,
+ "step": 203
+ },
+ {
+ "epoch": 2.8384279475982535,
+ "grad_norm": 0.021969519555568695,
+ "learning_rate": 0.0006,
+ "loss": 7.050841331481934,
+ "step": 204
+ },
+ {
+ "epoch": 2.8524017467248908,
+ "grad_norm": 0.021453695371747017,
+ "learning_rate": 0.0006,
+ "loss": 7.097085475921631,
+ "step": 205
+ },
+ {
+ "epoch": 2.8663755458515285,
+ "grad_norm": 0.022836022078990936,
+ "learning_rate": 0.0006,
+ "loss": 7.10405969619751,
+ "step": 206
+ },
+ {
+ "epoch": 2.880349344978166,
+ "grad_norm": 0.02430851384997368,
+ "learning_rate": 0.0006,
+ "loss": 7.045181751251221,
+ "step": 207
+ },
+ {
+ "epoch": 2.8943231441048036,
+ "grad_norm": 0.020166993141174316,
+ "learning_rate": 0.0006,
+ "loss": 7.068183898925781,
+ "step": 208
+ },
+ {
+ "epoch": 2.908296943231441,
+ "grad_norm": 0.0118505684658885,
+ "learning_rate": 0.0006,
+ "loss": 7.07132625579834,
+ "step": 209
+ },
+ {
+ "epoch": 2.9222707423580787,
+ "grad_norm": 0.023772427812218666,
+ "learning_rate": 0.0006,
+ "loss": 7.055238723754883,
+ "step": 210
+ },
+ {
+ "epoch": 2.936244541484716,
+ "grad_norm": 0.02350712940096855,
+ "learning_rate": 0.0006,
+ "loss": 7.055898666381836,
+ "step": 211
+ },
+ {
+ "epoch": 2.950218340611354,
+ "grad_norm": 0.017244907096028328,
+ "learning_rate": 0.0006,
+ "loss": 7.0099711418151855,
+ "step": 212
+ },
+ {
+ "epoch": 2.964192139737991,
+ "grad_norm": 0.021565575152635574,
+ "learning_rate": 0.0006,
+ "loss": 6.991751670837402,
+ "step": 213
+ },
+ {
+ "epoch": 2.978165938864629,
+ "grad_norm": 0.03241860866546631,
+ "learning_rate": 0.0006,
+ "loss": 7.005980491638184,
+ "step": 214
+ },
+ {
+ "epoch": 2.992139737991266,
+ "grad_norm": 0.04613726586103439,
+ "learning_rate": 0.0006,
+ "loss": 7.027859687805176,
+ "step": 215
+ },
+ {
+ "epoch": 3.0,
+ "grad_norm": 0.06404894590377808,
+ "learning_rate": 0.0006,
+ "loss": 7.123725891113281,
+ "step": 216
+ },
+ {
+ "epoch": 3.0,
+ "eval_loss": 7.070647716522217,
+ "eval_runtime": 58.5801,
+ "eval_samples_per_second": 41.687,
+ "eval_steps_per_second": 1.314,
+ "step": 216
+ },
+ {
+ "epoch": 3.0139737991266378,
+ "grad_norm": 0.05542432889342308,
+ "learning_rate": 0.0006,
+ "loss": 7.078299522399902,
+ "step": 217
+ },
+ {
+ "epoch": 3.027947598253275,
+ "grad_norm": 0.026484526693820953,
+ "learning_rate": 0.0006,
+ "loss": 6.961159706115723,
+ "step": 218
+ },
+ {
+ "epoch": 3.041921397379913,
+ "grad_norm": 0.028276391327381134,
+ "learning_rate": 0.0006,
+ "loss": 7.019961833953857,
+ "step": 219
+ },
+ {
+ "epoch": 3.05589519650655,
+ "grad_norm": 0.024486595764756203,
+ "learning_rate": 0.0006,
+ "loss": 7.039677619934082,
+ "step": 220
+ },
+ {
+ "epoch": 3.069868995633188,
+ "grad_norm": 0.017620893195271492,
+ "learning_rate": 0.0006,
+ "loss": 6.9270782470703125,
+ "step": 221
+ },
+ {
+ "epoch": 3.0838427947598253,
+ "grad_norm": 0.02547302283346653,
+ "learning_rate": 0.0006,
+ "loss": 6.924233436584473,
+ "step": 222
+ },
+ {
+ "epoch": 3.097816593886463,
+ "grad_norm": 0.02539009042084217,
+ "learning_rate": 0.0006,
+ "loss": 6.973979949951172,
+ "step": 223
+ },
+ {
+ "epoch": 3.1117903930131003,
+ "grad_norm": 0.03256227448582649,
+ "learning_rate": 0.0006,
+ "loss": 6.933725357055664,
+ "step": 224
+ },
+ {
+ "epoch": 3.125764192139738,
+ "grad_norm": 0.04651800915598869,
+ "learning_rate": 0.0006,
+ "loss": 6.951044082641602,
+ "step": 225
+ },
+ {
+ "epoch": 3.1397379912663754,
+ "grad_norm": 0.04822975769639015,
+ "learning_rate": 0.0006,
+ "loss": 6.9614410400390625,
+ "step": 226
+ },
+ {
+ "epoch": 3.153711790393013,
+ "grad_norm": 0.031165020540356636,
+ "learning_rate": 0.0006,
+ "loss": 6.91193151473999,
+ "step": 227
+ },
+ {
+ "epoch": 3.1676855895196505,
+ "grad_norm": 0.029754292219877243,
+ "learning_rate": 0.0006,
+ "loss": 6.91802978515625,
+ "step": 228
+ },
+ {
+ "epoch": 3.1816593886462883,
+ "grad_norm": 0.02193152718245983,
+ "learning_rate": 0.0006,
+ "loss": 6.928090572357178,
+ "step": 229
+ },
+ {
+ "epoch": 3.1956331877729256,
+ "grad_norm": 0.02428470179438591,
+ "learning_rate": 0.0006,
+ "loss": 6.90396785736084,
+ "step": 230
+ },
+ {
+ "epoch": 3.2096069868995634,
+ "grad_norm": 0.023374345153570175,
+ "learning_rate": 0.0006,
+ "loss": 6.885858535766602,
+ "step": 231
+ },
+ {
+ "epoch": 3.2235807860262007,
+ "grad_norm": 0.023405000567436218,
+ "learning_rate": 0.0006,
+ "loss": 6.8855767250061035,
+ "step": 232
+ },
+ {
+ "epoch": 3.2375545851528384,
+ "grad_norm": 0.017277134582400322,
+ "learning_rate": 0.0006,
+ "loss": 6.901449203491211,
+ "step": 233
+ },
+ {
+ "epoch": 3.251528384279476,
+ "grad_norm": 0.019496750086545944,
+ "learning_rate": 0.0006,
+ "loss": 6.885687828063965,
+ "step": 234
+ },
+ {
+ "epoch": 3.2655021834061135,
+ "grad_norm": 0.016893276944756508,
+ "learning_rate": 0.0006,
+ "loss": 6.872369766235352,
+ "step": 235
+ },
+ {
+ "epoch": 3.279475982532751,
+ "grad_norm": 0.02058715932071209,
+ "learning_rate": 0.0006,
+ "loss": 6.823280334472656,
+ "step": 236
+ },
+ {
+ "epoch": 3.2934497816593886,
+ "grad_norm": 0.015530755743384361,
+ "learning_rate": 0.0006,
+ "loss": 6.860499858856201,
+ "step": 237
+ },
+ {
+ "epoch": 3.3074235807860264,
+ "grad_norm": 0.01934926211833954,
+ "learning_rate": 0.0006,
+ "loss": 6.817448616027832,
+ "step": 238
+ },
+ {
+ "epoch": 3.3213973799126637,
+ "grad_norm": 0.017413552850484848,
+ "learning_rate": 0.0006,
+ "loss": 6.881955146789551,
+ "step": 239
+ },
+ {
+ "epoch": 3.3353711790393015,
+ "grad_norm": 0.026478875428438187,
+ "learning_rate": 0.0006,
+ "loss": 6.827404499053955,
+ "step": 240
+ },
+ {
+ "epoch": 3.349344978165939,
+ "grad_norm": 0.032611701637506485,
+ "learning_rate": 0.0006,
+ "loss": 6.78285551071167,
+ "step": 241
+ },
+ {
+ "epoch": 3.3633187772925766,
+ "grad_norm": 0.041900306940078735,
+ "learning_rate": 0.0006,
+ "loss": 6.783053874969482,
+ "step": 242
+ },
+ {
+ "epoch": 3.377292576419214,
+ "grad_norm": 0.04149497672915459,
+ "learning_rate": 0.0006,
+ "loss": 6.745220184326172,
+ "step": 243
+ },
+ {
+ "epoch": 3.3912663755458516,
+ "grad_norm": 0.023949826136231422,
+ "learning_rate": 0.0006,
+ "loss": 6.841533184051514,
+ "step": 244
+ },
+ {
+ "epoch": 3.405240174672489,
+ "grad_norm": 0.028601842001080513,
+ "learning_rate": 0.0006,
+ "loss": 6.852852821350098,
+ "step": 245
+ },
+ {
+ "epoch": 3.4192139737991267,
+ "grad_norm": 0.027365155518054962,
+ "learning_rate": 0.0006,
+ "loss": 6.74376106262207,
+ "step": 246
+ },
+ {
+ "epoch": 3.433187772925764,
+ "grad_norm": 0.02856568619608879,
+ "learning_rate": 0.0006,
+ "loss": 6.74906063079834,
+ "step": 247
+ },
+ {
+ "epoch": 3.447161572052402,
+ "grad_norm": 0.019172416999936104,
+ "learning_rate": 0.0006,
+ "loss": 6.809206008911133,
+ "step": 248
+ },
+ {
+ "epoch": 3.461135371179039,
+ "grad_norm": 0.02161746844649315,
+ "learning_rate": 0.0006,
+ "loss": 6.789888381958008,
+ "step": 249
+ },
+ {
+ "epoch": 3.475109170305677,
+ "grad_norm": 0.020273666828870773,
+ "learning_rate": 0.0006,
+ "loss": 6.73512077331543,
+ "step": 250
+ },
+ {
+ "epoch": 3.489082969432314,
+ "grad_norm": 0.022329283878207207,
+ "learning_rate": 0.0006,
+ "loss": 6.826148509979248,
+ "step": 251
+ },
+ {
+ "epoch": 3.503056768558952,
+ "grad_norm": 0.02284284122288227,
+ "learning_rate": 0.0006,
+ "loss": 6.748521327972412,
+ "step": 252
+ },
+ {
+ "epoch": 3.5170305676855893,
+ "grad_norm": 0.026368247345089912,
+ "learning_rate": 0.0006,
+ "loss": 6.645879745483398,
+ "step": 253
+ },
+ {
+ "epoch": 3.531004366812227,
+ "grad_norm": 0.03627244383096695,
+ "learning_rate": 0.0006,
+ "loss": 6.666064739227295,
+ "step": 254
+ },
+ {
+ "epoch": 3.544978165938865,
+ "grad_norm": 0.04140935093164444,
+ "learning_rate": 0.0006,
+ "loss": 6.712275981903076,
+ "step": 255
+ },
+ {
+ "epoch": 3.558951965065502,
+ "grad_norm": 0.0390462800860405,
+ "learning_rate": 0.0006,
+ "loss": 6.665900230407715,
+ "step": 256
+ },
+ {
+ "epoch": 3.5729257641921395,
+ "grad_norm": 0.05178583785891533,
+ "learning_rate": 0.0006,
+ "loss": 6.753298759460449,
+ "step": 257
+ },
+ {
+ "epoch": 3.5868995633187772,
+ "grad_norm": 0.054328061640262604,
+ "learning_rate": 0.0006,
+ "loss": 6.732633113861084,
+ "step": 258
+ },
+ {
+ "epoch": 3.600873362445415,
+ "grad_norm": 0.04590460658073425,
+ "learning_rate": 0.0006,
+ "loss": 6.65250301361084,
+ "step": 259
+ },
+ {
+ "epoch": 3.6148471615720523,
+ "grad_norm": 0.044331666082143784,
+ "learning_rate": 0.0006,
+ "loss": 6.68654203414917,
+ "step": 260
+ },
+ {
+ "epoch": 3.62882096069869,
+ "grad_norm": 0.05308730527758598,
+ "learning_rate": 0.0006,
+ "loss": 6.642482757568359,
+ "step": 261
+ },
+ {
+ "epoch": 3.6427947598253274,
+ "grad_norm": 0.03999016433954239,
+ "learning_rate": 0.0006,
+ "loss": 6.625227451324463,
+ "step": 262
+ },
+ {
+ "epoch": 3.656768558951965,
+ "grad_norm": 0.04088086634874344,
+ "learning_rate": 0.0006,
+ "loss": 6.67623233795166,
+ "step": 263
+ },
+ {
+ "epoch": 3.6707423580786025,
+ "grad_norm": 0.03615636005997658,
+ "learning_rate": 0.0006,
+ "loss": 6.6869001388549805,
+ "step": 264
+ },
+ {
+ "epoch": 3.6847161572052403,
+ "grad_norm": 0.02817360684275627,
+ "learning_rate": 0.0006,
+ "loss": 6.674601078033447,
+ "step": 265
+ },
+ {
+ "epoch": 3.6986899563318776,
+ "grad_norm": 0.032586123794317245,
+ "learning_rate": 0.0006,
+ "loss": 6.612994194030762,
+ "step": 266
+ },
+ {
+ "epoch": 3.7126637554585153,
+ "grad_norm": 0.031117867678403854,
+ "learning_rate": 0.0006,
+ "loss": 6.6244354248046875,
+ "step": 267
+ },
+ {
+ "epoch": 3.726637554585153,
+ "grad_norm": 0.028330687433481216,
+ "learning_rate": 0.0006,
+ "loss": 6.669557094573975,
+ "step": 268
+ },
+ {
+ "epoch": 3.7406113537117904,
+ "grad_norm": 0.021489202976226807,
+ "learning_rate": 0.0006,
+ "loss": 6.632035732269287,
+ "step": 269
+ },
+ {
+ "epoch": 3.7545851528384278,
+ "grad_norm": 0.024269424378871918,
+ "learning_rate": 0.0006,
+ "loss": 6.660867691040039,
+ "step": 270
+ },
+ {
+ "epoch": 3.7685589519650655,
+ "grad_norm": 0.02237873338162899,
+ "learning_rate": 0.0006,
+ "loss": 6.602587699890137,
+ "step": 271
+ },
+ {
+ "epoch": 3.7825327510917033,
+ "grad_norm": 0.026503929868340492,
+ "learning_rate": 0.0006,
+ "loss": 6.599200248718262,
+ "step": 272
+ },
+ {
+ "epoch": 3.7965065502183406,
+ "grad_norm": 0.03687124326825142,
+ "learning_rate": 0.0006,
+ "loss": 6.567296504974365,
+ "step": 273
+ },
+ {
+ "epoch": 3.810480349344978,
+ "grad_norm": 0.04708952456712723,
+ "learning_rate": 0.0006,
+ "loss": 6.621265411376953,
+ "step": 274
+ },
+ {
+ "epoch": 3.8244541484716157,
+ "grad_norm": 0.041846614331007004,
+ "learning_rate": 0.0006,
+ "loss": 6.550536155700684,
+ "step": 275
+ },
+ {
+ "epoch": 3.8384279475982535,
+ "grad_norm": 0.031249675899744034,
+ "learning_rate": 0.0006,
+ "loss": 6.54278564453125,
+ "step": 276
+ },
+ {
+ "epoch": 3.8524017467248908,
+ "grad_norm": 0.023082256317138672,
+ "learning_rate": 0.0006,
+ "loss": 6.544787406921387,
+ "step": 277
+ },
+ {
+ "epoch": 3.8663755458515285,
+ "grad_norm": 0.020995570346713066,
+ "learning_rate": 0.0006,
+ "loss": 6.518143653869629,
+ "step": 278
+ },
+ {
+ "epoch": 3.880349344978166,
+ "grad_norm": 0.020839324221014977,
+ "learning_rate": 0.0006,
+ "loss": 6.546667098999023,
+ "step": 279
+ },
+ {
+ "epoch": 3.8943231441048036,
+ "grad_norm": 0.018377432599663734,
+ "learning_rate": 0.0006,
+ "loss": 6.478307723999023,
+ "step": 280
+ },
+ {
+ "epoch": 3.908296943231441,
+ "grad_norm": 0.021713724359869957,
+ "learning_rate": 0.0006,
+ "loss": 6.485574245452881,
+ "step": 281
+ },
+ {
+ "epoch": 3.9222707423580787,
+ "grad_norm": 0.01359301246702671,
+ "learning_rate": 0.0006,
+ "loss": 6.570268630981445,
+ "step": 282
+ },
+ {
+ "epoch": 3.936244541484716,
+ "grad_norm": 0.016233332455158234,
+ "learning_rate": 0.0006,
+ "loss": 6.57150936126709,
+ "step": 283
+ },
+ {
+ "epoch": 3.950218340611354,
+ "grad_norm": 0.020687608048319817,
+ "learning_rate": 0.0006,
+ "loss": 6.527956008911133,
+ "step": 284
+ },
+ {
+ "epoch": 3.964192139737991,
+ "grad_norm": 0.02032964862883091,
+ "learning_rate": 0.0006,
+ "loss": 6.468169689178467,
+ "step": 285
+ },
+ {
+ "epoch": 3.978165938864629,
+ "grad_norm": 0.016676288098096848,
+ "learning_rate": 0.0006,
+ "loss": 6.448418617248535,
+ "step": 286
+ },
+ {
+ "epoch": 3.992139737991266,
+ "grad_norm": 0.0168539360165596,
+ "learning_rate": 0.0006,
+ "loss": 6.461180686950684,
+ "step": 287
+ },
+ {
+ "epoch": 4.0,
+ "grad_norm": 0.02666233666241169,
+ "learning_rate": 0.0006,
+ "loss": 6.344893932342529,
+ "step": 288
+ },
+ {
+ "epoch": 4.0,
+ "eval_loss": 6.5384979248046875,
+ "eval_runtime": 58.814,
+ "eval_samples_per_second": 41.521,
+ "eval_steps_per_second": 1.309,
+ "step": 288
+ },
+ {
+ "epoch": 4.013973799126638,
+ "grad_norm": 0.04154708981513977,
+ "learning_rate": 0.0006,
+ "loss": 6.41855525970459,
+ "step": 289
+ },
+ {
+ "epoch": 4.0279475982532755,
+ "grad_norm": 0.04895668104290962,
+ "learning_rate": 0.0006,
+ "loss": 6.499302864074707,
+ "step": 290
+ },
+ {
+ "epoch": 4.041921397379912,
+ "grad_norm": 0.034464482218027115,
+ "learning_rate": 0.0006,
+ "loss": 6.438611030578613,
+ "step": 291
+ },
+ {
+ "epoch": 4.05589519650655,
+ "grad_norm": 0.06845773756504059,
+ "learning_rate": 0.0006,
+ "loss": 6.538361072540283,
+ "step": 292
+ },
+ {
+ "epoch": 4.069868995633188,
+ "grad_norm": 0.053768858313560486,
+ "learning_rate": 0.0006,
+ "loss": 6.527524948120117,
+ "step": 293
+ },
+ {
+ "epoch": 4.083842794759826,
+ "grad_norm": 0.039037931710481644,
+ "learning_rate": 0.0006,
+ "loss": 6.503847122192383,
+ "step": 294
+ },
+ {
+ "epoch": 4.097816593886463,
+ "grad_norm": 0.03962196782231331,
+ "learning_rate": 0.0006,
+ "loss": 6.475178241729736,
+ "step": 295
+ },
+ {
+ "epoch": 4.1117903930131,
+ "grad_norm": 0.049101535230875015,
+ "learning_rate": 0.0006,
+ "loss": 6.474587917327881,
+ "step": 296
+ },
+ {
+ "epoch": 4.125764192139738,
+ "grad_norm": 0.02830282226204872,
+ "learning_rate": 0.0006,
+ "loss": 6.45065975189209,
+ "step": 297
+ },
+ {
+ "epoch": 4.139737991266376,
+ "grad_norm": 0.023153482005000114,
+ "learning_rate": 0.0006,
+ "loss": 6.5218706130981445,
+ "step": 298
+ },
+ {
+ "epoch": 4.153711790393013,
+ "grad_norm": 0.025064002722501755,
+ "learning_rate": 0.0006,
+ "loss": 6.456475257873535,
+ "step": 299
+ },
+ {
+ "epoch": 4.1676855895196505,
+ "grad_norm": 0.02357092685997486,
+ "learning_rate": 0.0006,
+ "loss": 6.485952377319336,
+ "step": 300
+ },
+ {
+ "epoch": 4.181659388646288,
+ "grad_norm": 0.021832305938005447,
+ "learning_rate": 0.0006,
+ "loss": 6.425506591796875,
+ "step": 301
+ },
+ {
+ "epoch": 4.195633187772926,
+ "grad_norm": 0.0227900929749012,
+ "learning_rate": 0.0006,
+ "loss": 6.439155578613281,
+ "step": 302
+ },
+ {
+ "epoch": 4.209606986899563,
+ "grad_norm": 0.015961607918143272,
+ "learning_rate": 0.0006,
+ "loss": 6.460862636566162,
+ "step": 303
+ },
+ {
+ "epoch": 4.223580786026201,
+ "grad_norm": 0.017876390367746353,
+ "learning_rate": 0.0006,
+ "loss": 6.44564151763916,
+ "step": 304
+ },
+ {
+ "epoch": 4.2375545851528384,
+ "grad_norm": 0.013703204691410065,
+ "learning_rate": 0.0006,
+ "loss": 6.418414115905762,
+ "step": 305
+ },
+ {
+ "epoch": 4.251528384279476,
+ "grad_norm": 0.015240306034684181,
+ "learning_rate": 0.0006,
+ "loss": 6.411238670349121,
+ "step": 306
+ },
+ {
+ "epoch": 4.265502183406113,
+ "grad_norm": 0.01587662845849991,
+ "learning_rate": 0.0006,
+ "loss": 6.395642280578613,
+ "step": 307
+ },
+ {
+ "epoch": 4.279475982532751,
+ "grad_norm": 0.014650700613856316,
+ "learning_rate": 0.0006,
+ "loss": 6.302793025970459,
+ "step": 308
+ },
+ {
+ "epoch": 4.293449781659389,
+ "grad_norm": 0.016533225774765015,
+ "learning_rate": 0.0006,
+ "loss": 6.380588054656982,
+ "step": 309
+ },
+ {
+ "epoch": 4.307423580786026,
+ "grad_norm": 0.019963741302490234,
+ "learning_rate": 0.0006,
+ "loss": 6.345336437225342,
+ "step": 310
+ },
+ {
+ "epoch": 4.321397379912664,
+ "grad_norm": 0.02237936295568943,
+ "learning_rate": 0.0006,
+ "loss": 6.387770652770996,
+ "step": 311
+ },
+ {
+ "epoch": 4.335371179039301,
+ "grad_norm": 0.026658328250050545,
+ "learning_rate": 0.0006,
+ "loss": 6.323662757873535,
+ "step": 312
+ },
+ {
+ "epoch": 4.349344978165939,
+ "grad_norm": 0.02852284163236618,
+ "learning_rate": 0.0006,
+ "loss": 6.382204055786133,
+ "step": 313
+ },
+ {
+ "epoch": 4.3633187772925766,
+ "grad_norm": 0.028815090656280518,
+ "learning_rate": 0.0006,
+ "loss": 6.402390480041504,
+ "step": 314
+ },
+ {
+ "epoch": 4.377292576419214,
+ "grad_norm": 0.029393529519438744,
+ "learning_rate": 0.0006,
+ "loss": 6.280278205871582,
+ "step": 315
+ },
+ {
+ "epoch": 4.391266375545851,
+ "grad_norm": 0.02186041697859764,
+ "learning_rate": 0.0006,
+ "loss": 6.302707672119141,
+ "step": 316
+ },
+ {
+ "epoch": 4.405240174672489,
+ "grad_norm": 0.02071218006312847,
+ "learning_rate": 0.0006,
+ "loss": 6.3365020751953125,
+ "step": 317
+ },
+ {
+ "epoch": 4.419213973799127,
+ "grad_norm": 0.02779117226600647,
+ "learning_rate": 0.0006,
+ "loss": 6.252157211303711,
+ "step": 318
+ },
+ {
+ "epoch": 4.4331877729257645,
+ "grad_norm": 0.03470654785633087,
+ "learning_rate": 0.0006,
+ "loss": 6.319070339202881,
+ "step": 319
+ },
+ {
+ "epoch": 4.447161572052401,
+ "grad_norm": 0.037067804485559464,
+ "learning_rate": 0.0006,
+ "loss": 6.314116477966309,
+ "step": 320
+ },
+ {
+ "epoch": 4.461135371179039,
+ "grad_norm": 0.032182756811380386,
+ "learning_rate": 0.0006,
+ "loss": 6.267298698425293,
+ "step": 321
+ },
+ {
+ "epoch": 4.475109170305677,
+ "grad_norm": 0.026305217295885086,
+ "learning_rate": 0.0006,
+ "loss": 6.331688404083252,
+ "step": 322
+ },
+ {
+ "epoch": 4.489082969432315,
+ "grad_norm": 0.027228357270359993,
+ "learning_rate": 0.0006,
+ "loss": 6.31139612197876,
+ "step": 323
+ },
+ {
+ "epoch": 4.503056768558952,
+ "grad_norm": 0.025781169533729553,
+ "learning_rate": 0.0006,
+ "loss": 6.194684982299805,
+ "step": 324
+ },
+ {
+ "epoch": 4.517030567685589,
+ "grad_norm": 0.029186977073550224,
+ "learning_rate": 0.0006,
+ "loss": 6.31306266784668,
+ "step": 325
+ },
+ {
+ "epoch": 4.531004366812227,
+ "grad_norm": 0.023608777672052383,
+ "learning_rate": 0.0006,
+ "loss": 6.285243034362793,
+ "step": 326
+ },
+ {
+ "epoch": 4.544978165938865,
+ "grad_norm": 0.02267594076693058,
+ "learning_rate": 0.0006,
+ "loss": 6.30342960357666,
+ "step": 327
+ },
+ {
+ "epoch": 4.558951965065502,
+ "grad_norm": 0.01926310732960701,
+ "learning_rate": 0.0006,
+ "loss": 6.280606269836426,
+ "step": 328
+ },
+ {
+ "epoch": 4.5729257641921395,
+ "grad_norm": 0.026484837755560875,
+ "learning_rate": 0.0006,
+ "loss": 6.237980365753174,
+ "step": 329
+ },
+ {
+ "epoch": 4.586899563318777,
+ "grad_norm": 0.026228854432702065,
+ "learning_rate": 0.0006,
+ "loss": 6.3018293380737305,
+ "step": 330
+ },
+ {
+ "epoch": 4.600873362445415,
+ "grad_norm": 0.022096967324614525,
+ "learning_rate": 0.0006,
+ "loss": 6.253863334655762,
+ "step": 331
+ },
+ {
+ "epoch": 4.614847161572053,
+ "grad_norm": 0.027223890647292137,
+ "learning_rate": 0.0006,
+ "loss": 6.166882038116455,
+ "step": 332
+ },
+ {
+ "epoch": 4.62882096069869,
+ "grad_norm": 0.03160124272108078,
+ "learning_rate": 0.0006,
+ "loss": 6.230309963226318,
+ "step": 333
+ },
+ {
+ "epoch": 4.642794759825327,
+ "grad_norm": 0.04110539332032204,
+ "learning_rate": 0.0006,
+ "loss": 6.2334818840026855,
+ "step": 334
+ },
+ {
+ "epoch": 4.656768558951965,
+ "grad_norm": 0.04355933889746666,
+ "learning_rate": 0.0006,
+ "loss": 6.160956382751465,
+ "step": 335
+ },
+ {
+ "epoch": 4.670742358078603,
+ "grad_norm": 0.03611086308956146,
+ "learning_rate": 0.0006,
+ "loss": 6.23399543762207,
+ "step": 336
+ },
+ {
+ "epoch": 4.68471615720524,
+ "grad_norm": 0.051811400800943375,
+ "learning_rate": 0.0006,
+ "loss": 6.241855144500732,
+ "step": 337
+ },
+ {
+ "epoch": 4.698689956331878,
+ "grad_norm": 0.057434193789958954,
+ "learning_rate": 0.0006,
+ "loss": 6.220993995666504,
+ "step": 338
+ },
+ {
+ "epoch": 4.712663755458515,
+ "grad_norm": 0.04420953616499901,
+ "learning_rate": 0.0006,
+ "loss": 6.300461292266846,
+ "step": 339
+ },
+ {
+ "epoch": 4.726637554585153,
+ "grad_norm": 0.03515457361936569,
+ "learning_rate": 0.0006,
+ "loss": 6.269349098205566,
+ "step": 340
+ },
+ {
+ "epoch": 4.74061135371179,
+ "grad_norm": 0.03222600743174553,
+ "learning_rate": 0.0006,
+ "loss": 6.248613357543945,
+ "step": 341
+ },
+ {
+ "epoch": 4.754585152838428,
+ "grad_norm": 0.0384046845138073,
+ "learning_rate": 0.0006,
+ "loss": 6.266979217529297,
+ "step": 342
+ },
+ {
+ "epoch": 4.7685589519650655,
+ "grad_norm": 0.0514867827296257,
+ "learning_rate": 0.0006,
+ "loss": 6.234103202819824,
+ "step": 343
+ },
+ {
+ "epoch": 4.782532751091703,
+ "grad_norm": 0.050747793167829514,
+ "learning_rate": 0.0006,
+ "loss": 6.322582244873047,
+ "step": 344
+ },
+ {
+ "epoch": 4.796506550218341,
+ "grad_norm": 0.040878988802433014,
+ "learning_rate": 0.0006,
+ "loss": 6.212509632110596,
+ "step": 345
+ },
+ {
+ "epoch": 4.810480349344978,
+ "grad_norm": 0.034655820578336716,
+ "learning_rate": 0.0006,
+ "loss": 6.265879154205322,
+ "step": 346
+ },
+ {
+ "epoch": 4.824454148471616,
+ "grad_norm": 0.04202224686741829,
+ "learning_rate": 0.0006,
+ "loss": 6.2896409034729,
+ "step": 347
+ },
+ {
+ "epoch": 4.8384279475982535,
+ "grad_norm": 0.03632277995347977,
+ "learning_rate": 0.0006,
+ "loss": 6.253917217254639,
+ "step": 348
+ },
+ {
+ "epoch": 4.85240174672489,
+ "grad_norm": 0.028645765036344528,
+ "learning_rate": 0.0006,
+ "loss": 6.207965850830078,
+ "step": 349
+ },
+ {
+ "epoch": 4.866375545851528,
+ "grad_norm": 0.025305170565843582,
+ "learning_rate": 0.0006,
+ "loss": 6.168068885803223,
+ "step": 350
+ },
+ {
+ "epoch": 4.880349344978166,
+ "grad_norm": 0.022137803956866264,
+ "learning_rate": 0.0006,
+ "loss": 6.235283374786377,
+ "step": 351
+ },
+ {
+ "epoch": 4.894323144104804,
+ "grad_norm": 0.01814538985490799,
+ "learning_rate": 0.0006,
+ "loss": 6.259641647338867,
+ "step": 352
+ },
+ {
+ "epoch": 4.908296943231441,
+ "grad_norm": 0.018465086817741394,
+ "learning_rate": 0.0006,
+ "loss": 6.111738681793213,
+ "step": 353
+ },
+ {
+ "epoch": 4.922270742358078,
+ "grad_norm": 0.018638933077454567,
+ "learning_rate": 0.0006,
+ "loss": 6.225642204284668,
+ "step": 354
+ },
+ {
+ "epoch": 4.936244541484716,
+ "grad_norm": 0.016461260616779327,
+ "learning_rate": 0.0006,
+ "loss": 6.140647888183594,
+ "step": 355
+ },
+ {
+ "epoch": 4.950218340611354,
+ "grad_norm": 0.015458385460078716,
+ "learning_rate": 0.0006,
+ "loss": 6.155224800109863,
+ "step": 356
+ },
+ {
+ "epoch": 4.964192139737992,
+ "grad_norm": 0.014934048056602478,
+ "learning_rate": 0.0006,
+ "loss": 6.101555824279785,
+ "step": 357
+ },
+ {
+ "epoch": 4.978165938864628,
+ "grad_norm": 0.013979545794427395,
+ "learning_rate": 0.0006,
+ "loss": 6.161718845367432,
+ "step": 358
+ },
+ {
+ "epoch": 4.992139737991266,
+ "grad_norm": 0.01640394888818264,
+ "learning_rate": 0.0006,
+ "loss": 6.055768966674805,
+ "step": 359
+ },
+ {
+ "epoch": 5.0,
+ "grad_norm": 0.015135680325329304,
+ "learning_rate": 0.0006,
+ "loss": 6.169695854187012,
+ "step": 360
+ },
+ {
+ "epoch": 5.0,
+ "eval_loss": 6.16408109664917,
+ "eval_runtime": 59.3649,
+ "eval_samples_per_second": 41.135,
+ "eval_steps_per_second": 1.297,
+ "step": 360
+ },
+ {
+ "epoch": 5.013973799126638,
+ "grad_norm": 0.012849048711359501,
+ "learning_rate": 0.0006,
+ "loss": 6.055830001831055,
+ "step": 361
+ },
+ {
+ "epoch": 5.0279475982532755,
+ "grad_norm": 0.018169576302170753,
+ "learning_rate": 0.0006,
+ "loss": 6.082655429840088,
+ "step": 362
+ },
+ {
+ "epoch": 5.041921397379912,
+ "grad_norm": 0.024293430149555206,
+ "learning_rate": 0.0006,
+ "loss": 6.109032154083252,
+ "step": 363
+ },
+ {
+ "epoch": 5.05589519650655,
+ "grad_norm": 0.03266787901520729,
+ "learning_rate": 0.0006,
+ "loss": 6.030613899230957,
+ "step": 364
+ },
+ {
+ "epoch": 5.069868995633188,
+ "grad_norm": 0.03572266176342964,
+ "learning_rate": 0.0006,
+ "loss": 6.050390720367432,
+ "step": 365
+ },
+ {
+ "epoch": 5.083842794759826,
+ "grad_norm": 0.032625213265419006,
+ "learning_rate": 0.0006,
+ "loss": 6.169313430786133,
+ "step": 366
+ },
+ {
+ "epoch": 5.097816593886463,
+ "grad_norm": 0.04051872715353966,
+ "learning_rate": 0.0006,
+ "loss": 6.10398006439209,
+ "step": 367
+ },
+ {
+ "epoch": 5.1117903930131,
+ "grad_norm": 0.0351213738322258,
+ "learning_rate": 0.0006,
+ "loss": 6.121973514556885,
+ "step": 368
+ },
+ {
+ "epoch": 5.125764192139738,
+ "grad_norm": 0.03523759916424751,
+ "learning_rate": 0.0006,
+ "loss": 6.0626654624938965,
+ "step": 369
+ },
+ {
+ "epoch": 5.139737991266376,
+ "grad_norm": 0.02706182189285755,
+ "learning_rate": 0.0006,
+ "loss": 6.049354076385498,
+ "step": 370
+ },
+ {
+ "epoch": 5.153711790393013,
+ "grad_norm": 0.024212589487433434,
+ "learning_rate": 0.0006,
+ "loss": 6.034826278686523,
+ "step": 371
+ },
+ {
+ "epoch": 5.1676855895196505,
+ "grad_norm": 0.024798082187771797,
+ "learning_rate": 0.0006,
+ "loss": 6.004058837890625,
+ "step": 372
+ },
+ {
+ "epoch": 5.181659388646288,
+ "grad_norm": 0.02325567416846752,
+ "learning_rate": 0.0006,
+ "loss": 5.986461639404297,
+ "step": 373
+ },
+ {
+ "epoch": 5.195633187772926,
+ "grad_norm": 0.019060570746660233,
+ "learning_rate": 0.0006,
+ "loss": 6.059736251831055,
+ "step": 374
+ },
+ {
+ "epoch": 5.209606986899563,
+ "grad_norm": 0.016822976991534233,
+ "learning_rate": 0.0006,
+ "loss": 6.072957515716553,
+ "step": 375
+ },
+ {
+ "epoch": 5.223580786026201,
+ "grad_norm": 0.017218658700585365,
+ "learning_rate": 0.0006,
+ "loss": 6.031739234924316,
+ "step": 376
+ },
+ {
+ "epoch": 5.2375545851528384,
+ "grad_norm": 0.014889383688569069,
+ "learning_rate": 0.0006,
+ "loss": 6.025674343109131,
+ "step": 377
+ },
+ {
+ "epoch": 5.251528384279476,
+ "grad_norm": 0.01708107627928257,
+ "learning_rate": 0.0006,
+ "loss": 6.090451240539551,
+ "step": 378
+ },
+ {
+ "epoch": 5.265502183406113,
+ "grad_norm": 0.020260317251086235,
+ "learning_rate": 0.0006,
+ "loss": 6.094257354736328,
+ "step": 379
+ },
+ {
+ "epoch": 5.279475982532751,
+ "grad_norm": 0.020110400393605232,
+ "learning_rate": 0.0006,
+ "loss": 6.021188735961914,
+ "step": 380
+ },
+ {
+ "epoch": 5.293449781659389,
+ "grad_norm": 0.02027830481529236,
+ "learning_rate": 0.0006,
+ "loss": 6.022156238555908,
+ "step": 381
+ },
+ {
+ "epoch": 5.307423580786026,
+ "grad_norm": 0.022747062146663666,
+ "learning_rate": 0.0006,
+ "loss": 6.011836051940918,
+ "step": 382
+ },
+ {
+ "epoch": 5.321397379912664,
+ "grad_norm": 0.01990230567753315,
+ "learning_rate": 0.0006,
+ "loss": 6.003054618835449,
+ "step": 383
+ },
+ {
+ "epoch": 5.335371179039301,
+ "grad_norm": 0.01596238650381565,
+ "learning_rate": 0.0006,
+ "loss": 5.991410732269287,
+ "step": 384
+ },
+ {
+ "epoch": 5.349344978165939,
+ "grad_norm": 0.015847105532884598,
+ "learning_rate": 0.0006,
+ "loss": 6.100622177124023,
+ "step": 385
+ },
+ {
+ "epoch": 5.3633187772925766,
+ "grad_norm": 0.016179397702217102,
+ "learning_rate": 0.0006,
+ "loss": 6.032659530639648,
+ "step": 386
+ },
+ {
+ "epoch": 5.377292576419214,
+ "grad_norm": 0.018256602808833122,
+ "learning_rate": 0.0006,
+ "loss": 5.961983680725098,
+ "step": 387
+ },
+ {
+ "epoch": 5.391266375545851,
+ "grad_norm": 0.02805912122130394,
+ "learning_rate": 0.0006,
+ "loss": 5.983541965484619,
+ "step": 388
+ },
+ {
+ "epoch": 5.405240174672489,
+ "grad_norm": 0.039082255214452744,
+ "learning_rate": 0.0006,
+ "loss": 6.021474361419678,
+ "step": 389
+ },
+ {
+ "epoch": 5.419213973799127,
+ "grad_norm": 0.036757659167051315,
+ "learning_rate": 0.0006,
+ "loss": 6.005046367645264,
+ "step": 390
+ },
+ {
+ "epoch": 5.4331877729257645,
+ "grad_norm": 0.035277366638183594,
+ "learning_rate": 0.0006,
+ "loss": 6.036296844482422,
+ "step": 391
+ },
+ {
+ "epoch": 5.447161572052401,
+ "grad_norm": 0.034404635429382324,
+ "learning_rate": 0.0006,
+ "loss": 5.994539260864258,
+ "step": 392
+ },
+ {
+ "epoch": 5.461135371179039,
+ "grad_norm": 0.0377206988632679,
+ "learning_rate": 0.0006,
+ "loss": 5.941534519195557,
+ "step": 393
+ },
+ {
+ "epoch": 5.475109170305677,
+ "grad_norm": 0.0389922633767128,
+ "learning_rate": 0.0006,
+ "loss": 5.983644485473633,
+ "step": 394
+ },
+ {
+ "epoch": 5.489082969432315,
+ "grad_norm": 0.04176778718829155,
+ "learning_rate": 0.0006,
+ "loss": 6.030922889709473,
+ "step": 395
+ },
+ {
+ "epoch": 5.503056768558952,
+ "grad_norm": 0.0337153784930706,
+ "learning_rate": 0.0006,
+ "loss": 5.961367130279541,
+ "step": 396
+ },
+ {
+ "epoch": 5.517030567685589,
+ "grad_norm": 0.036116816103458405,
+ "learning_rate": 0.0006,
+ "loss": 5.819280624389648,
+ "step": 397
+ },
+ {
+ "epoch": 5.531004366812227,
+ "grad_norm": 0.030724041163921356,
+ "learning_rate": 0.0006,
+ "loss": 5.906380653381348,
+ "step": 398
+ },
+ {
+ "epoch": 5.544978165938865,
+ "grad_norm": 0.030264202505350113,
+ "learning_rate": 0.0006,
+ "loss": 5.975833415985107,
+ "step": 399
+ },
+ {
+ "epoch": 5.558951965065502,
+ "grad_norm": 0.032096318900585175,
+ "learning_rate": 0.0006,
+ "loss": 6.001348972320557,
+ "step": 400
+ },
+ {
+ "epoch": 5.5729257641921395,
+ "grad_norm": 0.030579356476664543,
+ "learning_rate": 0.0006,
+ "loss": 5.928768634796143,
+ "step": 401
+ },
+ {
+ "epoch": 5.586899563318777,
+ "grad_norm": 0.028241973370313644,
+ "learning_rate": 0.0006,
+ "loss": 5.92582893371582,
+ "step": 402
+ },
+ {
+ "epoch": 5.600873362445415,
+ "grad_norm": 0.023086953908205032,
+ "learning_rate": 0.0006,
+ "loss": 5.892926216125488,
+ "step": 403
+ },
+ {
+ "epoch": 5.614847161572053,
+ "grad_norm": 0.02547992393374443,
+ "learning_rate": 0.0006,
+ "loss": 5.903195381164551,
+ "step": 404
+ },
+ {
+ "epoch": 5.62882096069869,
+ "grad_norm": 0.023089831694960594,
+ "learning_rate": 0.0006,
+ "loss": 5.912033557891846,
+ "step": 405
+ },
+ {
+ "epoch": 5.642794759825327,
+ "grad_norm": 0.025446368381381035,
+ "learning_rate": 0.0006,
+ "loss": 5.926138401031494,
+ "step": 406
+ },
+ {
+ "epoch": 5.656768558951965,
+ "grad_norm": 0.03748747706413269,
+ "learning_rate": 0.0006,
+ "loss": 5.913451194763184,
+ "step": 407
+ },
+ {
+ "epoch": 5.670742358078603,
+ "grad_norm": 0.03493810072541237,
+ "learning_rate": 0.0006,
+ "loss": 5.97633695602417,
+ "step": 408
+ },
+ {
+ "epoch": 5.68471615720524,
+ "grad_norm": 0.023131275549530983,
+ "learning_rate": 0.0006,
+ "loss": 5.899571418762207,
+ "step": 409
+ },
+ {
+ "epoch": 5.698689956331878,
+ "grad_norm": 0.02986014075577259,
+ "learning_rate": 0.0006,
+ "loss": 5.894903182983398,
+ "step": 410
+ },
+ {
+ "epoch": 5.712663755458515,
+ "grad_norm": 0.030171144753694534,
+ "learning_rate": 0.0006,
+ "loss": 5.876595497131348,
+ "step": 411
+ },
+ {
+ "epoch": 5.726637554585153,
+ "grad_norm": 0.029377546161413193,
+ "learning_rate": 0.0006,
+ "loss": 5.911637306213379,
+ "step": 412
+ },
+ {
+ "epoch": 5.74061135371179,
+ "grad_norm": 0.029829490929841995,
+ "learning_rate": 0.0006,
+ "loss": 5.819629669189453,
+ "step": 413
+ },
+ {
+ "epoch": 5.754585152838428,
+ "grad_norm": 0.024139299988746643,
+ "learning_rate": 0.0006,
+ "loss": 5.81728458404541,
+ "step": 414
+ },
+ {
+ "epoch": 5.7685589519650655,
+ "grad_norm": 0.021453432738780975,
+ "learning_rate": 0.0006,
+ "loss": 5.987326145172119,
+ "step": 415
+ },
+ {
+ "epoch": 5.782532751091703,
+ "grad_norm": 0.017942246049642563,
+ "learning_rate": 0.0006,
+ "loss": 5.893008232116699,
+ "step": 416
+ },
+ {
+ "epoch": 5.796506550218341,
+ "grad_norm": 0.019724637269973755,
+ "learning_rate": 0.0006,
+ "loss": 5.912441253662109,
+ "step": 417
+ },
+ {
+ "epoch": 5.810480349344978,
+ "grad_norm": 0.018597912043333054,
+ "learning_rate": 0.0006,
+ "loss": 5.860394477844238,
+ "step": 418
+ },
+ {
+ "epoch": 5.824454148471616,
+ "grad_norm": 0.016232412308454514,
+ "learning_rate": 0.0006,
+ "loss": 5.94991397857666,
+ "step": 419
+ },
+ {
+ "epoch": 5.8384279475982535,
+ "grad_norm": 0.01667204685509205,
+ "learning_rate": 0.0006,
+ "loss": 5.8506364822387695,
+ "step": 420
+ },
+ {
+ "epoch": 5.85240174672489,
+ "grad_norm": 0.014232151210308075,
+ "learning_rate": 0.0006,
+ "loss": 5.928424835205078,
+ "step": 421
+ },
+ {
+ "epoch": 5.866375545851528,
+ "grad_norm": 0.01570476032793522,
+ "learning_rate": 0.0006,
+ "loss": 5.953692436218262,
+ "step": 422
+ },
+ {
+ "epoch": 5.880349344978166,
+ "grad_norm": 0.01586179807782173,
+ "learning_rate": 0.0006,
+ "loss": 5.820014476776123,
+ "step": 423
+ },
+ {
+ "epoch": 5.894323144104804,
+ "grad_norm": 0.014729145914316177,
+ "learning_rate": 0.0006,
+ "loss": 5.860154628753662,
+ "step": 424
+ },
+ {
+ "epoch": 5.908296943231441,
+ "grad_norm": 0.013111459091305733,
+ "learning_rate": 0.0006,
+ "loss": 5.85988187789917,
+ "step": 425
+ },
+ {
+ "epoch": 5.922270742358078,
+ "grad_norm": 0.012937922962009907,
+ "learning_rate": 0.0006,
+ "loss": 5.756265640258789,
+ "step": 426
+ },
+ {
+ "epoch": 5.936244541484716,
+ "grad_norm": 0.01377453189343214,
+ "learning_rate": 0.0006,
+ "loss": 5.8333048820495605,
+ "step": 427
+ },
+ {
+ "epoch": 5.950218340611354,
+ "grad_norm": 0.014054795727133751,
+ "learning_rate": 0.0006,
+ "loss": 5.803333759307861,
+ "step": 428
+ },
+ {
+ "epoch": 5.964192139737992,
+ "grad_norm": 0.01674959622323513,
+ "learning_rate": 0.0006,
+ "loss": 5.8718485832214355,
+ "step": 429
+ },
+ {
+ "epoch": 5.978165938864628,
+ "grad_norm": 0.018733017146587372,
+ "learning_rate": 0.0006,
+ "loss": 5.875979900360107,
+ "step": 430
+ },
+ {
+ "epoch": 5.992139737991266,
+ "grad_norm": 0.02088983915746212,
+ "learning_rate": 0.0006,
+ "loss": 5.763634204864502,
+ "step": 431
+ },
+ {
+ "epoch": 6.0,
+ "grad_norm": 0.026705985888838768,
+ "learning_rate": 0.0006,
+ "loss": 5.6922478675842285,
+ "step": 432
+ },
+ {
+ "epoch": 6.0,
+ "eval_loss": 5.847958087921143,
+ "eval_runtime": 58.5451,
+ "eval_samples_per_second": 41.711,
+ "eval_steps_per_second": 1.315,
+ "step": 432
+ },
+ {
+ "epoch": 6.013973799126638,
+ "grad_norm": 0.03181544691324234,
+ "learning_rate": 0.0006,
+ "loss": 5.786368370056152,
+ "step": 433
+ },
+ {
+ "epoch": 6.0279475982532755,
+ "grad_norm": 0.03238112851977348,
+ "learning_rate": 0.0006,
+ "loss": 5.810310363769531,
+ "step": 434
+ },
+ {
+ "epoch": 6.041921397379912,
+ "grad_norm": 0.03916756808757782,
+ "learning_rate": 0.0006,
+ "loss": 5.7761993408203125,
+ "step": 435
+ },
+ {
+ "epoch": 6.05589519650655,
+ "grad_norm": 0.044009730219841,
+ "learning_rate": 0.0006,
+ "loss": 5.898112773895264,
+ "step": 436
+ },
+ {
+ "epoch": 6.069868995633188,
+ "grad_norm": 0.045235246419906616,
+ "learning_rate": 0.0006,
+ "loss": 5.876371383666992,
+ "step": 437
+ },
+ {
+ "epoch": 6.083842794759826,
+ "grad_norm": 0.050218384712934494,
+ "learning_rate": 0.0006,
+ "loss": 5.805103302001953,
+ "step": 438
+ },
+ {
+ "epoch": 6.097816593886463,
+ "grad_norm": 0.05444490164518356,
+ "learning_rate": 0.0006,
+ "loss": 5.826424598693848,
+ "step": 439
+ },
+ {
+ "epoch": 6.1117903930131,
+ "grad_norm": 0.04890982061624527,
+ "learning_rate": 0.0006,
+ "loss": 5.7735395431518555,
+ "step": 440
+ },
+ {
+ "epoch": 6.125764192139738,
+ "grad_norm": 0.05478639155626297,
+ "learning_rate": 0.0006,
+ "loss": 5.823677062988281,
+ "step": 441
+ },
+ {
+ "epoch": 6.139737991266376,
+ "grad_norm": 0.057562313973903656,
+ "learning_rate": 0.0006,
+ "loss": 5.889334678649902,
+ "step": 442
+ },
+ {
+ "epoch": 6.153711790393013,
+ "grad_norm": 0.06447703391313553,
+ "learning_rate": 0.0006,
+ "loss": 5.84621524810791,
+ "step": 443
+ },
+ {
+ "epoch": 6.1676855895196505,
+ "grad_norm": 0.048861872404813766,
+ "learning_rate": 0.0006,
+ "loss": 5.931595802307129,
+ "step": 444
+ },
+ {
+ "epoch": 6.181659388646288,
+ "grad_norm": 0.05521339178085327,
+ "learning_rate": 0.0006,
+ "loss": 5.8229780197143555,
+ "step": 445
+ },
+ {
+ "epoch": 6.195633187772926,
+ "grad_norm": 0.053666990250349045,
+ "learning_rate": 0.0006,
+ "loss": 5.879191875457764,
+ "step": 446
+ },
+ {
+ "epoch": 6.209606986899563,
+ "grad_norm": 0.0451025515794754,
+ "learning_rate": 0.0006,
+ "loss": 5.801628112792969,
+ "step": 447
+ },
+ {
+ "epoch": 6.223580786026201,
+ "grad_norm": 0.04291655868291855,
+ "learning_rate": 0.0006,
+ "loss": 5.914680480957031,
+ "step": 448
+ },
+ {
+ "epoch": 6.2375545851528384,
+ "grad_norm": 0.035102490335702896,
+ "learning_rate": 0.0006,
+ "loss": 5.843682765960693,
+ "step": 449
+ },
+ {
+ "epoch": 6.251528384279476,
+ "grad_norm": 0.03403995931148529,
+ "learning_rate": 0.0006,
+ "loss": 5.827154159545898,
+ "step": 450
+ },
+ {
+ "epoch": 6.265502183406113,
+ "grad_norm": 0.03444375470280647,
+ "learning_rate": 0.0006,
+ "loss": 5.800136566162109,
+ "step": 451
+ },
+ {
+ "epoch": 6.279475982532751,
+ "grad_norm": 0.03165159001946449,
+ "learning_rate": 0.0006,
+ "loss": 5.906252384185791,
+ "step": 452
+ },
+ {
+ "epoch": 6.293449781659389,
+ "grad_norm": 0.026153093203902245,
+ "learning_rate": 0.0006,
+ "loss": 5.7423200607299805,
+ "step": 453
+ },
+ {
+ "epoch": 6.307423580786026,
+ "grad_norm": 0.024357657879590988,
+ "learning_rate": 0.0006,
+ "loss": 5.701364517211914,
+ "step": 454
+ },
+ {
+ "epoch": 6.321397379912664,
+ "grad_norm": 0.021508049219846725,
+ "learning_rate": 0.0006,
+ "loss": 5.8328094482421875,
+ "step": 455
+ },
+ {
+ "epoch": 6.335371179039301,
+ "grad_norm": 0.017313921824097633,
+ "learning_rate": 0.0006,
+ "loss": 5.792145252227783,
+ "step": 456
+ },
+ {
+ "epoch": 6.349344978165939,
+ "grad_norm": 0.018563397228717804,
+ "learning_rate": 0.0006,
+ "loss": 5.732672214508057,
+ "step": 457
+ },
+ {
+ "epoch": 6.3633187772925766,
+ "grad_norm": 0.016568679362535477,
+ "learning_rate": 0.0006,
+ "loss": 5.81948184967041,
+ "step": 458
+ },
+ {
+ "epoch": 6.377292576419214,
+ "grad_norm": 0.014133770950138569,
+ "learning_rate": 0.0006,
+ "loss": 5.734982490539551,
+ "step": 459
+ },
+ {
+ "epoch": 6.391266375545851,
+ "grad_norm": 0.014114447869360447,
+ "learning_rate": 0.0006,
+ "loss": 5.70073127746582,
+ "step": 460
+ },
+ {
+ "epoch": 6.405240174672489,
+ "grad_norm": 0.012907393276691437,
+ "learning_rate": 0.0006,
+ "loss": 5.758626937866211,
+ "step": 461
+ },
+ {
+ "epoch": 6.419213973799127,
+ "grad_norm": 0.012673444114625454,
+ "learning_rate": 0.0006,
+ "loss": 5.775138854980469,
+ "step": 462
+ },
+ {
+ "epoch": 6.4331877729257645,
+ "grad_norm": 0.01311410591006279,
+ "learning_rate": 0.0006,
+ "loss": 5.70920991897583,
+ "step": 463
+ },
+ {
+ "epoch": 6.447161572052401,
+ "grad_norm": 0.012935544364154339,
+ "learning_rate": 0.0006,
+ "loss": 5.781479835510254,
+ "step": 464
+ },
+ {
+ "epoch": 6.461135371179039,
+ "grad_norm": 0.01439738366752863,
+ "learning_rate": 0.0006,
+ "loss": 5.691315650939941,
+ "step": 465
+ },
+ {
+ "epoch": 6.475109170305677,
+ "grad_norm": 0.01205186452716589,
+ "learning_rate": 0.0006,
+ "loss": 5.662499904632568,
+ "step": 466
+ },
+ {
+ "epoch": 6.489082969432315,
+ "grad_norm": 0.011575652286410332,
+ "learning_rate": 0.0006,
+ "loss": 5.6271772384643555,
+ "step": 467
+ },
+ {
+ "epoch": 6.503056768558952,
+ "grad_norm": 0.011672502383589745,
+ "learning_rate": 0.0006,
+ "loss": 5.761662006378174,
+ "step": 468
+ },
+ {
+ "epoch": 6.517030567685589,
+ "grad_norm": 0.011541751213371754,
+ "learning_rate": 0.0006,
+ "loss": 5.762078285217285,
+ "step": 469
+ },
+ {
+ "epoch": 6.531004366812227,
+ "grad_norm": 0.011396365240216255,
+ "learning_rate": 0.0006,
+ "loss": 5.67873477935791,
+ "step": 470
+ },
+ {
+ "epoch": 6.544978165938865,
+ "grad_norm": 0.010688499547541142,
+ "learning_rate": 0.0006,
+ "loss": 5.717051982879639,
+ "step": 471
+ },
+ {
+ "epoch": 6.558951965065502,
+ "grad_norm": 0.012224317528307438,
+ "learning_rate": 0.0006,
+ "loss": 5.707948684692383,
+ "step": 472
+ },
+ {
+ "epoch": 6.5729257641921395,
+ "grad_norm": 0.011856825090944767,
+ "learning_rate": 0.0006,
+ "loss": 5.70878791809082,
+ "step": 473
+ },
+ {
+ "epoch": 6.586899563318777,
+ "grad_norm": 0.01199064590036869,
+ "learning_rate": 0.0006,
+ "loss": 5.708697319030762,
+ "step": 474
+ },
+ {
+ "epoch": 6.600873362445415,
+ "grad_norm": 0.01219446212053299,
+ "learning_rate": 0.0006,
+ "loss": 5.6061577796936035,
+ "step": 475
+ },
+ {
+ "epoch": 6.614847161572053,
+ "grad_norm": 0.013397484086453915,
+ "learning_rate": 0.0006,
+ "loss": 5.624789714813232,
+ "step": 476
+ },
+ {
+ "epoch": 6.62882096069869,
+ "grad_norm": 0.01483136136084795,
+ "learning_rate": 0.0006,
+ "loss": 5.721141338348389,
+ "step": 477
+ },
+ {
+ "epoch": 6.642794759825327,
+ "grad_norm": 0.0194488987326622,
+ "learning_rate": 0.0006,
+ "loss": 5.604279518127441,
+ "step": 478
+ },
+ {
+ "epoch": 6.656768558951965,
+ "grad_norm": 0.02079896256327629,
+ "learning_rate": 0.0006,
+ "loss": 5.586322784423828,
+ "step": 479
+ },
+ {
+ "epoch": 6.670742358078603,
+ "grad_norm": 0.018996229395270348,
+ "learning_rate": 0.0006,
+ "loss": 5.5987772941589355,
+ "step": 480
+ },
+ {
+ "epoch": 6.68471615720524,
+ "grad_norm": 0.015779603272676468,
+ "learning_rate": 0.0006,
+ "loss": 5.588602066040039,
+ "step": 481
+ },
+ {
+ "epoch": 6.698689956331878,
+ "grad_norm": 0.015322973020374775,
+ "learning_rate": 0.0006,
+ "loss": 5.686467170715332,
+ "step": 482
+ },
+ {
+ "epoch": 6.712663755458515,
+ "grad_norm": 0.015282213687896729,
+ "learning_rate": 0.0006,
+ "loss": 5.664676666259766,
+ "step": 483
+ },
+ {
+ "epoch": 6.726637554585153,
+ "grad_norm": 0.016589272767305374,
+ "learning_rate": 0.0006,
+ "loss": 5.666739463806152,
+ "step": 484
+ },
+ {
+ "epoch": 6.74061135371179,
+ "grad_norm": 0.01811421848833561,
+ "learning_rate": 0.0006,
+ "loss": 5.647593021392822,
+ "step": 485
+ },
+ {
+ "epoch": 6.754585152838428,
+ "grad_norm": 0.019652029499411583,
+ "learning_rate": 0.0006,
+ "loss": 5.628726959228516,
+ "step": 486
+ },
+ {
+ "epoch": 6.7685589519650655,
+ "grad_norm": 0.02118665538728237,
+ "learning_rate": 0.0006,
+ "loss": 5.6580705642700195,
+ "step": 487
+ },
+ {
+ "epoch": 6.782532751091703,
+ "grad_norm": 0.02237832546234131,
+ "learning_rate": 0.0006,
+ "loss": 5.635931968688965,
+ "step": 488
+ },
+ {
+ "epoch": 6.796506550218341,
+ "grad_norm": 0.023897631093859673,
+ "learning_rate": 0.0006,
+ "loss": 5.517949104309082,
+ "step": 489
+ },
+ {
+ "epoch": 6.810480349344978,
+ "grad_norm": 0.02442227490246296,
+ "learning_rate": 0.0006,
+ "loss": 5.600021839141846,
+ "step": 490
+ },
+ {
+ "epoch": 6.824454148471616,
+ "grad_norm": 0.024675287306308746,
+ "learning_rate": 0.0006,
+ "loss": 5.6639909744262695,
+ "step": 491
+ },
+ {
+ "epoch": 6.8384279475982535,
+ "grad_norm": 0.030372392386198044,
+ "learning_rate": 0.0006,
+ "loss": 5.582888603210449,
+ "step": 492
+ },
+ {
+ "epoch": 6.85240174672489,
+ "grad_norm": 0.03237690031528473,
+ "learning_rate": 0.0006,
+ "loss": 5.573896884918213,
+ "step": 493
+ },
+ {
+ "epoch": 6.866375545851528,
+ "grad_norm": 0.03511589393019676,
+ "learning_rate": 0.0006,
+ "loss": 5.565878868103027,
+ "step": 494
+ },
+ {
+ "epoch": 6.880349344978166,
+ "grad_norm": 0.03883938118815422,
+ "learning_rate": 0.0006,
+ "loss": 5.58540678024292,
+ "step": 495
+ },
+ {
+ "epoch": 6.894323144104804,
+ "grad_norm": 0.04175502806901932,
+ "learning_rate": 0.0006,
+ "loss": 5.6095194816589355,
+ "step": 496
+ },
+ {
+ "epoch": 6.908296943231441,
+ "grad_norm": 0.04013441503047943,
+ "learning_rate": 0.0006,
+ "loss": 5.654401779174805,
+ "step": 497
+ },
+ {
+ "epoch": 6.922270742358078,
+ "grad_norm": 0.04618770629167557,
+ "learning_rate": 0.0006,
+ "loss": 5.6483473777771,
+ "step": 498
+ },
+ {
+ "epoch": 6.936244541484716,
+ "grad_norm": 0.04507308453321457,
+ "learning_rate": 0.0006,
+ "loss": 5.5856218338012695,
+ "step": 499
+ },
+ {
+ "epoch": 6.950218340611354,
+ "grad_norm": 0.03151383996009827,
+ "learning_rate": 0.0006,
+ "loss": 5.594086647033691,
+ "step": 500
+ },
+ {
+ "epoch": 6.964192139737992,
+ "grad_norm": 0.026883797720074654,
+ "learning_rate": 0.0006,
+ "loss": 5.607676029205322,
+ "step": 501
+ },
+ {
+ "epoch": 6.978165938864628,
+ "grad_norm": 0.02981836162507534,
+ "learning_rate": 0.0006,
+ "loss": 5.563666820526123,
+ "step": 502
+ },
+ {
+ "epoch": 6.992139737991266,
+ "grad_norm": 0.03233477845788002,
+ "learning_rate": 0.0006,
+ "loss": 5.667543411254883,
+ "step": 503
+ },
+ {
+ "epoch": 7.0,
+ "grad_norm": 0.029927456751465797,
+ "learning_rate": 0.0006,
+ "loss": 5.733482837677002,
+ "step": 504
+ },
+ {
+ "epoch": 7.0,
+ "eval_loss": 5.679076671600342,
+ "eval_runtime": 59.5965,
+ "eval_samples_per_second": 40.976,
+ "eval_steps_per_second": 1.292,
+ "step": 504
+ },
+ {
+ "epoch": 7.013973799126638,
+ "grad_norm": 0.027081597596406937,
+ "learning_rate": 0.0006,
+ "loss": 5.634487152099609,
+ "step": 505
+ },
+ {
+ "epoch": 7.0279475982532755,
+ "grad_norm": 0.031959421932697296,
+ "learning_rate": 0.0006,
+ "loss": 5.643294811248779,
+ "step": 506
+ },
+ {
+ "epoch": 7.041921397379912,
+ "grad_norm": 0.02802063524723053,
+ "learning_rate": 0.0006,
+ "loss": 5.573239326477051,
+ "step": 507
+ },
+ {
+ "epoch": 7.05589519650655,
+ "grad_norm": 0.032579705119132996,
+ "learning_rate": 0.0006,
+ "loss": 5.627150058746338,
+ "step": 508
+ },
+ {
+ "epoch": 7.069868995633188,
+ "grad_norm": 0.02503928542137146,
+ "learning_rate": 0.0006,
+ "loss": 5.6045308113098145,
+ "step": 509
+ },
+ {
+ "epoch": 7.083842794759826,
+ "grad_norm": 0.02421317622065544,
+ "learning_rate": 0.0006,
+ "loss": 5.522153377532959,
+ "step": 510
+ },
+ {
+ "epoch": 7.097816593886463,
+ "grad_norm": 0.021983426064252853,
+ "learning_rate": 0.0006,
+ "loss": 5.632939338684082,
+ "step": 511
+ },
+ {
+ "epoch": 7.1117903930131,
+ "grad_norm": 0.021933183073997498,
+ "learning_rate": 0.0006,
+ "loss": 5.592185974121094,
+ "step": 512
+ },
+ {
+ "epoch": 7.125764192139738,
+ "grad_norm": 0.02287031151354313,
+ "learning_rate": 0.0006,
+ "loss": 5.613700866699219,
+ "step": 513
+ },
+ {
+ "epoch": 7.139737991266376,
+ "grad_norm": 0.02281602844595909,
+ "learning_rate": 0.0006,
+ "loss": 5.551383018493652,
+ "step": 514
+ },
+ {
+ "epoch": 7.153711790393013,
+ "grad_norm": 0.021050360053777695,
+ "learning_rate": 0.0006,
+ "loss": 5.55275821685791,
+ "step": 515
+ },
+ {
+ "epoch": 7.1676855895196505,
+ "grad_norm": 0.018299926072359085,
+ "learning_rate": 0.0006,
+ "loss": 5.528225421905518,
+ "step": 516
+ },
+ {
+ "epoch": 7.181659388646288,
+ "grad_norm": 0.02024853602051735,
+ "learning_rate": 0.0006,
+ "loss": 5.522453784942627,
+ "step": 517
+ },
+ {
+ "epoch": 7.195633187772926,
+ "grad_norm": 0.02135239914059639,
+ "learning_rate": 0.0006,
+ "loss": 5.527522563934326,
+ "step": 518
+ },
+ {
+ "epoch": 7.209606986899563,
+ "grad_norm": 0.023394184187054634,
+ "learning_rate": 0.0006,
+ "loss": 5.4535017013549805,
+ "step": 519
+ },
+ {
+ "epoch": 7.223580786026201,
+ "grad_norm": 0.025952080264687538,
+ "learning_rate": 0.0006,
+ "loss": 5.5374674797058105,
+ "step": 520
+ },
+ {
+ "epoch": 7.2375545851528384,
+ "grad_norm": 0.022287718951702118,
+ "learning_rate": 0.0006,
+ "loss": 5.493753433227539,
+ "step": 521
+ },
+ {
+ "epoch": 7.251528384279476,
+ "grad_norm": 0.02018447406589985,
+ "learning_rate": 0.0006,
+ "loss": 5.510575294494629,
+ "step": 522
+ },
+ {
+ "epoch": 7.265502183406113,
+ "grad_norm": 0.02050507813692093,
+ "learning_rate": 0.0006,
+ "loss": 5.466026782989502,
+ "step": 523
+ },
+ {
+ "epoch": 7.279475982532751,
+ "grad_norm": 0.023688802495598793,
+ "learning_rate": 0.0006,
+ "loss": 5.485815525054932,
+ "step": 524
+ },
+ {
+ "epoch": 7.293449781659389,
+ "grad_norm": 0.021078843623399734,
+ "learning_rate": 0.0006,
+ "loss": 5.415581703186035,
+ "step": 525
+ },
+ {
+ "epoch": 7.307423580786026,
+ "grad_norm": 0.018266011029481888,
+ "learning_rate": 0.0006,
+ "loss": 5.402815818786621,
+ "step": 526
+ },
+ {
+ "epoch": 7.321397379912664,
+ "grad_norm": 0.019693493843078613,
+ "learning_rate": 0.0006,
+ "loss": 5.504674911499023,
+ "step": 527
+ },
+ {
+ "epoch": 7.335371179039301,
+ "grad_norm": 0.02031373605132103,
+ "learning_rate": 0.0006,
+ "loss": 5.503837585449219,
+ "step": 528
+ },
+ {
+ "epoch": 7.349344978165939,
+ "grad_norm": 0.017636626958847046,
+ "learning_rate": 0.0006,
+ "loss": 5.530580997467041,
+ "step": 529
+ },
+ {
+ "epoch": 7.3633187772925766,
+ "grad_norm": 0.01787244901061058,
+ "learning_rate": 0.0006,
+ "loss": 5.450218677520752,
+ "step": 530
+ },
+ {
+ "epoch": 7.377292576419214,
+ "grad_norm": 0.0170669574290514,
+ "learning_rate": 0.0006,
+ "loss": 5.481570243835449,
+ "step": 531
+ },
+ {
+ "epoch": 7.391266375545851,
+ "grad_norm": 0.01802165061235428,
+ "learning_rate": 0.0006,
+ "loss": 5.386394500732422,
+ "step": 532
+ },
+ {
+ "epoch": 7.405240174672489,
+ "grad_norm": 0.022949472069740295,
+ "learning_rate": 0.0006,
+ "loss": 5.424929618835449,
+ "step": 533
+ },
+ {
+ "epoch": 7.419213973799127,
+ "grad_norm": 0.034213390201330185,
+ "learning_rate": 0.0006,
+ "loss": 5.346663475036621,
+ "step": 534
+ },
+ {
+ "epoch": 7.4331877729257645,
+ "grad_norm": 0.03847593814134598,
+ "learning_rate": 0.0006,
+ "loss": 5.418482780456543,
+ "step": 535
+ },
+ {
+ "epoch": 7.447161572052401,
+ "grad_norm": 0.029393676668405533,
+ "learning_rate": 0.0006,
+ "loss": 5.456090927124023,
+ "step": 536
+ },
+ {
+ "epoch": 7.461135371179039,
+ "grad_norm": 0.03388667106628418,
+ "learning_rate": 0.0006,
+ "loss": 5.572983741760254,
+ "step": 537
+ },
+ {
+ "epoch": 7.475109170305677,
+ "grad_norm": 0.039690613746643066,
+ "learning_rate": 0.0006,
+ "loss": 5.414067268371582,
+ "step": 538
+ },
+ {
+ "epoch": 7.489082969432315,
+ "grad_norm": 0.03630776330828667,
+ "learning_rate": 0.0006,
+ "loss": 5.523739814758301,
+ "step": 539
+ },
+ {
+ "epoch": 7.503056768558952,
+ "grad_norm": 0.03356870263814926,
+ "learning_rate": 0.0006,
+ "loss": 5.444767951965332,
+ "step": 540
+ },
+ {
+ "epoch": 7.517030567685589,
+ "grad_norm": 0.030812304466962814,
+ "learning_rate": 0.0006,
+ "loss": 5.461244106292725,
+ "step": 541
+ },
+ {
+ "epoch": 7.531004366812227,
+ "grad_norm": 0.03319217637181282,
+ "learning_rate": 0.0006,
+ "loss": 5.502161026000977,
+ "step": 542
+ },
+ {
+ "epoch": 7.544978165938865,
+ "grad_norm": 0.032997481524944305,
+ "learning_rate": 0.0006,
+ "loss": 5.4162278175354,
+ "step": 543
+ },
+ {
+ "epoch": 7.558951965065502,
+ "grad_norm": 0.03364962339401245,
+ "learning_rate": 0.0006,
+ "loss": 5.415736198425293,
+ "step": 544
+ },
+ {
+ "epoch": 7.5729257641921395,
+ "grad_norm": 0.0344221405684948,
+ "learning_rate": 0.0006,
+ "loss": 5.541967868804932,
+ "step": 545
+ },
+ {
+ "epoch": 7.586899563318777,
+ "grad_norm": 0.029609503224492073,
+ "learning_rate": 0.0006,
+ "loss": 5.440614223480225,
+ "step": 546
+ },
+ {
+ "epoch": 7.600873362445415,
+ "grad_norm": 0.029055926948785782,
+ "learning_rate": 0.0006,
+ "loss": 5.462865352630615,
+ "step": 547
+ },
+ {
+ "epoch": 7.614847161572053,
+ "grad_norm": 0.02658848837018013,
+ "learning_rate": 0.0006,
+ "loss": 5.444467544555664,
+ "step": 548
+ },
+ {
+ "epoch": 7.62882096069869,
+ "grad_norm": 0.026276404038071632,
+ "learning_rate": 0.0006,
+ "loss": 5.462278842926025,
+ "step": 549
+ },
+ {
+ "epoch": 7.642794759825327,
+ "grad_norm": 0.0286889486014843,
+ "learning_rate": 0.0006,
+ "loss": 5.4849534034729,
+ "step": 550
+ },
+ {
+ "epoch": 7.656768558951965,
+ "grad_norm": 0.026849817484617233,
+ "learning_rate": 0.0006,
+ "loss": 5.431473731994629,
+ "step": 551
+ },
+ {
+ "epoch": 7.670742358078603,
+ "grad_norm": 0.02312396466732025,
+ "learning_rate": 0.0006,
+ "loss": 5.45462703704834,
+ "step": 552
+ },
+ {
+ "epoch": 7.68471615720524,
+ "grad_norm": 0.026197485625743866,
+ "learning_rate": 0.0006,
+ "loss": 5.452552318572998,
+ "step": 553
+ },
+ {
+ "epoch": 7.698689956331878,
+ "grad_norm": 0.02747255191206932,
+ "learning_rate": 0.0006,
+ "loss": 5.447169780731201,
+ "step": 554
+ },
+ {
+ "epoch": 7.712663755458515,
+ "grad_norm": 0.028123432770371437,
+ "learning_rate": 0.0006,
+ "loss": 5.394400596618652,
+ "step": 555
+ },
+ {
+ "epoch": 7.726637554585153,
+ "grad_norm": 0.02599870041012764,
+ "learning_rate": 0.0006,
+ "loss": 5.398341178894043,
+ "step": 556
+ },
+ {
+ "epoch": 7.74061135371179,
+ "grad_norm": 0.022171657532453537,
+ "learning_rate": 0.0006,
+ "loss": 5.368820667266846,
+ "step": 557
+ },
+ {
+ "epoch": 7.754585152838428,
+ "grad_norm": 0.022309480234980583,
+ "learning_rate": 0.0006,
+ "loss": 5.416772365570068,
+ "step": 558
+ },
+ {
+ "epoch": 7.7685589519650655,
+ "grad_norm": 0.024074165150523186,
+ "learning_rate": 0.0006,
+ "loss": 5.359002113342285,
+ "step": 559
+ },
+ {
+ "epoch": 7.782532751091703,
+ "grad_norm": 0.02636653557419777,
+ "learning_rate": 0.0006,
+ "loss": 5.484368324279785,
+ "step": 560
+ },
+ {
+ "epoch": 7.796506550218341,
+ "grad_norm": 0.02203752100467682,
+ "learning_rate": 0.0006,
+ "loss": 5.375498294830322,
+ "step": 561
+ },
+ {
+ "epoch": 7.810480349344978,
+ "grad_norm": 0.019924819469451904,
+ "learning_rate": 0.0006,
+ "loss": 5.334723472595215,
+ "step": 562
+ },
+ {
+ "epoch": 7.824454148471616,
+ "grad_norm": 0.018755966797471046,
+ "learning_rate": 0.0006,
+ "loss": 5.348781108856201,
+ "step": 563
+ },
+ {
+ "epoch": 7.8384279475982535,
+ "grad_norm": 0.016747845336794853,
+ "learning_rate": 0.0006,
+ "loss": 5.433079719543457,
+ "step": 564
+ },
+ {
+ "epoch": 7.85240174672489,
+ "grad_norm": 0.0170395877212286,
+ "learning_rate": 0.0006,
+ "loss": 5.346794128417969,
+ "step": 565
+ },
+ {
+ "epoch": 7.866375545851528,
+ "grad_norm": 0.016074176877737045,
+ "learning_rate": 0.0006,
+ "loss": 5.3277740478515625,
+ "step": 566
+ },
+ {
+ "epoch": 7.880349344978166,
+ "grad_norm": 0.014119806699454784,
+ "learning_rate": 0.0006,
+ "loss": 5.483551979064941,
+ "step": 567
+ },
+ {
+ "epoch": 7.894323144104804,
+ "grad_norm": 0.015270394273102283,
+ "learning_rate": 0.0006,
+ "loss": 5.378519535064697,
+ "step": 568
+ },
+ {
+ "epoch": 7.908296943231441,
+ "grad_norm": 0.015677539631724358,
+ "learning_rate": 0.0006,
+ "loss": 5.290210247039795,
+ "step": 569
+ },
+ {
+ "epoch": 7.922270742358078,
+ "grad_norm": 0.015930423513054848,
+ "learning_rate": 0.0006,
+ "loss": 5.427360534667969,
+ "step": 570
+ },
+ {
+ "epoch": 7.936244541484716,
+ "grad_norm": 0.016641564667224884,
+ "learning_rate": 0.0006,
+ "loss": 5.301599979400635,
+ "step": 571
+ },
+ {
+ "epoch": 7.950218340611354,
+ "grad_norm": 0.018293552100658417,
+ "learning_rate": 0.0006,
+ "loss": 5.277889251708984,
+ "step": 572
+ },
+ {
+ "epoch": 7.964192139737992,
+ "grad_norm": 0.017618799582123756,
+ "learning_rate": 0.0006,
+ "loss": 5.297194480895996,
+ "step": 573
+ },
+ {
+ "epoch": 7.978165938864628,
+ "grad_norm": 0.015034242533147335,
+ "learning_rate": 0.0006,
+ "loss": 5.42428731918335,
+ "step": 574
+ },
+ {
+ "epoch": 7.992139737991266,
+ "grad_norm": 0.01617511734366417,
+ "learning_rate": 0.0006,
+ "loss": 5.420044898986816,
+ "step": 575
+ },
+ {
+ "epoch": 8.0,
+ "grad_norm": 0.01837257295846939,
+ "learning_rate": 0.0006,
+ "loss": 5.240387916564941,
+ "step": 576
+ },
+ {
+ "epoch": 8.0,
+ "eval_loss": 5.423932075500488,
+ "eval_runtime": 58.0841,
+ "eval_samples_per_second": 42.042,
+ "eval_steps_per_second": 1.326,
+ "step": 576
+ },
+ {
+ "epoch": 8.013973799126637,
+ "grad_norm": 0.020953809842467308,
+ "learning_rate": 0.0006,
+ "loss": 5.4320783615112305,
+ "step": 577
+ },
+ {
+ "epoch": 8.027947598253276,
+ "grad_norm": 0.0227745920419693,
+ "learning_rate": 0.0006,
+ "loss": 5.310197830200195,
+ "step": 578
+ },
+ {
+ "epoch": 8.041921397379912,
+ "grad_norm": 0.021372603252530098,
+ "learning_rate": 0.0006,
+ "loss": 5.339468002319336,
+ "step": 579
+ },
+ {
+ "epoch": 8.055895196506551,
+ "grad_norm": 0.022401731461286545,
+ "learning_rate": 0.0006,
+ "loss": 5.279829978942871,
+ "step": 580
+ },
+ {
+ "epoch": 8.069868995633188,
+ "grad_norm": 0.023942379280924797,
+ "learning_rate": 0.0006,
+ "loss": 5.372323989868164,
+ "step": 581
+ },
+ {
+ "epoch": 8.083842794759825,
+ "grad_norm": 0.024380534887313843,
+ "learning_rate": 0.0006,
+ "loss": 5.253486633300781,
+ "step": 582
+ },
+ {
+ "epoch": 8.097816593886463,
+ "grad_norm": 0.03246625140309334,
+ "learning_rate": 0.0006,
+ "loss": 5.305774211883545,
+ "step": 583
+ },
+ {
+ "epoch": 8.1117903930131,
+ "grad_norm": 0.030399736016988754,
+ "learning_rate": 0.0006,
+ "loss": 5.370976448059082,
+ "step": 584
+ },
+ {
+ "epoch": 8.125764192139737,
+ "grad_norm": 0.027148913592100143,
+ "learning_rate": 0.0006,
+ "loss": 5.368010520935059,
+ "step": 585
+ },
+ {
+ "epoch": 8.139737991266376,
+ "grad_norm": 0.02925540879368782,
+ "learning_rate": 0.0006,
+ "loss": 5.303013801574707,
+ "step": 586
+ },
+ {
+ "epoch": 8.153711790393013,
+ "grad_norm": 0.02915453538298607,
+ "learning_rate": 0.0006,
+ "loss": 5.329667091369629,
+ "step": 587
+ },
+ {
+ "epoch": 8.167685589519651,
+ "grad_norm": 0.031732626259326935,
+ "learning_rate": 0.0006,
+ "loss": 5.3634934425354,
+ "step": 588
+ },
+ {
+ "epoch": 8.181659388646288,
+ "grad_norm": 0.036166731268167496,
+ "learning_rate": 0.0006,
+ "loss": 5.223980903625488,
+ "step": 589
+ },
+ {
+ "epoch": 8.195633187772925,
+ "grad_norm": 0.035654108971357346,
+ "learning_rate": 0.0006,
+ "loss": 5.361059188842773,
+ "step": 590
+ },
+ {
+ "epoch": 8.209606986899564,
+ "grad_norm": 0.0365324430167675,
+ "learning_rate": 0.0006,
+ "loss": 5.336151599884033,
+ "step": 591
+ },
+ {
+ "epoch": 8.2235807860262,
+ "grad_norm": 0.03578880429267883,
+ "learning_rate": 0.0006,
+ "loss": 5.434076309204102,
+ "step": 592
+ },
+ {
+ "epoch": 8.237554585152838,
+ "grad_norm": 0.035734012722969055,
+ "learning_rate": 0.0006,
+ "loss": 5.330893516540527,
+ "step": 593
+ },
+ {
+ "epoch": 8.251528384279476,
+ "grad_norm": 0.03196857124567032,
+ "learning_rate": 0.0006,
+ "loss": 5.306009769439697,
+ "step": 594
+ },
+ {
+ "epoch": 8.265502183406113,
+ "grad_norm": 0.032042086124420166,
+ "learning_rate": 0.0006,
+ "loss": 5.343267917633057,
+ "step": 595
+ },
+ {
+ "epoch": 8.279475982532752,
+ "grad_norm": 0.03160746395587921,
+ "learning_rate": 0.0006,
+ "loss": 5.2353715896606445,
+ "step": 596
+ },
+ {
+ "epoch": 8.293449781659389,
+ "grad_norm": 0.026689818128943443,
+ "learning_rate": 0.0006,
+ "loss": 5.254042625427246,
+ "step": 597
+ },
+ {
+ "epoch": 8.307423580786025,
+ "grad_norm": 0.02880188450217247,
+ "learning_rate": 0.0006,
+ "loss": 5.326833248138428,
+ "step": 598
+ },
+ {
+ "epoch": 8.321397379912664,
+ "grad_norm": 0.027516381815075874,
+ "learning_rate": 0.0006,
+ "loss": 5.342336177825928,
+ "step": 599
+ },
+ {
+ "epoch": 8.335371179039301,
+ "grad_norm": 0.027875030413269997,
+ "learning_rate": 0.0006,
+ "loss": 5.200719833374023,
+ "step": 600
+ },
+ {
+ "epoch": 8.34934497816594,
+ "grad_norm": 0.0268265288323164,
+ "learning_rate": 0.0006,
+ "loss": 5.249449729919434,
+ "step": 601
+ },
+ {
+ "epoch": 8.363318777292577,
+ "grad_norm": 0.024824608117341995,
+ "learning_rate": 0.0006,
+ "loss": 5.318952560424805,
+ "step": 602
+ },
+ {
+ "epoch": 8.377292576419213,
+ "grad_norm": 0.022990627214312553,
+ "learning_rate": 0.0006,
+ "loss": 5.244993209838867,
+ "step": 603
+ },
+ {
+ "epoch": 8.391266375545852,
+ "grad_norm": 0.022804604843258858,
+ "learning_rate": 0.0006,
+ "loss": 5.3236846923828125,
+ "step": 604
+ },
+ {
+ "epoch": 8.405240174672489,
+ "grad_norm": 0.02419872209429741,
+ "learning_rate": 0.0006,
+ "loss": 5.255486488342285,
+ "step": 605
+ },
+ {
+ "epoch": 8.419213973799126,
+ "grad_norm": 0.021952059119939804,
+ "learning_rate": 0.0006,
+ "loss": 5.3511552810668945,
+ "step": 606
+ },
+ {
+ "epoch": 8.433187772925764,
+ "grad_norm": 0.022375497967004776,
+ "learning_rate": 0.0006,
+ "loss": 5.294790267944336,
+ "step": 607
+ },
+ {
+ "epoch": 8.447161572052401,
+ "grad_norm": 0.019242815673351288,
+ "learning_rate": 0.0006,
+ "loss": 5.252618789672852,
+ "step": 608
+ },
+ {
+ "epoch": 8.46113537117904,
+ "grad_norm": 0.0173486340790987,
+ "learning_rate": 0.0006,
+ "loss": 5.30147647857666,
+ "step": 609
+ },
+ {
+ "epoch": 8.475109170305677,
+ "grad_norm": 0.01612604409456253,
+ "learning_rate": 0.0006,
+ "loss": 5.23960542678833,
+ "step": 610
+ },
+ {
+ "epoch": 8.489082969432314,
+ "grad_norm": 0.014293976128101349,
+ "learning_rate": 0.0006,
+ "loss": 5.119932651519775,
+ "step": 611
+ },
+ {
+ "epoch": 8.503056768558952,
+ "grad_norm": 0.014453536830842495,
+ "learning_rate": 0.0006,
+ "loss": 5.166794776916504,
+ "step": 612
+ },
+ {
+ "epoch": 8.51703056768559,
+ "grad_norm": 0.014685769565403461,
+ "learning_rate": 0.0006,
+ "loss": 5.264822959899902,
+ "step": 613
+ },
+ {
+ "epoch": 8.531004366812226,
+ "grad_norm": 0.01427740603685379,
+ "learning_rate": 0.0006,
+ "loss": 5.278433799743652,
+ "step": 614
+ },
+ {
+ "epoch": 8.544978165938865,
+ "grad_norm": 0.014864951372146606,
+ "learning_rate": 0.0006,
+ "loss": 5.214512348175049,
+ "step": 615
+ },
+ {
+ "epoch": 8.558951965065502,
+ "grad_norm": 0.01565164513885975,
+ "learning_rate": 0.0006,
+ "loss": 5.183066368103027,
+ "step": 616
+ },
+ {
+ "epoch": 8.57292576419214,
+ "grad_norm": 0.01779816672205925,
+ "learning_rate": 0.0006,
+ "loss": 5.163185119628906,
+ "step": 617
+ },
+ {
+ "epoch": 8.586899563318777,
+ "grad_norm": 0.017254827544093132,
+ "learning_rate": 0.0006,
+ "loss": 5.243555068969727,
+ "step": 618
+ },
+ {
+ "epoch": 8.600873362445414,
+ "grad_norm": 0.01650584116578102,
+ "learning_rate": 0.0006,
+ "loss": 5.234884738922119,
+ "step": 619
+ },
+ {
+ "epoch": 8.614847161572053,
+ "grad_norm": 0.017021115869283676,
+ "learning_rate": 0.0006,
+ "loss": 5.12271785736084,
+ "step": 620
+ },
+ {
+ "epoch": 8.62882096069869,
+ "grad_norm": 0.01773759163916111,
+ "learning_rate": 0.0006,
+ "loss": 5.242153167724609,
+ "step": 621
+ },
+ {
+ "epoch": 8.642794759825328,
+ "grad_norm": 0.015679042786359787,
+ "learning_rate": 0.0006,
+ "loss": 5.170779228210449,
+ "step": 622
+ },
+ {
+ "epoch": 8.656768558951965,
+ "grad_norm": 0.013760549947619438,
+ "learning_rate": 0.0006,
+ "loss": 5.238644599914551,
+ "step": 623
+ },
+ {
+ "epoch": 8.670742358078602,
+ "grad_norm": 0.014571522362530231,
+ "learning_rate": 0.0006,
+ "loss": 5.177056312561035,
+ "step": 624
+ },
+ {
+ "epoch": 8.68471615720524,
+ "grad_norm": 0.016209015622735023,
+ "learning_rate": 0.0006,
+ "loss": 5.163750648498535,
+ "step": 625
+ },
+ {
+ "epoch": 8.698689956331878,
+ "grad_norm": 0.016813941299915314,
+ "learning_rate": 0.0006,
+ "loss": 5.183428764343262,
+ "step": 626
+ },
+ {
+ "epoch": 8.712663755458514,
+ "grad_norm": 0.020985357463359833,
+ "learning_rate": 0.0006,
+ "loss": 5.229465007781982,
+ "step": 627
+ },
+ {
+ "epoch": 8.726637554585153,
+ "grad_norm": 0.02679632417857647,
+ "learning_rate": 0.0006,
+ "loss": 5.257368087768555,
+ "step": 628
+ },
+ {
+ "epoch": 8.74061135371179,
+ "grad_norm": 0.025756070390343666,
+ "learning_rate": 0.0006,
+ "loss": 5.253736972808838,
+ "step": 629
+ },
+ {
+ "epoch": 8.754585152838429,
+ "grad_norm": 0.02643490768969059,
+ "learning_rate": 0.0006,
+ "loss": 5.228633403778076,
+ "step": 630
+ },
+ {
+ "epoch": 8.768558951965066,
+ "grad_norm": 0.03303210437297821,
+ "learning_rate": 0.0006,
+ "loss": 5.163540363311768,
+ "step": 631
+ },
+ {
+ "epoch": 8.782532751091702,
+ "grad_norm": 0.030432431027293205,
+ "learning_rate": 0.0006,
+ "loss": 5.243169784545898,
+ "step": 632
+ },
+ {
+ "epoch": 8.796506550218341,
+ "grad_norm": 0.028631288558244705,
+ "learning_rate": 0.0006,
+ "loss": 5.166398048400879,
+ "step": 633
+ },
+ {
+ "epoch": 8.810480349344978,
+ "grad_norm": 0.026188310235738754,
+ "learning_rate": 0.0006,
+ "loss": 5.128122329711914,
+ "step": 634
+ },
+ {
+ "epoch": 8.824454148471617,
+ "grad_norm": 0.025308528915047646,
+ "learning_rate": 0.0006,
+ "loss": 5.211284637451172,
+ "step": 635
+ },
+ {
+ "epoch": 8.838427947598253,
+ "grad_norm": 0.020657729357481003,
+ "learning_rate": 0.0006,
+ "loss": 5.161575794219971,
+ "step": 636
+ },
+ {
+ "epoch": 8.85240174672489,
+ "grad_norm": 0.021901234984397888,
+ "learning_rate": 0.0006,
+ "loss": 5.20050048828125,
+ "step": 637
+ },
+ {
+ "epoch": 8.866375545851529,
+ "grad_norm": 0.021219318732619286,
+ "learning_rate": 0.0006,
+ "loss": 5.211699485778809,
+ "step": 638
+ },
+ {
+ "epoch": 8.880349344978166,
+ "grad_norm": 0.0200974028557539,
+ "learning_rate": 0.0006,
+ "loss": 5.092503547668457,
+ "step": 639
+ },
+ {
+ "epoch": 8.894323144104803,
+ "grad_norm": 0.023804882541298866,
+ "learning_rate": 0.0006,
+ "loss": 5.216068267822266,
+ "step": 640
+ },
+ {
+ "epoch": 8.908296943231441,
+ "grad_norm": 0.026088010519742966,
+ "learning_rate": 0.0006,
+ "loss": 5.155592918395996,
+ "step": 641
+ },
+ {
+ "epoch": 8.922270742358078,
+ "grad_norm": 0.02601276896893978,
+ "learning_rate": 0.0006,
+ "loss": 5.197238922119141,
+ "step": 642
+ },
+ {
+ "epoch": 8.936244541484717,
+ "grad_norm": 0.020387211814522743,
+ "learning_rate": 0.0006,
+ "loss": 5.1955437660217285,
+ "step": 643
+ },
+ {
+ "epoch": 8.950218340611354,
+ "grad_norm": 0.019214622676372528,
+ "learning_rate": 0.0006,
+ "loss": 5.206346035003662,
+ "step": 644
+ },
+ {
+ "epoch": 8.96419213973799,
+ "grad_norm": 0.019674314185976982,
+ "learning_rate": 0.0006,
+ "loss": 5.219857215881348,
+ "step": 645
+ },
+ {
+ "epoch": 8.97816593886463,
+ "grad_norm": 0.020318234339356422,
+ "learning_rate": 0.0006,
+ "loss": 5.191132545471191,
+ "step": 646
+ },
+ {
+ "epoch": 8.992139737991266,
+ "grad_norm": 0.021428676322102547,
+ "learning_rate": 0.0006,
+ "loss": 5.214695930480957,
+ "step": 647
+ },
+ {
+ "epoch": 9.0,
+ "grad_norm": 0.023235054686665535,
+ "learning_rate": 0.0006,
+ "loss": 5.229091644287109,
+ "step": 648
+ },
+ {
+ "epoch": 9.0,
+ "eval_loss": 5.289593696594238,
+ "eval_runtime": 56.9256,
+ "eval_samples_per_second": 42.898,
+ "eval_steps_per_second": 1.353,
+ "step": 648
+ },
+ {
+ "epoch": 9.013973799126637,
+ "grad_norm": 0.02320289984345436,
+ "learning_rate": 0.0006,
+ "loss": 4.983669281005859,
+ "step": 649
+ },
+ {
+ "epoch": 9.027947598253276,
+ "grad_norm": 0.018521282821893692,
+ "learning_rate": 0.0006,
+ "loss": 5.200575828552246,
+ "step": 650
+ },
+ {
+ "epoch": 9.041921397379912,
+ "grad_norm": 0.0180169939994812,
+ "learning_rate": 0.0006,
+ "loss": 5.148033618927002,
+ "step": 651
+ },
+ {
+ "epoch": 9.055895196506551,
+ "grad_norm": 0.019969860091805458,
+ "learning_rate": 0.0006,
+ "loss": 5.099125385284424,
+ "step": 652
+ },
+ {
+ "epoch": 9.069868995633188,
+ "grad_norm": 0.01643305830657482,
+ "learning_rate": 0.0006,
+ "loss": 5.092347145080566,
+ "step": 653
+ },
+ {
+ "epoch": 9.083842794759825,
+ "grad_norm": 0.016022363677620888,
+ "learning_rate": 0.0006,
+ "loss": 5.129279613494873,
+ "step": 654
+ },
+ {
+ "epoch": 9.097816593886463,
+ "grad_norm": 0.01657041162252426,
+ "learning_rate": 0.0006,
+ "loss": 5.159562587738037,
+ "step": 655
+ },
+ {
+ "epoch": 9.1117903930131,
+ "grad_norm": 0.017924228683114052,
+ "learning_rate": 0.0006,
+ "loss": 5.165497779846191,
+ "step": 656
+ },
+ {
+ "epoch": 9.125764192139737,
+ "grad_norm": 0.021592361852526665,
+ "learning_rate": 0.0006,
+ "loss": 5.204775810241699,
+ "step": 657
+ },
+ {
+ "epoch": 9.139737991266376,
+ "grad_norm": 0.026924120262265205,
+ "learning_rate": 0.0006,
+ "loss": 5.128296852111816,
+ "step": 658
+ },
+ {
+ "epoch": 9.153711790393013,
+ "grad_norm": 0.023156899958848953,
+ "learning_rate": 0.0006,
+ "loss": 5.1255340576171875,
+ "step": 659
+ },
+ {
+ "epoch": 9.167685589519651,
+ "grad_norm": 0.020670367404818535,
+ "learning_rate": 0.0006,
+ "loss": 5.128479957580566,
+ "step": 660
+ },
+ {
+ "epoch": 9.181659388646288,
+ "grad_norm": 0.025447173044085503,
+ "learning_rate": 0.0006,
+ "loss": 5.154821872711182,
+ "step": 661
+ },
+ {
+ "epoch": 9.195633187772925,
+ "grad_norm": 0.027940068393945694,
+ "learning_rate": 0.0006,
+ "loss": 5.143466472625732,
+ "step": 662
+ },
+ {
+ "epoch": 9.209606986899564,
+ "grad_norm": 0.026847844943404198,
+ "learning_rate": 0.0006,
+ "loss": 5.045665740966797,
+ "step": 663
+ },
+ {
+ "epoch": 9.2235807860262,
+ "grad_norm": 0.02340601570904255,
+ "learning_rate": 0.0006,
+ "loss": 5.21933126449585,
+ "step": 664
+ },
+ {
+ "epoch": 9.237554585152838,
+ "grad_norm": 0.02340371161699295,
+ "learning_rate": 0.0006,
+ "loss": 4.942949295043945,
+ "step": 665
+ },
+ {
+ "epoch": 9.251528384279476,
+ "grad_norm": 0.02221992425620556,
+ "learning_rate": 0.0006,
+ "loss": 5.197099208831787,
+ "step": 666
+ },
+ {
+ "epoch": 9.265502183406113,
+ "grad_norm": 0.023508677259087563,
+ "learning_rate": 0.0006,
+ "loss": 5.022897720336914,
+ "step": 667
+ },
+ {
+ "epoch": 9.279475982532752,
+ "grad_norm": 0.026120394468307495,
+ "learning_rate": 0.0006,
+ "loss": 5.087725639343262,
+ "step": 668
+ },
+ {
+ "epoch": 9.293449781659389,
+ "grad_norm": 0.027273228392004967,
+ "learning_rate": 0.0006,
+ "loss": 5.090963840484619,
+ "step": 669
+ },
+ {
+ "epoch": 9.307423580786025,
+ "grad_norm": 0.03241586685180664,
+ "learning_rate": 0.0006,
+ "loss": 5.047842979431152,
+ "step": 670
+ },
+ {
+ "epoch": 9.321397379912664,
+ "grad_norm": 0.030245667323470116,
+ "learning_rate": 0.0006,
+ "loss": 5.104803085327148,
+ "step": 671
+ },
+ {
+ "epoch": 9.335371179039301,
+ "grad_norm": 0.027698364108800888,
+ "learning_rate": 0.0006,
+ "loss": 5.0644097328186035,
+ "step": 672
+ },
+ {
+ "epoch": 9.34934497816594,
+ "grad_norm": 0.029692554846405983,
+ "learning_rate": 0.0006,
+ "loss": 5.111942291259766,
+ "step": 673
+ },
+ {
+ "epoch": 9.363318777292577,
+ "grad_norm": 0.03328656405210495,
+ "learning_rate": 0.0006,
+ "loss": 5.1574554443359375,
+ "step": 674
+ },
+ {
+ "epoch": 9.377292576419213,
+ "grad_norm": 0.031938180327415466,
+ "learning_rate": 0.0006,
+ "loss": 5.190371990203857,
+ "step": 675
+ },
+ {
+ "epoch": 9.391266375545852,
+ "grad_norm": 0.024858945980668068,
+ "learning_rate": 0.0006,
+ "loss": 5.121420860290527,
+ "step": 676
+ },
+ {
+ "epoch": 9.405240174672489,
+ "grad_norm": 0.023033203557133675,
+ "learning_rate": 0.0006,
+ "loss": 5.0974345207214355,
+ "step": 677
+ },
+ {
+ "epoch": 9.419213973799126,
+ "grad_norm": 0.022083545103669167,
+ "learning_rate": 0.0006,
+ "loss": 5.046319961547852,
+ "step": 678
+ },
+ {
+ "epoch": 9.433187772925764,
+ "grad_norm": 0.0228437427431345,
+ "learning_rate": 0.0006,
+ "loss": 5.119963645935059,
+ "step": 679
+ },
+ {
+ "epoch": 9.447161572052401,
+ "grad_norm": 0.0242351982742548,
+ "learning_rate": 0.0006,
+ "loss": 5.0444560050964355,
+ "step": 680
+ },
+ {
+ "epoch": 9.46113537117904,
+ "grad_norm": 0.02401185780763626,
+ "learning_rate": 0.0006,
+ "loss": 5.063238620758057,
+ "step": 681
+ },
+ {
+ "epoch": 9.475109170305677,
+ "grad_norm": 0.02560959942638874,
+ "learning_rate": 0.0006,
+ "loss": 5.035224914550781,
+ "step": 682
+ },
+ {
+ "epoch": 9.489082969432314,
+ "grad_norm": 0.024985626339912415,
+ "learning_rate": 0.0006,
+ "loss": 5.049284934997559,
+ "step": 683
+ },
+ {
+ "epoch": 9.503056768558952,
+ "grad_norm": 0.023178115487098694,
+ "learning_rate": 0.0006,
+ "loss": 5.2266621589660645,
+ "step": 684
+ },
+ {
+ "epoch": 9.51703056768559,
+ "grad_norm": 0.0215512253344059,
+ "learning_rate": 0.0006,
+ "loss": 5.1875715255737305,
+ "step": 685
+ },
+ {
+ "epoch": 9.531004366812226,
+ "grad_norm": 0.01662837713956833,
+ "learning_rate": 0.0006,
+ "loss": 5.137012481689453,
+ "step": 686
+ },
+ {
+ "epoch": 9.544978165938865,
+ "grad_norm": 0.01952764391899109,
+ "learning_rate": 0.0006,
+ "loss": 5.072500228881836,
+ "step": 687
+ },
+ {
+ "epoch": 9.558951965065502,
+ "grad_norm": 0.017521077767014503,
+ "learning_rate": 0.0006,
+ "loss": 5.045711994171143,
+ "step": 688
+ },
+ {
+ "epoch": 9.57292576419214,
+ "grad_norm": 0.017355265095829964,
+ "learning_rate": 0.0006,
+ "loss": 5.02881383895874,
+ "step": 689
+ },
+ {
+ "epoch": 9.586899563318777,
+ "grad_norm": 0.015587719157338142,
+ "learning_rate": 0.0006,
+ "loss": 5.132787704467773,
+ "step": 690
+ },
+ {
+ "epoch": 9.600873362445414,
+ "grad_norm": 0.017469845712184906,
+ "learning_rate": 0.0006,
+ "loss": 5.105447292327881,
+ "step": 691
+ },
+ {
+ "epoch": 9.614847161572053,
+ "grad_norm": 0.019630100578069687,
+ "learning_rate": 0.0006,
+ "loss": 5.032464981079102,
+ "step": 692
+ },
+ {
+ "epoch": 9.62882096069869,
+ "grad_norm": 0.01784994639456272,
+ "learning_rate": 0.0006,
+ "loss": 5.081345558166504,
+ "step": 693
+ },
+ {
+ "epoch": 9.642794759825328,
+ "grad_norm": 0.01872754842042923,
+ "learning_rate": 0.0006,
+ "loss": 4.9772114753723145,
+ "step": 694
+ },
+ {
+ "epoch": 9.656768558951965,
+ "grad_norm": 0.01999032311141491,
+ "learning_rate": 0.0006,
+ "loss": 5.089740753173828,
+ "step": 695
+ },
+ {
+ "epoch": 9.670742358078602,
+ "grad_norm": 0.01710767112672329,
+ "learning_rate": 0.0006,
+ "loss": 4.9869890213012695,
+ "step": 696
+ },
+ {
+ "epoch": 9.68471615720524,
+ "grad_norm": 0.015144161880016327,
+ "learning_rate": 0.0006,
+ "loss": 5.022564888000488,
+ "step": 697
+ },
+ {
+ "epoch": 9.698689956331878,
+ "grad_norm": 0.01618543453514576,
+ "learning_rate": 0.0006,
+ "loss": 5.035921573638916,
+ "step": 698
+ },
+ {
+ "epoch": 9.712663755458514,
+ "grad_norm": 0.014653387479484081,
+ "learning_rate": 0.0006,
+ "loss": 4.914456367492676,
+ "step": 699
+ },
+ {
+ "epoch": 9.726637554585153,
+ "grad_norm": 0.013434977270662785,
+ "learning_rate": 0.0006,
+ "loss": 5.063764572143555,
+ "step": 700
+ },
+ {
+ "epoch": 9.74061135371179,
+ "grad_norm": 0.015128864906728268,
+ "learning_rate": 0.0006,
+ "loss": 5.052367687225342,
+ "step": 701
+ },
+ {
+ "epoch": 9.754585152838429,
+ "grad_norm": 0.014210768043994904,
+ "learning_rate": 0.0006,
+ "loss": 5.016489028930664,
+ "step": 702
+ },
+ {
+ "epoch": 9.768558951965066,
+ "grad_norm": 0.015191680751740932,
+ "learning_rate": 0.0006,
+ "loss": 4.990176200866699,
+ "step": 703
+ },
+ {
+ "epoch": 9.782532751091702,
+ "grad_norm": 0.018285127356648445,
+ "learning_rate": 0.0006,
+ "loss": 5.015164375305176,
+ "step": 704
+ },
+ {
+ "epoch": 9.796506550218341,
+ "grad_norm": 0.0204079058021307,
+ "learning_rate": 0.0006,
+ "loss": 5.091142654418945,
+ "step": 705
+ },
+ {
+ "epoch": 9.810480349344978,
+ "grad_norm": 0.01980586163699627,
+ "learning_rate": 0.0006,
+ "loss": 5.03350830078125,
+ "step": 706
+ },
+ {
+ "epoch": 9.824454148471617,
+ "grad_norm": 0.02048729732632637,
+ "learning_rate": 0.0006,
+ "loss": 5.03611421585083,
+ "step": 707
+ },
+ {
+ "epoch": 9.838427947598253,
+ "grad_norm": 0.02058163844048977,
+ "learning_rate": 0.0006,
+ "loss": 4.9613447189331055,
+ "step": 708
+ },
+ {
+ "epoch": 9.85240174672489,
+ "grad_norm": 0.018004924058914185,
+ "learning_rate": 0.0006,
+ "loss": 4.966541767120361,
+ "step": 709
+ },
+ {
+ "epoch": 9.866375545851529,
+ "grad_norm": 0.0198152307420969,
+ "learning_rate": 0.0006,
+ "loss": 5.01750373840332,
+ "step": 710
+ },
+ {
+ "epoch": 9.880349344978166,
+ "grad_norm": 0.018532969057559967,
+ "learning_rate": 0.0006,
+ "loss": 5.097424507141113,
+ "step": 711
+ },
+ {
+ "epoch": 9.894323144104803,
+ "grad_norm": 0.018939971923828125,
+ "learning_rate": 0.0006,
+ "loss": 5.177223205566406,
+ "step": 712
+ },
+ {
+ "epoch": 9.908296943231441,
+ "grad_norm": 0.017969170585274696,
+ "learning_rate": 0.0006,
+ "loss": 5.118013858795166,
+ "step": 713
+ },
+ {
+ "epoch": 9.922270742358078,
+ "grad_norm": 0.018021270632743835,
+ "learning_rate": 0.0006,
+ "loss": 5.034787178039551,
+ "step": 714
+ },
+ {
+ "epoch": 9.936244541484717,
+ "grad_norm": 0.017009710893034935,
+ "learning_rate": 0.0006,
+ "loss": 5.003920555114746,
+ "step": 715
+ },
+ {
+ "epoch": 9.950218340611354,
+ "grad_norm": 0.017722049728035927,
+ "learning_rate": 0.0006,
+ "loss": 5.031939506530762,
+ "step": 716
+ },
+ {
+ "epoch": 9.96419213973799,
+ "grad_norm": 0.017134087160229683,
+ "learning_rate": 0.0006,
+ "loss": 5.093474388122559,
+ "step": 717
+ },
+ {
+ "epoch": 9.97816593886463,
+ "grad_norm": 0.018511613830924034,
+ "learning_rate": 0.0006,
+ "loss": 5.078604698181152,
+ "step": 718
+ },
+ {
+ "epoch": 9.992139737991266,
+ "grad_norm": 0.021206334233283997,
+ "learning_rate": 0.0006,
+ "loss": 5.052572727203369,
+ "step": 719
+ },
+ {
+ "epoch": 10.0,
+ "grad_norm": 0.02239062264561653,
+ "learning_rate": 0.0006,
+ "loss": 5.063547134399414,
+ "step": 720
+ },
+ {
+ "epoch": 10.0,
+ "eval_loss": 5.173044681549072,
+ "eval_runtime": 56.7287,
+ "eval_samples_per_second": 43.047,
+ "eval_steps_per_second": 1.357,
+ "step": 720
+ },
+ {
+ "epoch": 10.013973799126637,
+ "grad_norm": 0.020757969468832016,
+ "learning_rate": 0.0006,
+ "loss": 5.045648574829102,
+ "step": 721
+ },
+ {
+ "epoch": 10.027947598253276,
+ "grad_norm": 0.022697702050209045,
+ "learning_rate": 0.0006,
+ "loss": 4.993833065032959,
+ "step": 722
+ },
+ {
+ "epoch": 10.041921397379912,
+ "grad_norm": 0.02434331737458706,
+ "learning_rate": 0.0006,
+ "loss": 5.073219299316406,
+ "step": 723
+ },
+ {
+ "epoch": 10.055895196506551,
+ "grad_norm": 0.025499660521745682,
+ "learning_rate": 0.0006,
+ "loss": 5.019072532653809,
+ "step": 724
+ },
+ {
+ "epoch": 10.069868995633188,
+ "grad_norm": 0.029281053692102432,
+ "learning_rate": 0.0006,
+ "loss": 4.959851264953613,
+ "step": 725
+ },
+ {
+ "epoch": 10.083842794759825,
+ "grad_norm": 0.033878110349178314,
+ "learning_rate": 0.0006,
+ "loss": 5.132717609405518,
+ "step": 726
+ },
+ {
+ "epoch": 10.097816593886463,
+ "grad_norm": 0.03385764732956886,
+ "learning_rate": 0.0006,
+ "loss": 5.0482892990112305,
+ "step": 727
+ },
+ {
+ "epoch": 10.1117903930131,
+ "grad_norm": 0.03466648980975151,
+ "learning_rate": 0.0006,
+ "loss": 4.946855545043945,
+ "step": 728
+ },
+ {
+ "epoch": 10.125764192139737,
+ "grad_norm": 0.028546305373311043,
+ "learning_rate": 0.0006,
+ "loss": 5.027774333953857,
+ "step": 729
+ },
+ {
+ "epoch": 10.139737991266376,
+ "grad_norm": 0.030575869604945183,
+ "learning_rate": 0.0006,
+ "loss": 4.9768476486206055,
+ "step": 730
+ },
+ {
+ "epoch": 10.153711790393013,
+ "grad_norm": 0.03614303469657898,
+ "learning_rate": 0.0006,
+ "loss": 5.078113555908203,
+ "step": 731
+ },
+ {
+ "epoch": 10.167685589519651,
+ "grad_norm": 0.040991660207509995,
+ "learning_rate": 0.0006,
+ "loss": 5.033185958862305,
+ "step": 732
+ },
+ {
+ "epoch": 10.181659388646288,
+ "grad_norm": 0.04551699757575989,
+ "learning_rate": 0.0006,
+ "loss": 5.0820746421813965,
+ "step": 733
+ },
+ {
+ "epoch": 10.195633187772925,
+ "grad_norm": 0.040193989872932434,
+ "learning_rate": 0.0006,
+ "loss": 5.059700965881348,
+ "step": 734
+ },
+ {
+ "epoch": 10.209606986899564,
+ "grad_norm": 0.035851605236530304,
+ "learning_rate": 0.0006,
+ "loss": 5.142467498779297,
+ "step": 735
+ },
+ {
+ "epoch": 10.2235807860262,
+ "grad_norm": 0.034200169146060944,
+ "learning_rate": 0.0006,
+ "loss": 5.072312355041504,
+ "step": 736
+ },
+ {
+ "epoch": 10.237554585152838,
+ "grad_norm": 0.033621110022068024,
+ "learning_rate": 0.0006,
+ "loss": 5.08037805557251,
+ "step": 737
+ },
+ {
+ "epoch": 10.251528384279476,
+ "grad_norm": 0.037372369319200516,
+ "learning_rate": 0.0006,
+ "loss": 4.998425483703613,
+ "step": 738
+ },
+ {
+ "epoch": 10.265502183406113,
+ "grad_norm": 0.037044707685709,
+ "learning_rate": 0.0006,
+ "loss": 4.923993110656738,
+ "step": 739
+ },
+ {
+ "epoch": 10.279475982532752,
+ "grad_norm": 0.030454808846116066,
+ "learning_rate": 0.0006,
+ "loss": 5.0889997482299805,
+ "step": 740
+ },
+ {
+ "epoch": 10.293449781659389,
+ "grad_norm": 0.030969638377428055,
+ "learning_rate": 0.0006,
+ "loss": 4.984112739562988,
+ "step": 741
+ },
+ {
+ "epoch": 10.307423580786025,
+ "grad_norm": 0.0276983380317688,
+ "learning_rate": 0.0006,
+ "loss": 4.955412864685059,
+ "step": 742
+ },
+ {
+ "epoch": 10.321397379912664,
+ "grad_norm": 0.02716052532196045,
+ "learning_rate": 0.0006,
+ "loss": 5.0098876953125,
+ "step": 743
+ },
+ {
+ "epoch": 10.335371179039301,
+ "grad_norm": 0.02403552085161209,
+ "learning_rate": 0.0006,
+ "loss": 5.165160655975342,
+ "step": 744
+ },
+ {
+ "epoch": 10.34934497816594,
+ "grad_norm": 0.0257862601429224,
+ "learning_rate": 0.0006,
+ "loss": 5.001547813415527,
+ "step": 745
+ },
+ {
+ "epoch": 10.363318777292577,
+ "grad_norm": 0.02300378680229187,
+ "learning_rate": 0.0006,
+ "loss": 5.038888931274414,
+ "step": 746
+ },
+ {
+ "epoch": 10.377292576419213,
+ "grad_norm": 0.019766854122281075,
+ "learning_rate": 0.0006,
+ "loss": 4.924291610717773,
+ "step": 747
+ },
+ {
+ "epoch": 10.391266375545852,
+ "grad_norm": 0.01888395845890045,
+ "learning_rate": 0.0006,
+ "loss": 5.059971332550049,
+ "step": 748
+ },
+ {
+ "epoch": 10.405240174672489,
+ "grad_norm": 0.015577499754726887,
+ "learning_rate": 0.0006,
+ "loss": 4.987686634063721,
+ "step": 749
+ },
+ {
+ "epoch": 10.419213973799126,
+ "grad_norm": 0.016012346372008324,
+ "learning_rate": 0.0006,
+ "loss": 5.015157699584961,
+ "step": 750
+ },
+ {
+ "epoch": 10.433187772925764,
+ "grad_norm": 0.016277998685836792,
+ "learning_rate": 0.0006,
+ "loss": 5.004931449890137,
+ "step": 751
+ },
+ {
+ "epoch": 10.447161572052401,
+ "grad_norm": 0.017504561692476273,
+ "learning_rate": 0.0006,
+ "loss": 4.902827739715576,
+ "step": 752
+ },
+ {
+ "epoch": 10.46113537117904,
+ "grad_norm": 0.01569022797048092,
+ "learning_rate": 0.0006,
+ "loss": 4.954863548278809,
+ "step": 753
+ },
+ {
+ "epoch": 10.475109170305677,
+ "grad_norm": 0.014114422723650932,
+ "learning_rate": 0.0006,
+ "loss": 5.00656795501709,
+ "step": 754
+ },
+ {
+ "epoch": 10.489082969432314,
+ "grad_norm": 0.01573832333087921,
+ "learning_rate": 0.0006,
+ "loss": 4.978353023529053,
+ "step": 755
+ },
+ {
+ "epoch": 10.503056768558952,
+ "grad_norm": 0.016480036079883575,
+ "learning_rate": 0.0006,
+ "loss": 4.953327655792236,
+ "step": 756
+ },
+ {
+ "epoch": 10.51703056768559,
+ "grad_norm": 0.014008025638759136,
+ "learning_rate": 0.0006,
+ "loss": 5.050840377807617,
+ "step": 757
+ },
+ {
+ "epoch": 10.531004366812226,
+ "grad_norm": 0.013147998601198196,
+ "learning_rate": 0.0006,
+ "loss": 4.974964141845703,
+ "step": 758
+ },
+ {
+ "epoch": 10.544978165938865,
+ "grad_norm": 0.013807603158056736,
+ "learning_rate": 0.0006,
+ "loss": 4.927907943725586,
+ "step": 759
+ },
+ {
+ "epoch": 10.558951965065502,
+ "grad_norm": 0.013555224984884262,
+ "learning_rate": 0.0006,
+ "loss": 4.980690956115723,
+ "step": 760
+ },
+ {
+ "epoch": 10.57292576419214,
+ "grad_norm": 0.013045408762991428,
+ "learning_rate": 0.0006,
+ "loss": 4.9436116218566895,
+ "step": 761
+ },
+ {
+ "epoch": 10.586899563318777,
+ "grad_norm": 0.012858862057328224,
+ "learning_rate": 0.0006,
+ "loss": 4.9595770835876465,
+ "step": 762
+ },
+ {
+ "epoch": 10.600873362445414,
+ "grad_norm": 0.01468253880739212,
+ "learning_rate": 0.0006,
+ "loss": 4.90725040435791,
+ "step": 763
+ },
+ {
+ "epoch": 10.614847161572053,
+ "grad_norm": 0.013224679045379162,
+ "learning_rate": 0.0006,
+ "loss": 4.877615928649902,
+ "step": 764
+ },
+ {
+ "epoch": 10.62882096069869,
+ "grad_norm": 0.013555348850786686,
+ "learning_rate": 0.0006,
+ "loss": 4.9456610679626465,
+ "step": 765
+ },
+ {
+ "epoch": 10.642794759825328,
+ "grad_norm": 0.014387160539627075,
+ "learning_rate": 0.0006,
+ "loss": 4.9233717918396,
+ "step": 766
+ },
+ {
+ "epoch": 10.656768558951965,
+ "grad_norm": 0.014354088343679905,
+ "learning_rate": 0.0006,
+ "loss": 4.983511924743652,
+ "step": 767
+ },
+ {
+ "epoch": 10.670742358078602,
+ "grad_norm": 0.01332700252532959,
+ "learning_rate": 0.0006,
+ "loss": 4.934416770935059,
+ "step": 768
+ },
+ {
+ "epoch": 10.68471615720524,
+ "grad_norm": 0.011543313041329384,
+ "learning_rate": 0.0006,
+ "loss": 4.96388053894043,
+ "step": 769
+ },
+ {
+ "epoch": 10.698689956331878,
+ "grad_norm": 0.012330878525972366,
+ "learning_rate": 0.0006,
+ "loss": 4.946396350860596,
+ "step": 770
+ },
+ {
+ "epoch": 10.712663755458514,
+ "grad_norm": 0.01379779726266861,
+ "learning_rate": 0.0006,
+ "loss": 4.981019020080566,
+ "step": 771
+ },
+ {
+ "epoch": 10.726637554585153,
+ "grad_norm": 0.014612431637942791,
+ "learning_rate": 0.0006,
+ "loss": 4.894567966461182,
+ "step": 772
+ },
+ {
+ "epoch": 10.74061135371179,
+ "grad_norm": 0.013062899932265282,
+ "learning_rate": 0.0006,
+ "loss": 5.033038139343262,
+ "step": 773
+ },
+ {
+ "epoch": 10.754585152838429,
+ "grad_norm": 0.013898049481213093,
+ "learning_rate": 0.0006,
+ "loss": 4.823636054992676,
+ "step": 774
+ },
+ {
+ "epoch": 10.768558951965066,
+ "grad_norm": 0.01502019353210926,
+ "learning_rate": 0.0006,
+ "loss": 5.020156383514404,
+ "step": 775
+ },
+ {
+ "epoch": 10.782532751091702,
+ "grad_norm": 0.013810204342007637,
+ "learning_rate": 0.0006,
+ "loss": 4.935647010803223,
+ "step": 776
+ },
+ {
+ "epoch": 10.796506550218341,
+ "grad_norm": 0.013616513460874557,
+ "learning_rate": 0.0006,
+ "loss": 4.901486396789551,
+ "step": 777
+ },
+ {
+ "epoch": 10.810480349344978,
+ "grad_norm": 0.013601024635136127,
+ "learning_rate": 0.0006,
+ "loss": 4.931873321533203,
+ "step": 778
+ },
+ {
+ "epoch": 10.824454148471617,
+ "grad_norm": 0.013668350875377655,
+ "learning_rate": 0.0006,
+ "loss": 4.937256813049316,
+ "step": 779
+ },
+ {
+ "epoch": 10.838427947598253,
+ "grad_norm": 0.01581823080778122,
+ "learning_rate": 0.0006,
+ "loss": 4.9611406326293945,
+ "step": 780
+ },
+ {
+ "epoch": 10.85240174672489,
+ "grad_norm": 0.019478484988212585,
+ "learning_rate": 0.0006,
+ "loss": 4.9212236404418945,
+ "step": 781
+ },
+ {
+ "epoch": 10.866375545851529,
+ "grad_norm": 0.026921333745121956,
+ "learning_rate": 0.0006,
+ "loss": 4.88501501083374,
+ "step": 782
+ },
+ {
+ "epoch": 10.880349344978166,
+ "grad_norm": 0.03155914694070816,
+ "learning_rate": 0.0006,
+ "loss": 4.989967346191406,
+ "step": 783
+ },
+ {
+ "epoch": 10.894323144104803,
+ "grad_norm": 0.029689429327845573,
+ "learning_rate": 0.0006,
+ "loss": 4.936000823974609,
+ "step": 784
+ },
+ {
+ "epoch": 10.908296943231441,
+ "grad_norm": 0.026626063510775566,
+ "learning_rate": 0.0006,
+ "loss": 4.984785079956055,
+ "step": 785
+ },
+ {
+ "epoch": 10.922270742358078,
+ "grad_norm": 0.026168212294578552,
+ "learning_rate": 0.0006,
+ "loss": 4.908282279968262,
+ "step": 786
+ },
+ {
+ "epoch": 10.936244541484717,
+ "grad_norm": 0.02828158251941204,
+ "learning_rate": 0.0006,
+ "loss": 4.926390647888184,
+ "step": 787
+ },
+ {
+ "epoch": 10.950218340611354,
+ "grad_norm": 0.02649078331887722,
+ "learning_rate": 0.0006,
+ "loss": 5.00400972366333,
+ "step": 788
+ },
+ {
+ "epoch": 10.96419213973799,
+ "grad_norm": 0.0256856270134449,
+ "learning_rate": 0.0006,
+ "loss": 5.0317487716674805,
+ "step": 789
+ },
+ {
+ "epoch": 10.97816593886463,
+ "grad_norm": 0.024728883057832718,
+ "learning_rate": 0.0006,
+ "loss": 4.9299635887146,
+ "step": 790
+ },
+ {
+ "epoch": 10.992139737991266,
+ "grad_norm": 0.026241116225719452,
+ "learning_rate": 0.0006,
+ "loss": 4.947319984436035,
+ "step": 791
+ },
+ {
+ "epoch": 11.0,
+ "grad_norm": 0.023526743054389954,
+ "learning_rate": 0.0006,
+ "loss": 4.94429349899292,
+ "step": 792
+ },
+ {
+ "epoch": 11.0,
+ "eval_loss": 5.104345798492432,
+ "eval_runtime": 57.0955,
+ "eval_samples_per_second": 42.77,
+ "eval_steps_per_second": 1.349,
+ "step": 792
+ },
+ {
+ "epoch": 11.013973799126637,
+ "grad_norm": 0.0216788612306118,
+ "learning_rate": 0.0006,
+ "loss": 4.881950855255127,
+ "step": 793
+ },
+ {
+ "epoch": 11.027947598253276,
+ "grad_norm": 0.01830463856458664,
+ "learning_rate": 0.0006,
+ "loss": 4.971016883850098,
+ "step": 794
+ },
+ {
+ "epoch": 11.041921397379912,
+ "grad_norm": 0.02204521745443344,
+ "learning_rate": 0.0006,
+ "loss": 4.894623756408691,
+ "step": 795
+ },
+ {
+ "epoch": 11.055895196506551,
+ "grad_norm": 0.0210255216807127,
+ "learning_rate": 0.0006,
+ "loss": 4.834009170532227,
+ "step": 796
+ },
+ {
+ "epoch": 11.069868995633188,
+ "grad_norm": 0.019797448068857193,
+ "learning_rate": 0.0006,
+ "loss": 5.086714744567871,
+ "step": 797
+ },
+ {
+ "epoch": 11.083842794759825,
+ "grad_norm": 0.02036641724407673,
+ "learning_rate": 0.0006,
+ "loss": 4.920056343078613,
+ "step": 798
+ },
+ {
+ "epoch": 11.097816593886463,
+ "grad_norm": 0.01966066285967827,
+ "learning_rate": 0.0006,
+ "loss": 4.879859447479248,
+ "step": 799
+ },
+ {
+ "epoch": 11.1117903930131,
+ "grad_norm": 0.0202149897813797,
+ "learning_rate": 0.0006,
+ "loss": 4.933640480041504,
+ "step": 800
+ },
+ {
+ "epoch": 11.125764192139737,
+ "grad_norm": 0.01893039606511593,
+ "learning_rate": 0.0006,
+ "loss": 4.901970863342285,
+ "step": 801
+ },
+ {
+ "epoch": 11.139737991266376,
+ "grad_norm": 0.016040155664086342,
+ "learning_rate": 0.0006,
+ "loss": 4.905491828918457,
+ "step": 802
+ },
+ {
+ "epoch": 11.153711790393013,
+ "grad_norm": 0.01720350608229637,
+ "learning_rate": 0.0006,
+ "loss": 4.798816680908203,
+ "step": 803
+ },
+ {
+ "epoch": 11.167685589519651,
+ "grad_norm": 0.017002852633595467,
+ "learning_rate": 0.0006,
+ "loss": 4.851859092712402,
+ "step": 804
+ },
+ {
+ "epoch": 11.181659388646288,
+ "grad_norm": 0.015594680793583393,
+ "learning_rate": 0.0006,
+ "loss": 4.90328311920166,
+ "step": 805
+ },
+ {
+ "epoch": 11.195633187772925,
+ "grad_norm": 0.014904462732374668,
+ "learning_rate": 0.0006,
+ "loss": 4.904215335845947,
+ "step": 806
+ },
+ {
+ "epoch": 11.209606986899564,
+ "grad_norm": 0.01366228237748146,
+ "learning_rate": 0.0006,
+ "loss": 4.9813055992126465,
+ "step": 807
+ },
+ {
+ "epoch": 11.2235807860262,
+ "grad_norm": 0.014719638973474503,
+ "learning_rate": 0.0006,
+ "loss": 4.927306175231934,
+ "step": 808
+ },
+ {
+ "epoch": 11.237554585152838,
+ "grad_norm": 0.015016036108136177,
+ "learning_rate": 0.0006,
+ "loss": 4.874215126037598,
+ "step": 809
+ },
+ {
+ "epoch": 11.251528384279476,
+ "grad_norm": 0.014925424009561539,
+ "learning_rate": 0.0006,
+ "loss": 4.838257789611816,
+ "step": 810
+ },
+ {
+ "epoch": 11.265502183406113,
+ "grad_norm": 0.014335823245346546,
+ "learning_rate": 0.0006,
+ "loss": 4.940608024597168,
+ "step": 811
+ },
+ {
+ "epoch": 11.279475982532752,
+ "grad_norm": 0.014987512491643429,
+ "learning_rate": 0.0006,
+ "loss": 4.831531524658203,
+ "step": 812
+ },
+ {
+ "epoch": 11.293449781659389,
+ "grad_norm": 0.016087457537651062,
+ "learning_rate": 0.0006,
+ "loss": 4.863227367401123,
+ "step": 813
+ },
+ {
+ "epoch": 11.307423580786025,
+ "grad_norm": 0.01671300269663334,
+ "learning_rate": 0.0006,
+ "loss": 4.8535661697387695,
+ "step": 814
+ },
+ {
+ "epoch": 11.321397379912664,
+ "grad_norm": 0.015229142270982265,
+ "learning_rate": 0.0006,
+ "loss": 4.948635101318359,
+ "step": 815
+ },
+ {
+ "epoch": 11.335371179039301,
+ "grad_norm": 0.01532587967813015,
+ "learning_rate": 0.0006,
+ "loss": 4.9035797119140625,
+ "step": 816
+ },
+ {
+ "epoch": 11.34934497816594,
+ "grad_norm": 0.015715476125478745,
+ "learning_rate": 0.0006,
+ "loss": 4.894709587097168,
+ "step": 817
+ },
+ {
+ "epoch": 11.363318777292577,
+ "grad_norm": 0.014156977646052837,
+ "learning_rate": 0.0006,
+ "loss": 4.960206985473633,
+ "step": 818
+ },
+ {
+ "epoch": 11.377292576419213,
+ "grad_norm": 0.017076566815376282,
+ "learning_rate": 0.0006,
+ "loss": 5.005850791931152,
+ "step": 819
+ },
+ {
+ "epoch": 11.391266375545852,
+ "grad_norm": 0.0172084029763937,
+ "learning_rate": 0.0006,
+ "loss": 4.903355598449707,
+ "step": 820
+ },
+ {
+ "epoch": 11.405240174672489,
+ "grad_norm": 0.0180258359760046,
+ "learning_rate": 0.0006,
+ "loss": 4.842667579650879,
+ "step": 821
+ },
+ {
+ "epoch": 11.419213973799126,
+ "grad_norm": 0.022130120545625687,
+ "learning_rate": 0.0006,
+ "loss": 4.799970626831055,
+ "step": 822
+ },
+ {
+ "epoch": 11.433187772925764,
+ "grad_norm": 0.02384897693991661,
+ "learning_rate": 0.0006,
+ "loss": 4.860746383666992,
+ "step": 823
+ },
+ {
+ "epoch": 11.447161572052401,
+ "grad_norm": 0.021881572902202606,
+ "learning_rate": 0.0006,
+ "loss": 4.918107032775879,
+ "step": 824
+ },
+ {
+ "epoch": 11.46113537117904,
+ "grad_norm": 0.023085080087184906,
+ "learning_rate": 0.0006,
+ "loss": 4.878946304321289,
+ "step": 825
+ },
+ {
+ "epoch": 11.475109170305677,
+ "grad_norm": 0.02291012555360794,
+ "learning_rate": 0.0006,
+ "loss": 4.958887100219727,
+ "step": 826
+ },
+ {
+ "epoch": 11.489082969432314,
+ "grad_norm": 0.0199968870729208,
+ "learning_rate": 0.0006,
+ "loss": 4.80955696105957,
+ "step": 827
+ },
+ {
+ "epoch": 11.503056768558952,
+ "grad_norm": 0.018702229484915733,
+ "learning_rate": 0.0006,
+ "loss": 5.0276408195495605,
+ "step": 828
+ },
+ {
+ "epoch": 11.51703056768559,
+ "grad_norm": 0.017787311226129532,
+ "learning_rate": 0.0006,
+ "loss": 4.96018123626709,
+ "step": 829
+ },
+ {
+ "epoch": 11.531004366812226,
+ "grad_norm": 0.01704501546919346,
+ "learning_rate": 0.0006,
+ "loss": 4.914553165435791,
+ "step": 830
+ },
+ {
+ "epoch": 11.544978165938865,
+ "grad_norm": 0.017519677057862282,
+ "learning_rate": 0.0006,
+ "loss": 4.798098087310791,
+ "step": 831
+ },
+ {
+ "epoch": 11.558951965065502,
+ "grad_norm": 0.018511991947889328,
+ "learning_rate": 0.0006,
+ "loss": 4.817785263061523,
+ "step": 832
+ },
+ {
+ "epoch": 11.57292576419214,
+ "grad_norm": 0.01876644790172577,
+ "learning_rate": 0.0006,
+ "loss": 4.818995475769043,
+ "step": 833
+ },
+ {
+ "epoch": 11.586899563318777,
+ "grad_norm": 0.01753568835556507,
+ "learning_rate": 0.0006,
+ "loss": 4.811938285827637,
+ "step": 834
+ },
+ {
+ "epoch": 11.600873362445414,
+ "grad_norm": 0.017046531662344933,
+ "learning_rate": 0.0006,
+ "loss": 4.914674282073975,
+ "step": 835
+ },
+ {
+ "epoch": 11.614847161572053,
+ "grad_norm": 0.01802094094455242,
+ "learning_rate": 0.0006,
+ "loss": 4.993284225463867,
+ "step": 836
+ },
+ {
+ "epoch": 11.62882096069869,
+ "grad_norm": 0.019493764266371727,
+ "learning_rate": 0.0006,
+ "loss": 4.895628452301025,
+ "step": 837
+ },
+ {
+ "epoch": 11.642794759825328,
+ "grad_norm": 0.019461210817098618,
+ "learning_rate": 0.0006,
+ "loss": 4.869269371032715,
+ "step": 838
+ },
+ {
+ "epoch": 11.656768558951965,
+ "grad_norm": 0.01981567218899727,
+ "learning_rate": 0.0006,
+ "loss": 4.863052845001221,
+ "step": 839
+ },
+ {
+ "epoch": 11.670742358078602,
+ "grad_norm": 0.019894491881132126,
+ "learning_rate": 0.0006,
+ "loss": 4.8775787353515625,
+ "step": 840
+ },
+ {
+ "epoch": 11.68471615720524,
+ "grad_norm": 0.01875786855816841,
+ "learning_rate": 0.0006,
+ "loss": 4.817763328552246,
+ "step": 841
+ },
+ {
+ "epoch": 11.698689956331878,
+ "grad_norm": 0.020926695317029953,
+ "learning_rate": 0.0006,
+ "loss": 4.7470526695251465,
+ "step": 842
+ },
+ {
+ "epoch": 11.712663755458514,
+ "grad_norm": 0.021477103233337402,
+ "learning_rate": 0.0006,
+ "loss": 4.762682914733887,
+ "step": 843
+ },
+ {
+ "epoch": 11.726637554585153,
+ "grad_norm": 0.020628711208701134,
+ "learning_rate": 0.0006,
+ "loss": 4.807985782623291,
+ "step": 844
+ },
+ {
+ "epoch": 11.74061135371179,
+ "grad_norm": 0.021489612758159637,
+ "learning_rate": 0.0006,
+ "loss": 4.843091011047363,
+ "step": 845
+ },
+ {
+ "epoch": 11.754585152838429,
+ "grad_norm": 0.021194491535425186,
+ "learning_rate": 0.0006,
+ "loss": 4.8300251960754395,
+ "step": 846
+ },
+ {
+ "epoch": 11.768558951965066,
+ "grad_norm": 0.01833350397646427,
+ "learning_rate": 0.0006,
+ "loss": 5.000202178955078,
+ "step": 847
+ },
+ {
+ "epoch": 11.782532751091702,
+ "grad_norm": 0.018096931278705597,
+ "learning_rate": 0.0006,
+ "loss": 4.880648136138916,
+ "step": 848
+ },
+ {
+ "epoch": 11.796506550218341,
+ "grad_norm": 0.0197161752730608,
+ "learning_rate": 0.0006,
+ "loss": 4.788316249847412,
+ "step": 849
+ },
+ {
+ "epoch": 11.810480349344978,
+ "grad_norm": 0.0203181654214859,
+ "learning_rate": 0.0006,
+ "loss": 4.870944976806641,
+ "step": 850
+ },
+ {
+ "epoch": 11.824454148471617,
+ "grad_norm": 0.01908455230295658,
+ "learning_rate": 0.0006,
+ "loss": 4.890144348144531,
+ "step": 851
+ },
+ {
+ "epoch": 11.838427947598253,
+ "grad_norm": 0.021281801164150238,
+ "learning_rate": 0.0006,
+ "loss": 4.89578914642334,
+ "step": 852
+ },
+ {
+ "epoch": 11.85240174672489,
+ "grad_norm": 0.02124273031949997,
+ "learning_rate": 0.0006,
+ "loss": 4.883749008178711,
+ "step": 853
+ },
+ {
+ "epoch": 11.866375545851529,
+ "grad_norm": 0.020969398319721222,
+ "learning_rate": 0.0006,
+ "loss": 4.809142589569092,
+ "step": 854
+ },
+ {
+ "epoch": 11.880349344978166,
+ "grad_norm": 0.01967657171189785,
+ "learning_rate": 0.0006,
+ "loss": 4.821864604949951,
+ "step": 855
+ },
+ {
+ "epoch": 11.894323144104803,
+ "grad_norm": 0.020887333899736404,
+ "learning_rate": 0.0006,
+ "loss": 4.917985916137695,
+ "step": 856
+ },
+ {
+ "epoch": 11.908296943231441,
+ "grad_norm": 0.019203083589673042,
+ "learning_rate": 0.0006,
+ "loss": 4.824960708618164,
+ "step": 857
+ },
+ {
+ "epoch": 11.922270742358078,
+ "grad_norm": 0.020162401720881462,
+ "learning_rate": 0.0006,
+ "loss": 4.772617816925049,
+ "step": 858
+ },
+ {
+ "epoch": 11.936244541484717,
+ "grad_norm": 0.02251487225294113,
+ "learning_rate": 0.0006,
+ "loss": 4.735418319702148,
+ "step": 859
+ },
+ {
+ "epoch": 11.950218340611354,
+ "grad_norm": 0.020313527435064316,
+ "learning_rate": 0.0006,
+ "loss": 4.939233779907227,
+ "step": 860
+ },
+ {
+ "epoch": 11.96419213973799,
+ "grad_norm": 0.021635599434375763,
+ "learning_rate": 0.0006,
+ "loss": 4.861807346343994,
+ "step": 861
+ },
+ {
+ "epoch": 11.97816593886463,
+ "grad_norm": 0.02285311557352543,
+ "learning_rate": 0.0006,
+ "loss": 4.932732582092285,
+ "step": 862
+ },
+ {
+ "epoch": 11.992139737991266,
+ "grad_norm": 0.02072535641491413,
+ "learning_rate": 0.0006,
+ "loss": 4.796767711639404,
+ "step": 863
+ },
+ {
+ "epoch": 12.0,
+ "grad_norm": 0.020588304847478867,
+ "learning_rate": 0.0006,
+ "loss": 4.997537612915039,
+ "step": 864
+ },
+ {
+ "epoch": 12.0,
+ "eval_loss": 5.044651985168457,
+ "eval_runtime": 56.9145,
+ "eval_samples_per_second": 42.906,
+ "eval_steps_per_second": 1.353,
+ "step": 864
+ },
+ {
+ "epoch": 12.013973799126637,
+ "grad_norm": 0.017597166821360588,
+ "learning_rate": 0.0006,
+ "loss": 4.830524444580078,
+ "step": 865
+ },
+ {
+ "epoch": 12.027947598253276,
+ "grad_norm": 0.01716788113117218,
+ "learning_rate": 0.0006,
+ "loss": 4.894325256347656,
+ "step": 866
+ },
+ {
+ "epoch": 12.041921397379912,
+ "grad_norm": 0.017712457105517387,
+ "learning_rate": 0.0006,
+ "loss": 4.868297576904297,
+ "step": 867
+ },
+ {
+ "epoch": 12.055895196506551,
+ "grad_norm": 0.021486390382051468,
+ "learning_rate": 0.0006,
+ "loss": 4.922049522399902,
+ "step": 868
+ },
+ {
+ "epoch": 12.069868995633188,
+ "grad_norm": 0.024560654535889626,
+ "learning_rate": 0.0006,
+ "loss": 4.734241008758545,
+ "step": 869
+ },
+ {
+ "epoch": 12.083842794759825,
+ "grad_norm": 0.025870252400636673,
+ "learning_rate": 0.0006,
+ "loss": 4.868566989898682,
+ "step": 870
+ },
+ {
+ "epoch": 12.097816593886463,
+ "grad_norm": 0.0277389045804739,
+ "learning_rate": 0.0006,
+ "loss": 4.755006790161133,
+ "step": 871
+ },
+ {
+ "epoch": 12.1117903930131,
+ "grad_norm": 0.029100844636559486,
+ "learning_rate": 0.0006,
+ "loss": 4.909667015075684,
+ "step": 872
+ },
+ {
+ "epoch": 12.125764192139737,
+ "grad_norm": 0.031833067536354065,
+ "learning_rate": 0.0006,
+ "loss": 4.858674049377441,
+ "step": 873
+ },
+ {
+ "epoch": 12.139737991266376,
+ "grad_norm": 0.032709237188100815,
+ "learning_rate": 0.0006,
+ "loss": 4.823696136474609,
+ "step": 874
+ },
+ {
+ "epoch": 12.153711790393013,
+ "grad_norm": 0.02995380386710167,
+ "learning_rate": 0.0006,
+ "loss": 4.850235462188721,
+ "step": 875
+ },
+ {
+ "epoch": 12.167685589519651,
+ "grad_norm": 0.034098610281944275,
+ "learning_rate": 0.0006,
+ "loss": 4.8169989585876465,
+ "step": 876
+ },
+ {
+ "epoch": 12.181659388646288,
+ "grad_norm": 0.03377080336213112,
+ "learning_rate": 0.0006,
+ "loss": 4.90738582611084,
+ "step": 877
+ },
+ {
+ "epoch": 12.195633187772925,
+ "grad_norm": 0.029780825600028038,
+ "learning_rate": 0.0006,
+ "loss": 4.811163902282715,
+ "step": 878
+ },
+ {
+ "epoch": 12.209606986899564,
+ "grad_norm": 0.03118024580180645,
+ "learning_rate": 0.0006,
+ "loss": 4.861286640167236,
+ "step": 879
+ },
+ {
+ "epoch": 12.2235807860262,
+ "grad_norm": 0.033052004873752594,
+ "learning_rate": 0.0006,
+ "loss": 4.865021228790283,
+ "step": 880
+ },
+ {
+ "epoch": 12.237554585152838,
+ "grad_norm": 0.029070032760500908,
+ "learning_rate": 0.0006,
+ "loss": 4.947303771972656,
+ "step": 881
+ },
+ {
+ "epoch": 12.251528384279476,
+ "grad_norm": 0.028355715796351433,
+ "learning_rate": 0.0006,
+ "loss": 4.804648399353027,
+ "step": 882
+ },
+ {
+ "epoch": 12.265502183406113,
+ "grad_norm": 0.031257227063179016,
+ "learning_rate": 0.0006,
+ "loss": 4.757718086242676,
+ "step": 883
+ },
+ {
+ "epoch": 12.279475982532752,
+ "grad_norm": 0.031438373029232025,
+ "learning_rate": 0.0006,
+ "loss": 4.851268768310547,
+ "step": 884
+ },
+ {
+ "epoch": 12.293449781659389,
+ "grad_norm": 0.030492138117551804,
+ "learning_rate": 0.0006,
+ "loss": 4.7891035079956055,
+ "step": 885
+ },
+ {
+ "epoch": 12.307423580786025,
+ "grad_norm": 0.026115991175174713,
+ "learning_rate": 0.0006,
+ "loss": 4.86474609375,
+ "step": 886
+ },
+ {
+ "epoch": 12.321397379912664,
+ "grad_norm": 0.024184465408325195,
+ "learning_rate": 0.0006,
+ "loss": 4.860456943511963,
+ "step": 887
+ },
+ {
+ "epoch": 12.335371179039301,
+ "grad_norm": 0.025437982752919197,
+ "learning_rate": 0.0006,
+ "loss": 4.806251049041748,
+ "step": 888
+ },
+ {
+ "epoch": 12.34934497816594,
+ "grad_norm": 0.022390231490135193,
+ "learning_rate": 0.0006,
+ "loss": 4.807086944580078,
+ "step": 889
+ },
+ {
+ "epoch": 12.363318777292577,
+ "grad_norm": 0.020174294710159302,
+ "learning_rate": 0.0006,
+ "loss": 4.78292179107666,
+ "step": 890
+ },
+ {
+ "epoch": 12.377292576419213,
+ "grad_norm": 0.019734324887394905,
+ "learning_rate": 0.0006,
+ "loss": 4.784509658813477,
+ "step": 891
+ },
+ {
+ "epoch": 12.391266375545852,
+ "grad_norm": 0.016733523458242416,
+ "learning_rate": 0.0006,
+ "loss": 4.811755180358887,
+ "step": 892
+ },
+ {
+ "epoch": 12.405240174672489,
+ "grad_norm": 0.01570146717131138,
+ "learning_rate": 0.0006,
+ "loss": 4.720088005065918,
+ "step": 893
+ },
+ {
+ "epoch": 12.419213973799126,
+ "grad_norm": 0.01601393334567547,
+ "learning_rate": 0.0006,
+ "loss": 4.783543109893799,
+ "step": 894
+ },
+ {
+ "epoch": 12.433187772925764,
+ "grad_norm": 0.014035487547516823,
+ "learning_rate": 0.0006,
+ "loss": 4.848340034484863,
+ "step": 895
+ },
+ {
+ "epoch": 12.447161572052401,
+ "grad_norm": 0.015248487703502178,
+ "learning_rate": 0.0006,
+ "loss": 4.773287773132324,
+ "step": 896
+ },
+ {
+ "epoch": 12.46113537117904,
+ "grad_norm": 0.01532268151640892,
+ "learning_rate": 0.0006,
+ "loss": 4.799646854400635,
+ "step": 897
+ },
+ {
+ "epoch": 12.475109170305677,
+ "grad_norm": 0.013982534408569336,
+ "learning_rate": 0.0006,
+ "loss": 4.8090500831604,
+ "step": 898
+ },
+ {
+ "epoch": 12.489082969432314,
+ "grad_norm": 0.014081164263188839,
+ "learning_rate": 0.0006,
+ "loss": 4.751888751983643,
+ "step": 899
+ },
+ {
+ "epoch": 12.503056768558952,
+ "grad_norm": 0.01380133256316185,
+ "learning_rate": 0.0006,
+ "loss": 4.78144645690918,
+ "step": 900
+ },
+ {
+ "epoch": 12.51703056768559,
+ "grad_norm": 0.013555269688367844,
+ "learning_rate": 0.0006,
+ "loss": 4.779257774353027,
+ "step": 901
+ },
+ {
+ "epoch": 12.531004366812226,
+ "grad_norm": 0.013075701892375946,
+ "learning_rate": 0.0006,
+ "loss": 4.820705413818359,
+ "step": 902
+ },
+ {
+ "epoch": 12.544978165938865,
+ "grad_norm": 0.01276316400617361,
+ "learning_rate": 0.0006,
+ "loss": 4.836733818054199,
+ "step": 903
+ },
+ {
+ "epoch": 12.558951965065502,
+ "grad_norm": 0.013033892959356308,
+ "learning_rate": 0.0006,
+ "loss": 4.764178276062012,
+ "step": 904
+ },
+ {
+ "epoch": 12.57292576419214,
+ "grad_norm": 0.013996980153024197,
+ "learning_rate": 0.0006,
+ "loss": 4.767983436584473,
+ "step": 905
+ },
+ {
+ "epoch": 12.586899563318777,
+ "grad_norm": 0.015222184360027313,
+ "learning_rate": 0.0006,
+ "loss": 4.759721755981445,
+ "step": 906
+ },
+ {
+ "epoch": 12.600873362445414,
+ "grad_norm": 0.01623941957950592,
+ "learning_rate": 0.0006,
+ "loss": 4.849811553955078,
+ "step": 907
+ },
+ {
+ "epoch": 12.614847161572053,
+ "grad_norm": 0.015711182728409767,
+ "learning_rate": 0.0006,
+ "loss": 4.7829084396362305,
+ "step": 908
+ },
+ {
+ "epoch": 12.62882096069869,
+ "grad_norm": 0.014137467369437218,
+ "learning_rate": 0.0006,
+ "loss": 4.73321533203125,
+ "step": 909
+ },
+ {
+ "epoch": 12.642794759825328,
+ "grad_norm": 0.012738462537527084,
+ "learning_rate": 0.0006,
+ "loss": 4.870052337646484,
+ "step": 910
+ },
+ {
+ "epoch": 12.656768558951965,
+ "grad_norm": 0.013683130033314228,
+ "learning_rate": 0.0006,
+ "loss": 4.82658052444458,
+ "step": 911
+ },
+ {
+ "epoch": 12.670742358078602,
+ "grad_norm": 0.014409830793738365,
+ "learning_rate": 0.0006,
+ "loss": 4.707494735717773,
+ "step": 912
+ },
+ {
+ "epoch": 12.68471615720524,
+ "grad_norm": 0.013806809671223164,
+ "learning_rate": 0.0006,
+ "loss": 4.778928279876709,
+ "step": 913
+ },
+ {
+ "epoch": 12.698689956331878,
+ "grad_norm": 0.012857912108302116,
+ "learning_rate": 0.0006,
+ "loss": 4.890717506408691,
+ "step": 914
+ },
+ {
+ "epoch": 12.712663755458514,
+ "grad_norm": 0.013828758150339127,
+ "learning_rate": 0.0006,
+ "loss": 4.8357439041137695,
+ "step": 915
+ },
+ {
+ "epoch": 12.726637554585153,
+ "grad_norm": 0.01478351280093193,
+ "learning_rate": 0.0006,
+ "loss": 4.869986534118652,
+ "step": 916
+ },
+ {
+ "epoch": 12.74061135371179,
+ "grad_norm": 0.01326421182602644,
+ "learning_rate": 0.0006,
+ "loss": 4.830646514892578,
+ "step": 917
+ },
+ {
+ "epoch": 12.754585152838429,
+ "grad_norm": 0.013335433788597584,
+ "learning_rate": 0.0006,
+ "loss": 4.7786149978637695,
+ "step": 918
+ },
+ {
+ "epoch": 12.768558951965066,
+ "grad_norm": 0.014261147007346153,
+ "learning_rate": 0.0006,
+ "loss": 4.788693428039551,
+ "step": 919
+ },
+ {
+ "epoch": 12.782532751091702,
+ "grad_norm": 0.014274783432483673,
+ "learning_rate": 0.0006,
+ "loss": 4.77295446395874,
+ "step": 920
+ },
+ {
+ "epoch": 12.796506550218341,
+ "grad_norm": 0.015074674971401691,
+ "learning_rate": 0.0006,
+ "loss": 4.830754280090332,
+ "step": 921
+ },
+ {
+ "epoch": 12.810480349344978,
+ "grad_norm": 0.015599515289068222,
+ "learning_rate": 0.0006,
+ "loss": 4.837759017944336,
+ "step": 922
+ },
+ {
+ "epoch": 12.824454148471617,
+ "grad_norm": 0.014595242217183113,
+ "learning_rate": 0.0006,
+ "loss": 4.882965564727783,
+ "step": 923
+ },
+ {
+ "epoch": 12.838427947598253,
+ "grad_norm": 0.014951467514038086,
+ "learning_rate": 0.0006,
+ "loss": 4.709748268127441,
+ "step": 924
+ },
+ {
+ "epoch": 12.85240174672489,
+ "grad_norm": 0.01539295818656683,
+ "learning_rate": 0.0006,
+ "loss": 4.791160583496094,
+ "step": 925
+ },
+ {
+ "epoch": 12.866375545851529,
+ "grad_norm": 0.015987906605005264,
+ "learning_rate": 0.0006,
+ "loss": 4.661876678466797,
+ "step": 926
+ },
+ {
+ "epoch": 12.880349344978166,
+ "grad_norm": 0.016527341678738594,
+ "learning_rate": 0.0006,
+ "loss": 4.721952438354492,
+ "step": 927
+ },
+ {
+ "epoch": 12.894323144104803,
+ "grad_norm": 0.01637161523103714,
+ "learning_rate": 0.0006,
+ "loss": 4.704446792602539,
+ "step": 928
+ },
+ {
+ "epoch": 12.908296943231441,
+ "grad_norm": 0.015221050009131432,
+ "learning_rate": 0.0006,
+ "loss": 4.771907806396484,
+ "step": 929
+ },
+ {
+ "epoch": 12.922270742358078,
+ "grad_norm": 0.016242166981101036,
+ "learning_rate": 0.0006,
+ "loss": 4.806267738342285,
+ "step": 930
+ },
+ {
+ "epoch": 12.936244541484717,
+ "grad_norm": 0.01628357358276844,
+ "learning_rate": 0.0006,
+ "loss": 4.7252326011657715,
+ "step": 931
+ },
+ {
+ "epoch": 12.950218340611354,
+ "grad_norm": 0.017428133636713028,
+ "learning_rate": 0.0006,
+ "loss": 4.765689849853516,
+ "step": 932
+ },
+ {
+ "epoch": 12.96419213973799,
+ "grad_norm": 0.01870061084628105,
+ "learning_rate": 0.0006,
+ "loss": 4.814488887786865,
+ "step": 933
+ },
+ {
+ "epoch": 12.97816593886463,
+ "grad_norm": 0.019796060398221016,
+ "learning_rate": 0.0006,
+ "loss": 4.739185810089111,
+ "step": 934
+ },
+ {
+ "epoch": 12.992139737991266,
+ "grad_norm": 0.01985820196568966,
+ "learning_rate": 0.0006,
+ "loss": 4.810608863830566,
+ "step": 935
+ },
+ {
+ "epoch": 13.0,
+ "grad_norm": 0.020834477618336678,
+ "learning_rate": 0.0006,
+ "loss": 4.727858066558838,
+ "step": 936
+ },
+ {
+ "epoch": 13.0,
+ "eval_loss": 4.979116916656494,
+ "eval_runtime": 57.2877,
+ "eval_samples_per_second": 42.627,
+ "eval_steps_per_second": 1.344,
+ "step": 936
+ },
+ {
+ "epoch": 13.013973799126637,
+ "grad_norm": 0.02025519125163555,
+ "learning_rate": 0.0006,
+ "loss": 4.628875732421875,
+ "step": 937
+ },
+ {
+ "epoch": 13.027947598253276,
+ "grad_norm": 0.019596368074417114,
+ "learning_rate": 0.0006,
+ "loss": 4.694699287414551,
+ "step": 938
+ },
+ {
+ "epoch": 13.041921397379912,
+ "grad_norm": 0.020494934171438217,
+ "learning_rate": 0.0006,
+ "loss": 4.691821098327637,
+ "step": 939
+ },
+ {
+ "epoch": 13.055895196506551,
+ "grad_norm": 0.020121760666370392,
+ "learning_rate": 0.0006,
+ "loss": 4.71478271484375,
+ "step": 940
+ },
+ {
+ "epoch": 13.069868995633188,
+ "grad_norm": 0.021091405302286148,
+ "learning_rate": 0.0006,
+ "loss": 4.695915222167969,
+ "step": 941
+ },
+ {
+ "epoch": 13.083842794759825,
+ "grad_norm": 0.019787423312664032,
+ "learning_rate": 0.0006,
+ "loss": 4.8652448654174805,
+ "step": 942
+ },
+ {
+ "epoch": 13.097816593886463,
+ "grad_norm": 0.0185833927243948,
+ "learning_rate": 0.0006,
+ "loss": 4.7065887451171875,
+ "step": 943
+ },
+ {
+ "epoch": 13.1117903930131,
+ "grad_norm": 0.021220475435256958,
+ "learning_rate": 0.0006,
+ "loss": 4.741232395172119,
+ "step": 944
+ },
+ {
+ "epoch": 13.125764192139737,
+ "grad_norm": 0.02136622555553913,
+ "learning_rate": 0.0006,
+ "loss": 4.705967903137207,
+ "step": 945
+ },
+ {
+ "epoch": 13.139737991266376,
+ "grad_norm": 0.0191587433218956,
+ "learning_rate": 0.0006,
+ "loss": 4.758667945861816,
+ "step": 946
+ },
+ {
+ "epoch": 13.153711790393013,
+ "grad_norm": 0.018290022388100624,
+ "learning_rate": 0.0006,
+ "loss": 4.760666847229004,
+ "step": 947
+ },
+ {
+ "epoch": 13.167685589519651,
+ "grad_norm": 0.021704409271478653,
+ "learning_rate": 0.0006,
+ "loss": 4.750555992126465,
+ "step": 948
+ },
+ {
+ "epoch": 13.181659388646288,
+ "grad_norm": 0.021668918430805206,
+ "learning_rate": 0.0006,
+ "loss": 4.840903282165527,
+ "step": 949
+ },
+ {
+ "epoch": 13.195633187772925,
+ "grad_norm": 0.019497141242027283,
+ "learning_rate": 0.0006,
+ "loss": 4.738969326019287,
+ "step": 950
+ },
+ {
+ "epoch": 13.209606986899564,
+ "grad_norm": 0.018500596284866333,
+ "learning_rate": 0.0006,
+ "loss": 4.75154972076416,
+ "step": 951
+ },
+ {
+ "epoch": 13.2235807860262,
+ "grad_norm": 0.01872239261865616,
+ "learning_rate": 0.0006,
+ "loss": 4.811738014221191,
+ "step": 952
+ },
+ {
+ "epoch": 13.237554585152838,
+ "grad_norm": 0.020321547985076904,
+ "learning_rate": 0.0006,
+ "loss": 4.815384387969971,
+ "step": 953
+ },
+ {
+ "epoch": 13.251528384279476,
+ "grad_norm": 0.02283575013279915,
+ "learning_rate": 0.0006,
+ "loss": 4.74582052230835,
+ "step": 954
+ },
+ {
+ "epoch": 13.265502183406113,
+ "grad_norm": 0.02474398910999298,
+ "learning_rate": 0.0006,
+ "loss": 4.738375186920166,
+ "step": 955
+ },
+ {
+ "epoch": 13.279475982532752,
+ "grad_norm": 0.02065764181315899,
+ "learning_rate": 0.0006,
+ "loss": 4.746875286102295,
+ "step": 956
+ },
+ {
+ "epoch": 13.293449781659389,
+ "grad_norm": 0.020534520968794823,
+ "learning_rate": 0.0006,
+ "loss": 4.656914710998535,
+ "step": 957
+ },
+ {
+ "epoch": 13.307423580786025,
+ "grad_norm": 0.0206378772854805,
+ "learning_rate": 0.0006,
+ "loss": 4.842615604400635,
+ "step": 958
+ },
+ {
+ "epoch": 13.321397379912664,
+ "grad_norm": 0.017935654148459435,
+ "learning_rate": 0.0006,
+ "loss": 4.717827796936035,
+ "step": 959
+ },
+ {
+ "epoch": 13.335371179039301,
+ "grad_norm": 0.016958363354206085,
+ "learning_rate": 0.0006,
+ "loss": 4.681070327758789,
+ "step": 960
+ },
+ {
+ "epoch": 13.34934497816594,
+ "grad_norm": 0.015735412016510963,
+ "learning_rate": 0.0006,
+ "loss": 4.7957048416137695,
+ "step": 961
+ },
+ {
+ "epoch": 13.363318777292577,
+ "grad_norm": 0.01563699170947075,
+ "learning_rate": 0.0006,
+ "loss": 4.701192378997803,
+ "step": 962
+ },
+ {
+ "epoch": 13.377292576419213,
+ "grad_norm": 0.015207760035991669,
+ "learning_rate": 0.0006,
+ "loss": 4.73256254196167,
+ "step": 963
+ },
+ {
+ "epoch": 13.391266375545852,
+ "grad_norm": 0.015396883711218834,
+ "learning_rate": 0.0006,
+ "loss": 4.76412296295166,
+ "step": 964
+ },
+ {
+ "epoch": 13.405240174672489,
+ "grad_norm": 0.015690583735704422,
+ "learning_rate": 0.0006,
+ "loss": 4.791157245635986,
+ "step": 965
+ },
+ {
+ "epoch": 13.419213973799126,
+ "grad_norm": 0.01755724661052227,
+ "learning_rate": 0.0006,
+ "loss": 4.845666885375977,
+ "step": 966
+ },
+ {
+ "epoch": 13.433187772925764,
+ "grad_norm": 0.019443422555923462,
+ "learning_rate": 0.0006,
+ "loss": 4.5902299880981445,
+ "step": 967
+ },
+ {
+ "epoch": 13.447161572052401,
+ "grad_norm": 0.02118338644504547,
+ "learning_rate": 0.0006,
+ "loss": 4.740627288818359,
+ "step": 968
+ },
+ {
+ "epoch": 13.46113537117904,
+ "grad_norm": 0.02174839936196804,
+ "learning_rate": 0.0006,
+ "loss": 4.814560890197754,
+ "step": 969
+ },
+ {
+ "epoch": 13.475109170305677,
+ "grad_norm": 0.022860374301671982,
+ "learning_rate": 0.0006,
+ "loss": 4.66440486907959,
+ "step": 970
+ },
+ {
+ "epoch": 13.489082969432314,
+ "grad_norm": 0.024742677807807922,
+ "learning_rate": 0.0006,
+ "loss": 4.7451653480529785,
+ "step": 971
+ },
+ {
+ "epoch": 13.503056768558952,
+ "grad_norm": 0.027068907395005226,
+ "learning_rate": 0.0006,
+ "loss": 4.802584171295166,
+ "step": 972
+ },
+ {
+ "epoch": 13.51703056768559,
+ "grad_norm": 0.026316625997424126,
+ "learning_rate": 0.0006,
+ "loss": 4.842397689819336,
+ "step": 973
+ },
+ {
+ "epoch": 13.531004366812226,
+ "grad_norm": 0.024610962718725204,
+ "learning_rate": 0.0006,
+ "loss": 4.650721073150635,
+ "step": 974
+ },
+ {
+ "epoch": 13.544978165938865,
+ "grad_norm": 0.024122603237628937,
+ "learning_rate": 0.0006,
+ "loss": 4.6984357833862305,
+ "step": 975
+ },
+ {
+ "epoch": 13.558951965065502,
+ "grad_norm": 0.022699708119034767,
+ "learning_rate": 0.0006,
+ "loss": 4.793007850646973,
+ "step": 976
+ },
+ {
+ "epoch": 13.57292576419214,
+ "grad_norm": 0.021706534549593925,
+ "learning_rate": 0.0006,
+ "loss": 4.691362380981445,
+ "step": 977
+ },
+ {
+ "epoch": 13.586899563318777,
+ "grad_norm": 0.021363291889429092,
+ "learning_rate": 0.0006,
+ "loss": 4.76762580871582,
+ "step": 978
+ },
+ {
+ "epoch": 13.600873362445414,
+ "grad_norm": 0.023476677015423775,
+ "learning_rate": 0.0006,
+ "loss": 4.777202606201172,
+ "step": 979
+ },
+ {
+ "epoch": 13.614847161572053,
+ "grad_norm": 0.02457100711762905,
+ "learning_rate": 0.0006,
+ "loss": 4.590703010559082,
+ "step": 980
+ },
+ {
+ "epoch": 13.62882096069869,
+ "grad_norm": 0.02469099499285221,
+ "learning_rate": 0.0006,
+ "loss": 4.640185356140137,
+ "step": 981
+ },
+ {
+ "epoch": 13.642794759825328,
+ "grad_norm": 0.024914603680372238,
+ "learning_rate": 0.0006,
+ "loss": 4.772201061248779,
+ "step": 982
+ },
+ {
+ "epoch": 13.656768558951965,
+ "grad_norm": 0.025133173912763596,
+ "learning_rate": 0.0006,
+ "loss": 4.743337154388428,
+ "step": 983
+ },
+ {
+ "epoch": 13.670742358078602,
+ "grad_norm": 0.02034137211740017,
+ "learning_rate": 0.0006,
+ "loss": 4.752127647399902,
+ "step": 984
+ },
+ {
+ "epoch": 13.68471615720524,
+ "grad_norm": 0.019645661115646362,
+ "learning_rate": 0.0006,
+ "loss": 4.838453769683838,
+ "step": 985
+ },
+ {
+ "epoch": 13.698689956331878,
+ "grad_norm": 0.01900297962129116,
+ "learning_rate": 0.0006,
+ "loss": 4.804868221282959,
+ "step": 986
+ },
+ {
+ "epoch": 13.712663755458514,
+ "grad_norm": 0.020292341709136963,
+ "learning_rate": 0.0006,
+ "loss": 4.609743118286133,
+ "step": 987
+ },
+ {
+ "epoch": 13.726637554585153,
+ "grad_norm": 0.017014775425195694,
+ "learning_rate": 0.0006,
+ "loss": 4.740612983703613,
+ "step": 988
+ },
+ {
+ "epoch": 13.74061135371179,
+ "grad_norm": 0.015199844725430012,
+ "learning_rate": 0.0006,
+ "loss": 4.719701766967773,
+ "step": 989
+ },
+ {
+ "epoch": 13.754585152838429,
+ "grad_norm": 0.015125414356589317,
+ "learning_rate": 0.0006,
+ "loss": 4.686114311218262,
+ "step": 990
+ },
+ {
+ "epoch": 13.768558951965066,
+ "grad_norm": 0.014779540710151196,
+ "learning_rate": 0.0006,
+ "loss": 4.674347877502441,
+ "step": 991
+ },
+ {
+ "epoch": 13.782532751091702,
+ "grad_norm": 0.015255163423717022,
+ "learning_rate": 0.0006,
+ "loss": 4.6309309005737305,
+ "step": 992
+ },
+ {
+ "epoch": 13.796506550218341,
+ "grad_norm": 0.013692197389900684,
+ "learning_rate": 0.0006,
+ "loss": 4.698497295379639,
+ "step": 993
+ },
+ {
+ "epoch": 13.810480349344978,
+ "grad_norm": 0.014529623091220856,
+ "learning_rate": 0.0006,
+ "loss": 4.628443717956543,
+ "step": 994
+ },
+ {
+ "epoch": 13.824454148471617,
+ "grad_norm": 0.01576659083366394,
+ "learning_rate": 0.0006,
+ "loss": 4.788015842437744,
+ "step": 995
+ },
+ {
+ "epoch": 13.838427947598253,
+ "grad_norm": 0.015834277495741844,
+ "learning_rate": 0.0006,
+ "loss": 4.672703742980957,
+ "step": 996
+ },
+ {
+ "epoch": 13.85240174672489,
+ "grad_norm": 0.015539503656327724,
+ "learning_rate": 0.0006,
+ "loss": 4.731503486633301,
+ "step": 997
+ },
+ {
+ "epoch": 13.866375545851529,
+ "grad_norm": 0.013463746756315231,
+ "learning_rate": 0.0006,
+ "loss": 4.7006330490112305,
+ "step": 998
+ },
+ {
+ "epoch": 13.880349344978166,
+ "grad_norm": 0.016377726569771767,
+ "learning_rate": 0.0006,
+ "loss": 4.636074066162109,
+ "step": 999
+ },
+ {
+ "epoch": 13.894323144104803,
+ "grad_norm": 0.016660034656524658,
+ "learning_rate": 0.0006,
+ "loss": 4.742693901062012,
+ "step": 1000
+ },
+ {
+ "epoch": 13.908296943231441,
+ "grad_norm": 0.014902938157320023,
+ "learning_rate": 0.0006,
+ "loss": 4.769811630249023,
+ "step": 1001
+ },
+ {
+ "epoch": 13.922270742358078,
+ "grad_norm": 0.014883476309478283,
+ "learning_rate": 0.0006,
+ "loss": 4.731159210205078,
+ "step": 1002
+ },
+ {
+ "epoch": 13.936244541484717,
+ "grad_norm": 0.014494217932224274,
+ "learning_rate": 0.0006,
+ "loss": 4.659485816955566,
+ "step": 1003
+ },
+ {
+ "epoch": 13.950218340611354,
+ "grad_norm": 0.015426991507411003,
+ "learning_rate": 0.0006,
+ "loss": 4.755229949951172,
+ "step": 1004
+ },
+ {
+ "epoch": 13.96419213973799,
+ "grad_norm": 0.016648586839437485,
+ "learning_rate": 0.0006,
+ "loss": 4.74111270904541,
+ "step": 1005
+ },
+ {
+ "epoch": 13.97816593886463,
+ "grad_norm": 0.016980541869997978,
+ "learning_rate": 0.0006,
+ "loss": 4.74072265625,
+ "step": 1006
+ },
+ {
+ "epoch": 13.992139737991266,
+ "grad_norm": 0.018070152029395103,
+ "learning_rate": 0.0006,
+ "loss": 4.732975006103516,
+ "step": 1007
+ },
+ {
+ "epoch": 14.0,
+ "grad_norm": 0.019910665228962898,
+ "learning_rate": 0.0006,
+ "loss": 4.634697914123535,
+ "step": 1008
+ },
+ {
+ "epoch": 14.0,
+ "eval_loss": 4.850640296936035,
+ "eval_runtime": 57.3039,
+ "eval_samples_per_second": 42.615,
+ "eval_steps_per_second": 1.344,
+ "step": 1008
+ },
+ {
+ "epoch": 14.013973799126637,
+ "grad_norm": 0.01980053074657917,
+ "learning_rate": 0.0006,
+ "loss": 4.628744602203369,
+ "step": 1009
+ },
+ {
+ "epoch": 14.027947598253276,
+ "grad_norm": 0.02497623674571514,
+ "learning_rate": 0.0006,
+ "loss": 4.75930118560791,
+ "step": 1010
+ },
+ {
+ "epoch": 14.041921397379912,
+ "grad_norm": 0.02794097363948822,
+ "learning_rate": 0.0006,
+ "loss": 4.691340446472168,
+ "step": 1011
+ },
+ {
+ "epoch": 14.055895196506551,
+ "grad_norm": 0.023725535720586777,
+ "learning_rate": 0.0006,
+ "loss": 4.733664512634277,
+ "step": 1012
+ },
+ {
+ "epoch": 14.069868995633188,
+ "grad_norm": 0.022966474294662476,
+ "learning_rate": 0.0006,
+ "loss": 4.664808750152588,
+ "step": 1013
+ },
+ {
+ "epoch": 14.083842794759825,
+ "grad_norm": 0.02530563250184059,
+ "learning_rate": 0.0006,
+ "loss": 4.8047285079956055,
+ "step": 1014
+ },
+ {
+ "epoch": 14.097816593886463,
+ "grad_norm": 0.02491082064807415,
+ "learning_rate": 0.0006,
+ "loss": 4.667956352233887,
+ "step": 1015
+ },
+ {
+ "epoch": 14.1117903930131,
+ "grad_norm": 0.022110294550657272,
+ "learning_rate": 0.0006,
+ "loss": 4.666140556335449,
+ "step": 1016
+ },
+ {
+ "epoch": 14.125764192139737,
+ "grad_norm": 0.02065849117934704,
+ "learning_rate": 0.0006,
+ "loss": 4.581960678100586,
+ "step": 1017
+ },
+ {
+ "epoch": 14.139737991266376,
+ "grad_norm": 0.01941298507153988,
+ "learning_rate": 0.0006,
+ "loss": 4.614226341247559,
+ "step": 1018
+ },
+ {
+ "epoch": 14.153711790393013,
+ "grad_norm": 0.022441856563091278,
+ "learning_rate": 0.0006,
+ "loss": 4.638352394104004,
+ "step": 1019
+ },
+ {
+ "epoch": 14.167685589519651,
+ "grad_norm": 0.021519694477319717,
+ "learning_rate": 0.0006,
+ "loss": 4.700160980224609,
+ "step": 1020
+ },
+ {
+ "epoch": 14.181659388646288,
+ "grad_norm": 0.018956486135721207,
+ "learning_rate": 0.0006,
+ "loss": 4.649593353271484,
+ "step": 1021
+ },
+ {
+ "epoch": 14.195633187772925,
+ "grad_norm": 0.016681140288710594,
+ "learning_rate": 0.0006,
+ "loss": 4.603952407836914,
+ "step": 1022
+ },
+ {
+ "epoch": 14.209606986899564,
+ "grad_norm": 0.01788809522986412,
+ "learning_rate": 0.0006,
+ "loss": 4.669160842895508,
+ "step": 1023
+ },
+ {
+ "epoch": 14.2235807860262,
+ "grad_norm": 0.016047228127717972,
+ "learning_rate": 0.0006,
+ "loss": 4.6367950439453125,
+ "step": 1024
+ },
+ {
+ "epoch": 14.237554585152838,
+ "grad_norm": 0.01614239625632763,
+ "learning_rate": 0.0006,
+ "loss": 4.664529323577881,
+ "step": 1025
+ },
+ {
+ "epoch": 14.251528384279476,
+ "grad_norm": 0.016362549737095833,
+ "learning_rate": 0.0006,
+ "loss": 4.702373504638672,
+ "step": 1026
+ },
+ {
+ "epoch": 14.265502183406113,
+ "grad_norm": 0.015800409018993378,
+ "learning_rate": 0.0006,
+ "loss": 4.7594428062438965,
+ "step": 1027
+ },
+ {
+ "epoch": 14.279475982532752,
+ "grad_norm": 0.016730139032006264,
+ "learning_rate": 0.0006,
+ "loss": 4.677244186401367,
+ "step": 1028
+ },
+ {
+ "epoch": 14.293449781659389,
+ "grad_norm": 0.01661759801208973,
+ "learning_rate": 0.0006,
+ "loss": 4.633679389953613,
+ "step": 1029
+ },
+ {
+ "epoch": 14.307423580786025,
+ "grad_norm": 0.017673347145318985,
+ "learning_rate": 0.0006,
+ "loss": 4.649107456207275,
+ "step": 1030
+ },
+ {
+ "epoch": 14.321397379912664,
+ "grad_norm": 0.019243352115154266,
+ "learning_rate": 0.0006,
+ "loss": 4.611257553100586,
+ "step": 1031
+ },
+ {
+ "epoch": 14.335371179039301,
+ "grad_norm": 0.019817164167761803,
+ "learning_rate": 0.0006,
+ "loss": 4.588086128234863,
+ "step": 1032
+ },
+ {
+ "epoch": 14.34934497816594,
+ "grad_norm": 0.022292211651802063,
+ "learning_rate": 0.0006,
+ "loss": 4.783066749572754,
+ "step": 1033
+ },
+ {
+ "epoch": 14.363318777292577,
+ "grad_norm": 0.02140096202492714,
+ "learning_rate": 0.0006,
+ "loss": 4.604718208312988,
+ "step": 1034
+ },
+ {
+ "epoch": 14.377292576419213,
+ "grad_norm": 0.0187817495316267,
+ "learning_rate": 0.0006,
+ "loss": 4.721773147583008,
+ "step": 1035
+ },
+ {
+ "epoch": 14.391266375545852,
+ "grad_norm": 0.019879210740327835,
+ "learning_rate": 0.0006,
+ "loss": 4.789286136627197,
+ "step": 1036
+ },
+ {
+ "epoch": 14.405240174672489,
+ "grad_norm": 0.01885785162448883,
+ "learning_rate": 0.0006,
+ "loss": 4.669743537902832,
+ "step": 1037
+ },
+ {
+ "epoch": 14.419213973799126,
+ "grad_norm": 0.018324270844459534,
+ "learning_rate": 0.0006,
+ "loss": 4.65901517868042,
+ "step": 1038
+ },
+ {
+ "epoch": 14.433187772925764,
+ "grad_norm": 0.017664821818470955,
+ "learning_rate": 0.0006,
+ "loss": 4.667839527130127,
+ "step": 1039
+ },
+ {
+ "epoch": 14.447161572052401,
+ "grad_norm": 0.015702905133366585,
+ "learning_rate": 0.0006,
+ "loss": 4.64700984954834,
+ "step": 1040
+ },
+ {
+ "epoch": 14.46113537117904,
+ "grad_norm": 0.016674194484949112,
+ "learning_rate": 0.0006,
+ "loss": 4.694093227386475,
+ "step": 1041
+ },
+ {
+ "epoch": 14.475109170305677,
+ "grad_norm": 0.016869477927684784,
+ "learning_rate": 0.0006,
+ "loss": 4.619544982910156,
+ "step": 1042
+ },
+ {
+ "epoch": 14.489082969432314,
+ "grad_norm": 0.016557540744543076,
+ "learning_rate": 0.0006,
+ "loss": 4.568667411804199,
+ "step": 1043
+ },
+ {
+ "epoch": 14.503056768558952,
+ "grad_norm": 0.014500945806503296,
+ "learning_rate": 0.0006,
+ "loss": 4.561741828918457,
+ "step": 1044
+ },
+ {
+ "epoch": 14.51703056768559,
+ "grad_norm": 0.014228739775717258,
+ "learning_rate": 0.0006,
+ "loss": 4.654277324676514,
+ "step": 1045
+ },
+ {
+ "epoch": 14.531004366812226,
+ "grad_norm": 0.013894238509237766,
+ "learning_rate": 0.0006,
+ "loss": 4.5867719650268555,
+ "step": 1046
+ },
+ {
+ "epoch": 14.544978165938865,
+ "grad_norm": 0.01550068985670805,
+ "learning_rate": 0.0006,
+ "loss": 4.61168098449707,
+ "step": 1047
+ },
+ {
+ "epoch": 14.558951965065502,
+ "grad_norm": 0.015940772369503975,
+ "learning_rate": 0.0006,
+ "loss": 4.691247940063477,
+ "step": 1048
+ },
+ {
+ "epoch": 14.57292576419214,
+ "grad_norm": 0.01714535616338253,
+ "learning_rate": 0.0006,
+ "loss": 4.61183500289917,
+ "step": 1049
+ },
+ {
+ "epoch": 14.586899563318777,
+ "grad_norm": 0.017925038933753967,
+ "learning_rate": 0.0006,
+ "loss": 4.719022750854492,
+ "step": 1050
+ },
+ {
+ "epoch": 14.600873362445414,
+ "grad_norm": 0.018141552805900574,
+ "learning_rate": 0.0006,
+ "loss": 4.760281562805176,
+ "step": 1051
+ },
+ {
+ "epoch": 14.614847161572053,
+ "grad_norm": 0.017131328582763672,
+ "learning_rate": 0.0006,
+ "loss": 4.709680080413818,
+ "step": 1052
+ },
+ {
+ "epoch": 14.62882096069869,
+ "grad_norm": 0.018730048090219498,
+ "learning_rate": 0.0006,
+ "loss": 4.5994720458984375,
+ "step": 1053
+ },
+ {
+ "epoch": 14.642794759825328,
+ "grad_norm": 0.018894298002123833,
+ "learning_rate": 0.0006,
+ "loss": 4.669070720672607,
+ "step": 1054
+ },
+ {
+ "epoch": 14.656768558951965,
+ "grad_norm": 0.017643198370933533,
+ "learning_rate": 0.0006,
+ "loss": 4.641790390014648,
+ "step": 1055
+ },
+ {
+ "epoch": 14.670742358078602,
+ "grad_norm": 0.017647244036197662,
+ "learning_rate": 0.0006,
+ "loss": 4.653899192810059,
+ "step": 1056
+ },
+ {
+ "epoch": 14.68471615720524,
+ "grad_norm": 0.017304198816418648,
+ "learning_rate": 0.0006,
+ "loss": 4.7487473487854,
+ "step": 1057
+ },
+ {
+ "epoch": 14.698689956331878,
+ "grad_norm": 0.01638447679579258,
+ "learning_rate": 0.0006,
+ "loss": 4.71945858001709,
+ "step": 1058
+ },
+ {
+ "epoch": 14.712663755458514,
+ "grad_norm": 0.015892071649432182,
+ "learning_rate": 0.0006,
+ "loss": 4.744903564453125,
+ "step": 1059
+ },
+ {
+ "epoch": 14.726637554585153,
+ "grad_norm": 0.017420614138245583,
+ "learning_rate": 0.0006,
+ "loss": 4.781563758850098,
+ "step": 1060
+ },
+ {
+ "epoch": 14.74061135371179,
+ "grad_norm": 0.016498520970344543,
+ "learning_rate": 0.0006,
+ "loss": 4.622779846191406,
+ "step": 1061
+ },
+ {
+ "epoch": 14.754585152838429,
+ "grad_norm": 0.017482127994298935,
+ "learning_rate": 0.0006,
+ "loss": 4.62825870513916,
+ "step": 1062
+ },
+ {
+ "epoch": 14.768558951965066,
+ "grad_norm": 0.019018379971385002,
+ "learning_rate": 0.0006,
+ "loss": 4.693194389343262,
+ "step": 1063
+ },
+ {
+ "epoch": 14.782532751091702,
+ "grad_norm": 0.01993914321064949,
+ "learning_rate": 0.0006,
+ "loss": 4.742903709411621,
+ "step": 1064
+ },
+ {
+ "epoch": 14.796506550218341,
+ "grad_norm": 0.02301335707306862,
+ "learning_rate": 0.0006,
+ "loss": 4.613251686096191,
+ "step": 1065
+ },
+ {
+ "epoch": 14.810480349344978,
+ "grad_norm": 0.024088222533464432,
+ "learning_rate": 0.0006,
+ "loss": 4.594305992126465,
+ "step": 1066
+ },
+ {
+ "epoch": 14.824454148471617,
+ "grad_norm": 0.02493320405483246,
+ "learning_rate": 0.0006,
+ "loss": 4.700588226318359,
+ "step": 1067
+ },
+ {
+ "epoch": 14.838427947598253,
+ "grad_norm": 0.02501937560737133,
+ "learning_rate": 0.0006,
+ "loss": 4.601742744445801,
+ "step": 1068
+ },
+ {
+ "epoch": 14.85240174672489,
+ "grad_norm": 0.023490751162171364,
+ "learning_rate": 0.0006,
+ "loss": 4.589211463928223,
+ "step": 1069
+ },
+ {
+ "epoch": 14.866375545851529,
+ "grad_norm": 0.019695665687322617,
+ "learning_rate": 0.0006,
+ "loss": 4.635081768035889,
+ "step": 1070
+ },
+ {
+ "epoch": 14.880349344978166,
+ "grad_norm": 0.019276980310678482,
+ "learning_rate": 0.0006,
+ "loss": 4.737002372741699,
+ "step": 1071
+ },
+ {
+ "epoch": 14.894323144104803,
+ "grad_norm": 0.022430511191487312,
+ "learning_rate": 0.0006,
+ "loss": 4.808295726776123,
+ "step": 1072
+ },
+ {
+ "epoch": 14.908296943231441,
+ "grad_norm": 0.021543916314840317,
+ "learning_rate": 0.0006,
+ "loss": 4.559475421905518,
+ "step": 1073
+ },
+ {
+ "epoch": 14.922270742358078,
+ "grad_norm": 0.019885241985321045,
+ "learning_rate": 0.0006,
+ "loss": 4.647948741912842,
+ "step": 1074
+ },
+ {
+ "epoch": 14.936244541484717,
+ "grad_norm": 0.02115175500512123,
+ "learning_rate": 0.0006,
+ "loss": 4.680505752563477,
+ "step": 1075
+ },
+ {
+ "epoch": 14.950218340611354,
+ "grad_norm": 0.02156151458621025,
+ "learning_rate": 0.0006,
+ "loss": 4.673477649688721,
+ "step": 1076
+ },
+ {
+ "epoch": 14.96419213973799,
+ "grad_norm": 0.02184601128101349,
+ "learning_rate": 0.0006,
+ "loss": 4.644402027130127,
+ "step": 1077
+ },
+ {
+ "epoch": 14.97816593886463,
+ "grad_norm": 0.0211983360350132,
+ "learning_rate": 0.0006,
+ "loss": 4.821558475494385,
+ "step": 1078
+ },
+ {
+ "epoch": 14.992139737991266,
+ "grad_norm": 0.019636567682027817,
+ "learning_rate": 0.0006,
+ "loss": 4.578868865966797,
+ "step": 1079
+ },
+ {
+ "epoch": 15.0,
+ "grad_norm": 0.019997350871562958,
+ "learning_rate": 0.0006,
+ "loss": 4.691307544708252,
+ "step": 1080
+ },
+ {
+ "epoch": 15.0,
+ "eval_loss": 4.872478008270264,
+ "eval_runtime": 56.8513,
+ "eval_samples_per_second": 42.954,
+ "eval_steps_per_second": 1.354,
+ "step": 1080
+ },
+ {
+ "epoch": 15.013973799126637,
+ "grad_norm": 0.018830662593245506,
+ "learning_rate": 0.0006,
+ "loss": 4.761779308319092,
+ "step": 1081
+ },
+ {
+ "epoch": 15.027947598253276,
+ "grad_norm": 0.016876481473445892,
+ "learning_rate": 0.0006,
+ "loss": 4.61148738861084,
+ "step": 1082
+ },
+ {
+ "epoch": 15.041921397379912,
+ "grad_norm": 0.015550863929092884,
+ "learning_rate": 0.0006,
+ "loss": 4.676124095916748,
+ "step": 1083
+ },
+ {
+ "epoch": 15.055895196506551,
+ "grad_norm": 0.017178896814584732,
+ "learning_rate": 0.0006,
+ "loss": 4.588935852050781,
+ "step": 1084
+ },
+ {
+ "epoch": 15.069868995633188,
+ "grad_norm": 0.015465447679162025,
+ "learning_rate": 0.0006,
+ "loss": 4.605499267578125,
+ "step": 1085
+ },
+ {
+ "epoch": 15.083842794759825,
+ "grad_norm": 0.01606195978820324,
+ "learning_rate": 0.0006,
+ "loss": 4.723373889923096,
+ "step": 1086
+ },
+ {
+ "epoch": 15.097816593886463,
+ "grad_norm": 0.016521936282515526,
+ "learning_rate": 0.0006,
+ "loss": 4.646608352661133,
+ "step": 1087
+ },
+ {
+ "epoch": 15.1117903930131,
+ "grad_norm": 0.01589205674827099,
+ "learning_rate": 0.0006,
+ "loss": 4.498138427734375,
+ "step": 1088
+ },
+ {
+ "epoch": 15.125764192139737,
+ "grad_norm": 0.01524856686592102,
+ "learning_rate": 0.0006,
+ "loss": 4.594562530517578,
+ "step": 1089
+ },
+ {
+ "epoch": 15.139737991266376,
+ "grad_norm": 0.014714641496539116,
+ "learning_rate": 0.0006,
+ "loss": 4.630744934082031,
+ "step": 1090
+ },
+ {
+ "epoch": 15.153711790393013,
+ "grad_norm": 0.015333331190049648,
+ "learning_rate": 0.0006,
+ "loss": 4.682388782501221,
+ "step": 1091
+ },
+ {
+ "epoch": 15.167685589519651,
+ "grad_norm": 0.01550073828548193,
+ "learning_rate": 0.0006,
+ "loss": 4.692652702331543,
+ "step": 1092
+ },
+ {
+ "epoch": 15.181659388646288,
+ "grad_norm": 0.014783318154513836,
+ "learning_rate": 0.0006,
+ "loss": 4.6223835945129395,
+ "step": 1093
+ },
+ {
+ "epoch": 15.195633187772925,
+ "grad_norm": 0.014372751116752625,
+ "learning_rate": 0.0006,
+ "loss": 4.52687406539917,
+ "step": 1094
+ },
+ {
+ "epoch": 15.209606986899564,
+ "grad_norm": 0.014779130928218365,
+ "learning_rate": 0.0006,
+ "loss": 4.637009620666504,
+ "step": 1095
+ },
+ {
+ "epoch": 15.2235807860262,
+ "grad_norm": 0.015697073191404343,
+ "learning_rate": 0.0006,
+ "loss": 4.682981014251709,
+ "step": 1096
+ },
+ {
+ "epoch": 15.237554585152838,
+ "grad_norm": 0.016523664817214012,
+ "learning_rate": 0.0006,
+ "loss": 4.705722808837891,
+ "step": 1097
+ },
+ {
+ "epoch": 15.251528384279476,
+ "grad_norm": 0.017633996903896332,
+ "learning_rate": 0.0006,
+ "loss": 4.574686050415039,
+ "step": 1098
+ },
+ {
+ "epoch": 15.265502183406113,
+ "grad_norm": 0.019955076277256012,
+ "learning_rate": 0.0006,
+ "loss": 4.634450912475586,
+ "step": 1099
+ },
+ {
+ "epoch": 15.279475982532752,
+ "grad_norm": 0.02132618986070156,
+ "learning_rate": 0.0006,
+ "loss": 4.56110143661499,
+ "step": 1100
+ },
+ {
+ "epoch": 15.293449781659389,
+ "grad_norm": 0.0207593385130167,
+ "learning_rate": 0.0006,
+ "loss": 4.629396438598633,
+ "step": 1101
+ },
+ {
+ "epoch": 15.307423580786025,
+ "grad_norm": 0.017882540822029114,
+ "learning_rate": 0.0006,
+ "loss": 4.730963230133057,
+ "step": 1102
+ },
+ {
+ "epoch": 15.321397379912664,
+ "grad_norm": 0.01917876861989498,
+ "learning_rate": 0.0006,
+ "loss": 4.64585542678833,
+ "step": 1103
+ },
+ {
+ "epoch": 15.335371179039301,
+ "grad_norm": 0.01943974196910858,
+ "learning_rate": 0.0006,
+ "loss": 4.656181335449219,
+ "step": 1104
+ },
+ {
+ "epoch": 15.34934497816594,
+ "grad_norm": 0.019904715940356255,
+ "learning_rate": 0.0006,
+ "loss": 4.619329452514648,
+ "step": 1105
+ },
+ {
+ "epoch": 15.363318777292577,
+ "grad_norm": 0.019174346700310707,
+ "learning_rate": 0.0006,
+ "loss": 4.647892475128174,
+ "step": 1106
+ },
+ {
+ "epoch": 15.377292576419213,
+ "grad_norm": 0.018244758248329163,
+ "learning_rate": 0.0006,
+ "loss": 4.56166934967041,
+ "step": 1107
+ },
+ {
+ "epoch": 15.391266375545852,
+ "grad_norm": 0.018914205953478813,
+ "learning_rate": 0.0006,
+ "loss": 4.648063659667969,
+ "step": 1108
+ },
+ {
+ "epoch": 15.405240174672489,
+ "grad_norm": 0.01961071416735649,
+ "learning_rate": 0.0006,
+ "loss": 4.665525436401367,
+ "step": 1109
+ },
+ {
+ "epoch": 15.419213973799126,
+ "grad_norm": 0.01841311901807785,
+ "learning_rate": 0.0006,
+ "loss": 4.611342430114746,
+ "step": 1110
+ },
+ {
+ "epoch": 15.433187772925764,
+ "grad_norm": 0.01863427273929119,
+ "learning_rate": 0.0006,
+ "loss": 4.664256572723389,
+ "step": 1111
+ },
+ {
+ "epoch": 15.447161572052401,
+ "grad_norm": 0.020506877452135086,
+ "learning_rate": 0.0006,
+ "loss": 4.631237506866455,
+ "step": 1112
+ },
+ {
+ "epoch": 15.46113537117904,
+ "grad_norm": 0.02223833277821541,
+ "learning_rate": 0.0006,
+ "loss": 4.668654441833496,
+ "step": 1113
+ },
+ {
+ "epoch": 15.475109170305677,
+ "grad_norm": 0.023336702957749367,
+ "learning_rate": 0.0006,
+ "loss": 4.562833309173584,
+ "step": 1114
+ },
+ {
+ "epoch": 15.489082969432314,
+ "grad_norm": 0.023052413016557693,
+ "learning_rate": 0.0006,
+ "loss": 4.621532917022705,
+ "step": 1115
+ },
+ {
+ "epoch": 15.503056768558952,
+ "grad_norm": 0.023420924320816994,
+ "learning_rate": 0.0006,
+ "loss": 4.665465354919434,
+ "step": 1116
+ },
+ {
+ "epoch": 15.51703056768559,
+ "grad_norm": 0.024071281775832176,
+ "learning_rate": 0.0006,
+ "loss": 4.595980644226074,
+ "step": 1117
+ },
+ {
+ "epoch": 15.531004366812226,
+ "grad_norm": 0.023406682536005974,
+ "learning_rate": 0.0006,
+ "loss": 4.642635345458984,
+ "step": 1118
+ },
+ {
+ "epoch": 15.544978165938865,
+ "grad_norm": 0.02138647809624672,
+ "learning_rate": 0.0006,
+ "loss": 4.704513072967529,
+ "step": 1119
+ },
+ {
+ "epoch": 15.558951965065502,
+ "grad_norm": 0.0221384447067976,
+ "learning_rate": 0.0006,
+ "loss": 4.567503452301025,
+ "step": 1120
+ },
+ {
+ "epoch": 15.57292576419214,
+ "grad_norm": 0.020879851654171944,
+ "learning_rate": 0.0006,
+ "loss": 4.664057731628418,
+ "step": 1121
+ },
+ {
+ "epoch": 15.586899563318777,
+ "grad_norm": 0.021478967741131783,
+ "learning_rate": 0.0006,
+ "loss": 4.539627552032471,
+ "step": 1122
+ },
+ {
+ "epoch": 15.600873362445414,
+ "grad_norm": 0.022748656570911407,
+ "learning_rate": 0.0006,
+ "loss": 4.5654497146606445,
+ "step": 1123
+ },
+ {
+ "epoch": 15.614847161572053,
+ "grad_norm": 0.021870248019695282,
+ "learning_rate": 0.0006,
+ "loss": 4.56899356842041,
+ "step": 1124
+ },
+ {
+ "epoch": 15.62882096069869,
+ "grad_norm": 0.021711355075240135,
+ "learning_rate": 0.0006,
+ "loss": 4.657808303833008,
+ "step": 1125
+ },
+ {
+ "epoch": 15.642794759825328,
+ "grad_norm": 0.022810909897089005,
+ "learning_rate": 0.0006,
+ "loss": 4.6526689529418945,
+ "step": 1126
+ },
+ {
+ "epoch": 15.656768558951965,
+ "grad_norm": 0.022561343386769295,
+ "learning_rate": 0.0006,
+ "loss": 4.534887790679932,
+ "step": 1127
+ },
+ {
+ "epoch": 15.670742358078602,
+ "grad_norm": 0.019909802824258804,
+ "learning_rate": 0.0006,
+ "loss": 4.600790023803711,
+ "step": 1128
+ },
+ {
+ "epoch": 15.68471615720524,
+ "grad_norm": 0.020092325285077095,
+ "learning_rate": 0.0006,
+ "loss": 4.773830413818359,
+ "step": 1129
+ },
+ {
+ "epoch": 15.698689956331878,
+ "grad_norm": 0.018834445625543594,
+ "learning_rate": 0.0006,
+ "loss": 4.628868103027344,
+ "step": 1130
+ },
+ {
+ "epoch": 15.712663755458514,
+ "grad_norm": 0.017552798613905907,
+ "learning_rate": 0.0006,
+ "loss": 4.647500038146973,
+ "step": 1131
+ },
+ {
+ "epoch": 15.726637554585153,
+ "grad_norm": 0.01751251146197319,
+ "learning_rate": 0.0006,
+ "loss": 4.590545654296875,
+ "step": 1132
+ },
+ {
+ "epoch": 15.74061135371179,
+ "grad_norm": 0.01699099875986576,
+ "learning_rate": 0.0006,
+ "loss": 4.509699821472168,
+ "step": 1133
+ },
+ {
+ "epoch": 15.754585152838429,
+ "grad_norm": 0.015808649361133575,
+ "learning_rate": 0.0006,
+ "loss": 4.606992244720459,
+ "step": 1134
+ },
+ {
+ "epoch": 15.768558951965066,
+ "grad_norm": 0.01655443385243416,
+ "learning_rate": 0.0006,
+ "loss": 4.585219860076904,
+ "step": 1135
+ },
+ {
+ "epoch": 15.782532751091702,
+ "grad_norm": 0.01708163693547249,
+ "learning_rate": 0.0006,
+ "loss": 4.668147087097168,
+ "step": 1136
+ },
+ {
+ "epoch": 15.796506550218341,
+ "grad_norm": 0.017161913216114044,
+ "learning_rate": 0.0006,
+ "loss": 4.7170634269714355,
+ "step": 1137
+ },
+ {
+ "epoch": 15.810480349344978,
+ "grad_norm": 0.017262674868106842,
+ "learning_rate": 0.0006,
+ "loss": 4.572132110595703,
+ "step": 1138
+ },
+ {
+ "epoch": 15.824454148471617,
+ "grad_norm": 0.017713576555252075,
+ "learning_rate": 0.0006,
+ "loss": 4.702538967132568,
+ "step": 1139
+ },
+ {
+ "epoch": 15.838427947598253,
+ "grad_norm": 0.018730733543634415,
+ "learning_rate": 0.0006,
+ "loss": 4.629721641540527,
+ "step": 1140
+ },
+ {
+ "epoch": 15.85240174672489,
+ "grad_norm": 0.017960187047719955,
+ "learning_rate": 0.0006,
+ "loss": 4.661039352416992,
+ "step": 1141
+ },
+ {
+ "epoch": 15.866375545851529,
+ "grad_norm": 0.01750079356133938,
+ "learning_rate": 0.0006,
+ "loss": 4.503710746765137,
+ "step": 1142
+ },
+ {
+ "epoch": 15.880349344978166,
+ "grad_norm": 0.016630422323942184,
+ "learning_rate": 0.0006,
+ "loss": 4.6582136154174805,
+ "step": 1143
+ },
+ {
+ "epoch": 15.894323144104803,
+ "grad_norm": 0.015896832570433617,
+ "learning_rate": 0.0006,
+ "loss": 4.596693992614746,
+ "step": 1144
+ },
+ {
+ "epoch": 15.908296943231441,
+ "grad_norm": 0.01566407084465027,
+ "learning_rate": 0.0006,
+ "loss": 4.6337199211120605,
+ "step": 1145
+ },
+ {
+ "epoch": 15.922270742358078,
+ "grad_norm": 0.016823330894112587,
+ "learning_rate": 0.0006,
+ "loss": 4.686994552612305,
+ "step": 1146
+ },
+ {
+ "epoch": 15.936244541484717,
+ "grad_norm": 0.016208255663514137,
+ "learning_rate": 0.0006,
+ "loss": 4.6205549240112305,
+ "step": 1147
+ },
+ {
+ "epoch": 15.950218340611354,
+ "grad_norm": 0.0171707421541214,
+ "learning_rate": 0.0006,
+ "loss": 4.582968711853027,
+ "step": 1148
+ },
+ {
+ "epoch": 15.96419213973799,
+ "grad_norm": 0.017009446397423744,
+ "learning_rate": 0.0006,
+ "loss": 4.510274887084961,
+ "step": 1149
+ },
+ {
+ "epoch": 15.97816593886463,
+ "grad_norm": 0.01565811224281788,
+ "learning_rate": 0.0006,
+ "loss": 4.55834436416626,
+ "step": 1150
+ },
+ {
+ "epoch": 15.992139737991266,
+ "grad_norm": 0.015276037156581879,
+ "learning_rate": 0.0006,
+ "loss": 4.537870407104492,
+ "step": 1151
+ },
+ {
+ "epoch": 16.0,
+ "grad_norm": 0.017305882647633553,
+ "learning_rate": 0.0006,
+ "loss": 4.611305236816406,
+ "step": 1152
+ },
+ {
+ "epoch": 16.0,
+ "eval_loss": 4.775513172149658,
+ "eval_runtime": 56.7341,
+ "eval_samples_per_second": 43.043,
+ "eval_steps_per_second": 1.357,
+ "step": 1152
+ },
+ {
+ "epoch": 16.01397379912664,
+ "grad_norm": 0.016752436757087708,
+ "learning_rate": 0.0006,
+ "loss": 4.541312217712402,
+ "step": 1153
+ },
+ {
+ "epoch": 16.027947598253274,
+ "grad_norm": 0.01643134094774723,
+ "learning_rate": 0.0006,
+ "loss": 4.585402965545654,
+ "step": 1154
+ },
+ {
+ "epoch": 16.041921397379912,
+ "grad_norm": 0.01888720504939556,
+ "learning_rate": 0.0006,
+ "loss": 4.597825050354004,
+ "step": 1155
+ },
+ {
+ "epoch": 16.05589519650655,
+ "grad_norm": 0.022044632583856583,
+ "learning_rate": 0.0006,
+ "loss": 4.532772064208984,
+ "step": 1156
+ },
+ {
+ "epoch": 16.069868995633186,
+ "grad_norm": 0.022986987605690956,
+ "learning_rate": 0.0006,
+ "loss": 4.551351547241211,
+ "step": 1157
+ },
+ {
+ "epoch": 16.083842794759825,
+ "grad_norm": 0.020116791129112244,
+ "learning_rate": 0.0006,
+ "loss": 4.484149932861328,
+ "step": 1158
+ },
+ {
+ "epoch": 16.097816593886463,
+ "grad_norm": 0.021442247554659843,
+ "learning_rate": 0.0006,
+ "loss": 4.515316486358643,
+ "step": 1159
+ },
+ {
+ "epoch": 16.111790393013102,
+ "grad_norm": 0.020974772050976753,
+ "learning_rate": 0.0006,
+ "loss": 4.51606559753418,
+ "step": 1160
+ },
+ {
+ "epoch": 16.125764192139737,
+ "grad_norm": 0.021274864673614502,
+ "learning_rate": 0.0006,
+ "loss": 4.587329864501953,
+ "step": 1161
+ },
+ {
+ "epoch": 16.139737991266376,
+ "grad_norm": 0.022799784317612648,
+ "learning_rate": 0.0006,
+ "loss": 4.568149089813232,
+ "step": 1162
+ },
+ {
+ "epoch": 16.153711790393015,
+ "grad_norm": 0.021784933283925056,
+ "learning_rate": 0.0006,
+ "loss": 4.515336990356445,
+ "step": 1163
+ },
+ {
+ "epoch": 16.16768558951965,
+ "grad_norm": 0.024261271581053734,
+ "learning_rate": 0.0006,
+ "loss": 4.5127081871032715,
+ "step": 1164
+ },
+ {
+ "epoch": 16.18165938864629,
+ "grad_norm": 0.02248375490307808,
+ "learning_rate": 0.0006,
+ "loss": 4.754251003265381,
+ "step": 1165
+ },
+ {
+ "epoch": 16.195633187772927,
+ "grad_norm": 0.02136492356657982,
+ "learning_rate": 0.0006,
+ "loss": 4.58193826675415,
+ "step": 1166
+ },
+ {
+ "epoch": 16.209606986899562,
+ "grad_norm": 0.020727066323161125,
+ "learning_rate": 0.0006,
+ "loss": 4.564209938049316,
+ "step": 1167
+ },
+ {
+ "epoch": 16.2235807860262,
+ "grad_norm": 0.017789119854569435,
+ "learning_rate": 0.0006,
+ "loss": 4.698680877685547,
+ "step": 1168
+ },
+ {
+ "epoch": 16.23755458515284,
+ "grad_norm": 0.017603034153580666,
+ "learning_rate": 0.0006,
+ "loss": 4.501171588897705,
+ "step": 1169
+ },
+ {
+ "epoch": 16.251528384279474,
+ "grad_norm": 0.017345862463116646,
+ "learning_rate": 0.0006,
+ "loss": 4.593590259552002,
+ "step": 1170
+ },
+ {
+ "epoch": 16.265502183406113,
+ "grad_norm": 0.018047425895929337,
+ "learning_rate": 0.0006,
+ "loss": 4.597467422485352,
+ "step": 1171
+ },
+ {
+ "epoch": 16.27947598253275,
+ "grad_norm": 0.01947222836315632,
+ "learning_rate": 0.0006,
+ "loss": 4.5072221755981445,
+ "step": 1172
+ },
+ {
+ "epoch": 16.29344978165939,
+ "grad_norm": 0.019598500803112984,
+ "learning_rate": 0.0006,
+ "loss": 4.601351737976074,
+ "step": 1173
+ },
+ {
+ "epoch": 16.307423580786025,
+ "grad_norm": 0.020002514123916626,
+ "learning_rate": 0.0006,
+ "loss": 4.580431938171387,
+ "step": 1174
+ },
+ {
+ "epoch": 16.321397379912664,
+ "grad_norm": 0.0216530729085207,
+ "learning_rate": 0.0006,
+ "loss": 4.424681663513184,
+ "step": 1175
+ },
+ {
+ "epoch": 16.335371179039303,
+ "grad_norm": 0.01949235238134861,
+ "learning_rate": 0.0006,
+ "loss": 4.751883029937744,
+ "step": 1176
+ },
+ {
+ "epoch": 16.349344978165938,
+ "grad_norm": 0.017779016867280006,
+ "learning_rate": 0.0006,
+ "loss": 4.541490077972412,
+ "step": 1177
+ },
+ {
+ "epoch": 16.363318777292577,
+ "grad_norm": 0.01909121684730053,
+ "learning_rate": 0.0006,
+ "loss": 4.594141006469727,
+ "step": 1178
+ },
+ {
+ "epoch": 16.377292576419215,
+ "grad_norm": 0.019894849509000778,
+ "learning_rate": 0.0006,
+ "loss": 4.555070400238037,
+ "step": 1179
+ },
+ {
+ "epoch": 16.39126637554585,
+ "grad_norm": 0.01993032731115818,
+ "learning_rate": 0.0006,
+ "loss": 4.521401405334473,
+ "step": 1180
+ },
+ {
+ "epoch": 16.40524017467249,
+ "grad_norm": 0.0210788045078516,
+ "learning_rate": 0.0006,
+ "loss": 4.672842979431152,
+ "step": 1181
+ },
+ {
+ "epoch": 16.419213973799128,
+ "grad_norm": 0.02274598926305771,
+ "learning_rate": 0.0006,
+ "loss": 4.629504203796387,
+ "step": 1182
+ },
+ {
+ "epoch": 16.433187772925763,
+ "grad_norm": 0.020448120310902596,
+ "learning_rate": 0.0006,
+ "loss": 4.548209190368652,
+ "step": 1183
+ },
+ {
+ "epoch": 16.4471615720524,
+ "grad_norm": 0.017565643414855003,
+ "learning_rate": 0.0006,
+ "loss": 4.529054641723633,
+ "step": 1184
+ },
+ {
+ "epoch": 16.46113537117904,
+ "grad_norm": 0.019504237920045853,
+ "learning_rate": 0.0006,
+ "loss": 4.48182487487793,
+ "step": 1185
+ },
+ {
+ "epoch": 16.475109170305675,
+ "grad_norm": 0.016823742538690567,
+ "learning_rate": 0.0006,
+ "loss": 4.608696937561035,
+ "step": 1186
+ },
+ {
+ "epoch": 16.489082969432314,
+ "grad_norm": 0.01602894812822342,
+ "learning_rate": 0.0006,
+ "loss": 4.620062351226807,
+ "step": 1187
+ },
+ {
+ "epoch": 16.503056768558952,
+ "grad_norm": 0.016615642234683037,
+ "learning_rate": 0.0006,
+ "loss": 4.69378662109375,
+ "step": 1188
+ },
+ {
+ "epoch": 16.51703056768559,
+ "grad_norm": 0.018576730042696,
+ "learning_rate": 0.0006,
+ "loss": 4.537041664123535,
+ "step": 1189
+ },
+ {
+ "epoch": 16.531004366812226,
+ "grad_norm": 0.018611151725053787,
+ "learning_rate": 0.0006,
+ "loss": 4.496169090270996,
+ "step": 1190
+ },
+ {
+ "epoch": 16.544978165938865,
+ "grad_norm": 0.018622413277626038,
+ "learning_rate": 0.0006,
+ "loss": 4.557358264923096,
+ "step": 1191
+ },
+ {
+ "epoch": 16.558951965065503,
+ "grad_norm": 0.01823544315993786,
+ "learning_rate": 0.0006,
+ "loss": 4.730792999267578,
+ "step": 1192
+ },
+ {
+ "epoch": 16.57292576419214,
+ "grad_norm": 0.018796846270561218,
+ "learning_rate": 0.0006,
+ "loss": 4.781791687011719,
+ "step": 1193
+ },
+ {
+ "epoch": 16.586899563318777,
+ "grad_norm": 0.01728292554616928,
+ "learning_rate": 0.0006,
+ "loss": 4.515257835388184,
+ "step": 1194
+ },
+ {
+ "epoch": 16.600873362445416,
+ "grad_norm": 0.01699589006602764,
+ "learning_rate": 0.0006,
+ "loss": 4.5996904373168945,
+ "step": 1195
+ },
+ {
+ "epoch": 16.61484716157205,
+ "grad_norm": 0.015905970707535744,
+ "learning_rate": 0.0006,
+ "loss": 4.601301193237305,
+ "step": 1196
+ },
+ {
+ "epoch": 16.62882096069869,
+ "grad_norm": 0.01338571310043335,
+ "learning_rate": 0.0006,
+ "loss": 4.557708740234375,
+ "step": 1197
+ },
+ {
+ "epoch": 16.64279475982533,
+ "grad_norm": 0.015298855490982533,
+ "learning_rate": 0.0006,
+ "loss": 4.542287349700928,
+ "step": 1198
+ },
+ {
+ "epoch": 16.656768558951963,
+ "grad_norm": 0.014753367751836777,
+ "learning_rate": 0.0006,
+ "loss": 4.595787048339844,
+ "step": 1199
+ },
+ {
+ "epoch": 16.670742358078602,
+ "grad_norm": 0.015527483075857162,
+ "learning_rate": 0.0006,
+ "loss": 4.607529163360596,
+ "step": 1200
+ },
+ {
+ "epoch": 16.68471615720524,
+ "grad_norm": 0.015042162500321865,
+ "learning_rate": 0.0006,
+ "loss": 4.686626434326172,
+ "step": 1201
+ },
+ {
+ "epoch": 16.69868995633188,
+ "grad_norm": 0.0159393772482872,
+ "learning_rate": 0.0006,
+ "loss": 4.47656774520874,
+ "step": 1202
+ },
+ {
+ "epoch": 16.712663755458514,
+ "grad_norm": 0.01532787922769785,
+ "learning_rate": 0.0006,
+ "loss": 4.55678653717041,
+ "step": 1203
+ },
+ {
+ "epoch": 16.726637554585153,
+ "grad_norm": 0.016332164406776428,
+ "learning_rate": 0.0006,
+ "loss": 4.51441764831543,
+ "step": 1204
+ },
+ {
+ "epoch": 16.74061135371179,
+ "grad_norm": 0.018901726230978966,
+ "learning_rate": 0.0006,
+ "loss": 4.57227897644043,
+ "step": 1205
+ },
+ {
+ "epoch": 16.754585152838427,
+ "grad_norm": 0.018589433282613754,
+ "learning_rate": 0.0006,
+ "loss": 4.550070762634277,
+ "step": 1206
+ },
+ {
+ "epoch": 16.768558951965066,
+ "grad_norm": 0.01634034886956215,
+ "learning_rate": 0.0006,
+ "loss": 4.586910724639893,
+ "step": 1207
+ },
+ {
+ "epoch": 16.782532751091704,
+ "grad_norm": 0.017599433660507202,
+ "learning_rate": 0.0006,
+ "loss": 4.5891923904418945,
+ "step": 1208
+ },
+ {
+ "epoch": 16.79650655021834,
+ "grad_norm": 0.01878446340560913,
+ "learning_rate": 0.0006,
+ "loss": 4.57227087020874,
+ "step": 1209
+ },
+ {
+ "epoch": 16.810480349344978,
+ "grad_norm": 0.01803469844162464,
+ "learning_rate": 0.0006,
+ "loss": 4.583845138549805,
+ "step": 1210
+ },
+ {
+ "epoch": 16.824454148471617,
+ "grad_norm": 0.018505144864320755,
+ "learning_rate": 0.0006,
+ "loss": 4.561151504516602,
+ "step": 1211
+ },
+ {
+ "epoch": 16.83842794759825,
+ "grad_norm": 0.01920081488788128,
+ "learning_rate": 0.0006,
+ "loss": 4.506834030151367,
+ "step": 1212
+ },
+ {
+ "epoch": 16.85240174672489,
+ "grad_norm": 0.0177712831646204,
+ "learning_rate": 0.0006,
+ "loss": 4.557195663452148,
+ "step": 1213
+ },
+ {
+ "epoch": 16.86637554585153,
+ "grad_norm": 0.019113918766379356,
+ "learning_rate": 0.0006,
+ "loss": 4.651227951049805,
+ "step": 1214
+ },
+ {
+ "epoch": 16.880349344978168,
+ "grad_norm": 0.02012539468705654,
+ "learning_rate": 0.0006,
+ "loss": 4.4952239990234375,
+ "step": 1215
+ },
+ {
+ "epoch": 16.894323144104803,
+ "grad_norm": 0.019993646070361137,
+ "learning_rate": 0.0006,
+ "loss": 4.517433166503906,
+ "step": 1216
+ },
+ {
+ "epoch": 16.90829694323144,
+ "grad_norm": 0.020041441544890404,
+ "learning_rate": 0.0006,
+ "loss": 4.489407539367676,
+ "step": 1217
+ },
+ {
+ "epoch": 16.92227074235808,
+ "grad_norm": 0.0183611661195755,
+ "learning_rate": 0.0006,
+ "loss": 4.5435638427734375,
+ "step": 1218
+ },
+ {
+ "epoch": 16.936244541484715,
+ "grad_norm": 0.019138429313898087,
+ "learning_rate": 0.0006,
+ "loss": 4.545949459075928,
+ "step": 1219
+ },
+ {
+ "epoch": 16.950218340611354,
+ "grad_norm": 0.018554266542196274,
+ "learning_rate": 0.0006,
+ "loss": 4.579620361328125,
+ "step": 1220
+ },
+ {
+ "epoch": 16.964192139737992,
+ "grad_norm": 0.0193545650690794,
+ "learning_rate": 0.0006,
+ "loss": 4.588255405426025,
+ "step": 1221
+ },
+ {
+ "epoch": 16.978165938864628,
+ "grad_norm": 0.019389165565371513,
+ "learning_rate": 0.0006,
+ "loss": 4.520698547363281,
+ "step": 1222
+ },
+ {
+ "epoch": 16.992139737991266,
+ "grad_norm": 0.016601379960775375,
+ "learning_rate": 0.0006,
+ "loss": 4.4809064865112305,
+ "step": 1223
+ },
+ {
+ "epoch": 17.0,
+ "grad_norm": 0.01771303080022335,
+ "learning_rate": 0.0006,
+ "loss": 4.721122741699219,
+ "step": 1224
+ },
+ {
+ "epoch": 17.0,
+ "eval_loss": 4.7903666496276855,
+ "eval_runtime": 56.9826,
+ "eval_samples_per_second": 42.855,
+ "eval_steps_per_second": 1.351,
+ "step": 1224
+ },
+ {
+ "epoch": 17.01397379912664,
+ "grad_norm": 0.017720289528369904,
+ "learning_rate": 0.0006,
+ "loss": 4.523834705352783,
+ "step": 1225
+ },
+ {
+ "epoch": 17.027947598253274,
+ "grad_norm": 0.020137161016464233,
+ "learning_rate": 0.0006,
+ "loss": 4.574061870574951,
+ "step": 1226
+ },
+ {
+ "epoch": 17.041921397379912,
+ "grad_norm": 0.02002314105629921,
+ "learning_rate": 0.0006,
+ "loss": 4.512407302856445,
+ "step": 1227
+ },
+ {
+ "epoch": 17.05589519650655,
+ "grad_norm": 0.019628183916211128,
+ "learning_rate": 0.0006,
+ "loss": 4.502533435821533,
+ "step": 1228
+ },
+ {
+ "epoch": 17.069868995633186,
+ "grad_norm": 0.019808458164334297,
+ "learning_rate": 0.0006,
+ "loss": 4.463156700134277,
+ "step": 1229
+ },
+ {
+ "epoch": 17.083842794759825,
+ "grad_norm": 0.021250786259770393,
+ "learning_rate": 0.0006,
+ "loss": 4.622682094573975,
+ "step": 1230
+ },
+ {
+ "epoch": 17.097816593886463,
+ "grad_norm": 0.022031694650650024,
+ "learning_rate": 0.0006,
+ "loss": 4.451882839202881,
+ "step": 1231
+ },
+ {
+ "epoch": 17.111790393013102,
+ "grad_norm": 0.01898987777531147,
+ "learning_rate": 0.0006,
+ "loss": 4.529331684112549,
+ "step": 1232
+ },
+ {
+ "epoch": 17.125764192139737,
+ "grad_norm": 0.018749834969639778,
+ "learning_rate": 0.0006,
+ "loss": 4.435815334320068,
+ "step": 1233
+ },
+ {
+ "epoch": 17.139737991266376,
+ "grad_norm": 0.01896674931049347,
+ "learning_rate": 0.0006,
+ "loss": 4.484029293060303,
+ "step": 1234
+ },
+ {
+ "epoch": 17.153711790393015,
+ "grad_norm": 0.018286997452378273,
+ "learning_rate": 0.0006,
+ "loss": 4.523510932922363,
+ "step": 1235
+ },
+ {
+ "epoch": 17.16768558951965,
+ "grad_norm": 0.017273059114813805,
+ "learning_rate": 0.0006,
+ "loss": 4.451767444610596,
+ "step": 1236
+ },
+ {
+ "epoch": 17.18165938864629,
+ "grad_norm": 0.016499150544404984,
+ "learning_rate": 0.0006,
+ "loss": 4.5568647384643555,
+ "step": 1237
+ },
+ {
+ "epoch": 17.195633187772927,
+ "grad_norm": 0.015934469178318977,
+ "learning_rate": 0.0006,
+ "loss": 4.4639997482299805,
+ "step": 1238
+ },
+ {
+ "epoch": 17.209606986899562,
+ "grad_norm": 0.01714209094643593,
+ "learning_rate": 0.0006,
+ "loss": 4.547614097595215,
+ "step": 1239
+ },
+ {
+ "epoch": 17.2235807860262,
+ "grad_norm": 0.019167495891451836,
+ "learning_rate": 0.0006,
+ "loss": 4.571097373962402,
+ "step": 1240
+ },
+ {
+ "epoch": 17.23755458515284,
+ "grad_norm": 0.018276996910572052,
+ "learning_rate": 0.0006,
+ "loss": 4.489764213562012,
+ "step": 1241
+ },
+ {
+ "epoch": 17.251528384279474,
+ "grad_norm": 0.021296866238117218,
+ "learning_rate": 0.0006,
+ "loss": 4.526913642883301,
+ "step": 1242
+ },
+ {
+ "epoch": 17.265502183406113,
+ "grad_norm": 0.022606629878282547,
+ "learning_rate": 0.0006,
+ "loss": 4.521624565124512,
+ "step": 1243
+ },
+ {
+ "epoch": 17.27947598253275,
+ "grad_norm": 0.021204711869359016,
+ "learning_rate": 0.0006,
+ "loss": 4.500864028930664,
+ "step": 1244
+ },
+ {
+ "epoch": 17.29344978165939,
+ "grad_norm": 0.01988849975168705,
+ "learning_rate": 0.0006,
+ "loss": 4.45162296295166,
+ "step": 1245
+ },
+ {
+ "epoch": 17.307423580786025,
+ "grad_norm": 0.020268384367227554,
+ "learning_rate": 0.0006,
+ "loss": 4.536198616027832,
+ "step": 1246
+ },
+ {
+ "epoch": 17.321397379912664,
+ "grad_norm": 0.020185677334666252,
+ "learning_rate": 0.0006,
+ "loss": 4.440262794494629,
+ "step": 1247
+ },
+ {
+ "epoch": 17.335371179039303,
+ "grad_norm": 0.020106853917241096,
+ "learning_rate": 0.0006,
+ "loss": 4.475907325744629,
+ "step": 1248
+ },
+ {
+ "epoch": 17.349344978165938,
+ "grad_norm": 0.018883297219872475,
+ "learning_rate": 0.0006,
+ "loss": 4.640661716461182,
+ "step": 1249
+ },
+ {
+ "epoch": 17.363318777292577,
+ "grad_norm": 0.01852620206773281,
+ "learning_rate": 0.0006,
+ "loss": 4.578243255615234,
+ "step": 1250
+ },
+ {
+ "epoch": 17.377292576419215,
+ "grad_norm": 0.01886451430618763,
+ "learning_rate": 0.0006,
+ "loss": 4.546026229858398,
+ "step": 1251
+ },
+ {
+ "epoch": 17.39126637554585,
+ "grad_norm": 0.01777452416718006,
+ "learning_rate": 0.0006,
+ "loss": 4.521542549133301,
+ "step": 1252
+ },
+ {
+ "epoch": 17.40524017467249,
+ "grad_norm": 0.016626447439193726,
+ "learning_rate": 0.0006,
+ "loss": 4.52629280090332,
+ "step": 1253
+ },
+ {
+ "epoch": 17.419213973799128,
+ "grad_norm": 0.01599477231502533,
+ "learning_rate": 0.0006,
+ "loss": 4.4796037673950195,
+ "step": 1254
+ },
+ {
+ "epoch": 17.433187772925763,
+ "grad_norm": 0.017021458595991135,
+ "learning_rate": 0.0006,
+ "loss": 4.468813419342041,
+ "step": 1255
+ },
+ {
+ "epoch": 17.4471615720524,
+ "grad_norm": 0.016925333067774773,
+ "learning_rate": 0.0006,
+ "loss": 4.584898948669434,
+ "step": 1256
+ },
+ {
+ "epoch": 17.46113537117904,
+ "grad_norm": 0.01714998669922352,
+ "learning_rate": 0.0006,
+ "loss": 4.523544788360596,
+ "step": 1257
+ },
+ {
+ "epoch": 17.475109170305675,
+ "grad_norm": 0.01802288182079792,
+ "learning_rate": 0.0006,
+ "loss": 4.595581531524658,
+ "step": 1258
+ },
+ {
+ "epoch": 17.489082969432314,
+ "grad_norm": 0.0180289875715971,
+ "learning_rate": 0.0006,
+ "loss": 4.394440650939941,
+ "step": 1259
+ },
+ {
+ "epoch": 17.503056768558952,
+ "grad_norm": 0.018504245206713676,
+ "learning_rate": 0.0006,
+ "loss": 4.4794416427612305,
+ "step": 1260
+ },
+ {
+ "epoch": 17.51703056768559,
+ "grad_norm": 0.019973233342170715,
+ "learning_rate": 0.0006,
+ "loss": 4.4241437911987305,
+ "step": 1261
+ },
+ {
+ "epoch": 17.531004366812226,
+ "grad_norm": 0.020063083618879318,
+ "learning_rate": 0.0006,
+ "loss": 4.48068380355835,
+ "step": 1262
+ },
+ {
+ "epoch": 17.544978165938865,
+ "grad_norm": 0.02066640742123127,
+ "learning_rate": 0.0006,
+ "loss": 4.628993034362793,
+ "step": 1263
+ },
+ {
+ "epoch": 17.558951965065503,
+ "grad_norm": 0.02206416428089142,
+ "learning_rate": 0.0006,
+ "loss": 4.564790725708008,
+ "step": 1264
+ },
+ {
+ "epoch": 17.57292576419214,
+ "grad_norm": 0.020640335977077484,
+ "learning_rate": 0.0006,
+ "loss": 4.578705787658691,
+ "step": 1265
+ },
+ {
+ "epoch": 17.586899563318777,
+ "grad_norm": 0.019093405455350876,
+ "learning_rate": 0.0006,
+ "loss": 4.5023603439331055,
+ "step": 1266
+ },
+ {
+ "epoch": 17.600873362445416,
+ "grad_norm": 0.019459193572402,
+ "learning_rate": 0.0006,
+ "loss": 4.594531059265137,
+ "step": 1267
+ },
+ {
+ "epoch": 17.61484716157205,
+ "grad_norm": 0.02021171525120735,
+ "learning_rate": 0.0006,
+ "loss": 4.770299911499023,
+ "step": 1268
+ },
+ {
+ "epoch": 17.62882096069869,
+ "grad_norm": 0.021161191165447235,
+ "learning_rate": 0.0006,
+ "loss": 4.5380401611328125,
+ "step": 1269
+ },
+ {
+ "epoch": 17.64279475982533,
+ "grad_norm": 0.021013403311371803,
+ "learning_rate": 0.0006,
+ "loss": 4.455524921417236,
+ "step": 1270
+ },
+ {
+ "epoch": 17.656768558951963,
+ "grad_norm": 0.020336592569947243,
+ "learning_rate": 0.0006,
+ "loss": 4.533672332763672,
+ "step": 1271
+ },
+ {
+ "epoch": 17.670742358078602,
+ "grad_norm": 0.020159434527158737,
+ "learning_rate": 0.0006,
+ "loss": 4.596151351928711,
+ "step": 1272
+ },
+ {
+ "epoch": 17.68471615720524,
+ "grad_norm": 0.020432960242033005,
+ "learning_rate": 0.0006,
+ "loss": 4.55531644821167,
+ "step": 1273
+ },
+ {
+ "epoch": 17.69868995633188,
+ "grad_norm": 0.025466229766607285,
+ "learning_rate": 0.0006,
+ "loss": 4.566912651062012,
+ "step": 1274
+ },
+ {
+ "epoch": 17.712663755458514,
+ "grad_norm": 0.026537558063864708,
+ "learning_rate": 0.0006,
+ "loss": 4.526003837585449,
+ "step": 1275
+ },
+ {
+ "epoch": 17.726637554585153,
+ "grad_norm": 0.024397607892751694,
+ "learning_rate": 0.0006,
+ "loss": 4.541755676269531,
+ "step": 1276
+ },
+ {
+ "epoch": 17.74061135371179,
+ "grad_norm": 0.022062312811613083,
+ "learning_rate": 0.0006,
+ "loss": 4.6532793045043945,
+ "step": 1277
+ },
+ {
+ "epoch": 17.754585152838427,
+ "grad_norm": 0.02204282023012638,
+ "learning_rate": 0.0006,
+ "loss": 4.582825660705566,
+ "step": 1278
+ },
+ {
+ "epoch": 17.768558951965066,
+ "grad_norm": 0.020844416692852974,
+ "learning_rate": 0.0006,
+ "loss": 4.522336483001709,
+ "step": 1279
+ },
+ {
+ "epoch": 17.782532751091704,
+ "grad_norm": 0.019815703853964806,
+ "learning_rate": 0.0006,
+ "loss": 4.607398509979248,
+ "step": 1280
+ },
+ {
+ "epoch": 17.79650655021834,
+ "grad_norm": 0.016676954925060272,
+ "learning_rate": 0.0006,
+ "loss": 4.446206092834473,
+ "step": 1281
+ },
+ {
+ "epoch": 17.810480349344978,
+ "grad_norm": 0.0166452769190073,
+ "learning_rate": 0.0006,
+ "loss": 4.505931854248047,
+ "step": 1282
+ },
+ {
+ "epoch": 17.824454148471617,
+ "grad_norm": 0.015928907319903374,
+ "learning_rate": 0.0006,
+ "loss": 4.455267429351807,
+ "step": 1283
+ },
+ {
+ "epoch": 17.83842794759825,
+ "grad_norm": 0.015219560824334621,
+ "learning_rate": 0.0006,
+ "loss": 4.5623674392700195,
+ "step": 1284
+ },
+ {
+ "epoch": 17.85240174672489,
+ "grad_norm": 0.015125355683267117,
+ "learning_rate": 0.0006,
+ "loss": 4.554104804992676,
+ "step": 1285
+ },
+ {
+ "epoch": 17.86637554585153,
+ "grad_norm": 0.015612718649208546,
+ "learning_rate": 0.0006,
+ "loss": 4.484847068786621,
+ "step": 1286
+ },
+ {
+ "epoch": 17.880349344978168,
+ "grad_norm": 0.015215662308037281,
+ "learning_rate": 0.0006,
+ "loss": 4.4802093505859375,
+ "step": 1287
+ },
+ {
+ "epoch": 17.894323144104803,
+ "grad_norm": 0.01511684525758028,
+ "learning_rate": 0.0006,
+ "loss": 4.537298679351807,
+ "step": 1288
+ },
+ {
+ "epoch": 17.90829694323144,
+ "grad_norm": 0.013841322623193264,
+ "learning_rate": 0.0006,
+ "loss": 4.427488327026367,
+ "step": 1289
+ },
+ {
+ "epoch": 17.92227074235808,
+ "grad_norm": 0.015115964226424694,
+ "learning_rate": 0.0006,
+ "loss": 4.4547119140625,
+ "step": 1290
+ },
+ {
+ "epoch": 17.936244541484715,
+ "grad_norm": 0.016411006450653076,
+ "learning_rate": 0.0006,
+ "loss": 4.52994966506958,
+ "step": 1291
+ },
+ {
+ "epoch": 17.950218340611354,
+ "grad_norm": 0.017313152551651,
+ "learning_rate": 0.0006,
+ "loss": 4.5337653160095215,
+ "step": 1292
+ },
+ {
+ "epoch": 17.964192139737992,
+ "grad_norm": 0.016559796407818794,
+ "learning_rate": 0.0006,
+ "loss": 4.5277910232543945,
+ "step": 1293
+ },
+ {
+ "epoch": 17.978165938864628,
+ "grad_norm": 0.015854761004447937,
+ "learning_rate": 0.0006,
+ "loss": 4.6334309577941895,
+ "step": 1294
+ },
+ {
+ "epoch": 17.992139737991266,
+ "grad_norm": 0.015658238902688026,
+ "learning_rate": 0.0006,
+ "loss": 4.613372802734375,
+ "step": 1295
+ },
+ {
+ "epoch": 18.0,
+ "grad_norm": 0.0169155802577734,
+ "learning_rate": 0.0006,
+ "loss": 4.527356147766113,
+ "step": 1296
+ },
+ {
+ "epoch": 18.0,
+ "eval_loss": 4.764543056488037,
+ "eval_runtime": 56.5998,
+ "eval_samples_per_second": 43.145,
+ "eval_steps_per_second": 1.36,
+ "step": 1296
+ },
+ {
+ "epoch": 18.01397379912664,
+ "grad_norm": 0.015348963439464569,
+ "learning_rate": 0.0006,
+ "loss": 4.466423988342285,
+ "step": 1297
+ },
+ {
+ "epoch": 18.027947598253274,
+ "grad_norm": 0.01768588088452816,
+ "learning_rate": 0.0006,
+ "loss": 4.545933246612549,
+ "step": 1298
+ },
+ {
+ "epoch": 18.041921397379912,
+ "grad_norm": 0.017381226643919945,
+ "learning_rate": 0.0006,
+ "loss": 4.494060039520264,
+ "step": 1299
+ },
+ {
+ "epoch": 18.05589519650655,
+ "grad_norm": 0.019002335146069527,
+ "learning_rate": 0.0006,
+ "loss": 4.395200729370117,
+ "step": 1300
+ },
+ {
+ "epoch": 18.069868995633186,
+ "grad_norm": 0.01827932707965374,
+ "learning_rate": 0.0006,
+ "loss": 4.424848556518555,
+ "step": 1301
+ },
+ {
+ "epoch": 18.083842794759825,
+ "grad_norm": 0.016383031383156776,
+ "learning_rate": 0.0006,
+ "loss": 4.494722366333008,
+ "step": 1302
+ },
+ {
+ "epoch": 18.097816593886463,
+ "grad_norm": 0.016265157610177994,
+ "learning_rate": 0.0006,
+ "loss": 4.416522979736328,
+ "step": 1303
+ },
+ {
+ "epoch": 18.111790393013102,
+ "grad_norm": 0.01702386513352394,
+ "learning_rate": 0.0006,
+ "loss": 4.600103378295898,
+ "step": 1304
+ },
+ {
+ "epoch": 18.125764192139737,
+ "grad_norm": 0.018495498225092888,
+ "learning_rate": 0.0006,
+ "loss": 4.588401794433594,
+ "step": 1305
+ },
+ {
+ "epoch": 18.139737991266376,
+ "grad_norm": 0.017921684309840202,
+ "learning_rate": 0.0006,
+ "loss": 4.478075981140137,
+ "step": 1306
+ },
+ {
+ "epoch": 18.153711790393015,
+ "grad_norm": 0.01801278069615364,
+ "learning_rate": 0.0006,
+ "loss": 4.413186550140381,
+ "step": 1307
+ },
+ {
+ "epoch": 18.16768558951965,
+ "grad_norm": 0.017211275175213814,
+ "learning_rate": 0.0006,
+ "loss": 4.461341857910156,
+ "step": 1308
+ },
+ {
+ "epoch": 18.18165938864629,
+ "grad_norm": 0.017589328810572624,
+ "learning_rate": 0.0006,
+ "loss": 4.546969413757324,
+ "step": 1309
+ },
+ {
+ "epoch": 18.195633187772927,
+ "grad_norm": 0.01797698810696602,
+ "learning_rate": 0.0006,
+ "loss": 4.530797004699707,
+ "step": 1310
+ },
+ {
+ "epoch": 18.209606986899562,
+ "grad_norm": 0.020408619195222855,
+ "learning_rate": 0.0006,
+ "loss": 4.62999153137207,
+ "step": 1311
+ },
+ {
+ "epoch": 18.2235807860262,
+ "grad_norm": 0.01994038000702858,
+ "learning_rate": 0.0006,
+ "loss": 4.586811065673828,
+ "step": 1312
+ },
+ {
+ "epoch": 18.23755458515284,
+ "grad_norm": 0.019381340593099594,
+ "learning_rate": 0.0006,
+ "loss": 4.4769368171691895,
+ "step": 1313
+ },
+ {
+ "epoch": 18.251528384279474,
+ "grad_norm": 0.020484555512666702,
+ "learning_rate": 0.0006,
+ "loss": 4.524302959442139,
+ "step": 1314
+ },
+ {
+ "epoch": 18.265502183406113,
+ "grad_norm": 0.02426416426897049,
+ "learning_rate": 0.0006,
+ "loss": 4.561705589294434,
+ "step": 1315
+ },
+ {
+ "epoch": 18.27947598253275,
+ "grad_norm": 0.024748101830482483,
+ "learning_rate": 0.0006,
+ "loss": 4.4654316902160645,
+ "step": 1316
+ },
+ {
+ "epoch": 18.29344978165939,
+ "grad_norm": 0.020703328773379326,
+ "learning_rate": 0.0006,
+ "loss": 4.53033447265625,
+ "step": 1317
+ },
+ {
+ "epoch": 18.307423580786025,
+ "grad_norm": 0.01765458658337593,
+ "learning_rate": 0.0006,
+ "loss": 4.5154266357421875,
+ "step": 1318
+ },
+ {
+ "epoch": 18.321397379912664,
+ "grad_norm": 0.018646420910954475,
+ "learning_rate": 0.0006,
+ "loss": 4.5224714279174805,
+ "step": 1319
+ },
+ {
+ "epoch": 18.335371179039303,
+ "grad_norm": 0.01907474733889103,
+ "learning_rate": 0.0006,
+ "loss": 4.5055437088012695,
+ "step": 1320
+ },
+ {
+ "epoch": 18.349344978165938,
+ "grad_norm": 0.018179846927523613,
+ "learning_rate": 0.0006,
+ "loss": 4.462161064147949,
+ "step": 1321
+ },
+ {
+ "epoch": 18.363318777292577,
+ "grad_norm": 0.016697878018021584,
+ "learning_rate": 0.0006,
+ "loss": 4.473718643188477,
+ "step": 1322
+ },
+ {
+ "epoch": 18.377292576419215,
+ "grad_norm": 0.017504019662737846,
+ "learning_rate": 0.0006,
+ "loss": 4.474348068237305,
+ "step": 1323
+ },
+ {
+ "epoch": 18.39126637554585,
+ "grad_norm": 0.018889913335442543,
+ "learning_rate": 0.0006,
+ "loss": 4.4531636238098145,
+ "step": 1324
+ },
+ {
+ "epoch": 18.40524017467249,
+ "grad_norm": 0.019167251884937286,
+ "learning_rate": 0.0006,
+ "loss": 4.4278364181518555,
+ "step": 1325
+ },
+ {
+ "epoch": 18.419213973799128,
+ "grad_norm": 0.015895741060376167,
+ "learning_rate": 0.0006,
+ "loss": 4.474592208862305,
+ "step": 1326
+ },
+ {
+ "epoch": 18.433187772925763,
+ "grad_norm": 0.01531882956624031,
+ "learning_rate": 0.0006,
+ "loss": 4.403757572174072,
+ "step": 1327
+ },
+ {
+ "epoch": 18.4471615720524,
+ "grad_norm": 0.01565849967300892,
+ "learning_rate": 0.0006,
+ "loss": 4.507664680480957,
+ "step": 1328
+ },
+ {
+ "epoch": 18.46113537117904,
+ "grad_norm": 0.01602315716445446,
+ "learning_rate": 0.0006,
+ "loss": 4.5648908615112305,
+ "step": 1329
+ },
+ {
+ "epoch": 18.475109170305675,
+ "grad_norm": 0.016467366367578506,
+ "learning_rate": 0.0006,
+ "loss": 4.428762912750244,
+ "step": 1330
+ },
+ {
+ "epoch": 18.489082969432314,
+ "grad_norm": 0.01604127697646618,
+ "learning_rate": 0.0006,
+ "loss": 4.527771472930908,
+ "step": 1331
+ },
+ {
+ "epoch": 18.503056768558952,
+ "grad_norm": 0.018825586885213852,
+ "learning_rate": 0.0006,
+ "loss": 4.485386848449707,
+ "step": 1332
+ },
+ {
+ "epoch": 18.51703056768559,
+ "grad_norm": 0.019258007407188416,
+ "learning_rate": 0.0006,
+ "loss": 4.414972305297852,
+ "step": 1333
+ },
+ {
+ "epoch": 18.531004366812226,
+ "grad_norm": 0.018192021176218987,
+ "learning_rate": 0.0006,
+ "loss": 4.412166595458984,
+ "step": 1334
+ },
+ {
+ "epoch": 18.544978165938865,
+ "grad_norm": 0.016120698302984238,
+ "learning_rate": 0.0006,
+ "loss": 4.484317779541016,
+ "step": 1335
+ },
+ {
+ "epoch": 18.558951965065503,
+ "grad_norm": 0.016496622934937477,
+ "learning_rate": 0.0006,
+ "loss": 4.514986515045166,
+ "step": 1336
+ },
+ {
+ "epoch": 18.57292576419214,
+ "grad_norm": 0.01795116998255253,
+ "learning_rate": 0.0006,
+ "loss": 4.41473388671875,
+ "step": 1337
+ },
+ {
+ "epoch": 18.586899563318777,
+ "grad_norm": 0.019062276929616928,
+ "learning_rate": 0.0006,
+ "loss": 4.601996898651123,
+ "step": 1338
+ },
+ {
+ "epoch": 18.600873362445416,
+ "grad_norm": 0.017174631357192993,
+ "learning_rate": 0.0006,
+ "loss": 4.571218490600586,
+ "step": 1339
+ },
+ {
+ "epoch": 18.61484716157205,
+ "grad_norm": 0.014015606604516506,
+ "learning_rate": 0.0006,
+ "loss": 4.526606559753418,
+ "step": 1340
+ },
+ {
+ "epoch": 18.62882096069869,
+ "grad_norm": 0.015358510427176952,
+ "learning_rate": 0.0006,
+ "loss": 4.604286193847656,
+ "step": 1341
+ },
+ {
+ "epoch": 18.64279475982533,
+ "grad_norm": 0.01619391329586506,
+ "learning_rate": 0.0006,
+ "loss": 4.560883522033691,
+ "step": 1342
+ },
+ {
+ "epoch": 18.656768558951963,
+ "grad_norm": 0.015972964465618134,
+ "learning_rate": 0.0006,
+ "loss": 4.479755401611328,
+ "step": 1343
+ },
+ {
+ "epoch": 18.670742358078602,
+ "grad_norm": 0.016861025243997574,
+ "learning_rate": 0.0006,
+ "loss": 4.474471092224121,
+ "step": 1344
+ },
+ {
+ "epoch": 18.68471615720524,
+ "grad_norm": 0.01822078227996826,
+ "learning_rate": 0.0006,
+ "loss": 4.509489059448242,
+ "step": 1345
+ },
+ {
+ "epoch": 18.69868995633188,
+ "grad_norm": 0.019415007904171944,
+ "learning_rate": 0.0006,
+ "loss": 4.508941173553467,
+ "step": 1346
+ },
+ {
+ "epoch": 18.712663755458514,
+ "grad_norm": 0.019856126978993416,
+ "learning_rate": 0.0006,
+ "loss": 4.446670055389404,
+ "step": 1347
+ },
+ {
+ "epoch": 18.726637554585153,
+ "grad_norm": 0.01842520199716091,
+ "learning_rate": 0.0006,
+ "loss": 4.524422645568848,
+ "step": 1348
+ },
+ {
+ "epoch": 18.74061135371179,
+ "grad_norm": 0.017196480184793472,
+ "learning_rate": 0.0006,
+ "loss": 4.498466968536377,
+ "step": 1349
+ },
+ {
+ "epoch": 18.754585152838427,
+ "grad_norm": 0.01823010854423046,
+ "learning_rate": 0.0006,
+ "loss": 4.534684181213379,
+ "step": 1350
+ },
+ {
+ "epoch": 18.768558951965066,
+ "grad_norm": 0.020398065447807312,
+ "learning_rate": 0.0006,
+ "loss": 4.456494331359863,
+ "step": 1351
+ },
+ {
+ "epoch": 18.782532751091704,
+ "grad_norm": 0.022992262616753578,
+ "learning_rate": 0.0006,
+ "loss": 4.476374626159668,
+ "step": 1352
+ },
+ {
+ "epoch": 18.79650655021834,
+ "grad_norm": 0.02235420234501362,
+ "learning_rate": 0.0006,
+ "loss": 4.587776184082031,
+ "step": 1353
+ },
+ {
+ "epoch": 18.810480349344978,
+ "grad_norm": 0.02093626745045185,
+ "learning_rate": 0.0006,
+ "loss": 4.424172401428223,
+ "step": 1354
+ },
+ {
+ "epoch": 18.824454148471617,
+ "grad_norm": 0.01997302658855915,
+ "learning_rate": 0.0006,
+ "loss": 4.448066711425781,
+ "step": 1355
+ },
+ {
+ "epoch": 18.83842794759825,
+ "grad_norm": 0.018265459686517715,
+ "learning_rate": 0.0006,
+ "loss": 4.436610698699951,
+ "step": 1356
+ },
+ {
+ "epoch": 18.85240174672489,
+ "grad_norm": 0.01673833839595318,
+ "learning_rate": 0.0006,
+ "loss": 4.5502214431762695,
+ "step": 1357
+ },
+ {
+ "epoch": 18.86637554585153,
+ "grad_norm": 0.017908833920955658,
+ "learning_rate": 0.0006,
+ "loss": 4.550530433654785,
+ "step": 1358
+ },
+ {
+ "epoch": 18.880349344978168,
+ "grad_norm": 0.01796654611825943,
+ "learning_rate": 0.0006,
+ "loss": 4.531463146209717,
+ "step": 1359
+ },
+ {
+ "epoch": 18.894323144104803,
+ "grad_norm": 0.01646183803677559,
+ "learning_rate": 0.0006,
+ "loss": 4.498582363128662,
+ "step": 1360
+ },
+ {
+ "epoch": 18.90829694323144,
+ "grad_norm": 0.015430327504873276,
+ "learning_rate": 0.0006,
+ "loss": 4.52571439743042,
+ "step": 1361
+ },
+ {
+ "epoch": 18.92227074235808,
+ "grad_norm": 0.016527952626347542,
+ "learning_rate": 0.0006,
+ "loss": 4.547013759613037,
+ "step": 1362
+ },
+ {
+ "epoch": 18.936244541484715,
+ "grad_norm": 0.016263660043478012,
+ "learning_rate": 0.0006,
+ "loss": 4.468754291534424,
+ "step": 1363
+ },
+ {
+ "epoch": 18.950218340611354,
+ "grad_norm": 0.016347210854291916,
+ "learning_rate": 0.0006,
+ "loss": 4.450741767883301,
+ "step": 1364
+ },
+ {
+ "epoch": 18.964192139737992,
+ "grad_norm": 0.01701398566365242,
+ "learning_rate": 0.0006,
+ "loss": 4.534899711608887,
+ "step": 1365
+ },
+ {
+ "epoch": 18.978165938864628,
+ "grad_norm": 0.017693081870675087,
+ "learning_rate": 0.0006,
+ "loss": 4.460553169250488,
+ "step": 1366
+ },
+ {
+ "epoch": 18.992139737991266,
+ "grad_norm": 0.018746308982372284,
+ "learning_rate": 0.0006,
+ "loss": 4.538792610168457,
+ "step": 1367
+ },
+ {
+ "epoch": 19.0,
+ "grad_norm": 0.020099416375160217,
+ "learning_rate": 0.0006,
+ "loss": 4.449617385864258,
+ "step": 1368
+ },
+ {
+ "epoch": 19.0,
+ "eval_loss": 4.7756571769714355,
+ "eval_runtime": 57.1804,
+ "eval_samples_per_second": 42.707,
+ "eval_steps_per_second": 1.347,
+ "step": 1368
+ },
+ {
+ "epoch": 19.01397379912664,
+ "grad_norm": 0.018726933747529984,
+ "learning_rate": 0.0006,
+ "loss": 4.446059703826904,
+ "step": 1369
+ },
+ {
+ "epoch": 19.027947598253274,
+ "grad_norm": 0.01956087350845337,
+ "learning_rate": 0.0006,
+ "loss": 4.4520039558410645,
+ "step": 1370
+ },
+ {
+ "epoch": 19.041921397379912,
+ "grad_norm": 0.020092811435461044,
+ "learning_rate": 0.0006,
+ "loss": 4.543343544006348,
+ "step": 1371
+ },
+ {
+ "epoch": 19.05589519650655,
+ "grad_norm": 0.01843622885644436,
+ "learning_rate": 0.0006,
+ "loss": 4.535579204559326,
+ "step": 1372
+ },
+ {
+ "epoch": 19.069868995633186,
+ "grad_norm": 0.016989488154649734,
+ "learning_rate": 0.0006,
+ "loss": 4.52316951751709,
+ "step": 1373
+ },
+ {
+ "epoch": 19.083842794759825,
+ "grad_norm": 0.01805422455072403,
+ "learning_rate": 0.0006,
+ "loss": 4.5393266677856445,
+ "step": 1374
+ },
+ {
+ "epoch": 19.097816593886463,
+ "grad_norm": 0.01874629408121109,
+ "learning_rate": 0.0006,
+ "loss": 4.422059059143066,
+ "step": 1375
+ },
+ {
+ "epoch": 19.111790393013102,
+ "grad_norm": 0.01804221048951149,
+ "learning_rate": 0.0006,
+ "loss": 4.50508451461792,
+ "step": 1376
+ },
+ {
+ "epoch": 19.125764192139737,
+ "grad_norm": 0.01785474829375744,
+ "learning_rate": 0.0006,
+ "loss": 4.412599563598633,
+ "step": 1377
+ },
+ {
+ "epoch": 19.139737991266376,
+ "grad_norm": 0.01971791312098503,
+ "learning_rate": 0.0006,
+ "loss": 4.518318176269531,
+ "step": 1378
+ },
+ {
+ "epoch": 19.153711790393015,
+ "grad_norm": 0.02061633951961994,
+ "learning_rate": 0.0006,
+ "loss": 4.332393169403076,
+ "step": 1379
+ },
+ {
+ "epoch": 19.16768558951965,
+ "grad_norm": 0.020563429221510887,
+ "learning_rate": 0.0006,
+ "loss": 4.445863723754883,
+ "step": 1380
+ },
+ {
+ "epoch": 19.18165938864629,
+ "grad_norm": 0.019995713606476784,
+ "learning_rate": 0.0006,
+ "loss": 4.497312068939209,
+ "step": 1381
+ },
+ {
+ "epoch": 19.195633187772927,
+ "grad_norm": 0.019823070615530014,
+ "learning_rate": 0.0006,
+ "loss": 4.475085258483887,
+ "step": 1382
+ },
+ {
+ "epoch": 19.209606986899562,
+ "grad_norm": 0.019039355218410492,
+ "learning_rate": 0.0006,
+ "loss": 4.50432014465332,
+ "step": 1383
+ },
+ {
+ "epoch": 19.2235807860262,
+ "grad_norm": 0.017876973375678062,
+ "learning_rate": 0.0006,
+ "loss": 4.384788513183594,
+ "step": 1384
+ },
+ {
+ "epoch": 19.23755458515284,
+ "grad_norm": 0.018697699531912804,
+ "learning_rate": 0.0006,
+ "loss": 4.410778045654297,
+ "step": 1385
+ },
+ {
+ "epoch": 19.251528384279474,
+ "grad_norm": 0.018826806917786598,
+ "learning_rate": 0.0006,
+ "loss": 4.524034023284912,
+ "step": 1386
+ },
+ {
+ "epoch": 19.265502183406113,
+ "grad_norm": 0.01824226602911949,
+ "learning_rate": 0.0006,
+ "loss": 4.526393890380859,
+ "step": 1387
+ },
+ {
+ "epoch": 19.27947598253275,
+ "grad_norm": 0.01753208599984646,
+ "learning_rate": 0.0006,
+ "loss": 4.493703365325928,
+ "step": 1388
+ },
+ {
+ "epoch": 19.29344978165939,
+ "grad_norm": 0.01742432825267315,
+ "learning_rate": 0.0006,
+ "loss": 4.392831802368164,
+ "step": 1389
+ },
+ {
+ "epoch": 19.307423580786025,
+ "grad_norm": 0.017099356278777122,
+ "learning_rate": 0.0006,
+ "loss": 4.578668594360352,
+ "step": 1390
+ },
+ {
+ "epoch": 19.321397379912664,
+ "grad_norm": 0.017946023494005203,
+ "learning_rate": 0.0006,
+ "loss": 4.456718444824219,
+ "step": 1391
+ },
+ {
+ "epoch": 19.335371179039303,
+ "grad_norm": 0.017705943435430527,
+ "learning_rate": 0.0006,
+ "loss": 4.553682327270508,
+ "step": 1392
+ },
+ {
+ "epoch": 19.349344978165938,
+ "grad_norm": 0.016647523269057274,
+ "learning_rate": 0.0006,
+ "loss": 4.525016784667969,
+ "step": 1393
+ },
+ {
+ "epoch": 19.363318777292577,
+ "grad_norm": 0.015407416969537735,
+ "learning_rate": 0.0006,
+ "loss": 4.500231742858887,
+ "step": 1394
+ },
+ {
+ "epoch": 19.377292576419215,
+ "grad_norm": 0.017264019697904587,
+ "learning_rate": 0.0006,
+ "loss": 4.4615349769592285,
+ "step": 1395
+ },
+ {
+ "epoch": 19.39126637554585,
+ "grad_norm": 0.01871555484831333,
+ "learning_rate": 0.0006,
+ "loss": 4.281026840209961,
+ "step": 1396
+ },
+ {
+ "epoch": 19.40524017467249,
+ "grad_norm": 0.017361650243401527,
+ "learning_rate": 0.0006,
+ "loss": 4.479382514953613,
+ "step": 1397
+ },
+ {
+ "epoch": 19.419213973799128,
+ "grad_norm": 0.019781362265348434,
+ "learning_rate": 0.0006,
+ "loss": 4.402320861816406,
+ "step": 1398
+ },
+ {
+ "epoch": 19.433187772925763,
+ "grad_norm": 0.019043298438191414,
+ "learning_rate": 0.0006,
+ "loss": 4.476983070373535,
+ "step": 1399
+ },
+ {
+ "epoch": 19.4471615720524,
+ "grad_norm": 0.017137985676527023,
+ "learning_rate": 0.0006,
+ "loss": 4.418689250946045,
+ "step": 1400
+ },
+ {
+ "epoch": 19.46113537117904,
+ "grad_norm": 0.016481833532452583,
+ "learning_rate": 0.0006,
+ "loss": 4.579145431518555,
+ "step": 1401
+ },
+ {
+ "epoch": 19.475109170305675,
+ "grad_norm": 0.016507714986801147,
+ "learning_rate": 0.0006,
+ "loss": 4.519158363342285,
+ "step": 1402
+ },
+ {
+ "epoch": 19.489082969432314,
+ "grad_norm": 0.018092872574925423,
+ "learning_rate": 0.0006,
+ "loss": 4.513748645782471,
+ "step": 1403
+ },
+ {
+ "epoch": 19.503056768558952,
+ "grad_norm": 0.018153440207242966,
+ "learning_rate": 0.0006,
+ "loss": 4.645384311676025,
+ "step": 1404
+ },
+ {
+ "epoch": 19.51703056768559,
+ "grad_norm": 0.01657119207084179,
+ "learning_rate": 0.0006,
+ "loss": 4.399757385253906,
+ "step": 1405
+ },
+ {
+ "epoch": 19.531004366812226,
+ "grad_norm": 0.0156280305236578,
+ "learning_rate": 0.0006,
+ "loss": 4.467836380004883,
+ "step": 1406
+ },
+ {
+ "epoch": 19.544978165938865,
+ "grad_norm": 0.016838403418660164,
+ "learning_rate": 0.0006,
+ "loss": 4.529176235198975,
+ "step": 1407
+ },
+ {
+ "epoch": 19.558951965065503,
+ "grad_norm": 0.01743236929178238,
+ "learning_rate": 0.0006,
+ "loss": 4.440032482147217,
+ "step": 1408
+ },
+ {
+ "epoch": 19.57292576419214,
+ "grad_norm": 0.01553610060364008,
+ "learning_rate": 0.0006,
+ "loss": 4.425588607788086,
+ "step": 1409
+ },
+ {
+ "epoch": 19.586899563318777,
+ "grad_norm": 0.017259271815419197,
+ "learning_rate": 0.0006,
+ "loss": 4.405452728271484,
+ "step": 1410
+ },
+ {
+ "epoch": 19.600873362445416,
+ "grad_norm": 0.015568481758236885,
+ "learning_rate": 0.0006,
+ "loss": 4.444279193878174,
+ "step": 1411
+ },
+ {
+ "epoch": 19.61484716157205,
+ "grad_norm": 0.015331248752772808,
+ "learning_rate": 0.0006,
+ "loss": 4.451846122741699,
+ "step": 1412
+ },
+ {
+ "epoch": 19.62882096069869,
+ "grad_norm": 0.015667378902435303,
+ "learning_rate": 0.0006,
+ "loss": 4.533931255340576,
+ "step": 1413
+ },
+ {
+ "epoch": 19.64279475982533,
+ "grad_norm": 0.016177602112293243,
+ "learning_rate": 0.0006,
+ "loss": 4.467313289642334,
+ "step": 1414
+ },
+ {
+ "epoch": 19.656768558951963,
+ "grad_norm": 0.01733129471540451,
+ "learning_rate": 0.0006,
+ "loss": 4.496893882751465,
+ "step": 1415
+ },
+ {
+ "epoch": 19.670742358078602,
+ "grad_norm": 0.016539594158530235,
+ "learning_rate": 0.0006,
+ "loss": 4.401961326599121,
+ "step": 1416
+ },
+ {
+ "epoch": 19.68471615720524,
+ "grad_norm": 0.016058551147580147,
+ "learning_rate": 0.0006,
+ "loss": 4.503375053405762,
+ "step": 1417
+ },
+ {
+ "epoch": 19.69868995633188,
+ "grad_norm": 0.01715777814388275,
+ "learning_rate": 0.0006,
+ "loss": 4.3808369636535645,
+ "step": 1418
+ },
+ {
+ "epoch": 19.712663755458514,
+ "grad_norm": 0.01681119203567505,
+ "learning_rate": 0.0006,
+ "loss": 4.5469560623168945,
+ "step": 1419
+ },
+ {
+ "epoch": 19.726637554585153,
+ "grad_norm": 0.01634550653398037,
+ "learning_rate": 0.0006,
+ "loss": 4.328197956085205,
+ "step": 1420
+ },
+ {
+ "epoch": 19.74061135371179,
+ "grad_norm": 0.01549347210675478,
+ "learning_rate": 0.0006,
+ "loss": 4.534621238708496,
+ "step": 1421
+ },
+ {
+ "epoch": 19.754585152838427,
+ "grad_norm": 0.01654665172100067,
+ "learning_rate": 0.0006,
+ "loss": 4.4724225997924805,
+ "step": 1422
+ },
+ {
+ "epoch": 19.768558951965066,
+ "grad_norm": 0.017119025811553,
+ "learning_rate": 0.0006,
+ "loss": 4.3897247314453125,
+ "step": 1423
+ },
+ {
+ "epoch": 19.782532751091704,
+ "grad_norm": 0.01603223942220211,
+ "learning_rate": 0.0006,
+ "loss": 4.412731170654297,
+ "step": 1424
+ },
+ {
+ "epoch": 19.79650655021834,
+ "grad_norm": 0.016774067655205727,
+ "learning_rate": 0.0006,
+ "loss": 4.54707145690918,
+ "step": 1425
+ },
+ {
+ "epoch": 19.810480349344978,
+ "grad_norm": 0.016453025862574577,
+ "learning_rate": 0.0006,
+ "loss": 4.462501049041748,
+ "step": 1426
+ },
+ {
+ "epoch": 19.824454148471617,
+ "grad_norm": 0.01597629487514496,
+ "learning_rate": 0.0006,
+ "loss": 4.229758262634277,
+ "step": 1427
+ },
+ {
+ "epoch": 19.83842794759825,
+ "grad_norm": 0.017999034374952316,
+ "learning_rate": 0.0006,
+ "loss": 4.389263153076172,
+ "step": 1428
+ },
+ {
+ "epoch": 19.85240174672489,
+ "grad_norm": 0.02165898121893406,
+ "learning_rate": 0.0006,
+ "loss": 4.342130661010742,
+ "step": 1429
+ },
+ {
+ "epoch": 19.86637554585153,
+ "grad_norm": 0.02222822792828083,
+ "learning_rate": 0.0006,
+ "loss": 4.492661476135254,
+ "step": 1430
+ },
+ {
+ "epoch": 19.880349344978168,
+ "grad_norm": 0.019378066062927246,
+ "learning_rate": 0.0006,
+ "loss": 4.4614105224609375,
+ "step": 1431
+ },
+ {
+ "epoch": 19.894323144104803,
+ "grad_norm": 0.019908083602786064,
+ "learning_rate": 0.0006,
+ "loss": 4.370929718017578,
+ "step": 1432
+ },
+ {
+ "epoch": 19.90829694323144,
+ "grad_norm": 0.022296471521258354,
+ "learning_rate": 0.0006,
+ "loss": 4.438320159912109,
+ "step": 1433
+ },
+ {
+ "epoch": 19.92227074235808,
+ "grad_norm": 0.021810343489050865,
+ "learning_rate": 0.0006,
+ "loss": 4.462067127227783,
+ "step": 1434
+ },
+ {
+ "epoch": 19.936244541484715,
+ "grad_norm": 0.02129209227859974,
+ "learning_rate": 0.0006,
+ "loss": 4.541261196136475,
+ "step": 1435
+ },
+ {
+ "epoch": 19.950218340611354,
+ "grad_norm": 0.02079629711806774,
+ "learning_rate": 0.0006,
+ "loss": 4.36712646484375,
+ "step": 1436
+ },
+ {
+ "epoch": 19.964192139737992,
+ "grad_norm": 0.022501740604639053,
+ "learning_rate": 0.0006,
+ "loss": 4.5279083251953125,
+ "step": 1437
+ },
+ {
+ "epoch": 19.978165938864628,
+ "grad_norm": 0.019121676683425903,
+ "learning_rate": 0.0006,
+ "loss": 4.477789878845215,
+ "step": 1438
+ },
+ {
+ "epoch": 19.992139737991266,
+ "grad_norm": 0.017429588362574577,
+ "learning_rate": 0.0006,
+ "loss": 4.4023847579956055,
+ "step": 1439
+ },
+ {
+ "epoch": 20.0,
+ "grad_norm": 0.018857665359973907,
+ "learning_rate": 0.0006,
+ "loss": 4.494743347167969,
+ "step": 1440
+ },
+ {
+ "epoch": 20.0,
+ "eval_loss": 4.694555759429932,
+ "eval_runtime": 56.4146,
+ "eval_samples_per_second": 43.287,
+ "eval_steps_per_second": 1.365,
+ "step": 1440
+ },
+ {
+ "epoch": 20.01397379912664,
+ "grad_norm": 0.016505861654877663,
+ "learning_rate": 0.0006,
+ "loss": 4.506229877471924,
+ "step": 1441
+ },
+ {
+ "epoch": 20.027947598253274,
+ "grad_norm": 0.017571842297911644,
+ "learning_rate": 0.0006,
+ "loss": 4.384073257446289,
+ "step": 1442
+ },
+ {
+ "epoch": 20.041921397379912,
+ "grad_norm": 0.019398365169763565,
+ "learning_rate": 0.0006,
+ "loss": 4.482177734375,
+ "step": 1443
+ },
+ {
+ "epoch": 20.05589519650655,
+ "grad_norm": 0.02011062018573284,
+ "learning_rate": 0.0006,
+ "loss": 4.440773963928223,
+ "step": 1444
+ },
+ {
+ "epoch": 20.069868995633186,
+ "grad_norm": 0.019026953727006912,
+ "learning_rate": 0.0006,
+ "loss": 4.507168769836426,
+ "step": 1445
+ },
+ {
+ "epoch": 20.083842794759825,
+ "grad_norm": 0.020506031811237335,
+ "learning_rate": 0.0006,
+ "loss": 4.469079971313477,
+ "step": 1446
+ },
+ {
+ "epoch": 20.097816593886463,
+ "grad_norm": 0.019608162343502045,
+ "learning_rate": 0.0006,
+ "loss": 4.395927429199219,
+ "step": 1447
+ },
+ {
+ "epoch": 20.111790393013102,
+ "grad_norm": 0.019419124349951744,
+ "learning_rate": 0.0006,
+ "loss": 4.484235763549805,
+ "step": 1448
+ },
+ {
+ "epoch": 20.125764192139737,
+ "grad_norm": 0.021775128319859505,
+ "learning_rate": 0.0006,
+ "loss": 4.333173751831055,
+ "step": 1449
+ },
+ {
+ "epoch": 20.139737991266376,
+ "grad_norm": 0.025512272492051125,
+ "learning_rate": 0.0006,
+ "loss": 4.460362434387207,
+ "step": 1450
+ },
+ {
+ "epoch": 20.153711790393015,
+ "grad_norm": 0.025157401338219643,
+ "learning_rate": 0.0006,
+ "loss": 4.4361395835876465,
+ "step": 1451
+ },
+ {
+ "epoch": 20.16768558951965,
+ "grad_norm": 0.019862616434693336,
+ "learning_rate": 0.0006,
+ "loss": 4.426807403564453,
+ "step": 1452
+ },
+ {
+ "epoch": 20.18165938864629,
+ "grad_norm": 0.01932152360677719,
+ "learning_rate": 0.0006,
+ "loss": 4.375677108764648,
+ "step": 1453
+ },
+ {
+ "epoch": 20.195633187772927,
+ "grad_norm": 0.01760999858379364,
+ "learning_rate": 0.0006,
+ "loss": 4.42259407043457,
+ "step": 1454
+ },
+ {
+ "epoch": 20.209606986899562,
+ "grad_norm": 0.016998112201690674,
+ "learning_rate": 0.0006,
+ "loss": 4.481280326843262,
+ "step": 1455
+ },
+ {
+ "epoch": 20.2235807860262,
+ "grad_norm": 0.017199190333485603,
+ "learning_rate": 0.0006,
+ "loss": 4.496712684631348,
+ "step": 1456
+ },
+ {
+ "epoch": 20.23755458515284,
+ "grad_norm": 0.016833679750561714,
+ "learning_rate": 0.0006,
+ "loss": 4.40736722946167,
+ "step": 1457
+ },
+ {
+ "epoch": 20.251528384279474,
+ "grad_norm": 0.017546730116009712,
+ "learning_rate": 0.0006,
+ "loss": 4.4915266036987305,
+ "step": 1458
+ },
+ {
+ "epoch": 20.265502183406113,
+ "grad_norm": 0.016595078632235527,
+ "learning_rate": 0.0006,
+ "loss": 4.460815906524658,
+ "step": 1459
+ },
+ {
+ "epoch": 20.27947598253275,
+ "grad_norm": 0.017728568986058235,
+ "learning_rate": 0.0006,
+ "loss": 4.346048831939697,
+ "step": 1460
+ },
+ {
+ "epoch": 20.29344978165939,
+ "grad_norm": 0.01849197782576084,
+ "learning_rate": 0.0006,
+ "loss": 4.399681091308594,
+ "step": 1461
+ },
+ {
+ "epoch": 20.307423580786025,
+ "grad_norm": 0.018336588516831398,
+ "learning_rate": 0.0006,
+ "loss": 4.414183616638184,
+ "step": 1462
+ },
+ {
+ "epoch": 20.321397379912664,
+ "grad_norm": 0.018086519092321396,
+ "learning_rate": 0.0006,
+ "loss": 4.346586227416992,
+ "step": 1463
+ },
+ {
+ "epoch": 20.335371179039303,
+ "grad_norm": 0.019697299227118492,
+ "learning_rate": 0.0006,
+ "loss": 4.4420952796936035,
+ "step": 1464
+ },
+ {
+ "epoch": 20.349344978165938,
+ "grad_norm": 0.02198074571788311,
+ "learning_rate": 0.0006,
+ "loss": 4.588729381561279,
+ "step": 1465
+ },
+ {
+ "epoch": 20.363318777292577,
+ "grad_norm": 0.022813329473137856,
+ "learning_rate": 0.0006,
+ "loss": 4.49729585647583,
+ "step": 1466
+ },
+ {
+ "epoch": 20.377292576419215,
+ "grad_norm": 0.022006649523973465,
+ "learning_rate": 0.0006,
+ "loss": 4.475641250610352,
+ "step": 1467
+ },
+ {
+ "epoch": 20.39126637554585,
+ "grad_norm": 0.023188438266515732,
+ "learning_rate": 0.0006,
+ "loss": 4.404322624206543,
+ "step": 1468
+ },
+ {
+ "epoch": 20.40524017467249,
+ "grad_norm": 0.022226877510547638,
+ "learning_rate": 0.0006,
+ "loss": 4.50904655456543,
+ "step": 1469
+ },
+ {
+ "epoch": 20.419213973799128,
+ "grad_norm": 0.021082637831568718,
+ "learning_rate": 0.0006,
+ "loss": 4.339011192321777,
+ "step": 1470
+ },
+ {
+ "epoch": 20.433187772925763,
+ "grad_norm": 0.02013542130589485,
+ "learning_rate": 0.0006,
+ "loss": 4.516942024230957,
+ "step": 1471
+ },
+ {
+ "epoch": 20.4471615720524,
+ "grad_norm": 0.01898968778550625,
+ "learning_rate": 0.0006,
+ "loss": 4.315804481506348,
+ "step": 1472
+ },
+ {
+ "epoch": 20.46113537117904,
+ "grad_norm": 0.019450880587100983,
+ "learning_rate": 0.0006,
+ "loss": 4.4794921875,
+ "step": 1473
+ },
+ {
+ "epoch": 20.475109170305675,
+ "grad_norm": 0.018577802926301956,
+ "learning_rate": 0.0006,
+ "loss": 4.422323703765869,
+ "step": 1474
+ },
+ {
+ "epoch": 20.489082969432314,
+ "grad_norm": 0.01778412237763405,
+ "learning_rate": 0.0006,
+ "loss": 4.429789066314697,
+ "step": 1475
+ },
+ {
+ "epoch": 20.503056768558952,
+ "grad_norm": 0.017134476453065872,
+ "learning_rate": 0.0006,
+ "loss": 4.416459083557129,
+ "step": 1476
+ },
+ {
+ "epoch": 20.51703056768559,
+ "grad_norm": 0.017153557389974594,
+ "learning_rate": 0.0006,
+ "loss": 4.446170806884766,
+ "step": 1477
+ },
+ {
+ "epoch": 20.531004366812226,
+ "grad_norm": 0.014998532831668854,
+ "learning_rate": 0.0006,
+ "loss": 4.419447898864746,
+ "step": 1478
+ },
+ {
+ "epoch": 20.544978165938865,
+ "grad_norm": 0.015877850353717804,
+ "learning_rate": 0.0006,
+ "loss": 4.312338829040527,
+ "step": 1479
+ },
+ {
+ "epoch": 20.558951965065503,
+ "grad_norm": 0.01569489948451519,
+ "learning_rate": 0.0006,
+ "loss": 4.572905540466309,
+ "step": 1480
+ },
+ {
+ "epoch": 20.57292576419214,
+ "grad_norm": 0.01591474376618862,
+ "learning_rate": 0.0006,
+ "loss": 4.414907455444336,
+ "step": 1481
+ },
+ {
+ "epoch": 20.586899563318777,
+ "grad_norm": 0.015395903028547764,
+ "learning_rate": 0.0006,
+ "loss": 4.549131393432617,
+ "step": 1482
+ },
+ {
+ "epoch": 20.600873362445416,
+ "grad_norm": 0.015715358778834343,
+ "learning_rate": 0.0006,
+ "loss": 4.530194282531738,
+ "step": 1483
+ },
+ {
+ "epoch": 20.61484716157205,
+ "grad_norm": 0.016042975708842278,
+ "learning_rate": 0.0006,
+ "loss": 4.322149276733398,
+ "step": 1484
+ },
+ {
+ "epoch": 20.62882096069869,
+ "grad_norm": 0.01664186641573906,
+ "learning_rate": 0.0006,
+ "loss": 4.424866676330566,
+ "step": 1485
+ },
+ {
+ "epoch": 20.64279475982533,
+ "grad_norm": 0.01631534844636917,
+ "learning_rate": 0.0006,
+ "loss": 4.441752910614014,
+ "step": 1486
+ },
+ {
+ "epoch": 20.656768558951963,
+ "grad_norm": 0.014057058840990067,
+ "learning_rate": 0.0006,
+ "loss": 4.399982452392578,
+ "step": 1487
+ },
+ {
+ "epoch": 20.670742358078602,
+ "grad_norm": 0.015076273120939732,
+ "learning_rate": 0.0006,
+ "loss": 4.441324234008789,
+ "step": 1488
+ },
+ {
+ "epoch": 20.68471615720524,
+ "grad_norm": 0.014138277620077133,
+ "learning_rate": 0.0006,
+ "loss": 4.4312543869018555,
+ "step": 1489
+ },
+ {
+ "epoch": 20.69868995633188,
+ "grad_norm": 0.013762440532445908,
+ "learning_rate": 0.0006,
+ "loss": 4.387807846069336,
+ "step": 1490
+ },
+ {
+ "epoch": 20.712663755458514,
+ "grad_norm": 0.015058411285281181,
+ "learning_rate": 0.0006,
+ "loss": 4.594747543334961,
+ "step": 1491
+ },
+ {
+ "epoch": 20.726637554585153,
+ "grad_norm": 0.013827123679220676,
+ "learning_rate": 0.0006,
+ "loss": 4.434604167938232,
+ "step": 1492
+ },
+ {
+ "epoch": 20.74061135371179,
+ "grad_norm": 0.015315295197069645,
+ "learning_rate": 0.0006,
+ "loss": 4.588066101074219,
+ "step": 1493
+ },
+ {
+ "epoch": 20.754585152838427,
+ "grad_norm": 0.018678732216358185,
+ "learning_rate": 0.0006,
+ "loss": 4.3514204025268555,
+ "step": 1494
+ },
+ {
+ "epoch": 20.768558951965066,
+ "grad_norm": 0.019521350041031837,
+ "learning_rate": 0.0006,
+ "loss": 4.437658786773682,
+ "step": 1495
+ },
+ {
+ "epoch": 20.782532751091704,
+ "grad_norm": 0.018363403156399727,
+ "learning_rate": 0.0006,
+ "loss": 4.495002269744873,
+ "step": 1496
+ },
+ {
+ "epoch": 20.79650655021834,
+ "grad_norm": 0.016830725595355034,
+ "learning_rate": 0.0006,
+ "loss": 4.402471542358398,
+ "step": 1497
+ },
+ {
+ "epoch": 20.810480349344978,
+ "grad_norm": 0.01722072623670101,
+ "learning_rate": 0.0006,
+ "loss": 4.437653541564941,
+ "step": 1498
+ },
+ {
+ "epoch": 20.824454148471617,
+ "grad_norm": 0.017459599301218987,
+ "learning_rate": 0.0006,
+ "loss": 4.366603851318359,
+ "step": 1499
+ },
+ {
+ "epoch": 20.83842794759825,
+ "grad_norm": 0.018902219831943512,
+ "learning_rate": 0.0006,
+ "loss": 4.501561164855957,
+ "step": 1500
+ },
+ {
+ "epoch": 20.85240174672489,
+ "grad_norm": 0.020814096555113792,
+ "learning_rate": 0.0006,
+ "loss": 4.535295486450195,
+ "step": 1501
+ },
+ {
+ "epoch": 20.86637554585153,
+ "grad_norm": 0.018990976735949516,
+ "learning_rate": 0.0006,
+ "loss": 4.333580017089844,
+ "step": 1502
+ },
+ {
+ "epoch": 20.880349344978168,
+ "grad_norm": 0.01837880350649357,
+ "learning_rate": 0.0006,
+ "loss": 4.313388824462891,
+ "step": 1503
+ },
+ {
+ "epoch": 20.894323144104803,
+ "grad_norm": 0.019820360466837883,
+ "learning_rate": 0.0006,
+ "loss": 4.388157844543457,
+ "step": 1504
+ },
+ {
+ "epoch": 20.90829694323144,
+ "grad_norm": 0.020181143656373024,
+ "learning_rate": 0.0006,
+ "loss": 4.572224140167236,
+ "step": 1505
+ },
+ {
+ "epoch": 20.92227074235808,
+ "grad_norm": 0.021206077188253403,
+ "learning_rate": 0.0006,
+ "loss": 4.3127970695495605,
+ "step": 1506
+ },
+ {
+ "epoch": 20.936244541484715,
+ "grad_norm": 0.017962060868740082,
+ "learning_rate": 0.0006,
+ "loss": 4.384734630584717,
+ "step": 1507
+ },
+ {
+ "epoch": 20.950218340611354,
+ "grad_norm": 0.01737220585346222,
+ "learning_rate": 0.0006,
+ "loss": 4.4075751304626465,
+ "step": 1508
+ },
+ {
+ "epoch": 20.964192139737992,
+ "grad_norm": 0.01839268207550049,
+ "learning_rate": 0.0006,
+ "loss": 4.553779125213623,
+ "step": 1509
+ },
+ {
+ "epoch": 20.978165938864628,
+ "grad_norm": 0.020102640613913536,
+ "learning_rate": 0.0006,
+ "loss": 4.399398326873779,
+ "step": 1510
+ },
+ {
+ "epoch": 20.992139737991266,
+ "grad_norm": 0.018831463530659676,
+ "learning_rate": 0.0006,
+ "loss": 4.358499526977539,
+ "step": 1511
+ },
+ {
+ "epoch": 21.0,
+ "grad_norm": 0.018254734575748444,
+ "learning_rate": 0.0006,
+ "loss": 4.384119510650635,
+ "step": 1512
+ },
+ {
+ "epoch": 21.0,
+ "eval_loss": 4.7524237632751465,
+ "eval_runtime": 57.8812,
+ "eval_samples_per_second": 42.19,
+ "eval_steps_per_second": 1.33,
+ "step": 1512
+ },
+ {
+ "epoch": 21.01397379912664,
+ "grad_norm": 0.019728587940335274,
+ "learning_rate": 0.0006,
+ "loss": 4.32363224029541,
+ "step": 1513
+ },
+ {
+ "epoch": 21.027947598253274,
+ "grad_norm": 0.02216625027358532,
+ "learning_rate": 0.0006,
+ "loss": 4.436988353729248,
+ "step": 1514
+ },
+ {
+ "epoch": 21.041921397379912,
+ "grad_norm": 0.02227606810629368,
+ "learning_rate": 0.0006,
+ "loss": 4.406063556671143,
+ "step": 1515
+ },
+ {
+ "epoch": 21.05589519650655,
+ "grad_norm": 0.020187009125947952,
+ "learning_rate": 0.0006,
+ "loss": 4.324402332305908,
+ "step": 1516
+ },
+ {
+ "epoch": 21.069868995633186,
+ "grad_norm": 0.02095157466828823,
+ "learning_rate": 0.0006,
+ "loss": 4.456544876098633,
+ "step": 1517
+ },
+ {
+ "epoch": 21.083842794759825,
+ "grad_norm": 0.02146013453602791,
+ "learning_rate": 0.0006,
+ "loss": 4.368387222290039,
+ "step": 1518
+ },
+ {
+ "epoch": 21.097816593886463,
+ "grad_norm": 0.02162035182118416,
+ "learning_rate": 0.0006,
+ "loss": 4.48838996887207,
+ "step": 1519
+ },
+ {
+ "epoch": 21.111790393013102,
+ "grad_norm": 0.021745676174759865,
+ "learning_rate": 0.0006,
+ "loss": 4.30778694152832,
+ "step": 1520
+ },
+ {
+ "epoch": 21.125764192139737,
+ "grad_norm": 0.02173588052392006,
+ "learning_rate": 0.0006,
+ "loss": 4.433149814605713,
+ "step": 1521
+ },
+ {
+ "epoch": 21.139737991266376,
+ "grad_norm": 0.020950648933649063,
+ "learning_rate": 0.0006,
+ "loss": 4.520089149475098,
+ "step": 1522
+ },
+ {
+ "epoch": 21.153711790393015,
+ "grad_norm": 0.020684752613306046,
+ "learning_rate": 0.0006,
+ "loss": 4.372437477111816,
+ "step": 1523
+ },
+ {
+ "epoch": 21.16768558951965,
+ "grad_norm": 0.019472869113087654,
+ "learning_rate": 0.0006,
+ "loss": 4.437015533447266,
+ "step": 1524
+ },
+ {
+ "epoch": 21.18165938864629,
+ "grad_norm": 0.01846962980926037,
+ "learning_rate": 0.0006,
+ "loss": 4.449531555175781,
+ "step": 1525
+ },
+ {
+ "epoch": 21.195633187772927,
+ "grad_norm": 0.017377803102135658,
+ "learning_rate": 0.0006,
+ "loss": 4.458279132843018,
+ "step": 1526
+ },
+ {
+ "epoch": 21.209606986899562,
+ "grad_norm": 0.016002200543880463,
+ "learning_rate": 0.0006,
+ "loss": 4.266297817230225,
+ "step": 1527
+ },
+ {
+ "epoch": 21.2235807860262,
+ "grad_norm": 0.015575903467833996,
+ "learning_rate": 0.0006,
+ "loss": 4.342632293701172,
+ "step": 1528
+ },
+ {
+ "epoch": 21.23755458515284,
+ "grad_norm": 0.015634698793292046,
+ "learning_rate": 0.0006,
+ "loss": 4.338623523712158,
+ "step": 1529
+ },
+ {
+ "epoch": 21.251528384279474,
+ "grad_norm": 0.01449581328779459,
+ "learning_rate": 0.0006,
+ "loss": 4.354330062866211,
+ "step": 1530
+ },
+ {
+ "epoch": 21.265502183406113,
+ "grad_norm": 0.01487616915255785,
+ "learning_rate": 0.0006,
+ "loss": 4.486835479736328,
+ "step": 1531
+ },
+ {
+ "epoch": 21.27947598253275,
+ "grad_norm": 0.016346555203199387,
+ "learning_rate": 0.0006,
+ "loss": 4.407098770141602,
+ "step": 1532
+ },
+ {
+ "epoch": 21.29344978165939,
+ "grad_norm": 0.01682182401418686,
+ "learning_rate": 0.0006,
+ "loss": 4.436809539794922,
+ "step": 1533
+ },
+ {
+ "epoch": 21.307423580786025,
+ "grad_norm": 0.017704004421830177,
+ "learning_rate": 0.0006,
+ "loss": 4.3854780197143555,
+ "step": 1534
+ },
+ {
+ "epoch": 21.321397379912664,
+ "grad_norm": 0.018437420949339867,
+ "learning_rate": 0.0006,
+ "loss": 4.413283824920654,
+ "step": 1535
+ },
+ {
+ "epoch": 21.335371179039303,
+ "grad_norm": 0.016911081969738007,
+ "learning_rate": 0.0006,
+ "loss": 4.357306003570557,
+ "step": 1536
+ },
+ {
+ "epoch": 21.349344978165938,
+ "grad_norm": 0.01642223633825779,
+ "learning_rate": 0.0006,
+ "loss": 4.422322750091553,
+ "step": 1537
+ },
+ {
+ "epoch": 21.363318777292577,
+ "grad_norm": 0.01780877821147442,
+ "learning_rate": 0.0006,
+ "loss": 4.387620449066162,
+ "step": 1538
+ },
+ {
+ "epoch": 21.377292576419215,
+ "grad_norm": 0.01653975248336792,
+ "learning_rate": 0.0006,
+ "loss": 4.327267646789551,
+ "step": 1539
+ },
+ {
+ "epoch": 21.39126637554585,
+ "grad_norm": 0.01572156324982643,
+ "learning_rate": 0.0006,
+ "loss": 4.475979804992676,
+ "step": 1540
+ },
+ {
+ "epoch": 21.40524017467249,
+ "grad_norm": 0.017156066372990608,
+ "learning_rate": 0.0006,
+ "loss": 4.4011125564575195,
+ "step": 1541
+ },
+ {
+ "epoch": 21.419213973799128,
+ "grad_norm": 0.01788829080760479,
+ "learning_rate": 0.0006,
+ "loss": 4.4174089431762695,
+ "step": 1542
+ },
+ {
+ "epoch": 21.433187772925763,
+ "grad_norm": 0.01728985086083412,
+ "learning_rate": 0.0006,
+ "loss": 4.429244041442871,
+ "step": 1543
+ },
+ {
+ "epoch": 21.4471615720524,
+ "grad_norm": 0.016534466296434402,
+ "learning_rate": 0.0006,
+ "loss": 4.326068878173828,
+ "step": 1544
+ },
+ {
+ "epoch": 21.46113537117904,
+ "grad_norm": 0.016582584008574486,
+ "learning_rate": 0.0006,
+ "loss": 4.2975969314575195,
+ "step": 1545
+ },
+ {
+ "epoch": 21.475109170305675,
+ "grad_norm": 0.018034178763628006,
+ "learning_rate": 0.0006,
+ "loss": 4.501319408416748,
+ "step": 1546
+ },
+ {
+ "epoch": 21.489082969432314,
+ "grad_norm": 0.017827702686190605,
+ "learning_rate": 0.0006,
+ "loss": 4.502284526824951,
+ "step": 1547
+ },
+ {
+ "epoch": 21.503056768558952,
+ "grad_norm": 0.017182299867272377,
+ "learning_rate": 0.0006,
+ "loss": 4.380828857421875,
+ "step": 1548
+ },
+ {
+ "epoch": 21.51703056768559,
+ "grad_norm": 0.019101588055491447,
+ "learning_rate": 0.0006,
+ "loss": 4.31564998626709,
+ "step": 1549
+ },
+ {
+ "epoch": 21.531004366812226,
+ "grad_norm": 0.0204677302390337,
+ "learning_rate": 0.0006,
+ "loss": 4.3840131759643555,
+ "step": 1550
+ },
+ {
+ "epoch": 21.544978165938865,
+ "grad_norm": 0.0188993439078331,
+ "learning_rate": 0.0006,
+ "loss": 4.4257731437683105,
+ "step": 1551
+ },
+ {
+ "epoch": 21.558951965065503,
+ "grad_norm": 0.01730230636894703,
+ "learning_rate": 0.0006,
+ "loss": 4.445132255554199,
+ "step": 1552
+ },
+ {
+ "epoch": 21.57292576419214,
+ "grad_norm": 0.01801574043929577,
+ "learning_rate": 0.0006,
+ "loss": 4.427064895629883,
+ "step": 1553
+ },
+ {
+ "epoch": 21.586899563318777,
+ "grad_norm": 0.018417565152049065,
+ "learning_rate": 0.0006,
+ "loss": 4.319595813751221,
+ "step": 1554
+ },
+ {
+ "epoch": 21.600873362445416,
+ "grad_norm": 0.01683160290122032,
+ "learning_rate": 0.0006,
+ "loss": 4.349218368530273,
+ "step": 1555
+ },
+ {
+ "epoch": 21.61484716157205,
+ "grad_norm": 0.016973106190562248,
+ "learning_rate": 0.0006,
+ "loss": 4.375990390777588,
+ "step": 1556
+ },
+ {
+ "epoch": 21.62882096069869,
+ "grad_norm": 0.015688583254814148,
+ "learning_rate": 0.0006,
+ "loss": 4.408170223236084,
+ "step": 1557
+ },
+ {
+ "epoch": 21.64279475982533,
+ "grad_norm": 0.01573043316602707,
+ "learning_rate": 0.0006,
+ "loss": 4.416255950927734,
+ "step": 1558
+ },
+ {
+ "epoch": 21.656768558951963,
+ "grad_norm": 0.01744293048977852,
+ "learning_rate": 0.0006,
+ "loss": 4.477141380310059,
+ "step": 1559
+ },
+ {
+ "epoch": 21.670742358078602,
+ "grad_norm": 0.017503513023257256,
+ "learning_rate": 0.0006,
+ "loss": 4.3367486000061035,
+ "step": 1560
+ },
+ {
+ "epoch": 21.68471615720524,
+ "grad_norm": 0.016592150554060936,
+ "learning_rate": 0.0006,
+ "loss": 4.314785480499268,
+ "step": 1561
+ },
+ {
+ "epoch": 21.69868995633188,
+ "grad_norm": 0.015774084255099297,
+ "learning_rate": 0.0006,
+ "loss": 4.448209285736084,
+ "step": 1562
+ },
+ {
+ "epoch": 21.712663755458514,
+ "grad_norm": 0.015779150649905205,
+ "learning_rate": 0.0006,
+ "loss": 4.392690658569336,
+ "step": 1563
+ },
+ {
+ "epoch": 21.726637554585153,
+ "grad_norm": 0.015392648056149483,
+ "learning_rate": 0.0006,
+ "loss": 4.437741756439209,
+ "step": 1564
+ },
+ {
+ "epoch": 21.74061135371179,
+ "grad_norm": 0.01530836895108223,
+ "learning_rate": 0.0006,
+ "loss": 4.424595832824707,
+ "step": 1565
+ },
+ {
+ "epoch": 21.754585152838427,
+ "grad_norm": 0.015600779093801975,
+ "learning_rate": 0.0006,
+ "loss": 4.335417747497559,
+ "step": 1566
+ },
+ {
+ "epoch": 21.768558951965066,
+ "grad_norm": 0.014646518044173717,
+ "learning_rate": 0.0006,
+ "loss": 4.279081344604492,
+ "step": 1567
+ },
+ {
+ "epoch": 21.782532751091704,
+ "grad_norm": 0.014132864773273468,
+ "learning_rate": 0.0006,
+ "loss": 4.345950126647949,
+ "step": 1568
+ },
+ {
+ "epoch": 21.79650655021834,
+ "grad_norm": 0.01425724383443594,
+ "learning_rate": 0.0006,
+ "loss": 4.277588844299316,
+ "step": 1569
+ },
+ {
+ "epoch": 21.810480349344978,
+ "grad_norm": 0.015396572649478912,
+ "learning_rate": 0.0006,
+ "loss": 4.326196670532227,
+ "step": 1570
+ },
+ {
+ "epoch": 21.824454148471617,
+ "grad_norm": 0.01683669537305832,
+ "learning_rate": 0.0006,
+ "loss": 4.3294172286987305,
+ "step": 1571
+ },
+ {
+ "epoch": 21.83842794759825,
+ "grad_norm": 0.016365647315979004,
+ "learning_rate": 0.0006,
+ "loss": 4.523321628570557,
+ "step": 1572
+ },
+ {
+ "epoch": 21.85240174672489,
+ "grad_norm": 0.017142774537205696,
+ "learning_rate": 0.0006,
+ "loss": 4.475842475891113,
+ "step": 1573
+ },
+ {
+ "epoch": 21.86637554585153,
+ "grad_norm": 0.015950839966535568,
+ "learning_rate": 0.0006,
+ "loss": 4.397187232971191,
+ "step": 1574
+ },
+ {
+ "epoch": 21.880349344978168,
+ "grad_norm": 0.016180362552404404,
+ "learning_rate": 0.0006,
+ "loss": 4.414778709411621,
+ "step": 1575
+ },
+ {
+ "epoch": 21.894323144104803,
+ "grad_norm": 0.017615556716918945,
+ "learning_rate": 0.0006,
+ "loss": 4.348567962646484,
+ "step": 1576
+ },
+ {
+ "epoch": 21.90829694323144,
+ "grad_norm": 0.01896277628839016,
+ "learning_rate": 0.0006,
+ "loss": 4.431762218475342,
+ "step": 1577
+ },
+ {
+ "epoch": 21.92227074235808,
+ "grad_norm": 0.01990971714258194,
+ "learning_rate": 0.0006,
+ "loss": 4.431285858154297,
+ "step": 1578
+ },
+ {
+ "epoch": 21.936244541484715,
+ "grad_norm": 0.0201258547604084,
+ "learning_rate": 0.0006,
+ "loss": 4.309872627258301,
+ "step": 1579
+ },
+ {
+ "epoch": 21.950218340611354,
+ "grad_norm": 0.019620198756456375,
+ "learning_rate": 0.0006,
+ "loss": 4.484044551849365,
+ "step": 1580
+ },
+ {
+ "epoch": 21.964192139737992,
+ "grad_norm": 0.01771123707294464,
+ "learning_rate": 0.0006,
+ "loss": 4.368093013763428,
+ "step": 1581
+ },
+ {
+ "epoch": 21.978165938864628,
+ "grad_norm": 0.017501654103398323,
+ "learning_rate": 0.0006,
+ "loss": 4.372651100158691,
+ "step": 1582
+ },
+ {
+ "epoch": 21.992139737991266,
+ "grad_norm": 0.01693608984351158,
+ "learning_rate": 0.0006,
+ "loss": 4.379167556762695,
+ "step": 1583
+ },
+ {
+ "epoch": 22.0,
+ "grad_norm": 0.02002849616110325,
+ "learning_rate": 0.0006,
+ "loss": 4.329627990722656,
+ "step": 1584
+ },
+ {
+ "epoch": 22.0,
+ "eval_loss": 4.723696231842041,
+ "eval_runtime": 56.1758,
+ "eval_samples_per_second": 43.471,
+ "eval_steps_per_second": 1.371,
+ "step": 1584
+ },
+ {
+ "epoch": 22.01397379912664,
+ "grad_norm": 0.02030305378139019,
+ "learning_rate": 0.0006,
+ "loss": 4.355566024780273,
+ "step": 1585
+ },
+ {
+ "epoch": 22.027947598253274,
+ "grad_norm": 0.018638934940099716,
+ "learning_rate": 0.0006,
+ "loss": 4.314009189605713,
+ "step": 1586
+ },
+ {
+ "epoch": 22.041921397379912,
+ "grad_norm": 0.019917072728276253,
+ "learning_rate": 0.0006,
+ "loss": 4.367525577545166,
+ "step": 1587
+ },
+ {
+ "epoch": 22.05589519650655,
+ "grad_norm": 0.020785167813301086,
+ "learning_rate": 0.0006,
+ "loss": 4.375561714172363,
+ "step": 1588
+ },
+ {
+ "epoch": 22.069868995633186,
+ "grad_norm": 0.021214453503489494,
+ "learning_rate": 0.0006,
+ "loss": 4.418305397033691,
+ "step": 1589
+ },
+ {
+ "epoch": 22.083842794759825,
+ "grad_norm": 0.019051678478717804,
+ "learning_rate": 0.0006,
+ "loss": 4.215126991271973,
+ "step": 1590
+ },
+ {
+ "epoch": 22.097816593886463,
+ "grad_norm": 0.02323366142809391,
+ "learning_rate": 0.0006,
+ "loss": 4.306807994842529,
+ "step": 1591
+ },
+ {
+ "epoch": 22.111790393013102,
+ "grad_norm": 0.026186738163232803,
+ "learning_rate": 0.0006,
+ "loss": 4.408508777618408,
+ "step": 1592
+ },
+ {
+ "epoch": 22.125764192139737,
+ "grad_norm": 0.027361227199435234,
+ "learning_rate": 0.0006,
+ "loss": 4.329322814941406,
+ "step": 1593
+ },
+ {
+ "epoch": 22.139737991266376,
+ "grad_norm": 0.026564646512269974,
+ "learning_rate": 0.0006,
+ "loss": 4.346642017364502,
+ "step": 1594
+ },
+ {
+ "epoch": 22.153711790393015,
+ "grad_norm": 0.025073667988181114,
+ "learning_rate": 0.0006,
+ "loss": 4.3984198570251465,
+ "step": 1595
+ },
+ {
+ "epoch": 22.16768558951965,
+ "grad_norm": 0.02347794733941555,
+ "learning_rate": 0.0006,
+ "loss": 4.352538108825684,
+ "step": 1596
+ },
+ {
+ "epoch": 22.18165938864629,
+ "grad_norm": 0.02203468047082424,
+ "learning_rate": 0.0006,
+ "loss": 4.298148155212402,
+ "step": 1597
+ },
+ {
+ "epoch": 22.195633187772927,
+ "grad_norm": 0.020621957257390022,
+ "learning_rate": 0.0006,
+ "loss": 4.480309963226318,
+ "step": 1598
+ },
+ {
+ "epoch": 22.209606986899562,
+ "grad_norm": 0.023710055276751518,
+ "learning_rate": 0.0006,
+ "loss": 4.339317321777344,
+ "step": 1599
+ },
+ {
+ "epoch": 22.2235807860262,
+ "grad_norm": 0.0233482476323843,
+ "learning_rate": 0.0006,
+ "loss": 4.436896324157715,
+ "step": 1600
+ },
+ {
+ "epoch": 22.23755458515284,
+ "grad_norm": 0.020122403278946877,
+ "learning_rate": 0.0006,
+ "loss": 4.236002445220947,
+ "step": 1601
+ },
+ {
+ "epoch": 22.251528384279474,
+ "grad_norm": 0.01957266591489315,
+ "learning_rate": 0.0006,
+ "loss": 4.497328281402588,
+ "step": 1602
+ },
+ {
+ "epoch": 22.265502183406113,
+ "grad_norm": 0.018964895978569984,
+ "learning_rate": 0.0006,
+ "loss": 4.36360502243042,
+ "step": 1603
+ },
+ {
+ "epoch": 22.27947598253275,
+ "grad_norm": 0.018902769312262535,
+ "learning_rate": 0.0006,
+ "loss": 4.427093505859375,
+ "step": 1604
+ },
+ {
+ "epoch": 22.29344978165939,
+ "grad_norm": 0.018918119370937347,
+ "learning_rate": 0.0006,
+ "loss": 4.294950485229492,
+ "step": 1605
+ },
+ {
+ "epoch": 22.307423580786025,
+ "grad_norm": 0.019315775483846664,
+ "learning_rate": 0.0006,
+ "loss": 4.427952289581299,
+ "step": 1606
+ },
+ {
+ "epoch": 22.321397379912664,
+ "grad_norm": 0.018350522965192795,
+ "learning_rate": 0.0006,
+ "loss": 4.272945404052734,
+ "step": 1607
+ },
+ {
+ "epoch": 22.335371179039303,
+ "grad_norm": 0.01759534701704979,
+ "learning_rate": 0.0006,
+ "loss": 4.348023891448975,
+ "step": 1608
+ },
+ {
+ "epoch": 22.349344978165938,
+ "grad_norm": 0.01646057888865471,
+ "learning_rate": 0.0006,
+ "loss": 4.278518199920654,
+ "step": 1609
+ },
+ {
+ "epoch": 22.363318777292577,
+ "grad_norm": 0.015875060111284256,
+ "learning_rate": 0.0006,
+ "loss": 4.423527240753174,
+ "step": 1610
+ },
+ {
+ "epoch": 22.377292576419215,
+ "grad_norm": 0.01638529635965824,
+ "learning_rate": 0.0006,
+ "loss": 4.321723937988281,
+ "step": 1611
+ },
+ {
+ "epoch": 22.39126637554585,
+ "grad_norm": 0.016942061483860016,
+ "learning_rate": 0.0006,
+ "loss": 4.346164703369141,
+ "step": 1612
+ },
+ {
+ "epoch": 22.40524017467249,
+ "grad_norm": 0.015209621749818325,
+ "learning_rate": 0.0006,
+ "loss": 4.434340000152588,
+ "step": 1613
+ },
+ {
+ "epoch": 22.419213973799128,
+ "grad_norm": 0.014423009008169174,
+ "learning_rate": 0.0006,
+ "loss": 4.342560768127441,
+ "step": 1614
+ },
+ {
+ "epoch": 22.433187772925763,
+ "grad_norm": 0.014369525946676731,
+ "learning_rate": 0.0006,
+ "loss": 4.296586990356445,
+ "step": 1615
+ },
+ {
+ "epoch": 22.4471615720524,
+ "grad_norm": 0.01392540242522955,
+ "learning_rate": 0.0006,
+ "loss": 4.433176040649414,
+ "step": 1616
+ },
+ {
+ "epoch": 22.46113537117904,
+ "grad_norm": 0.013206899166107178,
+ "learning_rate": 0.0006,
+ "loss": 4.32962703704834,
+ "step": 1617
+ },
+ {
+ "epoch": 22.475109170305675,
+ "grad_norm": 0.013427077792584896,
+ "learning_rate": 0.0006,
+ "loss": 4.420437335968018,
+ "step": 1618
+ },
+ {
+ "epoch": 22.489082969432314,
+ "grad_norm": 0.014194196090102196,
+ "learning_rate": 0.0006,
+ "loss": 4.410478591918945,
+ "step": 1619
+ },
+ {
+ "epoch": 22.503056768558952,
+ "grad_norm": 0.01362364087253809,
+ "learning_rate": 0.0006,
+ "loss": 4.338932037353516,
+ "step": 1620
+ },
+ {
+ "epoch": 22.51703056768559,
+ "grad_norm": 0.014299589209258556,
+ "learning_rate": 0.0006,
+ "loss": 4.34089994430542,
+ "step": 1621
+ },
+ {
+ "epoch": 22.531004366812226,
+ "grad_norm": 0.014061295427381992,
+ "learning_rate": 0.0006,
+ "loss": 4.29201602935791,
+ "step": 1622
+ },
+ {
+ "epoch": 22.544978165938865,
+ "grad_norm": 0.013945229351520538,
+ "learning_rate": 0.0006,
+ "loss": 4.326052188873291,
+ "step": 1623
+ },
+ {
+ "epoch": 22.558951965065503,
+ "grad_norm": 0.01387068536132574,
+ "learning_rate": 0.0006,
+ "loss": 4.393521308898926,
+ "step": 1624
+ },
+ {
+ "epoch": 22.57292576419214,
+ "grad_norm": 0.013931898400187492,
+ "learning_rate": 0.0006,
+ "loss": 4.3739752769470215,
+ "step": 1625
+ },
+ {
+ "epoch": 22.586899563318777,
+ "grad_norm": 0.01497740764170885,
+ "learning_rate": 0.0006,
+ "loss": 4.3991618156433105,
+ "step": 1626
+ },
+ {
+ "epoch": 22.600873362445416,
+ "grad_norm": 0.015842905268073082,
+ "learning_rate": 0.0006,
+ "loss": 4.405340671539307,
+ "step": 1627
+ },
+ {
+ "epoch": 22.61484716157205,
+ "grad_norm": 0.01580633595585823,
+ "learning_rate": 0.0006,
+ "loss": 4.280102729797363,
+ "step": 1628
+ },
+ {
+ "epoch": 22.62882096069869,
+ "grad_norm": 0.016567382961511612,
+ "learning_rate": 0.0006,
+ "loss": 4.407657146453857,
+ "step": 1629
+ },
+ {
+ "epoch": 22.64279475982533,
+ "grad_norm": 0.01648622378706932,
+ "learning_rate": 0.0006,
+ "loss": 4.392590522766113,
+ "step": 1630
+ },
+ {
+ "epoch": 22.656768558951963,
+ "grad_norm": 0.017066307365894318,
+ "learning_rate": 0.0006,
+ "loss": 4.324862003326416,
+ "step": 1631
+ },
+ {
+ "epoch": 22.670742358078602,
+ "grad_norm": 0.01637062057852745,
+ "learning_rate": 0.0006,
+ "loss": 4.34598445892334,
+ "step": 1632
+ },
+ {
+ "epoch": 22.68471615720524,
+ "grad_norm": 0.018038960173726082,
+ "learning_rate": 0.0006,
+ "loss": 4.320711612701416,
+ "step": 1633
+ },
+ {
+ "epoch": 22.69868995633188,
+ "grad_norm": 0.018096428364515305,
+ "learning_rate": 0.0006,
+ "loss": 4.536527156829834,
+ "step": 1634
+ },
+ {
+ "epoch": 22.712663755458514,
+ "grad_norm": 0.017889291048049927,
+ "learning_rate": 0.0006,
+ "loss": 4.388131141662598,
+ "step": 1635
+ },
+ {
+ "epoch": 22.726637554585153,
+ "grad_norm": 0.018771521747112274,
+ "learning_rate": 0.0006,
+ "loss": 4.357668876647949,
+ "step": 1636
+ },
+ {
+ "epoch": 22.74061135371179,
+ "grad_norm": 0.02100873365998268,
+ "learning_rate": 0.0006,
+ "loss": 4.293972015380859,
+ "step": 1637
+ },
+ {
+ "epoch": 22.754585152838427,
+ "grad_norm": 0.017715785652399063,
+ "learning_rate": 0.0006,
+ "loss": 4.349849224090576,
+ "step": 1638
+ },
+ {
+ "epoch": 22.768558951965066,
+ "grad_norm": 0.015546333976089954,
+ "learning_rate": 0.0006,
+ "loss": 4.311728000640869,
+ "step": 1639
+ },
+ {
+ "epoch": 22.782532751091704,
+ "grad_norm": 0.01605990156531334,
+ "learning_rate": 0.0006,
+ "loss": 4.499123573303223,
+ "step": 1640
+ },
+ {
+ "epoch": 22.79650655021834,
+ "grad_norm": 0.016642142087221146,
+ "learning_rate": 0.0006,
+ "loss": 4.4319257736206055,
+ "step": 1641
+ },
+ {
+ "epoch": 22.810480349344978,
+ "grad_norm": 0.016120387241244316,
+ "learning_rate": 0.0006,
+ "loss": 4.4373650550842285,
+ "step": 1642
+ },
+ {
+ "epoch": 22.824454148471617,
+ "grad_norm": 0.015174890868365765,
+ "learning_rate": 0.0006,
+ "loss": 4.314082145690918,
+ "step": 1643
+ },
+ {
+ "epoch": 22.83842794759825,
+ "grad_norm": 0.01551714539527893,
+ "learning_rate": 0.0006,
+ "loss": 4.360363960266113,
+ "step": 1644
+ },
+ {
+ "epoch": 22.85240174672489,
+ "grad_norm": 0.014966742135584354,
+ "learning_rate": 0.0006,
+ "loss": 4.346253395080566,
+ "step": 1645
+ },
+ {
+ "epoch": 22.86637554585153,
+ "grad_norm": 0.014288028702139854,
+ "learning_rate": 0.0006,
+ "loss": 4.464366436004639,
+ "step": 1646
+ },
+ {
+ "epoch": 22.880349344978168,
+ "grad_norm": 0.015291682444512844,
+ "learning_rate": 0.0006,
+ "loss": 4.504101753234863,
+ "step": 1647
+ },
+ {
+ "epoch": 22.894323144104803,
+ "grad_norm": 0.014534426853060722,
+ "learning_rate": 0.0006,
+ "loss": 4.462172031402588,
+ "step": 1648
+ },
+ {
+ "epoch": 22.90829694323144,
+ "grad_norm": 0.014129819348454475,
+ "learning_rate": 0.0006,
+ "loss": 4.350126266479492,
+ "step": 1649
+ },
+ {
+ "epoch": 22.92227074235808,
+ "grad_norm": 0.014932668767869473,
+ "learning_rate": 0.0006,
+ "loss": 4.374392032623291,
+ "step": 1650
+ },
+ {
+ "epoch": 22.936244541484715,
+ "grad_norm": 0.014647853560745716,
+ "learning_rate": 0.0006,
+ "loss": 4.252264499664307,
+ "step": 1651
+ },
+ {
+ "epoch": 22.950218340611354,
+ "grad_norm": 0.013553360477089882,
+ "learning_rate": 0.0006,
+ "loss": 4.379729270935059,
+ "step": 1652
+ },
+ {
+ "epoch": 22.964192139737992,
+ "grad_norm": 0.014606194570660591,
+ "learning_rate": 0.0006,
+ "loss": 4.363372802734375,
+ "step": 1653
+ },
+ {
+ "epoch": 22.978165938864628,
+ "grad_norm": 0.015000631101429462,
+ "learning_rate": 0.0006,
+ "loss": 4.408048629760742,
+ "step": 1654
+ },
+ {
+ "epoch": 22.992139737991266,
+ "grad_norm": 0.015370816923677921,
+ "learning_rate": 0.0006,
+ "loss": 4.297451019287109,
+ "step": 1655
+ },
+ {
+ "epoch": 23.0,
+ "grad_norm": 0.01709144562482834,
+ "learning_rate": 0.0006,
+ "loss": 4.294776916503906,
+ "step": 1656
+ },
+ {
+ "epoch": 23.0,
+ "eval_loss": 4.678618907928467,
+ "eval_runtime": 56.1668,
+ "eval_samples_per_second": 43.478,
+ "eval_steps_per_second": 1.371,
+ "step": 1656
+ },
+ {
+ "epoch": 23.01397379912664,
+ "grad_norm": 0.01727290451526642,
+ "learning_rate": 0.0006,
+ "loss": 4.311701774597168,
+ "step": 1657
+ },
+ {
+ "epoch": 23.027947598253274,
+ "grad_norm": 0.020645759999752045,
+ "learning_rate": 0.0006,
+ "loss": 4.429897308349609,
+ "step": 1658
+ },
+ {
+ "epoch": 23.041921397379912,
+ "grad_norm": 0.02204226516187191,
+ "learning_rate": 0.0006,
+ "loss": 4.365151882171631,
+ "step": 1659
+ },
+ {
+ "epoch": 23.05589519650655,
+ "grad_norm": 0.023069269955158234,
+ "learning_rate": 0.0006,
+ "loss": 4.3710832595825195,
+ "step": 1660
+ },
+ {
+ "epoch": 23.069868995633186,
+ "grad_norm": 0.024151179939508438,
+ "learning_rate": 0.0006,
+ "loss": 4.403108596801758,
+ "step": 1661
+ },
+ {
+ "epoch": 23.083842794759825,
+ "grad_norm": 0.02511233650147915,
+ "learning_rate": 0.0006,
+ "loss": 4.414128303527832,
+ "step": 1662
+ },
+ {
+ "epoch": 23.097816593886463,
+ "grad_norm": 0.025802113115787506,
+ "learning_rate": 0.0006,
+ "loss": 4.388265609741211,
+ "step": 1663
+ },
+ {
+ "epoch": 23.111790393013102,
+ "grad_norm": 0.02377348765730858,
+ "learning_rate": 0.0006,
+ "loss": 4.357505798339844,
+ "step": 1664
+ },
+ {
+ "epoch": 23.125764192139737,
+ "grad_norm": 0.020876651629805565,
+ "learning_rate": 0.0006,
+ "loss": 4.23043966293335,
+ "step": 1665
+ },
+ {
+ "epoch": 23.139737991266376,
+ "grad_norm": 0.01959727331995964,
+ "learning_rate": 0.0006,
+ "loss": 4.280970096588135,
+ "step": 1666
+ },
+ {
+ "epoch": 23.153711790393015,
+ "grad_norm": 0.018678616732358932,
+ "learning_rate": 0.0006,
+ "loss": 4.346716403961182,
+ "step": 1667
+ },
+ {
+ "epoch": 23.16768558951965,
+ "grad_norm": 0.02108944021165371,
+ "learning_rate": 0.0006,
+ "loss": 4.277904510498047,
+ "step": 1668
+ },
+ {
+ "epoch": 23.18165938864629,
+ "grad_norm": 0.020221566781401634,
+ "learning_rate": 0.0006,
+ "loss": 4.367693901062012,
+ "step": 1669
+ },
+ {
+ "epoch": 23.195633187772927,
+ "grad_norm": 0.019599711522459984,
+ "learning_rate": 0.0006,
+ "loss": 4.257092475891113,
+ "step": 1670
+ },
+ {
+ "epoch": 23.209606986899562,
+ "grad_norm": 0.019833361729979515,
+ "learning_rate": 0.0006,
+ "loss": 4.540737628936768,
+ "step": 1671
+ },
+ {
+ "epoch": 23.2235807860262,
+ "grad_norm": 0.020352095365524292,
+ "learning_rate": 0.0006,
+ "loss": 4.360923767089844,
+ "step": 1672
+ },
+ {
+ "epoch": 23.23755458515284,
+ "grad_norm": 0.01939748041331768,
+ "learning_rate": 0.0006,
+ "loss": 4.393110275268555,
+ "step": 1673
+ },
+ {
+ "epoch": 23.251528384279474,
+ "grad_norm": 0.01884354278445244,
+ "learning_rate": 0.0006,
+ "loss": 4.383938789367676,
+ "step": 1674
+ },
+ {
+ "epoch": 23.265502183406113,
+ "grad_norm": 0.02008470520377159,
+ "learning_rate": 0.0006,
+ "loss": 4.346344947814941,
+ "step": 1675
+ },
+ {
+ "epoch": 23.27947598253275,
+ "grad_norm": 0.019065195694565773,
+ "learning_rate": 0.0006,
+ "loss": 4.36611795425415,
+ "step": 1676
+ },
+ {
+ "epoch": 23.29344978165939,
+ "grad_norm": 0.01938541606068611,
+ "learning_rate": 0.0006,
+ "loss": 4.4009599685668945,
+ "step": 1677
+ },
+ {
+ "epoch": 23.307423580786025,
+ "grad_norm": 0.019044723361730576,
+ "learning_rate": 0.0006,
+ "loss": 4.340169906616211,
+ "step": 1678
+ },
+ {
+ "epoch": 23.321397379912664,
+ "grad_norm": 0.018850378692150116,
+ "learning_rate": 0.0006,
+ "loss": 4.37674617767334,
+ "step": 1679
+ },
+ {
+ "epoch": 23.335371179039303,
+ "grad_norm": 0.019076095893979073,
+ "learning_rate": 0.0006,
+ "loss": 4.226874351501465,
+ "step": 1680
+ },
+ {
+ "epoch": 23.349344978165938,
+ "grad_norm": 0.017543038353323936,
+ "learning_rate": 0.0006,
+ "loss": 4.337751388549805,
+ "step": 1681
+ },
+ {
+ "epoch": 23.363318777292577,
+ "grad_norm": 0.017170218750834465,
+ "learning_rate": 0.0006,
+ "loss": 4.282512664794922,
+ "step": 1682
+ },
+ {
+ "epoch": 23.377292576419215,
+ "grad_norm": 0.01574796624481678,
+ "learning_rate": 0.0006,
+ "loss": 4.30638313293457,
+ "step": 1683
+ },
+ {
+ "epoch": 23.39126637554585,
+ "grad_norm": 0.015160846523940563,
+ "learning_rate": 0.0006,
+ "loss": 4.336716175079346,
+ "step": 1684
+ },
+ {
+ "epoch": 23.40524017467249,
+ "grad_norm": 0.01534526702016592,
+ "learning_rate": 0.0006,
+ "loss": 4.2174787521362305,
+ "step": 1685
+ },
+ {
+ "epoch": 23.419213973799128,
+ "grad_norm": 0.015289999544620514,
+ "learning_rate": 0.0006,
+ "loss": 4.272546291351318,
+ "step": 1686
+ },
+ {
+ "epoch": 23.433187772925763,
+ "grad_norm": 0.014894185587763786,
+ "learning_rate": 0.0006,
+ "loss": 4.296045303344727,
+ "step": 1687
+ },
+ {
+ "epoch": 23.4471615720524,
+ "grad_norm": 0.014869502745568752,
+ "learning_rate": 0.0006,
+ "loss": 4.425461769104004,
+ "step": 1688
+ },
+ {
+ "epoch": 23.46113537117904,
+ "grad_norm": 0.01474483497440815,
+ "learning_rate": 0.0006,
+ "loss": 4.361867427825928,
+ "step": 1689
+ },
+ {
+ "epoch": 23.475109170305675,
+ "grad_norm": 0.015299245715141296,
+ "learning_rate": 0.0006,
+ "loss": 4.364523410797119,
+ "step": 1690
+ },
+ {
+ "epoch": 23.489082969432314,
+ "grad_norm": 0.015201340429484844,
+ "learning_rate": 0.0006,
+ "loss": 4.285182952880859,
+ "step": 1691
+ },
+ {
+ "epoch": 23.503056768558952,
+ "grad_norm": 0.016057293862104416,
+ "learning_rate": 0.0006,
+ "loss": 4.284873962402344,
+ "step": 1692
+ },
+ {
+ "epoch": 23.51703056768559,
+ "grad_norm": 0.015871062874794006,
+ "learning_rate": 0.0006,
+ "loss": 4.326470851898193,
+ "step": 1693
+ },
+ {
+ "epoch": 23.531004366812226,
+ "grad_norm": 0.01586288772523403,
+ "learning_rate": 0.0006,
+ "loss": 4.324883460998535,
+ "step": 1694
+ },
+ {
+ "epoch": 23.544978165938865,
+ "grad_norm": 0.016092827543616295,
+ "learning_rate": 0.0006,
+ "loss": 4.3258867263793945,
+ "step": 1695
+ },
+ {
+ "epoch": 23.558951965065503,
+ "grad_norm": 0.01582016795873642,
+ "learning_rate": 0.0006,
+ "loss": 4.358892440795898,
+ "step": 1696
+ },
+ {
+ "epoch": 23.57292576419214,
+ "grad_norm": 0.01688452623784542,
+ "learning_rate": 0.0006,
+ "loss": 4.280842304229736,
+ "step": 1697
+ },
+ {
+ "epoch": 23.586899563318777,
+ "grad_norm": 0.015865886583924294,
+ "learning_rate": 0.0006,
+ "loss": 4.245567321777344,
+ "step": 1698
+ },
+ {
+ "epoch": 23.600873362445416,
+ "grad_norm": 0.015303662046790123,
+ "learning_rate": 0.0006,
+ "loss": 4.383112907409668,
+ "step": 1699
+ },
+ {
+ "epoch": 23.61484716157205,
+ "grad_norm": 0.015649516135454178,
+ "learning_rate": 0.0006,
+ "loss": 4.519082069396973,
+ "step": 1700
+ },
+ {
+ "epoch": 23.62882096069869,
+ "grad_norm": 0.0167120099067688,
+ "learning_rate": 0.0006,
+ "loss": 4.447847366333008,
+ "step": 1701
+ },
+ {
+ "epoch": 23.64279475982533,
+ "grad_norm": 0.016368716955184937,
+ "learning_rate": 0.0006,
+ "loss": 4.434186935424805,
+ "step": 1702
+ },
+ {
+ "epoch": 23.656768558951963,
+ "grad_norm": 0.017379503697156906,
+ "learning_rate": 0.0006,
+ "loss": 4.317594528198242,
+ "step": 1703
+ },
+ {
+ "epoch": 23.670742358078602,
+ "grad_norm": 0.018573811277747154,
+ "learning_rate": 0.0006,
+ "loss": 4.205793857574463,
+ "step": 1704
+ },
+ {
+ "epoch": 23.68471615720524,
+ "grad_norm": 0.01749090477824211,
+ "learning_rate": 0.0006,
+ "loss": 4.3985700607299805,
+ "step": 1705
+ },
+ {
+ "epoch": 23.69868995633188,
+ "grad_norm": 0.0165668074041605,
+ "learning_rate": 0.0006,
+ "loss": 4.329226970672607,
+ "step": 1706
+ },
+ {
+ "epoch": 23.712663755458514,
+ "grad_norm": 0.016913846135139465,
+ "learning_rate": 0.0006,
+ "loss": 4.388485908508301,
+ "step": 1707
+ },
+ {
+ "epoch": 23.726637554585153,
+ "grad_norm": 0.016813859343528748,
+ "learning_rate": 0.0006,
+ "loss": 4.3395915031433105,
+ "step": 1708
+ },
+ {
+ "epoch": 23.74061135371179,
+ "grad_norm": 0.01676975190639496,
+ "learning_rate": 0.0006,
+ "loss": 4.254745960235596,
+ "step": 1709
+ },
+ {
+ "epoch": 23.754585152838427,
+ "grad_norm": 0.016370350494980812,
+ "learning_rate": 0.0006,
+ "loss": 4.2612223625183105,
+ "step": 1710
+ },
+ {
+ "epoch": 23.768558951965066,
+ "grad_norm": 0.01696198247373104,
+ "learning_rate": 0.0006,
+ "loss": 4.327524662017822,
+ "step": 1711
+ },
+ {
+ "epoch": 23.782532751091704,
+ "grad_norm": 0.0189108457416296,
+ "learning_rate": 0.0006,
+ "loss": 4.325839519500732,
+ "step": 1712
+ },
+ {
+ "epoch": 23.79650655021834,
+ "grad_norm": 0.01862351782619953,
+ "learning_rate": 0.0006,
+ "loss": 4.333662509918213,
+ "step": 1713
+ },
+ {
+ "epoch": 23.810480349344978,
+ "grad_norm": 0.01655014418065548,
+ "learning_rate": 0.0006,
+ "loss": 4.39305305480957,
+ "step": 1714
+ },
+ {
+ "epoch": 23.824454148471617,
+ "grad_norm": 0.015899376943707466,
+ "learning_rate": 0.0006,
+ "loss": 4.390053749084473,
+ "step": 1715
+ },
+ {
+ "epoch": 23.83842794759825,
+ "grad_norm": 0.016519056633114815,
+ "learning_rate": 0.0006,
+ "loss": 4.255903244018555,
+ "step": 1716
+ },
+ {
+ "epoch": 23.85240174672489,
+ "grad_norm": 0.0167539119720459,
+ "learning_rate": 0.0006,
+ "loss": 4.255727767944336,
+ "step": 1717
+ },
+ {
+ "epoch": 23.86637554585153,
+ "grad_norm": 0.0171353816986084,
+ "learning_rate": 0.0006,
+ "loss": 4.376740455627441,
+ "step": 1718
+ },
+ {
+ "epoch": 23.880349344978168,
+ "grad_norm": 0.016719117760658264,
+ "learning_rate": 0.0006,
+ "loss": 4.3376688957214355,
+ "step": 1719
+ },
+ {
+ "epoch": 23.894323144104803,
+ "grad_norm": 0.01561494916677475,
+ "learning_rate": 0.0006,
+ "loss": 4.3345746994018555,
+ "step": 1720
+ },
+ {
+ "epoch": 23.90829694323144,
+ "grad_norm": 0.016303174197673798,
+ "learning_rate": 0.0006,
+ "loss": 4.332371234893799,
+ "step": 1721
+ },
+ {
+ "epoch": 23.92227074235808,
+ "grad_norm": 0.016722865402698517,
+ "learning_rate": 0.0006,
+ "loss": 4.267125129699707,
+ "step": 1722
+ },
+ {
+ "epoch": 23.936244541484715,
+ "grad_norm": 0.017072124406695366,
+ "learning_rate": 0.0006,
+ "loss": 4.322751998901367,
+ "step": 1723
+ },
+ {
+ "epoch": 23.950218340611354,
+ "grad_norm": 0.01699932850897312,
+ "learning_rate": 0.0006,
+ "loss": 4.378422737121582,
+ "step": 1724
+ },
+ {
+ "epoch": 23.964192139737992,
+ "grad_norm": 0.01742720976471901,
+ "learning_rate": 0.0006,
+ "loss": 4.345083236694336,
+ "step": 1725
+ },
+ {
+ "epoch": 23.978165938864628,
+ "grad_norm": 0.018797501921653748,
+ "learning_rate": 0.0006,
+ "loss": 4.290443420410156,
+ "step": 1726
+ },
+ {
+ "epoch": 23.992139737991266,
+ "grad_norm": 0.018090683966875076,
+ "learning_rate": 0.0006,
+ "loss": 4.377580642700195,
+ "step": 1727
+ },
+ {
+ "epoch": 24.0,
+ "grad_norm": 0.017305459827184677,
+ "learning_rate": 0.0006,
+ "loss": 4.364426612854004,
+ "step": 1728
+ },
+ {
+ "epoch": 24.0,
+ "eval_loss": 4.647762298583984,
+ "eval_runtime": 56.5402,
+ "eval_samples_per_second": 43.191,
+ "eval_steps_per_second": 1.362,
+ "step": 1728
+ },
+ {
+ "epoch": 24.01397379912664,
+ "grad_norm": 0.016493607312440872,
+ "learning_rate": 0.0006,
+ "loss": 4.252255916595459,
+ "step": 1729
+ },
+ {
+ "epoch": 24.027947598253274,
+ "grad_norm": 0.016529450193047523,
+ "learning_rate": 0.0006,
+ "loss": 4.352746963500977,
+ "step": 1730
+ },
+ {
+ "epoch": 24.041921397379912,
+ "grad_norm": 0.016940191388130188,
+ "learning_rate": 0.0006,
+ "loss": 4.301870822906494,
+ "step": 1731
+ },
+ {
+ "epoch": 24.05589519650655,
+ "grad_norm": 0.01752558909356594,
+ "learning_rate": 0.0006,
+ "loss": 4.2762451171875,
+ "step": 1732
+ },
+ {
+ "epoch": 24.069868995633186,
+ "grad_norm": 0.017112884670495987,
+ "learning_rate": 0.0006,
+ "loss": 4.2659149169921875,
+ "step": 1733
+ },
+ {
+ "epoch": 24.083842794759825,
+ "grad_norm": 0.01900586113333702,
+ "learning_rate": 0.0006,
+ "loss": 4.306058883666992,
+ "step": 1734
+ },
+ {
+ "epoch": 24.097816593886463,
+ "grad_norm": 0.019447680562734604,
+ "learning_rate": 0.0006,
+ "loss": 4.450128555297852,
+ "step": 1735
+ },
+ {
+ "epoch": 24.111790393013102,
+ "grad_norm": 0.01948186755180359,
+ "learning_rate": 0.0006,
+ "loss": 4.202869415283203,
+ "step": 1736
+ },
+ {
+ "epoch": 24.125764192139737,
+ "grad_norm": 0.01716383546590805,
+ "learning_rate": 0.0006,
+ "loss": 4.318876266479492,
+ "step": 1737
+ },
+ {
+ "epoch": 24.139737991266376,
+ "grad_norm": 0.017049988731741905,
+ "learning_rate": 0.0006,
+ "loss": 4.237238883972168,
+ "step": 1738
+ },
+ {
+ "epoch": 24.153711790393015,
+ "grad_norm": 0.015169923193752766,
+ "learning_rate": 0.0006,
+ "loss": 4.289426803588867,
+ "step": 1739
+ },
+ {
+ "epoch": 24.16768558951965,
+ "grad_norm": 0.014760667458176613,
+ "learning_rate": 0.0006,
+ "loss": 4.356924057006836,
+ "step": 1740
+ },
+ {
+ "epoch": 24.18165938864629,
+ "grad_norm": 0.015127750113606453,
+ "learning_rate": 0.0006,
+ "loss": 4.142916679382324,
+ "step": 1741
+ },
+ {
+ "epoch": 24.195633187772927,
+ "grad_norm": 0.01549555640667677,
+ "learning_rate": 0.0006,
+ "loss": 4.345563888549805,
+ "step": 1742
+ },
+ {
+ "epoch": 24.209606986899562,
+ "grad_norm": 0.016393091529607773,
+ "learning_rate": 0.0006,
+ "loss": 4.417495250701904,
+ "step": 1743
+ },
+ {
+ "epoch": 24.2235807860262,
+ "grad_norm": 0.01785266026854515,
+ "learning_rate": 0.0006,
+ "loss": 4.3553547859191895,
+ "step": 1744
+ },
+ {
+ "epoch": 24.23755458515284,
+ "grad_norm": 0.017574617639183998,
+ "learning_rate": 0.0006,
+ "loss": 4.241024494171143,
+ "step": 1745
+ },
+ {
+ "epoch": 24.251528384279474,
+ "grad_norm": 0.019060306251049042,
+ "learning_rate": 0.0006,
+ "loss": 4.3685150146484375,
+ "step": 1746
+ },
+ {
+ "epoch": 24.265502183406113,
+ "grad_norm": 0.0214633010327816,
+ "learning_rate": 0.0006,
+ "loss": 4.323866844177246,
+ "step": 1747
+ },
+ {
+ "epoch": 24.27947598253275,
+ "grad_norm": 0.021055003628134727,
+ "learning_rate": 0.0006,
+ "loss": 4.278757095336914,
+ "step": 1748
+ },
+ {
+ "epoch": 24.29344978165939,
+ "grad_norm": 0.018618572503328323,
+ "learning_rate": 0.0006,
+ "loss": 4.28849983215332,
+ "step": 1749
+ },
+ {
+ "epoch": 24.307423580786025,
+ "grad_norm": 0.018666435033082962,
+ "learning_rate": 0.0006,
+ "loss": 4.368590354919434,
+ "step": 1750
+ },
+ {
+ "epoch": 24.321397379912664,
+ "grad_norm": 0.020626146346330643,
+ "learning_rate": 0.0006,
+ "loss": 4.2951836585998535,
+ "step": 1751
+ },
+ {
+ "epoch": 24.335371179039303,
+ "grad_norm": 0.01929759792983532,
+ "learning_rate": 0.0006,
+ "loss": 4.3664703369140625,
+ "step": 1752
+ },
+ {
+ "epoch": 24.349344978165938,
+ "grad_norm": 0.020956002175807953,
+ "learning_rate": 0.0006,
+ "loss": 4.345308303833008,
+ "step": 1753
+ },
+ {
+ "epoch": 24.363318777292577,
+ "grad_norm": 0.019227957352995872,
+ "learning_rate": 0.0006,
+ "loss": 4.349849224090576,
+ "step": 1754
+ },
+ {
+ "epoch": 24.377292576419215,
+ "grad_norm": 0.015970097854733467,
+ "learning_rate": 0.0006,
+ "loss": 4.223012447357178,
+ "step": 1755
+ },
+ {
+ "epoch": 24.39126637554585,
+ "grad_norm": 0.015951646491885185,
+ "learning_rate": 0.0006,
+ "loss": 4.236420631408691,
+ "step": 1756
+ },
+ {
+ "epoch": 24.40524017467249,
+ "grad_norm": 0.016032515093684196,
+ "learning_rate": 0.0006,
+ "loss": 4.226899147033691,
+ "step": 1757
+ },
+ {
+ "epoch": 24.419213973799128,
+ "grad_norm": 0.01573132909834385,
+ "learning_rate": 0.0006,
+ "loss": 4.315916061401367,
+ "step": 1758
+ },
+ {
+ "epoch": 24.433187772925763,
+ "grad_norm": 0.015367005951702595,
+ "learning_rate": 0.0006,
+ "loss": 4.358939170837402,
+ "step": 1759
+ },
+ {
+ "epoch": 24.4471615720524,
+ "grad_norm": 0.01669907011091709,
+ "learning_rate": 0.0006,
+ "loss": 4.232147216796875,
+ "step": 1760
+ },
+ {
+ "epoch": 24.46113537117904,
+ "grad_norm": 0.018708152696490288,
+ "learning_rate": 0.0006,
+ "loss": 4.329852104187012,
+ "step": 1761
+ },
+ {
+ "epoch": 24.475109170305675,
+ "grad_norm": 0.019035987555980682,
+ "learning_rate": 0.0006,
+ "loss": 4.260340213775635,
+ "step": 1762
+ },
+ {
+ "epoch": 24.489082969432314,
+ "grad_norm": 0.01832144521176815,
+ "learning_rate": 0.0006,
+ "loss": 4.359543800354004,
+ "step": 1763
+ },
+ {
+ "epoch": 24.503056768558952,
+ "grad_norm": 0.015313294716179371,
+ "learning_rate": 0.0006,
+ "loss": 4.279088020324707,
+ "step": 1764
+ },
+ {
+ "epoch": 24.51703056768559,
+ "grad_norm": 0.015803923830389977,
+ "learning_rate": 0.0006,
+ "loss": 4.230682849884033,
+ "step": 1765
+ },
+ {
+ "epoch": 24.531004366812226,
+ "grad_norm": 0.016748374328017235,
+ "learning_rate": 0.0006,
+ "loss": 4.349661827087402,
+ "step": 1766
+ },
+ {
+ "epoch": 24.544978165938865,
+ "grad_norm": 0.015732428058981895,
+ "learning_rate": 0.0006,
+ "loss": 4.372441291809082,
+ "step": 1767
+ },
+ {
+ "epoch": 24.558951965065503,
+ "grad_norm": 0.016564829275012016,
+ "learning_rate": 0.0006,
+ "loss": 4.323090553283691,
+ "step": 1768
+ },
+ {
+ "epoch": 24.57292576419214,
+ "grad_norm": 0.01603010483086109,
+ "learning_rate": 0.0006,
+ "loss": 4.347898006439209,
+ "step": 1769
+ },
+ {
+ "epoch": 24.586899563318777,
+ "grad_norm": 0.01674809865653515,
+ "learning_rate": 0.0006,
+ "loss": 4.269165515899658,
+ "step": 1770
+ },
+ {
+ "epoch": 24.600873362445416,
+ "grad_norm": 0.017333930358290672,
+ "learning_rate": 0.0006,
+ "loss": 4.353298187255859,
+ "step": 1771
+ },
+ {
+ "epoch": 24.61484716157205,
+ "grad_norm": 0.017519580200314522,
+ "learning_rate": 0.0006,
+ "loss": 4.265199661254883,
+ "step": 1772
+ },
+ {
+ "epoch": 24.62882096069869,
+ "grad_norm": 0.017675306648015976,
+ "learning_rate": 0.0006,
+ "loss": 4.322596549987793,
+ "step": 1773
+ },
+ {
+ "epoch": 24.64279475982533,
+ "grad_norm": 0.017166534438729286,
+ "learning_rate": 0.0006,
+ "loss": 4.413900375366211,
+ "step": 1774
+ },
+ {
+ "epoch": 24.656768558951963,
+ "grad_norm": 0.016362346708774567,
+ "learning_rate": 0.0006,
+ "loss": 4.307236671447754,
+ "step": 1775
+ },
+ {
+ "epoch": 24.670742358078602,
+ "grad_norm": 0.016086872667074203,
+ "learning_rate": 0.0006,
+ "loss": 4.235316276550293,
+ "step": 1776
+ },
+ {
+ "epoch": 24.68471615720524,
+ "grad_norm": 0.01689537614583969,
+ "learning_rate": 0.0006,
+ "loss": 4.390591621398926,
+ "step": 1777
+ },
+ {
+ "epoch": 24.69868995633188,
+ "grad_norm": 0.016836725175380707,
+ "learning_rate": 0.0006,
+ "loss": 4.2990031242370605,
+ "step": 1778
+ },
+ {
+ "epoch": 24.712663755458514,
+ "grad_norm": 0.015627335757017136,
+ "learning_rate": 0.0006,
+ "loss": 4.166745185852051,
+ "step": 1779
+ },
+ {
+ "epoch": 24.726637554585153,
+ "grad_norm": 0.015007016249001026,
+ "learning_rate": 0.0006,
+ "loss": 4.295273303985596,
+ "step": 1780
+ },
+ {
+ "epoch": 24.74061135371179,
+ "grad_norm": 0.016084209084510803,
+ "learning_rate": 0.0006,
+ "loss": 4.342424392700195,
+ "step": 1781
+ },
+ {
+ "epoch": 24.754585152838427,
+ "grad_norm": 0.01595073565840721,
+ "learning_rate": 0.0006,
+ "loss": 4.3060102462768555,
+ "step": 1782
+ },
+ {
+ "epoch": 24.768558951965066,
+ "grad_norm": 0.016113461926579475,
+ "learning_rate": 0.0006,
+ "loss": 4.267908096313477,
+ "step": 1783
+ },
+ {
+ "epoch": 24.782532751091704,
+ "grad_norm": 0.01564556173980236,
+ "learning_rate": 0.0006,
+ "loss": 4.324642658233643,
+ "step": 1784
+ },
+ {
+ "epoch": 24.79650655021834,
+ "grad_norm": 0.01575268618762493,
+ "learning_rate": 0.0006,
+ "loss": 4.328181743621826,
+ "step": 1785
+ },
+ {
+ "epoch": 24.810480349344978,
+ "grad_norm": 0.016830824315547943,
+ "learning_rate": 0.0006,
+ "loss": 4.247906684875488,
+ "step": 1786
+ },
+ {
+ "epoch": 24.824454148471617,
+ "grad_norm": 0.017989547923207283,
+ "learning_rate": 0.0006,
+ "loss": 4.228281021118164,
+ "step": 1787
+ },
+ {
+ "epoch": 24.83842794759825,
+ "grad_norm": 0.016783086583018303,
+ "learning_rate": 0.0006,
+ "loss": 4.2880964279174805,
+ "step": 1788
+ },
+ {
+ "epoch": 24.85240174672489,
+ "grad_norm": 0.016467561945319176,
+ "learning_rate": 0.0006,
+ "loss": 4.36118745803833,
+ "step": 1789
+ },
+ {
+ "epoch": 24.86637554585153,
+ "grad_norm": 0.01780523732304573,
+ "learning_rate": 0.0006,
+ "loss": 4.172321319580078,
+ "step": 1790
+ },
+ {
+ "epoch": 24.880349344978168,
+ "grad_norm": 0.017102031037211418,
+ "learning_rate": 0.0006,
+ "loss": 4.352826118469238,
+ "step": 1791
+ },
+ {
+ "epoch": 24.894323144104803,
+ "grad_norm": 0.016409218311309814,
+ "learning_rate": 0.0006,
+ "loss": 4.212162017822266,
+ "step": 1792
+ },
+ {
+ "epoch": 24.90829694323144,
+ "grad_norm": 0.018660584464669228,
+ "learning_rate": 0.0006,
+ "loss": 4.346684455871582,
+ "step": 1793
+ },
+ {
+ "epoch": 24.92227074235808,
+ "grad_norm": 0.017389440909028053,
+ "learning_rate": 0.0006,
+ "loss": 4.197913646697998,
+ "step": 1794
+ },
+ {
+ "epoch": 24.936244541484715,
+ "grad_norm": 0.01596485823392868,
+ "learning_rate": 0.0006,
+ "loss": 4.370617866516113,
+ "step": 1795
+ },
+ {
+ "epoch": 24.950218340611354,
+ "grad_norm": 0.01709776371717453,
+ "learning_rate": 0.0006,
+ "loss": 4.313847064971924,
+ "step": 1796
+ },
+ {
+ "epoch": 24.964192139737992,
+ "grad_norm": 0.016571134328842163,
+ "learning_rate": 0.0006,
+ "loss": 4.1895856857299805,
+ "step": 1797
+ },
+ {
+ "epoch": 24.978165938864628,
+ "grad_norm": 0.016851048916578293,
+ "learning_rate": 0.0006,
+ "loss": 4.350772857666016,
+ "step": 1798
+ },
+ {
+ "epoch": 24.992139737991266,
+ "grad_norm": 0.017007440328598022,
+ "learning_rate": 0.0006,
+ "loss": 4.399008750915527,
+ "step": 1799
+ },
+ {
+ "epoch": 25.0,
+ "grad_norm": 0.01791212148964405,
+ "learning_rate": 0.0006,
+ "loss": 4.306121826171875,
+ "step": 1800
+ },
+ {
+ "epoch": 25.0,
+ "eval_loss": 4.636417865753174,
+ "eval_runtime": 56.8296,
+ "eval_samples_per_second": 42.971,
+ "eval_steps_per_second": 1.355,
+ "step": 1800
+ },
+ {
+ "epoch": 25.01397379912664,
+ "grad_norm": 0.01729617640376091,
+ "learning_rate": 0.0006,
+ "loss": 4.257018089294434,
+ "step": 1801
+ },
+ {
+ "epoch": 25.027947598253274,
+ "grad_norm": 0.017466643825173378,
+ "learning_rate": 0.0006,
+ "loss": 4.307605266571045,
+ "step": 1802
+ },
+ {
+ "epoch": 25.041921397379912,
+ "grad_norm": 0.018614279106259346,
+ "learning_rate": 0.0006,
+ "loss": 4.360141277313232,
+ "step": 1803
+ },
+ {
+ "epoch": 25.05589519650655,
+ "grad_norm": 0.02010771445930004,
+ "learning_rate": 0.0006,
+ "loss": 4.196428298950195,
+ "step": 1804
+ },
+ {
+ "epoch": 25.069868995633186,
+ "grad_norm": 0.02090228535234928,
+ "learning_rate": 0.0006,
+ "loss": 4.288947105407715,
+ "step": 1805
+ },
+ {
+ "epoch": 25.083842794759825,
+ "grad_norm": 0.02050255797803402,
+ "learning_rate": 0.0006,
+ "loss": 4.259582042694092,
+ "step": 1806
+ },
+ {
+ "epoch": 25.097816593886463,
+ "grad_norm": 0.021844089031219482,
+ "learning_rate": 0.0006,
+ "loss": 4.405138969421387,
+ "step": 1807
+ },
+ {
+ "epoch": 25.111790393013102,
+ "grad_norm": 0.022287018597126007,
+ "learning_rate": 0.0006,
+ "loss": 4.333793640136719,
+ "step": 1808
+ },
+ {
+ "epoch": 25.125764192139737,
+ "grad_norm": 0.02288329415023327,
+ "learning_rate": 0.0006,
+ "loss": 4.3910980224609375,
+ "step": 1809
+ },
+ {
+ "epoch": 25.139737991266376,
+ "grad_norm": 0.02373199723660946,
+ "learning_rate": 0.0006,
+ "loss": 4.32921028137207,
+ "step": 1810
+ },
+ {
+ "epoch": 25.153711790393015,
+ "grad_norm": 0.024960234761238098,
+ "learning_rate": 0.0006,
+ "loss": 4.295609474182129,
+ "step": 1811
+ },
+ {
+ "epoch": 25.16768558951965,
+ "grad_norm": 0.026743892580270767,
+ "learning_rate": 0.0006,
+ "loss": 4.315308094024658,
+ "step": 1812
+ },
+ {
+ "epoch": 25.18165938864629,
+ "grad_norm": 0.022138185799121857,
+ "learning_rate": 0.0006,
+ "loss": 4.360208511352539,
+ "step": 1813
+ },
+ {
+ "epoch": 25.195633187772927,
+ "grad_norm": 0.022120574489235878,
+ "learning_rate": 0.0006,
+ "loss": 4.1973772048950195,
+ "step": 1814
+ },
+ {
+ "epoch": 25.209606986899562,
+ "grad_norm": 0.02230502851307392,
+ "learning_rate": 0.0006,
+ "loss": 4.285558700561523,
+ "step": 1815
+ },
+ {
+ "epoch": 25.2235807860262,
+ "grad_norm": 0.022301986813545227,
+ "learning_rate": 0.0006,
+ "loss": 4.238343238830566,
+ "step": 1816
+ },
+ {
+ "epoch": 25.23755458515284,
+ "grad_norm": 0.020050447434186935,
+ "learning_rate": 0.0006,
+ "loss": 4.298235893249512,
+ "step": 1817
+ },
+ {
+ "epoch": 25.251528384279474,
+ "grad_norm": 0.01987724006175995,
+ "learning_rate": 0.0006,
+ "loss": 4.301384449005127,
+ "step": 1818
+ },
+ {
+ "epoch": 25.265502183406113,
+ "grad_norm": 0.020440855994820595,
+ "learning_rate": 0.0006,
+ "loss": 4.269218444824219,
+ "step": 1819
+ },
+ {
+ "epoch": 25.27947598253275,
+ "grad_norm": 0.01891893707215786,
+ "learning_rate": 0.0006,
+ "loss": 4.2401957511901855,
+ "step": 1820
+ },
+ {
+ "epoch": 25.29344978165939,
+ "grad_norm": 0.01903688907623291,
+ "learning_rate": 0.0006,
+ "loss": 4.139670372009277,
+ "step": 1821
+ },
+ {
+ "epoch": 25.307423580786025,
+ "grad_norm": 0.018856938928365707,
+ "learning_rate": 0.0006,
+ "loss": 4.233214378356934,
+ "step": 1822
+ },
+ {
+ "epoch": 25.321397379912664,
+ "grad_norm": 0.017791323363780975,
+ "learning_rate": 0.0006,
+ "loss": 4.2342023849487305,
+ "step": 1823
+ },
+ {
+ "epoch": 25.335371179039303,
+ "grad_norm": 0.017304565757513046,
+ "learning_rate": 0.0006,
+ "loss": 4.355198860168457,
+ "step": 1824
+ },
+ {
+ "epoch": 25.349344978165938,
+ "grad_norm": 0.01576206088066101,
+ "learning_rate": 0.0006,
+ "loss": 4.114147186279297,
+ "step": 1825
+ },
+ {
+ "epoch": 25.363318777292577,
+ "grad_norm": 0.015105154365301132,
+ "learning_rate": 0.0006,
+ "loss": 4.313795566558838,
+ "step": 1826
+ },
+ {
+ "epoch": 25.377292576419215,
+ "grad_norm": 0.01563587784767151,
+ "learning_rate": 0.0006,
+ "loss": 4.308282852172852,
+ "step": 1827
+ },
+ {
+ "epoch": 25.39126637554585,
+ "grad_norm": 0.015980906784534454,
+ "learning_rate": 0.0006,
+ "loss": 4.345149040222168,
+ "step": 1828
+ },
+ {
+ "epoch": 25.40524017467249,
+ "grad_norm": 0.01669284701347351,
+ "learning_rate": 0.0006,
+ "loss": 4.237611293792725,
+ "step": 1829
+ },
+ {
+ "epoch": 25.419213973799128,
+ "grad_norm": 0.016202064231038094,
+ "learning_rate": 0.0006,
+ "loss": 4.3072309494018555,
+ "step": 1830
+ },
+ {
+ "epoch": 25.433187772925763,
+ "grad_norm": 0.01478662807494402,
+ "learning_rate": 0.0006,
+ "loss": 4.276215076446533,
+ "step": 1831
+ },
+ {
+ "epoch": 25.4471615720524,
+ "grad_norm": 0.015168731100857258,
+ "learning_rate": 0.0006,
+ "loss": 4.37509822845459,
+ "step": 1832
+ },
+ {
+ "epoch": 25.46113537117904,
+ "grad_norm": 0.01648980937898159,
+ "learning_rate": 0.0006,
+ "loss": 4.349167346954346,
+ "step": 1833
+ },
+ {
+ "epoch": 25.475109170305675,
+ "grad_norm": 0.01708398200571537,
+ "learning_rate": 0.0006,
+ "loss": 4.346580505371094,
+ "step": 1834
+ },
+ {
+ "epoch": 25.489082969432314,
+ "grad_norm": 0.017370784655213356,
+ "learning_rate": 0.0006,
+ "loss": 4.299466133117676,
+ "step": 1835
+ },
+ {
+ "epoch": 25.503056768558952,
+ "grad_norm": 0.016248228028416634,
+ "learning_rate": 0.0006,
+ "loss": 4.184610366821289,
+ "step": 1836
+ },
+ {
+ "epoch": 25.51703056768559,
+ "grad_norm": 0.017106913030147552,
+ "learning_rate": 0.0006,
+ "loss": 4.366856098175049,
+ "step": 1837
+ },
+ {
+ "epoch": 25.531004366812226,
+ "grad_norm": 0.016071965917944908,
+ "learning_rate": 0.0006,
+ "loss": 4.297691345214844,
+ "step": 1838
+ },
+ {
+ "epoch": 25.544978165938865,
+ "grad_norm": 0.016451986506581306,
+ "learning_rate": 0.0006,
+ "loss": 4.229983329772949,
+ "step": 1839
+ },
+ {
+ "epoch": 25.558951965065503,
+ "grad_norm": 0.016195081174373627,
+ "learning_rate": 0.0006,
+ "loss": 4.259893417358398,
+ "step": 1840
+ },
+ {
+ "epoch": 25.57292576419214,
+ "grad_norm": 0.0165175162255764,
+ "learning_rate": 0.0006,
+ "loss": 4.3348236083984375,
+ "step": 1841
+ },
+ {
+ "epoch": 25.586899563318777,
+ "grad_norm": 0.01651517115533352,
+ "learning_rate": 0.0006,
+ "loss": 4.319541931152344,
+ "step": 1842
+ },
+ {
+ "epoch": 25.600873362445416,
+ "grad_norm": 0.016143690794706345,
+ "learning_rate": 0.0006,
+ "loss": 4.345652103424072,
+ "step": 1843
+ },
+ {
+ "epoch": 25.61484716157205,
+ "grad_norm": 0.015140696428716183,
+ "learning_rate": 0.0006,
+ "loss": 4.212047576904297,
+ "step": 1844
+ },
+ {
+ "epoch": 25.62882096069869,
+ "grad_norm": 0.014637443237006664,
+ "learning_rate": 0.0006,
+ "loss": 4.187453269958496,
+ "step": 1845
+ },
+ {
+ "epoch": 25.64279475982533,
+ "grad_norm": 0.014585614204406738,
+ "learning_rate": 0.0006,
+ "loss": 4.253936767578125,
+ "step": 1846
+ },
+ {
+ "epoch": 25.656768558951963,
+ "grad_norm": 0.015158289112150669,
+ "learning_rate": 0.0006,
+ "loss": 4.34612512588501,
+ "step": 1847
+ },
+ {
+ "epoch": 25.670742358078602,
+ "grad_norm": 0.015202849172055721,
+ "learning_rate": 0.0006,
+ "loss": 4.270634651184082,
+ "step": 1848
+ },
+ {
+ "epoch": 25.68471615720524,
+ "grad_norm": 0.015363575890660286,
+ "learning_rate": 0.0006,
+ "loss": 4.271080493927002,
+ "step": 1849
+ },
+ {
+ "epoch": 25.69868995633188,
+ "grad_norm": 0.014360226690769196,
+ "learning_rate": 0.0006,
+ "loss": 4.375033855438232,
+ "step": 1850
+ },
+ {
+ "epoch": 25.712663755458514,
+ "grad_norm": 0.014263181947171688,
+ "learning_rate": 0.0006,
+ "loss": 4.424138069152832,
+ "step": 1851
+ },
+ {
+ "epoch": 25.726637554585153,
+ "grad_norm": 0.014398688450455666,
+ "learning_rate": 0.0006,
+ "loss": 4.410154819488525,
+ "step": 1852
+ },
+ {
+ "epoch": 25.74061135371179,
+ "grad_norm": 0.016131488606333733,
+ "learning_rate": 0.0006,
+ "loss": 4.324014663696289,
+ "step": 1853
+ },
+ {
+ "epoch": 25.754585152838427,
+ "grad_norm": 0.017568735405802727,
+ "learning_rate": 0.0006,
+ "loss": 4.367861747741699,
+ "step": 1854
+ },
+ {
+ "epoch": 25.768558951965066,
+ "grad_norm": 0.017907511442899704,
+ "learning_rate": 0.0006,
+ "loss": 4.352965354919434,
+ "step": 1855
+ },
+ {
+ "epoch": 25.782532751091704,
+ "grad_norm": 0.016918016597628593,
+ "learning_rate": 0.0006,
+ "loss": 4.252397537231445,
+ "step": 1856
+ },
+ {
+ "epoch": 25.79650655021834,
+ "grad_norm": 0.017274610698223114,
+ "learning_rate": 0.0006,
+ "loss": 4.342309951782227,
+ "step": 1857
+ },
+ {
+ "epoch": 25.810480349344978,
+ "grad_norm": 0.017879825085401535,
+ "learning_rate": 0.0006,
+ "loss": 4.330987453460693,
+ "step": 1858
+ },
+ {
+ "epoch": 25.824454148471617,
+ "grad_norm": 0.01736099272966385,
+ "learning_rate": 0.0006,
+ "loss": 4.307255744934082,
+ "step": 1859
+ },
+ {
+ "epoch": 25.83842794759825,
+ "grad_norm": 0.016575824469327927,
+ "learning_rate": 0.0006,
+ "loss": 4.262016773223877,
+ "step": 1860
+ },
+ {
+ "epoch": 25.85240174672489,
+ "grad_norm": 0.015181140042841434,
+ "learning_rate": 0.0006,
+ "loss": 4.291863441467285,
+ "step": 1861
+ },
+ {
+ "epoch": 25.86637554585153,
+ "grad_norm": 0.015965687111020088,
+ "learning_rate": 0.0006,
+ "loss": 4.424836158752441,
+ "step": 1862
+ },
+ {
+ "epoch": 25.880349344978168,
+ "grad_norm": 0.01687219925224781,
+ "learning_rate": 0.0006,
+ "loss": 4.343531131744385,
+ "step": 1863
+ },
+ {
+ "epoch": 25.894323144104803,
+ "grad_norm": 0.015115504153072834,
+ "learning_rate": 0.0006,
+ "loss": 4.357808589935303,
+ "step": 1864
+ },
+ {
+ "epoch": 25.90829694323144,
+ "grad_norm": 0.014150998555123806,
+ "learning_rate": 0.0006,
+ "loss": 4.324550151824951,
+ "step": 1865
+ },
+ {
+ "epoch": 25.92227074235808,
+ "grad_norm": 0.014989510178565979,
+ "learning_rate": 0.0006,
+ "loss": 4.380428314208984,
+ "step": 1866
+ },
+ {
+ "epoch": 25.936244541484715,
+ "grad_norm": 0.01527960691601038,
+ "learning_rate": 0.0006,
+ "loss": 4.399374961853027,
+ "step": 1867
+ },
+ {
+ "epoch": 25.950218340611354,
+ "grad_norm": 0.015434633940458298,
+ "learning_rate": 0.0006,
+ "loss": 4.260984420776367,
+ "step": 1868
+ },
+ {
+ "epoch": 25.964192139737992,
+ "grad_norm": 0.015370228327810764,
+ "learning_rate": 0.0006,
+ "loss": 4.3340253829956055,
+ "step": 1869
+ },
+ {
+ "epoch": 25.978165938864628,
+ "grad_norm": 0.015348542481660843,
+ "learning_rate": 0.0006,
+ "loss": 4.369132995605469,
+ "step": 1870
+ },
+ {
+ "epoch": 25.992139737991266,
+ "grad_norm": 0.015354936942458153,
+ "learning_rate": 0.0006,
+ "loss": 4.381937026977539,
+ "step": 1871
+ },
+ {
+ "epoch": 26.0,
+ "grad_norm": 0.01796272210776806,
+ "learning_rate": 0.0006,
+ "loss": 4.286768436431885,
+ "step": 1872
+ },
+ {
+ "epoch": 26.0,
+ "eval_loss": 4.576984405517578,
+ "eval_runtime": 57.2091,
+ "eval_samples_per_second": 42.685,
+ "eval_steps_per_second": 1.346,
+ "step": 1872
+ },
+ {
+ "epoch": 26.01397379912664,
+ "grad_norm": 0.01748845726251602,
+ "learning_rate": 0.0006,
+ "loss": 4.337504863739014,
+ "step": 1873
+ },
+ {
+ "epoch": 26.027947598253274,
+ "grad_norm": 0.019244296476244926,
+ "learning_rate": 0.0006,
+ "loss": 4.264280796051025,
+ "step": 1874
+ },
+ {
+ "epoch": 26.041921397379912,
+ "grad_norm": 0.0215692650526762,
+ "learning_rate": 0.0006,
+ "loss": 4.350830078125,
+ "step": 1875
+ },
+ {
+ "epoch": 26.05589519650655,
+ "grad_norm": 0.022189252078533173,
+ "learning_rate": 0.0006,
+ "loss": 4.299098968505859,
+ "step": 1876
+ },
+ {
+ "epoch": 26.069868995633186,
+ "grad_norm": 0.022384043782949448,
+ "learning_rate": 0.0006,
+ "loss": 4.1185173988342285,
+ "step": 1877
+ },
+ {
+ "epoch": 26.083842794759825,
+ "grad_norm": 0.02076014317572117,
+ "learning_rate": 0.0006,
+ "loss": 4.297612190246582,
+ "step": 1878
+ },
+ {
+ "epoch": 26.097816593886463,
+ "grad_norm": 0.020843051373958588,
+ "learning_rate": 0.0006,
+ "loss": 4.294098854064941,
+ "step": 1879
+ },
+ {
+ "epoch": 26.111790393013102,
+ "grad_norm": 0.020654456689953804,
+ "learning_rate": 0.0006,
+ "loss": 4.267856597900391,
+ "step": 1880
+ },
+ {
+ "epoch": 26.125764192139737,
+ "grad_norm": 0.01860472559928894,
+ "learning_rate": 0.0006,
+ "loss": 4.284232139587402,
+ "step": 1881
+ },
+ {
+ "epoch": 26.139737991266376,
+ "grad_norm": 0.018719421699643135,
+ "learning_rate": 0.0006,
+ "loss": 4.277214527130127,
+ "step": 1882
+ },
+ {
+ "epoch": 26.153711790393015,
+ "grad_norm": 0.020903829485177994,
+ "learning_rate": 0.0006,
+ "loss": 4.254947662353516,
+ "step": 1883
+ },
+ {
+ "epoch": 26.16768558951965,
+ "grad_norm": 0.02020149491727352,
+ "learning_rate": 0.0006,
+ "loss": 4.2526936531066895,
+ "step": 1884
+ },
+ {
+ "epoch": 26.18165938864629,
+ "grad_norm": 0.0180392786860466,
+ "learning_rate": 0.0006,
+ "loss": 4.309922695159912,
+ "step": 1885
+ },
+ {
+ "epoch": 26.195633187772927,
+ "grad_norm": 0.018163125962018967,
+ "learning_rate": 0.0006,
+ "loss": 4.13860559463501,
+ "step": 1886
+ },
+ {
+ "epoch": 26.209606986899562,
+ "grad_norm": 0.01853291131556034,
+ "learning_rate": 0.0006,
+ "loss": 4.190926551818848,
+ "step": 1887
+ },
+ {
+ "epoch": 26.2235807860262,
+ "grad_norm": 0.01757677271962166,
+ "learning_rate": 0.0006,
+ "loss": 4.163999557495117,
+ "step": 1888
+ },
+ {
+ "epoch": 26.23755458515284,
+ "grad_norm": 0.01767992228269577,
+ "learning_rate": 0.0006,
+ "loss": 4.184430122375488,
+ "step": 1889
+ },
+ {
+ "epoch": 26.251528384279474,
+ "grad_norm": 0.018139973282814026,
+ "learning_rate": 0.0006,
+ "loss": 4.300766944885254,
+ "step": 1890
+ },
+ {
+ "epoch": 26.265502183406113,
+ "grad_norm": 0.020184461027383804,
+ "learning_rate": 0.0006,
+ "loss": 4.232537269592285,
+ "step": 1891
+ },
+ {
+ "epoch": 26.27947598253275,
+ "grad_norm": 0.02028856985270977,
+ "learning_rate": 0.0006,
+ "loss": 4.324883460998535,
+ "step": 1892
+ },
+ {
+ "epoch": 26.29344978165939,
+ "grad_norm": 0.017120616510510445,
+ "learning_rate": 0.0006,
+ "loss": 4.23493766784668,
+ "step": 1893
+ },
+ {
+ "epoch": 26.307423580786025,
+ "grad_norm": 0.016202978789806366,
+ "learning_rate": 0.0006,
+ "loss": 4.197602272033691,
+ "step": 1894
+ },
+ {
+ "epoch": 26.321397379912664,
+ "grad_norm": 0.017231421545147896,
+ "learning_rate": 0.0006,
+ "loss": 4.302001953125,
+ "step": 1895
+ },
+ {
+ "epoch": 26.335371179039303,
+ "grad_norm": 0.01725655607879162,
+ "learning_rate": 0.0006,
+ "loss": 4.356800079345703,
+ "step": 1896
+ },
+ {
+ "epoch": 26.349344978165938,
+ "grad_norm": 0.016657443717122078,
+ "learning_rate": 0.0006,
+ "loss": 4.266307830810547,
+ "step": 1897
+ },
+ {
+ "epoch": 26.363318777292577,
+ "grad_norm": 0.01616556942462921,
+ "learning_rate": 0.0006,
+ "loss": 4.307868003845215,
+ "step": 1898
+ },
+ {
+ "epoch": 26.377292576419215,
+ "grad_norm": 0.016621306538581848,
+ "learning_rate": 0.0006,
+ "loss": 4.172905921936035,
+ "step": 1899
+ },
+ {
+ "epoch": 26.39126637554585,
+ "grad_norm": 0.016243206337094307,
+ "learning_rate": 0.0006,
+ "loss": 4.162896156311035,
+ "step": 1900
+ },
+ {
+ "epoch": 26.40524017467249,
+ "grad_norm": 0.016663808375597,
+ "learning_rate": 0.0006,
+ "loss": 4.258232116699219,
+ "step": 1901
+ },
+ {
+ "epoch": 26.419213973799128,
+ "grad_norm": 0.01677478663623333,
+ "learning_rate": 0.0006,
+ "loss": 4.336090087890625,
+ "step": 1902
+ },
+ {
+ "epoch": 26.433187772925763,
+ "grad_norm": 0.01641303487122059,
+ "learning_rate": 0.0006,
+ "loss": 4.191476821899414,
+ "step": 1903
+ },
+ {
+ "epoch": 26.4471615720524,
+ "grad_norm": 0.016700323671102524,
+ "learning_rate": 0.0006,
+ "loss": 4.299427509307861,
+ "step": 1904
+ },
+ {
+ "epoch": 26.46113537117904,
+ "grad_norm": 0.017595035955309868,
+ "learning_rate": 0.0006,
+ "loss": 4.312828063964844,
+ "step": 1905
+ },
+ {
+ "epoch": 26.475109170305675,
+ "grad_norm": 0.020300284028053284,
+ "learning_rate": 0.0006,
+ "loss": 4.28331184387207,
+ "step": 1906
+ },
+ {
+ "epoch": 26.489082969432314,
+ "grad_norm": 0.02198321744799614,
+ "learning_rate": 0.0006,
+ "loss": 4.3280839920043945,
+ "step": 1907
+ },
+ {
+ "epoch": 26.503056768558952,
+ "grad_norm": 0.02003200724720955,
+ "learning_rate": 0.0006,
+ "loss": 4.325623989105225,
+ "step": 1908
+ },
+ {
+ "epoch": 26.51703056768559,
+ "grad_norm": 0.01870287023484707,
+ "learning_rate": 0.0006,
+ "loss": 4.251203536987305,
+ "step": 1909
+ },
+ {
+ "epoch": 26.531004366812226,
+ "grad_norm": 0.01835448481142521,
+ "learning_rate": 0.0006,
+ "loss": 4.335144996643066,
+ "step": 1910
+ },
+ {
+ "epoch": 26.544978165938865,
+ "grad_norm": 0.02046327292919159,
+ "learning_rate": 0.0006,
+ "loss": 4.238023281097412,
+ "step": 1911
+ },
+ {
+ "epoch": 26.558951965065503,
+ "grad_norm": 0.019771311432123184,
+ "learning_rate": 0.0006,
+ "loss": 4.353594779968262,
+ "step": 1912
+ },
+ {
+ "epoch": 26.57292576419214,
+ "grad_norm": 0.018090544268488884,
+ "learning_rate": 0.0006,
+ "loss": 4.304343223571777,
+ "step": 1913
+ },
+ {
+ "epoch": 26.586899563318777,
+ "grad_norm": 0.017489777877926826,
+ "learning_rate": 0.0006,
+ "loss": 4.204375267028809,
+ "step": 1914
+ },
+ {
+ "epoch": 26.600873362445416,
+ "grad_norm": 0.016785891726613045,
+ "learning_rate": 0.0006,
+ "loss": 4.286669731140137,
+ "step": 1915
+ },
+ {
+ "epoch": 26.61484716157205,
+ "grad_norm": 0.016382789239287376,
+ "learning_rate": 0.0006,
+ "loss": 4.347662448883057,
+ "step": 1916
+ },
+ {
+ "epoch": 26.62882096069869,
+ "grad_norm": 0.016581876203417778,
+ "learning_rate": 0.0006,
+ "loss": 4.2716193199157715,
+ "step": 1917
+ },
+ {
+ "epoch": 26.64279475982533,
+ "grad_norm": 0.015627184882760048,
+ "learning_rate": 0.0006,
+ "loss": 4.311489105224609,
+ "step": 1918
+ },
+ {
+ "epoch": 26.656768558951963,
+ "grad_norm": 0.015908483415842056,
+ "learning_rate": 0.0006,
+ "loss": 4.262099742889404,
+ "step": 1919
+ },
+ {
+ "epoch": 26.670742358078602,
+ "grad_norm": 0.015576236881315708,
+ "learning_rate": 0.0006,
+ "loss": 4.275871276855469,
+ "step": 1920
+ },
+ {
+ "epoch": 26.68471615720524,
+ "grad_norm": 0.014621064998209476,
+ "learning_rate": 0.0006,
+ "loss": 4.184260845184326,
+ "step": 1921
+ },
+ {
+ "epoch": 26.69868995633188,
+ "grad_norm": 0.014797811396420002,
+ "learning_rate": 0.0006,
+ "loss": 4.3412628173828125,
+ "step": 1922
+ },
+ {
+ "epoch": 26.712663755458514,
+ "grad_norm": 0.015610570088028908,
+ "learning_rate": 0.0006,
+ "loss": 4.181094646453857,
+ "step": 1923
+ },
+ {
+ "epoch": 26.726637554585153,
+ "grad_norm": 0.016048265621066093,
+ "learning_rate": 0.0006,
+ "loss": 4.365290641784668,
+ "step": 1924
+ },
+ {
+ "epoch": 26.74061135371179,
+ "grad_norm": 0.016604071483016014,
+ "learning_rate": 0.0006,
+ "loss": 4.2628912925720215,
+ "step": 1925
+ },
+ {
+ "epoch": 26.754585152838427,
+ "grad_norm": 0.017044221982359886,
+ "learning_rate": 0.0006,
+ "loss": 4.378512382507324,
+ "step": 1926
+ },
+ {
+ "epoch": 26.768558951965066,
+ "grad_norm": 0.01746384985744953,
+ "learning_rate": 0.0006,
+ "loss": 4.3888773918151855,
+ "step": 1927
+ },
+ {
+ "epoch": 26.782532751091704,
+ "grad_norm": 0.01572711206972599,
+ "learning_rate": 0.0006,
+ "loss": 4.189935684204102,
+ "step": 1928
+ },
+ {
+ "epoch": 26.79650655021834,
+ "grad_norm": 0.014976629987359047,
+ "learning_rate": 0.0006,
+ "loss": 4.280525207519531,
+ "step": 1929
+ },
+ {
+ "epoch": 26.810480349344978,
+ "grad_norm": 0.014804664999246597,
+ "learning_rate": 0.0006,
+ "loss": 4.270630836486816,
+ "step": 1930
+ },
+ {
+ "epoch": 26.824454148471617,
+ "grad_norm": 0.014641920104622841,
+ "learning_rate": 0.0006,
+ "loss": 4.319505214691162,
+ "step": 1931
+ },
+ {
+ "epoch": 26.83842794759825,
+ "grad_norm": 0.014928505755960941,
+ "learning_rate": 0.0006,
+ "loss": 4.180147647857666,
+ "step": 1932
+ },
+ {
+ "epoch": 26.85240174672489,
+ "grad_norm": 0.016130365431308746,
+ "learning_rate": 0.0006,
+ "loss": 4.260660648345947,
+ "step": 1933
+ },
+ {
+ "epoch": 26.86637554585153,
+ "grad_norm": 0.015642493963241577,
+ "learning_rate": 0.0006,
+ "loss": 4.304064750671387,
+ "step": 1934
+ },
+ {
+ "epoch": 26.880349344978168,
+ "grad_norm": 0.015988217666745186,
+ "learning_rate": 0.0006,
+ "loss": 4.2548017501831055,
+ "step": 1935
+ },
+ {
+ "epoch": 26.894323144104803,
+ "grad_norm": 0.016501398757100105,
+ "learning_rate": 0.0006,
+ "loss": 4.260058879852295,
+ "step": 1936
+ },
+ {
+ "epoch": 26.90829694323144,
+ "grad_norm": 0.01645040698349476,
+ "learning_rate": 0.0006,
+ "loss": 4.483081340789795,
+ "step": 1937
+ },
+ {
+ "epoch": 26.92227074235808,
+ "grad_norm": 0.016140863299369812,
+ "learning_rate": 0.0006,
+ "loss": 4.1776838302612305,
+ "step": 1938
+ },
+ {
+ "epoch": 26.936244541484715,
+ "grad_norm": 0.016681019216775894,
+ "learning_rate": 0.0006,
+ "loss": 4.224747657775879,
+ "step": 1939
+ },
+ {
+ "epoch": 26.950218340611354,
+ "grad_norm": 0.01650378853082657,
+ "learning_rate": 0.0006,
+ "loss": 4.305703163146973,
+ "step": 1940
+ },
+ {
+ "epoch": 26.964192139737992,
+ "grad_norm": 0.017051683738827705,
+ "learning_rate": 0.0006,
+ "loss": 4.285835266113281,
+ "step": 1941
+ },
+ {
+ "epoch": 26.978165938864628,
+ "grad_norm": 0.016894327476620674,
+ "learning_rate": 0.0006,
+ "loss": 4.318901538848877,
+ "step": 1942
+ },
+ {
+ "epoch": 26.992139737991266,
+ "grad_norm": 0.017313100397586823,
+ "learning_rate": 0.0006,
+ "loss": 4.281434535980225,
+ "step": 1943
+ },
+ {
+ "epoch": 27.0,
+ "grad_norm": 0.018450884148478508,
+ "learning_rate": 0.0006,
+ "loss": 4.335104942321777,
+ "step": 1944
+ },
+ {
+ "epoch": 27.0,
+ "eval_loss": 4.640687465667725,
+ "eval_runtime": 56.9624,
+ "eval_samples_per_second": 42.87,
+ "eval_steps_per_second": 1.352,
+ "step": 1944
+ },
+ {
+ "epoch": 27.01397379912664,
+ "grad_norm": 0.017755698412656784,
+ "learning_rate": 0.0006,
+ "loss": 4.278824329376221,
+ "step": 1945
+ },
+ {
+ "epoch": 27.027947598253274,
+ "grad_norm": 0.01724310778081417,
+ "learning_rate": 0.0006,
+ "loss": 4.311049461364746,
+ "step": 1946
+ },
+ {
+ "epoch": 27.041921397379912,
+ "grad_norm": 0.01690123789012432,
+ "learning_rate": 0.0006,
+ "loss": 4.261994361877441,
+ "step": 1947
+ },
+ {
+ "epoch": 27.05589519650655,
+ "grad_norm": 0.018406696617603302,
+ "learning_rate": 0.0006,
+ "loss": 4.1925153732299805,
+ "step": 1948
+ },
+ {
+ "epoch": 27.069868995633186,
+ "grad_norm": 0.019945867359638214,
+ "learning_rate": 0.0006,
+ "loss": 4.108537197113037,
+ "step": 1949
+ },
+ {
+ "epoch": 27.083842794759825,
+ "grad_norm": 0.020316628739237785,
+ "learning_rate": 0.0006,
+ "loss": 4.209043025970459,
+ "step": 1950
+ },
+ {
+ "epoch": 27.097816593886463,
+ "grad_norm": 0.019690370187163353,
+ "learning_rate": 0.0006,
+ "loss": 4.195789337158203,
+ "step": 1951
+ },
+ {
+ "epoch": 27.111790393013102,
+ "grad_norm": 0.01785232499241829,
+ "learning_rate": 0.0006,
+ "loss": 4.21466064453125,
+ "step": 1952
+ },
+ {
+ "epoch": 27.125764192139737,
+ "grad_norm": 0.0164767038077116,
+ "learning_rate": 0.0006,
+ "loss": 4.270247459411621,
+ "step": 1953
+ },
+ {
+ "epoch": 27.139737991266376,
+ "grad_norm": 0.018008455634117126,
+ "learning_rate": 0.0006,
+ "loss": 4.201042175292969,
+ "step": 1954
+ },
+ {
+ "epoch": 27.153711790393015,
+ "grad_norm": 0.01807340793311596,
+ "learning_rate": 0.0006,
+ "loss": 4.25289249420166,
+ "step": 1955
+ },
+ {
+ "epoch": 27.16768558951965,
+ "grad_norm": 0.016246909275650978,
+ "learning_rate": 0.0006,
+ "loss": 4.116283416748047,
+ "step": 1956
+ },
+ {
+ "epoch": 27.18165938864629,
+ "grad_norm": 0.017545776441693306,
+ "learning_rate": 0.0006,
+ "loss": 4.316999435424805,
+ "step": 1957
+ },
+ {
+ "epoch": 27.195633187772927,
+ "grad_norm": 0.017915375530719757,
+ "learning_rate": 0.0006,
+ "loss": 4.253859519958496,
+ "step": 1958
+ },
+ {
+ "epoch": 27.209606986899562,
+ "grad_norm": 0.018112223595380783,
+ "learning_rate": 0.0006,
+ "loss": 4.2541022300720215,
+ "step": 1959
+ },
+ {
+ "epoch": 27.2235807860262,
+ "grad_norm": 0.01846139505505562,
+ "learning_rate": 0.0006,
+ "loss": 4.20438289642334,
+ "step": 1960
+ },
+ {
+ "epoch": 27.23755458515284,
+ "grad_norm": 0.01943897269666195,
+ "learning_rate": 0.0006,
+ "loss": 4.2799577713012695,
+ "step": 1961
+ },
+ {
+ "epoch": 27.251528384279474,
+ "grad_norm": 0.018253348767757416,
+ "learning_rate": 0.0006,
+ "loss": 4.1720499992370605,
+ "step": 1962
+ },
+ {
+ "epoch": 27.265502183406113,
+ "grad_norm": 0.018292246386408806,
+ "learning_rate": 0.0006,
+ "loss": 4.28231143951416,
+ "step": 1963
+ },
+ {
+ "epoch": 27.27947598253275,
+ "grad_norm": 0.01943155936896801,
+ "learning_rate": 0.0006,
+ "loss": 4.314976692199707,
+ "step": 1964
+ },
+ {
+ "epoch": 27.29344978165939,
+ "grad_norm": 0.019064994528889656,
+ "learning_rate": 0.0006,
+ "loss": 4.25328254699707,
+ "step": 1965
+ },
+ {
+ "epoch": 27.307423580786025,
+ "grad_norm": 0.017423273995518684,
+ "learning_rate": 0.0006,
+ "loss": 4.195992469787598,
+ "step": 1966
+ },
+ {
+ "epoch": 27.321397379912664,
+ "grad_norm": 0.01677924580872059,
+ "learning_rate": 0.0006,
+ "loss": 4.260043144226074,
+ "step": 1967
+ },
+ {
+ "epoch": 27.335371179039303,
+ "grad_norm": 0.01720673404633999,
+ "learning_rate": 0.0006,
+ "loss": 4.222780704498291,
+ "step": 1968
+ },
+ {
+ "epoch": 27.349344978165938,
+ "grad_norm": 0.017300743609666824,
+ "learning_rate": 0.0006,
+ "loss": 4.109524250030518,
+ "step": 1969
+ },
+ {
+ "epoch": 27.363318777292577,
+ "grad_norm": 0.016065871343016624,
+ "learning_rate": 0.0006,
+ "loss": 4.36108922958374,
+ "step": 1970
+ },
+ {
+ "epoch": 27.377292576419215,
+ "grad_norm": 0.016626616939902306,
+ "learning_rate": 0.0006,
+ "loss": 4.220630645751953,
+ "step": 1971
+ },
+ {
+ "epoch": 27.39126637554585,
+ "grad_norm": 0.016482602804899216,
+ "learning_rate": 0.0006,
+ "loss": 4.344797134399414,
+ "step": 1972
+ },
+ {
+ "epoch": 27.40524017467249,
+ "grad_norm": 0.01727953553199768,
+ "learning_rate": 0.0006,
+ "loss": 4.296645164489746,
+ "step": 1973
+ },
+ {
+ "epoch": 27.419213973799128,
+ "grad_norm": 0.01522731315344572,
+ "learning_rate": 0.0006,
+ "loss": 4.223751068115234,
+ "step": 1974
+ },
+ {
+ "epoch": 27.433187772925763,
+ "grad_norm": 0.01581776700913906,
+ "learning_rate": 0.0006,
+ "loss": 4.351379871368408,
+ "step": 1975
+ },
+ {
+ "epoch": 27.4471615720524,
+ "grad_norm": 0.016394605860114098,
+ "learning_rate": 0.0006,
+ "loss": 4.2412919998168945,
+ "step": 1976
+ },
+ {
+ "epoch": 27.46113537117904,
+ "grad_norm": 0.01751169003546238,
+ "learning_rate": 0.0006,
+ "loss": 4.257606506347656,
+ "step": 1977
+ },
+ {
+ "epoch": 27.475109170305675,
+ "grad_norm": 0.017452310770750046,
+ "learning_rate": 0.0006,
+ "loss": 4.380134582519531,
+ "step": 1978
+ },
+ {
+ "epoch": 27.489082969432314,
+ "grad_norm": 0.017741898074746132,
+ "learning_rate": 0.0006,
+ "loss": 4.279632568359375,
+ "step": 1979
+ },
+ {
+ "epoch": 27.503056768558952,
+ "grad_norm": 0.017680590972304344,
+ "learning_rate": 0.0006,
+ "loss": 4.273406028747559,
+ "step": 1980
+ },
+ {
+ "epoch": 27.51703056768559,
+ "grad_norm": 0.01864718459546566,
+ "learning_rate": 0.0006,
+ "loss": 4.278583526611328,
+ "step": 1981
+ },
+ {
+ "epoch": 27.531004366812226,
+ "grad_norm": 0.019558541476726532,
+ "learning_rate": 0.0006,
+ "loss": 4.211919784545898,
+ "step": 1982
+ },
+ {
+ "epoch": 27.544978165938865,
+ "grad_norm": 0.019467400386929512,
+ "learning_rate": 0.0006,
+ "loss": 4.276822566986084,
+ "step": 1983
+ },
+ {
+ "epoch": 27.558951965065503,
+ "grad_norm": 0.01756172813475132,
+ "learning_rate": 0.0006,
+ "loss": 4.165379047393799,
+ "step": 1984
+ },
+ {
+ "epoch": 27.57292576419214,
+ "grad_norm": 0.017458263784646988,
+ "learning_rate": 0.0006,
+ "loss": 4.2992401123046875,
+ "step": 1985
+ },
+ {
+ "epoch": 27.586899563318777,
+ "grad_norm": 0.018089665099978447,
+ "learning_rate": 0.0006,
+ "loss": 4.243378639221191,
+ "step": 1986
+ },
+ {
+ "epoch": 27.600873362445416,
+ "grad_norm": 0.019535524770617485,
+ "learning_rate": 0.0006,
+ "loss": 4.321477890014648,
+ "step": 1987
+ },
+ {
+ "epoch": 27.61484716157205,
+ "grad_norm": 0.01997970975935459,
+ "learning_rate": 0.0006,
+ "loss": 4.123082637786865,
+ "step": 1988
+ },
+ {
+ "epoch": 27.62882096069869,
+ "grad_norm": 0.01913067139685154,
+ "learning_rate": 0.0006,
+ "loss": 4.289896488189697,
+ "step": 1989
+ },
+ {
+ "epoch": 27.64279475982533,
+ "grad_norm": 0.0169806070625782,
+ "learning_rate": 0.0006,
+ "loss": 4.247004985809326,
+ "step": 1990
+ },
+ {
+ "epoch": 27.656768558951963,
+ "grad_norm": 0.017315007746219635,
+ "learning_rate": 0.0006,
+ "loss": 4.2095441818237305,
+ "step": 1991
+ },
+ {
+ "epoch": 27.670742358078602,
+ "grad_norm": 0.015098759904503822,
+ "learning_rate": 0.0006,
+ "loss": 4.145560264587402,
+ "step": 1992
+ },
+ {
+ "epoch": 27.68471615720524,
+ "grad_norm": 0.016240514814853668,
+ "learning_rate": 0.0006,
+ "loss": 4.254927635192871,
+ "step": 1993
+ },
+ {
+ "epoch": 27.69868995633188,
+ "grad_norm": 0.01621108129620552,
+ "learning_rate": 0.0006,
+ "loss": 4.219725608825684,
+ "step": 1994
+ },
+ {
+ "epoch": 27.712663755458514,
+ "grad_norm": 0.015191009268164635,
+ "learning_rate": 0.0006,
+ "loss": 4.194345474243164,
+ "step": 1995
+ },
+ {
+ "epoch": 27.726637554585153,
+ "grad_norm": 0.014387983828783035,
+ "learning_rate": 0.0006,
+ "loss": 4.211467742919922,
+ "step": 1996
+ },
+ {
+ "epoch": 27.74061135371179,
+ "grad_norm": 0.015460561029613018,
+ "learning_rate": 0.0006,
+ "loss": 4.358700752258301,
+ "step": 1997
+ },
+ {
+ "epoch": 27.754585152838427,
+ "grad_norm": 0.015100076794624329,
+ "learning_rate": 0.0006,
+ "loss": 4.171223163604736,
+ "step": 1998
+ },
+ {
+ "epoch": 27.768558951965066,
+ "grad_norm": 0.015607891604304314,
+ "learning_rate": 0.0006,
+ "loss": 4.217874526977539,
+ "step": 1999
+ },
+ {
+ "epoch": 27.782532751091704,
+ "grad_norm": 0.015650689601898193,
+ "learning_rate": 0.0006,
+ "loss": 4.280780792236328,
+ "step": 2000
+ },
+ {
+ "epoch": 27.79650655021834,
+ "grad_norm": 0.01580994389951229,
+ "learning_rate": 0.0006,
+ "loss": 4.159212112426758,
+ "step": 2001
+ },
+ {
+ "epoch": 27.810480349344978,
+ "grad_norm": 0.016607413068413734,
+ "learning_rate": 0.0006,
+ "loss": 4.238853454589844,
+ "step": 2002
+ },
+ {
+ "epoch": 27.824454148471617,
+ "grad_norm": 0.017383860424160957,
+ "learning_rate": 0.0006,
+ "loss": 4.233944892883301,
+ "step": 2003
+ },
+ {
+ "epoch": 27.83842794759825,
+ "grad_norm": 0.014966659247875214,
+ "learning_rate": 0.0006,
+ "loss": 4.167392730712891,
+ "step": 2004
+ },
+ {
+ "epoch": 27.85240174672489,
+ "grad_norm": 0.014589878730475903,
+ "learning_rate": 0.0006,
+ "loss": 4.2390899658203125,
+ "step": 2005
+ },
+ {
+ "epoch": 27.86637554585153,
+ "grad_norm": 0.017179450020194054,
+ "learning_rate": 0.0006,
+ "loss": 4.365785598754883,
+ "step": 2006
+ },
+ {
+ "epoch": 27.880349344978168,
+ "grad_norm": 0.01893901452422142,
+ "learning_rate": 0.0006,
+ "loss": 4.289430141448975,
+ "step": 2007
+ },
+ {
+ "epoch": 27.894323144104803,
+ "grad_norm": 0.018619712442159653,
+ "learning_rate": 0.0006,
+ "loss": 4.310665130615234,
+ "step": 2008
+ },
+ {
+ "epoch": 27.90829694323144,
+ "grad_norm": 0.019180385395884514,
+ "learning_rate": 0.0006,
+ "loss": 4.27454137802124,
+ "step": 2009
+ },
+ {
+ "epoch": 27.92227074235808,
+ "grad_norm": 0.017811523750424385,
+ "learning_rate": 0.0006,
+ "loss": 4.312562465667725,
+ "step": 2010
+ },
+ {
+ "epoch": 27.936244541484715,
+ "grad_norm": 0.016583573073148727,
+ "learning_rate": 0.0006,
+ "loss": 4.202253341674805,
+ "step": 2011
+ },
+ {
+ "epoch": 27.950218340611354,
+ "grad_norm": 0.01658615842461586,
+ "learning_rate": 0.0006,
+ "loss": 4.183433532714844,
+ "step": 2012
+ },
+ {
+ "epoch": 27.964192139737992,
+ "grad_norm": 0.015327401459217072,
+ "learning_rate": 0.0006,
+ "loss": 4.2125139236450195,
+ "step": 2013
+ },
+ {
+ "epoch": 27.978165938864628,
+ "grad_norm": 0.015004601329565048,
+ "learning_rate": 0.0006,
+ "loss": 4.257392883300781,
+ "step": 2014
+ },
+ {
+ "epoch": 27.992139737991266,
+ "grad_norm": 0.015259161591529846,
+ "learning_rate": 0.0006,
+ "loss": 4.241818428039551,
+ "step": 2015
+ },
+ {
+ "epoch": 28.0,
+ "grad_norm": 0.016808776184916496,
+ "learning_rate": 0.0006,
+ "loss": 4.146241188049316,
+ "step": 2016
+ },
+ {
+ "epoch": 28.0,
+ "eval_loss": 4.663090705871582,
+ "eval_runtime": 56.4559,
+ "eval_samples_per_second": 43.255,
+ "eval_steps_per_second": 1.364,
+ "step": 2016
+ },
+ {
+ "epoch": 28.01397379912664,
+ "grad_norm": 0.015784382820129395,
+ "learning_rate": 0.0006,
+ "loss": 4.044191360473633,
+ "step": 2017
+ },
+ {
+ "epoch": 28.027947598253274,
+ "grad_norm": 0.015636710450053215,
+ "learning_rate": 0.0006,
+ "loss": 4.128754615783691,
+ "step": 2018
+ },
+ {
+ "epoch": 28.041921397379912,
+ "grad_norm": 0.015364975668489933,
+ "learning_rate": 0.0006,
+ "loss": 4.191835403442383,
+ "step": 2019
+ },
+ {
+ "epoch": 28.05589519650655,
+ "grad_norm": 0.015015700832009315,
+ "learning_rate": 0.0006,
+ "loss": 4.120022773742676,
+ "step": 2020
+ },
+ {
+ "epoch": 28.069868995633186,
+ "grad_norm": 0.016666799783706665,
+ "learning_rate": 0.0006,
+ "loss": 4.33906364440918,
+ "step": 2021
+ },
+ {
+ "epoch": 28.083842794759825,
+ "grad_norm": 0.01780104823410511,
+ "learning_rate": 0.0006,
+ "loss": 4.279458999633789,
+ "step": 2022
+ },
+ {
+ "epoch": 28.097816593886463,
+ "grad_norm": 0.01870882883667946,
+ "learning_rate": 0.0006,
+ "loss": 4.34728479385376,
+ "step": 2023
+ },
+ {
+ "epoch": 28.111790393013102,
+ "grad_norm": 0.018406381830573082,
+ "learning_rate": 0.0006,
+ "loss": 4.211311340332031,
+ "step": 2024
+ },
+ {
+ "epoch": 28.125764192139737,
+ "grad_norm": 0.017833448946475983,
+ "learning_rate": 0.0006,
+ "loss": 4.223254203796387,
+ "step": 2025
+ },
+ {
+ "epoch": 28.139737991266376,
+ "grad_norm": 0.017552688717842102,
+ "learning_rate": 0.0006,
+ "loss": 4.208823204040527,
+ "step": 2026
+ },
+ {
+ "epoch": 28.153711790393015,
+ "grad_norm": 0.018500229343771935,
+ "learning_rate": 0.0006,
+ "loss": 4.198025226593018,
+ "step": 2027
+ },
+ {
+ "epoch": 28.16768558951965,
+ "grad_norm": 0.01782156340777874,
+ "learning_rate": 0.0006,
+ "loss": 4.179978370666504,
+ "step": 2028
+ },
+ {
+ "epoch": 28.18165938864629,
+ "grad_norm": 0.019580967724323273,
+ "learning_rate": 0.0006,
+ "loss": 4.291114807128906,
+ "step": 2029
+ },
+ {
+ "epoch": 28.195633187772927,
+ "grad_norm": 0.020804036408662796,
+ "learning_rate": 0.0006,
+ "loss": 4.220607280731201,
+ "step": 2030
+ },
+ {
+ "epoch": 28.209606986899562,
+ "grad_norm": 0.019595693796873093,
+ "learning_rate": 0.0006,
+ "loss": 4.2905731201171875,
+ "step": 2031
+ },
+ {
+ "epoch": 28.2235807860262,
+ "grad_norm": 0.019247086718678474,
+ "learning_rate": 0.0006,
+ "loss": 4.2467360496521,
+ "step": 2032
+ },
+ {
+ "epoch": 28.23755458515284,
+ "grad_norm": 0.01958036608994007,
+ "learning_rate": 0.0006,
+ "loss": 4.286137580871582,
+ "step": 2033
+ },
+ {
+ "epoch": 28.251528384279474,
+ "grad_norm": 0.021354753524065018,
+ "learning_rate": 0.0006,
+ "loss": 4.234755039215088,
+ "step": 2034
+ },
+ {
+ "epoch": 28.265502183406113,
+ "grad_norm": 0.02206375077366829,
+ "learning_rate": 0.0006,
+ "loss": 4.149641036987305,
+ "step": 2035
+ },
+ {
+ "epoch": 28.27947598253275,
+ "grad_norm": 0.02201310358941555,
+ "learning_rate": 0.0006,
+ "loss": 4.1928629875183105,
+ "step": 2036
+ },
+ {
+ "epoch": 28.29344978165939,
+ "grad_norm": 0.01961950585246086,
+ "learning_rate": 0.0006,
+ "loss": 4.211404323577881,
+ "step": 2037
+ },
+ {
+ "epoch": 28.307423580786025,
+ "grad_norm": 0.019211795181035995,
+ "learning_rate": 0.0006,
+ "loss": 4.200976848602295,
+ "step": 2038
+ },
+ {
+ "epoch": 28.321397379912664,
+ "grad_norm": 0.01779370754957199,
+ "learning_rate": 0.0006,
+ "loss": 4.249148845672607,
+ "step": 2039
+ },
+ {
+ "epoch": 28.335371179039303,
+ "grad_norm": 0.018635908141732216,
+ "learning_rate": 0.0006,
+ "loss": 4.11806058883667,
+ "step": 2040
+ },
+ {
+ "epoch": 28.349344978165938,
+ "grad_norm": 0.019792694598436356,
+ "learning_rate": 0.0006,
+ "loss": 4.335224151611328,
+ "step": 2041
+ },
+ {
+ "epoch": 28.363318777292577,
+ "grad_norm": 0.021731717512011528,
+ "learning_rate": 0.0006,
+ "loss": 4.302778244018555,
+ "step": 2042
+ },
+ {
+ "epoch": 28.377292576419215,
+ "grad_norm": 0.020823447033762932,
+ "learning_rate": 0.0006,
+ "loss": 4.260610580444336,
+ "step": 2043
+ },
+ {
+ "epoch": 28.39126637554585,
+ "grad_norm": 0.019385412335395813,
+ "learning_rate": 0.0006,
+ "loss": 4.2247724533081055,
+ "step": 2044
+ },
+ {
+ "epoch": 28.40524017467249,
+ "grad_norm": 0.019539108499884605,
+ "learning_rate": 0.0006,
+ "loss": 4.236655235290527,
+ "step": 2045
+ },
+ {
+ "epoch": 28.419213973799128,
+ "grad_norm": 0.01942690648138523,
+ "learning_rate": 0.0006,
+ "loss": 4.251850605010986,
+ "step": 2046
+ },
+ {
+ "epoch": 28.433187772925763,
+ "grad_norm": 0.019667886197566986,
+ "learning_rate": 0.0006,
+ "loss": 4.222312927246094,
+ "step": 2047
+ },
+ {
+ "epoch": 28.4471615720524,
+ "grad_norm": 0.02029012329876423,
+ "learning_rate": 0.0006,
+ "loss": 4.246252059936523,
+ "step": 2048
+ },
+ {
+ "epoch": 28.46113537117904,
+ "grad_norm": 0.01784469373524189,
+ "learning_rate": 0.0006,
+ "loss": 4.192128658294678,
+ "step": 2049
+ },
+ {
+ "epoch": 28.475109170305675,
+ "grad_norm": 0.0160287544131279,
+ "learning_rate": 0.0006,
+ "loss": 4.208607196807861,
+ "step": 2050
+ },
+ {
+ "epoch": 28.489082969432314,
+ "grad_norm": 0.015348346903920174,
+ "learning_rate": 0.0006,
+ "loss": 4.236598014831543,
+ "step": 2051
+ },
+ {
+ "epoch": 28.503056768558952,
+ "grad_norm": 0.015198206529021263,
+ "learning_rate": 0.0006,
+ "loss": 4.235552787780762,
+ "step": 2052
+ },
+ {
+ "epoch": 28.51703056768559,
+ "grad_norm": 0.01523754745721817,
+ "learning_rate": 0.0006,
+ "loss": 4.1941728591918945,
+ "step": 2053
+ },
+ {
+ "epoch": 28.531004366812226,
+ "grad_norm": 0.015559614636003971,
+ "learning_rate": 0.0006,
+ "loss": 4.294757843017578,
+ "step": 2054
+ },
+ {
+ "epoch": 28.544978165938865,
+ "grad_norm": 0.01568358950316906,
+ "learning_rate": 0.0006,
+ "loss": 4.213065147399902,
+ "step": 2055
+ },
+ {
+ "epoch": 28.558951965065503,
+ "grad_norm": 0.015336516313254833,
+ "learning_rate": 0.0006,
+ "loss": 4.240293979644775,
+ "step": 2056
+ },
+ {
+ "epoch": 28.57292576419214,
+ "grad_norm": 0.015041496604681015,
+ "learning_rate": 0.0006,
+ "loss": 4.273397445678711,
+ "step": 2057
+ },
+ {
+ "epoch": 28.586899563318777,
+ "grad_norm": 0.01610128954052925,
+ "learning_rate": 0.0006,
+ "loss": 4.125369071960449,
+ "step": 2058
+ },
+ {
+ "epoch": 28.600873362445416,
+ "grad_norm": 0.015974976122379303,
+ "learning_rate": 0.0006,
+ "loss": 4.133459091186523,
+ "step": 2059
+ },
+ {
+ "epoch": 28.61484716157205,
+ "grad_norm": 0.015936799347400665,
+ "learning_rate": 0.0006,
+ "loss": 4.306667327880859,
+ "step": 2060
+ },
+ {
+ "epoch": 28.62882096069869,
+ "grad_norm": 0.01795247197151184,
+ "learning_rate": 0.0006,
+ "loss": 4.205156326293945,
+ "step": 2061
+ },
+ {
+ "epoch": 28.64279475982533,
+ "grad_norm": 0.019639814272522926,
+ "learning_rate": 0.0006,
+ "loss": 4.277614593505859,
+ "step": 2062
+ },
+ {
+ "epoch": 28.656768558951963,
+ "grad_norm": 0.01722300611436367,
+ "learning_rate": 0.0006,
+ "loss": 4.191032886505127,
+ "step": 2063
+ },
+ {
+ "epoch": 28.670742358078602,
+ "grad_norm": 0.01612633652985096,
+ "learning_rate": 0.0006,
+ "loss": 4.20717191696167,
+ "step": 2064
+ },
+ {
+ "epoch": 28.68471615720524,
+ "grad_norm": 0.01603279449045658,
+ "learning_rate": 0.0006,
+ "loss": 4.228884220123291,
+ "step": 2065
+ },
+ {
+ "epoch": 28.69868995633188,
+ "grad_norm": 0.016405927017331123,
+ "learning_rate": 0.0006,
+ "loss": 4.201292991638184,
+ "step": 2066
+ },
+ {
+ "epoch": 28.712663755458514,
+ "grad_norm": 0.016644183546304703,
+ "learning_rate": 0.0006,
+ "loss": 4.239282131195068,
+ "step": 2067
+ },
+ {
+ "epoch": 28.726637554585153,
+ "grad_norm": 0.0175587497651577,
+ "learning_rate": 0.0006,
+ "loss": 4.289038181304932,
+ "step": 2068
+ },
+ {
+ "epoch": 28.74061135371179,
+ "grad_norm": 0.018223397433757782,
+ "learning_rate": 0.0006,
+ "loss": 4.2266998291015625,
+ "step": 2069
+ },
+ {
+ "epoch": 28.754585152838427,
+ "grad_norm": 0.018680868670344353,
+ "learning_rate": 0.0006,
+ "loss": 4.348849296569824,
+ "step": 2070
+ },
+ {
+ "epoch": 28.768558951965066,
+ "grad_norm": 0.016726728528738022,
+ "learning_rate": 0.0006,
+ "loss": 4.292080879211426,
+ "step": 2071
+ },
+ {
+ "epoch": 28.782532751091704,
+ "grad_norm": 0.017386795952916145,
+ "learning_rate": 0.0006,
+ "loss": 4.208870887756348,
+ "step": 2072
+ },
+ {
+ "epoch": 28.79650655021834,
+ "grad_norm": 0.01865178905427456,
+ "learning_rate": 0.0006,
+ "loss": 4.054150581359863,
+ "step": 2073
+ },
+ {
+ "epoch": 28.810480349344978,
+ "grad_norm": 0.01776815392076969,
+ "learning_rate": 0.0006,
+ "loss": 4.202404499053955,
+ "step": 2074
+ },
+ {
+ "epoch": 28.824454148471617,
+ "grad_norm": 0.01643497310578823,
+ "learning_rate": 0.0006,
+ "loss": 4.131855010986328,
+ "step": 2075
+ },
+ {
+ "epoch": 28.83842794759825,
+ "grad_norm": 0.01574273779988289,
+ "learning_rate": 0.0006,
+ "loss": 4.296788692474365,
+ "step": 2076
+ },
+ {
+ "epoch": 28.85240174672489,
+ "grad_norm": 0.015613908879458904,
+ "learning_rate": 0.0006,
+ "loss": 4.210949897766113,
+ "step": 2077
+ },
+ {
+ "epoch": 28.86637554585153,
+ "grad_norm": 0.015543212182819843,
+ "learning_rate": 0.0006,
+ "loss": 4.218136310577393,
+ "step": 2078
+ },
+ {
+ "epoch": 28.880349344978168,
+ "grad_norm": 0.014681251719594002,
+ "learning_rate": 0.0006,
+ "loss": 4.19401741027832,
+ "step": 2079
+ },
+ {
+ "epoch": 28.894323144104803,
+ "grad_norm": 0.014393828809261322,
+ "learning_rate": 0.0006,
+ "loss": 4.198468208312988,
+ "step": 2080
+ },
+ {
+ "epoch": 28.90829694323144,
+ "grad_norm": 0.01580170728266239,
+ "learning_rate": 0.0006,
+ "loss": 4.224969863891602,
+ "step": 2081
+ },
+ {
+ "epoch": 28.92227074235808,
+ "grad_norm": 0.01484165620058775,
+ "learning_rate": 0.0006,
+ "loss": 4.12236213684082,
+ "step": 2082
+ },
+ {
+ "epoch": 28.936244541484715,
+ "grad_norm": 0.01497623696923256,
+ "learning_rate": 0.0006,
+ "loss": 4.24141788482666,
+ "step": 2083
+ },
+ {
+ "epoch": 28.950218340611354,
+ "grad_norm": 0.015080046840012074,
+ "learning_rate": 0.0006,
+ "loss": 4.284151077270508,
+ "step": 2084
+ },
+ {
+ "epoch": 28.964192139737992,
+ "grad_norm": 0.016207212582230568,
+ "learning_rate": 0.0006,
+ "loss": 4.298693656921387,
+ "step": 2085
+ },
+ {
+ "epoch": 28.978165938864628,
+ "grad_norm": 0.015474417246878147,
+ "learning_rate": 0.0006,
+ "loss": 4.2412261962890625,
+ "step": 2086
+ },
+ {
+ "epoch": 28.992139737991266,
+ "grad_norm": 0.01515351701527834,
+ "learning_rate": 0.0006,
+ "loss": 4.3054423332214355,
+ "step": 2087
+ },
+ {
+ "epoch": 29.0,
+ "grad_norm": 0.016411345452070236,
+ "learning_rate": 0.0006,
+ "loss": 4.104830741882324,
+ "step": 2088
+ },
+ {
+ "epoch": 29.0,
+ "eval_loss": 4.661567687988281,
+ "eval_runtime": 57.1244,
+ "eval_samples_per_second": 42.749,
+ "eval_steps_per_second": 1.348,
+ "step": 2088
+ },
+ {
+ "epoch": 29.01397379912664,
+ "grad_norm": 0.01665342226624489,
+ "learning_rate": 0.0006,
+ "loss": 4.311391830444336,
+ "step": 2089
+ },
+ {
+ "epoch": 29.027947598253274,
+ "grad_norm": 0.018053214997053146,
+ "learning_rate": 0.0006,
+ "loss": 4.135597229003906,
+ "step": 2090
+ },
+ {
+ "epoch": 29.041921397379912,
+ "grad_norm": 0.018213748931884766,
+ "learning_rate": 0.0006,
+ "loss": 4.303299903869629,
+ "step": 2091
+ },
+ {
+ "epoch": 29.05589519650655,
+ "grad_norm": 0.018460242077708244,
+ "learning_rate": 0.0006,
+ "loss": 4.20850944519043,
+ "step": 2092
+ },
+ {
+ "epoch": 29.069868995633186,
+ "grad_norm": 0.018365809693932533,
+ "learning_rate": 0.0006,
+ "loss": 4.275847434997559,
+ "step": 2093
+ },
+ {
+ "epoch": 29.083842794759825,
+ "grad_norm": 0.018346186727285385,
+ "learning_rate": 0.0006,
+ "loss": 4.202091217041016,
+ "step": 2094
+ },
+ {
+ "epoch": 29.097816593886463,
+ "grad_norm": 0.017920587211847305,
+ "learning_rate": 0.0006,
+ "loss": 4.182092666625977,
+ "step": 2095
+ },
+ {
+ "epoch": 29.111790393013102,
+ "grad_norm": 0.01812003180384636,
+ "learning_rate": 0.0006,
+ "loss": 4.259721755981445,
+ "step": 2096
+ },
+ {
+ "epoch": 29.125764192139737,
+ "grad_norm": 0.01700986735522747,
+ "learning_rate": 0.0006,
+ "loss": 4.067573070526123,
+ "step": 2097
+ },
+ {
+ "epoch": 29.139737991266376,
+ "grad_norm": 0.018069760873913765,
+ "learning_rate": 0.0006,
+ "loss": 4.234400749206543,
+ "step": 2098
+ },
+ {
+ "epoch": 29.153711790393015,
+ "grad_norm": 0.01905428245663643,
+ "learning_rate": 0.0006,
+ "loss": 4.16609001159668,
+ "step": 2099
+ },
+ {
+ "epoch": 29.16768558951965,
+ "grad_norm": 0.018711242824792862,
+ "learning_rate": 0.0006,
+ "loss": 4.249088764190674,
+ "step": 2100
+ },
+ {
+ "epoch": 29.18165938864629,
+ "grad_norm": 0.018721306696534157,
+ "learning_rate": 0.0006,
+ "loss": 4.14540958404541,
+ "step": 2101
+ },
+ {
+ "epoch": 29.195633187772927,
+ "grad_norm": 0.020139874890446663,
+ "learning_rate": 0.0006,
+ "loss": 4.049248695373535,
+ "step": 2102
+ },
+ {
+ "epoch": 29.209606986899562,
+ "grad_norm": 0.020849574357271194,
+ "learning_rate": 0.0006,
+ "loss": 4.235224723815918,
+ "step": 2103
+ },
+ {
+ "epoch": 29.2235807860262,
+ "grad_norm": 0.018693549558520317,
+ "learning_rate": 0.0006,
+ "loss": 4.232961177825928,
+ "step": 2104
+ },
+ {
+ "epoch": 29.23755458515284,
+ "grad_norm": 0.017889857292175293,
+ "learning_rate": 0.0006,
+ "loss": 4.171422004699707,
+ "step": 2105
+ },
+ {
+ "epoch": 29.251528384279474,
+ "grad_norm": 0.016690224409103394,
+ "learning_rate": 0.0006,
+ "loss": 4.244717597961426,
+ "step": 2106
+ },
+ {
+ "epoch": 29.265502183406113,
+ "grad_norm": 0.01708616502583027,
+ "learning_rate": 0.0006,
+ "loss": 4.125068664550781,
+ "step": 2107
+ },
+ {
+ "epoch": 29.27947598253275,
+ "grad_norm": 0.017928237095475197,
+ "learning_rate": 0.0006,
+ "loss": 4.208292007446289,
+ "step": 2108
+ },
+ {
+ "epoch": 29.29344978165939,
+ "grad_norm": 0.018250394612550735,
+ "learning_rate": 0.0006,
+ "loss": 4.118630886077881,
+ "step": 2109
+ },
+ {
+ "epoch": 29.307423580786025,
+ "grad_norm": 0.020160246640443802,
+ "learning_rate": 0.0006,
+ "loss": 4.206025123596191,
+ "step": 2110
+ },
+ {
+ "epoch": 29.321397379912664,
+ "grad_norm": 0.020045453682541847,
+ "learning_rate": 0.0006,
+ "loss": 4.2758684158325195,
+ "step": 2111
+ },
+ {
+ "epoch": 29.335371179039303,
+ "grad_norm": 0.019556907936930656,
+ "learning_rate": 0.0006,
+ "loss": 4.188272953033447,
+ "step": 2112
+ },
+ {
+ "epoch": 29.349344978165938,
+ "grad_norm": 0.019185200333595276,
+ "learning_rate": 0.0006,
+ "loss": 4.17451286315918,
+ "step": 2113
+ },
+ {
+ "epoch": 29.363318777292577,
+ "grad_norm": 0.01816629432141781,
+ "learning_rate": 0.0006,
+ "loss": 4.165246963500977,
+ "step": 2114
+ },
+ {
+ "epoch": 29.377292576419215,
+ "grad_norm": 0.01865146867930889,
+ "learning_rate": 0.0006,
+ "loss": 4.148184776306152,
+ "step": 2115
+ },
+ {
+ "epoch": 29.39126637554585,
+ "grad_norm": 0.018338464200496674,
+ "learning_rate": 0.0006,
+ "loss": 4.288936614990234,
+ "step": 2116
+ },
+ {
+ "epoch": 29.40524017467249,
+ "grad_norm": 0.018986187875270844,
+ "learning_rate": 0.0006,
+ "loss": 4.183379650115967,
+ "step": 2117
+ },
+ {
+ "epoch": 29.419213973799128,
+ "grad_norm": 0.018409637734293938,
+ "learning_rate": 0.0006,
+ "loss": 4.057574272155762,
+ "step": 2118
+ },
+ {
+ "epoch": 29.433187772925763,
+ "grad_norm": 0.016664542257785797,
+ "learning_rate": 0.0006,
+ "loss": 4.18517541885376,
+ "step": 2119
+ },
+ {
+ "epoch": 29.4471615720524,
+ "grad_norm": 0.016486089676618576,
+ "learning_rate": 0.0006,
+ "loss": 4.179367542266846,
+ "step": 2120
+ },
+ {
+ "epoch": 29.46113537117904,
+ "grad_norm": 0.01787523552775383,
+ "learning_rate": 0.0006,
+ "loss": 4.184126377105713,
+ "step": 2121
+ },
+ {
+ "epoch": 29.475109170305675,
+ "grad_norm": 0.016574613749980927,
+ "learning_rate": 0.0006,
+ "loss": 4.290170192718506,
+ "step": 2122
+ },
+ {
+ "epoch": 29.489082969432314,
+ "grad_norm": 0.01643376611173153,
+ "learning_rate": 0.0006,
+ "loss": 4.052913665771484,
+ "step": 2123
+ },
+ {
+ "epoch": 29.503056768558952,
+ "grad_norm": 0.016503969207406044,
+ "learning_rate": 0.0006,
+ "loss": 4.335184097290039,
+ "step": 2124
+ },
+ {
+ "epoch": 29.51703056768559,
+ "grad_norm": 0.017596479505300522,
+ "learning_rate": 0.0006,
+ "loss": 4.165539741516113,
+ "step": 2125
+ },
+ {
+ "epoch": 29.531004366812226,
+ "grad_norm": 0.017468582838773727,
+ "learning_rate": 0.0006,
+ "loss": 4.315964698791504,
+ "step": 2126
+ },
+ {
+ "epoch": 29.544978165938865,
+ "grad_norm": 0.01724618673324585,
+ "learning_rate": 0.0006,
+ "loss": 4.243681907653809,
+ "step": 2127
+ },
+ {
+ "epoch": 29.558951965065503,
+ "grad_norm": 0.018384616822004318,
+ "learning_rate": 0.0006,
+ "loss": 4.112447261810303,
+ "step": 2128
+ },
+ {
+ "epoch": 29.57292576419214,
+ "grad_norm": 0.01824074238538742,
+ "learning_rate": 0.0006,
+ "loss": 4.236065864562988,
+ "step": 2129
+ },
+ {
+ "epoch": 29.586899563318777,
+ "grad_norm": 0.01625971868634224,
+ "learning_rate": 0.0006,
+ "loss": 4.18143892288208,
+ "step": 2130
+ },
+ {
+ "epoch": 29.600873362445416,
+ "grad_norm": 0.015537494793534279,
+ "learning_rate": 0.0006,
+ "loss": 4.225147247314453,
+ "step": 2131
+ },
+ {
+ "epoch": 29.61484716157205,
+ "grad_norm": 0.015818974003195763,
+ "learning_rate": 0.0006,
+ "loss": 4.3100175857543945,
+ "step": 2132
+ },
+ {
+ "epoch": 29.62882096069869,
+ "grad_norm": 0.015909671783447266,
+ "learning_rate": 0.0006,
+ "loss": 4.169775009155273,
+ "step": 2133
+ },
+ {
+ "epoch": 29.64279475982533,
+ "grad_norm": 0.01769658550620079,
+ "learning_rate": 0.0006,
+ "loss": 4.278904914855957,
+ "step": 2134
+ },
+ {
+ "epoch": 29.656768558951963,
+ "grad_norm": 0.015641704201698303,
+ "learning_rate": 0.0006,
+ "loss": 4.2426228523254395,
+ "step": 2135
+ },
+ {
+ "epoch": 29.670742358078602,
+ "grad_norm": 0.01565658301115036,
+ "learning_rate": 0.0006,
+ "loss": 4.232687950134277,
+ "step": 2136
+ },
+ {
+ "epoch": 29.68471615720524,
+ "grad_norm": 0.016485940665006638,
+ "learning_rate": 0.0006,
+ "loss": 4.209839820861816,
+ "step": 2137
+ },
+ {
+ "epoch": 29.69868995633188,
+ "grad_norm": 0.017622729763388634,
+ "learning_rate": 0.0006,
+ "loss": 4.196017265319824,
+ "step": 2138
+ },
+ {
+ "epoch": 29.712663755458514,
+ "grad_norm": 0.018320564180612564,
+ "learning_rate": 0.0006,
+ "loss": 4.225683689117432,
+ "step": 2139
+ },
+ {
+ "epoch": 29.726637554585153,
+ "grad_norm": 0.018140794709324837,
+ "learning_rate": 0.0006,
+ "loss": 4.311944961547852,
+ "step": 2140
+ },
+ {
+ "epoch": 29.74061135371179,
+ "grad_norm": 0.016227100044488907,
+ "learning_rate": 0.0006,
+ "loss": 4.230257511138916,
+ "step": 2141
+ },
+ {
+ "epoch": 29.754585152838427,
+ "grad_norm": 0.015759488567709923,
+ "learning_rate": 0.0006,
+ "loss": 4.119174003601074,
+ "step": 2142
+ },
+ {
+ "epoch": 29.768558951965066,
+ "grad_norm": 0.01757766678929329,
+ "learning_rate": 0.0006,
+ "loss": 4.165826797485352,
+ "step": 2143
+ },
+ {
+ "epoch": 29.782532751091704,
+ "grad_norm": 0.019661923870444298,
+ "learning_rate": 0.0006,
+ "loss": 4.392203330993652,
+ "step": 2144
+ },
+ {
+ "epoch": 29.79650655021834,
+ "grad_norm": 0.019478755071759224,
+ "learning_rate": 0.0006,
+ "loss": 4.348371982574463,
+ "step": 2145
+ },
+ {
+ "epoch": 29.810480349344978,
+ "grad_norm": 0.019149569794535637,
+ "learning_rate": 0.0006,
+ "loss": 4.34104061126709,
+ "step": 2146
+ },
+ {
+ "epoch": 29.824454148471617,
+ "grad_norm": 0.017707521095871925,
+ "learning_rate": 0.0006,
+ "loss": 4.228226661682129,
+ "step": 2147
+ },
+ {
+ "epoch": 29.83842794759825,
+ "grad_norm": 0.015615378506481647,
+ "learning_rate": 0.0006,
+ "loss": 4.175541877746582,
+ "step": 2148
+ },
+ {
+ "epoch": 29.85240174672489,
+ "grad_norm": 0.016111129894852638,
+ "learning_rate": 0.0006,
+ "loss": 4.267936706542969,
+ "step": 2149
+ },
+ {
+ "epoch": 29.86637554585153,
+ "grad_norm": 0.01528779324144125,
+ "learning_rate": 0.0006,
+ "loss": 4.186014175415039,
+ "step": 2150
+ },
+ {
+ "epoch": 29.880349344978168,
+ "grad_norm": 0.014925646595656872,
+ "learning_rate": 0.0006,
+ "loss": 4.174860000610352,
+ "step": 2151
+ },
+ {
+ "epoch": 29.894323144104803,
+ "grad_norm": 0.016917673870921135,
+ "learning_rate": 0.0006,
+ "loss": 4.33908748626709,
+ "step": 2152
+ },
+ {
+ "epoch": 29.90829694323144,
+ "grad_norm": 0.016007075086236,
+ "learning_rate": 0.0006,
+ "loss": 4.266141414642334,
+ "step": 2153
+ },
+ {
+ "epoch": 29.92227074235808,
+ "grad_norm": 0.016178956255316734,
+ "learning_rate": 0.0006,
+ "loss": 4.271215915679932,
+ "step": 2154
+ },
+ {
+ "epoch": 29.936244541484715,
+ "grad_norm": 0.016180871054530144,
+ "learning_rate": 0.0006,
+ "loss": 4.278801441192627,
+ "step": 2155
+ },
+ {
+ "epoch": 29.950218340611354,
+ "grad_norm": 0.015335503034293652,
+ "learning_rate": 0.0006,
+ "loss": 4.194978713989258,
+ "step": 2156
+ },
+ {
+ "epoch": 29.964192139737992,
+ "grad_norm": 0.016580764204263687,
+ "learning_rate": 0.0006,
+ "loss": 4.290809631347656,
+ "step": 2157
+ },
+ {
+ "epoch": 29.978165938864628,
+ "grad_norm": 0.01623128168284893,
+ "learning_rate": 0.0006,
+ "loss": 4.205451488494873,
+ "step": 2158
+ },
+ {
+ "epoch": 29.992139737991266,
+ "grad_norm": 0.014904248528182507,
+ "learning_rate": 0.0006,
+ "loss": 4.24924373626709,
+ "step": 2159
+ },
+ {
+ "epoch": 30.0,
+ "grad_norm": 0.017089299857616425,
+ "learning_rate": 0.0006,
+ "loss": 4.222072601318359,
+ "step": 2160
+ },
+ {
+ "epoch": 30.0,
+ "eval_loss": 4.585411071777344,
+ "eval_runtime": 57.1164,
+ "eval_samples_per_second": 42.755,
+ "eval_steps_per_second": 1.348,
+ "step": 2160
+ },
+ {
+ "epoch": 30.01397379912664,
+ "grad_norm": 0.017226383090019226,
+ "learning_rate": 0.0006,
+ "loss": 4.19467830657959,
+ "step": 2161
+ },
+ {
+ "epoch": 30.027947598253274,
+ "grad_norm": 0.01635066606104374,
+ "learning_rate": 0.0006,
+ "loss": 4.176875591278076,
+ "step": 2162
+ },
+ {
+ "epoch": 30.041921397379912,
+ "grad_norm": 0.01735362410545349,
+ "learning_rate": 0.0006,
+ "loss": 4.13667106628418,
+ "step": 2163
+ },
+ {
+ "epoch": 30.05589519650655,
+ "grad_norm": 0.017047109082341194,
+ "learning_rate": 0.0006,
+ "loss": 4.136329650878906,
+ "step": 2164
+ },
+ {
+ "epoch": 30.069868995633186,
+ "grad_norm": 0.018116053193807602,
+ "learning_rate": 0.0006,
+ "loss": 4.228706359863281,
+ "step": 2165
+ },
+ {
+ "epoch": 30.083842794759825,
+ "grad_norm": 0.01791389286518097,
+ "learning_rate": 0.0006,
+ "loss": 4.1597747802734375,
+ "step": 2166
+ },
+ {
+ "epoch": 30.097816593886463,
+ "grad_norm": 0.017793502658605576,
+ "learning_rate": 0.0006,
+ "loss": 4.0807976722717285,
+ "step": 2167
+ },
+ {
+ "epoch": 30.111790393013102,
+ "grad_norm": 0.01827503927052021,
+ "learning_rate": 0.0006,
+ "loss": 4.170646667480469,
+ "step": 2168
+ },
+ {
+ "epoch": 30.125764192139737,
+ "grad_norm": 0.017228564247488976,
+ "learning_rate": 0.0006,
+ "loss": 4.146934509277344,
+ "step": 2169
+ },
+ {
+ "epoch": 30.139737991266376,
+ "grad_norm": 0.01757437363266945,
+ "learning_rate": 0.0006,
+ "loss": 4.1859130859375,
+ "step": 2170
+ },
+ {
+ "epoch": 30.153711790393015,
+ "grad_norm": 0.01701093092560768,
+ "learning_rate": 0.0006,
+ "loss": 4.126981258392334,
+ "step": 2171
+ },
+ {
+ "epoch": 30.16768558951965,
+ "grad_norm": 0.015220489352941513,
+ "learning_rate": 0.0006,
+ "loss": 4.164018630981445,
+ "step": 2172
+ },
+ {
+ "epoch": 30.18165938864629,
+ "grad_norm": 0.016254238784313202,
+ "learning_rate": 0.0006,
+ "loss": 4.258275508880615,
+ "step": 2173
+ },
+ {
+ "epoch": 30.195633187772927,
+ "grad_norm": 0.016609614714980125,
+ "learning_rate": 0.0006,
+ "loss": 4.222233772277832,
+ "step": 2174
+ },
+ {
+ "epoch": 30.209606986899562,
+ "grad_norm": 0.016531318426132202,
+ "learning_rate": 0.0006,
+ "loss": 4.144489288330078,
+ "step": 2175
+ },
+ {
+ "epoch": 30.2235807860262,
+ "grad_norm": 0.01567668654024601,
+ "learning_rate": 0.0006,
+ "loss": 4.181916236877441,
+ "step": 2176
+ },
+ {
+ "epoch": 30.23755458515284,
+ "grad_norm": 0.017195681110024452,
+ "learning_rate": 0.0006,
+ "loss": 4.189693927764893,
+ "step": 2177
+ },
+ {
+ "epoch": 30.251528384279474,
+ "grad_norm": 0.016882948577404022,
+ "learning_rate": 0.0006,
+ "loss": 4.21604061126709,
+ "step": 2178
+ },
+ {
+ "epoch": 30.265502183406113,
+ "grad_norm": 0.016261622309684753,
+ "learning_rate": 0.0006,
+ "loss": 4.227639198303223,
+ "step": 2179
+ },
+ {
+ "epoch": 30.27947598253275,
+ "grad_norm": 0.016411812976002693,
+ "learning_rate": 0.0006,
+ "loss": 4.130362510681152,
+ "step": 2180
+ },
+ {
+ "epoch": 30.29344978165939,
+ "grad_norm": 0.01732019893825054,
+ "learning_rate": 0.0006,
+ "loss": 4.230443000793457,
+ "step": 2181
+ },
+ {
+ "epoch": 30.307423580786025,
+ "grad_norm": 0.01825905591249466,
+ "learning_rate": 0.0006,
+ "loss": 4.175817489624023,
+ "step": 2182
+ },
+ {
+ "epoch": 30.321397379912664,
+ "grad_norm": 0.01887713558971882,
+ "learning_rate": 0.0006,
+ "loss": 4.169020652770996,
+ "step": 2183
+ },
+ {
+ "epoch": 30.335371179039303,
+ "grad_norm": 0.019391661509871483,
+ "learning_rate": 0.0006,
+ "loss": 4.176318645477295,
+ "step": 2184
+ },
+ {
+ "epoch": 30.349344978165938,
+ "grad_norm": 0.020840002223849297,
+ "learning_rate": 0.0006,
+ "loss": 4.174228668212891,
+ "step": 2185
+ },
+ {
+ "epoch": 30.363318777292577,
+ "grad_norm": 0.020193178206682205,
+ "learning_rate": 0.0006,
+ "loss": 4.127919673919678,
+ "step": 2186
+ },
+ {
+ "epoch": 30.377292576419215,
+ "grad_norm": 0.018449561670422554,
+ "learning_rate": 0.0006,
+ "loss": 4.229487895965576,
+ "step": 2187
+ },
+ {
+ "epoch": 30.39126637554585,
+ "grad_norm": 0.01711983233690262,
+ "learning_rate": 0.0006,
+ "loss": 4.201651573181152,
+ "step": 2188
+ },
+ {
+ "epoch": 30.40524017467249,
+ "grad_norm": 0.017293395474553108,
+ "learning_rate": 0.0006,
+ "loss": 4.168704032897949,
+ "step": 2189
+ },
+ {
+ "epoch": 30.419213973799128,
+ "grad_norm": 0.018526358529925346,
+ "learning_rate": 0.0006,
+ "loss": 4.171864032745361,
+ "step": 2190
+ },
+ {
+ "epoch": 30.433187772925763,
+ "grad_norm": 0.019287291914224625,
+ "learning_rate": 0.0006,
+ "loss": 4.039534568786621,
+ "step": 2191
+ },
+ {
+ "epoch": 30.4471615720524,
+ "grad_norm": 0.016902444884181023,
+ "learning_rate": 0.0006,
+ "loss": 4.282787799835205,
+ "step": 2192
+ },
+ {
+ "epoch": 30.46113537117904,
+ "grad_norm": 0.01656750962138176,
+ "learning_rate": 0.0006,
+ "loss": 4.213098526000977,
+ "step": 2193
+ },
+ {
+ "epoch": 30.475109170305675,
+ "grad_norm": 0.016805065795779228,
+ "learning_rate": 0.0006,
+ "loss": 4.1583476066589355,
+ "step": 2194
+ },
+ {
+ "epoch": 30.489082969432314,
+ "grad_norm": 0.017166638746857643,
+ "learning_rate": 0.0006,
+ "loss": 4.188634395599365,
+ "step": 2195
+ },
+ {
+ "epoch": 30.503056768558952,
+ "grad_norm": 0.01592106744647026,
+ "learning_rate": 0.0006,
+ "loss": 4.234043121337891,
+ "step": 2196
+ },
+ {
+ "epoch": 30.51703056768559,
+ "grad_norm": 0.015270989388227463,
+ "learning_rate": 0.0006,
+ "loss": 4.301420211791992,
+ "step": 2197
+ },
+ {
+ "epoch": 30.531004366812226,
+ "grad_norm": 0.015194724313914776,
+ "learning_rate": 0.0006,
+ "loss": 4.18703031539917,
+ "step": 2198
+ },
+ {
+ "epoch": 30.544978165938865,
+ "grad_norm": 0.015587746165692806,
+ "learning_rate": 0.0006,
+ "loss": 4.138928413391113,
+ "step": 2199
+ },
+ {
+ "epoch": 30.558951965065503,
+ "grad_norm": 0.016029570251703262,
+ "learning_rate": 0.0006,
+ "loss": 4.290863990783691,
+ "step": 2200
+ },
+ {
+ "epoch": 30.57292576419214,
+ "grad_norm": 0.015651309862732887,
+ "learning_rate": 0.0006,
+ "loss": 4.272062301635742,
+ "step": 2201
+ },
+ {
+ "epoch": 30.586899563318777,
+ "grad_norm": 0.015992306172847748,
+ "learning_rate": 0.0006,
+ "loss": 4.2487406730651855,
+ "step": 2202
+ },
+ {
+ "epoch": 30.600873362445416,
+ "grad_norm": 0.014969157986342907,
+ "learning_rate": 0.0006,
+ "loss": 4.115347862243652,
+ "step": 2203
+ },
+ {
+ "epoch": 30.61484716157205,
+ "grad_norm": 0.014639027416706085,
+ "learning_rate": 0.0006,
+ "loss": 4.2207489013671875,
+ "step": 2204
+ },
+ {
+ "epoch": 30.62882096069869,
+ "grad_norm": 0.015059413388371468,
+ "learning_rate": 0.0006,
+ "loss": 4.274044990539551,
+ "step": 2205
+ },
+ {
+ "epoch": 30.64279475982533,
+ "grad_norm": 0.0166641678661108,
+ "learning_rate": 0.0006,
+ "loss": 4.191373348236084,
+ "step": 2206
+ },
+ {
+ "epoch": 30.656768558951963,
+ "grad_norm": 0.017806345596909523,
+ "learning_rate": 0.0006,
+ "loss": 4.139800548553467,
+ "step": 2207
+ },
+ {
+ "epoch": 30.670742358078602,
+ "grad_norm": 0.018712421879172325,
+ "learning_rate": 0.0006,
+ "loss": 4.035965442657471,
+ "step": 2208
+ },
+ {
+ "epoch": 30.68471615720524,
+ "grad_norm": 0.018457243219017982,
+ "learning_rate": 0.0006,
+ "loss": 4.2579240798950195,
+ "step": 2209
+ },
+ {
+ "epoch": 30.69868995633188,
+ "grad_norm": 0.01920999586582184,
+ "learning_rate": 0.0006,
+ "loss": 4.131112098693848,
+ "step": 2210
+ },
+ {
+ "epoch": 30.712663755458514,
+ "grad_norm": 0.01729944534599781,
+ "learning_rate": 0.0006,
+ "loss": 4.254059314727783,
+ "step": 2211
+ },
+ {
+ "epoch": 30.726637554585153,
+ "grad_norm": 0.016914231702685356,
+ "learning_rate": 0.0006,
+ "loss": 4.233428955078125,
+ "step": 2212
+ },
+ {
+ "epoch": 30.74061135371179,
+ "grad_norm": 0.01913098618388176,
+ "learning_rate": 0.0006,
+ "loss": 4.23525333404541,
+ "step": 2213
+ },
+ {
+ "epoch": 30.754585152838427,
+ "grad_norm": 0.019150281324982643,
+ "learning_rate": 0.0006,
+ "loss": 4.213695049285889,
+ "step": 2214
+ },
+ {
+ "epoch": 30.768558951965066,
+ "grad_norm": 0.018947452306747437,
+ "learning_rate": 0.0006,
+ "loss": 4.151697158813477,
+ "step": 2215
+ },
+ {
+ "epoch": 30.782532751091704,
+ "grad_norm": 0.016058508306741714,
+ "learning_rate": 0.0006,
+ "loss": 4.220280170440674,
+ "step": 2216
+ },
+ {
+ "epoch": 30.79650655021834,
+ "grad_norm": 0.017488619312644005,
+ "learning_rate": 0.0006,
+ "loss": 4.154428482055664,
+ "step": 2217
+ },
+ {
+ "epoch": 30.810480349344978,
+ "grad_norm": 0.017355090007185936,
+ "learning_rate": 0.0006,
+ "loss": 4.256086349487305,
+ "step": 2218
+ },
+ {
+ "epoch": 30.824454148471617,
+ "grad_norm": 0.016636032611131668,
+ "learning_rate": 0.0006,
+ "loss": 4.227564334869385,
+ "step": 2219
+ },
+ {
+ "epoch": 30.83842794759825,
+ "grad_norm": 0.016273891553282738,
+ "learning_rate": 0.0006,
+ "loss": 4.200466156005859,
+ "step": 2220
+ },
+ {
+ "epoch": 30.85240174672489,
+ "grad_norm": 0.016939911991357803,
+ "learning_rate": 0.0006,
+ "loss": 4.1571221351623535,
+ "step": 2221
+ },
+ {
+ "epoch": 30.86637554585153,
+ "grad_norm": 0.018332522362470627,
+ "learning_rate": 0.0006,
+ "loss": 4.228398323059082,
+ "step": 2222
+ },
+ {
+ "epoch": 30.880349344978168,
+ "grad_norm": 0.019508114084601402,
+ "learning_rate": 0.0006,
+ "loss": 4.222960472106934,
+ "step": 2223
+ },
+ {
+ "epoch": 30.894323144104803,
+ "grad_norm": 0.018440278246998787,
+ "learning_rate": 0.0006,
+ "loss": 4.261051177978516,
+ "step": 2224
+ },
+ {
+ "epoch": 30.90829694323144,
+ "grad_norm": 0.018113229423761368,
+ "learning_rate": 0.0006,
+ "loss": 4.109235763549805,
+ "step": 2225
+ },
+ {
+ "epoch": 30.92227074235808,
+ "grad_norm": 0.015392083674669266,
+ "learning_rate": 0.0006,
+ "loss": 4.21950101852417,
+ "step": 2226
+ },
+ {
+ "epoch": 30.936244541484715,
+ "grad_norm": 0.01633290946483612,
+ "learning_rate": 0.0006,
+ "loss": 4.176053047180176,
+ "step": 2227
+ },
+ {
+ "epoch": 30.950218340611354,
+ "grad_norm": 0.016236383467912674,
+ "learning_rate": 0.0006,
+ "loss": 4.314309120178223,
+ "step": 2228
+ },
+ {
+ "epoch": 30.964192139737992,
+ "grad_norm": 0.015284373424947262,
+ "learning_rate": 0.0006,
+ "loss": 4.303133487701416,
+ "step": 2229
+ },
+ {
+ "epoch": 30.978165938864628,
+ "grad_norm": 0.015982897952198982,
+ "learning_rate": 0.0006,
+ "loss": 4.25739860534668,
+ "step": 2230
+ },
+ {
+ "epoch": 30.992139737991266,
+ "grad_norm": 0.016682198271155357,
+ "learning_rate": 0.0006,
+ "loss": 4.292210578918457,
+ "step": 2231
+ },
+ {
+ "epoch": 31.0,
+ "grad_norm": 0.017521467059850693,
+ "learning_rate": 0.0006,
+ "loss": 4.170637130737305,
+ "step": 2232
+ },
+ {
+ "epoch": 31.0,
+ "eval_loss": 4.642786979675293,
+ "eval_runtime": 57.029,
+ "eval_samples_per_second": 42.82,
+ "eval_steps_per_second": 1.35,
+ "step": 2232
+ },
+ {
+ "epoch": 31.01397379912664,
+ "grad_norm": 0.016607852652668953,
+ "learning_rate": 0.0006,
+ "loss": 4.106224536895752,
+ "step": 2233
+ },
+ {
+ "epoch": 31.027947598253274,
+ "grad_norm": 0.018448904156684875,
+ "learning_rate": 0.0006,
+ "loss": 4.224459648132324,
+ "step": 2234
+ },
+ {
+ "epoch": 31.041921397379912,
+ "grad_norm": 0.020481228828430176,
+ "learning_rate": 0.0006,
+ "loss": 4.305548667907715,
+ "step": 2235
+ },
+ {
+ "epoch": 31.05589519650655,
+ "grad_norm": 0.01999707892537117,
+ "learning_rate": 0.0006,
+ "loss": 4.253663539886475,
+ "step": 2236
+ },
+ {
+ "epoch": 31.069868995633186,
+ "grad_norm": 0.020428303629159927,
+ "learning_rate": 0.0006,
+ "loss": 4.166954040527344,
+ "step": 2237
+ },
+ {
+ "epoch": 31.083842794759825,
+ "grad_norm": 0.022608213126659393,
+ "learning_rate": 0.0006,
+ "loss": 4.172819137573242,
+ "step": 2238
+ },
+ {
+ "epoch": 31.097816593886463,
+ "grad_norm": 0.024669520556926727,
+ "learning_rate": 0.0006,
+ "loss": 4.106586456298828,
+ "step": 2239
+ },
+ {
+ "epoch": 31.111790393013102,
+ "grad_norm": 0.02247351035475731,
+ "learning_rate": 0.0006,
+ "loss": 4.160271644592285,
+ "step": 2240
+ },
+ {
+ "epoch": 31.125764192139737,
+ "grad_norm": 0.0207810141146183,
+ "learning_rate": 0.0006,
+ "loss": 4.272491931915283,
+ "step": 2241
+ },
+ {
+ "epoch": 31.139737991266376,
+ "grad_norm": 0.021576499566435814,
+ "learning_rate": 0.0006,
+ "loss": 4.171474456787109,
+ "step": 2242
+ },
+ {
+ "epoch": 31.153711790393015,
+ "grad_norm": 0.02337106689810753,
+ "learning_rate": 0.0006,
+ "loss": 4.094712734222412,
+ "step": 2243
+ },
+ {
+ "epoch": 31.16768558951965,
+ "grad_norm": 0.02219875156879425,
+ "learning_rate": 0.0006,
+ "loss": 4.1800994873046875,
+ "step": 2244
+ },
+ {
+ "epoch": 31.18165938864629,
+ "grad_norm": 0.022680504247546196,
+ "learning_rate": 0.0006,
+ "loss": 4.2054362297058105,
+ "step": 2245
+ },
+ {
+ "epoch": 31.195633187772927,
+ "grad_norm": 0.0208734143525362,
+ "learning_rate": 0.0006,
+ "loss": 4.06068229675293,
+ "step": 2246
+ },
+ {
+ "epoch": 31.209606986899562,
+ "grad_norm": 0.01811066083610058,
+ "learning_rate": 0.0006,
+ "loss": 4.2465128898620605,
+ "step": 2247
+ },
+ {
+ "epoch": 31.2235807860262,
+ "grad_norm": 0.018571637570858,
+ "learning_rate": 0.0006,
+ "loss": 4.191955089569092,
+ "step": 2248
+ },
+ {
+ "epoch": 31.23755458515284,
+ "grad_norm": 0.018598807975649834,
+ "learning_rate": 0.0006,
+ "loss": 4.200872421264648,
+ "step": 2249
+ },
+ {
+ "epoch": 31.251528384279474,
+ "grad_norm": 0.017577625811100006,
+ "learning_rate": 0.0006,
+ "loss": 4.148700714111328,
+ "step": 2250
+ },
+ {
+ "epoch": 31.265502183406113,
+ "grad_norm": 0.01817980408668518,
+ "learning_rate": 0.0006,
+ "loss": 4.256978988647461,
+ "step": 2251
+ },
+ {
+ "epoch": 31.27947598253275,
+ "grad_norm": 0.017983486875891685,
+ "learning_rate": 0.0006,
+ "loss": 4.175005912780762,
+ "step": 2252
+ },
+ {
+ "epoch": 31.29344978165939,
+ "grad_norm": 0.016751961782574654,
+ "learning_rate": 0.0006,
+ "loss": 4.237518310546875,
+ "step": 2253
+ },
+ {
+ "epoch": 31.307423580786025,
+ "grad_norm": 0.016407202929258347,
+ "learning_rate": 0.0006,
+ "loss": 4.197269439697266,
+ "step": 2254
+ },
+ {
+ "epoch": 31.321397379912664,
+ "grad_norm": 0.01578645594418049,
+ "learning_rate": 0.0006,
+ "loss": 4.179751396179199,
+ "step": 2255
+ },
+ {
+ "epoch": 31.335371179039303,
+ "grad_norm": 0.016001226380467415,
+ "learning_rate": 0.0006,
+ "loss": 4.08821964263916,
+ "step": 2256
+ },
+ {
+ "epoch": 31.349344978165938,
+ "grad_norm": 0.015122811309993267,
+ "learning_rate": 0.0006,
+ "loss": 4.047828674316406,
+ "step": 2257
+ },
+ {
+ "epoch": 31.363318777292577,
+ "grad_norm": 0.015399965457618237,
+ "learning_rate": 0.0006,
+ "loss": 4.1506524085998535,
+ "step": 2258
+ },
+ {
+ "epoch": 31.377292576419215,
+ "grad_norm": 0.01534605398774147,
+ "learning_rate": 0.0006,
+ "loss": 4.27952766418457,
+ "step": 2259
+ },
+ {
+ "epoch": 31.39126637554585,
+ "grad_norm": 0.015461381524801254,
+ "learning_rate": 0.0006,
+ "loss": 4.27102518081665,
+ "step": 2260
+ },
+ {
+ "epoch": 31.40524017467249,
+ "grad_norm": 0.016769153997302055,
+ "learning_rate": 0.0006,
+ "loss": 4.191592216491699,
+ "step": 2261
+ },
+ {
+ "epoch": 31.419213973799128,
+ "grad_norm": 0.016557959839701653,
+ "learning_rate": 0.0006,
+ "loss": 4.159705638885498,
+ "step": 2262
+ },
+ {
+ "epoch": 31.433187772925763,
+ "grad_norm": 0.01659569889307022,
+ "learning_rate": 0.0006,
+ "loss": 4.2136664390563965,
+ "step": 2263
+ },
+ {
+ "epoch": 31.4471615720524,
+ "grad_norm": 0.017101937904953957,
+ "learning_rate": 0.0006,
+ "loss": 4.2716498374938965,
+ "step": 2264
+ },
+ {
+ "epoch": 31.46113537117904,
+ "grad_norm": 0.01722985878586769,
+ "learning_rate": 0.0006,
+ "loss": 4.300534248352051,
+ "step": 2265
+ },
+ {
+ "epoch": 31.475109170305675,
+ "grad_norm": 0.016800185665488243,
+ "learning_rate": 0.0006,
+ "loss": 4.113240718841553,
+ "step": 2266
+ },
+ {
+ "epoch": 31.489082969432314,
+ "grad_norm": 0.01603875495493412,
+ "learning_rate": 0.0006,
+ "loss": 4.157105445861816,
+ "step": 2267
+ },
+ {
+ "epoch": 31.503056768558952,
+ "grad_norm": 0.015903932973742485,
+ "learning_rate": 0.0006,
+ "loss": 4.233669757843018,
+ "step": 2268
+ },
+ {
+ "epoch": 31.51703056768559,
+ "grad_norm": 0.0160983856767416,
+ "learning_rate": 0.0006,
+ "loss": 4.147704124450684,
+ "step": 2269
+ },
+ {
+ "epoch": 31.531004366812226,
+ "grad_norm": 0.01604730263352394,
+ "learning_rate": 0.0006,
+ "loss": 4.1465911865234375,
+ "step": 2270
+ },
+ {
+ "epoch": 31.544978165938865,
+ "grad_norm": 0.016205785796046257,
+ "learning_rate": 0.0006,
+ "loss": 4.106704235076904,
+ "step": 2271
+ },
+ {
+ "epoch": 31.558951965065503,
+ "grad_norm": 0.017008036375045776,
+ "learning_rate": 0.0006,
+ "loss": 4.191000938415527,
+ "step": 2272
+ },
+ {
+ "epoch": 31.57292576419214,
+ "grad_norm": 0.01699635200202465,
+ "learning_rate": 0.0006,
+ "loss": 4.187042236328125,
+ "step": 2273
+ },
+ {
+ "epoch": 31.586899563318777,
+ "grad_norm": 0.01692967116832733,
+ "learning_rate": 0.0006,
+ "loss": 4.208805561065674,
+ "step": 2274
+ },
+ {
+ "epoch": 31.600873362445416,
+ "grad_norm": 0.017608756199479103,
+ "learning_rate": 0.0006,
+ "loss": 4.200512886047363,
+ "step": 2275
+ },
+ {
+ "epoch": 31.61484716157205,
+ "grad_norm": 0.016590990126132965,
+ "learning_rate": 0.0006,
+ "loss": 4.194705009460449,
+ "step": 2276
+ },
+ {
+ "epoch": 31.62882096069869,
+ "grad_norm": 0.0170235987752676,
+ "learning_rate": 0.0006,
+ "loss": 4.210326194763184,
+ "step": 2277
+ },
+ {
+ "epoch": 31.64279475982533,
+ "grad_norm": 0.01604801043868065,
+ "learning_rate": 0.0006,
+ "loss": 4.196024417877197,
+ "step": 2278
+ },
+ {
+ "epoch": 31.656768558951963,
+ "grad_norm": 0.015694202855229378,
+ "learning_rate": 0.0006,
+ "loss": 4.178236484527588,
+ "step": 2279
+ },
+ {
+ "epoch": 31.670742358078602,
+ "grad_norm": 0.015473959036171436,
+ "learning_rate": 0.0006,
+ "loss": 4.259339809417725,
+ "step": 2280
+ },
+ {
+ "epoch": 31.68471615720524,
+ "grad_norm": 0.014689362607896328,
+ "learning_rate": 0.0006,
+ "loss": 3.9702324867248535,
+ "step": 2281
+ },
+ {
+ "epoch": 31.69868995633188,
+ "grad_norm": 0.015245123766362667,
+ "learning_rate": 0.0006,
+ "loss": 4.197025775909424,
+ "step": 2282
+ },
+ {
+ "epoch": 31.712663755458514,
+ "grad_norm": 0.016148347407579422,
+ "learning_rate": 0.0006,
+ "loss": 4.231856346130371,
+ "step": 2283
+ },
+ {
+ "epoch": 31.726637554585153,
+ "grad_norm": 0.015169818885624409,
+ "learning_rate": 0.0006,
+ "loss": 4.230169296264648,
+ "step": 2284
+ },
+ {
+ "epoch": 31.74061135371179,
+ "grad_norm": 0.014843013137578964,
+ "learning_rate": 0.0006,
+ "loss": 4.195633411407471,
+ "step": 2285
+ },
+ {
+ "epoch": 31.754585152838427,
+ "grad_norm": 0.015229340642690659,
+ "learning_rate": 0.0006,
+ "loss": 4.231280326843262,
+ "step": 2286
+ },
+ {
+ "epoch": 31.768558951965066,
+ "grad_norm": 0.01551708485931158,
+ "learning_rate": 0.0006,
+ "loss": 4.17064094543457,
+ "step": 2287
+ },
+ {
+ "epoch": 31.782532751091704,
+ "grad_norm": 0.015863575041294098,
+ "learning_rate": 0.0006,
+ "loss": 4.1579484939575195,
+ "step": 2288
+ },
+ {
+ "epoch": 31.79650655021834,
+ "grad_norm": 0.015118198469281197,
+ "learning_rate": 0.0006,
+ "loss": 4.1626787185668945,
+ "step": 2289
+ },
+ {
+ "epoch": 31.810480349344978,
+ "grad_norm": 0.015220776200294495,
+ "learning_rate": 0.0006,
+ "loss": 4.135746955871582,
+ "step": 2290
+ },
+ {
+ "epoch": 31.824454148471617,
+ "grad_norm": 0.014529784210026264,
+ "learning_rate": 0.0006,
+ "loss": 4.164931297302246,
+ "step": 2291
+ },
+ {
+ "epoch": 31.83842794759825,
+ "grad_norm": 0.015533296391367912,
+ "learning_rate": 0.0006,
+ "loss": 4.112464904785156,
+ "step": 2292
+ },
+ {
+ "epoch": 31.85240174672489,
+ "grad_norm": 0.016562335193157196,
+ "learning_rate": 0.0006,
+ "loss": 4.164898872375488,
+ "step": 2293
+ },
+ {
+ "epoch": 31.86637554585153,
+ "grad_norm": 0.016258113086223602,
+ "learning_rate": 0.0006,
+ "loss": 4.1979265213012695,
+ "step": 2294
+ },
+ {
+ "epoch": 31.880349344978168,
+ "grad_norm": 0.01558110024780035,
+ "learning_rate": 0.0006,
+ "loss": 4.174675941467285,
+ "step": 2295
+ },
+ {
+ "epoch": 31.894323144104803,
+ "grad_norm": 0.016412515193223953,
+ "learning_rate": 0.0006,
+ "loss": 4.217909336090088,
+ "step": 2296
+ },
+ {
+ "epoch": 31.90829694323144,
+ "grad_norm": 0.015118442475795746,
+ "learning_rate": 0.0006,
+ "loss": 4.137143611907959,
+ "step": 2297
+ },
+ {
+ "epoch": 31.92227074235808,
+ "grad_norm": 0.01506161317229271,
+ "learning_rate": 0.0006,
+ "loss": 4.196122169494629,
+ "step": 2298
+ },
+ {
+ "epoch": 31.936244541484715,
+ "grad_norm": 0.017165271565318108,
+ "learning_rate": 0.0006,
+ "loss": 4.14834451675415,
+ "step": 2299
+ },
+ {
+ "epoch": 31.950218340611354,
+ "grad_norm": 0.01696440577507019,
+ "learning_rate": 0.0006,
+ "loss": 4.218868732452393,
+ "step": 2300
+ },
+ {
+ "epoch": 31.964192139737992,
+ "grad_norm": 0.017349394038319588,
+ "learning_rate": 0.0006,
+ "loss": 4.2754011154174805,
+ "step": 2301
+ },
+ {
+ "epoch": 31.978165938864628,
+ "grad_norm": 0.018387921154499054,
+ "learning_rate": 0.0006,
+ "loss": 4.282623291015625,
+ "step": 2302
+ },
+ {
+ "epoch": 31.992139737991266,
+ "grad_norm": 0.018837926909327507,
+ "learning_rate": 0.0006,
+ "loss": 4.185425281524658,
+ "step": 2303
+ },
+ {
+ "epoch": 32.0,
+ "grad_norm": 0.018998464569449425,
+ "learning_rate": 0.0006,
+ "loss": 4.151291847229004,
+ "step": 2304
+ },
+ {
+ "epoch": 32.0,
+ "eval_loss": 4.522614002227783,
+ "eval_runtime": 57.0106,
+ "eval_samples_per_second": 42.834,
+ "eval_steps_per_second": 1.351,
+ "step": 2304
+ },
+ {
+ "epoch": 32.01397379912664,
+ "grad_norm": 0.019036833196878433,
+ "learning_rate": 0.0006,
+ "loss": 4.183202266693115,
+ "step": 2305
+ },
+ {
+ "epoch": 32.02794759825328,
+ "grad_norm": 0.02207397110760212,
+ "learning_rate": 0.0006,
+ "loss": 4.17875862121582,
+ "step": 2306
+ },
+ {
+ "epoch": 32.041921397379916,
+ "grad_norm": 0.022008279338479042,
+ "learning_rate": 0.0006,
+ "loss": 4.08042049407959,
+ "step": 2307
+ },
+ {
+ "epoch": 32.05589519650655,
+ "grad_norm": 0.022407017648220062,
+ "learning_rate": 0.0006,
+ "loss": 4.112189769744873,
+ "step": 2308
+ },
+ {
+ "epoch": 32.069868995633186,
+ "grad_norm": 0.021599747240543365,
+ "learning_rate": 0.0006,
+ "loss": 4.126455307006836,
+ "step": 2309
+ },
+ {
+ "epoch": 32.083842794759825,
+ "grad_norm": 0.020340966060757637,
+ "learning_rate": 0.0006,
+ "loss": 4.080538272857666,
+ "step": 2310
+ },
+ {
+ "epoch": 32.09781659388646,
+ "grad_norm": 0.020956892520189285,
+ "learning_rate": 0.0006,
+ "loss": 4.171943664550781,
+ "step": 2311
+ },
+ {
+ "epoch": 32.1117903930131,
+ "grad_norm": 0.019966667518019676,
+ "learning_rate": 0.0006,
+ "loss": 4.144007682800293,
+ "step": 2312
+ },
+ {
+ "epoch": 32.12576419213974,
+ "grad_norm": 0.018256952986121178,
+ "learning_rate": 0.0006,
+ "loss": 4.208107948303223,
+ "step": 2313
+ },
+ {
+ "epoch": 32.13973799126637,
+ "grad_norm": 0.017289170995354652,
+ "learning_rate": 0.0006,
+ "loss": 4.182024955749512,
+ "step": 2314
+ },
+ {
+ "epoch": 32.15371179039301,
+ "grad_norm": 0.017656700685620308,
+ "learning_rate": 0.0006,
+ "loss": 4.068373680114746,
+ "step": 2315
+ },
+ {
+ "epoch": 32.16768558951965,
+ "grad_norm": 0.017903361469507217,
+ "learning_rate": 0.0006,
+ "loss": 4.2345428466796875,
+ "step": 2316
+ },
+ {
+ "epoch": 32.18165938864629,
+ "grad_norm": 0.01785109005868435,
+ "learning_rate": 0.0006,
+ "loss": 4.219776153564453,
+ "step": 2317
+ },
+ {
+ "epoch": 32.19563318777293,
+ "grad_norm": 0.01887257769703865,
+ "learning_rate": 0.0006,
+ "loss": 4.139206409454346,
+ "step": 2318
+ },
+ {
+ "epoch": 32.209606986899566,
+ "grad_norm": 0.01924685761332512,
+ "learning_rate": 0.0006,
+ "loss": 4.121052265167236,
+ "step": 2319
+ },
+ {
+ "epoch": 32.223580786026204,
+ "grad_norm": 0.020574072375893593,
+ "learning_rate": 0.0006,
+ "loss": 4.13262414932251,
+ "step": 2320
+ },
+ {
+ "epoch": 32.237554585152836,
+ "grad_norm": 0.023849064484238625,
+ "learning_rate": 0.0006,
+ "loss": 4.239119529724121,
+ "step": 2321
+ },
+ {
+ "epoch": 32.251528384279474,
+ "grad_norm": 0.026239361613988876,
+ "learning_rate": 0.0006,
+ "loss": 4.248993873596191,
+ "step": 2322
+ },
+ {
+ "epoch": 32.26550218340611,
+ "grad_norm": 0.023980243131518364,
+ "learning_rate": 0.0006,
+ "loss": 4.159852504730225,
+ "step": 2323
+ },
+ {
+ "epoch": 32.27947598253275,
+ "grad_norm": 0.020442405715584755,
+ "learning_rate": 0.0006,
+ "loss": 4.133546829223633,
+ "step": 2324
+ },
+ {
+ "epoch": 32.29344978165939,
+ "grad_norm": 0.020661186426877975,
+ "learning_rate": 0.0006,
+ "loss": 4.078243732452393,
+ "step": 2325
+ },
+ {
+ "epoch": 32.30742358078603,
+ "grad_norm": 0.02046387456357479,
+ "learning_rate": 0.0006,
+ "loss": 4.172012805938721,
+ "step": 2326
+ },
+ {
+ "epoch": 32.32139737991266,
+ "grad_norm": 0.02089597098529339,
+ "learning_rate": 0.0006,
+ "loss": 4.0604705810546875,
+ "step": 2327
+ },
+ {
+ "epoch": 32.3353711790393,
+ "grad_norm": 0.01875591278076172,
+ "learning_rate": 0.0006,
+ "loss": 4.125911235809326,
+ "step": 2328
+ },
+ {
+ "epoch": 32.34934497816594,
+ "grad_norm": 0.018606796860694885,
+ "learning_rate": 0.0006,
+ "loss": 4.097362518310547,
+ "step": 2329
+ },
+ {
+ "epoch": 32.36331877729258,
+ "grad_norm": 0.01766626164317131,
+ "learning_rate": 0.0006,
+ "loss": 4.118616580963135,
+ "step": 2330
+ },
+ {
+ "epoch": 32.377292576419215,
+ "grad_norm": 0.017692960798740387,
+ "learning_rate": 0.0006,
+ "loss": 4.1457672119140625,
+ "step": 2331
+ },
+ {
+ "epoch": 32.391266375545854,
+ "grad_norm": 0.017619704827666283,
+ "learning_rate": 0.0006,
+ "loss": 4.175961494445801,
+ "step": 2332
+ },
+ {
+ "epoch": 32.40524017467249,
+ "grad_norm": 0.01784413494169712,
+ "learning_rate": 0.0006,
+ "loss": 4.051291465759277,
+ "step": 2333
+ },
+ {
+ "epoch": 32.419213973799124,
+ "grad_norm": 0.018508171662688255,
+ "learning_rate": 0.0006,
+ "loss": 4.0638556480407715,
+ "step": 2334
+ },
+ {
+ "epoch": 32.43318777292576,
+ "grad_norm": 0.01731436885893345,
+ "learning_rate": 0.0006,
+ "loss": 4.1805419921875,
+ "step": 2335
+ },
+ {
+ "epoch": 32.4471615720524,
+ "grad_norm": 0.015886498615145683,
+ "learning_rate": 0.0006,
+ "loss": 4.151026725769043,
+ "step": 2336
+ },
+ {
+ "epoch": 32.46113537117904,
+ "grad_norm": 0.016583820804953575,
+ "learning_rate": 0.0006,
+ "loss": 4.203604698181152,
+ "step": 2337
+ },
+ {
+ "epoch": 32.47510917030568,
+ "grad_norm": 0.0176097322255373,
+ "learning_rate": 0.0006,
+ "loss": 4.135159015655518,
+ "step": 2338
+ },
+ {
+ "epoch": 32.48908296943232,
+ "grad_norm": 0.016657903790473938,
+ "learning_rate": 0.0006,
+ "loss": 4.126681804656982,
+ "step": 2339
+ },
+ {
+ "epoch": 32.50305676855895,
+ "grad_norm": 0.016866130754351616,
+ "learning_rate": 0.0006,
+ "loss": 4.162341117858887,
+ "step": 2340
+ },
+ {
+ "epoch": 32.51703056768559,
+ "grad_norm": 0.016286680474877357,
+ "learning_rate": 0.0006,
+ "loss": 4.261801242828369,
+ "step": 2341
+ },
+ {
+ "epoch": 32.531004366812226,
+ "grad_norm": 0.01481365505605936,
+ "learning_rate": 0.0006,
+ "loss": 4.110016822814941,
+ "step": 2342
+ },
+ {
+ "epoch": 32.544978165938865,
+ "grad_norm": 0.016605013981461525,
+ "learning_rate": 0.0006,
+ "loss": 4.230024814605713,
+ "step": 2343
+ },
+ {
+ "epoch": 32.5589519650655,
+ "grad_norm": 0.01645175740122795,
+ "learning_rate": 0.0006,
+ "loss": 4.174612045288086,
+ "step": 2344
+ },
+ {
+ "epoch": 32.57292576419214,
+ "grad_norm": 0.014725026674568653,
+ "learning_rate": 0.0006,
+ "loss": 4.066371440887451,
+ "step": 2345
+ },
+ {
+ "epoch": 32.58689956331878,
+ "grad_norm": 0.015003837645053864,
+ "learning_rate": 0.0006,
+ "loss": 4.193343639373779,
+ "step": 2346
+ },
+ {
+ "epoch": 32.60087336244541,
+ "grad_norm": 0.015346302650868893,
+ "learning_rate": 0.0006,
+ "loss": 4.210659980773926,
+ "step": 2347
+ },
+ {
+ "epoch": 32.61484716157205,
+ "grad_norm": 0.015135962516069412,
+ "learning_rate": 0.0006,
+ "loss": 4.204618453979492,
+ "step": 2348
+ },
+ {
+ "epoch": 32.62882096069869,
+ "grad_norm": 0.01601956970989704,
+ "learning_rate": 0.0006,
+ "loss": 4.145564556121826,
+ "step": 2349
+ },
+ {
+ "epoch": 32.64279475982533,
+ "grad_norm": 0.01517851185053587,
+ "learning_rate": 0.0006,
+ "loss": 4.051673889160156,
+ "step": 2350
+ },
+ {
+ "epoch": 32.65676855895197,
+ "grad_norm": 0.014894292689859867,
+ "learning_rate": 0.0006,
+ "loss": 4.184818267822266,
+ "step": 2351
+ },
+ {
+ "epoch": 32.670742358078606,
+ "grad_norm": 0.015720834955573082,
+ "learning_rate": 0.0006,
+ "loss": 4.220036506652832,
+ "step": 2352
+ },
+ {
+ "epoch": 32.68471615720524,
+ "grad_norm": 0.01573433168232441,
+ "learning_rate": 0.0006,
+ "loss": 4.076773166656494,
+ "step": 2353
+ },
+ {
+ "epoch": 32.698689956331876,
+ "grad_norm": 0.015828052535653114,
+ "learning_rate": 0.0006,
+ "loss": 4.116329669952393,
+ "step": 2354
+ },
+ {
+ "epoch": 32.712663755458514,
+ "grad_norm": 0.015462259761989117,
+ "learning_rate": 0.0006,
+ "loss": 4.1155242919921875,
+ "step": 2355
+ },
+ {
+ "epoch": 32.72663755458515,
+ "grad_norm": 0.015763849020004272,
+ "learning_rate": 0.0006,
+ "loss": 4.169687747955322,
+ "step": 2356
+ },
+ {
+ "epoch": 32.74061135371179,
+ "grad_norm": 0.014935145154595375,
+ "learning_rate": 0.0006,
+ "loss": 4.2090911865234375,
+ "step": 2357
+ },
+ {
+ "epoch": 32.75458515283843,
+ "grad_norm": 0.015119184739887714,
+ "learning_rate": 0.0006,
+ "loss": 4.131838798522949,
+ "step": 2358
+ },
+ {
+ "epoch": 32.76855895196506,
+ "grad_norm": 0.014833525754511356,
+ "learning_rate": 0.0006,
+ "loss": 4.17618465423584,
+ "step": 2359
+ },
+ {
+ "epoch": 32.7825327510917,
+ "grad_norm": 0.01588359661400318,
+ "learning_rate": 0.0006,
+ "loss": 4.163723945617676,
+ "step": 2360
+ },
+ {
+ "epoch": 32.79650655021834,
+ "grad_norm": 0.0168188214302063,
+ "learning_rate": 0.0006,
+ "loss": 4.178189277648926,
+ "step": 2361
+ },
+ {
+ "epoch": 32.81048034934498,
+ "grad_norm": 0.01726095750927925,
+ "learning_rate": 0.0006,
+ "loss": 4.144223690032959,
+ "step": 2362
+ },
+ {
+ "epoch": 32.82445414847162,
+ "grad_norm": 0.01701410301029682,
+ "learning_rate": 0.0006,
+ "loss": 4.209896087646484,
+ "step": 2363
+ },
+ {
+ "epoch": 32.838427947598255,
+ "grad_norm": 0.01682264916598797,
+ "learning_rate": 0.0006,
+ "loss": 4.26223087310791,
+ "step": 2364
+ },
+ {
+ "epoch": 32.852401746724894,
+ "grad_norm": 0.015597456134855747,
+ "learning_rate": 0.0006,
+ "loss": 4.292664051055908,
+ "step": 2365
+ },
+ {
+ "epoch": 32.866375545851525,
+ "grad_norm": 0.015258579514920712,
+ "learning_rate": 0.0006,
+ "loss": 4.165006637573242,
+ "step": 2366
+ },
+ {
+ "epoch": 32.880349344978164,
+ "grad_norm": 0.01662200503051281,
+ "learning_rate": 0.0006,
+ "loss": 4.309786319732666,
+ "step": 2367
+ },
+ {
+ "epoch": 32.8943231441048,
+ "grad_norm": 0.015691161155700684,
+ "learning_rate": 0.0006,
+ "loss": 4.167202949523926,
+ "step": 2368
+ },
+ {
+ "epoch": 32.90829694323144,
+ "grad_norm": 0.014870637096464634,
+ "learning_rate": 0.0006,
+ "loss": 4.142415523529053,
+ "step": 2369
+ },
+ {
+ "epoch": 32.92227074235808,
+ "grad_norm": 0.01567252166569233,
+ "learning_rate": 0.0006,
+ "loss": 4.2312092781066895,
+ "step": 2370
+ },
+ {
+ "epoch": 32.93624454148472,
+ "grad_norm": 0.014948616735637188,
+ "learning_rate": 0.0006,
+ "loss": 4.1514129638671875,
+ "step": 2371
+ },
+ {
+ "epoch": 32.95021834061135,
+ "grad_norm": 0.014654500409960747,
+ "learning_rate": 0.0006,
+ "loss": 4.205010890960693,
+ "step": 2372
+ },
+ {
+ "epoch": 32.96419213973799,
+ "grad_norm": 0.015059148892760277,
+ "learning_rate": 0.0006,
+ "loss": 4.148405075073242,
+ "step": 2373
+ },
+ {
+ "epoch": 32.97816593886463,
+ "grad_norm": 0.015026576817035675,
+ "learning_rate": 0.0006,
+ "loss": 4.182535171508789,
+ "step": 2374
+ },
+ {
+ "epoch": 32.992139737991266,
+ "grad_norm": 0.016394320875406265,
+ "learning_rate": 0.0006,
+ "loss": 4.129051208496094,
+ "step": 2375
+ },
+ {
+ "epoch": 33.0,
+ "grad_norm": 0.01867043413221836,
+ "learning_rate": 0.0006,
+ "loss": 4.185556411743164,
+ "step": 2376
+ },
+ {
+ "epoch": 33.0,
+ "eval_loss": 4.640405654907227,
+ "eval_runtime": 56.5861,
+ "eval_samples_per_second": 43.155,
+ "eval_steps_per_second": 1.361,
+ "step": 2376
+ },
+ {
+ "epoch": 33.01397379912664,
+ "grad_norm": 0.018626727163791656,
+ "learning_rate": 0.0006,
+ "loss": 4.143031597137451,
+ "step": 2377
+ },
+ {
+ "epoch": 33.02794759825328,
+ "grad_norm": 0.019283363595604897,
+ "learning_rate": 0.0006,
+ "loss": 4.117845058441162,
+ "step": 2378
+ },
+ {
+ "epoch": 33.041921397379916,
+ "grad_norm": 0.018185822293162346,
+ "learning_rate": 0.0006,
+ "loss": 4.076862812042236,
+ "step": 2379
+ },
+ {
+ "epoch": 33.05589519650655,
+ "grad_norm": 0.019411196932196617,
+ "learning_rate": 0.0006,
+ "loss": 4.124156475067139,
+ "step": 2380
+ },
+ {
+ "epoch": 33.069868995633186,
+ "grad_norm": 0.019248466938734055,
+ "learning_rate": 0.0006,
+ "loss": 4.201735019683838,
+ "step": 2381
+ },
+ {
+ "epoch": 33.083842794759825,
+ "grad_norm": 0.018856720998883247,
+ "learning_rate": 0.0006,
+ "loss": 4.00296688079834,
+ "step": 2382
+ },
+ {
+ "epoch": 33.09781659388646,
+ "grad_norm": 0.018554294481873512,
+ "learning_rate": 0.0006,
+ "loss": 4.270577430725098,
+ "step": 2383
+ },
+ {
+ "epoch": 33.1117903930131,
+ "grad_norm": 0.019946224987506866,
+ "learning_rate": 0.0006,
+ "loss": 4.119589328765869,
+ "step": 2384
+ },
+ {
+ "epoch": 33.12576419213974,
+ "grad_norm": 0.021395670250058174,
+ "learning_rate": 0.0006,
+ "loss": 4.128847122192383,
+ "step": 2385
+ },
+ {
+ "epoch": 33.13973799126637,
+ "grad_norm": 0.02009674906730652,
+ "learning_rate": 0.0006,
+ "loss": 4.183228969573975,
+ "step": 2386
+ },
+ {
+ "epoch": 33.15371179039301,
+ "grad_norm": 0.01820223033428192,
+ "learning_rate": 0.0006,
+ "loss": 4.131601810455322,
+ "step": 2387
+ },
+ {
+ "epoch": 33.16768558951965,
+ "grad_norm": 0.018653156235814095,
+ "learning_rate": 0.0006,
+ "loss": 4.116430282592773,
+ "step": 2388
+ },
+ {
+ "epoch": 33.18165938864629,
+ "grad_norm": 0.016828063875436783,
+ "learning_rate": 0.0006,
+ "loss": 4.054663181304932,
+ "step": 2389
+ },
+ {
+ "epoch": 33.19563318777293,
+ "grad_norm": 0.017025409266352654,
+ "learning_rate": 0.0006,
+ "loss": 4.110044956207275,
+ "step": 2390
+ },
+ {
+ "epoch": 33.209606986899566,
+ "grad_norm": 0.016608452424407005,
+ "learning_rate": 0.0006,
+ "loss": 4.102668285369873,
+ "step": 2391
+ },
+ {
+ "epoch": 33.223580786026204,
+ "grad_norm": 0.01743577979505062,
+ "learning_rate": 0.0006,
+ "loss": 4.166791915893555,
+ "step": 2392
+ },
+ {
+ "epoch": 33.237554585152836,
+ "grad_norm": 0.017253734171390533,
+ "learning_rate": 0.0006,
+ "loss": 4.217832088470459,
+ "step": 2393
+ },
+ {
+ "epoch": 33.251528384279474,
+ "grad_norm": 0.017937595024704933,
+ "learning_rate": 0.0006,
+ "loss": 4.2173967361450195,
+ "step": 2394
+ },
+ {
+ "epoch": 33.26550218340611,
+ "grad_norm": 0.01985993981361389,
+ "learning_rate": 0.0006,
+ "loss": 4.17849588394165,
+ "step": 2395
+ },
+ {
+ "epoch": 33.27947598253275,
+ "grad_norm": 0.019243355840444565,
+ "learning_rate": 0.0006,
+ "loss": 4.086523056030273,
+ "step": 2396
+ },
+ {
+ "epoch": 33.29344978165939,
+ "grad_norm": 0.01924416422843933,
+ "learning_rate": 0.0006,
+ "loss": 4.128574371337891,
+ "step": 2397
+ },
+ {
+ "epoch": 33.30742358078603,
+ "grad_norm": 0.01983758620917797,
+ "learning_rate": 0.0006,
+ "loss": 4.061581611633301,
+ "step": 2398
+ },
+ {
+ "epoch": 33.32139737991266,
+ "grad_norm": 0.01930462196469307,
+ "learning_rate": 0.0006,
+ "loss": 4.091731071472168,
+ "step": 2399
+ },
+ {
+ "epoch": 33.3353711790393,
+ "grad_norm": 0.017988424748182297,
+ "learning_rate": 0.0006,
+ "loss": 4.155599117279053,
+ "step": 2400
+ },
+ {
+ "epoch": 33.34934497816594,
+ "grad_norm": 0.015194767154753208,
+ "learning_rate": 0.0006,
+ "loss": 4.144134044647217,
+ "step": 2401
+ },
+ {
+ "epoch": 33.36331877729258,
+ "grad_norm": 0.01536081824451685,
+ "learning_rate": 0.0006,
+ "loss": 4.127272129058838,
+ "step": 2402
+ },
+ {
+ "epoch": 33.377292576419215,
+ "grad_norm": 0.015431297942996025,
+ "learning_rate": 0.0006,
+ "loss": 4.3213019371032715,
+ "step": 2403
+ },
+ {
+ "epoch": 33.391266375545854,
+ "grad_norm": 0.015740280970931053,
+ "learning_rate": 0.0006,
+ "loss": 4.121467590332031,
+ "step": 2404
+ },
+ {
+ "epoch": 33.40524017467249,
+ "grad_norm": 0.015379384160041809,
+ "learning_rate": 0.0006,
+ "loss": 4.095216751098633,
+ "step": 2405
+ },
+ {
+ "epoch": 33.419213973799124,
+ "grad_norm": 0.01528073474764824,
+ "learning_rate": 0.0006,
+ "loss": 4.13047981262207,
+ "step": 2406
+ },
+ {
+ "epoch": 33.43318777292576,
+ "grad_norm": 0.015739979222416878,
+ "learning_rate": 0.0006,
+ "loss": 4.2533135414123535,
+ "step": 2407
+ },
+ {
+ "epoch": 33.4471615720524,
+ "grad_norm": 0.01519764307886362,
+ "learning_rate": 0.0006,
+ "loss": 4.131782054901123,
+ "step": 2408
+ },
+ {
+ "epoch": 33.46113537117904,
+ "grad_norm": 0.015425996854901314,
+ "learning_rate": 0.0006,
+ "loss": 4.02811336517334,
+ "step": 2409
+ },
+ {
+ "epoch": 33.47510917030568,
+ "grad_norm": 0.015378830023109913,
+ "learning_rate": 0.0006,
+ "loss": 4.08652925491333,
+ "step": 2410
+ },
+ {
+ "epoch": 33.48908296943232,
+ "grad_norm": 0.01473075058311224,
+ "learning_rate": 0.0006,
+ "loss": 4.214364528656006,
+ "step": 2411
+ },
+ {
+ "epoch": 33.50305676855895,
+ "grad_norm": 0.015141637064516544,
+ "learning_rate": 0.0006,
+ "loss": 4.155712127685547,
+ "step": 2412
+ },
+ {
+ "epoch": 33.51703056768559,
+ "grad_norm": 0.01641593500971794,
+ "learning_rate": 0.0006,
+ "loss": 4.097907066345215,
+ "step": 2413
+ },
+ {
+ "epoch": 33.531004366812226,
+ "grad_norm": 0.015182029455900192,
+ "learning_rate": 0.0006,
+ "loss": 4.238499641418457,
+ "step": 2414
+ },
+ {
+ "epoch": 33.544978165938865,
+ "grad_norm": 0.01643732376396656,
+ "learning_rate": 0.0006,
+ "loss": 4.0661492347717285,
+ "step": 2415
+ },
+ {
+ "epoch": 33.5589519650655,
+ "grad_norm": 0.015614228323101997,
+ "learning_rate": 0.0006,
+ "loss": 4.120795249938965,
+ "step": 2416
+ },
+ {
+ "epoch": 33.57292576419214,
+ "grad_norm": 0.016066187992691994,
+ "learning_rate": 0.0006,
+ "loss": 4.096465110778809,
+ "step": 2417
+ },
+ {
+ "epoch": 33.58689956331878,
+ "grad_norm": 0.017214342951774597,
+ "learning_rate": 0.0006,
+ "loss": 4.242628574371338,
+ "step": 2418
+ },
+ {
+ "epoch": 33.60087336244541,
+ "grad_norm": 0.01787586510181427,
+ "learning_rate": 0.0006,
+ "loss": 4.149589538574219,
+ "step": 2419
+ },
+ {
+ "epoch": 33.61484716157205,
+ "grad_norm": 0.019455241039395332,
+ "learning_rate": 0.0006,
+ "loss": 4.267796039581299,
+ "step": 2420
+ },
+ {
+ "epoch": 33.62882096069869,
+ "grad_norm": 0.01984674669802189,
+ "learning_rate": 0.0006,
+ "loss": 4.175220966339111,
+ "step": 2421
+ },
+ {
+ "epoch": 33.64279475982533,
+ "grad_norm": 0.021933717653155327,
+ "learning_rate": 0.0006,
+ "loss": 4.106647491455078,
+ "step": 2422
+ },
+ {
+ "epoch": 33.65676855895197,
+ "grad_norm": 0.02123589813709259,
+ "learning_rate": 0.0006,
+ "loss": 4.204996109008789,
+ "step": 2423
+ },
+ {
+ "epoch": 33.670742358078606,
+ "grad_norm": 0.019814589992165565,
+ "learning_rate": 0.0006,
+ "loss": 4.186171531677246,
+ "step": 2424
+ },
+ {
+ "epoch": 33.68471615720524,
+ "grad_norm": 0.02059728465974331,
+ "learning_rate": 0.0006,
+ "loss": 4.170509338378906,
+ "step": 2425
+ },
+ {
+ "epoch": 33.698689956331876,
+ "grad_norm": 0.018483608961105347,
+ "learning_rate": 0.0006,
+ "loss": 4.085585594177246,
+ "step": 2426
+ },
+ {
+ "epoch": 33.712663755458514,
+ "grad_norm": 0.018403852358460426,
+ "learning_rate": 0.0006,
+ "loss": 4.2551774978637695,
+ "step": 2427
+ },
+ {
+ "epoch": 33.72663755458515,
+ "grad_norm": 0.018653830513358116,
+ "learning_rate": 0.0006,
+ "loss": 4.091766357421875,
+ "step": 2428
+ },
+ {
+ "epoch": 33.74061135371179,
+ "grad_norm": 0.017692334949970245,
+ "learning_rate": 0.0006,
+ "loss": 4.100838661193848,
+ "step": 2429
+ },
+ {
+ "epoch": 33.75458515283843,
+ "grad_norm": 0.016029126942157745,
+ "learning_rate": 0.0006,
+ "loss": 4.1530680656433105,
+ "step": 2430
+ },
+ {
+ "epoch": 33.76855895196506,
+ "grad_norm": 0.016095684841275215,
+ "learning_rate": 0.0006,
+ "loss": 4.20544958114624,
+ "step": 2431
+ },
+ {
+ "epoch": 33.7825327510917,
+ "grad_norm": 0.016375798732042313,
+ "learning_rate": 0.0006,
+ "loss": 4.199177265167236,
+ "step": 2432
+ },
+ {
+ "epoch": 33.79650655021834,
+ "grad_norm": 0.01590786688029766,
+ "learning_rate": 0.0006,
+ "loss": 4.200799465179443,
+ "step": 2433
+ },
+ {
+ "epoch": 33.81048034934498,
+ "grad_norm": 0.014896097593009472,
+ "learning_rate": 0.0006,
+ "loss": 4.185850143432617,
+ "step": 2434
+ },
+ {
+ "epoch": 33.82445414847162,
+ "grad_norm": 0.01438513956964016,
+ "learning_rate": 0.0006,
+ "loss": 4.214962005615234,
+ "step": 2435
+ },
+ {
+ "epoch": 33.838427947598255,
+ "grad_norm": 0.014459339901804924,
+ "learning_rate": 0.0006,
+ "loss": 4.196972370147705,
+ "step": 2436
+ },
+ {
+ "epoch": 33.852401746724894,
+ "grad_norm": 0.014782169833779335,
+ "learning_rate": 0.0006,
+ "loss": 4.0885910987854,
+ "step": 2437
+ },
+ {
+ "epoch": 33.866375545851525,
+ "grad_norm": 0.01520370040088892,
+ "learning_rate": 0.0006,
+ "loss": 4.0323262214660645,
+ "step": 2438
+ },
+ {
+ "epoch": 33.880349344978164,
+ "grad_norm": 0.014056282117962837,
+ "learning_rate": 0.0006,
+ "loss": 4.1386518478393555,
+ "step": 2439
+ },
+ {
+ "epoch": 33.8943231441048,
+ "grad_norm": 0.013567990623414516,
+ "learning_rate": 0.0006,
+ "loss": 4.095754623413086,
+ "step": 2440
+ },
+ {
+ "epoch": 33.90829694323144,
+ "grad_norm": 0.013306716457009315,
+ "learning_rate": 0.0006,
+ "loss": 4.1623077392578125,
+ "step": 2441
+ },
+ {
+ "epoch": 33.92227074235808,
+ "grad_norm": 0.014047150500118732,
+ "learning_rate": 0.0006,
+ "loss": 4.142541885375977,
+ "step": 2442
+ },
+ {
+ "epoch": 33.93624454148472,
+ "grad_norm": 0.01377617847174406,
+ "learning_rate": 0.0006,
+ "loss": 4.268551826477051,
+ "step": 2443
+ },
+ {
+ "epoch": 33.95021834061135,
+ "grad_norm": 0.014671621844172478,
+ "learning_rate": 0.0006,
+ "loss": 4.198196887969971,
+ "step": 2444
+ },
+ {
+ "epoch": 33.96419213973799,
+ "grad_norm": 0.014806744642555714,
+ "learning_rate": 0.0006,
+ "loss": 4.057899475097656,
+ "step": 2445
+ },
+ {
+ "epoch": 33.97816593886463,
+ "grad_norm": 0.015734272077679634,
+ "learning_rate": 0.0006,
+ "loss": 4.163505554199219,
+ "step": 2446
+ },
+ {
+ "epoch": 33.992139737991266,
+ "grad_norm": 0.015250146389007568,
+ "learning_rate": 0.0006,
+ "loss": 4.076294898986816,
+ "step": 2447
+ },
+ {
+ "epoch": 34.0,
+ "grad_norm": 0.016733702272176743,
+ "learning_rate": 0.0006,
+ "loss": 4.138420104980469,
+ "step": 2448
+ },
+ {
+ "epoch": 34.0,
+ "eval_loss": 4.542503356933594,
+ "eval_runtime": 56.983,
+ "eval_samples_per_second": 42.855,
+ "eval_steps_per_second": 1.351,
+ "step": 2448
+ },
+ {
+ "epoch": 34.01397379912664,
+ "grad_norm": 0.015866613015532494,
+ "learning_rate": 0.0006,
+ "loss": 4.032081127166748,
+ "step": 2449
+ },
+ {
+ "epoch": 34.02794759825328,
+ "grad_norm": 0.01946570910513401,
+ "learning_rate": 0.0006,
+ "loss": 4.122320175170898,
+ "step": 2450
+ },
+ {
+ "epoch": 34.041921397379916,
+ "grad_norm": 0.020265938714146614,
+ "learning_rate": 0.0006,
+ "loss": 4.06460428237915,
+ "step": 2451
+ },
+ {
+ "epoch": 34.05589519650655,
+ "grad_norm": 0.018887171521782875,
+ "learning_rate": 0.0006,
+ "loss": 4.096906661987305,
+ "step": 2452
+ },
+ {
+ "epoch": 34.069868995633186,
+ "grad_norm": 0.017304692417383194,
+ "learning_rate": 0.0006,
+ "loss": 4.137261390686035,
+ "step": 2453
+ },
+ {
+ "epoch": 34.083842794759825,
+ "grad_norm": 0.017029082402586937,
+ "learning_rate": 0.0006,
+ "loss": 4.081112861633301,
+ "step": 2454
+ },
+ {
+ "epoch": 34.09781659388646,
+ "grad_norm": 0.016414813697338104,
+ "learning_rate": 0.0006,
+ "loss": 4.140783309936523,
+ "step": 2455
+ },
+ {
+ "epoch": 34.1117903930131,
+ "grad_norm": 0.018238281831145287,
+ "learning_rate": 0.0006,
+ "loss": 4.161116600036621,
+ "step": 2456
+ },
+ {
+ "epoch": 34.12576419213974,
+ "grad_norm": 0.019929232075810432,
+ "learning_rate": 0.0006,
+ "loss": 4.243722438812256,
+ "step": 2457
+ },
+ {
+ "epoch": 34.13973799126637,
+ "grad_norm": 0.019375786185264587,
+ "learning_rate": 0.0006,
+ "loss": 4.134873390197754,
+ "step": 2458
+ },
+ {
+ "epoch": 34.15371179039301,
+ "grad_norm": 0.018151137977838516,
+ "learning_rate": 0.0006,
+ "loss": 4.098745346069336,
+ "step": 2459
+ },
+ {
+ "epoch": 34.16768558951965,
+ "grad_norm": 0.017054934054613113,
+ "learning_rate": 0.0006,
+ "loss": 4.194662570953369,
+ "step": 2460
+ },
+ {
+ "epoch": 34.18165938864629,
+ "grad_norm": 0.018220555037260056,
+ "learning_rate": 0.0006,
+ "loss": 4.128866195678711,
+ "step": 2461
+ },
+ {
+ "epoch": 34.19563318777293,
+ "grad_norm": 0.01692879945039749,
+ "learning_rate": 0.0006,
+ "loss": 4.062477111816406,
+ "step": 2462
+ },
+ {
+ "epoch": 34.209606986899566,
+ "grad_norm": 0.01674199476838112,
+ "learning_rate": 0.0006,
+ "loss": 4.186149597167969,
+ "step": 2463
+ },
+ {
+ "epoch": 34.223580786026204,
+ "grad_norm": 0.016527455300092697,
+ "learning_rate": 0.0006,
+ "loss": 4.0734028816223145,
+ "step": 2464
+ },
+ {
+ "epoch": 34.237554585152836,
+ "grad_norm": 0.016448386013507843,
+ "learning_rate": 0.0006,
+ "loss": 4.098896026611328,
+ "step": 2465
+ },
+ {
+ "epoch": 34.251528384279474,
+ "grad_norm": 0.014644953422248363,
+ "learning_rate": 0.0006,
+ "loss": 4.1031904220581055,
+ "step": 2466
+ },
+ {
+ "epoch": 34.26550218340611,
+ "grad_norm": 0.016649341210722923,
+ "learning_rate": 0.0006,
+ "loss": 4.148791313171387,
+ "step": 2467
+ },
+ {
+ "epoch": 34.27947598253275,
+ "grad_norm": 0.017442714422941208,
+ "learning_rate": 0.0006,
+ "loss": 4.070799827575684,
+ "step": 2468
+ },
+ {
+ "epoch": 34.29344978165939,
+ "grad_norm": 0.017852013930678368,
+ "learning_rate": 0.0006,
+ "loss": 4.031924247741699,
+ "step": 2469
+ },
+ {
+ "epoch": 34.30742358078603,
+ "grad_norm": 0.016371270641684532,
+ "learning_rate": 0.0006,
+ "loss": 4.149420738220215,
+ "step": 2470
+ },
+ {
+ "epoch": 34.32139737991266,
+ "grad_norm": 0.015846814960241318,
+ "learning_rate": 0.0006,
+ "loss": 4.15427303314209,
+ "step": 2471
+ },
+ {
+ "epoch": 34.3353711790393,
+ "grad_norm": 0.017411163076758385,
+ "learning_rate": 0.0006,
+ "loss": 4.1304192543029785,
+ "step": 2472
+ },
+ {
+ "epoch": 34.34934497816594,
+ "grad_norm": 0.01780046708881855,
+ "learning_rate": 0.0006,
+ "loss": 4.191732406616211,
+ "step": 2473
+ },
+ {
+ "epoch": 34.36331877729258,
+ "grad_norm": 0.01958383060991764,
+ "learning_rate": 0.0006,
+ "loss": 4.068355083465576,
+ "step": 2474
+ },
+ {
+ "epoch": 34.377292576419215,
+ "grad_norm": 0.022173317149281502,
+ "learning_rate": 0.0006,
+ "loss": 4.182929039001465,
+ "step": 2475
+ },
+ {
+ "epoch": 34.391266375545854,
+ "grad_norm": 0.022389156743884087,
+ "learning_rate": 0.0006,
+ "loss": 4.110218048095703,
+ "step": 2476
+ },
+ {
+ "epoch": 34.40524017467249,
+ "grad_norm": 0.022478511556982994,
+ "learning_rate": 0.0006,
+ "loss": 4.095479488372803,
+ "step": 2477
+ },
+ {
+ "epoch": 34.419213973799124,
+ "grad_norm": 0.019504647701978683,
+ "learning_rate": 0.0006,
+ "loss": 4.101964950561523,
+ "step": 2478
+ },
+ {
+ "epoch": 34.43318777292576,
+ "grad_norm": 0.019378576427698135,
+ "learning_rate": 0.0006,
+ "loss": 4.14620304107666,
+ "step": 2479
+ },
+ {
+ "epoch": 34.4471615720524,
+ "grad_norm": 0.017993204295635223,
+ "learning_rate": 0.0006,
+ "loss": 4.183492660522461,
+ "step": 2480
+ },
+ {
+ "epoch": 34.46113537117904,
+ "grad_norm": 0.01814805157482624,
+ "learning_rate": 0.0006,
+ "loss": 4.151247978210449,
+ "step": 2481
+ },
+ {
+ "epoch": 34.47510917030568,
+ "grad_norm": 0.018622448667883873,
+ "learning_rate": 0.0006,
+ "loss": 4.020938396453857,
+ "step": 2482
+ },
+ {
+ "epoch": 34.48908296943232,
+ "grad_norm": 0.018360132351517677,
+ "learning_rate": 0.0006,
+ "loss": 4.064734935760498,
+ "step": 2483
+ },
+ {
+ "epoch": 34.50305676855895,
+ "grad_norm": 0.017715711146593094,
+ "learning_rate": 0.0006,
+ "loss": 4.185687065124512,
+ "step": 2484
+ },
+ {
+ "epoch": 34.51703056768559,
+ "grad_norm": 0.01714295893907547,
+ "learning_rate": 0.0006,
+ "loss": 4.1044721603393555,
+ "step": 2485
+ },
+ {
+ "epoch": 34.531004366812226,
+ "grad_norm": 0.016943803057074547,
+ "learning_rate": 0.0006,
+ "loss": 4.106837272644043,
+ "step": 2486
+ },
+ {
+ "epoch": 34.544978165938865,
+ "grad_norm": 0.0182423684746027,
+ "learning_rate": 0.0006,
+ "loss": 4.1041340827941895,
+ "step": 2487
+ },
+ {
+ "epoch": 34.5589519650655,
+ "grad_norm": 0.01636587455868721,
+ "learning_rate": 0.0006,
+ "loss": 3.9735546112060547,
+ "step": 2488
+ },
+ {
+ "epoch": 34.57292576419214,
+ "grad_norm": 0.014593546278774738,
+ "learning_rate": 0.0006,
+ "loss": 4.041980266571045,
+ "step": 2489
+ },
+ {
+ "epoch": 34.58689956331878,
+ "grad_norm": 0.015436965972185135,
+ "learning_rate": 0.0006,
+ "loss": 4.301148414611816,
+ "step": 2490
+ },
+ {
+ "epoch": 34.60087336244541,
+ "grad_norm": 0.015075269155204296,
+ "learning_rate": 0.0006,
+ "loss": 4.118527412414551,
+ "step": 2491
+ },
+ {
+ "epoch": 34.61484716157205,
+ "grad_norm": 0.01701514609158039,
+ "learning_rate": 0.0006,
+ "loss": 4.245968818664551,
+ "step": 2492
+ },
+ {
+ "epoch": 34.62882096069869,
+ "grad_norm": 0.018047887831926346,
+ "learning_rate": 0.0006,
+ "loss": 4.227261066436768,
+ "step": 2493
+ },
+ {
+ "epoch": 34.64279475982533,
+ "grad_norm": 0.018499085679650307,
+ "learning_rate": 0.0006,
+ "loss": 4.121360778808594,
+ "step": 2494
+ },
+ {
+ "epoch": 34.65676855895197,
+ "grad_norm": 0.018987711519002914,
+ "learning_rate": 0.0006,
+ "loss": 4.131564617156982,
+ "step": 2495
+ },
+ {
+ "epoch": 34.670742358078606,
+ "grad_norm": 0.01834476925432682,
+ "learning_rate": 0.0006,
+ "loss": 4.224941253662109,
+ "step": 2496
+ },
+ {
+ "epoch": 34.68471615720524,
+ "grad_norm": 0.017507942393422127,
+ "learning_rate": 0.0006,
+ "loss": 4.1394548416137695,
+ "step": 2497
+ },
+ {
+ "epoch": 34.698689956331876,
+ "grad_norm": 0.01680580899119377,
+ "learning_rate": 0.0006,
+ "loss": 4.1073431968688965,
+ "step": 2498
+ },
+ {
+ "epoch": 34.712663755458514,
+ "grad_norm": 0.015233514830470085,
+ "learning_rate": 0.0006,
+ "loss": 4.136994361877441,
+ "step": 2499
+ },
+ {
+ "epoch": 34.72663755458515,
+ "grad_norm": 0.015482917428016663,
+ "learning_rate": 0.0006,
+ "loss": 4.110719680786133,
+ "step": 2500
+ },
+ {
+ "epoch": 34.74061135371179,
+ "grad_norm": 0.015940461307764053,
+ "learning_rate": 0.0006,
+ "loss": 4.207851886749268,
+ "step": 2501
+ },
+ {
+ "epoch": 34.75458515283843,
+ "grad_norm": 0.015355597250163555,
+ "learning_rate": 0.0006,
+ "loss": 4.169469833374023,
+ "step": 2502
+ },
+ {
+ "epoch": 34.76855895196506,
+ "grad_norm": 0.015974000096321106,
+ "learning_rate": 0.0006,
+ "loss": 4.110061168670654,
+ "step": 2503
+ },
+ {
+ "epoch": 34.7825327510917,
+ "grad_norm": 0.016088927164673805,
+ "learning_rate": 0.0006,
+ "loss": 4.155452728271484,
+ "step": 2504
+ },
+ {
+ "epoch": 34.79650655021834,
+ "grad_norm": 0.016156120225787163,
+ "learning_rate": 0.0006,
+ "loss": 4.178092956542969,
+ "step": 2505
+ },
+ {
+ "epoch": 34.81048034934498,
+ "grad_norm": 0.014928647316992283,
+ "learning_rate": 0.0006,
+ "loss": 4.05457878112793,
+ "step": 2506
+ },
+ {
+ "epoch": 34.82445414847162,
+ "grad_norm": 0.015289328061044216,
+ "learning_rate": 0.0006,
+ "loss": 4.18869686126709,
+ "step": 2507
+ },
+ {
+ "epoch": 34.838427947598255,
+ "grad_norm": 0.015799039974808693,
+ "learning_rate": 0.0006,
+ "loss": 4.106073379516602,
+ "step": 2508
+ },
+ {
+ "epoch": 34.852401746724894,
+ "grad_norm": 0.015824418514966965,
+ "learning_rate": 0.0006,
+ "loss": 4.0634589195251465,
+ "step": 2509
+ },
+ {
+ "epoch": 34.866375545851525,
+ "grad_norm": 0.015321656130254269,
+ "learning_rate": 0.0006,
+ "loss": 4.093090057373047,
+ "step": 2510
+ },
+ {
+ "epoch": 34.880349344978164,
+ "grad_norm": 0.014380166307091713,
+ "learning_rate": 0.0006,
+ "loss": 4.1750640869140625,
+ "step": 2511
+ },
+ {
+ "epoch": 34.8943231441048,
+ "grad_norm": 0.014962567947804928,
+ "learning_rate": 0.0006,
+ "loss": 4.185277938842773,
+ "step": 2512
+ },
+ {
+ "epoch": 34.90829694323144,
+ "grad_norm": 0.014467543922364712,
+ "learning_rate": 0.0006,
+ "loss": 4.1600799560546875,
+ "step": 2513
+ },
+ {
+ "epoch": 34.92227074235808,
+ "grad_norm": 0.01545181218534708,
+ "learning_rate": 0.0006,
+ "loss": 4.096299648284912,
+ "step": 2514
+ },
+ {
+ "epoch": 34.93624454148472,
+ "grad_norm": 0.01605340465903282,
+ "learning_rate": 0.0006,
+ "loss": 4.203323841094971,
+ "step": 2515
+ },
+ {
+ "epoch": 34.95021834061135,
+ "grad_norm": 0.015203815884888172,
+ "learning_rate": 0.0006,
+ "loss": 4.201385021209717,
+ "step": 2516
+ },
+ {
+ "epoch": 34.96419213973799,
+ "grad_norm": 0.015149409882724285,
+ "learning_rate": 0.0006,
+ "loss": 4.088164329528809,
+ "step": 2517
+ },
+ {
+ "epoch": 34.97816593886463,
+ "grad_norm": 0.014885558746755123,
+ "learning_rate": 0.0006,
+ "loss": 4.073911666870117,
+ "step": 2518
+ },
+ {
+ "epoch": 34.992139737991266,
+ "grad_norm": 0.015003419481217861,
+ "learning_rate": 0.0006,
+ "loss": 4.157113552093506,
+ "step": 2519
+ },
+ {
+ "epoch": 35.0,
+ "grad_norm": 0.01641916297376156,
+ "learning_rate": 0.0006,
+ "loss": 3.970052719116211,
+ "step": 2520
+ },
+ {
+ "epoch": 35.0,
+ "eval_loss": 4.641692161560059,
+ "eval_runtime": 56.8836,
+ "eval_samples_per_second": 42.93,
+ "eval_steps_per_second": 1.354,
+ "step": 2520
+ },
+ {
+ "epoch": 35.01397379912664,
+ "grad_norm": 0.015894673764705658,
+ "learning_rate": 0.0006,
+ "loss": 4.104437351226807,
+ "step": 2521
+ },
+ {
+ "epoch": 35.02794759825328,
+ "grad_norm": 0.0173867866396904,
+ "learning_rate": 0.0006,
+ "loss": 4.091726303100586,
+ "step": 2522
+ },
+ {
+ "epoch": 35.041921397379916,
+ "grad_norm": 0.019272832199931145,
+ "learning_rate": 0.0006,
+ "loss": 4.217905044555664,
+ "step": 2523
+ },
+ {
+ "epoch": 35.05589519650655,
+ "grad_norm": 0.021820371970534325,
+ "learning_rate": 0.0006,
+ "loss": 4.111428260803223,
+ "step": 2524
+ },
+ {
+ "epoch": 35.069868995633186,
+ "grad_norm": 0.02340741828083992,
+ "learning_rate": 0.0006,
+ "loss": 4.115771293640137,
+ "step": 2525
+ },
+ {
+ "epoch": 35.083842794759825,
+ "grad_norm": 0.023760691285133362,
+ "learning_rate": 0.0006,
+ "loss": 4.132383346557617,
+ "step": 2526
+ },
+ {
+ "epoch": 35.09781659388646,
+ "grad_norm": 0.02073468267917633,
+ "learning_rate": 0.0006,
+ "loss": 4.272691249847412,
+ "step": 2527
+ },
+ {
+ "epoch": 35.1117903930131,
+ "grad_norm": 0.02197718247771263,
+ "learning_rate": 0.0006,
+ "loss": 4.105586528778076,
+ "step": 2528
+ },
+ {
+ "epoch": 35.12576419213974,
+ "grad_norm": 0.0200254675000906,
+ "learning_rate": 0.0006,
+ "loss": 3.968820095062256,
+ "step": 2529
+ },
+ {
+ "epoch": 35.13973799126637,
+ "grad_norm": 0.01929587684571743,
+ "learning_rate": 0.0006,
+ "loss": 4.045468807220459,
+ "step": 2530
+ },
+ {
+ "epoch": 35.15371179039301,
+ "grad_norm": 0.01886162906885147,
+ "learning_rate": 0.0006,
+ "loss": 4.185571670532227,
+ "step": 2531
+ },
+ {
+ "epoch": 35.16768558951965,
+ "grad_norm": 0.019789839163422585,
+ "learning_rate": 0.0006,
+ "loss": 3.9734551906585693,
+ "step": 2532
+ },
+ {
+ "epoch": 35.18165938864629,
+ "grad_norm": 0.0198142658919096,
+ "learning_rate": 0.0006,
+ "loss": 4.2662506103515625,
+ "step": 2533
+ },
+ {
+ "epoch": 35.19563318777293,
+ "grad_norm": 0.02087012305855751,
+ "learning_rate": 0.0006,
+ "loss": 4.175931930541992,
+ "step": 2534
+ },
+ {
+ "epoch": 35.209606986899566,
+ "grad_norm": 0.020294535905122757,
+ "learning_rate": 0.0006,
+ "loss": 4.060324668884277,
+ "step": 2535
+ },
+ {
+ "epoch": 35.223580786026204,
+ "grad_norm": 0.01908983662724495,
+ "learning_rate": 0.0006,
+ "loss": 4.217939853668213,
+ "step": 2536
+ },
+ {
+ "epoch": 35.237554585152836,
+ "grad_norm": 0.0178892333060503,
+ "learning_rate": 0.0006,
+ "loss": 4.190216541290283,
+ "step": 2537
+ },
+ {
+ "epoch": 35.251528384279474,
+ "grad_norm": 0.017892315983772278,
+ "learning_rate": 0.0006,
+ "loss": 4.067419052124023,
+ "step": 2538
+ },
+ {
+ "epoch": 35.26550218340611,
+ "grad_norm": 0.018556250259280205,
+ "learning_rate": 0.0006,
+ "loss": 4.105958461761475,
+ "step": 2539
+ },
+ {
+ "epoch": 35.27947598253275,
+ "grad_norm": 0.01944422349333763,
+ "learning_rate": 0.0006,
+ "loss": 4.155775547027588,
+ "step": 2540
+ },
+ {
+ "epoch": 35.29344978165939,
+ "grad_norm": 0.017333146184682846,
+ "learning_rate": 0.0006,
+ "loss": 4.244513511657715,
+ "step": 2541
+ },
+ {
+ "epoch": 35.30742358078603,
+ "grad_norm": 0.01639244332909584,
+ "learning_rate": 0.0006,
+ "loss": 4.167815685272217,
+ "step": 2542
+ },
+ {
+ "epoch": 35.32139737991266,
+ "grad_norm": 0.01610143668949604,
+ "learning_rate": 0.0006,
+ "loss": 4.121486663818359,
+ "step": 2543
+ },
+ {
+ "epoch": 35.3353711790393,
+ "grad_norm": 0.01656411588191986,
+ "learning_rate": 0.0006,
+ "loss": 4.1755499839782715,
+ "step": 2544
+ },
+ {
+ "epoch": 35.34934497816594,
+ "grad_norm": 0.015075297094881535,
+ "learning_rate": 0.0006,
+ "loss": 4.102992534637451,
+ "step": 2545
+ },
+ {
+ "epoch": 35.36331877729258,
+ "grad_norm": 0.01538170501589775,
+ "learning_rate": 0.0006,
+ "loss": 4.161603927612305,
+ "step": 2546
+ },
+ {
+ "epoch": 35.377292576419215,
+ "grad_norm": 0.01483312901109457,
+ "learning_rate": 0.0006,
+ "loss": 4.085355758666992,
+ "step": 2547
+ },
+ {
+ "epoch": 35.391266375545854,
+ "grad_norm": 0.014894738793373108,
+ "learning_rate": 0.0006,
+ "loss": 4.083585262298584,
+ "step": 2548
+ },
+ {
+ "epoch": 35.40524017467249,
+ "grad_norm": 0.01475780550390482,
+ "learning_rate": 0.0006,
+ "loss": 4.012645721435547,
+ "step": 2549
+ },
+ {
+ "epoch": 35.419213973799124,
+ "grad_norm": 0.014531653374433517,
+ "learning_rate": 0.0006,
+ "loss": 4.197113990783691,
+ "step": 2550
+ },
+ {
+ "epoch": 35.43318777292576,
+ "grad_norm": 0.015316493809223175,
+ "learning_rate": 0.0006,
+ "loss": 4.227970123291016,
+ "step": 2551
+ },
+ {
+ "epoch": 35.4471615720524,
+ "grad_norm": 0.014685571193695068,
+ "learning_rate": 0.0006,
+ "loss": 4.089415550231934,
+ "step": 2552
+ },
+ {
+ "epoch": 35.46113537117904,
+ "grad_norm": 0.015337035991251469,
+ "learning_rate": 0.0006,
+ "loss": 4.080583572387695,
+ "step": 2553
+ },
+ {
+ "epoch": 35.47510917030568,
+ "grad_norm": 0.016754208132624626,
+ "learning_rate": 0.0006,
+ "loss": 4.163935661315918,
+ "step": 2554
+ },
+ {
+ "epoch": 35.48908296943232,
+ "grad_norm": 0.017535798251628876,
+ "learning_rate": 0.0006,
+ "loss": 4.093027591705322,
+ "step": 2555
+ },
+ {
+ "epoch": 35.50305676855895,
+ "grad_norm": 0.017336975783109665,
+ "learning_rate": 0.0006,
+ "loss": 4.127961158752441,
+ "step": 2556
+ },
+ {
+ "epoch": 35.51703056768559,
+ "grad_norm": 0.016242152079939842,
+ "learning_rate": 0.0006,
+ "loss": 4.155593395233154,
+ "step": 2557
+ },
+ {
+ "epoch": 35.531004366812226,
+ "grad_norm": 0.015121322125196457,
+ "learning_rate": 0.0006,
+ "loss": 3.984004020690918,
+ "step": 2558
+ },
+ {
+ "epoch": 35.544978165938865,
+ "grad_norm": 0.015178884379565716,
+ "learning_rate": 0.0006,
+ "loss": 4.111564636230469,
+ "step": 2559
+ },
+ {
+ "epoch": 35.5589519650655,
+ "grad_norm": 0.014777131378650665,
+ "learning_rate": 0.0006,
+ "loss": 4.041402816772461,
+ "step": 2560
+ },
+ {
+ "epoch": 35.57292576419214,
+ "grad_norm": 0.015197676606476307,
+ "learning_rate": 0.0006,
+ "loss": 4.133706092834473,
+ "step": 2561
+ },
+ {
+ "epoch": 35.58689956331878,
+ "grad_norm": 0.015905972570180893,
+ "learning_rate": 0.0006,
+ "loss": 4.1069135665893555,
+ "step": 2562
+ },
+ {
+ "epoch": 35.60087336244541,
+ "grad_norm": 0.015942877158522606,
+ "learning_rate": 0.0006,
+ "loss": 4.0720624923706055,
+ "step": 2563
+ },
+ {
+ "epoch": 35.61484716157205,
+ "grad_norm": 0.016697583720088005,
+ "learning_rate": 0.0006,
+ "loss": 4.167448043823242,
+ "step": 2564
+ },
+ {
+ "epoch": 35.62882096069869,
+ "grad_norm": 0.015603577718138695,
+ "learning_rate": 0.0006,
+ "loss": 4.127327919006348,
+ "step": 2565
+ },
+ {
+ "epoch": 35.64279475982533,
+ "grad_norm": 0.017153460532426834,
+ "learning_rate": 0.0006,
+ "loss": 4.056543350219727,
+ "step": 2566
+ },
+ {
+ "epoch": 35.65676855895197,
+ "grad_norm": 0.016731958836317062,
+ "learning_rate": 0.0006,
+ "loss": 4.0681843757629395,
+ "step": 2567
+ },
+ {
+ "epoch": 35.670742358078606,
+ "grad_norm": 0.016167359426617622,
+ "learning_rate": 0.0006,
+ "loss": 4.174992084503174,
+ "step": 2568
+ },
+ {
+ "epoch": 35.68471615720524,
+ "grad_norm": 0.01615108922123909,
+ "learning_rate": 0.0006,
+ "loss": 4.1383161544799805,
+ "step": 2569
+ },
+ {
+ "epoch": 35.698689956331876,
+ "grad_norm": 0.015750816091895103,
+ "learning_rate": 0.0006,
+ "loss": 4.031595706939697,
+ "step": 2570
+ },
+ {
+ "epoch": 35.712663755458514,
+ "grad_norm": 0.015416189096868038,
+ "learning_rate": 0.0006,
+ "loss": 4.067498207092285,
+ "step": 2571
+ },
+ {
+ "epoch": 35.72663755458515,
+ "grad_norm": 0.01570792682468891,
+ "learning_rate": 0.0006,
+ "loss": 4.1338300704956055,
+ "step": 2572
+ },
+ {
+ "epoch": 35.74061135371179,
+ "grad_norm": 0.016367752104997635,
+ "learning_rate": 0.0006,
+ "loss": 4.152523040771484,
+ "step": 2573
+ },
+ {
+ "epoch": 35.75458515283843,
+ "grad_norm": 0.016349803656339645,
+ "learning_rate": 0.0006,
+ "loss": 4.101985454559326,
+ "step": 2574
+ },
+ {
+ "epoch": 35.76855895196506,
+ "grad_norm": 0.01663660630583763,
+ "learning_rate": 0.0006,
+ "loss": 4.190130710601807,
+ "step": 2575
+ },
+ {
+ "epoch": 35.7825327510917,
+ "grad_norm": 0.015599170699715614,
+ "learning_rate": 0.0006,
+ "loss": 4.0792436599731445,
+ "step": 2576
+ },
+ {
+ "epoch": 35.79650655021834,
+ "grad_norm": 0.015472427010536194,
+ "learning_rate": 0.0006,
+ "loss": 4.186446666717529,
+ "step": 2577
+ },
+ {
+ "epoch": 35.81048034934498,
+ "grad_norm": 0.015520439483225346,
+ "learning_rate": 0.0006,
+ "loss": 4.246635437011719,
+ "step": 2578
+ },
+ {
+ "epoch": 35.82445414847162,
+ "grad_norm": 0.016998281702399254,
+ "learning_rate": 0.0006,
+ "loss": 4.13435173034668,
+ "step": 2579
+ },
+ {
+ "epoch": 35.838427947598255,
+ "grad_norm": 0.016290361061692238,
+ "learning_rate": 0.0006,
+ "loss": 4.107428073883057,
+ "step": 2580
+ },
+ {
+ "epoch": 35.852401746724894,
+ "grad_norm": 0.0165516659617424,
+ "learning_rate": 0.0006,
+ "loss": 4.1612138748168945,
+ "step": 2581
+ },
+ {
+ "epoch": 35.866375545851525,
+ "grad_norm": 0.01736091636121273,
+ "learning_rate": 0.0006,
+ "loss": 4.1072869300842285,
+ "step": 2582
+ },
+ {
+ "epoch": 35.880349344978164,
+ "grad_norm": 0.01787031814455986,
+ "learning_rate": 0.0006,
+ "loss": 4.1076812744140625,
+ "step": 2583
+ },
+ {
+ "epoch": 35.8943231441048,
+ "grad_norm": 0.015437985770404339,
+ "learning_rate": 0.0006,
+ "loss": 4.097488880157471,
+ "step": 2584
+ },
+ {
+ "epoch": 35.90829694323144,
+ "grad_norm": 0.016306741163134575,
+ "learning_rate": 0.0006,
+ "loss": 4.105017185211182,
+ "step": 2585
+ },
+ {
+ "epoch": 35.92227074235808,
+ "grad_norm": 0.015723245218396187,
+ "learning_rate": 0.0006,
+ "loss": 4.205283164978027,
+ "step": 2586
+ },
+ {
+ "epoch": 35.93624454148472,
+ "grad_norm": 0.014626456424593925,
+ "learning_rate": 0.0006,
+ "loss": 4.086552619934082,
+ "step": 2587
+ },
+ {
+ "epoch": 35.95021834061135,
+ "grad_norm": 0.014348454773426056,
+ "learning_rate": 0.0006,
+ "loss": 4.171329975128174,
+ "step": 2588
+ },
+ {
+ "epoch": 35.96419213973799,
+ "grad_norm": 0.015598480589687824,
+ "learning_rate": 0.0006,
+ "loss": 4.140815258026123,
+ "step": 2589
+ },
+ {
+ "epoch": 35.97816593886463,
+ "grad_norm": 0.017126062884926796,
+ "learning_rate": 0.0006,
+ "loss": 4.157176971435547,
+ "step": 2590
+ },
+ {
+ "epoch": 35.992139737991266,
+ "grad_norm": 0.01860199309885502,
+ "learning_rate": 0.0006,
+ "loss": 4.252880096435547,
+ "step": 2591
+ },
+ {
+ "epoch": 36.0,
+ "grad_norm": 0.021780220791697502,
+ "learning_rate": 0.0006,
+ "loss": 4.249861717224121,
+ "step": 2592
+ },
+ {
+ "epoch": 36.0,
+ "eval_loss": 4.606466293334961,
+ "eval_runtime": 56.6217,
+ "eval_samples_per_second": 43.128,
+ "eval_steps_per_second": 1.36,
+ "step": 2592
+ },
+ {
+ "epoch": 36.01397379912664,
+ "grad_norm": 0.020892828702926636,
+ "learning_rate": 0.0006,
+ "loss": 4.102055549621582,
+ "step": 2593
+ },
+ {
+ "epoch": 36.02794759825328,
+ "grad_norm": 0.020306255668401718,
+ "learning_rate": 0.0006,
+ "loss": 4.184308052062988,
+ "step": 2594
+ },
+ {
+ "epoch": 36.041921397379916,
+ "grad_norm": 0.02025097794830799,
+ "learning_rate": 0.0006,
+ "loss": 3.9691858291625977,
+ "step": 2595
+ },
+ {
+ "epoch": 36.05589519650655,
+ "grad_norm": 0.020254382863640785,
+ "learning_rate": 0.0006,
+ "loss": 4.02957820892334,
+ "step": 2596
+ },
+ {
+ "epoch": 36.069868995633186,
+ "grad_norm": 0.019307196140289307,
+ "learning_rate": 0.0006,
+ "loss": 4.033181667327881,
+ "step": 2597
+ },
+ {
+ "epoch": 36.083842794759825,
+ "grad_norm": 0.0198476854711771,
+ "learning_rate": 0.0006,
+ "loss": 4.0816874504089355,
+ "step": 2598
+ },
+ {
+ "epoch": 36.09781659388646,
+ "grad_norm": 0.019916515797376633,
+ "learning_rate": 0.0006,
+ "loss": 4.164889812469482,
+ "step": 2599
+ },
+ {
+ "epoch": 36.1117903930131,
+ "grad_norm": 0.0189144778996706,
+ "learning_rate": 0.0006,
+ "loss": 4.038427352905273,
+ "step": 2600
+ },
+ {
+ "epoch": 36.12576419213974,
+ "grad_norm": 0.01754768192768097,
+ "learning_rate": 0.0006,
+ "loss": 4.010427474975586,
+ "step": 2601
+ },
+ {
+ "epoch": 36.13973799126637,
+ "grad_norm": 0.016862990334630013,
+ "learning_rate": 0.0006,
+ "loss": 4.035970211029053,
+ "step": 2602
+ },
+ {
+ "epoch": 36.15371179039301,
+ "grad_norm": 0.019616752862930298,
+ "learning_rate": 0.0006,
+ "loss": 4.079898834228516,
+ "step": 2603
+ },
+ {
+ "epoch": 36.16768558951965,
+ "grad_norm": 0.020131465047597885,
+ "learning_rate": 0.0006,
+ "loss": 4.144684791564941,
+ "step": 2604
+ },
+ {
+ "epoch": 36.18165938864629,
+ "grad_norm": 0.02072521299123764,
+ "learning_rate": 0.0006,
+ "loss": 4.013660430908203,
+ "step": 2605
+ },
+ {
+ "epoch": 36.19563318777293,
+ "grad_norm": 0.01924615167081356,
+ "learning_rate": 0.0006,
+ "loss": 4.071398735046387,
+ "step": 2606
+ },
+ {
+ "epoch": 36.209606986899566,
+ "grad_norm": 0.02013954520225525,
+ "learning_rate": 0.0006,
+ "loss": 4.127094268798828,
+ "step": 2607
+ },
+ {
+ "epoch": 36.223580786026204,
+ "grad_norm": 0.02042008936405182,
+ "learning_rate": 0.0006,
+ "loss": 4.154261589050293,
+ "step": 2608
+ },
+ {
+ "epoch": 36.237554585152836,
+ "grad_norm": 0.021998876705765724,
+ "learning_rate": 0.0006,
+ "loss": 4.204150199890137,
+ "step": 2609
+ },
+ {
+ "epoch": 36.251528384279474,
+ "grad_norm": 0.022946683689951897,
+ "learning_rate": 0.0006,
+ "loss": 4.013632774353027,
+ "step": 2610
+ },
+ {
+ "epoch": 36.26550218340611,
+ "grad_norm": 0.02061830274760723,
+ "learning_rate": 0.0006,
+ "loss": 4.171084403991699,
+ "step": 2611
+ },
+ {
+ "epoch": 36.27947598253275,
+ "grad_norm": 0.01883152686059475,
+ "learning_rate": 0.0006,
+ "loss": 4.149848461151123,
+ "step": 2612
+ },
+ {
+ "epoch": 36.29344978165939,
+ "grad_norm": 0.019945282489061356,
+ "learning_rate": 0.0006,
+ "loss": 4.174656867980957,
+ "step": 2613
+ },
+ {
+ "epoch": 36.30742358078603,
+ "grad_norm": 0.020751051604747772,
+ "learning_rate": 0.0006,
+ "loss": 4.080747127532959,
+ "step": 2614
+ },
+ {
+ "epoch": 36.32139737991266,
+ "grad_norm": 0.01879691146314144,
+ "learning_rate": 0.0006,
+ "loss": 4.057356357574463,
+ "step": 2615
+ },
+ {
+ "epoch": 36.3353711790393,
+ "grad_norm": 0.019037071615457535,
+ "learning_rate": 0.0006,
+ "loss": 4.204607009887695,
+ "step": 2616
+ },
+ {
+ "epoch": 36.34934497816594,
+ "grad_norm": 0.01872078701853752,
+ "learning_rate": 0.0006,
+ "loss": 4.050586700439453,
+ "step": 2617
+ },
+ {
+ "epoch": 36.36331877729258,
+ "grad_norm": 0.018092811107635498,
+ "learning_rate": 0.0006,
+ "loss": 4.102832794189453,
+ "step": 2618
+ },
+ {
+ "epoch": 36.377292576419215,
+ "grad_norm": 0.017010487616062164,
+ "learning_rate": 0.0006,
+ "loss": 4.191788673400879,
+ "step": 2619
+ },
+ {
+ "epoch": 36.391266375545854,
+ "grad_norm": 0.01691143773496151,
+ "learning_rate": 0.0006,
+ "loss": 4.160456657409668,
+ "step": 2620
+ },
+ {
+ "epoch": 36.40524017467249,
+ "grad_norm": 0.017046816647052765,
+ "learning_rate": 0.0006,
+ "loss": 4.073828220367432,
+ "step": 2621
+ },
+ {
+ "epoch": 36.419213973799124,
+ "grad_norm": 0.01707981713116169,
+ "learning_rate": 0.0006,
+ "loss": 4.0980095863342285,
+ "step": 2622
+ },
+ {
+ "epoch": 36.43318777292576,
+ "grad_norm": 0.01630452275276184,
+ "learning_rate": 0.0006,
+ "loss": 4.101321220397949,
+ "step": 2623
+ },
+ {
+ "epoch": 36.4471615720524,
+ "grad_norm": 0.01576915942132473,
+ "learning_rate": 0.0006,
+ "loss": 4.158249855041504,
+ "step": 2624
+ },
+ {
+ "epoch": 36.46113537117904,
+ "grad_norm": 0.016179712489247322,
+ "learning_rate": 0.0006,
+ "loss": 4.077852249145508,
+ "step": 2625
+ },
+ {
+ "epoch": 36.47510917030568,
+ "grad_norm": 0.01609169878065586,
+ "learning_rate": 0.0006,
+ "loss": 4.187317848205566,
+ "step": 2626
+ },
+ {
+ "epoch": 36.48908296943232,
+ "grad_norm": 0.017148254439234734,
+ "learning_rate": 0.0006,
+ "loss": 4.109071731567383,
+ "step": 2627
+ },
+ {
+ "epoch": 36.50305676855895,
+ "grad_norm": 0.015842624008655548,
+ "learning_rate": 0.0006,
+ "loss": 4.145096302032471,
+ "step": 2628
+ },
+ {
+ "epoch": 36.51703056768559,
+ "grad_norm": 0.014906194061040878,
+ "learning_rate": 0.0006,
+ "loss": 4.028440475463867,
+ "step": 2629
+ },
+ {
+ "epoch": 36.531004366812226,
+ "grad_norm": 0.01485038734972477,
+ "learning_rate": 0.0006,
+ "loss": 4.175201416015625,
+ "step": 2630
+ },
+ {
+ "epoch": 36.544978165938865,
+ "grad_norm": 0.015834150835871696,
+ "learning_rate": 0.0006,
+ "loss": 4.076434135437012,
+ "step": 2631
+ },
+ {
+ "epoch": 36.5589519650655,
+ "grad_norm": 0.015411358326673508,
+ "learning_rate": 0.0006,
+ "loss": 4.03912878036499,
+ "step": 2632
+ },
+ {
+ "epoch": 36.57292576419214,
+ "grad_norm": 0.014762846753001213,
+ "learning_rate": 0.0006,
+ "loss": 4.019827365875244,
+ "step": 2633
+ },
+ {
+ "epoch": 36.58689956331878,
+ "grad_norm": 0.015269497409462929,
+ "learning_rate": 0.0006,
+ "loss": 4.116957187652588,
+ "step": 2634
+ },
+ {
+ "epoch": 36.60087336244541,
+ "grad_norm": 0.015171617269515991,
+ "learning_rate": 0.0006,
+ "loss": 4.2069926261901855,
+ "step": 2635
+ },
+ {
+ "epoch": 36.61484716157205,
+ "grad_norm": 0.014770902693271637,
+ "learning_rate": 0.0006,
+ "loss": 4.092021942138672,
+ "step": 2636
+ },
+ {
+ "epoch": 36.62882096069869,
+ "grad_norm": 0.014957531355321407,
+ "learning_rate": 0.0006,
+ "loss": 4.16819953918457,
+ "step": 2637
+ },
+ {
+ "epoch": 36.64279475982533,
+ "grad_norm": 0.014774451963603497,
+ "learning_rate": 0.0006,
+ "loss": 4.046073913574219,
+ "step": 2638
+ },
+ {
+ "epoch": 36.65676855895197,
+ "grad_norm": 0.014320445246994495,
+ "learning_rate": 0.0006,
+ "loss": 3.977792501449585,
+ "step": 2639
+ },
+ {
+ "epoch": 36.670742358078606,
+ "grad_norm": 0.01413130946457386,
+ "learning_rate": 0.0006,
+ "loss": 4.130649566650391,
+ "step": 2640
+ },
+ {
+ "epoch": 36.68471615720524,
+ "grad_norm": 0.014336307533085346,
+ "learning_rate": 0.0006,
+ "loss": 4.161357402801514,
+ "step": 2641
+ },
+ {
+ "epoch": 36.698689956331876,
+ "grad_norm": 0.014516931027173996,
+ "learning_rate": 0.0006,
+ "loss": 4.14248514175415,
+ "step": 2642
+ },
+ {
+ "epoch": 36.712663755458514,
+ "grad_norm": 0.014377126470208168,
+ "learning_rate": 0.0006,
+ "loss": 4.062711715698242,
+ "step": 2643
+ },
+ {
+ "epoch": 36.72663755458515,
+ "grad_norm": 0.014332668855786324,
+ "learning_rate": 0.0006,
+ "loss": 4.06423807144165,
+ "step": 2644
+ },
+ {
+ "epoch": 36.74061135371179,
+ "grad_norm": 0.014030571095645428,
+ "learning_rate": 0.0006,
+ "loss": 4.178613662719727,
+ "step": 2645
+ },
+ {
+ "epoch": 36.75458515283843,
+ "grad_norm": 0.014376776292920113,
+ "learning_rate": 0.0006,
+ "loss": 4.073130130767822,
+ "step": 2646
+ },
+ {
+ "epoch": 36.76855895196506,
+ "grad_norm": 0.013827716931700706,
+ "learning_rate": 0.0006,
+ "loss": 4.043869972229004,
+ "step": 2647
+ },
+ {
+ "epoch": 36.7825327510917,
+ "grad_norm": 0.015548762865364552,
+ "learning_rate": 0.0006,
+ "loss": 4.134049415588379,
+ "step": 2648
+ },
+ {
+ "epoch": 36.79650655021834,
+ "grad_norm": 0.017041970044374466,
+ "learning_rate": 0.0006,
+ "loss": 4.205024242401123,
+ "step": 2649
+ },
+ {
+ "epoch": 36.81048034934498,
+ "grad_norm": 0.017638275399804115,
+ "learning_rate": 0.0006,
+ "loss": 4.199924468994141,
+ "step": 2650
+ },
+ {
+ "epoch": 36.82445414847162,
+ "grad_norm": 0.017771627753973007,
+ "learning_rate": 0.0006,
+ "loss": 4.15706729888916,
+ "step": 2651
+ },
+ {
+ "epoch": 36.838427947598255,
+ "grad_norm": 0.015933562070131302,
+ "learning_rate": 0.0006,
+ "loss": 4.220742225646973,
+ "step": 2652
+ },
+ {
+ "epoch": 36.852401746724894,
+ "grad_norm": 0.01668144389986992,
+ "learning_rate": 0.0006,
+ "loss": 4.161426544189453,
+ "step": 2653
+ },
+ {
+ "epoch": 36.866375545851525,
+ "grad_norm": 0.018014680594205856,
+ "learning_rate": 0.0006,
+ "loss": 4.183046817779541,
+ "step": 2654
+ },
+ {
+ "epoch": 36.880349344978164,
+ "grad_norm": 0.018808670341968536,
+ "learning_rate": 0.0006,
+ "loss": 4.105293273925781,
+ "step": 2655
+ },
+ {
+ "epoch": 36.8943231441048,
+ "grad_norm": 0.018003476783633232,
+ "learning_rate": 0.0006,
+ "loss": 4.103171348571777,
+ "step": 2656
+ },
+ {
+ "epoch": 36.90829694323144,
+ "grad_norm": 0.016865914687514305,
+ "learning_rate": 0.0006,
+ "loss": 4.125634670257568,
+ "step": 2657
+ },
+ {
+ "epoch": 36.92227074235808,
+ "grad_norm": 0.018371153622865677,
+ "learning_rate": 0.0006,
+ "loss": 4.192461013793945,
+ "step": 2658
+ },
+ {
+ "epoch": 36.93624454148472,
+ "grad_norm": 0.019045377150177956,
+ "learning_rate": 0.0006,
+ "loss": 4.198105812072754,
+ "step": 2659
+ },
+ {
+ "epoch": 36.95021834061135,
+ "grad_norm": 0.01594168320298195,
+ "learning_rate": 0.0006,
+ "loss": 4.169788360595703,
+ "step": 2660
+ },
+ {
+ "epoch": 36.96419213973799,
+ "grad_norm": 0.015645820647478104,
+ "learning_rate": 0.0006,
+ "loss": 4.047277450561523,
+ "step": 2661
+ },
+ {
+ "epoch": 36.97816593886463,
+ "grad_norm": 0.016313156113028526,
+ "learning_rate": 0.0006,
+ "loss": 4.087392807006836,
+ "step": 2662
+ },
+ {
+ "epoch": 36.992139737991266,
+ "grad_norm": 0.017050625756382942,
+ "learning_rate": 0.0006,
+ "loss": 4.0908002853393555,
+ "step": 2663
+ },
+ {
+ "epoch": 37.0,
+ "grad_norm": 0.017730968073010445,
+ "learning_rate": 0.0006,
+ "loss": 4.110957622528076,
+ "step": 2664
+ },
+ {
+ "epoch": 37.0,
+ "eval_loss": 4.563793182373047,
+ "eval_runtime": 56.5922,
+ "eval_samples_per_second": 43.151,
+ "eval_steps_per_second": 1.361,
+ "step": 2664
+ },
+ {
+ "epoch": 37.01397379912664,
+ "grad_norm": 0.015687910839915276,
+ "learning_rate": 0.0006,
+ "loss": 4.143589019775391,
+ "step": 2665
+ },
+ {
+ "epoch": 37.02794759825328,
+ "grad_norm": 0.017451366409659386,
+ "learning_rate": 0.0006,
+ "loss": 4.0326666831970215,
+ "step": 2666
+ },
+ {
+ "epoch": 37.041921397379916,
+ "grad_norm": 0.01909538544714451,
+ "learning_rate": 0.0006,
+ "loss": 3.9706249237060547,
+ "step": 2667
+ },
+ {
+ "epoch": 37.05589519650655,
+ "grad_norm": 0.01988379657268524,
+ "learning_rate": 0.0006,
+ "loss": 3.9531009197235107,
+ "step": 2668
+ },
+ {
+ "epoch": 37.069868995633186,
+ "grad_norm": 0.02035115472972393,
+ "learning_rate": 0.0006,
+ "loss": 3.962214469909668,
+ "step": 2669
+ },
+ {
+ "epoch": 37.083842794759825,
+ "grad_norm": 0.01952001079916954,
+ "learning_rate": 0.0006,
+ "loss": 4.182262897491455,
+ "step": 2670
+ },
+ {
+ "epoch": 37.09781659388646,
+ "grad_norm": 0.01854325644671917,
+ "learning_rate": 0.0006,
+ "loss": 4.127663612365723,
+ "step": 2671
+ },
+ {
+ "epoch": 37.1117903930131,
+ "grad_norm": 0.017757616937160492,
+ "learning_rate": 0.0006,
+ "loss": 4.100804805755615,
+ "step": 2672
+ },
+ {
+ "epoch": 37.12576419213974,
+ "grad_norm": 0.016712479293346405,
+ "learning_rate": 0.0006,
+ "loss": 4.026284217834473,
+ "step": 2673
+ },
+ {
+ "epoch": 37.13973799126637,
+ "grad_norm": 0.0168139711022377,
+ "learning_rate": 0.0006,
+ "loss": 4.183882713317871,
+ "step": 2674
+ },
+ {
+ "epoch": 37.15371179039301,
+ "grad_norm": 0.017060423269867897,
+ "learning_rate": 0.0006,
+ "loss": 4.131832599639893,
+ "step": 2675
+ },
+ {
+ "epoch": 37.16768558951965,
+ "grad_norm": 0.01726631447672844,
+ "learning_rate": 0.0006,
+ "loss": 3.9156057834625244,
+ "step": 2676
+ },
+ {
+ "epoch": 37.18165938864629,
+ "grad_norm": 0.016348622739315033,
+ "learning_rate": 0.0006,
+ "loss": 4.1731367111206055,
+ "step": 2677
+ },
+ {
+ "epoch": 37.19563318777293,
+ "grad_norm": 0.016640059649944305,
+ "learning_rate": 0.0006,
+ "loss": 4.10997200012207,
+ "step": 2678
+ },
+ {
+ "epoch": 37.209606986899566,
+ "grad_norm": 0.01601342298090458,
+ "learning_rate": 0.0006,
+ "loss": 4.081145286560059,
+ "step": 2679
+ },
+ {
+ "epoch": 37.223580786026204,
+ "grad_norm": 0.01735631190240383,
+ "learning_rate": 0.0006,
+ "loss": 4.202108860015869,
+ "step": 2680
+ },
+ {
+ "epoch": 37.237554585152836,
+ "grad_norm": 0.019164685159921646,
+ "learning_rate": 0.0006,
+ "loss": 4.104109764099121,
+ "step": 2681
+ },
+ {
+ "epoch": 37.251528384279474,
+ "grad_norm": 0.01881759613752365,
+ "learning_rate": 0.0006,
+ "loss": 4.145532131195068,
+ "step": 2682
+ },
+ {
+ "epoch": 37.26550218340611,
+ "grad_norm": 0.017656555399298668,
+ "learning_rate": 0.0006,
+ "loss": 4.012782096862793,
+ "step": 2683
+ },
+ {
+ "epoch": 37.27947598253275,
+ "grad_norm": 0.01624823734164238,
+ "learning_rate": 0.0006,
+ "loss": 4.001727104187012,
+ "step": 2684
+ },
+ {
+ "epoch": 37.29344978165939,
+ "grad_norm": 0.016973139718174934,
+ "learning_rate": 0.0006,
+ "loss": 4.095399856567383,
+ "step": 2685
+ },
+ {
+ "epoch": 37.30742358078603,
+ "grad_norm": 0.015503671951591969,
+ "learning_rate": 0.0006,
+ "loss": 4.034431457519531,
+ "step": 2686
+ },
+ {
+ "epoch": 37.32139737991266,
+ "grad_norm": 0.015370250679552555,
+ "learning_rate": 0.0006,
+ "loss": 4.075836181640625,
+ "step": 2687
+ },
+ {
+ "epoch": 37.3353711790393,
+ "grad_norm": 0.01564021222293377,
+ "learning_rate": 0.0006,
+ "loss": 4.125202178955078,
+ "step": 2688
+ },
+ {
+ "epoch": 37.34934497816594,
+ "grad_norm": 0.016717975959181786,
+ "learning_rate": 0.0006,
+ "loss": 4.153017044067383,
+ "step": 2689
+ },
+ {
+ "epoch": 37.36331877729258,
+ "grad_norm": 0.015263617038726807,
+ "learning_rate": 0.0006,
+ "loss": 4.021814823150635,
+ "step": 2690
+ },
+ {
+ "epoch": 37.377292576419215,
+ "grad_norm": 0.016497142612934113,
+ "learning_rate": 0.0006,
+ "loss": 4.069095611572266,
+ "step": 2691
+ },
+ {
+ "epoch": 37.391266375545854,
+ "grad_norm": 0.01669355481863022,
+ "learning_rate": 0.0006,
+ "loss": 4.174587726593018,
+ "step": 2692
+ },
+ {
+ "epoch": 37.40524017467249,
+ "grad_norm": 0.019695773720741272,
+ "learning_rate": 0.0006,
+ "loss": 4.0073041915893555,
+ "step": 2693
+ },
+ {
+ "epoch": 37.419213973799124,
+ "grad_norm": 0.023720644414424896,
+ "learning_rate": 0.0006,
+ "loss": 4.111948013305664,
+ "step": 2694
+ },
+ {
+ "epoch": 37.43318777292576,
+ "grad_norm": 0.02399350143969059,
+ "learning_rate": 0.0006,
+ "loss": 4.081164360046387,
+ "step": 2695
+ },
+ {
+ "epoch": 37.4471615720524,
+ "grad_norm": 0.019630271941423416,
+ "learning_rate": 0.0006,
+ "loss": 4.065731048583984,
+ "step": 2696
+ },
+ {
+ "epoch": 37.46113537117904,
+ "grad_norm": 0.019679008051753044,
+ "learning_rate": 0.0006,
+ "loss": 3.993852376937866,
+ "step": 2697
+ },
+ {
+ "epoch": 37.47510917030568,
+ "grad_norm": 0.020597299560904503,
+ "learning_rate": 0.0006,
+ "loss": 4.275555610656738,
+ "step": 2698
+ },
+ {
+ "epoch": 37.48908296943232,
+ "grad_norm": 0.01983889564871788,
+ "learning_rate": 0.0006,
+ "loss": 4.254708290100098,
+ "step": 2699
+ },
+ {
+ "epoch": 37.50305676855895,
+ "grad_norm": 0.019150668755173683,
+ "learning_rate": 0.0006,
+ "loss": 4.097885608673096,
+ "step": 2700
+ },
+ {
+ "epoch": 37.51703056768559,
+ "grad_norm": 0.018847715109586716,
+ "learning_rate": 0.0006,
+ "loss": 4.054081439971924,
+ "step": 2701
+ },
+ {
+ "epoch": 37.531004366812226,
+ "grad_norm": 0.019607288762927055,
+ "learning_rate": 0.0006,
+ "loss": 3.9891440868377686,
+ "step": 2702
+ },
+ {
+ "epoch": 37.544978165938865,
+ "grad_norm": 0.018863150849938393,
+ "learning_rate": 0.0006,
+ "loss": 4.084877014160156,
+ "step": 2703
+ },
+ {
+ "epoch": 37.5589519650655,
+ "grad_norm": 0.017340239137411118,
+ "learning_rate": 0.0006,
+ "loss": 4.113290786743164,
+ "step": 2704
+ },
+ {
+ "epoch": 37.57292576419214,
+ "grad_norm": 0.015628693625330925,
+ "learning_rate": 0.0006,
+ "loss": 4.1483001708984375,
+ "step": 2705
+ },
+ {
+ "epoch": 37.58689956331878,
+ "grad_norm": 0.016493283212184906,
+ "learning_rate": 0.0006,
+ "loss": 4.1040143966674805,
+ "step": 2706
+ },
+ {
+ "epoch": 37.60087336244541,
+ "grad_norm": 0.016732443124055862,
+ "learning_rate": 0.0006,
+ "loss": 4.07205057144165,
+ "step": 2707
+ },
+ {
+ "epoch": 37.61484716157205,
+ "grad_norm": 0.017176156863570213,
+ "learning_rate": 0.0006,
+ "loss": 4.1279754638671875,
+ "step": 2708
+ },
+ {
+ "epoch": 37.62882096069869,
+ "grad_norm": 0.015730608254671097,
+ "learning_rate": 0.0006,
+ "loss": 4.086658000946045,
+ "step": 2709
+ },
+ {
+ "epoch": 37.64279475982533,
+ "grad_norm": 0.016111090779304504,
+ "learning_rate": 0.0006,
+ "loss": 4.064335823059082,
+ "step": 2710
+ },
+ {
+ "epoch": 37.65676855895197,
+ "grad_norm": 0.01769379712641239,
+ "learning_rate": 0.0006,
+ "loss": 4.116047382354736,
+ "step": 2711
+ },
+ {
+ "epoch": 37.670742358078606,
+ "grad_norm": 0.01683102548122406,
+ "learning_rate": 0.0006,
+ "loss": 4.095986843109131,
+ "step": 2712
+ },
+ {
+ "epoch": 37.68471615720524,
+ "grad_norm": 0.015726450830698013,
+ "learning_rate": 0.0006,
+ "loss": 4.10775089263916,
+ "step": 2713
+ },
+ {
+ "epoch": 37.698689956331876,
+ "grad_norm": 0.014760235324501991,
+ "learning_rate": 0.0006,
+ "loss": 4.1234893798828125,
+ "step": 2714
+ },
+ {
+ "epoch": 37.712663755458514,
+ "grad_norm": 0.014933438040316105,
+ "learning_rate": 0.0006,
+ "loss": 4.079200744628906,
+ "step": 2715
+ },
+ {
+ "epoch": 37.72663755458515,
+ "grad_norm": 0.015723997727036476,
+ "learning_rate": 0.0006,
+ "loss": 4.07865047454834,
+ "step": 2716
+ },
+ {
+ "epoch": 37.74061135371179,
+ "grad_norm": 0.014539001509547234,
+ "learning_rate": 0.0006,
+ "loss": 4.036242485046387,
+ "step": 2717
+ },
+ {
+ "epoch": 37.75458515283843,
+ "grad_norm": 0.014844979159533978,
+ "learning_rate": 0.0006,
+ "loss": 4.1051483154296875,
+ "step": 2718
+ },
+ {
+ "epoch": 37.76855895196506,
+ "grad_norm": 0.015939844772219658,
+ "learning_rate": 0.0006,
+ "loss": 4.017253875732422,
+ "step": 2719
+ },
+ {
+ "epoch": 37.7825327510917,
+ "grad_norm": 0.014771555550396442,
+ "learning_rate": 0.0006,
+ "loss": 4.0694661140441895,
+ "step": 2720
+ },
+ {
+ "epoch": 37.79650655021834,
+ "grad_norm": 0.015777921304106712,
+ "learning_rate": 0.0006,
+ "loss": 3.947587490081787,
+ "step": 2721
+ },
+ {
+ "epoch": 37.81048034934498,
+ "grad_norm": 0.017451291903853416,
+ "learning_rate": 0.0006,
+ "loss": 4.093625068664551,
+ "step": 2722
+ },
+ {
+ "epoch": 37.82445414847162,
+ "grad_norm": 0.016829246655106544,
+ "learning_rate": 0.0006,
+ "loss": 4.09030818939209,
+ "step": 2723
+ },
+ {
+ "epoch": 37.838427947598255,
+ "grad_norm": 0.016984397545456886,
+ "learning_rate": 0.0006,
+ "loss": 4.140626907348633,
+ "step": 2724
+ },
+ {
+ "epoch": 37.852401746724894,
+ "grad_norm": 0.01732603646814823,
+ "learning_rate": 0.0006,
+ "loss": 4.128276824951172,
+ "step": 2725
+ },
+ {
+ "epoch": 37.866375545851525,
+ "grad_norm": 0.01744619384407997,
+ "learning_rate": 0.0006,
+ "loss": 4.103130340576172,
+ "step": 2726
+ },
+ {
+ "epoch": 37.880349344978164,
+ "grad_norm": 0.016331320628523827,
+ "learning_rate": 0.0006,
+ "loss": 4.037951469421387,
+ "step": 2727
+ },
+ {
+ "epoch": 37.8943231441048,
+ "grad_norm": 0.015919430181384087,
+ "learning_rate": 0.0006,
+ "loss": 4.121513366699219,
+ "step": 2728
+ },
+ {
+ "epoch": 37.90829694323144,
+ "grad_norm": 0.016153695061802864,
+ "learning_rate": 0.0006,
+ "loss": 4.099384307861328,
+ "step": 2729
+ },
+ {
+ "epoch": 37.92227074235808,
+ "grad_norm": 0.01628800667822361,
+ "learning_rate": 0.0006,
+ "loss": 4.092325687408447,
+ "step": 2730
+ },
+ {
+ "epoch": 37.93624454148472,
+ "grad_norm": 0.017285684123635292,
+ "learning_rate": 0.0006,
+ "loss": 3.9828906059265137,
+ "step": 2731
+ },
+ {
+ "epoch": 37.95021834061135,
+ "grad_norm": 0.016047554090619087,
+ "learning_rate": 0.0006,
+ "loss": 4.100419044494629,
+ "step": 2732
+ },
+ {
+ "epoch": 37.96419213973799,
+ "grad_norm": 0.01562986709177494,
+ "learning_rate": 0.0006,
+ "loss": 4.065832614898682,
+ "step": 2733
+ },
+ {
+ "epoch": 37.97816593886463,
+ "grad_norm": 0.01677163504064083,
+ "learning_rate": 0.0006,
+ "loss": 4.077518463134766,
+ "step": 2734
+ },
+ {
+ "epoch": 37.992139737991266,
+ "grad_norm": 0.01755492016673088,
+ "learning_rate": 0.0006,
+ "loss": 4.1076154708862305,
+ "step": 2735
+ },
+ {
+ "epoch": 38.0,
+ "grad_norm": 0.01929762214422226,
+ "learning_rate": 0.0006,
+ "loss": 4.09603214263916,
+ "step": 2736
+ },
+ {
+ "epoch": 38.0,
+ "eval_loss": 4.582442283630371,
+ "eval_runtime": 62.8649,
+ "eval_samples_per_second": 38.845,
+ "eval_steps_per_second": 1.225,
+ "step": 2736
+ },
+ {
+ "epoch": 38.01397379912664,
+ "grad_norm": 0.019129136577248573,
+ "learning_rate": 0.0006,
+ "loss": 4.095005989074707,
+ "step": 2737
+ },
+ {
+ "epoch": 38.02794759825328,
+ "grad_norm": 0.019167408347129822,
+ "learning_rate": 0.0006,
+ "loss": 4.054426670074463,
+ "step": 2738
+ },
+ {
+ "epoch": 38.041921397379916,
+ "grad_norm": 0.01789460889995098,
+ "learning_rate": 0.0006,
+ "loss": 4.0434417724609375,
+ "step": 2739
+ },
+ {
+ "epoch": 38.05589519650655,
+ "grad_norm": 0.015947027131915092,
+ "learning_rate": 0.0006,
+ "loss": 4.074516296386719,
+ "step": 2740
+ },
+ {
+ "epoch": 38.069868995633186,
+ "grad_norm": 0.016667252406477928,
+ "learning_rate": 0.0006,
+ "loss": 4.108545303344727,
+ "step": 2741
+ },
+ {
+ "epoch": 38.083842794759825,
+ "grad_norm": 0.016014399006962776,
+ "learning_rate": 0.0006,
+ "loss": 4.026925086975098,
+ "step": 2742
+ },
+ {
+ "epoch": 38.09781659388646,
+ "grad_norm": 0.015397695824503899,
+ "learning_rate": 0.0006,
+ "loss": 4.163970470428467,
+ "step": 2743
+ },
+ {
+ "epoch": 38.1117903930131,
+ "grad_norm": 0.01563289947807789,
+ "learning_rate": 0.0006,
+ "loss": 4.116578578948975,
+ "step": 2744
+ },
+ {
+ "epoch": 38.12576419213974,
+ "grad_norm": 0.015512318350374699,
+ "learning_rate": 0.0006,
+ "loss": 4.096311569213867,
+ "step": 2745
+ },
+ {
+ "epoch": 38.13973799126637,
+ "grad_norm": 0.015426642261445522,
+ "learning_rate": 0.0006,
+ "loss": 4.017065048217773,
+ "step": 2746
+ },
+ {
+ "epoch": 38.15371179039301,
+ "grad_norm": 0.015115435235202312,
+ "learning_rate": 0.0006,
+ "loss": 4.005434989929199,
+ "step": 2747
+ },
+ {
+ "epoch": 38.16768558951965,
+ "grad_norm": 0.015235894359648228,
+ "learning_rate": 0.0006,
+ "loss": 3.982743263244629,
+ "step": 2748
+ },
+ {
+ "epoch": 38.18165938864629,
+ "grad_norm": 0.016095101833343506,
+ "learning_rate": 0.0006,
+ "loss": 4.0849385261535645,
+ "step": 2749
+ },
+ {
+ "epoch": 38.19563318777293,
+ "grad_norm": 0.015201197937130928,
+ "learning_rate": 0.0006,
+ "loss": 4.103863716125488,
+ "step": 2750
+ },
+ {
+ "epoch": 38.209606986899566,
+ "grad_norm": 0.015687478706240654,
+ "learning_rate": 0.0006,
+ "loss": 4.12074089050293,
+ "step": 2751
+ },
+ {
+ "epoch": 38.223580786026204,
+ "grad_norm": 0.015201589092612267,
+ "learning_rate": 0.0006,
+ "loss": 4.056872367858887,
+ "step": 2752
+ },
+ {
+ "epoch": 38.237554585152836,
+ "grad_norm": 0.015154894441366196,
+ "learning_rate": 0.0006,
+ "loss": 4.089670181274414,
+ "step": 2753
+ },
+ {
+ "epoch": 38.251528384279474,
+ "grad_norm": 0.01586608588695526,
+ "learning_rate": 0.0006,
+ "loss": 4.155317306518555,
+ "step": 2754
+ },
+ {
+ "epoch": 38.26550218340611,
+ "grad_norm": 0.01609654724597931,
+ "learning_rate": 0.0006,
+ "loss": 3.995800495147705,
+ "step": 2755
+ },
+ {
+ "epoch": 38.27947598253275,
+ "grad_norm": 0.015495185740292072,
+ "learning_rate": 0.0006,
+ "loss": 3.995769739151001,
+ "step": 2756
+ },
+ {
+ "epoch": 38.29344978165939,
+ "grad_norm": 0.017042381688952446,
+ "learning_rate": 0.0006,
+ "loss": 4.132832050323486,
+ "step": 2757
+ },
+ {
+ "epoch": 38.30742358078603,
+ "grad_norm": 0.018706614151597023,
+ "learning_rate": 0.0006,
+ "loss": 4.038820743560791,
+ "step": 2758
+ },
+ {
+ "epoch": 38.32139737991266,
+ "grad_norm": 0.018316026777029037,
+ "learning_rate": 0.0006,
+ "loss": 3.972052574157715,
+ "step": 2759
+ },
+ {
+ "epoch": 38.3353711790393,
+ "grad_norm": 0.018375389277935028,
+ "learning_rate": 0.0006,
+ "loss": 4.152097702026367,
+ "step": 2760
+ },
+ {
+ "epoch": 38.34934497816594,
+ "grad_norm": 0.016879886388778687,
+ "learning_rate": 0.0006,
+ "loss": 4.020869731903076,
+ "step": 2761
+ },
+ {
+ "epoch": 38.36331877729258,
+ "grad_norm": 0.015728183090686798,
+ "learning_rate": 0.0006,
+ "loss": 4.008016586303711,
+ "step": 2762
+ },
+ {
+ "epoch": 38.377292576419215,
+ "grad_norm": 0.016147758811712265,
+ "learning_rate": 0.0006,
+ "loss": 4.035854339599609,
+ "step": 2763
+ },
+ {
+ "epoch": 38.391266375545854,
+ "grad_norm": 0.017774797976017,
+ "learning_rate": 0.0006,
+ "loss": 4.122506141662598,
+ "step": 2764
+ },
+ {
+ "epoch": 38.40524017467249,
+ "grad_norm": 0.016590069979429245,
+ "learning_rate": 0.0006,
+ "loss": 4.082847595214844,
+ "step": 2765
+ },
+ {
+ "epoch": 38.419213973799124,
+ "grad_norm": 0.016544528305530548,
+ "learning_rate": 0.0006,
+ "loss": 4.184270858764648,
+ "step": 2766
+ },
+ {
+ "epoch": 38.43318777292576,
+ "grad_norm": 0.017972925677895546,
+ "learning_rate": 0.0006,
+ "loss": 3.987558603286743,
+ "step": 2767
+ },
+ {
+ "epoch": 38.4471615720524,
+ "grad_norm": 0.018562227487564087,
+ "learning_rate": 0.0006,
+ "loss": 4.1193461418151855,
+ "step": 2768
+ },
+ {
+ "epoch": 38.46113537117904,
+ "grad_norm": 0.018896685913205147,
+ "learning_rate": 0.0006,
+ "loss": 4.057808876037598,
+ "step": 2769
+ },
+ {
+ "epoch": 38.47510917030568,
+ "grad_norm": 0.016793809831142426,
+ "learning_rate": 0.0006,
+ "loss": 4.069523334503174,
+ "step": 2770
+ },
+ {
+ "epoch": 38.48908296943232,
+ "grad_norm": 0.01698676310479641,
+ "learning_rate": 0.0006,
+ "loss": 3.997624158859253,
+ "step": 2771
+ },
+ {
+ "epoch": 38.50305676855895,
+ "grad_norm": 0.016784202307462692,
+ "learning_rate": 0.0006,
+ "loss": 4.174795627593994,
+ "step": 2772
+ },
+ {
+ "epoch": 38.51703056768559,
+ "grad_norm": 0.016785014420747757,
+ "learning_rate": 0.0006,
+ "loss": 3.999321222305298,
+ "step": 2773
+ },
+ {
+ "epoch": 38.531004366812226,
+ "grad_norm": 0.018615184351801872,
+ "learning_rate": 0.0006,
+ "loss": 4.191237926483154,
+ "step": 2774
+ },
+ {
+ "epoch": 38.544978165938865,
+ "grad_norm": 0.019090790301561356,
+ "learning_rate": 0.0006,
+ "loss": 4.106966018676758,
+ "step": 2775
+ },
+ {
+ "epoch": 38.5589519650655,
+ "grad_norm": 0.018298320472240448,
+ "learning_rate": 0.0006,
+ "loss": 4.063363552093506,
+ "step": 2776
+ },
+ {
+ "epoch": 38.57292576419214,
+ "grad_norm": 0.018541090190410614,
+ "learning_rate": 0.0006,
+ "loss": 4.101193428039551,
+ "step": 2777
+ },
+ {
+ "epoch": 38.58689956331878,
+ "grad_norm": 0.01849970407783985,
+ "learning_rate": 0.0006,
+ "loss": 4.071335792541504,
+ "step": 2778
+ },
+ {
+ "epoch": 38.60087336244541,
+ "grad_norm": 0.018309077247977257,
+ "learning_rate": 0.0006,
+ "loss": 4.245305061340332,
+ "step": 2779
+ },
+ {
+ "epoch": 38.61484716157205,
+ "grad_norm": 0.017681840807199478,
+ "learning_rate": 0.0006,
+ "loss": 4.081173419952393,
+ "step": 2780
+ },
+ {
+ "epoch": 38.62882096069869,
+ "grad_norm": 0.016286224126815796,
+ "learning_rate": 0.0006,
+ "loss": 4.024351596832275,
+ "step": 2781
+ },
+ {
+ "epoch": 38.64279475982533,
+ "grad_norm": 0.015837235376238823,
+ "learning_rate": 0.0006,
+ "loss": 4.028205394744873,
+ "step": 2782
+ },
+ {
+ "epoch": 38.65676855895197,
+ "grad_norm": 0.016062596812844276,
+ "learning_rate": 0.0006,
+ "loss": 4.131122589111328,
+ "step": 2783
+ },
+ {
+ "epoch": 38.670742358078606,
+ "grad_norm": 0.01711239665746689,
+ "learning_rate": 0.0006,
+ "loss": 4.124416351318359,
+ "step": 2784
+ },
+ {
+ "epoch": 38.68471615720524,
+ "grad_norm": 0.015304913744330406,
+ "learning_rate": 0.0006,
+ "loss": 4.056122779846191,
+ "step": 2785
+ },
+ {
+ "epoch": 38.698689956331876,
+ "grad_norm": 0.015971887856721878,
+ "learning_rate": 0.0006,
+ "loss": 4.132555961608887,
+ "step": 2786
+ },
+ {
+ "epoch": 38.712663755458514,
+ "grad_norm": 0.015787221491336823,
+ "learning_rate": 0.0006,
+ "loss": 4.051596164703369,
+ "step": 2787
+ },
+ {
+ "epoch": 38.72663755458515,
+ "grad_norm": 0.01605444774031639,
+ "learning_rate": 0.0006,
+ "loss": 4.191808700561523,
+ "step": 2788
+ },
+ {
+ "epoch": 38.74061135371179,
+ "grad_norm": 0.016383009031414986,
+ "learning_rate": 0.0006,
+ "loss": 4.0648417472839355,
+ "step": 2789
+ },
+ {
+ "epoch": 38.75458515283843,
+ "grad_norm": 0.016135554760694504,
+ "learning_rate": 0.0006,
+ "loss": 3.988563060760498,
+ "step": 2790
+ },
+ {
+ "epoch": 38.76855895196506,
+ "grad_norm": 0.016634158790111542,
+ "learning_rate": 0.0006,
+ "loss": 4.133493423461914,
+ "step": 2791
+ },
+ {
+ "epoch": 38.7825327510917,
+ "grad_norm": 0.01688210666179657,
+ "learning_rate": 0.0006,
+ "loss": 4.091277122497559,
+ "step": 2792
+ },
+ {
+ "epoch": 38.79650655021834,
+ "grad_norm": 0.016102122142910957,
+ "learning_rate": 0.0006,
+ "loss": 4.051219940185547,
+ "step": 2793
+ },
+ {
+ "epoch": 38.81048034934498,
+ "grad_norm": 0.016585038974881172,
+ "learning_rate": 0.0006,
+ "loss": 4.058587074279785,
+ "step": 2794
+ },
+ {
+ "epoch": 38.82445414847162,
+ "grad_norm": 0.017700951546430588,
+ "learning_rate": 0.0006,
+ "loss": 4.016030311584473,
+ "step": 2795
+ },
+ {
+ "epoch": 38.838427947598255,
+ "grad_norm": 0.017922203987836838,
+ "learning_rate": 0.0006,
+ "loss": 4.079977989196777,
+ "step": 2796
+ },
+ {
+ "epoch": 38.852401746724894,
+ "grad_norm": 0.017299890518188477,
+ "learning_rate": 0.0006,
+ "loss": 4.062984943389893,
+ "step": 2797
+ },
+ {
+ "epoch": 38.866375545851525,
+ "grad_norm": 0.018648581579327583,
+ "learning_rate": 0.0006,
+ "loss": 4.065717697143555,
+ "step": 2798
+ },
+ {
+ "epoch": 38.880349344978164,
+ "grad_norm": 0.019057920202612877,
+ "learning_rate": 0.0006,
+ "loss": 4.227892875671387,
+ "step": 2799
+ },
+ {
+ "epoch": 38.8943231441048,
+ "grad_norm": 0.018824422731995583,
+ "learning_rate": 0.0006,
+ "loss": 4.116058349609375,
+ "step": 2800
+ },
+ {
+ "epoch": 38.90829694323144,
+ "grad_norm": 0.019084978848695755,
+ "learning_rate": 0.0006,
+ "loss": 4.061677932739258,
+ "step": 2801
+ },
+ {
+ "epoch": 38.92227074235808,
+ "grad_norm": 0.017842981964349747,
+ "learning_rate": 0.0006,
+ "loss": 4.083813667297363,
+ "step": 2802
+ },
+ {
+ "epoch": 38.93624454148472,
+ "grad_norm": 0.01880858838558197,
+ "learning_rate": 0.0006,
+ "loss": 4.152561187744141,
+ "step": 2803
+ },
+ {
+ "epoch": 38.95021834061135,
+ "grad_norm": 0.017400087788701057,
+ "learning_rate": 0.0006,
+ "loss": 4.154635429382324,
+ "step": 2804
+ },
+ {
+ "epoch": 38.96419213973799,
+ "grad_norm": 0.017682507634162903,
+ "learning_rate": 0.0006,
+ "loss": 4.151128768920898,
+ "step": 2805
+ },
+ {
+ "epoch": 38.97816593886463,
+ "grad_norm": 0.015803052112460136,
+ "learning_rate": 0.0006,
+ "loss": 4.10498571395874,
+ "step": 2806
+ },
+ {
+ "epoch": 38.992139737991266,
+ "grad_norm": 0.01582511141896248,
+ "learning_rate": 0.0006,
+ "loss": 4.140159606933594,
+ "step": 2807
+ },
+ {
+ "epoch": 39.0,
+ "grad_norm": 0.01925666816532612,
+ "learning_rate": 0.0006,
+ "loss": 4.126394271850586,
+ "step": 2808
+ },
+ {
+ "epoch": 39.0,
+ "eval_loss": 4.615617275238037,
+ "eval_runtime": 56.8895,
+ "eval_samples_per_second": 42.925,
+ "eval_steps_per_second": 1.354,
+ "step": 2808
+ },
+ {
+ "epoch": 39.01397379912664,
+ "grad_norm": 0.01677853614091873,
+ "learning_rate": 0.0006,
+ "loss": 4.102395057678223,
+ "step": 2809
+ },
+ {
+ "epoch": 39.02794759825328,
+ "grad_norm": 0.017901137471199036,
+ "learning_rate": 0.0006,
+ "loss": 3.976609230041504,
+ "step": 2810
+ },
+ {
+ "epoch": 39.041921397379916,
+ "grad_norm": 0.017978668212890625,
+ "learning_rate": 0.0006,
+ "loss": 3.944416046142578,
+ "step": 2811
+ },
+ {
+ "epoch": 39.05589519650655,
+ "grad_norm": 0.019967492669820786,
+ "learning_rate": 0.0006,
+ "loss": 3.992234468460083,
+ "step": 2812
+ },
+ {
+ "epoch": 39.069868995633186,
+ "grad_norm": 0.018937544897198677,
+ "learning_rate": 0.0006,
+ "loss": 4.028573036193848,
+ "step": 2813
+ },
+ {
+ "epoch": 39.083842794759825,
+ "grad_norm": 0.016852952539920807,
+ "learning_rate": 0.0006,
+ "loss": 4.062442779541016,
+ "step": 2814
+ },
+ {
+ "epoch": 39.09781659388646,
+ "grad_norm": 0.016643647104501724,
+ "learning_rate": 0.0006,
+ "loss": 4.08871603012085,
+ "step": 2815
+ },
+ {
+ "epoch": 39.1117903930131,
+ "grad_norm": 0.016872866079211235,
+ "learning_rate": 0.0006,
+ "loss": 3.9514245986938477,
+ "step": 2816
+ },
+ {
+ "epoch": 39.12576419213974,
+ "grad_norm": 0.018045669421553612,
+ "learning_rate": 0.0006,
+ "loss": 4.038842678070068,
+ "step": 2817
+ },
+ {
+ "epoch": 39.13973799126637,
+ "grad_norm": 0.017853403463959694,
+ "learning_rate": 0.0006,
+ "loss": 3.93709135055542,
+ "step": 2818
+ },
+ {
+ "epoch": 39.15371179039301,
+ "grad_norm": 0.017507443204522133,
+ "learning_rate": 0.0006,
+ "loss": 3.98665714263916,
+ "step": 2819
+ },
+ {
+ "epoch": 39.16768558951965,
+ "grad_norm": 0.01766853965818882,
+ "learning_rate": 0.0006,
+ "loss": 3.937126636505127,
+ "step": 2820
+ },
+ {
+ "epoch": 39.18165938864629,
+ "grad_norm": 0.016660021618008614,
+ "learning_rate": 0.0006,
+ "loss": 3.8397014141082764,
+ "step": 2821
+ },
+ {
+ "epoch": 39.19563318777293,
+ "grad_norm": 0.017719309777021408,
+ "learning_rate": 0.0006,
+ "loss": 4.003579616546631,
+ "step": 2822
+ },
+ {
+ "epoch": 39.209606986899566,
+ "grad_norm": 0.018234414979815483,
+ "learning_rate": 0.0006,
+ "loss": 4.053859710693359,
+ "step": 2823
+ },
+ {
+ "epoch": 39.223580786026204,
+ "grad_norm": 0.018777024000883102,
+ "learning_rate": 0.0006,
+ "loss": 4.056182861328125,
+ "step": 2824
+ },
+ {
+ "epoch": 39.237554585152836,
+ "grad_norm": 0.017443502321839333,
+ "learning_rate": 0.0006,
+ "loss": 4.025068283081055,
+ "step": 2825
+ },
+ {
+ "epoch": 39.251528384279474,
+ "grad_norm": 0.016500331461429596,
+ "learning_rate": 0.0006,
+ "loss": 4.101631164550781,
+ "step": 2826
+ },
+ {
+ "epoch": 39.26550218340611,
+ "grad_norm": 0.016101544722914696,
+ "learning_rate": 0.0006,
+ "loss": 4.027082443237305,
+ "step": 2827
+ },
+ {
+ "epoch": 39.27947598253275,
+ "grad_norm": 0.015471171587705612,
+ "learning_rate": 0.0006,
+ "loss": 4.022909641265869,
+ "step": 2828
+ },
+ {
+ "epoch": 39.29344978165939,
+ "grad_norm": 0.015779748558998108,
+ "learning_rate": 0.0006,
+ "loss": 4.18796968460083,
+ "step": 2829
+ },
+ {
+ "epoch": 39.30742358078603,
+ "grad_norm": 0.015522679314017296,
+ "learning_rate": 0.0006,
+ "loss": 4.005777359008789,
+ "step": 2830
+ },
+ {
+ "epoch": 39.32139737991266,
+ "grad_norm": 0.015226504765450954,
+ "learning_rate": 0.0006,
+ "loss": 4.108416557312012,
+ "step": 2831
+ },
+ {
+ "epoch": 39.3353711790393,
+ "grad_norm": 0.015095778740942478,
+ "learning_rate": 0.0006,
+ "loss": 4.08029842376709,
+ "step": 2832
+ },
+ {
+ "epoch": 39.34934497816594,
+ "grad_norm": 0.015601031482219696,
+ "learning_rate": 0.0006,
+ "loss": 3.989492416381836,
+ "step": 2833
+ },
+ {
+ "epoch": 39.36331877729258,
+ "grad_norm": 0.014825052581727505,
+ "learning_rate": 0.0006,
+ "loss": 3.998520612716675,
+ "step": 2834
+ },
+ {
+ "epoch": 39.377292576419215,
+ "grad_norm": 0.014821100980043411,
+ "learning_rate": 0.0006,
+ "loss": 4.001777172088623,
+ "step": 2835
+ },
+ {
+ "epoch": 39.391266375545854,
+ "grad_norm": 0.015333754941821098,
+ "learning_rate": 0.0006,
+ "loss": 4.11985969543457,
+ "step": 2836
+ },
+ {
+ "epoch": 39.40524017467249,
+ "grad_norm": 0.015940699726343155,
+ "learning_rate": 0.0006,
+ "loss": 4.025373935699463,
+ "step": 2837
+ },
+ {
+ "epoch": 39.419213973799124,
+ "grad_norm": 0.015096287243068218,
+ "learning_rate": 0.0006,
+ "loss": 4.068551063537598,
+ "step": 2838
+ },
+ {
+ "epoch": 39.43318777292576,
+ "grad_norm": 0.01533548254519701,
+ "learning_rate": 0.0006,
+ "loss": 3.9800801277160645,
+ "step": 2839
+ },
+ {
+ "epoch": 39.4471615720524,
+ "grad_norm": 0.014648732729256153,
+ "learning_rate": 0.0006,
+ "loss": 3.9670190811157227,
+ "step": 2840
+ },
+ {
+ "epoch": 39.46113537117904,
+ "grad_norm": 0.015759671106934547,
+ "learning_rate": 0.0006,
+ "loss": 4.050443649291992,
+ "step": 2841
+ },
+ {
+ "epoch": 39.47510917030568,
+ "grad_norm": 0.016054073348641396,
+ "learning_rate": 0.0006,
+ "loss": 4.22734260559082,
+ "step": 2842
+ },
+ {
+ "epoch": 39.48908296943232,
+ "grad_norm": 0.0160264540463686,
+ "learning_rate": 0.0006,
+ "loss": 4.116174221038818,
+ "step": 2843
+ },
+ {
+ "epoch": 39.50305676855895,
+ "grad_norm": 0.015426610596477985,
+ "learning_rate": 0.0006,
+ "loss": 4.000178337097168,
+ "step": 2844
+ },
+ {
+ "epoch": 39.51703056768559,
+ "grad_norm": 0.01520642638206482,
+ "learning_rate": 0.0006,
+ "loss": 4.218410968780518,
+ "step": 2845
+ },
+ {
+ "epoch": 39.531004366812226,
+ "grad_norm": 0.015305949375033379,
+ "learning_rate": 0.0006,
+ "loss": 4.025891304016113,
+ "step": 2846
+ },
+ {
+ "epoch": 39.544978165938865,
+ "grad_norm": 0.015461564064025879,
+ "learning_rate": 0.0006,
+ "loss": 4.100950717926025,
+ "step": 2847
+ },
+ {
+ "epoch": 39.5589519650655,
+ "grad_norm": 0.01712625101208687,
+ "learning_rate": 0.0006,
+ "loss": 4.0620856285095215,
+ "step": 2848
+ },
+ {
+ "epoch": 39.57292576419214,
+ "grad_norm": 0.017133301123976707,
+ "learning_rate": 0.0006,
+ "loss": 4.097138404846191,
+ "step": 2849
+ },
+ {
+ "epoch": 39.58689956331878,
+ "grad_norm": 0.01683625392615795,
+ "learning_rate": 0.0006,
+ "loss": 4.084364891052246,
+ "step": 2850
+ },
+ {
+ "epoch": 39.60087336244541,
+ "grad_norm": 0.0165710486471653,
+ "learning_rate": 0.0006,
+ "loss": 4.042361259460449,
+ "step": 2851
+ },
+ {
+ "epoch": 39.61484716157205,
+ "grad_norm": 0.01660393364727497,
+ "learning_rate": 0.0006,
+ "loss": 4.1395487785339355,
+ "step": 2852
+ },
+ {
+ "epoch": 39.62882096069869,
+ "grad_norm": 0.017025845125317574,
+ "learning_rate": 0.0006,
+ "loss": 4.1207075119018555,
+ "step": 2853
+ },
+ {
+ "epoch": 39.64279475982533,
+ "grad_norm": 0.01660497486591339,
+ "learning_rate": 0.0006,
+ "loss": 4.103846073150635,
+ "step": 2854
+ },
+ {
+ "epoch": 39.65676855895197,
+ "grad_norm": 0.01607946865260601,
+ "learning_rate": 0.0006,
+ "loss": 4.189016342163086,
+ "step": 2855
+ },
+ {
+ "epoch": 39.670742358078606,
+ "grad_norm": 0.01614505983889103,
+ "learning_rate": 0.0006,
+ "loss": 4.079860687255859,
+ "step": 2856
+ },
+ {
+ "epoch": 39.68471615720524,
+ "grad_norm": 0.017085086554288864,
+ "learning_rate": 0.0006,
+ "loss": 3.9774293899536133,
+ "step": 2857
+ },
+ {
+ "epoch": 39.698689956331876,
+ "grad_norm": 0.01816520094871521,
+ "learning_rate": 0.0006,
+ "loss": 4.113666534423828,
+ "step": 2858
+ },
+ {
+ "epoch": 39.712663755458514,
+ "grad_norm": 0.016597526147961617,
+ "learning_rate": 0.0006,
+ "loss": 4.068242073059082,
+ "step": 2859
+ },
+ {
+ "epoch": 39.72663755458515,
+ "grad_norm": 0.01672235131263733,
+ "learning_rate": 0.0006,
+ "loss": 4.024444580078125,
+ "step": 2860
+ },
+ {
+ "epoch": 39.74061135371179,
+ "grad_norm": 0.020093556493520737,
+ "learning_rate": 0.0006,
+ "loss": 4.064124584197998,
+ "step": 2861
+ },
+ {
+ "epoch": 39.75458515283843,
+ "grad_norm": 0.023141304031014442,
+ "learning_rate": 0.0006,
+ "loss": 4.130439758300781,
+ "step": 2862
+ },
+ {
+ "epoch": 39.76855895196506,
+ "grad_norm": 0.02289372682571411,
+ "learning_rate": 0.0006,
+ "loss": 4.14794921875,
+ "step": 2863
+ },
+ {
+ "epoch": 39.7825327510917,
+ "grad_norm": 0.01919020712375641,
+ "learning_rate": 0.0006,
+ "loss": 4.123961925506592,
+ "step": 2864
+ },
+ {
+ "epoch": 39.79650655021834,
+ "grad_norm": 0.016932329162955284,
+ "learning_rate": 0.0006,
+ "loss": 4.0944132804870605,
+ "step": 2865
+ },
+ {
+ "epoch": 39.81048034934498,
+ "grad_norm": 0.016644535586237907,
+ "learning_rate": 0.0006,
+ "loss": 4.124269962310791,
+ "step": 2866
+ },
+ {
+ "epoch": 39.82445414847162,
+ "grad_norm": 0.016302216798067093,
+ "learning_rate": 0.0006,
+ "loss": 4.0267205238342285,
+ "step": 2867
+ },
+ {
+ "epoch": 39.838427947598255,
+ "grad_norm": 0.01673300378024578,
+ "learning_rate": 0.0006,
+ "loss": 4.10650634765625,
+ "step": 2868
+ },
+ {
+ "epoch": 39.852401746724894,
+ "grad_norm": 0.016581356525421143,
+ "learning_rate": 0.0006,
+ "loss": 4.1547956466674805,
+ "step": 2869
+ },
+ {
+ "epoch": 39.866375545851525,
+ "grad_norm": 0.01691797934472561,
+ "learning_rate": 0.0006,
+ "loss": 4.056936740875244,
+ "step": 2870
+ },
+ {
+ "epoch": 39.880349344978164,
+ "grad_norm": 0.01596726104617119,
+ "learning_rate": 0.0006,
+ "loss": 3.9891414642333984,
+ "step": 2871
+ },
+ {
+ "epoch": 39.8943231441048,
+ "grad_norm": 0.015615091659128666,
+ "learning_rate": 0.0006,
+ "loss": 4.137948036193848,
+ "step": 2872
+ },
+ {
+ "epoch": 39.90829694323144,
+ "grad_norm": 0.015703413635492325,
+ "learning_rate": 0.0006,
+ "loss": 3.9844930171966553,
+ "step": 2873
+ },
+ {
+ "epoch": 39.92227074235808,
+ "grad_norm": 0.015453094616532326,
+ "learning_rate": 0.0006,
+ "loss": 4.098443508148193,
+ "step": 2874
+ },
+ {
+ "epoch": 39.93624454148472,
+ "grad_norm": 0.01449158601462841,
+ "learning_rate": 0.0006,
+ "loss": 4.090852737426758,
+ "step": 2875
+ },
+ {
+ "epoch": 39.95021834061135,
+ "grad_norm": 0.01537515688687563,
+ "learning_rate": 0.0006,
+ "loss": 4.239612579345703,
+ "step": 2876
+ },
+ {
+ "epoch": 39.96419213973799,
+ "grad_norm": 0.015989921987056732,
+ "learning_rate": 0.0006,
+ "loss": 4.142073154449463,
+ "step": 2877
+ },
+ {
+ "epoch": 39.97816593886463,
+ "grad_norm": 0.015985611826181412,
+ "learning_rate": 0.0006,
+ "loss": 4.100076675415039,
+ "step": 2878
+ },
+ {
+ "epoch": 39.992139737991266,
+ "grad_norm": 0.015260833315551281,
+ "learning_rate": 0.0006,
+ "loss": 4.068576335906982,
+ "step": 2879
+ },
+ {
+ "epoch": 40.0,
+ "grad_norm": 0.018047941848635674,
+ "learning_rate": 0.0006,
+ "loss": 4.005878925323486,
+ "step": 2880
+ },
+ {
+ "epoch": 40.0,
+ "eval_loss": 4.622848987579346,
+ "eval_runtime": 56.9618,
+ "eval_samples_per_second": 42.871,
+ "eval_steps_per_second": 1.352,
+ "step": 2880
+ },
+ {
+ "epoch": 40.01397379912664,
+ "grad_norm": 0.015976838767528534,
+ "learning_rate": 0.0006,
+ "loss": 4.035689353942871,
+ "step": 2881
+ },
+ {
+ "epoch": 40.02794759825328,
+ "grad_norm": 0.015176220797002316,
+ "learning_rate": 0.0006,
+ "loss": 3.932197332382202,
+ "step": 2882
+ },
+ {
+ "epoch": 40.041921397379916,
+ "grad_norm": 0.016676228493452072,
+ "learning_rate": 0.0006,
+ "loss": 4.014171123504639,
+ "step": 2883
+ },
+ {
+ "epoch": 40.05589519650655,
+ "grad_norm": 0.01698281429708004,
+ "learning_rate": 0.0006,
+ "loss": 4.117893218994141,
+ "step": 2884
+ },
+ {
+ "epoch": 40.069868995633186,
+ "grad_norm": 0.017872486263513565,
+ "learning_rate": 0.0006,
+ "loss": 4.132119178771973,
+ "step": 2885
+ },
+ {
+ "epoch": 40.083842794759825,
+ "grad_norm": 0.017615554854273796,
+ "learning_rate": 0.0006,
+ "loss": 4.072000980377197,
+ "step": 2886
+ },
+ {
+ "epoch": 40.09781659388646,
+ "grad_norm": 0.017469394952058792,
+ "learning_rate": 0.0006,
+ "loss": 4.017726421356201,
+ "step": 2887
+ },
+ {
+ "epoch": 40.1117903930131,
+ "grad_norm": 0.01954762637615204,
+ "learning_rate": 0.0006,
+ "loss": 4.014443874359131,
+ "step": 2888
+ },
+ {
+ "epoch": 40.12576419213974,
+ "grad_norm": 0.018935853615403175,
+ "learning_rate": 0.0006,
+ "loss": 4.0042948722839355,
+ "step": 2889
+ },
+ {
+ "epoch": 40.13973799126637,
+ "grad_norm": 0.019382692873477936,
+ "learning_rate": 0.0006,
+ "loss": 4.06160831451416,
+ "step": 2890
+ },
+ {
+ "epoch": 40.15371179039301,
+ "grad_norm": 0.019677115604281425,
+ "learning_rate": 0.0006,
+ "loss": 4.13466215133667,
+ "step": 2891
+ },
+ {
+ "epoch": 40.16768558951965,
+ "grad_norm": 0.020365122705698013,
+ "learning_rate": 0.0006,
+ "loss": 4.095000267028809,
+ "step": 2892
+ },
+ {
+ "epoch": 40.18165938864629,
+ "grad_norm": 0.021842138841748238,
+ "learning_rate": 0.0006,
+ "loss": 3.9877853393554688,
+ "step": 2893
+ },
+ {
+ "epoch": 40.19563318777293,
+ "grad_norm": 0.02354881353676319,
+ "learning_rate": 0.0006,
+ "loss": 3.9430112838745117,
+ "step": 2894
+ },
+ {
+ "epoch": 40.209606986899566,
+ "grad_norm": 0.026602735742926598,
+ "learning_rate": 0.0006,
+ "loss": 4.012076377868652,
+ "step": 2895
+ },
+ {
+ "epoch": 40.223580786026204,
+ "grad_norm": 0.023976484313607216,
+ "learning_rate": 0.0006,
+ "loss": 3.9768595695495605,
+ "step": 2896
+ },
+ {
+ "epoch": 40.237554585152836,
+ "grad_norm": 0.02336616814136505,
+ "learning_rate": 0.0006,
+ "loss": 4.1096038818359375,
+ "step": 2897
+ },
+ {
+ "epoch": 40.251528384279474,
+ "grad_norm": 0.02426023595035076,
+ "learning_rate": 0.0006,
+ "loss": 4.157189846038818,
+ "step": 2898
+ },
+ {
+ "epoch": 40.26550218340611,
+ "grad_norm": 0.021572696045041084,
+ "learning_rate": 0.0006,
+ "loss": 3.979175090789795,
+ "step": 2899
+ },
+ {
+ "epoch": 40.27947598253275,
+ "grad_norm": 0.08300432562828064,
+ "learning_rate": 0.0006,
+ "loss": 4.092336654663086,
+ "step": 2900
+ },
+ {
+ "epoch": 40.29344978165939,
+ "grad_norm": 0.7323967814445496,
+ "learning_rate": 0.0006,
+ "loss": 5.337990760803223,
+ "step": 2901
+ },
+ {
+ "epoch": 40.30742358078603,
+ "grad_norm": 0.3363337814807892,
+ "learning_rate": 0.0006,
+ "loss": 6.40043830871582,
+ "step": 2902
+ },
+ {
+ "epoch": 40.32139737991266,
+ "grad_norm": 0.598686933517456,
+ "learning_rate": 0.0006,
+ "loss": 7.004765510559082,
+ "step": 2903
+ },
+ {
+ "epoch": 40.3353711790393,
+ "grad_norm": 0.5998364686965942,
+ "learning_rate": 0.0006,
+ "loss": 8.184747695922852,
+ "step": 2904
+ },
+ {
+ "epoch": 40.34934497816594,
+ "grad_norm": 1.408208966255188,
+ "learning_rate": 0.0006,
+ "loss": 8.88978385925293,
+ "step": 2905
+ },
+ {
+ "epoch": 40.36331877729258,
+ "grad_norm": 0.7829785943031311,
+ "learning_rate": 0.0006,
+ "loss": 11.13601303100586,
+ "step": 2906
+ },
+ {
+ "epoch": 40.377292576419215,
+ "grad_norm": 0.7769215106964111,
+ "learning_rate": 0.0006,
+ "loss": 9.90416145324707,
+ "step": 2907
+ },
+ {
+ "epoch": 40.391266375545854,
+ "grad_norm": 0.46508079767227173,
+ "learning_rate": 0.0006,
+ "loss": 9.647209167480469,
+ "step": 2908
+ },
+ {
+ "epoch": 40.40524017467249,
+ "grad_norm": 1.0228968858718872,
+ "learning_rate": 0.0006,
+ "loss": 11.685434341430664,
+ "step": 2909
+ },
+ {
+ "epoch": 40.419213973799124,
+ "grad_norm": 0.20553778111934662,
+ "learning_rate": 0.0006,
+ "loss": 8.621708869934082,
+ "step": 2910
+ },
+ {
+ "epoch": 40.43318777292576,
+ "grad_norm": 0.18466424942016602,
+ "learning_rate": 0.0006,
+ "loss": 8.590787887573242,
+ "step": 2911
+ },
+ {
+ "epoch": 40.4471615720524,
+ "grad_norm": 0.10458594560623169,
+ "learning_rate": 0.0006,
+ "loss": 8.324458122253418,
+ "step": 2912
+ },
+ {
+ "epoch": 40.46113537117904,
+ "grad_norm": 0.1096842959523201,
+ "learning_rate": 0.0006,
+ "loss": 8.300851821899414,
+ "step": 2913
+ },
+ {
+ "epoch": 40.47510917030568,
+ "grad_norm": 0.10501226782798767,
+ "learning_rate": 0.0006,
+ "loss": 8.21009635925293,
+ "step": 2914
+ },
+ {
+ "epoch": 40.48908296943232,
+ "grad_norm": 0.13704165816307068,
+ "learning_rate": 0.0006,
+ "loss": 8.141355514526367,
+ "step": 2915
+ },
+ {
+ "epoch": 40.50305676855895,
+ "grad_norm": 0.12129108607769012,
+ "learning_rate": 0.0006,
+ "loss": 8.074612617492676,
+ "step": 2916
+ },
+ {
+ "epoch": 40.51703056768559,
+ "grad_norm": 0.10028897225856781,
+ "learning_rate": 0.0006,
+ "loss": 8.00384521484375,
+ "step": 2917
+ },
+ {
+ "epoch": 40.531004366812226,
+ "grad_norm": 0.11764246970415115,
+ "learning_rate": 0.0006,
+ "loss": 8.078598022460938,
+ "step": 2918
+ },
+ {
+ "epoch": 40.544978165938865,
+ "grad_norm": 0.10081959515810013,
+ "learning_rate": 0.0006,
+ "loss": 8.008938789367676,
+ "step": 2919
+ },
+ {
+ "epoch": 40.5589519650655,
+ "grad_norm": 0.1003170907497406,
+ "learning_rate": 0.0006,
+ "loss": 7.962940216064453,
+ "step": 2920
+ },
+ {
+ "epoch": 40.57292576419214,
+ "grad_norm": 0.08763492107391357,
+ "learning_rate": 0.0006,
+ "loss": 7.909419536590576,
+ "step": 2921
+ },
+ {
+ "epoch": 40.58689956331878,
+ "grad_norm": 0.0636807233095169,
+ "learning_rate": 0.0006,
+ "loss": 7.878905773162842,
+ "step": 2922
+ },
+ {
+ "epoch": 40.60087336244541,
+ "grad_norm": 0.06005936488509178,
+ "learning_rate": 0.0006,
+ "loss": 7.850980758666992,
+ "step": 2923
+ },
+ {
+ "epoch": 40.61484716157205,
+ "grad_norm": 0.06415296345949173,
+ "learning_rate": 0.0006,
+ "loss": 7.830353260040283,
+ "step": 2924
+ },
+ {
+ "epoch": 40.62882096069869,
+ "grad_norm": 0.0732007846236229,
+ "learning_rate": 0.0006,
+ "loss": 7.830746650695801,
+ "step": 2925
+ },
+ {
+ "epoch": 40.64279475982533,
+ "grad_norm": 0.06628110259771347,
+ "learning_rate": 0.0006,
+ "loss": 7.82661771774292,
+ "step": 2926
+ },
+ {
+ "epoch": 40.65676855895197,
+ "grad_norm": 0.05544218048453331,
+ "learning_rate": 0.0006,
+ "loss": 7.814305782318115,
+ "step": 2927
+ },
+ {
+ "epoch": 40.670742358078606,
+ "grad_norm": 0.050327103585004807,
+ "learning_rate": 0.0006,
+ "loss": 7.764055252075195,
+ "step": 2928
+ },
+ {
+ "epoch": 40.68471615720524,
+ "grad_norm": 0.04918425530195236,
+ "learning_rate": 0.0006,
+ "loss": 7.744625091552734,
+ "step": 2929
+ },
+ {
+ "epoch": 40.698689956331876,
+ "grad_norm": 0.042862266302108765,
+ "learning_rate": 0.0006,
+ "loss": 7.764426231384277,
+ "step": 2930
+ },
+ {
+ "epoch": 40.712663755458514,
+ "grad_norm": 0.04338089004158974,
+ "learning_rate": 0.0006,
+ "loss": 7.712530136108398,
+ "step": 2931
+ },
+ {
+ "epoch": 40.72663755458515,
+ "grad_norm": 0.04127737879753113,
+ "learning_rate": 0.0006,
+ "loss": 7.742180347442627,
+ "step": 2932
+ },
+ {
+ "epoch": 40.74061135371179,
+ "grad_norm": 0.03358631581068039,
+ "learning_rate": 0.0006,
+ "loss": 7.714834213256836,
+ "step": 2933
+ },
+ {
+ "epoch": 40.75458515283843,
+ "grad_norm": 0.02845790982246399,
+ "learning_rate": 0.0006,
+ "loss": 7.7021026611328125,
+ "step": 2934
+ },
+ {
+ "epoch": 40.76855895196506,
+ "grad_norm": 0.03012419492006302,
+ "learning_rate": 0.0006,
+ "loss": 7.712377548217773,
+ "step": 2935
+ },
+ {
+ "epoch": 40.7825327510917,
+ "grad_norm": 0.032981179654598236,
+ "learning_rate": 0.0006,
+ "loss": 7.666385650634766,
+ "step": 2936
+ },
+ {
+ "epoch": 40.79650655021834,
+ "grad_norm": 0.03357074037194252,
+ "learning_rate": 0.0006,
+ "loss": 7.649097919464111,
+ "step": 2937
+ },
+ {
+ "epoch": 40.81048034934498,
+ "grad_norm": 0.030914708971977234,
+ "learning_rate": 0.0006,
+ "loss": 7.649196624755859,
+ "step": 2938
+ },
+ {
+ "epoch": 40.82445414847162,
+ "grad_norm": 0.026296131312847137,
+ "learning_rate": 0.0006,
+ "loss": 7.629964828491211,
+ "step": 2939
+ },
+ {
+ "epoch": 40.838427947598255,
+ "grad_norm": 0.026677245274186134,
+ "learning_rate": 0.0006,
+ "loss": 7.599973678588867,
+ "step": 2940
+ },
+ {
+ "epoch": 40.852401746724894,
+ "grad_norm": 0.023208219558000565,
+ "learning_rate": 0.0006,
+ "loss": 7.601624488830566,
+ "step": 2941
+ },
+ {
+ "epoch": 40.866375545851525,
+ "grad_norm": 0.020741475746035576,
+ "learning_rate": 0.0006,
+ "loss": 7.603902339935303,
+ "step": 2942
+ },
+ {
+ "epoch": 40.880349344978164,
+ "grad_norm": 0.019221751019358635,
+ "learning_rate": 0.0006,
+ "loss": 7.5535888671875,
+ "step": 2943
+ },
+ {
+ "epoch": 40.8943231441048,
+ "grad_norm": 0.019491951912641525,
+ "learning_rate": 0.0006,
+ "loss": 7.54597282409668,
+ "step": 2944
+ },
+ {
+ "epoch": 40.90829694323144,
+ "grad_norm": 0.019702572375535965,
+ "learning_rate": 0.0006,
+ "loss": 7.537567138671875,
+ "step": 2945
+ },
+ {
+ "epoch": 40.92227074235808,
+ "grad_norm": 0.01768800988793373,
+ "learning_rate": 0.0006,
+ "loss": 7.512800216674805,
+ "step": 2946
+ },
+ {
+ "epoch": 40.93624454148472,
+ "grad_norm": 0.01926138624548912,
+ "learning_rate": 0.0006,
+ "loss": 7.509932518005371,
+ "step": 2947
+ },
+ {
+ "epoch": 40.95021834061135,
+ "grad_norm": 0.019030006602406502,
+ "learning_rate": 0.0006,
+ "loss": 7.497983932495117,
+ "step": 2948
+ },
+ {
+ "epoch": 40.96419213973799,
+ "grad_norm": 0.017313841730356216,
+ "learning_rate": 0.0006,
+ "loss": 7.463518142700195,
+ "step": 2949
+ },
+ {
+ "epoch": 40.97816593886463,
+ "grad_norm": 0.01310762483626604,
+ "learning_rate": 0.0006,
+ "loss": 7.474668502807617,
+ "step": 2950
+ },
+ {
+ "epoch": 40.992139737991266,
+ "grad_norm": 0.013008282519876957,
+ "learning_rate": 0.0006,
+ "loss": 7.45982551574707,
+ "step": 2951
+ },
+ {
+ "epoch": 41.0,
+ "grad_norm": 0.01441363524645567,
+ "learning_rate": 0.0006,
+ "loss": 7.446529388427734,
+ "step": 2952
+ },
+ {
+ "epoch": 41.0,
+ "eval_loss": 7.486120223999023,
+ "eval_runtime": 70.4986,
+ "eval_samples_per_second": 34.639,
+ "eval_steps_per_second": 1.092,
+ "step": 2952
+ },
+ {
+ "epoch": 41.01397379912664,
+ "grad_norm": 0.014061855152249336,
+ "learning_rate": 0.0006,
+ "loss": 7.4692182540893555,
+ "step": 2953
+ },
+ {
+ "epoch": 41.02794759825328,
+ "grad_norm": 0.01587500050663948,
+ "learning_rate": 0.0006,
+ "loss": 7.4182024002075195,
+ "step": 2954
+ },
+ {
+ "epoch": 41.041921397379916,
+ "grad_norm": 0.014783984050154686,
+ "learning_rate": 0.0006,
+ "loss": 7.466778755187988,
+ "step": 2955
+ },
+ {
+ "epoch": 41.05589519650655,
+ "grad_norm": 0.011560854502022266,
+ "learning_rate": 0.0006,
+ "loss": 7.438002586364746,
+ "step": 2956
+ },
+ {
+ "epoch": 41.069868995633186,
+ "grad_norm": 0.00970957800745964,
+ "learning_rate": 0.0006,
+ "loss": 7.442694187164307,
+ "step": 2957
+ },
+ {
+ "epoch": 41.083842794759825,
+ "grad_norm": 0.009633459150791168,
+ "learning_rate": 0.0006,
+ "loss": 7.458884239196777,
+ "step": 2958
+ },
+ {
+ "epoch": 41.09781659388646,
+ "grad_norm": 0.010664350353181362,
+ "learning_rate": 0.0006,
+ "loss": 7.48378324508667,
+ "step": 2959
+ },
+ {
+ "epoch": 41.1117903930131,
+ "grad_norm": 0.011440591886639595,
+ "learning_rate": 0.0006,
+ "loss": 7.474484920501709,
+ "step": 2960
+ },
+ {
+ "epoch": 41.12576419213974,
+ "grad_norm": 0.01150908600538969,
+ "learning_rate": 0.0006,
+ "loss": 7.417902946472168,
+ "step": 2961
+ },
+ {
+ "epoch": 41.13973799126637,
+ "grad_norm": 0.010375697165727615,
+ "learning_rate": 0.0006,
+ "loss": 7.403279781341553,
+ "step": 2962
+ },
+ {
+ "epoch": 41.15371179039301,
+ "grad_norm": 0.009075703099370003,
+ "learning_rate": 0.0006,
+ "loss": 7.447347164154053,
+ "step": 2963
+ },
+ {
+ "epoch": 41.16768558951965,
+ "grad_norm": 0.008892177604138851,
+ "learning_rate": 0.0006,
+ "loss": 7.402681350708008,
+ "step": 2964
+ },
+ {
+ "epoch": 41.18165938864629,
+ "grad_norm": 0.009492084383964539,
+ "learning_rate": 0.0006,
+ "loss": 7.3871355056762695,
+ "step": 2965
+ },
+ {
+ "epoch": 41.19563318777293,
+ "grad_norm": 0.009139460511505604,
+ "learning_rate": 0.0006,
+ "loss": 7.419617176055908,
+ "step": 2966
+ },
+ {
+ "epoch": 41.209606986899566,
+ "grad_norm": 0.007959138602018356,
+ "learning_rate": 0.0006,
+ "loss": 7.374949932098389,
+ "step": 2967
+ },
+ {
+ "epoch": 41.223580786026204,
+ "grad_norm": 0.007571548223495483,
+ "learning_rate": 0.0006,
+ "loss": 7.4205241203308105,
+ "step": 2968
+ },
+ {
+ "epoch": 41.237554585152836,
+ "grad_norm": 0.007885400205850601,
+ "learning_rate": 0.0006,
+ "loss": 7.416543006896973,
+ "step": 2969
+ },
+ {
+ "epoch": 41.251528384279474,
+ "grad_norm": 0.008650158531963825,
+ "learning_rate": 0.0006,
+ "loss": 7.357326984405518,
+ "step": 2970
+ },
+ {
+ "epoch": 41.26550218340611,
+ "grad_norm": 0.007483262103050947,
+ "learning_rate": 0.0006,
+ "loss": 7.410891056060791,
+ "step": 2971
+ },
+ {
+ "epoch": 41.27947598253275,
+ "grad_norm": 0.00869277399033308,
+ "learning_rate": 0.0006,
+ "loss": 7.392127990722656,
+ "step": 2972
+ },
+ {
+ "epoch": 41.29344978165939,
+ "grad_norm": 0.00885078776627779,
+ "learning_rate": 0.0006,
+ "loss": 7.405449867248535,
+ "step": 2973
+ },
+ {
+ "epoch": 41.30742358078603,
+ "grad_norm": 0.00958226341754198,
+ "learning_rate": 0.0006,
+ "loss": 7.399868488311768,
+ "step": 2974
+ },
+ {
+ "epoch": 41.32139737991266,
+ "grad_norm": 0.006991599686443806,
+ "learning_rate": 0.0006,
+ "loss": 7.35369873046875,
+ "step": 2975
+ },
+ {
+ "epoch": 41.3353711790393,
+ "grad_norm": 0.007415142375975847,
+ "learning_rate": 0.0006,
+ "loss": 7.358353137969971,
+ "step": 2976
+ },
+ {
+ "epoch": 41.34934497816594,
+ "grad_norm": 0.007024371065199375,
+ "learning_rate": 0.0006,
+ "loss": 7.4070892333984375,
+ "step": 2977
+ },
+ {
+ "epoch": 41.36331877729258,
+ "grad_norm": 0.007234332151710987,
+ "learning_rate": 0.0006,
+ "loss": 7.379120349884033,
+ "step": 2978
+ },
+ {
+ "epoch": 41.377292576419215,
+ "grad_norm": 0.00798892229795456,
+ "learning_rate": 0.0006,
+ "loss": 7.365455627441406,
+ "step": 2979
+ },
+ {
+ "epoch": 41.391266375545854,
+ "grad_norm": 0.007303398102521896,
+ "learning_rate": 0.0006,
+ "loss": 7.341798782348633,
+ "step": 2980
+ },
+ {
+ "epoch": 41.40524017467249,
+ "grad_norm": 0.007119669578969479,
+ "learning_rate": 0.0006,
+ "loss": 7.363280296325684,
+ "step": 2981
+ },
+ {
+ "epoch": 41.419213973799124,
+ "grad_norm": 0.007457619532942772,
+ "learning_rate": 0.0006,
+ "loss": 7.39388370513916,
+ "step": 2982
+ },
+ {
+ "epoch": 41.43318777292576,
+ "grad_norm": 0.007905688136816025,
+ "learning_rate": 0.0006,
+ "loss": 7.380652904510498,
+ "step": 2983
+ },
+ {
+ "epoch": 41.4471615720524,
+ "grad_norm": 0.006212855689227581,
+ "learning_rate": 0.0006,
+ "loss": 7.398769378662109,
+ "step": 2984
+ },
+ {
+ "epoch": 41.46113537117904,
+ "grad_norm": 0.0070641543716192245,
+ "learning_rate": 0.0006,
+ "loss": 7.3503546714782715,
+ "step": 2985
+ },
+ {
+ "epoch": 41.47510917030568,
+ "grad_norm": 0.00787805300205946,
+ "learning_rate": 0.0006,
+ "loss": 7.347678184509277,
+ "step": 2986
+ },
+ {
+ "epoch": 41.48908296943232,
+ "grad_norm": 0.006760426796972752,
+ "learning_rate": 0.0006,
+ "loss": 7.380006790161133,
+ "step": 2987
+ },
+ {
+ "epoch": 41.50305676855895,
+ "grad_norm": 0.00772385997697711,
+ "learning_rate": 0.0006,
+ "loss": 7.363680362701416,
+ "step": 2988
+ },
+ {
+ "epoch": 41.51703056768559,
+ "grad_norm": 0.007373564876616001,
+ "learning_rate": 0.0006,
+ "loss": 7.333014488220215,
+ "step": 2989
+ },
+ {
+ "epoch": 41.531004366812226,
+ "grad_norm": 0.006717653013765812,
+ "learning_rate": 0.0006,
+ "loss": 7.360076427459717,
+ "step": 2990
+ },
+ {
+ "epoch": 41.544978165938865,
+ "grad_norm": 0.007191413082182407,
+ "learning_rate": 0.0006,
+ "loss": 7.397030830383301,
+ "step": 2991
+ },
+ {
+ "epoch": 41.5589519650655,
+ "grad_norm": 0.007831960916519165,
+ "learning_rate": 0.0006,
+ "loss": 7.369959831237793,
+ "step": 2992
+ },
+ {
+ "epoch": 41.57292576419214,
+ "grad_norm": 0.006947481073439121,
+ "learning_rate": 0.0006,
+ "loss": 7.333254337310791,
+ "step": 2993
+ },
+ {
+ "epoch": 41.58689956331878,
+ "grad_norm": 0.006243606563657522,
+ "learning_rate": 0.0006,
+ "loss": 7.327378273010254,
+ "step": 2994
+ },
+ {
+ "epoch": 41.60087336244541,
+ "grad_norm": 0.006298339460045099,
+ "learning_rate": 0.0006,
+ "loss": 7.323482036590576,
+ "step": 2995
+ },
+ {
+ "epoch": 41.61484716157205,
+ "grad_norm": 0.007587412837892771,
+ "learning_rate": 0.0006,
+ "loss": 7.313789367675781,
+ "step": 2996
+ },
+ {
+ "epoch": 41.62882096069869,
+ "grad_norm": 0.007140746805816889,
+ "learning_rate": 0.0006,
+ "loss": 7.356067657470703,
+ "step": 2997
+ },
+ {
+ "epoch": 41.64279475982533,
+ "grad_norm": 0.008315625600516796,
+ "learning_rate": 0.0006,
+ "loss": 7.3339643478393555,
+ "step": 2998
+ },
+ {
+ "epoch": 41.65676855895197,
+ "grad_norm": 0.006547102238982916,
+ "learning_rate": 0.0006,
+ "loss": 7.342191696166992,
+ "step": 2999
+ },
+ {
+ "epoch": 41.670742358078606,
+ "grad_norm": 0.006068122107535601,
+ "learning_rate": 0.0006,
+ "loss": 7.321015357971191,
+ "step": 3000
+ },
+ {
+ "epoch": 41.68471615720524,
+ "grad_norm": 0.007316991221159697,
+ "learning_rate": 0.0006,
+ "loss": 7.346490859985352,
+ "step": 3001
+ },
+ {
+ "epoch": 41.698689956331876,
+ "grad_norm": 0.0067388564348220825,
+ "learning_rate": 0.0006,
+ "loss": 7.340723991394043,
+ "step": 3002
+ },
+ {
+ "epoch": 41.712663755458514,
+ "grad_norm": 0.006839963141828775,
+ "learning_rate": 0.0006,
+ "loss": 7.329365253448486,
+ "step": 3003
+ },
+ {
+ "epoch": 41.72663755458515,
+ "grad_norm": 0.00805041566491127,
+ "learning_rate": 0.0006,
+ "loss": 7.298934459686279,
+ "step": 3004
+ },
+ {
+ "epoch": 41.74061135371179,
+ "grad_norm": 0.007197503466159105,
+ "learning_rate": 0.0006,
+ "loss": 7.331496238708496,
+ "step": 3005
+ },
+ {
+ "epoch": 41.75458515283843,
+ "grad_norm": 0.007220805156975985,
+ "learning_rate": 0.0006,
+ "loss": 7.317695140838623,
+ "step": 3006
+ },
+ {
+ "epoch": 41.76855895196506,
+ "grad_norm": 0.00874276738613844,
+ "learning_rate": 0.0006,
+ "loss": 7.272614479064941,
+ "step": 3007
+ },
+ {
+ "epoch": 41.7825327510917,
+ "grad_norm": 0.008386565372347832,
+ "learning_rate": 0.0006,
+ "loss": 7.295244216918945,
+ "step": 3008
+ },
+ {
+ "epoch": 41.79650655021834,
+ "grad_norm": 0.0077730282209813595,
+ "learning_rate": 0.0006,
+ "loss": 7.3395490646362305,
+ "step": 3009
+ },
+ {
+ "epoch": 41.81048034934498,
+ "grad_norm": 0.008441480807960033,
+ "learning_rate": 0.0006,
+ "loss": 7.319921493530273,
+ "step": 3010
+ },
+ {
+ "epoch": 41.82445414847162,
+ "grad_norm": 0.007398269604891539,
+ "learning_rate": 0.0006,
+ "loss": 7.3029465675354,
+ "step": 3011
+ },
+ {
+ "epoch": 41.838427947598255,
+ "grad_norm": 0.006878677289932966,
+ "learning_rate": 0.0006,
+ "loss": 7.318467140197754,
+ "step": 3012
+ },
+ {
+ "epoch": 41.852401746724894,
+ "grad_norm": 0.007616598624736071,
+ "learning_rate": 0.0006,
+ "loss": 7.3112077713012695,
+ "step": 3013
+ },
+ {
+ "epoch": 41.866375545851525,
+ "grad_norm": 0.011629699729382992,
+ "learning_rate": 0.0006,
+ "loss": 7.32523250579834,
+ "step": 3014
+ },
+ {
+ "epoch": 41.880349344978164,
+ "grad_norm": 0.011011249385774136,
+ "learning_rate": 0.0006,
+ "loss": 7.287247657775879,
+ "step": 3015
+ },
+ {
+ "epoch": 41.8943231441048,
+ "grad_norm": 0.010208295658230782,
+ "learning_rate": 0.0006,
+ "loss": 7.292117118835449,
+ "step": 3016
+ },
+ {
+ "epoch": 41.90829694323144,
+ "grad_norm": 0.008651095442473888,
+ "learning_rate": 0.0006,
+ "loss": 7.302201271057129,
+ "step": 3017
+ },
+ {
+ "epoch": 41.92227074235808,
+ "grad_norm": 0.006692920345813036,
+ "learning_rate": 0.0006,
+ "loss": 7.319568634033203,
+ "step": 3018
+ },
+ {
+ "epoch": 41.93624454148472,
+ "grad_norm": 0.010033348575234413,
+ "learning_rate": 0.0006,
+ "loss": 7.2738566398620605,
+ "step": 3019
+ },
+ {
+ "epoch": 41.95021834061135,
+ "grad_norm": 0.011790832504630089,
+ "learning_rate": 0.0006,
+ "loss": 7.298540115356445,
+ "step": 3020
+ },
+ {
+ "epoch": 41.96419213973799,
+ "grad_norm": 0.018622662872076035,
+ "learning_rate": 0.0006,
+ "loss": 7.267182350158691,
+ "step": 3021
+ },
+ {
+ "epoch": 41.97816593886463,
+ "grad_norm": 0.02520257607102394,
+ "learning_rate": 0.0006,
+ "loss": 7.261566162109375,
+ "step": 3022
+ },
+ {
+ "epoch": 41.992139737991266,
+ "grad_norm": 0.036332644522190094,
+ "learning_rate": 0.0006,
+ "loss": 7.270719051361084,
+ "step": 3023
+ },
+ {
+ "epoch": 42.0,
+ "grad_norm": 0.04717305675148964,
+ "learning_rate": 0.0006,
+ "loss": 7.293332099914551,
+ "step": 3024
+ },
+ {
+ "epoch": 42.0,
+ "eval_loss": 7.354280471801758,
+ "eval_runtime": 56.4353,
+ "eval_samples_per_second": 43.271,
+ "eval_steps_per_second": 1.364,
+ "step": 3024
+ },
+ {
+ "epoch": 42.01397379912664,
+ "grad_norm": 0.06661626696586609,
+ "learning_rate": 0.0006,
+ "loss": 7.338189601898193,
+ "step": 3025
+ },
+ {
+ "epoch": 42.02794759825328,
+ "grad_norm": 0.05452054366469383,
+ "learning_rate": 0.0006,
+ "loss": 7.316764831542969,
+ "step": 3026
+ },
+ {
+ "epoch": 42.041921397379916,
+ "grad_norm": 0.025638775900006294,
+ "learning_rate": 0.0006,
+ "loss": 7.34204626083374,
+ "step": 3027
+ },
+ {
+ "epoch": 42.05589519650655,
+ "grad_norm": 0.022398799657821655,
+ "learning_rate": 0.0006,
+ "loss": 7.262149333953857,
+ "step": 3028
+ },
+ {
+ "epoch": 42.069868995633186,
+ "grad_norm": 0.03713499382138252,
+ "learning_rate": 0.0006,
+ "loss": 7.27695369720459,
+ "step": 3029
+ },
+ {
+ "epoch": 42.083842794759825,
+ "grad_norm": 0.014949247241020203,
+ "learning_rate": 0.0006,
+ "loss": 7.296508312225342,
+ "step": 3030
+ },
+ {
+ "epoch": 42.09781659388646,
+ "grad_norm": 0.02703995071351528,
+ "learning_rate": 0.0006,
+ "loss": 7.258018493652344,
+ "step": 3031
+ },
+ {
+ "epoch": 42.1117903930131,
+ "grad_norm": 0.020352816209197044,
+ "learning_rate": 0.0006,
+ "loss": 7.241450309753418,
+ "step": 3032
+ },
+ {
+ "epoch": 42.12576419213974,
+ "grad_norm": 0.014578479342162609,
+ "learning_rate": 0.0006,
+ "loss": 7.257323265075684,
+ "step": 3033
+ },
+ {
+ "epoch": 42.13973799126637,
+ "grad_norm": 0.026053044945001602,
+ "learning_rate": 0.0006,
+ "loss": 7.254680633544922,
+ "step": 3034
+ },
+ {
+ "epoch": 42.15371179039301,
+ "grad_norm": 0.015120275318622589,
+ "learning_rate": 0.0006,
+ "loss": 7.261074066162109,
+ "step": 3035
+ },
+ {
+ "epoch": 42.16768558951965,
+ "grad_norm": 0.016822177916765213,
+ "learning_rate": 0.0006,
+ "loss": 7.279064178466797,
+ "step": 3036
+ },
+ {
+ "epoch": 42.18165938864629,
+ "grad_norm": 0.016367588192224503,
+ "learning_rate": 0.0006,
+ "loss": 7.251429557800293,
+ "step": 3037
+ },
+ {
+ "epoch": 42.19563318777293,
+ "grad_norm": 0.014350418001413345,
+ "learning_rate": 0.0006,
+ "loss": 7.259264945983887,
+ "step": 3038
+ },
+ {
+ "epoch": 42.209606986899566,
+ "grad_norm": 0.017078228294849396,
+ "learning_rate": 0.0006,
+ "loss": 7.227105140686035,
+ "step": 3039
+ },
+ {
+ "epoch": 42.223580786026204,
+ "grad_norm": 0.011759317480027676,
+ "learning_rate": 0.0006,
+ "loss": 7.275888442993164,
+ "step": 3040
+ },
+ {
+ "epoch": 42.237554585152836,
+ "grad_norm": 0.013019340112805367,
+ "learning_rate": 0.0006,
+ "loss": 7.2517008781433105,
+ "step": 3041
+ },
+ {
+ "epoch": 42.251528384279474,
+ "grad_norm": 0.01178441196680069,
+ "learning_rate": 0.0006,
+ "loss": 7.23576545715332,
+ "step": 3042
+ },
+ {
+ "epoch": 42.26550218340611,
+ "grad_norm": 0.014518891461193562,
+ "learning_rate": 0.0006,
+ "loss": 7.253895282745361,
+ "step": 3043
+ },
+ {
+ "epoch": 42.27947598253275,
+ "grad_norm": 0.018159950152039528,
+ "learning_rate": 0.0006,
+ "loss": 7.278058052062988,
+ "step": 3044
+ },
+ {
+ "epoch": 42.29344978165939,
+ "grad_norm": 0.007587796077132225,
+ "learning_rate": 0.0006,
+ "loss": 7.278578758239746,
+ "step": 3045
+ },
+ {
+ "epoch": 42.30742358078603,
+ "grad_norm": 0.015994979068636894,
+ "learning_rate": 0.0006,
+ "loss": 7.239841461181641,
+ "step": 3046
+ },
+ {
+ "epoch": 42.32139737991266,
+ "grad_norm": 0.017830168828368187,
+ "learning_rate": 0.0006,
+ "loss": 7.2443976402282715,
+ "step": 3047
+ },
+ {
+ "epoch": 42.3353711790393,
+ "grad_norm": 0.008738759905099869,
+ "learning_rate": 0.0006,
+ "loss": 7.232264518737793,
+ "step": 3048
+ },
+ {
+ "epoch": 42.34934497816594,
+ "grad_norm": 0.0175449438393116,
+ "learning_rate": 0.0006,
+ "loss": 7.228000640869141,
+ "step": 3049
+ },
+ {
+ "epoch": 42.36331877729258,
+ "grad_norm": 0.019611820578575134,
+ "learning_rate": 0.0006,
+ "loss": 7.234999179840088,
+ "step": 3050
+ },
+ {
+ "epoch": 42.377292576419215,
+ "grad_norm": 0.015302863903343678,
+ "learning_rate": 0.0006,
+ "loss": 7.242547988891602,
+ "step": 3051
+ },
+ {
+ "epoch": 42.391266375545854,
+ "grad_norm": 0.025919731706380844,
+ "learning_rate": 0.0006,
+ "loss": 7.216652870178223,
+ "step": 3052
+ },
+ {
+ "epoch": 42.40524017467249,
+ "grad_norm": 0.034587468951940536,
+ "learning_rate": 0.0006,
+ "loss": 7.245336532592773,
+ "step": 3053
+ },
+ {
+ "epoch": 42.419213973799124,
+ "grad_norm": 0.019198840484023094,
+ "learning_rate": 0.0006,
+ "loss": 7.256758689880371,
+ "step": 3054
+ },
+ {
+ "epoch": 42.43318777292576,
+ "grad_norm": 0.012316389009356499,
+ "learning_rate": 0.0006,
+ "loss": 7.209102630615234,
+ "step": 3055
+ },
+ {
+ "epoch": 42.4471615720524,
+ "grad_norm": 0.023512162268161774,
+ "learning_rate": 0.0006,
+ "loss": 7.250742435455322,
+ "step": 3056
+ },
+ {
+ "epoch": 42.46113537117904,
+ "grad_norm": 0.025457851588726044,
+ "learning_rate": 0.0006,
+ "loss": 7.263757705688477,
+ "step": 3057
+ },
+ {
+ "epoch": 42.47510917030568,
+ "grad_norm": 0.02192617766559124,
+ "learning_rate": 0.0006,
+ "loss": 7.246977806091309,
+ "step": 3058
+ },
+ {
+ "epoch": 42.48908296943232,
+ "grad_norm": 0.02541281282901764,
+ "learning_rate": 0.0006,
+ "loss": 7.232361316680908,
+ "step": 3059
+ },
+ {
+ "epoch": 42.50305676855895,
+ "grad_norm": 0.036353763192892075,
+ "learning_rate": 0.0006,
+ "loss": 7.242556571960449,
+ "step": 3060
+ },
+ {
+ "epoch": 42.51703056768559,
+ "grad_norm": 0.04555872827768326,
+ "learning_rate": 0.0006,
+ "loss": 7.228940963745117,
+ "step": 3061
+ },
+ {
+ "epoch": 42.531004366812226,
+ "grad_norm": 0.04563378915190697,
+ "learning_rate": 0.0006,
+ "loss": 7.235276222229004,
+ "step": 3062
+ },
+ {
+ "epoch": 42.544978165938865,
+ "grad_norm": 0.04275760427117348,
+ "learning_rate": 0.0006,
+ "loss": 7.238895416259766,
+ "step": 3063
+ },
+ {
+ "epoch": 42.5589519650655,
+ "grad_norm": 0.04535885155200958,
+ "learning_rate": 0.0006,
+ "loss": 7.205336570739746,
+ "step": 3064
+ },
+ {
+ "epoch": 42.57292576419214,
+ "grad_norm": 0.015534617006778717,
+ "learning_rate": 0.0006,
+ "loss": 7.238393783569336,
+ "step": 3065
+ },
+ {
+ "epoch": 42.58689956331878,
+ "grad_norm": 0.03286299481987953,
+ "learning_rate": 0.0006,
+ "loss": 7.216453552246094,
+ "step": 3066
+ },
+ {
+ "epoch": 42.60087336244541,
+ "grad_norm": 0.025453168898820877,
+ "learning_rate": 0.0006,
+ "loss": 7.203750133514404,
+ "step": 3067
+ },
+ {
+ "epoch": 42.61484716157205,
+ "grad_norm": 0.021501649171113968,
+ "learning_rate": 0.0006,
+ "loss": 7.241621971130371,
+ "step": 3068
+ },
+ {
+ "epoch": 42.62882096069869,
+ "grad_norm": 0.02743634767830372,
+ "learning_rate": 0.0006,
+ "loss": 7.22076940536499,
+ "step": 3069
+ },
+ {
+ "epoch": 42.64279475982533,
+ "grad_norm": 0.020890483632683754,
+ "learning_rate": 0.0006,
+ "loss": 7.218412399291992,
+ "step": 3070
+ },
+ {
+ "epoch": 42.65676855895197,
+ "grad_norm": 0.022499850019812584,
+ "learning_rate": 0.0006,
+ "loss": 7.217696189880371,
+ "step": 3071
+ },
+ {
+ "epoch": 42.670742358078606,
+ "grad_norm": 0.01653696969151497,
+ "learning_rate": 0.0006,
+ "loss": 7.156417369842529,
+ "step": 3072
+ },
+ {
+ "epoch": 42.68471615720524,
+ "grad_norm": 0.01834225095808506,
+ "learning_rate": 0.0006,
+ "loss": 7.211247444152832,
+ "step": 3073
+ },
+ {
+ "epoch": 42.698689956331876,
+ "grad_norm": 0.020137229934334755,
+ "learning_rate": 0.0006,
+ "loss": 7.232499599456787,
+ "step": 3074
+ },
+ {
+ "epoch": 42.712663755458514,
+ "grad_norm": 0.012671459466218948,
+ "learning_rate": 0.0006,
+ "loss": 7.172441482543945,
+ "step": 3075
+ },
+ {
+ "epoch": 42.72663755458515,
+ "grad_norm": 0.016803130507469177,
+ "learning_rate": 0.0006,
+ "loss": 7.152103424072266,
+ "step": 3076
+ },
+ {
+ "epoch": 42.74061135371179,
+ "grad_norm": 0.01623712107539177,
+ "learning_rate": 0.0006,
+ "loss": 7.187032699584961,
+ "step": 3077
+ },
+ {
+ "epoch": 42.75458515283843,
+ "grad_norm": 0.015938356518745422,
+ "learning_rate": 0.0006,
+ "loss": 7.171140670776367,
+ "step": 3078
+ },
+ {
+ "epoch": 42.76855895196506,
+ "grad_norm": 0.02580736204981804,
+ "learning_rate": 0.0006,
+ "loss": 7.204472541809082,
+ "step": 3079
+ },
+ {
+ "epoch": 42.7825327510917,
+ "grad_norm": 0.023372117429971695,
+ "learning_rate": 0.0006,
+ "loss": 7.165554523468018,
+ "step": 3080
+ },
+ {
+ "epoch": 42.79650655021834,
+ "grad_norm": 0.04440586268901825,
+ "learning_rate": 0.0006,
+ "loss": 7.169431209564209,
+ "step": 3081
+ },
+ {
+ "epoch": 42.81048034934498,
+ "grad_norm": 0.06797406077384949,
+ "learning_rate": 0.0006,
+ "loss": 7.168193340301514,
+ "step": 3082
+ },
+ {
+ "epoch": 42.82445414847162,
+ "grad_norm": 0.0914321169257164,
+ "learning_rate": 0.0006,
+ "loss": 7.204671859741211,
+ "step": 3083
+ },
+ {
+ "epoch": 42.838427947598255,
+ "grad_norm": 0.0506940595805645,
+ "learning_rate": 0.0006,
+ "loss": 7.15949821472168,
+ "step": 3084
+ },
+ {
+ "epoch": 42.852401746724894,
+ "grad_norm": 0.04819890484213829,
+ "learning_rate": 0.0006,
+ "loss": 7.192837238311768,
+ "step": 3085
+ },
+ {
+ "epoch": 42.866375545851525,
+ "grad_norm": 0.040044233202934265,
+ "learning_rate": 0.0006,
+ "loss": 7.1600165367126465,
+ "step": 3086
+ },
+ {
+ "epoch": 42.880349344978164,
+ "grad_norm": 0.02573299966752529,
+ "learning_rate": 0.0006,
+ "loss": 7.153427600860596,
+ "step": 3087
+ },
+ {
+ "epoch": 42.8943231441048,
+ "grad_norm": 0.046942535787820816,
+ "learning_rate": 0.0006,
+ "loss": 7.16767692565918,
+ "step": 3088
+ },
+ {
+ "epoch": 42.90829694323144,
+ "grad_norm": 0.02683161571621895,
+ "learning_rate": 0.0006,
+ "loss": 7.17144250869751,
+ "step": 3089
+ },
+ {
+ "epoch": 42.92227074235808,
+ "grad_norm": 0.05363674834370613,
+ "learning_rate": 0.0006,
+ "loss": 7.120128631591797,
+ "step": 3090
+ },
+ {
+ "epoch": 42.93624454148472,
+ "grad_norm": 0.023756247013807297,
+ "learning_rate": 0.0006,
+ "loss": 7.169703006744385,
+ "step": 3091
+ },
+ {
+ "epoch": 42.95021834061135,
+ "grad_norm": 0.04216783866286278,
+ "learning_rate": 0.0006,
+ "loss": 7.1206865310668945,
+ "step": 3092
+ },
+ {
+ "epoch": 42.96419213973799,
+ "grad_norm": 0.02219565585255623,
+ "learning_rate": 0.0006,
+ "loss": 7.157215118408203,
+ "step": 3093
+ },
+ {
+ "epoch": 42.97816593886463,
+ "grad_norm": 0.03527127206325531,
+ "learning_rate": 0.0006,
+ "loss": 7.166672706604004,
+ "step": 3094
+ },
+ {
+ "epoch": 42.992139737991266,
+ "grad_norm": 0.017748817801475525,
+ "learning_rate": 0.0006,
+ "loss": 7.154747009277344,
+ "step": 3095
+ },
+ {
+ "epoch": 43.0,
+ "grad_norm": 0.033165838569402695,
+ "learning_rate": 0.0006,
+ "loss": 7.150866508483887,
+ "step": 3096
+ },
+ {
+ "epoch": 43.0,
+ "eval_loss": 7.144403457641602,
+ "eval_runtime": 56.8853,
+ "eval_samples_per_second": 42.928,
+ "eval_steps_per_second": 1.354,
+ "step": 3096
+ },
+ {
+ "epoch": 43.01397379912664,
+ "grad_norm": 0.015199829824268818,
+ "learning_rate": 0.0006,
+ "loss": 7.122735023498535,
+ "step": 3097
+ },
+ {
+ "epoch": 43.02794759825328,
+ "grad_norm": 0.03220256045460701,
+ "learning_rate": 0.0006,
+ "loss": 7.143553733825684,
+ "step": 3098
+ },
+ {
+ "epoch": 43.041921397379916,
+ "grad_norm": 0.014687920920550823,
+ "learning_rate": 0.0006,
+ "loss": 7.080960750579834,
+ "step": 3099
+ },
+ {
+ "epoch": 43.05589519650655,
+ "grad_norm": 0.023522932082414627,
+ "learning_rate": 0.0006,
+ "loss": 7.137356758117676,
+ "step": 3100
+ },
+ {
+ "epoch": 43.069868995633186,
+ "grad_norm": 0.014808216132223606,
+ "learning_rate": 0.0006,
+ "loss": 7.09143590927124,
+ "step": 3101
+ },
+ {
+ "epoch": 43.083842794759825,
+ "grad_norm": 0.029339246451854706,
+ "learning_rate": 0.0006,
+ "loss": 7.122198104858398,
+ "step": 3102
+ },
+ {
+ "epoch": 43.09781659388646,
+ "grad_norm": 0.022953078150749207,
+ "learning_rate": 0.0006,
+ "loss": 7.092505931854248,
+ "step": 3103
+ },
+ {
+ "epoch": 43.1117903930131,
+ "grad_norm": 0.04590052738785744,
+ "learning_rate": 0.0006,
+ "loss": 7.113100051879883,
+ "step": 3104
+ },
+ {
+ "epoch": 43.12576419213974,
+ "grad_norm": 0.0524442233145237,
+ "learning_rate": 0.0006,
+ "loss": 7.147022247314453,
+ "step": 3105
+ },
+ {
+ "epoch": 43.13973799126637,
+ "grad_norm": 0.0315794013440609,
+ "learning_rate": 0.0006,
+ "loss": 7.094045162200928,
+ "step": 3106
+ },
+ {
+ "epoch": 43.15371179039301,
+ "grad_norm": 0.05112722888588905,
+ "learning_rate": 0.0006,
+ "loss": 7.054388523101807,
+ "step": 3107
+ },
+ {
+ "epoch": 43.16768558951965,
+ "grad_norm": 0.03146776929497719,
+ "learning_rate": 0.0006,
+ "loss": 7.116674423217773,
+ "step": 3108
+ },
+ {
+ "epoch": 43.18165938864629,
+ "grad_norm": 0.03391246497631073,
+ "learning_rate": 0.0006,
+ "loss": 7.150345802307129,
+ "step": 3109
+ },
+ {
+ "epoch": 43.19563318777293,
+ "grad_norm": 0.026957903057336807,
+ "learning_rate": 0.0006,
+ "loss": 7.058531284332275,
+ "step": 3110
+ },
+ {
+ "epoch": 43.209606986899566,
+ "grad_norm": 0.041532568633556366,
+ "learning_rate": 0.0006,
+ "loss": 7.085498809814453,
+ "step": 3111
+ },
+ {
+ "epoch": 43.223580786026204,
+ "grad_norm": 0.020001588389277458,
+ "learning_rate": 0.0006,
+ "loss": 7.131487846374512,
+ "step": 3112
+ },
+ {
+ "epoch": 43.237554585152836,
+ "grad_norm": 0.02268712781369686,
+ "learning_rate": 0.0006,
+ "loss": 7.0790863037109375,
+ "step": 3113
+ },
+ {
+ "epoch": 43.251528384279474,
+ "grad_norm": 0.023363925516605377,
+ "learning_rate": 0.0006,
+ "loss": 7.067816734313965,
+ "step": 3114
+ },
+ {
+ "epoch": 43.26550218340611,
+ "grad_norm": 0.019367508590221405,
+ "learning_rate": 0.0006,
+ "loss": 7.042041778564453,
+ "step": 3115
+ },
+ {
+ "epoch": 43.27947598253275,
+ "grad_norm": 0.02745090238749981,
+ "learning_rate": 0.0006,
+ "loss": 7.064733028411865,
+ "step": 3116
+ },
+ {
+ "epoch": 43.29344978165939,
+ "grad_norm": 0.02543802745640278,
+ "learning_rate": 0.0006,
+ "loss": 7.058948040008545,
+ "step": 3117
+ },
+ {
+ "epoch": 43.30742358078603,
+ "grad_norm": 0.018874092027544975,
+ "learning_rate": 0.0006,
+ "loss": 7.009946346282959,
+ "step": 3118
+ },
+ {
+ "epoch": 43.32139737991266,
+ "grad_norm": 0.024662822484970093,
+ "learning_rate": 0.0006,
+ "loss": 7.015913963317871,
+ "step": 3119
+ },
+ {
+ "epoch": 43.3353711790393,
+ "grad_norm": 0.02441416308283806,
+ "learning_rate": 0.0006,
+ "loss": 7.061622142791748,
+ "step": 3120
+ },
+ {
+ "epoch": 43.34934497816594,
+ "grad_norm": 0.031887322664260864,
+ "learning_rate": 0.0006,
+ "loss": 7.046883583068848,
+ "step": 3121
+ },
+ {
+ "epoch": 43.36331877729258,
+ "grad_norm": 0.021930739283561707,
+ "learning_rate": 0.0006,
+ "loss": 7.0309624671936035,
+ "step": 3122
+ },
+ {
+ "epoch": 43.377292576419215,
+ "grad_norm": 0.02722783200442791,
+ "learning_rate": 0.0006,
+ "loss": 7.072924613952637,
+ "step": 3123
+ },
+ {
+ "epoch": 43.391266375545854,
+ "grad_norm": 0.01483179721981287,
+ "learning_rate": 0.0006,
+ "loss": 7.028624534606934,
+ "step": 3124
+ },
+ {
+ "epoch": 43.40524017467249,
+ "grad_norm": 0.024793945252895355,
+ "learning_rate": 0.0006,
+ "loss": 7.001623153686523,
+ "step": 3125
+ },
+ {
+ "epoch": 43.419213973799124,
+ "grad_norm": 0.020148713141679764,
+ "learning_rate": 0.0006,
+ "loss": 6.9938812255859375,
+ "step": 3126
+ },
+ {
+ "epoch": 43.43318777292576,
+ "grad_norm": 0.02413850650191307,
+ "learning_rate": 0.0006,
+ "loss": 7.001243591308594,
+ "step": 3127
+ },
+ {
+ "epoch": 43.4471615720524,
+ "grad_norm": 0.028549203649163246,
+ "learning_rate": 0.0006,
+ "loss": 6.987671375274658,
+ "step": 3128
+ },
+ {
+ "epoch": 43.46113537117904,
+ "grad_norm": 0.025988785549998283,
+ "learning_rate": 0.0006,
+ "loss": 6.9551167488098145,
+ "step": 3129
+ },
+ {
+ "epoch": 43.47510917030568,
+ "grad_norm": 0.027140729129314423,
+ "learning_rate": 0.0006,
+ "loss": 6.9547810554504395,
+ "step": 3130
+ },
+ {
+ "epoch": 43.48908296943232,
+ "grad_norm": 0.030130082741379738,
+ "learning_rate": 0.0006,
+ "loss": 6.993186950683594,
+ "step": 3131
+ },
+ {
+ "epoch": 43.50305676855895,
+ "grad_norm": 0.025996994227170944,
+ "learning_rate": 0.0006,
+ "loss": 6.984624862670898,
+ "step": 3132
+ },
+ {
+ "epoch": 43.51703056768559,
+ "grad_norm": 0.02423008903861046,
+ "learning_rate": 0.0006,
+ "loss": 6.98813009262085,
+ "step": 3133
+ },
+ {
+ "epoch": 43.531004366812226,
+ "grad_norm": 0.029665837064385414,
+ "learning_rate": 0.0006,
+ "loss": 6.986610412597656,
+ "step": 3134
+ },
+ {
+ "epoch": 43.544978165938865,
+ "grad_norm": 0.028768595308065414,
+ "learning_rate": 0.0006,
+ "loss": 6.898797035217285,
+ "step": 3135
+ },
+ {
+ "epoch": 43.5589519650655,
+ "grad_norm": 0.03111208975315094,
+ "learning_rate": 0.0006,
+ "loss": 6.923741817474365,
+ "step": 3136
+ },
+ {
+ "epoch": 43.57292576419214,
+ "grad_norm": 0.0368187390267849,
+ "learning_rate": 0.0006,
+ "loss": 6.994343280792236,
+ "step": 3137
+ },
+ {
+ "epoch": 43.58689956331878,
+ "grad_norm": 0.04076763987541199,
+ "learning_rate": 0.0006,
+ "loss": 6.909458160400391,
+ "step": 3138
+ },
+ {
+ "epoch": 43.60087336244541,
+ "grad_norm": 0.03281061723828316,
+ "learning_rate": 0.0006,
+ "loss": 6.924851417541504,
+ "step": 3139
+ },
+ {
+ "epoch": 43.61484716157205,
+ "grad_norm": 0.025486988946795464,
+ "learning_rate": 0.0006,
+ "loss": 6.962926864624023,
+ "step": 3140
+ },
+ {
+ "epoch": 43.62882096069869,
+ "grad_norm": 0.04583485797047615,
+ "learning_rate": 0.0006,
+ "loss": 6.876555442810059,
+ "step": 3141
+ },
+ {
+ "epoch": 43.64279475982533,
+ "grad_norm": 0.05460785701870918,
+ "learning_rate": 0.0006,
+ "loss": 6.952932357788086,
+ "step": 3142
+ },
+ {
+ "epoch": 43.65676855895197,
+ "grad_norm": 0.040917906910181046,
+ "learning_rate": 0.0006,
+ "loss": 6.9624924659729,
+ "step": 3143
+ },
+ {
+ "epoch": 43.670742358078606,
+ "grad_norm": 0.04009668901562691,
+ "learning_rate": 0.0006,
+ "loss": 6.94033145904541,
+ "step": 3144
+ },
+ {
+ "epoch": 43.68471615720524,
+ "grad_norm": 0.04500625655055046,
+ "learning_rate": 0.0006,
+ "loss": 6.961839199066162,
+ "step": 3145
+ },
+ {
+ "epoch": 43.698689956331876,
+ "grad_norm": 0.03182258829474449,
+ "learning_rate": 0.0006,
+ "loss": 6.930580139160156,
+ "step": 3146
+ },
+ {
+ "epoch": 43.712663755458514,
+ "grad_norm": 0.053621504455804825,
+ "learning_rate": 0.0006,
+ "loss": 6.950235366821289,
+ "step": 3147
+ },
+ {
+ "epoch": 43.72663755458515,
+ "grad_norm": 0.031919289380311966,
+ "learning_rate": 0.0006,
+ "loss": 6.895111560821533,
+ "step": 3148
+ },
+ {
+ "epoch": 43.74061135371179,
+ "grad_norm": 0.042713891714811325,
+ "learning_rate": 0.0006,
+ "loss": 6.940885543823242,
+ "step": 3149
+ },
+ {
+ "epoch": 43.75458515283843,
+ "grad_norm": 0.02385653182864189,
+ "learning_rate": 0.0006,
+ "loss": 6.948801517486572,
+ "step": 3150
+ },
+ {
+ "epoch": 43.76855895196506,
+ "grad_norm": 0.04100663587450981,
+ "learning_rate": 0.0006,
+ "loss": 6.970411777496338,
+ "step": 3151
+ },
+ {
+ "epoch": 43.7825327510917,
+ "grad_norm": 0.02729583904147148,
+ "learning_rate": 0.0006,
+ "loss": 6.899379253387451,
+ "step": 3152
+ },
+ {
+ "epoch": 43.79650655021834,
+ "grad_norm": 0.03128620237112045,
+ "learning_rate": 0.0006,
+ "loss": 6.931319236755371,
+ "step": 3153
+ },
+ {
+ "epoch": 43.81048034934498,
+ "grad_norm": 0.025488659739494324,
+ "learning_rate": 0.0006,
+ "loss": 6.8486409187316895,
+ "step": 3154
+ },
+ {
+ "epoch": 43.82445414847162,
+ "grad_norm": 0.03679376840591431,
+ "learning_rate": 0.0006,
+ "loss": 6.840811252593994,
+ "step": 3155
+ },
+ {
+ "epoch": 43.838427947598255,
+ "grad_norm": 0.020578034222126007,
+ "learning_rate": 0.0006,
+ "loss": 6.9088616371154785,
+ "step": 3156
+ },
+ {
+ "epoch": 43.852401746724894,
+ "grad_norm": 0.02935841865837574,
+ "learning_rate": 0.0006,
+ "loss": 6.806676864624023,
+ "step": 3157
+ },
+ {
+ "epoch": 43.866375545851525,
+ "grad_norm": 0.023319244384765625,
+ "learning_rate": 0.0006,
+ "loss": 6.878457546234131,
+ "step": 3158
+ },
+ {
+ "epoch": 43.880349344978164,
+ "grad_norm": 0.02779589593410492,
+ "learning_rate": 0.0006,
+ "loss": 6.822447776794434,
+ "step": 3159
+ },
+ {
+ "epoch": 43.8943231441048,
+ "grad_norm": 0.01879825070500374,
+ "learning_rate": 0.0006,
+ "loss": 6.854121208190918,
+ "step": 3160
+ },
+ {
+ "epoch": 43.90829694323144,
+ "grad_norm": 0.023286139592528343,
+ "learning_rate": 0.0006,
+ "loss": 6.836827278137207,
+ "step": 3161
+ },
+ {
+ "epoch": 43.92227074235808,
+ "grad_norm": 0.030358297750353813,
+ "learning_rate": 0.0006,
+ "loss": 6.860037803649902,
+ "step": 3162
+ },
+ {
+ "epoch": 43.93624454148472,
+ "grad_norm": 0.029621105641126633,
+ "learning_rate": 0.0006,
+ "loss": 6.884025573730469,
+ "step": 3163
+ },
+ {
+ "epoch": 43.95021834061135,
+ "grad_norm": 0.029829824343323708,
+ "learning_rate": 0.0006,
+ "loss": 6.8702392578125,
+ "step": 3164
+ },
+ {
+ "epoch": 43.96419213973799,
+ "grad_norm": 0.036892473697662354,
+ "learning_rate": 0.0006,
+ "loss": 6.806652069091797,
+ "step": 3165
+ },
+ {
+ "epoch": 43.97816593886463,
+ "grad_norm": 0.03206709027290344,
+ "learning_rate": 0.0006,
+ "loss": 6.828495979309082,
+ "step": 3166
+ },
+ {
+ "epoch": 43.992139737991266,
+ "grad_norm": 0.02744000405073166,
+ "learning_rate": 0.0006,
+ "loss": 6.832002639770508,
+ "step": 3167
+ },
+ {
+ "epoch": 44.0,
+ "grad_norm": 0.02205660194158554,
+ "learning_rate": 0.0006,
+ "loss": 6.898828506469727,
+ "step": 3168
+ },
+ {
+ "epoch": 44.0,
+ "eval_loss": 6.8150553703308105,
+ "eval_runtime": 56.784,
+ "eval_samples_per_second": 43.005,
+ "eval_steps_per_second": 1.356,
+ "step": 3168
+ },
+ {
+ "epoch": 44.01397379912664,
+ "grad_norm": 0.030421260744333267,
+ "learning_rate": 0.0006,
+ "loss": 6.775057792663574,
+ "step": 3169
+ },
+ {
+ "epoch": 44.02794759825328,
+ "grad_norm": 0.0465611033141613,
+ "learning_rate": 0.0006,
+ "loss": 6.849390506744385,
+ "step": 3170
+ },
+ {
+ "epoch": 44.041921397379916,
+ "grad_norm": 0.05977972596883774,
+ "learning_rate": 0.0006,
+ "loss": 6.725576400756836,
+ "step": 3171
+ },
+ {
+ "epoch": 44.05589519650655,
+ "grad_norm": 0.05432697385549545,
+ "learning_rate": 0.0006,
+ "loss": 6.768024444580078,
+ "step": 3172
+ },
+ {
+ "epoch": 44.069868995633186,
+ "grad_norm": 0.04850967228412628,
+ "learning_rate": 0.0006,
+ "loss": 6.799457550048828,
+ "step": 3173
+ },
+ {
+ "epoch": 44.083842794759825,
+ "grad_norm": 0.06201786920428276,
+ "learning_rate": 0.0006,
+ "loss": 6.897234916687012,
+ "step": 3174
+ },
+ {
+ "epoch": 44.09781659388646,
+ "grad_norm": 0.04695591703057289,
+ "learning_rate": 0.0006,
+ "loss": 6.826470375061035,
+ "step": 3175
+ },
+ {
+ "epoch": 44.1117903930131,
+ "grad_norm": 0.03850164636969566,
+ "learning_rate": 0.0006,
+ "loss": 6.835901260375977,
+ "step": 3176
+ },
+ {
+ "epoch": 44.12576419213974,
+ "grad_norm": 0.04250155761837959,
+ "learning_rate": 0.0006,
+ "loss": 6.782310485839844,
+ "step": 3177
+ },
+ {
+ "epoch": 44.13973799126637,
+ "grad_norm": 0.041465479880571365,
+ "learning_rate": 0.0006,
+ "loss": 6.785732746124268,
+ "step": 3178
+ },
+ {
+ "epoch": 44.15371179039301,
+ "grad_norm": 0.02527807280421257,
+ "learning_rate": 0.0006,
+ "loss": 6.78061580657959,
+ "step": 3179
+ },
+ {
+ "epoch": 44.16768558951965,
+ "grad_norm": 0.03660374879837036,
+ "learning_rate": 0.0006,
+ "loss": 6.783992767333984,
+ "step": 3180
+ },
+ {
+ "epoch": 44.18165938864629,
+ "grad_norm": 0.027567964047193527,
+ "learning_rate": 0.0006,
+ "loss": 6.765413284301758,
+ "step": 3181
+ },
+ {
+ "epoch": 44.19563318777293,
+ "grad_norm": 0.024301424622535706,
+ "learning_rate": 0.0006,
+ "loss": 6.760368347167969,
+ "step": 3182
+ },
+ {
+ "epoch": 44.209606986899566,
+ "grad_norm": 0.0266821701079607,
+ "learning_rate": 0.0006,
+ "loss": 6.71417236328125,
+ "step": 3183
+ },
+ {
+ "epoch": 44.223580786026204,
+ "grad_norm": 0.028125090524554253,
+ "learning_rate": 0.0006,
+ "loss": 6.646126747131348,
+ "step": 3184
+ },
+ {
+ "epoch": 44.237554585152836,
+ "grad_norm": 0.0185737032443285,
+ "learning_rate": 0.0006,
+ "loss": 6.712944030761719,
+ "step": 3185
+ },
+ {
+ "epoch": 44.251528384279474,
+ "grad_norm": 0.025836674496531487,
+ "learning_rate": 0.0006,
+ "loss": 6.707030296325684,
+ "step": 3186
+ },
+ {
+ "epoch": 44.26550218340611,
+ "grad_norm": 0.029744870960712433,
+ "learning_rate": 0.0006,
+ "loss": 6.6271071434021,
+ "step": 3187
+ },
+ {
+ "epoch": 44.27947598253275,
+ "grad_norm": 0.03021426685154438,
+ "learning_rate": 0.0006,
+ "loss": 6.677826404571533,
+ "step": 3188
+ },
+ {
+ "epoch": 44.29344978165939,
+ "grad_norm": 0.032727014273405075,
+ "learning_rate": 0.0006,
+ "loss": 6.657764434814453,
+ "step": 3189
+ },
+ {
+ "epoch": 44.30742358078603,
+ "grad_norm": 0.03696778789162636,
+ "learning_rate": 0.0006,
+ "loss": 6.632883548736572,
+ "step": 3190
+ },
+ {
+ "epoch": 44.32139737991266,
+ "grad_norm": 0.034630827605724335,
+ "learning_rate": 0.0006,
+ "loss": 6.759583473205566,
+ "step": 3191
+ },
+ {
+ "epoch": 44.3353711790393,
+ "grad_norm": 0.030517758801579475,
+ "learning_rate": 0.0006,
+ "loss": 6.702268600463867,
+ "step": 3192
+ },
+ {
+ "epoch": 44.34934497816594,
+ "grad_norm": 0.04105190932750702,
+ "learning_rate": 0.0006,
+ "loss": 6.705792427062988,
+ "step": 3193
+ },
+ {
+ "epoch": 44.36331877729258,
+ "grad_norm": 0.06659174710512161,
+ "learning_rate": 0.0006,
+ "loss": 6.73862361907959,
+ "step": 3194
+ },
+ {
+ "epoch": 44.377292576419215,
+ "grad_norm": 0.08013619482517242,
+ "learning_rate": 0.0006,
+ "loss": 6.710282802581787,
+ "step": 3195
+ },
+ {
+ "epoch": 44.391266375545854,
+ "grad_norm": 0.049850963056087494,
+ "learning_rate": 0.0006,
+ "loss": 6.684974670410156,
+ "step": 3196
+ },
+ {
+ "epoch": 44.40524017467249,
+ "grad_norm": 0.061247408390045166,
+ "learning_rate": 0.0006,
+ "loss": 6.716034412384033,
+ "step": 3197
+ },
+ {
+ "epoch": 44.419213973799124,
+ "grad_norm": 0.05868857726454735,
+ "learning_rate": 0.0006,
+ "loss": 6.589107513427734,
+ "step": 3198
+ },
+ {
+ "epoch": 44.43318777292576,
+ "grad_norm": 0.048878561705350876,
+ "learning_rate": 0.0006,
+ "loss": 6.694855690002441,
+ "step": 3199
+ },
+ {
+ "epoch": 44.4471615720524,
+ "grad_norm": 0.028748324140906334,
+ "learning_rate": 0.0006,
+ "loss": 6.629969596862793,
+ "step": 3200
+ },
+ {
+ "epoch": 44.46113537117904,
+ "grad_norm": 0.04060306400060654,
+ "learning_rate": 0.0006,
+ "loss": 6.721431255340576,
+ "step": 3201
+ },
+ {
+ "epoch": 44.47510917030568,
+ "grad_norm": 0.03833824768662453,
+ "learning_rate": 0.0006,
+ "loss": 6.625029563903809,
+ "step": 3202
+ },
+ {
+ "epoch": 44.48908296943232,
+ "grad_norm": 0.03791997581720352,
+ "learning_rate": 0.0006,
+ "loss": 6.669187545776367,
+ "step": 3203
+ },
+ {
+ "epoch": 44.50305676855895,
+ "grad_norm": 0.03259684890508652,
+ "learning_rate": 0.0006,
+ "loss": 6.665643692016602,
+ "step": 3204
+ },
+ {
+ "epoch": 44.51703056768559,
+ "grad_norm": 0.040658626705408096,
+ "learning_rate": 0.0006,
+ "loss": 6.639796257019043,
+ "step": 3205
+ },
+ {
+ "epoch": 44.531004366812226,
+ "grad_norm": 0.03691193461418152,
+ "learning_rate": 0.0006,
+ "loss": 6.692244529724121,
+ "step": 3206
+ },
+ {
+ "epoch": 44.544978165938865,
+ "grad_norm": 0.02814812771975994,
+ "learning_rate": 0.0006,
+ "loss": 6.61978006362915,
+ "step": 3207
+ },
+ {
+ "epoch": 44.5589519650655,
+ "grad_norm": 0.022155074402689934,
+ "learning_rate": 0.0006,
+ "loss": 6.625905990600586,
+ "step": 3208
+ },
+ {
+ "epoch": 44.57292576419214,
+ "grad_norm": 0.030794885009527206,
+ "learning_rate": 0.0006,
+ "loss": 6.678500652313232,
+ "step": 3209
+ },
+ {
+ "epoch": 44.58689956331878,
+ "grad_norm": 0.03201768547296524,
+ "learning_rate": 0.0006,
+ "loss": 6.6015729904174805,
+ "step": 3210
+ },
+ {
+ "epoch": 44.60087336244541,
+ "grad_norm": 0.025138361379504204,
+ "learning_rate": 0.0006,
+ "loss": 6.568234443664551,
+ "step": 3211
+ },
+ {
+ "epoch": 44.61484716157205,
+ "grad_norm": 0.02969353087246418,
+ "learning_rate": 0.0006,
+ "loss": 6.588366508483887,
+ "step": 3212
+ },
+ {
+ "epoch": 44.62882096069869,
+ "grad_norm": 0.03136962279677391,
+ "learning_rate": 0.0006,
+ "loss": 6.479961395263672,
+ "step": 3213
+ },
+ {
+ "epoch": 44.64279475982533,
+ "grad_norm": 0.030178170651197433,
+ "learning_rate": 0.0006,
+ "loss": 6.622910976409912,
+ "step": 3214
+ },
+ {
+ "epoch": 44.65676855895197,
+ "grad_norm": 0.02682245895266533,
+ "learning_rate": 0.0006,
+ "loss": 6.597707748413086,
+ "step": 3215
+ },
+ {
+ "epoch": 44.670742358078606,
+ "grad_norm": 0.025154557079076767,
+ "learning_rate": 0.0006,
+ "loss": 6.630541801452637,
+ "step": 3216
+ },
+ {
+ "epoch": 44.68471615720524,
+ "grad_norm": 0.03564463183283806,
+ "learning_rate": 0.0006,
+ "loss": 6.542823791503906,
+ "step": 3217
+ },
+ {
+ "epoch": 44.698689956331876,
+ "grad_norm": 0.049477651715278625,
+ "learning_rate": 0.0006,
+ "loss": 6.646030426025391,
+ "step": 3218
+ },
+ {
+ "epoch": 44.712663755458514,
+ "grad_norm": 0.06536535918712616,
+ "learning_rate": 0.0006,
+ "loss": 6.595463275909424,
+ "step": 3219
+ },
+ {
+ "epoch": 44.72663755458515,
+ "grad_norm": 0.07710316777229309,
+ "learning_rate": 0.0006,
+ "loss": 6.55524206161499,
+ "step": 3220
+ },
+ {
+ "epoch": 44.74061135371179,
+ "grad_norm": 0.09571163356304169,
+ "learning_rate": 0.0006,
+ "loss": 6.659764289855957,
+ "step": 3221
+ },
+ {
+ "epoch": 44.75458515283843,
+ "grad_norm": 0.055076830089092255,
+ "learning_rate": 0.0006,
+ "loss": 6.658882141113281,
+ "step": 3222
+ },
+ {
+ "epoch": 44.76855895196506,
+ "grad_norm": 0.07671099156141281,
+ "learning_rate": 0.0006,
+ "loss": 6.605809211730957,
+ "step": 3223
+ },
+ {
+ "epoch": 44.7825327510917,
+ "grad_norm": 0.05295584350824356,
+ "learning_rate": 0.0006,
+ "loss": 6.641455173492432,
+ "step": 3224
+ },
+ {
+ "epoch": 44.79650655021834,
+ "grad_norm": 0.09128042310476303,
+ "learning_rate": 0.0006,
+ "loss": 6.691066741943359,
+ "step": 3225
+ },
+ {
+ "epoch": 44.81048034934498,
+ "grad_norm": 0.05362958461046219,
+ "learning_rate": 0.0006,
+ "loss": 6.573406219482422,
+ "step": 3226
+ },
+ {
+ "epoch": 44.82445414847162,
+ "grad_norm": 0.046910062432289124,
+ "learning_rate": 0.0006,
+ "loss": 6.550876617431641,
+ "step": 3227
+ },
+ {
+ "epoch": 44.838427947598255,
+ "grad_norm": 0.03494197130203247,
+ "learning_rate": 0.0006,
+ "loss": 6.540179252624512,
+ "step": 3228
+ },
+ {
+ "epoch": 44.852401746724894,
+ "grad_norm": 0.04716387391090393,
+ "learning_rate": 0.0006,
+ "loss": 6.490459442138672,
+ "step": 3229
+ },
+ {
+ "epoch": 44.866375545851525,
+ "grad_norm": 0.04190506786108017,
+ "learning_rate": 0.0006,
+ "loss": 6.509925365447998,
+ "step": 3230
+ },
+ {
+ "epoch": 44.880349344978164,
+ "grad_norm": 0.04749064892530441,
+ "learning_rate": 0.0006,
+ "loss": 6.478779315948486,
+ "step": 3231
+ },
+ {
+ "epoch": 44.8943231441048,
+ "grad_norm": 0.028002185747027397,
+ "learning_rate": 0.0006,
+ "loss": 6.528904914855957,
+ "step": 3232
+ },
+ {
+ "epoch": 44.90829694323144,
+ "grad_norm": 0.036473166197538376,
+ "learning_rate": 0.0006,
+ "loss": 6.549042701721191,
+ "step": 3233
+ },
+ {
+ "epoch": 44.92227074235808,
+ "grad_norm": 0.02543315477669239,
+ "learning_rate": 0.0006,
+ "loss": 6.498875141143799,
+ "step": 3234
+ },
+ {
+ "epoch": 44.93624454148472,
+ "grad_norm": 0.029906807467341423,
+ "learning_rate": 0.0006,
+ "loss": 6.453019142150879,
+ "step": 3235
+ },
+ {
+ "epoch": 44.95021834061135,
+ "grad_norm": 0.027231857180595398,
+ "learning_rate": 0.0006,
+ "loss": 6.428829193115234,
+ "step": 3236
+ },
+ {
+ "epoch": 44.96419213973799,
+ "grad_norm": 0.02506154775619507,
+ "learning_rate": 0.0006,
+ "loss": 6.420858383178711,
+ "step": 3237
+ },
+ {
+ "epoch": 44.97816593886463,
+ "grad_norm": 0.0269472673535347,
+ "learning_rate": 0.0006,
+ "loss": 6.484989643096924,
+ "step": 3238
+ },
+ {
+ "epoch": 44.992139737991266,
+ "grad_norm": 0.023410487920045853,
+ "learning_rate": 0.0006,
+ "loss": 6.454830169677734,
+ "step": 3239
+ },
+ {
+ "epoch": 45.0,
+ "grad_norm": 0.034175023436546326,
+ "learning_rate": 0.0006,
+ "loss": 6.430975437164307,
+ "step": 3240
+ },
+ {
+ "epoch": 45.0,
+ "eval_loss": 6.458896160125732,
+ "eval_runtime": 56.6934,
+ "eval_samples_per_second": 43.074,
+ "eval_steps_per_second": 1.358,
+ "step": 3240
+ },
+ {
+ "epoch": 45.01397379912664,
+ "grad_norm": 0.045031629502773285,
+ "learning_rate": 0.0006,
+ "loss": 6.4243011474609375,
+ "step": 3241
+ },
+ {
+ "epoch": 45.02794759825328,
+ "grad_norm": 0.04968734085559845,
+ "learning_rate": 0.0006,
+ "loss": 6.380091667175293,
+ "step": 3242
+ },
+ {
+ "epoch": 45.041921397379916,
+ "grad_norm": 0.047145187854766846,
+ "learning_rate": 0.0006,
+ "loss": 6.477178573608398,
+ "step": 3243
+ },
+ {
+ "epoch": 45.05589519650655,
+ "grad_norm": 0.04304948449134827,
+ "learning_rate": 0.0006,
+ "loss": 6.399726867675781,
+ "step": 3244
+ },
+ {
+ "epoch": 45.069868995633186,
+ "grad_norm": 0.04466010630130768,
+ "learning_rate": 0.0006,
+ "loss": 6.3415679931640625,
+ "step": 3245
+ },
+ {
+ "epoch": 45.083842794759825,
+ "grad_norm": 0.04305306077003479,
+ "learning_rate": 0.0006,
+ "loss": 6.3413310050964355,
+ "step": 3246
+ },
+ {
+ "epoch": 45.09781659388646,
+ "grad_norm": 0.043064575642347336,
+ "learning_rate": 0.0006,
+ "loss": 6.4838714599609375,
+ "step": 3247
+ },
+ {
+ "epoch": 45.1117903930131,
+ "grad_norm": 0.049503277987241745,
+ "learning_rate": 0.0006,
+ "loss": 6.38473653793335,
+ "step": 3248
+ },
+ {
+ "epoch": 45.12576419213974,
+ "grad_norm": 0.04613238573074341,
+ "learning_rate": 0.0006,
+ "loss": 6.370769500732422,
+ "step": 3249
+ },
+ {
+ "epoch": 45.13973799126637,
+ "grad_norm": 0.052119478583335876,
+ "learning_rate": 0.0006,
+ "loss": 6.362321853637695,
+ "step": 3250
+ },
+ {
+ "epoch": 45.15371179039301,
+ "grad_norm": 0.045110076665878296,
+ "learning_rate": 0.0006,
+ "loss": 6.369844913482666,
+ "step": 3251
+ },
+ {
+ "epoch": 45.16768558951965,
+ "grad_norm": 0.03798583149909973,
+ "learning_rate": 0.0006,
+ "loss": 6.324633598327637,
+ "step": 3252
+ },
+ {
+ "epoch": 45.18165938864629,
+ "grad_norm": 0.03744659945368767,
+ "learning_rate": 0.0006,
+ "loss": 6.349421501159668,
+ "step": 3253
+ },
+ {
+ "epoch": 45.19563318777293,
+ "grad_norm": 0.03767292574048042,
+ "learning_rate": 0.0006,
+ "loss": 6.334238052368164,
+ "step": 3254
+ },
+ {
+ "epoch": 45.209606986899566,
+ "grad_norm": 0.03577509894967079,
+ "learning_rate": 0.0006,
+ "loss": 6.3453826904296875,
+ "step": 3255
+ },
+ {
+ "epoch": 45.223580786026204,
+ "grad_norm": 0.03403240442276001,
+ "learning_rate": 0.0006,
+ "loss": 6.2836174964904785,
+ "step": 3256
+ },
+ {
+ "epoch": 45.237554585152836,
+ "grad_norm": 0.03517331928014755,
+ "learning_rate": 0.0006,
+ "loss": 6.281718730926514,
+ "step": 3257
+ },
+ {
+ "epoch": 45.251528384279474,
+ "grad_norm": 0.04409139230847359,
+ "learning_rate": 0.0006,
+ "loss": 6.320815086364746,
+ "step": 3258
+ },
+ {
+ "epoch": 45.26550218340611,
+ "grad_norm": 0.054750364273786545,
+ "learning_rate": 0.0006,
+ "loss": 6.343618392944336,
+ "step": 3259
+ },
+ {
+ "epoch": 45.27947598253275,
+ "grad_norm": 0.07612067461013794,
+ "learning_rate": 0.0006,
+ "loss": 6.310614585876465,
+ "step": 3260
+ },
+ {
+ "epoch": 45.29344978165939,
+ "grad_norm": 0.08438985049724579,
+ "learning_rate": 0.0006,
+ "loss": 6.2369184494018555,
+ "step": 3261
+ },
+ {
+ "epoch": 45.30742358078603,
+ "grad_norm": 0.0782201811671257,
+ "learning_rate": 0.0006,
+ "loss": 6.254452705383301,
+ "step": 3262
+ },
+ {
+ "epoch": 45.32139737991266,
+ "grad_norm": 0.09928355365991592,
+ "learning_rate": 0.0006,
+ "loss": 6.3112077713012695,
+ "step": 3263
+ },
+ {
+ "epoch": 45.3353711790393,
+ "grad_norm": 0.07233133912086487,
+ "learning_rate": 0.0006,
+ "loss": 6.408087253570557,
+ "step": 3264
+ },
+ {
+ "epoch": 45.34934497816594,
+ "grad_norm": 0.07478684186935425,
+ "learning_rate": 0.0006,
+ "loss": 6.3770036697387695,
+ "step": 3265
+ },
+ {
+ "epoch": 45.36331877729258,
+ "grad_norm": 0.06036846712231636,
+ "learning_rate": 0.0006,
+ "loss": 6.33865213394165,
+ "step": 3266
+ },
+ {
+ "epoch": 45.377292576419215,
+ "grad_norm": 0.05783458054065704,
+ "learning_rate": 0.0006,
+ "loss": 6.363203048706055,
+ "step": 3267
+ },
+ {
+ "epoch": 45.391266375545854,
+ "grad_norm": 0.05443909019231796,
+ "learning_rate": 0.0006,
+ "loss": 6.287069797515869,
+ "step": 3268
+ },
+ {
+ "epoch": 45.40524017467249,
+ "grad_norm": 0.0557909794151783,
+ "learning_rate": 0.0006,
+ "loss": 6.297784805297852,
+ "step": 3269
+ },
+ {
+ "epoch": 45.419213973799124,
+ "grad_norm": 0.03483372926712036,
+ "learning_rate": 0.0006,
+ "loss": 6.280734062194824,
+ "step": 3270
+ },
+ {
+ "epoch": 45.43318777292576,
+ "grad_norm": 0.0425543375313282,
+ "learning_rate": 0.0006,
+ "loss": 6.33565616607666,
+ "step": 3271
+ },
+ {
+ "epoch": 45.4471615720524,
+ "grad_norm": 0.03230159357190132,
+ "learning_rate": 0.0006,
+ "loss": 6.197698593139648,
+ "step": 3272
+ },
+ {
+ "epoch": 45.46113537117904,
+ "grad_norm": 0.036897290498018265,
+ "learning_rate": 0.0006,
+ "loss": 6.241239547729492,
+ "step": 3273
+ },
+ {
+ "epoch": 45.47510917030568,
+ "grad_norm": 0.031103838235139847,
+ "learning_rate": 0.0006,
+ "loss": 6.159435749053955,
+ "step": 3274
+ },
+ {
+ "epoch": 45.48908296943232,
+ "grad_norm": 0.027116449549794197,
+ "learning_rate": 0.0006,
+ "loss": 6.212225914001465,
+ "step": 3275
+ },
+ {
+ "epoch": 45.50305676855895,
+ "grad_norm": 0.029923981055617332,
+ "learning_rate": 0.0006,
+ "loss": 6.157269477844238,
+ "step": 3276
+ },
+ {
+ "epoch": 45.51703056768559,
+ "grad_norm": 0.02521538734436035,
+ "learning_rate": 0.0006,
+ "loss": 6.146918773651123,
+ "step": 3277
+ },
+ {
+ "epoch": 45.531004366812226,
+ "grad_norm": 0.02758920192718506,
+ "learning_rate": 0.0006,
+ "loss": 6.0446977615356445,
+ "step": 3278
+ },
+ {
+ "epoch": 45.544978165938865,
+ "grad_norm": 0.024328220635652542,
+ "learning_rate": 0.0006,
+ "loss": 6.099508285522461,
+ "step": 3279
+ },
+ {
+ "epoch": 45.5589519650655,
+ "grad_norm": 0.027425771579146385,
+ "learning_rate": 0.0006,
+ "loss": 6.084299087524414,
+ "step": 3280
+ },
+ {
+ "epoch": 45.57292576419214,
+ "grad_norm": 0.026927651837468147,
+ "learning_rate": 0.0006,
+ "loss": 6.146334171295166,
+ "step": 3281
+ },
+ {
+ "epoch": 45.58689956331878,
+ "grad_norm": 0.04054379090666771,
+ "learning_rate": 0.0006,
+ "loss": 6.1301493644714355,
+ "step": 3282
+ },
+ {
+ "epoch": 45.60087336244541,
+ "grad_norm": 0.04456290975213051,
+ "learning_rate": 0.0006,
+ "loss": 6.152511119842529,
+ "step": 3283
+ },
+ {
+ "epoch": 45.61484716157205,
+ "grad_norm": 0.052821505814790726,
+ "learning_rate": 0.0006,
+ "loss": 6.113544940948486,
+ "step": 3284
+ },
+ {
+ "epoch": 45.62882096069869,
+ "grad_norm": 0.07251739501953125,
+ "learning_rate": 0.0006,
+ "loss": 6.118298530578613,
+ "step": 3285
+ },
+ {
+ "epoch": 45.64279475982533,
+ "grad_norm": 0.09433556348085403,
+ "learning_rate": 0.0006,
+ "loss": 6.170897483825684,
+ "step": 3286
+ },
+ {
+ "epoch": 45.65676855895197,
+ "grad_norm": 0.057683493942022324,
+ "learning_rate": 0.0006,
+ "loss": 6.1747822761535645,
+ "step": 3287
+ },
+ {
+ "epoch": 45.670742358078606,
+ "grad_norm": 0.06363270431756973,
+ "learning_rate": 0.0006,
+ "loss": 6.1617889404296875,
+ "step": 3288
+ },
+ {
+ "epoch": 45.68471615720524,
+ "grad_norm": 0.07528169453144073,
+ "learning_rate": 0.0006,
+ "loss": 6.230060577392578,
+ "step": 3289
+ },
+ {
+ "epoch": 45.698689956331876,
+ "grad_norm": 0.07816170156002045,
+ "learning_rate": 0.0006,
+ "loss": 6.171257019042969,
+ "step": 3290
+ },
+ {
+ "epoch": 45.712663755458514,
+ "grad_norm": 0.05860164389014244,
+ "learning_rate": 0.0006,
+ "loss": 6.095484256744385,
+ "step": 3291
+ },
+ {
+ "epoch": 45.72663755458515,
+ "grad_norm": 0.04747429117560387,
+ "learning_rate": 0.0006,
+ "loss": 6.060067176818848,
+ "step": 3292
+ },
+ {
+ "epoch": 45.74061135371179,
+ "grad_norm": 0.04514474794268608,
+ "learning_rate": 0.0006,
+ "loss": 6.100218772888184,
+ "step": 3293
+ },
+ {
+ "epoch": 45.75458515283843,
+ "grad_norm": 0.05168681964278221,
+ "learning_rate": 0.0006,
+ "loss": 6.0621867179870605,
+ "step": 3294
+ },
+ {
+ "epoch": 45.76855895196506,
+ "grad_norm": 0.04665876924991608,
+ "learning_rate": 0.0006,
+ "loss": 6.118292808532715,
+ "step": 3295
+ },
+ {
+ "epoch": 45.7825327510917,
+ "grad_norm": 0.049374762922525406,
+ "learning_rate": 0.0006,
+ "loss": 6.098394393920898,
+ "step": 3296
+ },
+ {
+ "epoch": 45.79650655021834,
+ "grad_norm": 0.03793549910187721,
+ "learning_rate": 0.0006,
+ "loss": 6.069541931152344,
+ "step": 3297
+ },
+ {
+ "epoch": 45.81048034934498,
+ "grad_norm": 0.036362383514642715,
+ "learning_rate": 0.0006,
+ "loss": 6.1188435554504395,
+ "step": 3298
+ },
+ {
+ "epoch": 45.82445414847162,
+ "grad_norm": 0.03397097811102867,
+ "learning_rate": 0.0006,
+ "loss": 6.117589950561523,
+ "step": 3299
+ },
+ {
+ "epoch": 45.838427947598255,
+ "grad_norm": 0.030977753922343254,
+ "learning_rate": 0.0006,
+ "loss": 6.0236711502075195,
+ "step": 3300
+ },
+ {
+ "epoch": 45.852401746724894,
+ "grad_norm": 0.02800217643380165,
+ "learning_rate": 0.0006,
+ "loss": 6.034825325012207,
+ "step": 3301
+ },
+ {
+ "epoch": 45.866375545851525,
+ "grad_norm": 0.03129778802394867,
+ "learning_rate": 0.0006,
+ "loss": 5.980271339416504,
+ "step": 3302
+ },
+ {
+ "epoch": 45.880349344978164,
+ "grad_norm": 0.025886094197630882,
+ "learning_rate": 0.0006,
+ "loss": 5.966419219970703,
+ "step": 3303
+ },
+ {
+ "epoch": 45.8943231441048,
+ "grad_norm": 0.025712328031659126,
+ "learning_rate": 0.0006,
+ "loss": 5.938652515411377,
+ "step": 3304
+ },
+ {
+ "epoch": 45.90829694323144,
+ "grad_norm": 0.022508298978209496,
+ "learning_rate": 0.0006,
+ "loss": 5.982312202453613,
+ "step": 3305
+ },
+ {
+ "epoch": 45.92227074235808,
+ "grad_norm": 0.024880951270461082,
+ "learning_rate": 0.0006,
+ "loss": 5.931150436401367,
+ "step": 3306
+ },
+ {
+ "epoch": 45.93624454148472,
+ "grad_norm": 0.030007576569914818,
+ "learning_rate": 0.0006,
+ "loss": 5.936639308929443,
+ "step": 3307
+ },
+ {
+ "epoch": 45.95021834061135,
+ "grad_norm": 0.033314041793346405,
+ "learning_rate": 0.0006,
+ "loss": 5.944248676300049,
+ "step": 3308
+ },
+ {
+ "epoch": 45.96419213973799,
+ "grad_norm": 0.043078381568193436,
+ "learning_rate": 0.0006,
+ "loss": 5.971286773681641,
+ "step": 3309
+ },
+ {
+ "epoch": 45.97816593886463,
+ "grad_norm": 0.059296440333127975,
+ "learning_rate": 0.0006,
+ "loss": 5.910497665405273,
+ "step": 3310
+ },
+ {
+ "epoch": 45.992139737991266,
+ "grad_norm": 0.06232666224241257,
+ "learning_rate": 0.0006,
+ "loss": 5.929817199707031,
+ "step": 3311
+ },
+ {
+ "epoch": 46.0,
+ "grad_norm": 0.05303940549492836,
+ "learning_rate": 0.0006,
+ "loss": 5.910822868347168,
+ "step": 3312
+ },
+ {
+ "epoch": 46.0,
+ "eval_loss": 5.977851390838623,
+ "eval_runtime": 56.2435,
+ "eval_samples_per_second": 43.418,
+ "eval_steps_per_second": 1.369,
+ "step": 3312
+ },
+ {
+ "epoch": 46.01397379912664,
+ "grad_norm": 0.06943798810243607,
+ "learning_rate": 0.0006,
+ "loss": 5.854922294616699,
+ "step": 3313
+ },
+ {
+ "epoch": 46.02794759825328,
+ "grad_norm": 0.05236278846859932,
+ "learning_rate": 0.0006,
+ "loss": 5.876611232757568,
+ "step": 3314
+ },
+ {
+ "epoch": 46.041921397379916,
+ "grad_norm": 0.06240001320838928,
+ "learning_rate": 0.0006,
+ "loss": 5.878912448883057,
+ "step": 3315
+ },
+ {
+ "epoch": 46.05589519650655,
+ "grad_norm": 0.07401192933320999,
+ "learning_rate": 0.0006,
+ "loss": 5.985136032104492,
+ "step": 3316
+ },
+ {
+ "epoch": 46.069868995633186,
+ "grad_norm": 0.0826529711484909,
+ "learning_rate": 0.0006,
+ "loss": 5.939439296722412,
+ "step": 3317
+ },
+ {
+ "epoch": 46.083842794759825,
+ "grad_norm": 0.07999005168676376,
+ "learning_rate": 0.0006,
+ "loss": 5.982336044311523,
+ "step": 3318
+ },
+ {
+ "epoch": 46.09781659388646,
+ "grad_norm": 0.07014869153499603,
+ "learning_rate": 0.0006,
+ "loss": 5.896534442901611,
+ "step": 3319
+ },
+ {
+ "epoch": 46.1117903930131,
+ "grad_norm": 0.05795128643512726,
+ "learning_rate": 0.0006,
+ "loss": 5.843111991882324,
+ "step": 3320
+ },
+ {
+ "epoch": 46.12576419213974,
+ "grad_norm": 0.06374258548021317,
+ "learning_rate": 0.0006,
+ "loss": 5.78752326965332,
+ "step": 3321
+ },
+ {
+ "epoch": 46.13973799126637,
+ "grad_norm": 0.045524805784225464,
+ "learning_rate": 0.0006,
+ "loss": 5.86755895614624,
+ "step": 3322
+ },
+ {
+ "epoch": 46.15371179039301,
+ "grad_norm": 0.03773504123091698,
+ "learning_rate": 0.0006,
+ "loss": 5.83627462387085,
+ "step": 3323
+ },
+ {
+ "epoch": 46.16768558951965,
+ "grad_norm": 0.035815201699733734,
+ "learning_rate": 0.0006,
+ "loss": 5.798238277435303,
+ "step": 3324
+ },
+ {
+ "epoch": 46.18165938864629,
+ "grad_norm": 0.03756783902645111,
+ "learning_rate": 0.0006,
+ "loss": 5.845197677612305,
+ "step": 3325
+ },
+ {
+ "epoch": 46.19563318777293,
+ "grad_norm": 0.03569113463163376,
+ "learning_rate": 0.0006,
+ "loss": 5.888932704925537,
+ "step": 3326
+ },
+ {
+ "epoch": 46.209606986899566,
+ "grad_norm": 0.029467813670635223,
+ "learning_rate": 0.0006,
+ "loss": 5.792047500610352,
+ "step": 3327
+ },
+ {
+ "epoch": 46.223580786026204,
+ "grad_norm": 0.025698406621813774,
+ "learning_rate": 0.0006,
+ "loss": 5.783734321594238,
+ "step": 3328
+ },
+ {
+ "epoch": 46.237554585152836,
+ "grad_norm": 0.028426671400666237,
+ "learning_rate": 0.0006,
+ "loss": 5.762084007263184,
+ "step": 3329
+ },
+ {
+ "epoch": 46.251528384279474,
+ "grad_norm": 0.03099622018635273,
+ "learning_rate": 0.0006,
+ "loss": 5.719109058380127,
+ "step": 3330
+ },
+ {
+ "epoch": 46.26550218340611,
+ "grad_norm": 0.0310219656676054,
+ "learning_rate": 0.0006,
+ "loss": 5.737057209014893,
+ "step": 3331
+ },
+ {
+ "epoch": 46.27947598253275,
+ "grad_norm": 0.02713446132838726,
+ "learning_rate": 0.0006,
+ "loss": 5.758657455444336,
+ "step": 3332
+ },
+ {
+ "epoch": 46.29344978165939,
+ "grad_norm": 0.03099844604730606,
+ "learning_rate": 0.0006,
+ "loss": 5.742086410522461,
+ "step": 3333
+ },
+ {
+ "epoch": 46.30742358078603,
+ "grad_norm": 0.030258730053901672,
+ "learning_rate": 0.0006,
+ "loss": 5.767426490783691,
+ "step": 3334
+ },
+ {
+ "epoch": 46.32139737991266,
+ "grad_norm": 0.029034852981567383,
+ "learning_rate": 0.0006,
+ "loss": 5.651792049407959,
+ "step": 3335
+ },
+ {
+ "epoch": 46.3353711790393,
+ "grad_norm": 0.03138665854930878,
+ "learning_rate": 0.0006,
+ "loss": 5.664980888366699,
+ "step": 3336
+ },
+ {
+ "epoch": 46.34934497816594,
+ "grad_norm": 0.040546901524066925,
+ "learning_rate": 0.0006,
+ "loss": 5.665165901184082,
+ "step": 3337
+ },
+ {
+ "epoch": 46.36331877729258,
+ "grad_norm": 0.06395485252141953,
+ "learning_rate": 0.0006,
+ "loss": 5.7776713371276855,
+ "step": 3338
+ },
+ {
+ "epoch": 46.377292576419215,
+ "grad_norm": 0.10088056325912476,
+ "learning_rate": 0.0006,
+ "loss": 5.755645751953125,
+ "step": 3339
+ },
+ {
+ "epoch": 46.391266375545854,
+ "grad_norm": 0.09527938067913055,
+ "learning_rate": 0.0006,
+ "loss": 5.838809490203857,
+ "step": 3340
+ },
+ {
+ "epoch": 46.40524017467249,
+ "grad_norm": 0.0849648267030716,
+ "learning_rate": 0.0006,
+ "loss": 5.781932830810547,
+ "step": 3341
+ },
+ {
+ "epoch": 46.419213973799124,
+ "grad_norm": 0.10142666101455688,
+ "learning_rate": 0.0006,
+ "loss": 5.838970184326172,
+ "step": 3342
+ },
+ {
+ "epoch": 46.43318777292576,
+ "grad_norm": 0.42731454968452454,
+ "learning_rate": 0.0006,
+ "loss": 6.071902275085449,
+ "step": 3343
+ },
+ {
+ "epoch": 46.4471615720524,
+ "grad_norm": 0.13102488219738007,
+ "learning_rate": 0.0006,
+ "loss": 6.8635993003845215,
+ "step": 3344
+ },
+ {
+ "epoch": 46.46113537117904,
+ "grad_norm": 0.23691938817501068,
+ "learning_rate": 0.0006,
+ "loss": 7.313158988952637,
+ "step": 3345
+ },
+ {
+ "epoch": 46.47510917030568,
+ "grad_norm": 0.11670669168233871,
+ "learning_rate": 0.0006,
+ "loss": 7.240363121032715,
+ "step": 3346
+ },
+ {
+ "epoch": 46.48908296943232,
+ "grad_norm": 0.09985820949077606,
+ "learning_rate": 0.0006,
+ "loss": 7.113332271575928,
+ "step": 3347
+ },
+ {
+ "epoch": 46.50305676855895,
+ "grad_norm": 0.11406408250331879,
+ "learning_rate": 0.0006,
+ "loss": 7.051285743713379,
+ "step": 3348
+ },
+ {
+ "epoch": 46.51703056768559,
+ "grad_norm": 0.07900336384773254,
+ "learning_rate": 0.0006,
+ "loss": 7.014365196228027,
+ "step": 3349
+ },
+ {
+ "epoch": 46.531004366812226,
+ "grad_norm": 0.05380333960056305,
+ "learning_rate": 0.0006,
+ "loss": 6.8868818283081055,
+ "step": 3350
+ },
+ {
+ "epoch": 46.544978165938865,
+ "grad_norm": 0.05868449807167053,
+ "learning_rate": 0.0006,
+ "loss": 6.783599853515625,
+ "step": 3351
+ },
+ {
+ "epoch": 46.5589519650655,
+ "grad_norm": 0.04428785294294357,
+ "learning_rate": 0.0006,
+ "loss": 6.798511505126953,
+ "step": 3352
+ },
+ {
+ "epoch": 46.57292576419214,
+ "grad_norm": 0.055527277290821075,
+ "learning_rate": 0.0006,
+ "loss": 6.694507598876953,
+ "step": 3353
+ },
+ {
+ "epoch": 46.58689956331878,
+ "grad_norm": 0.03872432932257652,
+ "learning_rate": 0.0006,
+ "loss": 6.736631393432617,
+ "step": 3354
+ },
+ {
+ "epoch": 46.60087336244541,
+ "grad_norm": 0.04330779239535332,
+ "learning_rate": 0.0006,
+ "loss": 6.576806545257568,
+ "step": 3355
+ },
+ {
+ "epoch": 46.61484716157205,
+ "grad_norm": 0.036767736077308655,
+ "learning_rate": 0.0006,
+ "loss": 6.559871196746826,
+ "step": 3356
+ },
+ {
+ "epoch": 46.62882096069869,
+ "grad_norm": 0.039857108145952225,
+ "learning_rate": 0.0006,
+ "loss": 6.5749831199646,
+ "step": 3357
+ },
+ {
+ "epoch": 46.64279475982533,
+ "grad_norm": 0.048719972372055054,
+ "learning_rate": 0.0006,
+ "loss": 6.4915618896484375,
+ "step": 3358
+ },
+ {
+ "epoch": 46.65676855895197,
+ "grad_norm": 0.042849015444517136,
+ "learning_rate": 0.0006,
+ "loss": 6.413516044616699,
+ "step": 3359
+ },
+ {
+ "epoch": 46.670742358078606,
+ "grad_norm": 0.02871009334921837,
+ "learning_rate": 0.0006,
+ "loss": 6.320952415466309,
+ "step": 3360
+ },
+ {
+ "epoch": 46.68471615720524,
+ "grad_norm": 0.03193620592355728,
+ "learning_rate": 0.0006,
+ "loss": 6.4176435470581055,
+ "step": 3361
+ },
+ {
+ "epoch": 46.698689956331876,
+ "grad_norm": 0.027417024597525597,
+ "learning_rate": 0.0006,
+ "loss": 6.342319488525391,
+ "step": 3362
+ },
+ {
+ "epoch": 46.712663755458514,
+ "grad_norm": 0.03305937349796295,
+ "learning_rate": 0.0006,
+ "loss": 6.246112823486328,
+ "step": 3363
+ },
+ {
+ "epoch": 46.72663755458515,
+ "grad_norm": 0.031403183937072754,
+ "learning_rate": 0.0006,
+ "loss": 6.149503231048584,
+ "step": 3364
+ },
+ {
+ "epoch": 46.74061135371179,
+ "grad_norm": 0.031563788652420044,
+ "learning_rate": 0.0006,
+ "loss": 6.278808116912842,
+ "step": 3365
+ },
+ {
+ "epoch": 46.75458515283843,
+ "grad_norm": 0.028366589918732643,
+ "learning_rate": 0.0006,
+ "loss": 6.057036876678467,
+ "step": 3366
+ },
+ {
+ "epoch": 46.76855895196506,
+ "grad_norm": 0.023715803399682045,
+ "learning_rate": 0.0006,
+ "loss": 6.111324310302734,
+ "step": 3367
+ },
+ {
+ "epoch": 46.7825327510917,
+ "grad_norm": 0.026546811684966087,
+ "learning_rate": 0.0006,
+ "loss": 6.105669975280762,
+ "step": 3368
+ },
+ {
+ "epoch": 46.79650655021834,
+ "grad_norm": 0.023531019687652588,
+ "learning_rate": 0.0006,
+ "loss": 6.068072319030762,
+ "step": 3369
+ },
+ {
+ "epoch": 46.81048034934498,
+ "grad_norm": 0.024480128660798073,
+ "learning_rate": 0.0006,
+ "loss": 5.947896957397461,
+ "step": 3370
+ },
+ {
+ "epoch": 46.82445414847162,
+ "grad_norm": 0.02517310529947281,
+ "learning_rate": 0.0006,
+ "loss": 6.030037879943848,
+ "step": 3371
+ },
+ {
+ "epoch": 46.838427947598255,
+ "grad_norm": 0.026723403483629227,
+ "learning_rate": 0.0006,
+ "loss": 5.963776111602783,
+ "step": 3372
+ },
+ {
+ "epoch": 46.852401746724894,
+ "grad_norm": 0.03163504600524902,
+ "learning_rate": 0.0006,
+ "loss": 5.893466949462891,
+ "step": 3373
+ },
+ {
+ "epoch": 46.866375545851525,
+ "grad_norm": 0.04058683291077614,
+ "learning_rate": 0.0006,
+ "loss": 5.9202799797058105,
+ "step": 3374
+ },
+ {
+ "epoch": 46.880349344978164,
+ "grad_norm": 0.04964117333292961,
+ "learning_rate": 0.0006,
+ "loss": 5.9366559982299805,
+ "step": 3375
+ },
+ {
+ "epoch": 46.8943231441048,
+ "grad_norm": 0.055406033992767334,
+ "learning_rate": 0.0006,
+ "loss": 5.912230968475342,
+ "step": 3376
+ },
+ {
+ "epoch": 46.90829694323144,
+ "grad_norm": 0.04643625393509865,
+ "learning_rate": 0.0006,
+ "loss": 5.87260103225708,
+ "step": 3377
+ },
+ {
+ "epoch": 46.92227074235808,
+ "grad_norm": 0.029325876384973526,
+ "learning_rate": 0.0006,
+ "loss": 5.854763984680176,
+ "step": 3378
+ },
+ {
+ "epoch": 46.93624454148472,
+ "grad_norm": 0.03159382566809654,
+ "learning_rate": 0.0006,
+ "loss": 5.880882263183594,
+ "step": 3379
+ },
+ {
+ "epoch": 46.95021834061135,
+ "grad_norm": 0.030224012210965157,
+ "learning_rate": 0.0006,
+ "loss": 5.787360191345215,
+ "step": 3380
+ },
+ {
+ "epoch": 46.96419213973799,
+ "grad_norm": 0.028368130326271057,
+ "learning_rate": 0.0006,
+ "loss": 5.817195892333984,
+ "step": 3381
+ },
+ {
+ "epoch": 46.97816593886463,
+ "grad_norm": 0.029411399737000465,
+ "learning_rate": 0.0006,
+ "loss": 5.783403396606445,
+ "step": 3382
+ },
+ {
+ "epoch": 46.992139737991266,
+ "grad_norm": 0.029927462339401245,
+ "learning_rate": 0.0006,
+ "loss": 5.662444591522217,
+ "step": 3383
+ },
+ {
+ "epoch": 47.0,
+ "grad_norm": 0.02872961387038231,
+ "learning_rate": 0.0006,
+ "loss": 5.742781639099121,
+ "step": 3384
+ },
+ {
+ "epoch": 47.0,
+ "eval_loss": 5.870260238647461,
+ "eval_runtime": 57.3734,
+ "eval_samples_per_second": 42.563,
+ "eval_steps_per_second": 1.342,
+ "step": 3384
+ },
+ {
+ "epoch": 47.01397379912664,
+ "grad_norm": 0.028542131185531616,
+ "learning_rate": 0.0006,
+ "loss": 5.753256797790527,
+ "step": 3385
+ },
+ {
+ "epoch": 47.02794759825328,
+ "grad_norm": 0.03227119892835617,
+ "learning_rate": 0.0006,
+ "loss": 5.722246170043945,
+ "step": 3386
+ },
+ {
+ "epoch": 47.041921397379916,
+ "grad_norm": 0.031461723148822784,
+ "learning_rate": 0.0006,
+ "loss": 5.740520000457764,
+ "step": 3387
+ },
+ {
+ "epoch": 47.05589519650655,
+ "grad_norm": 0.03184220567345619,
+ "learning_rate": 0.0006,
+ "loss": 5.720890045166016,
+ "step": 3388
+ },
+ {
+ "epoch": 47.069868995633186,
+ "grad_norm": 0.030015580356121063,
+ "learning_rate": 0.0006,
+ "loss": 5.597814559936523,
+ "step": 3389
+ },
+ {
+ "epoch": 47.083842794759825,
+ "grad_norm": 0.029931779950857162,
+ "learning_rate": 0.0006,
+ "loss": 5.5412468910217285,
+ "step": 3390
+ },
+ {
+ "epoch": 47.09781659388646,
+ "grad_norm": 0.03206505626440048,
+ "learning_rate": 0.0006,
+ "loss": 5.6967597007751465,
+ "step": 3391
+ },
+ {
+ "epoch": 47.1117903930131,
+ "grad_norm": 0.03978161886334419,
+ "learning_rate": 0.0006,
+ "loss": 5.59661340713501,
+ "step": 3392
+ },
+ {
+ "epoch": 47.12576419213974,
+ "grad_norm": 0.04619080200791359,
+ "learning_rate": 0.0006,
+ "loss": 5.795788764953613,
+ "step": 3393
+ },
+ {
+ "epoch": 47.13973799126637,
+ "grad_norm": 0.049085382372140884,
+ "learning_rate": 0.0006,
+ "loss": 5.667335510253906,
+ "step": 3394
+ },
+ {
+ "epoch": 47.15371179039301,
+ "grad_norm": 0.04830858111381531,
+ "learning_rate": 0.0006,
+ "loss": 5.668806552886963,
+ "step": 3395
+ },
+ {
+ "epoch": 47.16768558951965,
+ "grad_norm": 0.0583178736269474,
+ "learning_rate": 0.0006,
+ "loss": 5.692285537719727,
+ "step": 3396
+ },
+ {
+ "epoch": 47.18165938864629,
+ "grad_norm": 0.05182253196835518,
+ "learning_rate": 0.0006,
+ "loss": 5.678645133972168,
+ "step": 3397
+ },
+ {
+ "epoch": 47.19563318777293,
+ "grad_norm": 0.04011905938386917,
+ "learning_rate": 0.0006,
+ "loss": 5.661550998687744,
+ "step": 3398
+ },
+ {
+ "epoch": 47.209606986899566,
+ "grad_norm": 0.03888486698269844,
+ "learning_rate": 0.0006,
+ "loss": 5.643001556396484,
+ "step": 3399
+ },
+ {
+ "epoch": 47.223580786026204,
+ "grad_norm": 0.030658265575766563,
+ "learning_rate": 0.0006,
+ "loss": 5.659090995788574,
+ "step": 3400
+ },
+ {
+ "epoch": 47.237554585152836,
+ "grad_norm": 0.03609256446361542,
+ "learning_rate": 0.0006,
+ "loss": 5.544501781463623,
+ "step": 3401
+ },
+ {
+ "epoch": 47.251528384279474,
+ "grad_norm": 0.033589769154787064,
+ "learning_rate": 0.0006,
+ "loss": 5.463878631591797,
+ "step": 3402
+ },
+ {
+ "epoch": 47.26550218340611,
+ "grad_norm": 0.035128287971019745,
+ "learning_rate": 0.0006,
+ "loss": 5.5680060386657715,
+ "step": 3403
+ },
+ {
+ "epoch": 47.27947598253275,
+ "grad_norm": 0.038684550672769547,
+ "learning_rate": 0.0006,
+ "loss": 5.587898254394531,
+ "step": 3404
+ },
+ {
+ "epoch": 47.29344978165939,
+ "grad_norm": 0.043566226959228516,
+ "learning_rate": 0.0006,
+ "loss": 5.580896854400635,
+ "step": 3405
+ },
+ {
+ "epoch": 47.30742358078603,
+ "grad_norm": 0.04428374394774437,
+ "learning_rate": 0.0006,
+ "loss": 5.603460311889648,
+ "step": 3406
+ },
+ {
+ "epoch": 47.32139737991266,
+ "grad_norm": 0.037368934601545334,
+ "learning_rate": 0.0006,
+ "loss": 5.564740180969238,
+ "step": 3407
+ },
+ {
+ "epoch": 47.3353711790393,
+ "grad_norm": 0.04092350974678993,
+ "learning_rate": 0.0006,
+ "loss": 5.584153652191162,
+ "step": 3408
+ },
+ {
+ "epoch": 47.34934497816594,
+ "grad_norm": 0.045031506568193436,
+ "learning_rate": 0.0006,
+ "loss": 5.601186275482178,
+ "step": 3409
+ },
+ {
+ "epoch": 47.36331877729258,
+ "grad_norm": 0.041830409318208694,
+ "learning_rate": 0.0006,
+ "loss": 5.512788772583008,
+ "step": 3410
+ },
+ {
+ "epoch": 47.377292576419215,
+ "grad_norm": 0.037633974105119705,
+ "learning_rate": 0.0006,
+ "loss": 5.503828525543213,
+ "step": 3411
+ },
+ {
+ "epoch": 47.391266375545854,
+ "grad_norm": 0.03473588079214096,
+ "learning_rate": 0.0006,
+ "loss": 5.570557594299316,
+ "step": 3412
+ },
+ {
+ "epoch": 47.40524017467249,
+ "grad_norm": 0.039325181394815445,
+ "learning_rate": 0.0006,
+ "loss": 5.427783966064453,
+ "step": 3413
+ },
+ {
+ "epoch": 47.419213973799124,
+ "grad_norm": 0.04034624621272087,
+ "learning_rate": 0.0006,
+ "loss": 5.572591304779053,
+ "step": 3414
+ },
+ {
+ "epoch": 47.43318777292576,
+ "grad_norm": 0.05245184153318405,
+ "learning_rate": 0.0006,
+ "loss": 5.505208969116211,
+ "step": 3415
+ },
+ {
+ "epoch": 47.4471615720524,
+ "grad_norm": 0.0568675734102726,
+ "learning_rate": 0.0006,
+ "loss": 5.589373588562012,
+ "step": 3416
+ },
+ {
+ "epoch": 47.46113537117904,
+ "grad_norm": 0.048336587846279144,
+ "learning_rate": 0.0006,
+ "loss": 5.522005081176758,
+ "step": 3417
+ },
+ {
+ "epoch": 47.47510917030568,
+ "grad_norm": 0.04194975271821022,
+ "learning_rate": 0.0006,
+ "loss": 5.387576103210449,
+ "step": 3418
+ },
+ {
+ "epoch": 47.48908296943232,
+ "grad_norm": 0.04019433259963989,
+ "learning_rate": 0.0006,
+ "loss": 5.522615432739258,
+ "step": 3419
+ },
+ {
+ "epoch": 47.50305676855895,
+ "grad_norm": 0.05709933117032051,
+ "learning_rate": 0.0006,
+ "loss": 5.420286178588867,
+ "step": 3420
+ },
+ {
+ "epoch": 47.51703056768559,
+ "grad_norm": 0.09130338579416275,
+ "learning_rate": 0.0006,
+ "loss": 5.484622478485107,
+ "step": 3421
+ },
+ {
+ "epoch": 47.531004366812226,
+ "grad_norm": 0.2757541835308075,
+ "learning_rate": 0.0006,
+ "loss": 5.821832656860352,
+ "step": 3422
+ },
+ {
+ "epoch": 47.544978165938865,
+ "grad_norm": 0.28292080760002136,
+ "learning_rate": 0.0006,
+ "loss": 6.579934120178223,
+ "step": 3423
+ },
+ {
+ "epoch": 47.5589519650655,
+ "grad_norm": 0.15293604135513306,
+ "learning_rate": 0.0006,
+ "loss": 6.478213310241699,
+ "step": 3424
+ },
+ {
+ "epoch": 47.57292576419214,
+ "grad_norm": 0.11284393817186356,
+ "learning_rate": 0.0006,
+ "loss": 6.4066386222839355,
+ "step": 3425
+ },
+ {
+ "epoch": 47.58689956331878,
+ "grad_norm": 0.10933073610067368,
+ "learning_rate": 0.0006,
+ "loss": 6.3306169509887695,
+ "step": 3426
+ },
+ {
+ "epoch": 47.60087336244541,
+ "grad_norm": 0.05837641656398773,
+ "learning_rate": 0.0006,
+ "loss": 6.250000953674316,
+ "step": 3427
+ },
+ {
+ "epoch": 47.61484716157205,
+ "grad_norm": 0.05987229570746422,
+ "learning_rate": 0.0006,
+ "loss": 6.152463436126709,
+ "step": 3428
+ },
+ {
+ "epoch": 47.62882096069869,
+ "grad_norm": 0.04064938426017761,
+ "learning_rate": 0.0006,
+ "loss": 6.117897033691406,
+ "step": 3429
+ },
+ {
+ "epoch": 47.64279475982533,
+ "grad_norm": 0.04565057530999184,
+ "learning_rate": 0.0006,
+ "loss": 6.042291641235352,
+ "step": 3430
+ },
+ {
+ "epoch": 47.65676855895197,
+ "grad_norm": 0.03475063294172287,
+ "learning_rate": 0.0006,
+ "loss": 6.0071868896484375,
+ "step": 3431
+ },
+ {
+ "epoch": 47.670742358078606,
+ "grad_norm": 0.03809220343828201,
+ "learning_rate": 0.0006,
+ "loss": 6.0128021240234375,
+ "step": 3432
+ },
+ {
+ "epoch": 47.68471615720524,
+ "grad_norm": 0.03307219594717026,
+ "learning_rate": 0.0006,
+ "loss": 5.894664287567139,
+ "step": 3433
+ },
+ {
+ "epoch": 47.698689956331876,
+ "grad_norm": 0.028824346140027046,
+ "learning_rate": 0.0006,
+ "loss": 5.914865493774414,
+ "step": 3434
+ },
+ {
+ "epoch": 47.712663755458514,
+ "grad_norm": 0.027923690155148506,
+ "learning_rate": 0.0006,
+ "loss": 5.851759910583496,
+ "step": 3435
+ },
+ {
+ "epoch": 47.72663755458515,
+ "grad_norm": 0.025492684915661812,
+ "learning_rate": 0.0006,
+ "loss": 5.749186038970947,
+ "step": 3436
+ },
+ {
+ "epoch": 47.74061135371179,
+ "grad_norm": 0.025105174630880356,
+ "learning_rate": 0.0006,
+ "loss": 5.803566932678223,
+ "step": 3437
+ },
+ {
+ "epoch": 47.75458515283843,
+ "grad_norm": 0.023298880085349083,
+ "learning_rate": 0.0006,
+ "loss": 5.76276159286499,
+ "step": 3438
+ },
+ {
+ "epoch": 47.76855895196506,
+ "grad_norm": 0.02468576282262802,
+ "learning_rate": 0.0006,
+ "loss": 5.833852767944336,
+ "step": 3439
+ },
+ {
+ "epoch": 47.7825327510917,
+ "grad_norm": 0.022730689495801926,
+ "learning_rate": 0.0006,
+ "loss": 5.736724376678467,
+ "step": 3440
+ },
+ {
+ "epoch": 47.79650655021834,
+ "grad_norm": 0.021335354074835777,
+ "learning_rate": 0.0006,
+ "loss": 5.644248008728027,
+ "step": 3441
+ },
+ {
+ "epoch": 47.81048034934498,
+ "grad_norm": 0.02288101427257061,
+ "learning_rate": 0.0006,
+ "loss": 5.7011942863464355,
+ "step": 3442
+ },
+ {
+ "epoch": 47.82445414847162,
+ "grad_norm": 0.01961585320532322,
+ "learning_rate": 0.0006,
+ "loss": 5.63215446472168,
+ "step": 3443
+ },
+ {
+ "epoch": 47.838427947598255,
+ "grad_norm": 0.017659205943346024,
+ "learning_rate": 0.0006,
+ "loss": 5.744845390319824,
+ "step": 3444
+ },
+ {
+ "epoch": 47.852401746724894,
+ "grad_norm": 0.02281784638762474,
+ "learning_rate": 0.0006,
+ "loss": 5.696993350982666,
+ "step": 3445
+ },
+ {
+ "epoch": 47.866375545851525,
+ "grad_norm": 0.016394242644309998,
+ "learning_rate": 0.0006,
+ "loss": 5.659870147705078,
+ "step": 3446
+ },
+ {
+ "epoch": 47.880349344978164,
+ "grad_norm": 0.01967368647456169,
+ "learning_rate": 0.0006,
+ "loss": 5.638072490692139,
+ "step": 3447
+ },
+ {
+ "epoch": 47.8943231441048,
+ "grad_norm": 0.017326267436146736,
+ "learning_rate": 0.0006,
+ "loss": 5.638836860656738,
+ "step": 3448
+ },
+ {
+ "epoch": 47.90829694323144,
+ "grad_norm": 0.017778463661670685,
+ "learning_rate": 0.0006,
+ "loss": 5.634466648101807,
+ "step": 3449
+ },
+ {
+ "epoch": 47.92227074235808,
+ "grad_norm": 0.017269758507609367,
+ "learning_rate": 0.0006,
+ "loss": 5.506277561187744,
+ "step": 3450
+ },
+ {
+ "epoch": 47.93624454148472,
+ "grad_norm": 0.01733289286494255,
+ "learning_rate": 0.0006,
+ "loss": 5.544455528259277,
+ "step": 3451
+ },
+ {
+ "epoch": 47.95021834061135,
+ "grad_norm": 0.017787795513868332,
+ "learning_rate": 0.0006,
+ "loss": 5.498943328857422,
+ "step": 3452
+ },
+ {
+ "epoch": 47.96419213973799,
+ "grad_norm": 0.015770932659506798,
+ "learning_rate": 0.0006,
+ "loss": 5.537227630615234,
+ "step": 3453
+ },
+ {
+ "epoch": 47.97816593886463,
+ "grad_norm": 0.016132809221744537,
+ "learning_rate": 0.0006,
+ "loss": 5.489081382751465,
+ "step": 3454
+ },
+ {
+ "epoch": 47.992139737991266,
+ "grad_norm": 0.01749664731323719,
+ "learning_rate": 0.0006,
+ "loss": 5.454263687133789,
+ "step": 3455
+ },
+ {
+ "epoch": 48.0,
+ "grad_norm": 0.016711296513676643,
+ "learning_rate": 0.0006,
+ "loss": 5.530783653259277,
+ "step": 3456
+ },
+ {
+ "epoch": 48.0,
+ "eval_loss": 5.596925258636475,
+ "eval_runtime": 56.9612,
+ "eval_samples_per_second": 42.871,
+ "eval_steps_per_second": 1.352,
+ "step": 3456
+ },
+ {
+ "epoch": 48.01397379912664,
+ "grad_norm": 0.01867641694843769,
+ "learning_rate": 0.0006,
+ "loss": 5.506853103637695,
+ "step": 3457
+ },
+ {
+ "epoch": 48.02794759825328,
+ "grad_norm": 0.023250792175531387,
+ "learning_rate": 0.0006,
+ "loss": 5.4074177742004395,
+ "step": 3458
+ },
+ {
+ "epoch": 48.041921397379916,
+ "grad_norm": 0.01970936730504036,
+ "learning_rate": 0.0006,
+ "loss": 5.508777618408203,
+ "step": 3459
+ },
+ {
+ "epoch": 48.05589519650655,
+ "grad_norm": 0.01879260502755642,
+ "learning_rate": 0.0006,
+ "loss": 5.412424087524414,
+ "step": 3460
+ },
+ {
+ "epoch": 48.069868995633186,
+ "grad_norm": 0.017420092597603798,
+ "learning_rate": 0.0006,
+ "loss": 5.43361759185791,
+ "step": 3461
+ },
+ {
+ "epoch": 48.083842794759825,
+ "grad_norm": 0.018707862123847008,
+ "learning_rate": 0.0006,
+ "loss": 5.476800918579102,
+ "step": 3462
+ },
+ {
+ "epoch": 48.09781659388646,
+ "grad_norm": 0.023185882717370987,
+ "learning_rate": 0.0006,
+ "loss": 5.465272426605225,
+ "step": 3463
+ },
+ {
+ "epoch": 48.1117903930131,
+ "grad_norm": 0.023570368066430092,
+ "learning_rate": 0.0006,
+ "loss": 5.391471862792969,
+ "step": 3464
+ },
+ {
+ "epoch": 48.12576419213974,
+ "grad_norm": 0.02651144750416279,
+ "learning_rate": 0.0006,
+ "loss": 5.422799110412598,
+ "step": 3465
+ },
+ {
+ "epoch": 48.13973799126637,
+ "grad_norm": 0.03316329047083855,
+ "learning_rate": 0.0006,
+ "loss": 5.3952202796936035,
+ "step": 3466
+ },
+ {
+ "epoch": 48.15371179039301,
+ "grad_norm": 0.04012278467416763,
+ "learning_rate": 0.0006,
+ "loss": 5.540853023529053,
+ "step": 3467
+ },
+ {
+ "epoch": 48.16768558951965,
+ "grad_norm": 0.04763203114271164,
+ "learning_rate": 0.0006,
+ "loss": 5.387890338897705,
+ "step": 3468
+ },
+ {
+ "epoch": 48.18165938864629,
+ "grad_norm": 0.05129165202379227,
+ "learning_rate": 0.0006,
+ "loss": 5.414722442626953,
+ "step": 3469
+ },
+ {
+ "epoch": 48.19563318777293,
+ "grad_norm": 0.04438071325421333,
+ "learning_rate": 0.0006,
+ "loss": 5.386244773864746,
+ "step": 3470
+ },
+ {
+ "epoch": 48.209606986899566,
+ "grad_norm": 0.040234919637441635,
+ "learning_rate": 0.0006,
+ "loss": 5.431617736816406,
+ "step": 3471
+ },
+ {
+ "epoch": 48.223580786026204,
+ "grad_norm": 0.03217008337378502,
+ "learning_rate": 0.0006,
+ "loss": 5.30168342590332,
+ "step": 3472
+ },
+ {
+ "epoch": 48.237554585152836,
+ "grad_norm": 0.03087974339723587,
+ "learning_rate": 0.0006,
+ "loss": 5.316389083862305,
+ "step": 3473
+ },
+ {
+ "epoch": 48.251528384279474,
+ "grad_norm": 0.02789347991347313,
+ "learning_rate": 0.0006,
+ "loss": 5.385970592498779,
+ "step": 3474
+ },
+ {
+ "epoch": 48.26550218340611,
+ "grad_norm": 0.024792097508907318,
+ "learning_rate": 0.0006,
+ "loss": 5.362329959869385,
+ "step": 3475
+ },
+ {
+ "epoch": 48.27947598253275,
+ "grad_norm": 0.025869522243738174,
+ "learning_rate": 0.0006,
+ "loss": 5.28400182723999,
+ "step": 3476
+ },
+ {
+ "epoch": 48.29344978165939,
+ "grad_norm": 0.021339882165193558,
+ "learning_rate": 0.0006,
+ "loss": 5.36669921875,
+ "step": 3477
+ },
+ {
+ "epoch": 48.30742358078603,
+ "grad_norm": 0.021462520584464073,
+ "learning_rate": 0.0006,
+ "loss": 5.311222076416016,
+ "step": 3478
+ },
+ {
+ "epoch": 48.32139737991266,
+ "grad_norm": 0.02193879708647728,
+ "learning_rate": 0.0006,
+ "loss": 5.291441917419434,
+ "step": 3479
+ },
+ {
+ "epoch": 48.3353711790393,
+ "grad_norm": 0.01963093690574169,
+ "learning_rate": 0.0006,
+ "loss": 5.342780113220215,
+ "step": 3480
+ },
+ {
+ "epoch": 48.34934497816594,
+ "grad_norm": 0.020634736865758896,
+ "learning_rate": 0.0006,
+ "loss": 5.144141674041748,
+ "step": 3481
+ },
+ {
+ "epoch": 48.36331877729258,
+ "grad_norm": 0.02126968279480934,
+ "learning_rate": 0.0006,
+ "loss": 5.232624530792236,
+ "step": 3482
+ },
+ {
+ "epoch": 48.377292576419215,
+ "grad_norm": 0.021203994750976562,
+ "learning_rate": 0.0006,
+ "loss": 5.290219783782959,
+ "step": 3483
+ },
+ {
+ "epoch": 48.391266375545854,
+ "grad_norm": 0.02193225920200348,
+ "learning_rate": 0.0006,
+ "loss": 5.29197359085083,
+ "step": 3484
+ },
+ {
+ "epoch": 48.40524017467249,
+ "grad_norm": 0.023174211382865906,
+ "learning_rate": 0.0006,
+ "loss": 5.204349517822266,
+ "step": 3485
+ },
+ {
+ "epoch": 48.419213973799124,
+ "grad_norm": 0.024161914363503456,
+ "learning_rate": 0.0006,
+ "loss": 5.339351654052734,
+ "step": 3486
+ },
+ {
+ "epoch": 48.43318777292576,
+ "grad_norm": 0.02221524342894554,
+ "learning_rate": 0.0006,
+ "loss": 5.263001441955566,
+ "step": 3487
+ },
+ {
+ "epoch": 48.4471615720524,
+ "grad_norm": 0.026632830500602722,
+ "learning_rate": 0.0006,
+ "loss": 5.221281051635742,
+ "step": 3488
+ },
+ {
+ "epoch": 48.46113537117904,
+ "grad_norm": 0.034923870116472244,
+ "learning_rate": 0.0006,
+ "loss": 5.296039581298828,
+ "step": 3489
+ },
+ {
+ "epoch": 48.47510917030568,
+ "grad_norm": 0.046184979379177094,
+ "learning_rate": 0.0006,
+ "loss": 5.212769985198975,
+ "step": 3490
+ },
+ {
+ "epoch": 48.48908296943232,
+ "grad_norm": 0.05154098570346832,
+ "learning_rate": 0.0006,
+ "loss": 5.337864875793457,
+ "step": 3491
+ },
+ {
+ "epoch": 48.50305676855895,
+ "grad_norm": 0.042967259883880615,
+ "learning_rate": 0.0006,
+ "loss": 5.267958641052246,
+ "step": 3492
+ },
+ {
+ "epoch": 48.51703056768559,
+ "grad_norm": 0.05356583371758461,
+ "learning_rate": 0.0006,
+ "loss": 5.1404218673706055,
+ "step": 3493
+ },
+ {
+ "epoch": 48.531004366812226,
+ "grad_norm": 0.08219032734632492,
+ "learning_rate": 0.0006,
+ "loss": 5.300687789916992,
+ "step": 3494
+ },
+ {
+ "epoch": 48.544978165938865,
+ "grad_norm": 0.19204089045524597,
+ "learning_rate": 0.0006,
+ "loss": 5.456025123596191,
+ "step": 3495
+ },
+ {
+ "epoch": 48.5589519650655,
+ "grad_norm": 0.25917696952819824,
+ "learning_rate": 0.0006,
+ "loss": 6.156761169433594,
+ "step": 3496
+ },
+ {
+ "epoch": 48.57292576419214,
+ "grad_norm": 0.1566283404827118,
+ "learning_rate": 0.0006,
+ "loss": 5.862346649169922,
+ "step": 3497
+ },
+ {
+ "epoch": 48.58689956331878,
+ "grad_norm": 0.1761486828327179,
+ "learning_rate": 0.0006,
+ "loss": 6.012795448303223,
+ "step": 3498
+ },
+ {
+ "epoch": 48.60087336244541,
+ "grad_norm": 0.09857621043920517,
+ "learning_rate": 0.0006,
+ "loss": 5.9248151779174805,
+ "step": 3499
+ },
+ {
+ "epoch": 48.61484716157205,
+ "grad_norm": 0.05968055501580238,
+ "learning_rate": 0.0006,
+ "loss": 5.780706882476807,
+ "step": 3500
+ },
+ {
+ "epoch": 48.62882096069869,
+ "grad_norm": 0.057433564215898514,
+ "learning_rate": 0.0006,
+ "loss": 5.645388603210449,
+ "step": 3501
+ },
+ {
+ "epoch": 48.64279475982533,
+ "grad_norm": 0.039659615606069565,
+ "learning_rate": 0.0006,
+ "loss": 5.695138931274414,
+ "step": 3502
+ },
+ {
+ "epoch": 48.65676855895197,
+ "grad_norm": 0.04019695147871971,
+ "learning_rate": 0.0006,
+ "loss": 5.672435283660889,
+ "step": 3503
+ },
+ {
+ "epoch": 48.670742358078606,
+ "grad_norm": 0.03033018670976162,
+ "learning_rate": 0.0006,
+ "loss": 5.682260513305664,
+ "step": 3504
+ },
+ {
+ "epoch": 48.68471615720524,
+ "grad_norm": 0.03215787187218666,
+ "learning_rate": 0.0006,
+ "loss": 5.604182243347168,
+ "step": 3505
+ },
+ {
+ "epoch": 48.698689956331876,
+ "grad_norm": 0.026865126565098763,
+ "learning_rate": 0.0006,
+ "loss": 5.64040470123291,
+ "step": 3506
+ },
+ {
+ "epoch": 48.712663755458514,
+ "grad_norm": 0.02831609919667244,
+ "learning_rate": 0.0006,
+ "loss": 5.552720069885254,
+ "step": 3507
+ },
+ {
+ "epoch": 48.72663755458515,
+ "grad_norm": 0.024226613342761993,
+ "learning_rate": 0.0006,
+ "loss": 5.556946277618408,
+ "step": 3508
+ },
+ {
+ "epoch": 48.74061135371179,
+ "grad_norm": 0.025393934920430183,
+ "learning_rate": 0.0006,
+ "loss": 5.476178169250488,
+ "step": 3509
+ },
+ {
+ "epoch": 48.75458515283843,
+ "grad_norm": 0.01984412968158722,
+ "learning_rate": 0.0006,
+ "loss": 5.471088409423828,
+ "step": 3510
+ },
+ {
+ "epoch": 48.76855895196506,
+ "grad_norm": 0.02035129815340042,
+ "learning_rate": 0.0006,
+ "loss": 5.438812732696533,
+ "step": 3511
+ },
+ {
+ "epoch": 48.7825327510917,
+ "grad_norm": 0.0208350270986557,
+ "learning_rate": 0.0006,
+ "loss": 5.381746292114258,
+ "step": 3512
+ },
+ {
+ "epoch": 48.79650655021834,
+ "grad_norm": 0.017783882096409798,
+ "learning_rate": 0.0006,
+ "loss": 5.351675510406494,
+ "step": 3513
+ },
+ {
+ "epoch": 48.81048034934498,
+ "grad_norm": 0.01671770215034485,
+ "learning_rate": 0.0006,
+ "loss": 5.329428672790527,
+ "step": 3514
+ },
+ {
+ "epoch": 48.82445414847162,
+ "grad_norm": 0.017891215160489082,
+ "learning_rate": 0.0006,
+ "loss": 5.212074279785156,
+ "step": 3515
+ },
+ {
+ "epoch": 48.838427947598255,
+ "grad_norm": 0.016847791150212288,
+ "learning_rate": 0.0006,
+ "loss": 5.299534797668457,
+ "step": 3516
+ },
+ {
+ "epoch": 48.852401746724894,
+ "grad_norm": 0.01593637838959694,
+ "learning_rate": 0.0006,
+ "loss": 5.30466890335083,
+ "step": 3517
+ },
+ {
+ "epoch": 48.866375545851525,
+ "grad_norm": 0.014713102020323277,
+ "learning_rate": 0.0006,
+ "loss": 5.294345378875732,
+ "step": 3518
+ },
+ {
+ "epoch": 48.880349344978164,
+ "grad_norm": 0.01433857623487711,
+ "learning_rate": 0.0006,
+ "loss": 5.2148542404174805,
+ "step": 3519
+ },
+ {
+ "epoch": 48.8943231441048,
+ "grad_norm": 0.014979379251599312,
+ "learning_rate": 0.0006,
+ "loss": 5.298582077026367,
+ "step": 3520
+ },
+ {
+ "epoch": 48.90829694323144,
+ "grad_norm": 0.013494192622601986,
+ "learning_rate": 0.0006,
+ "loss": 5.309783935546875,
+ "step": 3521
+ },
+ {
+ "epoch": 48.92227074235808,
+ "grad_norm": 0.014020012691617012,
+ "learning_rate": 0.0006,
+ "loss": 5.284879684448242,
+ "step": 3522
+ },
+ {
+ "epoch": 48.93624454148472,
+ "grad_norm": 0.01348058134317398,
+ "learning_rate": 0.0006,
+ "loss": 5.249116897583008,
+ "step": 3523
+ },
+ {
+ "epoch": 48.95021834061135,
+ "grad_norm": 0.013280218467116356,
+ "learning_rate": 0.0006,
+ "loss": 5.273489952087402,
+ "step": 3524
+ },
+ {
+ "epoch": 48.96419213973799,
+ "grad_norm": 0.012647110037505627,
+ "learning_rate": 0.0006,
+ "loss": 5.306652069091797,
+ "step": 3525
+ },
+ {
+ "epoch": 48.97816593886463,
+ "grad_norm": 0.012847724370658398,
+ "learning_rate": 0.0006,
+ "loss": 5.235745429992676,
+ "step": 3526
+ },
+ {
+ "epoch": 48.992139737991266,
+ "grad_norm": 0.011820166371762753,
+ "learning_rate": 0.0006,
+ "loss": 5.238091468811035,
+ "step": 3527
+ },
+ {
+ "epoch": 49.0,
+ "grad_norm": 0.01255585066974163,
+ "learning_rate": 0.0006,
+ "loss": 5.231295585632324,
+ "step": 3528
+ },
+ {
+ "epoch": 49.0,
+ "eval_loss": 5.394587993621826,
+ "eval_runtime": 56.3734,
+ "eval_samples_per_second": 43.318,
+ "eval_steps_per_second": 1.366,
+ "step": 3528
+ },
+ {
+ "epoch": 49.01397379912664,
+ "grad_norm": 0.011031972244381905,
+ "learning_rate": 0.0006,
+ "loss": 5.1354570388793945,
+ "step": 3529
+ },
+ {
+ "epoch": 49.02794759825328,
+ "grad_norm": 0.010631263256072998,
+ "learning_rate": 0.0006,
+ "loss": 5.177639961242676,
+ "step": 3530
+ },
+ {
+ "epoch": 49.041921397379916,
+ "grad_norm": 0.01133775431662798,
+ "learning_rate": 0.0006,
+ "loss": 5.283428192138672,
+ "step": 3531
+ },
+ {
+ "epoch": 49.05589519650655,
+ "grad_norm": 0.01154172234237194,
+ "learning_rate": 0.0006,
+ "loss": 5.206364154815674,
+ "step": 3532
+ },
+ {
+ "epoch": 49.069868995633186,
+ "grad_norm": 0.010192356072366238,
+ "learning_rate": 0.0006,
+ "loss": 5.2385735511779785,
+ "step": 3533
+ },
+ {
+ "epoch": 49.083842794759825,
+ "grad_norm": 0.012789200060069561,
+ "learning_rate": 0.0006,
+ "loss": 5.071800231933594,
+ "step": 3534
+ },
+ {
+ "epoch": 49.09781659388646,
+ "grad_norm": 0.011282820254564285,
+ "learning_rate": 0.0006,
+ "loss": 5.218020439147949,
+ "step": 3535
+ },
+ {
+ "epoch": 49.1117903930131,
+ "grad_norm": 0.012153448536992073,
+ "learning_rate": 0.0006,
+ "loss": 5.134698867797852,
+ "step": 3536
+ },
+ {
+ "epoch": 49.12576419213974,
+ "grad_norm": 0.011920543387532234,
+ "learning_rate": 0.0006,
+ "loss": 5.141620635986328,
+ "step": 3537
+ },
+ {
+ "epoch": 49.13973799126637,
+ "grad_norm": 0.010949329473078251,
+ "learning_rate": 0.0006,
+ "loss": 5.116259574890137,
+ "step": 3538
+ },
+ {
+ "epoch": 49.15371179039301,
+ "grad_norm": 0.012362448498606682,
+ "learning_rate": 0.0006,
+ "loss": 5.090177059173584,
+ "step": 3539
+ },
+ {
+ "epoch": 49.16768558951965,
+ "grad_norm": 0.011575071141123772,
+ "learning_rate": 0.0006,
+ "loss": 5.093441009521484,
+ "step": 3540
+ },
+ {
+ "epoch": 49.18165938864629,
+ "grad_norm": 0.011126236990094185,
+ "learning_rate": 0.0006,
+ "loss": 5.060235023498535,
+ "step": 3541
+ },
+ {
+ "epoch": 49.19563318777293,
+ "grad_norm": 0.0107401292771101,
+ "learning_rate": 0.0006,
+ "loss": 5.140704154968262,
+ "step": 3542
+ },
+ {
+ "epoch": 49.209606986899566,
+ "grad_norm": 0.011202906258404255,
+ "learning_rate": 0.0006,
+ "loss": 5.10855770111084,
+ "step": 3543
+ },
+ {
+ "epoch": 49.223580786026204,
+ "grad_norm": 0.011083454824984074,
+ "learning_rate": 0.0006,
+ "loss": 5.092751502990723,
+ "step": 3544
+ },
+ {
+ "epoch": 49.237554585152836,
+ "grad_norm": 0.009523800574243069,
+ "learning_rate": 0.0006,
+ "loss": 5.13733434677124,
+ "step": 3545
+ },
+ {
+ "epoch": 49.251528384279474,
+ "grad_norm": 0.011298353783786297,
+ "learning_rate": 0.0006,
+ "loss": 5.036048889160156,
+ "step": 3546
+ },
+ {
+ "epoch": 49.26550218340611,
+ "grad_norm": 0.010676483623683453,
+ "learning_rate": 0.0006,
+ "loss": 5.051901340484619,
+ "step": 3547
+ },
+ {
+ "epoch": 49.27947598253275,
+ "grad_norm": 0.011463399045169353,
+ "learning_rate": 0.0006,
+ "loss": 4.983170986175537,
+ "step": 3548
+ },
+ {
+ "epoch": 49.29344978165939,
+ "grad_norm": 0.010603402741253376,
+ "learning_rate": 0.0006,
+ "loss": 5.112311363220215,
+ "step": 3549
+ },
+ {
+ "epoch": 49.30742358078603,
+ "grad_norm": 0.010535670444369316,
+ "learning_rate": 0.0006,
+ "loss": 5.085227966308594,
+ "step": 3550
+ },
+ {
+ "epoch": 49.32139737991266,
+ "grad_norm": 0.011843928135931492,
+ "learning_rate": 0.0006,
+ "loss": 5.07575798034668,
+ "step": 3551
+ },
+ {
+ "epoch": 49.3353711790393,
+ "grad_norm": 0.010587952099740505,
+ "learning_rate": 0.0006,
+ "loss": 5.019237041473389,
+ "step": 3552
+ },
+ {
+ "epoch": 49.34934497816594,
+ "grad_norm": 0.012285659089684486,
+ "learning_rate": 0.0006,
+ "loss": 5.144603729248047,
+ "step": 3553
+ },
+ {
+ "epoch": 49.36331877729258,
+ "grad_norm": 0.013119431212544441,
+ "learning_rate": 0.0006,
+ "loss": 5.035256862640381,
+ "step": 3554
+ },
+ {
+ "epoch": 49.377292576419215,
+ "grad_norm": 0.012138537131249905,
+ "learning_rate": 0.0006,
+ "loss": 4.996667861938477,
+ "step": 3555
+ },
+ {
+ "epoch": 49.391266375545854,
+ "grad_norm": 0.012142804451286793,
+ "learning_rate": 0.0006,
+ "loss": 4.95058536529541,
+ "step": 3556
+ },
+ {
+ "epoch": 49.40524017467249,
+ "grad_norm": 0.012205676175653934,
+ "learning_rate": 0.0006,
+ "loss": 5.096931457519531,
+ "step": 3557
+ },
+ {
+ "epoch": 49.419213973799124,
+ "grad_norm": 0.012086744420230389,
+ "learning_rate": 0.0006,
+ "loss": 4.99384880065918,
+ "step": 3558
+ },
+ {
+ "epoch": 49.43318777292576,
+ "grad_norm": 0.012918103486299515,
+ "learning_rate": 0.0006,
+ "loss": 5.031294822692871,
+ "step": 3559
+ },
+ {
+ "epoch": 49.4471615720524,
+ "grad_norm": 0.012006713077425957,
+ "learning_rate": 0.0006,
+ "loss": 5.0547943115234375,
+ "step": 3560
+ },
+ {
+ "epoch": 49.46113537117904,
+ "grad_norm": 0.0120399734005332,
+ "learning_rate": 0.0006,
+ "loss": 5.014238357543945,
+ "step": 3561
+ },
+ {
+ "epoch": 49.47510917030568,
+ "grad_norm": 0.013306990265846252,
+ "learning_rate": 0.0006,
+ "loss": 4.972925186157227,
+ "step": 3562
+ },
+ {
+ "epoch": 49.48908296943232,
+ "grad_norm": 0.013878636993467808,
+ "learning_rate": 0.0006,
+ "loss": 5.079033851623535,
+ "step": 3563
+ },
+ {
+ "epoch": 49.50305676855895,
+ "grad_norm": 0.012454566545784473,
+ "learning_rate": 0.0006,
+ "loss": 4.93708610534668,
+ "step": 3564
+ },
+ {
+ "epoch": 49.51703056768559,
+ "grad_norm": 0.014075061306357384,
+ "learning_rate": 0.0006,
+ "loss": 4.965582370758057,
+ "step": 3565
+ },
+ {
+ "epoch": 49.531004366812226,
+ "grad_norm": 0.012525617145001888,
+ "learning_rate": 0.0006,
+ "loss": 4.968148708343506,
+ "step": 3566
+ },
+ {
+ "epoch": 49.544978165938865,
+ "grad_norm": 0.012418225407600403,
+ "learning_rate": 0.0006,
+ "loss": 5.06901741027832,
+ "step": 3567
+ },
+ {
+ "epoch": 49.5589519650655,
+ "grad_norm": 0.015034239739179611,
+ "learning_rate": 0.0006,
+ "loss": 4.984277725219727,
+ "step": 3568
+ },
+ {
+ "epoch": 49.57292576419214,
+ "grad_norm": 0.014329628087580204,
+ "learning_rate": 0.0006,
+ "loss": 4.951558589935303,
+ "step": 3569
+ },
+ {
+ "epoch": 49.58689956331878,
+ "grad_norm": 0.017769070342183113,
+ "learning_rate": 0.0006,
+ "loss": 4.966532230377197,
+ "step": 3570
+ },
+ {
+ "epoch": 49.60087336244541,
+ "grad_norm": 0.01909453794360161,
+ "learning_rate": 0.0006,
+ "loss": 4.919890880584717,
+ "step": 3571
+ },
+ {
+ "epoch": 49.61484716157205,
+ "grad_norm": 0.017519688233733177,
+ "learning_rate": 0.0006,
+ "loss": 5.031286716461182,
+ "step": 3572
+ },
+ {
+ "epoch": 49.62882096069869,
+ "grad_norm": 0.017143897712230682,
+ "learning_rate": 0.0006,
+ "loss": 4.969764709472656,
+ "step": 3573
+ },
+ {
+ "epoch": 49.64279475982533,
+ "grad_norm": 0.01688285917043686,
+ "learning_rate": 0.0006,
+ "loss": 5.0733442306518555,
+ "step": 3574
+ },
+ {
+ "epoch": 49.65676855895197,
+ "grad_norm": 0.018547311425209045,
+ "learning_rate": 0.0006,
+ "loss": 4.980863571166992,
+ "step": 3575
+ },
+ {
+ "epoch": 49.670742358078606,
+ "grad_norm": 0.018877137452363968,
+ "learning_rate": 0.0006,
+ "loss": 4.954646110534668,
+ "step": 3576
+ },
+ {
+ "epoch": 49.68471615720524,
+ "grad_norm": 0.018681600689888,
+ "learning_rate": 0.0006,
+ "loss": 4.959702014923096,
+ "step": 3577
+ },
+ {
+ "epoch": 49.698689956331876,
+ "grad_norm": 0.018947336822748184,
+ "learning_rate": 0.0006,
+ "loss": 5.070502758026123,
+ "step": 3578
+ },
+ {
+ "epoch": 49.712663755458514,
+ "grad_norm": 0.018817484378814697,
+ "learning_rate": 0.0006,
+ "loss": 5.056508541107178,
+ "step": 3579
+ },
+ {
+ "epoch": 49.72663755458515,
+ "grad_norm": 0.01931724324822426,
+ "learning_rate": 0.0006,
+ "loss": 4.922451972961426,
+ "step": 3580
+ },
+ {
+ "epoch": 49.74061135371179,
+ "grad_norm": 0.022047819569706917,
+ "learning_rate": 0.0006,
+ "loss": 5.053170204162598,
+ "step": 3581
+ },
+ {
+ "epoch": 49.75458515283843,
+ "grad_norm": 0.02279801294207573,
+ "learning_rate": 0.0006,
+ "loss": 4.8768439292907715,
+ "step": 3582
+ },
+ {
+ "epoch": 49.76855895196506,
+ "grad_norm": 0.020661601796746254,
+ "learning_rate": 0.0006,
+ "loss": 4.907928466796875,
+ "step": 3583
+ },
+ {
+ "epoch": 49.7825327510917,
+ "grad_norm": 0.020479509606957436,
+ "learning_rate": 0.0006,
+ "loss": 5.040694236755371,
+ "step": 3584
+ },
+ {
+ "epoch": 49.79650655021834,
+ "grad_norm": 0.021654162555933,
+ "learning_rate": 0.0006,
+ "loss": 5.018434524536133,
+ "step": 3585
+ },
+ {
+ "epoch": 49.81048034934498,
+ "grad_norm": 0.022486824542284012,
+ "learning_rate": 0.0006,
+ "loss": 5.086538314819336,
+ "step": 3586
+ },
+ {
+ "epoch": 49.82445414847162,
+ "grad_norm": 0.02141369879245758,
+ "learning_rate": 0.0006,
+ "loss": 5.035694122314453,
+ "step": 3587
+ },
+ {
+ "epoch": 49.838427947598255,
+ "grad_norm": 0.015926677733659744,
+ "learning_rate": 0.0006,
+ "loss": 4.905038833618164,
+ "step": 3588
+ },
+ {
+ "epoch": 49.852401746724894,
+ "grad_norm": 0.015775298699736595,
+ "learning_rate": 0.0006,
+ "loss": 5.003902435302734,
+ "step": 3589
+ },
+ {
+ "epoch": 49.866375545851525,
+ "grad_norm": 0.01872118003666401,
+ "learning_rate": 0.0006,
+ "loss": 4.891060829162598,
+ "step": 3590
+ },
+ {
+ "epoch": 49.880349344978164,
+ "grad_norm": 0.02066192403435707,
+ "learning_rate": 0.0006,
+ "loss": 5.002897262573242,
+ "step": 3591
+ },
+ {
+ "epoch": 49.8943231441048,
+ "grad_norm": 0.02348714880645275,
+ "learning_rate": 0.0006,
+ "loss": 4.9189958572387695,
+ "step": 3592
+ },
+ {
+ "epoch": 49.90829694323144,
+ "grad_norm": 0.020402103662490845,
+ "learning_rate": 0.0006,
+ "loss": 4.992661476135254,
+ "step": 3593
+ },
+ {
+ "epoch": 49.92227074235808,
+ "grad_norm": 0.020466072484850883,
+ "learning_rate": 0.0006,
+ "loss": 4.962968826293945,
+ "step": 3594
+ },
+ {
+ "epoch": 49.93624454148472,
+ "grad_norm": 0.01922314241528511,
+ "learning_rate": 0.0006,
+ "loss": 4.897730827331543,
+ "step": 3595
+ },
+ {
+ "epoch": 49.95021834061135,
+ "grad_norm": 0.017228754237294197,
+ "learning_rate": 0.0006,
+ "loss": 4.964842796325684,
+ "step": 3596
+ },
+ {
+ "epoch": 49.96419213973799,
+ "grad_norm": 0.016021743416786194,
+ "learning_rate": 0.0006,
+ "loss": 4.910758972167969,
+ "step": 3597
+ },
+ {
+ "epoch": 49.97816593886463,
+ "grad_norm": 0.017985103651881218,
+ "learning_rate": 0.0006,
+ "loss": 4.916072368621826,
+ "step": 3598
+ },
+ {
+ "epoch": 49.992139737991266,
+ "grad_norm": 0.018780240789055824,
+ "learning_rate": 0.0006,
+ "loss": 4.94658088684082,
+ "step": 3599
+ },
+ {
+ "epoch": 50.0,
+ "grad_norm": 0.01776314526796341,
+ "learning_rate": 0.0006,
+ "loss": 4.969357967376709,
+ "step": 3600
+ },
+ {
+ "epoch": 50.0,
+ "eval_loss": 5.141158580780029,
+ "eval_runtime": 56.2332,
+ "eval_samples_per_second": 43.426,
+ "eval_steps_per_second": 1.369,
+ "step": 3600
+ },
+ {
+ "epoch": 50.01397379912664,
+ "grad_norm": 0.016308696940541267,
+ "learning_rate": 0.0006,
+ "loss": 4.924664497375488,
+ "step": 3601
+ },
+ {
+ "epoch": 50.02794759825328,
+ "grad_norm": 0.018810199573636055,
+ "learning_rate": 0.0006,
+ "loss": 4.82293701171875,
+ "step": 3602
+ },
+ {
+ "epoch": 50.041921397379916,
+ "grad_norm": 0.02227160707116127,
+ "learning_rate": 0.0006,
+ "loss": 4.9612298011779785,
+ "step": 3603
+ },
+ {
+ "epoch": 50.05589519650655,
+ "grad_norm": 0.025266462936997414,
+ "learning_rate": 0.0006,
+ "loss": 4.963835716247559,
+ "step": 3604
+ },
+ {
+ "epoch": 50.069868995633186,
+ "grad_norm": 0.028588181361556053,
+ "learning_rate": 0.0006,
+ "loss": 4.992118835449219,
+ "step": 3605
+ },
+ {
+ "epoch": 50.083842794759825,
+ "grad_norm": 0.02739487960934639,
+ "learning_rate": 0.0006,
+ "loss": 4.915547847747803,
+ "step": 3606
+ },
+ {
+ "epoch": 50.09781659388646,
+ "grad_norm": 0.027000300586223602,
+ "learning_rate": 0.0006,
+ "loss": 4.883801460266113,
+ "step": 3607
+ },
+ {
+ "epoch": 50.1117903930131,
+ "grad_norm": 0.023468483239412308,
+ "learning_rate": 0.0006,
+ "loss": 5.006168365478516,
+ "step": 3608
+ },
+ {
+ "epoch": 50.12576419213974,
+ "grad_norm": 0.026753634214401245,
+ "learning_rate": 0.0006,
+ "loss": 4.92526912689209,
+ "step": 3609
+ },
+ {
+ "epoch": 50.13973799126637,
+ "grad_norm": 0.026816733181476593,
+ "learning_rate": 0.0006,
+ "loss": 4.847546100616455,
+ "step": 3610
+ },
+ {
+ "epoch": 50.15371179039301,
+ "grad_norm": 0.027362391352653503,
+ "learning_rate": 0.0006,
+ "loss": 4.862548351287842,
+ "step": 3611
+ },
+ {
+ "epoch": 50.16768558951965,
+ "grad_norm": 0.03450845181941986,
+ "learning_rate": 0.0006,
+ "loss": 4.963202953338623,
+ "step": 3612
+ },
+ {
+ "epoch": 50.18165938864629,
+ "grad_norm": 0.03801408037543297,
+ "learning_rate": 0.0006,
+ "loss": 4.770240306854248,
+ "step": 3613
+ },
+ {
+ "epoch": 50.19563318777293,
+ "grad_norm": 0.03579140082001686,
+ "learning_rate": 0.0006,
+ "loss": 4.797981262207031,
+ "step": 3614
+ },
+ {
+ "epoch": 50.209606986899566,
+ "grad_norm": 0.029393760487437248,
+ "learning_rate": 0.0006,
+ "loss": 4.798581600189209,
+ "step": 3615
+ },
+ {
+ "epoch": 50.223580786026204,
+ "grad_norm": 0.026111142709851265,
+ "learning_rate": 0.0006,
+ "loss": 4.990712642669678,
+ "step": 3616
+ },
+ {
+ "epoch": 50.237554585152836,
+ "grad_norm": 0.02213387005031109,
+ "learning_rate": 0.0006,
+ "loss": 4.865842342376709,
+ "step": 3617
+ },
+ {
+ "epoch": 50.251528384279474,
+ "grad_norm": 0.025450684130191803,
+ "learning_rate": 0.0006,
+ "loss": 4.903746128082275,
+ "step": 3618
+ },
+ {
+ "epoch": 50.26550218340611,
+ "grad_norm": 0.020915117114782333,
+ "learning_rate": 0.0006,
+ "loss": 4.951366424560547,
+ "step": 3619
+ },
+ {
+ "epoch": 50.27947598253275,
+ "grad_norm": 0.020972849801182747,
+ "learning_rate": 0.0006,
+ "loss": 4.783311367034912,
+ "step": 3620
+ },
+ {
+ "epoch": 50.29344978165939,
+ "grad_norm": 0.016956273466348648,
+ "learning_rate": 0.0006,
+ "loss": 4.955767631530762,
+ "step": 3621
+ },
+ {
+ "epoch": 50.30742358078603,
+ "grad_norm": 0.018045252189040184,
+ "learning_rate": 0.0006,
+ "loss": 4.869322776794434,
+ "step": 3622
+ },
+ {
+ "epoch": 50.32139737991266,
+ "grad_norm": 0.018199188634753227,
+ "learning_rate": 0.0006,
+ "loss": 4.95680046081543,
+ "step": 3623
+ },
+ {
+ "epoch": 50.3353711790393,
+ "grad_norm": 0.016688082367181778,
+ "learning_rate": 0.0006,
+ "loss": 4.783535957336426,
+ "step": 3624
+ },
+ {
+ "epoch": 50.34934497816594,
+ "grad_norm": 0.017540352419018745,
+ "learning_rate": 0.0006,
+ "loss": 4.8236083984375,
+ "step": 3625
+ },
+ {
+ "epoch": 50.36331877729258,
+ "grad_norm": 0.018054187297821045,
+ "learning_rate": 0.0006,
+ "loss": 4.764655113220215,
+ "step": 3626
+ },
+ {
+ "epoch": 50.377292576419215,
+ "grad_norm": 0.01741131953895092,
+ "learning_rate": 0.0006,
+ "loss": 4.875613689422607,
+ "step": 3627
+ },
+ {
+ "epoch": 50.391266375545854,
+ "grad_norm": 0.01682269014418125,
+ "learning_rate": 0.0006,
+ "loss": 4.818509101867676,
+ "step": 3628
+ },
+ {
+ "epoch": 50.40524017467249,
+ "grad_norm": 0.016652341932058334,
+ "learning_rate": 0.0006,
+ "loss": 4.974020481109619,
+ "step": 3629
+ },
+ {
+ "epoch": 50.419213973799124,
+ "grad_norm": 0.02035282365977764,
+ "learning_rate": 0.0006,
+ "loss": 4.893037796020508,
+ "step": 3630
+ },
+ {
+ "epoch": 50.43318777292576,
+ "grad_norm": 0.019907061010599136,
+ "learning_rate": 0.0006,
+ "loss": 4.848119735717773,
+ "step": 3631
+ },
+ {
+ "epoch": 50.4471615720524,
+ "grad_norm": 0.023861493915319443,
+ "learning_rate": 0.0006,
+ "loss": 4.819353103637695,
+ "step": 3632
+ },
+ {
+ "epoch": 50.46113537117904,
+ "grad_norm": 0.025254026055336,
+ "learning_rate": 0.0006,
+ "loss": 4.769312858581543,
+ "step": 3633
+ },
+ {
+ "epoch": 50.47510917030568,
+ "grad_norm": 0.026154937222599983,
+ "learning_rate": 0.0006,
+ "loss": 4.7586774826049805,
+ "step": 3634
+ },
+ {
+ "epoch": 50.48908296943232,
+ "grad_norm": 0.026342090219259262,
+ "learning_rate": 0.0006,
+ "loss": 4.861218452453613,
+ "step": 3635
+ },
+ {
+ "epoch": 50.50305676855895,
+ "grad_norm": 0.028169559314846992,
+ "learning_rate": 0.0006,
+ "loss": 4.818380355834961,
+ "step": 3636
+ },
+ {
+ "epoch": 50.51703056768559,
+ "grad_norm": 0.02820264920592308,
+ "learning_rate": 0.0006,
+ "loss": 4.892871856689453,
+ "step": 3637
+ },
+ {
+ "epoch": 50.531004366812226,
+ "grad_norm": 0.03345080465078354,
+ "learning_rate": 0.0006,
+ "loss": 4.811084747314453,
+ "step": 3638
+ },
+ {
+ "epoch": 50.544978165938865,
+ "grad_norm": 0.04005736857652664,
+ "learning_rate": 0.0006,
+ "loss": 4.765410900115967,
+ "step": 3639
+ },
+ {
+ "epoch": 50.5589519650655,
+ "grad_norm": 0.04333870857954025,
+ "learning_rate": 0.0006,
+ "loss": 4.956520080566406,
+ "step": 3640
+ },
+ {
+ "epoch": 50.57292576419214,
+ "grad_norm": 0.04216952994465828,
+ "learning_rate": 0.0006,
+ "loss": 4.793441295623779,
+ "step": 3641
+ },
+ {
+ "epoch": 50.58689956331878,
+ "grad_norm": 0.02936951443552971,
+ "learning_rate": 0.0006,
+ "loss": 4.8447957038879395,
+ "step": 3642
+ },
+ {
+ "epoch": 50.60087336244541,
+ "grad_norm": 0.026937026530504227,
+ "learning_rate": 0.0006,
+ "loss": 4.962215423583984,
+ "step": 3643
+ },
+ {
+ "epoch": 50.61484716157205,
+ "grad_norm": 0.02628890983760357,
+ "learning_rate": 0.0006,
+ "loss": 4.883061408996582,
+ "step": 3644
+ },
+ {
+ "epoch": 50.62882096069869,
+ "grad_norm": 0.025984834879636765,
+ "learning_rate": 0.0006,
+ "loss": 4.768662452697754,
+ "step": 3645
+ },
+ {
+ "epoch": 50.64279475982533,
+ "grad_norm": 0.023998012766242027,
+ "learning_rate": 0.0006,
+ "loss": 4.851325035095215,
+ "step": 3646
+ },
+ {
+ "epoch": 50.65676855895197,
+ "grad_norm": 0.02143184095621109,
+ "learning_rate": 0.0006,
+ "loss": 4.8274688720703125,
+ "step": 3647
+ },
+ {
+ "epoch": 50.670742358078606,
+ "grad_norm": 0.02488900162279606,
+ "learning_rate": 0.0006,
+ "loss": 4.831140518188477,
+ "step": 3648
+ },
+ {
+ "epoch": 50.68471615720524,
+ "grad_norm": 0.02330533228814602,
+ "learning_rate": 0.0006,
+ "loss": 4.850462913513184,
+ "step": 3649
+ },
+ {
+ "epoch": 50.698689956331876,
+ "grad_norm": 0.02121943049132824,
+ "learning_rate": 0.0006,
+ "loss": 4.870253086090088,
+ "step": 3650
+ },
+ {
+ "epoch": 50.712663755458514,
+ "grad_norm": 0.02080935798585415,
+ "learning_rate": 0.0006,
+ "loss": 4.798770904541016,
+ "step": 3651
+ },
+ {
+ "epoch": 50.72663755458515,
+ "grad_norm": 0.019047029316425323,
+ "learning_rate": 0.0006,
+ "loss": 4.8427300453186035,
+ "step": 3652
+ },
+ {
+ "epoch": 50.74061135371179,
+ "grad_norm": 0.01824103854596615,
+ "learning_rate": 0.0006,
+ "loss": 4.820980548858643,
+ "step": 3653
+ },
+ {
+ "epoch": 50.75458515283843,
+ "grad_norm": 0.018129967153072357,
+ "learning_rate": 0.0006,
+ "loss": 4.946732997894287,
+ "step": 3654
+ },
+ {
+ "epoch": 50.76855895196506,
+ "grad_norm": 0.01784815452992916,
+ "learning_rate": 0.0006,
+ "loss": 4.829516410827637,
+ "step": 3655
+ },
+ {
+ "epoch": 50.7825327510917,
+ "grad_norm": 0.020021555945277214,
+ "learning_rate": 0.0006,
+ "loss": 4.780428409576416,
+ "step": 3656
+ },
+ {
+ "epoch": 50.79650655021834,
+ "grad_norm": 0.01863253302872181,
+ "learning_rate": 0.0006,
+ "loss": 4.791567802429199,
+ "step": 3657
+ },
+ {
+ "epoch": 50.81048034934498,
+ "grad_norm": 0.019151683896780014,
+ "learning_rate": 0.0006,
+ "loss": 4.824892997741699,
+ "step": 3658
+ },
+ {
+ "epoch": 50.82445414847162,
+ "grad_norm": 0.01891648955643177,
+ "learning_rate": 0.0006,
+ "loss": 4.789605140686035,
+ "step": 3659
+ },
+ {
+ "epoch": 50.838427947598255,
+ "grad_norm": 0.016019510105252266,
+ "learning_rate": 0.0006,
+ "loss": 4.779555320739746,
+ "step": 3660
+ },
+ {
+ "epoch": 50.852401746724894,
+ "grad_norm": 0.016096491366624832,
+ "learning_rate": 0.0006,
+ "loss": 4.753842353820801,
+ "step": 3661
+ },
+ {
+ "epoch": 50.866375545851525,
+ "grad_norm": 0.01693476364016533,
+ "learning_rate": 0.0006,
+ "loss": 4.768477439880371,
+ "step": 3662
+ },
+ {
+ "epoch": 50.880349344978164,
+ "grad_norm": 0.01749575324356556,
+ "learning_rate": 0.0006,
+ "loss": 4.779143333435059,
+ "step": 3663
+ },
+ {
+ "epoch": 50.8943231441048,
+ "grad_norm": 0.016851043328642845,
+ "learning_rate": 0.0006,
+ "loss": 4.780387878417969,
+ "step": 3664
+ },
+ {
+ "epoch": 50.90829694323144,
+ "grad_norm": 0.019045202061533928,
+ "learning_rate": 0.0006,
+ "loss": 4.769913673400879,
+ "step": 3665
+ },
+ {
+ "epoch": 50.92227074235808,
+ "grad_norm": 0.01954123005270958,
+ "learning_rate": 0.0006,
+ "loss": 4.769009590148926,
+ "step": 3666
+ },
+ {
+ "epoch": 50.93624454148472,
+ "grad_norm": 0.022527096793055534,
+ "learning_rate": 0.0006,
+ "loss": 4.789215087890625,
+ "step": 3667
+ },
+ {
+ "epoch": 50.95021834061135,
+ "grad_norm": 0.02871634066104889,
+ "learning_rate": 0.0006,
+ "loss": 4.749727725982666,
+ "step": 3668
+ },
+ {
+ "epoch": 50.96419213973799,
+ "grad_norm": 0.04045053571462631,
+ "learning_rate": 0.0006,
+ "loss": 4.909038543701172,
+ "step": 3669
+ },
+ {
+ "epoch": 50.97816593886463,
+ "grad_norm": 0.04156879335641861,
+ "learning_rate": 0.0006,
+ "loss": 4.876831531524658,
+ "step": 3670
+ },
+ {
+ "epoch": 50.992139737991266,
+ "grad_norm": 0.02621397376060486,
+ "learning_rate": 0.0006,
+ "loss": 4.797173500061035,
+ "step": 3671
+ },
+ {
+ "epoch": 51.0,
+ "grad_norm": 0.03389597684144974,
+ "learning_rate": 0.0006,
+ "loss": 4.8428850173950195,
+ "step": 3672
+ },
+ {
+ "epoch": 51.0,
+ "eval_loss": 4.989261627197266,
+ "eval_runtime": 56.358,
+ "eval_samples_per_second": 43.33,
+ "eval_steps_per_second": 1.366,
+ "step": 3672
+ },
+ {
+ "epoch": 51.01397379912664,
+ "grad_norm": 0.03216972202062607,
+ "learning_rate": 0.0006,
+ "loss": 4.755102634429932,
+ "step": 3673
+ },
+ {
+ "epoch": 51.02794759825328,
+ "grad_norm": 0.03190374746918678,
+ "learning_rate": 0.0006,
+ "loss": 4.720424652099609,
+ "step": 3674
+ },
+ {
+ "epoch": 51.041921397379916,
+ "grad_norm": 0.031727299094200134,
+ "learning_rate": 0.0006,
+ "loss": 4.803277969360352,
+ "step": 3675
+ },
+ {
+ "epoch": 51.05589519650655,
+ "grad_norm": 0.02473079040646553,
+ "learning_rate": 0.0006,
+ "loss": 4.761087417602539,
+ "step": 3676
+ },
+ {
+ "epoch": 51.069868995633186,
+ "grad_norm": 0.020360209047794342,
+ "learning_rate": 0.0006,
+ "loss": 4.753884792327881,
+ "step": 3677
+ },
+ {
+ "epoch": 51.083842794759825,
+ "grad_norm": 0.02078923024237156,
+ "learning_rate": 0.0006,
+ "loss": 4.647130489349365,
+ "step": 3678
+ },
+ {
+ "epoch": 51.09781659388646,
+ "grad_norm": 0.019387200474739075,
+ "learning_rate": 0.0006,
+ "loss": 4.74135160446167,
+ "step": 3679
+ },
+ {
+ "epoch": 51.1117903930131,
+ "grad_norm": 0.01874949410557747,
+ "learning_rate": 0.0006,
+ "loss": 4.795329570770264,
+ "step": 3680
+ },
+ {
+ "epoch": 51.12576419213974,
+ "grad_norm": 0.019331837072968483,
+ "learning_rate": 0.0006,
+ "loss": 4.897967338562012,
+ "step": 3681
+ },
+ {
+ "epoch": 51.13973799126637,
+ "grad_norm": 0.017066050320863724,
+ "learning_rate": 0.0006,
+ "loss": 4.825351715087891,
+ "step": 3682
+ },
+ {
+ "epoch": 51.15371179039301,
+ "grad_norm": 0.01757810264825821,
+ "learning_rate": 0.0006,
+ "loss": 4.719335079193115,
+ "step": 3683
+ },
+ {
+ "epoch": 51.16768558951965,
+ "grad_norm": 0.018174266442656517,
+ "learning_rate": 0.0006,
+ "loss": 4.732680320739746,
+ "step": 3684
+ },
+ {
+ "epoch": 51.18165938864629,
+ "grad_norm": 0.018497727811336517,
+ "learning_rate": 0.0006,
+ "loss": 4.731050491333008,
+ "step": 3685
+ },
+ {
+ "epoch": 51.19563318777293,
+ "grad_norm": 0.0186338871717453,
+ "learning_rate": 0.0006,
+ "loss": 4.748035430908203,
+ "step": 3686
+ },
+ {
+ "epoch": 51.209606986899566,
+ "grad_norm": 0.018531711772084236,
+ "learning_rate": 0.0006,
+ "loss": 4.9007463455200195,
+ "step": 3687
+ },
+ {
+ "epoch": 51.223580786026204,
+ "grad_norm": 0.018221354112029076,
+ "learning_rate": 0.0006,
+ "loss": 4.84077262878418,
+ "step": 3688
+ },
+ {
+ "epoch": 51.237554585152836,
+ "grad_norm": 0.0203656405210495,
+ "learning_rate": 0.0006,
+ "loss": 4.73009729385376,
+ "step": 3689
+ },
+ {
+ "epoch": 51.251528384279474,
+ "grad_norm": 0.01964893378317356,
+ "learning_rate": 0.0006,
+ "loss": 4.772583961486816,
+ "step": 3690
+ },
+ {
+ "epoch": 51.26550218340611,
+ "grad_norm": 0.019606510177254677,
+ "learning_rate": 0.0006,
+ "loss": 4.599812030792236,
+ "step": 3691
+ },
+ {
+ "epoch": 51.27947598253275,
+ "grad_norm": 0.020501255989074707,
+ "learning_rate": 0.0006,
+ "loss": 4.809898853302002,
+ "step": 3692
+ },
+ {
+ "epoch": 51.29344978165939,
+ "grad_norm": 0.01870838552713394,
+ "learning_rate": 0.0006,
+ "loss": 4.741898059844971,
+ "step": 3693
+ },
+ {
+ "epoch": 51.30742358078603,
+ "grad_norm": 0.01802559196949005,
+ "learning_rate": 0.0006,
+ "loss": 4.746562480926514,
+ "step": 3694
+ },
+ {
+ "epoch": 51.32139737991266,
+ "grad_norm": 0.02053048275411129,
+ "learning_rate": 0.0006,
+ "loss": 4.69932746887207,
+ "step": 3695
+ },
+ {
+ "epoch": 51.3353711790393,
+ "grad_norm": 0.02661825157701969,
+ "learning_rate": 0.0006,
+ "loss": 4.68748664855957,
+ "step": 3696
+ },
+ {
+ "epoch": 51.34934497816594,
+ "grad_norm": 0.028269147500395775,
+ "learning_rate": 0.0006,
+ "loss": 4.731565475463867,
+ "step": 3697
+ },
+ {
+ "epoch": 51.36331877729258,
+ "grad_norm": 0.02650964818894863,
+ "learning_rate": 0.0006,
+ "loss": 4.733857154846191,
+ "step": 3698
+ },
+ {
+ "epoch": 51.377292576419215,
+ "grad_norm": 0.023696066811680794,
+ "learning_rate": 0.0006,
+ "loss": 4.812362194061279,
+ "step": 3699
+ },
+ {
+ "epoch": 51.391266375545854,
+ "grad_norm": 0.02365880273282528,
+ "learning_rate": 0.0006,
+ "loss": 4.733426094055176,
+ "step": 3700
+ },
+ {
+ "epoch": 51.40524017467249,
+ "grad_norm": 0.026249447837471962,
+ "learning_rate": 0.0006,
+ "loss": 4.715622901916504,
+ "step": 3701
+ },
+ {
+ "epoch": 51.419213973799124,
+ "grad_norm": 0.03301899880170822,
+ "learning_rate": 0.0006,
+ "loss": 4.7223663330078125,
+ "step": 3702
+ },
+ {
+ "epoch": 51.43318777292576,
+ "grad_norm": 0.044686660170555115,
+ "learning_rate": 0.0006,
+ "loss": 4.8121161460876465,
+ "step": 3703
+ },
+ {
+ "epoch": 51.4471615720524,
+ "grad_norm": 0.04733522981405258,
+ "learning_rate": 0.0006,
+ "loss": 4.841394424438477,
+ "step": 3704
+ },
+ {
+ "epoch": 51.46113537117904,
+ "grad_norm": 0.0395209975540638,
+ "learning_rate": 0.0006,
+ "loss": 4.773837089538574,
+ "step": 3705
+ },
+ {
+ "epoch": 51.47510917030568,
+ "grad_norm": 0.032790958881378174,
+ "learning_rate": 0.0006,
+ "loss": 4.833066463470459,
+ "step": 3706
+ },
+ {
+ "epoch": 51.48908296943232,
+ "grad_norm": 0.030578427016735077,
+ "learning_rate": 0.0006,
+ "loss": 4.685190200805664,
+ "step": 3707
+ },
+ {
+ "epoch": 51.50305676855895,
+ "grad_norm": 0.031619083136320114,
+ "learning_rate": 0.0006,
+ "loss": 4.740378379821777,
+ "step": 3708
+ },
+ {
+ "epoch": 51.51703056768559,
+ "grad_norm": 0.02385825850069523,
+ "learning_rate": 0.0006,
+ "loss": 4.812911510467529,
+ "step": 3709
+ },
+ {
+ "epoch": 51.531004366812226,
+ "grad_norm": 0.023523425683379173,
+ "learning_rate": 0.0006,
+ "loss": 4.669580459594727,
+ "step": 3710
+ },
+ {
+ "epoch": 51.544978165938865,
+ "grad_norm": 0.021316280588507652,
+ "learning_rate": 0.0006,
+ "loss": 4.76888370513916,
+ "step": 3711
+ },
+ {
+ "epoch": 51.5589519650655,
+ "grad_norm": 0.024663880467414856,
+ "learning_rate": 0.0006,
+ "loss": 4.702592372894287,
+ "step": 3712
+ },
+ {
+ "epoch": 51.57292576419214,
+ "grad_norm": 0.024173280224204063,
+ "learning_rate": 0.0006,
+ "loss": 4.710604667663574,
+ "step": 3713
+ },
+ {
+ "epoch": 51.58689956331878,
+ "grad_norm": 0.02754114381968975,
+ "learning_rate": 0.0006,
+ "loss": 4.746356964111328,
+ "step": 3714
+ },
+ {
+ "epoch": 51.60087336244541,
+ "grad_norm": 0.027903622016310692,
+ "learning_rate": 0.0006,
+ "loss": 4.7546491622924805,
+ "step": 3715
+ },
+ {
+ "epoch": 51.61484716157205,
+ "grad_norm": 0.024871617555618286,
+ "learning_rate": 0.0006,
+ "loss": 4.605746269226074,
+ "step": 3716
+ },
+ {
+ "epoch": 51.62882096069869,
+ "grad_norm": 0.021239031106233597,
+ "learning_rate": 0.0006,
+ "loss": 4.6942458152771,
+ "step": 3717
+ },
+ {
+ "epoch": 51.64279475982533,
+ "grad_norm": 0.01925622671842575,
+ "learning_rate": 0.0006,
+ "loss": 4.593558311462402,
+ "step": 3718
+ },
+ {
+ "epoch": 51.65676855895197,
+ "grad_norm": 0.01956283673644066,
+ "learning_rate": 0.0006,
+ "loss": 4.631172180175781,
+ "step": 3719
+ },
+ {
+ "epoch": 51.670742358078606,
+ "grad_norm": 0.02269871160387993,
+ "learning_rate": 0.0006,
+ "loss": 4.729236602783203,
+ "step": 3720
+ },
+ {
+ "epoch": 51.68471615720524,
+ "grad_norm": 0.02253916673362255,
+ "learning_rate": 0.0006,
+ "loss": 4.7191033363342285,
+ "step": 3721
+ },
+ {
+ "epoch": 51.698689956331876,
+ "grad_norm": 0.026231050491333008,
+ "learning_rate": 0.0006,
+ "loss": 4.811850547790527,
+ "step": 3722
+ },
+ {
+ "epoch": 51.712663755458514,
+ "grad_norm": 0.02558821812272072,
+ "learning_rate": 0.0006,
+ "loss": 4.681466102600098,
+ "step": 3723
+ },
+ {
+ "epoch": 51.72663755458515,
+ "grad_norm": 0.02124209702014923,
+ "learning_rate": 0.0006,
+ "loss": 4.693664073944092,
+ "step": 3724
+ },
+ {
+ "epoch": 51.74061135371179,
+ "grad_norm": 0.016638319939374924,
+ "learning_rate": 0.0006,
+ "loss": 4.771440029144287,
+ "step": 3725
+ },
+ {
+ "epoch": 51.75458515283843,
+ "grad_norm": 0.018048347905278206,
+ "learning_rate": 0.0006,
+ "loss": 4.643252372741699,
+ "step": 3726
+ },
+ {
+ "epoch": 51.76855895196506,
+ "grad_norm": 0.019904717803001404,
+ "learning_rate": 0.0006,
+ "loss": 4.66530704498291,
+ "step": 3727
+ },
+ {
+ "epoch": 51.7825327510917,
+ "grad_norm": 0.0194614976644516,
+ "learning_rate": 0.0006,
+ "loss": 4.711943626403809,
+ "step": 3728
+ },
+ {
+ "epoch": 51.79650655021834,
+ "grad_norm": 0.019175555557012558,
+ "learning_rate": 0.0006,
+ "loss": 4.721931457519531,
+ "step": 3729
+ },
+ {
+ "epoch": 51.81048034934498,
+ "grad_norm": 0.02093447372317314,
+ "learning_rate": 0.0006,
+ "loss": 4.633059024810791,
+ "step": 3730
+ },
+ {
+ "epoch": 51.82445414847162,
+ "grad_norm": 0.01930283196270466,
+ "learning_rate": 0.0006,
+ "loss": 4.659373760223389,
+ "step": 3731
+ },
+ {
+ "epoch": 51.838427947598255,
+ "grad_norm": 0.015350676141679287,
+ "learning_rate": 0.0006,
+ "loss": 4.763708114624023,
+ "step": 3732
+ },
+ {
+ "epoch": 51.852401746724894,
+ "grad_norm": 0.014139448292553425,
+ "learning_rate": 0.0006,
+ "loss": 4.753891468048096,
+ "step": 3733
+ },
+ {
+ "epoch": 51.866375545851525,
+ "grad_norm": 0.016215872019529343,
+ "learning_rate": 0.0006,
+ "loss": 4.658967018127441,
+ "step": 3734
+ },
+ {
+ "epoch": 51.880349344978164,
+ "grad_norm": 0.017328619956970215,
+ "learning_rate": 0.0006,
+ "loss": 4.696579933166504,
+ "step": 3735
+ },
+ {
+ "epoch": 51.8943231441048,
+ "grad_norm": 0.018103396520018578,
+ "learning_rate": 0.0006,
+ "loss": 4.627464771270752,
+ "step": 3736
+ },
+ {
+ "epoch": 51.90829694323144,
+ "grad_norm": 0.01715235412120819,
+ "learning_rate": 0.0006,
+ "loss": 4.7368011474609375,
+ "step": 3737
+ },
+ {
+ "epoch": 51.92227074235808,
+ "grad_norm": 0.014790094457566738,
+ "learning_rate": 0.0006,
+ "loss": 4.629441738128662,
+ "step": 3738
+ },
+ {
+ "epoch": 51.93624454148472,
+ "grad_norm": 0.017114296555519104,
+ "learning_rate": 0.0006,
+ "loss": 4.761256694793701,
+ "step": 3739
+ },
+ {
+ "epoch": 51.95021834061135,
+ "grad_norm": 0.019009845331311226,
+ "learning_rate": 0.0006,
+ "loss": 4.789155006408691,
+ "step": 3740
+ },
+ {
+ "epoch": 51.96419213973799,
+ "grad_norm": 0.019696692004799843,
+ "learning_rate": 0.0006,
+ "loss": 4.721269607543945,
+ "step": 3741
+ },
+ {
+ "epoch": 51.97816593886463,
+ "grad_norm": 0.020542822778224945,
+ "learning_rate": 0.0006,
+ "loss": 4.687038898468018,
+ "step": 3742
+ },
+ {
+ "epoch": 51.992139737991266,
+ "grad_norm": 0.021267160773277283,
+ "learning_rate": 0.0006,
+ "loss": 4.747860908508301,
+ "step": 3743
+ },
+ {
+ "epoch": 52.0,
+ "grad_norm": 0.025049567222595215,
+ "learning_rate": 0.0006,
+ "loss": 4.8473100662231445,
+ "step": 3744
+ },
+ {
+ "epoch": 52.0,
+ "eval_loss": 4.943043231964111,
+ "eval_runtime": 56.4929,
+ "eval_samples_per_second": 43.227,
+ "eval_steps_per_second": 1.363,
+ "step": 3744
+ },
+ {
+ "epoch": 52.01397379912664,
+ "grad_norm": 0.03054986707866192,
+ "learning_rate": 0.0006,
+ "loss": 4.856931686401367,
+ "step": 3745
+ },
+ {
+ "epoch": 52.02794759825328,
+ "grad_norm": 0.03838375583291054,
+ "learning_rate": 0.0006,
+ "loss": 4.699807167053223,
+ "step": 3746
+ },
+ {
+ "epoch": 52.041921397379916,
+ "grad_norm": 0.03383488208055496,
+ "learning_rate": 0.0006,
+ "loss": 4.6626787185668945,
+ "step": 3747
+ },
+ {
+ "epoch": 52.05589519650655,
+ "grad_norm": 0.0201561376452446,
+ "learning_rate": 0.0006,
+ "loss": 4.621831893920898,
+ "step": 3748
+ },
+ {
+ "epoch": 52.069868995633186,
+ "grad_norm": 0.028749438002705574,
+ "learning_rate": 0.0006,
+ "loss": 4.682540416717529,
+ "step": 3749
+ },
+ {
+ "epoch": 52.083842794759825,
+ "grad_norm": 0.02681174874305725,
+ "learning_rate": 0.0006,
+ "loss": 4.629997253417969,
+ "step": 3750
+ },
+ {
+ "epoch": 52.09781659388646,
+ "grad_norm": 0.024063318967819214,
+ "learning_rate": 0.0006,
+ "loss": 4.573021411895752,
+ "step": 3751
+ },
+ {
+ "epoch": 52.1117903930131,
+ "grad_norm": 0.02324330434203148,
+ "learning_rate": 0.0006,
+ "loss": 4.739465236663818,
+ "step": 3752
+ },
+ {
+ "epoch": 52.12576419213974,
+ "grad_norm": 0.0225426834076643,
+ "learning_rate": 0.0006,
+ "loss": 4.7335333824157715,
+ "step": 3753
+ },
+ {
+ "epoch": 52.13973799126637,
+ "grad_norm": 0.019593428820371628,
+ "learning_rate": 0.0006,
+ "loss": 4.704958438873291,
+ "step": 3754
+ },
+ {
+ "epoch": 52.15371179039301,
+ "grad_norm": 0.020777082070708275,
+ "learning_rate": 0.0006,
+ "loss": 4.638696670532227,
+ "step": 3755
+ },
+ {
+ "epoch": 52.16768558951965,
+ "grad_norm": 0.021525604650378227,
+ "learning_rate": 0.0006,
+ "loss": 4.643706798553467,
+ "step": 3756
+ },
+ {
+ "epoch": 52.18165938864629,
+ "grad_norm": 0.018943654373288155,
+ "learning_rate": 0.0006,
+ "loss": 4.5173211097717285,
+ "step": 3757
+ },
+ {
+ "epoch": 52.19563318777293,
+ "grad_norm": 0.01856580562889576,
+ "learning_rate": 0.0006,
+ "loss": 4.678788661956787,
+ "step": 3758
+ },
+ {
+ "epoch": 52.209606986899566,
+ "grad_norm": 0.019417259842157364,
+ "learning_rate": 0.0006,
+ "loss": 4.587244987487793,
+ "step": 3759
+ },
+ {
+ "epoch": 52.223580786026204,
+ "grad_norm": 0.01911911740899086,
+ "learning_rate": 0.0006,
+ "loss": 4.688503265380859,
+ "step": 3760
+ },
+ {
+ "epoch": 52.237554585152836,
+ "grad_norm": 0.021017879247665405,
+ "learning_rate": 0.0006,
+ "loss": 4.65829610824585,
+ "step": 3761
+ },
+ {
+ "epoch": 52.251528384279474,
+ "grad_norm": 0.02048875205218792,
+ "learning_rate": 0.0006,
+ "loss": 4.624055862426758,
+ "step": 3762
+ },
+ {
+ "epoch": 52.26550218340611,
+ "grad_norm": 0.01932772807776928,
+ "learning_rate": 0.0006,
+ "loss": 4.556562900543213,
+ "step": 3763
+ },
+ {
+ "epoch": 52.27947598253275,
+ "grad_norm": 0.01871614344418049,
+ "learning_rate": 0.0006,
+ "loss": 4.581201076507568,
+ "step": 3764
+ },
+ {
+ "epoch": 52.29344978165939,
+ "grad_norm": 0.017253106459975243,
+ "learning_rate": 0.0006,
+ "loss": 4.605219841003418,
+ "step": 3765
+ },
+ {
+ "epoch": 52.30742358078603,
+ "grad_norm": 0.01831238530576229,
+ "learning_rate": 0.0006,
+ "loss": 4.759900093078613,
+ "step": 3766
+ },
+ {
+ "epoch": 52.32139737991266,
+ "grad_norm": 0.017946207895874977,
+ "learning_rate": 0.0006,
+ "loss": 4.5598344802856445,
+ "step": 3767
+ },
+ {
+ "epoch": 52.3353711790393,
+ "grad_norm": 0.01680106669664383,
+ "learning_rate": 0.0006,
+ "loss": 4.605034828186035,
+ "step": 3768
+ },
+ {
+ "epoch": 52.34934497816594,
+ "grad_norm": 0.021482540294528008,
+ "learning_rate": 0.0006,
+ "loss": 4.620411396026611,
+ "step": 3769
+ },
+ {
+ "epoch": 52.36331877729258,
+ "grad_norm": 0.02210252732038498,
+ "learning_rate": 0.0006,
+ "loss": 4.667577743530273,
+ "step": 3770
+ },
+ {
+ "epoch": 52.377292576419215,
+ "grad_norm": 0.02237485907971859,
+ "learning_rate": 0.0006,
+ "loss": 4.628812789916992,
+ "step": 3771
+ },
+ {
+ "epoch": 52.391266375545854,
+ "grad_norm": 0.02196541801095009,
+ "learning_rate": 0.0006,
+ "loss": 4.572988510131836,
+ "step": 3772
+ },
+ {
+ "epoch": 52.40524017467249,
+ "grad_norm": 0.01930229924619198,
+ "learning_rate": 0.0006,
+ "loss": 4.588339805603027,
+ "step": 3773
+ },
+ {
+ "epoch": 52.419213973799124,
+ "grad_norm": 0.02222434990108013,
+ "learning_rate": 0.0006,
+ "loss": 4.766482353210449,
+ "step": 3774
+ },
+ {
+ "epoch": 52.43318777292576,
+ "grad_norm": 0.028806306421756744,
+ "learning_rate": 0.0006,
+ "loss": 4.617892742156982,
+ "step": 3775
+ },
+ {
+ "epoch": 52.4471615720524,
+ "grad_norm": 0.03306042402982712,
+ "learning_rate": 0.0006,
+ "loss": 4.665334701538086,
+ "step": 3776
+ },
+ {
+ "epoch": 52.46113537117904,
+ "grad_norm": 0.0358956903219223,
+ "learning_rate": 0.0006,
+ "loss": 4.566668510437012,
+ "step": 3777
+ },
+ {
+ "epoch": 52.47510917030568,
+ "grad_norm": 0.030406497418880463,
+ "learning_rate": 0.0006,
+ "loss": 4.513612270355225,
+ "step": 3778
+ },
+ {
+ "epoch": 52.48908296943232,
+ "grad_norm": 0.030295712873339653,
+ "learning_rate": 0.0006,
+ "loss": 4.640189170837402,
+ "step": 3779
+ },
+ {
+ "epoch": 52.50305676855895,
+ "grad_norm": 0.024981003254652023,
+ "learning_rate": 0.0006,
+ "loss": 4.57558012008667,
+ "step": 3780
+ },
+ {
+ "epoch": 52.51703056768559,
+ "grad_norm": 0.024990713223814964,
+ "learning_rate": 0.0006,
+ "loss": 4.708589553833008,
+ "step": 3781
+ },
+ {
+ "epoch": 52.531004366812226,
+ "grad_norm": 0.025034775957465172,
+ "learning_rate": 0.0006,
+ "loss": 4.70468282699585,
+ "step": 3782
+ },
+ {
+ "epoch": 52.544978165938865,
+ "grad_norm": 0.028243759647011757,
+ "learning_rate": 0.0006,
+ "loss": 4.712060451507568,
+ "step": 3783
+ },
+ {
+ "epoch": 52.5589519650655,
+ "grad_norm": 0.026984544470906258,
+ "learning_rate": 0.0006,
+ "loss": 4.505004405975342,
+ "step": 3784
+ },
+ {
+ "epoch": 52.57292576419214,
+ "grad_norm": 0.050843529403209686,
+ "learning_rate": 0.0006,
+ "loss": 4.593271255493164,
+ "step": 3785
+ },
+ {
+ "epoch": 52.58689956331878,
+ "grad_norm": 0.25422802567481995,
+ "learning_rate": 0.0006,
+ "loss": 4.7178802490234375,
+ "step": 3786
+ },
+ {
+ "epoch": 52.60087336244541,
+ "grad_norm": 0.6078406572341919,
+ "learning_rate": 0.0006,
+ "loss": 6.144372940063477,
+ "step": 3787
+ },
+ {
+ "epoch": 52.61484716157205,
+ "grad_norm": 0.25933101773262024,
+ "learning_rate": 0.0006,
+ "loss": 5.971083641052246,
+ "step": 3788
+ },
+ {
+ "epoch": 52.62882096069869,
+ "grad_norm": 0.294657826423645,
+ "learning_rate": 0.0006,
+ "loss": 6.204090118408203,
+ "step": 3789
+ },
+ {
+ "epoch": 52.64279475982533,
+ "grad_norm": 0.23641522228717804,
+ "learning_rate": 0.0006,
+ "loss": 6.869574546813965,
+ "step": 3790
+ },
+ {
+ "epoch": 52.65676855895197,
+ "grad_norm": 0.23717311024665833,
+ "learning_rate": 0.0006,
+ "loss": 6.739034175872803,
+ "step": 3791
+ },
+ {
+ "epoch": 52.670742358078606,
+ "grad_norm": 0.10300362855195999,
+ "learning_rate": 0.0006,
+ "loss": 6.49191951751709,
+ "step": 3792
+ },
+ {
+ "epoch": 52.68471615720524,
+ "grad_norm": 0.1484968513250351,
+ "learning_rate": 0.0006,
+ "loss": 6.6021270751953125,
+ "step": 3793
+ },
+ {
+ "epoch": 52.698689956331876,
+ "grad_norm": 0.08052115142345428,
+ "learning_rate": 0.0006,
+ "loss": 6.432796478271484,
+ "step": 3794
+ },
+ {
+ "epoch": 52.712663755458514,
+ "grad_norm": 0.06869253516197205,
+ "learning_rate": 0.0006,
+ "loss": 6.212784767150879,
+ "step": 3795
+ },
+ {
+ "epoch": 52.72663755458515,
+ "grad_norm": 0.05589582398533821,
+ "learning_rate": 0.0006,
+ "loss": 6.168288707733154,
+ "step": 3796
+ },
+ {
+ "epoch": 52.74061135371179,
+ "grad_norm": 0.04840940609574318,
+ "learning_rate": 0.0006,
+ "loss": 6.163174629211426,
+ "step": 3797
+ },
+ {
+ "epoch": 52.75458515283843,
+ "grad_norm": 0.04578601196408272,
+ "learning_rate": 0.0006,
+ "loss": 6.038390636444092,
+ "step": 3798
+ },
+ {
+ "epoch": 52.76855895196506,
+ "grad_norm": 0.04102053865790367,
+ "learning_rate": 0.0006,
+ "loss": 5.929539680480957,
+ "step": 3799
+ },
+ {
+ "epoch": 52.7825327510917,
+ "grad_norm": 0.038398485630750656,
+ "learning_rate": 0.0006,
+ "loss": 6.007378101348877,
+ "step": 3800
+ },
+ {
+ "epoch": 52.79650655021834,
+ "grad_norm": 0.03466642275452614,
+ "learning_rate": 0.0006,
+ "loss": 5.767291069030762,
+ "step": 3801
+ },
+ {
+ "epoch": 52.81048034934498,
+ "grad_norm": 0.03328905999660492,
+ "learning_rate": 0.0006,
+ "loss": 5.785032272338867,
+ "step": 3802
+ },
+ {
+ "epoch": 52.82445414847162,
+ "grad_norm": 0.03186681121587753,
+ "learning_rate": 0.0006,
+ "loss": 5.6942057609558105,
+ "step": 3803
+ },
+ {
+ "epoch": 52.838427947598255,
+ "grad_norm": 0.030553551390767097,
+ "learning_rate": 0.0006,
+ "loss": 5.680774688720703,
+ "step": 3804
+ },
+ {
+ "epoch": 52.852401746724894,
+ "grad_norm": 0.02578839473426342,
+ "learning_rate": 0.0006,
+ "loss": 5.703461647033691,
+ "step": 3805
+ },
+ {
+ "epoch": 52.866375545851525,
+ "grad_norm": 0.026261312887072563,
+ "learning_rate": 0.0006,
+ "loss": 5.642549514770508,
+ "step": 3806
+ },
+ {
+ "epoch": 52.880349344978164,
+ "grad_norm": 0.02782430872321129,
+ "learning_rate": 0.0006,
+ "loss": 5.563597679138184,
+ "step": 3807
+ },
+ {
+ "epoch": 52.8943231441048,
+ "grad_norm": 0.024467799812555313,
+ "learning_rate": 0.0006,
+ "loss": 5.583395957946777,
+ "step": 3808
+ },
+ {
+ "epoch": 52.90829694323144,
+ "grad_norm": 0.022602897137403488,
+ "learning_rate": 0.0006,
+ "loss": 5.50742769241333,
+ "step": 3809
+ },
+ {
+ "epoch": 52.92227074235808,
+ "grad_norm": 0.024292029440402985,
+ "learning_rate": 0.0006,
+ "loss": 5.367804527282715,
+ "step": 3810
+ },
+ {
+ "epoch": 52.93624454148472,
+ "grad_norm": 0.021062295883893967,
+ "learning_rate": 0.0006,
+ "loss": 5.423705101013184,
+ "step": 3811
+ },
+ {
+ "epoch": 52.95021834061135,
+ "grad_norm": 0.022226469591259956,
+ "learning_rate": 0.0006,
+ "loss": 5.344039440155029,
+ "step": 3812
+ },
+ {
+ "epoch": 52.96419213973799,
+ "grad_norm": 0.023691752925515175,
+ "learning_rate": 0.0006,
+ "loss": 5.2581892013549805,
+ "step": 3813
+ },
+ {
+ "epoch": 52.97816593886463,
+ "grad_norm": 0.02181999757885933,
+ "learning_rate": 0.0006,
+ "loss": 5.2710113525390625,
+ "step": 3814
+ },
+ {
+ "epoch": 52.992139737991266,
+ "grad_norm": 0.025295857340097427,
+ "learning_rate": 0.0006,
+ "loss": 5.278355598449707,
+ "step": 3815
+ },
+ {
+ "epoch": 53.0,
+ "grad_norm": 0.02216828241944313,
+ "learning_rate": 0.0006,
+ "loss": 5.127945899963379,
+ "step": 3816
+ },
+ {
+ "epoch": 53.0,
+ "eval_loss": 5.3411736488342285,
+ "eval_runtime": 56.2837,
+ "eval_samples_per_second": 43.387,
+ "eval_steps_per_second": 1.368,
+ "step": 3816
+ },
+ {
+ "epoch": 53.01397379912664,
+ "grad_norm": 0.022496026009321213,
+ "learning_rate": 0.0006,
+ "loss": 5.19655704498291,
+ "step": 3817
+ },
+ {
+ "epoch": 53.02794759825328,
+ "grad_norm": 0.03201168775558472,
+ "learning_rate": 0.0006,
+ "loss": 5.094621181488037,
+ "step": 3818
+ },
+ {
+ "epoch": 53.041921397379916,
+ "grad_norm": 0.02683178335428238,
+ "learning_rate": 0.0006,
+ "loss": 5.19899845123291,
+ "step": 3819
+ },
+ {
+ "epoch": 53.05589519650655,
+ "grad_norm": 0.022659169510006905,
+ "learning_rate": 0.0006,
+ "loss": 5.184103012084961,
+ "step": 3820
+ },
+ {
+ "epoch": 53.069868995633186,
+ "grad_norm": 0.024784443899989128,
+ "learning_rate": 0.0006,
+ "loss": 5.124286651611328,
+ "step": 3821
+ },
+ {
+ "epoch": 53.083842794759825,
+ "grad_norm": 0.028710370883345604,
+ "learning_rate": 0.0006,
+ "loss": 5.130426406860352,
+ "step": 3822
+ },
+ {
+ "epoch": 53.09781659388646,
+ "grad_norm": 0.02366088517010212,
+ "learning_rate": 0.0006,
+ "loss": 5.1453399658203125,
+ "step": 3823
+ },
+ {
+ "epoch": 53.1117903930131,
+ "grad_norm": 0.01813938096165657,
+ "learning_rate": 0.0006,
+ "loss": 5.05271053314209,
+ "step": 3824
+ },
+ {
+ "epoch": 53.12576419213974,
+ "grad_norm": 0.02051435597240925,
+ "learning_rate": 0.0006,
+ "loss": 5.023038864135742,
+ "step": 3825
+ },
+ {
+ "epoch": 53.13973799126637,
+ "grad_norm": 0.02185754105448723,
+ "learning_rate": 0.0006,
+ "loss": 4.8950018882751465,
+ "step": 3826
+ },
+ {
+ "epoch": 53.15371179039301,
+ "grad_norm": 0.017223743721842766,
+ "learning_rate": 0.0006,
+ "loss": 5.078621864318848,
+ "step": 3827
+ },
+ {
+ "epoch": 53.16768558951965,
+ "grad_norm": 0.01832217164337635,
+ "learning_rate": 0.0006,
+ "loss": 4.921341419219971,
+ "step": 3828
+ },
+ {
+ "epoch": 53.18165938864629,
+ "grad_norm": 0.02114005573093891,
+ "learning_rate": 0.0006,
+ "loss": 4.997865676879883,
+ "step": 3829
+ },
+ {
+ "epoch": 53.19563318777293,
+ "grad_norm": 0.021426431834697723,
+ "learning_rate": 0.0006,
+ "loss": 4.877785682678223,
+ "step": 3830
+ },
+ {
+ "epoch": 53.209606986899566,
+ "grad_norm": 0.021327560767531395,
+ "learning_rate": 0.0006,
+ "loss": 4.847512245178223,
+ "step": 3831
+ },
+ {
+ "epoch": 53.223580786026204,
+ "grad_norm": 0.027825530618429184,
+ "learning_rate": 0.0006,
+ "loss": 5.004823207855225,
+ "step": 3832
+ },
+ {
+ "epoch": 53.237554585152836,
+ "grad_norm": 0.03756481409072876,
+ "learning_rate": 0.0006,
+ "loss": 4.817290306091309,
+ "step": 3833
+ },
+ {
+ "epoch": 53.251528384279474,
+ "grad_norm": 0.04933721572160721,
+ "learning_rate": 0.0006,
+ "loss": 4.863317489624023,
+ "step": 3834
+ },
+ {
+ "epoch": 53.26550218340611,
+ "grad_norm": 0.04588694870471954,
+ "learning_rate": 0.0006,
+ "loss": 4.864742755889893,
+ "step": 3835
+ },
+ {
+ "epoch": 53.27947598253275,
+ "grad_norm": 0.02728264033794403,
+ "learning_rate": 0.0006,
+ "loss": 4.9068756103515625,
+ "step": 3836
+ },
+ {
+ "epoch": 53.29344978165939,
+ "grad_norm": 0.036240726709365845,
+ "learning_rate": 0.0006,
+ "loss": 4.902045249938965,
+ "step": 3837
+ },
+ {
+ "epoch": 53.30742358078603,
+ "grad_norm": 0.02703195810317993,
+ "learning_rate": 0.0006,
+ "loss": 4.908980369567871,
+ "step": 3838
+ },
+ {
+ "epoch": 53.32139737991266,
+ "grad_norm": 0.024894049391150475,
+ "learning_rate": 0.0006,
+ "loss": 4.794848442077637,
+ "step": 3839
+ },
+ {
+ "epoch": 53.3353711790393,
+ "grad_norm": 0.02389993891119957,
+ "learning_rate": 0.0006,
+ "loss": 4.8234734535217285,
+ "step": 3840
+ },
+ {
+ "epoch": 53.34934497816594,
+ "grad_norm": 0.02486400306224823,
+ "learning_rate": 0.0006,
+ "loss": 4.851463794708252,
+ "step": 3841
+ },
+ {
+ "epoch": 53.36331877729258,
+ "grad_norm": 0.02613813802599907,
+ "learning_rate": 0.0006,
+ "loss": 4.860444068908691,
+ "step": 3842
+ },
+ {
+ "epoch": 53.377292576419215,
+ "grad_norm": 0.0229744054377079,
+ "learning_rate": 0.0006,
+ "loss": 4.709606170654297,
+ "step": 3843
+ },
+ {
+ "epoch": 53.391266375545854,
+ "grad_norm": 0.02123042568564415,
+ "learning_rate": 0.0006,
+ "loss": 4.654613494873047,
+ "step": 3844
+ },
+ {
+ "epoch": 53.40524017467249,
+ "grad_norm": 0.017352338880300522,
+ "learning_rate": 0.0006,
+ "loss": 4.769761085510254,
+ "step": 3845
+ },
+ {
+ "epoch": 53.419213973799124,
+ "grad_norm": 0.019808724522590637,
+ "learning_rate": 0.0006,
+ "loss": 4.98514461517334,
+ "step": 3846
+ },
+ {
+ "epoch": 53.43318777292576,
+ "grad_norm": 0.018776152282953262,
+ "learning_rate": 0.0006,
+ "loss": 4.749990463256836,
+ "step": 3847
+ },
+ {
+ "epoch": 53.4471615720524,
+ "grad_norm": 0.02017521858215332,
+ "learning_rate": 0.0006,
+ "loss": 4.757381439208984,
+ "step": 3848
+ },
+ {
+ "epoch": 53.46113537117904,
+ "grad_norm": 0.019509416073560715,
+ "learning_rate": 0.0006,
+ "loss": 4.8111572265625,
+ "step": 3849
+ },
+ {
+ "epoch": 53.47510917030568,
+ "grad_norm": 0.015543580986559391,
+ "learning_rate": 0.0006,
+ "loss": 4.7501678466796875,
+ "step": 3850
+ },
+ {
+ "epoch": 53.48908296943232,
+ "grad_norm": 0.017941651865839958,
+ "learning_rate": 0.0006,
+ "loss": 4.701838493347168,
+ "step": 3851
+ },
+ {
+ "epoch": 53.50305676855895,
+ "grad_norm": 0.01678212359547615,
+ "learning_rate": 0.0006,
+ "loss": 4.787798881530762,
+ "step": 3852
+ },
+ {
+ "epoch": 53.51703056768559,
+ "grad_norm": 0.019050654023885727,
+ "learning_rate": 0.0006,
+ "loss": 4.754859924316406,
+ "step": 3853
+ },
+ {
+ "epoch": 53.531004366812226,
+ "grad_norm": 0.017406078055500984,
+ "learning_rate": 0.0006,
+ "loss": 4.752772808074951,
+ "step": 3854
+ },
+ {
+ "epoch": 53.544978165938865,
+ "grad_norm": 0.01773344725370407,
+ "learning_rate": 0.0006,
+ "loss": 4.651812553405762,
+ "step": 3855
+ },
+ {
+ "epoch": 53.5589519650655,
+ "grad_norm": 0.017221346497535706,
+ "learning_rate": 0.0006,
+ "loss": 4.792244911193848,
+ "step": 3856
+ },
+ {
+ "epoch": 53.57292576419214,
+ "grad_norm": 0.01722758449614048,
+ "learning_rate": 0.0006,
+ "loss": 4.7902960777282715,
+ "step": 3857
+ },
+ {
+ "epoch": 53.58689956331878,
+ "grad_norm": 0.020445192232728004,
+ "learning_rate": 0.0006,
+ "loss": 4.67392110824585,
+ "step": 3858
+ },
+ {
+ "epoch": 53.60087336244541,
+ "grad_norm": 0.02267114631831646,
+ "learning_rate": 0.0006,
+ "loss": 4.796089172363281,
+ "step": 3859
+ },
+ {
+ "epoch": 53.61484716157205,
+ "grad_norm": 0.027129942551255226,
+ "learning_rate": 0.0006,
+ "loss": 4.774303913116455,
+ "step": 3860
+ },
+ {
+ "epoch": 53.62882096069869,
+ "grad_norm": 0.025454839691519737,
+ "learning_rate": 0.0006,
+ "loss": 4.708813190460205,
+ "step": 3861
+ },
+ {
+ "epoch": 53.64279475982533,
+ "grad_norm": 0.019248757511377335,
+ "learning_rate": 0.0006,
+ "loss": 4.760467529296875,
+ "step": 3862
+ },
+ {
+ "epoch": 53.65676855895197,
+ "grad_norm": 0.01791176199913025,
+ "learning_rate": 0.0006,
+ "loss": 4.789203643798828,
+ "step": 3863
+ },
+ {
+ "epoch": 53.670742358078606,
+ "grad_norm": 0.016306612640619278,
+ "learning_rate": 0.0006,
+ "loss": 4.732605934143066,
+ "step": 3864
+ },
+ {
+ "epoch": 53.68471615720524,
+ "grad_norm": 0.017836976796388626,
+ "learning_rate": 0.0006,
+ "loss": 4.691871643066406,
+ "step": 3865
+ },
+ {
+ "epoch": 53.698689956331876,
+ "grad_norm": 0.015102861449122429,
+ "learning_rate": 0.0006,
+ "loss": 4.724156856536865,
+ "step": 3866
+ },
+ {
+ "epoch": 53.712663755458514,
+ "grad_norm": 0.017458142712712288,
+ "learning_rate": 0.0006,
+ "loss": 4.73524808883667,
+ "step": 3867
+ },
+ {
+ "epoch": 53.72663755458515,
+ "grad_norm": 0.01591932773590088,
+ "learning_rate": 0.0006,
+ "loss": 4.705888748168945,
+ "step": 3868
+ },
+ {
+ "epoch": 53.74061135371179,
+ "grad_norm": 0.01713593676686287,
+ "learning_rate": 0.0006,
+ "loss": 4.644516944885254,
+ "step": 3869
+ },
+ {
+ "epoch": 53.75458515283843,
+ "grad_norm": 0.01905885338783264,
+ "learning_rate": 0.0006,
+ "loss": 4.698416233062744,
+ "step": 3870
+ },
+ {
+ "epoch": 53.76855895196506,
+ "grad_norm": 0.01766437292098999,
+ "learning_rate": 0.0006,
+ "loss": 4.6796555519104,
+ "step": 3871
+ },
+ {
+ "epoch": 53.7825327510917,
+ "grad_norm": 0.01552243810147047,
+ "learning_rate": 0.0006,
+ "loss": 4.78138542175293,
+ "step": 3872
+ },
+ {
+ "epoch": 53.79650655021834,
+ "grad_norm": 0.014191331341862679,
+ "learning_rate": 0.0006,
+ "loss": 4.703588485717773,
+ "step": 3873
+ },
+ {
+ "epoch": 53.81048034934498,
+ "grad_norm": 0.015872279182076454,
+ "learning_rate": 0.0006,
+ "loss": 4.705611228942871,
+ "step": 3874
+ },
+ {
+ "epoch": 53.82445414847162,
+ "grad_norm": 0.01787560060620308,
+ "learning_rate": 0.0006,
+ "loss": 4.8227949142456055,
+ "step": 3875
+ },
+ {
+ "epoch": 53.838427947598255,
+ "grad_norm": 0.0158209390938282,
+ "learning_rate": 0.0006,
+ "loss": 4.645227432250977,
+ "step": 3876
+ },
+ {
+ "epoch": 53.852401746724894,
+ "grad_norm": 0.014362678863108158,
+ "learning_rate": 0.0006,
+ "loss": 4.7225422859191895,
+ "step": 3877
+ },
+ {
+ "epoch": 53.866375545851525,
+ "grad_norm": 0.014899161644279957,
+ "learning_rate": 0.0006,
+ "loss": 4.685914039611816,
+ "step": 3878
+ },
+ {
+ "epoch": 53.880349344978164,
+ "grad_norm": 0.01628505438566208,
+ "learning_rate": 0.0006,
+ "loss": 4.674065589904785,
+ "step": 3879
+ },
+ {
+ "epoch": 53.8943231441048,
+ "grad_norm": 0.01873331144452095,
+ "learning_rate": 0.0006,
+ "loss": 4.811744213104248,
+ "step": 3880
+ },
+ {
+ "epoch": 53.90829694323144,
+ "grad_norm": 0.02229069173336029,
+ "learning_rate": 0.0006,
+ "loss": 4.721514701843262,
+ "step": 3881
+ },
+ {
+ "epoch": 53.92227074235808,
+ "grad_norm": 0.028012916445732117,
+ "learning_rate": 0.0006,
+ "loss": 4.773603439331055,
+ "step": 3882
+ },
+ {
+ "epoch": 53.93624454148472,
+ "grad_norm": 0.036322783678770065,
+ "learning_rate": 0.0006,
+ "loss": 4.787134170532227,
+ "step": 3883
+ },
+ {
+ "epoch": 53.95021834061135,
+ "grad_norm": 0.048149142414331436,
+ "learning_rate": 0.0006,
+ "loss": 4.736196517944336,
+ "step": 3884
+ },
+ {
+ "epoch": 53.96419213973799,
+ "grad_norm": 0.051918916404247284,
+ "learning_rate": 0.0006,
+ "loss": 4.686728000640869,
+ "step": 3885
+ },
+ {
+ "epoch": 53.97816593886463,
+ "grad_norm": 0.0481342189013958,
+ "learning_rate": 0.0006,
+ "loss": 4.733730792999268,
+ "step": 3886
+ },
+ {
+ "epoch": 53.992139737991266,
+ "grad_norm": 0.045424818992614746,
+ "learning_rate": 0.0006,
+ "loss": 4.748723983764648,
+ "step": 3887
+ },
+ {
+ "epoch": 54.0,
+ "grad_norm": 0.04153316840529442,
+ "learning_rate": 0.0006,
+ "loss": 4.715220928192139,
+ "step": 3888
+ },
+ {
+ "epoch": 54.0,
+ "eval_loss": 4.964215278625488,
+ "eval_runtime": 56.7757,
+ "eval_samples_per_second": 43.011,
+ "eval_steps_per_second": 1.356,
+ "step": 3888
+ },
+ {
+ "epoch": 54.01397379912664,
+ "grad_norm": 0.03459750860929489,
+ "learning_rate": 0.0006,
+ "loss": 4.661557674407959,
+ "step": 3889
+ },
+ {
+ "epoch": 54.02794759825328,
+ "grad_norm": 0.025409208610653877,
+ "learning_rate": 0.0006,
+ "loss": 4.7065582275390625,
+ "step": 3890
+ },
+ {
+ "epoch": 54.041921397379916,
+ "grad_norm": 0.03202550485730171,
+ "learning_rate": 0.0006,
+ "loss": 4.751070976257324,
+ "step": 3891
+ },
+ {
+ "epoch": 54.05589519650655,
+ "grad_norm": 0.02968018501996994,
+ "learning_rate": 0.0006,
+ "loss": 4.635388374328613,
+ "step": 3892
+ },
+ {
+ "epoch": 54.069868995633186,
+ "grad_norm": 0.02757406234741211,
+ "learning_rate": 0.0006,
+ "loss": 4.706214427947998,
+ "step": 3893
+ },
+ {
+ "epoch": 54.083842794759825,
+ "grad_norm": 0.025223620235919952,
+ "learning_rate": 0.0006,
+ "loss": 4.703792572021484,
+ "step": 3894
+ },
+ {
+ "epoch": 54.09781659388646,
+ "grad_norm": 0.02431778982281685,
+ "learning_rate": 0.0006,
+ "loss": 4.658116340637207,
+ "step": 3895
+ },
+ {
+ "epoch": 54.1117903930131,
+ "grad_norm": 0.02792172320187092,
+ "learning_rate": 0.0006,
+ "loss": 4.638895034790039,
+ "step": 3896
+ },
+ {
+ "epoch": 54.12576419213974,
+ "grad_norm": 0.032253511250019073,
+ "learning_rate": 0.0006,
+ "loss": 4.693902492523193,
+ "step": 3897
+ },
+ {
+ "epoch": 54.13973799126637,
+ "grad_norm": 0.027044931426644325,
+ "learning_rate": 0.0006,
+ "loss": 4.575805187225342,
+ "step": 3898
+ },
+ {
+ "epoch": 54.15371179039301,
+ "grad_norm": 0.023270361125469208,
+ "learning_rate": 0.0006,
+ "loss": 4.67645263671875,
+ "step": 3899
+ },
+ {
+ "epoch": 54.16768558951965,
+ "grad_norm": 0.01929079182446003,
+ "learning_rate": 0.0006,
+ "loss": 4.68062686920166,
+ "step": 3900
+ },
+ {
+ "epoch": 54.18165938864629,
+ "grad_norm": 0.020343473181128502,
+ "learning_rate": 0.0006,
+ "loss": 4.708086967468262,
+ "step": 3901
+ },
+ {
+ "epoch": 54.19563318777293,
+ "grad_norm": 0.0211793202906847,
+ "learning_rate": 0.0006,
+ "loss": 4.677323341369629,
+ "step": 3902
+ },
+ {
+ "epoch": 54.209606986899566,
+ "grad_norm": 0.020341217517852783,
+ "learning_rate": 0.0006,
+ "loss": 4.760937690734863,
+ "step": 3903
+ },
+ {
+ "epoch": 54.223580786026204,
+ "grad_norm": 0.019421054050326347,
+ "learning_rate": 0.0006,
+ "loss": 4.508834362030029,
+ "step": 3904
+ },
+ {
+ "epoch": 54.237554585152836,
+ "grad_norm": 0.016701679676771164,
+ "learning_rate": 0.0006,
+ "loss": 4.664399147033691,
+ "step": 3905
+ },
+ {
+ "epoch": 54.251528384279474,
+ "grad_norm": 0.016561470925807953,
+ "learning_rate": 0.0006,
+ "loss": 4.742990493774414,
+ "step": 3906
+ },
+ {
+ "epoch": 54.26550218340611,
+ "grad_norm": 0.017446085810661316,
+ "learning_rate": 0.0006,
+ "loss": 4.63038444519043,
+ "step": 3907
+ },
+ {
+ "epoch": 54.27947598253275,
+ "grad_norm": 0.014340086840093136,
+ "learning_rate": 0.0006,
+ "loss": 4.688791275024414,
+ "step": 3908
+ },
+ {
+ "epoch": 54.29344978165939,
+ "grad_norm": 0.013580698519945145,
+ "learning_rate": 0.0006,
+ "loss": 4.625302314758301,
+ "step": 3909
+ },
+ {
+ "epoch": 54.30742358078603,
+ "grad_norm": 0.014931539073586464,
+ "learning_rate": 0.0006,
+ "loss": 4.6167497634887695,
+ "step": 3910
+ },
+ {
+ "epoch": 54.32139737991266,
+ "grad_norm": 0.013492150232195854,
+ "learning_rate": 0.0006,
+ "loss": 4.657521724700928,
+ "step": 3911
+ },
+ {
+ "epoch": 54.3353711790393,
+ "grad_norm": 0.015151413157582283,
+ "learning_rate": 0.0006,
+ "loss": 4.634521484375,
+ "step": 3912
+ },
+ {
+ "epoch": 54.34934497816594,
+ "grad_norm": 0.01302040833979845,
+ "learning_rate": 0.0006,
+ "loss": 4.718928337097168,
+ "step": 3913
+ },
+ {
+ "epoch": 54.36331877729258,
+ "grad_norm": 0.013433529995381832,
+ "learning_rate": 0.0006,
+ "loss": 4.584803581237793,
+ "step": 3914
+ },
+ {
+ "epoch": 54.377292576419215,
+ "grad_norm": 0.012987039051949978,
+ "learning_rate": 0.0006,
+ "loss": 4.6393585205078125,
+ "step": 3915
+ },
+ {
+ "epoch": 54.391266375545854,
+ "grad_norm": 0.013695075176656246,
+ "learning_rate": 0.0006,
+ "loss": 4.556313514709473,
+ "step": 3916
+ },
+ {
+ "epoch": 54.40524017467249,
+ "grad_norm": 0.013718255795538425,
+ "learning_rate": 0.0006,
+ "loss": 4.576223850250244,
+ "step": 3917
+ },
+ {
+ "epoch": 54.419213973799124,
+ "grad_norm": 0.013699112460017204,
+ "learning_rate": 0.0006,
+ "loss": 4.63889217376709,
+ "step": 3918
+ },
+ {
+ "epoch": 54.43318777292576,
+ "grad_norm": 0.013516264036297798,
+ "learning_rate": 0.0006,
+ "loss": 4.669532299041748,
+ "step": 3919
+ },
+ {
+ "epoch": 54.4471615720524,
+ "grad_norm": 0.013597294688224792,
+ "learning_rate": 0.0006,
+ "loss": 4.6644287109375,
+ "step": 3920
+ },
+ {
+ "epoch": 54.46113537117904,
+ "grad_norm": 0.013185849413275719,
+ "learning_rate": 0.0006,
+ "loss": 4.636900901794434,
+ "step": 3921
+ },
+ {
+ "epoch": 54.47510917030568,
+ "grad_norm": 0.014311990700662136,
+ "learning_rate": 0.0006,
+ "loss": 4.6047587394714355,
+ "step": 3922
+ },
+ {
+ "epoch": 54.48908296943232,
+ "grad_norm": 0.012426027096807957,
+ "learning_rate": 0.0006,
+ "loss": 4.6161394119262695,
+ "step": 3923
+ },
+ {
+ "epoch": 54.50305676855895,
+ "grad_norm": 0.01283493172377348,
+ "learning_rate": 0.0006,
+ "loss": 4.584190368652344,
+ "step": 3924
+ },
+ {
+ "epoch": 54.51703056768559,
+ "grad_norm": 0.01339943427592516,
+ "learning_rate": 0.0006,
+ "loss": 4.624184608459473,
+ "step": 3925
+ },
+ {
+ "epoch": 54.531004366812226,
+ "grad_norm": 0.012859724462032318,
+ "learning_rate": 0.0006,
+ "loss": 4.6174635887146,
+ "step": 3926
+ },
+ {
+ "epoch": 54.544978165938865,
+ "grad_norm": 0.014420023187994957,
+ "learning_rate": 0.0006,
+ "loss": 4.536544322967529,
+ "step": 3927
+ },
+ {
+ "epoch": 54.5589519650655,
+ "grad_norm": 0.014237185008823872,
+ "learning_rate": 0.0006,
+ "loss": 4.681514739990234,
+ "step": 3928
+ },
+ {
+ "epoch": 54.57292576419214,
+ "grad_norm": 0.013340328820049763,
+ "learning_rate": 0.0006,
+ "loss": 4.579689979553223,
+ "step": 3929
+ },
+ {
+ "epoch": 54.58689956331878,
+ "grad_norm": 0.013894530944526196,
+ "learning_rate": 0.0006,
+ "loss": 4.638958930969238,
+ "step": 3930
+ },
+ {
+ "epoch": 54.60087336244541,
+ "grad_norm": 0.016181141138076782,
+ "learning_rate": 0.0006,
+ "loss": 4.573508262634277,
+ "step": 3931
+ },
+ {
+ "epoch": 54.61484716157205,
+ "grad_norm": 0.013221992179751396,
+ "learning_rate": 0.0006,
+ "loss": 4.6229047775268555,
+ "step": 3932
+ },
+ {
+ "epoch": 54.62882096069869,
+ "grad_norm": 0.013241561129689217,
+ "learning_rate": 0.0006,
+ "loss": 4.641152858734131,
+ "step": 3933
+ },
+ {
+ "epoch": 54.64279475982533,
+ "grad_norm": 0.016237538307905197,
+ "learning_rate": 0.0006,
+ "loss": 4.6421217918396,
+ "step": 3934
+ },
+ {
+ "epoch": 54.65676855895197,
+ "grad_norm": 0.020866507664322853,
+ "learning_rate": 0.0006,
+ "loss": 4.646058082580566,
+ "step": 3935
+ },
+ {
+ "epoch": 54.670742358078606,
+ "grad_norm": 0.022182893007993698,
+ "learning_rate": 0.0006,
+ "loss": 4.554501533508301,
+ "step": 3936
+ },
+ {
+ "epoch": 54.68471615720524,
+ "grad_norm": 0.02254868485033512,
+ "learning_rate": 0.0006,
+ "loss": 4.563869953155518,
+ "step": 3937
+ },
+ {
+ "epoch": 54.698689956331876,
+ "grad_norm": 0.019730940461158752,
+ "learning_rate": 0.0006,
+ "loss": 4.592165946960449,
+ "step": 3938
+ },
+ {
+ "epoch": 54.712663755458514,
+ "grad_norm": 0.015146561898291111,
+ "learning_rate": 0.0006,
+ "loss": 4.6228742599487305,
+ "step": 3939
+ },
+ {
+ "epoch": 54.72663755458515,
+ "grad_norm": 0.01824716106057167,
+ "learning_rate": 0.0006,
+ "loss": 4.716236114501953,
+ "step": 3940
+ },
+ {
+ "epoch": 54.74061135371179,
+ "grad_norm": 0.017730550840497017,
+ "learning_rate": 0.0006,
+ "loss": 4.650232315063477,
+ "step": 3941
+ },
+ {
+ "epoch": 54.75458515283843,
+ "grad_norm": 0.014875595457851887,
+ "learning_rate": 0.0006,
+ "loss": 4.6377458572387695,
+ "step": 3942
+ },
+ {
+ "epoch": 54.76855895196506,
+ "grad_norm": 0.01613209955394268,
+ "learning_rate": 0.0006,
+ "loss": 4.5768537521362305,
+ "step": 3943
+ },
+ {
+ "epoch": 54.7825327510917,
+ "grad_norm": 0.017311280593276024,
+ "learning_rate": 0.0006,
+ "loss": 4.625744342803955,
+ "step": 3944
+ },
+ {
+ "epoch": 54.79650655021834,
+ "grad_norm": 0.018770085647702217,
+ "learning_rate": 0.0006,
+ "loss": 4.613787651062012,
+ "step": 3945
+ },
+ {
+ "epoch": 54.81048034934498,
+ "grad_norm": 0.0165257528424263,
+ "learning_rate": 0.0006,
+ "loss": 4.5991315841674805,
+ "step": 3946
+ },
+ {
+ "epoch": 54.82445414847162,
+ "grad_norm": 0.017433857545256615,
+ "learning_rate": 0.0006,
+ "loss": 4.513872146606445,
+ "step": 3947
+ },
+ {
+ "epoch": 54.838427947598255,
+ "grad_norm": 0.020126905292272568,
+ "learning_rate": 0.0006,
+ "loss": 4.616185188293457,
+ "step": 3948
+ },
+ {
+ "epoch": 54.852401746724894,
+ "grad_norm": 0.01819358579814434,
+ "learning_rate": 0.0006,
+ "loss": 4.535691261291504,
+ "step": 3949
+ },
+ {
+ "epoch": 54.866375545851525,
+ "grad_norm": 0.01899927295744419,
+ "learning_rate": 0.0006,
+ "loss": 4.595199108123779,
+ "step": 3950
+ },
+ {
+ "epoch": 54.880349344978164,
+ "grad_norm": 0.02005566656589508,
+ "learning_rate": 0.0006,
+ "loss": 4.633289337158203,
+ "step": 3951
+ },
+ {
+ "epoch": 54.8943231441048,
+ "grad_norm": 0.021570220589637756,
+ "learning_rate": 0.0006,
+ "loss": 4.592503547668457,
+ "step": 3952
+ },
+ {
+ "epoch": 54.90829694323144,
+ "grad_norm": 0.01883506216108799,
+ "learning_rate": 0.0006,
+ "loss": 4.6597795486450195,
+ "step": 3953
+ },
+ {
+ "epoch": 54.92227074235808,
+ "grad_norm": 0.018070021644234657,
+ "learning_rate": 0.0006,
+ "loss": 4.601474285125732,
+ "step": 3954
+ },
+ {
+ "epoch": 54.93624454148472,
+ "grad_norm": 0.018090378493070602,
+ "learning_rate": 0.0006,
+ "loss": 4.569846153259277,
+ "step": 3955
+ },
+ {
+ "epoch": 54.95021834061135,
+ "grad_norm": 0.018751563504338264,
+ "learning_rate": 0.0006,
+ "loss": 4.646064281463623,
+ "step": 3956
+ },
+ {
+ "epoch": 54.96419213973799,
+ "grad_norm": 0.017512254416942596,
+ "learning_rate": 0.0006,
+ "loss": 4.606670379638672,
+ "step": 3957
+ },
+ {
+ "epoch": 54.97816593886463,
+ "grad_norm": 0.018771110102534294,
+ "learning_rate": 0.0006,
+ "loss": 4.604315757751465,
+ "step": 3958
+ },
+ {
+ "epoch": 54.992139737991266,
+ "grad_norm": 0.01611141301691532,
+ "learning_rate": 0.0006,
+ "loss": 4.643245220184326,
+ "step": 3959
+ },
+ {
+ "epoch": 55.0,
+ "grad_norm": 0.0156728383153677,
+ "learning_rate": 0.0006,
+ "loss": 4.6739044189453125,
+ "step": 3960
+ },
+ {
+ "epoch": 55.0,
+ "eval_loss": 4.851032257080078,
+ "eval_runtime": 56.8699,
+ "eval_samples_per_second": 42.94,
+ "eval_steps_per_second": 1.354,
+ "step": 3960
+ },
+ {
+ "epoch": 55.01397379912664,
+ "grad_norm": 0.016214672476053238,
+ "learning_rate": 0.0006,
+ "loss": 4.571749687194824,
+ "step": 3961
+ },
+ {
+ "epoch": 55.02794759825328,
+ "grad_norm": 0.01649155467748642,
+ "learning_rate": 0.0006,
+ "loss": 4.553592205047607,
+ "step": 3962
+ },
+ {
+ "epoch": 55.041921397379916,
+ "grad_norm": 0.016793379560112953,
+ "learning_rate": 0.0006,
+ "loss": 4.650628566741943,
+ "step": 3963
+ },
+ {
+ "epoch": 55.05589519650655,
+ "grad_norm": 0.0191937405616045,
+ "learning_rate": 0.0006,
+ "loss": 4.620532989501953,
+ "step": 3964
+ },
+ {
+ "epoch": 55.069868995633186,
+ "grad_norm": 0.020976556465029716,
+ "learning_rate": 0.0006,
+ "loss": 4.6329545974731445,
+ "step": 3965
+ },
+ {
+ "epoch": 55.083842794759825,
+ "grad_norm": 0.023450477048754692,
+ "learning_rate": 0.0006,
+ "loss": 4.60828971862793,
+ "step": 3966
+ },
+ {
+ "epoch": 55.09781659388646,
+ "grad_norm": 0.026778312399983406,
+ "learning_rate": 0.0006,
+ "loss": 4.617481708526611,
+ "step": 3967
+ },
+ {
+ "epoch": 55.1117903930131,
+ "grad_norm": 0.02333568036556244,
+ "learning_rate": 0.0006,
+ "loss": 4.580722332000732,
+ "step": 3968
+ },
+ {
+ "epoch": 55.12576419213974,
+ "grad_norm": 0.018491310998797417,
+ "learning_rate": 0.0006,
+ "loss": 4.590953350067139,
+ "step": 3969
+ },
+ {
+ "epoch": 55.13973799126637,
+ "grad_norm": 0.020339803770184517,
+ "learning_rate": 0.0006,
+ "loss": 4.518374443054199,
+ "step": 3970
+ },
+ {
+ "epoch": 55.15371179039301,
+ "grad_norm": 0.01759706623852253,
+ "learning_rate": 0.0006,
+ "loss": 4.560522079467773,
+ "step": 3971
+ },
+ {
+ "epoch": 55.16768558951965,
+ "grad_norm": 0.015875980257987976,
+ "learning_rate": 0.0006,
+ "loss": 4.4795708656311035,
+ "step": 3972
+ },
+ {
+ "epoch": 55.18165938864629,
+ "grad_norm": 0.01650134287774563,
+ "learning_rate": 0.0006,
+ "loss": 4.57435941696167,
+ "step": 3973
+ },
+ {
+ "epoch": 55.19563318777293,
+ "grad_norm": 0.01734331250190735,
+ "learning_rate": 0.0006,
+ "loss": 4.661007881164551,
+ "step": 3974
+ },
+ {
+ "epoch": 55.209606986899566,
+ "grad_norm": 0.01651303470134735,
+ "learning_rate": 0.0006,
+ "loss": 4.511148452758789,
+ "step": 3975
+ },
+ {
+ "epoch": 55.223580786026204,
+ "grad_norm": 0.016189761459827423,
+ "learning_rate": 0.0006,
+ "loss": 4.592350006103516,
+ "step": 3976
+ },
+ {
+ "epoch": 55.237554585152836,
+ "grad_norm": 0.01833222061395645,
+ "learning_rate": 0.0006,
+ "loss": 4.506649971008301,
+ "step": 3977
+ },
+ {
+ "epoch": 55.251528384279474,
+ "grad_norm": 0.02433410845696926,
+ "learning_rate": 0.0006,
+ "loss": 4.49500036239624,
+ "step": 3978
+ },
+ {
+ "epoch": 55.26550218340611,
+ "grad_norm": 0.0314289927482605,
+ "learning_rate": 0.0006,
+ "loss": 4.507624626159668,
+ "step": 3979
+ },
+ {
+ "epoch": 55.27947598253275,
+ "grad_norm": 0.03013679012656212,
+ "learning_rate": 0.0006,
+ "loss": 4.533641815185547,
+ "step": 3980
+ },
+ {
+ "epoch": 55.29344978165939,
+ "grad_norm": 0.024524061009287834,
+ "learning_rate": 0.0006,
+ "loss": 4.5721116065979,
+ "step": 3981
+ },
+ {
+ "epoch": 55.30742358078603,
+ "grad_norm": 0.02640336938202381,
+ "learning_rate": 0.0006,
+ "loss": 4.652521133422852,
+ "step": 3982
+ },
+ {
+ "epoch": 55.32139737991266,
+ "grad_norm": 0.021309101954102516,
+ "learning_rate": 0.0006,
+ "loss": 4.600375175476074,
+ "step": 3983
+ },
+ {
+ "epoch": 55.3353711790393,
+ "grad_norm": 0.023008549585938454,
+ "learning_rate": 0.0006,
+ "loss": 4.53914737701416,
+ "step": 3984
+ },
+ {
+ "epoch": 55.34934497816594,
+ "grad_norm": 0.022491468116641045,
+ "learning_rate": 0.0006,
+ "loss": 4.661163330078125,
+ "step": 3985
+ },
+ {
+ "epoch": 55.36331877729258,
+ "grad_norm": 0.01904081553220749,
+ "learning_rate": 0.0006,
+ "loss": 4.5426177978515625,
+ "step": 3986
+ },
+ {
+ "epoch": 55.377292576419215,
+ "grad_norm": 0.019229518249630928,
+ "learning_rate": 0.0006,
+ "loss": 4.606066703796387,
+ "step": 3987
+ },
+ {
+ "epoch": 55.391266375545854,
+ "grad_norm": 0.01781383901834488,
+ "learning_rate": 0.0006,
+ "loss": 4.54642391204834,
+ "step": 3988
+ },
+ {
+ "epoch": 55.40524017467249,
+ "grad_norm": 0.0174600500613451,
+ "learning_rate": 0.0006,
+ "loss": 4.711891174316406,
+ "step": 3989
+ },
+ {
+ "epoch": 55.419213973799124,
+ "grad_norm": 0.018154339864850044,
+ "learning_rate": 0.0006,
+ "loss": 4.527122974395752,
+ "step": 3990
+ },
+ {
+ "epoch": 55.43318777292576,
+ "grad_norm": 0.01914648339152336,
+ "learning_rate": 0.0006,
+ "loss": 4.641975402832031,
+ "step": 3991
+ },
+ {
+ "epoch": 55.4471615720524,
+ "grad_norm": 0.017118625342845917,
+ "learning_rate": 0.0006,
+ "loss": 4.511425971984863,
+ "step": 3992
+ },
+ {
+ "epoch": 55.46113537117904,
+ "grad_norm": 0.01551489345729351,
+ "learning_rate": 0.0006,
+ "loss": 4.551141738891602,
+ "step": 3993
+ },
+ {
+ "epoch": 55.47510917030568,
+ "grad_norm": 0.015072065405547619,
+ "learning_rate": 0.0006,
+ "loss": 4.651115894317627,
+ "step": 3994
+ },
+ {
+ "epoch": 55.48908296943232,
+ "grad_norm": 0.017010986804962158,
+ "learning_rate": 0.0006,
+ "loss": 4.583282470703125,
+ "step": 3995
+ },
+ {
+ "epoch": 55.50305676855895,
+ "grad_norm": 0.01764831133186817,
+ "learning_rate": 0.0006,
+ "loss": 4.593364238739014,
+ "step": 3996
+ },
+ {
+ "epoch": 55.51703056768559,
+ "grad_norm": 0.01762884296476841,
+ "learning_rate": 0.0006,
+ "loss": 4.45671272277832,
+ "step": 3997
+ },
+ {
+ "epoch": 55.531004366812226,
+ "grad_norm": 0.01600457727909088,
+ "learning_rate": 0.0006,
+ "loss": 4.511260032653809,
+ "step": 3998
+ },
+ {
+ "epoch": 55.544978165938865,
+ "grad_norm": 0.015706021338701248,
+ "learning_rate": 0.0006,
+ "loss": 4.47593355178833,
+ "step": 3999
+ },
+ {
+ "epoch": 55.5589519650655,
+ "grad_norm": 0.017718996852636337,
+ "learning_rate": 0.0006,
+ "loss": 4.610543251037598,
+ "step": 4000
+ },
+ {
+ "epoch": 55.57292576419214,
+ "grad_norm": 0.015073757618665695,
+ "learning_rate": 0.0006,
+ "loss": 4.466620445251465,
+ "step": 4001
+ },
+ {
+ "epoch": 55.58689956331878,
+ "grad_norm": 0.01789247989654541,
+ "learning_rate": 0.0006,
+ "loss": 4.503409385681152,
+ "step": 4002
+ },
+ {
+ "epoch": 55.60087336244541,
+ "grad_norm": 0.022329598665237427,
+ "learning_rate": 0.0006,
+ "loss": 4.62099552154541,
+ "step": 4003
+ },
+ {
+ "epoch": 55.61484716157205,
+ "grad_norm": 0.02343829534947872,
+ "learning_rate": 0.0006,
+ "loss": 4.620813369750977,
+ "step": 4004
+ },
+ {
+ "epoch": 55.62882096069869,
+ "grad_norm": 0.023279687389731407,
+ "learning_rate": 0.0006,
+ "loss": 4.545765399932861,
+ "step": 4005
+ },
+ {
+ "epoch": 55.64279475982533,
+ "grad_norm": 0.020817464217543602,
+ "learning_rate": 0.0006,
+ "loss": 4.561470031738281,
+ "step": 4006
+ },
+ {
+ "epoch": 55.65676855895197,
+ "grad_norm": 0.020270058885216713,
+ "learning_rate": 0.0006,
+ "loss": 4.5875020027160645,
+ "step": 4007
+ },
+ {
+ "epoch": 55.670742358078606,
+ "grad_norm": 0.021085843443870544,
+ "learning_rate": 0.0006,
+ "loss": 4.504560470581055,
+ "step": 4008
+ },
+ {
+ "epoch": 55.68471615720524,
+ "grad_norm": 0.018507730215787888,
+ "learning_rate": 0.0006,
+ "loss": 4.520644664764404,
+ "step": 4009
+ },
+ {
+ "epoch": 55.698689956331876,
+ "grad_norm": 0.020571516826748848,
+ "learning_rate": 0.0006,
+ "loss": 4.539643287658691,
+ "step": 4010
+ },
+ {
+ "epoch": 55.712663755458514,
+ "grad_norm": 0.021484510973095894,
+ "learning_rate": 0.0006,
+ "loss": 4.628833293914795,
+ "step": 4011
+ },
+ {
+ "epoch": 55.72663755458515,
+ "grad_norm": 0.017693819478154182,
+ "learning_rate": 0.0006,
+ "loss": 4.6023783683776855,
+ "step": 4012
+ },
+ {
+ "epoch": 55.74061135371179,
+ "grad_norm": 0.020186059176921844,
+ "learning_rate": 0.0006,
+ "loss": 4.634884834289551,
+ "step": 4013
+ },
+ {
+ "epoch": 55.75458515283843,
+ "grad_norm": 0.01835780404508114,
+ "learning_rate": 0.0006,
+ "loss": 4.577791213989258,
+ "step": 4014
+ },
+ {
+ "epoch": 55.76855895196506,
+ "grad_norm": 0.018162988126277924,
+ "learning_rate": 0.0006,
+ "loss": 4.641597270965576,
+ "step": 4015
+ },
+ {
+ "epoch": 55.7825327510917,
+ "grad_norm": 0.018909504637122154,
+ "learning_rate": 0.0006,
+ "loss": 4.487447738647461,
+ "step": 4016
+ },
+ {
+ "epoch": 55.79650655021834,
+ "grad_norm": 0.018972251564264297,
+ "learning_rate": 0.0006,
+ "loss": 4.590317726135254,
+ "step": 4017
+ },
+ {
+ "epoch": 55.81048034934498,
+ "grad_norm": 0.01793607883155346,
+ "learning_rate": 0.0006,
+ "loss": 4.4994611740112305,
+ "step": 4018
+ },
+ {
+ "epoch": 55.82445414847162,
+ "grad_norm": 0.02123514749109745,
+ "learning_rate": 0.0006,
+ "loss": 4.479006767272949,
+ "step": 4019
+ },
+ {
+ "epoch": 55.838427947598255,
+ "grad_norm": 0.021246599033474922,
+ "learning_rate": 0.0006,
+ "loss": 4.4738640785217285,
+ "step": 4020
+ },
+ {
+ "epoch": 55.852401746724894,
+ "grad_norm": 0.020131602883338928,
+ "learning_rate": 0.0006,
+ "loss": 4.600308418273926,
+ "step": 4021
+ },
+ {
+ "epoch": 55.866375545851525,
+ "grad_norm": 0.019159460440278053,
+ "learning_rate": 0.0006,
+ "loss": 4.540908336639404,
+ "step": 4022
+ },
+ {
+ "epoch": 55.880349344978164,
+ "grad_norm": 0.023114705458283424,
+ "learning_rate": 0.0006,
+ "loss": 4.641247749328613,
+ "step": 4023
+ },
+ {
+ "epoch": 55.8943231441048,
+ "grad_norm": 0.02296135015785694,
+ "learning_rate": 0.0006,
+ "loss": 4.444624423980713,
+ "step": 4024
+ },
+ {
+ "epoch": 55.90829694323144,
+ "grad_norm": 0.02204655110836029,
+ "learning_rate": 0.0006,
+ "loss": 4.506349563598633,
+ "step": 4025
+ },
+ {
+ "epoch": 55.92227074235808,
+ "grad_norm": 0.019338225945830345,
+ "learning_rate": 0.0006,
+ "loss": 4.616047382354736,
+ "step": 4026
+ },
+ {
+ "epoch": 55.93624454148472,
+ "grad_norm": 0.020357884466648102,
+ "learning_rate": 0.0006,
+ "loss": 4.48170280456543,
+ "step": 4027
+ },
+ {
+ "epoch": 55.95021834061135,
+ "grad_norm": 0.02426772564649582,
+ "learning_rate": 0.0006,
+ "loss": 4.576877117156982,
+ "step": 4028
+ },
+ {
+ "epoch": 55.96419213973799,
+ "grad_norm": 0.023351816460490227,
+ "learning_rate": 0.0006,
+ "loss": 4.458885192871094,
+ "step": 4029
+ },
+ {
+ "epoch": 55.97816593886463,
+ "grad_norm": 0.019892243668437004,
+ "learning_rate": 0.0006,
+ "loss": 4.520229816436768,
+ "step": 4030
+ },
+ {
+ "epoch": 55.992139737991266,
+ "grad_norm": 0.018331089988350868,
+ "learning_rate": 0.0006,
+ "loss": 4.432201862335205,
+ "step": 4031
+ },
+ {
+ "epoch": 56.0,
+ "grad_norm": 0.021300828084349632,
+ "learning_rate": 0.0006,
+ "loss": 4.617072582244873,
+ "step": 4032
+ },
+ {
+ "epoch": 56.0,
+ "eval_loss": 4.811229705810547,
+ "eval_runtime": 56.4846,
+ "eval_samples_per_second": 43.233,
+ "eval_steps_per_second": 1.363,
+ "step": 4032
+ },
+ {
+ "epoch": 56.01397379912664,
+ "grad_norm": 0.02199401892721653,
+ "learning_rate": 0.0006,
+ "loss": 4.513965606689453,
+ "step": 4033
+ },
+ {
+ "epoch": 56.02794759825328,
+ "grad_norm": 0.022109858691692352,
+ "learning_rate": 0.0006,
+ "loss": 4.557568550109863,
+ "step": 4034
+ },
+ {
+ "epoch": 56.041921397379916,
+ "grad_norm": 0.020103694871068,
+ "learning_rate": 0.0006,
+ "loss": 4.569512844085693,
+ "step": 4035
+ },
+ {
+ "epoch": 56.05589519650655,
+ "grad_norm": 0.01898941583931446,
+ "learning_rate": 0.0006,
+ "loss": 4.623138904571533,
+ "step": 4036
+ },
+ {
+ "epoch": 56.069868995633186,
+ "grad_norm": 0.02227010577917099,
+ "learning_rate": 0.0006,
+ "loss": 4.433316230773926,
+ "step": 4037
+ },
+ {
+ "epoch": 56.083842794759825,
+ "grad_norm": 0.024683045223355293,
+ "learning_rate": 0.0006,
+ "loss": 4.458296775817871,
+ "step": 4038
+ },
+ {
+ "epoch": 56.09781659388646,
+ "grad_norm": 0.02900184690952301,
+ "learning_rate": 0.0006,
+ "loss": 4.531060218811035,
+ "step": 4039
+ },
+ {
+ "epoch": 56.1117903930131,
+ "grad_norm": 0.025557566434144974,
+ "learning_rate": 0.0006,
+ "loss": 4.56685733795166,
+ "step": 4040
+ },
+ {
+ "epoch": 56.12576419213974,
+ "grad_norm": 0.021414656192064285,
+ "learning_rate": 0.0006,
+ "loss": 4.469037055969238,
+ "step": 4041
+ },
+ {
+ "epoch": 56.13973799126637,
+ "grad_norm": 0.023262616246938705,
+ "learning_rate": 0.0006,
+ "loss": 4.493876934051514,
+ "step": 4042
+ },
+ {
+ "epoch": 56.15371179039301,
+ "grad_norm": 0.023011326789855957,
+ "learning_rate": 0.0006,
+ "loss": 4.4572906494140625,
+ "step": 4043
+ },
+ {
+ "epoch": 56.16768558951965,
+ "grad_norm": 0.02177615463733673,
+ "learning_rate": 0.0006,
+ "loss": 4.615389823913574,
+ "step": 4044
+ },
+ {
+ "epoch": 56.18165938864629,
+ "grad_norm": 0.021527713164687157,
+ "learning_rate": 0.0006,
+ "loss": 4.613762855529785,
+ "step": 4045
+ },
+ {
+ "epoch": 56.19563318777293,
+ "grad_norm": 0.01982325315475464,
+ "learning_rate": 0.0006,
+ "loss": 4.53670597076416,
+ "step": 4046
+ },
+ {
+ "epoch": 56.209606986899566,
+ "grad_norm": 0.022484654560685158,
+ "learning_rate": 0.0006,
+ "loss": 4.605565071105957,
+ "step": 4047
+ },
+ {
+ "epoch": 56.223580786026204,
+ "grad_norm": 0.02346140146255493,
+ "learning_rate": 0.0006,
+ "loss": 4.6265950202941895,
+ "step": 4048
+ },
+ {
+ "epoch": 56.237554585152836,
+ "grad_norm": 0.01876714453101158,
+ "learning_rate": 0.0006,
+ "loss": 4.481525421142578,
+ "step": 4049
+ },
+ {
+ "epoch": 56.251528384279474,
+ "grad_norm": 0.01926851086318493,
+ "learning_rate": 0.0006,
+ "loss": 4.630258083343506,
+ "step": 4050
+ },
+ {
+ "epoch": 56.26550218340611,
+ "grad_norm": 0.020858589559793472,
+ "learning_rate": 0.0006,
+ "loss": 4.5265913009643555,
+ "step": 4051
+ },
+ {
+ "epoch": 56.27947598253275,
+ "grad_norm": 0.01889280416071415,
+ "learning_rate": 0.0006,
+ "loss": 4.581275463104248,
+ "step": 4052
+ },
+ {
+ "epoch": 56.29344978165939,
+ "grad_norm": 0.01801002025604248,
+ "learning_rate": 0.0006,
+ "loss": 4.498557090759277,
+ "step": 4053
+ },
+ {
+ "epoch": 56.30742358078603,
+ "grad_norm": 0.02252180315554142,
+ "learning_rate": 0.0006,
+ "loss": 4.565267086029053,
+ "step": 4054
+ },
+ {
+ "epoch": 56.32139737991266,
+ "grad_norm": 0.02306070365011692,
+ "learning_rate": 0.0006,
+ "loss": 4.3591389656066895,
+ "step": 4055
+ },
+ {
+ "epoch": 56.3353711790393,
+ "grad_norm": 0.020292513072490692,
+ "learning_rate": 0.0006,
+ "loss": 4.407517433166504,
+ "step": 4056
+ },
+ {
+ "epoch": 56.34934497816594,
+ "grad_norm": 0.019154123961925507,
+ "learning_rate": 0.0006,
+ "loss": 4.482261657714844,
+ "step": 4057
+ },
+ {
+ "epoch": 56.36331877729258,
+ "grad_norm": 0.017338937148451805,
+ "learning_rate": 0.0006,
+ "loss": 4.4463582038879395,
+ "step": 4058
+ },
+ {
+ "epoch": 56.377292576419215,
+ "grad_norm": 0.01730559952557087,
+ "learning_rate": 0.0006,
+ "loss": 4.542474746704102,
+ "step": 4059
+ },
+ {
+ "epoch": 56.391266375545854,
+ "grad_norm": 0.017029426991939545,
+ "learning_rate": 0.0006,
+ "loss": 4.552488327026367,
+ "step": 4060
+ },
+ {
+ "epoch": 56.40524017467249,
+ "grad_norm": 0.01745498552918434,
+ "learning_rate": 0.0006,
+ "loss": 4.52928352355957,
+ "step": 4061
+ },
+ {
+ "epoch": 56.419213973799124,
+ "grad_norm": 0.0168515145778656,
+ "learning_rate": 0.0006,
+ "loss": 4.4516801834106445,
+ "step": 4062
+ },
+ {
+ "epoch": 56.43318777292576,
+ "grad_norm": 0.018545417115092278,
+ "learning_rate": 0.0006,
+ "loss": 4.569811820983887,
+ "step": 4063
+ },
+ {
+ "epoch": 56.4471615720524,
+ "grad_norm": 0.02227838523685932,
+ "learning_rate": 0.0006,
+ "loss": 4.412224769592285,
+ "step": 4064
+ },
+ {
+ "epoch": 56.46113537117904,
+ "grad_norm": 0.023338504135608673,
+ "learning_rate": 0.0006,
+ "loss": 4.490184307098389,
+ "step": 4065
+ },
+ {
+ "epoch": 56.47510917030568,
+ "grad_norm": 0.02257535606622696,
+ "learning_rate": 0.0006,
+ "loss": 4.479152679443359,
+ "step": 4066
+ },
+ {
+ "epoch": 56.48908296943232,
+ "grad_norm": 0.022737598046660423,
+ "learning_rate": 0.0006,
+ "loss": 4.418235778808594,
+ "step": 4067
+ },
+ {
+ "epoch": 56.50305676855895,
+ "grad_norm": 0.024603573605418205,
+ "learning_rate": 0.0006,
+ "loss": 4.45781135559082,
+ "step": 4068
+ },
+ {
+ "epoch": 56.51703056768559,
+ "grad_norm": 0.024953778833150864,
+ "learning_rate": 0.0006,
+ "loss": 4.593907833099365,
+ "step": 4069
+ },
+ {
+ "epoch": 56.531004366812226,
+ "grad_norm": 0.02195645309984684,
+ "learning_rate": 0.0006,
+ "loss": 4.467643737792969,
+ "step": 4070
+ },
+ {
+ "epoch": 56.544978165938865,
+ "grad_norm": 0.017126578837633133,
+ "learning_rate": 0.0006,
+ "loss": 4.555484771728516,
+ "step": 4071
+ },
+ {
+ "epoch": 56.5589519650655,
+ "grad_norm": 0.019482605159282684,
+ "learning_rate": 0.0006,
+ "loss": 4.430915832519531,
+ "step": 4072
+ },
+ {
+ "epoch": 56.57292576419214,
+ "grad_norm": 0.0219392292201519,
+ "learning_rate": 0.0006,
+ "loss": 4.593365669250488,
+ "step": 4073
+ },
+ {
+ "epoch": 56.58689956331878,
+ "grad_norm": 0.02420145645737648,
+ "learning_rate": 0.0006,
+ "loss": 4.491467475891113,
+ "step": 4074
+ },
+ {
+ "epoch": 56.60087336244541,
+ "grad_norm": 0.021523723378777504,
+ "learning_rate": 0.0006,
+ "loss": 4.463634490966797,
+ "step": 4075
+ },
+ {
+ "epoch": 56.61484716157205,
+ "grad_norm": 0.020242048427462578,
+ "learning_rate": 0.0006,
+ "loss": 4.577968597412109,
+ "step": 4076
+ },
+ {
+ "epoch": 56.62882096069869,
+ "grad_norm": 0.017771173268556595,
+ "learning_rate": 0.0006,
+ "loss": 4.568269729614258,
+ "step": 4077
+ },
+ {
+ "epoch": 56.64279475982533,
+ "grad_norm": 0.01598544418811798,
+ "learning_rate": 0.0006,
+ "loss": 4.476841926574707,
+ "step": 4078
+ },
+ {
+ "epoch": 56.65676855895197,
+ "grad_norm": 0.0168446097522974,
+ "learning_rate": 0.0006,
+ "loss": 4.480965614318848,
+ "step": 4079
+ },
+ {
+ "epoch": 56.670742358078606,
+ "grad_norm": 0.014948660507798195,
+ "learning_rate": 0.0006,
+ "loss": 4.535400390625,
+ "step": 4080
+ },
+ {
+ "epoch": 56.68471615720524,
+ "grad_norm": 0.014899153262376785,
+ "learning_rate": 0.0006,
+ "loss": 4.476048946380615,
+ "step": 4081
+ },
+ {
+ "epoch": 56.698689956331876,
+ "grad_norm": 0.014051870442926884,
+ "learning_rate": 0.0006,
+ "loss": 4.533884048461914,
+ "step": 4082
+ },
+ {
+ "epoch": 56.712663755458514,
+ "grad_norm": 0.014708572067320347,
+ "learning_rate": 0.0006,
+ "loss": 4.586690902709961,
+ "step": 4083
+ },
+ {
+ "epoch": 56.72663755458515,
+ "grad_norm": 0.014797299169003963,
+ "learning_rate": 0.0006,
+ "loss": 4.446291446685791,
+ "step": 4084
+ },
+ {
+ "epoch": 56.74061135371179,
+ "grad_norm": 0.015256541781127453,
+ "learning_rate": 0.0006,
+ "loss": 4.553572654724121,
+ "step": 4085
+ },
+ {
+ "epoch": 56.75458515283843,
+ "grad_norm": 0.016042491421103477,
+ "learning_rate": 0.0006,
+ "loss": 4.507801055908203,
+ "step": 4086
+ },
+ {
+ "epoch": 56.76855895196506,
+ "grad_norm": 0.018835173919796944,
+ "learning_rate": 0.0006,
+ "loss": 4.393911838531494,
+ "step": 4087
+ },
+ {
+ "epoch": 56.7825327510917,
+ "grad_norm": 0.021994033828377724,
+ "learning_rate": 0.0006,
+ "loss": 4.543368339538574,
+ "step": 4088
+ },
+ {
+ "epoch": 56.79650655021834,
+ "grad_norm": 0.0230990182608366,
+ "learning_rate": 0.0006,
+ "loss": 4.507420539855957,
+ "step": 4089
+ },
+ {
+ "epoch": 56.81048034934498,
+ "grad_norm": 0.018818842247128487,
+ "learning_rate": 0.0006,
+ "loss": 4.499941349029541,
+ "step": 4090
+ },
+ {
+ "epoch": 56.82445414847162,
+ "grad_norm": 0.018257278949022293,
+ "learning_rate": 0.0006,
+ "loss": 4.380057334899902,
+ "step": 4091
+ },
+ {
+ "epoch": 56.838427947598255,
+ "grad_norm": 0.020073018968105316,
+ "learning_rate": 0.0006,
+ "loss": 4.572811603546143,
+ "step": 4092
+ },
+ {
+ "epoch": 56.852401746724894,
+ "grad_norm": 0.02078658528625965,
+ "learning_rate": 0.0006,
+ "loss": 4.446835517883301,
+ "step": 4093
+ },
+ {
+ "epoch": 56.866375545851525,
+ "grad_norm": 0.01793118566274643,
+ "learning_rate": 0.0006,
+ "loss": 4.517854690551758,
+ "step": 4094
+ },
+ {
+ "epoch": 56.880349344978164,
+ "grad_norm": 0.01608281210064888,
+ "learning_rate": 0.0006,
+ "loss": 4.506482124328613,
+ "step": 4095
+ },
+ {
+ "epoch": 56.8943231441048,
+ "grad_norm": 0.017405368387699127,
+ "learning_rate": 0.0006,
+ "loss": 4.439718246459961,
+ "step": 4096
+ },
+ {
+ "epoch": 56.90829694323144,
+ "grad_norm": 0.017930863425135612,
+ "learning_rate": 0.0006,
+ "loss": 4.409365653991699,
+ "step": 4097
+ },
+ {
+ "epoch": 56.92227074235808,
+ "grad_norm": 0.015757670626044273,
+ "learning_rate": 0.0006,
+ "loss": 4.505547523498535,
+ "step": 4098
+ },
+ {
+ "epoch": 56.93624454148472,
+ "grad_norm": 0.01580345816910267,
+ "learning_rate": 0.0006,
+ "loss": 4.462774753570557,
+ "step": 4099
+ },
+ {
+ "epoch": 56.95021834061135,
+ "grad_norm": 0.01659180410206318,
+ "learning_rate": 0.0006,
+ "loss": 4.49345588684082,
+ "step": 4100
+ },
+ {
+ "epoch": 56.96419213973799,
+ "grad_norm": 0.01877259835600853,
+ "learning_rate": 0.0006,
+ "loss": 4.574443340301514,
+ "step": 4101
+ },
+ {
+ "epoch": 56.97816593886463,
+ "grad_norm": 0.01977839693427086,
+ "learning_rate": 0.0006,
+ "loss": 4.574034690856934,
+ "step": 4102
+ },
+ {
+ "epoch": 56.992139737991266,
+ "grad_norm": 0.020627597346901894,
+ "learning_rate": 0.0006,
+ "loss": 4.50969123840332,
+ "step": 4103
+ },
+ {
+ "epoch": 57.0,
+ "grad_norm": 0.025128807872533798,
+ "learning_rate": 0.0006,
+ "loss": 4.401662826538086,
+ "step": 4104
+ },
+ {
+ "epoch": 57.0,
+ "eval_loss": 4.779252052307129,
+ "eval_runtime": 56.8104,
+ "eval_samples_per_second": 42.985,
+ "eval_steps_per_second": 1.355,
+ "step": 4104
+ },
+ {
+ "epoch": 57.01397379912664,
+ "grad_norm": 0.027712570503354073,
+ "learning_rate": 0.0006,
+ "loss": 4.472955703735352,
+ "step": 4105
+ },
+ {
+ "epoch": 57.02794759825328,
+ "grad_norm": 0.02164197526872158,
+ "learning_rate": 0.0006,
+ "loss": 4.387596130371094,
+ "step": 4106
+ },
+ {
+ "epoch": 57.041921397379916,
+ "grad_norm": 0.017972752451896667,
+ "learning_rate": 0.0006,
+ "loss": 4.4326276779174805,
+ "step": 4107
+ },
+ {
+ "epoch": 57.05589519650655,
+ "grad_norm": 0.020811019465327263,
+ "learning_rate": 0.0006,
+ "loss": 4.452951431274414,
+ "step": 4108
+ },
+ {
+ "epoch": 57.069868995633186,
+ "grad_norm": 0.019913259893655777,
+ "learning_rate": 0.0006,
+ "loss": 4.448301315307617,
+ "step": 4109
+ },
+ {
+ "epoch": 57.083842794759825,
+ "grad_norm": 0.017409102991223335,
+ "learning_rate": 0.0006,
+ "loss": 4.4108357429504395,
+ "step": 4110
+ },
+ {
+ "epoch": 57.09781659388646,
+ "grad_norm": 0.019374269992113113,
+ "learning_rate": 0.0006,
+ "loss": 4.398412227630615,
+ "step": 4111
+ },
+ {
+ "epoch": 57.1117903930131,
+ "grad_norm": 0.015521776862442493,
+ "learning_rate": 0.0006,
+ "loss": 4.508594989776611,
+ "step": 4112
+ },
+ {
+ "epoch": 57.12576419213974,
+ "grad_norm": 0.017201920971274376,
+ "learning_rate": 0.0006,
+ "loss": 4.449558734893799,
+ "step": 4113
+ },
+ {
+ "epoch": 57.13973799126637,
+ "grad_norm": 0.018062541261315346,
+ "learning_rate": 0.0006,
+ "loss": 4.517874240875244,
+ "step": 4114
+ },
+ {
+ "epoch": 57.15371179039301,
+ "grad_norm": 0.015846073627471924,
+ "learning_rate": 0.0006,
+ "loss": 4.427501678466797,
+ "step": 4115
+ },
+ {
+ "epoch": 57.16768558951965,
+ "grad_norm": 0.01849362626671791,
+ "learning_rate": 0.0006,
+ "loss": 4.433181285858154,
+ "step": 4116
+ },
+ {
+ "epoch": 57.18165938864629,
+ "grad_norm": 0.01948804222047329,
+ "learning_rate": 0.0006,
+ "loss": 4.446122646331787,
+ "step": 4117
+ },
+ {
+ "epoch": 57.19563318777293,
+ "grad_norm": 0.017354389652609825,
+ "learning_rate": 0.0006,
+ "loss": 4.431153774261475,
+ "step": 4118
+ },
+ {
+ "epoch": 57.209606986899566,
+ "grad_norm": 0.015817373991012573,
+ "learning_rate": 0.0006,
+ "loss": 4.448566913604736,
+ "step": 4119
+ },
+ {
+ "epoch": 57.223580786026204,
+ "grad_norm": 0.01635785959661007,
+ "learning_rate": 0.0006,
+ "loss": 4.408031463623047,
+ "step": 4120
+ },
+ {
+ "epoch": 57.237554585152836,
+ "grad_norm": 0.014528338797390461,
+ "learning_rate": 0.0006,
+ "loss": 4.541668891906738,
+ "step": 4121
+ },
+ {
+ "epoch": 57.251528384279474,
+ "grad_norm": 0.01589174196124077,
+ "learning_rate": 0.0006,
+ "loss": 4.513199329376221,
+ "step": 4122
+ },
+ {
+ "epoch": 57.26550218340611,
+ "grad_norm": 0.016093695536255836,
+ "learning_rate": 0.0006,
+ "loss": 4.516178131103516,
+ "step": 4123
+ },
+ {
+ "epoch": 57.27947598253275,
+ "grad_norm": 0.01788361556828022,
+ "learning_rate": 0.0006,
+ "loss": 4.524880409240723,
+ "step": 4124
+ },
+ {
+ "epoch": 57.29344978165939,
+ "grad_norm": 0.01591576263308525,
+ "learning_rate": 0.0006,
+ "loss": 4.44764518737793,
+ "step": 4125
+ },
+ {
+ "epoch": 57.30742358078603,
+ "grad_norm": 0.015452570281922817,
+ "learning_rate": 0.0006,
+ "loss": 4.46844482421875,
+ "step": 4126
+ },
+ {
+ "epoch": 57.32139737991266,
+ "grad_norm": 0.016404492780566216,
+ "learning_rate": 0.0006,
+ "loss": 4.443678855895996,
+ "step": 4127
+ },
+ {
+ "epoch": 57.3353711790393,
+ "grad_norm": 0.019967148080468178,
+ "learning_rate": 0.0006,
+ "loss": 4.430230140686035,
+ "step": 4128
+ },
+ {
+ "epoch": 57.34934497816594,
+ "grad_norm": 0.02594015561044216,
+ "learning_rate": 0.0006,
+ "loss": 4.47061824798584,
+ "step": 4129
+ },
+ {
+ "epoch": 57.36331877729258,
+ "grad_norm": 0.031096890568733215,
+ "learning_rate": 0.0006,
+ "loss": 4.462275981903076,
+ "step": 4130
+ },
+ {
+ "epoch": 57.377292576419215,
+ "grad_norm": 0.02834898792207241,
+ "learning_rate": 0.0006,
+ "loss": 4.41143798828125,
+ "step": 4131
+ },
+ {
+ "epoch": 57.391266375545854,
+ "grad_norm": 0.023621153086423874,
+ "learning_rate": 0.0006,
+ "loss": 4.562548637390137,
+ "step": 4132
+ },
+ {
+ "epoch": 57.40524017467249,
+ "grad_norm": 0.02277296595275402,
+ "learning_rate": 0.0006,
+ "loss": 4.456247329711914,
+ "step": 4133
+ },
+ {
+ "epoch": 57.419213973799124,
+ "grad_norm": 0.019868768751621246,
+ "learning_rate": 0.0006,
+ "loss": 4.524300575256348,
+ "step": 4134
+ },
+ {
+ "epoch": 57.43318777292576,
+ "grad_norm": 0.01887078955769539,
+ "learning_rate": 0.0006,
+ "loss": 4.540097236633301,
+ "step": 4135
+ },
+ {
+ "epoch": 57.4471615720524,
+ "grad_norm": 0.024268802255392075,
+ "learning_rate": 0.0006,
+ "loss": 4.397233963012695,
+ "step": 4136
+ },
+ {
+ "epoch": 57.46113537117904,
+ "grad_norm": 0.025739721953868866,
+ "learning_rate": 0.0006,
+ "loss": 4.354194641113281,
+ "step": 4137
+ },
+ {
+ "epoch": 57.47510917030568,
+ "grad_norm": 0.02442290261387825,
+ "learning_rate": 0.0006,
+ "loss": 4.303822994232178,
+ "step": 4138
+ },
+ {
+ "epoch": 57.48908296943232,
+ "grad_norm": 0.022215209901332855,
+ "learning_rate": 0.0006,
+ "loss": 4.4870452880859375,
+ "step": 4139
+ },
+ {
+ "epoch": 57.50305676855895,
+ "grad_norm": 0.01789158768951893,
+ "learning_rate": 0.0006,
+ "loss": 4.343809127807617,
+ "step": 4140
+ },
+ {
+ "epoch": 57.51703056768559,
+ "grad_norm": 0.020381739363074303,
+ "learning_rate": 0.0006,
+ "loss": 4.53312873840332,
+ "step": 4141
+ },
+ {
+ "epoch": 57.531004366812226,
+ "grad_norm": 0.019585080444812775,
+ "learning_rate": 0.0006,
+ "loss": 4.510162353515625,
+ "step": 4142
+ },
+ {
+ "epoch": 57.544978165938865,
+ "grad_norm": 0.01985127478837967,
+ "learning_rate": 0.0006,
+ "loss": 4.3654022216796875,
+ "step": 4143
+ },
+ {
+ "epoch": 57.5589519650655,
+ "grad_norm": 0.01962905190885067,
+ "learning_rate": 0.0006,
+ "loss": 4.560830116271973,
+ "step": 4144
+ },
+ {
+ "epoch": 57.57292576419214,
+ "grad_norm": 0.02679629437625408,
+ "learning_rate": 0.0006,
+ "loss": 4.300804138183594,
+ "step": 4145
+ },
+ {
+ "epoch": 57.58689956331878,
+ "grad_norm": 0.039830856025218964,
+ "learning_rate": 0.0006,
+ "loss": 4.54150915145874,
+ "step": 4146
+ },
+ {
+ "epoch": 57.60087336244541,
+ "grad_norm": 0.049527477473020554,
+ "learning_rate": 0.0006,
+ "loss": 4.405792236328125,
+ "step": 4147
+ },
+ {
+ "epoch": 57.61484716157205,
+ "grad_norm": 0.042966946959495544,
+ "learning_rate": 0.0006,
+ "loss": 4.500643730163574,
+ "step": 4148
+ },
+ {
+ "epoch": 57.62882096069869,
+ "grad_norm": 0.08999177813529968,
+ "learning_rate": 0.0006,
+ "loss": 4.504790782928467,
+ "step": 4149
+ },
+ {
+ "epoch": 57.64279475982533,
+ "grad_norm": 0.44693124294281006,
+ "learning_rate": 0.0006,
+ "loss": 4.845344066619873,
+ "step": 4150
+ },
+ {
+ "epoch": 57.65676855895197,
+ "grad_norm": 0.7143144011497498,
+ "learning_rate": 0.0006,
+ "loss": 6.907845497131348,
+ "step": 4151
+ },
+ {
+ "epoch": 57.670742358078606,
+ "grad_norm": 0.25239595770835876,
+ "learning_rate": 0.0006,
+ "loss": 7.082579612731934,
+ "step": 4152
+ },
+ {
+ "epoch": 57.68471615720524,
+ "grad_norm": 0.26550790667533875,
+ "learning_rate": 0.0006,
+ "loss": 7.50180196762085,
+ "step": 4153
+ },
+ {
+ "epoch": 57.698689956331876,
+ "grad_norm": 0.14735378324985504,
+ "learning_rate": 0.0006,
+ "loss": 7.214590072631836,
+ "step": 4154
+ },
+ {
+ "epoch": 57.712663755458514,
+ "grad_norm": 0.17685334384441376,
+ "learning_rate": 0.0006,
+ "loss": 6.861320495605469,
+ "step": 4155
+ },
+ {
+ "epoch": 57.72663755458515,
+ "grad_norm": 0.18231339752674103,
+ "learning_rate": 0.0006,
+ "loss": 6.791839599609375,
+ "step": 4156
+ },
+ {
+ "epoch": 57.74061135371179,
+ "grad_norm": 0.09230359643697739,
+ "learning_rate": 0.0006,
+ "loss": 6.760223388671875,
+ "step": 4157
+ },
+ {
+ "epoch": 57.75458515283843,
+ "grad_norm": 0.06918249279260635,
+ "learning_rate": 0.0006,
+ "loss": 6.532999515533447,
+ "step": 4158
+ },
+ {
+ "epoch": 57.76855895196506,
+ "grad_norm": 0.0901007279753685,
+ "learning_rate": 0.0006,
+ "loss": 6.52907133102417,
+ "step": 4159
+ },
+ {
+ "epoch": 57.7825327510917,
+ "grad_norm": 0.05678229779005051,
+ "learning_rate": 0.0006,
+ "loss": 6.295290470123291,
+ "step": 4160
+ },
+ {
+ "epoch": 57.79650655021834,
+ "grad_norm": 0.04450210556387901,
+ "learning_rate": 0.0006,
+ "loss": 6.14900541305542,
+ "step": 4161
+ },
+ {
+ "epoch": 57.81048034934498,
+ "grad_norm": 0.049191512167453766,
+ "learning_rate": 0.0006,
+ "loss": 6.19650411605835,
+ "step": 4162
+ },
+ {
+ "epoch": 57.82445414847162,
+ "grad_norm": 0.03908967226743698,
+ "learning_rate": 0.0006,
+ "loss": 6.054948806762695,
+ "step": 4163
+ },
+ {
+ "epoch": 57.838427947598255,
+ "grad_norm": 0.04023387283086777,
+ "learning_rate": 0.0006,
+ "loss": 6.034873962402344,
+ "step": 4164
+ },
+ {
+ "epoch": 57.852401746724894,
+ "grad_norm": 0.03304268419742584,
+ "learning_rate": 0.0006,
+ "loss": 5.769852161407471,
+ "step": 4165
+ },
+ {
+ "epoch": 57.866375545851525,
+ "grad_norm": 0.03432450816035271,
+ "learning_rate": 0.0006,
+ "loss": 5.788785457611084,
+ "step": 4166
+ },
+ {
+ "epoch": 57.880349344978164,
+ "grad_norm": 0.029965840280056,
+ "learning_rate": 0.0006,
+ "loss": 5.815243721008301,
+ "step": 4167
+ },
+ {
+ "epoch": 57.8943231441048,
+ "grad_norm": 0.03637698292732239,
+ "learning_rate": 0.0006,
+ "loss": 5.7204060554504395,
+ "step": 4168
+ },
+ {
+ "epoch": 57.90829694323144,
+ "grad_norm": 0.03477517142891884,
+ "learning_rate": 0.0006,
+ "loss": 5.605930328369141,
+ "step": 4169
+ },
+ {
+ "epoch": 57.92227074235808,
+ "grad_norm": 0.05061859264969826,
+ "learning_rate": 0.0006,
+ "loss": 5.547054767608643,
+ "step": 4170
+ },
+ {
+ "epoch": 57.93624454148472,
+ "grad_norm": 0.06299655884504318,
+ "learning_rate": 0.0006,
+ "loss": 5.552166938781738,
+ "step": 4171
+ },
+ {
+ "epoch": 57.95021834061135,
+ "grad_norm": 0.04215948283672333,
+ "learning_rate": 0.0006,
+ "loss": 5.375457763671875,
+ "step": 4172
+ },
+ {
+ "epoch": 57.96419213973799,
+ "grad_norm": 0.0348566472530365,
+ "learning_rate": 0.0006,
+ "loss": 5.418606281280518,
+ "step": 4173
+ },
+ {
+ "epoch": 57.97816593886463,
+ "grad_norm": 0.03807530924677849,
+ "learning_rate": 0.0006,
+ "loss": 5.403156280517578,
+ "step": 4174
+ },
+ {
+ "epoch": 57.992139737991266,
+ "grad_norm": 0.02989993803203106,
+ "learning_rate": 0.0006,
+ "loss": 5.355951309204102,
+ "step": 4175
+ },
+ {
+ "epoch": 58.0,
+ "grad_norm": 0.034300558269023895,
+ "learning_rate": 0.0006,
+ "loss": 5.44920539855957,
+ "step": 4176
+ },
+ {
+ "epoch": 58.0,
+ "eval_loss": 5.416731834411621,
+ "eval_runtime": 56.7934,
+ "eval_samples_per_second": 42.998,
+ "eval_steps_per_second": 1.356,
+ "step": 4176
+ },
+ {
+ "epoch": 58.01397379912664,
+ "grad_norm": 0.033961448818445206,
+ "learning_rate": 0.0006,
+ "loss": 5.250707626342773,
+ "step": 4177
+ },
+ {
+ "epoch": 58.02794759825328,
+ "grad_norm": 0.027343502268195152,
+ "learning_rate": 0.0006,
+ "loss": 5.149238109588623,
+ "step": 4178
+ },
+ {
+ "epoch": 58.041921397379916,
+ "grad_norm": 0.03033650480210781,
+ "learning_rate": 0.0006,
+ "loss": 5.072463512420654,
+ "step": 4179
+ },
+ {
+ "epoch": 58.05589519650655,
+ "grad_norm": 0.028262868523597717,
+ "learning_rate": 0.0006,
+ "loss": 5.109417915344238,
+ "step": 4180
+ },
+ {
+ "epoch": 58.069868995633186,
+ "grad_norm": 0.02373034693300724,
+ "learning_rate": 0.0006,
+ "loss": 5.082464218139648,
+ "step": 4181
+ },
+ {
+ "epoch": 58.083842794759825,
+ "grad_norm": 0.025879688560962677,
+ "learning_rate": 0.0006,
+ "loss": 5.056792259216309,
+ "step": 4182
+ },
+ {
+ "epoch": 58.09781659388646,
+ "grad_norm": 0.026252275332808495,
+ "learning_rate": 0.0006,
+ "loss": 5.064239978790283,
+ "step": 4183
+ },
+ {
+ "epoch": 58.1117903930131,
+ "grad_norm": 0.028177792206406593,
+ "learning_rate": 0.0006,
+ "loss": 4.922310829162598,
+ "step": 4184
+ },
+ {
+ "epoch": 58.12576419213974,
+ "grad_norm": 0.02630820870399475,
+ "learning_rate": 0.0006,
+ "loss": 4.992884635925293,
+ "step": 4185
+ },
+ {
+ "epoch": 58.13973799126637,
+ "grad_norm": 0.024890903383493423,
+ "learning_rate": 0.0006,
+ "loss": 4.910036563873291,
+ "step": 4186
+ },
+ {
+ "epoch": 58.15371179039301,
+ "grad_norm": 0.02822992391884327,
+ "learning_rate": 0.0006,
+ "loss": 4.866640090942383,
+ "step": 4187
+ },
+ {
+ "epoch": 58.16768558951965,
+ "grad_norm": 0.03253607079386711,
+ "learning_rate": 0.0006,
+ "loss": 4.9112958908081055,
+ "step": 4188
+ },
+ {
+ "epoch": 58.18165938864629,
+ "grad_norm": 0.03765944764018059,
+ "learning_rate": 0.0006,
+ "loss": 4.774468421936035,
+ "step": 4189
+ },
+ {
+ "epoch": 58.19563318777293,
+ "grad_norm": 0.039874881505966187,
+ "learning_rate": 0.0006,
+ "loss": 4.831678867340088,
+ "step": 4190
+ },
+ {
+ "epoch": 58.209606986899566,
+ "grad_norm": 0.034595925360918045,
+ "learning_rate": 0.0006,
+ "loss": 4.8243794441223145,
+ "step": 4191
+ },
+ {
+ "epoch": 58.223580786026204,
+ "grad_norm": 0.04840189591050148,
+ "learning_rate": 0.0006,
+ "loss": 4.854156494140625,
+ "step": 4192
+ },
+ {
+ "epoch": 58.237554585152836,
+ "grad_norm": 0.06358753889799118,
+ "learning_rate": 0.0006,
+ "loss": 4.836430549621582,
+ "step": 4193
+ },
+ {
+ "epoch": 58.251528384279474,
+ "grad_norm": 0.0507555715739727,
+ "learning_rate": 0.0006,
+ "loss": 4.88357400894165,
+ "step": 4194
+ },
+ {
+ "epoch": 58.26550218340611,
+ "grad_norm": 0.04384690150618553,
+ "learning_rate": 0.0006,
+ "loss": 4.774049282073975,
+ "step": 4195
+ },
+ {
+ "epoch": 58.27947598253275,
+ "grad_norm": 0.03246928006410599,
+ "learning_rate": 0.0006,
+ "loss": 4.686605930328369,
+ "step": 4196
+ },
+ {
+ "epoch": 58.29344978165939,
+ "grad_norm": 0.028878789395093918,
+ "learning_rate": 0.0006,
+ "loss": 4.796070575714111,
+ "step": 4197
+ },
+ {
+ "epoch": 58.30742358078603,
+ "grad_norm": 0.028175361454486847,
+ "learning_rate": 0.0006,
+ "loss": 4.728490829467773,
+ "step": 4198
+ },
+ {
+ "epoch": 58.32139737991266,
+ "grad_norm": 0.029428014531731606,
+ "learning_rate": 0.0006,
+ "loss": 4.7159624099731445,
+ "step": 4199
+ },
+ {
+ "epoch": 58.3353711790393,
+ "grad_norm": 0.031227827072143555,
+ "learning_rate": 0.0006,
+ "loss": 4.702625274658203,
+ "step": 4200
+ },
+ {
+ "epoch": 58.34934497816594,
+ "grad_norm": 0.028843428939580917,
+ "learning_rate": 0.0006,
+ "loss": 4.811799049377441,
+ "step": 4201
+ },
+ {
+ "epoch": 58.36331877729258,
+ "grad_norm": 0.026780391111969948,
+ "learning_rate": 0.0006,
+ "loss": 4.81662654876709,
+ "step": 4202
+ },
+ {
+ "epoch": 58.377292576419215,
+ "grad_norm": 0.019699757918715477,
+ "learning_rate": 0.0006,
+ "loss": 4.713685989379883,
+ "step": 4203
+ },
+ {
+ "epoch": 58.391266375545854,
+ "grad_norm": 0.024389250203967094,
+ "learning_rate": 0.0006,
+ "loss": 4.72354793548584,
+ "step": 4204
+ },
+ {
+ "epoch": 58.40524017467249,
+ "grad_norm": 0.02218274027109146,
+ "learning_rate": 0.0006,
+ "loss": 4.674125671386719,
+ "step": 4205
+ },
+ {
+ "epoch": 58.419213973799124,
+ "grad_norm": 0.02041488140821457,
+ "learning_rate": 0.0006,
+ "loss": 4.7909698486328125,
+ "step": 4206
+ },
+ {
+ "epoch": 58.43318777292576,
+ "grad_norm": 0.018561935052275658,
+ "learning_rate": 0.0006,
+ "loss": 4.808770179748535,
+ "step": 4207
+ },
+ {
+ "epoch": 58.4471615720524,
+ "grad_norm": 0.01708846725523472,
+ "learning_rate": 0.0006,
+ "loss": 4.701626777648926,
+ "step": 4208
+ },
+ {
+ "epoch": 58.46113537117904,
+ "grad_norm": 0.0222734697163105,
+ "learning_rate": 0.0006,
+ "loss": 4.505410194396973,
+ "step": 4209
+ },
+ {
+ "epoch": 58.47510917030568,
+ "grad_norm": 0.03197057917714119,
+ "learning_rate": 0.0006,
+ "loss": 4.610360145568848,
+ "step": 4210
+ },
+ {
+ "epoch": 58.48908296943232,
+ "grad_norm": 0.06854520738124847,
+ "learning_rate": 0.0006,
+ "loss": 4.625649452209473,
+ "step": 4211
+ },
+ {
+ "epoch": 58.50305676855895,
+ "grad_norm": 0.09981521219015121,
+ "learning_rate": 0.0006,
+ "loss": 4.832150459289551,
+ "step": 4212
+ },
+ {
+ "epoch": 58.51703056768559,
+ "grad_norm": 0.046356040984392166,
+ "learning_rate": 0.0006,
+ "loss": 4.5641632080078125,
+ "step": 4213
+ },
+ {
+ "epoch": 58.531004366812226,
+ "grad_norm": 0.03882667422294617,
+ "learning_rate": 0.0006,
+ "loss": 4.666974067687988,
+ "step": 4214
+ },
+ {
+ "epoch": 58.544978165938865,
+ "grad_norm": 0.033203233033418655,
+ "learning_rate": 0.0006,
+ "loss": 4.629413604736328,
+ "step": 4215
+ },
+ {
+ "epoch": 58.5589519650655,
+ "grad_norm": 0.02806735597550869,
+ "learning_rate": 0.0006,
+ "loss": 4.763245582580566,
+ "step": 4216
+ },
+ {
+ "epoch": 58.57292576419214,
+ "grad_norm": 0.026728278025984764,
+ "learning_rate": 0.0006,
+ "loss": 4.629597187042236,
+ "step": 4217
+ },
+ {
+ "epoch": 58.58689956331878,
+ "grad_norm": 0.024380959570407867,
+ "learning_rate": 0.0006,
+ "loss": 4.633614540100098,
+ "step": 4218
+ },
+ {
+ "epoch": 58.60087336244541,
+ "grad_norm": 0.022312544286251068,
+ "learning_rate": 0.0006,
+ "loss": 4.671379566192627,
+ "step": 4219
+ },
+ {
+ "epoch": 58.61484716157205,
+ "grad_norm": 0.022049568593502045,
+ "learning_rate": 0.0006,
+ "loss": 4.661557197570801,
+ "step": 4220
+ },
+ {
+ "epoch": 58.62882096069869,
+ "grad_norm": 0.02119818702340126,
+ "learning_rate": 0.0006,
+ "loss": 4.688562393188477,
+ "step": 4221
+ },
+ {
+ "epoch": 58.64279475982533,
+ "grad_norm": 0.019769301638007164,
+ "learning_rate": 0.0006,
+ "loss": 4.607748031616211,
+ "step": 4222
+ },
+ {
+ "epoch": 58.65676855895197,
+ "grad_norm": 0.02124079130589962,
+ "learning_rate": 0.0006,
+ "loss": 4.570640563964844,
+ "step": 4223
+ },
+ {
+ "epoch": 58.670742358078606,
+ "grad_norm": 0.016344040632247925,
+ "learning_rate": 0.0006,
+ "loss": 4.676113128662109,
+ "step": 4224
+ },
+ {
+ "epoch": 58.68471615720524,
+ "grad_norm": 0.01735256239771843,
+ "learning_rate": 0.0006,
+ "loss": 4.62088680267334,
+ "step": 4225
+ },
+ {
+ "epoch": 58.698689956331876,
+ "grad_norm": 0.019323458895087242,
+ "learning_rate": 0.0006,
+ "loss": 4.579405307769775,
+ "step": 4226
+ },
+ {
+ "epoch": 58.712663755458514,
+ "grad_norm": 0.01760544814169407,
+ "learning_rate": 0.0006,
+ "loss": 4.625407695770264,
+ "step": 4227
+ },
+ {
+ "epoch": 58.72663755458515,
+ "grad_norm": 0.015593250282108784,
+ "learning_rate": 0.0006,
+ "loss": 4.6326117515563965,
+ "step": 4228
+ },
+ {
+ "epoch": 58.74061135371179,
+ "grad_norm": 0.014357523061335087,
+ "learning_rate": 0.0006,
+ "loss": 4.6869401931762695,
+ "step": 4229
+ },
+ {
+ "epoch": 58.75458515283843,
+ "grad_norm": 0.01604357920587063,
+ "learning_rate": 0.0006,
+ "loss": 4.690784931182861,
+ "step": 4230
+ },
+ {
+ "epoch": 58.76855895196506,
+ "grad_norm": 0.01673845760524273,
+ "learning_rate": 0.0006,
+ "loss": 4.513410568237305,
+ "step": 4231
+ },
+ {
+ "epoch": 58.7825327510917,
+ "grad_norm": 0.013657779432833195,
+ "learning_rate": 0.0006,
+ "loss": 4.637057781219482,
+ "step": 4232
+ },
+ {
+ "epoch": 58.79650655021834,
+ "grad_norm": 0.013373114168643951,
+ "learning_rate": 0.0006,
+ "loss": 4.719967842102051,
+ "step": 4233
+ },
+ {
+ "epoch": 58.81048034934498,
+ "grad_norm": 0.016345568001270294,
+ "learning_rate": 0.0006,
+ "loss": 4.604109764099121,
+ "step": 4234
+ },
+ {
+ "epoch": 58.82445414847162,
+ "grad_norm": 0.01573033444583416,
+ "learning_rate": 0.0006,
+ "loss": 4.6471428871154785,
+ "step": 4235
+ },
+ {
+ "epoch": 58.838427947598255,
+ "grad_norm": 0.01517449039965868,
+ "learning_rate": 0.0006,
+ "loss": 4.722684860229492,
+ "step": 4236
+ },
+ {
+ "epoch": 58.852401746724894,
+ "grad_norm": 0.0145102022215724,
+ "learning_rate": 0.0006,
+ "loss": 4.662420272827148,
+ "step": 4237
+ },
+ {
+ "epoch": 58.866375545851525,
+ "grad_norm": 0.01457217987626791,
+ "learning_rate": 0.0006,
+ "loss": 4.585169792175293,
+ "step": 4238
+ },
+ {
+ "epoch": 58.880349344978164,
+ "grad_norm": 0.01406773366034031,
+ "learning_rate": 0.0006,
+ "loss": 4.5413665771484375,
+ "step": 4239
+ },
+ {
+ "epoch": 58.8943231441048,
+ "grad_norm": 0.012653871439397335,
+ "learning_rate": 0.0006,
+ "loss": 4.553636074066162,
+ "step": 4240
+ },
+ {
+ "epoch": 58.90829694323144,
+ "grad_norm": 0.01261200476437807,
+ "learning_rate": 0.0006,
+ "loss": 4.534193992614746,
+ "step": 4241
+ },
+ {
+ "epoch": 58.92227074235808,
+ "grad_norm": 0.012955864891409874,
+ "learning_rate": 0.0006,
+ "loss": 4.697248458862305,
+ "step": 4242
+ },
+ {
+ "epoch": 58.93624454148472,
+ "grad_norm": 0.012134749442338943,
+ "learning_rate": 0.0006,
+ "loss": 4.557095527648926,
+ "step": 4243
+ },
+ {
+ "epoch": 58.95021834061135,
+ "grad_norm": 0.012164677493274212,
+ "learning_rate": 0.0006,
+ "loss": 4.623579025268555,
+ "step": 4244
+ },
+ {
+ "epoch": 58.96419213973799,
+ "grad_norm": 0.013802947476506233,
+ "learning_rate": 0.0006,
+ "loss": 4.562838077545166,
+ "step": 4245
+ },
+ {
+ "epoch": 58.97816593886463,
+ "grad_norm": 0.011420476250350475,
+ "learning_rate": 0.0006,
+ "loss": 4.62265682220459,
+ "step": 4246
+ },
+ {
+ "epoch": 58.992139737991266,
+ "grad_norm": 0.01169576682150364,
+ "learning_rate": 0.0006,
+ "loss": 4.727449417114258,
+ "step": 4247
+ },
+ {
+ "epoch": 59.0,
+ "grad_norm": 0.013577710837125778,
+ "learning_rate": 0.0006,
+ "loss": 4.536325454711914,
+ "step": 4248
+ },
+ {
+ "epoch": 59.0,
+ "eval_loss": 4.86527681350708,
+ "eval_runtime": 60.4463,
+ "eval_samples_per_second": 40.399,
+ "eval_steps_per_second": 1.274,
+ "step": 4248
+ },
+ {
+ "epoch": 59.01397379912664,
+ "grad_norm": 0.015251655131578445,
+ "learning_rate": 0.0006,
+ "loss": 4.634598731994629,
+ "step": 4249
+ },
+ {
+ "epoch": 59.02794759825328,
+ "grad_norm": 0.017414938658475876,
+ "learning_rate": 0.0006,
+ "loss": 4.521329879760742,
+ "step": 4250
+ },
+ {
+ "epoch": 59.041921397379916,
+ "grad_norm": 0.01656588353216648,
+ "learning_rate": 0.0006,
+ "loss": 4.446192264556885,
+ "step": 4251
+ },
+ {
+ "epoch": 59.05589519650655,
+ "grad_norm": 0.012038925662636757,
+ "learning_rate": 0.0006,
+ "loss": 4.523855209350586,
+ "step": 4252
+ },
+ {
+ "epoch": 59.069868995633186,
+ "grad_norm": 0.014403033070266247,
+ "learning_rate": 0.0006,
+ "loss": 4.561079978942871,
+ "step": 4253
+ },
+ {
+ "epoch": 59.083842794759825,
+ "grad_norm": 0.016243597492575645,
+ "learning_rate": 0.0006,
+ "loss": 4.673602104187012,
+ "step": 4254
+ },
+ {
+ "epoch": 59.09781659388646,
+ "grad_norm": 0.016492877155542374,
+ "learning_rate": 0.0006,
+ "loss": 4.564824104309082,
+ "step": 4255
+ },
+ {
+ "epoch": 59.1117903930131,
+ "grad_norm": 0.015980370342731476,
+ "learning_rate": 0.0006,
+ "loss": 4.339249134063721,
+ "step": 4256
+ },
+ {
+ "epoch": 59.12576419213974,
+ "grad_norm": 0.019687356427311897,
+ "learning_rate": 0.0006,
+ "loss": 4.481834888458252,
+ "step": 4257
+ },
+ {
+ "epoch": 59.13973799126637,
+ "grad_norm": 0.02983798086643219,
+ "learning_rate": 0.0006,
+ "loss": 4.4495697021484375,
+ "step": 4258
+ },
+ {
+ "epoch": 59.15371179039301,
+ "grad_norm": 0.04600981995463371,
+ "learning_rate": 0.0006,
+ "loss": 4.567838668823242,
+ "step": 4259
+ },
+ {
+ "epoch": 59.16768558951965,
+ "grad_norm": 0.04852001741528511,
+ "learning_rate": 0.0006,
+ "loss": 4.594701766967773,
+ "step": 4260
+ },
+ {
+ "epoch": 59.18165938864629,
+ "grad_norm": 0.03429504856467247,
+ "learning_rate": 0.0006,
+ "loss": 4.573735237121582,
+ "step": 4261
+ },
+ {
+ "epoch": 59.19563318777293,
+ "grad_norm": 0.029345160350203514,
+ "learning_rate": 0.0006,
+ "loss": 4.488258361816406,
+ "step": 4262
+ },
+ {
+ "epoch": 59.209606986899566,
+ "grad_norm": 0.01961623504757881,
+ "learning_rate": 0.0006,
+ "loss": 4.545222282409668,
+ "step": 4263
+ },
+ {
+ "epoch": 59.223580786026204,
+ "grad_norm": 0.020406130701303482,
+ "learning_rate": 0.0006,
+ "loss": 4.471898078918457,
+ "step": 4264
+ },
+ {
+ "epoch": 59.237554585152836,
+ "grad_norm": 0.020500272512435913,
+ "learning_rate": 0.0006,
+ "loss": 4.538337707519531,
+ "step": 4265
+ },
+ {
+ "epoch": 59.251528384279474,
+ "grad_norm": 0.020727017894387245,
+ "learning_rate": 0.0006,
+ "loss": 4.437284469604492,
+ "step": 4266
+ },
+ {
+ "epoch": 59.26550218340611,
+ "grad_norm": 0.03395998477935791,
+ "learning_rate": 0.0006,
+ "loss": 4.459124565124512,
+ "step": 4267
+ },
+ {
+ "epoch": 59.27947598253275,
+ "grad_norm": 0.04635605961084366,
+ "learning_rate": 0.0006,
+ "loss": 4.520228385925293,
+ "step": 4268
+ },
+ {
+ "epoch": 59.29344978165939,
+ "grad_norm": 0.04253006726503372,
+ "learning_rate": 0.0006,
+ "loss": 4.578673362731934,
+ "step": 4269
+ },
+ {
+ "epoch": 59.30742358078603,
+ "grad_norm": 0.02403509058058262,
+ "learning_rate": 0.0006,
+ "loss": 4.484639644622803,
+ "step": 4270
+ },
+ {
+ "epoch": 59.32139737991266,
+ "grad_norm": 0.02254190482199192,
+ "learning_rate": 0.0006,
+ "loss": 4.511314868927002,
+ "step": 4271
+ },
+ {
+ "epoch": 59.3353711790393,
+ "grad_norm": 0.021585190668702126,
+ "learning_rate": 0.0006,
+ "loss": 4.530239105224609,
+ "step": 4272
+ },
+ {
+ "epoch": 59.34934497816594,
+ "grad_norm": 0.021210316568613052,
+ "learning_rate": 0.0006,
+ "loss": 4.593514442443848,
+ "step": 4273
+ },
+ {
+ "epoch": 59.36331877729258,
+ "grad_norm": 0.019317107275128365,
+ "learning_rate": 0.0006,
+ "loss": 4.5591535568237305,
+ "step": 4274
+ },
+ {
+ "epoch": 59.377292576419215,
+ "grad_norm": 0.018059708178043365,
+ "learning_rate": 0.0006,
+ "loss": 4.4499592781066895,
+ "step": 4275
+ },
+ {
+ "epoch": 59.391266375545854,
+ "grad_norm": 0.017847446724772453,
+ "learning_rate": 0.0006,
+ "loss": 4.473999500274658,
+ "step": 4276
+ },
+ {
+ "epoch": 59.40524017467249,
+ "grad_norm": 0.014944756403565407,
+ "learning_rate": 0.0006,
+ "loss": 4.5934529304504395,
+ "step": 4277
+ },
+ {
+ "epoch": 59.419213973799124,
+ "grad_norm": 0.01663832925260067,
+ "learning_rate": 0.0006,
+ "loss": 4.504262924194336,
+ "step": 4278
+ },
+ {
+ "epoch": 59.43318777292576,
+ "grad_norm": 0.015571796335279942,
+ "learning_rate": 0.0006,
+ "loss": 4.512310981750488,
+ "step": 4279
+ },
+ {
+ "epoch": 59.4471615720524,
+ "grad_norm": 0.015896400436758995,
+ "learning_rate": 0.0006,
+ "loss": 4.452714920043945,
+ "step": 4280
+ },
+ {
+ "epoch": 59.46113537117904,
+ "grad_norm": 0.016010284423828125,
+ "learning_rate": 0.0006,
+ "loss": 4.6188063621521,
+ "step": 4281
+ },
+ {
+ "epoch": 59.47510917030568,
+ "grad_norm": 0.015075593255460262,
+ "learning_rate": 0.0006,
+ "loss": 4.477978229522705,
+ "step": 4282
+ },
+ {
+ "epoch": 59.48908296943232,
+ "grad_norm": 0.014698950573801994,
+ "learning_rate": 0.0006,
+ "loss": 4.544955730438232,
+ "step": 4283
+ },
+ {
+ "epoch": 59.50305676855895,
+ "grad_norm": 0.0136245833709836,
+ "learning_rate": 0.0006,
+ "loss": 4.530460357666016,
+ "step": 4284
+ },
+ {
+ "epoch": 59.51703056768559,
+ "grad_norm": 0.013118310831487179,
+ "learning_rate": 0.0006,
+ "loss": 4.614217758178711,
+ "step": 4285
+ },
+ {
+ "epoch": 59.531004366812226,
+ "grad_norm": 0.01295961532741785,
+ "learning_rate": 0.0006,
+ "loss": 4.43156099319458,
+ "step": 4286
+ },
+ {
+ "epoch": 59.544978165938865,
+ "grad_norm": 0.012429811991751194,
+ "learning_rate": 0.0006,
+ "loss": 4.517333030700684,
+ "step": 4287
+ },
+ {
+ "epoch": 59.5589519650655,
+ "grad_norm": 0.01261830423027277,
+ "learning_rate": 0.0006,
+ "loss": 4.569780349731445,
+ "step": 4288
+ },
+ {
+ "epoch": 59.57292576419214,
+ "grad_norm": 0.012394499965012074,
+ "learning_rate": 0.0006,
+ "loss": 4.5970869064331055,
+ "step": 4289
+ },
+ {
+ "epoch": 59.58689956331878,
+ "grad_norm": 0.01290759164839983,
+ "learning_rate": 0.0006,
+ "loss": 4.437100410461426,
+ "step": 4290
+ },
+ {
+ "epoch": 59.60087336244541,
+ "grad_norm": 0.012042169459164143,
+ "learning_rate": 0.0006,
+ "loss": 4.567404747009277,
+ "step": 4291
+ },
+ {
+ "epoch": 59.61484716157205,
+ "grad_norm": 0.012338520959019661,
+ "learning_rate": 0.0006,
+ "loss": 4.476193428039551,
+ "step": 4292
+ },
+ {
+ "epoch": 59.62882096069869,
+ "grad_norm": 0.01321258582174778,
+ "learning_rate": 0.0006,
+ "loss": 4.566680431365967,
+ "step": 4293
+ },
+ {
+ "epoch": 59.64279475982533,
+ "grad_norm": 0.013478122651576996,
+ "learning_rate": 0.0006,
+ "loss": 4.487398147583008,
+ "step": 4294
+ },
+ {
+ "epoch": 59.65676855895197,
+ "grad_norm": 0.01537603884935379,
+ "learning_rate": 0.0006,
+ "loss": 4.449337005615234,
+ "step": 4295
+ },
+ {
+ "epoch": 59.670742358078606,
+ "grad_norm": 0.017140565440058708,
+ "learning_rate": 0.0006,
+ "loss": 4.5174360275268555,
+ "step": 4296
+ },
+ {
+ "epoch": 59.68471615720524,
+ "grad_norm": 0.017931461334228516,
+ "learning_rate": 0.0006,
+ "loss": 4.522104740142822,
+ "step": 4297
+ },
+ {
+ "epoch": 59.698689956331876,
+ "grad_norm": 0.018037918955087662,
+ "learning_rate": 0.0006,
+ "loss": 4.496928691864014,
+ "step": 4298
+ },
+ {
+ "epoch": 59.712663755458514,
+ "grad_norm": 0.01684504561126232,
+ "learning_rate": 0.0006,
+ "loss": 4.5332417488098145,
+ "step": 4299
+ },
+ {
+ "epoch": 59.72663755458515,
+ "grad_norm": 0.016268383711576462,
+ "learning_rate": 0.0006,
+ "loss": 4.468893051147461,
+ "step": 4300
+ },
+ {
+ "epoch": 59.74061135371179,
+ "grad_norm": 0.014859385788440704,
+ "learning_rate": 0.0006,
+ "loss": 4.551092147827148,
+ "step": 4301
+ },
+ {
+ "epoch": 59.75458515283843,
+ "grad_norm": 0.013589047826826572,
+ "learning_rate": 0.0006,
+ "loss": 4.470579147338867,
+ "step": 4302
+ },
+ {
+ "epoch": 59.76855895196506,
+ "grad_norm": 0.015704551711678505,
+ "learning_rate": 0.0006,
+ "loss": 4.330068111419678,
+ "step": 4303
+ },
+ {
+ "epoch": 59.7825327510917,
+ "grad_norm": 0.017869921401143074,
+ "learning_rate": 0.0006,
+ "loss": 4.428255081176758,
+ "step": 4304
+ },
+ {
+ "epoch": 59.79650655021834,
+ "grad_norm": 0.019828537479043007,
+ "learning_rate": 0.0006,
+ "loss": 4.496023178100586,
+ "step": 4305
+ },
+ {
+ "epoch": 59.81048034934498,
+ "grad_norm": 0.015856236219406128,
+ "learning_rate": 0.0006,
+ "loss": 4.427152156829834,
+ "step": 4306
+ },
+ {
+ "epoch": 59.82445414847162,
+ "grad_norm": 0.013349304907023907,
+ "learning_rate": 0.0006,
+ "loss": 4.417333602905273,
+ "step": 4307
+ },
+ {
+ "epoch": 59.838427947598255,
+ "grad_norm": 0.014111637137830257,
+ "learning_rate": 0.0006,
+ "loss": 4.4507575035095215,
+ "step": 4308
+ },
+ {
+ "epoch": 59.852401746724894,
+ "grad_norm": 0.015186597593128681,
+ "learning_rate": 0.0006,
+ "loss": 4.375116348266602,
+ "step": 4309
+ },
+ {
+ "epoch": 59.866375545851525,
+ "grad_norm": 0.01604730449616909,
+ "learning_rate": 0.0006,
+ "loss": 4.604518890380859,
+ "step": 4310
+ },
+ {
+ "epoch": 59.880349344978164,
+ "grad_norm": 0.015308464877307415,
+ "learning_rate": 0.0006,
+ "loss": 4.507232189178467,
+ "step": 4311
+ },
+ {
+ "epoch": 59.8943231441048,
+ "grad_norm": 0.014017206616699696,
+ "learning_rate": 0.0006,
+ "loss": 4.398487091064453,
+ "step": 4312
+ },
+ {
+ "epoch": 59.90829694323144,
+ "grad_norm": 0.012911394238471985,
+ "learning_rate": 0.0006,
+ "loss": 4.567872047424316,
+ "step": 4313
+ },
+ {
+ "epoch": 59.92227074235808,
+ "grad_norm": 0.012775142677128315,
+ "learning_rate": 0.0006,
+ "loss": 4.562105655670166,
+ "step": 4314
+ },
+ {
+ "epoch": 59.93624454148472,
+ "grad_norm": 0.014750408008694649,
+ "learning_rate": 0.0006,
+ "loss": 4.456263542175293,
+ "step": 4315
+ },
+ {
+ "epoch": 59.95021834061135,
+ "grad_norm": 0.016726011410355568,
+ "learning_rate": 0.0006,
+ "loss": 4.360722064971924,
+ "step": 4316
+ },
+ {
+ "epoch": 59.96419213973799,
+ "grad_norm": 0.01734776981174946,
+ "learning_rate": 0.0006,
+ "loss": 4.572544574737549,
+ "step": 4317
+ },
+ {
+ "epoch": 59.97816593886463,
+ "grad_norm": 0.01822415366768837,
+ "learning_rate": 0.0006,
+ "loss": 4.490501880645752,
+ "step": 4318
+ },
+ {
+ "epoch": 59.992139737991266,
+ "grad_norm": 0.014638577587902546,
+ "learning_rate": 0.0006,
+ "loss": 4.440951824188232,
+ "step": 4319
+ },
+ {
+ "epoch": 60.0,
+ "grad_norm": 0.01255202479660511,
+ "learning_rate": 0.0006,
+ "loss": 4.559123516082764,
+ "step": 4320
+ },
+ {
+ "epoch": 60.0,
+ "eval_loss": 4.741089344024658,
+ "eval_runtime": 56.8997,
+ "eval_samples_per_second": 42.918,
+ "eval_steps_per_second": 1.353,
+ "step": 4320
+ },
+ {
+ "epoch": 60.01397379912664,
+ "grad_norm": 0.016548393294215202,
+ "learning_rate": 0.0006,
+ "loss": 4.442024230957031,
+ "step": 4321
+ },
+ {
+ "epoch": 60.02794759825328,
+ "grad_norm": 0.019759127870202065,
+ "learning_rate": 0.0006,
+ "loss": 4.394334316253662,
+ "step": 4322
+ },
+ {
+ "epoch": 60.041921397379916,
+ "grad_norm": 0.018325170502066612,
+ "learning_rate": 0.0006,
+ "loss": 4.318270683288574,
+ "step": 4323
+ },
+ {
+ "epoch": 60.05589519650655,
+ "grad_norm": 0.015289463102817535,
+ "learning_rate": 0.0006,
+ "loss": 4.318228721618652,
+ "step": 4324
+ },
+ {
+ "epoch": 60.069868995633186,
+ "grad_norm": 0.015424872748553753,
+ "learning_rate": 0.0006,
+ "loss": 4.3605499267578125,
+ "step": 4325
+ },
+ {
+ "epoch": 60.083842794759825,
+ "grad_norm": 0.014085669070482254,
+ "learning_rate": 0.0006,
+ "loss": 4.359585762023926,
+ "step": 4326
+ },
+ {
+ "epoch": 60.09781659388646,
+ "grad_norm": 0.014004210010170937,
+ "learning_rate": 0.0006,
+ "loss": 4.447783470153809,
+ "step": 4327
+ },
+ {
+ "epoch": 60.1117903930131,
+ "grad_norm": 0.013514582999050617,
+ "learning_rate": 0.0006,
+ "loss": 4.4850239753723145,
+ "step": 4328
+ },
+ {
+ "epoch": 60.12576419213974,
+ "grad_norm": 0.013049394823610783,
+ "learning_rate": 0.0006,
+ "loss": 4.528675556182861,
+ "step": 4329
+ },
+ {
+ "epoch": 60.13973799126637,
+ "grad_norm": 0.01378058921545744,
+ "learning_rate": 0.0006,
+ "loss": 4.444823741912842,
+ "step": 4330
+ },
+ {
+ "epoch": 60.15371179039301,
+ "grad_norm": 0.013752233237028122,
+ "learning_rate": 0.0006,
+ "loss": 4.5205206871032715,
+ "step": 4331
+ },
+ {
+ "epoch": 60.16768558951965,
+ "grad_norm": 0.013431803323328495,
+ "learning_rate": 0.0006,
+ "loss": 4.547320365905762,
+ "step": 4332
+ },
+ {
+ "epoch": 60.18165938864629,
+ "grad_norm": 0.014544196426868439,
+ "learning_rate": 0.0006,
+ "loss": 4.517143249511719,
+ "step": 4333
+ },
+ {
+ "epoch": 60.19563318777293,
+ "grad_norm": 0.01612176187336445,
+ "learning_rate": 0.0006,
+ "loss": 4.517232894897461,
+ "step": 4334
+ },
+ {
+ "epoch": 60.209606986899566,
+ "grad_norm": 0.01803239807486534,
+ "learning_rate": 0.0006,
+ "loss": 4.377574920654297,
+ "step": 4335
+ },
+ {
+ "epoch": 60.223580786026204,
+ "grad_norm": 0.014871489256620407,
+ "learning_rate": 0.0006,
+ "loss": 4.538285255432129,
+ "step": 4336
+ },
+ {
+ "epoch": 60.237554585152836,
+ "grad_norm": 0.013927377760410309,
+ "learning_rate": 0.0006,
+ "loss": 4.386175632476807,
+ "step": 4337
+ },
+ {
+ "epoch": 60.251528384279474,
+ "grad_norm": 0.014757219702005386,
+ "learning_rate": 0.0006,
+ "loss": 4.528553009033203,
+ "step": 4338
+ },
+ {
+ "epoch": 60.26550218340611,
+ "grad_norm": 0.01433873176574707,
+ "learning_rate": 0.0006,
+ "loss": 4.476929187774658,
+ "step": 4339
+ },
+ {
+ "epoch": 60.27947598253275,
+ "grad_norm": 0.015768803656101227,
+ "learning_rate": 0.0006,
+ "loss": 4.50970458984375,
+ "step": 4340
+ },
+ {
+ "epoch": 60.29344978165939,
+ "grad_norm": 0.015651429072022438,
+ "learning_rate": 0.0006,
+ "loss": 4.427133083343506,
+ "step": 4341
+ },
+ {
+ "epoch": 60.30742358078603,
+ "grad_norm": 0.01673000678420067,
+ "learning_rate": 0.0006,
+ "loss": 4.447723865509033,
+ "step": 4342
+ },
+ {
+ "epoch": 60.32139737991266,
+ "grad_norm": 0.018481194972991943,
+ "learning_rate": 0.0006,
+ "loss": 4.559832572937012,
+ "step": 4343
+ },
+ {
+ "epoch": 60.3353711790393,
+ "grad_norm": 0.016731027513742447,
+ "learning_rate": 0.0006,
+ "loss": 4.470893859863281,
+ "step": 4344
+ },
+ {
+ "epoch": 60.34934497816594,
+ "grad_norm": 0.01504011545330286,
+ "learning_rate": 0.0006,
+ "loss": 4.483014106750488,
+ "step": 4345
+ },
+ {
+ "epoch": 60.36331877729258,
+ "grad_norm": 0.01602456159889698,
+ "learning_rate": 0.0006,
+ "loss": 4.548428058624268,
+ "step": 4346
+ },
+ {
+ "epoch": 60.377292576419215,
+ "grad_norm": 0.016027364879846573,
+ "learning_rate": 0.0006,
+ "loss": 4.426779747009277,
+ "step": 4347
+ },
+ {
+ "epoch": 60.391266375545854,
+ "grad_norm": 0.01873827911913395,
+ "learning_rate": 0.0006,
+ "loss": 4.353686332702637,
+ "step": 4348
+ },
+ {
+ "epoch": 60.40524017467249,
+ "grad_norm": 0.01907406374812126,
+ "learning_rate": 0.0006,
+ "loss": 4.502945423126221,
+ "step": 4349
+ },
+ {
+ "epoch": 60.419213973799124,
+ "grad_norm": 0.016083354130387306,
+ "learning_rate": 0.0006,
+ "loss": 4.431731700897217,
+ "step": 4350
+ },
+ {
+ "epoch": 60.43318777292576,
+ "grad_norm": 0.015293709933757782,
+ "learning_rate": 0.0006,
+ "loss": 4.480112552642822,
+ "step": 4351
+ },
+ {
+ "epoch": 60.4471615720524,
+ "grad_norm": 0.01688101328909397,
+ "learning_rate": 0.0006,
+ "loss": 4.50853157043457,
+ "step": 4352
+ },
+ {
+ "epoch": 60.46113537117904,
+ "grad_norm": 0.02094118855893612,
+ "learning_rate": 0.0006,
+ "loss": 4.359930038452148,
+ "step": 4353
+ },
+ {
+ "epoch": 60.47510917030568,
+ "grad_norm": 0.018188107758760452,
+ "learning_rate": 0.0006,
+ "loss": 4.549592018127441,
+ "step": 4354
+ },
+ {
+ "epoch": 60.48908296943232,
+ "grad_norm": 0.01462319865822792,
+ "learning_rate": 0.0006,
+ "loss": 4.49467658996582,
+ "step": 4355
+ },
+ {
+ "epoch": 60.50305676855895,
+ "grad_norm": 0.017508579418063164,
+ "learning_rate": 0.0006,
+ "loss": 4.398318767547607,
+ "step": 4356
+ },
+ {
+ "epoch": 60.51703056768559,
+ "grad_norm": 0.018200315535068512,
+ "learning_rate": 0.0006,
+ "loss": 4.44169282913208,
+ "step": 4357
+ },
+ {
+ "epoch": 60.531004366812226,
+ "grad_norm": 0.01717468351125717,
+ "learning_rate": 0.0006,
+ "loss": 4.481400489807129,
+ "step": 4358
+ },
+ {
+ "epoch": 60.544978165938865,
+ "grad_norm": 0.019329151138663292,
+ "learning_rate": 0.0006,
+ "loss": 4.578839302062988,
+ "step": 4359
+ },
+ {
+ "epoch": 60.5589519650655,
+ "grad_norm": 0.016109555959701538,
+ "learning_rate": 0.0006,
+ "loss": 4.461043357849121,
+ "step": 4360
+ },
+ {
+ "epoch": 60.57292576419214,
+ "grad_norm": 0.01494457945227623,
+ "learning_rate": 0.0006,
+ "loss": 4.505096912384033,
+ "step": 4361
+ },
+ {
+ "epoch": 60.58689956331878,
+ "grad_norm": 0.01729399710893631,
+ "learning_rate": 0.0006,
+ "loss": 4.366747856140137,
+ "step": 4362
+ },
+ {
+ "epoch": 60.60087336244541,
+ "grad_norm": 0.01705821603536606,
+ "learning_rate": 0.0006,
+ "loss": 4.42534065246582,
+ "step": 4363
+ },
+ {
+ "epoch": 60.61484716157205,
+ "grad_norm": 0.018391354009509087,
+ "learning_rate": 0.0006,
+ "loss": 4.350294589996338,
+ "step": 4364
+ },
+ {
+ "epoch": 60.62882096069869,
+ "grad_norm": 0.016989829018712044,
+ "learning_rate": 0.0006,
+ "loss": 4.45905876159668,
+ "step": 4365
+ },
+ {
+ "epoch": 60.64279475982533,
+ "grad_norm": 0.017192212864756584,
+ "learning_rate": 0.0006,
+ "loss": 4.526487350463867,
+ "step": 4366
+ },
+ {
+ "epoch": 60.65676855895197,
+ "grad_norm": 0.018091315403580666,
+ "learning_rate": 0.0006,
+ "loss": 4.309333324432373,
+ "step": 4367
+ },
+ {
+ "epoch": 60.670742358078606,
+ "grad_norm": 0.014051459729671478,
+ "learning_rate": 0.0006,
+ "loss": 4.338171005249023,
+ "step": 4368
+ },
+ {
+ "epoch": 60.68471615720524,
+ "grad_norm": 0.013166418299078941,
+ "learning_rate": 0.0006,
+ "loss": 4.409758567810059,
+ "step": 4369
+ },
+ {
+ "epoch": 60.698689956331876,
+ "grad_norm": 0.014260428957641125,
+ "learning_rate": 0.0006,
+ "loss": 4.524338245391846,
+ "step": 4370
+ },
+ {
+ "epoch": 60.712663755458514,
+ "grad_norm": 0.017019255086779594,
+ "learning_rate": 0.0006,
+ "loss": 4.387436866760254,
+ "step": 4371
+ },
+ {
+ "epoch": 60.72663755458515,
+ "grad_norm": 0.017070675268769264,
+ "learning_rate": 0.0006,
+ "loss": 4.437321662902832,
+ "step": 4372
+ },
+ {
+ "epoch": 60.74061135371179,
+ "grad_norm": 0.015197164379060268,
+ "learning_rate": 0.0006,
+ "loss": 4.371205806732178,
+ "step": 4373
+ },
+ {
+ "epoch": 60.75458515283843,
+ "grad_norm": 0.014791185967624187,
+ "learning_rate": 0.0006,
+ "loss": 4.44439697265625,
+ "step": 4374
+ },
+ {
+ "epoch": 60.76855895196506,
+ "grad_norm": 0.016458261758089066,
+ "learning_rate": 0.0006,
+ "loss": 4.445584297180176,
+ "step": 4375
+ },
+ {
+ "epoch": 60.7825327510917,
+ "grad_norm": 0.017706342041492462,
+ "learning_rate": 0.0006,
+ "loss": 4.348708152770996,
+ "step": 4376
+ },
+ {
+ "epoch": 60.79650655021834,
+ "grad_norm": 0.017561476677656174,
+ "learning_rate": 0.0006,
+ "loss": 4.521929740905762,
+ "step": 4377
+ },
+ {
+ "epoch": 60.81048034934498,
+ "grad_norm": 0.016557445749640465,
+ "learning_rate": 0.0006,
+ "loss": 4.545806884765625,
+ "step": 4378
+ },
+ {
+ "epoch": 60.82445414847162,
+ "grad_norm": 0.016953222453594208,
+ "learning_rate": 0.0006,
+ "loss": 4.4230875968933105,
+ "step": 4379
+ },
+ {
+ "epoch": 60.838427947598255,
+ "grad_norm": 0.017094967886805534,
+ "learning_rate": 0.0006,
+ "loss": 4.423985481262207,
+ "step": 4380
+ },
+ {
+ "epoch": 60.852401746724894,
+ "grad_norm": 0.014787515625357628,
+ "learning_rate": 0.0006,
+ "loss": 4.39473819732666,
+ "step": 4381
+ },
+ {
+ "epoch": 60.866375545851525,
+ "grad_norm": 0.016208263114094734,
+ "learning_rate": 0.0006,
+ "loss": 4.3810038566589355,
+ "step": 4382
+ },
+ {
+ "epoch": 60.880349344978164,
+ "grad_norm": 0.017844321206212044,
+ "learning_rate": 0.0006,
+ "loss": 4.35086727142334,
+ "step": 4383
+ },
+ {
+ "epoch": 60.8943231441048,
+ "grad_norm": 0.017539294436573982,
+ "learning_rate": 0.0006,
+ "loss": 4.460260391235352,
+ "step": 4384
+ },
+ {
+ "epoch": 60.90829694323144,
+ "grad_norm": 0.01612042263150215,
+ "learning_rate": 0.0006,
+ "loss": 4.370500564575195,
+ "step": 4385
+ },
+ {
+ "epoch": 60.92227074235808,
+ "grad_norm": 0.015381601639091969,
+ "learning_rate": 0.0006,
+ "loss": 4.435868263244629,
+ "step": 4386
+ },
+ {
+ "epoch": 60.93624454148472,
+ "grad_norm": 0.01603585295379162,
+ "learning_rate": 0.0006,
+ "loss": 4.434070587158203,
+ "step": 4387
+ },
+ {
+ "epoch": 60.95021834061135,
+ "grad_norm": 0.016268130391836166,
+ "learning_rate": 0.0006,
+ "loss": 4.404343128204346,
+ "step": 4388
+ },
+ {
+ "epoch": 60.96419213973799,
+ "grad_norm": 0.014980032108724117,
+ "learning_rate": 0.0006,
+ "loss": 4.399213790893555,
+ "step": 4389
+ },
+ {
+ "epoch": 60.97816593886463,
+ "grad_norm": 0.016899267211556435,
+ "learning_rate": 0.0006,
+ "loss": 4.508069038391113,
+ "step": 4390
+ },
+ {
+ "epoch": 60.992139737991266,
+ "grad_norm": 0.013756908476352692,
+ "learning_rate": 0.0006,
+ "loss": 4.541858673095703,
+ "step": 4391
+ },
+ {
+ "epoch": 61.0,
+ "grad_norm": 0.01578413136303425,
+ "learning_rate": 0.0006,
+ "loss": 4.543513298034668,
+ "step": 4392
+ },
+ {
+ "epoch": 61.0,
+ "eval_loss": 4.756770610809326,
+ "eval_runtime": 56.4572,
+ "eval_samples_per_second": 43.254,
+ "eval_steps_per_second": 1.364,
+ "step": 4392
+ },
+ {
+ "epoch": 61.01397379912664,
+ "grad_norm": 0.013387506827712059,
+ "learning_rate": 0.0006,
+ "loss": 4.475942611694336,
+ "step": 4393
+ },
+ {
+ "epoch": 61.02794759825328,
+ "grad_norm": 0.014278904534876347,
+ "learning_rate": 0.0006,
+ "loss": 4.442403793334961,
+ "step": 4394
+ },
+ {
+ "epoch": 61.041921397379916,
+ "grad_norm": 0.014359553344547749,
+ "learning_rate": 0.0006,
+ "loss": 4.509463310241699,
+ "step": 4395
+ },
+ {
+ "epoch": 61.05589519650655,
+ "grad_norm": 0.01718050241470337,
+ "learning_rate": 0.0006,
+ "loss": 4.354458808898926,
+ "step": 4396
+ },
+ {
+ "epoch": 61.069868995633186,
+ "grad_norm": 0.015988312661647797,
+ "learning_rate": 0.0006,
+ "loss": 4.484508514404297,
+ "step": 4397
+ },
+ {
+ "epoch": 61.083842794759825,
+ "grad_norm": 0.015466735698282719,
+ "learning_rate": 0.0006,
+ "loss": 4.326908111572266,
+ "step": 4398
+ },
+ {
+ "epoch": 61.09781659388646,
+ "grad_norm": 0.014541332609951496,
+ "learning_rate": 0.0006,
+ "loss": 4.405823230743408,
+ "step": 4399
+ },
+ {
+ "epoch": 61.1117903930131,
+ "grad_norm": 0.014803987927734852,
+ "learning_rate": 0.0006,
+ "loss": 4.375386714935303,
+ "step": 4400
+ },
+ {
+ "epoch": 61.12576419213974,
+ "grad_norm": 0.013962093740701675,
+ "learning_rate": 0.0006,
+ "loss": 4.300838470458984,
+ "step": 4401
+ },
+ {
+ "epoch": 61.13973799126637,
+ "grad_norm": 0.015213954262435436,
+ "learning_rate": 0.0006,
+ "loss": 4.388306617736816,
+ "step": 4402
+ },
+ {
+ "epoch": 61.15371179039301,
+ "grad_norm": 0.018965506926178932,
+ "learning_rate": 0.0006,
+ "loss": 4.388340950012207,
+ "step": 4403
+ },
+ {
+ "epoch": 61.16768558951965,
+ "grad_norm": 0.01958410255610943,
+ "learning_rate": 0.0006,
+ "loss": 4.35012149810791,
+ "step": 4404
+ },
+ {
+ "epoch": 61.18165938864629,
+ "grad_norm": 0.01938101276755333,
+ "learning_rate": 0.0006,
+ "loss": 4.368372917175293,
+ "step": 4405
+ },
+ {
+ "epoch": 61.19563318777293,
+ "grad_norm": 0.016556516289711,
+ "learning_rate": 0.0006,
+ "loss": 4.486198425292969,
+ "step": 4406
+ },
+ {
+ "epoch": 61.209606986899566,
+ "grad_norm": 0.017031289637088776,
+ "learning_rate": 0.0006,
+ "loss": 4.421676158905029,
+ "step": 4407
+ },
+ {
+ "epoch": 61.223580786026204,
+ "grad_norm": 0.018023479729890823,
+ "learning_rate": 0.0006,
+ "loss": 4.38751220703125,
+ "step": 4408
+ },
+ {
+ "epoch": 61.237554585152836,
+ "grad_norm": 0.0178135447204113,
+ "learning_rate": 0.0006,
+ "loss": 4.336528301239014,
+ "step": 4409
+ },
+ {
+ "epoch": 61.251528384279474,
+ "grad_norm": 0.016782190650701523,
+ "learning_rate": 0.0006,
+ "loss": 4.386693954467773,
+ "step": 4410
+ },
+ {
+ "epoch": 61.26550218340611,
+ "grad_norm": 0.015494848601520061,
+ "learning_rate": 0.0006,
+ "loss": 4.308948040008545,
+ "step": 4411
+ },
+ {
+ "epoch": 61.27947598253275,
+ "grad_norm": 0.015002378262579441,
+ "learning_rate": 0.0006,
+ "loss": 4.314206123352051,
+ "step": 4412
+ },
+ {
+ "epoch": 61.29344978165939,
+ "grad_norm": 0.015798920765519142,
+ "learning_rate": 0.0006,
+ "loss": 4.502900123596191,
+ "step": 4413
+ },
+ {
+ "epoch": 61.30742358078603,
+ "grad_norm": 0.015444336459040642,
+ "learning_rate": 0.0006,
+ "loss": 4.473160743713379,
+ "step": 4414
+ },
+ {
+ "epoch": 61.32139737991266,
+ "grad_norm": 0.017025096341967583,
+ "learning_rate": 0.0006,
+ "loss": 4.460621356964111,
+ "step": 4415
+ },
+ {
+ "epoch": 61.3353711790393,
+ "grad_norm": 0.019502537325024605,
+ "learning_rate": 0.0006,
+ "loss": 4.385551452636719,
+ "step": 4416
+ },
+ {
+ "epoch": 61.34934497816594,
+ "grad_norm": 0.018529541790485382,
+ "learning_rate": 0.0006,
+ "loss": 4.425390720367432,
+ "step": 4417
+ },
+ {
+ "epoch": 61.36331877729258,
+ "grad_norm": 0.013155706226825714,
+ "learning_rate": 0.0006,
+ "loss": 4.467020034790039,
+ "step": 4418
+ },
+ {
+ "epoch": 61.377292576419215,
+ "grad_norm": 0.01518090907484293,
+ "learning_rate": 0.0006,
+ "loss": 4.369664192199707,
+ "step": 4419
+ },
+ {
+ "epoch": 61.391266375545854,
+ "grad_norm": 0.01626511849462986,
+ "learning_rate": 0.0006,
+ "loss": 4.463231086730957,
+ "step": 4420
+ },
+ {
+ "epoch": 61.40524017467249,
+ "grad_norm": 0.01701558195054531,
+ "learning_rate": 0.0006,
+ "loss": 4.43708610534668,
+ "step": 4421
+ },
+ {
+ "epoch": 61.419213973799124,
+ "grad_norm": 0.01761862449347973,
+ "learning_rate": 0.0006,
+ "loss": 4.3314208984375,
+ "step": 4422
+ },
+ {
+ "epoch": 61.43318777292576,
+ "grad_norm": 0.017084648832678795,
+ "learning_rate": 0.0006,
+ "loss": 4.39676570892334,
+ "step": 4423
+ },
+ {
+ "epoch": 61.4471615720524,
+ "grad_norm": 0.01667078025639057,
+ "learning_rate": 0.0006,
+ "loss": 4.3789873123168945,
+ "step": 4424
+ },
+ {
+ "epoch": 61.46113537117904,
+ "grad_norm": 0.01877436414361,
+ "learning_rate": 0.0006,
+ "loss": 4.415275573730469,
+ "step": 4425
+ },
+ {
+ "epoch": 61.47510917030568,
+ "grad_norm": 0.020250486209988594,
+ "learning_rate": 0.0006,
+ "loss": 4.478635787963867,
+ "step": 4426
+ },
+ {
+ "epoch": 61.48908296943232,
+ "grad_norm": 0.020244868472218513,
+ "learning_rate": 0.0006,
+ "loss": 4.296854019165039,
+ "step": 4427
+ },
+ {
+ "epoch": 61.50305676855895,
+ "grad_norm": 0.018723875284194946,
+ "learning_rate": 0.0006,
+ "loss": 4.3760881423950195,
+ "step": 4428
+ },
+ {
+ "epoch": 61.51703056768559,
+ "grad_norm": 0.015138164162635803,
+ "learning_rate": 0.0006,
+ "loss": 4.39974308013916,
+ "step": 4429
+ },
+ {
+ "epoch": 61.531004366812226,
+ "grad_norm": 0.015111138112843037,
+ "learning_rate": 0.0006,
+ "loss": 4.34659481048584,
+ "step": 4430
+ },
+ {
+ "epoch": 61.544978165938865,
+ "grad_norm": 0.0164666585624218,
+ "learning_rate": 0.0006,
+ "loss": 4.415737152099609,
+ "step": 4431
+ },
+ {
+ "epoch": 61.5589519650655,
+ "grad_norm": 0.020891312509775162,
+ "learning_rate": 0.0006,
+ "loss": 4.300319671630859,
+ "step": 4432
+ },
+ {
+ "epoch": 61.57292576419214,
+ "grad_norm": 0.02089156024158001,
+ "learning_rate": 0.0006,
+ "loss": 4.373029708862305,
+ "step": 4433
+ },
+ {
+ "epoch": 61.58689956331878,
+ "grad_norm": 0.020114559680223465,
+ "learning_rate": 0.0006,
+ "loss": 4.434499740600586,
+ "step": 4434
+ },
+ {
+ "epoch": 61.60087336244541,
+ "grad_norm": 0.01791365258395672,
+ "learning_rate": 0.0006,
+ "loss": 4.466872215270996,
+ "step": 4435
+ },
+ {
+ "epoch": 61.61484716157205,
+ "grad_norm": 0.01879458874464035,
+ "learning_rate": 0.0006,
+ "loss": 4.5176520347595215,
+ "step": 4436
+ },
+ {
+ "epoch": 61.62882096069869,
+ "grad_norm": 0.016371937468647957,
+ "learning_rate": 0.0006,
+ "loss": 4.3507585525512695,
+ "step": 4437
+ },
+ {
+ "epoch": 61.64279475982533,
+ "grad_norm": 0.01648143120110035,
+ "learning_rate": 0.0006,
+ "loss": 4.398778438568115,
+ "step": 4438
+ },
+ {
+ "epoch": 61.65676855895197,
+ "grad_norm": 0.017240576446056366,
+ "learning_rate": 0.0006,
+ "loss": 4.513933181762695,
+ "step": 4439
+ },
+ {
+ "epoch": 61.670742358078606,
+ "grad_norm": 0.01734284684062004,
+ "learning_rate": 0.0006,
+ "loss": 4.315343856811523,
+ "step": 4440
+ },
+ {
+ "epoch": 61.68471615720524,
+ "grad_norm": 0.01639772206544876,
+ "learning_rate": 0.0006,
+ "loss": 4.4641218185424805,
+ "step": 4441
+ },
+ {
+ "epoch": 61.698689956331876,
+ "grad_norm": 0.0164708960801363,
+ "learning_rate": 0.0006,
+ "loss": 4.439968585968018,
+ "step": 4442
+ },
+ {
+ "epoch": 61.712663755458514,
+ "grad_norm": 0.01757933758199215,
+ "learning_rate": 0.0006,
+ "loss": 4.408327102661133,
+ "step": 4443
+ },
+ {
+ "epoch": 61.72663755458515,
+ "grad_norm": 0.01789589412510395,
+ "learning_rate": 0.0006,
+ "loss": 4.478257179260254,
+ "step": 4444
+ },
+ {
+ "epoch": 61.74061135371179,
+ "grad_norm": 0.01822184957563877,
+ "learning_rate": 0.0006,
+ "loss": 4.412046909332275,
+ "step": 4445
+ },
+ {
+ "epoch": 61.75458515283843,
+ "grad_norm": 0.018624618649482727,
+ "learning_rate": 0.0006,
+ "loss": 4.427122116088867,
+ "step": 4446
+ },
+ {
+ "epoch": 61.76855895196506,
+ "grad_norm": 0.018674887716770172,
+ "learning_rate": 0.0006,
+ "loss": 4.429147720336914,
+ "step": 4447
+ },
+ {
+ "epoch": 61.7825327510917,
+ "grad_norm": 0.015317119657993317,
+ "learning_rate": 0.0006,
+ "loss": 4.2808451652526855,
+ "step": 4448
+ },
+ {
+ "epoch": 61.79650655021834,
+ "grad_norm": 0.015400845557451248,
+ "learning_rate": 0.0006,
+ "loss": 4.420193672180176,
+ "step": 4449
+ },
+ {
+ "epoch": 61.81048034934498,
+ "grad_norm": 0.016743650659918785,
+ "learning_rate": 0.0006,
+ "loss": 4.498254776000977,
+ "step": 4450
+ },
+ {
+ "epoch": 61.82445414847162,
+ "grad_norm": 0.014330781064927578,
+ "learning_rate": 0.0006,
+ "loss": 4.445384979248047,
+ "step": 4451
+ },
+ {
+ "epoch": 61.838427947598255,
+ "grad_norm": 0.015046479180455208,
+ "learning_rate": 0.0006,
+ "loss": 4.368041038513184,
+ "step": 4452
+ },
+ {
+ "epoch": 61.852401746724894,
+ "grad_norm": 0.01630168780684471,
+ "learning_rate": 0.0006,
+ "loss": 4.460142135620117,
+ "step": 4453
+ },
+ {
+ "epoch": 61.866375545851525,
+ "grad_norm": 0.01637885719537735,
+ "learning_rate": 0.0006,
+ "loss": 4.416606426239014,
+ "step": 4454
+ },
+ {
+ "epoch": 61.880349344978164,
+ "grad_norm": 0.016785888001322746,
+ "learning_rate": 0.0006,
+ "loss": 4.452545642852783,
+ "step": 4455
+ },
+ {
+ "epoch": 61.8943231441048,
+ "grad_norm": 0.015044555068016052,
+ "learning_rate": 0.0006,
+ "loss": 4.427334785461426,
+ "step": 4456
+ },
+ {
+ "epoch": 61.90829694323144,
+ "grad_norm": 0.013699211180210114,
+ "learning_rate": 0.0006,
+ "loss": 4.408761024475098,
+ "step": 4457
+ },
+ {
+ "epoch": 61.92227074235808,
+ "grad_norm": 0.014021803624927998,
+ "learning_rate": 0.0006,
+ "loss": 4.543607711791992,
+ "step": 4458
+ },
+ {
+ "epoch": 61.93624454148472,
+ "grad_norm": 0.015438460744917393,
+ "learning_rate": 0.0006,
+ "loss": 4.313363075256348,
+ "step": 4459
+ },
+ {
+ "epoch": 61.95021834061135,
+ "grad_norm": 0.017515428364276886,
+ "learning_rate": 0.0006,
+ "loss": 4.310412406921387,
+ "step": 4460
+ },
+ {
+ "epoch": 61.96419213973799,
+ "grad_norm": 0.016683880239725113,
+ "learning_rate": 0.0006,
+ "loss": 4.3378448486328125,
+ "step": 4461
+ },
+ {
+ "epoch": 61.97816593886463,
+ "grad_norm": 0.019384048879146576,
+ "learning_rate": 0.0006,
+ "loss": 4.450568675994873,
+ "step": 4462
+ },
+ {
+ "epoch": 61.992139737991266,
+ "grad_norm": 0.020651493221521378,
+ "learning_rate": 0.0006,
+ "loss": 4.453197479248047,
+ "step": 4463
+ },
+ {
+ "epoch": 62.0,
+ "grad_norm": 0.020991764962673187,
+ "learning_rate": 0.0006,
+ "loss": 4.441527366638184,
+ "step": 4464
+ },
+ {
+ "epoch": 62.0,
+ "eval_loss": 4.700193881988525,
+ "eval_runtime": 56.6098,
+ "eval_samples_per_second": 43.137,
+ "eval_steps_per_second": 1.36,
+ "step": 4464
+ },
+ {
+ "epoch": 62.01397379912664,
+ "grad_norm": 0.01811373233795166,
+ "learning_rate": 0.0006,
+ "loss": 4.37346076965332,
+ "step": 4465
+ },
+ {
+ "epoch": 62.02794759825328,
+ "grad_norm": 0.01667127199470997,
+ "learning_rate": 0.0006,
+ "loss": 4.237151145935059,
+ "step": 4466
+ },
+ {
+ "epoch": 62.041921397379916,
+ "grad_norm": 0.014971534721553326,
+ "learning_rate": 0.0006,
+ "loss": 4.367053031921387,
+ "step": 4467
+ },
+ {
+ "epoch": 62.05589519650655,
+ "grad_norm": 0.015655893832445145,
+ "learning_rate": 0.0006,
+ "loss": 4.437426567077637,
+ "step": 4468
+ },
+ {
+ "epoch": 62.069868995633186,
+ "grad_norm": 0.014935447834432125,
+ "learning_rate": 0.0006,
+ "loss": 4.3635125160217285,
+ "step": 4469
+ },
+ {
+ "epoch": 62.083842794759825,
+ "grad_norm": 0.015363802202045918,
+ "learning_rate": 0.0006,
+ "loss": 4.40492057800293,
+ "step": 4470
+ },
+ {
+ "epoch": 62.09781659388646,
+ "grad_norm": 0.0159201230853796,
+ "learning_rate": 0.0006,
+ "loss": 4.392058849334717,
+ "step": 4471
+ },
+ {
+ "epoch": 62.1117903930131,
+ "grad_norm": 0.014598241075873375,
+ "learning_rate": 0.0006,
+ "loss": 4.369944095611572,
+ "step": 4472
+ },
+ {
+ "epoch": 62.12576419213974,
+ "grad_norm": 0.015394841320812702,
+ "learning_rate": 0.0006,
+ "loss": 4.425097942352295,
+ "step": 4473
+ },
+ {
+ "epoch": 62.13973799126637,
+ "grad_norm": 0.015913628041744232,
+ "learning_rate": 0.0006,
+ "loss": 4.420596122741699,
+ "step": 4474
+ },
+ {
+ "epoch": 62.15371179039301,
+ "grad_norm": 0.01575258933007717,
+ "learning_rate": 0.0006,
+ "loss": 4.32290506362915,
+ "step": 4475
+ },
+ {
+ "epoch": 62.16768558951965,
+ "grad_norm": 0.0155344782397151,
+ "learning_rate": 0.0006,
+ "loss": 4.273068428039551,
+ "step": 4476
+ },
+ {
+ "epoch": 62.18165938864629,
+ "grad_norm": 0.014947040006518364,
+ "learning_rate": 0.0006,
+ "loss": 4.384868621826172,
+ "step": 4477
+ },
+ {
+ "epoch": 62.19563318777293,
+ "grad_norm": 0.01557591650635004,
+ "learning_rate": 0.0006,
+ "loss": 4.458410263061523,
+ "step": 4478
+ },
+ {
+ "epoch": 62.209606986899566,
+ "grad_norm": 0.015182198956608772,
+ "learning_rate": 0.0006,
+ "loss": 4.300738334655762,
+ "step": 4479
+ },
+ {
+ "epoch": 62.223580786026204,
+ "grad_norm": 0.01634717360138893,
+ "learning_rate": 0.0006,
+ "loss": 4.436875820159912,
+ "step": 4480
+ },
+ {
+ "epoch": 62.237554585152836,
+ "grad_norm": 0.018813718110322952,
+ "learning_rate": 0.0006,
+ "loss": 4.454665184020996,
+ "step": 4481
+ },
+ {
+ "epoch": 62.251528384279474,
+ "grad_norm": 0.020965630188584328,
+ "learning_rate": 0.0006,
+ "loss": 4.357341766357422,
+ "step": 4482
+ },
+ {
+ "epoch": 62.26550218340611,
+ "grad_norm": 0.020738353952765465,
+ "learning_rate": 0.0006,
+ "loss": 4.456644058227539,
+ "step": 4483
+ },
+ {
+ "epoch": 62.27947598253275,
+ "grad_norm": 0.018813546746969223,
+ "learning_rate": 0.0006,
+ "loss": 4.381941795349121,
+ "step": 4484
+ },
+ {
+ "epoch": 62.29344978165939,
+ "grad_norm": 0.015431761741638184,
+ "learning_rate": 0.0006,
+ "loss": 4.41126823425293,
+ "step": 4485
+ },
+ {
+ "epoch": 62.30742358078603,
+ "grad_norm": 0.016335977241396904,
+ "learning_rate": 0.0006,
+ "loss": 4.419488906860352,
+ "step": 4486
+ },
+ {
+ "epoch": 62.32139737991266,
+ "grad_norm": 0.01832406409084797,
+ "learning_rate": 0.0006,
+ "loss": 4.579026222229004,
+ "step": 4487
+ },
+ {
+ "epoch": 62.3353711790393,
+ "grad_norm": 0.01790742017328739,
+ "learning_rate": 0.0006,
+ "loss": 4.289964199066162,
+ "step": 4488
+ },
+ {
+ "epoch": 62.34934497816594,
+ "grad_norm": 0.017266202718019485,
+ "learning_rate": 0.0006,
+ "loss": 4.415959358215332,
+ "step": 4489
+ },
+ {
+ "epoch": 62.36331877729258,
+ "grad_norm": 0.015976132825016975,
+ "learning_rate": 0.0006,
+ "loss": 4.358587265014648,
+ "step": 4490
+ },
+ {
+ "epoch": 62.377292576419215,
+ "grad_norm": 0.016873881220817566,
+ "learning_rate": 0.0006,
+ "loss": 4.354607105255127,
+ "step": 4491
+ },
+ {
+ "epoch": 62.391266375545854,
+ "grad_norm": 0.01476745493710041,
+ "learning_rate": 0.0006,
+ "loss": 4.384904384613037,
+ "step": 4492
+ },
+ {
+ "epoch": 62.40524017467249,
+ "grad_norm": 0.013810782693326473,
+ "learning_rate": 0.0006,
+ "loss": 4.459230899810791,
+ "step": 4493
+ },
+ {
+ "epoch": 62.419213973799124,
+ "grad_norm": 0.015150421299040318,
+ "learning_rate": 0.0006,
+ "loss": 4.369471549987793,
+ "step": 4494
+ },
+ {
+ "epoch": 62.43318777292576,
+ "grad_norm": 0.014539610594511032,
+ "learning_rate": 0.0006,
+ "loss": 4.3564982414245605,
+ "step": 4495
+ },
+ {
+ "epoch": 62.4471615720524,
+ "grad_norm": 0.015554017387330532,
+ "learning_rate": 0.0006,
+ "loss": 4.414379119873047,
+ "step": 4496
+ },
+ {
+ "epoch": 62.46113537117904,
+ "grad_norm": 0.016608357429504395,
+ "learning_rate": 0.0006,
+ "loss": 4.311585426330566,
+ "step": 4497
+ },
+ {
+ "epoch": 62.47510917030568,
+ "grad_norm": 0.01885881833732128,
+ "learning_rate": 0.0006,
+ "loss": 4.306629180908203,
+ "step": 4498
+ },
+ {
+ "epoch": 62.48908296943232,
+ "grad_norm": 0.022128663957118988,
+ "learning_rate": 0.0006,
+ "loss": 4.345483779907227,
+ "step": 4499
+ },
+ {
+ "epoch": 62.50305676855895,
+ "grad_norm": 0.023401018232107162,
+ "learning_rate": 0.0006,
+ "loss": 4.398204326629639,
+ "step": 4500
+ },
+ {
+ "epoch": 62.51703056768559,
+ "grad_norm": 0.022080009803175926,
+ "learning_rate": 0.0006,
+ "loss": 4.375027179718018,
+ "step": 4501
+ },
+ {
+ "epoch": 62.531004366812226,
+ "grad_norm": 0.020004864782094955,
+ "learning_rate": 0.0006,
+ "loss": 4.33494758605957,
+ "step": 4502
+ },
+ {
+ "epoch": 62.544978165938865,
+ "grad_norm": 0.01786700263619423,
+ "learning_rate": 0.0006,
+ "loss": 4.4428300857543945,
+ "step": 4503
+ },
+ {
+ "epoch": 62.5589519650655,
+ "grad_norm": 0.01885388419032097,
+ "learning_rate": 0.0006,
+ "loss": 4.4138946533203125,
+ "step": 4504
+ },
+ {
+ "epoch": 62.57292576419214,
+ "grad_norm": 0.019890939816832542,
+ "learning_rate": 0.0006,
+ "loss": 4.278632164001465,
+ "step": 4505
+ },
+ {
+ "epoch": 62.58689956331878,
+ "grad_norm": 0.02109031192958355,
+ "learning_rate": 0.0006,
+ "loss": 4.3756327629089355,
+ "step": 4506
+ },
+ {
+ "epoch": 62.60087336244541,
+ "grad_norm": 0.020813899114727974,
+ "learning_rate": 0.0006,
+ "loss": 4.410213470458984,
+ "step": 4507
+ },
+ {
+ "epoch": 62.61484716157205,
+ "grad_norm": 0.021655689924955368,
+ "learning_rate": 0.0006,
+ "loss": 4.30633020401001,
+ "step": 4508
+ },
+ {
+ "epoch": 62.62882096069869,
+ "grad_norm": 0.018155649304389954,
+ "learning_rate": 0.0006,
+ "loss": 4.407284736633301,
+ "step": 4509
+ },
+ {
+ "epoch": 62.64279475982533,
+ "grad_norm": 0.017765069380402565,
+ "learning_rate": 0.0006,
+ "loss": 4.321715354919434,
+ "step": 4510
+ },
+ {
+ "epoch": 62.65676855895197,
+ "grad_norm": 0.017724554985761642,
+ "learning_rate": 0.0006,
+ "loss": 4.493371963500977,
+ "step": 4511
+ },
+ {
+ "epoch": 62.670742358078606,
+ "grad_norm": 0.01492092851549387,
+ "learning_rate": 0.0006,
+ "loss": 4.443025588989258,
+ "step": 4512
+ },
+ {
+ "epoch": 62.68471615720524,
+ "grad_norm": 0.01615324057638645,
+ "learning_rate": 0.0006,
+ "loss": 4.3289594650268555,
+ "step": 4513
+ },
+ {
+ "epoch": 62.698689956331876,
+ "grad_norm": 0.016569366678595543,
+ "learning_rate": 0.0006,
+ "loss": 4.369720458984375,
+ "step": 4514
+ },
+ {
+ "epoch": 62.712663755458514,
+ "grad_norm": 0.016509901732206345,
+ "learning_rate": 0.0006,
+ "loss": 4.292316436767578,
+ "step": 4515
+ },
+ {
+ "epoch": 62.72663755458515,
+ "grad_norm": 0.014270003885030746,
+ "learning_rate": 0.0006,
+ "loss": 4.349854469299316,
+ "step": 4516
+ },
+ {
+ "epoch": 62.74061135371179,
+ "grad_norm": 0.014235340990126133,
+ "learning_rate": 0.0006,
+ "loss": 4.473727226257324,
+ "step": 4517
+ },
+ {
+ "epoch": 62.75458515283843,
+ "grad_norm": 0.015163225121796131,
+ "learning_rate": 0.0006,
+ "loss": 4.452878475189209,
+ "step": 4518
+ },
+ {
+ "epoch": 62.76855895196506,
+ "grad_norm": 0.015748055651783943,
+ "learning_rate": 0.0006,
+ "loss": 4.521087169647217,
+ "step": 4519
+ },
+ {
+ "epoch": 62.7825327510917,
+ "grad_norm": 0.01661108434200287,
+ "learning_rate": 0.0006,
+ "loss": 4.443885326385498,
+ "step": 4520
+ },
+ {
+ "epoch": 62.79650655021834,
+ "grad_norm": 0.01610301062464714,
+ "learning_rate": 0.0006,
+ "loss": 4.312990188598633,
+ "step": 4521
+ },
+ {
+ "epoch": 62.81048034934498,
+ "grad_norm": 0.017007583752274513,
+ "learning_rate": 0.0006,
+ "loss": 4.408937454223633,
+ "step": 4522
+ },
+ {
+ "epoch": 62.82445414847162,
+ "grad_norm": 0.01749461516737938,
+ "learning_rate": 0.0006,
+ "loss": 4.346066951751709,
+ "step": 4523
+ },
+ {
+ "epoch": 62.838427947598255,
+ "grad_norm": 0.018243493512272835,
+ "learning_rate": 0.0006,
+ "loss": 4.357658386230469,
+ "step": 4524
+ },
+ {
+ "epoch": 62.852401746724894,
+ "grad_norm": 0.017499985173344612,
+ "learning_rate": 0.0006,
+ "loss": 4.374061584472656,
+ "step": 4525
+ },
+ {
+ "epoch": 62.866375545851525,
+ "grad_norm": 0.014536325819790363,
+ "learning_rate": 0.0006,
+ "loss": 4.189321994781494,
+ "step": 4526
+ },
+ {
+ "epoch": 62.880349344978164,
+ "grad_norm": 0.014728891663253307,
+ "learning_rate": 0.0006,
+ "loss": 4.425910949707031,
+ "step": 4527
+ },
+ {
+ "epoch": 62.8943231441048,
+ "grad_norm": 0.015013772062957287,
+ "learning_rate": 0.0006,
+ "loss": 4.337319850921631,
+ "step": 4528
+ },
+ {
+ "epoch": 62.90829694323144,
+ "grad_norm": 0.01626027375459671,
+ "learning_rate": 0.0006,
+ "loss": 4.330339431762695,
+ "step": 4529
+ },
+ {
+ "epoch": 62.92227074235808,
+ "grad_norm": 0.014773444272577763,
+ "learning_rate": 0.0006,
+ "loss": 4.4188737869262695,
+ "step": 4530
+ },
+ {
+ "epoch": 62.93624454148472,
+ "grad_norm": 0.01422932744026184,
+ "learning_rate": 0.0006,
+ "loss": 4.355986595153809,
+ "step": 4531
+ },
+ {
+ "epoch": 62.95021834061135,
+ "grad_norm": 0.012170110829174519,
+ "learning_rate": 0.0006,
+ "loss": 4.405722618103027,
+ "step": 4532
+ },
+ {
+ "epoch": 62.96419213973799,
+ "grad_norm": 0.013233253732323647,
+ "learning_rate": 0.0006,
+ "loss": 4.377286911010742,
+ "step": 4533
+ },
+ {
+ "epoch": 62.97816593886463,
+ "grad_norm": 0.014497299678623676,
+ "learning_rate": 0.0006,
+ "loss": 4.508796691894531,
+ "step": 4534
+ },
+ {
+ "epoch": 62.992139737991266,
+ "grad_norm": 0.014718741178512573,
+ "learning_rate": 0.0006,
+ "loss": 4.370919704437256,
+ "step": 4535
+ },
+ {
+ "epoch": 63.0,
+ "grad_norm": 0.015833502635359764,
+ "learning_rate": 0.0006,
+ "loss": 4.258753776550293,
+ "step": 4536
+ },
+ {
+ "epoch": 63.0,
+ "eval_loss": 4.714163303375244,
+ "eval_runtime": 57.0269,
+ "eval_samples_per_second": 42.822,
+ "eval_steps_per_second": 1.35,
+ "step": 4536
+ },
+ {
+ "epoch": 63.01397379912664,
+ "grad_norm": 0.014277510344982147,
+ "learning_rate": 0.0006,
+ "loss": 4.3112640380859375,
+ "step": 4537
+ },
+ {
+ "epoch": 63.02794759825328,
+ "grad_norm": 0.015377935022115707,
+ "learning_rate": 0.0006,
+ "loss": 4.342782974243164,
+ "step": 4538
+ },
+ {
+ "epoch": 63.041921397379916,
+ "grad_norm": 0.015406244434416294,
+ "learning_rate": 0.0006,
+ "loss": 4.474056720733643,
+ "step": 4539
+ },
+ {
+ "epoch": 63.05589519650655,
+ "grad_norm": 0.018121499568223953,
+ "learning_rate": 0.0006,
+ "loss": 4.397963523864746,
+ "step": 4540
+ },
+ {
+ "epoch": 63.069868995633186,
+ "grad_norm": 0.022634636610746384,
+ "learning_rate": 0.0006,
+ "loss": 4.306631088256836,
+ "step": 4541
+ },
+ {
+ "epoch": 63.083842794759825,
+ "grad_norm": 0.026448970660567284,
+ "learning_rate": 0.0006,
+ "loss": 4.283175468444824,
+ "step": 4542
+ },
+ {
+ "epoch": 63.09781659388646,
+ "grad_norm": 0.024104537442326546,
+ "learning_rate": 0.0006,
+ "loss": 4.315826416015625,
+ "step": 4543
+ },
+ {
+ "epoch": 63.1117903930131,
+ "grad_norm": 0.019040480256080627,
+ "learning_rate": 0.0006,
+ "loss": 4.44057035446167,
+ "step": 4544
+ },
+ {
+ "epoch": 63.12576419213974,
+ "grad_norm": 0.02006000280380249,
+ "learning_rate": 0.0006,
+ "loss": 4.405603408813477,
+ "step": 4545
+ },
+ {
+ "epoch": 63.13973799126637,
+ "grad_norm": 0.019980071112513542,
+ "learning_rate": 0.0006,
+ "loss": 4.393726348876953,
+ "step": 4546
+ },
+ {
+ "epoch": 63.15371179039301,
+ "grad_norm": 0.01641865260899067,
+ "learning_rate": 0.0006,
+ "loss": 4.272546768188477,
+ "step": 4547
+ },
+ {
+ "epoch": 63.16768558951965,
+ "grad_norm": 0.01890353485941887,
+ "learning_rate": 0.0006,
+ "loss": 4.382566928863525,
+ "step": 4548
+ },
+ {
+ "epoch": 63.18165938864629,
+ "grad_norm": 0.018370937556028366,
+ "learning_rate": 0.0006,
+ "loss": 4.423953533172607,
+ "step": 4549
+ },
+ {
+ "epoch": 63.19563318777293,
+ "grad_norm": 0.014985294081270695,
+ "learning_rate": 0.0006,
+ "loss": 4.308627128601074,
+ "step": 4550
+ },
+ {
+ "epoch": 63.209606986899566,
+ "grad_norm": 0.017513327300548553,
+ "learning_rate": 0.0006,
+ "loss": 4.248819351196289,
+ "step": 4551
+ },
+ {
+ "epoch": 63.223580786026204,
+ "grad_norm": 0.019464161247015,
+ "learning_rate": 0.0006,
+ "loss": 4.359729766845703,
+ "step": 4552
+ },
+ {
+ "epoch": 63.237554585152836,
+ "grad_norm": 0.018678035587072372,
+ "learning_rate": 0.0006,
+ "loss": 4.3299665451049805,
+ "step": 4553
+ },
+ {
+ "epoch": 63.251528384279474,
+ "grad_norm": 0.017909523099660873,
+ "learning_rate": 0.0006,
+ "loss": 4.246147155761719,
+ "step": 4554
+ },
+ {
+ "epoch": 63.26550218340611,
+ "grad_norm": 0.01747293211519718,
+ "learning_rate": 0.0006,
+ "loss": 4.41165018081665,
+ "step": 4555
+ },
+ {
+ "epoch": 63.27947598253275,
+ "grad_norm": 0.018670853227376938,
+ "learning_rate": 0.0006,
+ "loss": 4.409883499145508,
+ "step": 4556
+ },
+ {
+ "epoch": 63.29344978165939,
+ "grad_norm": 0.0189529862254858,
+ "learning_rate": 0.0006,
+ "loss": 4.418654918670654,
+ "step": 4557
+ },
+ {
+ "epoch": 63.30742358078603,
+ "grad_norm": 0.018394414335489273,
+ "learning_rate": 0.0006,
+ "loss": 4.38355827331543,
+ "step": 4558
+ },
+ {
+ "epoch": 63.32139737991266,
+ "grad_norm": 0.018317820504307747,
+ "learning_rate": 0.0006,
+ "loss": 4.456843376159668,
+ "step": 4559
+ },
+ {
+ "epoch": 63.3353711790393,
+ "grad_norm": 0.019225597381591797,
+ "learning_rate": 0.0006,
+ "loss": 4.232264518737793,
+ "step": 4560
+ },
+ {
+ "epoch": 63.34934497816594,
+ "grad_norm": 0.016663681715726852,
+ "learning_rate": 0.0006,
+ "loss": 4.393517017364502,
+ "step": 4561
+ },
+ {
+ "epoch": 63.36331877729258,
+ "grad_norm": 0.014975365251302719,
+ "learning_rate": 0.0006,
+ "loss": 4.375183582305908,
+ "step": 4562
+ },
+ {
+ "epoch": 63.377292576419215,
+ "grad_norm": 0.01771732047200203,
+ "learning_rate": 0.0006,
+ "loss": 4.3073410987854,
+ "step": 4563
+ },
+ {
+ "epoch": 63.391266375545854,
+ "grad_norm": 0.01663549244403839,
+ "learning_rate": 0.0006,
+ "loss": 4.311003684997559,
+ "step": 4564
+ },
+ {
+ "epoch": 63.40524017467249,
+ "grad_norm": 0.016301177442073822,
+ "learning_rate": 0.0006,
+ "loss": 4.326807022094727,
+ "step": 4565
+ },
+ {
+ "epoch": 63.419213973799124,
+ "grad_norm": 0.016781188547611237,
+ "learning_rate": 0.0006,
+ "loss": 4.306562423706055,
+ "step": 4566
+ },
+ {
+ "epoch": 63.43318777292576,
+ "grad_norm": 0.01767745055258274,
+ "learning_rate": 0.0006,
+ "loss": 4.367082595825195,
+ "step": 4567
+ },
+ {
+ "epoch": 63.4471615720524,
+ "grad_norm": 0.016728565096855164,
+ "learning_rate": 0.0006,
+ "loss": 4.327400207519531,
+ "step": 4568
+ },
+ {
+ "epoch": 63.46113537117904,
+ "grad_norm": 0.016289403662085533,
+ "learning_rate": 0.0006,
+ "loss": 4.392712116241455,
+ "step": 4569
+ },
+ {
+ "epoch": 63.47510917030568,
+ "grad_norm": 0.016233716160058975,
+ "learning_rate": 0.0006,
+ "loss": 4.354372024536133,
+ "step": 4570
+ },
+ {
+ "epoch": 63.48908296943232,
+ "grad_norm": 0.016547974199056625,
+ "learning_rate": 0.0006,
+ "loss": 4.309050559997559,
+ "step": 4571
+ },
+ {
+ "epoch": 63.50305676855895,
+ "grad_norm": 0.018466509878635406,
+ "learning_rate": 0.0006,
+ "loss": 4.318345069885254,
+ "step": 4572
+ },
+ {
+ "epoch": 63.51703056768559,
+ "grad_norm": 0.017685720697045326,
+ "learning_rate": 0.0006,
+ "loss": 4.399726867675781,
+ "step": 4573
+ },
+ {
+ "epoch": 63.531004366812226,
+ "grad_norm": 0.015801234170794487,
+ "learning_rate": 0.0006,
+ "loss": 4.328495502471924,
+ "step": 4574
+ },
+ {
+ "epoch": 63.544978165938865,
+ "grad_norm": 0.0188873540610075,
+ "learning_rate": 0.0006,
+ "loss": 4.253244400024414,
+ "step": 4575
+ },
+ {
+ "epoch": 63.5589519650655,
+ "grad_norm": 0.01868271268904209,
+ "learning_rate": 0.0006,
+ "loss": 4.463746070861816,
+ "step": 4576
+ },
+ {
+ "epoch": 63.57292576419214,
+ "grad_norm": 0.01844949834048748,
+ "learning_rate": 0.0006,
+ "loss": 4.400300025939941,
+ "step": 4577
+ },
+ {
+ "epoch": 63.58689956331878,
+ "grad_norm": 0.015256480313837528,
+ "learning_rate": 0.0006,
+ "loss": 4.445844650268555,
+ "step": 4578
+ },
+ {
+ "epoch": 63.60087336244541,
+ "grad_norm": 0.015305760316550732,
+ "learning_rate": 0.0006,
+ "loss": 4.317327499389648,
+ "step": 4579
+ },
+ {
+ "epoch": 63.61484716157205,
+ "grad_norm": 0.015200769528746605,
+ "learning_rate": 0.0006,
+ "loss": 4.25112247467041,
+ "step": 4580
+ },
+ {
+ "epoch": 63.62882096069869,
+ "grad_norm": 0.015919538214802742,
+ "learning_rate": 0.0006,
+ "loss": 4.277937889099121,
+ "step": 4581
+ },
+ {
+ "epoch": 63.64279475982533,
+ "grad_norm": 0.01572665013372898,
+ "learning_rate": 0.0006,
+ "loss": 4.302237510681152,
+ "step": 4582
+ },
+ {
+ "epoch": 63.65676855895197,
+ "grad_norm": 0.016428731381893158,
+ "learning_rate": 0.0006,
+ "loss": 4.36762809753418,
+ "step": 4583
+ },
+ {
+ "epoch": 63.670742358078606,
+ "grad_norm": 0.01706082746386528,
+ "learning_rate": 0.0006,
+ "loss": 4.236526966094971,
+ "step": 4584
+ },
+ {
+ "epoch": 63.68471615720524,
+ "grad_norm": 0.017850136384367943,
+ "learning_rate": 0.0006,
+ "loss": 4.264129638671875,
+ "step": 4585
+ },
+ {
+ "epoch": 63.698689956331876,
+ "grad_norm": 0.016079438850283623,
+ "learning_rate": 0.0006,
+ "loss": 4.449799060821533,
+ "step": 4586
+ },
+ {
+ "epoch": 63.712663755458514,
+ "grad_norm": 0.014996383339166641,
+ "learning_rate": 0.0006,
+ "loss": 4.404574394226074,
+ "step": 4587
+ },
+ {
+ "epoch": 63.72663755458515,
+ "grad_norm": 0.015910431742668152,
+ "learning_rate": 0.0006,
+ "loss": 4.348101615905762,
+ "step": 4588
+ },
+ {
+ "epoch": 63.74061135371179,
+ "grad_norm": 0.01630021072924137,
+ "learning_rate": 0.0006,
+ "loss": 4.3193817138671875,
+ "step": 4589
+ },
+ {
+ "epoch": 63.75458515283843,
+ "grad_norm": 0.014571773819625378,
+ "learning_rate": 0.0006,
+ "loss": 4.374402046203613,
+ "step": 4590
+ },
+ {
+ "epoch": 63.76855895196506,
+ "grad_norm": 0.01662345975637436,
+ "learning_rate": 0.0006,
+ "loss": 4.329075813293457,
+ "step": 4591
+ },
+ {
+ "epoch": 63.7825327510917,
+ "grad_norm": 0.016604198142886162,
+ "learning_rate": 0.0006,
+ "loss": 4.436126232147217,
+ "step": 4592
+ },
+ {
+ "epoch": 63.79650655021834,
+ "grad_norm": 0.016578884795308113,
+ "learning_rate": 0.0006,
+ "loss": 4.361468315124512,
+ "step": 4593
+ },
+ {
+ "epoch": 63.81048034934498,
+ "grad_norm": 0.016339825466275215,
+ "learning_rate": 0.0006,
+ "loss": 4.449892044067383,
+ "step": 4594
+ },
+ {
+ "epoch": 63.82445414847162,
+ "grad_norm": 0.016805455088615417,
+ "learning_rate": 0.0006,
+ "loss": 4.415866851806641,
+ "step": 4595
+ },
+ {
+ "epoch": 63.838427947598255,
+ "grad_norm": 0.017366213724017143,
+ "learning_rate": 0.0006,
+ "loss": 4.361730575561523,
+ "step": 4596
+ },
+ {
+ "epoch": 63.852401746724894,
+ "grad_norm": 0.015599003992974758,
+ "learning_rate": 0.0006,
+ "loss": 4.363637924194336,
+ "step": 4597
+ },
+ {
+ "epoch": 63.866375545851525,
+ "grad_norm": 0.015674632042646408,
+ "learning_rate": 0.0006,
+ "loss": 4.332045078277588,
+ "step": 4598
+ },
+ {
+ "epoch": 63.880349344978164,
+ "grad_norm": 0.015350443311035633,
+ "learning_rate": 0.0006,
+ "loss": 4.320063591003418,
+ "step": 4599
+ },
+ {
+ "epoch": 63.8943231441048,
+ "grad_norm": 0.015388146974146366,
+ "learning_rate": 0.0006,
+ "loss": 4.398199081420898,
+ "step": 4600
+ },
+ {
+ "epoch": 63.90829694323144,
+ "grad_norm": 0.01450140681117773,
+ "learning_rate": 0.0006,
+ "loss": 4.244076728820801,
+ "step": 4601
+ },
+ {
+ "epoch": 63.92227074235808,
+ "grad_norm": 0.01637491025030613,
+ "learning_rate": 0.0006,
+ "loss": 4.213549613952637,
+ "step": 4602
+ },
+ {
+ "epoch": 63.93624454148472,
+ "grad_norm": 0.0159814041107893,
+ "learning_rate": 0.0006,
+ "loss": 4.421775817871094,
+ "step": 4603
+ },
+ {
+ "epoch": 63.95021834061135,
+ "grad_norm": 0.017386827617883682,
+ "learning_rate": 0.0006,
+ "loss": 4.340023994445801,
+ "step": 4604
+ },
+ {
+ "epoch": 63.96419213973799,
+ "grad_norm": 0.019679656252264977,
+ "learning_rate": 0.0006,
+ "loss": 4.352534294128418,
+ "step": 4605
+ },
+ {
+ "epoch": 63.97816593886463,
+ "grad_norm": 0.022391516715288162,
+ "learning_rate": 0.0006,
+ "loss": 4.3538312911987305,
+ "step": 4606
+ },
+ {
+ "epoch": 63.992139737991266,
+ "grad_norm": 0.023281006142497063,
+ "learning_rate": 0.0006,
+ "loss": 4.432089328765869,
+ "step": 4607
+ },
+ {
+ "epoch": 64.0,
+ "grad_norm": 0.019916288554668427,
+ "learning_rate": 0.0006,
+ "loss": 4.466937065124512,
+ "step": 4608
+ },
+ {
+ "epoch": 64.0,
+ "eval_loss": 4.659008502960205,
+ "eval_runtime": 56.0019,
+ "eval_samples_per_second": 43.606,
+ "eval_steps_per_second": 1.375,
+ "step": 4608
+ },
+ {
+ "epoch": 64.01397379912663,
+ "grad_norm": 0.017351187765598297,
+ "learning_rate": 0.0006,
+ "loss": 4.294129371643066,
+ "step": 4609
+ },
+ {
+ "epoch": 64.02794759825328,
+ "grad_norm": 0.02405865117907524,
+ "learning_rate": 0.0006,
+ "loss": 4.266717910766602,
+ "step": 4610
+ },
+ {
+ "epoch": 64.04192139737991,
+ "grad_norm": 0.022158386185765266,
+ "learning_rate": 0.0006,
+ "loss": 4.395932197570801,
+ "step": 4611
+ },
+ {
+ "epoch": 64.05589519650655,
+ "grad_norm": 0.019238505512475967,
+ "learning_rate": 0.0006,
+ "loss": 4.266704082489014,
+ "step": 4612
+ },
+ {
+ "epoch": 64.06986899563319,
+ "grad_norm": 0.021670425310730934,
+ "learning_rate": 0.0006,
+ "loss": 4.3160552978515625,
+ "step": 4613
+ },
+ {
+ "epoch": 64.08384279475983,
+ "grad_norm": 0.02076641097664833,
+ "learning_rate": 0.0006,
+ "loss": 4.415626525878906,
+ "step": 4614
+ },
+ {
+ "epoch": 64.09781659388646,
+ "grad_norm": 0.018223078921437263,
+ "learning_rate": 0.0006,
+ "loss": 4.292387008666992,
+ "step": 4615
+ },
+ {
+ "epoch": 64.1117903930131,
+ "grad_norm": 0.017401937395334244,
+ "learning_rate": 0.0006,
+ "loss": 4.424524307250977,
+ "step": 4616
+ },
+ {
+ "epoch": 64.12576419213974,
+ "grad_norm": 0.01663743332028389,
+ "learning_rate": 0.0006,
+ "loss": 4.350032806396484,
+ "step": 4617
+ },
+ {
+ "epoch": 64.13973799126637,
+ "grad_norm": 0.018716629594564438,
+ "learning_rate": 0.0006,
+ "loss": 4.309157371520996,
+ "step": 4618
+ },
+ {
+ "epoch": 64.15371179039302,
+ "grad_norm": 0.019569825381040573,
+ "learning_rate": 0.0006,
+ "loss": 4.358292579650879,
+ "step": 4619
+ },
+ {
+ "epoch": 64.16768558951965,
+ "grad_norm": 0.018994422629475594,
+ "learning_rate": 0.0006,
+ "loss": 4.3338823318481445,
+ "step": 4620
+ },
+ {
+ "epoch": 64.18165938864628,
+ "grad_norm": 0.017690688371658325,
+ "learning_rate": 0.0006,
+ "loss": 4.414468765258789,
+ "step": 4621
+ },
+ {
+ "epoch": 64.19563318777293,
+ "grad_norm": 0.0170609001070261,
+ "learning_rate": 0.0006,
+ "loss": 4.4337568283081055,
+ "step": 4622
+ },
+ {
+ "epoch": 64.20960698689956,
+ "grad_norm": 0.016984131187200546,
+ "learning_rate": 0.0006,
+ "loss": 4.304839134216309,
+ "step": 4623
+ },
+ {
+ "epoch": 64.2235807860262,
+ "grad_norm": 0.01619867794215679,
+ "learning_rate": 0.0006,
+ "loss": 4.389623165130615,
+ "step": 4624
+ },
+ {
+ "epoch": 64.23755458515284,
+ "grad_norm": 0.015428110957145691,
+ "learning_rate": 0.0006,
+ "loss": 4.277515888214111,
+ "step": 4625
+ },
+ {
+ "epoch": 64.25152838427948,
+ "grad_norm": 0.015909532085061073,
+ "learning_rate": 0.0006,
+ "loss": 4.266029357910156,
+ "step": 4626
+ },
+ {
+ "epoch": 64.26550218340611,
+ "grad_norm": 0.014296560548245907,
+ "learning_rate": 0.0006,
+ "loss": 4.286067485809326,
+ "step": 4627
+ },
+ {
+ "epoch": 64.27947598253274,
+ "grad_norm": 0.014065971598029137,
+ "learning_rate": 0.0006,
+ "loss": 4.353528022766113,
+ "step": 4628
+ },
+ {
+ "epoch": 64.29344978165939,
+ "grad_norm": 0.014386293478310108,
+ "learning_rate": 0.0006,
+ "loss": 4.318715572357178,
+ "step": 4629
+ },
+ {
+ "epoch": 64.30742358078602,
+ "grad_norm": 0.014001123607158661,
+ "learning_rate": 0.0006,
+ "loss": 4.362497329711914,
+ "step": 4630
+ },
+ {
+ "epoch": 64.32139737991267,
+ "grad_norm": 0.013408979400992393,
+ "learning_rate": 0.0006,
+ "loss": 4.417397499084473,
+ "step": 4631
+ },
+ {
+ "epoch": 64.3353711790393,
+ "grad_norm": 0.014960093423724174,
+ "learning_rate": 0.0006,
+ "loss": 4.359710693359375,
+ "step": 4632
+ },
+ {
+ "epoch": 64.34934497816595,
+ "grad_norm": 0.014658777043223381,
+ "learning_rate": 0.0006,
+ "loss": 4.318822383880615,
+ "step": 4633
+ },
+ {
+ "epoch": 64.36331877729258,
+ "grad_norm": 0.013766744174063206,
+ "learning_rate": 0.0006,
+ "loss": 4.265679836273193,
+ "step": 4634
+ },
+ {
+ "epoch": 64.37729257641921,
+ "grad_norm": 0.013821255415678024,
+ "learning_rate": 0.0006,
+ "loss": 4.3452911376953125,
+ "step": 4635
+ },
+ {
+ "epoch": 64.39126637554585,
+ "grad_norm": 0.013541066087782383,
+ "learning_rate": 0.0006,
+ "loss": 4.429254531860352,
+ "step": 4636
+ },
+ {
+ "epoch": 64.40524017467249,
+ "grad_norm": 0.013985645025968552,
+ "learning_rate": 0.0006,
+ "loss": 4.304452896118164,
+ "step": 4637
+ },
+ {
+ "epoch": 64.41921397379913,
+ "grad_norm": 0.015079808421432972,
+ "learning_rate": 0.0006,
+ "loss": 4.296828746795654,
+ "step": 4638
+ },
+ {
+ "epoch": 64.43318777292576,
+ "grad_norm": 0.014896944165229797,
+ "learning_rate": 0.0006,
+ "loss": 4.299156665802002,
+ "step": 4639
+ },
+ {
+ "epoch": 64.44716157205241,
+ "grad_norm": 0.013770041987299919,
+ "learning_rate": 0.0006,
+ "loss": 4.351439952850342,
+ "step": 4640
+ },
+ {
+ "epoch": 64.46113537117904,
+ "grad_norm": 0.01531178504228592,
+ "learning_rate": 0.0006,
+ "loss": 4.373715877532959,
+ "step": 4641
+ },
+ {
+ "epoch": 64.47510917030567,
+ "grad_norm": 0.014024450443685055,
+ "learning_rate": 0.0006,
+ "loss": 4.368053436279297,
+ "step": 4642
+ },
+ {
+ "epoch": 64.48908296943232,
+ "grad_norm": 0.013864024542272091,
+ "learning_rate": 0.0006,
+ "loss": 4.378564834594727,
+ "step": 4643
+ },
+ {
+ "epoch": 64.50305676855895,
+ "grad_norm": 0.013925445266067982,
+ "learning_rate": 0.0006,
+ "loss": 4.261663436889648,
+ "step": 4644
+ },
+ {
+ "epoch": 64.5170305676856,
+ "grad_norm": 0.015011338517069817,
+ "learning_rate": 0.0006,
+ "loss": 4.337285995483398,
+ "step": 4645
+ },
+ {
+ "epoch": 64.53100436681223,
+ "grad_norm": 0.016941986978054047,
+ "learning_rate": 0.0006,
+ "loss": 4.336877822875977,
+ "step": 4646
+ },
+ {
+ "epoch": 64.54497816593886,
+ "grad_norm": 0.01719909906387329,
+ "learning_rate": 0.0006,
+ "loss": 4.149813175201416,
+ "step": 4647
+ },
+ {
+ "epoch": 64.5589519650655,
+ "grad_norm": 0.01608402095735073,
+ "learning_rate": 0.0006,
+ "loss": 4.191280364990234,
+ "step": 4648
+ },
+ {
+ "epoch": 64.57292576419214,
+ "grad_norm": 0.015284820459783077,
+ "learning_rate": 0.0006,
+ "loss": 4.35953426361084,
+ "step": 4649
+ },
+ {
+ "epoch": 64.58689956331878,
+ "grad_norm": 0.014855682849884033,
+ "learning_rate": 0.0006,
+ "loss": 4.308401107788086,
+ "step": 4650
+ },
+ {
+ "epoch": 64.60087336244541,
+ "grad_norm": 0.017553111538290977,
+ "learning_rate": 0.0006,
+ "loss": 4.279720306396484,
+ "step": 4651
+ },
+ {
+ "epoch": 64.61484716157206,
+ "grad_norm": 0.019506998360157013,
+ "learning_rate": 0.0006,
+ "loss": 4.319524765014648,
+ "step": 4652
+ },
+ {
+ "epoch": 64.62882096069869,
+ "grad_norm": 0.020293425768613815,
+ "learning_rate": 0.0006,
+ "loss": 4.340523719787598,
+ "step": 4653
+ },
+ {
+ "epoch": 64.64279475982532,
+ "grad_norm": 0.01664556749165058,
+ "learning_rate": 0.0006,
+ "loss": 4.296785831451416,
+ "step": 4654
+ },
+ {
+ "epoch": 64.65676855895197,
+ "grad_norm": 0.012592585757374763,
+ "learning_rate": 0.0006,
+ "loss": 4.464595317840576,
+ "step": 4655
+ },
+ {
+ "epoch": 64.6707423580786,
+ "grad_norm": 0.01549589168280363,
+ "learning_rate": 0.0006,
+ "loss": 4.275053024291992,
+ "step": 4656
+ },
+ {
+ "epoch": 64.68471615720524,
+ "grad_norm": 0.01705172471702099,
+ "learning_rate": 0.0006,
+ "loss": 4.388431549072266,
+ "step": 4657
+ },
+ {
+ "epoch": 64.69868995633188,
+ "grad_norm": 0.015438582748174667,
+ "learning_rate": 0.0006,
+ "loss": 4.312070846557617,
+ "step": 4658
+ },
+ {
+ "epoch": 64.71266375545852,
+ "grad_norm": 0.014448083005845547,
+ "learning_rate": 0.0006,
+ "loss": 4.237599849700928,
+ "step": 4659
+ },
+ {
+ "epoch": 64.72663755458515,
+ "grad_norm": 0.015267071314156055,
+ "learning_rate": 0.0006,
+ "loss": 4.29893684387207,
+ "step": 4660
+ },
+ {
+ "epoch": 64.74061135371178,
+ "grad_norm": 0.016615847125649452,
+ "learning_rate": 0.0006,
+ "loss": 4.432273864746094,
+ "step": 4661
+ },
+ {
+ "epoch": 64.75458515283843,
+ "grad_norm": 0.016584362834692,
+ "learning_rate": 0.0006,
+ "loss": 4.202266216278076,
+ "step": 4662
+ },
+ {
+ "epoch": 64.76855895196506,
+ "grad_norm": 0.018313994631171227,
+ "learning_rate": 0.0006,
+ "loss": 4.367966651916504,
+ "step": 4663
+ },
+ {
+ "epoch": 64.78253275109171,
+ "grad_norm": 0.017946481704711914,
+ "learning_rate": 0.0006,
+ "loss": 4.205172538757324,
+ "step": 4664
+ },
+ {
+ "epoch": 64.79650655021834,
+ "grad_norm": 0.017744354903697968,
+ "learning_rate": 0.0006,
+ "loss": 4.295661926269531,
+ "step": 4665
+ },
+ {
+ "epoch": 64.81048034934499,
+ "grad_norm": 0.016103824600577354,
+ "learning_rate": 0.0006,
+ "loss": 4.276543617248535,
+ "step": 4666
+ },
+ {
+ "epoch": 64.82445414847162,
+ "grad_norm": 0.014590012840926647,
+ "learning_rate": 0.0006,
+ "loss": 4.414247035980225,
+ "step": 4667
+ },
+ {
+ "epoch": 64.83842794759825,
+ "grad_norm": 0.01782386749982834,
+ "learning_rate": 0.0006,
+ "loss": 4.203220367431641,
+ "step": 4668
+ },
+ {
+ "epoch": 64.8524017467249,
+ "grad_norm": 0.022122733294963837,
+ "learning_rate": 0.0006,
+ "loss": 4.38078498840332,
+ "step": 4669
+ },
+ {
+ "epoch": 64.86637554585153,
+ "grad_norm": 0.0250293780118227,
+ "learning_rate": 0.0006,
+ "loss": 4.3310136795043945,
+ "step": 4670
+ },
+ {
+ "epoch": 64.88034934497817,
+ "grad_norm": 0.022438062354922295,
+ "learning_rate": 0.0006,
+ "loss": 4.428631782531738,
+ "step": 4671
+ },
+ {
+ "epoch": 64.8943231441048,
+ "grad_norm": 0.018518192693591118,
+ "learning_rate": 0.0006,
+ "loss": 4.428250312805176,
+ "step": 4672
+ },
+ {
+ "epoch": 64.90829694323143,
+ "grad_norm": 0.018105225637555122,
+ "learning_rate": 0.0006,
+ "loss": 4.336209297180176,
+ "step": 4673
+ },
+ {
+ "epoch": 64.92227074235808,
+ "grad_norm": 0.017564399167895317,
+ "learning_rate": 0.0006,
+ "loss": 4.416011810302734,
+ "step": 4674
+ },
+ {
+ "epoch": 64.93624454148471,
+ "grad_norm": 0.018747031688690186,
+ "learning_rate": 0.0006,
+ "loss": 4.419938087463379,
+ "step": 4675
+ },
+ {
+ "epoch": 64.95021834061136,
+ "grad_norm": 0.018686484545469284,
+ "learning_rate": 0.0006,
+ "loss": 4.347185134887695,
+ "step": 4676
+ },
+ {
+ "epoch": 64.96419213973799,
+ "grad_norm": 0.019329898059368134,
+ "learning_rate": 0.0006,
+ "loss": 4.343749046325684,
+ "step": 4677
+ },
+ {
+ "epoch": 64.97816593886463,
+ "grad_norm": 0.018517782911658287,
+ "learning_rate": 0.0006,
+ "loss": 4.264604568481445,
+ "step": 4678
+ },
+ {
+ "epoch": 64.99213973799127,
+ "grad_norm": 0.016579125076532364,
+ "learning_rate": 0.0006,
+ "loss": 4.303423881530762,
+ "step": 4679
+ },
+ {
+ "epoch": 65.0,
+ "grad_norm": 0.01617567613720894,
+ "learning_rate": 0.0006,
+ "loss": 4.326556205749512,
+ "step": 4680
+ },
+ {
+ "epoch": 65.0,
+ "eval_loss": 4.655026912689209,
+ "eval_runtime": 57.1299,
+ "eval_samples_per_second": 42.745,
+ "eval_steps_per_second": 1.348,
+ "step": 4680
+ },
+ {
+ "epoch": 65.01397379912663,
+ "grad_norm": 0.01607249490916729,
+ "learning_rate": 0.0006,
+ "loss": 4.276607036590576,
+ "step": 4681
+ },
+ {
+ "epoch": 65.02794759825328,
+ "grad_norm": 0.016779497265815735,
+ "learning_rate": 0.0006,
+ "loss": 4.412464618682861,
+ "step": 4682
+ },
+ {
+ "epoch": 65.04192139737991,
+ "grad_norm": 0.018161555752158165,
+ "learning_rate": 0.0006,
+ "loss": 4.199329376220703,
+ "step": 4683
+ },
+ {
+ "epoch": 65.05589519650655,
+ "grad_norm": 0.017508309334516525,
+ "learning_rate": 0.0006,
+ "loss": 4.217567443847656,
+ "step": 4684
+ },
+ {
+ "epoch": 65.06986899563319,
+ "grad_norm": 0.015716979280114174,
+ "learning_rate": 0.0006,
+ "loss": 4.322023391723633,
+ "step": 4685
+ },
+ {
+ "epoch": 65.08384279475983,
+ "grad_norm": 0.015113635919988155,
+ "learning_rate": 0.0006,
+ "loss": 4.333515167236328,
+ "step": 4686
+ },
+ {
+ "epoch": 65.09781659388646,
+ "grad_norm": 0.01437560748308897,
+ "learning_rate": 0.0006,
+ "loss": 4.295609474182129,
+ "step": 4687
+ },
+ {
+ "epoch": 65.1117903930131,
+ "grad_norm": 0.013074862770736217,
+ "learning_rate": 0.0006,
+ "loss": 4.314120769500732,
+ "step": 4688
+ },
+ {
+ "epoch": 65.12576419213974,
+ "grad_norm": 0.012908284552395344,
+ "learning_rate": 0.0006,
+ "loss": 4.216890811920166,
+ "step": 4689
+ },
+ {
+ "epoch": 65.13973799126637,
+ "grad_norm": 0.013001754879951477,
+ "learning_rate": 0.0006,
+ "loss": 4.287717342376709,
+ "step": 4690
+ },
+ {
+ "epoch": 65.15371179039302,
+ "grad_norm": 0.013694263994693756,
+ "learning_rate": 0.0006,
+ "loss": 4.333837985992432,
+ "step": 4691
+ },
+ {
+ "epoch": 65.16768558951965,
+ "grad_norm": 0.014461833983659744,
+ "learning_rate": 0.0006,
+ "loss": 4.323545455932617,
+ "step": 4692
+ },
+ {
+ "epoch": 65.18165938864628,
+ "grad_norm": 0.013670675456523895,
+ "learning_rate": 0.0006,
+ "loss": 4.3313493728637695,
+ "step": 4693
+ },
+ {
+ "epoch": 65.19563318777293,
+ "grad_norm": 0.013094350695610046,
+ "learning_rate": 0.0006,
+ "loss": 4.23858118057251,
+ "step": 4694
+ },
+ {
+ "epoch": 65.20960698689956,
+ "grad_norm": 0.013996944762766361,
+ "learning_rate": 0.0006,
+ "loss": 4.428705215454102,
+ "step": 4695
+ },
+ {
+ "epoch": 65.2235807860262,
+ "grad_norm": 0.013295048847794533,
+ "learning_rate": 0.0006,
+ "loss": 4.238950252532959,
+ "step": 4696
+ },
+ {
+ "epoch": 65.23755458515284,
+ "grad_norm": 0.013380616903305054,
+ "learning_rate": 0.0006,
+ "loss": 4.273185729980469,
+ "step": 4697
+ },
+ {
+ "epoch": 65.25152838427948,
+ "grad_norm": 0.015062890015542507,
+ "learning_rate": 0.0006,
+ "loss": 4.29472017288208,
+ "step": 4698
+ },
+ {
+ "epoch": 65.26550218340611,
+ "grad_norm": 0.017317263409495354,
+ "learning_rate": 0.0006,
+ "loss": 4.337954998016357,
+ "step": 4699
+ },
+ {
+ "epoch": 65.27947598253274,
+ "grad_norm": 0.017698992043733597,
+ "learning_rate": 0.0006,
+ "loss": 4.336437225341797,
+ "step": 4700
+ },
+ {
+ "epoch": 65.29344978165939,
+ "grad_norm": 0.01628575474023819,
+ "learning_rate": 0.0006,
+ "loss": 4.302372455596924,
+ "step": 4701
+ },
+ {
+ "epoch": 65.30742358078602,
+ "grad_norm": 0.014531854540109634,
+ "learning_rate": 0.0006,
+ "loss": 4.256450653076172,
+ "step": 4702
+ },
+ {
+ "epoch": 65.32139737991267,
+ "grad_norm": 0.014549105428159237,
+ "learning_rate": 0.0006,
+ "loss": 4.302240371704102,
+ "step": 4703
+ },
+ {
+ "epoch": 65.3353711790393,
+ "grad_norm": 0.014881663955748081,
+ "learning_rate": 0.0006,
+ "loss": 4.3316121101379395,
+ "step": 4704
+ },
+ {
+ "epoch": 65.34934497816595,
+ "grad_norm": 0.014513437636196613,
+ "learning_rate": 0.0006,
+ "loss": 4.441212177276611,
+ "step": 4705
+ },
+ {
+ "epoch": 65.36331877729258,
+ "grad_norm": 0.014122087508440018,
+ "learning_rate": 0.0006,
+ "loss": 4.343296527862549,
+ "step": 4706
+ },
+ {
+ "epoch": 65.37729257641921,
+ "grad_norm": 0.014817348681390285,
+ "learning_rate": 0.0006,
+ "loss": 4.195674896240234,
+ "step": 4707
+ },
+ {
+ "epoch": 65.39126637554585,
+ "grad_norm": 0.01787056028842926,
+ "learning_rate": 0.0006,
+ "loss": 4.3507914543151855,
+ "step": 4708
+ },
+ {
+ "epoch": 65.40524017467249,
+ "grad_norm": 0.015295136719942093,
+ "learning_rate": 0.0006,
+ "loss": 4.234372138977051,
+ "step": 4709
+ },
+ {
+ "epoch": 65.41921397379913,
+ "grad_norm": 0.014324234798550606,
+ "learning_rate": 0.0006,
+ "loss": 4.2907867431640625,
+ "step": 4710
+ },
+ {
+ "epoch": 65.43318777292576,
+ "grad_norm": 0.016090670600533485,
+ "learning_rate": 0.0006,
+ "loss": 4.177244186401367,
+ "step": 4711
+ },
+ {
+ "epoch": 65.44716157205241,
+ "grad_norm": 0.0174604132771492,
+ "learning_rate": 0.0006,
+ "loss": 4.335577964782715,
+ "step": 4712
+ },
+ {
+ "epoch": 65.46113537117904,
+ "grad_norm": 0.016731899231672287,
+ "learning_rate": 0.0006,
+ "loss": 4.30197811126709,
+ "step": 4713
+ },
+ {
+ "epoch": 65.47510917030567,
+ "grad_norm": 0.017423735931515694,
+ "learning_rate": 0.0006,
+ "loss": 4.3528852462768555,
+ "step": 4714
+ },
+ {
+ "epoch": 65.48908296943232,
+ "grad_norm": 0.01780976541340351,
+ "learning_rate": 0.0006,
+ "loss": 4.185036659240723,
+ "step": 4715
+ },
+ {
+ "epoch": 65.50305676855895,
+ "grad_norm": 0.019977200776338577,
+ "learning_rate": 0.0006,
+ "loss": 4.328766345977783,
+ "step": 4716
+ },
+ {
+ "epoch": 65.5170305676856,
+ "grad_norm": 0.020284287631511688,
+ "learning_rate": 0.0006,
+ "loss": 4.285494804382324,
+ "step": 4717
+ },
+ {
+ "epoch": 65.53100436681223,
+ "grad_norm": 0.019168002530932426,
+ "learning_rate": 0.0006,
+ "loss": 4.328560829162598,
+ "step": 4718
+ },
+ {
+ "epoch": 65.54497816593886,
+ "grad_norm": 0.019217776134610176,
+ "learning_rate": 0.0006,
+ "loss": 4.189998626708984,
+ "step": 4719
+ },
+ {
+ "epoch": 65.5589519650655,
+ "grad_norm": 0.018960488960146904,
+ "learning_rate": 0.0006,
+ "loss": 4.4995012283325195,
+ "step": 4720
+ },
+ {
+ "epoch": 65.57292576419214,
+ "grad_norm": 0.01941288262605667,
+ "learning_rate": 0.0006,
+ "loss": 4.26570987701416,
+ "step": 4721
+ },
+ {
+ "epoch": 65.58689956331878,
+ "grad_norm": 0.020263448357582092,
+ "learning_rate": 0.0006,
+ "loss": 4.411645889282227,
+ "step": 4722
+ },
+ {
+ "epoch": 65.60087336244541,
+ "grad_norm": 0.02144799567759037,
+ "learning_rate": 0.0006,
+ "loss": 4.163675308227539,
+ "step": 4723
+ },
+ {
+ "epoch": 65.61484716157206,
+ "grad_norm": 0.019790446385741234,
+ "learning_rate": 0.0006,
+ "loss": 4.2361063957214355,
+ "step": 4724
+ },
+ {
+ "epoch": 65.62882096069869,
+ "grad_norm": 0.01829805225133896,
+ "learning_rate": 0.0006,
+ "loss": 4.449810028076172,
+ "step": 4725
+ },
+ {
+ "epoch": 65.64279475982532,
+ "grad_norm": 0.01571105420589447,
+ "learning_rate": 0.0006,
+ "loss": 4.281406879425049,
+ "step": 4726
+ },
+ {
+ "epoch": 65.65676855895197,
+ "grad_norm": 0.016477147117257118,
+ "learning_rate": 0.0006,
+ "loss": 4.293408393859863,
+ "step": 4727
+ },
+ {
+ "epoch": 65.6707423580786,
+ "grad_norm": 0.016766032204031944,
+ "learning_rate": 0.0006,
+ "loss": 4.378252983093262,
+ "step": 4728
+ },
+ {
+ "epoch": 65.68471615720524,
+ "grad_norm": 0.0159933939576149,
+ "learning_rate": 0.0006,
+ "loss": 4.3065667152404785,
+ "step": 4729
+ },
+ {
+ "epoch": 65.69868995633188,
+ "grad_norm": 0.01658209227025509,
+ "learning_rate": 0.0006,
+ "loss": 4.174026966094971,
+ "step": 4730
+ },
+ {
+ "epoch": 65.71266375545852,
+ "grad_norm": 0.015081380493938923,
+ "learning_rate": 0.0006,
+ "loss": 4.360222339630127,
+ "step": 4731
+ },
+ {
+ "epoch": 65.72663755458515,
+ "grad_norm": 0.015110890381038189,
+ "learning_rate": 0.0006,
+ "loss": 4.218810558319092,
+ "step": 4732
+ },
+ {
+ "epoch": 65.74061135371178,
+ "grad_norm": 0.015300953760743141,
+ "learning_rate": 0.0006,
+ "loss": 4.331979274749756,
+ "step": 4733
+ },
+ {
+ "epoch": 65.75458515283843,
+ "grad_norm": 0.01715504191815853,
+ "learning_rate": 0.0006,
+ "loss": 4.349112033843994,
+ "step": 4734
+ },
+ {
+ "epoch": 65.76855895196506,
+ "grad_norm": 0.015870366245508194,
+ "learning_rate": 0.0006,
+ "loss": 4.329850196838379,
+ "step": 4735
+ },
+ {
+ "epoch": 65.78253275109171,
+ "grad_norm": 0.01541937980800867,
+ "learning_rate": 0.0006,
+ "loss": 4.3400163650512695,
+ "step": 4736
+ },
+ {
+ "epoch": 65.79650655021834,
+ "grad_norm": 0.01738942787051201,
+ "learning_rate": 0.0006,
+ "loss": 4.220798492431641,
+ "step": 4737
+ },
+ {
+ "epoch": 65.81048034934499,
+ "grad_norm": 0.01685990020632744,
+ "learning_rate": 0.0006,
+ "loss": 4.177238464355469,
+ "step": 4738
+ },
+ {
+ "epoch": 65.82445414847162,
+ "grad_norm": 0.016183465719223022,
+ "learning_rate": 0.0006,
+ "loss": 4.327216625213623,
+ "step": 4739
+ },
+ {
+ "epoch": 65.83842794759825,
+ "grad_norm": 0.01677694357931614,
+ "learning_rate": 0.0006,
+ "loss": 4.27766227722168,
+ "step": 4740
+ },
+ {
+ "epoch": 65.8524017467249,
+ "grad_norm": 0.01645066775381565,
+ "learning_rate": 0.0006,
+ "loss": 4.288347244262695,
+ "step": 4741
+ },
+ {
+ "epoch": 65.86637554585153,
+ "grad_norm": 0.018034635111689568,
+ "learning_rate": 0.0006,
+ "loss": 4.292458534240723,
+ "step": 4742
+ },
+ {
+ "epoch": 65.88034934497817,
+ "grad_norm": 0.017973346635699272,
+ "learning_rate": 0.0006,
+ "loss": 4.3117265701293945,
+ "step": 4743
+ },
+ {
+ "epoch": 65.8943231441048,
+ "grad_norm": 0.016605669632554054,
+ "learning_rate": 0.0006,
+ "loss": 4.280615329742432,
+ "step": 4744
+ },
+ {
+ "epoch": 65.90829694323143,
+ "grad_norm": 0.016298074275255203,
+ "learning_rate": 0.0006,
+ "loss": 4.301933288574219,
+ "step": 4745
+ },
+ {
+ "epoch": 65.92227074235808,
+ "grad_norm": 0.015383216552436352,
+ "learning_rate": 0.0006,
+ "loss": 4.352383136749268,
+ "step": 4746
+ },
+ {
+ "epoch": 65.93624454148471,
+ "grad_norm": 0.016222182661294937,
+ "learning_rate": 0.0006,
+ "loss": 4.305725574493408,
+ "step": 4747
+ },
+ {
+ "epoch": 65.95021834061136,
+ "grad_norm": 0.01691329851746559,
+ "learning_rate": 0.0006,
+ "loss": 4.29502010345459,
+ "step": 4748
+ },
+ {
+ "epoch": 65.96419213973799,
+ "grad_norm": 0.017637677490711212,
+ "learning_rate": 0.0006,
+ "loss": 4.3133039474487305,
+ "step": 4749
+ },
+ {
+ "epoch": 65.97816593886463,
+ "grad_norm": 0.021362047642469406,
+ "learning_rate": 0.0006,
+ "loss": 4.458834648132324,
+ "step": 4750
+ },
+ {
+ "epoch": 65.99213973799127,
+ "grad_norm": 0.023553011938929558,
+ "learning_rate": 0.0006,
+ "loss": 4.298727989196777,
+ "step": 4751
+ },
+ {
+ "epoch": 66.0,
+ "grad_norm": 0.02154950052499771,
+ "learning_rate": 0.0006,
+ "loss": 4.403392791748047,
+ "step": 4752
+ },
+ {
+ "epoch": 66.0,
+ "eval_loss": 4.724340438842773,
+ "eval_runtime": 56.9555,
+ "eval_samples_per_second": 42.876,
+ "eval_steps_per_second": 1.352,
+ "step": 4752
+ },
+ {
+ "epoch": 66.01397379912663,
+ "grad_norm": 0.01916784606873989,
+ "learning_rate": 0.0006,
+ "loss": 4.22931432723999,
+ "step": 4753
+ },
+ {
+ "epoch": 66.02794759825328,
+ "grad_norm": 0.017492054030299187,
+ "learning_rate": 0.0006,
+ "loss": 4.271522521972656,
+ "step": 4754
+ },
+ {
+ "epoch": 66.04192139737991,
+ "grad_norm": 0.018408095464110374,
+ "learning_rate": 0.0006,
+ "loss": 4.337705135345459,
+ "step": 4755
+ },
+ {
+ "epoch": 66.05589519650655,
+ "grad_norm": 0.019220108166337013,
+ "learning_rate": 0.0006,
+ "loss": 4.280340194702148,
+ "step": 4756
+ },
+ {
+ "epoch": 66.06986899563319,
+ "grad_norm": 0.01773735322058201,
+ "learning_rate": 0.0006,
+ "loss": 4.390201568603516,
+ "step": 4757
+ },
+ {
+ "epoch": 66.08384279475983,
+ "grad_norm": 0.01679125428199768,
+ "learning_rate": 0.0006,
+ "loss": 4.299282073974609,
+ "step": 4758
+ },
+ {
+ "epoch": 66.09781659388646,
+ "grad_norm": 0.01479059923440218,
+ "learning_rate": 0.0006,
+ "loss": 4.472630500793457,
+ "step": 4759
+ },
+ {
+ "epoch": 66.1117903930131,
+ "grad_norm": 0.015130104497075081,
+ "learning_rate": 0.0006,
+ "loss": 4.301194190979004,
+ "step": 4760
+ },
+ {
+ "epoch": 66.12576419213974,
+ "grad_norm": 0.0158979594707489,
+ "learning_rate": 0.0006,
+ "loss": 4.375751972198486,
+ "step": 4761
+ },
+ {
+ "epoch": 66.13973799126637,
+ "grad_norm": 0.014432408846914768,
+ "learning_rate": 0.0006,
+ "loss": 4.279788017272949,
+ "step": 4762
+ },
+ {
+ "epoch": 66.15371179039302,
+ "grad_norm": 0.013026784174144268,
+ "learning_rate": 0.0006,
+ "loss": 4.255035400390625,
+ "step": 4763
+ },
+ {
+ "epoch": 66.16768558951965,
+ "grad_norm": 0.014422653242945671,
+ "learning_rate": 0.0006,
+ "loss": 4.261832237243652,
+ "step": 4764
+ },
+ {
+ "epoch": 66.18165938864628,
+ "grad_norm": 0.014613019302487373,
+ "learning_rate": 0.0006,
+ "loss": 4.382184028625488,
+ "step": 4765
+ },
+ {
+ "epoch": 66.19563318777293,
+ "grad_norm": 0.014989894814789295,
+ "learning_rate": 0.0006,
+ "loss": 4.40528678894043,
+ "step": 4766
+ },
+ {
+ "epoch": 66.20960698689956,
+ "grad_norm": 0.018190350383520126,
+ "learning_rate": 0.0006,
+ "loss": 4.302414894104004,
+ "step": 4767
+ },
+ {
+ "epoch": 66.2235807860262,
+ "grad_norm": 0.020887890830636024,
+ "learning_rate": 0.0006,
+ "loss": 4.289920806884766,
+ "step": 4768
+ },
+ {
+ "epoch": 66.23755458515284,
+ "grad_norm": 0.02376876398921013,
+ "learning_rate": 0.0006,
+ "loss": 4.297855377197266,
+ "step": 4769
+ },
+ {
+ "epoch": 66.25152838427948,
+ "grad_norm": 0.02152269519865513,
+ "learning_rate": 0.0006,
+ "loss": 4.1972832679748535,
+ "step": 4770
+ },
+ {
+ "epoch": 66.26550218340611,
+ "grad_norm": 0.015572802163660526,
+ "learning_rate": 0.0006,
+ "loss": 4.307487487792969,
+ "step": 4771
+ },
+ {
+ "epoch": 66.27947598253274,
+ "grad_norm": 0.015228749252855778,
+ "learning_rate": 0.0006,
+ "loss": 4.103250503540039,
+ "step": 4772
+ },
+ {
+ "epoch": 66.29344978165939,
+ "grad_norm": 0.016925504431128502,
+ "learning_rate": 0.0006,
+ "loss": 4.247461318969727,
+ "step": 4773
+ },
+ {
+ "epoch": 66.30742358078602,
+ "grad_norm": 0.01613711006939411,
+ "learning_rate": 0.0006,
+ "loss": 4.4699273109436035,
+ "step": 4774
+ },
+ {
+ "epoch": 66.32139737991267,
+ "grad_norm": 0.018332140520215034,
+ "learning_rate": 0.0006,
+ "loss": 4.22969388961792,
+ "step": 4775
+ },
+ {
+ "epoch": 66.3353711790393,
+ "grad_norm": 0.019575074315071106,
+ "learning_rate": 0.0006,
+ "loss": 4.29397439956665,
+ "step": 4776
+ },
+ {
+ "epoch": 66.34934497816595,
+ "grad_norm": 0.018177419900894165,
+ "learning_rate": 0.0006,
+ "loss": 4.221899509429932,
+ "step": 4777
+ },
+ {
+ "epoch": 66.36331877729258,
+ "grad_norm": 0.014292889274656773,
+ "learning_rate": 0.0006,
+ "loss": 4.325457572937012,
+ "step": 4778
+ },
+ {
+ "epoch": 66.37729257641921,
+ "grad_norm": 0.015098796226084232,
+ "learning_rate": 0.0006,
+ "loss": 4.364837169647217,
+ "step": 4779
+ },
+ {
+ "epoch": 66.39126637554585,
+ "grad_norm": 0.013992635533213615,
+ "learning_rate": 0.0006,
+ "loss": 4.299082279205322,
+ "step": 4780
+ },
+ {
+ "epoch": 66.40524017467249,
+ "grad_norm": 0.01504511572420597,
+ "learning_rate": 0.0006,
+ "loss": 4.230877876281738,
+ "step": 4781
+ },
+ {
+ "epoch": 66.41921397379913,
+ "grad_norm": 0.014875639230012894,
+ "learning_rate": 0.0006,
+ "loss": 4.368405818939209,
+ "step": 4782
+ },
+ {
+ "epoch": 66.43318777292576,
+ "grad_norm": 0.014306887984275818,
+ "learning_rate": 0.0006,
+ "loss": 4.389184474945068,
+ "step": 4783
+ },
+ {
+ "epoch": 66.44716157205241,
+ "grad_norm": 0.012959563173353672,
+ "learning_rate": 0.0006,
+ "loss": 4.37934684753418,
+ "step": 4784
+ },
+ {
+ "epoch": 66.46113537117904,
+ "grad_norm": 0.013206703588366508,
+ "learning_rate": 0.0006,
+ "loss": 4.442799091339111,
+ "step": 4785
+ },
+ {
+ "epoch": 66.47510917030567,
+ "grad_norm": 0.014848909340798855,
+ "learning_rate": 0.0006,
+ "loss": 4.203481674194336,
+ "step": 4786
+ },
+ {
+ "epoch": 66.48908296943232,
+ "grad_norm": 0.015637682750821114,
+ "learning_rate": 0.0006,
+ "loss": 4.27552604675293,
+ "step": 4787
+ },
+ {
+ "epoch": 66.50305676855895,
+ "grad_norm": 0.016017816960811615,
+ "learning_rate": 0.0006,
+ "loss": 4.296402931213379,
+ "step": 4788
+ },
+ {
+ "epoch": 66.5170305676856,
+ "grad_norm": 0.01577814109623432,
+ "learning_rate": 0.0006,
+ "loss": 4.301733016967773,
+ "step": 4789
+ },
+ {
+ "epoch": 66.53100436681223,
+ "grad_norm": 0.015293175354599953,
+ "learning_rate": 0.0006,
+ "loss": 4.369877815246582,
+ "step": 4790
+ },
+ {
+ "epoch": 66.54497816593886,
+ "grad_norm": 0.014262278564274311,
+ "learning_rate": 0.0006,
+ "loss": 4.3314738273620605,
+ "step": 4791
+ },
+ {
+ "epoch": 66.5589519650655,
+ "grad_norm": 0.015123482793569565,
+ "learning_rate": 0.0006,
+ "loss": 4.313591957092285,
+ "step": 4792
+ },
+ {
+ "epoch": 66.57292576419214,
+ "grad_norm": 0.015458766371011734,
+ "learning_rate": 0.0006,
+ "loss": 4.3045854568481445,
+ "step": 4793
+ },
+ {
+ "epoch": 66.58689956331878,
+ "grad_norm": 0.017217233777046204,
+ "learning_rate": 0.0006,
+ "loss": 4.269944190979004,
+ "step": 4794
+ },
+ {
+ "epoch": 66.60087336244541,
+ "grad_norm": 0.01962282694876194,
+ "learning_rate": 0.0006,
+ "loss": 4.271735191345215,
+ "step": 4795
+ },
+ {
+ "epoch": 66.61484716157206,
+ "grad_norm": 0.01917066052556038,
+ "learning_rate": 0.0006,
+ "loss": 4.288168430328369,
+ "step": 4796
+ },
+ {
+ "epoch": 66.62882096069869,
+ "grad_norm": 0.018211524933576584,
+ "learning_rate": 0.0006,
+ "loss": 4.280660629272461,
+ "step": 4797
+ },
+ {
+ "epoch": 66.64279475982532,
+ "grad_norm": 0.01936524175107479,
+ "learning_rate": 0.0006,
+ "loss": 4.3710551261901855,
+ "step": 4798
+ },
+ {
+ "epoch": 66.65676855895197,
+ "grad_norm": 0.01882351003587246,
+ "learning_rate": 0.0006,
+ "loss": 4.342750549316406,
+ "step": 4799
+ },
+ {
+ "epoch": 66.6707423580786,
+ "grad_norm": 0.017149122431874275,
+ "learning_rate": 0.0006,
+ "loss": 4.32134485244751,
+ "step": 4800
+ },
+ {
+ "epoch": 66.68471615720524,
+ "grad_norm": 0.01699327677488327,
+ "learning_rate": 0.0006,
+ "loss": 4.248893737792969,
+ "step": 4801
+ },
+ {
+ "epoch": 66.69868995633188,
+ "grad_norm": 0.016873255372047424,
+ "learning_rate": 0.0006,
+ "loss": 4.277350425720215,
+ "step": 4802
+ },
+ {
+ "epoch": 66.71266375545852,
+ "grad_norm": 0.016592837870121002,
+ "learning_rate": 0.0006,
+ "loss": 4.221665859222412,
+ "step": 4803
+ },
+ {
+ "epoch": 66.72663755458515,
+ "grad_norm": 0.015887131914496422,
+ "learning_rate": 0.0006,
+ "loss": 4.328040599822998,
+ "step": 4804
+ },
+ {
+ "epoch": 66.74061135371178,
+ "grad_norm": 0.014057286083698273,
+ "learning_rate": 0.0006,
+ "loss": 4.3294548988342285,
+ "step": 4805
+ },
+ {
+ "epoch": 66.75458515283843,
+ "grad_norm": 0.013135905377566814,
+ "learning_rate": 0.0006,
+ "loss": 4.254383563995361,
+ "step": 4806
+ },
+ {
+ "epoch": 66.76855895196506,
+ "grad_norm": 0.013632557354867458,
+ "learning_rate": 0.0006,
+ "loss": 4.342815399169922,
+ "step": 4807
+ },
+ {
+ "epoch": 66.78253275109171,
+ "grad_norm": 0.015512688085436821,
+ "learning_rate": 0.0006,
+ "loss": 4.297835350036621,
+ "step": 4808
+ },
+ {
+ "epoch": 66.79650655021834,
+ "grad_norm": 0.01621832884848118,
+ "learning_rate": 0.0006,
+ "loss": 4.314773082733154,
+ "step": 4809
+ },
+ {
+ "epoch": 66.81048034934499,
+ "grad_norm": 0.016196032986044884,
+ "learning_rate": 0.0006,
+ "loss": 4.29209566116333,
+ "step": 4810
+ },
+ {
+ "epoch": 66.82445414847162,
+ "grad_norm": 0.015463353134691715,
+ "learning_rate": 0.0006,
+ "loss": 4.262795448303223,
+ "step": 4811
+ },
+ {
+ "epoch": 66.83842794759825,
+ "grad_norm": 0.016143787652254105,
+ "learning_rate": 0.0006,
+ "loss": 4.274284362792969,
+ "step": 4812
+ },
+ {
+ "epoch": 66.8524017467249,
+ "grad_norm": 0.01610061153769493,
+ "learning_rate": 0.0006,
+ "loss": 4.337160587310791,
+ "step": 4813
+ },
+ {
+ "epoch": 66.86637554585153,
+ "grad_norm": 0.01582838036119938,
+ "learning_rate": 0.0006,
+ "loss": 4.192723274230957,
+ "step": 4814
+ },
+ {
+ "epoch": 66.88034934497817,
+ "grad_norm": 0.015081540681421757,
+ "learning_rate": 0.0006,
+ "loss": 4.370708465576172,
+ "step": 4815
+ },
+ {
+ "epoch": 66.8943231441048,
+ "grad_norm": 0.015929723158478737,
+ "learning_rate": 0.0006,
+ "loss": 4.340244293212891,
+ "step": 4816
+ },
+ {
+ "epoch": 66.90829694323143,
+ "grad_norm": 0.01784711331129074,
+ "learning_rate": 0.0006,
+ "loss": 4.260406970977783,
+ "step": 4817
+ },
+ {
+ "epoch": 66.92227074235808,
+ "grad_norm": 0.017434703186154366,
+ "learning_rate": 0.0006,
+ "loss": 4.286900043487549,
+ "step": 4818
+ },
+ {
+ "epoch": 66.93624454148471,
+ "grad_norm": 0.016262540593743324,
+ "learning_rate": 0.0006,
+ "loss": 4.292142391204834,
+ "step": 4819
+ },
+ {
+ "epoch": 66.95021834061136,
+ "grad_norm": 0.016122829169034958,
+ "learning_rate": 0.0006,
+ "loss": 4.291485786437988,
+ "step": 4820
+ },
+ {
+ "epoch": 66.96419213973799,
+ "grad_norm": 0.014361178502440453,
+ "learning_rate": 0.0006,
+ "loss": 4.275425910949707,
+ "step": 4821
+ },
+ {
+ "epoch": 66.97816593886463,
+ "grad_norm": 0.01373987551778555,
+ "learning_rate": 0.0006,
+ "loss": 4.223120212554932,
+ "step": 4822
+ },
+ {
+ "epoch": 66.99213973799127,
+ "grad_norm": 0.015687376260757446,
+ "learning_rate": 0.0006,
+ "loss": 4.355381011962891,
+ "step": 4823
+ },
+ {
+ "epoch": 67.0,
+ "grad_norm": 0.01998465321958065,
+ "learning_rate": 0.0006,
+ "loss": 4.264053821563721,
+ "step": 4824
+ },
+ {
+ "epoch": 67.0,
+ "eval_loss": 4.662398338317871,
+ "eval_runtime": 56.7928,
+ "eval_samples_per_second": 42.998,
+ "eval_steps_per_second": 1.356,
+ "step": 4824
+ },
+ {
+ "epoch": 67.01397379912663,
+ "grad_norm": 0.021736836060881615,
+ "learning_rate": 0.0006,
+ "loss": 4.243354797363281,
+ "step": 4825
+ },
+ {
+ "epoch": 67.02794759825328,
+ "grad_norm": 0.021476037800312042,
+ "learning_rate": 0.0006,
+ "loss": 4.218207836151123,
+ "step": 4826
+ },
+ {
+ "epoch": 67.04192139737991,
+ "grad_norm": 0.017188388854265213,
+ "learning_rate": 0.0006,
+ "loss": 4.18048095703125,
+ "step": 4827
+ },
+ {
+ "epoch": 67.05589519650655,
+ "grad_norm": 0.017293719574809074,
+ "learning_rate": 0.0006,
+ "loss": 4.239145278930664,
+ "step": 4828
+ },
+ {
+ "epoch": 67.06986899563319,
+ "grad_norm": 0.017383819445967674,
+ "learning_rate": 0.0006,
+ "loss": 4.114765167236328,
+ "step": 4829
+ },
+ {
+ "epoch": 67.08384279475983,
+ "grad_norm": 0.01651529036462307,
+ "learning_rate": 0.0006,
+ "loss": 4.351023197174072,
+ "step": 4830
+ },
+ {
+ "epoch": 67.09781659388646,
+ "grad_norm": 0.015893027186393738,
+ "learning_rate": 0.0006,
+ "loss": 4.261432647705078,
+ "step": 4831
+ },
+ {
+ "epoch": 67.1117903930131,
+ "grad_norm": 0.013381517492234707,
+ "learning_rate": 0.0006,
+ "loss": 4.172064781188965,
+ "step": 4832
+ },
+ {
+ "epoch": 67.12576419213974,
+ "grad_norm": 0.014642654918134212,
+ "learning_rate": 0.0006,
+ "loss": 4.356171131134033,
+ "step": 4833
+ },
+ {
+ "epoch": 67.13973799126637,
+ "grad_norm": 0.01425282284617424,
+ "learning_rate": 0.0006,
+ "loss": 4.303495407104492,
+ "step": 4834
+ },
+ {
+ "epoch": 67.15371179039302,
+ "grad_norm": 0.014574599452316761,
+ "learning_rate": 0.0006,
+ "loss": 4.2850847244262695,
+ "step": 4835
+ },
+ {
+ "epoch": 67.16768558951965,
+ "grad_norm": 0.012676282785832882,
+ "learning_rate": 0.0006,
+ "loss": 4.297845840454102,
+ "step": 4836
+ },
+ {
+ "epoch": 67.18165938864628,
+ "grad_norm": 0.013948136940598488,
+ "learning_rate": 0.0006,
+ "loss": 4.324530601501465,
+ "step": 4837
+ },
+ {
+ "epoch": 67.19563318777293,
+ "grad_norm": 0.014195773750543594,
+ "learning_rate": 0.0006,
+ "loss": 4.31236457824707,
+ "step": 4838
+ },
+ {
+ "epoch": 67.20960698689956,
+ "grad_norm": 0.013631382025778294,
+ "learning_rate": 0.0006,
+ "loss": 4.086362838745117,
+ "step": 4839
+ },
+ {
+ "epoch": 67.2235807860262,
+ "grad_norm": 0.012798587791621685,
+ "learning_rate": 0.0006,
+ "loss": 4.328471660614014,
+ "step": 4840
+ },
+ {
+ "epoch": 67.23755458515284,
+ "grad_norm": 0.014443321153521538,
+ "learning_rate": 0.0006,
+ "loss": 4.278019905090332,
+ "step": 4841
+ },
+ {
+ "epoch": 67.25152838427948,
+ "grad_norm": 0.015355420298874378,
+ "learning_rate": 0.0006,
+ "loss": 4.333611011505127,
+ "step": 4842
+ },
+ {
+ "epoch": 67.26550218340611,
+ "grad_norm": 0.016487296670675278,
+ "learning_rate": 0.0006,
+ "loss": 4.351336479187012,
+ "step": 4843
+ },
+ {
+ "epoch": 67.27947598253274,
+ "grad_norm": 0.017422541975975037,
+ "learning_rate": 0.0006,
+ "loss": 4.386096477508545,
+ "step": 4844
+ },
+ {
+ "epoch": 67.29344978165939,
+ "grad_norm": 0.015790918841958046,
+ "learning_rate": 0.0006,
+ "loss": 4.301377773284912,
+ "step": 4845
+ },
+ {
+ "epoch": 67.30742358078602,
+ "grad_norm": 0.016427364200353622,
+ "learning_rate": 0.0006,
+ "loss": 4.2838945388793945,
+ "step": 4846
+ },
+ {
+ "epoch": 67.32139737991267,
+ "grad_norm": 0.017050761729478836,
+ "learning_rate": 0.0006,
+ "loss": 4.282417297363281,
+ "step": 4847
+ },
+ {
+ "epoch": 67.3353711790393,
+ "grad_norm": 0.015006253495812416,
+ "learning_rate": 0.0006,
+ "loss": 4.352236270904541,
+ "step": 4848
+ },
+ {
+ "epoch": 67.34934497816595,
+ "grad_norm": 0.012996627949178219,
+ "learning_rate": 0.0006,
+ "loss": 4.309242248535156,
+ "step": 4849
+ },
+ {
+ "epoch": 67.36331877729258,
+ "grad_norm": 0.01367615070194006,
+ "learning_rate": 0.0006,
+ "loss": 4.211801528930664,
+ "step": 4850
+ },
+ {
+ "epoch": 67.37729257641921,
+ "grad_norm": 0.014775963500142097,
+ "learning_rate": 0.0006,
+ "loss": 4.360859394073486,
+ "step": 4851
+ },
+ {
+ "epoch": 67.39126637554585,
+ "grad_norm": 0.015854304656386375,
+ "learning_rate": 0.0006,
+ "loss": 4.211830139160156,
+ "step": 4852
+ },
+ {
+ "epoch": 67.40524017467249,
+ "grad_norm": 0.016888894140720367,
+ "learning_rate": 0.0006,
+ "loss": 4.216549873352051,
+ "step": 4853
+ },
+ {
+ "epoch": 67.41921397379913,
+ "grad_norm": 0.0170894768089056,
+ "learning_rate": 0.0006,
+ "loss": 4.263486385345459,
+ "step": 4854
+ },
+ {
+ "epoch": 67.43318777292576,
+ "grad_norm": 0.015829868614673615,
+ "learning_rate": 0.0006,
+ "loss": 4.360196113586426,
+ "step": 4855
+ },
+ {
+ "epoch": 67.44716157205241,
+ "grad_norm": 0.015220202505588531,
+ "learning_rate": 0.0006,
+ "loss": 4.243794918060303,
+ "step": 4856
+ },
+ {
+ "epoch": 67.46113537117904,
+ "grad_norm": 0.01413388829678297,
+ "learning_rate": 0.0006,
+ "loss": 4.286624908447266,
+ "step": 4857
+ },
+ {
+ "epoch": 67.47510917030567,
+ "grad_norm": 0.01426626369357109,
+ "learning_rate": 0.0006,
+ "loss": 4.257316589355469,
+ "step": 4858
+ },
+ {
+ "epoch": 67.48908296943232,
+ "grad_norm": 0.015957845374941826,
+ "learning_rate": 0.0006,
+ "loss": 4.300443649291992,
+ "step": 4859
+ },
+ {
+ "epoch": 67.50305676855895,
+ "grad_norm": 0.015499051660299301,
+ "learning_rate": 0.0006,
+ "loss": 4.278162002563477,
+ "step": 4860
+ },
+ {
+ "epoch": 67.5170305676856,
+ "grad_norm": 0.015145624987781048,
+ "learning_rate": 0.0006,
+ "loss": 4.267950534820557,
+ "step": 4861
+ },
+ {
+ "epoch": 67.53100436681223,
+ "grad_norm": 0.014610803686082363,
+ "learning_rate": 0.0006,
+ "loss": 4.247499942779541,
+ "step": 4862
+ },
+ {
+ "epoch": 67.54497816593886,
+ "grad_norm": 0.014860909432172775,
+ "learning_rate": 0.0006,
+ "loss": 4.263200759887695,
+ "step": 4863
+ },
+ {
+ "epoch": 67.5589519650655,
+ "grad_norm": 0.01718984916806221,
+ "learning_rate": 0.0006,
+ "loss": 4.290019989013672,
+ "step": 4864
+ },
+ {
+ "epoch": 67.57292576419214,
+ "grad_norm": 0.01558963768184185,
+ "learning_rate": 0.0006,
+ "loss": 4.223816871643066,
+ "step": 4865
+ },
+ {
+ "epoch": 67.58689956331878,
+ "grad_norm": 0.014927188865840435,
+ "learning_rate": 0.0006,
+ "loss": 4.363586902618408,
+ "step": 4866
+ },
+ {
+ "epoch": 67.60087336244541,
+ "grad_norm": 0.01573595404624939,
+ "learning_rate": 0.0006,
+ "loss": 4.254765510559082,
+ "step": 4867
+ },
+ {
+ "epoch": 67.61484716157206,
+ "grad_norm": 0.015618929639458656,
+ "learning_rate": 0.0006,
+ "loss": 4.25072717666626,
+ "step": 4868
+ },
+ {
+ "epoch": 67.62882096069869,
+ "grad_norm": 0.01605590060353279,
+ "learning_rate": 0.0006,
+ "loss": 4.25208854675293,
+ "step": 4869
+ },
+ {
+ "epoch": 67.64279475982532,
+ "grad_norm": 0.01804601214826107,
+ "learning_rate": 0.0006,
+ "loss": 4.351354122161865,
+ "step": 4870
+ },
+ {
+ "epoch": 67.65676855895197,
+ "grad_norm": 0.020595047622919083,
+ "learning_rate": 0.0006,
+ "loss": 4.193732738494873,
+ "step": 4871
+ },
+ {
+ "epoch": 67.6707423580786,
+ "grad_norm": 0.023037966340780258,
+ "learning_rate": 0.0006,
+ "loss": 4.27754020690918,
+ "step": 4872
+ },
+ {
+ "epoch": 67.68471615720524,
+ "grad_norm": 0.020680895075201988,
+ "learning_rate": 0.0006,
+ "loss": 4.314730167388916,
+ "step": 4873
+ },
+ {
+ "epoch": 67.69868995633188,
+ "grad_norm": 0.01887434534728527,
+ "learning_rate": 0.0006,
+ "loss": 4.252305030822754,
+ "step": 4874
+ },
+ {
+ "epoch": 67.71266375545852,
+ "grad_norm": 0.016333596780896187,
+ "learning_rate": 0.0006,
+ "loss": 4.146123886108398,
+ "step": 4875
+ },
+ {
+ "epoch": 67.72663755458515,
+ "grad_norm": 0.015688810497522354,
+ "learning_rate": 0.0006,
+ "loss": 4.2814154624938965,
+ "step": 4876
+ },
+ {
+ "epoch": 67.74061135371178,
+ "grad_norm": 0.016270726919174194,
+ "learning_rate": 0.0006,
+ "loss": 4.323200225830078,
+ "step": 4877
+ },
+ {
+ "epoch": 67.75458515283843,
+ "grad_norm": 0.016824988648295403,
+ "learning_rate": 0.0006,
+ "loss": 4.269501686096191,
+ "step": 4878
+ },
+ {
+ "epoch": 67.76855895196506,
+ "grad_norm": 0.01730562001466751,
+ "learning_rate": 0.0006,
+ "loss": 4.187993049621582,
+ "step": 4879
+ },
+ {
+ "epoch": 67.78253275109171,
+ "grad_norm": 0.01546655036509037,
+ "learning_rate": 0.0006,
+ "loss": 4.340415000915527,
+ "step": 4880
+ },
+ {
+ "epoch": 67.79650655021834,
+ "grad_norm": 0.014173958450555801,
+ "learning_rate": 0.0006,
+ "loss": 4.217419624328613,
+ "step": 4881
+ },
+ {
+ "epoch": 67.81048034934499,
+ "grad_norm": 0.016810474917292595,
+ "learning_rate": 0.0006,
+ "loss": 4.352302074432373,
+ "step": 4882
+ },
+ {
+ "epoch": 67.82445414847162,
+ "grad_norm": 0.0175373163074255,
+ "learning_rate": 0.0006,
+ "loss": 4.20377254486084,
+ "step": 4883
+ },
+ {
+ "epoch": 67.83842794759825,
+ "grad_norm": 0.015734676271677017,
+ "learning_rate": 0.0006,
+ "loss": 4.147495746612549,
+ "step": 4884
+ },
+ {
+ "epoch": 67.8524017467249,
+ "grad_norm": 0.015091817826032639,
+ "learning_rate": 0.0006,
+ "loss": 4.272252082824707,
+ "step": 4885
+ },
+ {
+ "epoch": 67.86637554585153,
+ "grad_norm": 0.015000219456851482,
+ "learning_rate": 0.0006,
+ "loss": 4.302773952484131,
+ "step": 4886
+ },
+ {
+ "epoch": 67.88034934497817,
+ "grad_norm": 0.013827184215188026,
+ "learning_rate": 0.0006,
+ "loss": 4.337242126464844,
+ "step": 4887
+ },
+ {
+ "epoch": 67.8943231441048,
+ "grad_norm": 0.015308346599340439,
+ "learning_rate": 0.0006,
+ "loss": 4.140895366668701,
+ "step": 4888
+ },
+ {
+ "epoch": 67.90829694323143,
+ "grad_norm": 0.015356178395450115,
+ "learning_rate": 0.0006,
+ "loss": 4.312541484832764,
+ "step": 4889
+ },
+ {
+ "epoch": 67.92227074235808,
+ "grad_norm": 0.0137353939935565,
+ "learning_rate": 0.0006,
+ "loss": 4.173184394836426,
+ "step": 4890
+ },
+ {
+ "epoch": 67.93624454148471,
+ "grad_norm": 0.01332368515431881,
+ "learning_rate": 0.0006,
+ "loss": 4.191199779510498,
+ "step": 4891
+ },
+ {
+ "epoch": 67.95021834061136,
+ "grad_norm": 0.01488130260258913,
+ "learning_rate": 0.0006,
+ "loss": 4.331752300262451,
+ "step": 4892
+ },
+ {
+ "epoch": 67.96419213973799,
+ "grad_norm": 0.015449733473360538,
+ "learning_rate": 0.0006,
+ "loss": 4.2385454177856445,
+ "step": 4893
+ },
+ {
+ "epoch": 67.97816593886463,
+ "grad_norm": 0.01556103490293026,
+ "learning_rate": 0.0006,
+ "loss": 4.262353897094727,
+ "step": 4894
+ },
+ {
+ "epoch": 67.99213973799127,
+ "grad_norm": 0.014613849110901356,
+ "learning_rate": 0.0006,
+ "loss": 4.223945617675781,
+ "step": 4895
+ },
+ {
+ "epoch": 68.0,
+ "grad_norm": 0.015323462896049023,
+ "learning_rate": 0.0006,
+ "loss": 4.317168712615967,
+ "step": 4896
+ },
+ {
+ "epoch": 68.0,
+ "eval_loss": 4.671603679656982,
+ "eval_runtime": 56.7317,
+ "eval_samples_per_second": 43.045,
+ "eval_steps_per_second": 1.357,
+ "step": 4896
+ },
+ {
+ "epoch": 68.01397379912663,
+ "grad_norm": 0.01910698413848877,
+ "learning_rate": 0.0006,
+ "loss": 4.206167221069336,
+ "step": 4897
+ },
+ {
+ "epoch": 68.02794759825328,
+ "grad_norm": 0.025112485513091087,
+ "learning_rate": 0.0006,
+ "loss": 4.384706974029541,
+ "step": 4898
+ },
+ {
+ "epoch": 68.04192139737991,
+ "grad_norm": 0.02618720941245556,
+ "learning_rate": 0.0006,
+ "loss": 4.321046829223633,
+ "step": 4899
+ },
+ {
+ "epoch": 68.05589519650655,
+ "grad_norm": 0.020853538066148758,
+ "learning_rate": 0.0006,
+ "loss": 4.217638969421387,
+ "step": 4900
+ },
+ {
+ "epoch": 68.06986899563319,
+ "grad_norm": 0.020025717094540596,
+ "learning_rate": 0.0006,
+ "loss": 4.306824684143066,
+ "step": 4901
+ },
+ {
+ "epoch": 68.08384279475983,
+ "grad_norm": 0.023868650197982788,
+ "learning_rate": 0.0006,
+ "loss": 4.327455997467041,
+ "step": 4902
+ },
+ {
+ "epoch": 68.09781659388646,
+ "grad_norm": 0.022279081866145134,
+ "learning_rate": 0.0006,
+ "loss": 4.400699615478516,
+ "step": 4903
+ },
+ {
+ "epoch": 68.1117903930131,
+ "grad_norm": 0.018965687602758408,
+ "learning_rate": 0.0006,
+ "loss": 4.19097900390625,
+ "step": 4904
+ },
+ {
+ "epoch": 68.12576419213974,
+ "grad_norm": 0.01936378888785839,
+ "learning_rate": 0.0006,
+ "loss": 4.206711769104004,
+ "step": 4905
+ },
+ {
+ "epoch": 68.13973799126637,
+ "grad_norm": 0.01811872608959675,
+ "learning_rate": 0.0006,
+ "loss": 4.204622268676758,
+ "step": 4906
+ },
+ {
+ "epoch": 68.15371179039302,
+ "grad_norm": 0.018530402332544327,
+ "learning_rate": 0.0006,
+ "loss": 4.289602279663086,
+ "step": 4907
+ },
+ {
+ "epoch": 68.16768558951965,
+ "grad_norm": 0.01755962148308754,
+ "learning_rate": 0.0006,
+ "loss": 4.317145347595215,
+ "step": 4908
+ },
+ {
+ "epoch": 68.18165938864628,
+ "grad_norm": 0.020413702353835106,
+ "learning_rate": 0.0006,
+ "loss": 4.2882843017578125,
+ "step": 4909
+ },
+ {
+ "epoch": 68.19563318777293,
+ "grad_norm": 0.020064299926161766,
+ "learning_rate": 0.0006,
+ "loss": 4.248260021209717,
+ "step": 4910
+ },
+ {
+ "epoch": 68.20960698689956,
+ "grad_norm": 0.017415888607501984,
+ "learning_rate": 0.0006,
+ "loss": 4.370295524597168,
+ "step": 4911
+ },
+ {
+ "epoch": 68.2235807860262,
+ "grad_norm": 0.015005858615040779,
+ "learning_rate": 0.0006,
+ "loss": 4.319567680358887,
+ "step": 4912
+ },
+ {
+ "epoch": 68.23755458515284,
+ "grad_norm": 0.016349777579307556,
+ "learning_rate": 0.0006,
+ "loss": 4.301776885986328,
+ "step": 4913
+ },
+ {
+ "epoch": 68.25152838427948,
+ "grad_norm": 0.015779560431838036,
+ "learning_rate": 0.0006,
+ "loss": 4.271171569824219,
+ "step": 4914
+ },
+ {
+ "epoch": 68.26550218340611,
+ "grad_norm": 0.013237644918262959,
+ "learning_rate": 0.0006,
+ "loss": 4.324908256530762,
+ "step": 4915
+ },
+ {
+ "epoch": 68.27947598253274,
+ "grad_norm": 0.014076225459575653,
+ "learning_rate": 0.0006,
+ "loss": 4.177175998687744,
+ "step": 4916
+ },
+ {
+ "epoch": 68.29344978165939,
+ "grad_norm": 0.013906744308769703,
+ "learning_rate": 0.0006,
+ "loss": 4.11588191986084,
+ "step": 4917
+ },
+ {
+ "epoch": 68.30742358078602,
+ "grad_norm": 0.013939064927399158,
+ "learning_rate": 0.0006,
+ "loss": 4.183963775634766,
+ "step": 4918
+ },
+ {
+ "epoch": 68.32139737991267,
+ "grad_norm": 0.014572170563042164,
+ "learning_rate": 0.0006,
+ "loss": 4.19586181640625,
+ "step": 4919
+ },
+ {
+ "epoch": 68.3353711790393,
+ "grad_norm": 0.014085350558161736,
+ "learning_rate": 0.0006,
+ "loss": 4.170892238616943,
+ "step": 4920
+ },
+ {
+ "epoch": 68.34934497816595,
+ "grad_norm": 0.014357863925397396,
+ "learning_rate": 0.0006,
+ "loss": 4.321155548095703,
+ "step": 4921
+ },
+ {
+ "epoch": 68.36331877729258,
+ "grad_norm": 0.01461675763130188,
+ "learning_rate": 0.0006,
+ "loss": 4.293334007263184,
+ "step": 4922
+ },
+ {
+ "epoch": 68.37729257641921,
+ "grad_norm": 0.014828486368060112,
+ "learning_rate": 0.0006,
+ "loss": 4.1439714431762695,
+ "step": 4923
+ },
+ {
+ "epoch": 68.39126637554585,
+ "grad_norm": 0.013730363920331001,
+ "learning_rate": 0.0006,
+ "loss": 4.287973880767822,
+ "step": 4924
+ },
+ {
+ "epoch": 68.40524017467249,
+ "grad_norm": 0.013852203264832497,
+ "learning_rate": 0.0006,
+ "loss": 4.248824119567871,
+ "step": 4925
+ },
+ {
+ "epoch": 68.41921397379913,
+ "grad_norm": 0.014005862176418304,
+ "learning_rate": 0.0006,
+ "loss": 4.293768882751465,
+ "step": 4926
+ },
+ {
+ "epoch": 68.43318777292576,
+ "grad_norm": 0.014435851015150547,
+ "learning_rate": 0.0006,
+ "loss": 4.204813003540039,
+ "step": 4927
+ },
+ {
+ "epoch": 68.44716157205241,
+ "grad_norm": 0.015305934473872185,
+ "learning_rate": 0.0006,
+ "loss": 4.329364776611328,
+ "step": 4928
+ },
+ {
+ "epoch": 68.46113537117904,
+ "grad_norm": 0.014356814324855804,
+ "learning_rate": 0.0006,
+ "loss": 4.264206409454346,
+ "step": 4929
+ },
+ {
+ "epoch": 68.47510917030567,
+ "grad_norm": 0.01432347297668457,
+ "learning_rate": 0.0006,
+ "loss": 4.29440450668335,
+ "step": 4930
+ },
+ {
+ "epoch": 68.48908296943232,
+ "grad_norm": 0.012672432698309422,
+ "learning_rate": 0.0006,
+ "loss": 4.258024215698242,
+ "step": 4931
+ },
+ {
+ "epoch": 68.50305676855895,
+ "grad_norm": 0.014209977351129055,
+ "learning_rate": 0.0006,
+ "loss": 4.216668128967285,
+ "step": 4932
+ },
+ {
+ "epoch": 68.5170305676856,
+ "grad_norm": 0.014606663957238197,
+ "learning_rate": 0.0006,
+ "loss": 4.294184684753418,
+ "step": 4933
+ },
+ {
+ "epoch": 68.53100436681223,
+ "grad_norm": 0.014986990951001644,
+ "learning_rate": 0.0006,
+ "loss": 4.382591247558594,
+ "step": 4934
+ },
+ {
+ "epoch": 68.54497816593886,
+ "grad_norm": 0.013998783193528652,
+ "learning_rate": 0.0006,
+ "loss": 4.195143222808838,
+ "step": 4935
+ },
+ {
+ "epoch": 68.5589519650655,
+ "grad_norm": 0.015011102892458439,
+ "learning_rate": 0.0006,
+ "loss": 4.313785076141357,
+ "step": 4936
+ },
+ {
+ "epoch": 68.57292576419214,
+ "grad_norm": 0.01641209051012993,
+ "learning_rate": 0.0006,
+ "loss": 4.322803020477295,
+ "step": 4937
+ },
+ {
+ "epoch": 68.58689956331878,
+ "grad_norm": 0.016207464039325714,
+ "learning_rate": 0.0006,
+ "loss": 4.311431884765625,
+ "step": 4938
+ },
+ {
+ "epoch": 68.60087336244541,
+ "grad_norm": 0.01747811958193779,
+ "learning_rate": 0.0006,
+ "loss": 4.310845375061035,
+ "step": 4939
+ },
+ {
+ "epoch": 68.61484716157206,
+ "grad_norm": 0.019475717097520828,
+ "learning_rate": 0.0006,
+ "loss": 4.338624954223633,
+ "step": 4940
+ },
+ {
+ "epoch": 68.62882096069869,
+ "grad_norm": 0.018641719594597816,
+ "learning_rate": 0.0006,
+ "loss": 4.34913969039917,
+ "step": 4941
+ },
+ {
+ "epoch": 68.64279475982532,
+ "grad_norm": 0.018376432359218597,
+ "learning_rate": 0.0006,
+ "loss": 4.28464412689209,
+ "step": 4942
+ },
+ {
+ "epoch": 68.65676855895197,
+ "grad_norm": 0.017095010727643967,
+ "learning_rate": 0.0006,
+ "loss": 4.217957496643066,
+ "step": 4943
+ },
+ {
+ "epoch": 68.6707423580786,
+ "grad_norm": 0.016321983188390732,
+ "learning_rate": 0.0006,
+ "loss": 4.220042705535889,
+ "step": 4944
+ },
+ {
+ "epoch": 68.68471615720524,
+ "grad_norm": 0.017128000035881996,
+ "learning_rate": 0.0006,
+ "loss": 4.293500900268555,
+ "step": 4945
+ },
+ {
+ "epoch": 68.69868995633188,
+ "grad_norm": 0.015592411160469055,
+ "learning_rate": 0.0006,
+ "loss": 4.195798873901367,
+ "step": 4946
+ },
+ {
+ "epoch": 68.71266375545852,
+ "grad_norm": 0.016607079654932022,
+ "learning_rate": 0.0006,
+ "loss": 4.2654337882995605,
+ "step": 4947
+ },
+ {
+ "epoch": 68.72663755458515,
+ "grad_norm": 0.01545762736350298,
+ "learning_rate": 0.0006,
+ "loss": 4.303828239440918,
+ "step": 4948
+ },
+ {
+ "epoch": 68.74061135371178,
+ "grad_norm": 0.014328422956168652,
+ "learning_rate": 0.0006,
+ "loss": 4.29379940032959,
+ "step": 4949
+ },
+ {
+ "epoch": 68.75458515283843,
+ "grad_norm": 0.016516804695129395,
+ "learning_rate": 0.0006,
+ "loss": 4.2317214012146,
+ "step": 4950
+ },
+ {
+ "epoch": 68.76855895196506,
+ "grad_norm": 0.016324171796441078,
+ "learning_rate": 0.0006,
+ "loss": 4.278043746948242,
+ "step": 4951
+ },
+ {
+ "epoch": 68.78253275109171,
+ "grad_norm": 0.015511159785091877,
+ "learning_rate": 0.0006,
+ "loss": 4.188244342803955,
+ "step": 4952
+ },
+ {
+ "epoch": 68.79650655021834,
+ "grad_norm": 0.01676103286445141,
+ "learning_rate": 0.0006,
+ "loss": 4.2169976234436035,
+ "step": 4953
+ },
+ {
+ "epoch": 68.81048034934499,
+ "grad_norm": 0.015428312122821808,
+ "learning_rate": 0.0006,
+ "loss": 4.4004716873168945,
+ "step": 4954
+ },
+ {
+ "epoch": 68.82445414847162,
+ "grad_norm": 0.0162410456687212,
+ "learning_rate": 0.0006,
+ "loss": 4.282752513885498,
+ "step": 4955
+ },
+ {
+ "epoch": 68.83842794759825,
+ "grad_norm": 0.017578862607479095,
+ "learning_rate": 0.0006,
+ "loss": 4.098004341125488,
+ "step": 4956
+ },
+ {
+ "epoch": 68.8524017467249,
+ "grad_norm": 0.01805749163031578,
+ "learning_rate": 0.0006,
+ "loss": 4.277988433837891,
+ "step": 4957
+ },
+ {
+ "epoch": 68.86637554585153,
+ "grad_norm": 0.016728045418858528,
+ "learning_rate": 0.0006,
+ "loss": 4.29766321182251,
+ "step": 4958
+ },
+ {
+ "epoch": 68.88034934497817,
+ "grad_norm": 0.014899657107889652,
+ "learning_rate": 0.0006,
+ "loss": 4.329736709594727,
+ "step": 4959
+ },
+ {
+ "epoch": 68.8943231441048,
+ "grad_norm": 0.013441305607557297,
+ "learning_rate": 0.0006,
+ "loss": 4.2873215675354,
+ "step": 4960
+ },
+ {
+ "epoch": 68.90829694323143,
+ "grad_norm": 0.01554128248244524,
+ "learning_rate": 0.0006,
+ "loss": 4.278327941894531,
+ "step": 4961
+ },
+ {
+ "epoch": 68.92227074235808,
+ "grad_norm": 0.01806696318089962,
+ "learning_rate": 0.0006,
+ "loss": 4.38275146484375,
+ "step": 4962
+ },
+ {
+ "epoch": 68.93624454148471,
+ "grad_norm": 0.017025409266352654,
+ "learning_rate": 0.0006,
+ "loss": 4.210447311401367,
+ "step": 4963
+ },
+ {
+ "epoch": 68.95021834061136,
+ "grad_norm": 0.01597420684993267,
+ "learning_rate": 0.0006,
+ "loss": 4.309865951538086,
+ "step": 4964
+ },
+ {
+ "epoch": 68.96419213973799,
+ "grad_norm": 0.015835624188184738,
+ "learning_rate": 0.0006,
+ "loss": 4.225904941558838,
+ "step": 4965
+ },
+ {
+ "epoch": 68.97816593886463,
+ "grad_norm": 0.018170492723584175,
+ "learning_rate": 0.0006,
+ "loss": 4.248478889465332,
+ "step": 4966
+ },
+ {
+ "epoch": 68.99213973799127,
+ "grad_norm": 0.017372991889715195,
+ "learning_rate": 0.0006,
+ "loss": 4.283486843109131,
+ "step": 4967
+ },
+ {
+ "epoch": 69.0,
+ "grad_norm": 0.016305390745401382,
+ "learning_rate": 0.0006,
+ "loss": 4.273636817932129,
+ "step": 4968
+ },
+ {
+ "epoch": 69.0,
+ "eval_loss": 4.66873836517334,
+ "eval_runtime": 57.0791,
+ "eval_samples_per_second": 42.783,
+ "eval_steps_per_second": 1.349,
+ "step": 4968
+ },
+ {
+ "epoch": 69.01397379912663,
+ "grad_norm": 0.016567355021834373,
+ "learning_rate": 0.0006,
+ "loss": 4.269988059997559,
+ "step": 4969
+ },
+ {
+ "epoch": 69.02794759825328,
+ "grad_norm": 0.017476852983236313,
+ "learning_rate": 0.0006,
+ "loss": 4.2518720626831055,
+ "step": 4970
+ },
+ {
+ "epoch": 69.04192139737991,
+ "grad_norm": 0.017070338129997253,
+ "learning_rate": 0.0006,
+ "loss": 4.238024711608887,
+ "step": 4971
+ },
+ {
+ "epoch": 69.05589519650655,
+ "grad_norm": 0.01639373227953911,
+ "learning_rate": 0.0006,
+ "loss": 4.188558101654053,
+ "step": 4972
+ },
+ {
+ "epoch": 69.06986899563319,
+ "grad_norm": 0.01856113411486149,
+ "learning_rate": 0.0006,
+ "loss": 4.339556694030762,
+ "step": 4973
+ },
+ {
+ "epoch": 69.08384279475983,
+ "grad_norm": 0.01762578636407852,
+ "learning_rate": 0.0006,
+ "loss": 4.141518592834473,
+ "step": 4974
+ },
+ {
+ "epoch": 69.09781659388646,
+ "grad_norm": 0.015437949448823929,
+ "learning_rate": 0.0006,
+ "loss": 4.225461959838867,
+ "step": 4975
+ },
+ {
+ "epoch": 69.1117903930131,
+ "grad_norm": 0.015293503180146217,
+ "learning_rate": 0.0006,
+ "loss": 4.290884971618652,
+ "step": 4976
+ },
+ {
+ "epoch": 69.12576419213974,
+ "grad_norm": 0.016259880736470222,
+ "learning_rate": 0.0006,
+ "loss": 4.121919631958008,
+ "step": 4977
+ },
+ {
+ "epoch": 69.13973799126637,
+ "grad_norm": 0.017225584015250206,
+ "learning_rate": 0.0006,
+ "loss": 4.136716842651367,
+ "step": 4978
+ },
+ {
+ "epoch": 69.15371179039302,
+ "grad_norm": 0.01699289306998253,
+ "learning_rate": 0.0006,
+ "loss": 4.193601608276367,
+ "step": 4979
+ },
+ {
+ "epoch": 69.16768558951965,
+ "grad_norm": 0.015613161027431488,
+ "learning_rate": 0.0006,
+ "loss": 4.288212299346924,
+ "step": 4980
+ },
+ {
+ "epoch": 69.18165938864628,
+ "grad_norm": 0.01643933728337288,
+ "learning_rate": 0.0006,
+ "loss": 4.324707984924316,
+ "step": 4981
+ },
+ {
+ "epoch": 69.19563318777293,
+ "grad_norm": 0.01648905873298645,
+ "learning_rate": 0.0006,
+ "loss": 4.287806510925293,
+ "step": 4982
+ },
+ {
+ "epoch": 69.20960698689956,
+ "grad_norm": 0.015444950200617313,
+ "learning_rate": 0.0006,
+ "loss": 4.200622081756592,
+ "step": 4983
+ },
+ {
+ "epoch": 69.2235807860262,
+ "grad_norm": 0.016034258529543877,
+ "learning_rate": 0.0006,
+ "loss": 4.2193827629089355,
+ "step": 4984
+ },
+ {
+ "epoch": 69.23755458515284,
+ "grad_norm": 0.017984388396143913,
+ "learning_rate": 0.0006,
+ "loss": 4.18237829208374,
+ "step": 4985
+ },
+ {
+ "epoch": 69.25152838427948,
+ "grad_norm": 0.0195505041629076,
+ "learning_rate": 0.0006,
+ "loss": 4.216727256774902,
+ "step": 4986
+ },
+ {
+ "epoch": 69.26550218340611,
+ "grad_norm": 0.017779147252440453,
+ "learning_rate": 0.0006,
+ "loss": 4.278628826141357,
+ "step": 4987
+ },
+ {
+ "epoch": 69.27947598253274,
+ "grad_norm": 0.017119476571679115,
+ "learning_rate": 0.0006,
+ "loss": 4.2409138679504395,
+ "step": 4988
+ },
+ {
+ "epoch": 69.29344978165939,
+ "grad_norm": 0.016840364784002304,
+ "learning_rate": 0.0006,
+ "loss": 4.207575798034668,
+ "step": 4989
+ },
+ {
+ "epoch": 69.30742358078602,
+ "grad_norm": 0.014593573287129402,
+ "learning_rate": 0.0006,
+ "loss": 4.164905548095703,
+ "step": 4990
+ },
+ {
+ "epoch": 69.32139737991267,
+ "grad_norm": 0.013922630809247494,
+ "learning_rate": 0.0006,
+ "loss": 4.2696990966796875,
+ "step": 4991
+ },
+ {
+ "epoch": 69.3353711790393,
+ "grad_norm": 0.01577564887702465,
+ "learning_rate": 0.0006,
+ "loss": 4.201589107513428,
+ "step": 4992
+ },
+ {
+ "epoch": 69.34934497816595,
+ "grad_norm": 0.016392096877098083,
+ "learning_rate": 0.0006,
+ "loss": 4.248831748962402,
+ "step": 4993
+ },
+ {
+ "epoch": 69.36331877729258,
+ "grad_norm": 0.015961311757564545,
+ "learning_rate": 0.0006,
+ "loss": 4.181260108947754,
+ "step": 4994
+ },
+ {
+ "epoch": 69.37729257641921,
+ "grad_norm": 0.015350551344454288,
+ "learning_rate": 0.0006,
+ "loss": 4.230352878570557,
+ "step": 4995
+ },
+ {
+ "epoch": 69.39126637554585,
+ "grad_norm": 0.015870466828346252,
+ "learning_rate": 0.0006,
+ "loss": 4.298305511474609,
+ "step": 4996
+ },
+ {
+ "epoch": 69.40524017467249,
+ "grad_norm": 0.013716642744839191,
+ "learning_rate": 0.0006,
+ "loss": 4.220139503479004,
+ "step": 4997
+ },
+ {
+ "epoch": 69.41921397379913,
+ "grad_norm": 0.014431777410209179,
+ "learning_rate": 0.0006,
+ "loss": 4.288130760192871,
+ "step": 4998
+ },
+ {
+ "epoch": 69.43318777292576,
+ "grad_norm": 0.01426647324115038,
+ "learning_rate": 0.0006,
+ "loss": 4.295767784118652,
+ "step": 4999
+ },
+ {
+ "epoch": 69.44716157205241,
+ "grad_norm": 0.014286615885794163,
+ "learning_rate": 0.0006,
+ "loss": 4.38106632232666,
+ "step": 5000
+ },
+ {
+ "epoch": 69.46113537117904,
+ "grad_norm": 0.017990585416555405,
+ "learning_rate": 0.0006,
+ "loss": 4.1593017578125,
+ "step": 5001
+ },
+ {
+ "epoch": 69.47510917030567,
+ "grad_norm": 0.018846049904823303,
+ "learning_rate": 0.0006,
+ "loss": 4.216091632843018,
+ "step": 5002
+ },
+ {
+ "epoch": 69.48908296943232,
+ "grad_norm": 0.018179181963205338,
+ "learning_rate": 0.0006,
+ "loss": 4.240332126617432,
+ "step": 5003
+ },
+ {
+ "epoch": 69.50305676855895,
+ "grad_norm": 0.016693448647856712,
+ "learning_rate": 0.0006,
+ "loss": 4.256624221801758,
+ "step": 5004
+ },
+ {
+ "epoch": 69.5170305676856,
+ "grad_norm": 0.01718989387154579,
+ "learning_rate": 0.0006,
+ "loss": 4.300940990447998,
+ "step": 5005
+ },
+ {
+ "epoch": 69.53100436681223,
+ "grad_norm": 0.020236829295754433,
+ "learning_rate": 0.0006,
+ "loss": 4.289146423339844,
+ "step": 5006
+ },
+ {
+ "epoch": 69.54497816593886,
+ "grad_norm": 0.024860285222530365,
+ "learning_rate": 0.0006,
+ "loss": 4.216216087341309,
+ "step": 5007
+ },
+ {
+ "epoch": 69.5589519650655,
+ "grad_norm": 0.02559969387948513,
+ "learning_rate": 0.0006,
+ "loss": 4.262890815734863,
+ "step": 5008
+ },
+ {
+ "epoch": 69.57292576419214,
+ "grad_norm": 0.019180957227945328,
+ "learning_rate": 0.0006,
+ "loss": 4.230309009552002,
+ "step": 5009
+ },
+ {
+ "epoch": 69.58689956331878,
+ "grad_norm": 0.020442591980099678,
+ "learning_rate": 0.0006,
+ "loss": 4.291759490966797,
+ "step": 5010
+ },
+ {
+ "epoch": 69.60087336244541,
+ "grad_norm": 0.0212619137018919,
+ "learning_rate": 0.0006,
+ "loss": 4.3529839515686035,
+ "step": 5011
+ },
+ {
+ "epoch": 69.61484716157206,
+ "grad_norm": 0.02067175693809986,
+ "learning_rate": 0.0006,
+ "loss": 4.33197021484375,
+ "step": 5012
+ },
+ {
+ "epoch": 69.62882096069869,
+ "grad_norm": 0.01802872307598591,
+ "learning_rate": 0.0006,
+ "loss": 4.2101969718933105,
+ "step": 5013
+ },
+ {
+ "epoch": 69.64279475982532,
+ "grad_norm": 0.01787702925503254,
+ "learning_rate": 0.0006,
+ "loss": 4.270336627960205,
+ "step": 5014
+ },
+ {
+ "epoch": 69.65676855895197,
+ "grad_norm": 0.01586683839559555,
+ "learning_rate": 0.0006,
+ "loss": 4.222309589385986,
+ "step": 5015
+ },
+ {
+ "epoch": 69.6707423580786,
+ "grad_norm": 0.01717999391257763,
+ "learning_rate": 0.0006,
+ "loss": 4.253704071044922,
+ "step": 5016
+ },
+ {
+ "epoch": 69.68471615720524,
+ "grad_norm": 0.01594279147684574,
+ "learning_rate": 0.0006,
+ "loss": 4.344658851623535,
+ "step": 5017
+ },
+ {
+ "epoch": 69.69868995633188,
+ "grad_norm": 0.014419942162930965,
+ "learning_rate": 0.0006,
+ "loss": 4.2144880294799805,
+ "step": 5018
+ },
+ {
+ "epoch": 69.71266375545852,
+ "grad_norm": 0.015824224799871445,
+ "learning_rate": 0.0006,
+ "loss": 4.226252555847168,
+ "step": 5019
+ },
+ {
+ "epoch": 69.72663755458515,
+ "grad_norm": 0.013329979032278061,
+ "learning_rate": 0.0006,
+ "loss": 4.238587379455566,
+ "step": 5020
+ },
+ {
+ "epoch": 69.74061135371178,
+ "grad_norm": 0.013745979405939579,
+ "learning_rate": 0.0006,
+ "loss": 4.317209243774414,
+ "step": 5021
+ },
+ {
+ "epoch": 69.75458515283843,
+ "grad_norm": 0.014064167626202106,
+ "learning_rate": 0.0006,
+ "loss": 4.316982746124268,
+ "step": 5022
+ },
+ {
+ "epoch": 69.76855895196506,
+ "grad_norm": 0.014611301012337208,
+ "learning_rate": 0.0006,
+ "loss": 4.212728500366211,
+ "step": 5023
+ },
+ {
+ "epoch": 69.78253275109171,
+ "grad_norm": 0.014947416260838509,
+ "learning_rate": 0.0006,
+ "loss": 4.319550514221191,
+ "step": 5024
+ },
+ {
+ "epoch": 69.79650655021834,
+ "grad_norm": 0.014780526980757713,
+ "learning_rate": 0.0006,
+ "loss": 4.1888933181762695,
+ "step": 5025
+ },
+ {
+ "epoch": 69.81048034934499,
+ "grad_norm": 0.01611459068953991,
+ "learning_rate": 0.0006,
+ "loss": 4.239692211151123,
+ "step": 5026
+ },
+ {
+ "epoch": 69.82445414847162,
+ "grad_norm": 0.015395848080515862,
+ "learning_rate": 0.0006,
+ "loss": 4.308794021606445,
+ "step": 5027
+ },
+ {
+ "epoch": 69.83842794759825,
+ "grad_norm": 0.016113396733999252,
+ "learning_rate": 0.0006,
+ "loss": 4.328035354614258,
+ "step": 5028
+ },
+ {
+ "epoch": 69.8524017467249,
+ "grad_norm": 0.016386888921260834,
+ "learning_rate": 0.0006,
+ "loss": 4.272579193115234,
+ "step": 5029
+ },
+ {
+ "epoch": 69.86637554585153,
+ "grad_norm": 0.013675205409526825,
+ "learning_rate": 0.0006,
+ "loss": 4.226552963256836,
+ "step": 5030
+ },
+ {
+ "epoch": 69.88034934497817,
+ "grad_norm": 0.01472668256610632,
+ "learning_rate": 0.0006,
+ "loss": 4.224167346954346,
+ "step": 5031
+ },
+ {
+ "epoch": 69.8943231441048,
+ "grad_norm": 0.014958228915929794,
+ "learning_rate": 0.0006,
+ "loss": 4.232619285583496,
+ "step": 5032
+ },
+ {
+ "epoch": 69.90829694323143,
+ "grad_norm": 0.012926832772791386,
+ "learning_rate": 0.0006,
+ "loss": 4.27107048034668,
+ "step": 5033
+ },
+ {
+ "epoch": 69.92227074235808,
+ "grad_norm": 0.013138801790773869,
+ "learning_rate": 0.0006,
+ "loss": 4.288477420806885,
+ "step": 5034
+ },
+ {
+ "epoch": 69.93624454148471,
+ "grad_norm": 0.013415461406111717,
+ "learning_rate": 0.0006,
+ "loss": 4.2033257484436035,
+ "step": 5035
+ },
+ {
+ "epoch": 69.95021834061136,
+ "grad_norm": 0.013807233422994614,
+ "learning_rate": 0.0006,
+ "loss": 4.274764537811279,
+ "step": 5036
+ },
+ {
+ "epoch": 69.96419213973799,
+ "grad_norm": 0.013929193839430809,
+ "learning_rate": 0.0006,
+ "loss": 4.1854352951049805,
+ "step": 5037
+ },
+ {
+ "epoch": 69.97816593886463,
+ "grad_norm": 0.015497001819312572,
+ "learning_rate": 0.0006,
+ "loss": 4.226498126983643,
+ "step": 5038
+ },
+ {
+ "epoch": 69.99213973799127,
+ "grad_norm": 0.01874048449099064,
+ "learning_rate": 0.0006,
+ "loss": 4.304137706756592,
+ "step": 5039
+ },
+ {
+ "epoch": 70.0,
+ "grad_norm": 0.021745791658759117,
+ "learning_rate": 0.0006,
+ "loss": 4.206809997558594,
+ "step": 5040
+ },
+ {
+ "epoch": 70.0,
+ "eval_loss": 4.675539970397949,
+ "eval_runtime": 56.9034,
+ "eval_samples_per_second": 42.915,
+ "eval_steps_per_second": 1.353,
+ "step": 5040
+ },
+ {
+ "epoch": 70.01397379912663,
+ "grad_norm": 0.02124839462339878,
+ "learning_rate": 0.0006,
+ "loss": 4.198199272155762,
+ "step": 5041
+ },
+ {
+ "epoch": 70.02794759825328,
+ "grad_norm": 0.019432688131928444,
+ "learning_rate": 0.0006,
+ "loss": 4.187261581420898,
+ "step": 5042
+ },
+ {
+ "epoch": 70.04192139737991,
+ "grad_norm": 0.01769290305674076,
+ "learning_rate": 0.0006,
+ "loss": 4.155284881591797,
+ "step": 5043
+ },
+ {
+ "epoch": 70.05589519650655,
+ "grad_norm": 0.01751520484685898,
+ "learning_rate": 0.0006,
+ "loss": 4.386857986450195,
+ "step": 5044
+ },
+ {
+ "epoch": 70.06986899563319,
+ "grad_norm": 0.018138699233531952,
+ "learning_rate": 0.0006,
+ "loss": 4.117342472076416,
+ "step": 5045
+ },
+ {
+ "epoch": 70.08384279475983,
+ "grad_norm": 0.016331054270267487,
+ "learning_rate": 0.0006,
+ "loss": 4.225295066833496,
+ "step": 5046
+ },
+ {
+ "epoch": 70.09781659388646,
+ "grad_norm": 0.01658729277551174,
+ "learning_rate": 0.0006,
+ "loss": 4.143835067749023,
+ "step": 5047
+ },
+ {
+ "epoch": 70.1117903930131,
+ "grad_norm": 0.016367698088288307,
+ "learning_rate": 0.0006,
+ "loss": 4.1469268798828125,
+ "step": 5048
+ },
+ {
+ "epoch": 70.12576419213974,
+ "grad_norm": 0.015151629224419594,
+ "learning_rate": 0.0006,
+ "loss": 4.117629051208496,
+ "step": 5049
+ },
+ {
+ "epoch": 70.13973799126637,
+ "grad_norm": 0.016434503719210625,
+ "learning_rate": 0.0006,
+ "loss": 4.107848644256592,
+ "step": 5050
+ },
+ {
+ "epoch": 70.15371179039302,
+ "grad_norm": 0.017837485298514366,
+ "learning_rate": 0.0006,
+ "loss": 4.10837459564209,
+ "step": 5051
+ },
+ {
+ "epoch": 70.16768558951965,
+ "grad_norm": 0.0165961142629385,
+ "learning_rate": 0.0006,
+ "loss": 4.169414043426514,
+ "step": 5052
+ },
+ {
+ "epoch": 70.18165938864628,
+ "grad_norm": 0.0157014112919569,
+ "learning_rate": 0.0006,
+ "loss": 4.0629706382751465,
+ "step": 5053
+ },
+ {
+ "epoch": 70.19563318777293,
+ "grad_norm": 0.01581883803009987,
+ "learning_rate": 0.0006,
+ "loss": 4.220012664794922,
+ "step": 5054
+ },
+ {
+ "epoch": 70.20960698689956,
+ "grad_norm": 0.015285917557775974,
+ "learning_rate": 0.0006,
+ "loss": 4.275337219238281,
+ "step": 5055
+ },
+ {
+ "epoch": 70.2235807860262,
+ "grad_norm": 0.014984549954533577,
+ "learning_rate": 0.0006,
+ "loss": 4.284764289855957,
+ "step": 5056
+ },
+ {
+ "epoch": 70.23755458515284,
+ "grad_norm": 0.015172012150287628,
+ "learning_rate": 0.0006,
+ "loss": 4.234498023986816,
+ "step": 5057
+ },
+ {
+ "epoch": 70.25152838427948,
+ "grad_norm": 0.016846632584929466,
+ "learning_rate": 0.0006,
+ "loss": 4.181689262390137,
+ "step": 5058
+ },
+ {
+ "epoch": 70.26550218340611,
+ "grad_norm": 0.01920711249113083,
+ "learning_rate": 0.0006,
+ "loss": 4.235280513763428,
+ "step": 5059
+ },
+ {
+ "epoch": 70.27947598253274,
+ "grad_norm": 0.017765291035175323,
+ "learning_rate": 0.0006,
+ "loss": 4.238955497741699,
+ "step": 5060
+ },
+ {
+ "epoch": 70.29344978165939,
+ "grad_norm": 0.015436639077961445,
+ "learning_rate": 0.0006,
+ "loss": 4.232516288757324,
+ "step": 5061
+ },
+ {
+ "epoch": 70.30742358078602,
+ "grad_norm": 0.015660086646676064,
+ "learning_rate": 0.0006,
+ "loss": 4.260349273681641,
+ "step": 5062
+ },
+ {
+ "epoch": 70.32139737991267,
+ "grad_norm": 0.016095709055662155,
+ "learning_rate": 0.0006,
+ "loss": 4.168505668640137,
+ "step": 5063
+ },
+ {
+ "epoch": 70.3353711790393,
+ "grad_norm": 0.015001940540969372,
+ "learning_rate": 0.0006,
+ "loss": 4.161739349365234,
+ "step": 5064
+ },
+ {
+ "epoch": 70.34934497816595,
+ "grad_norm": 0.014668774791061878,
+ "learning_rate": 0.0006,
+ "loss": 4.181029796600342,
+ "step": 5065
+ },
+ {
+ "epoch": 70.36331877729258,
+ "grad_norm": 0.013266284950077534,
+ "learning_rate": 0.0006,
+ "loss": 4.3052873611450195,
+ "step": 5066
+ },
+ {
+ "epoch": 70.37729257641921,
+ "grad_norm": 0.015396138653159142,
+ "learning_rate": 0.0006,
+ "loss": 4.34413480758667,
+ "step": 5067
+ },
+ {
+ "epoch": 70.39126637554585,
+ "grad_norm": 0.015511523932218552,
+ "learning_rate": 0.0006,
+ "loss": 4.235037803649902,
+ "step": 5068
+ },
+ {
+ "epoch": 70.40524017467249,
+ "grad_norm": 0.014825738966464996,
+ "learning_rate": 0.0006,
+ "loss": 4.262448310852051,
+ "step": 5069
+ },
+ {
+ "epoch": 70.41921397379913,
+ "grad_norm": 0.015309952199459076,
+ "learning_rate": 0.0006,
+ "loss": 4.258711814880371,
+ "step": 5070
+ },
+ {
+ "epoch": 70.43318777292576,
+ "grad_norm": 0.017614055424928665,
+ "learning_rate": 0.0006,
+ "loss": 4.286028861999512,
+ "step": 5071
+ },
+ {
+ "epoch": 70.44716157205241,
+ "grad_norm": 0.0166039876639843,
+ "learning_rate": 0.0006,
+ "loss": 4.149197578430176,
+ "step": 5072
+ },
+ {
+ "epoch": 70.46113537117904,
+ "grad_norm": 0.016911419108510017,
+ "learning_rate": 0.0006,
+ "loss": 4.271981239318848,
+ "step": 5073
+ },
+ {
+ "epoch": 70.47510917030567,
+ "grad_norm": 0.016806820407509804,
+ "learning_rate": 0.0006,
+ "loss": 4.213315010070801,
+ "step": 5074
+ },
+ {
+ "epoch": 70.48908296943232,
+ "grad_norm": 0.015221742913126945,
+ "learning_rate": 0.0006,
+ "loss": 4.171091079711914,
+ "step": 5075
+ },
+ {
+ "epoch": 70.50305676855895,
+ "grad_norm": 0.014175347983837128,
+ "learning_rate": 0.0006,
+ "loss": 4.185309410095215,
+ "step": 5076
+ },
+ {
+ "epoch": 70.5170305676856,
+ "grad_norm": 0.013897798955440521,
+ "learning_rate": 0.0006,
+ "loss": 4.250119209289551,
+ "step": 5077
+ },
+ {
+ "epoch": 70.53100436681223,
+ "grad_norm": 0.014791003428399563,
+ "learning_rate": 0.0006,
+ "loss": 4.2504682540893555,
+ "step": 5078
+ },
+ {
+ "epoch": 70.54497816593886,
+ "grad_norm": 0.016700128093361855,
+ "learning_rate": 0.0006,
+ "loss": 4.235134124755859,
+ "step": 5079
+ },
+ {
+ "epoch": 70.5589519650655,
+ "grad_norm": 0.016679363325238228,
+ "learning_rate": 0.0006,
+ "loss": 4.32028865814209,
+ "step": 5080
+ },
+ {
+ "epoch": 70.57292576419214,
+ "grad_norm": 0.015117401257157326,
+ "learning_rate": 0.0006,
+ "loss": 4.210995674133301,
+ "step": 5081
+ },
+ {
+ "epoch": 70.58689956331878,
+ "grad_norm": 0.014441263861954212,
+ "learning_rate": 0.0006,
+ "loss": 4.27849006652832,
+ "step": 5082
+ },
+ {
+ "epoch": 70.60087336244541,
+ "grad_norm": 0.01523780357092619,
+ "learning_rate": 0.0006,
+ "loss": 4.197360992431641,
+ "step": 5083
+ },
+ {
+ "epoch": 70.61484716157206,
+ "grad_norm": 0.016201738268136978,
+ "learning_rate": 0.0006,
+ "loss": 4.28239107131958,
+ "step": 5084
+ },
+ {
+ "epoch": 70.62882096069869,
+ "grad_norm": 0.016003292053937912,
+ "learning_rate": 0.0006,
+ "loss": 4.209715843200684,
+ "step": 5085
+ },
+ {
+ "epoch": 70.64279475982532,
+ "grad_norm": 0.01472893450409174,
+ "learning_rate": 0.0006,
+ "loss": 4.169078826904297,
+ "step": 5086
+ },
+ {
+ "epoch": 70.65676855895197,
+ "grad_norm": 0.01604326069355011,
+ "learning_rate": 0.0006,
+ "loss": 4.090967178344727,
+ "step": 5087
+ },
+ {
+ "epoch": 70.6707423580786,
+ "grad_norm": 0.01598356105387211,
+ "learning_rate": 0.0006,
+ "loss": 4.16630220413208,
+ "step": 5088
+ },
+ {
+ "epoch": 70.68471615720524,
+ "grad_norm": 0.016988780349493027,
+ "learning_rate": 0.0006,
+ "loss": 4.117822170257568,
+ "step": 5089
+ },
+ {
+ "epoch": 70.69868995633188,
+ "grad_norm": 0.018582893535494804,
+ "learning_rate": 0.0006,
+ "loss": 4.192902088165283,
+ "step": 5090
+ },
+ {
+ "epoch": 70.71266375545852,
+ "grad_norm": 0.01793155074119568,
+ "learning_rate": 0.0006,
+ "loss": 4.314806938171387,
+ "step": 5091
+ },
+ {
+ "epoch": 70.72663755458515,
+ "grad_norm": 0.016286427155137062,
+ "learning_rate": 0.0006,
+ "loss": 4.2137556076049805,
+ "step": 5092
+ },
+ {
+ "epoch": 70.74061135371178,
+ "grad_norm": 0.014718376100063324,
+ "learning_rate": 0.0006,
+ "loss": 4.3073015213012695,
+ "step": 5093
+ },
+ {
+ "epoch": 70.75458515283843,
+ "grad_norm": 0.014675077050924301,
+ "learning_rate": 0.0006,
+ "loss": 4.160961151123047,
+ "step": 5094
+ },
+ {
+ "epoch": 70.76855895196506,
+ "grad_norm": 0.015550913289189339,
+ "learning_rate": 0.0006,
+ "loss": 4.322887420654297,
+ "step": 5095
+ },
+ {
+ "epoch": 70.78253275109171,
+ "grad_norm": 0.01453235000371933,
+ "learning_rate": 0.0006,
+ "loss": 4.2915873527526855,
+ "step": 5096
+ },
+ {
+ "epoch": 70.79650655021834,
+ "grad_norm": 0.01459164172410965,
+ "learning_rate": 0.0006,
+ "loss": 4.247288227081299,
+ "step": 5097
+ },
+ {
+ "epoch": 70.81048034934499,
+ "grad_norm": 0.015395080670714378,
+ "learning_rate": 0.0006,
+ "loss": 4.360955715179443,
+ "step": 5098
+ },
+ {
+ "epoch": 70.82445414847162,
+ "grad_norm": 0.01317683607339859,
+ "learning_rate": 0.0006,
+ "loss": 4.181969165802002,
+ "step": 5099
+ },
+ {
+ "epoch": 70.83842794759825,
+ "grad_norm": 0.013917932286858559,
+ "learning_rate": 0.0006,
+ "loss": 4.264626979827881,
+ "step": 5100
+ },
+ {
+ "epoch": 70.8524017467249,
+ "grad_norm": 0.015268695540726185,
+ "learning_rate": 0.0006,
+ "loss": 4.287857532501221,
+ "step": 5101
+ },
+ {
+ "epoch": 70.86637554585153,
+ "grad_norm": 0.015531329438090324,
+ "learning_rate": 0.0006,
+ "loss": 4.294188499450684,
+ "step": 5102
+ },
+ {
+ "epoch": 70.88034934497817,
+ "grad_norm": 0.015607129782438278,
+ "learning_rate": 0.0006,
+ "loss": 4.331040859222412,
+ "step": 5103
+ },
+ {
+ "epoch": 70.8943231441048,
+ "grad_norm": 0.015297149308025837,
+ "learning_rate": 0.0006,
+ "loss": 4.243650436401367,
+ "step": 5104
+ },
+ {
+ "epoch": 70.90829694323143,
+ "grad_norm": 0.014790025539696217,
+ "learning_rate": 0.0006,
+ "loss": 4.159859657287598,
+ "step": 5105
+ },
+ {
+ "epoch": 70.92227074235808,
+ "grad_norm": 0.013466274365782738,
+ "learning_rate": 0.0006,
+ "loss": 4.221992492675781,
+ "step": 5106
+ },
+ {
+ "epoch": 70.93624454148471,
+ "grad_norm": 0.01361898984760046,
+ "learning_rate": 0.0006,
+ "loss": 4.278751373291016,
+ "step": 5107
+ },
+ {
+ "epoch": 70.95021834061136,
+ "grad_norm": 0.01325751468539238,
+ "learning_rate": 0.0006,
+ "loss": 4.166001319885254,
+ "step": 5108
+ },
+ {
+ "epoch": 70.96419213973799,
+ "grad_norm": 0.012727087363600731,
+ "learning_rate": 0.0006,
+ "loss": 4.219997406005859,
+ "step": 5109
+ },
+ {
+ "epoch": 70.97816593886463,
+ "grad_norm": 0.013966867700219154,
+ "learning_rate": 0.0006,
+ "loss": 4.207491397857666,
+ "step": 5110
+ },
+ {
+ "epoch": 70.99213973799127,
+ "grad_norm": 0.015053587965667248,
+ "learning_rate": 0.0006,
+ "loss": 4.343245983123779,
+ "step": 5111
+ },
+ {
+ "epoch": 71.0,
+ "grad_norm": 0.017723368480801582,
+ "learning_rate": 0.0006,
+ "loss": 4.356122016906738,
+ "step": 5112
+ },
+ {
+ "epoch": 71.0,
+ "eval_loss": 4.598303318023682,
+ "eval_runtime": 56.7521,
+ "eval_samples_per_second": 43.029,
+ "eval_steps_per_second": 1.357,
+ "step": 5112
+ },
+ {
+ "epoch": 71.01397379912663,
+ "grad_norm": 0.018996279686689377,
+ "learning_rate": 0.0006,
+ "loss": 4.234772205352783,
+ "step": 5113
+ },
+ {
+ "epoch": 71.02794759825328,
+ "grad_norm": 0.02041555754840374,
+ "learning_rate": 0.0006,
+ "loss": 4.275816917419434,
+ "step": 5114
+ },
+ {
+ "epoch": 71.04192139737991,
+ "grad_norm": 0.021955285221338272,
+ "learning_rate": 0.0006,
+ "loss": 4.166038513183594,
+ "step": 5115
+ },
+ {
+ "epoch": 71.05589519650655,
+ "grad_norm": 0.021308740600943565,
+ "learning_rate": 0.0006,
+ "loss": 4.13063907623291,
+ "step": 5116
+ },
+ {
+ "epoch": 71.06986899563319,
+ "grad_norm": 0.01747054047882557,
+ "learning_rate": 0.0006,
+ "loss": 4.200522422790527,
+ "step": 5117
+ },
+ {
+ "epoch": 71.08384279475983,
+ "grad_norm": 0.018544532358646393,
+ "learning_rate": 0.0006,
+ "loss": 4.173915386199951,
+ "step": 5118
+ },
+ {
+ "epoch": 71.09781659388646,
+ "grad_norm": 0.0182361900806427,
+ "learning_rate": 0.0006,
+ "loss": 4.27392578125,
+ "step": 5119
+ },
+ {
+ "epoch": 71.1117903930131,
+ "grad_norm": 0.017062073573470116,
+ "learning_rate": 0.0006,
+ "loss": 4.219576358795166,
+ "step": 5120
+ },
+ {
+ "epoch": 71.12576419213974,
+ "grad_norm": 0.016289666295051575,
+ "learning_rate": 0.0006,
+ "loss": 4.133376598358154,
+ "step": 5121
+ },
+ {
+ "epoch": 71.13973799126637,
+ "grad_norm": 0.014864642173051834,
+ "learning_rate": 0.0006,
+ "loss": 4.168276786804199,
+ "step": 5122
+ },
+ {
+ "epoch": 71.15371179039302,
+ "grad_norm": 0.015309540554881096,
+ "learning_rate": 0.0006,
+ "loss": 4.175103187561035,
+ "step": 5123
+ },
+ {
+ "epoch": 71.16768558951965,
+ "grad_norm": 0.01556059904396534,
+ "learning_rate": 0.0006,
+ "loss": 4.289899826049805,
+ "step": 5124
+ },
+ {
+ "epoch": 71.18165938864628,
+ "grad_norm": 0.01479957066476345,
+ "learning_rate": 0.0006,
+ "loss": 4.341495037078857,
+ "step": 5125
+ },
+ {
+ "epoch": 71.19563318777293,
+ "grad_norm": 0.014762179926037788,
+ "learning_rate": 0.0006,
+ "loss": 4.252333641052246,
+ "step": 5126
+ },
+ {
+ "epoch": 71.20960698689956,
+ "grad_norm": 0.014897473156452179,
+ "learning_rate": 0.0006,
+ "loss": 4.2244181632995605,
+ "step": 5127
+ },
+ {
+ "epoch": 71.2235807860262,
+ "grad_norm": 0.012990163639187813,
+ "learning_rate": 0.0006,
+ "loss": 4.237288475036621,
+ "step": 5128
+ },
+ {
+ "epoch": 71.23755458515284,
+ "grad_norm": 0.012858416885137558,
+ "learning_rate": 0.0006,
+ "loss": 4.265738487243652,
+ "step": 5129
+ },
+ {
+ "epoch": 71.25152838427948,
+ "grad_norm": 0.013454788364470005,
+ "learning_rate": 0.0006,
+ "loss": 4.289131164550781,
+ "step": 5130
+ },
+ {
+ "epoch": 71.26550218340611,
+ "grad_norm": 0.013717238791286945,
+ "learning_rate": 0.0006,
+ "loss": 4.2844014167785645,
+ "step": 5131
+ },
+ {
+ "epoch": 71.27947598253274,
+ "grad_norm": 0.0135411461815238,
+ "learning_rate": 0.0006,
+ "loss": 4.139082908630371,
+ "step": 5132
+ },
+ {
+ "epoch": 71.29344978165939,
+ "grad_norm": 0.01406062301248312,
+ "learning_rate": 0.0006,
+ "loss": 4.226719856262207,
+ "step": 5133
+ },
+ {
+ "epoch": 71.30742358078602,
+ "grad_norm": 0.014027190394699574,
+ "learning_rate": 0.0006,
+ "loss": 4.191699981689453,
+ "step": 5134
+ },
+ {
+ "epoch": 71.32139737991267,
+ "grad_norm": 0.013809907250106335,
+ "learning_rate": 0.0006,
+ "loss": 4.242542266845703,
+ "step": 5135
+ },
+ {
+ "epoch": 71.3353711790393,
+ "grad_norm": 0.013507804833352566,
+ "learning_rate": 0.0006,
+ "loss": 4.068206310272217,
+ "step": 5136
+ },
+ {
+ "epoch": 71.34934497816595,
+ "grad_norm": 0.01432533748447895,
+ "learning_rate": 0.0006,
+ "loss": 4.378222942352295,
+ "step": 5137
+ },
+ {
+ "epoch": 71.36331877729258,
+ "grad_norm": 0.01753002591431141,
+ "learning_rate": 0.0006,
+ "loss": 4.174706935882568,
+ "step": 5138
+ },
+ {
+ "epoch": 71.37729257641921,
+ "grad_norm": 0.020048344507813454,
+ "learning_rate": 0.0006,
+ "loss": 4.19920015335083,
+ "step": 5139
+ },
+ {
+ "epoch": 71.39126637554585,
+ "grad_norm": 0.018551401793956757,
+ "learning_rate": 0.0006,
+ "loss": 4.175182342529297,
+ "step": 5140
+ },
+ {
+ "epoch": 71.40524017467249,
+ "grad_norm": 0.01598959229886532,
+ "learning_rate": 0.0006,
+ "loss": 4.2305169105529785,
+ "step": 5141
+ },
+ {
+ "epoch": 71.41921397379913,
+ "grad_norm": 0.016029560938477516,
+ "learning_rate": 0.0006,
+ "loss": 4.230090618133545,
+ "step": 5142
+ },
+ {
+ "epoch": 71.43318777292576,
+ "grad_norm": 0.01595657505095005,
+ "learning_rate": 0.0006,
+ "loss": 4.138402938842773,
+ "step": 5143
+ },
+ {
+ "epoch": 71.44716157205241,
+ "grad_norm": 0.01623230241239071,
+ "learning_rate": 0.0006,
+ "loss": 4.196106433868408,
+ "step": 5144
+ },
+ {
+ "epoch": 71.46113537117904,
+ "grad_norm": 0.014057597145438194,
+ "learning_rate": 0.0006,
+ "loss": 4.2599897384643555,
+ "step": 5145
+ },
+ {
+ "epoch": 71.47510917030567,
+ "grad_norm": 0.017161158844828606,
+ "learning_rate": 0.0006,
+ "loss": 4.183909893035889,
+ "step": 5146
+ },
+ {
+ "epoch": 71.48908296943232,
+ "grad_norm": 0.017038581892848015,
+ "learning_rate": 0.0006,
+ "loss": 4.373933792114258,
+ "step": 5147
+ },
+ {
+ "epoch": 71.50305676855895,
+ "grad_norm": 0.015851367264986038,
+ "learning_rate": 0.0006,
+ "loss": 4.2620086669921875,
+ "step": 5148
+ },
+ {
+ "epoch": 71.5170305676856,
+ "grad_norm": 0.014598303474485874,
+ "learning_rate": 0.0006,
+ "loss": 4.197151184082031,
+ "step": 5149
+ },
+ {
+ "epoch": 71.53100436681223,
+ "grad_norm": 0.01339802611619234,
+ "learning_rate": 0.0006,
+ "loss": 4.110706329345703,
+ "step": 5150
+ },
+ {
+ "epoch": 71.54497816593886,
+ "grad_norm": 0.015567775815725327,
+ "learning_rate": 0.0006,
+ "loss": 4.1210246086120605,
+ "step": 5151
+ },
+ {
+ "epoch": 71.5589519650655,
+ "grad_norm": 0.015616007149219513,
+ "learning_rate": 0.0006,
+ "loss": 4.129796028137207,
+ "step": 5152
+ },
+ {
+ "epoch": 71.57292576419214,
+ "grad_norm": 0.014198964461684227,
+ "learning_rate": 0.0006,
+ "loss": 4.193232536315918,
+ "step": 5153
+ },
+ {
+ "epoch": 71.58689956331878,
+ "grad_norm": 0.015226383693516254,
+ "learning_rate": 0.0006,
+ "loss": 4.223779678344727,
+ "step": 5154
+ },
+ {
+ "epoch": 71.60087336244541,
+ "grad_norm": 0.015610109083354473,
+ "learning_rate": 0.0006,
+ "loss": 4.223179817199707,
+ "step": 5155
+ },
+ {
+ "epoch": 71.61484716157206,
+ "grad_norm": 0.016034303233027458,
+ "learning_rate": 0.0006,
+ "loss": 4.196593284606934,
+ "step": 5156
+ },
+ {
+ "epoch": 71.62882096069869,
+ "grad_norm": 0.015648581087589264,
+ "learning_rate": 0.0006,
+ "loss": 4.341497421264648,
+ "step": 5157
+ },
+ {
+ "epoch": 71.64279475982532,
+ "grad_norm": 0.014117540791630745,
+ "learning_rate": 0.0006,
+ "loss": 4.194307327270508,
+ "step": 5158
+ },
+ {
+ "epoch": 71.65676855895197,
+ "grad_norm": 0.015612686984241009,
+ "learning_rate": 0.0006,
+ "loss": 4.290152549743652,
+ "step": 5159
+ },
+ {
+ "epoch": 71.6707423580786,
+ "grad_norm": 0.01722259446978569,
+ "learning_rate": 0.0006,
+ "loss": 4.258018493652344,
+ "step": 5160
+ },
+ {
+ "epoch": 71.68471615720524,
+ "grad_norm": 0.015923351049423218,
+ "learning_rate": 0.0006,
+ "loss": 4.21684455871582,
+ "step": 5161
+ },
+ {
+ "epoch": 71.69868995633188,
+ "grad_norm": 0.013663450255990028,
+ "learning_rate": 0.0006,
+ "loss": 4.175601959228516,
+ "step": 5162
+ },
+ {
+ "epoch": 71.71266375545852,
+ "grad_norm": 0.017075596377253532,
+ "learning_rate": 0.0006,
+ "loss": 4.205138683319092,
+ "step": 5163
+ },
+ {
+ "epoch": 71.72663755458515,
+ "grad_norm": 0.01723487675189972,
+ "learning_rate": 0.0006,
+ "loss": 4.08839225769043,
+ "step": 5164
+ },
+ {
+ "epoch": 71.74061135371178,
+ "grad_norm": 0.017355505377054214,
+ "learning_rate": 0.0006,
+ "loss": 4.165130615234375,
+ "step": 5165
+ },
+ {
+ "epoch": 71.75458515283843,
+ "grad_norm": 0.01868261583149433,
+ "learning_rate": 0.0006,
+ "loss": 4.16977071762085,
+ "step": 5166
+ },
+ {
+ "epoch": 71.76855895196506,
+ "grad_norm": 0.01897321827709675,
+ "learning_rate": 0.0006,
+ "loss": 4.173497200012207,
+ "step": 5167
+ },
+ {
+ "epoch": 71.78253275109171,
+ "grad_norm": 0.02045825682580471,
+ "learning_rate": 0.0006,
+ "loss": 4.176192283630371,
+ "step": 5168
+ },
+ {
+ "epoch": 71.79650655021834,
+ "grad_norm": 0.022895624861121178,
+ "learning_rate": 0.0006,
+ "loss": 4.163009166717529,
+ "step": 5169
+ },
+ {
+ "epoch": 71.81048034934499,
+ "grad_norm": 0.02062244527041912,
+ "learning_rate": 0.0006,
+ "loss": 4.156868934631348,
+ "step": 5170
+ },
+ {
+ "epoch": 71.82445414847162,
+ "grad_norm": 0.0176058579236269,
+ "learning_rate": 0.0006,
+ "loss": 4.154374122619629,
+ "step": 5171
+ },
+ {
+ "epoch": 71.83842794759825,
+ "grad_norm": 0.018049508333206177,
+ "learning_rate": 0.0006,
+ "loss": 4.296828269958496,
+ "step": 5172
+ },
+ {
+ "epoch": 71.8524017467249,
+ "grad_norm": 0.019466130062937737,
+ "learning_rate": 0.0006,
+ "loss": 4.19450569152832,
+ "step": 5173
+ },
+ {
+ "epoch": 71.86637554585153,
+ "grad_norm": 0.01838165521621704,
+ "learning_rate": 0.0006,
+ "loss": 4.11812686920166,
+ "step": 5174
+ },
+ {
+ "epoch": 71.88034934497817,
+ "grad_norm": 0.01608051173388958,
+ "learning_rate": 0.0006,
+ "loss": 4.240464210510254,
+ "step": 5175
+ },
+ {
+ "epoch": 71.8943231441048,
+ "grad_norm": 0.015807967633008957,
+ "learning_rate": 0.0006,
+ "loss": 4.275581359863281,
+ "step": 5176
+ },
+ {
+ "epoch": 71.90829694323143,
+ "grad_norm": 0.017366575077176094,
+ "learning_rate": 0.0006,
+ "loss": 4.064750671386719,
+ "step": 5177
+ },
+ {
+ "epoch": 71.92227074235808,
+ "grad_norm": 0.016241202130913734,
+ "learning_rate": 0.0006,
+ "loss": 4.203697681427002,
+ "step": 5178
+ },
+ {
+ "epoch": 71.93624454148471,
+ "grad_norm": 0.019105732440948486,
+ "learning_rate": 0.0006,
+ "loss": 4.140021800994873,
+ "step": 5179
+ },
+ {
+ "epoch": 71.95021834061136,
+ "grad_norm": 0.018492674455046654,
+ "learning_rate": 0.0006,
+ "loss": 4.2534074783325195,
+ "step": 5180
+ },
+ {
+ "epoch": 71.96419213973799,
+ "grad_norm": 0.019301170483231544,
+ "learning_rate": 0.0006,
+ "loss": 4.381392478942871,
+ "step": 5181
+ },
+ {
+ "epoch": 71.97816593886463,
+ "grad_norm": 0.019571911543607712,
+ "learning_rate": 0.0006,
+ "loss": 4.131056785583496,
+ "step": 5182
+ },
+ {
+ "epoch": 71.99213973799127,
+ "grad_norm": 0.01859748177230358,
+ "learning_rate": 0.0006,
+ "loss": 4.278930187225342,
+ "step": 5183
+ },
+ {
+ "epoch": 72.0,
+ "grad_norm": 0.02017994411289692,
+ "learning_rate": 0.0006,
+ "loss": 4.116464614868164,
+ "step": 5184
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 28800,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 400,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": true,
+ "should_log": false,
+ "should_save": false,
+ "should_training_stop": false
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 2.2101216201592013e+18,
+ "train_batch_size": 8,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-5184/training_args.bin b/runs/l2r50-i2-fulle-lm/checkpoint-5184/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..cad232831c222cc7f8b938b8bbe42f61e7d89c3a
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-5184/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:f7c40e3bfa2f5ac1450ca1243e3f2a895d44a06ed5995e6aefd97f78d229d8c8
+size 4792
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-5472/chat_template.jinja b/runs/l2r50-i2-fulle-lm/checkpoint-5472/chat_template.jinja
new file mode 100644
index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-5472/chat_template.jinja
@@ -0,0 +1,85 @@
+{%- if tools %}
+ {{- '<|im_start|>system\n' }}
+ {%- if messages[0].role == 'system' %}
+ {{- messages[0].content + '\n\n' }}
+ {%- endif %}
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }}
+ {%- for tool in tools %}
+ {{- "\n" }}
+ {{- tool | tojson }}
+ {%- endfor %}
+ {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }}
+{%- else %}
+ {%- if messages[0].role == 'system' %}
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
+ {%- endif %}
+{%- endif %}
+{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
+{%- for message in messages[::-1] %}
+ {%- set index = (messages|length - 1) - loop.index0 %}
+ {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %}
+ {%- set ns.multi_step_tool = false %}
+ {%- set ns.last_query_index = index %}
+ {%- endif %}
+{%- endfor %}
+{%- for message in messages %}
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
+ {%- elif message.role == "assistant" %}
+ {%- set content = message.content %}
+ {%- set reasoning_content = '' %}
+ {%- if message.reasoning_content is defined and message.reasoning_content is not none %}
+ {%- set reasoning_content = message.reasoning_content %}
+ {%- else %}
+ {%- if '' in message.content %}
+ {%- set content = message.content.split('')[-1].lstrip('\n') %}
+ {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %}
+ {%- endif %}
+ {%- endif %}
+ {%- if loop.index0 > ns.last_query_index %}
+ {%- if loop.last or (not loop.last and reasoning_content) %}
+ {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- if message.tool_calls %}
+ {%- for tool_call in message.tool_calls %}
+ {%- if (loop.first and content) or (not loop.first) %}
+ {{- '\n' }}
+ {%- endif %}
+ {%- if tool_call.function %}
+ {%- set tool_call = tool_call.function %}
+ {%- endif %}
+ {{- '\n{"name": "' }}
+ {{- tool_call.name }}
+ {{- '", "arguments": ' }}
+ {%- if tool_call.arguments is string %}
+ {{- tool_call.arguments }}
+ {%- else %}
+ {{- tool_call.arguments | tojson }}
+ {%- endif %}
+ {{- '}\n' }}
+ {%- endfor %}
+ {%- endif %}
+ {{- '<|im_end|>\n' }}
+ {%- elif message.role == "tool" %}
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
+ {{- '<|im_start|>user' }}
+ {%- endif %}
+ {{- '\n\n' }}
+ {{- message.content }}
+ {{- '\n' }}
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
+ {{- '<|im_end|>\n' }}
+ {%- endif %}
+ {%- endif %}
+{%- endfor %}
+{%- if add_generation_prompt %}
+ {{- '<|im_start|>assistant\n' }}
+ {%- if enable_thinking is defined and enable_thinking is false %}
+ {{- '\n\n\n\n' }}
+ {%- endif %}
+{%- endif %}
\ No newline at end of file
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-5472/config.json b/runs/l2r50-i2-fulle-lm/checkpoint-5472/config.json
new file mode 100644
index 0000000000000000000000000000000000000000..22f250953c6ce8fc91ee4411856dd73041a415e2
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-5472/config.json
@@ -0,0 +1,32 @@
+{
+ "architectures": [
+ "LlamaForCausalLM"
+ ],
+ "attention_bias": false,
+ "attention_dropout": 0.0,
+ "bos_token_id": null,
+ "dtype": "float32",
+ "eos_token_id": 151645,
+ "head_dim": 128,
+ "hidden_act": "silu",
+ "hidden_size": 512,
+ "initializer_range": 0.02,
+ "intermediate_size": 1536,
+ "max_position_embeddings": 2048,
+ "mlp_bias": false,
+ "model_type": "llama",
+ "num_attention_heads": 4,
+ "num_hidden_layers": 20,
+ "num_key_value_heads": 4,
+ "pad_token_id": 151645,
+ "pretraining_tp": 1,
+ "rms_norm_eps": 1e-06,
+ "rope_parameters": {
+ "rope_theta": 10000.0,
+ "rope_type": "default"
+ },
+ "tie_word_embeddings": true,
+ "transformers_version": "5.5.0",
+ "use_cache": false,
+ "vocab_size": 151673
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-5472/generation_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-5472/generation_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-5472/generation_config.json
@@ -0,0 +1,11 @@
+{
+ "_from_model_config": true,
+ "eos_token_id": [
+ 151645
+ ],
+ "output_attentions": false,
+ "output_hidden_states": false,
+ "pad_token_id": 151645,
+ "transformers_version": "5.5.0",
+ "use_cache": true
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-5472/model.safetensors b/runs/l2r50-i2-fulle-lm/checkpoint-5472/model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..c6b0ceeffaee9156184bcddec1e3318b24ac0b57
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-5472/model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:5d6f0df4c6768e67e40e8b29861a944d8646a122543c6f0481e8117c070a0767
+size 583360328
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-5472/optimizer.pt b/runs/l2r50-i2-fulle-lm/checkpoint-5472/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..c8babd07599f242150deac02aa8cad8e803e7b3d
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-5472/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:43ff9f559aba664ab73241955a5fb27153ab1285ca17c90bc5e46c614a4ad698
+size 1166833530
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-5472/rng_state_0.pth b/runs/l2r50-i2-fulle-lm/checkpoint-5472/rng_state_0.pth
new file mode 100644
index 0000000000000000000000000000000000000000..aff14f8ce1841e93fa324277ef2fcf672ce6f320
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-5472/rng_state_0.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:07649dfce15de07bfbfc987960ffc4fc589f7b76aa2b77ac1b4369c58a4f5a9f
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-5472/rng_state_1.pth b/runs/l2r50-i2-fulle-lm/checkpoint-5472/rng_state_1.pth
new file mode 100644
index 0000000000000000000000000000000000000000..e5641e294b0c012dcee9b441e5a618d691fac622
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-5472/rng_state_1.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:84c1f1ab67a99075a008b840c7088c557b3bd74f7085e4eb57ae5933f97e3096
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-5472/rng_state_2.pth b/runs/l2r50-i2-fulle-lm/checkpoint-5472/rng_state_2.pth
new file mode 100644
index 0000000000000000000000000000000000000000..a861486932b28d15d049b3b3b1850788353f236e
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-5472/rng_state_2.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:c246352697b88657a8da26487f5ced4f002f630719de68950b06addd63ed20c3
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-5472/rng_state_3.pth b/runs/l2r50-i2-fulle-lm/checkpoint-5472/rng_state_3.pth
new file mode 100644
index 0000000000000000000000000000000000000000..6165592a8f1c0944664783a010f42e4ab136086b
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-5472/rng_state_3.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:c4cf00673b1da661d86e94245cd02b21123f7422892904494df18097422747aa
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-5472/scheduler.pt b/runs/l2r50-i2-fulle-lm/checkpoint-5472/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..bde6e47e59bea2a2d1c7be49dbaa3a2d4dfadf37
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-5472/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:ec3fa8fa903a5a893f5b74cfd95272dfde1a7a2149fd0f3ecec1adb7304a55b8
+size 1064
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-5472/tokenizer.json b/runs/l2r50-i2-fulle-lm/checkpoint-5472/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..53786016f070e11184813f3c83101dc5e93e392e
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-5472/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:c3c63fda1646a555448e8f03662542e04b11d103721319e4d9b267d7da03ceb8
+size 11424463
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-5472/tokenizer_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-5472/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..98e2fb9b2cc54785dc9d871b8e503f8b8f2defe7
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-5472/tokenizer_config.json
@@ -0,0 +1,20 @@
+{
+ "add_prefix_space": false,
+ "backend": "tokenizers",
+ "bos_token": null,
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "<|im_end|>",
+ "errors": "replace",
+ "extra_special_tokens": [
+ "<|l2r_pred|>",
+ "<|r2l_pred|>",
+ "<|next_1_pred|>",
+ "<|next_2_pred|>"
+ ],
+ "is_local": false,
+ "model_max_length": 131072,
+ "pad_token": "<|im_end|>",
+ "split_special_tokens": false,
+ "tokenizer_class": "Qwen2Tokenizer",
+ "unk_token": null
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-5472/trainer_state.json b/runs/l2r50-i2-fulle-lm/checkpoint-5472/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..5887924a875c518f229ecdaa38b4d886282f279f
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-5472/trainer_state.json
@@ -0,0 +1,38938 @@
+{
+ "best_global_step": null,
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 76.0,
+ "eval_steps": 500,
+ "global_step": 5472,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.013973799126637555,
+ "grad_norm": 0.1739506870508194,
+ "learning_rate": 0.0,
+ "loss": 11.99299430847168,
+ "step": 1
+ },
+ {
+ "epoch": 0.02794759825327511,
+ "grad_norm": 0.17216597497463226,
+ "learning_rate": 5.999999999999999e-06,
+ "loss": 11.988704681396484,
+ "step": 2
+ },
+ {
+ "epoch": 0.04192139737991266,
+ "grad_norm": 0.17459715902805328,
+ "learning_rate": 1.1999999999999999e-05,
+ "loss": 11.945581436157227,
+ "step": 3
+ },
+ {
+ "epoch": 0.05589519650655022,
+ "grad_norm": 0.17577414214611053,
+ "learning_rate": 1.7999999999999997e-05,
+ "loss": 11.859772682189941,
+ "step": 4
+ },
+ {
+ "epoch": 0.06986899563318777,
+ "grad_norm": 0.16163651645183563,
+ "learning_rate": 2.3999999999999997e-05,
+ "loss": 11.744415283203125,
+ "step": 5
+ },
+ {
+ "epoch": 0.08384279475982533,
+ "grad_norm": 0.14409932494163513,
+ "learning_rate": 2.9999999999999997e-05,
+ "loss": 11.618099212646484,
+ "step": 6
+ },
+ {
+ "epoch": 0.09781659388646288,
+ "grad_norm": 0.13079848885536194,
+ "learning_rate": 3.5999999999999994e-05,
+ "loss": 11.505935668945312,
+ "step": 7
+ },
+ {
+ "epoch": 0.11179039301310044,
+ "grad_norm": 0.11964613944292068,
+ "learning_rate": 4.2e-05,
+ "loss": 11.416532516479492,
+ "step": 8
+ },
+ {
+ "epoch": 0.125764192139738,
+ "grad_norm": 0.11345142871141434,
+ "learning_rate": 4.7999999999999994e-05,
+ "loss": 11.33206558227539,
+ "step": 9
+ },
+ {
+ "epoch": 0.13973799126637554,
+ "grad_norm": 0.11033952981233597,
+ "learning_rate": 5.399999999999999e-05,
+ "loss": 11.259052276611328,
+ "step": 10
+ },
+ {
+ "epoch": 0.1537117903930131,
+ "grad_norm": 0.10776489973068237,
+ "learning_rate": 5.9999999999999995e-05,
+ "loss": 11.203384399414062,
+ "step": 11
+ },
+ {
+ "epoch": 0.16768558951965065,
+ "grad_norm": 0.10721946507692337,
+ "learning_rate": 6.599999999999999e-05,
+ "loss": 11.153295516967773,
+ "step": 12
+ },
+ {
+ "epoch": 0.18165938864628822,
+ "grad_norm": 0.10673430562019348,
+ "learning_rate": 7.199999999999999e-05,
+ "loss": 11.110200881958008,
+ "step": 13
+ },
+ {
+ "epoch": 0.19563318777292577,
+ "grad_norm": 0.10630254447460175,
+ "learning_rate": 7.8e-05,
+ "loss": 11.067477226257324,
+ "step": 14
+ },
+ {
+ "epoch": 0.2096069868995633,
+ "grad_norm": 0.10631615668535233,
+ "learning_rate": 8.4e-05,
+ "loss": 11.022941589355469,
+ "step": 15
+ },
+ {
+ "epoch": 0.22358078602620088,
+ "grad_norm": 0.10609443485736847,
+ "learning_rate": 8.999999999999999e-05,
+ "loss": 10.970746040344238,
+ "step": 16
+ },
+ {
+ "epoch": 0.23755458515283842,
+ "grad_norm": 0.10648232698440552,
+ "learning_rate": 9.599999999999999e-05,
+ "loss": 10.91267204284668,
+ "step": 17
+ },
+ {
+ "epoch": 0.251528384279476,
+ "grad_norm": 0.10584589838981628,
+ "learning_rate": 0.000102,
+ "loss": 10.854547500610352,
+ "step": 18
+ },
+ {
+ "epoch": 0.26550218340611353,
+ "grad_norm": 0.10591720044612885,
+ "learning_rate": 0.00010799999999999998,
+ "loss": 10.788917541503906,
+ "step": 19
+ },
+ {
+ "epoch": 0.2794759825327511,
+ "grad_norm": 0.10611672699451447,
+ "learning_rate": 0.00011399999999999999,
+ "loss": 10.717556953430176,
+ "step": 20
+ },
+ {
+ "epoch": 0.2934497816593886,
+ "grad_norm": 0.10455359518527985,
+ "learning_rate": 0.00011999999999999999,
+ "loss": 10.650186538696289,
+ "step": 21
+ },
+ {
+ "epoch": 0.3074235807860262,
+ "grad_norm": 0.1047525629401207,
+ "learning_rate": 0.00012599999999999997,
+ "loss": 10.56808090209961,
+ "step": 22
+ },
+ {
+ "epoch": 0.32139737991266376,
+ "grad_norm": 0.10511605441570282,
+ "learning_rate": 0.00013199999999999998,
+ "loss": 10.484028816223145,
+ "step": 23
+ },
+ {
+ "epoch": 0.3353711790393013,
+ "grad_norm": 0.10467177629470825,
+ "learning_rate": 0.000138,
+ "loss": 10.399947166442871,
+ "step": 24
+ },
+ {
+ "epoch": 0.34934497816593885,
+ "grad_norm": 0.10432136058807373,
+ "learning_rate": 0.00014399999999999998,
+ "loss": 10.312714576721191,
+ "step": 25
+ },
+ {
+ "epoch": 0.36331877729257644,
+ "grad_norm": 0.10527262091636658,
+ "learning_rate": 0.00015,
+ "loss": 10.212261199951172,
+ "step": 26
+ },
+ {
+ "epoch": 0.377292576419214,
+ "grad_norm": 0.103610560297966,
+ "learning_rate": 0.000156,
+ "loss": 10.132034301757812,
+ "step": 27
+ },
+ {
+ "epoch": 0.39126637554585153,
+ "grad_norm": 0.10553059726953506,
+ "learning_rate": 0.000162,
+ "loss": 10.020434379577637,
+ "step": 28
+ },
+ {
+ "epoch": 0.4052401746724891,
+ "grad_norm": 0.1037299782037735,
+ "learning_rate": 0.000168,
+ "loss": 9.934466361999512,
+ "step": 29
+ },
+ {
+ "epoch": 0.4192139737991266,
+ "grad_norm": 0.1042739674448967,
+ "learning_rate": 0.00017399999999999997,
+ "loss": 9.822539329528809,
+ "step": 30
+ },
+ {
+ "epoch": 0.4331877729257642,
+ "grad_norm": 0.10300865024328232,
+ "learning_rate": 0.00017999999999999998,
+ "loss": 9.734994888305664,
+ "step": 31
+ },
+ {
+ "epoch": 0.44716157205240176,
+ "grad_norm": 0.10291654616594315,
+ "learning_rate": 0.000186,
+ "loss": 9.624505996704102,
+ "step": 32
+ },
+ {
+ "epoch": 0.4611353711790393,
+ "grad_norm": 0.10210851579904556,
+ "learning_rate": 0.00019199999999999998,
+ "loss": 9.518205642700195,
+ "step": 33
+ },
+ {
+ "epoch": 0.47510917030567684,
+ "grad_norm": 0.10222529619932175,
+ "learning_rate": 0.000198,
+ "loss": 9.412222862243652,
+ "step": 34
+ },
+ {
+ "epoch": 0.4890829694323144,
+ "grad_norm": 0.09992019087076187,
+ "learning_rate": 0.000204,
+ "loss": 9.326566696166992,
+ "step": 35
+ },
+ {
+ "epoch": 0.503056768558952,
+ "grad_norm": 0.09931084513664246,
+ "learning_rate": 0.00020999999999999998,
+ "loss": 9.210111618041992,
+ "step": 36
+ },
+ {
+ "epoch": 0.5170305676855895,
+ "grad_norm": 0.09928256273269653,
+ "learning_rate": 0.00021599999999999996,
+ "loss": 9.100493431091309,
+ "step": 37
+ },
+ {
+ "epoch": 0.5310043668122271,
+ "grad_norm": 0.09843816608190536,
+ "learning_rate": 0.00022199999999999998,
+ "loss": 8.996543884277344,
+ "step": 38
+ },
+ {
+ "epoch": 0.5449781659388646,
+ "grad_norm": 0.09825374186038971,
+ "learning_rate": 0.00022799999999999999,
+ "loss": 8.888991355895996,
+ "step": 39
+ },
+ {
+ "epoch": 0.5589519650655022,
+ "grad_norm": 0.09602385759353638,
+ "learning_rate": 0.000234,
+ "loss": 8.802907943725586,
+ "step": 40
+ },
+ {
+ "epoch": 0.5729257641921397,
+ "grad_norm": 0.09414064139127731,
+ "learning_rate": 0.00023999999999999998,
+ "loss": 8.715423583984375,
+ "step": 41
+ },
+ {
+ "epoch": 0.5868995633187772,
+ "grad_norm": 0.09273893386125565,
+ "learning_rate": 0.00024599999999999996,
+ "loss": 8.61374282836914,
+ "step": 42
+ },
+ {
+ "epoch": 0.6008733624454149,
+ "grad_norm": 0.09079599380493164,
+ "learning_rate": 0.00025199999999999995,
+ "loss": 8.502904891967773,
+ "step": 43
+ },
+ {
+ "epoch": 0.6148471615720524,
+ "grad_norm": 0.08605825901031494,
+ "learning_rate": 0.000258,
+ "loss": 8.453008651733398,
+ "step": 44
+ },
+ {
+ "epoch": 0.62882096069869,
+ "grad_norm": 0.08338996022939682,
+ "learning_rate": 0.00026399999999999997,
+ "loss": 8.365163803100586,
+ "step": 45
+ },
+ {
+ "epoch": 0.6427947598253275,
+ "grad_norm": 0.08150867372751236,
+ "learning_rate": 0.00027,
+ "loss": 8.285372734069824,
+ "step": 46
+ },
+ {
+ "epoch": 0.6567685589519651,
+ "grad_norm": 0.07794807851314545,
+ "learning_rate": 0.000276,
+ "loss": 8.185811996459961,
+ "step": 47
+ },
+ {
+ "epoch": 0.6707423580786026,
+ "grad_norm": 0.07592720538377762,
+ "learning_rate": 0.00028199999999999997,
+ "loss": 8.098482131958008,
+ "step": 48
+ },
+ {
+ "epoch": 0.6847161572052402,
+ "grad_norm": 0.0709897130727768,
+ "learning_rate": 0.00028799999999999995,
+ "loss": 8.042091369628906,
+ "step": 49
+ },
+ {
+ "epoch": 0.6986899563318777,
+ "grad_norm": 0.06582305580377579,
+ "learning_rate": 0.000294,
+ "loss": 7.972844123840332,
+ "step": 50
+ },
+ {
+ "epoch": 0.7126637554585152,
+ "grad_norm": 0.059452351182699203,
+ "learning_rate": 0.0003,
+ "loss": 7.924579620361328,
+ "step": 51
+ },
+ {
+ "epoch": 0.7266375545851529,
+ "grad_norm": 0.052426718175411224,
+ "learning_rate": 0.00030599999999999996,
+ "loss": 7.876336097717285,
+ "step": 52
+ },
+ {
+ "epoch": 0.7406113537117904,
+ "grad_norm": 0.04768150672316551,
+ "learning_rate": 0.000312,
+ "loss": 7.8102707862854,
+ "step": 53
+ },
+ {
+ "epoch": 0.754585152838428,
+ "grad_norm": 0.03970075398683548,
+ "learning_rate": 0.000318,
+ "loss": 7.791913986206055,
+ "step": 54
+ },
+ {
+ "epoch": 0.7685589519650655,
+ "grad_norm": 0.03559752553701401,
+ "learning_rate": 0.000324,
+ "loss": 7.756960868835449,
+ "step": 55
+ },
+ {
+ "epoch": 0.7825327510917031,
+ "grad_norm": 0.029321353882551193,
+ "learning_rate": 0.00033,
+ "loss": 7.724878787994385,
+ "step": 56
+ },
+ {
+ "epoch": 0.7965065502183406,
+ "grad_norm": 0.023485427722334862,
+ "learning_rate": 0.000336,
+ "loss": 7.72553014755249,
+ "step": 57
+ },
+ {
+ "epoch": 0.8104803493449781,
+ "grad_norm": 0.019319474697113037,
+ "learning_rate": 0.00034199999999999996,
+ "loss": 7.709493160247803,
+ "step": 58
+ },
+ {
+ "epoch": 0.8244541484716157,
+ "grad_norm": 0.016705691814422607,
+ "learning_rate": 0.00034799999999999995,
+ "loss": 7.6873297691345215,
+ "step": 59
+ },
+ {
+ "epoch": 0.8384279475982532,
+ "grad_norm": 0.017667734995484352,
+ "learning_rate": 0.00035399999999999993,
+ "loss": 7.677059173583984,
+ "step": 60
+ },
+ {
+ "epoch": 0.8524017467248908,
+ "grad_norm": 0.01671992614865303,
+ "learning_rate": 0.00035999999999999997,
+ "loss": 7.643512725830078,
+ "step": 61
+ },
+ {
+ "epoch": 0.8663755458515284,
+ "grad_norm": 0.012288172729313374,
+ "learning_rate": 0.00036599999999999995,
+ "loss": 7.658343315124512,
+ "step": 62
+ },
+ {
+ "epoch": 0.880349344978166,
+ "grad_norm": 0.012004776857793331,
+ "learning_rate": 0.000372,
+ "loss": 7.649759292602539,
+ "step": 63
+ },
+ {
+ "epoch": 0.8943231441048035,
+ "grad_norm": 0.012487445957958698,
+ "learning_rate": 0.00037799999999999997,
+ "loss": 7.631503105163574,
+ "step": 64
+ },
+ {
+ "epoch": 0.9082969432314411,
+ "grad_norm": 0.01489241048693657,
+ "learning_rate": 0.00038399999999999996,
+ "loss": 7.647110462188721,
+ "step": 65
+ },
+ {
+ "epoch": 0.9222707423580786,
+ "grad_norm": 0.01428980939090252,
+ "learning_rate": 0.00039,
+ "loss": 7.636983871459961,
+ "step": 66
+ },
+ {
+ "epoch": 0.9362445414847161,
+ "grad_norm": 0.010711144655942917,
+ "learning_rate": 0.000396,
+ "loss": 7.621130466461182,
+ "step": 67
+ },
+ {
+ "epoch": 0.9502183406113537,
+ "grad_norm": 0.010178886353969574,
+ "learning_rate": 0.000402,
+ "loss": 7.593924522399902,
+ "step": 68
+ },
+ {
+ "epoch": 0.9641921397379912,
+ "grad_norm": 0.010604874230921268,
+ "learning_rate": 0.000408,
+ "loss": 7.6015496253967285,
+ "step": 69
+ },
+ {
+ "epoch": 0.9781659388646288,
+ "grad_norm": 0.009751426056027412,
+ "learning_rate": 0.0004139999999999999,
+ "loss": 7.571964263916016,
+ "step": 70
+ },
+ {
+ "epoch": 0.9921397379912664,
+ "grad_norm": 0.009816481731832027,
+ "learning_rate": 0.00041999999999999996,
+ "loss": 7.57344913482666,
+ "step": 71
+ },
+ {
+ "epoch": 1.0,
+ "grad_norm": 0.008665637113153934,
+ "learning_rate": 0.00042599999999999995,
+ "loss": 7.579050064086914,
+ "step": 72
+ },
+ {
+ "epoch": 1.0,
+ "eval_loss": 7.581204891204834,
+ "eval_runtime": 59.0091,
+ "eval_samples_per_second": 41.383,
+ "eval_steps_per_second": 1.305,
+ "step": 72
+ },
+ {
+ "epoch": 1.0139737991266375,
+ "grad_norm": 0.009684201329946518,
+ "learning_rate": 0.00043199999999999993,
+ "loss": 7.568594932556152,
+ "step": 73
+ },
+ {
+ "epoch": 1.027947598253275,
+ "grad_norm": 0.025783710181713104,
+ "learning_rate": 0.00043799999999999997,
+ "loss": 7.578948497772217,
+ "step": 74
+ },
+ {
+ "epoch": 1.0419213973799126,
+ "grad_norm": 0.0327819399535656,
+ "learning_rate": 0.00044399999999999995,
+ "loss": 7.548173904418945,
+ "step": 75
+ },
+ {
+ "epoch": 1.0558951965065502,
+ "grad_norm": 0.021604498848319054,
+ "learning_rate": 0.00045,
+ "loss": 7.552953720092773,
+ "step": 76
+ },
+ {
+ "epoch": 1.0698689956331877,
+ "grad_norm": 0.010230105370283127,
+ "learning_rate": 0.00045599999999999997,
+ "loss": 7.5213212966918945,
+ "step": 77
+ },
+ {
+ "epoch": 1.0838427947598253,
+ "grad_norm": 0.023315589874982834,
+ "learning_rate": 0.00046199999999999995,
+ "loss": 7.501640319824219,
+ "step": 78
+ },
+ {
+ "epoch": 1.0978165938864628,
+ "grad_norm": 0.01109944935888052,
+ "learning_rate": 0.000468,
+ "loss": 7.539519786834717,
+ "step": 79
+ },
+ {
+ "epoch": 1.1117903930131003,
+ "grad_norm": 0.026093758642673492,
+ "learning_rate": 0.000474,
+ "loss": 7.525928974151611,
+ "step": 80
+ },
+ {
+ "epoch": 1.125764192139738,
+ "grad_norm": 0.03274895250797272,
+ "learning_rate": 0.00047999999999999996,
+ "loss": 7.535602569580078,
+ "step": 81
+ },
+ {
+ "epoch": 1.1397379912663754,
+ "grad_norm": 0.01543602254241705,
+ "learning_rate": 0.000486,
+ "loss": 7.513683795928955,
+ "step": 82
+ },
+ {
+ "epoch": 1.1537117903930132,
+ "grad_norm": 0.028377249836921692,
+ "learning_rate": 0.0004919999999999999,
+ "loss": 7.49384880065918,
+ "step": 83
+ },
+ {
+ "epoch": 1.1676855895196507,
+ "grad_norm": 0.04121481627225876,
+ "learning_rate": 0.000498,
+ "loss": 7.488447189331055,
+ "step": 84
+ },
+ {
+ "epoch": 1.1816593886462883,
+ "grad_norm": 0.013282302767038345,
+ "learning_rate": 0.0005039999999999999,
+ "loss": 7.463010787963867,
+ "step": 85
+ },
+ {
+ "epoch": 1.1956331877729258,
+ "grad_norm": 0.03200364485383034,
+ "learning_rate": 0.0005099999999999999,
+ "loss": 7.473515033721924,
+ "step": 86
+ },
+ {
+ "epoch": 1.2096069868995634,
+ "grad_norm": 0.020768029615283012,
+ "learning_rate": 0.000516,
+ "loss": 7.464688777923584,
+ "step": 87
+ },
+ {
+ "epoch": 1.223580786026201,
+ "grad_norm": 0.021715788170695305,
+ "learning_rate": 0.000522,
+ "loss": 7.4835028648376465,
+ "step": 88
+ },
+ {
+ "epoch": 1.2375545851528384,
+ "grad_norm": 0.026255754753947258,
+ "learning_rate": 0.0005279999999999999,
+ "loss": 7.443392753601074,
+ "step": 89
+ },
+ {
+ "epoch": 1.251528384279476,
+ "grad_norm": 0.022325951606035233,
+ "learning_rate": 0.000534,
+ "loss": 7.457373142242432,
+ "step": 90
+ },
+ {
+ "epoch": 1.2655021834061135,
+ "grad_norm": 0.009270387701690197,
+ "learning_rate": 0.00054,
+ "loss": 7.455986976623535,
+ "step": 91
+ },
+ {
+ "epoch": 1.279475982532751,
+ "grad_norm": 0.02369544468820095,
+ "learning_rate": 0.0005459999999999999,
+ "loss": 7.442216873168945,
+ "step": 92
+ },
+ {
+ "epoch": 1.2934497816593886,
+ "grad_norm": 0.008474506437778473,
+ "learning_rate": 0.000552,
+ "loss": 7.4294915199279785,
+ "step": 93
+ },
+ {
+ "epoch": 1.3074235807860262,
+ "grad_norm": 0.017043249681591988,
+ "learning_rate": 0.000558,
+ "loss": 7.406932830810547,
+ "step": 94
+ },
+ {
+ "epoch": 1.3213973799126637,
+ "grad_norm": 0.011813412420451641,
+ "learning_rate": 0.0005639999999999999,
+ "loss": 7.415900230407715,
+ "step": 95
+ },
+ {
+ "epoch": 1.3353711790393012,
+ "grad_norm": 0.01086883433163166,
+ "learning_rate": 0.00057,
+ "loss": 7.452998638153076,
+ "step": 96
+ },
+ {
+ "epoch": 1.3493449781659388,
+ "grad_norm": 0.014700951054692268,
+ "learning_rate": 0.0005759999999999999,
+ "loss": 7.410249710083008,
+ "step": 97
+ },
+ {
+ "epoch": 1.3633187772925766,
+ "grad_norm": 0.014512976631522179,
+ "learning_rate": 0.0005819999999999999,
+ "loss": 7.429132461547852,
+ "step": 98
+ },
+ {
+ "epoch": 1.3772925764192139,
+ "grad_norm": 0.012519920244812965,
+ "learning_rate": 0.000588,
+ "loss": 7.430817604064941,
+ "step": 99
+ },
+ {
+ "epoch": 1.3912663755458516,
+ "grad_norm": 0.007132664322853088,
+ "learning_rate": 0.0005939999999999999,
+ "loss": 7.405234336853027,
+ "step": 100
+ },
+ {
+ "epoch": 1.405240174672489,
+ "grad_norm": 0.019489388912916183,
+ "learning_rate": 0.0006,
+ "loss": 7.427791595458984,
+ "step": 101
+ },
+ {
+ "epoch": 1.4192139737991267,
+ "grad_norm": 0.027252664789557457,
+ "learning_rate": 0.0006,
+ "loss": 7.41761589050293,
+ "step": 102
+ },
+ {
+ "epoch": 1.4331877729257643,
+ "grad_norm": 0.02501637488603592,
+ "learning_rate": 0.0006,
+ "loss": 7.416735649108887,
+ "step": 103
+ },
+ {
+ "epoch": 1.4471615720524018,
+ "grad_norm": 0.00989772193133831,
+ "learning_rate": 0.0006,
+ "loss": 7.3756632804870605,
+ "step": 104
+ },
+ {
+ "epoch": 1.4611353711790394,
+ "grad_norm": 0.01622949168086052,
+ "learning_rate": 0.0006,
+ "loss": 7.400811672210693,
+ "step": 105
+ },
+ {
+ "epoch": 1.475109170305677,
+ "grad_norm": 0.012407763861119747,
+ "learning_rate": 0.0006,
+ "loss": 7.37675666809082,
+ "step": 106
+ },
+ {
+ "epoch": 1.4890829694323144,
+ "grad_norm": 0.010353565216064453,
+ "learning_rate": 0.0006,
+ "loss": 7.4013142585754395,
+ "step": 107
+ },
+ {
+ "epoch": 1.503056768558952,
+ "grad_norm": 0.0120553532615304,
+ "learning_rate": 0.0006,
+ "loss": 7.379873275756836,
+ "step": 108
+ },
+ {
+ "epoch": 1.5170305676855895,
+ "grad_norm": 0.01610225811600685,
+ "learning_rate": 0.0006,
+ "loss": 7.371631145477295,
+ "step": 109
+ },
+ {
+ "epoch": 1.531004366812227,
+ "grad_norm": 0.01159879844635725,
+ "learning_rate": 0.0006,
+ "loss": 7.389007568359375,
+ "step": 110
+ },
+ {
+ "epoch": 1.5449781659388646,
+ "grad_norm": 0.008013816550374031,
+ "learning_rate": 0.0006,
+ "loss": 7.395920753479004,
+ "step": 111
+ },
+ {
+ "epoch": 1.5589519650655022,
+ "grad_norm": 0.016334809362888336,
+ "learning_rate": 0.0006,
+ "loss": 7.393121719360352,
+ "step": 112
+ },
+ {
+ "epoch": 1.5729257641921397,
+ "grad_norm": 0.018821721896529198,
+ "learning_rate": 0.0006,
+ "loss": 7.367543697357178,
+ "step": 113
+ },
+ {
+ "epoch": 1.5868995633187772,
+ "grad_norm": 0.014088994823396206,
+ "learning_rate": 0.0006,
+ "loss": 7.385537624359131,
+ "step": 114
+ },
+ {
+ "epoch": 1.600873362445415,
+ "grad_norm": 0.005864050704985857,
+ "learning_rate": 0.0006,
+ "loss": 7.365346908569336,
+ "step": 115
+ },
+ {
+ "epoch": 1.6148471615720523,
+ "grad_norm": 0.011745430529117584,
+ "learning_rate": 0.0006,
+ "loss": 7.390227317810059,
+ "step": 116
+ },
+ {
+ "epoch": 1.62882096069869,
+ "grad_norm": 0.017205573618412018,
+ "learning_rate": 0.0006,
+ "loss": 7.3558549880981445,
+ "step": 117
+ },
+ {
+ "epoch": 1.6427947598253274,
+ "grad_norm": 0.021256346255540848,
+ "learning_rate": 0.0006,
+ "loss": 7.377346992492676,
+ "step": 118
+ },
+ {
+ "epoch": 1.6567685589519652,
+ "grad_norm": 0.019335204735398293,
+ "learning_rate": 0.0006,
+ "loss": 7.346194267272949,
+ "step": 119
+ },
+ {
+ "epoch": 1.6707423580786025,
+ "grad_norm": 0.008134003728628159,
+ "learning_rate": 0.0006,
+ "loss": 7.3063578605651855,
+ "step": 120
+ },
+ {
+ "epoch": 1.6847161572052403,
+ "grad_norm": 0.017613831907510757,
+ "learning_rate": 0.0006,
+ "loss": 7.338171005249023,
+ "step": 121
+ },
+ {
+ "epoch": 1.6986899563318776,
+ "grad_norm": 0.00890734139829874,
+ "learning_rate": 0.0006,
+ "loss": 7.355384826660156,
+ "step": 122
+ },
+ {
+ "epoch": 1.7126637554585153,
+ "grad_norm": 0.013448155485093594,
+ "learning_rate": 0.0006,
+ "loss": 7.344296455383301,
+ "step": 123
+ },
+ {
+ "epoch": 1.726637554585153,
+ "grad_norm": 0.009241633117198944,
+ "learning_rate": 0.0006,
+ "loss": 7.3454060554504395,
+ "step": 124
+ },
+ {
+ "epoch": 1.7406113537117904,
+ "grad_norm": 0.018561311066150665,
+ "learning_rate": 0.0006,
+ "loss": 7.32348108291626,
+ "step": 125
+ },
+ {
+ "epoch": 1.754585152838428,
+ "grad_norm": 0.025667704641819,
+ "learning_rate": 0.0006,
+ "loss": 7.356142520904541,
+ "step": 126
+ },
+ {
+ "epoch": 1.7685589519650655,
+ "grad_norm": 0.036232151091098785,
+ "learning_rate": 0.0006,
+ "loss": 7.363539695739746,
+ "step": 127
+ },
+ {
+ "epoch": 1.782532751091703,
+ "grad_norm": 0.059483714401721954,
+ "learning_rate": 0.0006,
+ "loss": 7.402331352233887,
+ "step": 128
+ },
+ {
+ "epoch": 1.7965065502183406,
+ "grad_norm": 0.05365009233355522,
+ "learning_rate": 0.0006,
+ "loss": 7.40457820892334,
+ "step": 129
+ },
+ {
+ "epoch": 1.8104803493449781,
+ "grad_norm": 0.02738415263593197,
+ "learning_rate": 0.0006,
+ "loss": 7.3756208419799805,
+ "step": 130
+ },
+ {
+ "epoch": 1.8244541484716157,
+ "grad_norm": 0.035017091780900955,
+ "learning_rate": 0.0006,
+ "loss": 7.379263877868652,
+ "step": 131
+ },
+ {
+ "epoch": 1.8384279475982532,
+ "grad_norm": 0.02819863148033619,
+ "learning_rate": 0.0006,
+ "loss": 7.327673435211182,
+ "step": 132
+ },
+ {
+ "epoch": 1.8524017467248908,
+ "grad_norm": 0.017754238098859787,
+ "learning_rate": 0.0006,
+ "loss": 7.333420276641846,
+ "step": 133
+ },
+ {
+ "epoch": 1.8663755458515285,
+ "grad_norm": 0.024685733020305634,
+ "learning_rate": 0.0006,
+ "loss": 7.34493350982666,
+ "step": 134
+ },
+ {
+ "epoch": 1.8803493449781659,
+ "grad_norm": 0.019426532089710236,
+ "learning_rate": 0.0006,
+ "loss": 7.328448295593262,
+ "step": 135
+ },
+ {
+ "epoch": 1.8943231441048036,
+ "grad_norm": 0.02106391079723835,
+ "learning_rate": 0.0006,
+ "loss": 7.327917098999023,
+ "step": 136
+ },
+ {
+ "epoch": 1.908296943231441,
+ "grad_norm": 0.02056879736483097,
+ "learning_rate": 0.0006,
+ "loss": 7.3311991691589355,
+ "step": 137
+ },
+ {
+ "epoch": 1.9222707423580787,
+ "grad_norm": 0.016865141689777374,
+ "learning_rate": 0.0006,
+ "loss": 7.312766075134277,
+ "step": 138
+ },
+ {
+ "epoch": 1.936244541484716,
+ "grad_norm": 0.017549855634570122,
+ "learning_rate": 0.0006,
+ "loss": 7.31191349029541,
+ "step": 139
+ },
+ {
+ "epoch": 1.9502183406113538,
+ "grad_norm": 0.015930157154798508,
+ "learning_rate": 0.0006,
+ "loss": 7.330901622772217,
+ "step": 140
+ },
+ {
+ "epoch": 1.9641921397379911,
+ "grad_norm": 0.013283558189868927,
+ "learning_rate": 0.0006,
+ "loss": 7.298318386077881,
+ "step": 141
+ },
+ {
+ "epoch": 1.9781659388646289,
+ "grad_norm": 0.012023469433188438,
+ "learning_rate": 0.0006,
+ "loss": 7.282337188720703,
+ "step": 142
+ },
+ {
+ "epoch": 1.9921397379912664,
+ "grad_norm": 0.01459511648863554,
+ "learning_rate": 0.0006,
+ "loss": 7.282924175262451,
+ "step": 143
+ },
+ {
+ "epoch": 2.0,
+ "grad_norm": 0.017474235966801643,
+ "learning_rate": 0.0006,
+ "loss": 7.254333972930908,
+ "step": 144
+ },
+ {
+ "epoch": 2.0,
+ "eval_loss": 7.318708896636963,
+ "eval_runtime": 59.6555,
+ "eval_samples_per_second": 40.935,
+ "eval_steps_per_second": 1.291,
+ "step": 144
+ },
+ {
+ "epoch": 2.0139737991266378,
+ "grad_norm": 0.019757507368922234,
+ "learning_rate": 0.0006,
+ "loss": 7.30288553237915,
+ "step": 145
+ },
+ {
+ "epoch": 2.027947598253275,
+ "grad_norm": 0.014571248553693295,
+ "learning_rate": 0.0006,
+ "loss": 7.274681568145752,
+ "step": 146
+ },
+ {
+ "epoch": 2.041921397379913,
+ "grad_norm": 0.015744198113679886,
+ "learning_rate": 0.0006,
+ "loss": 7.278632640838623,
+ "step": 147
+ },
+ {
+ "epoch": 2.05589519650655,
+ "grad_norm": 0.028048858046531677,
+ "learning_rate": 0.0006,
+ "loss": 7.26558780670166,
+ "step": 148
+ },
+ {
+ "epoch": 2.069868995633188,
+ "grad_norm": 0.06547307223081589,
+ "learning_rate": 0.0006,
+ "loss": 7.32567834854126,
+ "step": 149
+ },
+ {
+ "epoch": 2.0838427947598253,
+ "grad_norm": 0.05969763547182083,
+ "learning_rate": 0.0006,
+ "loss": 7.395836353302002,
+ "step": 150
+ },
+ {
+ "epoch": 2.097816593886463,
+ "grad_norm": 0.032574281096458435,
+ "learning_rate": 0.0006,
+ "loss": 7.318150997161865,
+ "step": 151
+ },
+ {
+ "epoch": 2.1117903930131003,
+ "grad_norm": 0.029452962800860405,
+ "learning_rate": 0.0006,
+ "loss": 7.308066368103027,
+ "step": 152
+ },
+ {
+ "epoch": 2.125764192139738,
+ "grad_norm": 0.03161991387605667,
+ "learning_rate": 0.0006,
+ "loss": 7.306290626525879,
+ "step": 153
+ },
+ {
+ "epoch": 2.1397379912663754,
+ "grad_norm": 0.027096295729279518,
+ "learning_rate": 0.0006,
+ "loss": 7.297796249389648,
+ "step": 154
+ },
+ {
+ "epoch": 2.153711790393013,
+ "grad_norm": 0.028657056391239166,
+ "learning_rate": 0.0006,
+ "loss": 7.317424774169922,
+ "step": 155
+ },
+ {
+ "epoch": 2.1676855895196505,
+ "grad_norm": 0.02046995982527733,
+ "learning_rate": 0.0006,
+ "loss": 7.297986030578613,
+ "step": 156
+ },
+ {
+ "epoch": 2.1816593886462883,
+ "grad_norm": 0.02220369316637516,
+ "learning_rate": 0.0006,
+ "loss": 7.268237590789795,
+ "step": 157
+ },
+ {
+ "epoch": 2.1956331877729256,
+ "grad_norm": 0.018381357192993164,
+ "learning_rate": 0.0006,
+ "loss": 7.259415626525879,
+ "step": 158
+ },
+ {
+ "epoch": 2.2096069868995634,
+ "grad_norm": 0.0204413291066885,
+ "learning_rate": 0.0006,
+ "loss": 7.258173942565918,
+ "step": 159
+ },
+ {
+ "epoch": 2.2235807860262007,
+ "grad_norm": 0.015874288976192474,
+ "learning_rate": 0.0006,
+ "loss": 7.281834602355957,
+ "step": 160
+ },
+ {
+ "epoch": 2.2375545851528384,
+ "grad_norm": 0.01634068600833416,
+ "learning_rate": 0.0006,
+ "loss": 7.2469868659973145,
+ "step": 161
+ },
+ {
+ "epoch": 2.251528384279476,
+ "grad_norm": 0.01728747971355915,
+ "learning_rate": 0.0006,
+ "loss": 7.280995845794678,
+ "step": 162
+ },
+ {
+ "epoch": 2.2655021834061135,
+ "grad_norm": 0.016341188922524452,
+ "learning_rate": 0.0006,
+ "loss": 7.273357391357422,
+ "step": 163
+ },
+ {
+ "epoch": 2.279475982532751,
+ "grad_norm": 0.016125807538628578,
+ "learning_rate": 0.0006,
+ "loss": 7.273510932922363,
+ "step": 164
+ },
+ {
+ "epoch": 2.2934497816593886,
+ "grad_norm": 0.01629287749528885,
+ "learning_rate": 0.0006,
+ "loss": 7.267029762268066,
+ "step": 165
+ },
+ {
+ "epoch": 2.3074235807860264,
+ "grad_norm": 0.015296131372451782,
+ "learning_rate": 0.0006,
+ "loss": 7.211989402770996,
+ "step": 166
+ },
+ {
+ "epoch": 2.3213973799126637,
+ "grad_norm": 0.01709570363163948,
+ "learning_rate": 0.0006,
+ "loss": 7.242851734161377,
+ "step": 167
+ },
+ {
+ "epoch": 2.3353711790393015,
+ "grad_norm": 0.012552388943731785,
+ "learning_rate": 0.0006,
+ "loss": 7.241696834564209,
+ "step": 168
+ },
+ {
+ "epoch": 2.349344978165939,
+ "grad_norm": 0.014571291394531727,
+ "learning_rate": 0.0006,
+ "loss": 7.270681381225586,
+ "step": 169
+ },
+ {
+ "epoch": 2.3633187772925766,
+ "grad_norm": 0.013389437459409237,
+ "learning_rate": 0.0006,
+ "loss": 7.191600799560547,
+ "step": 170
+ },
+ {
+ "epoch": 2.377292576419214,
+ "grad_norm": 0.011895314790308475,
+ "learning_rate": 0.0006,
+ "loss": 7.213610649108887,
+ "step": 171
+ },
+ {
+ "epoch": 2.3912663755458516,
+ "grad_norm": 0.01345642190426588,
+ "learning_rate": 0.0006,
+ "loss": 7.193787574768066,
+ "step": 172
+ },
+ {
+ "epoch": 2.405240174672489,
+ "grad_norm": 0.0091794328764081,
+ "learning_rate": 0.0006,
+ "loss": 7.233364105224609,
+ "step": 173
+ },
+ {
+ "epoch": 2.4192139737991267,
+ "grad_norm": 0.013767086900770664,
+ "learning_rate": 0.0006,
+ "loss": 7.217272758483887,
+ "step": 174
+ },
+ {
+ "epoch": 2.433187772925764,
+ "grad_norm": 0.015134960412979126,
+ "learning_rate": 0.0006,
+ "loss": 7.192791938781738,
+ "step": 175
+ },
+ {
+ "epoch": 2.447161572052402,
+ "grad_norm": 0.01790648326277733,
+ "learning_rate": 0.0006,
+ "loss": 7.229467391967773,
+ "step": 176
+ },
+ {
+ "epoch": 2.461135371179039,
+ "grad_norm": 0.015362377278506756,
+ "learning_rate": 0.0006,
+ "loss": 7.180346488952637,
+ "step": 177
+ },
+ {
+ "epoch": 2.475109170305677,
+ "grad_norm": 0.010505498386919498,
+ "learning_rate": 0.0006,
+ "loss": 7.188833236694336,
+ "step": 178
+ },
+ {
+ "epoch": 2.489082969432314,
+ "grad_norm": 0.009129747748374939,
+ "learning_rate": 0.0006,
+ "loss": 7.203153133392334,
+ "step": 179
+ },
+ {
+ "epoch": 2.503056768558952,
+ "grad_norm": 0.01147883478552103,
+ "learning_rate": 0.0006,
+ "loss": 7.163824558258057,
+ "step": 180
+ },
+ {
+ "epoch": 2.5170305676855893,
+ "grad_norm": 0.011338942684233189,
+ "learning_rate": 0.0006,
+ "loss": 7.16287899017334,
+ "step": 181
+ },
+ {
+ "epoch": 2.531004366812227,
+ "grad_norm": 0.013277675025165081,
+ "learning_rate": 0.0006,
+ "loss": 7.155905723571777,
+ "step": 182
+ },
+ {
+ "epoch": 2.544978165938865,
+ "grad_norm": 0.015269347466528416,
+ "learning_rate": 0.0006,
+ "loss": 7.189371109008789,
+ "step": 183
+ },
+ {
+ "epoch": 2.558951965065502,
+ "grad_norm": 0.017905596643686295,
+ "learning_rate": 0.0006,
+ "loss": 7.177547931671143,
+ "step": 184
+ },
+ {
+ "epoch": 2.5729257641921395,
+ "grad_norm": 0.024431610479950905,
+ "learning_rate": 0.0006,
+ "loss": 7.192169189453125,
+ "step": 185
+ },
+ {
+ "epoch": 2.5868995633187772,
+ "grad_norm": 0.03656961768865585,
+ "learning_rate": 0.0006,
+ "loss": 7.124387741088867,
+ "step": 186
+ },
+ {
+ "epoch": 2.600873362445415,
+ "grad_norm": 0.04337048903107643,
+ "learning_rate": 0.0006,
+ "loss": 7.140231609344482,
+ "step": 187
+ },
+ {
+ "epoch": 2.6148471615720523,
+ "grad_norm": 0.04047228768467903,
+ "learning_rate": 0.0006,
+ "loss": 7.171927452087402,
+ "step": 188
+ },
+ {
+ "epoch": 2.62882096069869,
+ "grad_norm": 0.07825770974159241,
+ "learning_rate": 0.0006,
+ "loss": 7.207631587982178,
+ "step": 189
+ },
+ {
+ "epoch": 2.6427947598253274,
+ "grad_norm": 0.07118063420057297,
+ "learning_rate": 0.0006,
+ "loss": 7.209488391876221,
+ "step": 190
+ },
+ {
+ "epoch": 2.656768558951965,
+ "grad_norm": 0.05153293535113335,
+ "learning_rate": 0.0006,
+ "loss": 7.14793062210083,
+ "step": 191
+ },
+ {
+ "epoch": 2.6707423580786025,
+ "grad_norm": 0.03318839520215988,
+ "learning_rate": 0.0006,
+ "loss": 7.163010597229004,
+ "step": 192
+ },
+ {
+ "epoch": 2.6847161572052403,
+ "grad_norm": 0.029333539307117462,
+ "learning_rate": 0.0006,
+ "loss": 7.159955978393555,
+ "step": 193
+ },
+ {
+ "epoch": 2.6986899563318776,
+ "grad_norm": 0.029405880719423294,
+ "learning_rate": 0.0006,
+ "loss": 7.136178493499756,
+ "step": 194
+ },
+ {
+ "epoch": 2.7126637554585153,
+ "grad_norm": 0.037815988063812256,
+ "learning_rate": 0.0006,
+ "loss": 7.183186054229736,
+ "step": 195
+ },
+ {
+ "epoch": 2.726637554585153,
+ "grad_norm": 0.020307250320911407,
+ "learning_rate": 0.0006,
+ "loss": 7.133164882659912,
+ "step": 196
+ },
+ {
+ "epoch": 2.7406113537117904,
+ "grad_norm": 0.026048069819808006,
+ "learning_rate": 0.0006,
+ "loss": 7.124977111816406,
+ "step": 197
+ },
+ {
+ "epoch": 2.7545851528384278,
+ "grad_norm": 0.027316724881529808,
+ "learning_rate": 0.0006,
+ "loss": 7.07265567779541,
+ "step": 198
+ },
+ {
+ "epoch": 2.7685589519650655,
+ "grad_norm": 0.021629009395837784,
+ "learning_rate": 0.0006,
+ "loss": 7.150020599365234,
+ "step": 199
+ },
+ {
+ "epoch": 2.7825327510917033,
+ "grad_norm": 0.01605929434299469,
+ "learning_rate": 0.0006,
+ "loss": 7.10919189453125,
+ "step": 200
+ },
+ {
+ "epoch": 2.7965065502183406,
+ "grad_norm": 0.020800089463591576,
+ "learning_rate": 0.0006,
+ "loss": 7.132045269012451,
+ "step": 201
+ },
+ {
+ "epoch": 2.810480349344978,
+ "grad_norm": 0.019225774332880974,
+ "learning_rate": 0.0006,
+ "loss": 7.128364562988281,
+ "step": 202
+ },
+ {
+ "epoch": 2.8244541484716157,
+ "grad_norm": 0.01561372634023428,
+ "learning_rate": 0.0006,
+ "loss": 7.083669662475586,
+ "step": 203
+ },
+ {
+ "epoch": 2.8384279475982535,
+ "grad_norm": 0.021969519555568695,
+ "learning_rate": 0.0006,
+ "loss": 7.050841331481934,
+ "step": 204
+ },
+ {
+ "epoch": 2.8524017467248908,
+ "grad_norm": 0.021453695371747017,
+ "learning_rate": 0.0006,
+ "loss": 7.097085475921631,
+ "step": 205
+ },
+ {
+ "epoch": 2.8663755458515285,
+ "grad_norm": 0.022836022078990936,
+ "learning_rate": 0.0006,
+ "loss": 7.10405969619751,
+ "step": 206
+ },
+ {
+ "epoch": 2.880349344978166,
+ "grad_norm": 0.02430851384997368,
+ "learning_rate": 0.0006,
+ "loss": 7.045181751251221,
+ "step": 207
+ },
+ {
+ "epoch": 2.8943231441048036,
+ "grad_norm": 0.020166993141174316,
+ "learning_rate": 0.0006,
+ "loss": 7.068183898925781,
+ "step": 208
+ },
+ {
+ "epoch": 2.908296943231441,
+ "grad_norm": 0.0118505684658885,
+ "learning_rate": 0.0006,
+ "loss": 7.07132625579834,
+ "step": 209
+ },
+ {
+ "epoch": 2.9222707423580787,
+ "grad_norm": 0.023772427812218666,
+ "learning_rate": 0.0006,
+ "loss": 7.055238723754883,
+ "step": 210
+ },
+ {
+ "epoch": 2.936244541484716,
+ "grad_norm": 0.02350712940096855,
+ "learning_rate": 0.0006,
+ "loss": 7.055898666381836,
+ "step": 211
+ },
+ {
+ "epoch": 2.950218340611354,
+ "grad_norm": 0.017244907096028328,
+ "learning_rate": 0.0006,
+ "loss": 7.0099711418151855,
+ "step": 212
+ },
+ {
+ "epoch": 2.964192139737991,
+ "grad_norm": 0.021565575152635574,
+ "learning_rate": 0.0006,
+ "loss": 6.991751670837402,
+ "step": 213
+ },
+ {
+ "epoch": 2.978165938864629,
+ "grad_norm": 0.03241860866546631,
+ "learning_rate": 0.0006,
+ "loss": 7.005980491638184,
+ "step": 214
+ },
+ {
+ "epoch": 2.992139737991266,
+ "grad_norm": 0.04613726586103439,
+ "learning_rate": 0.0006,
+ "loss": 7.027859687805176,
+ "step": 215
+ },
+ {
+ "epoch": 3.0,
+ "grad_norm": 0.06404894590377808,
+ "learning_rate": 0.0006,
+ "loss": 7.123725891113281,
+ "step": 216
+ },
+ {
+ "epoch": 3.0,
+ "eval_loss": 7.070647716522217,
+ "eval_runtime": 58.5801,
+ "eval_samples_per_second": 41.687,
+ "eval_steps_per_second": 1.314,
+ "step": 216
+ },
+ {
+ "epoch": 3.0139737991266378,
+ "grad_norm": 0.05542432889342308,
+ "learning_rate": 0.0006,
+ "loss": 7.078299522399902,
+ "step": 217
+ },
+ {
+ "epoch": 3.027947598253275,
+ "grad_norm": 0.026484526693820953,
+ "learning_rate": 0.0006,
+ "loss": 6.961159706115723,
+ "step": 218
+ },
+ {
+ "epoch": 3.041921397379913,
+ "grad_norm": 0.028276391327381134,
+ "learning_rate": 0.0006,
+ "loss": 7.019961833953857,
+ "step": 219
+ },
+ {
+ "epoch": 3.05589519650655,
+ "grad_norm": 0.024486595764756203,
+ "learning_rate": 0.0006,
+ "loss": 7.039677619934082,
+ "step": 220
+ },
+ {
+ "epoch": 3.069868995633188,
+ "grad_norm": 0.017620893195271492,
+ "learning_rate": 0.0006,
+ "loss": 6.9270782470703125,
+ "step": 221
+ },
+ {
+ "epoch": 3.0838427947598253,
+ "grad_norm": 0.02547302283346653,
+ "learning_rate": 0.0006,
+ "loss": 6.924233436584473,
+ "step": 222
+ },
+ {
+ "epoch": 3.097816593886463,
+ "grad_norm": 0.02539009042084217,
+ "learning_rate": 0.0006,
+ "loss": 6.973979949951172,
+ "step": 223
+ },
+ {
+ "epoch": 3.1117903930131003,
+ "grad_norm": 0.03256227448582649,
+ "learning_rate": 0.0006,
+ "loss": 6.933725357055664,
+ "step": 224
+ },
+ {
+ "epoch": 3.125764192139738,
+ "grad_norm": 0.04651800915598869,
+ "learning_rate": 0.0006,
+ "loss": 6.951044082641602,
+ "step": 225
+ },
+ {
+ "epoch": 3.1397379912663754,
+ "grad_norm": 0.04822975769639015,
+ "learning_rate": 0.0006,
+ "loss": 6.9614410400390625,
+ "step": 226
+ },
+ {
+ "epoch": 3.153711790393013,
+ "grad_norm": 0.031165020540356636,
+ "learning_rate": 0.0006,
+ "loss": 6.91193151473999,
+ "step": 227
+ },
+ {
+ "epoch": 3.1676855895196505,
+ "grad_norm": 0.029754292219877243,
+ "learning_rate": 0.0006,
+ "loss": 6.91802978515625,
+ "step": 228
+ },
+ {
+ "epoch": 3.1816593886462883,
+ "grad_norm": 0.02193152718245983,
+ "learning_rate": 0.0006,
+ "loss": 6.928090572357178,
+ "step": 229
+ },
+ {
+ "epoch": 3.1956331877729256,
+ "grad_norm": 0.02428470179438591,
+ "learning_rate": 0.0006,
+ "loss": 6.90396785736084,
+ "step": 230
+ },
+ {
+ "epoch": 3.2096069868995634,
+ "grad_norm": 0.023374345153570175,
+ "learning_rate": 0.0006,
+ "loss": 6.885858535766602,
+ "step": 231
+ },
+ {
+ "epoch": 3.2235807860262007,
+ "grad_norm": 0.023405000567436218,
+ "learning_rate": 0.0006,
+ "loss": 6.8855767250061035,
+ "step": 232
+ },
+ {
+ "epoch": 3.2375545851528384,
+ "grad_norm": 0.017277134582400322,
+ "learning_rate": 0.0006,
+ "loss": 6.901449203491211,
+ "step": 233
+ },
+ {
+ "epoch": 3.251528384279476,
+ "grad_norm": 0.019496750086545944,
+ "learning_rate": 0.0006,
+ "loss": 6.885687828063965,
+ "step": 234
+ },
+ {
+ "epoch": 3.2655021834061135,
+ "grad_norm": 0.016893276944756508,
+ "learning_rate": 0.0006,
+ "loss": 6.872369766235352,
+ "step": 235
+ },
+ {
+ "epoch": 3.279475982532751,
+ "grad_norm": 0.02058715932071209,
+ "learning_rate": 0.0006,
+ "loss": 6.823280334472656,
+ "step": 236
+ },
+ {
+ "epoch": 3.2934497816593886,
+ "grad_norm": 0.015530755743384361,
+ "learning_rate": 0.0006,
+ "loss": 6.860499858856201,
+ "step": 237
+ },
+ {
+ "epoch": 3.3074235807860264,
+ "grad_norm": 0.01934926211833954,
+ "learning_rate": 0.0006,
+ "loss": 6.817448616027832,
+ "step": 238
+ },
+ {
+ "epoch": 3.3213973799126637,
+ "grad_norm": 0.017413552850484848,
+ "learning_rate": 0.0006,
+ "loss": 6.881955146789551,
+ "step": 239
+ },
+ {
+ "epoch": 3.3353711790393015,
+ "grad_norm": 0.026478875428438187,
+ "learning_rate": 0.0006,
+ "loss": 6.827404499053955,
+ "step": 240
+ },
+ {
+ "epoch": 3.349344978165939,
+ "grad_norm": 0.032611701637506485,
+ "learning_rate": 0.0006,
+ "loss": 6.78285551071167,
+ "step": 241
+ },
+ {
+ "epoch": 3.3633187772925766,
+ "grad_norm": 0.041900306940078735,
+ "learning_rate": 0.0006,
+ "loss": 6.783053874969482,
+ "step": 242
+ },
+ {
+ "epoch": 3.377292576419214,
+ "grad_norm": 0.04149497672915459,
+ "learning_rate": 0.0006,
+ "loss": 6.745220184326172,
+ "step": 243
+ },
+ {
+ "epoch": 3.3912663755458516,
+ "grad_norm": 0.023949826136231422,
+ "learning_rate": 0.0006,
+ "loss": 6.841533184051514,
+ "step": 244
+ },
+ {
+ "epoch": 3.405240174672489,
+ "grad_norm": 0.028601842001080513,
+ "learning_rate": 0.0006,
+ "loss": 6.852852821350098,
+ "step": 245
+ },
+ {
+ "epoch": 3.4192139737991267,
+ "grad_norm": 0.027365155518054962,
+ "learning_rate": 0.0006,
+ "loss": 6.74376106262207,
+ "step": 246
+ },
+ {
+ "epoch": 3.433187772925764,
+ "grad_norm": 0.02856568619608879,
+ "learning_rate": 0.0006,
+ "loss": 6.74906063079834,
+ "step": 247
+ },
+ {
+ "epoch": 3.447161572052402,
+ "grad_norm": 0.019172416999936104,
+ "learning_rate": 0.0006,
+ "loss": 6.809206008911133,
+ "step": 248
+ },
+ {
+ "epoch": 3.461135371179039,
+ "grad_norm": 0.02161746844649315,
+ "learning_rate": 0.0006,
+ "loss": 6.789888381958008,
+ "step": 249
+ },
+ {
+ "epoch": 3.475109170305677,
+ "grad_norm": 0.020273666828870773,
+ "learning_rate": 0.0006,
+ "loss": 6.73512077331543,
+ "step": 250
+ },
+ {
+ "epoch": 3.489082969432314,
+ "grad_norm": 0.022329283878207207,
+ "learning_rate": 0.0006,
+ "loss": 6.826148509979248,
+ "step": 251
+ },
+ {
+ "epoch": 3.503056768558952,
+ "grad_norm": 0.02284284122288227,
+ "learning_rate": 0.0006,
+ "loss": 6.748521327972412,
+ "step": 252
+ },
+ {
+ "epoch": 3.5170305676855893,
+ "grad_norm": 0.026368247345089912,
+ "learning_rate": 0.0006,
+ "loss": 6.645879745483398,
+ "step": 253
+ },
+ {
+ "epoch": 3.531004366812227,
+ "grad_norm": 0.03627244383096695,
+ "learning_rate": 0.0006,
+ "loss": 6.666064739227295,
+ "step": 254
+ },
+ {
+ "epoch": 3.544978165938865,
+ "grad_norm": 0.04140935093164444,
+ "learning_rate": 0.0006,
+ "loss": 6.712275981903076,
+ "step": 255
+ },
+ {
+ "epoch": 3.558951965065502,
+ "grad_norm": 0.0390462800860405,
+ "learning_rate": 0.0006,
+ "loss": 6.665900230407715,
+ "step": 256
+ },
+ {
+ "epoch": 3.5729257641921395,
+ "grad_norm": 0.05178583785891533,
+ "learning_rate": 0.0006,
+ "loss": 6.753298759460449,
+ "step": 257
+ },
+ {
+ "epoch": 3.5868995633187772,
+ "grad_norm": 0.054328061640262604,
+ "learning_rate": 0.0006,
+ "loss": 6.732633113861084,
+ "step": 258
+ },
+ {
+ "epoch": 3.600873362445415,
+ "grad_norm": 0.04590460658073425,
+ "learning_rate": 0.0006,
+ "loss": 6.65250301361084,
+ "step": 259
+ },
+ {
+ "epoch": 3.6148471615720523,
+ "grad_norm": 0.044331666082143784,
+ "learning_rate": 0.0006,
+ "loss": 6.68654203414917,
+ "step": 260
+ },
+ {
+ "epoch": 3.62882096069869,
+ "grad_norm": 0.05308730527758598,
+ "learning_rate": 0.0006,
+ "loss": 6.642482757568359,
+ "step": 261
+ },
+ {
+ "epoch": 3.6427947598253274,
+ "grad_norm": 0.03999016433954239,
+ "learning_rate": 0.0006,
+ "loss": 6.625227451324463,
+ "step": 262
+ },
+ {
+ "epoch": 3.656768558951965,
+ "grad_norm": 0.04088086634874344,
+ "learning_rate": 0.0006,
+ "loss": 6.67623233795166,
+ "step": 263
+ },
+ {
+ "epoch": 3.6707423580786025,
+ "grad_norm": 0.03615636005997658,
+ "learning_rate": 0.0006,
+ "loss": 6.6869001388549805,
+ "step": 264
+ },
+ {
+ "epoch": 3.6847161572052403,
+ "grad_norm": 0.02817360684275627,
+ "learning_rate": 0.0006,
+ "loss": 6.674601078033447,
+ "step": 265
+ },
+ {
+ "epoch": 3.6986899563318776,
+ "grad_norm": 0.032586123794317245,
+ "learning_rate": 0.0006,
+ "loss": 6.612994194030762,
+ "step": 266
+ },
+ {
+ "epoch": 3.7126637554585153,
+ "grad_norm": 0.031117867678403854,
+ "learning_rate": 0.0006,
+ "loss": 6.6244354248046875,
+ "step": 267
+ },
+ {
+ "epoch": 3.726637554585153,
+ "grad_norm": 0.028330687433481216,
+ "learning_rate": 0.0006,
+ "loss": 6.669557094573975,
+ "step": 268
+ },
+ {
+ "epoch": 3.7406113537117904,
+ "grad_norm": 0.021489202976226807,
+ "learning_rate": 0.0006,
+ "loss": 6.632035732269287,
+ "step": 269
+ },
+ {
+ "epoch": 3.7545851528384278,
+ "grad_norm": 0.024269424378871918,
+ "learning_rate": 0.0006,
+ "loss": 6.660867691040039,
+ "step": 270
+ },
+ {
+ "epoch": 3.7685589519650655,
+ "grad_norm": 0.02237873338162899,
+ "learning_rate": 0.0006,
+ "loss": 6.602587699890137,
+ "step": 271
+ },
+ {
+ "epoch": 3.7825327510917033,
+ "grad_norm": 0.026503929868340492,
+ "learning_rate": 0.0006,
+ "loss": 6.599200248718262,
+ "step": 272
+ },
+ {
+ "epoch": 3.7965065502183406,
+ "grad_norm": 0.03687124326825142,
+ "learning_rate": 0.0006,
+ "loss": 6.567296504974365,
+ "step": 273
+ },
+ {
+ "epoch": 3.810480349344978,
+ "grad_norm": 0.04708952456712723,
+ "learning_rate": 0.0006,
+ "loss": 6.621265411376953,
+ "step": 274
+ },
+ {
+ "epoch": 3.8244541484716157,
+ "grad_norm": 0.041846614331007004,
+ "learning_rate": 0.0006,
+ "loss": 6.550536155700684,
+ "step": 275
+ },
+ {
+ "epoch": 3.8384279475982535,
+ "grad_norm": 0.031249675899744034,
+ "learning_rate": 0.0006,
+ "loss": 6.54278564453125,
+ "step": 276
+ },
+ {
+ "epoch": 3.8524017467248908,
+ "grad_norm": 0.023082256317138672,
+ "learning_rate": 0.0006,
+ "loss": 6.544787406921387,
+ "step": 277
+ },
+ {
+ "epoch": 3.8663755458515285,
+ "grad_norm": 0.020995570346713066,
+ "learning_rate": 0.0006,
+ "loss": 6.518143653869629,
+ "step": 278
+ },
+ {
+ "epoch": 3.880349344978166,
+ "grad_norm": 0.020839324221014977,
+ "learning_rate": 0.0006,
+ "loss": 6.546667098999023,
+ "step": 279
+ },
+ {
+ "epoch": 3.8943231441048036,
+ "grad_norm": 0.018377432599663734,
+ "learning_rate": 0.0006,
+ "loss": 6.478307723999023,
+ "step": 280
+ },
+ {
+ "epoch": 3.908296943231441,
+ "grad_norm": 0.021713724359869957,
+ "learning_rate": 0.0006,
+ "loss": 6.485574245452881,
+ "step": 281
+ },
+ {
+ "epoch": 3.9222707423580787,
+ "grad_norm": 0.01359301246702671,
+ "learning_rate": 0.0006,
+ "loss": 6.570268630981445,
+ "step": 282
+ },
+ {
+ "epoch": 3.936244541484716,
+ "grad_norm": 0.016233332455158234,
+ "learning_rate": 0.0006,
+ "loss": 6.57150936126709,
+ "step": 283
+ },
+ {
+ "epoch": 3.950218340611354,
+ "grad_norm": 0.020687608048319817,
+ "learning_rate": 0.0006,
+ "loss": 6.527956008911133,
+ "step": 284
+ },
+ {
+ "epoch": 3.964192139737991,
+ "grad_norm": 0.02032964862883091,
+ "learning_rate": 0.0006,
+ "loss": 6.468169689178467,
+ "step": 285
+ },
+ {
+ "epoch": 3.978165938864629,
+ "grad_norm": 0.016676288098096848,
+ "learning_rate": 0.0006,
+ "loss": 6.448418617248535,
+ "step": 286
+ },
+ {
+ "epoch": 3.992139737991266,
+ "grad_norm": 0.0168539360165596,
+ "learning_rate": 0.0006,
+ "loss": 6.461180686950684,
+ "step": 287
+ },
+ {
+ "epoch": 4.0,
+ "grad_norm": 0.02666233666241169,
+ "learning_rate": 0.0006,
+ "loss": 6.344893932342529,
+ "step": 288
+ },
+ {
+ "epoch": 4.0,
+ "eval_loss": 6.5384979248046875,
+ "eval_runtime": 58.814,
+ "eval_samples_per_second": 41.521,
+ "eval_steps_per_second": 1.309,
+ "step": 288
+ },
+ {
+ "epoch": 4.013973799126638,
+ "grad_norm": 0.04154708981513977,
+ "learning_rate": 0.0006,
+ "loss": 6.41855525970459,
+ "step": 289
+ },
+ {
+ "epoch": 4.0279475982532755,
+ "grad_norm": 0.04895668104290962,
+ "learning_rate": 0.0006,
+ "loss": 6.499302864074707,
+ "step": 290
+ },
+ {
+ "epoch": 4.041921397379912,
+ "grad_norm": 0.034464482218027115,
+ "learning_rate": 0.0006,
+ "loss": 6.438611030578613,
+ "step": 291
+ },
+ {
+ "epoch": 4.05589519650655,
+ "grad_norm": 0.06845773756504059,
+ "learning_rate": 0.0006,
+ "loss": 6.538361072540283,
+ "step": 292
+ },
+ {
+ "epoch": 4.069868995633188,
+ "grad_norm": 0.053768858313560486,
+ "learning_rate": 0.0006,
+ "loss": 6.527524948120117,
+ "step": 293
+ },
+ {
+ "epoch": 4.083842794759826,
+ "grad_norm": 0.039037931710481644,
+ "learning_rate": 0.0006,
+ "loss": 6.503847122192383,
+ "step": 294
+ },
+ {
+ "epoch": 4.097816593886463,
+ "grad_norm": 0.03962196782231331,
+ "learning_rate": 0.0006,
+ "loss": 6.475178241729736,
+ "step": 295
+ },
+ {
+ "epoch": 4.1117903930131,
+ "grad_norm": 0.049101535230875015,
+ "learning_rate": 0.0006,
+ "loss": 6.474587917327881,
+ "step": 296
+ },
+ {
+ "epoch": 4.125764192139738,
+ "grad_norm": 0.02830282226204872,
+ "learning_rate": 0.0006,
+ "loss": 6.45065975189209,
+ "step": 297
+ },
+ {
+ "epoch": 4.139737991266376,
+ "grad_norm": 0.023153482005000114,
+ "learning_rate": 0.0006,
+ "loss": 6.5218706130981445,
+ "step": 298
+ },
+ {
+ "epoch": 4.153711790393013,
+ "grad_norm": 0.025064002722501755,
+ "learning_rate": 0.0006,
+ "loss": 6.456475257873535,
+ "step": 299
+ },
+ {
+ "epoch": 4.1676855895196505,
+ "grad_norm": 0.02357092685997486,
+ "learning_rate": 0.0006,
+ "loss": 6.485952377319336,
+ "step": 300
+ },
+ {
+ "epoch": 4.181659388646288,
+ "grad_norm": 0.021832305938005447,
+ "learning_rate": 0.0006,
+ "loss": 6.425506591796875,
+ "step": 301
+ },
+ {
+ "epoch": 4.195633187772926,
+ "grad_norm": 0.0227900929749012,
+ "learning_rate": 0.0006,
+ "loss": 6.439155578613281,
+ "step": 302
+ },
+ {
+ "epoch": 4.209606986899563,
+ "grad_norm": 0.015961607918143272,
+ "learning_rate": 0.0006,
+ "loss": 6.460862636566162,
+ "step": 303
+ },
+ {
+ "epoch": 4.223580786026201,
+ "grad_norm": 0.017876390367746353,
+ "learning_rate": 0.0006,
+ "loss": 6.44564151763916,
+ "step": 304
+ },
+ {
+ "epoch": 4.2375545851528384,
+ "grad_norm": 0.013703204691410065,
+ "learning_rate": 0.0006,
+ "loss": 6.418414115905762,
+ "step": 305
+ },
+ {
+ "epoch": 4.251528384279476,
+ "grad_norm": 0.015240306034684181,
+ "learning_rate": 0.0006,
+ "loss": 6.411238670349121,
+ "step": 306
+ },
+ {
+ "epoch": 4.265502183406113,
+ "grad_norm": 0.01587662845849991,
+ "learning_rate": 0.0006,
+ "loss": 6.395642280578613,
+ "step": 307
+ },
+ {
+ "epoch": 4.279475982532751,
+ "grad_norm": 0.014650700613856316,
+ "learning_rate": 0.0006,
+ "loss": 6.302793025970459,
+ "step": 308
+ },
+ {
+ "epoch": 4.293449781659389,
+ "grad_norm": 0.016533225774765015,
+ "learning_rate": 0.0006,
+ "loss": 6.380588054656982,
+ "step": 309
+ },
+ {
+ "epoch": 4.307423580786026,
+ "grad_norm": 0.019963741302490234,
+ "learning_rate": 0.0006,
+ "loss": 6.345336437225342,
+ "step": 310
+ },
+ {
+ "epoch": 4.321397379912664,
+ "grad_norm": 0.02237936295568943,
+ "learning_rate": 0.0006,
+ "loss": 6.387770652770996,
+ "step": 311
+ },
+ {
+ "epoch": 4.335371179039301,
+ "grad_norm": 0.026658328250050545,
+ "learning_rate": 0.0006,
+ "loss": 6.323662757873535,
+ "step": 312
+ },
+ {
+ "epoch": 4.349344978165939,
+ "grad_norm": 0.02852284163236618,
+ "learning_rate": 0.0006,
+ "loss": 6.382204055786133,
+ "step": 313
+ },
+ {
+ "epoch": 4.3633187772925766,
+ "grad_norm": 0.028815090656280518,
+ "learning_rate": 0.0006,
+ "loss": 6.402390480041504,
+ "step": 314
+ },
+ {
+ "epoch": 4.377292576419214,
+ "grad_norm": 0.029393529519438744,
+ "learning_rate": 0.0006,
+ "loss": 6.280278205871582,
+ "step": 315
+ },
+ {
+ "epoch": 4.391266375545851,
+ "grad_norm": 0.02186041697859764,
+ "learning_rate": 0.0006,
+ "loss": 6.302707672119141,
+ "step": 316
+ },
+ {
+ "epoch": 4.405240174672489,
+ "grad_norm": 0.02071218006312847,
+ "learning_rate": 0.0006,
+ "loss": 6.3365020751953125,
+ "step": 317
+ },
+ {
+ "epoch": 4.419213973799127,
+ "grad_norm": 0.02779117226600647,
+ "learning_rate": 0.0006,
+ "loss": 6.252157211303711,
+ "step": 318
+ },
+ {
+ "epoch": 4.4331877729257645,
+ "grad_norm": 0.03470654785633087,
+ "learning_rate": 0.0006,
+ "loss": 6.319070339202881,
+ "step": 319
+ },
+ {
+ "epoch": 4.447161572052401,
+ "grad_norm": 0.037067804485559464,
+ "learning_rate": 0.0006,
+ "loss": 6.314116477966309,
+ "step": 320
+ },
+ {
+ "epoch": 4.461135371179039,
+ "grad_norm": 0.032182756811380386,
+ "learning_rate": 0.0006,
+ "loss": 6.267298698425293,
+ "step": 321
+ },
+ {
+ "epoch": 4.475109170305677,
+ "grad_norm": 0.026305217295885086,
+ "learning_rate": 0.0006,
+ "loss": 6.331688404083252,
+ "step": 322
+ },
+ {
+ "epoch": 4.489082969432315,
+ "grad_norm": 0.027228357270359993,
+ "learning_rate": 0.0006,
+ "loss": 6.31139612197876,
+ "step": 323
+ },
+ {
+ "epoch": 4.503056768558952,
+ "grad_norm": 0.025781169533729553,
+ "learning_rate": 0.0006,
+ "loss": 6.194684982299805,
+ "step": 324
+ },
+ {
+ "epoch": 4.517030567685589,
+ "grad_norm": 0.029186977073550224,
+ "learning_rate": 0.0006,
+ "loss": 6.31306266784668,
+ "step": 325
+ },
+ {
+ "epoch": 4.531004366812227,
+ "grad_norm": 0.023608777672052383,
+ "learning_rate": 0.0006,
+ "loss": 6.285243034362793,
+ "step": 326
+ },
+ {
+ "epoch": 4.544978165938865,
+ "grad_norm": 0.02267594076693058,
+ "learning_rate": 0.0006,
+ "loss": 6.30342960357666,
+ "step": 327
+ },
+ {
+ "epoch": 4.558951965065502,
+ "grad_norm": 0.01926310732960701,
+ "learning_rate": 0.0006,
+ "loss": 6.280606269836426,
+ "step": 328
+ },
+ {
+ "epoch": 4.5729257641921395,
+ "grad_norm": 0.026484837755560875,
+ "learning_rate": 0.0006,
+ "loss": 6.237980365753174,
+ "step": 329
+ },
+ {
+ "epoch": 4.586899563318777,
+ "grad_norm": 0.026228854432702065,
+ "learning_rate": 0.0006,
+ "loss": 6.3018293380737305,
+ "step": 330
+ },
+ {
+ "epoch": 4.600873362445415,
+ "grad_norm": 0.022096967324614525,
+ "learning_rate": 0.0006,
+ "loss": 6.253863334655762,
+ "step": 331
+ },
+ {
+ "epoch": 4.614847161572053,
+ "grad_norm": 0.027223890647292137,
+ "learning_rate": 0.0006,
+ "loss": 6.166882038116455,
+ "step": 332
+ },
+ {
+ "epoch": 4.62882096069869,
+ "grad_norm": 0.03160124272108078,
+ "learning_rate": 0.0006,
+ "loss": 6.230309963226318,
+ "step": 333
+ },
+ {
+ "epoch": 4.642794759825327,
+ "grad_norm": 0.04110539332032204,
+ "learning_rate": 0.0006,
+ "loss": 6.2334818840026855,
+ "step": 334
+ },
+ {
+ "epoch": 4.656768558951965,
+ "grad_norm": 0.04355933889746666,
+ "learning_rate": 0.0006,
+ "loss": 6.160956382751465,
+ "step": 335
+ },
+ {
+ "epoch": 4.670742358078603,
+ "grad_norm": 0.03611086308956146,
+ "learning_rate": 0.0006,
+ "loss": 6.23399543762207,
+ "step": 336
+ },
+ {
+ "epoch": 4.68471615720524,
+ "grad_norm": 0.051811400800943375,
+ "learning_rate": 0.0006,
+ "loss": 6.241855144500732,
+ "step": 337
+ },
+ {
+ "epoch": 4.698689956331878,
+ "grad_norm": 0.057434193789958954,
+ "learning_rate": 0.0006,
+ "loss": 6.220993995666504,
+ "step": 338
+ },
+ {
+ "epoch": 4.712663755458515,
+ "grad_norm": 0.04420953616499901,
+ "learning_rate": 0.0006,
+ "loss": 6.300461292266846,
+ "step": 339
+ },
+ {
+ "epoch": 4.726637554585153,
+ "grad_norm": 0.03515457361936569,
+ "learning_rate": 0.0006,
+ "loss": 6.269349098205566,
+ "step": 340
+ },
+ {
+ "epoch": 4.74061135371179,
+ "grad_norm": 0.03222600743174553,
+ "learning_rate": 0.0006,
+ "loss": 6.248613357543945,
+ "step": 341
+ },
+ {
+ "epoch": 4.754585152838428,
+ "grad_norm": 0.0384046845138073,
+ "learning_rate": 0.0006,
+ "loss": 6.266979217529297,
+ "step": 342
+ },
+ {
+ "epoch": 4.7685589519650655,
+ "grad_norm": 0.0514867827296257,
+ "learning_rate": 0.0006,
+ "loss": 6.234103202819824,
+ "step": 343
+ },
+ {
+ "epoch": 4.782532751091703,
+ "grad_norm": 0.050747793167829514,
+ "learning_rate": 0.0006,
+ "loss": 6.322582244873047,
+ "step": 344
+ },
+ {
+ "epoch": 4.796506550218341,
+ "grad_norm": 0.040878988802433014,
+ "learning_rate": 0.0006,
+ "loss": 6.212509632110596,
+ "step": 345
+ },
+ {
+ "epoch": 4.810480349344978,
+ "grad_norm": 0.034655820578336716,
+ "learning_rate": 0.0006,
+ "loss": 6.265879154205322,
+ "step": 346
+ },
+ {
+ "epoch": 4.824454148471616,
+ "grad_norm": 0.04202224686741829,
+ "learning_rate": 0.0006,
+ "loss": 6.2896409034729,
+ "step": 347
+ },
+ {
+ "epoch": 4.8384279475982535,
+ "grad_norm": 0.03632277995347977,
+ "learning_rate": 0.0006,
+ "loss": 6.253917217254639,
+ "step": 348
+ },
+ {
+ "epoch": 4.85240174672489,
+ "grad_norm": 0.028645765036344528,
+ "learning_rate": 0.0006,
+ "loss": 6.207965850830078,
+ "step": 349
+ },
+ {
+ "epoch": 4.866375545851528,
+ "grad_norm": 0.025305170565843582,
+ "learning_rate": 0.0006,
+ "loss": 6.168068885803223,
+ "step": 350
+ },
+ {
+ "epoch": 4.880349344978166,
+ "grad_norm": 0.022137803956866264,
+ "learning_rate": 0.0006,
+ "loss": 6.235283374786377,
+ "step": 351
+ },
+ {
+ "epoch": 4.894323144104804,
+ "grad_norm": 0.01814538985490799,
+ "learning_rate": 0.0006,
+ "loss": 6.259641647338867,
+ "step": 352
+ },
+ {
+ "epoch": 4.908296943231441,
+ "grad_norm": 0.018465086817741394,
+ "learning_rate": 0.0006,
+ "loss": 6.111738681793213,
+ "step": 353
+ },
+ {
+ "epoch": 4.922270742358078,
+ "grad_norm": 0.018638933077454567,
+ "learning_rate": 0.0006,
+ "loss": 6.225642204284668,
+ "step": 354
+ },
+ {
+ "epoch": 4.936244541484716,
+ "grad_norm": 0.016461260616779327,
+ "learning_rate": 0.0006,
+ "loss": 6.140647888183594,
+ "step": 355
+ },
+ {
+ "epoch": 4.950218340611354,
+ "grad_norm": 0.015458385460078716,
+ "learning_rate": 0.0006,
+ "loss": 6.155224800109863,
+ "step": 356
+ },
+ {
+ "epoch": 4.964192139737992,
+ "grad_norm": 0.014934048056602478,
+ "learning_rate": 0.0006,
+ "loss": 6.101555824279785,
+ "step": 357
+ },
+ {
+ "epoch": 4.978165938864628,
+ "grad_norm": 0.013979545794427395,
+ "learning_rate": 0.0006,
+ "loss": 6.161718845367432,
+ "step": 358
+ },
+ {
+ "epoch": 4.992139737991266,
+ "grad_norm": 0.01640394888818264,
+ "learning_rate": 0.0006,
+ "loss": 6.055768966674805,
+ "step": 359
+ },
+ {
+ "epoch": 5.0,
+ "grad_norm": 0.015135680325329304,
+ "learning_rate": 0.0006,
+ "loss": 6.169695854187012,
+ "step": 360
+ },
+ {
+ "epoch": 5.0,
+ "eval_loss": 6.16408109664917,
+ "eval_runtime": 59.3649,
+ "eval_samples_per_second": 41.135,
+ "eval_steps_per_second": 1.297,
+ "step": 360
+ },
+ {
+ "epoch": 5.013973799126638,
+ "grad_norm": 0.012849048711359501,
+ "learning_rate": 0.0006,
+ "loss": 6.055830001831055,
+ "step": 361
+ },
+ {
+ "epoch": 5.0279475982532755,
+ "grad_norm": 0.018169576302170753,
+ "learning_rate": 0.0006,
+ "loss": 6.082655429840088,
+ "step": 362
+ },
+ {
+ "epoch": 5.041921397379912,
+ "grad_norm": 0.024293430149555206,
+ "learning_rate": 0.0006,
+ "loss": 6.109032154083252,
+ "step": 363
+ },
+ {
+ "epoch": 5.05589519650655,
+ "grad_norm": 0.03266787901520729,
+ "learning_rate": 0.0006,
+ "loss": 6.030613899230957,
+ "step": 364
+ },
+ {
+ "epoch": 5.069868995633188,
+ "grad_norm": 0.03572266176342964,
+ "learning_rate": 0.0006,
+ "loss": 6.050390720367432,
+ "step": 365
+ },
+ {
+ "epoch": 5.083842794759826,
+ "grad_norm": 0.032625213265419006,
+ "learning_rate": 0.0006,
+ "loss": 6.169313430786133,
+ "step": 366
+ },
+ {
+ "epoch": 5.097816593886463,
+ "grad_norm": 0.04051872715353966,
+ "learning_rate": 0.0006,
+ "loss": 6.10398006439209,
+ "step": 367
+ },
+ {
+ "epoch": 5.1117903930131,
+ "grad_norm": 0.0351213738322258,
+ "learning_rate": 0.0006,
+ "loss": 6.121973514556885,
+ "step": 368
+ },
+ {
+ "epoch": 5.125764192139738,
+ "grad_norm": 0.03523759916424751,
+ "learning_rate": 0.0006,
+ "loss": 6.0626654624938965,
+ "step": 369
+ },
+ {
+ "epoch": 5.139737991266376,
+ "grad_norm": 0.02706182189285755,
+ "learning_rate": 0.0006,
+ "loss": 6.049354076385498,
+ "step": 370
+ },
+ {
+ "epoch": 5.153711790393013,
+ "grad_norm": 0.024212589487433434,
+ "learning_rate": 0.0006,
+ "loss": 6.034826278686523,
+ "step": 371
+ },
+ {
+ "epoch": 5.1676855895196505,
+ "grad_norm": 0.024798082187771797,
+ "learning_rate": 0.0006,
+ "loss": 6.004058837890625,
+ "step": 372
+ },
+ {
+ "epoch": 5.181659388646288,
+ "grad_norm": 0.02325567416846752,
+ "learning_rate": 0.0006,
+ "loss": 5.986461639404297,
+ "step": 373
+ },
+ {
+ "epoch": 5.195633187772926,
+ "grad_norm": 0.019060570746660233,
+ "learning_rate": 0.0006,
+ "loss": 6.059736251831055,
+ "step": 374
+ },
+ {
+ "epoch": 5.209606986899563,
+ "grad_norm": 0.016822976991534233,
+ "learning_rate": 0.0006,
+ "loss": 6.072957515716553,
+ "step": 375
+ },
+ {
+ "epoch": 5.223580786026201,
+ "grad_norm": 0.017218658700585365,
+ "learning_rate": 0.0006,
+ "loss": 6.031739234924316,
+ "step": 376
+ },
+ {
+ "epoch": 5.2375545851528384,
+ "grad_norm": 0.014889383688569069,
+ "learning_rate": 0.0006,
+ "loss": 6.025674343109131,
+ "step": 377
+ },
+ {
+ "epoch": 5.251528384279476,
+ "grad_norm": 0.01708107627928257,
+ "learning_rate": 0.0006,
+ "loss": 6.090451240539551,
+ "step": 378
+ },
+ {
+ "epoch": 5.265502183406113,
+ "grad_norm": 0.020260317251086235,
+ "learning_rate": 0.0006,
+ "loss": 6.094257354736328,
+ "step": 379
+ },
+ {
+ "epoch": 5.279475982532751,
+ "grad_norm": 0.020110400393605232,
+ "learning_rate": 0.0006,
+ "loss": 6.021188735961914,
+ "step": 380
+ },
+ {
+ "epoch": 5.293449781659389,
+ "grad_norm": 0.02027830481529236,
+ "learning_rate": 0.0006,
+ "loss": 6.022156238555908,
+ "step": 381
+ },
+ {
+ "epoch": 5.307423580786026,
+ "grad_norm": 0.022747062146663666,
+ "learning_rate": 0.0006,
+ "loss": 6.011836051940918,
+ "step": 382
+ },
+ {
+ "epoch": 5.321397379912664,
+ "grad_norm": 0.01990230567753315,
+ "learning_rate": 0.0006,
+ "loss": 6.003054618835449,
+ "step": 383
+ },
+ {
+ "epoch": 5.335371179039301,
+ "grad_norm": 0.01596238650381565,
+ "learning_rate": 0.0006,
+ "loss": 5.991410732269287,
+ "step": 384
+ },
+ {
+ "epoch": 5.349344978165939,
+ "grad_norm": 0.015847105532884598,
+ "learning_rate": 0.0006,
+ "loss": 6.100622177124023,
+ "step": 385
+ },
+ {
+ "epoch": 5.3633187772925766,
+ "grad_norm": 0.016179397702217102,
+ "learning_rate": 0.0006,
+ "loss": 6.032659530639648,
+ "step": 386
+ },
+ {
+ "epoch": 5.377292576419214,
+ "grad_norm": 0.018256602808833122,
+ "learning_rate": 0.0006,
+ "loss": 5.961983680725098,
+ "step": 387
+ },
+ {
+ "epoch": 5.391266375545851,
+ "grad_norm": 0.02805912122130394,
+ "learning_rate": 0.0006,
+ "loss": 5.983541965484619,
+ "step": 388
+ },
+ {
+ "epoch": 5.405240174672489,
+ "grad_norm": 0.039082255214452744,
+ "learning_rate": 0.0006,
+ "loss": 6.021474361419678,
+ "step": 389
+ },
+ {
+ "epoch": 5.419213973799127,
+ "grad_norm": 0.036757659167051315,
+ "learning_rate": 0.0006,
+ "loss": 6.005046367645264,
+ "step": 390
+ },
+ {
+ "epoch": 5.4331877729257645,
+ "grad_norm": 0.035277366638183594,
+ "learning_rate": 0.0006,
+ "loss": 6.036296844482422,
+ "step": 391
+ },
+ {
+ "epoch": 5.447161572052401,
+ "grad_norm": 0.034404635429382324,
+ "learning_rate": 0.0006,
+ "loss": 5.994539260864258,
+ "step": 392
+ },
+ {
+ "epoch": 5.461135371179039,
+ "grad_norm": 0.0377206988632679,
+ "learning_rate": 0.0006,
+ "loss": 5.941534519195557,
+ "step": 393
+ },
+ {
+ "epoch": 5.475109170305677,
+ "grad_norm": 0.0389922633767128,
+ "learning_rate": 0.0006,
+ "loss": 5.983644485473633,
+ "step": 394
+ },
+ {
+ "epoch": 5.489082969432315,
+ "grad_norm": 0.04176778718829155,
+ "learning_rate": 0.0006,
+ "loss": 6.030922889709473,
+ "step": 395
+ },
+ {
+ "epoch": 5.503056768558952,
+ "grad_norm": 0.0337153784930706,
+ "learning_rate": 0.0006,
+ "loss": 5.961367130279541,
+ "step": 396
+ },
+ {
+ "epoch": 5.517030567685589,
+ "grad_norm": 0.036116816103458405,
+ "learning_rate": 0.0006,
+ "loss": 5.819280624389648,
+ "step": 397
+ },
+ {
+ "epoch": 5.531004366812227,
+ "grad_norm": 0.030724041163921356,
+ "learning_rate": 0.0006,
+ "loss": 5.906380653381348,
+ "step": 398
+ },
+ {
+ "epoch": 5.544978165938865,
+ "grad_norm": 0.030264202505350113,
+ "learning_rate": 0.0006,
+ "loss": 5.975833415985107,
+ "step": 399
+ },
+ {
+ "epoch": 5.558951965065502,
+ "grad_norm": 0.032096318900585175,
+ "learning_rate": 0.0006,
+ "loss": 6.001348972320557,
+ "step": 400
+ },
+ {
+ "epoch": 5.5729257641921395,
+ "grad_norm": 0.030579356476664543,
+ "learning_rate": 0.0006,
+ "loss": 5.928768634796143,
+ "step": 401
+ },
+ {
+ "epoch": 5.586899563318777,
+ "grad_norm": 0.028241973370313644,
+ "learning_rate": 0.0006,
+ "loss": 5.92582893371582,
+ "step": 402
+ },
+ {
+ "epoch": 5.600873362445415,
+ "grad_norm": 0.023086953908205032,
+ "learning_rate": 0.0006,
+ "loss": 5.892926216125488,
+ "step": 403
+ },
+ {
+ "epoch": 5.614847161572053,
+ "grad_norm": 0.02547992393374443,
+ "learning_rate": 0.0006,
+ "loss": 5.903195381164551,
+ "step": 404
+ },
+ {
+ "epoch": 5.62882096069869,
+ "grad_norm": 0.023089831694960594,
+ "learning_rate": 0.0006,
+ "loss": 5.912033557891846,
+ "step": 405
+ },
+ {
+ "epoch": 5.642794759825327,
+ "grad_norm": 0.025446368381381035,
+ "learning_rate": 0.0006,
+ "loss": 5.926138401031494,
+ "step": 406
+ },
+ {
+ "epoch": 5.656768558951965,
+ "grad_norm": 0.03748747706413269,
+ "learning_rate": 0.0006,
+ "loss": 5.913451194763184,
+ "step": 407
+ },
+ {
+ "epoch": 5.670742358078603,
+ "grad_norm": 0.03493810072541237,
+ "learning_rate": 0.0006,
+ "loss": 5.97633695602417,
+ "step": 408
+ },
+ {
+ "epoch": 5.68471615720524,
+ "grad_norm": 0.023131275549530983,
+ "learning_rate": 0.0006,
+ "loss": 5.899571418762207,
+ "step": 409
+ },
+ {
+ "epoch": 5.698689956331878,
+ "grad_norm": 0.02986014075577259,
+ "learning_rate": 0.0006,
+ "loss": 5.894903182983398,
+ "step": 410
+ },
+ {
+ "epoch": 5.712663755458515,
+ "grad_norm": 0.030171144753694534,
+ "learning_rate": 0.0006,
+ "loss": 5.876595497131348,
+ "step": 411
+ },
+ {
+ "epoch": 5.726637554585153,
+ "grad_norm": 0.029377546161413193,
+ "learning_rate": 0.0006,
+ "loss": 5.911637306213379,
+ "step": 412
+ },
+ {
+ "epoch": 5.74061135371179,
+ "grad_norm": 0.029829490929841995,
+ "learning_rate": 0.0006,
+ "loss": 5.819629669189453,
+ "step": 413
+ },
+ {
+ "epoch": 5.754585152838428,
+ "grad_norm": 0.024139299988746643,
+ "learning_rate": 0.0006,
+ "loss": 5.81728458404541,
+ "step": 414
+ },
+ {
+ "epoch": 5.7685589519650655,
+ "grad_norm": 0.021453432738780975,
+ "learning_rate": 0.0006,
+ "loss": 5.987326145172119,
+ "step": 415
+ },
+ {
+ "epoch": 5.782532751091703,
+ "grad_norm": 0.017942246049642563,
+ "learning_rate": 0.0006,
+ "loss": 5.893008232116699,
+ "step": 416
+ },
+ {
+ "epoch": 5.796506550218341,
+ "grad_norm": 0.019724637269973755,
+ "learning_rate": 0.0006,
+ "loss": 5.912441253662109,
+ "step": 417
+ },
+ {
+ "epoch": 5.810480349344978,
+ "grad_norm": 0.018597912043333054,
+ "learning_rate": 0.0006,
+ "loss": 5.860394477844238,
+ "step": 418
+ },
+ {
+ "epoch": 5.824454148471616,
+ "grad_norm": 0.016232412308454514,
+ "learning_rate": 0.0006,
+ "loss": 5.94991397857666,
+ "step": 419
+ },
+ {
+ "epoch": 5.8384279475982535,
+ "grad_norm": 0.01667204685509205,
+ "learning_rate": 0.0006,
+ "loss": 5.8506364822387695,
+ "step": 420
+ },
+ {
+ "epoch": 5.85240174672489,
+ "grad_norm": 0.014232151210308075,
+ "learning_rate": 0.0006,
+ "loss": 5.928424835205078,
+ "step": 421
+ },
+ {
+ "epoch": 5.866375545851528,
+ "grad_norm": 0.01570476032793522,
+ "learning_rate": 0.0006,
+ "loss": 5.953692436218262,
+ "step": 422
+ },
+ {
+ "epoch": 5.880349344978166,
+ "grad_norm": 0.01586179807782173,
+ "learning_rate": 0.0006,
+ "loss": 5.820014476776123,
+ "step": 423
+ },
+ {
+ "epoch": 5.894323144104804,
+ "grad_norm": 0.014729145914316177,
+ "learning_rate": 0.0006,
+ "loss": 5.860154628753662,
+ "step": 424
+ },
+ {
+ "epoch": 5.908296943231441,
+ "grad_norm": 0.013111459091305733,
+ "learning_rate": 0.0006,
+ "loss": 5.85988187789917,
+ "step": 425
+ },
+ {
+ "epoch": 5.922270742358078,
+ "grad_norm": 0.012937922962009907,
+ "learning_rate": 0.0006,
+ "loss": 5.756265640258789,
+ "step": 426
+ },
+ {
+ "epoch": 5.936244541484716,
+ "grad_norm": 0.01377453189343214,
+ "learning_rate": 0.0006,
+ "loss": 5.8333048820495605,
+ "step": 427
+ },
+ {
+ "epoch": 5.950218340611354,
+ "grad_norm": 0.014054795727133751,
+ "learning_rate": 0.0006,
+ "loss": 5.803333759307861,
+ "step": 428
+ },
+ {
+ "epoch": 5.964192139737992,
+ "grad_norm": 0.01674959622323513,
+ "learning_rate": 0.0006,
+ "loss": 5.8718485832214355,
+ "step": 429
+ },
+ {
+ "epoch": 5.978165938864628,
+ "grad_norm": 0.018733017146587372,
+ "learning_rate": 0.0006,
+ "loss": 5.875979900360107,
+ "step": 430
+ },
+ {
+ "epoch": 5.992139737991266,
+ "grad_norm": 0.02088983915746212,
+ "learning_rate": 0.0006,
+ "loss": 5.763634204864502,
+ "step": 431
+ },
+ {
+ "epoch": 6.0,
+ "grad_norm": 0.026705985888838768,
+ "learning_rate": 0.0006,
+ "loss": 5.6922478675842285,
+ "step": 432
+ },
+ {
+ "epoch": 6.0,
+ "eval_loss": 5.847958087921143,
+ "eval_runtime": 58.5451,
+ "eval_samples_per_second": 41.711,
+ "eval_steps_per_second": 1.315,
+ "step": 432
+ },
+ {
+ "epoch": 6.013973799126638,
+ "grad_norm": 0.03181544691324234,
+ "learning_rate": 0.0006,
+ "loss": 5.786368370056152,
+ "step": 433
+ },
+ {
+ "epoch": 6.0279475982532755,
+ "grad_norm": 0.03238112851977348,
+ "learning_rate": 0.0006,
+ "loss": 5.810310363769531,
+ "step": 434
+ },
+ {
+ "epoch": 6.041921397379912,
+ "grad_norm": 0.03916756808757782,
+ "learning_rate": 0.0006,
+ "loss": 5.7761993408203125,
+ "step": 435
+ },
+ {
+ "epoch": 6.05589519650655,
+ "grad_norm": 0.044009730219841,
+ "learning_rate": 0.0006,
+ "loss": 5.898112773895264,
+ "step": 436
+ },
+ {
+ "epoch": 6.069868995633188,
+ "grad_norm": 0.045235246419906616,
+ "learning_rate": 0.0006,
+ "loss": 5.876371383666992,
+ "step": 437
+ },
+ {
+ "epoch": 6.083842794759826,
+ "grad_norm": 0.050218384712934494,
+ "learning_rate": 0.0006,
+ "loss": 5.805103302001953,
+ "step": 438
+ },
+ {
+ "epoch": 6.097816593886463,
+ "grad_norm": 0.05444490164518356,
+ "learning_rate": 0.0006,
+ "loss": 5.826424598693848,
+ "step": 439
+ },
+ {
+ "epoch": 6.1117903930131,
+ "grad_norm": 0.04890982061624527,
+ "learning_rate": 0.0006,
+ "loss": 5.7735395431518555,
+ "step": 440
+ },
+ {
+ "epoch": 6.125764192139738,
+ "grad_norm": 0.05478639155626297,
+ "learning_rate": 0.0006,
+ "loss": 5.823677062988281,
+ "step": 441
+ },
+ {
+ "epoch": 6.139737991266376,
+ "grad_norm": 0.057562313973903656,
+ "learning_rate": 0.0006,
+ "loss": 5.889334678649902,
+ "step": 442
+ },
+ {
+ "epoch": 6.153711790393013,
+ "grad_norm": 0.06447703391313553,
+ "learning_rate": 0.0006,
+ "loss": 5.84621524810791,
+ "step": 443
+ },
+ {
+ "epoch": 6.1676855895196505,
+ "grad_norm": 0.048861872404813766,
+ "learning_rate": 0.0006,
+ "loss": 5.931595802307129,
+ "step": 444
+ },
+ {
+ "epoch": 6.181659388646288,
+ "grad_norm": 0.05521339178085327,
+ "learning_rate": 0.0006,
+ "loss": 5.8229780197143555,
+ "step": 445
+ },
+ {
+ "epoch": 6.195633187772926,
+ "grad_norm": 0.053666990250349045,
+ "learning_rate": 0.0006,
+ "loss": 5.879191875457764,
+ "step": 446
+ },
+ {
+ "epoch": 6.209606986899563,
+ "grad_norm": 0.0451025515794754,
+ "learning_rate": 0.0006,
+ "loss": 5.801628112792969,
+ "step": 447
+ },
+ {
+ "epoch": 6.223580786026201,
+ "grad_norm": 0.04291655868291855,
+ "learning_rate": 0.0006,
+ "loss": 5.914680480957031,
+ "step": 448
+ },
+ {
+ "epoch": 6.2375545851528384,
+ "grad_norm": 0.035102490335702896,
+ "learning_rate": 0.0006,
+ "loss": 5.843682765960693,
+ "step": 449
+ },
+ {
+ "epoch": 6.251528384279476,
+ "grad_norm": 0.03403995931148529,
+ "learning_rate": 0.0006,
+ "loss": 5.827154159545898,
+ "step": 450
+ },
+ {
+ "epoch": 6.265502183406113,
+ "grad_norm": 0.03444375470280647,
+ "learning_rate": 0.0006,
+ "loss": 5.800136566162109,
+ "step": 451
+ },
+ {
+ "epoch": 6.279475982532751,
+ "grad_norm": 0.03165159001946449,
+ "learning_rate": 0.0006,
+ "loss": 5.906252384185791,
+ "step": 452
+ },
+ {
+ "epoch": 6.293449781659389,
+ "grad_norm": 0.026153093203902245,
+ "learning_rate": 0.0006,
+ "loss": 5.7423200607299805,
+ "step": 453
+ },
+ {
+ "epoch": 6.307423580786026,
+ "grad_norm": 0.024357657879590988,
+ "learning_rate": 0.0006,
+ "loss": 5.701364517211914,
+ "step": 454
+ },
+ {
+ "epoch": 6.321397379912664,
+ "grad_norm": 0.021508049219846725,
+ "learning_rate": 0.0006,
+ "loss": 5.8328094482421875,
+ "step": 455
+ },
+ {
+ "epoch": 6.335371179039301,
+ "grad_norm": 0.017313921824097633,
+ "learning_rate": 0.0006,
+ "loss": 5.792145252227783,
+ "step": 456
+ },
+ {
+ "epoch": 6.349344978165939,
+ "grad_norm": 0.018563397228717804,
+ "learning_rate": 0.0006,
+ "loss": 5.732672214508057,
+ "step": 457
+ },
+ {
+ "epoch": 6.3633187772925766,
+ "grad_norm": 0.016568679362535477,
+ "learning_rate": 0.0006,
+ "loss": 5.81948184967041,
+ "step": 458
+ },
+ {
+ "epoch": 6.377292576419214,
+ "grad_norm": 0.014133770950138569,
+ "learning_rate": 0.0006,
+ "loss": 5.734982490539551,
+ "step": 459
+ },
+ {
+ "epoch": 6.391266375545851,
+ "grad_norm": 0.014114447869360447,
+ "learning_rate": 0.0006,
+ "loss": 5.70073127746582,
+ "step": 460
+ },
+ {
+ "epoch": 6.405240174672489,
+ "grad_norm": 0.012907393276691437,
+ "learning_rate": 0.0006,
+ "loss": 5.758626937866211,
+ "step": 461
+ },
+ {
+ "epoch": 6.419213973799127,
+ "grad_norm": 0.012673444114625454,
+ "learning_rate": 0.0006,
+ "loss": 5.775138854980469,
+ "step": 462
+ },
+ {
+ "epoch": 6.4331877729257645,
+ "grad_norm": 0.01311410591006279,
+ "learning_rate": 0.0006,
+ "loss": 5.70920991897583,
+ "step": 463
+ },
+ {
+ "epoch": 6.447161572052401,
+ "grad_norm": 0.012935544364154339,
+ "learning_rate": 0.0006,
+ "loss": 5.781479835510254,
+ "step": 464
+ },
+ {
+ "epoch": 6.461135371179039,
+ "grad_norm": 0.01439738366752863,
+ "learning_rate": 0.0006,
+ "loss": 5.691315650939941,
+ "step": 465
+ },
+ {
+ "epoch": 6.475109170305677,
+ "grad_norm": 0.01205186452716589,
+ "learning_rate": 0.0006,
+ "loss": 5.662499904632568,
+ "step": 466
+ },
+ {
+ "epoch": 6.489082969432315,
+ "grad_norm": 0.011575652286410332,
+ "learning_rate": 0.0006,
+ "loss": 5.6271772384643555,
+ "step": 467
+ },
+ {
+ "epoch": 6.503056768558952,
+ "grad_norm": 0.011672502383589745,
+ "learning_rate": 0.0006,
+ "loss": 5.761662006378174,
+ "step": 468
+ },
+ {
+ "epoch": 6.517030567685589,
+ "grad_norm": 0.011541751213371754,
+ "learning_rate": 0.0006,
+ "loss": 5.762078285217285,
+ "step": 469
+ },
+ {
+ "epoch": 6.531004366812227,
+ "grad_norm": 0.011396365240216255,
+ "learning_rate": 0.0006,
+ "loss": 5.67873477935791,
+ "step": 470
+ },
+ {
+ "epoch": 6.544978165938865,
+ "grad_norm": 0.010688499547541142,
+ "learning_rate": 0.0006,
+ "loss": 5.717051982879639,
+ "step": 471
+ },
+ {
+ "epoch": 6.558951965065502,
+ "grad_norm": 0.012224317528307438,
+ "learning_rate": 0.0006,
+ "loss": 5.707948684692383,
+ "step": 472
+ },
+ {
+ "epoch": 6.5729257641921395,
+ "grad_norm": 0.011856825090944767,
+ "learning_rate": 0.0006,
+ "loss": 5.70878791809082,
+ "step": 473
+ },
+ {
+ "epoch": 6.586899563318777,
+ "grad_norm": 0.01199064590036869,
+ "learning_rate": 0.0006,
+ "loss": 5.708697319030762,
+ "step": 474
+ },
+ {
+ "epoch": 6.600873362445415,
+ "grad_norm": 0.01219446212053299,
+ "learning_rate": 0.0006,
+ "loss": 5.6061577796936035,
+ "step": 475
+ },
+ {
+ "epoch": 6.614847161572053,
+ "grad_norm": 0.013397484086453915,
+ "learning_rate": 0.0006,
+ "loss": 5.624789714813232,
+ "step": 476
+ },
+ {
+ "epoch": 6.62882096069869,
+ "grad_norm": 0.01483136136084795,
+ "learning_rate": 0.0006,
+ "loss": 5.721141338348389,
+ "step": 477
+ },
+ {
+ "epoch": 6.642794759825327,
+ "grad_norm": 0.0194488987326622,
+ "learning_rate": 0.0006,
+ "loss": 5.604279518127441,
+ "step": 478
+ },
+ {
+ "epoch": 6.656768558951965,
+ "grad_norm": 0.02079896256327629,
+ "learning_rate": 0.0006,
+ "loss": 5.586322784423828,
+ "step": 479
+ },
+ {
+ "epoch": 6.670742358078603,
+ "grad_norm": 0.018996229395270348,
+ "learning_rate": 0.0006,
+ "loss": 5.5987772941589355,
+ "step": 480
+ },
+ {
+ "epoch": 6.68471615720524,
+ "grad_norm": 0.015779603272676468,
+ "learning_rate": 0.0006,
+ "loss": 5.588602066040039,
+ "step": 481
+ },
+ {
+ "epoch": 6.698689956331878,
+ "grad_norm": 0.015322973020374775,
+ "learning_rate": 0.0006,
+ "loss": 5.686467170715332,
+ "step": 482
+ },
+ {
+ "epoch": 6.712663755458515,
+ "grad_norm": 0.015282213687896729,
+ "learning_rate": 0.0006,
+ "loss": 5.664676666259766,
+ "step": 483
+ },
+ {
+ "epoch": 6.726637554585153,
+ "grad_norm": 0.016589272767305374,
+ "learning_rate": 0.0006,
+ "loss": 5.666739463806152,
+ "step": 484
+ },
+ {
+ "epoch": 6.74061135371179,
+ "grad_norm": 0.01811421848833561,
+ "learning_rate": 0.0006,
+ "loss": 5.647593021392822,
+ "step": 485
+ },
+ {
+ "epoch": 6.754585152838428,
+ "grad_norm": 0.019652029499411583,
+ "learning_rate": 0.0006,
+ "loss": 5.628726959228516,
+ "step": 486
+ },
+ {
+ "epoch": 6.7685589519650655,
+ "grad_norm": 0.02118665538728237,
+ "learning_rate": 0.0006,
+ "loss": 5.6580705642700195,
+ "step": 487
+ },
+ {
+ "epoch": 6.782532751091703,
+ "grad_norm": 0.02237832546234131,
+ "learning_rate": 0.0006,
+ "loss": 5.635931968688965,
+ "step": 488
+ },
+ {
+ "epoch": 6.796506550218341,
+ "grad_norm": 0.023897631093859673,
+ "learning_rate": 0.0006,
+ "loss": 5.517949104309082,
+ "step": 489
+ },
+ {
+ "epoch": 6.810480349344978,
+ "grad_norm": 0.02442227490246296,
+ "learning_rate": 0.0006,
+ "loss": 5.600021839141846,
+ "step": 490
+ },
+ {
+ "epoch": 6.824454148471616,
+ "grad_norm": 0.024675287306308746,
+ "learning_rate": 0.0006,
+ "loss": 5.6639909744262695,
+ "step": 491
+ },
+ {
+ "epoch": 6.8384279475982535,
+ "grad_norm": 0.030372392386198044,
+ "learning_rate": 0.0006,
+ "loss": 5.582888603210449,
+ "step": 492
+ },
+ {
+ "epoch": 6.85240174672489,
+ "grad_norm": 0.03237690031528473,
+ "learning_rate": 0.0006,
+ "loss": 5.573896884918213,
+ "step": 493
+ },
+ {
+ "epoch": 6.866375545851528,
+ "grad_norm": 0.03511589393019676,
+ "learning_rate": 0.0006,
+ "loss": 5.565878868103027,
+ "step": 494
+ },
+ {
+ "epoch": 6.880349344978166,
+ "grad_norm": 0.03883938118815422,
+ "learning_rate": 0.0006,
+ "loss": 5.58540678024292,
+ "step": 495
+ },
+ {
+ "epoch": 6.894323144104804,
+ "grad_norm": 0.04175502806901932,
+ "learning_rate": 0.0006,
+ "loss": 5.6095194816589355,
+ "step": 496
+ },
+ {
+ "epoch": 6.908296943231441,
+ "grad_norm": 0.04013441503047943,
+ "learning_rate": 0.0006,
+ "loss": 5.654401779174805,
+ "step": 497
+ },
+ {
+ "epoch": 6.922270742358078,
+ "grad_norm": 0.04618770629167557,
+ "learning_rate": 0.0006,
+ "loss": 5.6483473777771,
+ "step": 498
+ },
+ {
+ "epoch": 6.936244541484716,
+ "grad_norm": 0.04507308453321457,
+ "learning_rate": 0.0006,
+ "loss": 5.5856218338012695,
+ "step": 499
+ },
+ {
+ "epoch": 6.950218340611354,
+ "grad_norm": 0.03151383996009827,
+ "learning_rate": 0.0006,
+ "loss": 5.594086647033691,
+ "step": 500
+ },
+ {
+ "epoch": 6.964192139737992,
+ "grad_norm": 0.026883797720074654,
+ "learning_rate": 0.0006,
+ "loss": 5.607676029205322,
+ "step": 501
+ },
+ {
+ "epoch": 6.978165938864628,
+ "grad_norm": 0.02981836162507534,
+ "learning_rate": 0.0006,
+ "loss": 5.563666820526123,
+ "step": 502
+ },
+ {
+ "epoch": 6.992139737991266,
+ "grad_norm": 0.03233477845788002,
+ "learning_rate": 0.0006,
+ "loss": 5.667543411254883,
+ "step": 503
+ },
+ {
+ "epoch": 7.0,
+ "grad_norm": 0.029927456751465797,
+ "learning_rate": 0.0006,
+ "loss": 5.733482837677002,
+ "step": 504
+ },
+ {
+ "epoch": 7.0,
+ "eval_loss": 5.679076671600342,
+ "eval_runtime": 59.5965,
+ "eval_samples_per_second": 40.976,
+ "eval_steps_per_second": 1.292,
+ "step": 504
+ },
+ {
+ "epoch": 7.013973799126638,
+ "grad_norm": 0.027081597596406937,
+ "learning_rate": 0.0006,
+ "loss": 5.634487152099609,
+ "step": 505
+ },
+ {
+ "epoch": 7.0279475982532755,
+ "grad_norm": 0.031959421932697296,
+ "learning_rate": 0.0006,
+ "loss": 5.643294811248779,
+ "step": 506
+ },
+ {
+ "epoch": 7.041921397379912,
+ "grad_norm": 0.02802063524723053,
+ "learning_rate": 0.0006,
+ "loss": 5.573239326477051,
+ "step": 507
+ },
+ {
+ "epoch": 7.05589519650655,
+ "grad_norm": 0.032579705119132996,
+ "learning_rate": 0.0006,
+ "loss": 5.627150058746338,
+ "step": 508
+ },
+ {
+ "epoch": 7.069868995633188,
+ "grad_norm": 0.02503928542137146,
+ "learning_rate": 0.0006,
+ "loss": 5.6045308113098145,
+ "step": 509
+ },
+ {
+ "epoch": 7.083842794759826,
+ "grad_norm": 0.02421317622065544,
+ "learning_rate": 0.0006,
+ "loss": 5.522153377532959,
+ "step": 510
+ },
+ {
+ "epoch": 7.097816593886463,
+ "grad_norm": 0.021983426064252853,
+ "learning_rate": 0.0006,
+ "loss": 5.632939338684082,
+ "step": 511
+ },
+ {
+ "epoch": 7.1117903930131,
+ "grad_norm": 0.021933183073997498,
+ "learning_rate": 0.0006,
+ "loss": 5.592185974121094,
+ "step": 512
+ },
+ {
+ "epoch": 7.125764192139738,
+ "grad_norm": 0.02287031151354313,
+ "learning_rate": 0.0006,
+ "loss": 5.613700866699219,
+ "step": 513
+ },
+ {
+ "epoch": 7.139737991266376,
+ "grad_norm": 0.02281602844595909,
+ "learning_rate": 0.0006,
+ "loss": 5.551383018493652,
+ "step": 514
+ },
+ {
+ "epoch": 7.153711790393013,
+ "grad_norm": 0.021050360053777695,
+ "learning_rate": 0.0006,
+ "loss": 5.55275821685791,
+ "step": 515
+ },
+ {
+ "epoch": 7.1676855895196505,
+ "grad_norm": 0.018299926072359085,
+ "learning_rate": 0.0006,
+ "loss": 5.528225421905518,
+ "step": 516
+ },
+ {
+ "epoch": 7.181659388646288,
+ "grad_norm": 0.02024853602051735,
+ "learning_rate": 0.0006,
+ "loss": 5.522453784942627,
+ "step": 517
+ },
+ {
+ "epoch": 7.195633187772926,
+ "grad_norm": 0.02135239914059639,
+ "learning_rate": 0.0006,
+ "loss": 5.527522563934326,
+ "step": 518
+ },
+ {
+ "epoch": 7.209606986899563,
+ "grad_norm": 0.023394184187054634,
+ "learning_rate": 0.0006,
+ "loss": 5.4535017013549805,
+ "step": 519
+ },
+ {
+ "epoch": 7.223580786026201,
+ "grad_norm": 0.025952080264687538,
+ "learning_rate": 0.0006,
+ "loss": 5.5374674797058105,
+ "step": 520
+ },
+ {
+ "epoch": 7.2375545851528384,
+ "grad_norm": 0.022287718951702118,
+ "learning_rate": 0.0006,
+ "loss": 5.493753433227539,
+ "step": 521
+ },
+ {
+ "epoch": 7.251528384279476,
+ "grad_norm": 0.02018447406589985,
+ "learning_rate": 0.0006,
+ "loss": 5.510575294494629,
+ "step": 522
+ },
+ {
+ "epoch": 7.265502183406113,
+ "grad_norm": 0.02050507813692093,
+ "learning_rate": 0.0006,
+ "loss": 5.466026782989502,
+ "step": 523
+ },
+ {
+ "epoch": 7.279475982532751,
+ "grad_norm": 0.023688802495598793,
+ "learning_rate": 0.0006,
+ "loss": 5.485815525054932,
+ "step": 524
+ },
+ {
+ "epoch": 7.293449781659389,
+ "grad_norm": 0.021078843623399734,
+ "learning_rate": 0.0006,
+ "loss": 5.415581703186035,
+ "step": 525
+ },
+ {
+ "epoch": 7.307423580786026,
+ "grad_norm": 0.018266011029481888,
+ "learning_rate": 0.0006,
+ "loss": 5.402815818786621,
+ "step": 526
+ },
+ {
+ "epoch": 7.321397379912664,
+ "grad_norm": 0.019693493843078613,
+ "learning_rate": 0.0006,
+ "loss": 5.504674911499023,
+ "step": 527
+ },
+ {
+ "epoch": 7.335371179039301,
+ "grad_norm": 0.02031373605132103,
+ "learning_rate": 0.0006,
+ "loss": 5.503837585449219,
+ "step": 528
+ },
+ {
+ "epoch": 7.349344978165939,
+ "grad_norm": 0.017636626958847046,
+ "learning_rate": 0.0006,
+ "loss": 5.530580997467041,
+ "step": 529
+ },
+ {
+ "epoch": 7.3633187772925766,
+ "grad_norm": 0.01787244901061058,
+ "learning_rate": 0.0006,
+ "loss": 5.450218677520752,
+ "step": 530
+ },
+ {
+ "epoch": 7.377292576419214,
+ "grad_norm": 0.0170669574290514,
+ "learning_rate": 0.0006,
+ "loss": 5.481570243835449,
+ "step": 531
+ },
+ {
+ "epoch": 7.391266375545851,
+ "grad_norm": 0.01802165061235428,
+ "learning_rate": 0.0006,
+ "loss": 5.386394500732422,
+ "step": 532
+ },
+ {
+ "epoch": 7.405240174672489,
+ "grad_norm": 0.022949472069740295,
+ "learning_rate": 0.0006,
+ "loss": 5.424929618835449,
+ "step": 533
+ },
+ {
+ "epoch": 7.419213973799127,
+ "grad_norm": 0.034213390201330185,
+ "learning_rate": 0.0006,
+ "loss": 5.346663475036621,
+ "step": 534
+ },
+ {
+ "epoch": 7.4331877729257645,
+ "grad_norm": 0.03847593814134598,
+ "learning_rate": 0.0006,
+ "loss": 5.418482780456543,
+ "step": 535
+ },
+ {
+ "epoch": 7.447161572052401,
+ "grad_norm": 0.029393676668405533,
+ "learning_rate": 0.0006,
+ "loss": 5.456090927124023,
+ "step": 536
+ },
+ {
+ "epoch": 7.461135371179039,
+ "grad_norm": 0.03388667106628418,
+ "learning_rate": 0.0006,
+ "loss": 5.572983741760254,
+ "step": 537
+ },
+ {
+ "epoch": 7.475109170305677,
+ "grad_norm": 0.039690613746643066,
+ "learning_rate": 0.0006,
+ "loss": 5.414067268371582,
+ "step": 538
+ },
+ {
+ "epoch": 7.489082969432315,
+ "grad_norm": 0.03630776330828667,
+ "learning_rate": 0.0006,
+ "loss": 5.523739814758301,
+ "step": 539
+ },
+ {
+ "epoch": 7.503056768558952,
+ "grad_norm": 0.03356870263814926,
+ "learning_rate": 0.0006,
+ "loss": 5.444767951965332,
+ "step": 540
+ },
+ {
+ "epoch": 7.517030567685589,
+ "grad_norm": 0.030812304466962814,
+ "learning_rate": 0.0006,
+ "loss": 5.461244106292725,
+ "step": 541
+ },
+ {
+ "epoch": 7.531004366812227,
+ "grad_norm": 0.03319217637181282,
+ "learning_rate": 0.0006,
+ "loss": 5.502161026000977,
+ "step": 542
+ },
+ {
+ "epoch": 7.544978165938865,
+ "grad_norm": 0.032997481524944305,
+ "learning_rate": 0.0006,
+ "loss": 5.4162278175354,
+ "step": 543
+ },
+ {
+ "epoch": 7.558951965065502,
+ "grad_norm": 0.03364962339401245,
+ "learning_rate": 0.0006,
+ "loss": 5.415736198425293,
+ "step": 544
+ },
+ {
+ "epoch": 7.5729257641921395,
+ "grad_norm": 0.0344221405684948,
+ "learning_rate": 0.0006,
+ "loss": 5.541967868804932,
+ "step": 545
+ },
+ {
+ "epoch": 7.586899563318777,
+ "grad_norm": 0.029609503224492073,
+ "learning_rate": 0.0006,
+ "loss": 5.440614223480225,
+ "step": 546
+ },
+ {
+ "epoch": 7.600873362445415,
+ "grad_norm": 0.029055926948785782,
+ "learning_rate": 0.0006,
+ "loss": 5.462865352630615,
+ "step": 547
+ },
+ {
+ "epoch": 7.614847161572053,
+ "grad_norm": 0.02658848837018013,
+ "learning_rate": 0.0006,
+ "loss": 5.444467544555664,
+ "step": 548
+ },
+ {
+ "epoch": 7.62882096069869,
+ "grad_norm": 0.026276404038071632,
+ "learning_rate": 0.0006,
+ "loss": 5.462278842926025,
+ "step": 549
+ },
+ {
+ "epoch": 7.642794759825327,
+ "grad_norm": 0.0286889486014843,
+ "learning_rate": 0.0006,
+ "loss": 5.4849534034729,
+ "step": 550
+ },
+ {
+ "epoch": 7.656768558951965,
+ "grad_norm": 0.026849817484617233,
+ "learning_rate": 0.0006,
+ "loss": 5.431473731994629,
+ "step": 551
+ },
+ {
+ "epoch": 7.670742358078603,
+ "grad_norm": 0.02312396466732025,
+ "learning_rate": 0.0006,
+ "loss": 5.45462703704834,
+ "step": 552
+ },
+ {
+ "epoch": 7.68471615720524,
+ "grad_norm": 0.026197485625743866,
+ "learning_rate": 0.0006,
+ "loss": 5.452552318572998,
+ "step": 553
+ },
+ {
+ "epoch": 7.698689956331878,
+ "grad_norm": 0.02747255191206932,
+ "learning_rate": 0.0006,
+ "loss": 5.447169780731201,
+ "step": 554
+ },
+ {
+ "epoch": 7.712663755458515,
+ "grad_norm": 0.028123432770371437,
+ "learning_rate": 0.0006,
+ "loss": 5.394400596618652,
+ "step": 555
+ },
+ {
+ "epoch": 7.726637554585153,
+ "grad_norm": 0.02599870041012764,
+ "learning_rate": 0.0006,
+ "loss": 5.398341178894043,
+ "step": 556
+ },
+ {
+ "epoch": 7.74061135371179,
+ "grad_norm": 0.022171657532453537,
+ "learning_rate": 0.0006,
+ "loss": 5.368820667266846,
+ "step": 557
+ },
+ {
+ "epoch": 7.754585152838428,
+ "grad_norm": 0.022309480234980583,
+ "learning_rate": 0.0006,
+ "loss": 5.416772365570068,
+ "step": 558
+ },
+ {
+ "epoch": 7.7685589519650655,
+ "grad_norm": 0.024074165150523186,
+ "learning_rate": 0.0006,
+ "loss": 5.359002113342285,
+ "step": 559
+ },
+ {
+ "epoch": 7.782532751091703,
+ "grad_norm": 0.02636653557419777,
+ "learning_rate": 0.0006,
+ "loss": 5.484368324279785,
+ "step": 560
+ },
+ {
+ "epoch": 7.796506550218341,
+ "grad_norm": 0.02203752100467682,
+ "learning_rate": 0.0006,
+ "loss": 5.375498294830322,
+ "step": 561
+ },
+ {
+ "epoch": 7.810480349344978,
+ "grad_norm": 0.019924819469451904,
+ "learning_rate": 0.0006,
+ "loss": 5.334723472595215,
+ "step": 562
+ },
+ {
+ "epoch": 7.824454148471616,
+ "grad_norm": 0.018755966797471046,
+ "learning_rate": 0.0006,
+ "loss": 5.348781108856201,
+ "step": 563
+ },
+ {
+ "epoch": 7.8384279475982535,
+ "grad_norm": 0.016747845336794853,
+ "learning_rate": 0.0006,
+ "loss": 5.433079719543457,
+ "step": 564
+ },
+ {
+ "epoch": 7.85240174672489,
+ "grad_norm": 0.0170395877212286,
+ "learning_rate": 0.0006,
+ "loss": 5.346794128417969,
+ "step": 565
+ },
+ {
+ "epoch": 7.866375545851528,
+ "grad_norm": 0.016074176877737045,
+ "learning_rate": 0.0006,
+ "loss": 5.3277740478515625,
+ "step": 566
+ },
+ {
+ "epoch": 7.880349344978166,
+ "grad_norm": 0.014119806699454784,
+ "learning_rate": 0.0006,
+ "loss": 5.483551979064941,
+ "step": 567
+ },
+ {
+ "epoch": 7.894323144104804,
+ "grad_norm": 0.015270394273102283,
+ "learning_rate": 0.0006,
+ "loss": 5.378519535064697,
+ "step": 568
+ },
+ {
+ "epoch": 7.908296943231441,
+ "grad_norm": 0.015677539631724358,
+ "learning_rate": 0.0006,
+ "loss": 5.290210247039795,
+ "step": 569
+ },
+ {
+ "epoch": 7.922270742358078,
+ "grad_norm": 0.015930423513054848,
+ "learning_rate": 0.0006,
+ "loss": 5.427360534667969,
+ "step": 570
+ },
+ {
+ "epoch": 7.936244541484716,
+ "grad_norm": 0.016641564667224884,
+ "learning_rate": 0.0006,
+ "loss": 5.301599979400635,
+ "step": 571
+ },
+ {
+ "epoch": 7.950218340611354,
+ "grad_norm": 0.018293552100658417,
+ "learning_rate": 0.0006,
+ "loss": 5.277889251708984,
+ "step": 572
+ },
+ {
+ "epoch": 7.964192139737992,
+ "grad_norm": 0.017618799582123756,
+ "learning_rate": 0.0006,
+ "loss": 5.297194480895996,
+ "step": 573
+ },
+ {
+ "epoch": 7.978165938864628,
+ "grad_norm": 0.015034242533147335,
+ "learning_rate": 0.0006,
+ "loss": 5.42428731918335,
+ "step": 574
+ },
+ {
+ "epoch": 7.992139737991266,
+ "grad_norm": 0.01617511734366417,
+ "learning_rate": 0.0006,
+ "loss": 5.420044898986816,
+ "step": 575
+ },
+ {
+ "epoch": 8.0,
+ "grad_norm": 0.01837257295846939,
+ "learning_rate": 0.0006,
+ "loss": 5.240387916564941,
+ "step": 576
+ },
+ {
+ "epoch": 8.0,
+ "eval_loss": 5.423932075500488,
+ "eval_runtime": 58.0841,
+ "eval_samples_per_second": 42.042,
+ "eval_steps_per_second": 1.326,
+ "step": 576
+ },
+ {
+ "epoch": 8.013973799126637,
+ "grad_norm": 0.020953809842467308,
+ "learning_rate": 0.0006,
+ "loss": 5.4320783615112305,
+ "step": 577
+ },
+ {
+ "epoch": 8.027947598253276,
+ "grad_norm": 0.0227745920419693,
+ "learning_rate": 0.0006,
+ "loss": 5.310197830200195,
+ "step": 578
+ },
+ {
+ "epoch": 8.041921397379912,
+ "grad_norm": 0.021372603252530098,
+ "learning_rate": 0.0006,
+ "loss": 5.339468002319336,
+ "step": 579
+ },
+ {
+ "epoch": 8.055895196506551,
+ "grad_norm": 0.022401731461286545,
+ "learning_rate": 0.0006,
+ "loss": 5.279829978942871,
+ "step": 580
+ },
+ {
+ "epoch": 8.069868995633188,
+ "grad_norm": 0.023942379280924797,
+ "learning_rate": 0.0006,
+ "loss": 5.372323989868164,
+ "step": 581
+ },
+ {
+ "epoch": 8.083842794759825,
+ "grad_norm": 0.024380534887313843,
+ "learning_rate": 0.0006,
+ "loss": 5.253486633300781,
+ "step": 582
+ },
+ {
+ "epoch": 8.097816593886463,
+ "grad_norm": 0.03246625140309334,
+ "learning_rate": 0.0006,
+ "loss": 5.305774211883545,
+ "step": 583
+ },
+ {
+ "epoch": 8.1117903930131,
+ "grad_norm": 0.030399736016988754,
+ "learning_rate": 0.0006,
+ "loss": 5.370976448059082,
+ "step": 584
+ },
+ {
+ "epoch": 8.125764192139737,
+ "grad_norm": 0.027148913592100143,
+ "learning_rate": 0.0006,
+ "loss": 5.368010520935059,
+ "step": 585
+ },
+ {
+ "epoch": 8.139737991266376,
+ "grad_norm": 0.02925540879368782,
+ "learning_rate": 0.0006,
+ "loss": 5.303013801574707,
+ "step": 586
+ },
+ {
+ "epoch": 8.153711790393013,
+ "grad_norm": 0.02915453538298607,
+ "learning_rate": 0.0006,
+ "loss": 5.329667091369629,
+ "step": 587
+ },
+ {
+ "epoch": 8.167685589519651,
+ "grad_norm": 0.031732626259326935,
+ "learning_rate": 0.0006,
+ "loss": 5.3634934425354,
+ "step": 588
+ },
+ {
+ "epoch": 8.181659388646288,
+ "grad_norm": 0.036166731268167496,
+ "learning_rate": 0.0006,
+ "loss": 5.223980903625488,
+ "step": 589
+ },
+ {
+ "epoch": 8.195633187772925,
+ "grad_norm": 0.035654108971357346,
+ "learning_rate": 0.0006,
+ "loss": 5.361059188842773,
+ "step": 590
+ },
+ {
+ "epoch": 8.209606986899564,
+ "grad_norm": 0.0365324430167675,
+ "learning_rate": 0.0006,
+ "loss": 5.336151599884033,
+ "step": 591
+ },
+ {
+ "epoch": 8.2235807860262,
+ "grad_norm": 0.03578880429267883,
+ "learning_rate": 0.0006,
+ "loss": 5.434076309204102,
+ "step": 592
+ },
+ {
+ "epoch": 8.237554585152838,
+ "grad_norm": 0.035734012722969055,
+ "learning_rate": 0.0006,
+ "loss": 5.330893516540527,
+ "step": 593
+ },
+ {
+ "epoch": 8.251528384279476,
+ "grad_norm": 0.03196857124567032,
+ "learning_rate": 0.0006,
+ "loss": 5.306009769439697,
+ "step": 594
+ },
+ {
+ "epoch": 8.265502183406113,
+ "grad_norm": 0.032042086124420166,
+ "learning_rate": 0.0006,
+ "loss": 5.343267917633057,
+ "step": 595
+ },
+ {
+ "epoch": 8.279475982532752,
+ "grad_norm": 0.03160746395587921,
+ "learning_rate": 0.0006,
+ "loss": 5.2353715896606445,
+ "step": 596
+ },
+ {
+ "epoch": 8.293449781659389,
+ "grad_norm": 0.026689818128943443,
+ "learning_rate": 0.0006,
+ "loss": 5.254042625427246,
+ "step": 597
+ },
+ {
+ "epoch": 8.307423580786025,
+ "grad_norm": 0.02880188450217247,
+ "learning_rate": 0.0006,
+ "loss": 5.326833248138428,
+ "step": 598
+ },
+ {
+ "epoch": 8.321397379912664,
+ "grad_norm": 0.027516381815075874,
+ "learning_rate": 0.0006,
+ "loss": 5.342336177825928,
+ "step": 599
+ },
+ {
+ "epoch": 8.335371179039301,
+ "grad_norm": 0.027875030413269997,
+ "learning_rate": 0.0006,
+ "loss": 5.200719833374023,
+ "step": 600
+ },
+ {
+ "epoch": 8.34934497816594,
+ "grad_norm": 0.0268265288323164,
+ "learning_rate": 0.0006,
+ "loss": 5.249449729919434,
+ "step": 601
+ },
+ {
+ "epoch": 8.363318777292577,
+ "grad_norm": 0.024824608117341995,
+ "learning_rate": 0.0006,
+ "loss": 5.318952560424805,
+ "step": 602
+ },
+ {
+ "epoch": 8.377292576419213,
+ "grad_norm": 0.022990627214312553,
+ "learning_rate": 0.0006,
+ "loss": 5.244993209838867,
+ "step": 603
+ },
+ {
+ "epoch": 8.391266375545852,
+ "grad_norm": 0.022804604843258858,
+ "learning_rate": 0.0006,
+ "loss": 5.3236846923828125,
+ "step": 604
+ },
+ {
+ "epoch": 8.405240174672489,
+ "grad_norm": 0.02419872209429741,
+ "learning_rate": 0.0006,
+ "loss": 5.255486488342285,
+ "step": 605
+ },
+ {
+ "epoch": 8.419213973799126,
+ "grad_norm": 0.021952059119939804,
+ "learning_rate": 0.0006,
+ "loss": 5.3511552810668945,
+ "step": 606
+ },
+ {
+ "epoch": 8.433187772925764,
+ "grad_norm": 0.022375497967004776,
+ "learning_rate": 0.0006,
+ "loss": 5.294790267944336,
+ "step": 607
+ },
+ {
+ "epoch": 8.447161572052401,
+ "grad_norm": 0.019242815673351288,
+ "learning_rate": 0.0006,
+ "loss": 5.252618789672852,
+ "step": 608
+ },
+ {
+ "epoch": 8.46113537117904,
+ "grad_norm": 0.0173486340790987,
+ "learning_rate": 0.0006,
+ "loss": 5.30147647857666,
+ "step": 609
+ },
+ {
+ "epoch": 8.475109170305677,
+ "grad_norm": 0.01612604409456253,
+ "learning_rate": 0.0006,
+ "loss": 5.23960542678833,
+ "step": 610
+ },
+ {
+ "epoch": 8.489082969432314,
+ "grad_norm": 0.014293976128101349,
+ "learning_rate": 0.0006,
+ "loss": 5.119932651519775,
+ "step": 611
+ },
+ {
+ "epoch": 8.503056768558952,
+ "grad_norm": 0.014453536830842495,
+ "learning_rate": 0.0006,
+ "loss": 5.166794776916504,
+ "step": 612
+ },
+ {
+ "epoch": 8.51703056768559,
+ "grad_norm": 0.014685769565403461,
+ "learning_rate": 0.0006,
+ "loss": 5.264822959899902,
+ "step": 613
+ },
+ {
+ "epoch": 8.531004366812226,
+ "grad_norm": 0.01427740603685379,
+ "learning_rate": 0.0006,
+ "loss": 5.278433799743652,
+ "step": 614
+ },
+ {
+ "epoch": 8.544978165938865,
+ "grad_norm": 0.014864951372146606,
+ "learning_rate": 0.0006,
+ "loss": 5.214512348175049,
+ "step": 615
+ },
+ {
+ "epoch": 8.558951965065502,
+ "grad_norm": 0.01565164513885975,
+ "learning_rate": 0.0006,
+ "loss": 5.183066368103027,
+ "step": 616
+ },
+ {
+ "epoch": 8.57292576419214,
+ "grad_norm": 0.01779816672205925,
+ "learning_rate": 0.0006,
+ "loss": 5.163185119628906,
+ "step": 617
+ },
+ {
+ "epoch": 8.586899563318777,
+ "grad_norm": 0.017254827544093132,
+ "learning_rate": 0.0006,
+ "loss": 5.243555068969727,
+ "step": 618
+ },
+ {
+ "epoch": 8.600873362445414,
+ "grad_norm": 0.01650584116578102,
+ "learning_rate": 0.0006,
+ "loss": 5.234884738922119,
+ "step": 619
+ },
+ {
+ "epoch": 8.614847161572053,
+ "grad_norm": 0.017021115869283676,
+ "learning_rate": 0.0006,
+ "loss": 5.12271785736084,
+ "step": 620
+ },
+ {
+ "epoch": 8.62882096069869,
+ "grad_norm": 0.01773759163916111,
+ "learning_rate": 0.0006,
+ "loss": 5.242153167724609,
+ "step": 621
+ },
+ {
+ "epoch": 8.642794759825328,
+ "grad_norm": 0.015679042786359787,
+ "learning_rate": 0.0006,
+ "loss": 5.170779228210449,
+ "step": 622
+ },
+ {
+ "epoch": 8.656768558951965,
+ "grad_norm": 0.013760549947619438,
+ "learning_rate": 0.0006,
+ "loss": 5.238644599914551,
+ "step": 623
+ },
+ {
+ "epoch": 8.670742358078602,
+ "grad_norm": 0.014571522362530231,
+ "learning_rate": 0.0006,
+ "loss": 5.177056312561035,
+ "step": 624
+ },
+ {
+ "epoch": 8.68471615720524,
+ "grad_norm": 0.016209015622735023,
+ "learning_rate": 0.0006,
+ "loss": 5.163750648498535,
+ "step": 625
+ },
+ {
+ "epoch": 8.698689956331878,
+ "grad_norm": 0.016813941299915314,
+ "learning_rate": 0.0006,
+ "loss": 5.183428764343262,
+ "step": 626
+ },
+ {
+ "epoch": 8.712663755458514,
+ "grad_norm": 0.020985357463359833,
+ "learning_rate": 0.0006,
+ "loss": 5.229465007781982,
+ "step": 627
+ },
+ {
+ "epoch": 8.726637554585153,
+ "grad_norm": 0.02679632417857647,
+ "learning_rate": 0.0006,
+ "loss": 5.257368087768555,
+ "step": 628
+ },
+ {
+ "epoch": 8.74061135371179,
+ "grad_norm": 0.025756070390343666,
+ "learning_rate": 0.0006,
+ "loss": 5.253736972808838,
+ "step": 629
+ },
+ {
+ "epoch": 8.754585152838429,
+ "grad_norm": 0.02643490768969059,
+ "learning_rate": 0.0006,
+ "loss": 5.228633403778076,
+ "step": 630
+ },
+ {
+ "epoch": 8.768558951965066,
+ "grad_norm": 0.03303210437297821,
+ "learning_rate": 0.0006,
+ "loss": 5.163540363311768,
+ "step": 631
+ },
+ {
+ "epoch": 8.782532751091702,
+ "grad_norm": 0.030432431027293205,
+ "learning_rate": 0.0006,
+ "loss": 5.243169784545898,
+ "step": 632
+ },
+ {
+ "epoch": 8.796506550218341,
+ "grad_norm": 0.028631288558244705,
+ "learning_rate": 0.0006,
+ "loss": 5.166398048400879,
+ "step": 633
+ },
+ {
+ "epoch": 8.810480349344978,
+ "grad_norm": 0.026188310235738754,
+ "learning_rate": 0.0006,
+ "loss": 5.128122329711914,
+ "step": 634
+ },
+ {
+ "epoch": 8.824454148471617,
+ "grad_norm": 0.025308528915047646,
+ "learning_rate": 0.0006,
+ "loss": 5.211284637451172,
+ "step": 635
+ },
+ {
+ "epoch": 8.838427947598253,
+ "grad_norm": 0.020657729357481003,
+ "learning_rate": 0.0006,
+ "loss": 5.161575794219971,
+ "step": 636
+ },
+ {
+ "epoch": 8.85240174672489,
+ "grad_norm": 0.021901234984397888,
+ "learning_rate": 0.0006,
+ "loss": 5.20050048828125,
+ "step": 637
+ },
+ {
+ "epoch": 8.866375545851529,
+ "grad_norm": 0.021219318732619286,
+ "learning_rate": 0.0006,
+ "loss": 5.211699485778809,
+ "step": 638
+ },
+ {
+ "epoch": 8.880349344978166,
+ "grad_norm": 0.0200974028557539,
+ "learning_rate": 0.0006,
+ "loss": 5.092503547668457,
+ "step": 639
+ },
+ {
+ "epoch": 8.894323144104803,
+ "grad_norm": 0.023804882541298866,
+ "learning_rate": 0.0006,
+ "loss": 5.216068267822266,
+ "step": 640
+ },
+ {
+ "epoch": 8.908296943231441,
+ "grad_norm": 0.026088010519742966,
+ "learning_rate": 0.0006,
+ "loss": 5.155592918395996,
+ "step": 641
+ },
+ {
+ "epoch": 8.922270742358078,
+ "grad_norm": 0.02601276896893978,
+ "learning_rate": 0.0006,
+ "loss": 5.197238922119141,
+ "step": 642
+ },
+ {
+ "epoch": 8.936244541484717,
+ "grad_norm": 0.020387211814522743,
+ "learning_rate": 0.0006,
+ "loss": 5.1955437660217285,
+ "step": 643
+ },
+ {
+ "epoch": 8.950218340611354,
+ "grad_norm": 0.019214622676372528,
+ "learning_rate": 0.0006,
+ "loss": 5.206346035003662,
+ "step": 644
+ },
+ {
+ "epoch": 8.96419213973799,
+ "grad_norm": 0.019674314185976982,
+ "learning_rate": 0.0006,
+ "loss": 5.219857215881348,
+ "step": 645
+ },
+ {
+ "epoch": 8.97816593886463,
+ "grad_norm": 0.020318234339356422,
+ "learning_rate": 0.0006,
+ "loss": 5.191132545471191,
+ "step": 646
+ },
+ {
+ "epoch": 8.992139737991266,
+ "grad_norm": 0.021428676322102547,
+ "learning_rate": 0.0006,
+ "loss": 5.214695930480957,
+ "step": 647
+ },
+ {
+ "epoch": 9.0,
+ "grad_norm": 0.023235054686665535,
+ "learning_rate": 0.0006,
+ "loss": 5.229091644287109,
+ "step": 648
+ },
+ {
+ "epoch": 9.0,
+ "eval_loss": 5.289593696594238,
+ "eval_runtime": 56.9256,
+ "eval_samples_per_second": 42.898,
+ "eval_steps_per_second": 1.353,
+ "step": 648
+ },
+ {
+ "epoch": 9.013973799126637,
+ "grad_norm": 0.02320289984345436,
+ "learning_rate": 0.0006,
+ "loss": 4.983669281005859,
+ "step": 649
+ },
+ {
+ "epoch": 9.027947598253276,
+ "grad_norm": 0.018521282821893692,
+ "learning_rate": 0.0006,
+ "loss": 5.200575828552246,
+ "step": 650
+ },
+ {
+ "epoch": 9.041921397379912,
+ "grad_norm": 0.0180169939994812,
+ "learning_rate": 0.0006,
+ "loss": 5.148033618927002,
+ "step": 651
+ },
+ {
+ "epoch": 9.055895196506551,
+ "grad_norm": 0.019969860091805458,
+ "learning_rate": 0.0006,
+ "loss": 5.099125385284424,
+ "step": 652
+ },
+ {
+ "epoch": 9.069868995633188,
+ "grad_norm": 0.01643305830657482,
+ "learning_rate": 0.0006,
+ "loss": 5.092347145080566,
+ "step": 653
+ },
+ {
+ "epoch": 9.083842794759825,
+ "grad_norm": 0.016022363677620888,
+ "learning_rate": 0.0006,
+ "loss": 5.129279613494873,
+ "step": 654
+ },
+ {
+ "epoch": 9.097816593886463,
+ "grad_norm": 0.01657041162252426,
+ "learning_rate": 0.0006,
+ "loss": 5.159562587738037,
+ "step": 655
+ },
+ {
+ "epoch": 9.1117903930131,
+ "grad_norm": 0.017924228683114052,
+ "learning_rate": 0.0006,
+ "loss": 5.165497779846191,
+ "step": 656
+ },
+ {
+ "epoch": 9.125764192139737,
+ "grad_norm": 0.021592361852526665,
+ "learning_rate": 0.0006,
+ "loss": 5.204775810241699,
+ "step": 657
+ },
+ {
+ "epoch": 9.139737991266376,
+ "grad_norm": 0.026924120262265205,
+ "learning_rate": 0.0006,
+ "loss": 5.128296852111816,
+ "step": 658
+ },
+ {
+ "epoch": 9.153711790393013,
+ "grad_norm": 0.023156899958848953,
+ "learning_rate": 0.0006,
+ "loss": 5.1255340576171875,
+ "step": 659
+ },
+ {
+ "epoch": 9.167685589519651,
+ "grad_norm": 0.020670367404818535,
+ "learning_rate": 0.0006,
+ "loss": 5.128479957580566,
+ "step": 660
+ },
+ {
+ "epoch": 9.181659388646288,
+ "grad_norm": 0.025447173044085503,
+ "learning_rate": 0.0006,
+ "loss": 5.154821872711182,
+ "step": 661
+ },
+ {
+ "epoch": 9.195633187772925,
+ "grad_norm": 0.027940068393945694,
+ "learning_rate": 0.0006,
+ "loss": 5.143466472625732,
+ "step": 662
+ },
+ {
+ "epoch": 9.209606986899564,
+ "grad_norm": 0.026847844943404198,
+ "learning_rate": 0.0006,
+ "loss": 5.045665740966797,
+ "step": 663
+ },
+ {
+ "epoch": 9.2235807860262,
+ "grad_norm": 0.02340601570904255,
+ "learning_rate": 0.0006,
+ "loss": 5.21933126449585,
+ "step": 664
+ },
+ {
+ "epoch": 9.237554585152838,
+ "grad_norm": 0.02340371161699295,
+ "learning_rate": 0.0006,
+ "loss": 4.942949295043945,
+ "step": 665
+ },
+ {
+ "epoch": 9.251528384279476,
+ "grad_norm": 0.02221992425620556,
+ "learning_rate": 0.0006,
+ "loss": 5.197099208831787,
+ "step": 666
+ },
+ {
+ "epoch": 9.265502183406113,
+ "grad_norm": 0.023508677259087563,
+ "learning_rate": 0.0006,
+ "loss": 5.022897720336914,
+ "step": 667
+ },
+ {
+ "epoch": 9.279475982532752,
+ "grad_norm": 0.026120394468307495,
+ "learning_rate": 0.0006,
+ "loss": 5.087725639343262,
+ "step": 668
+ },
+ {
+ "epoch": 9.293449781659389,
+ "grad_norm": 0.027273228392004967,
+ "learning_rate": 0.0006,
+ "loss": 5.090963840484619,
+ "step": 669
+ },
+ {
+ "epoch": 9.307423580786025,
+ "grad_norm": 0.03241586685180664,
+ "learning_rate": 0.0006,
+ "loss": 5.047842979431152,
+ "step": 670
+ },
+ {
+ "epoch": 9.321397379912664,
+ "grad_norm": 0.030245667323470116,
+ "learning_rate": 0.0006,
+ "loss": 5.104803085327148,
+ "step": 671
+ },
+ {
+ "epoch": 9.335371179039301,
+ "grad_norm": 0.027698364108800888,
+ "learning_rate": 0.0006,
+ "loss": 5.0644097328186035,
+ "step": 672
+ },
+ {
+ "epoch": 9.34934497816594,
+ "grad_norm": 0.029692554846405983,
+ "learning_rate": 0.0006,
+ "loss": 5.111942291259766,
+ "step": 673
+ },
+ {
+ "epoch": 9.363318777292577,
+ "grad_norm": 0.03328656405210495,
+ "learning_rate": 0.0006,
+ "loss": 5.1574554443359375,
+ "step": 674
+ },
+ {
+ "epoch": 9.377292576419213,
+ "grad_norm": 0.031938180327415466,
+ "learning_rate": 0.0006,
+ "loss": 5.190371990203857,
+ "step": 675
+ },
+ {
+ "epoch": 9.391266375545852,
+ "grad_norm": 0.024858945980668068,
+ "learning_rate": 0.0006,
+ "loss": 5.121420860290527,
+ "step": 676
+ },
+ {
+ "epoch": 9.405240174672489,
+ "grad_norm": 0.023033203557133675,
+ "learning_rate": 0.0006,
+ "loss": 5.0974345207214355,
+ "step": 677
+ },
+ {
+ "epoch": 9.419213973799126,
+ "grad_norm": 0.022083545103669167,
+ "learning_rate": 0.0006,
+ "loss": 5.046319961547852,
+ "step": 678
+ },
+ {
+ "epoch": 9.433187772925764,
+ "grad_norm": 0.0228437427431345,
+ "learning_rate": 0.0006,
+ "loss": 5.119963645935059,
+ "step": 679
+ },
+ {
+ "epoch": 9.447161572052401,
+ "grad_norm": 0.0242351982742548,
+ "learning_rate": 0.0006,
+ "loss": 5.0444560050964355,
+ "step": 680
+ },
+ {
+ "epoch": 9.46113537117904,
+ "grad_norm": 0.02401185780763626,
+ "learning_rate": 0.0006,
+ "loss": 5.063238620758057,
+ "step": 681
+ },
+ {
+ "epoch": 9.475109170305677,
+ "grad_norm": 0.02560959942638874,
+ "learning_rate": 0.0006,
+ "loss": 5.035224914550781,
+ "step": 682
+ },
+ {
+ "epoch": 9.489082969432314,
+ "grad_norm": 0.024985626339912415,
+ "learning_rate": 0.0006,
+ "loss": 5.049284934997559,
+ "step": 683
+ },
+ {
+ "epoch": 9.503056768558952,
+ "grad_norm": 0.023178115487098694,
+ "learning_rate": 0.0006,
+ "loss": 5.2266621589660645,
+ "step": 684
+ },
+ {
+ "epoch": 9.51703056768559,
+ "grad_norm": 0.0215512253344059,
+ "learning_rate": 0.0006,
+ "loss": 5.1875715255737305,
+ "step": 685
+ },
+ {
+ "epoch": 9.531004366812226,
+ "grad_norm": 0.01662837713956833,
+ "learning_rate": 0.0006,
+ "loss": 5.137012481689453,
+ "step": 686
+ },
+ {
+ "epoch": 9.544978165938865,
+ "grad_norm": 0.01952764391899109,
+ "learning_rate": 0.0006,
+ "loss": 5.072500228881836,
+ "step": 687
+ },
+ {
+ "epoch": 9.558951965065502,
+ "grad_norm": 0.017521077767014503,
+ "learning_rate": 0.0006,
+ "loss": 5.045711994171143,
+ "step": 688
+ },
+ {
+ "epoch": 9.57292576419214,
+ "grad_norm": 0.017355265095829964,
+ "learning_rate": 0.0006,
+ "loss": 5.02881383895874,
+ "step": 689
+ },
+ {
+ "epoch": 9.586899563318777,
+ "grad_norm": 0.015587719157338142,
+ "learning_rate": 0.0006,
+ "loss": 5.132787704467773,
+ "step": 690
+ },
+ {
+ "epoch": 9.600873362445414,
+ "grad_norm": 0.017469845712184906,
+ "learning_rate": 0.0006,
+ "loss": 5.105447292327881,
+ "step": 691
+ },
+ {
+ "epoch": 9.614847161572053,
+ "grad_norm": 0.019630100578069687,
+ "learning_rate": 0.0006,
+ "loss": 5.032464981079102,
+ "step": 692
+ },
+ {
+ "epoch": 9.62882096069869,
+ "grad_norm": 0.01784994639456272,
+ "learning_rate": 0.0006,
+ "loss": 5.081345558166504,
+ "step": 693
+ },
+ {
+ "epoch": 9.642794759825328,
+ "grad_norm": 0.01872754842042923,
+ "learning_rate": 0.0006,
+ "loss": 4.9772114753723145,
+ "step": 694
+ },
+ {
+ "epoch": 9.656768558951965,
+ "grad_norm": 0.01999032311141491,
+ "learning_rate": 0.0006,
+ "loss": 5.089740753173828,
+ "step": 695
+ },
+ {
+ "epoch": 9.670742358078602,
+ "grad_norm": 0.01710767112672329,
+ "learning_rate": 0.0006,
+ "loss": 4.9869890213012695,
+ "step": 696
+ },
+ {
+ "epoch": 9.68471615720524,
+ "grad_norm": 0.015144161880016327,
+ "learning_rate": 0.0006,
+ "loss": 5.022564888000488,
+ "step": 697
+ },
+ {
+ "epoch": 9.698689956331878,
+ "grad_norm": 0.01618543453514576,
+ "learning_rate": 0.0006,
+ "loss": 5.035921573638916,
+ "step": 698
+ },
+ {
+ "epoch": 9.712663755458514,
+ "grad_norm": 0.014653387479484081,
+ "learning_rate": 0.0006,
+ "loss": 4.914456367492676,
+ "step": 699
+ },
+ {
+ "epoch": 9.726637554585153,
+ "grad_norm": 0.013434977270662785,
+ "learning_rate": 0.0006,
+ "loss": 5.063764572143555,
+ "step": 700
+ },
+ {
+ "epoch": 9.74061135371179,
+ "grad_norm": 0.015128864906728268,
+ "learning_rate": 0.0006,
+ "loss": 5.052367687225342,
+ "step": 701
+ },
+ {
+ "epoch": 9.754585152838429,
+ "grad_norm": 0.014210768043994904,
+ "learning_rate": 0.0006,
+ "loss": 5.016489028930664,
+ "step": 702
+ },
+ {
+ "epoch": 9.768558951965066,
+ "grad_norm": 0.015191680751740932,
+ "learning_rate": 0.0006,
+ "loss": 4.990176200866699,
+ "step": 703
+ },
+ {
+ "epoch": 9.782532751091702,
+ "grad_norm": 0.018285127356648445,
+ "learning_rate": 0.0006,
+ "loss": 5.015164375305176,
+ "step": 704
+ },
+ {
+ "epoch": 9.796506550218341,
+ "grad_norm": 0.0204079058021307,
+ "learning_rate": 0.0006,
+ "loss": 5.091142654418945,
+ "step": 705
+ },
+ {
+ "epoch": 9.810480349344978,
+ "grad_norm": 0.01980586163699627,
+ "learning_rate": 0.0006,
+ "loss": 5.03350830078125,
+ "step": 706
+ },
+ {
+ "epoch": 9.824454148471617,
+ "grad_norm": 0.02048729732632637,
+ "learning_rate": 0.0006,
+ "loss": 5.03611421585083,
+ "step": 707
+ },
+ {
+ "epoch": 9.838427947598253,
+ "grad_norm": 0.02058163844048977,
+ "learning_rate": 0.0006,
+ "loss": 4.9613447189331055,
+ "step": 708
+ },
+ {
+ "epoch": 9.85240174672489,
+ "grad_norm": 0.018004924058914185,
+ "learning_rate": 0.0006,
+ "loss": 4.966541767120361,
+ "step": 709
+ },
+ {
+ "epoch": 9.866375545851529,
+ "grad_norm": 0.0198152307420969,
+ "learning_rate": 0.0006,
+ "loss": 5.01750373840332,
+ "step": 710
+ },
+ {
+ "epoch": 9.880349344978166,
+ "grad_norm": 0.018532969057559967,
+ "learning_rate": 0.0006,
+ "loss": 5.097424507141113,
+ "step": 711
+ },
+ {
+ "epoch": 9.894323144104803,
+ "grad_norm": 0.018939971923828125,
+ "learning_rate": 0.0006,
+ "loss": 5.177223205566406,
+ "step": 712
+ },
+ {
+ "epoch": 9.908296943231441,
+ "grad_norm": 0.017969170585274696,
+ "learning_rate": 0.0006,
+ "loss": 5.118013858795166,
+ "step": 713
+ },
+ {
+ "epoch": 9.922270742358078,
+ "grad_norm": 0.018021270632743835,
+ "learning_rate": 0.0006,
+ "loss": 5.034787178039551,
+ "step": 714
+ },
+ {
+ "epoch": 9.936244541484717,
+ "grad_norm": 0.017009710893034935,
+ "learning_rate": 0.0006,
+ "loss": 5.003920555114746,
+ "step": 715
+ },
+ {
+ "epoch": 9.950218340611354,
+ "grad_norm": 0.017722049728035927,
+ "learning_rate": 0.0006,
+ "loss": 5.031939506530762,
+ "step": 716
+ },
+ {
+ "epoch": 9.96419213973799,
+ "grad_norm": 0.017134087160229683,
+ "learning_rate": 0.0006,
+ "loss": 5.093474388122559,
+ "step": 717
+ },
+ {
+ "epoch": 9.97816593886463,
+ "grad_norm": 0.018511613830924034,
+ "learning_rate": 0.0006,
+ "loss": 5.078604698181152,
+ "step": 718
+ },
+ {
+ "epoch": 9.992139737991266,
+ "grad_norm": 0.021206334233283997,
+ "learning_rate": 0.0006,
+ "loss": 5.052572727203369,
+ "step": 719
+ },
+ {
+ "epoch": 10.0,
+ "grad_norm": 0.02239062264561653,
+ "learning_rate": 0.0006,
+ "loss": 5.063547134399414,
+ "step": 720
+ },
+ {
+ "epoch": 10.0,
+ "eval_loss": 5.173044681549072,
+ "eval_runtime": 56.7287,
+ "eval_samples_per_second": 43.047,
+ "eval_steps_per_second": 1.357,
+ "step": 720
+ },
+ {
+ "epoch": 10.013973799126637,
+ "grad_norm": 0.020757969468832016,
+ "learning_rate": 0.0006,
+ "loss": 5.045648574829102,
+ "step": 721
+ },
+ {
+ "epoch": 10.027947598253276,
+ "grad_norm": 0.022697702050209045,
+ "learning_rate": 0.0006,
+ "loss": 4.993833065032959,
+ "step": 722
+ },
+ {
+ "epoch": 10.041921397379912,
+ "grad_norm": 0.02434331737458706,
+ "learning_rate": 0.0006,
+ "loss": 5.073219299316406,
+ "step": 723
+ },
+ {
+ "epoch": 10.055895196506551,
+ "grad_norm": 0.025499660521745682,
+ "learning_rate": 0.0006,
+ "loss": 5.019072532653809,
+ "step": 724
+ },
+ {
+ "epoch": 10.069868995633188,
+ "grad_norm": 0.029281053692102432,
+ "learning_rate": 0.0006,
+ "loss": 4.959851264953613,
+ "step": 725
+ },
+ {
+ "epoch": 10.083842794759825,
+ "grad_norm": 0.033878110349178314,
+ "learning_rate": 0.0006,
+ "loss": 5.132717609405518,
+ "step": 726
+ },
+ {
+ "epoch": 10.097816593886463,
+ "grad_norm": 0.03385764732956886,
+ "learning_rate": 0.0006,
+ "loss": 5.0482892990112305,
+ "step": 727
+ },
+ {
+ "epoch": 10.1117903930131,
+ "grad_norm": 0.03466648980975151,
+ "learning_rate": 0.0006,
+ "loss": 4.946855545043945,
+ "step": 728
+ },
+ {
+ "epoch": 10.125764192139737,
+ "grad_norm": 0.028546305373311043,
+ "learning_rate": 0.0006,
+ "loss": 5.027774333953857,
+ "step": 729
+ },
+ {
+ "epoch": 10.139737991266376,
+ "grad_norm": 0.030575869604945183,
+ "learning_rate": 0.0006,
+ "loss": 4.9768476486206055,
+ "step": 730
+ },
+ {
+ "epoch": 10.153711790393013,
+ "grad_norm": 0.03614303469657898,
+ "learning_rate": 0.0006,
+ "loss": 5.078113555908203,
+ "step": 731
+ },
+ {
+ "epoch": 10.167685589519651,
+ "grad_norm": 0.040991660207509995,
+ "learning_rate": 0.0006,
+ "loss": 5.033185958862305,
+ "step": 732
+ },
+ {
+ "epoch": 10.181659388646288,
+ "grad_norm": 0.04551699757575989,
+ "learning_rate": 0.0006,
+ "loss": 5.0820746421813965,
+ "step": 733
+ },
+ {
+ "epoch": 10.195633187772925,
+ "grad_norm": 0.040193989872932434,
+ "learning_rate": 0.0006,
+ "loss": 5.059700965881348,
+ "step": 734
+ },
+ {
+ "epoch": 10.209606986899564,
+ "grad_norm": 0.035851605236530304,
+ "learning_rate": 0.0006,
+ "loss": 5.142467498779297,
+ "step": 735
+ },
+ {
+ "epoch": 10.2235807860262,
+ "grad_norm": 0.034200169146060944,
+ "learning_rate": 0.0006,
+ "loss": 5.072312355041504,
+ "step": 736
+ },
+ {
+ "epoch": 10.237554585152838,
+ "grad_norm": 0.033621110022068024,
+ "learning_rate": 0.0006,
+ "loss": 5.08037805557251,
+ "step": 737
+ },
+ {
+ "epoch": 10.251528384279476,
+ "grad_norm": 0.037372369319200516,
+ "learning_rate": 0.0006,
+ "loss": 4.998425483703613,
+ "step": 738
+ },
+ {
+ "epoch": 10.265502183406113,
+ "grad_norm": 0.037044707685709,
+ "learning_rate": 0.0006,
+ "loss": 4.923993110656738,
+ "step": 739
+ },
+ {
+ "epoch": 10.279475982532752,
+ "grad_norm": 0.030454808846116066,
+ "learning_rate": 0.0006,
+ "loss": 5.0889997482299805,
+ "step": 740
+ },
+ {
+ "epoch": 10.293449781659389,
+ "grad_norm": 0.030969638377428055,
+ "learning_rate": 0.0006,
+ "loss": 4.984112739562988,
+ "step": 741
+ },
+ {
+ "epoch": 10.307423580786025,
+ "grad_norm": 0.0276983380317688,
+ "learning_rate": 0.0006,
+ "loss": 4.955412864685059,
+ "step": 742
+ },
+ {
+ "epoch": 10.321397379912664,
+ "grad_norm": 0.02716052532196045,
+ "learning_rate": 0.0006,
+ "loss": 5.0098876953125,
+ "step": 743
+ },
+ {
+ "epoch": 10.335371179039301,
+ "grad_norm": 0.02403552085161209,
+ "learning_rate": 0.0006,
+ "loss": 5.165160655975342,
+ "step": 744
+ },
+ {
+ "epoch": 10.34934497816594,
+ "grad_norm": 0.0257862601429224,
+ "learning_rate": 0.0006,
+ "loss": 5.001547813415527,
+ "step": 745
+ },
+ {
+ "epoch": 10.363318777292577,
+ "grad_norm": 0.02300378680229187,
+ "learning_rate": 0.0006,
+ "loss": 5.038888931274414,
+ "step": 746
+ },
+ {
+ "epoch": 10.377292576419213,
+ "grad_norm": 0.019766854122281075,
+ "learning_rate": 0.0006,
+ "loss": 4.924291610717773,
+ "step": 747
+ },
+ {
+ "epoch": 10.391266375545852,
+ "grad_norm": 0.01888395845890045,
+ "learning_rate": 0.0006,
+ "loss": 5.059971332550049,
+ "step": 748
+ },
+ {
+ "epoch": 10.405240174672489,
+ "grad_norm": 0.015577499754726887,
+ "learning_rate": 0.0006,
+ "loss": 4.987686634063721,
+ "step": 749
+ },
+ {
+ "epoch": 10.419213973799126,
+ "grad_norm": 0.016012346372008324,
+ "learning_rate": 0.0006,
+ "loss": 5.015157699584961,
+ "step": 750
+ },
+ {
+ "epoch": 10.433187772925764,
+ "grad_norm": 0.016277998685836792,
+ "learning_rate": 0.0006,
+ "loss": 5.004931449890137,
+ "step": 751
+ },
+ {
+ "epoch": 10.447161572052401,
+ "grad_norm": 0.017504561692476273,
+ "learning_rate": 0.0006,
+ "loss": 4.902827739715576,
+ "step": 752
+ },
+ {
+ "epoch": 10.46113537117904,
+ "grad_norm": 0.01569022797048092,
+ "learning_rate": 0.0006,
+ "loss": 4.954863548278809,
+ "step": 753
+ },
+ {
+ "epoch": 10.475109170305677,
+ "grad_norm": 0.014114422723650932,
+ "learning_rate": 0.0006,
+ "loss": 5.00656795501709,
+ "step": 754
+ },
+ {
+ "epoch": 10.489082969432314,
+ "grad_norm": 0.01573832333087921,
+ "learning_rate": 0.0006,
+ "loss": 4.978353023529053,
+ "step": 755
+ },
+ {
+ "epoch": 10.503056768558952,
+ "grad_norm": 0.016480036079883575,
+ "learning_rate": 0.0006,
+ "loss": 4.953327655792236,
+ "step": 756
+ },
+ {
+ "epoch": 10.51703056768559,
+ "grad_norm": 0.014008025638759136,
+ "learning_rate": 0.0006,
+ "loss": 5.050840377807617,
+ "step": 757
+ },
+ {
+ "epoch": 10.531004366812226,
+ "grad_norm": 0.013147998601198196,
+ "learning_rate": 0.0006,
+ "loss": 4.974964141845703,
+ "step": 758
+ },
+ {
+ "epoch": 10.544978165938865,
+ "grad_norm": 0.013807603158056736,
+ "learning_rate": 0.0006,
+ "loss": 4.927907943725586,
+ "step": 759
+ },
+ {
+ "epoch": 10.558951965065502,
+ "grad_norm": 0.013555224984884262,
+ "learning_rate": 0.0006,
+ "loss": 4.980690956115723,
+ "step": 760
+ },
+ {
+ "epoch": 10.57292576419214,
+ "grad_norm": 0.013045408762991428,
+ "learning_rate": 0.0006,
+ "loss": 4.9436116218566895,
+ "step": 761
+ },
+ {
+ "epoch": 10.586899563318777,
+ "grad_norm": 0.012858862057328224,
+ "learning_rate": 0.0006,
+ "loss": 4.9595770835876465,
+ "step": 762
+ },
+ {
+ "epoch": 10.600873362445414,
+ "grad_norm": 0.01468253880739212,
+ "learning_rate": 0.0006,
+ "loss": 4.90725040435791,
+ "step": 763
+ },
+ {
+ "epoch": 10.614847161572053,
+ "grad_norm": 0.013224679045379162,
+ "learning_rate": 0.0006,
+ "loss": 4.877615928649902,
+ "step": 764
+ },
+ {
+ "epoch": 10.62882096069869,
+ "grad_norm": 0.013555348850786686,
+ "learning_rate": 0.0006,
+ "loss": 4.9456610679626465,
+ "step": 765
+ },
+ {
+ "epoch": 10.642794759825328,
+ "grad_norm": 0.014387160539627075,
+ "learning_rate": 0.0006,
+ "loss": 4.9233717918396,
+ "step": 766
+ },
+ {
+ "epoch": 10.656768558951965,
+ "grad_norm": 0.014354088343679905,
+ "learning_rate": 0.0006,
+ "loss": 4.983511924743652,
+ "step": 767
+ },
+ {
+ "epoch": 10.670742358078602,
+ "grad_norm": 0.01332700252532959,
+ "learning_rate": 0.0006,
+ "loss": 4.934416770935059,
+ "step": 768
+ },
+ {
+ "epoch": 10.68471615720524,
+ "grad_norm": 0.011543313041329384,
+ "learning_rate": 0.0006,
+ "loss": 4.96388053894043,
+ "step": 769
+ },
+ {
+ "epoch": 10.698689956331878,
+ "grad_norm": 0.012330878525972366,
+ "learning_rate": 0.0006,
+ "loss": 4.946396350860596,
+ "step": 770
+ },
+ {
+ "epoch": 10.712663755458514,
+ "grad_norm": 0.01379779726266861,
+ "learning_rate": 0.0006,
+ "loss": 4.981019020080566,
+ "step": 771
+ },
+ {
+ "epoch": 10.726637554585153,
+ "grad_norm": 0.014612431637942791,
+ "learning_rate": 0.0006,
+ "loss": 4.894567966461182,
+ "step": 772
+ },
+ {
+ "epoch": 10.74061135371179,
+ "grad_norm": 0.013062899932265282,
+ "learning_rate": 0.0006,
+ "loss": 5.033038139343262,
+ "step": 773
+ },
+ {
+ "epoch": 10.754585152838429,
+ "grad_norm": 0.013898049481213093,
+ "learning_rate": 0.0006,
+ "loss": 4.823636054992676,
+ "step": 774
+ },
+ {
+ "epoch": 10.768558951965066,
+ "grad_norm": 0.01502019353210926,
+ "learning_rate": 0.0006,
+ "loss": 5.020156383514404,
+ "step": 775
+ },
+ {
+ "epoch": 10.782532751091702,
+ "grad_norm": 0.013810204342007637,
+ "learning_rate": 0.0006,
+ "loss": 4.935647010803223,
+ "step": 776
+ },
+ {
+ "epoch": 10.796506550218341,
+ "grad_norm": 0.013616513460874557,
+ "learning_rate": 0.0006,
+ "loss": 4.901486396789551,
+ "step": 777
+ },
+ {
+ "epoch": 10.810480349344978,
+ "grad_norm": 0.013601024635136127,
+ "learning_rate": 0.0006,
+ "loss": 4.931873321533203,
+ "step": 778
+ },
+ {
+ "epoch": 10.824454148471617,
+ "grad_norm": 0.013668350875377655,
+ "learning_rate": 0.0006,
+ "loss": 4.937256813049316,
+ "step": 779
+ },
+ {
+ "epoch": 10.838427947598253,
+ "grad_norm": 0.01581823080778122,
+ "learning_rate": 0.0006,
+ "loss": 4.9611406326293945,
+ "step": 780
+ },
+ {
+ "epoch": 10.85240174672489,
+ "grad_norm": 0.019478484988212585,
+ "learning_rate": 0.0006,
+ "loss": 4.9212236404418945,
+ "step": 781
+ },
+ {
+ "epoch": 10.866375545851529,
+ "grad_norm": 0.026921333745121956,
+ "learning_rate": 0.0006,
+ "loss": 4.88501501083374,
+ "step": 782
+ },
+ {
+ "epoch": 10.880349344978166,
+ "grad_norm": 0.03155914694070816,
+ "learning_rate": 0.0006,
+ "loss": 4.989967346191406,
+ "step": 783
+ },
+ {
+ "epoch": 10.894323144104803,
+ "grad_norm": 0.029689429327845573,
+ "learning_rate": 0.0006,
+ "loss": 4.936000823974609,
+ "step": 784
+ },
+ {
+ "epoch": 10.908296943231441,
+ "grad_norm": 0.026626063510775566,
+ "learning_rate": 0.0006,
+ "loss": 4.984785079956055,
+ "step": 785
+ },
+ {
+ "epoch": 10.922270742358078,
+ "grad_norm": 0.026168212294578552,
+ "learning_rate": 0.0006,
+ "loss": 4.908282279968262,
+ "step": 786
+ },
+ {
+ "epoch": 10.936244541484717,
+ "grad_norm": 0.02828158251941204,
+ "learning_rate": 0.0006,
+ "loss": 4.926390647888184,
+ "step": 787
+ },
+ {
+ "epoch": 10.950218340611354,
+ "grad_norm": 0.02649078331887722,
+ "learning_rate": 0.0006,
+ "loss": 5.00400972366333,
+ "step": 788
+ },
+ {
+ "epoch": 10.96419213973799,
+ "grad_norm": 0.0256856270134449,
+ "learning_rate": 0.0006,
+ "loss": 5.0317487716674805,
+ "step": 789
+ },
+ {
+ "epoch": 10.97816593886463,
+ "grad_norm": 0.024728883057832718,
+ "learning_rate": 0.0006,
+ "loss": 4.9299635887146,
+ "step": 790
+ },
+ {
+ "epoch": 10.992139737991266,
+ "grad_norm": 0.026241116225719452,
+ "learning_rate": 0.0006,
+ "loss": 4.947319984436035,
+ "step": 791
+ },
+ {
+ "epoch": 11.0,
+ "grad_norm": 0.023526743054389954,
+ "learning_rate": 0.0006,
+ "loss": 4.94429349899292,
+ "step": 792
+ },
+ {
+ "epoch": 11.0,
+ "eval_loss": 5.104345798492432,
+ "eval_runtime": 57.0955,
+ "eval_samples_per_second": 42.77,
+ "eval_steps_per_second": 1.349,
+ "step": 792
+ },
+ {
+ "epoch": 11.013973799126637,
+ "grad_norm": 0.0216788612306118,
+ "learning_rate": 0.0006,
+ "loss": 4.881950855255127,
+ "step": 793
+ },
+ {
+ "epoch": 11.027947598253276,
+ "grad_norm": 0.01830463856458664,
+ "learning_rate": 0.0006,
+ "loss": 4.971016883850098,
+ "step": 794
+ },
+ {
+ "epoch": 11.041921397379912,
+ "grad_norm": 0.02204521745443344,
+ "learning_rate": 0.0006,
+ "loss": 4.894623756408691,
+ "step": 795
+ },
+ {
+ "epoch": 11.055895196506551,
+ "grad_norm": 0.0210255216807127,
+ "learning_rate": 0.0006,
+ "loss": 4.834009170532227,
+ "step": 796
+ },
+ {
+ "epoch": 11.069868995633188,
+ "grad_norm": 0.019797448068857193,
+ "learning_rate": 0.0006,
+ "loss": 5.086714744567871,
+ "step": 797
+ },
+ {
+ "epoch": 11.083842794759825,
+ "grad_norm": 0.02036641724407673,
+ "learning_rate": 0.0006,
+ "loss": 4.920056343078613,
+ "step": 798
+ },
+ {
+ "epoch": 11.097816593886463,
+ "grad_norm": 0.01966066285967827,
+ "learning_rate": 0.0006,
+ "loss": 4.879859447479248,
+ "step": 799
+ },
+ {
+ "epoch": 11.1117903930131,
+ "grad_norm": 0.0202149897813797,
+ "learning_rate": 0.0006,
+ "loss": 4.933640480041504,
+ "step": 800
+ },
+ {
+ "epoch": 11.125764192139737,
+ "grad_norm": 0.01893039606511593,
+ "learning_rate": 0.0006,
+ "loss": 4.901970863342285,
+ "step": 801
+ },
+ {
+ "epoch": 11.139737991266376,
+ "grad_norm": 0.016040155664086342,
+ "learning_rate": 0.0006,
+ "loss": 4.905491828918457,
+ "step": 802
+ },
+ {
+ "epoch": 11.153711790393013,
+ "grad_norm": 0.01720350608229637,
+ "learning_rate": 0.0006,
+ "loss": 4.798816680908203,
+ "step": 803
+ },
+ {
+ "epoch": 11.167685589519651,
+ "grad_norm": 0.017002852633595467,
+ "learning_rate": 0.0006,
+ "loss": 4.851859092712402,
+ "step": 804
+ },
+ {
+ "epoch": 11.181659388646288,
+ "grad_norm": 0.015594680793583393,
+ "learning_rate": 0.0006,
+ "loss": 4.90328311920166,
+ "step": 805
+ },
+ {
+ "epoch": 11.195633187772925,
+ "grad_norm": 0.014904462732374668,
+ "learning_rate": 0.0006,
+ "loss": 4.904215335845947,
+ "step": 806
+ },
+ {
+ "epoch": 11.209606986899564,
+ "grad_norm": 0.01366228237748146,
+ "learning_rate": 0.0006,
+ "loss": 4.9813055992126465,
+ "step": 807
+ },
+ {
+ "epoch": 11.2235807860262,
+ "grad_norm": 0.014719638973474503,
+ "learning_rate": 0.0006,
+ "loss": 4.927306175231934,
+ "step": 808
+ },
+ {
+ "epoch": 11.237554585152838,
+ "grad_norm": 0.015016036108136177,
+ "learning_rate": 0.0006,
+ "loss": 4.874215126037598,
+ "step": 809
+ },
+ {
+ "epoch": 11.251528384279476,
+ "grad_norm": 0.014925424009561539,
+ "learning_rate": 0.0006,
+ "loss": 4.838257789611816,
+ "step": 810
+ },
+ {
+ "epoch": 11.265502183406113,
+ "grad_norm": 0.014335823245346546,
+ "learning_rate": 0.0006,
+ "loss": 4.940608024597168,
+ "step": 811
+ },
+ {
+ "epoch": 11.279475982532752,
+ "grad_norm": 0.014987512491643429,
+ "learning_rate": 0.0006,
+ "loss": 4.831531524658203,
+ "step": 812
+ },
+ {
+ "epoch": 11.293449781659389,
+ "grad_norm": 0.016087457537651062,
+ "learning_rate": 0.0006,
+ "loss": 4.863227367401123,
+ "step": 813
+ },
+ {
+ "epoch": 11.307423580786025,
+ "grad_norm": 0.01671300269663334,
+ "learning_rate": 0.0006,
+ "loss": 4.8535661697387695,
+ "step": 814
+ },
+ {
+ "epoch": 11.321397379912664,
+ "grad_norm": 0.015229142270982265,
+ "learning_rate": 0.0006,
+ "loss": 4.948635101318359,
+ "step": 815
+ },
+ {
+ "epoch": 11.335371179039301,
+ "grad_norm": 0.01532587967813015,
+ "learning_rate": 0.0006,
+ "loss": 4.9035797119140625,
+ "step": 816
+ },
+ {
+ "epoch": 11.34934497816594,
+ "grad_norm": 0.015715476125478745,
+ "learning_rate": 0.0006,
+ "loss": 4.894709587097168,
+ "step": 817
+ },
+ {
+ "epoch": 11.363318777292577,
+ "grad_norm": 0.014156977646052837,
+ "learning_rate": 0.0006,
+ "loss": 4.960206985473633,
+ "step": 818
+ },
+ {
+ "epoch": 11.377292576419213,
+ "grad_norm": 0.017076566815376282,
+ "learning_rate": 0.0006,
+ "loss": 5.005850791931152,
+ "step": 819
+ },
+ {
+ "epoch": 11.391266375545852,
+ "grad_norm": 0.0172084029763937,
+ "learning_rate": 0.0006,
+ "loss": 4.903355598449707,
+ "step": 820
+ },
+ {
+ "epoch": 11.405240174672489,
+ "grad_norm": 0.0180258359760046,
+ "learning_rate": 0.0006,
+ "loss": 4.842667579650879,
+ "step": 821
+ },
+ {
+ "epoch": 11.419213973799126,
+ "grad_norm": 0.022130120545625687,
+ "learning_rate": 0.0006,
+ "loss": 4.799970626831055,
+ "step": 822
+ },
+ {
+ "epoch": 11.433187772925764,
+ "grad_norm": 0.02384897693991661,
+ "learning_rate": 0.0006,
+ "loss": 4.860746383666992,
+ "step": 823
+ },
+ {
+ "epoch": 11.447161572052401,
+ "grad_norm": 0.021881572902202606,
+ "learning_rate": 0.0006,
+ "loss": 4.918107032775879,
+ "step": 824
+ },
+ {
+ "epoch": 11.46113537117904,
+ "grad_norm": 0.023085080087184906,
+ "learning_rate": 0.0006,
+ "loss": 4.878946304321289,
+ "step": 825
+ },
+ {
+ "epoch": 11.475109170305677,
+ "grad_norm": 0.02291012555360794,
+ "learning_rate": 0.0006,
+ "loss": 4.958887100219727,
+ "step": 826
+ },
+ {
+ "epoch": 11.489082969432314,
+ "grad_norm": 0.0199968870729208,
+ "learning_rate": 0.0006,
+ "loss": 4.80955696105957,
+ "step": 827
+ },
+ {
+ "epoch": 11.503056768558952,
+ "grad_norm": 0.018702229484915733,
+ "learning_rate": 0.0006,
+ "loss": 5.0276408195495605,
+ "step": 828
+ },
+ {
+ "epoch": 11.51703056768559,
+ "grad_norm": 0.017787311226129532,
+ "learning_rate": 0.0006,
+ "loss": 4.96018123626709,
+ "step": 829
+ },
+ {
+ "epoch": 11.531004366812226,
+ "grad_norm": 0.01704501546919346,
+ "learning_rate": 0.0006,
+ "loss": 4.914553165435791,
+ "step": 830
+ },
+ {
+ "epoch": 11.544978165938865,
+ "grad_norm": 0.017519677057862282,
+ "learning_rate": 0.0006,
+ "loss": 4.798098087310791,
+ "step": 831
+ },
+ {
+ "epoch": 11.558951965065502,
+ "grad_norm": 0.018511991947889328,
+ "learning_rate": 0.0006,
+ "loss": 4.817785263061523,
+ "step": 832
+ },
+ {
+ "epoch": 11.57292576419214,
+ "grad_norm": 0.01876644790172577,
+ "learning_rate": 0.0006,
+ "loss": 4.818995475769043,
+ "step": 833
+ },
+ {
+ "epoch": 11.586899563318777,
+ "grad_norm": 0.01753568835556507,
+ "learning_rate": 0.0006,
+ "loss": 4.811938285827637,
+ "step": 834
+ },
+ {
+ "epoch": 11.600873362445414,
+ "grad_norm": 0.017046531662344933,
+ "learning_rate": 0.0006,
+ "loss": 4.914674282073975,
+ "step": 835
+ },
+ {
+ "epoch": 11.614847161572053,
+ "grad_norm": 0.01802094094455242,
+ "learning_rate": 0.0006,
+ "loss": 4.993284225463867,
+ "step": 836
+ },
+ {
+ "epoch": 11.62882096069869,
+ "grad_norm": 0.019493764266371727,
+ "learning_rate": 0.0006,
+ "loss": 4.895628452301025,
+ "step": 837
+ },
+ {
+ "epoch": 11.642794759825328,
+ "grad_norm": 0.019461210817098618,
+ "learning_rate": 0.0006,
+ "loss": 4.869269371032715,
+ "step": 838
+ },
+ {
+ "epoch": 11.656768558951965,
+ "grad_norm": 0.01981567218899727,
+ "learning_rate": 0.0006,
+ "loss": 4.863052845001221,
+ "step": 839
+ },
+ {
+ "epoch": 11.670742358078602,
+ "grad_norm": 0.019894491881132126,
+ "learning_rate": 0.0006,
+ "loss": 4.8775787353515625,
+ "step": 840
+ },
+ {
+ "epoch": 11.68471615720524,
+ "grad_norm": 0.01875786855816841,
+ "learning_rate": 0.0006,
+ "loss": 4.817763328552246,
+ "step": 841
+ },
+ {
+ "epoch": 11.698689956331878,
+ "grad_norm": 0.020926695317029953,
+ "learning_rate": 0.0006,
+ "loss": 4.7470526695251465,
+ "step": 842
+ },
+ {
+ "epoch": 11.712663755458514,
+ "grad_norm": 0.021477103233337402,
+ "learning_rate": 0.0006,
+ "loss": 4.762682914733887,
+ "step": 843
+ },
+ {
+ "epoch": 11.726637554585153,
+ "grad_norm": 0.020628711208701134,
+ "learning_rate": 0.0006,
+ "loss": 4.807985782623291,
+ "step": 844
+ },
+ {
+ "epoch": 11.74061135371179,
+ "grad_norm": 0.021489612758159637,
+ "learning_rate": 0.0006,
+ "loss": 4.843091011047363,
+ "step": 845
+ },
+ {
+ "epoch": 11.754585152838429,
+ "grad_norm": 0.021194491535425186,
+ "learning_rate": 0.0006,
+ "loss": 4.8300251960754395,
+ "step": 846
+ },
+ {
+ "epoch": 11.768558951965066,
+ "grad_norm": 0.01833350397646427,
+ "learning_rate": 0.0006,
+ "loss": 5.000202178955078,
+ "step": 847
+ },
+ {
+ "epoch": 11.782532751091702,
+ "grad_norm": 0.018096931278705597,
+ "learning_rate": 0.0006,
+ "loss": 4.880648136138916,
+ "step": 848
+ },
+ {
+ "epoch": 11.796506550218341,
+ "grad_norm": 0.0197161752730608,
+ "learning_rate": 0.0006,
+ "loss": 4.788316249847412,
+ "step": 849
+ },
+ {
+ "epoch": 11.810480349344978,
+ "grad_norm": 0.0203181654214859,
+ "learning_rate": 0.0006,
+ "loss": 4.870944976806641,
+ "step": 850
+ },
+ {
+ "epoch": 11.824454148471617,
+ "grad_norm": 0.01908455230295658,
+ "learning_rate": 0.0006,
+ "loss": 4.890144348144531,
+ "step": 851
+ },
+ {
+ "epoch": 11.838427947598253,
+ "grad_norm": 0.021281801164150238,
+ "learning_rate": 0.0006,
+ "loss": 4.89578914642334,
+ "step": 852
+ },
+ {
+ "epoch": 11.85240174672489,
+ "grad_norm": 0.02124273031949997,
+ "learning_rate": 0.0006,
+ "loss": 4.883749008178711,
+ "step": 853
+ },
+ {
+ "epoch": 11.866375545851529,
+ "grad_norm": 0.020969398319721222,
+ "learning_rate": 0.0006,
+ "loss": 4.809142589569092,
+ "step": 854
+ },
+ {
+ "epoch": 11.880349344978166,
+ "grad_norm": 0.01967657171189785,
+ "learning_rate": 0.0006,
+ "loss": 4.821864604949951,
+ "step": 855
+ },
+ {
+ "epoch": 11.894323144104803,
+ "grad_norm": 0.020887333899736404,
+ "learning_rate": 0.0006,
+ "loss": 4.917985916137695,
+ "step": 856
+ },
+ {
+ "epoch": 11.908296943231441,
+ "grad_norm": 0.019203083589673042,
+ "learning_rate": 0.0006,
+ "loss": 4.824960708618164,
+ "step": 857
+ },
+ {
+ "epoch": 11.922270742358078,
+ "grad_norm": 0.020162401720881462,
+ "learning_rate": 0.0006,
+ "loss": 4.772617816925049,
+ "step": 858
+ },
+ {
+ "epoch": 11.936244541484717,
+ "grad_norm": 0.02251487225294113,
+ "learning_rate": 0.0006,
+ "loss": 4.735418319702148,
+ "step": 859
+ },
+ {
+ "epoch": 11.950218340611354,
+ "grad_norm": 0.020313527435064316,
+ "learning_rate": 0.0006,
+ "loss": 4.939233779907227,
+ "step": 860
+ },
+ {
+ "epoch": 11.96419213973799,
+ "grad_norm": 0.021635599434375763,
+ "learning_rate": 0.0006,
+ "loss": 4.861807346343994,
+ "step": 861
+ },
+ {
+ "epoch": 11.97816593886463,
+ "grad_norm": 0.02285311557352543,
+ "learning_rate": 0.0006,
+ "loss": 4.932732582092285,
+ "step": 862
+ },
+ {
+ "epoch": 11.992139737991266,
+ "grad_norm": 0.02072535641491413,
+ "learning_rate": 0.0006,
+ "loss": 4.796767711639404,
+ "step": 863
+ },
+ {
+ "epoch": 12.0,
+ "grad_norm": 0.020588304847478867,
+ "learning_rate": 0.0006,
+ "loss": 4.997537612915039,
+ "step": 864
+ },
+ {
+ "epoch": 12.0,
+ "eval_loss": 5.044651985168457,
+ "eval_runtime": 56.9145,
+ "eval_samples_per_second": 42.906,
+ "eval_steps_per_second": 1.353,
+ "step": 864
+ },
+ {
+ "epoch": 12.013973799126637,
+ "grad_norm": 0.017597166821360588,
+ "learning_rate": 0.0006,
+ "loss": 4.830524444580078,
+ "step": 865
+ },
+ {
+ "epoch": 12.027947598253276,
+ "grad_norm": 0.01716788113117218,
+ "learning_rate": 0.0006,
+ "loss": 4.894325256347656,
+ "step": 866
+ },
+ {
+ "epoch": 12.041921397379912,
+ "grad_norm": 0.017712457105517387,
+ "learning_rate": 0.0006,
+ "loss": 4.868297576904297,
+ "step": 867
+ },
+ {
+ "epoch": 12.055895196506551,
+ "grad_norm": 0.021486390382051468,
+ "learning_rate": 0.0006,
+ "loss": 4.922049522399902,
+ "step": 868
+ },
+ {
+ "epoch": 12.069868995633188,
+ "grad_norm": 0.024560654535889626,
+ "learning_rate": 0.0006,
+ "loss": 4.734241008758545,
+ "step": 869
+ },
+ {
+ "epoch": 12.083842794759825,
+ "grad_norm": 0.025870252400636673,
+ "learning_rate": 0.0006,
+ "loss": 4.868566989898682,
+ "step": 870
+ },
+ {
+ "epoch": 12.097816593886463,
+ "grad_norm": 0.0277389045804739,
+ "learning_rate": 0.0006,
+ "loss": 4.755006790161133,
+ "step": 871
+ },
+ {
+ "epoch": 12.1117903930131,
+ "grad_norm": 0.029100844636559486,
+ "learning_rate": 0.0006,
+ "loss": 4.909667015075684,
+ "step": 872
+ },
+ {
+ "epoch": 12.125764192139737,
+ "grad_norm": 0.031833067536354065,
+ "learning_rate": 0.0006,
+ "loss": 4.858674049377441,
+ "step": 873
+ },
+ {
+ "epoch": 12.139737991266376,
+ "grad_norm": 0.032709237188100815,
+ "learning_rate": 0.0006,
+ "loss": 4.823696136474609,
+ "step": 874
+ },
+ {
+ "epoch": 12.153711790393013,
+ "grad_norm": 0.02995380386710167,
+ "learning_rate": 0.0006,
+ "loss": 4.850235462188721,
+ "step": 875
+ },
+ {
+ "epoch": 12.167685589519651,
+ "grad_norm": 0.034098610281944275,
+ "learning_rate": 0.0006,
+ "loss": 4.8169989585876465,
+ "step": 876
+ },
+ {
+ "epoch": 12.181659388646288,
+ "grad_norm": 0.03377080336213112,
+ "learning_rate": 0.0006,
+ "loss": 4.90738582611084,
+ "step": 877
+ },
+ {
+ "epoch": 12.195633187772925,
+ "grad_norm": 0.029780825600028038,
+ "learning_rate": 0.0006,
+ "loss": 4.811163902282715,
+ "step": 878
+ },
+ {
+ "epoch": 12.209606986899564,
+ "grad_norm": 0.03118024580180645,
+ "learning_rate": 0.0006,
+ "loss": 4.861286640167236,
+ "step": 879
+ },
+ {
+ "epoch": 12.2235807860262,
+ "grad_norm": 0.033052004873752594,
+ "learning_rate": 0.0006,
+ "loss": 4.865021228790283,
+ "step": 880
+ },
+ {
+ "epoch": 12.237554585152838,
+ "grad_norm": 0.029070032760500908,
+ "learning_rate": 0.0006,
+ "loss": 4.947303771972656,
+ "step": 881
+ },
+ {
+ "epoch": 12.251528384279476,
+ "grad_norm": 0.028355715796351433,
+ "learning_rate": 0.0006,
+ "loss": 4.804648399353027,
+ "step": 882
+ },
+ {
+ "epoch": 12.265502183406113,
+ "grad_norm": 0.031257227063179016,
+ "learning_rate": 0.0006,
+ "loss": 4.757718086242676,
+ "step": 883
+ },
+ {
+ "epoch": 12.279475982532752,
+ "grad_norm": 0.031438373029232025,
+ "learning_rate": 0.0006,
+ "loss": 4.851268768310547,
+ "step": 884
+ },
+ {
+ "epoch": 12.293449781659389,
+ "grad_norm": 0.030492138117551804,
+ "learning_rate": 0.0006,
+ "loss": 4.7891035079956055,
+ "step": 885
+ },
+ {
+ "epoch": 12.307423580786025,
+ "grad_norm": 0.026115991175174713,
+ "learning_rate": 0.0006,
+ "loss": 4.86474609375,
+ "step": 886
+ },
+ {
+ "epoch": 12.321397379912664,
+ "grad_norm": 0.024184465408325195,
+ "learning_rate": 0.0006,
+ "loss": 4.860456943511963,
+ "step": 887
+ },
+ {
+ "epoch": 12.335371179039301,
+ "grad_norm": 0.025437982752919197,
+ "learning_rate": 0.0006,
+ "loss": 4.806251049041748,
+ "step": 888
+ },
+ {
+ "epoch": 12.34934497816594,
+ "grad_norm": 0.022390231490135193,
+ "learning_rate": 0.0006,
+ "loss": 4.807086944580078,
+ "step": 889
+ },
+ {
+ "epoch": 12.363318777292577,
+ "grad_norm": 0.020174294710159302,
+ "learning_rate": 0.0006,
+ "loss": 4.78292179107666,
+ "step": 890
+ },
+ {
+ "epoch": 12.377292576419213,
+ "grad_norm": 0.019734324887394905,
+ "learning_rate": 0.0006,
+ "loss": 4.784509658813477,
+ "step": 891
+ },
+ {
+ "epoch": 12.391266375545852,
+ "grad_norm": 0.016733523458242416,
+ "learning_rate": 0.0006,
+ "loss": 4.811755180358887,
+ "step": 892
+ },
+ {
+ "epoch": 12.405240174672489,
+ "grad_norm": 0.01570146717131138,
+ "learning_rate": 0.0006,
+ "loss": 4.720088005065918,
+ "step": 893
+ },
+ {
+ "epoch": 12.419213973799126,
+ "grad_norm": 0.01601393334567547,
+ "learning_rate": 0.0006,
+ "loss": 4.783543109893799,
+ "step": 894
+ },
+ {
+ "epoch": 12.433187772925764,
+ "grad_norm": 0.014035487547516823,
+ "learning_rate": 0.0006,
+ "loss": 4.848340034484863,
+ "step": 895
+ },
+ {
+ "epoch": 12.447161572052401,
+ "grad_norm": 0.015248487703502178,
+ "learning_rate": 0.0006,
+ "loss": 4.773287773132324,
+ "step": 896
+ },
+ {
+ "epoch": 12.46113537117904,
+ "grad_norm": 0.01532268151640892,
+ "learning_rate": 0.0006,
+ "loss": 4.799646854400635,
+ "step": 897
+ },
+ {
+ "epoch": 12.475109170305677,
+ "grad_norm": 0.013982534408569336,
+ "learning_rate": 0.0006,
+ "loss": 4.8090500831604,
+ "step": 898
+ },
+ {
+ "epoch": 12.489082969432314,
+ "grad_norm": 0.014081164263188839,
+ "learning_rate": 0.0006,
+ "loss": 4.751888751983643,
+ "step": 899
+ },
+ {
+ "epoch": 12.503056768558952,
+ "grad_norm": 0.01380133256316185,
+ "learning_rate": 0.0006,
+ "loss": 4.78144645690918,
+ "step": 900
+ },
+ {
+ "epoch": 12.51703056768559,
+ "grad_norm": 0.013555269688367844,
+ "learning_rate": 0.0006,
+ "loss": 4.779257774353027,
+ "step": 901
+ },
+ {
+ "epoch": 12.531004366812226,
+ "grad_norm": 0.013075701892375946,
+ "learning_rate": 0.0006,
+ "loss": 4.820705413818359,
+ "step": 902
+ },
+ {
+ "epoch": 12.544978165938865,
+ "grad_norm": 0.01276316400617361,
+ "learning_rate": 0.0006,
+ "loss": 4.836733818054199,
+ "step": 903
+ },
+ {
+ "epoch": 12.558951965065502,
+ "grad_norm": 0.013033892959356308,
+ "learning_rate": 0.0006,
+ "loss": 4.764178276062012,
+ "step": 904
+ },
+ {
+ "epoch": 12.57292576419214,
+ "grad_norm": 0.013996980153024197,
+ "learning_rate": 0.0006,
+ "loss": 4.767983436584473,
+ "step": 905
+ },
+ {
+ "epoch": 12.586899563318777,
+ "grad_norm": 0.015222184360027313,
+ "learning_rate": 0.0006,
+ "loss": 4.759721755981445,
+ "step": 906
+ },
+ {
+ "epoch": 12.600873362445414,
+ "grad_norm": 0.01623941957950592,
+ "learning_rate": 0.0006,
+ "loss": 4.849811553955078,
+ "step": 907
+ },
+ {
+ "epoch": 12.614847161572053,
+ "grad_norm": 0.015711182728409767,
+ "learning_rate": 0.0006,
+ "loss": 4.7829084396362305,
+ "step": 908
+ },
+ {
+ "epoch": 12.62882096069869,
+ "grad_norm": 0.014137467369437218,
+ "learning_rate": 0.0006,
+ "loss": 4.73321533203125,
+ "step": 909
+ },
+ {
+ "epoch": 12.642794759825328,
+ "grad_norm": 0.012738462537527084,
+ "learning_rate": 0.0006,
+ "loss": 4.870052337646484,
+ "step": 910
+ },
+ {
+ "epoch": 12.656768558951965,
+ "grad_norm": 0.013683130033314228,
+ "learning_rate": 0.0006,
+ "loss": 4.82658052444458,
+ "step": 911
+ },
+ {
+ "epoch": 12.670742358078602,
+ "grad_norm": 0.014409830793738365,
+ "learning_rate": 0.0006,
+ "loss": 4.707494735717773,
+ "step": 912
+ },
+ {
+ "epoch": 12.68471615720524,
+ "grad_norm": 0.013806809671223164,
+ "learning_rate": 0.0006,
+ "loss": 4.778928279876709,
+ "step": 913
+ },
+ {
+ "epoch": 12.698689956331878,
+ "grad_norm": 0.012857912108302116,
+ "learning_rate": 0.0006,
+ "loss": 4.890717506408691,
+ "step": 914
+ },
+ {
+ "epoch": 12.712663755458514,
+ "grad_norm": 0.013828758150339127,
+ "learning_rate": 0.0006,
+ "loss": 4.8357439041137695,
+ "step": 915
+ },
+ {
+ "epoch": 12.726637554585153,
+ "grad_norm": 0.01478351280093193,
+ "learning_rate": 0.0006,
+ "loss": 4.869986534118652,
+ "step": 916
+ },
+ {
+ "epoch": 12.74061135371179,
+ "grad_norm": 0.01326421182602644,
+ "learning_rate": 0.0006,
+ "loss": 4.830646514892578,
+ "step": 917
+ },
+ {
+ "epoch": 12.754585152838429,
+ "grad_norm": 0.013335433788597584,
+ "learning_rate": 0.0006,
+ "loss": 4.7786149978637695,
+ "step": 918
+ },
+ {
+ "epoch": 12.768558951965066,
+ "grad_norm": 0.014261147007346153,
+ "learning_rate": 0.0006,
+ "loss": 4.788693428039551,
+ "step": 919
+ },
+ {
+ "epoch": 12.782532751091702,
+ "grad_norm": 0.014274783432483673,
+ "learning_rate": 0.0006,
+ "loss": 4.77295446395874,
+ "step": 920
+ },
+ {
+ "epoch": 12.796506550218341,
+ "grad_norm": 0.015074674971401691,
+ "learning_rate": 0.0006,
+ "loss": 4.830754280090332,
+ "step": 921
+ },
+ {
+ "epoch": 12.810480349344978,
+ "grad_norm": 0.015599515289068222,
+ "learning_rate": 0.0006,
+ "loss": 4.837759017944336,
+ "step": 922
+ },
+ {
+ "epoch": 12.824454148471617,
+ "grad_norm": 0.014595242217183113,
+ "learning_rate": 0.0006,
+ "loss": 4.882965564727783,
+ "step": 923
+ },
+ {
+ "epoch": 12.838427947598253,
+ "grad_norm": 0.014951467514038086,
+ "learning_rate": 0.0006,
+ "loss": 4.709748268127441,
+ "step": 924
+ },
+ {
+ "epoch": 12.85240174672489,
+ "grad_norm": 0.01539295818656683,
+ "learning_rate": 0.0006,
+ "loss": 4.791160583496094,
+ "step": 925
+ },
+ {
+ "epoch": 12.866375545851529,
+ "grad_norm": 0.015987906605005264,
+ "learning_rate": 0.0006,
+ "loss": 4.661876678466797,
+ "step": 926
+ },
+ {
+ "epoch": 12.880349344978166,
+ "grad_norm": 0.016527341678738594,
+ "learning_rate": 0.0006,
+ "loss": 4.721952438354492,
+ "step": 927
+ },
+ {
+ "epoch": 12.894323144104803,
+ "grad_norm": 0.01637161523103714,
+ "learning_rate": 0.0006,
+ "loss": 4.704446792602539,
+ "step": 928
+ },
+ {
+ "epoch": 12.908296943231441,
+ "grad_norm": 0.015221050009131432,
+ "learning_rate": 0.0006,
+ "loss": 4.771907806396484,
+ "step": 929
+ },
+ {
+ "epoch": 12.922270742358078,
+ "grad_norm": 0.016242166981101036,
+ "learning_rate": 0.0006,
+ "loss": 4.806267738342285,
+ "step": 930
+ },
+ {
+ "epoch": 12.936244541484717,
+ "grad_norm": 0.01628357358276844,
+ "learning_rate": 0.0006,
+ "loss": 4.7252326011657715,
+ "step": 931
+ },
+ {
+ "epoch": 12.950218340611354,
+ "grad_norm": 0.017428133636713028,
+ "learning_rate": 0.0006,
+ "loss": 4.765689849853516,
+ "step": 932
+ },
+ {
+ "epoch": 12.96419213973799,
+ "grad_norm": 0.01870061084628105,
+ "learning_rate": 0.0006,
+ "loss": 4.814488887786865,
+ "step": 933
+ },
+ {
+ "epoch": 12.97816593886463,
+ "grad_norm": 0.019796060398221016,
+ "learning_rate": 0.0006,
+ "loss": 4.739185810089111,
+ "step": 934
+ },
+ {
+ "epoch": 12.992139737991266,
+ "grad_norm": 0.01985820196568966,
+ "learning_rate": 0.0006,
+ "loss": 4.810608863830566,
+ "step": 935
+ },
+ {
+ "epoch": 13.0,
+ "grad_norm": 0.020834477618336678,
+ "learning_rate": 0.0006,
+ "loss": 4.727858066558838,
+ "step": 936
+ },
+ {
+ "epoch": 13.0,
+ "eval_loss": 4.979116916656494,
+ "eval_runtime": 57.2877,
+ "eval_samples_per_second": 42.627,
+ "eval_steps_per_second": 1.344,
+ "step": 936
+ },
+ {
+ "epoch": 13.013973799126637,
+ "grad_norm": 0.02025519125163555,
+ "learning_rate": 0.0006,
+ "loss": 4.628875732421875,
+ "step": 937
+ },
+ {
+ "epoch": 13.027947598253276,
+ "grad_norm": 0.019596368074417114,
+ "learning_rate": 0.0006,
+ "loss": 4.694699287414551,
+ "step": 938
+ },
+ {
+ "epoch": 13.041921397379912,
+ "grad_norm": 0.020494934171438217,
+ "learning_rate": 0.0006,
+ "loss": 4.691821098327637,
+ "step": 939
+ },
+ {
+ "epoch": 13.055895196506551,
+ "grad_norm": 0.020121760666370392,
+ "learning_rate": 0.0006,
+ "loss": 4.71478271484375,
+ "step": 940
+ },
+ {
+ "epoch": 13.069868995633188,
+ "grad_norm": 0.021091405302286148,
+ "learning_rate": 0.0006,
+ "loss": 4.695915222167969,
+ "step": 941
+ },
+ {
+ "epoch": 13.083842794759825,
+ "grad_norm": 0.019787423312664032,
+ "learning_rate": 0.0006,
+ "loss": 4.8652448654174805,
+ "step": 942
+ },
+ {
+ "epoch": 13.097816593886463,
+ "grad_norm": 0.0185833927243948,
+ "learning_rate": 0.0006,
+ "loss": 4.7065887451171875,
+ "step": 943
+ },
+ {
+ "epoch": 13.1117903930131,
+ "grad_norm": 0.021220475435256958,
+ "learning_rate": 0.0006,
+ "loss": 4.741232395172119,
+ "step": 944
+ },
+ {
+ "epoch": 13.125764192139737,
+ "grad_norm": 0.02136622555553913,
+ "learning_rate": 0.0006,
+ "loss": 4.705967903137207,
+ "step": 945
+ },
+ {
+ "epoch": 13.139737991266376,
+ "grad_norm": 0.0191587433218956,
+ "learning_rate": 0.0006,
+ "loss": 4.758667945861816,
+ "step": 946
+ },
+ {
+ "epoch": 13.153711790393013,
+ "grad_norm": 0.018290022388100624,
+ "learning_rate": 0.0006,
+ "loss": 4.760666847229004,
+ "step": 947
+ },
+ {
+ "epoch": 13.167685589519651,
+ "grad_norm": 0.021704409271478653,
+ "learning_rate": 0.0006,
+ "loss": 4.750555992126465,
+ "step": 948
+ },
+ {
+ "epoch": 13.181659388646288,
+ "grad_norm": 0.021668918430805206,
+ "learning_rate": 0.0006,
+ "loss": 4.840903282165527,
+ "step": 949
+ },
+ {
+ "epoch": 13.195633187772925,
+ "grad_norm": 0.019497141242027283,
+ "learning_rate": 0.0006,
+ "loss": 4.738969326019287,
+ "step": 950
+ },
+ {
+ "epoch": 13.209606986899564,
+ "grad_norm": 0.018500596284866333,
+ "learning_rate": 0.0006,
+ "loss": 4.75154972076416,
+ "step": 951
+ },
+ {
+ "epoch": 13.2235807860262,
+ "grad_norm": 0.01872239261865616,
+ "learning_rate": 0.0006,
+ "loss": 4.811738014221191,
+ "step": 952
+ },
+ {
+ "epoch": 13.237554585152838,
+ "grad_norm": 0.020321547985076904,
+ "learning_rate": 0.0006,
+ "loss": 4.815384387969971,
+ "step": 953
+ },
+ {
+ "epoch": 13.251528384279476,
+ "grad_norm": 0.02283575013279915,
+ "learning_rate": 0.0006,
+ "loss": 4.74582052230835,
+ "step": 954
+ },
+ {
+ "epoch": 13.265502183406113,
+ "grad_norm": 0.02474398910999298,
+ "learning_rate": 0.0006,
+ "loss": 4.738375186920166,
+ "step": 955
+ },
+ {
+ "epoch": 13.279475982532752,
+ "grad_norm": 0.02065764181315899,
+ "learning_rate": 0.0006,
+ "loss": 4.746875286102295,
+ "step": 956
+ },
+ {
+ "epoch": 13.293449781659389,
+ "grad_norm": 0.020534520968794823,
+ "learning_rate": 0.0006,
+ "loss": 4.656914710998535,
+ "step": 957
+ },
+ {
+ "epoch": 13.307423580786025,
+ "grad_norm": 0.0206378772854805,
+ "learning_rate": 0.0006,
+ "loss": 4.842615604400635,
+ "step": 958
+ },
+ {
+ "epoch": 13.321397379912664,
+ "grad_norm": 0.017935654148459435,
+ "learning_rate": 0.0006,
+ "loss": 4.717827796936035,
+ "step": 959
+ },
+ {
+ "epoch": 13.335371179039301,
+ "grad_norm": 0.016958363354206085,
+ "learning_rate": 0.0006,
+ "loss": 4.681070327758789,
+ "step": 960
+ },
+ {
+ "epoch": 13.34934497816594,
+ "grad_norm": 0.015735412016510963,
+ "learning_rate": 0.0006,
+ "loss": 4.7957048416137695,
+ "step": 961
+ },
+ {
+ "epoch": 13.363318777292577,
+ "grad_norm": 0.01563699170947075,
+ "learning_rate": 0.0006,
+ "loss": 4.701192378997803,
+ "step": 962
+ },
+ {
+ "epoch": 13.377292576419213,
+ "grad_norm": 0.015207760035991669,
+ "learning_rate": 0.0006,
+ "loss": 4.73256254196167,
+ "step": 963
+ },
+ {
+ "epoch": 13.391266375545852,
+ "grad_norm": 0.015396883711218834,
+ "learning_rate": 0.0006,
+ "loss": 4.76412296295166,
+ "step": 964
+ },
+ {
+ "epoch": 13.405240174672489,
+ "grad_norm": 0.015690583735704422,
+ "learning_rate": 0.0006,
+ "loss": 4.791157245635986,
+ "step": 965
+ },
+ {
+ "epoch": 13.419213973799126,
+ "grad_norm": 0.01755724661052227,
+ "learning_rate": 0.0006,
+ "loss": 4.845666885375977,
+ "step": 966
+ },
+ {
+ "epoch": 13.433187772925764,
+ "grad_norm": 0.019443422555923462,
+ "learning_rate": 0.0006,
+ "loss": 4.5902299880981445,
+ "step": 967
+ },
+ {
+ "epoch": 13.447161572052401,
+ "grad_norm": 0.02118338644504547,
+ "learning_rate": 0.0006,
+ "loss": 4.740627288818359,
+ "step": 968
+ },
+ {
+ "epoch": 13.46113537117904,
+ "grad_norm": 0.02174839936196804,
+ "learning_rate": 0.0006,
+ "loss": 4.814560890197754,
+ "step": 969
+ },
+ {
+ "epoch": 13.475109170305677,
+ "grad_norm": 0.022860374301671982,
+ "learning_rate": 0.0006,
+ "loss": 4.66440486907959,
+ "step": 970
+ },
+ {
+ "epoch": 13.489082969432314,
+ "grad_norm": 0.024742677807807922,
+ "learning_rate": 0.0006,
+ "loss": 4.7451653480529785,
+ "step": 971
+ },
+ {
+ "epoch": 13.503056768558952,
+ "grad_norm": 0.027068907395005226,
+ "learning_rate": 0.0006,
+ "loss": 4.802584171295166,
+ "step": 972
+ },
+ {
+ "epoch": 13.51703056768559,
+ "grad_norm": 0.026316625997424126,
+ "learning_rate": 0.0006,
+ "loss": 4.842397689819336,
+ "step": 973
+ },
+ {
+ "epoch": 13.531004366812226,
+ "grad_norm": 0.024610962718725204,
+ "learning_rate": 0.0006,
+ "loss": 4.650721073150635,
+ "step": 974
+ },
+ {
+ "epoch": 13.544978165938865,
+ "grad_norm": 0.024122603237628937,
+ "learning_rate": 0.0006,
+ "loss": 4.6984357833862305,
+ "step": 975
+ },
+ {
+ "epoch": 13.558951965065502,
+ "grad_norm": 0.022699708119034767,
+ "learning_rate": 0.0006,
+ "loss": 4.793007850646973,
+ "step": 976
+ },
+ {
+ "epoch": 13.57292576419214,
+ "grad_norm": 0.021706534549593925,
+ "learning_rate": 0.0006,
+ "loss": 4.691362380981445,
+ "step": 977
+ },
+ {
+ "epoch": 13.586899563318777,
+ "grad_norm": 0.021363291889429092,
+ "learning_rate": 0.0006,
+ "loss": 4.76762580871582,
+ "step": 978
+ },
+ {
+ "epoch": 13.600873362445414,
+ "grad_norm": 0.023476677015423775,
+ "learning_rate": 0.0006,
+ "loss": 4.777202606201172,
+ "step": 979
+ },
+ {
+ "epoch": 13.614847161572053,
+ "grad_norm": 0.02457100711762905,
+ "learning_rate": 0.0006,
+ "loss": 4.590703010559082,
+ "step": 980
+ },
+ {
+ "epoch": 13.62882096069869,
+ "grad_norm": 0.02469099499285221,
+ "learning_rate": 0.0006,
+ "loss": 4.640185356140137,
+ "step": 981
+ },
+ {
+ "epoch": 13.642794759825328,
+ "grad_norm": 0.024914603680372238,
+ "learning_rate": 0.0006,
+ "loss": 4.772201061248779,
+ "step": 982
+ },
+ {
+ "epoch": 13.656768558951965,
+ "grad_norm": 0.025133173912763596,
+ "learning_rate": 0.0006,
+ "loss": 4.743337154388428,
+ "step": 983
+ },
+ {
+ "epoch": 13.670742358078602,
+ "grad_norm": 0.02034137211740017,
+ "learning_rate": 0.0006,
+ "loss": 4.752127647399902,
+ "step": 984
+ },
+ {
+ "epoch": 13.68471615720524,
+ "grad_norm": 0.019645661115646362,
+ "learning_rate": 0.0006,
+ "loss": 4.838453769683838,
+ "step": 985
+ },
+ {
+ "epoch": 13.698689956331878,
+ "grad_norm": 0.01900297962129116,
+ "learning_rate": 0.0006,
+ "loss": 4.804868221282959,
+ "step": 986
+ },
+ {
+ "epoch": 13.712663755458514,
+ "grad_norm": 0.020292341709136963,
+ "learning_rate": 0.0006,
+ "loss": 4.609743118286133,
+ "step": 987
+ },
+ {
+ "epoch": 13.726637554585153,
+ "grad_norm": 0.017014775425195694,
+ "learning_rate": 0.0006,
+ "loss": 4.740612983703613,
+ "step": 988
+ },
+ {
+ "epoch": 13.74061135371179,
+ "grad_norm": 0.015199844725430012,
+ "learning_rate": 0.0006,
+ "loss": 4.719701766967773,
+ "step": 989
+ },
+ {
+ "epoch": 13.754585152838429,
+ "grad_norm": 0.015125414356589317,
+ "learning_rate": 0.0006,
+ "loss": 4.686114311218262,
+ "step": 990
+ },
+ {
+ "epoch": 13.768558951965066,
+ "grad_norm": 0.014779540710151196,
+ "learning_rate": 0.0006,
+ "loss": 4.674347877502441,
+ "step": 991
+ },
+ {
+ "epoch": 13.782532751091702,
+ "grad_norm": 0.015255163423717022,
+ "learning_rate": 0.0006,
+ "loss": 4.6309309005737305,
+ "step": 992
+ },
+ {
+ "epoch": 13.796506550218341,
+ "grad_norm": 0.013692197389900684,
+ "learning_rate": 0.0006,
+ "loss": 4.698497295379639,
+ "step": 993
+ },
+ {
+ "epoch": 13.810480349344978,
+ "grad_norm": 0.014529623091220856,
+ "learning_rate": 0.0006,
+ "loss": 4.628443717956543,
+ "step": 994
+ },
+ {
+ "epoch": 13.824454148471617,
+ "grad_norm": 0.01576659083366394,
+ "learning_rate": 0.0006,
+ "loss": 4.788015842437744,
+ "step": 995
+ },
+ {
+ "epoch": 13.838427947598253,
+ "grad_norm": 0.015834277495741844,
+ "learning_rate": 0.0006,
+ "loss": 4.672703742980957,
+ "step": 996
+ },
+ {
+ "epoch": 13.85240174672489,
+ "grad_norm": 0.015539503656327724,
+ "learning_rate": 0.0006,
+ "loss": 4.731503486633301,
+ "step": 997
+ },
+ {
+ "epoch": 13.866375545851529,
+ "grad_norm": 0.013463746756315231,
+ "learning_rate": 0.0006,
+ "loss": 4.7006330490112305,
+ "step": 998
+ },
+ {
+ "epoch": 13.880349344978166,
+ "grad_norm": 0.016377726569771767,
+ "learning_rate": 0.0006,
+ "loss": 4.636074066162109,
+ "step": 999
+ },
+ {
+ "epoch": 13.894323144104803,
+ "grad_norm": 0.016660034656524658,
+ "learning_rate": 0.0006,
+ "loss": 4.742693901062012,
+ "step": 1000
+ },
+ {
+ "epoch": 13.908296943231441,
+ "grad_norm": 0.014902938157320023,
+ "learning_rate": 0.0006,
+ "loss": 4.769811630249023,
+ "step": 1001
+ },
+ {
+ "epoch": 13.922270742358078,
+ "grad_norm": 0.014883476309478283,
+ "learning_rate": 0.0006,
+ "loss": 4.731159210205078,
+ "step": 1002
+ },
+ {
+ "epoch": 13.936244541484717,
+ "grad_norm": 0.014494217932224274,
+ "learning_rate": 0.0006,
+ "loss": 4.659485816955566,
+ "step": 1003
+ },
+ {
+ "epoch": 13.950218340611354,
+ "grad_norm": 0.015426991507411003,
+ "learning_rate": 0.0006,
+ "loss": 4.755229949951172,
+ "step": 1004
+ },
+ {
+ "epoch": 13.96419213973799,
+ "grad_norm": 0.016648586839437485,
+ "learning_rate": 0.0006,
+ "loss": 4.74111270904541,
+ "step": 1005
+ },
+ {
+ "epoch": 13.97816593886463,
+ "grad_norm": 0.016980541869997978,
+ "learning_rate": 0.0006,
+ "loss": 4.74072265625,
+ "step": 1006
+ },
+ {
+ "epoch": 13.992139737991266,
+ "grad_norm": 0.018070152029395103,
+ "learning_rate": 0.0006,
+ "loss": 4.732975006103516,
+ "step": 1007
+ },
+ {
+ "epoch": 14.0,
+ "grad_norm": 0.019910665228962898,
+ "learning_rate": 0.0006,
+ "loss": 4.634697914123535,
+ "step": 1008
+ },
+ {
+ "epoch": 14.0,
+ "eval_loss": 4.850640296936035,
+ "eval_runtime": 57.3039,
+ "eval_samples_per_second": 42.615,
+ "eval_steps_per_second": 1.344,
+ "step": 1008
+ },
+ {
+ "epoch": 14.013973799126637,
+ "grad_norm": 0.01980053074657917,
+ "learning_rate": 0.0006,
+ "loss": 4.628744602203369,
+ "step": 1009
+ },
+ {
+ "epoch": 14.027947598253276,
+ "grad_norm": 0.02497623674571514,
+ "learning_rate": 0.0006,
+ "loss": 4.75930118560791,
+ "step": 1010
+ },
+ {
+ "epoch": 14.041921397379912,
+ "grad_norm": 0.02794097363948822,
+ "learning_rate": 0.0006,
+ "loss": 4.691340446472168,
+ "step": 1011
+ },
+ {
+ "epoch": 14.055895196506551,
+ "grad_norm": 0.023725535720586777,
+ "learning_rate": 0.0006,
+ "loss": 4.733664512634277,
+ "step": 1012
+ },
+ {
+ "epoch": 14.069868995633188,
+ "grad_norm": 0.022966474294662476,
+ "learning_rate": 0.0006,
+ "loss": 4.664808750152588,
+ "step": 1013
+ },
+ {
+ "epoch": 14.083842794759825,
+ "grad_norm": 0.02530563250184059,
+ "learning_rate": 0.0006,
+ "loss": 4.8047285079956055,
+ "step": 1014
+ },
+ {
+ "epoch": 14.097816593886463,
+ "grad_norm": 0.02491082064807415,
+ "learning_rate": 0.0006,
+ "loss": 4.667956352233887,
+ "step": 1015
+ },
+ {
+ "epoch": 14.1117903930131,
+ "grad_norm": 0.022110294550657272,
+ "learning_rate": 0.0006,
+ "loss": 4.666140556335449,
+ "step": 1016
+ },
+ {
+ "epoch": 14.125764192139737,
+ "grad_norm": 0.02065849117934704,
+ "learning_rate": 0.0006,
+ "loss": 4.581960678100586,
+ "step": 1017
+ },
+ {
+ "epoch": 14.139737991266376,
+ "grad_norm": 0.01941298507153988,
+ "learning_rate": 0.0006,
+ "loss": 4.614226341247559,
+ "step": 1018
+ },
+ {
+ "epoch": 14.153711790393013,
+ "grad_norm": 0.022441856563091278,
+ "learning_rate": 0.0006,
+ "loss": 4.638352394104004,
+ "step": 1019
+ },
+ {
+ "epoch": 14.167685589519651,
+ "grad_norm": 0.021519694477319717,
+ "learning_rate": 0.0006,
+ "loss": 4.700160980224609,
+ "step": 1020
+ },
+ {
+ "epoch": 14.181659388646288,
+ "grad_norm": 0.018956486135721207,
+ "learning_rate": 0.0006,
+ "loss": 4.649593353271484,
+ "step": 1021
+ },
+ {
+ "epoch": 14.195633187772925,
+ "grad_norm": 0.016681140288710594,
+ "learning_rate": 0.0006,
+ "loss": 4.603952407836914,
+ "step": 1022
+ },
+ {
+ "epoch": 14.209606986899564,
+ "grad_norm": 0.01788809522986412,
+ "learning_rate": 0.0006,
+ "loss": 4.669160842895508,
+ "step": 1023
+ },
+ {
+ "epoch": 14.2235807860262,
+ "grad_norm": 0.016047228127717972,
+ "learning_rate": 0.0006,
+ "loss": 4.6367950439453125,
+ "step": 1024
+ },
+ {
+ "epoch": 14.237554585152838,
+ "grad_norm": 0.01614239625632763,
+ "learning_rate": 0.0006,
+ "loss": 4.664529323577881,
+ "step": 1025
+ },
+ {
+ "epoch": 14.251528384279476,
+ "grad_norm": 0.016362549737095833,
+ "learning_rate": 0.0006,
+ "loss": 4.702373504638672,
+ "step": 1026
+ },
+ {
+ "epoch": 14.265502183406113,
+ "grad_norm": 0.015800409018993378,
+ "learning_rate": 0.0006,
+ "loss": 4.7594428062438965,
+ "step": 1027
+ },
+ {
+ "epoch": 14.279475982532752,
+ "grad_norm": 0.016730139032006264,
+ "learning_rate": 0.0006,
+ "loss": 4.677244186401367,
+ "step": 1028
+ },
+ {
+ "epoch": 14.293449781659389,
+ "grad_norm": 0.01661759801208973,
+ "learning_rate": 0.0006,
+ "loss": 4.633679389953613,
+ "step": 1029
+ },
+ {
+ "epoch": 14.307423580786025,
+ "grad_norm": 0.017673347145318985,
+ "learning_rate": 0.0006,
+ "loss": 4.649107456207275,
+ "step": 1030
+ },
+ {
+ "epoch": 14.321397379912664,
+ "grad_norm": 0.019243352115154266,
+ "learning_rate": 0.0006,
+ "loss": 4.611257553100586,
+ "step": 1031
+ },
+ {
+ "epoch": 14.335371179039301,
+ "grad_norm": 0.019817164167761803,
+ "learning_rate": 0.0006,
+ "loss": 4.588086128234863,
+ "step": 1032
+ },
+ {
+ "epoch": 14.34934497816594,
+ "grad_norm": 0.022292211651802063,
+ "learning_rate": 0.0006,
+ "loss": 4.783066749572754,
+ "step": 1033
+ },
+ {
+ "epoch": 14.363318777292577,
+ "grad_norm": 0.02140096202492714,
+ "learning_rate": 0.0006,
+ "loss": 4.604718208312988,
+ "step": 1034
+ },
+ {
+ "epoch": 14.377292576419213,
+ "grad_norm": 0.0187817495316267,
+ "learning_rate": 0.0006,
+ "loss": 4.721773147583008,
+ "step": 1035
+ },
+ {
+ "epoch": 14.391266375545852,
+ "grad_norm": 0.019879210740327835,
+ "learning_rate": 0.0006,
+ "loss": 4.789286136627197,
+ "step": 1036
+ },
+ {
+ "epoch": 14.405240174672489,
+ "grad_norm": 0.01885785162448883,
+ "learning_rate": 0.0006,
+ "loss": 4.669743537902832,
+ "step": 1037
+ },
+ {
+ "epoch": 14.419213973799126,
+ "grad_norm": 0.018324270844459534,
+ "learning_rate": 0.0006,
+ "loss": 4.65901517868042,
+ "step": 1038
+ },
+ {
+ "epoch": 14.433187772925764,
+ "grad_norm": 0.017664821818470955,
+ "learning_rate": 0.0006,
+ "loss": 4.667839527130127,
+ "step": 1039
+ },
+ {
+ "epoch": 14.447161572052401,
+ "grad_norm": 0.015702905133366585,
+ "learning_rate": 0.0006,
+ "loss": 4.64700984954834,
+ "step": 1040
+ },
+ {
+ "epoch": 14.46113537117904,
+ "grad_norm": 0.016674194484949112,
+ "learning_rate": 0.0006,
+ "loss": 4.694093227386475,
+ "step": 1041
+ },
+ {
+ "epoch": 14.475109170305677,
+ "grad_norm": 0.016869477927684784,
+ "learning_rate": 0.0006,
+ "loss": 4.619544982910156,
+ "step": 1042
+ },
+ {
+ "epoch": 14.489082969432314,
+ "grad_norm": 0.016557540744543076,
+ "learning_rate": 0.0006,
+ "loss": 4.568667411804199,
+ "step": 1043
+ },
+ {
+ "epoch": 14.503056768558952,
+ "grad_norm": 0.014500945806503296,
+ "learning_rate": 0.0006,
+ "loss": 4.561741828918457,
+ "step": 1044
+ },
+ {
+ "epoch": 14.51703056768559,
+ "grad_norm": 0.014228739775717258,
+ "learning_rate": 0.0006,
+ "loss": 4.654277324676514,
+ "step": 1045
+ },
+ {
+ "epoch": 14.531004366812226,
+ "grad_norm": 0.013894238509237766,
+ "learning_rate": 0.0006,
+ "loss": 4.5867719650268555,
+ "step": 1046
+ },
+ {
+ "epoch": 14.544978165938865,
+ "grad_norm": 0.01550068985670805,
+ "learning_rate": 0.0006,
+ "loss": 4.61168098449707,
+ "step": 1047
+ },
+ {
+ "epoch": 14.558951965065502,
+ "grad_norm": 0.015940772369503975,
+ "learning_rate": 0.0006,
+ "loss": 4.691247940063477,
+ "step": 1048
+ },
+ {
+ "epoch": 14.57292576419214,
+ "grad_norm": 0.01714535616338253,
+ "learning_rate": 0.0006,
+ "loss": 4.61183500289917,
+ "step": 1049
+ },
+ {
+ "epoch": 14.586899563318777,
+ "grad_norm": 0.017925038933753967,
+ "learning_rate": 0.0006,
+ "loss": 4.719022750854492,
+ "step": 1050
+ },
+ {
+ "epoch": 14.600873362445414,
+ "grad_norm": 0.018141552805900574,
+ "learning_rate": 0.0006,
+ "loss": 4.760281562805176,
+ "step": 1051
+ },
+ {
+ "epoch": 14.614847161572053,
+ "grad_norm": 0.017131328582763672,
+ "learning_rate": 0.0006,
+ "loss": 4.709680080413818,
+ "step": 1052
+ },
+ {
+ "epoch": 14.62882096069869,
+ "grad_norm": 0.018730048090219498,
+ "learning_rate": 0.0006,
+ "loss": 4.5994720458984375,
+ "step": 1053
+ },
+ {
+ "epoch": 14.642794759825328,
+ "grad_norm": 0.018894298002123833,
+ "learning_rate": 0.0006,
+ "loss": 4.669070720672607,
+ "step": 1054
+ },
+ {
+ "epoch": 14.656768558951965,
+ "grad_norm": 0.017643198370933533,
+ "learning_rate": 0.0006,
+ "loss": 4.641790390014648,
+ "step": 1055
+ },
+ {
+ "epoch": 14.670742358078602,
+ "grad_norm": 0.017647244036197662,
+ "learning_rate": 0.0006,
+ "loss": 4.653899192810059,
+ "step": 1056
+ },
+ {
+ "epoch": 14.68471615720524,
+ "grad_norm": 0.017304198816418648,
+ "learning_rate": 0.0006,
+ "loss": 4.7487473487854,
+ "step": 1057
+ },
+ {
+ "epoch": 14.698689956331878,
+ "grad_norm": 0.01638447679579258,
+ "learning_rate": 0.0006,
+ "loss": 4.71945858001709,
+ "step": 1058
+ },
+ {
+ "epoch": 14.712663755458514,
+ "grad_norm": 0.015892071649432182,
+ "learning_rate": 0.0006,
+ "loss": 4.744903564453125,
+ "step": 1059
+ },
+ {
+ "epoch": 14.726637554585153,
+ "grad_norm": 0.017420614138245583,
+ "learning_rate": 0.0006,
+ "loss": 4.781563758850098,
+ "step": 1060
+ },
+ {
+ "epoch": 14.74061135371179,
+ "grad_norm": 0.016498520970344543,
+ "learning_rate": 0.0006,
+ "loss": 4.622779846191406,
+ "step": 1061
+ },
+ {
+ "epoch": 14.754585152838429,
+ "grad_norm": 0.017482127994298935,
+ "learning_rate": 0.0006,
+ "loss": 4.62825870513916,
+ "step": 1062
+ },
+ {
+ "epoch": 14.768558951965066,
+ "grad_norm": 0.019018379971385002,
+ "learning_rate": 0.0006,
+ "loss": 4.693194389343262,
+ "step": 1063
+ },
+ {
+ "epoch": 14.782532751091702,
+ "grad_norm": 0.01993914321064949,
+ "learning_rate": 0.0006,
+ "loss": 4.742903709411621,
+ "step": 1064
+ },
+ {
+ "epoch": 14.796506550218341,
+ "grad_norm": 0.02301335707306862,
+ "learning_rate": 0.0006,
+ "loss": 4.613251686096191,
+ "step": 1065
+ },
+ {
+ "epoch": 14.810480349344978,
+ "grad_norm": 0.024088222533464432,
+ "learning_rate": 0.0006,
+ "loss": 4.594305992126465,
+ "step": 1066
+ },
+ {
+ "epoch": 14.824454148471617,
+ "grad_norm": 0.02493320405483246,
+ "learning_rate": 0.0006,
+ "loss": 4.700588226318359,
+ "step": 1067
+ },
+ {
+ "epoch": 14.838427947598253,
+ "grad_norm": 0.02501937560737133,
+ "learning_rate": 0.0006,
+ "loss": 4.601742744445801,
+ "step": 1068
+ },
+ {
+ "epoch": 14.85240174672489,
+ "grad_norm": 0.023490751162171364,
+ "learning_rate": 0.0006,
+ "loss": 4.589211463928223,
+ "step": 1069
+ },
+ {
+ "epoch": 14.866375545851529,
+ "grad_norm": 0.019695665687322617,
+ "learning_rate": 0.0006,
+ "loss": 4.635081768035889,
+ "step": 1070
+ },
+ {
+ "epoch": 14.880349344978166,
+ "grad_norm": 0.019276980310678482,
+ "learning_rate": 0.0006,
+ "loss": 4.737002372741699,
+ "step": 1071
+ },
+ {
+ "epoch": 14.894323144104803,
+ "grad_norm": 0.022430511191487312,
+ "learning_rate": 0.0006,
+ "loss": 4.808295726776123,
+ "step": 1072
+ },
+ {
+ "epoch": 14.908296943231441,
+ "grad_norm": 0.021543916314840317,
+ "learning_rate": 0.0006,
+ "loss": 4.559475421905518,
+ "step": 1073
+ },
+ {
+ "epoch": 14.922270742358078,
+ "grad_norm": 0.019885241985321045,
+ "learning_rate": 0.0006,
+ "loss": 4.647948741912842,
+ "step": 1074
+ },
+ {
+ "epoch": 14.936244541484717,
+ "grad_norm": 0.02115175500512123,
+ "learning_rate": 0.0006,
+ "loss": 4.680505752563477,
+ "step": 1075
+ },
+ {
+ "epoch": 14.950218340611354,
+ "grad_norm": 0.02156151458621025,
+ "learning_rate": 0.0006,
+ "loss": 4.673477649688721,
+ "step": 1076
+ },
+ {
+ "epoch": 14.96419213973799,
+ "grad_norm": 0.02184601128101349,
+ "learning_rate": 0.0006,
+ "loss": 4.644402027130127,
+ "step": 1077
+ },
+ {
+ "epoch": 14.97816593886463,
+ "grad_norm": 0.0211983360350132,
+ "learning_rate": 0.0006,
+ "loss": 4.821558475494385,
+ "step": 1078
+ },
+ {
+ "epoch": 14.992139737991266,
+ "grad_norm": 0.019636567682027817,
+ "learning_rate": 0.0006,
+ "loss": 4.578868865966797,
+ "step": 1079
+ },
+ {
+ "epoch": 15.0,
+ "grad_norm": 0.019997350871562958,
+ "learning_rate": 0.0006,
+ "loss": 4.691307544708252,
+ "step": 1080
+ },
+ {
+ "epoch": 15.0,
+ "eval_loss": 4.872478008270264,
+ "eval_runtime": 56.8513,
+ "eval_samples_per_second": 42.954,
+ "eval_steps_per_second": 1.354,
+ "step": 1080
+ },
+ {
+ "epoch": 15.013973799126637,
+ "grad_norm": 0.018830662593245506,
+ "learning_rate": 0.0006,
+ "loss": 4.761779308319092,
+ "step": 1081
+ },
+ {
+ "epoch": 15.027947598253276,
+ "grad_norm": 0.016876481473445892,
+ "learning_rate": 0.0006,
+ "loss": 4.61148738861084,
+ "step": 1082
+ },
+ {
+ "epoch": 15.041921397379912,
+ "grad_norm": 0.015550863929092884,
+ "learning_rate": 0.0006,
+ "loss": 4.676124095916748,
+ "step": 1083
+ },
+ {
+ "epoch": 15.055895196506551,
+ "grad_norm": 0.017178896814584732,
+ "learning_rate": 0.0006,
+ "loss": 4.588935852050781,
+ "step": 1084
+ },
+ {
+ "epoch": 15.069868995633188,
+ "grad_norm": 0.015465447679162025,
+ "learning_rate": 0.0006,
+ "loss": 4.605499267578125,
+ "step": 1085
+ },
+ {
+ "epoch": 15.083842794759825,
+ "grad_norm": 0.01606195978820324,
+ "learning_rate": 0.0006,
+ "loss": 4.723373889923096,
+ "step": 1086
+ },
+ {
+ "epoch": 15.097816593886463,
+ "grad_norm": 0.016521936282515526,
+ "learning_rate": 0.0006,
+ "loss": 4.646608352661133,
+ "step": 1087
+ },
+ {
+ "epoch": 15.1117903930131,
+ "grad_norm": 0.01589205674827099,
+ "learning_rate": 0.0006,
+ "loss": 4.498138427734375,
+ "step": 1088
+ },
+ {
+ "epoch": 15.125764192139737,
+ "grad_norm": 0.01524856686592102,
+ "learning_rate": 0.0006,
+ "loss": 4.594562530517578,
+ "step": 1089
+ },
+ {
+ "epoch": 15.139737991266376,
+ "grad_norm": 0.014714641496539116,
+ "learning_rate": 0.0006,
+ "loss": 4.630744934082031,
+ "step": 1090
+ },
+ {
+ "epoch": 15.153711790393013,
+ "grad_norm": 0.015333331190049648,
+ "learning_rate": 0.0006,
+ "loss": 4.682388782501221,
+ "step": 1091
+ },
+ {
+ "epoch": 15.167685589519651,
+ "grad_norm": 0.01550073828548193,
+ "learning_rate": 0.0006,
+ "loss": 4.692652702331543,
+ "step": 1092
+ },
+ {
+ "epoch": 15.181659388646288,
+ "grad_norm": 0.014783318154513836,
+ "learning_rate": 0.0006,
+ "loss": 4.6223835945129395,
+ "step": 1093
+ },
+ {
+ "epoch": 15.195633187772925,
+ "grad_norm": 0.014372751116752625,
+ "learning_rate": 0.0006,
+ "loss": 4.52687406539917,
+ "step": 1094
+ },
+ {
+ "epoch": 15.209606986899564,
+ "grad_norm": 0.014779130928218365,
+ "learning_rate": 0.0006,
+ "loss": 4.637009620666504,
+ "step": 1095
+ },
+ {
+ "epoch": 15.2235807860262,
+ "grad_norm": 0.015697073191404343,
+ "learning_rate": 0.0006,
+ "loss": 4.682981014251709,
+ "step": 1096
+ },
+ {
+ "epoch": 15.237554585152838,
+ "grad_norm": 0.016523664817214012,
+ "learning_rate": 0.0006,
+ "loss": 4.705722808837891,
+ "step": 1097
+ },
+ {
+ "epoch": 15.251528384279476,
+ "grad_norm": 0.017633996903896332,
+ "learning_rate": 0.0006,
+ "loss": 4.574686050415039,
+ "step": 1098
+ },
+ {
+ "epoch": 15.265502183406113,
+ "grad_norm": 0.019955076277256012,
+ "learning_rate": 0.0006,
+ "loss": 4.634450912475586,
+ "step": 1099
+ },
+ {
+ "epoch": 15.279475982532752,
+ "grad_norm": 0.02132618986070156,
+ "learning_rate": 0.0006,
+ "loss": 4.56110143661499,
+ "step": 1100
+ },
+ {
+ "epoch": 15.293449781659389,
+ "grad_norm": 0.0207593385130167,
+ "learning_rate": 0.0006,
+ "loss": 4.629396438598633,
+ "step": 1101
+ },
+ {
+ "epoch": 15.307423580786025,
+ "grad_norm": 0.017882540822029114,
+ "learning_rate": 0.0006,
+ "loss": 4.730963230133057,
+ "step": 1102
+ },
+ {
+ "epoch": 15.321397379912664,
+ "grad_norm": 0.01917876861989498,
+ "learning_rate": 0.0006,
+ "loss": 4.64585542678833,
+ "step": 1103
+ },
+ {
+ "epoch": 15.335371179039301,
+ "grad_norm": 0.01943974196910858,
+ "learning_rate": 0.0006,
+ "loss": 4.656181335449219,
+ "step": 1104
+ },
+ {
+ "epoch": 15.34934497816594,
+ "grad_norm": 0.019904715940356255,
+ "learning_rate": 0.0006,
+ "loss": 4.619329452514648,
+ "step": 1105
+ },
+ {
+ "epoch": 15.363318777292577,
+ "grad_norm": 0.019174346700310707,
+ "learning_rate": 0.0006,
+ "loss": 4.647892475128174,
+ "step": 1106
+ },
+ {
+ "epoch": 15.377292576419213,
+ "grad_norm": 0.018244758248329163,
+ "learning_rate": 0.0006,
+ "loss": 4.56166934967041,
+ "step": 1107
+ },
+ {
+ "epoch": 15.391266375545852,
+ "grad_norm": 0.018914205953478813,
+ "learning_rate": 0.0006,
+ "loss": 4.648063659667969,
+ "step": 1108
+ },
+ {
+ "epoch": 15.405240174672489,
+ "grad_norm": 0.01961071416735649,
+ "learning_rate": 0.0006,
+ "loss": 4.665525436401367,
+ "step": 1109
+ },
+ {
+ "epoch": 15.419213973799126,
+ "grad_norm": 0.01841311901807785,
+ "learning_rate": 0.0006,
+ "loss": 4.611342430114746,
+ "step": 1110
+ },
+ {
+ "epoch": 15.433187772925764,
+ "grad_norm": 0.01863427273929119,
+ "learning_rate": 0.0006,
+ "loss": 4.664256572723389,
+ "step": 1111
+ },
+ {
+ "epoch": 15.447161572052401,
+ "grad_norm": 0.020506877452135086,
+ "learning_rate": 0.0006,
+ "loss": 4.631237506866455,
+ "step": 1112
+ },
+ {
+ "epoch": 15.46113537117904,
+ "grad_norm": 0.02223833277821541,
+ "learning_rate": 0.0006,
+ "loss": 4.668654441833496,
+ "step": 1113
+ },
+ {
+ "epoch": 15.475109170305677,
+ "grad_norm": 0.023336702957749367,
+ "learning_rate": 0.0006,
+ "loss": 4.562833309173584,
+ "step": 1114
+ },
+ {
+ "epoch": 15.489082969432314,
+ "grad_norm": 0.023052413016557693,
+ "learning_rate": 0.0006,
+ "loss": 4.621532917022705,
+ "step": 1115
+ },
+ {
+ "epoch": 15.503056768558952,
+ "grad_norm": 0.023420924320816994,
+ "learning_rate": 0.0006,
+ "loss": 4.665465354919434,
+ "step": 1116
+ },
+ {
+ "epoch": 15.51703056768559,
+ "grad_norm": 0.024071281775832176,
+ "learning_rate": 0.0006,
+ "loss": 4.595980644226074,
+ "step": 1117
+ },
+ {
+ "epoch": 15.531004366812226,
+ "grad_norm": 0.023406682536005974,
+ "learning_rate": 0.0006,
+ "loss": 4.642635345458984,
+ "step": 1118
+ },
+ {
+ "epoch": 15.544978165938865,
+ "grad_norm": 0.02138647809624672,
+ "learning_rate": 0.0006,
+ "loss": 4.704513072967529,
+ "step": 1119
+ },
+ {
+ "epoch": 15.558951965065502,
+ "grad_norm": 0.0221384447067976,
+ "learning_rate": 0.0006,
+ "loss": 4.567503452301025,
+ "step": 1120
+ },
+ {
+ "epoch": 15.57292576419214,
+ "grad_norm": 0.020879851654171944,
+ "learning_rate": 0.0006,
+ "loss": 4.664057731628418,
+ "step": 1121
+ },
+ {
+ "epoch": 15.586899563318777,
+ "grad_norm": 0.021478967741131783,
+ "learning_rate": 0.0006,
+ "loss": 4.539627552032471,
+ "step": 1122
+ },
+ {
+ "epoch": 15.600873362445414,
+ "grad_norm": 0.022748656570911407,
+ "learning_rate": 0.0006,
+ "loss": 4.5654497146606445,
+ "step": 1123
+ },
+ {
+ "epoch": 15.614847161572053,
+ "grad_norm": 0.021870248019695282,
+ "learning_rate": 0.0006,
+ "loss": 4.56899356842041,
+ "step": 1124
+ },
+ {
+ "epoch": 15.62882096069869,
+ "grad_norm": 0.021711355075240135,
+ "learning_rate": 0.0006,
+ "loss": 4.657808303833008,
+ "step": 1125
+ },
+ {
+ "epoch": 15.642794759825328,
+ "grad_norm": 0.022810909897089005,
+ "learning_rate": 0.0006,
+ "loss": 4.6526689529418945,
+ "step": 1126
+ },
+ {
+ "epoch": 15.656768558951965,
+ "grad_norm": 0.022561343386769295,
+ "learning_rate": 0.0006,
+ "loss": 4.534887790679932,
+ "step": 1127
+ },
+ {
+ "epoch": 15.670742358078602,
+ "grad_norm": 0.019909802824258804,
+ "learning_rate": 0.0006,
+ "loss": 4.600790023803711,
+ "step": 1128
+ },
+ {
+ "epoch": 15.68471615720524,
+ "grad_norm": 0.020092325285077095,
+ "learning_rate": 0.0006,
+ "loss": 4.773830413818359,
+ "step": 1129
+ },
+ {
+ "epoch": 15.698689956331878,
+ "grad_norm": 0.018834445625543594,
+ "learning_rate": 0.0006,
+ "loss": 4.628868103027344,
+ "step": 1130
+ },
+ {
+ "epoch": 15.712663755458514,
+ "grad_norm": 0.017552798613905907,
+ "learning_rate": 0.0006,
+ "loss": 4.647500038146973,
+ "step": 1131
+ },
+ {
+ "epoch": 15.726637554585153,
+ "grad_norm": 0.01751251146197319,
+ "learning_rate": 0.0006,
+ "loss": 4.590545654296875,
+ "step": 1132
+ },
+ {
+ "epoch": 15.74061135371179,
+ "grad_norm": 0.01699099875986576,
+ "learning_rate": 0.0006,
+ "loss": 4.509699821472168,
+ "step": 1133
+ },
+ {
+ "epoch": 15.754585152838429,
+ "grad_norm": 0.015808649361133575,
+ "learning_rate": 0.0006,
+ "loss": 4.606992244720459,
+ "step": 1134
+ },
+ {
+ "epoch": 15.768558951965066,
+ "grad_norm": 0.01655443385243416,
+ "learning_rate": 0.0006,
+ "loss": 4.585219860076904,
+ "step": 1135
+ },
+ {
+ "epoch": 15.782532751091702,
+ "grad_norm": 0.01708163693547249,
+ "learning_rate": 0.0006,
+ "loss": 4.668147087097168,
+ "step": 1136
+ },
+ {
+ "epoch": 15.796506550218341,
+ "grad_norm": 0.017161913216114044,
+ "learning_rate": 0.0006,
+ "loss": 4.7170634269714355,
+ "step": 1137
+ },
+ {
+ "epoch": 15.810480349344978,
+ "grad_norm": 0.017262674868106842,
+ "learning_rate": 0.0006,
+ "loss": 4.572132110595703,
+ "step": 1138
+ },
+ {
+ "epoch": 15.824454148471617,
+ "grad_norm": 0.017713576555252075,
+ "learning_rate": 0.0006,
+ "loss": 4.702538967132568,
+ "step": 1139
+ },
+ {
+ "epoch": 15.838427947598253,
+ "grad_norm": 0.018730733543634415,
+ "learning_rate": 0.0006,
+ "loss": 4.629721641540527,
+ "step": 1140
+ },
+ {
+ "epoch": 15.85240174672489,
+ "grad_norm": 0.017960187047719955,
+ "learning_rate": 0.0006,
+ "loss": 4.661039352416992,
+ "step": 1141
+ },
+ {
+ "epoch": 15.866375545851529,
+ "grad_norm": 0.01750079356133938,
+ "learning_rate": 0.0006,
+ "loss": 4.503710746765137,
+ "step": 1142
+ },
+ {
+ "epoch": 15.880349344978166,
+ "grad_norm": 0.016630422323942184,
+ "learning_rate": 0.0006,
+ "loss": 4.6582136154174805,
+ "step": 1143
+ },
+ {
+ "epoch": 15.894323144104803,
+ "grad_norm": 0.015896832570433617,
+ "learning_rate": 0.0006,
+ "loss": 4.596693992614746,
+ "step": 1144
+ },
+ {
+ "epoch": 15.908296943231441,
+ "grad_norm": 0.01566407084465027,
+ "learning_rate": 0.0006,
+ "loss": 4.6337199211120605,
+ "step": 1145
+ },
+ {
+ "epoch": 15.922270742358078,
+ "grad_norm": 0.016823330894112587,
+ "learning_rate": 0.0006,
+ "loss": 4.686994552612305,
+ "step": 1146
+ },
+ {
+ "epoch": 15.936244541484717,
+ "grad_norm": 0.016208255663514137,
+ "learning_rate": 0.0006,
+ "loss": 4.6205549240112305,
+ "step": 1147
+ },
+ {
+ "epoch": 15.950218340611354,
+ "grad_norm": 0.0171707421541214,
+ "learning_rate": 0.0006,
+ "loss": 4.582968711853027,
+ "step": 1148
+ },
+ {
+ "epoch": 15.96419213973799,
+ "grad_norm": 0.017009446397423744,
+ "learning_rate": 0.0006,
+ "loss": 4.510274887084961,
+ "step": 1149
+ },
+ {
+ "epoch": 15.97816593886463,
+ "grad_norm": 0.01565811224281788,
+ "learning_rate": 0.0006,
+ "loss": 4.55834436416626,
+ "step": 1150
+ },
+ {
+ "epoch": 15.992139737991266,
+ "grad_norm": 0.015276037156581879,
+ "learning_rate": 0.0006,
+ "loss": 4.537870407104492,
+ "step": 1151
+ },
+ {
+ "epoch": 16.0,
+ "grad_norm": 0.017305882647633553,
+ "learning_rate": 0.0006,
+ "loss": 4.611305236816406,
+ "step": 1152
+ },
+ {
+ "epoch": 16.0,
+ "eval_loss": 4.775513172149658,
+ "eval_runtime": 56.7341,
+ "eval_samples_per_second": 43.043,
+ "eval_steps_per_second": 1.357,
+ "step": 1152
+ },
+ {
+ "epoch": 16.01397379912664,
+ "grad_norm": 0.016752436757087708,
+ "learning_rate": 0.0006,
+ "loss": 4.541312217712402,
+ "step": 1153
+ },
+ {
+ "epoch": 16.027947598253274,
+ "grad_norm": 0.01643134094774723,
+ "learning_rate": 0.0006,
+ "loss": 4.585402965545654,
+ "step": 1154
+ },
+ {
+ "epoch": 16.041921397379912,
+ "grad_norm": 0.01888720504939556,
+ "learning_rate": 0.0006,
+ "loss": 4.597825050354004,
+ "step": 1155
+ },
+ {
+ "epoch": 16.05589519650655,
+ "grad_norm": 0.022044632583856583,
+ "learning_rate": 0.0006,
+ "loss": 4.532772064208984,
+ "step": 1156
+ },
+ {
+ "epoch": 16.069868995633186,
+ "grad_norm": 0.022986987605690956,
+ "learning_rate": 0.0006,
+ "loss": 4.551351547241211,
+ "step": 1157
+ },
+ {
+ "epoch": 16.083842794759825,
+ "grad_norm": 0.020116791129112244,
+ "learning_rate": 0.0006,
+ "loss": 4.484149932861328,
+ "step": 1158
+ },
+ {
+ "epoch": 16.097816593886463,
+ "grad_norm": 0.021442247554659843,
+ "learning_rate": 0.0006,
+ "loss": 4.515316486358643,
+ "step": 1159
+ },
+ {
+ "epoch": 16.111790393013102,
+ "grad_norm": 0.020974772050976753,
+ "learning_rate": 0.0006,
+ "loss": 4.51606559753418,
+ "step": 1160
+ },
+ {
+ "epoch": 16.125764192139737,
+ "grad_norm": 0.021274864673614502,
+ "learning_rate": 0.0006,
+ "loss": 4.587329864501953,
+ "step": 1161
+ },
+ {
+ "epoch": 16.139737991266376,
+ "grad_norm": 0.022799784317612648,
+ "learning_rate": 0.0006,
+ "loss": 4.568149089813232,
+ "step": 1162
+ },
+ {
+ "epoch": 16.153711790393015,
+ "grad_norm": 0.021784933283925056,
+ "learning_rate": 0.0006,
+ "loss": 4.515336990356445,
+ "step": 1163
+ },
+ {
+ "epoch": 16.16768558951965,
+ "grad_norm": 0.024261271581053734,
+ "learning_rate": 0.0006,
+ "loss": 4.5127081871032715,
+ "step": 1164
+ },
+ {
+ "epoch": 16.18165938864629,
+ "grad_norm": 0.02248375490307808,
+ "learning_rate": 0.0006,
+ "loss": 4.754251003265381,
+ "step": 1165
+ },
+ {
+ "epoch": 16.195633187772927,
+ "grad_norm": 0.02136492356657982,
+ "learning_rate": 0.0006,
+ "loss": 4.58193826675415,
+ "step": 1166
+ },
+ {
+ "epoch": 16.209606986899562,
+ "grad_norm": 0.020727066323161125,
+ "learning_rate": 0.0006,
+ "loss": 4.564209938049316,
+ "step": 1167
+ },
+ {
+ "epoch": 16.2235807860262,
+ "grad_norm": 0.017789119854569435,
+ "learning_rate": 0.0006,
+ "loss": 4.698680877685547,
+ "step": 1168
+ },
+ {
+ "epoch": 16.23755458515284,
+ "grad_norm": 0.017603034153580666,
+ "learning_rate": 0.0006,
+ "loss": 4.501171588897705,
+ "step": 1169
+ },
+ {
+ "epoch": 16.251528384279474,
+ "grad_norm": 0.017345862463116646,
+ "learning_rate": 0.0006,
+ "loss": 4.593590259552002,
+ "step": 1170
+ },
+ {
+ "epoch": 16.265502183406113,
+ "grad_norm": 0.018047425895929337,
+ "learning_rate": 0.0006,
+ "loss": 4.597467422485352,
+ "step": 1171
+ },
+ {
+ "epoch": 16.27947598253275,
+ "grad_norm": 0.01947222836315632,
+ "learning_rate": 0.0006,
+ "loss": 4.5072221755981445,
+ "step": 1172
+ },
+ {
+ "epoch": 16.29344978165939,
+ "grad_norm": 0.019598500803112984,
+ "learning_rate": 0.0006,
+ "loss": 4.601351737976074,
+ "step": 1173
+ },
+ {
+ "epoch": 16.307423580786025,
+ "grad_norm": 0.020002514123916626,
+ "learning_rate": 0.0006,
+ "loss": 4.580431938171387,
+ "step": 1174
+ },
+ {
+ "epoch": 16.321397379912664,
+ "grad_norm": 0.0216530729085207,
+ "learning_rate": 0.0006,
+ "loss": 4.424681663513184,
+ "step": 1175
+ },
+ {
+ "epoch": 16.335371179039303,
+ "grad_norm": 0.01949235238134861,
+ "learning_rate": 0.0006,
+ "loss": 4.751883029937744,
+ "step": 1176
+ },
+ {
+ "epoch": 16.349344978165938,
+ "grad_norm": 0.017779016867280006,
+ "learning_rate": 0.0006,
+ "loss": 4.541490077972412,
+ "step": 1177
+ },
+ {
+ "epoch": 16.363318777292577,
+ "grad_norm": 0.01909121684730053,
+ "learning_rate": 0.0006,
+ "loss": 4.594141006469727,
+ "step": 1178
+ },
+ {
+ "epoch": 16.377292576419215,
+ "grad_norm": 0.019894849509000778,
+ "learning_rate": 0.0006,
+ "loss": 4.555070400238037,
+ "step": 1179
+ },
+ {
+ "epoch": 16.39126637554585,
+ "grad_norm": 0.01993032731115818,
+ "learning_rate": 0.0006,
+ "loss": 4.521401405334473,
+ "step": 1180
+ },
+ {
+ "epoch": 16.40524017467249,
+ "grad_norm": 0.0210788045078516,
+ "learning_rate": 0.0006,
+ "loss": 4.672842979431152,
+ "step": 1181
+ },
+ {
+ "epoch": 16.419213973799128,
+ "grad_norm": 0.02274598926305771,
+ "learning_rate": 0.0006,
+ "loss": 4.629504203796387,
+ "step": 1182
+ },
+ {
+ "epoch": 16.433187772925763,
+ "grad_norm": 0.020448120310902596,
+ "learning_rate": 0.0006,
+ "loss": 4.548209190368652,
+ "step": 1183
+ },
+ {
+ "epoch": 16.4471615720524,
+ "grad_norm": 0.017565643414855003,
+ "learning_rate": 0.0006,
+ "loss": 4.529054641723633,
+ "step": 1184
+ },
+ {
+ "epoch": 16.46113537117904,
+ "grad_norm": 0.019504237920045853,
+ "learning_rate": 0.0006,
+ "loss": 4.48182487487793,
+ "step": 1185
+ },
+ {
+ "epoch": 16.475109170305675,
+ "grad_norm": 0.016823742538690567,
+ "learning_rate": 0.0006,
+ "loss": 4.608696937561035,
+ "step": 1186
+ },
+ {
+ "epoch": 16.489082969432314,
+ "grad_norm": 0.01602894812822342,
+ "learning_rate": 0.0006,
+ "loss": 4.620062351226807,
+ "step": 1187
+ },
+ {
+ "epoch": 16.503056768558952,
+ "grad_norm": 0.016615642234683037,
+ "learning_rate": 0.0006,
+ "loss": 4.69378662109375,
+ "step": 1188
+ },
+ {
+ "epoch": 16.51703056768559,
+ "grad_norm": 0.018576730042696,
+ "learning_rate": 0.0006,
+ "loss": 4.537041664123535,
+ "step": 1189
+ },
+ {
+ "epoch": 16.531004366812226,
+ "grad_norm": 0.018611151725053787,
+ "learning_rate": 0.0006,
+ "loss": 4.496169090270996,
+ "step": 1190
+ },
+ {
+ "epoch": 16.544978165938865,
+ "grad_norm": 0.018622413277626038,
+ "learning_rate": 0.0006,
+ "loss": 4.557358264923096,
+ "step": 1191
+ },
+ {
+ "epoch": 16.558951965065503,
+ "grad_norm": 0.01823544315993786,
+ "learning_rate": 0.0006,
+ "loss": 4.730792999267578,
+ "step": 1192
+ },
+ {
+ "epoch": 16.57292576419214,
+ "grad_norm": 0.018796846270561218,
+ "learning_rate": 0.0006,
+ "loss": 4.781791687011719,
+ "step": 1193
+ },
+ {
+ "epoch": 16.586899563318777,
+ "grad_norm": 0.01728292554616928,
+ "learning_rate": 0.0006,
+ "loss": 4.515257835388184,
+ "step": 1194
+ },
+ {
+ "epoch": 16.600873362445416,
+ "grad_norm": 0.01699589006602764,
+ "learning_rate": 0.0006,
+ "loss": 4.5996904373168945,
+ "step": 1195
+ },
+ {
+ "epoch": 16.61484716157205,
+ "grad_norm": 0.015905970707535744,
+ "learning_rate": 0.0006,
+ "loss": 4.601301193237305,
+ "step": 1196
+ },
+ {
+ "epoch": 16.62882096069869,
+ "grad_norm": 0.01338571310043335,
+ "learning_rate": 0.0006,
+ "loss": 4.557708740234375,
+ "step": 1197
+ },
+ {
+ "epoch": 16.64279475982533,
+ "grad_norm": 0.015298855490982533,
+ "learning_rate": 0.0006,
+ "loss": 4.542287349700928,
+ "step": 1198
+ },
+ {
+ "epoch": 16.656768558951963,
+ "grad_norm": 0.014753367751836777,
+ "learning_rate": 0.0006,
+ "loss": 4.595787048339844,
+ "step": 1199
+ },
+ {
+ "epoch": 16.670742358078602,
+ "grad_norm": 0.015527483075857162,
+ "learning_rate": 0.0006,
+ "loss": 4.607529163360596,
+ "step": 1200
+ },
+ {
+ "epoch": 16.68471615720524,
+ "grad_norm": 0.015042162500321865,
+ "learning_rate": 0.0006,
+ "loss": 4.686626434326172,
+ "step": 1201
+ },
+ {
+ "epoch": 16.69868995633188,
+ "grad_norm": 0.0159393772482872,
+ "learning_rate": 0.0006,
+ "loss": 4.47656774520874,
+ "step": 1202
+ },
+ {
+ "epoch": 16.712663755458514,
+ "grad_norm": 0.01532787922769785,
+ "learning_rate": 0.0006,
+ "loss": 4.55678653717041,
+ "step": 1203
+ },
+ {
+ "epoch": 16.726637554585153,
+ "grad_norm": 0.016332164406776428,
+ "learning_rate": 0.0006,
+ "loss": 4.51441764831543,
+ "step": 1204
+ },
+ {
+ "epoch": 16.74061135371179,
+ "grad_norm": 0.018901726230978966,
+ "learning_rate": 0.0006,
+ "loss": 4.57227897644043,
+ "step": 1205
+ },
+ {
+ "epoch": 16.754585152838427,
+ "grad_norm": 0.018589433282613754,
+ "learning_rate": 0.0006,
+ "loss": 4.550070762634277,
+ "step": 1206
+ },
+ {
+ "epoch": 16.768558951965066,
+ "grad_norm": 0.01634034886956215,
+ "learning_rate": 0.0006,
+ "loss": 4.586910724639893,
+ "step": 1207
+ },
+ {
+ "epoch": 16.782532751091704,
+ "grad_norm": 0.017599433660507202,
+ "learning_rate": 0.0006,
+ "loss": 4.5891923904418945,
+ "step": 1208
+ },
+ {
+ "epoch": 16.79650655021834,
+ "grad_norm": 0.01878446340560913,
+ "learning_rate": 0.0006,
+ "loss": 4.57227087020874,
+ "step": 1209
+ },
+ {
+ "epoch": 16.810480349344978,
+ "grad_norm": 0.01803469844162464,
+ "learning_rate": 0.0006,
+ "loss": 4.583845138549805,
+ "step": 1210
+ },
+ {
+ "epoch": 16.824454148471617,
+ "grad_norm": 0.018505144864320755,
+ "learning_rate": 0.0006,
+ "loss": 4.561151504516602,
+ "step": 1211
+ },
+ {
+ "epoch": 16.83842794759825,
+ "grad_norm": 0.01920081488788128,
+ "learning_rate": 0.0006,
+ "loss": 4.506834030151367,
+ "step": 1212
+ },
+ {
+ "epoch": 16.85240174672489,
+ "grad_norm": 0.0177712831646204,
+ "learning_rate": 0.0006,
+ "loss": 4.557195663452148,
+ "step": 1213
+ },
+ {
+ "epoch": 16.86637554585153,
+ "grad_norm": 0.019113918766379356,
+ "learning_rate": 0.0006,
+ "loss": 4.651227951049805,
+ "step": 1214
+ },
+ {
+ "epoch": 16.880349344978168,
+ "grad_norm": 0.02012539468705654,
+ "learning_rate": 0.0006,
+ "loss": 4.4952239990234375,
+ "step": 1215
+ },
+ {
+ "epoch": 16.894323144104803,
+ "grad_norm": 0.019993646070361137,
+ "learning_rate": 0.0006,
+ "loss": 4.517433166503906,
+ "step": 1216
+ },
+ {
+ "epoch": 16.90829694323144,
+ "grad_norm": 0.020041441544890404,
+ "learning_rate": 0.0006,
+ "loss": 4.489407539367676,
+ "step": 1217
+ },
+ {
+ "epoch": 16.92227074235808,
+ "grad_norm": 0.0183611661195755,
+ "learning_rate": 0.0006,
+ "loss": 4.5435638427734375,
+ "step": 1218
+ },
+ {
+ "epoch": 16.936244541484715,
+ "grad_norm": 0.019138429313898087,
+ "learning_rate": 0.0006,
+ "loss": 4.545949459075928,
+ "step": 1219
+ },
+ {
+ "epoch": 16.950218340611354,
+ "grad_norm": 0.018554266542196274,
+ "learning_rate": 0.0006,
+ "loss": 4.579620361328125,
+ "step": 1220
+ },
+ {
+ "epoch": 16.964192139737992,
+ "grad_norm": 0.0193545650690794,
+ "learning_rate": 0.0006,
+ "loss": 4.588255405426025,
+ "step": 1221
+ },
+ {
+ "epoch": 16.978165938864628,
+ "grad_norm": 0.019389165565371513,
+ "learning_rate": 0.0006,
+ "loss": 4.520698547363281,
+ "step": 1222
+ },
+ {
+ "epoch": 16.992139737991266,
+ "grad_norm": 0.016601379960775375,
+ "learning_rate": 0.0006,
+ "loss": 4.4809064865112305,
+ "step": 1223
+ },
+ {
+ "epoch": 17.0,
+ "grad_norm": 0.01771303080022335,
+ "learning_rate": 0.0006,
+ "loss": 4.721122741699219,
+ "step": 1224
+ },
+ {
+ "epoch": 17.0,
+ "eval_loss": 4.7903666496276855,
+ "eval_runtime": 56.9826,
+ "eval_samples_per_second": 42.855,
+ "eval_steps_per_second": 1.351,
+ "step": 1224
+ },
+ {
+ "epoch": 17.01397379912664,
+ "grad_norm": 0.017720289528369904,
+ "learning_rate": 0.0006,
+ "loss": 4.523834705352783,
+ "step": 1225
+ },
+ {
+ "epoch": 17.027947598253274,
+ "grad_norm": 0.020137161016464233,
+ "learning_rate": 0.0006,
+ "loss": 4.574061870574951,
+ "step": 1226
+ },
+ {
+ "epoch": 17.041921397379912,
+ "grad_norm": 0.02002314105629921,
+ "learning_rate": 0.0006,
+ "loss": 4.512407302856445,
+ "step": 1227
+ },
+ {
+ "epoch": 17.05589519650655,
+ "grad_norm": 0.019628183916211128,
+ "learning_rate": 0.0006,
+ "loss": 4.502533435821533,
+ "step": 1228
+ },
+ {
+ "epoch": 17.069868995633186,
+ "grad_norm": 0.019808458164334297,
+ "learning_rate": 0.0006,
+ "loss": 4.463156700134277,
+ "step": 1229
+ },
+ {
+ "epoch": 17.083842794759825,
+ "grad_norm": 0.021250786259770393,
+ "learning_rate": 0.0006,
+ "loss": 4.622682094573975,
+ "step": 1230
+ },
+ {
+ "epoch": 17.097816593886463,
+ "grad_norm": 0.022031694650650024,
+ "learning_rate": 0.0006,
+ "loss": 4.451882839202881,
+ "step": 1231
+ },
+ {
+ "epoch": 17.111790393013102,
+ "grad_norm": 0.01898987777531147,
+ "learning_rate": 0.0006,
+ "loss": 4.529331684112549,
+ "step": 1232
+ },
+ {
+ "epoch": 17.125764192139737,
+ "grad_norm": 0.018749834969639778,
+ "learning_rate": 0.0006,
+ "loss": 4.435815334320068,
+ "step": 1233
+ },
+ {
+ "epoch": 17.139737991266376,
+ "grad_norm": 0.01896674931049347,
+ "learning_rate": 0.0006,
+ "loss": 4.484029293060303,
+ "step": 1234
+ },
+ {
+ "epoch": 17.153711790393015,
+ "grad_norm": 0.018286997452378273,
+ "learning_rate": 0.0006,
+ "loss": 4.523510932922363,
+ "step": 1235
+ },
+ {
+ "epoch": 17.16768558951965,
+ "grad_norm": 0.017273059114813805,
+ "learning_rate": 0.0006,
+ "loss": 4.451767444610596,
+ "step": 1236
+ },
+ {
+ "epoch": 17.18165938864629,
+ "grad_norm": 0.016499150544404984,
+ "learning_rate": 0.0006,
+ "loss": 4.5568647384643555,
+ "step": 1237
+ },
+ {
+ "epoch": 17.195633187772927,
+ "grad_norm": 0.015934469178318977,
+ "learning_rate": 0.0006,
+ "loss": 4.4639997482299805,
+ "step": 1238
+ },
+ {
+ "epoch": 17.209606986899562,
+ "grad_norm": 0.01714209094643593,
+ "learning_rate": 0.0006,
+ "loss": 4.547614097595215,
+ "step": 1239
+ },
+ {
+ "epoch": 17.2235807860262,
+ "grad_norm": 0.019167495891451836,
+ "learning_rate": 0.0006,
+ "loss": 4.571097373962402,
+ "step": 1240
+ },
+ {
+ "epoch": 17.23755458515284,
+ "grad_norm": 0.018276996910572052,
+ "learning_rate": 0.0006,
+ "loss": 4.489764213562012,
+ "step": 1241
+ },
+ {
+ "epoch": 17.251528384279474,
+ "grad_norm": 0.021296866238117218,
+ "learning_rate": 0.0006,
+ "loss": 4.526913642883301,
+ "step": 1242
+ },
+ {
+ "epoch": 17.265502183406113,
+ "grad_norm": 0.022606629878282547,
+ "learning_rate": 0.0006,
+ "loss": 4.521624565124512,
+ "step": 1243
+ },
+ {
+ "epoch": 17.27947598253275,
+ "grad_norm": 0.021204711869359016,
+ "learning_rate": 0.0006,
+ "loss": 4.500864028930664,
+ "step": 1244
+ },
+ {
+ "epoch": 17.29344978165939,
+ "grad_norm": 0.01988849975168705,
+ "learning_rate": 0.0006,
+ "loss": 4.45162296295166,
+ "step": 1245
+ },
+ {
+ "epoch": 17.307423580786025,
+ "grad_norm": 0.020268384367227554,
+ "learning_rate": 0.0006,
+ "loss": 4.536198616027832,
+ "step": 1246
+ },
+ {
+ "epoch": 17.321397379912664,
+ "grad_norm": 0.020185677334666252,
+ "learning_rate": 0.0006,
+ "loss": 4.440262794494629,
+ "step": 1247
+ },
+ {
+ "epoch": 17.335371179039303,
+ "grad_norm": 0.020106853917241096,
+ "learning_rate": 0.0006,
+ "loss": 4.475907325744629,
+ "step": 1248
+ },
+ {
+ "epoch": 17.349344978165938,
+ "grad_norm": 0.018883297219872475,
+ "learning_rate": 0.0006,
+ "loss": 4.640661716461182,
+ "step": 1249
+ },
+ {
+ "epoch": 17.363318777292577,
+ "grad_norm": 0.01852620206773281,
+ "learning_rate": 0.0006,
+ "loss": 4.578243255615234,
+ "step": 1250
+ },
+ {
+ "epoch": 17.377292576419215,
+ "grad_norm": 0.01886451430618763,
+ "learning_rate": 0.0006,
+ "loss": 4.546026229858398,
+ "step": 1251
+ },
+ {
+ "epoch": 17.39126637554585,
+ "grad_norm": 0.01777452416718006,
+ "learning_rate": 0.0006,
+ "loss": 4.521542549133301,
+ "step": 1252
+ },
+ {
+ "epoch": 17.40524017467249,
+ "grad_norm": 0.016626447439193726,
+ "learning_rate": 0.0006,
+ "loss": 4.52629280090332,
+ "step": 1253
+ },
+ {
+ "epoch": 17.419213973799128,
+ "grad_norm": 0.01599477231502533,
+ "learning_rate": 0.0006,
+ "loss": 4.4796037673950195,
+ "step": 1254
+ },
+ {
+ "epoch": 17.433187772925763,
+ "grad_norm": 0.017021458595991135,
+ "learning_rate": 0.0006,
+ "loss": 4.468813419342041,
+ "step": 1255
+ },
+ {
+ "epoch": 17.4471615720524,
+ "grad_norm": 0.016925333067774773,
+ "learning_rate": 0.0006,
+ "loss": 4.584898948669434,
+ "step": 1256
+ },
+ {
+ "epoch": 17.46113537117904,
+ "grad_norm": 0.01714998669922352,
+ "learning_rate": 0.0006,
+ "loss": 4.523544788360596,
+ "step": 1257
+ },
+ {
+ "epoch": 17.475109170305675,
+ "grad_norm": 0.01802288182079792,
+ "learning_rate": 0.0006,
+ "loss": 4.595581531524658,
+ "step": 1258
+ },
+ {
+ "epoch": 17.489082969432314,
+ "grad_norm": 0.0180289875715971,
+ "learning_rate": 0.0006,
+ "loss": 4.394440650939941,
+ "step": 1259
+ },
+ {
+ "epoch": 17.503056768558952,
+ "grad_norm": 0.018504245206713676,
+ "learning_rate": 0.0006,
+ "loss": 4.4794416427612305,
+ "step": 1260
+ },
+ {
+ "epoch": 17.51703056768559,
+ "grad_norm": 0.019973233342170715,
+ "learning_rate": 0.0006,
+ "loss": 4.4241437911987305,
+ "step": 1261
+ },
+ {
+ "epoch": 17.531004366812226,
+ "grad_norm": 0.020063083618879318,
+ "learning_rate": 0.0006,
+ "loss": 4.48068380355835,
+ "step": 1262
+ },
+ {
+ "epoch": 17.544978165938865,
+ "grad_norm": 0.02066640742123127,
+ "learning_rate": 0.0006,
+ "loss": 4.628993034362793,
+ "step": 1263
+ },
+ {
+ "epoch": 17.558951965065503,
+ "grad_norm": 0.02206416428089142,
+ "learning_rate": 0.0006,
+ "loss": 4.564790725708008,
+ "step": 1264
+ },
+ {
+ "epoch": 17.57292576419214,
+ "grad_norm": 0.020640335977077484,
+ "learning_rate": 0.0006,
+ "loss": 4.578705787658691,
+ "step": 1265
+ },
+ {
+ "epoch": 17.586899563318777,
+ "grad_norm": 0.019093405455350876,
+ "learning_rate": 0.0006,
+ "loss": 4.5023603439331055,
+ "step": 1266
+ },
+ {
+ "epoch": 17.600873362445416,
+ "grad_norm": 0.019459193572402,
+ "learning_rate": 0.0006,
+ "loss": 4.594531059265137,
+ "step": 1267
+ },
+ {
+ "epoch": 17.61484716157205,
+ "grad_norm": 0.02021171525120735,
+ "learning_rate": 0.0006,
+ "loss": 4.770299911499023,
+ "step": 1268
+ },
+ {
+ "epoch": 17.62882096069869,
+ "grad_norm": 0.021161191165447235,
+ "learning_rate": 0.0006,
+ "loss": 4.5380401611328125,
+ "step": 1269
+ },
+ {
+ "epoch": 17.64279475982533,
+ "grad_norm": 0.021013403311371803,
+ "learning_rate": 0.0006,
+ "loss": 4.455524921417236,
+ "step": 1270
+ },
+ {
+ "epoch": 17.656768558951963,
+ "grad_norm": 0.020336592569947243,
+ "learning_rate": 0.0006,
+ "loss": 4.533672332763672,
+ "step": 1271
+ },
+ {
+ "epoch": 17.670742358078602,
+ "grad_norm": 0.020159434527158737,
+ "learning_rate": 0.0006,
+ "loss": 4.596151351928711,
+ "step": 1272
+ },
+ {
+ "epoch": 17.68471615720524,
+ "grad_norm": 0.020432960242033005,
+ "learning_rate": 0.0006,
+ "loss": 4.55531644821167,
+ "step": 1273
+ },
+ {
+ "epoch": 17.69868995633188,
+ "grad_norm": 0.025466229766607285,
+ "learning_rate": 0.0006,
+ "loss": 4.566912651062012,
+ "step": 1274
+ },
+ {
+ "epoch": 17.712663755458514,
+ "grad_norm": 0.026537558063864708,
+ "learning_rate": 0.0006,
+ "loss": 4.526003837585449,
+ "step": 1275
+ },
+ {
+ "epoch": 17.726637554585153,
+ "grad_norm": 0.024397607892751694,
+ "learning_rate": 0.0006,
+ "loss": 4.541755676269531,
+ "step": 1276
+ },
+ {
+ "epoch": 17.74061135371179,
+ "grad_norm": 0.022062312811613083,
+ "learning_rate": 0.0006,
+ "loss": 4.6532793045043945,
+ "step": 1277
+ },
+ {
+ "epoch": 17.754585152838427,
+ "grad_norm": 0.02204282023012638,
+ "learning_rate": 0.0006,
+ "loss": 4.582825660705566,
+ "step": 1278
+ },
+ {
+ "epoch": 17.768558951965066,
+ "grad_norm": 0.020844416692852974,
+ "learning_rate": 0.0006,
+ "loss": 4.522336483001709,
+ "step": 1279
+ },
+ {
+ "epoch": 17.782532751091704,
+ "grad_norm": 0.019815703853964806,
+ "learning_rate": 0.0006,
+ "loss": 4.607398509979248,
+ "step": 1280
+ },
+ {
+ "epoch": 17.79650655021834,
+ "grad_norm": 0.016676954925060272,
+ "learning_rate": 0.0006,
+ "loss": 4.446206092834473,
+ "step": 1281
+ },
+ {
+ "epoch": 17.810480349344978,
+ "grad_norm": 0.0166452769190073,
+ "learning_rate": 0.0006,
+ "loss": 4.505931854248047,
+ "step": 1282
+ },
+ {
+ "epoch": 17.824454148471617,
+ "grad_norm": 0.015928907319903374,
+ "learning_rate": 0.0006,
+ "loss": 4.455267429351807,
+ "step": 1283
+ },
+ {
+ "epoch": 17.83842794759825,
+ "grad_norm": 0.015219560824334621,
+ "learning_rate": 0.0006,
+ "loss": 4.5623674392700195,
+ "step": 1284
+ },
+ {
+ "epoch": 17.85240174672489,
+ "grad_norm": 0.015125355683267117,
+ "learning_rate": 0.0006,
+ "loss": 4.554104804992676,
+ "step": 1285
+ },
+ {
+ "epoch": 17.86637554585153,
+ "grad_norm": 0.015612718649208546,
+ "learning_rate": 0.0006,
+ "loss": 4.484847068786621,
+ "step": 1286
+ },
+ {
+ "epoch": 17.880349344978168,
+ "grad_norm": 0.015215662308037281,
+ "learning_rate": 0.0006,
+ "loss": 4.4802093505859375,
+ "step": 1287
+ },
+ {
+ "epoch": 17.894323144104803,
+ "grad_norm": 0.01511684525758028,
+ "learning_rate": 0.0006,
+ "loss": 4.537298679351807,
+ "step": 1288
+ },
+ {
+ "epoch": 17.90829694323144,
+ "grad_norm": 0.013841322623193264,
+ "learning_rate": 0.0006,
+ "loss": 4.427488327026367,
+ "step": 1289
+ },
+ {
+ "epoch": 17.92227074235808,
+ "grad_norm": 0.015115964226424694,
+ "learning_rate": 0.0006,
+ "loss": 4.4547119140625,
+ "step": 1290
+ },
+ {
+ "epoch": 17.936244541484715,
+ "grad_norm": 0.016411006450653076,
+ "learning_rate": 0.0006,
+ "loss": 4.52994966506958,
+ "step": 1291
+ },
+ {
+ "epoch": 17.950218340611354,
+ "grad_norm": 0.017313152551651,
+ "learning_rate": 0.0006,
+ "loss": 4.5337653160095215,
+ "step": 1292
+ },
+ {
+ "epoch": 17.964192139737992,
+ "grad_norm": 0.016559796407818794,
+ "learning_rate": 0.0006,
+ "loss": 4.5277910232543945,
+ "step": 1293
+ },
+ {
+ "epoch": 17.978165938864628,
+ "grad_norm": 0.015854761004447937,
+ "learning_rate": 0.0006,
+ "loss": 4.6334309577941895,
+ "step": 1294
+ },
+ {
+ "epoch": 17.992139737991266,
+ "grad_norm": 0.015658238902688026,
+ "learning_rate": 0.0006,
+ "loss": 4.613372802734375,
+ "step": 1295
+ },
+ {
+ "epoch": 18.0,
+ "grad_norm": 0.0169155802577734,
+ "learning_rate": 0.0006,
+ "loss": 4.527356147766113,
+ "step": 1296
+ },
+ {
+ "epoch": 18.0,
+ "eval_loss": 4.764543056488037,
+ "eval_runtime": 56.5998,
+ "eval_samples_per_second": 43.145,
+ "eval_steps_per_second": 1.36,
+ "step": 1296
+ },
+ {
+ "epoch": 18.01397379912664,
+ "grad_norm": 0.015348963439464569,
+ "learning_rate": 0.0006,
+ "loss": 4.466423988342285,
+ "step": 1297
+ },
+ {
+ "epoch": 18.027947598253274,
+ "grad_norm": 0.01768588088452816,
+ "learning_rate": 0.0006,
+ "loss": 4.545933246612549,
+ "step": 1298
+ },
+ {
+ "epoch": 18.041921397379912,
+ "grad_norm": 0.017381226643919945,
+ "learning_rate": 0.0006,
+ "loss": 4.494060039520264,
+ "step": 1299
+ },
+ {
+ "epoch": 18.05589519650655,
+ "grad_norm": 0.019002335146069527,
+ "learning_rate": 0.0006,
+ "loss": 4.395200729370117,
+ "step": 1300
+ },
+ {
+ "epoch": 18.069868995633186,
+ "grad_norm": 0.01827932707965374,
+ "learning_rate": 0.0006,
+ "loss": 4.424848556518555,
+ "step": 1301
+ },
+ {
+ "epoch": 18.083842794759825,
+ "grad_norm": 0.016383031383156776,
+ "learning_rate": 0.0006,
+ "loss": 4.494722366333008,
+ "step": 1302
+ },
+ {
+ "epoch": 18.097816593886463,
+ "grad_norm": 0.016265157610177994,
+ "learning_rate": 0.0006,
+ "loss": 4.416522979736328,
+ "step": 1303
+ },
+ {
+ "epoch": 18.111790393013102,
+ "grad_norm": 0.01702386513352394,
+ "learning_rate": 0.0006,
+ "loss": 4.600103378295898,
+ "step": 1304
+ },
+ {
+ "epoch": 18.125764192139737,
+ "grad_norm": 0.018495498225092888,
+ "learning_rate": 0.0006,
+ "loss": 4.588401794433594,
+ "step": 1305
+ },
+ {
+ "epoch": 18.139737991266376,
+ "grad_norm": 0.017921684309840202,
+ "learning_rate": 0.0006,
+ "loss": 4.478075981140137,
+ "step": 1306
+ },
+ {
+ "epoch": 18.153711790393015,
+ "grad_norm": 0.01801278069615364,
+ "learning_rate": 0.0006,
+ "loss": 4.413186550140381,
+ "step": 1307
+ },
+ {
+ "epoch": 18.16768558951965,
+ "grad_norm": 0.017211275175213814,
+ "learning_rate": 0.0006,
+ "loss": 4.461341857910156,
+ "step": 1308
+ },
+ {
+ "epoch": 18.18165938864629,
+ "grad_norm": 0.017589328810572624,
+ "learning_rate": 0.0006,
+ "loss": 4.546969413757324,
+ "step": 1309
+ },
+ {
+ "epoch": 18.195633187772927,
+ "grad_norm": 0.01797698810696602,
+ "learning_rate": 0.0006,
+ "loss": 4.530797004699707,
+ "step": 1310
+ },
+ {
+ "epoch": 18.209606986899562,
+ "grad_norm": 0.020408619195222855,
+ "learning_rate": 0.0006,
+ "loss": 4.62999153137207,
+ "step": 1311
+ },
+ {
+ "epoch": 18.2235807860262,
+ "grad_norm": 0.01994038000702858,
+ "learning_rate": 0.0006,
+ "loss": 4.586811065673828,
+ "step": 1312
+ },
+ {
+ "epoch": 18.23755458515284,
+ "grad_norm": 0.019381340593099594,
+ "learning_rate": 0.0006,
+ "loss": 4.4769368171691895,
+ "step": 1313
+ },
+ {
+ "epoch": 18.251528384279474,
+ "grad_norm": 0.020484555512666702,
+ "learning_rate": 0.0006,
+ "loss": 4.524302959442139,
+ "step": 1314
+ },
+ {
+ "epoch": 18.265502183406113,
+ "grad_norm": 0.02426416426897049,
+ "learning_rate": 0.0006,
+ "loss": 4.561705589294434,
+ "step": 1315
+ },
+ {
+ "epoch": 18.27947598253275,
+ "grad_norm": 0.024748101830482483,
+ "learning_rate": 0.0006,
+ "loss": 4.4654316902160645,
+ "step": 1316
+ },
+ {
+ "epoch": 18.29344978165939,
+ "grad_norm": 0.020703328773379326,
+ "learning_rate": 0.0006,
+ "loss": 4.53033447265625,
+ "step": 1317
+ },
+ {
+ "epoch": 18.307423580786025,
+ "grad_norm": 0.01765458658337593,
+ "learning_rate": 0.0006,
+ "loss": 4.5154266357421875,
+ "step": 1318
+ },
+ {
+ "epoch": 18.321397379912664,
+ "grad_norm": 0.018646420910954475,
+ "learning_rate": 0.0006,
+ "loss": 4.5224714279174805,
+ "step": 1319
+ },
+ {
+ "epoch": 18.335371179039303,
+ "grad_norm": 0.01907474733889103,
+ "learning_rate": 0.0006,
+ "loss": 4.5055437088012695,
+ "step": 1320
+ },
+ {
+ "epoch": 18.349344978165938,
+ "grad_norm": 0.018179846927523613,
+ "learning_rate": 0.0006,
+ "loss": 4.462161064147949,
+ "step": 1321
+ },
+ {
+ "epoch": 18.363318777292577,
+ "grad_norm": 0.016697878018021584,
+ "learning_rate": 0.0006,
+ "loss": 4.473718643188477,
+ "step": 1322
+ },
+ {
+ "epoch": 18.377292576419215,
+ "grad_norm": 0.017504019662737846,
+ "learning_rate": 0.0006,
+ "loss": 4.474348068237305,
+ "step": 1323
+ },
+ {
+ "epoch": 18.39126637554585,
+ "grad_norm": 0.018889913335442543,
+ "learning_rate": 0.0006,
+ "loss": 4.4531636238098145,
+ "step": 1324
+ },
+ {
+ "epoch": 18.40524017467249,
+ "grad_norm": 0.019167251884937286,
+ "learning_rate": 0.0006,
+ "loss": 4.4278364181518555,
+ "step": 1325
+ },
+ {
+ "epoch": 18.419213973799128,
+ "grad_norm": 0.015895741060376167,
+ "learning_rate": 0.0006,
+ "loss": 4.474592208862305,
+ "step": 1326
+ },
+ {
+ "epoch": 18.433187772925763,
+ "grad_norm": 0.01531882956624031,
+ "learning_rate": 0.0006,
+ "loss": 4.403757572174072,
+ "step": 1327
+ },
+ {
+ "epoch": 18.4471615720524,
+ "grad_norm": 0.01565849967300892,
+ "learning_rate": 0.0006,
+ "loss": 4.507664680480957,
+ "step": 1328
+ },
+ {
+ "epoch": 18.46113537117904,
+ "grad_norm": 0.01602315716445446,
+ "learning_rate": 0.0006,
+ "loss": 4.5648908615112305,
+ "step": 1329
+ },
+ {
+ "epoch": 18.475109170305675,
+ "grad_norm": 0.016467366367578506,
+ "learning_rate": 0.0006,
+ "loss": 4.428762912750244,
+ "step": 1330
+ },
+ {
+ "epoch": 18.489082969432314,
+ "grad_norm": 0.01604127697646618,
+ "learning_rate": 0.0006,
+ "loss": 4.527771472930908,
+ "step": 1331
+ },
+ {
+ "epoch": 18.503056768558952,
+ "grad_norm": 0.018825586885213852,
+ "learning_rate": 0.0006,
+ "loss": 4.485386848449707,
+ "step": 1332
+ },
+ {
+ "epoch": 18.51703056768559,
+ "grad_norm": 0.019258007407188416,
+ "learning_rate": 0.0006,
+ "loss": 4.414972305297852,
+ "step": 1333
+ },
+ {
+ "epoch": 18.531004366812226,
+ "grad_norm": 0.018192021176218987,
+ "learning_rate": 0.0006,
+ "loss": 4.412166595458984,
+ "step": 1334
+ },
+ {
+ "epoch": 18.544978165938865,
+ "grad_norm": 0.016120698302984238,
+ "learning_rate": 0.0006,
+ "loss": 4.484317779541016,
+ "step": 1335
+ },
+ {
+ "epoch": 18.558951965065503,
+ "grad_norm": 0.016496622934937477,
+ "learning_rate": 0.0006,
+ "loss": 4.514986515045166,
+ "step": 1336
+ },
+ {
+ "epoch": 18.57292576419214,
+ "grad_norm": 0.01795116998255253,
+ "learning_rate": 0.0006,
+ "loss": 4.41473388671875,
+ "step": 1337
+ },
+ {
+ "epoch": 18.586899563318777,
+ "grad_norm": 0.019062276929616928,
+ "learning_rate": 0.0006,
+ "loss": 4.601996898651123,
+ "step": 1338
+ },
+ {
+ "epoch": 18.600873362445416,
+ "grad_norm": 0.017174631357192993,
+ "learning_rate": 0.0006,
+ "loss": 4.571218490600586,
+ "step": 1339
+ },
+ {
+ "epoch": 18.61484716157205,
+ "grad_norm": 0.014015606604516506,
+ "learning_rate": 0.0006,
+ "loss": 4.526606559753418,
+ "step": 1340
+ },
+ {
+ "epoch": 18.62882096069869,
+ "grad_norm": 0.015358510427176952,
+ "learning_rate": 0.0006,
+ "loss": 4.604286193847656,
+ "step": 1341
+ },
+ {
+ "epoch": 18.64279475982533,
+ "grad_norm": 0.01619391329586506,
+ "learning_rate": 0.0006,
+ "loss": 4.560883522033691,
+ "step": 1342
+ },
+ {
+ "epoch": 18.656768558951963,
+ "grad_norm": 0.015972964465618134,
+ "learning_rate": 0.0006,
+ "loss": 4.479755401611328,
+ "step": 1343
+ },
+ {
+ "epoch": 18.670742358078602,
+ "grad_norm": 0.016861025243997574,
+ "learning_rate": 0.0006,
+ "loss": 4.474471092224121,
+ "step": 1344
+ },
+ {
+ "epoch": 18.68471615720524,
+ "grad_norm": 0.01822078227996826,
+ "learning_rate": 0.0006,
+ "loss": 4.509489059448242,
+ "step": 1345
+ },
+ {
+ "epoch": 18.69868995633188,
+ "grad_norm": 0.019415007904171944,
+ "learning_rate": 0.0006,
+ "loss": 4.508941173553467,
+ "step": 1346
+ },
+ {
+ "epoch": 18.712663755458514,
+ "grad_norm": 0.019856126978993416,
+ "learning_rate": 0.0006,
+ "loss": 4.446670055389404,
+ "step": 1347
+ },
+ {
+ "epoch": 18.726637554585153,
+ "grad_norm": 0.01842520199716091,
+ "learning_rate": 0.0006,
+ "loss": 4.524422645568848,
+ "step": 1348
+ },
+ {
+ "epoch": 18.74061135371179,
+ "grad_norm": 0.017196480184793472,
+ "learning_rate": 0.0006,
+ "loss": 4.498466968536377,
+ "step": 1349
+ },
+ {
+ "epoch": 18.754585152838427,
+ "grad_norm": 0.01823010854423046,
+ "learning_rate": 0.0006,
+ "loss": 4.534684181213379,
+ "step": 1350
+ },
+ {
+ "epoch": 18.768558951965066,
+ "grad_norm": 0.020398065447807312,
+ "learning_rate": 0.0006,
+ "loss": 4.456494331359863,
+ "step": 1351
+ },
+ {
+ "epoch": 18.782532751091704,
+ "grad_norm": 0.022992262616753578,
+ "learning_rate": 0.0006,
+ "loss": 4.476374626159668,
+ "step": 1352
+ },
+ {
+ "epoch": 18.79650655021834,
+ "grad_norm": 0.02235420234501362,
+ "learning_rate": 0.0006,
+ "loss": 4.587776184082031,
+ "step": 1353
+ },
+ {
+ "epoch": 18.810480349344978,
+ "grad_norm": 0.02093626745045185,
+ "learning_rate": 0.0006,
+ "loss": 4.424172401428223,
+ "step": 1354
+ },
+ {
+ "epoch": 18.824454148471617,
+ "grad_norm": 0.01997302658855915,
+ "learning_rate": 0.0006,
+ "loss": 4.448066711425781,
+ "step": 1355
+ },
+ {
+ "epoch": 18.83842794759825,
+ "grad_norm": 0.018265459686517715,
+ "learning_rate": 0.0006,
+ "loss": 4.436610698699951,
+ "step": 1356
+ },
+ {
+ "epoch": 18.85240174672489,
+ "grad_norm": 0.01673833839595318,
+ "learning_rate": 0.0006,
+ "loss": 4.5502214431762695,
+ "step": 1357
+ },
+ {
+ "epoch": 18.86637554585153,
+ "grad_norm": 0.017908833920955658,
+ "learning_rate": 0.0006,
+ "loss": 4.550530433654785,
+ "step": 1358
+ },
+ {
+ "epoch": 18.880349344978168,
+ "grad_norm": 0.01796654611825943,
+ "learning_rate": 0.0006,
+ "loss": 4.531463146209717,
+ "step": 1359
+ },
+ {
+ "epoch": 18.894323144104803,
+ "grad_norm": 0.01646183803677559,
+ "learning_rate": 0.0006,
+ "loss": 4.498582363128662,
+ "step": 1360
+ },
+ {
+ "epoch": 18.90829694323144,
+ "grad_norm": 0.015430327504873276,
+ "learning_rate": 0.0006,
+ "loss": 4.52571439743042,
+ "step": 1361
+ },
+ {
+ "epoch": 18.92227074235808,
+ "grad_norm": 0.016527952626347542,
+ "learning_rate": 0.0006,
+ "loss": 4.547013759613037,
+ "step": 1362
+ },
+ {
+ "epoch": 18.936244541484715,
+ "grad_norm": 0.016263660043478012,
+ "learning_rate": 0.0006,
+ "loss": 4.468754291534424,
+ "step": 1363
+ },
+ {
+ "epoch": 18.950218340611354,
+ "grad_norm": 0.016347210854291916,
+ "learning_rate": 0.0006,
+ "loss": 4.450741767883301,
+ "step": 1364
+ },
+ {
+ "epoch": 18.964192139737992,
+ "grad_norm": 0.01701398566365242,
+ "learning_rate": 0.0006,
+ "loss": 4.534899711608887,
+ "step": 1365
+ },
+ {
+ "epoch": 18.978165938864628,
+ "grad_norm": 0.017693081870675087,
+ "learning_rate": 0.0006,
+ "loss": 4.460553169250488,
+ "step": 1366
+ },
+ {
+ "epoch": 18.992139737991266,
+ "grad_norm": 0.018746308982372284,
+ "learning_rate": 0.0006,
+ "loss": 4.538792610168457,
+ "step": 1367
+ },
+ {
+ "epoch": 19.0,
+ "grad_norm": 0.020099416375160217,
+ "learning_rate": 0.0006,
+ "loss": 4.449617385864258,
+ "step": 1368
+ },
+ {
+ "epoch": 19.0,
+ "eval_loss": 4.7756571769714355,
+ "eval_runtime": 57.1804,
+ "eval_samples_per_second": 42.707,
+ "eval_steps_per_second": 1.347,
+ "step": 1368
+ },
+ {
+ "epoch": 19.01397379912664,
+ "grad_norm": 0.018726933747529984,
+ "learning_rate": 0.0006,
+ "loss": 4.446059703826904,
+ "step": 1369
+ },
+ {
+ "epoch": 19.027947598253274,
+ "grad_norm": 0.01956087350845337,
+ "learning_rate": 0.0006,
+ "loss": 4.4520039558410645,
+ "step": 1370
+ },
+ {
+ "epoch": 19.041921397379912,
+ "grad_norm": 0.020092811435461044,
+ "learning_rate": 0.0006,
+ "loss": 4.543343544006348,
+ "step": 1371
+ },
+ {
+ "epoch": 19.05589519650655,
+ "grad_norm": 0.01843622885644436,
+ "learning_rate": 0.0006,
+ "loss": 4.535579204559326,
+ "step": 1372
+ },
+ {
+ "epoch": 19.069868995633186,
+ "grad_norm": 0.016989488154649734,
+ "learning_rate": 0.0006,
+ "loss": 4.52316951751709,
+ "step": 1373
+ },
+ {
+ "epoch": 19.083842794759825,
+ "grad_norm": 0.01805422455072403,
+ "learning_rate": 0.0006,
+ "loss": 4.5393266677856445,
+ "step": 1374
+ },
+ {
+ "epoch": 19.097816593886463,
+ "grad_norm": 0.01874629408121109,
+ "learning_rate": 0.0006,
+ "loss": 4.422059059143066,
+ "step": 1375
+ },
+ {
+ "epoch": 19.111790393013102,
+ "grad_norm": 0.01804221048951149,
+ "learning_rate": 0.0006,
+ "loss": 4.50508451461792,
+ "step": 1376
+ },
+ {
+ "epoch": 19.125764192139737,
+ "grad_norm": 0.01785474829375744,
+ "learning_rate": 0.0006,
+ "loss": 4.412599563598633,
+ "step": 1377
+ },
+ {
+ "epoch": 19.139737991266376,
+ "grad_norm": 0.01971791312098503,
+ "learning_rate": 0.0006,
+ "loss": 4.518318176269531,
+ "step": 1378
+ },
+ {
+ "epoch": 19.153711790393015,
+ "grad_norm": 0.02061633951961994,
+ "learning_rate": 0.0006,
+ "loss": 4.332393169403076,
+ "step": 1379
+ },
+ {
+ "epoch": 19.16768558951965,
+ "grad_norm": 0.020563429221510887,
+ "learning_rate": 0.0006,
+ "loss": 4.445863723754883,
+ "step": 1380
+ },
+ {
+ "epoch": 19.18165938864629,
+ "grad_norm": 0.019995713606476784,
+ "learning_rate": 0.0006,
+ "loss": 4.497312068939209,
+ "step": 1381
+ },
+ {
+ "epoch": 19.195633187772927,
+ "grad_norm": 0.019823070615530014,
+ "learning_rate": 0.0006,
+ "loss": 4.475085258483887,
+ "step": 1382
+ },
+ {
+ "epoch": 19.209606986899562,
+ "grad_norm": 0.019039355218410492,
+ "learning_rate": 0.0006,
+ "loss": 4.50432014465332,
+ "step": 1383
+ },
+ {
+ "epoch": 19.2235807860262,
+ "grad_norm": 0.017876973375678062,
+ "learning_rate": 0.0006,
+ "loss": 4.384788513183594,
+ "step": 1384
+ },
+ {
+ "epoch": 19.23755458515284,
+ "grad_norm": 0.018697699531912804,
+ "learning_rate": 0.0006,
+ "loss": 4.410778045654297,
+ "step": 1385
+ },
+ {
+ "epoch": 19.251528384279474,
+ "grad_norm": 0.018826806917786598,
+ "learning_rate": 0.0006,
+ "loss": 4.524034023284912,
+ "step": 1386
+ },
+ {
+ "epoch": 19.265502183406113,
+ "grad_norm": 0.01824226602911949,
+ "learning_rate": 0.0006,
+ "loss": 4.526393890380859,
+ "step": 1387
+ },
+ {
+ "epoch": 19.27947598253275,
+ "grad_norm": 0.01753208599984646,
+ "learning_rate": 0.0006,
+ "loss": 4.493703365325928,
+ "step": 1388
+ },
+ {
+ "epoch": 19.29344978165939,
+ "grad_norm": 0.01742432825267315,
+ "learning_rate": 0.0006,
+ "loss": 4.392831802368164,
+ "step": 1389
+ },
+ {
+ "epoch": 19.307423580786025,
+ "grad_norm": 0.017099356278777122,
+ "learning_rate": 0.0006,
+ "loss": 4.578668594360352,
+ "step": 1390
+ },
+ {
+ "epoch": 19.321397379912664,
+ "grad_norm": 0.017946023494005203,
+ "learning_rate": 0.0006,
+ "loss": 4.456718444824219,
+ "step": 1391
+ },
+ {
+ "epoch": 19.335371179039303,
+ "grad_norm": 0.017705943435430527,
+ "learning_rate": 0.0006,
+ "loss": 4.553682327270508,
+ "step": 1392
+ },
+ {
+ "epoch": 19.349344978165938,
+ "grad_norm": 0.016647523269057274,
+ "learning_rate": 0.0006,
+ "loss": 4.525016784667969,
+ "step": 1393
+ },
+ {
+ "epoch": 19.363318777292577,
+ "grad_norm": 0.015407416969537735,
+ "learning_rate": 0.0006,
+ "loss": 4.500231742858887,
+ "step": 1394
+ },
+ {
+ "epoch": 19.377292576419215,
+ "grad_norm": 0.017264019697904587,
+ "learning_rate": 0.0006,
+ "loss": 4.4615349769592285,
+ "step": 1395
+ },
+ {
+ "epoch": 19.39126637554585,
+ "grad_norm": 0.01871555484831333,
+ "learning_rate": 0.0006,
+ "loss": 4.281026840209961,
+ "step": 1396
+ },
+ {
+ "epoch": 19.40524017467249,
+ "grad_norm": 0.017361650243401527,
+ "learning_rate": 0.0006,
+ "loss": 4.479382514953613,
+ "step": 1397
+ },
+ {
+ "epoch": 19.419213973799128,
+ "grad_norm": 0.019781362265348434,
+ "learning_rate": 0.0006,
+ "loss": 4.402320861816406,
+ "step": 1398
+ },
+ {
+ "epoch": 19.433187772925763,
+ "grad_norm": 0.019043298438191414,
+ "learning_rate": 0.0006,
+ "loss": 4.476983070373535,
+ "step": 1399
+ },
+ {
+ "epoch": 19.4471615720524,
+ "grad_norm": 0.017137985676527023,
+ "learning_rate": 0.0006,
+ "loss": 4.418689250946045,
+ "step": 1400
+ },
+ {
+ "epoch": 19.46113537117904,
+ "grad_norm": 0.016481833532452583,
+ "learning_rate": 0.0006,
+ "loss": 4.579145431518555,
+ "step": 1401
+ },
+ {
+ "epoch": 19.475109170305675,
+ "grad_norm": 0.016507714986801147,
+ "learning_rate": 0.0006,
+ "loss": 4.519158363342285,
+ "step": 1402
+ },
+ {
+ "epoch": 19.489082969432314,
+ "grad_norm": 0.018092872574925423,
+ "learning_rate": 0.0006,
+ "loss": 4.513748645782471,
+ "step": 1403
+ },
+ {
+ "epoch": 19.503056768558952,
+ "grad_norm": 0.018153440207242966,
+ "learning_rate": 0.0006,
+ "loss": 4.645384311676025,
+ "step": 1404
+ },
+ {
+ "epoch": 19.51703056768559,
+ "grad_norm": 0.01657119207084179,
+ "learning_rate": 0.0006,
+ "loss": 4.399757385253906,
+ "step": 1405
+ },
+ {
+ "epoch": 19.531004366812226,
+ "grad_norm": 0.0156280305236578,
+ "learning_rate": 0.0006,
+ "loss": 4.467836380004883,
+ "step": 1406
+ },
+ {
+ "epoch": 19.544978165938865,
+ "grad_norm": 0.016838403418660164,
+ "learning_rate": 0.0006,
+ "loss": 4.529176235198975,
+ "step": 1407
+ },
+ {
+ "epoch": 19.558951965065503,
+ "grad_norm": 0.01743236929178238,
+ "learning_rate": 0.0006,
+ "loss": 4.440032482147217,
+ "step": 1408
+ },
+ {
+ "epoch": 19.57292576419214,
+ "grad_norm": 0.01553610060364008,
+ "learning_rate": 0.0006,
+ "loss": 4.425588607788086,
+ "step": 1409
+ },
+ {
+ "epoch": 19.586899563318777,
+ "grad_norm": 0.017259271815419197,
+ "learning_rate": 0.0006,
+ "loss": 4.405452728271484,
+ "step": 1410
+ },
+ {
+ "epoch": 19.600873362445416,
+ "grad_norm": 0.015568481758236885,
+ "learning_rate": 0.0006,
+ "loss": 4.444279193878174,
+ "step": 1411
+ },
+ {
+ "epoch": 19.61484716157205,
+ "grad_norm": 0.015331248752772808,
+ "learning_rate": 0.0006,
+ "loss": 4.451846122741699,
+ "step": 1412
+ },
+ {
+ "epoch": 19.62882096069869,
+ "grad_norm": 0.015667378902435303,
+ "learning_rate": 0.0006,
+ "loss": 4.533931255340576,
+ "step": 1413
+ },
+ {
+ "epoch": 19.64279475982533,
+ "grad_norm": 0.016177602112293243,
+ "learning_rate": 0.0006,
+ "loss": 4.467313289642334,
+ "step": 1414
+ },
+ {
+ "epoch": 19.656768558951963,
+ "grad_norm": 0.01733129471540451,
+ "learning_rate": 0.0006,
+ "loss": 4.496893882751465,
+ "step": 1415
+ },
+ {
+ "epoch": 19.670742358078602,
+ "grad_norm": 0.016539594158530235,
+ "learning_rate": 0.0006,
+ "loss": 4.401961326599121,
+ "step": 1416
+ },
+ {
+ "epoch": 19.68471615720524,
+ "grad_norm": 0.016058551147580147,
+ "learning_rate": 0.0006,
+ "loss": 4.503375053405762,
+ "step": 1417
+ },
+ {
+ "epoch": 19.69868995633188,
+ "grad_norm": 0.01715777814388275,
+ "learning_rate": 0.0006,
+ "loss": 4.3808369636535645,
+ "step": 1418
+ },
+ {
+ "epoch": 19.712663755458514,
+ "grad_norm": 0.01681119203567505,
+ "learning_rate": 0.0006,
+ "loss": 4.5469560623168945,
+ "step": 1419
+ },
+ {
+ "epoch": 19.726637554585153,
+ "grad_norm": 0.01634550653398037,
+ "learning_rate": 0.0006,
+ "loss": 4.328197956085205,
+ "step": 1420
+ },
+ {
+ "epoch": 19.74061135371179,
+ "grad_norm": 0.01549347210675478,
+ "learning_rate": 0.0006,
+ "loss": 4.534621238708496,
+ "step": 1421
+ },
+ {
+ "epoch": 19.754585152838427,
+ "grad_norm": 0.01654665172100067,
+ "learning_rate": 0.0006,
+ "loss": 4.4724225997924805,
+ "step": 1422
+ },
+ {
+ "epoch": 19.768558951965066,
+ "grad_norm": 0.017119025811553,
+ "learning_rate": 0.0006,
+ "loss": 4.3897247314453125,
+ "step": 1423
+ },
+ {
+ "epoch": 19.782532751091704,
+ "grad_norm": 0.01603223942220211,
+ "learning_rate": 0.0006,
+ "loss": 4.412731170654297,
+ "step": 1424
+ },
+ {
+ "epoch": 19.79650655021834,
+ "grad_norm": 0.016774067655205727,
+ "learning_rate": 0.0006,
+ "loss": 4.54707145690918,
+ "step": 1425
+ },
+ {
+ "epoch": 19.810480349344978,
+ "grad_norm": 0.016453025862574577,
+ "learning_rate": 0.0006,
+ "loss": 4.462501049041748,
+ "step": 1426
+ },
+ {
+ "epoch": 19.824454148471617,
+ "grad_norm": 0.01597629487514496,
+ "learning_rate": 0.0006,
+ "loss": 4.229758262634277,
+ "step": 1427
+ },
+ {
+ "epoch": 19.83842794759825,
+ "grad_norm": 0.017999034374952316,
+ "learning_rate": 0.0006,
+ "loss": 4.389263153076172,
+ "step": 1428
+ },
+ {
+ "epoch": 19.85240174672489,
+ "grad_norm": 0.02165898121893406,
+ "learning_rate": 0.0006,
+ "loss": 4.342130661010742,
+ "step": 1429
+ },
+ {
+ "epoch": 19.86637554585153,
+ "grad_norm": 0.02222822792828083,
+ "learning_rate": 0.0006,
+ "loss": 4.492661476135254,
+ "step": 1430
+ },
+ {
+ "epoch": 19.880349344978168,
+ "grad_norm": 0.019378066062927246,
+ "learning_rate": 0.0006,
+ "loss": 4.4614105224609375,
+ "step": 1431
+ },
+ {
+ "epoch": 19.894323144104803,
+ "grad_norm": 0.019908083602786064,
+ "learning_rate": 0.0006,
+ "loss": 4.370929718017578,
+ "step": 1432
+ },
+ {
+ "epoch": 19.90829694323144,
+ "grad_norm": 0.022296471521258354,
+ "learning_rate": 0.0006,
+ "loss": 4.438320159912109,
+ "step": 1433
+ },
+ {
+ "epoch": 19.92227074235808,
+ "grad_norm": 0.021810343489050865,
+ "learning_rate": 0.0006,
+ "loss": 4.462067127227783,
+ "step": 1434
+ },
+ {
+ "epoch": 19.936244541484715,
+ "grad_norm": 0.02129209227859974,
+ "learning_rate": 0.0006,
+ "loss": 4.541261196136475,
+ "step": 1435
+ },
+ {
+ "epoch": 19.950218340611354,
+ "grad_norm": 0.02079629711806774,
+ "learning_rate": 0.0006,
+ "loss": 4.36712646484375,
+ "step": 1436
+ },
+ {
+ "epoch": 19.964192139737992,
+ "grad_norm": 0.022501740604639053,
+ "learning_rate": 0.0006,
+ "loss": 4.5279083251953125,
+ "step": 1437
+ },
+ {
+ "epoch": 19.978165938864628,
+ "grad_norm": 0.019121676683425903,
+ "learning_rate": 0.0006,
+ "loss": 4.477789878845215,
+ "step": 1438
+ },
+ {
+ "epoch": 19.992139737991266,
+ "grad_norm": 0.017429588362574577,
+ "learning_rate": 0.0006,
+ "loss": 4.4023847579956055,
+ "step": 1439
+ },
+ {
+ "epoch": 20.0,
+ "grad_norm": 0.018857665359973907,
+ "learning_rate": 0.0006,
+ "loss": 4.494743347167969,
+ "step": 1440
+ },
+ {
+ "epoch": 20.0,
+ "eval_loss": 4.694555759429932,
+ "eval_runtime": 56.4146,
+ "eval_samples_per_second": 43.287,
+ "eval_steps_per_second": 1.365,
+ "step": 1440
+ },
+ {
+ "epoch": 20.01397379912664,
+ "grad_norm": 0.016505861654877663,
+ "learning_rate": 0.0006,
+ "loss": 4.506229877471924,
+ "step": 1441
+ },
+ {
+ "epoch": 20.027947598253274,
+ "grad_norm": 0.017571842297911644,
+ "learning_rate": 0.0006,
+ "loss": 4.384073257446289,
+ "step": 1442
+ },
+ {
+ "epoch": 20.041921397379912,
+ "grad_norm": 0.019398365169763565,
+ "learning_rate": 0.0006,
+ "loss": 4.482177734375,
+ "step": 1443
+ },
+ {
+ "epoch": 20.05589519650655,
+ "grad_norm": 0.02011062018573284,
+ "learning_rate": 0.0006,
+ "loss": 4.440773963928223,
+ "step": 1444
+ },
+ {
+ "epoch": 20.069868995633186,
+ "grad_norm": 0.019026953727006912,
+ "learning_rate": 0.0006,
+ "loss": 4.507168769836426,
+ "step": 1445
+ },
+ {
+ "epoch": 20.083842794759825,
+ "grad_norm": 0.020506031811237335,
+ "learning_rate": 0.0006,
+ "loss": 4.469079971313477,
+ "step": 1446
+ },
+ {
+ "epoch": 20.097816593886463,
+ "grad_norm": 0.019608162343502045,
+ "learning_rate": 0.0006,
+ "loss": 4.395927429199219,
+ "step": 1447
+ },
+ {
+ "epoch": 20.111790393013102,
+ "grad_norm": 0.019419124349951744,
+ "learning_rate": 0.0006,
+ "loss": 4.484235763549805,
+ "step": 1448
+ },
+ {
+ "epoch": 20.125764192139737,
+ "grad_norm": 0.021775128319859505,
+ "learning_rate": 0.0006,
+ "loss": 4.333173751831055,
+ "step": 1449
+ },
+ {
+ "epoch": 20.139737991266376,
+ "grad_norm": 0.025512272492051125,
+ "learning_rate": 0.0006,
+ "loss": 4.460362434387207,
+ "step": 1450
+ },
+ {
+ "epoch": 20.153711790393015,
+ "grad_norm": 0.025157401338219643,
+ "learning_rate": 0.0006,
+ "loss": 4.4361395835876465,
+ "step": 1451
+ },
+ {
+ "epoch": 20.16768558951965,
+ "grad_norm": 0.019862616434693336,
+ "learning_rate": 0.0006,
+ "loss": 4.426807403564453,
+ "step": 1452
+ },
+ {
+ "epoch": 20.18165938864629,
+ "grad_norm": 0.01932152360677719,
+ "learning_rate": 0.0006,
+ "loss": 4.375677108764648,
+ "step": 1453
+ },
+ {
+ "epoch": 20.195633187772927,
+ "grad_norm": 0.01760999858379364,
+ "learning_rate": 0.0006,
+ "loss": 4.42259407043457,
+ "step": 1454
+ },
+ {
+ "epoch": 20.209606986899562,
+ "grad_norm": 0.016998112201690674,
+ "learning_rate": 0.0006,
+ "loss": 4.481280326843262,
+ "step": 1455
+ },
+ {
+ "epoch": 20.2235807860262,
+ "grad_norm": 0.017199190333485603,
+ "learning_rate": 0.0006,
+ "loss": 4.496712684631348,
+ "step": 1456
+ },
+ {
+ "epoch": 20.23755458515284,
+ "grad_norm": 0.016833679750561714,
+ "learning_rate": 0.0006,
+ "loss": 4.40736722946167,
+ "step": 1457
+ },
+ {
+ "epoch": 20.251528384279474,
+ "grad_norm": 0.017546730116009712,
+ "learning_rate": 0.0006,
+ "loss": 4.4915266036987305,
+ "step": 1458
+ },
+ {
+ "epoch": 20.265502183406113,
+ "grad_norm": 0.016595078632235527,
+ "learning_rate": 0.0006,
+ "loss": 4.460815906524658,
+ "step": 1459
+ },
+ {
+ "epoch": 20.27947598253275,
+ "grad_norm": 0.017728568986058235,
+ "learning_rate": 0.0006,
+ "loss": 4.346048831939697,
+ "step": 1460
+ },
+ {
+ "epoch": 20.29344978165939,
+ "grad_norm": 0.01849197782576084,
+ "learning_rate": 0.0006,
+ "loss": 4.399681091308594,
+ "step": 1461
+ },
+ {
+ "epoch": 20.307423580786025,
+ "grad_norm": 0.018336588516831398,
+ "learning_rate": 0.0006,
+ "loss": 4.414183616638184,
+ "step": 1462
+ },
+ {
+ "epoch": 20.321397379912664,
+ "grad_norm": 0.018086519092321396,
+ "learning_rate": 0.0006,
+ "loss": 4.346586227416992,
+ "step": 1463
+ },
+ {
+ "epoch": 20.335371179039303,
+ "grad_norm": 0.019697299227118492,
+ "learning_rate": 0.0006,
+ "loss": 4.4420952796936035,
+ "step": 1464
+ },
+ {
+ "epoch": 20.349344978165938,
+ "grad_norm": 0.02198074571788311,
+ "learning_rate": 0.0006,
+ "loss": 4.588729381561279,
+ "step": 1465
+ },
+ {
+ "epoch": 20.363318777292577,
+ "grad_norm": 0.022813329473137856,
+ "learning_rate": 0.0006,
+ "loss": 4.49729585647583,
+ "step": 1466
+ },
+ {
+ "epoch": 20.377292576419215,
+ "grad_norm": 0.022006649523973465,
+ "learning_rate": 0.0006,
+ "loss": 4.475641250610352,
+ "step": 1467
+ },
+ {
+ "epoch": 20.39126637554585,
+ "grad_norm": 0.023188438266515732,
+ "learning_rate": 0.0006,
+ "loss": 4.404322624206543,
+ "step": 1468
+ },
+ {
+ "epoch": 20.40524017467249,
+ "grad_norm": 0.022226877510547638,
+ "learning_rate": 0.0006,
+ "loss": 4.50904655456543,
+ "step": 1469
+ },
+ {
+ "epoch": 20.419213973799128,
+ "grad_norm": 0.021082637831568718,
+ "learning_rate": 0.0006,
+ "loss": 4.339011192321777,
+ "step": 1470
+ },
+ {
+ "epoch": 20.433187772925763,
+ "grad_norm": 0.02013542130589485,
+ "learning_rate": 0.0006,
+ "loss": 4.516942024230957,
+ "step": 1471
+ },
+ {
+ "epoch": 20.4471615720524,
+ "grad_norm": 0.01898968778550625,
+ "learning_rate": 0.0006,
+ "loss": 4.315804481506348,
+ "step": 1472
+ },
+ {
+ "epoch": 20.46113537117904,
+ "grad_norm": 0.019450880587100983,
+ "learning_rate": 0.0006,
+ "loss": 4.4794921875,
+ "step": 1473
+ },
+ {
+ "epoch": 20.475109170305675,
+ "grad_norm": 0.018577802926301956,
+ "learning_rate": 0.0006,
+ "loss": 4.422323703765869,
+ "step": 1474
+ },
+ {
+ "epoch": 20.489082969432314,
+ "grad_norm": 0.01778412237763405,
+ "learning_rate": 0.0006,
+ "loss": 4.429789066314697,
+ "step": 1475
+ },
+ {
+ "epoch": 20.503056768558952,
+ "grad_norm": 0.017134476453065872,
+ "learning_rate": 0.0006,
+ "loss": 4.416459083557129,
+ "step": 1476
+ },
+ {
+ "epoch": 20.51703056768559,
+ "grad_norm": 0.017153557389974594,
+ "learning_rate": 0.0006,
+ "loss": 4.446170806884766,
+ "step": 1477
+ },
+ {
+ "epoch": 20.531004366812226,
+ "grad_norm": 0.014998532831668854,
+ "learning_rate": 0.0006,
+ "loss": 4.419447898864746,
+ "step": 1478
+ },
+ {
+ "epoch": 20.544978165938865,
+ "grad_norm": 0.015877850353717804,
+ "learning_rate": 0.0006,
+ "loss": 4.312338829040527,
+ "step": 1479
+ },
+ {
+ "epoch": 20.558951965065503,
+ "grad_norm": 0.01569489948451519,
+ "learning_rate": 0.0006,
+ "loss": 4.572905540466309,
+ "step": 1480
+ },
+ {
+ "epoch": 20.57292576419214,
+ "grad_norm": 0.01591474376618862,
+ "learning_rate": 0.0006,
+ "loss": 4.414907455444336,
+ "step": 1481
+ },
+ {
+ "epoch": 20.586899563318777,
+ "grad_norm": 0.015395903028547764,
+ "learning_rate": 0.0006,
+ "loss": 4.549131393432617,
+ "step": 1482
+ },
+ {
+ "epoch": 20.600873362445416,
+ "grad_norm": 0.015715358778834343,
+ "learning_rate": 0.0006,
+ "loss": 4.530194282531738,
+ "step": 1483
+ },
+ {
+ "epoch": 20.61484716157205,
+ "grad_norm": 0.016042975708842278,
+ "learning_rate": 0.0006,
+ "loss": 4.322149276733398,
+ "step": 1484
+ },
+ {
+ "epoch": 20.62882096069869,
+ "grad_norm": 0.01664186641573906,
+ "learning_rate": 0.0006,
+ "loss": 4.424866676330566,
+ "step": 1485
+ },
+ {
+ "epoch": 20.64279475982533,
+ "grad_norm": 0.01631534844636917,
+ "learning_rate": 0.0006,
+ "loss": 4.441752910614014,
+ "step": 1486
+ },
+ {
+ "epoch": 20.656768558951963,
+ "grad_norm": 0.014057058840990067,
+ "learning_rate": 0.0006,
+ "loss": 4.399982452392578,
+ "step": 1487
+ },
+ {
+ "epoch": 20.670742358078602,
+ "grad_norm": 0.015076273120939732,
+ "learning_rate": 0.0006,
+ "loss": 4.441324234008789,
+ "step": 1488
+ },
+ {
+ "epoch": 20.68471615720524,
+ "grad_norm": 0.014138277620077133,
+ "learning_rate": 0.0006,
+ "loss": 4.4312543869018555,
+ "step": 1489
+ },
+ {
+ "epoch": 20.69868995633188,
+ "grad_norm": 0.013762440532445908,
+ "learning_rate": 0.0006,
+ "loss": 4.387807846069336,
+ "step": 1490
+ },
+ {
+ "epoch": 20.712663755458514,
+ "grad_norm": 0.015058411285281181,
+ "learning_rate": 0.0006,
+ "loss": 4.594747543334961,
+ "step": 1491
+ },
+ {
+ "epoch": 20.726637554585153,
+ "grad_norm": 0.013827123679220676,
+ "learning_rate": 0.0006,
+ "loss": 4.434604167938232,
+ "step": 1492
+ },
+ {
+ "epoch": 20.74061135371179,
+ "grad_norm": 0.015315295197069645,
+ "learning_rate": 0.0006,
+ "loss": 4.588066101074219,
+ "step": 1493
+ },
+ {
+ "epoch": 20.754585152838427,
+ "grad_norm": 0.018678732216358185,
+ "learning_rate": 0.0006,
+ "loss": 4.3514204025268555,
+ "step": 1494
+ },
+ {
+ "epoch": 20.768558951965066,
+ "grad_norm": 0.019521350041031837,
+ "learning_rate": 0.0006,
+ "loss": 4.437658786773682,
+ "step": 1495
+ },
+ {
+ "epoch": 20.782532751091704,
+ "grad_norm": 0.018363403156399727,
+ "learning_rate": 0.0006,
+ "loss": 4.495002269744873,
+ "step": 1496
+ },
+ {
+ "epoch": 20.79650655021834,
+ "grad_norm": 0.016830725595355034,
+ "learning_rate": 0.0006,
+ "loss": 4.402471542358398,
+ "step": 1497
+ },
+ {
+ "epoch": 20.810480349344978,
+ "grad_norm": 0.01722072623670101,
+ "learning_rate": 0.0006,
+ "loss": 4.437653541564941,
+ "step": 1498
+ },
+ {
+ "epoch": 20.824454148471617,
+ "grad_norm": 0.017459599301218987,
+ "learning_rate": 0.0006,
+ "loss": 4.366603851318359,
+ "step": 1499
+ },
+ {
+ "epoch": 20.83842794759825,
+ "grad_norm": 0.018902219831943512,
+ "learning_rate": 0.0006,
+ "loss": 4.501561164855957,
+ "step": 1500
+ },
+ {
+ "epoch": 20.85240174672489,
+ "grad_norm": 0.020814096555113792,
+ "learning_rate": 0.0006,
+ "loss": 4.535295486450195,
+ "step": 1501
+ },
+ {
+ "epoch": 20.86637554585153,
+ "grad_norm": 0.018990976735949516,
+ "learning_rate": 0.0006,
+ "loss": 4.333580017089844,
+ "step": 1502
+ },
+ {
+ "epoch": 20.880349344978168,
+ "grad_norm": 0.01837880350649357,
+ "learning_rate": 0.0006,
+ "loss": 4.313388824462891,
+ "step": 1503
+ },
+ {
+ "epoch": 20.894323144104803,
+ "grad_norm": 0.019820360466837883,
+ "learning_rate": 0.0006,
+ "loss": 4.388157844543457,
+ "step": 1504
+ },
+ {
+ "epoch": 20.90829694323144,
+ "grad_norm": 0.020181143656373024,
+ "learning_rate": 0.0006,
+ "loss": 4.572224140167236,
+ "step": 1505
+ },
+ {
+ "epoch": 20.92227074235808,
+ "grad_norm": 0.021206077188253403,
+ "learning_rate": 0.0006,
+ "loss": 4.3127970695495605,
+ "step": 1506
+ },
+ {
+ "epoch": 20.936244541484715,
+ "grad_norm": 0.017962060868740082,
+ "learning_rate": 0.0006,
+ "loss": 4.384734630584717,
+ "step": 1507
+ },
+ {
+ "epoch": 20.950218340611354,
+ "grad_norm": 0.01737220585346222,
+ "learning_rate": 0.0006,
+ "loss": 4.4075751304626465,
+ "step": 1508
+ },
+ {
+ "epoch": 20.964192139737992,
+ "grad_norm": 0.01839268207550049,
+ "learning_rate": 0.0006,
+ "loss": 4.553779125213623,
+ "step": 1509
+ },
+ {
+ "epoch": 20.978165938864628,
+ "grad_norm": 0.020102640613913536,
+ "learning_rate": 0.0006,
+ "loss": 4.399398326873779,
+ "step": 1510
+ },
+ {
+ "epoch": 20.992139737991266,
+ "grad_norm": 0.018831463530659676,
+ "learning_rate": 0.0006,
+ "loss": 4.358499526977539,
+ "step": 1511
+ },
+ {
+ "epoch": 21.0,
+ "grad_norm": 0.018254734575748444,
+ "learning_rate": 0.0006,
+ "loss": 4.384119510650635,
+ "step": 1512
+ },
+ {
+ "epoch": 21.0,
+ "eval_loss": 4.7524237632751465,
+ "eval_runtime": 57.8812,
+ "eval_samples_per_second": 42.19,
+ "eval_steps_per_second": 1.33,
+ "step": 1512
+ },
+ {
+ "epoch": 21.01397379912664,
+ "grad_norm": 0.019728587940335274,
+ "learning_rate": 0.0006,
+ "loss": 4.32363224029541,
+ "step": 1513
+ },
+ {
+ "epoch": 21.027947598253274,
+ "grad_norm": 0.02216625027358532,
+ "learning_rate": 0.0006,
+ "loss": 4.436988353729248,
+ "step": 1514
+ },
+ {
+ "epoch": 21.041921397379912,
+ "grad_norm": 0.02227606810629368,
+ "learning_rate": 0.0006,
+ "loss": 4.406063556671143,
+ "step": 1515
+ },
+ {
+ "epoch": 21.05589519650655,
+ "grad_norm": 0.020187009125947952,
+ "learning_rate": 0.0006,
+ "loss": 4.324402332305908,
+ "step": 1516
+ },
+ {
+ "epoch": 21.069868995633186,
+ "grad_norm": 0.02095157466828823,
+ "learning_rate": 0.0006,
+ "loss": 4.456544876098633,
+ "step": 1517
+ },
+ {
+ "epoch": 21.083842794759825,
+ "grad_norm": 0.02146013453602791,
+ "learning_rate": 0.0006,
+ "loss": 4.368387222290039,
+ "step": 1518
+ },
+ {
+ "epoch": 21.097816593886463,
+ "grad_norm": 0.02162035182118416,
+ "learning_rate": 0.0006,
+ "loss": 4.48838996887207,
+ "step": 1519
+ },
+ {
+ "epoch": 21.111790393013102,
+ "grad_norm": 0.021745676174759865,
+ "learning_rate": 0.0006,
+ "loss": 4.30778694152832,
+ "step": 1520
+ },
+ {
+ "epoch": 21.125764192139737,
+ "grad_norm": 0.02173588052392006,
+ "learning_rate": 0.0006,
+ "loss": 4.433149814605713,
+ "step": 1521
+ },
+ {
+ "epoch": 21.139737991266376,
+ "grad_norm": 0.020950648933649063,
+ "learning_rate": 0.0006,
+ "loss": 4.520089149475098,
+ "step": 1522
+ },
+ {
+ "epoch": 21.153711790393015,
+ "grad_norm": 0.020684752613306046,
+ "learning_rate": 0.0006,
+ "loss": 4.372437477111816,
+ "step": 1523
+ },
+ {
+ "epoch": 21.16768558951965,
+ "grad_norm": 0.019472869113087654,
+ "learning_rate": 0.0006,
+ "loss": 4.437015533447266,
+ "step": 1524
+ },
+ {
+ "epoch": 21.18165938864629,
+ "grad_norm": 0.01846962980926037,
+ "learning_rate": 0.0006,
+ "loss": 4.449531555175781,
+ "step": 1525
+ },
+ {
+ "epoch": 21.195633187772927,
+ "grad_norm": 0.017377803102135658,
+ "learning_rate": 0.0006,
+ "loss": 4.458279132843018,
+ "step": 1526
+ },
+ {
+ "epoch": 21.209606986899562,
+ "grad_norm": 0.016002200543880463,
+ "learning_rate": 0.0006,
+ "loss": 4.266297817230225,
+ "step": 1527
+ },
+ {
+ "epoch": 21.2235807860262,
+ "grad_norm": 0.015575903467833996,
+ "learning_rate": 0.0006,
+ "loss": 4.342632293701172,
+ "step": 1528
+ },
+ {
+ "epoch": 21.23755458515284,
+ "grad_norm": 0.015634698793292046,
+ "learning_rate": 0.0006,
+ "loss": 4.338623523712158,
+ "step": 1529
+ },
+ {
+ "epoch": 21.251528384279474,
+ "grad_norm": 0.01449581328779459,
+ "learning_rate": 0.0006,
+ "loss": 4.354330062866211,
+ "step": 1530
+ },
+ {
+ "epoch": 21.265502183406113,
+ "grad_norm": 0.01487616915255785,
+ "learning_rate": 0.0006,
+ "loss": 4.486835479736328,
+ "step": 1531
+ },
+ {
+ "epoch": 21.27947598253275,
+ "grad_norm": 0.016346555203199387,
+ "learning_rate": 0.0006,
+ "loss": 4.407098770141602,
+ "step": 1532
+ },
+ {
+ "epoch": 21.29344978165939,
+ "grad_norm": 0.01682182401418686,
+ "learning_rate": 0.0006,
+ "loss": 4.436809539794922,
+ "step": 1533
+ },
+ {
+ "epoch": 21.307423580786025,
+ "grad_norm": 0.017704004421830177,
+ "learning_rate": 0.0006,
+ "loss": 4.3854780197143555,
+ "step": 1534
+ },
+ {
+ "epoch": 21.321397379912664,
+ "grad_norm": 0.018437420949339867,
+ "learning_rate": 0.0006,
+ "loss": 4.413283824920654,
+ "step": 1535
+ },
+ {
+ "epoch": 21.335371179039303,
+ "grad_norm": 0.016911081969738007,
+ "learning_rate": 0.0006,
+ "loss": 4.357306003570557,
+ "step": 1536
+ },
+ {
+ "epoch": 21.349344978165938,
+ "grad_norm": 0.01642223633825779,
+ "learning_rate": 0.0006,
+ "loss": 4.422322750091553,
+ "step": 1537
+ },
+ {
+ "epoch": 21.363318777292577,
+ "grad_norm": 0.01780877821147442,
+ "learning_rate": 0.0006,
+ "loss": 4.387620449066162,
+ "step": 1538
+ },
+ {
+ "epoch": 21.377292576419215,
+ "grad_norm": 0.01653975248336792,
+ "learning_rate": 0.0006,
+ "loss": 4.327267646789551,
+ "step": 1539
+ },
+ {
+ "epoch": 21.39126637554585,
+ "grad_norm": 0.01572156324982643,
+ "learning_rate": 0.0006,
+ "loss": 4.475979804992676,
+ "step": 1540
+ },
+ {
+ "epoch": 21.40524017467249,
+ "grad_norm": 0.017156066372990608,
+ "learning_rate": 0.0006,
+ "loss": 4.4011125564575195,
+ "step": 1541
+ },
+ {
+ "epoch": 21.419213973799128,
+ "grad_norm": 0.01788829080760479,
+ "learning_rate": 0.0006,
+ "loss": 4.4174089431762695,
+ "step": 1542
+ },
+ {
+ "epoch": 21.433187772925763,
+ "grad_norm": 0.01728985086083412,
+ "learning_rate": 0.0006,
+ "loss": 4.429244041442871,
+ "step": 1543
+ },
+ {
+ "epoch": 21.4471615720524,
+ "grad_norm": 0.016534466296434402,
+ "learning_rate": 0.0006,
+ "loss": 4.326068878173828,
+ "step": 1544
+ },
+ {
+ "epoch": 21.46113537117904,
+ "grad_norm": 0.016582584008574486,
+ "learning_rate": 0.0006,
+ "loss": 4.2975969314575195,
+ "step": 1545
+ },
+ {
+ "epoch": 21.475109170305675,
+ "grad_norm": 0.018034178763628006,
+ "learning_rate": 0.0006,
+ "loss": 4.501319408416748,
+ "step": 1546
+ },
+ {
+ "epoch": 21.489082969432314,
+ "grad_norm": 0.017827702686190605,
+ "learning_rate": 0.0006,
+ "loss": 4.502284526824951,
+ "step": 1547
+ },
+ {
+ "epoch": 21.503056768558952,
+ "grad_norm": 0.017182299867272377,
+ "learning_rate": 0.0006,
+ "loss": 4.380828857421875,
+ "step": 1548
+ },
+ {
+ "epoch": 21.51703056768559,
+ "grad_norm": 0.019101588055491447,
+ "learning_rate": 0.0006,
+ "loss": 4.31564998626709,
+ "step": 1549
+ },
+ {
+ "epoch": 21.531004366812226,
+ "grad_norm": 0.0204677302390337,
+ "learning_rate": 0.0006,
+ "loss": 4.3840131759643555,
+ "step": 1550
+ },
+ {
+ "epoch": 21.544978165938865,
+ "grad_norm": 0.0188993439078331,
+ "learning_rate": 0.0006,
+ "loss": 4.4257731437683105,
+ "step": 1551
+ },
+ {
+ "epoch": 21.558951965065503,
+ "grad_norm": 0.01730230636894703,
+ "learning_rate": 0.0006,
+ "loss": 4.445132255554199,
+ "step": 1552
+ },
+ {
+ "epoch": 21.57292576419214,
+ "grad_norm": 0.01801574043929577,
+ "learning_rate": 0.0006,
+ "loss": 4.427064895629883,
+ "step": 1553
+ },
+ {
+ "epoch": 21.586899563318777,
+ "grad_norm": 0.018417565152049065,
+ "learning_rate": 0.0006,
+ "loss": 4.319595813751221,
+ "step": 1554
+ },
+ {
+ "epoch": 21.600873362445416,
+ "grad_norm": 0.01683160290122032,
+ "learning_rate": 0.0006,
+ "loss": 4.349218368530273,
+ "step": 1555
+ },
+ {
+ "epoch": 21.61484716157205,
+ "grad_norm": 0.016973106190562248,
+ "learning_rate": 0.0006,
+ "loss": 4.375990390777588,
+ "step": 1556
+ },
+ {
+ "epoch": 21.62882096069869,
+ "grad_norm": 0.015688583254814148,
+ "learning_rate": 0.0006,
+ "loss": 4.408170223236084,
+ "step": 1557
+ },
+ {
+ "epoch": 21.64279475982533,
+ "grad_norm": 0.01573043316602707,
+ "learning_rate": 0.0006,
+ "loss": 4.416255950927734,
+ "step": 1558
+ },
+ {
+ "epoch": 21.656768558951963,
+ "grad_norm": 0.01744293048977852,
+ "learning_rate": 0.0006,
+ "loss": 4.477141380310059,
+ "step": 1559
+ },
+ {
+ "epoch": 21.670742358078602,
+ "grad_norm": 0.017503513023257256,
+ "learning_rate": 0.0006,
+ "loss": 4.3367486000061035,
+ "step": 1560
+ },
+ {
+ "epoch": 21.68471615720524,
+ "grad_norm": 0.016592150554060936,
+ "learning_rate": 0.0006,
+ "loss": 4.314785480499268,
+ "step": 1561
+ },
+ {
+ "epoch": 21.69868995633188,
+ "grad_norm": 0.015774084255099297,
+ "learning_rate": 0.0006,
+ "loss": 4.448209285736084,
+ "step": 1562
+ },
+ {
+ "epoch": 21.712663755458514,
+ "grad_norm": 0.015779150649905205,
+ "learning_rate": 0.0006,
+ "loss": 4.392690658569336,
+ "step": 1563
+ },
+ {
+ "epoch": 21.726637554585153,
+ "grad_norm": 0.015392648056149483,
+ "learning_rate": 0.0006,
+ "loss": 4.437741756439209,
+ "step": 1564
+ },
+ {
+ "epoch": 21.74061135371179,
+ "grad_norm": 0.01530836895108223,
+ "learning_rate": 0.0006,
+ "loss": 4.424595832824707,
+ "step": 1565
+ },
+ {
+ "epoch": 21.754585152838427,
+ "grad_norm": 0.015600779093801975,
+ "learning_rate": 0.0006,
+ "loss": 4.335417747497559,
+ "step": 1566
+ },
+ {
+ "epoch": 21.768558951965066,
+ "grad_norm": 0.014646518044173717,
+ "learning_rate": 0.0006,
+ "loss": 4.279081344604492,
+ "step": 1567
+ },
+ {
+ "epoch": 21.782532751091704,
+ "grad_norm": 0.014132864773273468,
+ "learning_rate": 0.0006,
+ "loss": 4.345950126647949,
+ "step": 1568
+ },
+ {
+ "epoch": 21.79650655021834,
+ "grad_norm": 0.01425724383443594,
+ "learning_rate": 0.0006,
+ "loss": 4.277588844299316,
+ "step": 1569
+ },
+ {
+ "epoch": 21.810480349344978,
+ "grad_norm": 0.015396572649478912,
+ "learning_rate": 0.0006,
+ "loss": 4.326196670532227,
+ "step": 1570
+ },
+ {
+ "epoch": 21.824454148471617,
+ "grad_norm": 0.01683669537305832,
+ "learning_rate": 0.0006,
+ "loss": 4.3294172286987305,
+ "step": 1571
+ },
+ {
+ "epoch": 21.83842794759825,
+ "grad_norm": 0.016365647315979004,
+ "learning_rate": 0.0006,
+ "loss": 4.523321628570557,
+ "step": 1572
+ },
+ {
+ "epoch": 21.85240174672489,
+ "grad_norm": 0.017142774537205696,
+ "learning_rate": 0.0006,
+ "loss": 4.475842475891113,
+ "step": 1573
+ },
+ {
+ "epoch": 21.86637554585153,
+ "grad_norm": 0.015950839966535568,
+ "learning_rate": 0.0006,
+ "loss": 4.397187232971191,
+ "step": 1574
+ },
+ {
+ "epoch": 21.880349344978168,
+ "grad_norm": 0.016180362552404404,
+ "learning_rate": 0.0006,
+ "loss": 4.414778709411621,
+ "step": 1575
+ },
+ {
+ "epoch": 21.894323144104803,
+ "grad_norm": 0.017615556716918945,
+ "learning_rate": 0.0006,
+ "loss": 4.348567962646484,
+ "step": 1576
+ },
+ {
+ "epoch": 21.90829694323144,
+ "grad_norm": 0.01896277628839016,
+ "learning_rate": 0.0006,
+ "loss": 4.431762218475342,
+ "step": 1577
+ },
+ {
+ "epoch": 21.92227074235808,
+ "grad_norm": 0.01990971714258194,
+ "learning_rate": 0.0006,
+ "loss": 4.431285858154297,
+ "step": 1578
+ },
+ {
+ "epoch": 21.936244541484715,
+ "grad_norm": 0.0201258547604084,
+ "learning_rate": 0.0006,
+ "loss": 4.309872627258301,
+ "step": 1579
+ },
+ {
+ "epoch": 21.950218340611354,
+ "grad_norm": 0.019620198756456375,
+ "learning_rate": 0.0006,
+ "loss": 4.484044551849365,
+ "step": 1580
+ },
+ {
+ "epoch": 21.964192139737992,
+ "grad_norm": 0.01771123707294464,
+ "learning_rate": 0.0006,
+ "loss": 4.368093013763428,
+ "step": 1581
+ },
+ {
+ "epoch": 21.978165938864628,
+ "grad_norm": 0.017501654103398323,
+ "learning_rate": 0.0006,
+ "loss": 4.372651100158691,
+ "step": 1582
+ },
+ {
+ "epoch": 21.992139737991266,
+ "grad_norm": 0.01693608984351158,
+ "learning_rate": 0.0006,
+ "loss": 4.379167556762695,
+ "step": 1583
+ },
+ {
+ "epoch": 22.0,
+ "grad_norm": 0.02002849616110325,
+ "learning_rate": 0.0006,
+ "loss": 4.329627990722656,
+ "step": 1584
+ },
+ {
+ "epoch": 22.0,
+ "eval_loss": 4.723696231842041,
+ "eval_runtime": 56.1758,
+ "eval_samples_per_second": 43.471,
+ "eval_steps_per_second": 1.371,
+ "step": 1584
+ },
+ {
+ "epoch": 22.01397379912664,
+ "grad_norm": 0.02030305378139019,
+ "learning_rate": 0.0006,
+ "loss": 4.355566024780273,
+ "step": 1585
+ },
+ {
+ "epoch": 22.027947598253274,
+ "grad_norm": 0.018638934940099716,
+ "learning_rate": 0.0006,
+ "loss": 4.314009189605713,
+ "step": 1586
+ },
+ {
+ "epoch": 22.041921397379912,
+ "grad_norm": 0.019917072728276253,
+ "learning_rate": 0.0006,
+ "loss": 4.367525577545166,
+ "step": 1587
+ },
+ {
+ "epoch": 22.05589519650655,
+ "grad_norm": 0.020785167813301086,
+ "learning_rate": 0.0006,
+ "loss": 4.375561714172363,
+ "step": 1588
+ },
+ {
+ "epoch": 22.069868995633186,
+ "grad_norm": 0.021214453503489494,
+ "learning_rate": 0.0006,
+ "loss": 4.418305397033691,
+ "step": 1589
+ },
+ {
+ "epoch": 22.083842794759825,
+ "grad_norm": 0.019051678478717804,
+ "learning_rate": 0.0006,
+ "loss": 4.215126991271973,
+ "step": 1590
+ },
+ {
+ "epoch": 22.097816593886463,
+ "grad_norm": 0.02323366142809391,
+ "learning_rate": 0.0006,
+ "loss": 4.306807994842529,
+ "step": 1591
+ },
+ {
+ "epoch": 22.111790393013102,
+ "grad_norm": 0.026186738163232803,
+ "learning_rate": 0.0006,
+ "loss": 4.408508777618408,
+ "step": 1592
+ },
+ {
+ "epoch": 22.125764192139737,
+ "grad_norm": 0.027361227199435234,
+ "learning_rate": 0.0006,
+ "loss": 4.329322814941406,
+ "step": 1593
+ },
+ {
+ "epoch": 22.139737991266376,
+ "grad_norm": 0.026564646512269974,
+ "learning_rate": 0.0006,
+ "loss": 4.346642017364502,
+ "step": 1594
+ },
+ {
+ "epoch": 22.153711790393015,
+ "grad_norm": 0.025073667988181114,
+ "learning_rate": 0.0006,
+ "loss": 4.3984198570251465,
+ "step": 1595
+ },
+ {
+ "epoch": 22.16768558951965,
+ "grad_norm": 0.02347794733941555,
+ "learning_rate": 0.0006,
+ "loss": 4.352538108825684,
+ "step": 1596
+ },
+ {
+ "epoch": 22.18165938864629,
+ "grad_norm": 0.02203468047082424,
+ "learning_rate": 0.0006,
+ "loss": 4.298148155212402,
+ "step": 1597
+ },
+ {
+ "epoch": 22.195633187772927,
+ "grad_norm": 0.020621957257390022,
+ "learning_rate": 0.0006,
+ "loss": 4.480309963226318,
+ "step": 1598
+ },
+ {
+ "epoch": 22.209606986899562,
+ "grad_norm": 0.023710055276751518,
+ "learning_rate": 0.0006,
+ "loss": 4.339317321777344,
+ "step": 1599
+ },
+ {
+ "epoch": 22.2235807860262,
+ "grad_norm": 0.0233482476323843,
+ "learning_rate": 0.0006,
+ "loss": 4.436896324157715,
+ "step": 1600
+ },
+ {
+ "epoch": 22.23755458515284,
+ "grad_norm": 0.020122403278946877,
+ "learning_rate": 0.0006,
+ "loss": 4.236002445220947,
+ "step": 1601
+ },
+ {
+ "epoch": 22.251528384279474,
+ "grad_norm": 0.01957266591489315,
+ "learning_rate": 0.0006,
+ "loss": 4.497328281402588,
+ "step": 1602
+ },
+ {
+ "epoch": 22.265502183406113,
+ "grad_norm": 0.018964895978569984,
+ "learning_rate": 0.0006,
+ "loss": 4.36360502243042,
+ "step": 1603
+ },
+ {
+ "epoch": 22.27947598253275,
+ "grad_norm": 0.018902769312262535,
+ "learning_rate": 0.0006,
+ "loss": 4.427093505859375,
+ "step": 1604
+ },
+ {
+ "epoch": 22.29344978165939,
+ "grad_norm": 0.018918119370937347,
+ "learning_rate": 0.0006,
+ "loss": 4.294950485229492,
+ "step": 1605
+ },
+ {
+ "epoch": 22.307423580786025,
+ "grad_norm": 0.019315775483846664,
+ "learning_rate": 0.0006,
+ "loss": 4.427952289581299,
+ "step": 1606
+ },
+ {
+ "epoch": 22.321397379912664,
+ "grad_norm": 0.018350522965192795,
+ "learning_rate": 0.0006,
+ "loss": 4.272945404052734,
+ "step": 1607
+ },
+ {
+ "epoch": 22.335371179039303,
+ "grad_norm": 0.01759534701704979,
+ "learning_rate": 0.0006,
+ "loss": 4.348023891448975,
+ "step": 1608
+ },
+ {
+ "epoch": 22.349344978165938,
+ "grad_norm": 0.01646057888865471,
+ "learning_rate": 0.0006,
+ "loss": 4.278518199920654,
+ "step": 1609
+ },
+ {
+ "epoch": 22.363318777292577,
+ "grad_norm": 0.015875060111284256,
+ "learning_rate": 0.0006,
+ "loss": 4.423527240753174,
+ "step": 1610
+ },
+ {
+ "epoch": 22.377292576419215,
+ "grad_norm": 0.01638529635965824,
+ "learning_rate": 0.0006,
+ "loss": 4.321723937988281,
+ "step": 1611
+ },
+ {
+ "epoch": 22.39126637554585,
+ "grad_norm": 0.016942061483860016,
+ "learning_rate": 0.0006,
+ "loss": 4.346164703369141,
+ "step": 1612
+ },
+ {
+ "epoch": 22.40524017467249,
+ "grad_norm": 0.015209621749818325,
+ "learning_rate": 0.0006,
+ "loss": 4.434340000152588,
+ "step": 1613
+ },
+ {
+ "epoch": 22.419213973799128,
+ "grad_norm": 0.014423009008169174,
+ "learning_rate": 0.0006,
+ "loss": 4.342560768127441,
+ "step": 1614
+ },
+ {
+ "epoch": 22.433187772925763,
+ "grad_norm": 0.014369525946676731,
+ "learning_rate": 0.0006,
+ "loss": 4.296586990356445,
+ "step": 1615
+ },
+ {
+ "epoch": 22.4471615720524,
+ "grad_norm": 0.01392540242522955,
+ "learning_rate": 0.0006,
+ "loss": 4.433176040649414,
+ "step": 1616
+ },
+ {
+ "epoch": 22.46113537117904,
+ "grad_norm": 0.013206899166107178,
+ "learning_rate": 0.0006,
+ "loss": 4.32962703704834,
+ "step": 1617
+ },
+ {
+ "epoch": 22.475109170305675,
+ "grad_norm": 0.013427077792584896,
+ "learning_rate": 0.0006,
+ "loss": 4.420437335968018,
+ "step": 1618
+ },
+ {
+ "epoch": 22.489082969432314,
+ "grad_norm": 0.014194196090102196,
+ "learning_rate": 0.0006,
+ "loss": 4.410478591918945,
+ "step": 1619
+ },
+ {
+ "epoch": 22.503056768558952,
+ "grad_norm": 0.01362364087253809,
+ "learning_rate": 0.0006,
+ "loss": 4.338932037353516,
+ "step": 1620
+ },
+ {
+ "epoch": 22.51703056768559,
+ "grad_norm": 0.014299589209258556,
+ "learning_rate": 0.0006,
+ "loss": 4.34089994430542,
+ "step": 1621
+ },
+ {
+ "epoch": 22.531004366812226,
+ "grad_norm": 0.014061295427381992,
+ "learning_rate": 0.0006,
+ "loss": 4.29201602935791,
+ "step": 1622
+ },
+ {
+ "epoch": 22.544978165938865,
+ "grad_norm": 0.013945229351520538,
+ "learning_rate": 0.0006,
+ "loss": 4.326052188873291,
+ "step": 1623
+ },
+ {
+ "epoch": 22.558951965065503,
+ "grad_norm": 0.01387068536132574,
+ "learning_rate": 0.0006,
+ "loss": 4.393521308898926,
+ "step": 1624
+ },
+ {
+ "epoch": 22.57292576419214,
+ "grad_norm": 0.013931898400187492,
+ "learning_rate": 0.0006,
+ "loss": 4.3739752769470215,
+ "step": 1625
+ },
+ {
+ "epoch": 22.586899563318777,
+ "grad_norm": 0.01497740764170885,
+ "learning_rate": 0.0006,
+ "loss": 4.3991618156433105,
+ "step": 1626
+ },
+ {
+ "epoch": 22.600873362445416,
+ "grad_norm": 0.015842905268073082,
+ "learning_rate": 0.0006,
+ "loss": 4.405340671539307,
+ "step": 1627
+ },
+ {
+ "epoch": 22.61484716157205,
+ "grad_norm": 0.01580633595585823,
+ "learning_rate": 0.0006,
+ "loss": 4.280102729797363,
+ "step": 1628
+ },
+ {
+ "epoch": 22.62882096069869,
+ "grad_norm": 0.016567382961511612,
+ "learning_rate": 0.0006,
+ "loss": 4.407657146453857,
+ "step": 1629
+ },
+ {
+ "epoch": 22.64279475982533,
+ "grad_norm": 0.01648622378706932,
+ "learning_rate": 0.0006,
+ "loss": 4.392590522766113,
+ "step": 1630
+ },
+ {
+ "epoch": 22.656768558951963,
+ "grad_norm": 0.017066307365894318,
+ "learning_rate": 0.0006,
+ "loss": 4.324862003326416,
+ "step": 1631
+ },
+ {
+ "epoch": 22.670742358078602,
+ "grad_norm": 0.01637062057852745,
+ "learning_rate": 0.0006,
+ "loss": 4.34598445892334,
+ "step": 1632
+ },
+ {
+ "epoch": 22.68471615720524,
+ "grad_norm": 0.018038960173726082,
+ "learning_rate": 0.0006,
+ "loss": 4.320711612701416,
+ "step": 1633
+ },
+ {
+ "epoch": 22.69868995633188,
+ "grad_norm": 0.018096428364515305,
+ "learning_rate": 0.0006,
+ "loss": 4.536527156829834,
+ "step": 1634
+ },
+ {
+ "epoch": 22.712663755458514,
+ "grad_norm": 0.017889291048049927,
+ "learning_rate": 0.0006,
+ "loss": 4.388131141662598,
+ "step": 1635
+ },
+ {
+ "epoch": 22.726637554585153,
+ "grad_norm": 0.018771521747112274,
+ "learning_rate": 0.0006,
+ "loss": 4.357668876647949,
+ "step": 1636
+ },
+ {
+ "epoch": 22.74061135371179,
+ "grad_norm": 0.02100873365998268,
+ "learning_rate": 0.0006,
+ "loss": 4.293972015380859,
+ "step": 1637
+ },
+ {
+ "epoch": 22.754585152838427,
+ "grad_norm": 0.017715785652399063,
+ "learning_rate": 0.0006,
+ "loss": 4.349849224090576,
+ "step": 1638
+ },
+ {
+ "epoch": 22.768558951965066,
+ "grad_norm": 0.015546333976089954,
+ "learning_rate": 0.0006,
+ "loss": 4.311728000640869,
+ "step": 1639
+ },
+ {
+ "epoch": 22.782532751091704,
+ "grad_norm": 0.01605990156531334,
+ "learning_rate": 0.0006,
+ "loss": 4.499123573303223,
+ "step": 1640
+ },
+ {
+ "epoch": 22.79650655021834,
+ "grad_norm": 0.016642142087221146,
+ "learning_rate": 0.0006,
+ "loss": 4.4319257736206055,
+ "step": 1641
+ },
+ {
+ "epoch": 22.810480349344978,
+ "grad_norm": 0.016120387241244316,
+ "learning_rate": 0.0006,
+ "loss": 4.4373650550842285,
+ "step": 1642
+ },
+ {
+ "epoch": 22.824454148471617,
+ "grad_norm": 0.015174890868365765,
+ "learning_rate": 0.0006,
+ "loss": 4.314082145690918,
+ "step": 1643
+ },
+ {
+ "epoch": 22.83842794759825,
+ "grad_norm": 0.01551714539527893,
+ "learning_rate": 0.0006,
+ "loss": 4.360363960266113,
+ "step": 1644
+ },
+ {
+ "epoch": 22.85240174672489,
+ "grad_norm": 0.014966742135584354,
+ "learning_rate": 0.0006,
+ "loss": 4.346253395080566,
+ "step": 1645
+ },
+ {
+ "epoch": 22.86637554585153,
+ "grad_norm": 0.014288028702139854,
+ "learning_rate": 0.0006,
+ "loss": 4.464366436004639,
+ "step": 1646
+ },
+ {
+ "epoch": 22.880349344978168,
+ "grad_norm": 0.015291682444512844,
+ "learning_rate": 0.0006,
+ "loss": 4.504101753234863,
+ "step": 1647
+ },
+ {
+ "epoch": 22.894323144104803,
+ "grad_norm": 0.014534426853060722,
+ "learning_rate": 0.0006,
+ "loss": 4.462172031402588,
+ "step": 1648
+ },
+ {
+ "epoch": 22.90829694323144,
+ "grad_norm": 0.014129819348454475,
+ "learning_rate": 0.0006,
+ "loss": 4.350126266479492,
+ "step": 1649
+ },
+ {
+ "epoch": 22.92227074235808,
+ "grad_norm": 0.014932668767869473,
+ "learning_rate": 0.0006,
+ "loss": 4.374392032623291,
+ "step": 1650
+ },
+ {
+ "epoch": 22.936244541484715,
+ "grad_norm": 0.014647853560745716,
+ "learning_rate": 0.0006,
+ "loss": 4.252264499664307,
+ "step": 1651
+ },
+ {
+ "epoch": 22.950218340611354,
+ "grad_norm": 0.013553360477089882,
+ "learning_rate": 0.0006,
+ "loss": 4.379729270935059,
+ "step": 1652
+ },
+ {
+ "epoch": 22.964192139737992,
+ "grad_norm": 0.014606194570660591,
+ "learning_rate": 0.0006,
+ "loss": 4.363372802734375,
+ "step": 1653
+ },
+ {
+ "epoch": 22.978165938864628,
+ "grad_norm": 0.015000631101429462,
+ "learning_rate": 0.0006,
+ "loss": 4.408048629760742,
+ "step": 1654
+ },
+ {
+ "epoch": 22.992139737991266,
+ "grad_norm": 0.015370816923677921,
+ "learning_rate": 0.0006,
+ "loss": 4.297451019287109,
+ "step": 1655
+ },
+ {
+ "epoch": 23.0,
+ "grad_norm": 0.01709144562482834,
+ "learning_rate": 0.0006,
+ "loss": 4.294776916503906,
+ "step": 1656
+ },
+ {
+ "epoch": 23.0,
+ "eval_loss": 4.678618907928467,
+ "eval_runtime": 56.1668,
+ "eval_samples_per_second": 43.478,
+ "eval_steps_per_second": 1.371,
+ "step": 1656
+ },
+ {
+ "epoch": 23.01397379912664,
+ "grad_norm": 0.01727290451526642,
+ "learning_rate": 0.0006,
+ "loss": 4.311701774597168,
+ "step": 1657
+ },
+ {
+ "epoch": 23.027947598253274,
+ "grad_norm": 0.020645759999752045,
+ "learning_rate": 0.0006,
+ "loss": 4.429897308349609,
+ "step": 1658
+ },
+ {
+ "epoch": 23.041921397379912,
+ "grad_norm": 0.02204226516187191,
+ "learning_rate": 0.0006,
+ "loss": 4.365151882171631,
+ "step": 1659
+ },
+ {
+ "epoch": 23.05589519650655,
+ "grad_norm": 0.023069269955158234,
+ "learning_rate": 0.0006,
+ "loss": 4.3710832595825195,
+ "step": 1660
+ },
+ {
+ "epoch": 23.069868995633186,
+ "grad_norm": 0.024151179939508438,
+ "learning_rate": 0.0006,
+ "loss": 4.403108596801758,
+ "step": 1661
+ },
+ {
+ "epoch": 23.083842794759825,
+ "grad_norm": 0.02511233650147915,
+ "learning_rate": 0.0006,
+ "loss": 4.414128303527832,
+ "step": 1662
+ },
+ {
+ "epoch": 23.097816593886463,
+ "grad_norm": 0.025802113115787506,
+ "learning_rate": 0.0006,
+ "loss": 4.388265609741211,
+ "step": 1663
+ },
+ {
+ "epoch": 23.111790393013102,
+ "grad_norm": 0.02377348765730858,
+ "learning_rate": 0.0006,
+ "loss": 4.357505798339844,
+ "step": 1664
+ },
+ {
+ "epoch": 23.125764192139737,
+ "grad_norm": 0.020876651629805565,
+ "learning_rate": 0.0006,
+ "loss": 4.23043966293335,
+ "step": 1665
+ },
+ {
+ "epoch": 23.139737991266376,
+ "grad_norm": 0.01959727331995964,
+ "learning_rate": 0.0006,
+ "loss": 4.280970096588135,
+ "step": 1666
+ },
+ {
+ "epoch": 23.153711790393015,
+ "grad_norm": 0.018678616732358932,
+ "learning_rate": 0.0006,
+ "loss": 4.346716403961182,
+ "step": 1667
+ },
+ {
+ "epoch": 23.16768558951965,
+ "grad_norm": 0.02108944021165371,
+ "learning_rate": 0.0006,
+ "loss": 4.277904510498047,
+ "step": 1668
+ },
+ {
+ "epoch": 23.18165938864629,
+ "grad_norm": 0.020221566781401634,
+ "learning_rate": 0.0006,
+ "loss": 4.367693901062012,
+ "step": 1669
+ },
+ {
+ "epoch": 23.195633187772927,
+ "grad_norm": 0.019599711522459984,
+ "learning_rate": 0.0006,
+ "loss": 4.257092475891113,
+ "step": 1670
+ },
+ {
+ "epoch": 23.209606986899562,
+ "grad_norm": 0.019833361729979515,
+ "learning_rate": 0.0006,
+ "loss": 4.540737628936768,
+ "step": 1671
+ },
+ {
+ "epoch": 23.2235807860262,
+ "grad_norm": 0.020352095365524292,
+ "learning_rate": 0.0006,
+ "loss": 4.360923767089844,
+ "step": 1672
+ },
+ {
+ "epoch": 23.23755458515284,
+ "grad_norm": 0.01939748041331768,
+ "learning_rate": 0.0006,
+ "loss": 4.393110275268555,
+ "step": 1673
+ },
+ {
+ "epoch": 23.251528384279474,
+ "grad_norm": 0.01884354278445244,
+ "learning_rate": 0.0006,
+ "loss": 4.383938789367676,
+ "step": 1674
+ },
+ {
+ "epoch": 23.265502183406113,
+ "grad_norm": 0.02008470520377159,
+ "learning_rate": 0.0006,
+ "loss": 4.346344947814941,
+ "step": 1675
+ },
+ {
+ "epoch": 23.27947598253275,
+ "grad_norm": 0.019065195694565773,
+ "learning_rate": 0.0006,
+ "loss": 4.36611795425415,
+ "step": 1676
+ },
+ {
+ "epoch": 23.29344978165939,
+ "grad_norm": 0.01938541606068611,
+ "learning_rate": 0.0006,
+ "loss": 4.4009599685668945,
+ "step": 1677
+ },
+ {
+ "epoch": 23.307423580786025,
+ "grad_norm": 0.019044723361730576,
+ "learning_rate": 0.0006,
+ "loss": 4.340169906616211,
+ "step": 1678
+ },
+ {
+ "epoch": 23.321397379912664,
+ "grad_norm": 0.018850378692150116,
+ "learning_rate": 0.0006,
+ "loss": 4.37674617767334,
+ "step": 1679
+ },
+ {
+ "epoch": 23.335371179039303,
+ "grad_norm": 0.019076095893979073,
+ "learning_rate": 0.0006,
+ "loss": 4.226874351501465,
+ "step": 1680
+ },
+ {
+ "epoch": 23.349344978165938,
+ "grad_norm": 0.017543038353323936,
+ "learning_rate": 0.0006,
+ "loss": 4.337751388549805,
+ "step": 1681
+ },
+ {
+ "epoch": 23.363318777292577,
+ "grad_norm": 0.017170218750834465,
+ "learning_rate": 0.0006,
+ "loss": 4.282512664794922,
+ "step": 1682
+ },
+ {
+ "epoch": 23.377292576419215,
+ "grad_norm": 0.01574796624481678,
+ "learning_rate": 0.0006,
+ "loss": 4.30638313293457,
+ "step": 1683
+ },
+ {
+ "epoch": 23.39126637554585,
+ "grad_norm": 0.015160846523940563,
+ "learning_rate": 0.0006,
+ "loss": 4.336716175079346,
+ "step": 1684
+ },
+ {
+ "epoch": 23.40524017467249,
+ "grad_norm": 0.01534526702016592,
+ "learning_rate": 0.0006,
+ "loss": 4.2174787521362305,
+ "step": 1685
+ },
+ {
+ "epoch": 23.419213973799128,
+ "grad_norm": 0.015289999544620514,
+ "learning_rate": 0.0006,
+ "loss": 4.272546291351318,
+ "step": 1686
+ },
+ {
+ "epoch": 23.433187772925763,
+ "grad_norm": 0.014894185587763786,
+ "learning_rate": 0.0006,
+ "loss": 4.296045303344727,
+ "step": 1687
+ },
+ {
+ "epoch": 23.4471615720524,
+ "grad_norm": 0.014869502745568752,
+ "learning_rate": 0.0006,
+ "loss": 4.425461769104004,
+ "step": 1688
+ },
+ {
+ "epoch": 23.46113537117904,
+ "grad_norm": 0.01474483497440815,
+ "learning_rate": 0.0006,
+ "loss": 4.361867427825928,
+ "step": 1689
+ },
+ {
+ "epoch": 23.475109170305675,
+ "grad_norm": 0.015299245715141296,
+ "learning_rate": 0.0006,
+ "loss": 4.364523410797119,
+ "step": 1690
+ },
+ {
+ "epoch": 23.489082969432314,
+ "grad_norm": 0.015201340429484844,
+ "learning_rate": 0.0006,
+ "loss": 4.285182952880859,
+ "step": 1691
+ },
+ {
+ "epoch": 23.503056768558952,
+ "grad_norm": 0.016057293862104416,
+ "learning_rate": 0.0006,
+ "loss": 4.284873962402344,
+ "step": 1692
+ },
+ {
+ "epoch": 23.51703056768559,
+ "grad_norm": 0.015871062874794006,
+ "learning_rate": 0.0006,
+ "loss": 4.326470851898193,
+ "step": 1693
+ },
+ {
+ "epoch": 23.531004366812226,
+ "grad_norm": 0.01586288772523403,
+ "learning_rate": 0.0006,
+ "loss": 4.324883460998535,
+ "step": 1694
+ },
+ {
+ "epoch": 23.544978165938865,
+ "grad_norm": 0.016092827543616295,
+ "learning_rate": 0.0006,
+ "loss": 4.3258867263793945,
+ "step": 1695
+ },
+ {
+ "epoch": 23.558951965065503,
+ "grad_norm": 0.01582016795873642,
+ "learning_rate": 0.0006,
+ "loss": 4.358892440795898,
+ "step": 1696
+ },
+ {
+ "epoch": 23.57292576419214,
+ "grad_norm": 0.01688452623784542,
+ "learning_rate": 0.0006,
+ "loss": 4.280842304229736,
+ "step": 1697
+ },
+ {
+ "epoch": 23.586899563318777,
+ "grad_norm": 0.015865886583924294,
+ "learning_rate": 0.0006,
+ "loss": 4.245567321777344,
+ "step": 1698
+ },
+ {
+ "epoch": 23.600873362445416,
+ "grad_norm": 0.015303662046790123,
+ "learning_rate": 0.0006,
+ "loss": 4.383112907409668,
+ "step": 1699
+ },
+ {
+ "epoch": 23.61484716157205,
+ "grad_norm": 0.015649516135454178,
+ "learning_rate": 0.0006,
+ "loss": 4.519082069396973,
+ "step": 1700
+ },
+ {
+ "epoch": 23.62882096069869,
+ "grad_norm": 0.0167120099067688,
+ "learning_rate": 0.0006,
+ "loss": 4.447847366333008,
+ "step": 1701
+ },
+ {
+ "epoch": 23.64279475982533,
+ "grad_norm": 0.016368716955184937,
+ "learning_rate": 0.0006,
+ "loss": 4.434186935424805,
+ "step": 1702
+ },
+ {
+ "epoch": 23.656768558951963,
+ "grad_norm": 0.017379503697156906,
+ "learning_rate": 0.0006,
+ "loss": 4.317594528198242,
+ "step": 1703
+ },
+ {
+ "epoch": 23.670742358078602,
+ "grad_norm": 0.018573811277747154,
+ "learning_rate": 0.0006,
+ "loss": 4.205793857574463,
+ "step": 1704
+ },
+ {
+ "epoch": 23.68471615720524,
+ "grad_norm": 0.01749090477824211,
+ "learning_rate": 0.0006,
+ "loss": 4.3985700607299805,
+ "step": 1705
+ },
+ {
+ "epoch": 23.69868995633188,
+ "grad_norm": 0.0165668074041605,
+ "learning_rate": 0.0006,
+ "loss": 4.329226970672607,
+ "step": 1706
+ },
+ {
+ "epoch": 23.712663755458514,
+ "grad_norm": 0.016913846135139465,
+ "learning_rate": 0.0006,
+ "loss": 4.388485908508301,
+ "step": 1707
+ },
+ {
+ "epoch": 23.726637554585153,
+ "grad_norm": 0.016813859343528748,
+ "learning_rate": 0.0006,
+ "loss": 4.3395915031433105,
+ "step": 1708
+ },
+ {
+ "epoch": 23.74061135371179,
+ "grad_norm": 0.01676975190639496,
+ "learning_rate": 0.0006,
+ "loss": 4.254745960235596,
+ "step": 1709
+ },
+ {
+ "epoch": 23.754585152838427,
+ "grad_norm": 0.016370350494980812,
+ "learning_rate": 0.0006,
+ "loss": 4.2612223625183105,
+ "step": 1710
+ },
+ {
+ "epoch": 23.768558951965066,
+ "grad_norm": 0.01696198247373104,
+ "learning_rate": 0.0006,
+ "loss": 4.327524662017822,
+ "step": 1711
+ },
+ {
+ "epoch": 23.782532751091704,
+ "grad_norm": 0.0189108457416296,
+ "learning_rate": 0.0006,
+ "loss": 4.325839519500732,
+ "step": 1712
+ },
+ {
+ "epoch": 23.79650655021834,
+ "grad_norm": 0.01862351782619953,
+ "learning_rate": 0.0006,
+ "loss": 4.333662509918213,
+ "step": 1713
+ },
+ {
+ "epoch": 23.810480349344978,
+ "grad_norm": 0.01655014418065548,
+ "learning_rate": 0.0006,
+ "loss": 4.39305305480957,
+ "step": 1714
+ },
+ {
+ "epoch": 23.824454148471617,
+ "grad_norm": 0.015899376943707466,
+ "learning_rate": 0.0006,
+ "loss": 4.390053749084473,
+ "step": 1715
+ },
+ {
+ "epoch": 23.83842794759825,
+ "grad_norm": 0.016519056633114815,
+ "learning_rate": 0.0006,
+ "loss": 4.255903244018555,
+ "step": 1716
+ },
+ {
+ "epoch": 23.85240174672489,
+ "grad_norm": 0.0167539119720459,
+ "learning_rate": 0.0006,
+ "loss": 4.255727767944336,
+ "step": 1717
+ },
+ {
+ "epoch": 23.86637554585153,
+ "grad_norm": 0.0171353816986084,
+ "learning_rate": 0.0006,
+ "loss": 4.376740455627441,
+ "step": 1718
+ },
+ {
+ "epoch": 23.880349344978168,
+ "grad_norm": 0.016719117760658264,
+ "learning_rate": 0.0006,
+ "loss": 4.3376688957214355,
+ "step": 1719
+ },
+ {
+ "epoch": 23.894323144104803,
+ "grad_norm": 0.01561494916677475,
+ "learning_rate": 0.0006,
+ "loss": 4.3345746994018555,
+ "step": 1720
+ },
+ {
+ "epoch": 23.90829694323144,
+ "grad_norm": 0.016303174197673798,
+ "learning_rate": 0.0006,
+ "loss": 4.332371234893799,
+ "step": 1721
+ },
+ {
+ "epoch": 23.92227074235808,
+ "grad_norm": 0.016722865402698517,
+ "learning_rate": 0.0006,
+ "loss": 4.267125129699707,
+ "step": 1722
+ },
+ {
+ "epoch": 23.936244541484715,
+ "grad_norm": 0.017072124406695366,
+ "learning_rate": 0.0006,
+ "loss": 4.322751998901367,
+ "step": 1723
+ },
+ {
+ "epoch": 23.950218340611354,
+ "grad_norm": 0.01699932850897312,
+ "learning_rate": 0.0006,
+ "loss": 4.378422737121582,
+ "step": 1724
+ },
+ {
+ "epoch": 23.964192139737992,
+ "grad_norm": 0.01742720976471901,
+ "learning_rate": 0.0006,
+ "loss": 4.345083236694336,
+ "step": 1725
+ },
+ {
+ "epoch": 23.978165938864628,
+ "grad_norm": 0.018797501921653748,
+ "learning_rate": 0.0006,
+ "loss": 4.290443420410156,
+ "step": 1726
+ },
+ {
+ "epoch": 23.992139737991266,
+ "grad_norm": 0.018090683966875076,
+ "learning_rate": 0.0006,
+ "loss": 4.377580642700195,
+ "step": 1727
+ },
+ {
+ "epoch": 24.0,
+ "grad_norm": 0.017305459827184677,
+ "learning_rate": 0.0006,
+ "loss": 4.364426612854004,
+ "step": 1728
+ },
+ {
+ "epoch": 24.0,
+ "eval_loss": 4.647762298583984,
+ "eval_runtime": 56.5402,
+ "eval_samples_per_second": 43.191,
+ "eval_steps_per_second": 1.362,
+ "step": 1728
+ },
+ {
+ "epoch": 24.01397379912664,
+ "grad_norm": 0.016493607312440872,
+ "learning_rate": 0.0006,
+ "loss": 4.252255916595459,
+ "step": 1729
+ },
+ {
+ "epoch": 24.027947598253274,
+ "grad_norm": 0.016529450193047523,
+ "learning_rate": 0.0006,
+ "loss": 4.352746963500977,
+ "step": 1730
+ },
+ {
+ "epoch": 24.041921397379912,
+ "grad_norm": 0.016940191388130188,
+ "learning_rate": 0.0006,
+ "loss": 4.301870822906494,
+ "step": 1731
+ },
+ {
+ "epoch": 24.05589519650655,
+ "grad_norm": 0.01752558909356594,
+ "learning_rate": 0.0006,
+ "loss": 4.2762451171875,
+ "step": 1732
+ },
+ {
+ "epoch": 24.069868995633186,
+ "grad_norm": 0.017112884670495987,
+ "learning_rate": 0.0006,
+ "loss": 4.2659149169921875,
+ "step": 1733
+ },
+ {
+ "epoch": 24.083842794759825,
+ "grad_norm": 0.01900586113333702,
+ "learning_rate": 0.0006,
+ "loss": 4.306058883666992,
+ "step": 1734
+ },
+ {
+ "epoch": 24.097816593886463,
+ "grad_norm": 0.019447680562734604,
+ "learning_rate": 0.0006,
+ "loss": 4.450128555297852,
+ "step": 1735
+ },
+ {
+ "epoch": 24.111790393013102,
+ "grad_norm": 0.01948186755180359,
+ "learning_rate": 0.0006,
+ "loss": 4.202869415283203,
+ "step": 1736
+ },
+ {
+ "epoch": 24.125764192139737,
+ "grad_norm": 0.01716383546590805,
+ "learning_rate": 0.0006,
+ "loss": 4.318876266479492,
+ "step": 1737
+ },
+ {
+ "epoch": 24.139737991266376,
+ "grad_norm": 0.017049988731741905,
+ "learning_rate": 0.0006,
+ "loss": 4.237238883972168,
+ "step": 1738
+ },
+ {
+ "epoch": 24.153711790393015,
+ "grad_norm": 0.015169923193752766,
+ "learning_rate": 0.0006,
+ "loss": 4.289426803588867,
+ "step": 1739
+ },
+ {
+ "epoch": 24.16768558951965,
+ "grad_norm": 0.014760667458176613,
+ "learning_rate": 0.0006,
+ "loss": 4.356924057006836,
+ "step": 1740
+ },
+ {
+ "epoch": 24.18165938864629,
+ "grad_norm": 0.015127750113606453,
+ "learning_rate": 0.0006,
+ "loss": 4.142916679382324,
+ "step": 1741
+ },
+ {
+ "epoch": 24.195633187772927,
+ "grad_norm": 0.01549555640667677,
+ "learning_rate": 0.0006,
+ "loss": 4.345563888549805,
+ "step": 1742
+ },
+ {
+ "epoch": 24.209606986899562,
+ "grad_norm": 0.016393091529607773,
+ "learning_rate": 0.0006,
+ "loss": 4.417495250701904,
+ "step": 1743
+ },
+ {
+ "epoch": 24.2235807860262,
+ "grad_norm": 0.01785266026854515,
+ "learning_rate": 0.0006,
+ "loss": 4.3553547859191895,
+ "step": 1744
+ },
+ {
+ "epoch": 24.23755458515284,
+ "grad_norm": 0.017574617639183998,
+ "learning_rate": 0.0006,
+ "loss": 4.241024494171143,
+ "step": 1745
+ },
+ {
+ "epoch": 24.251528384279474,
+ "grad_norm": 0.019060306251049042,
+ "learning_rate": 0.0006,
+ "loss": 4.3685150146484375,
+ "step": 1746
+ },
+ {
+ "epoch": 24.265502183406113,
+ "grad_norm": 0.0214633010327816,
+ "learning_rate": 0.0006,
+ "loss": 4.323866844177246,
+ "step": 1747
+ },
+ {
+ "epoch": 24.27947598253275,
+ "grad_norm": 0.021055003628134727,
+ "learning_rate": 0.0006,
+ "loss": 4.278757095336914,
+ "step": 1748
+ },
+ {
+ "epoch": 24.29344978165939,
+ "grad_norm": 0.018618572503328323,
+ "learning_rate": 0.0006,
+ "loss": 4.28849983215332,
+ "step": 1749
+ },
+ {
+ "epoch": 24.307423580786025,
+ "grad_norm": 0.018666435033082962,
+ "learning_rate": 0.0006,
+ "loss": 4.368590354919434,
+ "step": 1750
+ },
+ {
+ "epoch": 24.321397379912664,
+ "grad_norm": 0.020626146346330643,
+ "learning_rate": 0.0006,
+ "loss": 4.2951836585998535,
+ "step": 1751
+ },
+ {
+ "epoch": 24.335371179039303,
+ "grad_norm": 0.01929759792983532,
+ "learning_rate": 0.0006,
+ "loss": 4.3664703369140625,
+ "step": 1752
+ },
+ {
+ "epoch": 24.349344978165938,
+ "grad_norm": 0.020956002175807953,
+ "learning_rate": 0.0006,
+ "loss": 4.345308303833008,
+ "step": 1753
+ },
+ {
+ "epoch": 24.363318777292577,
+ "grad_norm": 0.019227957352995872,
+ "learning_rate": 0.0006,
+ "loss": 4.349849224090576,
+ "step": 1754
+ },
+ {
+ "epoch": 24.377292576419215,
+ "grad_norm": 0.015970097854733467,
+ "learning_rate": 0.0006,
+ "loss": 4.223012447357178,
+ "step": 1755
+ },
+ {
+ "epoch": 24.39126637554585,
+ "grad_norm": 0.015951646491885185,
+ "learning_rate": 0.0006,
+ "loss": 4.236420631408691,
+ "step": 1756
+ },
+ {
+ "epoch": 24.40524017467249,
+ "grad_norm": 0.016032515093684196,
+ "learning_rate": 0.0006,
+ "loss": 4.226899147033691,
+ "step": 1757
+ },
+ {
+ "epoch": 24.419213973799128,
+ "grad_norm": 0.01573132909834385,
+ "learning_rate": 0.0006,
+ "loss": 4.315916061401367,
+ "step": 1758
+ },
+ {
+ "epoch": 24.433187772925763,
+ "grad_norm": 0.015367005951702595,
+ "learning_rate": 0.0006,
+ "loss": 4.358939170837402,
+ "step": 1759
+ },
+ {
+ "epoch": 24.4471615720524,
+ "grad_norm": 0.01669907011091709,
+ "learning_rate": 0.0006,
+ "loss": 4.232147216796875,
+ "step": 1760
+ },
+ {
+ "epoch": 24.46113537117904,
+ "grad_norm": 0.018708152696490288,
+ "learning_rate": 0.0006,
+ "loss": 4.329852104187012,
+ "step": 1761
+ },
+ {
+ "epoch": 24.475109170305675,
+ "grad_norm": 0.019035987555980682,
+ "learning_rate": 0.0006,
+ "loss": 4.260340213775635,
+ "step": 1762
+ },
+ {
+ "epoch": 24.489082969432314,
+ "grad_norm": 0.01832144521176815,
+ "learning_rate": 0.0006,
+ "loss": 4.359543800354004,
+ "step": 1763
+ },
+ {
+ "epoch": 24.503056768558952,
+ "grad_norm": 0.015313294716179371,
+ "learning_rate": 0.0006,
+ "loss": 4.279088020324707,
+ "step": 1764
+ },
+ {
+ "epoch": 24.51703056768559,
+ "grad_norm": 0.015803923830389977,
+ "learning_rate": 0.0006,
+ "loss": 4.230682849884033,
+ "step": 1765
+ },
+ {
+ "epoch": 24.531004366812226,
+ "grad_norm": 0.016748374328017235,
+ "learning_rate": 0.0006,
+ "loss": 4.349661827087402,
+ "step": 1766
+ },
+ {
+ "epoch": 24.544978165938865,
+ "grad_norm": 0.015732428058981895,
+ "learning_rate": 0.0006,
+ "loss": 4.372441291809082,
+ "step": 1767
+ },
+ {
+ "epoch": 24.558951965065503,
+ "grad_norm": 0.016564829275012016,
+ "learning_rate": 0.0006,
+ "loss": 4.323090553283691,
+ "step": 1768
+ },
+ {
+ "epoch": 24.57292576419214,
+ "grad_norm": 0.01603010483086109,
+ "learning_rate": 0.0006,
+ "loss": 4.347898006439209,
+ "step": 1769
+ },
+ {
+ "epoch": 24.586899563318777,
+ "grad_norm": 0.01674809865653515,
+ "learning_rate": 0.0006,
+ "loss": 4.269165515899658,
+ "step": 1770
+ },
+ {
+ "epoch": 24.600873362445416,
+ "grad_norm": 0.017333930358290672,
+ "learning_rate": 0.0006,
+ "loss": 4.353298187255859,
+ "step": 1771
+ },
+ {
+ "epoch": 24.61484716157205,
+ "grad_norm": 0.017519580200314522,
+ "learning_rate": 0.0006,
+ "loss": 4.265199661254883,
+ "step": 1772
+ },
+ {
+ "epoch": 24.62882096069869,
+ "grad_norm": 0.017675306648015976,
+ "learning_rate": 0.0006,
+ "loss": 4.322596549987793,
+ "step": 1773
+ },
+ {
+ "epoch": 24.64279475982533,
+ "grad_norm": 0.017166534438729286,
+ "learning_rate": 0.0006,
+ "loss": 4.413900375366211,
+ "step": 1774
+ },
+ {
+ "epoch": 24.656768558951963,
+ "grad_norm": 0.016362346708774567,
+ "learning_rate": 0.0006,
+ "loss": 4.307236671447754,
+ "step": 1775
+ },
+ {
+ "epoch": 24.670742358078602,
+ "grad_norm": 0.016086872667074203,
+ "learning_rate": 0.0006,
+ "loss": 4.235316276550293,
+ "step": 1776
+ },
+ {
+ "epoch": 24.68471615720524,
+ "grad_norm": 0.01689537614583969,
+ "learning_rate": 0.0006,
+ "loss": 4.390591621398926,
+ "step": 1777
+ },
+ {
+ "epoch": 24.69868995633188,
+ "grad_norm": 0.016836725175380707,
+ "learning_rate": 0.0006,
+ "loss": 4.2990031242370605,
+ "step": 1778
+ },
+ {
+ "epoch": 24.712663755458514,
+ "grad_norm": 0.015627335757017136,
+ "learning_rate": 0.0006,
+ "loss": 4.166745185852051,
+ "step": 1779
+ },
+ {
+ "epoch": 24.726637554585153,
+ "grad_norm": 0.015007016249001026,
+ "learning_rate": 0.0006,
+ "loss": 4.295273303985596,
+ "step": 1780
+ },
+ {
+ "epoch": 24.74061135371179,
+ "grad_norm": 0.016084209084510803,
+ "learning_rate": 0.0006,
+ "loss": 4.342424392700195,
+ "step": 1781
+ },
+ {
+ "epoch": 24.754585152838427,
+ "grad_norm": 0.01595073565840721,
+ "learning_rate": 0.0006,
+ "loss": 4.3060102462768555,
+ "step": 1782
+ },
+ {
+ "epoch": 24.768558951965066,
+ "grad_norm": 0.016113461926579475,
+ "learning_rate": 0.0006,
+ "loss": 4.267908096313477,
+ "step": 1783
+ },
+ {
+ "epoch": 24.782532751091704,
+ "grad_norm": 0.01564556173980236,
+ "learning_rate": 0.0006,
+ "loss": 4.324642658233643,
+ "step": 1784
+ },
+ {
+ "epoch": 24.79650655021834,
+ "grad_norm": 0.01575268618762493,
+ "learning_rate": 0.0006,
+ "loss": 4.328181743621826,
+ "step": 1785
+ },
+ {
+ "epoch": 24.810480349344978,
+ "grad_norm": 0.016830824315547943,
+ "learning_rate": 0.0006,
+ "loss": 4.247906684875488,
+ "step": 1786
+ },
+ {
+ "epoch": 24.824454148471617,
+ "grad_norm": 0.017989547923207283,
+ "learning_rate": 0.0006,
+ "loss": 4.228281021118164,
+ "step": 1787
+ },
+ {
+ "epoch": 24.83842794759825,
+ "grad_norm": 0.016783086583018303,
+ "learning_rate": 0.0006,
+ "loss": 4.2880964279174805,
+ "step": 1788
+ },
+ {
+ "epoch": 24.85240174672489,
+ "grad_norm": 0.016467561945319176,
+ "learning_rate": 0.0006,
+ "loss": 4.36118745803833,
+ "step": 1789
+ },
+ {
+ "epoch": 24.86637554585153,
+ "grad_norm": 0.01780523732304573,
+ "learning_rate": 0.0006,
+ "loss": 4.172321319580078,
+ "step": 1790
+ },
+ {
+ "epoch": 24.880349344978168,
+ "grad_norm": 0.017102031037211418,
+ "learning_rate": 0.0006,
+ "loss": 4.352826118469238,
+ "step": 1791
+ },
+ {
+ "epoch": 24.894323144104803,
+ "grad_norm": 0.016409218311309814,
+ "learning_rate": 0.0006,
+ "loss": 4.212162017822266,
+ "step": 1792
+ },
+ {
+ "epoch": 24.90829694323144,
+ "grad_norm": 0.018660584464669228,
+ "learning_rate": 0.0006,
+ "loss": 4.346684455871582,
+ "step": 1793
+ },
+ {
+ "epoch": 24.92227074235808,
+ "grad_norm": 0.017389440909028053,
+ "learning_rate": 0.0006,
+ "loss": 4.197913646697998,
+ "step": 1794
+ },
+ {
+ "epoch": 24.936244541484715,
+ "grad_norm": 0.01596485823392868,
+ "learning_rate": 0.0006,
+ "loss": 4.370617866516113,
+ "step": 1795
+ },
+ {
+ "epoch": 24.950218340611354,
+ "grad_norm": 0.01709776371717453,
+ "learning_rate": 0.0006,
+ "loss": 4.313847064971924,
+ "step": 1796
+ },
+ {
+ "epoch": 24.964192139737992,
+ "grad_norm": 0.016571134328842163,
+ "learning_rate": 0.0006,
+ "loss": 4.1895856857299805,
+ "step": 1797
+ },
+ {
+ "epoch": 24.978165938864628,
+ "grad_norm": 0.016851048916578293,
+ "learning_rate": 0.0006,
+ "loss": 4.350772857666016,
+ "step": 1798
+ },
+ {
+ "epoch": 24.992139737991266,
+ "grad_norm": 0.017007440328598022,
+ "learning_rate": 0.0006,
+ "loss": 4.399008750915527,
+ "step": 1799
+ },
+ {
+ "epoch": 25.0,
+ "grad_norm": 0.01791212148964405,
+ "learning_rate": 0.0006,
+ "loss": 4.306121826171875,
+ "step": 1800
+ },
+ {
+ "epoch": 25.0,
+ "eval_loss": 4.636417865753174,
+ "eval_runtime": 56.8296,
+ "eval_samples_per_second": 42.971,
+ "eval_steps_per_second": 1.355,
+ "step": 1800
+ },
+ {
+ "epoch": 25.01397379912664,
+ "grad_norm": 0.01729617640376091,
+ "learning_rate": 0.0006,
+ "loss": 4.257018089294434,
+ "step": 1801
+ },
+ {
+ "epoch": 25.027947598253274,
+ "grad_norm": 0.017466643825173378,
+ "learning_rate": 0.0006,
+ "loss": 4.307605266571045,
+ "step": 1802
+ },
+ {
+ "epoch": 25.041921397379912,
+ "grad_norm": 0.018614279106259346,
+ "learning_rate": 0.0006,
+ "loss": 4.360141277313232,
+ "step": 1803
+ },
+ {
+ "epoch": 25.05589519650655,
+ "grad_norm": 0.02010771445930004,
+ "learning_rate": 0.0006,
+ "loss": 4.196428298950195,
+ "step": 1804
+ },
+ {
+ "epoch": 25.069868995633186,
+ "grad_norm": 0.02090228535234928,
+ "learning_rate": 0.0006,
+ "loss": 4.288947105407715,
+ "step": 1805
+ },
+ {
+ "epoch": 25.083842794759825,
+ "grad_norm": 0.02050255797803402,
+ "learning_rate": 0.0006,
+ "loss": 4.259582042694092,
+ "step": 1806
+ },
+ {
+ "epoch": 25.097816593886463,
+ "grad_norm": 0.021844089031219482,
+ "learning_rate": 0.0006,
+ "loss": 4.405138969421387,
+ "step": 1807
+ },
+ {
+ "epoch": 25.111790393013102,
+ "grad_norm": 0.022287018597126007,
+ "learning_rate": 0.0006,
+ "loss": 4.333793640136719,
+ "step": 1808
+ },
+ {
+ "epoch": 25.125764192139737,
+ "grad_norm": 0.02288329415023327,
+ "learning_rate": 0.0006,
+ "loss": 4.3910980224609375,
+ "step": 1809
+ },
+ {
+ "epoch": 25.139737991266376,
+ "grad_norm": 0.02373199723660946,
+ "learning_rate": 0.0006,
+ "loss": 4.32921028137207,
+ "step": 1810
+ },
+ {
+ "epoch": 25.153711790393015,
+ "grad_norm": 0.024960234761238098,
+ "learning_rate": 0.0006,
+ "loss": 4.295609474182129,
+ "step": 1811
+ },
+ {
+ "epoch": 25.16768558951965,
+ "grad_norm": 0.026743892580270767,
+ "learning_rate": 0.0006,
+ "loss": 4.315308094024658,
+ "step": 1812
+ },
+ {
+ "epoch": 25.18165938864629,
+ "grad_norm": 0.022138185799121857,
+ "learning_rate": 0.0006,
+ "loss": 4.360208511352539,
+ "step": 1813
+ },
+ {
+ "epoch": 25.195633187772927,
+ "grad_norm": 0.022120574489235878,
+ "learning_rate": 0.0006,
+ "loss": 4.1973772048950195,
+ "step": 1814
+ },
+ {
+ "epoch": 25.209606986899562,
+ "grad_norm": 0.02230502851307392,
+ "learning_rate": 0.0006,
+ "loss": 4.285558700561523,
+ "step": 1815
+ },
+ {
+ "epoch": 25.2235807860262,
+ "grad_norm": 0.022301986813545227,
+ "learning_rate": 0.0006,
+ "loss": 4.238343238830566,
+ "step": 1816
+ },
+ {
+ "epoch": 25.23755458515284,
+ "grad_norm": 0.020050447434186935,
+ "learning_rate": 0.0006,
+ "loss": 4.298235893249512,
+ "step": 1817
+ },
+ {
+ "epoch": 25.251528384279474,
+ "grad_norm": 0.01987724006175995,
+ "learning_rate": 0.0006,
+ "loss": 4.301384449005127,
+ "step": 1818
+ },
+ {
+ "epoch": 25.265502183406113,
+ "grad_norm": 0.020440855994820595,
+ "learning_rate": 0.0006,
+ "loss": 4.269218444824219,
+ "step": 1819
+ },
+ {
+ "epoch": 25.27947598253275,
+ "grad_norm": 0.01891893707215786,
+ "learning_rate": 0.0006,
+ "loss": 4.2401957511901855,
+ "step": 1820
+ },
+ {
+ "epoch": 25.29344978165939,
+ "grad_norm": 0.01903688907623291,
+ "learning_rate": 0.0006,
+ "loss": 4.139670372009277,
+ "step": 1821
+ },
+ {
+ "epoch": 25.307423580786025,
+ "grad_norm": 0.018856938928365707,
+ "learning_rate": 0.0006,
+ "loss": 4.233214378356934,
+ "step": 1822
+ },
+ {
+ "epoch": 25.321397379912664,
+ "grad_norm": 0.017791323363780975,
+ "learning_rate": 0.0006,
+ "loss": 4.2342023849487305,
+ "step": 1823
+ },
+ {
+ "epoch": 25.335371179039303,
+ "grad_norm": 0.017304565757513046,
+ "learning_rate": 0.0006,
+ "loss": 4.355198860168457,
+ "step": 1824
+ },
+ {
+ "epoch": 25.349344978165938,
+ "grad_norm": 0.01576206088066101,
+ "learning_rate": 0.0006,
+ "loss": 4.114147186279297,
+ "step": 1825
+ },
+ {
+ "epoch": 25.363318777292577,
+ "grad_norm": 0.015105154365301132,
+ "learning_rate": 0.0006,
+ "loss": 4.313795566558838,
+ "step": 1826
+ },
+ {
+ "epoch": 25.377292576419215,
+ "grad_norm": 0.01563587784767151,
+ "learning_rate": 0.0006,
+ "loss": 4.308282852172852,
+ "step": 1827
+ },
+ {
+ "epoch": 25.39126637554585,
+ "grad_norm": 0.015980906784534454,
+ "learning_rate": 0.0006,
+ "loss": 4.345149040222168,
+ "step": 1828
+ },
+ {
+ "epoch": 25.40524017467249,
+ "grad_norm": 0.01669284701347351,
+ "learning_rate": 0.0006,
+ "loss": 4.237611293792725,
+ "step": 1829
+ },
+ {
+ "epoch": 25.419213973799128,
+ "grad_norm": 0.016202064231038094,
+ "learning_rate": 0.0006,
+ "loss": 4.3072309494018555,
+ "step": 1830
+ },
+ {
+ "epoch": 25.433187772925763,
+ "grad_norm": 0.01478662807494402,
+ "learning_rate": 0.0006,
+ "loss": 4.276215076446533,
+ "step": 1831
+ },
+ {
+ "epoch": 25.4471615720524,
+ "grad_norm": 0.015168731100857258,
+ "learning_rate": 0.0006,
+ "loss": 4.37509822845459,
+ "step": 1832
+ },
+ {
+ "epoch": 25.46113537117904,
+ "grad_norm": 0.01648980937898159,
+ "learning_rate": 0.0006,
+ "loss": 4.349167346954346,
+ "step": 1833
+ },
+ {
+ "epoch": 25.475109170305675,
+ "grad_norm": 0.01708398200571537,
+ "learning_rate": 0.0006,
+ "loss": 4.346580505371094,
+ "step": 1834
+ },
+ {
+ "epoch": 25.489082969432314,
+ "grad_norm": 0.017370784655213356,
+ "learning_rate": 0.0006,
+ "loss": 4.299466133117676,
+ "step": 1835
+ },
+ {
+ "epoch": 25.503056768558952,
+ "grad_norm": 0.016248228028416634,
+ "learning_rate": 0.0006,
+ "loss": 4.184610366821289,
+ "step": 1836
+ },
+ {
+ "epoch": 25.51703056768559,
+ "grad_norm": 0.017106913030147552,
+ "learning_rate": 0.0006,
+ "loss": 4.366856098175049,
+ "step": 1837
+ },
+ {
+ "epoch": 25.531004366812226,
+ "grad_norm": 0.016071965917944908,
+ "learning_rate": 0.0006,
+ "loss": 4.297691345214844,
+ "step": 1838
+ },
+ {
+ "epoch": 25.544978165938865,
+ "grad_norm": 0.016451986506581306,
+ "learning_rate": 0.0006,
+ "loss": 4.229983329772949,
+ "step": 1839
+ },
+ {
+ "epoch": 25.558951965065503,
+ "grad_norm": 0.016195081174373627,
+ "learning_rate": 0.0006,
+ "loss": 4.259893417358398,
+ "step": 1840
+ },
+ {
+ "epoch": 25.57292576419214,
+ "grad_norm": 0.0165175162255764,
+ "learning_rate": 0.0006,
+ "loss": 4.3348236083984375,
+ "step": 1841
+ },
+ {
+ "epoch": 25.586899563318777,
+ "grad_norm": 0.01651517115533352,
+ "learning_rate": 0.0006,
+ "loss": 4.319541931152344,
+ "step": 1842
+ },
+ {
+ "epoch": 25.600873362445416,
+ "grad_norm": 0.016143690794706345,
+ "learning_rate": 0.0006,
+ "loss": 4.345652103424072,
+ "step": 1843
+ },
+ {
+ "epoch": 25.61484716157205,
+ "grad_norm": 0.015140696428716183,
+ "learning_rate": 0.0006,
+ "loss": 4.212047576904297,
+ "step": 1844
+ },
+ {
+ "epoch": 25.62882096069869,
+ "grad_norm": 0.014637443237006664,
+ "learning_rate": 0.0006,
+ "loss": 4.187453269958496,
+ "step": 1845
+ },
+ {
+ "epoch": 25.64279475982533,
+ "grad_norm": 0.014585614204406738,
+ "learning_rate": 0.0006,
+ "loss": 4.253936767578125,
+ "step": 1846
+ },
+ {
+ "epoch": 25.656768558951963,
+ "grad_norm": 0.015158289112150669,
+ "learning_rate": 0.0006,
+ "loss": 4.34612512588501,
+ "step": 1847
+ },
+ {
+ "epoch": 25.670742358078602,
+ "grad_norm": 0.015202849172055721,
+ "learning_rate": 0.0006,
+ "loss": 4.270634651184082,
+ "step": 1848
+ },
+ {
+ "epoch": 25.68471615720524,
+ "grad_norm": 0.015363575890660286,
+ "learning_rate": 0.0006,
+ "loss": 4.271080493927002,
+ "step": 1849
+ },
+ {
+ "epoch": 25.69868995633188,
+ "grad_norm": 0.014360226690769196,
+ "learning_rate": 0.0006,
+ "loss": 4.375033855438232,
+ "step": 1850
+ },
+ {
+ "epoch": 25.712663755458514,
+ "grad_norm": 0.014263181947171688,
+ "learning_rate": 0.0006,
+ "loss": 4.424138069152832,
+ "step": 1851
+ },
+ {
+ "epoch": 25.726637554585153,
+ "grad_norm": 0.014398688450455666,
+ "learning_rate": 0.0006,
+ "loss": 4.410154819488525,
+ "step": 1852
+ },
+ {
+ "epoch": 25.74061135371179,
+ "grad_norm": 0.016131488606333733,
+ "learning_rate": 0.0006,
+ "loss": 4.324014663696289,
+ "step": 1853
+ },
+ {
+ "epoch": 25.754585152838427,
+ "grad_norm": 0.017568735405802727,
+ "learning_rate": 0.0006,
+ "loss": 4.367861747741699,
+ "step": 1854
+ },
+ {
+ "epoch": 25.768558951965066,
+ "grad_norm": 0.017907511442899704,
+ "learning_rate": 0.0006,
+ "loss": 4.352965354919434,
+ "step": 1855
+ },
+ {
+ "epoch": 25.782532751091704,
+ "grad_norm": 0.016918016597628593,
+ "learning_rate": 0.0006,
+ "loss": 4.252397537231445,
+ "step": 1856
+ },
+ {
+ "epoch": 25.79650655021834,
+ "grad_norm": 0.017274610698223114,
+ "learning_rate": 0.0006,
+ "loss": 4.342309951782227,
+ "step": 1857
+ },
+ {
+ "epoch": 25.810480349344978,
+ "grad_norm": 0.017879825085401535,
+ "learning_rate": 0.0006,
+ "loss": 4.330987453460693,
+ "step": 1858
+ },
+ {
+ "epoch": 25.824454148471617,
+ "grad_norm": 0.01736099272966385,
+ "learning_rate": 0.0006,
+ "loss": 4.307255744934082,
+ "step": 1859
+ },
+ {
+ "epoch": 25.83842794759825,
+ "grad_norm": 0.016575824469327927,
+ "learning_rate": 0.0006,
+ "loss": 4.262016773223877,
+ "step": 1860
+ },
+ {
+ "epoch": 25.85240174672489,
+ "grad_norm": 0.015181140042841434,
+ "learning_rate": 0.0006,
+ "loss": 4.291863441467285,
+ "step": 1861
+ },
+ {
+ "epoch": 25.86637554585153,
+ "grad_norm": 0.015965687111020088,
+ "learning_rate": 0.0006,
+ "loss": 4.424836158752441,
+ "step": 1862
+ },
+ {
+ "epoch": 25.880349344978168,
+ "grad_norm": 0.01687219925224781,
+ "learning_rate": 0.0006,
+ "loss": 4.343531131744385,
+ "step": 1863
+ },
+ {
+ "epoch": 25.894323144104803,
+ "grad_norm": 0.015115504153072834,
+ "learning_rate": 0.0006,
+ "loss": 4.357808589935303,
+ "step": 1864
+ },
+ {
+ "epoch": 25.90829694323144,
+ "grad_norm": 0.014150998555123806,
+ "learning_rate": 0.0006,
+ "loss": 4.324550151824951,
+ "step": 1865
+ },
+ {
+ "epoch": 25.92227074235808,
+ "grad_norm": 0.014989510178565979,
+ "learning_rate": 0.0006,
+ "loss": 4.380428314208984,
+ "step": 1866
+ },
+ {
+ "epoch": 25.936244541484715,
+ "grad_norm": 0.01527960691601038,
+ "learning_rate": 0.0006,
+ "loss": 4.399374961853027,
+ "step": 1867
+ },
+ {
+ "epoch": 25.950218340611354,
+ "grad_norm": 0.015434633940458298,
+ "learning_rate": 0.0006,
+ "loss": 4.260984420776367,
+ "step": 1868
+ },
+ {
+ "epoch": 25.964192139737992,
+ "grad_norm": 0.015370228327810764,
+ "learning_rate": 0.0006,
+ "loss": 4.3340253829956055,
+ "step": 1869
+ },
+ {
+ "epoch": 25.978165938864628,
+ "grad_norm": 0.015348542481660843,
+ "learning_rate": 0.0006,
+ "loss": 4.369132995605469,
+ "step": 1870
+ },
+ {
+ "epoch": 25.992139737991266,
+ "grad_norm": 0.015354936942458153,
+ "learning_rate": 0.0006,
+ "loss": 4.381937026977539,
+ "step": 1871
+ },
+ {
+ "epoch": 26.0,
+ "grad_norm": 0.01796272210776806,
+ "learning_rate": 0.0006,
+ "loss": 4.286768436431885,
+ "step": 1872
+ },
+ {
+ "epoch": 26.0,
+ "eval_loss": 4.576984405517578,
+ "eval_runtime": 57.2091,
+ "eval_samples_per_second": 42.685,
+ "eval_steps_per_second": 1.346,
+ "step": 1872
+ },
+ {
+ "epoch": 26.01397379912664,
+ "grad_norm": 0.01748845726251602,
+ "learning_rate": 0.0006,
+ "loss": 4.337504863739014,
+ "step": 1873
+ },
+ {
+ "epoch": 26.027947598253274,
+ "grad_norm": 0.019244296476244926,
+ "learning_rate": 0.0006,
+ "loss": 4.264280796051025,
+ "step": 1874
+ },
+ {
+ "epoch": 26.041921397379912,
+ "grad_norm": 0.0215692650526762,
+ "learning_rate": 0.0006,
+ "loss": 4.350830078125,
+ "step": 1875
+ },
+ {
+ "epoch": 26.05589519650655,
+ "grad_norm": 0.022189252078533173,
+ "learning_rate": 0.0006,
+ "loss": 4.299098968505859,
+ "step": 1876
+ },
+ {
+ "epoch": 26.069868995633186,
+ "grad_norm": 0.022384043782949448,
+ "learning_rate": 0.0006,
+ "loss": 4.1185173988342285,
+ "step": 1877
+ },
+ {
+ "epoch": 26.083842794759825,
+ "grad_norm": 0.02076014317572117,
+ "learning_rate": 0.0006,
+ "loss": 4.297612190246582,
+ "step": 1878
+ },
+ {
+ "epoch": 26.097816593886463,
+ "grad_norm": 0.020843051373958588,
+ "learning_rate": 0.0006,
+ "loss": 4.294098854064941,
+ "step": 1879
+ },
+ {
+ "epoch": 26.111790393013102,
+ "grad_norm": 0.020654456689953804,
+ "learning_rate": 0.0006,
+ "loss": 4.267856597900391,
+ "step": 1880
+ },
+ {
+ "epoch": 26.125764192139737,
+ "grad_norm": 0.01860472559928894,
+ "learning_rate": 0.0006,
+ "loss": 4.284232139587402,
+ "step": 1881
+ },
+ {
+ "epoch": 26.139737991266376,
+ "grad_norm": 0.018719421699643135,
+ "learning_rate": 0.0006,
+ "loss": 4.277214527130127,
+ "step": 1882
+ },
+ {
+ "epoch": 26.153711790393015,
+ "grad_norm": 0.020903829485177994,
+ "learning_rate": 0.0006,
+ "loss": 4.254947662353516,
+ "step": 1883
+ },
+ {
+ "epoch": 26.16768558951965,
+ "grad_norm": 0.02020149491727352,
+ "learning_rate": 0.0006,
+ "loss": 4.2526936531066895,
+ "step": 1884
+ },
+ {
+ "epoch": 26.18165938864629,
+ "grad_norm": 0.0180392786860466,
+ "learning_rate": 0.0006,
+ "loss": 4.309922695159912,
+ "step": 1885
+ },
+ {
+ "epoch": 26.195633187772927,
+ "grad_norm": 0.018163125962018967,
+ "learning_rate": 0.0006,
+ "loss": 4.13860559463501,
+ "step": 1886
+ },
+ {
+ "epoch": 26.209606986899562,
+ "grad_norm": 0.01853291131556034,
+ "learning_rate": 0.0006,
+ "loss": 4.190926551818848,
+ "step": 1887
+ },
+ {
+ "epoch": 26.2235807860262,
+ "grad_norm": 0.01757677271962166,
+ "learning_rate": 0.0006,
+ "loss": 4.163999557495117,
+ "step": 1888
+ },
+ {
+ "epoch": 26.23755458515284,
+ "grad_norm": 0.01767992228269577,
+ "learning_rate": 0.0006,
+ "loss": 4.184430122375488,
+ "step": 1889
+ },
+ {
+ "epoch": 26.251528384279474,
+ "grad_norm": 0.018139973282814026,
+ "learning_rate": 0.0006,
+ "loss": 4.300766944885254,
+ "step": 1890
+ },
+ {
+ "epoch": 26.265502183406113,
+ "grad_norm": 0.020184461027383804,
+ "learning_rate": 0.0006,
+ "loss": 4.232537269592285,
+ "step": 1891
+ },
+ {
+ "epoch": 26.27947598253275,
+ "grad_norm": 0.02028856985270977,
+ "learning_rate": 0.0006,
+ "loss": 4.324883460998535,
+ "step": 1892
+ },
+ {
+ "epoch": 26.29344978165939,
+ "grad_norm": 0.017120616510510445,
+ "learning_rate": 0.0006,
+ "loss": 4.23493766784668,
+ "step": 1893
+ },
+ {
+ "epoch": 26.307423580786025,
+ "grad_norm": 0.016202978789806366,
+ "learning_rate": 0.0006,
+ "loss": 4.197602272033691,
+ "step": 1894
+ },
+ {
+ "epoch": 26.321397379912664,
+ "grad_norm": 0.017231421545147896,
+ "learning_rate": 0.0006,
+ "loss": 4.302001953125,
+ "step": 1895
+ },
+ {
+ "epoch": 26.335371179039303,
+ "grad_norm": 0.01725655607879162,
+ "learning_rate": 0.0006,
+ "loss": 4.356800079345703,
+ "step": 1896
+ },
+ {
+ "epoch": 26.349344978165938,
+ "grad_norm": 0.016657443717122078,
+ "learning_rate": 0.0006,
+ "loss": 4.266307830810547,
+ "step": 1897
+ },
+ {
+ "epoch": 26.363318777292577,
+ "grad_norm": 0.01616556942462921,
+ "learning_rate": 0.0006,
+ "loss": 4.307868003845215,
+ "step": 1898
+ },
+ {
+ "epoch": 26.377292576419215,
+ "grad_norm": 0.016621306538581848,
+ "learning_rate": 0.0006,
+ "loss": 4.172905921936035,
+ "step": 1899
+ },
+ {
+ "epoch": 26.39126637554585,
+ "grad_norm": 0.016243206337094307,
+ "learning_rate": 0.0006,
+ "loss": 4.162896156311035,
+ "step": 1900
+ },
+ {
+ "epoch": 26.40524017467249,
+ "grad_norm": 0.016663808375597,
+ "learning_rate": 0.0006,
+ "loss": 4.258232116699219,
+ "step": 1901
+ },
+ {
+ "epoch": 26.419213973799128,
+ "grad_norm": 0.01677478663623333,
+ "learning_rate": 0.0006,
+ "loss": 4.336090087890625,
+ "step": 1902
+ },
+ {
+ "epoch": 26.433187772925763,
+ "grad_norm": 0.01641303487122059,
+ "learning_rate": 0.0006,
+ "loss": 4.191476821899414,
+ "step": 1903
+ },
+ {
+ "epoch": 26.4471615720524,
+ "grad_norm": 0.016700323671102524,
+ "learning_rate": 0.0006,
+ "loss": 4.299427509307861,
+ "step": 1904
+ },
+ {
+ "epoch": 26.46113537117904,
+ "grad_norm": 0.017595035955309868,
+ "learning_rate": 0.0006,
+ "loss": 4.312828063964844,
+ "step": 1905
+ },
+ {
+ "epoch": 26.475109170305675,
+ "grad_norm": 0.020300284028053284,
+ "learning_rate": 0.0006,
+ "loss": 4.28331184387207,
+ "step": 1906
+ },
+ {
+ "epoch": 26.489082969432314,
+ "grad_norm": 0.02198321744799614,
+ "learning_rate": 0.0006,
+ "loss": 4.3280839920043945,
+ "step": 1907
+ },
+ {
+ "epoch": 26.503056768558952,
+ "grad_norm": 0.02003200724720955,
+ "learning_rate": 0.0006,
+ "loss": 4.325623989105225,
+ "step": 1908
+ },
+ {
+ "epoch": 26.51703056768559,
+ "grad_norm": 0.01870287023484707,
+ "learning_rate": 0.0006,
+ "loss": 4.251203536987305,
+ "step": 1909
+ },
+ {
+ "epoch": 26.531004366812226,
+ "grad_norm": 0.01835448481142521,
+ "learning_rate": 0.0006,
+ "loss": 4.335144996643066,
+ "step": 1910
+ },
+ {
+ "epoch": 26.544978165938865,
+ "grad_norm": 0.02046327292919159,
+ "learning_rate": 0.0006,
+ "loss": 4.238023281097412,
+ "step": 1911
+ },
+ {
+ "epoch": 26.558951965065503,
+ "grad_norm": 0.019771311432123184,
+ "learning_rate": 0.0006,
+ "loss": 4.353594779968262,
+ "step": 1912
+ },
+ {
+ "epoch": 26.57292576419214,
+ "grad_norm": 0.018090544268488884,
+ "learning_rate": 0.0006,
+ "loss": 4.304343223571777,
+ "step": 1913
+ },
+ {
+ "epoch": 26.586899563318777,
+ "grad_norm": 0.017489777877926826,
+ "learning_rate": 0.0006,
+ "loss": 4.204375267028809,
+ "step": 1914
+ },
+ {
+ "epoch": 26.600873362445416,
+ "grad_norm": 0.016785891726613045,
+ "learning_rate": 0.0006,
+ "loss": 4.286669731140137,
+ "step": 1915
+ },
+ {
+ "epoch": 26.61484716157205,
+ "grad_norm": 0.016382789239287376,
+ "learning_rate": 0.0006,
+ "loss": 4.347662448883057,
+ "step": 1916
+ },
+ {
+ "epoch": 26.62882096069869,
+ "grad_norm": 0.016581876203417778,
+ "learning_rate": 0.0006,
+ "loss": 4.2716193199157715,
+ "step": 1917
+ },
+ {
+ "epoch": 26.64279475982533,
+ "grad_norm": 0.015627184882760048,
+ "learning_rate": 0.0006,
+ "loss": 4.311489105224609,
+ "step": 1918
+ },
+ {
+ "epoch": 26.656768558951963,
+ "grad_norm": 0.015908483415842056,
+ "learning_rate": 0.0006,
+ "loss": 4.262099742889404,
+ "step": 1919
+ },
+ {
+ "epoch": 26.670742358078602,
+ "grad_norm": 0.015576236881315708,
+ "learning_rate": 0.0006,
+ "loss": 4.275871276855469,
+ "step": 1920
+ },
+ {
+ "epoch": 26.68471615720524,
+ "grad_norm": 0.014621064998209476,
+ "learning_rate": 0.0006,
+ "loss": 4.184260845184326,
+ "step": 1921
+ },
+ {
+ "epoch": 26.69868995633188,
+ "grad_norm": 0.014797811396420002,
+ "learning_rate": 0.0006,
+ "loss": 4.3412628173828125,
+ "step": 1922
+ },
+ {
+ "epoch": 26.712663755458514,
+ "grad_norm": 0.015610570088028908,
+ "learning_rate": 0.0006,
+ "loss": 4.181094646453857,
+ "step": 1923
+ },
+ {
+ "epoch": 26.726637554585153,
+ "grad_norm": 0.016048265621066093,
+ "learning_rate": 0.0006,
+ "loss": 4.365290641784668,
+ "step": 1924
+ },
+ {
+ "epoch": 26.74061135371179,
+ "grad_norm": 0.016604071483016014,
+ "learning_rate": 0.0006,
+ "loss": 4.2628912925720215,
+ "step": 1925
+ },
+ {
+ "epoch": 26.754585152838427,
+ "grad_norm": 0.017044221982359886,
+ "learning_rate": 0.0006,
+ "loss": 4.378512382507324,
+ "step": 1926
+ },
+ {
+ "epoch": 26.768558951965066,
+ "grad_norm": 0.01746384985744953,
+ "learning_rate": 0.0006,
+ "loss": 4.3888773918151855,
+ "step": 1927
+ },
+ {
+ "epoch": 26.782532751091704,
+ "grad_norm": 0.01572711206972599,
+ "learning_rate": 0.0006,
+ "loss": 4.189935684204102,
+ "step": 1928
+ },
+ {
+ "epoch": 26.79650655021834,
+ "grad_norm": 0.014976629987359047,
+ "learning_rate": 0.0006,
+ "loss": 4.280525207519531,
+ "step": 1929
+ },
+ {
+ "epoch": 26.810480349344978,
+ "grad_norm": 0.014804664999246597,
+ "learning_rate": 0.0006,
+ "loss": 4.270630836486816,
+ "step": 1930
+ },
+ {
+ "epoch": 26.824454148471617,
+ "grad_norm": 0.014641920104622841,
+ "learning_rate": 0.0006,
+ "loss": 4.319505214691162,
+ "step": 1931
+ },
+ {
+ "epoch": 26.83842794759825,
+ "grad_norm": 0.014928505755960941,
+ "learning_rate": 0.0006,
+ "loss": 4.180147647857666,
+ "step": 1932
+ },
+ {
+ "epoch": 26.85240174672489,
+ "grad_norm": 0.016130365431308746,
+ "learning_rate": 0.0006,
+ "loss": 4.260660648345947,
+ "step": 1933
+ },
+ {
+ "epoch": 26.86637554585153,
+ "grad_norm": 0.015642493963241577,
+ "learning_rate": 0.0006,
+ "loss": 4.304064750671387,
+ "step": 1934
+ },
+ {
+ "epoch": 26.880349344978168,
+ "grad_norm": 0.015988217666745186,
+ "learning_rate": 0.0006,
+ "loss": 4.2548017501831055,
+ "step": 1935
+ },
+ {
+ "epoch": 26.894323144104803,
+ "grad_norm": 0.016501398757100105,
+ "learning_rate": 0.0006,
+ "loss": 4.260058879852295,
+ "step": 1936
+ },
+ {
+ "epoch": 26.90829694323144,
+ "grad_norm": 0.01645040698349476,
+ "learning_rate": 0.0006,
+ "loss": 4.483081340789795,
+ "step": 1937
+ },
+ {
+ "epoch": 26.92227074235808,
+ "grad_norm": 0.016140863299369812,
+ "learning_rate": 0.0006,
+ "loss": 4.1776838302612305,
+ "step": 1938
+ },
+ {
+ "epoch": 26.936244541484715,
+ "grad_norm": 0.016681019216775894,
+ "learning_rate": 0.0006,
+ "loss": 4.224747657775879,
+ "step": 1939
+ },
+ {
+ "epoch": 26.950218340611354,
+ "grad_norm": 0.01650378853082657,
+ "learning_rate": 0.0006,
+ "loss": 4.305703163146973,
+ "step": 1940
+ },
+ {
+ "epoch": 26.964192139737992,
+ "grad_norm": 0.017051683738827705,
+ "learning_rate": 0.0006,
+ "loss": 4.285835266113281,
+ "step": 1941
+ },
+ {
+ "epoch": 26.978165938864628,
+ "grad_norm": 0.016894327476620674,
+ "learning_rate": 0.0006,
+ "loss": 4.318901538848877,
+ "step": 1942
+ },
+ {
+ "epoch": 26.992139737991266,
+ "grad_norm": 0.017313100397586823,
+ "learning_rate": 0.0006,
+ "loss": 4.281434535980225,
+ "step": 1943
+ },
+ {
+ "epoch": 27.0,
+ "grad_norm": 0.018450884148478508,
+ "learning_rate": 0.0006,
+ "loss": 4.335104942321777,
+ "step": 1944
+ },
+ {
+ "epoch": 27.0,
+ "eval_loss": 4.640687465667725,
+ "eval_runtime": 56.9624,
+ "eval_samples_per_second": 42.87,
+ "eval_steps_per_second": 1.352,
+ "step": 1944
+ },
+ {
+ "epoch": 27.01397379912664,
+ "grad_norm": 0.017755698412656784,
+ "learning_rate": 0.0006,
+ "loss": 4.278824329376221,
+ "step": 1945
+ },
+ {
+ "epoch": 27.027947598253274,
+ "grad_norm": 0.01724310778081417,
+ "learning_rate": 0.0006,
+ "loss": 4.311049461364746,
+ "step": 1946
+ },
+ {
+ "epoch": 27.041921397379912,
+ "grad_norm": 0.01690123789012432,
+ "learning_rate": 0.0006,
+ "loss": 4.261994361877441,
+ "step": 1947
+ },
+ {
+ "epoch": 27.05589519650655,
+ "grad_norm": 0.018406696617603302,
+ "learning_rate": 0.0006,
+ "loss": 4.1925153732299805,
+ "step": 1948
+ },
+ {
+ "epoch": 27.069868995633186,
+ "grad_norm": 0.019945867359638214,
+ "learning_rate": 0.0006,
+ "loss": 4.108537197113037,
+ "step": 1949
+ },
+ {
+ "epoch": 27.083842794759825,
+ "grad_norm": 0.020316628739237785,
+ "learning_rate": 0.0006,
+ "loss": 4.209043025970459,
+ "step": 1950
+ },
+ {
+ "epoch": 27.097816593886463,
+ "grad_norm": 0.019690370187163353,
+ "learning_rate": 0.0006,
+ "loss": 4.195789337158203,
+ "step": 1951
+ },
+ {
+ "epoch": 27.111790393013102,
+ "grad_norm": 0.01785232499241829,
+ "learning_rate": 0.0006,
+ "loss": 4.21466064453125,
+ "step": 1952
+ },
+ {
+ "epoch": 27.125764192139737,
+ "grad_norm": 0.0164767038077116,
+ "learning_rate": 0.0006,
+ "loss": 4.270247459411621,
+ "step": 1953
+ },
+ {
+ "epoch": 27.139737991266376,
+ "grad_norm": 0.018008455634117126,
+ "learning_rate": 0.0006,
+ "loss": 4.201042175292969,
+ "step": 1954
+ },
+ {
+ "epoch": 27.153711790393015,
+ "grad_norm": 0.01807340793311596,
+ "learning_rate": 0.0006,
+ "loss": 4.25289249420166,
+ "step": 1955
+ },
+ {
+ "epoch": 27.16768558951965,
+ "grad_norm": 0.016246909275650978,
+ "learning_rate": 0.0006,
+ "loss": 4.116283416748047,
+ "step": 1956
+ },
+ {
+ "epoch": 27.18165938864629,
+ "grad_norm": 0.017545776441693306,
+ "learning_rate": 0.0006,
+ "loss": 4.316999435424805,
+ "step": 1957
+ },
+ {
+ "epoch": 27.195633187772927,
+ "grad_norm": 0.017915375530719757,
+ "learning_rate": 0.0006,
+ "loss": 4.253859519958496,
+ "step": 1958
+ },
+ {
+ "epoch": 27.209606986899562,
+ "grad_norm": 0.018112223595380783,
+ "learning_rate": 0.0006,
+ "loss": 4.2541022300720215,
+ "step": 1959
+ },
+ {
+ "epoch": 27.2235807860262,
+ "grad_norm": 0.01846139505505562,
+ "learning_rate": 0.0006,
+ "loss": 4.20438289642334,
+ "step": 1960
+ },
+ {
+ "epoch": 27.23755458515284,
+ "grad_norm": 0.01943897269666195,
+ "learning_rate": 0.0006,
+ "loss": 4.2799577713012695,
+ "step": 1961
+ },
+ {
+ "epoch": 27.251528384279474,
+ "grad_norm": 0.018253348767757416,
+ "learning_rate": 0.0006,
+ "loss": 4.1720499992370605,
+ "step": 1962
+ },
+ {
+ "epoch": 27.265502183406113,
+ "grad_norm": 0.018292246386408806,
+ "learning_rate": 0.0006,
+ "loss": 4.28231143951416,
+ "step": 1963
+ },
+ {
+ "epoch": 27.27947598253275,
+ "grad_norm": 0.01943155936896801,
+ "learning_rate": 0.0006,
+ "loss": 4.314976692199707,
+ "step": 1964
+ },
+ {
+ "epoch": 27.29344978165939,
+ "grad_norm": 0.019064994528889656,
+ "learning_rate": 0.0006,
+ "loss": 4.25328254699707,
+ "step": 1965
+ },
+ {
+ "epoch": 27.307423580786025,
+ "grad_norm": 0.017423273995518684,
+ "learning_rate": 0.0006,
+ "loss": 4.195992469787598,
+ "step": 1966
+ },
+ {
+ "epoch": 27.321397379912664,
+ "grad_norm": 0.01677924580872059,
+ "learning_rate": 0.0006,
+ "loss": 4.260043144226074,
+ "step": 1967
+ },
+ {
+ "epoch": 27.335371179039303,
+ "grad_norm": 0.01720673404633999,
+ "learning_rate": 0.0006,
+ "loss": 4.222780704498291,
+ "step": 1968
+ },
+ {
+ "epoch": 27.349344978165938,
+ "grad_norm": 0.017300743609666824,
+ "learning_rate": 0.0006,
+ "loss": 4.109524250030518,
+ "step": 1969
+ },
+ {
+ "epoch": 27.363318777292577,
+ "grad_norm": 0.016065871343016624,
+ "learning_rate": 0.0006,
+ "loss": 4.36108922958374,
+ "step": 1970
+ },
+ {
+ "epoch": 27.377292576419215,
+ "grad_norm": 0.016626616939902306,
+ "learning_rate": 0.0006,
+ "loss": 4.220630645751953,
+ "step": 1971
+ },
+ {
+ "epoch": 27.39126637554585,
+ "grad_norm": 0.016482602804899216,
+ "learning_rate": 0.0006,
+ "loss": 4.344797134399414,
+ "step": 1972
+ },
+ {
+ "epoch": 27.40524017467249,
+ "grad_norm": 0.01727953553199768,
+ "learning_rate": 0.0006,
+ "loss": 4.296645164489746,
+ "step": 1973
+ },
+ {
+ "epoch": 27.419213973799128,
+ "grad_norm": 0.01522731315344572,
+ "learning_rate": 0.0006,
+ "loss": 4.223751068115234,
+ "step": 1974
+ },
+ {
+ "epoch": 27.433187772925763,
+ "grad_norm": 0.01581776700913906,
+ "learning_rate": 0.0006,
+ "loss": 4.351379871368408,
+ "step": 1975
+ },
+ {
+ "epoch": 27.4471615720524,
+ "grad_norm": 0.016394605860114098,
+ "learning_rate": 0.0006,
+ "loss": 4.2412919998168945,
+ "step": 1976
+ },
+ {
+ "epoch": 27.46113537117904,
+ "grad_norm": 0.01751169003546238,
+ "learning_rate": 0.0006,
+ "loss": 4.257606506347656,
+ "step": 1977
+ },
+ {
+ "epoch": 27.475109170305675,
+ "grad_norm": 0.017452310770750046,
+ "learning_rate": 0.0006,
+ "loss": 4.380134582519531,
+ "step": 1978
+ },
+ {
+ "epoch": 27.489082969432314,
+ "grad_norm": 0.017741898074746132,
+ "learning_rate": 0.0006,
+ "loss": 4.279632568359375,
+ "step": 1979
+ },
+ {
+ "epoch": 27.503056768558952,
+ "grad_norm": 0.017680590972304344,
+ "learning_rate": 0.0006,
+ "loss": 4.273406028747559,
+ "step": 1980
+ },
+ {
+ "epoch": 27.51703056768559,
+ "grad_norm": 0.01864718459546566,
+ "learning_rate": 0.0006,
+ "loss": 4.278583526611328,
+ "step": 1981
+ },
+ {
+ "epoch": 27.531004366812226,
+ "grad_norm": 0.019558541476726532,
+ "learning_rate": 0.0006,
+ "loss": 4.211919784545898,
+ "step": 1982
+ },
+ {
+ "epoch": 27.544978165938865,
+ "grad_norm": 0.019467400386929512,
+ "learning_rate": 0.0006,
+ "loss": 4.276822566986084,
+ "step": 1983
+ },
+ {
+ "epoch": 27.558951965065503,
+ "grad_norm": 0.01756172813475132,
+ "learning_rate": 0.0006,
+ "loss": 4.165379047393799,
+ "step": 1984
+ },
+ {
+ "epoch": 27.57292576419214,
+ "grad_norm": 0.017458263784646988,
+ "learning_rate": 0.0006,
+ "loss": 4.2992401123046875,
+ "step": 1985
+ },
+ {
+ "epoch": 27.586899563318777,
+ "grad_norm": 0.018089665099978447,
+ "learning_rate": 0.0006,
+ "loss": 4.243378639221191,
+ "step": 1986
+ },
+ {
+ "epoch": 27.600873362445416,
+ "grad_norm": 0.019535524770617485,
+ "learning_rate": 0.0006,
+ "loss": 4.321477890014648,
+ "step": 1987
+ },
+ {
+ "epoch": 27.61484716157205,
+ "grad_norm": 0.01997970975935459,
+ "learning_rate": 0.0006,
+ "loss": 4.123082637786865,
+ "step": 1988
+ },
+ {
+ "epoch": 27.62882096069869,
+ "grad_norm": 0.01913067139685154,
+ "learning_rate": 0.0006,
+ "loss": 4.289896488189697,
+ "step": 1989
+ },
+ {
+ "epoch": 27.64279475982533,
+ "grad_norm": 0.0169806070625782,
+ "learning_rate": 0.0006,
+ "loss": 4.247004985809326,
+ "step": 1990
+ },
+ {
+ "epoch": 27.656768558951963,
+ "grad_norm": 0.017315007746219635,
+ "learning_rate": 0.0006,
+ "loss": 4.2095441818237305,
+ "step": 1991
+ },
+ {
+ "epoch": 27.670742358078602,
+ "grad_norm": 0.015098759904503822,
+ "learning_rate": 0.0006,
+ "loss": 4.145560264587402,
+ "step": 1992
+ },
+ {
+ "epoch": 27.68471615720524,
+ "grad_norm": 0.016240514814853668,
+ "learning_rate": 0.0006,
+ "loss": 4.254927635192871,
+ "step": 1993
+ },
+ {
+ "epoch": 27.69868995633188,
+ "grad_norm": 0.01621108129620552,
+ "learning_rate": 0.0006,
+ "loss": 4.219725608825684,
+ "step": 1994
+ },
+ {
+ "epoch": 27.712663755458514,
+ "grad_norm": 0.015191009268164635,
+ "learning_rate": 0.0006,
+ "loss": 4.194345474243164,
+ "step": 1995
+ },
+ {
+ "epoch": 27.726637554585153,
+ "grad_norm": 0.014387983828783035,
+ "learning_rate": 0.0006,
+ "loss": 4.211467742919922,
+ "step": 1996
+ },
+ {
+ "epoch": 27.74061135371179,
+ "grad_norm": 0.015460561029613018,
+ "learning_rate": 0.0006,
+ "loss": 4.358700752258301,
+ "step": 1997
+ },
+ {
+ "epoch": 27.754585152838427,
+ "grad_norm": 0.015100076794624329,
+ "learning_rate": 0.0006,
+ "loss": 4.171223163604736,
+ "step": 1998
+ },
+ {
+ "epoch": 27.768558951965066,
+ "grad_norm": 0.015607891604304314,
+ "learning_rate": 0.0006,
+ "loss": 4.217874526977539,
+ "step": 1999
+ },
+ {
+ "epoch": 27.782532751091704,
+ "grad_norm": 0.015650689601898193,
+ "learning_rate": 0.0006,
+ "loss": 4.280780792236328,
+ "step": 2000
+ },
+ {
+ "epoch": 27.79650655021834,
+ "grad_norm": 0.01580994389951229,
+ "learning_rate": 0.0006,
+ "loss": 4.159212112426758,
+ "step": 2001
+ },
+ {
+ "epoch": 27.810480349344978,
+ "grad_norm": 0.016607413068413734,
+ "learning_rate": 0.0006,
+ "loss": 4.238853454589844,
+ "step": 2002
+ },
+ {
+ "epoch": 27.824454148471617,
+ "grad_norm": 0.017383860424160957,
+ "learning_rate": 0.0006,
+ "loss": 4.233944892883301,
+ "step": 2003
+ },
+ {
+ "epoch": 27.83842794759825,
+ "grad_norm": 0.014966659247875214,
+ "learning_rate": 0.0006,
+ "loss": 4.167392730712891,
+ "step": 2004
+ },
+ {
+ "epoch": 27.85240174672489,
+ "grad_norm": 0.014589878730475903,
+ "learning_rate": 0.0006,
+ "loss": 4.2390899658203125,
+ "step": 2005
+ },
+ {
+ "epoch": 27.86637554585153,
+ "grad_norm": 0.017179450020194054,
+ "learning_rate": 0.0006,
+ "loss": 4.365785598754883,
+ "step": 2006
+ },
+ {
+ "epoch": 27.880349344978168,
+ "grad_norm": 0.01893901452422142,
+ "learning_rate": 0.0006,
+ "loss": 4.289430141448975,
+ "step": 2007
+ },
+ {
+ "epoch": 27.894323144104803,
+ "grad_norm": 0.018619712442159653,
+ "learning_rate": 0.0006,
+ "loss": 4.310665130615234,
+ "step": 2008
+ },
+ {
+ "epoch": 27.90829694323144,
+ "grad_norm": 0.019180385395884514,
+ "learning_rate": 0.0006,
+ "loss": 4.27454137802124,
+ "step": 2009
+ },
+ {
+ "epoch": 27.92227074235808,
+ "grad_norm": 0.017811523750424385,
+ "learning_rate": 0.0006,
+ "loss": 4.312562465667725,
+ "step": 2010
+ },
+ {
+ "epoch": 27.936244541484715,
+ "grad_norm": 0.016583573073148727,
+ "learning_rate": 0.0006,
+ "loss": 4.202253341674805,
+ "step": 2011
+ },
+ {
+ "epoch": 27.950218340611354,
+ "grad_norm": 0.01658615842461586,
+ "learning_rate": 0.0006,
+ "loss": 4.183433532714844,
+ "step": 2012
+ },
+ {
+ "epoch": 27.964192139737992,
+ "grad_norm": 0.015327401459217072,
+ "learning_rate": 0.0006,
+ "loss": 4.2125139236450195,
+ "step": 2013
+ },
+ {
+ "epoch": 27.978165938864628,
+ "grad_norm": 0.015004601329565048,
+ "learning_rate": 0.0006,
+ "loss": 4.257392883300781,
+ "step": 2014
+ },
+ {
+ "epoch": 27.992139737991266,
+ "grad_norm": 0.015259161591529846,
+ "learning_rate": 0.0006,
+ "loss": 4.241818428039551,
+ "step": 2015
+ },
+ {
+ "epoch": 28.0,
+ "grad_norm": 0.016808776184916496,
+ "learning_rate": 0.0006,
+ "loss": 4.146241188049316,
+ "step": 2016
+ },
+ {
+ "epoch": 28.0,
+ "eval_loss": 4.663090705871582,
+ "eval_runtime": 56.4559,
+ "eval_samples_per_second": 43.255,
+ "eval_steps_per_second": 1.364,
+ "step": 2016
+ },
+ {
+ "epoch": 28.01397379912664,
+ "grad_norm": 0.015784382820129395,
+ "learning_rate": 0.0006,
+ "loss": 4.044191360473633,
+ "step": 2017
+ },
+ {
+ "epoch": 28.027947598253274,
+ "grad_norm": 0.015636710450053215,
+ "learning_rate": 0.0006,
+ "loss": 4.128754615783691,
+ "step": 2018
+ },
+ {
+ "epoch": 28.041921397379912,
+ "grad_norm": 0.015364975668489933,
+ "learning_rate": 0.0006,
+ "loss": 4.191835403442383,
+ "step": 2019
+ },
+ {
+ "epoch": 28.05589519650655,
+ "grad_norm": 0.015015700832009315,
+ "learning_rate": 0.0006,
+ "loss": 4.120022773742676,
+ "step": 2020
+ },
+ {
+ "epoch": 28.069868995633186,
+ "grad_norm": 0.016666799783706665,
+ "learning_rate": 0.0006,
+ "loss": 4.33906364440918,
+ "step": 2021
+ },
+ {
+ "epoch": 28.083842794759825,
+ "grad_norm": 0.01780104823410511,
+ "learning_rate": 0.0006,
+ "loss": 4.279458999633789,
+ "step": 2022
+ },
+ {
+ "epoch": 28.097816593886463,
+ "grad_norm": 0.01870882883667946,
+ "learning_rate": 0.0006,
+ "loss": 4.34728479385376,
+ "step": 2023
+ },
+ {
+ "epoch": 28.111790393013102,
+ "grad_norm": 0.018406381830573082,
+ "learning_rate": 0.0006,
+ "loss": 4.211311340332031,
+ "step": 2024
+ },
+ {
+ "epoch": 28.125764192139737,
+ "grad_norm": 0.017833448946475983,
+ "learning_rate": 0.0006,
+ "loss": 4.223254203796387,
+ "step": 2025
+ },
+ {
+ "epoch": 28.139737991266376,
+ "grad_norm": 0.017552688717842102,
+ "learning_rate": 0.0006,
+ "loss": 4.208823204040527,
+ "step": 2026
+ },
+ {
+ "epoch": 28.153711790393015,
+ "grad_norm": 0.018500229343771935,
+ "learning_rate": 0.0006,
+ "loss": 4.198025226593018,
+ "step": 2027
+ },
+ {
+ "epoch": 28.16768558951965,
+ "grad_norm": 0.01782156340777874,
+ "learning_rate": 0.0006,
+ "loss": 4.179978370666504,
+ "step": 2028
+ },
+ {
+ "epoch": 28.18165938864629,
+ "grad_norm": 0.019580967724323273,
+ "learning_rate": 0.0006,
+ "loss": 4.291114807128906,
+ "step": 2029
+ },
+ {
+ "epoch": 28.195633187772927,
+ "grad_norm": 0.020804036408662796,
+ "learning_rate": 0.0006,
+ "loss": 4.220607280731201,
+ "step": 2030
+ },
+ {
+ "epoch": 28.209606986899562,
+ "grad_norm": 0.019595693796873093,
+ "learning_rate": 0.0006,
+ "loss": 4.2905731201171875,
+ "step": 2031
+ },
+ {
+ "epoch": 28.2235807860262,
+ "grad_norm": 0.019247086718678474,
+ "learning_rate": 0.0006,
+ "loss": 4.2467360496521,
+ "step": 2032
+ },
+ {
+ "epoch": 28.23755458515284,
+ "grad_norm": 0.01958036608994007,
+ "learning_rate": 0.0006,
+ "loss": 4.286137580871582,
+ "step": 2033
+ },
+ {
+ "epoch": 28.251528384279474,
+ "grad_norm": 0.021354753524065018,
+ "learning_rate": 0.0006,
+ "loss": 4.234755039215088,
+ "step": 2034
+ },
+ {
+ "epoch": 28.265502183406113,
+ "grad_norm": 0.02206375077366829,
+ "learning_rate": 0.0006,
+ "loss": 4.149641036987305,
+ "step": 2035
+ },
+ {
+ "epoch": 28.27947598253275,
+ "grad_norm": 0.02201310358941555,
+ "learning_rate": 0.0006,
+ "loss": 4.1928629875183105,
+ "step": 2036
+ },
+ {
+ "epoch": 28.29344978165939,
+ "grad_norm": 0.01961950585246086,
+ "learning_rate": 0.0006,
+ "loss": 4.211404323577881,
+ "step": 2037
+ },
+ {
+ "epoch": 28.307423580786025,
+ "grad_norm": 0.019211795181035995,
+ "learning_rate": 0.0006,
+ "loss": 4.200976848602295,
+ "step": 2038
+ },
+ {
+ "epoch": 28.321397379912664,
+ "grad_norm": 0.01779370754957199,
+ "learning_rate": 0.0006,
+ "loss": 4.249148845672607,
+ "step": 2039
+ },
+ {
+ "epoch": 28.335371179039303,
+ "grad_norm": 0.018635908141732216,
+ "learning_rate": 0.0006,
+ "loss": 4.11806058883667,
+ "step": 2040
+ },
+ {
+ "epoch": 28.349344978165938,
+ "grad_norm": 0.019792694598436356,
+ "learning_rate": 0.0006,
+ "loss": 4.335224151611328,
+ "step": 2041
+ },
+ {
+ "epoch": 28.363318777292577,
+ "grad_norm": 0.021731717512011528,
+ "learning_rate": 0.0006,
+ "loss": 4.302778244018555,
+ "step": 2042
+ },
+ {
+ "epoch": 28.377292576419215,
+ "grad_norm": 0.020823447033762932,
+ "learning_rate": 0.0006,
+ "loss": 4.260610580444336,
+ "step": 2043
+ },
+ {
+ "epoch": 28.39126637554585,
+ "grad_norm": 0.019385412335395813,
+ "learning_rate": 0.0006,
+ "loss": 4.2247724533081055,
+ "step": 2044
+ },
+ {
+ "epoch": 28.40524017467249,
+ "grad_norm": 0.019539108499884605,
+ "learning_rate": 0.0006,
+ "loss": 4.236655235290527,
+ "step": 2045
+ },
+ {
+ "epoch": 28.419213973799128,
+ "grad_norm": 0.01942690648138523,
+ "learning_rate": 0.0006,
+ "loss": 4.251850605010986,
+ "step": 2046
+ },
+ {
+ "epoch": 28.433187772925763,
+ "grad_norm": 0.019667886197566986,
+ "learning_rate": 0.0006,
+ "loss": 4.222312927246094,
+ "step": 2047
+ },
+ {
+ "epoch": 28.4471615720524,
+ "grad_norm": 0.02029012329876423,
+ "learning_rate": 0.0006,
+ "loss": 4.246252059936523,
+ "step": 2048
+ },
+ {
+ "epoch": 28.46113537117904,
+ "grad_norm": 0.01784469373524189,
+ "learning_rate": 0.0006,
+ "loss": 4.192128658294678,
+ "step": 2049
+ },
+ {
+ "epoch": 28.475109170305675,
+ "grad_norm": 0.0160287544131279,
+ "learning_rate": 0.0006,
+ "loss": 4.208607196807861,
+ "step": 2050
+ },
+ {
+ "epoch": 28.489082969432314,
+ "grad_norm": 0.015348346903920174,
+ "learning_rate": 0.0006,
+ "loss": 4.236598014831543,
+ "step": 2051
+ },
+ {
+ "epoch": 28.503056768558952,
+ "grad_norm": 0.015198206529021263,
+ "learning_rate": 0.0006,
+ "loss": 4.235552787780762,
+ "step": 2052
+ },
+ {
+ "epoch": 28.51703056768559,
+ "grad_norm": 0.01523754745721817,
+ "learning_rate": 0.0006,
+ "loss": 4.1941728591918945,
+ "step": 2053
+ },
+ {
+ "epoch": 28.531004366812226,
+ "grad_norm": 0.015559614636003971,
+ "learning_rate": 0.0006,
+ "loss": 4.294757843017578,
+ "step": 2054
+ },
+ {
+ "epoch": 28.544978165938865,
+ "grad_norm": 0.01568358950316906,
+ "learning_rate": 0.0006,
+ "loss": 4.213065147399902,
+ "step": 2055
+ },
+ {
+ "epoch": 28.558951965065503,
+ "grad_norm": 0.015336516313254833,
+ "learning_rate": 0.0006,
+ "loss": 4.240293979644775,
+ "step": 2056
+ },
+ {
+ "epoch": 28.57292576419214,
+ "grad_norm": 0.015041496604681015,
+ "learning_rate": 0.0006,
+ "loss": 4.273397445678711,
+ "step": 2057
+ },
+ {
+ "epoch": 28.586899563318777,
+ "grad_norm": 0.01610128954052925,
+ "learning_rate": 0.0006,
+ "loss": 4.125369071960449,
+ "step": 2058
+ },
+ {
+ "epoch": 28.600873362445416,
+ "grad_norm": 0.015974976122379303,
+ "learning_rate": 0.0006,
+ "loss": 4.133459091186523,
+ "step": 2059
+ },
+ {
+ "epoch": 28.61484716157205,
+ "grad_norm": 0.015936799347400665,
+ "learning_rate": 0.0006,
+ "loss": 4.306667327880859,
+ "step": 2060
+ },
+ {
+ "epoch": 28.62882096069869,
+ "grad_norm": 0.01795247197151184,
+ "learning_rate": 0.0006,
+ "loss": 4.205156326293945,
+ "step": 2061
+ },
+ {
+ "epoch": 28.64279475982533,
+ "grad_norm": 0.019639814272522926,
+ "learning_rate": 0.0006,
+ "loss": 4.277614593505859,
+ "step": 2062
+ },
+ {
+ "epoch": 28.656768558951963,
+ "grad_norm": 0.01722300611436367,
+ "learning_rate": 0.0006,
+ "loss": 4.191032886505127,
+ "step": 2063
+ },
+ {
+ "epoch": 28.670742358078602,
+ "grad_norm": 0.01612633652985096,
+ "learning_rate": 0.0006,
+ "loss": 4.20717191696167,
+ "step": 2064
+ },
+ {
+ "epoch": 28.68471615720524,
+ "grad_norm": 0.01603279449045658,
+ "learning_rate": 0.0006,
+ "loss": 4.228884220123291,
+ "step": 2065
+ },
+ {
+ "epoch": 28.69868995633188,
+ "grad_norm": 0.016405927017331123,
+ "learning_rate": 0.0006,
+ "loss": 4.201292991638184,
+ "step": 2066
+ },
+ {
+ "epoch": 28.712663755458514,
+ "grad_norm": 0.016644183546304703,
+ "learning_rate": 0.0006,
+ "loss": 4.239282131195068,
+ "step": 2067
+ },
+ {
+ "epoch": 28.726637554585153,
+ "grad_norm": 0.0175587497651577,
+ "learning_rate": 0.0006,
+ "loss": 4.289038181304932,
+ "step": 2068
+ },
+ {
+ "epoch": 28.74061135371179,
+ "grad_norm": 0.018223397433757782,
+ "learning_rate": 0.0006,
+ "loss": 4.2266998291015625,
+ "step": 2069
+ },
+ {
+ "epoch": 28.754585152838427,
+ "grad_norm": 0.018680868670344353,
+ "learning_rate": 0.0006,
+ "loss": 4.348849296569824,
+ "step": 2070
+ },
+ {
+ "epoch": 28.768558951965066,
+ "grad_norm": 0.016726728528738022,
+ "learning_rate": 0.0006,
+ "loss": 4.292080879211426,
+ "step": 2071
+ },
+ {
+ "epoch": 28.782532751091704,
+ "grad_norm": 0.017386795952916145,
+ "learning_rate": 0.0006,
+ "loss": 4.208870887756348,
+ "step": 2072
+ },
+ {
+ "epoch": 28.79650655021834,
+ "grad_norm": 0.01865178905427456,
+ "learning_rate": 0.0006,
+ "loss": 4.054150581359863,
+ "step": 2073
+ },
+ {
+ "epoch": 28.810480349344978,
+ "grad_norm": 0.01776815392076969,
+ "learning_rate": 0.0006,
+ "loss": 4.202404499053955,
+ "step": 2074
+ },
+ {
+ "epoch": 28.824454148471617,
+ "grad_norm": 0.01643497310578823,
+ "learning_rate": 0.0006,
+ "loss": 4.131855010986328,
+ "step": 2075
+ },
+ {
+ "epoch": 28.83842794759825,
+ "grad_norm": 0.01574273779988289,
+ "learning_rate": 0.0006,
+ "loss": 4.296788692474365,
+ "step": 2076
+ },
+ {
+ "epoch": 28.85240174672489,
+ "grad_norm": 0.015613908879458904,
+ "learning_rate": 0.0006,
+ "loss": 4.210949897766113,
+ "step": 2077
+ },
+ {
+ "epoch": 28.86637554585153,
+ "grad_norm": 0.015543212182819843,
+ "learning_rate": 0.0006,
+ "loss": 4.218136310577393,
+ "step": 2078
+ },
+ {
+ "epoch": 28.880349344978168,
+ "grad_norm": 0.014681251719594002,
+ "learning_rate": 0.0006,
+ "loss": 4.19401741027832,
+ "step": 2079
+ },
+ {
+ "epoch": 28.894323144104803,
+ "grad_norm": 0.014393828809261322,
+ "learning_rate": 0.0006,
+ "loss": 4.198468208312988,
+ "step": 2080
+ },
+ {
+ "epoch": 28.90829694323144,
+ "grad_norm": 0.01580170728266239,
+ "learning_rate": 0.0006,
+ "loss": 4.224969863891602,
+ "step": 2081
+ },
+ {
+ "epoch": 28.92227074235808,
+ "grad_norm": 0.01484165620058775,
+ "learning_rate": 0.0006,
+ "loss": 4.12236213684082,
+ "step": 2082
+ },
+ {
+ "epoch": 28.936244541484715,
+ "grad_norm": 0.01497623696923256,
+ "learning_rate": 0.0006,
+ "loss": 4.24141788482666,
+ "step": 2083
+ },
+ {
+ "epoch": 28.950218340611354,
+ "grad_norm": 0.015080046840012074,
+ "learning_rate": 0.0006,
+ "loss": 4.284151077270508,
+ "step": 2084
+ },
+ {
+ "epoch": 28.964192139737992,
+ "grad_norm": 0.016207212582230568,
+ "learning_rate": 0.0006,
+ "loss": 4.298693656921387,
+ "step": 2085
+ },
+ {
+ "epoch": 28.978165938864628,
+ "grad_norm": 0.015474417246878147,
+ "learning_rate": 0.0006,
+ "loss": 4.2412261962890625,
+ "step": 2086
+ },
+ {
+ "epoch": 28.992139737991266,
+ "grad_norm": 0.01515351701527834,
+ "learning_rate": 0.0006,
+ "loss": 4.3054423332214355,
+ "step": 2087
+ },
+ {
+ "epoch": 29.0,
+ "grad_norm": 0.016411345452070236,
+ "learning_rate": 0.0006,
+ "loss": 4.104830741882324,
+ "step": 2088
+ },
+ {
+ "epoch": 29.0,
+ "eval_loss": 4.661567687988281,
+ "eval_runtime": 57.1244,
+ "eval_samples_per_second": 42.749,
+ "eval_steps_per_second": 1.348,
+ "step": 2088
+ },
+ {
+ "epoch": 29.01397379912664,
+ "grad_norm": 0.01665342226624489,
+ "learning_rate": 0.0006,
+ "loss": 4.311391830444336,
+ "step": 2089
+ },
+ {
+ "epoch": 29.027947598253274,
+ "grad_norm": 0.018053214997053146,
+ "learning_rate": 0.0006,
+ "loss": 4.135597229003906,
+ "step": 2090
+ },
+ {
+ "epoch": 29.041921397379912,
+ "grad_norm": 0.018213748931884766,
+ "learning_rate": 0.0006,
+ "loss": 4.303299903869629,
+ "step": 2091
+ },
+ {
+ "epoch": 29.05589519650655,
+ "grad_norm": 0.018460242077708244,
+ "learning_rate": 0.0006,
+ "loss": 4.20850944519043,
+ "step": 2092
+ },
+ {
+ "epoch": 29.069868995633186,
+ "grad_norm": 0.018365809693932533,
+ "learning_rate": 0.0006,
+ "loss": 4.275847434997559,
+ "step": 2093
+ },
+ {
+ "epoch": 29.083842794759825,
+ "grad_norm": 0.018346186727285385,
+ "learning_rate": 0.0006,
+ "loss": 4.202091217041016,
+ "step": 2094
+ },
+ {
+ "epoch": 29.097816593886463,
+ "grad_norm": 0.017920587211847305,
+ "learning_rate": 0.0006,
+ "loss": 4.182092666625977,
+ "step": 2095
+ },
+ {
+ "epoch": 29.111790393013102,
+ "grad_norm": 0.01812003180384636,
+ "learning_rate": 0.0006,
+ "loss": 4.259721755981445,
+ "step": 2096
+ },
+ {
+ "epoch": 29.125764192139737,
+ "grad_norm": 0.01700986735522747,
+ "learning_rate": 0.0006,
+ "loss": 4.067573070526123,
+ "step": 2097
+ },
+ {
+ "epoch": 29.139737991266376,
+ "grad_norm": 0.018069760873913765,
+ "learning_rate": 0.0006,
+ "loss": 4.234400749206543,
+ "step": 2098
+ },
+ {
+ "epoch": 29.153711790393015,
+ "grad_norm": 0.01905428245663643,
+ "learning_rate": 0.0006,
+ "loss": 4.16609001159668,
+ "step": 2099
+ },
+ {
+ "epoch": 29.16768558951965,
+ "grad_norm": 0.018711242824792862,
+ "learning_rate": 0.0006,
+ "loss": 4.249088764190674,
+ "step": 2100
+ },
+ {
+ "epoch": 29.18165938864629,
+ "grad_norm": 0.018721306696534157,
+ "learning_rate": 0.0006,
+ "loss": 4.14540958404541,
+ "step": 2101
+ },
+ {
+ "epoch": 29.195633187772927,
+ "grad_norm": 0.020139874890446663,
+ "learning_rate": 0.0006,
+ "loss": 4.049248695373535,
+ "step": 2102
+ },
+ {
+ "epoch": 29.209606986899562,
+ "grad_norm": 0.020849574357271194,
+ "learning_rate": 0.0006,
+ "loss": 4.235224723815918,
+ "step": 2103
+ },
+ {
+ "epoch": 29.2235807860262,
+ "grad_norm": 0.018693549558520317,
+ "learning_rate": 0.0006,
+ "loss": 4.232961177825928,
+ "step": 2104
+ },
+ {
+ "epoch": 29.23755458515284,
+ "grad_norm": 0.017889857292175293,
+ "learning_rate": 0.0006,
+ "loss": 4.171422004699707,
+ "step": 2105
+ },
+ {
+ "epoch": 29.251528384279474,
+ "grad_norm": 0.016690224409103394,
+ "learning_rate": 0.0006,
+ "loss": 4.244717597961426,
+ "step": 2106
+ },
+ {
+ "epoch": 29.265502183406113,
+ "grad_norm": 0.01708616502583027,
+ "learning_rate": 0.0006,
+ "loss": 4.125068664550781,
+ "step": 2107
+ },
+ {
+ "epoch": 29.27947598253275,
+ "grad_norm": 0.017928237095475197,
+ "learning_rate": 0.0006,
+ "loss": 4.208292007446289,
+ "step": 2108
+ },
+ {
+ "epoch": 29.29344978165939,
+ "grad_norm": 0.018250394612550735,
+ "learning_rate": 0.0006,
+ "loss": 4.118630886077881,
+ "step": 2109
+ },
+ {
+ "epoch": 29.307423580786025,
+ "grad_norm": 0.020160246640443802,
+ "learning_rate": 0.0006,
+ "loss": 4.206025123596191,
+ "step": 2110
+ },
+ {
+ "epoch": 29.321397379912664,
+ "grad_norm": 0.020045453682541847,
+ "learning_rate": 0.0006,
+ "loss": 4.2758684158325195,
+ "step": 2111
+ },
+ {
+ "epoch": 29.335371179039303,
+ "grad_norm": 0.019556907936930656,
+ "learning_rate": 0.0006,
+ "loss": 4.188272953033447,
+ "step": 2112
+ },
+ {
+ "epoch": 29.349344978165938,
+ "grad_norm": 0.019185200333595276,
+ "learning_rate": 0.0006,
+ "loss": 4.17451286315918,
+ "step": 2113
+ },
+ {
+ "epoch": 29.363318777292577,
+ "grad_norm": 0.01816629432141781,
+ "learning_rate": 0.0006,
+ "loss": 4.165246963500977,
+ "step": 2114
+ },
+ {
+ "epoch": 29.377292576419215,
+ "grad_norm": 0.01865146867930889,
+ "learning_rate": 0.0006,
+ "loss": 4.148184776306152,
+ "step": 2115
+ },
+ {
+ "epoch": 29.39126637554585,
+ "grad_norm": 0.018338464200496674,
+ "learning_rate": 0.0006,
+ "loss": 4.288936614990234,
+ "step": 2116
+ },
+ {
+ "epoch": 29.40524017467249,
+ "grad_norm": 0.018986187875270844,
+ "learning_rate": 0.0006,
+ "loss": 4.183379650115967,
+ "step": 2117
+ },
+ {
+ "epoch": 29.419213973799128,
+ "grad_norm": 0.018409637734293938,
+ "learning_rate": 0.0006,
+ "loss": 4.057574272155762,
+ "step": 2118
+ },
+ {
+ "epoch": 29.433187772925763,
+ "grad_norm": 0.016664542257785797,
+ "learning_rate": 0.0006,
+ "loss": 4.18517541885376,
+ "step": 2119
+ },
+ {
+ "epoch": 29.4471615720524,
+ "grad_norm": 0.016486089676618576,
+ "learning_rate": 0.0006,
+ "loss": 4.179367542266846,
+ "step": 2120
+ },
+ {
+ "epoch": 29.46113537117904,
+ "grad_norm": 0.01787523552775383,
+ "learning_rate": 0.0006,
+ "loss": 4.184126377105713,
+ "step": 2121
+ },
+ {
+ "epoch": 29.475109170305675,
+ "grad_norm": 0.016574613749980927,
+ "learning_rate": 0.0006,
+ "loss": 4.290170192718506,
+ "step": 2122
+ },
+ {
+ "epoch": 29.489082969432314,
+ "grad_norm": 0.01643376611173153,
+ "learning_rate": 0.0006,
+ "loss": 4.052913665771484,
+ "step": 2123
+ },
+ {
+ "epoch": 29.503056768558952,
+ "grad_norm": 0.016503969207406044,
+ "learning_rate": 0.0006,
+ "loss": 4.335184097290039,
+ "step": 2124
+ },
+ {
+ "epoch": 29.51703056768559,
+ "grad_norm": 0.017596479505300522,
+ "learning_rate": 0.0006,
+ "loss": 4.165539741516113,
+ "step": 2125
+ },
+ {
+ "epoch": 29.531004366812226,
+ "grad_norm": 0.017468582838773727,
+ "learning_rate": 0.0006,
+ "loss": 4.315964698791504,
+ "step": 2126
+ },
+ {
+ "epoch": 29.544978165938865,
+ "grad_norm": 0.01724618673324585,
+ "learning_rate": 0.0006,
+ "loss": 4.243681907653809,
+ "step": 2127
+ },
+ {
+ "epoch": 29.558951965065503,
+ "grad_norm": 0.018384616822004318,
+ "learning_rate": 0.0006,
+ "loss": 4.112447261810303,
+ "step": 2128
+ },
+ {
+ "epoch": 29.57292576419214,
+ "grad_norm": 0.01824074238538742,
+ "learning_rate": 0.0006,
+ "loss": 4.236065864562988,
+ "step": 2129
+ },
+ {
+ "epoch": 29.586899563318777,
+ "grad_norm": 0.01625971868634224,
+ "learning_rate": 0.0006,
+ "loss": 4.18143892288208,
+ "step": 2130
+ },
+ {
+ "epoch": 29.600873362445416,
+ "grad_norm": 0.015537494793534279,
+ "learning_rate": 0.0006,
+ "loss": 4.225147247314453,
+ "step": 2131
+ },
+ {
+ "epoch": 29.61484716157205,
+ "grad_norm": 0.015818974003195763,
+ "learning_rate": 0.0006,
+ "loss": 4.3100175857543945,
+ "step": 2132
+ },
+ {
+ "epoch": 29.62882096069869,
+ "grad_norm": 0.015909671783447266,
+ "learning_rate": 0.0006,
+ "loss": 4.169775009155273,
+ "step": 2133
+ },
+ {
+ "epoch": 29.64279475982533,
+ "grad_norm": 0.01769658550620079,
+ "learning_rate": 0.0006,
+ "loss": 4.278904914855957,
+ "step": 2134
+ },
+ {
+ "epoch": 29.656768558951963,
+ "grad_norm": 0.015641704201698303,
+ "learning_rate": 0.0006,
+ "loss": 4.2426228523254395,
+ "step": 2135
+ },
+ {
+ "epoch": 29.670742358078602,
+ "grad_norm": 0.01565658301115036,
+ "learning_rate": 0.0006,
+ "loss": 4.232687950134277,
+ "step": 2136
+ },
+ {
+ "epoch": 29.68471615720524,
+ "grad_norm": 0.016485940665006638,
+ "learning_rate": 0.0006,
+ "loss": 4.209839820861816,
+ "step": 2137
+ },
+ {
+ "epoch": 29.69868995633188,
+ "grad_norm": 0.017622729763388634,
+ "learning_rate": 0.0006,
+ "loss": 4.196017265319824,
+ "step": 2138
+ },
+ {
+ "epoch": 29.712663755458514,
+ "grad_norm": 0.018320564180612564,
+ "learning_rate": 0.0006,
+ "loss": 4.225683689117432,
+ "step": 2139
+ },
+ {
+ "epoch": 29.726637554585153,
+ "grad_norm": 0.018140794709324837,
+ "learning_rate": 0.0006,
+ "loss": 4.311944961547852,
+ "step": 2140
+ },
+ {
+ "epoch": 29.74061135371179,
+ "grad_norm": 0.016227100044488907,
+ "learning_rate": 0.0006,
+ "loss": 4.230257511138916,
+ "step": 2141
+ },
+ {
+ "epoch": 29.754585152838427,
+ "grad_norm": 0.015759488567709923,
+ "learning_rate": 0.0006,
+ "loss": 4.119174003601074,
+ "step": 2142
+ },
+ {
+ "epoch": 29.768558951965066,
+ "grad_norm": 0.01757766678929329,
+ "learning_rate": 0.0006,
+ "loss": 4.165826797485352,
+ "step": 2143
+ },
+ {
+ "epoch": 29.782532751091704,
+ "grad_norm": 0.019661923870444298,
+ "learning_rate": 0.0006,
+ "loss": 4.392203330993652,
+ "step": 2144
+ },
+ {
+ "epoch": 29.79650655021834,
+ "grad_norm": 0.019478755071759224,
+ "learning_rate": 0.0006,
+ "loss": 4.348371982574463,
+ "step": 2145
+ },
+ {
+ "epoch": 29.810480349344978,
+ "grad_norm": 0.019149569794535637,
+ "learning_rate": 0.0006,
+ "loss": 4.34104061126709,
+ "step": 2146
+ },
+ {
+ "epoch": 29.824454148471617,
+ "grad_norm": 0.017707521095871925,
+ "learning_rate": 0.0006,
+ "loss": 4.228226661682129,
+ "step": 2147
+ },
+ {
+ "epoch": 29.83842794759825,
+ "grad_norm": 0.015615378506481647,
+ "learning_rate": 0.0006,
+ "loss": 4.175541877746582,
+ "step": 2148
+ },
+ {
+ "epoch": 29.85240174672489,
+ "grad_norm": 0.016111129894852638,
+ "learning_rate": 0.0006,
+ "loss": 4.267936706542969,
+ "step": 2149
+ },
+ {
+ "epoch": 29.86637554585153,
+ "grad_norm": 0.01528779324144125,
+ "learning_rate": 0.0006,
+ "loss": 4.186014175415039,
+ "step": 2150
+ },
+ {
+ "epoch": 29.880349344978168,
+ "grad_norm": 0.014925646595656872,
+ "learning_rate": 0.0006,
+ "loss": 4.174860000610352,
+ "step": 2151
+ },
+ {
+ "epoch": 29.894323144104803,
+ "grad_norm": 0.016917673870921135,
+ "learning_rate": 0.0006,
+ "loss": 4.33908748626709,
+ "step": 2152
+ },
+ {
+ "epoch": 29.90829694323144,
+ "grad_norm": 0.016007075086236,
+ "learning_rate": 0.0006,
+ "loss": 4.266141414642334,
+ "step": 2153
+ },
+ {
+ "epoch": 29.92227074235808,
+ "grad_norm": 0.016178956255316734,
+ "learning_rate": 0.0006,
+ "loss": 4.271215915679932,
+ "step": 2154
+ },
+ {
+ "epoch": 29.936244541484715,
+ "grad_norm": 0.016180871054530144,
+ "learning_rate": 0.0006,
+ "loss": 4.278801441192627,
+ "step": 2155
+ },
+ {
+ "epoch": 29.950218340611354,
+ "grad_norm": 0.015335503034293652,
+ "learning_rate": 0.0006,
+ "loss": 4.194978713989258,
+ "step": 2156
+ },
+ {
+ "epoch": 29.964192139737992,
+ "grad_norm": 0.016580764204263687,
+ "learning_rate": 0.0006,
+ "loss": 4.290809631347656,
+ "step": 2157
+ },
+ {
+ "epoch": 29.978165938864628,
+ "grad_norm": 0.01623128168284893,
+ "learning_rate": 0.0006,
+ "loss": 4.205451488494873,
+ "step": 2158
+ },
+ {
+ "epoch": 29.992139737991266,
+ "grad_norm": 0.014904248528182507,
+ "learning_rate": 0.0006,
+ "loss": 4.24924373626709,
+ "step": 2159
+ },
+ {
+ "epoch": 30.0,
+ "grad_norm": 0.017089299857616425,
+ "learning_rate": 0.0006,
+ "loss": 4.222072601318359,
+ "step": 2160
+ },
+ {
+ "epoch": 30.0,
+ "eval_loss": 4.585411071777344,
+ "eval_runtime": 57.1164,
+ "eval_samples_per_second": 42.755,
+ "eval_steps_per_second": 1.348,
+ "step": 2160
+ },
+ {
+ "epoch": 30.01397379912664,
+ "grad_norm": 0.017226383090019226,
+ "learning_rate": 0.0006,
+ "loss": 4.19467830657959,
+ "step": 2161
+ },
+ {
+ "epoch": 30.027947598253274,
+ "grad_norm": 0.01635066606104374,
+ "learning_rate": 0.0006,
+ "loss": 4.176875591278076,
+ "step": 2162
+ },
+ {
+ "epoch": 30.041921397379912,
+ "grad_norm": 0.01735362410545349,
+ "learning_rate": 0.0006,
+ "loss": 4.13667106628418,
+ "step": 2163
+ },
+ {
+ "epoch": 30.05589519650655,
+ "grad_norm": 0.017047109082341194,
+ "learning_rate": 0.0006,
+ "loss": 4.136329650878906,
+ "step": 2164
+ },
+ {
+ "epoch": 30.069868995633186,
+ "grad_norm": 0.018116053193807602,
+ "learning_rate": 0.0006,
+ "loss": 4.228706359863281,
+ "step": 2165
+ },
+ {
+ "epoch": 30.083842794759825,
+ "grad_norm": 0.01791389286518097,
+ "learning_rate": 0.0006,
+ "loss": 4.1597747802734375,
+ "step": 2166
+ },
+ {
+ "epoch": 30.097816593886463,
+ "grad_norm": 0.017793502658605576,
+ "learning_rate": 0.0006,
+ "loss": 4.0807976722717285,
+ "step": 2167
+ },
+ {
+ "epoch": 30.111790393013102,
+ "grad_norm": 0.01827503927052021,
+ "learning_rate": 0.0006,
+ "loss": 4.170646667480469,
+ "step": 2168
+ },
+ {
+ "epoch": 30.125764192139737,
+ "grad_norm": 0.017228564247488976,
+ "learning_rate": 0.0006,
+ "loss": 4.146934509277344,
+ "step": 2169
+ },
+ {
+ "epoch": 30.139737991266376,
+ "grad_norm": 0.01757437363266945,
+ "learning_rate": 0.0006,
+ "loss": 4.1859130859375,
+ "step": 2170
+ },
+ {
+ "epoch": 30.153711790393015,
+ "grad_norm": 0.01701093092560768,
+ "learning_rate": 0.0006,
+ "loss": 4.126981258392334,
+ "step": 2171
+ },
+ {
+ "epoch": 30.16768558951965,
+ "grad_norm": 0.015220489352941513,
+ "learning_rate": 0.0006,
+ "loss": 4.164018630981445,
+ "step": 2172
+ },
+ {
+ "epoch": 30.18165938864629,
+ "grad_norm": 0.016254238784313202,
+ "learning_rate": 0.0006,
+ "loss": 4.258275508880615,
+ "step": 2173
+ },
+ {
+ "epoch": 30.195633187772927,
+ "grad_norm": 0.016609614714980125,
+ "learning_rate": 0.0006,
+ "loss": 4.222233772277832,
+ "step": 2174
+ },
+ {
+ "epoch": 30.209606986899562,
+ "grad_norm": 0.016531318426132202,
+ "learning_rate": 0.0006,
+ "loss": 4.144489288330078,
+ "step": 2175
+ },
+ {
+ "epoch": 30.2235807860262,
+ "grad_norm": 0.01567668654024601,
+ "learning_rate": 0.0006,
+ "loss": 4.181916236877441,
+ "step": 2176
+ },
+ {
+ "epoch": 30.23755458515284,
+ "grad_norm": 0.017195681110024452,
+ "learning_rate": 0.0006,
+ "loss": 4.189693927764893,
+ "step": 2177
+ },
+ {
+ "epoch": 30.251528384279474,
+ "grad_norm": 0.016882948577404022,
+ "learning_rate": 0.0006,
+ "loss": 4.21604061126709,
+ "step": 2178
+ },
+ {
+ "epoch": 30.265502183406113,
+ "grad_norm": 0.016261622309684753,
+ "learning_rate": 0.0006,
+ "loss": 4.227639198303223,
+ "step": 2179
+ },
+ {
+ "epoch": 30.27947598253275,
+ "grad_norm": 0.016411812976002693,
+ "learning_rate": 0.0006,
+ "loss": 4.130362510681152,
+ "step": 2180
+ },
+ {
+ "epoch": 30.29344978165939,
+ "grad_norm": 0.01732019893825054,
+ "learning_rate": 0.0006,
+ "loss": 4.230443000793457,
+ "step": 2181
+ },
+ {
+ "epoch": 30.307423580786025,
+ "grad_norm": 0.01825905591249466,
+ "learning_rate": 0.0006,
+ "loss": 4.175817489624023,
+ "step": 2182
+ },
+ {
+ "epoch": 30.321397379912664,
+ "grad_norm": 0.01887713558971882,
+ "learning_rate": 0.0006,
+ "loss": 4.169020652770996,
+ "step": 2183
+ },
+ {
+ "epoch": 30.335371179039303,
+ "grad_norm": 0.019391661509871483,
+ "learning_rate": 0.0006,
+ "loss": 4.176318645477295,
+ "step": 2184
+ },
+ {
+ "epoch": 30.349344978165938,
+ "grad_norm": 0.020840002223849297,
+ "learning_rate": 0.0006,
+ "loss": 4.174228668212891,
+ "step": 2185
+ },
+ {
+ "epoch": 30.363318777292577,
+ "grad_norm": 0.020193178206682205,
+ "learning_rate": 0.0006,
+ "loss": 4.127919673919678,
+ "step": 2186
+ },
+ {
+ "epoch": 30.377292576419215,
+ "grad_norm": 0.018449561670422554,
+ "learning_rate": 0.0006,
+ "loss": 4.229487895965576,
+ "step": 2187
+ },
+ {
+ "epoch": 30.39126637554585,
+ "grad_norm": 0.01711983233690262,
+ "learning_rate": 0.0006,
+ "loss": 4.201651573181152,
+ "step": 2188
+ },
+ {
+ "epoch": 30.40524017467249,
+ "grad_norm": 0.017293395474553108,
+ "learning_rate": 0.0006,
+ "loss": 4.168704032897949,
+ "step": 2189
+ },
+ {
+ "epoch": 30.419213973799128,
+ "grad_norm": 0.018526358529925346,
+ "learning_rate": 0.0006,
+ "loss": 4.171864032745361,
+ "step": 2190
+ },
+ {
+ "epoch": 30.433187772925763,
+ "grad_norm": 0.019287291914224625,
+ "learning_rate": 0.0006,
+ "loss": 4.039534568786621,
+ "step": 2191
+ },
+ {
+ "epoch": 30.4471615720524,
+ "grad_norm": 0.016902444884181023,
+ "learning_rate": 0.0006,
+ "loss": 4.282787799835205,
+ "step": 2192
+ },
+ {
+ "epoch": 30.46113537117904,
+ "grad_norm": 0.01656750962138176,
+ "learning_rate": 0.0006,
+ "loss": 4.213098526000977,
+ "step": 2193
+ },
+ {
+ "epoch": 30.475109170305675,
+ "grad_norm": 0.016805065795779228,
+ "learning_rate": 0.0006,
+ "loss": 4.1583476066589355,
+ "step": 2194
+ },
+ {
+ "epoch": 30.489082969432314,
+ "grad_norm": 0.017166638746857643,
+ "learning_rate": 0.0006,
+ "loss": 4.188634395599365,
+ "step": 2195
+ },
+ {
+ "epoch": 30.503056768558952,
+ "grad_norm": 0.01592106744647026,
+ "learning_rate": 0.0006,
+ "loss": 4.234043121337891,
+ "step": 2196
+ },
+ {
+ "epoch": 30.51703056768559,
+ "grad_norm": 0.015270989388227463,
+ "learning_rate": 0.0006,
+ "loss": 4.301420211791992,
+ "step": 2197
+ },
+ {
+ "epoch": 30.531004366812226,
+ "grad_norm": 0.015194724313914776,
+ "learning_rate": 0.0006,
+ "loss": 4.18703031539917,
+ "step": 2198
+ },
+ {
+ "epoch": 30.544978165938865,
+ "grad_norm": 0.015587746165692806,
+ "learning_rate": 0.0006,
+ "loss": 4.138928413391113,
+ "step": 2199
+ },
+ {
+ "epoch": 30.558951965065503,
+ "grad_norm": 0.016029570251703262,
+ "learning_rate": 0.0006,
+ "loss": 4.290863990783691,
+ "step": 2200
+ },
+ {
+ "epoch": 30.57292576419214,
+ "grad_norm": 0.015651309862732887,
+ "learning_rate": 0.0006,
+ "loss": 4.272062301635742,
+ "step": 2201
+ },
+ {
+ "epoch": 30.586899563318777,
+ "grad_norm": 0.015992306172847748,
+ "learning_rate": 0.0006,
+ "loss": 4.2487406730651855,
+ "step": 2202
+ },
+ {
+ "epoch": 30.600873362445416,
+ "grad_norm": 0.014969157986342907,
+ "learning_rate": 0.0006,
+ "loss": 4.115347862243652,
+ "step": 2203
+ },
+ {
+ "epoch": 30.61484716157205,
+ "grad_norm": 0.014639027416706085,
+ "learning_rate": 0.0006,
+ "loss": 4.2207489013671875,
+ "step": 2204
+ },
+ {
+ "epoch": 30.62882096069869,
+ "grad_norm": 0.015059413388371468,
+ "learning_rate": 0.0006,
+ "loss": 4.274044990539551,
+ "step": 2205
+ },
+ {
+ "epoch": 30.64279475982533,
+ "grad_norm": 0.0166641678661108,
+ "learning_rate": 0.0006,
+ "loss": 4.191373348236084,
+ "step": 2206
+ },
+ {
+ "epoch": 30.656768558951963,
+ "grad_norm": 0.017806345596909523,
+ "learning_rate": 0.0006,
+ "loss": 4.139800548553467,
+ "step": 2207
+ },
+ {
+ "epoch": 30.670742358078602,
+ "grad_norm": 0.018712421879172325,
+ "learning_rate": 0.0006,
+ "loss": 4.035965442657471,
+ "step": 2208
+ },
+ {
+ "epoch": 30.68471615720524,
+ "grad_norm": 0.018457243219017982,
+ "learning_rate": 0.0006,
+ "loss": 4.2579240798950195,
+ "step": 2209
+ },
+ {
+ "epoch": 30.69868995633188,
+ "grad_norm": 0.01920999586582184,
+ "learning_rate": 0.0006,
+ "loss": 4.131112098693848,
+ "step": 2210
+ },
+ {
+ "epoch": 30.712663755458514,
+ "grad_norm": 0.01729944534599781,
+ "learning_rate": 0.0006,
+ "loss": 4.254059314727783,
+ "step": 2211
+ },
+ {
+ "epoch": 30.726637554585153,
+ "grad_norm": 0.016914231702685356,
+ "learning_rate": 0.0006,
+ "loss": 4.233428955078125,
+ "step": 2212
+ },
+ {
+ "epoch": 30.74061135371179,
+ "grad_norm": 0.01913098618388176,
+ "learning_rate": 0.0006,
+ "loss": 4.23525333404541,
+ "step": 2213
+ },
+ {
+ "epoch": 30.754585152838427,
+ "grad_norm": 0.019150281324982643,
+ "learning_rate": 0.0006,
+ "loss": 4.213695049285889,
+ "step": 2214
+ },
+ {
+ "epoch": 30.768558951965066,
+ "grad_norm": 0.018947452306747437,
+ "learning_rate": 0.0006,
+ "loss": 4.151697158813477,
+ "step": 2215
+ },
+ {
+ "epoch": 30.782532751091704,
+ "grad_norm": 0.016058508306741714,
+ "learning_rate": 0.0006,
+ "loss": 4.220280170440674,
+ "step": 2216
+ },
+ {
+ "epoch": 30.79650655021834,
+ "grad_norm": 0.017488619312644005,
+ "learning_rate": 0.0006,
+ "loss": 4.154428482055664,
+ "step": 2217
+ },
+ {
+ "epoch": 30.810480349344978,
+ "grad_norm": 0.017355090007185936,
+ "learning_rate": 0.0006,
+ "loss": 4.256086349487305,
+ "step": 2218
+ },
+ {
+ "epoch": 30.824454148471617,
+ "grad_norm": 0.016636032611131668,
+ "learning_rate": 0.0006,
+ "loss": 4.227564334869385,
+ "step": 2219
+ },
+ {
+ "epoch": 30.83842794759825,
+ "grad_norm": 0.016273891553282738,
+ "learning_rate": 0.0006,
+ "loss": 4.200466156005859,
+ "step": 2220
+ },
+ {
+ "epoch": 30.85240174672489,
+ "grad_norm": 0.016939911991357803,
+ "learning_rate": 0.0006,
+ "loss": 4.1571221351623535,
+ "step": 2221
+ },
+ {
+ "epoch": 30.86637554585153,
+ "grad_norm": 0.018332522362470627,
+ "learning_rate": 0.0006,
+ "loss": 4.228398323059082,
+ "step": 2222
+ },
+ {
+ "epoch": 30.880349344978168,
+ "grad_norm": 0.019508114084601402,
+ "learning_rate": 0.0006,
+ "loss": 4.222960472106934,
+ "step": 2223
+ },
+ {
+ "epoch": 30.894323144104803,
+ "grad_norm": 0.018440278246998787,
+ "learning_rate": 0.0006,
+ "loss": 4.261051177978516,
+ "step": 2224
+ },
+ {
+ "epoch": 30.90829694323144,
+ "grad_norm": 0.018113229423761368,
+ "learning_rate": 0.0006,
+ "loss": 4.109235763549805,
+ "step": 2225
+ },
+ {
+ "epoch": 30.92227074235808,
+ "grad_norm": 0.015392083674669266,
+ "learning_rate": 0.0006,
+ "loss": 4.21950101852417,
+ "step": 2226
+ },
+ {
+ "epoch": 30.936244541484715,
+ "grad_norm": 0.01633290946483612,
+ "learning_rate": 0.0006,
+ "loss": 4.176053047180176,
+ "step": 2227
+ },
+ {
+ "epoch": 30.950218340611354,
+ "grad_norm": 0.016236383467912674,
+ "learning_rate": 0.0006,
+ "loss": 4.314309120178223,
+ "step": 2228
+ },
+ {
+ "epoch": 30.964192139737992,
+ "grad_norm": 0.015284373424947262,
+ "learning_rate": 0.0006,
+ "loss": 4.303133487701416,
+ "step": 2229
+ },
+ {
+ "epoch": 30.978165938864628,
+ "grad_norm": 0.015982897952198982,
+ "learning_rate": 0.0006,
+ "loss": 4.25739860534668,
+ "step": 2230
+ },
+ {
+ "epoch": 30.992139737991266,
+ "grad_norm": 0.016682198271155357,
+ "learning_rate": 0.0006,
+ "loss": 4.292210578918457,
+ "step": 2231
+ },
+ {
+ "epoch": 31.0,
+ "grad_norm": 0.017521467059850693,
+ "learning_rate": 0.0006,
+ "loss": 4.170637130737305,
+ "step": 2232
+ },
+ {
+ "epoch": 31.0,
+ "eval_loss": 4.642786979675293,
+ "eval_runtime": 57.029,
+ "eval_samples_per_second": 42.82,
+ "eval_steps_per_second": 1.35,
+ "step": 2232
+ },
+ {
+ "epoch": 31.01397379912664,
+ "grad_norm": 0.016607852652668953,
+ "learning_rate": 0.0006,
+ "loss": 4.106224536895752,
+ "step": 2233
+ },
+ {
+ "epoch": 31.027947598253274,
+ "grad_norm": 0.018448904156684875,
+ "learning_rate": 0.0006,
+ "loss": 4.224459648132324,
+ "step": 2234
+ },
+ {
+ "epoch": 31.041921397379912,
+ "grad_norm": 0.020481228828430176,
+ "learning_rate": 0.0006,
+ "loss": 4.305548667907715,
+ "step": 2235
+ },
+ {
+ "epoch": 31.05589519650655,
+ "grad_norm": 0.01999707892537117,
+ "learning_rate": 0.0006,
+ "loss": 4.253663539886475,
+ "step": 2236
+ },
+ {
+ "epoch": 31.069868995633186,
+ "grad_norm": 0.020428303629159927,
+ "learning_rate": 0.0006,
+ "loss": 4.166954040527344,
+ "step": 2237
+ },
+ {
+ "epoch": 31.083842794759825,
+ "grad_norm": 0.022608213126659393,
+ "learning_rate": 0.0006,
+ "loss": 4.172819137573242,
+ "step": 2238
+ },
+ {
+ "epoch": 31.097816593886463,
+ "grad_norm": 0.024669520556926727,
+ "learning_rate": 0.0006,
+ "loss": 4.106586456298828,
+ "step": 2239
+ },
+ {
+ "epoch": 31.111790393013102,
+ "grad_norm": 0.02247351035475731,
+ "learning_rate": 0.0006,
+ "loss": 4.160271644592285,
+ "step": 2240
+ },
+ {
+ "epoch": 31.125764192139737,
+ "grad_norm": 0.0207810141146183,
+ "learning_rate": 0.0006,
+ "loss": 4.272491931915283,
+ "step": 2241
+ },
+ {
+ "epoch": 31.139737991266376,
+ "grad_norm": 0.021576499566435814,
+ "learning_rate": 0.0006,
+ "loss": 4.171474456787109,
+ "step": 2242
+ },
+ {
+ "epoch": 31.153711790393015,
+ "grad_norm": 0.02337106689810753,
+ "learning_rate": 0.0006,
+ "loss": 4.094712734222412,
+ "step": 2243
+ },
+ {
+ "epoch": 31.16768558951965,
+ "grad_norm": 0.02219875156879425,
+ "learning_rate": 0.0006,
+ "loss": 4.1800994873046875,
+ "step": 2244
+ },
+ {
+ "epoch": 31.18165938864629,
+ "grad_norm": 0.022680504247546196,
+ "learning_rate": 0.0006,
+ "loss": 4.2054362297058105,
+ "step": 2245
+ },
+ {
+ "epoch": 31.195633187772927,
+ "grad_norm": 0.0208734143525362,
+ "learning_rate": 0.0006,
+ "loss": 4.06068229675293,
+ "step": 2246
+ },
+ {
+ "epoch": 31.209606986899562,
+ "grad_norm": 0.01811066083610058,
+ "learning_rate": 0.0006,
+ "loss": 4.2465128898620605,
+ "step": 2247
+ },
+ {
+ "epoch": 31.2235807860262,
+ "grad_norm": 0.018571637570858,
+ "learning_rate": 0.0006,
+ "loss": 4.191955089569092,
+ "step": 2248
+ },
+ {
+ "epoch": 31.23755458515284,
+ "grad_norm": 0.018598807975649834,
+ "learning_rate": 0.0006,
+ "loss": 4.200872421264648,
+ "step": 2249
+ },
+ {
+ "epoch": 31.251528384279474,
+ "grad_norm": 0.017577625811100006,
+ "learning_rate": 0.0006,
+ "loss": 4.148700714111328,
+ "step": 2250
+ },
+ {
+ "epoch": 31.265502183406113,
+ "grad_norm": 0.01817980408668518,
+ "learning_rate": 0.0006,
+ "loss": 4.256978988647461,
+ "step": 2251
+ },
+ {
+ "epoch": 31.27947598253275,
+ "grad_norm": 0.017983486875891685,
+ "learning_rate": 0.0006,
+ "loss": 4.175005912780762,
+ "step": 2252
+ },
+ {
+ "epoch": 31.29344978165939,
+ "grad_norm": 0.016751961782574654,
+ "learning_rate": 0.0006,
+ "loss": 4.237518310546875,
+ "step": 2253
+ },
+ {
+ "epoch": 31.307423580786025,
+ "grad_norm": 0.016407202929258347,
+ "learning_rate": 0.0006,
+ "loss": 4.197269439697266,
+ "step": 2254
+ },
+ {
+ "epoch": 31.321397379912664,
+ "grad_norm": 0.01578645594418049,
+ "learning_rate": 0.0006,
+ "loss": 4.179751396179199,
+ "step": 2255
+ },
+ {
+ "epoch": 31.335371179039303,
+ "grad_norm": 0.016001226380467415,
+ "learning_rate": 0.0006,
+ "loss": 4.08821964263916,
+ "step": 2256
+ },
+ {
+ "epoch": 31.349344978165938,
+ "grad_norm": 0.015122811309993267,
+ "learning_rate": 0.0006,
+ "loss": 4.047828674316406,
+ "step": 2257
+ },
+ {
+ "epoch": 31.363318777292577,
+ "grad_norm": 0.015399965457618237,
+ "learning_rate": 0.0006,
+ "loss": 4.1506524085998535,
+ "step": 2258
+ },
+ {
+ "epoch": 31.377292576419215,
+ "grad_norm": 0.01534605398774147,
+ "learning_rate": 0.0006,
+ "loss": 4.27952766418457,
+ "step": 2259
+ },
+ {
+ "epoch": 31.39126637554585,
+ "grad_norm": 0.015461381524801254,
+ "learning_rate": 0.0006,
+ "loss": 4.27102518081665,
+ "step": 2260
+ },
+ {
+ "epoch": 31.40524017467249,
+ "grad_norm": 0.016769153997302055,
+ "learning_rate": 0.0006,
+ "loss": 4.191592216491699,
+ "step": 2261
+ },
+ {
+ "epoch": 31.419213973799128,
+ "grad_norm": 0.016557959839701653,
+ "learning_rate": 0.0006,
+ "loss": 4.159705638885498,
+ "step": 2262
+ },
+ {
+ "epoch": 31.433187772925763,
+ "grad_norm": 0.01659569889307022,
+ "learning_rate": 0.0006,
+ "loss": 4.2136664390563965,
+ "step": 2263
+ },
+ {
+ "epoch": 31.4471615720524,
+ "grad_norm": 0.017101937904953957,
+ "learning_rate": 0.0006,
+ "loss": 4.2716498374938965,
+ "step": 2264
+ },
+ {
+ "epoch": 31.46113537117904,
+ "grad_norm": 0.01722985878586769,
+ "learning_rate": 0.0006,
+ "loss": 4.300534248352051,
+ "step": 2265
+ },
+ {
+ "epoch": 31.475109170305675,
+ "grad_norm": 0.016800185665488243,
+ "learning_rate": 0.0006,
+ "loss": 4.113240718841553,
+ "step": 2266
+ },
+ {
+ "epoch": 31.489082969432314,
+ "grad_norm": 0.01603875495493412,
+ "learning_rate": 0.0006,
+ "loss": 4.157105445861816,
+ "step": 2267
+ },
+ {
+ "epoch": 31.503056768558952,
+ "grad_norm": 0.015903932973742485,
+ "learning_rate": 0.0006,
+ "loss": 4.233669757843018,
+ "step": 2268
+ },
+ {
+ "epoch": 31.51703056768559,
+ "grad_norm": 0.0160983856767416,
+ "learning_rate": 0.0006,
+ "loss": 4.147704124450684,
+ "step": 2269
+ },
+ {
+ "epoch": 31.531004366812226,
+ "grad_norm": 0.01604730263352394,
+ "learning_rate": 0.0006,
+ "loss": 4.1465911865234375,
+ "step": 2270
+ },
+ {
+ "epoch": 31.544978165938865,
+ "grad_norm": 0.016205785796046257,
+ "learning_rate": 0.0006,
+ "loss": 4.106704235076904,
+ "step": 2271
+ },
+ {
+ "epoch": 31.558951965065503,
+ "grad_norm": 0.017008036375045776,
+ "learning_rate": 0.0006,
+ "loss": 4.191000938415527,
+ "step": 2272
+ },
+ {
+ "epoch": 31.57292576419214,
+ "grad_norm": 0.01699635200202465,
+ "learning_rate": 0.0006,
+ "loss": 4.187042236328125,
+ "step": 2273
+ },
+ {
+ "epoch": 31.586899563318777,
+ "grad_norm": 0.01692967116832733,
+ "learning_rate": 0.0006,
+ "loss": 4.208805561065674,
+ "step": 2274
+ },
+ {
+ "epoch": 31.600873362445416,
+ "grad_norm": 0.017608756199479103,
+ "learning_rate": 0.0006,
+ "loss": 4.200512886047363,
+ "step": 2275
+ },
+ {
+ "epoch": 31.61484716157205,
+ "grad_norm": 0.016590990126132965,
+ "learning_rate": 0.0006,
+ "loss": 4.194705009460449,
+ "step": 2276
+ },
+ {
+ "epoch": 31.62882096069869,
+ "grad_norm": 0.0170235987752676,
+ "learning_rate": 0.0006,
+ "loss": 4.210326194763184,
+ "step": 2277
+ },
+ {
+ "epoch": 31.64279475982533,
+ "grad_norm": 0.01604801043868065,
+ "learning_rate": 0.0006,
+ "loss": 4.196024417877197,
+ "step": 2278
+ },
+ {
+ "epoch": 31.656768558951963,
+ "grad_norm": 0.015694202855229378,
+ "learning_rate": 0.0006,
+ "loss": 4.178236484527588,
+ "step": 2279
+ },
+ {
+ "epoch": 31.670742358078602,
+ "grad_norm": 0.015473959036171436,
+ "learning_rate": 0.0006,
+ "loss": 4.259339809417725,
+ "step": 2280
+ },
+ {
+ "epoch": 31.68471615720524,
+ "grad_norm": 0.014689362607896328,
+ "learning_rate": 0.0006,
+ "loss": 3.9702324867248535,
+ "step": 2281
+ },
+ {
+ "epoch": 31.69868995633188,
+ "grad_norm": 0.015245123766362667,
+ "learning_rate": 0.0006,
+ "loss": 4.197025775909424,
+ "step": 2282
+ },
+ {
+ "epoch": 31.712663755458514,
+ "grad_norm": 0.016148347407579422,
+ "learning_rate": 0.0006,
+ "loss": 4.231856346130371,
+ "step": 2283
+ },
+ {
+ "epoch": 31.726637554585153,
+ "grad_norm": 0.015169818885624409,
+ "learning_rate": 0.0006,
+ "loss": 4.230169296264648,
+ "step": 2284
+ },
+ {
+ "epoch": 31.74061135371179,
+ "grad_norm": 0.014843013137578964,
+ "learning_rate": 0.0006,
+ "loss": 4.195633411407471,
+ "step": 2285
+ },
+ {
+ "epoch": 31.754585152838427,
+ "grad_norm": 0.015229340642690659,
+ "learning_rate": 0.0006,
+ "loss": 4.231280326843262,
+ "step": 2286
+ },
+ {
+ "epoch": 31.768558951965066,
+ "grad_norm": 0.01551708485931158,
+ "learning_rate": 0.0006,
+ "loss": 4.17064094543457,
+ "step": 2287
+ },
+ {
+ "epoch": 31.782532751091704,
+ "grad_norm": 0.015863575041294098,
+ "learning_rate": 0.0006,
+ "loss": 4.1579484939575195,
+ "step": 2288
+ },
+ {
+ "epoch": 31.79650655021834,
+ "grad_norm": 0.015118198469281197,
+ "learning_rate": 0.0006,
+ "loss": 4.1626787185668945,
+ "step": 2289
+ },
+ {
+ "epoch": 31.810480349344978,
+ "grad_norm": 0.015220776200294495,
+ "learning_rate": 0.0006,
+ "loss": 4.135746955871582,
+ "step": 2290
+ },
+ {
+ "epoch": 31.824454148471617,
+ "grad_norm": 0.014529784210026264,
+ "learning_rate": 0.0006,
+ "loss": 4.164931297302246,
+ "step": 2291
+ },
+ {
+ "epoch": 31.83842794759825,
+ "grad_norm": 0.015533296391367912,
+ "learning_rate": 0.0006,
+ "loss": 4.112464904785156,
+ "step": 2292
+ },
+ {
+ "epoch": 31.85240174672489,
+ "grad_norm": 0.016562335193157196,
+ "learning_rate": 0.0006,
+ "loss": 4.164898872375488,
+ "step": 2293
+ },
+ {
+ "epoch": 31.86637554585153,
+ "grad_norm": 0.016258113086223602,
+ "learning_rate": 0.0006,
+ "loss": 4.1979265213012695,
+ "step": 2294
+ },
+ {
+ "epoch": 31.880349344978168,
+ "grad_norm": 0.01558110024780035,
+ "learning_rate": 0.0006,
+ "loss": 4.174675941467285,
+ "step": 2295
+ },
+ {
+ "epoch": 31.894323144104803,
+ "grad_norm": 0.016412515193223953,
+ "learning_rate": 0.0006,
+ "loss": 4.217909336090088,
+ "step": 2296
+ },
+ {
+ "epoch": 31.90829694323144,
+ "grad_norm": 0.015118442475795746,
+ "learning_rate": 0.0006,
+ "loss": 4.137143611907959,
+ "step": 2297
+ },
+ {
+ "epoch": 31.92227074235808,
+ "grad_norm": 0.01506161317229271,
+ "learning_rate": 0.0006,
+ "loss": 4.196122169494629,
+ "step": 2298
+ },
+ {
+ "epoch": 31.936244541484715,
+ "grad_norm": 0.017165271565318108,
+ "learning_rate": 0.0006,
+ "loss": 4.14834451675415,
+ "step": 2299
+ },
+ {
+ "epoch": 31.950218340611354,
+ "grad_norm": 0.01696440577507019,
+ "learning_rate": 0.0006,
+ "loss": 4.218868732452393,
+ "step": 2300
+ },
+ {
+ "epoch": 31.964192139737992,
+ "grad_norm": 0.017349394038319588,
+ "learning_rate": 0.0006,
+ "loss": 4.2754011154174805,
+ "step": 2301
+ },
+ {
+ "epoch": 31.978165938864628,
+ "grad_norm": 0.018387921154499054,
+ "learning_rate": 0.0006,
+ "loss": 4.282623291015625,
+ "step": 2302
+ },
+ {
+ "epoch": 31.992139737991266,
+ "grad_norm": 0.018837926909327507,
+ "learning_rate": 0.0006,
+ "loss": 4.185425281524658,
+ "step": 2303
+ },
+ {
+ "epoch": 32.0,
+ "grad_norm": 0.018998464569449425,
+ "learning_rate": 0.0006,
+ "loss": 4.151291847229004,
+ "step": 2304
+ },
+ {
+ "epoch": 32.0,
+ "eval_loss": 4.522614002227783,
+ "eval_runtime": 57.0106,
+ "eval_samples_per_second": 42.834,
+ "eval_steps_per_second": 1.351,
+ "step": 2304
+ },
+ {
+ "epoch": 32.01397379912664,
+ "grad_norm": 0.019036833196878433,
+ "learning_rate": 0.0006,
+ "loss": 4.183202266693115,
+ "step": 2305
+ },
+ {
+ "epoch": 32.02794759825328,
+ "grad_norm": 0.02207397110760212,
+ "learning_rate": 0.0006,
+ "loss": 4.17875862121582,
+ "step": 2306
+ },
+ {
+ "epoch": 32.041921397379916,
+ "grad_norm": 0.022008279338479042,
+ "learning_rate": 0.0006,
+ "loss": 4.08042049407959,
+ "step": 2307
+ },
+ {
+ "epoch": 32.05589519650655,
+ "grad_norm": 0.022407017648220062,
+ "learning_rate": 0.0006,
+ "loss": 4.112189769744873,
+ "step": 2308
+ },
+ {
+ "epoch": 32.069868995633186,
+ "grad_norm": 0.021599747240543365,
+ "learning_rate": 0.0006,
+ "loss": 4.126455307006836,
+ "step": 2309
+ },
+ {
+ "epoch": 32.083842794759825,
+ "grad_norm": 0.020340966060757637,
+ "learning_rate": 0.0006,
+ "loss": 4.080538272857666,
+ "step": 2310
+ },
+ {
+ "epoch": 32.09781659388646,
+ "grad_norm": 0.020956892520189285,
+ "learning_rate": 0.0006,
+ "loss": 4.171943664550781,
+ "step": 2311
+ },
+ {
+ "epoch": 32.1117903930131,
+ "grad_norm": 0.019966667518019676,
+ "learning_rate": 0.0006,
+ "loss": 4.144007682800293,
+ "step": 2312
+ },
+ {
+ "epoch": 32.12576419213974,
+ "grad_norm": 0.018256952986121178,
+ "learning_rate": 0.0006,
+ "loss": 4.208107948303223,
+ "step": 2313
+ },
+ {
+ "epoch": 32.13973799126637,
+ "grad_norm": 0.017289170995354652,
+ "learning_rate": 0.0006,
+ "loss": 4.182024955749512,
+ "step": 2314
+ },
+ {
+ "epoch": 32.15371179039301,
+ "grad_norm": 0.017656700685620308,
+ "learning_rate": 0.0006,
+ "loss": 4.068373680114746,
+ "step": 2315
+ },
+ {
+ "epoch": 32.16768558951965,
+ "grad_norm": 0.017903361469507217,
+ "learning_rate": 0.0006,
+ "loss": 4.2345428466796875,
+ "step": 2316
+ },
+ {
+ "epoch": 32.18165938864629,
+ "grad_norm": 0.01785109005868435,
+ "learning_rate": 0.0006,
+ "loss": 4.219776153564453,
+ "step": 2317
+ },
+ {
+ "epoch": 32.19563318777293,
+ "grad_norm": 0.01887257769703865,
+ "learning_rate": 0.0006,
+ "loss": 4.139206409454346,
+ "step": 2318
+ },
+ {
+ "epoch": 32.209606986899566,
+ "grad_norm": 0.01924685761332512,
+ "learning_rate": 0.0006,
+ "loss": 4.121052265167236,
+ "step": 2319
+ },
+ {
+ "epoch": 32.223580786026204,
+ "grad_norm": 0.020574072375893593,
+ "learning_rate": 0.0006,
+ "loss": 4.13262414932251,
+ "step": 2320
+ },
+ {
+ "epoch": 32.237554585152836,
+ "grad_norm": 0.023849064484238625,
+ "learning_rate": 0.0006,
+ "loss": 4.239119529724121,
+ "step": 2321
+ },
+ {
+ "epoch": 32.251528384279474,
+ "grad_norm": 0.026239361613988876,
+ "learning_rate": 0.0006,
+ "loss": 4.248993873596191,
+ "step": 2322
+ },
+ {
+ "epoch": 32.26550218340611,
+ "grad_norm": 0.023980243131518364,
+ "learning_rate": 0.0006,
+ "loss": 4.159852504730225,
+ "step": 2323
+ },
+ {
+ "epoch": 32.27947598253275,
+ "grad_norm": 0.020442405715584755,
+ "learning_rate": 0.0006,
+ "loss": 4.133546829223633,
+ "step": 2324
+ },
+ {
+ "epoch": 32.29344978165939,
+ "grad_norm": 0.020661186426877975,
+ "learning_rate": 0.0006,
+ "loss": 4.078243732452393,
+ "step": 2325
+ },
+ {
+ "epoch": 32.30742358078603,
+ "grad_norm": 0.02046387456357479,
+ "learning_rate": 0.0006,
+ "loss": 4.172012805938721,
+ "step": 2326
+ },
+ {
+ "epoch": 32.32139737991266,
+ "grad_norm": 0.02089597098529339,
+ "learning_rate": 0.0006,
+ "loss": 4.0604705810546875,
+ "step": 2327
+ },
+ {
+ "epoch": 32.3353711790393,
+ "grad_norm": 0.01875591278076172,
+ "learning_rate": 0.0006,
+ "loss": 4.125911235809326,
+ "step": 2328
+ },
+ {
+ "epoch": 32.34934497816594,
+ "grad_norm": 0.018606796860694885,
+ "learning_rate": 0.0006,
+ "loss": 4.097362518310547,
+ "step": 2329
+ },
+ {
+ "epoch": 32.36331877729258,
+ "grad_norm": 0.01766626164317131,
+ "learning_rate": 0.0006,
+ "loss": 4.118616580963135,
+ "step": 2330
+ },
+ {
+ "epoch": 32.377292576419215,
+ "grad_norm": 0.017692960798740387,
+ "learning_rate": 0.0006,
+ "loss": 4.1457672119140625,
+ "step": 2331
+ },
+ {
+ "epoch": 32.391266375545854,
+ "grad_norm": 0.017619704827666283,
+ "learning_rate": 0.0006,
+ "loss": 4.175961494445801,
+ "step": 2332
+ },
+ {
+ "epoch": 32.40524017467249,
+ "grad_norm": 0.01784413494169712,
+ "learning_rate": 0.0006,
+ "loss": 4.051291465759277,
+ "step": 2333
+ },
+ {
+ "epoch": 32.419213973799124,
+ "grad_norm": 0.018508171662688255,
+ "learning_rate": 0.0006,
+ "loss": 4.0638556480407715,
+ "step": 2334
+ },
+ {
+ "epoch": 32.43318777292576,
+ "grad_norm": 0.01731436885893345,
+ "learning_rate": 0.0006,
+ "loss": 4.1805419921875,
+ "step": 2335
+ },
+ {
+ "epoch": 32.4471615720524,
+ "grad_norm": 0.015886498615145683,
+ "learning_rate": 0.0006,
+ "loss": 4.151026725769043,
+ "step": 2336
+ },
+ {
+ "epoch": 32.46113537117904,
+ "grad_norm": 0.016583820804953575,
+ "learning_rate": 0.0006,
+ "loss": 4.203604698181152,
+ "step": 2337
+ },
+ {
+ "epoch": 32.47510917030568,
+ "grad_norm": 0.0176097322255373,
+ "learning_rate": 0.0006,
+ "loss": 4.135159015655518,
+ "step": 2338
+ },
+ {
+ "epoch": 32.48908296943232,
+ "grad_norm": 0.016657903790473938,
+ "learning_rate": 0.0006,
+ "loss": 4.126681804656982,
+ "step": 2339
+ },
+ {
+ "epoch": 32.50305676855895,
+ "grad_norm": 0.016866130754351616,
+ "learning_rate": 0.0006,
+ "loss": 4.162341117858887,
+ "step": 2340
+ },
+ {
+ "epoch": 32.51703056768559,
+ "grad_norm": 0.016286680474877357,
+ "learning_rate": 0.0006,
+ "loss": 4.261801242828369,
+ "step": 2341
+ },
+ {
+ "epoch": 32.531004366812226,
+ "grad_norm": 0.01481365505605936,
+ "learning_rate": 0.0006,
+ "loss": 4.110016822814941,
+ "step": 2342
+ },
+ {
+ "epoch": 32.544978165938865,
+ "grad_norm": 0.016605013981461525,
+ "learning_rate": 0.0006,
+ "loss": 4.230024814605713,
+ "step": 2343
+ },
+ {
+ "epoch": 32.5589519650655,
+ "grad_norm": 0.01645175740122795,
+ "learning_rate": 0.0006,
+ "loss": 4.174612045288086,
+ "step": 2344
+ },
+ {
+ "epoch": 32.57292576419214,
+ "grad_norm": 0.014725026674568653,
+ "learning_rate": 0.0006,
+ "loss": 4.066371440887451,
+ "step": 2345
+ },
+ {
+ "epoch": 32.58689956331878,
+ "grad_norm": 0.015003837645053864,
+ "learning_rate": 0.0006,
+ "loss": 4.193343639373779,
+ "step": 2346
+ },
+ {
+ "epoch": 32.60087336244541,
+ "grad_norm": 0.015346302650868893,
+ "learning_rate": 0.0006,
+ "loss": 4.210659980773926,
+ "step": 2347
+ },
+ {
+ "epoch": 32.61484716157205,
+ "grad_norm": 0.015135962516069412,
+ "learning_rate": 0.0006,
+ "loss": 4.204618453979492,
+ "step": 2348
+ },
+ {
+ "epoch": 32.62882096069869,
+ "grad_norm": 0.01601956970989704,
+ "learning_rate": 0.0006,
+ "loss": 4.145564556121826,
+ "step": 2349
+ },
+ {
+ "epoch": 32.64279475982533,
+ "grad_norm": 0.01517851185053587,
+ "learning_rate": 0.0006,
+ "loss": 4.051673889160156,
+ "step": 2350
+ },
+ {
+ "epoch": 32.65676855895197,
+ "grad_norm": 0.014894292689859867,
+ "learning_rate": 0.0006,
+ "loss": 4.184818267822266,
+ "step": 2351
+ },
+ {
+ "epoch": 32.670742358078606,
+ "grad_norm": 0.015720834955573082,
+ "learning_rate": 0.0006,
+ "loss": 4.220036506652832,
+ "step": 2352
+ },
+ {
+ "epoch": 32.68471615720524,
+ "grad_norm": 0.01573433168232441,
+ "learning_rate": 0.0006,
+ "loss": 4.076773166656494,
+ "step": 2353
+ },
+ {
+ "epoch": 32.698689956331876,
+ "grad_norm": 0.015828052535653114,
+ "learning_rate": 0.0006,
+ "loss": 4.116329669952393,
+ "step": 2354
+ },
+ {
+ "epoch": 32.712663755458514,
+ "grad_norm": 0.015462259761989117,
+ "learning_rate": 0.0006,
+ "loss": 4.1155242919921875,
+ "step": 2355
+ },
+ {
+ "epoch": 32.72663755458515,
+ "grad_norm": 0.015763849020004272,
+ "learning_rate": 0.0006,
+ "loss": 4.169687747955322,
+ "step": 2356
+ },
+ {
+ "epoch": 32.74061135371179,
+ "grad_norm": 0.014935145154595375,
+ "learning_rate": 0.0006,
+ "loss": 4.2090911865234375,
+ "step": 2357
+ },
+ {
+ "epoch": 32.75458515283843,
+ "grad_norm": 0.015119184739887714,
+ "learning_rate": 0.0006,
+ "loss": 4.131838798522949,
+ "step": 2358
+ },
+ {
+ "epoch": 32.76855895196506,
+ "grad_norm": 0.014833525754511356,
+ "learning_rate": 0.0006,
+ "loss": 4.17618465423584,
+ "step": 2359
+ },
+ {
+ "epoch": 32.7825327510917,
+ "grad_norm": 0.01588359661400318,
+ "learning_rate": 0.0006,
+ "loss": 4.163723945617676,
+ "step": 2360
+ },
+ {
+ "epoch": 32.79650655021834,
+ "grad_norm": 0.0168188214302063,
+ "learning_rate": 0.0006,
+ "loss": 4.178189277648926,
+ "step": 2361
+ },
+ {
+ "epoch": 32.81048034934498,
+ "grad_norm": 0.01726095750927925,
+ "learning_rate": 0.0006,
+ "loss": 4.144223690032959,
+ "step": 2362
+ },
+ {
+ "epoch": 32.82445414847162,
+ "grad_norm": 0.01701410301029682,
+ "learning_rate": 0.0006,
+ "loss": 4.209896087646484,
+ "step": 2363
+ },
+ {
+ "epoch": 32.838427947598255,
+ "grad_norm": 0.01682264916598797,
+ "learning_rate": 0.0006,
+ "loss": 4.26223087310791,
+ "step": 2364
+ },
+ {
+ "epoch": 32.852401746724894,
+ "grad_norm": 0.015597456134855747,
+ "learning_rate": 0.0006,
+ "loss": 4.292664051055908,
+ "step": 2365
+ },
+ {
+ "epoch": 32.866375545851525,
+ "grad_norm": 0.015258579514920712,
+ "learning_rate": 0.0006,
+ "loss": 4.165006637573242,
+ "step": 2366
+ },
+ {
+ "epoch": 32.880349344978164,
+ "grad_norm": 0.01662200503051281,
+ "learning_rate": 0.0006,
+ "loss": 4.309786319732666,
+ "step": 2367
+ },
+ {
+ "epoch": 32.8943231441048,
+ "grad_norm": 0.015691161155700684,
+ "learning_rate": 0.0006,
+ "loss": 4.167202949523926,
+ "step": 2368
+ },
+ {
+ "epoch": 32.90829694323144,
+ "grad_norm": 0.014870637096464634,
+ "learning_rate": 0.0006,
+ "loss": 4.142415523529053,
+ "step": 2369
+ },
+ {
+ "epoch": 32.92227074235808,
+ "grad_norm": 0.01567252166569233,
+ "learning_rate": 0.0006,
+ "loss": 4.2312092781066895,
+ "step": 2370
+ },
+ {
+ "epoch": 32.93624454148472,
+ "grad_norm": 0.014948616735637188,
+ "learning_rate": 0.0006,
+ "loss": 4.1514129638671875,
+ "step": 2371
+ },
+ {
+ "epoch": 32.95021834061135,
+ "grad_norm": 0.014654500409960747,
+ "learning_rate": 0.0006,
+ "loss": 4.205010890960693,
+ "step": 2372
+ },
+ {
+ "epoch": 32.96419213973799,
+ "grad_norm": 0.015059148892760277,
+ "learning_rate": 0.0006,
+ "loss": 4.148405075073242,
+ "step": 2373
+ },
+ {
+ "epoch": 32.97816593886463,
+ "grad_norm": 0.015026576817035675,
+ "learning_rate": 0.0006,
+ "loss": 4.182535171508789,
+ "step": 2374
+ },
+ {
+ "epoch": 32.992139737991266,
+ "grad_norm": 0.016394320875406265,
+ "learning_rate": 0.0006,
+ "loss": 4.129051208496094,
+ "step": 2375
+ },
+ {
+ "epoch": 33.0,
+ "grad_norm": 0.01867043413221836,
+ "learning_rate": 0.0006,
+ "loss": 4.185556411743164,
+ "step": 2376
+ },
+ {
+ "epoch": 33.0,
+ "eval_loss": 4.640405654907227,
+ "eval_runtime": 56.5861,
+ "eval_samples_per_second": 43.155,
+ "eval_steps_per_second": 1.361,
+ "step": 2376
+ },
+ {
+ "epoch": 33.01397379912664,
+ "grad_norm": 0.018626727163791656,
+ "learning_rate": 0.0006,
+ "loss": 4.143031597137451,
+ "step": 2377
+ },
+ {
+ "epoch": 33.02794759825328,
+ "grad_norm": 0.019283363595604897,
+ "learning_rate": 0.0006,
+ "loss": 4.117845058441162,
+ "step": 2378
+ },
+ {
+ "epoch": 33.041921397379916,
+ "grad_norm": 0.018185822293162346,
+ "learning_rate": 0.0006,
+ "loss": 4.076862812042236,
+ "step": 2379
+ },
+ {
+ "epoch": 33.05589519650655,
+ "grad_norm": 0.019411196932196617,
+ "learning_rate": 0.0006,
+ "loss": 4.124156475067139,
+ "step": 2380
+ },
+ {
+ "epoch": 33.069868995633186,
+ "grad_norm": 0.019248466938734055,
+ "learning_rate": 0.0006,
+ "loss": 4.201735019683838,
+ "step": 2381
+ },
+ {
+ "epoch": 33.083842794759825,
+ "grad_norm": 0.018856720998883247,
+ "learning_rate": 0.0006,
+ "loss": 4.00296688079834,
+ "step": 2382
+ },
+ {
+ "epoch": 33.09781659388646,
+ "grad_norm": 0.018554294481873512,
+ "learning_rate": 0.0006,
+ "loss": 4.270577430725098,
+ "step": 2383
+ },
+ {
+ "epoch": 33.1117903930131,
+ "grad_norm": 0.019946224987506866,
+ "learning_rate": 0.0006,
+ "loss": 4.119589328765869,
+ "step": 2384
+ },
+ {
+ "epoch": 33.12576419213974,
+ "grad_norm": 0.021395670250058174,
+ "learning_rate": 0.0006,
+ "loss": 4.128847122192383,
+ "step": 2385
+ },
+ {
+ "epoch": 33.13973799126637,
+ "grad_norm": 0.02009674906730652,
+ "learning_rate": 0.0006,
+ "loss": 4.183228969573975,
+ "step": 2386
+ },
+ {
+ "epoch": 33.15371179039301,
+ "grad_norm": 0.01820223033428192,
+ "learning_rate": 0.0006,
+ "loss": 4.131601810455322,
+ "step": 2387
+ },
+ {
+ "epoch": 33.16768558951965,
+ "grad_norm": 0.018653156235814095,
+ "learning_rate": 0.0006,
+ "loss": 4.116430282592773,
+ "step": 2388
+ },
+ {
+ "epoch": 33.18165938864629,
+ "grad_norm": 0.016828063875436783,
+ "learning_rate": 0.0006,
+ "loss": 4.054663181304932,
+ "step": 2389
+ },
+ {
+ "epoch": 33.19563318777293,
+ "grad_norm": 0.017025409266352654,
+ "learning_rate": 0.0006,
+ "loss": 4.110044956207275,
+ "step": 2390
+ },
+ {
+ "epoch": 33.209606986899566,
+ "grad_norm": 0.016608452424407005,
+ "learning_rate": 0.0006,
+ "loss": 4.102668285369873,
+ "step": 2391
+ },
+ {
+ "epoch": 33.223580786026204,
+ "grad_norm": 0.01743577979505062,
+ "learning_rate": 0.0006,
+ "loss": 4.166791915893555,
+ "step": 2392
+ },
+ {
+ "epoch": 33.237554585152836,
+ "grad_norm": 0.017253734171390533,
+ "learning_rate": 0.0006,
+ "loss": 4.217832088470459,
+ "step": 2393
+ },
+ {
+ "epoch": 33.251528384279474,
+ "grad_norm": 0.017937595024704933,
+ "learning_rate": 0.0006,
+ "loss": 4.2173967361450195,
+ "step": 2394
+ },
+ {
+ "epoch": 33.26550218340611,
+ "grad_norm": 0.01985993981361389,
+ "learning_rate": 0.0006,
+ "loss": 4.17849588394165,
+ "step": 2395
+ },
+ {
+ "epoch": 33.27947598253275,
+ "grad_norm": 0.019243355840444565,
+ "learning_rate": 0.0006,
+ "loss": 4.086523056030273,
+ "step": 2396
+ },
+ {
+ "epoch": 33.29344978165939,
+ "grad_norm": 0.01924416422843933,
+ "learning_rate": 0.0006,
+ "loss": 4.128574371337891,
+ "step": 2397
+ },
+ {
+ "epoch": 33.30742358078603,
+ "grad_norm": 0.01983758620917797,
+ "learning_rate": 0.0006,
+ "loss": 4.061581611633301,
+ "step": 2398
+ },
+ {
+ "epoch": 33.32139737991266,
+ "grad_norm": 0.01930462196469307,
+ "learning_rate": 0.0006,
+ "loss": 4.091731071472168,
+ "step": 2399
+ },
+ {
+ "epoch": 33.3353711790393,
+ "grad_norm": 0.017988424748182297,
+ "learning_rate": 0.0006,
+ "loss": 4.155599117279053,
+ "step": 2400
+ },
+ {
+ "epoch": 33.34934497816594,
+ "grad_norm": 0.015194767154753208,
+ "learning_rate": 0.0006,
+ "loss": 4.144134044647217,
+ "step": 2401
+ },
+ {
+ "epoch": 33.36331877729258,
+ "grad_norm": 0.01536081824451685,
+ "learning_rate": 0.0006,
+ "loss": 4.127272129058838,
+ "step": 2402
+ },
+ {
+ "epoch": 33.377292576419215,
+ "grad_norm": 0.015431297942996025,
+ "learning_rate": 0.0006,
+ "loss": 4.3213019371032715,
+ "step": 2403
+ },
+ {
+ "epoch": 33.391266375545854,
+ "grad_norm": 0.015740280970931053,
+ "learning_rate": 0.0006,
+ "loss": 4.121467590332031,
+ "step": 2404
+ },
+ {
+ "epoch": 33.40524017467249,
+ "grad_norm": 0.015379384160041809,
+ "learning_rate": 0.0006,
+ "loss": 4.095216751098633,
+ "step": 2405
+ },
+ {
+ "epoch": 33.419213973799124,
+ "grad_norm": 0.01528073474764824,
+ "learning_rate": 0.0006,
+ "loss": 4.13047981262207,
+ "step": 2406
+ },
+ {
+ "epoch": 33.43318777292576,
+ "grad_norm": 0.015739979222416878,
+ "learning_rate": 0.0006,
+ "loss": 4.2533135414123535,
+ "step": 2407
+ },
+ {
+ "epoch": 33.4471615720524,
+ "grad_norm": 0.01519764307886362,
+ "learning_rate": 0.0006,
+ "loss": 4.131782054901123,
+ "step": 2408
+ },
+ {
+ "epoch": 33.46113537117904,
+ "grad_norm": 0.015425996854901314,
+ "learning_rate": 0.0006,
+ "loss": 4.02811336517334,
+ "step": 2409
+ },
+ {
+ "epoch": 33.47510917030568,
+ "grad_norm": 0.015378830023109913,
+ "learning_rate": 0.0006,
+ "loss": 4.08652925491333,
+ "step": 2410
+ },
+ {
+ "epoch": 33.48908296943232,
+ "grad_norm": 0.01473075058311224,
+ "learning_rate": 0.0006,
+ "loss": 4.214364528656006,
+ "step": 2411
+ },
+ {
+ "epoch": 33.50305676855895,
+ "grad_norm": 0.015141637064516544,
+ "learning_rate": 0.0006,
+ "loss": 4.155712127685547,
+ "step": 2412
+ },
+ {
+ "epoch": 33.51703056768559,
+ "grad_norm": 0.01641593500971794,
+ "learning_rate": 0.0006,
+ "loss": 4.097907066345215,
+ "step": 2413
+ },
+ {
+ "epoch": 33.531004366812226,
+ "grad_norm": 0.015182029455900192,
+ "learning_rate": 0.0006,
+ "loss": 4.238499641418457,
+ "step": 2414
+ },
+ {
+ "epoch": 33.544978165938865,
+ "grad_norm": 0.01643732376396656,
+ "learning_rate": 0.0006,
+ "loss": 4.0661492347717285,
+ "step": 2415
+ },
+ {
+ "epoch": 33.5589519650655,
+ "grad_norm": 0.015614228323101997,
+ "learning_rate": 0.0006,
+ "loss": 4.120795249938965,
+ "step": 2416
+ },
+ {
+ "epoch": 33.57292576419214,
+ "grad_norm": 0.016066187992691994,
+ "learning_rate": 0.0006,
+ "loss": 4.096465110778809,
+ "step": 2417
+ },
+ {
+ "epoch": 33.58689956331878,
+ "grad_norm": 0.017214342951774597,
+ "learning_rate": 0.0006,
+ "loss": 4.242628574371338,
+ "step": 2418
+ },
+ {
+ "epoch": 33.60087336244541,
+ "grad_norm": 0.01787586510181427,
+ "learning_rate": 0.0006,
+ "loss": 4.149589538574219,
+ "step": 2419
+ },
+ {
+ "epoch": 33.61484716157205,
+ "grad_norm": 0.019455241039395332,
+ "learning_rate": 0.0006,
+ "loss": 4.267796039581299,
+ "step": 2420
+ },
+ {
+ "epoch": 33.62882096069869,
+ "grad_norm": 0.01984674669802189,
+ "learning_rate": 0.0006,
+ "loss": 4.175220966339111,
+ "step": 2421
+ },
+ {
+ "epoch": 33.64279475982533,
+ "grad_norm": 0.021933717653155327,
+ "learning_rate": 0.0006,
+ "loss": 4.106647491455078,
+ "step": 2422
+ },
+ {
+ "epoch": 33.65676855895197,
+ "grad_norm": 0.02123589813709259,
+ "learning_rate": 0.0006,
+ "loss": 4.204996109008789,
+ "step": 2423
+ },
+ {
+ "epoch": 33.670742358078606,
+ "grad_norm": 0.019814589992165565,
+ "learning_rate": 0.0006,
+ "loss": 4.186171531677246,
+ "step": 2424
+ },
+ {
+ "epoch": 33.68471615720524,
+ "grad_norm": 0.02059728465974331,
+ "learning_rate": 0.0006,
+ "loss": 4.170509338378906,
+ "step": 2425
+ },
+ {
+ "epoch": 33.698689956331876,
+ "grad_norm": 0.018483608961105347,
+ "learning_rate": 0.0006,
+ "loss": 4.085585594177246,
+ "step": 2426
+ },
+ {
+ "epoch": 33.712663755458514,
+ "grad_norm": 0.018403852358460426,
+ "learning_rate": 0.0006,
+ "loss": 4.2551774978637695,
+ "step": 2427
+ },
+ {
+ "epoch": 33.72663755458515,
+ "grad_norm": 0.018653830513358116,
+ "learning_rate": 0.0006,
+ "loss": 4.091766357421875,
+ "step": 2428
+ },
+ {
+ "epoch": 33.74061135371179,
+ "grad_norm": 0.017692334949970245,
+ "learning_rate": 0.0006,
+ "loss": 4.100838661193848,
+ "step": 2429
+ },
+ {
+ "epoch": 33.75458515283843,
+ "grad_norm": 0.016029126942157745,
+ "learning_rate": 0.0006,
+ "loss": 4.1530680656433105,
+ "step": 2430
+ },
+ {
+ "epoch": 33.76855895196506,
+ "grad_norm": 0.016095684841275215,
+ "learning_rate": 0.0006,
+ "loss": 4.20544958114624,
+ "step": 2431
+ },
+ {
+ "epoch": 33.7825327510917,
+ "grad_norm": 0.016375798732042313,
+ "learning_rate": 0.0006,
+ "loss": 4.199177265167236,
+ "step": 2432
+ },
+ {
+ "epoch": 33.79650655021834,
+ "grad_norm": 0.01590786688029766,
+ "learning_rate": 0.0006,
+ "loss": 4.200799465179443,
+ "step": 2433
+ },
+ {
+ "epoch": 33.81048034934498,
+ "grad_norm": 0.014896097593009472,
+ "learning_rate": 0.0006,
+ "loss": 4.185850143432617,
+ "step": 2434
+ },
+ {
+ "epoch": 33.82445414847162,
+ "grad_norm": 0.01438513956964016,
+ "learning_rate": 0.0006,
+ "loss": 4.214962005615234,
+ "step": 2435
+ },
+ {
+ "epoch": 33.838427947598255,
+ "grad_norm": 0.014459339901804924,
+ "learning_rate": 0.0006,
+ "loss": 4.196972370147705,
+ "step": 2436
+ },
+ {
+ "epoch": 33.852401746724894,
+ "grad_norm": 0.014782169833779335,
+ "learning_rate": 0.0006,
+ "loss": 4.0885910987854,
+ "step": 2437
+ },
+ {
+ "epoch": 33.866375545851525,
+ "grad_norm": 0.01520370040088892,
+ "learning_rate": 0.0006,
+ "loss": 4.0323262214660645,
+ "step": 2438
+ },
+ {
+ "epoch": 33.880349344978164,
+ "grad_norm": 0.014056282117962837,
+ "learning_rate": 0.0006,
+ "loss": 4.1386518478393555,
+ "step": 2439
+ },
+ {
+ "epoch": 33.8943231441048,
+ "grad_norm": 0.013567990623414516,
+ "learning_rate": 0.0006,
+ "loss": 4.095754623413086,
+ "step": 2440
+ },
+ {
+ "epoch": 33.90829694323144,
+ "grad_norm": 0.013306716457009315,
+ "learning_rate": 0.0006,
+ "loss": 4.1623077392578125,
+ "step": 2441
+ },
+ {
+ "epoch": 33.92227074235808,
+ "grad_norm": 0.014047150500118732,
+ "learning_rate": 0.0006,
+ "loss": 4.142541885375977,
+ "step": 2442
+ },
+ {
+ "epoch": 33.93624454148472,
+ "grad_norm": 0.01377617847174406,
+ "learning_rate": 0.0006,
+ "loss": 4.268551826477051,
+ "step": 2443
+ },
+ {
+ "epoch": 33.95021834061135,
+ "grad_norm": 0.014671621844172478,
+ "learning_rate": 0.0006,
+ "loss": 4.198196887969971,
+ "step": 2444
+ },
+ {
+ "epoch": 33.96419213973799,
+ "grad_norm": 0.014806744642555714,
+ "learning_rate": 0.0006,
+ "loss": 4.057899475097656,
+ "step": 2445
+ },
+ {
+ "epoch": 33.97816593886463,
+ "grad_norm": 0.015734272077679634,
+ "learning_rate": 0.0006,
+ "loss": 4.163505554199219,
+ "step": 2446
+ },
+ {
+ "epoch": 33.992139737991266,
+ "grad_norm": 0.015250146389007568,
+ "learning_rate": 0.0006,
+ "loss": 4.076294898986816,
+ "step": 2447
+ },
+ {
+ "epoch": 34.0,
+ "grad_norm": 0.016733702272176743,
+ "learning_rate": 0.0006,
+ "loss": 4.138420104980469,
+ "step": 2448
+ },
+ {
+ "epoch": 34.0,
+ "eval_loss": 4.542503356933594,
+ "eval_runtime": 56.983,
+ "eval_samples_per_second": 42.855,
+ "eval_steps_per_second": 1.351,
+ "step": 2448
+ },
+ {
+ "epoch": 34.01397379912664,
+ "grad_norm": 0.015866613015532494,
+ "learning_rate": 0.0006,
+ "loss": 4.032081127166748,
+ "step": 2449
+ },
+ {
+ "epoch": 34.02794759825328,
+ "grad_norm": 0.01946570910513401,
+ "learning_rate": 0.0006,
+ "loss": 4.122320175170898,
+ "step": 2450
+ },
+ {
+ "epoch": 34.041921397379916,
+ "grad_norm": 0.020265938714146614,
+ "learning_rate": 0.0006,
+ "loss": 4.06460428237915,
+ "step": 2451
+ },
+ {
+ "epoch": 34.05589519650655,
+ "grad_norm": 0.018887171521782875,
+ "learning_rate": 0.0006,
+ "loss": 4.096906661987305,
+ "step": 2452
+ },
+ {
+ "epoch": 34.069868995633186,
+ "grad_norm": 0.017304692417383194,
+ "learning_rate": 0.0006,
+ "loss": 4.137261390686035,
+ "step": 2453
+ },
+ {
+ "epoch": 34.083842794759825,
+ "grad_norm": 0.017029082402586937,
+ "learning_rate": 0.0006,
+ "loss": 4.081112861633301,
+ "step": 2454
+ },
+ {
+ "epoch": 34.09781659388646,
+ "grad_norm": 0.016414813697338104,
+ "learning_rate": 0.0006,
+ "loss": 4.140783309936523,
+ "step": 2455
+ },
+ {
+ "epoch": 34.1117903930131,
+ "grad_norm": 0.018238281831145287,
+ "learning_rate": 0.0006,
+ "loss": 4.161116600036621,
+ "step": 2456
+ },
+ {
+ "epoch": 34.12576419213974,
+ "grad_norm": 0.019929232075810432,
+ "learning_rate": 0.0006,
+ "loss": 4.243722438812256,
+ "step": 2457
+ },
+ {
+ "epoch": 34.13973799126637,
+ "grad_norm": 0.019375786185264587,
+ "learning_rate": 0.0006,
+ "loss": 4.134873390197754,
+ "step": 2458
+ },
+ {
+ "epoch": 34.15371179039301,
+ "grad_norm": 0.018151137977838516,
+ "learning_rate": 0.0006,
+ "loss": 4.098745346069336,
+ "step": 2459
+ },
+ {
+ "epoch": 34.16768558951965,
+ "grad_norm": 0.017054934054613113,
+ "learning_rate": 0.0006,
+ "loss": 4.194662570953369,
+ "step": 2460
+ },
+ {
+ "epoch": 34.18165938864629,
+ "grad_norm": 0.018220555037260056,
+ "learning_rate": 0.0006,
+ "loss": 4.128866195678711,
+ "step": 2461
+ },
+ {
+ "epoch": 34.19563318777293,
+ "grad_norm": 0.01692879945039749,
+ "learning_rate": 0.0006,
+ "loss": 4.062477111816406,
+ "step": 2462
+ },
+ {
+ "epoch": 34.209606986899566,
+ "grad_norm": 0.01674199476838112,
+ "learning_rate": 0.0006,
+ "loss": 4.186149597167969,
+ "step": 2463
+ },
+ {
+ "epoch": 34.223580786026204,
+ "grad_norm": 0.016527455300092697,
+ "learning_rate": 0.0006,
+ "loss": 4.0734028816223145,
+ "step": 2464
+ },
+ {
+ "epoch": 34.237554585152836,
+ "grad_norm": 0.016448386013507843,
+ "learning_rate": 0.0006,
+ "loss": 4.098896026611328,
+ "step": 2465
+ },
+ {
+ "epoch": 34.251528384279474,
+ "grad_norm": 0.014644953422248363,
+ "learning_rate": 0.0006,
+ "loss": 4.1031904220581055,
+ "step": 2466
+ },
+ {
+ "epoch": 34.26550218340611,
+ "grad_norm": 0.016649341210722923,
+ "learning_rate": 0.0006,
+ "loss": 4.148791313171387,
+ "step": 2467
+ },
+ {
+ "epoch": 34.27947598253275,
+ "grad_norm": 0.017442714422941208,
+ "learning_rate": 0.0006,
+ "loss": 4.070799827575684,
+ "step": 2468
+ },
+ {
+ "epoch": 34.29344978165939,
+ "grad_norm": 0.017852013930678368,
+ "learning_rate": 0.0006,
+ "loss": 4.031924247741699,
+ "step": 2469
+ },
+ {
+ "epoch": 34.30742358078603,
+ "grad_norm": 0.016371270641684532,
+ "learning_rate": 0.0006,
+ "loss": 4.149420738220215,
+ "step": 2470
+ },
+ {
+ "epoch": 34.32139737991266,
+ "grad_norm": 0.015846814960241318,
+ "learning_rate": 0.0006,
+ "loss": 4.15427303314209,
+ "step": 2471
+ },
+ {
+ "epoch": 34.3353711790393,
+ "grad_norm": 0.017411163076758385,
+ "learning_rate": 0.0006,
+ "loss": 4.1304192543029785,
+ "step": 2472
+ },
+ {
+ "epoch": 34.34934497816594,
+ "grad_norm": 0.01780046708881855,
+ "learning_rate": 0.0006,
+ "loss": 4.191732406616211,
+ "step": 2473
+ },
+ {
+ "epoch": 34.36331877729258,
+ "grad_norm": 0.01958383060991764,
+ "learning_rate": 0.0006,
+ "loss": 4.068355083465576,
+ "step": 2474
+ },
+ {
+ "epoch": 34.377292576419215,
+ "grad_norm": 0.022173317149281502,
+ "learning_rate": 0.0006,
+ "loss": 4.182929039001465,
+ "step": 2475
+ },
+ {
+ "epoch": 34.391266375545854,
+ "grad_norm": 0.022389156743884087,
+ "learning_rate": 0.0006,
+ "loss": 4.110218048095703,
+ "step": 2476
+ },
+ {
+ "epoch": 34.40524017467249,
+ "grad_norm": 0.022478511556982994,
+ "learning_rate": 0.0006,
+ "loss": 4.095479488372803,
+ "step": 2477
+ },
+ {
+ "epoch": 34.419213973799124,
+ "grad_norm": 0.019504647701978683,
+ "learning_rate": 0.0006,
+ "loss": 4.101964950561523,
+ "step": 2478
+ },
+ {
+ "epoch": 34.43318777292576,
+ "grad_norm": 0.019378576427698135,
+ "learning_rate": 0.0006,
+ "loss": 4.14620304107666,
+ "step": 2479
+ },
+ {
+ "epoch": 34.4471615720524,
+ "grad_norm": 0.017993204295635223,
+ "learning_rate": 0.0006,
+ "loss": 4.183492660522461,
+ "step": 2480
+ },
+ {
+ "epoch": 34.46113537117904,
+ "grad_norm": 0.01814805157482624,
+ "learning_rate": 0.0006,
+ "loss": 4.151247978210449,
+ "step": 2481
+ },
+ {
+ "epoch": 34.47510917030568,
+ "grad_norm": 0.018622448667883873,
+ "learning_rate": 0.0006,
+ "loss": 4.020938396453857,
+ "step": 2482
+ },
+ {
+ "epoch": 34.48908296943232,
+ "grad_norm": 0.018360132351517677,
+ "learning_rate": 0.0006,
+ "loss": 4.064734935760498,
+ "step": 2483
+ },
+ {
+ "epoch": 34.50305676855895,
+ "grad_norm": 0.017715711146593094,
+ "learning_rate": 0.0006,
+ "loss": 4.185687065124512,
+ "step": 2484
+ },
+ {
+ "epoch": 34.51703056768559,
+ "grad_norm": 0.01714295893907547,
+ "learning_rate": 0.0006,
+ "loss": 4.1044721603393555,
+ "step": 2485
+ },
+ {
+ "epoch": 34.531004366812226,
+ "grad_norm": 0.016943803057074547,
+ "learning_rate": 0.0006,
+ "loss": 4.106837272644043,
+ "step": 2486
+ },
+ {
+ "epoch": 34.544978165938865,
+ "grad_norm": 0.0182423684746027,
+ "learning_rate": 0.0006,
+ "loss": 4.1041340827941895,
+ "step": 2487
+ },
+ {
+ "epoch": 34.5589519650655,
+ "grad_norm": 0.01636587455868721,
+ "learning_rate": 0.0006,
+ "loss": 3.9735546112060547,
+ "step": 2488
+ },
+ {
+ "epoch": 34.57292576419214,
+ "grad_norm": 0.014593546278774738,
+ "learning_rate": 0.0006,
+ "loss": 4.041980266571045,
+ "step": 2489
+ },
+ {
+ "epoch": 34.58689956331878,
+ "grad_norm": 0.015436965972185135,
+ "learning_rate": 0.0006,
+ "loss": 4.301148414611816,
+ "step": 2490
+ },
+ {
+ "epoch": 34.60087336244541,
+ "grad_norm": 0.015075269155204296,
+ "learning_rate": 0.0006,
+ "loss": 4.118527412414551,
+ "step": 2491
+ },
+ {
+ "epoch": 34.61484716157205,
+ "grad_norm": 0.01701514609158039,
+ "learning_rate": 0.0006,
+ "loss": 4.245968818664551,
+ "step": 2492
+ },
+ {
+ "epoch": 34.62882096069869,
+ "grad_norm": 0.018047887831926346,
+ "learning_rate": 0.0006,
+ "loss": 4.227261066436768,
+ "step": 2493
+ },
+ {
+ "epoch": 34.64279475982533,
+ "grad_norm": 0.018499085679650307,
+ "learning_rate": 0.0006,
+ "loss": 4.121360778808594,
+ "step": 2494
+ },
+ {
+ "epoch": 34.65676855895197,
+ "grad_norm": 0.018987711519002914,
+ "learning_rate": 0.0006,
+ "loss": 4.131564617156982,
+ "step": 2495
+ },
+ {
+ "epoch": 34.670742358078606,
+ "grad_norm": 0.01834476925432682,
+ "learning_rate": 0.0006,
+ "loss": 4.224941253662109,
+ "step": 2496
+ },
+ {
+ "epoch": 34.68471615720524,
+ "grad_norm": 0.017507942393422127,
+ "learning_rate": 0.0006,
+ "loss": 4.1394548416137695,
+ "step": 2497
+ },
+ {
+ "epoch": 34.698689956331876,
+ "grad_norm": 0.01680580899119377,
+ "learning_rate": 0.0006,
+ "loss": 4.1073431968688965,
+ "step": 2498
+ },
+ {
+ "epoch": 34.712663755458514,
+ "grad_norm": 0.015233514830470085,
+ "learning_rate": 0.0006,
+ "loss": 4.136994361877441,
+ "step": 2499
+ },
+ {
+ "epoch": 34.72663755458515,
+ "grad_norm": 0.015482917428016663,
+ "learning_rate": 0.0006,
+ "loss": 4.110719680786133,
+ "step": 2500
+ },
+ {
+ "epoch": 34.74061135371179,
+ "grad_norm": 0.015940461307764053,
+ "learning_rate": 0.0006,
+ "loss": 4.207851886749268,
+ "step": 2501
+ },
+ {
+ "epoch": 34.75458515283843,
+ "grad_norm": 0.015355597250163555,
+ "learning_rate": 0.0006,
+ "loss": 4.169469833374023,
+ "step": 2502
+ },
+ {
+ "epoch": 34.76855895196506,
+ "grad_norm": 0.015974000096321106,
+ "learning_rate": 0.0006,
+ "loss": 4.110061168670654,
+ "step": 2503
+ },
+ {
+ "epoch": 34.7825327510917,
+ "grad_norm": 0.016088927164673805,
+ "learning_rate": 0.0006,
+ "loss": 4.155452728271484,
+ "step": 2504
+ },
+ {
+ "epoch": 34.79650655021834,
+ "grad_norm": 0.016156120225787163,
+ "learning_rate": 0.0006,
+ "loss": 4.178092956542969,
+ "step": 2505
+ },
+ {
+ "epoch": 34.81048034934498,
+ "grad_norm": 0.014928647316992283,
+ "learning_rate": 0.0006,
+ "loss": 4.05457878112793,
+ "step": 2506
+ },
+ {
+ "epoch": 34.82445414847162,
+ "grad_norm": 0.015289328061044216,
+ "learning_rate": 0.0006,
+ "loss": 4.18869686126709,
+ "step": 2507
+ },
+ {
+ "epoch": 34.838427947598255,
+ "grad_norm": 0.015799039974808693,
+ "learning_rate": 0.0006,
+ "loss": 4.106073379516602,
+ "step": 2508
+ },
+ {
+ "epoch": 34.852401746724894,
+ "grad_norm": 0.015824418514966965,
+ "learning_rate": 0.0006,
+ "loss": 4.0634589195251465,
+ "step": 2509
+ },
+ {
+ "epoch": 34.866375545851525,
+ "grad_norm": 0.015321656130254269,
+ "learning_rate": 0.0006,
+ "loss": 4.093090057373047,
+ "step": 2510
+ },
+ {
+ "epoch": 34.880349344978164,
+ "grad_norm": 0.014380166307091713,
+ "learning_rate": 0.0006,
+ "loss": 4.1750640869140625,
+ "step": 2511
+ },
+ {
+ "epoch": 34.8943231441048,
+ "grad_norm": 0.014962567947804928,
+ "learning_rate": 0.0006,
+ "loss": 4.185277938842773,
+ "step": 2512
+ },
+ {
+ "epoch": 34.90829694323144,
+ "grad_norm": 0.014467543922364712,
+ "learning_rate": 0.0006,
+ "loss": 4.1600799560546875,
+ "step": 2513
+ },
+ {
+ "epoch": 34.92227074235808,
+ "grad_norm": 0.01545181218534708,
+ "learning_rate": 0.0006,
+ "loss": 4.096299648284912,
+ "step": 2514
+ },
+ {
+ "epoch": 34.93624454148472,
+ "grad_norm": 0.01605340465903282,
+ "learning_rate": 0.0006,
+ "loss": 4.203323841094971,
+ "step": 2515
+ },
+ {
+ "epoch": 34.95021834061135,
+ "grad_norm": 0.015203815884888172,
+ "learning_rate": 0.0006,
+ "loss": 4.201385021209717,
+ "step": 2516
+ },
+ {
+ "epoch": 34.96419213973799,
+ "grad_norm": 0.015149409882724285,
+ "learning_rate": 0.0006,
+ "loss": 4.088164329528809,
+ "step": 2517
+ },
+ {
+ "epoch": 34.97816593886463,
+ "grad_norm": 0.014885558746755123,
+ "learning_rate": 0.0006,
+ "loss": 4.073911666870117,
+ "step": 2518
+ },
+ {
+ "epoch": 34.992139737991266,
+ "grad_norm": 0.015003419481217861,
+ "learning_rate": 0.0006,
+ "loss": 4.157113552093506,
+ "step": 2519
+ },
+ {
+ "epoch": 35.0,
+ "grad_norm": 0.01641916297376156,
+ "learning_rate": 0.0006,
+ "loss": 3.970052719116211,
+ "step": 2520
+ },
+ {
+ "epoch": 35.0,
+ "eval_loss": 4.641692161560059,
+ "eval_runtime": 56.8836,
+ "eval_samples_per_second": 42.93,
+ "eval_steps_per_second": 1.354,
+ "step": 2520
+ },
+ {
+ "epoch": 35.01397379912664,
+ "grad_norm": 0.015894673764705658,
+ "learning_rate": 0.0006,
+ "loss": 4.104437351226807,
+ "step": 2521
+ },
+ {
+ "epoch": 35.02794759825328,
+ "grad_norm": 0.0173867866396904,
+ "learning_rate": 0.0006,
+ "loss": 4.091726303100586,
+ "step": 2522
+ },
+ {
+ "epoch": 35.041921397379916,
+ "grad_norm": 0.019272832199931145,
+ "learning_rate": 0.0006,
+ "loss": 4.217905044555664,
+ "step": 2523
+ },
+ {
+ "epoch": 35.05589519650655,
+ "grad_norm": 0.021820371970534325,
+ "learning_rate": 0.0006,
+ "loss": 4.111428260803223,
+ "step": 2524
+ },
+ {
+ "epoch": 35.069868995633186,
+ "grad_norm": 0.02340741828083992,
+ "learning_rate": 0.0006,
+ "loss": 4.115771293640137,
+ "step": 2525
+ },
+ {
+ "epoch": 35.083842794759825,
+ "grad_norm": 0.023760691285133362,
+ "learning_rate": 0.0006,
+ "loss": 4.132383346557617,
+ "step": 2526
+ },
+ {
+ "epoch": 35.09781659388646,
+ "grad_norm": 0.02073468267917633,
+ "learning_rate": 0.0006,
+ "loss": 4.272691249847412,
+ "step": 2527
+ },
+ {
+ "epoch": 35.1117903930131,
+ "grad_norm": 0.02197718247771263,
+ "learning_rate": 0.0006,
+ "loss": 4.105586528778076,
+ "step": 2528
+ },
+ {
+ "epoch": 35.12576419213974,
+ "grad_norm": 0.0200254675000906,
+ "learning_rate": 0.0006,
+ "loss": 3.968820095062256,
+ "step": 2529
+ },
+ {
+ "epoch": 35.13973799126637,
+ "grad_norm": 0.01929587684571743,
+ "learning_rate": 0.0006,
+ "loss": 4.045468807220459,
+ "step": 2530
+ },
+ {
+ "epoch": 35.15371179039301,
+ "grad_norm": 0.01886162906885147,
+ "learning_rate": 0.0006,
+ "loss": 4.185571670532227,
+ "step": 2531
+ },
+ {
+ "epoch": 35.16768558951965,
+ "grad_norm": 0.019789839163422585,
+ "learning_rate": 0.0006,
+ "loss": 3.9734551906585693,
+ "step": 2532
+ },
+ {
+ "epoch": 35.18165938864629,
+ "grad_norm": 0.0198142658919096,
+ "learning_rate": 0.0006,
+ "loss": 4.2662506103515625,
+ "step": 2533
+ },
+ {
+ "epoch": 35.19563318777293,
+ "grad_norm": 0.02087012305855751,
+ "learning_rate": 0.0006,
+ "loss": 4.175931930541992,
+ "step": 2534
+ },
+ {
+ "epoch": 35.209606986899566,
+ "grad_norm": 0.020294535905122757,
+ "learning_rate": 0.0006,
+ "loss": 4.060324668884277,
+ "step": 2535
+ },
+ {
+ "epoch": 35.223580786026204,
+ "grad_norm": 0.01908983662724495,
+ "learning_rate": 0.0006,
+ "loss": 4.217939853668213,
+ "step": 2536
+ },
+ {
+ "epoch": 35.237554585152836,
+ "grad_norm": 0.0178892333060503,
+ "learning_rate": 0.0006,
+ "loss": 4.190216541290283,
+ "step": 2537
+ },
+ {
+ "epoch": 35.251528384279474,
+ "grad_norm": 0.017892315983772278,
+ "learning_rate": 0.0006,
+ "loss": 4.067419052124023,
+ "step": 2538
+ },
+ {
+ "epoch": 35.26550218340611,
+ "grad_norm": 0.018556250259280205,
+ "learning_rate": 0.0006,
+ "loss": 4.105958461761475,
+ "step": 2539
+ },
+ {
+ "epoch": 35.27947598253275,
+ "grad_norm": 0.01944422349333763,
+ "learning_rate": 0.0006,
+ "loss": 4.155775547027588,
+ "step": 2540
+ },
+ {
+ "epoch": 35.29344978165939,
+ "grad_norm": 0.017333146184682846,
+ "learning_rate": 0.0006,
+ "loss": 4.244513511657715,
+ "step": 2541
+ },
+ {
+ "epoch": 35.30742358078603,
+ "grad_norm": 0.01639244332909584,
+ "learning_rate": 0.0006,
+ "loss": 4.167815685272217,
+ "step": 2542
+ },
+ {
+ "epoch": 35.32139737991266,
+ "grad_norm": 0.01610143668949604,
+ "learning_rate": 0.0006,
+ "loss": 4.121486663818359,
+ "step": 2543
+ },
+ {
+ "epoch": 35.3353711790393,
+ "grad_norm": 0.01656411588191986,
+ "learning_rate": 0.0006,
+ "loss": 4.1755499839782715,
+ "step": 2544
+ },
+ {
+ "epoch": 35.34934497816594,
+ "grad_norm": 0.015075297094881535,
+ "learning_rate": 0.0006,
+ "loss": 4.102992534637451,
+ "step": 2545
+ },
+ {
+ "epoch": 35.36331877729258,
+ "grad_norm": 0.01538170501589775,
+ "learning_rate": 0.0006,
+ "loss": 4.161603927612305,
+ "step": 2546
+ },
+ {
+ "epoch": 35.377292576419215,
+ "grad_norm": 0.01483312901109457,
+ "learning_rate": 0.0006,
+ "loss": 4.085355758666992,
+ "step": 2547
+ },
+ {
+ "epoch": 35.391266375545854,
+ "grad_norm": 0.014894738793373108,
+ "learning_rate": 0.0006,
+ "loss": 4.083585262298584,
+ "step": 2548
+ },
+ {
+ "epoch": 35.40524017467249,
+ "grad_norm": 0.01475780550390482,
+ "learning_rate": 0.0006,
+ "loss": 4.012645721435547,
+ "step": 2549
+ },
+ {
+ "epoch": 35.419213973799124,
+ "grad_norm": 0.014531653374433517,
+ "learning_rate": 0.0006,
+ "loss": 4.197113990783691,
+ "step": 2550
+ },
+ {
+ "epoch": 35.43318777292576,
+ "grad_norm": 0.015316493809223175,
+ "learning_rate": 0.0006,
+ "loss": 4.227970123291016,
+ "step": 2551
+ },
+ {
+ "epoch": 35.4471615720524,
+ "grad_norm": 0.014685571193695068,
+ "learning_rate": 0.0006,
+ "loss": 4.089415550231934,
+ "step": 2552
+ },
+ {
+ "epoch": 35.46113537117904,
+ "grad_norm": 0.015337035991251469,
+ "learning_rate": 0.0006,
+ "loss": 4.080583572387695,
+ "step": 2553
+ },
+ {
+ "epoch": 35.47510917030568,
+ "grad_norm": 0.016754208132624626,
+ "learning_rate": 0.0006,
+ "loss": 4.163935661315918,
+ "step": 2554
+ },
+ {
+ "epoch": 35.48908296943232,
+ "grad_norm": 0.017535798251628876,
+ "learning_rate": 0.0006,
+ "loss": 4.093027591705322,
+ "step": 2555
+ },
+ {
+ "epoch": 35.50305676855895,
+ "grad_norm": 0.017336975783109665,
+ "learning_rate": 0.0006,
+ "loss": 4.127961158752441,
+ "step": 2556
+ },
+ {
+ "epoch": 35.51703056768559,
+ "grad_norm": 0.016242152079939842,
+ "learning_rate": 0.0006,
+ "loss": 4.155593395233154,
+ "step": 2557
+ },
+ {
+ "epoch": 35.531004366812226,
+ "grad_norm": 0.015121322125196457,
+ "learning_rate": 0.0006,
+ "loss": 3.984004020690918,
+ "step": 2558
+ },
+ {
+ "epoch": 35.544978165938865,
+ "grad_norm": 0.015178884379565716,
+ "learning_rate": 0.0006,
+ "loss": 4.111564636230469,
+ "step": 2559
+ },
+ {
+ "epoch": 35.5589519650655,
+ "grad_norm": 0.014777131378650665,
+ "learning_rate": 0.0006,
+ "loss": 4.041402816772461,
+ "step": 2560
+ },
+ {
+ "epoch": 35.57292576419214,
+ "grad_norm": 0.015197676606476307,
+ "learning_rate": 0.0006,
+ "loss": 4.133706092834473,
+ "step": 2561
+ },
+ {
+ "epoch": 35.58689956331878,
+ "grad_norm": 0.015905972570180893,
+ "learning_rate": 0.0006,
+ "loss": 4.1069135665893555,
+ "step": 2562
+ },
+ {
+ "epoch": 35.60087336244541,
+ "grad_norm": 0.015942877158522606,
+ "learning_rate": 0.0006,
+ "loss": 4.0720624923706055,
+ "step": 2563
+ },
+ {
+ "epoch": 35.61484716157205,
+ "grad_norm": 0.016697583720088005,
+ "learning_rate": 0.0006,
+ "loss": 4.167448043823242,
+ "step": 2564
+ },
+ {
+ "epoch": 35.62882096069869,
+ "grad_norm": 0.015603577718138695,
+ "learning_rate": 0.0006,
+ "loss": 4.127327919006348,
+ "step": 2565
+ },
+ {
+ "epoch": 35.64279475982533,
+ "grad_norm": 0.017153460532426834,
+ "learning_rate": 0.0006,
+ "loss": 4.056543350219727,
+ "step": 2566
+ },
+ {
+ "epoch": 35.65676855895197,
+ "grad_norm": 0.016731958836317062,
+ "learning_rate": 0.0006,
+ "loss": 4.0681843757629395,
+ "step": 2567
+ },
+ {
+ "epoch": 35.670742358078606,
+ "grad_norm": 0.016167359426617622,
+ "learning_rate": 0.0006,
+ "loss": 4.174992084503174,
+ "step": 2568
+ },
+ {
+ "epoch": 35.68471615720524,
+ "grad_norm": 0.01615108922123909,
+ "learning_rate": 0.0006,
+ "loss": 4.1383161544799805,
+ "step": 2569
+ },
+ {
+ "epoch": 35.698689956331876,
+ "grad_norm": 0.015750816091895103,
+ "learning_rate": 0.0006,
+ "loss": 4.031595706939697,
+ "step": 2570
+ },
+ {
+ "epoch": 35.712663755458514,
+ "grad_norm": 0.015416189096868038,
+ "learning_rate": 0.0006,
+ "loss": 4.067498207092285,
+ "step": 2571
+ },
+ {
+ "epoch": 35.72663755458515,
+ "grad_norm": 0.01570792682468891,
+ "learning_rate": 0.0006,
+ "loss": 4.1338300704956055,
+ "step": 2572
+ },
+ {
+ "epoch": 35.74061135371179,
+ "grad_norm": 0.016367752104997635,
+ "learning_rate": 0.0006,
+ "loss": 4.152523040771484,
+ "step": 2573
+ },
+ {
+ "epoch": 35.75458515283843,
+ "grad_norm": 0.016349803656339645,
+ "learning_rate": 0.0006,
+ "loss": 4.101985454559326,
+ "step": 2574
+ },
+ {
+ "epoch": 35.76855895196506,
+ "grad_norm": 0.01663660630583763,
+ "learning_rate": 0.0006,
+ "loss": 4.190130710601807,
+ "step": 2575
+ },
+ {
+ "epoch": 35.7825327510917,
+ "grad_norm": 0.015599170699715614,
+ "learning_rate": 0.0006,
+ "loss": 4.0792436599731445,
+ "step": 2576
+ },
+ {
+ "epoch": 35.79650655021834,
+ "grad_norm": 0.015472427010536194,
+ "learning_rate": 0.0006,
+ "loss": 4.186446666717529,
+ "step": 2577
+ },
+ {
+ "epoch": 35.81048034934498,
+ "grad_norm": 0.015520439483225346,
+ "learning_rate": 0.0006,
+ "loss": 4.246635437011719,
+ "step": 2578
+ },
+ {
+ "epoch": 35.82445414847162,
+ "grad_norm": 0.016998281702399254,
+ "learning_rate": 0.0006,
+ "loss": 4.13435173034668,
+ "step": 2579
+ },
+ {
+ "epoch": 35.838427947598255,
+ "grad_norm": 0.016290361061692238,
+ "learning_rate": 0.0006,
+ "loss": 4.107428073883057,
+ "step": 2580
+ },
+ {
+ "epoch": 35.852401746724894,
+ "grad_norm": 0.0165516659617424,
+ "learning_rate": 0.0006,
+ "loss": 4.1612138748168945,
+ "step": 2581
+ },
+ {
+ "epoch": 35.866375545851525,
+ "grad_norm": 0.01736091636121273,
+ "learning_rate": 0.0006,
+ "loss": 4.1072869300842285,
+ "step": 2582
+ },
+ {
+ "epoch": 35.880349344978164,
+ "grad_norm": 0.01787031814455986,
+ "learning_rate": 0.0006,
+ "loss": 4.1076812744140625,
+ "step": 2583
+ },
+ {
+ "epoch": 35.8943231441048,
+ "grad_norm": 0.015437985770404339,
+ "learning_rate": 0.0006,
+ "loss": 4.097488880157471,
+ "step": 2584
+ },
+ {
+ "epoch": 35.90829694323144,
+ "grad_norm": 0.016306741163134575,
+ "learning_rate": 0.0006,
+ "loss": 4.105017185211182,
+ "step": 2585
+ },
+ {
+ "epoch": 35.92227074235808,
+ "grad_norm": 0.015723245218396187,
+ "learning_rate": 0.0006,
+ "loss": 4.205283164978027,
+ "step": 2586
+ },
+ {
+ "epoch": 35.93624454148472,
+ "grad_norm": 0.014626456424593925,
+ "learning_rate": 0.0006,
+ "loss": 4.086552619934082,
+ "step": 2587
+ },
+ {
+ "epoch": 35.95021834061135,
+ "grad_norm": 0.014348454773426056,
+ "learning_rate": 0.0006,
+ "loss": 4.171329975128174,
+ "step": 2588
+ },
+ {
+ "epoch": 35.96419213973799,
+ "grad_norm": 0.015598480589687824,
+ "learning_rate": 0.0006,
+ "loss": 4.140815258026123,
+ "step": 2589
+ },
+ {
+ "epoch": 35.97816593886463,
+ "grad_norm": 0.017126062884926796,
+ "learning_rate": 0.0006,
+ "loss": 4.157176971435547,
+ "step": 2590
+ },
+ {
+ "epoch": 35.992139737991266,
+ "grad_norm": 0.01860199309885502,
+ "learning_rate": 0.0006,
+ "loss": 4.252880096435547,
+ "step": 2591
+ },
+ {
+ "epoch": 36.0,
+ "grad_norm": 0.021780220791697502,
+ "learning_rate": 0.0006,
+ "loss": 4.249861717224121,
+ "step": 2592
+ },
+ {
+ "epoch": 36.0,
+ "eval_loss": 4.606466293334961,
+ "eval_runtime": 56.6217,
+ "eval_samples_per_second": 43.128,
+ "eval_steps_per_second": 1.36,
+ "step": 2592
+ },
+ {
+ "epoch": 36.01397379912664,
+ "grad_norm": 0.020892828702926636,
+ "learning_rate": 0.0006,
+ "loss": 4.102055549621582,
+ "step": 2593
+ },
+ {
+ "epoch": 36.02794759825328,
+ "grad_norm": 0.020306255668401718,
+ "learning_rate": 0.0006,
+ "loss": 4.184308052062988,
+ "step": 2594
+ },
+ {
+ "epoch": 36.041921397379916,
+ "grad_norm": 0.02025097794830799,
+ "learning_rate": 0.0006,
+ "loss": 3.9691858291625977,
+ "step": 2595
+ },
+ {
+ "epoch": 36.05589519650655,
+ "grad_norm": 0.020254382863640785,
+ "learning_rate": 0.0006,
+ "loss": 4.02957820892334,
+ "step": 2596
+ },
+ {
+ "epoch": 36.069868995633186,
+ "grad_norm": 0.019307196140289307,
+ "learning_rate": 0.0006,
+ "loss": 4.033181667327881,
+ "step": 2597
+ },
+ {
+ "epoch": 36.083842794759825,
+ "grad_norm": 0.0198476854711771,
+ "learning_rate": 0.0006,
+ "loss": 4.0816874504089355,
+ "step": 2598
+ },
+ {
+ "epoch": 36.09781659388646,
+ "grad_norm": 0.019916515797376633,
+ "learning_rate": 0.0006,
+ "loss": 4.164889812469482,
+ "step": 2599
+ },
+ {
+ "epoch": 36.1117903930131,
+ "grad_norm": 0.0189144778996706,
+ "learning_rate": 0.0006,
+ "loss": 4.038427352905273,
+ "step": 2600
+ },
+ {
+ "epoch": 36.12576419213974,
+ "grad_norm": 0.01754768192768097,
+ "learning_rate": 0.0006,
+ "loss": 4.010427474975586,
+ "step": 2601
+ },
+ {
+ "epoch": 36.13973799126637,
+ "grad_norm": 0.016862990334630013,
+ "learning_rate": 0.0006,
+ "loss": 4.035970211029053,
+ "step": 2602
+ },
+ {
+ "epoch": 36.15371179039301,
+ "grad_norm": 0.019616752862930298,
+ "learning_rate": 0.0006,
+ "loss": 4.079898834228516,
+ "step": 2603
+ },
+ {
+ "epoch": 36.16768558951965,
+ "grad_norm": 0.020131465047597885,
+ "learning_rate": 0.0006,
+ "loss": 4.144684791564941,
+ "step": 2604
+ },
+ {
+ "epoch": 36.18165938864629,
+ "grad_norm": 0.02072521299123764,
+ "learning_rate": 0.0006,
+ "loss": 4.013660430908203,
+ "step": 2605
+ },
+ {
+ "epoch": 36.19563318777293,
+ "grad_norm": 0.01924615167081356,
+ "learning_rate": 0.0006,
+ "loss": 4.071398735046387,
+ "step": 2606
+ },
+ {
+ "epoch": 36.209606986899566,
+ "grad_norm": 0.02013954520225525,
+ "learning_rate": 0.0006,
+ "loss": 4.127094268798828,
+ "step": 2607
+ },
+ {
+ "epoch": 36.223580786026204,
+ "grad_norm": 0.02042008936405182,
+ "learning_rate": 0.0006,
+ "loss": 4.154261589050293,
+ "step": 2608
+ },
+ {
+ "epoch": 36.237554585152836,
+ "grad_norm": 0.021998876705765724,
+ "learning_rate": 0.0006,
+ "loss": 4.204150199890137,
+ "step": 2609
+ },
+ {
+ "epoch": 36.251528384279474,
+ "grad_norm": 0.022946683689951897,
+ "learning_rate": 0.0006,
+ "loss": 4.013632774353027,
+ "step": 2610
+ },
+ {
+ "epoch": 36.26550218340611,
+ "grad_norm": 0.02061830274760723,
+ "learning_rate": 0.0006,
+ "loss": 4.171084403991699,
+ "step": 2611
+ },
+ {
+ "epoch": 36.27947598253275,
+ "grad_norm": 0.01883152686059475,
+ "learning_rate": 0.0006,
+ "loss": 4.149848461151123,
+ "step": 2612
+ },
+ {
+ "epoch": 36.29344978165939,
+ "grad_norm": 0.019945282489061356,
+ "learning_rate": 0.0006,
+ "loss": 4.174656867980957,
+ "step": 2613
+ },
+ {
+ "epoch": 36.30742358078603,
+ "grad_norm": 0.020751051604747772,
+ "learning_rate": 0.0006,
+ "loss": 4.080747127532959,
+ "step": 2614
+ },
+ {
+ "epoch": 36.32139737991266,
+ "grad_norm": 0.01879691146314144,
+ "learning_rate": 0.0006,
+ "loss": 4.057356357574463,
+ "step": 2615
+ },
+ {
+ "epoch": 36.3353711790393,
+ "grad_norm": 0.019037071615457535,
+ "learning_rate": 0.0006,
+ "loss": 4.204607009887695,
+ "step": 2616
+ },
+ {
+ "epoch": 36.34934497816594,
+ "grad_norm": 0.01872078701853752,
+ "learning_rate": 0.0006,
+ "loss": 4.050586700439453,
+ "step": 2617
+ },
+ {
+ "epoch": 36.36331877729258,
+ "grad_norm": 0.018092811107635498,
+ "learning_rate": 0.0006,
+ "loss": 4.102832794189453,
+ "step": 2618
+ },
+ {
+ "epoch": 36.377292576419215,
+ "grad_norm": 0.017010487616062164,
+ "learning_rate": 0.0006,
+ "loss": 4.191788673400879,
+ "step": 2619
+ },
+ {
+ "epoch": 36.391266375545854,
+ "grad_norm": 0.01691143773496151,
+ "learning_rate": 0.0006,
+ "loss": 4.160456657409668,
+ "step": 2620
+ },
+ {
+ "epoch": 36.40524017467249,
+ "grad_norm": 0.017046816647052765,
+ "learning_rate": 0.0006,
+ "loss": 4.073828220367432,
+ "step": 2621
+ },
+ {
+ "epoch": 36.419213973799124,
+ "grad_norm": 0.01707981713116169,
+ "learning_rate": 0.0006,
+ "loss": 4.0980095863342285,
+ "step": 2622
+ },
+ {
+ "epoch": 36.43318777292576,
+ "grad_norm": 0.01630452275276184,
+ "learning_rate": 0.0006,
+ "loss": 4.101321220397949,
+ "step": 2623
+ },
+ {
+ "epoch": 36.4471615720524,
+ "grad_norm": 0.01576915942132473,
+ "learning_rate": 0.0006,
+ "loss": 4.158249855041504,
+ "step": 2624
+ },
+ {
+ "epoch": 36.46113537117904,
+ "grad_norm": 0.016179712489247322,
+ "learning_rate": 0.0006,
+ "loss": 4.077852249145508,
+ "step": 2625
+ },
+ {
+ "epoch": 36.47510917030568,
+ "grad_norm": 0.01609169878065586,
+ "learning_rate": 0.0006,
+ "loss": 4.187317848205566,
+ "step": 2626
+ },
+ {
+ "epoch": 36.48908296943232,
+ "grad_norm": 0.017148254439234734,
+ "learning_rate": 0.0006,
+ "loss": 4.109071731567383,
+ "step": 2627
+ },
+ {
+ "epoch": 36.50305676855895,
+ "grad_norm": 0.015842624008655548,
+ "learning_rate": 0.0006,
+ "loss": 4.145096302032471,
+ "step": 2628
+ },
+ {
+ "epoch": 36.51703056768559,
+ "grad_norm": 0.014906194061040878,
+ "learning_rate": 0.0006,
+ "loss": 4.028440475463867,
+ "step": 2629
+ },
+ {
+ "epoch": 36.531004366812226,
+ "grad_norm": 0.01485038734972477,
+ "learning_rate": 0.0006,
+ "loss": 4.175201416015625,
+ "step": 2630
+ },
+ {
+ "epoch": 36.544978165938865,
+ "grad_norm": 0.015834150835871696,
+ "learning_rate": 0.0006,
+ "loss": 4.076434135437012,
+ "step": 2631
+ },
+ {
+ "epoch": 36.5589519650655,
+ "grad_norm": 0.015411358326673508,
+ "learning_rate": 0.0006,
+ "loss": 4.03912878036499,
+ "step": 2632
+ },
+ {
+ "epoch": 36.57292576419214,
+ "grad_norm": 0.014762846753001213,
+ "learning_rate": 0.0006,
+ "loss": 4.019827365875244,
+ "step": 2633
+ },
+ {
+ "epoch": 36.58689956331878,
+ "grad_norm": 0.015269497409462929,
+ "learning_rate": 0.0006,
+ "loss": 4.116957187652588,
+ "step": 2634
+ },
+ {
+ "epoch": 36.60087336244541,
+ "grad_norm": 0.015171617269515991,
+ "learning_rate": 0.0006,
+ "loss": 4.2069926261901855,
+ "step": 2635
+ },
+ {
+ "epoch": 36.61484716157205,
+ "grad_norm": 0.014770902693271637,
+ "learning_rate": 0.0006,
+ "loss": 4.092021942138672,
+ "step": 2636
+ },
+ {
+ "epoch": 36.62882096069869,
+ "grad_norm": 0.014957531355321407,
+ "learning_rate": 0.0006,
+ "loss": 4.16819953918457,
+ "step": 2637
+ },
+ {
+ "epoch": 36.64279475982533,
+ "grad_norm": 0.014774451963603497,
+ "learning_rate": 0.0006,
+ "loss": 4.046073913574219,
+ "step": 2638
+ },
+ {
+ "epoch": 36.65676855895197,
+ "grad_norm": 0.014320445246994495,
+ "learning_rate": 0.0006,
+ "loss": 3.977792501449585,
+ "step": 2639
+ },
+ {
+ "epoch": 36.670742358078606,
+ "grad_norm": 0.01413130946457386,
+ "learning_rate": 0.0006,
+ "loss": 4.130649566650391,
+ "step": 2640
+ },
+ {
+ "epoch": 36.68471615720524,
+ "grad_norm": 0.014336307533085346,
+ "learning_rate": 0.0006,
+ "loss": 4.161357402801514,
+ "step": 2641
+ },
+ {
+ "epoch": 36.698689956331876,
+ "grad_norm": 0.014516931027173996,
+ "learning_rate": 0.0006,
+ "loss": 4.14248514175415,
+ "step": 2642
+ },
+ {
+ "epoch": 36.712663755458514,
+ "grad_norm": 0.014377126470208168,
+ "learning_rate": 0.0006,
+ "loss": 4.062711715698242,
+ "step": 2643
+ },
+ {
+ "epoch": 36.72663755458515,
+ "grad_norm": 0.014332668855786324,
+ "learning_rate": 0.0006,
+ "loss": 4.06423807144165,
+ "step": 2644
+ },
+ {
+ "epoch": 36.74061135371179,
+ "grad_norm": 0.014030571095645428,
+ "learning_rate": 0.0006,
+ "loss": 4.178613662719727,
+ "step": 2645
+ },
+ {
+ "epoch": 36.75458515283843,
+ "grad_norm": 0.014376776292920113,
+ "learning_rate": 0.0006,
+ "loss": 4.073130130767822,
+ "step": 2646
+ },
+ {
+ "epoch": 36.76855895196506,
+ "grad_norm": 0.013827716931700706,
+ "learning_rate": 0.0006,
+ "loss": 4.043869972229004,
+ "step": 2647
+ },
+ {
+ "epoch": 36.7825327510917,
+ "grad_norm": 0.015548762865364552,
+ "learning_rate": 0.0006,
+ "loss": 4.134049415588379,
+ "step": 2648
+ },
+ {
+ "epoch": 36.79650655021834,
+ "grad_norm": 0.017041970044374466,
+ "learning_rate": 0.0006,
+ "loss": 4.205024242401123,
+ "step": 2649
+ },
+ {
+ "epoch": 36.81048034934498,
+ "grad_norm": 0.017638275399804115,
+ "learning_rate": 0.0006,
+ "loss": 4.199924468994141,
+ "step": 2650
+ },
+ {
+ "epoch": 36.82445414847162,
+ "grad_norm": 0.017771627753973007,
+ "learning_rate": 0.0006,
+ "loss": 4.15706729888916,
+ "step": 2651
+ },
+ {
+ "epoch": 36.838427947598255,
+ "grad_norm": 0.015933562070131302,
+ "learning_rate": 0.0006,
+ "loss": 4.220742225646973,
+ "step": 2652
+ },
+ {
+ "epoch": 36.852401746724894,
+ "grad_norm": 0.01668144389986992,
+ "learning_rate": 0.0006,
+ "loss": 4.161426544189453,
+ "step": 2653
+ },
+ {
+ "epoch": 36.866375545851525,
+ "grad_norm": 0.018014680594205856,
+ "learning_rate": 0.0006,
+ "loss": 4.183046817779541,
+ "step": 2654
+ },
+ {
+ "epoch": 36.880349344978164,
+ "grad_norm": 0.018808670341968536,
+ "learning_rate": 0.0006,
+ "loss": 4.105293273925781,
+ "step": 2655
+ },
+ {
+ "epoch": 36.8943231441048,
+ "grad_norm": 0.018003476783633232,
+ "learning_rate": 0.0006,
+ "loss": 4.103171348571777,
+ "step": 2656
+ },
+ {
+ "epoch": 36.90829694323144,
+ "grad_norm": 0.016865914687514305,
+ "learning_rate": 0.0006,
+ "loss": 4.125634670257568,
+ "step": 2657
+ },
+ {
+ "epoch": 36.92227074235808,
+ "grad_norm": 0.018371153622865677,
+ "learning_rate": 0.0006,
+ "loss": 4.192461013793945,
+ "step": 2658
+ },
+ {
+ "epoch": 36.93624454148472,
+ "grad_norm": 0.019045377150177956,
+ "learning_rate": 0.0006,
+ "loss": 4.198105812072754,
+ "step": 2659
+ },
+ {
+ "epoch": 36.95021834061135,
+ "grad_norm": 0.01594168320298195,
+ "learning_rate": 0.0006,
+ "loss": 4.169788360595703,
+ "step": 2660
+ },
+ {
+ "epoch": 36.96419213973799,
+ "grad_norm": 0.015645820647478104,
+ "learning_rate": 0.0006,
+ "loss": 4.047277450561523,
+ "step": 2661
+ },
+ {
+ "epoch": 36.97816593886463,
+ "grad_norm": 0.016313156113028526,
+ "learning_rate": 0.0006,
+ "loss": 4.087392807006836,
+ "step": 2662
+ },
+ {
+ "epoch": 36.992139737991266,
+ "grad_norm": 0.017050625756382942,
+ "learning_rate": 0.0006,
+ "loss": 4.0908002853393555,
+ "step": 2663
+ },
+ {
+ "epoch": 37.0,
+ "grad_norm": 0.017730968073010445,
+ "learning_rate": 0.0006,
+ "loss": 4.110957622528076,
+ "step": 2664
+ },
+ {
+ "epoch": 37.0,
+ "eval_loss": 4.563793182373047,
+ "eval_runtime": 56.5922,
+ "eval_samples_per_second": 43.151,
+ "eval_steps_per_second": 1.361,
+ "step": 2664
+ },
+ {
+ "epoch": 37.01397379912664,
+ "grad_norm": 0.015687910839915276,
+ "learning_rate": 0.0006,
+ "loss": 4.143589019775391,
+ "step": 2665
+ },
+ {
+ "epoch": 37.02794759825328,
+ "grad_norm": 0.017451366409659386,
+ "learning_rate": 0.0006,
+ "loss": 4.0326666831970215,
+ "step": 2666
+ },
+ {
+ "epoch": 37.041921397379916,
+ "grad_norm": 0.01909538544714451,
+ "learning_rate": 0.0006,
+ "loss": 3.9706249237060547,
+ "step": 2667
+ },
+ {
+ "epoch": 37.05589519650655,
+ "grad_norm": 0.01988379657268524,
+ "learning_rate": 0.0006,
+ "loss": 3.9531009197235107,
+ "step": 2668
+ },
+ {
+ "epoch": 37.069868995633186,
+ "grad_norm": 0.02035115472972393,
+ "learning_rate": 0.0006,
+ "loss": 3.962214469909668,
+ "step": 2669
+ },
+ {
+ "epoch": 37.083842794759825,
+ "grad_norm": 0.01952001079916954,
+ "learning_rate": 0.0006,
+ "loss": 4.182262897491455,
+ "step": 2670
+ },
+ {
+ "epoch": 37.09781659388646,
+ "grad_norm": 0.01854325644671917,
+ "learning_rate": 0.0006,
+ "loss": 4.127663612365723,
+ "step": 2671
+ },
+ {
+ "epoch": 37.1117903930131,
+ "grad_norm": 0.017757616937160492,
+ "learning_rate": 0.0006,
+ "loss": 4.100804805755615,
+ "step": 2672
+ },
+ {
+ "epoch": 37.12576419213974,
+ "grad_norm": 0.016712479293346405,
+ "learning_rate": 0.0006,
+ "loss": 4.026284217834473,
+ "step": 2673
+ },
+ {
+ "epoch": 37.13973799126637,
+ "grad_norm": 0.0168139711022377,
+ "learning_rate": 0.0006,
+ "loss": 4.183882713317871,
+ "step": 2674
+ },
+ {
+ "epoch": 37.15371179039301,
+ "grad_norm": 0.017060423269867897,
+ "learning_rate": 0.0006,
+ "loss": 4.131832599639893,
+ "step": 2675
+ },
+ {
+ "epoch": 37.16768558951965,
+ "grad_norm": 0.01726631447672844,
+ "learning_rate": 0.0006,
+ "loss": 3.9156057834625244,
+ "step": 2676
+ },
+ {
+ "epoch": 37.18165938864629,
+ "grad_norm": 0.016348622739315033,
+ "learning_rate": 0.0006,
+ "loss": 4.1731367111206055,
+ "step": 2677
+ },
+ {
+ "epoch": 37.19563318777293,
+ "grad_norm": 0.016640059649944305,
+ "learning_rate": 0.0006,
+ "loss": 4.10997200012207,
+ "step": 2678
+ },
+ {
+ "epoch": 37.209606986899566,
+ "grad_norm": 0.01601342298090458,
+ "learning_rate": 0.0006,
+ "loss": 4.081145286560059,
+ "step": 2679
+ },
+ {
+ "epoch": 37.223580786026204,
+ "grad_norm": 0.01735631190240383,
+ "learning_rate": 0.0006,
+ "loss": 4.202108860015869,
+ "step": 2680
+ },
+ {
+ "epoch": 37.237554585152836,
+ "grad_norm": 0.019164685159921646,
+ "learning_rate": 0.0006,
+ "loss": 4.104109764099121,
+ "step": 2681
+ },
+ {
+ "epoch": 37.251528384279474,
+ "grad_norm": 0.01881759613752365,
+ "learning_rate": 0.0006,
+ "loss": 4.145532131195068,
+ "step": 2682
+ },
+ {
+ "epoch": 37.26550218340611,
+ "grad_norm": 0.017656555399298668,
+ "learning_rate": 0.0006,
+ "loss": 4.012782096862793,
+ "step": 2683
+ },
+ {
+ "epoch": 37.27947598253275,
+ "grad_norm": 0.01624823734164238,
+ "learning_rate": 0.0006,
+ "loss": 4.001727104187012,
+ "step": 2684
+ },
+ {
+ "epoch": 37.29344978165939,
+ "grad_norm": 0.016973139718174934,
+ "learning_rate": 0.0006,
+ "loss": 4.095399856567383,
+ "step": 2685
+ },
+ {
+ "epoch": 37.30742358078603,
+ "grad_norm": 0.015503671951591969,
+ "learning_rate": 0.0006,
+ "loss": 4.034431457519531,
+ "step": 2686
+ },
+ {
+ "epoch": 37.32139737991266,
+ "grad_norm": 0.015370250679552555,
+ "learning_rate": 0.0006,
+ "loss": 4.075836181640625,
+ "step": 2687
+ },
+ {
+ "epoch": 37.3353711790393,
+ "grad_norm": 0.01564021222293377,
+ "learning_rate": 0.0006,
+ "loss": 4.125202178955078,
+ "step": 2688
+ },
+ {
+ "epoch": 37.34934497816594,
+ "grad_norm": 0.016717975959181786,
+ "learning_rate": 0.0006,
+ "loss": 4.153017044067383,
+ "step": 2689
+ },
+ {
+ "epoch": 37.36331877729258,
+ "grad_norm": 0.015263617038726807,
+ "learning_rate": 0.0006,
+ "loss": 4.021814823150635,
+ "step": 2690
+ },
+ {
+ "epoch": 37.377292576419215,
+ "grad_norm": 0.016497142612934113,
+ "learning_rate": 0.0006,
+ "loss": 4.069095611572266,
+ "step": 2691
+ },
+ {
+ "epoch": 37.391266375545854,
+ "grad_norm": 0.01669355481863022,
+ "learning_rate": 0.0006,
+ "loss": 4.174587726593018,
+ "step": 2692
+ },
+ {
+ "epoch": 37.40524017467249,
+ "grad_norm": 0.019695773720741272,
+ "learning_rate": 0.0006,
+ "loss": 4.0073041915893555,
+ "step": 2693
+ },
+ {
+ "epoch": 37.419213973799124,
+ "grad_norm": 0.023720644414424896,
+ "learning_rate": 0.0006,
+ "loss": 4.111948013305664,
+ "step": 2694
+ },
+ {
+ "epoch": 37.43318777292576,
+ "grad_norm": 0.02399350143969059,
+ "learning_rate": 0.0006,
+ "loss": 4.081164360046387,
+ "step": 2695
+ },
+ {
+ "epoch": 37.4471615720524,
+ "grad_norm": 0.019630271941423416,
+ "learning_rate": 0.0006,
+ "loss": 4.065731048583984,
+ "step": 2696
+ },
+ {
+ "epoch": 37.46113537117904,
+ "grad_norm": 0.019679008051753044,
+ "learning_rate": 0.0006,
+ "loss": 3.993852376937866,
+ "step": 2697
+ },
+ {
+ "epoch": 37.47510917030568,
+ "grad_norm": 0.020597299560904503,
+ "learning_rate": 0.0006,
+ "loss": 4.275555610656738,
+ "step": 2698
+ },
+ {
+ "epoch": 37.48908296943232,
+ "grad_norm": 0.01983889564871788,
+ "learning_rate": 0.0006,
+ "loss": 4.254708290100098,
+ "step": 2699
+ },
+ {
+ "epoch": 37.50305676855895,
+ "grad_norm": 0.019150668755173683,
+ "learning_rate": 0.0006,
+ "loss": 4.097885608673096,
+ "step": 2700
+ },
+ {
+ "epoch": 37.51703056768559,
+ "grad_norm": 0.018847715109586716,
+ "learning_rate": 0.0006,
+ "loss": 4.054081439971924,
+ "step": 2701
+ },
+ {
+ "epoch": 37.531004366812226,
+ "grad_norm": 0.019607288762927055,
+ "learning_rate": 0.0006,
+ "loss": 3.9891440868377686,
+ "step": 2702
+ },
+ {
+ "epoch": 37.544978165938865,
+ "grad_norm": 0.018863150849938393,
+ "learning_rate": 0.0006,
+ "loss": 4.084877014160156,
+ "step": 2703
+ },
+ {
+ "epoch": 37.5589519650655,
+ "grad_norm": 0.017340239137411118,
+ "learning_rate": 0.0006,
+ "loss": 4.113290786743164,
+ "step": 2704
+ },
+ {
+ "epoch": 37.57292576419214,
+ "grad_norm": 0.015628693625330925,
+ "learning_rate": 0.0006,
+ "loss": 4.1483001708984375,
+ "step": 2705
+ },
+ {
+ "epoch": 37.58689956331878,
+ "grad_norm": 0.016493283212184906,
+ "learning_rate": 0.0006,
+ "loss": 4.1040143966674805,
+ "step": 2706
+ },
+ {
+ "epoch": 37.60087336244541,
+ "grad_norm": 0.016732443124055862,
+ "learning_rate": 0.0006,
+ "loss": 4.07205057144165,
+ "step": 2707
+ },
+ {
+ "epoch": 37.61484716157205,
+ "grad_norm": 0.017176156863570213,
+ "learning_rate": 0.0006,
+ "loss": 4.1279754638671875,
+ "step": 2708
+ },
+ {
+ "epoch": 37.62882096069869,
+ "grad_norm": 0.015730608254671097,
+ "learning_rate": 0.0006,
+ "loss": 4.086658000946045,
+ "step": 2709
+ },
+ {
+ "epoch": 37.64279475982533,
+ "grad_norm": 0.016111090779304504,
+ "learning_rate": 0.0006,
+ "loss": 4.064335823059082,
+ "step": 2710
+ },
+ {
+ "epoch": 37.65676855895197,
+ "grad_norm": 0.01769379712641239,
+ "learning_rate": 0.0006,
+ "loss": 4.116047382354736,
+ "step": 2711
+ },
+ {
+ "epoch": 37.670742358078606,
+ "grad_norm": 0.01683102548122406,
+ "learning_rate": 0.0006,
+ "loss": 4.095986843109131,
+ "step": 2712
+ },
+ {
+ "epoch": 37.68471615720524,
+ "grad_norm": 0.015726450830698013,
+ "learning_rate": 0.0006,
+ "loss": 4.10775089263916,
+ "step": 2713
+ },
+ {
+ "epoch": 37.698689956331876,
+ "grad_norm": 0.014760235324501991,
+ "learning_rate": 0.0006,
+ "loss": 4.1234893798828125,
+ "step": 2714
+ },
+ {
+ "epoch": 37.712663755458514,
+ "grad_norm": 0.014933438040316105,
+ "learning_rate": 0.0006,
+ "loss": 4.079200744628906,
+ "step": 2715
+ },
+ {
+ "epoch": 37.72663755458515,
+ "grad_norm": 0.015723997727036476,
+ "learning_rate": 0.0006,
+ "loss": 4.07865047454834,
+ "step": 2716
+ },
+ {
+ "epoch": 37.74061135371179,
+ "grad_norm": 0.014539001509547234,
+ "learning_rate": 0.0006,
+ "loss": 4.036242485046387,
+ "step": 2717
+ },
+ {
+ "epoch": 37.75458515283843,
+ "grad_norm": 0.014844979159533978,
+ "learning_rate": 0.0006,
+ "loss": 4.1051483154296875,
+ "step": 2718
+ },
+ {
+ "epoch": 37.76855895196506,
+ "grad_norm": 0.015939844772219658,
+ "learning_rate": 0.0006,
+ "loss": 4.017253875732422,
+ "step": 2719
+ },
+ {
+ "epoch": 37.7825327510917,
+ "grad_norm": 0.014771555550396442,
+ "learning_rate": 0.0006,
+ "loss": 4.0694661140441895,
+ "step": 2720
+ },
+ {
+ "epoch": 37.79650655021834,
+ "grad_norm": 0.015777921304106712,
+ "learning_rate": 0.0006,
+ "loss": 3.947587490081787,
+ "step": 2721
+ },
+ {
+ "epoch": 37.81048034934498,
+ "grad_norm": 0.017451291903853416,
+ "learning_rate": 0.0006,
+ "loss": 4.093625068664551,
+ "step": 2722
+ },
+ {
+ "epoch": 37.82445414847162,
+ "grad_norm": 0.016829246655106544,
+ "learning_rate": 0.0006,
+ "loss": 4.09030818939209,
+ "step": 2723
+ },
+ {
+ "epoch": 37.838427947598255,
+ "grad_norm": 0.016984397545456886,
+ "learning_rate": 0.0006,
+ "loss": 4.140626907348633,
+ "step": 2724
+ },
+ {
+ "epoch": 37.852401746724894,
+ "grad_norm": 0.01732603646814823,
+ "learning_rate": 0.0006,
+ "loss": 4.128276824951172,
+ "step": 2725
+ },
+ {
+ "epoch": 37.866375545851525,
+ "grad_norm": 0.01744619384407997,
+ "learning_rate": 0.0006,
+ "loss": 4.103130340576172,
+ "step": 2726
+ },
+ {
+ "epoch": 37.880349344978164,
+ "grad_norm": 0.016331320628523827,
+ "learning_rate": 0.0006,
+ "loss": 4.037951469421387,
+ "step": 2727
+ },
+ {
+ "epoch": 37.8943231441048,
+ "grad_norm": 0.015919430181384087,
+ "learning_rate": 0.0006,
+ "loss": 4.121513366699219,
+ "step": 2728
+ },
+ {
+ "epoch": 37.90829694323144,
+ "grad_norm": 0.016153695061802864,
+ "learning_rate": 0.0006,
+ "loss": 4.099384307861328,
+ "step": 2729
+ },
+ {
+ "epoch": 37.92227074235808,
+ "grad_norm": 0.01628800667822361,
+ "learning_rate": 0.0006,
+ "loss": 4.092325687408447,
+ "step": 2730
+ },
+ {
+ "epoch": 37.93624454148472,
+ "grad_norm": 0.017285684123635292,
+ "learning_rate": 0.0006,
+ "loss": 3.9828906059265137,
+ "step": 2731
+ },
+ {
+ "epoch": 37.95021834061135,
+ "grad_norm": 0.016047554090619087,
+ "learning_rate": 0.0006,
+ "loss": 4.100419044494629,
+ "step": 2732
+ },
+ {
+ "epoch": 37.96419213973799,
+ "grad_norm": 0.01562986709177494,
+ "learning_rate": 0.0006,
+ "loss": 4.065832614898682,
+ "step": 2733
+ },
+ {
+ "epoch": 37.97816593886463,
+ "grad_norm": 0.01677163504064083,
+ "learning_rate": 0.0006,
+ "loss": 4.077518463134766,
+ "step": 2734
+ },
+ {
+ "epoch": 37.992139737991266,
+ "grad_norm": 0.01755492016673088,
+ "learning_rate": 0.0006,
+ "loss": 4.1076154708862305,
+ "step": 2735
+ },
+ {
+ "epoch": 38.0,
+ "grad_norm": 0.01929762214422226,
+ "learning_rate": 0.0006,
+ "loss": 4.09603214263916,
+ "step": 2736
+ },
+ {
+ "epoch": 38.0,
+ "eval_loss": 4.582442283630371,
+ "eval_runtime": 62.8649,
+ "eval_samples_per_second": 38.845,
+ "eval_steps_per_second": 1.225,
+ "step": 2736
+ },
+ {
+ "epoch": 38.01397379912664,
+ "grad_norm": 0.019129136577248573,
+ "learning_rate": 0.0006,
+ "loss": 4.095005989074707,
+ "step": 2737
+ },
+ {
+ "epoch": 38.02794759825328,
+ "grad_norm": 0.019167408347129822,
+ "learning_rate": 0.0006,
+ "loss": 4.054426670074463,
+ "step": 2738
+ },
+ {
+ "epoch": 38.041921397379916,
+ "grad_norm": 0.01789460889995098,
+ "learning_rate": 0.0006,
+ "loss": 4.0434417724609375,
+ "step": 2739
+ },
+ {
+ "epoch": 38.05589519650655,
+ "grad_norm": 0.015947027131915092,
+ "learning_rate": 0.0006,
+ "loss": 4.074516296386719,
+ "step": 2740
+ },
+ {
+ "epoch": 38.069868995633186,
+ "grad_norm": 0.016667252406477928,
+ "learning_rate": 0.0006,
+ "loss": 4.108545303344727,
+ "step": 2741
+ },
+ {
+ "epoch": 38.083842794759825,
+ "grad_norm": 0.016014399006962776,
+ "learning_rate": 0.0006,
+ "loss": 4.026925086975098,
+ "step": 2742
+ },
+ {
+ "epoch": 38.09781659388646,
+ "grad_norm": 0.015397695824503899,
+ "learning_rate": 0.0006,
+ "loss": 4.163970470428467,
+ "step": 2743
+ },
+ {
+ "epoch": 38.1117903930131,
+ "grad_norm": 0.01563289947807789,
+ "learning_rate": 0.0006,
+ "loss": 4.116578578948975,
+ "step": 2744
+ },
+ {
+ "epoch": 38.12576419213974,
+ "grad_norm": 0.015512318350374699,
+ "learning_rate": 0.0006,
+ "loss": 4.096311569213867,
+ "step": 2745
+ },
+ {
+ "epoch": 38.13973799126637,
+ "grad_norm": 0.015426642261445522,
+ "learning_rate": 0.0006,
+ "loss": 4.017065048217773,
+ "step": 2746
+ },
+ {
+ "epoch": 38.15371179039301,
+ "grad_norm": 0.015115435235202312,
+ "learning_rate": 0.0006,
+ "loss": 4.005434989929199,
+ "step": 2747
+ },
+ {
+ "epoch": 38.16768558951965,
+ "grad_norm": 0.015235894359648228,
+ "learning_rate": 0.0006,
+ "loss": 3.982743263244629,
+ "step": 2748
+ },
+ {
+ "epoch": 38.18165938864629,
+ "grad_norm": 0.016095101833343506,
+ "learning_rate": 0.0006,
+ "loss": 4.0849385261535645,
+ "step": 2749
+ },
+ {
+ "epoch": 38.19563318777293,
+ "grad_norm": 0.015201197937130928,
+ "learning_rate": 0.0006,
+ "loss": 4.103863716125488,
+ "step": 2750
+ },
+ {
+ "epoch": 38.209606986899566,
+ "grad_norm": 0.015687478706240654,
+ "learning_rate": 0.0006,
+ "loss": 4.12074089050293,
+ "step": 2751
+ },
+ {
+ "epoch": 38.223580786026204,
+ "grad_norm": 0.015201589092612267,
+ "learning_rate": 0.0006,
+ "loss": 4.056872367858887,
+ "step": 2752
+ },
+ {
+ "epoch": 38.237554585152836,
+ "grad_norm": 0.015154894441366196,
+ "learning_rate": 0.0006,
+ "loss": 4.089670181274414,
+ "step": 2753
+ },
+ {
+ "epoch": 38.251528384279474,
+ "grad_norm": 0.01586608588695526,
+ "learning_rate": 0.0006,
+ "loss": 4.155317306518555,
+ "step": 2754
+ },
+ {
+ "epoch": 38.26550218340611,
+ "grad_norm": 0.01609654724597931,
+ "learning_rate": 0.0006,
+ "loss": 3.995800495147705,
+ "step": 2755
+ },
+ {
+ "epoch": 38.27947598253275,
+ "grad_norm": 0.015495185740292072,
+ "learning_rate": 0.0006,
+ "loss": 3.995769739151001,
+ "step": 2756
+ },
+ {
+ "epoch": 38.29344978165939,
+ "grad_norm": 0.017042381688952446,
+ "learning_rate": 0.0006,
+ "loss": 4.132832050323486,
+ "step": 2757
+ },
+ {
+ "epoch": 38.30742358078603,
+ "grad_norm": 0.018706614151597023,
+ "learning_rate": 0.0006,
+ "loss": 4.038820743560791,
+ "step": 2758
+ },
+ {
+ "epoch": 38.32139737991266,
+ "grad_norm": 0.018316026777029037,
+ "learning_rate": 0.0006,
+ "loss": 3.972052574157715,
+ "step": 2759
+ },
+ {
+ "epoch": 38.3353711790393,
+ "grad_norm": 0.018375389277935028,
+ "learning_rate": 0.0006,
+ "loss": 4.152097702026367,
+ "step": 2760
+ },
+ {
+ "epoch": 38.34934497816594,
+ "grad_norm": 0.016879886388778687,
+ "learning_rate": 0.0006,
+ "loss": 4.020869731903076,
+ "step": 2761
+ },
+ {
+ "epoch": 38.36331877729258,
+ "grad_norm": 0.015728183090686798,
+ "learning_rate": 0.0006,
+ "loss": 4.008016586303711,
+ "step": 2762
+ },
+ {
+ "epoch": 38.377292576419215,
+ "grad_norm": 0.016147758811712265,
+ "learning_rate": 0.0006,
+ "loss": 4.035854339599609,
+ "step": 2763
+ },
+ {
+ "epoch": 38.391266375545854,
+ "grad_norm": 0.017774797976017,
+ "learning_rate": 0.0006,
+ "loss": 4.122506141662598,
+ "step": 2764
+ },
+ {
+ "epoch": 38.40524017467249,
+ "grad_norm": 0.016590069979429245,
+ "learning_rate": 0.0006,
+ "loss": 4.082847595214844,
+ "step": 2765
+ },
+ {
+ "epoch": 38.419213973799124,
+ "grad_norm": 0.016544528305530548,
+ "learning_rate": 0.0006,
+ "loss": 4.184270858764648,
+ "step": 2766
+ },
+ {
+ "epoch": 38.43318777292576,
+ "grad_norm": 0.017972925677895546,
+ "learning_rate": 0.0006,
+ "loss": 3.987558603286743,
+ "step": 2767
+ },
+ {
+ "epoch": 38.4471615720524,
+ "grad_norm": 0.018562227487564087,
+ "learning_rate": 0.0006,
+ "loss": 4.1193461418151855,
+ "step": 2768
+ },
+ {
+ "epoch": 38.46113537117904,
+ "grad_norm": 0.018896685913205147,
+ "learning_rate": 0.0006,
+ "loss": 4.057808876037598,
+ "step": 2769
+ },
+ {
+ "epoch": 38.47510917030568,
+ "grad_norm": 0.016793809831142426,
+ "learning_rate": 0.0006,
+ "loss": 4.069523334503174,
+ "step": 2770
+ },
+ {
+ "epoch": 38.48908296943232,
+ "grad_norm": 0.01698676310479641,
+ "learning_rate": 0.0006,
+ "loss": 3.997624158859253,
+ "step": 2771
+ },
+ {
+ "epoch": 38.50305676855895,
+ "grad_norm": 0.016784202307462692,
+ "learning_rate": 0.0006,
+ "loss": 4.174795627593994,
+ "step": 2772
+ },
+ {
+ "epoch": 38.51703056768559,
+ "grad_norm": 0.016785014420747757,
+ "learning_rate": 0.0006,
+ "loss": 3.999321222305298,
+ "step": 2773
+ },
+ {
+ "epoch": 38.531004366812226,
+ "grad_norm": 0.018615184351801872,
+ "learning_rate": 0.0006,
+ "loss": 4.191237926483154,
+ "step": 2774
+ },
+ {
+ "epoch": 38.544978165938865,
+ "grad_norm": 0.019090790301561356,
+ "learning_rate": 0.0006,
+ "loss": 4.106966018676758,
+ "step": 2775
+ },
+ {
+ "epoch": 38.5589519650655,
+ "grad_norm": 0.018298320472240448,
+ "learning_rate": 0.0006,
+ "loss": 4.063363552093506,
+ "step": 2776
+ },
+ {
+ "epoch": 38.57292576419214,
+ "grad_norm": 0.018541090190410614,
+ "learning_rate": 0.0006,
+ "loss": 4.101193428039551,
+ "step": 2777
+ },
+ {
+ "epoch": 38.58689956331878,
+ "grad_norm": 0.01849970407783985,
+ "learning_rate": 0.0006,
+ "loss": 4.071335792541504,
+ "step": 2778
+ },
+ {
+ "epoch": 38.60087336244541,
+ "grad_norm": 0.018309077247977257,
+ "learning_rate": 0.0006,
+ "loss": 4.245305061340332,
+ "step": 2779
+ },
+ {
+ "epoch": 38.61484716157205,
+ "grad_norm": 0.017681840807199478,
+ "learning_rate": 0.0006,
+ "loss": 4.081173419952393,
+ "step": 2780
+ },
+ {
+ "epoch": 38.62882096069869,
+ "grad_norm": 0.016286224126815796,
+ "learning_rate": 0.0006,
+ "loss": 4.024351596832275,
+ "step": 2781
+ },
+ {
+ "epoch": 38.64279475982533,
+ "grad_norm": 0.015837235376238823,
+ "learning_rate": 0.0006,
+ "loss": 4.028205394744873,
+ "step": 2782
+ },
+ {
+ "epoch": 38.65676855895197,
+ "grad_norm": 0.016062596812844276,
+ "learning_rate": 0.0006,
+ "loss": 4.131122589111328,
+ "step": 2783
+ },
+ {
+ "epoch": 38.670742358078606,
+ "grad_norm": 0.01711239665746689,
+ "learning_rate": 0.0006,
+ "loss": 4.124416351318359,
+ "step": 2784
+ },
+ {
+ "epoch": 38.68471615720524,
+ "grad_norm": 0.015304913744330406,
+ "learning_rate": 0.0006,
+ "loss": 4.056122779846191,
+ "step": 2785
+ },
+ {
+ "epoch": 38.698689956331876,
+ "grad_norm": 0.015971887856721878,
+ "learning_rate": 0.0006,
+ "loss": 4.132555961608887,
+ "step": 2786
+ },
+ {
+ "epoch": 38.712663755458514,
+ "grad_norm": 0.015787221491336823,
+ "learning_rate": 0.0006,
+ "loss": 4.051596164703369,
+ "step": 2787
+ },
+ {
+ "epoch": 38.72663755458515,
+ "grad_norm": 0.01605444774031639,
+ "learning_rate": 0.0006,
+ "loss": 4.191808700561523,
+ "step": 2788
+ },
+ {
+ "epoch": 38.74061135371179,
+ "grad_norm": 0.016383009031414986,
+ "learning_rate": 0.0006,
+ "loss": 4.0648417472839355,
+ "step": 2789
+ },
+ {
+ "epoch": 38.75458515283843,
+ "grad_norm": 0.016135554760694504,
+ "learning_rate": 0.0006,
+ "loss": 3.988563060760498,
+ "step": 2790
+ },
+ {
+ "epoch": 38.76855895196506,
+ "grad_norm": 0.016634158790111542,
+ "learning_rate": 0.0006,
+ "loss": 4.133493423461914,
+ "step": 2791
+ },
+ {
+ "epoch": 38.7825327510917,
+ "grad_norm": 0.01688210666179657,
+ "learning_rate": 0.0006,
+ "loss": 4.091277122497559,
+ "step": 2792
+ },
+ {
+ "epoch": 38.79650655021834,
+ "grad_norm": 0.016102122142910957,
+ "learning_rate": 0.0006,
+ "loss": 4.051219940185547,
+ "step": 2793
+ },
+ {
+ "epoch": 38.81048034934498,
+ "grad_norm": 0.016585038974881172,
+ "learning_rate": 0.0006,
+ "loss": 4.058587074279785,
+ "step": 2794
+ },
+ {
+ "epoch": 38.82445414847162,
+ "grad_norm": 0.017700951546430588,
+ "learning_rate": 0.0006,
+ "loss": 4.016030311584473,
+ "step": 2795
+ },
+ {
+ "epoch": 38.838427947598255,
+ "grad_norm": 0.017922203987836838,
+ "learning_rate": 0.0006,
+ "loss": 4.079977989196777,
+ "step": 2796
+ },
+ {
+ "epoch": 38.852401746724894,
+ "grad_norm": 0.017299890518188477,
+ "learning_rate": 0.0006,
+ "loss": 4.062984943389893,
+ "step": 2797
+ },
+ {
+ "epoch": 38.866375545851525,
+ "grad_norm": 0.018648581579327583,
+ "learning_rate": 0.0006,
+ "loss": 4.065717697143555,
+ "step": 2798
+ },
+ {
+ "epoch": 38.880349344978164,
+ "grad_norm": 0.019057920202612877,
+ "learning_rate": 0.0006,
+ "loss": 4.227892875671387,
+ "step": 2799
+ },
+ {
+ "epoch": 38.8943231441048,
+ "grad_norm": 0.018824422731995583,
+ "learning_rate": 0.0006,
+ "loss": 4.116058349609375,
+ "step": 2800
+ },
+ {
+ "epoch": 38.90829694323144,
+ "grad_norm": 0.019084978848695755,
+ "learning_rate": 0.0006,
+ "loss": 4.061677932739258,
+ "step": 2801
+ },
+ {
+ "epoch": 38.92227074235808,
+ "grad_norm": 0.017842981964349747,
+ "learning_rate": 0.0006,
+ "loss": 4.083813667297363,
+ "step": 2802
+ },
+ {
+ "epoch": 38.93624454148472,
+ "grad_norm": 0.01880858838558197,
+ "learning_rate": 0.0006,
+ "loss": 4.152561187744141,
+ "step": 2803
+ },
+ {
+ "epoch": 38.95021834061135,
+ "grad_norm": 0.017400087788701057,
+ "learning_rate": 0.0006,
+ "loss": 4.154635429382324,
+ "step": 2804
+ },
+ {
+ "epoch": 38.96419213973799,
+ "grad_norm": 0.017682507634162903,
+ "learning_rate": 0.0006,
+ "loss": 4.151128768920898,
+ "step": 2805
+ },
+ {
+ "epoch": 38.97816593886463,
+ "grad_norm": 0.015803052112460136,
+ "learning_rate": 0.0006,
+ "loss": 4.10498571395874,
+ "step": 2806
+ },
+ {
+ "epoch": 38.992139737991266,
+ "grad_norm": 0.01582511141896248,
+ "learning_rate": 0.0006,
+ "loss": 4.140159606933594,
+ "step": 2807
+ },
+ {
+ "epoch": 39.0,
+ "grad_norm": 0.01925666816532612,
+ "learning_rate": 0.0006,
+ "loss": 4.126394271850586,
+ "step": 2808
+ },
+ {
+ "epoch": 39.0,
+ "eval_loss": 4.615617275238037,
+ "eval_runtime": 56.8895,
+ "eval_samples_per_second": 42.925,
+ "eval_steps_per_second": 1.354,
+ "step": 2808
+ },
+ {
+ "epoch": 39.01397379912664,
+ "grad_norm": 0.01677853614091873,
+ "learning_rate": 0.0006,
+ "loss": 4.102395057678223,
+ "step": 2809
+ },
+ {
+ "epoch": 39.02794759825328,
+ "grad_norm": 0.017901137471199036,
+ "learning_rate": 0.0006,
+ "loss": 3.976609230041504,
+ "step": 2810
+ },
+ {
+ "epoch": 39.041921397379916,
+ "grad_norm": 0.017978668212890625,
+ "learning_rate": 0.0006,
+ "loss": 3.944416046142578,
+ "step": 2811
+ },
+ {
+ "epoch": 39.05589519650655,
+ "grad_norm": 0.019967492669820786,
+ "learning_rate": 0.0006,
+ "loss": 3.992234468460083,
+ "step": 2812
+ },
+ {
+ "epoch": 39.069868995633186,
+ "grad_norm": 0.018937544897198677,
+ "learning_rate": 0.0006,
+ "loss": 4.028573036193848,
+ "step": 2813
+ },
+ {
+ "epoch": 39.083842794759825,
+ "grad_norm": 0.016852952539920807,
+ "learning_rate": 0.0006,
+ "loss": 4.062442779541016,
+ "step": 2814
+ },
+ {
+ "epoch": 39.09781659388646,
+ "grad_norm": 0.016643647104501724,
+ "learning_rate": 0.0006,
+ "loss": 4.08871603012085,
+ "step": 2815
+ },
+ {
+ "epoch": 39.1117903930131,
+ "grad_norm": 0.016872866079211235,
+ "learning_rate": 0.0006,
+ "loss": 3.9514245986938477,
+ "step": 2816
+ },
+ {
+ "epoch": 39.12576419213974,
+ "grad_norm": 0.018045669421553612,
+ "learning_rate": 0.0006,
+ "loss": 4.038842678070068,
+ "step": 2817
+ },
+ {
+ "epoch": 39.13973799126637,
+ "grad_norm": 0.017853403463959694,
+ "learning_rate": 0.0006,
+ "loss": 3.93709135055542,
+ "step": 2818
+ },
+ {
+ "epoch": 39.15371179039301,
+ "grad_norm": 0.017507443204522133,
+ "learning_rate": 0.0006,
+ "loss": 3.98665714263916,
+ "step": 2819
+ },
+ {
+ "epoch": 39.16768558951965,
+ "grad_norm": 0.01766853965818882,
+ "learning_rate": 0.0006,
+ "loss": 3.937126636505127,
+ "step": 2820
+ },
+ {
+ "epoch": 39.18165938864629,
+ "grad_norm": 0.016660021618008614,
+ "learning_rate": 0.0006,
+ "loss": 3.8397014141082764,
+ "step": 2821
+ },
+ {
+ "epoch": 39.19563318777293,
+ "grad_norm": 0.017719309777021408,
+ "learning_rate": 0.0006,
+ "loss": 4.003579616546631,
+ "step": 2822
+ },
+ {
+ "epoch": 39.209606986899566,
+ "grad_norm": 0.018234414979815483,
+ "learning_rate": 0.0006,
+ "loss": 4.053859710693359,
+ "step": 2823
+ },
+ {
+ "epoch": 39.223580786026204,
+ "grad_norm": 0.018777024000883102,
+ "learning_rate": 0.0006,
+ "loss": 4.056182861328125,
+ "step": 2824
+ },
+ {
+ "epoch": 39.237554585152836,
+ "grad_norm": 0.017443502321839333,
+ "learning_rate": 0.0006,
+ "loss": 4.025068283081055,
+ "step": 2825
+ },
+ {
+ "epoch": 39.251528384279474,
+ "grad_norm": 0.016500331461429596,
+ "learning_rate": 0.0006,
+ "loss": 4.101631164550781,
+ "step": 2826
+ },
+ {
+ "epoch": 39.26550218340611,
+ "grad_norm": 0.016101544722914696,
+ "learning_rate": 0.0006,
+ "loss": 4.027082443237305,
+ "step": 2827
+ },
+ {
+ "epoch": 39.27947598253275,
+ "grad_norm": 0.015471171587705612,
+ "learning_rate": 0.0006,
+ "loss": 4.022909641265869,
+ "step": 2828
+ },
+ {
+ "epoch": 39.29344978165939,
+ "grad_norm": 0.015779748558998108,
+ "learning_rate": 0.0006,
+ "loss": 4.18796968460083,
+ "step": 2829
+ },
+ {
+ "epoch": 39.30742358078603,
+ "grad_norm": 0.015522679314017296,
+ "learning_rate": 0.0006,
+ "loss": 4.005777359008789,
+ "step": 2830
+ },
+ {
+ "epoch": 39.32139737991266,
+ "grad_norm": 0.015226504765450954,
+ "learning_rate": 0.0006,
+ "loss": 4.108416557312012,
+ "step": 2831
+ },
+ {
+ "epoch": 39.3353711790393,
+ "grad_norm": 0.015095778740942478,
+ "learning_rate": 0.0006,
+ "loss": 4.08029842376709,
+ "step": 2832
+ },
+ {
+ "epoch": 39.34934497816594,
+ "grad_norm": 0.015601031482219696,
+ "learning_rate": 0.0006,
+ "loss": 3.989492416381836,
+ "step": 2833
+ },
+ {
+ "epoch": 39.36331877729258,
+ "grad_norm": 0.014825052581727505,
+ "learning_rate": 0.0006,
+ "loss": 3.998520612716675,
+ "step": 2834
+ },
+ {
+ "epoch": 39.377292576419215,
+ "grad_norm": 0.014821100980043411,
+ "learning_rate": 0.0006,
+ "loss": 4.001777172088623,
+ "step": 2835
+ },
+ {
+ "epoch": 39.391266375545854,
+ "grad_norm": 0.015333754941821098,
+ "learning_rate": 0.0006,
+ "loss": 4.11985969543457,
+ "step": 2836
+ },
+ {
+ "epoch": 39.40524017467249,
+ "grad_norm": 0.015940699726343155,
+ "learning_rate": 0.0006,
+ "loss": 4.025373935699463,
+ "step": 2837
+ },
+ {
+ "epoch": 39.419213973799124,
+ "grad_norm": 0.015096287243068218,
+ "learning_rate": 0.0006,
+ "loss": 4.068551063537598,
+ "step": 2838
+ },
+ {
+ "epoch": 39.43318777292576,
+ "grad_norm": 0.01533548254519701,
+ "learning_rate": 0.0006,
+ "loss": 3.9800801277160645,
+ "step": 2839
+ },
+ {
+ "epoch": 39.4471615720524,
+ "grad_norm": 0.014648732729256153,
+ "learning_rate": 0.0006,
+ "loss": 3.9670190811157227,
+ "step": 2840
+ },
+ {
+ "epoch": 39.46113537117904,
+ "grad_norm": 0.015759671106934547,
+ "learning_rate": 0.0006,
+ "loss": 4.050443649291992,
+ "step": 2841
+ },
+ {
+ "epoch": 39.47510917030568,
+ "grad_norm": 0.016054073348641396,
+ "learning_rate": 0.0006,
+ "loss": 4.22734260559082,
+ "step": 2842
+ },
+ {
+ "epoch": 39.48908296943232,
+ "grad_norm": 0.0160264540463686,
+ "learning_rate": 0.0006,
+ "loss": 4.116174221038818,
+ "step": 2843
+ },
+ {
+ "epoch": 39.50305676855895,
+ "grad_norm": 0.015426610596477985,
+ "learning_rate": 0.0006,
+ "loss": 4.000178337097168,
+ "step": 2844
+ },
+ {
+ "epoch": 39.51703056768559,
+ "grad_norm": 0.01520642638206482,
+ "learning_rate": 0.0006,
+ "loss": 4.218410968780518,
+ "step": 2845
+ },
+ {
+ "epoch": 39.531004366812226,
+ "grad_norm": 0.015305949375033379,
+ "learning_rate": 0.0006,
+ "loss": 4.025891304016113,
+ "step": 2846
+ },
+ {
+ "epoch": 39.544978165938865,
+ "grad_norm": 0.015461564064025879,
+ "learning_rate": 0.0006,
+ "loss": 4.100950717926025,
+ "step": 2847
+ },
+ {
+ "epoch": 39.5589519650655,
+ "grad_norm": 0.01712625101208687,
+ "learning_rate": 0.0006,
+ "loss": 4.0620856285095215,
+ "step": 2848
+ },
+ {
+ "epoch": 39.57292576419214,
+ "grad_norm": 0.017133301123976707,
+ "learning_rate": 0.0006,
+ "loss": 4.097138404846191,
+ "step": 2849
+ },
+ {
+ "epoch": 39.58689956331878,
+ "grad_norm": 0.01683625392615795,
+ "learning_rate": 0.0006,
+ "loss": 4.084364891052246,
+ "step": 2850
+ },
+ {
+ "epoch": 39.60087336244541,
+ "grad_norm": 0.0165710486471653,
+ "learning_rate": 0.0006,
+ "loss": 4.042361259460449,
+ "step": 2851
+ },
+ {
+ "epoch": 39.61484716157205,
+ "grad_norm": 0.01660393364727497,
+ "learning_rate": 0.0006,
+ "loss": 4.1395487785339355,
+ "step": 2852
+ },
+ {
+ "epoch": 39.62882096069869,
+ "grad_norm": 0.017025845125317574,
+ "learning_rate": 0.0006,
+ "loss": 4.1207075119018555,
+ "step": 2853
+ },
+ {
+ "epoch": 39.64279475982533,
+ "grad_norm": 0.01660497486591339,
+ "learning_rate": 0.0006,
+ "loss": 4.103846073150635,
+ "step": 2854
+ },
+ {
+ "epoch": 39.65676855895197,
+ "grad_norm": 0.01607946865260601,
+ "learning_rate": 0.0006,
+ "loss": 4.189016342163086,
+ "step": 2855
+ },
+ {
+ "epoch": 39.670742358078606,
+ "grad_norm": 0.01614505983889103,
+ "learning_rate": 0.0006,
+ "loss": 4.079860687255859,
+ "step": 2856
+ },
+ {
+ "epoch": 39.68471615720524,
+ "grad_norm": 0.017085086554288864,
+ "learning_rate": 0.0006,
+ "loss": 3.9774293899536133,
+ "step": 2857
+ },
+ {
+ "epoch": 39.698689956331876,
+ "grad_norm": 0.01816520094871521,
+ "learning_rate": 0.0006,
+ "loss": 4.113666534423828,
+ "step": 2858
+ },
+ {
+ "epoch": 39.712663755458514,
+ "grad_norm": 0.016597526147961617,
+ "learning_rate": 0.0006,
+ "loss": 4.068242073059082,
+ "step": 2859
+ },
+ {
+ "epoch": 39.72663755458515,
+ "grad_norm": 0.01672235131263733,
+ "learning_rate": 0.0006,
+ "loss": 4.024444580078125,
+ "step": 2860
+ },
+ {
+ "epoch": 39.74061135371179,
+ "grad_norm": 0.020093556493520737,
+ "learning_rate": 0.0006,
+ "loss": 4.064124584197998,
+ "step": 2861
+ },
+ {
+ "epoch": 39.75458515283843,
+ "grad_norm": 0.023141304031014442,
+ "learning_rate": 0.0006,
+ "loss": 4.130439758300781,
+ "step": 2862
+ },
+ {
+ "epoch": 39.76855895196506,
+ "grad_norm": 0.02289372682571411,
+ "learning_rate": 0.0006,
+ "loss": 4.14794921875,
+ "step": 2863
+ },
+ {
+ "epoch": 39.7825327510917,
+ "grad_norm": 0.01919020712375641,
+ "learning_rate": 0.0006,
+ "loss": 4.123961925506592,
+ "step": 2864
+ },
+ {
+ "epoch": 39.79650655021834,
+ "grad_norm": 0.016932329162955284,
+ "learning_rate": 0.0006,
+ "loss": 4.0944132804870605,
+ "step": 2865
+ },
+ {
+ "epoch": 39.81048034934498,
+ "grad_norm": 0.016644535586237907,
+ "learning_rate": 0.0006,
+ "loss": 4.124269962310791,
+ "step": 2866
+ },
+ {
+ "epoch": 39.82445414847162,
+ "grad_norm": 0.016302216798067093,
+ "learning_rate": 0.0006,
+ "loss": 4.0267205238342285,
+ "step": 2867
+ },
+ {
+ "epoch": 39.838427947598255,
+ "grad_norm": 0.01673300378024578,
+ "learning_rate": 0.0006,
+ "loss": 4.10650634765625,
+ "step": 2868
+ },
+ {
+ "epoch": 39.852401746724894,
+ "grad_norm": 0.016581356525421143,
+ "learning_rate": 0.0006,
+ "loss": 4.1547956466674805,
+ "step": 2869
+ },
+ {
+ "epoch": 39.866375545851525,
+ "grad_norm": 0.01691797934472561,
+ "learning_rate": 0.0006,
+ "loss": 4.056936740875244,
+ "step": 2870
+ },
+ {
+ "epoch": 39.880349344978164,
+ "grad_norm": 0.01596726104617119,
+ "learning_rate": 0.0006,
+ "loss": 3.9891414642333984,
+ "step": 2871
+ },
+ {
+ "epoch": 39.8943231441048,
+ "grad_norm": 0.015615091659128666,
+ "learning_rate": 0.0006,
+ "loss": 4.137948036193848,
+ "step": 2872
+ },
+ {
+ "epoch": 39.90829694323144,
+ "grad_norm": 0.015703413635492325,
+ "learning_rate": 0.0006,
+ "loss": 3.9844930171966553,
+ "step": 2873
+ },
+ {
+ "epoch": 39.92227074235808,
+ "grad_norm": 0.015453094616532326,
+ "learning_rate": 0.0006,
+ "loss": 4.098443508148193,
+ "step": 2874
+ },
+ {
+ "epoch": 39.93624454148472,
+ "grad_norm": 0.01449158601462841,
+ "learning_rate": 0.0006,
+ "loss": 4.090852737426758,
+ "step": 2875
+ },
+ {
+ "epoch": 39.95021834061135,
+ "grad_norm": 0.01537515688687563,
+ "learning_rate": 0.0006,
+ "loss": 4.239612579345703,
+ "step": 2876
+ },
+ {
+ "epoch": 39.96419213973799,
+ "grad_norm": 0.015989921987056732,
+ "learning_rate": 0.0006,
+ "loss": 4.142073154449463,
+ "step": 2877
+ },
+ {
+ "epoch": 39.97816593886463,
+ "grad_norm": 0.015985611826181412,
+ "learning_rate": 0.0006,
+ "loss": 4.100076675415039,
+ "step": 2878
+ },
+ {
+ "epoch": 39.992139737991266,
+ "grad_norm": 0.015260833315551281,
+ "learning_rate": 0.0006,
+ "loss": 4.068576335906982,
+ "step": 2879
+ },
+ {
+ "epoch": 40.0,
+ "grad_norm": 0.018047941848635674,
+ "learning_rate": 0.0006,
+ "loss": 4.005878925323486,
+ "step": 2880
+ },
+ {
+ "epoch": 40.0,
+ "eval_loss": 4.622848987579346,
+ "eval_runtime": 56.9618,
+ "eval_samples_per_second": 42.871,
+ "eval_steps_per_second": 1.352,
+ "step": 2880
+ },
+ {
+ "epoch": 40.01397379912664,
+ "grad_norm": 0.015976838767528534,
+ "learning_rate": 0.0006,
+ "loss": 4.035689353942871,
+ "step": 2881
+ },
+ {
+ "epoch": 40.02794759825328,
+ "grad_norm": 0.015176220797002316,
+ "learning_rate": 0.0006,
+ "loss": 3.932197332382202,
+ "step": 2882
+ },
+ {
+ "epoch": 40.041921397379916,
+ "grad_norm": 0.016676228493452072,
+ "learning_rate": 0.0006,
+ "loss": 4.014171123504639,
+ "step": 2883
+ },
+ {
+ "epoch": 40.05589519650655,
+ "grad_norm": 0.01698281429708004,
+ "learning_rate": 0.0006,
+ "loss": 4.117893218994141,
+ "step": 2884
+ },
+ {
+ "epoch": 40.069868995633186,
+ "grad_norm": 0.017872486263513565,
+ "learning_rate": 0.0006,
+ "loss": 4.132119178771973,
+ "step": 2885
+ },
+ {
+ "epoch": 40.083842794759825,
+ "grad_norm": 0.017615554854273796,
+ "learning_rate": 0.0006,
+ "loss": 4.072000980377197,
+ "step": 2886
+ },
+ {
+ "epoch": 40.09781659388646,
+ "grad_norm": 0.017469394952058792,
+ "learning_rate": 0.0006,
+ "loss": 4.017726421356201,
+ "step": 2887
+ },
+ {
+ "epoch": 40.1117903930131,
+ "grad_norm": 0.01954762637615204,
+ "learning_rate": 0.0006,
+ "loss": 4.014443874359131,
+ "step": 2888
+ },
+ {
+ "epoch": 40.12576419213974,
+ "grad_norm": 0.018935853615403175,
+ "learning_rate": 0.0006,
+ "loss": 4.0042948722839355,
+ "step": 2889
+ },
+ {
+ "epoch": 40.13973799126637,
+ "grad_norm": 0.019382692873477936,
+ "learning_rate": 0.0006,
+ "loss": 4.06160831451416,
+ "step": 2890
+ },
+ {
+ "epoch": 40.15371179039301,
+ "grad_norm": 0.019677115604281425,
+ "learning_rate": 0.0006,
+ "loss": 4.13466215133667,
+ "step": 2891
+ },
+ {
+ "epoch": 40.16768558951965,
+ "grad_norm": 0.020365122705698013,
+ "learning_rate": 0.0006,
+ "loss": 4.095000267028809,
+ "step": 2892
+ },
+ {
+ "epoch": 40.18165938864629,
+ "grad_norm": 0.021842138841748238,
+ "learning_rate": 0.0006,
+ "loss": 3.9877853393554688,
+ "step": 2893
+ },
+ {
+ "epoch": 40.19563318777293,
+ "grad_norm": 0.02354881353676319,
+ "learning_rate": 0.0006,
+ "loss": 3.9430112838745117,
+ "step": 2894
+ },
+ {
+ "epoch": 40.209606986899566,
+ "grad_norm": 0.026602735742926598,
+ "learning_rate": 0.0006,
+ "loss": 4.012076377868652,
+ "step": 2895
+ },
+ {
+ "epoch": 40.223580786026204,
+ "grad_norm": 0.023976484313607216,
+ "learning_rate": 0.0006,
+ "loss": 3.9768595695495605,
+ "step": 2896
+ },
+ {
+ "epoch": 40.237554585152836,
+ "grad_norm": 0.02336616814136505,
+ "learning_rate": 0.0006,
+ "loss": 4.1096038818359375,
+ "step": 2897
+ },
+ {
+ "epoch": 40.251528384279474,
+ "grad_norm": 0.02426023595035076,
+ "learning_rate": 0.0006,
+ "loss": 4.157189846038818,
+ "step": 2898
+ },
+ {
+ "epoch": 40.26550218340611,
+ "grad_norm": 0.021572696045041084,
+ "learning_rate": 0.0006,
+ "loss": 3.979175090789795,
+ "step": 2899
+ },
+ {
+ "epoch": 40.27947598253275,
+ "grad_norm": 0.08300432562828064,
+ "learning_rate": 0.0006,
+ "loss": 4.092336654663086,
+ "step": 2900
+ },
+ {
+ "epoch": 40.29344978165939,
+ "grad_norm": 0.7323967814445496,
+ "learning_rate": 0.0006,
+ "loss": 5.337990760803223,
+ "step": 2901
+ },
+ {
+ "epoch": 40.30742358078603,
+ "grad_norm": 0.3363337814807892,
+ "learning_rate": 0.0006,
+ "loss": 6.40043830871582,
+ "step": 2902
+ },
+ {
+ "epoch": 40.32139737991266,
+ "grad_norm": 0.598686933517456,
+ "learning_rate": 0.0006,
+ "loss": 7.004765510559082,
+ "step": 2903
+ },
+ {
+ "epoch": 40.3353711790393,
+ "grad_norm": 0.5998364686965942,
+ "learning_rate": 0.0006,
+ "loss": 8.184747695922852,
+ "step": 2904
+ },
+ {
+ "epoch": 40.34934497816594,
+ "grad_norm": 1.408208966255188,
+ "learning_rate": 0.0006,
+ "loss": 8.88978385925293,
+ "step": 2905
+ },
+ {
+ "epoch": 40.36331877729258,
+ "grad_norm": 0.7829785943031311,
+ "learning_rate": 0.0006,
+ "loss": 11.13601303100586,
+ "step": 2906
+ },
+ {
+ "epoch": 40.377292576419215,
+ "grad_norm": 0.7769215106964111,
+ "learning_rate": 0.0006,
+ "loss": 9.90416145324707,
+ "step": 2907
+ },
+ {
+ "epoch": 40.391266375545854,
+ "grad_norm": 0.46508079767227173,
+ "learning_rate": 0.0006,
+ "loss": 9.647209167480469,
+ "step": 2908
+ },
+ {
+ "epoch": 40.40524017467249,
+ "grad_norm": 1.0228968858718872,
+ "learning_rate": 0.0006,
+ "loss": 11.685434341430664,
+ "step": 2909
+ },
+ {
+ "epoch": 40.419213973799124,
+ "grad_norm": 0.20553778111934662,
+ "learning_rate": 0.0006,
+ "loss": 8.621708869934082,
+ "step": 2910
+ },
+ {
+ "epoch": 40.43318777292576,
+ "grad_norm": 0.18466424942016602,
+ "learning_rate": 0.0006,
+ "loss": 8.590787887573242,
+ "step": 2911
+ },
+ {
+ "epoch": 40.4471615720524,
+ "grad_norm": 0.10458594560623169,
+ "learning_rate": 0.0006,
+ "loss": 8.324458122253418,
+ "step": 2912
+ },
+ {
+ "epoch": 40.46113537117904,
+ "grad_norm": 0.1096842959523201,
+ "learning_rate": 0.0006,
+ "loss": 8.300851821899414,
+ "step": 2913
+ },
+ {
+ "epoch": 40.47510917030568,
+ "grad_norm": 0.10501226782798767,
+ "learning_rate": 0.0006,
+ "loss": 8.21009635925293,
+ "step": 2914
+ },
+ {
+ "epoch": 40.48908296943232,
+ "grad_norm": 0.13704165816307068,
+ "learning_rate": 0.0006,
+ "loss": 8.141355514526367,
+ "step": 2915
+ },
+ {
+ "epoch": 40.50305676855895,
+ "grad_norm": 0.12129108607769012,
+ "learning_rate": 0.0006,
+ "loss": 8.074612617492676,
+ "step": 2916
+ },
+ {
+ "epoch": 40.51703056768559,
+ "grad_norm": 0.10028897225856781,
+ "learning_rate": 0.0006,
+ "loss": 8.00384521484375,
+ "step": 2917
+ },
+ {
+ "epoch": 40.531004366812226,
+ "grad_norm": 0.11764246970415115,
+ "learning_rate": 0.0006,
+ "loss": 8.078598022460938,
+ "step": 2918
+ },
+ {
+ "epoch": 40.544978165938865,
+ "grad_norm": 0.10081959515810013,
+ "learning_rate": 0.0006,
+ "loss": 8.008938789367676,
+ "step": 2919
+ },
+ {
+ "epoch": 40.5589519650655,
+ "grad_norm": 0.1003170907497406,
+ "learning_rate": 0.0006,
+ "loss": 7.962940216064453,
+ "step": 2920
+ },
+ {
+ "epoch": 40.57292576419214,
+ "grad_norm": 0.08763492107391357,
+ "learning_rate": 0.0006,
+ "loss": 7.909419536590576,
+ "step": 2921
+ },
+ {
+ "epoch": 40.58689956331878,
+ "grad_norm": 0.0636807233095169,
+ "learning_rate": 0.0006,
+ "loss": 7.878905773162842,
+ "step": 2922
+ },
+ {
+ "epoch": 40.60087336244541,
+ "grad_norm": 0.06005936488509178,
+ "learning_rate": 0.0006,
+ "loss": 7.850980758666992,
+ "step": 2923
+ },
+ {
+ "epoch": 40.61484716157205,
+ "grad_norm": 0.06415296345949173,
+ "learning_rate": 0.0006,
+ "loss": 7.830353260040283,
+ "step": 2924
+ },
+ {
+ "epoch": 40.62882096069869,
+ "grad_norm": 0.0732007846236229,
+ "learning_rate": 0.0006,
+ "loss": 7.830746650695801,
+ "step": 2925
+ },
+ {
+ "epoch": 40.64279475982533,
+ "grad_norm": 0.06628110259771347,
+ "learning_rate": 0.0006,
+ "loss": 7.82661771774292,
+ "step": 2926
+ },
+ {
+ "epoch": 40.65676855895197,
+ "grad_norm": 0.05544218048453331,
+ "learning_rate": 0.0006,
+ "loss": 7.814305782318115,
+ "step": 2927
+ },
+ {
+ "epoch": 40.670742358078606,
+ "grad_norm": 0.050327103585004807,
+ "learning_rate": 0.0006,
+ "loss": 7.764055252075195,
+ "step": 2928
+ },
+ {
+ "epoch": 40.68471615720524,
+ "grad_norm": 0.04918425530195236,
+ "learning_rate": 0.0006,
+ "loss": 7.744625091552734,
+ "step": 2929
+ },
+ {
+ "epoch": 40.698689956331876,
+ "grad_norm": 0.042862266302108765,
+ "learning_rate": 0.0006,
+ "loss": 7.764426231384277,
+ "step": 2930
+ },
+ {
+ "epoch": 40.712663755458514,
+ "grad_norm": 0.04338089004158974,
+ "learning_rate": 0.0006,
+ "loss": 7.712530136108398,
+ "step": 2931
+ },
+ {
+ "epoch": 40.72663755458515,
+ "grad_norm": 0.04127737879753113,
+ "learning_rate": 0.0006,
+ "loss": 7.742180347442627,
+ "step": 2932
+ },
+ {
+ "epoch": 40.74061135371179,
+ "grad_norm": 0.03358631581068039,
+ "learning_rate": 0.0006,
+ "loss": 7.714834213256836,
+ "step": 2933
+ },
+ {
+ "epoch": 40.75458515283843,
+ "grad_norm": 0.02845790982246399,
+ "learning_rate": 0.0006,
+ "loss": 7.7021026611328125,
+ "step": 2934
+ },
+ {
+ "epoch": 40.76855895196506,
+ "grad_norm": 0.03012419492006302,
+ "learning_rate": 0.0006,
+ "loss": 7.712377548217773,
+ "step": 2935
+ },
+ {
+ "epoch": 40.7825327510917,
+ "grad_norm": 0.032981179654598236,
+ "learning_rate": 0.0006,
+ "loss": 7.666385650634766,
+ "step": 2936
+ },
+ {
+ "epoch": 40.79650655021834,
+ "grad_norm": 0.03357074037194252,
+ "learning_rate": 0.0006,
+ "loss": 7.649097919464111,
+ "step": 2937
+ },
+ {
+ "epoch": 40.81048034934498,
+ "grad_norm": 0.030914708971977234,
+ "learning_rate": 0.0006,
+ "loss": 7.649196624755859,
+ "step": 2938
+ },
+ {
+ "epoch": 40.82445414847162,
+ "grad_norm": 0.026296131312847137,
+ "learning_rate": 0.0006,
+ "loss": 7.629964828491211,
+ "step": 2939
+ },
+ {
+ "epoch": 40.838427947598255,
+ "grad_norm": 0.026677245274186134,
+ "learning_rate": 0.0006,
+ "loss": 7.599973678588867,
+ "step": 2940
+ },
+ {
+ "epoch": 40.852401746724894,
+ "grad_norm": 0.023208219558000565,
+ "learning_rate": 0.0006,
+ "loss": 7.601624488830566,
+ "step": 2941
+ },
+ {
+ "epoch": 40.866375545851525,
+ "grad_norm": 0.020741475746035576,
+ "learning_rate": 0.0006,
+ "loss": 7.603902339935303,
+ "step": 2942
+ },
+ {
+ "epoch": 40.880349344978164,
+ "grad_norm": 0.019221751019358635,
+ "learning_rate": 0.0006,
+ "loss": 7.5535888671875,
+ "step": 2943
+ },
+ {
+ "epoch": 40.8943231441048,
+ "grad_norm": 0.019491951912641525,
+ "learning_rate": 0.0006,
+ "loss": 7.54597282409668,
+ "step": 2944
+ },
+ {
+ "epoch": 40.90829694323144,
+ "grad_norm": 0.019702572375535965,
+ "learning_rate": 0.0006,
+ "loss": 7.537567138671875,
+ "step": 2945
+ },
+ {
+ "epoch": 40.92227074235808,
+ "grad_norm": 0.01768800988793373,
+ "learning_rate": 0.0006,
+ "loss": 7.512800216674805,
+ "step": 2946
+ },
+ {
+ "epoch": 40.93624454148472,
+ "grad_norm": 0.01926138624548912,
+ "learning_rate": 0.0006,
+ "loss": 7.509932518005371,
+ "step": 2947
+ },
+ {
+ "epoch": 40.95021834061135,
+ "grad_norm": 0.019030006602406502,
+ "learning_rate": 0.0006,
+ "loss": 7.497983932495117,
+ "step": 2948
+ },
+ {
+ "epoch": 40.96419213973799,
+ "grad_norm": 0.017313841730356216,
+ "learning_rate": 0.0006,
+ "loss": 7.463518142700195,
+ "step": 2949
+ },
+ {
+ "epoch": 40.97816593886463,
+ "grad_norm": 0.01310762483626604,
+ "learning_rate": 0.0006,
+ "loss": 7.474668502807617,
+ "step": 2950
+ },
+ {
+ "epoch": 40.992139737991266,
+ "grad_norm": 0.013008282519876957,
+ "learning_rate": 0.0006,
+ "loss": 7.45982551574707,
+ "step": 2951
+ },
+ {
+ "epoch": 41.0,
+ "grad_norm": 0.01441363524645567,
+ "learning_rate": 0.0006,
+ "loss": 7.446529388427734,
+ "step": 2952
+ },
+ {
+ "epoch": 41.0,
+ "eval_loss": 7.486120223999023,
+ "eval_runtime": 70.4986,
+ "eval_samples_per_second": 34.639,
+ "eval_steps_per_second": 1.092,
+ "step": 2952
+ },
+ {
+ "epoch": 41.01397379912664,
+ "grad_norm": 0.014061855152249336,
+ "learning_rate": 0.0006,
+ "loss": 7.4692182540893555,
+ "step": 2953
+ },
+ {
+ "epoch": 41.02794759825328,
+ "grad_norm": 0.01587500050663948,
+ "learning_rate": 0.0006,
+ "loss": 7.4182024002075195,
+ "step": 2954
+ },
+ {
+ "epoch": 41.041921397379916,
+ "grad_norm": 0.014783984050154686,
+ "learning_rate": 0.0006,
+ "loss": 7.466778755187988,
+ "step": 2955
+ },
+ {
+ "epoch": 41.05589519650655,
+ "grad_norm": 0.011560854502022266,
+ "learning_rate": 0.0006,
+ "loss": 7.438002586364746,
+ "step": 2956
+ },
+ {
+ "epoch": 41.069868995633186,
+ "grad_norm": 0.00970957800745964,
+ "learning_rate": 0.0006,
+ "loss": 7.442694187164307,
+ "step": 2957
+ },
+ {
+ "epoch": 41.083842794759825,
+ "grad_norm": 0.009633459150791168,
+ "learning_rate": 0.0006,
+ "loss": 7.458884239196777,
+ "step": 2958
+ },
+ {
+ "epoch": 41.09781659388646,
+ "grad_norm": 0.010664350353181362,
+ "learning_rate": 0.0006,
+ "loss": 7.48378324508667,
+ "step": 2959
+ },
+ {
+ "epoch": 41.1117903930131,
+ "grad_norm": 0.011440591886639595,
+ "learning_rate": 0.0006,
+ "loss": 7.474484920501709,
+ "step": 2960
+ },
+ {
+ "epoch": 41.12576419213974,
+ "grad_norm": 0.01150908600538969,
+ "learning_rate": 0.0006,
+ "loss": 7.417902946472168,
+ "step": 2961
+ },
+ {
+ "epoch": 41.13973799126637,
+ "grad_norm": 0.010375697165727615,
+ "learning_rate": 0.0006,
+ "loss": 7.403279781341553,
+ "step": 2962
+ },
+ {
+ "epoch": 41.15371179039301,
+ "grad_norm": 0.009075703099370003,
+ "learning_rate": 0.0006,
+ "loss": 7.447347164154053,
+ "step": 2963
+ },
+ {
+ "epoch": 41.16768558951965,
+ "grad_norm": 0.008892177604138851,
+ "learning_rate": 0.0006,
+ "loss": 7.402681350708008,
+ "step": 2964
+ },
+ {
+ "epoch": 41.18165938864629,
+ "grad_norm": 0.009492084383964539,
+ "learning_rate": 0.0006,
+ "loss": 7.3871355056762695,
+ "step": 2965
+ },
+ {
+ "epoch": 41.19563318777293,
+ "grad_norm": 0.009139460511505604,
+ "learning_rate": 0.0006,
+ "loss": 7.419617176055908,
+ "step": 2966
+ },
+ {
+ "epoch": 41.209606986899566,
+ "grad_norm": 0.007959138602018356,
+ "learning_rate": 0.0006,
+ "loss": 7.374949932098389,
+ "step": 2967
+ },
+ {
+ "epoch": 41.223580786026204,
+ "grad_norm": 0.007571548223495483,
+ "learning_rate": 0.0006,
+ "loss": 7.4205241203308105,
+ "step": 2968
+ },
+ {
+ "epoch": 41.237554585152836,
+ "grad_norm": 0.007885400205850601,
+ "learning_rate": 0.0006,
+ "loss": 7.416543006896973,
+ "step": 2969
+ },
+ {
+ "epoch": 41.251528384279474,
+ "grad_norm": 0.008650158531963825,
+ "learning_rate": 0.0006,
+ "loss": 7.357326984405518,
+ "step": 2970
+ },
+ {
+ "epoch": 41.26550218340611,
+ "grad_norm": 0.007483262103050947,
+ "learning_rate": 0.0006,
+ "loss": 7.410891056060791,
+ "step": 2971
+ },
+ {
+ "epoch": 41.27947598253275,
+ "grad_norm": 0.00869277399033308,
+ "learning_rate": 0.0006,
+ "loss": 7.392127990722656,
+ "step": 2972
+ },
+ {
+ "epoch": 41.29344978165939,
+ "grad_norm": 0.00885078776627779,
+ "learning_rate": 0.0006,
+ "loss": 7.405449867248535,
+ "step": 2973
+ },
+ {
+ "epoch": 41.30742358078603,
+ "grad_norm": 0.00958226341754198,
+ "learning_rate": 0.0006,
+ "loss": 7.399868488311768,
+ "step": 2974
+ },
+ {
+ "epoch": 41.32139737991266,
+ "grad_norm": 0.006991599686443806,
+ "learning_rate": 0.0006,
+ "loss": 7.35369873046875,
+ "step": 2975
+ },
+ {
+ "epoch": 41.3353711790393,
+ "grad_norm": 0.007415142375975847,
+ "learning_rate": 0.0006,
+ "loss": 7.358353137969971,
+ "step": 2976
+ },
+ {
+ "epoch": 41.34934497816594,
+ "grad_norm": 0.007024371065199375,
+ "learning_rate": 0.0006,
+ "loss": 7.4070892333984375,
+ "step": 2977
+ },
+ {
+ "epoch": 41.36331877729258,
+ "grad_norm": 0.007234332151710987,
+ "learning_rate": 0.0006,
+ "loss": 7.379120349884033,
+ "step": 2978
+ },
+ {
+ "epoch": 41.377292576419215,
+ "grad_norm": 0.00798892229795456,
+ "learning_rate": 0.0006,
+ "loss": 7.365455627441406,
+ "step": 2979
+ },
+ {
+ "epoch": 41.391266375545854,
+ "grad_norm": 0.007303398102521896,
+ "learning_rate": 0.0006,
+ "loss": 7.341798782348633,
+ "step": 2980
+ },
+ {
+ "epoch": 41.40524017467249,
+ "grad_norm": 0.007119669578969479,
+ "learning_rate": 0.0006,
+ "loss": 7.363280296325684,
+ "step": 2981
+ },
+ {
+ "epoch": 41.419213973799124,
+ "grad_norm": 0.007457619532942772,
+ "learning_rate": 0.0006,
+ "loss": 7.39388370513916,
+ "step": 2982
+ },
+ {
+ "epoch": 41.43318777292576,
+ "grad_norm": 0.007905688136816025,
+ "learning_rate": 0.0006,
+ "loss": 7.380652904510498,
+ "step": 2983
+ },
+ {
+ "epoch": 41.4471615720524,
+ "grad_norm": 0.006212855689227581,
+ "learning_rate": 0.0006,
+ "loss": 7.398769378662109,
+ "step": 2984
+ },
+ {
+ "epoch": 41.46113537117904,
+ "grad_norm": 0.0070641543716192245,
+ "learning_rate": 0.0006,
+ "loss": 7.3503546714782715,
+ "step": 2985
+ },
+ {
+ "epoch": 41.47510917030568,
+ "grad_norm": 0.00787805300205946,
+ "learning_rate": 0.0006,
+ "loss": 7.347678184509277,
+ "step": 2986
+ },
+ {
+ "epoch": 41.48908296943232,
+ "grad_norm": 0.006760426796972752,
+ "learning_rate": 0.0006,
+ "loss": 7.380006790161133,
+ "step": 2987
+ },
+ {
+ "epoch": 41.50305676855895,
+ "grad_norm": 0.00772385997697711,
+ "learning_rate": 0.0006,
+ "loss": 7.363680362701416,
+ "step": 2988
+ },
+ {
+ "epoch": 41.51703056768559,
+ "grad_norm": 0.007373564876616001,
+ "learning_rate": 0.0006,
+ "loss": 7.333014488220215,
+ "step": 2989
+ },
+ {
+ "epoch": 41.531004366812226,
+ "grad_norm": 0.006717653013765812,
+ "learning_rate": 0.0006,
+ "loss": 7.360076427459717,
+ "step": 2990
+ },
+ {
+ "epoch": 41.544978165938865,
+ "grad_norm": 0.007191413082182407,
+ "learning_rate": 0.0006,
+ "loss": 7.397030830383301,
+ "step": 2991
+ },
+ {
+ "epoch": 41.5589519650655,
+ "grad_norm": 0.007831960916519165,
+ "learning_rate": 0.0006,
+ "loss": 7.369959831237793,
+ "step": 2992
+ },
+ {
+ "epoch": 41.57292576419214,
+ "grad_norm": 0.006947481073439121,
+ "learning_rate": 0.0006,
+ "loss": 7.333254337310791,
+ "step": 2993
+ },
+ {
+ "epoch": 41.58689956331878,
+ "grad_norm": 0.006243606563657522,
+ "learning_rate": 0.0006,
+ "loss": 7.327378273010254,
+ "step": 2994
+ },
+ {
+ "epoch": 41.60087336244541,
+ "grad_norm": 0.006298339460045099,
+ "learning_rate": 0.0006,
+ "loss": 7.323482036590576,
+ "step": 2995
+ },
+ {
+ "epoch": 41.61484716157205,
+ "grad_norm": 0.007587412837892771,
+ "learning_rate": 0.0006,
+ "loss": 7.313789367675781,
+ "step": 2996
+ },
+ {
+ "epoch": 41.62882096069869,
+ "grad_norm": 0.007140746805816889,
+ "learning_rate": 0.0006,
+ "loss": 7.356067657470703,
+ "step": 2997
+ },
+ {
+ "epoch": 41.64279475982533,
+ "grad_norm": 0.008315625600516796,
+ "learning_rate": 0.0006,
+ "loss": 7.3339643478393555,
+ "step": 2998
+ },
+ {
+ "epoch": 41.65676855895197,
+ "grad_norm": 0.006547102238982916,
+ "learning_rate": 0.0006,
+ "loss": 7.342191696166992,
+ "step": 2999
+ },
+ {
+ "epoch": 41.670742358078606,
+ "grad_norm": 0.006068122107535601,
+ "learning_rate": 0.0006,
+ "loss": 7.321015357971191,
+ "step": 3000
+ },
+ {
+ "epoch": 41.68471615720524,
+ "grad_norm": 0.007316991221159697,
+ "learning_rate": 0.0006,
+ "loss": 7.346490859985352,
+ "step": 3001
+ },
+ {
+ "epoch": 41.698689956331876,
+ "grad_norm": 0.0067388564348220825,
+ "learning_rate": 0.0006,
+ "loss": 7.340723991394043,
+ "step": 3002
+ },
+ {
+ "epoch": 41.712663755458514,
+ "grad_norm": 0.006839963141828775,
+ "learning_rate": 0.0006,
+ "loss": 7.329365253448486,
+ "step": 3003
+ },
+ {
+ "epoch": 41.72663755458515,
+ "grad_norm": 0.00805041566491127,
+ "learning_rate": 0.0006,
+ "loss": 7.298934459686279,
+ "step": 3004
+ },
+ {
+ "epoch": 41.74061135371179,
+ "grad_norm": 0.007197503466159105,
+ "learning_rate": 0.0006,
+ "loss": 7.331496238708496,
+ "step": 3005
+ },
+ {
+ "epoch": 41.75458515283843,
+ "grad_norm": 0.007220805156975985,
+ "learning_rate": 0.0006,
+ "loss": 7.317695140838623,
+ "step": 3006
+ },
+ {
+ "epoch": 41.76855895196506,
+ "grad_norm": 0.00874276738613844,
+ "learning_rate": 0.0006,
+ "loss": 7.272614479064941,
+ "step": 3007
+ },
+ {
+ "epoch": 41.7825327510917,
+ "grad_norm": 0.008386565372347832,
+ "learning_rate": 0.0006,
+ "loss": 7.295244216918945,
+ "step": 3008
+ },
+ {
+ "epoch": 41.79650655021834,
+ "grad_norm": 0.0077730282209813595,
+ "learning_rate": 0.0006,
+ "loss": 7.3395490646362305,
+ "step": 3009
+ },
+ {
+ "epoch": 41.81048034934498,
+ "grad_norm": 0.008441480807960033,
+ "learning_rate": 0.0006,
+ "loss": 7.319921493530273,
+ "step": 3010
+ },
+ {
+ "epoch": 41.82445414847162,
+ "grad_norm": 0.007398269604891539,
+ "learning_rate": 0.0006,
+ "loss": 7.3029465675354,
+ "step": 3011
+ },
+ {
+ "epoch": 41.838427947598255,
+ "grad_norm": 0.006878677289932966,
+ "learning_rate": 0.0006,
+ "loss": 7.318467140197754,
+ "step": 3012
+ },
+ {
+ "epoch": 41.852401746724894,
+ "grad_norm": 0.007616598624736071,
+ "learning_rate": 0.0006,
+ "loss": 7.3112077713012695,
+ "step": 3013
+ },
+ {
+ "epoch": 41.866375545851525,
+ "grad_norm": 0.011629699729382992,
+ "learning_rate": 0.0006,
+ "loss": 7.32523250579834,
+ "step": 3014
+ },
+ {
+ "epoch": 41.880349344978164,
+ "grad_norm": 0.011011249385774136,
+ "learning_rate": 0.0006,
+ "loss": 7.287247657775879,
+ "step": 3015
+ },
+ {
+ "epoch": 41.8943231441048,
+ "grad_norm": 0.010208295658230782,
+ "learning_rate": 0.0006,
+ "loss": 7.292117118835449,
+ "step": 3016
+ },
+ {
+ "epoch": 41.90829694323144,
+ "grad_norm": 0.008651095442473888,
+ "learning_rate": 0.0006,
+ "loss": 7.302201271057129,
+ "step": 3017
+ },
+ {
+ "epoch": 41.92227074235808,
+ "grad_norm": 0.006692920345813036,
+ "learning_rate": 0.0006,
+ "loss": 7.319568634033203,
+ "step": 3018
+ },
+ {
+ "epoch": 41.93624454148472,
+ "grad_norm": 0.010033348575234413,
+ "learning_rate": 0.0006,
+ "loss": 7.2738566398620605,
+ "step": 3019
+ },
+ {
+ "epoch": 41.95021834061135,
+ "grad_norm": 0.011790832504630089,
+ "learning_rate": 0.0006,
+ "loss": 7.298540115356445,
+ "step": 3020
+ },
+ {
+ "epoch": 41.96419213973799,
+ "grad_norm": 0.018622662872076035,
+ "learning_rate": 0.0006,
+ "loss": 7.267182350158691,
+ "step": 3021
+ },
+ {
+ "epoch": 41.97816593886463,
+ "grad_norm": 0.02520257607102394,
+ "learning_rate": 0.0006,
+ "loss": 7.261566162109375,
+ "step": 3022
+ },
+ {
+ "epoch": 41.992139737991266,
+ "grad_norm": 0.036332644522190094,
+ "learning_rate": 0.0006,
+ "loss": 7.270719051361084,
+ "step": 3023
+ },
+ {
+ "epoch": 42.0,
+ "grad_norm": 0.04717305675148964,
+ "learning_rate": 0.0006,
+ "loss": 7.293332099914551,
+ "step": 3024
+ },
+ {
+ "epoch": 42.0,
+ "eval_loss": 7.354280471801758,
+ "eval_runtime": 56.4353,
+ "eval_samples_per_second": 43.271,
+ "eval_steps_per_second": 1.364,
+ "step": 3024
+ },
+ {
+ "epoch": 42.01397379912664,
+ "grad_norm": 0.06661626696586609,
+ "learning_rate": 0.0006,
+ "loss": 7.338189601898193,
+ "step": 3025
+ },
+ {
+ "epoch": 42.02794759825328,
+ "grad_norm": 0.05452054366469383,
+ "learning_rate": 0.0006,
+ "loss": 7.316764831542969,
+ "step": 3026
+ },
+ {
+ "epoch": 42.041921397379916,
+ "grad_norm": 0.025638775900006294,
+ "learning_rate": 0.0006,
+ "loss": 7.34204626083374,
+ "step": 3027
+ },
+ {
+ "epoch": 42.05589519650655,
+ "grad_norm": 0.022398799657821655,
+ "learning_rate": 0.0006,
+ "loss": 7.262149333953857,
+ "step": 3028
+ },
+ {
+ "epoch": 42.069868995633186,
+ "grad_norm": 0.03713499382138252,
+ "learning_rate": 0.0006,
+ "loss": 7.27695369720459,
+ "step": 3029
+ },
+ {
+ "epoch": 42.083842794759825,
+ "grad_norm": 0.014949247241020203,
+ "learning_rate": 0.0006,
+ "loss": 7.296508312225342,
+ "step": 3030
+ },
+ {
+ "epoch": 42.09781659388646,
+ "grad_norm": 0.02703995071351528,
+ "learning_rate": 0.0006,
+ "loss": 7.258018493652344,
+ "step": 3031
+ },
+ {
+ "epoch": 42.1117903930131,
+ "grad_norm": 0.020352816209197044,
+ "learning_rate": 0.0006,
+ "loss": 7.241450309753418,
+ "step": 3032
+ },
+ {
+ "epoch": 42.12576419213974,
+ "grad_norm": 0.014578479342162609,
+ "learning_rate": 0.0006,
+ "loss": 7.257323265075684,
+ "step": 3033
+ },
+ {
+ "epoch": 42.13973799126637,
+ "grad_norm": 0.026053044945001602,
+ "learning_rate": 0.0006,
+ "loss": 7.254680633544922,
+ "step": 3034
+ },
+ {
+ "epoch": 42.15371179039301,
+ "grad_norm": 0.015120275318622589,
+ "learning_rate": 0.0006,
+ "loss": 7.261074066162109,
+ "step": 3035
+ },
+ {
+ "epoch": 42.16768558951965,
+ "grad_norm": 0.016822177916765213,
+ "learning_rate": 0.0006,
+ "loss": 7.279064178466797,
+ "step": 3036
+ },
+ {
+ "epoch": 42.18165938864629,
+ "grad_norm": 0.016367588192224503,
+ "learning_rate": 0.0006,
+ "loss": 7.251429557800293,
+ "step": 3037
+ },
+ {
+ "epoch": 42.19563318777293,
+ "grad_norm": 0.014350418001413345,
+ "learning_rate": 0.0006,
+ "loss": 7.259264945983887,
+ "step": 3038
+ },
+ {
+ "epoch": 42.209606986899566,
+ "grad_norm": 0.017078228294849396,
+ "learning_rate": 0.0006,
+ "loss": 7.227105140686035,
+ "step": 3039
+ },
+ {
+ "epoch": 42.223580786026204,
+ "grad_norm": 0.011759317480027676,
+ "learning_rate": 0.0006,
+ "loss": 7.275888442993164,
+ "step": 3040
+ },
+ {
+ "epoch": 42.237554585152836,
+ "grad_norm": 0.013019340112805367,
+ "learning_rate": 0.0006,
+ "loss": 7.2517008781433105,
+ "step": 3041
+ },
+ {
+ "epoch": 42.251528384279474,
+ "grad_norm": 0.01178441196680069,
+ "learning_rate": 0.0006,
+ "loss": 7.23576545715332,
+ "step": 3042
+ },
+ {
+ "epoch": 42.26550218340611,
+ "grad_norm": 0.014518891461193562,
+ "learning_rate": 0.0006,
+ "loss": 7.253895282745361,
+ "step": 3043
+ },
+ {
+ "epoch": 42.27947598253275,
+ "grad_norm": 0.018159950152039528,
+ "learning_rate": 0.0006,
+ "loss": 7.278058052062988,
+ "step": 3044
+ },
+ {
+ "epoch": 42.29344978165939,
+ "grad_norm": 0.007587796077132225,
+ "learning_rate": 0.0006,
+ "loss": 7.278578758239746,
+ "step": 3045
+ },
+ {
+ "epoch": 42.30742358078603,
+ "grad_norm": 0.015994979068636894,
+ "learning_rate": 0.0006,
+ "loss": 7.239841461181641,
+ "step": 3046
+ },
+ {
+ "epoch": 42.32139737991266,
+ "grad_norm": 0.017830168828368187,
+ "learning_rate": 0.0006,
+ "loss": 7.2443976402282715,
+ "step": 3047
+ },
+ {
+ "epoch": 42.3353711790393,
+ "grad_norm": 0.008738759905099869,
+ "learning_rate": 0.0006,
+ "loss": 7.232264518737793,
+ "step": 3048
+ },
+ {
+ "epoch": 42.34934497816594,
+ "grad_norm": 0.0175449438393116,
+ "learning_rate": 0.0006,
+ "loss": 7.228000640869141,
+ "step": 3049
+ },
+ {
+ "epoch": 42.36331877729258,
+ "grad_norm": 0.019611820578575134,
+ "learning_rate": 0.0006,
+ "loss": 7.234999179840088,
+ "step": 3050
+ },
+ {
+ "epoch": 42.377292576419215,
+ "grad_norm": 0.015302863903343678,
+ "learning_rate": 0.0006,
+ "loss": 7.242547988891602,
+ "step": 3051
+ },
+ {
+ "epoch": 42.391266375545854,
+ "grad_norm": 0.025919731706380844,
+ "learning_rate": 0.0006,
+ "loss": 7.216652870178223,
+ "step": 3052
+ },
+ {
+ "epoch": 42.40524017467249,
+ "grad_norm": 0.034587468951940536,
+ "learning_rate": 0.0006,
+ "loss": 7.245336532592773,
+ "step": 3053
+ },
+ {
+ "epoch": 42.419213973799124,
+ "grad_norm": 0.019198840484023094,
+ "learning_rate": 0.0006,
+ "loss": 7.256758689880371,
+ "step": 3054
+ },
+ {
+ "epoch": 42.43318777292576,
+ "grad_norm": 0.012316389009356499,
+ "learning_rate": 0.0006,
+ "loss": 7.209102630615234,
+ "step": 3055
+ },
+ {
+ "epoch": 42.4471615720524,
+ "grad_norm": 0.023512162268161774,
+ "learning_rate": 0.0006,
+ "loss": 7.250742435455322,
+ "step": 3056
+ },
+ {
+ "epoch": 42.46113537117904,
+ "grad_norm": 0.025457851588726044,
+ "learning_rate": 0.0006,
+ "loss": 7.263757705688477,
+ "step": 3057
+ },
+ {
+ "epoch": 42.47510917030568,
+ "grad_norm": 0.02192617766559124,
+ "learning_rate": 0.0006,
+ "loss": 7.246977806091309,
+ "step": 3058
+ },
+ {
+ "epoch": 42.48908296943232,
+ "grad_norm": 0.02541281282901764,
+ "learning_rate": 0.0006,
+ "loss": 7.232361316680908,
+ "step": 3059
+ },
+ {
+ "epoch": 42.50305676855895,
+ "grad_norm": 0.036353763192892075,
+ "learning_rate": 0.0006,
+ "loss": 7.242556571960449,
+ "step": 3060
+ },
+ {
+ "epoch": 42.51703056768559,
+ "grad_norm": 0.04555872827768326,
+ "learning_rate": 0.0006,
+ "loss": 7.228940963745117,
+ "step": 3061
+ },
+ {
+ "epoch": 42.531004366812226,
+ "grad_norm": 0.04563378915190697,
+ "learning_rate": 0.0006,
+ "loss": 7.235276222229004,
+ "step": 3062
+ },
+ {
+ "epoch": 42.544978165938865,
+ "grad_norm": 0.04275760427117348,
+ "learning_rate": 0.0006,
+ "loss": 7.238895416259766,
+ "step": 3063
+ },
+ {
+ "epoch": 42.5589519650655,
+ "grad_norm": 0.04535885155200958,
+ "learning_rate": 0.0006,
+ "loss": 7.205336570739746,
+ "step": 3064
+ },
+ {
+ "epoch": 42.57292576419214,
+ "grad_norm": 0.015534617006778717,
+ "learning_rate": 0.0006,
+ "loss": 7.238393783569336,
+ "step": 3065
+ },
+ {
+ "epoch": 42.58689956331878,
+ "grad_norm": 0.03286299481987953,
+ "learning_rate": 0.0006,
+ "loss": 7.216453552246094,
+ "step": 3066
+ },
+ {
+ "epoch": 42.60087336244541,
+ "grad_norm": 0.025453168898820877,
+ "learning_rate": 0.0006,
+ "loss": 7.203750133514404,
+ "step": 3067
+ },
+ {
+ "epoch": 42.61484716157205,
+ "grad_norm": 0.021501649171113968,
+ "learning_rate": 0.0006,
+ "loss": 7.241621971130371,
+ "step": 3068
+ },
+ {
+ "epoch": 42.62882096069869,
+ "grad_norm": 0.02743634767830372,
+ "learning_rate": 0.0006,
+ "loss": 7.22076940536499,
+ "step": 3069
+ },
+ {
+ "epoch": 42.64279475982533,
+ "grad_norm": 0.020890483632683754,
+ "learning_rate": 0.0006,
+ "loss": 7.218412399291992,
+ "step": 3070
+ },
+ {
+ "epoch": 42.65676855895197,
+ "grad_norm": 0.022499850019812584,
+ "learning_rate": 0.0006,
+ "loss": 7.217696189880371,
+ "step": 3071
+ },
+ {
+ "epoch": 42.670742358078606,
+ "grad_norm": 0.01653696969151497,
+ "learning_rate": 0.0006,
+ "loss": 7.156417369842529,
+ "step": 3072
+ },
+ {
+ "epoch": 42.68471615720524,
+ "grad_norm": 0.01834225095808506,
+ "learning_rate": 0.0006,
+ "loss": 7.211247444152832,
+ "step": 3073
+ },
+ {
+ "epoch": 42.698689956331876,
+ "grad_norm": 0.020137229934334755,
+ "learning_rate": 0.0006,
+ "loss": 7.232499599456787,
+ "step": 3074
+ },
+ {
+ "epoch": 42.712663755458514,
+ "grad_norm": 0.012671459466218948,
+ "learning_rate": 0.0006,
+ "loss": 7.172441482543945,
+ "step": 3075
+ },
+ {
+ "epoch": 42.72663755458515,
+ "grad_norm": 0.016803130507469177,
+ "learning_rate": 0.0006,
+ "loss": 7.152103424072266,
+ "step": 3076
+ },
+ {
+ "epoch": 42.74061135371179,
+ "grad_norm": 0.01623712107539177,
+ "learning_rate": 0.0006,
+ "loss": 7.187032699584961,
+ "step": 3077
+ },
+ {
+ "epoch": 42.75458515283843,
+ "grad_norm": 0.015938356518745422,
+ "learning_rate": 0.0006,
+ "loss": 7.171140670776367,
+ "step": 3078
+ },
+ {
+ "epoch": 42.76855895196506,
+ "grad_norm": 0.02580736204981804,
+ "learning_rate": 0.0006,
+ "loss": 7.204472541809082,
+ "step": 3079
+ },
+ {
+ "epoch": 42.7825327510917,
+ "grad_norm": 0.023372117429971695,
+ "learning_rate": 0.0006,
+ "loss": 7.165554523468018,
+ "step": 3080
+ },
+ {
+ "epoch": 42.79650655021834,
+ "grad_norm": 0.04440586268901825,
+ "learning_rate": 0.0006,
+ "loss": 7.169431209564209,
+ "step": 3081
+ },
+ {
+ "epoch": 42.81048034934498,
+ "grad_norm": 0.06797406077384949,
+ "learning_rate": 0.0006,
+ "loss": 7.168193340301514,
+ "step": 3082
+ },
+ {
+ "epoch": 42.82445414847162,
+ "grad_norm": 0.0914321169257164,
+ "learning_rate": 0.0006,
+ "loss": 7.204671859741211,
+ "step": 3083
+ },
+ {
+ "epoch": 42.838427947598255,
+ "grad_norm": 0.0506940595805645,
+ "learning_rate": 0.0006,
+ "loss": 7.15949821472168,
+ "step": 3084
+ },
+ {
+ "epoch": 42.852401746724894,
+ "grad_norm": 0.04819890484213829,
+ "learning_rate": 0.0006,
+ "loss": 7.192837238311768,
+ "step": 3085
+ },
+ {
+ "epoch": 42.866375545851525,
+ "grad_norm": 0.040044233202934265,
+ "learning_rate": 0.0006,
+ "loss": 7.1600165367126465,
+ "step": 3086
+ },
+ {
+ "epoch": 42.880349344978164,
+ "grad_norm": 0.02573299966752529,
+ "learning_rate": 0.0006,
+ "loss": 7.153427600860596,
+ "step": 3087
+ },
+ {
+ "epoch": 42.8943231441048,
+ "grad_norm": 0.046942535787820816,
+ "learning_rate": 0.0006,
+ "loss": 7.16767692565918,
+ "step": 3088
+ },
+ {
+ "epoch": 42.90829694323144,
+ "grad_norm": 0.02683161571621895,
+ "learning_rate": 0.0006,
+ "loss": 7.17144250869751,
+ "step": 3089
+ },
+ {
+ "epoch": 42.92227074235808,
+ "grad_norm": 0.05363674834370613,
+ "learning_rate": 0.0006,
+ "loss": 7.120128631591797,
+ "step": 3090
+ },
+ {
+ "epoch": 42.93624454148472,
+ "grad_norm": 0.023756247013807297,
+ "learning_rate": 0.0006,
+ "loss": 7.169703006744385,
+ "step": 3091
+ },
+ {
+ "epoch": 42.95021834061135,
+ "grad_norm": 0.04216783866286278,
+ "learning_rate": 0.0006,
+ "loss": 7.1206865310668945,
+ "step": 3092
+ },
+ {
+ "epoch": 42.96419213973799,
+ "grad_norm": 0.02219565585255623,
+ "learning_rate": 0.0006,
+ "loss": 7.157215118408203,
+ "step": 3093
+ },
+ {
+ "epoch": 42.97816593886463,
+ "grad_norm": 0.03527127206325531,
+ "learning_rate": 0.0006,
+ "loss": 7.166672706604004,
+ "step": 3094
+ },
+ {
+ "epoch": 42.992139737991266,
+ "grad_norm": 0.017748817801475525,
+ "learning_rate": 0.0006,
+ "loss": 7.154747009277344,
+ "step": 3095
+ },
+ {
+ "epoch": 43.0,
+ "grad_norm": 0.033165838569402695,
+ "learning_rate": 0.0006,
+ "loss": 7.150866508483887,
+ "step": 3096
+ },
+ {
+ "epoch": 43.0,
+ "eval_loss": 7.144403457641602,
+ "eval_runtime": 56.8853,
+ "eval_samples_per_second": 42.928,
+ "eval_steps_per_second": 1.354,
+ "step": 3096
+ },
+ {
+ "epoch": 43.01397379912664,
+ "grad_norm": 0.015199829824268818,
+ "learning_rate": 0.0006,
+ "loss": 7.122735023498535,
+ "step": 3097
+ },
+ {
+ "epoch": 43.02794759825328,
+ "grad_norm": 0.03220256045460701,
+ "learning_rate": 0.0006,
+ "loss": 7.143553733825684,
+ "step": 3098
+ },
+ {
+ "epoch": 43.041921397379916,
+ "grad_norm": 0.014687920920550823,
+ "learning_rate": 0.0006,
+ "loss": 7.080960750579834,
+ "step": 3099
+ },
+ {
+ "epoch": 43.05589519650655,
+ "grad_norm": 0.023522932082414627,
+ "learning_rate": 0.0006,
+ "loss": 7.137356758117676,
+ "step": 3100
+ },
+ {
+ "epoch": 43.069868995633186,
+ "grad_norm": 0.014808216132223606,
+ "learning_rate": 0.0006,
+ "loss": 7.09143590927124,
+ "step": 3101
+ },
+ {
+ "epoch": 43.083842794759825,
+ "grad_norm": 0.029339246451854706,
+ "learning_rate": 0.0006,
+ "loss": 7.122198104858398,
+ "step": 3102
+ },
+ {
+ "epoch": 43.09781659388646,
+ "grad_norm": 0.022953078150749207,
+ "learning_rate": 0.0006,
+ "loss": 7.092505931854248,
+ "step": 3103
+ },
+ {
+ "epoch": 43.1117903930131,
+ "grad_norm": 0.04590052738785744,
+ "learning_rate": 0.0006,
+ "loss": 7.113100051879883,
+ "step": 3104
+ },
+ {
+ "epoch": 43.12576419213974,
+ "grad_norm": 0.0524442233145237,
+ "learning_rate": 0.0006,
+ "loss": 7.147022247314453,
+ "step": 3105
+ },
+ {
+ "epoch": 43.13973799126637,
+ "grad_norm": 0.0315794013440609,
+ "learning_rate": 0.0006,
+ "loss": 7.094045162200928,
+ "step": 3106
+ },
+ {
+ "epoch": 43.15371179039301,
+ "grad_norm": 0.05112722888588905,
+ "learning_rate": 0.0006,
+ "loss": 7.054388523101807,
+ "step": 3107
+ },
+ {
+ "epoch": 43.16768558951965,
+ "grad_norm": 0.03146776929497719,
+ "learning_rate": 0.0006,
+ "loss": 7.116674423217773,
+ "step": 3108
+ },
+ {
+ "epoch": 43.18165938864629,
+ "grad_norm": 0.03391246497631073,
+ "learning_rate": 0.0006,
+ "loss": 7.150345802307129,
+ "step": 3109
+ },
+ {
+ "epoch": 43.19563318777293,
+ "grad_norm": 0.026957903057336807,
+ "learning_rate": 0.0006,
+ "loss": 7.058531284332275,
+ "step": 3110
+ },
+ {
+ "epoch": 43.209606986899566,
+ "grad_norm": 0.041532568633556366,
+ "learning_rate": 0.0006,
+ "loss": 7.085498809814453,
+ "step": 3111
+ },
+ {
+ "epoch": 43.223580786026204,
+ "grad_norm": 0.020001588389277458,
+ "learning_rate": 0.0006,
+ "loss": 7.131487846374512,
+ "step": 3112
+ },
+ {
+ "epoch": 43.237554585152836,
+ "grad_norm": 0.02268712781369686,
+ "learning_rate": 0.0006,
+ "loss": 7.0790863037109375,
+ "step": 3113
+ },
+ {
+ "epoch": 43.251528384279474,
+ "grad_norm": 0.023363925516605377,
+ "learning_rate": 0.0006,
+ "loss": 7.067816734313965,
+ "step": 3114
+ },
+ {
+ "epoch": 43.26550218340611,
+ "grad_norm": 0.019367508590221405,
+ "learning_rate": 0.0006,
+ "loss": 7.042041778564453,
+ "step": 3115
+ },
+ {
+ "epoch": 43.27947598253275,
+ "grad_norm": 0.02745090238749981,
+ "learning_rate": 0.0006,
+ "loss": 7.064733028411865,
+ "step": 3116
+ },
+ {
+ "epoch": 43.29344978165939,
+ "grad_norm": 0.02543802745640278,
+ "learning_rate": 0.0006,
+ "loss": 7.058948040008545,
+ "step": 3117
+ },
+ {
+ "epoch": 43.30742358078603,
+ "grad_norm": 0.018874092027544975,
+ "learning_rate": 0.0006,
+ "loss": 7.009946346282959,
+ "step": 3118
+ },
+ {
+ "epoch": 43.32139737991266,
+ "grad_norm": 0.024662822484970093,
+ "learning_rate": 0.0006,
+ "loss": 7.015913963317871,
+ "step": 3119
+ },
+ {
+ "epoch": 43.3353711790393,
+ "grad_norm": 0.02441416308283806,
+ "learning_rate": 0.0006,
+ "loss": 7.061622142791748,
+ "step": 3120
+ },
+ {
+ "epoch": 43.34934497816594,
+ "grad_norm": 0.031887322664260864,
+ "learning_rate": 0.0006,
+ "loss": 7.046883583068848,
+ "step": 3121
+ },
+ {
+ "epoch": 43.36331877729258,
+ "grad_norm": 0.021930739283561707,
+ "learning_rate": 0.0006,
+ "loss": 7.0309624671936035,
+ "step": 3122
+ },
+ {
+ "epoch": 43.377292576419215,
+ "grad_norm": 0.02722783200442791,
+ "learning_rate": 0.0006,
+ "loss": 7.072924613952637,
+ "step": 3123
+ },
+ {
+ "epoch": 43.391266375545854,
+ "grad_norm": 0.01483179721981287,
+ "learning_rate": 0.0006,
+ "loss": 7.028624534606934,
+ "step": 3124
+ },
+ {
+ "epoch": 43.40524017467249,
+ "grad_norm": 0.024793945252895355,
+ "learning_rate": 0.0006,
+ "loss": 7.001623153686523,
+ "step": 3125
+ },
+ {
+ "epoch": 43.419213973799124,
+ "grad_norm": 0.020148713141679764,
+ "learning_rate": 0.0006,
+ "loss": 6.9938812255859375,
+ "step": 3126
+ },
+ {
+ "epoch": 43.43318777292576,
+ "grad_norm": 0.02413850650191307,
+ "learning_rate": 0.0006,
+ "loss": 7.001243591308594,
+ "step": 3127
+ },
+ {
+ "epoch": 43.4471615720524,
+ "grad_norm": 0.028549203649163246,
+ "learning_rate": 0.0006,
+ "loss": 6.987671375274658,
+ "step": 3128
+ },
+ {
+ "epoch": 43.46113537117904,
+ "grad_norm": 0.025988785549998283,
+ "learning_rate": 0.0006,
+ "loss": 6.9551167488098145,
+ "step": 3129
+ },
+ {
+ "epoch": 43.47510917030568,
+ "grad_norm": 0.027140729129314423,
+ "learning_rate": 0.0006,
+ "loss": 6.9547810554504395,
+ "step": 3130
+ },
+ {
+ "epoch": 43.48908296943232,
+ "grad_norm": 0.030130082741379738,
+ "learning_rate": 0.0006,
+ "loss": 6.993186950683594,
+ "step": 3131
+ },
+ {
+ "epoch": 43.50305676855895,
+ "grad_norm": 0.025996994227170944,
+ "learning_rate": 0.0006,
+ "loss": 6.984624862670898,
+ "step": 3132
+ },
+ {
+ "epoch": 43.51703056768559,
+ "grad_norm": 0.02423008903861046,
+ "learning_rate": 0.0006,
+ "loss": 6.98813009262085,
+ "step": 3133
+ },
+ {
+ "epoch": 43.531004366812226,
+ "grad_norm": 0.029665837064385414,
+ "learning_rate": 0.0006,
+ "loss": 6.986610412597656,
+ "step": 3134
+ },
+ {
+ "epoch": 43.544978165938865,
+ "grad_norm": 0.028768595308065414,
+ "learning_rate": 0.0006,
+ "loss": 6.898797035217285,
+ "step": 3135
+ },
+ {
+ "epoch": 43.5589519650655,
+ "grad_norm": 0.03111208975315094,
+ "learning_rate": 0.0006,
+ "loss": 6.923741817474365,
+ "step": 3136
+ },
+ {
+ "epoch": 43.57292576419214,
+ "grad_norm": 0.0368187390267849,
+ "learning_rate": 0.0006,
+ "loss": 6.994343280792236,
+ "step": 3137
+ },
+ {
+ "epoch": 43.58689956331878,
+ "grad_norm": 0.04076763987541199,
+ "learning_rate": 0.0006,
+ "loss": 6.909458160400391,
+ "step": 3138
+ },
+ {
+ "epoch": 43.60087336244541,
+ "grad_norm": 0.03281061723828316,
+ "learning_rate": 0.0006,
+ "loss": 6.924851417541504,
+ "step": 3139
+ },
+ {
+ "epoch": 43.61484716157205,
+ "grad_norm": 0.025486988946795464,
+ "learning_rate": 0.0006,
+ "loss": 6.962926864624023,
+ "step": 3140
+ },
+ {
+ "epoch": 43.62882096069869,
+ "grad_norm": 0.04583485797047615,
+ "learning_rate": 0.0006,
+ "loss": 6.876555442810059,
+ "step": 3141
+ },
+ {
+ "epoch": 43.64279475982533,
+ "grad_norm": 0.05460785701870918,
+ "learning_rate": 0.0006,
+ "loss": 6.952932357788086,
+ "step": 3142
+ },
+ {
+ "epoch": 43.65676855895197,
+ "grad_norm": 0.040917906910181046,
+ "learning_rate": 0.0006,
+ "loss": 6.9624924659729,
+ "step": 3143
+ },
+ {
+ "epoch": 43.670742358078606,
+ "grad_norm": 0.04009668901562691,
+ "learning_rate": 0.0006,
+ "loss": 6.94033145904541,
+ "step": 3144
+ },
+ {
+ "epoch": 43.68471615720524,
+ "grad_norm": 0.04500625655055046,
+ "learning_rate": 0.0006,
+ "loss": 6.961839199066162,
+ "step": 3145
+ },
+ {
+ "epoch": 43.698689956331876,
+ "grad_norm": 0.03182258829474449,
+ "learning_rate": 0.0006,
+ "loss": 6.930580139160156,
+ "step": 3146
+ },
+ {
+ "epoch": 43.712663755458514,
+ "grad_norm": 0.053621504455804825,
+ "learning_rate": 0.0006,
+ "loss": 6.950235366821289,
+ "step": 3147
+ },
+ {
+ "epoch": 43.72663755458515,
+ "grad_norm": 0.031919289380311966,
+ "learning_rate": 0.0006,
+ "loss": 6.895111560821533,
+ "step": 3148
+ },
+ {
+ "epoch": 43.74061135371179,
+ "grad_norm": 0.042713891714811325,
+ "learning_rate": 0.0006,
+ "loss": 6.940885543823242,
+ "step": 3149
+ },
+ {
+ "epoch": 43.75458515283843,
+ "grad_norm": 0.02385653182864189,
+ "learning_rate": 0.0006,
+ "loss": 6.948801517486572,
+ "step": 3150
+ },
+ {
+ "epoch": 43.76855895196506,
+ "grad_norm": 0.04100663587450981,
+ "learning_rate": 0.0006,
+ "loss": 6.970411777496338,
+ "step": 3151
+ },
+ {
+ "epoch": 43.7825327510917,
+ "grad_norm": 0.02729583904147148,
+ "learning_rate": 0.0006,
+ "loss": 6.899379253387451,
+ "step": 3152
+ },
+ {
+ "epoch": 43.79650655021834,
+ "grad_norm": 0.03128620237112045,
+ "learning_rate": 0.0006,
+ "loss": 6.931319236755371,
+ "step": 3153
+ },
+ {
+ "epoch": 43.81048034934498,
+ "grad_norm": 0.025488659739494324,
+ "learning_rate": 0.0006,
+ "loss": 6.8486409187316895,
+ "step": 3154
+ },
+ {
+ "epoch": 43.82445414847162,
+ "grad_norm": 0.03679376840591431,
+ "learning_rate": 0.0006,
+ "loss": 6.840811252593994,
+ "step": 3155
+ },
+ {
+ "epoch": 43.838427947598255,
+ "grad_norm": 0.020578034222126007,
+ "learning_rate": 0.0006,
+ "loss": 6.9088616371154785,
+ "step": 3156
+ },
+ {
+ "epoch": 43.852401746724894,
+ "grad_norm": 0.02935841865837574,
+ "learning_rate": 0.0006,
+ "loss": 6.806676864624023,
+ "step": 3157
+ },
+ {
+ "epoch": 43.866375545851525,
+ "grad_norm": 0.023319244384765625,
+ "learning_rate": 0.0006,
+ "loss": 6.878457546234131,
+ "step": 3158
+ },
+ {
+ "epoch": 43.880349344978164,
+ "grad_norm": 0.02779589593410492,
+ "learning_rate": 0.0006,
+ "loss": 6.822447776794434,
+ "step": 3159
+ },
+ {
+ "epoch": 43.8943231441048,
+ "grad_norm": 0.01879825070500374,
+ "learning_rate": 0.0006,
+ "loss": 6.854121208190918,
+ "step": 3160
+ },
+ {
+ "epoch": 43.90829694323144,
+ "grad_norm": 0.023286139592528343,
+ "learning_rate": 0.0006,
+ "loss": 6.836827278137207,
+ "step": 3161
+ },
+ {
+ "epoch": 43.92227074235808,
+ "grad_norm": 0.030358297750353813,
+ "learning_rate": 0.0006,
+ "loss": 6.860037803649902,
+ "step": 3162
+ },
+ {
+ "epoch": 43.93624454148472,
+ "grad_norm": 0.029621105641126633,
+ "learning_rate": 0.0006,
+ "loss": 6.884025573730469,
+ "step": 3163
+ },
+ {
+ "epoch": 43.95021834061135,
+ "grad_norm": 0.029829824343323708,
+ "learning_rate": 0.0006,
+ "loss": 6.8702392578125,
+ "step": 3164
+ },
+ {
+ "epoch": 43.96419213973799,
+ "grad_norm": 0.036892473697662354,
+ "learning_rate": 0.0006,
+ "loss": 6.806652069091797,
+ "step": 3165
+ },
+ {
+ "epoch": 43.97816593886463,
+ "grad_norm": 0.03206709027290344,
+ "learning_rate": 0.0006,
+ "loss": 6.828495979309082,
+ "step": 3166
+ },
+ {
+ "epoch": 43.992139737991266,
+ "grad_norm": 0.02744000405073166,
+ "learning_rate": 0.0006,
+ "loss": 6.832002639770508,
+ "step": 3167
+ },
+ {
+ "epoch": 44.0,
+ "grad_norm": 0.02205660194158554,
+ "learning_rate": 0.0006,
+ "loss": 6.898828506469727,
+ "step": 3168
+ },
+ {
+ "epoch": 44.0,
+ "eval_loss": 6.8150553703308105,
+ "eval_runtime": 56.784,
+ "eval_samples_per_second": 43.005,
+ "eval_steps_per_second": 1.356,
+ "step": 3168
+ },
+ {
+ "epoch": 44.01397379912664,
+ "grad_norm": 0.030421260744333267,
+ "learning_rate": 0.0006,
+ "loss": 6.775057792663574,
+ "step": 3169
+ },
+ {
+ "epoch": 44.02794759825328,
+ "grad_norm": 0.0465611033141613,
+ "learning_rate": 0.0006,
+ "loss": 6.849390506744385,
+ "step": 3170
+ },
+ {
+ "epoch": 44.041921397379916,
+ "grad_norm": 0.05977972596883774,
+ "learning_rate": 0.0006,
+ "loss": 6.725576400756836,
+ "step": 3171
+ },
+ {
+ "epoch": 44.05589519650655,
+ "grad_norm": 0.05432697385549545,
+ "learning_rate": 0.0006,
+ "loss": 6.768024444580078,
+ "step": 3172
+ },
+ {
+ "epoch": 44.069868995633186,
+ "grad_norm": 0.04850967228412628,
+ "learning_rate": 0.0006,
+ "loss": 6.799457550048828,
+ "step": 3173
+ },
+ {
+ "epoch": 44.083842794759825,
+ "grad_norm": 0.06201786920428276,
+ "learning_rate": 0.0006,
+ "loss": 6.897234916687012,
+ "step": 3174
+ },
+ {
+ "epoch": 44.09781659388646,
+ "grad_norm": 0.04695591703057289,
+ "learning_rate": 0.0006,
+ "loss": 6.826470375061035,
+ "step": 3175
+ },
+ {
+ "epoch": 44.1117903930131,
+ "grad_norm": 0.03850164636969566,
+ "learning_rate": 0.0006,
+ "loss": 6.835901260375977,
+ "step": 3176
+ },
+ {
+ "epoch": 44.12576419213974,
+ "grad_norm": 0.04250155761837959,
+ "learning_rate": 0.0006,
+ "loss": 6.782310485839844,
+ "step": 3177
+ },
+ {
+ "epoch": 44.13973799126637,
+ "grad_norm": 0.041465479880571365,
+ "learning_rate": 0.0006,
+ "loss": 6.785732746124268,
+ "step": 3178
+ },
+ {
+ "epoch": 44.15371179039301,
+ "grad_norm": 0.02527807280421257,
+ "learning_rate": 0.0006,
+ "loss": 6.78061580657959,
+ "step": 3179
+ },
+ {
+ "epoch": 44.16768558951965,
+ "grad_norm": 0.03660374879837036,
+ "learning_rate": 0.0006,
+ "loss": 6.783992767333984,
+ "step": 3180
+ },
+ {
+ "epoch": 44.18165938864629,
+ "grad_norm": 0.027567964047193527,
+ "learning_rate": 0.0006,
+ "loss": 6.765413284301758,
+ "step": 3181
+ },
+ {
+ "epoch": 44.19563318777293,
+ "grad_norm": 0.024301424622535706,
+ "learning_rate": 0.0006,
+ "loss": 6.760368347167969,
+ "step": 3182
+ },
+ {
+ "epoch": 44.209606986899566,
+ "grad_norm": 0.0266821701079607,
+ "learning_rate": 0.0006,
+ "loss": 6.71417236328125,
+ "step": 3183
+ },
+ {
+ "epoch": 44.223580786026204,
+ "grad_norm": 0.028125090524554253,
+ "learning_rate": 0.0006,
+ "loss": 6.646126747131348,
+ "step": 3184
+ },
+ {
+ "epoch": 44.237554585152836,
+ "grad_norm": 0.0185737032443285,
+ "learning_rate": 0.0006,
+ "loss": 6.712944030761719,
+ "step": 3185
+ },
+ {
+ "epoch": 44.251528384279474,
+ "grad_norm": 0.025836674496531487,
+ "learning_rate": 0.0006,
+ "loss": 6.707030296325684,
+ "step": 3186
+ },
+ {
+ "epoch": 44.26550218340611,
+ "grad_norm": 0.029744870960712433,
+ "learning_rate": 0.0006,
+ "loss": 6.6271071434021,
+ "step": 3187
+ },
+ {
+ "epoch": 44.27947598253275,
+ "grad_norm": 0.03021426685154438,
+ "learning_rate": 0.0006,
+ "loss": 6.677826404571533,
+ "step": 3188
+ },
+ {
+ "epoch": 44.29344978165939,
+ "grad_norm": 0.032727014273405075,
+ "learning_rate": 0.0006,
+ "loss": 6.657764434814453,
+ "step": 3189
+ },
+ {
+ "epoch": 44.30742358078603,
+ "grad_norm": 0.03696778789162636,
+ "learning_rate": 0.0006,
+ "loss": 6.632883548736572,
+ "step": 3190
+ },
+ {
+ "epoch": 44.32139737991266,
+ "grad_norm": 0.034630827605724335,
+ "learning_rate": 0.0006,
+ "loss": 6.759583473205566,
+ "step": 3191
+ },
+ {
+ "epoch": 44.3353711790393,
+ "grad_norm": 0.030517758801579475,
+ "learning_rate": 0.0006,
+ "loss": 6.702268600463867,
+ "step": 3192
+ },
+ {
+ "epoch": 44.34934497816594,
+ "grad_norm": 0.04105190932750702,
+ "learning_rate": 0.0006,
+ "loss": 6.705792427062988,
+ "step": 3193
+ },
+ {
+ "epoch": 44.36331877729258,
+ "grad_norm": 0.06659174710512161,
+ "learning_rate": 0.0006,
+ "loss": 6.73862361907959,
+ "step": 3194
+ },
+ {
+ "epoch": 44.377292576419215,
+ "grad_norm": 0.08013619482517242,
+ "learning_rate": 0.0006,
+ "loss": 6.710282802581787,
+ "step": 3195
+ },
+ {
+ "epoch": 44.391266375545854,
+ "grad_norm": 0.049850963056087494,
+ "learning_rate": 0.0006,
+ "loss": 6.684974670410156,
+ "step": 3196
+ },
+ {
+ "epoch": 44.40524017467249,
+ "grad_norm": 0.061247408390045166,
+ "learning_rate": 0.0006,
+ "loss": 6.716034412384033,
+ "step": 3197
+ },
+ {
+ "epoch": 44.419213973799124,
+ "grad_norm": 0.05868857726454735,
+ "learning_rate": 0.0006,
+ "loss": 6.589107513427734,
+ "step": 3198
+ },
+ {
+ "epoch": 44.43318777292576,
+ "grad_norm": 0.048878561705350876,
+ "learning_rate": 0.0006,
+ "loss": 6.694855690002441,
+ "step": 3199
+ },
+ {
+ "epoch": 44.4471615720524,
+ "grad_norm": 0.028748324140906334,
+ "learning_rate": 0.0006,
+ "loss": 6.629969596862793,
+ "step": 3200
+ },
+ {
+ "epoch": 44.46113537117904,
+ "grad_norm": 0.04060306400060654,
+ "learning_rate": 0.0006,
+ "loss": 6.721431255340576,
+ "step": 3201
+ },
+ {
+ "epoch": 44.47510917030568,
+ "grad_norm": 0.03833824768662453,
+ "learning_rate": 0.0006,
+ "loss": 6.625029563903809,
+ "step": 3202
+ },
+ {
+ "epoch": 44.48908296943232,
+ "grad_norm": 0.03791997581720352,
+ "learning_rate": 0.0006,
+ "loss": 6.669187545776367,
+ "step": 3203
+ },
+ {
+ "epoch": 44.50305676855895,
+ "grad_norm": 0.03259684890508652,
+ "learning_rate": 0.0006,
+ "loss": 6.665643692016602,
+ "step": 3204
+ },
+ {
+ "epoch": 44.51703056768559,
+ "grad_norm": 0.040658626705408096,
+ "learning_rate": 0.0006,
+ "loss": 6.639796257019043,
+ "step": 3205
+ },
+ {
+ "epoch": 44.531004366812226,
+ "grad_norm": 0.03691193461418152,
+ "learning_rate": 0.0006,
+ "loss": 6.692244529724121,
+ "step": 3206
+ },
+ {
+ "epoch": 44.544978165938865,
+ "grad_norm": 0.02814812771975994,
+ "learning_rate": 0.0006,
+ "loss": 6.61978006362915,
+ "step": 3207
+ },
+ {
+ "epoch": 44.5589519650655,
+ "grad_norm": 0.022155074402689934,
+ "learning_rate": 0.0006,
+ "loss": 6.625905990600586,
+ "step": 3208
+ },
+ {
+ "epoch": 44.57292576419214,
+ "grad_norm": 0.030794885009527206,
+ "learning_rate": 0.0006,
+ "loss": 6.678500652313232,
+ "step": 3209
+ },
+ {
+ "epoch": 44.58689956331878,
+ "grad_norm": 0.03201768547296524,
+ "learning_rate": 0.0006,
+ "loss": 6.6015729904174805,
+ "step": 3210
+ },
+ {
+ "epoch": 44.60087336244541,
+ "grad_norm": 0.025138361379504204,
+ "learning_rate": 0.0006,
+ "loss": 6.568234443664551,
+ "step": 3211
+ },
+ {
+ "epoch": 44.61484716157205,
+ "grad_norm": 0.02969353087246418,
+ "learning_rate": 0.0006,
+ "loss": 6.588366508483887,
+ "step": 3212
+ },
+ {
+ "epoch": 44.62882096069869,
+ "grad_norm": 0.03136962279677391,
+ "learning_rate": 0.0006,
+ "loss": 6.479961395263672,
+ "step": 3213
+ },
+ {
+ "epoch": 44.64279475982533,
+ "grad_norm": 0.030178170651197433,
+ "learning_rate": 0.0006,
+ "loss": 6.622910976409912,
+ "step": 3214
+ },
+ {
+ "epoch": 44.65676855895197,
+ "grad_norm": 0.02682245895266533,
+ "learning_rate": 0.0006,
+ "loss": 6.597707748413086,
+ "step": 3215
+ },
+ {
+ "epoch": 44.670742358078606,
+ "grad_norm": 0.025154557079076767,
+ "learning_rate": 0.0006,
+ "loss": 6.630541801452637,
+ "step": 3216
+ },
+ {
+ "epoch": 44.68471615720524,
+ "grad_norm": 0.03564463183283806,
+ "learning_rate": 0.0006,
+ "loss": 6.542823791503906,
+ "step": 3217
+ },
+ {
+ "epoch": 44.698689956331876,
+ "grad_norm": 0.049477651715278625,
+ "learning_rate": 0.0006,
+ "loss": 6.646030426025391,
+ "step": 3218
+ },
+ {
+ "epoch": 44.712663755458514,
+ "grad_norm": 0.06536535918712616,
+ "learning_rate": 0.0006,
+ "loss": 6.595463275909424,
+ "step": 3219
+ },
+ {
+ "epoch": 44.72663755458515,
+ "grad_norm": 0.07710316777229309,
+ "learning_rate": 0.0006,
+ "loss": 6.55524206161499,
+ "step": 3220
+ },
+ {
+ "epoch": 44.74061135371179,
+ "grad_norm": 0.09571163356304169,
+ "learning_rate": 0.0006,
+ "loss": 6.659764289855957,
+ "step": 3221
+ },
+ {
+ "epoch": 44.75458515283843,
+ "grad_norm": 0.055076830089092255,
+ "learning_rate": 0.0006,
+ "loss": 6.658882141113281,
+ "step": 3222
+ },
+ {
+ "epoch": 44.76855895196506,
+ "grad_norm": 0.07671099156141281,
+ "learning_rate": 0.0006,
+ "loss": 6.605809211730957,
+ "step": 3223
+ },
+ {
+ "epoch": 44.7825327510917,
+ "grad_norm": 0.05295584350824356,
+ "learning_rate": 0.0006,
+ "loss": 6.641455173492432,
+ "step": 3224
+ },
+ {
+ "epoch": 44.79650655021834,
+ "grad_norm": 0.09128042310476303,
+ "learning_rate": 0.0006,
+ "loss": 6.691066741943359,
+ "step": 3225
+ },
+ {
+ "epoch": 44.81048034934498,
+ "grad_norm": 0.05362958461046219,
+ "learning_rate": 0.0006,
+ "loss": 6.573406219482422,
+ "step": 3226
+ },
+ {
+ "epoch": 44.82445414847162,
+ "grad_norm": 0.046910062432289124,
+ "learning_rate": 0.0006,
+ "loss": 6.550876617431641,
+ "step": 3227
+ },
+ {
+ "epoch": 44.838427947598255,
+ "grad_norm": 0.03494197130203247,
+ "learning_rate": 0.0006,
+ "loss": 6.540179252624512,
+ "step": 3228
+ },
+ {
+ "epoch": 44.852401746724894,
+ "grad_norm": 0.04716387391090393,
+ "learning_rate": 0.0006,
+ "loss": 6.490459442138672,
+ "step": 3229
+ },
+ {
+ "epoch": 44.866375545851525,
+ "grad_norm": 0.04190506786108017,
+ "learning_rate": 0.0006,
+ "loss": 6.509925365447998,
+ "step": 3230
+ },
+ {
+ "epoch": 44.880349344978164,
+ "grad_norm": 0.04749064892530441,
+ "learning_rate": 0.0006,
+ "loss": 6.478779315948486,
+ "step": 3231
+ },
+ {
+ "epoch": 44.8943231441048,
+ "grad_norm": 0.028002185747027397,
+ "learning_rate": 0.0006,
+ "loss": 6.528904914855957,
+ "step": 3232
+ },
+ {
+ "epoch": 44.90829694323144,
+ "grad_norm": 0.036473166197538376,
+ "learning_rate": 0.0006,
+ "loss": 6.549042701721191,
+ "step": 3233
+ },
+ {
+ "epoch": 44.92227074235808,
+ "grad_norm": 0.02543315477669239,
+ "learning_rate": 0.0006,
+ "loss": 6.498875141143799,
+ "step": 3234
+ },
+ {
+ "epoch": 44.93624454148472,
+ "grad_norm": 0.029906807467341423,
+ "learning_rate": 0.0006,
+ "loss": 6.453019142150879,
+ "step": 3235
+ },
+ {
+ "epoch": 44.95021834061135,
+ "grad_norm": 0.027231857180595398,
+ "learning_rate": 0.0006,
+ "loss": 6.428829193115234,
+ "step": 3236
+ },
+ {
+ "epoch": 44.96419213973799,
+ "grad_norm": 0.02506154775619507,
+ "learning_rate": 0.0006,
+ "loss": 6.420858383178711,
+ "step": 3237
+ },
+ {
+ "epoch": 44.97816593886463,
+ "grad_norm": 0.0269472673535347,
+ "learning_rate": 0.0006,
+ "loss": 6.484989643096924,
+ "step": 3238
+ },
+ {
+ "epoch": 44.992139737991266,
+ "grad_norm": 0.023410487920045853,
+ "learning_rate": 0.0006,
+ "loss": 6.454830169677734,
+ "step": 3239
+ },
+ {
+ "epoch": 45.0,
+ "grad_norm": 0.034175023436546326,
+ "learning_rate": 0.0006,
+ "loss": 6.430975437164307,
+ "step": 3240
+ },
+ {
+ "epoch": 45.0,
+ "eval_loss": 6.458896160125732,
+ "eval_runtime": 56.6934,
+ "eval_samples_per_second": 43.074,
+ "eval_steps_per_second": 1.358,
+ "step": 3240
+ },
+ {
+ "epoch": 45.01397379912664,
+ "grad_norm": 0.045031629502773285,
+ "learning_rate": 0.0006,
+ "loss": 6.4243011474609375,
+ "step": 3241
+ },
+ {
+ "epoch": 45.02794759825328,
+ "grad_norm": 0.04968734085559845,
+ "learning_rate": 0.0006,
+ "loss": 6.380091667175293,
+ "step": 3242
+ },
+ {
+ "epoch": 45.041921397379916,
+ "grad_norm": 0.047145187854766846,
+ "learning_rate": 0.0006,
+ "loss": 6.477178573608398,
+ "step": 3243
+ },
+ {
+ "epoch": 45.05589519650655,
+ "grad_norm": 0.04304948449134827,
+ "learning_rate": 0.0006,
+ "loss": 6.399726867675781,
+ "step": 3244
+ },
+ {
+ "epoch": 45.069868995633186,
+ "grad_norm": 0.04466010630130768,
+ "learning_rate": 0.0006,
+ "loss": 6.3415679931640625,
+ "step": 3245
+ },
+ {
+ "epoch": 45.083842794759825,
+ "grad_norm": 0.04305306077003479,
+ "learning_rate": 0.0006,
+ "loss": 6.3413310050964355,
+ "step": 3246
+ },
+ {
+ "epoch": 45.09781659388646,
+ "grad_norm": 0.043064575642347336,
+ "learning_rate": 0.0006,
+ "loss": 6.4838714599609375,
+ "step": 3247
+ },
+ {
+ "epoch": 45.1117903930131,
+ "grad_norm": 0.049503277987241745,
+ "learning_rate": 0.0006,
+ "loss": 6.38473653793335,
+ "step": 3248
+ },
+ {
+ "epoch": 45.12576419213974,
+ "grad_norm": 0.04613238573074341,
+ "learning_rate": 0.0006,
+ "loss": 6.370769500732422,
+ "step": 3249
+ },
+ {
+ "epoch": 45.13973799126637,
+ "grad_norm": 0.052119478583335876,
+ "learning_rate": 0.0006,
+ "loss": 6.362321853637695,
+ "step": 3250
+ },
+ {
+ "epoch": 45.15371179039301,
+ "grad_norm": 0.045110076665878296,
+ "learning_rate": 0.0006,
+ "loss": 6.369844913482666,
+ "step": 3251
+ },
+ {
+ "epoch": 45.16768558951965,
+ "grad_norm": 0.03798583149909973,
+ "learning_rate": 0.0006,
+ "loss": 6.324633598327637,
+ "step": 3252
+ },
+ {
+ "epoch": 45.18165938864629,
+ "grad_norm": 0.03744659945368767,
+ "learning_rate": 0.0006,
+ "loss": 6.349421501159668,
+ "step": 3253
+ },
+ {
+ "epoch": 45.19563318777293,
+ "grad_norm": 0.03767292574048042,
+ "learning_rate": 0.0006,
+ "loss": 6.334238052368164,
+ "step": 3254
+ },
+ {
+ "epoch": 45.209606986899566,
+ "grad_norm": 0.03577509894967079,
+ "learning_rate": 0.0006,
+ "loss": 6.3453826904296875,
+ "step": 3255
+ },
+ {
+ "epoch": 45.223580786026204,
+ "grad_norm": 0.03403240442276001,
+ "learning_rate": 0.0006,
+ "loss": 6.2836174964904785,
+ "step": 3256
+ },
+ {
+ "epoch": 45.237554585152836,
+ "grad_norm": 0.03517331928014755,
+ "learning_rate": 0.0006,
+ "loss": 6.281718730926514,
+ "step": 3257
+ },
+ {
+ "epoch": 45.251528384279474,
+ "grad_norm": 0.04409139230847359,
+ "learning_rate": 0.0006,
+ "loss": 6.320815086364746,
+ "step": 3258
+ },
+ {
+ "epoch": 45.26550218340611,
+ "grad_norm": 0.054750364273786545,
+ "learning_rate": 0.0006,
+ "loss": 6.343618392944336,
+ "step": 3259
+ },
+ {
+ "epoch": 45.27947598253275,
+ "grad_norm": 0.07612067461013794,
+ "learning_rate": 0.0006,
+ "loss": 6.310614585876465,
+ "step": 3260
+ },
+ {
+ "epoch": 45.29344978165939,
+ "grad_norm": 0.08438985049724579,
+ "learning_rate": 0.0006,
+ "loss": 6.2369184494018555,
+ "step": 3261
+ },
+ {
+ "epoch": 45.30742358078603,
+ "grad_norm": 0.0782201811671257,
+ "learning_rate": 0.0006,
+ "loss": 6.254452705383301,
+ "step": 3262
+ },
+ {
+ "epoch": 45.32139737991266,
+ "grad_norm": 0.09928355365991592,
+ "learning_rate": 0.0006,
+ "loss": 6.3112077713012695,
+ "step": 3263
+ },
+ {
+ "epoch": 45.3353711790393,
+ "grad_norm": 0.07233133912086487,
+ "learning_rate": 0.0006,
+ "loss": 6.408087253570557,
+ "step": 3264
+ },
+ {
+ "epoch": 45.34934497816594,
+ "grad_norm": 0.07478684186935425,
+ "learning_rate": 0.0006,
+ "loss": 6.3770036697387695,
+ "step": 3265
+ },
+ {
+ "epoch": 45.36331877729258,
+ "grad_norm": 0.06036846712231636,
+ "learning_rate": 0.0006,
+ "loss": 6.33865213394165,
+ "step": 3266
+ },
+ {
+ "epoch": 45.377292576419215,
+ "grad_norm": 0.05783458054065704,
+ "learning_rate": 0.0006,
+ "loss": 6.363203048706055,
+ "step": 3267
+ },
+ {
+ "epoch": 45.391266375545854,
+ "grad_norm": 0.05443909019231796,
+ "learning_rate": 0.0006,
+ "loss": 6.287069797515869,
+ "step": 3268
+ },
+ {
+ "epoch": 45.40524017467249,
+ "grad_norm": 0.0557909794151783,
+ "learning_rate": 0.0006,
+ "loss": 6.297784805297852,
+ "step": 3269
+ },
+ {
+ "epoch": 45.419213973799124,
+ "grad_norm": 0.03483372926712036,
+ "learning_rate": 0.0006,
+ "loss": 6.280734062194824,
+ "step": 3270
+ },
+ {
+ "epoch": 45.43318777292576,
+ "grad_norm": 0.0425543375313282,
+ "learning_rate": 0.0006,
+ "loss": 6.33565616607666,
+ "step": 3271
+ },
+ {
+ "epoch": 45.4471615720524,
+ "grad_norm": 0.03230159357190132,
+ "learning_rate": 0.0006,
+ "loss": 6.197698593139648,
+ "step": 3272
+ },
+ {
+ "epoch": 45.46113537117904,
+ "grad_norm": 0.036897290498018265,
+ "learning_rate": 0.0006,
+ "loss": 6.241239547729492,
+ "step": 3273
+ },
+ {
+ "epoch": 45.47510917030568,
+ "grad_norm": 0.031103838235139847,
+ "learning_rate": 0.0006,
+ "loss": 6.159435749053955,
+ "step": 3274
+ },
+ {
+ "epoch": 45.48908296943232,
+ "grad_norm": 0.027116449549794197,
+ "learning_rate": 0.0006,
+ "loss": 6.212225914001465,
+ "step": 3275
+ },
+ {
+ "epoch": 45.50305676855895,
+ "grad_norm": 0.029923981055617332,
+ "learning_rate": 0.0006,
+ "loss": 6.157269477844238,
+ "step": 3276
+ },
+ {
+ "epoch": 45.51703056768559,
+ "grad_norm": 0.02521538734436035,
+ "learning_rate": 0.0006,
+ "loss": 6.146918773651123,
+ "step": 3277
+ },
+ {
+ "epoch": 45.531004366812226,
+ "grad_norm": 0.02758920192718506,
+ "learning_rate": 0.0006,
+ "loss": 6.0446977615356445,
+ "step": 3278
+ },
+ {
+ "epoch": 45.544978165938865,
+ "grad_norm": 0.024328220635652542,
+ "learning_rate": 0.0006,
+ "loss": 6.099508285522461,
+ "step": 3279
+ },
+ {
+ "epoch": 45.5589519650655,
+ "grad_norm": 0.027425771579146385,
+ "learning_rate": 0.0006,
+ "loss": 6.084299087524414,
+ "step": 3280
+ },
+ {
+ "epoch": 45.57292576419214,
+ "grad_norm": 0.026927651837468147,
+ "learning_rate": 0.0006,
+ "loss": 6.146334171295166,
+ "step": 3281
+ },
+ {
+ "epoch": 45.58689956331878,
+ "grad_norm": 0.04054379090666771,
+ "learning_rate": 0.0006,
+ "loss": 6.1301493644714355,
+ "step": 3282
+ },
+ {
+ "epoch": 45.60087336244541,
+ "grad_norm": 0.04456290975213051,
+ "learning_rate": 0.0006,
+ "loss": 6.152511119842529,
+ "step": 3283
+ },
+ {
+ "epoch": 45.61484716157205,
+ "grad_norm": 0.052821505814790726,
+ "learning_rate": 0.0006,
+ "loss": 6.113544940948486,
+ "step": 3284
+ },
+ {
+ "epoch": 45.62882096069869,
+ "grad_norm": 0.07251739501953125,
+ "learning_rate": 0.0006,
+ "loss": 6.118298530578613,
+ "step": 3285
+ },
+ {
+ "epoch": 45.64279475982533,
+ "grad_norm": 0.09433556348085403,
+ "learning_rate": 0.0006,
+ "loss": 6.170897483825684,
+ "step": 3286
+ },
+ {
+ "epoch": 45.65676855895197,
+ "grad_norm": 0.057683493942022324,
+ "learning_rate": 0.0006,
+ "loss": 6.1747822761535645,
+ "step": 3287
+ },
+ {
+ "epoch": 45.670742358078606,
+ "grad_norm": 0.06363270431756973,
+ "learning_rate": 0.0006,
+ "loss": 6.1617889404296875,
+ "step": 3288
+ },
+ {
+ "epoch": 45.68471615720524,
+ "grad_norm": 0.07528169453144073,
+ "learning_rate": 0.0006,
+ "loss": 6.230060577392578,
+ "step": 3289
+ },
+ {
+ "epoch": 45.698689956331876,
+ "grad_norm": 0.07816170156002045,
+ "learning_rate": 0.0006,
+ "loss": 6.171257019042969,
+ "step": 3290
+ },
+ {
+ "epoch": 45.712663755458514,
+ "grad_norm": 0.05860164389014244,
+ "learning_rate": 0.0006,
+ "loss": 6.095484256744385,
+ "step": 3291
+ },
+ {
+ "epoch": 45.72663755458515,
+ "grad_norm": 0.04747429117560387,
+ "learning_rate": 0.0006,
+ "loss": 6.060067176818848,
+ "step": 3292
+ },
+ {
+ "epoch": 45.74061135371179,
+ "grad_norm": 0.04514474794268608,
+ "learning_rate": 0.0006,
+ "loss": 6.100218772888184,
+ "step": 3293
+ },
+ {
+ "epoch": 45.75458515283843,
+ "grad_norm": 0.05168681964278221,
+ "learning_rate": 0.0006,
+ "loss": 6.0621867179870605,
+ "step": 3294
+ },
+ {
+ "epoch": 45.76855895196506,
+ "grad_norm": 0.04665876924991608,
+ "learning_rate": 0.0006,
+ "loss": 6.118292808532715,
+ "step": 3295
+ },
+ {
+ "epoch": 45.7825327510917,
+ "grad_norm": 0.049374762922525406,
+ "learning_rate": 0.0006,
+ "loss": 6.098394393920898,
+ "step": 3296
+ },
+ {
+ "epoch": 45.79650655021834,
+ "grad_norm": 0.03793549910187721,
+ "learning_rate": 0.0006,
+ "loss": 6.069541931152344,
+ "step": 3297
+ },
+ {
+ "epoch": 45.81048034934498,
+ "grad_norm": 0.036362383514642715,
+ "learning_rate": 0.0006,
+ "loss": 6.1188435554504395,
+ "step": 3298
+ },
+ {
+ "epoch": 45.82445414847162,
+ "grad_norm": 0.03397097811102867,
+ "learning_rate": 0.0006,
+ "loss": 6.117589950561523,
+ "step": 3299
+ },
+ {
+ "epoch": 45.838427947598255,
+ "grad_norm": 0.030977753922343254,
+ "learning_rate": 0.0006,
+ "loss": 6.0236711502075195,
+ "step": 3300
+ },
+ {
+ "epoch": 45.852401746724894,
+ "grad_norm": 0.02800217643380165,
+ "learning_rate": 0.0006,
+ "loss": 6.034825325012207,
+ "step": 3301
+ },
+ {
+ "epoch": 45.866375545851525,
+ "grad_norm": 0.03129778802394867,
+ "learning_rate": 0.0006,
+ "loss": 5.980271339416504,
+ "step": 3302
+ },
+ {
+ "epoch": 45.880349344978164,
+ "grad_norm": 0.025886094197630882,
+ "learning_rate": 0.0006,
+ "loss": 5.966419219970703,
+ "step": 3303
+ },
+ {
+ "epoch": 45.8943231441048,
+ "grad_norm": 0.025712328031659126,
+ "learning_rate": 0.0006,
+ "loss": 5.938652515411377,
+ "step": 3304
+ },
+ {
+ "epoch": 45.90829694323144,
+ "grad_norm": 0.022508298978209496,
+ "learning_rate": 0.0006,
+ "loss": 5.982312202453613,
+ "step": 3305
+ },
+ {
+ "epoch": 45.92227074235808,
+ "grad_norm": 0.024880951270461082,
+ "learning_rate": 0.0006,
+ "loss": 5.931150436401367,
+ "step": 3306
+ },
+ {
+ "epoch": 45.93624454148472,
+ "grad_norm": 0.030007576569914818,
+ "learning_rate": 0.0006,
+ "loss": 5.936639308929443,
+ "step": 3307
+ },
+ {
+ "epoch": 45.95021834061135,
+ "grad_norm": 0.033314041793346405,
+ "learning_rate": 0.0006,
+ "loss": 5.944248676300049,
+ "step": 3308
+ },
+ {
+ "epoch": 45.96419213973799,
+ "grad_norm": 0.043078381568193436,
+ "learning_rate": 0.0006,
+ "loss": 5.971286773681641,
+ "step": 3309
+ },
+ {
+ "epoch": 45.97816593886463,
+ "grad_norm": 0.059296440333127975,
+ "learning_rate": 0.0006,
+ "loss": 5.910497665405273,
+ "step": 3310
+ },
+ {
+ "epoch": 45.992139737991266,
+ "grad_norm": 0.06232666224241257,
+ "learning_rate": 0.0006,
+ "loss": 5.929817199707031,
+ "step": 3311
+ },
+ {
+ "epoch": 46.0,
+ "grad_norm": 0.05303940549492836,
+ "learning_rate": 0.0006,
+ "loss": 5.910822868347168,
+ "step": 3312
+ },
+ {
+ "epoch": 46.0,
+ "eval_loss": 5.977851390838623,
+ "eval_runtime": 56.2435,
+ "eval_samples_per_second": 43.418,
+ "eval_steps_per_second": 1.369,
+ "step": 3312
+ },
+ {
+ "epoch": 46.01397379912664,
+ "grad_norm": 0.06943798810243607,
+ "learning_rate": 0.0006,
+ "loss": 5.854922294616699,
+ "step": 3313
+ },
+ {
+ "epoch": 46.02794759825328,
+ "grad_norm": 0.05236278846859932,
+ "learning_rate": 0.0006,
+ "loss": 5.876611232757568,
+ "step": 3314
+ },
+ {
+ "epoch": 46.041921397379916,
+ "grad_norm": 0.06240001320838928,
+ "learning_rate": 0.0006,
+ "loss": 5.878912448883057,
+ "step": 3315
+ },
+ {
+ "epoch": 46.05589519650655,
+ "grad_norm": 0.07401192933320999,
+ "learning_rate": 0.0006,
+ "loss": 5.985136032104492,
+ "step": 3316
+ },
+ {
+ "epoch": 46.069868995633186,
+ "grad_norm": 0.0826529711484909,
+ "learning_rate": 0.0006,
+ "loss": 5.939439296722412,
+ "step": 3317
+ },
+ {
+ "epoch": 46.083842794759825,
+ "grad_norm": 0.07999005168676376,
+ "learning_rate": 0.0006,
+ "loss": 5.982336044311523,
+ "step": 3318
+ },
+ {
+ "epoch": 46.09781659388646,
+ "grad_norm": 0.07014869153499603,
+ "learning_rate": 0.0006,
+ "loss": 5.896534442901611,
+ "step": 3319
+ },
+ {
+ "epoch": 46.1117903930131,
+ "grad_norm": 0.05795128643512726,
+ "learning_rate": 0.0006,
+ "loss": 5.843111991882324,
+ "step": 3320
+ },
+ {
+ "epoch": 46.12576419213974,
+ "grad_norm": 0.06374258548021317,
+ "learning_rate": 0.0006,
+ "loss": 5.78752326965332,
+ "step": 3321
+ },
+ {
+ "epoch": 46.13973799126637,
+ "grad_norm": 0.045524805784225464,
+ "learning_rate": 0.0006,
+ "loss": 5.86755895614624,
+ "step": 3322
+ },
+ {
+ "epoch": 46.15371179039301,
+ "grad_norm": 0.03773504123091698,
+ "learning_rate": 0.0006,
+ "loss": 5.83627462387085,
+ "step": 3323
+ },
+ {
+ "epoch": 46.16768558951965,
+ "grad_norm": 0.035815201699733734,
+ "learning_rate": 0.0006,
+ "loss": 5.798238277435303,
+ "step": 3324
+ },
+ {
+ "epoch": 46.18165938864629,
+ "grad_norm": 0.03756783902645111,
+ "learning_rate": 0.0006,
+ "loss": 5.845197677612305,
+ "step": 3325
+ },
+ {
+ "epoch": 46.19563318777293,
+ "grad_norm": 0.03569113463163376,
+ "learning_rate": 0.0006,
+ "loss": 5.888932704925537,
+ "step": 3326
+ },
+ {
+ "epoch": 46.209606986899566,
+ "grad_norm": 0.029467813670635223,
+ "learning_rate": 0.0006,
+ "loss": 5.792047500610352,
+ "step": 3327
+ },
+ {
+ "epoch": 46.223580786026204,
+ "grad_norm": 0.025698406621813774,
+ "learning_rate": 0.0006,
+ "loss": 5.783734321594238,
+ "step": 3328
+ },
+ {
+ "epoch": 46.237554585152836,
+ "grad_norm": 0.028426671400666237,
+ "learning_rate": 0.0006,
+ "loss": 5.762084007263184,
+ "step": 3329
+ },
+ {
+ "epoch": 46.251528384279474,
+ "grad_norm": 0.03099622018635273,
+ "learning_rate": 0.0006,
+ "loss": 5.719109058380127,
+ "step": 3330
+ },
+ {
+ "epoch": 46.26550218340611,
+ "grad_norm": 0.0310219656676054,
+ "learning_rate": 0.0006,
+ "loss": 5.737057209014893,
+ "step": 3331
+ },
+ {
+ "epoch": 46.27947598253275,
+ "grad_norm": 0.02713446132838726,
+ "learning_rate": 0.0006,
+ "loss": 5.758657455444336,
+ "step": 3332
+ },
+ {
+ "epoch": 46.29344978165939,
+ "grad_norm": 0.03099844604730606,
+ "learning_rate": 0.0006,
+ "loss": 5.742086410522461,
+ "step": 3333
+ },
+ {
+ "epoch": 46.30742358078603,
+ "grad_norm": 0.030258730053901672,
+ "learning_rate": 0.0006,
+ "loss": 5.767426490783691,
+ "step": 3334
+ },
+ {
+ "epoch": 46.32139737991266,
+ "grad_norm": 0.029034852981567383,
+ "learning_rate": 0.0006,
+ "loss": 5.651792049407959,
+ "step": 3335
+ },
+ {
+ "epoch": 46.3353711790393,
+ "grad_norm": 0.03138665854930878,
+ "learning_rate": 0.0006,
+ "loss": 5.664980888366699,
+ "step": 3336
+ },
+ {
+ "epoch": 46.34934497816594,
+ "grad_norm": 0.040546901524066925,
+ "learning_rate": 0.0006,
+ "loss": 5.665165901184082,
+ "step": 3337
+ },
+ {
+ "epoch": 46.36331877729258,
+ "grad_norm": 0.06395485252141953,
+ "learning_rate": 0.0006,
+ "loss": 5.7776713371276855,
+ "step": 3338
+ },
+ {
+ "epoch": 46.377292576419215,
+ "grad_norm": 0.10088056325912476,
+ "learning_rate": 0.0006,
+ "loss": 5.755645751953125,
+ "step": 3339
+ },
+ {
+ "epoch": 46.391266375545854,
+ "grad_norm": 0.09527938067913055,
+ "learning_rate": 0.0006,
+ "loss": 5.838809490203857,
+ "step": 3340
+ },
+ {
+ "epoch": 46.40524017467249,
+ "grad_norm": 0.0849648267030716,
+ "learning_rate": 0.0006,
+ "loss": 5.781932830810547,
+ "step": 3341
+ },
+ {
+ "epoch": 46.419213973799124,
+ "grad_norm": 0.10142666101455688,
+ "learning_rate": 0.0006,
+ "loss": 5.838970184326172,
+ "step": 3342
+ },
+ {
+ "epoch": 46.43318777292576,
+ "grad_norm": 0.42731454968452454,
+ "learning_rate": 0.0006,
+ "loss": 6.071902275085449,
+ "step": 3343
+ },
+ {
+ "epoch": 46.4471615720524,
+ "grad_norm": 0.13102488219738007,
+ "learning_rate": 0.0006,
+ "loss": 6.8635993003845215,
+ "step": 3344
+ },
+ {
+ "epoch": 46.46113537117904,
+ "grad_norm": 0.23691938817501068,
+ "learning_rate": 0.0006,
+ "loss": 7.313158988952637,
+ "step": 3345
+ },
+ {
+ "epoch": 46.47510917030568,
+ "grad_norm": 0.11670669168233871,
+ "learning_rate": 0.0006,
+ "loss": 7.240363121032715,
+ "step": 3346
+ },
+ {
+ "epoch": 46.48908296943232,
+ "grad_norm": 0.09985820949077606,
+ "learning_rate": 0.0006,
+ "loss": 7.113332271575928,
+ "step": 3347
+ },
+ {
+ "epoch": 46.50305676855895,
+ "grad_norm": 0.11406408250331879,
+ "learning_rate": 0.0006,
+ "loss": 7.051285743713379,
+ "step": 3348
+ },
+ {
+ "epoch": 46.51703056768559,
+ "grad_norm": 0.07900336384773254,
+ "learning_rate": 0.0006,
+ "loss": 7.014365196228027,
+ "step": 3349
+ },
+ {
+ "epoch": 46.531004366812226,
+ "grad_norm": 0.05380333960056305,
+ "learning_rate": 0.0006,
+ "loss": 6.8868818283081055,
+ "step": 3350
+ },
+ {
+ "epoch": 46.544978165938865,
+ "grad_norm": 0.05868449807167053,
+ "learning_rate": 0.0006,
+ "loss": 6.783599853515625,
+ "step": 3351
+ },
+ {
+ "epoch": 46.5589519650655,
+ "grad_norm": 0.04428785294294357,
+ "learning_rate": 0.0006,
+ "loss": 6.798511505126953,
+ "step": 3352
+ },
+ {
+ "epoch": 46.57292576419214,
+ "grad_norm": 0.055527277290821075,
+ "learning_rate": 0.0006,
+ "loss": 6.694507598876953,
+ "step": 3353
+ },
+ {
+ "epoch": 46.58689956331878,
+ "grad_norm": 0.03872432932257652,
+ "learning_rate": 0.0006,
+ "loss": 6.736631393432617,
+ "step": 3354
+ },
+ {
+ "epoch": 46.60087336244541,
+ "grad_norm": 0.04330779239535332,
+ "learning_rate": 0.0006,
+ "loss": 6.576806545257568,
+ "step": 3355
+ },
+ {
+ "epoch": 46.61484716157205,
+ "grad_norm": 0.036767736077308655,
+ "learning_rate": 0.0006,
+ "loss": 6.559871196746826,
+ "step": 3356
+ },
+ {
+ "epoch": 46.62882096069869,
+ "grad_norm": 0.039857108145952225,
+ "learning_rate": 0.0006,
+ "loss": 6.5749831199646,
+ "step": 3357
+ },
+ {
+ "epoch": 46.64279475982533,
+ "grad_norm": 0.048719972372055054,
+ "learning_rate": 0.0006,
+ "loss": 6.4915618896484375,
+ "step": 3358
+ },
+ {
+ "epoch": 46.65676855895197,
+ "grad_norm": 0.042849015444517136,
+ "learning_rate": 0.0006,
+ "loss": 6.413516044616699,
+ "step": 3359
+ },
+ {
+ "epoch": 46.670742358078606,
+ "grad_norm": 0.02871009334921837,
+ "learning_rate": 0.0006,
+ "loss": 6.320952415466309,
+ "step": 3360
+ },
+ {
+ "epoch": 46.68471615720524,
+ "grad_norm": 0.03193620592355728,
+ "learning_rate": 0.0006,
+ "loss": 6.4176435470581055,
+ "step": 3361
+ },
+ {
+ "epoch": 46.698689956331876,
+ "grad_norm": 0.027417024597525597,
+ "learning_rate": 0.0006,
+ "loss": 6.342319488525391,
+ "step": 3362
+ },
+ {
+ "epoch": 46.712663755458514,
+ "grad_norm": 0.03305937349796295,
+ "learning_rate": 0.0006,
+ "loss": 6.246112823486328,
+ "step": 3363
+ },
+ {
+ "epoch": 46.72663755458515,
+ "grad_norm": 0.031403183937072754,
+ "learning_rate": 0.0006,
+ "loss": 6.149503231048584,
+ "step": 3364
+ },
+ {
+ "epoch": 46.74061135371179,
+ "grad_norm": 0.031563788652420044,
+ "learning_rate": 0.0006,
+ "loss": 6.278808116912842,
+ "step": 3365
+ },
+ {
+ "epoch": 46.75458515283843,
+ "grad_norm": 0.028366589918732643,
+ "learning_rate": 0.0006,
+ "loss": 6.057036876678467,
+ "step": 3366
+ },
+ {
+ "epoch": 46.76855895196506,
+ "grad_norm": 0.023715803399682045,
+ "learning_rate": 0.0006,
+ "loss": 6.111324310302734,
+ "step": 3367
+ },
+ {
+ "epoch": 46.7825327510917,
+ "grad_norm": 0.026546811684966087,
+ "learning_rate": 0.0006,
+ "loss": 6.105669975280762,
+ "step": 3368
+ },
+ {
+ "epoch": 46.79650655021834,
+ "grad_norm": 0.023531019687652588,
+ "learning_rate": 0.0006,
+ "loss": 6.068072319030762,
+ "step": 3369
+ },
+ {
+ "epoch": 46.81048034934498,
+ "grad_norm": 0.024480128660798073,
+ "learning_rate": 0.0006,
+ "loss": 5.947896957397461,
+ "step": 3370
+ },
+ {
+ "epoch": 46.82445414847162,
+ "grad_norm": 0.02517310529947281,
+ "learning_rate": 0.0006,
+ "loss": 6.030037879943848,
+ "step": 3371
+ },
+ {
+ "epoch": 46.838427947598255,
+ "grad_norm": 0.026723403483629227,
+ "learning_rate": 0.0006,
+ "loss": 5.963776111602783,
+ "step": 3372
+ },
+ {
+ "epoch": 46.852401746724894,
+ "grad_norm": 0.03163504600524902,
+ "learning_rate": 0.0006,
+ "loss": 5.893466949462891,
+ "step": 3373
+ },
+ {
+ "epoch": 46.866375545851525,
+ "grad_norm": 0.04058683291077614,
+ "learning_rate": 0.0006,
+ "loss": 5.9202799797058105,
+ "step": 3374
+ },
+ {
+ "epoch": 46.880349344978164,
+ "grad_norm": 0.04964117333292961,
+ "learning_rate": 0.0006,
+ "loss": 5.9366559982299805,
+ "step": 3375
+ },
+ {
+ "epoch": 46.8943231441048,
+ "grad_norm": 0.055406033992767334,
+ "learning_rate": 0.0006,
+ "loss": 5.912230968475342,
+ "step": 3376
+ },
+ {
+ "epoch": 46.90829694323144,
+ "grad_norm": 0.04643625393509865,
+ "learning_rate": 0.0006,
+ "loss": 5.87260103225708,
+ "step": 3377
+ },
+ {
+ "epoch": 46.92227074235808,
+ "grad_norm": 0.029325876384973526,
+ "learning_rate": 0.0006,
+ "loss": 5.854763984680176,
+ "step": 3378
+ },
+ {
+ "epoch": 46.93624454148472,
+ "grad_norm": 0.03159382566809654,
+ "learning_rate": 0.0006,
+ "loss": 5.880882263183594,
+ "step": 3379
+ },
+ {
+ "epoch": 46.95021834061135,
+ "grad_norm": 0.030224012210965157,
+ "learning_rate": 0.0006,
+ "loss": 5.787360191345215,
+ "step": 3380
+ },
+ {
+ "epoch": 46.96419213973799,
+ "grad_norm": 0.028368130326271057,
+ "learning_rate": 0.0006,
+ "loss": 5.817195892333984,
+ "step": 3381
+ },
+ {
+ "epoch": 46.97816593886463,
+ "grad_norm": 0.029411399737000465,
+ "learning_rate": 0.0006,
+ "loss": 5.783403396606445,
+ "step": 3382
+ },
+ {
+ "epoch": 46.992139737991266,
+ "grad_norm": 0.029927462339401245,
+ "learning_rate": 0.0006,
+ "loss": 5.662444591522217,
+ "step": 3383
+ },
+ {
+ "epoch": 47.0,
+ "grad_norm": 0.02872961387038231,
+ "learning_rate": 0.0006,
+ "loss": 5.742781639099121,
+ "step": 3384
+ },
+ {
+ "epoch": 47.0,
+ "eval_loss": 5.870260238647461,
+ "eval_runtime": 57.3734,
+ "eval_samples_per_second": 42.563,
+ "eval_steps_per_second": 1.342,
+ "step": 3384
+ },
+ {
+ "epoch": 47.01397379912664,
+ "grad_norm": 0.028542131185531616,
+ "learning_rate": 0.0006,
+ "loss": 5.753256797790527,
+ "step": 3385
+ },
+ {
+ "epoch": 47.02794759825328,
+ "grad_norm": 0.03227119892835617,
+ "learning_rate": 0.0006,
+ "loss": 5.722246170043945,
+ "step": 3386
+ },
+ {
+ "epoch": 47.041921397379916,
+ "grad_norm": 0.031461723148822784,
+ "learning_rate": 0.0006,
+ "loss": 5.740520000457764,
+ "step": 3387
+ },
+ {
+ "epoch": 47.05589519650655,
+ "grad_norm": 0.03184220567345619,
+ "learning_rate": 0.0006,
+ "loss": 5.720890045166016,
+ "step": 3388
+ },
+ {
+ "epoch": 47.069868995633186,
+ "grad_norm": 0.030015580356121063,
+ "learning_rate": 0.0006,
+ "loss": 5.597814559936523,
+ "step": 3389
+ },
+ {
+ "epoch": 47.083842794759825,
+ "grad_norm": 0.029931779950857162,
+ "learning_rate": 0.0006,
+ "loss": 5.5412468910217285,
+ "step": 3390
+ },
+ {
+ "epoch": 47.09781659388646,
+ "grad_norm": 0.03206505626440048,
+ "learning_rate": 0.0006,
+ "loss": 5.6967597007751465,
+ "step": 3391
+ },
+ {
+ "epoch": 47.1117903930131,
+ "grad_norm": 0.03978161886334419,
+ "learning_rate": 0.0006,
+ "loss": 5.59661340713501,
+ "step": 3392
+ },
+ {
+ "epoch": 47.12576419213974,
+ "grad_norm": 0.04619080200791359,
+ "learning_rate": 0.0006,
+ "loss": 5.795788764953613,
+ "step": 3393
+ },
+ {
+ "epoch": 47.13973799126637,
+ "grad_norm": 0.049085382372140884,
+ "learning_rate": 0.0006,
+ "loss": 5.667335510253906,
+ "step": 3394
+ },
+ {
+ "epoch": 47.15371179039301,
+ "grad_norm": 0.04830858111381531,
+ "learning_rate": 0.0006,
+ "loss": 5.668806552886963,
+ "step": 3395
+ },
+ {
+ "epoch": 47.16768558951965,
+ "grad_norm": 0.0583178736269474,
+ "learning_rate": 0.0006,
+ "loss": 5.692285537719727,
+ "step": 3396
+ },
+ {
+ "epoch": 47.18165938864629,
+ "grad_norm": 0.05182253196835518,
+ "learning_rate": 0.0006,
+ "loss": 5.678645133972168,
+ "step": 3397
+ },
+ {
+ "epoch": 47.19563318777293,
+ "grad_norm": 0.04011905938386917,
+ "learning_rate": 0.0006,
+ "loss": 5.661550998687744,
+ "step": 3398
+ },
+ {
+ "epoch": 47.209606986899566,
+ "grad_norm": 0.03888486698269844,
+ "learning_rate": 0.0006,
+ "loss": 5.643001556396484,
+ "step": 3399
+ },
+ {
+ "epoch": 47.223580786026204,
+ "grad_norm": 0.030658265575766563,
+ "learning_rate": 0.0006,
+ "loss": 5.659090995788574,
+ "step": 3400
+ },
+ {
+ "epoch": 47.237554585152836,
+ "grad_norm": 0.03609256446361542,
+ "learning_rate": 0.0006,
+ "loss": 5.544501781463623,
+ "step": 3401
+ },
+ {
+ "epoch": 47.251528384279474,
+ "grad_norm": 0.033589769154787064,
+ "learning_rate": 0.0006,
+ "loss": 5.463878631591797,
+ "step": 3402
+ },
+ {
+ "epoch": 47.26550218340611,
+ "grad_norm": 0.035128287971019745,
+ "learning_rate": 0.0006,
+ "loss": 5.5680060386657715,
+ "step": 3403
+ },
+ {
+ "epoch": 47.27947598253275,
+ "grad_norm": 0.038684550672769547,
+ "learning_rate": 0.0006,
+ "loss": 5.587898254394531,
+ "step": 3404
+ },
+ {
+ "epoch": 47.29344978165939,
+ "grad_norm": 0.043566226959228516,
+ "learning_rate": 0.0006,
+ "loss": 5.580896854400635,
+ "step": 3405
+ },
+ {
+ "epoch": 47.30742358078603,
+ "grad_norm": 0.04428374394774437,
+ "learning_rate": 0.0006,
+ "loss": 5.603460311889648,
+ "step": 3406
+ },
+ {
+ "epoch": 47.32139737991266,
+ "grad_norm": 0.037368934601545334,
+ "learning_rate": 0.0006,
+ "loss": 5.564740180969238,
+ "step": 3407
+ },
+ {
+ "epoch": 47.3353711790393,
+ "grad_norm": 0.04092350974678993,
+ "learning_rate": 0.0006,
+ "loss": 5.584153652191162,
+ "step": 3408
+ },
+ {
+ "epoch": 47.34934497816594,
+ "grad_norm": 0.045031506568193436,
+ "learning_rate": 0.0006,
+ "loss": 5.601186275482178,
+ "step": 3409
+ },
+ {
+ "epoch": 47.36331877729258,
+ "grad_norm": 0.041830409318208694,
+ "learning_rate": 0.0006,
+ "loss": 5.512788772583008,
+ "step": 3410
+ },
+ {
+ "epoch": 47.377292576419215,
+ "grad_norm": 0.037633974105119705,
+ "learning_rate": 0.0006,
+ "loss": 5.503828525543213,
+ "step": 3411
+ },
+ {
+ "epoch": 47.391266375545854,
+ "grad_norm": 0.03473588079214096,
+ "learning_rate": 0.0006,
+ "loss": 5.570557594299316,
+ "step": 3412
+ },
+ {
+ "epoch": 47.40524017467249,
+ "grad_norm": 0.039325181394815445,
+ "learning_rate": 0.0006,
+ "loss": 5.427783966064453,
+ "step": 3413
+ },
+ {
+ "epoch": 47.419213973799124,
+ "grad_norm": 0.04034624621272087,
+ "learning_rate": 0.0006,
+ "loss": 5.572591304779053,
+ "step": 3414
+ },
+ {
+ "epoch": 47.43318777292576,
+ "grad_norm": 0.05245184153318405,
+ "learning_rate": 0.0006,
+ "loss": 5.505208969116211,
+ "step": 3415
+ },
+ {
+ "epoch": 47.4471615720524,
+ "grad_norm": 0.0568675734102726,
+ "learning_rate": 0.0006,
+ "loss": 5.589373588562012,
+ "step": 3416
+ },
+ {
+ "epoch": 47.46113537117904,
+ "grad_norm": 0.048336587846279144,
+ "learning_rate": 0.0006,
+ "loss": 5.522005081176758,
+ "step": 3417
+ },
+ {
+ "epoch": 47.47510917030568,
+ "grad_norm": 0.04194975271821022,
+ "learning_rate": 0.0006,
+ "loss": 5.387576103210449,
+ "step": 3418
+ },
+ {
+ "epoch": 47.48908296943232,
+ "grad_norm": 0.04019433259963989,
+ "learning_rate": 0.0006,
+ "loss": 5.522615432739258,
+ "step": 3419
+ },
+ {
+ "epoch": 47.50305676855895,
+ "grad_norm": 0.05709933117032051,
+ "learning_rate": 0.0006,
+ "loss": 5.420286178588867,
+ "step": 3420
+ },
+ {
+ "epoch": 47.51703056768559,
+ "grad_norm": 0.09130338579416275,
+ "learning_rate": 0.0006,
+ "loss": 5.484622478485107,
+ "step": 3421
+ },
+ {
+ "epoch": 47.531004366812226,
+ "grad_norm": 0.2757541835308075,
+ "learning_rate": 0.0006,
+ "loss": 5.821832656860352,
+ "step": 3422
+ },
+ {
+ "epoch": 47.544978165938865,
+ "grad_norm": 0.28292080760002136,
+ "learning_rate": 0.0006,
+ "loss": 6.579934120178223,
+ "step": 3423
+ },
+ {
+ "epoch": 47.5589519650655,
+ "grad_norm": 0.15293604135513306,
+ "learning_rate": 0.0006,
+ "loss": 6.478213310241699,
+ "step": 3424
+ },
+ {
+ "epoch": 47.57292576419214,
+ "grad_norm": 0.11284393817186356,
+ "learning_rate": 0.0006,
+ "loss": 6.4066386222839355,
+ "step": 3425
+ },
+ {
+ "epoch": 47.58689956331878,
+ "grad_norm": 0.10933073610067368,
+ "learning_rate": 0.0006,
+ "loss": 6.3306169509887695,
+ "step": 3426
+ },
+ {
+ "epoch": 47.60087336244541,
+ "grad_norm": 0.05837641656398773,
+ "learning_rate": 0.0006,
+ "loss": 6.250000953674316,
+ "step": 3427
+ },
+ {
+ "epoch": 47.61484716157205,
+ "grad_norm": 0.05987229570746422,
+ "learning_rate": 0.0006,
+ "loss": 6.152463436126709,
+ "step": 3428
+ },
+ {
+ "epoch": 47.62882096069869,
+ "grad_norm": 0.04064938426017761,
+ "learning_rate": 0.0006,
+ "loss": 6.117897033691406,
+ "step": 3429
+ },
+ {
+ "epoch": 47.64279475982533,
+ "grad_norm": 0.04565057530999184,
+ "learning_rate": 0.0006,
+ "loss": 6.042291641235352,
+ "step": 3430
+ },
+ {
+ "epoch": 47.65676855895197,
+ "grad_norm": 0.03475063294172287,
+ "learning_rate": 0.0006,
+ "loss": 6.0071868896484375,
+ "step": 3431
+ },
+ {
+ "epoch": 47.670742358078606,
+ "grad_norm": 0.03809220343828201,
+ "learning_rate": 0.0006,
+ "loss": 6.0128021240234375,
+ "step": 3432
+ },
+ {
+ "epoch": 47.68471615720524,
+ "grad_norm": 0.03307219594717026,
+ "learning_rate": 0.0006,
+ "loss": 5.894664287567139,
+ "step": 3433
+ },
+ {
+ "epoch": 47.698689956331876,
+ "grad_norm": 0.028824346140027046,
+ "learning_rate": 0.0006,
+ "loss": 5.914865493774414,
+ "step": 3434
+ },
+ {
+ "epoch": 47.712663755458514,
+ "grad_norm": 0.027923690155148506,
+ "learning_rate": 0.0006,
+ "loss": 5.851759910583496,
+ "step": 3435
+ },
+ {
+ "epoch": 47.72663755458515,
+ "grad_norm": 0.025492684915661812,
+ "learning_rate": 0.0006,
+ "loss": 5.749186038970947,
+ "step": 3436
+ },
+ {
+ "epoch": 47.74061135371179,
+ "grad_norm": 0.025105174630880356,
+ "learning_rate": 0.0006,
+ "loss": 5.803566932678223,
+ "step": 3437
+ },
+ {
+ "epoch": 47.75458515283843,
+ "grad_norm": 0.023298880085349083,
+ "learning_rate": 0.0006,
+ "loss": 5.76276159286499,
+ "step": 3438
+ },
+ {
+ "epoch": 47.76855895196506,
+ "grad_norm": 0.02468576282262802,
+ "learning_rate": 0.0006,
+ "loss": 5.833852767944336,
+ "step": 3439
+ },
+ {
+ "epoch": 47.7825327510917,
+ "grad_norm": 0.022730689495801926,
+ "learning_rate": 0.0006,
+ "loss": 5.736724376678467,
+ "step": 3440
+ },
+ {
+ "epoch": 47.79650655021834,
+ "grad_norm": 0.021335354074835777,
+ "learning_rate": 0.0006,
+ "loss": 5.644248008728027,
+ "step": 3441
+ },
+ {
+ "epoch": 47.81048034934498,
+ "grad_norm": 0.02288101427257061,
+ "learning_rate": 0.0006,
+ "loss": 5.7011942863464355,
+ "step": 3442
+ },
+ {
+ "epoch": 47.82445414847162,
+ "grad_norm": 0.01961585320532322,
+ "learning_rate": 0.0006,
+ "loss": 5.63215446472168,
+ "step": 3443
+ },
+ {
+ "epoch": 47.838427947598255,
+ "grad_norm": 0.017659205943346024,
+ "learning_rate": 0.0006,
+ "loss": 5.744845390319824,
+ "step": 3444
+ },
+ {
+ "epoch": 47.852401746724894,
+ "grad_norm": 0.02281784638762474,
+ "learning_rate": 0.0006,
+ "loss": 5.696993350982666,
+ "step": 3445
+ },
+ {
+ "epoch": 47.866375545851525,
+ "grad_norm": 0.016394242644309998,
+ "learning_rate": 0.0006,
+ "loss": 5.659870147705078,
+ "step": 3446
+ },
+ {
+ "epoch": 47.880349344978164,
+ "grad_norm": 0.01967368647456169,
+ "learning_rate": 0.0006,
+ "loss": 5.638072490692139,
+ "step": 3447
+ },
+ {
+ "epoch": 47.8943231441048,
+ "grad_norm": 0.017326267436146736,
+ "learning_rate": 0.0006,
+ "loss": 5.638836860656738,
+ "step": 3448
+ },
+ {
+ "epoch": 47.90829694323144,
+ "grad_norm": 0.017778463661670685,
+ "learning_rate": 0.0006,
+ "loss": 5.634466648101807,
+ "step": 3449
+ },
+ {
+ "epoch": 47.92227074235808,
+ "grad_norm": 0.017269758507609367,
+ "learning_rate": 0.0006,
+ "loss": 5.506277561187744,
+ "step": 3450
+ },
+ {
+ "epoch": 47.93624454148472,
+ "grad_norm": 0.01733289286494255,
+ "learning_rate": 0.0006,
+ "loss": 5.544455528259277,
+ "step": 3451
+ },
+ {
+ "epoch": 47.95021834061135,
+ "grad_norm": 0.017787795513868332,
+ "learning_rate": 0.0006,
+ "loss": 5.498943328857422,
+ "step": 3452
+ },
+ {
+ "epoch": 47.96419213973799,
+ "grad_norm": 0.015770932659506798,
+ "learning_rate": 0.0006,
+ "loss": 5.537227630615234,
+ "step": 3453
+ },
+ {
+ "epoch": 47.97816593886463,
+ "grad_norm": 0.016132809221744537,
+ "learning_rate": 0.0006,
+ "loss": 5.489081382751465,
+ "step": 3454
+ },
+ {
+ "epoch": 47.992139737991266,
+ "grad_norm": 0.01749664731323719,
+ "learning_rate": 0.0006,
+ "loss": 5.454263687133789,
+ "step": 3455
+ },
+ {
+ "epoch": 48.0,
+ "grad_norm": 0.016711296513676643,
+ "learning_rate": 0.0006,
+ "loss": 5.530783653259277,
+ "step": 3456
+ },
+ {
+ "epoch": 48.0,
+ "eval_loss": 5.596925258636475,
+ "eval_runtime": 56.9612,
+ "eval_samples_per_second": 42.871,
+ "eval_steps_per_second": 1.352,
+ "step": 3456
+ },
+ {
+ "epoch": 48.01397379912664,
+ "grad_norm": 0.01867641694843769,
+ "learning_rate": 0.0006,
+ "loss": 5.506853103637695,
+ "step": 3457
+ },
+ {
+ "epoch": 48.02794759825328,
+ "grad_norm": 0.023250792175531387,
+ "learning_rate": 0.0006,
+ "loss": 5.4074177742004395,
+ "step": 3458
+ },
+ {
+ "epoch": 48.041921397379916,
+ "grad_norm": 0.01970936730504036,
+ "learning_rate": 0.0006,
+ "loss": 5.508777618408203,
+ "step": 3459
+ },
+ {
+ "epoch": 48.05589519650655,
+ "grad_norm": 0.01879260502755642,
+ "learning_rate": 0.0006,
+ "loss": 5.412424087524414,
+ "step": 3460
+ },
+ {
+ "epoch": 48.069868995633186,
+ "grad_norm": 0.017420092597603798,
+ "learning_rate": 0.0006,
+ "loss": 5.43361759185791,
+ "step": 3461
+ },
+ {
+ "epoch": 48.083842794759825,
+ "grad_norm": 0.018707862123847008,
+ "learning_rate": 0.0006,
+ "loss": 5.476800918579102,
+ "step": 3462
+ },
+ {
+ "epoch": 48.09781659388646,
+ "grad_norm": 0.023185882717370987,
+ "learning_rate": 0.0006,
+ "loss": 5.465272426605225,
+ "step": 3463
+ },
+ {
+ "epoch": 48.1117903930131,
+ "grad_norm": 0.023570368066430092,
+ "learning_rate": 0.0006,
+ "loss": 5.391471862792969,
+ "step": 3464
+ },
+ {
+ "epoch": 48.12576419213974,
+ "grad_norm": 0.02651144750416279,
+ "learning_rate": 0.0006,
+ "loss": 5.422799110412598,
+ "step": 3465
+ },
+ {
+ "epoch": 48.13973799126637,
+ "grad_norm": 0.03316329047083855,
+ "learning_rate": 0.0006,
+ "loss": 5.3952202796936035,
+ "step": 3466
+ },
+ {
+ "epoch": 48.15371179039301,
+ "grad_norm": 0.04012278467416763,
+ "learning_rate": 0.0006,
+ "loss": 5.540853023529053,
+ "step": 3467
+ },
+ {
+ "epoch": 48.16768558951965,
+ "grad_norm": 0.04763203114271164,
+ "learning_rate": 0.0006,
+ "loss": 5.387890338897705,
+ "step": 3468
+ },
+ {
+ "epoch": 48.18165938864629,
+ "grad_norm": 0.05129165202379227,
+ "learning_rate": 0.0006,
+ "loss": 5.414722442626953,
+ "step": 3469
+ },
+ {
+ "epoch": 48.19563318777293,
+ "grad_norm": 0.04438071325421333,
+ "learning_rate": 0.0006,
+ "loss": 5.386244773864746,
+ "step": 3470
+ },
+ {
+ "epoch": 48.209606986899566,
+ "grad_norm": 0.040234919637441635,
+ "learning_rate": 0.0006,
+ "loss": 5.431617736816406,
+ "step": 3471
+ },
+ {
+ "epoch": 48.223580786026204,
+ "grad_norm": 0.03217008337378502,
+ "learning_rate": 0.0006,
+ "loss": 5.30168342590332,
+ "step": 3472
+ },
+ {
+ "epoch": 48.237554585152836,
+ "grad_norm": 0.03087974339723587,
+ "learning_rate": 0.0006,
+ "loss": 5.316389083862305,
+ "step": 3473
+ },
+ {
+ "epoch": 48.251528384279474,
+ "grad_norm": 0.02789347991347313,
+ "learning_rate": 0.0006,
+ "loss": 5.385970592498779,
+ "step": 3474
+ },
+ {
+ "epoch": 48.26550218340611,
+ "grad_norm": 0.024792097508907318,
+ "learning_rate": 0.0006,
+ "loss": 5.362329959869385,
+ "step": 3475
+ },
+ {
+ "epoch": 48.27947598253275,
+ "grad_norm": 0.025869522243738174,
+ "learning_rate": 0.0006,
+ "loss": 5.28400182723999,
+ "step": 3476
+ },
+ {
+ "epoch": 48.29344978165939,
+ "grad_norm": 0.021339882165193558,
+ "learning_rate": 0.0006,
+ "loss": 5.36669921875,
+ "step": 3477
+ },
+ {
+ "epoch": 48.30742358078603,
+ "grad_norm": 0.021462520584464073,
+ "learning_rate": 0.0006,
+ "loss": 5.311222076416016,
+ "step": 3478
+ },
+ {
+ "epoch": 48.32139737991266,
+ "grad_norm": 0.02193879708647728,
+ "learning_rate": 0.0006,
+ "loss": 5.291441917419434,
+ "step": 3479
+ },
+ {
+ "epoch": 48.3353711790393,
+ "grad_norm": 0.01963093690574169,
+ "learning_rate": 0.0006,
+ "loss": 5.342780113220215,
+ "step": 3480
+ },
+ {
+ "epoch": 48.34934497816594,
+ "grad_norm": 0.020634736865758896,
+ "learning_rate": 0.0006,
+ "loss": 5.144141674041748,
+ "step": 3481
+ },
+ {
+ "epoch": 48.36331877729258,
+ "grad_norm": 0.02126968279480934,
+ "learning_rate": 0.0006,
+ "loss": 5.232624530792236,
+ "step": 3482
+ },
+ {
+ "epoch": 48.377292576419215,
+ "grad_norm": 0.021203994750976562,
+ "learning_rate": 0.0006,
+ "loss": 5.290219783782959,
+ "step": 3483
+ },
+ {
+ "epoch": 48.391266375545854,
+ "grad_norm": 0.02193225920200348,
+ "learning_rate": 0.0006,
+ "loss": 5.29197359085083,
+ "step": 3484
+ },
+ {
+ "epoch": 48.40524017467249,
+ "grad_norm": 0.023174211382865906,
+ "learning_rate": 0.0006,
+ "loss": 5.204349517822266,
+ "step": 3485
+ },
+ {
+ "epoch": 48.419213973799124,
+ "grad_norm": 0.024161914363503456,
+ "learning_rate": 0.0006,
+ "loss": 5.339351654052734,
+ "step": 3486
+ },
+ {
+ "epoch": 48.43318777292576,
+ "grad_norm": 0.02221524342894554,
+ "learning_rate": 0.0006,
+ "loss": 5.263001441955566,
+ "step": 3487
+ },
+ {
+ "epoch": 48.4471615720524,
+ "grad_norm": 0.026632830500602722,
+ "learning_rate": 0.0006,
+ "loss": 5.221281051635742,
+ "step": 3488
+ },
+ {
+ "epoch": 48.46113537117904,
+ "grad_norm": 0.034923870116472244,
+ "learning_rate": 0.0006,
+ "loss": 5.296039581298828,
+ "step": 3489
+ },
+ {
+ "epoch": 48.47510917030568,
+ "grad_norm": 0.046184979379177094,
+ "learning_rate": 0.0006,
+ "loss": 5.212769985198975,
+ "step": 3490
+ },
+ {
+ "epoch": 48.48908296943232,
+ "grad_norm": 0.05154098570346832,
+ "learning_rate": 0.0006,
+ "loss": 5.337864875793457,
+ "step": 3491
+ },
+ {
+ "epoch": 48.50305676855895,
+ "grad_norm": 0.042967259883880615,
+ "learning_rate": 0.0006,
+ "loss": 5.267958641052246,
+ "step": 3492
+ },
+ {
+ "epoch": 48.51703056768559,
+ "grad_norm": 0.05356583371758461,
+ "learning_rate": 0.0006,
+ "loss": 5.1404218673706055,
+ "step": 3493
+ },
+ {
+ "epoch": 48.531004366812226,
+ "grad_norm": 0.08219032734632492,
+ "learning_rate": 0.0006,
+ "loss": 5.300687789916992,
+ "step": 3494
+ },
+ {
+ "epoch": 48.544978165938865,
+ "grad_norm": 0.19204089045524597,
+ "learning_rate": 0.0006,
+ "loss": 5.456025123596191,
+ "step": 3495
+ },
+ {
+ "epoch": 48.5589519650655,
+ "grad_norm": 0.25917696952819824,
+ "learning_rate": 0.0006,
+ "loss": 6.156761169433594,
+ "step": 3496
+ },
+ {
+ "epoch": 48.57292576419214,
+ "grad_norm": 0.1566283404827118,
+ "learning_rate": 0.0006,
+ "loss": 5.862346649169922,
+ "step": 3497
+ },
+ {
+ "epoch": 48.58689956331878,
+ "grad_norm": 0.1761486828327179,
+ "learning_rate": 0.0006,
+ "loss": 6.012795448303223,
+ "step": 3498
+ },
+ {
+ "epoch": 48.60087336244541,
+ "grad_norm": 0.09857621043920517,
+ "learning_rate": 0.0006,
+ "loss": 5.9248151779174805,
+ "step": 3499
+ },
+ {
+ "epoch": 48.61484716157205,
+ "grad_norm": 0.05968055501580238,
+ "learning_rate": 0.0006,
+ "loss": 5.780706882476807,
+ "step": 3500
+ },
+ {
+ "epoch": 48.62882096069869,
+ "grad_norm": 0.057433564215898514,
+ "learning_rate": 0.0006,
+ "loss": 5.645388603210449,
+ "step": 3501
+ },
+ {
+ "epoch": 48.64279475982533,
+ "grad_norm": 0.039659615606069565,
+ "learning_rate": 0.0006,
+ "loss": 5.695138931274414,
+ "step": 3502
+ },
+ {
+ "epoch": 48.65676855895197,
+ "grad_norm": 0.04019695147871971,
+ "learning_rate": 0.0006,
+ "loss": 5.672435283660889,
+ "step": 3503
+ },
+ {
+ "epoch": 48.670742358078606,
+ "grad_norm": 0.03033018670976162,
+ "learning_rate": 0.0006,
+ "loss": 5.682260513305664,
+ "step": 3504
+ },
+ {
+ "epoch": 48.68471615720524,
+ "grad_norm": 0.03215787187218666,
+ "learning_rate": 0.0006,
+ "loss": 5.604182243347168,
+ "step": 3505
+ },
+ {
+ "epoch": 48.698689956331876,
+ "grad_norm": 0.026865126565098763,
+ "learning_rate": 0.0006,
+ "loss": 5.64040470123291,
+ "step": 3506
+ },
+ {
+ "epoch": 48.712663755458514,
+ "grad_norm": 0.02831609919667244,
+ "learning_rate": 0.0006,
+ "loss": 5.552720069885254,
+ "step": 3507
+ },
+ {
+ "epoch": 48.72663755458515,
+ "grad_norm": 0.024226613342761993,
+ "learning_rate": 0.0006,
+ "loss": 5.556946277618408,
+ "step": 3508
+ },
+ {
+ "epoch": 48.74061135371179,
+ "grad_norm": 0.025393934920430183,
+ "learning_rate": 0.0006,
+ "loss": 5.476178169250488,
+ "step": 3509
+ },
+ {
+ "epoch": 48.75458515283843,
+ "grad_norm": 0.01984412968158722,
+ "learning_rate": 0.0006,
+ "loss": 5.471088409423828,
+ "step": 3510
+ },
+ {
+ "epoch": 48.76855895196506,
+ "grad_norm": 0.02035129815340042,
+ "learning_rate": 0.0006,
+ "loss": 5.438812732696533,
+ "step": 3511
+ },
+ {
+ "epoch": 48.7825327510917,
+ "grad_norm": 0.0208350270986557,
+ "learning_rate": 0.0006,
+ "loss": 5.381746292114258,
+ "step": 3512
+ },
+ {
+ "epoch": 48.79650655021834,
+ "grad_norm": 0.017783882096409798,
+ "learning_rate": 0.0006,
+ "loss": 5.351675510406494,
+ "step": 3513
+ },
+ {
+ "epoch": 48.81048034934498,
+ "grad_norm": 0.01671770215034485,
+ "learning_rate": 0.0006,
+ "loss": 5.329428672790527,
+ "step": 3514
+ },
+ {
+ "epoch": 48.82445414847162,
+ "grad_norm": 0.017891215160489082,
+ "learning_rate": 0.0006,
+ "loss": 5.212074279785156,
+ "step": 3515
+ },
+ {
+ "epoch": 48.838427947598255,
+ "grad_norm": 0.016847791150212288,
+ "learning_rate": 0.0006,
+ "loss": 5.299534797668457,
+ "step": 3516
+ },
+ {
+ "epoch": 48.852401746724894,
+ "grad_norm": 0.01593637838959694,
+ "learning_rate": 0.0006,
+ "loss": 5.30466890335083,
+ "step": 3517
+ },
+ {
+ "epoch": 48.866375545851525,
+ "grad_norm": 0.014713102020323277,
+ "learning_rate": 0.0006,
+ "loss": 5.294345378875732,
+ "step": 3518
+ },
+ {
+ "epoch": 48.880349344978164,
+ "grad_norm": 0.01433857623487711,
+ "learning_rate": 0.0006,
+ "loss": 5.2148542404174805,
+ "step": 3519
+ },
+ {
+ "epoch": 48.8943231441048,
+ "grad_norm": 0.014979379251599312,
+ "learning_rate": 0.0006,
+ "loss": 5.298582077026367,
+ "step": 3520
+ },
+ {
+ "epoch": 48.90829694323144,
+ "grad_norm": 0.013494192622601986,
+ "learning_rate": 0.0006,
+ "loss": 5.309783935546875,
+ "step": 3521
+ },
+ {
+ "epoch": 48.92227074235808,
+ "grad_norm": 0.014020012691617012,
+ "learning_rate": 0.0006,
+ "loss": 5.284879684448242,
+ "step": 3522
+ },
+ {
+ "epoch": 48.93624454148472,
+ "grad_norm": 0.01348058134317398,
+ "learning_rate": 0.0006,
+ "loss": 5.249116897583008,
+ "step": 3523
+ },
+ {
+ "epoch": 48.95021834061135,
+ "grad_norm": 0.013280218467116356,
+ "learning_rate": 0.0006,
+ "loss": 5.273489952087402,
+ "step": 3524
+ },
+ {
+ "epoch": 48.96419213973799,
+ "grad_norm": 0.012647110037505627,
+ "learning_rate": 0.0006,
+ "loss": 5.306652069091797,
+ "step": 3525
+ },
+ {
+ "epoch": 48.97816593886463,
+ "grad_norm": 0.012847724370658398,
+ "learning_rate": 0.0006,
+ "loss": 5.235745429992676,
+ "step": 3526
+ },
+ {
+ "epoch": 48.992139737991266,
+ "grad_norm": 0.011820166371762753,
+ "learning_rate": 0.0006,
+ "loss": 5.238091468811035,
+ "step": 3527
+ },
+ {
+ "epoch": 49.0,
+ "grad_norm": 0.01255585066974163,
+ "learning_rate": 0.0006,
+ "loss": 5.231295585632324,
+ "step": 3528
+ },
+ {
+ "epoch": 49.0,
+ "eval_loss": 5.394587993621826,
+ "eval_runtime": 56.3734,
+ "eval_samples_per_second": 43.318,
+ "eval_steps_per_second": 1.366,
+ "step": 3528
+ },
+ {
+ "epoch": 49.01397379912664,
+ "grad_norm": 0.011031972244381905,
+ "learning_rate": 0.0006,
+ "loss": 5.1354570388793945,
+ "step": 3529
+ },
+ {
+ "epoch": 49.02794759825328,
+ "grad_norm": 0.010631263256072998,
+ "learning_rate": 0.0006,
+ "loss": 5.177639961242676,
+ "step": 3530
+ },
+ {
+ "epoch": 49.041921397379916,
+ "grad_norm": 0.01133775431662798,
+ "learning_rate": 0.0006,
+ "loss": 5.283428192138672,
+ "step": 3531
+ },
+ {
+ "epoch": 49.05589519650655,
+ "grad_norm": 0.01154172234237194,
+ "learning_rate": 0.0006,
+ "loss": 5.206364154815674,
+ "step": 3532
+ },
+ {
+ "epoch": 49.069868995633186,
+ "grad_norm": 0.010192356072366238,
+ "learning_rate": 0.0006,
+ "loss": 5.2385735511779785,
+ "step": 3533
+ },
+ {
+ "epoch": 49.083842794759825,
+ "grad_norm": 0.012789200060069561,
+ "learning_rate": 0.0006,
+ "loss": 5.071800231933594,
+ "step": 3534
+ },
+ {
+ "epoch": 49.09781659388646,
+ "grad_norm": 0.011282820254564285,
+ "learning_rate": 0.0006,
+ "loss": 5.218020439147949,
+ "step": 3535
+ },
+ {
+ "epoch": 49.1117903930131,
+ "grad_norm": 0.012153448536992073,
+ "learning_rate": 0.0006,
+ "loss": 5.134698867797852,
+ "step": 3536
+ },
+ {
+ "epoch": 49.12576419213974,
+ "grad_norm": 0.011920543387532234,
+ "learning_rate": 0.0006,
+ "loss": 5.141620635986328,
+ "step": 3537
+ },
+ {
+ "epoch": 49.13973799126637,
+ "grad_norm": 0.010949329473078251,
+ "learning_rate": 0.0006,
+ "loss": 5.116259574890137,
+ "step": 3538
+ },
+ {
+ "epoch": 49.15371179039301,
+ "grad_norm": 0.012362448498606682,
+ "learning_rate": 0.0006,
+ "loss": 5.090177059173584,
+ "step": 3539
+ },
+ {
+ "epoch": 49.16768558951965,
+ "grad_norm": 0.011575071141123772,
+ "learning_rate": 0.0006,
+ "loss": 5.093441009521484,
+ "step": 3540
+ },
+ {
+ "epoch": 49.18165938864629,
+ "grad_norm": 0.011126236990094185,
+ "learning_rate": 0.0006,
+ "loss": 5.060235023498535,
+ "step": 3541
+ },
+ {
+ "epoch": 49.19563318777293,
+ "grad_norm": 0.0107401292771101,
+ "learning_rate": 0.0006,
+ "loss": 5.140704154968262,
+ "step": 3542
+ },
+ {
+ "epoch": 49.209606986899566,
+ "grad_norm": 0.011202906258404255,
+ "learning_rate": 0.0006,
+ "loss": 5.10855770111084,
+ "step": 3543
+ },
+ {
+ "epoch": 49.223580786026204,
+ "grad_norm": 0.011083454824984074,
+ "learning_rate": 0.0006,
+ "loss": 5.092751502990723,
+ "step": 3544
+ },
+ {
+ "epoch": 49.237554585152836,
+ "grad_norm": 0.009523800574243069,
+ "learning_rate": 0.0006,
+ "loss": 5.13733434677124,
+ "step": 3545
+ },
+ {
+ "epoch": 49.251528384279474,
+ "grad_norm": 0.011298353783786297,
+ "learning_rate": 0.0006,
+ "loss": 5.036048889160156,
+ "step": 3546
+ },
+ {
+ "epoch": 49.26550218340611,
+ "grad_norm": 0.010676483623683453,
+ "learning_rate": 0.0006,
+ "loss": 5.051901340484619,
+ "step": 3547
+ },
+ {
+ "epoch": 49.27947598253275,
+ "grad_norm": 0.011463399045169353,
+ "learning_rate": 0.0006,
+ "loss": 4.983170986175537,
+ "step": 3548
+ },
+ {
+ "epoch": 49.29344978165939,
+ "grad_norm": 0.010603402741253376,
+ "learning_rate": 0.0006,
+ "loss": 5.112311363220215,
+ "step": 3549
+ },
+ {
+ "epoch": 49.30742358078603,
+ "grad_norm": 0.010535670444369316,
+ "learning_rate": 0.0006,
+ "loss": 5.085227966308594,
+ "step": 3550
+ },
+ {
+ "epoch": 49.32139737991266,
+ "grad_norm": 0.011843928135931492,
+ "learning_rate": 0.0006,
+ "loss": 5.07575798034668,
+ "step": 3551
+ },
+ {
+ "epoch": 49.3353711790393,
+ "grad_norm": 0.010587952099740505,
+ "learning_rate": 0.0006,
+ "loss": 5.019237041473389,
+ "step": 3552
+ },
+ {
+ "epoch": 49.34934497816594,
+ "grad_norm": 0.012285659089684486,
+ "learning_rate": 0.0006,
+ "loss": 5.144603729248047,
+ "step": 3553
+ },
+ {
+ "epoch": 49.36331877729258,
+ "grad_norm": 0.013119431212544441,
+ "learning_rate": 0.0006,
+ "loss": 5.035256862640381,
+ "step": 3554
+ },
+ {
+ "epoch": 49.377292576419215,
+ "grad_norm": 0.012138537131249905,
+ "learning_rate": 0.0006,
+ "loss": 4.996667861938477,
+ "step": 3555
+ },
+ {
+ "epoch": 49.391266375545854,
+ "grad_norm": 0.012142804451286793,
+ "learning_rate": 0.0006,
+ "loss": 4.95058536529541,
+ "step": 3556
+ },
+ {
+ "epoch": 49.40524017467249,
+ "grad_norm": 0.012205676175653934,
+ "learning_rate": 0.0006,
+ "loss": 5.096931457519531,
+ "step": 3557
+ },
+ {
+ "epoch": 49.419213973799124,
+ "grad_norm": 0.012086744420230389,
+ "learning_rate": 0.0006,
+ "loss": 4.99384880065918,
+ "step": 3558
+ },
+ {
+ "epoch": 49.43318777292576,
+ "grad_norm": 0.012918103486299515,
+ "learning_rate": 0.0006,
+ "loss": 5.031294822692871,
+ "step": 3559
+ },
+ {
+ "epoch": 49.4471615720524,
+ "grad_norm": 0.012006713077425957,
+ "learning_rate": 0.0006,
+ "loss": 5.0547943115234375,
+ "step": 3560
+ },
+ {
+ "epoch": 49.46113537117904,
+ "grad_norm": 0.0120399734005332,
+ "learning_rate": 0.0006,
+ "loss": 5.014238357543945,
+ "step": 3561
+ },
+ {
+ "epoch": 49.47510917030568,
+ "grad_norm": 0.013306990265846252,
+ "learning_rate": 0.0006,
+ "loss": 4.972925186157227,
+ "step": 3562
+ },
+ {
+ "epoch": 49.48908296943232,
+ "grad_norm": 0.013878636993467808,
+ "learning_rate": 0.0006,
+ "loss": 5.079033851623535,
+ "step": 3563
+ },
+ {
+ "epoch": 49.50305676855895,
+ "grad_norm": 0.012454566545784473,
+ "learning_rate": 0.0006,
+ "loss": 4.93708610534668,
+ "step": 3564
+ },
+ {
+ "epoch": 49.51703056768559,
+ "grad_norm": 0.014075061306357384,
+ "learning_rate": 0.0006,
+ "loss": 4.965582370758057,
+ "step": 3565
+ },
+ {
+ "epoch": 49.531004366812226,
+ "grad_norm": 0.012525617145001888,
+ "learning_rate": 0.0006,
+ "loss": 4.968148708343506,
+ "step": 3566
+ },
+ {
+ "epoch": 49.544978165938865,
+ "grad_norm": 0.012418225407600403,
+ "learning_rate": 0.0006,
+ "loss": 5.06901741027832,
+ "step": 3567
+ },
+ {
+ "epoch": 49.5589519650655,
+ "grad_norm": 0.015034239739179611,
+ "learning_rate": 0.0006,
+ "loss": 4.984277725219727,
+ "step": 3568
+ },
+ {
+ "epoch": 49.57292576419214,
+ "grad_norm": 0.014329628087580204,
+ "learning_rate": 0.0006,
+ "loss": 4.951558589935303,
+ "step": 3569
+ },
+ {
+ "epoch": 49.58689956331878,
+ "grad_norm": 0.017769070342183113,
+ "learning_rate": 0.0006,
+ "loss": 4.966532230377197,
+ "step": 3570
+ },
+ {
+ "epoch": 49.60087336244541,
+ "grad_norm": 0.01909453794360161,
+ "learning_rate": 0.0006,
+ "loss": 4.919890880584717,
+ "step": 3571
+ },
+ {
+ "epoch": 49.61484716157205,
+ "grad_norm": 0.017519688233733177,
+ "learning_rate": 0.0006,
+ "loss": 5.031286716461182,
+ "step": 3572
+ },
+ {
+ "epoch": 49.62882096069869,
+ "grad_norm": 0.017143897712230682,
+ "learning_rate": 0.0006,
+ "loss": 4.969764709472656,
+ "step": 3573
+ },
+ {
+ "epoch": 49.64279475982533,
+ "grad_norm": 0.01688285917043686,
+ "learning_rate": 0.0006,
+ "loss": 5.0733442306518555,
+ "step": 3574
+ },
+ {
+ "epoch": 49.65676855895197,
+ "grad_norm": 0.018547311425209045,
+ "learning_rate": 0.0006,
+ "loss": 4.980863571166992,
+ "step": 3575
+ },
+ {
+ "epoch": 49.670742358078606,
+ "grad_norm": 0.018877137452363968,
+ "learning_rate": 0.0006,
+ "loss": 4.954646110534668,
+ "step": 3576
+ },
+ {
+ "epoch": 49.68471615720524,
+ "grad_norm": 0.018681600689888,
+ "learning_rate": 0.0006,
+ "loss": 4.959702014923096,
+ "step": 3577
+ },
+ {
+ "epoch": 49.698689956331876,
+ "grad_norm": 0.018947336822748184,
+ "learning_rate": 0.0006,
+ "loss": 5.070502758026123,
+ "step": 3578
+ },
+ {
+ "epoch": 49.712663755458514,
+ "grad_norm": 0.018817484378814697,
+ "learning_rate": 0.0006,
+ "loss": 5.056508541107178,
+ "step": 3579
+ },
+ {
+ "epoch": 49.72663755458515,
+ "grad_norm": 0.01931724324822426,
+ "learning_rate": 0.0006,
+ "loss": 4.922451972961426,
+ "step": 3580
+ },
+ {
+ "epoch": 49.74061135371179,
+ "grad_norm": 0.022047819569706917,
+ "learning_rate": 0.0006,
+ "loss": 5.053170204162598,
+ "step": 3581
+ },
+ {
+ "epoch": 49.75458515283843,
+ "grad_norm": 0.02279801294207573,
+ "learning_rate": 0.0006,
+ "loss": 4.8768439292907715,
+ "step": 3582
+ },
+ {
+ "epoch": 49.76855895196506,
+ "grad_norm": 0.020661601796746254,
+ "learning_rate": 0.0006,
+ "loss": 4.907928466796875,
+ "step": 3583
+ },
+ {
+ "epoch": 49.7825327510917,
+ "grad_norm": 0.020479509606957436,
+ "learning_rate": 0.0006,
+ "loss": 5.040694236755371,
+ "step": 3584
+ },
+ {
+ "epoch": 49.79650655021834,
+ "grad_norm": 0.021654162555933,
+ "learning_rate": 0.0006,
+ "loss": 5.018434524536133,
+ "step": 3585
+ },
+ {
+ "epoch": 49.81048034934498,
+ "grad_norm": 0.022486824542284012,
+ "learning_rate": 0.0006,
+ "loss": 5.086538314819336,
+ "step": 3586
+ },
+ {
+ "epoch": 49.82445414847162,
+ "grad_norm": 0.02141369879245758,
+ "learning_rate": 0.0006,
+ "loss": 5.035694122314453,
+ "step": 3587
+ },
+ {
+ "epoch": 49.838427947598255,
+ "grad_norm": 0.015926677733659744,
+ "learning_rate": 0.0006,
+ "loss": 4.905038833618164,
+ "step": 3588
+ },
+ {
+ "epoch": 49.852401746724894,
+ "grad_norm": 0.015775298699736595,
+ "learning_rate": 0.0006,
+ "loss": 5.003902435302734,
+ "step": 3589
+ },
+ {
+ "epoch": 49.866375545851525,
+ "grad_norm": 0.01872118003666401,
+ "learning_rate": 0.0006,
+ "loss": 4.891060829162598,
+ "step": 3590
+ },
+ {
+ "epoch": 49.880349344978164,
+ "grad_norm": 0.02066192403435707,
+ "learning_rate": 0.0006,
+ "loss": 5.002897262573242,
+ "step": 3591
+ },
+ {
+ "epoch": 49.8943231441048,
+ "grad_norm": 0.02348714880645275,
+ "learning_rate": 0.0006,
+ "loss": 4.9189958572387695,
+ "step": 3592
+ },
+ {
+ "epoch": 49.90829694323144,
+ "grad_norm": 0.020402103662490845,
+ "learning_rate": 0.0006,
+ "loss": 4.992661476135254,
+ "step": 3593
+ },
+ {
+ "epoch": 49.92227074235808,
+ "grad_norm": 0.020466072484850883,
+ "learning_rate": 0.0006,
+ "loss": 4.962968826293945,
+ "step": 3594
+ },
+ {
+ "epoch": 49.93624454148472,
+ "grad_norm": 0.01922314241528511,
+ "learning_rate": 0.0006,
+ "loss": 4.897730827331543,
+ "step": 3595
+ },
+ {
+ "epoch": 49.95021834061135,
+ "grad_norm": 0.017228754237294197,
+ "learning_rate": 0.0006,
+ "loss": 4.964842796325684,
+ "step": 3596
+ },
+ {
+ "epoch": 49.96419213973799,
+ "grad_norm": 0.016021743416786194,
+ "learning_rate": 0.0006,
+ "loss": 4.910758972167969,
+ "step": 3597
+ },
+ {
+ "epoch": 49.97816593886463,
+ "grad_norm": 0.017985103651881218,
+ "learning_rate": 0.0006,
+ "loss": 4.916072368621826,
+ "step": 3598
+ },
+ {
+ "epoch": 49.992139737991266,
+ "grad_norm": 0.018780240789055824,
+ "learning_rate": 0.0006,
+ "loss": 4.94658088684082,
+ "step": 3599
+ },
+ {
+ "epoch": 50.0,
+ "grad_norm": 0.01776314526796341,
+ "learning_rate": 0.0006,
+ "loss": 4.969357967376709,
+ "step": 3600
+ },
+ {
+ "epoch": 50.0,
+ "eval_loss": 5.141158580780029,
+ "eval_runtime": 56.2332,
+ "eval_samples_per_second": 43.426,
+ "eval_steps_per_second": 1.369,
+ "step": 3600
+ },
+ {
+ "epoch": 50.01397379912664,
+ "grad_norm": 0.016308696940541267,
+ "learning_rate": 0.0006,
+ "loss": 4.924664497375488,
+ "step": 3601
+ },
+ {
+ "epoch": 50.02794759825328,
+ "grad_norm": 0.018810199573636055,
+ "learning_rate": 0.0006,
+ "loss": 4.82293701171875,
+ "step": 3602
+ },
+ {
+ "epoch": 50.041921397379916,
+ "grad_norm": 0.02227160707116127,
+ "learning_rate": 0.0006,
+ "loss": 4.9612298011779785,
+ "step": 3603
+ },
+ {
+ "epoch": 50.05589519650655,
+ "grad_norm": 0.025266462936997414,
+ "learning_rate": 0.0006,
+ "loss": 4.963835716247559,
+ "step": 3604
+ },
+ {
+ "epoch": 50.069868995633186,
+ "grad_norm": 0.028588181361556053,
+ "learning_rate": 0.0006,
+ "loss": 4.992118835449219,
+ "step": 3605
+ },
+ {
+ "epoch": 50.083842794759825,
+ "grad_norm": 0.02739487960934639,
+ "learning_rate": 0.0006,
+ "loss": 4.915547847747803,
+ "step": 3606
+ },
+ {
+ "epoch": 50.09781659388646,
+ "grad_norm": 0.027000300586223602,
+ "learning_rate": 0.0006,
+ "loss": 4.883801460266113,
+ "step": 3607
+ },
+ {
+ "epoch": 50.1117903930131,
+ "grad_norm": 0.023468483239412308,
+ "learning_rate": 0.0006,
+ "loss": 5.006168365478516,
+ "step": 3608
+ },
+ {
+ "epoch": 50.12576419213974,
+ "grad_norm": 0.026753634214401245,
+ "learning_rate": 0.0006,
+ "loss": 4.92526912689209,
+ "step": 3609
+ },
+ {
+ "epoch": 50.13973799126637,
+ "grad_norm": 0.026816733181476593,
+ "learning_rate": 0.0006,
+ "loss": 4.847546100616455,
+ "step": 3610
+ },
+ {
+ "epoch": 50.15371179039301,
+ "grad_norm": 0.027362391352653503,
+ "learning_rate": 0.0006,
+ "loss": 4.862548351287842,
+ "step": 3611
+ },
+ {
+ "epoch": 50.16768558951965,
+ "grad_norm": 0.03450845181941986,
+ "learning_rate": 0.0006,
+ "loss": 4.963202953338623,
+ "step": 3612
+ },
+ {
+ "epoch": 50.18165938864629,
+ "grad_norm": 0.03801408037543297,
+ "learning_rate": 0.0006,
+ "loss": 4.770240306854248,
+ "step": 3613
+ },
+ {
+ "epoch": 50.19563318777293,
+ "grad_norm": 0.03579140082001686,
+ "learning_rate": 0.0006,
+ "loss": 4.797981262207031,
+ "step": 3614
+ },
+ {
+ "epoch": 50.209606986899566,
+ "grad_norm": 0.029393760487437248,
+ "learning_rate": 0.0006,
+ "loss": 4.798581600189209,
+ "step": 3615
+ },
+ {
+ "epoch": 50.223580786026204,
+ "grad_norm": 0.026111142709851265,
+ "learning_rate": 0.0006,
+ "loss": 4.990712642669678,
+ "step": 3616
+ },
+ {
+ "epoch": 50.237554585152836,
+ "grad_norm": 0.02213387005031109,
+ "learning_rate": 0.0006,
+ "loss": 4.865842342376709,
+ "step": 3617
+ },
+ {
+ "epoch": 50.251528384279474,
+ "grad_norm": 0.025450684130191803,
+ "learning_rate": 0.0006,
+ "loss": 4.903746128082275,
+ "step": 3618
+ },
+ {
+ "epoch": 50.26550218340611,
+ "grad_norm": 0.020915117114782333,
+ "learning_rate": 0.0006,
+ "loss": 4.951366424560547,
+ "step": 3619
+ },
+ {
+ "epoch": 50.27947598253275,
+ "grad_norm": 0.020972849801182747,
+ "learning_rate": 0.0006,
+ "loss": 4.783311367034912,
+ "step": 3620
+ },
+ {
+ "epoch": 50.29344978165939,
+ "grad_norm": 0.016956273466348648,
+ "learning_rate": 0.0006,
+ "loss": 4.955767631530762,
+ "step": 3621
+ },
+ {
+ "epoch": 50.30742358078603,
+ "grad_norm": 0.018045252189040184,
+ "learning_rate": 0.0006,
+ "loss": 4.869322776794434,
+ "step": 3622
+ },
+ {
+ "epoch": 50.32139737991266,
+ "grad_norm": 0.018199188634753227,
+ "learning_rate": 0.0006,
+ "loss": 4.95680046081543,
+ "step": 3623
+ },
+ {
+ "epoch": 50.3353711790393,
+ "grad_norm": 0.016688082367181778,
+ "learning_rate": 0.0006,
+ "loss": 4.783535957336426,
+ "step": 3624
+ },
+ {
+ "epoch": 50.34934497816594,
+ "grad_norm": 0.017540352419018745,
+ "learning_rate": 0.0006,
+ "loss": 4.8236083984375,
+ "step": 3625
+ },
+ {
+ "epoch": 50.36331877729258,
+ "grad_norm": 0.018054187297821045,
+ "learning_rate": 0.0006,
+ "loss": 4.764655113220215,
+ "step": 3626
+ },
+ {
+ "epoch": 50.377292576419215,
+ "grad_norm": 0.01741131953895092,
+ "learning_rate": 0.0006,
+ "loss": 4.875613689422607,
+ "step": 3627
+ },
+ {
+ "epoch": 50.391266375545854,
+ "grad_norm": 0.01682269014418125,
+ "learning_rate": 0.0006,
+ "loss": 4.818509101867676,
+ "step": 3628
+ },
+ {
+ "epoch": 50.40524017467249,
+ "grad_norm": 0.016652341932058334,
+ "learning_rate": 0.0006,
+ "loss": 4.974020481109619,
+ "step": 3629
+ },
+ {
+ "epoch": 50.419213973799124,
+ "grad_norm": 0.02035282365977764,
+ "learning_rate": 0.0006,
+ "loss": 4.893037796020508,
+ "step": 3630
+ },
+ {
+ "epoch": 50.43318777292576,
+ "grad_norm": 0.019907061010599136,
+ "learning_rate": 0.0006,
+ "loss": 4.848119735717773,
+ "step": 3631
+ },
+ {
+ "epoch": 50.4471615720524,
+ "grad_norm": 0.023861493915319443,
+ "learning_rate": 0.0006,
+ "loss": 4.819353103637695,
+ "step": 3632
+ },
+ {
+ "epoch": 50.46113537117904,
+ "grad_norm": 0.025254026055336,
+ "learning_rate": 0.0006,
+ "loss": 4.769312858581543,
+ "step": 3633
+ },
+ {
+ "epoch": 50.47510917030568,
+ "grad_norm": 0.026154937222599983,
+ "learning_rate": 0.0006,
+ "loss": 4.7586774826049805,
+ "step": 3634
+ },
+ {
+ "epoch": 50.48908296943232,
+ "grad_norm": 0.026342090219259262,
+ "learning_rate": 0.0006,
+ "loss": 4.861218452453613,
+ "step": 3635
+ },
+ {
+ "epoch": 50.50305676855895,
+ "grad_norm": 0.028169559314846992,
+ "learning_rate": 0.0006,
+ "loss": 4.818380355834961,
+ "step": 3636
+ },
+ {
+ "epoch": 50.51703056768559,
+ "grad_norm": 0.02820264920592308,
+ "learning_rate": 0.0006,
+ "loss": 4.892871856689453,
+ "step": 3637
+ },
+ {
+ "epoch": 50.531004366812226,
+ "grad_norm": 0.03345080465078354,
+ "learning_rate": 0.0006,
+ "loss": 4.811084747314453,
+ "step": 3638
+ },
+ {
+ "epoch": 50.544978165938865,
+ "grad_norm": 0.04005736857652664,
+ "learning_rate": 0.0006,
+ "loss": 4.765410900115967,
+ "step": 3639
+ },
+ {
+ "epoch": 50.5589519650655,
+ "grad_norm": 0.04333870857954025,
+ "learning_rate": 0.0006,
+ "loss": 4.956520080566406,
+ "step": 3640
+ },
+ {
+ "epoch": 50.57292576419214,
+ "grad_norm": 0.04216952994465828,
+ "learning_rate": 0.0006,
+ "loss": 4.793441295623779,
+ "step": 3641
+ },
+ {
+ "epoch": 50.58689956331878,
+ "grad_norm": 0.02936951443552971,
+ "learning_rate": 0.0006,
+ "loss": 4.8447957038879395,
+ "step": 3642
+ },
+ {
+ "epoch": 50.60087336244541,
+ "grad_norm": 0.026937026530504227,
+ "learning_rate": 0.0006,
+ "loss": 4.962215423583984,
+ "step": 3643
+ },
+ {
+ "epoch": 50.61484716157205,
+ "grad_norm": 0.02628890983760357,
+ "learning_rate": 0.0006,
+ "loss": 4.883061408996582,
+ "step": 3644
+ },
+ {
+ "epoch": 50.62882096069869,
+ "grad_norm": 0.025984834879636765,
+ "learning_rate": 0.0006,
+ "loss": 4.768662452697754,
+ "step": 3645
+ },
+ {
+ "epoch": 50.64279475982533,
+ "grad_norm": 0.023998012766242027,
+ "learning_rate": 0.0006,
+ "loss": 4.851325035095215,
+ "step": 3646
+ },
+ {
+ "epoch": 50.65676855895197,
+ "grad_norm": 0.02143184095621109,
+ "learning_rate": 0.0006,
+ "loss": 4.8274688720703125,
+ "step": 3647
+ },
+ {
+ "epoch": 50.670742358078606,
+ "grad_norm": 0.02488900162279606,
+ "learning_rate": 0.0006,
+ "loss": 4.831140518188477,
+ "step": 3648
+ },
+ {
+ "epoch": 50.68471615720524,
+ "grad_norm": 0.02330533228814602,
+ "learning_rate": 0.0006,
+ "loss": 4.850462913513184,
+ "step": 3649
+ },
+ {
+ "epoch": 50.698689956331876,
+ "grad_norm": 0.02121943049132824,
+ "learning_rate": 0.0006,
+ "loss": 4.870253086090088,
+ "step": 3650
+ },
+ {
+ "epoch": 50.712663755458514,
+ "grad_norm": 0.02080935798585415,
+ "learning_rate": 0.0006,
+ "loss": 4.798770904541016,
+ "step": 3651
+ },
+ {
+ "epoch": 50.72663755458515,
+ "grad_norm": 0.019047029316425323,
+ "learning_rate": 0.0006,
+ "loss": 4.8427300453186035,
+ "step": 3652
+ },
+ {
+ "epoch": 50.74061135371179,
+ "grad_norm": 0.01824103854596615,
+ "learning_rate": 0.0006,
+ "loss": 4.820980548858643,
+ "step": 3653
+ },
+ {
+ "epoch": 50.75458515283843,
+ "grad_norm": 0.018129967153072357,
+ "learning_rate": 0.0006,
+ "loss": 4.946732997894287,
+ "step": 3654
+ },
+ {
+ "epoch": 50.76855895196506,
+ "grad_norm": 0.01784815452992916,
+ "learning_rate": 0.0006,
+ "loss": 4.829516410827637,
+ "step": 3655
+ },
+ {
+ "epoch": 50.7825327510917,
+ "grad_norm": 0.020021555945277214,
+ "learning_rate": 0.0006,
+ "loss": 4.780428409576416,
+ "step": 3656
+ },
+ {
+ "epoch": 50.79650655021834,
+ "grad_norm": 0.01863253302872181,
+ "learning_rate": 0.0006,
+ "loss": 4.791567802429199,
+ "step": 3657
+ },
+ {
+ "epoch": 50.81048034934498,
+ "grad_norm": 0.019151683896780014,
+ "learning_rate": 0.0006,
+ "loss": 4.824892997741699,
+ "step": 3658
+ },
+ {
+ "epoch": 50.82445414847162,
+ "grad_norm": 0.01891648955643177,
+ "learning_rate": 0.0006,
+ "loss": 4.789605140686035,
+ "step": 3659
+ },
+ {
+ "epoch": 50.838427947598255,
+ "grad_norm": 0.016019510105252266,
+ "learning_rate": 0.0006,
+ "loss": 4.779555320739746,
+ "step": 3660
+ },
+ {
+ "epoch": 50.852401746724894,
+ "grad_norm": 0.016096491366624832,
+ "learning_rate": 0.0006,
+ "loss": 4.753842353820801,
+ "step": 3661
+ },
+ {
+ "epoch": 50.866375545851525,
+ "grad_norm": 0.01693476364016533,
+ "learning_rate": 0.0006,
+ "loss": 4.768477439880371,
+ "step": 3662
+ },
+ {
+ "epoch": 50.880349344978164,
+ "grad_norm": 0.01749575324356556,
+ "learning_rate": 0.0006,
+ "loss": 4.779143333435059,
+ "step": 3663
+ },
+ {
+ "epoch": 50.8943231441048,
+ "grad_norm": 0.016851043328642845,
+ "learning_rate": 0.0006,
+ "loss": 4.780387878417969,
+ "step": 3664
+ },
+ {
+ "epoch": 50.90829694323144,
+ "grad_norm": 0.019045202061533928,
+ "learning_rate": 0.0006,
+ "loss": 4.769913673400879,
+ "step": 3665
+ },
+ {
+ "epoch": 50.92227074235808,
+ "grad_norm": 0.01954123005270958,
+ "learning_rate": 0.0006,
+ "loss": 4.769009590148926,
+ "step": 3666
+ },
+ {
+ "epoch": 50.93624454148472,
+ "grad_norm": 0.022527096793055534,
+ "learning_rate": 0.0006,
+ "loss": 4.789215087890625,
+ "step": 3667
+ },
+ {
+ "epoch": 50.95021834061135,
+ "grad_norm": 0.02871634066104889,
+ "learning_rate": 0.0006,
+ "loss": 4.749727725982666,
+ "step": 3668
+ },
+ {
+ "epoch": 50.96419213973799,
+ "grad_norm": 0.04045053571462631,
+ "learning_rate": 0.0006,
+ "loss": 4.909038543701172,
+ "step": 3669
+ },
+ {
+ "epoch": 50.97816593886463,
+ "grad_norm": 0.04156879335641861,
+ "learning_rate": 0.0006,
+ "loss": 4.876831531524658,
+ "step": 3670
+ },
+ {
+ "epoch": 50.992139737991266,
+ "grad_norm": 0.02621397376060486,
+ "learning_rate": 0.0006,
+ "loss": 4.797173500061035,
+ "step": 3671
+ },
+ {
+ "epoch": 51.0,
+ "grad_norm": 0.03389597684144974,
+ "learning_rate": 0.0006,
+ "loss": 4.8428850173950195,
+ "step": 3672
+ },
+ {
+ "epoch": 51.0,
+ "eval_loss": 4.989261627197266,
+ "eval_runtime": 56.358,
+ "eval_samples_per_second": 43.33,
+ "eval_steps_per_second": 1.366,
+ "step": 3672
+ },
+ {
+ "epoch": 51.01397379912664,
+ "grad_norm": 0.03216972202062607,
+ "learning_rate": 0.0006,
+ "loss": 4.755102634429932,
+ "step": 3673
+ },
+ {
+ "epoch": 51.02794759825328,
+ "grad_norm": 0.03190374746918678,
+ "learning_rate": 0.0006,
+ "loss": 4.720424652099609,
+ "step": 3674
+ },
+ {
+ "epoch": 51.041921397379916,
+ "grad_norm": 0.031727299094200134,
+ "learning_rate": 0.0006,
+ "loss": 4.803277969360352,
+ "step": 3675
+ },
+ {
+ "epoch": 51.05589519650655,
+ "grad_norm": 0.02473079040646553,
+ "learning_rate": 0.0006,
+ "loss": 4.761087417602539,
+ "step": 3676
+ },
+ {
+ "epoch": 51.069868995633186,
+ "grad_norm": 0.020360209047794342,
+ "learning_rate": 0.0006,
+ "loss": 4.753884792327881,
+ "step": 3677
+ },
+ {
+ "epoch": 51.083842794759825,
+ "grad_norm": 0.02078923024237156,
+ "learning_rate": 0.0006,
+ "loss": 4.647130489349365,
+ "step": 3678
+ },
+ {
+ "epoch": 51.09781659388646,
+ "grad_norm": 0.019387200474739075,
+ "learning_rate": 0.0006,
+ "loss": 4.74135160446167,
+ "step": 3679
+ },
+ {
+ "epoch": 51.1117903930131,
+ "grad_norm": 0.01874949410557747,
+ "learning_rate": 0.0006,
+ "loss": 4.795329570770264,
+ "step": 3680
+ },
+ {
+ "epoch": 51.12576419213974,
+ "grad_norm": 0.019331837072968483,
+ "learning_rate": 0.0006,
+ "loss": 4.897967338562012,
+ "step": 3681
+ },
+ {
+ "epoch": 51.13973799126637,
+ "grad_norm": 0.017066050320863724,
+ "learning_rate": 0.0006,
+ "loss": 4.825351715087891,
+ "step": 3682
+ },
+ {
+ "epoch": 51.15371179039301,
+ "grad_norm": 0.01757810264825821,
+ "learning_rate": 0.0006,
+ "loss": 4.719335079193115,
+ "step": 3683
+ },
+ {
+ "epoch": 51.16768558951965,
+ "grad_norm": 0.018174266442656517,
+ "learning_rate": 0.0006,
+ "loss": 4.732680320739746,
+ "step": 3684
+ },
+ {
+ "epoch": 51.18165938864629,
+ "grad_norm": 0.018497727811336517,
+ "learning_rate": 0.0006,
+ "loss": 4.731050491333008,
+ "step": 3685
+ },
+ {
+ "epoch": 51.19563318777293,
+ "grad_norm": 0.0186338871717453,
+ "learning_rate": 0.0006,
+ "loss": 4.748035430908203,
+ "step": 3686
+ },
+ {
+ "epoch": 51.209606986899566,
+ "grad_norm": 0.018531711772084236,
+ "learning_rate": 0.0006,
+ "loss": 4.9007463455200195,
+ "step": 3687
+ },
+ {
+ "epoch": 51.223580786026204,
+ "grad_norm": 0.018221354112029076,
+ "learning_rate": 0.0006,
+ "loss": 4.84077262878418,
+ "step": 3688
+ },
+ {
+ "epoch": 51.237554585152836,
+ "grad_norm": 0.0203656405210495,
+ "learning_rate": 0.0006,
+ "loss": 4.73009729385376,
+ "step": 3689
+ },
+ {
+ "epoch": 51.251528384279474,
+ "grad_norm": 0.01964893378317356,
+ "learning_rate": 0.0006,
+ "loss": 4.772583961486816,
+ "step": 3690
+ },
+ {
+ "epoch": 51.26550218340611,
+ "grad_norm": 0.019606510177254677,
+ "learning_rate": 0.0006,
+ "loss": 4.599812030792236,
+ "step": 3691
+ },
+ {
+ "epoch": 51.27947598253275,
+ "grad_norm": 0.020501255989074707,
+ "learning_rate": 0.0006,
+ "loss": 4.809898853302002,
+ "step": 3692
+ },
+ {
+ "epoch": 51.29344978165939,
+ "grad_norm": 0.01870838552713394,
+ "learning_rate": 0.0006,
+ "loss": 4.741898059844971,
+ "step": 3693
+ },
+ {
+ "epoch": 51.30742358078603,
+ "grad_norm": 0.01802559196949005,
+ "learning_rate": 0.0006,
+ "loss": 4.746562480926514,
+ "step": 3694
+ },
+ {
+ "epoch": 51.32139737991266,
+ "grad_norm": 0.02053048275411129,
+ "learning_rate": 0.0006,
+ "loss": 4.69932746887207,
+ "step": 3695
+ },
+ {
+ "epoch": 51.3353711790393,
+ "grad_norm": 0.02661825157701969,
+ "learning_rate": 0.0006,
+ "loss": 4.68748664855957,
+ "step": 3696
+ },
+ {
+ "epoch": 51.34934497816594,
+ "grad_norm": 0.028269147500395775,
+ "learning_rate": 0.0006,
+ "loss": 4.731565475463867,
+ "step": 3697
+ },
+ {
+ "epoch": 51.36331877729258,
+ "grad_norm": 0.02650964818894863,
+ "learning_rate": 0.0006,
+ "loss": 4.733857154846191,
+ "step": 3698
+ },
+ {
+ "epoch": 51.377292576419215,
+ "grad_norm": 0.023696066811680794,
+ "learning_rate": 0.0006,
+ "loss": 4.812362194061279,
+ "step": 3699
+ },
+ {
+ "epoch": 51.391266375545854,
+ "grad_norm": 0.02365880273282528,
+ "learning_rate": 0.0006,
+ "loss": 4.733426094055176,
+ "step": 3700
+ },
+ {
+ "epoch": 51.40524017467249,
+ "grad_norm": 0.026249447837471962,
+ "learning_rate": 0.0006,
+ "loss": 4.715622901916504,
+ "step": 3701
+ },
+ {
+ "epoch": 51.419213973799124,
+ "grad_norm": 0.03301899880170822,
+ "learning_rate": 0.0006,
+ "loss": 4.7223663330078125,
+ "step": 3702
+ },
+ {
+ "epoch": 51.43318777292576,
+ "grad_norm": 0.044686660170555115,
+ "learning_rate": 0.0006,
+ "loss": 4.8121161460876465,
+ "step": 3703
+ },
+ {
+ "epoch": 51.4471615720524,
+ "grad_norm": 0.04733522981405258,
+ "learning_rate": 0.0006,
+ "loss": 4.841394424438477,
+ "step": 3704
+ },
+ {
+ "epoch": 51.46113537117904,
+ "grad_norm": 0.0395209975540638,
+ "learning_rate": 0.0006,
+ "loss": 4.773837089538574,
+ "step": 3705
+ },
+ {
+ "epoch": 51.47510917030568,
+ "grad_norm": 0.032790958881378174,
+ "learning_rate": 0.0006,
+ "loss": 4.833066463470459,
+ "step": 3706
+ },
+ {
+ "epoch": 51.48908296943232,
+ "grad_norm": 0.030578427016735077,
+ "learning_rate": 0.0006,
+ "loss": 4.685190200805664,
+ "step": 3707
+ },
+ {
+ "epoch": 51.50305676855895,
+ "grad_norm": 0.031619083136320114,
+ "learning_rate": 0.0006,
+ "loss": 4.740378379821777,
+ "step": 3708
+ },
+ {
+ "epoch": 51.51703056768559,
+ "grad_norm": 0.02385825850069523,
+ "learning_rate": 0.0006,
+ "loss": 4.812911510467529,
+ "step": 3709
+ },
+ {
+ "epoch": 51.531004366812226,
+ "grad_norm": 0.023523425683379173,
+ "learning_rate": 0.0006,
+ "loss": 4.669580459594727,
+ "step": 3710
+ },
+ {
+ "epoch": 51.544978165938865,
+ "grad_norm": 0.021316280588507652,
+ "learning_rate": 0.0006,
+ "loss": 4.76888370513916,
+ "step": 3711
+ },
+ {
+ "epoch": 51.5589519650655,
+ "grad_norm": 0.024663880467414856,
+ "learning_rate": 0.0006,
+ "loss": 4.702592372894287,
+ "step": 3712
+ },
+ {
+ "epoch": 51.57292576419214,
+ "grad_norm": 0.024173280224204063,
+ "learning_rate": 0.0006,
+ "loss": 4.710604667663574,
+ "step": 3713
+ },
+ {
+ "epoch": 51.58689956331878,
+ "grad_norm": 0.02754114381968975,
+ "learning_rate": 0.0006,
+ "loss": 4.746356964111328,
+ "step": 3714
+ },
+ {
+ "epoch": 51.60087336244541,
+ "grad_norm": 0.027903622016310692,
+ "learning_rate": 0.0006,
+ "loss": 4.7546491622924805,
+ "step": 3715
+ },
+ {
+ "epoch": 51.61484716157205,
+ "grad_norm": 0.024871617555618286,
+ "learning_rate": 0.0006,
+ "loss": 4.605746269226074,
+ "step": 3716
+ },
+ {
+ "epoch": 51.62882096069869,
+ "grad_norm": 0.021239031106233597,
+ "learning_rate": 0.0006,
+ "loss": 4.6942458152771,
+ "step": 3717
+ },
+ {
+ "epoch": 51.64279475982533,
+ "grad_norm": 0.01925622671842575,
+ "learning_rate": 0.0006,
+ "loss": 4.593558311462402,
+ "step": 3718
+ },
+ {
+ "epoch": 51.65676855895197,
+ "grad_norm": 0.01956283673644066,
+ "learning_rate": 0.0006,
+ "loss": 4.631172180175781,
+ "step": 3719
+ },
+ {
+ "epoch": 51.670742358078606,
+ "grad_norm": 0.02269871160387993,
+ "learning_rate": 0.0006,
+ "loss": 4.729236602783203,
+ "step": 3720
+ },
+ {
+ "epoch": 51.68471615720524,
+ "grad_norm": 0.02253916673362255,
+ "learning_rate": 0.0006,
+ "loss": 4.7191033363342285,
+ "step": 3721
+ },
+ {
+ "epoch": 51.698689956331876,
+ "grad_norm": 0.026231050491333008,
+ "learning_rate": 0.0006,
+ "loss": 4.811850547790527,
+ "step": 3722
+ },
+ {
+ "epoch": 51.712663755458514,
+ "grad_norm": 0.02558821812272072,
+ "learning_rate": 0.0006,
+ "loss": 4.681466102600098,
+ "step": 3723
+ },
+ {
+ "epoch": 51.72663755458515,
+ "grad_norm": 0.02124209702014923,
+ "learning_rate": 0.0006,
+ "loss": 4.693664073944092,
+ "step": 3724
+ },
+ {
+ "epoch": 51.74061135371179,
+ "grad_norm": 0.016638319939374924,
+ "learning_rate": 0.0006,
+ "loss": 4.771440029144287,
+ "step": 3725
+ },
+ {
+ "epoch": 51.75458515283843,
+ "grad_norm": 0.018048347905278206,
+ "learning_rate": 0.0006,
+ "loss": 4.643252372741699,
+ "step": 3726
+ },
+ {
+ "epoch": 51.76855895196506,
+ "grad_norm": 0.019904717803001404,
+ "learning_rate": 0.0006,
+ "loss": 4.66530704498291,
+ "step": 3727
+ },
+ {
+ "epoch": 51.7825327510917,
+ "grad_norm": 0.0194614976644516,
+ "learning_rate": 0.0006,
+ "loss": 4.711943626403809,
+ "step": 3728
+ },
+ {
+ "epoch": 51.79650655021834,
+ "grad_norm": 0.019175555557012558,
+ "learning_rate": 0.0006,
+ "loss": 4.721931457519531,
+ "step": 3729
+ },
+ {
+ "epoch": 51.81048034934498,
+ "grad_norm": 0.02093447372317314,
+ "learning_rate": 0.0006,
+ "loss": 4.633059024810791,
+ "step": 3730
+ },
+ {
+ "epoch": 51.82445414847162,
+ "grad_norm": 0.01930283196270466,
+ "learning_rate": 0.0006,
+ "loss": 4.659373760223389,
+ "step": 3731
+ },
+ {
+ "epoch": 51.838427947598255,
+ "grad_norm": 0.015350676141679287,
+ "learning_rate": 0.0006,
+ "loss": 4.763708114624023,
+ "step": 3732
+ },
+ {
+ "epoch": 51.852401746724894,
+ "grad_norm": 0.014139448292553425,
+ "learning_rate": 0.0006,
+ "loss": 4.753891468048096,
+ "step": 3733
+ },
+ {
+ "epoch": 51.866375545851525,
+ "grad_norm": 0.016215872019529343,
+ "learning_rate": 0.0006,
+ "loss": 4.658967018127441,
+ "step": 3734
+ },
+ {
+ "epoch": 51.880349344978164,
+ "grad_norm": 0.017328619956970215,
+ "learning_rate": 0.0006,
+ "loss": 4.696579933166504,
+ "step": 3735
+ },
+ {
+ "epoch": 51.8943231441048,
+ "grad_norm": 0.018103396520018578,
+ "learning_rate": 0.0006,
+ "loss": 4.627464771270752,
+ "step": 3736
+ },
+ {
+ "epoch": 51.90829694323144,
+ "grad_norm": 0.01715235412120819,
+ "learning_rate": 0.0006,
+ "loss": 4.7368011474609375,
+ "step": 3737
+ },
+ {
+ "epoch": 51.92227074235808,
+ "grad_norm": 0.014790094457566738,
+ "learning_rate": 0.0006,
+ "loss": 4.629441738128662,
+ "step": 3738
+ },
+ {
+ "epoch": 51.93624454148472,
+ "grad_norm": 0.017114296555519104,
+ "learning_rate": 0.0006,
+ "loss": 4.761256694793701,
+ "step": 3739
+ },
+ {
+ "epoch": 51.95021834061135,
+ "grad_norm": 0.019009845331311226,
+ "learning_rate": 0.0006,
+ "loss": 4.789155006408691,
+ "step": 3740
+ },
+ {
+ "epoch": 51.96419213973799,
+ "grad_norm": 0.019696692004799843,
+ "learning_rate": 0.0006,
+ "loss": 4.721269607543945,
+ "step": 3741
+ },
+ {
+ "epoch": 51.97816593886463,
+ "grad_norm": 0.020542822778224945,
+ "learning_rate": 0.0006,
+ "loss": 4.687038898468018,
+ "step": 3742
+ },
+ {
+ "epoch": 51.992139737991266,
+ "grad_norm": 0.021267160773277283,
+ "learning_rate": 0.0006,
+ "loss": 4.747860908508301,
+ "step": 3743
+ },
+ {
+ "epoch": 52.0,
+ "grad_norm": 0.025049567222595215,
+ "learning_rate": 0.0006,
+ "loss": 4.8473100662231445,
+ "step": 3744
+ },
+ {
+ "epoch": 52.0,
+ "eval_loss": 4.943043231964111,
+ "eval_runtime": 56.4929,
+ "eval_samples_per_second": 43.227,
+ "eval_steps_per_second": 1.363,
+ "step": 3744
+ },
+ {
+ "epoch": 52.01397379912664,
+ "grad_norm": 0.03054986707866192,
+ "learning_rate": 0.0006,
+ "loss": 4.856931686401367,
+ "step": 3745
+ },
+ {
+ "epoch": 52.02794759825328,
+ "grad_norm": 0.03838375583291054,
+ "learning_rate": 0.0006,
+ "loss": 4.699807167053223,
+ "step": 3746
+ },
+ {
+ "epoch": 52.041921397379916,
+ "grad_norm": 0.03383488208055496,
+ "learning_rate": 0.0006,
+ "loss": 4.6626787185668945,
+ "step": 3747
+ },
+ {
+ "epoch": 52.05589519650655,
+ "grad_norm": 0.0201561376452446,
+ "learning_rate": 0.0006,
+ "loss": 4.621831893920898,
+ "step": 3748
+ },
+ {
+ "epoch": 52.069868995633186,
+ "grad_norm": 0.028749438002705574,
+ "learning_rate": 0.0006,
+ "loss": 4.682540416717529,
+ "step": 3749
+ },
+ {
+ "epoch": 52.083842794759825,
+ "grad_norm": 0.02681174874305725,
+ "learning_rate": 0.0006,
+ "loss": 4.629997253417969,
+ "step": 3750
+ },
+ {
+ "epoch": 52.09781659388646,
+ "grad_norm": 0.024063318967819214,
+ "learning_rate": 0.0006,
+ "loss": 4.573021411895752,
+ "step": 3751
+ },
+ {
+ "epoch": 52.1117903930131,
+ "grad_norm": 0.02324330434203148,
+ "learning_rate": 0.0006,
+ "loss": 4.739465236663818,
+ "step": 3752
+ },
+ {
+ "epoch": 52.12576419213974,
+ "grad_norm": 0.0225426834076643,
+ "learning_rate": 0.0006,
+ "loss": 4.7335333824157715,
+ "step": 3753
+ },
+ {
+ "epoch": 52.13973799126637,
+ "grad_norm": 0.019593428820371628,
+ "learning_rate": 0.0006,
+ "loss": 4.704958438873291,
+ "step": 3754
+ },
+ {
+ "epoch": 52.15371179039301,
+ "grad_norm": 0.020777082070708275,
+ "learning_rate": 0.0006,
+ "loss": 4.638696670532227,
+ "step": 3755
+ },
+ {
+ "epoch": 52.16768558951965,
+ "grad_norm": 0.021525604650378227,
+ "learning_rate": 0.0006,
+ "loss": 4.643706798553467,
+ "step": 3756
+ },
+ {
+ "epoch": 52.18165938864629,
+ "grad_norm": 0.018943654373288155,
+ "learning_rate": 0.0006,
+ "loss": 4.5173211097717285,
+ "step": 3757
+ },
+ {
+ "epoch": 52.19563318777293,
+ "grad_norm": 0.01856580562889576,
+ "learning_rate": 0.0006,
+ "loss": 4.678788661956787,
+ "step": 3758
+ },
+ {
+ "epoch": 52.209606986899566,
+ "grad_norm": 0.019417259842157364,
+ "learning_rate": 0.0006,
+ "loss": 4.587244987487793,
+ "step": 3759
+ },
+ {
+ "epoch": 52.223580786026204,
+ "grad_norm": 0.01911911740899086,
+ "learning_rate": 0.0006,
+ "loss": 4.688503265380859,
+ "step": 3760
+ },
+ {
+ "epoch": 52.237554585152836,
+ "grad_norm": 0.021017879247665405,
+ "learning_rate": 0.0006,
+ "loss": 4.65829610824585,
+ "step": 3761
+ },
+ {
+ "epoch": 52.251528384279474,
+ "grad_norm": 0.02048875205218792,
+ "learning_rate": 0.0006,
+ "loss": 4.624055862426758,
+ "step": 3762
+ },
+ {
+ "epoch": 52.26550218340611,
+ "grad_norm": 0.01932772807776928,
+ "learning_rate": 0.0006,
+ "loss": 4.556562900543213,
+ "step": 3763
+ },
+ {
+ "epoch": 52.27947598253275,
+ "grad_norm": 0.01871614344418049,
+ "learning_rate": 0.0006,
+ "loss": 4.581201076507568,
+ "step": 3764
+ },
+ {
+ "epoch": 52.29344978165939,
+ "grad_norm": 0.017253106459975243,
+ "learning_rate": 0.0006,
+ "loss": 4.605219841003418,
+ "step": 3765
+ },
+ {
+ "epoch": 52.30742358078603,
+ "grad_norm": 0.01831238530576229,
+ "learning_rate": 0.0006,
+ "loss": 4.759900093078613,
+ "step": 3766
+ },
+ {
+ "epoch": 52.32139737991266,
+ "grad_norm": 0.017946207895874977,
+ "learning_rate": 0.0006,
+ "loss": 4.5598344802856445,
+ "step": 3767
+ },
+ {
+ "epoch": 52.3353711790393,
+ "grad_norm": 0.01680106669664383,
+ "learning_rate": 0.0006,
+ "loss": 4.605034828186035,
+ "step": 3768
+ },
+ {
+ "epoch": 52.34934497816594,
+ "grad_norm": 0.021482540294528008,
+ "learning_rate": 0.0006,
+ "loss": 4.620411396026611,
+ "step": 3769
+ },
+ {
+ "epoch": 52.36331877729258,
+ "grad_norm": 0.02210252732038498,
+ "learning_rate": 0.0006,
+ "loss": 4.667577743530273,
+ "step": 3770
+ },
+ {
+ "epoch": 52.377292576419215,
+ "grad_norm": 0.02237485907971859,
+ "learning_rate": 0.0006,
+ "loss": 4.628812789916992,
+ "step": 3771
+ },
+ {
+ "epoch": 52.391266375545854,
+ "grad_norm": 0.02196541801095009,
+ "learning_rate": 0.0006,
+ "loss": 4.572988510131836,
+ "step": 3772
+ },
+ {
+ "epoch": 52.40524017467249,
+ "grad_norm": 0.01930229924619198,
+ "learning_rate": 0.0006,
+ "loss": 4.588339805603027,
+ "step": 3773
+ },
+ {
+ "epoch": 52.419213973799124,
+ "grad_norm": 0.02222434990108013,
+ "learning_rate": 0.0006,
+ "loss": 4.766482353210449,
+ "step": 3774
+ },
+ {
+ "epoch": 52.43318777292576,
+ "grad_norm": 0.028806306421756744,
+ "learning_rate": 0.0006,
+ "loss": 4.617892742156982,
+ "step": 3775
+ },
+ {
+ "epoch": 52.4471615720524,
+ "grad_norm": 0.03306042402982712,
+ "learning_rate": 0.0006,
+ "loss": 4.665334701538086,
+ "step": 3776
+ },
+ {
+ "epoch": 52.46113537117904,
+ "grad_norm": 0.0358956903219223,
+ "learning_rate": 0.0006,
+ "loss": 4.566668510437012,
+ "step": 3777
+ },
+ {
+ "epoch": 52.47510917030568,
+ "grad_norm": 0.030406497418880463,
+ "learning_rate": 0.0006,
+ "loss": 4.513612270355225,
+ "step": 3778
+ },
+ {
+ "epoch": 52.48908296943232,
+ "grad_norm": 0.030295712873339653,
+ "learning_rate": 0.0006,
+ "loss": 4.640189170837402,
+ "step": 3779
+ },
+ {
+ "epoch": 52.50305676855895,
+ "grad_norm": 0.024981003254652023,
+ "learning_rate": 0.0006,
+ "loss": 4.57558012008667,
+ "step": 3780
+ },
+ {
+ "epoch": 52.51703056768559,
+ "grad_norm": 0.024990713223814964,
+ "learning_rate": 0.0006,
+ "loss": 4.708589553833008,
+ "step": 3781
+ },
+ {
+ "epoch": 52.531004366812226,
+ "grad_norm": 0.025034775957465172,
+ "learning_rate": 0.0006,
+ "loss": 4.70468282699585,
+ "step": 3782
+ },
+ {
+ "epoch": 52.544978165938865,
+ "grad_norm": 0.028243759647011757,
+ "learning_rate": 0.0006,
+ "loss": 4.712060451507568,
+ "step": 3783
+ },
+ {
+ "epoch": 52.5589519650655,
+ "grad_norm": 0.026984544470906258,
+ "learning_rate": 0.0006,
+ "loss": 4.505004405975342,
+ "step": 3784
+ },
+ {
+ "epoch": 52.57292576419214,
+ "grad_norm": 0.050843529403209686,
+ "learning_rate": 0.0006,
+ "loss": 4.593271255493164,
+ "step": 3785
+ },
+ {
+ "epoch": 52.58689956331878,
+ "grad_norm": 0.25422802567481995,
+ "learning_rate": 0.0006,
+ "loss": 4.7178802490234375,
+ "step": 3786
+ },
+ {
+ "epoch": 52.60087336244541,
+ "grad_norm": 0.6078406572341919,
+ "learning_rate": 0.0006,
+ "loss": 6.144372940063477,
+ "step": 3787
+ },
+ {
+ "epoch": 52.61484716157205,
+ "grad_norm": 0.25933101773262024,
+ "learning_rate": 0.0006,
+ "loss": 5.971083641052246,
+ "step": 3788
+ },
+ {
+ "epoch": 52.62882096069869,
+ "grad_norm": 0.294657826423645,
+ "learning_rate": 0.0006,
+ "loss": 6.204090118408203,
+ "step": 3789
+ },
+ {
+ "epoch": 52.64279475982533,
+ "grad_norm": 0.23641522228717804,
+ "learning_rate": 0.0006,
+ "loss": 6.869574546813965,
+ "step": 3790
+ },
+ {
+ "epoch": 52.65676855895197,
+ "grad_norm": 0.23717311024665833,
+ "learning_rate": 0.0006,
+ "loss": 6.739034175872803,
+ "step": 3791
+ },
+ {
+ "epoch": 52.670742358078606,
+ "grad_norm": 0.10300362855195999,
+ "learning_rate": 0.0006,
+ "loss": 6.49191951751709,
+ "step": 3792
+ },
+ {
+ "epoch": 52.68471615720524,
+ "grad_norm": 0.1484968513250351,
+ "learning_rate": 0.0006,
+ "loss": 6.6021270751953125,
+ "step": 3793
+ },
+ {
+ "epoch": 52.698689956331876,
+ "grad_norm": 0.08052115142345428,
+ "learning_rate": 0.0006,
+ "loss": 6.432796478271484,
+ "step": 3794
+ },
+ {
+ "epoch": 52.712663755458514,
+ "grad_norm": 0.06869253516197205,
+ "learning_rate": 0.0006,
+ "loss": 6.212784767150879,
+ "step": 3795
+ },
+ {
+ "epoch": 52.72663755458515,
+ "grad_norm": 0.05589582398533821,
+ "learning_rate": 0.0006,
+ "loss": 6.168288707733154,
+ "step": 3796
+ },
+ {
+ "epoch": 52.74061135371179,
+ "grad_norm": 0.04840940609574318,
+ "learning_rate": 0.0006,
+ "loss": 6.163174629211426,
+ "step": 3797
+ },
+ {
+ "epoch": 52.75458515283843,
+ "grad_norm": 0.04578601196408272,
+ "learning_rate": 0.0006,
+ "loss": 6.038390636444092,
+ "step": 3798
+ },
+ {
+ "epoch": 52.76855895196506,
+ "grad_norm": 0.04102053865790367,
+ "learning_rate": 0.0006,
+ "loss": 5.929539680480957,
+ "step": 3799
+ },
+ {
+ "epoch": 52.7825327510917,
+ "grad_norm": 0.038398485630750656,
+ "learning_rate": 0.0006,
+ "loss": 6.007378101348877,
+ "step": 3800
+ },
+ {
+ "epoch": 52.79650655021834,
+ "grad_norm": 0.03466642275452614,
+ "learning_rate": 0.0006,
+ "loss": 5.767291069030762,
+ "step": 3801
+ },
+ {
+ "epoch": 52.81048034934498,
+ "grad_norm": 0.03328905999660492,
+ "learning_rate": 0.0006,
+ "loss": 5.785032272338867,
+ "step": 3802
+ },
+ {
+ "epoch": 52.82445414847162,
+ "grad_norm": 0.03186681121587753,
+ "learning_rate": 0.0006,
+ "loss": 5.6942057609558105,
+ "step": 3803
+ },
+ {
+ "epoch": 52.838427947598255,
+ "grad_norm": 0.030553551390767097,
+ "learning_rate": 0.0006,
+ "loss": 5.680774688720703,
+ "step": 3804
+ },
+ {
+ "epoch": 52.852401746724894,
+ "grad_norm": 0.02578839473426342,
+ "learning_rate": 0.0006,
+ "loss": 5.703461647033691,
+ "step": 3805
+ },
+ {
+ "epoch": 52.866375545851525,
+ "grad_norm": 0.026261312887072563,
+ "learning_rate": 0.0006,
+ "loss": 5.642549514770508,
+ "step": 3806
+ },
+ {
+ "epoch": 52.880349344978164,
+ "grad_norm": 0.02782430872321129,
+ "learning_rate": 0.0006,
+ "loss": 5.563597679138184,
+ "step": 3807
+ },
+ {
+ "epoch": 52.8943231441048,
+ "grad_norm": 0.024467799812555313,
+ "learning_rate": 0.0006,
+ "loss": 5.583395957946777,
+ "step": 3808
+ },
+ {
+ "epoch": 52.90829694323144,
+ "grad_norm": 0.022602897137403488,
+ "learning_rate": 0.0006,
+ "loss": 5.50742769241333,
+ "step": 3809
+ },
+ {
+ "epoch": 52.92227074235808,
+ "grad_norm": 0.024292029440402985,
+ "learning_rate": 0.0006,
+ "loss": 5.367804527282715,
+ "step": 3810
+ },
+ {
+ "epoch": 52.93624454148472,
+ "grad_norm": 0.021062295883893967,
+ "learning_rate": 0.0006,
+ "loss": 5.423705101013184,
+ "step": 3811
+ },
+ {
+ "epoch": 52.95021834061135,
+ "grad_norm": 0.022226469591259956,
+ "learning_rate": 0.0006,
+ "loss": 5.344039440155029,
+ "step": 3812
+ },
+ {
+ "epoch": 52.96419213973799,
+ "grad_norm": 0.023691752925515175,
+ "learning_rate": 0.0006,
+ "loss": 5.2581892013549805,
+ "step": 3813
+ },
+ {
+ "epoch": 52.97816593886463,
+ "grad_norm": 0.02181999757885933,
+ "learning_rate": 0.0006,
+ "loss": 5.2710113525390625,
+ "step": 3814
+ },
+ {
+ "epoch": 52.992139737991266,
+ "grad_norm": 0.025295857340097427,
+ "learning_rate": 0.0006,
+ "loss": 5.278355598449707,
+ "step": 3815
+ },
+ {
+ "epoch": 53.0,
+ "grad_norm": 0.02216828241944313,
+ "learning_rate": 0.0006,
+ "loss": 5.127945899963379,
+ "step": 3816
+ },
+ {
+ "epoch": 53.0,
+ "eval_loss": 5.3411736488342285,
+ "eval_runtime": 56.2837,
+ "eval_samples_per_second": 43.387,
+ "eval_steps_per_second": 1.368,
+ "step": 3816
+ },
+ {
+ "epoch": 53.01397379912664,
+ "grad_norm": 0.022496026009321213,
+ "learning_rate": 0.0006,
+ "loss": 5.19655704498291,
+ "step": 3817
+ },
+ {
+ "epoch": 53.02794759825328,
+ "grad_norm": 0.03201168775558472,
+ "learning_rate": 0.0006,
+ "loss": 5.094621181488037,
+ "step": 3818
+ },
+ {
+ "epoch": 53.041921397379916,
+ "grad_norm": 0.02683178335428238,
+ "learning_rate": 0.0006,
+ "loss": 5.19899845123291,
+ "step": 3819
+ },
+ {
+ "epoch": 53.05589519650655,
+ "grad_norm": 0.022659169510006905,
+ "learning_rate": 0.0006,
+ "loss": 5.184103012084961,
+ "step": 3820
+ },
+ {
+ "epoch": 53.069868995633186,
+ "grad_norm": 0.024784443899989128,
+ "learning_rate": 0.0006,
+ "loss": 5.124286651611328,
+ "step": 3821
+ },
+ {
+ "epoch": 53.083842794759825,
+ "grad_norm": 0.028710370883345604,
+ "learning_rate": 0.0006,
+ "loss": 5.130426406860352,
+ "step": 3822
+ },
+ {
+ "epoch": 53.09781659388646,
+ "grad_norm": 0.02366088517010212,
+ "learning_rate": 0.0006,
+ "loss": 5.1453399658203125,
+ "step": 3823
+ },
+ {
+ "epoch": 53.1117903930131,
+ "grad_norm": 0.01813938096165657,
+ "learning_rate": 0.0006,
+ "loss": 5.05271053314209,
+ "step": 3824
+ },
+ {
+ "epoch": 53.12576419213974,
+ "grad_norm": 0.02051435597240925,
+ "learning_rate": 0.0006,
+ "loss": 5.023038864135742,
+ "step": 3825
+ },
+ {
+ "epoch": 53.13973799126637,
+ "grad_norm": 0.02185754105448723,
+ "learning_rate": 0.0006,
+ "loss": 4.8950018882751465,
+ "step": 3826
+ },
+ {
+ "epoch": 53.15371179039301,
+ "grad_norm": 0.017223743721842766,
+ "learning_rate": 0.0006,
+ "loss": 5.078621864318848,
+ "step": 3827
+ },
+ {
+ "epoch": 53.16768558951965,
+ "grad_norm": 0.01832217164337635,
+ "learning_rate": 0.0006,
+ "loss": 4.921341419219971,
+ "step": 3828
+ },
+ {
+ "epoch": 53.18165938864629,
+ "grad_norm": 0.02114005573093891,
+ "learning_rate": 0.0006,
+ "loss": 4.997865676879883,
+ "step": 3829
+ },
+ {
+ "epoch": 53.19563318777293,
+ "grad_norm": 0.021426431834697723,
+ "learning_rate": 0.0006,
+ "loss": 4.877785682678223,
+ "step": 3830
+ },
+ {
+ "epoch": 53.209606986899566,
+ "grad_norm": 0.021327560767531395,
+ "learning_rate": 0.0006,
+ "loss": 4.847512245178223,
+ "step": 3831
+ },
+ {
+ "epoch": 53.223580786026204,
+ "grad_norm": 0.027825530618429184,
+ "learning_rate": 0.0006,
+ "loss": 5.004823207855225,
+ "step": 3832
+ },
+ {
+ "epoch": 53.237554585152836,
+ "grad_norm": 0.03756481409072876,
+ "learning_rate": 0.0006,
+ "loss": 4.817290306091309,
+ "step": 3833
+ },
+ {
+ "epoch": 53.251528384279474,
+ "grad_norm": 0.04933721572160721,
+ "learning_rate": 0.0006,
+ "loss": 4.863317489624023,
+ "step": 3834
+ },
+ {
+ "epoch": 53.26550218340611,
+ "grad_norm": 0.04588694870471954,
+ "learning_rate": 0.0006,
+ "loss": 4.864742755889893,
+ "step": 3835
+ },
+ {
+ "epoch": 53.27947598253275,
+ "grad_norm": 0.02728264033794403,
+ "learning_rate": 0.0006,
+ "loss": 4.9068756103515625,
+ "step": 3836
+ },
+ {
+ "epoch": 53.29344978165939,
+ "grad_norm": 0.036240726709365845,
+ "learning_rate": 0.0006,
+ "loss": 4.902045249938965,
+ "step": 3837
+ },
+ {
+ "epoch": 53.30742358078603,
+ "grad_norm": 0.02703195810317993,
+ "learning_rate": 0.0006,
+ "loss": 4.908980369567871,
+ "step": 3838
+ },
+ {
+ "epoch": 53.32139737991266,
+ "grad_norm": 0.024894049391150475,
+ "learning_rate": 0.0006,
+ "loss": 4.794848442077637,
+ "step": 3839
+ },
+ {
+ "epoch": 53.3353711790393,
+ "grad_norm": 0.02389993891119957,
+ "learning_rate": 0.0006,
+ "loss": 4.8234734535217285,
+ "step": 3840
+ },
+ {
+ "epoch": 53.34934497816594,
+ "grad_norm": 0.02486400306224823,
+ "learning_rate": 0.0006,
+ "loss": 4.851463794708252,
+ "step": 3841
+ },
+ {
+ "epoch": 53.36331877729258,
+ "grad_norm": 0.02613813802599907,
+ "learning_rate": 0.0006,
+ "loss": 4.860444068908691,
+ "step": 3842
+ },
+ {
+ "epoch": 53.377292576419215,
+ "grad_norm": 0.0229744054377079,
+ "learning_rate": 0.0006,
+ "loss": 4.709606170654297,
+ "step": 3843
+ },
+ {
+ "epoch": 53.391266375545854,
+ "grad_norm": 0.02123042568564415,
+ "learning_rate": 0.0006,
+ "loss": 4.654613494873047,
+ "step": 3844
+ },
+ {
+ "epoch": 53.40524017467249,
+ "grad_norm": 0.017352338880300522,
+ "learning_rate": 0.0006,
+ "loss": 4.769761085510254,
+ "step": 3845
+ },
+ {
+ "epoch": 53.419213973799124,
+ "grad_norm": 0.019808724522590637,
+ "learning_rate": 0.0006,
+ "loss": 4.98514461517334,
+ "step": 3846
+ },
+ {
+ "epoch": 53.43318777292576,
+ "grad_norm": 0.018776152282953262,
+ "learning_rate": 0.0006,
+ "loss": 4.749990463256836,
+ "step": 3847
+ },
+ {
+ "epoch": 53.4471615720524,
+ "grad_norm": 0.02017521858215332,
+ "learning_rate": 0.0006,
+ "loss": 4.757381439208984,
+ "step": 3848
+ },
+ {
+ "epoch": 53.46113537117904,
+ "grad_norm": 0.019509416073560715,
+ "learning_rate": 0.0006,
+ "loss": 4.8111572265625,
+ "step": 3849
+ },
+ {
+ "epoch": 53.47510917030568,
+ "grad_norm": 0.015543580986559391,
+ "learning_rate": 0.0006,
+ "loss": 4.7501678466796875,
+ "step": 3850
+ },
+ {
+ "epoch": 53.48908296943232,
+ "grad_norm": 0.017941651865839958,
+ "learning_rate": 0.0006,
+ "loss": 4.701838493347168,
+ "step": 3851
+ },
+ {
+ "epoch": 53.50305676855895,
+ "grad_norm": 0.01678212359547615,
+ "learning_rate": 0.0006,
+ "loss": 4.787798881530762,
+ "step": 3852
+ },
+ {
+ "epoch": 53.51703056768559,
+ "grad_norm": 0.019050654023885727,
+ "learning_rate": 0.0006,
+ "loss": 4.754859924316406,
+ "step": 3853
+ },
+ {
+ "epoch": 53.531004366812226,
+ "grad_norm": 0.017406078055500984,
+ "learning_rate": 0.0006,
+ "loss": 4.752772808074951,
+ "step": 3854
+ },
+ {
+ "epoch": 53.544978165938865,
+ "grad_norm": 0.01773344725370407,
+ "learning_rate": 0.0006,
+ "loss": 4.651812553405762,
+ "step": 3855
+ },
+ {
+ "epoch": 53.5589519650655,
+ "grad_norm": 0.017221346497535706,
+ "learning_rate": 0.0006,
+ "loss": 4.792244911193848,
+ "step": 3856
+ },
+ {
+ "epoch": 53.57292576419214,
+ "grad_norm": 0.01722758449614048,
+ "learning_rate": 0.0006,
+ "loss": 4.7902960777282715,
+ "step": 3857
+ },
+ {
+ "epoch": 53.58689956331878,
+ "grad_norm": 0.020445192232728004,
+ "learning_rate": 0.0006,
+ "loss": 4.67392110824585,
+ "step": 3858
+ },
+ {
+ "epoch": 53.60087336244541,
+ "grad_norm": 0.02267114631831646,
+ "learning_rate": 0.0006,
+ "loss": 4.796089172363281,
+ "step": 3859
+ },
+ {
+ "epoch": 53.61484716157205,
+ "grad_norm": 0.027129942551255226,
+ "learning_rate": 0.0006,
+ "loss": 4.774303913116455,
+ "step": 3860
+ },
+ {
+ "epoch": 53.62882096069869,
+ "grad_norm": 0.025454839691519737,
+ "learning_rate": 0.0006,
+ "loss": 4.708813190460205,
+ "step": 3861
+ },
+ {
+ "epoch": 53.64279475982533,
+ "grad_norm": 0.019248757511377335,
+ "learning_rate": 0.0006,
+ "loss": 4.760467529296875,
+ "step": 3862
+ },
+ {
+ "epoch": 53.65676855895197,
+ "grad_norm": 0.01791176199913025,
+ "learning_rate": 0.0006,
+ "loss": 4.789203643798828,
+ "step": 3863
+ },
+ {
+ "epoch": 53.670742358078606,
+ "grad_norm": 0.016306612640619278,
+ "learning_rate": 0.0006,
+ "loss": 4.732605934143066,
+ "step": 3864
+ },
+ {
+ "epoch": 53.68471615720524,
+ "grad_norm": 0.017836976796388626,
+ "learning_rate": 0.0006,
+ "loss": 4.691871643066406,
+ "step": 3865
+ },
+ {
+ "epoch": 53.698689956331876,
+ "grad_norm": 0.015102861449122429,
+ "learning_rate": 0.0006,
+ "loss": 4.724156856536865,
+ "step": 3866
+ },
+ {
+ "epoch": 53.712663755458514,
+ "grad_norm": 0.017458142712712288,
+ "learning_rate": 0.0006,
+ "loss": 4.73524808883667,
+ "step": 3867
+ },
+ {
+ "epoch": 53.72663755458515,
+ "grad_norm": 0.01591932773590088,
+ "learning_rate": 0.0006,
+ "loss": 4.705888748168945,
+ "step": 3868
+ },
+ {
+ "epoch": 53.74061135371179,
+ "grad_norm": 0.01713593676686287,
+ "learning_rate": 0.0006,
+ "loss": 4.644516944885254,
+ "step": 3869
+ },
+ {
+ "epoch": 53.75458515283843,
+ "grad_norm": 0.01905885338783264,
+ "learning_rate": 0.0006,
+ "loss": 4.698416233062744,
+ "step": 3870
+ },
+ {
+ "epoch": 53.76855895196506,
+ "grad_norm": 0.01766437292098999,
+ "learning_rate": 0.0006,
+ "loss": 4.6796555519104,
+ "step": 3871
+ },
+ {
+ "epoch": 53.7825327510917,
+ "grad_norm": 0.01552243810147047,
+ "learning_rate": 0.0006,
+ "loss": 4.78138542175293,
+ "step": 3872
+ },
+ {
+ "epoch": 53.79650655021834,
+ "grad_norm": 0.014191331341862679,
+ "learning_rate": 0.0006,
+ "loss": 4.703588485717773,
+ "step": 3873
+ },
+ {
+ "epoch": 53.81048034934498,
+ "grad_norm": 0.015872279182076454,
+ "learning_rate": 0.0006,
+ "loss": 4.705611228942871,
+ "step": 3874
+ },
+ {
+ "epoch": 53.82445414847162,
+ "grad_norm": 0.01787560060620308,
+ "learning_rate": 0.0006,
+ "loss": 4.8227949142456055,
+ "step": 3875
+ },
+ {
+ "epoch": 53.838427947598255,
+ "grad_norm": 0.0158209390938282,
+ "learning_rate": 0.0006,
+ "loss": 4.645227432250977,
+ "step": 3876
+ },
+ {
+ "epoch": 53.852401746724894,
+ "grad_norm": 0.014362678863108158,
+ "learning_rate": 0.0006,
+ "loss": 4.7225422859191895,
+ "step": 3877
+ },
+ {
+ "epoch": 53.866375545851525,
+ "grad_norm": 0.014899161644279957,
+ "learning_rate": 0.0006,
+ "loss": 4.685914039611816,
+ "step": 3878
+ },
+ {
+ "epoch": 53.880349344978164,
+ "grad_norm": 0.01628505438566208,
+ "learning_rate": 0.0006,
+ "loss": 4.674065589904785,
+ "step": 3879
+ },
+ {
+ "epoch": 53.8943231441048,
+ "grad_norm": 0.01873331144452095,
+ "learning_rate": 0.0006,
+ "loss": 4.811744213104248,
+ "step": 3880
+ },
+ {
+ "epoch": 53.90829694323144,
+ "grad_norm": 0.02229069173336029,
+ "learning_rate": 0.0006,
+ "loss": 4.721514701843262,
+ "step": 3881
+ },
+ {
+ "epoch": 53.92227074235808,
+ "grad_norm": 0.028012916445732117,
+ "learning_rate": 0.0006,
+ "loss": 4.773603439331055,
+ "step": 3882
+ },
+ {
+ "epoch": 53.93624454148472,
+ "grad_norm": 0.036322783678770065,
+ "learning_rate": 0.0006,
+ "loss": 4.787134170532227,
+ "step": 3883
+ },
+ {
+ "epoch": 53.95021834061135,
+ "grad_norm": 0.048149142414331436,
+ "learning_rate": 0.0006,
+ "loss": 4.736196517944336,
+ "step": 3884
+ },
+ {
+ "epoch": 53.96419213973799,
+ "grad_norm": 0.051918916404247284,
+ "learning_rate": 0.0006,
+ "loss": 4.686728000640869,
+ "step": 3885
+ },
+ {
+ "epoch": 53.97816593886463,
+ "grad_norm": 0.0481342189013958,
+ "learning_rate": 0.0006,
+ "loss": 4.733730792999268,
+ "step": 3886
+ },
+ {
+ "epoch": 53.992139737991266,
+ "grad_norm": 0.045424818992614746,
+ "learning_rate": 0.0006,
+ "loss": 4.748723983764648,
+ "step": 3887
+ },
+ {
+ "epoch": 54.0,
+ "grad_norm": 0.04153316840529442,
+ "learning_rate": 0.0006,
+ "loss": 4.715220928192139,
+ "step": 3888
+ },
+ {
+ "epoch": 54.0,
+ "eval_loss": 4.964215278625488,
+ "eval_runtime": 56.7757,
+ "eval_samples_per_second": 43.011,
+ "eval_steps_per_second": 1.356,
+ "step": 3888
+ },
+ {
+ "epoch": 54.01397379912664,
+ "grad_norm": 0.03459750860929489,
+ "learning_rate": 0.0006,
+ "loss": 4.661557674407959,
+ "step": 3889
+ },
+ {
+ "epoch": 54.02794759825328,
+ "grad_norm": 0.025409208610653877,
+ "learning_rate": 0.0006,
+ "loss": 4.7065582275390625,
+ "step": 3890
+ },
+ {
+ "epoch": 54.041921397379916,
+ "grad_norm": 0.03202550485730171,
+ "learning_rate": 0.0006,
+ "loss": 4.751070976257324,
+ "step": 3891
+ },
+ {
+ "epoch": 54.05589519650655,
+ "grad_norm": 0.02968018501996994,
+ "learning_rate": 0.0006,
+ "loss": 4.635388374328613,
+ "step": 3892
+ },
+ {
+ "epoch": 54.069868995633186,
+ "grad_norm": 0.02757406234741211,
+ "learning_rate": 0.0006,
+ "loss": 4.706214427947998,
+ "step": 3893
+ },
+ {
+ "epoch": 54.083842794759825,
+ "grad_norm": 0.025223620235919952,
+ "learning_rate": 0.0006,
+ "loss": 4.703792572021484,
+ "step": 3894
+ },
+ {
+ "epoch": 54.09781659388646,
+ "grad_norm": 0.02431778982281685,
+ "learning_rate": 0.0006,
+ "loss": 4.658116340637207,
+ "step": 3895
+ },
+ {
+ "epoch": 54.1117903930131,
+ "grad_norm": 0.02792172320187092,
+ "learning_rate": 0.0006,
+ "loss": 4.638895034790039,
+ "step": 3896
+ },
+ {
+ "epoch": 54.12576419213974,
+ "grad_norm": 0.032253511250019073,
+ "learning_rate": 0.0006,
+ "loss": 4.693902492523193,
+ "step": 3897
+ },
+ {
+ "epoch": 54.13973799126637,
+ "grad_norm": 0.027044931426644325,
+ "learning_rate": 0.0006,
+ "loss": 4.575805187225342,
+ "step": 3898
+ },
+ {
+ "epoch": 54.15371179039301,
+ "grad_norm": 0.023270361125469208,
+ "learning_rate": 0.0006,
+ "loss": 4.67645263671875,
+ "step": 3899
+ },
+ {
+ "epoch": 54.16768558951965,
+ "grad_norm": 0.01929079182446003,
+ "learning_rate": 0.0006,
+ "loss": 4.68062686920166,
+ "step": 3900
+ },
+ {
+ "epoch": 54.18165938864629,
+ "grad_norm": 0.020343473181128502,
+ "learning_rate": 0.0006,
+ "loss": 4.708086967468262,
+ "step": 3901
+ },
+ {
+ "epoch": 54.19563318777293,
+ "grad_norm": 0.0211793202906847,
+ "learning_rate": 0.0006,
+ "loss": 4.677323341369629,
+ "step": 3902
+ },
+ {
+ "epoch": 54.209606986899566,
+ "grad_norm": 0.020341217517852783,
+ "learning_rate": 0.0006,
+ "loss": 4.760937690734863,
+ "step": 3903
+ },
+ {
+ "epoch": 54.223580786026204,
+ "grad_norm": 0.019421054050326347,
+ "learning_rate": 0.0006,
+ "loss": 4.508834362030029,
+ "step": 3904
+ },
+ {
+ "epoch": 54.237554585152836,
+ "grad_norm": 0.016701679676771164,
+ "learning_rate": 0.0006,
+ "loss": 4.664399147033691,
+ "step": 3905
+ },
+ {
+ "epoch": 54.251528384279474,
+ "grad_norm": 0.016561470925807953,
+ "learning_rate": 0.0006,
+ "loss": 4.742990493774414,
+ "step": 3906
+ },
+ {
+ "epoch": 54.26550218340611,
+ "grad_norm": 0.017446085810661316,
+ "learning_rate": 0.0006,
+ "loss": 4.63038444519043,
+ "step": 3907
+ },
+ {
+ "epoch": 54.27947598253275,
+ "grad_norm": 0.014340086840093136,
+ "learning_rate": 0.0006,
+ "loss": 4.688791275024414,
+ "step": 3908
+ },
+ {
+ "epoch": 54.29344978165939,
+ "grad_norm": 0.013580698519945145,
+ "learning_rate": 0.0006,
+ "loss": 4.625302314758301,
+ "step": 3909
+ },
+ {
+ "epoch": 54.30742358078603,
+ "grad_norm": 0.014931539073586464,
+ "learning_rate": 0.0006,
+ "loss": 4.6167497634887695,
+ "step": 3910
+ },
+ {
+ "epoch": 54.32139737991266,
+ "grad_norm": 0.013492150232195854,
+ "learning_rate": 0.0006,
+ "loss": 4.657521724700928,
+ "step": 3911
+ },
+ {
+ "epoch": 54.3353711790393,
+ "grad_norm": 0.015151413157582283,
+ "learning_rate": 0.0006,
+ "loss": 4.634521484375,
+ "step": 3912
+ },
+ {
+ "epoch": 54.34934497816594,
+ "grad_norm": 0.01302040833979845,
+ "learning_rate": 0.0006,
+ "loss": 4.718928337097168,
+ "step": 3913
+ },
+ {
+ "epoch": 54.36331877729258,
+ "grad_norm": 0.013433529995381832,
+ "learning_rate": 0.0006,
+ "loss": 4.584803581237793,
+ "step": 3914
+ },
+ {
+ "epoch": 54.377292576419215,
+ "grad_norm": 0.012987039051949978,
+ "learning_rate": 0.0006,
+ "loss": 4.6393585205078125,
+ "step": 3915
+ },
+ {
+ "epoch": 54.391266375545854,
+ "grad_norm": 0.013695075176656246,
+ "learning_rate": 0.0006,
+ "loss": 4.556313514709473,
+ "step": 3916
+ },
+ {
+ "epoch": 54.40524017467249,
+ "grad_norm": 0.013718255795538425,
+ "learning_rate": 0.0006,
+ "loss": 4.576223850250244,
+ "step": 3917
+ },
+ {
+ "epoch": 54.419213973799124,
+ "grad_norm": 0.013699112460017204,
+ "learning_rate": 0.0006,
+ "loss": 4.63889217376709,
+ "step": 3918
+ },
+ {
+ "epoch": 54.43318777292576,
+ "grad_norm": 0.013516264036297798,
+ "learning_rate": 0.0006,
+ "loss": 4.669532299041748,
+ "step": 3919
+ },
+ {
+ "epoch": 54.4471615720524,
+ "grad_norm": 0.013597294688224792,
+ "learning_rate": 0.0006,
+ "loss": 4.6644287109375,
+ "step": 3920
+ },
+ {
+ "epoch": 54.46113537117904,
+ "grad_norm": 0.013185849413275719,
+ "learning_rate": 0.0006,
+ "loss": 4.636900901794434,
+ "step": 3921
+ },
+ {
+ "epoch": 54.47510917030568,
+ "grad_norm": 0.014311990700662136,
+ "learning_rate": 0.0006,
+ "loss": 4.6047587394714355,
+ "step": 3922
+ },
+ {
+ "epoch": 54.48908296943232,
+ "grad_norm": 0.012426027096807957,
+ "learning_rate": 0.0006,
+ "loss": 4.6161394119262695,
+ "step": 3923
+ },
+ {
+ "epoch": 54.50305676855895,
+ "grad_norm": 0.01283493172377348,
+ "learning_rate": 0.0006,
+ "loss": 4.584190368652344,
+ "step": 3924
+ },
+ {
+ "epoch": 54.51703056768559,
+ "grad_norm": 0.01339943427592516,
+ "learning_rate": 0.0006,
+ "loss": 4.624184608459473,
+ "step": 3925
+ },
+ {
+ "epoch": 54.531004366812226,
+ "grad_norm": 0.012859724462032318,
+ "learning_rate": 0.0006,
+ "loss": 4.6174635887146,
+ "step": 3926
+ },
+ {
+ "epoch": 54.544978165938865,
+ "grad_norm": 0.014420023187994957,
+ "learning_rate": 0.0006,
+ "loss": 4.536544322967529,
+ "step": 3927
+ },
+ {
+ "epoch": 54.5589519650655,
+ "grad_norm": 0.014237185008823872,
+ "learning_rate": 0.0006,
+ "loss": 4.681514739990234,
+ "step": 3928
+ },
+ {
+ "epoch": 54.57292576419214,
+ "grad_norm": 0.013340328820049763,
+ "learning_rate": 0.0006,
+ "loss": 4.579689979553223,
+ "step": 3929
+ },
+ {
+ "epoch": 54.58689956331878,
+ "grad_norm": 0.013894530944526196,
+ "learning_rate": 0.0006,
+ "loss": 4.638958930969238,
+ "step": 3930
+ },
+ {
+ "epoch": 54.60087336244541,
+ "grad_norm": 0.016181141138076782,
+ "learning_rate": 0.0006,
+ "loss": 4.573508262634277,
+ "step": 3931
+ },
+ {
+ "epoch": 54.61484716157205,
+ "grad_norm": 0.013221992179751396,
+ "learning_rate": 0.0006,
+ "loss": 4.6229047775268555,
+ "step": 3932
+ },
+ {
+ "epoch": 54.62882096069869,
+ "grad_norm": 0.013241561129689217,
+ "learning_rate": 0.0006,
+ "loss": 4.641152858734131,
+ "step": 3933
+ },
+ {
+ "epoch": 54.64279475982533,
+ "grad_norm": 0.016237538307905197,
+ "learning_rate": 0.0006,
+ "loss": 4.6421217918396,
+ "step": 3934
+ },
+ {
+ "epoch": 54.65676855895197,
+ "grad_norm": 0.020866507664322853,
+ "learning_rate": 0.0006,
+ "loss": 4.646058082580566,
+ "step": 3935
+ },
+ {
+ "epoch": 54.670742358078606,
+ "grad_norm": 0.022182893007993698,
+ "learning_rate": 0.0006,
+ "loss": 4.554501533508301,
+ "step": 3936
+ },
+ {
+ "epoch": 54.68471615720524,
+ "grad_norm": 0.02254868485033512,
+ "learning_rate": 0.0006,
+ "loss": 4.563869953155518,
+ "step": 3937
+ },
+ {
+ "epoch": 54.698689956331876,
+ "grad_norm": 0.019730940461158752,
+ "learning_rate": 0.0006,
+ "loss": 4.592165946960449,
+ "step": 3938
+ },
+ {
+ "epoch": 54.712663755458514,
+ "grad_norm": 0.015146561898291111,
+ "learning_rate": 0.0006,
+ "loss": 4.6228742599487305,
+ "step": 3939
+ },
+ {
+ "epoch": 54.72663755458515,
+ "grad_norm": 0.01824716106057167,
+ "learning_rate": 0.0006,
+ "loss": 4.716236114501953,
+ "step": 3940
+ },
+ {
+ "epoch": 54.74061135371179,
+ "grad_norm": 0.017730550840497017,
+ "learning_rate": 0.0006,
+ "loss": 4.650232315063477,
+ "step": 3941
+ },
+ {
+ "epoch": 54.75458515283843,
+ "grad_norm": 0.014875595457851887,
+ "learning_rate": 0.0006,
+ "loss": 4.6377458572387695,
+ "step": 3942
+ },
+ {
+ "epoch": 54.76855895196506,
+ "grad_norm": 0.01613209955394268,
+ "learning_rate": 0.0006,
+ "loss": 4.5768537521362305,
+ "step": 3943
+ },
+ {
+ "epoch": 54.7825327510917,
+ "grad_norm": 0.017311280593276024,
+ "learning_rate": 0.0006,
+ "loss": 4.625744342803955,
+ "step": 3944
+ },
+ {
+ "epoch": 54.79650655021834,
+ "grad_norm": 0.018770085647702217,
+ "learning_rate": 0.0006,
+ "loss": 4.613787651062012,
+ "step": 3945
+ },
+ {
+ "epoch": 54.81048034934498,
+ "grad_norm": 0.0165257528424263,
+ "learning_rate": 0.0006,
+ "loss": 4.5991315841674805,
+ "step": 3946
+ },
+ {
+ "epoch": 54.82445414847162,
+ "grad_norm": 0.017433857545256615,
+ "learning_rate": 0.0006,
+ "loss": 4.513872146606445,
+ "step": 3947
+ },
+ {
+ "epoch": 54.838427947598255,
+ "grad_norm": 0.020126905292272568,
+ "learning_rate": 0.0006,
+ "loss": 4.616185188293457,
+ "step": 3948
+ },
+ {
+ "epoch": 54.852401746724894,
+ "grad_norm": 0.01819358579814434,
+ "learning_rate": 0.0006,
+ "loss": 4.535691261291504,
+ "step": 3949
+ },
+ {
+ "epoch": 54.866375545851525,
+ "grad_norm": 0.01899927295744419,
+ "learning_rate": 0.0006,
+ "loss": 4.595199108123779,
+ "step": 3950
+ },
+ {
+ "epoch": 54.880349344978164,
+ "grad_norm": 0.02005566656589508,
+ "learning_rate": 0.0006,
+ "loss": 4.633289337158203,
+ "step": 3951
+ },
+ {
+ "epoch": 54.8943231441048,
+ "grad_norm": 0.021570220589637756,
+ "learning_rate": 0.0006,
+ "loss": 4.592503547668457,
+ "step": 3952
+ },
+ {
+ "epoch": 54.90829694323144,
+ "grad_norm": 0.01883506216108799,
+ "learning_rate": 0.0006,
+ "loss": 4.6597795486450195,
+ "step": 3953
+ },
+ {
+ "epoch": 54.92227074235808,
+ "grad_norm": 0.018070021644234657,
+ "learning_rate": 0.0006,
+ "loss": 4.601474285125732,
+ "step": 3954
+ },
+ {
+ "epoch": 54.93624454148472,
+ "grad_norm": 0.018090378493070602,
+ "learning_rate": 0.0006,
+ "loss": 4.569846153259277,
+ "step": 3955
+ },
+ {
+ "epoch": 54.95021834061135,
+ "grad_norm": 0.018751563504338264,
+ "learning_rate": 0.0006,
+ "loss": 4.646064281463623,
+ "step": 3956
+ },
+ {
+ "epoch": 54.96419213973799,
+ "grad_norm": 0.017512254416942596,
+ "learning_rate": 0.0006,
+ "loss": 4.606670379638672,
+ "step": 3957
+ },
+ {
+ "epoch": 54.97816593886463,
+ "grad_norm": 0.018771110102534294,
+ "learning_rate": 0.0006,
+ "loss": 4.604315757751465,
+ "step": 3958
+ },
+ {
+ "epoch": 54.992139737991266,
+ "grad_norm": 0.01611141301691532,
+ "learning_rate": 0.0006,
+ "loss": 4.643245220184326,
+ "step": 3959
+ },
+ {
+ "epoch": 55.0,
+ "grad_norm": 0.0156728383153677,
+ "learning_rate": 0.0006,
+ "loss": 4.6739044189453125,
+ "step": 3960
+ },
+ {
+ "epoch": 55.0,
+ "eval_loss": 4.851032257080078,
+ "eval_runtime": 56.8699,
+ "eval_samples_per_second": 42.94,
+ "eval_steps_per_second": 1.354,
+ "step": 3960
+ },
+ {
+ "epoch": 55.01397379912664,
+ "grad_norm": 0.016214672476053238,
+ "learning_rate": 0.0006,
+ "loss": 4.571749687194824,
+ "step": 3961
+ },
+ {
+ "epoch": 55.02794759825328,
+ "grad_norm": 0.01649155467748642,
+ "learning_rate": 0.0006,
+ "loss": 4.553592205047607,
+ "step": 3962
+ },
+ {
+ "epoch": 55.041921397379916,
+ "grad_norm": 0.016793379560112953,
+ "learning_rate": 0.0006,
+ "loss": 4.650628566741943,
+ "step": 3963
+ },
+ {
+ "epoch": 55.05589519650655,
+ "grad_norm": 0.0191937405616045,
+ "learning_rate": 0.0006,
+ "loss": 4.620532989501953,
+ "step": 3964
+ },
+ {
+ "epoch": 55.069868995633186,
+ "grad_norm": 0.020976556465029716,
+ "learning_rate": 0.0006,
+ "loss": 4.6329545974731445,
+ "step": 3965
+ },
+ {
+ "epoch": 55.083842794759825,
+ "grad_norm": 0.023450477048754692,
+ "learning_rate": 0.0006,
+ "loss": 4.60828971862793,
+ "step": 3966
+ },
+ {
+ "epoch": 55.09781659388646,
+ "grad_norm": 0.026778312399983406,
+ "learning_rate": 0.0006,
+ "loss": 4.617481708526611,
+ "step": 3967
+ },
+ {
+ "epoch": 55.1117903930131,
+ "grad_norm": 0.02333568036556244,
+ "learning_rate": 0.0006,
+ "loss": 4.580722332000732,
+ "step": 3968
+ },
+ {
+ "epoch": 55.12576419213974,
+ "grad_norm": 0.018491310998797417,
+ "learning_rate": 0.0006,
+ "loss": 4.590953350067139,
+ "step": 3969
+ },
+ {
+ "epoch": 55.13973799126637,
+ "grad_norm": 0.020339803770184517,
+ "learning_rate": 0.0006,
+ "loss": 4.518374443054199,
+ "step": 3970
+ },
+ {
+ "epoch": 55.15371179039301,
+ "grad_norm": 0.01759706623852253,
+ "learning_rate": 0.0006,
+ "loss": 4.560522079467773,
+ "step": 3971
+ },
+ {
+ "epoch": 55.16768558951965,
+ "grad_norm": 0.015875980257987976,
+ "learning_rate": 0.0006,
+ "loss": 4.4795708656311035,
+ "step": 3972
+ },
+ {
+ "epoch": 55.18165938864629,
+ "grad_norm": 0.01650134287774563,
+ "learning_rate": 0.0006,
+ "loss": 4.57435941696167,
+ "step": 3973
+ },
+ {
+ "epoch": 55.19563318777293,
+ "grad_norm": 0.01734331250190735,
+ "learning_rate": 0.0006,
+ "loss": 4.661007881164551,
+ "step": 3974
+ },
+ {
+ "epoch": 55.209606986899566,
+ "grad_norm": 0.01651303470134735,
+ "learning_rate": 0.0006,
+ "loss": 4.511148452758789,
+ "step": 3975
+ },
+ {
+ "epoch": 55.223580786026204,
+ "grad_norm": 0.016189761459827423,
+ "learning_rate": 0.0006,
+ "loss": 4.592350006103516,
+ "step": 3976
+ },
+ {
+ "epoch": 55.237554585152836,
+ "grad_norm": 0.01833222061395645,
+ "learning_rate": 0.0006,
+ "loss": 4.506649971008301,
+ "step": 3977
+ },
+ {
+ "epoch": 55.251528384279474,
+ "grad_norm": 0.02433410845696926,
+ "learning_rate": 0.0006,
+ "loss": 4.49500036239624,
+ "step": 3978
+ },
+ {
+ "epoch": 55.26550218340611,
+ "grad_norm": 0.0314289927482605,
+ "learning_rate": 0.0006,
+ "loss": 4.507624626159668,
+ "step": 3979
+ },
+ {
+ "epoch": 55.27947598253275,
+ "grad_norm": 0.03013679012656212,
+ "learning_rate": 0.0006,
+ "loss": 4.533641815185547,
+ "step": 3980
+ },
+ {
+ "epoch": 55.29344978165939,
+ "grad_norm": 0.024524061009287834,
+ "learning_rate": 0.0006,
+ "loss": 4.5721116065979,
+ "step": 3981
+ },
+ {
+ "epoch": 55.30742358078603,
+ "grad_norm": 0.02640336938202381,
+ "learning_rate": 0.0006,
+ "loss": 4.652521133422852,
+ "step": 3982
+ },
+ {
+ "epoch": 55.32139737991266,
+ "grad_norm": 0.021309101954102516,
+ "learning_rate": 0.0006,
+ "loss": 4.600375175476074,
+ "step": 3983
+ },
+ {
+ "epoch": 55.3353711790393,
+ "grad_norm": 0.023008549585938454,
+ "learning_rate": 0.0006,
+ "loss": 4.53914737701416,
+ "step": 3984
+ },
+ {
+ "epoch": 55.34934497816594,
+ "grad_norm": 0.022491468116641045,
+ "learning_rate": 0.0006,
+ "loss": 4.661163330078125,
+ "step": 3985
+ },
+ {
+ "epoch": 55.36331877729258,
+ "grad_norm": 0.01904081553220749,
+ "learning_rate": 0.0006,
+ "loss": 4.5426177978515625,
+ "step": 3986
+ },
+ {
+ "epoch": 55.377292576419215,
+ "grad_norm": 0.019229518249630928,
+ "learning_rate": 0.0006,
+ "loss": 4.606066703796387,
+ "step": 3987
+ },
+ {
+ "epoch": 55.391266375545854,
+ "grad_norm": 0.01781383901834488,
+ "learning_rate": 0.0006,
+ "loss": 4.54642391204834,
+ "step": 3988
+ },
+ {
+ "epoch": 55.40524017467249,
+ "grad_norm": 0.0174600500613451,
+ "learning_rate": 0.0006,
+ "loss": 4.711891174316406,
+ "step": 3989
+ },
+ {
+ "epoch": 55.419213973799124,
+ "grad_norm": 0.018154339864850044,
+ "learning_rate": 0.0006,
+ "loss": 4.527122974395752,
+ "step": 3990
+ },
+ {
+ "epoch": 55.43318777292576,
+ "grad_norm": 0.01914648339152336,
+ "learning_rate": 0.0006,
+ "loss": 4.641975402832031,
+ "step": 3991
+ },
+ {
+ "epoch": 55.4471615720524,
+ "grad_norm": 0.017118625342845917,
+ "learning_rate": 0.0006,
+ "loss": 4.511425971984863,
+ "step": 3992
+ },
+ {
+ "epoch": 55.46113537117904,
+ "grad_norm": 0.01551489345729351,
+ "learning_rate": 0.0006,
+ "loss": 4.551141738891602,
+ "step": 3993
+ },
+ {
+ "epoch": 55.47510917030568,
+ "grad_norm": 0.015072065405547619,
+ "learning_rate": 0.0006,
+ "loss": 4.651115894317627,
+ "step": 3994
+ },
+ {
+ "epoch": 55.48908296943232,
+ "grad_norm": 0.017010986804962158,
+ "learning_rate": 0.0006,
+ "loss": 4.583282470703125,
+ "step": 3995
+ },
+ {
+ "epoch": 55.50305676855895,
+ "grad_norm": 0.01764831133186817,
+ "learning_rate": 0.0006,
+ "loss": 4.593364238739014,
+ "step": 3996
+ },
+ {
+ "epoch": 55.51703056768559,
+ "grad_norm": 0.01762884296476841,
+ "learning_rate": 0.0006,
+ "loss": 4.45671272277832,
+ "step": 3997
+ },
+ {
+ "epoch": 55.531004366812226,
+ "grad_norm": 0.01600457727909088,
+ "learning_rate": 0.0006,
+ "loss": 4.511260032653809,
+ "step": 3998
+ },
+ {
+ "epoch": 55.544978165938865,
+ "grad_norm": 0.015706021338701248,
+ "learning_rate": 0.0006,
+ "loss": 4.47593355178833,
+ "step": 3999
+ },
+ {
+ "epoch": 55.5589519650655,
+ "grad_norm": 0.017718996852636337,
+ "learning_rate": 0.0006,
+ "loss": 4.610543251037598,
+ "step": 4000
+ },
+ {
+ "epoch": 55.57292576419214,
+ "grad_norm": 0.015073757618665695,
+ "learning_rate": 0.0006,
+ "loss": 4.466620445251465,
+ "step": 4001
+ },
+ {
+ "epoch": 55.58689956331878,
+ "grad_norm": 0.01789247989654541,
+ "learning_rate": 0.0006,
+ "loss": 4.503409385681152,
+ "step": 4002
+ },
+ {
+ "epoch": 55.60087336244541,
+ "grad_norm": 0.022329598665237427,
+ "learning_rate": 0.0006,
+ "loss": 4.62099552154541,
+ "step": 4003
+ },
+ {
+ "epoch": 55.61484716157205,
+ "grad_norm": 0.02343829534947872,
+ "learning_rate": 0.0006,
+ "loss": 4.620813369750977,
+ "step": 4004
+ },
+ {
+ "epoch": 55.62882096069869,
+ "grad_norm": 0.023279687389731407,
+ "learning_rate": 0.0006,
+ "loss": 4.545765399932861,
+ "step": 4005
+ },
+ {
+ "epoch": 55.64279475982533,
+ "grad_norm": 0.020817464217543602,
+ "learning_rate": 0.0006,
+ "loss": 4.561470031738281,
+ "step": 4006
+ },
+ {
+ "epoch": 55.65676855895197,
+ "grad_norm": 0.020270058885216713,
+ "learning_rate": 0.0006,
+ "loss": 4.5875020027160645,
+ "step": 4007
+ },
+ {
+ "epoch": 55.670742358078606,
+ "grad_norm": 0.021085843443870544,
+ "learning_rate": 0.0006,
+ "loss": 4.504560470581055,
+ "step": 4008
+ },
+ {
+ "epoch": 55.68471615720524,
+ "grad_norm": 0.018507730215787888,
+ "learning_rate": 0.0006,
+ "loss": 4.520644664764404,
+ "step": 4009
+ },
+ {
+ "epoch": 55.698689956331876,
+ "grad_norm": 0.020571516826748848,
+ "learning_rate": 0.0006,
+ "loss": 4.539643287658691,
+ "step": 4010
+ },
+ {
+ "epoch": 55.712663755458514,
+ "grad_norm": 0.021484510973095894,
+ "learning_rate": 0.0006,
+ "loss": 4.628833293914795,
+ "step": 4011
+ },
+ {
+ "epoch": 55.72663755458515,
+ "grad_norm": 0.017693819478154182,
+ "learning_rate": 0.0006,
+ "loss": 4.6023783683776855,
+ "step": 4012
+ },
+ {
+ "epoch": 55.74061135371179,
+ "grad_norm": 0.020186059176921844,
+ "learning_rate": 0.0006,
+ "loss": 4.634884834289551,
+ "step": 4013
+ },
+ {
+ "epoch": 55.75458515283843,
+ "grad_norm": 0.01835780404508114,
+ "learning_rate": 0.0006,
+ "loss": 4.577791213989258,
+ "step": 4014
+ },
+ {
+ "epoch": 55.76855895196506,
+ "grad_norm": 0.018162988126277924,
+ "learning_rate": 0.0006,
+ "loss": 4.641597270965576,
+ "step": 4015
+ },
+ {
+ "epoch": 55.7825327510917,
+ "grad_norm": 0.018909504637122154,
+ "learning_rate": 0.0006,
+ "loss": 4.487447738647461,
+ "step": 4016
+ },
+ {
+ "epoch": 55.79650655021834,
+ "grad_norm": 0.018972251564264297,
+ "learning_rate": 0.0006,
+ "loss": 4.590317726135254,
+ "step": 4017
+ },
+ {
+ "epoch": 55.81048034934498,
+ "grad_norm": 0.01793607883155346,
+ "learning_rate": 0.0006,
+ "loss": 4.4994611740112305,
+ "step": 4018
+ },
+ {
+ "epoch": 55.82445414847162,
+ "grad_norm": 0.02123514749109745,
+ "learning_rate": 0.0006,
+ "loss": 4.479006767272949,
+ "step": 4019
+ },
+ {
+ "epoch": 55.838427947598255,
+ "grad_norm": 0.021246599033474922,
+ "learning_rate": 0.0006,
+ "loss": 4.4738640785217285,
+ "step": 4020
+ },
+ {
+ "epoch": 55.852401746724894,
+ "grad_norm": 0.020131602883338928,
+ "learning_rate": 0.0006,
+ "loss": 4.600308418273926,
+ "step": 4021
+ },
+ {
+ "epoch": 55.866375545851525,
+ "grad_norm": 0.019159460440278053,
+ "learning_rate": 0.0006,
+ "loss": 4.540908336639404,
+ "step": 4022
+ },
+ {
+ "epoch": 55.880349344978164,
+ "grad_norm": 0.023114705458283424,
+ "learning_rate": 0.0006,
+ "loss": 4.641247749328613,
+ "step": 4023
+ },
+ {
+ "epoch": 55.8943231441048,
+ "grad_norm": 0.02296135015785694,
+ "learning_rate": 0.0006,
+ "loss": 4.444624423980713,
+ "step": 4024
+ },
+ {
+ "epoch": 55.90829694323144,
+ "grad_norm": 0.02204655110836029,
+ "learning_rate": 0.0006,
+ "loss": 4.506349563598633,
+ "step": 4025
+ },
+ {
+ "epoch": 55.92227074235808,
+ "grad_norm": 0.019338225945830345,
+ "learning_rate": 0.0006,
+ "loss": 4.616047382354736,
+ "step": 4026
+ },
+ {
+ "epoch": 55.93624454148472,
+ "grad_norm": 0.020357884466648102,
+ "learning_rate": 0.0006,
+ "loss": 4.48170280456543,
+ "step": 4027
+ },
+ {
+ "epoch": 55.95021834061135,
+ "grad_norm": 0.02426772564649582,
+ "learning_rate": 0.0006,
+ "loss": 4.576877117156982,
+ "step": 4028
+ },
+ {
+ "epoch": 55.96419213973799,
+ "grad_norm": 0.023351816460490227,
+ "learning_rate": 0.0006,
+ "loss": 4.458885192871094,
+ "step": 4029
+ },
+ {
+ "epoch": 55.97816593886463,
+ "grad_norm": 0.019892243668437004,
+ "learning_rate": 0.0006,
+ "loss": 4.520229816436768,
+ "step": 4030
+ },
+ {
+ "epoch": 55.992139737991266,
+ "grad_norm": 0.018331089988350868,
+ "learning_rate": 0.0006,
+ "loss": 4.432201862335205,
+ "step": 4031
+ },
+ {
+ "epoch": 56.0,
+ "grad_norm": 0.021300828084349632,
+ "learning_rate": 0.0006,
+ "loss": 4.617072582244873,
+ "step": 4032
+ },
+ {
+ "epoch": 56.0,
+ "eval_loss": 4.811229705810547,
+ "eval_runtime": 56.4846,
+ "eval_samples_per_second": 43.233,
+ "eval_steps_per_second": 1.363,
+ "step": 4032
+ },
+ {
+ "epoch": 56.01397379912664,
+ "grad_norm": 0.02199401892721653,
+ "learning_rate": 0.0006,
+ "loss": 4.513965606689453,
+ "step": 4033
+ },
+ {
+ "epoch": 56.02794759825328,
+ "grad_norm": 0.022109858691692352,
+ "learning_rate": 0.0006,
+ "loss": 4.557568550109863,
+ "step": 4034
+ },
+ {
+ "epoch": 56.041921397379916,
+ "grad_norm": 0.020103694871068,
+ "learning_rate": 0.0006,
+ "loss": 4.569512844085693,
+ "step": 4035
+ },
+ {
+ "epoch": 56.05589519650655,
+ "grad_norm": 0.01898941583931446,
+ "learning_rate": 0.0006,
+ "loss": 4.623138904571533,
+ "step": 4036
+ },
+ {
+ "epoch": 56.069868995633186,
+ "grad_norm": 0.02227010577917099,
+ "learning_rate": 0.0006,
+ "loss": 4.433316230773926,
+ "step": 4037
+ },
+ {
+ "epoch": 56.083842794759825,
+ "grad_norm": 0.024683045223355293,
+ "learning_rate": 0.0006,
+ "loss": 4.458296775817871,
+ "step": 4038
+ },
+ {
+ "epoch": 56.09781659388646,
+ "grad_norm": 0.02900184690952301,
+ "learning_rate": 0.0006,
+ "loss": 4.531060218811035,
+ "step": 4039
+ },
+ {
+ "epoch": 56.1117903930131,
+ "grad_norm": 0.025557566434144974,
+ "learning_rate": 0.0006,
+ "loss": 4.56685733795166,
+ "step": 4040
+ },
+ {
+ "epoch": 56.12576419213974,
+ "grad_norm": 0.021414656192064285,
+ "learning_rate": 0.0006,
+ "loss": 4.469037055969238,
+ "step": 4041
+ },
+ {
+ "epoch": 56.13973799126637,
+ "grad_norm": 0.023262616246938705,
+ "learning_rate": 0.0006,
+ "loss": 4.493876934051514,
+ "step": 4042
+ },
+ {
+ "epoch": 56.15371179039301,
+ "grad_norm": 0.023011326789855957,
+ "learning_rate": 0.0006,
+ "loss": 4.4572906494140625,
+ "step": 4043
+ },
+ {
+ "epoch": 56.16768558951965,
+ "grad_norm": 0.02177615463733673,
+ "learning_rate": 0.0006,
+ "loss": 4.615389823913574,
+ "step": 4044
+ },
+ {
+ "epoch": 56.18165938864629,
+ "grad_norm": 0.021527713164687157,
+ "learning_rate": 0.0006,
+ "loss": 4.613762855529785,
+ "step": 4045
+ },
+ {
+ "epoch": 56.19563318777293,
+ "grad_norm": 0.01982325315475464,
+ "learning_rate": 0.0006,
+ "loss": 4.53670597076416,
+ "step": 4046
+ },
+ {
+ "epoch": 56.209606986899566,
+ "grad_norm": 0.022484654560685158,
+ "learning_rate": 0.0006,
+ "loss": 4.605565071105957,
+ "step": 4047
+ },
+ {
+ "epoch": 56.223580786026204,
+ "grad_norm": 0.02346140146255493,
+ "learning_rate": 0.0006,
+ "loss": 4.6265950202941895,
+ "step": 4048
+ },
+ {
+ "epoch": 56.237554585152836,
+ "grad_norm": 0.01876714453101158,
+ "learning_rate": 0.0006,
+ "loss": 4.481525421142578,
+ "step": 4049
+ },
+ {
+ "epoch": 56.251528384279474,
+ "grad_norm": 0.01926851086318493,
+ "learning_rate": 0.0006,
+ "loss": 4.630258083343506,
+ "step": 4050
+ },
+ {
+ "epoch": 56.26550218340611,
+ "grad_norm": 0.020858589559793472,
+ "learning_rate": 0.0006,
+ "loss": 4.5265913009643555,
+ "step": 4051
+ },
+ {
+ "epoch": 56.27947598253275,
+ "grad_norm": 0.01889280416071415,
+ "learning_rate": 0.0006,
+ "loss": 4.581275463104248,
+ "step": 4052
+ },
+ {
+ "epoch": 56.29344978165939,
+ "grad_norm": 0.01801002025604248,
+ "learning_rate": 0.0006,
+ "loss": 4.498557090759277,
+ "step": 4053
+ },
+ {
+ "epoch": 56.30742358078603,
+ "grad_norm": 0.02252180315554142,
+ "learning_rate": 0.0006,
+ "loss": 4.565267086029053,
+ "step": 4054
+ },
+ {
+ "epoch": 56.32139737991266,
+ "grad_norm": 0.02306070365011692,
+ "learning_rate": 0.0006,
+ "loss": 4.3591389656066895,
+ "step": 4055
+ },
+ {
+ "epoch": 56.3353711790393,
+ "grad_norm": 0.020292513072490692,
+ "learning_rate": 0.0006,
+ "loss": 4.407517433166504,
+ "step": 4056
+ },
+ {
+ "epoch": 56.34934497816594,
+ "grad_norm": 0.019154123961925507,
+ "learning_rate": 0.0006,
+ "loss": 4.482261657714844,
+ "step": 4057
+ },
+ {
+ "epoch": 56.36331877729258,
+ "grad_norm": 0.017338937148451805,
+ "learning_rate": 0.0006,
+ "loss": 4.4463582038879395,
+ "step": 4058
+ },
+ {
+ "epoch": 56.377292576419215,
+ "grad_norm": 0.01730559952557087,
+ "learning_rate": 0.0006,
+ "loss": 4.542474746704102,
+ "step": 4059
+ },
+ {
+ "epoch": 56.391266375545854,
+ "grad_norm": 0.017029426991939545,
+ "learning_rate": 0.0006,
+ "loss": 4.552488327026367,
+ "step": 4060
+ },
+ {
+ "epoch": 56.40524017467249,
+ "grad_norm": 0.01745498552918434,
+ "learning_rate": 0.0006,
+ "loss": 4.52928352355957,
+ "step": 4061
+ },
+ {
+ "epoch": 56.419213973799124,
+ "grad_norm": 0.0168515145778656,
+ "learning_rate": 0.0006,
+ "loss": 4.4516801834106445,
+ "step": 4062
+ },
+ {
+ "epoch": 56.43318777292576,
+ "grad_norm": 0.018545417115092278,
+ "learning_rate": 0.0006,
+ "loss": 4.569811820983887,
+ "step": 4063
+ },
+ {
+ "epoch": 56.4471615720524,
+ "grad_norm": 0.02227838523685932,
+ "learning_rate": 0.0006,
+ "loss": 4.412224769592285,
+ "step": 4064
+ },
+ {
+ "epoch": 56.46113537117904,
+ "grad_norm": 0.023338504135608673,
+ "learning_rate": 0.0006,
+ "loss": 4.490184307098389,
+ "step": 4065
+ },
+ {
+ "epoch": 56.47510917030568,
+ "grad_norm": 0.02257535606622696,
+ "learning_rate": 0.0006,
+ "loss": 4.479152679443359,
+ "step": 4066
+ },
+ {
+ "epoch": 56.48908296943232,
+ "grad_norm": 0.022737598046660423,
+ "learning_rate": 0.0006,
+ "loss": 4.418235778808594,
+ "step": 4067
+ },
+ {
+ "epoch": 56.50305676855895,
+ "grad_norm": 0.024603573605418205,
+ "learning_rate": 0.0006,
+ "loss": 4.45781135559082,
+ "step": 4068
+ },
+ {
+ "epoch": 56.51703056768559,
+ "grad_norm": 0.024953778833150864,
+ "learning_rate": 0.0006,
+ "loss": 4.593907833099365,
+ "step": 4069
+ },
+ {
+ "epoch": 56.531004366812226,
+ "grad_norm": 0.02195645309984684,
+ "learning_rate": 0.0006,
+ "loss": 4.467643737792969,
+ "step": 4070
+ },
+ {
+ "epoch": 56.544978165938865,
+ "grad_norm": 0.017126578837633133,
+ "learning_rate": 0.0006,
+ "loss": 4.555484771728516,
+ "step": 4071
+ },
+ {
+ "epoch": 56.5589519650655,
+ "grad_norm": 0.019482605159282684,
+ "learning_rate": 0.0006,
+ "loss": 4.430915832519531,
+ "step": 4072
+ },
+ {
+ "epoch": 56.57292576419214,
+ "grad_norm": 0.0219392292201519,
+ "learning_rate": 0.0006,
+ "loss": 4.593365669250488,
+ "step": 4073
+ },
+ {
+ "epoch": 56.58689956331878,
+ "grad_norm": 0.02420145645737648,
+ "learning_rate": 0.0006,
+ "loss": 4.491467475891113,
+ "step": 4074
+ },
+ {
+ "epoch": 56.60087336244541,
+ "grad_norm": 0.021523723378777504,
+ "learning_rate": 0.0006,
+ "loss": 4.463634490966797,
+ "step": 4075
+ },
+ {
+ "epoch": 56.61484716157205,
+ "grad_norm": 0.020242048427462578,
+ "learning_rate": 0.0006,
+ "loss": 4.577968597412109,
+ "step": 4076
+ },
+ {
+ "epoch": 56.62882096069869,
+ "grad_norm": 0.017771173268556595,
+ "learning_rate": 0.0006,
+ "loss": 4.568269729614258,
+ "step": 4077
+ },
+ {
+ "epoch": 56.64279475982533,
+ "grad_norm": 0.01598544418811798,
+ "learning_rate": 0.0006,
+ "loss": 4.476841926574707,
+ "step": 4078
+ },
+ {
+ "epoch": 56.65676855895197,
+ "grad_norm": 0.0168446097522974,
+ "learning_rate": 0.0006,
+ "loss": 4.480965614318848,
+ "step": 4079
+ },
+ {
+ "epoch": 56.670742358078606,
+ "grad_norm": 0.014948660507798195,
+ "learning_rate": 0.0006,
+ "loss": 4.535400390625,
+ "step": 4080
+ },
+ {
+ "epoch": 56.68471615720524,
+ "grad_norm": 0.014899153262376785,
+ "learning_rate": 0.0006,
+ "loss": 4.476048946380615,
+ "step": 4081
+ },
+ {
+ "epoch": 56.698689956331876,
+ "grad_norm": 0.014051870442926884,
+ "learning_rate": 0.0006,
+ "loss": 4.533884048461914,
+ "step": 4082
+ },
+ {
+ "epoch": 56.712663755458514,
+ "grad_norm": 0.014708572067320347,
+ "learning_rate": 0.0006,
+ "loss": 4.586690902709961,
+ "step": 4083
+ },
+ {
+ "epoch": 56.72663755458515,
+ "grad_norm": 0.014797299169003963,
+ "learning_rate": 0.0006,
+ "loss": 4.446291446685791,
+ "step": 4084
+ },
+ {
+ "epoch": 56.74061135371179,
+ "grad_norm": 0.015256541781127453,
+ "learning_rate": 0.0006,
+ "loss": 4.553572654724121,
+ "step": 4085
+ },
+ {
+ "epoch": 56.75458515283843,
+ "grad_norm": 0.016042491421103477,
+ "learning_rate": 0.0006,
+ "loss": 4.507801055908203,
+ "step": 4086
+ },
+ {
+ "epoch": 56.76855895196506,
+ "grad_norm": 0.018835173919796944,
+ "learning_rate": 0.0006,
+ "loss": 4.393911838531494,
+ "step": 4087
+ },
+ {
+ "epoch": 56.7825327510917,
+ "grad_norm": 0.021994033828377724,
+ "learning_rate": 0.0006,
+ "loss": 4.543368339538574,
+ "step": 4088
+ },
+ {
+ "epoch": 56.79650655021834,
+ "grad_norm": 0.0230990182608366,
+ "learning_rate": 0.0006,
+ "loss": 4.507420539855957,
+ "step": 4089
+ },
+ {
+ "epoch": 56.81048034934498,
+ "grad_norm": 0.018818842247128487,
+ "learning_rate": 0.0006,
+ "loss": 4.499941349029541,
+ "step": 4090
+ },
+ {
+ "epoch": 56.82445414847162,
+ "grad_norm": 0.018257278949022293,
+ "learning_rate": 0.0006,
+ "loss": 4.380057334899902,
+ "step": 4091
+ },
+ {
+ "epoch": 56.838427947598255,
+ "grad_norm": 0.020073018968105316,
+ "learning_rate": 0.0006,
+ "loss": 4.572811603546143,
+ "step": 4092
+ },
+ {
+ "epoch": 56.852401746724894,
+ "grad_norm": 0.02078658528625965,
+ "learning_rate": 0.0006,
+ "loss": 4.446835517883301,
+ "step": 4093
+ },
+ {
+ "epoch": 56.866375545851525,
+ "grad_norm": 0.01793118566274643,
+ "learning_rate": 0.0006,
+ "loss": 4.517854690551758,
+ "step": 4094
+ },
+ {
+ "epoch": 56.880349344978164,
+ "grad_norm": 0.01608281210064888,
+ "learning_rate": 0.0006,
+ "loss": 4.506482124328613,
+ "step": 4095
+ },
+ {
+ "epoch": 56.8943231441048,
+ "grad_norm": 0.017405368387699127,
+ "learning_rate": 0.0006,
+ "loss": 4.439718246459961,
+ "step": 4096
+ },
+ {
+ "epoch": 56.90829694323144,
+ "grad_norm": 0.017930863425135612,
+ "learning_rate": 0.0006,
+ "loss": 4.409365653991699,
+ "step": 4097
+ },
+ {
+ "epoch": 56.92227074235808,
+ "grad_norm": 0.015757670626044273,
+ "learning_rate": 0.0006,
+ "loss": 4.505547523498535,
+ "step": 4098
+ },
+ {
+ "epoch": 56.93624454148472,
+ "grad_norm": 0.01580345816910267,
+ "learning_rate": 0.0006,
+ "loss": 4.462774753570557,
+ "step": 4099
+ },
+ {
+ "epoch": 56.95021834061135,
+ "grad_norm": 0.01659180410206318,
+ "learning_rate": 0.0006,
+ "loss": 4.49345588684082,
+ "step": 4100
+ },
+ {
+ "epoch": 56.96419213973799,
+ "grad_norm": 0.01877259835600853,
+ "learning_rate": 0.0006,
+ "loss": 4.574443340301514,
+ "step": 4101
+ },
+ {
+ "epoch": 56.97816593886463,
+ "grad_norm": 0.01977839693427086,
+ "learning_rate": 0.0006,
+ "loss": 4.574034690856934,
+ "step": 4102
+ },
+ {
+ "epoch": 56.992139737991266,
+ "grad_norm": 0.020627597346901894,
+ "learning_rate": 0.0006,
+ "loss": 4.50969123840332,
+ "step": 4103
+ },
+ {
+ "epoch": 57.0,
+ "grad_norm": 0.025128807872533798,
+ "learning_rate": 0.0006,
+ "loss": 4.401662826538086,
+ "step": 4104
+ },
+ {
+ "epoch": 57.0,
+ "eval_loss": 4.779252052307129,
+ "eval_runtime": 56.8104,
+ "eval_samples_per_second": 42.985,
+ "eval_steps_per_second": 1.355,
+ "step": 4104
+ },
+ {
+ "epoch": 57.01397379912664,
+ "grad_norm": 0.027712570503354073,
+ "learning_rate": 0.0006,
+ "loss": 4.472955703735352,
+ "step": 4105
+ },
+ {
+ "epoch": 57.02794759825328,
+ "grad_norm": 0.02164197526872158,
+ "learning_rate": 0.0006,
+ "loss": 4.387596130371094,
+ "step": 4106
+ },
+ {
+ "epoch": 57.041921397379916,
+ "grad_norm": 0.017972752451896667,
+ "learning_rate": 0.0006,
+ "loss": 4.4326276779174805,
+ "step": 4107
+ },
+ {
+ "epoch": 57.05589519650655,
+ "grad_norm": 0.020811019465327263,
+ "learning_rate": 0.0006,
+ "loss": 4.452951431274414,
+ "step": 4108
+ },
+ {
+ "epoch": 57.069868995633186,
+ "grad_norm": 0.019913259893655777,
+ "learning_rate": 0.0006,
+ "loss": 4.448301315307617,
+ "step": 4109
+ },
+ {
+ "epoch": 57.083842794759825,
+ "grad_norm": 0.017409102991223335,
+ "learning_rate": 0.0006,
+ "loss": 4.4108357429504395,
+ "step": 4110
+ },
+ {
+ "epoch": 57.09781659388646,
+ "grad_norm": 0.019374269992113113,
+ "learning_rate": 0.0006,
+ "loss": 4.398412227630615,
+ "step": 4111
+ },
+ {
+ "epoch": 57.1117903930131,
+ "grad_norm": 0.015521776862442493,
+ "learning_rate": 0.0006,
+ "loss": 4.508594989776611,
+ "step": 4112
+ },
+ {
+ "epoch": 57.12576419213974,
+ "grad_norm": 0.017201920971274376,
+ "learning_rate": 0.0006,
+ "loss": 4.449558734893799,
+ "step": 4113
+ },
+ {
+ "epoch": 57.13973799126637,
+ "grad_norm": 0.018062541261315346,
+ "learning_rate": 0.0006,
+ "loss": 4.517874240875244,
+ "step": 4114
+ },
+ {
+ "epoch": 57.15371179039301,
+ "grad_norm": 0.015846073627471924,
+ "learning_rate": 0.0006,
+ "loss": 4.427501678466797,
+ "step": 4115
+ },
+ {
+ "epoch": 57.16768558951965,
+ "grad_norm": 0.01849362626671791,
+ "learning_rate": 0.0006,
+ "loss": 4.433181285858154,
+ "step": 4116
+ },
+ {
+ "epoch": 57.18165938864629,
+ "grad_norm": 0.01948804222047329,
+ "learning_rate": 0.0006,
+ "loss": 4.446122646331787,
+ "step": 4117
+ },
+ {
+ "epoch": 57.19563318777293,
+ "grad_norm": 0.017354389652609825,
+ "learning_rate": 0.0006,
+ "loss": 4.431153774261475,
+ "step": 4118
+ },
+ {
+ "epoch": 57.209606986899566,
+ "grad_norm": 0.015817373991012573,
+ "learning_rate": 0.0006,
+ "loss": 4.448566913604736,
+ "step": 4119
+ },
+ {
+ "epoch": 57.223580786026204,
+ "grad_norm": 0.01635785959661007,
+ "learning_rate": 0.0006,
+ "loss": 4.408031463623047,
+ "step": 4120
+ },
+ {
+ "epoch": 57.237554585152836,
+ "grad_norm": 0.014528338797390461,
+ "learning_rate": 0.0006,
+ "loss": 4.541668891906738,
+ "step": 4121
+ },
+ {
+ "epoch": 57.251528384279474,
+ "grad_norm": 0.01589174196124077,
+ "learning_rate": 0.0006,
+ "loss": 4.513199329376221,
+ "step": 4122
+ },
+ {
+ "epoch": 57.26550218340611,
+ "grad_norm": 0.016093695536255836,
+ "learning_rate": 0.0006,
+ "loss": 4.516178131103516,
+ "step": 4123
+ },
+ {
+ "epoch": 57.27947598253275,
+ "grad_norm": 0.01788361556828022,
+ "learning_rate": 0.0006,
+ "loss": 4.524880409240723,
+ "step": 4124
+ },
+ {
+ "epoch": 57.29344978165939,
+ "grad_norm": 0.01591576263308525,
+ "learning_rate": 0.0006,
+ "loss": 4.44764518737793,
+ "step": 4125
+ },
+ {
+ "epoch": 57.30742358078603,
+ "grad_norm": 0.015452570281922817,
+ "learning_rate": 0.0006,
+ "loss": 4.46844482421875,
+ "step": 4126
+ },
+ {
+ "epoch": 57.32139737991266,
+ "grad_norm": 0.016404492780566216,
+ "learning_rate": 0.0006,
+ "loss": 4.443678855895996,
+ "step": 4127
+ },
+ {
+ "epoch": 57.3353711790393,
+ "grad_norm": 0.019967148080468178,
+ "learning_rate": 0.0006,
+ "loss": 4.430230140686035,
+ "step": 4128
+ },
+ {
+ "epoch": 57.34934497816594,
+ "grad_norm": 0.02594015561044216,
+ "learning_rate": 0.0006,
+ "loss": 4.47061824798584,
+ "step": 4129
+ },
+ {
+ "epoch": 57.36331877729258,
+ "grad_norm": 0.031096890568733215,
+ "learning_rate": 0.0006,
+ "loss": 4.462275981903076,
+ "step": 4130
+ },
+ {
+ "epoch": 57.377292576419215,
+ "grad_norm": 0.02834898792207241,
+ "learning_rate": 0.0006,
+ "loss": 4.41143798828125,
+ "step": 4131
+ },
+ {
+ "epoch": 57.391266375545854,
+ "grad_norm": 0.023621153086423874,
+ "learning_rate": 0.0006,
+ "loss": 4.562548637390137,
+ "step": 4132
+ },
+ {
+ "epoch": 57.40524017467249,
+ "grad_norm": 0.02277296595275402,
+ "learning_rate": 0.0006,
+ "loss": 4.456247329711914,
+ "step": 4133
+ },
+ {
+ "epoch": 57.419213973799124,
+ "grad_norm": 0.019868768751621246,
+ "learning_rate": 0.0006,
+ "loss": 4.524300575256348,
+ "step": 4134
+ },
+ {
+ "epoch": 57.43318777292576,
+ "grad_norm": 0.01887078955769539,
+ "learning_rate": 0.0006,
+ "loss": 4.540097236633301,
+ "step": 4135
+ },
+ {
+ "epoch": 57.4471615720524,
+ "grad_norm": 0.024268802255392075,
+ "learning_rate": 0.0006,
+ "loss": 4.397233963012695,
+ "step": 4136
+ },
+ {
+ "epoch": 57.46113537117904,
+ "grad_norm": 0.025739721953868866,
+ "learning_rate": 0.0006,
+ "loss": 4.354194641113281,
+ "step": 4137
+ },
+ {
+ "epoch": 57.47510917030568,
+ "grad_norm": 0.02442290261387825,
+ "learning_rate": 0.0006,
+ "loss": 4.303822994232178,
+ "step": 4138
+ },
+ {
+ "epoch": 57.48908296943232,
+ "grad_norm": 0.022215209901332855,
+ "learning_rate": 0.0006,
+ "loss": 4.4870452880859375,
+ "step": 4139
+ },
+ {
+ "epoch": 57.50305676855895,
+ "grad_norm": 0.01789158768951893,
+ "learning_rate": 0.0006,
+ "loss": 4.343809127807617,
+ "step": 4140
+ },
+ {
+ "epoch": 57.51703056768559,
+ "grad_norm": 0.020381739363074303,
+ "learning_rate": 0.0006,
+ "loss": 4.53312873840332,
+ "step": 4141
+ },
+ {
+ "epoch": 57.531004366812226,
+ "grad_norm": 0.019585080444812775,
+ "learning_rate": 0.0006,
+ "loss": 4.510162353515625,
+ "step": 4142
+ },
+ {
+ "epoch": 57.544978165938865,
+ "grad_norm": 0.01985127478837967,
+ "learning_rate": 0.0006,
+ "loss": 4.3654022216796875,
+ "step": 4143
+ },
+ {
+ "epoch": 57.5589519650655,
+ "grad_norm": 0.01962905190885067,
+ "learning_rate": 0.0006,
+ "loss": 4.560830116271973,
+ "step": 4144
+ },
+ {
+ "epoch": 57.57292576419214,
+ "grad_norm": 0.02679629437625408,
+ "learning_rate": 0.0006,
+ "loss": 4.300804138183594,
+ "step": 4145
+ },
+ {
+ "epoch": 57.58689956331878,
+ "grad_norm": 0.039830856025218964,
+ "learning_rate": 0.0006,
+ "loss": 4.54150915145874,
+ "step": 4146
+ },
+ {
+ "epoch": 57.60087336244541,
+ "grad_norm": 0.049527477473020554,
+ "learning_rate": 0.0006,
+ "loss": 4.405792236328125,
+ "step": 4147
+ },
+ {
+ "epoch": 57.61484716157205,
+ "grad_norm": 0.042966946959495544,
+ "learning_rate": 0.0006,
+ "loss": 4.500643730163574,
+ "step": 4148
+ },
+ {
+ "epoch": 57.62882096069869,
+ "grad_norm": 0.08999177813529968,
+ "learning_rate": 0.0006,
+ "loss": 4.504790782928467,
+ "step": 4149
+ },
+ {
+ "epoch": 57.64279475982533,
+ "grad_norm": 0.44693124294281006,
+ "learning_rate": 0.0006,
+ "loss": 4.845344066619873,
+ "step": 4150
+ },
+ {
+ "epoch": 57.65676855895197,
+ "grad_norm": 0.7143144011497498,
+ "learning_rate": 0.0006,
+ "loss": 6.907845497131348,
+ "step": 4151
+ },
+ {
+ "epoch": 57.670742358078606,
+ "grad_norm": 0.25239595770835876,
+ "learning_rate": 0.0006,
+ "loss": 7.082579612731934,
+ "step": 4152
+ },
+ {
+ "epoch": 57.68471615720524,
+ "grad_norm": 0.26550790667533875,
+ "learning_rate": 0.0006,
+ "loss": 7.50180196762085,
+ "step": 4153
+ },
+ {
+ "epoch": 57.698689956331876,
+ "grad_norm": 0.14735378324985504,
+ "learning_rate": 0.0006,
+ "loss": 7.214590072631836,
+ "step": 4154
+ },
+ {
+ "epoch": 57.712663755458514,
+ "grad_norm": 0.17685334384441376,
+ "learning_rate": 0.0006,
+ "loss": 6.861320495605469,
+ "step": 4155
+ },
+ {
+ "epoch": 57.72663755458515,
+ "grad_norm": 0.18231339752674103,
+ "learning_rate": 0.0006,
+ "loss": 6.791839599609375,
+ "step": 4156
+ },
+ {
+ "epoch": 57.74061135371179,
+ "grad_norm": 0.09230359643697739,
+ "learning_rate": 0.0006,
+ "loss": 6.760223388671875,
+ "step": 4157
+ },
+ {
+ "epoch": 57.75458515283843,
+ "grad_norm": 0.06918249279260635,
+ "learning_rate": 0.0006,
+ "loss": 6.532999515533447,
+ "step": 4158
+ },
+ {
+ "epoch": 57.76855895196506,
+ "grad_norm": 0.0901007279753685,
+ "learning_rate": 0.0006,
+ "loss": 6.52907133102417,
+ "step": 4159
+ },
+ {
+ "epoch": 57.7825327510917,
+ "grad_norm": 0.05678229779005051,
+ "learning_rate": 0.0006,
+ "loss": 6.295290470123291,
+ "step": 4160
+ },
+ {
+ "epoch": 57.79650655021834,
+ "grad_norm": 0.04450210556387901,
+ "learning_rate": 0.0006,
+ "loss": 6.14900541305542,
+ "step": 4161
+ },
+ {
+ "epoch": 57.81048034934498,
+ "grad_norm": 0.049191512167453766,
+ "learning_rate": 0.0006,
+ "loss": 6.19650411605835,
+ "step": 4162
+ },
+ {
+ "epoch": 57.82445414847162,
+ "grad_norm": 0.03908967226743698,
+ "learning_rate": 0.0006,
+ "loss": 6.054948806762695,
+ "step": 4163
+ },
+ {
+ "epoch": 57.838427947598255,
+ "grad_norm": 0.04023387283086777,
+ "learning_rate": 0.0006,
+ "loss": 6.034873962402344,
+ "step": 4164
+ },
+ {
+ "epoch": 57.852401746724894,
+ "grad_norm": 0.03304268419742584,
+ "learning_rate": 0.0006,
+ "loss": 5.769852161407471,
+ "step": 4165
+ },
+ {
+ "epoch": 57.866375545851525,
+ "grad_norm": 0.03432450816035271,
+ "learning_rate": 0.0006,
+ "loss": 5.788785457611084,
+ "step": 4166
+ },
+ {
+ "epoch": 57.880349344978164,
+ "grad_norm": 0.029965840280056,
+ "learning_rate": 0.0006,
+ "loss": 5.815243721008301,
+ "step": 4167
+ },
+ {
+ "epoch": 57.8943231441048,
+ "grad_norm": 0.03637698292732239,
+ "learning_rate": 0.0006,
+ "loss": 5.7204060554504395,
+ "step": 4168
+ },
+ {
+ "epoch": 57.90829694323144,
+ "grad_norm": 0.03477517142891884,
+ "learning_rate": 0.0006,
+ "loss": 5.605930328369141,
+ "step": 4169
+ },
+ {
+ "epoch": 57.92227074235808,
+ "grad_norm": 0.05061859264969826,
+ "learning_rate": 0.0006,
+ "loss": 5.547054767608643,
+ "step": 4170
+ },
+ {
+ "epoch": 57.93624454148472,
+ "grad_norm": 0.06299655884504318,
+ "learning_rate": 0.0006,
+ "loss": 5.552166938781738,
+ "step": 4171
+ },
+ {
+ "epoch": 57.95021834061135,
+ "grad_norm": 0.04215948283672333,
+ "learning_rate": 0.0006,
+ "loss": 5.375457763671875,
+ "step": 4172
+ },
+ {
+ "epoch": 57.96419213973799,
+ "grad_norm": 0.0348566472530365,
+ "learning_rate": 0.0006,
+ "loss": 5.418606281280518,
+ "step": 4173
+ },
+ {
+ "epoch": 57.97816593886463,
+ "grad_norm": 0.03807530924677849,
+ "learning_rate": 0.0006,
+ "loss": 5.403156280517578,
+ "step": 4174
+ },
+ {
+ "epoch": 57.992139737991266,
+ "grad_norm": 0.02989993803203106,
+ "learning_rate": 0.0006,
+ "loss": 5.355951309204102,
+ "step": 4175
+ },
+ {
+ "epoch": 58.0,
+ "grad_norm": 0.034300558269023895,
+ "learning_rate": 0.0006,
+ "loss": 5.44920539855957,
+ "step": 4176
+ },
+ {
+ "epoch": 58.0,
+ "eval_loss": 5.416731834411621,
+ "eval_runtime": 56.7934,
+ "eval_samples_per_second": 42.998,
+ "eval_steps_per_second": 1.356,
+ "step": 4176
+ },
+ {
+ "epoch": 58.01397379912664,
+ "grad_norm": 0.033961448818445206,
+ "learning_rate": 0.0006,
+ "loss": 5.250707626342773,
+ "step": 4177
+ },
+ {
+ "epoch": 58.02794759825328,
+ "grad_norm": 0.027343502268195152,
+ "learning_rate": 0.0006,
+ "loss": 5.149238109588623,
+ "step": 4178
+ },
+ {
+ "epoch": 58.041921397379916,
+ "grad_norm": 0.03033650480210781,
+ "learning_rate": 0.0006,
+ "loss": 5.072463512420654,
+ "step": 4179
+ },
+ {
+ "epoch": 58.05589519650655,
+ "grad_norm": 0.028262868523597717,
+ "learning_rate": 0.0006,
+ "loss": 5.109417915344238,
+ "step": 4180
+ },
+ {
+ "epoch": 58.069868995633186,
+ "grad_norm": 0.02373034693300724,
+ "learning_rate": 0.0006,
+ "loss": 5.082464218139648,
+ "step": 4181
+ },
+ {
+ "epoch": 58.083842794759825,
+ "grad_norm": 0.025879688560962677,
+ "learning_rate": 0.0006,
+ "loss": 5.056792259216309,
+ "step": 4182
+ },
+ {
+ "epoch": 58.09781659388646,
+ "grad_norm": 0.026252275332808495,
+ "learning_rate": 0.0006,
+ "loss": 5.064239978790283,
+ "step": 4183
+ },
+ {
+ "epoch": 58.1117903930131,
+ "grad_norm": 0.028177792206406593,
+ "learning_rate": 0.0006,
+ "loss": 4.922310829162598,
+ "step": 4184
+ },
+ {
+ "epoch": 58.12576419213974,
+ "grad_norm": 0.02630820870399475,
+ "learning_rate": 0.0006,
+ "loss": 4.992884635925293,
+ "step": 4185
+ },
+ {
+ "epoch": 58.13973799126637,
+ "grad_norm": 0.024890903383493423,
+ "learning_rate": 0.0006,
+ "loss": 4.910036563873291,
+ "step": 4186
+ },
+ {
+ "epoch": 58.15371179039301,
+ "grad_norm": 0.02822992391884327,
+ "learning_rate": 0.0006,
+ "loss": 4.866640090942383,
+ "step": 4187
+ },
+ {
+ "epoch": 58.16768558951965,
+ "grad_norm": 0.03253607079386711,
+ "learning_rate": 0.0006,
+ "loss": 4.9112958908081055,
+ "step": 4188
+ },
+ {
+ "epoch": 58.18165938864629,
+ "grad_norm": 0.03765944764018059,
+ "learning_rate": 0.0006,
+ "loss": 4.774468421936035,
+ "step": 4189
+ },
+ {
+ "epoch": 58.19563318777293,
+ "grad_norm": 0.039874881505966187,
+ "learning_rate": 0.0006,
+ "loss": 4.831678867340088,
+ "step": 4190
+ },
+ {
+ "epoch": 58.209606986899566,
+ "grad_norm": 0.034595925360918045,
+ "learning_rate": 0.0006,
+ "loss": 4.8243794441223145,
+ "step": 4191
+ },
+ {
+ "epoch": 58.223580786026204,
+ "grad_norm": 0.04840189591050148,
+ "learning_rate": 0.0006,
+ "loss": 4.854156494140625,
+ "step": 4192
+ },
+ {
+ "epoch": 58.237554585152836,
+ "grad_norm": 0.06358753889799118,
+ "learning_rate": 0.0006,
+ "loss": 4.836430549621582,
+ "step": 4193
+ },
+ {
+ "epoch": 58.251528384279474,
+ "grad_norm": 0.0507555715739727,
+ "learning_rate": 0.0006,
+ "loss": 4.88357400894165,
+ "step": 4194
+ },
+ {
+ "epoch": 58.26550218340611,
+ "grad_norm": 0.04384690150618553,
+ "learning_rate": 0.0006,
+ "loss": 4.774049282073975,
+ "step": 4195
+ },
+ {
+ "epoch": 58.27947598253275,
+ "grad_norm": 0.03246928006410599,
+ "learning_rate": 0.0006,
+ "loss": 4.686605930328369,
+ "step": 4196
+ },
+ {
+ "epoch": 58.29344978165939,
+ "grad_norm": 0.028878789395093918,
+ "learning_rate": 0.0006,
+ "loss": 4.796070575714111,
+ "step": 4197
+ },
+ {
+ "epoch": 58.30742358078603,
+ "grad_norm": 0.028175361454486847,
+ "learning_rate": 0.0006,
+ "loss": 4.728490829467773,
+ "step": 4198
+ },
+ {
+ "epoch": 58.32139737991266,
+ "grad_norm": 0.029428014531731606,
+ "learning_rate": 0.0006,
+ "loss": 4.7159624099731445,
+ "step": 4199
+ },
+ {
+ "epoch": 58.3353711790393,
+ "grad_norm": 0.031227827072143555,
+ "learning_rate": 0.0006,
+ "loss": 4.702625274658203,
+ "step": 4200
+ },
+ {
+ "epoch": 58.34934497816594,
+ "grad_norm": 0.028843428939580917,
+ "learning_rate": 0.0006,
+ "loss": 4.811799049377441,
+ "step": 4201
+ },
+ {
+ "epoch": 58.36331877729258,
+ "grad_norm": 0.026780391111969948,
+ "learning_rate": 0.0006,
+ "loss": 4.81662654876709,
+ "step": 4202
+ },
+ {
+ "epoch": 58.377292576419215,
+ "grad_norm": 0.019699757918715477,
+ "learning_rate": 0.0006,
+ "loss": 4.713685989379883,
+ "step": 4203
+ },
+ {
+ "epoch": 58.391266375545854,
+ "grad_norm": 0.024389250203967094,
+ "learning_rate": 0.0006,
+ "loss": 4.72354793548584,
+ "step": 4204
+ },
+ {
+ "epoch": 58.40524017467249,
+ "grad_norm": 0.02218274027109146,
+ "learning_rate": 0.0006,
+ "loss": 4.674125671386719,
+ "step": 4205
+ },
+ {
+ "epoch": 58.419213973799124,
+ "grad_norm": 0.02041488140821457,
+ "learning_rate": 0.0006,
+ "loss": 4.7909698486328125,
+ "step": 4206
+ },
+ {
+ "epoch": 58.43318777292576,
+ "grad_norm": 0.018561935052275658,
+ "learning_rate": 0.0006,
+ "loss": 4.808770179748535,
+ "step": 4207
+ },
+ {
+ "epoch": 58.4471615720524,
+ "grad_norm": 0.01708846725523472,
+ "learning_rate": 0.0006,
+ "loss": 4.701626777648926,
+ "step": 4208
+ },
+ {
+ "epoch": 58.46113537117904,
+ "grad_norm": 0.0222734697163105,
+ "learning_rate": 0.0006,
+ "loss": 4.505410194396973,
+ "step": 4209
+ },
+ {
+ "epoch": 58.47510917030568,
+ "grad_norm": 0.03197057917714119,
+ "learning_rate": 0.0006,
+ "loss": 4.610360145568848,
+ "step": 4210
+ },
+ {
+ "epoch": 58.48908296943232,
+ "grad_norm": 0.06854520738124847,
+ "learning_rate": 0.0006,
+ "loss": 4.625649452209473,
+ "step": 4211
+ },
+ {
+ "epoch": 58.50305676855895,
+ "grad_norm": 0.09981521219015121,
+ "learning_rate": 0.0006,
+ "loss": 4.832150459289551,
+ "step": 4212
+ },
+ {
+ "epoch": 58.51703056768559,
+ "grad_norm": 0.046356040984392166,
+ "learning_rate": 0.0006,
+ "loss": 4.5641632080078125,
+ "step": 4213
+ },
+ {
+ "epoch": 58.531004366812226,
+ "grad_norm": 0.03882667422294617,
+ "learning_rate": 0.0006,
+ "loss": 4.666974067687988,
+ "step": 4214
+ },
+ {
+ "epoch": 58.544978165938865,
+ "grad_norm": 0.033203233033418655,
+ "learning_rate": 0.0006,
+ "loss": 4.629413604736328,
+ "step": 4215
+ },
+ {
+ "epoch": 58.5589519650655,
+ "grad_norm": 0.02806735597550869,
+ "learning_rate": 0.0006,
+ "loss": 4.763245582580566,
+ "step": 4216
+ },
+ {
+ "epoch": 58.57292576419214,
+ "grad_norm": 0.026728278025984764,
+ "learning_rate": 0.0006,
+ "loss": 4.629597187042236,
+ "step": 4217
+ },
+ {
+ "epoch": 58.58689956331878,
+ "grad_norm": 0.024380959570407867,
+ "learning_rate": 0.0006,
+ "loss": 4.633614540100098,
+ "step": 4218
+ },
+ {
+ "epoch": 58.60087336244541,
+ "grad_norm": 0.022312544286251068,
+ "learning_rate": 0.0006,
+ "loss": 4.671379566192627,
+ "step": 4219
+ },
+ {
+ "epoch": 58.61484716157205,
+ "grad_norm": 0.022049568593502045,
+ "learning_rate": 0.0006,
+ "loss": 4.661557197570801,
+ "step": 4220
+ },
+ {
+ "epoch": 58.62882096069869,
+ "grad_norm": 0.02119818702340126,
+ "learning_rate": 0.0006,
+ "loss": 4.688562393188477,
+ "step": 4221
+ },
+ {
+ "epoch": 58.64279475982533,
+ "grad_norm": 0.019769301638007164,
+ "learning_rate": 0.0006,
+ "loss": 4.607748031616211,
+ "step": 4222
+ },
+ {
+ "epoch": 58.65676855895197,
+ "grad_norm": 0.02124079130589962,
+ "learning_rate": 0.0006,
+ "loss": 4.570640563964844,
+ "step": 4223
+ },
+ {
+ "epoch": 58.670742358078606,
+ "grad_norm": 0.016344040632247925,
+ "learning_rate": 0.0006,
+ "loss": 4.676113128662109,
+ "step": 4224
+ },
+ {
+ "epoch": 58.68471615720524,
+ "grad_norm": 0.01735256239771843,
+ "learning_rate": 0.0006,
+ "loss": 4.62088680267334,
+ "step": 4225
+ },
+ {
+ "epoch": 58.698689956331876,
+ "grad_norm": 0.019323458895087242,
+ "learning_rate": 0.0006,
+ "loss": 4.579405307769775,
+ "step": 4226
+ },
+ {
+ "epoch": 58.712663755458514,
+ "grad_norm": 0.01760544814169407,
+ "learning_rate": 0.0006,
+ "loss": 4.625407695770264,
+ "step": 4227
+ },
+ {
+ "epoch": 58.72663755458515,
+ "grad_norm": 0.015593250282108784,
+ "learning_rate": 0.0006,
+ "loss": 4.6326117515563965,
+ "step": 4228
+ },
+ {
+ "epoch": 58.74061135371179,
+ "grad_norm": 0.014357523061335087,
+ "learning_rate": 0.0006,
+ "loss": 4.6869401931762695,
+ "step": 4229
+ },
+ {
+ "epoch": 58.75458515283843,
+ "grad_norm": 0.01604357920587063,
+ "learning_rate": 0.0006,
+ "loss": 4.690784931182861,
+ "step": 4230
+ },
+ {
+ "epoch": 58.76855895196506,
+ "grad_norm": 0.01673845760524273,
+ "learning_rate": 0.0006,
+ "loss": 4.513410568237305,
+ "step": 4231
+ },
+ {
+ "epoch": 58.7825327510917,
+ "grad_norm": 0.013657779432833195,
+ "learning_rate": 0.0006,
+ "loss": 4.637057781219482,
+ "step": 4232
+ },
+ {
+ "epoch": 58.79650655021834,
+ "grad_norm": 0.013373114168643951,
+ "learning_rate": 0.0006,
+ "loss": 4.719967842102051,
+ "step": 4233
+ },
+ {
+ "epoch": 58.81048034934498,
+ "grad_norm": 0.016345568001270294,
+ "learning_rate": 0.0006,
+ "loss": 4.604109764099121,
+ "step": 4234
+ },
+ {
+ "epoch": 58.82445414847162,
+ "grad_norm": 0.01573033444583416,
+ "learning_rate": 0.0006,
+ "loss": 4.6471428871154785,
+ "step": 4235
+ },
+ {
+ "epoch": 58.838427947598255,
+ "grad_norm": 0.01517449039965868,
+ "learning_rate": 0.0006,
+ "loss": 4.722684860229492,
+ "step": 4236
+ },
+ {
+ "epoch": 58.852401746724894,
+ "grad_norm": 0.0145102022215724,
+ "learning_rate": 0.0006,
+ "loss": 4.662420272827148,
+ "step": 4237
+ },
+ {
+ "epoch": 58.866375545851525,
+ "grad_norm": 0.01457217987626791,
+ "learning_rate": 0.0006,
+ "loss": 4.585169792175293,
+ "step": 4238
+ },
+ {
+ "epoch": 58.880349344978164,
+ "grad_norm": 0.01406773366034031,
+ "learning_rate": 0.0006,
+ "loss": 4.5413665771484375,
+ "step": 4239
+ },
+ {
+ "epoch": 58.8943231441048,
+ "grad_norm": 0.012653871439397335,
+ "learning_rate": 0.0006,
+ "loss": 4.553636074066162,
+ "step": 4240
+ },
+ {
+ "epoch": 58.90829694323144,
+ "grad_norm": 0.01261200476437807,
+ "learning_rate": 0.0006,
+ "loss": 4.534193992614746,
+ "step": 4241
+ },
+ {
+ "epoch": 58.92227074235808,
+ "grad_norm": 0.012955864891409874,
+ "learning_rate": 0.0006,
+ "loss": 4.697248458862305,
+ "step": 4242
+ },
+ {
+ "epoch": 58.93624454148472,
+ "grad_norm": 0.012134749442338943,
+ "learning_rate": 0.0006,
+ "loss": 4.557095527648926,
+ "step": 4243
+ },
+ {
+ "epoch": 58.95021834061135,
+ "grad_norm": 0.012164677493274212,
+ "learning_rate": 0.0006,
+ "loss": 4.623579025268555,
+ "step": 4244
+ },
+ {
+ "epoch": 58.96419213973799,
+ "grad_norm": 0.013802947476506233,
+ "learning_rate": 0.0006,
+ "loss": 4.562838077545166,
+ "step": 4245
+ },
+ {
+ "epoch": 58.97816593886463,
+ "grad_norm": 0.011420476250350475,
+ "learning_rate": 0.0006,
+ "loss": 4.62265682220459,
+ "step": 4246
+ },
+ {
+ "epoch": 58.992139737991266,
+ "grad_norm": 0.01169576682150364,
+ "learning_rate": 0.0006,
+ "loss": 4.727449417114258,
+ "step": 4247
+ },
+ {
+ "epoch": 59.0,
+ "grad_norm": 0.013577710837125778,
+ "learning_rate": 0.0006,
+ "loss": 4.536325454711914,
+ "step": 4248
+ },
+ {
+ "epoch": 59.0,
+ "eval_loss": 4.86527681350708,
+ "eval_runtime": 60.4463,
+ "eval_samples_per_second": 40.399,
+ "eval_steps_per_second": 1.274,
+ "step": 4248
+ },
+ {
+ "epoch": 59.01397379912664,
+ "grad_norm": 0.015251655131578445,
+ "learning_rate": 0.0006,
+ "loss": 4.634598731994629,
+ "step": 4249
+ },
+ {
+ "epoch": 59.02794759825328,
+ "grad_norm": 0.017414938658475876,
+ "learning_rate": 0.0006,
+ "loss": 4.521329879760742,
+ "step": 4250
+ },
+ {
+ "epoch": 59.041921397379916,
+ "grad_norm": 0.01656588353216648,
+ "learning_rate": 0.0006,
+ "loss": 4.446192264556885,
+ "step": 4251
+ },
+ {
+ "epoch": 59.05589519650655,
+ "grad_norm": 0.012038925662636757,
+ "learning_rate": 0.0006,
+ "loss": 4.523855209350586,
+ "step": 4252
+ },
+ {
+ "epoch": 59.069868995633186,
+ "grad_norm": 0.014403033070266247,
+ "learning_rate": 0.0006,
+ "loss": 4.561079978942871,
+ "step": 4253
+ },
+ {
+ "epoch": 59.083842794759825,
+ "grad_norm": 0.016243597492575645,
+ "learning_rate": 0.0006,
+ "loss": 4.673602104187012,
+ "step": 4254
+ },
+ {
+ "epoch": 59.09781659388646,
+ "grad_norm": 0.016492877155542374,
+ "learning_rate": 0.0006,
+ "loss": 4.564824104309082,
+ "step": 4255
+ },
+ {
+ "epoch": 59.1117903930131,
+ "grad_norm": 0.015980370342731476,
+ "learning_rate": 0.0006,
+ "loss": 4.339249134063721,
+ "step": 4256
+ },
+ {
+ "epoch": 59.12576419213974,
+ "grad_norm": 0.019687356427311897,
+ "learning_rate": 0.0006,
+ "loss": 4.481834888458252,
+ "step": 4257
+ },
+ {
+ "epoch": 59.13973799126637,
+ "grad_norm": 0.02983798086643219,
+ "learning_rate": 0.0006,
+ "loss": 4.4495697021484375,
+ "step": 4258
+ },
+ {
+ "epoch": 59.15371179039301,
+ "grad_norm": 0.04600981995463371,
+ "learning_rate": 0.0006,
+ "loss": 4.567838668823242,
+ "step": 4259
+ },
+ {
+ "epoch": 59.16768558951965,
+ "grad_norm": 0.04852001741528511,
+ "learning_rate": 0.0006,
+ "loss": 4.594701766967773,
+ "step": 4260
+ },
+ {
+ "epoch": 59.18165938864629,
+ "grad_norm": 0.03429504856467247,
+ "learning_rate": 0.0006,
+ "loss": 4.573735237121582,
+ "step": 4261
+ },
+ {
+ "epoch": 59.19563318777293,
+ "grad_norm": 0.029345160350203514,
+ "learning_rate": 0.0006,
+ "loss": 4.488258361816406,
+ "step": 4262
+ },
+ {
+ "epoch": 59.209606986899566,
+ "grad_norm": 0.01961623504757881,
+ "learning_rate": 0.0006,
+ "loss": 4.545222282409668,
+ "step": 4263
+ },
+ {
+ "epoch": 59.223580786026204,
+ "grad_norm": 0.020406130701303482,
+ "learning_rate": 0.0006,
+ "loss": 4.471898078918457,
+ "step": 4264
+ },
+ {
+ "epoch": 59.237554585152836,
+ "grad_norm": 0.020500272512435913,
+ "learning_rate": 0.0006,
+ "loss": 4.538337707519531,
+ "step": 4265
+ },
+ {
+ "epoch": 59.251528384279474,
+ "grad_norm": 0.020727017894387245,
+ "learning_rate": 0.0006,
+ "loss": 4.437284469604492,
+ "step": 4266
+ },
+ {
+ "epoch": 59.26550218340611,
+ "grad_norm": 0.03395998477935791,
+ "learning_rate": 0.0006,
+ "loss": 4.459124565124512,
+ "step": 4267
+ },
+ {
+ "epoch": 59.27947598253275,
+ "grad_norm": 0.04635605961084366,
+ "learning_rate": 0.0006,
+ "loss": 4.520228385925293,
+ "step": 4268
+ },
+ {
+ "epoch": 59.29344978165939,
+ "grad_norm": 0.04253006726503372,
+ "learning_rate": 0.0006,
+ "loss": 4.578673362731934,
+ "step": 4269
+ },
+ {
+ "epoch": 59.30742358078603,
+ "grad_norm": 0.02403509058058262,
+ "learning_rate": 0.0006,
+ "loss": 4.484639644622803,
+ "step": 4270
+ },
+ {
+ "epoch": 59.32139737991266,
+ "grad_norm": 0.02254190482199192,
+ "learning_rate": 0.0006,
+ "loss": 4.511314868927002,
+ "step": 4271
+ },
+ {
+ "epoch": 59.3353711790393,
+ "grad_norm": 0.021585190668702126,
+ "learning_rate": 0.0006,
+ "loss": 4.530239105224609,
+ "step": 4272
+ },
+ {
+ "epoch": 59.34934497816594,
+ "grad_norm": 0.021210316568613052,
+ "learning_rate": 0.0006,
+ "loss": 4.593514442443848,
+ "step": 4273
+ },
+ {
+ "epoch": 59.36331877729258,
+ "grad_norm": 0.019317107275128365,
+ "learning_rate": 0.0006,
+ "loss": 4.5591535568237305,
+ "step": 4274
+ },
+ {
+ "epoch": 59.377292576419215,
+ "grad_norm": 0.018059708178043365,
+ "learning_rate": 0.0006,
+ "loss": 4.4499592781066895,
+ "step": 4275
+ },
+ {
+ "epoch": 59.391266375545854,
+ "grad_norm": 0.017847446724772453,
+ "learning_rate": 0.0006,
+ "loss": 4.473999500274658,
+ "step": 4276
+ },
+ {
+ "epoch": 59.40524017467249,
+ "grad_norm": 0.014944756403565407,
+ "learning_rate": 0.0006,
+ "loss": 4.5934529304504395,
+ "step": 4277
+ },
+ {
+ "epoch": 59.419213973799124,
+ "grad_norm": 0.01663832925260067,
+ "learning_rate": 0.0006,
+ "loss": 4.504262924194336,
+ "step": 4278
+ },
+ {
+ "epoch": 59.43318777292576,
+ "grad_norm": 0.015571796335279942,
+ "learning_rate": 0.0006,
+ "loss": 4.512310981750488,
+ "step": 4279
+ },
+ {
+ "epoch": 59.4471615720524,
+ "grad_norm": 0.015896400436758995,
+ "learning_rate": 0.0006,
+ "loss": 4.452714920043945,
+ "step": 4280
+ },
+ {
+ "epoch": 59.46113537117904,
+ "grad_norm": 0.016010284423828125,
+ "learning_rate": 0.0006,
+ "loss": 4.6188063621521,
+ "step": 4281
+ },
+ {
+ "epoch": 59.47510917030568,
+ "grad_norm": 0.015075593255460262,
+ "learning_rate": 0.0006,
+ "loss": 4.477978229522705,
+ "step": 4282
+ },
+ {
+ "epoch": 59.48908296943232,
+ "grad_norm": 0.014698950573801994,
+ "learning_rate": 0.0006,
+ "loss": 4.544955730438232,
+ "step": 4283
+ },
+ {
+ "epoch": 59.50305676855895,
+ "grad_norm": 0.0136245833709836,
+ "learning_rate": 0.0006,
+ "loss": 4.530460357666016,
+ "step": 4284
+ },
+ {
+ "epoch": 59.51703056768559,
+ "grad_norm": 0.013118310831487179,
+ "learning_rate": 0.0006,
+ "loss": 4.614217758178711,
+ "step": 4285
+ },
+ {
+ "epoch": 59.531004366812226,
+ "grad_norm": 0.01295961532741785,
+ "learning_rate": 0.0006,
+ "loss": 4.43156099319458,
+ "step": 4286
+ },
+ {
+ "epoch": 59.544978165938865,
+ "grad_norm": 0.012429811991751194,
+ "learning_rate": 0.0006,
+ "loss": 4.517333030700684,
+ "step": 4287
+ },
+ {
+ "epoch": 59.5589519650655,
+ "grad_norm": 0.01261830423027277,
+ "learning_rate": 0.0006,
+ "loss": 4.569780349731445,
+ "step": 4288
+ },
+ {
+ "epoch": 59.57292576419214,
+ "grad_norm": 0.012394499965012074,
+ "learning_rate": 0.0006,
+ "loss": 4.5970869064331055,
+ "step": 4289
+ },
+ {
+ "epoch": 59.58689956331878,
+ "grad_norm": 0.01290759164839983,
+ "learning_rate": 0.0006,
+ "loss": 4.437100410461426,
+ "step": 4290
+ },
+ {
+ "epoch": 59.60087336244541,
+ "grad_norm": 0.012042169459164143,
+ "learning_rate": 0.0006,
+ "loss": 4.567404747009277,
+ "step": 4291
+ },
+ {
+ "epoch": 59.61484716157205,
+ "grad_norm": 0.012338520959019661,
+ "learning_rate": 0.0006,
+ "loss": 4.476193428039551,
+ "step": 4292
+ },
+ {
+ "epoch": 59.62882096069869,
+ "grad_norm": 0.01321258582174778,
+ "learning_rate": 0.0006,
+ "loss": 4.566680431365967,
+ "step": 4293
+ },
+ {
+ "epoch": 59.64279475982533,
+ "grad_norm": 0.013478122651576996,
+ "learning_rate": 0.0006,
+ "loss": 4.487398147583008,
+ "step": 4294
+ },
+ {
+ "epoch": 59.65676855895197,
+ "grad_norm": 0.01537603884935379,
+ "learning_rate": 0.0006,
+ "loss": 4.449337005615234,
+ "step": 4295
+ },
+ {
+ "epoch": 59.670742358078606,
+ "grad_norm": 0.017140565440058708,
+ "learning_rate": 0.0006,
+ "loss": 4.5174360275268555,
+ "step": 4296
+ },
+ {
+ "epoch": 59.68471615720524,
+ "grad_norm": 0.017931461334228516,
+ "learning_rate": 0.0006,
+ "loss": 4.522104740142822,
+ "step": 4297
+ },
+ {
+ "epoch": 59.698689956331876,
+ "grad_norm": 0.018037918955087662,
+ "learning_rate": 0.0006,
+ "loss": 4.496928691864014,
+ "step": 4298
+ },
+ {
+ "epoch": 59.712663755458514,
+ "grad_norm": 0.01684504561126232,
+ "learning_rate": 0.0006,
+ "loss": 4.5332417488098145,
+ "step": 4299
+ },
+ {
+ "epoch": 59.72663755458515,
+ "grad_norm": 0.016268383711576462,
+ "learning_rate": 0.0006,
+ "loss": 4.468893051147461,
+ "step": 4300
+ },
+ {
+ "epoch": 59.74061135371179,
+ "grad_norm": 0.014859385788440704,
+ "learning_rate": 0.0006,
+ "loss": 4.551092147827148,
+ "step": 4301
+ },
+ {
+ "epoch": 59.75458515283843,
+ "grad_norm": 0.013589047826826572,
+ "learning_rate": 0.0006,
+ "loss": 4.470579147338867,
+ "step": 4302
+ },
+ {
+ "epoch": 59.76855895196506,
+ "grad_norm": 0.015704551711678505,
+ "learning_rate": 0.0006,
+ "loss": 4.330068111419678,
+ "step": 4303
+ },
+ {
+ "epoch": 59.7825327510917,
+ "grad_norm": 0.017869921401143074,
+ "learning_rate": 0.0006,
+ "loss": 4.428255081176758,
+ "step": 4304
+ },
+ {
+ "epoch": 59.79650655021834,
+ "grad_norm": 0.019828537479043007,
+ "learning_rate": 0.0006,
+ "loss": 4.496023178100586,
+ "step": 4305
+ },
+ {
+ "epoch": 59.81048034934498,
+ "grad_norm": 0.015856236219406128,
+ "learning_rate": 0.0006,
+ "loss": 4.427152156829834,
+ "step": 4306
+ },
+ {
+ "epoch": 59.82445414847162,
+ "grad_norm": 0.013349304907023907,
+ "learning_rate": 0.0006,
+ "loss": 4.417333602905273,
+ "step": 4307
+ },
+ {
+ "epoch": 59.838427947598255,
+ "grad_norm": 0.014111637137830257,
+ "learning_rate": 0.0006,
+ "loss": 4.4507575035095215,
+ "step": 4308
+ },
+ {
+ "epoch": 59.852401746724894,
+ "grad_norm": 0.015186597593128681,
+ "learning_rate": 0.0006,
+ "loss": 4.375116348266602,
+ "step": 4309
+ },
+ {
+ "epoch": 59.866375545851525,
+ "grad_norm": 0.01604730449616909,
+ "learning_rate": 0.0006,
+ "loss": 4.604518890380859,
+ "step": 4310
+ },
+ {
+ "epoch": 59.880349344978164,
+ "grad_norm": 0.015308464877307415,
+ "learning_rate": 0.0006,
+ "loss": 4.507232189178467,
+ "step": 4311
+ },
+ {
+ "epoch": 59.8943231441048,
+ "grad_norm": 0.014017206616699696,
+ "learning_rate": 0.0006,
+ "loss": 4.398487091064453,
+ "step": 4312
+ },
+ {
+ "epoch": 59.90829694323144,
+ "grad_norm": 0.012911394238471985,
+ "learning_rate": 0.0006,
+ "loss": 4.567872047424316,
+ "step": 4313
+ },
+ {
+ "epoch": 59.92227074235808,
+ "grad_norm": 0.012775142677128315,
+ "learning_rate": 0.0006,
+ "loss": 4.562105655670166,
+ "step": 4314
+ },
+ {
+ "epoch": 59.93624454148472,
+ "grad_norm": 0.014750408008694649,
+ "learning_rate": 0.0006,
+ "loss": 4.456263542175293,
+ "step": 4315
+ },
+ {
+ "epoch": 59.95021834061135,
+ "grad_norm": 0.016726011410355568,
+ "learning_rate": 0.0006,
+ "loss": 4.360722064971924,
+ "step": 4316
+ },
+ {
+ "epoch": 59.96419213973799,
+ "grad_norm": 0.01734776981174946,
+ "learning_rate": 0.0006,
+ "loss": 4.572544574737549,
+ "step": 4317
+ },
+ {
+ "epoch": 59.97816593886463,
+ "grad_norm": 0.01822415366768837,
+ "learning_rate": 0.0006,
+ "loss": 4.490501880645752,
+ "step": 4318
+ },
+ {
+ "epoch": 59.992139737991266,
+ "grad_norm": 0.014638577587902546,
+ "learning_rate": 0.0006,
+ "loss": 4.440951824188232,
+ "step": 4319
+ },
+ {
+ "epoch": 60.0,
+ "grad_norm": 0.01255202479660511,
+ "learning_rate": 0.0006,
+ "loss": 4.559123516082764,
+ "step": 4320
+ },
+ {
+ "epoch": 60.0,
+ "eval_loss": 4.741089344024658,
+ "eval_runtime": 56.8997,
+ "eval_samples_per_second": 42.918,
+ "eval_steps_per_second": 1.353,
+ "step": 4320
+ },
+ {
+ "epoch": 60.01397379912664,
+ "grad_norm": 0.016548393294215202,
+ "learning_rate": 0.0006,
+ "loss": 4.442024230957031,
+ "step": 4321
+ },
+ {
+ "epoch": 60.02794759825328,
+ "grad_norm": 0.019759127870202065,
+ "learning_rate": 0.0006,
+ "loss": 4.394334316253662,
+ "step": 4322
+ },
+ {
+ "epoch": 60.041921397379916,
+ "grad_norm": 0.018325170502066612,
+ "learning_rate": 0.0006,
+ "loss": 4.318270683288574,
+ "step": 4323
+ },
+ {
+ "epoch": 60.05589519650655,
+ "grad_norm": 0.015289463102817535,
+ "learning_rate": 0.0006,
+ "loss": 4.318228721618652,
+ "step": 4324
+ },
+ {
+ "epoch": 60.069868995633186,
+ "grad_norm": 0.015424872748553753,
+ "learning_rate": 0.0006,
+ "loss": 4.3605499267578125,
+ "step": 4325
+ },
+ {
+ "epoch": 60.083842794759825,
+ "grad_norm": 0.014085669070482254,
+ "learning_rate": 0.0006,
+ "loss": 4.359585762023926,
+ "step": 4326
+ },
+ {
+ "epoch": 60.09781659388646,
+ "grad_norm": 0.014004210010170937,
+ "learning_rate": 0.0006,
+ "loss": 4.447783470153809,
+ "step": 4327
+ },
+ {
+ "epoch": 60.1117903930131,
+ "grad_norm": 0.013514582999050617,
+ "learning_rate": 0.0006,
+ "loss": 4.4850239753723145,
+ "step": 4328
+ },
+ {
+ "epoch": 60.12576419213974,
+ "grad_norm": 0.013049394823610783,
+ "learning_rate": 0.0006,
+ "loss": 4.528675556182861,
+ "step": 4329
+ },
+ {
+ "epoch": 60.13973799126637,
+ "grad_norm": 0.01378058921545744,
+ "learning_rate": 0.0006,
+ "loss": 4.444823741912842,
+ "step": 4330
+ },
+ {
+ "epoch": 60.15371179039301,
+ "grad_norm": 0.013752233237028122,
+ "learning_rate": 0.0006,
+ "loss": 4.5205206871032715,
+ "step": 4331
+ },
+ {
+ "epoch": 60.16768558951965,
+ "grad_norm": 0.013431803323328495,
+ "learning_rate": 0.0006,
+ "loss": 4.547320365905762,
+ "step": 4332
+ },
+ {
+ "epoch": 60.18165938864629,
+ "grad_norm": 0.014544196426868439,
+ "learning_rate": 0.0006,
+ "loss": 4.517143249511719,
+ "step": 4333
+ },
+ {
+ "epoch": 60.19563318777293,
+ "grad_norm": 0.01612176187336445,
+ "learning_rate": 0.0006,
+ "loss": 4.517232894897461,
+ "step": 4334
+ },
+ {
+ "epoch": 60.209606986899566,
+ "grad_norm": 0.01803239807486534,
+ "learning_rate": 0.0006,
+ "loss": 4.377574920654297,
+ "step": 4335
+ },
+ {
+ "epoch": 60.223580786026204,
+ "grad_norm": 0.014871489256620407,
+ "learning_rate": 0.0006,
+ "loss": 4.538285255432129,
+ "step": 4336
+ },
+ {
+ "epoch": 60.237554585152836,
+ "grad_norm": 0.013927377760410309,
+ "learning_rate": 0.0006,
+ "loss": 4.386175632476807,
+ "step": 4337
+ },
+ {
+ "epoch": 60.251528384279474,
+ "grad_norm": 0.014757219702005386,
+ "learning_rate": 0.0006,
+ "loss": 4.528553009033203,
+ "step": 4338
+ },
+ {
+ "epoch": 60.26550218340611,
+ "grad_norm": 0.01433873176574707,
+ "learning_rate": 0.0006,
+ "loss": 4.476929187774658,
+ "step": 4339
+ },
+ {
+ "epoch": 60.27947598253275,
+ "grad_norm": 0.015768803656101227,
+ "learning_rate": 0.0006,
+ "loss": 4.50970458984375,
+ "step": 4340
+ },
+ {
+ "epoch": 60.29344978165939,
+ "grad_norm": 0.015651429072022438,
+ "learning_rate": 0.0006,
+ "loss": 4.427133083343506,
+ "step": 4341
+ },
+ {
+ "epoch": 60.30742358078603,
+ "grad_norm": 0.01673000678420067,
+ "learning_rate": 0.0006,
+ "loss": 4.447723865509033,
+ "step": 4342
+ },
+ {
+ "epoch": 60.32139737991266,
+ "grad_norm": 0.018481194972991943,
+ "learning_rate": 0.0006,
+ "loss": 4.559832572937012,
+ "step": 4343
+ },
+ {
+ "epoch": 60.3353711790393,
+ "grad_norm": 0.016731027513742447,
+ "learning_rate": 0.0006,
+ "loss": 4.470893859863281,
+ "step": 4344
+ },
+ {
+ "epoch": 60.34934497816594,
+ "grad_norm": 0.01504011545330286,
+ "learning_rate": 0.0006,
+ "loss": 4.483014106750488,
+ "step": 4345
+ },
+ {
+ "epoch": 60.36331877729258,
+ "grad_norm": 0.01602456159889698,
+ "learning_rate": 0.0006,
+ "loss": 4.548428058624268,
+ "step": 4346
+ },
+ {
+ "epoch": 60.377292576419215,
+ "grad_norm": 0.016027364879846573,
+ "learning_rate": 0.0006,
+ "loss": 4.426779747009277,
+ "step": 4347
+ },
+ {
+ "epoch": 60.391266375545854,
+ "grad_norm": 0.01873827911913395,
+ "learning_rate": 0.0006,
+ "loss": 4.353686332702637,
+ "step": 4348
+ },
+ {
+ "epoch": 60.40524017467249,
+ "grad_norm": 0.01907406374812126,
+ "learning_rate": 0.0006,
+ "loss": 4.502945423126221,
+ "step": 4349
+ },
+ {
+ "epoch": 60.419213973799124,
+ "grad_norm": 0.016083354130387306,
+ "learning_rate": 0.0006,
+ "loss": 4.431731700897217,
+ "step": 4350
+ },
+ {
+ "epoch": 60.43318777292576,
+ "grad_norm": 0.015293709933757782,
+ "learning_rate": 0.0006,
+ "loss": 4.480112552642822,
+ "step": 4351
+ },
+ {
+ "epoch": 60.4471615720524,
+ "grad_norm": 0.01688101328909397,
+ "learning_rate": 0.0006,
+ "loss": 4.50853157043457,
+ "step": 4352
+ },
+ {
+ "epoch": 60.46113537117904,
+ "grad_norm": 0.02094118855893612,
+ "learning_rate": 0.0006,
+ "loss": 4.359930038452148,
+ "step": 4353
+ },
+ {
+ "epoch": 60.47510917030568,
+ "grad_norm": 0.018188107758760452,
+ "learning_rate": 0.0006,
+ "loss": 4.549592018127441,
+ "step": 4354
+ },
+ {
+ "epoch": 60.48908296943232,
+ "grad_norm": 0.01462319865822792,
+ "learning_rate": 0.0006,
+ "loss": 4.49467658996582,
+ "step": 4355
+ },
+ {
+ "epoch": 60.50305676855895,
+ "grad_norm": 0.017508579418063164,
+ "learning_rate": 0.0006,
+ "loss": 4.398318767547607,
+ "step": 4356
+ },
+ {
+ "epoch": 60.51703056768559,
+ "grad_norm": 0.018200315535068512,
+ "learning_rate": 0.0006,
+ "loss": 4.44169282913208,
+ "step": 4357
+ },
+ {
+ "epoch": 60.531004366812226,
+ "grad_norm": 0.01717468351125717,
+ "learning_rate": 0.0006,
+ "loss": 4.481400489807129,
+ "step": 4358
+ },
+ {
+ "epoch": 60.544978165938865,
+ "grad_norm": 0.019329151138663292,
+ "learning_rate": 0.0006,
+ "loss": 4.578839302062988,
+ "step": 4359
+ },
+ {
+ "epoch": 60.5589519650655,
+ "grad_norm": 0.016109555959701538,
+ "learning_rate": 0.0006,
+ "loss": 4.461043357849121,
+ "step": 4360
+ },
+ {
+ "epoch": 60.57292576419214,
+ "grad_norm": 0.01494457945227623,
+ "learning_rate": 0.0006,
+ "loss": 4.505096912384033,
+ "step": 4361
+ },
+ {
+ "epoch": 60.58689956331878,
+ "grad_norm": 0.01729399710893631,
+ "learning_rate": 0.0006,
+ "loss": 4.366747856140137,
+ "step": 4362
+ },
+ {
+ "epoch": 60.60087336244541,
+ "grad_norm": 0.01705821603536606,
+ "learning_rate": 0.0006,
+ "loss": 4.42534065246582,
+ "step": 4363
+ },
+ {
+ "epoch": 60.61484716157205,
+ "grad_norm": 0.018391354009509087,
+ "learning_rate": 0.0006,
+ "loss": 4.350294589996338,
+ "step": 4364
+ },
+ {
+ "epoch": 60.62882096069869,
+ "grad_norm": 0.016989829018712044,
+ "learning_rate": 0.0006,
+ "loss": 4.45905876159668,
+ "step": 4365
+ },
+ {
+ "epoch": 60.64279475982533,
+ "grad_norm": 0.017192212864756584,
+ "learning_rate": 0.0006,
+ "loss": 4.526487350463867,
+ "step": 4366
+ },
+ {
+ "epoch": 60.65676855895197,
+ "grad_norm": 0.018091315403580666,
+ "learning_rate": 0.0006,
+ "loss": 4.309333324432373,
+ "step": 4367
+ },
+ {
+ "epoch": 60.670742358078606,
+ "grad_norm": 0.014051459729671478,
+ "learning_rate": 0.0006,
+ "loss": 4.338171005249023,
+ "step": 4368
+ },
+ {
+ "epoch": 60.68471615720524,
+ "grad_norm": 0.013166418299078941,
+ "learning_rate": 0.0006,
+ "loss": 4.409758567810059,
+ "step": 4369
+ },
+ {
+ "epoch": 60.698689956331876,
+ "grad_norm": 0.014260428957641125,
+ "learning_rate": 0.0006,
+ "loss": 4.524338245391846,
+ "step": 4370
+ },
+ {
+ "epoch": 60.712663755458514,
+ "grad_norm": 0.017019255086779594,
+ "learning_rate": 0.0006,
+ "loss": 4.387436866760254,
+ "step": 4371
+ },
+ {
+ "epoch": 60.72663755458515,
+ "grad_norm": 0.017070675268769264,
+ "learning_rate": 0.0006,
+ "loss": 4.437321662902832,
+ "step": 4372
+ },
+ {
+ "epoch": 60.74061135371179,
+ "grad_norm": 0.015197164379060268,
+ "learning_rate": 0.0006,
+ "loss": 4.371205806732178,
+ "step": 4373
+ },
+ {
+ "epoch": 60.75458515283843,
+ "grad_norm": 0.014791185967624187,
+ "learning_rate": 0.0006,
+ "loss": 4.44439697265625,
+ "step": 4374
+ },
+ {
+ "epoch": 60.76855895196506,
+ "grad_norm": 0.016458261758089066,
+ "learning_rate": 0.0006,
+ "loss": 4.445584297180176,
+ "step": 4375
+ },
+ {
+ "epoch": 60.7825327510917,
+ "grad_norm": 0.017706342041492462,
+ "learning_rate": 0.0006,
+ "loss": 4.348708152770996,
+ "step": 4376
+ },
+ {
+ "epoch": 60.79650655021834,
+ "grad_norm": 0.017561476677656174,
+ "learning_rate": 0.0006,
+ "loss": 4.521929740905762,
+ "step": 4377
+ },
+ {
+ "epoch": 60.81048034934498,
+ "grad_norm": 0.016557445749640465,
+ "learning_rate": 0.0006,
+ "loss": 4.545806884765625,
+ "step": 4378
+ },
+ {
+ "epoch": 60.82445414847162,
+ "grad_norm": 0.016953222453594208,
+ "learning_rate": 0.0006,
+ "loss": 4.4230875968933105,
+ "step": 4379
+ },
+ {
+ "epoch": 60.838427947598255,
+ "grad_norm": 0.017094967886805534,
+ "learning_rate": 0.0006,
+ "loss": 4.423985481262207,
+ "step": 4380
+ },
+ {
+ "epoch": 60.852401746724894,
+ "grad_norm": 0.014787515625357628,
+ "learning_rate": 0.0006,
+ "loss": 4.39473819732666,
+ "step": 4381
+ },
+ {
+ "epoch": 60.866375545851525,
+ "grad_norm": 0.016208263114094734,
+ "learning_rate": 0.0006,
+ "loss": 4.3810038566589355,
+ "step": 4382
+ },
+ {
+ "epoch": 60.880349344978164,
+ "grad_norm": 0.017844321206212044,
+ "learning_rate": 0.0006,
+ "loss": 4.35086727142334,
+ "step": 4383
+ },
+ {
+ "epoch": 60.8943231441048,
+ "grad_norm": 0.017539294436573982,
+ "learning_rate": 0.0006,
+ "loss": 4.460260391235352,
+ "step": 4384
+ },
+ {
+ "epoch": 60.90829694323144,
+ "grad_norm": 0.01612042263150215,
+ "learning_rate": 0.0006,
+ "loss": 4.370500564575195,
+ "step": 4385
+ },
+ {
+ "epoch": 60.92227074235808,
+ "grad_norm": 0.015381601639091969,
+ "learning_rate": 0.0006,
+ "loss": 4.435868263244629,
+ "step": 4386
+ },
+ {
+ "epoch": 60.93624454148472,
+ "grad_norm": 0.01603585295379162,
+ "learning_rate": 0.0006,
+ "loss": 4.434070587158203,
+ "step": 4387
+ },
+ {
+ "epoch": 60.95021834061135,
+ "grad_norm": 0.016268130391836166,
+ "learning_rate": 0.0006,
+ "loss": 4.404343128204346,
+ "step": 4388
+ },
+ {
+ "epoch": 60.96419213973799,
+ "grad_norm": 0.014980032108724117,
+ "learning_rate": 0.0006,
+ "loss": 4.399213790893555,
+ "step": 4389
+ },
+ {
+ "epoch": 60.97816593886463,
+ "grad_norm": 0.016899267211556435,
+ "learning_rate": 0.0006,
+ "loss": 4.508069038391113,
+ "step": 4390
+ },
+ {
+ "epoch": 60.992139737991266,
+ "grad_norm": 0.013756908476352692,
+ "learning_rate": 0.0006,
+ "loss": 4.541858673095703,
+ "step": 4391
+ },
+ {
+ "epoch": 61.0,
+ "grad_norm": 0.01578413136303425,
+ "learning_rate": 0.0006,
+ "loss": 4.543513298034668,
+ "step": 4392
+ },
+ {
+ "epoch": 61.0,
+ "eval_loss": 4.756770610809326,
+ "eval_runtime": 56.4572,
+ "eval_samples_per_second": 43.254,
+ "eval_steps_per_second": 1.364,
+ "step": 4392
+ },
+ {
+ "epoch": 61.01397379912664,
+ "grad_norm": 0.013387506827712059,
+ "learning_rate": 0.0006,
+ "loss": 4.475942611694336,
+ "step": 4393
+ },
+ {
+ "epoch": 61.02794759825328,
+ "grad_norm": 0.014278904534876347,
+ "learning_rate": 0.0006,
+ "loss": 4.442403793334961,
+ "step": 4394
+ },
+ {
+ "epoch": 61.041921397379916,
+ "grad_norm": 0.014359553344547749,
+ "learning_rate": 0.0006,
+ "loss": 4.509463310241699,
+ "step": 4395
+ },
+ {
+ "epoch": 61.05589519650655,
+ "grad_norm": 0.01718050241470337,
+ "learning_rate": 0.0006,
+ "loss": 4.354458808898926,
+ "step": 4396
+ },
+ {
+ "epoch": 61.069868995633186,
+ "grad_norm": 0.015988312661647797,
+ "learning_rate": 0.0006,
+ "loss": 4.484508514404297,
+ "step": 4397
+ },
+ {
+ "epoch": 61.083842794759825,
+ "grad_norm": 0.015466735698282719,
+ "learning_rate": 0.0006,
+ "loss": 4.326908111572266,
+ "step": 4398
+ },
+ {
+ "epoch": 61.09781659388646,
+ "grad_norm": 0.014541332609951496,
+ "learning_rate": 0.0006,
+ "loss": 4.405823230743408,
+ "step": 4399
+ },
+ {
+ "epoch": 61.1117903930131,
+ "grad_norm": 0.014803987927734852,
+ "learning_rate": 0.0006,
+ "loss": 4.375386714935303,
+ "step": 4400
+ },
+ {
+ "epoch": 61.12576419213974,
+ "grad_norm": 0.013962093740701675,
+ "learning_rate": 0.0006,
+ "loss": 4.300838470458984,
+ "step": 4401
+ },
+ {
+ "epoch": 61.13973799126637,
+ "grad_norm": 0.015213954262435436,
+ "learning_rate": 0.0006,
+ "loss": 4.388306617736816,
+ "step": 4402
+ },
+ {
+ "epoch": 61.15371179039301,
+ "grad_norm": 0.018965506926178932,
+ "learning_rate": 0.0006,
+ "loss": 4.388340950012207,
+ "step": 4403
+ },
+ {
+ "epoch": 61.16768558951965,
+ "grad_norm": 0.01958410255610943,
+ "learning_rate": 0.0006,
+ "loss": 4.35012149810791,
+ "step": 4404
+ },
+ {
+ "epoch": 61.18165938864629,
+ "grad_norm": 0.01938101276755333,
+ "learning_rate": 0.0006,
+ "loss": 4.368372917175293,
+ "step": 4405
+ },
+ {
+ "epoch": 61.19563318777293,
+ "grad_norm": 0.016556516289711,
+ "learning_rate": 0.0006,
+ "loss": 4.486198425292969,
+ "step": 4406
+ },
+ {
+ "epoch": 61.209606986899566,
+ "grad_norm": 0.017031289637088776,
+ "learning_rate": 0.0006,
+ "loss": 4.421676158905029,
+ "step": 4407
+ },
+ {
+ "epoch": 61.223580786026204,
+ "grad_norm": 0.018023479729890823,
+ "learning_rate": 0.0006,
+ "loss": 4.38751220703125,
+ "step": 4408
+ },
+ {
+ "epoch": 61.237554585152836,
+ "grad_norm": 0.0178135447204113,
+ "learning_rate": 0.0006,
+ "loss": 4.336528301239014,
+ "step": 4409
+ },
+ {
+ "epoch": 61.251528384279474,
+ "grad_norm": 0.016782190650701523,
+ "learning_rate": 0.0006,
+ "loss": 4.386693954467773,
+ "step": 4410
+ },
+ {
+ "epoch": 61.26550218340611,
+ "grad_norm": 0.015494848601520061,
+ "learning_rate": 0.0006,
+ "loss": 4.308948040008545,
+ "step": 4411
+ },
+ {
+ "epoch": 61.27947598253275,
+ "grad_norm": 0.015002378262579441,
+ "learning_rate": 0.0006,
+ "loss": 4.314206123352051,
+ "step": 4412
+ },
+ {
+ "epoch": 61.29344978165939,
+ "grad_norm": 0.015798920765519142,
+ "learning_rate": 0.0006,
+ "loss": 4.502900123596191,
+ "step": 4413
+ },
+ {
+ "epoch": 61.30742358078603,
+ "grad_norm": 0.015444336459040642,
+ "learning_rate": 0.0006,
+ "loss": 4.473160743713379,
+ "step": 4414
+ },
+ {
+ "epoch": 61.32139737991266,
+ "grad_norm": 0.017025096341967583,
+ "learning_rate": 0.0006,
+ "loss": 4.460621356964111,
+ "step": 4415
+ },
+ {
+ "epoch": 61.3353711790393,
+ "grad_norm": 0.019502537325024605,
+ "learning_rate": 0.0006,
+ "loss": 4.385551452636719,
+ "step": 4416
+ },
+ {
+ "epoch": 61.34934497816594,
+ "grad_norm": 0.018529541790485382,
+ "learning_rate": 0.0006,
+ "loss": 4.425390720367432,
+ "step": 4417
+ },
+ {
+ "epoch": 61.36331877729258,
+ "grad_norm": 0.013155706226825714,
+ "learning_rate": 0.0006,
+ "loss": 4.467020034790039,
+ "step": 4418
+ },
+ {
+ "epoch": 61.377292576419215,
+ "grad_norm": 0.01518090907484293,
+ "learning_rate": 0.0006,
+ "loss": 4.369664192199707,
+ "step": 4419
+ },
+ {
+ "epoch": 61.391266375545854,
+ "grad_norm": 0.01626511849462986,
+ "learning_rate": 0.0006,
+ "loss": 4.463231086730957,
+ "step": 4420
+ },
+ {
+ "epoch": 61.40524017467249,
+ "grad_norm": 0.01701558195054531,
+ "learning_rate": 0.0006,
+ "loss": 4.43708610534668,
+ "step": 4421
+ },
+ {
+ "epoch": 61.419213973799124,
+ "grad_norm": 0.01761862449347973,
+ "learning_rate": 0.0006,
+ "loss": 4.3314208984375,
+ "step": 4422
+ },
+ {
+ "epoch": 61.43318777292576,
+ "grad_norm": 0.017084648832678795,
+ "learning_rate": 0.0006,
+ "loss": 4.39676570892334,
+ "step": 4423
+ },
+ {
+ "epoch": 61.4471615720524,
+ "grad_norm": 0.01667078025639057,
+ "learning_rate": 0.0006,
+ "loss": 4.3789873123168945,
+ "step": 4424
+ },
+ {
+ "epoch": 61.46113537117904,
+ "grad_norm": 0.01877436414361,
+ "learning_rate": 0.0006,
+ "loss": 4.415275573730469,
+ "step": 4425
+ },
+ {
+ "epoch": 61.47510917030568,
+ "grad_norm": 0.020250486209988594,
+ "learning_rate": 0.0006,
+ "loss": 4.478635787963867,
+ "step": 4426
+ },
+ {
+ "epoch": 61.48908296943232,
+ "grad_norm": 0.020244868472218513,
+ "learning_rate": 0.0006,
+ "loss": 4.296854019165039,
+ "step": 4427
+ },
+ {
+ "epoch": 61.50305676855895,
+ "grad_norm": 0.018723875284194946,
+ "learning_rate": 0.0006,
+ "loss": 4.3760881423950195,
+ "step": 4428
+ },
+ {
+ "epoch": 61.51703056768559,
+ "grad_norm": 0.015138164162635803,
+ "learning_rate": 0.0006,
+ "loss": 4.39974308013916,
+ "step": 4429
+ },
+ {
+ "epoch": 61.531004366812226,
+ "grad_norm": 0.015111138112843037,
+ "learning_rate": 0.0006,
+ "loss": 4.34659481048584,
+ "step": 4430
+ },
+ {
+ "epoch": 61.544978165938865,
+ "grad_norm": 0.0164666585624218,
+ "learning_rate": 0.0006,
+ "loss": 4.415737152099609,
+ "step": 4431
+ },
+ {
+ "epoch": 61.5589519650655,
+ "grad_norm": 0.020891312509775162,
+ "learning_rate": 0.0006,
+ "loss": 4.300319671630859,
+ "step": 4432
+ },
+ {
+ "epoch": 61.57292576419214,
+ "grad_norm": 0.02089156024158001,
+ "learning_rate": 0.0006,
+ "loss": 4.373029708862305,
+ "step": 4433
+ },
+ {
+ "epoch": 61.58689956331878,
+ "grad_norm": 0.020114559680223465,
+ "learning_rate": 0.0006,
+ "loss": 4.434499740600586,
+ "step": 4434
+ },
+ {
+ "epoch": 61.60087336244541,
+ "grad_norm": 0.01791365258395672,
+ "learning_rate": 0.0006,
+ "loss": 4.466872215270996,
+ "step": 4435
+ },
+ {
+ "epoch": 61.61484716157205,
+ "grad_norm": 0.01879458874464035,
+ "learning_rate": 0.0006,
+ "loss": 4.5176520347595215,
+ "step": 4436
+ },
+ {
+ "epoch": 61.62882096069869,
+ "grad_norm": 0.016371937468647957,
+ "learning_rate": 0.0006,
+ "loss": 4.3507585525512695,
+ "step": 4437
+ },
+ {
+ "epoch": 61.64279475982533,
+ "grad_norm": 0.01648143120110035,
+ "learning_rate": 0.0006,
+ "loss": 4.398778438568115,
+ "step": 4438
+ },
+ {
+ "epoch": 61.65676855895197,
+ "grad_norm": 0.017240576446056366,
+ "learning_rate": 0.0006,
+ "loss": 4.513933181762695,
+ "step": 4439
+ },
+ {
+ "epoch": 61.670742358078606,
+ "grad_norm": 0.01734284684062004,
+ "learning_rate": 0.0006,
+ "loss": 4.315343856811523,
+ "step": 4440
+ },
+ {
+ "epoch": 61.68471615720524,
+ "grad_norm": 0.01639772206544876,
+ "learning_rate": 0.0006,
+ "loss": 4.4641218185424805,
+ "step": 4441
+ },
+ {
+ "epoch": 61.698689956331876,
+ "grad_norm": 0.0164708960801363,
+ "learning_rate": 0.0006,
+ "loss": 4.439968585968018,
+ "step": 4442
+ },
+ {
+ "epoch": 61.712663755458514,
+ "grad_norm": 0.01757933758199215,
+ "learning_rate": 0.0006,
+ "loss": 4.408327102661133,
+ "step": 4443
+ },
+ {
+ "epoch": 61.72663755458515,
+ "grad_norm": 0.01789589412510395,
+ "learning_rate": 0.0006,
+ "loss": 4.478257179260254,
+ "step": 4444
+ },
+ {
+ "epoch": 61.74061135371179,
+ "grad_norm": 0.01822184957563877,
+ "learning_rate": 0.0006,
+ "loss": 4.412046909332275,
+ "step": 4445
+ },
+ {
+ "epoch": 61.75458515283843,
+ "grad_norm": 0.018624618649482727,
+ "learning_rate": 0.0006,
+ "loss": 4.427122116088867,
+ "step": 4446
+ },
+ {
+ "epoch": 61.76855895196506,
+ "grad_norm": 0.018674887716770172,
+ "learning_rate": 0.0006,
+ "loss": 4.429147720336914,
+ "step": 4447
+ },
+ {
+ "epoch": 61.7825327510917,
+ "grad_norm": 0.015317119657993317,
+ "learning_rate": 0.0006,
+ "loss": 4.2808451652526855,
+ "step": 4448
+ },
+ {
+ "epoch": 61.79650655021834,
+ "grad_norm": 0.015400845557451248,
+ "learning_rate": 0.0006,
+ "loss": 4.420193672180176,
+ "step": 4449
+ },
+ {
+ "epoch": 61.81048034934498,
+ "grad_norm": 0.016743650659918785,
+ "learning_rate": 0.0006,
+ "loss": 4.498254776000977,
+ "step": 4450
+ },
+ {
+ "epoch": 61.82445414847162,
+ "grad_norm": 0.014330781064927578,
+ "learning_rate": 0.0006,
+ "loss": 4.445384979248047,
+ "step": 4451
+ },
+ {
+ "epoch": 61.838427947598255,
+ "grad_norm": 0.015046479180455208,
+ "learning_rate": 0.0006,
+ "loss": 4.368041038513184,
+ "step": 4452
+ },
+ {
+ "epoch": 61.852401746724894,
+ "grad_norm": 0.01630168780684471,
+ "learning_rate": 0.0006,
+ "loss": 4.460142135620117,
+ "step": 4453
+ },
+ {
+ "epoch": 61.866375545851525,
+ "grad_norm": 0.01637885719537735,
+ "learning_rate": 0.0006,
+ "loss": 4.416606426239014,
+ "step": 4454
+ },
+ {
+ "epoch": 61.880349344978164,
+ "grad_norm": 0.016785888001322746,
+ "learning_rate": 0.0006,
+ "loss": 4.452545642852783,
+ "step": 4455
+ },
+ {
+ "epoch": 61.8943231441048,
+ "grad_norm": 0.015044555068016052,
+ "learning_rate": 0.0006,
+ "loss": 4.427334785461426,
+ "step": 4456
+ },
+ {
+ "epoch": 61.90829694323144,
+ "grad_norm": 0.013699211180210114,
+ "learning_rate": 0.0006,
+ "loss": 4.408761024475098,
+ "step": 4457
+ },
+ {
+ "epoch": 61.92227074235808,
+ "grad_norm": 0.014021803624927998,
+ "learning_rate": 0.0006,
+ "loss": 4.543607711791992,
+ "step": 4458
+ },
+ {
+ "epoch": 61.93624454148472,
+ "grad_norm": 0.015438460744917393,
+ "learning_rate": 0.0006,
+ "loss": 4.313363075256348,
+ "step": 4459
+ },
+ {
+ "epoch": 61.95021834061135,
+ "grad_norm": 0.017515428364276886,
+ "learning_rate": 0.0006,
+ "loss": 4.310412406921387,
+ "step": 4460
+ },
+ {
+ "epoch": 61.96419213973799,
+ "grad_norm": 0.016683880239725113,
+ "learning_rate": 0.0006,
+ "loss": 4.3378448486328125,
+ "step": 4461
+ },
+ {
+ "epoch": 61.97816593886463,
+ "grad_norm": 0.019384048879146576,
+ "learning_rate": 0.0006,
+ "loss": 4.450568675994873,
+ "step": 4462
+ },
+ {
+ "epoch": 61.992139737991266,
+ "grad_norm": 0.020651493221521378,
+ "learning_rate": 0.0006,
+ "loss": 4.453197479248047,
+ "step": 4463
+ },
+ {
+ "epoch": 62.0,
+ "grad_norm": 0.020991764962673187,
+ "learning_rate": 0.0006,
+ "loss": 4.441527366638184,
+ "step": 4464
+ },
+ {
+ "epoch": 62.0,
+ "eval_loss": 4.700193881988525,
+ "eval_runtime": 56.6098,
+ "eval_samples_per_second": 43.137,
+ "eval_steps_per_second": 1.36,
+ "step": 4464
+ },
+ {
+ "epoch": 62.01397379912664,
+ "grad_norm": 0.01811373233795166,
+ "learning_rate": 0.0006,
+ "loss": 4.37346076965332,
+ "step": 4465
+ },
+ {
+ "epoch": 62.02794759825328,
+ "grad_norm": 0.01667127199470997,
+ "learning_rate": 0.0006,
+ "loss": 4.237151145935059,
+ "step": 4466
+ },
+ {
+ "epoch": 62.041921397379916,
+ "grad_norm": 0.014971534721553326,
+ "learning_rate": 0.0006,
+ "loss": 4.367053031921387,
+ "step": 4467
+ },
+ {
+ "epoch": 62.05589519650655,
+ "grad_norm": 0.015655893832445145,
+ "learning_rate": 0.0006,
+ "loss": 4.437426567077637,
+ "step": 4468
+ },
+ {
+ "epoch": 62.069868995633186,
+ "grad_norm": 0.014935447834432125,
+ "learning_rate": 0.0006,
+ "loss": 4.3635125160217285,
+ "step": 4469
+ },
+ {
+ "epoch": 62.083842794759825,
+ "grad_norm": 0.015363802202045918,
+ "learning_rate": 0.0006,
+ "loss": 4.40492057800293,
+ "step": 4470
+ },
+ {
+ "epoch": 62.09781659388646,
+ "grad_norm": 0.0159201230853796,
+ "learning_rate": 0.0006,
+ "loss": 4.392058849334717,
+ "step": 4471
+ },
+ {
+ "epoch": 62.1117903930131,
+ "grad_norm": 0.014598241075873375,
+ "learning_rate": 0.0006,
+ "loss": 4.369944095611572,
+ "step": 4472
+ },
+ {
+ "epoch": 62.12576419213974,
+ "grad_norm": 0.015394841320812702,
+ "learning_rate": 0.0006,
+ "loss": 4.425097942352295,
+ "step": 4473
+ },
+ {
+ "epoch": 62.13973799126637,
+ "grad_norm": 0.015913628041744232,
+ "learning_rate": 0.0006,
+ "loss": 4.420596122741699,
+ "step": 4474
+ },
+ {
+ "epoch": 62.15371179039301,
+ "grad_norm": 0.01575258933007717,
+ "learning_rate": 0.0006,
+ "loss": 4.32290506362915,
+ "step": 4475
+ },
+ {
+ "epoch": 62.16768558951965,
+ "grad_norm": 0.0155344782397151,
+ "learning_rate": 0.0006,
+ "loss": 4.273068428039551,
+ "step": 4476
+ },
+ {
+ "epoch": 62.18165938864629,
+ "grad_norm": 0.014947040006518364,
+ "learning_rate": 0.0006,
+ "loss": 4.384868621826172,
+ "step": 4477
+ },
+ {
+ "epoch": 62.19563318777293,
+ "grad_norm": 0.01557591650635004,
+ "learning_rate": 0.0006,
+ "loss": 4.458410263061523,
+ "step": 4478
+ },
+ {
+ "epoch": 62.209606986899566,
+ "grad_norm": 0.015182198956608772,
+ "learning_rate": 0.0006,
+ "loss": 4.300738334655762,
+ "step": 4479
+ },
+ {
+ "epoch": 62.223580786026204,
+ "grad_norm": 0.01634717360138893,
+ "learning_rate": 0.0006,
+ "loss": 4.436875820159912,
+ "step": 4480
+ },
+ {
+ "epoch": 62.237554585152836,
+ "grad_norm": 0.018813718110322952,
+ "learning_rate": 0.0006,
+ "loss": 4.454665184020996,
+ "step": 4481
+ },
+ {
+ "epoch": 62.251528384279474,
+ "grad_norm": 0.020965630188584328,
+ "learning_rate": 0.0006,
+ "loss": 4.357341766357422,
+ "step": 4482
+ },
+ {
+ "epoch": 62.26550218340611,
+ "grad_norm": 0.020738353952765465,
+ "learning_rate": 0.0006,
+ "loss": 4.456644058227539,
+ "step": 4483
+ },
+ {
+ "epoch": 62.27947598253275,
+ "grad_norm": 0.018813546746969223,
+ "learning_rate": 0.0006,
+ "loss": 4.381941795349121,
+ "step": 4484
+ },
+ {
+ "epoch": 62.29344978165939,
+ "grad_norm": 0.015431761741638184,
+ "learning_rate": 0.0006,
+ "loss": 4.41126823425293,
+ "step": 4485
+ },
+ {
+ "epoch": 62.30742358078603,
+ "grad_norm": 0.016335977241396904,
+ "learning_rate": 0.0006,
+ "loss": 4.419488906860352,
+ "step": 4486
+ },
+ {
+ "epoch": 62.32139737991266,
+ "grad_norm": 0.01832406409084797,
+ "learning_rate": 0.0006,
+ "loss": 4.579026222229004,
+ "step": 4487
+ },
+ {
+ "epoch": 62.3353711790393,
+ "grad_norm": 0.01790742017328739,
+ "learning_rate": 0.0006,
+ "loss": 4.289964199066162,
+ "step": 4488
+ },
+ {
+ "epoch": 62.34934497816594,
+ "grad_norm": 0.017266202718019485,
+ "learning_rate": 0.0006,
+ "loss": 4.415959358215332,
+ "step": 4489
+ },
+ {
+ "epoch": 62.36331877729258,
+ "grad_norm": 0.015976132825016975,
+ "learning_rate": 0.0006,
+ "loss": 4.358587265014648,
+ "step": 4490
+ },
+ {
+ "epoch": 62.377292576419215,
+ "grad_norm": 0.016873881220817566,
+ "learning_rate": 0.0006,
+ "loss": 4.354607105255127,
+ "step": 4491
+ },
+ {
+ "epoch": 62.391266375545854,
+ "grad_norm": 0.01476745493710041,
+ "learning_rate": 0.0006,
+ "loss": 4.384904384613037,
+ "step": 4492
+ },
+ {
+ "epoch": 62.40524017467249,
+ "grad_norm": 0.013810782693326473,
+ "learning_rate": 0.0006,
+ "loss": 4.459230899810791,
+ "step": 4493
+ },
+ {
+ "epoch": 62.419213973799124,
+ "grad_norm": 0.015150421299040318,
+ "learning_rate": 0.0006,
+ "loss": 4.369471549987793,
+ "step": 4494
+ },
+ {
+ "epoch": 62.43318777292576,
+ "grad_norm": 0.014539610594511032,
+ "learning_rate": 0.0006,
+ "loss": 4.3564982414245605,
+ "step": 4495
+ },
+ {
+ "epoch": 62.4471615720524,
+ "grad_norm": 0.015554017387330532,
+ "learning_rate": 0.0006,
+ "loss": 4.414379119873047,
+ "step": 4496
+ },
+ {
+ "epoch": 62.46113537117904,
+ "grad_norm": 0.016608357429504395,
+ "learning_rate": 0.0006,
+ "loss": 4.311585426330566,
+ "step": 4497
+ },
+ {
+ "epoch": 62.47510917030568,
+ "grad_norm": 0.01885881833732128,
+ "learning_rate": 0.0006,
+ "loss": 4.306629180908203,
+ "step": 4498
+ },
+ {
+ "epoch": 62.48908296943232,
+ "grad_norm": 0.022128663957118988,
+ "learning_rate": 0.0006,
+ "loss": 4.345483779907227,
+ "step": 4499
+ },
+ {
+ "epoch": 62.50305676855895,
+ "grad_norm": 0.023401018232107162,
+ "learning_rate": 0.0006,
+ "loss": 4.398204326629639,
+ "step": 4500
+ },
+ {
+ "epoch": 62.51703056768559,
+ "grad_norm": 0.022080009803175926,
+ "learning_rate": 0.0006,
+ "loss": 4.375027179718018,
+ "step": 4501
+ },
+ {
+ "epoch": 62.531004366812226,
+ "grad_norm": 0.020004864782094955,
+ "learning_rate": 0.0006,
+ "loss": 4.33494758605957,
+ "step": 4502
+ },
+ {
+ "epoch": 62.544978165938865,
+ "grad_norm": 0.01786700263619423,
+ "learning_rate": 0.0006,
+ "loss": 4.4428300857543945,
+ "step": 4503
+ },
+ {
+ "epoch": 62.5589519650655,
+ "grad_norm": 0.01885388419032097,
+ "learning_rate": 0.0006,
+ "loss": 4.4138946533203125,
+ "step": 4504
+ },
+ {
+ "epoch": 62.57292576419214,
+ "grad_norm": 0.019890939816832542,
+ "learning_rate": 0.0006,
+ "loss": 4.278632164001465,
+ "step": 4505
+ },
+ {
+ "epoch": 62.58689956331878,
+ "grad_norm": 0.02109031192958355,
+ "learning_rate": 0.0006,
+ "loss": 4.3756327629089355,
+ "step": 4506
+ },
+ {
+ "epoch": 62.60087336244541,
+ "grad_norm": 0.020813899114727974,
+ "learning_rate": 0.0006,
+ "loss": 4.410213470458984,
+ "step": 4507
+ },
+ {
+ "epoch": 62.61484716157205,
+ "grad_norm": 0.021655689924955368,
+ "learning_rate": 0.0006,
+ "loss": 4.30633020401001,
+ "step": 4508
+ },
+ {
+ "epoch": 62.62882096069869,
+ "grad_norm": 0.018155649304389954,
+ "learning_rate": 0.0006,
+ "loss": 4.407284736633301,
+ "step": 4509
+ },
+ {
+ "epoch": 62.64279475982533,
+ "grad_norm": 0.017765069380402565,
+ "learning_rate": 0.0006,
+ "loss": 4.321715354919434,
+ "step": 4510
+ },
+ {
+ "epoch": 62.65676855895197,
+ "grad_norm": 0.017724554985761642,
+ "learning_rate": 0.0006,
+ "loss": 4.493371963500977,
+ "step": 4511
+ },
+ {
+ "epoch": 62.670742358078606,
+ "grad_norm": 0.01492092851549387,
+ "learning_rate": 0.0006,
+ "loss": 4.443025588989258,
+ "step": 4512
+ },
+ {
+ "epoch": 62.68471615720524,
+ "grad_norm": 0.01615324057638645,
+ "learning_rate": 0.0006,
+ "loss": 4.3289594650268555,
+ "step": 4513
+ },
+ {
+ "epoch": 62.698689956331876,
+ "grad_norm": 0.016569366678595543,
+ "learning_rate": 0.0006,
+ "loss": 4.369720458984375,
+ "step": 4514
+ },
+ {
+ "epoch": 62.712663755458514,
+ "grad_norm": 0.016509901732206345,
+ "learning_rate": 0.0006,
+ "loss": 4.292316436767578,
+ "step": 4515
+ },
+ {
+ "epoch": 62.72663755458515,
+ "grad_norm": 0.014270003885030746,
+ "learning_rate": 0.0006,
+ "loss": 4.349854469299316,
+ "step": 4516
+ },
+ {
+ "epoch": 62.74061135371179,
+ "grad_norm": 0.014235340990126133,
+ "learning_rate": 0.0006,
+ "loss": 4.473727226257324,
+ "step": 4517
+ },
+ {
+ "epoch": 62.75458515283843,
+ "grad_norm": 0.015163225121796131,
+ "learning_rate": 0.0006,
+ "loss": 4.452878475189209,
+ "step": 4518
+ },
+ {
+ "epoch": 62.76855895196506,
+ "grad_norm": 0.015748055651783943,
+ "learning_rate": 0.0006,
+ "loss": 4.521087169647217,
+ "step": 4519
+ },
+ {
+ "epoch": 62.7825327510917,
+ "grad_norm": 0.01661108434200287,
+ "learning_rate": 0.0006,
+ "loss": 4.443885326385498,
+ "step": 4520
+ },
+ {
+ "epoch": 62.79650655021834,
+ "grad_norm": 0.01610301062464714,
+ "learning_rate": 0.0006,
+ "loss": 4.312990188598633,
+ "step": 4521
+ },
+ {
+ "epoch": 62.81048034934498,
+ "grad_norm": 0.017007583752274513,
+ "learning_rate": 0.0006,
+ "loss": 4.408937454223633,
+ "step": 4522
+ },
+ {
+ "epoch": 62.82445414847162,
+ "grad_norm": 0.01749461516737938,
+ "learning_rate": 0.0006,
+ "loss": 4.346066951751709,
+ "step": 4523
+ },
+ {
+ "epoch": 62.838427947598255,
+ "grad_norm": 0.018243493512272835,
+ "learning_rate": 0.0006,
+ "loss": 4.357658386230469,
+ "step": 4524
+ },
+ {
+ "epoch": 62.852401746724894,
+ "grad_norm": 0.017499985173344612,
+ "learning_rate": 0.0006,
+ "loss": 4.374061584472656,
+ "step": 4525
+ },
+ {
+ "epoch": 62.866375545851525,
+ "grad_norm": 0.014536325819790363,
+ "learning_rate": 0.0006,
+ "loss": 4.189321994781494,
+ "step": 4526
+ },
+ {
+ "epoch": 62.880349344978164,
+ "grad_norm": 0.014728891663253307,
+ "learning_rate": 0.0006,
+ "loss": 4.425910949707031,
+ "step": 4527
+ },
+ {
+ "epoch": 62.8943231441048,
+ "grad_norm": 0.015013772062957287,
+ "learning_rate": 0.0006,
+ "loss": 4.337319850921631,
+ "step": 4528
+ },
+ {
+ "epoch": 62.90829694323144,
+ "grad_norm": 0.01626027375459671,
+ "learning_rate": 0.0006,
+ "loss": 4.330339431762695,
+ "step": 4529
+ },
+ {
+ "epoch": 62.92227074235808,
+ "grad_norm": 0.014773444272577763,
+ "learning_rate": 0.0006,
+ "loss": 4.4188737869262695,
+ "step": 4530
+ },
+ {
+ "epoch": 62.93624454148472,
+ "grad_norm": 0.01422932744026184,
+ "learning_rate": 0.0006,
+ "loss": 4.355986595153809,
+ "step": 4531
+ },
+ {
+ "epoch": 62.95021834061135,
+ "grad_norm": 0.012170110829174519,
+ "learning_rate": 0.0006,
+ "loss": 4.405722618103027,
+ "step": 4532
+ },
+ {
+ "epoch": 62.96419213973799,
+ "grad_norm": 0.013233253732323647,
+ "learning_rate": 0.0006,
+ "loss": 4.377286911010742,
+ "step": 4533
+ },
+ {
+ "epoch": 62.97816593886463,
+ "grad_norm": 0.014497299678623676,
+ "learning_rate": 0.0006,
+ "loss": 4.508796691894531,
+ "step": 4534
+ },
+ {
+ "epoch": 62.992139737991266,
+ "grad_norm": 0.014718741178512573,
+ "learning_rate": 0.0006,
+ "loss": 4.370919704437256,
+ "step": 4535
+ },
+ {
+ "epoch": 63.0,
+ "grad_norm": 0.015833502635359764,
+ "learning_rate": 0.0006,
+ "loss": 4.258753776550293,
+ "step": 4536
+ },
+ {
+ "epoch": 63.0,
+ "eval_loss": 4.714163303375244,
+ "eval_runtime": 57.0269,
+ "eval_samples_per_second": 42.822,
+ "eval_steps_per_second": 1.35,
+ "step": 4536
+ },
+ {
+ "epoch": 63.01397379912664,
+ "grad_norm": 0.014277510344982147,
+ "learning_rate": 0.0006,
+ "loss": 4.3112640380859375,
+ "step": 4537
+ },
+ {
+ "epoch": 63.02794759825328,
+ "grad_norm": 0.015377935022115707,
+ "learning_rate": 0.0006,
+ "loss": 4.342782974243164,
+ "step": 4538
+ },
+ {
+ "epoch": 63.041921397379916,
+ "grad_norm": 0.015406244434416294,
+ "learning_rate": 0.0006,
+ "loss": 4.474056720733643,
+ "step": 4539
+ },
+ {
+ "epoch": 63.05589519650655,
+ "grad_norm": 0.018121499568223953,
+ "learning_rate": 0.0006,
+ "loss": 4.397963523864746,
+ "step": 4540
+ },
+ {
+ "epoch": 63.069868995633186,
+ "grad_norm": 0.022634636610746384,
+ "learning_rate": 0.0006,
+ "loss": 4.306631088256836,
+ "step": 4541
+ },
+ {
+ "epoch": 63.083842794759825,
+ "grad_norm": 0.026448970660567284,
+ "learning_rate": 0.0006,
+ "loss": 4.283175468444824,
+ "step": 4542
+ },
+ {
+ "epoch": 63.09781659388646,
+ "grad_norm": 0.024104537442326546,
+ "learning_rate": 0.0006,
+ "loss": 4.315826416015625,
+ "step": 4543
+ },
+ {
+ "epoch": 63.1117903930131,
+ "grad_norm": 0.019040480256080627,
+ "learning_rate": 0.0006,
+ "loss": 4.44057035446167,
+ "step": 4544
+ },
+ {
+ "epoch": 63.12576419213974,
+ "grad_norm": 0.02006000280380249,
+ "learning_rate": 0.0006,
+ "loss": 4.405603408813477,
+ "step": 4545
+ },
+ {
+ "epoch": 63.13973799126637,
+ "grad_norm": 0.019980071112513542,
+ "learning_rate": 0.0006,
+ "loss": 4.393726348876953,
+ "step": 4546
+ },
+ {
+ "epoch": 63.15371179039301,
+ "grad_norm": 0.01641865260899067,
+ "learning_rate": 0.0006,
+ "loss": 4.272546768188477,
+ "step": 4547
+ },
+ {
+ "epoch": 63.16768558951965,
+ "grad_norm": 0.01890353485941887,
+ "learning_rate": 0.0006,
+ "loss": 4.382566928863525,
+ "step": 4548
+ },
+ {
+ "epoch": 63.18165938864629,
+ "grad_norm": 0.018370937556028366,
+ "learning_rate": 0.0006,
+ "loss": 4.423953533172607,
+ "step": 4549
+ },
+ {
+ "epoch": 63.19563318777293,
+ "grad_norm": 0.014985294081270695,
+ "learning_rate": 0.0006,
+ "loss": 4.308627128601074,
+ "step": 4550
+ },
+ {
+ "epoch": 63.209606986899566,
+ "grad_norm": 0.017513327300548553,
+ "learning_rate": 0.0006,
+ "loss": 4.248819351196289,
+ "step": 4551
+ },
+ {
+ "epoch": 63.223580786026204,
+ "grad_norm": 0.019464161247015,
+ "learning_rate": 0.0006,
+ "loss": 4.359729766845703,
+ "step": 4552
+ },
+ {
+ "epoch": 63.237554585152836,
+ "grad_norm": 0.018678035587072372,
+ "learning_rate": 0.0006,
+ "loss": 4.3299665451049805,
+ "step": 4553
+ },
+ {
+ "epoch": 63.251528384279474,
+ "grad_norm": 0.017909523099660873,
+ "learning_rate": 0.0006,
+ "loss": 4.246147155761719,
+ "step": 4554
+ },
+ {
+ "epoch": 63.26550218340611,
+ "grad_norm": 0.01747293211519718,
+ "learning_rate": 0.0006,
+ "loss": 4.41165018081665,
+ "step": 4555
+ },
+ {
+ "epoch": 63.27947598253275,
+ "grad_norm": 0.018670853227376938,
+ "learning_rate": 0.0006,
+ "loss": 4.409883499145508,
+ "step": 4556
+ },
+ {
+ "epoch": 63.29344978165939,
+ "grad_norm": 0.0189529862254858,
+ "learning_rate": 0.0006,
+ "loss": 4.418654918670654,
+ "step": 4557
+ },
+ {
+ "epoch": 63.30742358078603,
+ "grad_norm": 0.018394414335489273,
+ "learning_rate": 0.0006,
+ "loss": 4.38355827331543,
+ "step": 4558
+ },
+ {
+ "epoch": 63.32139737991266,
+ "grad_norm": 0.018317820504307747,
+ "learning_rate": 0.0006,
+ "loss": 4.456843376159668,
+ "step": 4559
+ },
+ {
+ "epoch": 63.3353711790393,
+ "grad_norm": 0.019225597381591797,
+ "learning_rate": 0.0006,
+ "loss": 4.232264518737793,
+ "step": 4560
+ },
+ {
+ "epoch": 63.34934497816594,
+ "grad_norm": 0.016663681715726852,
+ "learning_rate": 0.0006,
+ "loss": 4.393517017364502,
+ "step": 4561
+ },
+ {
+ "epoch": 63.36331877729258,
+ "grad_norm": 0.014975365251302719,
+ "learning_rate": 0.0006,
+ "loss": 4.375183582305908,
+ "step": 4562
+ },
+ {
+ "epoch": 63.377292576419215,
+ "grad_norm": 0.01771732047200203,
+ "learning_rate": 0.0006,
+ "loss": 4.3073410987854,
+ "step": 4563
+ },
+ {
+ "epoch": 63.391266375545854,
+ "grad_norm": 0.01663549244403839,
+ "learning_rate": 0.0006,
+ "loss": 4.311003684997559,
+ "step": 4564
+ },
+ {
+ "epoch": 63.40524017467249,
+ "grad_norm": 0.016301177442073822,
+ "learning_rate": 0.0006,
+ "loss": 4.326807022094727,
+ "step": 4565
+ },
+ {
+ "epoch": 63.419213973799124,
+ "grad_norm": 0.016781188547611237,
+ "learning_rate": 0.0006,
+ "loss": 4.306562423706055,
+ "step": 4566
+ },
+ {
+ "epoch": 63.43318777292576,
+ "grad_norm": 0.01767745055258274,
+ "learning_rate": 0.0006,
+ "loss": 4.367082595825195,
+ "step": 4567
+ },
+ {
+ "epoch": 63.4471615720524,
+ "grad_norm": 0.016728565096855164,
+ "learning_rate": 0.0006,
+ "loss": 4.327400207519531,
+ "step": 4568
+ },
+ {
+ "epoch": 63.46113537117904,
+ "grad_norm": 0.016289403662085533,
+ "learning_rate": 0.0006,
+ "loss": 4.392712116241455,
+ "step": 4569
+ },
+ {
+ "epoch": 63.47510917030568,
+ "grad_norm": 0.016233716160058975,
+ "learning_rate": 0.0006,
+ "loss": 4.354372024536133,
+ "step": 4570
+ },
+ {
+ "epoch": 63.48908296943232,
+ "grad_norm": 0.016547974199056625,
+ "learning_rate": 0.0006,
+ "loss": 4.309050559997559,
+ "step": 4571
+ },
+ {
+ "epoch": 63.50305676855895,
+ "grad_norm": 0.018466509878635406,
+ "learning_rate": 0.0006,
+ "loss": 4.318345069885254,
+ "step": 4572
+ },
+ {
+ "epoch": 63.51703056768559,
+ "grad_norm": 0.017685720697045326,
+ "learning_rate": 0.0006,
+ "loss": 4.399726867675781,
+ "step": 4573
+ },
+ {
+ "epoch": 63.531004366812226,
+ "grad_norm": 0.015801234170794487,
+ "learning_rate": 0.0006,
+ "loss": 4.328495502471924,
+ "step": 4574
+ },
+ {
+ "epoch": 63.544978165938865,
+ "grad_norm": 0.0188873540610075,
+ "learning_rate": 0.0006,
+ "loss": 4.253244400024414,
+ "step": 4575
+ },
+ {
+ "epoch": 63.5589519650655,
+ "grad_norm": 0.01868271268904209,
+ "learning_rate": 0.0006,
+ "loss": 4.463746070861816,
+ "step": 4576
+ },
+ {
+ "epoch": 63.57292576419214,
+ "grad_norm": 0.01844949834048748,
+ "learning_rate": 0.0006,
+ "loss": 4.400300025939941,
+ "step": 4577
+ },
+ {
+ "epoch": 63.58689956331878,
+ "grad_norm": 0.015256480313837528,
+ "learning_rate": 0.0006,
+ "loss": 4.445844650268555,
+ "step": 4578
+ },
+ {
+ "epoch": 63.60087336244541,
+ "grad_norm": 0.015305760316550732,
+ "learning_rate": 0.0006,
+ "loss": 4.317327499389648,
+ "step": 4579
+ },
+ {
+ "epoch": 63.61484716157205,
+ "grad_norm": 0.015200769528746605,
+ "learning_rate": 0.0006,
+ "loss": 4.25112247467041,
+ "step": 4580
+ },
+ {
+ "epoch": 63.62882096069869,
+ "grad_norm": 0.015919538214802742,
+ "learning_rate": 0.0006,
+ "loss": 4.277937889099121,
+ "step": 4581
+ },
+ {
+ "epoch": 63.64279475982533,
+ "grad_norm": 0.01572665013372898,
+ "learning_rate": 0.0006,
+ "loss": 4.302237510681152,
+ "step": 4582
+ },
+ {
+ "epoch": 63.65676855895197,
+ "grad_norm": 0.016428731381893158,
+ "learning_rate": 0.0006,
+ "loss": 4.36762809753418,
+ "step": 4583
+ },
+ {
+ "epoch": 63.670742358078606,
+ "grad_norm": 0.01706082746386528,
+ "learning_rate": 0.0006,
+ "loss": 4.236526966094971,
+ "step": 4584
+ },
+ {
+ "epoch": 63.68471615720524,
+ "grad_norm": 0.017850136384367943,
+ "learning_rate": 0.0006,
+ "loss": 4.264129638671875,
+ "step": 4585
+ },
+ {
+ "epoch": 63.698689956331876,
+ "grad_norm": 0.016079438850283623,
+ "learning_rate": 0.0006,
+ "loss": 4.449799060821533,
+ "step": 4586
+ },
+ {
+ "epoch": 63.712663755458514,
+ "grad_norm": 0.014996383339166641,
+ "learning_rate": 0.0006,
+ "loss": 4.404574394226074,
+ "step": 4587
+ },
+ {
+ "epoch": 63.72663755458515,
+ "grad_norm": 0.015910431742668152,
+ "learning_rate": 0.0006,
+ "loss": 4.348101615905762,
+ "step": 4588
+ },
+ {
+ "epoch": 63.74061135371179,
+ "grad_norm": 0.01630021072924137,
+ "learning_rate": 0.0006,
+ "loss": 4.3193817138671875,
+ "step": 4589
+ },
+ {
+ "epoch": 63.75458515283843,
+ "grad_norm": 0.014571773819625378,
+ "learning_rate": 0.0006,
+ "loss": 4.374402046203613,
+ "step": 4590
+ },
+ {
+ "epoch": 63.76855895196506,
+ "grad_norm": 0.01662345975637436,
+ "learning_rate": 0.0006,
+ "loss": 4.329075813293457,
+ "step": 4591
+ },
+ {
+ "epoch": 63.7825327510917,
+ "grad_norm": 0.016604198142886162,
+ "learning_rate": 0.0006,
+ "loss": 4.436126232147217,
+ "step": 4592
+ },
+ {
+ "epoch": 63.79650655021834,
+ "grad_norm": 0.016578884795308113,
+ "learning_rate": 0.0006,
+ "loss": 4.361468315124512,
+ "step": 4593
+ },
+ {
+ "epoch": 63.81048034934498,
+ "grad_norm": 0.016339825466275215,
+ "learning_rate": 0.0006,
+ "loss": 4.449892044067383,
+ "step": 4594
+ },
+ {
+ "epoch": 63.82445414847162,
+ "grad_norm": 0.016805455088615417,
+ "learning_rate": 0.0006,
+ "loss": 4.415866851806641,
+ "step": 4595
+ },
+ {
+ "epoch": 63.838427947598255,
+ "grad_norm": 0.017366213724017143,
+ "learning_rate": 0.0006,
+ "loss": 4.361730575561523,
+ "step": 4596
+ },
+ {
+ "epoch": 63.852401746724894,
+ "grad_norm": 0.015599003992974758,
+ "learning_rate": 0.0006,
+ "loss": 4.363637924194336,
+ "step": 4597
+ },
+ {
+ "epoch": 63.866375545851525,
+ "grad_norm": 0.015674632042646408,
+ "learning_rate": 0.0006,
+ "loss": 4.332045078277588,
+ "step": 4598
+ },
+ {
+ "epoch": 63.880349344978164,
+ "grad_norm": 0.015350443311035633,
+ "learning_rate": 0.0006,
+ "loss": 4.320063591003418,
+ "step": 4599
+ },
+ {
+ "epoch": 63.8943231441048,
+ "grad_norm": 0.015388146974146366,
+ "learning_rate": 0.0006,
+ "loss": 4.398199081420898,
+ "step": 4600
+ },
+ {
+ "epoch": 63.90829694323144,
+ "grad_norm": 0.01450140681117773,
+ "learning_rate": 0.0006,
+ "loss": 4.244076728820801,
+ "step": 4601
+ },
+ {
+ "epoch": 63.92227074235808,
+ "grad_norm": 0.01637491025030613,
+ "learning_rate": 0.0006,
+ "loss": 4.213549613952637,
+ "step": 4602
+ },
+ {
+ "epoch": 63.93624454148472,
+ "grad_norm": 0.0159814041107893,
+ "learning_rate": 0.0006,
+ "loss": 4.421775817871094,
+ "step": 4603
+ },
+ {
+ "epoch": 63.95021834061135,
+ "grad_norm": 0.017386827617883682,
+ "learning_rate": 0.0006,
+ "loss": 4.340023994445801,
+ "step": 4604
+ },
+ {
+ "epoch": 63.96419213973799,
+ "grad_norm": 0.019679656252264977,
+ "learning_rate": 0.0006,
+ "loss": 4.352534294128418,
+ "step": 4605
+ },
+ {
+ "epoch": 63.97816593886463,
+ "grad_norm": 0.022391516715288162,
+ "learning_rate": 0.0006,
+ "loss": 4.3538312911987305,
+ "step": 4606
+ },
+ {
+ "epoch": 63.992139737991266,
+ "grad_norm": 0.023281006142497063,
+ "learning_rate": 0.0006,
+ "loss": 4.432089328765869,
+ "step": 4607
+ },
+ {
+ "epoch": 64.0,
+ "grad_norm": 0.019916288554668427,
+ "learning_rate": 0.0006,
+ "loss": 4.466937065124512,
+ "step": 4608
+ },
+ {
+ "epoch": 64.0,
+ "eval_loss": 4.659008502960205,
+ "eval_runtime": 56.0019,
+ "eval_samples_per_second": 43.606,
+ "eval_steps_per_second": 1.375,
+ "step": 4608
+ },
+ {
+ "epoch": 64.01397379912663,
+ "grad_norm": 0.017351187765598297,
+ "learning_rate": 0.0006,
+ "loss": 4.294129371643066,
+ "step": 4609
+ },
+ {
+ "epoch": 64.02794759825328,
+ "grad_norm": 0.02405865117907524,
+ "learning_rate": 0.0006,
+ "loss": 4.266717910766602,
+ "step": 4610
+ },
+ {
+ "epoch": 64.04192139737991,
+ "grad_norm": 0.022158386185765266,
+ "learning_rate": 0.0006,
+ "loss": 4.395932197570801,
+ "step": 4611
+ },
+ {
+ "epoch": 64.05589519650655,
+ "grad_norm": 0.019238505512475967,
+ "learning_rate": 0.0006,
+ "loss": 4.266704082489014,
+ "step": 4612
+ },
+ {
+ "epoch": 64.06986899563319,
+ "grad_norm": 0.021670425310730934,
+ "learning_rate": 0.0006,
+ "loss": 4.3160552978515625,
+ "step": 4613
+ },
+ {
+ "epoch": 64.08384279475983,
+ "grad_norm": 0.02076641097664833,
+ "learning_rate": 0.0006,
+ "loss": 4.415626525878906,
+ "step": 4614
+ },
+ {
+ "epoch": 64.09781659388646,
+ "grad_norm": 0.018223078921437263,
+ "learning_rate": 0.0006,
+ "loss": 4.292387008666992,
+ "step": 4615
+ },
+ {
+ "epoch": 64.1117903930131,
+ "grad_norm": 0.017401937395334244,
+ "learning_rate": 0.0006,
+ "loss": 4.424524307250977,
+ "step": 4616
+ },
+ {
+ "epoch": 64.12576419213974,
+ "grad_norm": 0.01663743332028389,
+ "learning_rate": 0.0006,
+ "loss": 4.350032806396484,
+ "step": 4617
+ },
+ {
+ "epoch": 64.13973799126637,
+ "grad_norm": 0.018716629594564438,
+ "learning_rate": 0.0006,
+ "loss": 4.309157371520996,
+ "step": 4618
+ },
+ {
+ "epoch": 64.15371179039302,
+ "grad_norm": 0.019569825381040573,
+ "learning_rate": 0.0006,
+ "loss": 4.358292579650879,
+ "step": 4619
+ },
+ {
+ "epoch": 64.16768558951965,
+ "grad_norm": 0.018994422629475594,
+ "learning_rate": 0.0006,
+ "loss": 4.3338823318481445,
+ "step": 4620
+ },
+ {
+ "epoch": 64.18165938864628,
+ "grad_norm": 0.017690688371658325,
+ "learning_rate": 0.0006,
+ "loss": 4.414468765258789,
+ "step": 4621
+ },
+ {
+ "epoch": 64.19563318777293,
+ "grad_norm": 0.0170609001070261,
+ "learning_rate": 0.0006,
+ "loss": 4.4337568283081055,
+ "step": 4622
+ },
+ {
+ "epoch": 64.20960698689956,
+ "grad_norm": 0.016984131187200546,
+ "learning_rate": 0.0006,
+ "loss": 4.304839134216309,
+ "step": 4623
+ },
+ {
+ "epoch": 64.2235807860262,
+ "grad_norm": 0.01619867794215679,
+ "learning_rate": 0.0006,
+ "loss": 4.389623165130615,
+ "step": 4624
+ },
+ {
+ "epoch": 64.23755458515284,
+ "grad_norm": 0.015428110957145691,
+ "learning_rate": 0.0006,
+ "loss": 4.277515888214111,
+ "step": 4625
+ },
+ {
+ "epoch": 64.25152838427948,
+ "grad_norm": 0.015909532085061073,
+ "learning_rate": 0.0006,
+ "loss": 4.266029357910156,
+ "step": 4626
+ },
+ {
+ "epoch": 64.26550218340611,
+ "grad_norm": 0.014296560548245907,
+ "learning_rate": 0.0006,
+ "loss": 4.286067485809326,
+ "step": 4627
+ },
+ {
+ "epoch": 64.27947598253274,
+ "grad_norm": 0.014065971598029137,
+ "learning_rate": 0.0006,
+ "loss": 4.353528022766113,
+ "step": 4628
+ },
+ {
+ "epoch": 64.29344978165939,
+ "grad_norm": 0.014386293478310108,
+ "learning_rate": 0.0006,
+ "loss": 4.318715572357178,
+ "step": 4629
+ },
+ {
+ "epoch": 64.30742358078602,
+ "grad_norm": 0.014001123607158661,
+ "learning_rate": 0.0006,
+ "loss": 4.362497329711914,
+ "step": 4630
+ },
+ {
+ "epoch": 64.32139737991267,
+ "grad_norm": 0.013408979400992393,
+ "learning_rate": 0.0006,
+ "loss": 4.417397499084473,
+ "step": 4631
+ },
+ {
+ "epoch": 64.3353711790393,
+ "grad_norm": 0.014960093423724174,
+ "learning_rate": 0.0006,
+ "loss": 4.359710693359375,
+ "step": 4632
+ },
+ {
+ "epoch": 64.34934497816595,
+ "grad_norm": 0.014658777043223381,
+ "learning_rate": 0.0006,
+ "loss": 4.318822383880615,
+ "step": 4633
+ },
+ {
+ "epoch": 64.36331877729258,
+ "grad_norm": 0.013766744174063206,
+ "learning_rate": 0.0006,
+ "loss": 4.265679836273193,
+ "step": 4634
+ },
+ {
+ "epoch": 64.37729257641921,
+ "grad_norm": 0.013821255415678024,
+ "learning_rate": 0.0006,
+ "loss": 4.3452911376953125,
+ "step": 4635
+ },
+ {
+ "epoch": 64.39126637554585,
+ "grad_norm": 0.013541066087782383,
+ "learning_rate": 0.0006,
+ "loss": 4.429254531860352,
+ "step": 4636
+ },
+ {
+ "epoch": 64.40524017467249,
+ "grad_norm": 0.013985645025968552,
+ "learning_rate": 0.0006,
+ "loss": 4.304452896118164,
+ "step": 4637
+ },
+ {
+ "epoch": 64.41921397379913,
+ "grad_norm": 0.015079808421432972,
+ "learning_rate": 0.0006,
+ "loss": 4.296828746795654,
+ "step": 4638
+ },
+ {
+ "epoch": 64.43318777292576,
+ "grad_norm": 0.014896944165229797,
+ "learning_rate": 0.0006,
+ "loss": 4.299156665802002,
+ "step": 4639
+ },
+ {
+ "epoch": 64.44716157205241,
+ "grad_norm": 0.013770041987299919,
+ "learning_rate": 0.0006,
+ "loss": 4.351439952850342,
+ "step": 4640
+ },
+ {
+ "epoch": 64.46113537117904,
+ "grad_norm": 0.01531178504228592,
+ "learning_rate": 0.0006,
+ "loss": 4.373715877532959,
+ "step": 4641
+ },
+ {
+ "epoch": 64.47510917030567,
+ "grad_norm": 0.014024450443685055,
+ "learning_rate": 0.0006,
+ "loss": 4.368053436279297,
+ "step": 4642
+ },
+ {
+ "epoch": 64.48908296943232,
+ "grad_norm": 0.013864024542272091,
+ "learning_rate": 0.0006,
+ "loss": 4.378564834594727,
+ "step": 4643
+ },
+ {
+ "epoch": 64.50305676855895,
+ "grad_norm": 0.013925445266067982,
+ "learning_rate": 0.0006,
+ "loss": 4.261663436889648,
+ "step": 4644
+ },
+ {
+ "epoch": 64.5170305676856,
+ "grad_norm": 0.015011338517069817,
+ "learning_rate": 0.0006,
+ "loss": 4.337285995483398,
+ "step": 4645
+ },
+ {
+ "epoch": 64.53100436681223,
+ "grad_norm": 0.016941986978054047,
+ "learning_rate": 0.0006,
+ "loss": 4.336877822875977,
+ "step": 4646
+ },
+ {
+ "epoch": 64.54497816593886,
+ "grad_norm": 0.01719909906387329,
+ "learning_rate": 0.0006,
+ "loss": 4.149813175201416,
+ "step": 4647
+ },
+ {
+ "epoch": 64.5589519650655,
+ "grad_norm": 0.01608402095735073,
+ "learning_rate": 0.0006,
+ "loss": 4.191280364990234,
+ "step": 4648
+ },
+ {
+ "epoch": 64.57292576419214,
+ "grad_norm": 0.015284820459783077,
+ "learning_rate": 0.0006,
+ "loss": 4.35953426361084,
+ "step": 4649
+ },
+ {
+ "epoch": 64.58689956331878,
+ "grad_norm": 0.014855682849884033,
+ "learning_rate": 0.0006,
+ "loss": 4.308401107788086,
+ "step": 4650
+ },
+ {
+ "epoch": 64.60087336244541,
+ "grad_norm": 0.017553111538290977,
+ "learning_rate": 0.0006,
+ "loss": 4.279720306396484,
+ "step": 4651
+ },
+ {
+ "epoch": 64.61484716157206,
+ "grad_norm": 0.019506998360157013,
+ "learning_rate": 0.0006,
+ "loss": 4.319524765014648,
+ "step": 4652
+ },
+ {
+ "epoch": 64.62882096069869,
+ "grad_norm": 0.020293425768613815,
+ "learning_rate": 0.0006,
+ "loss": 4.340523719787598,
+ "step": 4653
+ },
+ {
+ "epoch": 64.64279475982532,
+ "grad_norm": 0.01664556749165058,
+ "learning_rate": 0.0006,
+ "loss": 4.296785831451416,
+ "step": 4654
+ },
+ {
+ "epoch": 64.65676855895197,
+ "grad_norm": 0.012592585757374763,
+ "learning_rate": 0.0006,
+ "loss": 4.464595317840576,
+ "step": 4655
+ },
+ {
+ "epoch": 64.6707423580786,
+ "grad_norm": 0.01549589168280363,
+ "learning_rate": 0.0006,
+ "loss": 4.275053024291992,
+ "step": 4656
+ },
+ {
+ "epoch": 64.68471615720524,
+ "grad_norm": 0.01705172471702099,
+ "learning_rate": 0.0006,
+ "loss": 4.388431549072266,
+ "step": 4657
+ },
+ {
+ "epoch": 64.69868995633188,
+ "grad_norm": 0.015438582748174667,
+ "learning_rate": 0.0006,
+ "loss": 4.312070846557617,
+ "step": 4658
+ },
+ {
+ "epoch": 64.71266375545852,
+ "grad_norm": 0.014448083005845547,
+ "learning_rate": 0.0006,
+ "loss": 4.237599849700928,
+ "step": 4659
+ },
+ {
+ "epoch": 64.72663755458515,
+ "grad_norm": 0.015267071314156055,
+ "learning_rate": 0.0006,
+ "loss": 4.29893684387207,
+ "step": 4660
+ },
+ {
+ "epoch": 64.74061135371178,
+ "grad_norm": 0.016615847125649452,
+ "learning_rate": 0.0006,
+ "loss": 4.432273864746094,
+ "step": 4661
+ },
+ {
+ "epoch": 64.75458515283843,
+ "grad_norm": 0.016584362834692,
+ "learning_rate": 0.0006,
+ "loss": 4.202266216278076,
+ "step": 4662
+ },
+ {
+ "epoch": 64.76855895196506,
+ "grad_norm": 0.018313994631171227,
+ "learning_rate": 0.0006,
+ "loss": 4.367966651916504,
+ "step": 4663
+ },
+ {
+ "epoch": 64.78253275109171,
+ "grad_norm": 0.017946481704711914,
+ "learning_rate": 0.0006,
+ "loss": 4.205172538757324,
+ "step": 4664
+ },
+ {
+ "epoch": 64.79650655021834,
+ "grad_norm": 0.017744354903697968,
+ "learning_rate": 0.0006,
+ "loss": 4.295661926269531,
+ "step": 4665
+ },
+ {
+ "epoch": 64.81048034934499,
+ "grad_norm": 0.016103824600577354,
+ "learning_rate": 0.0006,
+ "loss": 4.276543617248535,
+ "step": 4666
+ },
+ {
+ "epoch": 64.82445414847162,
+ "grad_norm": 0.014590012840926647,
+ "learning_rate": 0.0006,
+ "loss": 4.414247035980225,
+ "step": 4667
+ },
+ {
+ "epoch": 64.83842794759825,
+ "grad_norm": 0.01782386749982834,
+ "learning_rate": 0.0006,
+ "loss": 4.203220367431641,
+ "step": 4668
+ },
+ {
+ "epoch": 64.8524017467249,
+ "grad_norm": 0.022122733294963837,
+ "learning_rate": 0.0006,
+ "loss": 4.38078498840332,
+ "step": 4669
+ },
+ {
+ "epoch": 64.86637554585153,
+ "grad_norm": 0.0250293780118227,
+ "learning_rate": 0.0006,
+ "loss": 4.3310136795043945,
+ "step": 4670
+ },
+ {
+ "epoch": 64.88034934497817,
+ "grad_norm": 0.022438062354922295,
+ "learning_rate": 0.0006,
+ "loss": 4.428631782531738,
+ "step": 4671
+ },
+ {
+ "epoch": 64.8943231441048,
+ "grad_norm": 0.018518192693591118,
+ "learning_rate": 0.0006,
+ "loss": 4.428250312805176,
+ "step": 4672
+ },
+ {
+ "epoch": 64.90829694323143,
+ "grad_norm": 0.018105225637555122,
+ "learning_rate": 0.0006,
+ "loss": 4.336209297180176,
+ "step": 4673
+ },
+ {
+ "epoch": 64.92227074235808,
+ "grad_norm": 0.017564399167895317,
+ "learning_rate": 0.0006,
+ "loss": 4.416011810302734,
+ "step": 4674
+ },
+ {
+ "epoch": 64.93624454148471,
+ "grad_norm": 0.018747031688690186,
+ "learning_rate": 0.0006,
+ "loss": 4.419938087463379,
+ "step": 4675
+ },
+ {
+ "epoch": 64.95021834061136,
+ "grad_norm": 0.018686484545469284,
+ "learning_rate": 0.0006,
+ "loss": 4.347185134887695,
+ "step": 4676
+ },
+ {
+ "epoch": 64.96419213973799,
+ "grad_norm": 0.019329898059368134,
+ "learning_rate": 0.0006,
+ "loss": 4.343749046325684,
+ "step": 4677
+ },
+ {
+ "epoch": 64.97816593886463,
+ "grad_norm": 0.018517782911658287,
+ "learning_rate": 0.0006,
+ "loss": 4.264604568481445,
+ "step": 4678
+ },
+ {
+ "epoch": 64.99213973799127,
+ "grad_norm": 0.016579125076532364,
+ "learning_rate": 0.0006,
+ "loss": 4.303423881530762,
+ "step": 4679
+ },
+ {
+ "epoch": 65.0,
+ "grad_norm": 0.01617567613720894,
+ "learning_rate": 0.0006,
+ "loss": 4.326556205749512,
+ "step": 4680
+ },
+ {
+ "epoch": 65.0,
+ "eval_loss": 4.655026912689209,
+ "eval_runtime": 57.1299,
+ "eval_samples_per_second": 42.745,
+ "eval_steps_per_second": 1.348,
+ "step": 4680
+ },
+ {
+ "epoch": 65.01397379912663,
+ "grad_norm": 0.01607249490916729,
+ "learning_rate": 0.0006,
+ "loss": 4.276607036590576,
+ "step": 4681
+ },
+ {
+ "epoch": 65.02794759825328,
+ "grad_norm": 0.016779497265815735,
+ "learning_rate": 0.0006,
+ "loss": 4.412464618682861,
+ "step": 4682
+ },
+ {
+ "epoch": 65.04192139737991,
+ "grad_norm": 0.018161555752158165,
+ "learning_rate": 0.0006,
+ "loss": 4.199329376220703,
+ "step": 4683
+ },
+ {
+ "epoch": 65.05589519650655,
+ "grad_norm": 0.017508309334516525,
+ "learning_rate": 0.0006,
+ "loss": 4.217567443847656,
+ "step": 4684
+ },
+ {
+ "epoch": 65.06986899563319,
+ "grad_norm": 0.015716979280114174,
+ "learning_rate": 0.0006,
+ "loss": 4.322023391723633,
+ "step": 4685
+ },
+ {
+ "epoch": 65.08384279475983,
+ "grad_norm": 0.015113635919988155,
+ "learning_rate": 0.0006,
+ "loss": 4.333515167236328,
+ "step": 4686
+ },
+ {
+ "epoch": 65.09781659388646,
+ "grad_norm": 0.01437560748308897,
+ "learning_rate": 0.0006,
+ "loss": 4.295609474182129,
+ "step": 4687
+ },
+ {
+ "epoch": 65.1117903930131,
+ "grad_norm": 0.013074862770736217,
+ "learning_rate": 0.0006,
+ "loss": 4.314120769500732,
+ "step": 4688
+ },
+ {
+ "epoch": 65.12576419213974,
+ "grad_norm": 0.012908284552395344,
+ "learning_rate": 0.0006,
+ "loss": 4.216890811920166,
+ "step": 4689
+ },
+ {
+ "epoch": 65.13973799126637,
+ "grad_norm": 0.013001754879951477,
+ "learning_rate": 0.0006,
+ "loss": 4.287717342376709,
+ "step": 4690
+ },
+ {
+ "epoch": 65.15371179039302,
+ "grad_norm": 0.013694263994693756,
+ "learning_rate": 0.0006,
+ "loss": 4.333837985992432,
+ "step": 4691
+ },
+ {
+ "epoch": 65.16768558951965,
+ "grad_norm": 0.014461833983659744,
+ "learning_rate": 0.0006,
+ "loss": 4.323545455932617,
+ "step": 4692
+ },
+ {
+ "epoch": 65.18165938864628,
+ "grad_norm": 0.013670675456523895,
+ "learning_rate": 0.0006,
+ "loss": 4.3313493728637695,
+ "step": 4693
+ },
+ {
+ "epoch": 65.19563318777293,
+ "grad_norm": 0.013094350695610046,
+ "learning_rate": 0.0006,
+ "loss": 4.23858118057251,
+ "step": 4694
+ },
+ {
+ "epoch": 65.20960698689956,
+ "grad_norm": 0.013996944762766361,
+ "learning_rate": 0.0006,
+ "loss": 4.428705215454102,
+ "step": 4695
+ },
+ {
+ "epoch": 65.2235807860262,
+ "grad_norm": 0.013295048847794533,
+ "learning_rate": 0.0006,
+ "loss": 4.238950252532959,
+ "step": 4696
+ },
+ {
+ "epoch": 65.23755458515284,
+ "grad_norm": 0.013380616903305054,
+ "learning_rate": 0.0006,
+ "loss": 4.273185729980469,
+ "step": 4697
+ },
+ {
+ "epoch": 65.25152838427948,
+ "grad_norm": 0.015062890015542507,
+ "learning_rate": 0.0006,
+ "loss": 4.29472017288208,
+ "step": 4698
+ },
+ {
+ "epoch": 65.26550218340611,
+ "grad_norm": 0.017317263409495354,
+ "learning_rate": 0.0006,
+ "loss": 4.337954998016357,
+ "step": 4699
+ },
+ {
+ "epoch": 65.27947598253274,
+ "grad_norm": 0.017698992043733597,
+ "learning_rate": 0.0006,
+ "loss": 4.336437225341797,
+ "step": 4700
+ },
+ {
+ "epoch": 65.29344978165939,
+ "grad_norm": 0.01628575474023819,
+ "learning_rate": 0.0006,
+ "loss": 4.302372455596924,
+ "step": 4701
+ },
+ {
+ "epoch": 65.30742358078602,
+ "grad_norm": 0.014531854540109634,
+ "learning_rate": 0.0006,
+ "loss": 4.256450653076172,
+ "step": 4702
+ },
+ {
+ "epoch": 65.32139737991267,
+ "grad_norm": 0.014549105428159237,
+ "learning_rate": 0.0006,
+ "loss": 4.302240371704102,
+ "step": 4703
+ },
+ {
+ "epoch": 65.3353711790393,
+ "grad_norm": 0.014881663955748081,
+ "learning_rate": 0.0006,
+ "loss": 4.3316121101379395,
+ "step": 4704
+ },
+ {
+ "epoch": 65.34934497816595,
+ "grad_norm": 0.014513437636196613,
+ "learning_rate": 0.0006,
+ "loss": 4.441212177276611,
+ "step": 4705
+ },
+ {
+ "epoch": 65.36331877729258,
+ "grad_norm": 0.014122087508440018,
+ "learning_rate": 0.0006,
+ "loss": 4.343296527862549,
+ "step": 4706
+ },
+ {
+ "epoch": 65.37729257641921,
+ "grad_norm": 0.014817348681390285,
+ "learning_rate": 0.0006,
+ "loss": 4.195674896240234,
+ "step": 4707
+ },
+ {
+ "epoch": 65.39126637554585,
+ "grad_norm": 0.01787056028842926,
+ "learning_rate": 0.0006,
+ "loss": 4.3507914543151855,
+ "step": 4708
+ },
+ {
+ "epoch": 65.40524017467249,
+ "grad_norm": 0.015295136719942093,
+ "learning_rate": 0.0006,
+ "loss": 4.234372138977051,
+ "step": 4709
+ },
+ {
+ "epoch": 65.41921397379913,
+ "grad_norm": 0.014324234798550606,
+ "learning_rate": 0.0006,
+ "loss": 4.2907867431640625,
+ "step": 4710
+ },
+ {
+ "epoch": 65.43318777292576,
+ "grad_norm": 0.016090670600533485,
+ "learning_rate": 0.0006,
+ "loss": 4.177244186401367,
+ "step": 4711
+ },
+ {
+ "epoch": 65.44716157205241,
+ "grad_norm": 0.0174604132771492,
+ "learning_rate": 0.0006,
+ "loss": 4.335577964782715,
+ "step": 4712
+ },
+ {
+ "epoch": 65.46113537117904,
+ "grad_norm": 0.016731899231672287,
+ "learning_rate": 0.0006,
+ "loss": 4.30197811126709,
+ "step": 4713
+ },
+ {
+ "epoch": 65.47510917030567,
+ "grad_norm": 0.017423735931515694,
+ "learning_rate": 0.0006,
+ "loss": 4.3528852462768555,
+ "step": 4714
+ },
+ {
+ "epoch": 65.48908296943232,
+ "grad_norm": 0.01780976541340351,
+ "learning_rate": 0.0006,
+ "loss": 4.185036659240723,
+ "step": 4715
+ },
+ {
+ "epoch": 65.50305676855895,
+ "grad_norm": 0.019977200776338577,
+ "learning_rate": 0.0006,
+ "loss": 4.328766345977783,
+ "step": 4716
+ },
+ {
+ "epoch": 65.5170305676856,
+ "grad_norm": 0.020284287631511688,
+ "learning_rate": 0.0006,
+ "loss": 4.285494804382324,
+ "step": 4717
+ },
+ {
+ "epoch": 65.53100436681223,
+ "grad_norm": 0.019168002530932426,
+ "learning_rate": 0.0006,
+ "loss": 4.328560829162598,
+ "step": 4718
+ },
+ {
+ "epoch": 65.54497816593886,
+ "grad_norm": 0.019217776134610176,
+ "learning_rate": 0.0006,
+ "loss": 4.189998626708984,
+ "step": 4719
+ },
+ {
+ "epoch": 65.5589519650655,
+ "grad_norm": 0.018960488960146904,
+ "learning_rate": 0.0006,
+ "loss": 4.4995012283325195,
+ "step": 4720
+ },
+ {
+ "epoch": 65.57292576419214,
+ "grad_norm": 0.01941288262605667,
+ "learning_rate": 0.0006,
+ "loss": 4.26570987701416,
+ "step": 4721
+ },
+ {
+ "epoch": 65.58689956331878,
+ "grad_norm": 0.020263448357582092,
+ "learning_rate": 0.0006,
+ "loss": 4.411645889282227,
+ "step": 4722
+ },
+ {
+ "epoch": 65.60087336244541,
+ "grad_norm": 0.02144799567759037,
+ "learning_rate": 0.0006,
+ "loss": 4.163675308227539,
+ "step": 4723
+ },
+ {
+ "epoch": 65.61484716157206,
+ "grad_norm": 0.019790446385741234,
+ "learning_rate": 0.0006,
+ "loss": 4.2361063957214355,
+ "step": 4724
+ },
+ {
+ "epoch": 65.62882096069869,
+ "grad_norm": 0.01829805225133896,
+ "learning_rate": 0.0006,
+ "loss": 4.449810028076172,
+ "step": 4725
+ },
+ {
+ "epoch": 65.64279475982532,
+ "grad_norm": 0.01571105420589447,
+ "learning_rate": 0.0006,
+ "loss": 4.281406879425049,
+ "step": 4726
+ },
+ {
+ "epoch": 65.65676855895197,
+ "grad_norm": 0.016477147117257118,
+ "learning_rate": 0.0006,
+ "loss": 4.293408393859863,
+ "step": 4727
+ },
+ {
+ "epoch": 65.6707423580786,
+ "grad_norm": 0.016766032204031944,
+ "learning_rate": 0.0006,
+ "loss": 4.378252983093262,
+ "step": 4728
+ },
+ {
+ "epoch": 65.68471615720524,
+ "grad_norm": 0.0159933939576149,
+ "learning_rate": 0.0006,
+ "loss": 4.3065667152404785,
+ "step": 4729
+ },
+ {
+ "epoch": 65.69868995633188,
+ "grad_norm": 0.01658209227025509,
+ "learning_rate": 0.0006,
+ "loss": 4.174026966094971,
+ "step": 4730
+ },
+ {
+ "epoch": 65.71266375545852,
+ "grad_norm": 0.015081380493938923,
+ "learning_rate": 0.0006,
+ "loss": 4.360222339630127,
+ "step": 4731
+ },
+ {
+ "epoch": 65.72663755458515,
+ "grad_norm": 0.015110890381038189,
+ "learning_rate": 0.0006,
+ "loss": 4.218810558319092,
+ "step": 4732
+ },
+ {
+ "epoch": 65.74061135371178,
+ "grad_norm": 0.015300953760743141,
+ "learning_rate": 0.0006,
+ "loss": 4.331979274749756,
+ "step": 4733
+ },
+ {
+ "epoch": 65.75458515283843,
+ "grad_norm": 0.01715504191815853,
+ "learning_rate": 0.0006,
+ "loss": 4.349112033843994,
+ "step": 4734
+ },
+ {
+ "epoch": 65.76855895196506,
+ "grad_norm": 0.015870366245508194,
+ "learning_rate": 0.0006,
+ "loss": 4.329850196838379,
+ "step": 4735
+ },
+ {
+ "epoch": 65.78253275109171,
+ "grad_norm": 0.01541937980800867,
+ "learning_rate": 0.0006,
+ "loss": 4.3400163650512695,
+ "step": 4736
+ },
+ {
+ "epoch": 65.79650655021834,
+ "grad_norm": 0.01738942787051201,
+ "learning_rate": 0.0006,
+ "loss": 4.220798492431641,
+ "step": 4737
+ },
+ {
+ "epoch": 65.81048034934499,
+ "grad_norm": 0.01685990020632744,
+ "learning_rate": 0.0006,
+ "loss": 4.177238464355469,
+ "step": 4738
+ },
+ {
+ "epoch": 65.82445414847162,
+ "grad_norm": 0.016183465719223022,
+ "learning_rate": 0.0006,
+ "loss": 4.327216625213623,
+ "step": 4739
+ },
+ {
+ "epoch": 65.83842794759825,
+ "grad_norm": 0.01677694357931614,
+ "learning_rate": 0.0006,
+ "loss": 4.27766227722168,
+ "step": 4740
+ },
+ {
+ "epoch": 65.8524017467249,
+ "grad_norm": 0.01645066775381565,
+ "learning_rate": 0.0006,
+ "loss": 4.288347244262695,
+ "step": 4741
+ },
+ {
+ "epoch": 65.86637554585153,
+ "grad_norm": 0.018034635111689568,
+ "learning_rate": 0.0006,
+ "loss": 4.292458534240723,
+ "step": 4742
+ },
+ {
+ "epoch": 65.88034934497817,
+ "grad_norm": 0.017973346635699272,
+ "learning_rate": 0.0006,
+ "loss": 4.3117265701293945,
+ "step": 4743
+ },
+ {
+ "epoch": 65.8943231441048,
+ "grad_norm": 0.016605669632554054,
+ "learning_rate": 0.0006,
+ "loss": 4.280615329742432,
+ "step": 4744
+ },
+ {
+ "epoch": 65.90829694323143,
+ "grad_norm": 0.016298074275255203,
+ "learning_rate": 0.0006,
+ "loss": 4.301933288574219,
+ "step": 4745
+ },
+ {
+ "epoch": 65.92227074235808,
+ "grad_norm": 0.015383216552436352,
+ "learning_rate": 0.0006,
+ "loss": 4.352383136749268,
+ "step": 4746
+ },
+ {
+ "epoch": 65.93624454148471,
+ "grad_norm": 0.016222182661294937,
+ "learning_rate": 0.0006,
+ "loss": 4.305725574493408,
+ "step": 4747
+ },
+ {
+ "epoch": 65.95021834061136,
+ "grad_norm": 0.01691329851746559,
+ "learning_rate": 0.0006,
+ "loss": 4.29502010345459,
+ "step": 4748
+ },
+ {
+ "epoch": 65.96419213973799,
+ "grad_norm": 0.017637677490711212,
+ "learning_rate": 0.0006,
+ "loss": 4.3133039474487305,
+ "step": 4749
+ },
+ {
+ "epoch": 65.97816593886463,
+ "grad_norm": 0.021362047642469406,
+ "learning_rate": 0.0006,
+ "loss": 4.458834648132324,
+ "step": 4750
+ },
+ {
+ "epoch": 65.99213973799127,
+ "grad_norm": 0.023553011938929558,
+ "learning_rate": 0.0006,
+ "loss": 4.298727989196777,
+ "step": 4751
+ },
+ {
+ "epoch": 66.0,
+ "grad_norm": 0.02154950052499771,
+ "learning_rate": 0.0006,
+ "loss": 4.403392791748047,
+ "step": 4752
+ },
+ {
+ "epoch": 66.0,
+ "eval_loss": 4.724340438842773,
+ "eval_runtime": 56.9555,
+ "eval_samples_per_second": 42.876,
+ "eval_steps_per_second": 1.352,
+ "step": 4752
+ },
+ {
+ "epoch": 66.01397379912663,
+ "grad_norm": 0.01916784606873989,
+ "learning_rate": 0.0006,
+ "loss": 4.22931432723999,
+ "step": 4753
+ },
+ {
+ "epoch": 66.02794759825328,
+ "grad_norm": 0.017492054030299187,
+ "learning_rate": 0.0006,
+ "loss": 4.271522521972656,
+ "step": 4754
+ },
+ {
+ "epoch": 66.04192139737991,
+ "grad_norm": 0.018408095464110374,
+ "learning_rate": 0.0006,
+ "loss": 4.337705135345459,
+ "step": 4755
+ },
+ {
+ "epoch": 66.05589519650655,
+ "grad_norm": 0.019220108166337013,
+ "learning_rate": 0.0006,
+ "loss": 4.280340194702148,
+ "step": 4756
+ },
+ {
+ "epoch": 66.06986899563319,
+ "grad_norm": 0.01773735322058201,
+ "learning_rate": 0.0006,
+ "loss": 4.390201568603516,
+ "step": 4757
+ },
+ {
+ "epoch": 66.08384279475983,
+ "grad_norm": 0.01679125428199768,
+ "learning_rate": 0.0006,
+ "loss": 4.299282073974609,
+ "step": 4758
+ },
+ {
+ "epoch": 66.09781659388646,
+ "grad_norm": 0.01479059923440218,
+ "learning_rate": 0.0006,
+ "loss": 4.472630500793457,
+ "step": 4759
+ },
+ {
+ "epoch": 66.1117903930131,
+ "grad_norm": 0.015130104497075081,
+ "learning_rate": 0.0006,
+ "loss": 4.301194190979004,
+ "step": 4760
+ },
+ {
+ "epoch": 66.12576419213974,
+ "grad_norm": 0.0158979594707489,
+ "learning_rate": 0.0006,
+ "loss": 4.375751972198486,
+ "step": 4761
+ },
+ {
+ "epoch": 66.13973799126637,
+ "grad_norm": 0.014432408846914768,
+ "learning_rate": 0.0006,
+ "loss": 4.279788017272949,
+ "step": 4762
+ },
+ {
+ "epoch": 66.15371179039302,
+ "grad_norm": 0.013026784174144268,
+ "learning_rate": 0.0006,
+ "loss": 4.255035400390625,
+ "step": 4763
+ },
+ {
+ "epoch": 66.16768558951965,
+ "grad_norm": 0.014422653242945671,
+ "learning_rate": 0.0006,
+ "loss": 4.261832237243652,
+ "step": 4764
+ },
+ {
+ "epoch": 66.18165938864628,
+ "grad_norm": 0.014613019302487373,
+ "learning_rate": 0.0006,
+ "loss": 4.382184028625488,
+ "step": 4765
+ },
+ {
+ "epoch": 66.19563318777293,
+ "grad_norm": 0.014989894814789295,
+ "learning_rate": 0.0006,
+ "loss": 4.40528678894043,
+ "step": 4766
+ },
+ {
+ "epoch": 66.20960698689956,
+ "grad_norm": 0.018190350383520126,
+ "learning_rate": 0.0006,
+ "loss": 4.302414894104004,
+ "step": 4767
+ },
+ {
+ "epoch": 66.2235807860262,
+ "grad_norm": 0.020887890830636024,
+ "learning_rate": 0.0006,
+ "loss": 4.289920806884766,
+ "step": 4768
+ },
+ {
+ "epoch": 66.23755458515284,
+ "grad_norm": 0.02376876398921013,
+ "learning_rate": 0.0006,
+ "loss": 4.297855377197266,
+ "step": 4769
+ },
+ {
+ "epoch": 66.25152838427948,
+ "grad_norm": 0.02152269519865513,
+ "learning_rate": 0.0006,
+ "loss": 4.1972832679748535,
+ "step": 4770
+ },
+ {
+ "epoch": 66.26550218340611,
+ "grad_norm": 0.015572802163660526,
+ "learning_rate": 0.0006,
+ "loss": 4.307487487792969,
+ "step": 4771
+ },
+ {
+ "epoch": 66.27947598253274,
+ "grad_norm": 0.015228749252855778,
+ "learning_rate": 0.0006,
+ "loss": 4.103250503540039,
+ "step": 4772
+ },
+ {
+ "epoch": 66.29344978165939,
+ "grad_norm": 0.016925504431128502,
+ "learning_rate": 0.0006,
+ "loss": 4.247461318969727,
+ "step": 4773
+ },
+ {
+ "epoch": 66.30742358078602,
+ "grad_norm": 0.01613711006939411,
+ "learning_rate": 0.0006,
+ "loss": 4.4699273109436035,
+ "step": 4774
+ },
+ {
+ "epoch": 66.32139737991267,
+ "grad_norm": 0.018332140520215034,
+ "learning_rate": 0.0006,
+ "loss": 4.22969388961792,
+ "step": 4775
+ },
+ {
+ "epoch": 66.3353711790393,
+ "grad_norm": 0.019575074315071106,
+ "learning_rate": 0.0006,
+ "loss": 4.29397439956665,
+ "step": 4776
+ },
+ {
+ "epoch": 66.34934497816595,
+ "grad_norm": 0.018177419900894165,
+ "learning_rate": 0.0006,
+ "loss": 4.221899509429932,
+ "step": 4777
+ },
+ {
+ "epoch": 66.36331877729258,
+ "grad_norm": 0.014292889274656773,
+ "learning_rate": 0.0006,
+ "loss": 4.325457572937012,
+ "step": 4778
+ },
+ {
+ "epoch": 66.37729257641921,
+ "grad_norm": 0.015098796226084232,
+ "learning_rate": 0.0006,
+ "loss": 4.364837169647217,
+ "step": 4779
+ },
+ {
+ "epoch": 66.39126637554585,
+ "grad_norm": 0.013992635533213615,
+ "learning_rate": 0.0006,
+ "loss": 4.299082279205322,
+ "step": 4780
+ },
+ {
+ "epoch": 66.40524017467249,
+ "grad_norm": 0.01504511572420597,
+ "learning_rate": 0.0006,
+ "loss": 4.230877876281738,
+ "step": 4781
+ },
+ {
+ "epoch": 66.41921397379913,
+ "grad_norm": 0.014875639230012894,
+ "learning_rate": 0.0006,
+ "loss": 4.368405818939209,
+ "step": 4782
+ },
+ {
+ "epoch": 66.43318777292576,
+ "grad_norm": 0.014306887984275818,
+ "learning_rate": 0.0006,
+ "loss": 4.389184474945068,
+ "step": 4783
+ },
+ {
+ "epoch": 66.44716157205241,
+ "grad_norm": 0.012959563173353672,
+ "learning_rate": 0.0006,
+ "loss": 4.37934684753418,
+ "step": 4784
+ },
+ {
+ "epoch": 66.46113537117904,
+ "grad_norm": 0.013206703588366508,
+ "learning_rate": 0.0006,
+ "loss": 4.442799091339111,
+ "step": 4785
+ },
+ {
+ "epoch": 66.47510917030567,
+ "grad_norm": 0.014848909340798855,
+ "learning_rate": 0.0006,
+ "loss": 4.203481674194336,
+ "step": 4786
+ },
+ {
+ "epoch": 66.48908296943232,
+ "grad_norm": 0.015637682750821114,
+ "learning_rate": 0.0006,
+ "loss": 4.27552604675293,
+ "step": 4787
+ },
+ {
+ "epoch": 66.50305676855895,
+ "grad_norm": 0.016017816960811615,
+ "learning_rate": 0.0006,
+ "loss": 4.296402931213379,
+ "step": 4788
+ },
+ {
+ "epoch": 66.5170305676856,
+ "grad_norm": 0.01577814109623432,
+ "learning_rate": 0.0006,
+ "loss": 4.301733016967773,
+ "step": 4789
+ },
+ {
+ "epoch": 66.53100436681223,
+ "grad_norm": 0.015293175354599953,
+ "learning_rate": 0.0006,
+ "loss": 4.369877815246582,
+ "step": 4790
+ },
+ {
+ "epoch": 66.54497816593886,
+ "grad_norm": 0.014262278564274311,
+ "learning_rate": 0.0006,
+ "loss": 4.3314738273620605,
+ "step": 4791
+ },
+ {
+ "epoch": 66.5589519650655,
+ "grad_norm": 0.015123482793569565,
+ "learning_rate": 0.0006,
+ "loss": 4.313591957092285,
+ "step": 4792
+ },
+ {
+ "epoch": 66.57292576419214,
+ "grad_norm": 0.015458766371011734,
+ "learning_rate": 0.0006,
+ "loss": 4.3045854568481445,
+ "step": 4793
+ },
+ {
+ "epoch": 66.58689956331878,
+ "grad_norm": 0.017217233777046204,
+ "learning_rate": 0.0006,
+ "loss": 4.269944190979004,
+ "step": 4794
+ },
+ {
+ "epoch": 66.60087336244541,
+ "grad_norm": 0.01962282694876194,
+ "learning_rate": 0.0006,
+ "loss": 4.271735191345215,
+ "step": 4795
+ },
+ {
+ "epoch": 66.61484716157206,
+ "grad_norm": 0.01917066052556038,
+ "learning_rate": 0.0006,
+ "loss": 4.288168430328369,
+ "step": 4796
+ },
+ {
+ "epoch": 66.62882096069869,
+ "grad_norm": 0.018211524933576584,
+ "learning_rate": 0.0006,
+ "loss": 4.280660629272461,
+ "step": 4797
+ },
+ {
+ "epoch": 66.64279475982532,
+ "grad_norm": 0.01936524175107479,
+ "learning_rate": 0.0006,
+ "loss": 4.3710551261901855,
+ "step": 4798
+ },
+ {
+ "epoch": 66.65676855895197,
+ "grad_norm": 0.01882351003587246,
+ "learning_rate": 0.0006,
+ "loss": 4.342750549316406,
+ "step": 4799
+ },
+ {
+ "epoch": 66.6707423580786,
+ "grad_norm": 0.017149122431874275,
+ "learning_rate": 0.0006,
+ "loss": 4.32134485244751,
+ "step": 4800
+ },
+ {
+ "epoch": 66.68471615720524,
+ "grad_norm": 0.01699327677488327,
+ "learning_rate": 0.0006,
+ "loss": 4.248893737792969,
+ "step": 4801
+ },
+ {
+ "epoch": 66.69868995633188,
+ "grad_norm": 0.016873255372047424,
+ "learning_rate": 0.0006,
+ "loss": 4.277350425720215,
+ "step": 4802
+ },
+ {
+ "epoch": 66.71266375545852,
+ "grad_norm": 0.016592837870121002,
+ "learning_rate": 0.0006,
+ "loss": 4.221665859222412,
+ "step": 4803
+ },
+ {
+ "epoch": 66.72663755458515,
+ "grad_norm": 0.015887131914496422,
+ "learning_rate": 0.0006,
+ "loss": 4.328040599822998,
+ "step": 4804
+ },
+ {
+ "epoch": 66.74061135371178,
+ "grad_norm": 0.014057286083698273,
+ "learning_rate": 0.0006,
+ "loss": 4.3294548988342285,
+ "step": 4805
+ },
+ {
+ "epoch": 66.75458515283843,
+ "grad_norm": 0.013135905377566814,
+ "learning_rate": 0.0006,
+ "loss": 4.254383563995361,
+ "step": 4806
+ },
+ {
+ "epoch": 66.76855895196506,
+ "grad_norm": 0.013632557354867458,
+ "learning_rate": 0.0006,
+ "loss": 4.342815399169922,
+ "step": 4807
+ },
+ {
+ "epoch": 66.78253275109171,
+ "grad_norm": 0.015512688085436821,
+ "learning_rate": 0.0006,
+ "loss": 4.297835350036621,
+ "step": 4808
+ },
+ {
+ "epoch": 66.79650655021834,
+ "grad_norm": 0.01621832884848118,
+ "learning_rate": 0.0006,
+ "loss": 4.314773082733154,
+ "step": 4809
+ },
+ {
+ "epoch": 66.81048034934499,
+ "grad_norm": 0.016196032986044884,
+ "learning_rate": 0.0006,
+ "loss": 4.29209566116333,
+ "step": 4810
+ },
+ {
+ "epoch": 66.82445414847162,
+ "grad_norm": 0.015463353134691715,
+ "learning_rate": 0.0006,
+ "loss": 4.262795448303223,
+ "step": 4811
+ },
+ {
+ "epoch": 66.83842794759825,
+ "grad_norm": 0.016143787652254105,
+ "learning_rate": 0.0006,
+ "loss": 4.274284362792969,
+ "step": 4812
+ },
+ {
+ "epoch": 66.8524017467249,
+ "grad_norm": 0.01610061153769493,
+ "learning_rate": 0.0006,
+ "loss": 4.337160587310791,
+ "step": 4813
+ },
+ {
+ "epoch": 66.86637554585153,
+ "grad_norm": 0.01582838036119938,
+ "learning_rate": 0.0006,
+ "loss": 4.192723274230957,
+ "step": 4814
+ },
+ {
+ "epoch": 66.88034934497817,
+ "grad_norm": 0.015081540681421757,
+ "learning_rate": 0.0006,
+ "loss": 4.370708465576172,
+ "step": 4815
+ },
+ {
+ "epoch": 66.8943231441048,
+ "grad_norm": 0.015929723158478737,
+ "learning_rate": 0.0006,
+ "loss": 4.340244293212891,
+ "step": 4816
+ },
+ {
+ "epoch": 66.90829694323143,
+ "grad_norm": 0.01784711331129074,
+ "learning_rate": 0.0006,
+ "loss": 4.260406970977783,
+ "step": 4817
+ },
+ {
+ "epoch": 66.92227074235808,
+ "grad_norm": 0.017434703186154366,
+ "learning_rate": 0.0006,
+ "loss": 4.286900043487549,
+ "step": 4818
+ },
+ {
+ "epoch": 66.93624454148471,
+ "grad_norm": 0.016262540593743324,
+ "learning_rate": 0.0006,
+ "loss": 4.292142391204834,
+ "step": 4819
+ },
+ {
+ "epoch": 66.95021834061136,
+ "grad_norm": 0.016122829169034958,
+ "learning_rate": 0.0006,
+ "loss": 4.291485786437988,
+ "step": 4820
+ },
+ {
+ "epoch": 66.96419213973799,
+ "grad_norm": 0.014361178502440453,
+ "learning_rate": 0.0006,
+ "loss": 4.275425910949707,
+ "step": 4821
+ },
+ {
+ "epoch": 66.97816593886463,
+ "grad_norm": 0.01373987551778555,
+ "learning_rate": 0.0006,
+ "loss": 4.223120212554932,
+ "step": 4822
+ },
+ {
+ "epoch": 66.99213973799127,
+ "grad_norm": 0.015687376260757446,
+ "learning_rate": 0.0006,
+ "loss": 4.355381011962891,
+ "step": 4823
+ },
+ {
+ "epoch": 67.0,
+ "grad_norm": 0.01998465321958065,
+ "learning_rate": 0.0006,
+ "loss": 4.264053821563721,
+ "step": 4824
+ },
+ {
+ "epoch": 67.0,
+ "eval_loss": 4.662398338317871,
+ "eval_runtime": 56.7928,
+ "eval_samples_per_second": 42.998,
+ "eval_steps_per_second": 1.356,
+ "step": 4824
+ },
+ {
+ "epoch": 67.01397379912663,
+ "grad_norm": 0.021736836060881615,
+ "learning_rate": 0.0006,
+ "loss": 4.243354797363281,
+ "step": 4825
+ },
+ {
+ "epoch": 67.02794759825328,
+ "grad_norm": 0.021476037800312042,
+ "learning_rate": 0.0006,
+ "loss": 4.218207836151123,
+ "step": 4826
+ },
+ {
+ "epoch": 67.04192139737991,
+ "grad_norm": 0.017188388854265213,
+ "learning_rate": 0.0006,
+ "loss": 4.18048095703125,
+ "step": 4827
+ },
+ {
+ "epoch": 67.05589519650655,
+ "grad_norm": 0.017293719574809074,
+ "learning_rate": 0.0006,
+ "loss": 4.239145278930664,
+ "step": 4828
+ },
+ {
+ "epoch": 67.06986899563319,
+ "grad_norm": 0.017383819445967674,
+ "learning_rate": 0.0006,
+ "loss": 4.114765167236328,
+ "step": 4829
+ },
+ {
+ "epoch": 67.08384279475983,
+ "grad_norm": 0.01651529036462307,
+ "learning_rate": 0.0006,
+ "loss": 4.351023197174072,
+ "step": 4830
+ },
+ {
+ "epoch": 67.09781659388646,
+ "grad_norm": 0.015893027186393738,
+ "learning_rate": 0.0006,
+ "loss": 4.261432647705078,
+ "step": 4831
+ },
+ {
+ "epoch": 67.1117903930131,
+ "grad_norm": 0.013381517492234707,
+ "learning_rate": 0.0006,
+ "loss": 4.172064781188965,
+ "step": 4832
+ },
+ {
+ "epoch": 67.12576419213974,
+ "grad_norm": 0.014642654918134212,
+ "learning_rate": 0.0006,
+ "loss": 4.356171131134033,
+ "step": 4833
+ },
+ {
+ "epoch": 67.13973799126637,
+ "grad_norm": 0.01425282284617424,
+ "learning_rate": 0.0006,
+ "loss": 4.303495407104492,
+ "step": 4834
+ },
+ {
+ "epoch": 67.15371179039302,
+ "grad_norm": 0.014574599452316761,
+ "learning_rate": 0.0006,
+ "loss": 4.2850847244262695,
+ "step": 4835
+ },
+ {
+ "epoch": 67.16768558951965,
+ "grad_norm": 0.012676282785832882,
+ "learning_rate": 0.0006,
+ "loss": 4.297845840454102,
+ "step": 4836
+ },
+ {
+ "epoch": 67.18165938864628,
+ "grad_norm": 0.013948136940598488,
+ "learning_rate": 0.0006,
+ "loss": 4.324530601501465,
+ "step": 4837
+ },
+ {
+ "epoch": 67.19563318777293,
+ "grad_norm": 0.014195773750543594,
+ "learning_rate": 0.0006,
+ "loss": 4.31236457824707,
+ "step": 4838
+ },
+ {
+ "epoch": 67.20960698689956,
+ "grad_norm": 0.013631382025778294,
+ "learning_rate": 0.0006,
+ "loss": 4.086362838745117,
+ "step": 4839
+ },
+ {
+ "epoch": 67.2235807860262,
+ "grad_norm": 0.012798587791621685,
+ "learning_rate": 0.0006,
+ "loss": 4.328471660614014,
+ "step": 4840
+ },
+ {
+ "epoch": 67.23755458515284,
+ "grad_norm": 0.014443321153521538,
+ "learning_rate": 0.0006,
+ "loss": 4.278019905090332,
+ "step": 4841
+ },
+ {
+ "epoch": 67.25152838427948,
+ "grad_norm": 0.015355420298874378,
+ "learning_rate": 0.0006,
+ "loss": 4.333611011505127,
+ "step": 4842
+ },
+ {
+ "epoch": 67.26550218340611,
+ "grad_norm": 0.016487296670675278,
+ "learning_rate": 0.0006,
+ "loss": 4.351336479187012,
+ "step": 4843
+ },
+ {
+ "epoch": 67.27947598253274,
+ "grad_norm": 0.017422541975975037,
+ "learning_rate": 0.0006,
+ "loss": 4.386096477508545,
+ "step": 4844
+ },
+ {
+ "epoch": 67.29344978165939,
+ "grad_norm": 0.015790918841958046,
+ "learning_rate": 0.0006,
+ "loss": 4.301377773284912,
+ "step": 4845
+ },
+ {
+ "epoch": 67.30742358078602,
+ "grad_norm": 0.016427364200353622,
+ "learning_rate": 0.0006,
+ "loss": 4.2838945388793945,
+ "step": 4846
+ },
+ {
+ "epoch": 67.32139737991267,
+ "grad_norm": 0.017050761729478836,
+ "learning_rate": 0.0006,
+ "loss": 4.282417297363281,
+ "step": 4847
+ },
+ {
+ "epoch": 67.3353711790393,
+ "grad_norm": 0.015006253495812416,
+ "learning_rate": 0.0006,
+ "loss": 4.352236270904541,
+ "step": 4848
+ },
+ {
+ "epoch": 67.34934497816595,
+ "grad_norm": 0.012996627949178219,
+ "learning_rate": 0.0006,
+ "loss": 4.309242248535156,
+ "step": 4849
+ },
+ {
+ "epoch": 67.36331877729258,
+ "grad_norm": 0.01367615070194006,
+ "learning_rate": 0.0006,
+ "loss": 4.211801528930664,
+ "step": 4850
+ },
+ {
+ "epoch": 67.37729257641921,
+ "grad_norm": 0.014775963500142097,
+ "learning_rate": 0.0006,
+ "loss": 4.360859394073486,
+ "step": 4851
+ },
+ {
+ "epoch": 67.39126637554585,
+ "grad_norm": 0.015854304656386375,
+ "learning_rate": 0.0006,
+ "loss": 4.211830139160156,
+ "step": 4852
+ },
+ {
+ "epoch": 67.40524017467249,
+ "grad_norm": 0.016888894140720367,
+ "learning_rate": 0.0006,
+ "loss": 4.216549873352051,
+ "step": 4853
+ },
+ {
+ "epoch": 67.41921397379913,
+ "grad_norm": 0.0170894768089056,
+ "learning_rate": 0.0006,
+ "loss": 4.263486385345459,
+ "step": 4854
+ },
+ {
+ "epoch": 67.43318777292576,
+ "grad_norm": 0.015829868614673615,
+ "learning_rate": 0.0006,
+ "loss": 4.360196113586426,
+ "step": 4855
+ },
+ {
+ "epoch": 67.44716157205241,
+ "grad_norm": 0.015220202505588531,
+ "learning_rate": 0.0006,
+ "loss": 4.243794918060303,
+ "step": 4856
+ },
+ {
+ "epoch": 67.46113537117904,
+ "grad_norm": 0.01413388829678297,
+ "learning_rate": 0.0006,
+ "loss": 4.286624908447266,
+ "step": 4857
+ },
+ {
+ "epoch": 67.47510917030567,
+ "grad_norm": 0.01426626369357109,
+ "learning_rate": 0.0006,
+ "loss": 4.257316589355469,
+ "step": 4858
+ },
+ {
+ "epoch": 67.48908296943232,
+ "grad_norm": 0.015957845374941826,
+ "learning_rate": 0.0006,
+ "loss": 4.300443649291992,
+ "step": 4859
+ },
+ {
+ "epoch": 67.50305676855895,
+ "grad_norm": 0.015499051660299301,
+ "learning_rate": 0.0006,
+ "loss": 4.278162002563477,
+ "step": 4860
+ },
+ {
+ "epoch": 67.5170305676856,
+ "grad_norm": 0.015145624987781048,
+ "learning_rate": 0.0006,
+ "loss": 4.267950534820557,
+ "step": 4861
+ },
+ {
+ "epoch": 67.53100436681223,
+ "grad_norm": 0.014610803686082363,
+ "learning_rate": 0.0006,
+ "loss": 4.247499942779541,
+ "step": 4862
+ },
+ {
+ "epoch": 67.54497816593886,
+ "grad_norm": 0.014860909432172775,
+ "learning_rate": 0.0006,
+ "loss": 4.263200759887695,
+ "step": 4863
+ },
+ {
+ "epoch": 67.5589519650655,
+ "grad_norm": 0.01718984916806221,
+ "learning_rate": 0.0006,
+ "loss": 4.290019989013672,
+ "step": 4864
+ },
+ {
+ "epoch": 67.57292576419214,
+ "grad_norm": 0.01558963768184185,
+ "learning_rate": 0.0006,
+ "loss": 4.223816871643066,
+ "step": 4865
+ },
+ {
+ "epoch": 67.58689956331878,
+ "grad_norm": 0.014927188865840435,
+ "learning_rate": 0.0006,
+ "loss": 4.363586902618408,
+ "step": 4866
+ },
+ {
+ "epoch": 67.60087336244541,
+ "grad_norm": 0.01573595404624939,
+ "learning_rate": 0.0006,
+ "loss": 4.254765510559082,
+ "step": 4867
+ },
+ {
+ "epoch": 67.61484716157206,
+ "grad_norm": 0.015618929639458656,
+ "learning_rate": 0.0006,
+ "loss": 4.25072717666626,
+ "step": 4868
+ },
+ {
+ "epoch": 67.62882096069869,
+ "grad_norm": 0.01605590060353279,
+ "learning_rate": 0.0006,
+ "loss": 4.25208854675293,
+ "step": 4869
+ },
+ {
+ "epoch": 67.64279475982532,
+ "grad_norm": 0.01804601214826107,
+ "learning_rate": 0.0006,
+ "loss": 4.351354122161865,
+ "step": 4870
+ },
+ {
+ "epoch": 67.65676855895197,
+ "grad_norm": 0.020595047622919083,
+ "learning_rate": 0.0006,
+ "loss": 4.193732738494873,
+ "step": 4871
+ },
+ {
+ "epoch": 67.6707423580786,
+ "grad_norm": 0.023037966340780258,
+ "learning_rate": 0.0006,
+ "loss": 4.27754020690918,
+ "step": 4872
+ },
+ {
+ "epoch": 67.68471615720524,
+ "grad_norm": 0.020680895075201988,
+ "learning_rate": 0.0006,
+ "loss": 4.314730167388916,
+ "step": 4873
+ },
+ {
+ "epoch": 67.69868995633188,
+ "grad_norm": 0.01887434534728527,
+ "learning_rate": 0.0006,
+ "loss": 4.252305030822754,
+ "step": 4874
+ },
+ {
+ "epoch": 67.71266375545852,
+ "grad_norm": 0.016333596780896187,
+ "learning_rate": 0.0006,
+ "loss": 4.146123886108398,
+ "step": 4875
+ },
+ {
+ "epoch": 67.72663755458515,
+ "grad_norm": 0.015688810497522354,
+ "learning_rate": 0.0006,
+ "loss": 4.2814154624938965,
+ "step": 4876
+ },
+ {
+ "epoch": 67.74061135371178,
+ "grad_norm": 0.016270726919174194,
+ "learning_rate": 0.0006,
+ "loss": 4.323200225830078,
+ "step": 4877
+ },
+ {
+ "epoch": 67.75458515283843,
+ "grad_norm": 0.016824988648295403,
+ "learning_rate": 0.0006,
+ "loss": 4.269501686096191,
+ "step": 4878
+ },
+ {
+ "epoch": 67.76855895196506,
+ "grad_norm": 0.01730562001466751,
+ "learning_rate": 0.0006,
+ "loss": 4.187993049621582,
+ "step": 4879
+ },
+ {
+ "epoch": 67.78253275109171,
+ "grad_norm": 0.01546655036509037,
+ "learning_rate": 0.0006,
+ "loss": 4.340415000915527,
+ "step": 4880
+ },
+ {
+ "epoch": 67.79650655021834,
+ "grad_norm": 0.014173958450555801,
+ "learning_rate": 0.0006,
+ "loss": 4.217419624328613,
+ "step": 4881
+ },
+ {
+ "epoch": 67.81048034934499,
+ "grad_norm": 0.016810474917292595,
+ "learning_rate": 0.0006,
+ "loss": 4.352302074432373,
+ "step": 4882
+ },
+ {
+ "epoch": 67.82445414847162,
+ "grad_norm": 0.0175373163074255,
+ "learning_rate": 0.0006,
+ "loss": 4.20377254486084,
+ "step": 4883
+ },
+ {
+ "epoch": 67.83842794759825,
+ "grad_norm": 0.015734676271677017,
+ "learning_rate": 0.0006,
+ "loss": 4.147495746612549,
+ "step": 4884
+ },
+ {
+ "epoch": 67.8524017467249,
+ "grad_norm": 0.015091817826032639,
+ "learning_rate": 0.0006,
+ "loss": 4.272252082824707,
+ "step": 4885
+ },
+ {
+ "epoch": 67.86637554585153,
+ "grad_norm": 0.015000219456851482,
+ "learning_rate": 0.0006,
+ "loss": 4.302773952484131,
+ "step": 4886
+ },
+ {
+ "epoch": 67.88034934497817,
+ "grad_norm": 0.013827184215188026,
+ "learning_rate": 0.0006,
+ "loss": 4.337242126464844,
+ "step": 4887
+ },
+ {
+ "epoch": 67.8943231441048,
+ "grad_norm": 0.015308346599340439,
+ "learning_rate": 0.0006,
+ "loss": 4.140895366668701,
+ "step": 4888
+ },
+ {
+ "epoch": 67.90829694323143,
+ "grad_norm": 0.015356178395450115,
+ "learning_rate": 0.0006,
+ "loss": 4.312541484832764,
+ "step": 4889
+ },
+ {
+ "epoch": 67.92227074235808,
+ "grad_norm": 0.0137353939935565,
+ "learning_rate": 0.0006,
+ "loss": 4.173184394836426,
+ "step": 4890
+ },
+ {
+ "epoch": 67.93624454148471,
+ "grad_norm": 0.01332368515431881,
+ "learning_rate": 0.0006,
+ "loss": 4.191199779510498,
+ "step": 4891
+ },
+ {
+ "epoch": 67.95021834061136,
+ "grad_norm": 0.01488130260258913,
+ "learning_rate": 0.0006,
+ "loss": 4.331752300262451,
+ "step": 4892
+ },
+ {
+ "epoch": 67.96419213973799,
+ "grad_norm": 0.015449733473360538,
+ "learning_rate": 0.0006,
+ "loss": 4.2385454177856445,
+ "step": 4893
+ },
+ {
+ "epoch": 67.97816593886463,
+ "grad_norm": 0.01556103490293026,
+ "learning_rate": 0.0006,
+ "loss": 4.262353897094727,
+ "step": 4894
+ },
+ {
+ "epoch": 67.99213973799127,
+ "grad_norm": 0.014613849110901356,
+ "learning_rate": 0.0006,
+ "loss": 4.223945617675781,
+ "step": 4895
+ },
+ {
+ "epoch": 68.0,
+ "grad_norm": 0.015323462896049023,
+ "learning_rate": 0.0006,
+ "loss": 4.317168712615967,
+ "step": 4896
+ },
+ {
+ "epoch": 68.0,
+ "eval_loss": 4.671603679656982,
+ "eval_runtime": 56.7317,
+ "eval_samples_per_second": 43.045,
+ "eval_steps_per_second": 1.357,
+ "step": 4896
+ },
+ {
+ "epoch": 68.01397379912663,
+ "grad_norm": 0.01910698413848877,
+ "learning_rate": 0.0006,
+ "loss": 4.206167221069336,
+ "step": 4897
+ },
+ {
+ "epoch": 68.02794759825328,
+ "grad_norm": 0.025112485513091087,
+ "learning_rate": 0.0006,
+ "loss": 4.384706974029541,
+ "step": 4898
+ },
+ {
+ "epoch": 68.04192139737991,
+ "grad_norm": 0.02618720941245556,
+ "learning_rate": 0.0006,
+ "loss": 4.321046829223633,
+ "step": 4899
+ },
+ {
+ "epoch": 68.05589519650655,
+ "grad_norm": 0.020853538066148758,
+ "learning_rate": 0.0006,
+ "loss": 4.217638969421387,
+ "step": 4900
+ },
+ {
+ "epoch": 68.06986899563319,
+ "grad_norm": 0.020025717094540596,
+ "learning_rate": 0.0006,
+ "loss": 4.306824684143066,
+ "step": 4901
+ },
+ {
+ "epoch": 68.08384279475983,
+ "grad_norm": 0.023868650197982788,
+ "learning_rate": 0.0006,
+ "loss": 4.327455997467041,
+ "step": 4902
+ },
+ {
+ "epoch": 68.09781659388646,
+ "grad_norm": 0.022279081866145134,
+ "learning_rate": 0.0006,
+ "loss": 4.400699615478516,
+ "step": 4903
+ },
+ {
+ "epoch": 68.1117903930131,
+ "grad_norm": 0.018965687602758408,
+ "learning_rate": 0.0006,
+ "loss": 4.19097900390625,
+ "step": 4904
+ },
+ {
+ "epoch": 68.12576419213974,
+ "grad_norm": 0.01936378888785839,
+ "learning_rate": 0.0006,
+ "loss": 4.206711769104004,
+ "step": 4905
+ },
+ {
+ "epoch": 68.13973799126637,
+ "grad_norm": 0.01811872608959675,
+ "learning_rate": 0.0006,
+ "loss": 4.204622268676758,
+ "step": 4906
+ },
+ {
+ "epoch": 68.15371179039302,
+ "grad_norm": 0.018530402332544327,
+ "learning_rate": 0.0006,
+ "loss": 4.289602279663086,
+ "step": 4907
+ },
+ {
+ "epoch": 68.16768558951965,
+ "grad_norm": 0.01755962148308754,
+ "learning_rate": 0.0006,
+ "loss": 4.317145347595215,
+ "step": 4908
+ },
+ {
+ "epoch": 68.18165938864628,
+ "grad_norm": 0.020413702353835106,
+ "learning_rate": 0.0006,
+ "loss": 4.2882843017578125,
+ "step": 4909
+ },
+ {
+ "epoch": 68.19563318777293,
+ "grad_norm": 0.020064299926161766,
+ "learning_rate": 0.0006,
+ "loss": 4.248260021209717,
+ "step": 4910
+ },
+ {
+ "epoch": 68.20960698689956,
+ "grad_norm": 0.017415888607501984,
+ "learning_rate": 0.0006,
+ "loss": 4.370295524597168,
+ "step": 4911
+ },
+ {
+ "epoch": 68.2235807860262,
+ "grad_norm": 0.015005858615040779,
+ "learning_rate": 0.0006,
+ "loss": 4.319567680358887,
+ "step": 4912
+ },
+ {
+ "epoch": 68.23755458515284,
+ "grad_norm": 0.016349777579307556,
+ "learning_rate": 0.0006,
+ "loss": 4.301776885986328,
+ "step": 4913
+ },
+ {
+ "epoch": 68.25152838427948,
+ "grad_norm": 0.015779560431838036,
+ "learning_rate": 0.0006,
+ "loss": 4.271171569824219,
+ "step": 4914
+ },
+ {
+ "epoch": 68.26550218340611,
+ "grad_norm": 0.013237644918262959,
+ "learning_rate": 0.0006,
+ "loss": 4.324908256530762,
+ "step": 4915
+ },
+ {
+ "epoch": 68.27947598253274,
+ "grad_norm": 0.014076225459575653,
+ "learning_rate": 0.0006,
+ "loss": 4.177175998687744,
+ "step": 4916
+ },
+ {
+ "epoch": 68.29344978165939,
+ "grad_norm": 0.013906744308769703,
+ "learning_rate": 0.0006,
+ "loss": 4.11588191986084,
+ "step": 4917
+ },
+ {
+ "epoch": 68.30742358078602,
+ "grad_norm": 0.013939064927399158,
+ "learning_rate": 0.0006,
+ "loss": 4.183963775634766,
+ "step": 4918
+ },
+ {
+ "epoch": 68.32139737991267,
+ "grad_norm": 0.014572170563042164,
+ "learning_rate": 0.0006,
+ "loss": 4.19586181640625,
+ "step": 4919
+ },
+ {
+ "epoch": 68.3353711790393,
+ "grad_norm": 0.014085350558161736,
+ "learning_rate": 0.0006,
+ "loss": 4.170892238616943,
+ "step": 4920
+ },
+ {
+ "epoch": 68.34934497816595,
+ "grad_norm": 0.014357863925397396,
+ "learning_rate": 0.0006,
+ "loss": 4.321155548095703,
+ "step": 4921
+ },
+ {
+ "epoch": 68.36331877729258,
+ "grad_norm": 0.01461675763130188,
+ "learning_rate": 0.0006,
+ "loss": 4.293334007263184,
+ "step": 4922
+ },
+ {
+ "epoch": 68.37729257641921,
+ "grad_norm": 0.014828486368060112,
+ "learning_rate": 0.0006,
+ "loss": 4.1439714431762695,
+ "step": 4923
+ },
+ {
+ "epoch": 68.39126637554585,
+ "grad_norm": 0.013730363920331001,
+ "learning_rate": 0.0006,
+ "loss": 4.287973880767822,
+ "step": 4924
+ },
+ {
+ "epoch": 68.40524017467249,
+ "grad_norm": 0.013852203264832497,
+ "learning_rate": 0.0006,
+ "loss": 4.248824119567871,
+ "step": 4925
+ },
+ {
+ "epoch": 68.41921397379913,
+ "grad_norm": 0.014005862176418304,
+ "learning_rate": 0.0006,
+ "loss": 4.293768882751465,
+ "step": 4926
+ },
+ {
+ "epoch": 68.43318777292576,
+ "grad_norm": 0.014435851015150547,
+ "learning_rate": 0.0006,
+ "loss": 4.204813003540039,
+ "step": 4927
+ },
+ {
+ "epoch": 68.44716157205241,
+ "grad_norm": 0.015305934473872185,
+ "learning_rate": 0.0006,
+ "loss": 4.329364776611328,
+ "step": 4928
+ },
+ {
+ "epoch": 68.46113537117904,
+ "grad_norm": 0.014356814324855804,
+ "learning_rate": 0.0006,
+ "loss": 4.264206409454346,
+ "step": 4929
+ },
+ {
+ "epoch": 68.47510917030567,
+ "grad_norm": 0.01432347297668457,
+ "learning_rate": 0.0006,
+ "loss": 4.29440450668335,
+ "step": 4930
+ },
+ {
+ "epoch": 68.48908296943232,
+ "grad_norm": 0.012672432698309422,
+ "learning_rate": 0.0006,
+ "loss": 4.258024215698242,
+ "step": 4931
+ },
+ {
+ "epoch": 68.50305676855895,
+ "grad_norm": 0.014209977351129055,
+ "learning_rate": 0.0006,
+ "loss": 4.216668128967285,
+ "step": 4932
+ },
+ {
+ "epoch": 68.5170305676856,
+ "grad_norm": 0.014606663957238197,
+ "learning_rate": 0.0006,
+ "loss": 4.294184684753418,
+ "step": 4933
+ },
+ {
+ "epoch": 68.53100436681223,
+ "grad_norm": 0.014986990951001644,
+ "learning_rate": 0.0006,
+ "loss": 4.382591247558594,
+ "step": 4934
+ },
+ {
+ "epoch": 68.54497816593886,
+ "grad_norm": 0.013998783193528652,
+ "learning_rate": 0.0006,
+ "loss": 4.195143222808838,
+ "step": 4935
+ },
+ {
+ "epoch": 68.5589519650655,
+ "grad_norm": 0.015011102892458439,
+ "learning_rate": 0.0006,
+ "loss": 4.313785076141357,
+ "step": 4936
+ },
+ {
+ "epoch": 68.57292576419214,
+ "grad_norm": 0.01641209051012993,
+ "learning_rate": 0.0006,
+ "loss": 4.322803020477295,
+ "step": 4937
+ },
+ {
+ "epoch": 68.58689956331878,
+ "grad_norm": 0.016207464039325714,
+ "learning_rate": 0.0006,
+ "loss": 4.311431884765625,
+ "step": 4938
+ },
+ {
+ "epoch": 68.60087336244541,
+ "grad_norm": 0.01747811958193779,
+ "learning_rate": 0.0006,
+ "loss": 4.310845375061035,
+ "step": 4939
+ },
+ {
+ "epoch": 68.61484716157206,
+ "grad_norm": 0.019475717097520828,
+ "learning_rate": 0.0006,
+ "loss": 4.338624954223633,
+ "step": 4940
+ },
+ {
+ "epoch": 68.62882096069869,
+ "grad_norm": 0.018641719594597816,
+ "learning_rate": 0.0006,
+ "loss": 4.34913969039917,
+ "step": 4941
+ },
+ {
+ "epoch": 68.64279475982532,
+ "grad_norm": 0.018376432359218597,
+ "learning_rate": 0.0006,
+ "loss": 4.28464412689209,
+ "step": 4942
+ },
+ {
+ "epoch": 68.65676855895197,
+ "grad_norm": 0.017095010727643967,
+ "learning_rate": 0.0006,
+ "loss": 4.217957496643066,
+ "step": 4943
+ },
+ {
+ "epoch": 68.6707423580786,
+ "grad_norm": 0.016321983188390732,
+ "learning_rate": 0.0006,
+ "loss": 4.220042705535889,
+ "step": 4944
+ },
+ {
+ "epoch": 68.68471615720524,
+ "grad_norm": 0.017128000035881996,
+ "learning_rate": 0.0006,
+ "loss": 4.293500900268555,
+ "step": 4945
+ },
+ {
+ "epoch": 68.69868995633188,
+ "grad_norm": 0.015592411160469055,
+ "learning_rate": 0.0006,
+ "loss": 4.195798873901367,
+ "step": 4946
+ },
+ {
+ "epoch": 68.71266375545852,
+ "grad_norm": 0.016607079654932022,
+ "learning_rate": 0.0006,
+ "loss": 4.2654337882995605,
+ "step": 4947
+ },
+ {
+ "epoch": 68.72663755458515,
+ "grad_norm": 0.01545762736350298,
+ "learning_rate": 0.0006,
+ "loss": 4.303828239440918,
+ "step": 4948
+ },
+ {
+ "epoch": 68.74061135371178,
+ "grad_norm": 0.014328422956168652,
+ "learning_rate": 0.0006,
+ "loss": 4.29379940032959,
+ "step": 4949
+ },
+ {
+ "epoch": 68.75458515283843,
+ "grad_norm": 0.016516804695129395,
+ "learning_rate": 0.0006,
+ "loss": 4.2317214012146,
+ "step": 4950
+ },
+ {
+ "epoch": 68.76855895196506,
+ "grad_norm": 0.016324171796441078,
+ "learning_rate": 0.0006,
+ "loss": 4.278043746948242,
+ "step": 4951
+ },
+ {
+ "epoch": 68.78253275109171,
+ "grad_norm": 0.015511159785091877,
+ "learning_rate": 0.0006,
+ "loss": 4.188244342803955,
+ "step": 4952
+ },
+ {
+ "epoch": 68.79650655021834,
+ "grad_norm": 0.01676103286445141,
+ "learning_rate": 0.0006,
+ "loss": 4.2169976234436035,
+ "step": 4953
+ },
+ {
+ "epoch": 68.81048034934499,
+ "grad_norm": 0.015428312122821808,
+ "learning_rate": 0.0006,
+ "loss": 4.4004716873168945,
+ "step": 4954
+ },
+ {
+ "epoch": 68.82445414847162,
+ "grad_norm": 0.0162410456687212,
+ "learning_rate": 0.0006,
+ "loss": 4.282752513885498,
+ "step": 4955
+ },
+ {
+ "epoch": 68.83842794759825,
+ "grad_norm": 0.017578862607479095,
+ "learning_rate": 0.0006,
+ "loss": 4.098004341125488,
+ "step": 4956
+ },
+ {
+ "epoch": 68.8524017467249,
+ "grad_norm": 0.01805749163031578,
+ "learning_rate": 0.0006,
+ "loss": 4.277988433837891,
+ "step": 4957
+ },
+ {
+ "epoch": 68.86637554585153,
+ "grad_norm": 0.016728045418858528,
+ "learning_rate": 0.0006,
+ "loss": 4.29766321182251,
+ "step": 4958
+ },
+ {
+ "epoch": 68.88034934497817,
+ "grad_norm": 0.014899657107889652,
+ "learning_rate": 0.0006,
+ "loss": 4.329736709594727,
+ "step": 4959
+ },
+ {
+ "epoch": 68.8943231441048,
+ "grad_norm": 0.013441305607557297,
+ "learning_rate": 0.0006,
+ "loss": 4.2873215675354,
+ "step": 4960
+ },
+ {
+ "epoch": 68.90829694323143,
+ "grad_norm": 0.01554128248244524,
+ "learning_rate": 0.0006,
+ "loss": 4.278327941894531,
+ "step": 4961
+ },
+ {
+ "epoch": 68.92227074235808,
+ "grad_norm": 0.01806696318089962,
+ "learning_rate": 0.0006,
+ "loss": 4.38275146484375,
+ "step": 4962
+ },
+ {
+ "epoch": 68.93624454148471,
+ "grad_norm": 0.017025409266352654,
+ "learning_rate": 0.0006,
+ "loss": 4.210447311401367,
+ "step": 4963
+ },
+ {
+ "epoch": 68.95021834061136,
+ "grad_norm": 0.01597420684993267,
+ "learning_rate": 0.0006,
+ "loss": 4.309865951538086,
+ "step": 4964
+ },
+ {
+ "epoch": 68.96419213973799,
+ "grad_norm": 0.015835624188184738,
+ "learning_rate": 0.0006,
+ "loss": 4.225904941558838,
+ "step": 4965
+ },
+ {
+ "epoch": 68.97816593886463,
+ "grad_norm": 0.018170492723584175,
+ "learning_rate": 0.0006,
+ "loss": 4.248478889465332,
+ "step": 4966
+ },
+ {
+ "epoch": 68.99213973799127,
+ "grad_norm": 0.017372991889715195,
+ "learning_rate": 0.0006,
+ "loss": 4.283486843109131,
+ "step": 4967
+ },
+ {
+ "epoch": 69.0,
+ "grad_norm": 0.016305390745401382,
+ "learning_rate": 0.0006,
+ "loss": 4.273636817932129,
+ "step": 4968
+ },
+ {
+ "epoch": 69.0,
+ "eval_loss": 4.66873836517334,
+ "eval_runtime": 57.0791,
+ "eval_samples_per_second": 42.783,
+ "eval_steps_per_second": 1.349,
+ "step": 4968
+ },
+ {
+ "epoch": 69.01397379912663,
+ "grad_norm": 0.016567355021834373,
+ "learning_rate": 0.0006,
+ "loss": 4.269988059997559,
+ "step": 4969
+ },
+ {
+ "epoch": 69.02794759825328,
+ "grad_norm": 0.017476852983236313,
+ "learning_rate": 0.0006,
+ "loss": 4.2518720626831055,
+ "step": 4970
+ },
+ {
+ "epoch": 69.04192139737991,
+ "grad_norm": 0.017070338129997253,
+ "learning_rate": 0.0006,
+ "loss": 4.238024711608887,
+ "step": 4971
+ },
+ {
+ "epoch": 69.05589519650655,
+ "grad_norm": 0.01639373227953911,
+ "learning_rate": 0.0006,
+ "loss": 4.188558101654053,
+ "step": 4972
+ },
+ {
+ "epoch": 69.06986899563319,
+ "grad_norm": 0.01856113411486149,
+ "learning_rate": 0.0006,
+ "loss": 4.339556694030762,
+ "step": 4973
+ },
+ {
+ "epoch": 69.08384279475983,
+ "grad_norm": 0.01762578636407852,
+ "learning_rate": 0.0006,
+ "loss": 4.141518592834473,
+ "step": 4974
+ },
+ {
+ "epoch": 69.09781659388646,
+ "grad_norm": 0.015437949448823929,
+ "learning_rate": 0.0006,
+ "loss": 4.225461959838867,
+ "step": 4975
+ },
+ {
+ "epoch": 69.1117903930131,
+ "grad_norm": 0.015293503180146217,
+ "learning_rate": 0.0006,
+ "loss": 4.290884971618652,
+ "step": 4976
+ },
+ {
+ "epoch": 69.12576419213974,
+ "grad_norm": 0.016259880736470222,
+ "learning_rate": 0.0006,
+ "loss": 4.121919631958008,
+ "step": 4977
+ },
+ {
+ "epoch": 69.13973799126637,
+ "grad_norm": 0.017225584015250206,
+ "learning_rate": 0.0006,
+ "loss": 4.136716842651367,
+ "step": 4978
+ },
+ {
+ "epoch": 69.15371179039302,
+ "grad_norm": 0.01699289306998253,
+ "learning_rate": 0.0006,
+ "loss": 4.193601608276367,
+ "step": 4979
+ },
+ {
+ "epoch": 69.16768558951965,
+ "grad_norm": 0.015613161027431488,
+ "learning_rate": 0.0006,
+ "loss": 4.288212299346924,
+ "step": 4980
+ },
+ {
+ "epoch": 69.18165938864628,
+ "grad_norm": 0.01643933728337288,
+ "learning_rate": 0.0006,
+ "loss": 4.324707984924316,
+ "step": 4981
+ },
+ {
+ "epoch": 69.19563318777293,
+ "grad_norm": 0.01648905873298645,
+ "learning_rate": 0.0006,
+ "loss": 4.287806510925293,
+ "step": 4982
+ },
+ {
+ "epoch": 69.20960698689956,
+ "grad_norm": 0.015444950200617313,
+ "learning_rate": 0.0006,
+ "loss": 4.200622081756592,
+ "step": 4983
+ },
+ {
+ "epoch": 69.2235807860262,
+ "grad_norm": 0.016034258529543877,
+ "learning_rate": 0.0006,
+ "loss": 4.2193827629089355,
+ "step": 4984
+ },
+ {
+ "epoch": 69.23755458515284,
+ "grad_norm": 0.017984388396143913,
+ "learning_rate": 0.0006,
+ "loss": 4.18237829208374,
+ "step": 4985
+ },
+ {
+ "epoch": 69.25152838427948,
+ "grad_norm": 0.0195505041629076,
+ "learning_rate": 0.0006,
+ "loss": 4.216727256774902,
+ "step": 4986
+ },
+ {
+ "epoch": 69.26550218340611,
+ "grad_norm": 0.017779147252440453,
+ "learning_rate": 0.0006,
+ "loss": 4.278628826141357,
+ "step": 4987
+ },
+ {
+ "epoch": 69.27947598253274,
+ "grad_norm": 0.017119476571679115,
+ "learning_rate": 0.0006,
+ "loss": 4.2409138679504395,
+ "step": 4988
+ },
+ {
+ "epoch": 69.29344978165939,
+ "grad_norm": 0.016840364784002304,
+ "learning_rate": 0.0006,
+ "loss": 4.207575798034668,
+ "step": 4989
+ },
+ {
+ "epoch": 69.30742358078602,
+ "grad_norm": 0.014593573287129402,
+ "learning_rate": 0.0006,
+ "loss": 4.164905548095703,
+ "step": 4990
+ },
+ {
+ "epoch": 69.32139737991267,
+ "grad_norm": 0.013922630809247494,
+ "learning_rate": 0.0006,
+ "loss": 4.2696990966796875,
+ "step": 4991
+ },
+ {
+ "epoch": 69.3353711790393,
+ "grad_norm": 0.01577564887702465,
+ "learning_rate": 0.0006,
+ "loss": 4.201589107513428,
+ "step": 4992
+ },
+ {
+ "epoch": 69.34934497816595,
+ "grad_norm": 0.016392096877098083,
+ "learning_rate": 0.0006,
+ "loss": 4.248831748962402,
+ "step": 4993
+ },
+ {
+ "epoch": 69.36331877729258,
+ "grad_norm": 0.015961311757564545,
+ "learning_rate": 0.0006,
+ "loss": 4.181260108947754,
+ "step": 4994
+ },
+ {
+ "epoch": 69.37729257641921,
+ "grad_norm": 0.015350551344454288,
+ "learning_rate": 0.0006,
+ "loss": 4.230352878570557,
+ "step": 4995
+ },
+ {
+ "epoch": 69.39126637554585,
+ "grad_norm": 0.015870466828346252,
+ "learning_rate": 0.0006,
+ "loss": 4.298305511474609,
+ "step": 4996
+ },
+ {
+ "epoch": 69.40524017467249,
+ "grad_norm": 0.013716642744839191,
+ "learning_rate": 0.0006,
+ "loss": 4.220139503479004,
+ "step": 4997
+ },
+ {
+ "epoch": 69.41921397379913,
+ "grad_norm": 0.014431777410209179,
+ "learning_rate": 0.0006,
+ "loss": 4.288130760192871,
+ "step": 4998
+ },
+ {
+ "epoch": 69.43318777292576,
+ "grad_norm": 0.01426647324115038,
+ "learning_rate": 0.0006,
+ "loss": 4.295767784118652,
+ "step": 4999
+ },
+ {
+ "epoch": 69.44716157205241,
+ "grad_norm": 0.014286615885794163,
+ "learning_rate": 0.0006,
+ "loss": 4.38106632232666,
+ "step": 5000
+ },
+ {
+ "epoch": 69.46113537117904,
+ "grad_norm": 0.017990585416555405,
+ "learning_rate": 0.0006,
+ "loss": 4.1593017578125,
+ "step": 5001
+ },
+ {
+ "epoch": 69.47510917030567,
+ "grad_norm": 0.018846049904823303,
+ "learning_rate": 0.0006,
+ "loss": 4.216091632843018,
+ "step": 5002
+ },
+ {
+ "epoch": 69.48908296943232,
+ "grad_norm": 0.018179181963205338,
+ "learning_rate": 0.0006,
+ "loss": 4.240332126617432,
+ "step": 5003
+ },
+ {
+ "epoch": 69.50305676855895,
+ "grad_norm": 0.016693448647856712,
+ "learning_rate": 0.0006,
+ "loss": 4.256624221801758,
+ "step": 5004
+ },
+ {
+ "epoch": 69.5170305676856,
+ "grad_norm": 0.01718989387154579,
+ "learning_rate": 0.0006,
+ "loss": 4.300940990447998,
+ "step": 5005
+ },
+ {
+ "epoch": 69.53100436681223,
+ "grad_norm": 0.020236829295754433,
+ "learning_rate": 0.0006,
+ "loss": 4.289146423339844,
+ "step": 5006
+ },
+ {
+ "epoch": 69.54497816593886,
+ "grad_norm": 0.024860285222530365,
+ "learning_rate": 0.0006,
+ "loss": 4.216216087341309,
+ "step": 5007
+ },
+ {
+ "epoch": 69.5589519650655,
+ "grad_norm": 0.02559969387948513,
+ "learning_rate": 0.0006,
+ "loss": 4.262890815734863,
+ "step": 5008
+ },
+ {
+ "epoch": 69.57292576419214,
+ "grad_norm": 0.019180957227945328,
+ "learning_rate": 0.0006,
+ "loss": 4.230309009552002,
+ "step": 5009
+ },
+ {
+ "epoch": 69.58689956331878,
+ "grad_norm": 0.020442591980099678,
+ "learning_rate": 0.0006,
+ "loss": 4.291759490966797,
+ "step": 5010
+ },
+ {
+ "epoch": 69.60087336244541,
+ "grad_norm": 0.0212619137018919,
+ "learning_rate": 0.0006,
+ "loss": 4.3529839515686035,
+ "step": 5011
+ },
+ {
+ "epoch": 69.61484716157206,
+ "grad_norm": 0.02067175693809986,
+ "learning_rate": 0.0006,
+ "loss": 4.33197021484375,
+ "step": 5012
+ },
+ {
+ "epoch": 69.62882096069869,
+ "grad_norm": 0.01802872307598591,
+ "learning_rate": 0.0006,
+ "loss": 4.2101969718933105,
+ "step": 5013
+ },
+ {
+ "epoch": 69.64279475982532,
+ "grad_norm": 0.01787702925503254,
+ "learning_rate": 0.0006,
+ "loss": 4.270336627960205,
+ "step": 5014
+ },
+ {
+ "epoch": 69.65676855895197,
+ "grad_norm": 0.01586683839559555,
+ "learning_rate": 0.0006,
+ "loss": 4.222309589385986,
+ "step": 5015
+ },
+ {
+ "epoch": 69.6707423580786,
+ "grad_norm": 0.01717999391257763,
+ "learning_rate": 0.0006,
+ "loss": 4.253704071044922,
+ "step": 5016
+ },
+ {
+ "epoch": 69.68471615720524,
+ "grad_norm": 0.01594279147684574,
+ "learning_rate": 0.0006,
+ "loss": 4.344658851623535,
+ "step": 5017
+ },
+ {
+ "epoch": 69.69868995633188,
+ "grad_norm": 0.014419942162930965,
+ "learning_rate": 0.0006,
+ "loss": 4.2144880294799805,
+ "step": 5018
+ },
+ {
+ "epoch": 69.71266375545852,
+ "grad_norm": 0.015824224799871445,
+ "learning_rate": 0.0006,
+ "loss": 4.226252555847168,
+ "step": 5019
+ },
+ {
+ "epoch": 69.72663755458515,
+ "grad_norm": 0.013329979032278061,
+ "learning_rate": 0.0006,
+ "loss": 4.238587379455566,
+ "step": 5020
+ },
+ {
+ "epoch": 69.74061135371178,
+ "grad_norm": 0.013745979405939579,
+ "learning_rate": 0.0006,
+ "loss": 4.317209243774414,
+ "step": 5021
+ },
+ {
+ "epoch": 69.75458515283843,
+ "grad_norm": 0.014064167626202106,
+ "learning_rate": 0.0006,
+ "loss": 4.316982746124268,
+ "step": 5022
+ },
+ {
+ "epoch": 69.76855895196506,
+ "grad_norm": 0.014611301012337208,
+ "learning_rate": 0.0006,
+ "loss": 4.212728500366211,
+ "step": 5023
+ },
+ {
+ "epoch": 69.78253275109171,
+ "grad_norm": 0.014947416260838509,
+ "learning_rate": 0.0006,
+ "loss": 4.319550514221191,
+ "step": 5024
+ },
+ {
+ "epoch": 69.79650655021834,
+ "grad_norm": 0.014780526980757713,
+ "learning_rate": 0.0006,
+ "loss": 4.1888933181762695,
+ "step": 5025
+ },
+ {
+ "epoch": 69.81048034934499,
+ "grad_norm": 0.01611459068953991,
+ "learning_rate": 0.0006,
+ "loss": 4.239692211151123,
+ "step": 5026
+ },
+ {
+ "epoch": 69.82445414847162,
+ "grad_norm": 0.015395848080515862,
+ "learning_rate": 0.0006,
+ "loss": 4.308794021606445,
+ "step": 5027
+ },
+ {
+ "epoch": 69.83842794759825,
+ "grad_norm": 0.016113396733999252,
+ "learning_rate": 0.0006,
+ "loss": 4.328035354614258,
+ "step": 5028
+ },
+ {
+ "epoch": 69.8524017467249,
+ "grad_norm": 0.016386888921260834,
+ "learning_rate": 0.0006,
+ "loss": 4.272579193115234,
+ "step": 5029
+ },
+ {
+ "epoch": 69.86637554585153,
+ "grad_norm": 0.013675205409526825,
+ "learning_rate": 0.0006,
+ "loss": 4.226552963256836,
+ "step": 5030
+ },
+ {
+ "epoch": 69.88034934497817,
+ "grad_norm": 0.01472668256610632,
+ "learning_rate": 0.0006,
+ "loss": 4.224167346954346,
+ "step": 5031
+ },
+ {
+ "epoch": 69.8943231441048,
+ "grad_norm": 0.014958228915929794,
+ "learning_rate": 0.0006,
+ "loss": 4.232619285583496,
+ "step": 5032
+ },
+ {
+ "epoch": 69.90829694323143,
+ "grad_norm": 0.012926832772791386,
+ "learning_rate": 0.0006,
+ "loss": 4.27107048034668,
+ "step": 5033
+ },
+ {
+ "epoch": 69.92227074235808,
+ "grad_norm": 0.013138801790773869,
+ "learning_rate": 0.0006,
+ "loss": 4.288477420806885,
+ "step": 5034
+ },
+ {
+ "epoch": 69.93624454148471,
+ "grad_norm": 0.013415461406111717,
+ "learning_rate": 0.0006,
+ "loss": 4.2033257484436035,
+ "step": 5035
+ },
+ {
+ "epoch": 69.95021834061136,
+ "grad_norm": 0.013807233422994614,
+ "learning_rate": 0.0006,
+ "loss": 4.274764537811279,
+ "step": 5036
+ },
+ {
+ "epoch": 69.96419213973799,
+ "grad_norm": 0.013929193839430809,
+ "learning_rate": 0.0006,
+ "loss": 4.1854352951049805,
+ "step": 5037
+ },
+ {
+ "epoch": 69.97816593886463,
+ "grad_norm": 0.015497001819312572,
+ "learning_rate": 0.0006,
+ "loss": 4.226498126983643,
+ "step": 5038
+ },
+ {
+ "epoch": 69.99213973799127,
+ "grad_norm": 0.01874048449099064,
+ "learning_rate": 0.0006,
+ "loss": 4.304137706756592,
+ "step": 5039
+ },
+ {
+ "epoch": 70.0,
+ "grad_norm": 0.021745791658759117,
+ "learning_rate": 0.0006,
+ "loss": 4.206809997558594,
+ "step": 5040
+ },
+ {
+ "epoch": 70.0,
+ "eval_loss": 4.675539970397949,
+ "eval_runtime": 56.9034,
+ "eval_samples_per_second": 42.915,
+ "eval_steps_per_second": 1.353,
+ "step": 5040
+ },
+ {
+ "epoch": 70.01397379912663,
+ "grad_norm": 0.02124839462339878,
+ "learning_rate": 0.0006,
+ "loss": 4.198199272155762,
+ "step": 5041
+ },
+ {
+ "epoch": 70.02794759825328,
+ "grad_norm": 0.019432688131928444,
+ "learning_rate": 0.0006,
+ "loss": 4.187261581420898,
+ "step": 5042
+ },
+ {
+ "epoch": 70.04192139737991,
+ "grad_norm": 0.01769290305674076,
+ "learning_rate": 0.0006,
+ "loss": 4.155284881591797,
+ "step": 5043
+ },
+ {
+ "epoch": 70.05589519650655,
+ "grad_norm": 0.01751520484685898,
+ "learning_rate": 0.0006,
+ "loss": 4.386857986450195,
+ "step": 5044
+ },
+ {
+ "epoch": 70.06986899563319,
+ "grad_norm": 0.018138699233531952,
+ "learning_rate": 0.0006,
+ "loss": 4.117342472076416,
+ "step": 5045
+ },
+ {
+ "epoch": 70.08384279475983,
+ "grad_norm": 0.016331054270267487,
+ "learning_rate": 0.0006,
+ "loss": 4.225295066833496,
+ "step": 5046
+ },
+ {
+ "epoch": 70.09781659388646,
+ "grad_norm": 0.01658729277551174,
+ "learning_rate": 0.0006,
+ "loss": 4.143835067749023,
+ "step": 5047
+ },
+ {
+ "epoch": 70.1117903930131,
+ "grad_norm": 0.016367698088288307,
+ "learning_rate": 0.0006,
+ "loss": 4.1469268798828125,
+ "step": 5048
+ },
+ {
+ "epoch": 70.12576419213974,
+ "grad_norm": 0.015151629224419594,
+ "learning_rate": 0.0006,
+ "loss": 4.117629051208496,
+ "step": 5049
+ },
+ {
+ "epoch": 70.13973799126637,
+ "grad_norm": 0.016434503719210625,
+ "learning_rate": 0.0006,
+ "loss": 4.107848644256592,
+ "step": 5050
+ },
+ {
+ "epoch": 70.15371179039302,
+ "grad_norm": 0.017837485298514366,
+ "learning_rate": 0.0006,
+ "loss": 4.10837459564209,
+ "step": 5051
+ },
+ {
+ "epoch": 70.16768558951965,
+ "grad_norm": 0.0165961142629385,
+ "learning_rate": 0.0006,
+ "loss": 4.169414043426514,
+ "step": 5052
+ },
+ {
+ "epoch": 70.18165938864628,
+ "grad_norm": 0.0157014112919569,
+ "learning_rate": 0.0006,
+ "loss": 4.0629706382751465,
+ "step": 5053
+ },
+ {
+ "epoch": 70.19563318777293,
+ "grad_norm": 0.01581883803009987,
+ "learning_rate": 0.0006,
+ "loss": 4.220012664794922,
+ "step": 5054
+ },
+ {
+ "epoch": 70.20960698689956,
+ "grad_norm": 0.015285917557775974,
+ "learning_rate": 0.0006,
+ "loss": 4.275337219238281,
+ "step": 5055
+ },
+ {
+ "epoch": 70.2235807860262,
+ "grad_norm": 0.014984549954533577,
+ "learning_rate": 0.0006,
+ "loss": 4.284764289855957,
+ "step": 5056
+ },
+ {
+ "epoch": 70.23755458515284,
+ "grad_norm": 0.015172012150287628,
+ "learning_rate": 0.0006,
+ "loss": 4.234498023986816,
+ "step": 5057
+ },
+ {
+ "epoch": 70.25152838427948,
+ "grad_norm": 0.016846632584929466,
+ "learning_rate": 0.0006,
+ "loss": 4.181689262390137,
+ "step": 5058
+ },
+ {
+ "epoch": 70.26550218340611,
+ "grad_norm": 0.01920711249113083,
+ "learning_rate": 0.0006,
+ "loss": 4.235280513763428,
+ "step": 5059
+ },
+ {
+ "epoch": 70.27947598253274,
+ "grad_norm": 0.017765291035175323,
+ "learning_rate": 0.0006,
+ "loss": 4.238955497741699,
+ "step": 5060
+ },
+ {
+ "epoch": 70.29344978165939,
+ "grad_norm": 0.015436639077961445,
+ "learning_rate": 0.0006,
+ "loss": 4.232516288757324,
+ "step": 5061
+ },
+ {
+ "epoch": 70.30742358078602,
+ "grad_norm": 0.015660086646676064,
+ "learning_rate": 0.0006,
+ "loss": 4.260349273681641,
+ "step": 5062
+ },
+ {
+ "epoch": 70.32139737991267,
+ "grad_norm": 0.016095709055662155,
+ "learning_rate": 0.0006,
+ "loss": 4.168505668640137,
+ "step": 5063
+ },
+ {
+ "epoch": 70.3353711790393,
+ "grad_norm": 0.015001940540969372,
+ "learning_rate": 0.0006,
+ "loss": 4.161739349365234,
+ "step": 5064
+ },
+ {
+ "epoch": 70.34934497816595,
+ "grad_norm": 0.014668774791061878,
+ "learning_rate": 0.0006,
+ "loss": 4.181029796600342,
+ "step": 5065
+ },
+ {
+ "epoch": 70.36331877729258,
+ "grad_norm": 0.013266284950077534,
+ "learning_rate": 0.0006,
+ "loss": 4.3052873611450195,
+ "step": 5066
+ },
+ {
+ "epoch": 70.37729257641921,
+ "grad_norm": 0.015396138653159142,
+ "learning_rate": 0.0006,
+ "loss": 4.34413480758667,
+ "step": 5067
+ },
+ {
+ "epoch": 70.39126637554585,
+ "grad_norm": 0.015511523932218552,
+ "learning_rate": 0.0006,
+ "loss": 4.235037803649902,
+ "step": 5068
+ },
+ {
+ "epoch": 70.40524017467249,
+ "grad_norm": 0.014825738966464996,
+ "learning_rate": 0.0006,
+ "loss": 4.262448310852051,
+ "step": 5069
+ },
+ {
+ "epoch": 70.41921397379913,
+ "grad_norm": 0.015309952199459076,
+ "learning_rate": 0.0006,
+ "loss": 4.258711814880371,
+ "step": 5070
+ },
+ {
+ "epoch": 70.43318777292576,
+ "grad_norm": 0.017614055424928665,
+ "learning_rate": 0.0006,
+ "loss": 4.286028861999512,
+ "step": 5071
+ },
+ {
+ "epoch": 70.44716157205241,
+ "grad_norm": 0.0166039876639843,
+ "learning_rate": 0.0006,
+ "loss": 4.149197578430176,
+ "step": 5072
+ },
+ {
+ "epoch": 70.46113537117904,
+ "grad_norm": 0.016911419108510017,
+ "learning_rate": 0.0006,
+ "loss": 4.271981239318848,
+ "step": 5073
+ },
+ {
+ "epoch": 70.47510917030567,
+ "grad_norm": 0.016806820407509804,
+ "learning_rate": 0.0006,
+ "loss": 4.213315010070801,
+ "step": 5074
+ },
+ {
+ "epoch": 70.48908296943232,
+ "grad_norm": 0.015221742913126945,
+ "learning_rate": 0.0006,
+ "loss": 4.171091079711914,
+ "step": 5075
+ },
+ {
+ "epoch": 70.50305676855895,
+ "grad_norm": 0.014175347983837128,
+ "learning_rate": 0.0006,
+ "loss": 4.185309410095215,
+ "step": 5076
+ },
+ {
+ "epoch": 70.5170305676856,
+ "grad_norm": 0.013897798955440521,
+ "learning_rate": 0.0006,
+ "loss": 4.250119209289551,
+ "step": 5077
+ },
+ {
+ "epoch": 70.53100436681223,
+ "grad_norm": 0.014791003428399563,
+ "learning_rate": 0.0006,
+ "loss": 4.2504682540893555,
+ "step": 5078
+ },
+ {
+ "epoch": 70.54497816593886,
+ "grad_norm": 0.016700128093361855,
+ "learning_rate": 0.0006,
+ "loss": 4.235134124755859,
+ "step": 5079
+ },
+ {
+ "epoch": 70.5589519650655,
+ "grad_norm": 0.016679363325238228,
+ "learning_rate": 0.0006,
+ "loss": 4.32028865814209,
+ "step": 5080
+ },
+ {
+ "epoch": 70.57292576419214,
+ "grad_norm": 0.015117401257157326,
+ "learning_rate": 0.0006,
+ "loss": 4.210995674133301,
+ "step": 5081
+ },
+ {
+ "epoch": 70.58689956331878,
+ "grad_norm": 0.014441263861954212,
+ "learning_rate": 0.0006,
+ "loss": 4.27849006652832,
+ "step": 5082
+ },
+ {
+ "epoch": 70.60087336244541,
+ "grad_norm": 0.01523780357092619,
+ "learning_rate": 0.0006,
+ "loss": 4.197360992431641,
+ "step": 5083
+ },
+ {
+ "epoch": 70.61484716157206,
+ "grad_norm": 0.016201738268136978,
+ "learning_rate": 0.0006,
+ "loss": 4.28239107131958,
+ "step": 5084
+ },
+ {
+ "epoch": 70.62882096069869,
+ "grad_norm": 0.016003292053937912,
+ "learning_rate": 0.0006,
+ "loss": 4.209715843200684,
+ "step": 5085
+ },
+ {
+ "epoch": 70.64279475982532,
+ "grad_norm": 0.01472893450409174,
+ "learning_rate": 0.0006,
+ "loss": 4.169078826904297,
+ "step": 5086
+ },
+ {
+ "epoch": 70.65676855895197,
+ "grad_norm": 0.01604326069355011,
+ "learning_rate": 0.0006,
+ "loss": 4.090967178344727,
+ "step": 5087
+ },
+ {
+ "epoch": 70.6707423580786,
+ "grad_norm": 0.01598356105387211,
+ "learning_rate": 0.0006,
+ "loss": 4.16630220413208,
+ "step": 5088
+ },
+ {
+ "epoch": 70.68471615720524,
+ "grad_norm": 0.016988780349493027,
+ "learning_rate": 0.0006,
+ "loss": 4.117822170257568,
+ "step": 5089
+ },
+ {
+ "epoch": 70.69868995633188,
+ "grad_norm": 0.018582893535494804,
+ "learning_rate": 0.0006,
+ "loss": 4.192902088165283,
+ "step": 5090
+ },
+ {
+ "epoch": 70.71266375545852,
+ "grad_norm": 0.01793155074119568,
+ "learning_rate": 0.0006,
+ "loss": 4.314806938171387,
+ "step": 5091
+ },
+ {
+ "epoch": 70.72663755458515,
+ "grad_norm": 0.016286427155137062,
+ "learning_rate": 0.0006,
+ "loss": 4.2137556076049805,
+ "step": 5092
+ },
+ {
+ "epoch": 70.74061135371178,
+ "grad_norm": 0.014718376100063324,
+ "learning_rate": 0.0006,
+ "loss": 4.3073015213012695,
+ "step": 5093
+ },
+ {
+ "epoch": 70.75458515283843,
+ "grad_norm": 0.014675077050924301,
+ "learning_rate": 0.0006,
+ "loss": 4.160961151123047,
+ "step": 5094
+ },
+ {
+ "epoch": 70.76855895196506,
+ "grad_norm": 0.015550913289189339,
+ "learning_rate": 0.0006,
+ "loss": 4.322887420654297,
+ "step": 5095
+ },
+ {
+ "epoch": 70.78253275109171,
+ "grad_norm": 0.01453235000371933,
+ "learning_rate": 0.0006,
+ "loss": 4.2915873527526855,
+ "step": 5096
+ },
+ {
+ "epoch": 70.79650655021834,
+ "grad_norm": 0.01459164172410965,
+ "learning_rate": 0.0006,
+ "loss": 4.247288227081299,
+ "step": 5097
+ },
+ {
+ "epoch": 70.81048034934499,
+ "grad_norm": 0.015395080670714378,
+ "learning_rate": 0.0006,
+ "loss": 4.360955715179443,
+ "step": 5098
+ },
+ {
+ "epoch": 70.82445414847162,
+ "grad_norm": 0.01317683607339859,
+ "learning_rate": 0.0006,
+ "loss": 4.181969165802002,
+ "step": 5099
+ },
+ {
+ "epoch": 70.83842794759825,
+ "grad_norm": 0.013917932286858559,
+ "learning_rate": 0.0006,
+ "loss": 4.264626979827881,
+ "step": 5100
+ },
+ {
+ "epoch": 70.8524017467249,
+ "grad_norm": 0.015268695540726185,
+ "learning_rate": 0.0006,
+ "loss": 4.287857532501221,
+ "step": 5101
+ },
+ {
+ "epoch": 70.86637554585153,
+ "grad_norm": 0.015531329438090324,
+ "learning_rate": 0.0006,
+ "loss": 4.294188499450684,
+ "step": 5102
+ },
+ {
+ "epoch": 70.88034934497817,
+ "grad_norm": 0.015607129782438278,
+ "learning_rate": 0.0006,
+ "loss": 4.331040859222412,
+ "step": 5103
+ },
+ {
+ "epoch": 70.8943231441048,
+ "grad_norm": 0.015297149308025837,
+ "learning_rate": 0.0006,
+ "loss": 4.243650436401367,
+ "step": 5104
+ },
+ {
+ "epoch": 70.90829694323143,
+ "grad_norm": 0.014790025539696217,
+ "learning_rate": 0.0006,
+ "loss": 4.159859657287598,
+ "step": 5105
+ },
+ {
+ "epoch": 70.92227074235808,
+ "grad_norm": 0.013466274365782738,
+ "learning_rate": 0.0006,
+ "loss": 4.221992492675781,
+ "step": 5106
+ },
+ {
+ "epoch": 70.93624454148471,
+ "grad_norm": 0.01361898984760046,
+ "learning_rate": 0.0006,
+ "loss": 4.278751373291016,
+ "step": 5107
+ },
+ {
+ "epoch": 70.95021834061136,
+ "grad_norm": 0.01325751468539238,
+ "learning_rate": 0.0006,
+ "loss": 4.166001319885254,
+ "step": 5108
+ },
+ {
+ "epoch": 70.96419213973799,
+ "grad_norm": 0.012727087363600731,
+ "learning_rate": 0.0006,
+ "loss": 4.219997406005859,
+ "step": 5109
+ },
+ {
+ "epoch": 70.97816593886463,
+ "grad_norm": 0.013966867700219154,
+ "learning_rate": 0.0006,
+ "loss": 4.207491397857666,
+ "step": 5110
+ },
+ {
+ "epoch": 70.99213973799127,
+ "grad_norm": 0.015053587965667248,
+ "learning_rate": 0.0006,
+ "loss": 4.343245983123779,
+ "step": 5111
+ },
+ {
+ "epoch": 71.0,
+ "grad_norm": 0.017723368480801582,
+ "learning_rate": 0.0006,
+ "loss": 4.356122016906738,
+ "step": 5112
+ },
+ {
+ "epoch": 71.0,
+ "eval_loss": 4.598303318023682,
+ "eval_runtime": 56.7521,
+ "eval_samples_per_second": 43.029,
+ "eval_steps_per_second": 1.357,
+ "step": 5112
+ },
+ {
+ "epoch": 71.01397379912663,
+ "grad_norm": 0.018996279686689377,
+ "learning_rate": 0.0006,
+ "loss": 4.234772205352783,
+ "step": 5113
+ },
+ {
+ "epoch": 71.02794759825328,
+ "grad_norm": 0.02041555754840374,
+ "learning_rate": 0.0006,
+ "loss": 4.275816917419434,
+ "step": 5114
+ },
+ {
+ "epoch": 71.04192139737991,
+ "grad_norm": 0.021955285221338272,
+ "learning_rate": 0.0006,
+ "loss": 4.166038513183594,
+ "step": 5115
+ },
+ {
+ "epoch": 71.05589519650655,
+ "grad_norm": 0.021308740600943565,
+ "learning_rate": 0.0006,
+ "loss": 4.13063907623291,
+ "step": 5116
+ },
+ {
+ "epoch": 71.06986899563319,
+ "grad_norm": 0.01747054047882557,
+ "learning_rate": 0.0006,
+ "loss": 4.200522422790527,
+ "step": 5117
+ },
+ {
+ "epoch": 71.08384279475983,
+ "grad_norm": 0.018544532358646393,
+ "learning_rate": 0.0006,
+ "loss": 4.173915386199951,
+ "step": 5118
+ },
+ {
+ "epoch": 71.09781659388646,
+ "grad_norm": 0.0182361900806427,
+ "learning_rate": 0.0006,
+ "loss": 4.27392578125,
+ "step": 5119
+ },
+ {
+ "epoch": 71.1117903930131,
+ "grad_norm": 0.017062073573470116,
+ "learning_rate": 0.0006,
+ "loss": 4.219576358795166,
+ "step": 5120
+ },
+ {
+ "epoch": 71.12576419213974,
+ "grad_norm": 0.016289666295051575,
+ "learning_rate": 0.0006,
+ "loss": 4.133376598358154,
+ "step": 5121
+ },
+ {
+ "epoch": 71.13973799126637,
+ "grad_norm": 0.014864642173051834,
+ "learning_rate": 0.0006,
+ "loss": 4.168276786804199,
+ "step": 5122
+ },
+ {
+ "epoch": 71.15371179039302,
+ "grad_norm": 0.015309540554881096,
+ "learning_rate": 0.0006,
+ "loss": 4.175103187561035,
+ "step": 5123
+ },
+ {
+ "epoch": 71.16768558951965,
+ "grad_norm": 0.01556059904396534,
+ "learning_rate": 0.0006,
+ "loss": 4.289899826049805,
+ "step": 5124
+ },
+ {
+ "epoch": 71.18165938864628,
+ "grad_norm": 0.01479957066476345,
+ "learning_rate": 0.0006,
+ "loss": 4.341495037078857,
+ "step": 5125
+ },
+ {
+ "epoch": 71.19563318777293,
+ "grad_norm": 0.014762179926037788,
+ "learning_rate": 0.0006,
+ "loss": 4.252333641052246,
+ "step": 5126
+ },
+ {
+ "epoch": 71.20960698689956,
+ "grad_norm": 0.014897473156452179,
+ "learning_rate": 0.0006,
+ "loss": 4.2244181632995605,
+ "step": 5127
+ },
+ {
+ "epoch": 71.2235807860262,
+ "grad_norm": 0.012990163639187813,
+ "learning_rate": 0.0006,
+ "loss": 4.237288475036621,
+ "step": 5128
+ },
+ {
+ "epoch": 71.23755458515284,
+ "grad_norm": 0.012858416885137558,
+ "learning_rate": 0.0006,
+ "loss": 4.265738487243652,
+ "step": 5129
+ },
+ {
+ "epoch": 71.25152838427948,
+ "grad_norm": 0.013454788364470005,
+ "learning_rate": 0.0006,
+ "loss": 4.289131164550781,
+ "step": 5130
+ },
+ {
+ "epoch": 71.26550218340611,
+ "grad_norm": 0.013717238791286945,
+ "learning_rate": 0.0006,
+ "loss": 4.2844014167785645,
+ "step": 5131
+ },
+ {
+ "epoch": 71.27947598253274,
+ "grad_norm": 0.0135411461815238,
+ "learning_rate": 0.0006,
+ "loss": 4.139082908630371,
+ "step": 5132
+ },
+ {
+ "epoch": 71.29344978165939,
+ "grad_norm": 0.01406062301248312,
+ "learning_rate": 0.0006,
+ "loss": 4.226719856262207,
+ "step": 5133
+ },
+ {
+ "epoch": 71.30742358078602,
+ "grad_norm": 0.014027190394699574,
+ "learning_rate": 0.0006,
+ "loss": 4.191699981689453,
+ "step": 5134
+ },
+ {
+ "epoch": 71.32139737991267,
+ "grad_norm": 0.013809907250106335,
+ "learning_rate": 0.0006,
+ "loss": 4.242542266845703,
+ "step": 5135
+ },
+ {
+ "epoch": 71.3353711790393,
+ "grad_norm": 0.013507804833352566,
+ "learning_rate": 0.0006,
+ "loss": 4.068206310272217,
+ "step": 5136
+ },
+ {
+ "epoch": 71.34934497816595,
+ "grad_norm": 0.01432533748447895,
+ "learning_rate": 0.0006,
+ "loss": 4.378222942352295,
+ "step": 5137
+ },
+ {
+ "epoch": 71.36331877729258,
+ "grad_norm": 0.01753002591431141,
+ "learning_rate": 0.0006,
+ "loss": 4.174706935882568,
+ "step": 5138
+ },
+ {
+ "epoch": 71.37729257641921,
+ "grad_norm": 0.020048344507813454,
+ "learning_rate": 0.0006,
+ "loss": 4.19920015335083,
+ "step": 5139
+ },
+ {
+ "epoch": 71.39126637554585,
+ "grad_norm": 0.018551401793956757,
+ "learning_rate": 0.0006,
+ "loss": 4.175182342529297,
+ "step": 5140
+ },
+ {
+ "epoch": 71.40524017467249,
+ "grad_norm": 0.01598959229886532,
+ "learning_rate": 0.0006,
+ "loss": 4.2305169105529785,
+ "step": 5141
+ },
+ {
+ "epoch": 71.41921397379913,
+ "grad_norm": 0.016029560938477516,
+ "learning_rate": 0.0006,
+ "loss": 4.230090618133545,
+ "step": 5142
+ },
+ {
+ "epoch": 71.43318777292576,
+ "grad_norm": 0.01595657505095005,
+ "learning_rate": 0.0006,
+ "loss": 4.138402938842773,
+ "step": 5143
+ },
+ {
+ "epoch": 71.44716157205241,
+ "grad_norm": 0.01623230241239071,
+ "learning_rate": 0.0006,
+ "loss": 4.196106433868408,
+ "step": 5144
+ },
+ {
+ "epoch": 71.46113537117904,
+ "grad_norm": 0.014057597145438194,
+ "learning_rate": 0.0006,
+ "loss": 4.2599897384643555,
+ "step": 5145
+ },
+ {
+ "epoch": 71.47510917030567,
+ "grad_norm": 0.017161158844828606,
+ "learning_rate": 0.0006,
+ "loss": 4.183909893035889,
+ "step": 5146
+ },
+ {
+ "epoch": 71.48908296943232,
+ "grad_norm": 0.017038581892848015,
+ "learning_rate": 0.0006,
+ "loss": 4.373933792114258,
+ "step": 5147
+ },
+ {
+ "epoch": 71.50305676855895,
+ "grad_norm": 0.015851367264986038,
+ "learning_rate": 0.0006,
+ "loss": 4.2620086669921875,
+ "step": 5148
+ },
+ {
+ "epoch": 71.5170305676856,
+ "grad_norm": 0.014598303474485874,
+ "learning_rate": 0.0006,
+ "loss": 4.197151184082031,
+ "step": 5149
+ },
+ {
+ "epoch": 71.53100436681223,
+ "grad_norm": 0.01339802611619234,
+ "learning_rate": 0.0006,
+ "loss": 4.110706329345703,
+ "step": 5150
+ },
+ {
+ "epoch": 71.54497816593886,
+ "grad_norm": 0.015567775815725327,
+ "learning_rate": 0.0006,
+ "loss": 4.1210246086120605,
+ "step": 5151
+ },
+ {
+ "epoch": 71.5589519650655,
+ "grad_norm": 0.015616007149219513,
+ "learning_rate": 0.0006,
+ "loss": 4.129796028137207,
+ "step": 5152
+ },
+ {
+ "epoch": 71.57292576419214,
+ "grad_norm": 0.014198964461684227,
+ "learning_rate": 0.0006,
+ "loss": 4.193232536315918,
+ "step": 5153
+ },
+ {
+ "epoch": 71.58689956331878,
+ "grad_norm": 0.015226383693516254,
+ "learning_rate": 0.0006,
+ "loss": 4.223779678344727,
+ "step": 5154
+ },
+ {
+ "epoch": 71.60087336244541,
+ "grad_norm": 0.015610109083354473,
+ "learning_rate": 0.0006,
+ "loss": 4.223179817199707,
+ "step": 5155
+ },
+ {
+ "epoch": 71.61484716157206,
+ "grad_norm": 0.016034303233027458,
+ "learning_rate": 0.0006,
+ "loss": 4.196593284606934,
+ "step": 5156
+ },
+ {
+ "epoch": 71.62882096069869,
+ "grad_norm": 0.015648581087589264,
+ "learning_rate": 0.0006,
+ "loss": 4.341497421264648,
+ "step": 5157
+ },
+ {
+ "epoch": 71.64279475982532,
+ "grad_norm": 0.014117540791630745,
+ "learning_rate": 0.0006,
+ "loss": 4.194307327270508,
+ "step": 5158
+ },
+ {
+ "epoch": 71.65676855895197,
+ "grad_norm": 0.015612686984241009,
+ "learning_rate": 0.0006,
+ "loss": 4.290152549743652,
+ "step": 5159
+ },
+ {
+ "epoch": 71.6707423580786,
+ "grad_norm": 0.01722259446978569,
+ "learning_rate": 0.0006,
+ "loss": 4.258018493652344,
+ "step": 5160
+ },
+ {
+ "epoch": 71.68471615720524,
+ "grad_norm": 0.015923351049423218,
+ "learning_rate": 0.0006,
+ "loss": 4.21684455871582,
+ "step": 5161
+ },
+ {
+ "epoch": 71.69868995633188,
+ "grad_norm": 0.013663450255990028,
+ "learning_rate": 0.0006,
+ "loss": 4.175601959228516,
+ "step": 5162
+ },
+ {
+ "epoch": 71.71266375545852,
+ "grad_norm": 0.017075596377253532,
+ "learning_rate": 0.0006,
+ "loss": 4.205138683319092,
+ "step": 5163
+ },
+ {
+ "epoch": 71.72663755458515,
+ "grad_norm": 0.01723487675189972,
+ "learning_rate": 0.0006,
+ "loss": 4.08839225769043,
+ "step": 5164
+ },
+ {
+ "epoch": 71.74061135371178,
+ "grad_norm": 0.017355505377054214,
+ "learning_rate": 0.0006,
+ "loss": 4.165130615234375,
+ "step": 5165
+ },
+ {
+ "epoch": 71.75458515283843,
+ "grad_norm": 0.01868261583149433,
+ "learning_rate": 0.0006,
+ "loss": 4.16977071762085,
+ "step": 5166
+ },
+ {
+ "epoch": 71.76855895196506,
+ "grad_norm": 0.01897321827709675,
+ "learning_rate": 0.0006,
+ "loss": 4.173497200012207,
+ "step": 5167
+ },
+ {
+ "epoch": 71.78253275109171,
+ "grad_norm": 0.02045825682580471,
+ "learning_rate": 0.0006,
+ "loss": 4.176192283630371,
+ "step": 5168
+ },
+ {
+ "epoch": 71.79650655021834,
+ "grad_norm": 0.022895624861121178,
+ "learning_rate": 0.0006,
+ "loss": 4.163009166717529,
+ "step": 5169
+ },
+ {
+ "epoch": 71.81048034934499,
+ "grad_norm": 0.02062244527041912,
+ "learning_rate": 0.0006,
+ "loss": 4.156868934631348,
+ "step": 5170
+ },
+ {
+ "epoch": 71.82445414847162,
+ "grad_norm": 0.0176058579236269,
+ "learning_rate": 0.0006,
+ "loss": 4.154374122619629,
+ "step": 5171
+ },
+ {
+ "epoch": 71.83842794759825,
+ "grad_norm": 0.018049508333206177,
+ "learning_rate": 0.0006,
+ "loss": 4.296828269958496,
+ "step": 5172
+ },
+ {
+ "epoch": 71.8524017467249,
+ "grad_norm": 0.019466130062937737,
+ "learning_rate": 0.0006,
+ "loss": 4.19450569152832,
+ "step": 5173
+ },
+ {
+ "epoch": 71.86637554585153,
+ "grad_norm": 0.01838165521621704,
+ "learning_rate": 0.0006,
+ "loss": 4.11812686920166,
+ "step": 5174
+ },
+ {
+ "epoch": 71.88034934497817,
+ "grad_norm": 0.01608051173388958,
+ "learning_rate": 0.0006,
+ "loss": 4.240464210510254,
+ "step": 5175
+ },
+ {
+ "epoch": 71.8943231441048,
+ "grad_norm": 0.015807967633008957,
+ "learning_rate": 0.0006,
+ "loss": 4.275581359863281,
+ "step": 5176
+ },
+ {
+ "epoch": 71.90829694323143,
+ "grad_norm": 0.017366575077176094,
+ "learning_rate": 0.0006,
+ "loss": 4.064750671386719,
+ "step": 5177
+ },
+ {
+ "epoch": 71.92227074235808,
+ "grad_norm": 0.016241202130913734,
+ "learning_rate": 0.0006,
+ "loss": 4.203697681427002,
+ "step": 5178
+ },
+ {
+ "epoch": 71.93624454148471,
+ "grad_norm": 0.019105732440948486,
+ "learning_rate": 0.0006,
+ "loss": 4.140021800994873,
+ "step": 5179
+ },
+ {
+ "epoch": 71.95021834061136,
+ "grad_norm": 0.018492674455046654,
+ "learning_rate": 0.0006,
+ "loss": 4.2534074783325195,
+ "step": 5180
+ },
+ {
+ "epoch": 71.96419213973799,
+ "grad_norm": 0.019301170483231544,
+ "learning_rate": 0.0006,
+ "loss": 4.381392478942871,
+ "step": 5181
+ },
+ {
+ "epoch": 71.97816593886463,
+ "grad_norm": 0.019571911543607712,
+ "learning_rate": 0.0006,
+ "loss": 4.131056785583496,
+ "step": 5182
+ },
+ {
+ "epoch": 71.99213973799127,
+ "grad_norm": 0.01859748177230358,
+ "learning_rate": 0.0006,
+ "loss": 4.278930187225342,
+ "step": 5183
+ },
+ {
+ "epoch": 72.0,
+ "grad_norm": 0.02017994411289692,
+ "learning_rate": 0.0006,
+ "loss": 4.116464614868164,
+ "step": 5184
+ },
+ {
+ "epoch": 72.0,
+ "eval_loss": 4.60784912109375,
+ "eval_runtime": 56.3242,
+ "eval_samples_per_second": 43.356,
+ "eval_steps_per_second": 1.367,
+ "step": 5184
+ },
+ {
+ "epoch": 72.01397379912663,
+ "grad_norm": 0.030090296640992165,
+ "learning_rate": 0.0006,
+ "loss": 4.218588352203369,
+ "step": 5185
+ },
+ {
+ "epoch": 72.02794759825328,
+ "grad_norm": 1.330443024635315,
+ "learning_rate": 0.0006,
+ "loss": 4.417870044708252,
+ "step": 5186
+ },
+ {
+ "epoch": 72.04192139737991,
+ "grad_norm": 0.6991518139839172,
+ "learning_rate": 0.0006,
+ "loss": 8.414234161376953,
+ "step": 5187
+ },
+ {
+ "epoch": 72.05589519650655,
+ "grad_norm": 0.2631438374519348,
+ "learning_rate": 0.0006,
+ "loss": 7.41729736328125,
+ "step": 5188
+ },
+ {
+ "epoch": 72.06986899563319,
+ "grad_norm": 0.12391754984855652,
+ "learning_rate": 0.0006,
+ "loss": 7.209164619445801,
+ "step": 5189
+ },
+ {
+ "epoch": 72.08384279475983,
+ "grad_norm": 0.13519170880317688,
+ "learning_rate": 0.0006,
+ "loss": 6.943539619445801,
+ "step": 5190
+ },
+ {
+ "epoch": 72.09781659388646,
+ "grad_norm": 1.258439302444458,
+ "learning_rate": 0.0006,
+ "loss": 10.267714500427246,
+ "step": 5191
+ },
+ {
+ "epoch": 72.1117903930131,
+ "grad_norm": 1.0296908617019653,
+ "learning_rate": 0.0006,
+ "loss": 9.51768970489502,
+ "step": 5192
+ },
+ {
+ "epoch": 72.12576419213974,
+ "grad_norm": 0.12273016571998596,
+ "learning_rate": 0.0006,
+ "loss": 7.634477615356445,
+ "step": 5193
+ },
+ {
+ "epoch": 72.13973799126637,
+ "grad_norm": 0.06170198693871498,
+ "learning_rate": 0.0006,
+ "loss": 7.592115879058838,
+ "step": 5194
+ },
+ {
+ "epoch": 72.15371179039302,
+ "grad_norm": 0.06295108795166016,
+ "learning_rate": 0.0006,
+ "loss": 7.576314449310303,
+ "step": 5195
+ },
+ {
+ "epoch": 72.16768558951965,
+ "grad_norm": 0.0666993111371994,
+ "learning_rate": 0.0006,
+ "loss": 7.53862190246582,
+ "step": 5196
+ },
+ {
+ "epoch": 72.18165938864628,
+ "grad_norm": 0.052215222269296646,
+ "learning_rate": 0.0006,
+ "loss": 7.502930641174316,
+ "step": 5197
+ },
+ {
+ "epoch": 72.19563318777293,
+ "grad_norm": 0.04978961497545242,
+ "learning_rate": 0.0006,
+ "loss": 7.471630096435547,
+ "step": 5198
+ },
+ {
+ "epoch": 72.20960698689956,
+ "grad_norm": 0.07275046408176422,
+ "learning_rate": 0.0006,
+ "loss": 7.496924877166748,
+ "step": 5199
+ },
+ {
+ "epoch": 72.2235807860262,
+ "grad_norm": 0.04491819068789482,
+ "learning_rate": 0.0006,
+ "loss": 7.424494743347168,
+ "step": 5200
+ },
+ {
+ "epoch": 72.23755458515284,
+ "grad_norm": 0.043718695640563965,
+ "learning_rate": 0.0006,
+ "loss": 7.441090106964111,
+ "step": 5201
+ },
+ {
+ "epoch": 72.25152838427948,
+ "grad_norm": 0.04483497887849808,
+ "learning_rate": 0.0006,
+ "loss": 7.3753862380981445,
+ "step": 5202
+ },
+ {
+ "epoch": 72.26550218340611,
+ "grad_norm": 0.03463944420218468,
+ "learning_rate": 0.0006,
+ "loss": 7.375782012939453,
+ "step": 5203
+ },
+ {
+ "epoch": 72.27947598253274,
+ "grad_norm": 0.0397045724093914,
+ "learning_rate": 0.0006,
+ "loss": 7.358898639678955,
+ "step": 5204
+ },
+ {
+ "epoch": 72.29344978165939,
+ "grad_norm": 0.029339438304305077,
+ "learning_rate": 0.0006,
+ "loss": 7.3774871826171875,
+ "step": 5205
+ },
+ {
+ "epoch": 72.30742358078602,
+ "grad_norm": 0.025469699874520302,
+ "learning_rate": 0.0006,
+ "loss": 7.324203014373779,
+ "step": 5206
+ },
+ {
+ "epoch": 72.32139737991267,
+ "grad_norm": 0.030497785657644272,
+ "learning_rate": 0.0006,
+ "loss": 7.329351902008057,
+ "step": 5207
+ },
+ {
+ "epoch": 72.3353711790393,
+ "grad_norm": 0.03382078558206558,
+ "learning_rate": 0.0006,
+ "loss": 7.289940357208252,
+ "step": 5208
+ },
+ {
+ "epoch": 72.34934497816595,
+ "grad_norm": 0.019869405776262283,
+ "learning_rate": 0.0006,
+ "loss": 7.272582054138184,
+ "step": 5209
+ },
+ {
+ "epoch": 72.36331877729258,
+ "grad_norm": 0.02385753206908703,
+ "learning_rate": 0.0006,
+ "loss": 7.2654266357421875,
+ "step": 5210
+ },
+ {
+ "epoch": 72.37729257641921,
+ "grad_norm": 0.02794312685728073,
+ "learning_rate": 0.0006,
+ "loss": 7.250753402709961,
+ "step": 5211
+ },
+ {
+ "epoch": 72.39126637554585,
+ "grad_norm": 0.02445952221751213,
+ "learning_rate": 0.0006,
+ "loss": 7.293970108032227,
+ "step": 5212
+ },
+ {
+ "epoch": 72.40524017467249,
+ "grad_norm": 0.019660986959934235,
+ "learning_rate": 0.0006,
+ "loss": 7.1948771476745605,
+ "step": 5213
+ },
+ {
+ "epoch": 72.41921397379913,
+ "grad_norm": 0.024110861122608185,
+ "learning_rate": 0.0006,
+ "loss": 7.224499225616455,
+ "step": 5214
+ },
+ {
+ "epoch": 72.43318777292576,
+ "grad_norm": 0.021630609408020973,
+ "learning_rate": 0.0006,
+ "loss": 7.1736884117126465,
+ "step": 5215
+ },
+ {
+ "epoch": 72.44716157205241,
+ "grad_norm": 0.02185880020260811,
+ "learning_rate": 0.0006,
+ "loss": 7.167229652404785,
+ "step": 5216
+ },
+ {
+ "epoch": 72.46113537117904,
+ "grad_norm": 0.018913568928837776,
+ "learning_rate": 0.0006,
+ "loss": 7.196551322937012,
+ "step": 5217
+ },
+ {
+ "epoch": 72.47510917030567,
+ "grad_norm": 0.01802486553788185,
+ "learning_rate": 0.0006,
+ "loss": 7.144286632537842,
+ "step": 5218
+ },
+ {
+ "epoch": 72.48908296943232,
+ "grad_norm": 0.020386485382914543,
+ "learning_rate": 0.0006,
+ "loss": 7.196260929107666,
+ "step": 5219
+ },
+ {
+ "epoch": 72.50305676855895,
+ "grad_norm": 0.020077696070075035,
+ "learning_rate": 0.0006,
+ "loss": 7.138718605041504,
+ "step": 5220
+ },
+ {
+ "epoch": 72.5170305676856,
+ "grad_norm": 0.017224183306097984,
+ "learning_rate": 0.0006,
+ "loss": 7.119478702545166,
+ "step": 5221
+ },
+ {
+ "epoch": 72.53100436681223,
+ "grad_norm": 0.020847434177994728,
+ "learning_rate": 0.0006,
+ "loss": 7.068922996520996,
+ "step": 5222
+ },
+ {
+ "epoch": 72.54497816593886,
+ "grad_norm": 0.021764913573861122,
+ "learning_rate": 0.0006,
+ "loss": 7.071352958679199,
+ "step": 5223
+ },
+ {
+ "epoch": 72.5589519650655,
+ "grad_norm": 0.02697215974330902,
+ "learning_rate": 0.0006,
+ "loss": 7.033344268798828,
+ "step": 5224
+ },
+ {
+ "epoch": 72.57292576419214,
+ "grad_norm": 0.042392998933792114,
+ "learning_rate": 0.0006,
+ "loss": 7.0589919090271,
+ "step": 5225
+ },
+ {
+ "epoch": 72.58689956331878,
+ "grad_norm": 0.06616316735744476,
+ "learning_rate": 0.0006,
+ "loss": 7.060244083404541,
+ "step": 5226
+ },
+ {
+ "epoch": 72.60087336244541,
+ "grad_norm": 0.0828910544514656,
+ "learning_rate": 0.0006,
+ "loss": 7.057173728942871,
+ "step": 5227
+ },
+ {
+ "epoch": 72.61484716157206,
+ "grad_norm": 0.046767015010118484,
+ "learning_rate": 0.0006,
+ "loss": 7.010563850402832,
+ "step": 5228
+ },
+ {
+ "epoch": 72.62882096069869,
+ "grad_norm": 0.03831760957837105,
+ "learning_rate": 0.0006,
+ "loss": 6.9028825759887695,
+ "step": 5229
+ },
+ {
+ "epoch": 72.64279475982532,
+ "grad_norm": 0.05109567195177078,
+ "learning_rate": 0.0006,
+ "loss": 6.923257827758789,
+ "step": 5230
+ },
+ {
+ "epoch": 72.65676855895197,
+ "grad_norm": 0.02897934429347515,
+ "learning_rate": 0.0006,
+ "loss": 6.823014259338379,
+ "step": 5231
+ },
+ {
+ "epoch": 72.6707423580786,
+ "grad_norm": 0.03769192099571228,
+ "learning_rate": 0.0006,
+ "loss": 6.87416410446167,
+ "step": 5232
+ },
+ {
+ "epoch": 72.68471615720524,
+ "grad_norm": 0.03677637130022049,
+ "learning_rate": 0.0006,
+ "loss": 6.7649030685424805,
+ "step": 5233
+ },
+ {
+ "epoch": 72.69868995633188,
+ "grad_norm": 0.0371948778629303,
+ "learning_rate": 0.0006,
+ "loss": 6.749821662902832,
+ "step": 5234
+ },
+ {
+ "epoch": 72.71266375545852,
+ "grad_norm": 0.03057795763015747,
+ "learning_rate": 0.0006,
+ "loss": 6.748945236206055,
+ "step": 5235
+ },
+ {
+ "epoch": 72.72663755458515,
+ "grad_norm": 0.028109321370720863,
+ "learning_rate": 0.0006,
+ "loss": 6.663548946380615,
+ "step": 5236
+ },
+ {
+ "epoch": 72.74061135371178,
+ "grad_norm": 0.02442420832812786,
+ "learning_rate": 0.0006,
+ "loss": 6.636251449584961,
+ "step": 5237
+ },
+ {
+ "epoch": 72.75458515283843,
+ "grad_norm": 0.031841423362493515,
+ "learning_rate": 0.0006,
+ "loss": 6.606883525848389,
+ "step": 5238
+ },
+ {
+ "epoch": 72.76855895196506,
+ "grad_norm": 0.028100458905100822,
+ "learning_rate": 0.0006,
+ "loss": 6.603745937347412,
+ "step": 5239
+ },
+ {
+ "epoch": 72.78253275109171,
+ "grad_norm": 0.03222901374101639,
+ "learning_rate": 0.0006,
+ "loss": 6.463825225830078,
+ "step": 5240
+ },
+ {
+ "epoch": 72.79650655021834,
+ "grad_norm": 0.04766521975398064,
+ "learning_rate": 0.0006,
+ "loss": 6.43734073638916,
+ "step": 5241
+ },
+ {
+ "epoch": 72.81048034934499,
+ "grad_norm": 0.09058783948421478,
+ "learning_rate": 0.0006,
+ "loss": 6.448520183563232,
+ "step": 5242
+ },
+ {
+ "epoch": 72.82445414847162,
+ "grad_norm": 0.08995158225297928,
+ "learning_rate": 0.0006,
+ "loss": 6.522093296051025,
+ "step": 5243
+ },
+ {
+ "epoch": 72.83842794759825,
+ "grad_norm": 0.05491040274500847,
+ "learning_rate": 0.0006,
+ "loss": 6.37354040145874,
+ "step": 5244
+ },
+ {
+ "epoch": 72.8524017467249,
+ "grad_norm": 0.05198514088988304,
+ "learning_rate": 0.0006,
+ "loss": 6.362032413482666,
+ "step": 5245
+ },
+ {
+ "epoch": 72.86637554585153,
+ "grad_norm": 0.05464765429496765,
+ "learning_rate": 0.0006,
+ "loss": 6.326773643493652,
+ "step": 5246
+ },
+ {
+ "epoch": 72.88034934497817,
+ "grad_norm": 0.07337361574172974,
+ "learning_rate": 0.0006,
+ "loss": 6.297532081604004,
+ "step": 5247
+ },
+ {
+ "epoch": 72.8943231441048,
+ "grad_norm": 0.10061230510473251,
+ "learning_rate": 0.0006,
+ "loss": 6.359557628631592,
+ "step": 5248
+ },
+ {
+ "epoch": 72.90829694323143,
+ "grad_norm": 0.10005959123373032,
+ "learning_rate": 0.0006,
+ "loss": 6.422270774841309,
+ "step": 5249
+ },
+ {
+ "epoch": 72.92227074235808,
+ "grad_norm": 0.08989301323890686,
+ "learning_rate": 0.0006,
+ "loss": 6.289371490478516,
+ "step": 5250
+ },
+ {
+ "epoch": 72.93624454148471,
+ "grad_norm": 0.06371799111366272,
+ "learning_rate": 0.0006,
+ "loss": 6.199618339538574,
+ "step": 5251
+ },
+ {
+ "epoch": 72.95021834061136,
+ "grad_norm": 0.05932944267988205,
+ "learning_rate": 0.0006,
+ "loss": 6.111152648925781,
+ "step": 5252
+ },
+ {
+ "epoch": 72.96419213973799,
+ "grad_norm": 0.04735502973198891,
+ "learning_rate": 0.0006,
+ "loss": 6.026026248931885,
+ "step": 5253
+ },
+ {
+ "epoch": 72.97816593886463,
+ "grad_norm": 0.04270688816905022,
+ "learning_rate": 0.0006,
+ "loss": 6.01606559753418,
+ "step": 5254
+ },
+ {
+ "epoch": 72.99213973799127,
+ "grad_norm": 0.03774143010377884,
+ "learning_rate": 0.0006,
+ "loss": 5.8935441970825195,
+ "step": 5255
+ },
+ {
+ "epoch": 73.0,
+ "grad_norm": 0.04313631355762482,
+ "learning_rate": 0.0006,
+ "loss": 5.842061996459961,
+ "step": 5256
+ },
+ {
+ "epoch": 73.0,
+ "eval_loss": 5.9985785484313965,
+ "eval_runtime": 56.3891,
+ "eval_samples_per_second": 43.306,
+ "eval_steps_per_second": 1.366,
+ "step": 5256
+ },
+ {
+ "epoch": 73.01397379912663,
+ "grad_norm": 0.045038655400276184,
+ "learning_rate": 0.0006,
+ "loss": 5.8209662437438965,
+ "step": 5257
+ },
+ {
+ "epoch": 73.02794759825328,
+ "grad_norm": 0.03706910088658333,
+ "learning_rate": 0.0006,
+ "loss": 5.806189060211182,
+ "step": 5258
+ },
+ {
+ "epoch": 73.04192139737991,
+ "grad_norm": 0.04051685333251953,
+ "learning_rate": 0.0006,
+ "loss": 5.706869125366211,
+ "step": 5259
+ },
+ {
+ "epoch": 73.05589519650655,
+ "grad_norm": 0.04849153012037277,
+ "learning_rate": 0.0006,
+ "loss": 5.696385383605957,
+ "step": 5260
+ },
+ {
+ "epoch": 73.06986899563319,
+ "grad_norm": 0.06124403700232506,
+ "learning_rate": 0.0006,
+ "loss": 5.618223667144775,
+ "step": 5261
+ },
+ {
+ "epoch": 73.08384279475983,
+ "grad_norm": 0.07375937700271606,
+ "learning_rate": 0.0006,
+ "loss": 5.631045341491699,
+ "step": 5262
+ },
+ {
+ "epoch": 73.09781659388646,
+ "grad_norm": 0.07413844019174576,
+ "learning_rate": 0.0006,
+ "loss": 5.6532182693481445,
+ "step": 5263
+ },
+ {
+ "epoch": 73.1117903930131,
+ "grad_norm": 0.0633770152926445,
+ "learning_rate": 0.0006,
+ "loss": 5.57209587097168,
+ "step": 5264
+ },
+ {
+ "epoch": 73.12576419213974,
+ "grad_norm": 0.05436766892671585,
+ "learning_rate": 0.0006,
+ "loss": 5.53944730758667,
+ "step": 5265
+ },
+ {
+ "epoch": 73.13973799126637,
+ "grad_norm": 0.04639481008052826,
+ "learning_rate": 0.0006,
+ "loss": 5.502948760986328,
+ "step": 5266
+ },
+ {
+ "epoch": 73.15371179039302,
+ "grad_norm": 0.07556558400392532,
+ "learning_rate": 0.0006,
+ "loss": 5.429069995880127,
+ "step": 5267
+ },
+ {
+ "epoch": 73.16768558951965,
+ "grad_norm": 0.1127929836511612,
+ "learning_rate": 0.0006,
+ "loss": 5.545061111450195,
+ "step": 5268
+ },
+ {
+ "epoch": 73.18165938864628,
+ "grad_norm": 0.07620745897293091,
+ "learning_rate": 0.0006,
+ "loss": 5.507102966308594,
+ "step": 5269
+ },
+ {
+ "epoch": 73.19563318777293,
+ "grad_norm": 0.05076538026332855,
+ "learning_rate": 0.0006,
+ "loss": 5.329682350158691,
+ "step": 5270
+ },
+ {
+ "epoch": 73.20960698689956,
+ "grad_norm": 0.05407547950744629,
+ "learning_rate": 0.0006,
+ "loss": 5.3048553466796875,
+ "step": 5271
+ },
+ {
+ "epoch": 73.2235807860262,
+ "grad_norm": 0.059095583856105804,
+ "learning_rate": 0.0006,
+ "loss": 5.292568683624268,
+ "step": 5272
+ },
+ {
+ "epoch": 73.23755458515284,
+ "grad_norm": 0.06193305924534798,
+ "learning_rate": 0.0006,
+ "loss": 5.1870927810668945,
+ "step": 5273
+ },
+ {
+ "epoch": 73.25152838427948,
+ "grad_norm": 0.062261562794446945,
+ "learning_rate": 0.0006,
+ "loss": 5.256819248199463,
+ "step": 5274
+ },
+ {
+ "epoch": 73.26550218340611,
+ "grad_norm": 0.0732516348361969,
+ "learning_rate": 0.0006,
+ "loss": 5.256464004516602,
+ "step": 5275
+ },
+ {
+ "epoch": 73.27947598253274,
+ "grad_norm": 0.12177979946136475,
+ "learning_rate": 0.0006,
+ "loss": 5.284239768981934,
+ "step": 5276
+ },
+ {
+ "epoch": 73.29344978165939,
+ "grad_norm": 0.10459653288125992,
+ "learning_rate": 0.0006,
+ "loss": 5.269713401794434,
+ "step": 5277
+ },
+ {
+ "epoch": 73.30742358078602,
+ "grad_norm": 0.07854694873094559,
+ "learning_rate": 0.0006,
+ "loss": 5.3245391845703125,
+ "step": 5278
+ },
+ {
+ "epoch": 73.32139737991267,
+ "grad_norm": 0.06030094623565674,
+ "learning_rate": 0.0006,
+ "loss": 5.033702850341797,
+ "step": 5279
+ },
+ {
+ "epoch": 73.3353711790393,
+ "grad_norm": 0.05168461427092552,
+ "learning_rate": 0.0006,
+ "loss": 5.140530109405518,
+ "step": 5280
+ },
+ {
+ "epoch": 73.34934497816595,
+ "grad_norm": 0.041064050048589706,
+ "learning_rate": 0.0006,
+ "loss": 5.1085896492004395,
+ "step": 5281
+ },
+ {
+ "epoch": 73.36331877729258,
+ "grad_norm": 0.03788219019770622,
+ "learning_rate": 0.0006,
+ "loss": 5.045009613037109,
+ "step": 5282
+ },
+ {
+ "epoch": 73.37729257641921,
+ "grad_norm": 0.03851071000099182,
+ "learning_rate": 0.0006,
+ "loss": 5.055008888244629,
+ "step": 5283
+ },
+ {
+ "epoch": 73.39126637554585,
+ "grad_norm": 0.03448539599776268,
+ "learning_rate": 0.0006,
+ "loss": 5.10657262802124,
+ "step": 5284
+ },
+ {
+ "epoch": 73.40524017467249,
+ "grad_norm": 0.026618091389536858,
+ "learning_rate": 0.0006,
+ "loss": 5.0342912673950195,
+ "step": 5285
+ },
+ {
+ "epoch": 73.41921397379913,
+ "grad_norm": 0.03060489147901535,
+ "learning_rate": 0.0006,
+ "loss": 5.022700786590576,
+ "step": 5286
+ },
+ {
+ "epoch": 73.43318777292576,
+ "grad_norm": 0.03362000361084938,
+ "learning_rate": 0.0006,
+ "loss": 4.891584396362305,
+ "step": 5287
+ },
+ {
+ "epoch": 73.44716157205241,
+ "grad_norm": 0.03887505829334259,
+ "learning_rate": 0.0006,
+ "loss": 4.772258758544922,
+ "step": 5288
+ },
+ {
+ "epoch": 73.46113537117904,
+ "grad_norm": 0.07157225161790848,
+ "learning_rate": 0.0006,
+ "loss": 4.997142791748047,
+ "step": 5289
+ },
+ {
+ "epoch": 73.47510917030567,
+ "grad_norm": 0.10180624574422836,
+ "learning_rate": 0.0006,
+ "loss": 4.908722877502441,
+ "step": 5290
+ },
+ {
+ "epoch": 73.48908296943232,
+ "grad_norm": 0.06529122591018677,
+ "learning_rate": 0.0006,
+ "loss": 4.98001766204834,
+ "step": 5291
+ },
+ {
+ "epoch": 73.50305676855895,
+ "grad_norm": 0.055171865969896317,
+ "learning_rate": 0.0006,
+ "loss": 4.877438068389893,
+ "step": 5292
+ },
+ {
+ "epoch": 73.5170305676856,
+ "grad_norm": 0.0318770632147789,
+ "learning_rate": 0.0006,
+ "loss": 4.836934566497803,
+ "step": 5293
+ },
+ {
+ "epoch": 73.53100436681223,
+ "grad_norm": 0.042072564363479614,
+ "learning_rate": 0.0006,
+ "loss": 4.919489860534668,
+ "step": 5294
+ },
+ {
+ "epoch": 73.54497816593886,
+ "grad_norm": 0.029197754338383675,
+ "learning_rate": 0.0006,
+ "loss": 4.805579662322998,
+ "step": 5295
+ },
+ {
+ "epoch": 73.5589519650655,
+ "grad_norm": 0.03249497339129448,
+ "learning_rate": 0.0006,
+ "loss": 4.704806327819824,
+ "step": 5296
+ },
+ {
+ "epoch": 73.57292576419214,
+ "grad_norm": 0.025050945580005646,
+ "learning_rate": 0.0006,
+ "loss": 4.936512470245361,
+ "step": 5297
+ },
+ {
+ "epoch": 73.58689956331878,
+ "grad_norm": 0.029073381796479225,
+ "learning_rate": 0.0006,
+ "loss": 4.757592678070068,
+ "step": 5298
+ },
+ {
+ "epoch": 73.60087336244541,
+ "grad_norm": 0.03465264290571213,
+ "learning_rate": 0.0006,
+ "loss": 4.729528427124023,
+ "step": 5299
+ },
+ {
+ "epoch": 73.61484716157206,
+ "grad_norm": 0.054795749485492706,
+ "learning_rate": 0.0006,
+ "loss": 4.77252197265625,
+ "step": 5300
+ },
+ {
+ "epoch": 73.62882096069869,
+ "grad_norm": 0.18129196763038635,
+ "learning_rate": 0.0006,
+ "loss": 4.901700019836426,
+ "step": 5301
+ },
+ {
+ "epoch": 73.64279475982532,
+ "grad_norm": 0.23274236917495728,
+ "learning_rate": 0.0006,
+ "loss": 5.207432270050049,
+ "step": 5302
+ },
+ {
+ "epoch": 73.65676855895197,
+ "grad_norm": 0.18162772059440613,
+ "learning_rate": 0.0006,
+ "loss": 5.326632976531982,
+ "step": 5303
+ },
+ {
+ "epoch": 73.6707423580786,
+ "grad_norm": 0.11858413368463516,
+ "learning_rate": 0.0006,
+ "loss": 5.3730058670043945,
+ "step": 5304
+ },
+ {
+ "epoch": 73.68471615720524,
+ "grad_norm": 0.06031268090009689,
+ "learning_rate": 0.0006,
+ "loss": 5.330936431884766,
+ "step": 5305
+ },
+ {
+ "epoch": 73.69868995633188,
+ "grad_norm": 0.059304628521203995,
+ "learning_rate": 0.0006,
+ "loss": 5.280514717102051,
+ "step": 5306
+ },
+ {
+ "epoch": 73.71266375545852,
+ "grad_norm": 0.04702131822705269,
+ "learning_rate": 0.0006,
+ "loss": 5.288274765014648,
+ "step": 5307
+ },
+ {
+ "epoch": 73.72663755458515,
+ "grad_norm": 0.037792228162288666,
+ "learning_rate": 0.0006,
+ "loss": 5.173289775848389,
+ "step": 5308
+ },
+ {
+ "epoch": 73.74061135371178,
+ "grad_norm": 0.037763576954603195,
+ "learning_rate": 0.0006,
+ "loss": 5.084990978240967,
+ "step": 5309
+ },
+ {
+ "epoch": 73.75458515283843,
+ "grad_norm": 0.03337346389889717,
+ "learning_rate": 0.0006,
+ "loss": 5.121417045593262,
+ "step": 5310
+ },
+ {
+ "epoch": 73.76855895196506,
+ "grad_norm": 0.03184802085161209,
+ "learning_rate": 0.0006,
+ "loss": 5.208918571472168,
+ "step": 5311
+ },
+ {
+ "epoch": 73.78253275109171,
+ "grad_norm": 0.02832048572599888,
+ "learning_rate": 0.0006,
+ "loss": 4.923543453216553,
+ "step": 5312
+ },
+ {
+ "epoch": 73.79650655021834,
+ "grad_norm": 0.028297102078795433,
+ "learning_rate": 0.0006,
+ "loss": 4.944367408752441,
+ "step": 5313
+ },
+ {
+ "epoch": 73.81048034934499,
+ "grad_norm": 0.027701659128069878,
+ "learning_rate": 0.0006,
+ "loss": 4.858168601989746,
+ "step": 5314
+ },
+ {
+ "epoch": 73.82445414847162,
+ "grad_norm": 0.02214639261364937,
+ "learning_rate": 0.0006,
+ "loss": 4.887052536010742,
+ "step": 5315
+ },
+ {
+ "epoch": 73.83842794759825,
+ "grad_norm": 0.01995093561708927,
+ "learning_rate": 0.0006,
+ "loss": 4.885797500610352,
+ "step": 5316
+ },
+ {
+ "epoch": 73.8524017467249,
+ "grad_norm": 0.020473985001444817,
+ "learning_rate": 0.0006,
+ "loss": 4.917325019836426,
+ "step": 5317
+ },
+ {
+ "epoch": 73.86637554585153,
+ "grad_norm": 0.0192868635058403,
+ "learning_rate": 0.0006,
+ "loss": 4.801746845245361,
+ "step": 5318
+ },
+ {
+ "epoch": 73.88034934497817,
+ "grad_norm": 0.017108267173171043,
+ "learning_rate": 0.0006,
+ "loss": 4.893620491027832,
+ "step": 5319
+ },
+ {
+ "epoch": 73.8943231441048,
+ "grad_norm": 0.01718071475625038,
+ "learning_rate": 0.0006,
+ "loss": 4.940695285797119,
+ "step": 5320
+ },
+ {
+ "epoch": 73.90829694323143,
+ "grad_norm": 0.017821557819843292,
+ "learning_rate": 0.0006,
+ "loss": 4.767613410949707,
+ "step": 5321
+ },
+ {
+ "epoch": 73.92227074235808,
+ "grad_norm": 0.015262950211763382,
+ "learning_rate": 0.0006,
+ "loss": 4.910315036773682,
+ "step": 5322
+ },
+ {
+ "epoch": 73.93624454148471,
+ "grad_norm": 0.01769006811082363,
+ "learning_rate": 0.0006,
+ "loss": 4.7743449211120605,
+ "step": 5323
+ },
+ {
+ "epoch": 73.95021834061136,
+ "grad_norm": 0.014460788108408451,
+ "learning_rate": 0.0006,
+ "loss": 4.743494987487793,
+ "step": 5324
+ },
+ {
+ "epoch": 73.96419213973799,
+ "grad_norm": 0.01403993833810091,
+ "learning_rate": 0.0006,
+ "loss": 4.666640281677246,
+ "step": 5325
+ },
+ {
+ "epoch": 73.97816593886463,
+ "grad_norm": 0.013997809961438179,
+ "learning_rate": 0.0006,
+ "loss": 4.698747158050537,
+ "step": 5326
+ },
+ {
+ "epoch": 73.99213973799127,
+ "grad_norm": 0.014244182966649532,
+ "learning_rate": 0.0006,
+ "loss": 4.745250701904297,
+ "step": 5327
+ },
+ {
+ "epoch": 74.0,
+ "grad_norm": 0.015144186094403267,
+ "learning_rate": 0.0006,
+ "loss": 4.667946815490723,
+ "step": 5328
+ },
+ {
+ "epoch": 74.0,
+ "eval_loss": 5.005440711975098,
+ "eval_runtime": 56.7836,
+ "eval_samples_per_second": 43.005,
+ "eval_steps_per_second": 1.356,
+ "step": 5328
+ },
+ {
+ "epoch": 74.01397379912663,
+ "grad_norm": 0.020384902134537697,
+ "learning_rate": 0.0006,
+ "loss": 4.626955509185791,
+ "step": 5329
+ },
+ {
+ "epoch": 74.02794759825328,
+ "grad_norm": 0.03624720871448517,
+ "learning_rate": 0.0006,
+ "loss": 4.713998317718506,
+ "step": 5330
+ },
+ {
+ "epoch": 74.04192139737991,
+ "grad_norm": 0.05701754242181778,
+ "learning_rate": 0.0006,
+ "loss": 4.717874526977539,
+ "step": 5331
+ },
+ {
+ "epoch": 74.05589519650655,
+ "grad_norm": 0.03942621126770973,
+ "learning_rate": 0.0006,
+ "loss": 4.7026190757751465,
+ "step": 5332
+ },
+ {
+ "epoch": 74.06986899563319,
+ "grad_norm": 0.017803272232413292,
+ "learning_rate": 0.0006,
+ "loss": 4.702548027038574,
+ "step": 5333
+ },
+ {
+ "epoch": 74.08384279475983,
+ "grad_norm": 0.017024677246809006,
+ "learning_rate": 0.0006,
+ "loss": 4.69752311706543,
+ "step": 5334
+ },
+ {
+ "epoch": 74.09781659388646,
+ "grad_norm": 0.016467854380607605,
+ "learning_rate": 0.0006,
+ "loss": 4.624876976013184,
+ "step": 5335
+ },
+ {
+ "epoch": 74.1117903930131,
+ "grad_norm": 0.016569215804338455,
+ "learning_rate": 0.0006,
+ "loss": 4.645040512084961,
+ "step": 5336
+ },
+ {
+ "epoch": 74.12576419213974,
+ "grad_norm": 0.012872067280113697,
+ "learning_rate": 0.0006,
+ "loss": 4.64582633972168,
+ "step": 5337
+ },
+ {
+ "epoch": 74.13973799126637,
+ "grad_norm": 0.015689387917518616,
+ "learning_rate": 0.0006,
+ "loss": 4.596040725708008,
+ "step": 5338
+ },
+ {
+ "epoch": 74.15371179039302,
+ "grad_norm": 0.013082082383334637,
+ "learning_rate": 0.0006,
+ "loss": 4.648033142089844,
+ "step": 5339
+ },
+ {
+ "epoch": 74.16768558951965,
+ "grad_norm": 0.013728172518312931,
+ "learning_rate": 0.0006,
+ "loss": 4.657873153686523,
+ "step": 5340
+ },
+ {
+ "epoch": 74.18165938864628,
+ "grad_norm": 0.013120594434440136,
+ "learning_rate": 0.0006,
+ "loss": 4.696074485778809,
+ "step": 5341
+ },
+ {
+ "epoch": 74.19563318777293,
+ "grad_norm": 0.01431245356798172,
+ "learning_rate": 0.0006,
+ "loss": 4.5679121017456055,
+ "step": 5342
+ },
+ {
+ "epoch": 74.20960698689956,
+ "grad_norm": 0.013994511216878891,
+ "learning_rate": 0.0006,
+ "loss": 4.5236406326293945,
+ "step": 5343
+ },
+ {
+ "epoch": 74.2235807860262,
+ "grad_norm": 0.011971707455813885,
+ "learning_rate": 0.0006,
+ "loss": 4.626547813415527,
+ "step": 5344
+ },
+ {
+ "epoch": 74.23755458515284,
+ "grad_norm": 0.012805287726223469,
+ "learning_rate": 0.0006,
+ "loss": 4.556386947631836,
+ "step": 5345
+ },
+ {
+ "epoch": 74.25152838427948,
+ "grad_norm": 0.011694584973156452,
+ "learning_rate": 0.0006,
+ "loss": 4.656987190246582,
+ "step": 5346
+ },
+ {
+ "epoch": 74.26550218340611,
+ "grad_norm": 0.013396821916103363,
+ "learning_rate": 0.0006,
+ "loss": 4.526467323303223,
+ "step": 5347
+ },
+ {
+ "epoch": 74.27947598253274,
+ "grad_norm": 0.014398031868040562,
+ "learning_rate": 0.0006,
+ "loss": 4.595100402832031,
+ "step": 5348
+ },
+ {
+ "epoch": 74.29344978165939,
+ "grad_norm": 0.02165800705552101,
+ "learning_rate": 0.0006,
+ "loss": 4.460549354553223,
+ "step": 5349
+ },
+ {
+ "epoch": 74.30742358078602,
+ "grad_norm": 0.03276512026786804,
+ "learning_rate": 0.0006,
+ "loss": 4.498189449310303,
+ "step": 5350
+ },
+ {
+ "epoch": 74.32139737991267,
+ "grad_norm": 0.04616224020719528,
+ "learning_rate": 0.0006,
+ "loss": 4.5268659591674805,
+ "step": 5351
+ },
+ {
+ "epoch": 74.3353711790393,
+ "grad_norm": 0.03863786533474922,
+ "learning_rate": 0.0006,
+ "loss": 4.554747104644775,
+ "step": 5352
+ },
+ {
+ "epoch": 74.34934497816595,
+ "grad_norm": 0.01749865710735321,
+ "learning_rate": 0.0006,
+ "loss": 4.6656036376953125,
+ "step": 5353
+ },
+ {
+ "epoch": 74.36331877729258,
+ "grad_norm": 0.021814396604895592,
+ "learning_rate": 0.0006,
+ "loss": 4.607456207275391,
+ "step": 5354
+ },
+ {
+ "epoch": 74.37729257641921,
+ "grad_norm": 0.017018496990203857,
+ "learning_rate": 0.0006,
+ "loss": 4.54205322265625,
+ "step": 5355
+ },
+ {
+ "epoch": 74.39126637554585,
+ "grad_norm": 0.018297340720891953,
+ "learning_rate": 0.0006,
+ "loss": 4.608870506286621,
+ "step": 5356
+ },
+ {
+ "epoch": 74.40524017467249,
+ "grad_norm": 0.01473482046276331,
+ "learning_rate": 0.0006,
+ "loss": 4.384891510009766,
+ "step": 5357
+ },
+ {
+ "epoch": 74.41921397379913,
+ "grad_norm": 0.01579415239393711,
+ "learning_rate": 0.0006,
+ "loss": 4.588859558105469,
+ "step": 5358
+ },
+ {
+ "epoch": 74.43318777292576,
+ "grad_norm": 0.013906026259064674,
+ "learning_rate": 0.0006,
+ "loss": 4.525175094604492,
+ "step": 5359
+ },
+ {
+ "epoch": 74.44716157205241,
+ "grad_norm": 0.014163361862301826,
+ "learning_rate": 0.0006,
+ "loss": 4.495802879333496,
+ "step": 5360
+ },
+ {
+ "epoch": 74.46113537117904,
+ "grad_norm": 0.014009335078299046,
+ "learning_rate": 0.0006,
+ "loss": 4.5960869789123535,
+ "step": 5361
+ },
+ {
+ "epoch": 74.47510917030567,
+ "grad_norm": 0.012765258550643921,
+ "learning_rate": 0.0006,
+ "loss": 4.515678405761719,
+ "step": 5362
+ },
+ {
+ "epoch": 74.48908296943232,
+ "grad_norm": 0.014071918092668056,
+ "learning_rate": 0.0006,
+ "loss": 4.394025802612305,
+ "step": 5363
+ },
+ {
+ "epoch": 74.50305676855895,
+ "grad_norm": 0.01244515273720026,
+ "learning_rate": 0.0006,
+ "loss": 4.576985836029053,
+ "step": 5364
+ },
+ {
+ "epoch": 74.5170305676856,
+ "grad_norm": 0.012975298799574375,
+ "learning_rate": 0.0006,
+ "loss": 4.540657997131348,
+ "step": 5365
+ },
+ {
+ "epoch": 74.53100436681223,
+ "grad_norm": 0.012380947358906269,
+ "learning_rate": 0.0006,
+ "loss": 4.55731725692749,
+ "step": 5366
+ },
+ {
+ "epoch": 74.54497816593886,
+ "grad_norm": 0.012577617540955544,
+ "learning_rate": 0.0006,
+ "loss": 4.500924110412598,
+ "step": 5367
+ },
+ {
+ "epoch": 74.5589519650655,
+ "grad_norm": 0.012089254334568977,
+ "learning_rate": 0.0006,
+ "loss": 4.561795234680176,
+ "step": 5368
+ },
+ {
+ "epoch": 74.57292576419214,
+ "grad_norm": 0.012024011462926865,
+ "learning_rate": 0.0006,
+ "loss": 4.430635452270508,
+ "step": 5369
+ },
+ {
+ "epoch": 74.58689956331878,
+ "grad_norm": 0.012361960485577583,
+ "learning_rate": 0.0006,
+ "loss": 4.4664106369018555,
+ "step": 5370
+ },
+ {
+ "epoch": 74.60087336244541,
+ "grad_norm": 0.012149893678724766,
+ "learning_rate": 0.0006,
+ "loss": 4.522271156311035,
+ "step": 5371
+ },
+ {
+ "epoch": 74.61484716157206,
+ "grad_norm": 0.01166840922087431,
+ "learning_rate": 0.0006,
+ "loss": 4.3870463371276855,
+ "step": 5372
+ },
+ {
+ "epoch": 74.62882096069869,
+ "grad_norm": 0.01160407718271017,
+ "learning_rate": 0.0006,
+ "loss": 4.504255771636963,
+ "step": 5373
+ },
+ {
+ "epoch": 74.64279475982532,
+ "grad_norm": 0.011259177699685097,
+ "learning_rate": 0.0006,
+ "loss": 4.507267951965332,
+ "step": 5374
+ },
+ {
+ "epoch": 74.65676855895197,
+ "grad_norm": 0.012285715900361538,
+ "learning_rate": 0.0006,
+ "loss": 4.472966194152832,
+ "step": 5375
+ },
+ {
+ "epoch": 74.6707423580786,
+ "grad_norm": 0.011214440688490868,
+ "learning_rate": 0.0006,
+ "loss": 4.5840253829956055,
+ "step": 5376
+ },
+ {
+ "epoch": 74.68471615720524,
+ "grad_norm": 0.012394879944622517,
+ "learning_rate": 0.0006,
+ "loss": 4.417830944061279,
+ "step": 5377
+ },
+ {
+ "epoch": 74.69868995633188,
+ "grad_norm": 0.012360049411654472,
+ "learning_rate": 0.0006,
+ "loss": 4.535454750061035,
+ "step": 5378
+ },
+ {
+ "epoch": 74.71266375545852,
+ "grad_norm": 0.011961652897298336,
+ "learning_rate": 0.0006,
+ "loss": 4.403409957885742,
+ "step": 5379
+ },
+ {
+ "epoch": 74.72663755458515,
+ "grad_norm": 0.012078274972736835,
+ "learning_rate": 0.0006,
+ "loss": 4.448366641998291,
+ "step": 5380
+ },
+ {
+ "epoch": 74.74061135371178,
+ "grad_norm": 0.013239333406090736,
+ "learning_rate": 0.0006,
+ "loss": 4.543450355529785,
+ "step": 5381
+ },
+ {
+ "epoch": 74.75458515283843,
+ "grad_norm": 0.01476441603153944,
+ "learning_rate": 0.0006,
+ "loss": 4.488677978515625,
+ "step": 5382
+ },
+ {
+ "epoch": 74.76855895196506,
+ "grad_norm": 0.013968009501695633,
+ "learning_rate": 0.0006,
+ "loss": 4.603639602661133,
+ "step": 5383
+ },
+ {
+ "epoch": 74.78253275109171,
+ "grad_norm": 0.012322458438575268,
+ "learning_rate": 0.0006,
+ "loss": 4.637874603271484,
+ "step": 5384
+ },
+ {
+ "epoch": 74.79650655021834,
+ "grad_norm": 0.013352593407034874,
+ "learning_rate": 0.0006,
+ "loss": 4.474833965301514,
+ "step": 5385
+ },
+ {
+ "epoch": 74.81048034934499,
+ "grad_norm": 0.014671728946268559,
+ "learning_rate": 0.0006,
+ "loss": 4.516847610473633,
+ "step": 5386
+ },
+ {
+ "epoch": 74.82445414847162,
+ "grad_norm": 0.017163321375846863,
+ "learning_rate": 0.0006,
+ "loss": 4.462325096130371,
+ "step": 5387
+ },
+ {
+ "epoch": 74.83842794759825,
+ "grad_norm": 0.02012511156499386,
+ "learning_rate": 0.0006,
+ "loss": 4.355982780456543,
+ "step": 5388
+ },
+ {
+ "epoch": 74.8524017467249,
+ "grad_norm": 0.01881697215139866,
+ "learning_rate": 0.0006,
+ "loss": 4.451120376586914,
+ "step": 5389
+ },
+ {
+ "epoch": 74.86637554585153,
+ "grad_norm": 0.017279407009482384,
+ "learning_rate": 0.0006,
+ "loss": 4.461202621459961,
+ "step": 5390
+ },
+ {
+ "epoch": 74.88034934497817,
+ "grad_norm": 0.015470407903194427,
+ "learning_rate": 0.0006,
+ "loss": 4.4968671798706055,
+ "step": 5391
+ },
+ {
+ "epoch": 74.8943231441048,
+ "grad_norm": 0.01411783043295145,
+ "learning_rate": 0.0006,
+ "loss": 4.558697700500488,
+ "step": 5392
+ },
+ {
+ "epoch": 74.90829694323143,
+ "grad_norm": 0.017178812995553017,
+ "learning_rate": 0.0006,
+ "loss": 4.254153251647949,
+ "step": 5393
+ },
+ {
+ "epoch": 74.92227074235808,
+ "grad_norm": 0.019193004816770554,
+ "learning_rate": 0.0006,
+ "loss": 4.394747257232666,
+ "step": 5394
+ },
+ {
+ "epoch": 74.93624454148471,
+ "grad_norm": 0.018942881375551224,
+ "learning_rate": 0.0006,
+ "loss": 4.459539413452148,
+ "step": 5395
+ },
+ {
+ "epoch": 74.95021834061136,
+ "grad_norm": 0.01641223020851612,
+ "learning_rate": 0.0006,
+ "loss": 4.485820770263672,
+ "step": 5396
+ },
+ {
+ "epoch": 74.96419213973799,
+ "grad_norm": 0.014485753141343594,
+ "learning_rate": 0.0006,
+ "loss": 4.5068583488464355,
+ "step": 5397
+ },
+ {
+ "epoch": 74.97816593886463,
+ "grad_norm": 0.018242299556732178,
+ "learning_rate": 0.0006,
+ "loss": 4.508835792541504,
+ "step": 5398
+ },
+ {
+ "epoch": 74.99213973799127,
+ "grad_norm": 0.016491873189806938,
+ "learning_rate": 0.0006,
+ "loss": 4.4760870933532715,
+ "step": 5399
+ },
+ {
+ "epoch": 75.0,
+ "grad_norm": 0.014543353579938412,
+ "learning_rate": 0.0006,
+ "loss": 4.540217399597168,
+ "step": 5400
+ },
+ {
+ "epoch": 75.0,
+ "eval_loss": 4.7617573738098145,
+ "eval_runtime": 57.321,
+ "eval_samples_per_second": 42.602,
+ "eval_steps_per_second": 1.343,
+ "step": 5400
+ },
+ {
+ "epoch": 75.01397379912663,
+ "grad_norm": 0.014417114667594433,
+ "learning_rate": 0.0006,
+ "loss": 4.332956790924072,
+ "step": 5401
+ },
+ {
+ "epoch": 75.02794759825328,
+ "grad_norm": 0.01607639715075493,
+ "learning_rate": 0.0006,
+ "loss": 4.352453708648682,
+ "step": 5402
+ },
+ {
+ "epoch": 75.04192139737991,
+ "grad_norm": 0.015438318252563477,
+ "learning_rate": 0.0006,
+ "loss": 4.387139320373535,
+ "step": 5403
+ },
+ {
+ "epoch": 75.05589519650655,
+ "grad_norm": 0.015539007261395454,
+ "learning_rate": 0.0006,
+ "loss": 4.453171730041504,
+ "step": 5404
+ },
+ {
+ "epoch": 75.06986899563319,
+ "grad_norm": 0.017361903563141823,
+ "learning_rate": 0.0006,
+ "loss": 4.512143135070801,
+ "step": 5405
+ },
+ {
+ "epoch": 75.08384279475983,
+ "grad_norm": 0.01815599389374256,
+ "learning_rate": 0.0006,
+ "loss": 4.508997440338135,
+ "step": 5406
+ },
+ {
+ "epoch": 75.09781659388646,
+ "grad_norm": 0.01620607078075409,
+ "learning_rate": 0.0006,
+ "loss": 4.322444915771484,
+ "step": 5407
+ },
+ {
+ "epoch": 75.1117903930131,
+ "grad_norm": 0.014773265458643436,
+ "learning_rate": 0.0006,
+ "loss": 4.359251976013184,
+ "step": 5408
+ },
+ {
+ "epoch": 75.12576419213974,
+ "grad_norm": 0.014099023304879665,
+ "learning_rate": 0.0006,
+ "loss": 4.313380241394043,
+ "step": 5409
+ },
+ {
+ "epoch": 75.13973799126637,
+ "grad_norm": 0.014387735165655613,
+ "learning_rate": 0.0006,
+ "loss": 4.435023307800293,
+ "step": 5410
+ },
+ {
+ "epoch": 75.15371179039302,
+ "grad_norm": 0.014886162243783474,
+ "learning_rate": 0.0006,
+ "loss": 4.335939407348633,
+ "step": 5411
+ },
+ {
+ "epoch": 75.16768558951965,
+ "grad_norm": 0.015766102820634842,
+ "learning_rate": 0.0006,
+ "loss": 4.408674240112305,
+ "step": 5412
+ },
+ {
+ "epoch": 75.18165938864628,
+ "grad_norm": 0.016383284702897072,
+ "learning_rate": 0.0006,
+ "loss": 4.385453224182129,
+ "step": 5413
+ },
+ {
+ "epoch": 75.19563318777293,
+ "grad_norm": 0.013535384088754654,
+ "learning_rate": 0.0006,
+ "loss": 4.292785167694092,
+ "step": 5414
+ },
+ {
+ "epoch": 75.20960698689956,
+ "grad_norm": 0.01354036945849657,
+ "learning_rate": 0.0006,
+ "loss": 4.306252479553223,
+ "step": 5415
+ },
+ {
+ "epoch": 75.2235807860262,
+ "grad_norm": 0.015818078070878983,
+ "learning_rate": 0.0006,
+ "loss": 4.425487995147705,
+ "step": 5416
+ },
+ {
+ "epoch": 75.23755458515284,
+ "grad_norm": 0.014363568276166916,
+ "learning_rate": 0.0006,
+ "loss": 4.395087242126465,
+ "step": 5417
+ },
+ {
+ "epoch": 75.25152838427948,
+ "grad_norm": 0.013895470649003983,
+ "learning_rate": 0.0006,
+ "loss": 4.364090442657471,
+ "step": 5418
+ },
+ {
+ "epoch": 75.26550218340611,
+ "grad_norm": 0.014178668148815632,
+ "learning_rate": 0.0006,
+ "loss": 4.44619083404541,
+ "step": 5419
+ },
+ {
+ "epoch": 75.27947598253274,
+ "grad_norm": 0.0152514623478055,
+ "learning_rate": 0.0006,
+ "loss": 4.348445892333984,
+ "step": 5420
+ },
+ {
+ "epoch": 75.29344978165939,
+ "grad_norm": 0.015397852286696434,
+ "learning_rate": 0.0006,
+ "loss": 4.476832389831543,
+ "step": 5421
+ },
+ {
+ "epoch": 75.30742358078602,
+ "grad_norm": 0.018009863793849945,
+ "learning_rate": 0.0006,
+ "loss": 4.431162357330322,
+ "step": 5422
+ },
+ {
+ "epoch": 75.32139737991267,
+ "grad_norm": 0.017038367688655853,
+ "learning_rate": 0.0006,
+ "loss": 4.474787712097168,
+ "step": 5423
+ },
+ {
+ "epoch": 75.3353711790393,
+ "grad_norm": 0.013763288035988808,
+ "learning_rate": 0.0006,
+ "loss": 4.470419883728027,
+ "step": 5424
+ },
+ {
+ "epoch": 75.34934497816595,
+ "grad_norm": 0.013972110114991665,
+ "learning_rate": 0.0006,
+ "loss": 4.373147964477539,
+ "step": 5425
+ },
+ {
+ "epoch": 75.36331877729258,
+ "grad_norm": 0.017197875306010246,
+ "learning_rate": 0.0006,
+ "loss": 4.494963645935059,
+ "step": 5426
+ },
+ {
+ "epoch": 75.37729257641921,
+ "grad_norm": 0.01687047816812992,
+ "learning_rate": 0.0006,
+ "loss": 4.435934066772461,
+ "step": 5427
+ },
+ {
+ "epoch": 75.39126637554585,
+ "grad_norm": 0.01578272506594658,
+ "learning_rate": 0.0006,
+ "loss": 4.37918758392334,
+ "step": 5428
+ },
+ {
+ "epoch": 75.40524017467249,
+ "grad_norm": 0.015956247225403786,
+ "learning_rate": 0.0006,
+ "loss": 4.455668926239014,
+ "step": 5429
+ },
+ {
+ "epoch": 75.41921397379913,
+ "grad_norm": 0.013495155610144138,
+ "learning_rate": 0.0006,
+ "loss": 4.4271039962768555,
+ "step": 5430
+ },
+ {
+ "epoch": 75.43318777292576,
+ "grad_norm": 0.012851621024310589,
+ "learning_rate": 0.0006,
+ "loss": 4.213107109069824,
+ "step": 5431
+ },
+ {
+ "epoch": 75.44716157205241,
+ "grad_norm": 0.013351158238947392,
+ "learning_rate": 0.0006,
+ "loss": 4.3070783615112305,
+ "step": 5432
+ },
+ {
+ "epoch": 75.46113537117904,
+ "grad_norm": 0.014765217900276184,
+ "learning_rate": 0.0006,
+ "loss": 4.385021209716797,
+ "step": 5433
+ },
+ {
+ "epoch": 75.47510917030567,
+ "grad_norm": 0.016091614961624146,
+ "learning_rate": 0.0006,
+ "loss": 4.312644004821777,
+ "step": 5434
+ },
+ {
+ "epoch": 75.48908296943232,
+ "grad_norm": 0.01612013578414917,
+ "learning_rate": 0.0006,
+ "loss": 4.4482011795043945,
+ "step": 5435
+ },
+ {
+ "epoch": 75.50305676855895,
+ "grad_norm": 0.013157603330910206,
+ "learning_rate": 0.0006,
+ "loss": 4.427262306213379,
+ "step": 5436
+ },
+ {
+ "epoch": 75.5170305676856,
+ "grad_norm": 0.016039809212088585,
+ "learning_rate": 0.0006,
+ "loss": 4.396635055541992,
+ "step": 5437
+ },
+ {
+ "epoch": 75.53100436681223,
+ "grad_norm": 0.01794283092021942,
+ "learning_rate": 0.0006,
+ "loss": 4.476870059967041,
+ "step": 5438
+ },
+ {
+ "epoch": 75.54497816593886,
+ "grad_norm": 0.017030486837029457,
+ "learning_rate": 0.0006,
+ "loss": 4.361156463623047,
+ "step": 5439
+ },
+ {
+ "epoch": 75.5589519650655,
+ "grad_norm": 0.017915161326527596,
+ "learning_rate": 0.0006,
+ "loss": 4.286147117614746,
+ "step": 5440
+ },
+ {
+ "epoch": 75.57292576419214,
+ "grad_norm": 0.017514433711767197,
+ "learning_rate": 0.0006,
+ "loss": 4.394091606140137,
+ "step": 5441
+ },
+ {
+ "epoch": 75.58689956331878,
+ "grad_norm": 0.015128589235246181,
+ "learning_rate": 0.0006,
+ "loss": 4.383126735687256,
+ "step": 5442
+ },
+ {
+ "epoch": 75.60087336244541,
+ "grad_norm": 0.0202616099268198,
+ "learning_rate": 0.0006,
+ "loss": 4.336188316345215,
+ "step": 5443
+ },
+ {
+ "epoch": 75.61484716157206,
+ "grad_norm": 0.019978471100330353,
+ "learning_rate": 0.0006,
+ "loss": 4.461453437805176,
+ "step": 5444
+ },
+ {
+ "epoch": 75.62882096069869,
+ "grad_norm": 0.01738741435110569,
+ "learning_rate": 0.0006,
+ "loss": 4.371197700500488,
+ "step": 5445
+ },
+ {
+ "epoch": 75.64279475982532,
+ "grad_norm": 0.016157809644937515,
+ "learning_rate": 0.0006,
+ "loss": 4.3312835693359375,
+ "step": 5446
+ },
+ {
+ "epoch": 75.65676855895197,
+ "grad_norm": 0.016394613310694695,
+ "learning_rate": 0.0006,
+ "loss": 4.380190849304199,
+ "step": 5447
+ },
+ {
+ "epoch": 75.6707423580786,
+ "grad_norm": 0.014727243222296238,
+ "learning_rate": 0.0006,
+ "loss": 4.308028221130371,
+ "step": 5448
+ },
+ {
+ "epoch": 75.68471615720524,
+ "grad_norm": 0.01695958711206913,
+ "learning_rate": 0.0006,
+ "loss": 4.399503707885742,
+ "step": 5449
+ },
+ {
+ "epoch": 75.69868995633188,
+ "grad_norm": 0.015720795840024948,
+ "learning_rate": 0.0006,
+ "loss": 4.342249870300293,
+ "step": 5450
+ },
+ {
+ "epoch": 75.71266375545852,
+ "grad_norm": 0.01574772410094738,
+ "learning_rate": 0.0006,
+ "loss": 4.331503868103027,
+ "step": 5451
+ },
+ {
+ "epoch": 75.72663755458515,
+ "grad_norm": 0.014812648296356201,
+ "learning_rate": 0.0006,
+ "loss": 4.332413673400879,
+ "step": 5452
+ },
+ {
+ "epoch": 75.74061135371178,
+ "grad_norm": 0.012850426137447357,
+ "learning_rate": 0.0006,
+ "loss": 4.2884063720703125,
+ "step": 5453
+ },
+ {
+ "epoch": 75.75458515283843,
+ "grad_norm": 0.012876970693469048,
+ "learning_rate": 0.0006,
+ "loss": 4.399855613708496,
+ "step": 5454
+ },
+ {
+ "epoch": 75.76855895196506,
+ "grad_norm": 0.012958863750100136,
+ "learning_rate": 0.0006,
+ "loss": 4.411185264587402,
+ "step": 5455
+ },
+ {
+ "epoch": 75.78253275109171,
+ "grad_norm": 0.012197135016322136,
+ "learning_rate": 0.0006,
+ "loss": 4.272603988647461,
+ "step": 5456
+ },
+ {
+ "epoch": 75.79650655021834,
+ "grad_norm": 0.013483566232025623,
+ "learning_rate": 0.0006,
+ "loss": 4.3564348220825195,
+ "step": 5457
+ },
+ {
+ "epoch": 75.81048034934499,
+ "grad_norm": 0.012365429662168026,
+ "learning_rate": 0.0006,
+ "loss": 4.354825973510742,
+ "step": 5458
+ },
+ {
+ "epoch": 75.82445414847162,
+ "grad_norm": 0.012238141149282455,
+ "learning_rate": 0.0006,
+ "loss": 4.440438747406006,
+ "step": 5459
+ },
+ {
+ "epoch": 75.83842794759825,
+ "grad_norm": 0.013380305841565132,
+ "learning_rate": 0.0006,
+ "loss": 4.2813215255737305,
+ "step": 5460
+ },
+ {
+ "epoch": 75.8524017467249,
+ "grad_norm": 0.01443779468536377,
+ "learning_rate": 0.0006,
+ "loss": 4.289078235626221,
+ "step": 5461
+ },
+ {
+ "epoch": 75.86637554585153,
+ "grad_norm": 0.012646574527025223,
+ "learning_rate": 0.0006,
+ "loss": 4.374177932739258,
+ "step": 5462
+ },
+ {
+ "epoch": 75.88034934497817,
+ "grad_norm": 0.012403388507664204,
+ "learning_rate": 0.0006,
+ "loss": 4.343016147613525,
+ "step": 5463
+ },
+ {
+ "epoch": 75.8943231441048,
+ "grad_norm": 0.01319252047687769,
+ "learning_rate": 0.0006,
+ "loss": 4.309957027435303,
+ "step": 5464
+ },
+ {
+ "epoch": 75.90829694323143,
+ "grad_norm": 0.012597428634762764,
+ "learning_rate": 0.0006,
+ "loss": 4.403654098510742,
+ "step": 5465
+ },
+ {
+ "epoch": 75.92227074235808,
+ "grad_norm": 0.012724226340651512,
+ "learning_rate": 0.0006,
+ "loss": 4.326869010925293,
+ "step": 5466
+ },
+ {
+ "epoch": 75.93624454148471,
+ "grad_norm": 0.012698772363364697,
+ "learning_rate": 0.0006,
+ "loss": 4.268430233001709,
+ "step": 5467
+ },
+ {
+ "epoch": 75.95021834061136,
+ "grad_norm": 0.01376490481197834,
+ "learning_rate": 0.0006,
+ "loss": 4.42446231842041,
+ "step": 5468
+ },
+ {
+ "epoch": 75.96419213973799,
+ "grad_norm": 0.013175144791603088,
+ "learning_rate": 0.0006,
+ "loss": 4.374675750732422,
+ "step": 5469
+ },
+ {
+ "epoch": 75.97816593886463,
+ "grad_norm": 0.012731322087347507,
+ "learning_rate": 0.0006,
+ "loss": 4.372966766357422,
+ "step": 5470
+ },
+ {
+ "epoch": 75.99213973799127,
+ "grad_norm": 0.01360505260527134,
+ "learning_rate": 0.0006,
+ "loss": 4.3095903396606445,
+ "step": 5471
+ },
+ {
+ "epoch": 76.0,
+ "grad_norm": 0.014011901803314686,
+ "learning_rate": 0.0006,
+ "loss": 4.26954460144043,
+ "step": 5472
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 28800,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 400,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": true,
+ "should_log": false,
+ "should_save": false,
+ "should_training_stop": false
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 2.33290615461249e+18,
+ "train_batch_size": 8,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-5472/training_args.bin b/runs/l2r50-i2-fulle-lm/checkpoint-5472/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..cad232831c222cc7f8b938b8bbe42f61e7d89c3a
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-5472/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:f7c40e3bfa2f5ac1450ca1243e3f2a895d44a06ed5995e6aefd97f78d229d8c8
+size 4792
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-576/chat_template.jinja b/runs/l2r50-i2-fulle-lm/checkpoint-576/chat_template.jinja
new file mode 100644
index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-576/chat_template.jinja
@@ -0,0 +1,85 @@
+{%- if tools %}
+ {{- '<|im_start|>system\n' }}
+ {%- if messages[0].role == 'system' %}
+ {{- messages[0].content + '\n\n' }}
+ {%- endif %}
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }}
+ {%- for tool in tools %}
+ {{- "\n" }}
+ {{- tool | tojson }}
+ {%- endfor %}
+ {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }}
+{%- else %}
+ {%- if messages[0].role == 'system' %}
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
+ {%- endif %}
+{%- endif %}
+{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
+{%- for message in messages[::-1] %}
+ {%- set index = (messages|length - 1) - loop.index0 %}
+ {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %}
+ {%- set ns.multi_step_tool = false %}
+ {%- set ns.last_query_index = index %}
+ {%- endif %}
+{%- endfor %}
+{%- for message in messages %}
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
+ {%- elif message.role == "assistant" %}
+ {%- set content = message.content %}
+ {%- set reasoning_content = '' %}
+ {%- if message.reasoning_content is defined and message.reasoning_content is not none %}
+ {%- set reasoning_content = message.reasoning_content %}
+ {%- else %}
+ {%- if '' in message.content %}
+ {%- set content = message.content.split('')[-1].lstrip('\n') %}
+ {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %}
+ {%- endif %}
+ {%- endif %}
+ {%- if loop.index0 > ns.last_query_index %}
+ {%- if loop.last or (not loop.last and reasoning_content) %}
+ {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- if message.tool_calls %}
+ {%- for tool_call in message.tool_calls %}
+ {%- if (loop.first and content) or (not loop.first) %}
+ {{- '\n' }}
+ {%- endif %}
+ {%- if tool_call.function %}
+ {%- set tool_call = tool_call.function %}
+ {%- endif %}
+ {{- '\n{"name": "' }}
+ {{- tool_call.name }}
+ {{- '", "arguments": ' }}
+ {%- if tool_call.arguments is string %}
+ {{- tool_call.arguments }}
+ {%- else %}
+ {{- tool_call.arguments | tojson }}
+ {%- endif %}
+ {{- '}\n' }}
+ {%- endfor %}
+ {%- endif %}
+ {{- '<|im_end|>\n' }}
+ {%- elif message.role == "tool" %}
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
+ {{- '<|im_start|>user' }}
+ {%- endif %}
+ {{- '\n\n' }}
+ {{- message.content }}
+ {{- '\n' }}
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
+ {{- '<|im_end|>\n' }}
+ {%- endif %}
+ {%- endif %}
+{%- endfor %}
+{%- if add_generation_prompt %}
+ {{- '<|im_start|>assistant\n' }}
+ {%- if enable_thinking is defined and enable_thinking is false %}
+ {{- '\n\n\n\n' }}
+ {%- endif %}
+{%- endif %}
\ No newline at end of file
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-576/config.json b/runs/l2r50-i2-fulle-lm/checkpoint-576/config.json
new file mode 100644
index 0000000000000000000000000000000000000000..22f250953c6ce8fc91ee4411856dd73041a415e2
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-576/config.json
@@ -0,0 +1,32 @@
+{
+ "architectures": [
+ "LlamaForCausalLM"
+ ],
+ "attention_bias": false,
+ "attention_dropout": 0.0,
+ "bos_token_id": null,
+ "dtype": "float32",
+ "eos_token_id": 151645,
+ "head_dim": 128,
+ "hidden_act": "silu",
+ "hidden_size": 512,
+ "initializer_range": 0.02,
+ "intermediate_size": 1536,
+ "max_position_embeddings": 2048,
+ "mlp_bias": false,
+ "model_type": "llama",
+ "num_attention_heads": 4,
+ "num_hidden_layers": 20,
+ "num_key_value_heads": 4,
+ "pad_token_id": 151645,
+ "pretraining_tp": 1,
+ "rms_norm_eps": 1e-06,
+ "rope_parameters": {
+ "rope_theta": 10000.0,
+ "rope_type": "default"
+ },
+ "tie_word_embeddings": true,
+ "transformers_version": "5.5.0",
+ "use_cache": false,
+ "vocab_size": 151673
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-576/generation_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-576/generation_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-576/generation_config.json
@@ -0,0 +1,11 @@
+{
+ "_from_model_config": true,
+ "eos_token_id": [
+ 151645
+ ],
+ "output_attentions": false,
+ "output_hidden_states": false,
+ "pad_token_id": 151645,
+ "transformers_version": "5.5.0",
+ "use_cache": true
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-576/model.safetensors b/runs/l2r50-i2-fulle-lm/checkpoint-576/model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..4024774d232f7817ed4648e017722ba7708e5810
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-576/model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:a5f0056e676bbae080ef7dddaf146e32ce36d296fe6594a7327e78b97be583a2
+size 583360328
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-576/optimizer.pt b/runs/l2r50-i2-fulle-lm/checkpoint-576/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..65065a006c00793f3c83b355a9dff47bb804851b
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-576/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:1ed71010198015ed1b30f0465125e93c395a3cfe700f8c774df921c6b6f48403
+size 1166833530
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-576/rng_state_0.pth b/runs/l2r50-i2-fulle-lm/checkpoint-576/rng_state_0.pth
new file mode 100644
index 0000000000000000000000000000000000000000..112b041e1d36ec1494d292f68d4ed5a663a7760c
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-576/rng_state_0.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:16f4eb7f2a605fbab21186f755abda641fe39281304fff9e3cd8d1f68ad970b0
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-576/rng_state_1.pth b/runs/l2r50-i2-fulle-lm/checkpoint-576/rng_state_1.pth
new file mode 100644
index 0000000000000000000000000000000000000000..5b7627425b53209442125436df436e61e1a84ab1
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-576/rng_state_1.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:f3ecca285a09feeb269b66fce026683d05f7aded4f691e94faec2607d18d85d8
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-576/rng_state_2.pth b/runs/l2r50-i2-fulle-lm/checkpoint-576/rng_state_2.pth
new file mode 100644
index 0000000000000000000000000000000000000000..bef918e95a7666f775e2c4b5e8379f56bf2b1155
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-576/rng_state_2.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:c164f926b8a1118b5abd6db76ed1ce8a68afbda1403f138f3634fa1a937f448d
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-576/rng_state_3.pth b/runs/l2r50-i2-fulle-lm/checkpoint-576/rng_state_3.pth
new file mode 100644
index 0000000000000000000000000000000000000000..1277d7e0e0037e1991a7937de0c59863261bf553
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-576/rng_state_3.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:4b4476d524aa476d8f8153629400bf093c94260913a40a182ea50a014b63af36
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-576/scheduler.pt b/runs/l2r50-i2-fulle-lm/checkpoint-576/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..dbc454a52c5d1edb241846f7cbf86268ced2c2ff
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-576/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:5f28b8022dab4a12fbbad2bccde532c0f7c60e469bde0bf66d40f7d043ac1727
+size 1064
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-576/tokenizer.json b/runs/l2r50-i2-fulle-lm/checkpoint-576/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..53786016f070e11184813f3c83101dc5e93e392e
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-576/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:c3c63fda1646a555448e8f03662542e04b11d103721319e4d9b267d7da03ceb8
+size 11424463
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-576/tokenizer_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-576/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..98e2fb9b2cc54785dc9d871b8e503f8b8f2defe7
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-576/tokenizer_config.json
@@ -0,0 +1,20 @@
+{
+ "add_prefix_space": false,
+ "backend": "tokenizers",
+ "bos_token": null,
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "<|im_end|>",
+ "errors": "replace",
+ "extra_special_tokens": [
+ "<|l2r_pred|>",
+ "<|r2l_pred|>",
+ "<|next_1_pred|>",
+ "<|next_2_pred|>"
+ ],
+ "is_local": false,
+ "model_max_length": 131072,
+ "pad_token": "<|im_end|>",
+ "split_special_tokens": false,
+ "tokenizer_class": "Qwen2Tokenizer",
+ "unk_token": null
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-576/trainer_state.json b/runs/l2r50-i2-fulle-lm/checkpoint-576/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..b176ef27d117d084ce630325cf416e0d647090c2
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-576/trainer_state.json
@@ -0,0 +1,4122 @@
+{
+ "best_global_step": null,
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 8.0,
+ "eval_steps": 500,
+ "global_step": 576,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.013973799126637555,
+ "grad_norm": 0.1739506870508194,
+ "learning_rate": 0.0,
+ "loss": 11.99299430847168,
+ "step": 1
+ },
+ {
+ "epoch": 0.02794759825327511,
+ "grad_norm": 0.17216597497463226,
+ "learning_rate": 5.999999999999999e-06,
+ "loss": 11.988704681396484,
+ "step": 2
+ },
+ {
+ "epoch": 0.04192139737991266,
+ "grad_norm": 0.17459715902805328,
+ "learning_rate": 1.1999999999999999e-05,
+ "loss": 11.945581436157227,
+ "step": 3
+ },
+ {
+ "epoch": 0.05589519650655022,
+ "grad_norm": 0.17577414214611053,
+ "learning_rate": 1.7999999999999997e-05,
+ "loss": 11.859772682189941,
+ "step": 4
+ },
+ {
+ "epoch": 0.06986899563318777,
+ "grad_norm": 0.16163651645183563,
+ "learning_rate": 2.3999999999999997e-05,
+ "loss": 11.744415283203125,
+ "step": 5
+ },
+ {
+ "epoch": 0.08384279475982533,
+ "grad_norm": 0.14409932494163513,
+ "learning_rate": 2.9999999999999997e-05,
+ "loss": 11.618099212646484,
+ "step": 6
+ },
+ {
+ "epoch": 0.09781659388646288,
+ "grad_norm": 0.13079848885536194,
+ "learning_rate": 3.5999999999999994e-05,
+ "loss": 11.505935668945312,
+ "step": 7
+ },
+ {
+ "epoch": 0.11179039301310044,
+ "grad_norm": 0.11964613944292068,
+ "learning_rate": 4.2e-05,
+ "loss": 11.416532516479492,
+ "step": 8
+ },
+ {
+ "epoch": 0.125764192139738,
+ "grad_norm": 0.11345142871141434,
+ "learning_rate": 4.7999999999999994e-05,
+ "loss": 11.33206558227539,
+ "step": 9
+ },
+ {
+ "epoch": 0.13973799126637554,
+ "grad_norm": 0.11033952981233597,
+ "learning_rate": 5.399999999999999e-05,
+ "loss": 11.259052276611328,
+ "step": 10
+ },
+ {
+ "epoch": 0.1537117903930131,
+ "grad_norm": 0.10776489973068237,
+ "learning_rate": 5.9999999999999995e-05,
+ "loss": 11.203384399414062,
+ "step": 11
+ },
+ {
+ "epoch": 0.16768558951965065,
+ "grad_norm": 0.10721946507692337,
+ "learning_rate": 6.599999999999999e-05,
+ "loss": 11.153295516967773,
+ "step": 12
+ },
+ {
+ "epoch": 0.18165938864628822,
+ "grad_norm": 0.10673430562019348,
+ "learning_rate": 7.199999999999999e-05,
+ "loss": 11.110200881958008,
+ "step": 13
+ },
+ {
+ "epoch": 0.19563318777292577,
+ "grad_norm": 0.10630254447460175,
+ "learning_rate": 7.8e-05,
+ "loss": 11.067477226257324,
+ "step": 14
+ },
+ {
+ "epoch": 0.2096069868995633,
+ "grad_norm": 0.10631615668535233,
+ "learning_rate": 8.4e-05,
+ "loss": 11.022941589355469,
+ "step": 15
+ },
+ {
+ "epoch": 0.22358078602620088,
+ "grad_norm": 0.10609443485736847,
+ "learning_rate": 8.999999999999999e-05,
+ "loss": 10.970746040344238,
+ "step": 16
+ },
+ {
+ "epoch": 0.23755458515283842,
+ "grad_norm": 0.10648232698440552,
+ "learning_rate": 9.599999999999999e-05,
+ "loss": 10.91267204284668,
+ "step": 17
+ },
+ {
+ "epoch": 0.251528384279476,
+ "grad_norm": 0.10584589838981628,
+ "learning_rate": 0.000102,
+ "loss": 10.854547500610352,
+ "step": 18
+ },
+ {
+ "epoch": 0.26550218340611353,
+ "grad_norm": 0.10591720044612885,
+ "learning_rate": 0.00010799999999999998,
+ "loss": 10.788917541503906,
+ "step": 19
+ },
+ {
+ "epoch": 0.2794759825327511,
+ "grad_norm": 0.10611672699451447,
+ "learning_rate": 0.00011399999999999999,
+ "loss": 10.717556953430176,
+ "step": 20
+ },
+ {
+ "epoch": 0.2934497816593886,
+ "grad_norm": 0.10455359518527985,
+ "learning_rate": 0.00011999999999999999,
+ "loss": 10.650186538696289,
+ "step": 21
+ },
+ {
+ "epoch": 0.3074235807860262,
+ "grad_norm": 0.1047525629401207,
+ "learning_rate": 0.00012599999999999997,
+ "loss": 10.56808090209961,
+ "step": 22
+ },
+ {
+ "epoch": 0.32139737991266376,
+ "grad_norm": 0.10511605441570282,
+ "learning_rate": 0.00013199999999999998,
+ "loss": 10.484028816223145,
+ "step": 23
+ },
+ {
+ "epoch": 0.3353711790393013,
+ "grad_norm": 0.10467177629470825,
+ "learning_rate": 0.000138,
+ "loss": 10.399947166442871,
+ "step": 24
+ },
+ {
+ "epoch": 0.34934497816593885,
+ "grad_norm": 0.10432136058807373,
+ "learning_rate": 0.00014399999999999998,
+ "loss": 10.312714576721191,
+ "step": 25
+ },
+ {
+ "epoch": 0.36331877729257644,
+ "grad_norm": 0.10527262091636658,
+ "learning_rate": 0.00015,
+ "loss": 10.212261199951172,
+ "step": 26
+ },
+ {
+ "epoch": 0.377292576419214,
+ "grad_norm": 0.103610560297966,
+ "learning_rate": 0.000156,
+ "loss": 10.132034301757812,
+ "step": 27
+ },
+ {
+ "epoch": 0.39126637554585153,
+ "grad_norm": 0.10553059726953506,
+ "learning_rate": 0.000162,
+ "loss": 10.020434379577637,
+ "step": 28
+ },
+ {
+ "epoch": 0.4052401746724891,
+ "grad_norm": 0.1037299782037735,
+ "learning_rate": 0.000168,
+ "loss": 9.934466361999512,
+ "step": 29
+ },
+ {
+ "epoch": 0.4192139737991266,
+ "grad_norm": 0.1042739674448967,
+ "learning_rate": 0.00017399999999999997,
+ "loss": 9.822539329528809,
+ "step": 30
+ },
+ {
+ "epoch": 0.4331877729257642,
+ "grad_norm": 0.10300865024328232,
+ "learning_rate": 0.00017999999999999998,
+ "loss": 9.734994888305664,
+ "step": 31
+ },
+ {
+ "epoch": 0.44716157205240176,
+ "grad_norm": 0.10291654616594315,
+ "learning_rate": 0.000186,
+ "loss": 9.624505996704102,
+ "step": 32
+ },
+ {
+ "epoch": 0.4611353711790393,
+ "grad_norm": 0.10210851579904556,
+ "learning_rate": 0.00019199999999999998,
+ "loss": 9.518205642700195,
+ "step": 33
+ },
+ {
+ "epoch": 0.47510917030567684,
+ "grad_norm": 0.10222529619932175,
+ "learning_rate": 0.000198,
+ "loss": 9.412222862243652,
+ "step": 34
+ },
+ {
+ "epoch": 0.4890829694323144,
+ "grad_norm": 0.09992019087076187,
+ "learning_rate": 0.000204,
+ "loss": 9.326566696166992,
+ "step": 35
+ },
+ {
+ "epoch": 0.503056768558952,
+ "grad_norm": 0.09931084513664246,
+ "learning_rate": 0.00020999999999999998,
+ "loss": 9.210111618041992,
+ "step": 36
+ },
+ {
+ "epoch": 0.5170305676855895,
+ "grad_norm": 0.09928256273269653,
+ "learning_rate": 0.00021599999999999996,
+ "loss": 9.100493431091309,
+ "step": 37
+ },
+ {
+ "epoch": 0.5310043668122271,
+ "grad_norm": 0.09843816608190536,
+ "learning_rate": 0.00022199999999999998,
+ "loss": 8.996543884277344,
+ "step": 38
+ },
+ {
+ "epoch": 0.5449781659388646,
+ "grad_norm": 0.09825374186038971,
+ "learning_rate": 0.00022799999999999999,
+ "loss": 8.888991355895996,
+ "step": 39
+ },
+ {
+ "epoch": 0.5589519650655022,
+ "grad_norm": 0.09602385759353638,
+ "learning_rate": 0.000234,
+ "loss": 8.802907943725586,
+ "step": 40
+ },
+ {
+ "epoch": 0.5729257641921397,
+ "grad_norm": 0.09414064139127731,
+ "learning_rate": 0.00023999999999999998,
+ "loss": 8.715423583984375,
+ "step": 41
+ },
+ {
+ "epoch": 0.5868995633187772,
+ "grad_norm": 0.09273893386125565,
+ "learning_rate": 0.00024599999999999996,
+ "loss": 8.61374282836914,
+ "step": 42
+ },
+ {
+ "epoch": 0.6008733624454149,
+ "grad_norm": 0.09079599380493164,
+ "learning_rate": 0.00025199999999999995,
+ "loss": 8.502904891967773,
+ "step": 43
+ },
+ {
+ "epoch": 0.6148471615720524,
+ "grad_norm": 0.08605825901031494,
+ "learning_rate": 0.000258,
+ "loss": 8.453008651733398,
+ "step": 44
+ },
+ {
+ "epoch": 0.62882096069869,
+ "grad_norm": 0.08338996022939682,
+ "learning_rate": 0.00026399999999999997,
+ "loss": 8.365163803100586,
+ "step": 45
+ },
+ {
+ "epoch": 0.6427947598253275,
+ "grad_norm": 0.08150867372751236,
+ "learning_rate": 0.00027,
+ "loss": 8.285372734069824,
+ "step": 46
+ },
+ {
+ "epoch": 0.6567685589519651,
+ "grad_norm": 0.07794807851314545,
+ "learning_rate": 0.000276,
+ "loss": 8.185811996459961,
+ "step": 47
+ },
+ {
+ "epoch": 0.6707423580786026,
+ "grad_norm": 0.07592720538377762,
+ "learning_rate": 0.00028199999999999997,
+ "loss": 8.098482131958008,
+ "step": 48
+ },
+ {
+ "epoch": 0.6847161572052402,
+ "grad_norm": 0.0709897130727768,
+ "learning_rate": 0.00028799999999999995,
+ "loss": 8.042091369628906,
+ "step": 49
+ },
+ {
+ "epoch": 0.6986899563318777,
+ "grad_norm": 0.06582305580377579,
+ "learning_rate": 0.000294,
+ "loss": 7.972844123840332,
+ "step": 50
+ },
+ {
+ "epoch": 0.7126637554585152,
+ "grad_norm": 0.059452351182699203,
+ "learning_rate": 0.0003,
+ "loss": 7.924579620361328,
+ "step": 51
+ },
+ {
+ "epoch": 0.7266375545851529,
+ "grad_norm": 0.052426718175411224,
+ "learning_rate": 0.00030599999999999996,
+ "loss": 7.876336097717285,
+ "step": 52
+ },
+ {
+ "epoch": 0.7406113537117904,
+ "grad_norm": 0.04768150672316551,
+ "learning_rate": 0.000312,
+ "loss": 7.8102707862854,
+ "step": 53
+ },
+ {
+ "epoch": 0.754585152838428,
+ "grad_norm": 0.03970075398683548,
+ "learning_rate": 0.000318,
+ "loss": 7.791913986206055,
+ "step": 54
+ },
+ {
+ "epoch": 0.7685589519650655,
+ "grad_norm": 0.03559752553701401,
+ "learning_rate": 0.000324,
+ "loss": 7.756960868835449,
+ "step": 55
+ },
+ {
+ "epoch": 0.7825327510917031,
+ "grad_norm": 0.029321353882551193,
+ "learning_rate": 0.00033,
+ "loss": 7.724878787994385,
+ "step": 56
+ },
+ {
+ "epoch": 0.7965065502183406,
+ "grad_norm": 0.023485427722334862,
+ "learning_rate": 0.000336,
+ "loss": 7.72553014755249,
+ "step": 57
+ },
+ {
+ "epoch": 0.8104803493449781,
+ "grad_norm": 0.019319474697113037,
+ "learning_rate": 0.00034199999999999996,
+ "loss": 7.709493160247803,
+ "step": 58
+ },
+ {
+ "epoch": 0.8244541484716157,
+ "grad_norm": 0.016705691814422607,
+ "learning_rate": 0.00034799999999999995,
+ "loss": 7.6873297691345215,
+ "step": 59
+ },
+ {
+ "epoch": 0.8384279475982532,
+ "grad_norm": 0.017667734995484352,
+ "learning_rate": 0.00035399999999999993,
+ "loss": 7.677059173583984,
+ "step": 60
+ },
+ {
+ "epoch": 0.8524017467248908,
+ "grad_norm": 0.01671992614865303,
+ "learning_rate": 0.00035999999999999997,
+ "loss": 7.643512725830078,
+ "step": 61
+ },
+ {
+ "epoch": 0.8663755458515284,
+ "grad_norm": 0.012288172729313374,
+ "learning_rate": 0.00036599999999999995,
+ "loss": 7.658343315124512,
+ "step": 62
+ },
+ {
+ "epoch": 0.880349344978166,
+ "grad_norm": 0.012004776857793331,
+ "learning_rate": 0.000372,
+ "loss": 7.649759292602539,
+ "step": 63
+ },
+ {
+ "epoch": 0.8943231441048035,
+ "grad_norm": 0.012487445957958698,
+ "learning_rate": 0.00037799999999999997,
+ "loss": 7.631503105163574,
+ "step": 64
+ },
+ {
+ "epoch": 0.9082969432314411,
+ "grad_norm": 0.01489241048693657,
+ "learning_rate": 0.00038399999999999996,
+ "loss": 7.647110462188721,
+ "step": 65
+ },
+ {
+ "epoch": 0.9222707423580786,
+ "grad_norm": 0.01428980939090252,
+ "learning_rate": 0.00039,
+ "loss": 7.636983871459961,
+ "step": 66
+ },
+ {
+ "epoch": 0.9362445414847161,
+ "grad_norm": 0.010711144655942917,
+ "learning_rate": 0.000396,
+ "loss": 7.621130466461182,
+ "step": 67
+ },
+ {
+ "epoch": 0.9502183406113537,
+ "grad_norm": 0.010178886353969574,
+ "learning_rate": 0.000402,
+ "loss": 7.593924522399902,
+ "step": 68
+ },
+ {
+ "epoch": 0.9641921397379912,
+ "grad_norm": 0.010604874230921268,
+ "learning_rate": 0.000408,
+ "loss": 7.6015496253967285,
+ "step": 69
+ },
+ {
+ "epoch": 0.9781659388646288,
+ "grad_norm": 0.009751426056027412,
+ "learning_rate": 0.0004139999999999999,
+ "loss": 7.571964263916016,
+ "step": 70
+ },
+ {
+ "epoch": 0.9921397379912664,
+ "grad_norm": 0.009816481731832027,
+ "learning_rate": 0.00041999999999999996,
+ "loss": 7.57344913482666,
+ "step": 71
+ },
+ {
+ "epoch": 1.0,
+ "grad_norm": 0.008665637113153934,
+ "learning_rate": 0.00042599999999999995,
+ "loss": 7.579050064086914,
+ "step": 72
+ },
+ {
+ "epoch": 1.0,
+ "eval_loss": 7.581204891204834,
+ "eval_runtime": 59.0091,
+ "eval_samples_per_second": 41.383,
+ "eval_steps_per_second": 1.305,
+ "step": 72
+ },
+ {
+ "epoch": 1.0139737991266375,
+ "grad_norm": 0.009684201329946518,
+ "learning_rate": 0.00043199999999999993,
+ "loss": 7.568594932556152,
+ "step": 73
+ },
+ {
+ "epoch": 1.027947598253275,
+ "grad_norm": 0.025783710181713104,
+ "learning_rate": 0.00043799999999999997,
+ "loss": 7.578948497772217,
+ "step": 74
+ },
+ {
+ "epoch": 1.0419213973799126,
+ "grad_norm": 0.0327819399535656,
+ "learning_rate": 0.00044399999999999995,
+ "loss": 7.548173904418945,
+ "step": 75
+ },
+ {
+ "epoch": 1.0558951965065502,
+ "grad_norm": 0.021604498848319054,
+ "learning_rate": 0.00045,
+ "loss": 7.552953720092773,
+ "step": 76
+ },
+ {
+ "epoch": 1.0698689956331877,
+ "grad_norm": 0.010230105370283127,
+ "learning_rate": 0.00045599999999999997,
+ "loss": 7.5213212966918945,
+ "step": 77
+ },
+ {
+ "epoch": 1.0838427947598253,
+ "grad_norm": 0.023315589874982834,
+ "learning_rate": 0.00046199999999999995,
+ "loss": 7.501640319824219,
+ "step": 78
+ },
+ {
+ "epoch": 1.0978165938864628,
+ "grad_norm": 0.01109944935888052,
+ "learning_rate": 0.000468,
+ "loss": 7.539519786834717,
+ "step": 79
+ },
+ {
+ "epoch": 1.1117903930131003,
+ "grad_norm": 0.026093758642673492,
+ "learning_rate": 0.000474,
+ "loss": 7.525928974151611,
+ "step": 80
+ },
+ {
+ "epoch": 1.125764192139738,
+ "grad_norm": 0.03274895250797272,
+ "learning_rate": 0.00047999999999999996,
+ "loss": 7.535602569580078,
+ "step": 81
+ },
+ {
+ "epoch": 1.1397379912663754,
+ "grad_norm": 0.01543602254241705,
+ "learning_rate": 0.000486,
+ "loss": 7.513683795928955,
+ "step": 82
+ },
+ {
+ "epoch": 1.1537117903930132,
+ "grad_norm": 0.028377249836921692,
+ "learning_rate": 0.0004919999999999999,
+ "loss": 7.49384880065918,
+ "step": 83
+ },
+ {
+ "epoch": 1.1676855895196507,
+ "grad_norm": 0.04121481627225876,
+ "learning_rate": 0.000498,
+ "loss": 7.488447189331055,
+ "step": 84
+ },
+ {
+ "epoch": 1.1816593886462883,
+ "grad_norm": 0.013282302767038345,
+ "learning_rate": 0.0005039999999999999,
+ "loss": 7.463010787963867,
+ "step": 85
+ },
+ {
+ "epoch": 1.1956331877729258,
+ "grad_norm": 0.03200364485383034,
+ "learning_rate": 0.0005099999999999999,
+ "loss": 7.473515033721924,
+ "step": 86
+ },
+ {
+ "epoch": 1.2096069868995634,
+ "grad_norm": 0.020768029615283012,
+ "learning_rate": 0.000516,
+ "loss": 7.464688777923584,
+ "step": 87
+ },
+ {
+ "epoch": 1.223580786026201,
+ "grad_norm": 0.021715788170695305,
+ "learning_rate": 0.000522,
+ "loss": 7.4835028648376465,
+ "step": 88
+ },
+ {
+ "epoch": 1.2375545851528384,
+ "grad_norm": 0.026255754753947258,
+ "learning_rate": 0.0005279999999999999,
+ "loss": 7.443392753601074,
+ "step": 89
+ },
+ {
+ "epoch": 1.251528384279476,
+ "grad_norm": 0.022325951606035233,
+ "learning_rate": 0.000534,
+ "loss": 7.457373142242432,
+ "step": 90
+ },
+ {
+ "epoch": 1.2655021834061135,
+ "grad_norm": 0.009270387701690197,
+ "learning_rate": 0.00054,
+ "loss": 7.455986976623535,
+ "step": 91
+ },
+ {
+ "epoch": 1.279475982532751,
+ "grad_norm": 0.02369544468820095,
+ "learning_rate": 0.0005459999999999999,
+ "loss": 7.442216873168945,
+ "step": 92
+ },
+ {
+ "epoch": 1.2934497816593886,
+ "grad_norm": 0.008474506437778473,
+ "learning_rate": 0.000552,
+ "loss": 7.4294915199279785,
+ "step": 93
+ },
+ {
+ "epoch": 1.3074235807860262,
+ "grad_norm": 0.017043249681591988,
+ "learning_rate": 0.000558,
+ "loss": 7.406932830810547,
+ "step": 94
+ },
+ {
+ "epoch": 1.3213973799126637,
+ "grad_norm": 0.011813412420451641,
+ "learning_rate": 0.0005639999999999999,
+ "loss": 7.415900230407715,
+ "step": 95
+ },
+ {
+ "epoch": 1.3353711790393012,
+ "grad_norm": 0.01086883433163166,
+ "learning_rate": 0.00057,
+ "loss": 7.452998638153076,
+ "step": 96
+ },
+ {
+ "epoch": 1.3493449781659388,
+ "grad_norm": 0.014700951054692268,
+ "learning_rate": 0.0005759999999999999,
+ "loss": 7.410249710083008,
+ "step": 97
+ },
+ {
+ "epoch": 1.3633187772925766,
+ "grad_norm": 0.014512976631522179,
+ "learning_rate": 0.0005819999999999999,
+ "loss": 7.429132461547852,
+ "step": 98
+ },
+ {
+ "epoch": 1.3772925764192139,
+ "grad_norm": 0.012519920244812965,
+ "learning_rate": 0.000588,
+ "loss": 7.430817604064941,
+ "step": 99
+ },
+ {
+ "epoch": 1.3912663755458516,
+ "grad_norm": 0.007132664322853088,
+ "learning_rate": 0.0005939999999999999,
+ "loss": 7.405234336853027,
+ "step": 100
+ },
+ {
+ "epoch": 1.405240174672489,
+ "grad_norm": 0.019489388912916183,
+ "learning_rate": 0.0006,
+ "loss": 7.427791595458984,
+ "step": 101
+ },
+ {
+ "epoch": 1.4192139737991267,
+ "grad_norm": 0.027252664789557457,
+ "learning_rate": 0.0006,
+ "loss": 7.41761589050293,
+ "step": 102
+ },
+ {
+ "epoch": 1.4331877729257643,
+ "grad_norm": 0.02501637488603592,
+ "learning_rate": 0.0006,
+ "loss": 7.416735649108887,
+ "step": 103
+ },
+ {
+ "epoch": 1.4471615720524018,
+ "grad_norm": 0.00989772193133831,
+ "learning_rate": 0.0006,
+ "loss": 7.3756632804870605,
+ "step": 104
+ },
+ {
+ "epoch": 1.4611353711790394,
+ "grad_norm": 0.01622949168086052,
+ "learning_rate": 0.0006,
+ "loss": 7.400811672210693,
+ "step": 105
+ },
+ {
+ "epoch": 1.475109170305677,
+ "grad_norm": 0.012407763861119747,
+ "learning_rate": 0.0006,
+ "loss": 7.37675666809082,
+ "step": 106
+ },
+ {
+ "epoch": 1.4890829694323144,
+ "grad_norm": 0.010353565216064453,
+ "learning_rate": 0.0006,
+ "loss": 7.4013142585754395,
+ "step": 107
+ },
+ {
+ "epoch": 1.503056768558952,
+ "grad_norm": 0.0120553532615304,
+ "learning_rate": 0.0006,
+ "loss": 7.379873275756836,
+ "step": 108
+ },
+ {
+ "epoch": 1.5170305676855895,
+ "grad_norm": 0.01610225811600685,
+ "learning_rate": 0.0006,
+ "loss": 7.371631145477295,
+ "step": 109
+ },
+ {
+ "epoch": 1.531004366812227,
+ "grad_norm": 0.01159879844635725,
+ "learning_rate": 0.0006,
+ "loss": 7.389007568359375,
+ "step": 110
+ },
+ {
+ "epoch": 1.5449781659388646,
+ "grad_norm": 0.008013816550374031,
+ "learning_rate": 0.0006,
+ "loss": 7.395920753479004,
+ "step": 111
+ },
+ {
+ "epoch": 1.5589519650655022,
+ "grad_norm": 0.016334809362888336,
+ "learning_rate": 0.0006,
+ "loss": 7.393121719360352,
+ "step": 112
+ },
+ {
+ "epoch": 1.5729257641921397,
+ "grad_norm": 0.018821721896529198,
+ "learning_rate": 0.0006,
+ "loss": 7.367543697357178,
+ "step": 113
+ },
+ {
+ "epoch": 1.5868995633187772,
+ "grad_norm": 0.014088994823396206,
+ "learning_rate": 0.0006,
+ "loss": 7.385537624359131,
+ "step": 114
+ },
+ {
+ "epoch": 1.600873362445415,
+ "grad_norm": 0.005864050704985857,
+ "learning_rate": 0.0006,
+ "loss": 7.365346908569336,
+ "step": 115
+ },
+ {
+ "epoch": 1.6148471615720523,
+ "grad_norm": 0.011745430529117584,
+ "learning_rate": 0.0006,
+ "loss": 7.390227317810059,
+ "step": 116
+ },
+ {
+ "epoch": 1.62882096069869,
+ "grad_norm": 0.017205573618412018,
+ "learning_rate": 0.0006,
+ "loss": 7.3558549880981445,
+ "step": 117
+ },
+ {
+ "epoch": 1.6427947598253274,
+ "grad_norm": 0.021256346255540848,
+ "learning_rate": 0.0006,
+ "loss": 7.377346992492676,
+ "step": 118
+ },
+ {
+ "epoch": 1.6567685589519652,
+ "grad_norm": 0.019335204735398293,
+ "learning_rate": 0.0006,
+ "loss": 7.346194267272949,
+ "step": 119
+ },
+ {
+ "epoch": 1.6707423580786025,
+ "grad_norm": 0.008134003728628159,
+ "learning_rate": 0.0006,
+ "loss": 7.3063578605651855,
+ "step": 120
+ },
+ {
+ "epoch": 1.6847161572052403,
+ "grad_norm": 0.017613831907510757,
+ "learning_rate": 0.0006,
+ "loss": 7.338171005249023,
+ "step": 121
+ },
+ {
+ "epoch": 1.6986899563318776,
+ "grad_norm": 0.00890734139829874,
+ "learning_rate": 0.0006,
+ "loss": 7.355384826660156,
+ "step": 122
+ },
+ {
+ "epoch": 1.7126637554585153,
+ "grad_norm": 0.013448155485093594,
+ "learning_rate": 0.0006,
+ "loss": 7.344296455383301,
+ "step": 123
+ },
+ {
+ "epoch": 1.726637554585153,
+ "grad_norm": 0.009241633117198944,
+ "learning_rate": 0.0006,
+ "loss": 7.3454060554504395,
+ "step": 124
+ },
+ {
+ "epoch": 1.7406113537117904,
+ "grad_norm": 0.018561311066150665,
+ "learning_rate": 0.0006,
+ "loss": 7.32348108291626,
+ "step": 125
+ },
+ {
+ "epoch": 1.754585152838428,
+ "grad_norm": 0.025667704641819,
+ "learning_rate": 0.0006,
+ "loss": 7.356142520904541,
+ "step": 126
+ },
+ {
+ "epoch": 1.7685589519650655,
+ "grad_norm": 0.036232151091098785,
+ "learning_rate": 0.0006,
+ "loss": 7.363539695739746,
+ "step": 127
+ },
+ {
+ "epoch": 1.782532751091703,
+ "grad_norm": 0.059483714401721954,
+ "learning_rate": 0.0006,
+ "loss": 7.402331352233887,
+ "step": 128
+ },
+ {
+ "epoch": 1.7965065502183406,
+ "grad_norm": 0.05365009233355522,
+ "learning_rate": 0.0006,
+ "loss": 7.40457820892334,
+ "step": 129
+ },
+ {
+ "epoch": 1.8104803493449781,
+ "grad_norm": 0.02738415263593197,
+ "learning_rate": 0.0006,
+ "loss": 7.3756208419799805,
+ "step": 130
+ },
+ {
+ "epoch": 1.8244541484716157,
+ "grad_norm": 0.035017091780900955,
+ "learning_rate": 0.0006,
+ "loss": 7.379263877868652,
+ "step": 131
+ },
+ {
+ "epoch": 1.8384279475982532,
+ "grad_norm": 0.02819863148033619,
+ "learning_rate": 0.0006,
+ "loss": 7.327673435211182,
+ "step": 132
+ },
+ {
+ "epoch": 1.8524017467248908,
+ "grad_norm": 0.017754238098859787,
+ "learning_rate": 0.0006,
+ "loss": 7.333420276641846,
+ "step": 133
+ },
+ {
+ "epoch": 1.8663755458515285,
+ "grad_norm": 0.024685733020305634,
+ "learning_rate": 0.0006,
+ "loss": 7.34493350982666,
+ "step": 134
+ },
+ {
+ "epoch": 1.8803493449781659,
+ "grad_norm": 0.019426532089710236,
+ "learning_rate": 0.0006,
+ "loss": 7.328448295593262,
+ "step": 135
+ },
+ {
+ "epoch": 1.8943231441048036,
+ "grad_norm": 0.02106391079723835,
+ "learning_rate": 0.0006,
+ "loss": 7.327917098999023,
+ "step": 136
+ },
+ {
+ "epoch": 1.908296943231441,
+ "grad_norm": 0.02056879736483097,
+ "learning_rate": 0.0006,
+ "loss": 7.3311991691589355,
+ "step": 137
+ },
+ {
+ "epoch": 1.9222707423580787,
+ "grad_norm": 0.016865141689777374,
+ "learning_rate": 0.0006,
+ "loss": 7.312766075134277,
+ "step": 138
+ },
+ {
+ "epoch": 1.936244541484716,
+ "grad_norm": 0.017549855634570122,
+ "learning_rate": 0.0006,
+ "loss": 7.31191349029541,
+ "step": 139
+ },
+ {
+ "epoch": 1.9502183406113538,
+ "grad_norm": 0.015930157154798508,
+ "learning_rate": 0.0006,
+ "loss": 7.330901622772217,
+ "step": 140
+ },
+ {
+ "epoch": 1.9641921397379911,
+ "grad_norm": 0.013283558189868927,
+ "learning_rate": 0.0006,
+ "loss": 7.298318386077881,
+ "step": 141
+ },
+ {
+ "epoch": 1.9781659388646289,
+ "grad_norm": 0.012023469433188438,
+ "learning_rate": 0.0006,
+ "loss": 7.282337188720703,
+ "step": 142
+ },
+ {
+ "epoch": 1.9921397379912664,
+ "grad_norm": 0.01459511648863554,
+ "learning_rate": 0.0006,
+ "loss": 7.282924175262451,
+ "step": 143
+ },
+ {
+ "epoch": 2.0,
+ "grad_norm": 0.017474235966801643,
+ "learning_rate": 0.0006,
+ "loss": 7.254333972930908,
+ "step": 144
+ },
+ {
+ "epoch": 2.0,
+ "eval_loss": 7.318708896636963,
+ "eval_runtime": 59.6555,
+ "eval_samples_per_second": 40.935,
+ "eval_steps_per_second": 1.291,
+ "step": 144
+ },
+ {
+ "epoch": 2.0139737991266378,
+ "grad_norm": 0.019757507368922234,
+ "learning_rate": 0.0006,
+ "loss": 7.30288553237915,
+ "step": 145
+ },
+ {
+ "epoch": 2.027947598253275,
+ "grad_norm": 0.014571248553693295,
+ "learning_rate": 0.0006,
+ "loss": 7.274681568145752,
+ "step": 146
+ },
+ {
+ "epoch": 2.041921397379913,
+ "grad_norm": 0.015744198113679886,
+ "learning_rate": 0.0006,
+ "loss": 7.278632640838623,
+ "step": 147
+ },
+ {
+ "epoch": 2.05589519650655,
+ "grad_norm": 0.028048858046531677,
+ "learning_rate": 0.0006,
+ "loss": 7.26558780670166,
+ "step": 148
+ },
+ {
+ "epoch": 2.069868995633188,
+ "grad_norm": 0.06547307223081589,
+ "learning_rate": 0.0006,
+ "loss": 7.32567834854126,
+ "step": 149
+ },
+ {
+ "epoch": 2.0838427947598253,
+ "grad_norm": 0.05969763547182083,
+ "learning_rate": 0.0006,
+ "loss": 7.395836353302002,
+ "step": 150
+ },
+ {
+ "epoch": 2.097816593886463,
+ "grad_norm": 0.032574281096458435,
+ "learning_rate": 0.0006,
+ "loss": 7.318150997161865,
+ "step": 151
+ },
+ {
+ "epoch": 2.1117903930131003,
+ "grad_norm": 0.029452962800860405,
+ "learning_rate": 0.0006,
+ "loss": 7.308066368103027,
+ "step": 152
+ },
+ {
+ "epoch": 2.125764192139738,
+ "grad_norm": 0.03161991387605667,
+ "learning_rate": 0.0006,
+ "loss": 7.306290626525879,
+ "step": 153
+ },
+ {
+ "epoch": 2.1397379912663754,
+ "grad_norm": 0.027096295729279518,
+ "learning_rate": 0.0006,
+ "loss": 7.297796249389648,
+ "step": 154
+ },
+ {
+ "epoch": 2.153711790393013,
+ "grad_norm": 0.028657056391239166,
+ "learning_rate": 0.0006,
+ "loss": 7.317424774169922,
+ "step": 155
+ },
+ {
+ "epoch": 2.1676855895196505,
+ "grad_norm": 0.02046995982527733,
+ "learning_rate": 0.0006,
+ "loss": 7.297986030578613,
+ "step": 156
+ },
+ {
+ "epoch": 2.1816593886462883,
+ "grad_norm": 0.02220369316637516,
+ "learning_rate": 0.0006,
+ "loss": 7.268237590789795,
+ "step": 157
+ },
+ {
+ "epoch": 2.1956331877729256,
+ "grad_norm": 0.018381357192993164,
+ "learning_rate": 0.0006,
+ "loss": 7.259415626525879,
+ "step": 158
+ },
+ {
+ "epoch": 2.2096069868995634,
+ "grad_norm": 0.0204413291066885,
+ "learning_rate": 0.0006,
+ "loss": 7.258173942565918,
+ "step": 159
+ },
+ {
+ "epoch": 2.2235807860262007,
+ "grad_norm": 0.015874288976192474,
+ "learning_rate": 0.0006,
+ "loss": 7.281834602355957,
+ "step": 160
+ },
+ {
+ "epoch": 2.2375545851528384,
+ "grad_norm": 0.01634068600833416,
+ "learning_rate": 0.0006,
+ "loss": 7.2469868659973145,
+ "step": 161
+ },
+ {
+ "epoch": 2.251528384279476,
+ "grad_norm": 0.01728747971355915,
+ "learning_rate": 0.0006,
+ "loss": 7.280995845794678,
+ "step": 162
+ },
+ {
+ "epoch": 2.2655021834061135,
+ "grad_norm": 0.016341188922524452,
+ "learning_rate": 0.0006,
+ "loss": 7.273357391357422,
+ "step": 163
+ },
+ {
+ "epoch": 2.279475982532751,
+ "grad_norm": 0.016125807538628578,
+ "learning_rate": 0.0006,
+ "loss": 7.273510932922363,
+ "step": 164
+ },
+ {
+ "epoch": 2.2934497816593886,
+ "grad_norm": 0.01629287749528885,
+ "learning_rate": 0.0006,
+ "loss": 7.267029762268066,
+ "step": 165
+ },
+ {
+ "epoch": 2.3074235807860264,
+ "grad_norm": 0.015296131372451782,
+ "learning_rate": 0.0006,
+ "loss": 7.211989402770996,
+ "step": 166
+ },
+ {
+ "epoch": 2.3213973799126637,
+ "grad_norm": 0.01709570363163948,
+ "learning_rate": 0.0006,
+ "loss": 7.242851734161377,
+ "step": 167
+ },
+ {
+ "epoch": 2.3353711790393015,
+ "grad_norm": 0.012552388943731785,
+ "learning_rate": 0.0006,
+ "loss": 7.241696834564209,
+ "step": 168
+ },
+ {
+ "epoch": 2.349344978165939,
+ "grad_norm": 0.014571291394531727,
+ "learning_rate": 0.0006,
+ "loss": 7.270681381225586,
+ "step": 169
+ },
+ {
+ "epoch": 2.3633187772925766,
+ "grad_norm": 0.013389437459409237,
+ "learning_rate": 0.0006,
+ "loss": 7.191600799560547,
+ "step": 170
+ },
+ {
+ "epoch": 2.377292576419214,
+ "grad_norm": 0.011895314790308475,
+ "learning_rate": 0.0006,
+ "loss": 7.213610649108887,
+ "step": 171
+ },
+ {
+ "epoch": 2.3912663755458516,
+ "grad_norm": 0.01345642190426588,
+ "learning_rate": 0.0006,
+ "loss": 7.193787574768066,
+ "step": 172
+ },
+ {
+ "epoch": 2.405240174672489,
+ "grad_norm": 0.0091794328764081,
+ "learning_rate": 0.0006,
+ "loss": 7.233364105224609,
+ "step": 173
+ },
+ {
+ "epoch": 2.4192139737991267,
+ "grad_norm": 0.013767086900770664,
+ "learning_rate": 0.0006,
+ "loss": 7.217272758483887,
+ "step": 174
+ },
+ {
+ "epoch": 2.433187772925764,
+ "grad_norm": 0.015134960412979126,
+ "learning_rate": 0.0006,
+ "loss": 7.192791938781738,
+ "step": 175
+ },
+ {
+ "epoch": 2.447161572052402,
+ "grad_norm": 0.01790648326277733,
+ "learning_rate": 0.0006,
+ "loss": 7.229467391967773,
+ "step": 176
+ },
+ {
+ "epoch": 2.461135371179039,
+ "grad_norm": 0.015362377278506756,
+ "learning_rate": 0.0006,
+ "loss": 7.180346488952637,
+ "step": 177
+ },
+ {
+ "epoch": 2.475109170305677,
+ "grad_norm": 0.010505498386919498,
+ "learning_rate": 0.0006,
+ "loss": 7.188833236694336,
+ "step": 178
+ },
+ {
+ "epoch": 2.489082969432314,
+ "grad_norm": 0.009129747748374939,
+ "learning_rate": 0.0006,
+ "loss": 7.203153133392334,
+ "step": 179
+ },
+ {
+ "epoch": 2.503056768558952,
+ "grad_norm": 0.01147883478552103,
+ "learning_rate": 0.0006,
+ "loss": 7.163824558258057,
+ "step": 180
+ },
+ {
+ "epoch": 2.5170305676855893,
+ "grad_norm": 0.011338942684233189,
+ "learning_rate": 0.0006,
+ "loss": 7.16287899017334,
+ "step": 181
+ },
+ {
+ "epoch": 2.531004366812227,
+ "grad_norm": 0.013277675025165081,
+ "learning_rate": 0.0006,
+ "loss": 7.155905723571777,
+ "step": 182
+ },
+ {
+ "epoch": 2.544978165938865,
+ "grad_norm": 0.015269347466528416,
+ "learning_rate": 0.0006,
+ "loss": 7.189371109008789,
+ "step": 183
+ },
+ {
+ "epoch": 2.558951965065502,
+ "grad_norm": 0.017905596643686295,
+ "learning_rate": 0.0006,
+ "loss": 7.177547931671143,
+ "step": 184
+ },
+ {
+ "epoch": 2.5729257641921395,
+ "grad_norm": 0.024431610479950905,
+ "learning_rate": 0.0006,
+ "loss": 7.192169189453125,
+ "step": 185
+ },
+ {
+ "epoch": 2.5868995633187772,
+ "grad_norm": 0.03656961768865585,
+ "learning_rate": 0.0006,
+ "loss": 7.124387741088867,
+ "step": 186
+ },
+ {
+ "epoch": 2.600873362445415,
+ "grad_norm": 0.04337048903107643,
+ "learning_rate": 0.0006,
+ "loss": 7.140231609344482,
+ "step": 187
+ },
+ {
+ "epoch": 2.6148471615720523,
+ "grad_norm": 0.04047228768467903,
+ "learning_rate": 0.0006,
+ "loss": 7.171927452087402,
+ "step": 188
+ },
+ {
+ "epoch": 2.62882096069869,
+ "grad_norm": 0.07825770974159241,
+ "learning_rate": 0.0006,
+ "loss": 7.207631587982178,
+ "step": 189
+ },
+ {
+ "epoch": 2.6427947598253274,
+ "grad_norm": 0.07118063420057297,
+ "learning_rate": 0.0006,
+ "loss": 7.209488391876221,
+ "step": 190
+ },
+ {
+ "epoch": 2.656768558951965,
+ "grad_norm": 0.05153293535113335,
+ "learning_rate": 0.0006,
+ "loss": 7.14793062210083,
+ "step": 191
+ },
+ {
+ "epoch": 2.6707423580786025,
+ "grad_norm": 0.03318839520215988,
+ "learning_rate": 0.0006,
+ "loss": 7.163010597229004,
+ "step": 192
+ },
+ {
+ "epoch": 2.6847161572052403,
+ "grad_norm": 0.029333539307117462,
+ "learning_rate": 0.0006,
+ "loss": 7.159955978393555,
+ "step": 193
+ },
+ {
+ "epoch": 2.6986899563318776,
+ "grad_norm": 0.029405880719423294,
+ "learning_rate": 0.0006,
+ "loss": 7.136178493499756,
+ "step": 194
+ },
+ {
+ "epoch": 2.7126637554585153,
+ "grad_norm": 0.037815988063812256,
+ "learning_rate": 0.0006,
+ "loss": 7.183186054229736,
+ "step": 195
+ },
+ {
+ "epoch": 2.726637554585153,
+ "grad_norm": 0.020307250320911407,
+ "learning_rate": 0.0006,
+ "loss": 7.133164882659912,
+ "step": 196
+ },
+ {
+ "epoch": 2.7406113537117904,
+ "grad_norm": 0.026048069819808006,
+ "learning_rate": 0.0006,
+ "loss": 7.124977111816406,
+ "step": 197
+ },
+ {
+ "epoch": 2.7545851528384278,
+ "grad_norm": 0.027316724881529808,
+ "learning_rate": 0.0006,
+ "loss": 7.07265567779541,
+ "step": 198
+ },
+ {
+ "epoch": 2.7685589519650655,
+ "grad_norm": 0.021629009395837784,
+ "learning_rate": 0.0006,
+ "loss": 7.150020599365234,
+ "step": 199
+ },
+ {
+ "epoch": 2.7825327510917033,
+ "grad_norm": 0.01605929434299469,
+ "learning_rate": 0.0006,
+ "loss": 7.10919189453125,
+ "step": 200
+ },
+ {
+ "epoch": 2.7965065502183406,
+ "grad_norm": 0.020800089463591576,
+ "learning_rate": 0.0006,
+ "loss": 7.132045269012451,
+ "step": 201
+ },
+ {
+ "epoch": 2.810480349344978,
+ "grad_norm": 0.019225774332880974,
+ "learning_rate": 0.0006,
+ "loss": 7.128364562988281,
+ "step": 202
+ },
+ {
+ "epoch": 2.8244541484716157,
+ "grad_norm": 0.01561372634023428,
+ "learning_rate": 0.0006,
+ "loss": 7.083669662475586,
+ "step": 203
+ },
+ {
+ "epoch": 2.8384279475982535,
+ "grad_norm": 0.021969519555568695,
+ "learning_rate": 0.0006,
+ "loss": 7.050841331481934,
+ "step": 204
+ },
+ {
+ "epoch": 2.8524017467248908,
+ "grad_norm": 0.021453695371747017,
+ "learning_rate": 0.0006,
+ "loss": 7.097085475921631,
+ "step": 205
+ },
+ {
+ "epoch": 2.8663755458515285,
+ "grad_norm": 0.022836022078990936,
+ "learning_rate": 0.0006,
+ "loss": 7.10405969619751,
+ "step": 206
+ },
+ {
+ "epoch": 2.880349344978166,
+ "grad_norm": 0.02430851384997368,
+ "learning_rate": 0.0006,
+ "loss": 7.045181751251221,
+ "step": 207
+ },
+ {
+ "epoch": 2.8943231441048036,
+ "grad_norm": 0.020166993141174316,
+ "learning_rate": 0.0006,
+ "loss": 7.068183898925781,
+ "step": 208
+ },
+ {
+ "epoch": 2.908296943231441,
+ "grad_norm": 0.0118505684658885,
+ "learning_rate": 0.0006,
+ "loss": 7.07132625579834,
+ "step": 209
+ },
+ {
+ "epoch": 2.9222707423580787,
+ "grad_norm": 0.023772427812218666,
+ "learning_rate": 0.0006,
+ "loss": 7.055238723754883,
+ "step": 210
+ },
+ {
+ "epoch": 2.936244541484716,
+ "grad_norm": 0.02350712940096855,
+ "learning_rate": 0.0006,
+ "loss": 7.055898666381836,
+ "step": 211
+ },
+ {
+ "epoch": 2.950218340611354,
+ "grad_norm": 0.017244907096028328,
+ "learning_rate": 0.0006,
+ "loss": 7.0099711418151855,
+ "step": 212
+ },
+ {
+ "epoch": 2.964192139737991,
+ "grad_norm": 0.021565575152635574,
+ "learning_rate": 0.0006,
+ "loss": 6.991751670837402,
+ "step": 213
+ },
+ {
+ "epoch": 2.978165938864629,
+ "grad_norm": 0.03241860866546631,
+ "learning_rate": 0.0006,
+ "loss": 7.005980491638184,
+ "step": 214
+ },
+ {
+ "epoch": 2.992139737991266,
+ "grad_norm": 0.04613726586103439,
+ "learning_rate": 0.0006,
+ "loss": 7.027859687805176,
+ "step": 215
+ },
+ {
+ "epoch": 3.0,
+ "grad_norm": 0.06404894590377808,
+ "learning_rate": 0.0006,
+ "loss": 7.123725891113281,
+ "step": 216
+ },
+ {
+ "epoch": 3.0,
+ "eval_loss": 7.070647716522217,
+ "eval_runtime": 58.5801,
+ "eval_samples_per_second": 41.687,
+ "eval_steps_per_second": 1.314,
+ "step": 216
+ },
+ {
+ "epoch": 3.0139737991266378,
+ "grad_norm": 0.05542432889342308,
+ "learning_rate": 0.0006,
+ "loss": 7.078299522399902,
+ "step": 217
+ },
+ {
+ "epoch": 3.027947598253275,
+ "grad_norm": 0.026484526693820953,
+ "learning_rate": 0.0006,
+ "loss": 6.961159706115723,
+ "step": 218
+ },
+ {
+ "epoch": 3.041921397379913,
+ "grad_norm": 0.028276391327381134,
+ "learning_rate": 0.0006,
+ "loss": 7.019961833953857,
+ "step": 219
+ },
+ {
+ "epoch": 3.05589519650655,
+ "grad_norm": 0.024486595764756203,
+ "learning_rate": 0.0006,
+ "loss": 7.039677619934082,
+ "step": 220
+ },
+ {
+ "epoch": 3.069868995633188,
+ "grad_norm": 0.017620893195271492,
+ "learning_rate": 0.0006,
+ "loss": 6.9270782470703125,
+ "step": 221
+ },
+ {
+ "epoch": 3.0838427947598253,
+ "grad_norm": 0.02547302283346653,
+ "learning_rate": 0.0006,
+ "loss": 6.924233436584473,
+ "step": 222
+ },
+ {
+ "epoch": 3.097816593886463,
+ "grad_norm": 0.02539009042084217,
+ "learning_rate": 0.0006,
+ "loss": 6.973979949951172,
+ "step": 223
+ },
+ {
+ "epoch": 3.1117903930131003,
+ "grad_norm": 0.03256227448582649,
+ "learning_rate": 0.0006,
+ "loss": 6.933725357055664,
+ "step": 224
+ },
+ {
+ "epoch": 3.125764192139738,
+ "grad_norm": 0.04651800915598869,
+ "learning_rate": 0.0006,
+ "loss": 6.951044082641602,
+ "step": 225
+ },
+ {
+ "epoch": 3.1397379912663754,
+ "grad_norm": 0.04822975769639015,
+ "learning_rate": 0.0006,
+ "loss": 6.9614410400390625,
+ "step": 226
+ },
+ {
+ "epoch": 3.153711790393013,
+ "grad_norm": 0.031165020540356636,
+ "learning_rate": 0.0006,
+ "loss": 6.91193151473999,
+ "step": 227
+ },
+ {
+ "epoch": 3.1676855895196505,
+ "grad_norm": 0.029754292219877243,
+ "learning_rate": 0.0006,
+ "loss": 6.91802978515625,
+ "step": 228
+ },
+ {
+ "epoch": 3.1816593886462883,
+ "grad_norm": 0.02193152718245983,
+ "learning_rate": 0.0006,
+ "loss": 6.928090572357178,
+ "step": 229
+ },
+ {
+ "epoch": 3.1956331877729256,
+ "grad_norm": 0.02428470179438591,
+ "learning_rate": 0.0006,
+ "loss": 6.90396785736084,
+ "step": 230
+ },
+ {
+ "epoch": 3.2096069868995634,
+ "grad_norm": 0.023374345153570175,
+ "learning_rate": 0.0006,
+ "loss": 6.885858535766602,
+ "step": 231
+ },
+ {
+ "epoch": 3.2235807860262007,
+ "grad_norm": 0.023405000567436218,
+ "learning_rate": 0.0006,
+ "loss": 6.8855767250061035,
+ "step": 232
+ },
+ {
+ "epoch": 3.2375545851528384,
+ "grad_norm": 0.017277134582400322,
+ "learning_rate": 0.0006,
+ "loss": 6.901449203491211,
+ "step": 233
+ },
+ {
+ "epoch": 3.251528384279476,
+ "grad_norm": 0.019496750086545944,
+ "learning_rate": 0.0006,
+ "loss": 6.885687828063965,
+ "step": 234
+ },
+ {
+ "epoch": 3.2655021834061135,
+ "grad_norm": 0.016893276944756508,
+ "learning_rate": 0.0006,
+ "loss": 6.872369766235352,
+ "step": 235
+ },
+ {
+ "epoch": 3.279475982532751,
+ "grad_norm": 0.02058715932071209,
+ "learning_rate": 0.0006,
+ "loss": 6.823280334472656,
+ "step": 236
+ },
+ {
+ "epoch": 3.2934497816593886,
+ "grad_norm": 0.015530755743384361,
+ "learning_rate": 0.0006,
+ "loss": 6.860499858856201,
+ "step": 237
+ },
+ {
+ "epoch": 3.3074235807860264,
+ "grad_norm": 0.01934926211833954,
+ "learning_rate": 0.0006,
+ "loss": 6.817448616027832,
+ "step": 238
+ },
+ {
+ "epoch": 3.3213973799126637,
+ "grad_norm": 0.017413552850484848,
+ "learning_rate": 0.0006,
+ "loss": 6.881955146789551,
+ "step": 239
+ },
+ {
+ "epoch": 3.3353711790393015,
+ "grad_norm": 0.026478875428438187,
+ "learning_rate": 0.0006,
+ "loss": 6.827404499053955,
+ "step": 240
+ },
+ {
+ "epoch": 3.349344978165939,
+ "grad_norm": 0.032611701637506485,
+ "learning_rate": 0.0006,
+ "loss": 6.78285551071167,
+ "step": 241
+ },
+ {
+ "epoch": 3.3633187772925766,
+ "grad_norm": 0.041900306940078735,
+ "learning_rate": 0.0006,
+ "loss": 6.783053874969482,
+ "step": 242
+ },
+ {
+ "epoch": 3.377292576419214,
+ "grad_norm": 0.04149497672915459,
+ "learning_rate": 0.0006,
+ "loss": 6.745220184326172,
+ "step": 243
+ },
+ {
+ "epoch": 3.3912663755458516,
+ "grad_norm": 0.023949826136231422,
+ "learning_rate": 0.0006,
+ "loss": 6.841533184051514,
+ "step": 244
+ },
+ {
+ "epoch": 3.405240174672489,
+ "grad_norm": 0.028601842001080513,
+ "learning_rate": 0.0006,
+ "loss": 6.852852821350098,
+ "step": 245
+ },
+ {
+ "epoch": 3.4192139737991267,
+ "grad_norm": 0.027365155518054962,
+ "learning_rate": 0.0006,
+ "loss": 6.74376106262207,
+ "step": 246
+ },
+ {
+ "epoch": 3.433187772925764,
+ "grad_norm": 0.02856568619608879,
+ "learning_rate": 0.0006,
+ "loss": 6.74906063079834,
+ "step": 247
+ },
+ {
+ "epoch": 3.447161572052402,
+ "grad_norm": 0.019172416999936104,
+ "learning_rate": 0.0006,
+ "loss": 6.809206008911133,
+ "step": 248
+ },
+ {
+ "epoch": 3.461135371179039,
+ "grad_norm": 0.02161746844649315,
+ "learning_rate": 0.0006,
+ "loss": 6.789888381958008,
+ "step": 249
+ },
+ {
+ "epoch": 3.475109170305677,
+ "grad_norm": 0.020273666828870773,
+ "learning_rate": 0.0006,
+ "loss": 6.73512077331543,
+ "step": 250
+ },
+ {
+ "epoch": 3.489082969432314,
+ "grad_norm": 0.022329283878207207,
+ "learning_rate": 0.0006,
+ "loss": 6.826148509979248,
+ "step": 251
+ },
+ {
+ "epoch": 3.503056768558952,
+ "grad_norm": 0.02284284122288227,
+ "learning_rate": 0.0006,
+ "loss": 6.748521327972412,
+ "step": 252
+ },
+ {
+ "epoch": 3.5170305676855893,
+ "grad_norm": 0.026368247345089912,
+ "learning_rate": 0.0006,
+ "loss": 6.645879745483398,
+ "step": 253
+ },
+ {
+ "epoch": 3.531004366812227,
+ "grad_norm": 0.03627244383096695,
+ "learning_rate": 0.0006,
+ "loss": 6.666064739227295,
+ "step": 254
+ },
+ {
+ "epoch": 3.544978165938865,
+ "grad_norm": 0.04140935093164444,
+ "learning_rate": 0.0006,
+ "loss": 6.712275981903076,
+ "step": 255
+ },
+ {
+ "epoch": 3.558951965065502,
+ "grad_norm": 0.0390462800860405,
+ "learning_rate": 0.0006,
+ "loss": 6.665900230407715,
+ "step": 256
+ },
+ {
+ "epoch": 3.5729257641921395,
+ "grad_norm": 0.05178583785891533,
+ "learning_rate": 0.0006,
+ "loss": 6.753298759460449,
+ "step": 257
+ },
+ {
+ "epoch": 3.5868995633187772,
+ "grad_norm": 0.054328061640262604,
+ "learning_rate": 0.0006,
+ "loss": 6.732633113861084,
+ "step": 258
+ },
+ {
+ "epoch": 3.600873362445415,
+ "grad_norm": 0.04590460658073425,
+ "learning_rate": 0.0006,
+ "loss": 6.65250301361084,
+ "step": 259
+ },
+ {
+ "epoch": 3.6148471615720523,
+ "grad_norm": 0.044331666082143784,
+ "learning_rate": 0.0006,
+ "loss": 6.68654203414917,
+ "step": 260
+ },
+ {
+ "epoch": 3.62882096069869,
+ "grad_norm": 0.05308730527758598,
+ "learning_rate": 0.0006,
+ "loss": 6.642482757568359,
+ "step": 261
+ },
+ {
+ "epoch": 3.6427947598253274,
+ "grad_norm": 0.03999016433954239,
+ "learning_rate": 0.0006,
+ "loss": 6.625227451324463,
+ "step": 262
+ },
+ {
+ "epoch": 3.656768558951965,
+ "grad_norm": 0.04088086634874344,
+ "learning_rate": 0.0006,
+ "loss": 6.67623233795166,
+ "step": 263
+ },
+ {
+ "epoch": 3.6707423580786025,
+ "grad_norm": 0.03615636005997658,
+ "learning_rate": 0.0006,
+ "loss": 6.6869001388549805,
+ "step": 264
+ },
+ {
+ "epoch": 3.6847161572052403,
+ "grad_norm": 0.02817360684275627,
+ "learning_rate": 0.0006,
+ "loss": 6.674601078033447,
+ "step": 265
+ },
+ {
+ "epoch": 3.6986899563318776,
+ "grad_norm": 0.032586123794317245,
+ "learning_rate": 0.0006,
+ "loss": 6.612994194030762,
+ "step": 266
+ },
+ {
+ "epoch": 3.7126637554585153,
+ "grad_norm": 0.031117867678403854,
+ "learning_rate": 0.0006,
+ "loss": 6.6244354248046875,
+ "step": 267
+ },
+ {
+ "epoch": 3.726637554585153,
+ "grad_norm": 0.028330687433481216,
+ "learning_rate": 0.0006,
+ "loss": 6.669557094573975,
+ "step": 268
+ },
+ {
+ "epoch": 3.7406113537117904,
+ "grad_norm": 0.021489202976226807,
+ "learning_rate": 0.0006,
+ "loss": 6.632035732269287,
+ "step": 269
+ },
+ {
+ "epoch": 3.7545851528384278,
+ "grad_norm": 0.024269424378871918,
+ "learning_rate": 0.0006,
+ "loss": 6.660867691040039,
+ "step": 270
+ },
+ {
+ "epoch": 3.7685589519650655,
+ "grad_norm": 0.02237873338162899,
+ "learning_rate": 0.0006,
+ "loss": 6.602587699890137,
+ "step": 271
+ },
+ {
+ "epoch": 3.7825327510917033,
+ "grad_norm": 0.026503929868340492,
+ "learning_rate": 0.0006,
+ "loss": 6.599200248718262,
+ "step": 272
+ },
+ {
+ "epoch": 3.7965065502183406,
+ "grad_norm": 0.03687124326825142,
+ "learning_rate": 0.0006,
+ "loss": 6.567296504974365,
+ "step": 273
+ },
+ {
+ "epoch": 3.810480349344978,
+ "grad_norm": 0.04708952456712723,
+ "learning_rate": 0.0006,
+ "loss": 6.621265411376953,
+ "step": 274
+ },
+ {
+ "epoch": 3.8244541484716157,
+ "grad_norm": 0.041846614331007004,
+ "learning_rate": 0.0006,
+ "loss": 6.550536155700684,
+ "step": 275
+ },
+ {
+ "epoch": 3.8384279475982535,
+ "grad_norm": 0.031249675899744034,
+ "learning_rate": 0.0006,
+ "loss": 6.54278564453125,
+ "step": 276
+ },
+ {
+ "epoch": 3.8524017467248908,
+ "grad_norm": 0.023082256317138672,
+ "learning_rate": 0.0006,
+ "loss": 6.544787406921387,
+ "step": 277
+ },
+ {
+ "epoch": 3.8663755458515285,
+ "grad_norm": 0.020995570346713066,
+ "learning_rate": 0.0006,
+ "loss": 6.518143653869629,
+ "step": 278
+ },
+ {
+ "epoch": 3.880349344978166,
+ "grad_norm": 0.020839324221014977,
+ "learning_rate": 0.0006,
+ "loss": 6.546667098999023,
+ "step": 279
+ },
+ {
+ "epoch": 3.8943231441048036,
+ "grad_norm": 0.018377432599663734,
+ "learning_rate": 0.0006,
+ "loss": 6.478307723999023,
+ "step": 280
+ },
+ {
+ "epoch": 3.908296943231441,
+ "grad_norm": 0.021713724359869957,
+ "learning_rate": 0.0006,
+ "loss": 6.485574245452881,
+ "step": 281
+ },
+ {
+ "epoch": 3.9222707423580787,
+ "grad_norm": 0.01359301246702671,
+ "learning_rate": 0.0006,
+ "loss": 6.570268630981445,
+ "step": 282
+ },
+ {
+ "epoch": 3.936244541484716,
+ "grad_norm": 0.016233332455158234,
+ "learning_rate": 0.0006,
+ "loss": 6.57150936126709,
+ "step": 283
+ },
+ {
+ "epoch": 3.950218340611354,
+ "grad_norm": 0.020687608048319817,
+ "learning_rate": 0.0006,
+ "loss": 6.527956008911133,
+ "step": 284
+ },
+ {
+ "epoch": 3.964192139737991,
+ "grad_norm": 0.02032964862883091,
+ "learning_rate": 0.0006,
+ "loss": 6.468169689178467,
+ "step": 285
+ },
+ {
+ "epoch": 3.978165938864629,
+ "grad_norm": 0.016676288098096848,
+ "learning_rate": 0.0006,
+ "loss": 6.448418617248535,
+ "step": 286
+ },
+ {
+ "epoch": 3.992139737991266,
+ "grad_norm": 0.0168539360165596,
+ "learning_rate": 0.0006,
+ "loss": 6.461180686950684,
+ "step": 287
+ },
+ {
+ "epoch": 4.0,
+ "grad_norm": 0.02666233666241169,
+ "learning_rate": 0.0006,
+ "loss": 6.344893932342529,
+ "step": 288
+ },
+ {
+ "epoch": 4.0,
+ "eval_loss": 6.5384979248046875,
+ "eval_runtime": 58.814,
+ "eval_samples_per_second": 41.521,
+ "eval_steps_per_second": 1.309,
+ "step": 288
+ },
+ {
+ "epoch": 4.013973799126638,
+ "grad_norm": 0.04154708981513977,
+ "learning_rate": 0.0006,
+ "loss": 6.41855525970459,
+ "step": 289
+ },
+ {
+ "epoch": 4.0279475982532755,
+ "grad_norm": 0.04895668104290962,
+ "learning_rate": 0.0006,
+ "loss": 6.499302864074707,
+ "step": 290
+ },
+ {
+ "epoch": 4.041921397379912,
+ "grad_norm": 0.034464482218027115,
+ "learning_rate": 0.0006,
+ "loss": 6.438611030578613,
+ "step": 291
+ },
+ {
+ "epoch": 4.05589519650655,
+ "grad_norm": 0.06845773756504059,
+ "learning_rate": 0.0006,
+ "loss": 6.538361072540283,
+ "step": 292
+ },
+ {
+ "epoch": 4.069868995633188,
+ "grad_norm": 0.053768858313560486,
+ "learning_rate": 0.0006,
+ "loss": 6.527524948120117,
+ "step": 293
+ },
+ {
+ "epoch": 4.083842794759826,
+ "grad_norm": 0.039037931710481644,
+ "learning_rate": 0.0006,
+ "loss": 6.503847122192383,
+ "step": 294
+ },
+ {
+ "epoch": 4.097816593886463,
+ "grad_norm": 0.03962196782231331,
+ "learning_rate": 0.0006,
+ "loss": 6.475178241729736,
+ "step": 295
+ },
+ {
+ "epoch": 4.1117903930131,
+ "grad_norm": 0.049101535230875015,
+ "learning_rate": 0.0006,
+ "loss": 6.474587917327881,
+ "step": 296
+ },
+ {
+ "epoch": 4.125764192139738,
+ "grad_norm": 0.02830282226204872,
+ "learning_rate": 0.0006,
+ "loss": 6.45065975189209,
+ "step": 297
+ },
+ {
+ "epoch": 4.139737991266376,
+ "grad_norm": 0.023153482005000114,
+ "learning_rate": 0.0006,
+ "loss": 6.5218706130981445,
+ "step": 298
+ },
+ {
+ "epoch": 4.153711790393013,
+ "grad_norm": 0.025064002722501755,
+ "learning_rate": 0.0006,
+ "loss": 6.456475257873535,
+ "step": 299
+ },
+ {
+ "epoch": 4.1676855895196505,
+ "grad_norm": 0.02357092685997486,
+ "learning_rate": 0.0006,
+ "loss": 6.485952377319336,
+ "step": 300
+ },
+ {
+ "epoch": 4.181659388646288,
+ "grad_norm": 0.021832305938005447,
+ "learning_rate": 0.0006,
+ "loss": 6.425506591796875,
+ "step": 301
+ },
+ {
+ "epoch": 4.195633187772926,
+ "grad_norm": 0.0227900929749012,
+ "learning_rate": 0.0006,
+ "loss": 6.439155578613281,
+ "step": 302
+ },
+ {
+ "epoch": 4.209606986899563,
+ "grad_norm": 0.015961607918143272,
+ "learning_rate": 0.0006,
+ "loss": 6.460862636566162,
+ "step": 303
+ },
+ {
+ "epoch": 4.223580786026201,
+ "grad_norm": 0.017876390367746353,
+ "learning_rate": 0.0006,
+ "loss": 6.44564151763916,
+ "step": 304
+ },
+ {
+ "epoch": 4.2375545851528384,
+ "grad_norm": 0.013703204691410065,
+ "learning_rate": 0.0006,
+ "loss": 6.418414115905762,
+ "step": 305
+ },
+ {
+ "epoch": 4.251528384279476,
+ "grad_norm": 0.015240306034684181,
+ "learning_rate": 0.0006,
+ "loss": 6.411238670349121,
+ "step": 306
+ },
+ {
+ "epoch": 4.265502183406113,
+ "grad_norm": 0.01587662845849991,
+ "learning_rate": 0.0006,
+ "loss": 6.395642280578613,
+ "step": 307
+ },
+ {
+ "epoch": 4.279475982532751,
+ "grad_norm": 0.014650700613856316,
+ "learning_rate": 0.0006,
+ "loss": 6.302793025970459,
+ "step": 308
+ },
+ {
+ "epoch": 4.293449781659389,
+ "grad_norm": 0.016533225774765015,
+ "learning_rate": 0.0006,
+ "loss": 6.380588054656982,
+ "step": 309
+ },
+ {
+ "epoch": 4.307423580786026,
+ "grad_norm": 0.019963741302490234,
+ "learning_rate": 0.0006,
+ "loss": 6.345336437225342,
+ "step": 310
+ },
+ {
+ "epoch": 4.321397379912664,
+ "grad_norm": 0.02237936295568943,
+ "learning_rate": 0.0006,
+ "loss": 6.387770652770996,
+ "step": 311
+ },
+ {
+ "epoch": 4.335371179039301,
+ "grad_norm": 0.026658328250050545,
+ "learning_rate": 0.0006,
+ "loss": 6.323662757873535,
+ "step": 312
+ },
+ {
+ "epoch": 4.349344978165939,
+ "grad_norm": 0.02852284163236618,
+ "learning_rate": 0.0006,
+ "loss": 6.382204055786133,
+ "step": 313
+ },
+ {
+ "epoch": 4.3633187772925766,
+ "grad_norm": 0.028815090656280518,
+ "learning_rate": 0.0006,
+ "loss": 6.402390480041504,
+ "step": 314
+ },
+ {
+ "epoch": 4.377292576419214,
+ "grad_norm": 0.029393529519438744,
+ "learning_rate": 0.0006,
+ "loss": 6.280278205871582,
+ "step": 315
+ },
+ {
+ "epoch": 4.391266375545851,
+ "grad_norm": 0.02186041697859764,
+ "learning_rate": 0.0006,
+ "loss": 6.302707672119141,
+ "step": 316
+ },
+ {
+ "epoch": 4.405240174672489,
+ "grad_norm": 0.02071218006312847,
+ "learning_rate": 0.0006,
+ "loss": 6.3365020751953125,
+ "step": 317
+ },
+ {
+ "epoch": 4.419213973799127,
+ "grad_norm": 0.02779117226600647,
+ "learning_rate": 0.0006,
+ "loss": 6.252157211303711,
+ "step": 318
+ },
+ {
+ "epoch": 4.4331877729257645,
+ "grad_norm": 0.03470654785633087,
+ "learning_rate": 0.0006,
+ "loss": 6.319070339202881,
+ "step": 319
+ },
+ {
+ "epoch": 4.447161572052401,
+ "grad_norm": 0.037067804485559464,
+ "learning_rate": 0.0006,
+ "loss": 6.314116477966309,
+ "step": 320
+ },
+ {
+ "epoch": 4.461135371179039,
+ "grad_norm": 0.032182756811380386,
+ "learning_rate": 0.0006,
+ "loss": 6.267298698425293,
+ "step": 321
+ },
+ {
+ "epoch": 4.475109170305677,
+ "grad_norm": 0.026305217295885086,
+ "learning_rate": 0.0006,
+ "loss": 6.331688404083252,
+ "step": 322
+ },
+ {
+ "epoch": 4.489082969432315,
+ "grad_norm": 0.027228357270359993,
+ "learning_rate": 0.0006,
+ "loss": 6.31139612197876,
+ "step": 323
+ },
+ {
+ "epoch": 4.503056768558952,
+ "grad_norm": 0.025781169533729553,
+ "learning_rate": 0.0006,
+ "loss": 6.194684982299805,
+ "step": 324
+ },
+ {
+ "epoch": 4.517030567685589,
+ "grad_norm": 0.029186977073550224,
+ "learning_rate": 0.0006,
+ "loss": 6.31306266784668,
+ "step": 325
+ },
+ {
+ "epoch": 4.531004366812227,
+ "grad_norm": 0.023608777672052383,
+ "learning_rate": 0.0006,
+ "loss": 6.285243034362793,
+ "step": 326
+ },
+ {
+ "epoch": 4.544978165938865,
+ "grad_norm": 0.02267594076693058,
+ "learning_rate": 0.0006,
+ "loss": 6.30342960357666,
+ "step": 327
+ },
+ {
+ "epoch": 4.558951965065502,
+ "grad_norm": 0.01926310732960701,
+ "learning_rate": 0.0006,
+ "loss": 6.280606269836426,
+ "step": 328
+ },
+ {
+ "epoch": 4.5729257641921395,
+ "grad_norm": 0.026484837755560875,
+ "learning_rate": 0.0006,
+ "loss": 6.237980365753174,
+ "step": 329
+ },
+ {
+ "epoch": 4.586899563318777,
+ "grad_norm": 0.026228854432702065,
+ "learning_rate": 0.0006,
+ "loss": 6.3018293380737305,
+ "step": 330
+ },
+ {
+ "epoch": 4.600873362445415,
+ "grad_norm": 0.022096967324614525,
+ "learning_rate": 0.0006,
+ "loss": 6.253863334655762,
+ "step": 331
+ },
+ {
+ "epoch": 4.614847161572053,
+ "grad_norm": 0.027223890647292137,
+ "learning_rate": 0.0006,
+ "loss": 6.166882038116455,
+ "step": 332
+ },
+ {
+ "epoch": 4.62882096069869,
+ "grad_norm": 0.03160124272108078,
+ "learning_rate": 0.0006,
+ "loss": 6.230309963226318,
+ "step": 333
+ },
+ {
+ "epoch": 4.642794759825327,
+ "grad_norm": 0.04110539332032204,
+ "learning_rate": 0.0006,
+ "loss": 6.2334818840026855,
+ "step": 334
+ },
+ {
+ "epoch": 4.656768558951965,
+ "grad_norm": 0.04355933889746666,
+ "learning_rate": 0.0006,
+ "loss": 6.160956382751465,
+ "step": 335
+ },
+ {
+ "epoch": 4.670742358078603,
+ "grad_norm": 0.03611086308956146,
+ "learning_rate": 0.0006,
+ "loss": 6.23399543762207,
+ "step": 336
+ },
+ {
+ "epoch": 4.68471615720524,
+ "grad_norm": 0.051811400800943375,
+ "learning_rate": 0.0006,
+ "loss": 6.241855144500732,
+ "step": 337
+ },
+ {
+ "epoch": 4.698689956331878,
+ "grad_norm": 0.057434193789958954,
+ "learning_rate": 0.0006,
+ "loss": 6.220993995666504,
+ "step": 338
+ },
+ {
+ "epoch": 4.712663755458515,
+ "grad_norm": 0.04420953616499901,
+ "learning_rate": 0.0006,
+ "loss": 6.300461292266846,
+ "step": 339
+ },
+ {
+ "epoch": 4.726637554585153,
+ "grad_norm": 0.03515457361936569,
+ "learning_rate": 0.0006,
+ "loss": 6.269349098205566,
+ "step": 340
+ },
+ {
+ "epoch": 4.74061135371179,
+ "grad_norm": 0.03222600743174553,
+ "learning_rate": 0.0006,
+ "loss": 6.248613357543945,
+ "step": 341
+ },
+ {
+ "epoch": 4.754585152838428,
+ "grad_norm": 0.0384046845138073,
+ "learning_rate": 0.0006,
+ "loss": 6.266979217529297,
+ "step": 342
+ },
+ {
+ "epoch": 4.7685589519650655,
+ "grad_norm": 0.0514867827296257,
+ "learning_rate": 0.0006,
+ "loss": 6.234103202819824,
+ "step": 343
+ },
+ {
+ "epoch": 4.782532751091703,
+ "grad_norm": 0.050747793167829514,
+ "learning_rate": 0.0006,
+ "loss": 6.322582244873047,
+ "step": 344
+ },
+ {
+ "epoch": 4.796506550218341,
+ "grad_norm": 0.040878988802433014,
+ "learning_rate": 0.0006,
+ "loss": 6.212509632110596,
+ "step": 345
+ },
+ {
+ "epoch": 4.810480349344978,
+ "grad_norm": 0.034655820578336716,
+ "learning_rate": 0.0006,
+ "loss": 6.265879154205322,
+ "step": 346
+ },
+ {
+ "epoch": 4.824454148471616,
+ "grad_norm": 0.04202224686741829,
+ "learning_rate": 0.0006,
+ "loss": 6.2896409034729,
+ "step": 347
+ },
+ {
+ "epoch": 4.8384279475982535,
+ "grad_norm": 0.03632277995347977,
+ "learning_rate": 0.0006,
+ "loss": 6.253917217254639,
+ "step": 348
+ },
+ {
+ "epoch": 4.85240174672489,
+ "grad_norm": 0.028645765036344528,
+ "learning_rate": 0.0006,
+ "loss": 6.207965850830078,
+ "step": 349
+ },
+ {
+ "epoch": 4.866375545851528,
+ "grad_norm": 0.025305170565843582,
+ "learning_rate": 0.0006,
+ "loss": 6.168068885803223,
+ "step": 350
+ },
+ {
+ "epoch": 4.880349344978166,
+ "grad_norm": 0.022137803956866264,
+ "learning_rate": 0.0006,
+ "loss": 6.235283374786377,
+ "step": 351
+ },
+ {
+ "epoch": 4.894323144104804,
+ "grad_norm": 0.01814538985490799,
+ "learning_rate": 0.0006,
+ "loss": 6.259641647338867,
+ "step": 352
+ },
+ {
+ "epoch": 4.908296943231441,
+ "grad_norm": 0.018465086817741394,
+ "learning_rate": 0.0006,
+ "loss": 6.111738681793213,
+ "step": 353
+ },
+ {
+ "epoch": 4.922270742358078,
+ "grad_norm": 0.018638933077454567,
+ "learning_rate": 0.0006,
+ "loss": 6.225642204284668,
+ "step": 354
+ },
+ {
+ "epoch": 4.936244541484716,
+ "grad_norm": 0.016461260616779327,
+ "learning_rate": 0.0006,
+ "loss": 6.140647888183594,
+ "step": 355
+ },
+ {
+ "epoch": 4.950218340611354,
+ "grad_norm": 0.015458385460078716,
+ "learning_rate": 0.0006,
+ "loss": 6.155224800109863,
+ "step": 356
+ },
+ {
+ "epoch": 4.964192139737992,
+ "grad_norm": 0.014934048056602478,
+ "learning_rate": 0.0006,
+ "loss": 6.101555824279785,
+ "step": 357
+ },
+ {
+ "epoch": 4.978165938864628,
+ "grad_norm": 0.013979545794427395,
+ "learning_rate": 0.0006,
+ "loss": 6.161718845367432,
+ "step": 358
+ },
+ {
+ "epoch": 4.992139737991266,
+ "grad_norm": 0.01640394888818264,
+ "learning_rate": 0.0006,
+ "loss": 6.055768966674805,
+ "step": 359
+ },
+ {
+ "epoch": 5.0,
+ "grad_norm": 0.015135680325329304,
+ "learning_rate": 0.0006,
+ "loss": 6.169695854187012,
+ "step": 360
+ },
+ {
+ "epoch": 5.0,
+ "eval_loss": 6.16408109664917,
+ "eval_runtime": 59.3649,
+ "eval_samples_per_second": 41.135,
+ "eval_steps_per_second": 1.297,
+ "step": 360
+ },
+ {
+ "epoch": 5.013973799126638,
+ "grad_norm": 0.012849048711359501,
+ "learning_rate": 0.0006,
+ "loss": 6.055830001831055,
+ "step": 361
+ },
+ {
+ "epoch": 5.0279475982532755,
+ "grad_norm": 0.018169576302170753,
+ "learning_rate": 0.0006,
+ "loss": 6.082655429840088,
+ "step": 362
+ },
+ {
+ "epoch": 5.041921397379912,
+ "grad_norm": 0.024293430149555206,
+ "learning_rate": 0.0006,
+ "loss": 6.109032154083252,
+ "step": 363
+ },
+ {
+ "epoch": 5.05589519650655,
+ "grad_norm": 0.03266787901520729,
+ "learning_rate": 0.0006,
+ "loss": 6.030613899230957,
+ "step": 364
+ },
+ {
+ "epoch": 5.069868995633188,
+ "grad_norm": 0.03572266176342964,
+ "learning_rate": 0.0006,
+ "loss": 6.050390720367432,
+ "step": 365
+ },
+ {
+ "epoch": 5.083842794759826,
+ "grad_norm": 0.032625213265419006,
+ "learning_rate": 0.0006,
+ "loss": 6.169313430786133,
+ "step": 366
+ },
+ {
+ "epoch": 5.097816593886463,
+ "grad_norm": 0.04051872715353966,
+ "learning_rate": 0.0006,
+ "loss": 6.10398006439209,
+ "step": 367
+ },
+ {
+ "epoch": 5.1117903930131,
+ "grad_norm": 0.0351213738322258,
+ "learning_rate": 0.0006,
+ "loss": 6.121973514556885,
+ "step": 368
+ },
+ {
+ "epoch": 5.125764192139738,
+ "grad_norm": 0.03523759916424751,
+ "learning_rate": 0.0006,
+ "loss": 6.0626654624938965,
+ "step": 369
+ },
+ {
+ "epoch": 5.139737991266376,
+ "grad_norm": 0.02706182189285755,
+ "learning_rate": 0.0006,
+ "loss": 6.049354076385498,
+ "step": 370
+ },
+ {
+ "epoch": 5.153711790393013,
+ "grad_norm": 0.024212589487433434,
+ "learning_rate": 0.0006,
+ "loss": 6.034826278686523,
+ "step": 371
+ },
+ {
+ "epoch": 5.1676855895196505,
+ "grad_norm": 0.024798082187771797,
+ "learning_rate": 0.0006,
+ "loss": 6.004058837890625,
+ "step": 372
+ },
+ {
+ "epoch": 5.181659388646288,
+ "grad_norm": 0.02325567416846752,
+ "learning_rate": 0.0006,
+ "loss": 5.986461639404297,
+ "step": 373
+ },
+ {
+ "epoch": 5.195633187772926,
+ "grad_norm": 0.019060570746660233,
+ "learning_rate": 0.0006,
+ "loss": 6.059736251831055,
+ "step": 374
+ },
+ {
+ "epoch": 5.209606986899563,
+ "grad_norm": 0.016822976991534233,
+ "learning_rate": 0.0006,
+ "loss": 6.072957515716553,
+ "step": 375
+ },
+ {
+ "epoch": 5.223580786026201,
+ "grad_norm": 0.017218658700585365,
+ "learning_rate": 0.0006,
+ "loss": 6.031739234924316,
+ "step": 376
+ },
+ {
+ "epoch": 5.2375545851528384,
+ "grad_norm": 0.014889383688569069,
+ "learning_rate": 0.0006,
+ "loss": 6.025674343109131,
+ "step": 377
+ },
+ {
+ "epoch": 5.251528384279476,
+ "grad_norm": 0.01708107627928257,
+ "learning_rate": 0.0006,
+ "loss": 6.090451240539551,
+ "step": 378
+ },
+ {
+ "epoch": 5.265502183406113,
+ "grad_norm": 0.020260317251086235,
+ "learning_rate": 0.0006,
+ "loss": 6.094257354736328,
+ "step": 379
+ },
+ {
+ "epoch": 5.279475982532751,
+ "grad_norm": 0.020110400393605232,
+ "learning_rate": 0.0006,
+ "loss": 6.021188735961914,
+ "step": 380
+ },
+ {
+ "epoch": 5.293449781659389,
+ "grad_norm": 0.02027830481529236,
+ "learning_rate": 0.0006,
+ "loss": 6.022156238555908,
+ "step": 381
+ },
+ {
+ "epoch": 5.307423580786026,
+ "grad_norm": 0.022747062146663666,
+ "learning_rate": 0.0006,
+ "loss": 6.011836051940918,
+ "step": 382
+ },
+ {
+ "epoch": 5.321397379912664,
+ "grad_norm": 0.01990230567753315,
+ "learning_rate": 0.0006,
+ "loss": 6.003054618835449,
+ "step": 383
+ },
+ {
+ "epoch": 5.335371179039301,
+ "grad_norm": 0.01596238650381565,
+ "learning_rate": 0.0006,
+ "loss": 5.991410732269287,
+ "step": 384
+ },
+ {
+ "epoch": 5.349344978165939,
+ "grad_norm": 0.015847105532884598,
+ "learning_rate": 0.0006,
+ "loss": 6.100622177124023,
+ "step": 385
+ },
+ {
+ "epoch": 5.3633187772925766,
+ "grad_norm": 0.016179397702217102,
+ "learning_rate": 0.0006,
+ "loss": 6.032659530639648,
+ "step": 386
+ },
+ {
+ "epoch": 5.377292576419214,
+ "grad_norm": 0.018256602808833122,
+ "learning_rate": 0.0006,
+ "loss": 5.961983680725098,
+ "step": 387
+ },
+ {
+ "epoch": 5.391266375545851,
+ "grad_norm": 0.02805912122130394,
+ "learning_rate": 0.0006,
+ "loss": 5.983541965484619,
+ "step": 388
+ },
+ {
+ "epoch": 5.405240174672489,
+ "grad_norm": 0.039082255214452744,
+ "learning_rate": 0.0006,
+ "loss": 6.021474361419678,
+ "step": 389
+ },
+ {
+ "epoch": 5.419213973799127,
+ "grad_norm": 0.036757659167051315,
+ "learning_rate": 0.0006,
+ "loss": 6.005046367645264,
+ "step": 390
+ },
+ {
+ "epoch": 5.4331877729257645,
+ "grad_norm": 0.035277366638183594,
+ "learning_rate": 0.0006,
+ "loss": 6.036296844482422,
+ "step": 391
+ },
+ {
+ "epoch": 5.447161572052401,
+ "grad_norm": 0.034404635429382324,
+ "learning_rate": 0.0006,
+ "loss": 5.994539260864258,
+ "step": 392
+ },
+ {
+ "epoch": 5.461135371179039,
+ "grad_norm": 0.0377206988632679,
+ "learning_rate": 0.0006,
+ "loss": 5.941534519195557,
+ "step": 393
+ },
+ {
+ "epoch": 5.475109170305677,
+ "grad_norm": 0.0389922633767128,
+ "learning_rate": 0.0006,
+ "loss": 5.983644485473633,
+ "step": 394
+ },
+ {
+ "epoch": 5.489082969432315,
+ "grad_norm": 0.04176778718829155,
+ "learning_rate": 0.0006,
+ "loss": 6.030922889709473,
+ "step": 395
+ },
+ {
+ "epoch": 5.503056768558952,
+ "grad_norm": 0.0337153784930706,
+ "learning_rate": 0.0006,
+ "loss": 5.961367130279541,
+ "step": 396
+ },
+ {
+ "epoch": 5.517030567685589,
+ "grad_norm": 0.036116816103458405,
+ "learning_rate": 0.0006,
+ "loss": 5.819280624389648,
+ "step": 397
+ },
+ {
+ "epoch": 5.531004366812227,
+ "grad_norm": 0.030724041163921356,
+ "learning_rate": 0.0006,
+ "loss": 5.906380653381348,
+ "step": 398
+ },
+ {
+ "epoch": 5.544978165938865,
+ "grad_norm": 0.030264202505350113,
+ "learning_rate": 0.0006,
+ "loss": 5.975833415985107,
+ "step": 399
+ },
+ {
+ "epoch": 5.558951965065502,
+ "grad_norm": 0.032096318900585175,
+ "learning_rate": 0.0006,
+ "loss": 6.001348972320557,
+ "step": 400
+ },
+ {
+ "epoch": 5.5729257641921395,
+ "grad_norm": 0.030579356476664543,
+ "learning_rate": 0.0006,
+ "loss": 5.928768634796143,
+ "step": 401
+ },
+ {
+ "epoch": 5.586899563318777,
+ "grad_norm": 0.028241973370313644,
+ "learning_rate": 0.0006,
+ "loss": 5.92582893371582,
+ "step": 402
+ },
+ {
+ "epoch": 5.600873362445415,
+ "grad_norm": 0.023086953908205032,
+ "learning_rate": 0.0006,
+ "loss": 5.892926216125488,
+ "step": 403
+ },
+ {
+ "epoch": 5.614847161572053,
+ "grad_norm": 0.02547992393374443,
+ "learning_rate": 0.0006,
+ "loss": 5.903195381164551,
+ "step": 404
+ },
+ {
+ "epoch": 5.62882096069869,
+ "grad_norm": 0.023089831694960594,
+ "learning_rate": 0.0006,
+ "loss": 5.912033557891846,
+ "step": 405
+ },
+ {
+ "epoch": 5.642794759825327,
+ "grad_norm": 0.025446368381381035,
+ "learning_rate": 0.0006,
+ "loss": 5.926138401031494,
+ "step": 406
+ },
+ {
+ "epoch": 5.656768558951965,
+ "grad_norm": 0.03748747706413269,
+ "learning_rate": 0.0006,
+ "loss": 5.913451194763184,
+ "step": 407
+ },
+ {
+ "epoch": 5.670742358078603,
+ "grad_norm": 0.03493810072541237,
+ "learning_rate": 0.0006,
+ "loss": 5.97633695602417,
+ "step": 408
+ },
+ {
+ "epoch": 5.68471615720524,
+ "grad_norm": 0.023131275549530983,
+ "learning_rate": 0.0006,
+ "loss": 5.899571418762207,
+ "step": 409
+ },
+ {
+ "epoch": 5.698689956331878,
+ "grad_norm": 0.02986014075577259,
+ "learning_rate": 0.0006,
+ "loss": 5.894903182983398,
+ "step": 410
+ },
+ {
+ "epoch": 5.712663755458515,
+ "grad_norm": 0.030171144753694534,
+ "learning_rate": 0.0006,
+ "loss": 5.876595497131348,
+ "step": 411
+ },
+ {
+ "epoch": 5.726637554585153,
+ "grad_norm": 0.029377546161413193,
+ "learning_rate": 0.0006,
+ "loss": 5.911637306213379,
+ "step": 412
+ },
+ {
+ "epoch": 5.74061135371179,
+ "grad_norm": 0.029829490929841995,
+ "learning_rate": 0.0006,
+ "loss": 5.819629669189453,
+ "step": 413
+ },
+ {
+ "epoch": 5.754585152838428,
+ "grad_norm": 0.024139299988746643,
+ "learning_rate": 0.0006,
+ "loss": 5.81728458404541,
+ "step": 414
+ },
+ {
+ "epoch": 5.7685589519650655,
+ "grad_norm": 0.021453432738780975,
+ "learning_rate": 0.0006,
+ "loss": 5.987326145172119,
+ "step": 415
+ },
+ {
+ "epoch": 5.782532751091703,
+ "grad_norm": 0.017942246049642563,
+ "learning_rate": 0.0006,
+ "loss": 5.893008232116699,
+ "step": 416
+ },
+ {
+ "epoch": 5.796506550218341,
+ "grad_norm": 0.019724637269973755,
+ "learning_rate": 0.0006,
+ "loss": 5.912441253662109,
+ "step": 417
+ },
+ {
+ "epoch": 5.810480349344978,
+ "grad_norm": 0.018597912043333054,
+ "learning_rate": 0.0006,
+ "loss": 5.860394477844238,
+ "step": 418
+ },
+ {
+ "epoch": 5.824454148471616,
+ "grad_norm": 0.016232412308454514,
+ "learning_rate": 0.0006,
+ "loss": 5.94991397857666,
+ "step": 419
+ },
+ {
+ "epoch": 5.8384279475982535,
+ "grad_norm": 0.01667204685509205,
+ "learning_rate": 0.0006,
+ "loss": 5.8506364822387695,
+ "step": 420
+ },
+ {
+ "epoch": 5.85240174672489,
+ "grad_norm": 0.014232151210308075,
+ "learning_rate": 0.0006,
+ "loss": 5.928424835205078,
+ "step": 421
+ },
+ {
+ "epoch": 5.866375545851528,
+ "grad_norm": 0.01570476032793522,
+ "learning_rate": 0.0006,
+ "loss": 5.953692436218262,
+ "step": 422
+ },
+ {
+ "epoch": 5.880349344978166,
+ "grad_norm": 0.01586179807782173,
+ "learning_rate": 0.0006,
+ "loss": 5.820014476776123,
+ "step": 423
+ },
+ {
+ "epoch": 5.894323144104804,
+ "grad_norm": 0.014729145914316177,
+ "learning_rate": 0.0006,
+ "loss": 5.860154628753662,
+ "step": 424
+ },
+ {
+ "epoch": 5.908296943231441,
+ "grad_norm": 0.013111459091305733,
+ "learning_rate": 0.0006,
+ "loss": 5.85988187789917,
+ "step": 425
+ },
+ {
+ "epoch": 5.922270742358078,
+ "grad_norm": 0.012937922962009907,
+ "learning_rate": 0.0006,
+ "loss": 5.756265640258789,
+ "step": 426
+ },
+ {
+ "epoch": 5.936244541484716,
+ "grad_norm": 0.01377453189343214,
+ "learning_rate": 0.0006,
+ "loss": 5.8333048820495605,
+ "step": 427
+ },
+ {
+ "epoch": 5.950218340611354,
+ "grad_norm": 0.014054795727133751,
+ "learning_rate": 0.0006,
+ "loss": 5.803333759307861,
+ "step": 428
+ },
+ {
+ "epoch": 5.964192139737992,
+ "grad_norm": 0.01674959622323513,
+ "learning_rate": 0.0006,
+ "loss": 5.8718485832214355,
+ "step": 429
+ },
+ {
+ "epoch": 5.978165938864628,
+ "grad_norm": 0.018733017146587372,
+ "learning_rate": 0.0006,
+ "loss": 5.875979900360107,
+ "step": 430
+ },
+ {
+ "epoch": 5.992139737991266,
+ "grad_norm": 0.02088983915746212,
+ "learning_rate": 0.0006,
+ "loss": 5.763634204864502,
+ "step": 431
+ },
+ {
+ "epoch": 6.0,
+ "grad_norm": 0.026705985888838768,
+ "learning_rate": 0.0006,
+ "loss": 5.6922478675842285,
+ "step": 432
+ },
+ {
+ "epoch": 6.0,
+ "eval_loss": 5.847958087921143,
+ "eval_runtime": 58.5451,
+ "eval_samples_per_second": 41.711,
+ "eval_steps_per_second": 1.315,
+ "step": 432
+ },
+ {
+ "epoch": 6.013973799126638,
+ "grad_norm": 0.03181544691324234,
+ "learning_rate": 0.0006,
+ "loss": 5.786368370056152,
+ "step": 433
+ },
+ {
+ "epoch": 6.0279475982532755,
+ "grad_norm": 0.03238112851977348,
+ "learning_rate": 0.0006,
+ "loss": 5.810310363769531,
+ "step": 434
+ },
+ {
+ "epoch": 6.041921397379912,
+ "grad_norm": 0.03916756808757782,
+ "learning_rate": 0.0006,
+ "loss": 5.7761993408203125,
+ "step": 435
+ },
+ {
+ "epoch": 6.05589519650655,
+ "grad_norm": 0.044009730219841,
+ "learning_rate": 0.0006,
+ "loss": 5.898112773895264,
+ "step": 436
+ },
+ {
+ "epoch": 6.069868995633188,
+ "grad_norm": 0.045235246419906616,
+ "learning_rate": 0.0006,
+ "loss": 5.876371383666992,
+ "step": 437
+ },
+ {
+ "epoch": 6.083842794759826,
+ "grad_norm": 0.050218384712934494,
+ "learning_rate": 0.0006,
+ "loss": 5.805103302001953,
+ "step": 438
+ },
+ {
+ "epoch": 6.097816593886463,
+ "grad_norm": 0.05444490164518356,
+ "learning_rate": 0.0006,
+ "loss": 5.826424598693848,
+ "step": 439
+ },
+ {
+ "epoch": 6.1117903930131,
+ "grad_norm": 0.04890982061624527,
+ "learning_rate": 0.0006,
+ "loss": 5.7735395431518555,
+ "step": 440
+ },
+ {
+ "epoch": 6.125764192139738,
+ "grad_norm": 0.05478639155626297,
+ "learning_rate": 0.0006,
+ "loss": 5.823677062988281,
+ "step": 441
+ },
+ {
+ "epoch": 6.139737991266376,
+ "grad_norm": 0.057562313973903656,
+ "learning_rate": 0.0006,
+ "loss": 5.889334678649902,
+ "step": 442
+ },
+ {
+ "epoch": 6.153711790393013,
+ "grad_norm": 0.06447703391313553,
+ "learning_rate": 0.0006,
+ "loss": 5.84621524810791,
+ "step": 443
+ },
+ {
+ "epoch": 6.1676855895196505,
+ "grad_norm": 0.048861872404813766,
+ "learning_rate": 0.0006,
+ "loss": 5.931595802307129,
+ "step": 444
+ },
+ {
+ "epoch": 6.181659388646288,
+ "grad_norm": 0.05521339178085327,
+ "learning_rate": 0.0006,
+ "loss": 5.8229780197143555,
+ "step": 445
+ },
+ {
+ "epoch": 6.195633187772926,
+ "grad_norm": 0.053666990250349045,
+ "learning_rate": 0.0006,
+ "loss": 5.879191875457764,
+ "step": 446
+ },
+ {
+ "epoch": 6.209606986899563,
+ "grad_norm": 0.0451025515794754,
+ "learning_rate": 0.0006,
+ "loss": 5.801628112792969,
+ "step": 447
+ },
+ {
+ "epoch": 6.223580786026201,
+ "grad_norm": 0.04291655868291855,
+ "learning_rate": 0.0006,
+ "loss": 5.914680480957031,
+ "step": 448
+ },
+ {
+ "epoch": 6.2375545851528384,
+ "grad_norm": 0.035102490335702896,
+ "learning_rate": 0.0006,
+ "loss": 5.843682765960693,
+ "step": 449
+ },
+ {
+ "epoch": 6.251528384279476,
+ "grad_norm": 0.03403995931148529,
+ "learning_rate": 0.0006,
+ "loss": 5.827154159545898,
+ "step": 450
+ },
+ {
+ "epoch": 6.265502183406113,
+ "grad_norm": 0.03444375470280647,
+ "learning_rate": 0.0006,
+ "loss": 5.800136566162109,
+ "step": 451
+ },
+ {
+ "epoch": 6.279475982532751,
+ "grad_norm": 0.03165159001946449,
+ "learning_rate": 0.0006,
+ "loss": 5.906252384185791,
+ "step": 452
+ },
+ {
+ "epoch": 6.293449781659389,
+ "grad_norm": 0.026153093203902245,
+ "learning_rate": 0.0006,
+ "loss": 5.7423200607299805,
+ "step": 453
+ },
+ {
+ "epoch": 6.307423580786026,
+ "grad_norm": 0.024357657879590988,
+ "learning_rate": 0.0006,
+ "loss": 5.701364517211914,
+ "step": 454
+ },
+ {
+ "epoch": 6.321397379912664,
+ "grad_norm": 0.021508049219846725,
+ "learning_rate": 0.0006,
+ "loss": 5.8328094482421875,
+ "step": 455
+ },
+ {
+ "epoch": 6.335371179039301,
+ "grad_norm": 0.017313921824097633,
+ "learning_rate": 0.0006,
+ "loss": 5.792145252227783,
+ "step": 456
+ },
+ {
+ "epoch": 6.349344978165939,
+ "grad_norm": 0.018563397228717804,
+ "learning_rate": 0.0006,
+ "loss": 5.732672214508057,
+ "step": 457
+ },
+ {
+ "epoch": 6.3633187772925766,
+ "grad_norm": 0.016568679362535477,
+ "learning_rate": 0.0006,
+ "loss": 5.81948184967041,
+ "step": 458
+ },
+ {
+ "epoch": 6.377292576419214,
+ "grad_norm": 0.014133770950138569,
+ "learning_rate": 0.0006,
+ "loss": 5.734982490539551,
+ "step": 459
+ },
+ {
+ "epoch": 6.391266375545851,
+ "grad_norm": 0.014114447869360447,
+ "learning_rate": 0.0006,
+ "loss": 5.70073127746582,
+ "step": 460
+ },
+ {
+ "epoch": 6.405240174672489,
+ "grad_norm": 0.012907393276691437,
+ "learning_rate": 0.0006,
+ "loss": 5.758626937866211,
+ "step": 461
+ },
+ {
+ "epoch": 6.419213973799127,
+ "grad_norm": 0.012673444114625454,
+ "learning_rate": 0.0006,
+ "loss": 5.775138854980469,
+ "step": 462
+ },
+ {
+ "epoch": 6.4331877729257645,
+ "grad_norm": 0.01311410591006279,
+ "learning_rate": 0.0006,
+ "loss": 5.70920991897583,
+ "step": 463
+ },
+ {
+ "epoch": 6.447161572052401,
+ "grad_norm": 0.012935544364154339,
+ "learning_rate": 0.0006,
+ "loss": 5.781479835510254,
+ "step": 464
+ },
+ {
+ "epoch": 6.461135371179039,
+ "grad_norm": 0.01439738366752863,
+ "learning_rate": 0.0006,
+ "loss": 5.691315650939941,
+ "step": 465
+ },
+ {
+ "epoch": 6.475109170305677,
+ "grad_norm": 0.01205186452716589,
+ "learning_rate": 0.0006,
+ "loss": 5.662499904632568,
+ "step": 466
+ },
+ {
+ "epoch": 6.489082969432315,
+ "grad_norm": 0.011575652286410332,
+ "learning_rate": 0.0006,
+ "loss": 5.6271772384643555,
+ "step": 467
+ },
+ {
+ "epoch": 6.503056768558952,
+ "grad_norm": 0.011672502383589745,
+ "learning_rate": 0.0006,
+ "loss": 5.761662006378174,
+ "step": 468
+ },
+ {
+ "epoch": 6.517030567685589,
+ "grad_norm": 0.011541751213371754,
+ "learning_rate": 0.0006,
+ "loss": 5.762078285217285,
+ "step": 469
+ },
+ {
+ "epoch": 6.531004366812227,
+ "grad_norm": 0.011396365240216255,
+ "learning_rate": 0.0006,
+ "loss": 5.67873477935791,
+ "step": 470
+ },
+ {
+ "epoch": 6.544978165938865,
+ "grad_norm": 0.010688499547541142,
+ "learning_rate": 0.0006,
+ "loss": 5.717051982879639,
+ "step": 471
+ },
+ {
+ "epoch": 6.558951965065502,
+ "grad_norm": 0.012224317528307438,
+ "learning_rate": 0.0006,
+ "loss": 5.707948684692383,
+ "step": 472
+ },
+ {
+ "epoch": 6.5729257641921395,
+ "grad_norm": 0.011856825090944767,
+ "learning_rate": 0.0006,
+ "loss": 5.70878791809082,
+ "step": 473
+ },
+ {
+ "epoch": 6.586899563318777,
+ "grad_norm": 0.01199064590036869,
+ "learning_rate": 0.0006,
+ "loss": 5.708697319030762,
+ "step": 474
+ },
+ {
+ "epoch": 6.600873362445415,
+ "grad_norm": 0.01219446212053299,
+ "learning_rate": 0.0006,
+ "loss": 5.6061577796936035,
+ "step": 475
+ },
+ {
+ "epoch": 6.614847161572053,
+ "grad_norm": 0.013397484086453915,
+ "learning_rate": 0.0006,
+ "loss": 5.624789714813232,
+ "step": 476
+ },
+ {
+ "epoch": 6.62882096069869,
+ "grad_norm": 0.01483136136084795,
+ "learning_rate": 0.0006,
+ "loss": 5.721141338348389,
+ "step": 477
+ },
+ {
+ "epoch": 6.642794759825327,
+ "grad_norm": 0.0194488987326622,
+ "learning_rate": 0.0006,
+ "loss": 5.604279518127441,
+ "step": 478
+ },
+ {
+ "epoch": 6.656768558951965,
+ "grad_norm": 0.02079896256327629,
+ "learning_rate": 0.0006,
+ "loss": 5.586322784423828,
+ "step": 479
+ },
+ {
+ "epoch": 6.670742358078603,
+ "grad_norm": 0.018996229395270348,
+ "learning_rate": 0.0006,
+ "loss": 5.5987772941589355,
+ "step": 480
+ },
+ {
+ "epoch": 6.68471615720524,
+ "grad_norm": 0.015779603272676468,
+ "learning_rate": 0.0006,
+ "loss": 5.588602066040039,
+ "step": 481
+ },
+ {
+ "epoch": 6.698689956331878,
+ "grad_norm": 0.015322973020374775,
+ "learning_rate": 0.0006,
+ "loss": 5.686467170715332,
+ "step": 482
+ },
+ {
+ "epoch": 6.712663755458515,
+ "grad_norm": 0.015282213687896729,
+ "learning_rate": 0.0006,
+ "loss": 5.664676666259766,
+ "step": 483
+ },
+ {
+ "epoch": 6.726637554585153,
+ "grad_norm": 0.016589272767305374,
+ "learning_rate": 0.0006,
+ "loss": 5.666739463806152,
+ "step": 484
+ },
+ {
+ "epoch": 6.74061135371179,
+ "grad_norm": 0.01811421848833561,
+ "learning_rate": 0.0006,
+ "loss": 5.647593021392822,
+ "step": 485
+ },
+ {
+ "epoch": 6.754585152838428,
+ "grad_norm": 0.019652029499411583,
+ "learning_rate": 0.0006,
+ "loss": 5.628726959228516,
+ "step": 486
+ },
+ {
+ "epoch": 6.7685589519650655,
+ "grad_norm": 0.02118665538728237,
+ "learning_rate": 0.0006,
+ "loss": 5.6580705642700195,
+ "step": 487
+ },
+ {
+ "epoch": 6.782532751091703,
+ "grad_norm": 0.02237832546234131,
+ "learning_rate": 0.0006,
+ "loss": 5.635931968688965,
+ "step": 488
+ },
+ {
+ "epoch": 6.796506550218341,
+ "grad_norm": 0.023897631093859673,
+ "learning_rate": 0.0006,
+ "loss": 5.517949104309082,
+ "step": 489
+ },
+ {
+ "epoch": 6.810480349344978,
+ "grad_norm": 0.02442227490246296,
+ "learning_rate": 0.0006,
+ "loss": 5.600021839141846,
+ "step": 490
+ },
+ {
+ "epoch": 6.824454148471616,
+ "grad_norm": 0.024675287306308746,
+ "learning_rate": 0.0006,
+ "loss": 5.6639909744262695,
+ "step": 491
+ },
+ {
+ "epoch": 6.8384279475982535,
+ "grad_norm": 0.030372392386198044,
+ "learning_rate": 0.0006,
+ "loss": 5.582888603210449,
+ "step": 492
+ },
+ {
+ "epoch": 6.85240174672489,
+ "grad_norm": 0.03237690031528473,
+ "learning_rate": 0.0006,
+ "loss": 5.573896884918213,
+ "step": 493
+ },
+ {
+ "epoch": 6.866375545851528,
+ "grad_norm": 0.03511589393019676,
+ "learning_rate": 0.0006,
+ "loss": 5.565878868103027,
+ "step": 494
+ },
+ {
+ "epoch": 6.880349344978166,
+ "grad_norm": 0.03883938118815422,
+ "learning_rate": 0.0006,
+ "loss": 5.58540678024292,
+ "step": 495
+ },
+ {
+ "epoch": 6.894323144104804,
+ "grad_norm": 0.04175502806901932,
+ "learning_rate": 0.0006,
+ "loss": 5.6095194816589355,
+ "step": 496
+ },
+ {
+ "epoch": 6.908296943231441,
+ "grad_norm": 0.04013441503047943,
+ "learning_rate": 0.0006,
+ "loss": 5.654401779174805,
+ "step": 497
+ },
+ {
+ "epoch": 6.922270742358078,
+ "grad_norm": 0.04618770629167557,
+ "learning_rate": 0.0006,
+ "loss": 5.6483473777771,
+ "step": 498
+ },
+ {
+ "epoch": 6.936244541484716,
+ "grad_norm": 0.04507308453321457,
+ "learning_rate": 0.0006,
+ "loss": 5.5856218338012695,
+ "step": 499
+ },
+ {
+ "epoch": 6.950218340611354,
+ "grad_norm": 0.03151383996009827,
+ "learning_rate": 0.0006,
+ "loss": 5.594086647033691,
+ "step": 500
+ },
+ {
+ "epoch": 6.964192139737992,
+ "grad_norm": 0.026883797720074654,
+ "learning_rate": 0.0006,
+ "loss": 5.607676029205322,
+ "step": 501
+ },
+ {
+ "epoch": 6.978165938864628,
+ "grad_norm": 0.02981836162507534,
+ "learning_rate": 0.0006,
+ "loss": 5.563666820526123,
+ "step": 502
+ },
+ {
+ "epoch": 6.992139737991266,
+ "grad_norm": 0.03233477845788002,
+ "learning_rate": 0.0006,
+ "loss": 5.667543411254883,
+ "step": 503
+ },
+ {
+ "epoch": 7.0,
+ "grad_norm": 0.029927456751465797,
+ "learning_rate": 0.0006,
+ "loss": 5.733482837677002,
+ "step": 504
+ },
+ {
+ "epoch": 7.0,
+ "eval_loss": 5.679076671600342,
+ "eval_runtime": 59.5965,
+ "eval_samples_per_second": 40.976,
+ "eval_steps_per_second": 1.292,
+ "step": 504
+ },
+ {
+ "epoch": 7.013973799126638,
+ "grad_norm": 0.027081597596406937,
+ "learning_rate": 0.0006,
+ "loss": 5.634487152099609,
+ "step": 505
+ },
+ {
+ "epoch": 7.0279475982532755,
+ "grad_norm": 0.031959421932697296,
+ "learning_rate": 0.0006,
+ "loss": 5.643294811248779,
+ "step": 506
+ },
+ {
+ "epoch": 7.041921397379912,
+ "grad_norm": 0.02802063524723053,
+ "learning_rate": 0.0006,
+ "loss": 5.573239326477051,
+ "step": 507
+ },
+ {
+ "epoch": 7.05589519650655,
+ "grad_norm": 0.032579705119132996,
+ "learning_rate": 0.0006,
+ "loss": 5.627150058746338,
+ "step": 508
+ },
+ {
+ "epoch": 7.069868995633188,
+ "grad_norm": 0.02503928542137146,
+ "learning_rate": 0.0006,
+ "loss": 5.6045308113098145,
+ "step": 509
+ },
+ {
+ "epoch": 7.083842794759826,
+ "grad_norm": 0.02421317622065544,
+ "learning_rate": 0.0006,
+ "loss": 5.522153377532959,
+ "step": 510
+ },
+ {
+ "epoch": 7.097816593886463,
+ "grad_norm": 0.021983426064252853,
+ "learning_rate": 0.0006,
+ "loss": 5.632939338684082,
+ "step": 511
+ },
+ {
+ "epoch": 7.1117903930131,
+ "grad_norm": 0.021933183073997498,
+ "learning_rate": 0.0006,
+ "loss": 5.592185974121094,
+ "step": 512
+ },
+ {
+ "epoch": 7.125764192139738,
+ "grad_norm": 0.02287031151354313,
+ "learning_rate": 0.0006,
+ "loss": 5.613700866699219,
+ "step": 513
+ },
+ {
+ "epoch": 7.139737991266376,
+ "grad_norm": 0.02281602844595909,
+ "learning_rate": 0.0006,
+ "loss": 5.551383018493652,
+ "step": 514
+ },
+ {
+ "epoch": 7.153711790393013,
+ "grad_norm": 0.021050360053777695,
+ "learning_rate": 0.0006,
+ "loss": 5.55275821685791,
+ "step": 515
+ },
+ {
+ "epoch": 7.1676855895196505,
+ "grad_norm": 0.018299926072359085,
+ "learning_rate": 0.0006,
+ "loss": 5.528225421905518,
+ "step": 516
+ },
+ {
+ "epoch": 7.181659388646288,
+ "grad_norm": 0.02024853602051735,
+ "learning_rate": 0.0006,
+ "loss": 5.522453784942627,
+ "step": 517
+ },
+ {
+ "epoch": 7.195633187772926,
+ "grad_norm": 0.02135239914059639,
+ "learning_rate": 0.0006,
+ "loss": 5.527522563934326,
+ "step": 518
+ },
+ {
+ "epoch": 7.209606986899563,
+ "grad_norm": 0.023394184187054634,
+ "learning_rate": 0.0006,
+ "loss": 5.4535017013549805,
+ "step": 519
+ },
+ {
+ "epoch": 7.223580786026201,
+ "grad_norm": 0.025952080264687538,
+ "learning_rate": 0.0006,
+ "loss": 5.5374674797058105,
+ "step": 520
+ },
+ {
+ "epoch": 7.2375545851528384,
+ "grad_norm": 0.022287718951702118,
+ "learning_rate": 0.0006,
+ "loss": 5.493753433227539,
+ "step": 521
+ },
+ {
+ "epoch": 7.251528384279476,
+ "grad_norm": 0.02018447406589985,
+ "learning_rate": 0.0006,
+ "loss": 5.510575294494629,
+ "step": 522
+ },
+ {
+ "epoch": 7.265502183406113,
+ "grad_norm": 0.02050507813692093,
+ "learning_rate": 0.0006,
+ "loss": 5.466026782989502,
+ "step": 523
+ },
+ {
+ "epoch": 7.279475982532751,
+ "grad_norm": 0.023688802495598793,
+ "learning_rate": 0.0006,
+ "loss": 5.485815525054932,
+ "step": 524
+ },
+ {
+ "epoch": 7.293449781659389,
+ "grad_norm": 0.021078843623399734,
+ "learning_rate": 0.0006,
+ "loss": 5.415581703186035,
+ "step": 525
+ },
+ {
+ "epoch": 7.307423580786026,
+ "grad_norm": 0.018266011029481888,
+ "learning_rate": 0.0006,
+ "loss": 5.402815818786621,
+ "step": 526
+ },
+ {
+ "epoch": 7.321397379912664,
+ "grad_norm": 0.019693493843078613,
+ "learning_rate": 0.0006,
+ "loss": 5.504674911499023,
+ "step": 527
+ },
+ {
+ "epoch": 7.335371179039301,
+ "grad_norm": 0.02031373605132103,
+ "learning_rate": 0.0006,
+ "loss": 5.503837585449219,
+ "step": 528
+ },
+ {
+ "epoch": 7.349344978165939,
+ "grad_norm": 0.017636626958847046,
+ "learning_rate": 0.0006,
+ "loss": 5.530580997467041,
+ "step": 529
+ },
+ {
+ "epoch": 7.3633187772925766,
+ "grad_norm": 0.01787244901061058,
+ "learning_rate": 0.0006,
+ "loss": 5.450218677520752,
+ "step": 530
+ },
+ {
+ "epoch": 7.377292576419214,
+ "grad_norm": 0.0170669574290514,
+ "learning_rate": 0.0006,
+ "loss": 5.481570243835449,
+ "step": 531
+ },
+ {
+ "epoch": 7.391266375545851,
+ "grad_norm": 0.01802165061235428,
+ "learning_rate": 0.0006,
+ "loss": 5.386394500732422,
+ "step": 532
+ },
+ {
+ "epoch": 7.405240174672489,
+ "grad_norm": 0.022949472069740295,
+ "learning_rate": 0.0006,
+ "loss": 5.424929618835449,
+ "step": 533
+ },
+ {
+ "epoch": 7.419213973799127,
+ "grad_norm": 0.034213390201330185,
+ "learning_rate": 0.0006,
+ "loss": 5.346663475036621,
+ "step": 534
+ },
+ {
+ "epoch": 7.4331877729257645,
+ "grad_norm": 0.03847593814134598,
+ "learning_rate": 0.0006,
+ "loss": 5.418482780456543,
+ "step": 535
+ },
+ {
+ "epoch": 7.447161572052401,
+ "grad_norm": 0.029393676668405533,
+ "learning_rate": 0.0006,
+ "loss": 5.456090927124023,
+ "step": 536
+ },
+ {
+ "epoch": 7.461135371179039,
+ "grad_norm": 0.03388667106628418,
+ "learning_rate": 0.0006,
+ "loss": 5.572983741760254,
+ "step": 537
+ },
+ {
+ "epoch": 7.475109170305677,
+ "grad_norm": 0.039690613746643066,
+ "learning_rate": 0.0006,
+ "loss": 5.414067268371582,
+ "step": 538
+ },
+ {
+ "epoch": 7.489082969432315,
+ "grad_norm": 0.03630776330828667,
+ "learning_rate": 0.0006,
+ "loss": 5.523739814758301,
+ "step": 539
+ },
+ {
+ "epoch": 7.503056768558952,
+ "grad_norm": 0.03356870263814926,
+ "learning_rate": 0.0006,
+ "loss": 5.444767951965332,
+ "step": 540
+ },
+ {
+ "epoch": 7.517030567685589,
+ "grad_norm": 0.030812304466962814,
+ "learning_rate": 0.0006,
+ "loss": 5.461244106292725,
+ "step": 541
+ },
+ {
+ "epoch": 7.531004366812227,
+ "grad_norm": 0.03319217637181282,
+ "learning_rate": 0.0006,
+ "loss": 5.502161026000977,
+ "step": 542
+ },
+ {
+ "epoch": 7.544978165938865,
+ "grad_norm": 0.032997481524944305,
+ "learning_rate": 0.0006,
+ "loss": 5.4162278175354,
+ "step": 543
+ },
+ {
+ "epoch": 7.558951965065502,
+ "grad_norm": 0.03364962339401245,
+ "learning_rate": 0.0006,
+ "loss": 5.415736198425293,
+ "step": 544
+ },
+ {
+ "epoch": 7.5729257641921395,
+ "grad_norm": 0.0344221405684948,
+ "learning_rate": 0.0006,
+ "loss": 5.541967868804932,
+ "step": 545
+ },
+ {
+ "epoch": 7.586899563318777,
+ "grad_norm": 0.029609503224492073,
+ "learning_rate": 0.0006,
+ "loss": 5.440614223480225,
+ "step": 546
+ },
+ {
+ "epoch": 7.600873362445415,
+ "grad_norm": 0.029055926948785782,
+ "learning_rate": 0.0006,
+ "loss": 5.462865352630615,
+ "step": 547
+ },
+ {
+ "epoch": 7.614847161572053,
+ "grad_norm": 0.02658848837018013,
+ "learning_rate": 0.0006,
+ "loss": 5.444467544555664,
+ "step": 548
+ },
+ {
+ "epoch": 7.62882096069869,
+ "grad_norm": 0.026276404038071632,
+ "learning_rate": 0.0006,
+ "loss": 5.462278842926025,
+ "step": 549
+ },
+ {
+ "epoch": 7.642794759825327,
+ "grad_norm": 0.0286889486014843,
+ "learning_rate": 0.0006,
+ "loss": 5.4849534034729,
+ "step": 550
+ },
+ {
+ "epoch": 7.656768558951965,
+ "grad_norm": 0.026849817484617233,
+ "learning_rate": 0.0006,
+ "loss": 5.431473731994629,
+ "step": 551
+ },
+ {
+ "epoch": 7.670742358078603,
+ "grad_norm": 0.02312396466732025,
+ "learning_rate": 0.0006,
+ "loss": 5.45462703704834,
+ "step": 552
+ },
+ {
+ "epoch": 7.68471615720524,
+ "grad_norm": 0.026197485625743866,
+ "learning_rate": 0.0006,
+ "loss": 5.452552318572998,
+ "step": 553
+ },
+ {
+ "epoch": 7.698689956331878,
+ "grad_norm": 0.02747255191206932,
+ "learning_rate": 0.0006,
+ "loss": 5.447169780731201,
+ "step": 554
+ },
+ {
+ "epoch": 7.712663755458515,
+ "grad_norm": 0.028123432770371437,
+ "learning_rate": 0.0006,
+ "loss": 5.394400596618652,
+ "step": 555
+ },
+ {
+ "epoch": 7.726637554585153,
+ "grad_norm": 0.02599870041012764,
+ "learning_rate": 0.0006,
+ "loss": 5.398341178894043,
+ "step": 556
+ },
+ {
+ "epoch": 7.74061135371179,
+ "grad_norm": 0.022171657532453537,
+ "learning_rate": 0.0006,
+ "loss": 5.368820667266846,
+ "step": 557
+ },
+ {
+ "epoch": 7.754585152838428,
+ "grad_norm": 0.022309480234980583,
+ "learning_rate": 0.0006,
+ "loss": 5.416772365570068,
+ "step": 558
+ },
+ {
+ "epoch": 7.7685589519650655,
+ "grad_norm": 0.024074165150523186,
+ "learning_rate": 0.0006,
+ "loss": 5.359002113342285,
+ "step": 559
+ },
+ {
+ "epoch": 7.782532751091703,
+ "grad_norm": 0.02636653557419777,
+ "learning_rate": 0.0006,
+ "loss": 5.484368324279785,
+ "step": 560
+ },
+ {
+ "epoch": 7.796506550218341,
+ "grad_norm": 0.02203752100467682,
+ "learning_rate": 0.0006,
+ "loss": 5.375498294830322,
+ "step": 561
+ },
+ {
+ "epoch": 7.810480349344978,
+ "grad_norm": 0.019924819469451904,
+ "learning_rate": 0.0006,
+ "loss": 5.334723472595215,
+ "step": 562
+ },
+ {
+ "epoch": 7.824454148471616,
+ "grad_norm": 0.018755966797471046,
+ "learning_rate": 0.0006,
+ "loss": 5.348781108856201,
+ "step": 563
+ },
+ {
+ "epoch": 7.8384279475982535,
+ "grad_norm": 0.016747845336794853,
+ "learning_rate": 0.0006,
+ "loss": 5.433079719543457,
+ "step": 564
+ },
+ {
+ "epoch": 7.85240174672489,
+ "grad_norm": 0.0170395877212286,
+ "learning_rate": 0.0006,
+ "loss": 5.346794128417969,
+ "step": 565
+ },
+ {
+ "epoch": 7.866375545851528,
+ "grad_norm": 0.016074176877737045,
+ "learning_rate": 0.0006,
+ "loss": 5.3277740478515625,
+ "step": 566
+ },
+ {
+ "epoch": 7.880349344978166,
+ "grad_norm": 0.014119806699454784,
+ "learning_rate": 0.0006,
+ "loss": 5.483551979064941,
+ "step": 567
+ },
+ {
+ "epoch": 7.894323144104804,
+ "grad_norm": 0.015270394273102283,
+ "learning_rate": 0.0006,
+ "loss": 5.378519535064697,
+ "step": 568
+ },
+ {
+ "epoch": 7.908296943231441,
+ "grad_norm": 0.015677539631724358,
+ "learning_rate": 0.0006,
+ "loss": 5.290210247039795,
+ "step": 569
+ },
+ {
+ "epoch": 7.922270742358078,
+ "grad_norm": 0.015930423513054848,
+ "learning_rate": 0.0006,
+ "loss": 5.427360534667969,
+ "step": 570
+ },
+ {
+ "epoch": 7.936244541484716,
+ "grad_norm": 0.016641564667224884,
+ "learning_rate": 0.0006,
+ "loss": 5.301599979400635,
+ "step": 571
+ },
+ {
+ "epoch": 7.950218340611354,
+ "grad_norm": 0.018293552100658417,
+ "learning_rate": 0.0006,
+ "loss": 5.277889251708984,
+ "step": 572
+ },
+ {
+ "epoch": 7.964192139737992,
+ "grad_norm": 0.017618799582123756,
+ "learning_rate": 0.0006,
+ "loss": 5.297194480895996,
+ "step": 573
+ },
+ {
+ "epoch": 7.978165938864628,
+ "grad_norm": 0.015034242533147335,
+ "learning_rate": 0.0006,
+ "loss": 5.42428731918335,
+ "step": 574
+ },
+ {
+ "epoch": 7.992139737991266,
+ "grad_norm": 0.01617511734366417,
+ "learning_rate": 0.0006,
+ "loss": 5.420044898986816,
+ "step": 575
+ },
+ {
+ "epoch": 8.0,
+ "grad_norm": 0.01837257295846939,
+ "learning_rate": 0.0006,
+ "loss": 5.240387916564941,
+ "step": 576
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 28800,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 400,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": true,
+ "should_log": false,
+ "should_save": false,
+ "should_training_stop": false
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 2.4556906890657792e+17,
+ "train_batch_size": 8,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-576/training_args.bin b/runs/l2r50-i2-fulle-lm/checkpoint-576/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..cad232831c222cc7f8b938b8bbe42f61e7d89c3a
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-576/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:f7c40e3bfa2f5ac1450ca1243e3f2a895d44a06ed5995e6aefd97f78d229d8c8
+size 4792
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-5760/chat_template.jinja b/runs/l2r50-i2-fulle-lm/checkpoint-5760/chat_template.jinja
new file mode 100644
index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-5760/chat_template.jinja
@@ -0,0 +1,85 @@
+{%- if tools %}
+ {{- '<|im_start|>system\n' }}
+ {%- if messages[0].role == 'system' %}
+ {{- messages[0].content + '\n\n' }}
+ {%- endif %}
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }}
+ {%- for tool in tools %}
+ {{- "\n" }}
+ {{- tool | tojson }}
+ {%- endfor %}
+ {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }}
+{%- else %}
+ {%- if messages[0].role == 'system' %}
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
+ {%- endif %}
+{%- endif %}
+{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
+{%- for message in messages[::-1] %}
+ {%- set index = (messages|length - 1) - loop.index0 %}
+ {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %}
+ {%- set ns.multi_step_tool = false %}
+ {%- set ns.last_query_index = index %}
+ {%- endif %}
+{%- endfor %}
+{%- for message in messages %}
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
+ {%- elif message.role == "assistant" %}
+ {%- set content = message.content %}
+ {%- set reasoning_content = '' %}
+ {%- if message.reasoning_content is defined and message.reasoning_content is not none %}
+ {%- set reasoning_content = message.reasoning_content %}
+ {%- else %}
+ {%- if '' in message.content %}
+ {%- set content = message.content.split('')[-1].lstrip('\n') %}
+ {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %}
+ {%- endif %}
+ {%- endif %}
+ {%- if loop.index0 > ns.last_query_index %}
+ {%- if loop.last or (not loop.last and reasoning_content) %}
+ {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- if message.tool_calls %}
+ {%- for tool_call in message.tool_calls %}
+ {%- if (loop.first and content) or (not loop.first) %}
+ {{- '\n' }}
+ {%- endif %}
+ {%- if tool_call.function %}
+ {%- set tool_call = tool_call.function %}
+ {%- endif %}
+ {{- '\n{"name": "' }}
+ {{- tool_call.name }}
+ {{- '", "arguments": ' }}
+ {%- if tool_call.arguments is string %}
+ {{- tool_call.arguments }}
+ {%- else %}
+ {{- tool_call.arguments | tojson }}
+ {%- endif %}
+ {{- '}\n' }}
+ {%- endfor %}
+ {%- endif %}
+ {{- '<|im_end|>\n' }}
+ {%- elif message.role == "tool" %}
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
+ {{- '<|im_start|>user' }}
+ {%- endif %}
+ {{- '\n\n' }}
+ {{- message.content }}
+ {{- '\n' }}
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
+ {{- '<|im_end|>\n' }}
+ {%- endif %}
+ {%- endif %}
+{%- endfor %}
+{%- if add_generation_prompt %}
+ {{- '<|im_start|>assistant\n' }}
+ {%- if enable_thinking is defined and enable_thinking is false %}
+ {{- '\n\n\n\n' }}
+ {%- endif %}
+{%- endif %}
\ No newline at end of file
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-5760/config.json b/runs/l2r50-i2-fulle-lm/checkpoint-5760/config.json
new file mode 100644
index 0000000000000000000000000000000000000000..22f250953c6ce8fc91ee4411856dd73041a415e2
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-5760/config.json
@@ -0,0 +1,32 @@
+{
+ "architectures": [
+ "LlamaForCausalLM"
+ ],
+ "attention_bias": false,
+ "attention_dropout": 0.0,
+ "bos_token_id": null,
+ "dtype": "float32",
+ "eos_token_id": 151645,
+ "head_dim": 128,
+ "hidden_act": "silu",
+ "hidden_size": 512,
+ "initializer_range": 0.02,
+ "intermediate_size": 1536,
+ "max_position_embeddings": 2048,
+ "mlp_bias": false,
+ "model_type": "llama",
+ "num_attention_heads": 4,
+ "num_hidden_layers": 20,
+ "num_key_value_heads": 4,
+ "pad_token_id": 151645,
+ "pretraining_tp": 1,
+ "rms_norm_eps": 1e-06,
+ "rope_parameters": {
+ "rope_theta": 10000.0,
+ "rope_type": "default"
+ },
+ "tie_word_embeddings": true,
+ "transformers_version": "5.5.0",
+ "use_cache": false,
+ "vocab_size": 151673
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-5760/generation_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-5760/generation_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-5760/generation_config.json
@@ -0,0 +1,11 @@
+{
+ "_from_model_config": true,
+ "eos_token_id": [
+ 151645
+ ],
+ "output_attentions": false,
+ "output_hidden_states": false,
+ "pad_token_id": 151645,
+ "transformers_version": "5.5.0",
+ "use_cache": true
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-5760/model.safetensors b/runs/l2r50-i2-fulle-lm/checkpoint-5760/model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..0c32862e1ebfabd231188cab38c0f5fa4ff1aeda
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-5760/model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:44c96e9d16f833e2f029812687786b0c582ce8043d4854729e759d91e4ae5379
+size 583360328
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-5760/optimizer.pt b/runs/l2r50-i2-fulle-lm/checkpoint-5760/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..eca2cb0191673025aff2d7629b3647f70531dc3f
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-5760/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:8927b1690f33447659d31c9807fb5f51447a30aba0302a83a57d0e44d2673dc8
+size 1166833530
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-5760/rng_state_0.pth b/runs/l2r50-i2-fulle-lm/checkpoint-5760/rng_state_0.pth
new file mode 100644
index 0000000000000000000000000000000000000000..a285714799f193b1956b942912d36ac13f0bf7ce
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-5760/rng_state_0.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:0d4b2db3d31ab56483f8ea333d381b8d330a869c9278fe352f6cbfd94741b766
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-5760/rng_state_1.pth b/runs/l2r50-i2-fulle-lm/checkpoint-5760/rng_state_1.pth
new file mode 100644
index 0000000000000000000000000000000000000000..2e2e42c1aab8de141eefa799062d967e975b6218
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-5760/rng_state_1.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:1322bae719fba49418ac08f03776405c298a53bdd11d4473f4a6a187dc25005e
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-5760/rng_state_2.pth b/runs/l2r50-i2-fulle-lm/checkpoint-5760/rng_state_2.pth
new file mode 100644
index 0000000000000000000000000000000000000000..f2cefc91b060c9f1a03313ee35052d85ae1fb721
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-5760/rng_state_2.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:4f9d34a9ab58ec8211a1996d0791f9c8c7a6150afc12696230c213c6c98fd63b
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-5760/rng_state_3.pth b/runs/l2r50-i2-fulle-lm/checkpoint-5760/rng_state_3.pth
new file mode 100644
index 0000000000000000000000000000000000000000..8ad2dc824fec084bf990db595b420b6fcc0aff87
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-5760/rng_state_3.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:73b5916c48044cca60d11397b7319bdd569e58f962208e54da1468f10cf9a5f4
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-5760/scheduler.pt b/runs/l2r50-i2-fulle-lm/checkpoint-5760/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..146fac8459db1a89ef1d0b02179a40f96280dd47
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-5760/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:a202cad58a75ceda9326b04df06bcdb148f3de25b706c12232648bdbcae46e67
+size 1064
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-5760/tokenizer.json b/runs/l2r50-i2-fulle-lm/checkpoint-5760/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..53786016f070e11184813f3c83101dc5e93e392e
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-5760/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:c3c63fda1646a555448e8f03662542e04b11d103721319e4d9b267d7da03ceb8
+size 11424463
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-5760/tokenizer_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-5760/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..98e2fb9b2cc54785dc9d871b8e503f8b8f2defe7
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-5760/tokenizer_config.json
@@ -0,0 +1,20 @@
+{
+ "add_prefix_space": false,
+ "backend": "tokenizers",
+ "bos_token": null,
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "<|im_end|>",
+ "errors": "replace",
+ "extra_special_tokens": [
+ "<|l2r_pred|>",
+ "<|r2l_pred|>",
+ "<|next_1_pred|>",
+ "<|next_2_pred|>"
+ ],
+ "is_local": false,
+ "model_max_length": 131072,
+ "pad_token": "<|im_end|>",
+ "split_special_tokens": false,
+ "tokenizer_class": "Qwen2Tokenizer",
+ "unk_token": null
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-5760/trainer_state.json b/runs/l2r50-i2-fulle-lm/checkpoint-5760/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..493d844116fb14245dcf82d371caeffd003511cd
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-5760/trainer_state.json
@@ -0,0 +1,40986 @@
+{
+ "best_global_step": null,
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 80.0,
+ "eval_steps": 500,
+ "global_step": 5760,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.013973799126637555,
+ "grad_norm": 0.1739506870508194,
+ "learning_rate": 0.0,
+ "loss": 11.99299430847168,
+ "step": 1
+ },
+ {
+ "epoch": 0.02794759825327511,
+ "grad_norm": 0.17216597497463226,
+ "learning_rate": 5.999999999999999e-06,
+ "loss": 11.988704681396484,
+ "step": 2
+ },
+ {
+ "epoch": 0.04192139737991266,
+ "grad_norm": 0.17459715902805328,
+ "learning_rate": 1.1999999999999999e-05,
+ "loss": 11.945581436157227,
+ "step": 3
+ },
+ {
+ "epoch": 0.05589519650655022,
+ "grad_norm": 0.17577414214611053,
+ "learning_rate": 1.7999999999999997e-05,
+ "loss": 11.859772682189941,
+ "step": 4
+ },
+ {
+ "epoch": 0.06986899563318777,
+ "grad_norm": 0.16163651645183563,
+ "learning_rate": 2.3999999999999997e-05,
+ "loss": 11.744415283203125,
+ "step": 5
+ },
+ {
+ "epoch": 0.08384279475982533,
+ "grad_norm": 0.14409932494163513,
+ "learning_rate": 2.9999999999999997e-05,
+ "loss": 11.618099212646484,
+ "step": 6
+ },
+ {
+ "epoch": 0.09781659388646288,
+ "grad_norm": 0.13079848885536194,
+ "learning_rate": 3.5999999999999994e-05,
+ "loss": 11.505935668945312,
+ "step": 7
+ },
+ {
+ "epoch": 0.11179039301310044,
+ "grad_norm": 0.11964613944292068,
+ "learning_rate": 4.2e-05,
+ "loss": 11.416532516479492,
+ "step": 8
+ },
+ {
+ "epoch": 0.125764192139738,
+ "grad_norm": 0.11345142871141434,
+ "learning_rate": 4.7999999999999994e-05,
+ "loss": 11.33206558227539,
+ "step": 9
+ },
+ {
+ "epoch": 0.13973799126637554,
+ "grad_norm": 0.11033952981233597,
+ "learning_rate": 5.399999999999999e-05,
+ "loss": 11.259052276611328,
+ "step": 10
+ },
+ {
+ "epoch": 0.1537117903930131,
+ "grad_norm": 0.10776489973068237,
+ "learning_rate": 5.9999999999999995e-05,
+ "loss": 11.203384399414062,
+ "step": 11
+ },
+ {
+ "epoch": 0.16768558951965065,
+ "grad_norm": 0.10721946507692337,
+ "learning_rate": 6.599999999999999e-05,
+ "loss": 11.153295516967773,
+ "step": 12
+ },
+ {
+ "epoch": 0.18165938864628822,
+ "grad_norm": 0.10673430562019348,
+ "learning_rate": 7.199999999999999e-05,
+ "loss": 11.110200881958008,
+ "step": 13
+ },
+ {
+ "epoch": 0.19563318777292577,
+ "grad_norm": 0.10630254447460175,
+ "learning_rate": 7.8e-05,
+ "loss": 11.067477226257324,
+ "step": 14
+ },
+ {
+ "epoch": 0.2096069868995633,
+ "grad_norm": 0.10631615668535233,
+ "learning_rate": 8.4e-05,
+ "loss": 11.022941589355469,
+ "step": 15
+ },
+ {
+ "epoch": 0.22358078602620088,
+ "grad_norm": 0.10609443485736847,
+ "learning_rate": 8.999999999999999e-05,
+ "loss": 10.970746040344238,
+ "step": 16
+ },
+ {
+ "epoch": 0.23755458515283842,
+ "grad_norm": 0.10648232698440552,
+ "learning_rate": 9.599999999999999e-05,
+ "loss": 10.91267204284668,
+ "step": 17
+ },
+ {
+ "epoch": 0.251528384279476,
+ "grad_norm": 0.10584589838981628,
+ "learning_rate": 0.000102,
+ "loss": 10.854547500610352,
+ "step": 18
+ },
+ {
+ "epoch": 0.26550218340611353,
+ "grad_norm": 0.10591720044612885,
+ "learning_rate": 0.00010799999999999998,
+ "loss": 10.788917541503906,
+ "step": 19
+ },
+ {
+ "epoch": 0.2794759825327511,
+ "grad_norm": 0.10611672699451447,
+ "learning_rate": 0.00011399999999999999,
+ "loss": 10.717556953430176,
+ "step": 20
+ },
+ {
+ "epoch": 0.2934497816593886,
+ "grad_norm": 0.10455359518527985,
+ "learning_rate": 0.00011999999999999999,
+ "loss": 10.650186538696289,
+ "step": 21
+ },
+ {
+ "epoch": 0.3074235807860262,
+ "grad_norm": 0.1047525629401207,
+ "learning_rate": 0.00012599999999999997,
+ "loss": 10.56808090209961,
+ "step": 22
+ },
+ {
+ "epoch": 0.32139737991266376,
+ "grad_norm": 0.10511605441570282,
+ "learning_rate": 0.00013199999999999998,
+ "loss": 10.484028816223145,
+ "step": 23
+ },
+ {
+ "epoch": 0.3353711790393013,
+ "grad_norm": 0.10467177629470825,
+ "learning_rate": 0.000138,
+ "loss": 10.399947166442871,
+ "step": 24
+ },
+ {
+ "epoch": 0.34934497816593885,
+ "grad_norm": 0.10432136058807373,
+ "learning_rate": 0.00014399999999999998,
+ "loss": 10.312714576721191,
+ "step": 25
+ },
+ {
+ "epoch": 0.36331877729257644,
+ "grad_norm": 0.10527262091636658,
+ "learning_rate": 0.00015,
+ "loss": 10.212261199951172,
+ "step": 26
+ },
+ {
+ "epoch": 0.377292576419214,
+ "grad_norm": 0.103610560297966,
+ "learning_rate": 0.000156,
+ "loss": 10.132034301757812,
+ "step": 27
+ },
+ {
+ "epoch": 0.39126637554585153,
+ "grad_norm": 0.10553059726953506,
+ "learning_rate": 0.000162,
+ "loss": 10.020434379577637,
+ "step": 28
+ },
+ {
+ "epoch": 0.4052401746724891,
+ "grad_norm": 0.1037299782037735,
+ "learning_rate": 0.000168,
+ "loss": 9.934466361999512,
+ "step": 29
+ },
+ {
+ "epoch": 0.4192139737991266,
+ "grad_norm": 0.1042739674448967,
+ "learning_rate": 0.00017399999999999997,
+ "loss": 9.822539329528809,
+ "step": 30
+ },
+ {
+ "epoch": 0.4331877729257642,
+ "grad_norm": 0.10300865024328232,
+ "learning_rate": 0.00017999999999999998,
+ "loss": 9.734994888305664,
+ "step": 31
+ },
+ {
+ "epoch": 0.44716157205240176,
+ "grad_norm": 0.10291654616594315,
+ "learning_rate": 0.000186,
+ "loss": 9.624505996704102,
+ "step": 32
+ },
+ {
+ "epoch": 0.4611353711790393,
+ "grad_norm": 0.10210851579904556,
+ "learning_rate": 0.00019199999999999998,
+ "loss": 9.518205642700195,
+ "step": 33
+ },
+ {
+ "epoch": 0.47510917030567684,
+ "grad_norm": 0.10222529619932175,
+ "learning_rate": 0.000198,
+ "loss": 9.412222862243652,
+ "step": 34
+ },
+ {
+ "epoch": 0.4890829694323144,
+ "grad_norm": 0.09992019087076187,
+ "learning_rate": 0.000204,
+ "loss": 9.326566696166992,
+ "step": 35
+ },
+ {
+ "epoch": 0.503056768558952,
+ "grad_norm": 0.09931084513664246,
+ "learning_rate": 0.00020999999999999998,
+ "loss": 9.210111618041992,
+ "step": 36
+ },
+ {
+ "epoch": 0.5170305676855895,
+ "grad_norm": 0.09928256273269653,
+ "learning_rate": 0.00021599999999999996,
+ "loss": 9.100493431091309,
+ "step": 37
+ },
+ {
+ "epoch": 0.5310043668122271,
+ "grad_norm": 0.09843816608190536,
+ "learning_rate": 0.00022199999999999998,
+ "loss": 8.996543884277344,
+ "step": 38
+ },
+ {
+ "epoch": 0.5449781659388646,
+ "grad_norm": 0.09825374186038971,
+ "learning_rate": 0.00022799999999999999,
+ "loss": 8.888991355895996,
+ "step": 39
+ },
+ {
+ "epoch": 0.5589519650655022,
+ "grad_norm": 0.09602385759353638,
+ "learning_rate": 0.000234,
+ "loss": 8.802907943725586,
+ "step": 40
+ },
+ {
+ "epoch": 0.5729257641921397,
+ "grad_norm": 0.09414064139127731,
+ "learning_rate": 0.00023999999999999998,
+ "loss": 8.715423583984375,
+ "step": 41
+ },
+ {
+ "epoch": 0.5868995633187772,
+ "grad_norm": 0.09273893386125565,
+ "learning_rate": 0.00024599999999999996,
+ "loss": 8.61374282836914,
+ "step": 42
+ },
+ {
+ "epoch": 0.6008733624454149,
+ "grad_norm": 0.09079599380493164,
+ "learning_rate": 0.00025199999999999995,
+ "loss": 8.502904891967773,
+ "step": 43
+ },
+ {
+ "epoch": 0.6148471615720524,
+ "grad_norm": 0.08605825901031494,
+ "learning_rate": 0.000258,
+ "loss": 8.453008651733398,
+ "step": 44
+ },
+ {
+ "epoch": 0.62882096069869,
+ "grad_norm": 0.08338996022939682,
+ "learning_rate": 0.00026399999999999997,
+ "loss": 8.365163803100586,
+ "step": 45
+ },
+ {
+ "epoch": 0.6427947598253275,
+ "grad_norm": 0.08150867372751236,
+ "learning_rate": 0.00027,
+ "loss": 8.285372734069824,
+ "step": 46
+ },
+ {
+ "epoch": 0.6567685589519651,
+ "grad_norm": 0.07794807851314545,
+ "learning_rate": 0.000276,
+ "loss": 8.185811996459961,
+ "step": 47
+ },
+ {
+ "epoch": 0.6707423580786026,
+ "grad_norm": 0.07592720538377762,
+ "learning_rate": 0.00028199999999999997,
+ "loss": 8.098482131958008,
+ "step": 48
+ },
+ {
+ "epoch": 0.6847161572052402,
+ "grad_norm": 0.0709897130727768,
+ "learning_rate": 0.00028799999999999995,
+ "loss": 8.042091369628906,
+ "step": 49
+ },
+ {
+ "epoch": 0.6986899563318777,
+ "grad_norm": 0.06582305580377579,
+ "learning_rate": 0.000294,
+ "loss": 7.972844123840332,
+ "step": 50
+ },
+ {
+ "epoch": 0.7126637554585152,
+ "grad_norm": 0.059452351182699203,
+ "learning_rate": 0.0003,
+ "loss": 7.924579620361328,
+ "step": 51
+ },
+ {
+ "epoch": 0.7266375545851529,
+ "grad_norm": 0.052426718175411224,
+ "learning_rate": 0.00030599999999999996,
+ "loss": 7.876336097717285,
+ "step": 52
+ },
+ {
+ "epoch": 0.7406113537117904,
+ "grad_norm": 0.04768150672316551,
+ "learning_rate": 0.000312,
+ "loss": 7.8102707862854,
+ "step": 53
+ },
+ {
+ "epoch": 0.754585152838428,
+ "grad_norm": 0.03970075398683548,
+ "learning_rate": 0.000318,
+ "loss": 7.791913986206055,
+ "step": 54
+ },
+ {
+ "epoch": 0.7685589519650655,
+ "grad_norm": 0.03559752553701401,
+ "learning_rate": 0.000324,
+ "loss": 7.756960868835449,
+ "step": 55
+ },
+ {
+ "epoch": 0.7825327510917031,
+ "grad_norm": 0.029321353882551193,
+ "learning_rate": 0.00033,
+ "loss": 7.724878787994385,
+ "step": 56
+ },
+ {
+ "epoch": 0.7965065502183406,
+ "grad_norm": 0.023485427722334862,
+ "learning_rate": 0.000336,
+ "loss": 7.72553014755249,
+ "step": 57
+ },
+ {
+ "epoch": 0.8104803493449781,
+ "grad_norm": 0.019319474697113037,
+ "learning_rate": 0.00034199999999999996,
+ "loss": 7.709493160247803,
+ "step": 58
+ },
+ {
+ "epoch": 0.8244541484716157,
+ "grad_norm": 0.016705691814422607,
+ "learning_rate": 0.00034799999999999995,
+ "loss": 7.6873297691345215,
+ "step": 59
+ },
+ {
+ "epoch": 0.8384279475982532,
+ "grad_norm": 0.017667734995484352,
+ "learning_rate": 0.00035399999999999993,
+ "loss": 7.677059173583984,
+ "step": 60
+ },
+ {
+ "epoch": 0.8524017467248908,
+ "grad_norm": 0.01671992614865303,
+ "learning_rate": 0.00035999999999999997,
+ "loss": 7.643512725830078,
+ "step": 61
+ },
+ {
+ "epoch": 0.8663755458515284,
+ "grad_norm": 0.012288172729313374,
+ "learning_rate": 0.00036599999999999995,
+ "loss": 7.658343315124512,
+ "step": 62
+ },
+ {
+ "epoch": 0.880349344978166,
+ "grad_norm": 0.012004776857793331,
+ "learning_rate": 0.000372,
+ "loss": 7.649759292602539,
+ "step": 63
+ },
+ {
+ "epoch": 0.8943231441048035,
+ "grad_norm": 0.012487445957958698,
+ "learning_rate": 0.00037799999999999997,
+ "loss": 7.631503105163574,
+ "step": 64
+ },
+ {
+ "epoch": 0.9082969432314411,
+ "grad_norm": 0.01489241048693657,
+ "learning_rate": 0.00038399999999999996,
+ "loss": 7.647110462188721,
+ "step": 65
+ },
+ {
+ "epoch": 0.9222707423580786,
+ "grad_norm": 0.01428980939090252,
+ "learning_rate": 0.00039,
+ "loss": 7.636983871459961,
+ "step": 66
+ },
+ {
+ "epoch": 0.9362445414847161,
+ "grad_norm": 0.010711144655942917,
+ "learning_rate": 0.000396,
+ "loss": 7.621130466461182,
+ "step": 67
+ },
+ {
+ "epoch": 0.9502183406113537,
+ "grad_norm": 0.010178886353969574,
+ "learning_rate": 0.000402,
+ "loss": 7.593924522399902,
+ "step": 68
+ },
+ {
+ "epoch": 0.9641921397379912,
+ "grad_norm": 0.010604874230921268,
+ "learning_rate": 0.000408,
+ "loss": 7.6015496253967285,
+ "step": 69
+ },
+ {
+ "epoch": 0.9781659388646288,
+ "grad_norm": 0.009751426056027412,
+ "learning_rate": 0.0004139999999999999,
+ "loss": 7.571964263916016,
+ "step": 70
+ },
+ {
+ "epoch": 0.9921397379912664,
+ "grad_norm": 0.009816481731832027,
+ "learning_rate": 0.00041999999999999996,
+ "loss": 7.57344913482666,
+ "step": 71
+ },
+ {
+ "epoch": 1.0,
+ "grad_norm": 0.008665637113153934,
+ "learning_rate": 0.00042599999999999995,
+ "loss": 7.579050064086914,
+ "step": 72
+ },
+ {
+ "epoch": 1.0,
+ "eval_loss": 7.581204891204834,
+ "eval_runtime": 59.0091,
+ "eval_samples_per_second": 41.383,
+ "eval_steps_per_second": 1.305,
+ "step": 72
+ },
+ {
+ "epoch": 1.0139737991266375,
+ "grad_norm": 0.009684201329946518,
+ "learning_rate": 0.00043199999999999993,
+ "loss": 7.568594932556152,
+ "step": 73
+ },
+ {
+ "epoch": 1.027947598253275,
+ "grad_norm": 0.025783710181713104,
+ "learning_rate": 0.00043799999999999997,
+ "loss": 7.578948497772217,
+ "step": 74
+ },
+ {
+ "epoch": 1.0419213973799126,
+ "grad_norm": 0.0327819399535656,
+ "learning_rate": 0.00044399999999999995,
+ "loss": 7.548173904418945,
+ "step": 75
+ },
+ {
+ "epoch": 1.0558951965065502,
+ "grad_norm": 0.021604498848319054,
+ "learning_rate": 0.00045,
+ "loss": 7.552953720092773,
+ "step": 76
+ },
+ {
+ "epoch": 1.0698689956331877,
+ "grad_norm": 0.010230105370283127,
+ "learning_rate": 0.00045599999999999997,
+ "loss": 7.5213212966918945,
+ "step": 77
+ },
+ {
+ "epoch": 1.0838427947598253,
+ "grad_norm": 0.023315589874982834,
+ "learning_rate": 0.00046199999999999995,
+ "loss": 7.501640319824219,
+ "step": 78
+ },
+ {
+ "epoch": 1.0978165938864628,
+ "grad_norm": 0.01109944935888052,
+ "learning_rate": 0.000468,
+ "loss": 7.539519786834717,
+ "step": 79
+ },
+ {
+ "epoch": 1.1117903930131003,
+ "grad_norm": 0.026093758642673492,
+ "learning_rate": 0.000474,
+ "loss": 7.525928974151611,
+ "step": 80
+ },
+ {
+ "epoch": 1.125764192139738,
+ "grad_norm": 0.03274895250797272,
+ "learning_rate": 0.00047999999999999996,
+ "loss": 7.535602569580078,
+ "step": 81
+ },
+ {
+ "epoch": 1.1397379912663754,
+ "grad_norm": 0.01543602254241705,
+ "learning_rate": 0.000486,
+ "loss": 7.513683795928955,
+ "step": 82
+ },
+ {
+ "epoch": 1.1537117903930132,
+ "grad_norm": 0.028377249836921692,
+ "learning_rate": 0.0004919999999999999,
+ "loss": 7.49384880065918,
+ "step": 83
+ },
+ {
+ "epoch": 1.1676855895196507,
+ "grad_norm": 0.04121481627225876,
+ "learning_rate": 0.000498,
+ "loss": 7.488447189331055,
+ "step": 84
+ },
+ {
+ "epoch": 1.1816593886462883,
+ "grad_norm": 0.013282302767038345,
+ "learning_rate": 0.0005039999999999999,
+ "loss": 7.463010787963867,
+ "step": 85
+ },
+ {
+ "epoch": 1.1956331877729258,
+ "grad_norm": 0.03200364485383034,
+ "learning_rate": 0.0005099999999999999,
+ "loss": 7.473515033721924,
+ "step": 86
+ },
+ {
+ "epoch": 1.2096069868995634,
+ "grad_norm": 0.020768029615283012,
+ "learning_rate": 0.000516,
+ "loss": 7.464688777923584,
+ "step": 87
+ },
+ {
+ "epoch": 1.223580786026201,
+ "grad_norm": 0.021715788170695305,
+ "learning_rate": 0.000522,
+ "loss": 7.4835028648376465,
+ "step": 88
+ },
+ {
+ "epoch": 1.2375545851528384,
+ "grad_norm": 0.026255754753947258,
+ "learning_rate": 0.0005279999999999999,
+ "loss": 7.443392753601074,
+ "step": 89
+ },
+ {
+ "epoch": 1.251528384279476,
+ "grad_norm": 0.022325951606035233,
+ "learning_rate": 0.000534,
+ "loss": 7.457373142242432,
+ "step": 90
+ },
+ {
+ "epoch": 1.2655021834061135,
+ "grad_norm": 0.009270387701690197,
+ "learning_rate": 0.00054,
+ "loss": 7.455986976623535,
+ "step": 91
+ },
+ {
+ "epoch": 1.279475982532751,
+ "grad_norm": 0.02369544468820095,
+ "learning_rate": 0.0005459999999999999,
+ "loss": 7.442216873168945,
+ "step": 92
+ },
+ {
+ "epoch": 1.2934497816593886,
+ "grad_norm": 0.008474506437778473,
+ "learning_rate": 0.000552,
+ "loss": 7.4294915199279785,
+ "step": 93
+ },
+ {
+ "epoch": 1.3074235807860262,
+ "grad_norm": 0.017043249681591988,
+ "learning_rate": 0.000558,
+ "loss": 7.406932830810547,
+ "step": 94
+ },
+ {
+ "epoch": 1.3213973799126637,
+ "grad_norm": 0.011813412420451641,
+ "learning_rate": 0.0005639999999999999,
+ "loss": 7.415900230407715,
+ "step": 95
+ },
+ {
+ "epoch": 1.3353711790393012,
+ "grad_norm": 0.01086883433163166,
+ "learning_rate": 0.00057,
+ "loss": 7.452998638153076,
+ "step": 96
+ },
+ {
+ "epoch": 1.3493449781659388,
+ "grad_norm": 0.014700951054692268,
+ "learning_rate": 0.0005759999999999999,
+ "loss": 7.410249710083008,
+ "step": 97
+ },
+ {
+ "epoch": 1.3633187772925766,
+ "grad_norm": 0.014512976631522179,
+ "learning_rate": 0.0005819999999999999,
+ "loss": 7.429132461547852,
+ "step": 98
+ },
+ {
+ "epoch": 1.3772925764192139,
+ "grad_norm": 0.012519920244812965,
+ "learning_rate": 0.000588,
+ "loss": 7.430817604064941,
+ "step": 99
+ },
+ {
+ "epoch": 1.3912663755458516,
+ "grad_norm": 0.007132664322853088,
+ "learning_rate": 0.0005939999999999999,
+ "loss": 7.405234336853027,
+ "step": 100
+ },
+ {
+ "epoch": 1.405240174672489,
+ "grad_norm": 0.019489388912916183,
+ "learning_rate": 0.0006,
+ "loss": 7.427791595458984,
+ "step": 101
+ },
+ {
+ "epoch": 1.4192139737991267,
+ "grad_norm": 0.027252664789557457,
+ "learning_rate": 0.0006,
+ "loss": 7.41761589050293,
+ "step": 102
+ },
+ {
+ "epoch": 1.4331877729257643,
+ "grad_norm": 0.02501637488603592,
+ "learning_rate": 0.0006,
+ "loss": 7.416735649108887,
+ "step": 103
+ },
+ {
+ "epoch": 1.4471615720524018,
+ "grad_norm": 0.00989772193133831,
+ "learning_rate": 0.0006,
+ "loss": 7.3756632804870605,
+ "step": 104
+ },
+ {
+ "epoch": 1.4611353711790394,
+ "grad_norm": 0.01622949168086052,
+ "learning_rate": 0.0006,
+ "loss": 7.400811672210693,
+ "step": 105
+ },
+ {
+ "epoch": 1.475109170305677,
+ "grad_norm": 0.012407763861119747,
+ "learning_rate": 0.0006,
+ "loss": 7.37675666809082,
+ "step": 106
+ },
+ {
+ "epoch": 1.4890829694323144,
+ "grad_norm": 0.010353565216064453,
+ "learning_rate": 0.0006,
+ "loss": 7.4013142585754395,
+ "step": 107
+ },
+ {
+ "epoch": 1.503056768558952,
+ "grad_norm": 0.0120553532615304,
+ "learning_rate": 0.0006,
+ "loss": 7.379873275756836,
+ "step": 108
+ },
+ {
+ "epoch": 1.5170305676855895,
+ "grad_norm": 0.01610225811600685,
+ "learning_rate": 0.0006,
+ "loss": 7.371631145477295,
+ "step": 109
+ },
+ {
+ "epoch": 1.531004366812227,
+ "grad_norm": 0.01159879844635725,
+ "learning_rate": 0.0006,
+ "loss": 7.389007568359375,
+ "step": 110
+ },
+ {
+ "epoch": 1.5449781659388646,
+ "grad_norm": 0.008013816550374031,
+ "learning_rate": 0.0006,
+ "loss": 7.395920753479004,
+ "step": 111
+ },
+ {
+ "epoch": 1.5589519650655022,
+ "grad_norm": 0.016334809362888336,
+ "learning_rate": 0.0006,
+ "loss": 7.393121719360352,
+ "step": 112
+ },
+ {
+ "epoch": 1.5729257641921397,
+ "grad_norm": 0.018821721896529198,
+ "learning_rate": 0.0006,
+ "loss": 7.367543697357178,
+ "step": 113
+ },
+ {
+ "epoch": 1.5868995633187772,
+ "grad_norm": 0.014088994823396206,
+ "learning_rate": 0.0006,
+ "loss": 7.385537624359131,
+ "step": 114
+ },
+ {
+ "epoch": 1.600873362445415,
+ "grad_norm": 0.005864050704985857,
+ "learning_rate": 0.0006,
+ "loss": 7.365346908569336,
+ "step": 115
+ },
+ {
+ "epoch": 1.6148471615720523,
+ "grad_norm": 0.011745430529117584,
+ "learning_rate": 0.0006,
+ "loss": 7.390227317810059,
+ "step": 116
+ },
+ {
+ "epoch": 1.62882096069869,
+ "grad_norm": 0.017205573618412018,
+ "learning_rate": 0.0006,
+ "loss": 7.3558549880981445,
+ "step": 117
+ },
+ {
+ "epoch": 1.6427947598253274,
+ "grad_norm": 0.021256346255540848,
+ "learning_rate": 0.0006,
+ "loss": 7.377346992492676,
+ "step": 118
+ },
+ {
+ "epoch": 1.6567685589519652,
+ "grad_norm": 0.019335204735398293,
+ "learning_rate": 0.0006,
+ "loss": 7.346194267272949,
+ "step": 119
+ },
+ {
+ "epoch": 1.6707423580786025,
+ "grad_norm": 0.008134003728628159,
+ "learning_rate": 0.0006,
+ "loss": 7.3063578605651855,
+ "step": 120
+ },
+ {
+ "epoch": 1.6847161572052403,
+ "grad_norm": 0.017613831907510757,
+ "learning_rate": 0.0006,
+ "loss": 7.338171005249023,
+ "step": 121
+ },
+ {
+ "epoch": 1.6986899563318776,
+ "grad_norm": 0.00890734139829874,
+ "learning_rate": 0.0006,
+ "loss": 7.355384826660156,
+ "step": 122
+ },
+ {
+ "epoch": 1.7126637554585153,
+ "grad_norm": 0.013448155485093594,
+ "learning_rate": 0.0006,
+ "loss": 7.344296455383301,
+ "step": 123
+ },
+ {
+ "epoch": 1.726637554585153,
+ "grad_norm": 0.009241633117198944,
+ "learning_rate": 0.0006,
+ "loss": 7.3454060554504395,
+ "step": 124
+ },
+ {
+ "epoch": 1.7406113537117904,
+ "grad_norm": 0.018561311066150665,
+ "learning_rate": 0.0006,
+ "loss": 7.32348108291626,
+ "step": 125
+ },
+ {
+ "epoch": 1.754585152838428,
+ "grad_norm": 0.025667704641819,
+ "learning_rate": 0.0006,
+ "loss": 7.356142520904541,
+ "step": 126
+ },
+ {
+ "epoch": 1.7685589519650655,
+ "grad_norm": 0.036232151091098785,
+ "learning_rate": 0.0006,
+ "loss": 7.363539695739746,
+ "step": 127
+ },
+ {
+ "epoch": 1.782532751091703,
+ "grad_norm": 0.059483714401721954,
+ "learning_rate": 0.0006,
+ "loss": 7.402331352233887,
+ "step": 128
+ },
+ {
+ "epoch": 1.7965065502183406,
+ "grad_norm": 0.05365009233355522,
+ "learning_rate": 0.0006,
+ "loss": 7.40457820892334,
+ "step": 129
+ },
+ {
+ "epoch": 1.8104803493449781,
+ "grad_norm": 0.02738415263593197,
+ "learning_rate": 0.0006,
+ "loss": 7.3756208419799805,
+ "step": 130
+ },
+ {
+ "epoch": 1.8244541484716157,
+ "grad_norm": 0.035017091780900955,
+ "learning_rate": 0.0006,
+ "loss": 7.379263877868652,
+ "step": 131
+ },
+ {
+ "epoch": 1.8384279475982532,
+ "grad_norm": 0.02819863148033619,
+ "learning_rate": 0.0006,
+ "loss": 7.327673435211182,
+ "step": 132
+ },
+ {
+ "epoch": 1.8524017467248908,
+ "grad_norm": 0.017754238098859787,
+ "learning_rate": 0.0006,
+ "loss": 7.333420276641846,
+ "step": 133
+ },
+ {
+ "epoch": 1.8663755458515285,
+ "grad_norm": 0.024685733020305634,
+ "learning_rate": 0.0006,
+ "loss": 7.34493350982666,
+ "step": 134
+ },
+ {
+ "epoch": 1.8803493449781659,
+ "grad_norm": 0.019426532089710236,
+ "learning_rate": 0.0006,
+ "loss": 7.328448295593262,
+ "step": 135
+ },
+ {
+ "epoch": 1.8943231441048036,
+ "grad_norm": 0.02106391079723835,
+ "learning_rate": 0.0006,
+ "loss": 7.327917098999023,
+ "step": 136
+ },
+ {
+ "epoch": 1.908296943231441,
+ "grad_norm": 0.02056879736483097,
+ "learning_rate": 0.0006,
+ "loss": 7.3311991691589355,
+ "step": 137
+ },
+ {
+ "epoch": 1.9222707423580787,
+ "grad_norm": 0.016865141689777374,
+ "learning_rate": 0.0006,
+ "loss": 7.312766075134277,
+ "step": 138
+ },
+ {
+ "epoch": 1.936244541484716,
+ "grad_norm": 0.017549855634570122,
+ "learning_rate": 0.0006,
+ "loss": 7.31191349029541,
+ "step": 139
+ },
+ {
+ "epoch": 1.9502183406113538,
+ "grad_norm": 0.015930157154798508,
+ "learning_rate": 0.0006,
+ "loss": 7.330901622772217,
+ "step": 140
+ },
+ {
+ "epoch": 1.9641921397379911,
+ "grad_norm": 0.013283558189868927,
+ "learning_rate": 0.0006,
+ "loss": 7.298318386077881,
+ "step": 141
+ },
+ {
+ "epoch": 1.9781659388646289,
+ "grad_norm": 0.012023469433188438,
+ "learning_rate": 0.0006,
+ "loss": 7.282337188720703,
+ "step": 142
+ },
+ {
+ "epoch": 1.9921397379912664,
+ "grad_norm": 0.01459511648863554,
+ "learning_rate": 0.0006,
+ "loss": 7.282924175262451,
+ "step": 143
+ },
+ {
+ "epoch": 2.0,
+ "grad_norm": 0.017474235966801643,
+ "learning_rate": 0.0006,
+ "loss": 7.254333972930908,
+ "step": 144
+ },
+ {
+ "epoch": 2.0,
+ "eval_loss": 7.318708896636963,
+ "eval_runtime": 59.6555,
+ "eval_samples_per_second": 40.935,
+ "eval_steps_per_second": 1.291,
+ "step": 144
+ },
+ {
+ "epoch": 2.0139737991266378,
+ "grad_norm": 0.019757507368922234,
+ "learning_rate": 0.0006,
+ "loss": 7.30288553237915,
+ "step": 145
+ },
+ {
+ "epoch": 2.027947598253275,
+ "grad_norm": 0.014571248553693295,
+ "learning_rate": 0.0006,
+ "loss": 7.274681568145752,
+ "step": 146
+ },
+ {
+ "epoch": 2.041921397379913,
+ "grad_norm": 0.015744198113679886,
+ "learning_rate": 0.0006,
+ "loss": 7.278632640838623,
+ "step": 147
+ },
+ {
+ "epoch": 2.05589519650655,
+ "grad_norm": 0.028048858046531677,
+ "learning_rate": 0.0006,
+ "loss": 7.26558780670166,
+ "step": 148
+ },
+ {
+ "epoch": 2.069868995633188,
+ "grad_norm": 0.06547307223081589,
+ "learning_rate": 0.0006,
+ "loss": 7.32567834854126,
+ "step": 149
+ },
+ {
+ "epoch": 2.0838427947598253,
+ "grad_norm": 0.05969763547182083,
+ "learning_rate": 0.0006,
+ "loss": 7.395836353302002,
+ "step": 150
+ },
+ {
+ "epoch": 2.097816593886463,
+ "grad_norm": 0.032574281096458435,
+ "learning_rate": 0.0006,
+ "loss": 7.318150997161865,
+ "step": 151
+ },
+ {
+ "epoch": 2.1117903930131003,
+ "grad_norm": 0.029452962800860405,
+ "learning_rate": 0.0006,
+ "loss": 7.308066368103027,
+ "step": 152
+ },
+ {
+ "epoch": 2.125764192139738,
+ "grad_norm": 0.03161991387605667,
+ "learning_rate": 0.0006,
+ "loss": 7.306290626525879,
+ "step": 153
+ },
+ {
+ "epoch": 2.1397379912663754,
+ "grad_norm": 0.027096295729279518,
+ "learning_rate": 0.0006,
+ "loss": 7.297796249389648,
+ "step": 154
+ },
+ {
+ "epoch": 2.153711790393013,
+ "grad_norm": 0.028657056391239166,
+ "learning_rate": 0.0006,
+ "loss": 7.317424774169922,
+ "step": 155
+ },
+ {
+ "epoch": 2.1676855895196505,
+ "grad_norm": 0.02046995982527733,
+ "learning_rate": 0.0006,
+ "loss": 7.297986030578613,
+ "step": 156
+ },
+ {
+ "epoch": 2.1816593886462883,
+ "grad_norm": 0.02220369316637516,
+ "learning_rate": 0.0006,
+ "loss": 7.268237590789795,
+ "step": 157
+ },
+ {
+ "epoch": 2.1956331877729256,
+ "grad_norm": 0.018381357192993164,
+ "learning_rate": 0.0006,
+ "loss": 7.259415626525879,
+ "step": 158
+ },
+ {
+ "epoch": 2.2096069868995634,
+ "grad_norm": 0.0204413291066885,
+ "learning_rate": 0.0006,
+ "loss": 7.258173942565918,
+ "step": 159
+ },
+ {
+ "epoch": 2.2235807860262007,
+ "grad_norm": 0.015874288976192474,
+ "learning_rate": 0.0006,
+ "loss": 7.281834602355957,
+ "step": 160
+ },
+ {
+ "epoch": 2.2375545851528384,
+ "grad_norm": 0.01634068600833416,
+ "learning_rate": 0.0006,
+ "loss": 7.2469868659973145,
+ "step": 161
+ },
+ {
+ "epoch": 2.251528384279476,
+ "grad_norm": 0.01728747971355915,
+ "learning_rate": 0.0006,
+ "loss": 7.280995845794678,
+ "step": 162
+ },
+ {
+ "epoch": 2.2655021834061135,
+ "grad_norm": 0.016341188922524452,
+ "learning_rate": 0.0006,
+ "loss": 7.273357391357422,
+ "step": 163
+ },
+ {
+ "epoch": 2.279475982532751,
+ "grad_norm": 0.016125807538628578,
+ "learning_rate": 0.0006,
+ "loss": 7.273510932922363,
+ "step": 164
+ },
+ {
+ "epoch": 2.2934497816593886,
+ "grad_norm": 0.01629287749528885,
+ "learning_rate": 0.0006,
+ "loss": 7.267029762268066,
+ "step": 165
+ },
+ {
+ "epoch": 2.3074235807860264,
+ "grad_norm": 0.015296131372451782,
+ "learning_rate": 0.0006,
+ "loss": 7.211989402770996,
+ "step": 166
+ },
+ {
+ "epoch": 2.3213973799126637,
+ "grad_norm": 0.01709570363163948,
+ "learning_rate": 0.0006,
+ "loss": 7.242851734161377,
+ "step": 167
+ },
+ {
+ "epoch": 2.3353711790393015,
+ "grad_norm": 0.012552388943731785,
+ "learning_rate": 0.0006,
+ "loss": 7.241696834564209,
+ "step": 168
+ },
+ {
+ "epoch": 2.349344978165939,
+ "grad_norm": 0.014571291394531727,
+ "learning_rate": 0.0006,
+ "loss": 7.270681381225586,
+ "step": 169
+ },
+ {
+ "epoch": 2.3633187772925766,
+ "grad_norm": 0.013389437459409237,
+ "learning_rate": 0.0006,
+ "loss": 7.191600799560547,
+ "step": 170
+ },
+ {
+ "epoch": 2.377292576419214,
+ "grad_norm": 0.011895314790308475,
+ "learning_rate": 0.0006,
+ "loss": 7.213610649108887,
+ "step": 171
+ },
+ {
+ "epoch": 2.3912663755458516,
+ "grad_norm": 0.01345642190426588,
+ "learning_rate": 0.0006,
+ "loss": 7.193787574768066,
+ "step": 172
+ },
+ {
+ "epoch": 2.405240174672489,
+ "grad_norm": 0.0091794328764081,
+ "learning_rate": 0.0006,
+ "loss": 7.233364105224609,
+ "step": 173
+ },
+ {
+ "epoch": 2.4192139737991267,
+ "grad_norm": 0.013767086900770664,
+ "learning_rate": 0.0006,
+ "loss": 7.217272758483887,
+ "step": 174
+ },
+ {
+ "epoch": 2.433187772925764,
+ "grad_norm": 0.015134960412979126,
+ "learning_rate": 0.0006,
+ "loss": 7.192791938781738,
+ "step": 175
+ },
+ {
+ "epoch": 2.447161572052402,
+ "grad_norm": 0.01790648326277733,
+ "learning_rate": 0.0006,
+ "loss": 7.229467391967773,
+ "step": 176
+ },
+ {
+ "epoch": 2.461135371179039,
+ "grad_norm": 0.015362377278506756,
+ "learning_rate": 0.0006,
+ "loss": 7.180346488952637,
+ "step": 177
+ },
+ {
+ "epoch": 2.475109170305677,
+ "grad_norm": 0.010505498386919498,
+ "learning_rate": 0.0006,
+ "loss": 7.188833236694336,
+ "step": 178
+ },
+ {
+ "epoch": 2.489082969432314,
+ "grad_norm": 0.009129747748374939,
+ "learning_rate": 0.0006,
+ "loss": 7.203153133392334,
+ "step": 179
+ },
+ {
+ "epoch": 2.503056768558952,
+ "grad_norm": 0.01147883478552103,
+ "learning_rate": 0.0006,
+ "loss": 7.163824558258057,
+ "step": 180
+ },
+ {
+ "epoch": 2.5170305676855893,
+ "grad_norm": 0.011338942684233189,
+ "learning_rate": 0.0006,
+ "loss": 7.16287899017334,
+ "step": 181
+ },
+ {
+ "epoch": 2.531004366812227,
+ "grad_norm": 0.013277675025165081,
+ "learning_rate": 0.0006,
+ "loss": 7.155905723571777,
+ "step": 182
+ },
+ {
+ "epoch": 2.544978165938865,
+ "grad_norm": 0.015269347466528416,
+ "learning_rate": 0.0006,
+ "loss": 7.189371109008789,
+ "step": 183
+ },
+ {
+ "epoch": 2.558951965065502,
+ "grad_norm": 0.017905596643686295,
+ "learning_rate": 0.0006,
+ "loss": 7.177547931671143,
+ "step": 184
+ },
+ {
+ "epoch": 2.5729257641921395,
+ "grad_norm": 0.024431610479950905,
+ "learning_rate": 0.0006,
+ "loss": 7.192169189453125,
+ "step": 185
+ },
+ {
+ "epoch": 2.5868995633187772,
+ "grad_norm": 0.03656961768865585,
+ "learning_rate": 0.0006,
+ "loss": 7.124387741088867,
+ "step": 186
+ },
+ {
+ "epoch": 2.600873362445415,
+ "grad_norm": 0.04337048903107643,
+ "learning_rate": 0.0006,
+ "loss": 7.140231609344482,
+ "step": 187
+ },
+ {
+ "epoch": 2.6148471615720523,
+ "grad_norm": 0.04047228768467903,
+ "learning_rate": 0.0006,
+ "loss": 7.171927452087402,
+ "step": 188
+ },
+ {
+ "epoch": 2.62882096069869,
+ "grad_norm": 0.07825770974159241,
+ "learning_rate": 0.0006,
+ "loss": 7.207631587982178,
+ "step": 189
+ },
+ {
+ "epoch": 2.6427947598253274,
+ "grad_norm": 0.07118063420057297,
+ "learning_rate": 0.0006,
+ "loss": 7.209488391876221,
+ "step": 190
+ },
+ {
+ "epoch": 2.656768558951965,
+ "grad_norm": 0.05153293535113335,
+ "learning_rate": 0.0006,
+ "loss": 7.14793062210083,
+ "step": 191
+ },
+ {
+ "epoch": 2.6707423580786025,
+ "grad_norm": 0.03318839520215988,
+ "learning_rate": 0.0006,
+ "loss": 7.163010597229004,
+ "step": 192
+ },
+ {
+ "epoch": 2.6847161572052403,
+ "grad_norm": 0.029333539307117462,
+ "learning_rate": 0.0006,
+ "loss": 7.159955978393555,
+ "step": 193
+ },
+ {
+ "epoch": 2.6986899563318776,
+ "grad_norm": 0.029405880719423294,
+ "learning_rate": 0.0006,
+ "loss": 7.136178493499756,
+ "step": 194
+ },
+ {
+ "epoch": 2.7126637554585153,
+ "grad_norm": 0.037815988063812256,
+ "learning_rate": 0.0006,
+ "loss": 7.183186054229736,
+ "step": 195
+ },
+ {
+ "epoch": 2.726637554585153,
+ "grad_norm": 0.020307250320911407,
+ "learning_rate": 0.0006,
+ "loss": 7.133164882659912,
+ "step": 196
+ },
+ {
+ "epoch": 2.7406113537117904,
+ "grad_norm": 0.026048069819808006,
+ "learning_rate": 0.0006,
+ "loss": 7.124977111816406,
+ "step": 197
+ },
+ {
+ "epoch": 2.7545851528384278,
+ "grad_norm": 0.027316724881529808,
+ "learning_rate": 0.0006,
+ "loss": 7.07265567779541,
+ "step": 198
+ },
+ {
+ "epoch": 2.7685589519650655,
+ "grad_norm": 0.021629009395837784,
+ "learning_rate": 0.0006,
+ "loss": 7.150020599365234,
+ "step": 199
+ },
+ {
+ "epoch": 2.7825327510917033,
+ "grad_norm": 0.01605929434299469,
+ "learning_rate": 0.0006,
+ "loss": 7.10919189453125,
+ "step": 200
+ },
+ {
+ "epoch": 2.7965065502183406,
+ "grad_norm": 0.020800089463591576,
+ "learning_rate": 0.0006,
+ "loss": 7.132045269012451,
+ "step": 201
+ },
+ {
+ "epoch": 2.810480349344978,
+ "grad_norm": 0.019225774332880974,
+ "learning_rate": 0.0006,
+ "loss": 7.128364562988281,
+ "step": 202
+ },
+ {
+ "epoch": 2.8244541484716157,
+ "grad_norm": 0.01561372634023428,
+ "learning_rate": 0.0006,
+ "loss": 7.083669662475586,
+ "step": 203
+ },
+ {
+ "epoch": 2.8384279475982535,
+ "grad_norm": 0.021969519555568695,
+ "learning_rate": 0.0006,
+ "loss": 7.050841331481934,
+ "step": 204
+ },
+ {
+ "epoch": 2.8524017467248908,
+ "grad_norm": 0.021453695371747017,
+ "learning_rate": 0.0006,
+ "loss": 7.097085475921631,
+ "step": 205
+ },
+ {
+ "epoch": 2.8663755458515285,
+ "grad_norm": 0.022836022078990936,
+ "learning_rate": 0.0006,
+ "loss": 7.10405969619751,
+ "step": 206
+ },
+ {
+ "epoch": 2.880349344978166,
+ "grad_norm": 0.02430851384997368,
+ "learning_rate": 0.0006,
+ "loss": 7.045181751251221,
+ "step": 207
+ },
+ {
+ "epoch": 2.8943231441048036,
+ "grad_norm": 0.020166993141174316,
+ "learning_rate": 0.0006,
+ "loss": 7.068183898925781,
+ "step": 208
+ },
+ {
+ "epoch": 2.908296943231441,
+ "grad_norm": 0.0118505684658885,
+ "learning_rate": 0.0006,
+ "loss": 7.07132625579834,
+ "step": 209
+ },
+ {
+ "epoch": 2.9222707423580787,
+ "grad_norm": 0.023772427812218666,
+ "learning_rate": 0.0006,
+ "loss": 7.055238723754883,
+ "step": 210
+ },
+ {
+ "epoch": 2.936244541484716,
+ "grad_norm": 0.02350712940096855,
+ "learning_rate": 0.0006,
+ "loss": 7.055898666381836,
+ "step": 211
+ },
+ {
+ "epoch": 2.950218340611354,
+ "grad_norm": 0.017244907096028328,
+ "learning_rate": 0.0006,
+ "loss": 7.0099711418151855,
+ "step": 212
+ },
+ {
+ "epoch": 2.964192139737991,
+ "grad_norm": 0.021565575152635574,
+ "learning_rate": 0.0006,
+ "loss": 6.991751670837402,
+ "step": 213
+ },
+ {
+ "epoch": 2.978165938864629,
+ "grad_norm": 0.03241860866546631,
+ "learning_rate": 0.0006,
+ "loss": 7.005980491638184,
+ "step": 214
+ },
+ {
+ "epoch": 2.992139737991266,
+ "grad_norm": 0.04613726586103439,
+ "learning_rate": 0.0006,
+ "loss": 7.027859687805176,
+ "step": 215
+ },
+ {
+ "epoch": 3.0,
+ "grad_norm": 0.06404894590377808,
+ "learning_rate": 0.0006,
+ "loss": 7.123725891113281,
+ "step": 216
+ },
+ {
+ "epoch": 3.0,
+ "eval_loss": 7.070647716522217,
+ "eval_runtime": 58.5801,
+ "eval_samples_per_second": 41.687,
+ "eval_steps_per_second": 1.314,
+ "step": 216
+ },
+ {
+ "epoch": 3.0139737991266378,
+ "grad_norm": 0.05542432889342308,
+ "learning_rate": 0.0006,
+ "loss": 7.078299522399902,
+ "step": 217
+ },
+ {
+ "epoch": 3.027947598253275,
+ "grad_norm": 0.026484526693820953,
+ "learning_rate": 0.0006,
+ "loss": 6.961159706115723,
+ "step": 218
+ },
+ {
+ "epoch": 3.041921397379913,
+ "grad_norm": 0.028276391327381134,
+ "learning_rate": 0.0006,
+ "loss": 7.019961833953857,
+ "step": 219
+ },
+ {
+ "epoch": 3.05589519650655,
+ "grad_norm": 0.024486595764756203,
+ "learning_rate": 0.0006,
+ "loss": 7.039677619934082,
+ "step": 220
+ },
+ {
+ "epoch": 3.069868995633188,
+ "grad_norm": 0.017620893195271492,
+ "learning_rate": 0.0006,
+ "loss": 6.9270782470703125,
+ "step": 221
+ },
+ {
+ "epoch": 3.0838427947598253,
+ "grad_norm": 0.02547302283346653,
+ "learning_rate": 0.0006,
+ "loss": 6.924233436584473,
+ "step": 222
+ },
+ {
+ "epoch": 3.097816593886463,
+ "grad_norm": 0.02539009042084217,
+ "learning_rate": 0.0006,
+ "loss": 6.973979949951172,
+ "step": 223
+ },
+ {
+ "epoch": 3.1117903930131003,
+ "grad_norm": 0.03256227448582649,
+ "learning_rate": 0.0006,
+ "loss": 6.933725357055664,
+ "step": 224
+ },
+ {
+ "epoch": 3.125764192139738,
+ "grad_norm": 0.04651800915598869,
+ "learning_rate": 0.0006,
+ "loss": 6.951044082641602,
+ "step": 225
+ },
+ {
+ "epoch": 3.1397379912663754,
+ "grad_norm": 0.04822975769639015,
+ "learning_rate": 0.0006,
+ "loss": 6.9614410400390625,
+ "step": 226
+ },
+ {
+ "epoch": 3.153711790393013,
+ "grad_norm": 0.031165020540356636,
+ "learning_rate": 0.0006,
+ "loss": 6.91193151473999,
+ "step": 227
+ },
+ {
+ "epoch": 3.1676855895196505,
+ "grad_norm": 0.029754292219877243,
+ "learning_rate": 0.0006,
+ "loss": 6.91802978515625,
+ "step": 228
+ },
+ {
+ "epoch": 3.1816593886462883,
+ "grad_norm": 0.02193152718245983,
+ "learning_rate": 0.0006,
+ "loss": 6.928090572357178,
+ "step": 229
+ },
+ {
+ "epoch": 3.1956331877729256,
+ "grad_norm": 0.02428470179438591,
+ "learning_rate": 0.0006,
+ "loss": 6.90396785736084,
+ "step": 230
+ },
+ {
+ "epoch": 3.2096069868995634,
+ "grad_norm": 0.023374345153570175,
+ "learning_rate": 0.0006,
+ "loss": 6.885858535766602,
+ "step": 231
+ },
+ {
+ "epoch": 3.2235807860262007,
+ "grad_norm": 0.023405000567436218,
+ "learning_rate": 0.0006,
+ "loss": 6.8855767250061035,
+ "step": 232
+ },
+ {
+ "epoch": 3.2375545851528384,
+ "grad_norm": 0.017277134582400322,
+ "learning_rate": 0.0006,
+ "loss": 6.901449203491211,
+ "step": 233
+ },
+ {
+ "epoch": 3.251528384279476,
+ "grad_norm": 0.019496750086545944,
+ "learning_rate": 0.0006,
+ "loss": 6.885687828063965,
+ "step": 234
+ },
+ {
+ "epoch": 3.2655021834061135,
+ "grad_norm": 0.016893276944756508,
+ "learning_rate": 0.0006,
+ "loss": 6.872369766235352,
+ "step": 235
+ },
+ {
+ "epoch": 3.279475982532751,
+ "grad_norm": 0.02058715932071209,
+ "learning_rate": 0.0006,
+ "loss": 6.823280334472656,
+ "step": 236
+ },
+ {
+ "epoch": 3.2934497816593886,
+ "grad_norm": 0.015530755743384361,
+ "learning_rate": 0.0006,
+ "loss": 6.860499858856201,
+ "step": 237
+ },
+ {
+ "epoch": 3.3074235807860264,
+ "grad_norm": 0.01934926211833954,
+ "learning_rate": 0.0006,
+ "loss": 6.817448616027832,
+ "step": 238
+ },
+ {
+ "epoch": 3.3213973799126637,
+ "grad_norm": 0.017413552850484848,
+ "learning_rate": 0.0006,
+ "loss": 6.881955146789551,
+ "step": 239
+ },
+ {
+ "epoch": 3.3353711790393015,
+ "grad_norm": 0.026478875428438187,
+ "learning_rate": 0.0006,
+ "loss": 6.827404499053955,
+ "step": 240
+ },
+ {
+ "epoch": 3.349344978165939,
+ "grad_norm": 0.032611701637506485,
+ "learning_rate": 0.0006,
+ "loss": 6.78285551071167,
+ "step": 241
+ },
+ {
+ "epoch": 3.3633187772925766,
+ "grad_norm": 0.041900306940078735,
+ "learning_rate": 0.0006,
+ "loss": 6.783053874969482,
+ "step": 242
+ },
+ {
+ "epoch": 3.377292576419214,
+ "grad_norm": 0.04149497672915459,
+ "learning_rate": 0.0006,
+ "loss": 6.745220184326172,
+ "step": 243
+ },
+ {
+ "epoch": 3.3912663755458516,
+ "grad_norm": 0.023949826136231422,
+ "learning_rate": 0.0006,
+ "loss": 6.841533184051514,
+ "step": 244
+ },
+ {
+ "epoch": 3.405240174672489,
+ "grad_norm": 0.028601842001080513,
+ "learning_rate": 0.0006,
+ "loss": 6.852852821350098,
+ "step": 245
+ },
+ {
+ "epoch": 3.4192139737991267,
+ "grad_norm": 0.027365155518054962,
+ "learning_rate": 0.0006,
+ "loss": 6.74376106262207,
+ "step": 246
+ },
+ {
+ "epoch": 3.433187772925764,
+ "grad_norm": 0.02856568619608879,
+ "learning_rate": 0.0006,
+ "loss": 6.74906063079834,
+ "step": 247
+ },
+ {
+ "epoch": 3.447161572052402,
+ "grad_norm": 0.019172416999936104,
+ "learning_rate": 0.0006,
+ "loss": 6.809206008911133,
+ "step": 248
+ },
+ {
+ "epoch": 3.461135371179039,
+ "grad_norm": 0.02161746844649315,
+ "learning_rate": 0.0006,
+ "loss": 6.789888381958008,
+ "step": 249
+ },
+ {
+ "epoch": 3.475109170305677,
+ "grad_norm": 0.020273666828870773,
+ "learning_rate": 0.0006,
+ "loss": 6.73512077331543,
+ "step": 250
+ },
+ {
+ "epoch": 3.489082969432314,
+ "grad_norm": 0.022329283878207207,
+ "learning_rate": 0.0006,
+ "loss": 6.826148509979248,
+ "step": 251
+ },
+ {
+ "epoch": 3.503056768558952,
+ "grad_norm": 0.02284284122288227,
+ "learning_rate": 0.0006,
+ "loss": 6.748521327972412,
+ "step": 252
+ },
+ {
+ "epoch": 3.5170305676855893,
+ "grad_norm": 0.026368247345089912,
+ "learning_rate": 0.0006,
+ "loss": 6.645879745483398,
+ "step": 253
+ },
+ {
+ "epoch": 3.531004366812227,
+ "grad_norm": 0.03627244383096695,
+ "learning_rate": 0.0006,
+ "loss": 6.666064739227295,
+ "step": 254
+ },
+ {
+ "epoch": 3.544978165938865,
+ "grad_norm": 0.04140935093164444,
+ "learning_rate": 0.0006,
+ "loss": 6.712275981903076,
+ "step": 255
+ },
+ {
+ "epoch": 3.558951965065502,
+ "grad_norm": 0.0390462800860405,
+ "learning_rate": 0.0006,
+ "loss": 6.665900230407715,
+ "step": 256
+ },
+ {
+ "epoch": 3.5729257641921395,
+ "grad_norm": 0.05178583785891533,
+ "learning_rate": 0.0006,
+ "loss": 6.753298759460449,
+ "step": 257
+ },
+ {
+ "epoch": 3.5868995633187772,
+ "grad_norm": 0.054328061640262604,
+ "learning_rate": 0.0006,
+ "loss": 6.732633113861084,
+ "step": 258
+ },
+ {
+ "epoch": 3.600873362445415,
+ "grad_norm": 0.04590460658073425,
+ "learning_rate": 0.0006,
+ "loss": 6.65250301361084,
+ "step": 259
+ },
+ {
+ "epoch": 3.6148471615720523,
+ "grad_norm": 0.044331666082143784,
+ "learning_rate": 0.0006,
+ "loss": 6.68654203414917,
+ "step": 260
+ },
+ {
+ "epoch": 3.62882096069869,
+ "grad_norm": 0.05308730527758598,
+ "learning_rate": 0.0006,
+ "loss": 6.642482757568359,
+ "step": 261
+ },
+ {
+ "epoch": 3.6427947598253274,
+ "grad_norm": 0.03999016433954239,
+ "learning_rate": 0.0006,
+ "loss": 6.625227451324463,
+ "step": 262
+ },
+ {
+ "epoch": 3.656768558951965,
+ "grad_norm": 0.04088086634874344,
+ "learning_rate": 0.0006,
+ "loss": 6.67623233795166,
+ "step": 263
+ },
+ {
+ "epoch": 3.6707423580786025,
+ "grad_norm": 0.03615636005997658,
+ "learning_rate": 0.0006,
+ "loss": 6.6869001388549805,
+ "step": 264
+ },
+ {
+ "epoch": 3.6847161572052403,
+ "grad_norm": 0.02817360684275627,
+ "learning_rate": 0.0006,
+ "loss": 6.674601078033447,
+ "step": 265
+ },
+ {
+ "epoch": 3.6986899563318776,
+ "grad_norm": 0.032586123794317245,
+ "learning_rate": 0.0006,
+ "loss": 6.612994194030762,
+ "step": 266
+ },
+ {
+ "epoch": 3.7126637554585153,
+ "grad_norm": 0.031117867678403854,
+ "learning_rate": 0.0006,
+ "loss": 6.6244354248046875,
+ "step": 267
+ },
+ {
+ "epoch": 3.726637554585153,
+ "grad_norm": 0.028330687433481216,
+ "learning_rate": 0.0006,
+ "loss": 6.669557094573975,
+ "step": 268
+ },
+ {
+ "epoch": 3.7406113537117904,
+ "grad_norm": 0.021489202976226807,
+ "learning_rate": 0.0006,
+ "loss": 6.632035732269287,
+ "step": 269
+ },
+ {
+ "epoch": 3.7545851528384278,
+ "grad_norm": 0.024269424378871918,
+ "learning_rate": 0.0006,
+ "loss": 6.660867691040039,
+ "step": 270
+ },
+ {
+ "epoch": 3.7685589519650655,
+ "grad_norm": 0.02237873338162899,
+ "learning_rate": 0.0006,
+ "loss": 6.602587699890137,
+ "step": 271
+ },
+ {
+ "epoch": 3.7825327510917033,
+ "grad_norm": 0.026503929868340492,
+ "learning_rate": 0.0006,
+ "loss": 6.599200248718262,
+ "step": 272
+ },
+ {
+ "epoch": 3.7965065502183406,
+ "grad_norm": 0.03687124326825142,
+ "learning_rate": 0.0006,
+ "loss": 6.567296504974365,
+ "step": 273
+ },
+ {
+ "epoch": 3.810480349344978,
+ "grad_norm": 0.04708952456712723,
+ "learning_rate": 0.0006,
+ "loss": 6.621265411376953,
+ "step": 274
+ },
+ {
+ "epoch": 3.8244541484716157,
+ "grad_norm": 0.041846614331007004,
+ "learning_rate": 0.0006,
+ "loss": 6.550536155700684,
+ "step": 275
+ },
+ {
+ "epoch": 3.8384279475982535,
+ "grad_norm": 0.031249675899744034,
+ "learning_rate": 0.0006,
+ "loss": 6.54278564453125,
+ "step": 276
+ },
+ {
+ "epoch": 3.8524017467248908,
+ "grad_norm": 0.023082256317138672,
+ "learning_rate": 0.0006,
+ "loss": 6.544787406921387,
+ "step": 277
+ },
+ {
+ "epoch": 3.8663755458515285,
+ "grad_norm": 0.020995570346713066,
+ "learning_rate": 0.0006,
+ "loss": 6.518143653869629,
+ "step": 278
+ },
+ {
+ "epoch": 3.880349344978166,
+ "grad_norm": 0.020839324221014977,
+ "learning_rate": 0.0006,
+ "loss": 6.546667098999023,
+ "step": 279
+ },
+ {
+ "epoch": 3.8943231441048036,
+ "grad_norm": 0.018377432599663734,
+ "learning_rate": 0.0006,
+ "loss": 6.478307723999023,
+ "step": 280
+ },
+ {
+ "epoch": 3.908296943231441,
+ "grad_norm": 0.021713724359869957,
+ "learning_rate": 0.0006,
+ "loss": 6.485574245452881,
+ "step": 281
+ },
+ {
+ "epoch": 3.9222707423580787,
+ "grad_norm": 0.01359301246702671,
+ "learning_rate": 0.0006,
+ "loss": 6.570268630981445,
+ "step": 282
+ },
+ {
+ "epoch": 3.936244541484716,
+ "grad_norm": 0.016233332455158234,
+ "learning_rate": 0.0006,
+ "loss": 6.57150936126709,
+ "step": 283
+ },
+ {
+ "epoch": 3.950218340611354,
+ "grad_norm": 0.020687608048319817,
+ "learning_rate": 0.0006,
+ "loss": 6.527956008911133,
+ "step": 284
+ },
+ {
+ "epoch": 3.964192139737991,
+ "grad_norm": 0.02032964862883091,
+ "learning_rate": 0.0006,
+ "loss": 6.468169689178467,
+ "step": 285
+ },
+ {
+ "epoch": 3.978165938864629,
+ "grad_norm": 0.016676288098096848,
+ "learning_rate": 0.0006,
+ "loss": 6.448418617248535,
+ "step": 286
+ },
+ {
+ "epoch": 3.992139737991266,
+ "grad_norm": 0.0168539360165596,
+ "learning_rate": 0.0006,
+ "loss": 6.461180686950684,
+ "step": 287
+ },
+ {
+ "epoch": 4.0,
+ "grad_norm": 0.02666233666241169,
+ "learning_rate": 0.0006,
+ "loss": 6.344893932342529,
+ "step": 288
+ },
+ {
+ "epoch": 4.0,
+ "eval_loss": 6.5384979248046875,
+ "eval_runtime": 58.814,
+ "eval_samples_per_second": 41.521,
+ "eval_steps_per_second": 1.309,
+ "step": 288
+ },
+ {
+ "epoch": 4.013973799126638,
+ "grad_norm": 0.04154708981513977,
+ "learning_rate": 0.0006,
+ "loss": 6.41855525970459,
+ "step": 289
+ },
+ {
+ "epoch": 4.0279475982532755,
+ "grad_norm": 0.04895668104290962,
+ "learning_rate": 0.0006,
+ "loss": 6.499302864074707,
+ "step": 290
+ },
+ {
+ "epoch": 4.041921397379912,
+ "grad_norm": 0.034464482218027115,
+ "learning_rate": 0.0006,
+ "loss": 6.438611030578613,
+ "step": 291
+ },
+ {
+ "epoch": 4.05589519650655,
+ "grad_norm": 0.06845773756504059,
+ "learning_rate": 0.0006,
+ "loss": 6.538361072540283,
+ "step": 292
+ },
+ {
+ "epoch": 4.069868995633188,
+ "grad_norm": 0.053768858313560486,
+ "learning_rate": 0.0006,
+ "loss": 6.527524948120117,
+ "step": 293
+ },
+ {
+ "epoch": 4.083842794759826,
+ "grad_norm": 0.039037931710481644,
+ "learning_rate": 0.0006,
+ "loss": 6.503847122192383,
+ "step": 294
+ },
+ {
+ "epoch": 4.097816593886463,
+ "grad_norm": 0.03962196782231331,
+ "learning_rate": 0.0006,
+ "loss": 6.475178241729736,
+ "step": 295
+ },
+ {
+ "epoch": 4.1117903930131,
+ "grad_norm": 0.049101535230875015,
+ "learning_rate": 0.0006,
+ "loss": 6.474587917327881,
+ "step": 296
+ },
+ {
+ "epoch": 4.125764192139738,
+ "grad_norm": 0.02830282226204872,
+ "learning_rate": 0.0006,
+ "loss": 6.45065975189209,
+ "step": 297
+ },
+ {
+ "epoch": 4.139737991266376,
+ "grad_norm": 0.023153482005000114,
+ "learning_rate": 0.0006,
+ "loss": 6.5218706130981445,
+ "step": 298
+ },
+ {
+ "epoch": 4.153711790393013,
+ "grad_norm": 0.025064002722501755,
+ "learning_rate": 0.0006,
+ "loss": 6.456475257873535,
+ "step": 299
+ },
+ {
+ "epoch": 4.1676855895196505,
+ "grad_norm": 0.02357092685997486,
+ "learning_rate": 0.0006,
+ "loss": 6.485952377319336,
+ "step": 300
+ },
+ {
+ "epoch": 4.181659388646288,
+ "grad_norm": 0.021832305938005447,
+ "learning_rate": 0.0006,
+ "loss": 6.425506591796875,
+ "step": 301
+ },
+ {
+ "epoch": 4.195633187772926,
+ "grad_norm": 0.0227900929749012,
+ "learning_rate": 0.0006,
+ "loss": 6.439155578613281,
+ "step": 302
+ },
+ {
+ "epoch": 4.209606986899563,
+ "grad_norm": 0.015961607918143272,
+ "learning_rate": 0.0006,
+ "loss": 6.460862636566162,
+ "step": 303
+ },
+ {
+ "epoch": 4.223580786026201,
+ "grad_norm": 0.017876390367746353,
+ "learning_rate": 0.0006,
+ "loss": 6.44564151763916,
+ "step": 304
+ },
+ {
+ "epoch": 4.2375545851528384,
+ "grad_norm": 0.013703204691410065,
+ "learning_rate": 0.0006,
+ "loss": 6.418414115905762,
+ "step": 305
+ },
+ {
+ "epoch": 4.251528384279476,
+ "grad_norm": 0.015240306034684181,
+ "learning_rate": 0.0006,
+ "loss": 6.411238670349121,
+ "step": 306
+ },
+ {
+ "epoch": 4.265502183406113,
+ "grad_norm": 0.01587662845849991,
+ "learning_rate": 0.0006,
+ "loss": 6.395642280578613,
+ "step": 307
+ },
+ {
+ "epoch": 4.279475982532751,
+ "grad_norm": 0.014650700613856316,
+ "learning_rate": 0.0006,
+ "loss": 6.302793025970459,
+ "step": 308
+ },
+ {
+ "epoch": 4.293449781659389,
+ "grad_norm": 0.016533225774765015,
+ "learning_rate": 0.0006,
+ "loss": 6.380588054656982,
+ "step": 309
+ },
+ {
+ "epoch": 4.307423580786026,
+ "grad_norm": 0.019963741302490234,
+ "learning_rate": 0.0006,
+ "loss": 6.345336437225342,
+ "step": 310
+ },
+ {
+ "epoch": 4.321397379912664,
+ "grad_norm": 0.02237936295568943,
+ "learning_rate": 0.0006,
+ "loss": 6.387770652770996,
+ "step": 311
+ },
+ {
+ "epoch": 4.335371179039301,
+ "grad_norm": 0.026658328250050545,
+ "learning_rate": 0.0006,
+ "loss": 6.323662757873535,
+ "step": 312
+ },
+ {
+ "epoch": 4.349344978165939,
+ "grad_norm": 0.02852284163236618,
+ "learning_rate": 0.0006,
+ "loss": 6.382204055786133,
+ "step": 313
+ },
+ {
+ "epoch": 4.3633187772925766,
+ "grad_norm": 0.028815090656280518,
+ "learning_rate": 0.0006,
+ "loss": 6.402390480041504,
+ "step": 314
+ },
+ {
+ "epoch": 4.377292576419214,
+ "grad_norm": 0.029393529519438744,
+ "learning_rate": 0.0006,
+ "loss": 6.280278205871582,
+ "step": 315
+ },
+ {
+ "epoch": 4.391266375545851,
+ "grad_norm": 0.02186041697859764,
+ "learning_rate": 0.0006,
+ "loss": 6.302707672119141,
+ "step": 316
+ },
+ {
+ "epoch": 4.405240174672489,
+ "grad_norm": 0.02071218006312847,
+ "learning_rate": 0.0006,
+ "loss": 6.3365020751953125,
+ "step": 317
+ },
+ {
+ "epoch": 4.419213973799127,
+ "grad_norm": 0.02779117226600647,
+ "learning_rate": 0.0006,
+ "loss": 6.252157211303711,
+ "step": 318
+ },
+ {
+ "epoch": 4.4331877729257645,
+ "grad_norm": 0.03470654785633087,
+ "learning_rate": 0.0006,
+ "loss": 6.319070339202881,
+ "step": 319
+ },
+ {
+ "epoch": 4.447161572052401,
+ "grad_norm": 0.037067804485559464,
+ "learning_rate": 0.0006,
+ "loss": 6.314116477966309,
+ "step": 320
+ },
+ {
+ "epoch": 4.461135371179039,
+ "grad_norm": 0.032182756811380386,
+ "learning_rate": 0.0006,
+ "loss": 6.267298698425293,
+ "step": 321
+ },
+ {
+ "epoch": 4.475109170305677,
+ "grad_norm": 0.026305217295885086,
+ "learning_rate": 0.0006,
+ "loss": 6.331688404083252,
+ "step": 322
+ },
+ {
+ "epoch": 4.489082969432315,
+ "grad_norm": 0.027228357270359993,
+ "learning_rate": 0.0006,
+ "loss": 6.31139612197876,
+ "step": 323
+ },
+ {
+ "epoch": 4.503056768558952,
+ "grad_norm": 0.025781169533729553,
+ "learning_rate": 0.0006,
+ "loss": 6.194684982299805,
+ "step": 324
+ },
+ {
+ "epoch": 4.517030567685589,
+ "grad_norm": 0.029186977073550224,
+ "learning_rate": 0.0006,
+ "loss": 6.31306266784668,
+ "step": 325
+ },
+ {
+ "epoch": 4.531004366812227,
+ "grad_norm": 0.023608777672052383,
+ "learning_rate": 0.0006,
+ "loss": 6.285243034362793,
+ "step": 326
+ },
+ {
+ "epoch": 4.544978165938865,
+ "grad_norm": 0.02267594076693058,
+ "learning_rate": 0.0006,
+ "loss": 6.30342960357666,
+ "step": 327
+ },
+ {
+ "epoch": 4.558951965065502,
+ "grad_norm": 0.01926310732960701,
+ "learning_rate": 0.0006,
+ "loss": 6.280606269836426,
+ "step": 328
+ },
+ {
+ "epoch": 4.5729257641921395,
+ "grad_norm": 0.026484837755560875,
+ "learning_rate": 0.0006,
+ "loss": 6.237980365753174,
+ "step": 329
+ },
+ {
+ "epoch": 4.586899563318777,
+ "grad_norm": 0.026228854432702065,
+ "learning_rate": 0.0006,
+ "loss": 6.3018293380737305,
+ "step": 330
+ },
+ {
+ "epoch": 4.600873362445415,
+ "grad_norm": 0.022096967324614525,
+ "learning_rate": 0.0006,
+ "loss": 6.253863334655762,
+ "step": 331
+ },
+ {
+ "epoch": 4.614847161572053,
+ "grad_norm": 0.027223890647292137,
+ "learning_rate": 0.0006,
+ "loss": 6.166882038116455,
+ "step": 332
+ },
+ {
+ "epoch": 4.62882096069869,
+ "grad_norm": 0.03160124272108078,
+ "learning_rate": 0.0006,
+ "loss": 6.230309963226318,
+ "step": 333
+ },
+ {
+ "epoch": 4.642794759825327,
+ "grad_norm": 0.04110539332032204,
+ "learning_rate": 0.0006,
+ "loss": 6.2334818840026855,
+ "step": 334
+ },
+ {
+ "epoch": 4.656768558951965,
+ "grad_norm": 0.04355933889746666,
+ "learning_rate": 0.0006,
+ "loss": 6.160956382751465,
+ "step": 335
+ },
+ {
+ "epoch": 4.670742358078603,
+ "grad_norm": 0.03611086308956146,
+ "learning_rate": 0.0006,
+ "loss": 6.23399543762207,
+ "step": 336
+ },
+ {
+ "epoch": 4.68471615720524,
+ "grad_norm": 0.051811400800943375,
+ "learning_rate": 0.0006,
+ "loss": 6.241855144500732,
+ "step": 337
+ },
+ {
+ "epoch": 4.698689956331878,
+ "grad_norm": 0.057434193789958954,
+ "learning_rate": 0.0006,
+ "loss": 6.220993995666504,
+ "step": 338
+ },
+ {
+ "epoch": 4.712663755458515,
+ "grad_norm": 0.04420953616499901,
+ "learning_rate": 0.0006,
+ "loss": 6.300461292266846,
+ "step": 339
+ },
+ {
+ "epoch": 4.726637554585153,
+ "grad_norm": 0.03515457361936569,
+ "learning_rate": 0.0006,
+ "loss": 6.269349098205566,
+ "step": 340
+ },
+ {
+ "epoch": 4.74061135371179,
+ "grad_norm": 0.03222600743174553,
+ "learning_rate": 0.0006,
+ "loss": 6.248613357543945,
+ "step": 341
+ },
+ {
+ "epoch": 4.754585152838428,
+ "grad_norm": 0.0384046845138073,
+ "learning_rate": 0.0006,
+ "loss": 6.266979217529297,
+ "step": 342
+ },
+ {
+ "epoch": 4.7685589519650655,
+ "grad_norm": 0.0514867827296257,
+ "learning_rate": 0.0006,
+ "loss": 6.234103202819824,
+ "step": 343
+ },
+ {
+ "epoch": 4.782532751091703,
+ "grad_norm": 0.050747793167829514,
+ "learning_rate": 0.0006,
+ "loss": 6.322582244873047,
+ "step": 344
+ },
+ {
+ "epoch": 4.796506550218341,
+ "grad_norm": 0.040878988802433014,
+ "learning_rate": 0.0006,
+ "loss": 6.212509632110596,
+ "step": 345
+ },
+ {
+ "epoch": 4.810480349344978,
+ "grad_norm": 0.034655820578336716,
+ "learning_rate": 0.0006,
+ "loss": 6.265879154205322,
+ "step": 346
+ },
+ {
+ "epoch": 4.824454148471616,
+ "grad_norm": 0.04202224686741829,
+ "learning_rate": 0.0006,
+ "loss": 6.2896409034729,
+ "step": 347
+ },
+ {
+ "epoch": 4.8384279475982535,
+ "grad_norm": 0.03632277995347977,
+ "learning_rate": 0.0006,
+ "loss": 6.253917217254639,
+ "step": 348
+ },
+ {
+ "epoch": 4.85240174672489,
+ "grad_norm": 0.028645765036344528,
+ "learning_rate": 0.0006,
+ "loss": 6.207965850830078,
+ "step": 349
+ },
+ {
+ "epoch": 4.866375545851528,
+ "grad_norm": 0.025305170565843582,
+ "learning_rate": 0.0006,
+ "loss": 6.168068885803223,
+ "step": 350
+ },
+ {
+ "epoch": 4.880349344978166,
+ "grad_norm": 0.022137803956866264,
+ "learning_rate": 0.0006,
+ "loss": 6.235283374786377,
+ "step": 351
+ },
+ {
+ "epoch": 4.894323144104804,
+ "grad_norm": 0.01814538985490799,
+ "learning_rate": 0.0006,
+ "loss": 6.259641647338867,
+ "step": 352
+ },
+ {
+ "epoch": 4.908296943231441,
+ "grad_norm": 0.018465086817741394,
+ "learning_rate": 0.0006,
+ "loss": 6.111738681793213,
+ "step": 353
+ },
+ {
+ "epoch": 4.922270742358078,
+ "grad_norm": 0.018638933077454567,
+ "learning_rate": 0.0006,
+ "loss": 6.225642204284668,
+ "step": 354
+ },
+ {
+ "epoch": 4.936244541484716,
+ "grad_norm": 0.016461260616779327,
+ "learning_rate": 0.0006,
+ "loss": 6.140647888183594,
+ "step": 355
+ },
+ {
+ "epoch": 4.950218340611354,
+ "grad_norm": 0.015458385460078716,
+ "learning_rate": 0.0006,
+ "loss": 6.155224800109863,
+ "step": 356
+ },
+ {
+ "epoch": 4.964192139737992,
+ "grad_norm": 0.014934048056602478,
+ "learning_rate": 0.0006,
+ "loss": 6.101555824279785,
+ "step": 357
+ },
+ {
+ "epoch": 4.978165938864628,
+ "grad_norm": 0.013979545794427395,
+ "learning_rate": 0.0006,
+ "loss": 6.161718845367432,
+ "step": 358
+ },
+ {
+ "epoch": 4.992139737991266,
+ "grad_norm": 0.01640394888818264,
+ "learning_rate": 0.0006,
+ "loss": 6.055768966674805,
+ "step": 359
+ },
+ {
+ "epoch": 5.0,
+ "grad_norm": 0.015135680325329304,
+ "learning_rate": 0.0006,
+ "loss": 6.169695854187012,
+ "step": 360
+ },
+ {
+ "epoch": 5.0,
+ "eval_loss": 6.16408109664917,
+ "eval_runtime": 59.3649,
+ "eval_samples_per_second": 41.135,
+ "eval_steps_per_second": 1.297,
+ "step": 360
+ },
+ {
+ "epoch": 5.013973799126638,
+ "grad_norm": 0.012849048711359501,
+ "learning_rate": 0.0006,
+ "loss": 6.055830001831055,
+ "step": 361
+ },
+ {
+ "epoch": 5.0279475982532755,
+ "grad_norm": 0.018169576302170753,
+ "learning_rate": 0.0006,
+ "loss": 6.082655429840088,
+ "step": 362
+ },
+ {
+ "epoch": 5.041921397379912,
+ "grad_norm": 0.024293430149555206,
+ "learning_rate": 0.0006,
+ "loss": 6.109032154083252,
+ "step": 363
+ },
+ {
+ "epoch": 5.05589519650655,
+ "grad_norm": 0.03266787901520729,
+ "learning_rate": 0.0006,
+ "loss": 6.030613899230957,
+ "step": 364
+ },
+ {
+ "epoch": 5.069868995633188,
+ "grad_norm": 0.03572266176342964,
+ "learning_rate": 0.0006,
+ "loss": 6.050390720367432,
+ "step": 365
+ },
+ {
+ "epoch": 5.083842794759826,
+ "grad_norm": 0.032625213265419006,
+ "learning_rate": 0.0006,
+ "loss": 6.169313430786133,
+ "step": 366
+ },
+ {
+ "epoch": 5.097816593886463,
+ "grad_norm": 0.04051872715353966,
+ "learning_rate": 0.0006,
+ "loss": 6.10398006439209,
+ "step": 367
+ },
+ {
+ "epoch": 5.1117903930131,
+ "grad_norm": 0.0351213738322258,
+ "learning_rate": 0.0006,
+ "loss": 6.121973514556885,
+ "step": 368
+ },
+ {
+ "epoch": 5.125764192139738,
+ "grad_norm": 0.03523759916424751,
+ "learning_rate": 0.0006,
+ "loss": 6.0626654624938965,
+ "step": 369
+ },
+ {
+ "epoch": 5.139737991266376,
+ "grad_norm": 0.02706182189285755,
+ "learning_rate": 0.0006,
+ "loss": 6.049354076385498,
+ "step": 370
+ },
+ {
+ "epoch": 5.153711790393013,
+ "grad_norm": 0.024212589487433434,
+ "learning_rate": 0.0006,
+ "loss": 6.034826278686523,
+ "step": 371
+ },
+ {
+ "epoch": 5.1676855895196505,
+ "grad_norm": 0.024798082187771797,
+ "learning_rate": 0.0006,
+ "loss": 6.004058837890625,
+ "step": 372
+ },
+ {
+ "epoch": 5.181659388646288,
+ "grad_norm": 0.02325567416846752,
+ "learning_rate": 0.0006,
+ "loss": 5.986461639404297,
+ "step": 373
+ },
+ {
+ "epoch": 5.195633187772926,
+ "grad_norm": 0.019060570746660233,
+ "learning_rate": 0.0006,
+ "loss": 6.059736251831055,
+ "step": 374
+ },
+ {
+ "epoch": 5.209606986899563,
+ "grad_norm": 0.016822976991534233,
+ "learning_rate": 0.0006,
+ "loss": 6.072957515716553,
+ "step": 375
+ },
+ {
+ "epoch": 5.223580786026201,
+ "grad_norm": 0.017218658700585365,
+ "learning_rate": 0.0006,
+ "loss": 6.031739234924316,
+ "step": 376
+ },
+ {
+ "epoch": 5.2375545851528384,
+ "grad_norm": 0.014889383688569069,
+ "learning_rate": 0.0006,
+ "loss": 6.025674343109131,
+ "step": 377
+ },
+ {
+ "epoch": 5.251528384279476,
+ "grad_norm": 0.01708107627928257,
+ "learning_rate": 0.0006,
+ "loss": 6.090451240539551,
+ "step": 378
+ },
+ {
+ "epoch": 5.265502183406113,
+ "grad_norm": 0.020260317251086235,
+ "learning_rate": 0.0006,
+ "loss": 6.094257354736328,
+ "step": 379
+ },
+ {
+ "epoch": 5.279475982532751,
+ "grad_norm": 0.020110400393605232,
+ "learning_rate": 0.0006,
+ "loss": 6.021188735961914,
+ "step": 380
+ },
+ {
+ "epoch": 5.293449781659389,
+ "grad_norm": 0.02027830481529236,
+ "learning_rate": 0.0006,
+ "loss": 6.022156238555908,
+ "step": 381
+ },
+ {
+ "epoch": 5.307423580786026,
+ "grad_norm": 0.022747062146663666,
+ "learning_rate": 0.0006,
+ "loss": 6.011836051940918,
+ "step": 382
+ },
+ {
+ "epoch": 5.321397379912664,
+ "grad_norm": 0.01990230567753315,
+ "learning_rate": 0.0006,
+ "loss": 6.003054618835449,
+ "step": 383
+ },
+ {
+ "epoch": 5.335371179039301,
+ "grad_norm": 0.01596238650381565,
+ "learning_rate": 0.0006,
+ "loss": 5.991410732269287,
+ "step": 384
+ },
+ {
+ "epoch": 5.349344978165939,
+ "grad_norm": 0.015847105532884598,
+ "learning_rate": 0.0006,
+ "loss": 6.100622177124023,
+ "step": 385
+ },
+ {
+ "epoch": 5.3633187772925766,
+ "grad_norm": 0.016179397702217102,
+ "learning_rate": 0.0006,
+ "loss": 6.032659530639648,
+ "step": 386
+ },
+ {
+ "epoch": 5.377292576419214,
+ "grad_norm": 0.018256602808833122,
+ "learning_rate": 0.0006,
+ "loss": 5.961983680725098,
+ "step": 387
+ },
+ {
+ "epoch": 5.391266375545851,
+ "grad_norm": 0.02805912122130394,
+ "learning_rate": 0.0006,
+ "loss": 5.983541965484619,
+ "step": 388
+ },
+ {
+ "epoch": 5.405240174672489,
+ "grad_norm": 0.039082255214452744,
+ "learning_rate": 0.0006,
+ "loss": 6.021474361419678,
+ "step": 389
+ },
+ {
+ "epoch": 5.419213973799127,
+ "grad_norm": 0.036757659167051315,
+ "learning_rate": 0.0006,
+ "loss": 6.005046367645264,
+ "step": 390
+ },
+ {
+ "epoch": 5.4331877729257645,
+ "grad_norm": 0.035277366638183594,
+ "learning_rate": 0.0006,
+ "loss": 6.036296844482422,
+ "step": 391
+ },
+ {
+ "epoch": 5.447161572052401,
+ "grad_norm": 0.034404635429382324,
+ "learning_rate": 0.0006,
+ "loss": 5.994539260864258,
+ "step": 392
+ },
+ {
+ "epoch": 5.461135371179039,
+ "grad_norm": 0.0377206988632679,
+ "learning_rate": 0.0006,
+ "loss": 5.941534519195557,
+ "step": 393
+ },
+ {
+ "epoch": 5.475109170305677,
+ "grad_norm": 0.0389922633767128,
+ "learning_rate": 0.0006,
+ "loss": 5.983644485473633,
+ "step": 394
+ },
+ {
+ "epoch": 5.489082969432315,
+ "grad_norm": 0.04176778718829155,
+ "learning_rate": 0.0006,
+ "loss": 6.030922889709473,
+ "step": 395
+ },
+ {
+ "epoch": 5.503056768558952,
+ "grad_norm": 0.0337153784930706,
+ "learning_rate": 0.0006,
+ "loss": 5.961367130279541,
+ "step": 396
+ },
+ {
+ "epoch": 5.517030567685589,
+ "grad_norm": 0.036116816103458405,
+ "learning_rate": 0.0006,
+ "loss": 5.819280624389648,
+ "step": 397
+ },
+ {
+ "epoch": 5.531004366812227,
+ "grad_norm": 0.030724041163921356,
+ "learning_rate": 0.0006,
+ "loss": 5.906380653381348,
+ "step": 398
+ },
+ {
+ "epoch": 5.544978165938865,
+ "grad_norm": 0.030264202505350113,
+ "learning_rate": 0.0006,
+ "loss": 5.975833415985107,
+ "step": 399
+ },
+ {
+ "epoch": 5.558951965065502,
+ "grad_norm": 0.032096318900585175,
+ "learning_rate": 0.0006,
+ "loss": 6.001348972320557,
+ "step": 400
+ },
+ {
+ "epoch": 5.5729257641921395,
+ "grad_norm": 0.030579356476664543,
+ "learning_rate": 0.0006,
+ "loss": 5.928768634796143,
+ "step": 401
+ },
+ {
+ "epoch": 5.586899563318777,
+ "grad_norm": 0.028241973370313644,
+ "learning_rate": 0.0006,
+ "loss": 5.92582893371582,
+ "step": 402
+ },
+ {
+ "epoch": 5.600873362445415,
+ "grad_norm": 0.023086953908205032,
+ "learning_rate": 0.0006,
+ "loss": 5.892926216125488,
+ "step": 403
+ },
+ {
+ "epoch": 5.614847161572053,
+ "grad_norm": 0.02547992393374443,
+ "learning_rate": 0.0006,
+ "loss": 5.903195381164551,
+ "step": 404
+ },
+ {
+ "epoch": 5.62882096069869,
+ "grad_norm": 0.023089831694960594,
+ "learning_rate": 0.0006,
+ "loss": 5.912033557891846,
+ "step": 405
+ },
+ {
+ "epoch": 5.642794759825327,
+ "grad_norm": 0.025446368381381035,
+ "learning_rate": 0.0006,
+ "loss": 5.926138401031494,
+ "step": 406
+ },
+ {
+ "epoch": 5.656768558951965,
+ "grad_norm": 0.03748747706413269,
+ "learning_rate": 0.0006,
+ "loss": 5.913451194763184,
+ "step": 407
+ },
+ {
+ "epoch": 5.670742358078603,
+ "grad_norm": 0.03493810072541237,
+ "learning_rate": 0.0006,
+ "loss": 5.97633695602417,
+ "step": 408
+ },
+ {
+ "epoch": 5.68471615720524,
+ "grad_norm": 0.023131275549530983,
+ "learning_rate": 0.0006,
+ "loss": 5.899571418762207,
+ "step": 409
+ },
+ {
+ "epoch": 5.698689956331878,
+ "grad_norm": 0.02986014075577259,
+ "learning_rate": 0.0006,
+ "loss": 5.894903182983398,
+ "step": 410
+ },
+ {
+ "epoch": 5.712663755458515,
+ "grad_norm": 0.030171144753694534,
+ "learning_rate": 0.0006,
+ "loss": 5.876595497131348,
+ "step": 411
+ },
+ {
+ "epoch": 5.726637554585153,
+ "grad_norm": 0.029377546161413193,
+ "learning_rate": 0.0006,
+ "loss": 5.911637306213379,
+ "step": 412
+ },
+ {
+ "epoch": 5.74061135371179,
+ "grad_norm": 0.029829490929841995,
+ "learning_rate": 0.0006,
+ "loss": 5.819629669189453,
+ "step": 413
+ },
+ {
+ "epoch": 5.754585152838428,
+ "grad_norm": 0.024139299988746643,
+ "learning_rate": 0.0006,
+ "loss": 5.81728458404541,
+ "step": 414
+ },
+ {
+ "epoch": 5.7685589519650655,
+ "grad_norm": 0.021453432738780975,
+ "learning_rate": 0.0006,
+ "loss": 5.987326145172119,
+ "step": 415
+ },
+ {
+ "epoch": 5.782532751091703,
+ "grad_norm": 0.017942246049642563,
+ "learning_rate": 0.0006,
+ "loss": 5.893008232116699,
+ "step": 416
+ },
+ {
+ "epoch": 5.796506550218341,
+ "grad_norm": 0.019724637269973755,
+ "learning_rate": 0.0006,
+ "loss": 5.912441253662109,
+ "step": 417
+ },
+ {
+ "epoch": 5.810480349344978,
+ "grad_norm": 0.018597912043333054,
+ "learning_rate": 0.0006,
+ "loss": 5.860394477844238,
+ "step": 418
+ },
+ {
+ "epoch": 5.824454148471616,
+ "grad_norm": 0.016232412308454514,
+ "learning_rate": 0.0006,
+ "loss": 5.94991397857666,
+ "step": 419
+ },
+ {
+ "epoch": 5.8384279475982535,
+ "grad_norm": 0.01667204685509205,
+ "learning_rate": 0.0006,
+ "loss": 5.8506364822387695,
+ "step": 420
+ },
+ {
+ "epoch": 5.85240174672489,
+ "grad_norm": 0.014232151210308075,
+ "learning_rate": 0.0006,
+ "loss": 5.928424835205078,
+ "step": 421
+ },
+ {
+ "epoch": 5.866375545851528,
+ "grad_norm": 0.01570476032793522,
+ "learning_rate": 0.0006,
+ "loss": 5.953692436218262,
+ "step": 422
+ },
+ {
+ "epoch": 5.880349344978166,
+ "grad_norm": 0.01586179807782173,
+ "learning_rate": 0.0006,
+ "loss": 5.820014476776123,
+ "step": 423
+ },
+ {
+ "epoch": 5.894323144104804,
+ "grad_norm": 0.014729145914316177,
+ "learning_rate": 0.0006,
+ "loss": 5.860154628753662,
+ "step": 424
+ },
+ {
+ "epoch": 5.908296943231441,
+ "grad_norm": 0.013111459091305733,
+ "learning_rate": 0.0006,
+ "loss": 5.85988187789917,
+ "step": 425
+ },
+ {
+ "epoch": 5.922270742358078,
+ "grad_norm": 0.012937922962009907,
+ "learning_rate": 0.0006,
+ "loss": 5.756265640258789,
+ "step": 426
+ },
+ {
+ "epoch": 5.936244541484716,
+ "grad_norm": 0.01377453189343214,
+ "learning_rate": 0.0006,
+ "loss": 5.8333048820495605,
+ "step": 427
+ },
+ {
+ "epoch": 5.950218340611354,
+ "grad_norm": 0.014054795727133751,
+ "learning_rate": 0.0006,
+ "loss": 5.803333759307861,
+ "step": 428
+ },
+ {
+ "epoch": 5.964192139737992,
+ "grad_norm": 0.01674959622323513,
+ "learning_rate": 0.0006,
+ "loss": 5.8718485832214355,
+ "step": 429
+ },
+ {
+ "epoch": 5.978165938864628,
+ "grad_norm": 0.018733017146587372,
+ "learning_rate": 0.0006,
+ "loss": 5.875979900360107,
+ "step": 430
+ },
+ {
+ "epoch": 5.992139737991266,
+ "grad_norm": 0.02088983915746212,
+ "learning_rate": 0.0006,
+ "loss": 5.763634204864502,
+ "step": 431
+ },
+ {
+ "epoch": 6.0,
+ "grad_norm": 0.026705985888838768,
+ "learning_rate": 0.0006,
+ "loss": 5.6922478675842285,
+ "step": 432
+ },
+ {
+ "epoch": 6.0,
+ "eval_loss": 5.847958087921143,
+ "eval_runtime": 58.5451,
+ "eval_samples_per_second": 41.711,
+ "eval_steps_per_second": 1.315,
+ "step": 432
+ },
+ {
+ "epoch": 6.013973799126638,
+ "grad_norm": 0.03181544691324234,
+ "learning_rate": 0.0006,
+ "loss": 5.786368370056152,
+ "step": 433
+ },
+ {
+ "epoch": 6.0279475982532755,
+ "grad_norm": 0.03238112851977348,
+ "learning_rate": 0.0006,
+ "loss": 5.810310363769531,
+ "step": 434
+ },
+ {
+ "epoch": 6.041921397379912,
+ "grad_norm": 0.03916756808757782,
+ "learning_rate": 0.0006,
+ "loss": 5.7761993408203125,
+ "step": 435
+ },
+ {
+ "epoch": 6.05589519650655,
+ "grad_norm": 0.044009730219841,
+ "learning_rate": 0.0006,
+ "loss": 5.898112773895264,
+ "step": 436
+ },
+ {
+ "epoch": 6.069868995633188,
+ "grad_norm": 0.045235246419906616,
+ "learning_rate": 0.0006,
+ "loss": 5.876371383666992,
+ "step": 437
+ },
+ {
+ "epoch": 6.083842794759826,
+ "grad_norm": 0.050218384712934494,
+ "learning_rate": 0.0006,
+ "loss": 5.805103302001953,
+ "step": 438
+ },
+ {
+ "epoch": 6.097816593886463,
+ "grad_norm": 0.05444490164518356,
+ "learning_rate": 0.0006,
+ "loss": 5.826424598693848,
+ "step": 439
+ },
+ {
+ "epoch": 6.1117903930131,
+ "grad_norm": 0.04890982061624527,
+ "learning_rate": 0.0006,
+ "loss": 5.7735395431518555,
+ "step": 440
+ },
+ {
+ "epoch": 6.125764192139738,
+ "grad_norm": 0.05478639155626297,
+ "learning_rate": 0.0006,
+ "loss": 5.823677062988281,
+ "step": 441
+ },
+ {
+ "epoch": 6.139737991266376,
+ "grad_norm": 0.057562313973903656,
+ "learning_rate": 0.0006,
+ "loss": 5.889334678649902,
+ "step": 442
+ },
+ {
+ "epoch": 6.153711790393013,
+ "grad_norm": 0.06447703391313553,
+ "learning_rate": 0.0006,
+ "loss": 5.84621524810791,
+ "step": 443
+ },
+ {
+ "epoch": 6.1676855895196505,
+ "grad_norm": 0.048861872404813766,
+ "learning_rate": 0.0006,
+ "loss": 5.931595802307129,
+ "step": 444
+ },
+ {
+ "epoch": 6.181659388646288,
+ "grad_norm": 0.05521339178085327,
+ "learning_rate": 0.0006,
+ "loss": 5.8229780197143555,
+ "step": 445
+ },
+ {
+ "epoch": 6.195633187772926,
+ "grad_norm": 0.053666990250349045,
+ "learning_rate": 0.0006,
+ "loss": 5.879191875457764,
+ "step": 446
+ },
+ {
+ "epoch": 6.209606986899563,
+ "grad_norm": 0.0451025515794754,
+ "learning_rate": 0.0006,
+ "loss": 5.801628112792969,
+ "step": 447
+ },
+ {
+ "epoch": 6.223580786026201,
+ "grad_norm": 0.04291655868291855,
+ "learning_rate": 0.0006,
+ "loss": 5.914680480957031,
+ "step": 448
+ },
+ {
+ "epoch": 6.2375545851528384,
+ "grad_norm": 0.035102490335702896,
+ "learning_rate": 0.0006,
+ "loss": 5.843682765960693,
+ "step": 449
+ },
+ {
+ "epoch": 6.251528384279476,
+ "grad_norm": 0.03403995931148529,
+ "learning_rate": 0.0006,
+ "loss": 5.827154159545898,
+ "step": 450
+ },
+ {
+ "epoch": 6.265502183406113,
+ "grad_norm": 0.03444375470280647,
+ "learning_rate": 0.0006,
+ "loss": 5.800136566162109,
+ "step": 451
+ },
+ {
+ "epoch": 6.279475982532751,
+ "grad_norm": 0.03165159001946449,
+ "learning_rate": 0.0006,
+ "loss": 5.906252384185791,
+ "step": 452
+ },
+ {
+ "epoch": 6.293449781659389,
+ "grad_norm": 0.026153093203902245,
+ "learning_rate": 0.0006,
+ "loss": 5.7423200607299805,
+ "step": 453
+ },
+ {
+ "epoch": 6.307423580786026,
+ "grad_norm": 0.024357657879590988,
+ "learning_rate": 0.0006,
+ "loss": 5.701364517211914,
+ "step": 454
+ },
+ {
+ "epoch": 6.321397379912664,
+ "grad_norm": 0.021508049219846725,
+ "learning_rate": 0.0006,
+ "loss": 5.8328094482421875,
+ "step": 455
+ },
+ {
+ "epoch": 6.335371179039301,
+ "grad_norm": 0.017313921824097633,
+ "learning_rate": 0.0006,
+ "loss": 5.792145252227783,
+ "step": 456
+ },
+ {
+ "epoch": 6.349344978165939,
+ "grad_norm": 0.018563397228717804,
+ "learning_rate": 0.0006,
+ "loss": 5.732672214508057,
+ "step": 457
+ },
+ {
+ "epoch": 6.3633187772925766,
+ "grad_norm": 0.016568679362535477,
+ "learning_rate": 0.0006,
+ "loss": 5.81948184967041,
+ "step": 458
+ },
+ {
+ "epoch": 6.377292576419214,
+ "grad_norm": 0.014133770950138569,
+ "learning_rate": 0.0006,
+ "loss": 5.734982490539551,
+ "step": 459
+ },
+ {
+ "epoch": 6.391266375545851,
+ "grad_norm": 0.014114447869360447,
+ "learning_rate": 0.0006,
+ "loss": 5.70073127746582,
+ "step": 460
+ },
+ {
+ "epoch": 6.405240174672489,
+ "grad_norm": 0.012907393276691437,
+ "learning_rate": 0.0006,
+ "loss": 5.758626937866211,
+ "step": 461
+ },
+ {
+ "epoch": 6.419213973799127,
+ "grad_norm": 0.012673444114625454,
+ "learning_rate": 0.0006,
+ "loss": 5.775138854980469,
+ "step": 462
+ },
+ {
+ "epoch": 6.4331877729257645,
+ "grad_norm": 0.01311410591006279,
+ "learning_rate": 0.0006,
+ "loss": 5.70920991897583,
+ "step": 463
+ },
+ {
+ "epoch": 6.447161572052401,
+ "grad_norm": 0.012935544364154339,
+ "learning_rate": 0.0006,
+ "loss": 5.781479835510254,
+ "step": 464
+ },
+ {
+ "epoch": 6.461135371179039,
+ "grad_norm": 0.01439738366752863,
+ "learning_rate": 0.0006,
+ "loss": 5.691315650939941,
+ "step": 465
+ },
+ {
+ "epoch": 6.475109170305677,
+ "grad_norm": 0.01205186452716589,
+ "learning_rate": 0.0006,
+ "loss": 5.662499904632568,
+ "step": 466
+ },
+ {
+ "epoch": 6.489082969432315,
+ "grad_norm": 0.011575652286410332,
+ "learning_rate": 0.0006,
+ "loss": 5.6271772384643555,
+ "step": 467
+ },
+ {
+ "epoch": 6.503056768558952,
+ "grad_norm": 0.011672502383589745,
+ "learning_rate": 0.0006,
+ "loss": 5.761662006378174,
+ "step": 468
+ },
+ {
+ "epoch": 6.517030567685589,
+ "grad_norm": 0.011541751213371754,
+ "learning_rate": 0.0006,
+ "loss": 5.762078285217285,
+ "step": 469
+ },
+ {
+ "epoch": 6.531004366812227,
+ "grad_norm": 0.011396365240216255,
+ "learning_rate": 0.0006,
+ "loss": 5.67873477935791,
+ "step": 470
+ },
+ {
+ "epoch": 6.544978165938865,
+ "grad_norm": 0.010688499547541142,
+ "learning_rate": 0.0006,
+ "loss": 5.717051982879639,
+ "step": 471
+ },
+ {
+ "epoch": 6.558951965065502,
+ "grad_norm": 0.012224317528307438,
+ "learning_rate": 0.0006,
+ "loss": 5.707948684692383,
+ "step": 472
+ },
+ {
+ "epoch": 6.5729257641921395,
+ "grad_norm": 0.011856825090944767,
+ "learning_rate": 0.0006,
+ "loss": 5.70878791809082,
+ "step": 473
+ },
+ {
+ "epoch": 6.586899563318777,
+ "grad_norm": 0.01199064590036869,
+ "learning_rate": 0.0006,
+ "loss": 5.708697319030762,
+ "step": 474
+ },
+ {
+ "epoch": 6.600873362445415,
+ "grad_norm": 0.01219446212053299,
+ "learning_rate": 0.0006,
+ "loss": 5.6061577796936035,
+ "step": 475
+ },
+ {
+ "epoch": 6.614847161572053,
+ "grad_norm": 0.013397484086453915,
+ "learning_rate": 0.0006,
+ "loss": 5.624789714813232,
+ "step": 476
+ },
+ {
+ "epoch": 6.62882096069869,
+ "grad_norm": 0.01483136136084795,
+ "learning_rate": 0.0006,
+ "loss": 5.721141338348389,
+ "step": 477
+ },
+ {
+ "epoch": 6.642794759825327,
+ "grad_norm": 0.0194488987326622,
+ "learning_rate": 0.0006,
+ "loss": 5.604279518127441,
+ "step": 478
+ },
+ {
+ "epoch": 6.656768558951965,
+ "grad_norm": 0.02079896256327629,
+ "learning_rate": 0.0006,
+ "loss": 5.586322784423828,
+ "step": 479
+ },
+ {
+ "epoch": 6.670742358078603,
+ "grad_norm": 0.018996229395270348,
+ "learning_rate": 0.0006,
+ "loss": 5.5987772941589355,
+ "step": 480
+ },
+ {
+ "epoch": 6.68471615720524,
+ "grad_norm": 0.015779603272676468,
+ "learning_rate": 0.0006,
+ "loss": 5.588602066040039,
+ "step": 481
+ },
+ {
+ "epoch": 6.698689956331878,
+ "grad_norm": 0.015322973020374775,
+ "learning_rate": 0.0006,
+ "loss": 5.686467170715332,
+ "step": 482
+ },
+ {
+ "epoch": 6.712663755458515,
+ "grad_norm": 0.015282213687896729,
+ "learning_rate": 0.0006,
+ "loss": 5.664676666259766,
+ "step": 483
+ },
+ {
+ "epoch": 6.726637554585153,
+ "grad_norm": 0.016589272767305374,
+ "learning_rate": 0.0006,
+ "loss": 5.666739463806152,
+ "step": 484
+ },
+ {
+ "epoch": 6.74061135371179,
+ "grad_norm": 0.01811421848833561,
+ "learning_rate": 0.0006,
+ "loss": 5.647593021392822,
+ "step": 485
+ },
+ {
+ "epoch": 6.754585152838428,
+ "grad_norm": 0.019652029499411583,
+ "learning_rate": 0.0006,
+ "loss": 5.628726959228516,
+ "step": 486
+ },
+ {
+ "epoch": 6.7685589519650655,
+ "grad_norm": 0.02118665538728237,
+ "learning_rate": 0.0006,
+ "loss": 5.6580705642700195,
+ "step": 487
+ },
+ {
+ "epoch": 6.782532751091703,
+ "grad_norm": 0.02237832546234131,
+ "learning_rate": 0.0006,
+ "loss": 5.635931968688965,
+ "step": 488
+ },
+ {
+ "epoch": 6.796506550218341,
+ "grad_norm": 0.023897631093859673,
+ "learning_rate": 0.0006,
+ "loss": 5.517949104309082,
+ "step": 489
+ },
+ {
+ "epoch": 6.810480349344978,
+ "grad_norm": 0.02442227490246296,
+ "learning_rate": 0.0006,
+ "loss": 5.600021839141846,
+ "step": 490
+ },
+ {
+ "epoch": 6.824454148471616,
+ "grad_norm": 0.024675287306308746,
+ "learning_rate": 0.0006,
+ "loss": 5.6639909744262695,
+ "step": 491
+ },
+ {
+ "epoch": 6.8384279475982535,
+ "grad_norm": 0.030372392386198044,
+ "learning_rate": 0.0006,
+ "loss": 5.582888603210449,
+ "step": 492
+ },
+ {
+ "epoch": 6.85240174672489,
+ "grad_norm": 0.03237690031528473,
+ "learning_rate": 0.0006,
+ "loss": 5.573896884918213,
+ "step": 493
+ },
+ {
+ "epoch": 6.866375545851528,
+ "grad_norm": 0.03511589393019676,
+ "learning_rate": 0.0006,
+ "loss": 5.565878868103027,
+ "step": 494
+ },
+ {
+ "epoch": 6.880349344978166,
+ "grad_norm": 0.03883938118815422,
+ "learning_rate": 0.0006,
+ "loss": 5.58540678024292,
+ "step": 495
+ },
+ {
+ "epoch": 6.894323144104804,
+ "grad_norm": 0.04175502806901932,
+ "learning_rate": 0.0006,
+ "loss": 5.6095194816589355,
+ "step": 496
+ },
+ {
+ "epoch": 6.908296943231441,
+ "grad_norm": 0.04013441503047943,
+ "learning_rate": 0.0006,
+ "loss": 5.654401779174805,
+ "step": 497
+ },
+ {
+ "epoch": 6.922270742358078,
+ "grad_norm": 0.04618770629167557,
+ "learning_rate": 0.0006,
+ "loss": 5.6483473777771,
+ "step": 498
+ },
+ {
+ "epoch": 6.936244541484716,
+ "grad_norm": 0.04507308453321457,
+ "learning_rate": 0.0006,
+ "loss": 5.5856218338012695,
+ "step": 499
+ },
+ {
+ "epoch": 6.950218340611354,
+ "grad_norm": 0.03151383996009827,
+ "learning_rate": 0.0006,
+ "loss": 5.594086647033691,
+ "step": 500
+ },
+ {
+ "epoch": 6.964192139737992,
+ "grad_norm": 0.026883797720074654,
+ "learning_rate": 0.0006,
+ "loss": 5.607676029205322,
+ "step": 501
+ },
+ {
+ "epoch": 6.978165938864628,
+ "grad_norm": 0.02981836162507534,
+ "learning_rate": 0.0006,
+ "loss": 5.563666820526123,
+ "step": 502
+ },
+ {
+ "epoch": 6.992139737991266,
+ "grad_norm": 0.03233477845788002,
+ "learning_rate": 0.0006,
+ "loss": 5.667543411254883,
+ "step": 503
+ },
+ {
+ "epoch": 7.0,
+ "grad_norm": 0.029927456751465797,
+ "learning_rate": 0.0006,
+ "loss": 5.733482837677002,
+ "step": 504
+ },
+ {
+ "epoch": 7.0,
+ "eval_loss": 5.679076671600342,
+ "eval_runtime": 59.5965,
+ "eval_samples_per_second": 40.976,
+ "eval_steps_per_second": 1.292,
+ "step": 504
+ },
+ {
+ "epoch": 7.013973799126638,
+ "grad_norm": 0.027081597596406937,
+ "learning_rate": 0.0006,
+ "loss": 5.634487152099609,
+ "step": 505
+ },
+ {
+ "epoch": 7.0279475982532755,
+ "grad_norm": 0.031959421932697296,
+ "learning_rate": 0.0006,
+ "loss": 5.643294811248779,
+ "step": 506
+ },
+ {
+ "epoch": 7.041921397379912,
+ "grad_norm": 0.02802063524723053,
+ "learning_rate": 0.0006,
+ "loss": 5.573239326477051,
+ "step": 507
+ },
+ {
+ "epoch": 7.05589519650655,
+ "grad_norm": 0.032579705119132996,
+ "learning_rate": 0.0006,
+ "loss": 5.627150058746338,
+ "step": 508
+ },
+ {
+ "epoch": 7.069868995633188,
+ "grad_norm": 0.02503928542137146,
+ "learning_rate": 0.0006,
+ "loss": 5.6045308113098145,
+ "step": 509
+ },
+ {
+ "epoch": 7.083842794759826,
+ "grad_norm": 0.02421317622065544,
+ "learning_rate": 0.0006,
+ "loss": 5.522153377532959,
+ "step": 510
+ },
+ {
+ "epoch": 7.097816593886463,
+ "grad_norm": 0.021983426064252853,
+ "learning_rate": 0.0006,
+ "loss": 5.632939338684082,
+ "step": 511
+ },
+ {
+ "epoch": 7.1117903930131,
+ "grad_norm": 0.021933183073997498,
+ "learning_rate": 0.0006,
+ "loss": 5.592185974121094,
+ "step": 512
+ },
+ {
+ "epoch": 7.125764192139738,
+ "grad_norm": 0.02287031151354313,
+ "learning_rate": 0.0006,
+ "loss": 5.613700866699219,
+ "step": 513
+ },
+ {
+ "epoch": 7.139737991266376,
+ "grad_norm": 0.02281602844595909,
+ "learning_rate": 0.0006,
+ "loss": 5.551383018493652,
+ "step": 514
+ },
+ {
+ "epoch": 7.153711790393013,
+ "grad_norm": 0.021050360053777695,
+ "learning_rate": 0.0006,
+ "loss": 5.55275821685791,
+ "step": 515
+ },
+ {
+ "epoch": 7.1676855895196505,
+ "grad_norm": 0.018299926072359085,
+ "learning_rate": 0.0006,
+ "loss": 5.528225421905518,
+ "step": 516
+ },
+ {
+ "epoch": 7.181659388646288,
+ "grad_norm": 0.02024853602051735,
+ "learning_rate": 0.0006,
+ "loss": 5.522453784942627,
+ "step": 517
+ },
+ {
+ "epoch": 7.195633187772926,
+ "grad_norm": 0.02135239914059639,
+ "learning_rate": 0.0006,
+ "loss": 5.527522563934326,
+ "step": 518
+ },
+ {
+ "epoch": 7.209606986899563,
+ "grad_norm": 0.023394184187054634,
+ "learning_rate": 0.0006,
+ "loss": 5.4535017013549805,
+ "step": 519
+ },
+ {
+ "epoch": 7.223580786026201,
+ "grad_norm": 0.025952080264687538,
+ "learning_rate": 0.0006,
+ "loss": 5.5374674797058105,
+ "step": 520
+ },
+ {
+ "epoch": 7.2375545851528384,
+ "grad_norm": 0.022287718951702118,
+ "learning_rate": 0.0006,
+ "loss": 5.493753433227539,
+ "step": 521
+ },
+ {
+ "epoch": 7.251528384279476,
+ "grad_norm": 0.02018447406589985,
+ "learning_rate": 0.0006,
+ "loss": 5.510575294494629,
+ "step": 522
+ },
+ {
+ "epoch": 7.265502183406113,
+ "grad_norm": 0.02050507813692093,
+ "learning_rate": 0.0006,
+ "loss": 5.466026782989502,
+ "step": 523
+ },
+ {
+ "epoch": 7.279475982532751,
+ "grad_norm": 0.023688802495598793,
+ "learning_rate": 0.0006,
+ "loss": 5.485815525054932,
+ "step": 524
+ },
+ {
+ "epoch": 7.293449781659389,
+ "grad_norm": 0.021078843623399734,
+ "learning_rate": 0.0006,
+ "loss": 5.415581703186035,
+ "step": 525
+ },
+ {
+ "epoch": 7.307423580786026,
+ "grad_norm": 0.018266011029481888,
+ "learning_rate": 0.0006,
+ "loss": 5.402815818786621,
+ "step": 526
+ },
+ {
+ "epoch": 7.321397379912664,
+ "grad_norm": 0.019693493843078613,
+ "learning_rate": 0.0006,
+ "loss": 5.504674911499023,
+ "step": 527
+ },
+ {
+ "epoch": 7.335371179039301,
+ "grad_norm": 0.02031373605132103,
+ "learning_rate": 0.0006,
+ "loss": 5.503837585449219,
+ "step": 528
+ },
+ {
+ "epoch": 7.349344978165939,
+ "grad_norm": 0.017636626958847046,
+ "learning_rate": 0.0006,
+ "loss": 5.530580997467041,
+ "step": 529
+ },
+ {
+ "epoch": 7.3633187772925766,
+ "grad_norm": 0.01787244901061058,
+ "learning_rate": 0.0006,
+ "loss": 5.450218677520752,
+ "step": 530
+ },
+ {
+ "epoch": 7.377292576419214,
+ "grad_norm": 0.0170669574290514,
+ "learning_rate": 0.0006,
+ "loss": 5.481570243835449,
+ "step": 531
+ },
+ {
+ "epoch": 7.391266375545851,
+ "grad_norm": 0.01802165061235428,
+ "learning_rate": 0.0006,
+ "loss": 5.386394500732422,
+ "step": 532
+ },
+ {
+ "epoch": 7.405240174672489,
+ "grad_norm": 0.022949472069740295,
+ "learning_rate": 0.0006,
+ "loss": 5.424929618835449,
+ "step": 533
+ },
+ {
+ "epoch": 7.419213973799127,
+ "grad_norm": 0.034213390201330185,
+ "learning_rate": 0.0006,
+ "loss": 5.346663475036621,
+ "step": 534
+ },
+ {
+ "epoch": 7.4331877729257645,
+ "grad_norm": 0.03847593814134598,
+ "learning_rate": 0.0006,
+ "loss": 5.418482780456543,
+ "step": 535
+ },
+ {
+ "epoch": 7.447161572052401,
+ "grad_norm": 0.029393676668405533,
+ "learning_rate": 0.0006,
+ "loss": 5.456090927124023,
+ "step": 536
+ },
+ {
+ "epoch": 7.461135371179039,
+ "grad_norm": 0.03388667106628418,
+ "learning_rate": 0.0006,
+ "loss": 5.572983741760254,
+ "step": 537
+ },
+ {
+ "epoch": 7.475109170305677,
+ "grad_norm": 0.039690613746643066,
+ "learning_rate": 0.0006,
+ "loss": 5.414067268371582,
+ "step": 538
+ },
+ {
+ "epoch": 7.489082969432315,
+ "grad_norm": 0.03630776330828667,
+ "learning_rate": 0.0006,
+ "loss": 5.523739814758301,
+ "step": 539
+ },
+ {
+ "epoch": 7.503056768558952,
+ "grad_norm": 0.03356870263814926,
+ "learning_rate": 0.0006,
+ "loss": 5.444767951965332,
+ "step": 540
+ },
+ {
+ "epoch": 7.517030567685589,
+ "grad_norm": 0.030812304466962814,
+ "learning_rate": 0.0006,
+ "loss": 5.461244106292725,
+ "step": 541
+ },
+ {
+ "epoch": 7.531004366812227,
+ "grad_norm": 0.03319217637181282,
+ "learning_rate": 0.0006,
+ "loss": 5.502161026000977,
+ "step": 542
+ },
+ {
+ "epoch": 7.544978165938865,
+ "grad_norm": 0.032997481524944305,
+ "learning_rate": 0.0006,
+ "loss": 5.4162278175354,
+ "step": 543
+ },
+ {
+ "epoch": 7.558951965065502,
+ "grad_norm": 0.03364962339401245,
+ "learning_rate": 0.0006,
+ "loss": 5.415736198425293,
+ "step": 544
+ },
+ {
+ "epoch": 7.5729257641921395,
+ "grad_norm": 0.0344221405684948,
+ "learning_rate": 0.0006,
+ "loss": 5.541967868804932,
+ "step": 545
+ },
+ {
+ "epoch": 7.586899563318777,
+ "grad_norm": 0.029609503224492073,
+ "learning_rate": 0.0006,
+ "loss": 5.440614223480225,
+ "step": 546
+ },
+ {
+ "epoch": 7.600873362445415,
+ "grad_norm": 0.029055926948785782,
+ "learning_rate": 0.0006,
+ "loss": 5.462865352630615,
+ "step": 547
+ },
+ {
+ "epoch": 7.614847161572053,
+ "grad_norm": 0.02658848837018013,
+ "learning_rate": 0.0006,
+ "loss": 5.444467544555664,
+ "step": 548
+ },
+ {
+ "epoch": 7.62882096069869,
+ "grad_norm": 0.026276404038071632,
+ "learning_rate": 0.0006,
+ "loss": 5.462278842926025,
+ "step": 549
+ },
+ {
+ "epoch": 7.642794759825327,
+ "grad_norm": 0.0286889486014843,
+ "learning_rate": 0.0006,
+ "loss": 5.4849534034729,
+ "step": 550
+ },
+ {
+ "epoch": 7.656768558951965,
+ "grad_norm": 0.026849817484617233,
+ "learning_rate": 0.0006,
+ "loss": 5.431473731994629,
+ "step": 551
+ },
+ {
+ "epoch": 7.670742358078603,
+ "grad_norm": 0.02312396466732025,
+ "learning_rate": 0.0006,
+ "loss": 5.45462703704834,
+ "step": 552
+ },
+ {
+ "epoch": 7.68471615720524,
+ "grad_norm": 0.026197485625743866,
+ "learning_rate": 0.0006,
+ "loss": 5.452552318572998,
+ "step": 553
+ },
+ {
+ "epoch": 7.698689956331878,
+ "grad_norm": 0.02747255191206932,
+ "learning_rate": 0.0006,
+ "loss": 5.447169780731201,
+ "step": 554
+ },
+ {
+ "epoch": 7.712663755458515,
+ "grad_norm": 0.028123432770371437,
+ "learning_rate": 0.0006,
+ "loss": 5.394400596618652,
+ "step": 555
+ },
+ {
+ "epoch": 7.726637554585153,
+ "grad_norm": 0.02599870041012764,
+ "learning_rate": 0.0006,
+ "loss": 5.398341178894043,
+ "step": 556
+ },
+ {
+ "epoch": 7.74061135371179,
+ "grad_norm": 0.022171657532453537,
+ "learning_rate": 0.0006,
+ "loss": 5.368820667266846,
+ "step": 557
+ },
+ {
+ "epoch": 7.754585152838428,
+ "grad_norm": 0.022309480234980583,
+ "learning_rate": 0.0006,
+ "loss": 5.416772365570068,
+ "step": 558
+ },
+ {
+ "epoch": 7.7685589519650655,
+ "grad_norm": 0.024074165150523186,
+ "learning_rate": 0.0006,
+ "loss": 5.359002113342285,
+ "step": 559
+ },
+ {
+ "epoch": 7.782532751091703,
+ "grad_norm": 0.02636653557419777,
+ "learning_rate": 0.0006,
+ "loss": 5.484368324279785,
+ "step": 560
+ },
+ {
+ "epoch": 7.796506550218341,
+ "grad_norm": 0.02203752100467682,
+ "learning_rate": 0.0006,
+ "loss": 5.375498294830322,
+ "step": 561
+ },
+ {
+ "epoch": 7.810480349344978,
+ "grad_norm": 0.019924819469451904,
+ "learning_rate": 0.0006,
+ "loss": 5.334723472595215,
+ "step": 562
+ },
+ {
+ "epoch": 7.824454148471616,
+ "grad_norm": 0.018755966797471046,
+ "learning_rate": 0.0006,
+ "loss": 5.348781108856201,
+ "step": 563
+ },
+ {
+ "epoch": 7.8384279475982535,
+ "grad_norm": 0.016747845336794853,
+ "learning_rate": 0.0006,
+ "loss": 5.433079719543457,
+ "step": 564
+ },
+ {
+ "epoch": 7.85240174672489,
+ "grad_norm": 0.0170395877212286,
+ "learning_rate": 0.0006,
+ "loss": 5.346794128417969,
+ "step": 565
+ },
+ {
+ "epoch": 7.866375545851528,
+ "grad_norm": 0.016074176877737045,
+ "learning_rate": 0.0006,
+ "loss": 5.3277740478515625,
+ "step": 566
+ },
+ {
+ "epoch": 7.880349344978166,
+ "grad_norm": 0.014119806699454784,
+ "learning_rate": 0.0006,
+ "loss": 5.483551979064941,
+ "step": 567
+ },
+ {
+ "epoch": 7.894323144104804,
+ "grad_norm": 0.015270394273102283,
+ "learning_rate": 0.0006,
+ "loss": 5.378519535064697,
+ "step": 568
+ },
+ {
+ "epoch": 7.908296943231441,
+ "grad_norm": 0.015677539631724358,
+ "learning_rate": 0.0006,
+ "loss": 5.290210247039795,
+ "step": 569
+ },
+ {
+ "epoch": 7.922270742358078,
+ "grad_norm": 0.015930423513054848,
+ "learning_rate": 0.0006,
+ "loss": 5.427360534667969,
+ "step": 570
+ },
+ {
+ "epoch": 7.936244541484716,
+ "grad_norm": 0.016641564667224884,
+ "learning_rate": 0.0006,
+ "loss": 5.301599979400635,
+ "step": 571
+ },
+ {
+ "epoch": 7.950218340611354,
+ "grad_norm": 0.018293552100658417,
+ "learning_rate": 0.0006,
+ "loss": 5.277889251708984,
+ "step": 572
+ },
+ {
+ "epoch": 7.964192139737992,
+ "grad_norm": 0.017618799582123756,
+ "learning_rate": 0.0006,
+ "loss": 5.297194480895996,
+ "step": 573
+ },
+ {
+ "epoch": 7.978165938864628,
+ "grad_norm": 0.015034242533147335,
+ "learning_rate": 0.0006,
+ "loss": 5.42428731918335,
+ "step": 574
+ },
+ {
+ "epoch": 7.992139737991266,
+ "grad_norm": 0.01617511734366417,
+ "learning_rate": 0.0006,
+ "loss": 5.420044898986816,
+ "step": 575
+ },
+ {
+ "epoch": 8.0,
+ "grad_norm": 0.01837257295846939,
+ "learning_rate": 0.0006,
+ "loss": 5.240387916564941,
+ "step": 576
+ },
+ {
+ "epoch": 8.0,
+ "eval_loss": 5.423932075500488,
+ "eval_runtime": 58.0841,
+ "eval_samples_per_second": 42.042,
+ "eval_steps_per_second": 1.326,
+ "step": 576
+ },
+ {
+ "epoch": 8.013973799126637,
+ "grad_norm": 0.020953809842467308,
+ "learning_rate": 0.0006,
+ "loss": 5.4320783615112305,
+ "step": 577
+ },
+ {
+ "epoch": 8.027947598253276,
+ "grad_norm": 0.0227745920419693,
+ "learning_rate": 0.0006,
+ "loss": 5.310197830200195,
+ "step": 578
+ },
+ {
+ "epoch": 8.041921397379912,
+ "grad_norm": 0.021372603252530098,
+ "learning_rate": 0.0006,
+ "loss": 5.339468002319336,
+ "step": 579
+ },
+ {
+ "epoch": 8.055895196506551,
+ "grad_norm": 0.022401731461286545,
+ "learning_rate": 0.0006,
+ "loss": 5.279829978942871,
+ "step": 580
+ },
+ {
+ "epoch": 8.069868995633188,
+ "grad_norm": 0.023942379280924797,
+ "learning_rate": 0.0006,
+ "loss": 5.372323989868164,
+ "step": 581
+ },
+ {
+ "epoch": 8.083842794759825,
+ "grad_norm": 0.024380534887313843,
+ "learning_rate": 0.0006,
+ "loss": 5.253486633300781,
+ "step": 582
+ },
+ {
+ "epoch": 8.097816593886463,
+ "grad_norm": 0.03246625140309334,
+ "learning_rate": 0.0006,
+ "loss": 5.305774211883545,
+ "step": 583
+ },
+ {
+ "epoch": 8.1117903930131,
+ "grad_norm": 0.030399736016988754,
+ "learning_rate": 0.0006,
+ "loss": 5.370976448059082,
+ "step": 584
+ },
+ {
+ "epoch": 8.125764192139737,
+ "grad_norm": 0.027148913592100143,
+ "learning_rate": 0.0006,
+ "loss": 5.368010520935059,
+ "step": 585
+ },
+ {
+ "epoch": 8.139737991266376,
+ "grad_norm": 0.02925540879368782,
+ "learning_rate": 0.0006,
+ "loss": 5.303013801574707,
+ "step": 586
+ },
+ {
+ "epoch": 8.153711790393013,
+ "grad_norm": 0.02915453538298607,
+ "learning_rate": 0.0006,
+ "loss": 5.329667091369629,
+ "step": 587
+ },
+ {
+ "epoch": 8.167685589519651,
+ "grad_norm": 0.031732626259326935,
+ "learning_rate": 0.0006,
+ "loss": 5.3634934425354,
+ "step": 588
+ },
+ {
+ "epoch": 8.181659388646288,
+ "grad_norm": 0.036166731268167496,
+ "learning_rate": 0.0006,
+ "loss": 5.223980903625488,
+ "step": 589
+ },
+ {
+ "epoch": 8.195633187772925,
+ "grad_norm": 0.035654108971357346,
+ "learning_rate": 0.0006,
+ "loss": 5.361059188842773,
+ "step": 590
+ },
+ {
+ "epoch": 8.209606986899564,
+ "grad_norm": 0.0365324430167675,
+ "learning_rate": 0.0006,
+ "loss": 5.336151599884033,
+ "step": 591
+ },
+ {
+ "epoch": 8.2235807860262,
+ "grad_norm": 0.03578880429267883,
+ "learning_rate": 0.0006,
+ "loss": 5.434076309204102,
+ "step": 592
+ },
+ {
+ "epoch": 8.237554585152838,
+ "grad_norm": 0.035734012722969055,
+ "learning_rate": 0.0006,
+ "loss": 5.330893516540527,
+ "step": 593
+ },
+ {
+ "epoch": 8.251528384279476,
+ "grad_norm": 0.03196857124567032,
+ "learning_rate": 0.0006,
+ "loss": 5.306009769439697,
+ "step": 594
+ },
+ {
+ "epoch": 8.265502183406113,
+ "grad_norm": 0.032042086124420166,
+ "learning_rate": 0.0006,
+ "loss": 5.343267917633057,
+ "step": 595
+ },
+ {
+ "epoch": 8.279475982532752,
+ "grad_norm": 0.03160746395587921,
+ "learning_rate": 0.0006,
+ "loss": 5.2353715896606445,
+ "step": 596
+ },
+ {
+ "epoch": 8.293449781659389,
+ "grad_norm": 0.026689818128943443,
+ "learning_rate": 0.0006,
+ "loss": 5.254042625427246,
+ "step": 597
+ },
+ {
+ "epoch": 8.307423580786025,
+ "grad_norm": 0.02880188450217247,
+ "learning_rate": 0.0006,
+ "loss": 5.326833248138428,
+ "step": 598
+ },
+ {
+ "epoch": 8.321397379912664,
+ "grad_norm": 0.027516381815075874,
+ "learning_rate": 0.0006,
+ "loss": 5.342336177825928,
+ "step": 599
+ },
+ {
+ "epoch": 8.335371179039301,
+ "grad_norm": 0.027875030413269997,
+ "learning_rate": 0.0006,
+ "loss": 5.200719833374023,
+ "step": 600
+ },
+ {
+ "epoch": 8.34934497816594,
+ "grad_norm": 0.0268265288323164,
+ "learning_rate": 0.0006,
+ "loss": 5.249449729919434,
+ "step": 601
+ },
+ {
+ "epoch": 8.363318777292577,
+ "grad_norm": 0.024824608117341995,
+ "learning_rate": 0.0006,
+ "loss": 5.318952560424805,
+ "step": 602
+ },
+ {
+ "epoch": 8.377292576419213,
+ "grad_norm": 0.022990627214312553,
+ "learning_rate": 0.0006,
+ "loss": 5.244993209838867,
+ "step": 603
+ },
+ {
+ "epoch": 8.391266375545852,
+ "grad_norm": 0.022804604843258858,
+ "learning_rate": 0.0006,
+ "loss": 5.3236846923828125,
+ "step": 604
+ },
+ {
+ "epoch": 8.405240174672489,
+ "grad_norm": 0.02419872209429741,
+ "learning_rate": 0.0006,
+ "loss": 5.255486488342285,
+ "step": 605
+ },
+ {
+ "epoch": 8.419213973799126,
+ "grad_norm": 0.021952059119939804,
+ "learning_rate": 0.0006,
+ "loss": 5.3511552810668945,
+ "step": 606
+ },
+ {
+ "epoch": 8.433187772925764,
+ "grad_norm": 0.022375497967004776,
+ "learning_rate": 0.0006,
+ "loss": 5.294790267944336,
+ "step": 607
+ },
+ {
+ "epoch": 8.447161572052401,
+ "grad_norm": 0.019242815673351288,
+ "learning_rate": 0.0006,
+ "loss": 5.252618789672852,
+ "step": 608
+ },
+ {
+ "epoch": 8.46113537117904,
+ "grad_norm": 0.0173486340790987,
+ "learning_rate": 0.0006,
+ "loss": 5.30147647857666,
+ "step": 609
+ },
+ {
+ "epoch": 8.475109170305677,
+ "grad_norm": 0.01612604409456253,
+ "learning_rate": 0.0006,
+ "loss": 5.23960542678833,
+ "step": 610
+ },
+ {
+ "epoch": 8.489082969432314,
+ "grad_norm": 0.014293976128101349,
+ "learning_rate": 0.0006,
+ "loss": 5.119932651519775,
+ "step": 611
+ },
+ {
+ "epoch": 8.503056768558952,
+ "grad_norm": 0.014453536830842495,
+ "learning_rate": 0.0006,
+ "loss": 5.166794776916504,
+ "step": 612
+ },
+ {
+ "epoch": 8.51703056768559,
+ "grad_norm": 0.014685769565403461,
+ "learning_rate": 0.0006,
+ "loss": 5.264822959899902,
+ "step": 613
+ },
+ {
+ "epoch": 8.531004366812226,
+ "grad_norm": 0.01427740603685379,
+ "learning_rate": 0.0006,
+ "loss": 5.278433799743652,
+ "step": 614
+ },
+ {
+ "epoch": 8.544978165938865,
+ "grad_norm": 0.014864951372146606,
+ "learning_rate": 0.0006,
+ "loss": 5.214512348175049,
+ "step": 615
+ },
+ {
+ "epoch": 8.558951965065502,
+ "grad_norm": 0.01565164513885975,
+ "learning_rate": 0.0006,
+ "loss": 5.183066368103027,
+ "step": 616
+ },
+ {
+ "epoch": 8.57292576419214,
+ "grad_norm": 0.01779816672205925,
+ "learning_rate": 0.0006,
+ "loss": 5.163185119628906,
+ "step": 617
+ },
+ {
+ "epoch": 8.586899563318777,
+ "grad_norm": 0.017254827544093132,
+ "learning_rate": 0.0006,
+ "loss": 5.243555068969727,
+ "step": 618
+ },
+ {
+ "epoch": 8.600873362445414,
+ "grad_norm": 0.01650584116578102,
+ "learning_rate": 0.0006,
+ "loss": 5.234884738922119,
+ "step": 619
+ },
+ {
+ "epoch": 8.614847161572053,
+ "grad_norm": 0.017021115869283676,
+ "learning_rate": 0.0006,
+ "loss": 5.12271785736084,
+ "step": 620
+ },
+ {
+ "epoch": 8.62882096069869,
+ "grad_norm": 0.01773759163916111,
+ "learning_rate": 0.0006,
+ "loss": 5.242153167724609,
+ "step": 621
+ },
+ {
+ "epoch": 8.642794759825328,
+ "grad_norm": 0.015679042786359787,
+ "learning_rate": 0.0006,
+ "loss": 5.170779228210449,
+ "step": 622
+ },
+ {
+ "epoch": 8.656768558951965,
+ "grad_norm": 0.013760549947619438,
+ "learning_rate": 0.0006,
+ "loss": 5.238644599914551,
+ "step": 623
+ },
+ {
+ "epoch": 8.670742358078602,
+ "grad_norm": 0.014571522362530231,
+ "learning_rate": 0.0006,
+ "loss": 5.177056312561035,
+ "step": 624
+ },
+ {
+ "epoch": 8.68471615720524,
+ "grad_norm": 0.016209015622735023,
+ "learning_rate": 0.0006,
+ "loss": 5.163750648498535,
+ "step": 625
+ },
+ {
+ "epoch": 8.698689956331878,
+ "grad_norm": 0.016813941299915314,
+ "learning_rate": 0.0006,
+ "loss": 5.183428764343262,
+ "step": 626
+ },
+ {
+ "epoch": 8.712663755458514,
+ "grad_norm": 0.020985357463359833,
+ "learning_rate": 0.0006,
+ "loss": 5.229465007781982,
+ "step": 627
+ },
+ {
+ "epoch": 8.726637554585153,
+ "grad_norm": 0.02679632417857647,
+ "learning_rate": 0.0006,
+ "loss": 5.257368087768555,
+ "step": 628
+ },
+ {
+ "epoch": 8.74061135371179,
+ "grad_norm": 0.025756070390343666,
+ "learning_rate": 0.0006,
+ "loss": 5.253736972808838,
+ "step": 629
+ },
+ {
+ "epoch": 8.754585152838429,
+ "grad_norm": 0.02643490768969059,
+ "learning_rate": 0.0006,
+ "loss": 5.228633403778076,
+ "step": 630
+ },
+ {
+ "epoch": 8.768558951965066,
+ "grad_norm": 0.03303210437297821,
+ "learning_rate": 0.0006,
+ "loss": 5.163540363311768,
+ "step": 631
+ },
+ {
+ "epoch": 8.782532751091702,
+ "grad_norm": 0.030432431027293205,
+ "learning_rate": 0.0006,
+ "loss": 5.243169784545898,
+ "step": 632
+ },
+ {
+ "epoch": 8.796506550218341,
+ "grad_norm": 0.028631288558244705,
+ "learning_rate": 0.0006,
+ "loss": 5.166398048400879,
+ "step": 633
+ },
+ {
+ "epoch": 8.810480349344978,
+ "grad_norm": 0.026188310235738754,
+ "learning_rate": 0.0006,
+ "loss": 5.128122329711914,
+ "step": 634
+ },
+ {
+ "epoch": 8.824454148471617,
+ "grad_norm": 0.025308528915047646,
+ "learning_rate": 0.0006,
+ "loss": 5.211284637451172,
+ "step": 635
+ },
+ {
+ "epoch": 8.838427947598253,
+ "grad_norm": 0.020657729357481003,
+ "learning_rate": 0.0006,
+ "loss": 5.161575794219971,
+ "step": 636
+ },
+ {
+ "epoch": 8.85240174672489,
+ "grad_norm": 0.021901234984397888,
+ "learning_rate": 0.0006,
+ "loss": 5.20050048828125,
+ "step": 637
+ },
+ {
+ "epoch": 8.866375545851529,
+ "grad_norm": 0.021219318732619286,
+ "learning_rate": 0.0006,
+ "loss": 5.211699485778809,
+ "step": 638
+ },
+ {
+ "epoch": 8.880349344978166,
+ "grad_norm": 0.0200974028557539,
+ "learning_rate": 0.0006,
+ "loss": 5.092503547668457,
+ "step": 639
+ },
+ {
+ "epoch": 8.894323144104803,
+ "grad_norm": 0.023804882541298866,
+ "learning_rate": 0.0006,
+ "loss": 5.216068267822266,
+ "step": 640
+ },
+ {
+ "epoch": 8.908296943231441,
+ "grad_norm": 0.026088010519742966,
+ "learning_rate": 0.0006,
+ "loss": 5.155592918395996,
+ "step": 641
+ },
+ {
+ "epoch": 8.922270742358078,
+ "grad_norm": 0.02601276896893978,
+ "learning_rate": 0.0006,
+ "loss": 5.197238922119141,
+ "step": 642
+ },
+ {
+ "epoch": 8.936244541484717,
+ "grad_norm": 0.020387211814522743,
+ "learning_rate": 0.0006,
+ "loss": 5.1955437660217285,
+ "step": 643
+ },
+ {
+ "epoch": 8.950218340611354,
+ "grad_norm": 0.019214622676372528,
+ "learning_rate": 0.0006,
+ "loss": 5.206346035003662,
+ "step": 644
+ },
+ {
+ "epoch": 8.96419213973799,
+ "grad_norm": 0.019674314185976982,
+ "learning_rate": 0.0006,
+ "loss": 5.219857215881348,
+ "step": 645
+ },
+ {
+ "epoch": 8.97816593886463,
+ "grad_norm": 0.020318234339356422,
+ "learning_rate": 0.0006,
+ "loss": 5.191132545471191,
+ "step": 646
+ },
+ {
+ "epoch": 8.992139737991266,
+ "grad_norm": 0.021428676322102547,
+ "learning_rate": 0.0006,
+ "loss": 5.214695930480957,
+ "step": 647
+ },
+ {
+ "epoch": 9.0,
+ "grad_norm": 0.023235054686665535,
+ "learning_rate": 0.0006,
+ "loss": 5.229091644287109,
+ "step": 648
+ },
+ {
+ "epoch": 9.0,
+ "eval_loss": 5.289593696594238,
+ "eval_runtime": 56.9256,
+ "eval_samples_per_second": 42.898,
+ "eval_steps_per_second": 1.353,
+ "step": 648
+ },
+ {
+ "epoch": 9.013973799126637,
+ "grad_norm": 0.02320289984345436,
+ "learning_rate": 0.0006,
+ "loss": 4.983669281005859,
+ "step": 649
+ },
+ {
+ "epoch": 9.027947598253276,
+ "grad_norm": 0.018521282821893692,
+ "learning_rate": 0.0006,
+ "loss": 5.200575828552246,
+ "step": 650
+ },
+ {
+ "epoch": 9.041921397379912,
+ "grad_norm": 0.0180169939994812,
+ "learning_rate": 0.0006,
+ "loss": 5.148033618927002,
+ "step": 651
+ },
+ {
+ "epoch": 9.055895196506551,
+ "grad_norm": 0.019969860091805458,
+ "learning_rate": 0.0006,
+ "loss": 5.099125385284424,
+ "step": 652
+ },
+ {
+ "epoch": 9.069868995633188,
+ "grad_norm": 0.01643305830657482,
+ "learning_rate": 0.0006,
+ "loss": 5.092347145080566,
+ "step": 653
+ },
+ {
+ "epoch": 9.083842794759825,
+ "grad_norm": 0.016022363677620888,
+ "learning_rate": 0.0006,
+ "loss": 5.129279613494873,
+ "step": 654
+ },
+ {
+ "epoch": 9.097816593886463,
+ "grad_norm": 0.01657041162252426,
+ "learning_rate": 0.0006,
+ "loss": 5.159562587738037,
+ "step": 655
+ },
+ {
+ "epoch": 9.1117903930131,
+ "grad_norm": 0.017924228683114052,
+ "learning_rate": 0.0006,
+ "loss": 5.165497779846191,
+ "step": 656
+ },
+ {
+ "epoch": 9.125764192139737,
+ "grad_norm": 0.021592361852526665,
+ "learning_rate": 0.0006,
+ "loss": 5.204775810241699,
+ "step": 657
+ },
+ {
+ "epoch": 9.139737991266376,
+ "grad_norm": 0.026924120262265205,
+ "learning_rate": 0.0006,
+ "loss": 5.128296852111816,
+ "step": 658
+ },
+ {
+ "epoch": 9.153711790393013,
+ "grad_norm": 0.023156899958848953,
+ "learning_rate": 0.0006,
+ "loss": 5.1255340576171875,
+ "step": 659
+ },
+ {
+ "epoch": 9.167685589519651,
+ "grad_norm": 0.020670367404818535,
+ "learning_rate": 0.0006,
+ "loss": 5.128479957580566,
+ "step": 660
+ },
+ {
+ "epoch": 9.181659388646288,
+ "grad_norm": 0.025447173044085503,
+ "learning_rate": 0.0006,
+ "loss": 5.154821872711182,
+ "step": 661
+ },
+ {
+ "epoch": 9.195633187772925,
+ "grad_norm": 0.027940068393945694,
+ "learning_rate": 0.0006,
+ "loss": 5.143466472625732,
+ "step": 662
+ },
+ {
+ "epoch": 9.209606986899564,
+ "grad_norm": 0.026847844943404198,
+ "learning_rate": 0.0006,
+ "loss": 5.045665740966797,
+ "step": 663
+ },
+ {
+ "epoch": 9.2235807860262,
+ "grad_norm": 0.02340601570904255,
+ "learning_rate": 0.0006,
+ "loss": 5.21933126449585,
+ "step": 664
+ },
+ {
+ "epoch": 9.237554585152838,
+ "grad_norm": 0.02340371161699295,
+ "learning_rate": 0.0006,
+ "loss": 4.942949295043945,
+ "step": 665
+ },
+ {
+ "epoch": 9.251528384279476,
+ "grad_norm": 0.02221992425620556,
+ "learning_rate": 0.0006,
+ "loss": 5.197099208831787,
+ "step": 666
+ },
+ {
+ "epoch": 9.265502183406113,
+ "grad_norm": 0.023508677259087563,
+ "learning_rate": 0.0006,
+ "loss": 5.022897720336914,
+ "step": 667
+ },
+ {
+ "epoch": 9.279475982532752,
+ "grad_norm": 0.026120394468307495,
+ "learning_rate": 0.0006,
+ "loss": 5.087725639343262,
+ "step": 668
+ },
+ {
+ "epoch": 9.293449781659389,
+ "grad_norm": 0.027273228392004967,
+ "learning_rate": 0.0006,
+ "loss": 5.090963840484619,
+ "step": 669
+ },
+ {
+ "epoch": 9.307423580786025,
+ "grad_norm": 0.03241586685180664,
+ "learning_rate": 0.0006,
+ "loss": 5.047842979431152,
+ "step": 670
+ },
+ {
+ "epoch": 9.321397379912664,
+ "grad_norm": 0.030245667323470116,
+ "learning_rate": 0.0006,
+ "loss": 5.104803085327148,
+ "step": 671
+ },
+ {
+ "epoch": 9.335371179039301,
+ "grad_norm": 0.027698364108800888,
+ "learning_rate": 0.0006,
+ "loss": 5.0644097328186035,
+ "step": 672
+ },
+ {
+ "epoch": 9.34934497816594,
+ "grad_norm": 0.029692554846405983,
+ "learning_rate": 0.0006,
+ "loss": 5.111942291259766,
+ "step": 673
+ },
+ {
+ "epoch": 9.363318777292577,
+ "grad_norm": 0.03328656405210495,
+ "learning_rate": 0.0006,
+ "loss": 5.1574554443359375,
+ "step": 674
+ },
+ {
+ "epoch": 9.377292576419213,
+ "grad_norm": 0.031938180327415466,
+ "learning_rate": 0.0006,
+ "loss": 5.190371990203857,
+ "step": 675
+ },
+ {
+ "epoch": 9.391266375545852,
+ "grad_norm": 0.024858945980668068,
+ "learning_rate": 0.0006,
+ "loss": 5.121420860290527,
+ "step": 676
+ },
+ {
+ "epoch": 9.405240174672489,
+ "grad_norm": 0.023033203557133675,
+ "learning_rate": 0.0006,
+ "loss": 5.0974345207214355,
+ "step": 677
+ },
+ {
+ "epoch": 9.419213973799126,
+ "grad_norm": 0.022083545103669167,
+ "learning_rate": 0.0006,
+ "loss": 5.046319961547852,
+ "step": 678
+ },
+ {
+ "epoch": 9.433187772925764,
+ "grad_norm": 0.0228437427431345,
+ "learning_rate": 0.0006,
+ "loss": 5.119963645935059,
+ "step": 679
+ },
+ {
+ "epoch": 9.447161572052401,
+ "grad_norm": 0.0242351982742548,
+ "learning_rate": 0.0006,
+ "loss": 5.0444560050964355,
+ "step": 680
+ },
+ {
+ "epoch": 9.46113537117904,
+ "grad_norm": 0.02401185780763626,
+ "learning_rate": 0.0006,
+ "loss": 5.063238620758057,
+ "step": 681
+ },
+ {
+ "epoch": 9.475109170305677,
+ "grad_norm": 0.02560959942638874,
+ "learning_rate": 0.0006,
+ "loss": 5.035224914550781,
+ "step": 682
+ },
+ {
+ "epoch": 9.489082969432314,
+ "grad_norm": 0.024985626339912415,
+ "learning_rate": 0.0006,
+ "loss": 5.049284934997559,
+ "step": 683
+ },
+ {
+ "epoch": 9.503056768558952,
+ "grad_norm": 0.023178115487098694,
+ "learning_rate": 0.0006,
+ "loss": 5.2266621589660645,
+ "step": 684
+ },
+ {
+ "epoch": 9.51703056768559,
+ "grad_norm": 0.0215512253344059,
+ "learning_rate": 0.0006,
+ "loss": 5.1875715255737305,
+ "step": 685
+ },
+ {
+ "epoch": 9.531004366812226,
+ "grad_norm": 0.01662837713956833,
+ "learning_rate": 0.0006,
+ "loss": 5.137012481689453,
+ "step": 686
+ },
+ {
+ "epoch": 9.544978165938865,
+ "grad_norm": 0.01952764391899109,
+ "learning_rate": 0.0006,
+ "loss": 5.072500228881836,
+ "step": 687
+ },
+ {
+ "epoch": 9.558951965065502,
+ "grad_norm": 0.017521077767014503,
+ "learning_rate": 0.0006,
+ "loss": 5.045711994171143,
+ "step": 688
+ },
+ {
+ "epoch": 9.57292576419214,
+ "grad_norm": 0.017355265095829964,
+ "learning_rate": 0.0006,
+ "loss": 5.02881383895874,
+ "step": 689
+ },
+ {
+ "epoch": 9.586899563318777,
+ "grad_norm": 0.015587719157338142,
+ "learning_rate": 0.0006,
+ "loss": 5.132787704467773,
+ "step": 690
+ },
+ {
+ "epoch": 9.600873362445414,
+ "grad_norm": 0.017469845712184906,
+ "learning_rate": 0.0006,
+ "loss": 5.105447292327881,
+ "step": 691
+ },
+ {
+ "epoch": 9.614847161572053,
+ "grad_norm": 0.019630100578069687,
+ "learning_rate": 0.0006,
+ "loss": 5.032464981079102,
+ "step": 692
+ },
+ {
+ "epoch": 9.62882096069869,
+ "grad_norm": 0.01784994639456272,
+ "learning_rate": 0.0006,
+ "loss": 5.081345558166504,
+ "step": 693
+ },
+ {
+ "epoch": 9.642794759825328,
+ "grad_norm": 0.01872754842042923,
+ "learning_rate": 0.0006,
+ "loss": 4.9772114753723145,
+ "step": 694
+ },
+ {
+ "epoch": 9.656768558951965,
+ "grad_norm": 0.01999032311141491,
+ "learning_rate": 0.0006,
+ "loss": 5.089740753173828,
+ "step": 695
+ },
+ {
+ "epoch": 9.670742358078602,
+ "grad_norm": 0.01710767112672329,
+ "learning_rate": 0.0006,
+ "loss": 4.9869890213012695,
+ "step": 696
+ },
+ {
+ "epoch": 9.68471615720524,
+ "grad_norm": 0.015144161880016327,
+ "learning_rate": 0.0006,
+ "loss": 5.022564888000488,
+ "step": 697
+ },
+ {
+ "epoch": 9.698689956331878,
+ "grad_norm": 0.01618543453514576,
+ "learning_rate": 0.0006,
+ "loss": 5.035921573638916,
+ "step": 698
+ },
+ {
+ "epoch": 9.712663755458514,
+ "grad_norm": 0.014653387479484081,
+ "learning_rate": 0.0006,
+ "loss": 4.914456367492676,
+ "step": 699
+ },
+ {
+ "epoch": 9.726637554585153,
+ "grad_norm": 0.013434977270662785,
+ "learning_rate": 0.0006,
+ "loss": 5.063764572143555,
+ "step": 700
+ },
+ {
+ "epoch": 9.74061135371179,
+ "grad_norm": 0.015128864906728268,
+ "learning_rate": 0.0006,
+ "loss": 5.052367687225342,
+ "step": 701
+ },
+ {
+ "epoch": 9.754585152838429,
+ "grad_norm": 0.014210768043994904,
+ "learning_rate": 0.0006,
+ "loss": 5.016489028930664,
+ "step": 702
+ },
+ {
+ "epoch": 9.768558951965066,
+ "grad_norm": 0.015191680751740932,
+ "learning_rate": 0.0006,
+ "loss": 4.990176200866699,
+ "step": 703
+ },
+ {
+ "epoch": 9.782532751091702,
+ "grad_norm": 0.018285127356648445,
+ "learning_rate": 0.0006,
+ "loss": 5.015164375305176,
+ "step": 704
+ },
+ {
+ "epoch": 9.796506550218341,
+ "grad_norm": 0.0204079058021307,
+ "learning_rate": 0.0006,
+ "loss": 5.091142654418945,
+ "step": 705
+ },
+ {
+ "epoch": 9.810480349344978,
+ "grad_norm": 0.01980586163699627,
+ "learning_rate": 0.0006,
+ "loss": 5.03350830078125,
+ "step": 706
+ },
+ {
+ "epoch": 9.824454148471617,
+ "grad_norm": 0.02048729732632637,
+ "learning_rate": 0.0006,
+ "loss": 5.03611421585083,
+ "step": 707
+ },
+ {
+ "epoch": 9.838427947598253,
+ "grad_norm": 0.02058163844048977,
+ "learning_rate": 0.0006,
+ "loss": 4.9613447189331055,
+ "step": 708
+ },
+ {
+ "epoch": 9.85240174672489,
+ "grad_norm": 0.018004924058914185,
+ "learning_rate": 0.0006,
+ "loss": 4.966541767120361,
+ "step": 709
+ },
+ {
+ "epoch": 9.866375545851529,
+ "grad_norm": 0.0198152307420969,
+ "learning_rate": 0.0006,
+ "loss": 5.01750373840332,
+ "step": 710
+ },
+ {
+ "epoch": 9.880349344978166,
+ "grad_norm": 0.018532969057559967,
+ "learning_rate": 0.0006,
+ "loss": 5.097424507141113,
+ "step": 711
+ },
+ {
+ "epoch": 9.894323144104803,
+ "grad_norm": 0.018939971923828125,
+ "learning_rate": 0.0006,
+ "loss": 5.177223205566406,
+ "step": 712
+ },
+ {
+ "epoch": 9.908296943231441,
+ "grad_norm": 0.017969170585274696,
+ "learning_rate": 0.0006,
+ "loss": 5.118013858795166,
+ "step": 713
+ },
+ {
+ "epoch": 9.922270742358078,
+ "grad_norm": 0.018021270632743835,
+ "learning_rate": 0.0006,
+ "loss": 5.034787178039551,
+ "step": 714
+ },
+ {
+ "epoch": 9.936244541484717,
+ "grad_norm": 0.017009710893034935,
+ "learning_rate": 0.0006,
+ "loss": 5.003920555114746,
+ "step": 715
+ },
+ {
+ "epoch": 9.950218340611354,
+ "grad_norm": 0.017722049728035927,
+ "learning_rate": 0.0006,
+ "loss": 5.031939506530762,
+ "step": 716
+ },
+ {
+ "epoch": 9.96419213973799,
+ "grad_norm": 0.017134087160229683,
+ "learning_rate": 0.0006,
+ "loss": 5.093474388122559,
+ "step": 717
+ },
+ {
+ "epoch": 9.97816593886463,
+ "grad_norm": 0.018511613830924034,
+ "learning_rate": 0.0006,
+ "loss": 5.078604698181152,
+ "step": 718
+ },
+ {
+ "epoch": 9.992139737991266,
+ "grad_norm": 0.021206334233283997,
+ "learning_rate": 0.0006,
+ "loss": 5.052572727203369,
+ "step": 719
+ },
+ {
+ "epoch": 10.0,
+ "grad_norm": 0.02239062264561653,
+ "learning_rate": 0.0006,
+ "loss": 5.063547134399414,
+ "step": 720
+ },
+ {
+ "epoch": 10.0,
+ "eval_loss": 5.173044681549072,
+ "eval_runtime": 56.7287,
+ "eval_samples_per_second": 43.047,
+ "eval_steps_per_second": 1.357,
+ "step": 720
+ },
+ {
+ "epoch": 10.013973799126637,
+ "grad_norm": 0.020757969468832016,
+ "learning_rate": 0.0006,
+ "loss": 5.045648574829102,
+ "step": 721
+ },
+ {
+ "epoch": 10.027947598253276,
+ "grad_norm": 0.022697702050209045,
+ "learning_rate": 0.0006,
+ "loss": 4.993833065032959,
+ "step": 722
+ },
+ {
+ "epoch": 10.041921397379912,
+ "grad_norm": 0.02434331737458706,
+ "learning_rate": 0.0006,
+ "loss": 5.073219299316406,
+ "step": 723
+ },
+ {
+ "epoch": 10.055895196506551,
+ "grad_norm": 0.025499660521745682,
+ "learning_rate": 0.0006,
+ "loss": 5.019072532653809,
+ "step": 724
+ },
+ {
+ "epoch": 10.069868995633188,
+ "grad_norm": 0.029281053692102432,
+ "learning_rate": 0.0006,
+ "loss": 4.959851264953613,
+ "step": 725
+ },
+ {
+ "epoch": 10.083842794759825,
+ "grad_norm": 0.033878110349178314,
+ "learning_rate": 0.0006,
+ "loss": 5.132717609405518,
+ "step": 726
+ },
+ {
+ "epoch": 10.097816593886463,
+ "grad_norm": 0.03385764732956886,
+ "learning_rate": 0.0006,
+ "loss": 5.0482892990112305,
+ "step": 727
+ },
+ {
+ "epoch": 10.1117903930131,
+ "grad_norm": 0.03466648980975151,
+ "learning_rate": 0.0006,
+ "loss": 4.946855545043945,
+ "step": 728
+ },
+ {
+ "epoch": 10.125764192139737,
+ "grad_norm": 0.028546305373311043,
+ "learning_rate": 0.0006,
+ "loss": 5.027774333953857,
+ "step": 729
+ },
+ {
+ "epoch": 10.139737991266376,
+ "grad_norm": 0.030575869604945183,
+ "learning_rate": 0.0006,
+ "loss": 4.9768476486206055,
+ "step": 730
+ },
+ {
+ "epoch": 10.153711790393013,
+ "grad_norm": 0.03614303469657898,
+ "learning_rate": 0.0006,
+ "loss": 5.078113555908203,
+ "step": 731
+ },
+ {
+ "epoch": 10.167685589519651,
+ "grad_norm": 0.040991660207509995,
+ "learning_rate": 0.0006,
+ "loss": 5.033185958862305,
+ "step": 732
+ },
+ {
+ "epoch": 10.181659388646288,
+ "grad_norm": 0.04551699757575989,
+ "learning_rate": 0.0006,
+ "loss": 5.0820746421813965,
+ "step": 733
+ },
+ {
+ "epoch": 10.195633187772925,
+ "grad_norm": 0.040193989872932434,
+ "learning_rate": 0.0006,
+ "loss": 5.059700965881348,
+ "step": 734
+ },
+ {
+ "epoch": 10.209606986899564,
+ "grad_norm": 0.035851605236530304,
+ "learning_rate": 0.0006,
+ "loss": 5.142467498779297,
+ "step": 735
+ },
+ {
+ "epoch": 10.2235807860262,
+ "grad_norm": 0.034200169146060944,
+ "learning_rate": 0.0006,
+ "loss": 5.072312355041504,
+ "step": 736
+ },
+ {
+ "epoch": 10.237554585152838,
+ "grad_norm": 0.033621110022068024,
+ "learning_rate": 0.0006,
+ "loss": 5.08037805557251,
+ "step": 737
+ },
+ {
+ "epoch": 10.251528384279476,
+ "grad_norm": 0.037372369319200516,
+ "learning_rate": 0.0006,
+ "loss": 4.998425483703613,
+ "step": 738
+ },
+ {
+ "epoch": 10.265502183406113,
+ "grad_norm": 0.037044707685709,
+ "learning_rate": 0.0006,
+ "loss": 4.923993110656738,
+ "step": 739
+ },
+ {
+ "epoch": 10.279475982532752,
+ "grad_norm": 0.030454808846116066,
+ "learning_rate": 0.0006,
+ "loss": 5.0889997482299805,
+ "step": 740
+ },
+ {
+ "epoch": 10.293449781659389,
+ "grad_norm": 0.030969638377428055,
+ "learning_rate": 0.0006,
+ "loss": 4.984112739562988,
+ "step": 741
+ },
+ {
+ "epoch": 10.307423580786025,
+ "grad_norm": 0.0276983380317688,
+ "learning_rate": 0.0006,
+ "loss": 4.955412864685059,
+ "step": 742
+ },
+ {
+ "epoch": 10.321397379912664,
+ "grad_norm": 0.02716052532196045,
+ "learning_rate": 0.0006,
+ "loss": 5.0098876953125,
+ "step": 743
+ },
+ {
+ "epoch": 10.335371179039301,
+ "grad_norm": 0.02403552085161209,
+ "learning_rate": 0.0006,
+ "loss": 5.165160655975342,
+ "step": 744
+ },
+ {
+ "epoch": 10.34934497816594,
+ "grad_norm": 0.0257862601429224,
+ "learning_rate": 0.0006,
+ "loss": 5.001547813415527,
+ "step": 745
+ },
+ {
+ "epoch": 10.363318777292577,
+ "grad_norm": 0.02300378680229187,
+ "learning_rate": 0.0006,
+ "loss": 5.038888931274414,
+ "step": 746
+ },
+ {
+ "epoch": 10.377292576419213,
+ "grad_norm": 0.019766854122281075,
+ "learning_rate": 0.0006,
+ "loss": 4.924291610717773,
+ "step": 747
+ },
+ {
+ "epoch": 10.391266375545852,
+ "grad_norm": 0.01888395845890045,
+ "learning_rate": 0.0006,
+ "loss": 5.059971332550049,
+ "step": 748
+ },
+ {
+ "epoch": 10.405240174672489,
+ "grad_norm": 0.015577499754726887,
+ "learning_rate": 0.0006,
+ "loss": 4.987686634063721,
+ "step": 749
+ },
+ {
+ "epoch": 10.419213973799126,
+ "grad_norm": 0.016012346372008324,
+ "learning_rate": 0.0006,
+ "loss": 5.015157699584961,
+ "step": 750
+ },
+ {
+ "epoch": 10.433187772925764,
+ "grad_norm": 0.016277998685836792,
+ "learning_rate": 0.0006,
+ "loss": 5.004931449890137,
+ "step": 751
+ },
+ {
+ "epoch": 10.447161572052401,
+ "grad_norm": 0.017504561692476273,
+ "learning_rate": 0.0006,
+ "loss": 4.902827739715576,
+ "step": 752
+ },
+ {
+ "epoch": 10.46113537117904,
+ "grad_norm": 0.01569022797048092,
+ "learning_rate": 0.0006,
+ "loss": 4.954863548278809,
+ "step": 753
+ },
+ {
+ "epoch": 10.475109170305677,
+ "grad_norm": 0.014114422723650932,
+ "learning_rate": 0.0006,
+ "loss": 5.00656795501709,
+ "step": 754
+ },
+ {
+ "epoch": 10.489082969432314,
+ "grad_norm": 0.01573832333087921,
+ "learning_rate": 0.0006,
+ "loss": 4.978353023529053,
+ "step": 755
+ },
+ {
+ "epoch": 10.503056768558952,
+ "grad_norm": 0.016480036079883575,
+ "learning_rate": 0.0006,
+ "loss": 4.953327655792236,
+ "step": 756
+ },
+ {
+ "epoch": 10.51703056768559,
+ "grad_norm": 0.014008025638759136,
+ "learning_rate": 0.0006,
+ "loss": 5.050840377807617,
+ "step": 757
+ },
+ {
+ "epoch": 10.531004366812226,
+ "grad_norm": 0.013147998601198196,
+ "learning_rate": 0.0006,
+ "loss": 4.974964141845703,
+ "step": 758
+ },
+ {
+ "epoch": 10.544978165938865,
+ "grad_norm": 0.013807603158056736,
+ "learning_rate": 0.0006,
+ "loss": 4.927907943725586,
+ "step": 759
+ },
+ {
+ "epoch": 10.558951965065502,
+ "grad_norm": 0.013555224984884262,
+ "learning_rate": 0.0006,
+ "loss": 4.980690956115723,
+ "step": 760
+ },
+ {
+ "epoch": 10.57292576419214,
+ "grad_norm": 0.013045408762991428,
+ "learning_rate": 0.0006,
+ "loss": 4.9436116218566895,
+ "step": 761
+ },
+ {
+ "epoch": 10.586899563318777,
+ "grad_norm": 0.012858862057328224,
+ "learning_rate": 0.0006,
+ "loss": 4.9595770835876465,
+ "step": 762
+ },
+ {
+ "epoch": 10.600873362445414,
+ "grad_norm": 0.01468253880739212,
+ "learning_rate": 0.0006,
+ "loss": 4.90725040435791,
+ "step": 763
+ },
+ {
+ "epoch": 10.614847161572053,
+ "grad_norm": 0.013224679045379162,
+ "learning_rate": 0.0006,
+ "loss": 4.877615928649902,
+ "step": 764
+ },
+ {
+ "epoch": 10.62882096069869,
+ "grad_norm": 0.013555348850786686,
+ "learning_rate": 0.0006,
+ "loss": 4.9456610679626465,
+ "step": 765
+ },
+ {
+ "epoch": 10.642794759825328,
+ "grad_norm": 0.014387160539627075,
+ "learning_rate": 0.0006,
+ "loss": 4.9233717918396,
+ "step": 766
+ },
+ {
+ "epoch": 10.656768558951965,
+ "grad_norm": 0.014354088343679905,
+ "learning_rate": 0.0006,
+ "loss": 4.983511924743652,
+ "step": 767
+ },
+ {
+ "epoch": 10.670742358078602,
+ "grad_norm": 0.01332700252532959,
+ "learning_rate": 0.0006,
+ "loss": 4.934416770935059,
+ "step": 768
+ },
+ {
+ "epoch": 10.68471615720524,
+ "grad_norm": 0.011543313041329384,
+ "learning_rate": 0.0006,
+ "loss": 4.96388053894043,
+ "step": 769
+ },
+ {
+ "epoch": 10.698689956331878,
+ "grad_norm": 0.012330878525972366,
+ "learning_rate": 0.0006,
+ "loss": 4.946396350860596,
+ "step": 770
+ },
+ {
+ "epoch": 10.712663755458514,
+ "grad_norm": 0.01379779726266861,
+ "learning_rate": 0.0006,
+ "loss": 4.981019020080566,
+ "step": 771
+ },
+ {
+ "epoch": 10.726637554585153,
+ "grad_norm": 0.014612431637942791,
+ "learning_rate": 0.0006,
+ "loss": 4.894567966461182,
+ "step": 772
+ },
+ {
+ "epoch": 10.74061135371179,
+ "grad_norm": 0.013062899932265282,
+ "learning_rate": 0.0006,
+ "loss": 5.033038139343262,
+ "step": 773
+ },
+ {
+ "epoch": 10.754585152838429,
+ "grad_norm": 0.013898049481213093,
+ "learning_rate": 0.0006,
+ "loss": 4.823636054992676,
+ "step": 774
+ },
+ {
+ "epoch": 10.768558951965066,
+ "grad_norm": 0.01502019353210926,
+ "learning_rate": 0.0006,
+ "loss": 5.020156383514404,
+ "step": 775
+ },
+ {
+ "epoch": 10.782532751091702,
+ "grad_norm": 0.013810204342007637,
+ "learning_rate": 0.0006,
+ "loss": 4.935647010803223,
+ "step": 776
+ },
+ {
+ "epoch": 10.796506550218341,
+ "grad_norm": 0.013616513460874557,
+ "learning_rate": 0.0006,
+ "loss": 4.901486396789551,
+ "step": 777
+ },
+ {
+ "epoch": 10.810480349344978,
+ "grad_norm": 0.013601024635136127,
+ "learning_rate": 0.0006,
+ "loss": 4.931873321533203,
+ "step": 778
+ },
+ {
+ "epoch": 10.824454148471617,
+ "grad_norm": 0.013668350875377655,
+ "learning_rate": 0.0006,
+ "loss": 4.937256813049316,
+ "step": 779
+ },
+ {
+ "epoch": 10.838427947598253,
+ "grad_norm": 0.01581823080778122,
+ "learning_rate": 0.0006,
+ "loss": 4.9611406326293945,
+ "step": 780
+ },
+ {
+ "epoch": 10.85240174672489,
+ "grad_norm": 0.019478484988212585,
+ "learning_rate": 0.0006,
+ "loss": 4.9212236404418945,
+ "step": 781
+ },
+ {
+ "epoch": 10.866375545851529,
+ "grad_norm": 0.026921333745121956,
+ "learning_rate": 0.0006,
+ "loss": 4.88501501083374,
+ "step": 782
+ },
+ {
+ "epoch": 10.880349344978166,
+ "grad_norm": 0.03155914694070816,
+ "learning_rate": 0.0006,
+ "loss": 4.989967346191406,
+ "step": 783
+ },
+ {
+ "epoch": 10.894323144104803,
+ "grad_norm": 0.029689429327845573,
+ "learning_rate": 0.0006,
+ "loss": 4.936000823974609,
+ "step": 784
+ },
+ {
+ "epoch": 10.908296943231441,
+ "grad_norm": 0.026626063510775566,
+ "learning_rate": 0.0006,
+ "loss": 4.984785079956055,
+ "step": 785
+ },
+ {
+ "epoch": 10.922270742358078,
+ "grad_norm": 0.026168212294578552,
+ "learning_rate": 0.0006,
+ "loss": 4.908282279968262,
+ "step": 786
+ },
+ {
+ "epoch": 10.936244541484717,
+ "grad_norm": 0.02828158251941204,
+ "learning_rate": 0.0006,
+ "loss": 4.926390647888184,
+ "step": 787
+ },
+ {
+ "epoch": 10.950218340611354,
+ "grad_norm": 0.02649078331887722,
+ "learning_rate": 0.0006,
+ "loss": 5.00400972366333,
+ "step": 788
+ },
+ {
+ "epoch": 10.96419213973799,
+ "grad_norm": 0.0256856270134449,
+ "learning_rate": 0.0006,
+ "loss": 5.0317487716674805,
+ "step": 789
+ },
+ {
+ "epoch": 10.97816593886463,
+ "grad_norm": 0.024728883057832718,
+ "learning_rate": 0.0006,
+ "loss": 4.9299635887146,
+ "step": 790
+ },
+ {
+ "epoch": 10.992139737991266,
+ "grad_norm": 0.026241116225719452,
+ "learning_rate": 0.0006,
+ "loss": 4.947319984436035,
+ "step": 791
+ },
+ {
+ "epoch": 11.0,
+ "grad_norm": 0.023526743054389954,
+ "learning_rate": 0.0006,
+ "loss": 4.94429349899292,
+ "step": 792
+ },
+ {
+ "epoch": 11.0,
+ "eval_loss": 5.104345798492432,
+ "eval_runtime": 57.0955,
+ "eval_samples_per_second": 42.77,
+ "eval_steps_per_second": 1.349,
+ "step": 792
+ },
+ {
+ "epoch": 11.013973799126637,
+ "grad_norm": 0.0216788612306118,
+ "learning_rate": 0.0006,
+ "loss": 4.881950855255127,
+ "step": 793
+ },
+ {
+ "epoch": 11.027947598253276,
+ "grad_norm": 0.01830463856458664,
+ "learning_rate": 0.0006,
+ "loss": 4.971016883850098,
+ "step": 794
+ },
+ {
+ "epoch": 11.041921397379912,
+ "grad_norm": 0.02204521745443344,
+ "learning_rate": 0.0006,
+ "loss": 4.894623756408691,
+ "step": 795
+ },
+ {
+ "epoch": 11.055895196506551,
+ "grad_norm": 0.0210255216807127,
+ "learning_rate": 0.0006,
+ "loss": 4.834009170532227,
+ "step": 796
+ },
+ {
+ "epoch": 11.069868995633188,
+ "grad_norm": 0.019797448068857193,
+ "learning_rate": 0.0006,
+ "loss": 5.086714744567871,
+ "step": 797
+ },
+ {
+ "epoch": 11.083842794759825,
+ "grad_norm": 0.02036641724407673,
+ "learning_rate": 0.0006,
+ "loss": 4.920056343078613,
+ "step": 798
+ },
+ {
+ "epoch": 11.097816593886463,
+ "grad_norm": 0.01966066285967827,
+ "learning_rate": 0.0006,
+ "loss": 4.879859447479248,
+ "step": 799
+ },
+ {
+ "epoch": 11.1117903930131,
+ "grad_norm": 0.0202149897813797,
+ "learning_rate": 0.0006,
+ "loss": 4.933640480041504,
+ "step": 800
+ },
+ {
+ "epoch": 11.125764192139737,
+ "grad_norm": 0.01893039606511593,
+ "learning_rate": 0.0006,
+ "loss": 4.901970863342285,
+ "step": 801
+ },
+ {
+ "epoch": 11.139737991266376,
+ "grad_norm": 0.016040155664086342,
+ "learning_rate": 0.0006,
+ "loss": 4.905491828918457,
+ "step": 802
+ },
+ {
+ "epoch": 11.153711790393013,
+ "grad_norm": 0.01720350608229637,
+ "learning_rate": 0.0006,
+ "loss": 4.798816680908203,
+ "step": 803
+ },
+ {
+ "epoch": 11.167685589519651,
+ "grad_norm": 0.017002852633595467,
+ "learning_rate": 0.0006,
+ "loss": 4.851859092712402,
+ "step": 804
+ },
+ {
+ "epoch": 11.181659388646288,
+ "grad_norm": 0.015594680793583393,
+ "learning_rate": 0.0006,
+ "loss": 4.90328311920166,
+ "step": 805
+ },
+ {
+ "epoch": 11.195633187772925,
+ "grad_norm": 0.014904462732374668,
+ "learning_rate": 0.0006,
+ "loss": 4.904215335845947,
+ "step": 806
+ },
+ {
+ "epoch": 11.209606986899564,
+ "grad_norm": 0.01366228237748146,
+ "learning_rate": 0.0006,
+ "loss": 4.9813055992126465,
+ "step": 807
+ },
+ {
+ "epoch": 11.2235807860262,
+ "grad_norm": 0.014719638973474503,
+ "learning_rate": 0.0006,
+ "loss": 4.927306175231934,
+ "step": 808
+ },
+ {
+ "epoch": 11.237554585152838,
+ "grad_norm": 0.015016036108136177,
+ "learning_rate": 0.0006,
+ "loss": 4.874215126037598,
+ "step": 809
+ },
+ {
+ "epoch": 11.251528384279476,
+ "grad_norm": 0.014925424009561539,
+ "learning_rate": 0.0006,
+ "loss": 4.838257789611816,
+ "step": 810
+ },
+ {
+ "epoch": 11.265502183406113,
+ "grad_norm": 0.014335823245346546,
+ "learning_rate": 0.0006,
+ "loss": 4.940608024597168,
+ "step": 811
+ },
+ {
+ "epoch": 11.279475982532752,
+ "grad_norm": 0.014987512491643429,
+ "learning_rate": 0.0006,
+ "loss": 4.831531524658203,
+ "step": 812
+ },
+ {
+ "epoch": 11.293449781659389,
+ "grad_norm": 0.016087457537651062,
+ "learning_rate": 0.0006,
+ "loss": 4.863227367401123,
+ "step": 813
+ },
+ {
+ "epoch": 11.307423580786025,
+ "grad_norm": 0.01671300269663334,
+ "learning_rate": 0.0006,
+ "loss": 4.8535661697387695,
+ "step": 814
+ },
+ {
+ "epoch": 11.321397379912664,
+ "grad_norm": 0.015229142270982265,
+ "learning_rate": 0.0006,
+ "loss": 4.948635101318359,
+ "step": 815
+ },
+ {
+ "epoch": 11.335371179039301,
+ "grad_norm": 0.01532587967813015,
+ "learning_rate": 0.0006,
+ "loss": 4.9035797119140625,
+ "step": 816
+ },
+ {
+ "epoch": 11.34934497816594,
+ "grad_norm": 0.015715476125478745,
+ "learning_rate": 0.0006,
+ "loss": 4.894709587097168,
+ "step": 817
+ },
+ {
+ "epoch": 11.363318777292577,
+ "grad_norm": 0.014156977646052837,
+ "learning_rate": 0.0006,
+ "loss": 4.960206985473633,
+ "step": 818
+ },
+ {
+ "epoch": 11.377292576419213,
+ "grad_norm": 0.017076566815376282,
+ "learning_rate": 0.0006,
+ "loss": 5.005850791931152,
+ "step": 819
+ },
+ {
+ "epoch": 11.391266375545852,
+ "grad_norm": 0.0172084029763937,
+ "learning_rate": 0.0006,
+ "loss": 4.903355598449707,
+ "step": 820
+ },
+ {
+ "epoch": 11.405240174672489,
+ "grad_norm": 0.0180258359760046,
+ "learning_rate": 0.0006,
+ "loss": 4.842667579650879,
+ "step": 821
+ },
+ {
+ "epoch": 11.419213973799126,
+ "grad_norm": 0.022130120545625687,
+ "learning_rate": 0.0006,
+ "loss": 4.799970626831055,
+ "step": 822
+ },
+ {
+ "epoch": 11.433187772925764,
+ "grad_norm": 0.02384897693991661,
+ "learning_rate": 0.0006,
+ "loss": 4.860746383666992,
+ "step": 823
+ },
+ {
+ "epoch": 11.447161572052401,
+ "grad_norm": 0.021881572902202606,
+ "learning_rate": 0.0006,
+ "loss": 4.918107032775879,
+ "step": 824
+ },
+ {
+ "epoch": 11.46113537117904,
+ "grad_norm": 0.023085080087184906,
+ "learning_rate": 0.0006,
+ "loss": 4.878946304321289,
+ "step": 825
+ },
+ {
+ "epoch": 11.475109170305677,
+ "grad_norm": 0.02291012555360794,
+ "learning_rate": 0.0006,
+ "loss": 4.958887100219727,
+ "step": 826
+ },
+ {
+ "epoch": 11.489082969432314,
+ "grad_norm": 0.0199968870729208,
+ "learning_rate": 0.0006,
+ "loss": 4.80955696105957,
+ "step": 827
+ },
+ {
+ "epoch": 11.503056768558952,
+ "grad_norm": 0.018702229484915733,
+ "learning_rate": 0.0006,
+ "loss": 5.0276408195495605,
+ "step": 828
+ },
+ {
+ "epoch": 11.51703056768559,
+ "grad_norm": 0.017787311226129532,
+ "learning_rate": 0.0006,
+ "loss": 4.96018123626709,
+ "step": 829
+ },
+ {
+ "epoch": 11.531004366812226,
+ "grad_norm": 0.01704501546919346,
+ "learning_rate": 0.0006,
+ "loss": 4.914553165435791,
+ "step": 830
+ },
+ {
+ "epoch": 11.544978165938865,
+ "grad_norm": 0.017519677057862282,
+ "learning_rate": 0.0006,
+ "loss": 4.798098087310791,
+ "step": 831
+ },
+ {
+ "epoch": 11.558951965065502,
+ "grad_norm": 0.018511991947889328,
+ "learning_rate": 0.0006,
+ "loss": 4.817785263061523,
+ "step": 832
+ },
+ {
+ "epoch": 11.57292576419214,
+ "grad_norm": 0.01876644790172577,
+ "learning_rate": 0.0006,
+ "loss": 4.818995475769043,
+ "step": 833
+ },
+ {
+ "epoch": 11.586899563318777,
+ "grad_norm": 0.01753568835556507,
+ "learning_rate": 0.0006,
+ "loss": 4.811938285827637,
+ "step": 834
+ },
+ {
+ "epoch": 11.600873362445414,
+ "grad_norm": 0.017046531662344933,
+ "learning_rate": 0.0006,
+ "loss": 4.914674282073975,
+ "step": 835
+ },
+ {
+ "epoch": 11.614847161572053,
+ "grad_norm": 0.01802094094455242,
+ "learning_rate": 0.0006,
+ "loss": 4.993284225463867,
+ "step": 836
+ },
+ {
+ "epoch": 11.62882096069869,
+ "grad_norm": 0.019493764266371727,
+ "learning_rate": 0.0006,
+ "loss": 4.895628452301025,
+ "step": 837
+ },
+ {
+ "epoch": 11.642794759825328,
+ "grad_norm": 0.019461210817098618,
+ "learning_rate": 0.0006,
+ "loss": 4.869269371032715,
+ "step": 838
+ },
+ {
+ "epoch": 11.656768558951965,
+ "grad_norm": 0.01981567218899727,
+ "learning_rate": 0.0006,
+ "loss": 4.863052845001221,
+ "step": 839
+ },
+ {
+ "epoch": 11.670742358078602,
+ "grad_norm": 0.019894491881132126,
+ "learning_rate": 0.0006,
+ "loss": 4.8775787353515625,
+ "step": 840
+ },
+ {
+ "epoch": 11.68471615720524,
+ "grad_norm": 0.01875786855816841,
+ "learning_rate": 0.0006,
+ "loss": 4.817763328552246,
+ "step": 841
+ },
+ {
+ "epoch": 11.698689956331878,
+ "grad_norm": 0.020926695317029953,
+ "learning_rate": 0.0006,
+ "loss": 4.7470526695251465,
+ "step": 842
+ },
+ {
+ "epoch": 11.712663755458514,
+ "grad_norm": 0.021477103233337402,
+ "learning_rate": 0.0006,
+ "loss": 4.762682914733887,
+ "step": 843
+ },
+ {
+ "epoch": 11.726637554585153,
+ "grad_norm": 0.020628711208701134,
+ "learning_rate": 0.0006,
+ "loss": 4.807985782623291,
+ "step": 844
+ },
+ {
+ "epoch": 11.74061135371179,
+ "grad_norm": 0.021489612758159637,
+ "learning_rate": 0.0006,
+ "loss": 4.843091011047363,
+ "step": 845
+ },
+ {
+ "epoch": 11.754585152838429,
+ "grad_norm": 0.021194491535425186,
+ "learning_rate": 0.0006,
+ "loss": 4.8300251960754395,
+ "step": 846
+ },
+ {
+ "epoch": 11.768558951965066,
+ "grad_norm": 0.01833350397646427,
+ "learning_rate": 0.0006,
+ "loss": 5.000202178955078,
+ "step": 847
+ },
+ {
+ "epoch": 11.782532751091702,
+ "grad_norm": 0.018096931278705597,
+ "learning_rate": 0.0006,
+ "loss": 4.880648136138916,
+ "step": 848
+ },
+ {
+ "epoch": 11.796506550218341,
+ "grad_norm": 0.0197161752730608,
+ "learning_rate": 0.0006,
+ "loss": 4.788316249847412,
+ "step": 849
+ },
+ {
+ "epoch": 11.810480349344978,
+ "grad_norm": 0.0203181654214859,
+ "learning_rate": 0.0006,
+ "loss": 4.870944976806641,
+ "step": 850
+ },
+ {
+ "epoch": 11.824454148471617,
+ "grad_norm": 0.01908455230295658,
+ "learning_rate": 0.0006,
+ "loss": 4.890144348144531,
+ "step": 851
+ },
+ {
+ "epoch": 11.838427947598253,
+ "grad_norm": 0.021281801164150238,
+ "learning_rate": 0.0006,
+ "loss": 4.89578914642334,
+ "step": 852
+ },
+ {
+ "epoch": 11.85240174672489,
+ "grad_norm": 0.02124273031949997,
+ "learning_rate": 0.0006,
+ "loss": 4.883749008178711,
+ "step": 853
+ },
+ {
+ "epoch": 11.866375545851529,
+ "grad_norm": 0.020969398319721222,
+ "learning_rate": 0.0006,
+ "loss": 4.809142589569092,
+ "step": 854
+ },
+ {
+ "epoch": 11.880349344978166,
+ "grad_norm": 0.01967657171189785,
+ "learning_rate": 0.0006,
+ "loss": 4.821864604949951,
+ "step": 855
+ },
+ {
+ "epoch": 11.894323144104803,
+ "grad_norm": 0.020887333899736404,
+ "learning_rate": 0.0006,
+ "loss": 4.917985916137695,
+ "step": 856
+ },
+ {
+ "epoch": 11.908296943231441,
+ "grad_norm": 0.019203083589673042,
+ "learning_rate": 0.0006,
+ "loss": 4.824960708618164,
+ "step": 857
+ },
+ {
+ "epoch": 11.922270742358078,
+ "grad_norm": 0.020162401720881462,
+ "learning_rate": 0.0006,
+ "loss": 4.772617816925049,
+ "step": 858
+ },
+ {
+ "epoch": 11.936244541484717,
+ "grad_norm": 0.02251487225294113,
+ "learning_rate": 0.0006,
+ "loss": 4.735418319702148,
+ "step": 859
+ },
+ {
+ "epoch": 11.950218340611354,
+ "grad_norm": 0.020313527435064316,
+ "learning_rate": 0.0006,
+ "loss": 4.939233779907227,
+ "step": 860
+ },
+ {
+ "epoch": 11.96419213973799,
+ "grad_norm": 0.021635599434375763,
+ "learning_rate": 0.0006,
+ "loss": 4.861807346343994,
+ "step": 861
+ },
+ {
+ "epoch": 11.97816593886463,
+ "grad_norm": 0.02285311557352543,
+ "learning_rate": 0.0006,
+ "loss": 4.932732582092285,
+ "step": 862
+ },
+ {
+ "epoch": 11.992139737991266,
+ "grad_norm": 0.02072535641491413,
+ "learning_rate": 0.0006,
+ "loss": 4.796767711639404,
+ "step": 863
+ },
+ {
+ "epoch": 12.0,
+ "grad_norm": 0.020588304847478867,
+ "learning_rate": 0.0006,
+ "loss": 4.997537612915039,
+ "step": 864
+ },
+ {
+ "epoch": 12.0,
+ "eval_loss": 5.044651985168457,
+ "eval_runtime": 56.9145,
+ "eval_samples_per_second": 42.906,
+ "eval_steps_per_second": 1.353,
+ "step": 864
+ },
+ {
+ "epoch": 12.013973799126637,
+ "grad_norm": 0.017597166821360588,
+ "learning_rate": 0.0006,
+ "loss": 4.830524444580078,
+ "step": 865
+ },
+ {
+ "epoch": 12.027947598253276,
+ "grad_norm": 0.01716788113117218,
+ "learning_rate": 0.0006,
+ "loss": 4.894325256347656,
+ "step": 866
+ },
+ {
+ "epoch": 12.041921397379912,
+ "grad_norm": 0.017712457105517387,
+ "learning_rate": 0.0006,
+ "loss": 4.868297576904297,
+ "step": 867
+ },
+ {
+ "epoch": 12.055895196506551,
+ "grad_norm": 0.021486390382051468,
+ "learning_rate": 0.0006,
+ "loss": 4.922049522399902,
+ "step": 868
+ },
+ {
+ "epoch": 12.069868995633188,
+ "grad_norm": 0.024560654535889626,
+ "learning_rate": 0.0006,
+ "loss": 4.734241008758545,
+ "step": 869
+ },
+ {
+ "epoch": 12.083842794759825,
+ "grad_norm": 0.025870252400636673,
+ "learning_rate": 0.0006,
+ "loss": 4.868566989898682,
+ "step": 870
+ },
+ {
+ "epoch": 12.097816593886463,
+ "grad_norm": 0.0277389045804739,
+ "learning_rate": 0.0006,
+ "loss": 4.755006790161133,
+ "step": 871
+ },
+ {
+ "epoch": 12.1117903930131,
+ "grad_norm": 0.029100844636559486,
+ "learning_rate": 0.0006,
+ "loss": 4.909667015075684,
+ "step": 872
+ },
+ {
+ "epoch": 12.125764192139737,
+ "grad_norm": 0.031833067536354065,
+ "learning_rate": 0.0006,
+ "loss": 4.858674049377441,
+ "step": 873
+ },
+ {
+ "epoch": 12.139737991266376,
+ "grad_norm": 0.032709237188100815,
+ "learning_rate": 0.0006,
+ "loss": 4.823696136474609,
+ "step": 874
+ },
+ {
+ "epoch": 12.153711790393013,
+ "grad_norm": 0.02995380386710167,
+ "learning_rate": 0.0006,
+ "loss": 4.850235462188721,
+ "step": 875
+ },
+ {
+ "epoch": 12.167685589519651,
+ "grad_norm": 0.034098610281944275,
+ "learning_rate": 0.0006,
+ "loss": 4.8169989585876465,
+ "step": 876
+ },
+ {
+ "epoch": 12.181659388646288,
+ "grad_norm": 0.03377080336213112,
+ "learning_rate": 0.0006,
+ "loss": 4.90738582611084,
+ "step": 877
+ },
+ {
+ "epoch": 12.195633187772925,
+ "grad_norm": 0.029780825600028038,
+ "learning_rate": 0.0006,
+ "loss": 4.811163902282715,
+ "step": 878
+ },
+ {
+ "epoch": 12.209606986899564,
+ "grad_norm": 0.03118024580180645,
+ "learning_rate": 0.0006,
+ "loss": 4.861286640167236,
+ "step": 879
+ },
+ {
+ "epoch": 12.2235807860262,
+ "grad_norm": 0.033052004873752594,
+ "learning_rate": 0.0006,
+ "loss": 4.865021228790283,
+ "step": 880
+ },
+ {
+ "epoch": 12.237554585152838,
+ "grad_norm": 0.029070032760500908,
+ "learning_rate": 0.0006,
+ "loss": 4.947303771972656,
+ "step": 881
+ },
+ {
+ "epoch": 12.251528384279476,
+ "grad_norm": 0.028355715796351433,
+ "learning_rate": 0.0006,
+ "loss": 4.804648399353027,
+ "step": 882
+ },
+ {
+ "epoch": 12.265502183406113,
+ "grad_norm": 0.031257227063179016,
+ "learning_rate": 0.0006,
+ "loss": 4.757718086242676,
+ "step": 883
+ },
+ {
+ "epoch": 12.279475982532752,
+ "grad_norm": 0.031438373029232025,
+ "learning_rate": 0.0006,
+ "loss": 4.851268768310547,
+ "step": 884
+ },
+ {
+ "epoch": 12.293449781659389,
+ "grad_norm": 0.030492138117551804,
+ "learning_rate": 0.0006,
+ "loss": 4.7891035079956055,
+ "step": 885
+ },
+ {
+ "epoch": 12.307423580786025,
+ "grad_norm": 0.026115991175174713,
+ "learning_rate": 0.0006,
+ "loss": 4.86474609375,
+ "step": 886
+ },
+ {
+ "epoch": 12.321397379912664,
+ "grad_norm": 0.024184465408325195,
+ "learning_rate": 0.0006,
+ "loss": 4.860456943511963,
+ "step": 887
+ },
+ {
+ "epoch": 12.335371179039301,
+ "grad_norm": 0.025437982752919197,
+ "learning_rate": 0.0006,
+ "loss": 4.806251049041748,
+ "step": 888
+ },
+ {
+ "epoch": 12.34934497816594,
+ "grad_norm": 0.022390231490135193,
+ "learning_rate": 0.0006,
+ "loss": 4.807086944580078,
+ "step": 889
+ },
+ {
+ "epoch": 12.363318777292577,
+ "grad_norm": 0.020174294710159302,
+ "learning_rate": 0.0006,
+ "loss": 4.78292179107666,
+ "step": 890
+ },
+ {
+ "epoch": 12.377292576419213,
+ "grad_norm": 0.019734324887394905,
+ "learning_rate": 0.0006,
+ "loss": 4.784509658813477,
+ "step": 891
+ },
+ {
+ "epoch": 12.391266375545852,
+ "grad_norm": 0.016733523458242416,
+ "learning_rate": 0.0006,
+ "loss": 4.811755180358887,
+ "step": 892
+ },
+ {
+ "epoch": 12.405240174672489,
+ "grad_norm": 0.01570146717131138,
+ "learning_rate": 0.0006,
+ "loss": 4.720088005065918,
+ "step": 893
+ },
+ {
+ "epoch": 12.419213973799126,
+ "grad_norm": 0.01601393334567547,
+ "learning_rate": 0.0006,
+ "loss": 4.783543109893799,
+ "step": 894
+ },
+ {
+ "epoch": 12.433187772925764,
+ "grad_norm": 0.014035487547516823,
+ "learning_rate": 0.0006,
+ "loss": 4.848340034484863,
+ "step": 895
+ },
+ {
+ "epoch": 12.447161572052401,
+ "grad_norm": 0.015248487703502178,
+ "learning_rate": 0.0006,
+ "loss": 4.773287773132324,
+ "step": 896
+ },
+ {
+ "epoch": 12.46113537117904,
+ "grad_norm": 0.01532268151640892,
+ "learning_rate": 0.0006,
+ "loss": 4.799646854400635,
+ "step": 897
+ },
+ {
+ "epoch": 12.475109170305677,
+ "grad_norm": 0.013982534408569336,
+ "learning_rate": 0.0006,
+ "loss": 4.8090500831604,
+ "step": 898
+ },
+ {
+ "epoch": 12.489082969432314,
+ "grad_norm": 0.014081164263188839,
+ "learning_rate": 0.0006,
+ "loss": 4.751888751983643,
+ "step": 899
+ },
+ {
+ "epoch": 12.503056768558952,
+ "grad_norm": 0.01380133256316185,
+ "learning_rate": 0.0006,
+ "loss": 4.78144645690918,
+ "step": 900
+ },
+ {
+ "epoch": 12.51703056768559,
+ "grad_norm": 0.013555269688367844,
+ "learning_rate": 0.0006,
+ "loss": 4.779257774353027,
+ "step": 901
+ },
+ {
+ "epoch": 12.531004366812226,
+ "grad_norm": 0.013075701892375946,
+ "learning_rate": 0.0006,
+ "loss": 4.820705413818359,
+ "step": 902
+ },
+ {
+ "epoch": 12.544978165938865,
+ "grad_norm": 0.01276316400617361,
+ "learning_rate": 0.0006,
+ "loss": 4.836733818054199,
+ "step": 903
+ },
+ {
+ "epoch": 12.558951965065502,
+ "grad_norm": 0.013033892959356308,
+ "learning_rate": 0.0006,
+ "loss": 4.764178276062012,
+ "step": 904
+ },
+ {
+ "epoch": 12.57292576419214,
+ "grad_norm": 0.013996980153024197,
+ "learning_rate": 0.0006,
+ "loss": 4.767983436584473,
+ "step": 905
+ },
+ {
+ "epoch": 12.586899563318777,
+ "grad_norm": 0.015222184360027313,
+ "learning_rate": 0.0006,
+ "loss": 4.759721755981445,
+ "step": 906
+ },
+ {
+ "epoch": 12.600873362445414,
+ "grad_norm": 0.01623941957950592,
+ "learning_rate": 0.0006,
+ "loss": 4.849811553955078,
+ "step": 907
+ },
+ {
+ "epoch": 12.614847161572053,
+ "grad_norm": 0.015711182728409767,
+ "learning_rate": 0.0006,
+ "loss": 4.7829084396362305,
+ "step": 908
+ },
+ {
+ "epoch": 12.62882096069869,
+ "grad_norm": 0.014137467369437218,
+ "learning_rate": 0.0006,
+ "loss": 4.73321533203125,
+ "step": 909
+ },
+ {
+ "epoch": 12.642794759825328,
+ "grad_norm": 0.012738462537527084,
+ "learning_rate": 0.0006,
+ "loss": 4.870052337646484,
+ "step": 910
+ },
+ {
+ "epoch": 12.656768558951965,
+ "grad_norm": 0.013683130033314228,
+ "learning_rate": 0.0006,
+ "loss": 4.82658052444458,
+ "step": 911
+ },
+ {
+ "epoch": 12.670742358078602,
+ "grad_norm": 0.014409830793738365,
+ "learning_rate": 0.0006,
+ "loss": 4.707494735717773,
+ "step": 912
+ },
+ {
+ "epoch": 12.68471615720524,
+ "grad_norm": 0.013806809671223164,
+ "learning_rate": 0.0006,
+ "loss": 4.778928279876709,
+ "step": 913
+ },
+ {
+ "epoch": 12.698689956331878,
+ "grad_norm": 0.012857912108302116,
+ "learning_rate": 0.0006,
+ "loss": 4.890717506408691,
+ "step": 914
+ },
+ {
+ "epoch": 12.712663755458514,
+ "grad_norm": 0.013828758150339127,
+ "learning_rate": 0.0006,
+ "loss": 4.8357439041137695,
+ "step": 915
+ },
+ {
+ "epoch": 12.726637554585153,
+ "grad_norm": 0.01478351280093193,
+ "learning_rate": 0.0006,
+ "loss": 4.869986534118652,
+ "step": 916
+ },
+ {
+ "epoch": 12.74061135371179,
+ "grad_norm": 0.01326421182602644,
+ "learning_rate": 0.0006,
+ "loss": 4.830646514892578,
+ "step": 917
+ },
+ {
+ "epoch": 12.754585152838429,
+ "grad_norm": 0.013335433788597584,
+ "learning_rate": 0.0006,
+ "loss": 4.7786149978637695,
+ "step": 918
+ },
+ {
+ "epoch": 12.768558951965066,
+ "grad_norm": 0.014261147007346153,
+ "learning_rate": 0.0006,
+ "loss": 4.788693428039551,
+ "step": 919
+ },
+ {
+ "epoch": 12.782532751091702,
+ "grad_norm": 0.014274783432483673,
+ "learning_rate": 0.0006,
+ "loss": 4.77295446395874,
+ "step": 920
+ },
+ {
+ "epoch": 12.796506550218341,
+ "grad_norm": 0.015074674971401691,
+ "learning_rate": 0.0006,
+ "loss": 4.830754280090332,
+ "step": 921
+ },
+ {
+ "epoch": 12.810480349344978,
+ "grad_norm": 0.015599515289068222,
+ "learning_rate": 0.0006,
+ "loss": 4.837759017944336,
+ "step": 922
+ },
+ {
+ "epoch": 12.824454148471617,
+ "grad_norm": 0.014595242217183113,
+ "learning_rate": 0.0006,
+ "loss": 4.882965564727783,
+ "step": 923
+ },
+ {
+ "epoch": 12.838427947598253,
+ "grad_norm": 0.014951467514038086,
+ "learning_rate": 0.0006,
+ "loss": 4.709748268127441,
+ "step": 924
+ },
+ {
+ "epoch": 12.85240174672489,
+ "grad_norm": 0.01539295818656683,
+ "learning_rate": 0.0006,
+ "loss": 4.791160583496094,
+ "step": 925
+ },
+ {
+ "epoch": 12.866375545851529,
+ "grad_norm": 0.015987906605005264,
+ "learning_rate": 0.0006,
+ "loss": 4.661876678466797,
+ "step": 926
+ },
+ {
+ "epoch": 12.880349344978166,
+ "grad_norm": 0.016527341678738594,
+ "learning_rate": 0.0006,
+ "loss": 4.721952438354492,
+ "step": 927
+ },
+ {
+ "epoch": 12.894323144104803,
+ "grad_norm": 0.01637161523103714,
+ "learning_rate": 0.0006,
+ "loss": 4.704446792602539,
+ "step": 928
+ },
+ {
+ "epoch": 12.908296943231441,
+ "grad_norm": 0.015221050009131432,
+ "learning_rate": 0.0006,
+ "loss": 4.771907806396484,
+ "step": 929
+ },
+ {
+ "epoch": 12.922270742358078,
+ "grad_norm": 0.016242166981101036,
+ "learning_rate": 0.0006,
+ "loss": 4.806267738342285,
+ "step": 930
+ },
+ {
+ "epoch": 12.936244541484717,
+ "grad_norm": 0.01628357358276844,
+ "learning_rate": 0.0006,
+ "loss": 4.7252326011657715,
+ "step": 931
+ },
+ {
+ "epoch": 12.950218340611354,
+ "grad_norm": 0.017428133636713028,
+ "learning_rate": 0.0006,
+ "loss": 4.765689849853516,
+ "step": 932
+ },
+ {
+ "epoch": 12.96419213973799,
+ "grad_norm": 0.01870061084628105,
+ "learning_rate": 0.0006,
+ "loss": 4.814488887786865,
+ "step": 933
+ },
+ {
+ "epoch": 12.97816593886463,
+ "grad_norm": 0.019796060398221016,
+ "learning_rate": 0.0006,
+ "loss": 4.739185810089111,
+ "step": 934
+ },
+ {
+ "epoch": 12.992139737991266,
+ "grad_norm": 0.01985820196568966,
+ "learning_rate": 0.0006,
+ "loss": 4.810608863830566,
+ "step": 935
+ },
+ {
+ "epoch": 13.0,
+ "grad_norm": 0.020834477618336678,
+ "learning_rate": 0.0006,
+ "loss": 4.727858066558838,
+ "step": 936
+ },
+ {
+ "epoch": 13.0,
+ "eval_loss": 4.979116916656494,
+ "eval_runtime": 57.2877,
+ "eval_samples_per_second": 42.627,
+ "eval_steps_per_second": 1.344,
+ "step": 936
+ },
+ {
+ "epoch": 13.013973799126637,
+ "grad_norm": 0.02025519125163555,
+ "learning_rate": 0.0006,
+ "loss": 4.628875732421875,
+ "step": 937
+ },
+ {
+ "epoch": 13.027947598253276,
+ "grad_norm": 0.019596368074417114,
+ "learning_rate": 0.0006,
+ "loss": 4.694699287414551,
+ "step": 938
+ },
+ {
+ "epoch": 13.041921397379912,
+ "grad_norm": 0.020494934171438217,
+ "learning_rate": 0.0006,
+ "loss": 4.691821098327637,
+ "step": 939
+ },
+ {
+ "epoch": 13.055895196506551,
+ "grad_norm": 0.020121760666370392,
+ "learning_rate": 0.0006,
+ "loss": 4.71478271484375,
+ "step": 940
+ },
+ {
+ "epoch": 13.069868995633188,
+ "grad_norm": 0.021091405302286148,
+ "learning_rate": 0.0006,
+ "loss": 4.695915222167969,
+ "step": 941
+ },
+ {
+ "epoch": 13.083842794759825,
+ "grad_norm": 0.019787423312664032,
+ "learning_rate": 0.0006,
+ "loss": 4.8652448654174805,
+ "step": 942
+ },
+ {
+ "epoch": 13.097816593886463,
+ "grad_norm": 0.0185833927243948,
+ "learning_rate": 0.0006,
+ "loss": 4.7065887451171875,
+ "step": 943
+ },
+ {
+ "epoch": 13.1117903930131,
+ "grad_norm": 0.021220475435256958,
+ "learning_rate": 0.0006,
+ "loss": 4.741232395172119,
+ "step": 944
+ },
+ {
+ "epoch": 13.125764192139737,
+ "grad_norm": 0.02136622555553913,
+ "learning_rate": 0.0006,
+ "loss": 4.705967903137207,
+ "step": 945
+ },
+ {
+ "epoch": 13.139737991266376,
+ "grad_norm": 0.0191587433218956,
+ "learning_rate": 0.0006,
+ "loss": 4.758667945861816,
+ "step": 946
+ },
+ {
+ "epoch": 13.153711790393013,
+ "grad_norm": 0.018290022388100624,
+ "learning_rate": 0.0006,
+ "loss": 4.760666847229004,
+ "step": 947
+ },
+ {
+ "epoch": 13.167685589519651,
+ "grad_norm": 0.021704409271478653,
+ "learning_rate": 0.0006,
+ "loss": 4.750555992126465,
+ "step": 948
+ },
+ {
+ "epoch": 13.181659388646288,
+ "grad_norm": 0.021668918430805206,
+ "learning_rate": 0.0006,
+ "loss": 4.840903282165527,
+ "step": 949
+ },
+ {
+ "epoch": 13.195633187772925,
+ "grad_norm": 0.019497141242027283,
+ "learning_rate": 0.0006,
+ "loss": 4.738969326019287,
+ "step": 950
+ },
+ {
+ "epoch": 13.209606986899564,
+ "grad_norm": 0.018500596284866333,
+ "learning_rate": 0.0006,
+ "loss": 4.75154972076416,
+ "step": 951
+ },
+ {
+ "epoch": 13.2235807860262,
+ "grad_norm": 0.01872239261865616,
+ "learning_rate": 0.0006,
+ "loss": 4.811738014221191,
+ "step": 952
+ },
+ {
+ "epoch": 13.237554585152838,
+ "grad_norm": 0.020321547985076904,
+ "learning_rate": 0.0006,
+ "loss": 4.815384387969971,
+ "step": 953
+ },
+ {
+ "epoch": 13.251528384279476,
+ "grad_norm": 0.02283575013279915,
+ "learning_rate": 0.0006,
+ "loss": 4.74582052230835,
+ "step": 954
+ },
+ {
+ "epoch": 13.265502183406113,
+ "grad_norm": 0.02474398910999298,
+ "learning_rate": 0.0006,
+ "loss": 4.738375186920166,
+ "step": 955
+ },
+ {
+ "epoch": 13.279475982532752,
+ "grad_norm": 0.02065764181315899,
+ "learning_rate": 0.0006,
+ "loss": 4.746875286102295,
+ "step": 956
+ },
+ {
+ "epoch": 13.293449781659389,
+ "grad_norm": 0.020534520968794823,
+ "learning_rate": 0.0006,
+ "loss": 4.656914710998535,
+ "step": 957
+ },
+ {
+ "epoch": 13.307423580786025,
+ "grad_norm": 0.0206378772854805,
+ "learning_rate": 0.0006,
+ "loss": 4.842615604400635,
+ "step": 958
+ },
+ {
+ "epoch": 13.321397379912664,
+ "grad_norm": 0.017935654148459435,
+ "learning_rate": 0.0006,
+ "loss": 4.717827796936035,
+ "step": 959
+ },
+ {
+ "epoch": 13.335371179039301,
+ "grad_norm": 0.016958363354206085,
+ "learning_rate": 0.0006,
+ "loss": 4.681070327758789,
+ "step": 960
+ },
+ {
+ "epoch": 13.34934497816594,
+ "grad_norm": 0.015735412016510963,
+ "learning_rate": 0.0006,
+ "loss": 4.7957048416137695,
+ "step": 961
+ },
+ {
+ "epoch": 13.363318777292577,
+ "grad_norm": 0.01563699170947075,
+ "learning_rate": 0.0006,
+ "loss": 4.701192378997803,
+ "step": 962
+ },
+ {
+ "epoch": 13.377292576419213,
+ "grad_norm": 0.015207760035991669,
+ "learning_rate": 0.0006,
+ "loss": 4.73256254196167,
+ "step": 963
+ },
+ {
+ "epoch": 13.391266375545852,
+ "grad_norm": 0.015396883711218834,
+ "learning_rate": 0.0006,
+ "loss": 4.76412296295166,
+ "step": 964
+ },
+ {
+ "epoch": 13.405240174672489,
+ "grad_norm": 0.015690583735704422,
+ "learning_rate": 0.0006,
+ "loss": 4.791157245635986,
+ "step": 965
+ },
+ {
+ "epoch": 13.419213973799126,
+ "grad_norm": 0.01755724661052227,
+ "learning_rate": 0.0006,
+ "loss": 4.845666885375977,
+ "step": 966
+ },
+ {
+ "epoch": 13.433187772925764,
+ "grad_norm": 0.019443422555923462,
+ "learning_rate": 0.0006,
+ "loss": 4.5902299880981445,
+ "step": 967
+ },
+ {
+ "epoch": 13.447161572052401,
+ "grad_norm": 0.02118338644504547,
+ "learning_rate": 0.0006,
+ "loss": 4.740627288818359,
+ "step": 968
+ },
+ {
+ "epoch": 13.46113537117904,
+ "grad_norm": 0.02174839936196804,
+ "learning_rate": 0.0006,
+ "loss": 4.814560890197754,
+ "step": 969
+ },
+ {
+ "epoch": 13.475109170305677,
+ "grad_norm": 0.022860374301671982,
+ "learning_rate": 0.0006,
+ "loss": 4.66440486907959,
+ "step": 970
+ },
+ {
+ "epoch": 13.489082969432314,
+ "grad_norm": 0.024742677807807922,
+ "learning_rate": 0.0006,
+ "loss": 4.7451653480529785,
+ "step": 971
+ },
+ {
+ "epoch": 13.503056768558952,
+ "grad_norm": 0.027068907395005226,
+ "learning_rate": 0.0006,
+ "loss": 4.802584171295166,
+ "step": 972
+ },
+ {
+ "epoch": 13.51703056768559,
+ "grad_norm": 0.026316625997424126,
+ "learning_rate": 0.0006,
+ "loss": 4.842397689819336,
+ "step": 973
+ },
+ {
+ "epoch": 13.531004366812226,
+ "grad_norm": 0.024610962718725204,
+ "learning_rate": 0.0006,
+ "loss": 4.650721073150635,
+ "step": 974
+ },
+ {
+ "epoch": 13.544978165938865,
+ "grad_norm": 0.024122603237628937,
+ "learning_rate": 0.0006,
+ "loss": 4.6984357833862305,
+ "step": 975
+ },
+ {
+ "epoch": 13.558951965065502,
+ "grad_norm": 0.022699708119034767,
+ "learning_rate": 0.0006,
+ "loss": 4.793007850646973,
+ "step": 976
+ },
+ {
+ "epoch": 13.57292576419214,
+ "grad_norm": 0.021706534549593925,
+ "learning_rate": 0.0006,
+ "loss": 4.691362380981445,
+ "step": 977
+ },
+ {
+ "epoch": 13.586899563318777,
+ "grad_norm": 0.021363291889429092,
+ "learning_rate": 0.0006,
+ "loss": 4.76762580871582,
+ "step": 978
+ },
+ {
+ "epoch": 13.600873362445414,
+ "grad_norm": 0.023476677015423775,
+ "learning_rate": 0.0006,
+ "loss": 4.777202606201172,
+ "step": 979
+ },
+ {
+ "epoch": 13.614847161572053,
+ "grad_norm": 0.02457100711762905,
+ "learning_rate": 0.0006,
+ "loss": 4.590703010559082,
+ "step": 980
+ },
+ {
+ "epoch": 13.62882096069869,
+ "grad_norm": 0.02469099499285221,
+ "learning_rate": 0.0006,
+ "loss": 4.640185356140137,
+ "step": 981
+ },
+ {
+ "epoch": 13.642794759825328,
+ "grad_norm": 0.024914603680372238,
+ "learning_rate": 0.0006,
+ "loss": 4.772201061248779,
+ "step": 982
+ },
+ {
+ "epoch": 13.656768558951965,
+ "grad_norm": 0.025133173912763596,
+ "learning_rate": 0.0006,
+ "loss": 4.743337154388428,
+ "step": 983
+ },
+ {
+ "epoch": 13.670742358078602,
+ "grad_norm": 0.02034137211740017,
+ "learning_rate": 0.0006,
+ "loss": 4.752127647399902,
+ "step": 984
+ },
+ {
+ "epoch": 13.68471615720524,
+ "grad_norm": 0.019645661115646362,
+ "learning_rate": 0.0006,
+ "loss": 4.838453769683838,
+ "step": 985
+ },
+ {
+ "epoch": 13.698689956331878,
+ "grad_norm": 0.01900297962129116,
+ "learning_rate": 0.0006,
+ "loss": 4.804868221282959,
+ "step": 986
+ },
+ {
+ "epoch": 13.712663755458514,
+ "grad_norm": 0.020292341709136963,
+ "learning_rate": 0.0006,
+ "loss": 4.609743118286133,
+ "step": 987
+ },
+ {
+ "epoch": 13.726637554585153,
+ "grad_norm": 0.017014775425195694,
+ "learning_rate": 0.0006,
+ "loss": 4.740612983703613,
+ "step": 988
+ },
+ {
+ "epoch": 13.74061135371179,
+ "grad_norm": 0.015199844725430012,
+ "learning_rate": 0.0006,
+ "loss": 4.719701766967773,
+ "step": 989
+ },
+ {
+ "epoch": 13.754585152838429,
+ "grad_norm": 0.015125414356589317,
+ "learning_rate": 0.0006,
+ "loss": 4.686114311218262,
+ "step": 990
+ },
+ {
+ "epoch": 13.768558951965066,
+ "grad_norm": 0.014779540710151196,
+ "learning_rate": 0.0006,
+ "loss": 4.674347877502441,
+ "step": 991
+ },
+ {
+ "epoch": 13.782532751091702,
+ "grad_norm": 0.015255163423717022,
+ "learning_rate": 0.0006,
+ "loss": 4.6309309005737305,
+ "step": 992
+ },
+ {
+ "epoch": 13.796506550218341,
+ "grad_norm": 0.013692197389900684,
+ "learning_rate": 0.0006,
+ "loss": 4.698497295379639,
+ "step": 993
+ },
+ {
+ "epoch": 13.810480349344978,
+ "grad_norm": 0.014529623091220856,
+ "learning_rate": 0.0006,
+ "loss": 4.628443717956543,
+ "step": 994
+ },
+ {
+ "epoch": 13.824454148471617,
+ "grad_norm": 0.01576659083366394,
+ "learning_rate": 0.0006,
+ "loss": 4.788015842437744,
+ "step": 995
+ },
+ {
+ "epoch": 13.838427947598253,
+ "grad_norm": 0.015834277495741844,
+ "learning_rate": 0.0006,
+ "loss": 4.672703742980957,
+ "step": 996
+ },
+ {
+ "epoch": 13.85240174672489,
+ "grad_norm": 0.015539503656327724,
+ "learning_rate": 0.0006,
+ "loss": 4.731503486633301,
+ "step": 997
+ },
+ {
+ "epoch": 13.866375545851529,
+ "grad_norm": 0.013463746756315231,
+ "learning_rate": 0.0006,
+ "loss": 4.7006330490112305,
+ "step": 998
+ },
+ {
+ "epoch": 13.880349344978166,
+ "grad_norm": 0.016377726569771767,
+ "learning_rate": 0.0006,
+ "loss": 4.636074066162109,
+ "step": 999
+ },
+ {
+ "epoch": 13.894323144104803,
+ "grad_norm": 0.016660034656524658,
+ "learning_rate": 0.0006,
+ "loss": 4.742693901062012,
+ "step": 1000
+ },
+ {
+ "epoch": 13.908296943231441,
+ "grad_norm": 0.014902938157320023,
+ "learning_rate": 0.0006,
+ "loss": 4.769811630249023,
+ "step": 1001
+ },
+ {
+ "epoch": 13.922270742358078,
+ "grad_norm": 0.014883476309478283,
+ "learning_rate": 0.0006,
+ "loss": 4.731159210205078,
+ "step": 1002
+ },
+ {
+ "epoch": 13.936244541484717,
+ "grad_norm": 0.014494217932224274,
+ "learning_rate": 0.0006,
+ "loss": 4.659485816955566,
+ "step": 1003
+ },
+ {
+ "epoch": 13.950218340611354,
+ "grad_norm": 0.015426991507411003,
+ "learning_rate": 0.0006,
+ "loss": 4.755229949951172,
+ "step": 1004
+ },
+ {
+ "epoch": 13.96419213973799,
+ "grad_norm": 0.016648586839437485,
+ "learning_rate": 0.0006,
+ "loss": 4.74111270904541,
+ "step": 1005
+ },
+ {
+ "epoch": 13.97816593886463,
+ "grad_norm": 0.016980541869997978,
+ "learning_rate": 0.0006,
+ "loss": 4.74072265625,
+ "step": 1006
+ },
+ {
+ "epoch": 13.992139737991266,
+ "grad_norm": 0.018070152029395103,
+ "learning_rate": 0.0006,
+ "loss": 4.732975006103516,
+ "step": 1007
+ },
+ {
+ "epoch": 14.0,
+ "grad_norm": 0.019910665228962898,
+ "learning_rate": 0.0006,
+ "loss": 4.634697914123535,
+ "step": 1008
+ },
+ {
+ "epoch": 14.0,
+ "eval_loss": 4.850640296936035,
+ "eval_runtime": 57.3039,
+ "eval_samples_per_second": 42.615,
+ "eval_steps_per_second": 1.344,
+ "step": 1008
+ },
+ {
+ "epoch": 14.013973799126637,
+ "grad_norm": 0.01980053074657917,
+ "learning_rate": 0.0006,
+ "loss": 4.628744602203369,
+ "step": 1009
+ },
+ {
+ "epoch": 14.027947598253276,
+ "grad_norm": 0.02497623674571514,
+ "learning_rate": 0.0006,
+ "loss": 4.75930118560791,
+ "step": 1010
+ },
+ {
+ "epoch": 14.041921397379912,
+ "grad_norm": 0.02794097363948822,
+ "learning_rate": 0.0006,
+ "loss": 4.691340446472168,
+ "step": 1011
+ },
+ {
+ "epoch": 14.055895196506551,
+ "grad_norm": 0.023725535720586777,
+ "learning_rate": 0.0006,
+ "loss": 4.733664512634277,
+ "step": 1012
+ },
+ {
+ "epoch": 14.069868995633188,
+ "grad_norm": 0.022966474294662476,
+ "learning_rate": 0.0006,
+ "loss": 4.664808750152588,
+ "step": 1013
+ },
+ {
+ "epoch": 14.083842794759825,
+ "grad_norm": 0.02530563250184059,
+ "learning_rate": 0.0006,
+ "loss": 4.8047285079956055,
+ "step": 1014
+ },
+ {
+ "epoch": 14.097816593886463,
+ "grad_norm": 0.02491082064807415,
+ "learning_rate": 0.0006,
+ "loss": 4.667956352233887,
+ "step": 1015
+ },
+ {
+ "epoch": 14.1117903930131,
+ "grad_norm": 0.022110294550657272,
+ "learning_rate": 0.0006,
+ "loss": 4.666140556335449,
+ "step": 1016
+ },
+ {
+ "epoch": 14.125764192139737,
+ "grad_norm": 0.02065849117934704,
+ "learning_rate": 0.0006,
+ "loss": 4.581960678100586,
+ "step": 1017
+ },
+ {
+ "epoch": 14.139737991266376,
+ "grad_norm": 0.01941298507153988,
+ "learning_rate": 0.0006,
+ "loss": 4.614226341247559,
+ "step": 1018
+ },
+ {
+ "epoch": 14.153711790393013,
+ "grad_norm": 0.022441856563091278,
+ "learning_rate": 0.0006,
+ "loss": 4.638352394104004,
+ "step": 1019
+ },
+ {
+ "epoch": 14.167685589519651,
+ "grad_norm": 0.021519694477319717,
+ "learning_rate": 0.0006,
+ "loss": 4.700160980224609,
+ "step": 1020
+ },
+ {
+ "epoch": 14.181659388646288,
+ "grad_norm": 0.018956486135721207,
+ "learning_rate": 0.0006,
+ "loss": 4.649593353271484,
+ "step": 1021
+ },
+ {
+ "epoch": 14.195633187772925,
+ "grad_norm": 0.016681140288710594,
+ "learning_rate": 0.0006,
+ "loss": 4.603952407836914,
+ "step": 1022
+ },
+ {
+ "epoch": 14.209606986899564,
+ "grad_norm": 0.01788809522986412,
+ "learning_rate": 0.0006,
+ "loss": 4.669160842895508,
+ "step": 1023
+ },
+ {
+ "epoch": 14.2235807860262,
+ "grad_norm": 0.016047228127717972,
+ "learning_rate": 0.0006,
+ "loss": 4.6367950439453125,
+ "step": 1024
+ },
+ {
+ "epoch": 14.237554585152838,
+ "grad_norm": 0.01614239625632763,
+ "learning_rate": 0.0006,
+ "loss": 4.664529323577881,
+ "step": 1025
+ },
+ {
+ "epoch": 14.251528384279476,
+ "grad_norm": 0.016362549737095833,
+ "learning_rate": 0.0006,
+ "loss": 4.702373504638672,
+ "step": 1026
+ },
+ {
+ "epoch": 14.265502183406113,
+ "grad_norm": 0.015800409018993378,
+ "learning_rate": 0.0006,
+ "loss": 4.7594428062438965,
+ "step": 1027
+ },
+ {
+ "epoch": 14.279475982532752,
+ "grad_norm": 0.016730139032006264,
+ "learning_rate": 0.0006,
+ "loss": 4.677244186401367,
+ "step": 1028
+ },
+ {
+ "epoch": 14.293449781659389,
+ "grad_norm": 0.01661759801208973,
+ "learning_rate": 0.0006,
+ "loss": 4.633679389953613,
+ "step": 1029
+ },
+ {
+ "epoch": 14.307423580786025,
+ "grad_norm": 0.017673347145318985,
+ "learning_rate": 0.0006,
+ "loss": 4.649107456207275,
+ "step": 1030
+ },
+ {
+ "epoch": 14.321397379912664,
+ "grad_norm": 0.019243352115154266,
+ "learning_rate": 0.0006,
+ "loss": 4.611257553100586,
+ "step": 1031
+ },
+ {
+ "epoch": 14.335371179039301,
+ "grad_norm": 0.019817164167761803,
+ "learning_rate": 0.0006,
+ "loss": 4.588086128234863,
+ "step": 1032
+ },
+ {
+ "epoch": 14.34934497816594,
+ "grad_norm": 0.022292211651802063,
+ "learning_rate": 0.0006,
+ "loss": 4.783066749572754,
+ "step": 1033
+ },
+ {
+ "epoch": 14.363318777292577,
+ "grad_norm": 0.02140096202492714,
+ "learning_rate": 0.0006,
+ "loss": 4.604718208312988,
+ "step": 1034
+ },
+ {
+ "epoch": 14.377292576419213,
+ "grad_norm": 0.0187817495316267,
+ "learning_rate": 0.0006,
+ "loss": 4.721773147583008,
+ "step": 1035
+ },
+ {
+ "epoch": 14.391266375545852,
+ "grad_norm": 0.019879210740327835,
+ "learning_rate": 0.0006,
+ "loss": 4.789286136627197,
+ "step": 1036
+ },
+ {
+ "epoch": 14.405240174672489,
+ "grad_norm": 0.01885785162448883,
+ "learning_rate": 0.0006,
+ "loss": 4.669743537902832,
+ "step": 1037
+ },
+ {
+ "epoch": 14.419213973799126,
+ "grad_norm": 0.018324270844459534,
+ "learning_rate": 0.0006,
+ "loss": 4.65901517868042,
+ "step": 1038
+ },
+ {
+ "epoch": 14.433187772925764,
+ "grad_norm": 0.017664821818470955,
+ "learning_rate": 0.0006,
+ "loss": 4.667839527130127,
+ "step": 1039
+ },
+ {
+ "epoch": 14.447161572052401,
+ "grad_norm": 0.015702905133366585,
+ "learning_rate": 0.0006,
+ "loss": 4.64700984954834,
+ "step": 1040
+ },
+ {
+ "epoch": 14.46113537117904,
+ "grad_norm": 0.016674194484949112,
+ "learning_rate": 0.0006,
+ "loss": 4.694093227386475,
+ "step": 1041
+ },
+ {
+ "epoch": 14.475109170305677,
+ "grad_norm": 0.016869477927684784,
+ "learning_rate": 0.0006,
+ "loss": 4.619544982910156,
+ "step": 1042
+ },
+ {
+ "epoch": 14.489082969432314,
+ "grad_norm": 0.016557540744543076,
+ "learning_rate": 0.0006,
+ "loss": 4.568667411804199,
+ "step": 1043
+ },
+ {
+ "epoch": 14.503056768558952,
+ "grad_norm": 0.014500945806503296,
+ "learning_rate": 0.0006,
+ "loss": 4.561741828918457,
+ "step": 1044
+ },
+ {
+ "epoch": 14.51703056768559,
+ "grad_norm": 0.014228739775717258,
+ "learning_rate": 0.0006,
+ "loss": 4.654277324676514,
+ "step": 1045
+ },
+ {
+ "epoch": 14.531004366812226,
+ "grad_norm": 0.013894238509237766,
+ "learning_rate": 0.0006,
+ "loss": 4.5867719650268555,
+ "step": 1046
+ },
+ {
+ "epoch": 14.544978165938865,
+ "grad_norm": 0.01550068985670805,
+ "learning_rate": 0.0006,
+ "loss": 4.61168098449707,
+ "step": 1047
+ },
+ {
+ "epoch": 14.558951965065502,
+ "grad_norm": 0.015940772369503975,
+ "learning_rate": 0.0006,
+ "loss": 4.691247940063477,
+ "step": 1048
+ },
+ {
+ "epoch": 14.57292576419214,
+ "grad_norm": 0.01714535616338253,
+ "learning_rate": 0.0006,
+ "loss": 4.61183500289917,
+ "step": 1049
+ },
+ {
+ "epoch": 14.586899563318777,
+ "grad_norm": 0.017925038933753967,
+ "learning_rate": 0.0006,
+ "loss": 4.719022750854492,
+ "step": 1050
+ },
+ {
+ "epoch": 14.600873362445414,
+ "grad_norm": 0.018141552805900574,
+ "learning_rate": 0.0006,
+ "loss": 4.760281562805176,
+ "step": 1051
+ },
+ {
+ "epoch": 14.614847161572053,
+ "grad_norm": 0.017131328582763672,
+ "learning_rate": 0.0006,
+ "loss": 4.709680080413818,
+ "step": 1052
+ },
+ {
+ "epoch": 14.62882096069869,
+ "grad_norm": 0.018730048090219498,
+ "learning_rate": 0.0006,
+ "loss": 4.5994720458984375,
+ "step": 1053
+ },
+ {
+ "epoch": 14.642794759825328,
+ "grad_norm": 0.018894298002123833,
+ "learning_rate": 0.0006,
+ "loss": 4.669070720672607,
+ "step": 1054
+ },
+ {
+ "epoch": 14.656768558951965,
+ "grad_norm": 0.017643198370933533,
+ "learning_rate": 0.0006,
+ "loss": 4.641790390014648,
+ "step": 1055
+ },
+ {
+ "epoch": 14.670742358078602,
+ "grad_norm": 0.017647244036197662,
+ "learning_rate": 0.0006,
+ "loss": 4.653899192810059,
+ "step": 1056
+ },
+ {
+ "epoch": 14.68471615720524,
+ "grad_norm": 0.017304198816418648,
+ "learning_rate": 0.0006,
+ "loss": 4.7487473487854,
+ "step": 1057
+ },
+ {
+ "epoch": 14.698689956331878,
+ "grad_norm": 0.01638447679579258,
+ "learning_rate": 0.0006,
+ "loss": 4.71945858001709,
+ "step": 1058
+ },
+ {
+ "epoch": 14.712663755458514,
+ "grad_norm": 0.015892071649432182,
+ "learning_rate": 0.0006,
+ "loss": 4.744903564453125,
+ "step": 1059
+ },
+ {
+ "epoch": 14.726637554585153,
+ "grad_norm": 0.017420614138245583,
+ "learning_rate": 0.0006,
+ "loss": 4.781563758850098,
+ "step": 1060
+ },
+ {
+ "epoch": 14.74061135371179,
+ "grad_norm": 0.016498520970344543,
+ "learning_rate": 0.0006,
+ "loss": 4.622779846191406,
+ "step": 1061
+ },
+ {
+ "epoch": 14.754585152838429,
+ "grad_norm": 0.017482127994298935,
+ "learning_rate": 0.0006,
+ "loss": 4.62825870513916,
+ "step": 1062
+ },
+ {
+ "epoch": 14.768558951965066,
+ "grad_norm": 0.019018379971385002,
+ "learning_rate": 0.0006,
+ "loss": 4.693194389343262,
+ "step": 1063
+ },
+ {
+ "epoch": 14.782532751091702,
+ "grad_norm": 0.01993914321064949,
+ "learning_rate": 0.0006,
+ "loss": 4.742903709411621,
+ "step": 1064
+ },
+ {
+ "epoch": 14.796506550218341,
+ "grad_norm": 0.02301335707306862,
+ "learning_rate": 0.0006,
+ "loss": 4.613251686096191,
+ "step": 1065
+ },
+ {
+ "epoch": 14.810480349344978,
+ "grad_norm": 0.024088222533464432,
+ "learning_rate": 0.0006,
+ "loss": 4.594305992126465,
+ "step": 1066
+ },
+ {
+ "epoch": 14.824454148471617,
+ "grad_norm": 0.02493320405483246,
+ "learning_rate": 0.0006,
+ "loss": 4.700588226318359,
+ "step": 1067
+ },
+ {
+ "epoch": 14.838427947598253,
+ "grad_norm": 0.02501937560737133,
+ "learning_rate": 0.0006,
+ "loss": 4.601742744445801,
+ "step": 1068
+ },
+ {
+ "epoch": 14.85240174672489,
+ "grad_norm": 0.023490751162171364,
+ "learning_rate": 0.0006,
+ "loss": 4.589211463928223,
+ "step": 1069
+ },
+ {
+ "epoch": 14.866375545851529,
+ "grad_norm": 0.019695665687322617,
+ "learning_rate": 0.0006,
+ "loss": 4.635081768035889,
+ "step": 1070
+ },
+ {
+ "epoch": 14.880349344978166,
+ "grad_norm": 0.019276980310678482,
+ "learning_rate": 0.0006,
+ "loss": 4.737002372741699,
+ "step": 1071
+ },
+ {
+ "epoch": 14.894323144104803,
+ "grad_norm": 0.022430511191487312,
+ "learning_rate": 0.0006,
+ "loss": 4.808295726776123,
+ "step": 1072
+ },
+ {
+ "epoch": 14.908296943231441,
+ "grad_norm": 0.021543916314840317,
+ "learning_rate": 0.0006,
+ "loss": 4.559475421905518,
+ "step": 1073
+ },
+ {
+ "epoch": 14.922270742358078,
+ "grad_norm": 0.019885241985321045,
+ "learning_rate": 0.0006,
+ "loss": 4.647948741912842,
+ "step": 1074
+ },
+ {
+ "epoch": 14.936244541484717,
+ "grad_norm": 0.02115175500512123,
+ "learning_rate": 0.0006,
+ "loss": 4.680505752563477,
+ "step": 1075
+ },
+ {
+ "epoch": 14.950218340611354,
+ "grad_norm": 0.02156151458621025,
+ "learning_rate": 0.0006,
+ "loss": 4.673477649688721,
+ "step": 1076
+ },
+ {
+ "epoch": 14.96419213973799,
+ "grad_norm": 0.02184601128101349,
+ "learning_rate": 0.0006,
+ "loss": 4.644402027130127,
+ "step": 1077
+ },
+ {
+ "epoch": 14.97816593886463,
+ "grad_norm": 0.0211983360350132,
+ "learning_rate": 0.0006,
+ "loss": 4.821558475494385,
+ "step": 1078
+ },
+ {
+ "epoch": 14.992139737991266,
+ "grad_norm": 0.019636567682027817,
+ "learning_rate": 0.0006,
+ "loss": 4.578868865966797,
+ "step": 1079
+ },
+ {
+ "epoch": 15.0,
+ "grad_norm": 0.019997350871562958,
+ "learning_rate": 0.0006,
+ "loss": 4.691307544708252,
+ "step": 1080
+ },
+ {
+ "epoch": 15.0,
+ "eval_loss": 4.872478008270264,
+ "eval_runtime": 56.8513,
+ "eval_samples_per_second": 42.954,
+ "eval_steps_per_second": 1.354,
+ "step": 1080
+ },
+ {
+ "epoch": 15.013973799126637,
+ "grad_norm": 0.018830662593245506,
+ "learning_rate": 0.0006,
+ "loss": 4.761779308319092,
+ "step": 1081
+ },
+ {
+ "epoch": 15.027947598253276,
+ "grad_norm": 0.016876481473445892,
+ "learning_rate": 0.0006,
+ "loss": 4.61148738861084,
+ "step": 1082
+ },
+ {
+ "epoch": 15.041921397379912,
+ "grad_norm": 0.015550863929092884,
+ "learning_rate": 0.0006,
+ "loss": 4.676124095916748,
+ "step": 1083
+ },
+ {
+ "epoch": 15.055895196506551,
+ "grad_norm": 0.017178896814584732,
+ "learning_rate": 0.0006,
+ "loss": 4.588935852050781,
+ "step": 1084
+ },
+ {
+ "epoch": 15.069868995633188,
+ "grad_norm": 0.015465447679162025,
+ "learning_rate": 0.0006,
+ "loss": 4.605499267578125,
+ "step": 1085
+ },
+ {
+ "epoch": 15.083842794759825,
+ "grad_norm": 0.01606195978820324,
+ "learning_rate": 0.0006,
+ "loss": 4.723373889923096,
+ "step": 1086
+ },
+ {
+ "epoch": 15.097816593886463,
+ "grad_norm": 0.016521936282515526,
+ "learning_rate": 0.0006,
+ "loss": 4.646608352661133,
+ "step": 1087
+ },
+ {
+ "epoch": 15.1117903930131,
+ "grad_norm": 0.01589205674827099,
+ "learning_rate": 0.0006,
+ "loss": 4.498138427734375,
+ "step": 1088
+ },
+ {
+ "epoch": 15.125764192139737,
+ "grad_norm": 0.01524856686592102,
+ "learning_rate": 0.0006,
+ "loss": 4.594562530517578,
+ "step": 1089
+ },
+ {
+ "epoch": 15.139737991266376,
+ "grad_norm": 0.014714641496539116,
+ "learning_rate": 0.0006,
+ "loss": 4.630744934082031,
+ "step": 1090
+ },
+ {
+ "epoch": 15.153711790393013,
+ "grad_norm": 0.015333331190049648,
+ "learning_rate": 0.0006,
+ "loss": 4.682388782501221,
+ "step": 1091
+ },
+ {
+ "epoch": 15.167685589519651,
+ "grad_norm": 0.01550073828548193,
+ "learning_rate": 0.0006,
+ "loss": 4.692652702331543,
+ "step": 1092
+ },
+ {
+ "epoch": 15.181659388646288,
+ "grad_norm": 0.014783318154513836,
+ "learning_rate": 0.0006,
+ "loss": 4.6223835945129395,
+ "step": 1093
+ },
+ {
+ "epoch": 15.195633187772925,
+ "grad_norm": 0.014372751116752625,
+ "learning_rate": 0.0006,
+ "loss": 4.52687406539917,
+ "step": 1094
+ },
+ {
+ "epoch": 15.209606986899564,
+ "grad_norm": 0.014779130928218365,
+ "learning_rate": 0.0006,
+ "loss": 4.637009620666504,
+ "step": 1095
+ },
+ {
+ "epoch": 15.2235807860262,
+ "grad_norm": 0.015697073191404343,
+ "learning_rate": 0.0006,
+ "loss": 4.682981014251709,
+ "step": 1096
+ },
+ {
+ "epoch": 15.237554585152838,
+ "grad_norm": 0.016523664817214012,
+ "learning_rate": 0.0006,
+ "loss": 4.705722808837891,
+ "step": 1097
+ },
+ {
+ "epoch": 15.251528384279476,
+ "grad_norm": 0.017633996903896332,
+ "learning_rate": 0.0006,
+ "loss": 4.574686050415039,
+ "step": 1098
+ },
+ {
+ "epoch": 15.265502183406113,
+ "grad_norm": 0.019955076277256012,
+ "learning_rate": 0.0006,
+ "loss": 4.634450912475586,
+ "step": 1099
+ },
+ {
+ "epoch": 15.279475982532752,
+ "grad_norm": 0.02132618986070156,
+ "learning_rate": 0.0006,
+ "loss": 4.56110143661499,
+ "step": 1100
+ },
+ {
+ "epoch": 15.293449781659389,
+ "grad_norm": 0.0207593385130167,
+ "learning_rate": 0.0006,
+ "loss": 4.629396438598633,
+ "step": 1101
+ },
+ {
+ "epoch": 15.307423580786025,
+ "grad_norm": 0.017882540822029114,
+ "learning_rate": 0.0006,
+ "loss": 4.730963230133057,
+ "step": 1102
+ },
+ {
+ "epoch": 15.321397379912664,
+ "grad_norm": 0.01917876861989498,
+ "learning_rate": 0.0006,
+ "loss": 4.64585542678833,
+ "step": 1103
+ },
+ {
+ "epoch": 15.335371179039301,
+ "grad_norm": 0.01943974196910858,
+ "learning_rate": 0.0006,
+ "loss": 4.656181335449219,
+ "step": 1104
+ },
+ {
+ "epoch": 15.34934497816594,
+ "grad_norm": 0.019904715940356255,
+ "learning_rate": 0.0006,
+ "loss": 4.619329452514648,
+ "step": 1105
+ },
+ {
+ "epoch": 15.363318777292577,
+ "grad_norm": 0.019174346700310707,
+ "learning_rate": 0.0006,
+ "loss": 4.647892475128174,
+ "step": 1106
+ },
+ {
+ "epoch": 15.377292576419213,
+ "grad_norm": 0.018244758248329163,
+ "learning_rate": 0.0006,
+ "loss": 4.56166934967041,
+ "step": 1107
+ },
+ {
+ "epoch": 15.391266375545852,
+ "grad_norm": 0.018914205953478813,
+ "learning_rate": 0.0006,
+ "loss": 4.648063659667969,
+ "step": 1108
+ },
+ {
+ "epoch": 15.405240174672489,
+ "grad_norm": 0.01961071416735649,
+ "learning_rate": 0.0006,
+ "loss": 4.665525436401367,
+ "step": 1109
+ },
+ {
+ "epoch": 15.419213973799126,
+ "grad_norm": 0.01841311901807785,
+ "learning_rate": 0.0006,
+ "loss": 4.611342430114746,
+ "step": 1110
+ },
+ {
+ "epoch": 15.433187772925764,
+ "grad_norm": 0.01863427273929119,
+ "learning_rate": 0.0006,
+ "loss": 4.664256572723389,
+ "step": 1111
+ },
+ {
+ "epoch": 15.447161572052401,
+ "grad_norm": 0.020506877452135086,
+ "learning_rate": 0.0006,
+ "loss": 4.631237506866455,
+ "step": 1112
+ },
+ {
+ "epoch": 15.46113537117904,
+ "grad_norm": 0.02223833277821541,
+ "learning_rate": 0.0006,
+ "loss": 4.668654441833496,
+ "step": 1113
+ },
+ {
+ "epoch": 15.475109170305677,
+ "grad_norm": 0.023336702957749367,
+ "learning_rate": 0.0006,
+ "loss": 4.562833309173584,
+ "step": 1114
+ },
+ {
+ "epoch": 15.489082969432314,
+ "grad_norm": 0.023052413016557693,
+ "learning_rate": 0.0006,
+ "loss": 4.621532917022705,
+ "step": 1115
+ },
+ {
+ "epoch": 15.503056768558952,
+ "grad_norm": 0.023420924320816994,
+ "learning_rate": 0.0006,
+ "loss": 4.665465354919434,
+ "step": 1116
+ },
+ {
+ "epoch": 15.51703056768559,
+ "grad_norm": 0.024071281775832176,
+ "learning_rate": 0.0006,
+ "loss": 4.595980644226074,
+ "step": 1117
+ },
+ {
+ "epoch": 15.531004366812226,
+ "grad_norm": 0.023406682536005974,
+ "learning_rate": 0.0006,
+ "loss": 4.642635345458984,
+ "step": 1118
+ },
+ {
+ "epoch": 15.544978165938865,
+ "grad_norm": 0.02138647809624672,
+ "learning_rate": 0.0006,
+ "loss": 4.704513072967529,
+ "step": 1119
+ },
+ {
+ "epoch": 15.558951965065502,
+ "grad_norm": 0.0221384447067976,
+ "learning_rate": 0.0006,
+ "loss": 4.567503452301025,
+ "step": 1120
+ },
+ {
+ "epoch": 15.57292576419214,
+ "grad_norm": 0.020879851654171944,
+ "learning_rate": 0.0006,
+ "loss": 4.664057731628418,
+ "step": 1121
+ },
+ {
+ "epoch": 15.586899563318777,
+ "grad_norm": 0.021478967741131783,
+ "learning_rate": 0.0006,
+ "loss": 4.539627552032471,
+ "step": 1122
+ },
+ {
+ "epoch": 15.600873362445414,
+ "grad_norm": 0.022748656570911407,
+ "learning_rate": 0.0006,
+ "loss": 4.5654497146606445,
+ "step": 1123
+ },
+ {
+ "epoch": 15.614847161572053,
+ "grad_norm": 0.021870248019695282,
+ "learning_rate": 0.0006,
+ "loss": 4.56899356842041,
+ "step": 1124
+ },
+ {
+ "epoch": 15.62882096069869,
+ "grad_norm": 0.021711355075240135,
+ "learning_rate": 0.0006,
+ "loss": 4.657808303833008,
+ "step": 1125
+ },
+ {
+ "epoch": 15.642794759825328,
+ "grad_norm": 0.022810909897089005,
+ "learning_rate": 0.0006,
+ "loss": 4.6526689529418945,
+ "step": 1126
+ },
+ {
+ "epoch": 15.656768558951965,
+ "grad_norm": 0.022561343386769295,
+ "learning_rate": 0.0006,
+ "loss": 4.534887790679932,
+ "step": 1127
+ },
+ {
+ "epoch": 15.670742358078602,
+ "grad_norm": 0.019909802824258804,
+ "learning_rate": 0.0006,
+ "loss": 4.600790023803711,
+ "step": 1128
+ },
+ {
+ "epoch": 15.68471615720524,
+ "grad_norm": 0.020092325285077095,
+ "learning_rate": 0.0006,
+ "loss": 4.773830413818359,
+ "step": 1129
+ },
+ {
+ "epoch": 15.698689956331878,
+ "grad_norm": 0.018834445625543594,
+ "learning_rate": 0.0006,
+ "loss": 4.628868103027344,
+ "step": 1130
+ },
+ {
+ "epoch": 15.712663755458514,
+ "grad_norm": 0.017552798613905907,
+ "learning_rate": 0.0006,
+ "loss": 4.647500038146973,
+ "step": 1131
+ },
+ {
+ "epoch": 15.726637554585153,
+ "grad_norm": 0.01751251146197319,
+ "learning_rate": 0.0006,
+ "loss": 4.590545654296875,
+ "step": 1132
+ },
+ {
+ "epoch": 15.74061135371179,
+ "grad_norm": 0.01699099875986576,
+ "learning_rate": 0.0006,
+ "loss": 4.509699821472168,
+ "step": 1133
+ },
+ {
+ "epoch": 15.754585152838429,
+ "grad_norm": 0.015808649361133575,
+ "learning_rate": 0.0006,
+ "loss": 4.606992244720459,
+ "step": 1134
+ },
+ {
+ "epoch": 15.768558951965066,
+ "grad_norm": 0.01655443385243416,
+ "learning_rate": 0.0006,
+ "loss": 4.585219860076904,
+ "step": 1135
+ },
+ {
+ "epoch": 15.782532751091702,
+ "grad_norm": 0.01708163693547249,
+ "learning_rate": 0.0006,
+ "loss": 4.668147087097168,
+ "step": 1136
+ },
+ {
+ "epoch": 15.796506550218341,
+ "grad_norm": 0.017161913216114044,
+ "learning_rate": 0.0006,
+ "loss": 4.7170634269714355,
+ "step": 1137
+ },
+ {
+ "epoch": 15.810480349344978,
+ "grad_norm": 0.017262674868106842,
+ "learning_rate": 0.0006,
+ "loss": 4.572132110595703,
+ "step": 1138
+ },
+ {
+ "epoch": 15.824454148471617,
+ "grad_norm": 0.017713576555252075,
+ "learning_rate": 0.0006,
+ "loss": 4.702538967132568,
+ "step": 1139
+ },
+ {
+ "epoch": 15.838427947598253,
+ "grad_norm": 0.018730733543634415,
+ "learning_rate": 0.0006,
+ "loss": 4.629721641540527,
+ "step": 1140
+ },
+ {
+ "epoch": 15.85240174672489,
+ "grad_norm": 0.017960187047719955,
+ "learning_rate": 0.0006,
+ "loss": 4.661039352416992,
+ "step": 1141
+ },
+ {
+ "epoch": 15.866375545851529,
+ "grad_norm": 0.01750079356133938,
+ "learning_rate": 0.0006,
+ "loss": 4.503710746765137,
+ "step": 1142
+ },
+ {
+ "epoch": 15.880349344978166,
+ "grad_norm": 0.016630422323942184,
+ "learning_rate": 0.0006,
+ "loss": 4.6582136154174805,
+ "step": 1143
+ },
+ {
+ "epoch": 15.894323144104803,
+ "grad_norm": 0.015896832570433617,
+ "learning_rate": 0.0006,
+ "loss": 4.596693992614746,
+ "step": 1144
+ },
+ {
+ "epoch": 15.908296943231441,
+ "grad_norm": 0.01566407084465027,
+ "learning_rate": 0.0006,
+ "loss": 4.6337199211120605,
+ "step": 1145
+ },
+ {
+ "epoch": 15.922270742358078,
+ "grad_norm": 0.016823330894112587,
+ "learning_rate": 0.0006,
+ "loss": 4.686994552612305,
+ "step": 1146
+ },
+ {
+ "epoch": 15.936244541484717,
+ "grad_norm": 0.016208255663514137,
+ "learning_rate": 0.0006,
+ "loss": 4.6205549240112305,
+ "step": 1147
+ },
+ {
+ "epoch": 15.950218340611354,
+ "grad_norm": 0.0171707421541214,
+ "learning_rate": 0.0006,
+ "loss": 4.582968711853027,
+ "step": 1148
+ },
+ {
+ "epoch": 15.96419213973799,
+ "grad_norm": 0.017009446397423744,
+ "learning_rate": 0.0006,
+ "loss": 4.510274887084961,
+ "step": 1149
+ },
+ {
+ "epoch": 15.97816593886463,
+ "grad_norm": 0.01565811224281788,
+ "learning_rate": 0.0006,
+ "loss": 4.55834436416626,
+ "step": 1150
+ },
+ {
+ "epoch": 15.992139737991266,
+ "grad_norm": 0.015276037156581879,
+ "learning_rate": 0.0006,
+ "loss": 4.537870407104492,
+ "step": 1151
+ },
+ {
+ "epoch": 16.0,
+ "grad_norm": 0.017305882647633553,
+ "learning_rate": 0.0006,
+ "loss": 4.611305236816406,
+ "step": 1152
+ },
+ {
+ "epoch": 16.0,
+ "eval_loss": 4.775513172149658,
+ "eval_runtime": 56.7341,
+ "eval_samples_per_second": 43.043,
+ "eval_steps_per_second": 1.357,
+ "step": 1152
+ },
+ {
+ "epoch": 16.01397379912664,
+ "grad_norm": 0.016752436757087708,
+ "learning_rate": 0.0006,
+ "loss": 4.541312217712402,
+ "step": 1153
+ },
+ {
+ "epoch": 16.027947598253274,
+ "grad_norm": 0.01643134094774723,
+ "learning_rate": 0.0006,
+ "loss": 4.585402965545654,
+ "step": 1154
+ },
+ {
+ "epoch": 16.041921397379912,
+ "grad_norm": 0.01888720504939556,
+ "learning_rate": 0.0006,
+ "loss": 4.597825050354004,
+ "step": 1155
+ },
+ {
+ "epoch": 16.05589519650655,
+ "grad_norm": 0.022044632583856583,
+ "learning_rate": 0.0006,
+ "loss": 4.532772064208984,
+ "step": 1156
+ },
+ {
+ "epoch": 16.069868995633186,
+ "grad_norm": 0.022986987605690956,
+ "learning_rate": 0.0006,
+ "loss": 4.551351547241211,
+ "step": 1157
+ },
+ {
+ "epoch": 16.083842794759825,
+ "grad_norm": 0.020116791129112244,
+ "learning_rate": 0.0006,
+ "loss": 4.484149932861328,
+ "step": 1158
+ },
+ {
+ "epoch": 16.097816593886463,
+ "grad_norm": 0.021442247554659843,
+ "learning_rate": 0.0006,
+ "loss": 4.515316486358643,
+ "step": 1159
+ },
+ {
+ "epoch": 16.111790393013102,
+ "grad_norm": 0.020974772050976753,
+ "learning_rate": 0.0006,
+ "loss": 4.51606559753418,
+ "step": 1160
+ },
+ {
+ "epoch": 16.125764192139737,
+ "grad_norm": 0.021274864673614502,
+ "learning_rate": 0.0006,
+ "loss": 4.587329864501953,
+ "step": 1161
+ },
+ {
+ "epoch": 16.139737991266376,
+ "grad_norm": 0.022799784317612648,
+ "learning_rate": 0.0006,
+ "loss": 4.568149089813232,
+ "step": 1162
+ },
+ {
+ "epoch": 16.153711790393015,
+ "grad_norm": 0.021784933283925056,
+ "learning_rate": 0.0006,
+ "loss": 4.515336990356445,
+ "step": 1163
+ },
+ {
+ "epoch": 16.16768558951965,
+ "grad_norm": 0.024261271581053734,
+ "learning_rate": 0.0006,
+ "loss": 4.5127081871032715,
+ "step": 1164
+ },
+ {
+ "epoch": 16.18165938864629,
+ "grad_norm": 0.02248375490307808,
+ "learning_rate": 0.0006,
+ "loss": 4.754251003265381,
+ "step": 1165
+ },
+ {
+ "epoch": 16.195633187772927,
+ "grad_norm": 0.02136492356657982,
+ "learning_rate": 0.0006,
+ "loss": 4.58193826675415,
+ "step": 1166
+ },
+ {
+ "epoch": 16.209606986899562,
+ "grad_norm": 0.020727066323161125,
+ "learning_rate": 0.0006,
+ "loss": 4.564209938049316,
+ "step": 1167
+ },
+ {
+ "epoch": 16.2235807860262,
+ "grad_norm": 0.017789119854569435,
+ "learning_rate": 0.0006,
+ "loss": 4.698680877685547,
+ "step": 1168
+ },
+ {
+ "epoch": 16.23755458515284,
+ "grad_norm": 0.017603034153580666,
+ "learning_rate": 0.0006,
+ "loss": 4.501171588897705,
+ "step": 1169
+ },
+ {
+ "epoch": 16.251528384279474,
+ "grad_norm": 0.017345862463116646,
+ "learning_rate": 0.0006,
+ "loss": 4.593590259552002,
+ "step": 1170
+ },
+ {
+ "epoch": 16.265502183406113,
+ "grad_norm": 0.018047425895929337,
+ "learning_rate": 0.0006,
+ "loss": 4.597467422485352,
+ "step": 1171
+ },
+ {
+ "epoch": 16.27947598253275,
+ "grad_norm": 0.01947222836315632,
+ "learning_rate": 0.0006,
+ "loss": 4.5072221755981445,
+ "step": 1172
+ },
+ {
+ "epoch": 16.29344978165939,
+ "grad_norm": 0.019598500803112984,
+ "learning_rate": 0.0006,
+ "loss": 4.601351737976074,
+ "step": 1173
+ },
+ {
+ "epoch": 16.307423580786025,
+ "grad_norm": 0.020002514123916626,
+ "learning_rate": 0.0006,
+ "loss": 4.580431938171387,
+ "step": 1174
+ },
+ {
+ "epoch": 16.321397379912664,
+ "grad_norm": 0.0216530729085207,
+ "learning_rate": 0.0006,
+ "loss": 4.424681663513184,
+ "step": 1175
+ },
+ {
+ "epoch": 16.335371179039303,
+ "grad_norm": 0.01949235238134861,
+ "learning_rate": 0.0006,
+ "loss": 4.751883029937744,
+ "step": 1176
+ },
+ {
+ "epoch": 16.349344978165938,
+ "grad_norm": 0.017779016867280006,
+ "learning_rate": 0.0006,
+ "loss": 4.541490077972412,
+ "step": 1177
+ },
+ {
+ "epoch": 16.363318777292577,
+ "grad_norm": 0.01909121684730053,
+ "learning_rate": 0.0006,
+ "loss": 4.594141006469727,
+ "step": 1178
+ },
+ {
+ "epoch": 16.377292576419215,
+ "grad_norm": 0.019894849509000778,
+ "learning_rate": 0.0006,
+ "loss": 4.555070400238037,
+ "step": 1179
+ },
+ {
+ "epoch": 16.39126637554585,
+ "grad_norm": 0.01993032731115818,
+ "learning_rate": 0.0006,
+ "loss": 4.521401405334473,
+ "step": 1180
+ },
+ {
+ "epoch": 16.40524017467249,
+ "grad_norm": 0.0210788045078516,
+ "learning_rate": 0.0006,
+ "loss": 4.672842979431152,
+ "step": 1181
+ },
+ {
+ "epoch": 16.419213973799128,
+ "grad_norm": 0.02274598926305771,
+ "learning_rate": 0.0006,
+ "loss": 4.629504203796387,
+ "step": 1182
+ },
+ {
+ "epoch": 16.433187772925763,
+ "grad_norm": 0.020448120310902596,
+ "learning_rate": 0.0006,
+ "loss": 4.548209190368652,
+ "step": 1183
+ },
+ {
+ "epoch": 16.4471615720524,
+ "grad_norm": 0.017565643414855003,
+ "learning_rate": 0.0006,
+ "loss": 4.529054641723633,
+ "step": 1184
+ },
+ {
+ "epoch": 16.46113537117904,
+ "grad_norm": 0.019504237920045853,
+ "learning_rate": 0.0006,
+ "loss": 4.48182487487793,
+ "step": 1185
+ },
+ {
+ "epoch": 16.475109170305675,
+ "grad_norm": 0.016823742538690567,
+ "learning_rate": 0.0006,
+ "loss": 4.608696937561035,
+ "step": 1186
+ },
+ {
+ "epoch": 16.489082969432314,
+ "grad_norm": 0.01602894812822342,
+ "learning_rate": 0.0006,
+ "loss": 4.620062351226807,
+ "step": 1187
+ },
+ {
+ "epoch": 16.503056768558952,
+ "grad_norm": 0.016615642234683037,
+ "learning_rate": 0.0006,
+ "loss": 4.69378662109375,
+ "step": 1188
+ },
+ {
+ "epoch": 16.51703056768559,
+ "grad_norm": 0.018576730042696,
+ "learning_rate": 0.0006,
+ "loss": 4.537041664123535,
+ "step": 1189
+ },
+ {
+ "epoch": 16.531004366812226,
+ "grad_norm": 0.018611151725053787,
+ "learning_rate": 0.0006,
+ "loss": 4.496169090270996,
+ "step": 1190
+ },
+ {
+ "epoch": 16.544978165938865,
+ "grad_norm": 0.018622413277626038,
+ "learning_rate": 0.0006,
+ "loss": 4.557358264923096,
+ "step": 1191
+ },
+ {
+ "epoch": 16.558951965065503,
+ "grad_norm": 0.01823544315993786,
+ "learning_rate": 0.0006,
+ "loss": 4.730792999267578,
+ "step": 1192
+ },
+ {
+ "epoch": 16.57292576419214,
+ "grad_norm": 0.018796846270561218,
+ "learning_rate": 0.0006,
+ "loss": 4.781791687011719,
+ "step": 1193
+ },
+ {
+ "epoch": 16.586899563318777,
+ "grad_norm": 0.01728292554616928,
+ "learning_rate": 0.0006,
+ "loss": 4.515257835388184,
+ "step": 1194
+ },
+ {
+ "epoch": 16.600873362445416,
+ "grad_norm": 0.01699589006602764,
+ "learning_rate": 0.0006,
+ "loss": 4.5996904373168945,
+ "step": 1195
+ },
+ {
+ "epoch": 16.61484716157205,
+ "grad_norm": 0.015905970707535744,
+ "learning_rate": 0.0006,
+ "loss": 4.601301193237305,
+ "step": 1196
+ },
+ {
+ "epoch": 16.62882096069869,
+ "grad_norm": 0.01338571310043335,
+ "learning_rate": 0.0006,
+ "loss": 4.557708740234375,
+ "step": 1197
+ },
+ {
+ "epoch": 16.64279475982533,
+ "grad_norm": 0.015298855490982533,
+ "learning_rate": 0.0006,
+ "loss": 4.542287349700928,
+ "step": 1198
+ },
+ {
+ "epoch": 16.656768558951963,
+ "grad_norm": 0.014753367751836777,
+ "learning_rate": 0.0006,
+ "loss": 4.595787048339844,
+ "step": 1199
+ },
+ {
+ "epoch": 16.670742358078602,
+ "grad_norm": 0.015527483075857162,
+ "learning_rate": 0.0006,
+ "loss": 4.607529163360596,
+ "step": 1200
+ },
+ {
+ "epoch": 16.68471615720524,
+ "grad_norm": 0.015042162500321865,
+ "learning_rate": 0.0006,
+ "loss": 4.686626434326172,
+ "step": 1201
+ },
+ {
+ "epoch": 16.69868995633188,
+ "grad_norm": 0.0159393772482872,
+ "learning_rate": 0.0006,
+ "loss": 4.47656774520874,
+ "step": 1202
+ },
+ {
+ "epoch": 16.712663755458514,
+ "grad_norm": 0.01532787922769785,
+ "learning_rate": 0.0006,
+ "loss": 4.55678653717041,
+ "step": 1203
+ },
+ {
+ "epoch": 16.726637554585153,
+ "grad_norm": 0.016332164406776428,
+ "learning_rate": 0.0006,
+ "loss": 4.51441764831543,
+ "step": 1204
+ },
+ {
+ "epoch": 16.74061135371179,
+ "grad_norm": 0.018901726230978966,
+ "learning_rate": 0.0006,
+ "loss": 4.57227897644043,
+ "step": 1205
+ },
+ {
+ "epoch": 16.754585152838427,
+ "grad_norm": 0.018589433282613754,
+ "learning_rate": 0.0006,
+ "loss": 4.550070762634277,
+ "step": 1206
+ },
+ {
+ "epoch": 16.768558951965066,
+ "grad_norm": 0.01634034886956215,
+ "learning_rate": 0.0006,
+ "loss": 4.586910724639893,
+ "step": 1207
+ },
+ {
+ "epoch": 16.782532751091704,
+ "grad_norm": 0.017599433660507202,
+ "learning_rate": 0.0006,
+ "loss": 4.5891923904418945,
+ "step": 1208
+ },
+ {
+ "epoch": 16.79650655021834,
+ "grad_norm": 0.01878446340560913,
+ "learning_rate": 0.0006,
+ "loss": 4.57227087020874,
+ "step": 1209
+ },
+ {
+ "epoch": 16.810480349344978,
+ "grad_norm": 0.01803469844162464,
+ "learning_rate": 0.0006,
+ "loss": 4.583845138549805,
+ "step": 1210
+ },
+ {
+ "epoch": 16.824454148471617,
+ "grad_norm": 0.018505144864320755,
+ "learning_rate": 0.0006,
+ "loss": 4.561151504516602,
+ "step": 1211
+ },
+ {
+ "epoch": 16.83842794759825,
+ "grad_norm": 0.01920081488788128,
+ "learning_rate": 0.0006,
+ "loss": 4.506834030151367,
+ "step": 1212
+ },
+ {
+ "epoch": 16.85240174672489,
+ "grad_norm": 0.0177712831646204,
+ "learning_rate": 0.0006,
+ "loss": 4.557195663452148,
+ "step": 1213
+ },
+ {
+ "epoch": 16.86637554585153,
+ "grad_norm": 0.019113918766379356,
+ "learning_rate": 0.0006,
+ "loss": 4.651227951049805,
+ "step": 1214
+ },
+ {
+ "epoch": 16.880349344978168,
+ "grad_norm": 0.02012539468705654,
+ "learning_rate": 0.0006,
+ "loss": 4.4952239990234375,
+ "step": 1215
+ },
+ {
+ "epoch": 16.894323144104803,
+ "grad_norm": 0.019993646070361137,
+ "learning_rate": 0.0006,
+ "loss": 4.517433166503906,
+ "step": 1216
+ },
+ {
+ "epoch": 16.90829694323144,
+ "grad_norm": 0.020041441544890404,
+ "learning_rate": 0.0006,
+ "loss": 4.489407539367676,
+ "step": 1217
+ },
+ {
+ "epoch": 16.92227074235808,
+ "grad_norm": 0.0183611661195755,
+ "learning_rate": 0.0006,
+ "loss": 4.5435638427734375,
+ "step": 1218
+ },
+ {
+ "epoch": 16.936244541484715,
+ "grad_norm": 0.019138429313898087,
+ "learning_rate": 0.0006,
+ "loss": 4.545949459075928,
+ "step": 1219
+ },
+ {
+ "epoch": 16.950218340611354,
+ "grad_norm": 0.018554266542196274,
+ "learning_rate": 0.0006,
+ "loss": 4.579620361328125,
+ "step": 1220
+ },
+ {
+ "epoch": 16.964192139737992,
+ "grad_norm": 0.0193545650690794,
+ "learning_rate": 0.0006,
+ "loss": 4.588255405426025,
+ "step": 1221
+ },
+ {
+ "epoch": 16.978165938864628,
+ "grad_norm": 0.019389165565371513,
+ "learning_rate": 0.0006,
+ "loss": 4.520698547363281,
+ "step": 1222
+ },
+ {
+ "epoch": 16.992139737991266,
+ "grad_norm": 0.016601379960775375,
+ "learning_rate": 0.0006,
+ "loss": 4.4809064865112305,
+ "step": 1223
+ },
+ {
+ "epoch": 17.0,
+ "grad_norm": 0.01771303080022335,
+ "learning_rate": 0.0006,
+ "loss": 4.721122741699219,
+ "step": 1224
+ },
+ {
+ "epoch": 17.0,
+ "eval_loss": 4.7903666496276855,
+ "eval_runtime": 56.9826,
+ "eval_samples_per_second": 42.855,
+ "eval_steps_per_second": 1.351,
+ "step": 1224
+ },
+ {
+ "epoch": 17.01397379912664,
+ "grad_norm": 0.017720289528369904,
+ "learning_rate": 0.0006,
+ "loss": 4.523834705352783,
+ "step": 1225
+ },
+ {
+ "epoch": 17.027947598253274,
+ "grad_norm": 0.020137161016464233,
+ "learning_rate": 0.0006,
+ "loss": 4.574061870574951,
+ "step": 1226
+ },
+ {
+ "epoch": 17.041921397379912,
+ "grad_norm": 0.02002314105629921,
+ "learning_rate": 0.0006,
+ "loss": 4.512407302856445,
+ "step": 1227
+ },
+ {
+ "epoch": 17.05589519650655,
+ "grad_norm": 0.019628183916211128,
+ "learning_rate": 0.0006,
+ "loss": 4.502533435821533,
+ "step": 1228
+ },
+ {
+ "epoch": 17.069868995633186,
+ "grad_norm": 0.019808458164334297,
+ "learning_rate": 0.0006,
+ "loss": 4.463156700134277,
+ "step": 1229
+ },
+ {
+ "epoch": 17.083842794759825,
+ "grad_norm": 0.021250786259770393,
+ "learning_rate": 0.0006,
+ "loss": 4.622682094573975,
+ "step": 1230
+ },
+ {
+ "epoch": 17.097816593886463,
+ "grad_norm": 0.022031694650650024,
+ "learning_rate": 0.0006,
+ "loss": 4.451882839202881,
+ "step": 1231
+ },
+ {
+ "epoch": 17.111790393013102,
+ "grad_norm": 0.01898987777531147,
+ "learning_rate": 0.0006,
+ "loss": 4.529331684112549,
+ "step": 1232
+ },
+ {
+ "epoch": 17.125764192139737,
+ "grad_norm": 0.018749834969639778,
+ "learning_rate": 0.0006,
+ "loss": 4.435815334320068,
+ "step": 1233
+ },
+ {
+ "epoch": 17.139737991266376,
+ "grad_norm": 0.01896674931049347,
+ "learning_rate": 0.0006,
+ "loss": 4.484029293060303,
+ "step": 1234
+ },
+ {
+ "epoch": 17.153711790393015,
+ "grad_norm": 0.018286997452378273,
+ "learning_rate": 0.0006,
+ "loss": 4.523510932922363,
+ "step": 1235
+ },
+ {
+ "epoch": 17.16768558951965,
+ "grad_norm": 0.017273059114813805,
+ "learning_rate": 0.0006,
+ "loss": 4.451767444610596,
+ "step": 1236
+ },
+ {
+ "epoch": 17.18165938864629,
+ "grad_norm": 0.016499150544404984,
+ "learning_rate": 0.0006,
+ "loss": 4.5568647384643555,
+ "step": 1237
+ },
+ {
+ "epoch": 17.195633187772927,
+ "grad_norm": 0.015934469178318977,
+ "learning_rate": 0.0006,
+ "loss": 4.4639997482299805,
+ "step": 1238
+ },
+ {
+ "epoch": 17.209606986899562,
+ "grad_norm": 0.01714209094643593,
+ "learning_rate": 0.0006,
+ "loss": 4.547614097595215,
+ "step": 1239
+ },
+ {
+ "epoch": 17.2235807860262,
+ "grad_norm": 0.019167495891451836,
+ "learning_rate": 0.0006,
+ "loss": 4.571097373962402,
+ "step": 1240
+ },
+ {
+ "epoch": 17.23755458515284,
+ "grad_norm": 0.018276996910572052,
+ "learning_rate": 0.0006,
+ "loss": 4.489764213562012,
+ "step": 1241
+ },
+ {
+ "epoch": 17.251528384279474,
+ "grad_norm": 0.021296866238117218,
+ "learning_rate": 0.0006,
+ "loss": 4.526913642883301,
+ "step": 1242
+ },
+ {
+ "epoch": 17.265502183406113,
+ "grad_norm": 0.022606629878282547,
+ "learning_rate": 0.0006,
+ "loss": 4.521624565124512,
+ "step": 1243
+ },
+ {
+ "epoch": 17.27947598253275,
+ "grad_norm": 0.021204711869359016,
+ "learning_rate": 0.0006,
+ "loss": 4.500864028930664,
+ "step": 1244
+ },
+ {
+ "epoch": 17.29344978165939,
+ "grad_norm": 0.01988849975168705,
+ "learning_rate": 0.0006,
+ "loss": 4.45162296295166,
+ "step": 1245
+ },
+ {
+ "epoch": 17.307423580786025,
+ "grad_norm": 0.020268384367227554,
+ "learning_rate": 0.0006,
+ "loss": 4.536198616027832,
+ "step": 1246
+ },
+ {
+ "epoch": 17.321397379912664,
+ "grad_norm": 0.020185677334666252,
+ "learning_rate": 0.0006,
+ "loss": 4.440262794494629,
+ "step": 1247
+ },
+ {
+ "epoch": 17.335371179039303,
+ "grad_norm": 0.020106853917241096,
+ "learning_rate": 0.0006,
+ "loss": 4.475907325744629,
+ "step": 1248
+ },
+ {
+ "epoch": 17.349344978165938,
+ "grad_norm": 0.018883297219872475,
+ "learning_rate": 0.0006,
+ "loss": 4.640661716461182,
+ "step": 1249
+ },
+ {
+ "epoch": 17.363318777292577,
+ "grad_norm": 0.01852620206773281,
+ "learning_rate": 0.0006,
+ "loss": 4.578243255615234,
+ "step": 1250
+ },
+ {
+ "epoch": 17.377292576419215,
+ "grad_norm": 0.01886451430618763,
+ "learning_rate": 0.0006,
+ "loss": 4.546026229858398,
+ "step": 1251
+ },
+ {
+ "epoch": 17.39126637554585,
+ "grad_norm": 0.01777452416718006,
+ "learning_rate": 0.0006,
+ "loss": 4.521542549133301,
+ "step": 1252
+ },
+ {
+ "epoch": 17.40524017467249,
+ "grad_norm": 0.016626447439193726,
+ "learning_rate": 0.0006,
+ "loss": 4.52629280090332,
+ "step": 1253
+ },
+ {
+ "epoch": 17.419213973799128,
+ "grad_norm": 0.01599477231502533,
+ "learning_rate": 0.0006,
+ "loss": 4.4796037673950195,
+ "step": 1254
+ },
+ {
+ "epoch": 17.433187772925763,
+ "grad_norm": 0.017021458595991135,
+ "learning_rate": 0.0006,
+ "loss": 4.468813419342041,
+ "step": 1255
+ },
+ {
+ "epoch": 17.4471615720524,
+ "grad_norm": 0.016925333067774773,
+ "learning_rate": 0.0006,
+ "loss": 4.584898948669434,
+ "step": 1256
+ },
+ {
+ "epoch": 17.46113537117904,
+ "grad_norm": 0.01714998669922352,
+ "learning_rate": 0.0006,
+ "loss": 4.523544788360596,
+ "step": 1257
+ },
+ {
+ "epoch": 17.475109170305675,
+ "grad_norm": 0.01802288182079792,
+ "learning_rate": 0.0006,
+ "loss": 4.595581531524658,
+ "step": 1258
+ },
+ {
+ "epoch": 17.489082969432314,
+ "grad_norm": 0.0180289875715971,
+ "learning_rate": 0.0006,
+ "loss": 4.394440650939941,
+ "step": 1259
+ },
+ {
+ "epoch": 17.503056768558952,
+ "grad_norm": 0.018504245206713676,
+ "learning_rate": 0.0006,
+ "loss": 4.4794416427612305,
+ "step": 1260
+ },
+ {
+ "epoch": 17.51703056768559,
+ "grad_norm": 0.019973233342170715,
+ "learning_rate": 0.0006,
+ "loss": 4.4241437911987305,
+ "step": 1261
+ },
+ {
+ "epoch": 17.531004366812226,
+ "grad_norm": 0.020063083618879318,
+ "learning_rate": 0.0006,
+ "loss": 4.48068380355835,
+ "step": 1262
+ },
+ {
+ "epoch": 17.544978165938865,
+ "grad_norm": 0.02066640742123127,
+ "learning_rate": 0.0006,
+ "loss": 4.628993034362793,
+ "step": 1263
+ },
+ {
+ "epoch": 17.558951965065503,
+ "grad_norm": 0.02206416428089142,
+ "learning_rate": 0.0006,
+ "loss": 4.564790725708008,
+ "step": 1264
+ },
+ {
+ "epoch": 17.57292576419214,
+ "grad_norm": 0.020640335977077484,
+ "learning_rate": 0.0006,
+ "loss": 4.578705787658691,
+ "step": 1265
+ },
+ {
+ "epoch": 17.586899563318777,
+ "grad_norm": 0.019093405455350876,
+ "learning_rate": 0.0006,
+ "loss": 4.5023603439331055,
+ "step": 1266
+ },
+ {
+ "epoch": 17.600873362445416,
+ "grad_norm": 0.019459193572402,
+ "learning_rate": 0.0006,
+ "loss": 4.594531059265137,
+ "step": 1267
+ },
+ {
+ "epoch": 17.61484716157205,
+ "grad_norm": 0.02021171525120735,
+ "learning_rate": 0.0006,
+ "loss": 4.770299911499023,
+ "step": 1268
+ },
+ {
+ "epoch": 17.62882096069869,
+ "grad_norm": 0.021161191165447235,
+ "learning_rate": 0.0006,
+ "loss": 4.5380401611328125,
+ "step": 1269
+ },
+ {
+ "epoch": 17.64279475982533,
+ "grad_norm": 0.021013403311371803,
+ "learning_rate": 0.0006,
+ "loss": 4.455524921417236,
+ "step": 1270
+ },
+ {
+ "epoch": 17.656768558951963,
+ "grad_norm": 0.020336592569947243,
+ "learning_rate": 0.0006,
+ "loss": 4.533672332763672,
+ "step": 1271
+ },
+ {
+ "epoch": 17.670742358078602,
+ "grad_norm": 0.020159434527158737,
+ "learning_rate": 0.0006,
+ "loss": 4.596151351928711,
+ "step": 1272
+ },
+ {
+ "epoch": 17.68471615720524,
+ "grad_norm": 0.020432960242033005,
+ "learning_rate": 0.0006,
+ "loss": 4.55531644821167,
+ "step": 1273
+ },
+ {
+ "epoch": 17.69868995633188,
+ "grad_norm": 0.025466229766607285,
+ "learning_rate": 0.0006,
+ "loss": 4.566912651062012,
+ "step": 1274
+ },
+ {
+ "epoch": 17.712663755458514,
+ "grad_norm": 0.026537558063864708,
+ "learning_rate": 0.0006,
+ "loss": 4.526003837585449,
+ "step": 1275
+ },
+ {
+ "epoch": 17.726637554585153,
+ "grad_norm": 0.024397607892751694,
+ "learning_rate": 0.0006,
+ "loss": 4.541755676269531,
+ "step": 1276
+ },
+ {
+ "epoch": 17.74061135371179,
+ "grad_norm": 0.022062312811613083,
+ "learning_rate": 0.0006,
+ "loss": 4.6532793045043945,
+ "step": 1277
+ },
+ {
+ "epoch": 17.754585152838427,
+ "grad_norm": 0.02204282023012638,
+ "learning_rate": 0.0006,
+ "loss": 4.582825660705566,
+ "step": 1278
+ },
+ {
+ "epoch": 17.768558951965066,
+ "grad_norm": 0.020844416692852974,
+ "learning_rate": 0.0006,
+ "loss": 4.522336483001709,
+ "step": 1279
+ },
+ {
+ "epoch": 17.782532751091704,
+ "grad_norm": 0.019815703853964806,
+ "learning_rate": 0.0006,
+ "loss": 4.607398509979248,
+ "step": 1280
+ },
+ {
+ "epoch": 17.79650655021834,
+ "grad_norm": 0.016676954925060272,
+ "learning_rate": 0.0006,
+ "loss": 4.446206092834473,
+ "step": 1281
+ },
+ {
+ "epoch": 17.810480349344978,
+ "grad_norm": 0.0166452769190073,
+ "learning_rate": 0.0006,
+ "loss": 4.505931854248047,
+ "step": 1282
+ },
+ {
+ "epoch": 17.824454148471617,
+ "grad_norm": 0.015928907319903374,
+ "learning_rate": 0.0006,
+ "loss": 4.455267429351807,
+ "step": 1283
+ },
+ {
+ "epoch": 17.83842794759825,
+ "grad_norm": 0.015219560824334621,
+ "learning_rate": 0.0006,
+ "loss": 4.5623674392700195,
+ "step": 1284
+ },
+ {
+ "epoch": 17.85240174672489,
+ "grad_norm": 0.015125355683267117,
+ "learning_rate": 0.0006,
+ "loss": 4.554104804992676,
+ "step": 1285
+ },
+ {
+ "epoch": 17.86637554585153,
+ "grad_norm": 0.015612718649208546,
+ "learning_rate": 0.0006,
+ "loss": 4.484847068786621,
+ "step": 1286
+ },
+ {
+ "epoch": 17.880349344978168,
+ "grad_norm": 0.015215662308037281,
+ "learning_rate": 0.0006,
+ "loss": 4.4802093505859375,
+ "step": 1287
+ },
+ {
+ "epoch": 17.894323144104803,
+ "grad_norm": 0.01511684525758028,
+ "learning_rate": 0.0006,
+ "loss": 4.537298679351807,
+ "step": 1288
+ },
+ {
+ "epoch": 17.90829694323144,
+ "grad_norm": 0.013841322623193264,
+ "learning_rate": 0.0006,
+ "loss": 4.427488327026367,
+ "step": 1289
+ },
+ {
+ "epoch": 17.92227074235808,
+ "grad_norm": 0.015115964226424694,
+ "learning_rate": 0.0006,
+ "loss": 4.4547119140625,
+ "step": 1290
+ },
+ {
+ "epoch": 17.936244541484715,
+ "grad_norm": 0.016411006450653076,
+ "learning_rate": 0.0006,
+ "loss": 4.52994966506958,
+ "step": 1291
+ },
+ {
+ "epoch": 17.950218340611354,
+ "grad_norm": 0.017313152551651,
+ "learning_rate": 0.0006,
+ "loss": 4.5337653160095215,
+ "step": 1292
+ },
+ {
+ "epoch": 17.964192139737992,
+ "grad_norm": 0.016559796407818794,
+ "learning_rate": 0.0006,
+ "loss": 4.5277910232543945,
+ "step": 1293
+ },
+ {
+ "epoch": 17.978165938864628,
+ "grad_norm": 0.015854761004447937,
+ "learning_rate": 0.0006,
+ "loss": 4.6334309577941895,
+ "step": 1294
+ },
+ {
+ "epoch": 17.992139737991266,
+ "grad_norm": 0.015658238902688026,
+ "learning_rate": 0.0006,
+ "loss": 4.613372802734375,
+ "step": 1295
+ },
+ {
+ "epoch": 18.0,
+ "grad_norm": 0.0169155802577734,
+ "learning_rate": 0.0006,
+ "loss": 4.527356147766113,
+ "step": 1296
+ },
+ {
+ "epoch": 18.0,
+ "eval_loss": 4.764543056488037,
+ "eval_runtime": 56.5998,
+ "eval_samples_per_second": 43.145,
+ "eval_steps_per_second": 1.36,
+ "step": 1296
+ },
+ {
+ "epoch": 18.01397379912664,
+ "grad_norm": 0.015348963439464569,
+ "learning_rate": 0.0006,
+ "loss": 4.466423988342285,
+ "step": 1297
+ },
+ {
+ "epoch": 18.027947598253274,
+ "grad_norm": 0.01768588088452816,
+ "learning_rate": 0.0006,
+ "loss": 4.545933246612549,
+ "step": 1298
+ },
+ {
+ "epoch": 18.041921397379912,
+ "grad_norm": 0.017381226643919945,
+ "learning_rate": 0.0006,
+ "loss": 4.494060039520264,
+ "step": 1299
+ },
+ {
+ "epoch": 18.05589519650655,
+ "grad_norm": 0.019002335146069527,
+ "learning_rate": 0.0006,
+ "loss": 4.395200729370117,
+ "step": 1300
+ },
+ {
+ "epoch": 18.069868995633186,
+ "grad_norm": 0.01827932707965374,
+ "learning_rate": 0.0006,
+ "loss": 4.424848556518555,
+ "step": 1301
+ },
+ {
+ "epoch": 18.083842794759825,
+ "grad_norm": 0.016383031383156776,
+ "learning_rate": 0.0006,
+ "loss": 4.494722366333008,
+ "step": 1302
+ },
+ {
+ "epoch": 18.097816593886463,
+ "grad_norm": 0.016265157610177994,
+ "learning_rate": 0.0006,
+ "loss": 4.416522979736328,
+ "step": 1303
+ },
+ {
+ "epoch": 18.111790393013102,
+ "grad_norm": 0.01702386513352394,
+ "learning_rate": 0.0006,
+ "loss": 4.600103378295898,
+ "step": 1304
+ },
+ {
+ "epoch": 18.125764192139737,
+ "grad_norm": 0.018495498225092888,
+ "learning_rate": 0.0006,
+ "loss": 4.588401794433594,
+ "step": 1305
+ },
+ {
+ "epoch": 18.139737991266376,
+ "grad_norm": 0.017921684309840202,
+ "learning_rate": 0.0006,
+ "loss": 4.478075981140137,
+ "step": 1306
+ },
+ {
+ "epoch": 18.153711790393015,
+ "grad_norm": 0.01801278069615364,
+ "learning_rate": 0.0006,
+ "loss": 4.413186550140381,
+ "step": 1307
+ },
+ {
+ "epoch": 18.16768558951965,
+ "grad_norm": 0.017211275175213814,
+ "learning_rate": 0.0006,
+ "loss": 4.461341857910156,
+ "step": 1308
+ },
+ {
+ "epoch": 18.18165938864629,
+ "grad_norm": 0.017589328810572624,
+ "learning_rate": 0.0006,
+ "loss": 4.546969413757324,
+ "step": 1309
+ },
+ {
+ "epoch": 18.195633187772927,
+ "grad_norm": 0.01797698810696602,
+ "learning_rate": 0.0006,
+ "loss": 4.530797004699707,
+ "step": 1310
+ },
+ {
+ "epoch": 18.209606986899562,
+ "grad_norm": 0.020408619195222855,
+ "learning_rate": 0.0006,
+ "loss": 4.62999153137207,
+ "step": 1311
+ },
+ {
+ "epoch": 18.2235807860262,
+ "grad_norm": 0.01994038000702858,
+ "learning_rate": 0.0006,
+ "loss": 4.586811065673828,
+ "step": 1312
+ },
+ {
+ "epoch": 18.23755458515284,
+ "grad_norm": 0.019381340593099594,
+ "learning_rate": 0.0006,
+ "loss": 4.4769368171691895,
+ "step": 1313
+ },
+ {
+ "epoch": 18.251528384279474,
+ "grad_norm": 0.020484555512666702,
+ "learning_rate": 0.0006,
+ "loss": 4.524302959442139,
+ "step": 1314
+ },
+ {
+ "epoch": 18.265502183406113,
+ "grad_norm": 0.02426416426897049,
+ "learning_rate": 0.0006,
+ "loss": 4.561705589294434,
+ "step": 1315
+ },
+ {
+ "epoch": 18.27947598253275,
+ "grad_norm": 0.024748101830482483,
+ "learning_rate": 0.0006,
+ "loss": 4.4654316902160645,
+ "step": 1316
+ },
+ {
+ "epoch": 18.29344978165939,
+ "grad_norm": 0.020703328773379326,
+ "learning_rate": 0.0006,
+ "loss": 4.53033447265625,
+ "step": 1317
+ },
+ {
+ "epoch": 18.307423580786025,
+ "grad_norm": 0.01765458658337593,
+ "learning_rate": 0.0006,
+ "loss": 4.5154266357421875,
+ "step": 1318
+ },
+ {
+ "epoch": 18.321397379912664,
+ "grad_norm": 0.018646420910954475,
+ "learning_rate": 0.0006,
+ "loss": 4.5224714279174805,
+ "step": 1319
+ },
+ {
+ "epoch": 18.335371179039303,
+ "grad_norm": 0.01907474733889103,
+ "learning_rate": 0.0006,
+ "loss": 4.5055437088012695,
+ "step": 1320
+ },
+ {
+ "epoch": 18.349344978165938,
+ "grad_norm": 0.018179846927523613,
+ "learning_rate": 0.0006,
+ "loss": 4.462161064147949,
+ "step": 1321
+ },
+ {
+ "epoch": 18.363318777292577,
+ "grad_norm": 0.016697878018021584,
+ "learning_rate": 0.0006,
+ "loss": 4.473718643188477,
+ "step": 1322
+ },
+ {
+ "epoch": 18.377292576419215,
+ "grad_norm": 0.017504019662737846,
+ "learning_rate": 0.0006,
+ "loss": 4.474348068237305,
+ "step": 1323
+ },
+ {
+ "epoch": 18.39126637554585,
+ "grad_norm": 0.018889913335442543,
+ "learning_rate": 0.0006,
+ "loss": 4.4531636238098145,
+ "step": 1324
+ },
+ {
+ "epoch": 18.40524017467249,
+ "grad_norm": 0.019167251884937286,
+ "learning_rate": 0.0006,
+ "loss": 4.4278364181518555,
+ "step": 1325
+ },
+ {
+ "epoch": 18.419213973799128,
+ "grad_norm": 0.015895741060376167,
+ "learning_rate": 0.0006,
+ "loss": 4.474592208862305,
+ "step": 1326
+ },
+ {
+ "epoch": 18.433187772925763,
+ "grad_norm": 0.01531882956624031,
+ "learning_rate": 0.0006,
+ "loss": 4.403757572174072,
+ "step": 1327
+ },
+ {
+ "epoch": 18.4471615720524,
+ "grad_norm": 0.01565849967300892,
+ "learning_rate": 0.0006,
+ "loss": 4.507664680480957,
+ "step": 1328
+ },
+ {
+ "epoch": 18.46113537117904,
+ "grad_norm": 0.01602315716445446,
+ "learning_rate": 0.0006,
+ "loss": 4.5648908615112305,
+ "step": 1329
+ },
+ {
+ "epoch": 18.475109170305675,
+ "grad_norm": 0.016467366367578506,
+ "learning_rate": 0.0006,
+ "loss": 4.428762912750244,
+ "step": 1330
+ },
+ {
+ "epoch": 18.489082969432314,
+ "grad_norm": 0.01604127697646618,
+ "learning_rate": 0.0006,
+ "loss": 4.527771472930908,
+ "step": 1331
+ },
+ {
+ "epoch": 18.503056768558952,
+ "grad_norm": 0.018825586885213852,
+ "learning_rate": 0.0006,
+ "loss": 4.485386848449707,
+ "step": 1332
+ },
+ {
+ "epoch": 18.51703056768559,
+ "grad_norm": 0.019258007407188416,
+ "learning_rate": 0.0006,
+ "loss": 4.414972305297852,
+ "step": 1333
+ },
+ {
+ "epoch": 18.531004366812226,
+ "grad_norm": 0.018192021176218987,
+ "learning_rate": 0.0006,
+ "loss": 4.412166595458984,
+ "step": 1334
+ },
+ {
+ "epoch": 18.544978165938865,
+ "grad_norm": 0.016120698302984238,
+ "learning_rate": 0.0006,
+ "loss": 4.484317779541016,
+ "step": 1335
+ },
+ {
+ "epoch": 18.558951965065503,
+ "grad_norm": 0.016496622934937477,
+ "learning_rate": 0.0006,
+ "loss": 4.514986515045166,
+ "step": 1336
+ },
+ {
+ "epoch": 18.57292576419214,
+ "grad_norm": 0.01795116998255253,
+ "learning_rate": 0.0006,
+ "loss": 4.41473388671875,
+ "step": 1337
+ },
+ {
+ "epoch": 18.586899563318777,
+ "grad_norm": 0.019062276929616928,
+ "learning_rate": 0.0006,
+ "loss": 4.601996898651123,
+ "step": 1338
+ },
+ {
+ "epoch": 18.600873362445416,
+ "grad_norm": 0.017174631357192993,
+ "learning_rate": 0.0006,
+ "loss": 4.571218490600586,
+ "step": 1339
+ },
+ {
+ "epoch": 18.61484716157205,
+ "grad_norm": 0.014015606604516506,
+ "learning_rate": 0.0006,
+ "loss": 4.526606559753418,
+ "step": 1340
+ },
+ {
+ "epoch": 18.62882096069869,
+ "grad_norm": 0.015358510427176952,
+ "learning_rate": 0.0006,
+ "loss": 4.604286193847656,
+ "step": 1341
+ },
+ {
+ "epoch": 18.64279475982533,
+ "grad_norm": 0.01619391329586506,
+ "learning_rate": 0.0006,
+ "loss": 4.560883522033691,
+ "step": 1342
+ },
+ {
+ "epoch": 18.656768558951963,
+ "grad_norm": 0.015972964465618134,
+ "learning_rate": 0.0006,
+ "loss": 4.479755401611328,
+ "step": 1343
+ },
+ {
+ "epoch": 18.670742358078602,
+ "grad_norm": 0.016861025243997574,
+ "learning_rate": 0.0006,
+ "loss": 4.474471092224121,
+ "step": 1344
+ },
+ {
+ "epoch": 18.68471615720524,
+ "grad_norm": 0.01822078227996826,
+ "learning_rate": 0.0006,
+ "loss": 4.509489059448242,
+ "step": 1345
+ },
+ {
+ "epoch": 18.69868995633188,
+ "grad_norm": 0.019415007904171944,
+ "learning_rate": 0.0006,
+ "loss": 4.508941173553467,
+ "step": 1346
+ },
+ {
+ "epoch": 18.712663755458514,
+ "grad_norm": 0.019856126978993416,
+ "learning_rate": 0.0006,
+ "loss": 4.446670055389404,
+ "step": 1347
+ },
+ {
+ "epoch": 18.726637554585153,
+ "grad_norm": 0.01842520199716091,
+ "learning_rate": 0.0006,
+ "loss": 4.524422645568848,
+ "step": 1348
+ },
+ {
+ "epoch": 18.74061135371179,
+ "grad_norm": 0.017196480184793472,
+ "learning_rate": 0.0006,
+ "loss": 4.498466968536377,
+ "step": 1349
+ },
+ {
+ "epoch": 18.754585152838427,
+ "grad_norm": 0.01823010854423046,
+ "learning_rate": 0.0006,
+ "loss": 4.534684181213379,
+ "step": 1350
+ },
+ {
+ "epoch": 18.768558951965066,
+ "grad_norm": 0.020398065447807312,
+ "learning_rate": 0.0006,
+ "loss": 4.456494331359863,
+ "step": 1351
+ },
+ {
+ "epoch": 18.782532751091704,
+ "grad_norm": 0.022992262616753578,
+ "learning_rate": 0.0006,
+ "loss": 4.476374626159668,
+ "step": 1352
+ },
+ {
+ "epoch": 18.79650655021834,
+ "grad_norm": 0.02235420234501362,
+ "learning_rate": 0.0006,
+ "loss": 4.587776184082031,
+ "step": 1353
+ },
+ {
+ "epoch": 18.810480349344978,
+ "grad_norm": 0.02093626745045185,
+ "learning_rate": 0.0006,
+ "loss": 4.424172401428223,
+ "step": 1354
+ },
+ {
+ "epoch": 18.824454148471617,
+ "grad_norm": 0.01997302658855915,
+ "learning_rate": 0.0006,
+ "loss": 4.448066711425781,
+ "step": 1355
+ },
+ {
+ "epoch": 18.83842794759825,
+ "grad_norm": 0.018265459686517715,
+ "learning_rate": 0.0006,
+ "loss": 4.436610698699951,
+ "step": 1356
+ },
+ {
+ "epoch": 18.85240174672489,
+ "grad_norm": 0.01673833839595318,
+ "learning_rate": 0.0006,
+ "loss": 4.5502214431762695,
+ "step": 1357
+ },
+ {
+ "epoch": 18.86637554585153,
+ "grad_norm": 0.017908833920955658,
+ "learning_rate": 0.0006,
+ "loss": 4.550530433654785,
+ "step": 1358
+ },
+ {
+ "epoch": 18.880349344978168,
+ "grad_norm": 0.01796654611825943,
+ "learning_rate": 0.0006,
+ "loss": 4.531463146209717,
+ "step": 1359
+ },
+ {
+ "epoch": 18.894323144104803,
+ "grad_norm": 0.01646183803677559,
+ "learning_rate": 0.0006,
+ "loss": 4.498582363128662,
+ "step": 1360
+ },
+ {
+ "epoch": 18.90829694323144,
+ "grad_norm": 0.015430327504873276,
+ "learning_rate": 0.0006,
+ "loss": 4.52571439743042,
+ "step": 1361
+ },
+ {
+ "epoch": 18.92227074235808,
+ "grad_norm": 0.016527952626347542,
+ "learning_rate": 0.0006,
+ "loss": 4.547013759613037,
+ "step": 1362
+ },
+ {
+ "epoch": 18.936244541484715,
+ "grad_norm": 0.016263660043478012,
+ "learning_rate": 0.0006,
+ "loss": 4.468754291534424,
+ "step": 1363
+ },
+ {
+ "epoch": 18.950218340611354,
+ "grad_norm": 0.016347210854291916,
+ "learning_rate": 0.0006,
+ "loss": 4.450741767883301,
+ "step": 1364
+ },
+ {
+ "epoch": 18.964192139737992,
+ "grad_norm": 0.01701398566365242,
+ "learning_rate": 0.0006,
+ "loss": 4.534899711608887,
+ "step": 1365
+ },
+ {
+ "epoch": 18.978165938864628,
+ "grad_norm": 0.017693081870675087,
+ "learning_rate": 0.0006,
+ "loss": 4.460553169250488,
+ "step": 1366
+ },
+ {
+ "epoch": 18.992139737991266,
+ "grad_norm": 0.018746308982372284,
+ "learning_rate": 0.0006,
+ "loss": 4.538792610168457,
+ "step": 1367
+ },
+ {
+ "epoch": 19.0,
+ "grad_norm": 0.020099416375160217,
+ "learning_rate": 0.0006,
+ "loss": 4.449617385864258,
+ "step": 1368
+ },
+ {
+ "epoch": 19.0,
+ "eval_loss": 4.7756571769714355,
+ "eval_runtime": 57.1804,
+ "eval_samples_per_second": 42.707,
+ "eval_steps_per_second": 1.347,
+ "step": 1368
+ },
+ {
+ "epoch": 19.01397379912664,
+ "grad_norm": 0.018726933747529984,
+ "learning_rate": 0.0006,
+ "loss": 4.446059703826904,
+ "step": 1369
+ },
+ {
+ "epoch": 19.027947598253274,
+ "grad_norm": 0.01956087350845337,
+ "learning_rate": 0.0006,
+ "loss": 4.4520039558410645,
+ "step": 1370
+ },
+ {
+ "epoch": 19.041921397379912,
+ "grad_norm": 0.020092811435461044,
+ "learning_rate": 0.0006,
+ "loss": 4.543343544006348,
+ "step": 1371
+ },
+ {
+ "epoch": 19.05589519650655,
+ "grad_norm": 0.01843622885644436,
+ "learning_rate": 0.0006,
+ "loss": 4.535579204559326,
+ "step": 1372
+ },
+ {
+ "epoch": 19.069868995633186,
+ "grad_norm": 0.016989488154649734,
+ "learning_rate": 0.0006,
+ "loss": 4.52316951751709,
+ "step": 1373
+ },
+ {
+ "epoch": 19.083842794759825,
+ "grad_norm": 0.01805422455072403,
+ "learning_rate": 0.0006,
+ "loss": 4.5393266677856445,
+ "step": 1374
+ },
+ {
+ "epoch": 19.097816593886463,
+ "grad_norm": 0.01874629408121109,
+ "learning_rate": 0.0006,
+ "loss": 4.422059059143066,
+ "step": 1375
+ },
+ {
+ "epoch": 19.111790393013102,
+ "grad_norm": 0.01804221048951149,
+ "learning_rate": 0.0006,
+ "loss": 4.50508451461792,
+ "step": 1376
+ },
+ {
+ "epoch": 19.125764192139737,
+ "grad_norm": 0.01785474829375744,
+ "learning_rate": 0.0006,
+ "loss": 4.412599563598633,
+ "step": 1377
+ },
+ {
+ "epoch": 19.139737991266376,
+ "grad_norm": 0.01971791312098503,
+ "learning_rate": 0.0006,
+ "loss": 4.518318176269531,
+ "step": 1378
+ },
+ {
+ "epoch": 19.153711790393015,
+ "grad_norm": 0.02061633951961994,
+ "learning_rate": 0.0006,
+ "loss": 4.332393169403076,
+ "step": 1379
+ },
+ {
+ "epoch": 19.16768558951965,
+ "grad_norm": 0.020563429221510887,
+ "learning_rate": 0.0006,
+ "loss": 4.445863723754883,
+ "step": 1380
+ },
+ {
+ "epoch": 19.18165938864629,
+ "grad_norm": 0.019995713606476784,
+ "learning_rate": 0.0006,
+ "loss": 4.497312068939209,
+ "step": 1381
+ },
+ {
+ "epoch": 19.195633187772927,
+ "grad_norm": 0.019823070615530014,
+ "learning_rate": 0.0006,
+ "loss": 4.475085258483887,
+ "step": 1382
+ },
+ {
+ "epoch": 19.209606986899562,
+ "grad_norm": 0.019039355218410492,
+ "learning_rate": 0.0006,
+ "loss": 4.50432014465332,
+ "step": 1383
+ },
+ {
+ "epoch": 19.2235807860262,
+ "grad_norm": 0.017876973375678062,
+ "learning_rate": 0.0006,
+ "loss": 4.384788513183594,
+ "step": 1384
+ },
+ {
+ "epoch": 19.23755458515284,
+ "grad_norm": 0.018697699531912804,
+ "learning_rate": 0.0006,
+ "loss": 4.410778045654297,
+ "step": 1385
+ },
+ {
+ "epoch": 19.251528384279474,
+ "grad_norm": 0.018826806917786598,
+ "learning_rate": 0.0006,
+ "loss": 4.524034023284912,
+ "step": 1386
+ },
+ {
+ "epoch": 19.265502183406113,
+ "grad_norm": 0.01824226602911949,
+ "learning_rate": 0.0006,
+ "loss": 4.526393890380859,
+ "step": 1387
+ },
+ {
+ "epoch": 19.27947598253275,
+ "grad_norm": 0.01753208599984646,
+ "learning_rate": 0.0006,
+ "loss": 4.493703365325928,
+ "step": 1388
+ },
+ {
+ "epoch": 19.29344978165939,
+ "grad_norm": 0.01742432825267315,
+ "learning_rate": 0.0006,
+ "loss": 4.392831802368164,
+ "step": 1389
+ },
+ {
+ "epoch": 19.307423580786025,
+ "grad_norm": 0.017099356278777122,
+ "learning_rate": 0.0006,
+ "loss": 4.578668594360352,
+ "step": 1390
+ },
+ {
+ "epoch": 19.321397379912664,
+ "grad_norm": 0.017946023494005203,
+ "learning_rate": 0.0006,
+ "loss": 4.456718444824219,
+ "step": 1391
+ },
+ {
+ "epoch": 19.335371179039303,
+ "grad_norm": 0.017705943435430527,
+ "learning_rate": 0.0006,
+ "loss": 4.553682327270508,
+ "step": 1392
+ },
+ {
+ "epoch": 19.349344978165938,
+ "grad_norm": 0.016647523269057274,
+ "learning_rate": 0.0006,
+ "loss": 4.525016784667969,
+ "step": 1393
+ },
+ {
+ "epoch": 19.363318777292577,
+ "grad_norm": 0.015407416969537735,
+ "learning_rate": 0.0006,
+ "loss": 4.500231742858887,
+ "step": 1394
+ },
+ {
+ "epoch": 19.377292576419215,
+ "grad_norm": 0.017264019697904587,
+ "learning_rate": 0.0006,
+ "loss": 4.4615349769592285,
+ "step": 1395
+ },
+ {
+ "epoch": 19.39126637554585,
+ "grad_norm": 0.01871555484831333,
+ "learning_rate": 0.0006,
+ "loss": 4.281026840209961,
+ "step": 1396
+ },
+ {
+ "epoch": 19.40524017467249,
+ "grad_norm": 0.017361650243401527,
+ "learning_rate": 0.0006,
+ "loss": 4.479382514953613,
+ "step": 1397
+ },
+ {
+ "epoch": 19.419213973799128,
+ "grad_norm": 0.019781362265348434,
+ "learning_rate": 0.0006,
+ "loss": 4.402320861816406,
+ "step": 1398
+ },
+ {
+ "epoch": 19.433187772925763,
+ "grad_norm": 0.019043298438191414,
+ "learning_rate": 0.0006,
+ "loss": 4.476983070373535,
+ "step": 1399
+ },
+ {
+ "epoch": 19.4471615720524,
+ "grad_norm": 0.017137985676527023,
+ "learning_rate": 0.0006,
+ "loss": 4.418689250946045,
+ "step": 1400
+ },
+ {
+ "epoch": 19.46113537117904,
+ "grad_norm": 0.016481833532452583,
+ "learning_rate": 0.0006,
+ "loss": 4.579145431518555,
+ "step": 1401
+ },
+ {
+ "epoch": 19.475109170305675,
+ "grad_norm": 0.016507714986801147,
+ "learning_rate": 0.0006,
+ "loss": 4.519158363342285,
+ "step": 1402
+ },
+ {
+ "epoch": 19.489082969432314,
+ "grad_norm": 0.018092872574925423,
+ "learning_rate": 0.0006,
+ "loss": 4.513748645782471,
+ "step": 1403
+ },
+ {
+ "epoch": 19.503056768558952,
+ "grad_norm": 0.018153440207242966,
+ "learning_rate": 0.0006,
+ "loss": 4.645384311676025,
+ "step": 1404
+ },
+ {
+ "epoch": 19.51703056768559,
+ "grad_norm": 0.01657119207084179,
+ "learning_rate": 0.0006,
+ "loss": 4.399757385253906,
+ "step": 1405
+ },
+ {
+ "epoch": 19.531004366812226,
+ "grad_norm": 0.0156280305236578,
+ "learning_rate": 0.0006,
+ "loss": 4.467836380004883,
+ "step": 1406
+ },
+ {
+ "epoch": 19.544978165938865,
+ "grad_norm": 0.016838403418660164,
+ "learning_rate": 0.0006,
+ "loss": 4.529176235198975,
+ "step": 1407
+ },
+ {
+ "epoch": 19.558951965065503,
+ "grad_norm": 0.01743236929178238,
+ "learning_rate": 0.0006,
+ "loss": 4.440032482147217,
+ "step": 1408
+ },
+ {
+ "epoch": 19.57292576419214,
+ "grad_norm": 0.01553610060364008,
+ "learning_rate": 0.0006,
+ "loss": 4.425588607788086,
+ "step": 1409
+ },
+ {
+ "epoch": 19.586899563318777,
+ "grad_norm": 0.017259271815419197,
+ "learning_rate": 0.0006,
+ "loss": 4.405452728271484,
+ "step": 1410
+ },
+ {
+ "epoch": 19.600873362445416,
+ "grad_norm": 0.015568481758236885,
+ "learning_rate": 0.0006,
+ "loss": 4.444279193878174,
+ "step": 1411
+ },
+ {
+ "epoch": 19.61484716157205,
+ "grad_norm": 0.015331248752772808,
+ "learning_rate": 0.0006,
+ "loss": 4.451846122741699,
+ "step": 1412
+ },
+ {
+ "epoch": 19.62882096069869,
+ "grad_norm": 0.015667378902435303,
+ "learning_rate": 0.0006,
+ "loss": 4.533931255340576,
+ "step": 1413
+ },
+ {
+ "epoch": 19.64279475982533,
+ "grad_norm": 0.016177602112293243,
+ "learning_rate": 0.0006,
+ "loss": 4.467313289642334,
+ "step": 1414
+ },
+ {
+ "epoch": 19.656768558951963,
+ "grad_norm": 0.01733129471540451,
+ "learning_rate": 0.0006,
+ "loss": 4.496893882751465,
+ "step": 1415
+ },
+ {
+ "epoch": 19.670742358078602,
+ "grad_norm": 0.016539594158530235,
+ "learning_rate": 0.0006,
+ "loss": 4.401961326599121,
+ "step": 1416
+ },
+ {
+ "epoch": 19.68471615720524,
+ "grad_norm": 0.016058551147580147,
+ "learning_rate": 0.0006,
+ "loss": 4.503375053405762,
+ "step": 1417
+ },
+ {
+ "epoch": 19.69868995633188,
+ "grad_norm": 0.01715777814388275,
+ "learning_rate": 0.0006,
+ "loss": 4.3808369636535645,
+ "step": 1418
+ },
+ {
+ "epoch": 19.712663755458514,
+ "grad_norm": 0.01681119203567505,
+ "learning_rate": 0.0006,
+ "loss": 4.5469560623168945,
+ "step": 1419
+ },
+ {
+ "epoch": 19.726637554585153,
+ "grad_norm": 0.01634550653398037,
+ "learning_rate": 0.0006,
+ "loss": 4.328197956085205,
+ "step": 1420
+ },
+ {
+ "epoch": 19.74061135371179,
+ "grad_norm": 0.01549347210675478,
+ "learning_rate": 0.0006,
+ "loss": 4.534621238708496,
+ "step": 1421
+ },
+ {
+ "epoch": 19.754585152838427,
+ "grad_norm": 0.01654665172100067,
+ "learning_rate": 0.0006,
+ "loss": 4.4724225997924805,
+ "step": 1422
+ },
+ {
+ "epoch": 19.768558951965066,
+ "grad_norm": 0.017119025811553,
+ "learning_rate": 0.0006,
+ "loss": 4.3897247314453125,
+ "step": 1423
+ },
+ {
+ "epoch": 19.782532751091704,
+ "grad_norm": 0.01603223942220211,
+ "learning_rate": 0.0006,
+ "loss": 4.412731170654297,
+ "step": 1424
+ },
+ {
+ "epoch": 19.79650655021834,
+ "grad_norm": 0.016774067655205727,
+ "learning_rate": 0.0006,
+ "loss": 4.54707145690918,
+ "step": 1425
+ },
+ {
+ "epoch": 19.810480349344978,
+ "grad_norm": 0.016453025862574577,
+ "learning_rate": 0.0006,
+ "loss": 4.462501049041748,
+ "step": 1426
+ },
+ {
+ "epoch": 19.824454148471617,
+ "grad_norm": 0.01597629487514496,
+ "learning_rate": 0.0006,
+ "loss": 4.229758262634277,
+ "step": 1427
+ },
+ {
+ "epoch": 19.83842794759825,
+ "grad_norm": 0.017999034374952316,
+ "learning_rate": 0.0006,
+ "loss": 4.389263153076172,
+ "step": 1428
+ },
+ {
+ "epoch": 19.85240174672489,
+ "grad_norm": 0.02165898121893406,
+ "learning_rate": 0.0006,
+ "loss": 4.342130661010742,
+ "step": 1429
+ },
+ {
+ "epoch": 19.86637554585153,
+ "grad_norm": 0.02222822792828083,
+ "learning_rate": 0.0006,
+ "loss": 4.492661476135254,
+ "step": 1430
+ },
+ {
+ "epoch": 19.880349344978168,
+ "grad_norm": 0.019378066062927246,
+ "learning_rate": 0.0006,
+ "loss": 4.4614105224609375,
+ "step": 1431
+ },
+ {
+ "epoch": 19.894323144104803,
+ "grad_norm": 0.019908083602786064,
+ "learning_rate": 0.0006,
+ "loss": 4.370929718017578,
+ "step": 1432
+ },
+ {
+ "epoch": 19.90829694323144,
+ "grad_norm": 0.022296471521258354,
+ "learning_rate": 0.0006,
+ "loss": 4.438320159912109,
+ "step": 1433
+ },
+ {
+ "epoch": 19.92227074235808,
+ "grad_norm": 0.021810343489050865,
+ "learning_rate": 0.0006,
+ "loss": 4.462067127227783,
+ "step": 1434
+ },
+ {
+ "epoch": 19.936244541484715,
+ "grad_norm": 0.02129209227859974,
+ "learning_rate": 0.0006,
+ "loss": 4.541261196136475,
+ "step": 1435
+ },
+ {
+ "epoch": 19.950218340611354,
+ "grad_norm": 0.02079629711806774,
+ "learning_rate": 0.0006,
+ "loss": 4.36712646484375,
+ "step": 1436
+ },
+ {
+ "epoch": 19.964192139737992,
+ "grad_norm": 0.022501740604639053,
+ "learning_rate": 0.0006,
+ "loss": 4.5279083251953125,
+ "step": 1437
+ },
+ {
+ "epoch": 19.978165938864628,
+ "grad_norm": 0.019121676683425903,
+ "learning_rate": 0.0006,
+ "loss": 4.477789878845215,
+ "step": 1438
+ },
+ {
+ "epoch": 19.992139737991266,
+ "grad_norm": 0.017429588362574577,
+ "learning_rate": 0.0006,
+ "loss": 4.4023847579956055,
+ "step": 1439
+ },
+ {
+ "epoch": 20.0,
+ "grad_norm": 0.018857665359973907,
+ "learning_rate": 0.0006,
+ "loss": 4.494743347167969,
+ "step": 1440
+ },
+ {
+ "epoch": 20.0,
+ "eval_loss": 4.694555759429932,
+ "eval_runtime": 56.4146,
+ "eval_samples_per_second": 43.287,
+ "eval_steps_per_second": 1.365,
+ "step": 1440
+ },
+ {
+ "epoch": 20.01397379912664,
+ "grad_norm": 0.016505861654877663,
+ "learning_rate": 0.0006,
+ "loss": 4.506229877471924,
+ "step": 1441
+ },
+ {
+ "epoch": 20.027947598253274,
+ "grad_norm": 0.017571842297911644,
+ "learning_rate": 0.0006,
+ "loss": 4.384073257446289,
+ "step": 1442
+ },
+ {
+ "epoch": 20.041921397379912,
+ "grad_norm": 0.019398365169763565,
+ "learning_rate": 0.0006,
+ "loss": 4.482177734375,
+ "step": 1443
+ },
+ {
+ "epoch": 20.05589519650655,
+ "grad_norm": 0.02011062018573284,
+ "learning_rate": 0.0006,
+ "loss": 4.440773963928223,
+ "step": 1444
+ },
+ {
+ "epoch": 20.069868995633186,
+ "grad_norm": 0.019026953727006912,
+ "learning_rate": 0.0006,
+ "loss": 4.507168769836426,
+ "step": 1445
+ },
+ {
+ "epoch": 20.083842794759825,
+ "grad_norm": 0.020506031811237335,
+ "learning_rate": 0.0006,
+ "loss": 4.469079971313477,
+ "step": 1446
+ },
+ {
+ "epoch": 20.097816593886463,
+ "grad_norm": 0.019608162343502045,
+ "learning_rate": 0.0006,
+ "loss": 4.395927429199219,
+ "step": 1447
+ },
+ {
+ "epoch": 20.111790393013102,
+ "grad_norm": 0.019419124349951744,
+ "learning_rate": 0.0006,
+ "loss": 4.484235763549805,
+ "step": 1448
+ },
+ {
+ "epoch": 20.125764192139737,
+ "grad_norm": 0.021775128319859505,
+ "learning_rate": 0.0006,
+ "loss": 4.333173751831055,
+ "step": 1449
+ },
+ {
+ "epoch": 20.139737991266376,
+ "grad_norm": 0.025512272492051125,
+ "learning_rate": 0.0006,
+ "loss": 4.460362434387207,
+ "step": 1450
+ },
+ {
+ "epoch": 20.153711790393015,
+ "grad_norm": 0.025157401338219643,
+ "learning_rate": 0.0006,
+ "loss": 4.4361395835876465,
+ "step": 1451
+ },
+ {
+ "epoch": 20.16768558951965,
+ "grad_norm": 0.019862616434693336,
+ "learning_rate": 0.0006,
+ "loss": 4.426807403564453,
+ "step": 1452
+ },
+ {
+ "epoch": 20.18165938864629,
+ "grad_norm": 0.01932152360677719,
+ "learning_rate": 0.0006,
+ "loss": 4.375677108764648,
+ "step": 1453
+ },
+ {
+ "epoch": 20.195633187772927,
+ "grad_norm": 0.01760999858379364,
+ "learning_rate": 0.0006,
+ "loss": 4.42259407043457,
+ "step": 1454
+ },
+ {
+ "epoch": 20.209606986899562,
+ "grad_norm": 0.016998112201690674,
+ "learning_rate": 0.0006,
+ "loss": 4.481280326843262,
+ "step": 1455
+ },
+ {
+ "epoch": 20.2235807860262,
+ "grad_norm": 0.017199190333485603,
+ "learning_rate": 0.0006,
+ "loss": 4.496712684631348,
+ "step": 1456
+ },
+ {
+ "epoch": 20.23755458515284,
+ "grad_norm": 0.016833679750561714,
+ "learning_rate": 0.0006,
+ "loss": 4.40736722946167,
+ "step": 1457
+ },
+ {
+ "epoch": 20.251528384279474,
+ "grad_norm": 0.017546730116009712,
+ "learning_rate": 0.0006,
+ "loss": 4.4915266036987305,
+ "step": 1458
+ },
+ {
+ "epoch": 20.265502183406113,
+ "grad_norm": 0.016595078632235527,
+ "learning_rate": 0.0006,
+ "loss": 4.460815906524658,
+ "step": 1459
+ },
+ {
+ "epoch": 20.27947598253275,
+ "grad_norm": 0.017728568986058235,
+ "learning_rate": 0.0006,
+ "loss": 4.346048831939697,
+ "step": 1460
+ },
+ {
+ "epoch": 20.29344978165939,
+ "grad_norm": 0.01849197782576084,
+ "learning_rate": 0.0006,
+ "loss": 4.399681091308594,
+ "step": 1461
+ },
+ {
+ "epoch": 20.307423580786025,
+ "grad_norm": 0.018336588516831398,
+ "learning_rate": 0.0006,
+ "loss": 4.414183616638184,
+ "step": 1462
+ },
+ {
+ "epoch": 20.321397379912664,
+ "grad_norm": 0.018086519092321396,
+ "learning_rate": 0.0006,
+ "loss": 4.346586227416992,
+ "step": 1463
+ },
+ {
+ "epoch": 20.335371179039303,
+ "grad_norm": 0.019697299227118492,
+ "learning_rate": 0.0006,
+ "loss": 4.4420952796936035,
+ "step": 1464
+ },
+ {
+ "epoch": 20.349344978165938,
+ "grad_norm": 0.02198074571788311,
+ "learning_rate": 0.0006,
+ "loss": 4.588729381561279,
+ "step": 1465
+ },
+ {
+ "epoch": 20.363318777292577,
+ "grad_norm": 0.022813329473137856,
+ "learning_rate": 0.0006,
+ "loss": 4.49729585647583,
+ "step": 1466
+ },
+ {
+ "epoch": 20.377292576419215,
+ "grad_norm": 0.022006649523973465,
+ "learning_rate": 0.0006,
+ "loss": 4.475641250610352,
+ "step": 1467
+ },
+ {
+ "epoch": 20.39126637554585,
+ "grad_norm": 0.023188438266515732,
+ "learning_rate": 0.0006,
+ "loss": 4.404322624206543,
+ "step": 1468
+ },
+ {
+ "epoch": 20.40524017467249,
+ "grad_norm": 0.022226877510547638,
+ "learning_rate": 0.0006,
+ "loss": 4.50904655456543,
+ "step": 1469
+ },
+ {
+ "epoch": 20.419213973799128,
+ "grad_norm": 0.021082637831568718,
+ "learning_rate": 0.0006,
+ "loss": 4.339011192321777,
+ "step": 1470
+ },
+ {
+ "epoch": 20.433187772925763,
+ "grad_norm": 0.02013542130589485,
+ "learning_rate": 0.0006,
+ "loss": 4.516942024230957,
+ "step": 1471
+ },
+ {
+ "epoch": 20.4471615720524,
+ "grad_norm": 0.01898968778550625,
+ "learning_rate": 0.0006,
+ "loss": 4.315804481506348,
+ "step": 1472
+ },
+ {
+ "epoch": 20.46113537117904,
+ "grad_norm": 0.019450880587100983,
+ "learning_rate": 0.0006,
+ "loss": 4.4794921875,
+ "step": 1473
+ },
+ {
+ "epoch": 20.475109170305675,
+ "grad_norm": 0.018577802926301956,
+ "learning_rate": 0.0006,
+ "loss": 4.422323703765869,
+ "step": 1474
+ },
+ {
+ "epoch": 20.489082969432314,
+ "grad_norm": 0.01778412237763405,
+ "learning_rate": 0.0006,
+ "loss": 4.429789066314697,
+ "step": 1475
+ },
+ {
+ "epoch": 20.503056768558952,
+ "grad_norm": 0.017134476453065872,
+ "learning_rate": 0.0006,
+ "loss": 4.416459083557129,
+ "step": 1476
+ },
+ {
+ "epoch": 20.51703056768559,
+ "grad_norm": 0.017153557389974594,
+ "learning_rate": 0.0006,
+ "loss": 4.446170806884766,
+ "step": 1477
+ },
+ {
+ "epoch": 20.531004366812226,
+ "grad_norm": 0.014998532831668854,
+ "learning_rate": 0.0006,
+ "loss": 4.419447898864746,
+ "step": 1478
+ },
+ {
+ "epoch": 20.544978165938865,
+ "grad_norm": 0.015877850353717804,
+ "learning_rate": 0.0006,
+ "loss": 4.312338829040527,
+ "step": 1479
+ },
+ {
+ "epoch": 20.558951965065503,
+ "grad_norm": 0.01569489948451519,
+ "learning_rate": 0.0006,
+ "loss": 4.572905540466309,
+ "step": 1480
+ },
+ {
+ "epoch": 20.57292576419214,
+ "grad_norm": 0.01591474376618862,
+ "learning_rate": 0.0006,
+ "loss": 4.414907455444336,
+ "step": 1481
+ },
+ {
+ "epoch": 20.586899563318777,
+ "grad_norm": 0.015395903028547764,
+ "learning_rate": 0.0006,
+ "loss": 4.549131393432617,
+ "step": 1482
+ },
+ {
+ "epoch": 20.600873362445416,
+ "grad_norm": 0.015715358778834343,
+ "learning_rate": 0.0006,
+ "loss": 4.530194282531738,
+ "step": 1483
+ },
+ {
+ "epoch": 20.61484716157205,
+ "grad_norm": 0.016042975708842278,
+ "learning_rate": 0.0006,
+ "loss": 4.322149276733398,
+ "step": 1484
+ },
+ {
+ "epoch": 20.62882096069869,
+ "grad_norm": 0.01664186641573906,
+ "learning_rate": 0.0006,
+ "loss": 4.424866676330566,
+ "step": 1485
+ },
+ {
+ "epoch": 20.64279475982533,
+ "grad_norm": 0.01631534844636917,
+ "learning_rate": 0.0006,
+ "loss": 4.441752910614014,
+ "step": 1486
+ },
+ {
+ "epoch": 20.656768558951963,
+ "grad_norm": 0.014057058840990067,
+ "learning_rate": 0.0006,
+ "loss": 4.399982452392578,
+ "step": 1487
+ },
+ {
+ "epoch": 20.670742358078602,
+ "grad_norm": 0.015076273120939732,
+ "learning_rate": 0.0006,
+ "loss": 4.441324234008789,
+ "step": 1488
+ },
+ {
+ "epoch": 20.68471615720524,
+ "grad_norm": 0.014138277620077133,
+ "learning_rate": 0.0006,
+ "loss": 4.4312543869018555,
+ "step": 1489
+ },
+ {
+ "epoch": 20.69868995633188,
+ "grad_norm": 0.013762440532445908,
+ "learning_rate": 0.0006,
+ "loss": 4.387807846069336,
+ "step": 1490
+ },
+ {
+ "epoch": 20.712663755458514,
+ "grad_norm": 0.015058411285281181,
+ "learning_rate": 0.0006,
+ "loss": 4.594747543334961,
+ "step": 1491
+ },
+ {
+ "epoch": 20.726637554585153,
+ "grad_norm": 0.013827123679220676,
+ "learning_rate": 0.0006,
+ "loss": 4.434604167938232,
+ "step": 1492
+ },
+ {
+ "epoch": 20.74061135371179,
+ "grad_norm": 0.015315295197069645,
+ "learning_rate": 0.0006,
+ "loss": 4.588066101074219,
+ "step": 1493
+ },
+ {
+ "epoch": 20.754585152838427,
+ "grad_norm": 0.018678732216358185,
+ "learning_rate": 0.0006,
+ "loss": 4.3514204025268555,
+ "step": 1494
+ },
+ {
+ "epoch": 20.768558951965066,
+ "grad_norm": 0.019521350041031837,
+ "learning_rate": 0.0006,
+ "loss": 4.437658786773682,
+ "step": 1495
+ },
+ {
+ "epoch": 20.782532751091704,
+ "grad_norm": 0.018363403156399727,
+ "learning_rate": 0.0006,
+ "loss": 4.495002269744873,
+ "step": 1496
+ },
+ {
+ "epoch": 20.79650655021834,
+ "grad_norm": 0.016830725595355034,
+ "learning_rate": 0.0006,
+ "loss": 4.402471542358398,
+ "step": 1497
+ },
+ {
+ "epoch": 20.810480349344978,
+ "grad_norm": 0.01722072623670101,
+ "learning_rate": 0.0006,
+ "loss": 4.437653541564941,
+ "step": 1498
+ },
+ {
+ "epoch": 20.824454148471617,
+ "grad_norm": 0.017459599301218987,
+ "learning_rate": 0.0006,
+ "loss": 4.366603851318359,
+ "step": 1499
+ },
+ {
+ "epoch": 20.83842794759825,
+ "grad_norm": 0.018902219831943512,
+ "learning_rate": 0.0006,
+ "loss": 4.501561164855957,
+ "step": 1500
+ },
+ {
+ "epoch": 20.85240174672489,
+ "grad_norm": 0.020814096555113792,
+ "learning_rate": 0.0006,
+ "loss": 4.535295486450195,
+ "step": 1501
+ },
+ {
+ "epoch": 20.86637554585153,
+ "grad_norm": 0.018990976735949516,
+ "learning_rate": 0.0006,
+ "loss": 4.333580017089844,
+ "step": 1502
+ },
+ {
+ "epoch": 20.880349344978168,
+ "grad_norm": 0.01837880350649357,
+ "learning_rate": 0.0006,
+ "loss": 4.313388824462891,
+ "step": 1503
+ },
+ {
+ "epoch": 20.894323144104803,
+ "grad_norm": 0.019820360466837883,
+ "learning_rate": 0.0006,
+ "loss": 4.388157844543457,
+ "step": 1504
+ },
+ {
+ "epoch": 20.90829694323144,
+ "grad_norm": 0.020181143656373024,
+ "learning_rate": 0.0006,
+ "loss": 4.572224140167236,
+ "step": 1505
+ },
+ {
+ "epoch": 20.92227074235808,
+ "grad_norm": 0.021206077188253403,
+ "learning_rate": 0.0006,
+ "loss": 4.3127970695495605,
+ "step": 1506
+ },
+ {
+ "epoch": 20.936244541484715,
+ "grad_norm": 0.017962060868740082,
+ "learning_rate": 0.0006,
+ "loss": 4.384734630584717,
+ "step": 1507
+ },
+ {
+ "epoch": 20.950218340611354,
+ "grad_norm": 0.01737220585346222,
+ "learning_rate": 0.0006,
+ "loss": 4.4075751304626465,
+ "step": 1508
+ },
+ {
+ "epoch": 20.964192139737992,
+ "grad_norm": 0.01839268207550049,
+ "learning_rate": 0.0006,
+ "loss": 4.553779125213623,
+ "step": 1509
+ },
+ {
+ "epoch": 20.978165938864628,
+ "grad_norm": 0.020102640613913536,
+ "learning_rate": 0.0006,
+ "loss": 4.399398326873779,
+ "step": 1510
+ },
+ {
+ "epoch": 20.992139737991266,
+ "grad_norm": 0.018831463530659676,
+ "learning_rate": 0.0006,
+ "loss": 4.358499526977539,
+ "step": 1511
+ },
+ {
+ "epoch": 21.0,
+ "grad_norm": 0.018254734575748444,
+ "learning_rate": 0.0006,
+ "loss": 4.384119510650635,
+ "step": 1512
+ },
+ {
+ "epoch": 21.0,
+ "eval_loss": 4.7524237632751465,
+ "eval_runtime": 57.8812,
+ "eval_samples_per_second": 42.19,
+ "eval_steps_per_second": 1.33,
+ "step": 1512
+ },
+ {
+ "epoch": 21.01397379912664,
+ "grad_norm": 0.019728587940335274,
+ "learning_rate": 0.0006,
+ "loss": 4.32363224029541,
+ "step": 1513
+ },
+ {
+ "epoch": 21.027947598253274,
+ "grad_norm": 0.02216625027358532,
+ "learning_rate": 0.0006,
+ "loss": 4.436988353729248,
+ "step": 1514
+ },
+ {
+ "epoch": 21.041921397379912,
+ "grad_norm": 0.02227606810629368,
+ "learning_rate": 0.0006,
+ "loss": 4.406063556671143,
+ "step": 1515
+ },
+ {
+ "epoch": 21.05589519650655,
+ "grad_norm": 0.020187009125947952,
+ "learning_rate": 0.0006,
+ "loss": 4.324402332305908,
+ "step": 1516
+ },
+ {
+ "epoch": 21.069868995633186,
+ "grad_norm": 0.02095157466828823,
+ "learning_rate": 0.0006,
+ "loss": 4.456544876098633,
+ "step": 1517
+ },
+ {
+ "epoch": 21.083842794759825,
+ "grad_norm": 0.02146013453602791,
+ "learning_rate": 0.0006,
+ "loss": 4.368387222290039,
+ "step": 1518
+ },
+ {
+ "epoch": 21.097816593886463,
+ "grad_norm": 0.02162035182118416,
+ "learning_rate": 0.0006,
+ "loss": 4.48838996887207,
+ "step": 1519
+ },
+ {
+ "epoch": 21.111790393013102,
+ "grad_norm": 0.021745676174759865,
+ "learning_rate": 0.0006,
+ "loss": 4.30778694152832,
+ "step": 1520
+ },
+ {
+ "epoch": 21.125764192139737,
+ "grad_norm": 0.02173588052392006,
+ "learning_rate": 0.0006,
+ "loss": 4.433149814605713,
+ "step": 1521
+ },
+ {
+ "epoch": 21.139737991266376,
+ "grad_norm": 0.020950648933649063,
+ "learning_rate": 0.0006,
+ "loss": 4.520089149475098,
+ "step": 1522
+ },
+ {
+ "epoch": 21.153711790393015,
+ "grad_norm": 0.020684752613306046,
+ "learning_rate": 0.0006,
+ "loss": 4.372437477111816,
+ "step": 1523
+ },
+ {
+ "epoch": 21.16768558951965,
+ "grad_norm": 0.019472869113087654,
+ "learning_rate": 0.0006,
+ "loss": 4.437015533447266,
+ "step": 1524
+ },
+ {
+ "epoch": 21.18165938864629,
+ "grad_norm": 0.01846962980926037,
+ "learning_rate": 0.0006,
+ "loss": 4.449531555175781,
+ "step": 1525
+ },
+ {
+ "epoch": 21.195633187772927,
+ "grad_norm": 0.017377803102135658,
+ "learning_rate": 0.0006,
+ "loss": 4.458279132843018,
+ "step": 1526
+ },
+ {
+ "epoch": 21.209606986899562,
+ "grad_norm": 0.016002200543880463,
+ "learning_rate": 0.0006,
+ "loss": 4.266297817230225,
+ "step": 1527
+ },
+ {
+ "epoch": 21.2235807860262,
+ "grad_norm": 0.015575903467833996,
+ "learning_rate": 0.0006,
+ "loss": 4.342632293701172,
+ "step": 1528
+ },
+ {
+ "epoch": 21.23755458515284,
+ "grad_norm": 0.015634698793292046,
+ "learning_rate": 0.0006,
+ "loss": 4.338623523712158,
+ "step": 1529
+ },
+ {
+ "epoch": 21.251528384279474,
+ "grad_norm": 0.01449581328779459,
+ "learning_rate": 0.0006,
+ "loss": 4.354330062866211,
+ "step": 1530
+ },
+ {
+ "epoch": 21.265502183406113,
+ "grad_norm": 0.01487616915255785,
+ "learning_rate": 0.0006,
+ "loss": 4.486835479736328,
+ "step": 1531
+ },
+ {
+ "epoch": 21.27947598253275,
+ "grad_norm": 0.016346555203199387,
+ "learning_rate": 0.0006,
+ "loss": 4.407098770141602,
+ "step": 1532
+ },
+ {
+ "epoch": 21.29344978165939,
+ "grad_norm": 0.01682182401418686,
+ "learning_rate": 0.0006,
+ "loss": 4.436809539794922,
+ "step": 1533
+ },
+ {
+ "epoch": 21.307423580786025,
+ "grad_norm": 0.017704004421830177,
+ "learning_rate": 0.0006,
+ "loss": 4.3854780197143555,
+ "step": 1534
+ },
+ {
+ "epoch": 21.321397379912664,
+ "grad_norm": 0.018437420949339867,
+ "learning_rate": 0.0006,
+ "loss": 4.413283824920654,
+ "step": 1535
+ },
+ {
+ "epoch": 21.335371179039303,
+ "grad_norm": 0.016911081969738007,
+ "learning_rate": 0.0006,
+ "loss": 4.357306003570557,
+ "step": 1536
+ },
+ {
+ "epoch": 21.349344978165938,
+ "grad_norm": 0.01642223633825779,
+ "learning_rate": 0.0006,
+ "loss": 4.422322750091553,
+ "step": 1537
+ },
+ {
+ "epoch": 21.363318777292577,
+ "grad_norm": 0.01780877821147442,
+ "learning_rate": 0.0006,
+ "loss": 4.387620449066162,
+ "step": 1538
+ },
+ {
+ "epoch": 21.377292576419215,
+ "grad_norm": 0.01653975248336792,
+ "learning_rate": 0.0006,
+ "loss": 4.327267646789551,
+ "step": 1539
+ },
+ {
+ "epoch": 21.39126637554585,
+ "grad_norm": 0.01572156324982643,
+ "learning_rate": 0.0006,
+ "loss": 4.475979804992676,
+ "step": 1540
+ },
+ {
+ "epoch": 21.40524017467249,
+ "grad_norm": 0.017156066372990608,
+ "learning_rate": 0.0006,
+ "loss": 4.4011125564575195,
+ "step": 1541
+ },
+ {
+ "epoch": 21.419213973799128,
+ "grad_norm": 0.01788829080760479,
+ "learning_rate": 0.0006,
+ "loss": 4.4174089431762695,
+ "step": 1542
+ },
+ {
+ "epoch": 21.433187772925763,
+ "grad_norm": 0.01728985086083412,
+ "learning_rate": 0.0006,
+ "loss": 4.429244041442871,
+ "step": 1543
+ },
+ {
+ "epoch": 21.4471615720524,
+ "grad_norm": 0.016534466296434402,
+ "learning_rate": 0.0006,
+ "loss": 4.326068878173828,
+ "step": 1544
+ },
+ {
+ "epoch": 21.46113537117904,
+ "grad_norm": 0.016582584008574486,
+ "learning_rate": 0.0006,
+ "loss": 4.2975969314575195,
+ "step": 1545
+ },
+ {
+ "epoch": 21.475109170305675,
+ "grad_norm": 0.018034178763628006,
+ "learning_rate": 0.0006,
+ "loss": 4.501319408416748,
+ "step": 1546
+ },
+ {
+ "epoch": 21.489082969432314,
+ "grad_norm": 0.017827702686190605,
+ "learning_rate": 0.0006,
+ "loss": 4.502284526824951,
+ "step": 1547
+ },
+ {
+ "epoch": 21.503056768558952,
+ "grad_norm": 0.017182299867272377,
+ "learning_rate": 0.0006,
+ "loss": 4.380828857421875,
+ "step": 1548
+ },
+ {
+ "epoch": 21.51703056768559,
+ "grad_norm": 0.019101588055491447,
+ "learning_rate": 0.0006,
+ "loss": 4.31564998626709,
+ "step": 1549
+ },
+ {
+ "epoch": 21.531004366812226,
+ "grad_norm": 0.0204677302390337,
+ "learning_rate": 0.0006,
+ "loss": 4.3840131759643555,
+ "step": 1550
+ },
+ {
+ "epoch": 21.544978165938865,
+ "grad_norm": 0.0188993439078331,
+ "learning_rate": 0.0006,
+ "loss": 4.4257731437683105,
+ "step": 1551
+ },
+ {
+ "epoch": 21.558951965065503,
+ "grad_norm": 0.01730230636894703,
+ "learning_rate": 0.0006,
+ "loss": 4.445132255554199,
+ "step": 1552
+ },
+ {
+ "epoch": 21.57292576419214,
+ "grad_norm": 0.01801574043929577,
+ "learning_rate": 0.0006,
+ "loss": 4.427064895629883,
+ "step": 1553
+ },
+ {
+ "epoch": 21.586899563318777,
+ "grad_norm": 0.018417565152049065,
+ "learning_rate": 0.0006,
+ "loss": 4.319595813751221,
+ "step": 1554
+ },
+ {
+ "epoch": 21.600873362445416,
+ "grad_norm": 0.01683160290122032,
+ "learning_rate": 0.0006,
+ "loss": 4.349218368530273,
+ "step": 1555
+ },
+ {
+ "epoch": 21.61484716157205,
+ "grad_norm": 0.016973106190562248,
+ "learning_rate": 0.0006,
+ "loss": 4.375990390777588,
+ "step": 1556
+ },
+ {
+ "epoch": 21.62882096069869,
+ "grad_norm": 0.015688583254814148,
+ "learning_rate": 0.0006,
+ "loss": 4.408170223236084,
+ "step": 1557
+ },
+ {
+ "epoch": 21.64279475982533,
+ "grad_norm": 0.01573043316602707,
+ "learning_rate": 0.0006,
+ "loss": 4.416255950927734,
+ "step": 1558
+ },
+ {
+ "epoch": 21.656768558951963,
+ "grad_norm": 0.01744293048977852,
+ "learning_rate": 0.0006,
+ "loss": 4.477141380310059,
+ "step": 1559
+ },
+ {
+ "epoch": 21.670742358078602,
+ "grad_norm": 0.017503513023257256,
+ "learning_rate": 0.0006,
+ "loss": 4.3367486000061035,
+ "step": 1560
+ },
+ {
+ "epoch": 21.68471615720524,
+ "grad_norm": 0.016592150554060936,
+ "learning_rate": 0.0006,
+ "loss": 4.314785480499268,
+ "step": 1561
+ },
+ {
+ "epoch": 21.69868995633188,
+ "grad_norm": 0.015774084255099297,
+ "learning_rate": 0.0006,
+ "loss": 4.448209285736084,
+ "step": 1562
+ },
+ {
+ "epoch": 21.712663755458514,
+ "grad_norm": 0.015779150649905205,
+ "learning_rate": 0.0006,
+ "loss": 4.392690658569336,
+ "step": 1563
+ },
+ {
+ "epoch": 21.726637554585153,
+ "grad_norm": 0.015392648056149483,
+ "learning_rate": 0.0006,
+ "loss": 4.437741756439209,
+ "step": 1564
+ },
+ {
+ "epoch": 21.74061135371179,
+ "grad_norm": 0.01530836895108223,
+ "learning_rate": 0.0006,
+ "loss": 4.424595832824707,
+ "step": 1565
+ },
+ {
+ "epoch": 21.754585152838427,
+ "grad_norm": 0.015600779093801975,
+ "learning_rate": 0.0006,
+ "loss": 4.335417747497559,
+ "step": 1566
+ },
+ {
+ "epoch": 21.768558951965066,
+ "grad_norm": 0.014646518044173717,
+ "learning_rate": 0.0006,
+ "loss": 4.279081344604492,
+ "step": 1567
+ },
+ {
+ "epoch": 21.782532751091704,
+ "grad_norm": 0.014132864773273468,
+ "learning_rate": 0.0006,
+ "loss": 4.345950126647949,
+ "step": 1568
+ },
+ {
+ "epoch": 21.79650655021834,
+ "grad_norm": 0.01425724383443594,
+ "learning_rate": 0.0006,
+ "loss": 4.277588844299316,
+ "step": 1569
+ },
+ {
+ "epoch": 21.810480349344978,
+ "grad_norm": 0.015396572649478912,
+ "learning_rate": 0.0006,
+ "loss": 4.326196670532227,
+ "step": 1570
+ },
+ {
+ "epoch": 21.824454148471617,
+ "grad_norm": 0.01683669537305832,
+ "learning_rate": 0.0006,
+ "loss": 4.3294172286987305,
+ "step": 1571
+ },
+ {
+ "epoch": 21.83842794759825,
+ "grad_norm": 0.016365647315979004,
+ "learning_rate": 0.0006,
+ "loss": 4.523321628570557,
+ "step": 1572
+ },
+ {
+ "epoch": 21.85240174672489,
+ "grad_norm": 0.017142774537205696,
+ "learning_rate": 0.0006,
+ "loss": 4.475842475891113,
+ "step": 1573
+ },
+ {
+ "epoch": 21.86637554585153,
+ "grad_norm": 0.015950839966535568,
+ "learning_rate": 0.0006,
+ "loss": 4.397187232971191,
+ "step": 1574
+ },
+ {
+ "epoch": 21.880349344978168,
+ "grad_norm": 0.016180362552404404,
+ "learning_rate": 0.0006,
+ "loss": 4.414778709411621,
+ "step": 1575
+ },
+ {
+ "epoch": 21.894323144104803,
+ "grad_norm": 0.017615556716918945,
+ "learning_rate": 0.0006,
+ "loss": 4.348567962646484,
+ "step": 1576
+ },
+ {
+ "epoch": 21.90829694323144,
+ "grad_norm": 0.01896277628839016,
+ "learning_rate": 0.0006,
+ "loss": 4.431762218475342,
+ "step": 1577
+ },
+ {
+ "epoch": 21.92227074235808,
+ "grad_norm": 0.01990971714258194,
+ "learning_rate": 0.0006,
+ "loss": 4.431285858154297,
+ "step": 1578
+ },
+ {
+ "epoch": 21.936244541484715,
+ "grad_norm": 0.0201258547604084,
+ "learning_rate": 0.0006,
+ "loss": 4.309872627258301,
+ "step": 1579
+ },
+ {
+ "epoch": 21.950218340611354,
+ "grad_norm": 0.019620198756456375,
+ "learning_rate": 0.0006,
+ "loss": 4.484044551849365,
+ "step": 1580
+ },
+ {
+ "epoch": 21.964192139737992,
+ "grad_norm": 0.01771123707294464,
+ "learning_rate": 0.0006,
+ "loss": 4.368093013763428,
+ "step": 1581
+ },
+ {
+ "epoch": 21.978165938864628,
+ "grad_norm": 0.017501654103398323,
+ "learning_rate": 0.0006,
+ "loss": 4.372651100158691,
+ "step": 1582
+ },
+ {
+ "epoch": 21.992139737991266,
+ "grad_norm": 0.01693608984351158,
+ "learning_rate": 0.0006,
+ "loss": 4.379167556762695,
+ "step": 1583
+ },
+ {
+ "epoch": 22.0,
+ "grad_norm": 0.02002849616110325,
+ "learning_rate": 0.0006,
+ "loss": 4.329627990722656,
+ "step": 1584
+ },
+ {
+ "epoch": 22.0,
+ "eval_loss": 4.723696231842041,
+ "eval_runtime": 56.1758,
+ "eval_samples_per_second": 43.471,
+ "eval_steps_per_second": 1.371,
+ "step": 1584
+ },
+ {
+ "epoch": 22.01397379912664,
+ "grad_norm": 0.02030305378139019,
+ "learning_rate": 0.0006,
+ "loss": 4.355566024780273,
+ "step": 1585
+ },
+ {
+ "epoch": 22.027947598253274,
+ "grad_norm": 0.018638934940099716,
+ "learning_rate": 0.0006,
+ "loss": 4.314009189605713,
+ "step": 1586
+ },
+ {
+ "epoch": 22.041921397379912,
+ "grad_norm": 0.019917072728276253,
+ "learning_rate": 0.0006,
+ "loss": 4.367525577545166,
+ "step": 1587
+ },
+ {
+ "epoch": 22.05589519650655,
+ "grad_norm": 0.020785167813301086,
+ "learning_rate": 0.0006,
+ "loss": 4.375561714172363,
+ "step": 1588
+ },
+ {
+ "epoch": 22.069868995633186,
+ "grad_norm": 0.021214453503489494,
+ "learning_rate": 0.0006,
+ "loss": 4.418305397033691,
+ "step": 1589
+ },
+ {
+ "epoch": 22.083842794759825,
+ "grad_norm": 0.019051678478717804,
+ "learning_rate": 0.0006,
+ "loss": 4.215126991271973,
+ "step": 1590
+ },
+ {
+ "epoch": 22.097816593886463,
+ "grad_norm": 0.02323366142809391,
+ "learning_rate": 0.0006,
+ "loss": 4.306807994842529,
+ "step": 1591
+ },
+ {
+ "epoch": 22.111790393013102,
+ "grad_norm": 0.026186738163232803,
+ "learning_rate": 0.0006,
+ "loss": 4.408508777618408,
+ "step": 1592
+ },
+ {
+ "epoch": 22.125764192139737,
+ "grad_norm": 0.027361227199435234,
+ "learning_rate": 0.0006,
+ "loss": 4.329322814941406,
+ "step": 1593
+ },
+ {
+ "epoch": 22.139737991266376,
+ "grad_norm": 0.026564646512269974,
+ "learning_rate": 0.0006,
+ "loss": 4.346642017364502,
+ "step": 1594
+ },
+ {
+ "epoch": 22.153711790393015,
+ "grad_norm": 0.025073667988181114,
+ "learning_rate": 0.0006,
+ "loss": 4.3984198570251465,
+ "step": 1595
+ },
+ {
+ "epoch": 22.16768558951965,
+ "grad_norm": 0.02347794733941555,
+ "learning_rate": 0.0006,
+ "loss": 4.352538108825684,
+ "step": 1596
+ },
+ {
+ "epoch": 22.18165938864629,
+ "grad_norm": 0.02203468047082424,
+ "learning_rate": 0.0006,
+ "loss": 4.298148155212402,
+ "step": 1597
+ },
+ {
+ "epoch": 22.195633187772927,
+ "grad_norm": 0.020621957257390022,
+ "learning_rate": 0.0006,
+ "loss": 4.480309963226318,
+ "step": 1598
+ },
+ {
+ "epoch": 22.209606986899562,
+ "grad_norm": 0.023710055276751518,
+ "learning_rate": 0.0006,
+ "loss": 4.339317321777344,
+ "step": 1599
+ },
+ {
+ "epoch": 22.2235807860262,
+ "grad_norm": 0.0233482476323843,
+ "learning_rate": 0.0006,
+ "loss": 4.436896324157715,
+ "step": 1600
+ },
+ {
+ "epoch": 22.23755458515284,
+ "grad_norm": 0.020122403278946877,
+ "learning_rate": 0.0006,
+ "loss": 4.236002445220947,
+ "step": 1601
+ },
+ {
+ "epoch": 22.251528384279474,
+ "grad_norm": 0.01957266591489315,
+ "learning_rate": 0.0006,
+ "loss": 4.497328281402588,
+ "step": 1602
+ },
+ {
+ "epoch": 22.265502183406113,
+ "grad_norm": 0.018964895978569984,
+ "learning_rate": 0.0006,
+ "loss": 4.36360502243042,
+ "step": 1603
+ },
+ {
+ "epoch": 22.27947598253275,
+ "grad_norm": 0.018902769312262535,
+ "learning_rate": 0.0006,
+ "loss": 4.427093505859375,
+ "step": 1604
+ },
+ {
+ "epoch": 22.29344978165939,
+ "grad_norm": 0.018918119370937347,
+ "learning_rate": 0.0006,
+ "loss": 4.294950485229492,
+ "step": 1605
+ },
+ {
+ "epoch": 22.307423580786025,
+ "grad_norm": 0.019315775483846664,
+ "learning_rate": 0.0006,
+ "loss": 4.427952289581299,
+ "step": 1606
+ },
+ {
+ "epoch": 22.321397379912664,
+ "grad_norm": 0.018350522965192795,
+ "learning_rate": 0.0006,
+ "loss": 4.272945404052734,
+ "step": 1607
+ },
+ {
+ "epoch": 22.335371179039303,
+ "grad_norm": 0.01759534701704979,
+ "learning_rate": 0.0006,
+ "loss": 4.348023891448975,
+ "step": 1608
+ },
+ {
+ "epoch": 22.349344978165938,
+ "grad_norm": 0.01646057888865471,
+ "learning_rate": 0.0006,
+ "loss": 4.278518199920654,
+ "step": 1609
+ },
+ {
+ "epoch": 22.363318777292577,
+ "grad_norm": 0.015875060111284256,
+ "learning_rate": 0.0006,
+ "loss": 4.423527240753174,
+ "step": 1610
+ },
+ {
+ "epoch": 22.377292576419215,
+ "grad_norm": 0.01638529635965824,
+ "learning_rate": 0.0006,
+ "loss": 4.321723937988281,
+ "step": 1611
+ },
+ {
+ "epoch": 22.39126637554585,
+ "grad_norm": 0.016942061483860016,
+ "learning_rate": 0.0006,
+ "loss": 4.346164703369141,
+ "step": 1612
+ },
+ {
+ "epoch": 22.40524017467249,
+ "grad_norm": 0.015209621749818325,
+ "learning_rate": 0.0006,
+ "loss": 4.434340000152588,
+ "step": 1613
+ },
+ {
+ "epoch": 22.419213973799128,
+ "grad_norm": 0.014423009008169174,
+ "learning_rate": 0.0006,
+ "loss": 4.342560768127441,
+ "step": 1614
+ },
+ {
+ "epoch": 22.433187772925763,
+ "grad_norm": 0.014369525946676731,
+ "learning_rate": 0.0006,
+ "loss": 4.296586990356445,
+ "step": 1615
+ },
+ {
+ "epoch": 22.4471615720524,
+ "grad_norm": 0.01392540242522955,
+ "learning_rate": 0.0006,
+ "loss": 4.433176040649414,
+ "step": 1616
+ },
+ {
+ "epoch": 22.46113537117904,
+ "grad_norm": 0.013206899166107178,
+ "learning_rate": 0.0006,
+ "loss": 4.32962703704834,
+ "step": 1617
+ },
+ {
+ "epoch": 22.475109170305675,
+ "grad_norm": 0.013427077792584896,
+ "learning_rate": 0.0006,
+ "loss": 4.420437335968018,
+ "step": 1618
+ },
+ {
+ "epoch": 22.489082969432314,
+ "grad_norm": 0.014194196090102196,
+ "learning_rate": 0.0006,
+ "loss": 4.410478591918945,
+ "step": 1619
+ },
+ {
+ "epoch": 22.503056768558952,
+ "grad_norm": 0.01362364087253809,
+ "learning_rate": 0.0006,
+ "loss": 4.338932037353516,
+ "step": 1620
+ },
+ {
+ "epoch": 22.51703056768559,
+ "grad_norm": 0.014299589209258556,
+ "learning_rate": 0.0006,
+ "loss": 4.34089994430542,
+ "step": 1621
+ },
+ {
+ "epoch": 22.531004366812226,
+ "grad_norm": 0.014061295427381992,
+ "learning_rate": 0.0006,
+ "loss": 4.29201602935791,
+ "step": 1622
+ },
+ {
+ "epoch": 22.544978165938865,
+ "grad_norm": 0.013945229351520538,
+ "learning_rate": 0.0006,
+ "loss": 4.326052188873291,
+ "step": 1623
+ },
+ {
+ "epoch": 22.558951965065503,
+ "grad_norm": 0.01387068536132574,
+ "learning_rate": 0.0006,
+ "loss": 4.393521308898926,
+ "step": 1624
+ },
+ {
+ "epoch": 22.57292576419214,
+ "grad_norm": 0.013931898400187492,
+ "learning_rate": 0.0006,
+ "loss": 4.3739752769470215,
+ "step": 1625
+ },
+ {
+ "epoch": 22.586899563318777,
+ "grad_norm": 0.01497740764170885,
+ "learning_rate": 0.0006,
+ "loss": 4.3991618156433105,
+ "step": 1626
+ },
+ {
+ "epoch": 22.600873362445416,
+ "grad_norm": 0.015842905268073082,
+ "learning_rate": 0.0006,
+ "loss": 4.405340671539307,
+ "step": 1627
+ },
+ {
+ "epoch": 22.61484716157205,
+ "grad_norm": 0.01580633595585823,
+ "learning_rate": 0.0006,
+ "loss": 4.280102729797363,
+ "step": 1628
+ },
+ {
+ "epoch": 22.62882096069869,
+ "grad_norm": 0.016567382961511612,
+ "learning_rate": 0.0006,
+ "loss": 4.407657146453857,
+ "step": 1629
+ },
+ {
+ "epoch": 22.64279475982533,
+ "grad_norm": 0.01648622378706932,
+ "learning_rate": 0.0006,
+ "loss": 4.392590522766113,
+ "step": 1630
+ },
+ {
+ "epoch": 22.656768558951963,
+ "grad_norm": 0.017066307365894318,
+ "learning_rate": 0.0006,
+ "loss": 4.324862003326416,
+ "step": 1631
+ },
+ {
+ "epoch": 22.670742358078602,
+ "grad_norm": 0.01637062057852745,
+ "learning_rate": 0.0006,
+ "loss": 4.34598445892334,
+ "step": 1632
+ },
+ {
+ "epoch": 22.68471615720524,
+ "grad_norm": 0.018038960173726082,
+ "learning_rate": 0.0006,
+ "loss": 4.320711612701416,
+ "step": 1633
+ },
+ {
+ "epoch": 22.69868995633188,
+ "grad_norm": 0.018096428364515305,
+ "learning_rate": 0.0006,
+ "loss": 4.536527156829834,
+ "step": 1634
+ },
+ {
+ "epoch": 22.712663755458514,
+ "grad_norm": 0.017889291048049927,
+ "learning_rate": 0.0006,
+ "loss": 4.388131141662598,
+ "step": 1635
+ },
+ {
+ "epoch": 22.726637554585153,
+ "grad_norm": 0.018771521747112274,
+ "learning_rate": 0.0006,
+ "loss": 4.357668876647949,
+ "step": 1636
+ },
+ {
+ "epoch": 22.74061135371179,
+ "grad_norm": 0.02100873365998268,
+ "learning_rate": 0.0006,
+ "loss": 4.293972015380859,
+ "step": 1637
+ },
+ {
+ "epoch": 22.754585152838427,
+ "grad_norm": 0.017715785652399063,
+ "learning_rate": 0.0006,
+ "loss": 4.349849224090576,
+ "step": 1638
+ },
+ {
+ "epoch": 22.768558951965066,
+ "grad_norm": 0.015546333976089954,
+ "learning_rate": 0.0006,
+ "loss": 4.311728000640869,
+ "step": 1639
+ },
+ {
+ "epoch": 22.782532751091704,
+ "grad_norm": 0.01605990156531334,
+ "learning_rate": 0.0006,
+ "loss": 4.499123573303223,
+ "step": 1640
+ },
+ {
+ "epoch": 22.79650655021834,
+ "grad_norm": 0.016642142087221146,
+ "learning_rate": 0.0006,
+ "loss": 4.4319257736206055,
+ "step": 1641
+ },
+ {
+ "epoch": 22.810480349344978,
+ "grad_norm": 0.016120387241244316,
+ "learning_rate": 0.0006,
+ "loss": 4.4373650550842285,
+ "step": 1642
+ },
+ {
+ "epoch": 22.824454148471617,
+ "grad_norm": 0.015174890868365765,
+ "learning_rate": 0.0006,
+ "loss": 4.314082145690918,
+ "step": 1643
+ },
+ {
+ "epoch": 22.83842794759825,
+ "grad_norm": 0.01551714539527893,
+ "learning_rate": 0.0006,
+ "loss": 4.360363960266113,
+ "step": 1644
+ },
+ {
+ "epoch": 22.85240174672489,
+ "grad_norm": 0.014966742135584354,
+ "learning_rate": 0.0006,
+ "loss": 4.346253395080566,
+ "step": 1645
+ },
+ {
+ "epoch": 22.86637554585153,
+ "grad_norm": 0.014288028702139854,
+ "learning_rate": 0.0006,
+ "loss": 4.464366436004639,
+ "step": 1646
+ },
+ {
+ "epoch": 22.880349344978168,
+ "grad_norm": 0.015291682444512844,
+ "learning_rate": 0.0006,
+ "loss": 4.504101753234863,
+ "step": 1647
+ },
+ {
+ "epoch": 22.894323144104803,
+ "grad_norm": 0.014534426853060722,
+ "learning_rate": 0.0006,
+ "loss": 4.462172031402588,
+ "step": 1648
+ },
+ {
+ "epoch": 22.90829694323144,
+ "grad_norm": 0.014129819348454475,
+ "learning_rate": 0.0006,
+ "loss": 4.350126266479492,
+ "step": 1649
+ },
+ {
+ "epoch": 22.92227074235808,
+ "grad_norm": 0.014932668767869473,
+ "learning_rate": 0.0006,
+ "loss": 4.374392032623291,
+ "step": 1650
+ },
+ {
+ "epoch": 22.936244541484715,
+ "grad_norm": 0.014647853560745716,
+ "learning_rate": 0.0006,
+ "loss": 4.252264499664307,
+ "step": 1651
+ },
+ {
+ "epoch": 22.950218340611354,
+ "grad_norm": 0.013553360477089882,
+ "learning_rate": 0.0006,
+ "loss": 4.379729270935059,
+ "step": 1652
+ },
+ {
+ "epoch": 22.964192139737992,
+ "grad_norm": 0.014606194570660591,
+ "learning_rate": 0.0006,
+ "loss": 4.363372802734375,
+ "step": 1653
+ },
+ {
+ "epoch": 22.978165938864628,
+ "grad_norm": 0.015000631101429462,
+ "learning_rate": 0.0006,
+ "loss": 4.408048629760742,
+ "step": 1654
+ },
+ {
+ "epoch": 22.992139737991266,
+ "grad_norm": 0.015370816923677921,
+ "learning_rate": 0.0006,
+ "loss": 4.297451019287109,
+ "step": 1655
+ },
+ {
+ "epoch": 23.0,
+ "grad_norm": 0.01709144562482834,
+ "learning_rate": 0.0006,
+ "loss": 4.294776916503906,
+ "step": 1656
+ },
+ {
+ "epoch": 23.0,
+ "eval_loss": 4.678618907928467,
+ "eval_runtime": 56.1668,
+ "eval_samples_per_second": 43.478,
+ "eval_steps_per_second": 1.371,
+ "step": 1656
+ },
+ {
+ "epoch": 23.01397379912664,
+ "grad_norm": 0.01727290451526642,
+ "learning_rate": 0.0006,
+ "loss": 4.311701774597168,
+ "step": 1657
+ },
+ {
+ "epoch": 23.027947598253274,
+ "grad_norm": 0.020645759999752045,
+ "learning_rate": 0.0006,
+ "loss": 4.429897308349609,
+ "step": 1658
+ },
+ {
+ "epoch": 23.041921397379912,
+ "grad_norm": 0.02204226516187191,
+ "learning_rate": 0.0006,
+ "loss": 4.365151882171631,
+ "step": 1659
+ },
+ {
+ "epoch": 23.05589519650655,
+ "grad_norm": 0.023069269955158234,
+ "learning_rate": 0.0006,
+ "loss": 4.3710832595825195,
+ "step": 1660
+ },
+ {
+ "epoch": 23.069868995633186,
+ "grad_norm": 0.024151179939508438,
+ "learning_rate": 0.0006,
+ "loss": 4.403108596801758,
+ "step": 1661
+ },
+ {
+ "epoch": 23.083842794759825,
+ "grad_norm": 0.02511233650147915,
+ "learning_rate": 0.0006,
+ "loss": 4.414128303527832,
+ "step": 1662
+ },
+ {
+ "epoch": 23.097816593886463,
+ "grad_norm": 0.025802113115787506,
+ "learning_rate": 0.0006,
+ "loss": 4.388265609741211,
+ "step": 1663
+ },
+ {
+ "epoch": 23.111790393013102,
+ "grad_norm": 0.02377348765730858,
+ "learning_rate": 0.0006,
+ "loss": 4.357505798339844,
+ "step": 1664
+ },
+ {
+ "epoch": 23.125764192139737,
+ "grad_norm": 0.020876651629805565,
+ "learning_rate": 0.0006,
+ "loss": 4.23043966293335,
+ "step": 1665
+ },
+ {
+ "epoch": 23.139737991266376,
+ "grad_norm": 0.01959727331995964,
+ "learning_rate": 0.0006,
+ "loss": 4.280970096588135,
+ "step": 1666
+ },
+ {
+ "epoch": 23.153711790393015,
+ "grad_norm": 0.018678616732358932,
+ "learning_rate": 0.0006,
+ "loss": 4.346716403961182,
+ "step": 1667
+ },
+ {
+ "epoch": 23.16768558951965,
+ "grad_norm": 0.02108944021165371,
+ "learning_rate": 0.0006,
+ "loss": 4.277904510498047,
+ "step": 1668
+ },
+ {
+ "epoch": 23.18165938864629,
+ "grad_norm": 0.020221566781401634,
+ "learning_rate": 0.0006,
+ "loss": 4.367693901062012,
+ "step": 1669
+ },
+ {
+ "epoch": 23.195633187772927,
+ "grad_norm": 0.019599711522459984,
+ "learning_rate": 0.0006,
+ "loss": 4.257092475891113,
+ "step": 1670
+ },
+ {
+ "epoch": 23.209606986899562,
+ "grad_norm": 0.019833361729979515,
+ "learning_rate": 0.0006,
+ "loss": 4.540737628936768,
+ "step": 1671
+ },
+ {
+ "epoch": 23.2235807860262,
+ "grad_norm": 0.020352095365524292,
+ "learning_rate": 0.0006,
+ "loss": 4.360923767089844,
+ "step": 1672
+ },
+ {
+ "epoch": 23.23755458515284,
+ "grad_norm": 0.01939748041331768,
+ "learning_rate": 0.0006,
+ "loss": 4.393110275268555,
+ "step": 1673
+ },
+ {
+ "epoch": 23.251528384279474,
+ "grad_norm": 0.01884354278445244,
+ "learning_rate": 0.0006,
+ "loss": 4.383938789367676,
+ "step": 1674
+ },
+ {
+ "epoch": 23.265502183406113,
+ "grad_norm": 0.02008470520377159,
+ "learning_rate": 0.0006,
+ "loss": 4.346344947814941,
+ "step": 1675
+ },
+ {
+ "epoch": 23.27947598253275,
+ "grad_norm": 0.019065195694565773,
+ "learning_rate": 0.0006,
+ "loss": 4.36611795425415,
+ "step": 1676
+ },
+ {
+ "epoch": 23.29344978165939,
+ "grad_norm": 0.01938541606068611,
+ "learning_rate": 0.0006,
+ "loss": 4.4009599685668945,
+ "step": 1677
+ },
+ {
+ "epoch": 23.307423580786025,
+ "grad_norm": 0.019044723361730576,
+ "learning_rate": 0.0006,
+ "loss": 4.340169906616211,
+ "step": 1678
+ },
+ {
+ "epoch": 23.321397379912664,
+ "grad_norm": 0.018850378692150116,
+ "learning_rate": 0.0006,
+ "loss": 4.37674617767334,
+ "step": 1679
+ },
+ {
+ "epoch": 23.335371179039303,
+ "grad_norm": 0.019076095893979073,
+ "learning_rate": 0.0006,
+ "loss": 4.226874351501465,
+ "step": 1680
+ },
+ {
+ "epoch": 23.349344978165938,
+ "grad_norm": 0.017543038353323936,
+ "learning_rate": 0.0006,
+ "loss": 4.337751388549805,
+ "step": 1681
+ },
+ {
+ "epoch": 23.363318777292577,
+ "grad_norm": 0.017170218750834465,
+ "learning_rate": 0.0006,
+ "loss": 4.282512664794922,
+ "step": 1682
+ },
+ {
+ "epoch": 23.377292576419215,
+ "grad_norm": 0.01574796624481678,
+ "learning_rate": 0.0006,
+ "loss": 4.30638313293457,
+ "step": 1683
+ },
+ {
+ "epoch": 23.39126637554585,
+ "grad_norm": 0.015160846523940563,
+ "learning_rate": 0.0006,
+ "loss": 4.336716175079346,
+ "step": 1684
+ },
+ {
+ "epoch": 23.40524017467249,
+ "grad_norm": 0.01534526702016592,
+ "learning_rate": 0.0006,
+ "loss": 4.2174787521362305,
+ "step": 1685
+ },
+ {
+ "epoch": 23.419213973799128,
+ "grad_norm": 0.015289999544620514,
+ "learning_rate": 0.0006,
+ "loss": 4.272546291351318,
+ "step": 1686
+ },
+ {
+ "epoch": 23.433187772925763,
+ "grad_norm": 0.014894185587763786,
+ "learning_rate": 0.0006,
+ "loss": 4.296045303344727,
+ "step": 1687
+ },
+ {
+ "epoch": 23.4471615720524,
+ "grad_norm": 0.014869502745568752,
+ "learning_rate": 0.0006,
+ "loss": 4.425461769104004,
+ "step": 1688
+ },
+ {
+ "epoch": 23.46113537117904,
+ "grad_norm": 0.01474483497440815,
+ "learning_rate": 0.0006,
+ "loss": 4.361867427825928,
+ "step": 1689
+ },
+ {
+ "epoch": 23.475109170305675,
+ "grad_norm": 0.015299245715141296,
+ "learning_rate": 0.0006,
+ "loss": 4.364523410797119,
+ "step": 1690
+ },
+ {
+ "epoch": 23.489082969432314,
+ "grad_norm": 0.015201340429484844,
+ "learning_rate": 0.0006,
+ "loss": 4.285182952880859,
+ "step": 1691
+ },
+ {
+ "epoch": 23.503056768558952,
+ "grad_norm": 0.016057293862104416,
+ "learning_rate": 0.0006,
+ "loss": 4.284873962402344,
+ "step": 1692
+ },
+ {
+ "epoch": 23.51703056768559,
+ "grad_norm": 0.015871062874794006,
+ "learning_rate": 0.0006,
+ "loss": 4.326470851898193,
+ "step": 1693
+ },
+ {
+ "epoch": 23.531004366812226,
+ "grad_norm": 0.01586288772523403,
+ "learning_rate": 0.0006,
+ "loss": 4.324883460998535,
+ "step": 1694
+ },
+ {
+ "epoch": 23.544978165938865,
+ "grad_norm": 0.016092827543616295,
+ "learning_rate": 0.0006,
+ "loss": 4.3258867263793945,
+ "step": 1695
+ },
+ {
+ "epoch": 23.558951965065503,
+ "grad_norm": 0.01582016795873642,
+ "learning_rate": 0.0006,
+ "loss": 4.358892440795898,
+ "step": 1696
+ },
+ {
+ "epoch": 23.57292576419214,
+ "grad_norm": 0.01688452623784542,
+ "learning_rate": 0.0006,
+ "loss": 4.280842304229736,
+ "step": 1697
+ },
+ {
+ "epoch": 23.586899563318777,
+ "grad_norm": 0.015865886583924294,
+ "learning_rate": 0.0006,
+ "loss": 4.245567321777344,
+ "step": 1698
+ },
+ {
+ "epoch": 23.600873362445416,
+ "grad_norm": 0.015303662046790123,
+ "learning_rate": 0.0006,
+ "loss": 4.383112907409668,
+ "step": 1699
+ },
+ {
+ "epoch": 23.61484716157205,
+ "grad_norm": 0.015649516135454178,
+ "learning_rate": 0.0006,
+ "loss": 4.519082069396973,
+ "step": 1700
+ },
+ {
+ "epoch": 23.62882096069869,
+ "grad_norm": 0.0167120099067688,
+ "learning_rate": 0.0006,
+ "loss": 4.447847366333008,
+ "step": 1701
+ },
+ {
+ "epoch": 23.64279475982533,
+ "grad_norm": 0.016368716955184937,
+ "learning_rate": 0.0006,
+ "loss": 4.434186935424805,
+ "step": 1702
+ },
+ {
+ "epoch": 23.656768558951963,
+ "grad_norm": 0.017379503697156906,
+ "learning_rate": 0.0006,
+ "loss": 4.317594528198242,
+ "step": 1703
+ },
+ {
+ "epoch": 23.670742358078602,
+ "grad_norm": 0.018573811277747154,
+ "learning_rate": 0.0006,
+ "loss": 4.205793857574463,
+ "step": 1704
+ },
+ {
+ "epoch": 23.68471615720524,
+ "grad_norm": 0.01749090477824211,
+ "learning_rate": 0.0006,
+ "loss": 4.3985700607299805,
+ "step": 1705
+ },
+ {
+ "epoch": 23.69868995633188,
+ "grad_norm": 0.0165668074041605,
+ "learning_rate": 0.0006,
+ "loss": 4.329226970672607,
+ "step": 1706
+ },
+ {
+ "epoch": 23.712663755458514,
+ "grad_norm": 0.016913846135139465,
+ "learning_rate": 0.0006,
+ "loss": 4.388485908508301,
+ "step": 1707
+ },
+ {
+ "epoch": 23.726637554585153,
+ "grad_norm": 0.016813859343528748,
+ "learning_rate": 0.0006,
+ "loss": 4.3395915031433105,
+ "step": 1708
+ },
+ {
+ "epoch": 23.74061135371179,
+ "grad_norm": 0.01676975190639496,
+ "learning_rate": 0.0006,
+ "loss": 4.254745960235596,
+ "step": 1709
+ },
+ {
+ "epoch": 23.754585152838427,
+ "grad_norm": 0.016370350494980812,
+ "learning_rate": 0.0006,
+ "loss": 4.2612223625183105,
+ "step": 1710
+ },
+ {
+ "epoch": 23.768558951965066,
+ "grad_norm": 0.01696198247373104,
+ "learning_rate": 0.0006,
+ "loss": 4.327524662017822,
+ "step": 1711
+ },
+ {
+ "epoch": 23.782532751091704,
+ "grad_norm": 0.0189108457416296,
+ "learning_rate": 0.0006,
+ "loss": 4.325839519500732,
+ "step": 1712
+ },
+ {
+ "epoch": 23.79650655021834,
+ "grad_norm": 0.01862351782619953,
+ "learning_rate": 0.0006,
+ "loss": 4.333662509918213,
+ "step": 1713
+ },
+ {
+ "epoch": 23.810480349344978,
+ "grad_norm": 0.01655014418065548,
+ "learning_rate": 0.0006,
+ "loss": 4.39305305480957,
+ "step": 1714
+ },
+ {
+ "epoch": 23.824454148471617,
+ "grad_norm": 0.015899376943707466,
+ "learning_rate": 0.0006,
+ "loss": 4.390053749084473,
+ "step": 1715
+ },
+ {
+ "epoch": 23.83842794759825,
+ "grad_norm": 0.016519056633114815,
+ "learning_rate": 0.0006,
+ "loss": 4.255903244018555,
+ "step": 1716
+ },
+ {
+ "epoch": 23.85240174672489,
+ "grad_norm": 0.0167539119720459,
+ "learning_rate": 0.0006,
+ "loss": 4.255727767944336,
+ "step": 1717
+ },
+ {
+ "epoch": 23.86637554585153,
+ "grad_norm": 0.0171353816986084,
+ "learning_rate": 0.0006,
+ "loss": 4.376740455627441,
+ "step": 1718
+ },
+ {
+ "epoch": 23.880349344978168,
+ "grad_norm": 0.016719117760658264,
+ "learning_rate": 0.0006,
+ "loss": 4.3376688957214355,
+ "step": 1719
+ },
+ {
+ "epoch": 23.894323144104803,
+ "grad_norm": 0.01561494916677475,
+ "learning_rate": 0.0006,
+ "loss": 4.3345746994018555,
+ "step": 1720
+ },
+ {
+ "epoch": 23.90829694323144,
+ "grad_norm": 0.016303174197673798,
+ "learning_rate": 0.0006,
+ "loss": 4.332371234893799,
+ "step": 1721
+ },
+ {
+ "epoch": 23.92227074235808,
+ "grad_norm": 0.016722865402698517,
+ "learning_rate": 0.0006,
+ "loss": 4.267125129699707,
+ "step": 1722
+ },
+ {
+ "epoch": 23.936244541484715,
+ "grad_norm": 0.017072124406695366,
+ "learning_rate": 0.0006,
+ "loss": 4.322751998901367,
+ "step": 1723
+ },
+ {
+ "epoch": 23.950218340611354,
+ "grad_norm": 0.01699932850897312,
+ "learning_rate": 0.0006,
+ "loss": 4.378422737121582,
+ "step": 1724
+ },
+ {
+ "epoch": 23.964192139737992,
+ "grad_norm": 0.01742720976471901,
+ "learning_rate": 0.0006,
+ "loss": 4.345083236694336,
+ "step": 1725
+ },
+ {
+ "epoch": 23.978165938864628,
+ "grad_norm": 0.018797501921653748,
+ "learning_rate": 0.0006,
+ "loss": 4.290443420410156,
+ "step": 1726
+ },
+ {
+ "epoch": 23.992139737991266,
+ "grad_norm": 0.018090683966875076,
+ "learning_rate": 0.0006,
+ "loss": 4.377580642700195,
+ "step": 1727
+ },
+ {
+ "epoch": 24.0,
+ "grad_norm": 0.017305459827184677,
+ "learning_rate": 0.0006,
+ "loss": 4.364426612854004,
+ "step": 1728
+ },
+ {
+ "epoch": 24.0,
+ "eval_loss": 4.647762298583984,
+ "eval_runtime": 56.5402,
+ "eval_samples_per_second": 43.191,
+ "eval_steps_per_second": 1.362,
+ "step": 1728
+ },
+ {
+ "epoch": 24.01397379912664,
+ "grad_norm": 0.016493607312440872,
+ "learning_rate": 0.0006,
+ "loss": 4.252255916595459,
+ "step": 1729
+ },
+ {
+ "epoch": 24.027947598253274,
+ "grad_norm": 0.016529450193047523,
+ "learning_rate": 0.0006,
+ "loss": 4.352746963500977,
+ "step": 1730
+ },
+ {
+ "epoch": 24.041921397379912,
+ "grad_norm": 0.016940191388130188,
+ "learning_rate": 0.0006,
+ "loss": 4.301870822906494,
+ "step": 1731
+ },
+ {
+ "epoch": 24.05589519650655,
+ "grad_norm": 0.01752558909356594,
+ "learning_rate": 0.0006,
+ "loss": 4.2762451171875,
+ "step": 1732
+ },
+ {
+ "epoch": 24.069868995633186,
+ "grad_norm": 0.017112884670495987,
+ "learning_rate": 0.0006,
+ "loss": 4.2659149169921875,
+ "step": 1733
+ },
+ {
+ "epoch": 24.083842794759825,
+ "grad_norm": 0.01900586113333702,
+ "learning_rate": 0.0006,
+ "loss": 4.306058883666992,
+ "step": 1734
+ },
+ {
+ "epoch": 24.097816593886463,
+ "grad_norm": 0.019447680562734604,
+ "learning_rate": 0.0006,
+ "loss": 4.450128555297852,
+ "step": 1735
+ },
+ {
+ "epoch": 24.111790393013102,
+ "grad_norm": 0.01948186755180359,
+ "learning_rate": 0.0006,
+ "loss": 4.202869415283203,
+ "step": 1736
+ },
+ {
+ "epoch": 24.125764192139737,
+ "grad_norm": 0.01716383546590805,
+ "learning_rate": 0.0006,
+ "loss": 4.318876266479492,
+ "step": 1737
+ },
+ {
+ "epoch": 24.139737991266376,
+ "grad_norm": 0.017049988731741905,
+ "learning_rate": 0.0006,
+ "loss": 4.237238883972168,
+ "step": 1738
+ },
+ {
+ "epoch": 24.153711790393015,
+ "grad_norm": 0.015169923193752766,
+ "learning_rate": 0.0006,
+ "loss": 4.289426803588867,
+ "step": 1739
+ },
+ {
+ "epoch": 24.16768558951965,
+ "grad_norm": 0.014760667458176613,
+ "learning_rate": 0.0006,
+ "loss": 4.356924057006836,
+ "step": 1740
+ },
+ {
+ "epoch": 24.18165938864629,
+ "grad_norm": 0.015127750113606453,
+ "learning_rate": 0.0006,
+ "loss": 4.142916679382324,
+ "step": 1741
+ },
+ {
+ "epoch": 24.195633187772927,
+ "grad_norm": 0.01549555640667677,
+ "learning_rate": 0.0006,
+ "loss": 4.345563888549805,
+ "step": 1742
+ },
+ {
+ "epoch": 24.209606986899562,
+ "grad_norm": 0.016393091529607773,
+ "learning_rate": 0.0006,
+ "loss": 4.417495250701904,
+ "step": 1743
+ },
+ {
+ "epoch": 24.2235807860262,
+ "grad_norm": 0.01785266026854515,
+ "learning_rate": 0.0006,
+ "loss": 4.3553547859191895,
+ "step": 1744
+ },
+ {
+ "epoch": 24.23755458515284,
+ "grad_norm": 0.017574617639183998,
+ "learning_rate": 0.0006,
+ "loss": 4.241024494171143,
+ "step": 1745
+ },
+ {
+ "epoch": 24.251528384279474,
+ "grad_norm": 0.019060306251049042,
+ "learning_rate": 0.0006,
+ "loss": 4.3685150146484375,
+ "step": 1746
+ },
+ {
+ "epoch": 24.265502183406113,
+ "grad_norm": 0.0214633010327816,
+ "learning_rate": 0.0006,
+ "loss": 4.323866844177246,
+ "step": 1747
+ },
+ {
+ "epoch": 24.27947598253275,
+ "grad_norm": 0.021055003628134727,
+ "learning_rate": 0.0006,
+ "loss": 4.278757095336914,
+ "step": 1748
+ },
+ {
+ "epoch": 24.29344978165939,
+ "grad_norm": 0.018618572503328323,
+ "learning_rate": 0.0006,
+ "loss": 4.28849983215332,
+ "step": 1749
+ },
+ {
+ "epoch": 24.307423580786025,
+ "grad_norm": 0.018666435033082962,
+ "learning_rate": 0.0006,
+ "loss": 4.368590354919434,
+ "step": 1750
+ },
+ {
+ "epoch": 24.321397379912664,
+ "grad_norm": 0.020626146346330643,
+ "learning_rate": 0.0006,
+ "loss": 4.2951836585998535,
+ "step": 1751
+ },
+ {
+ "epoch": 24.335371179039303,
+ "grad_norm": 0.01929759792983532,
+ "learning_rate": 0.0006,
+ "loss": 4.3664703369140625,
+ "step": 1752
+ },
+ {
+ "epoch": 24.349344978165938,
+ "grad_norm": 0.020956002175807953,
+ "learning_rate": 0.0006,
+ "loss": 4.345308303833008,
+ "step": 1753
+ },
+ {
+ "epoch": 24.363318777292577,
+ "grad_norm": 0.019227957352995872,
+ "learning_rate": 0.0006,
+ "loss": 4.349849224090576,
+ "step": 1754
+ },
+ {
+ "epoch": 24.377292576419215,
+ "grad_norm": 0.015970097854733467,
+ "learning_rate": 0.0006,
+ "loss": 4.223012447357178,
+ "step": 1755
+ },
+ {
+ "epoch": 24.39126637554585,
+ "grad_norm": 0.015951646491885185,
+ "learning_rate": 0.0006,
+ "loss": 4.236420631408691,
+ "step": 1756
+ },
+ {
+ "epoch": 24.40524017467249,
+ "grad_norm": 0.016032515093684196,
+ "learning_rate": 0.0006,
+ "loss": 4.226899147033691,
+ "step": 1757
+ },
+ {
+ "epoch": 24.419213973799128,
+ "grad_norm": 0.01573132909834385,
+ "learning_rate": 0.0006,
+ "loss": 4.315916061401367,
+ "step": 1758
+ },
+ {
+ "epoch": 24.433187772925763,
+ "grad_norm": 0.015367005951702595,
+ "learning_rate": 0.0006,
+ "loss": 4.358939170837402,
+ "step": 1759
+ },
+ {
+ "epoch": 24.4471615720524,
+ "grad_norm": 0.01669907011091709,
+ "learning_rate": 0.0006,
+ "loss": 4.232147216796875,
+ "step": 1760
+ },
+ {
+ "epoch": 24.46113537117904,
+ "grad_norm": 0.018708152696490288,
+ "learning_rate": 0.0006,
+ "loss": 4.329852104187012,
+ "step": 1761
+ },
+ {
+ "epoch": 24.475109170305675,
+ "grad_norm": 0.019035987555980682,
+ "learning_rate": 0.0006,
+ "loss": 4.260340213775635,
+ "step": 1762
+ },
+ {
+ "epoch": 24.489082969432314,
+ "grad_norm": 0.01832144521176815,
+ "learning_rate": 0.0006,
+ "loss": 4.359543800354004,
+ "step": 1763
+ },
+ {
+ "epoch": 24.503056768558952,
+ "grad_norm": 0.015313294716179371,
+ "learning_rate": 0.0006,
+ "loss": 4.279088020324707,
+ "step": 1764
+ },
+ {
+ "epoch": 24.51703056768559,
+ "grad_norm": 0.015803923830389977,
+ "learning_rate": 0.0006,
+ "loss": 4.230682849884033,
+ "step": 1765
+ },
+ {
+ "epoch": 24.531004366812226,
+ "grad_norm": 0.016748374328017235,
+ "learning_rate": 0.0006,
+ "loss": 4.349661827087402,
+ "step": 1766
+ },
+ {
+ "epoch": 24.544978165938865,
+ "grad_norm": 0.015732428058981895,
+ "learning_rate": 0.0006,
+ "loss": 4.372441291809082,
+ "step": 1767
+ },
+ {
+ "epoch": 24.558951965065503,
+ "grad_norm": 0.016564829275012016,
+ "learning_rate": 0.0006,
+ "loss": 4.323090553283691,
+ "step": 1768
+ },
+ {
+ "epoch": 24.57292576419214,
+ "grad_norm": 0.01603010483086109,
+ "learning_rate": 0.0006,
+ "loss": 4.347898006439209,
+ "step": 1769
+ },
+ {
+ "epoch": 24.586899563318777,
+ "grad_norm": 0.01674809865653515,
+ "learning_rate": 0.0006,
+ "loss": 4.269165515899658,
+ "step": 1770
+ },
+ {
+ "epoch": 24.600873362445416,
+ "grad_norm": 0.017333930358290672,
+ "learning_rate": 0.0006,
+ "loss": 4.353298187255859,
+ "step": 1771
+ },
+ {
+ "epoch": 24.61484716157205,
+ "grad_norm": 0.017519580200314522,
+ "learning_rate": 0.0006,
+ "loss": 4.265199661254883,
+ "step": 1772
+ },
+ {
+ "epoch": 24.62882096069869,
+ "grad_norm": 0.017675306648015976,
+ "learning_rate": 0.0006,
+ "loss": 4.322596549987793,
+ "step": 1773
+ },
+ {
+ "epoch": 24.64279475982533,
+ "grad_norm": 0.017166534438729286,
+ "learning_rate": 0.0006,
+ "loss": 4.413900375366211,
+ "step": 1774
+ },
+ {
+ "epoch": 24.656768558951963,
+ "grad_norm": 0.016362346708774567,
+ "learning_rate": 0.0006,
+ "loss": 4.307236671447754,
+ "step": 1775
+ },
+ {
+ "epoch": 24.670742358078602,
+ "grad_norm": 0.016086872667074203,
+ "learning_rate": 0.0006,
+ "loss": 4.235316276550293,
+ "step": 1776
+ },
+ {
+ "epoch": 24.68471615720524,
+ "grad_norm": 0.01689537614583969,
+ "learning_rate": 0.0006,
+ "loss": 4.390591621398926,
+ "step": 1777
+ },
+ {
+ "epoch": 24.69868995633188,
+ "grad_norm": 0.016836725175380707,
+ "learning_rate": 0.0006,
+ "loss": 4.2990031242370605,
+ "step": 1778
+ },
+ {
+ "epoch": 24.712663755458514,
+ "grad_norm": 0.015627335757017136,
+ "learning_rate": 0.0006,
+ "loss": 4.166745185852051,
+ "step": 1779
+ },
+ {
+ "epoch": 24.726637554585153,
+ "grad_norm": 0.015007016249001026,
+ "learning_rate": 0.0006,
+ "loss": 4.295273303985596,
+ "step": 1780
+ },
+ {
+ "epoch": 24.74061135371179,
+ "grad_norm": 0.016084209084510803,
+ "learning_rate": 0.0006,
+ "loss": 4.342424392700195,
+ "step": 1781
+ },
+ {
+ "epoch": 24.754585152838427,
+ "grad_norm": 0.01595073565840721,
+ "learning_rate": 0.0006,
+ "loss": 4.3060102462768555,
+ "step": 1782
+ },
+ {
+ "epoch": 24.768558951965066,
+ "grad_norm": 0.016113461926579475,
+ "learning_rate": 0.0006,
+ "loss": 4.267908096313477,
+ "step": 1783
+ },
+ {
+ "epoch": 24.782532751091704,
+ "grad_norm": 0.01564556173980236,
+ "learning_rate": 0.0006,
+ "loss": 4.324642658233643,
+ "step": 1784
+ },
+ {
+ "epoch": 24.79650655021834,
+ "grad_norm": 0.01575268618762493,
+ "learning_rate": 0.0006,
+ "loss": 4.328181743621826,
+ "step": 1785
+ },
+ {
+ "epoch": 24.810480349344978,
+ "grad_norm": 0.016830824315547943,
+ "learning_rate": 0.0006,
+ "loss": 4.247906684875488,
+ "step": 1786
+ },
+ {
+ "epoch": 24.824454148471617,
+ "grad_norm": 0.017989547923207283,
+ "learning_rate": 0.0006,
+ "loss": 4.228281021118164,
+ "step": 1787
+ },
+ {
+ "epoch": 24.83842794759825,
+ "grad_norm": 0.016783086583018303,
+ "learning_rate": 0.0006,
+ "loss": 4.2880964279174805,
+ "step": 1788
+ },
+ {
+ "epoch": 24.85240174672489,
+ "grad_norm": 0.016467561945319176,
+ "learning_rate": 0.0006,
+ "loss": 4.36118745803833,
+ "step": 1789
+ },
+ {
+ "epoch": 24.86637554585153,
+ "grad_norm": 0.01780523732304573,
+ "learning_rate": 0.0006,
+ "loss": 4.172321319580078,
+ "step": 1790
+ },
+ {
+ "epoch": 24.880349344978168,
+ "grad_norm": 0.017102031037211418,
+ "learning_rate": 0.0006,
+ "loss": 4.352826118469238,
+ "step": 1791
+ },
+ {
+ "epoch": 24.894323144104803,
+ "grad_norm": 0.016409218311309814,
+ "learning_rate": 0.0006,
+ "loss": 4.212162017822266,
+ "step": 1792
+ },
+ {
+ "epoch": 24.90829694323144,
+ "grad_norm": 0.018660584464669228,
+ "learning_rate": 0.0006,
+ "loss": 4.346684455871582,
+ "step": 1793
+ },
+ {
+ "epoch": 24.92227074235808,
+ "grad_norm": 0.017389440909028053,
+ "learning_rate": 0.0006,
+ "loss": 4.197913646697998,
+ "step": 1794
+ },
+ {
+ "epoch": 24.936244541484715,
+ "grad_norm": 0.01596485823392868,
+ "learning_rate": 0.0006,
+ "loss": 4.370617866516113,
+ "step": 1795
+ },
+ {
+ "epoch": 24.950218340611354,
+ "grad_norm": 0.01709776371717453,
+ "learning_rate": 0.0006,
+ "loss": 4.313847064971924,
+ "step": 1796
+ },
+ {
+ "epoch": 24.964192139737992,
+ "grad_norm": 0.016571134328842163,
+ "learning_rate": 0.0006,
+ "loss": 4.1895856857299805,
+ "step": 1797
+ },
+ {
+ "epoch": 24.978165938864628,
+ "grad_norm": 0.016851048916578293,
+ "learning_rate": 0.0006,
+ "loss": 4.350772857666016,
+ "step": 1798
+ },
+ {
+ "epoch": 24.992139737991266,
+ "grad_norm": 0.017007440328598022,
+ "learning_rate": 0.0006,
+ "loss": 4.399008750915527,
+ "step": 1799
+ },
+ {
+ "epoch": 25.0,
+ "grad_norm": 0.01791212148964405,
+ "learning_rate": 0.0006,
+ "loss": 4.306121826171875,
+ "step": 1800
+ },
+ {
+ "epoch": 25.0,
+ "eval_loss": 4.636417865753174,
+ "eval_runtime": 56.8296,
+ "eval_samples_per_second": 42.971,
+ "eval_steps_per_second": 1.355,
+ "step": 1800
+ },
+ {
+ "epoch": 25.01397379912664,
+ "grad_norm": 0.01729617640376091,
+ "learning_rate": 0.0006,
+ "loss": 4.257018089294434,
+ "step": 1801
+ },
+ {
+ "epoch": 25.027947598253274,
+ "grad_norm": 0.017466643825173378,
+ "learning_rate": 0.0006,
+ "loss": 4.307605266571045,
+ "step": 1802
+ },
+ {
+ "epoch": 25.041921397379912,
+ "grad_norm": 0.018614279106259346,
+ "learning_rate": 0.0006,
+ "loss": 4.360141277313232,
+ "step": 1803
+ },
+ {
+ "epoch": 25.05589519650655,
+ "grad_norm": 0.02010771445930004,
+ "learning_rate": 0.0006,
+ "loss": 4.196428298950195,
+ "step": 1804
+ },
+ {
+ "epoch": 25.069868995633186,
+ "grad_norm": 0.02090228535234928,
+ "learning_rate": 0.0006,
+ "loss": 4.288947105407715,
+ "step": 1805
+ },
+ {
+ "epoch": 25.083842794759825,
+ "grad_norm": 0.02050255797803402,
+ "learning_rate": 0.0006,
+ "loss": 4.259582042694092,
+ "step": 1806
+ },
+ {
+ "epoch": 25.097816593886463,
+ "grad_norm": 0.021844089031219482,
+ "learning_rate": 0.0006,
+ "loss": 4.405138969421387,
+ "step": 1807
+ },
+ {
+ "epoch": 25.111790393013102,
+ "grad_norm": 0.022287018597126007,
+ "learning_rate": 0.0006,
+ "loss": 4.333793640136719,
+ "step": 1808
+ },
+ {
+ "epoch": 25.125764192139737,
+ "grad_norm": 0.02288329415023327,
+ "learning_rate": 0.0006,
+ "loss": 4.3910980224609375,
+ "step": 1809
+ },
+ {
+ "epoch": 25.139737991266376,
+ "grad_norm": 0.02373199723660946,
+ "learning_rate": 0.0006,
+ "loss": 4.32921028137207,
+ "step": 1810
+ },
+ {
+ "epoch": 25.153711790393015,
+ "grad_norm": 0.024960234761238098,
+ "learning_rate": 0.0006,
+ "loss": 4.295609474182129,
+ "step": 1811
+ },
+ {
+ "epoch": 25.16768558951965,
+ "grad_norm": 0.026743892580270767,
+ "learning_rate": 0.0006,
+ "loss": 4.315308094024658,
+ "step": 1812
+ },
+ {
+ "epoch": 25.18165938864629,
+ "grad_norm": 0.022138185799121857,
+ "learning_rate": 0.0006,
+ "loss": 4.360208511352539,
+ "step": 1813
+ },
+ {
+ "epoch": 25.195633187772927,
+ "grad_norm": 0.022120574489235878,
+ "learning_rate": 0.0006,
+ "loss": 4.1973772048950195,
+ "step": 1814
+ },
+ {
+ "epoch": 25.209606986899562,
+ "grad_norm": 0.02230502851307392,
+ "learning_rate": 0.0006,
+ "loss": 4.285558700561523,
+ "step": 1815
+ },
+ {
+ "epoch": 25.2235807860262,
+ "grad_norm": 0.022301986813545227,
+ "learning_rate": 0.0006,
+ "loss": 4.238343238830566,
+ "step": 1816
+ },
+ {
+ "epoch": 25.23755458515284,
+ "grad_norm": 0.020050447434186935,
+ "learning_rate": 0.0006,
+ "loss": 4.298235893249512,
+ "step": 1817
+ },
+ {
+ "epoch": 25.251528384279474,
+ "grad_norm": 0.01987724006175995,
+ "learning_rate": 0.0006,
+ "loss": 4.301384449005127,
+ "step": 1818
+ },
+ {
+ "epoch": 25.265502183406113,
+ "grad_norm": 0.020440855994820595,
+ "learning_rate": 0.0006,
+ "loss": 4.269218444824219,
+ "step": 1819
+ },
+ {
+ "epoch": 25.27947598253275,
+ "grad_norm": 0.01891893707215786,
+ "learning_rate": 0.0006,
+ "loss": 4.2401957511901855,
+ "step": 1820
+ },
+ {
+ "epoch": 25.29344978165939,
+ "grad_norm": 0.01903688907623291,
+ "learning_rate": 0.0006,
+ "loss": 4.139670372009277,
+ "step": 1821
+ },
+ {
+ "epoch": 25.307423580786025,
+ "grad_norm": 0.018856938928365707,
+ "learning_rate": 0.0006,
+ "loss": 4.233214378356934,
+ "step": 1822
+ },
+ {
+ "epoch": 25.321397379912664,
+ "grad_norm": 0.017791323363780975,
+ "learning_rate": 0.0006,
+ "loss": 4.2342023849487305,
+ "step": 1823
+ },
+ {
+ "epoch": 25.335371179039303,
+ "grad_norm": 0.017304565757513046,
+ "learning_rate": 0.0006,
+ "loss": 4.355198860168457,
+ "step": 1824
+ },
+ {
+ "epoch": 25.349344978165938,
+ "grad_norm": 0.01576206088066101,
+ "learning_rate": 0.0006,
+ "loss": 4.114147186279297,
+ "step": 1825
+ },
+ {
+ "epoch": 25.363318777292577,
+ "grad_norm": 0.015105154365301132,
+ "learning_rate": 0.0006,
+ "loss": 4.313795566558838,
+ "step": 1826
+ },
+ {
+ "epoch": 25.377292576419215,
+ "grad_norm": 0.01563587784767151,
+ "learning_rate": 0.0006,
+ "loss": 4.308282852172852,
+ "step": 1827
+ },
+ {
+ "epoch": 25.39126637554585,
+ "grad_norm": 0.015980906784534454,
+ "learning_rate": 0.0006,
+ "loss": 4.345149040222168,
+ "step": 1828
+ },
+ {
+ "epoch": 25.40524017467249,
+ "grad_norm": 0.01669284701347351,
+ "learning_rate": 0.0006,
+ "loss": 4.237611293792725,
+ "step": 1829
+ },
+ {
+ "epoch": 25.419213973799128,
+ "grad_norm": 0.016202064231038094,
+ "learning_rate": 0.0006,
+ "loss": 4.3072309494018555,
+ "step": 1830
+ },
+ {
+ "epoch": 25.433187772925763,
+ "grad_norm": 0.01478662807494402,
+ "learning_rate": 0.0006,
+ "loss": 4.276215076446533,
+ "step": 1831
+ },
+ {
+ "epoch": 25.4471615720524,
+ "grad_norm": 0.015168731100857258,
+ "learning_rate": 0.0006,
+ "loss": 4.37509822845459,
+ "step": 1832
+ },
+ {
+ "epoch": 25.46113537117904,
+ "grad_norm": 0.01648980937898159,
+ "learning_rate": 0.0006,
+ "loss": 4.349167346954346,
+ "step": 1833
+ },
+ {
+ "epoch": 25.475109170305675,
+ "grad_norm": 0.01708398200571537,
+ "learning_rate": 0.0006,
+ "loss": 4.346580505371094,
+ "step": 1834
+ },
+ {
+ "epoch": 25.489082969432314,
+ "grad_norm": 0.017370784655213356,
+ "learning_rate": 0.0006,
+ "loss": 4.299466133117676,
+ "step": 1835
+ },
+ {
+ "epoch": 25.503056768558952,
+ "grad_norm": 0.016248228028416634,
+ "learning_rate": 0.0006,
+ "loss": 4.184610366821289,
+ "step": 1836
+ },
+ {
+ "epoch": 25.51703056768559,
+ "grad_norm": 0.017106913030147552,
+ "learning_rate": 0.0006,
+ "loss": 4.366856098175049,
+ "step": 1837
+ },
+ {
+ "epoch": 25.531004366812226,
+ "grad_norm": 0.016071965917944908,
+ "learning_rate": 0.0006,
+ "loss": 4.297691345214844,
+ "step": 1838
+ },
+ {
+ "epoch": 25.544978165938865,
+ "grad_norm": 0.016451986506581306,
+ "learning_rate": 0.0006,
+ "loss": 4.229983329772949,
+ "step": 1839
+ },
+ {
+ "epoch": 25.558951965065503,
+ "grad_norm": 0.016195081174373627,
+ "learning_rate": 0.0006,
+ "loss": 4.259893417358398,
+ "step": 1840
+ },
+ {
+ "epoch": 25.57292576419214,
+ "grad_norm": 0.0165175162255764,
+ "learning_rate": 0.0006,
+ "loss": 4.3348236083984375,
+ "step": 1841
+ },
+ {
+ "epoch": 25.586899563318777,
+ "grad_norm": 0.01651517115533352,
+ "learning_rate": 0.0006,
+ "loss": 4.319541931152344,
+ "step": 1842
+ },
+ {
+ "epoch": 25.600873362445416,
+ "grad_norm": 0.016143690794706345,
+ "learning_rate": 0.0006,
+ "loss": 4.345652103424072,
+ "step": 1843
+ },
+ {
+ "epoch": 25.61484716157205,
+ "grad_norm": 0.015140696428716183,
+ "learning_rate": 0.0006,
+ "loss": 4.212047576904297,
+ "step": 1844
+ },
+ {
+ "epoch": 25.62882096069869,
+ "grad_norm": 0.014637443237006664,
+ "learning_rate": 0.0006,
+ "loss": 4.187453269958496,
+ "step": 1845
+ },
+ {
+ "epoch": 25.64279475982533,
+ "grad_norm": 0.014585614204406738,
+ "learning_rate": 0.0006,
+ "loss": 4.253936767578125,
+ "step": 1846
+ },
+ {
+ "epoch": 25.656768558951963,
+ "grad_norm": 0.015158289112150669,
+ "learning_rate": 0.0006,
+ "loss": 4.34612512588501,
+ "step": 1847
+ },
+ {
+ "epoch": 25.670742358078602,
+ "grad_norm": 0.015202849172055721,
+ "learning_rate": 0.0006,
+ "loss": 4.270634651184082,
+ "step": 1848
+ },
+ {
+ "epoch": 25.68471615720524,
+ "grad_norm": 0.015363575890660286,
+ "learning_rate": 0.0006,
+ "loss": 4.271080493927002,
+ "step": 1849
+ },
+ {
+ "epoch": 25.69868995633188,
+ "grad_norm": 0.014360226690769196,
+ "learning_rate": 0.0006,
+ "loss": 4.375033855438232,
+ "step": 1850
+ },
+ {
+ "epoch": 25.712663755458514,
+ "grad_norm": 0.014263181947171688,
+ "learning_rate": 0.0006,
+ "loss": 4.424138069152832,
+ "step": 1851
+ },
+ {
+ "epoch": 25.726637554585153,
+ "grad_norm": 0.014398688450455666,
+ "learning_rate": 0.0006,
+ "loss": 4.410154819488525,
+ "step": 1852
+ },
+ {
+ "epoch": 25.74061135371179,
+ "grad_norm": 0.016131488606333733,
+ "learning_rate": 0.0006,
+ "loss": 4.324014663696289,
+ "step": 1853
+ },
+ {
+ "epoch": 25.754585152838427,
+ "grad_norm": 0.017568735405802727,
+ "learning_rate": 0.0006,
+ "loss": 4.367861747741699,
+ "step": 1854
+ },
+ {
+ "epoch": 25.768558951965066,
+ "grad_norm": 0.017907511442899704,
+ "learning_rate": 0.0006,
+ "loss": 4.352965354919434,
+ "step": 1855
+ },
+ {
+ "epoch": 25.782532751091704,
+ "grad_norm": 0.016918016597628593,
+ "learning_rate": 0.0006,
+ "loss": 4.252397537231445,
+ "step": 1856
+ },
+ {
+ "epoch": 25.79650655021834,
+ "grad_norm": 0.017274610698223114,
+ "learning_rate": 0.0006,
+ "loss": 4.342309951782227,
+ "step": 1857
+ },
+ {
+ "epoch": 25.810480349344978,
+ "grad_norm": 0.017879825085401535,
+ "learning_rate": 0.0006,
+ "loss": 4.330987453460693,
+ "step": 1858
+ },
+ {
+ "epoch": 25.824454148471617,
+ "grad_norm": 0.01736099272966385,
+ "learning_rate": 0.0006,
+ "loss": 4.307255744934082,
+ "step": 1859
+ },
+ {
+ "epoch": 25.83842794759825,
+ "grad_norm": 0.016575824469327927,
+ "learning_rate": 0.0006,
+ "loss": 4.262016773223877,
+ "step": 1860
+ },
+ {
+ "epoch": 25.85240174672489,
+ "grad_norm": 0.015181140042841434,
+ "learning_rate": 0.0006,
+ "loss": 4.291863441467285,
+ "step": 1861
+ },
+ {
+ "epoch": 25.86637554585153,
+ "grad_norm": 0.015965687111020088,
+ "learning_rate": 0.0006,
+ "loss": 4.424836158752441,
+ "step": 1862
+ },
+ {
+ "epoch": 25.880349344978168,
+ "grad_norm": 0.01687219925224781,
+ "learning_rate": 0.0006,
+ "loss": 4.343531131744385,
+ "step": 1863
+ },
+ {
+ "epoch": 25.894323144104803,
+ "grad_norm": 0.015115504153072834,
+ "learning_rate": 0.0006,
+ "loss": 4.357808589935303,
+ "step": 1864
+ },
+ {
+ "epoch": 25.90829694323144,
+ "grad_norm": 0.014150998555123806,
+ "learning_rate": 0.0006,
+ "loss": 4.324550151824951,
+ "step": 1865
+ },
+ {
+ "epoch": 25.92227074235808,
+ "grad_norm": 0.014989510178565979,
+ "learning_rate": 0.0006,
+ "loss": 4.380428314208984,
+ "step": 1866
+ },
+ {
+ "epoch": 25.936244541484715,
+ "grad_norm": 0.01527960691601038,
+ "learning_rate": 0.0006,
+ "loss": 4.399374961853027,
+ "step": 1867
+ },
+ {
+ "epoch": 25.950218340611354,
+ "grad_norm": 0.015434633940458298,
+ "learning_rate": 0.0006,
+ "loss": 4.260984420776367,
+ "step": 1868
+ },
+ {
+ "epoch": 25.964192139737992,
+ "grad_norm": 0.015370228327810764,
+ "learning_rate": 0.0006,
+ "loss": 4.3340253829956055,
+ "step": 1869
+ },
+ {
+ "epoch": 25.978165938864628,
+ "grad_norm": 0.015348542481660843,
+ "learning_rate": 0.0006,
+ "loss": 4.369132995605469,
+ "step": 1870
+ },
+ {
+ "epoch": 25.992139737991266,
+ "grad_norm": 0.015354936942458153,
+ "learning_rate": 0.0006,
+ "loss": 4.381937026977539,
+ "step": 1871
+ },
+ {
+ "epoch": 26.0,
+ "grad_norm": 0.01796272210776806,
+ "learning_rate": 0.0006,
+ "loss": 4.286768436431885,
+ "step": 1872
+ },
+ {
+ "epoch": 26.0,
+ "eval_loss": 4.576984405517578,
+ "eval_runtime": 57.2091,
+ "eval_samples_per_second": 42.685,
+ "eval_steps_per_second": 1.346,
+ "step": 1872
+ },
+ {
+ "epoch": 26.01397379912664,
+ "grad_norm": 0.01748845726251602,
+ "learning_rate": 0.0006,
+ "loss": 4.337504863739014,
+ "step": 1873
+ },
+ {
+ "epoch": 26.027947598253274,
+ "grad_norm": 0.019244296476244926,
+ "learning_rate": 0.0006,
+ "loss": 4.264280796051025,
+ "step": 1874
+ },
+ {
+ "epoch": 26.041921397379912,
+ "grad_norm": 0.0215692650526762,
+ "learning_rate": 0.0006,
+ "loss": 4.350830078125,
+ "step": 1875
+ },
+ {
+ "epoch": 26.05589519650655,
+ "grad_norm": 0.022189252078533173,
+ "learning_rate": 0.0006,
+ "loss": 4.299098968505859,
+ "step": 1876
+ },
+ {
+ "epoch": 26.069868995633186,
+ "grad_norm": 0.022384043782949448,
+ "learning_rate": 0.0006,
+ "loss": 4.1185173988342285,
+ "step": 1877
+ },
+ {
+ "epoch": 26.083842794759825,
+ "grad_norm": 0.02076014317572117,
+ "learning_rate": 0.0006,
+ "loss": 4.297612190246582,
+ "step": 1878
+ },
+ {
+ "epoch": 26.097816593886463,
+ "grad_norm": 0.020843051373958588,
+ "learning_rate": 0.0006,
+ "loss": 4.294098854064941,
+ "step": 1879
+ },
+ {
+ "epoch": 26.111790393013102,
+ "grad_norm": 0.020654456689953804,
+ "learning_rate": 0.0006,
+ "loss": 4.267856597900391,
+ "step": 1880
+ },
+ {
+ "epoch": 26.125764192139737,
+ "grad_norm": 0.01860472559928894,
+ "learning_rate": 0.0006,
+ "loss": 4.284232139587402,
+ "step": 1881
+ },
+ {
+ "epoch": 26.139737991266376,
+ "grad_norm": 0.018719421699643135,
+ "learning_rate": 0.0006,
+ "loss": 4.277214527130127,
+ "step": 1882
+ },
+ {
+ "epoch": 26.153711790393015,
+ "grad_norm": 0.020903829485177994,
+ "learning_rate": 0.0006,
+ "loss": 4.254947662353516,
+ "step": 1883
+ },
+ {
+ "epoch": 26.16768558951965,
+ "grad_norm": 0.02020149491727352,
+ "learning_rate": 0.0006,
+ "loss": 4.2526936531066895,
+ "step": 1884
+ },
+ {
+ "epoch": 26.18165938864629,
+ "grad_norm": 0.0180392786860466,
+ "learning_rate": 0.0006,
+ "loss": 4.309922695159912,
+ "step": 1885
+ },
+ {
+ "epoch": 26.195633187772927,
+ "grad_norm": 0.018163125962018967,
+ "learning_rate": 0.0006,
+ "loss": 4.13860559463501,
+ "step": 1886
+ },
+ {
+ "epoch": 26.209606986899562,
+ "grad_norm": 0.01853291131556034,
+ "learning_rate": 0.0006,
+ "loss": 4.190926551818848,
+ "step": 1887
+ },
+ {
+ "epoch": 26.2235807860262,
+ "grad_norm": 0.01757677271962166,
+ "learning_rate": 0.0006,
+ "loss": 4.163999557495117,
+ "step": 1888
+ },
+ {
+ "epoch": 26.23755458515284,
+ "grad_norm": 0.01767992228269577,
+ "learning_rate": 0.0006,
+ "loss": 4.184430122375488,
+ "step": 1889
+ },
+ {
+ "epoch": 26.251528384279474,
+ "grad_norm": 0.018139973282814026,
+ "learning_rate": 0.0006,
+ "loss": 4.300766944885254,
+ "step": 1890
+ },
+ {
+ "epoch": 26.265502183406113,
+ "grad_norm": 0.020184461027383804,
+ "learning_rate": 0.0006,
+ "loss": 4.232537269592285,
+ "step": 1891
+ },
+ {
+ "epoch": 26.27947598253275,
+ "grad_norm": 0.02028856985270977,
+ "learning_rate": 0.0006,
+ "loss": 4.324883460998535,
+ "step": 1892
+ },
+ {
+ "epoch": 26.29344978165939,
+ "grad_norm": 0.017120616510510445,
+ "learning_rate": 0.0006,
+ "loss": 4.23493766784668,
+ "step": 1893
+ },
+ {
+ "epoch": 26.307423580786025,
+ "grad_norm": 0.016202978789806366,
+ "learning_rate": 0.0006,
+ "loss": 4.197602272033691,
+ "step": 1894
+ },
+ {
+ "epoch": 26.321397379912664,
+ "grad_norm": 0.017231421545147896,
+ "learning_rate": 0.0006,
+ "loss": 4.302001953125,
+ "step": 1895
+ },
+ {
+ "epoch": 26.335371179039303,
+ "grad_norm": 0.01725655607879162,
+ "learning_rate": 0.0006,
+ "loss": 4.356800079345703,
+ "step": 1896
+ },
+ {
+ "epoch": 26.349344978165938,
+ "grad_norm": 0.016657443717122078,
+ "learning_rate": 0.0006,
+ "loss": 4.266307830810547,
+ "step": 1897
+ },
+ {
+ "epoch": 26.363318777292577,
+ "grad_norm": 0.01616556942462921,
+ "learning_rate": 0.0006,
+ "loss": 4.307868003845215,
+ "step": 1898
+ },
+ {
+ "epoch": 26.377292576419215,
+ "grad_norm": 0.016621306538581848,
+ "learning_rate": 0.0006,
+ "loss": 4.172905921936035,
+ "step": 1899
+ },
+ {
+ "epoch": 26.39126637554585,
+ "grad_norm": 0.016243206337094307,
+ "learning_rate": 0.0006,
+ "loss": 4.162896156311035,
+ "step": 1900
+ },
+ {
+ "epoch": 26.40524017467249,
+ "grad_norm": 0.016663808375597,
+ "learning_rate": 0.0006,
+ "loss": 4.258232116699219,
+ "step": 1901
+ },
+ {
+ "epoch": 26.419213973799128,
+ "grad_norm": 0.01677478663623333,
+ "learning_rate": 0.0006,
+ "loss": 4.336090087890625,
+ "step": 1902
+ },
+ {
+ "epoch": 26.433187772925763,
+ "grad_norm": 0.01641303487122059,
+ "learning_rate": 0.0006,
+ "loss": 4.191476821899414,
+ "step": 1903
+ },
+ {
+ "epoch": 26.4471615720524,
+ "grad_norm": 0.016700323671102524,
+ "learning_rate": 0.0006,
+ "loss": 4.299427509307861,
+ "step": 1904
+ },
+ {
+ "epoch": 26.46113537117904,
+ "grad_norm": 0.017595035955309868,
+ "learning_rate": 0.0006,
+ "loss": 4.312828063964844,
+ "step": 1905
+ },
+ {
+ "epoch": 26.475109170305675,
+ "grad_norm": 0.020300284028053284,
+ "learning_rate": 0.0006,
+ "loss": 4.28331184387207,
+ "step": 1906
+ },
+ {
+ "epoch": 26.489082969432314,
+ "grad_norm": 0.02198321744799614,
+ "learning_rate": 0.0006,
+ "loss": 4.3280839920043945,
+ "step": 1907
+ },
+ {
+ "epoch": 26.503056768558952,
+ "grad_norm": 0.02003200724720955,
+ "learning_rate": 0.0006,
+ "loss": 4.325623989105225,
+ "step": 1908
+ },
+ {
+ "epoch": 26.51703056768559,
+ "grad_norm": 0.01870287023484707,
+ "learning_rate": 0.0006,
+ "loss": 4.251203536987305,
+ "step": 1909
+ },
+ {
+ "epoch": 26.531004366812226,
+ "grad_norm": 0.01835448481142521,
+ "learning_rate": 0.0006,
+ "loss": 4.335144996643066,
+ "step": 1910
+ },
+ {
+ "epoch": 26.544978165938865,
+ "grad_norm": 0.02046327292919159,
+ "learning_rate": 0.0006,
+ "loss": 4.238023281097412,
+ "step": 1911
+ },
+ {
+ "epoch": 26.558951965065503,
+ "grad_norm": 0.019771311432123184,
+ "learning_rate": 0.0006,
+ "loss": 4.353594779968262,
+ "step": 1912
+ },
+ {
+ "epoch": 26.57292576419214,
+ "grad_norm": 0.018090544268488884,
+ "learning_rate": 0.0006,
+ "loss": 4.304343223571777,
+ "step": 1913
+ },
+ {
+ "epoch": 26.586899563318777,
+ "grad_norm": 0.017489777877926826,
+ "learning_rate": 0.0006,
+ "loss": 4.204375267028809,
+ "step": 1914
+ },
+ {
+ "epoch": 26.600873362445416,
+ "grad_norm": 0.016785891726613045,
+ "learning_rate": 0.0006,
+ "loss": 4.286669731140137,
+ "step": 1915
+ },
+ {
+ "epoch": 26.61484716157205,
+ "grad_norm": 0.016382789239287376,
+ "learning_rate": 0.0006,
+ "loss": 4.347662448883057,
+ "step": 1916
+ },
+ {
+ "epoch": 26.62882096069869,
+ "grad_norm": 0.016581876203417778,
+ "learning_rate": 0.0006,
+ "loss": 4.2716193199157715,
+ "step": 1917
+ },
+ {
+ "epoch": 26.64279475982533,
+ "grad_norm": 0.015627184882760048,
+ "learning_rate": 0.0006,
+ "loss": 4.311489105224609,
+ "step": 1918
+ },
+ {
+ "epoch": 26.656768558951963,
+ "grad_norm": 0.015908483415842056,
+ "learning_rate": 0.0006,
+ "loss": 4.262099742889404,
+ "step": 1919
+ },
+ {
+ "epoch": 26.670742358078602,
+ "grad_norm": 0.015576236881315708,
+ "learning_rate": 0.0006,
+ "loss": 4.275871276855469,
+ "step": 1920
+ },
+ {
+ "epoch": 26.68471615720524,
+ "grad_norm": 0.014621064998209476,
+ "learning_rate": 0.0006,
+ "loss": 4.184260845184326,
+ "step": 1921
+ },
+ {
+ "epoch": 26.69868995633188,
+ "grad_norm": 0.014797811396420002,
+ "learning_rate": 0.0006,
+ "loss": 4.3412628173828125,
+ "step": 1922
+ },
+ {
+ "epoch": 26.712663755458514,
+ "grad_norm": 0.015610570088028908,
+ "learning_rate": 0.0006,
+ "loss": 4.181094646453857,
+ "step": 1923
+ },
+ {
+ "epoch": 26.726637554585153,
+ "grad_norm": 0.016048265621066093,
+ "learning_rate": 0.0006,
+ "loss": 4.365290641784668,
+ "step": 1924
+ },
+ {
+ "epoch": 26.74061135371179,
+ "grad_norm": 0.016604071483016014,
+ "learning_rate": 0.0006,
+ "loss": 4.2628912925720215,
+ "step": 1925
+ },
+ {
+ "epoch": 26.754585152838427,
+ "grad_norm": 0.017044221982359886,
+ "learning_rate": 0.0006,
+ "loss": 4.378512382507324,
+ "step": 1926
+ },
+ {
+ "epoch": 26.768558951965066,
+ "grad_norm": 0.01746384985744953,
+ "learning_rate": 0.0006,
+ "loss": 4.3888773918151855,
+ "step": 1927
+ },
+ {
+ "epoch": 26.782532751091704,
+ "grad_norm": 0.01572711206972599,
+ "learning_rate": 0.0006,
+ "loss": 4.189935684204102,
+ "step": 1928
+ },
+ {
+ "epoch": 26.79650655021834,
+ "grad_norm": 0.014976629987359047,
+ "learning_rate": 0.0006,
+ "loss": 4.280525207519531,
+ "step": 1929
+ },
+ {
+ "epoch": 26.810480349344978,
+ "grad_norm": 0.014804664999246597,
+ "learning_rate": 0.0006,
+ "loss": 4.270630836486816,
+ "step": 1930
+ },
+ {
+ "epoch": 26.824454148471617,
+ "grad_norm": 0.014641920104622841,
+ "learning_rate": 0.0006,
+ "loss": 4.319505214691162,
+ "step": 1931
+ },
+ {
+ "epoch": 26.83842794759825,
+ "grad_norm": 0.014928505755960941,
+ "learning_rate": 0.0006,
+ "loss": 4.180147647857666,
+ "step": 1932
+ },
+ {
+ "epoch": 26.85240174672489,
+ "grad_norm": 0.016130365431308746,
+ "learning_rate": 0.0006,
+ "loss": 4.260660648345947,
+ "step": 1933
+ },
+ {
+ "epoch": 26.86637554585153,
+ "grad_norm": 0.015642493963241577,
+ "learning_rate": 0.0006,
+ "loss": 4.304064750671387,
+ "step": 1934
+ },
+ {
+ "epoch": 26.880349344978168,
+ "grad_norm": 0.015988217666745186,
+ "learning_rate": 0.0006,
+ "loss": 4.2548017501831055,
+ "step": 1935
+ },
+ {
+ "epoch": 26.894323144104803,
+ "grad_norm": 0.016501398757100105,
+ "learning_rate": 0.0006,
+ "loss": 4.260058879852295,
+ "step": 1936
+ },
+ {
+ "epoch": 26.90829694323144,
+ "grad_norm": 0.01645040698349476,
+ "learning_rate": 0.0006,
+ "loss": 4.483081340789795,
+ "step": 1937
+ },
+ {
+ "epoch": 26.92227074235808,
+ "grad_norm": 0.016140863299369812,
+ "learning_rate": 0.0006,
+ "loss": 4.1776838302612305,
+ "step": 1938
+ },
+ {
+ "epoch": 26.936244541484715,
+ "grad_norm": 0.016681019216775894,
+ "learning_rate": 0.0006,
+ "loss": 4.224747657775879,
+ "step": 1939
+ },
+ {
+ "epoch": 26.950218340611354,
+ "grad_norm": 0.01650378853082657,
+ "learning_rate": 0.0006,
+ "loss": 4.305703163146973,
+ "step": 1940
+ },
+ {
+ "epoch": 26.964192139737992,
+ "grad_norm": 0.017051683738827705,
+ "learning_rate": 0.0006,
+ "loss": 4.285835266113281,
+ "step": 1941
+ },
+ {
+ "epoch": 26.978165938864628,
+ "grad_norm": 0.016894327476620674,
+ "learning_rate": 0.0006,
+ "loss": 4.318901538848877,
+ "step": 1942
+ },
+ {
+ "epoch": 26.992139737991266,
+ "grad_norm": 0.017313100397586823,
+ "learning_rate": 0.0006,
+ "loss": 4.281434535980225,
+ "step": 1943
+ },
+ {
+ "epoch": 27.0,
+ "grad_norm": 0.018450884148478508,
+ "learning_rate": 0.0006,
+ "loss": 4.335104942321777,
+ "step": 1944
+ },
+ {
+ "epoch": 27.0,
+ "eval_loss": 4.640687465667725,
+ "eval_runtime": 56.9624,
+ "eval_samples_per_second": 42.87,
+ "eval_steps_per_second": 1.352,
+ "step": 1944
+ },
+ {
+ "epoch": 27.01397379912664,
+ "grad_norm": 0.017755698412656784,
+ "learning_rate": 0.0006,
+ "loss": 4.278824329376221,
+ "step": 1945
+ },
+ {
+ "epoch": 27.027947598253274,
+ "grad_norm": 0.01724310778081417,
+ "learning_rate": 0.0006,
+ "loss": 4.311049461364746,
+ "step": 1946
+ },
+ {
+ "epoch": 27.041921397379912,
+ "grad_norm": 0.01690123789012432,
+ "learning_rate": 0.0006,
+ "loss": 4.261994361877441,
+ "step": 1947
+ },
+ {
+ "epoch": 27.05589519650655,
+ "grad_norm": 0.018406696617603302,
+ "learning_rate": 0.0006,
+ "loss": 4.1925153732299805,
+ "step": 1948
+ },
+ {
+ "epoch": 27.069868995633186,
+ "grad_norm": 0.019945867359638214,
+ "learning_rate": 0.0006,
+ "loss": 4.108537197113037,
+ "step": 1949
+ },
+ {
+ "epoch": 27.083842794759825,
+ "grad_norm": 0.020316628739237785,
+ "learning_rate": 0.0006,
+ "loss": 4.209043025970459,
+ "step": 1950
+ },
+ {
+ "epoch": 27.097816593886463,
+ "grad_norm": 0.019690370187163353,
+ "learning_rate": 0.0006,
+ "loss": 4.195789337158203,
+ "step": 1951
+ },
+ {
+ "epoch": 27.111790393013102,
+ "grad_norm": 0.01785232499241829,
+ "learning_rate": 0.0006,
+ "loss": 4.21466064453125,
+ "step": 1952
+ },
+ {
+ "epoch": 27.125764192139737,
+ "grad_norm": 0.0164767038077116,
+ "learning_rate": 0.0006,
+ "loss": 4.270247459411621,
+ "step": 1953
+ },
+ {
+ "epoch": 27.139737991266376,
+ "grad_norm": 0.018008455634117126,
+ "learning_rate": 0.0006,
+ "loss": 4.201042175292969,
+ "step": 1954
+ },
+ {
+ "epoch": 27.153711790393015,
+ "grad_norm": 0.01807340793311596,
+ "learning_rate": 0.0006,
+ "loss": 4.25289249420166,
+ "step": 1955
+ },
+ {
+ "epoch": 27.16768558951965,
+ "grad_norm": 0.016246909275650978,
+ "learning_rate": 0.0006,
+ "loss": 4.116283416748047,
+ "step": 1956
+ },
+ {
+ "epoch": 27.18165938864629,
+ "grad_norm": 0.017545776441693306,
+ "learning_rate": 0.0006,
+ "loss": 4.316999435424805,
+ "step": 1957
+ },
+ {
+ "epoch": 27.195633187772927,
+ "grad_norm": 0.017915375530719757,
+ "learning_rate": 0.0006,
+ "loss": 4.253859519958496,
+ "step": 1958
+ },
+ {
+ "epoch": 27.209606986899562,
+ "grad_norm": 0.018112223595380783,
+ "learning_rate": 0.0006,
+ "loss": 4.2541022300720215,
+ "step": 1959
+ },
+ {
+ "epoch": 27.2235807860262,
+ "grad_norm": 0.01846139505505562,
+ "learning_rate": 0.0006,
+ "loss": 4.20438289642334,
+ "step": 1960
+ },
+ {
+ "epoch": 27.23755458515284,
+ "grad_norm": 0.01943897269666195,
+ "learning_rate": 0.0006,
+ "loss": 4.2799577713012695,
+ "step": 1961
+ },
+ {
+ "epoch": 27.251528384279474,
+ "grad_norm": 0.018253348767757416,
+ "learning_rate": 0.0006,
+ "loss": 4.1720499992370605,
+ "step": 1962
+ },
+ {
+ "epoch": 27.265502183406113,
+ "grad_norm": 0.018292246386408806,
+ "learning_rate": 0.0006,
+ "loss": 4.28231143951416,
+ "step": 1963
+ },
+ {
+ "epoch": 27.27947598253275,
+ "grad_norm": 0.01943155936896801,
+ "learning_rate": 0.0006,
+ "loss": 4.314976692199707,
+ "step": 1964
+ },
+ {
+ "epoch": 27.29344978165939,
+ "grad_norm": 0.019064994528889656,
+ "learning_rate": 0.0006,
+ "loss": 4.25328254699707,
+ "step": 1965
+ },
+ {
+ "epoch": 27.307423580786025,
+ "grad_norm": 0.017423273995518684,
+ "learning_rate": 0.0006,
+ "loss": 4.195992469787598,
+ "step": 1966
+ },
+ {
+ "epoch": 27.321397379912664,
+ "grad_norm": 0.01677924580872059,
+ "learning_rate": 0.0006,
+ "loss": 4.260043144226074,
+ "step": 1967
+ },
+ {
+ "epoch": 27.335371179039303,
+ "grad_norm": 0.01720673404633999,
+ "learning_rate": 0.0006,
+ "loss": 4.222780704498291,
+ "step": 1968
+ },
+ {
+ "epoch": 27.349344978165938,
+ "grad_norm": 0.017300743609666824,
+ "learning_rate": 0.0006,
+ "loss": 4.109524250030518,
+ "step": 1969
+ },
+ {
+ "epoch": 27.363318777292577,
+ "grad_norm": 0.016065871343016624,
+ "learning_rate": 0.0006,
+ "loss": 4.36108922958374,
+ "step": 1970
+ },
+ {
+ "epoch": 27.377292576419215,
+ "grad_norm": 0.016626616939902306,
+ "learning_rate": 0.0006,
+ "loss": 4.220630645751953,
+ "step": 1971
+ },
+ {
+ "epoch": 27.39126637554585,
+ "grad_norm": 0.016482602804899216,
+ "learning_rate": 0.0006,
+ "loss": 4.344797134399414,
+ "step": 1972
+ },
+ {
+ "epoch": 27.40524017467249,
+ "grad_norm": 0.01727953553199768,
+ "learning_rate": 0.0006,
+ "loss": 4.296645164489746,
+ "step": 1973
+ },
+ {
+ "epoch": 27.419213973799128,
+ "grad_norm": 0.01522731315344572,
+ "learning_rate": 0.0006,
+ "loss": 4.223751068115234,
+ "step": 1974
+ },
+ {
+ "epoch": 27.433187772925763,
+ "grad_norm": 0.01581776700913906,
+ "learning_rate": 0.0006,
+ "loss": 4.351379871368408,
+ "step": 1975
+ },
+ {
+ "epoch": 27.4471615720524,
+ "grad_norm": 0.016394605860114098,
+ "learning_rate": 0.0006,
+ "loss": 4.2412919998168945,
+ "step": 1976
+ },
+ {
+ "epoch": 27.46113537117904,
+ "grad_norm": 0.01751169003546238,
+ "learning_rate": 0.0006,
+ "loss": 4.257606506347656,
+ "step": 1977
+ },
+ {
+ "epoch": 27.475109170305675,
+ "grad_norm": 0.017452310770750046,
+ "learning_rate": 0.0006,
+ "loss": 4.380134582519531,
+ "step": 1978
+ },
+ {
+ "epoch": 27.489082969432314,
+ "grad_norm": 0.017741898074746132,
+ "learning_rate": 0.0006,
+ "loss": 4.279632568359375,
+ "step": 1979
+ },
+ {
+ "epoch": 27.503056768558952,
+ "grad_norm": 0.017680590972304344,
+ "learning_rate": 0.0006,
+ "loss": 4.273406028747559,
+ "step": 1980
+ },
+ {
+ "epoch": 27.51703056768559,
+ "grad_norm": 0.01864718459546566,
+ "learning_rate": 0.0006,
+ "loss": 4.278583526611328,
+ "step": 1981
+ },
+ {
+ "epoch": 27.531004366812226,
+ "grad_norm": 0.019558541476726532,
+ "learning_rate": 0.0006,
+ "loss": 4.211919784545898,
+ "step": 1982
+ },
+ {
+ "epoch": 27.544978165938865,
+ "grad_norm": 0.019467400386929512,
+ "learning_rate": 0.0006,
+ "loss": 4.276822566986084,
+ "step": 1983
+ },
+ {
+ "epoch": 27.558951965065503,
+ "grad_norm": 0.01756172813475132,
+ "learning_rate": 0.0006,
+ "loss": 4.165379047393799,
+ "step": 1984
+ },
+ {
+ "epoch": 27.57292576419214,
+ "grad_norm": 0.017458263784646988,
+ "learning_rate": 0.0006,
+ "loss": 4.2992401123046875,
+ "step": 1985
+ },
+ {
+ "epoch": 27.586899563318777,
+ "grad_norm": 0.018089665099978447,
+ "learning_rate": 0.0006,
+ "loss": 4.243378639221191,
+ "step": 1986
+ },
+ {
+ "epoch": 27.600873362445416,
+ "grad_norm": 0.019535524770617485,
+ "learning_rate": 0.0006,
+ "loss": 4.321477890014648,
+ "step": 1987
+ },
+ {
+ "epoch": 27.61484716157205,
+ "grad_norm": 0.01997970975935459,
+ "learning_rate": 0.0006,
+ "loss": 4.123082637786865,
+ "step": 1988
+ },
+ {
+ "epoch": 27.62882096069869,
+ "grad_norm": 0.01913067139685154,
+ "learning_rate": 0.0006,
+ "loss": 4.289896488189697,
+ "step": 1989
+ },
+ {
+ "epoch": 27.64279475982533,
+ "grad_norm": 0.0169806070625782,
+ "learning_rate": 0.0006,
+ "loss": 4.247004985809326,
+ "step": 1990
+ },
+ {
+ "epoch": 27.656768558951963,
+ "grad_norm": 0.017315007746219635,
+ "learning_rate": 0.0006,
+ "loss": 4.2095441818237305,
+ "step": 1991
+ },
+ {
+ "epoch": 27.670742358078602,
+ "grad_norm": 0.015098759904503822,
+ "learning_rate": 0.0006,
+ "loss": 4.145560264587402,
+ "step": 1992
+ },
+ {
+ "epoch": 27.68471615720524,
+ "grad_norm": 0.016240514814853668,
+ "learning_rate": 0.0006,
+ "loss": 4.254927635192871,
+ "step": 1993
+ },
+ {
+ "epoch": 27.69868995633188,
+ "grad_norm": 0.01621108129620552,
+ "learning_rate": 0.0006,
+ "loss": 4.219725608825684,
+ "step": 1994
+ },
+ {
+ "epoch": 27.712663755458514,
+ "grad_norm": 0.015191009268164635,
+ "learning_rate": 0.0006,
+ "loss": 4.194345474243164,
+ "step": 1995
+ },
+ {
+ "epoch": 27.726637554585153,
+ "grad_norm": 0.014387983828783035,
+ "learning_rate": 0.0006,
+ "loss": 4.211467742919922,
+ "step": 1996
+ },
+ {
+ "epoch": 27.74061135371179,
+ "grad_norm": 0.015460561029613018,
+ "learning_rate": 0.0006,
+ "loss": 4.358700752258301,
+ "step": 1997
+ },
+ {
+ "epoch": 27.754585152838427,
+ "grad_norm": 0.015100076794624329,
+ "learning_rate": 0.0006,
+ "loss": 4.171223163604736,
+ "step": 1998
+ },
+ {
+ "epoch": 27.768558951965066,
+ "grad_norm": 0.015607891604304314,
+ "learning_rate": 0.0006,
+ "loss": 4.217874526977539,
+ "step": 1999
+ },
+ {
+ "epoch": 27.782532751091704,
+ "grad_norm": 0.015650689601898193,
+ "learning_rate": 0.0006,
+ "loss": 4.280780792236328,
+ "step": 2000
+ },
+ {
+ "epoch": 27.79650655021834,
+ "grad_norm": 0.01580994389951229,
+ "learning_rate": 0.0006,
+ "loss": 4.159212112426758,
+ "step": 2001
+ },
+ {
+ "epoch": 27.810480349344978,
+ "grad_norm": 0.016607413068413734,
+ "learning_rate": 0.0006,
+ "loss": 4.238853454589844,
+ "step": 2002
+ },
+ {
+ "epoch": 27.824454148471617,
+ "grad_norm": 0.017383860424160957,
+ "learning_rate": 0.0006,
+ "loss": 4.233944892883301,
+ "step": 2003
+ },
+ {
+ "epoch": 27.83842794759825,
+ "grad_norm": 0.014966659247875214,
+ "learning_rate": 0.0006,
+ "loss": 4.167392730712891,
+ "step": 2004
+ },
+ {
+ "epoch": 27.85240174672489,
+ "grad_norm": 0.014589878730475903,
+ "learning_rate": 0.0006,
+ "loss": 4.2390899658203125,
+ "step": 2005
+ },
+ {
+ "epoch": 27.86637554585153,
+ "grad_norm": 0.017179450020194054,
+ "learning_rate": 0.0006,
+ "loss": 4.365785598754883,
+ "step": 2006
+ },
+ {
+ "epoch": 27.880349344978168,
+ "grad_norm": 0.01893901452422142,
+ "learning_rate": 0.0006,
+ "loss": 4.289430141448975,
+ "step": 2007
+ },
+ {
+ "epoch": 27.894323144104803,
+ "grad_norm": 0.018619712442159653,
+ "learning_rate": 0.0006,
+ "loss": 4.310665130615234,
+ "step": 2008
+ },
+ {
+ "epoch": 27.90829694323144,
+ "grad_norm": 0.019180385395884514,
+ "learning_rate": 0.0006,
+ "loss": 4.27454137802124,
+ "step": 2009
+ },
+ {
+ "epoch": 27.92227074235808,
+ "grad_norm": 0.017811523750424385,
+ "learning_rate": 0.0006,
+ "loss": 4.312562465667725,
+ "step": 2010
+ },
+ {
+ "epoch": 27.936244541484715,
+ "grad_norm": 0.016583573073148727,
+ "learning_rate": 0.0006,
+ "loss": 4.202253341674805,
+ "step": 2011
+ },
+ {
+ "epoch": 27.950218340611354,
+ "grad_norm": 0.01658615842461586,
+ "learning_rate": 0.0006,
+ "loss": 4.183433532714844,
+ "step": 2012
+ },
+ {
+ "epoch": 27.964192139737992,
+ "grad_norm": 0.015327401459217072,
+ "learning_rate": 0.0006,
+ "loss": 4.2125139236450195,
+ "step": 2013
+ },
+ {
+ "epoch": 27.978165938864628,
+ "grad_norm": 0.015004601329565048,
+ "learning_rate": 0.0006,
+ "loss": 4.257392883300781,
+ "step": 2014
+ },
+ {
+ "epoch": 27.992139737991266,
+ "grad_norm": 0.015259161591529846,
+ "learning_rate": 0.0006,
+ "loss": 4.241818428039551,
+ "step": 2015
+ },
+ {
+ "epoch": 28.0,
+ "grad_norm": 0.016808776184916496,
+ "learning_rate": 0.0006,
+ "loss": 4.146241188049316,
+ "step": 2016
+ },
+ {
+ "epoch": 28.0,
+ "eval_loss": 4.663090705871582,
+ "eval_runtime": 56.4559,
+ "eval_samples_per_second": 43.255,
+ "eval_steps_per_second": 1.364,
+ "step": 2016
+ },
+ {
+ "epoch": 28.01397379912664,
+ "grad_norm": 0.015784382820129395,
+ "learning_rate": 0.0006,
+ "loss": 4.044191360473633,
+ "step": 2017
+ },
+ {
+ "epoch": 28.027947598253274,
+ "grad_norm": 0.015636710450053215,
+ "learning_rate": 0.0006,
+ "loss": 4.128754615783691,
+ "step": 2018
+ },
+ {
+ "epoch": 28.041921397379912,
+ "grad_norm": 0.015364975668489933,
+ "learning_rate": 0.0006,
+ "loss": 4.191835403442383,
+ "step": 2019
+ },
+ {
+ "epoch": 28.05589519650655,
+ "grad_norm": 0.015015700832009315,
+ "learning_rate": 0.0006,
+ "loss": 4.120022773742676,
+ "step": 2020
+ },
+ {
+ "epoch": 28.069868995633186,
+ "grad_norm": 0.016666799783706665,
+ "learning_rate": 0.0006,
+ "loss": 4.33906364440918,
+ "step": 2021
+ },
+ {
+ "epoch": 28.083842794759825,
+ "grad_norm": 0.01780104823410511,
+ "learning_rate": 0.0006,
+ "loss": 4.279458999633789,
+ "step": 2022
+ },
+ {
+ "epoch": 28.097816593886463,
+ "grad_norm": 0.01870882883667946,
+ "learning_rate": 0.0006,
+ "loss": 4.34728479385376,
+ "step": 2023
+ },
+ {
+ "epoch": 28.111790393013102,
+ "grad_norm": 0.018406381830573082,
+ "learning_rate": 0.0006,
+ "loss": 4.211311340332031,
+ "step": 2024
+ },
+ {
+ "epoch": 28.125764192139737,
+ "grad_norm": 0.017833448946475983,
+ "learning_rate": 0.0006,
+ "loss": 4.223254203796387,
+ "step": 2025
+ },
+ {
+ "epoch": 28.139737991266376,
+ "grad_norm": 0.017552688717842102,
+ "learning_rate": 0.0006,
+ "loss": 4.208823204040527,
+ "step": 2026
+ },
+ {
+ "epoch": 28.153711790393015,
+ "grad_norm": 0.018500229343771935,
+ "learning_rate": 0.0006,
+ "loss": 4.198025226593018,
+ "step": 2027
+ },
+ {
+ "epoch": 28.16768558951965,
+ "grad_norm": 0.01782156340777874,
+ "learning_rate": 0.0006,
+ "loss": 4.179978370666504,
+ "step": 2028
+ },
+ {
+ "epoch": 28.18165938864629,
+ "grad_norm": 0.019580967724323273,
+ "learning_rate": 0.0006,
+ "loss": 4.291114807128906,
+ "step": 2029
+ },
+ {
+ "epoch": 28.195633187772927,
+ "grad_norm": 0.020804036408662796,
+ "learning_rate": 0.0006,
+ "loss": 4.220607280731201,
+ "step": 2030
+ },
+ {
+ "epoch": 28.209606986899562,
+ "grad_norm": 0.019595693796873093,
+ "learning_rate": 0.0006,
+ "loss": 4.2905731201171875,
+ "step": 2031
+ },
+ {
+ "epoch": 28.2235807860262,
+ "grad_norm": 0.019247086718678474,
+ "learning_rate": 0.0006,
+ "loss": 4.2467360496521,
+ "step": 2032
+ },
+ {
+ "epoch": 28.23755458515284,
+ "grad_norm": 0.01958036608994007,
+ "learning_rate": 0.0006,
+ "loss": 4.286137580871582,
+ "step": 2033
+ },
+ {
+ "epoch": 28.251528384279474,
+ "grad_norm": 0.021354753524065018,
+ "learning_rate": 0.0006,
+ "loss": 4.234755039215088,
+ "step": 2034
+ },
+ {
+ "epoch": 28.265502183406113,
+ "grad_norm": 0.02206375077366829,
+ "learning_rate": 0.0006,
+ "loss": 4.149641036987305,
+ "step": 2035
+ },
+ {
+ "epoch": 28.27947598253275,
+ "grad_norm": 0.02201310358941555,
+ "learning_rate": 0.0006,
+ "loss": 4.1928629875183105,
+ "step": 2036
+ },
+ {
+ "epoch": 28.29344978165939,
+ "grad_norm": 0.01961950585246086,
+ "learning_rate": 0.0006,
+ "loss": 4.211404323577881,
+ "step": 2037
+ },
+ {
+ "epoch": 28.307423580786025,
+ "grad_norm": 0.019211795181035995,
+ "learning_rate": 0.0006,
+ "loss": 4.200976848602295,
+ "step": 2038
+ },
+ {
+ "epoch": 28.321397379912664,
+ "grad_norm": 0.01779370754957199,
+ "learning_rate": 0.0006,
+ "loss": 4.249148845672607,
+ "step": 2039
+ },
+ {
+ "epoch": 28.335371179039303,
+ "grad_norm": 0.018635908141732216,
+ "learning_rate": 0.0006,
+ "loss": 4.11806058883667,
+ "step": 2040
+ },
+ {
+ "epoch": 28.349344978165938,
+ "grad_norm": 0.019792694598436356,
+ "learning_rate": 0.0006,
+ "loss": 4.335224151611328,
+ "step": 2041
+ },
+ {
+ "epoch": 28.363318777292577,
+ "grad_norm": 0.021731717512011528,
+ "learning_rate": 0.0006,
+ "loss": 4.302778244018555,
+ "step": 2042
+ },
+ {
+ "epoch": 28.377292576419215,
+ "grad_norm": 0.020823447033762932,
+ "learning_rate": 0.0006,
+ "loss": 4.260610580444336,
+ "step": 2043
+ },
+ {
+ "epoch": 28.39126637554585,
+ "grad_norm": 0.019385412335395813,
+ "learning_rate": 0.0006,
+ "loss": 4.2247724533081055,
+ "step": 2044
+ },
+ {
+ "epoch": 28.40524017467249,
+ "grad_norm": 0.019539108499884605,
+ "learning_rate": 0.0006,
+ "loss": 4.236655235290527,
+ "step": 2045
+ },
+ {
+ "epoch": 28.419213973799128,
+ "grad_norm": 0.01942690648138523,
+ "learning_rate": 0.0006,
+ "loss": 4.251850605010986,
+ "step": 2046
+ },
+ {
+ "epoch": 28.433187772925763,
+ "grad_norm": 0.019667886197566986,
+ "learning_rate": 0.0006,
+ "loss": 4.222312927246094,
+ "step": 2047
+ },
+ {
+ "epoch": 28.4471615720524,
+ "grad_norm": 0.02029012329876423,
+ "learning_rate": 0.0006,
+ "loss": 4.246252059936523,
+ "step": 2048
+ },
+ {
+ "epoch": 28.46113537117904,
+ "grad_norm": 0.01784469373524189,
+ "learning_rate": 0.0006,
+ "loss": 4.192128658294678,
+ "step": 2049
+ },
+ {
+ "epoch": 28.475109170305675,
+ "grad_norm": 0.0160287544131279,
+ "learning_rate": 0.0006,
+ "loss": 4.208607196807861,
+ "step": 2050
+ },
+ {
+ "epoch": 28.489082969432314,
+ "grad_norm": 0.015348346903920174,
+ "learning_rate": 0.0006,
+ "loss": 4.236598014831543,
+ "step": 2051
+ },
+ {
+ "epoch": 28.503056768558952,
+ "grad_norm": 0.015198206529021263,
+ "learning_rate": 0.0006,
+ "loss": 4.235552787780762,
+ "step": 2052
+ },
+ {
+ "epoch": 28.51703056768559,
+ "grad_norm": 0.01523754745721817,
+ "learning_rate": 0.0006,
+ "loss": 4.1941728591918945,
+ "step": 2053
+ },
+ {
+ "epoch": 28.531004366812226,
+ "grad_norm": 0.015559614636003971,
+ "learning_rate": 0.0006,
+ "loss": 4.294757843017578,
+ "step": 2054
+ },
+ {
+ "epoch": 28.544978165938865,
+ "grad_norm": 0.01568358950316906,
+ "learning_rate": 0.0006,
+ "loss": 4.213065147399902,
+ "step": 2055
+ },
+ {
+ "epoch": 28.558951965065503,
+ "grad_norm": 0.015336516313254833,
+ "learning_rate": 0.0006,
+ "loss": 4.240293979644775,
+ "step": 2056
+ },
+ {
+ "epoch": 28.57292576419214,
+ "grad_norm": 0.015041496604681015,
+ "learning_rate": 0.0006,
+ "loss": 4.273397445678711,
+ "step": 2057
+ },
+ {
+ "epoch": 28.586899563318777,
+ "grad_norm": 0.01610128954052925,
+ "learning_rate": 0.0006,
+ "loss": 4.125369071960449,
+ "step": 2058
+ },
+ {
+ "epoch": 28.600873362445416,
+ "grad_norm": 0.015974976122379303,
+ "learning_rate": 0.0006,
+ "loss": 4.133459091186523,
+ "step": 2059
+ },
+ {
+ "epoch": 28.61484716157205,
+ "grad_norm": 0.015936799347400665,
+ "learning_rate": 0.0006,
+ "loss": 4.306667327880859,
+ "step": 2060
+ },
+ {
+ "epoch": 28.62882096069869,
+ "grad_norm": 0.01795247197151184,
+ "learning_rate": 0.0006,
+ "loss": 4.205156326293945,
+ "step": 2061
+ },
+ {
+ "epoch": 28.64279475982533,
+ "grad_norm": 0.019639814272522926,
+ "learning_rate": 0.0006,
+ "loss": 4.277614593505859,
+ "step": 2062
+ },
+ {
+ "epoch": 28.656768558951963,
+ "grad_norm": 0.01722300611436367,
+ "learning_rate": 0.0006,
+ "loss": 4.191032886505127,
+ "step": 2063
+ },
+ {
+ "epoch": 28.670742358078602,
+ "grad_norm": 0.01612633652985096,
+ "learning_rate": 0.0006,
+ "loss": 4.20717191696167,
+ "step": 2064
+ },
+ {
+ "epoch": 28.68471615720524,
+ "grad_norm": 0.01603279449045658,
+ "learning_rate": 0.0006,
+ "loss": 4.228884220123291,
+ "step": 2065
+ },
+ {
+ "epoch": 28.69868995633188,
+ "grad_norm": 0.016405927017331123,
+ "learning_rate": 0.0006,
+ "loss": 4.201292991638184,
+ "step": 2066
+ },
+ {
+ "epoch": 28.712663755458514,
+ "grad_norm": 0.016644183546304703,
+ "learning_rate": 0.0006,
+ "loss": 4.239282131195068,
+ "step": 2067
+ },
+ {
+ "epoch": 28.726637554585153,
+ "grad_norm": 0.0175587497651577,
+ "learning_rate": 0.0006,
+ "loss": 4.289038181304932,
+ "step": 2068
+ },
+ {
+ "epoch": 28.74061135371179,
+ "grad_norm": 0.018223397433757782,
+ "learning_rate": 0.0006,
+ "loss": 4.2266998291015625,
+ "step": 2069
+ },
+ {
+ "epoch": 28.754585152838427,
+ "grad_norm": 0.018680868670344353,
+ "learning_rate": 0.0006,
+ "loss": 4.348849296569824,
+ "step": 2070
+ },
+ {
+ "epoch": 28.768558951965066,
+ "grad_norm": 0.016726728528738022,
+ "learning_rate": 0.0006,
+ "loss": 4.292080879211426,
+ "step": 2071
+ },
+ {
+ "epoch": 28.782532751091704,
+ "grad_norm": 0.017386795952916145,
+ "learning_rate": 0.0006,
+ "loss": 4.208870887756348,
+ "step": 2072
+ },
+ {
+ "epoch": 28.79650655021834,
+ "grad_norm": 0.01865178905427456,
+ "learning_rate": 0.0006,
+ "loss": 4.054150581359863,
+ "step": 2073
+ },
+ {
+ "epoch": 28.810480349344978,
+ "grad_norm": 0.01776815392076969,
+ "learning_rate": 0.0006,
+ "loss": 4.202404499053955,
+ "step": 2074
+ },
+ {
+ "epoch": 28.824454148471617,
+ "grad_norm": 0.01643497310578823,
+ "learning_rate": 0.0006,
+ "loss": 4.131855010986328,
+ "step": 2075
+ },
+ {
+ "epoch": 28.83842794759825,
+ "grad_norm": 0.01574273779988289,
+ "learning_rate": 0.0006,
+ "loss": 4.296788692474365,
+ "step": 2076
+ },
+ {
+ "epoch": 28.85240174672489,
+ "grad_norm": 0.015613908879458904,
+ "learning_rate": 0.0006,
+ "loss": 4.210949897766113,
+ "step": 2077
+ },
+ {
+ "epoch": 28.86637554585153,
+ "grad_norm": 0.015543212182819843,
+ "learning_rate": 0.0006,
+ "loss": 4.218136310577393,
+ "step": 2078
+ },
+ {
+ "epoch": 28.880349344978168,
+ "grad_norm": 0.014681251719594002,
+ "learning_rate": 0.0006,
+ "loss": 4.19401741027832,
+ "step": 2079
+ },
+ {
+ "epoch": 28.894323144104803,
+ "grad_norm": 0.014393828809261322,
+ "learning_rate": 0.0006,
+ "loss": 4.198468208312988,
+ "step": 2080
+ },
+ {
+ "epoch": 28.90829694323144,
+ "grad_norm": 0.01580170728266239,
+ "learning_rate": 0.0006,
+ "loss": 4.224969863891602,
+ "step": 2081
+ },
+ {
+ "epoch": 28.92227074235808,
+ "grad_norm": 0.01484165620058775,
+ "learning_rate": 0.0006,
+ "loss": 4.12236213684082,
+ "step": 2082
+ },
+ {
+ "epoch": 28.936244541484715,
+ "grad_norm": 0.01497623696923256,
+ "learning_rate": 0.0006,
+ "loss": 4.24141788482666,
+ "step": 2083
+ },
+ {
+ "epoch": 28.950218340611354,
+ "grad_norm": 0.015080046840012074,
+ "learning_rate": 0.0006,
+ "loss": 4.284151077270508,
+ "step": 2084
+ },
+ {
+ "epoch": 28.964192139737992,
+ "grad_norm": 0.016207212582230568,
+ "learning_rate": 0.0006,
+ "loss": 4.298693656921387,
+ "step": 2085
+ },
+ {
+ "epoch": 28.978165938864628,
+ "grad_norm": 0.015474417246878147,
+ "learning_rate": 0.0006,
+ "loss": 4.2412261962890625,
+ "step": 2086
+ },
+ {
+ "epoch": 28.992139737991266,
+ "grad_norm": 0.01515351701527834,
+ "learning_rate": 0.0006,
+ "loss": 4.3054423332214355,
+ "step": 2087
+ },
+ {
+ "epoch": 29.0,
+ "grad_norm": 0.016411345452070236,
+ "learning_rate": 0.0006,
+ "loss": 4.104830741882324,
+ "step": 2088
+ },
+ {
+ "epoch": 29.0,
+ "eval_loss": 4.661567687988281,
+ "eval_runtime": 57.1244,
+ "eval_samples_per_second": 42.749,
+ "eval_steps_per_second": 1.348,
+ "step": 2088
+ },
+ {
+ "epoch": 29.01397379912664,
+ "grad_norm": 0.01665342226624489,
+ "learning_rate": 0.0006,
+ "loss": 4.311391830444336,
+ "step": 2089
+ },
+ {
+ "epoch": 29.027947598253274,
+ "grad_norm": 0.018053214997053146,
+ "learning_rate": 0.0006,
+ "loss": 4.135597229003906,
+ "step": 2090
+ },
+ {
+ "epoch": 29.041921397379912,
+ "grad_norm": 0.018213748931884766,
+ "learning_rate": 0.0006,
+ "loss": 4.303299903869629,
+ "step": 2091
+ },
+ {
+ "epoch": 29.05589519650655,
+ "grad_norm": 0.018460242077708244,
+ "learning_rate": 0.0006,
+ "loss": 4.20850944519043,
+ "step": 2092
+ },
+ {
+ "epoch": 29.069868995633186,
+ "grad_norm": 0.018365809693932533,
+ "learning_rate": 0.0006,
+ "loss": 4.275847434997559,
+ "step": 2093
+ },
+ {
+ "epoch": 29.083842794759825,
+ "grad_norm": 0.018346186727285385,
+ "learning_rate": 0.0006,
+ "loss": 4.202091217041016,
+ "step": 2094
+ },
+ {
+ "epoch": 29.097816593886463,
+ "grad_norm": 0.017920587211847305,
+ "learning_rate": 0.0006,
+ "loss": 4.182092666625977,
+ "step": 2095
+ },
+ {
+ "epoch": 29.111790393013102,
+ "grad_norm": 0.01812003180384636,
+ "learning_rate": 0.0006,
+ "loss": 4.259721755981445,
+ "step": 2096
+ },
+ {
+ "epoch": 29.125764192139737,
+ "grad_norm": 0.01700986735522747,
+ "learning_rate": 0.0006,
+ "loss": 4.067573070526123,
+ "step": 2097
+ },
+ {
+ "epoch": 29.139737991266376,
+ "grad_norm": 0.018069760873913765,
+ "learning_rate": 0.0006,
+ "loss": 4.234400749206543,
+ "step": 2098
+ },
+ {
+ "epoch": 29.153711790393015,
+ "grad_norm": 0.01905428245663643,
+ "learning_rate": 0.0006,
+ "loss": 4.16609001159668,
+ "step": 2099
+ },
+ {
+ "epoch": 29.16768558951965,
+ "grad_norm": 0.018711242824792862,
+ "learning_rate": 0.0006,
+ "loss": 4.249088764190674,
+ "step": 2100
+ },
+ {
+ "epoch": 29.18165938864629,
+ "grad_norm": 0.018721306696534157,
+ "learning_rate": 0.0006,
+ "loss": 4.14540958404541,
+ "step": 2101
+ },
+ {
+ "epoch": 29.195633187772927,
+ "grad_norm": 0.020139874890446663,
+ "learning_rate": 0.0006,
+ "loss": 4.049248695373535,
+ "step": 2102
+ },
+ {
+ "epoch": 29.209606986899562,
+ "grad_norm": 0.020849574357271194,
+ "learning_rate": 0.0006,
+ "loss": 4.235224723815918,
+ "step": 2103
+ },
+ {
+ "epoch": 29.2235807860262,
+ "grad_norm": 0.018693549558520317,
+ "learning_rate": 0.0006,
+ "loss": 4.232961177825928,
+ "step": 2104
+ },
+ {
+ "epoch": 29.23755458515284,
+ "grad_norm": 0.017889857292175293,
+ "learning_rate": 0.0006,
+ "loss": 4.171422004699707,
+ "step": 2105
+ },
+ {
+ "epoch": 29.251528384279474,
+ "grad_norm": 0.016690224409103394,
+ "learning_rate": 0.0006,
+ "loss": 4.244717597961426,
+ "step": 2106
+ },
+ {
+ "epoch": 29.265502183406113,
+ "grad_norm": 0.01708616502583027,
+ "learning_rate": 0.0006,
+ "loss": 4.125068664550781,
+ "step": 2107
+ },
+ {
+ "epoch": 29.27947598253275,
+ "grad_norm": 0.017928237095475197,
+ "learning_rate": 0.0006,
+ "loss": 4.208292007446289,
+ "step": 2108
+ },
+ {
+ "epoch": 29.29344978165939,
+ "grad_norm": 0.018250394612550735,
+ "learning_rate": 0.0006,
+ "loss": 4.118630886077881,
+ "step": 2109
+ },
+ {
+ "epoch": 29.307423580786025,
+ "grad_norm": 0.020160246640443802,
+ "learning_rate": 0.0006,
+ "loss": 4.206025123596191,
+ "step": 2110
+ },
+ {
+ "epoch": 29.321397379912664,
+ "grad_norm": 0.020045453682541847,
+ "learning_rate": 0.0006,
+ "loss": 4.2758684158325195,
+ "step": 2111
+ },
+ {
+ "epoch": 29.335371179039303,
+ "grad_norm": 0.019556907936930656,
+ "learning_rate": 0.0006,
+ "loss": 4.188272953033447,
+ "step": 2112
+ },
+ {
+ "epoch": 29.349344978165938,
+ "grad_norm": 0.019185200333595276,
+ "learning_rate": 0.0006,
+ "loss": 4.17451286315918,
+ "step": 2113
+ },
+ {
+ "epoch": 29.363318777292577,
+ "grad_norm": 0.01816629432141781,
+ "learning_rate": 0.0006,
+ "loss": 4.165246963500977,
+ "step": 2114
+ },
+ {
+ "epoch": 29.377292576419215,
+ "grad_norm": 0.01865146867930889,
+ "learning_rate": 0.0006,
+ "loss": 4.148184776306152,
+ "step": 2115
+ },
+ {
+ "epoch": 29.39126637554585,
+ "grad_norm": 0.018338464200496674,
+ "learning_rate": 0.0006,
+ "loss": 4.288936614990234,
+ "step": 2116
+ },
+ {
+ "epoch": 29.40524017467249,
+ "grad_norm": 0.018986187875270844,
+ "learning_rate": 0.0006,
+ "loss": 4.183379650115967,
+ "step": 2117
+ },
+ {
+ "epoch": 29.419213973799128,
+ "grad_norm": 0.018409637734293938,
+ "learning_rate": 0.0006,
+ "loss": 4.057574272155762,
+ "step": 2118
+ },
+ {
+ "epoch": 29.433187772925763,
+ "grad_norm": 0.016664542257785797,
+ "learning_rate": 0.0006,
+ "loss": 4.18517541885376,
+ "step": 2119
+ },
+ {
+ "epoch": 29.4471615720524,
+ "grad_norm": 0.016486089676618576,
+ "learning_rate": 0.0006,
+ "loss": 4.179367542266846,
+ "step": 2120
+ },
+ {
+ "epoch": 29.46113537117904,
+ "grad_norm": 0.01787523552775383,
+ "learning_rate": 0.0006,
+ "loss": 4.184126377105713,
+ "step": 2121
+ },
+ {
+ "epoch": 29.475109170305675,
+ "grad_norm": 0.016574613749980927,
+ "learning_rate": 0.0006,
+ "loss": 4.290170192718506,
+ "step": 2122
+ },
+ {
+ "epoch": 29.489082969432314,
+ "grad_norm": 0.01643376611173153,
+ "learning_rate": 0.0006,
+ "loss": 4.052913665771484,
+ "step": 2123
+ },
+ {
+ "epoch": 29.503056768558952,
+ "grad_norm": 0.016503969207406044,
+ "learning_rate": 0.0006,
+ "loss": 4.335184097290039,
+ "step": 2124
+ },
+ {
+ "epoch": 29.51703056768559,
+ "grad_norm": 0.017596479505300522,
+ "learning_rate": 0.0006,
+ "loss": 4.165539741516113,
+ "step": 2125
+ },
+ {
+ "epoch": 29.531004366812226,
+ "grad_norm": 0.017468582838773727,
+ "learning_rate": 0.0006,
+ "loss": 4.315964698791504,
+ "step": 2126
+ },
+ {
+ "epoch": 29.544978165938865,
+ "grad_norm": 0.01724618673324585,
+ "learning_rate": 0.0006,
+ "loss": 4.243681907653809,
+ "step": 2127
+ },
+ {
+ "epoch": 29.558951965065503,
+ "grad_norm": 0.018384616822004318,
+ "learning_rate": 0.0006,
+ "loss": 4.112447261810303,
+ "step": 2128
+ },
+ {
+ "epoch": 29.57292576419214,
+ "grad_norm": 0.01824074238538742,
+ "learning_rate": 0.0006,
+ "loss": 4.236065864562988,
+ "step": 2129
+ },
+ {
+ "epoch": 29.586899563318777,
+ "grad_norm": 0.01625971868634224,
+ "learning_rate": 0.0006,
+ "loss": 4.18143892288208,
+ "step": 2130
+ },
+ {
+ "epoch": 29.600873362445416,
+ "grad_norm": 0.015537494793534279,
+ "learning_rate": 0.0006,
+ "loss": 4.225147247314453,
+ "step": 2131
+ },
+ {
+ "epoch": 29.61484716157205,
+ "grad_norm": 0.015818974003195763,
+ "learning_rate": 0.0006,
+ "loss": 4.3100175857543945,
+ "step": 2132
+ },
+ {
+ "epoch": 29.62882096069869,
+ "grad_norm": 0.015909671783447266,
+ "learning_rate": 0.0006,
+ "loss": 4.169775009155273,
+ "step": 2133
+ },
+ {
+ "epoch": 29.64279475982533,
+ "grad_norm": 0.01769658550620079,
+ "learning_rate": 0.0006,
+ "loss": 4.278904914855957,
+ "step": 2134
+ },
+ {
+ "epoch": 29.656768558951963,
+ "grad_norm": 0.015641704201698303,
+ "learning_rate": 0.0006,
+ "loss": 4.2426228523254395,
+ "step": 2135
+ },
+ {
+ "epoch": 29.670742358078602,
+ "grad_norm": 0.01565658301115036,
+ "learning_rate": 0.0006,
+ "loss": 4.232687950134277,
+ "step": 2136
+ },
+ {
+ "epoch": 29.68471615720524,
+ "grad_norm": 0.016485940665006638,
+ "learning_rate": 0.0006,
+ "loss": 4.209839820861816,
+ "step": 2137
+ },
+ {
+ "epoch": 29.69868995633188,
+ "grad_norm": 0.017622729763388634,
+ "learning_rate": 0.0006,
+ "loss": 4.196017265319824,
+ "step": 2138
+ },
+ {
+ "epoch": 29.712663755458514,
+ "grad_norm": 0.018320564180612564,
+ "learning_rate": 0.0006,
+ "loss": 4.225683689117432,
+ "step": 2139
+ },
+ {
+ "epoch": 29.726637554585153,
+ "grad_norm": 0.018140794709324837,
+ "learning_rate": 0.0006,
+ "loss": 4.311944961547852,
+ "step": 2140
+ },
+ {
+ "epoch": 29.74061135371179,
+ "grad_norm": 0.016227100044488907,
+ "learning_rate": 0.0006,
+ "loss": 4.230257511138916,
+ "step": 2141
+ },
+ {
+ "epoch": 29.754585152838427,
+ "grad_norm": 0.015759488567709923,
+ "learning_rate": 0.0006,
+ "loss": 4.119174003601074,
+ "step": 2142
+ },
+ {
+ "epoch": 29.768558951965066,
+ "grad_norm": 0.01757766678929329,
+ "learning_rate": 0.0006,
+ "loss": 4.165826797485352,
+ "step": 2143
+ },
+ {
+ "epoch": 29.782532751091704,
+ "grad_norm": 0.019661923870444298,
+ "learning_rate": 0.0006,
+ "loss": 4.392203330993652,
+ "step": 2144
+ },
+ {
+ "epoch": 29.79650655021834,
+ "grad_norm": 0.019478755071759224,
+ "learning_rate": 0.0006,
+ "loss": 4.348371982574463,
+ "step": 2145
+ },
+ {
+ "epoch": 29.810480349344978,
+ "grad_norm": 0.019149569794535637,
+ "learning_rate": 0.0006,
+ "loss": 4.34104061126709,
+ "step": 2146
+ },
+ {
+ "epoch": 29.824454148471617,
+ "grad_norm": 0.017707521095871925,
+ "learning_rate": 0.0006,
+ "loss": 4.228226661682129,
+ "step": 2147
+ },
+ {
+ "epoch": 29.83842794759825,
+ "grad_norm": 0.015615378506481647,
+ "learning_rate": 0.0006,
+ "loss": 4.175541877746582,
+ "step": 2148
+ },
+ {
+ "epoch": 29.85240174672489,
+ "grad_norm": 0.016111129894852638,
+ "learning_rate": 0.0006,
+ "loss": 4.267936706542969,
+ "step": 2149
+ },
+ {
+ "epoch": 29.86637554585153,
+ "grad_norm": 0.01528779324144125,
+ "learning_rate": 0.0006,
+ "loss": 4.186014175415039,
+ "step": 2150
+ },
+ {
+ "epoch": 29.880349344978168,
+ "grad_norm": 0.014925646595656872,
+ "learning_rate": 0.0006,
+ "loss": 4.174860000610352,
+ "step": 2151
+ },
+ {
+ "epoch": 29.894323144104803,
+ "grad_norm": 0.016917673870921135,
+ "learning_rate": 0.0006,
+ "loss": 4.33908748626709,
+ "step": 2152
+ },
+ {
+ "epoch": 29.90829694323144,
+ "grad_norm": 0.016007075086236,
+ "learning_rate": 0.0006,
+ "loss": 4.266141414642334,
+ "step": 2153
+ },
+ {
+ "epoch": 29.92227074235808,
+ "grad_norm": 0.016178956255316734,
+ "learning_rate": 0.0006,
+ "loss": 4.271215915679932,
+ "step": 2154
+ },
+ {
+ "epoch": 29.936244541484715,
+ "grad_norm": 0.016180871054530144,
+ "learning_rate": 0.0006,
+ "loss": 4.278801441192627,
+ "step": 2155
+ },
+ {
+ "epoch": 29.950218340611354,
+ "grad_norm": 0.015335503034293652,
+ "learning_rate": 0.0006,
+ "loss": 4.194978713989258,
+ "step": 2156
+ },
+ {
+ "epoch": 29.964192139737992,
+ "grad_norm": 0.016580764204263687,
+ "learning_rate": 0.0006,
+ "loss": 4.290809631347656,
+ "step": 2157
+ },
+ {
+ "epoch": 29.978165938864628,
+ "grad_norm": 0.01623128168284893,
+ "learning_rate": 0.0006,
+ "loss": 4.205451488494873,
+ "step": 2158
+ },
+ {
+ "epoch": 29.992139737991266,
+ "grad_norm": 0.014904248528182507,
+ "learning_rate": 0.0006,
+ "loss": 4.24924373626709,
+ "step": 2159
+ },
+ {
+ "epoch": 30.0,
+ "grad_norm": 0.017089299857616425,
+ "learning_rate": 0.0006,
+ "loss": 4.222072601318359,
+ "step": 2160
+ },
+ {
+ "epoch": 30.0,
+ "eval_loss": 4.585411071777344,
+ "eval_runtime": 57.1164,
+ "eval_samples_per_second": 42.755,
+ "eval_steps_per_second": 1.348,
+ "step": 2160
+ },
+ {
+ "epoch": 30.01397379912664,
+ "grad_norm": 0.017226383090019226,
+ "learning_rate": 0.0006,
+ "loss": 4.19467830657959,
+ "step": 2161
+ },
+ {
+ "epoch": 30.027947598253274,
+ "grad_norm": 0.01635066606104374,
+ "learning_rate": 0.0006,
+ "loss": 4.176875591278076,
+ "step": 2162
+ },
+ {
+ "epoch": 30.041921397379912,
+ "grad_norm": 0.01735362410545349,
+ "learning_rate": 0.0006,
+ "loss": 4.13667106628418,
+ "step": 2163
+ },
+ {
+ "epoch": 30.05589519650655,
+ "grad_norm": 0.017047109082341194,
+ "learning_rate": 0.0006,
+ "loss": 4.136329650878906,
+ "step": 2164
+ },
+ {
+ "epoch": 30.069868995633186,
+ "grad_norm": 0.018116053193807602,
+ "learning_rate": 0.0006,
+ "loss": 4.228706359863281,
+ "step": 2165
+ },
+ {
+ "epoch": 30.083842794759825,
+ "grad_norm": 0.01791389286518097,
+ "learning_rate": 0.0006,
+ "loss": 4.1597747802734375,
+ "step": 2166
+ },
+ {
+ "epoch": 30.097816593886463,
+ "grad_norm": 0.017793502658605576,
+ "learning_rate": 0.0006,
+ "loss": 4.0807976722717285,
+ "step": 2167
+ },
+ {
+ "epoch": 30.111790393013102,
+ "grad_norm": 0.01827503927052021,
+ "learning_rate": 0.0006,
+ "loss": 4.170646667480469,
+ "step": 2168
+ },
+ {
+ "epoch": 30.125764192139737,
+ "grad_norm": 0.017228564247488976,
+ "learning_rate": 0.0006,
+ "loss": 4.146934509277344,
+ "step": 2169
+ },
+ {
+ "epoch": 30.139737991266376,
+ "grad_norm": 0.01757437363266945,
+ "learning_rate": 0.0006,
+ "loss": 4.1859130859375,
+ "step": 2170
+ },
+ {
+ "epoch": 30.153711790393015,
+ "grad_norm": 0.01701093092560768,
+ "learning_rate": 0.0006,
+ "loss": 4.126981258392334,
+ "step": 2171
+ },
+ {
+ "epoch": 30.16768558951965,
+ "grad_norm": 0.015220489352941513,
+ "learning_rate": 0.0006,
+ "loss": 4.164018630981445,
+ "step": 2172
+ },
+ {
+ "epoch": 30.18165938864629,
+ "grad_norm": 0.016254238784313202,
+ "learning_rate": 0.0006,
+ "loss": 4.258275508880615,
+ "step": 2173
+ },
+ {
+ "epoch": 30.195633187772927,
+ "grad_norm": 0.016609614714980125,
+ "learning_rate": 0.0006,
+ "loss": 4.222233772277832,
+ "step": 2174
+ },
+ {
+ "epoch": 30.209606986899562,
+ "grad_norm": 0.016531318426132202,
+ "learning_rate": 0.0006,
+ "loss": 4.144489288330078,
+ "step": 2175
+ },
+ {
+ "epoch": 30.2235807860262,
+ "grad_norm": 0.01567668654024601,
+ "learning_rate": 0.0006,
+ "loss": 4.181916236877441,
+ "step": 2176
+ },
+ {
+ "epoch": 30.23755458515284,
+ "grad_norm": 0.017195681110024452,
+ "learning_rate": 0.0006,
+ "loss": 4.189693927764893,
+ "step": 2177
+ },
+ {
+ "epoch": 30.251528384279474,
+ "grad_norm": 0.016882948577404022,
+ "learning_rate": 0.0006,
+ "loss": 4.21604061126709,
+ "step": 2178
+ },
+ {
+ "epoch": 30.265502183406113,
+ "grad_norm": 0.016261622309684753,
+ "learning_rate": 0.0006,
+ "loss": 4.227639198303223,
+ "step": 2179
+ },
+ {
+ "epoch": 30.27947598253275,
+ "grad_norm": 0.016411812976002693,
+ "learning_rate": 0.0006,
+ "loss": 4.130362510681152,
+ "step": 2180
+ },
+ {
+ "epoch": 30.29344978165939,
+ "grad_norm": 0.01732019893825054,
+ "learning_rate": 0.0006,
+ "loss": 4.230443000793457,
+ "step": 2181
+ },
+ {
+ "epoch": 30.307423580786025,
+ "grad_norm": 0.01825905591249466,
+ "learning_rate": 0.0006,
+ "loss": 4.175817489624023,
+ "step": 2182
+ },
+ {
+ "epoch": 30.321397379912664,
+ "grad_norm": 0.01887713558971882,
+ "learning_rate": 0.0006,
+ "loss": 4.169020652770996,
+ "step": 2183
+ },
+ {
+ "epoch": 30.335371179039303,
+ "grad_norm": 0.019391661509871483,
+ "learning_rate": 0.0006,
+ "loss": 4.176318645477295,
+ "step": 2184
+ },
+ {
+ "epoch": 30.349344978165938,
+ "grad_norm": 0.020840002223849297,
+ "learning_rate": 0.0006,
+ "loss": 4.174228668212891,
+ "step": 2185
+ },
+ {
+ "epoch": 30.363318777292577,
+ "grad_norm": 0.020193178206682205,
+ "learning_rate": 0.0006,
+ "loss": 4.127919673919678,
+ "step": 2186
+ },
+ {
+ "epoch": 30.377292576419215,
+ "grad_norm": 0.018449561670422554,
+ "learning_rate": 0.0006,
+ "loss": 4.229487895965576,
+ "step": 2187
+ },
+ {
+ "epoch": 30.39126637554585,
+ "grad_norm": 0.01711983233690262,
+ "learning_rate": 0.0006,
+ "loss": 4.201651573181152,
+ "step": 2188
+ },
+ {
+ "epoch": 30.40524017467249,
+ "grad_norm": 0.017293395474553108,
+ "learning_rate": 0.0006,
+ "loss": 4.168704032897949,
+ "step": 2189
+ },
+ {
+ "epoch": 30.419213973799128,
+ "grad_norm": 0.018526358529925346,
+ "learning_rate": 0.0006,
+ "loss": 4.171864032745361,
+ "step": 2190
+ },
+ {
+ "epoch": 30.433187772925763,
+ "grad_norm": 0.019287291914224625,
+ "learning_rate": 0.0006,
+ "loss": 4.039534568786621,
+ "step": 2191
+ },
+ {
+ "epoch": 30.4471615720524,
+ "grad_norm": 0.016902444884181023,
+ "learning_rate": 0.0006,
+ "loss": 4.282787799835205,
+ "step": 2192
+ },
+ {
+ "epoch": 30.46113537117904,
+ "grad_norm": 0.01656750962138176,
+ "learning_rate": 0.0006,
+ "loss": 4.213098526000977,
+ "step": 2193
+ },
+ {
+ "epoch": 30.475109170305675,
+ "grad_norm": 0.016805065795779228,
+ "learning_rate": 0.0006,
+ "loss": 4.1583476066589355,
+ "step": 2194
+ },
+ {
+ "epoch": 30.489082969432314,
+ "grad_norm": 0.017166638746857643,
+ "learning_rate": 0.0006,
+ "loss": 4.188634395599365,
+ "step": 2195
+ },
+ {
+ "epoch": 30.503056768558952,
+ "grad_norm": 0.01592106744647026,
+ "learning_rate": 0.0006,
+ "loss": 4.234043121337891,
+ "step": 2196
+ },
+ {
+ "epoch": 30.51703056768559,
+ "grad_norm": 0.015270989388227463,
+ "learning_rate": 0.0006,
+ "loss": 4.301420211791992,
+ "step": 2197
+ },
+ {
+ "epoch": 30.531004366812226,
+ "grad_norm": 0.015194724313914776,
+ "learning_rate": 0.0006,
+ "loss": 4.18703031539917,
+ "step": 2198
+ },
+ {
+ "epoch": 30.544978165938865,
+ "grad_norm": 0.015587746165692806,
+ "learning_rate": 0.0006,
+ "loss": 4.138928413391113,
+ "step": 2199
+ },
+ {
+ "epoch": 30.558951965065503,
+ "grad_norm": 0.016029570251703262,
+ "learning_rate": 0.0006,
+ "loss": 4.290863990783691,
+ "step": 2200
+ },
+ {
+ "epoch": 30.57292576419214,
+ "grad_norm": 0.015651309862732887,
+ "learning_rate": 0.0006,
+ "loss": 4.272062301635742,
+ "step": 2201
+ },
+ {
+ "epoch": 30.586899563318777,
+ "grad_norm": 0.015992306172847748,
+ "learning_rate": 0.0006,
+ "loss": 4.2487406730651855,
+ "step": 2202
+ },
+ {
+ "epoch": 30.600873362445416,
+ "grad_norm": 0.014969157986342907,
+ "learning_rate": 0.0006,
+ "loss": 4.115347862243652,
+ "step": 2203
+ },
+ {
+ "epoch": 30.61484716157205,
+ "grad_norm": 0.014639027416706085,
+ "learning_rate": 0.0006,
+ "loss": 4.2207489013671875,
+ "step": 2204
+ },
+ {
+ "epoch": 30.62882096069869,
+ "grad_norm": 0.015059413388371468,
+ "learning_rate": 0.0006,
+ "loss": 4.274044990539551,
+ "step": 2205
+ },
+ {
+ "epoch": 30.64279475982533,
+ "grad_norm": 0.0166641678661108,
+ "learning_rate": 0.0006,
+ "loss": 4.191373348236084,
+ "step": 2206
+ },
+ {
+ "epoch": 30.656768558951963,
+ "grad_norm": 0.017806345596909523,
+ "learning_rate": 0.0006,
+ "loss": 4.139800548553467,
+ "step": 2207
+ },
+ {
+ "epoch": 30.670742358078602,
+ "grad_norm": 0.018712421879172325,
+ "learning_rate": 0.0006,
+ "loss": 4.035965442657471,
+ "step": 2208
+ },
+ {
+ "epoch": 30.68471615720524,
+ "grad_norm": 0.018457243219017982,
+ "learning_rate": 0.0006,
+ "loss": 4.2579240798950195,
+ "step": 2209
+ },
+ {
+ "epoch": 30.69868995633188,
+ "grad_norm": 0.01920999586582184,
+ "learning_rate": 0.0006,
+ "loss": 4.131112098693848,
+ "step": 2210
+ },
+ {
+ "epoch": 30.712663755458514,
+ "grad_norm": 0.01729944534599781,
+ "learning_rate": 0.0006,
+ "loss": 4.254059314727783,
+ "step": 2211
+ },
+ {
+ "epoch": 30.726637554585153,
+ "grad_norm": 0.016914231702685356,
+ "learning_rate": 0.0006,
+ "loss": 4.233428955078125,
+ "step": 2212
+ },
+ {
+ "epoch": 30.74061135371179,
+ "grad_norm": 0.01913098618388176,
+ "learning_rate": 0.0006,
+ "loss": 4.23525333404541,
+ "step": 2213
+ },
+ {
+ "epoch": 30.754585152838427,
+ "grad_norm": 0.019150281324982643,
+ "learning_rate": 0.0006,
+ "loss": 4.213695049285889,
+ "step": 2214
+ },
+ {
+ "epoch": 30.768558951965066,
+ "grad_norm": 0.018947452306747437,
+ "learning_rate": 0.0006,
+ "loss": 4.151697158813477,
+ "step": 2215
+ },
+ {
+ "epoch": 30.782532751091704,
+ "grad_norm": 0.016058508306741714,
+ "learning_rate": 0.0006,
+ "loss": 4.220280170440674,
+ "step": 2216
+ },
+ {
+ "epoch": 30.79650655021834,
+ "grad_norm": 0.017488619312644005,
+ "learning_rate": 0.0006,
+ "loss": 4.154428482055664,
+ "step": 2217
+ },
+ {
+ "epoch": 30.810480349344978,
+ "grad_norm": 0.017355090007185936,
+ "learning_rate": 0.0006,
+ "loss": 4.256086349487305,
+ "step": 2218
+ },
+ {
+ "epoch": 30.824454148471617,
+ "grad_norm": 0.016636032611131668,
+ "learning_rate": 0.0006,
+ "loss": 4.227564334869385,
+ "step": 2219
+ },
+ {
+ "epoch": 30.83842794759825,
+ "grad_norm": 0.016273891553282738,
+ "learning_rate": 0.0006,
+ "loss": 4.200466156005859,
+ "step": 2220
+ },
+ {
+ "epoch": 30.85240174672489,
+ "grad_norm": 0.016939911991357803,
+ "learning_rate": 0.0006,
+ "loss": 4.1571221351623535,
+ "step": 2221
+ },
+ {
+ "epoch": 30.86637554585153,
+ "grad_norm": 0.018332522362470627,
+ "learning_rate": 0.0006,
+ "loss": 4.228398323059082,
+ "step": 2222
+ },
+ {
+ "epoch": 30.880349344978168,
+ "grad_norm": 0.019508114084601402,
+ "learning_rate": 0.0006,
+ "loss": 4.222960472106934,
+ "step": 2223
+ },
+ {
+ "epoch": 30.894323144104803,
+ "grad_norm": 0.018440278246998787,
+ "learning_rate": 0.0006,
+ "loss": 4.261051177978516,
+ "step": 2224
+ },
+ {
+ "epoch": 30.90829694323144,
+ "grad_norm": 0.018113229423761368,
+ "learning_rate": 0.0006,
+ "loss": 4.109235763549805,
+ "step": 2225
+ },
+ {
+ "epoch": 30.92227074235808,
+ "grad_norm": 0.015392083674669266,
+ "learning_rate": 0.0006,
+ "loss": 4.21950101852417,
+ "step": 2226
+ },
+ {
+ "epoch": 30.936244541484715,
+ "grad_norm": 0.01633290946483612,
+ "learning_rate": 0.0006,
+ "loss": 4.176053047180176,
+ "step": 2227
+ },
+ {
+ "epoch": 30.950218340611354,
+ "grad_norm": 0.016236383467912674,
+ "learning_rate": 0.0006,
+ "loss": 4.314309120178223,
+ "step": 2228
+ },
+ {
+ "epoch": 30.964192139737992,
+ "grad_norm": 0.015284373424947262,
+ "learning_rate": 0.0006,
+ "loss": 4.303133487701416,
+ "step": 2229
+ },
+ {
+ "epoch": 30.978165938864628,
+ "grad_norm": 0.015982897952198982,
+ "learning_rate": 0.0006,
+ "loss": 4.25739860534668,
+ "step": 2230
+ },
+ {
+ "epoch": 30.992139737991266,
+ "grad_norm": 0.016682198271155357,
+ "learning_rate": 0.0006,
+ "loss": 4.292210578918457,
+ "step": 2231
+ },
+ {
+ "epoch": 31.0,
+ "grad_norm": 0.017521467059850693,
+ "learning_rate": 0.0006,
+ "loss": 4.170637130737305,
+ "step": 2232
+ },
+ {
+ "epoch": 31.0,
+ "eval_loss": 4.642786979675293,
+ "eval_runtime": 57.029,
+ "eval_samples_per_second": 42.82,
+ "eval_steps_per_second": 1.35,
+ "step": 2232
+ },
+ {
+ "epoch": 31.01397379912664,
+ "grad_norm": 0.016607852652668953,
+ "learning_rate": 0.0006,
+ "loss": 4.106224536895752,
+ "step": 2233
+ },
+ {
+ "epoch": 31.027947598253274,
+ "grad_norm": 0.018448904156684875,
+ "learning_rate": 0.0006,
+ "loss": 4.224459648132324,
+ "step": 2234
+ },
+ {
+ "epoch": 31.041921397379912,
+ "grad_norm": 0.020481228828430176,
+ "learning_rate": 0.0006,
+ "loss": 4.305548667907715,
+ "step": 2235
+ },
+ {
+ "epoch": 31.05589519650655,
+ "grad_norm": 0.01999707892537117,
+ "learning_rate": 0.0006,
+ "loss": 4.253663539886475,
+ "step": 2236
+ },
+ {
+ "epoch": 31.069868995633186,
+ "grad_norm": 0.020428303629159927,
+ "learning_rate": 0.0006,
+ "loss": 4.166954040527344,
+ "step": 2237
+ },
+ {
+ "epoch": 31.083842794759825,
+ "grad_norm": 0.022608213126659393,
+ "learning_rate": 0.0006,
+ "loss": 4.172819137573242,
+ "step": 2238
+ },
+ {
+ "epoch": 31.097816593886463,
+ "grad_norm": 0.024669520556926727,
+ "learning_rate": 0.0006,
+ "loss": 4.106586456298828,
+ "step": 2239
+ },
+ {
+ "epoch": 31.111790393013102,
+ "grad_norm": 0.02247351035475731,
+ "learning_rate": 0.0006,
+ "loss": 4.160271644592285,
+ "step": 2240
+ },
+ {
+ "epoch": 31.125764192139737,
+ "grad_norm": 0.0207810141146183,
+ "learning_rate": 0.0006,
+ "loss": 4.272491931915283,
+ "step": 2241
+ },
+ {
+ "epoch": 31.139737991266376,
+ "grad_norm": 0.021576499566435814,
+ "learning_rate": 0.0006,
+ "loss": 4.171474456787109,
+ "step": 2242
+ },
+ {
+ "epoch": 31.153711790393015,
+ "grad_norm": 0.02337106689810753,
+ "learning_rate": 0.0006,
+ "loss": 4.094712734222412,
+ "step": 2243
+ },
+ {
+ "epoch": 31.16768558951965,
+ "grad_norm": 0.02219875156879425,
+ "learning_rate": 0.0006,
+ "loss": 4.1800994873046875,
+ "step": 2244
+ },
+ {
+ "epoch": 31.18165938864629,
+ "grad_norm": 0.022680504247546196,
+ "learning_rate": 0.0006,
+ "loss": 4.2054362297058105,
+ "step": 2245
+ },
+ {
+ "epoch": 31.195633187772927,
+ "grad_norm": 0.0208734143525362,
+ "learning_rate": 0.0006,
+ "loss": 4.06068229675293,
+ "step": 2246
+ },
+ {
+ "epoch": 31.209606986899562,
+ "grad_norm": 0.01811066083610058,
+ "learning_rate": 0.0006,
+ "loss": 4.2465128898620605,
+ "step": 2247
+ },
+ {
+ "epoch": 31.2235807860262,
+ "grad_norm": 0.018571637570858,
+ "learning_rate": 0.0006,
+ "loss": 4.191955089569092,
+ "step": 2248
+ },
+ {
+ "epoch": 31.23755458515284,
+ "grad_norm": 0.018598807975649834,
+ "learning_rate": 0.0006,
+ "loss": 4.200872421264648,
+ "step": 2249
+ },
+ {
+ "epoch": 31.251528384279474,
+ "grad_norm": 0.017577625811100006,
+ "learning_rate": 0.0006,
+ "loss": 4.148700714111328,
+ "step": 2250
+ },
+ {
+ "epoch": 31.265502183406113,
+ "grad_norm": 0.01817980408668518,
+ "learning_rate": 0.0006,
+ "loss": 4.256978988647461,
+ "step": 2251
+ },
+ {
+ "epoch": 31.27947598253275,
+ "grad_norm": 0.017983486875891685,
+ "learning_rate": 0.0006,
+ "loss": 4.175005912780762,
+ "step": 2252
+ },
+ {
+ "epoch": 31.29344978165939,
+ "grad_norm": 0.016751961782574654,
+ "learning_rate": 0.0006,
+ "loss": 4.237518310546875,
+ "step": 2253
+ },
+ {
+ "epoch": 31.307423580786025,
+ "grad_norm": 0.016407202929258347,
+ "learning_rate": 0.0006,
+ "loss": 4.197269439697266,
+ "step": 2254
+ },
+ {
+ "epoch": 31.321397379912664,
+ "grad_norm": 0.01578645594418049,
+ "learning_rate": 0.0006,
+ "loss": 4.179751396179199,
+ "step": 2255
+ },
+ {
+ "epoch": 31.335371179039303,
+ "grad_norm": 0.016001226380467415,
+ "learning_rate": 0.0006,
+ "loss": 4.08821964263916,
+ "step": 2256
+ },
+ {
+ "epoch": 31.349344978165938,
+ "grad_norm": 0.015122811309993267,
+ "learning_rate": 0.0006,
+ "loss": 4.047828674316406,
+ "step": 2257
+ },
+ {
+ "epoch": 31.363318777292577,
+ "grad_norm": 0.015399965457618237,
+ "learning_rate": 0.0006,
+ "loss": 4.1506524085998535,
+ "step": 2258
+ },
+ {
+ "epoch": 31.377292576419215,
+ "grad_norm": 0.01534605398774147,
+ "learning_rate": 0.0006,
+ "loss": 4.27952766418457,
+ "step": 2259
+ },
+ {
+ "epoch": 31.39126637554585,
+ "grad_norm": 0.015461381524801254,
+ "learning_rate": 0.0006,
+ "loss": 4.27102518081665,
+ "step": 2260
+ },
+ {
+ "epoch": 31.40524017467249,
+ "grad_norm": 0.016769153997302055,
+ "learning_rate": 0.0006,
+ "loss": 4.191592216491699,
+ "step": 2261
+ },
+ {
+ "epoch": 31.419213973799128,
+ "grad_norm": 0.016557959839701653,
+ "learning_rate": 0.0006,
+ "loss": 4.159705638885498,
+ "step": 2262
+ },
+ {
+ "epoch": 31.433187772925763,
+ "grad_norm": 0.01659569889307022,
+ "learning_rate": 0.0006,
+ "loss": 4.2136664390563965,
+ "step": 2263
+ },
+ {
+ "epoch": 31.4471615720524,
+ "grad_norm": 0.017101937904953957,
+ "learning_rate": 0.0006,
+ "loss": 4.2716498374938965,
+ "step": 2264
+ },
+ {
+ "epoch": 31.46113537117904,
+ "grad_norm": 0.01722985878586769,
+ "learning_rate": 0.0006,
+ "loss": 4.300534248352051,
+ "step": 2265
+ },
+ {
+ "epoch": 31.475109170305675,
+ "grad_norm": 0.016800185665488243,
+ "learning_rate": 0.0006,
+ "loss": 4.113240718841553,
+ "step": 2266
+ },
+ {
+ "epoch": 31.489082969432314,
+ "grad_norm": 0.01603875495493412,
+ "learning_rate": 0.0006,
+ "loss": 4.157105445861816,
+ "step": 2267
+ },
+ {
+ "epoch": 31.503056768558952,
+ "grad_norm": 0.015903932973742485,
+ "learning_rate": 0.0006,
+ "loss": 4.233669757843018,
+ "step": 2268
+ },
+ {
+ "epoch": 31.51703056768559,
+ "grad_norm": 0.0160983856767416,
+ "learning_rate": 0.0006,
+ "loss": 4.147704124450684,
+ "step": 2269
+ },
+ {
+ "epoch": 31.531004366812226,
+ "grad_norm": 0.01604730263352394,
+ "learning_rate": 0.0006,
+ "loss": 4.1465911865234375,
+ "step": 2270
+ },
+ {
+ "epoch": 31.544978165938865,
+ "grad_norm": 0.016205785796046257,
+ "learning_rate": 0.0006,
+ "loss": 4.106704235076904,
+ "step": 2271
+ },
+ {
+ "epoch": 31.558951965065503,
+ "grad_norm": 0.017008036375045776,
+ "learning_rate": 0.0006,
+ "loss": 4.191000938415527,
+ "step": 2272
+ },
+ {
+ "epoch": 31.57292576419214,
+ "grad_norm": 0.01699635200202465,
+ "learning_rate": 0.0006,
+ "loss": 4.187042236328125,
+ "step": 2273
+ },
+ {
+ "epoch": 31.586899563318777,
+ "grad_norm": 0.01692967116832733,
+ "learning_rate": 0.0006,
+ "loss": 4.208805561065674,
+ "step": 2274
+ },
+ {
+ "epoch": 31.600873362445416,
+ "grad_norm": 0.017608756199479103,
+ "learning_rate": 0.0006,
+ "loss": 4.200512886047363,
+ "step": 2275
+ },
+ {
+ "epoch": 31.61484716157205,
+ "grad_norm": 0.016590990126132965,
+ "learning_rate": 0.0006,
+ "loss": 4.194705009460449,
+ "step": 2276
+ },
+ {
+ "epoch": 31.62882096069869,
+ "grad_norm": 0.0170235987752676,
+ "learning_rate": 0.0006,
+ "loss": 4.210326194763184,
+ "step": 2277
+ },
+ {
+ "epoch": 31.64279475982533,
+ "grad_norm": 0.01604801043868065,
+ "learning_rate": 0.0006,
+ "loss": 4.196024417877197,
+ "step": 2278
+ },
+ {
+ "epoch": 31.656768558951963,
+ "grad_norm": 0.015694202855229378,
+ "learning_rate": 0.0006,
+ "loss": 4.178236484527588,
+ "step": 2279
+ },
+ {
+ "epoch": 31.670742358078602,
+ "grad_norm": 0.015473959036171436,
+ "learning_rate": 0.0006,
+ "loss": 4.259339809417725,
+ "step": 2280
+ },
+ {
+ "epoch": 31.68471615720524,
+ "grad_norm": 0.014689362607896328,
+ "learning_rate": 0.0006,
+ "loss": 3.9702324867248535,
+ "step": 2281
+ },
+ {
+ "epoch": 31.69868995633188,
+ "grad_norm": 0.015245123766362667,
+ "learning_rate": 0.0006,
+ "loss": 4.197025775909424,
+ "step": 2282
+ },
+ {
+ "epoch": 31.712663755458514,
+ "grad_norm": 0.016148347407579422,
+ "learning_rate": 0.0006,
+ "loss": 4.231856346130371,
+ "step": 2283
+ },
+ {
+ "epoch": 31.726637554585153,
+ "grad_norm": 0.015169818885624409,
+ "learning_rate": 0.0006,
+ "loss": 4.230169296264648,
+ "step": 2284
+ },
+ {
+ "epoch": 31.74061135371179,
+ "grad_norm": 0.014843013137578964,
+ "learning_rate": 0.0006,
+ "loss": 4.195633411407471,
+ "step": 2285
+ },
+ {
+ "epoch": 31.754585152838427,
+ "grad_norm": 0.015229340642690659,
+ "learning_rate": 0.0006,
+ "loss": 4.231280326843262,
+ "step": 2286
+ },
+ {
+ "epoch": 31.768558951965066,
+ "grad_norm": 0.01551708485931158,
+ "learning_rate": 0.0006,
+ "loss": 4.17064094543457,
+ "step": 2287
+ },
+ {
+ "epoch": 31.782532751091704,
+ "grad_norm": 0.015863575041294098,
+ "learning_rate": 0.0006,
+ "loss": 4.1579484939575195,
+ "step": 2288
+ },
+ {
+ "epoch": 31.79650655021834,
+ "grad_norm": 0.015118198469281197,
+ "learning_rate": 0.0006,
+ "loss": 4.1626787185668945,
+ "step": 2289
+ },
+ {
+ "epoch": 31.810480349344978,
+ "grad_norm": 0.015220776200294495,
+ "learning_rate": 0.0006,
+ "loss": 4.135746955871582,
+ "step": 2290
+ },
+ {
+ "epoch": 31.824454148471617,
+ "grad_norm": 0.014529784210026264,
+ "learning_rate": 0.0006,
+ "loss": 4.164931297302246,
+ "step": 2291
+ },
+ {
+ "epoch": 31.83842794759825,
+ "grad_norm": 0.015533296391367912,
+ "learning_rate": 0.0006,
+ "loss": 4.112464904785156,
+ "step": 2292
+ },
+ {
+ "epoch": 31.85240174672489,
+ "grad_norm": 0.016562335193157196,
+ "learning_rate": 0.0006,
+ "loss": 4.164898872375488,
+ "step": 2293
+ },
+ {
+ "epoch": 31.86637554585153,
+ "grad_norm": 0.016258113086223602,
+ "learning_rate": 0.0006,
+ "loss": 4.1979265213012695,
+ "step": 2294
+ },
+ {
+ "epoch": 31.880349344978168,
+ "grad_norm": 0.01558110024780035,
+ "learning_rate": 0.0006,
+ "loss": 4.174675941467285,
+ "step": 2295
+ },
+ {
+ "epoch": 31.894323144104803,
+ "grad_norm": 0.016412515193223953,
+ "learning_rate": 0.0006,
+ "loss": 4.217909336090088,
+ "step": 2296
+ },
+ {
+ "epoch": 31.90829694323144,
+ "grad_norm": 0.015118442475795746,
+ "learning_rate": 0.0006,
+ "loss": 4.137143611907959,
+ "step": 2297
+ },
+ {
+ "epoch": 31.92227074235808,
+ "grad_norm": 0.01506161317229271,
+ "learning_rate": 0.0006,
+ "loss": 4.196122169494629,
+ "step": 2298
+ },
+ {
+ "epoch": 31.936244541484715,
+ "grad_norm": 0.017165271565318108,
+ "learning_rate": 0.0006,
+ "loss": 4.14834451675415,
+ "step": 2299
+ },
+ {
+ "epoch": 31.950218340611354,
+ "grad_norm": 0.01696440577507019,
+ "learning_rate": 0.0006,
+ "loss": 4.218868732452393,
+ "step": 2300
+ },
+ {
+ "epoch": 31.964192139737992,
+ "grad_norm": 0.017349394038319588,
+ "learning_rate": 0.0006,
+ "loss": 4.2754011154174805,
+ "step": 2301
+ },
+ {
+ "epoch": 31.978165938864628,
+ "grad_norm": 0.018387921154499054,
+ "learning_rate": 0.0006,
+ "loss": 4.282623291015625,
+ "step": 2302
+ },
+ {
+ "epoch": 31.992139737991266,
+ "grad_norm": 0.018837926909327507,
+ "learning_rate": 0.0006,
+ "loss": 4.185425281524658,
+ "step": 2303
+ },
+ {
+ "epoch": 32.0,
+ "grad_norm": 0.018998464569449425,
+ "learning_rate": 0.0006,
+ "loss": 4.151291847229004,
+ "step": 2304
+ },
+ {
+ "epoch": 32.0,
+ "eval_loss": 4.522614002227783,
+ "eval_runtime": 57.0106,
+ "eval_samples_per_second": 42.834,
+ "eval_steps_per_second": 1.351,
+ "step": 2304
+ },
+ {
+ "epoch": 32.01397379912664,
+ "grad_norm": 0.019036833196878433,
+ "learning_rate": 0.0006,
+ "loss": 4.183202266693115,
+ "step": 2305
+ },
+ {
+ "epoch": 32.02794759825328,
+ "grad_norm": 0.02207397110760212,
+ "learning_rate": 0.0006,
+ "loss": 4.17875862121582,
+ "step": 2306
+ },
+ {
+ "epoch": 32.041921397379916,
+ "grad_norm": 0.022008279338479042,
+ "learning_rate": 0.0006,
+ "loss": 4.08042049407959,
+ "step": 2307
+ },
+ {
+ "epoch": 32.05589519650655,
+ "grad_norm": 0.022407017648220062,
+ "learning_rate": 0.0006,
+ "loss": 4.112189769744873,
+ "step": 2308
+ },
+ {
+ "epoch": 32.069868995633186,
+ "grad_norm": 0.021599747240543365,
+ "learning_rate": 0.0006,
+ "loss": 4.126455307006836,
+ "step": 2309
+ },
+ {
+ "epoch": 32.083842794759825,
+ "grad_norm": 0.020340966060757637,
+ "learning_rate": 0.0006,
+ "loss": 4.080538272857666,
+ "step": 2310
+ },
+ {
+ "epoch": 32.09781659388646,
+ "grad_norm": 0.020956892520189285,
+ "learning_rate": 0.0006,
+ "loss": 4.171943664550781,
+ "step": 2311
+ },
+ {
+ "epoch": 32.1117903930131,
+ "grad_norm": 0.019966667518019676,
+ "learning_rate": 0.0006,
+ "loss": 4.144007682800293,
+ "step": 2312
+ },
+ {
+ "epoch": 32.12576419213974,
+ "grad_norm": 0.018256952986121178,
+ "learning_rate": 0.0006,
+ "loss": 4.208107948303223,
+ "step": 2313
+ },
+ {
+ "epoch": 32.13973799126637,
+ "grad_norm": 0.017289170995354652,
+ "learning_rate": 0.0006,
+ "loss": 4.182024955749512,
+ "step": 2314
+ },
+ {
+ "epoch": 32.15371179039301,
+ "grad_norm": 0.017656700685620308,
+ "learning_rate": 0.0006,
+ "loss": 4.068373680114746,
+ "step": 2315
+ },
+ {
+ "epoch": 32.16768558951965,
+ "grad_norm": 0.017903361469507217,
+ "learning_rate": 0.0006,
+ "loss": 4.2345428466796875,
+ "step": 2316
+ },
+ {
+ "epoch": 32.18165938864629,
+ "grad_norm": 0.01785109005868435,
+ "learning_rate": 0.0006,
+ "loss": 4.219776153564453,
+ "step": 2317
+ },
+ {
+ "epoch": 32.19563318777293,
+ "grad_norm": 0.01887257769703865,
+ "learning_rate": 0.0006,
+ "loss": 4.139206409454346,
+ "step": 2318
+ },
+ {
+ "epoch": 32.209606986899566,
+ "grad_norm": 0.01924685761332512,
+ "learning_rate": 0.0006,
+ "loss": 4.121052265167236,
+ "step": 2319
+ },
+ {
+ "epoch": 32.223580786026204,
+ "grad_norm": 0.020574072375893593,
+ "learning_rate": 0.0006,
+ "loss": 4.13262414932251,
+ "step": 2320
+ },
+ {
+ "epoch": 32.237554585152836,
+ "grad_norm": 0.023849064484238625,
+ "learning_rate": 0.0006,
+ "loss": 4.239119529724121,
+ "step": 2321
+ },
+ {
+ "epoch": 32.251528384279474,
+ "grad_norm": 0.026239361613988876,
+ "learning_rate": 0.0006,
+ "loss": 4.248993873596191,
+ "step": 2322
+ },
+ {
+ "epoch": 32.26550218340611,
+ "grad_norm": 0.023980243131518364,
+ "learning_rate": 0.0006,
+ "loss": 4.159852504730225,
+ "step": 2323
+ },
+ {
+ "epoch": 32.27947598253275,
+ "grad_norm": 0.020442405715584755,
+ "learning_rate": 0.0006,
+ "loss": 4.133546829223633,
+ "step": 2324
+ },
+ {
+ "epoch": 32.29344978165939,
+ "grad_norm": 0.020661186426877975,
+ "learning_rate": 0.0006,
+ "loss": 4.078243732452393,
+ "step": 2325
+ },
+ {
+ "epoch": 32.30742358078603,
+ "grad_norm": 0.02046387456357479,
+ "learning_rate": 0.0006,
+ "loss": 4.172012805938721,
+ "step": 2326
+ },
+ {
+ "epoch": 32.32139737991266,
+ "grad_norm": 0.02089597098529339,
+ "learning_rate": 0.0006,
+ "loss": 4.0604705810546875,
+ "step": 2327
+ },
+ {
+ "epoch": 32.3353711790393,
+ "grad_norm": 0.01875591278076172,
+ "learning_rate": 0.0006,
+ "loss": 4.125911235809326,
+ "step": 2328
+ },
+ {
+ "epoch": 32.34934497816594,
+ "grad_norm": 0.018606796860694885,
+ "learning_rate": 0.0006,
+ "loss": 4.097362518310547,
+ "step": 2329
+ },
+ {
+ "epoch": 32.36331877729258,
+ "grad_norm": 0.01766626164317131,
+ "learning_rate": 0.0006,
+ "loss": 4.118616580963135,
+ "step": 2330
+ },
+ {
+ "epoch": 32.377292576419215,
+ "grad_norm": 0.017692960798740387,
+ "learning_rate": 0.0006,
+ "loss": 4.1457672119140625,
+ "step": 2331
+ },
+ {
+ "epoch": 32.391266375545854,
+ "grad_norm": 0.017619704827666283,
+ "learning_rate": 0.0006,
+ "loss": 4.175961494445801,
+ "step": 2332
+ },
+ {
+ "epoch": 32.40524017467249,
+ "grad_norm": 0.01784413494169712,
+ "learning_rate": 0.0006,
+ "loss": 4.051291465759277,
+ "step": 2333
+ },
+ {
+ "epoch": 32.419213973799124,
+ "grad_norm": 0.018508171662688255,
+ "learning_rate": 0.0006,
+ "loss": 4.0638556480407715,
+ "step": 2334
+ },
+ {
+ "epoch": 32.43318777292576,
+ "grad_norm": 0.01731436885893345,
+ "learning_rate": 0.0006,
+ "loss": 4.1805419921875,
+ "step": 2335
+ },
+ {
+ "epoch": 32.4471615720524,
+ "grad_norm": 0.015886498615145683,
+ "learning_rate": 0.0006,
+ "loss": 4.151026725769043,
+ "step": 2336
+ },
+ {
+ "epoch": 32.46113537117904,
+ "grad_norm": 0.016583820804953575,
+ "learning_rate": 0.0006,
+ "loss": 4.203604698181152,
+ "step": 2337
+ },
+ {
+ "epoch": 32.47510917030568,
+ "grad_norm": 0.0176097322255373,
+ "learning_rate": 0.0006,
+ "loss": 4.135159015655518,
+ "step": 2338
+ },
+ {
+ "epoch": 32.48908296943232,
+ "grad_norm": 0.016657903790473938,
+ "learning_rate": 0.0006,
+ "loss": 4.126681804656982,
+ "step": 2339
+ },
+ {
+ "epoch": 32.50305676855895,
+ "grad_norm": 0.016866130754351616,
+ "learning_rate": 0.0006,
+ "loss": 4.162341117858887,
+ "step": 2340
+ },
+ {
+ "epoch": 32.51703056768559,
+ "grad_norm": 0.016286680474877357,
+ "learning_rate": 0.0006,
+ "loss": 4.261801242828369,
+ "step": 2341
+ },
+ {
+ "epoch": 32.531004366812226,
+ "grad_norm": 0.01481365505605936,
+ "learning_rate": 0.0006,
+ "loss": 4.110016822814941,
+ "step": 2342
+ },
+ {
+ "epoch": 32.544978165938865,
+ "grad_norm": 0.016605013981461525,
+ "learning_rate": 0.0006,
+ "loss": 4.230024814605713,
+ "step": 2343
+ },
+ {
+ "epoch": 32.5589519650655,
+ "grad_norm": 0.01645175740122795,
+ "learning_rate": 0.0006,
+ "loss": 4.174612045288086,
+ "step": 2344
+ },
+ {
+ "epoch": 32.57292576419214,
+ "grad_norm": 0.014725026674568653,
+ "learning_rate": 0.0006,
+ "loss": 4.066371440887451,
+ "step": 2345
+ },
+ {
+ "epoch": 32.58689956331878,
+ "grad_norm": 0.015003837645053864,
+ "learning_rate": 0.0006,
+ "loss": 4.193343639373779,
+ "step": 2346
+ },
+ {
+ "epoch": 32.60087336244541,
+ "grad_norm": 0.015346302650868893,
+ "learning_rate": 0.0006,
+ "loss": 4.210659980773926,
+ "step": 2347
+ },
+ {
+ "epoch": 32.61484716157205,
+ "grad_norm": 0.015135962516069412,
+ "learning_rate": 0.0006,
+ "loss": 4.204618453979492,
+ "step": 2348
+ },
+ {
+ "epoch": 32.62882096069869,
+ "grad_norm": 0.01601956970989704,
+ "learning_rate": 0.0006,
+ "loss": 4.145564556121826,
+ "step": 2349
+ },
+ {
+ "epoch": 32.64279475982533,
+ "grad_norm": 0.01517851185053587,
+ "learning_rate": 0.0006,
+ "loss": 4.051673889160156,
+ "step": 2350
+ },
+ {
+ "epoch": 32.65676855895197,
+ "grad_norm": 0.014894292689859867,
+ "learning_rate": 0.0006,
+ "loss": 4.184818267822266,
+ "step": 2351
+ },
+ {
+ "epoch": 32.670742358078606,
+ "grad_norm": 0.015720834955573082,
+ "learning_rate": 0.0006,
+ "loss": 4.220036506652832,
+ "step": 2352
+ },
+ {
+ "epoch": 32.68471615720524,
+ "grad_norm": 0.01573433168232441,
+ "learning_rate": 0.0006,
+ "loss": 4.076773166656494,
+ "step": 2353
+ },
+ {
+ "epoch": 32.698689956331876,
+ "grad_norm": 0.015828052535653114,
+ "learning_rate": 0.0006,
+ "loss": 4.116329669952393,
+ "step": 2354
+ },
+ {
+ "epoch": 32.712663755458514,
+ "grad_norm": 0.015462259761989117,
+ "learning_rate": 0.0006,
+ "loss": 4.1155242919921875,
+ "step": 2355
+ },
+ {
+ "epoch": 32.72663755458515,
+ "grad_norm": 0.015763849020004272,
+ "learning_rate": 0.0006,
+ "loss": 4.169687747955322,
+ "step": 2356
+ },
+ {
+ "epoch": 32.74061135371179,
+ "grad_norm": 0.014935145154595375,
+ "learning_rate": 0.0006,
+ "loss": 4.2090911865234375,
+ "step": 2357
+ },
+ {
+ "epoch": 32.75458515283843,
+ "grad_norm": 0.015119184739887714,
+ "learning_rate": 0.0006,
+ "loss": 4.131838798522949,
+ "step": 2358
+ },
+ {
+ "epoch": 32.76855895196506,
+ "grad_norm": 0.014833525754511356,
+ "learning_rate": 0.0006,
+ "loss": 4.17618465423584,
+ "step": 2359
+ },
+ {
+ "epoch": 32.7825327510917,
+ "grad_norm": 0.01588359661400318,
+ "learning_rate": 0.0006,
+ "loss": 4.163723945617676,
+ "step": 2360
+ },
+ {
+ "epoch": 32.79650655021834,
+ "grad_norm": 0.0168188214302063,
+ "learning_rate": 0.0006,
+ "loss": 4.178189277648926,
+ "step": 2361
+ },
+ {
+ "epoch": 32.81048034934498,
+ "grad_norm": 0.01726095750927925,
+ "learning_rate": 0.0006,
+ "loss": 4.144223690032959,
+ "step": 2362
+ },
+ {
+ "epoch": 32.82445414847162,
+ "grad_norm": 0.01701410301029682,
+ "learning_rate": 0.0006,
+ "loss": 4.209896087646484,
+ "step": 2363
+ },
+ {
+ "epoch": 32.838427947598255,
+ "grad_norm": 0.01682264916598797,
+ "learning_rate": 0.0006,
+ "loss": 4.26223087310791,
+ "step": 2364
+ },
+ {
+ "epoch": 32.852401746724894,
+ "grad_norm": 0.015597456134855747,
+ "learning_rate": 0.0006,
+ "loss": 4.292664051055908,
+ "step": 2365
+ },
+ {
+ "epoch": 32.866375545851525,
+ "grad_norm": 0.015258579514920712,
+ "learning_rate": 0.0006,
+ "loss": 4.165006637573242,
+ "step": 2366
+ },
+ {
+ "epoch": 32.880349344978164,
+ "grad_norm": 0.01662200503051281,
+ "learning_rate": 0.0006,
+ "loss": 4.309786319732666,
+ "step": 2367
+ },
+ {
+ "epoch": 32.8943231441048,
+ "grad_norm": 0.015691161155700684,
+ "learning_rate": 0.0006,
+ "loss": 4.167202949523926,
+ "step": 2368
+ },
+ {
+ "epoch": 32.90829694323144,
+ "grad_norm": 0.014870637096464634,
+ "learning_rate": 0.0006,
+ "loss": 4.142415523529053,
+ "step": 2369
+ },
+ {
+ "epoch": 32.92227074235808,
+ "grad_norm": 0.01567252166569233,
+ "learning_rate": 0.0006,
+ "loss": 4.2312092781066895,
+ "step": 2370
+ },
+ {
+ "epoch": 32.93624454148472,
+ "grad_norm": 0.014948616735637188,
+ "learning_rate": 0.0006,
+ "loss": 4.1514129638671875,
+ "step": 2371
+ },
+ {
+ "epoch": 32.95021834061135,
+ "grad_norm": 0.014654500409960747,
+ "learning_rate": 0.0006,
+ "loss": 4.205010890960693,
+ "step": 2372
+ },
+ {
+ "epoch": 32.96419213973799,
+ "grad_norm": 0.015059148892760277,
+ "learning_rate": 0.0006,
+ "loss": 4.148405075073242,
+ "step": 2373
+ },
+ {
+ "epoch": 32.97816593886463,
+ "grad_norm": 0.015026576817035675,
+ "learning_rate": 0.0006,
+ "loss": 4.182535171508789,
+ "step": 2374
+ },
+ {
+ "epoch": 32.992139737991266,
+ "grad_norm": 0.016394320875406265,
+ "learning_rate": 0.0006,
+ "loss": 4.129051208496094,
+ "step": 2375
+ },
+ {
+ "epoch": 33.0,
+ "grad_norm": 0.01867043413221836,
+ "learning_rate": 0.0006,
+ "loss": 4.185556411743164,
+ "step": 2376
+ },
+ {
+ "epoch": 33.0,
+ "eval_loss": 4.640405654907227,
+ "eval_runtime": 56.5861,
+ "eval_samples_per_second": 43.155,
+ "eval_steps_per_second": 1.361,
+ "step": 2376
+ },
+ {
+ "epoch": 33.01397379912664,
+ "grad_norm": 0.018626727163791656,
+ "learning_rate": 0.0006,
+ "loss": 4.143031597137451,
+ "step": 2377
+ },
+ {
+ "epoch": 33.02794759825328,
+ "grad_norm": 0.019283363595604897,
+ "learning_rate": 0.0006,
+ "loss": 4.117845058441162,
+ "step": 2378
+ },
+ {
+ "epoch": 33.041921397379916,
+ "grad_norm": 0.018185822293162346,
+ "learning_rate": 0.0006,
+ "loss": 4.076862812042236,
+ "step": 2379
+ },
+ {
+ "epoch": 33.05589519650655,
+ "grad_norm": 0.019411196932196617,
+ "learning_rate": 0.0006,
+ "loss": 4.124156475067139,
+ "step": 2380
+ },
+ {
+ "epoch": 33.069868995633186,
+ "grad_norm": 0.019248466938734055,
+ "learning_rate": 0.0006,
+ "loss": 4.201735019683838,
+ "step": 2381
+ },
+ {
+ "epoch": 33.083842794759825,
+ "grad_norm": 0.018856720998883247,
+ "learning_rate": 0.0006,
+ "loss": 4.00296688079834,
+ "step": 2382
+ },
+ {
+ "epoch": 33.09781659388646,
+ "grad_norm": 0.018554294481873512,
+ "learning_rate": 0.0006,
+ "loss": 4.270577430725098,
+ "step": 2383
+ },
+ {
+ "epoch": 33.1117903930131,
+ "grad_norm": 0.019946224987506866,
+ "learning_rate": 0.0006,
+ "loss": 4.119589328765869,
+ "step": 2384
+ },
+ {
+ "epoch": 33.12576419213974,
+ "grad_norm": 0.021395670250058174,
+ "learning_rate": 0.0006,
+ "loss": 4.128847122192383,
+ "step": 2385
+ },
+ {
+ "epoch": 33.13973799126637,
+ "grad_norm": 0.02009674906730652,
+ "learning_rate": 0.0006,
+ "loss": 4.183228969573975,
+ "step": 2386
+ },
+ {
+ "epoch": 33.15371179039301,
+ "grad_norm": 0.01820223033428192,
+ "learning_rate": 0.0006,
+ "loss": 4.131601810455322,
+ "step": 2387
+ },
+ {
+ "epoch": 33.16768558951965,
+ "grad_norm": 0.018653156235814095,
+ "learning_rate": 0.0006,
+ "loss": 4.116430282592773,
+ "step": 2388
+ },
+ {
+ "epoch": 33.18165938864629,
+ "grad_norm": 0.016828063875436783,
+ "learning_rate": 0.0006,
+ "loss": 4.054663181304932,
+ "step": 2389
+ },
+ {
+ "epoch": 33.19563318777293,
+ "grad_norm": 0.017025409266352654,
+ "learning_rate": 0.0006,
+ "loss": 4.110044956207275,
+ "step": 2390
+ },
+ {
+ "epoch": 33.209606986899566,
+ "grad_norm": 0.016608452424407005,
+ "learning_rate": 0.0006,
+ "loss": 4.102668285369873,
+ "step": 2391
+ },
+ {
+ "epoch": 33.223580786026204,
+ "grad_norm": 0.01743577979505062,
+ "learning_rate": 0.0006,
+ "loss": 4.166791915893555,
+ "step": 2392
+ },
+ {
+ "epoch": 33.237554585152836,
+ "grad_norm": 0.017253734171390533,
+ "learning_rate": 0.0006,
+ "loss": 4.217832088470459,
+ "step": 2393
+ },
+ {
+ "epoch": 33.251528384279474,
+ "grad_norm": 0.017937595024704933,
+ "learning_rate": 0.0006,
+ "loss": 4.2173967361450195,
+ "step": 2394
+ },
+ {
+ "epoch": 33.26550218340611,
+ "grad_norm": 0.01985993981361389,
+ "learning_rate": 0.0006,
+ "loss": 4.17849588394165,
+ "step": 2395
+ },
+ {
+ "epoch": 33.27947598253275,
+ "grad_norm": 0.019243355840444565,
+ "learning_rate": 0.0006,
+ "loss": 4.086523056030273,
+ "step": 2396
+ },
+ {
+ "epoch": 33.29344978165939,
+ "grad_norm": 0.01924416422843933,
+ "learning_rate": 0.0006,
+ "loss": 4.128574371337891,
+ "step": 2397
+ },
+ {
+ "epoch": 33.30742358078603,
+ "grad_norm": 0.01983758620917797,
+ "learning_rate": 0.0006,
+ "loss": 4.061581611633301,
+ "step": 2398
+ },
+ {
+ "epoch": 33.32139737991266,
+ "grad_norm": 0.01930462196469307,
+ "learning_rate": 0.0006,
+ "loss": 4.091731071472168,
+ "step": 2399
+ },
+ {
+ "epoch": 33.3353711790393,
+ "grad_norm": 0.017988424748182297,
+ "learning_rate": 0.0006,
+ "loss": 4.155599117279053,
+ "step": 2400
+ },
+ {
+ "epoch": 33.34934497816594,
+ "grad_norm": 0.015194767154753208,
+ "learning_rate": 0.0006,
+ "loss": 4.144134044647217,
+ "step": 2401
+ },
+ {
+ "epoch": 33.36331877729258,
+ "grad_norm": 0.01536081824451685,
+ "learning_rate": 0.0006,
+ "loss": 4.127272129058838,
+ "step": 2402
+ },
+ {
+ "epoch": 33.377292576419215,
+ "grad_norm": 0.015431297942996025,
+ "learning_rate": 0.0006,
+ "loss": 4.3213019371032715,
+ "step": 2403
+ },
+ {
+ "epoch": 33.391266375545854,
+ "grad_norm": 0.015740280970931053,
+ "learning_rate": 0.0006,
+ "loss": 4.121467590332031,
+ "step": 2404
+ },
+ {
+ "epoch": 33.40524017467249,
+ "grad_norm": 0.015379384160041809,
+ "learning_rate": 0.0006,
+ "loss": 4.095216751098633,
+ "step": 2405
+ },
+ {
+ "epoch": 33.419213973799124,
+ "grad_norm": 0.01528073474764824,
+ "learning_rate": 0.0006,
+ "loss": 4.13047981262207,
+ "step": 2406
+ },
+ {
+ "epoch": 33.43318777292576,
+ "grad_norm": 0.015739979222416878,
+ "learning_rate": 0.0006,
+ "loss": 4.2533135414123535,
+ "step": 2407
+ },
+ {
+ "epoch": 33.4471615720524,
+ "grad_norm": 0.01519764307886362,
+ "learning_rate": 0.0006,
+ "loss": 4.131782054901123,
+ "step": 2408
+ },
+ {
+ "epoch": 33.46113537117904,
+ "grad_norm": 0.015425996854901314,
+ "learning_rate": 0.0006,
+ "loss": 4.02811336517334,
+ "step": 2409
+ },
+ {
+ "epoch": 33.47510917030568,
+ "grad_norm": 0.015378830023109913,
+ "learning_rate": 0.0006,
+ "loss": 4.08652925491333,
+ "step": 2410
+ },
+ {
+ "epoch": 33.48908296943232,
+ "grad_norm": 0.01473075058311224,
+ "learning_rate": 0.0006,
+ "loss": 4.214364528656006,
+ "step": 2411
+ },
+ {
+ "epoch": 33.50305676855895,
+ "grad_norm": 0.015141637064516544,
+ "learning_rate": 0.0006,
+ "loss": 4.155712127685547,
+ "step": 2412
+ },
+ {
+ "epoch": 33.51703056768559,
+ "grad_norm": 0.01641593500971794,
+ "learning_rate": 0.0006,
+ "loss": 4.097907066345215,
+ "step": 2413
+ },
+ {
+ "epoch": 33.531004366812226,
+ "grad_norm": 0.015182029455900192,
+ "learning_rate": 0.0006,
+ "loss": 4.238499641418457,
+ "step": 2414
+ },
+ {
+ "epoch": 33.544978165938865,
+ "grad_norm": 0.01643732376396656,
+ "learning_rate": 0.0006,
+ "loss": 4.0661492347717285,
+ "step": 2415
+ },
+ {
+ "epoch": 33.5589519650655,
+ "grad_norm": 0.015614228323101997,
+ "learning_rate": 0.0006,
+ "loss": 4.120795249938965,
+ "step": 2416
+ },
+ {
+ "epoch": 33.57292576419214,
+ "grad_norm": 0.016066187992691994,
+ "learning_rate": 0.0006,
+ "loss": 4.096465110778809,
+ "step": 2417
+ },
+ {
+ "epoch": 33.58689956331878,
+ "grad_norm": 0.017214342951774597,
+ "learning_rate": 0.0006,
+ "loss": 4.242628574371338,
+ "step": 2418
+ },
+ {
+ "epoch": 33.60087336244541,
+ "grad_norm": 0.01787586510181427,
+ "learning_rate": 0.0006,
+ "loss": 4.149589538574219,
+ "step": 2419
+ },
+ {
+ "epoch": 33.61484716157205,
+ "grad_norm": 0.019455241039395332,
+ "learning_rate": 0.0006,
+ "loss": 4.267796039581299,
+ "step": 2420
+ },
+ {
+ "epoch": 33.62882096069869,
+ "grad_norm": 0.01984674669802189,
+ "learning_rate": 0.0006,
+ "loss": 4.175220966339111,
+ "step": 2421
+ },
+ {
+ "epoch": 33.64279475982533,
+ "grad_norm": 0.021933717653155327,
+ "learning_rate": 0.0006,
+ "loss": 4.106647491455078,
+ "step": 2422
+ },
+ {
+ "epoch": 33.65676855895197,
+ "grad_norm": 0.02123589813709259,
+ "learning_rate": 0.0006,
+ "loss": 4.204996109008789,
+ "step": 2423
+ },
+ {
+ "epoch": 33.670742358078606,
+ "grad_norm": 0.019814589992165565,
+ "learning_rate": 0.0006,
+ "loss": 4.186171531677246,
+ "step": 2424
+ },
+ {
+ "epoch": 33.68471615720524,
+ "grad_norm": 0.02059728465974331,
+ "learning_rate": 0.0006,
+ "loss": 4.170509338378906,
+ "step": 2425
+ },
+ {
+ "epoch": 33.698689956331876,
+ "grad_norm": 0.018483608961105347,
+ "learning_rate": 0.0006,
+ "loss": 4.085585594177246,
+ "step": 2426
+ },
+ {
+ "epoch": 33.712663755458514,
+ "grad_norm": 0.018403852358460426,
+ "learning_rate": 0.0006,
+ "loss": 4.2551774978637695,
+ "step": 2427
+ },
+ {
+ "epoch": 33.72663755458515,
+ "grad_norm": 0.018653830513358116,
+ "learning_rate": 0.0006,
+ "loss": 4.091766357421875,
+ "step": 2428
+ },
+ {
+ "epoch": 33.74061135371179,
+ "grad_norm": 0.017692334949970245,
+ "learning_rate": 0.0006,
+ "loss": 4.100838661193848,
+ "step": 2429
+ },
+ {
+ "epoch": 33.75458515283843,
+ "grad_norm": 0.016029126942157745,
+ "learning_rate": 0.0006,
+ "loss": 4.1530680656433105,
+ "step": 2430
+ },
+ {
+ "epoch": 33.76855895196506,
+ "grad_norm": 0.016095684841275215,
+ "learning_rate": 0.0006,
+ "loss": 4.20544958114624,
+ "step": 2431
+ },
+ {
+ "epoch": 33.7825327510917,
+ "grad_norm": 0.016375798732042313,
+ "learning_rate": 0.0006,
+ "loss": 4.199177265167236,
+ "step": 2432
+ },
+ {
+ "epoch": 33.79650655021834,
+ "grad_norm": 0.01590786688029766,
+ "learning_rate": 0.0006,
+ "loss": 4.200799465179443,
+ "step": 2433
+ },
+ {
+ "epoch": 33.81048034934498,
+ "grad_norm": 0.014896097593009472,
+ "learning_rate": 0.0006,
+ "loss": 4.185850143432617,
+ "step": 2434
+ },
+ {
+ "epoch": 33.82445414847162,
+ "grad_norm": 0.01438513956964016,
+ "learning_rate": 0.0006,
+ "loss": 4.214962005615234,
+ "step": 2435
+ },
+ {
+ "epoch": 33.838427947598255,
+ "grad_norm": 0.014459339901804924,
+ "learning_rate": 0.0006,
+ "loss": 4.196972370147705,
+ "step": 2436
+ },
+ {
+ "epoch": 33.852401746724894,
+ "grad_norm": 0.014782169833779335,
+ "learning_rate": 0.0006,
+ "loss": 4.0885910987854,
+ "step": 2437
+ },
+ {
+ "epoch": 33.866375545851525,
+ "grad_norm": 0.01520370040088892,
+ "learning_rate": 0.0006,
+ "loss": 4.0323262214660645,
+ "step": 2438
+ },
+ {
+ "epoch": 33.880349344978164,
+ "grad_norm": 0.014056282117962837,
+ "learning_rate": 0.0006,
+ "loss": 4.1386518478393555,
+ "step": 2439
+ },
+ {
+ "epoch": 33.8943231441048,
+ "grad_norm": 0.013567990623414516,
+ "learning_rate": 0.0006,
+ "loss": 4.095754623413086,
+ "step": 2440
+ },
+ {
+ "epoch": 33.90829694323144,
+ "grad_norm": 0.013306716457009315,
+ "learning_rate": 0.0006,
+ "loss": 4.1623077392578125,
+ "step": 2441
+ },
+ {
+ "epoch": 33.92227074235808,
+ "grad_norm": 0.014047150500118732,
+ "learning_rate": 0.0006,
+ "loss": 4.142541885375977,
+ "step": 2442
+ },
+ {
+ "epoch": 33.93624454148472,
+ "grad_norm": 0.01377617847174406,
+ "learning_rate": 0.0006,
+ "loss": 4.268551826477051,
+ "step": 2443
+ },
+ {
+ "epoch": 33.95021834061135,
+ "grad_norm": 0.014671621844172478,
+ "learning_rate": 0.0006,
+ "loss": 4.198196887969971,
+ "step": 2444
+ },
+ {
+ "epoch": 33.96419213973799,
+ "grad_norm": 0.014806744642555714,
+ "learning_rate": 0.0006,
+ "loss": 4.057899475097656,
+ "step": 2445
+ },
+ {
+ "epoch": 33.97816593886463,
+ "grad_norm": 0.015734272077679634,
+ "learning_rate": 0.0006,
+ "loss": 4.163505554199219,
+ "step": 2446
+ },
+ {
+ "epoch": 33.992139737991266,
+ "grad_norm": 0.015250146389007568,
+ "learning_rate": 0.0006,
+ "loss": 4.076294898986816,
+ "step": 2447
+ },
+ {
+ "epoch": 34.0,
+ "grad_norm": 0.016733702272176743,
+ "learning_rate": 0.0006,
+ "loss": 4.138420104980469,
+ "step": 2448
+ },
+ {
+ "epoch": 34.0,
+ "eval_loss": 4.542503356933594,
+ "eval_runtime": 56.983,
+ "eval_samples_per_second": 42.855,
+ "eval_steps_per_second": 1.351,
+ "step": 2448
+ },
+ {
+ "epoch": 34.01397379912664,
+ "grad_norm": 0.015866613015532494,
+ "learning_rate": 0.0006,
+ "loss": 4.032081127166748,
+ "step": 2449
+ },
+ {
+ "epoch": 34.02794759825328,
+ "grad_norm": 0.01946570910513401,
+ "learning_rate": 0.0006,
+ "loss": 4.122320175170898,
+ "step": 2450
+ },
+ {
+ "epoch": 34.041921397379916,
+ "grad_norm": 0.020265938714146614,
+ "learning_rate": 0.0006,
+ "loss": 4.06460428237915,
+ "step": 2451
+ },
+ {
+ "epoch": 34.05589519650655,
+ "grad_norm": 0.018887171521782875,
+ "learning_rate": 0.0006,
+ "loss": 4.096906661987305,
+ "step": 2452
+ },
+ {
+ "epoch": 34.069868995633186,
+ "grad_norm": 0.017304692417383194,
+ "learning_rate": 0.0006,
+ "loss": 4.137261390686035,
+ "step": 2453
+ },
+ {
+ "epoch": 34.083842794759825,
+ "grad_norm": 0.017029082402586937,
+ "learning_rate": 0.0006,
+ "loss": 4.081112861633301,
+ "step": 2454
+ },
+ {
+ "epoch": 34.09781659388646,
+ "grad_norm": 0.016414813697338104,
+ "learning_rate": 0.0006,
+ "loss": 4.140783309936523,
+ "step": 2455
+ },
+ {
+ "epoch": 34.1117903930131,
+ "grad_norm": 0.018238281831145287,
+ "learning_rate": 0.0006,
+ "loss": 4.161116600036621,
+ "step": 2456
+ },
+ {
+ "epoch": 34.12576419213974,
+ "grad_norm": 0.019929232075810432,
+ "learning_rate": 0.0006,
+ "loss": 4.243722438812256,
+ "step": 2457
+ },
+ {
+ "epoch": 34.13973799126637,
+ "grad_norm": 0.019375786185264587,
+ "learning_rate": 0.0006,
+ "loss": 4.134873390197754,
+ "step": 2458
+ },
+ {
+ "epoch": 34.15371179039301,
+ "grad_norm": 0.018151137977838516,
+ "learning_rate": 0.0006,
+ "loss": 4.098745346069336,
+ "step": 2459
+ },
+ {
+ "epoch": 34.16768558951965,
+ "grad_norm": 0.017054934054613113,
+ "learning_rate": 0.0006,
+ "loss": 4.194662570953369,
+ "step": 2460
+ },
+ {
+ "epoch": 34.18165938864629,
+ "grad_norm": 0.018220555037260056,
+ "learning_rate": 0.0006,
+ "loss": 4.128866195678711,
+ "step": 2461
+ },
+ {
+ "epoch": 34.19563318777293,
+ "grad_norm": 0.01692879945039749,
+ "learning_rate": 0.0006,
+ "loss": 4.062477111816406,
+ "step": 2462
+ },
+ {
+ "epoch": 34.209606986899566,
+ "grad_norm": 0.01674199476838112,
+ "learning_rate": 0.0006,
+ "loss": 4.186149597167969,
+ "step": 2463
+ },
+ {
+ "epoch": 34.223580786026204,
+ "grad_norm": 0.016527455300092697,
+ "learning_rate": 0.0006,
+ "loss": 4.0734028816223145,
+ "step": 2464
+ },
+ {
+ "epoch": 34.237554585152836,
+ "grad_norm": 0.016448386013507843,
+ "learning_rate": 0.0006,
+ "loss": 4.098896026611328,
+ "step": 2465
+ },
+ {
+ "epoch": 34.251528384279474,
+ "grad_norm": 0.014644953422248363,
+ "learning_rate": 0.0006,
+ "loss": 4.1031904220581055,
+ "step": 2466
+ },
+ {
+ "epoch": 34.26550218340611,
+ "grad_norm": 0.016649341210722923,
+ "learning_rate": 0.0006,
+ "loss": 4.148791313171387,
+ "step": 2467
+ },
+ {
+ "epoch": 34.27947598253275,
+ "grad_norm": 0.017442714422941208,
+ "learning_rate": 0.0006,
+ "loss": 4.070799827575684,
+ "step": 2468
+ },
+ {
+ "epoch": 34.29344978165939,
+ "grad_norm": 0.017852013930678368,
+ "learning_rate": 0.0006,
+ "loss": 4.031924247741699,
+ "step": 2469
+ },
+ {
+ "epoch": 34.30742358078603,
+ "grad_norm": 0.016371270641684532,
+ "learning_rate": 0.0006,
+ "loss": 4.149420738220215,
+ "step": 2470
+ },
+ {
+ "epoch": 34.32139737991266,
+ "grad_norm": 0.015846814960241318,
+ "learning_rate": 0.0006,
+ "loss": 4.15427303314209,
+ "step": 2471
+ },
+ {
+ "epoch": 34.3353711790393,
+ "grad_norm": 0.017411163076758385,
+ "learning_rate": 0.0006,
+ "loss": 4.1304192543029785,
+ "step": 2472
+ },
+ {
+ "epoch": 34.34934497816594,
+ "grad_norm": 0.01780046708881855,
+ "learning_rate": 0.0006,
+ "loss": 4.191732406616211,
+ "step": 2473
+ },
+ {
+ "epoch": 34.36331877729258,
+ "grad_norm": 0.01958383060991764,
+ "learning_rate": 0.0006,
+ "loss": 4.068355083465576,
+ "step": 2474
+ },
+ {
+ "epoch": 34.377292576419215,
+ "grad_norm": 0.022173317149281502,
+ "learning_rate": 0.0006,
+ "loss": 4.182929039001465,
+ "step": 2475
+ },
+ {
+ "epoch": 34.391266375545854,
+ "grad_norm": 0.022389156743884087,
+ "learning_rate": 0.0006,
+ "loss": 4.110218048095703,
+ "step": 2476
+ },
+ {
+ "epoch": 34.40524017467249,
+ "grad_norm": 0.022478511556982994,
+ "learning_rate": 0.0006,
+ "loss": 4.095479488372803,
+ "step": 2477
+ },
+ {
+ "epoch": 34.419213973799124,
+ "grad_norm": 0.019504647701978683,
+ "learning_rate": 0.0006,
+ "loss": 4.101964950561523,
+ "step": 2478
+ },
+ {
+ "epoch": 34.43318777292576,
+ "grad_norm": 0.019378576427698135,
+ "learning_rate": 0.0006,
+ "loss": 4.14620304107666,
+ "step": 2479
+ },
+ {
+ "epoch": 34.4471615720524,
+ "grad_norm": 0.017993204295635223,
+ "learning_rate": 0.0006,
+ "loss": 4.183492660522461,
+ "step": 2480
+ },
+ {
+ "epoch": 34.46113537117904,
+ "grad_norm": 0.01814805157482624,
+ "learning_rate": 0.0006,
+ "loss": 4.151247978210449,
+ "step": 2481
+ },
+ {
+ "epoch": 34.47510917030568,
+ "grad_norm": 0.018622448667883873,
+ "learning_rate": 0.0006,
+ "loss": 4.020938396453857,
+ "step": 2482
+ },
+ {
+ "epoch": 34.48908296943232,
+ "grad_norm": 0.018360132351517677,
+ "learning_rate": 0.0006,
+ "loss": 4.064734935760498,
+ "step": 2483
+ },
+ {
+ "epoch": 34.50305676855895,
+ "grad_norm": 0.017715711146593094,
+ "learning_rate": 0.0006,
+ "loss": 4.185687065124512,
+ "step": 2484
+ },
+ {
+ "epoch": 34.51703056768559,
+ "grad_norm": 0.01714295893907547,
+ "learning_rate": 0.0006,
+ "loss": 4.1044721603393555,
+ "step": 2485
+ },
+ {
+ "epoch": 34.531004366812226,
+ "grad_norm": 0.016943803057074547,
+ "learning_rate": 0.0006,
+ "loss": 4.106837272644043,
+ "step": 2486
+ },
+ {
+ "epoch": 34.544978165938865,
+ "grad_norm": 0.0182423684746027,
+ "learning_rate": 0.0006,
+ "loss": 4.1041340827941895,
+ "step": 2487
+ },
+ {
+ "epoch": 34.5589519650655,
+ "grad_norm": 0.01636587455868721,
+ "learning_rate": 0.0006,
+ "loss": 3.9735546112060547,
+ "step": 2488
+ },
+ {
+ "epoch": 34.57292576419214,
+ "grad_norm": 0.014593546278774738,
+ "learning_rate": 0.0006,
+ "loss": 4.041980266571045,
+ "step": 2489
+ },
+ {
+ "epoch": 34.58689956331878,
+ "grad_norm": 0.015436965972185135,
+ "learning_rate": 0.0006,
+ "loss": 4.301148414611816,
+ "step": 2490
+ },
+ {
+ "epoch": 34.60087336244541,
+ "grad_norm": 0.015075269155204296,
+ "learning_rate": 0.0006,
+ "loss": 4.118527412414551,
+ "step": 2491
+ },
+ {
+ "epoch": 34.61484716157205,
+ "grad_norm": 0.01701514609158039,
+ "learning_rate": 0.0006,
+ "loss": 4.245968818664551,
+ "step": 2492
+ },
+ {
+ "epoch": 34.62882096069869,
+ "grad_norm": 0.018047887831926346,
+ "learning_rate": 0.0006,
+ "loss": 4.227261066436768,
+ "step": 2493
+ },
+ {
+ "epoch": 34.64279475982533,
+ "grad_norm": 0.018499085679650307,
+ "learning_rate": 0.0006,
+ "loss": 4.121360778808594,
+ "step": 2494
+ },
+ {
+ "epoch": 34.65676855895197,
+ "grad_norm": 0.018987711519002914,
+ "learning_rate": 0.0006,
+ "loss": 4.131564617156982,
+ "step": 2495
+ },
+ {
+ "epoch": 34.670742358078606,
+ "grad_norm": 0.01834476925432682,
+ "learning_rate": 0.0006,
+ "loss": 4.224941253662109,
+ "step": 2496
+ },
+ {
+ "epoch": 34.68471615720524,
+ "grad_norm": 0.017507942393422127,
+ "learning_rate": 0.0006,
+ "loss": 4.1394548416137695,
+ "step": 2497
+ },
+ {
+ "epoch": 34.698689956331876,
+ "grad_norm": 0.01680580899119377,
+ "learning_rate": 0.0006,
+ "loss": 4.1073431968688965,
+ "step": 2498
+ },
+ {
+ "epoch": 34.712663755458514,
+ "grad_norm": 0.015233514830470085,
+ "learning_rate": 0.0006,
+ "loss": 4.136994361877441,
+ "step": 2499
+ },
+ {
+ "epoch": 34.72663755458515,
+ "grad_norm": 0.015482917428016663,
+ "learning_rate": 0.0006,
+ "loss": 4.110719680786133,
+ "step": 2500
+ },
+ {
+ "epoch": 34.74061135371179,
+ "grad_norm": 0.015940461307764053,
+ "learning_rate": 0.0006,
+ "loss": 4.207851886749268,
+ "step": 2501
+ },
+ {
+ "epoch": 34.75458515283843,
+ "grad_norm": 0.015355597250163555,
+ "learning_rate": 0.0006,
+ "loss": 4.169469833374023,
+ "step": 2502
+ },
+ {
+ "epoch": 34.76855895196506,
+ "grad_norm": 0.015974000096321106,
+ "learning_rate": 0.0006,
+ "loss": 4.110061168670654,
+ "step": 2503
+ },
+ {
+ "epoch": 34.7825327510917,
+ "grad_norm": 0.016088927164673805,
+ "learning_rate": 0.0006,
+ "loss": 4.155452728271484,
+ "step": 2504
+ },
+ {
+ "epoch": 34.79650655021834,
+ "grad_norm": 0.016156120225787163,
+ "learning_rate": 0.0006,
+ "loss": 4.178092956542969,
+ "step": 2505
+ },
+ {
+ "epoch": 34.81048034934498,
+ "grad_norm": 0.014928647316992283,
+ "learning_rate": 0.0006,
+ "loss": 4.05457878112793,
+ "step": 2506
+ },
+ {
+ "epoch": 34.82445414847162,
+ "grad_norm": 0.015289328061044216,
+ "learning_rate": 0.0006,
+ "loss": 4.18869686126709,
+ "step": 2507
+ },
+ {
+ "epoch": 34.838427947598255,
+ "grad_norm": 0.015799039974808693,
+ "learning_rate": 0.0006,
+ "loss": 4.106073379516602,
+ "step": 2508
+ },
+ {
+ "epoch": 34.852401746724894,
+ "grad_norm": 0.015824418514966965,
+ "learning_rate": 0.0006,
+ "loss": 4.0634589195251465,
+ "step": 2509
+ },
+ {
+ "epoch": 34.866375545851525,
+ "grad_norm": 0.015321656130254269,
+ "learning_rate": 0.0006,
+ "loss": 4.093090057373047,
+ "step": 2510
+ },
+ {
+ "epoch": 34.880349344978164,
+ "grad_norm": 0.014380166307091713,
+ "learning_rate": 0.0006,
+ "loss": 4.1750640869140625,
+ "step": 2511
+ },
+ {
+ "epoch": 34.8943231441048,
+ "grad_norm": 0.014962567947804928,
+ "learning_rate": 0.0006,
+ "loss": 4.185277938842773,
+ "step": 2512
+ },
+ {
+ "epoch": 34.90829694323144,
+ "grad_norm": 0.014467543922364712,
+ "learning_rate": 0.0006,
+ "loss": 4.1600799560546875,
+ "step": 2513
+ },
+ {
+ "epoch": 34.92227074235808,
+ "grad_norm": 0.01545181218534708,
+ "learning_rate": 0.0006,
+ "loss": 4.096299648284912,
+ "step": 2514
+ },
+ {
+ "epoch": 34.93624454148472,
+ "grad_norm": 0.01605340465903282,
+ "learning_rate": 0.0006,
+ "loss": 4.203323841094971,
+ "step": 2515
+ },
+ {
+ "epoch": 34.95021834061135,
+ "grad_norm": 0.015203815884888172,
+ "learning_rate": 0.0006,
+ "loss": 4.201385021209717,
+ "step": 2516
+ },
+ {
+ "epoch": 34.96419213973799,
+ "grad_norm": 0.015149409882724285,
+ "learning_rate": 0.0006,
+ "loss": 4.088164329528809,
+ "step": 2517
+ },
+ {
+ "epoch": 34.97816593886463,
+ "grad_norm": 0.014885558746755123,
+ "learning_rate": 0.0006,
+ "loss": 4.073911666870117,
+ "step": 2518
+ },
+ {
+ "epoch": 34.992139737991266,
+ "grad_norm": 0.015003419481217861,
+ "learning_rate": 0.0006,
+ "loss": 4.157113552093506,
+ "step": 2519
+ },
+ {
+ "epoch": 35.0,
+ "grad_norm": 0.01641916297376156,
+ "learning_rate": 0.0006,
+ "loss": 3.970052719116211,
+ "step": 2520
+ },
+ {
+ "epoch": 35.0,
+ "eval_loss": 4.641692161560059,
+ "eval_runtime": 56.8836,
+ "eval_samples_per_second": 42.93,
+ "eval_steps_per_second": 1.354,
+ "step": 2520
+ },
+ {
+ "epoch": 35.01397379912664,
+ "grad_norm": 0.015894673764705658,
+ "learning_rate": 0.0006,
+ "loss": 4.104437351226807,
+ "step": 2521
+ },
+ {
+ "epoch": 35.02794759825328,
+ "grad_norm": 0.0173867866396904,
+ "learning_rate": 0.0006,
+ "loss": 4.091726303100586,
+ "step": 2522
+ },
+ {
+ "epoch": 35.041921397379916,
+ "grad_norm": 0.019272832199931145,
+ "learning_rate": 0.0006,
+ "loss": 4.217905044555664,
+ "step": 2523
+ },
+ {
+ "epoch": 35.05589519650655,
+ "grad_norm": 0.021820371970534325,
+ "learning_rate": 0.0006,
+ "loss": 4.111428260803223,
+ "step": 2524
+ },
+ {
+ "epoch": 35.069868995633186,
+ "grad_norm": 0.02340741828083992,
+ "learning_rate": 0.0006,
+ "loss": 4.115771293640137,
+ "step": 2525
+ },
+ {
+ "epoch": 35.083842794759825,
+ "grad_norm": 0.023760691285133362,
+ "learning_rate": 0.0006,
+ "loss": 4.132383346557617,
+ "step": 2526
+ },
+ {
+ "epoch": 35.09781659388646,
+ "grad_norm": 0.02073468267917633,
+ "learning_rate": 0.0006,
+ "loss": 4.272691249847412,
+ "step": 2527
+ },
+ {
+ "epoch": 35.1117903930131,
+ "grad_norm": 0.02197718247771263,
+ "learning_rate": 0.0006,
+ "loss": 4.105586528778076,
+ "step": 2528
+ },
+ {
+ "epoch": 35.12576419213974,
+ "grad_norm": 0.0200254675000906,
+ "learning_rate": 0.0006,
+ "loss": 3.968820095062256,
+ "step": 2529
+ },
+ {
+ "epoch": 35.13973799126637,
+ "grad_norm": 0.01929587684571743,
+ "learning_rate": 0.0006,
+ "loss": 4.045468807220459,
+ "step": 2530
+ },
+ {
+ "epoch": 35.15371179039301,
+ "grad_norm": 0.01886162906885147,
+ "learning_rate": 0.0006,
+ "loss": 4.185571670532227,
+ "step": 2531
+ },
+ {
+ "epoch": 35.16768558951965,
+ "grad_norm": 0.019789839163422585,
+ "learning_rate": 0.0006,
+ "loss": 3.9734551906585693,
+ "step": 2532
+ },
+ {
+ "epoch": 35.18165938864629,
+ "grad_norm": 0.0198142658919096,
+ "learning_rate": 0.0006,
+ "loss": 4.2662506103515625,
+ "step": 2533
+ },
+ {
+ "epoch": 35.19563318777293,
+ "grad_norm": 0.02087012305855751,
+ "learning_rate": 0.0006,
+ "loss": 4.175931930541992,
+ "step": 2534
+ },
+ {
+ "epoch": 35.209606986899566,
+ "grad_norm": 0.020294535905122757,
+ "learning_rate": 0.0006,
+ "loss": 4.060324668884277,
+ "step": 2535
+ },
+ {
+ "epoch": 35.223580786026204,
+ "grad_norm": 0.01908983662724495,
+ "learning_rate": 0.0006,
+ "loss": 4.217939853668213,
+ "step": 2536
+ },
+ {
+ "epoch": 35.237554585152836,
+ "grad_norm": 0.0178892333060503,
+ "learning_rate": 0.0006,
+ "loss": 4.190216541290283,
+ "step": 2537
+ },
+ {
+ "epoch": 35.251528384279474,
+ "grad_norm": 0.017892315983772278,
+ "learning_rate": 0.0006,
+ "loss": 4.067419052124023,
+ "step": 2538
+ },
+ {
+ "epoch": 35.26550218340611,
+ "grad_norm": 0.018556250259280205,
+ "learning_rate": 0.0006,
+ "loss": 4.105958461761475,
+ "step": 2539
+ },
+ {
+ "epoch": 35.27947598253275,
+ "grad_norm": 0.01944422349333763,
+ "learning_rate": 0.0006,
+ "loss": 4.155775547027588,
+ "step": 2540
+ },
+ {
+ "epoch": 35.29344978165939,
+ "grad_norm": 0.017333146184682846,
+ "learning_rate": 0.0006,
+ "loss": 4.244513511657715,
+ "step": 2541
+ },
+ {
+ "epoch": 35.30742358078603,
+ "grad_norm": 0.01639244332909584,
+ "learning_rate": 0.0006,
+ "loss": 4.167815685272217,
+ "step": 2542
+ },
+ {
+ "epoch": 35.32139737991266,
+ "grad_norm": 0.01610143668949604,
+ "learning_rate": 0.0006,
+ "loss": 4.121486663818359,
+ "step": 2543
+ },
+ {
+ "epoch": 35.3353711790393,
+ "grad_norm": 0.01656411588191986,
+ "learning_rate": 0.0006,
+ "loss": 4.1755499839782715,
+ "step": 2544
+ },
+ {
+ "epoch": 35.34934497816594,
+ "grad_norm": 0.015075297094881535,
+ "learning_rate": 0.0006,
+ "loss": 4.102992534637451,
+ "step": 2545
+ },
+ {
+ "epoch": 35.36331877729258,
+ "grad_norm": 0.01538170501589775,
+ "learning_rate": 0.0006,
+ "loss": 4.161603927612305,
+ "step": 2546
+ },
+ {
+ "epoch": 35.377292576419215,
+ "grad_norm": 0.01483312901109457,
+ "learning_rate": 0.0006,
+ "loss": 4.085355758666992,
+ "step": 2547
+ },
+ {
+ "epoch": 35.391266375545854,
+ "grad_norm": 0.014894738793373108,
+ "learning_rate": 0.0006,
+ "loss": 4.083585262298584,
+ "step": 2548
+ },
+ {
+ "epoch": 35.40524017467249,
+ "grad_norm": 0.01475780550390482,
+ "learning_rate": 0.0006,
+ "loss": 4.012645721435547,
+ "step": 2549
+ },
+ {
+ "epoch": 35.419213973799124,
+ "grad_norm": 0.014531653374433517,
+ "learning_rate": 0.0006,
+ "loss": 4.197113990783691,
+ "step": 2550
+ },
+ {
+ "epoch": 35.43318777292576,
+ "grad_norm": 0.015316493809223175,
+ "learning_rate": 0.0006,
+ "loss": 4.227970123291016,
+ "step": 2551
+ },
+ {
+ "epoch": 35.4471615720524,
+ "grad_norm": 0.014685571193695068,
+ "learning_rate": 0.0006,
+ "loss": 4.089415550231934,
+ "step": 2552
+ },
+ {
+ "epoch": 35.46113537117904,
+ "grad_norm": 0.015337035991251469,
+ "learning_rate": 0.0006,
+ "loss": 4.080583572387695,
+ "step": 2553
+ },
+ {
+ "epoch": 35.47510917030568,
+ "grad_norm": 0.016754208132624626,
+ "learning_rate": 0.0006,
+ "loss": 4.163935661315918,
+ "step": 2554
+ },
+ {
+ "epoch": 35.48908296943232,
+ "grad_norm": 0.017535798251628876,
+ "learning_rate": 0.0006,
+ "loss": 4.093027591705322,
+ "step": 2555
+ },
+ {
+ "epoch": 35.50305676855895,
+ "grad_norm": 0.017336975783109665,
+ "learning_rate": 0.0006,
+ "loss": 4.127961158752441,
+ "step": 2556
+ },
+ {
+ "epoch": 35.51703056768559,
+ "grad_norm": 0.016242152079939842,
+ "learning_rate": 0.0006,
+ "loss": 4.155593395233154,
+ "step": 2557
+ },
+ {
+ "epoch": 35.531004366812226,
+ "grad_norm": 0.015121322125196457,
+ "learning_rate": 0.0006,
+ "loss": 3.984004020690918,
+ "step": 2558
+ },
+ {
+ "epoch": 35.544978165938865,
+ "grad_norm": 0.015178884379565716,
+ "learning_rate": 0.0006,
+ "loss": 4.111564636230469,
+ "step": 2559
+ },
+ {
+ "epoch": 35.5589519650655,
+ "grad_norm": 0.014777131378650665,
+ "learning_rate": 0.0006,
+ "loss": 4.041402816772461,
+ "step": 2560
+ },
+ {
+ "epoch": 35.57292576419214,
+ "grad_norm": 0.015197676606476307,
+ "learning_rate": 0.0006,
+ "loss": 4.133706092834473,
+ "step": 2561
+ },
+ {
+ "epoch": 35.58689956331878,
+ "grad_norm": 0.015905972570180893,
+ "learning_rate": 0.0006,
+ "loss": 4.1069135665893555,
+ "step": 2562
+ },
+ {
+ "epoch": 35.60087336244541,
+ "grad_norm": 0.015942877158522606,
+ "learning_rate": 0.0006,
+ "loss": 4.0720624923706055,
+ "step": 2563
+ },
+ {
+ "epoch": 35.61484716157205,
+ "grad_norm": 0.016697583720088005,
+ "learning_rate": 0.0006,
+ "loss": 4.167448043823242,
+ "step": 2564
+ },
+ {
+ "epoch": 35.62882096069869,
+ "grad_norm": 0.015603577718138695,
+ "learning_rate": 0.0006,
+ "loss": 4.127327919006348,
+ "step": 2565
+ },
+ {
+ "epoch": 35.64279475982533,
+ "grad_norm": 0.017153460532426834,
+ "learning_rate": 0.0006,
+ "loss": 4.056543350219727,
+ "step": 2566
+ },
+ {
+ "epoch": 35.65676855895197,
+ "grad_norm": 0.016731958836317062,
+ "learning_rate": 0.0006,
+ "loss": 4.0681843757629395,
+ "step": 2567
+ },
+ {
+ "epoch": 35.670742358078606,
+ "grad_norm": 0.016167359426617622,
+ "learning_rate": 0.0006,
+ "loss": 4.174992084503174,
+ "step": 2568
+ },
+ {
+ "epoch": 35.68471615720524,
+ "grad_norm": 0.01615108922123909,
+ "learning_rate": 0.0006,
+ "loss": 4.1383161544799805,
+ "step": 2569
+ },
+ {
+ "epoch": 35.698689956331876,
+ "grad_norm": 0.015750816091895103,
+ "learning_rate": 0.0006,
+ "loss": 4.031595706939697,
+ "step": 2570
+ },
+ {
+ "epoch": 35.712663755458514,
+ "grad_norm": 0.015416189096868038,
+ "learning_rate": 0.0006,
+ "loss": 4.067498207092285,
+ "step": 2571
+ },
+ {
+ "epoch": 35.72663755458515,
+ "grad_norm": 0.01570792682468891,
+ "learning_rate": 0.0006,
+ "loss": 4.1338300704956055,
+ "step": 2572
+ },
+ {
+ "epoch": 35.74061135371179,
+ "grad_norm": 0.016367752104997635,
+ "learning_rate": 0.0006,
+ "loss": 4.152523040771484,
+ "step": 2573
+ },
+ {
+ "epoch": 35.75458515283843,
+ "grad_norm": 0.016349803656339645,
+ "learning_rate": 0.0006,
+ "loss": 4.101985454559326,
+ "step": 2574
+ },
+ {
+ "epoch": 35.76855895196506,
+ "grad_norm": 0.01663660630583763,
+ "learning_rate": 0.0006,
+ "loss": 4.190130710601807,
+ "step": 2575
+ },
+ {
+ "epoch": 35.7825327510917,
+ "grad_norm": 0.015599170699715614,
+ "learning_rate": 0.0006,
+ "loss": 4.0792436599731445,
+ "step": 2576
+ },
+ {
+ "epoch": 35.79650655021834,
+ "grad_norm": 0.015472427010536194,
+ "learning_rate": 0.0006,
+ "loss": 4.186446666717529,
+ "step": 2577
+ },
+ {
+ "epoch": 35.81048034934498,
+ "grad_norm": 0.015520439483225346,
+ "learning_rate": 0.0006,
+ "loss": 4.246635437011719,
+ "step": 2578
+ },
+ {
+ "epoch": 35.82445414847162,
+ "grad_norm": 0.016998281702399254,
+ "learning_rate": 0.0006,
+ "loss": 4.13435173034668,
+ "step": 2579
+ },
+ {
+ "epoch": 35.838427947598255,
+ "grad_norm": 0.016290361061692238,
+ "learning_rate": 0.0006,
+ "loss": 4.107428073883057,
+ "step": 2580
+ },
+ {
+ "epoch": 35.852401746724894,
+ "grad_norm": 0.0165516659617424,
+ "learning_rate": 0.0006,
+ "loss": 4.1612138748168945,
+ "step": 2581
+ },
+ {
+ "epoch": 35.866375545851525,
+ "grad_norm": 0.01736091636121273,
+ "learning_rate": 0.0006,
+ "loss": 4.1072869300842285,
+ "step": 2582
+ },
+ {
+ "epoch": 35.880349344978164,
+ "grad_norm": 0.01787031814455986,
+ "learning_rate": 0.0006,
+ "loss": 4.1076812744140625,
+ "step": 2583
+ },
+ {
+ "epoch": 35.8943231441048,
+ "grad_norm": 0.015437985770404339,
+ "learning_rate": 0.0006,
+ "loss": 4.097488880157471,
+ "step": 2584
+ },
+ {
+ "epoch": 35.90829694323144,
+ "grad_norm": 0.016306741163134575,
+ "learning_rate": 0.0006,
+ "loss": 4.105017185211182,
+ "step": 2585
+ },
+ {
+ "epoch": 35.92227074235808,
+ "grad_norm": 0.015723245218396187,
+ "learning_rate": 0.0006,
+ "loss": 4.205283164978027,
+ "step": 2586
+ },
+ {
+ "epoch": 35.93624454148472,
+ "grad_norm": 0.014626456424593925,
+ "learning_rate": 0.0006,
+ "loss": 4.086552619934082,
+ "step": 2587
+ },
+ {
+ "epoch": 35.95021834061135,
+ "grad_norm": 0.014348454773426056,
+ "learning_rate": 0.0006,
+ "loss": 4.171329975128174,
+ "step": 2588
+ },
+ {
+ "epoch": 35.96419213973799,
+ "grad_norm": 0.015598480589687824,
+ "learning_rate": 0.0006,
+ "loss": 4.140815258026123,
+ "step": 2589
+ },
+ {
+ "epoch": 35.97816593886463,
+ "grad_norm": 0.017126062884926796,
+ "learning_rate": 0.0006,
+ "loss": 4.157176971435547,
+ "step": 2590
+ },
+ {
+ "epoch": 35.992139737991266,
+ "grad_norm": 0.01860199309885502,
+ "learning_rate": 0.0006,
+ "loss": 4.252880096435547,
+ "step": 2591
+ },
+ {
+ "epoch": 36.0,
+ "grad_norm": 0.021780220791697502,
+ "learning_rate": 0.0006,
+ "loss": 4.249861717224121,
+ "step": 2592
+ },
+ {
+ "epoch": 36.0,
+ "eval_loss": 4.606466293334961,
+ "eval_runtime": 56.6217,
+ "eval_samples_per_second": 43.128,
+ "eval_steps_per_second": 1.36,
+ "step": 2592
+ },
+ {
+ "epoch": 36.01397379912664,
+ "grad_norm": 0.020892828702926636,
+ "learning_rate": 0.0006,
+ "loss": 4.102055549621582,
+ "step": 2593
+ },
+ {
+ "epoch": 36.02794759825328,
+ "grad_norm": 0.020306255668401718,
+ "learning_rate": 0.0006,
+ "loss": 4.184308052062988,
+ "step": 2594
+ },
+ {
+ "epoch": 36.041921397379916,
+ "grad_norm": 0.02025097794830799,
+ "learning_rate": 0.0006,
+ "loss": 3.9691858291625977,
+ "step": 2595
+ },
+ {
+ "epoch": 36.05589519650655,
+ "grad_norm": 0.020254382863640785,
+ "learning_rate": 0.0006,
+ "loss": 4.02957820892334,
+ "step": 2596
+ },
+ {
+ "epoch": 36.069868995633186,
+ "grad_norm": 0.019307196140289307,
+ "learning_rate": 0.0006,
+ "loss": 4.033181667327881,
+ "step": 2597
+ },
+ {
+ "epoch": 36.083842794759825,
+ "grad_norm": 0.0198476854711771,
+ "learning_rate": 0.0006,
+ "loss": 4.0816874504089355,
+ "step": 2598
+ },
+ {
+ "epoch": 36.09781659388646,
+ "grad_norm": 0.019916515797376633,
+ "learning_rate": 0.0006,
+ "loss": 4.164889812469482,
+ "step": 2599
+ },
+ {
+ "epoch": 36.1117903930131,
+ "grad_norm": 0.0189144778996706,
+ "learning_rate": 0.0006,
+ "loss": 4.038427352905273,
+ "step": 2600
+ },
+ {
+ "epoch": 36.12576419213974,
+ "grad_norm": 0.01754768192768097,
+ "learning_rate": 0.0006,
+ "loss": 4.010427474975586,
+ "step": 2601
+ },
+ {
+ "epoch": 36.13973799126637,
+ "grad_norm": 0.016862990334630013,
+ "learning_rate": 0.0006,
+ "loss": 4.035970211029053,
+ "step": 2602
+ },
+ {
+ "epoch": 36.15371179039301,
+ "grad_norm": 0.019616752862930298,
+ "learning_rate": 0.0006,
+ "loss": 4.079898834228516,
+ "step": 2603
+ },
+ {
+ "epoch": 36.16768558951965,
+ "grad_norm": 0.020131465047597885,
+ "learning_rate": 0.0006,
+ "loss": 4.144684791564941,
+ "step": 2604
+ },
+ {
+ "epoch": 36.18165938864629,
+ "grad_norm": 0.02072521299123764,
+ "learning_rate": 0.0006,
+ "loss": 4.013660430908203,
+ "step": 2605
+ },
+ {
+ "epoch": 36.19563318777293,
+ "grad_norm": 0.01924615167081356,
+ "learning_rate": 0.0006,
+ "loss": 4.071398735046387,
+ "step": 2606
+ },
+ {
+ "epoch": 36.209606986899566,
+ "grad_norm": 0.02013954520225525,
+ "learning_rate": 0.0006,
+ "loss": 4.127094268798828,
+ "step": 2607
+ },
+ {
+ "epoch": 36.223580786026204,
+ "grad_norm": 0.02042008936405182,
+ "learning_rate": 0.0006,
+ "loss": 4.154261589050293,
+ "step": 2608
+ },
+ {
+ "epoch": 36.237554585152836,
+ "grad_norm": 0.021998876705765724,
+ "learning_rate": 0.0006,
+ "loss": 4.204150199890137,
+ "step": 2609
+ },
+ {
+ "epoch": 36.251528384279474,
+ "grad_norm": 0.022946683689951897,
+ "learning_rate": 0.0006,
+ "loss": 4.013632774353027,
+ "step": 2610
+ },
+ {
+ "epoch": 36.26550218340611,
+ "grad_norm": 0.02061830274760723,
+ "learning_rate": 0.0006,
+ "loss": 4.171084403991699,
+ "step": 2611
+ },
+ {
+ "epoch": 36.27947598253275,
+ "grad_norm": 0.01883152686059475,
+ "learning_rate": 0.0006,
+ "loss": 4.149848461151123,
+ "step": 2612
+ },
+ {
+ "epoch": 36.29344978165939,
+ "grad_norm": 0.019945282489061356,
+ "learning_rate": 0.0006,
+ "loss": 4.174656867980957,
+ "step": 2613
+ },
+ {
+ "epoch": 36.30742358078603,
+ "grad_norm": 0.020751051604747772,
+ "learning_rate": 0.0006,
+ "loss": 4.080747127532959,
+ "step": 2614
+ },
+ {
+ "epoch": 36.32139737991266,
+ "grad_norm": 0.01879691146314144,
+ "learning_rate": 0.0006,
+ "loss": 4.057356357574463,
+ "step": 2615
+ },
+ {
+ "epoch": 36.3353711790393,
+ "grad_norm": 0.019037071615457535,
+ "learning_rate": 0.0006,
+ "loss": 4.204607009887695,
+ "step": 2616
+ },
+ {
+ "epoch": 36.34934497816594,
+ "grad_norm": 0.01872078701853752,
+ "learning_rate": 0.0006,
+ "loss": 4.050586700439453,
+ "step": 2617
+ },
+ {
+ "epoch": 36.36331877729258,
+ "grad_norm": 0.018092811107635498,
+ "learning_rate": 0.0006,
+ "loss": 4.102832794189453,
+ "step": 2618
+ },
+ {
+ "epoch": 36.377292576419215,
+ "grad_norm": 0.017010487616062164,
+ "learning_rate": 0.0006,
+ "loss": 4.191788673400879,
+ "step": 2619
+ },
+ {
+ "epoch": 36.391266375545854,
+ "grad_norm": 0.01691143773496151,
+ "learning_rate": 0.0006,
+ "loss": 4.160456657409668,
+ "step": 2620
+ },
+ {
+ "epoch": 36.40524017467249,
+ "grad_norm": 0.017046816647052765,
+ "learning_rate": 0.0006,
+ "loss": 4.073828220367432,
+ "step": 2621
+ },
+ {
+ "epoch": 36.419213973799124,
+ "grad_norm": 0.01707981713116169,
+ "learning_rate": 0.0006,
+ "loss": 4.0980095863342285,
+ "step": 2622
+ },
+ {
+ "epoch": 36.43318777292576,
+ "grad_norm": 0.01630452275276184,
+ "learning_rate": 0.0006,
+ "loss": 4.101321220397949,
+ "step": 2623
+ },
+ {
+ "epoch": 36.4471615720524,
+ "grad_norm": 0.01576915942132473,
+ "learning_rate": 0.0006,
+ "loss": 4.158249855041504,
+ "step": 2624
+ },
+ {
+ "epoch": 36.46113537117904,
+ "grad_norm": 0.016179712489247322,
+ "learning_rate": 0.0006,
+ "loss": 4.077852249145508,
+ "step": 2625
+ },
+ {
+ "epoch": 36.47510917030568,
+ "grad_norm": 0.01609169878065586,
+ "learning_rate": 0.0006,
+ "loss": 4.187317848205566,
+ "step": 2626
+ },
+ {
+ "epoch": 36.48908296943232,
+ "grad_norm": 0.017148254439234734,
+ "learning_rate": 0.0006,
+ "loss": 4.109071731567383,
+ "step": 2627
+ },
+ {
+ "epoch": 36.50305676855895,
+ "grad_norm": 0.015842624008655548,
+ "learning_rate": 0.0006,
+ "loss": 4.145096302032471,
+ "step": 2628
+ },
+ {
+ "epoch": 36.51703056768559,
+ "grad_norm": 0.014906194061040878,
+ "learning_rate": 0.0006,
+ "loss": 4.028440475463867,
+ "step": 2629
+ },
+ {
+ "epoch": 36.531004366812226,
+ "grad_norm": 0.01485038734972477,
+ "learning_rate": 0.0006,
+ "loss": 4.175201416015625,
+ "step": 2630
+ },
+ {
+ "epoch": 36.544978165938865,
+ "grad_norm": 0.015834150835871696,
+ "learning_rate": 0.0006,
+ "loss": 4.076434135437012,
+ "step": 2631
+ },
+ {
+ "epoch": 36.5589519650655,
+ "grad_norm": 0.015411358326673508,
+ "learning_rate": 0.0006,
+ "loss": 4.03912878036499,
+ "step": 2632
+ },
+ {
+ "epoch": 36.57292576419214,
+ "grad_norm": 0.014762846753001213,
+ "learning_rate": 0.0006,
+ "loss": 4.019827365875244,
+ "step": 2633
+ },
+ {
+ "epoch": 36.58689956331878,
+ "grad_norm": 0.015269497409462929,
+ "learning_rate": 0.0006,
+ "loss": 4.116957187652588,
+ "step": 2634
+ },
+ {
+ "epoch": 36.60087336244541,
+ "grad_norm": 0.015171617269515991,
+ "learning_rate": 0.0006,
+ "loss": 4.2069926261901855,
+ "step": 2635
+ },
+ {
+ "epoch": 36.61484716157205,
+ "grad_norm": 0.014770902693271637,
+ "learning_rate": 0.0006,
+ "loss": 4.092021942138672,
+ "step": 2636
+ },
+ {
+ "epoch": 36.62882096069869,
+ "grad_norm": 0.014957531355321407,
+ "learning_rate": 0.0006,
+ "loss": 4.16819953918457,
+ "step": 2637
+ },
+ {
+ "epoch": 36.64279475982533,
+ "grad_norm": 0.014774451963603497,
+ "learning_rate": 0.0006,
+ "loss": 4.046073913574219,
+ "step": 2638
+ },
+ {
+ "epoch": 36.65676855895197,
+ "grad_norm": 0.014320445246994495,
+ "learning_rate": 0.0006,
+ "loss": 3.977792501449585,
+ "step": 2639
+ },
+ {
+ "epoch": 36.670742358078606,
+ "grad_norm": 0.01413130946457386,
+ "learning_rate": 0.0006,
+ "loss": 4.130649566650391,
+ "step": 2640
+ },
+ {
+ "epoch": 36.68471615720524,
+ "grad_norm": 0.014336307533085346,
+ "learning_rate": 0.0006,
+ "loss": 4.161357402801514,
+ "step": 2641
+ },
+ {
+ "epoch": 36.698689956331876,
+ "grad_norm": 0.014516931027173996,
+ "learning_rate": 0.0006,
+ "loss": 4.14248514175415,
+ "step": 2642
+ },
+ {
+ "epoch": 36.712663755458514,
+ "grad_norm": 0.014377126470208168,
+ "learning_rate": 0.0006,
+ "loss": 4.062711715698242,
+ "step": 2643
+ },
+ {
+ "epoch": 36.72663755458515,
+ "grad_norm": 0.014332668855786324,
+ "learning_rate": 0.0006,
+ "loss": 4.06423807144165,
+ "step": 2644
+ },
+ {
+ "epoch": 36.74061135371179,
+ "grad_norm": 0.014030571095645428,
+ "learning_rate": 0.0006,
+ "loss": 4.178613662719727,
+ "step": 2645
+ },
+ {
+ "epoch": 36.75458515283843,
+ "grad_norm": 0.014376776292920113,
+ "learning_rate": 0.0006,
+ "loss": 4.073130130767822,
+ "step": 2646
+ },
+ {
+ "epoch": 36.76855895196506,
+ "grad_norm": 0.013827716931700706,
+ "learning_rate": 0.0006,
+ "loss": 4.043869972229004,
+ "step": 2647
+ },
+ {
+ "epoch": 36.7825327510917,
+ "grad_norm": 0.015548762865364552,
+ "learning_rate": 0.0006,
+ "loss": 4.134049415588379,
+ "step": 2648
+ },
+ {
+ "epoch": 36.79650655021834,
+ "grad_norm": 0.017041970044374466,
+ "learning_rate": 0.0006,
+ "loss": 4.205024242401123,
+ "step": 2649
+ },
+ {
+ "epoch": 36.81048034934498,
+ "grad_norm": 0.017638275399804115,
+ "learning_rate": 0.0006,
+ "loss": 4.199924468994141,
+ "step": 2650
+ },
+ {
+ "epoch": 36.82445414847162,
+ "grad_norm": 0.017771627753973007,
+ "learning_rate": 0.0006,
+ "loss": 4.15706729888916,
+ "step": 2651
+ },
+ {
+ "epoch": 36.838427947598255,
+ "grad_norm": 0.015933562070131302,
+ "learning_rate": 0.0006,
+ "loss": 4.220742225646973,
+ "step": 2652
+ },
+ {
+ "epoch": 36.852401746724894,
+ "grad_norm": 0.01668144389986992,
+ "learning_rate": 0.0006,
+ "loss": 4.161426544189453,
+ "step": 2653
+ },
+ {
+ "epoch": 36.866375545851525,
+ "grad_norm": 0.018014680594205856,
+ "learning_rate": 0.0006,
+ "loss": 4.183046817779541,
+ "step": 2654
+ },
+ {
+ "epoch": 36.880349344978164,
+ "grad_norm": 0.018808670341968536,
+ "learning_rate": 0.0006,
+ "loss": 4.105293273925781,
+ "step": 2655
+ },
+ {
+ "epoch": 36.8943231441048,
+ "grad_norm": 0.018003476783633232,
+ "learning_rate": 0.0006,
+ "loss": 4.103171348571777,
+ "step": 2656
+ },
+ {
+ "epoch": 36.90829694323144,
+ "grad_norm": 0.016865914687514305,
+ "learning_rate": 0.0006,
+ "loss": 4.125634670257568,
+ "step": 2657
+ },
+ {
+ "epoch": 36.92227074235808,
+ "grad_norm": 0.018371153622865677,
+ "learning_rate": 0.0006,
+ "loss": 4.192461013793945,
+ "step": 2658
+ },
+ {
+ "epoch": 36.93624454148472,
+ "grad_norm": 0.019045377150177956,
+ "learning_rate": 0.0006,
+ "loss": 4.198105812072754,
+ "step": 2659
+ },
+ {
+ "epoch": 36.95021834061135,
+ "grad_norm": 0.01594168320298195,
+ "learning_rate": 0.0006,
+ "loss": 4.169788360595703,
+ "step": 2660
+ },
+ {
+ "epoch": 36.96419213973799,
+ "grad_norm": 0.015645820647478104,
+ "learning_rate": 0.0006,
+ "loss": 4.047277450561523,
+ "step": 2661
+ },
+ {
+ "epoch": 36.97816593886463,
+ "grad_norm": 0.016313156113028526,
+ "learning_rate": 0.0006,
+ "loss": 4.087392807006836,
+ "step": 2662
+ },
+ {
+ "epoch": 36.992139737991266,
+ "grad_norm": 0.017050625756382942,
+ "learning_rate": 0.0006,
+ "loss": 4.0908002853393555,
+ "step": 2663
+ },
+ {
+ "epoch": 37.0,
+ "grad_norm": 0.017730968073010445,
+ "learning_rate": 0.0006,
+ "loss": 4.110957622528076,
+ "step": 2664
+ },
+ {
+ "epoch": 37.0,
+ "eval_loss": 4.563793182373047,
+ "eval_runtime": 56.5922,
+ "eval_samples_per_second": 43.151,
+ "eval_steps_per_second": 1.361,
+ "step": 2664
+ },
+ {
+ "epoch": 37.01397379912664,
+ "grad_norm": 0.015687910839915276,
+ "learning_rate": 0.0006,
+ "loss": 4.143589019775391,
+ "step": 2665
+ },
+ {
+ "epoch": 37.02794759825328,
+ "grad_norm": 0.017451366409659386,
+ "learning_rate": 0.0006,
+ "loss": 4.0326666831970215,
+ "step": 2666
+ },
+ {
+ "epoch": 37.041921397379916,
+ "grad_norm": 0.01909538544714451,
+ "learning_rate": 0.0006,
+ "loss": 3.9706249237060547,
+ "step": 2667
+ },
+ {
+ "epoch": 37.05589519650655,
+ "grad_norm": 0.01988379657268524,
+ "learning_rate": 0.0006,
+ "loss": 3.9531009197235107,
+ "step": 2668
+ },
+ {
+ "epoch": 37.069868995633186,
+ "grad_norm": 0.02035115472972393,
+ "learning_rate": 0.0006,
+ "loss": 3.962214469909668,
+ "step": 2669
+ },
+ {
+ "epoch": 37.083842794759825,
+ "grad_norm": 0.01952001079916954,
+ "learning_rate": 0.0006,
+ "loss": 4.182262897491455,
+ "step": 2670
+ },
+ {
+ "epoch": 37.09781659388646,
+ "grad_norm": 0.01854325644671917,
+ "learning_rate": 0.0006,
+ "loss": 4.127663612365723,
+ "step": 2671
+ },
+ {
+ "epoch": 37.1117903930131,
+ "grad_norm": 0.017757616937160492,
+ "learning_rate": 0.0006,
+ "loss": 4.100804805755615,
+ "step": 2672
+ },
+ {
+ "epoch": 37.12576419213974,
+ "grad_norm": 0.016712479293346405,
+ "learning_rate": 0.0006,
+ "loss": 4.026284217834473,
+ "step": 2673
+ },
+ {
+ "epoch": 37.13973799126637,
+ "grad_norm": 0.0168139711022377,
+ "learning_rate": 0.0006,
+ "loss": 4.183882713317871,
+ "step": 2674
+ },
+ {
+ "epoch": 37.15371179039301,
+ "grad_norm": 0.017060423269867897,
+ "learning_rate": 0.0006,
+ "loss": 4.131832599639893,
+ "step": 2675
+ },
+ {
+ "epoch": 37.16768558951965,
+ "grad_norm": 0.01726631447672844,
+ "learning_rate": 0.0006,
+ "loss": 3.9156057834625244,
+ "step": 2676
+ },
+ {
+ "epoch": 37.18165938864629,
+ "grad_norm": 0.016348622739315033,
+ "learning_rate": 0.0006,
+ "loss": 4.1731367111206055,
+ "step": 2677
+ },
+ {
+ "epoch": 37.19563318777293,
+ "grad_norm": 0.016640059649944305,
+ "learning_rate": 0.0006,
+ "loss": 4.10997200012207,
+ "step": 2678
+ },
+ {
+ "epoch": 37.209606986899566,
+ "grad_norm": 0.01601342298090458,
+ "learning_rate": 0.0006,
+ "loss": 4.081145286560059,
+ "step": 2679
+ },
+ {
+ "epoch": 37.223580786026204,
+ "grad_norm": 0.01735631190240383,
+ "learning_rate": 0.0006,
+ "loss": 4.202108860015869,
+ "step": 2680
+ },
+ {
+ "epoch": 37.237554585152836,
+ "grad_norm": 0.019164685159921646,
+ "learning_rate": 0.0006,
+ "loss": 4.104109764099121,
+ "step": 2681
+ },
+ {
+ "epoch": 37.251528384279474,
+ "grad_norm": 0.01881759613752365,
+ "learning_rate": 0.0006,
+ "loss": 4.145532131195068,
+ "step": 2682
+ },
+ {
+ "epoch": 37.26550218340611,
+ "grad_norm": 0.017656555399298668,
+ "learning_rate": 0.0006,
+ "loss": 4.012782096862793,
+ "step": 2683
+ },
+ {
+ "epoch": 37.27947598253275,
+ "grad_norm": 0.01624823734164238,
+ "learning_rate": 0.0006,
+ "loss": 4.001727104187012,
+ "step": 2684
+ },
+ {
+ "epoch": 37.29344978165939,
+ "grad_norm": 0.016973139718174934,
+ "learning_rate": 0.0006,
+ "loss": 4.095399856567383,
+ "step": 2685
+ },
+ {
+ "epoch": 37.30742358078603,
+ "grad_norm": 0.015503671951591969,
+ "learning_rate": 0.0006,
+ "loss": 4.034431457519531,
+ "step": 2686
+ },
+ {
+ "epoch": 37.32139737991266,
+ "grad_norm": 0.015370250679552555,
+ "learning_rate": 0.0006,
+ "loss": 4.075836181640625,
+ "step": 2687
+ },
+ {
+ "epoch": 37.3353711790393,
+ "grad_norm": 0.01564021222293377,
+ "learning_rate": 0.0006,
+ "loss": 4.125202178955078,
+ "step": 2688
+ },
+ {
+ "epoch": 37.34934497816594,
+ "grad_norm": 0.016717975959181786,
+ "learning_rate": 0.0006,
+ "loss": 4.153017044067383,
+ "step": 2689
+ },
+ {
+ "epoch": 37.36331877729258,
+ "grad_norm": 0.015263617038726807,
+ "learning_rate": 0.0006,
+ "loss": 4.021814823150635,
+ "step": 2690
+ },
+ {
+ "epoch": 37.377292576419215,
+ "grad_norm": 0.016497142612934113,
+ "learning_rate": 0.0006,
+ "loss": 4.069095611572266,
+ "step": 2691
+ },
+ {
+ "epoch": 37.391266375545854,
+ "grad_norm": 0.01669355481863022,
+ "learning_rate": 0.0006,
+ "loss": 4.174587726593018,
+ "step": 2692
+ },
+ {
+ "epoch": 37.40524017467249,
+ "grad_norm": 0.019695773720741272,
+ "learning_rate": 0.0006,
+ "loss": 4.0073041915893555,
+ "step": 2693
+ },
+ {
+ "epoch": 37.419213973799124,
+ "grad_norm": 0.023720644414424896,
+ "learning_rate": 0.0006,
+ "loss": 4.111948013305664,
+ "step": 2694
+ },
+ {
+ "epoch": 37.43318777292576,
+ "grad_norm": 0.02399350143969059,
+ "learning_rate": 0.0006,
+ "loss": 4.081164360046387,
+ "step": 2695
+ },
+ {
+ "epoch": 37.4471615720524,
+ "grad_norm": 0.019630271941423416,
+ "learning_rate": 0.0006,
+ "loss": 4.065731048583984,
+ "step": 2696
+ },
+ {
+ "epoch": 37.46113537117904,
+ "grad_norm": 0.019679008051753044,
+ "learning_rate": 0.0006,
+ "loss": 3.993852376937866,
+ "step": 2697
+ },
+ {
+ "epoch": 37.47510917030568,
+ "grad_norm": 0.020597299560904503,
+ "learning_rate": 0.0006,
+ "loss": 4.275555610656738,
+ "step": 2698
+ },
+ {
+ "epoch": 37.48908296943232,
+ "grad_norm": 0.01983889564871788,
+ "learning_rate": 0.0006,
+ "loss": 4.254708290100098,
+ "step": 2699
+ },
+ {
+ "epoch": 37.50305676855895,
+ "grad_norm": 0.019150668755173683,
+ "learning_rate": 0.0006,
+ "loss": 4.097885608673096,
+ "step": 2700
+ },
+ {
+ "epoch": 37.51703056768559,
+ "grad_norm": 0.018847715109586716,
+ "learning_rate": 0.0006,
+ "loss": 4.054081439971924,
+ "step": 2701
+ },
+ {
+ "epoch": 37.531004366812226,
+ "grad_norm": 0.019607288762927055,
+ "learning_rate": 0.0006,
+ "loss": 3.9891440868377686,
+ "step": 2702
+ },
+ {
+ "epoch": 37.544978165938865,
+ "grad_norm": 0.018863150849938393,
+ "learning_rate": 0.0006,
+ "loss": 4.084877014160156,
+ "step": 2703
+ },
+ {
+ "epoch": 37.5589519650655,
+ "grad_norm": 0.017340239137411118,
+ "learning_rate": 0.0006,
+ "loss": 4.113290786743164,
+ "step": 2704
+ },
+ {
+ "epoch": 37.57292576419214,
+ "grad_norm": 0.015628693625330925,
+ "learning_rate": 0.0006,
+ "loss": 4.1483001708984375,
+ "step": 2705
+ },
+ {
+ "epoch": 37.58689956331878,
+ "grad_norm": 0.016493283212184906,
+ "learning_rate": 0.0006,
+ "loss": 4.1040143966674805,
+ "step": 2706
+ },
+ {
+ "epoch": 37.60087336244541,
+ "grad_norm": 0.016732443124055862,
+ "learning_rate": 0.0006,
+ "loss": 4.07205057144165,
+ "step": 2707
+ },
+ {
+ "epoch": 37.61484716157205,
+ "grad_norm": 0.017176156863570213,
+ "learning_rate": 0.0006,
+ "loss": 4.1279754638671875,
+ "step": 2708
+ },
+ {
+ "epoch": 37.62882096069869,
+ "grad_norm": 0.015730608254671097,
+ "learning_rate": 0.0006,
+ "loss": 4.086658000946045,
+ "step": 2709
+ },
+ {
+ "epoch": 37.64279475982533,
+ "grad_norm": 0.016111090779304504,
+ "learning_rate": 0.0006,
+ "loss": 4.064335823059082,
+ "step": 2710
+ },
+ {
+ "epoch": 37.65676855895197,
+ "grad_norm": 0.01769379712641239,
+ "learning_rate": 0.0006,
+ "loss": 4.116047382354736,
+ "step": 2711
+ },
+ {
+ "epoch": 37.670742358078606,
+ "grad_norm": 0.01683102548122406,
+ "learning_rate": 0.0006,
+ "loss": 4.095986843109131,
+ "step": 2712
+ },
+ {
+ "epoch": 37.68471615720524,
+ "grad_norm": 0.015726450830698013,
+ "learning_rate": 0.0006,
+ "loss": 4.10775089263916,
+ "step": 2713
+ },
+ {
+ "epoch": 37.698689956331876,
+ "grad_norm": 0.014760235324501991,
+ "learning_rate": 0.0006,
+ "loss": 4.1234893798828125,
+ "step": 2714
+ },
+ {
+ "epoch": 37.712663755458514,
+ "grad_norm": 0.014933438040316105,
+ "learning_rate": 0.0006,
+ "loss": 4.079200744628906,
+ "step": 2715
+ },
+ {
+ "epoch": 37.72663755458515,
+ "grad_norm": 0.015723997727036476,
+ "learning_rate": 0.0006,
+ "loss": 4.07865047454834,
+ "step": 2716
+ },
+ {
+ "epoch": 37.74061135371179,
+ "grad_norm": 0.014539001509547234,
+ "learning_rate": 0.0006,
+ "loss": 4.036242485046387,
+ "step": 2717
+ },
+ {
+ "epoch": 37.75458515283843,
+ "grad_norm": 0.014844979159533978,
+ "learning_rate": 0.0006,
+ "loss": 4.1051483154296875,
+ "step": 2718
+ },
+ {
+ "epoch": 37.76855895196506,
+ "grad_norm": 0.015939844772219658,
+ "learning_rate": 0.0006,
+ "loss": 4.017253875732422,
+ "step": 2719
+ },
+ {
+ "epoch": 37.7825327510917,
+ "grad_norm": 0.014771555550396442,
+ "learning_rate": 0.0006,
+ "loss": 4.0694661140441895,
+ "step": 2720
+ },
+ {
+ "epoch": 37.79650655021834,
+ "grad_norm": 0.015777921304106712,
+ "learning_rate": 0.0006,
+ "loss": 3.947587490081787,
+ "step": 2721
+ },
+ {
+ "epoch": 37.81048034934498,
+ "grad_norm": 0.017451291903853416,
+ "learning_rate": 0.0006,
+ "loss": 4.093625068664551,
+ "step": 2722
+ },
+ {
+ "epoch": 37.82445414847162,
+ "grad_norm": 0.016829246655106544,
+ "learning_rate": 0.0006,
+ "loss": 4.09030818939209,
+ "step": 2723
+ },
+ {
+ "epoch": 37.838427947598255,
+ "grad_norm": 0.016984397545456886,
+ "learning_rate": 0.0006,
+ "loss": 4.140626907348633,
+ "step": 2724
+ },
+ {
+ "epoch": 37.852401746724894,
+ "grad_norm": 0.01732603646814823,
+ "learning_rate": 0.0006,
+ "loss": 4.128276824951172,
+ "step": 2725
+ },
+ {
+ "epoch": 37.866375545851525,
+ "grad_norm": 0.01744619384407997,
+ "learning_rate": 0.0006,
+ "loss": 4.103130340576172,
+ "step": 2726
+ },
+ {
+ "epoch": 37.880349344978164,
+ "grad_norm": 0.016331320628523827,
+ "learning_rate": 0.0006,
+ "loss": 4.037951469421387,
+ "step": 2727
+ },
+ {
+ "epoch": 37.8943231441048,
+ "grad_norm": 0.015919430181384087,
+ "learning_rate": 0.0006,
+ "loss": 4.121513366699219,
+ "step": 2728
+ },
+ {
+ "epoch": 37.90829694323144,
+ "grad_norm": 0.016153695061802864,
+ "learning_rate": 0.0006,
+ "loss": 4.099384307861328,
+ "step": 2729
+ },
+ {
+ "epoch": 37.92227074235808,
+ "grad_norm": 0.01628800667822361,
+ "learning_rate": 0.0006,
+ "loss": 4.092325687408447,
+ "step": 2730
+ },
+ {
+ "epoch": 37.93624454148472,
+ "grad_norm": 0.017285684123635292,
+ "learning_rate": 0.0006,
+ "loss": 3.9828906059265137,
+ "step": 2731
+ },
+ {
+ "epoch": 37.95021834061135,
+ "grad_norm": 0.016047554090619087,
+ "learning_rate": 0.0006,
+ "loss": 4.100419044494629,
+ "step": 2732
+ },
+ {
+ "epoch": 37.96419213973799,
+ "grad_norm": 0.01562986709177494,
+ "learning_rate": 0.0006,
+ "loss": 4.065832614898682,
+ "step": 2733
+ },
+ {
+ "epoch": 37.97816593886463,
+ "grad_norm": 0.01677163504064083,
+ "learning_rate": 0.0006,
+ "loss": 4.077518463134766,
+ "step": 2734
+ },
+ {
+ "epoch": 37.992139737991266,
+ "grad_norm": 0.01755492016673088,
+ "learning_rate": 0.0006,
+ "loss": 4.1076154708862305,
+ "step": 2735
+ },
+ {
+ "epoch": 38.0,
+ "grad_norm": 0.01929762214422226,
+ "learning_rate": 0.0006,
+ "loss": 4.09603214263916,
+ "step": 2736
+ },
+ {
+ "epoch": 38.0,
+ "eval_loss": 4.582442283630371,
+ "eval_runtime": 62.8649,
+ "eval_samples_per_second": 38.845,
+ "eval_steps_per_second": 1.225,
+ "step": 2736
+ },
+ {
+ "epoch": 38.01397379912664,
+ "grad_norm": 0.019129136577248573,
+ "learning_rate": 0.0006,
+ "loss": 4.095005989074707,
+ "step": 2737
+ },
+ {
+ "epoch": 38.02794759825328,
+ "grad_norm": 0.019167408347129822,
+ "learning_rate": 0.0006,
+ "loss": 4.054426670074463,
+ "step": 2738
+ },
+ {
+ "epoch": 38.041921397379916,
+ "grad_norm": 0.01789460889995098,
+ "learning_rate": 0.0006,
+ "loss": 4.0434417724609375,
+ "step": 2739
+ },
+ {
+ "epoch": 38.05589519650655,
+ "grad_norm": 0.015947027131915092,
+ "learning_rate": 0.0006,
+ "loss": 4.074516296386719,
+ "step": 2740
+ },
+ {
+ "epoch": 38.069868995633186,
+ "grad_norm": 0.016667252406477928,
+ "learning_rate": 0.0006,
+ "loss": 4.108545303344727,
+ "step": 2741
+ },
+ {
+ "epoch": 38.083842794759825,
+ "grad_norm": 0.016014399006962776,
+ "learning_rate": 0.0006,
+ "loss": 4.026925086975098,
+ "step": 2742
+ },
+ {
+ "epoch": 38.09781659388646,
+ "grad_norm": 0.015397695824503899,
+ "learning_rate": 0.0006,
+ "loss": 4.163970470428467,
+ "step": 2743
+ },
+ {
+ "epoch": 38.1117903930131,
+ "grad_norm": 0.01563289947807789,
+ "learning_rate": 0.0006,
+ "loss": 4.116578578948975,
+ "step": 2744
+ },
+ {
+ "epoch": 38.12576419213974,
+ "grad_norm": 0.015512318350374699,
+ "learning_rate": 0.0006,
+ "loss": 4.096311569213867,
+ "step": 2745
+ },
+ {
+ "epoch": 38.13973799126637,
+ "grad_norm": 0.015426642261445522,
+ "learning_rate": 0.0006,
+ "loss": 4.017065048217773,
+ "step": 2746
+ },
+ {
+ "epoch": 38.15371179039301,
+ "grad_norm": 0.015115435235202312,
+ "learning_rate": 0.0006,
+ "loss": 4.005434989929199,
+ "step": 2747
+ },
+ {
+ "epoch": 38.16768558951965,
+ "grad_norm": 0.015235894359648228,
+ "learning_rate": 0.0006,
+ "loss": 3.982743263244629,
+ "step": 2748
+ },
+ {
+ "epoch": 38.18165938864629,
+ "grad_norm": 0.016095101833343506,
+ "learning_rate": 0.0006,
+ "loss": 4.0849385261535645,
+ "step": 2749
+ },
+ {
+ "epoch": 38.19563318777293,
+ "grad_norm": 0.015201197937130928,
+ "learning_rate": 0.0006,
+ "loss": 4.103863716125488,
+ "step": 2750
+ },
+ {
+ "epoch": 38.209606986899566,
+ "grad_norm": 0.015687478706240654,
+ "learning_rate": 0.0006,
+ "loss": 4.12074089050293,
+ "step": 2751
+ },
+ {
+ "epoch": 38.223580786026204,
+ "grad_norm": 0.015201589092612267,
+ "learning_rate": 0.0006,
+ "loss": 4.056872367858887,
+ "step": 2752
+ },
+ {
+ "epoch": 38.237554585152836,
+ "grad_norm": 0.015154894441366196,
+ "learning_rate": 0.0006,
+ "loss": 4.089670181274414,
+ "step": 2753
+ },
+ {
+ "epoch": 38.251528384279474,
+ "grad_norm": 0.01586608588695526,
+ "learning_rate": 0.0006,
+ "loss": 4.155317306518555,
+ "step": 2754
+ },
+ {
+ "epoch": 38.26550218340611,
+ "grad_norm": 0.01609654724597931,
+ "learning_rate": 0.0006,
+ "loss": 3.995800495147705,
+ "step": 2755
+ },
+ {
+ "epoch": 38.27947598253275,
+ "grad_norm": 0.015495185740292072,
+ "learning_rate": 0.0006,
+ "loss": 3.995769739151001,
+ "step": 2756
+ },
+ {
+ "epoch": 38.29344978165939,
+ "grad_norm": 0.017042381688952446,
+ "learning_rate": 0.0006,
+ "loss": 4.132832050323486,
+ "step": 2757
+ },
+ {
+ "epoch": 38.30742358078603,
+ "grad_norm": 0.018706614151597023,
+ "learning_rate": 0.0006,
+ "loss": 4.038820743560791,
+ "step": 2758
+ },
+ {
+ "epoch": 38.32139737991266,
+ "grad_norm": 0.018316026777029037,
+ "learning_rate": 0.0006,
+ "loss": 3.972052574157715,
+ "step": 2759
+ },
+ {
+ "epoch": 38.3353711790393,
+ "grad_norm": 0.018375389277935028,
+ "learning_rate": 0.0006,
+ "loss": 4.152097702026367,
+ "step": 2760
+ },
+ {
+ "epoch": 38.34934497816594,
+ "grad_norm": 0.016879886388778687,
+ "learning_rate": 0.0006,
+ "loss": 4.020869731903076,
+ "step": 2761
+ },
+ {
+ "epoch": 38.36331877729258,
+ "grad_norm": 0.015728183090686798,
+ "learning_rate": 0.0006,
+ "loss": 4.008016586303711,
+ "step": 2762
+ },
+ {
+ "epoch": 38.377292576419215,
+ "grad_norm": 0.016147758811712265,
+ "learning_rate": 0.0006,
+ "loss": 4.035854339599609,
+ "step": 2763
+ },
+ {
+ "epoch": 38.391266375545854,
+ "grad_norm": 0.017774797976017,
+ "learning_rate": 0.0006,
+ "loss": 4.122506141662598,
+ "step": 2764
+ },
+ {
+ "epoch": 38.40524017467249,
+ "grad_norm": 0.016590069979429245,
+ "learning_rate": 0.0006,
+ "loss": 4.082847595214844,
+ "step": 2765
+ },
+ {
+ "epoch": 38.419213973799124,
+ "grad_norm": 0.016544528305530548,
+ "learning_rate": 0.0006,
+ "loss": 4.184270858764648,
+ "step": 2766
+ },
+ {
+ "epoch": 38.43318777292576,
+ "grad_norm": 0.017972925677895546,
+ "learning_rate": 0.0006,
+ "loss": 3.987558603286743,
+ "step": 2767
+ },
+ {
+ "epoch": 38.4471615720524,
+ "grad_norm": 0.018562227487564087,
+ "learning_rate": 0.0006,
+ "loss": 4.1193461418151855,
+ "step": 2768
+ },
+ {
+ "epoch": 38.46113537117904,
+ "grad_norm": 0.018896685913205147,
+ "learning_rate": 0.0006,
+ "loss": 4.057808876037598,
+ "step": 2769
+ },
+ {
+ "epoch": 38.47510917030568,
+ "grad_norm": 0.016793809831142426,
+ "learning_rate": 0.0006,
+ "loss": 4.069523334503174,
+ "step": 2770
+ },
+ {
+ "epoch": 38.48908296943232,
+ "grad_norm": 0.01698676310479641,
+ "learning_rate": 0.0006,
+ "loss": 3.997624158859253,
+ "step": 2771
+ },
+ {
+ "epoch": 38.50305676855895,
+ "grad_norm": 0.016784202307462692,
+ "learning_rate": 0.0006,
+ "loss": 4.174795627593994,
+ "step": 2772
+ },
+ {
+ "epoch": 38.51703056768559,
+ "grad_norm": 0.016785014420747757,
+ "learning_rate": 0.0006,
+ "loss": 3.999321222305298,
+ "step": 2773
+ },
+ {
+ "epoch": 38.531004366812226,
+ "grad_norm": 0.018615184351801872,
+ "learning_rate": 0.0006,
+ "loss": 4.191237926483154,
+ "step": 2774
+ },
+ {
+ "epoch": 38.544978165938865,
+ "grad_norm": 0.019090790301561356,
+ "learning_rate": 0.0006,
+ "loss": 4.106966018676758,
+ "step": 2775
+ },
+ {
+ "epoch": 38.5589519650655,
+ "grad_norm": 0.018298320472240448,
+ "learning_rate": 0.0006,
+ "loss": 4.063363552093506,
+ "step": 2776
+ },
+ {
+ "epoch": 38.57292576419214,
+ "grad_norm": 0.018541090190410614,
+ "learning_rate": 0.0006,
+ "loss": 4.101193428039551,
+ "step": 2777
+ },
+ {
+ "epoch": 38.58689956331878,
+ "grad_norm": 0.01849970407783985,
+ "learning_rate": 0.0006,
+ "loss": 4.071335792541504,
+ "step": 2778
+ },
+ {
+ "epoch": 38.60087336244541,
+ "grad_norm": 0.018309077247977257,
+ "learning_rate": 0.0006,
+ "loss": 4.245305061340332,
+ "step": 2779
+ },
+ {
+ "epoch": 38.61484716157205,
+ "grad_norm": 0.017681840807199478,
+ "learning_rate": 0.0006,
+ "loss": 4.081173419952393,
+ "step": 2780
+ },
+ {
+ "epoch": 38.62882096069869,
+ "grad_norm": 0.016286224126815796,
+ "learning_rate": 0.0006,
+ "loss": 4.024351596832275,
+ "step": 2781
+ },
+ {
+ "epoch": 38.64279475982533,
+ "grad_norm": 0.015837235376238823,
+ "learning_rate": 0.0006,
+ "loss": 4.028205394744873,
+ "step": 2782
+ },
+ {
+ "epoch": 38.65676855895197,
+ "grad_norm": 0.016062596812844276,
+ "learning_rate": 0.0006,
+ "loss": 4.131122589111328,
+ "step": 2783
+ },
+ {
+ "epoch": 38.670742358078606,
+ "grad_norm": 0.01711239665746689,
+ "learning_rate": 0.0006,
+ "loss": 4.124416351318359,
+ "step": 2784
+ },
+ {
+ "epoch": 38.68471615720524,
+ "grad_norm": 0.015304913744330406,
+ "learning_rate": 0.0006,
+ "loss": 4.056122779846191,
+ "step": 2785
+ },
+ {
+ "epoch": 38.698689956331876,
+ "grad_norm": 0.015971887856721878,
+ "learning_rate": 0.0006,
+ "loss": 4.132555961608887,
+ "step": 2786
+ },
+ {
+ "epoch": 38.712663755458514,
+ "grad_norm": 0.015787221491336823,
+ "learning_rate": 0.0006,
+ "loss": 4.051596164703369,
+ "step": 2787
+ },
+ {
+ "epoch": 38.72663755458515,
+ "grad_norm": 0.01605444774031639,
+ "learning_rate": 0.0006,
+ "loss": 4.191808700561523,
+ "step": 2788
+ },
+ {
+ "epoch": 38.74061135371179,
+ "grad_norm": 0.016383009031414986,
+ "learning_rate": 0.0006,
+ "loss": 4.0648417472839355,
+ "step": 2789
+ },
+ {
+ "epoch": 38.75458515283843,
+ "grad_norm": 0.016135554760694504,
+ "learning_rate": 0.0006,
+ "loss": 3.988563060760498,
+ "step": 2790
+ },
+ {
+ "epoch": 38.76855895196506,
+ "grad_norm": 0.016634158790111542,
+ "learning_rate": 0.0006,
+ "loss": 4.133493423461914,
+ "step": 2791
+ },
+ {
+ "epoch": 38.7825327510917,
+ "grad_norm": 0.01688210666179657,
+ "learning_rate": 0.0006,
+ "loss": 4.091277122497559,
+ "step": 2792
+ },
+ {
+ "epoch": 38.79650655021834,
+ "grad_norm": 0.016102122142910957,
+ "learning_rate": 0.0006,
+ "loss": 4.051219940185547,
+ "step": 2793
+ },
+ {
+ "epoch": 38.81048034934498,
+ "grad_norm": 0.016585038974881172,
+ "learning_rate": 0.0006,
+ "loss": 4.058587074279785,
+ "step": 2794
+ },
+ {
+ "epoch": 38.82445414847162,
+ "grad_norm": 0.017700951546430588,
+ "learning_rate": 0.0006,
+ "loss": 4.016030311584473,
+ "step": 2795
+ },
+ {
+ "epoch": 38.838427947598255,
+ "grad_norm": 0.017922203987836838,
+ "learning_rate": 0.0006,
+ "loss": 4.079977989196777,
+ "step": 2796
+ },
+ {
+ "epoch": 38.852401746724894,
+ "grad_norm": 0.017299890518188477,
+ "learning_rate": 0.0006,
+ "loss": 4.062984943389893,
+ "step": 2797
+ },
+ {
+ "epoch": 38.866375545851525,
+ "grad_norm": 0.018648581579327583,
+ "learning_rate": 0.0006,
+ "loss": 4.065717697143555,
+ "step": 2798
+ },
+ {
+ "epoch": 38.880349344978164,
+ "grad_norm": 0.019057920202612877,
+ "learning_rate": 0.0006,
+ "loss": 4.227892875671387,
+ "step": 2799
+ },
+ {
+ "epoch": 38.8943231441048,
+ "grad_norm": 0.018824422731995583,
+ "learning_rate": 0.0006,
+ "loss": 4.116058349609375,
+ "step": 2800
+ },
+ {
+ "epoch": 38.90829694323144,
+ "grad_norm": 0.019084978848695755,
+ "learning_rate": 0.0006,
+ "loss": 4.061677932739258,
+ "step": 2801
+ },
+ {
+ "epoch": 38.92227074235808,
+ "grad_norm": 0.017842981964349747,
+ "learning_rate": 0.0006,
+ "loss": 4.083813667297363,
+ "step": 2802
+ },
+ {
+ "epoch": 38.93624454148472,
+ "grad_norm": 0.01880858838558197,
+ "learning_rate": 0.0006,
+ "loss": 4.152561187744141,
+ "step": 2803
+ },
+ {
+ "epoch": 38.95021834061135,
+ "grad_norm": 0.017400087788701057,
+ "learning_rate": 0.0006,
+ "loss": 4.154635429382324,
+ "step": 2804
+ },
+ {
+ "epoch": 38.96419213973799,
+ "grad_norm": 0.017682507634162903,
+ "learning_rate": 0.0006,
+ "loss": 4.151128768920898,
+ "step": 2805
+ },
+ {
+ "epoch": 38.97816593886463,
+ "grad_norm": 0.015803052112460136,
+ "learning_rate": 0.0006,
+ "loss": 4.10498571395874,
+ "step": 2806
+ },
+ {
+ "epoch": 38.992139737991266,
+ "grad_norm": 0.01582511141896248,
+ "learning_rate": 0.0006,
+ "loss": 4.140159606933594,
+ "step": 2807
+ },
+ {
+ "epoch": 39.0,
+ "grad_norm": 0.01925666816532612,
+ "learning_rate": 0.0006,
+ "loss": 4.126394271850586,
+ "step": 2808
+ },
+ {
+ "epoch": 39.0,
+ "eval_loss": 4.615617275238037,
+ "eval_runtime": 56.8895,
+ "eval_samples_per_second": 42.925,
+ "eval_steps_per_second": 1.354,
+ "step": 2808
+ },
+ {
+ "epoch": 39.01397379912664,
+ "grad_norm": 0.01677853614091873,
+ "learning_rate": 0.0006,
+ "loss": 4.102395057678223,
+ "step": 2809
+ },
+ {
+ "epoch": 39.02794759825328,
+ "grad_norm": 0.017901137471199036,
+ "learning_rate": 0.0006,
+ "loss": 3.976609230041504,
+ "step": 2810
+ },
+ {
+ "epoch": 39.041921397379916,
+ "grad_norm": 0.017978668212890625,
+ "learning_rate": 0.0006,
+ "loss": 3.944416046142578,
+ "step": 2811
+ },
+ {
+ "epoch": 39.05589519650655,
+ "grad_norm": 0.019967492669820786,
+ "learning_rate": 0.0006,
+ "loss": 3.992234468460083,
+ "step": 2812
+ },
+ {
+ "epoch": 39.069868995633186,
+ "grad_norm": 0.018937544897198677,
+ "learning_rate": 0.0006,
+ "loss": 4.028573036193848,
+ "step": 2813
+ },
+ {
+ "epoch": 39.083842794759825,
+ "grad_norm": 0.016852952539920807,
+ "learning_rate": 0.0006,
+ "loss": 4.062442779541016,
+ "step": 2814
+ },
+ {
+ "epoch": 39.09781659388646,
+ "grad_norm": 0.016643647104501724,
+ "learning_rate": 0.0006,
+ "loss": 4.08871603012085,
+ "step": 2815
+ },
+ {
+ "epoch": 39.1117903930131,
+ "grad_norm": 0.016872866079211235,
+ "learning_rate": 0.0006,
+ "loss": 3.9514245986938477,
+ "step": 2816
+ },
+ {
+ "epoch": 39.12576419213974,
+ "grad_norm": 0.018045669421553612,
+ "learning_rate": 0.0006,
+ "loss": 4.038842678070068,
+ "step": 2817
+ },
+ {
+ "epoch": 39.13973799126637,
+ "grad_norm": 0.017853403463959694,
+ "learning_rate": 0.0006,
+ "loss": 3.93709135055542,
+ "step": 2818
+ },
+ {
+ "epoch": 39.15371179039301,
+ "grad_norm": 0.017507443204522133,
+ "learning_rate": 0.0006,
+ "loss": 3.98665714263916,
+ "step": 2819
+ },
+ {
+ "epoch": 39.16768558951965,
+ "grad_norm": 0.01766853965818882,
+ "learning_rate": 0.0006,
+ "loss": 3.937126636505127,
+ "step": 2820
+ },
+ {
+ "epoch": 39.18165938864629,
+ "grad_norm": 0.016660021618008614,
+ "learning_rate": 0.0006,
+ "loss": 3.8397014141082764,
+ "step": 2821
+ },
+ {
+ "epoch": 39.19563318777293,
+ "grad_norm": 0.017719309777021408,
+ "learning_rate": 0.0006,
+ "loss": 4.003579616546631,
+ "step": 2822
+ },
+ {
+ "epoch": 39.209606986899566,
+ "grad_norm": 0.018234414979815483,
+ "learning_rate": 0.0006,
+ "loss": 4.053859710693359,
+ "step": 2823
+ },
+ {
+ "epoch": 39.223580786026204,
+ "grad_norm": 0.018777024000883102,
+ "learning_rate": 0.0006,
+ "loss": 4.056182861328125,
+ "step": 2824
+ },
+ {
+ "epoch": 39.237554585152836,
+ "grad_norm": 0.017443502321839333,
+ "learning_rate": 0.0006,
+ "loss": 4.025068283081055,
+ "step": 2825
+ },
+ {
+ "epoch": 39.251528384279474,
+ "grad_norm": 0.016500331461429596,
+ "learning_rate": 0.0006,
+ "loss": 4.101631164550781,
+ "step": 2826
+ },
+ {
+ "epoch": 39.26550218340611,
+ "grad_norm": 0.016101544722914696,
+ "learning_rate": 0.0006,
+ "loss": 4.027082443237305,
+ "step": 2827
+ },
+ {
+ "epoch": 39.27947598253275,
+ "grad_norm": 0.015471171587705612,
+ "learning_rate": 0.0006,
+ "loss": 4.022909641265869,
+ "step": 2828
+ },
+ {
+ "epoch": 39.29344978165939,
+ "grad_norm": 0.015779748558998108,
+ "learning_rate": 0.0006,
+ "loss": 4.18796968460083,
+ "step": 2829
+ },
+ {
+ "epoch": 39.30742358078603,
+ "grad_norm": 0.015522679314017296,
+ "learning_rate": 0.0006,
+ "loss": 4.005777359008789,
+ "step": 2830
+ },
+ {
+ "epoch": 39.32139737991266,
+ "grad_norm": 0.015226504765450954,
+ "learning_rate": 0.0006,
+ "loss": 4.108416557312012,
+ "step": 2831
+ },
+ {
+ "epoch": 39.3353711790393,
+ "grad_norm": 0.015095778740942478,
+ "learning_rate": 0.0006,
+ "loss": 4.08029842376709,
+ "step": 2832
+ },
+ {
+ "epoch": 39.34934497816594,
+ "grad_norm": 0.015601031482219696,
+ "learning_rate": 0.0006,
+ "loss": 3.989492416381836,
+ "step": 2833
+ },
+ {
+ "epoch": 39.36331877729258,
+ "grad_norm": 0.014825052581727505,
+ "learning_rate": 0.0006,
+ "loss": 3.998520612716675,
+ "step": 2834
+ },
+ {
+ "epoch": 39.377292576419215,
+ "grad_norm": 0.014821100980043411,
+ "learning_rate": 0.0006,
+ "loss": 4.001777172088623,
+ "step": 2835
+ },
+ {
+ "epoch": 39.391266375545854,
+ "grad_norm": 0.015333754941821098,
+ "learning_rate": 0.0006,
+ "loss": 4.11985969543457,
+ "step": 2836
+ },
+ {
+ "epoch": 39.40524017467249,
+ "grad_norm": 0.015940699726343155,
+ "learning_rate": 0.0006,
+ "loss": 4.025373935699463,
+ "step": 2837
+ },
+ {
+ "epoch": 39.419213973799124,
+ "grad_norm": 0.015096287243068218,
+ "learning_rate": 0.0006,
+ "loss": 4.068551063537598,
+ "step": 2838
+ },
+ {
+ "epoch": 39.43318777292576,
+ "grad_norm": 0.01533548254519701,
+ "learning_rate": 0.0006,
+ "loss": 3.9800801277160645,
+ "step": 2839
+ },
+ {
+ "epoch": 39.4471615720524,
+ "grad_norm": 0.014648732729256153,
+ "learning_rate": 0.0006,
+ "loss": 3.9670190811157227,
+ "step": 2840
+ },
+ {
+ "epoch": 39.46113537117904,
+ "grad_norm": 0.015759671106934547,
+ "learning_rate": 0.0006,
+ "loss": 4.050443649291992,
+ "step": 2841
+ },
+ {
+ "epoch": 39.47510917030568,
+ "grad_norm": 0.016054073348641396,
+ "learning_rate": 0.0006,
+ "loss": 4.22734260559082,
+ "step": 2842
+ },
+ {
+ "epoch": 39.48908296943232,
+ "grad_norm": 0.0160264540463686,
+ "learning_rate": 0.0006,
+ "loss": 4.116174221038818,
+ "step": 2843
+ },
+ {
+ "epoch": 39.50305676855895,
+ "grad_norm": 0.015426610596477985,
+ "learning_rate": 0.0006,
+ "loss": 4.000178337097168,
+ "step": 2844
+ },
+ {
+ "epoch": 39.51703056768559,
+ "grad_norm": 0.01520642638206482,
+ "learning_rate": 0.0006,
+ "loss": 4.218410968780518,
+ "step": 2845
+ },
+ {
+ "epoch": 39.531004366812226,
+ "grad_norm": 0.015305949375033379,
+ "learning_rate": 0.0006,
+ "loss": 4.025891304016113,
+ "step": 2846
+ },
+ {
+ "epoch": 39.544978165938865,
+ "grad_norm": 0.015461564064025879,
+ "learning_rate": 0.0006,
+ "loss": 4.100950717926025,
+ "step": 2847
+ },
+ {
+ "epoch": 39.5589519650655,
+ "grad_norm": 0.01712625101208687,
+ "learning_rate": 0.0006,
+ "loss": 4.0620856285095215,
+ "step": 2848
+ },
+ {
+ "epoch": 39.57292576419214,
+ "grad_norm": 0.017133301123976707,
+ "learning_rate": 0.0006,
+ "loss": 4.097138404846191,
+ "step": 2849
+ },
+ {
+ "epoch": 39.58689956331878,
+ "grad_norm": 0.01683625392615795,
+ "learning_rate": 0.0006,
+ "loss": 4.084364891052246,
+ "step": 2850
+ },
+ {
+ "epoch": 39.60087336244541,
+ "grad_norm": 0.0165710486471653,
+ "learning_rate": 0.0006,
+ "loss": 4.042361259460449,
+ "step": 2851
+ },
+ {
+ "epoch": 39.61484716157205,
+ "grad_norm": 0.01660393364727497,
+ "learning_rate": 0.0006,
+ "loss": 4.1395487785339355,
+ "step": 2852
+ },
+ {
+ "epoch": 39.62882096069869,
+ "grad_norm": 0.017025845125317574,
+ "learning_rate": 0.0006,
+ "loss": 4.1207075119018555,
+ "step": 2853
+ },
+ {
+ "epoch": 39.64279475982533,
+ "grad_norm": 0.01660497486591339,
+ "learning_rate": 0.0006,
+ "loss": 4.103846073150635,
+ "step": 2854
+ },
+ {
+ "epoch": 39.65676855895197,
+ "grad_norm": 0.01607946865260601,
+ "learning_rate": 0.0006,
+ "loss": 4.189016342163086,
+ "step": 2855
+ },
+ {
+ "epoch": 39.670742358078606,
+ "grad_norm": 0.01614505983889103,
+ "learning_rate": 0.0006,
+ "loss": 4.079860687255859,
+ "step": 2856
+ },
+ {
+ "epoch": 39.68471615720524,
+ "grad_norm": 0.017085086554288864,
+ "learning_rate": 0.0006,
+ "loss": 3.9774293899536133,
+ "step": 2857
+ },
+ {
+ "epoch": 39.698689956331876,
+ "grad_norm": 0.01816520094871521,
+ "learning_rate": 0.0006,
+ "loss": 4.113666534423828,
+ "step": 2858
+ },
+ {
+ "epoch": 39.712663755458514,
+ "grad_norm": 0.016597526147961617,
+ "learning_rate": 0.0006,
+ "loss": 4.068242073059082,
+ "step": 2859
+ },
+ {
+ "epoch": 39.72663755458515,
+ "grad_norm": 0.01672235131263733,
+ "learning_rate": 0.0006,
+ "loss": 4.024444580078125,
+ "step": 2860
+ },
+ {
+ "epoch": 39.74061135371179,
+ "grad_norm": 0.020093556493520737,
+ "learning_rate": 0.0006,
+ "loss": 4.064124584197998,
+ "step": 2861
+ },
+ {
+ "epoch": 39.75458515283843,
+ "grad_norm": 0.023141304031014442,
+ "learning_rate": 0.0006,
+ "loss": 4.130439758300781,
+ "step": 2862
+ },
+ {
+ "epoch": 39.76855895196506,
+ "grad_norm": 0.02289372682571411,
+ "learning_rate": 0.0006,
+ "loss": 4.14794921875,
+ "step": 2863
+ },
+ {
+ "epoch": 39.7825327510917,
+ "grad_norm": 0.01919020712375641,
+ "learning_rate": 0.0006,
+ "loss": 4.123961925506592,
+ "step": 2864
+ },
+ {
+ "epoch": 39.79650655021834,
+ "grad_norm": 0.016932329162955284,
+ "learning_rate": 0.0006,
+ "loss": 4.0944132804870605,
+ "step": 2865
+ },
+ {
+ "epoch": 39.81048034934498,
+ "grad_norm": 0.016644535586237907,
+ "learning_rate": 0.0006,
+ "loss": 4.124269962310791,
+ "step": 2866
+ },
+ {
+ "epoch": 39.82445414847162,
+ "grad_norm": 0.016302216798067093,
+ "learning_rate": 0.0006,
+ "loss": 4.0267205238342285,
+ "step": 2867
+ },
+ {
+ "epoch": 39.838427947598255,
+ "grad_norm": 0.01673300378024578,
+ "learning_rate": 0.0006,
+ "loss": 4.10650634765625,
+ "step": 2868
+ },
+ {
+ "epoch": 39.852401746724894,
+ "grad_norm": 0.016581356525421143,
+ "learning_rate": 0.0006,
+ "loss": 4.1547956466674805,
+ "step": 2869
+ },
+ {
+ "epoch": 39.866375545851525,
+ "grad_norm": 0.01691797934472561,
+ "learning_rate": 0.0006,
+ "loss": 4.056936740875244,
+ "step": 2870
+ },
+ {
+ "epoch": 39.880349344978164,
+ "grad_norm": 0.01596726104617119,
+ "learning_rate": 0.0006,
+ "loss": 3.9891414642333984,
+ "step": 2871
+ },
+ {
+ "epoch": 39.8943231441048,
+ "grad_norm": 0.015615091659128666,
+ "learning_rate": 0.0006,
+ "loss": 4.137948036193848,
+ "step": 2872
+ },
+ {
+ "epoch": 39.90829694323144,
+ "grad_norm": 0.015703413635492325,
+ "learning_rate": 0.0006,
+ "loss": 3.9844930171966553,
+ "step": 2873
+ },
+ {
+ "epoch": 39.92227074235808,
+ "grad_norm": 0.015453094616532326,
+ "learning_rate": 0.0006,
+ "loss": 4.098443508148193,
+ "step": 2874
+ },
+ {
+ "epoch": 39.93624454148472,
+ "grad_norm": 0.01449158601462841,
+ "learning_rate": 0.0006,
+ "loss": 4.090852737426758,
+ "step": 2875
+ },
+ {
+ "epoch": 39.95021834061135,
+ "grad_norm": 0.01537515688687563,
+ "learning_rate": 0.0006,
+ "loss": 4.239612579345703,
+ "step": 2876
+ },
+ {
+ "epoch": 39.96419213973799,
+ "grad_norm": 0.015989921987056732,
+ "learning_rate": 0.0006,
+ "loss": 4.142073154449463,
+ "step": 2877
+ },
+ {
+ "epoch": 39.97816593886463,
+ "grad_norm": 0.015985611826181412,
+ "learning_rate": 0.0006,
+ "loss": 4.100076675415039,
+ "step": 2878
+ },
+ {
+ "epoch": 39.992139737991266,
+ "grad_norm": 0.015260833315551281,
+ "learning_rate": 0.0006,
+ "loss": 4.068576335906982,
+ "step": 2879
+ },
+ {
+ "epoch": 40.0,
+ "grad_norm": 0.018047941848635674,
+ "learning_rate": 0.0006,
+ "loss": 4.005878925323486,
+ "step": 2880
+ },
+ {
+ "epoch": 40.0,
+ "eval_loss": 4.622848987579346,
+ "eval_runtime": 56.9618,
+ "eval_samples_per_second": 42.871,
+ "eval_steps_per_second": 1.352,
+ "step": 2880
+ },
+ {
+ "epoch": 40.01397379912664,
+ "grad_norm": 0.015976838767528534,
+ "learning_rate": 0.0006,
+ "loss": 4.035689353942871,
+ "step": 2881
+ },
+ {
+ "epoch": 40.02794759825328,
+ "grad_norm": 0.015176220797002316,
+ "learning_rate": 0.0006,
+ "loss": 3.932197332382202,
+ "step": 2882
+ },
+ {
+ "epoch": 40.041921397379916,
+ "grad_norm": 0.016676228493452072,
+ "learning_rate": 0.0006,
+ "loss": 4.014171123504639,
+ "step": 2883
+ },
+ {
+ "epoch": 40.05589519650655,
+ "grad_norm": 0.01698281429708004,
+ "learning_rate": 0.0006,
+ "loss": 4.117893218994141,
+ "step": 2884
+ },
+ {
+ "epoch": 40.069868995633186,
+ "grad_norm": 0.017872486263513565,
+ "learning_rate": 0.0006,
+ "loss": 4.132119178771973,
+ "step": 2885
+ },
+ {
+ "epoch": 40.083842794759825,
+ "grad_norm": 0.017615554854273796,
+ "learning_rate": 0.0006,
+ "loss": 4.072000980377197,
+ "step": 2886
+ },
+ {
+ "epoch": 40.09781659388646,
+ "grad_norm": 0.017469394952058792,
+ "learning_rate": 0.0006,
+ "loss": 4.017726421356201,
+ "step": 2887
+ },
+ {
+ "epoch": 40.1117903930131,
+ "grad_norm": 0.01954762637615204,
+ "learning_rate": 0.0006,
+ "loss": 4.014443874359131,
+ "step": 2888
+ },
+ {
+ "epoch": 40.12576419213974,
+ "grad_norm": 0.018935853615403175,
+ "learning_rate": 0.0006,
+ "loss": 4.0042948722839355,
+ "step": 2889
+ },
+ {
+ "epoch": 40.13973799126637,
+ "grad_norm": 0.019382692873477936,
+ "learning_rate": 0.0006,
+ "loss": 4.06160831451416,
+ "step": 2890
+ },
+ {
+ "epoch": 40.15371179039301,
+ "grad_norm": 0.019677115604281425,
+ "learning_rate": 0.0006,
+ "loss": 4.13466215133667,
+ "step": 2891
+ },
+ {
+ "epoch": 40.16768558951965,
+ "grad_norm": 0.020365122705698013,
+ "learning_rate": 0.0006,
+ "loss": 4.095000267028809,
+ "step": 2892
+ },
+ {
+ "epoch": 40.18165938864629,
+ "grad_norm": 0.021842138841748238,
+ "learning_rate": 0.0006,
+ "loss": 3.9877853393554688,
+ "step": 2893
+ },
+ {
+ "epoch": 40.19563318777293,
+ "grad_norm": 0.02354881353676319,
+ "learning_rate": 0.0006,
+ "loss": 3.9430112838745117,
+ "step": 2894
+ },
+ {
+ "epoch": 40.209606986899566,
+ "grad_norm": 0.026602735742926598,
+ "learning_rate": 0.0006,
+ "loss": 4.012076377868652,
+ "step": 2895
+ },
+ {
+ "epoch": 40.223580786026204,
+ "grad_norm": 0.023976484313607216,
+ "learning_rate": 0.0006,
+ "loss": 3.9768595695495605,
+ "step": 2896
+ },
+ {
+ "epoch": 40.237554585152836,
+ "grad_norm": 0.02336616814136505,
+ "learning_rate": 0.0006,
+ "loss": 4.1096038818359375,
+ "step": 2897
+ },
+ {
+ "epoch": 40.251528384279474,
+ "grad_norm": 0.02426023595035076,
+ "learning_rate": 0.0006,
+ "loss": 4.157189846038818,
+ "step": 2898
+ },
+ {
+ "epoch": 40.26550218340611,
+ "grad_norm": 0.021572696045041084,
+ "learning_rate": 0.0006,
+ "loss": 3.979175090789795,
+ "step": 2899
+ },
+ {
+ "epoch": 40.27947598253275,
+ "grad_norm": 0.08300432562828064,
+ "learning_rate": 0.0006,
+ "loss": 4.092336654663086,
+ "step": 2900
+ },
+ {
+ "epoch": 40.29344978165939,
+ "grad_norm": 0.7323967814445496,
+ "learning_rate": 0.0006,
+ "loss": 5.337990760803223,
+ "step": 2901
+ },
+ {
+ "epoch": 40.30742358078603,
+ "grad_norm": 0.3363337814807892,
+ "learning_rate": 0.0006,
+ "loss": 6.40043830871582,
+ "step": 2902
+ },
+ {
+ "epoch": 40.32139737991266,
+ "grad_norm": 0.598686933517456,
+ "learning_rate": 0.0006,
+ "loss": 7.004765510559082,
+ "step": 2903
+ },
+ {
+ "epoch": 40.3353711790393,
+ "grad_norm": 0.5998364686965942,
+ "learning_rate": 0.0006,
+ "loss": 8.184747695922852,
+ "step": 2904
+ },
+ {
+ "epoch": 40.34934497816594,
+ "grad_norm": 1.408208966255188,
+ "learning_rate": 0.0006,
+ "loss": 8.88978385925293,
+ "step": 2905
+ },
+ {
+ "epoch": 40.36331877729258,
+ "grad_norm": 0.7829785943031311,
+ "learning_rate": 0.0006,
+ "loss": 11.13601303100586,
+ "step": 2906
+ },
+ {
+ "epoch": 40.377292576419215,
+ "grad_norm": 0.7769215106964111,
+ "learning_rate": 0.0006,
+ "loss": 9.90416145324707,
+ "step": 2907
+ },
+ {
+ "epoch": 40.391266375545854,
+ "grad_norm": 0.46508079767227173,
+ "learning_rate": 0.0006,
+ "loss": 9.647209167480469,
+ "step": 2908
+ },
+ {
+ "epoch": 40.40524017467249,
+ "grad_norm": 1.0228968858718872,
+ "learning_rate": 0.0006,
+ "loss": 11.685434341430664,
+ "step": 2909
+ },
+ {
+ "epoch": 40.419213973799124,
+ "grad_norm": 0.20553778111934662,
+ "learning_rate": 0.0006,
+ "loss": 8.621708869934082,
+ "step": 2910
+ },
+ {
+ "epoch": 40.43318777292576,
+ "grad_norm": 0.18466424942016602,
+ "learning_rate": 0.0006,
+ "loss": 8.590787887573242,
+ "step": 2911
+ },
+ {
+ "epoch": 40.4471615720524,
+ "grad_norm": 0.10458594560623169,
+ "learning_rate": 0.0006,
+ "loss": 8.324458122253418,
+ "step": 2912
+ },
+ {
+ "epoch": 40.46113537117904,
+ "grad_norm": 0.1096842959523201,
+ "learning_rate": 0.0006,
+ "loss": 8.300851821899414,
+ "step": 2913
+ },
+ {
+ "epoch": 40.47510917030568,
+ "grad_norm": 0.10501226782798767,
+ "learning_rate": 0.0006,
+ "loss": 8.21009635925293,
+ "step": 2914
+ },
+ {
+ "epoch": 40.48908296943232,
+ "grad_norm": 0.13704165816307068,
+ "learning_rate": 0.0006,
+ "loss": 8.141355514526367,
+ "step": 2915
+ },
+ {
+ "epoch": 40.50305676855895,
+ "grad_norm": 0.12129108607769012,
+ "learning_rate": 0.0006,
+ "loss": 8.074612617492676,
+ "step": 2916
+ },
+ {
+ "epoch": 40.51703056768559,
+ "grad_norm": 0.10028897225856781,
+ "learning_rate": 0.0006,
+ "loss": 8.00384521484375,
+ "step": 2917
+ },
+ {
+ "epoch": 40.531004366812226,
+ "grad_norm": 0.11764246970415115,
+ "learning_rate": 0.0006,
+ "loss": 8.078598022460938,
+ "step": 2918
+ },
+ {
+ "epoch": 40.544978165938865,
+ "grad_norm": 0.10081959515810013,
+ "learning_rate": 0.0006,
+ "loss": 8.008938789367676,
+ "step": 2919
+ },
+ {
+ "epoch": 40.5589519650655,
+ "grad_norm": 0.1003170907497406,
+ "learning_rate": 0.0006,
+ "loss": 7.962940216064453,
+ "step": 2920
+ },
+ {
+ "epoch": 40.57292576419214,
+ "grad_norm": 0.08763492107391357,
+ "learning_rate": 0.0006,
+ "loss": 7.909419536590576,
+ "step": 2921
+ },
+ {
+ "epoch": 40.58689956331878,
+ "grad_norm": 0.0636807233095169,
+ "learning_rate": 0.0006,
+ "loss": 7.878905773162842,
+ "step": 2922
+ },
+ {
+ "epoch": 40.60087336244541,
+ "grad_norm": 0.06005936488509178,
+ "learning_rate": 0.0006,
+ "loss": 7.850980758666992,
+ "step": 2923
+ },
+ {
+ "epoch": 40.61484716157205,
+ "grad_norm": 0.06415296345949173,
+ "learning_rate": 0.0006,
+ "loss": 7.830353260040283,
+ "step": 2924
+ },
+ {
+ "epoch": 40.62882096069869,
+ "grad_norm": 0.0732007846236229,
+ "learning_rate": 0.0006,
+ "loss": 7.830746650695801,
+ "step": 2925
+ },
+ {
+ "epoch": 40.64279475982533,
+ "grad_norm": 0.06628110259771347,
+ "learning_rate": 0.0006,
+ "loss": 7.82661771774292,
+ "step": 2926
+ },
+ {
+ "epoch": 40.65676855895197,
+ "grad_norm": 0.05544218048453331,
+ "learning_rate": 0.0006,
+ "loss": 7.814305782318115,
+ "step": 2927
+ },
+ {
+ "epoch": 40.670742358078606,
+ "grad_norm": 0.050327103585004807,
+ "learning_rate": 0.0006,
+ "loss": 7.764055252075195,
+ "step": 2928
+ },
+ {
+ "epoch": 40.68471615720524,
+ "grad_norm": 0.04918425530195236,
+ "learning_rate": 0.0006,
+ "loss": 7.744625091552734,
+ "step": 2929
+ },
+ {
+ "epoch": 40.698689956331876,
+ "grad_norm": 0.042862266302108765,
+ "learning_rate": 0.0006,
+ "loss": 7.764426231384277,
+ "step": 2930
+ },
+ {
+ "epoch": 40.712663755458514,
+ "grad_norm": 0.04338089004158974,
+ "learning_rate": 0.0006,
+ "loss": 7.712530136108398,
+ "step": 2931
+ },
+ {
+ "epoch": 40.72663755458515,
+ "grad_norm": 0.04127737879753113,
+ "learning_rate": 0.0006,
+ "loss": 7.742180347442627,
+ "step": 2932
+ },
+ {
+ "epoch": 40.74061135371179,
+ "grad_norm": 0.03358631581068039,
+ "learning_rate": 0.0006,
+ "loss": 7.714834213256836,
+ "step": 2933
+ },
+ {
+ "epoch": 40.75458515283843,
+ "grad_norm": 0.02845790982246399,
+ "learning_rate": 0.0006,
+ "loss": 7.7021026611328125,
+ "step": 2934
+ },
+ {
+ "epoch": 40.76855895196506,
+ "grad_norm": 0.03012419492006302,
+ "learning_rate": 0.0006,
+ "loss": 7.712377548217773,
+ "step": 2935
+ },
+ {
+ "epoch": 40.7825327510917,
+ "grad_norm": 0.032981179654598236,
+ "learning_rate": 0.0006,
+ "loss": 7.666385650634766,
+ "step": 2936
+ },
+ {
+ "epoch": 40.79650655021834,
+ "grad_norm": 0.03357074037194252,
+ "learning_rate": 0.0006,
+ "loss": 7.649097919464111,
+ "step": 2937
+ },
+ {
+ "epoch": 40.81048034934498,
+ "grad_norm": 0.030914708971977234,
+ "learning_rate": 0.0006,
+ "loss": 7.649196624755859,
+ "step": 2938
+ },
+ {
+ "epoch": 40.82445414847162,
+ "grad_norm": 0.026296131312847137,
+ "learning_rate": 0.0006,
+ "loss": 7.629964828491211,
+ "step": 2939
+ },
+ {
+ "epoch": 40.838427947598255,
+ "grad_norm": 0.026677245274186134,
+ "learning_rate": 0.0006,
+ "loss": 7.599973678588867,
+ "step": 2940
+ },
+ {
+ "epoch": 40.852401746724894,
+ "grad_norm": 0.023208219558000565,
+ "learning_rate": 0.0006,
+ "loss": 7.601624488830566,
+ "step": 2941
+ },
+ {
+ "epoch": 40.866375545851525,
+ "grad_norm": 0.020741475746035576,
+ "learning_rate": 0.0006,
+ "loss": 7.603902339935303,
+ "step": 2942
+ },
+ {
+ "epoch": 40.880349344978164,
+ "grad_norm": 0.019221751019358635,
+ "learning_rate": 0.0006,
+ "loss": 7.5535888671875,
+ "step": 2943
+ },
+ {
+ "epoch": 40.8943231441048,
+ "grad_norm": 0.019491951912641525,
+ "learning_rate": 0.0006,
+ "loss": 7.54597282409668,
+ "step": 2944
+ },
+ {
+ "epoch": 40.90829694323144,
+ "grad_norm": 0.019702572375535965,
+ "learning_rate": 0.0006,
+ "loss": 7.537567138671875,
+ "step": 2945
+ },
+ {
+ "epoch": 40.92227074235808,
+ "grad_norm": 0.01768800988793373,
+ "learning_rate": 0.0006,
+ "loss": 7.512800216674805,
+ "step": 2946
+ },
+ {
+ "epoch": 40.93624454148472,
+ "grad_norm": 0.01926138624548912,
+ "learning_rate": 0.0006,
+ "loss": 7.509932518005371,
+ "step": 2947
+ },
+ {
+ "epoch": 40.95021834061135,
+ "grad_norm": 0.019030006602406502,
+ "learning_rate": 0.0006,
+ "loss": 7.497983932495117,
+ "step": 2948
+ },
+ {
+ "epoch": 40.96419213973799,
+ "grad_norm": 0.017313841730356216,
+ "learning_rate": 0.0006,
+ "loss": 7.463518142700195,
+ "step": 2949
+ },
+ {
+ "epoch": 40.97816593886463,
+ "grad_norm": 0.01310762483626604,
+ "learning_rate": 0.0006,
+ "loss": 7.474668502807617,
+ "step": 2950
+ },
+ {
+ "epoch": 40.992139737991266,
+ "grad_norm": 0.013008282519876957,
+ "learning_rate": 0.0006,
+ "loss": 7.45982551574707,
+ "step": 2951
+ },
+ {
+ "epoch": 41.0,
+ "grad_norm": 0.01441363524645567,
+ "learning_rate": 0.0006,
+ "loss": 7.446529388427734,
+ "step": 2952
+ },
+ {
+ "epoch": 41.0,
+ "eval_loss": 7.486120223999023,
+ "eval_runtime": 70.4986,
+ "eval_samples_per_second": 34.639,
+ "eval_steps_per_second": 1.092,
+ "step": 2952
+ },
+ {
+ "epoch": 41.01397379912664,
+ "grad_norm": 0.014061855152249336,
+ "learning_rate": 0.0006,
+ "loss": 7.4692182540893555,
+ "step": 2953
+ },
+ {
+ "epoch": 41.02794759825328,
+ "grad_norm": 0.01587500050663948,
+ "learning_rate": 0.0006,
+ "loss": 7.4182024002075195,
+ "step": 2954
+ },
+ {
+ "epoch": 41.041921397379916,
+ "grad_norm": 0.014783984050154686,
+ "learning_rate": 0.0006,
+ "loss": 7.466778755187988,
+ "step": 2955
+ },
+ {
+ "epoch": 41.05589519650655,
+ "grad_norm": 0.011560854502022266,
+ "learning_rate": 0.0006,
+ "loss": 7.438002586364746,
+ "step": 2956
+ },
+ {
+ "epoch": 41.069868995633186,
+ "grad_norm": 0.00970957800745964,
+ "learning_rate": 0.0006,
+ "loss": 7.442694187164307,
+ "step": 2957
+ },
+ {
+ "epoch": 41.083842794759825,
+ "grad_norm": 0.009633459150791168,
+ "learning_rate": 0.0006,
+ "loss": 7.458884239196777,
+ "step": 2958
+ },
+ {
+ "epoch": 41.09781659388646,
+ "grad_norm": 0.010664350353181362,
+ "learning_rate": 0.0006,
+ "loss": 7.48378324508667,
+ "step": 2959
+ },
+ {
+ "epoch": 41.1117903930131,
+ "grad_norm": 0.011440591886639595,
+ "learning_rate": 0.0006,
+ "loss": 7.474484920501709,
+ "step": 2960
+ },
+ {
+ "epoch": 41.12576419213974,
+ "grad_norm": 0.01150908600538969,
+ "learning_rate": 0.0006,
+ "loss": 7.417902946472168,
+ "step": 2961
+ },
+ {
+ "epoch": 41.13973799126637,
+ "grad_norm": 0.010375697165727615,
+ "learning_rate": 0.0006,
+ "loss": 7.403279781341553,
+ "step": 2962
+ },
+ {
+ "epoch": 41.15371179039301,
+ "grad_norm": 0.009075703099370003,
+ "learning_rate": 0.0006,
+ "loss": 7.447347164154053,
+ "step": 2963
+ },
+ {
+ "epoch": 41.16768558951965,
+ "grad_norm": 0.008892177604138851,
+ "learning_rate": 0.0006,
+ "loss": 7.402681350708008,
+ "step": 2964
+ },
+ {
+ "epoch": 41.18165938864629,
+ "grad_norm": 0.009492084383964539,
+ "learning_rate": 0.0006,
+ "loss": 7.3871355056762695,
+ "step": 2965
+ },
+ {
+ "epoch": 41.19563318777293,
+ "grad_norm": 0.009139460511505604,
+ "learning_rate": 0.0006,
+ "loss": 7.419617176055908,
+ "step": 2966
+ },
+ {
+ "epoch": 41.209606986899566,
+ "grad_norm": 0.007959138602018356,
+ "learning_rate": 0.0006,
+ "loss": 7.374949932098389,
+ "step": 2967
+ },
+ {
+ "epoch": 41.223580786026204,
+ "grad_norm": 0.007571548223495483,
+ "learning_rate": 0.0006,
+ "loss": 7.4205241203308105,
+ "step": 2968
+ },
+ {
+ "epoch": 41.237554585152836,
+ "grad_norm": 0.007885400205850601,
+ "learning_rate": 0.0006,
+ "loss": 7.416543006896973,
+ "step": 2969
+ },
+ {
+ "epoch": 41.251528384279474,
+ "grad_norm": 0.008650158531963825,
+ "learning_rate": 0.0006,
+ "loss": 7.357326984405518,
+ "step": 2970
+ },
+ {
+ "epoch": 41.26550218340611,
+ "grad_norm": 0.007483262103050947,
+ "learning_rate": 0.0006,
+ "loss": 7.410891056060791,
+ "step": 2971
+ },
+ {
+ "epoch": 41.27947598253275,
+ "grad_norm": 0.00869277399033308,
+ "learning_rate": 0.0006,
+ "loss": 7.392127990722656,
+ "step": 2972
+ },
+ {
+ "epoch": 41.29344978165939,
+ "grad_norm": 0.00885078776627779,
+ "learning_rate": 0.0006,
+ "loss": 7.405449867248535,
+ "step": 2973
+ },
+ {
+ "epoch": 41.30742358078603,
+ "grad_norm": 0.00958226341754198,
+ "learning_rate": 0.0006,
+ "loss": 7.399868488311768,
+ "step": 2974
+ },
+ {
+ "epoch": 41.32139737991266,
+ "grad_norm": 0.006991599686443806,
+ "learning_rate": 0.0006,
+ "loss": 7.35369873046875,
+ "step": 2975
+ },
+ {
+ "epoch": 41.3353711790393,
+ "grad_norm": 0.007415142375975847,
+ "learning_rate": 0.0006,
+ "loss": 7.358353137969971,
+ "step": 2976
+ },
+ {
+ "epoch": 41.34934497816594,
+ "grad_norm": 0.007024371065199375,
+ "learning_rate": 0.0006,
+ "loss": 7.4070892333984375,
+ "step": 2977
+ },
+ {
+ "epoch": 41.36331877729258,
+ "grad_norm": 0.007234332151710987,
+ "learning_rate": 0.0006,
+ "loss": 7.379120349884033,
+ "step": 2978
+ },
+ {
+ "epoch": 41.377292576419215,
+ "grad_norm": 0.00798892229795456,
+ "learning_rate": 0.0006,
+ "loss": 7.365455627441406,
+ "step": 2979
+ },
+ {
+ "epoch": 41.391266375545854,
+ "grad_norm": 0.007303398102521896,
+ "learning_rate": 0.0006,
+ "loss": 7.341798782348633,
+ "step": 2980
+ },
+ {
+ "epoch": 41.40524017467249,
+ "grad_norm": 0.007119669578969479,
+ "learning_rate": 0.0006,
+ "loss": 7.363280296325684,
+ "step": 2981
+ },
+ {
+ "epoch": 41.419213973799124,
+ "grad_norm": 0.007457619532942772,
+ "learning_rate": 0.0006,
+ "loss": 7.39388370513916,
+ "step": 2982
+ },
+ {
+ "epoch": 41.43318777292576,
+ "grad_norm": 0.007905688136816025,
+ "learning_rate": 0.0006,
+ "loss": 7.380652904510498,
+ "step": 2983
+ },
+ {
+ "epoch": 41.4471615720524,
+ "grad_norm": 0.006212855689227581,
+ "learning_rate": 0.0006,
+ "loss": 7.398769378662109,
+ "step": 2984
+ },
+ {
+ "epoch": 41.46113537117904,
+ "grad_norm": 0.0070641543716192245,
+ "learning_rate": 0.0006,
+ "loss": 7.3503546714782715,
+ "step": 2985
+ },
+ {
+ "epoch": 41.47510917030568,
+ "grad_norm": 0.00787805300205946,
+ "learning_rate": 0.0006,
+ "loss": 7.347678184509277,
+ "step": 2986
+ },
+ {
+ "epoch": 41.48908296943232,
+ "grad_norm": 0.006760426796972752,
+ "learning_rate": 0.0006,
+ "loss": 7.380006790161133,
+ "step": 2987
+ },
+ {
+ "epoch": 41.50305676855895,
+ "grad_norm": 0.00772385997697711,
+ "learning_rate": 0.0006,
+ "loss": 7.363680362701416,
+ "step": 2988
+ },
+ {
+ "epoch": 41.51703056768559,
+ "grad_norm": 0.007373564876616001,
+ "learning_rate": 0.0006,
+ "loss": 7.333014488220215,
+ "step": 2989
+ },
+ {
+ "epoch": 41.531004366812226,
+ "grad_norm": 0.006717653013765812,
+ "learning_rate": 0.0006,
+ "loss": 7.360076427459717,
+ "step": 2990
+ },
+ {
+ "epoch": 41.544978165938865,
+ "grad_norm": 0.007191413082182407,
+ "learning_rate": 0.0006,
+ "loss": 7.397030830383301,
+ "step": 2991
+ },
+ {
+ "epoch": 41.5589519650655,
+ "grad_norm": 0.007831960916519165,
+ "learning_rate": 0.0006,
+ "loss": 7.369959831237793,
+ "step": 2992
+ },
+ {
+ "epoch": 41.57292576419214,
+ "grad_norm": 0.006947481073439121,
+ "learning_rate": 0.0006,
+ "loss": 7.333254337310791,
+ "step": 2993
+ },
+ {
+ "epoch": 41.58689956331878,
+ "grad_norm": 0.006243606563657522,
+ "learning_rate": 0.0006,
+ "loss": 7.327378273010254,
+ "step": 2994
+ },
+ {
+ "epoch": 41.60087336244541,
+ "grad_norm": 0.006298339460045099,
+ "learning_rate": 0.0006,
+ "loss": 7.323482036590576,
+ "step": 2995
+ },
+ {
+ "epoch": 41.61484716157205,
+ "grad_norm": 0.007587412837892771,
+ "learning_rate": 0.0006,
+ "loss": 7.313789367675781,
+ "step": 2996
+ },
+ {
+ "epoch": 41.62882096069869,
+ "grad_norm": 0.007140746805816889,
+ "learning_rate": 0.0006,
+ "loss": 7.356067657470703,
+ "step": 2997
+ },
+ {
+ "epoch": 41.64279475982533,
+ "grad_norm": 0.008315625600516796,
+ "learning_rate": 0.0006,
+ "loss": 7.3339643478393555,
+ "step": 2998
+ },
+ {
+ "epoch": 41.65676855895197,
+ "grad_norm": 0.006547102238982916,
+ "learning_rate": 0.0006,
+ "loss": 7.342191696166992,
+ "step": 2999
+ },
+ {
+ "epoch": 41.670742358078606,
+ "grad_norm": 0.006068122107535601,
+ "learning_rate": 0.0006,
+ "loss": 7.321015357971191,
+ "step": 3000
+ },
+ {
+ "epoch": 41.68471615720524,
+ "grad_norm": 0.007316991221159697,
+ "learning_rate": 0.0006,
+ "loss": 7.346490859985352,
+ "step": 3001
+ },
+ {
+ "epoch": 41.698689956331876,
+ "grad_norm": 0.0067388564348220825,
+ "learning_rate": 0.0006,
+ "loss": 7.340723991394043,
+ "step": 3002
+ },
+ {
+ "epoch": 41.712663755458514,
+ "grad_norm": 0.006839963141828775,
+ "learning_rate": 0.0006,
+ "loss": 7.329365253448486,
+ "step": 3003
+ },
+ {
+ "epoch": 41.72663755458515,
+ "grad_norm": 0.00805041566491127,
+ "learning_rate": 0.0006,
+ "loss": 7.298934459686279,
+ "step": 3004
+ },
+ {
+ "epoch": 41.74061135371179,
+ "grad_norm": 0.007197503466159105,
+ "learning_rate": 0.0006,
+ "loss": 7.331496238708496,
+ "step": 3005
+ },
+ {
+ "epoch": 41.75458515283843,
+ "grad_norm": 0.007220805156975985,
+ "learning_rate": 0.0006,
+ "loss": 7.317695140838623,
+ "step": 3006
+ },
+ {
+ "epoch": 41.76855895196506,
+ "grad_norm": 0.00874276738613844,
+ "learning_rate": 0.0006,
+ "loss": 7.272614479064941,
+ "step": 3007
+ },
+ {
+ "epoch": 41.7825327510917,
+ "grad_norm": 0.008386565372347832,
+ "learning_rate": 0.0006,
+ "loss": 7.295244216918945,
+ "step": 3008
+ },
+ {
+ "epoch": 41.79650655021834,
+ "grad_norm": 0.0077730282209813595,
+ "learning_rate": 0.0006,
+ "loss": 7.3395490646362305,
+ "step": 3009
+ },
+ {
+ "epoch": 41.81048034934498,
+ "grad_norm": 0.008441480807960033,
+ "learning_rate": 0.0006,
+ "loss": 7.319921493530273,
+ "step": 3010
+ },
+ {
+ "epoch": 41.82445414847162,
+ "grad_norm": 0.007398269604891539,
+ "learning_rate": 0.0006,
+ "loss": 7.3029465675354,
+ "step": 3011
+ },
+ {
+ "epoch": 41.838427947598255,
+ "grad_norm": 0.006878677289932966,
+ "learning_rate": 0.0006,
+ "loss": 7.318467140197754,
+ "step": 3012
+ },
+ {
+ "epoch": 41.852401746724894,
+ "grad_norm": 0.007616598624736071,
+ "learning_rate": 0.0006,
+ "loss": 7.3112077713012695,
+ "step": 3013
+ },
+ {
+ "epoch": 41.866375545851525,
+ "grad_norm": 0.011629699729382992,
+ "learning_rate": 0.0006,
+ "loss": 7.32523250579834,
+ "step": 3014
+ },
+ {
+ "epoch": 41.880349344978164,
+ "grad_norm": 0.011011249385774136,
+ "learning_rate": 0.0006,
+ "loss": 7.287247657775879,
+ "step": 3015
+ },
+ {
+ "epoch": 41.8943231441048,
+ "grad_norm": 0.010208295658230782,
+ "learning_rate": 0.0006,
+ "loss": 7.292117118835449,
+ "step": 3016
+ },
+ {
+ "epoch": 41.90829694323144,
+ "grad_norm": 0.008651095442473888,
+ "learning_rate": 0.0006,
+ "loss": 7.302201271057129,
+ "step": 3017
+ },
+ {
+ "epoch": 41.92227074235808,
+ "grad_norm": 0.006692920345813036,
+ "learning_rate": 0.0006,
+ "loss": 7.319568634033203,
+ "step": 3018
+ },
+ {
+ "epoch": 41.93624454148472,
+ "grad_norm": 0.010033348575234413,
+ "learning_rate": 0.0006,
+ "loss": 7.2738566398620605,
+ "step": 3019
+ },
+ {
+ "epoch": 41.95021834061135,
+ "grad_norm": 0.011790832504630089,
+ "learning_rate": 0.0006,
+ "loss": 7.298540115356445,
+ "step": 3020
+ },
+ {
+ "epoch": 41.96419213973799,
+ "grad_norm": 0.018622662872076035,
+ "learning_rate": 0.0006,
+ "loss": 7.267182350158691,
+ "step": 3021
+ },
+ {
+ "epoch": 41.97816593886463,
+ "grad_norm": 0.02520257607102394,
+ "learning_rate": 0.0006,
+ "loss": 7.261566162109375,
+ "step": 3022
+ },
+ {
+ "epoch": 41.992139737991266,
+ "grad_norm": 0.036332644522190094,
+ "learning_rate": 0.0006,
+ "loss": 7.270719051361084,
+ "step": 3023
+ },
+ {
+ "epoch": 42.0,
+ "grad_norm": 0.04717305675148964,
+ "learning_rate": 0.0006,
+ "loss": 7.293332099914551,
+ "step": 3024
+ },
+ {
+ "epoch": 42.0,
+ "eval_loss": 7.354280471801758,
+ "eval_runtime": 56.4353,
+ "eval_samples_per_second": 43.271,
+ "eval_steps_per_second": 1.364,
+ "step": 3024
+ },
+ {
+ "epoch": 42.01397379912664,
+ "grad_norm": 0.06661626696586609,
+ "learning_rate": 0.0006,
+ "loss": 7.338189601898193,
+ "step": 3025
+ },
+ {
+ "epoch": 42.02794759825328,
+ "grad_norm": 0.05452054366469383,
+ "learning_rate": 0.0006,
+ "loss": 7.316764831542969,
+ "step": 3026
+ },
+ {
+ "epoch": 42.041921397379916,
+ "grad_norm": 0.025638775900006294,
+ "learning_rate": 0.0006,
+ "loss": 7.34204626083374,
+ "step": 3027
+ },
+ {
+ "epoch": 42.05589519650655,
+ "grad_norm": 0.022398799657821655,
+ "learning_rate": 0.0006,
+ "loss": 7.262149333953857,
+ "step": 3028
+ },
+ {
+ "epoch": 42.069868995633186,
+ "grad_norm": 0.03713499382138252,
+ "learning_rate": 0.0006,
+ "loss": 7.27695369720459,
+ "step": 3029
+ },
+ {
+ "epoch": 42.083842794759825,
+ "grad_norm": 0.014949247241020203,
+ "learning_rate": 0.0006,
+ "loss": 7.296508312225342,
+ "step": 3030
+ },
+ {
+ "epoch": 42.09781659388646,
+ "grad_norm": 0.02703995071351528,
+ "learning_rate": 0.0006,
+ "loss": 7.258018493652344,
+ "step": 3031
+ },
+ {
+ "epoch": 42.1117903930131,
+ "grad_norm": 0.020352816209197044,
+ "learning_rate": 0.0006,
+ "loss": 7.241450309753418,
+ "step": 3032
+ },
+ {
+ "epoch": 42.12576419213974,
+ "grad_norm": 0.014578479342162609,
+ "learning_rate": 0.0006,
+ "loss": 7.257323265075684,
+ "step": 3033
+ },
+ {
+ "epoch": 42.13973799126637,
+ "grad_norm": 0.026053044945001602,
+ "learning_rate": 0.0006,
+ "loss": 7.254680633544922,
+ "step": 3034
+ },
+ {
+ "epoch": 42.15371179039301,
+ "grad_norm": 0.015120275318622589,
+ "learning_rate": 0.0006,
+ "loss": 7.261074066162109,
+ "step": 3035
+ },
+ {
+ "epoch": 42.16768558951965,
+ "grad_norm": 0.016822177916765213,
+ "learning_rate": 0.0006,
+ "loss": 7.279064178466797,
+ "step": 3036
+ },
+ {
+ "epoch": 42.18165938864629,
+ "grad_norm": 0.016367588192224503,
+ "learning_rate": 0.0006,
+ "loss": 7.251429557800293,
+ "step": 3037
+ },
+ {
+ "epoch": 42.19563318777293,
+ "grad_norm": 0.014350418001413345,
+ "learning_rate": 0.0006,
+ "loss": 7.259264945983887,
+ "step": 3038
+ },
+ {
+ "epoch": 42.209606986899566,
+ "grad_norm": 0.017078228294849396,
+ "learning_rate": 0.0006,
+ "loss": 7.227105140686035,
+ "step": 3039
+ },
+ {
+ "epoch": 42.223580786026204,
+ "grad_norm": 0.011759317480027676,
+ "learning_rate": 0.0006,
+ "loss": 7.275888442993164,
+ "step": 3040
+ },
+ {
+ "epoch": 42.237554585152836,
+ "grad_norm": 0.013019340112805367,
+ "learning_rate": 0.0006,
+ "loss": 7.2517008781433105,
+ "step": 3041
+ },
+ {
+ "epoch": 42.251528384279474,
+ "grad_norm": 0.01178441196680069,
+ "learning_rate": 0.0006,
+ "loss": 7.23576545715332,
+ "step": 3042
+ },
+ {
+ "epoch": 42.26550218340611,
+ "grad_norm": 0.014518891461193562,
+ "learning_rate": 0.0006,
+ "loss": 7.253895282745361,
+ "step": 3043
+ },
+ {
+ "epoch": 42.27947598253275,
+ "grad_norm": 0.018159950152039528,
+ "learning_rate": 0.0006,
+ "loss": 7.278058052062988,
+ "step": 3044
+ },
+ {
+ "epoch": 42.29344978165939,
+ "grad_norm": 0.007587796077132225,
+ "learning_rate": 0.0006,
+ "loss": 7.278578758239746,
+ "step": 3045
+ },
+ {
+ "epoch": 42.30742358078603,
+ "grad_norm": 0.015994979068636894,
+ "learning_rate": 0.0006,
+ "loss": 7.239841461181641,
+ "step": 3046
+ },
+ {
+ "epoch": 42.32139737991266,
+ "grad_norm": 0.017830168828368187,
+ "learning_rate": 0.0006,
+ "loss": 7.2443976402282715,
+ "step": 3047
+ },
+ {
+ "epoch": 42.3353711790393,
+ "grad_norm": 0.008738759905099869,
+ "learning_rate": 0.0006,
+ "loss": 7.232264518737793,
+ "step": 3048
+ },
+ {
+ "epoch": 42.34934497816594,
+ "grad_norm": 0.0175449438393116,
+ "learning_rate": 0.0006,
+ "loss": 7.228000640869141,
+ "step": 3049
+ },
+ {
+ "epoch": 42.36331877729258,
+ "grad_norm": 0.019611820578575134,
+ "learning_rate": 0.0006,
+ "loss": 7.234999179840088,
+ "step": 3050
+ },
+ {
+ "epoch": 42.377292576419215,
+ "grad_norm": 0.015302863903343678,
+ "learning_rate": 0.0006,
+ "loss": 7.242547988891602,
+ "step": 3051
+ },
+ {
+ "epoch": 42.391266375545854,
+ "grad_norm": 0.025919731706380844,
+ "learning_rate": 0.0006,
+ "loss": 7.216652870178223,
+ "step": 3052
+ },
+ {
+ "epoch": 42.40524017467249,
+ "grad_norm": 0.034587468951940536,
+ "learning_rate": 0.0006,
+ "loss": 7.245336532592773,
+ "step": 3053
+ },
+ {
+ "epoch": 42.419213973799124,
+ "grad_norm": 0.019198840484023094,
+ "learning_rate": 0.0006,
+ "loss": 7.256758689880371,
+ "step": 3054
+ },
+ {
+ "epoch": 42.43318777292576,
+ "grad_norm": 0.012316389009356499,
+ "learning_rate": 0.0006,
+ "loss": 7.209102630615234,
+ "step": 3055
+ },
+ {
+ "epoch": 42.4471615720524,
+ "grad_norm": 0.023512162268161774,
+ "learning_rate": 0.0006,
+ "loss": 7.250742435455322,
+ "step": 3056
+ },
+ {
+ "epoch": 42.46113537117904,
+ "grad_norm": 0.025457851588726044,
+ "learning_rate": 0.0006,
+ "loss": 7.263757705688477,
+ "step": 3057
+ },
+ {
+ "epoch": 42.47510917030568,
+ "grad_norm": 0.02192617766559124,
+ "learning_rate": 0.0006,
+ "loss": 7.246977806091309,
+ "step": 3058
+ },
+ {
+ "epoch": 42.48908296943232,
+ "grad_norm": 0.02541281282901764,
+ "learning_rate": 0.0006,
+ "loss": 7.232361316680908,
+ "step": 3059
+ },
+ {
+ "epoch": 42.50305676855895,
+ "grad_norm": 0.036353763192892075,
+ "learning_rate": 0.0006,
+ "loss": 7.242556571960449,
+ "step": 3060
+ },
+ {
+ "epoch": 42.51703056768559,
+ "grad_norm": 0.04555872827768326,
+ "learning_rate": 0.0006,
+ "loss": 7.228940963745117,
+ "step": 3061
+ },
+ {
+ "epoch": 42.531004366812226,
+ "grad_norm": 0.04563378915190697,
+ "learning_rate": 0.0006,
+ "loss": 7.235276222229004,
+ "step": 3062
+ },
+ {
+ "epoch": 42.544978165938865,
+ "grad_norm": 0.04275760427117348,
+ "learning_rate": 0.0006,
+ "loss": 7.238895416259766,
+ "step": 3063
+ },
+ {
+ "epoch": 42.5589519650655,
+ "grad_norm": 0.04535885155200958,
+ "learning_rate": 0.0006,
+ "loss": 7.205336570739746,
+ "step": 3064
+ },
+ {
+ "epoch": 42.57292576419214,
+ "grad_norm": 0.015534617006778717,
+ "learning_rate": 0.0006,
+ "loss": 7.238393783569336,
+ "step": 3065
+ },
+ {
+ "epoch": 42.58689956331878,
+ "grad_norm": 0.03286299481987953,
+ "learning_rate": 0.0006,
+ "loss": 7.216453552246094,
+ "step": 3066
+ },
+ {
+ "epoch": 42.60087336244541,
+ "grad_norm": 0.025453168898820877,
+ "learning_rate": 0.0006,
+ "loss": 7.203750133514404,
+ "step": 3067
+ },
+ {
+ "epoch": 42.61484716157205,
+ "grad_norm": 0.021501649171113968,
+ "learning_rate": 0.0006,
+ "loss": 7.241621971130371,
+ "step": 3068
+ },
+ {
+ "epoch": 42.62882096069869,
+ "grad_norm": 0.02743634767830372,
+ "learning_rate": 0.0006,
+ "loss": 7.22076940536499,
+ "step": 3069
+ },
+ {
+ "epoch": 42.64279475982533,
+ "grad_norm": 0.020890483632683754,
+ "learning_rate": 0.0006,
+ "loss": 7.218412399291992,
+ "step": 3070
+ },
+ {
+ "epoch": 42.65676855895197,
+ "grad_norm": 0.022499850019812584,
+ "learning_rate": 0.0006,
+ "loss": 7.217696189880371,
+ "step": 3071
+ },
+ {
+ "epoch": 42.670742358078606,
+ "grad_norm": 0.01653696969151497,
+ "learning_rate": 0.0006,
+ "loss": 7.156417369842529,
+ "step": 3072
+ },
+ {
+ "epoch": 42.68471615720524,
+ "grad_norm": 0.01834225095808506,
+ "learning_rate": 0.0006,
+ "loss": 7.211247444152832,
+ "step": 3073
+ },
+ {
+ "epoch": 42.698689956331876,
+ "grad_norm": 0.020137229934334755,
+ "learning_rate": 0.0006,
+ "loss": 7.232499599456787,
+ "step": 3074
+ },
+ {
+ "epoch": 42.712663755458514,
+ "grad_norm": 0.012671459466218948,
+ "learning_rate": 0.0006,
+ "loss": 7.172441482543945,
+ "step": 3075
+ },
+ {
+ "epoch": 42.72663755458515,
+ "grad_norm": 0.016803130507469177,
+ "learning_rate": 0.0006,
+ "loss": 7.152103424072266,
+ "step": 3076
+ },
+ {
+ "epoch": 42.74061135371179,
+ "grad_norm": 0.01623712107539177,
+ "learning_rate": 0.0006,
+ "loss": 7.187032699584961,
+ "step": 3077
+ },
+ {
+ "epoch": 42.75458515283843,
+ "grad_norm": 0.015938356518745422,
+ "learning_rate": 0.0006,
+ "loss": 7.171140670776367,
+ "step": 3078
+ },
+ {
+ "epoch": 42.76855895196506,
+ "grad_norm": 0.02580736204981804,
+ "learning_rate": 0.0006,
+ "loss": 7.204472541809082,
+ "step": 3079
+ },
+ {
+ "epoch": 42.7825327510917,
+ "grad_norm": 0.023372117429971695,
+ "learning_rate": 0.0006,
+ "loss": 7.165554523468018,
+ "step": 3080
+ },
+ {
+ "epoch": 42.79650655021834,
+ "grad_norm": 0.04440586268901825,
+ "learning_rate": 0.0006,
+ "loss": 7.169431209564209,
+ "step": 3081
+ },
+ {
+ "epoch": 42.81048034934498,
+ "grad_norm": 0.06797406077384949,
+ "learning_rate": 0.0006,
+ "loss": 7.168193340301514,
+ "step": 3082
+ },
+ {
+ "epoch": 42.82445414847162,
+ "grad_norm": 0.0914321169257164,
+ "learning_rate": 0.0006,
+ "loss": 7.204671859741211,
+ "step": 3083
+ },
+ {
+ "epoch": 42.838427947598255,
+ "grad_norm": 0.0506940595805645,
+ "learning_rate": 0.0006,
+ "loss": 7.15949821472168,
+ "step": 3084
+ },
+ {
+ "epoch": 42.852401746724894,
+ "grad_norm": 0.04819890484213829,
+ "learning_rate": 0.0006,
+ "loss": 7.192837238311768,
+ "step": 3085
+ },
+ {
+ "epoch": 42.866375545851525,
+ "grad_norm": 0.040044233202934265,
+ "learning_rate": 0.0006,
+ "loss": 7.1600165367126465,
+ "step": 3086
+ },
+ {
+ "epoch": 42.880349344978164,
+ "grad_norm": 0.02573299966752529,
+ "learning_rate": 0.0006,
+ "loss": 7.153427600860596,
+ "step": 3087
+ },
+ {
+ "epoch": 42.8943231441048,
+ "grad_norm": 0.046942535787820816,
+ "learning_rate": 0.0006,
+ "loss": 7.16767692565918,
+ "step": 3088
+ },
+ {
+ "epoch": 42.90829694323144,
+ "grad_norm": 0.02683161571621895,
+ "learning_rate": 0.0006,
+ "loss": 7.17144250869751,
+ "step": 3089
+ },
+ {
+ "epoch": 42.92227074235808,
+ "grad_norm": 0.05363674834370613,
+ "learning_rate": 0.0006,
+ "loss": 7.120128631591797,
+ "step": 3090
+ },
+ {
+ "epoch": 42.93624454148472,
+ "grad_norm": 0.023756247013807297,
+ "learning_rate": 0.0006,
+ "loss": 7.169703006744385,
+ "step": 3091
+ },
+ {
+ "epoch": 42.95021834061135,
+ "grad_norm": 0.04216783866286278,
+ "learning_rate": 0.0006,
+ "loss": 7.1206865310668945,
+ "step": 3092
+ },
+ {
+ "epoch": 42.96419213973799,
+ "grad_norm": 0.02219565585255623,
+ "learning_rate": 0.0006,
+ "loss": 7.157215118408203,
+ "step": 3093
+ },
+ {
+ "epoch": 42.97816593886463,
+ "grad_norm": 0.03527127206325531,
+ "learning_rate": 0.0006,
+ "loss": 7.166672706604004,
+ "step": 3094
+ },
+ {
+ "epoch": 42.992139737991266,
+ "grad_norm": 0.017748817801475525,
+ "learning_rate": 0.0006,
+ "loss": 7.154747009277344,
+ "step": 3095
+ },
+ {
+ "epoch": 43.0,
+ "grad_norm": 0.033165838569402695,
+ "learning_rate": 0.0006,
+ "loss": 7.150866508483887,
+ "step": 3096
+ },
+ {
+ "epoch": 43.0,
+ "eval_loss": 7.144403457641602,
+ "eval_runtime": 56.8853,
+ "eval_samples_per_second": 42.928,
+ "eval_steps_per_second": 1.354,
+ "step": 3096
+ },
+ {
+ "epoch": 43.01397379912664,
+ "grad_norm": 0.015199829824268818,
+ "learning_rate": 0.0006,
+ "loss": 7.122735023498535,
+ "step": 3097
+ },
+ {
+ "epoch": 43.02794759825328,
+ "grad_norm": 0.03220256045460701,
+ "learning_rate": 0.0006,
+ "loss": 7.143553733825684,
+ "step": 3098
+ },
+ {
+ "epoch": 43.041921397379916,
+ "grad_norm": 0.014687920920550823,
+ "learning_rate": 0.0006,
+ "loss": 7.080960750579834,
+ "step": 3099
+ },
+ {
+ "epoch": 43.05589519650655,
+ "grad_norm": 0.023522932082414627,
+ "learning_rate": 0.0006,
+ "loss": 7.137356758117676,
+ "step": 3100
+ },
+ {
+ "epoch": 43.069868995633186,
+ "grad_norm": 0.014808216132223606,
+ "learning_rate": 0.0006,
+ "loss": 7.09143590927124,
+ "step": 3101
+ },
+ {
+ "epoch": 43.083842794759825,
+ "grad_norm": 0.029339246451854706,
+ "learning_rate": 0.0006,
+ "loss": 7.122198104858398,
+ "step": 3102
+ },
+ {
+ "epoch": 43.09781659388646,
+ "grad_norm": 0.022953078150749207,
+ "learning_rate": 0.0006,
+ "loss": 7.092505931854248,
+ "step": 3103
+ },
+ {
+ "epoch": 43.1117903930131,
+ "grad_norm": 0.04590052738785744,
+ "learning_rate": 0.0006,
+ "loss": 7.113100051879883,
+ "step": 3104
+ },
+ {
+ "epoch": 43.12576419213974,
+ "grad_norm": 0.0524442233145237,
+ "learning_rate": 0.0006,
+ "loss": 7.147022247314453,
+ "step": 3105
+ },
+ {
+ "epoch": 43.13973799126637,
+ "grad_norm": 0.0315794013440609,
+ "learning_rate": 0.0006,
+ "loss": 7.094045162200928,
+ "step": 3106
+ },
+ {
+ "epoch": 43.15371179039301,
+ "grad_norm": 0.05112722888588905,
+ "learning_rate": 0.0006,
+ "loss": 7.054388523101807,
+ "step": 3107
+ },
+ {
+ "epoch": 43.16768558951965,
+ "grad_norm": 0.03146776929497719,
+ "learning_rate": 0.0006,
+ "loss": 7.116674423217773,
+ "step": 3108
+ },
+ {
+ "epoch": 43.18165938864629,
+ "grad_norm": 0.03391246497631073,
+ "learning_rate": 0.0006,
+ "loss": 7.150345802307129,
+ "step": 3109
+ },
+ {
+ "epoch": 43.19563318777293,
+ "grad_norm": 0.026957903057336807,
+ "learning_rate": 0.0006,
+ "loss": 7.058531284332275,
+ "step": 3110
+ },
+ {
+ "epoch": 43.209606986899566,
+ "grad_norm": 0.041532568633556366,
+ "learning_rate": 0.0006,
+ "loss": 7.085498809814453,
+ "step": 3111
+ },
+ {
+ "epoch": 43.223580786026204,
+ "grad_norm": 0.020001588389277458,
+ "learning_rate": 0.0006,
+ "loss": 7.131487846374512,
+ "step": 3112
+ },
+ {
+ "epoch": 43.237554585152836,
+ "grad_norm": 0.02268712781369686,
+ "learning_rate": 0.0006,
+ "loss": 7.0790863037109375,
+ "step": 3113
+ },
+ {
+ "epoch": 43.251528384279474,
+ "grad_norm": 0.023363925516605377,
+ "learning_rate": 0.0006,
+ "loss": 7.067816734313965,
+ "step": 3114
+ },
+ {
+ "epoch": 43.26550218340611,
+ "grad_norm": 0.019367508590221405,
+ "learning_rate": 0.0006,
+ "loss": 7.042041778564453,
+ "step": 3115
+ },
+ {
+ "epoch": 43.27947598253275,
+ "grad_norm": 0.02745090238749981,
+ "learning_rate": 0.0006,
+ "loss": 7.064733028411865,
+ "step": 3116
+ },
+ {
+ "epoch": 43.29344978165939,
+ "grad_norm": 0.02543802745640278,
+ "learning_rate": 0.0006,
+ "loss": 7.058948040008545,
+ "step": 3117
+ },
+ {
+ "epoch": 43.30742358078603,
+ "grad_norm": 0.018874092027544975,
+ "learning_rate": 0.0006,
+ "loss": 7.009946346282959,
+ "step": 3118
+ },
+ {
+ "epoch": 43.32139737991266,
+ "grad_norm": 0.024662822484970093,
+ "learning_rate": 0.0006,
+ "loss": 7.015913963317871,
+ "step": 3119
+ },
+ {
+ "epoch": 43.3353711790393,
+ "grad_norm": 0.02441416308283806,
+ "learning_rate": 0.0006,
+ "loss": 7.061622142791748,
+ "step": 3120
+ },
+ {
+ "epoch": 43.34934497816594,
+ "grad_norm": 0.031887322664260864,
+ "learning_rate": 0.0006,
+ "loss": 7.046883583068848,
+ "step": 3121
+ },
+ {
+ "epoch": 43.36331877729258,
+ "grad_norm": 0.021930739283561707,
+ "learning_rate": 0.0006,
+ "loss": 7.0309624671936035,
+ "step": 3122
+ },
+ {
+ "epoch": 43.377292576419215,
+ "grad_norm": 0.02722783200442791,
+ "learning_rate": 0.0006,
+ "loss": 7.072924613952637,
+ "step": 3123
+ },
+ {
+ "epoch": 43.391266375545854,
+ "grad_norm": 0.01483179721981287,
+ "learning_rate": 0.0006,
+ "loss": 7.028624534606934,
+ "step": 3124
+ },
+ {
+ "epoch": 43.40524017467249,
+ "grad_norm": 0.024793945252895355,
+ "learning_rate": 0.0006,
+ "loss": 7.001623153686523,
+ "step": 3125
+ },
+ {
+ "epoch": 43.419213973799124,
+ "grad_norm": 0.020148713141679764,
+ "learning_rate": 0.0006,
+ "loss": 6.9938812255859375,
+ "step": 3126
+ },
+ {
+ "epoch": 43.43318777292576,
+ "grad_norm": 0.02413850650191307,
+ "learning_rate": 0.0006,
+ "loss": 7.001243591308594,
+ "step": 3127
+ },
+ {
+ "epoch": 43.4471615720524,
+ "grad_norm": 0.028549203649163246,
+ "learning_rate": 0.0006,
+ "loss": 6.987671375274658,
+ "step": 3128
+ },
+ {
+ "epoch": 43.46113537117904,
+ "grad_norm": 0.025988785549998283,
+ "learning_rate": 0.0006,
+ "loss": 6.9551167488098145,
+ "step": 3129
+ },
+ {
+ "epoch": 43.47510917030568,
+ "grad_norm": 0.027140729129314423,
+ "learning_rate": 0.0006,
+ "loss": 6.9547810554504395,
+ "step": 3130
+ },
+ {
+ "epoch": 43.48908296943232,
+ "grad_norm": 0.030130082741379738,
+ "learning_rate": 0.0006,
+ "loss": 6.993186950683594,
+ "step": 3131
+ },
+ {
+ "epoch": 43.50305676855895,
+ "grad_norm": 0.025996994227170944,
+ "learning_rate": 0.0006,
+ "loss": 6.984624862670898,
+ "step": 3132
+ },
+ {
+ "epoch": 43.51703056768559,
+ "grad_norm": 0.02423008903861046,
+ "learning_rate": 0.0006,
+ "loss": 6.98813009262085,
+ "step": 3133
+ },
+ {
+ "epoch": 43.531004366812226,
+ "grad_norm": 0.029665837064385414,
+ "learning_rate": 0.0006,
+ "loss": 6.986610412597656,
+ "step": 3134
+ },
+ {
+ "epoch": 43.544978165938865,
+ "grad_norm": 0.028768595308065414,
+ "learning_rate": 0.0006,
+ "loss": 6.898797035217285,
+ "step": 3135
+ },
+ {
+ "epoch": 43.5589519650655,
+ "grad_norm": 0.03111208975315094,
+ "learning_rate": 0.0006,
+ "loss": 6.923741817474365,
+ "step": 3136
+ },
+ {
+ "epoch": 43.57292576419214,
+ "grad_norm": 0.0368187390267849,
+ "learning_rate": 0.0006,
+ "loss": 6.994343280792236,
+ "step": 3137
+ },
+ {
+ "epoch": 43.58689956331878,
+ "grad_norm": 0.04076763987541199,
+ "learning_rate": 0.0006,
+ "loss": 6.909458160400391,
+ "step": 3138
+ },
+ {
+ "epoch": 43.60087336244541,
+ "grad_norm": 0.03281061723828316,
+ "learning_rate": 0.0006,
+ "loss": 6.924851417541504,
+ "step": 3139
+ },
+ {
+ "epoch": 43.61484716157205,
+ "grad_norm": 0.025486988946795464,
+ "learning_rate": 0.0006,
+ "loss": 6.962926864624023,
+ "step": 3140
+ },
+ {
+ "epoch": 43.62882096069869,
+ "grad_norm": 0.04583485797047615,
+ "learning_rate": 0.0006,
+ "loss": 6.876555442810059,
+ "step": 3141
+ },
+ {
+ "epoch": 43.64279475982533,
+ "grad_norm": 0.05460785701870918,
+ "learning_rate": 0.0006,
+ "loss": 6.952932357788086,
+ "step": 3142
+ },
+ {
+ "epoch": 43.65676855895197,
+ "grad_norm": 0.040917906910181046,
+ "learning_rate": 0.0006,
+ "loss": 6.9624924659729,
+ "step": 3143
+ },
+ {
+ "epoch": 43.670742358078606,
+ "grad_norm": 0.04009668901562691,
+ "learning_rate": 0.0006,
+ "loss": 6.94033145904541,
+ "step": 3144
+ },
+ {
+ "epoch": 43.68471615720524,
+ "grad_norm": 0.04500625655055046,
+ "learning_rate": 0.0006,
+ "loss": 6.961839199066162,
+ "step": 3145
+ },
+ {
+ "epoch": 43.698689956331876,
+ "grad_norm": 0.03182258829474449,
+ "learning_rate": 0.0006,
+ "loss": 6.930580139160156,
+ "step": 3146
+ },
+ {
+ "epoch": 43.712663755458514,
+ "grad_norm": 0.053621504455804825,
+ "learning_rate": 0.0006,
+ "loss": 6.950235366821289,
+ "step": 3147
+ },
+ {
+ "epoch": 43.72663755458515,
+ "grad_norm": 0.031919289380311966,
+ "learning_rate": 0.0006,
+ "loss": 6.895111560821533,
+ "step": 3148
+ },
+ {
+ "epoch": 43.74061135371179,
+ "grad_norm": 0.042713891714811325,
+ "learning_rate": 0.0006,
+ "loss": 6.940885543823242,
+ "step": 3149
+ },
+ {
+ "epoch": 43.75458515283843,
+ "grad_norm": 0.02385653182864189,
+ "learning_rate": 0.0006,
+ "loss": 6.948801517486572,
+ "step": 3150
+ },
+ {
+ "epoch": 43.76855895196506,
+ "grad_norm": 0.04100663587450981,
+ "learning_rate": 0.0006,
+ "loss": 6.970411777496338,
+ "step": 3151
+ },
+ {
+ "epoch": 43.7825327510917,
+ "grad_norm": 0.02729583904147148,
+ "learning_rate": 0.0006,
+ "loss": 6.899379253387451,
+ "step": 3152
+ },
+ {
+ "epoch": 43.79650655021834,
+ "grad_norm": 0.03128620237112045,
+ "learning_rate": 0.0006,
+ "loss": 6.931319236755371,
+ "step": 3153
+ },
+ {
+ "epoch": 43.81048034934498,
+ "grad_norm": 0.025488659739494324,
+ "learning_rate": 0.0006,
+ "loss": 6.8486409187316895,
+ "step": 3154
+ },
+ {
+ "epoch": 43.82445414847162,
+ "grad_norm": 0.03679376840591431,
+ "learning_rate": 0.0006,
+ "loss": 6.840811252593994,
+ "step": 3155
+ },
+ {
+ "epoch": 43.838427947598255,
+ "grad_norm": 0.020578034222126007,
+ "learning_rate": 0.0006,
+ "loss": 6.9088616371154785,
+ "step": 3156
+ },
+ {
+ "epoch": 43.852401746724894,
+ "grad_norm": 0.02935841865837574,
+ "learning_rate": 0.0006,
+ "loss": 6.806676864624023,
+ "step": 3157
+ },
+ {
+ "epoch": 43.866375545851525,
+ "grad_norm": 0.023319244384765625,
+ "learning_rate": 0.0006,
+ "loss": 6.878457546234131,
+ "step": 3158
+ },
+ {
+ "epoch": 43.880349344978164,
+ "grad_norm": 0.02779589593410492,
+ "learning_rate": 0.0006,
+ "loss": 6.822447776794434,
+ "step": 3159
+ },
+ {
+ "epoch": 43.8943231441048,
+ "grad_norm": 0.01879825070500374,
+ "learning_rate": 0.0006,
+ "loss": 6.854121208190918,
+ "step": 3160
+ },
+ {
+ "epoch": 43.90829694323144,
+ "grad_norm": 0.023286139592528343,
+ "learning_rate": 0.0006,
+ "loss": 6.836827278137207,
+ "step": 3161
+ },
+ {
+ "epoch": 43.92227074235808,
+ "grad_norm": 0.030358297750353813,
+ "learning_rate": 0.0006,
+ "loss": 6.860037803649902,
+ "step": 3162
+ },
+ {
+ "epoch": 43.93624454148472,
+ "grad_norm": 0.029621105641126633,
+ "learning_rate": 0.0006,
+ "loss": 6.884025573730469,
+ "step": 3163
+ },
+ {
+ "epoch": 43.95021834061135,
+ "grad_norm": 0.029829824343323708,
+ "learning_rate": 0.0006,
+ "loss": 6.8702392578125,
+ "step": 3164
+ },
+ {
+ "epoch": 43.96419213973799,
+ "grad_norm": 0.036892473697662354,
+ "learning_rate": 0.0006,
+ "loss": 6.806652069091797,
+ "step": 3165
+ },
+ {
+ "epoch": 43.97816593886463,
+ "grad_norm": 0.03206709027290344,
+ "learning_rate": 0.0006,
+ "loss": 6.828495979309082,
+ "step": 3166
+ },
+ {
+ "epoch": 43.992139737991266,
+ "grad_norm": 0.02744000405073166,
+ "learning_rate": 0.0006,
+ "loss": 6.832002639770508,
+ "step": 3167
+ },
+ {
+ "epoch": 44.0,
+ "grad_norm": 0.02205660194158554,
+ "learning_rate": 0.0006,
+ "loss": 6.898828506469727,
+ "step": 3168
+ },
+ {
+ "epoch": 44.0,
+ "eval_loss": 6.8150553703308105,
+ "eval_runtime": 56.784,
+ "eval_samples_per_second": 43.005,
+ "eval_steps_per_second": 1.356,
+ "step": 3168
+ },
+ {
+ "epoch": 44.01397379912664,
+ "grad_norm": 0.030421260744333267,
+ "learning_rate": 0.0006,
+ "loss": 6.775057792663574,
+ "step": 3169
+ },
+ {
+ "epoch": 44.02794759825328,
+ "grad_norm": 0.0465611033141613,
+ "learning_rate": 0.0006,
+ "loss": 6.849390506744385,
+ "step": 3170
+ },
+ {
+ "epoch": 44.041921397379916,
+ "grad_norm": 0.05977972596883774,
+ "learning_rate": 0.0006,
+ "loss": 6.725576400756836,
+ "step": 3171
+ },
+ {
+ "epoch": 44.05589519650655,
+ "grad_norm": 0.05432697385549545,
+ "learning_rate": 0.0006,
+ "loss": 6.768024444580078,
+ "step": 3172
+ },
+ {
+ "epoch": 44.069868995633186,
+ "grad_norm": 0.04850967228412628,
+ "learning_rate": 0.0006,
+ "loss": 6.799457550048828,
+ "step": 3173
+ },
+ {
+ "epoch": 44.083842794759825,
+ "grad_norm": 0.06201786920428276,
+ "learning_rate": 0.0006,
+ "loss": 6.897234916687012,
+ "step": 3174
+ },
+ {
+ "epoch": 44.09781659388646,
+ "grad_norm": 0.04695591703057289,
+ "learning_rate": 0.0006,
+ "loss": 6.826470375061035,
+ "step": 3175
+ },
+ {
+ "epoch": 44.1117903930131,
+ "grad_norm": 0.03850164636969566,
+ "learning_rate": 0.0006,
+ "loss": 6.835901260375977,
+ "step": 3176
+ },
+ {
+ "epoch": 44.12576419213974,
+ "grad_norm": 0.04250155761837959,
+ "learning_rate": 0.0006,
+ "loss": 6.782310485839844,
+ "step": 3177
+ },
+ {
+ "epoch": 44.13973799126637,
+ "grad_norm": 0.041465479880571365,
+ "learning_rate": 0.0006,
+ "loss": 6.785732746124268,
+ "step": 3178
+ },
+ {
+ "epoch": 44.15371179039301,
+ "grad_norm": 0.02527807280421257,
+ "learning_rate": 0.0006,
+ "loss": 6.78061580657959,
+ "step": 3179
+ },
+ {
+ "epoch": 44.16768558951965,
+ "grad_norm": 0.03660374879837036,
+ "learning_rate": 0.0006,
+ "loss": 6.783992767333984,
+ "step": 3180
+ },
+ {
+ "epoch": 44.18165938864629,
+ "grad_norm": 0.027567964047193527,
+ "learning_rate": 0.0006,
+ "loss": 6.765413284301758,
+ "step": 3181
+ },
+ {
+ "epoch": 44.19563318777293,
+ "grad_norm": 0.024301424622535706,
+ "learning_rate": 0.0006,
+ "loss": 6.760368347167969,
+ "step": 3182
+ },
+ {
+ "epoch": 44.209606986899566,
+ "grad_norm": 0.0266821701079607,
+ "learning_rate": 0.0006,
+ "loss": 6.71417236328125,
+ "step": 3183
+ },
+ {
+ "epoch": 44.223580786026204,
+ "grad_norm": 0.028125090524554253,
+ "learning_rate": 0.0006,
+ "loss": 6.646126747131348,
+ "step": 3184
+ },
+ {
+ "epoch": 44.237554585152836,
+ "grad_norm": 0.0185737032443285,
+ "learning_rate": 0.0006,
+ "loss": 6.712944030761719,
+ "step": 3185
+ },
+ {
+ "epoch": 44.251528384279474,
+ "grad_norm": 0.025836674496531487,
+ "learning_rate": 0.0006,
+ "loss": 6.707030296325684,
+ "step": 3186
+ },
+ {
+ "epoch": 44.26550218340611,
+ "grad_norm": 0.029744870960712433,
+ "learning_rate": 0.0006,
+ "loss": 6.6271071434021,
+ "step": 3187
+ },
+ {
+ "epoch": 44.27947598253275,
+ "grad_norm": 0.03021426685154438,
+ "learning_rate": 0.0006,
+ "loss": 6.677826404571533,
+ "step": 3188
+ },
+ {
+ "epoch": 44.29344978165939,
+ "grad_norm": 0.032727014273405075,
+ "learning_rate": 0.0006,
+ "loss": 6.657764434814453,
+ "step": 3189
+ },
+ {
+ "epoch": 44.30742358078603,
+ "grad_norm": 0.03696778789162636,
+ "learning_rate": 0.0006,
+ "loss": 6.632883548736572,
+ "step": 3190
+ },
+ {
+ "epoch": 44.32139737991266,
+ "grad_norm": 0.034630827605724335,
+ "learning_rate": 0.0006,
+ "loss": 6.759583473205566,
+ "step": 3191
+ },
+ {
+ "epoch": 44.3353711790393,
+ "grad_norm": 0.030517758801579475,
+ "learning_rate": 0.0006,
+ "loss": 6.702268600463867,
+ "step": 3192
+ },
+ {
+ "epoch": 44.34934497816594,
+ "grad_norm": 0.04105190932750702,
+ "learning_rate": 0.0006,
+ "loss": 6.705792427062988,
+ "step": 3193
+ },
+ {
+ "epoch": 44.36331877729258,
+ "grad_norm": 0.06659174710512161,
+ "learning_rate": 0.0006,
+ "loss": 6.73862361907959,
+ "step": 3194
+ },
+ {
+ "epoch": 44.377292576419215,
+ "grad_norm": 0.08013619482517242,
+ "learning_rate": 0.0006,
+ "loss": 6.710282802581787,
+ "step": 3195
+ },
+ {
+ "epoch": 44.391266375545854,
+ "grad_norm": 0.049850963056087494,
+ "learning_rate": 0.0006,
+ "loss": 6.684974670410156,
+ "step": 3196
+ },
+ {
+ "epoch": 44.40524017467249,
+ "grad_norm": 0.061247408390045166,
+ "learning_rate": 0.0006,
+ "loss": 6.716034412384033,
+ "step": 3197
+ },
+ {
+ "epoch": 44.419213973799124,
+ "grad_norm": 0.05868857726454735,
+ "learning_rate": 0.0006,
+ "loss": 6.589107513427734,
+ "step": 3198
+ },
+ {
+ "epoch": 44.43318777292576,
+ "grad_norm": 0.048878561705350876,
+ "learning_rate": 0.0006,
+ "loss": 6.694855690002441,
+ "step": 3199
+ },
+ {
+ "epoch": 44.4471615720524,
+ "grad_norm": 0.028748324140906334,
+ "learning_rate": 0.0006,
+ "loss": 6.629969596862793,
+ "step": 3200
+ },
+ {
+ "epoch": 44.46113537117904,
+ "grad_norm": 0.04060306400060654,
+ "learning_rate": 0.0006,
+ "loss": 6.721431255340576,
+ "step": 3201
+ },
+ {
+ "epoch": 44.47510917030568,
+ "grad_norm": 0.03833824768662453,
+ "learning_rate": 0.0006,
+ "loss": 6.625029563903809,
+ "step": 3202
+ },
+ {
+ "epoch": 44.48908296943232,
+ "grad_norm": 0.03791997581720352,
+ "learning_rate": 0.0006,
+ "loss": 6.669187545776367,
+ "step": 3203
+ },
+ {
+ "epoch": 44.50305676855895,
+ "grad_norm": 0.03259684890508652,
+ "learning_rate": 0.0006,
+ "loss": 6.665643692016602,
+ "step": 3204
+ },
+ {
+ "epoch": 44.51703056768559,
+ "grad_norm": 0.040658626705408096,
+ "learning_rate": 0.0006,
+ "loss": 6.639796257019043,
+ "step": 3205
+ },
+ {
+ "epoch": 44.531004366812226,
+ "grad_norm": 0.03691193461418152,
+ "learning_rate": 0.0006,
+ "loss": 6.692244529724121,
+ "step": 3206
+ },
+ {
+ "epoch": 44.544978165938865,
+ "grad_norm": 0.02814812771975994,
+ "learning_rate": 0.0006,
+ "loss": 6.61978006362915,
+ "step": 3207
+ },
+ {
+ "epoch": 44.5589519650655,
+ "grad_norm": 0.022155074402689934,
+ "learning_rate": 0.0006,
+ "loss": 6.625905990600586,
+ "step": 3208
+ },
+ {
+ "epoch": 44.57292576419214,
+ "grad_norm": 0.030794885009527206,
+ "learning_rate": 0.0006,
+ "loss": 6.678500652313232,
+ "step": 3209
+ },
+ {
+ "epoch": 44.58689956331878,
+ "grad_norm": 0.03201768547296524,
+ "learning_rate": 0.0006,
+ "loss": 6.6015729904174805,
+ "step": 3210
+ },
+ {
+ "epoch": 44.60087336244541,
+ "grad_norm": 0.025138361379504204,
+ "learning_rate": 0.0006,
+ "loss": 6.568234443664551,
+ "step": 3211
+ },
+ {
+ "epoch": 44.61484716157205,
+ "grad_norm": 0.02969353087246418,
+ "learning_rate": 0.0006,
+ "loss": 6.588366508483887,
+ "step": 3212
+ },
+ {
+ "epoch": 44.62882096069869,
+ "grad_norm": 0.03136962279677391,
+ "learning_rate": 0.0006,
+ "loss": 6.479961395263672,
+ "step": 3213
+ },
+ {
+ "epoch": 44.64279475982533,
+ "grad_norm": 0.030178170651197433,
+ "learning_rate": 0.0006,
+ "loss": 6.622910976409912,
+ "step": 3214
+ },
+ {
+ "epoch": 44.65676855895197,
+ "grad_norm": 0.02682245895266533,
+ "learning_rate": 0.0006,
+ "loss": 6.597707748413086,
+ "step": 3215
+ },
+ {
+ "epoch": 44.670742358078606,
+ "grad_norm": 0.025154557079076767,
+ "learning_rate": 0.0006,
+ "loss": 6.630541801452637,
+ "step": 3216
+ },
+ {
+ "epoch": 44.68471615720524,
+ "grad_norm": 0.03564463183283806,
+ "learning_rate": 0.0006,
+ "loss": 6.542823791503906,
+ "step": 3217
+ },
+ {
+ "epoch": 44.698689956331876,
+ "grad_norm": 0.049477651715278625,
+ "learning_rate": 0.0006,
+ "loss": 6.646030426025391,
+ "step": 3218
+ },
+ {
+ "epoch": 44.712663755458514,
+ "grad_norm": 0.06536535918712616,
+ "learning_rate": 0.0006,
+ "loss": 6.595463275909424,
+ "step": 3219
+ },
+ {
+ "epoch": 44.72663755458515,
+ "grad_norm": 0.07710316777229309,
+ "learning_rate": 0.0006,
+ "loss": 6.55524206161499,
+ "step": 3220
+ },
+ {
+ "epoch": 44.74061135371179,
+ "grad_norm": 0.09571163356304169,
+ "learning_rate": 0.0006,
+ "loss": 6.659764289855957,
+ "step": 3221
+ },
+ {
+ "epoch": 44.75458515283843,
+ "grad_norm": 0.055076830089092255,
+ "learning_rate": 0.0006,
+ "loss": 6.658882141113281,
+ "step": 3222
+ },
+ {
+ "epoch": 44.76855895196506,
+ "grad_norm": 0.07671099156141281,
+ "learning_rate": 0.0006,
+ "loss": 6.605809211730957,
+ "step": 3223
+ },
+ {
+ "epoch": 44.7825327510917,
+ "grad_norm": 0.05295584350824356,
+ "learning_rate": 0.0006,
+ "loss": 6.641455173492432,
+ "step": 3224
+ },
+ {
+ "epoch": 44.79650655021834,
+ "grad_norm": 0.09128042310476303,
+ "learning_rate": 0.0006,
+ "loss": 6.691066741943359,
+ "step": 3225
+ },
+ {
+ "epoch": 44.81048034934498,
+ "grad_norm": 0.05362958461046219,
+ "learning_rate": 0.0006,
+ "loss": 6.573406219482422,
+ "step": 3226
+ },
+ {
+ "epoch": 44.82445414847162,
+ "grad_norm": 0.046910062432289124,
+ "learning_rate": 0.0006,
+ "loss": 6.550876617431641,
+ "step": 3227
+ },
+ {
+ "epoch": 44.838427947598255,
+ "grad_norm": 0.03494197130203247,
+ "learning_rate": 0.0006,
+ "loss": 6.540179252624512,
+ "step": 3228
+ },
+ {
+ "epoch": 44.852401746724894,
+ "grad_norm": 0.04716387391090393,
+ "learning_rate": 0.0006,
+ "loss": 6.490459442138672,
+ "step": 3229
+ },
+ {
+ "epoch": 44.866375545851525,
+ "grad_norm": 0.04190506786108017,
+ "learning_rate": 0.0006,
+ "loss": 6.509925365447998,
+ "step": 3230
+ },
+ {
+ "epoch": 44.880349344978164,
+ "grad_norm": 0.04749064892530441,
+ "learning_rate": 0.0006,
+ "loss": 6.478779315948486,
+ "step": 3231
+ },
+ {
+ "epoch": 44.8943231441048,
+ "grad_norm": 0.028002185747027397,
+ "learning_rate": 0.0006,
+ "loss": 6.528904914855957,
+ "step": 3232
+ },
+ {
+ "epoch": 44.90829694323144,
+ "grad_norm": 0.036473166197538376,
+ "learning_rate": 0.0006,
+ "loss": 6.549042701721191,
+ "step": 3233
+ },
+ {
+ "epoch": 44.92227074235808,
+ "grad_norm": 0.02543315477669239,
+ "learning_rate": 0.0006,
+ "loss": 6.498875141143799,
+ "step": 3234
+ },
+ {
+ "epoch": 44.93624454148472,
+ "grad_norm": 0.029906807467341423,
+ "learning_rate": 0.0006,
+ "loss": 6.453019142150879,
+ "step": 3235
+ },
+ {
+ "epoch": 44.95021834061135,
+ "grad_norm": 0.027231857180595398,
+ "learning_rate": 0.0006,
+ "loss": 6.428829193115234,
+ "step": 3236
+ },
+ {
+ "epoch": 44.96419213973799,
+ "grad_norm": 0.02506154775619507,
+ "learning_rate": 0.0006,
+ "loss": 6.420858383178711,
+ "step": 3237
+ },
+ {
+ "epoch": 44.97816593886463,
+ "grad_norm": 0.0269472673535347,
+ "learning_rate": 0.0006,
+ "loss": 6.484989643096924,
+ "step": 3238
+ },
+ {
+ "epoch": 44.992139737991266,
+ "grad_norm": 0.023410487920045853,
+ "learning_rate": 0.0006,
+ "loss": 6.454830169677734,
+ "step": 3239
+ },
+ {
+ "epoch": 45.0,
+ "grad_norm": 0.034175023436546326,
+ "learning_rate": 0.0006,
+ "loss": 6.430975437164307,
+ "step": 3240
+ },
+ {
+ "epoch": 45.0,
+ "eval_loss": 6.458896160125732,
+ "eval_runtime": 56.6934,
+ "eval_samples_per_second": 43.074,
+ "eval_steps_per_second": 1.358,
+ "step": 3240
+ },
+ {
+ "epoch": 45.01397379912664,
+ "grad_norm": 0.045031629502773285,
+ "learning_rate": 0.0006,
+ "loss": 6.4243011474609375,
+ "step": 3241
+ },
+ {
+ "epoch": 45.02794759825328,
+ "grad_norm": 0.04968734085559845,
+ "learning_rate": 0.0006,
+ "loss": 6.380091667175293,
+ "step": 3242
+ },
+ {
+ "epoch": 45.041921397379916,
+ "grad_norm": 0.047145187854766846,
+ "learning_rate": 0.0006,
+ "loss": 6.477178573608398,
+ "step": 3243
+ },
+ {
+ "epoch": 45.05589519650655,
+ "grad_norm": 0.04304948449134827,
+ "learning_rate": 0.0006,
+ "loss": 6.399726867675781,
+ "step": 3244
+ },
+ {
+ "epoch": 45.069868995633186,
+ "grad_norm": 0.04466010630130768,
+ "learning_rate": 0.0006,
+ "loss": 6.3415679931640625,
+ "step": 3245
+ },
+ {
+ "epoch": 45.083842794759825,
+ "grad_norm": 0.04305306077003479,
+ "learning_rate": 0.0006,
+ "loss": 6.3413310050964355,
+ "step": 3246
+ },
+ {
+ "epoch": 45.09781659388646,
+ "grad_norm": 0.043064575642347336,
+ "learning_rate": 0.0006,
+ "loss": 6.4838714599609375,
+ "step": 3247
+ },
+ {
+ "epoch": 45.1117903930131,
+ "grad_norm": 0.049503277987241745,
+ "learning_rate": 0.0006,
+ "loss": 6.38473653793335,
+ "step": 3248
+ },
+ {
+ "epoch": 45.12576419213974,
+ "grad_norm": 0.04613238573074341,
+ "learning_rate": 0.0006,
+ "loss": 6.370769500732422,
+ "step": 3249
+ },
+ {
+ "epoch": 45.13973799126637,
+ "grad_norm": 0.052119478583335876,
+ "learning_rate": 0.0006,
+ "loss": 6.362321853637695,
+ "step": 3250
+ },
+ {
+ "epoch": 45.15371179039301,
+ "grad_norm": 0.045110076665878296,
+ "learning_rate": 0.0006,
+ "loss": 6.369844913482666,
+ "step": 3251
+ },
+ {
+ "epoch": 45.16768558951965,
+ "grad_norm": 0.03798583149909973,
+ "learning_rate": 0.0006,
+ "loss": 6.324633598327637,
+ "step": 3252
+ },
+ {
+ "epoch": 45.18165938864629,
+ "grad_norm": 0.03744659945368767,
+ "learning_rate": 0.0006,
+ "loss": 6.349421501159668,
+ "step": 3253
+ },
+ {
+ "epoch": 45.19563318777293,
+ "grad_norm": 0.03767292574048042,
+ "learning_rate": 0.0006,
+ "loss": 6.334238052368164,
+ "step": 3254
+ },
+ {
+ "epoch": 45.209606986899566,
+ "grad_norm": 0.03577509894967079,
+ "learning_rate": 0.0006,
+ "loss": 6.3453826904296875,
+ "step": 3255
+ },
+ {
+ "epoch": 45.223580786026204,
+ "grad_norm": 0.03403240442276001,
+ "learning_rate": 0.0006,
+ "loss": 6.2836174964904785,
+ "step": 3256
+ },
+ {
+ "epoch": 45.237554585152836,
+ "grad_norm": 0.03517331928014755,
+ "learning_rate": 0.0006,
+ "loss": 6.281718730926514,
+ "step": 3257
+ },
+ {
+ "epoch": 45.251528384279474,
+ "grad_norm": 0.04409139230847359,
+ "learning_rate": 0.0006,
+ "loss": 6.320815086364746,
+ "step": 3258
+ },
+ {
+ "epoch": 45.26550218340611,
+ "grad_norm": 0.054750364273786545,
+ "learning_rate": 0.0006,
+ "loss": 6.343618392944336,
+ "step": 3259
+ },
+ {
+ "epoch": 45.27947598253275,
+ "grad_norm": 0.07612067461013794,
+ "learning_rate": 0.0006,
+ "loss": 6.310614585876465,
+ "step": 3260
+ },
+ {
+ "epoch": 45.29344978165939,
+ "grad_norm": 0.08438985049724579,
+ "learning_rate": 0.0006,
+ "loss": 6.2369184494018555,
+ "step": 3261
+ },
+ {
+ "epoch": 45.30742358078603,
+ "grad_norm": 0.0782201811671257,
+ "learning_rate": 0.0006,
+ "loss": 6.254452705383301,
+ "step": 3262
+ },
+ {
+ "epoch": 45.32139737991266,
+ "grad_norm": 0.09928355365991592,
+ "learning_rate": 0.0006,
+ "loss": 6.3112077713012695,
+ "step": 3263
+ },
+ {
+ "epoch": 45.3353711790393,
+ "grad_norm": 0.07233133912086487,
+ "learning_rate": 0.0006,
+ "loss": 6.408087253570557,
+ "step": 3264
+ },
+ {
+ "epoch": 45.34934497816594,
+ "grad_norm": 0.07478684186935425,
+ "learning_rate": 0.0006,
+ "loss": 6.3770036697387695,
+ "step": 3265
+ },
+ {
+ "epoch": 45.36331877729258,
+ "grad_norm": 0.06036846712231636,
+ "learning_rate": 0.0006,
+ "loss": 6.33865213394165,
+ "step": 3266
+ },
+ {
+ "epoch": 45.377292576419215,
+ "grad_norm": 0.05783458054065704,
+ "learning_rate": 0.0006,
+ "loss": 6.363203048706055,
+ "step": 3267
+ },
+ {
+ "epoch": 45.391266375545854,
+ "grad_norm": 0.05443909019231796,
+ "learning_rate": 0.0006,
+ "loss": 6.287069797515869,
+ "step": 3268
+ },
+ {
+ "epoch": 45.40524017467249,
+ "grad_norm": 0.0557909794151783,
+ "learning_rate": 0.0006,
+ "loss": 6.297784805297852,
+ "step": 3269
+ },
+ {
+ "epoch": 45.419213973799124,
+ "grad_norm": 0.03483372926712036,
+ "learning_rate": 0.0006,
+ "loss": 6.280734062194824,
+ "step": 3270
+ },
+ {
+ "epoch": 45.43318777292576,
+ "grad_norm": 0.0425543375313282,
+ "learning_rate": 0.0006,
+ "loss": 6.33565616607666,
+ "step": 3271
+ },
+ {
+ "epoch": 45.4471615720524,
+ "grad_norm": 0.03230159357190132,
+ "learning_rate": 0.0006,
+ "loss": 6.197698593139648,
+ "step": 3272
+ },
+ {
+ "epoch": 45.46113537117904,
+ "grad_norm": 0.036897290498018265,
+ "learning_rate": 0.0006,
+ "loss": 6.241239547729492,
+ "step": 3273
+ },
+ {
+ "epoch": 45.47510917030568,
+ "grad_norm": 0.031103838235139847,
+ "learning_rate": 0.0006,
+ "loss": 6.159435749053955,
+ "step": 3274
+ },
+ {
+ "epoch": 45.48908296943232,
+ "grad_norm": 0.027116449549794197,
+ "learning_rate": 0.0006,
+ "loss": 6.212225914001465,
+ "step": 3275
+ },
+ {
+ "epoch": 45.50305676855895,
+ "grad_norm": 0.029923981055617332,
+ "learning_rate": 0.0006,
+ "loss": 6.157269477844238,
+ "step": 3276
+ },
+ {
+ "epoch": 45.51703056768559,
+ "grad_norm": 0.02521538734436035,
+ "learning_rate": 0.0006,
+ "loss": 6.146918773651123,
+ "step": 3277
+ },
+ {
+ "epoch": 45.531004366812226,
+ "grad_norm": 0.02758920192718506,
+ "learning_rate": 0.0006,
+ "loss": 6.0446977615356445,
+ "step": 3278
+ },
+ {
+ "epoch": 45.544978165938865,
+ "grad_norm": 0.024328220635652542,
+ "learning_rate": 0.0006,
+ "loss": 6.099508285522461,
+ "step": 3279
+ },
+ {
+ "epoch": 45.5589519650655,
+ "grad_norm": 0.027425771579146385,
+ "learning_rate": 0.0006,
+ "loss": 6.084299087524414,
+ "step": 3280
+ },
+ {
+ "epoch": 45.57292576419214,
+ "grad_norm": 0.026927651837468147,
+ "learning_rate": 0.0006,
+ "loss": 6.146334171295166,
+ "step": 3281
+ },
+ {
+ "epoch": 45.58689956331878,
+ "grad_norm": 0.04054379090666771,
+ "learning_rate": 0.0006,
+ "loss": 6.1301493644714355,
+ "step": 3282
+ },
+ {
+ "epoch": 45.60087336244541,
+ "grad_norm": 0.04456290975213051,
+ "learning_rate": 0.0006,
+ "loss": 6.152511119842529,
+ "step": 3283
+ },
+ {
+ "epoch": 45.61484716157205,
+ "grad_norm": 0.052821505814790726,
+ "learning_rate": 0.0006,
+ "loss": 6.113544940948486,
+ "step": 3284
+ },
+ {
+ "epoch": 45.62882096069869,
+ "grad_norm": 0.07251739501953125,
+ "learning_rate": 0.0006,
+ "loss": 6.118298530578613,
+ "step": 3285
+ },
+ {
+ "epoch": 45.64279475982533,
+ "grad_norm": 0.09433556348085403,
+ "learning_rate": 0.0006,
+ "loss": 6.170897483825684,
+ "step": 3286
+ },
+ {
+ "epoch": 45.65676855895197,
+ "grad_norm": 0.057683493942022324,
+ "learning_rate": 0.0006,
+ "loss": 6.1747822761535645,
+ "step": 3287
+ },
+ {
+ "epoch": 45.670742358078606,
+ "grad_norm": 0.06363270431756973,
+ "learning_rate": 0.0006,
+ "loss": 6.1617889404296875,
+ "step": 3288
+ },
+ {
+ "epoch": 45.68471615720524,
+ "grad_norm": 0.07528169453144073,
+ "learning_rate": 0.0006,
+ "loss": 6.230060577392578,
+ "step": 3289
+ },
+ {
+ "epoch": 45.698689956331876,
+ "grad_norm": 0.07816170156002045,
+ "learning_rate": 0.0006,
+ "loss": 6.171257019042969,
+ "step": 3290
+ },
+ {
+ "epoch": 45.712663755458514,
+ "grad_norm": 0.05860164389014244,
+ "learning_rate": 0.0006,
+ "loss": 6.095484256744385,
+ "step": 3291
+ },
+ {
+ "epoch": 45.72663755458515,
+ "grad_norm": 0.04747429117560387,
+ "learning_rate": 0.0006,
+ "loss": 6.060067176818848,
+ "step": 3292
+ },
+ {
+ "epoch": 45.74061135371179,
+ "grad_norm": 0.04514474794268608,
+ "learning_rate": 0.0006,
+ "loss": 6.100218772888184,
+ "step": 3293
+ },
+ {
+ "epoch": 45.75458515283843,
+ "grad_norm": 0.05168681964278221,
+ "learning_rate": 0.0006,
+ "loss": 6.0621867179870605,
+ "step": 3294
+ },
+ {
+ "epoch": 45.76855895196506,
+ "grad_norm": 0.04665876924991608,
+ "learning_rate": 0.0006,
+ "loss": 6.118292808532715,
+ "step": 3295
+ },
+ {
+ "epoch": 45.7825327510917,
+ "grad_norm": 0.049374762922525406,
+ "learning_rate": 0.0006,
+ "loss": 6.098394393920898,
+ "step": 3296
+ },
+ {
+ "epoch": 45.79650655021834,
+ "grad_norm": 0.03793549910187721,
+ "learning_rate": 0.0006,
+ "loss": 6.069541931152344,
+ "step": 3297
+ },
+ {
+ "epoch": 45.81048034934498,
+ "grad_norm": 0.036362383514642715,
+ "learning_rate": 0.0006,
+ "loss": 6.1188435554504395,
+ "step": 3298
+ },
+ {
+ "epoch": 45.82445414847162,
+ "grad_norm": 0.03397097811102867,
+ "learning_rate": 0.0006,
+ "loss": 6.117589950561523,
+ "step": 3299
+ },
+ {
+ "epoch": 45.838427947598255,
+ "grad_norm": 0.030977753922343254,
+ "learning_rate": 0.0006,
+ "loss": 6.0236711502075195,
+ "step": 3300
+ },
+ {
+ "epoch": 45.852401746724894,
+ "grad_norm": 0.02800217643380165,
+ "learning_rate": 0.0006,
+ "loss": 6.034825325012207,
+ "step": 3301
+ },
+ {
+ "epoch": 45.866375545851525,
+ "grad_norm": 0.03129778802394867,
+ "learning_rate": 0.0006,
+ "loss": 5.980271339416504,
+ "step": 3302
+ },
+ {
+ "epoch": 45.880349344978164,
+ "grad_norm": 0.025886094197630882,
+ "learning_rate": 0.0006,
+ "loss": 5.966419219970703,
+ "step": 3303
+ },
+ {
+ "epoch": 45.8943231441048,
+ "grad_norm": 0.025712328031659126,
+ "learning_rate": 0.0006,
+ "loss": 5.938652515411377,
+ "step": 3304
+ },
+ {
+ "epoch": 45.90829694323144,
+ "grad_norm": 0.022508298978209496,
+ "learning_rate": 0.0006,
+ "loss": 5.982312202453613,
+ "step": 3305
+ },
+ {
+ "epoch": 45.92227074235808,
+ "grad_norm": 0.024880951270461082,
+ "learning_rate": 0.0006,
+ "loss": 5.931150436401367,
+ "step": 3306
+ },
+ {
+ "epoch": 45.93624454148472,
+ "grad_norm": 0.030007576569914818,
+ "learning_rate": 0.0006,
+ "loss": 5.936639308929443,
+ "step": 3307
+ },
+ {
+ "epoch": 45.95021834061135,
+ "grad_norm": 0.033314041793346405,
+ "learning_rate": 0.0006,
+ "loss": 5.944248676300049,
+ "step": 3308
+ },
+ {
+ "epoch": 45.96419213973799,
+ "grad_norm": 0.043078381568193436,
+ "learning_rate": 0.0006,
+ "loss": 5.971286773681641,
+ "step": 3309
+ },
+ {
+ "epoch": 45.97816593886463,
+ "grad_norm": 0.059296440333127975,
+ "learning_rate": 0.0006,
+ "loss": 5.910497665405273,
+ "step": 3310
+ },
+ {
+ "epoch": 45.992139737991266,
+ "grad_norm": 0.06232666224241257,
+ "learning_rate": 0.0006,
+ "loss": 5.929817199707031,
+ "step": 3311
+ },
+ {
+ "epoch": 46.0,
+ "grad_norm": 0.05303940549492836,
+ "learning_rate": 0.0006,
+ "loss": 5.910822868347168,
+ "step": 3312
+ },
+ {
+ "epoch": 46.0,
+ "eval_loss": 5.977851390838623,
+ "eval_runtime": 56.2435,
+ "eval_samples_per_second": 43.418,
+ "eval_steps_per_second": 1.369,
+ "step": 3312
+ },
+ {
+ "epoch": 46.01397379912664,
+ "grad_norm": 0.06943798810243607,
+ "learning_rate": 0.0006,
+ "loss": 5.854922294616699,
+ "step": 3313
+ },
+ {
+ "epoch": 46.02794759825328,
+ "grad_norm": 0.05236278846859932,
+ "learning_rate": 0.0006,
+ "loss": 5.876611232757568,
+ "step": 3314
+ },
+ {
+ "epoch": 46.041921397379916,
+ "grad_norm": 0.06240001320838928,
+ "learning_rate": 0.0006,
+ "loss": 5.878912448883057,
+ "step": 3315
+ },
+ {
+ "epoch": 46.05589519650655,
+ "grad_norm": 0.07401192933320999,
+ "learning_rate": 0.0006,
+ "loss": 5.985136032104492,
+ "step": 3316
+ },
+ {
+ "epoch": 46.069868995633186,
+ "grad_norm": 0.0826529711484909,
+ "learning_rate": 0.0006,
+ "loss": 5.939439296722412,
+ "step": 3317
+ },
+ {
+ "epoch": 46.083842794759825,
+ "grad_norm": 0.07999005168676376,
+ "learning_rate": 0.0006,
+ "loss": 5.982336044311523,
+ "step": 3318
+ },
+ {
+ "epoch": 46.09781659388646,
+ "grad_norm": 0.07014869153499603,
+ "learning_rate": 0.0006,
+ "loss": 5.896534442901611,
+ "step": 3319
+ },
+ {
+ "epoch": 46.1117903930131,
+ "grad_norm": 0.05795128643512726,
+ "learning_rate": 0.0006,
+ "loss": 5.843111991882324,
+ "step": 3320
+ },
+ {
+ "epoch": 46.12576419213974,
+ "grad_norm": 0.06374258548021317,
+ "learning_rate": 0.0006,
+ "loss": 5.78752326965332,
+ "step": 3321
+ },
+ {
+ "epoch": 46.13973799126637,
+ "grad_norm": 0.045524805784225464,
+ "learning_rate": 0.0006,
+ "loss": 5.86755895614624,
+ "step": 3322
+ },
+ {
+ "epoch": 46.15371179039301,
+ "grad_norm": 0.03773504123091698,
+ "learning_rate": 0.0006,
+ "loss": 5.83627462387085,
+ "step": 3323
+ },
+ {
+ "epoch": 46.16768558951965,
+ "grad_norm": 0.035815201699733734,
+ "learning_rate": 0.0006,
+ "loss": 5.798238277435303,
+ "step": 3324
+ },
+ {
+ "epoch": 46.18165938864629,
+ "grad_norm": 0.03756783902645111,
+ "learning_rate": 0.0006,
+ "loss": 5.845197677612305,
+ "step": 3325
+ },
+ {
+ "epoch": 46.19563318777293,
+ "grad_norm": 0.03569113463163376,
+ "learning_rate": 0.0006,
+ "loss": 5.888932704925537,
+ "step": 3326
+ },
+ {
+ "epoch": 46.209606986899566,
+ "grad_norm": 0.029467813670635223,
+ "learning_rate": 0.0006,
+ "loss": 5.792047500610352,
+ "step": 3327
+ },
+ {
+ "epoch": 46.223580786026204,
+ "grad_norm": 0.025698406621813774,
+ "learning_rate": 0.0006,
+ "loss": 5.783734321594238,
+ "step": 3328
+ },
+ {
+ "epoch": 46.237554585152836,
+ "grad_norm": 0.028426671400666237,
+ "learning_rate": 0.0006,
+ "loss": 5.762084007263184,
+ "step": 3329
+ },
+ {
+ "epoch": 46.251528384279474,
+ "grad_norm": 0.03099622018635273,
+ "learning_rate": 0.0006,
+ "loss": 5.719109058380127,
+ "step": 3330
+ },
+ {
+ "epoch": 46.26550218340611,
+ "grad_norm": 0.0310219656676054,
+ "learning_rate": 0.0006,
+ "loss": 5.737057209014893,
+ "step": 3331
+ },
+ {
+ "epoch": 46.27947598253275,
+ "grad_norm": 0.02713446132838726,
+ "learning_rate": 0.0006,
+ "loss": 5.758657455444336,
+ "step": 3332
+ },
+ {
+ "epoch": 46.29344978165939,
+ "grad_norm": 0.03099844604730606,
+ "learning_rate": 0.0006,
+ "loss": 5.742086410522461,
+ "step": 3333
+ },
+ {
+ "epoch": 46.30742358078603,
+ "grad_norm": 0.030258730053901672,
+ "learning_rate": 0.0006,
+ "loss": 5.767426490783691,
+ "step": 3334
+ },
+ {
+ "epoch": 46.32139737991266,
+ "grad_norm": 0.029034852981567383,
+ "learning_rate": 0.0006,
+ "loss": 5.651792049407959,
+ "step": 3335
+ },
+ {
+ "epoch": 46.3353711790393,
+ "grad_norm": 0.03138665854930878,
+ "learning_rate": 0.0006,
+ "loss": 5.664980888366699,
+ "step": 3336
+ },
+ {
+ "epoch": 46.34934497816594,
+ "grad_norm": 0.040546901524066925,
+ "learning_rate": 0.0006,
+ "loss": 5.665165901184082,
+ "step": 3337
+ },
+ {
+ "epoch": 46.36331877729258,
+ "grad_norm": 0.06395485252141953,
+ "learning_rate": 0.0006,
+ "loss": 5.7776713371276855,
+ "step": 3338
+ },
+ {
+ "epoch": 46.377292576419215,
+ "grad_norm": 0.10088056325912476,
+ "learning_rate": 0.0006,
+ "loss": 5.755645751953125,
+ "step": 3339
+ },
+ {
+ "epoch": 46.391266375545854,
+ "grad_norm": 0.09527938067913055,
+ "learning_rate": 0.0006,
+ "loss": 5.838809490203857,
+ "step": 3340
+ },
+ {
+ "epoch": 46.40524017467249,
+ "grad_norm": 0.0849648267030716,
+ "learning_rate": 0.0006,
+ "loss": 5.781932830810547,
+ "step": 3341
+ },
+ {
+ "epoch": 46.419213973799124,
+ "grad_norm": 0.10142666101455688,
+ "learning_rate": 0.0006,
+ "loss": 5.838970184326172,
+ "step": 3342
+ },
+ {
+ "epoch": 46.43318777292576,
+ "grad_norm": 0.42731454968452454,
+ "learning_rate": 0.0006,
+ "loss": 6.071902275085449,
+ "step": 3343
+ },
+ {
+ "epoch": 46.4471615720524,
+ "grad_norm": 0.13102488219738007,
+ "learning_rate": 0.0006,
+ "loss": 6.8635993003845215,
+ "step": 3344
+ },
+ {
+ "epoch": 46.46113537117904,
+ "grad_norm": 0.23691938817501068,
+ "learning_rate": 0.0006,
+ "loss": 7.313158988952637,
+ "step": 3345
+ },
+ {
+ "epoch": 46.47510917030568,
+ "grad_norm": 0.11670669168233871,
+ "learning_rate": 0.0006,
+ "loss": 7.240363121032715,
+ "step": 3346
+ },
+ {
+ "epoch": 46.48908296943232,
+ "grad_norm": 0.09985820949077606,
+ "learning_rate": 0.0006,
+ "loss": 7.113332271575928,
+ "step": 3347
+ },
+ {
+ "epoch": 46.50305676855895,
+ "grad_norm": 0.11406408250331879,
+ "learning_rate": 0.0006,
+ "loss": 7.051285743713379,
+ "step": 3348
+ },
+ {
+ "epoch": 46.51703056768559,
+ "grad_norm": 0.07900336384773254,
+ "learning_rate": 0.0006,
+ "loss": 7.014365196228027,
+ "step": 3349
+ },
+ {
+ "epoch": 46.531004366812226,
+ "grad_norm": 0.05380333960056305,
+ "learning_rate": 0.0006,
+ "loss": 6.8868818283081055,
+ "step": 3350
+ },
+ {
+ "epoch": 46.544978165938865,
+ "grad_norm": 0.05868449807167053,
+ "learning_rate": 0.0006,
+ "loss": 6.783599853515625,
+ "step": 3351
+ },
+ {
+ "epoch": 46.5589519650655,
+ "grad_norm": 0.04428785294294357,
+ "learning_rate": 0.0006,
+ "loss": 6.798511505126953,
+ "step": 3352
+ },
+ {
+ "epoch": 46.57292576419214,
+ "grad_norm": 0.055527277290821075,
+ "learning_rate": 0.0006,
+ "loss": 6.694507598876953,
+ "step": 3353
+ },
+ {
+ "epoch": 46.58689956331878,
+ "grad_norm": 0.03872432932257652,
+ "learning_rate": 0.0006,
+ "loss": 6.736631393432617,
+ "step": 3354
+ },
+ {
+ "epoch": 46.60087336244541,
+ "grad_norm": 0.04330779239535332,
+ "learning_rate": 0.0006,
+ "loss": 6.576806545257568,
+ "step": 3355
+ },
+ {
+ "epoch": 46.61484716157205,
+ "grad_norm": 0.036767736077308655,
+ "learning_rate": 0.0006,
+ "loss": 6.559871196746826,
+ "step": 3356
+ },
+ {
+ "epoch": 46.62882096069869,
+ "grad_norm": 0.039857108145952225,
+ "learning_rate": 0.0006,
+ "loss": 6.5749831199646,
+ "step": 3357
+ },
+ {
+ "epoch": 46.64279475982533,
+ "grad_norm": 0.048719972372055054,
+ "learning_rate": 0.0006,
+ "loss": 6.4915618896484375,
+ "step": 3358
+ },
+ {
+ "epoch": 46.65676855895197,
+ "grad_norm": 0.042849015444517136,
+ "learning_rate": 0.0006,
+ "loss": 6.413516044616699,
+ "step": 3359
+ },
+ {
+ "epoch": 46.670742358078606,
+ "grad_norm": 0.02871009334921837,
+ "learning_rate": 0.0006,
+ "loss": 6.320952415466309,
+ "step": 3360
+ },
+ {
+ "epoch": 46.68471615720524,
+ "grad_norm": 0.03193620592355728,
+ "learning_rate": 0.0006,
+ "loss": 6.4176435470581055,
+ "step": 3361
+ },
+ {
+ "epoch": 46.698689956331876,
+ "grad_norm": 0.027417024597525597,
+ "learning_rate": 0.0006,
+ "loss": 6.342319488525391,
+ "step": 3362
+ },
+ {
+ "epoch": 46.712663755458514,
+ "grad_norm": 0.03305937349796295,
+ "learning_rate": 0.0006,
+ "loss": 6.246112823486328,
+ "step": 3363
+ },
+ {
+ "epoch": 46.72663755458515,
+ "grad_norm": 0.031403183937072754,
+ "learning_rate": 0.0006,
+ "loss": 6.149503231048584,
+ "step": 3364
+ },
+ {
+ "epoch": 46.74061135371179,
+ "grad_norm": 0.031563788652420044,
+ "learning_rate": 0.0006,
+ "loss": 6.278808116912842,
+ "step": 3365
+ },
+ {
+ "epoch": 46.75458515283843,
+ "grad_norm": 0.028366589918732643,
+ "learning_rate": 0.0006,
+ "loss": 6.057036876678467,
+ "step": 3366
+ },
+ {
+ "epoch": 46.76855895196506,
+ "grad_norm": 0.023715803399682045,
+ "learning_rate": 0.0006,
+ "loss": 6.111324310302734,
+ "step": 3367
+ },
+ {
+ "epoch": 46.7825327510917,
+ "grad_norm": 0.026546811684966087,
+ "learning_rate": 0.0006,
+ "loss": 6.105669975280762,
+ "step": 3368
+ },
+ {
+ "epoch": 46.79650655021834,
+ "grad_norm": 0.023531019687652588,
+ "learning_rate": 0.0006,
+ "loss": 6.068072319030762,
+ "step": 3369
+ },
+ {
+ "epoch": 46.81048034934498,
+ "grad_norm": 0.024480128660798073,
+ "learning_rate": 0.0006,
+ "loss": 5.947896957397461,
+ "step": 3370
+ },
+ {
+ "epoch": 46.82445414847162,
+ "grad_norm": 0.02517310529947281,
+ "learning_rate": 0.0006,
+ "loss": 6.030037879943848,
+ "step": 3371
+ },
+ {
+ "epoch": 46.838427947598255,
+ "grad_norm": 0.026723403483629227,
+ "learning_rate": 0.0006,
+ "loss": 5.963776111602783,
+ "step": 3372
+ },
+ {
+ "epoch": 46.852401746724894,
+ "grad_norm": 0.03163504600524902,
+ "learning_rate": 0.0006,
+ "loss": 5.893466949462891,
+ "step": 3373
+ },
+ {
+ "epoch": 46.866375545851525,
+ "grad_norm": 0.04058683291077614,
+ "learning_rate": 0.0006,
+ "loss": 5.9202799797058105,
+ "step": 3374
+ },
+ {
+ "epoch": 46.880349344978164,
+ "grad_norm": 0.04964117333292961,
+ "learning_rate": 0.0006,
+ "loss": 5.9366559982299805,
+ "step": 3375
+ },
+ {
+ "epoch": 46.8943231441048,
+ "grad_norm": 0.055406033992767334,
+ "learning_rate": 0.0006,
+ "loss": 5.912230968475342,
+ "step": 3376
+ },
+ {
+ "epoch": 46.90829694323144,
+ "grad_norm": 0.04643625393509865,
+ "learning_rate": 0.0006,
+ "loss": 5.87260103225708,
+ "step": 3377
+ },
+ {
+ "epoch": 46.92227074235808,
+ "grad_norm": 0.029325876384973526,
+ "learning_rate": 0.0006,
+ "loss": 5.854763984680176,
+ "step": 3378
+ },
+ {
+ "epoch": 46.93624454148472,
+ "grad_norm": 0.03159382566809654,
+ "learning_rate": 0.0006,
+ "loss": 5.880882263183594,
+ "step": 3379
+ },
+ {
+ "epoch": 46.95021834061135,
+ "grad_norm": 0.030224012210965157,
+ "learning_rate": 0.0006,
+ "loss": 5.787360191345215,
+ "step": 3380
+ },
+ {
+ "epoch": 46.96419213973799,
+ "grad_norm": 0.028368130326271057,
+ "learning_rate": 0.0006,
+ "loss": 5.817195892333984,
+ "step": 3381
+ },
+ {
+ "epoch": 46.97816593886463,
+ "grad_norm": 0.029411399737000465,
+ "learning_rate": 0.0006,
+ "loss": 5.783403396606445,
+ "step": 3382
+ },
+ {
+ "epoch": 46.992139737991266,
+ "grad_norm": 0.029927462339401245,
+ "learning_rate": 0.0006,
+ "loss": 5.662444591522217,
+ "step": 3383
+ },
+ {
+ "epoch": 47.0,
+ "grad_norm": 0.02872961387038231,
+ "learning_rate": 0.0006,
+ "loss": 5.742781639099121,
+ "step": 3384
+ },
+ {
+ "epoch": 47.0,
+ "eval_loss": 5.870260238647461,
+ "eval_runtime": 57.3734,
+ "eval_samples_per_second": 42.563,
+ "eval_steps_per_second": 1.342,
+ "step": 3384
+ },
+ {
+ "epoch": 47.01397379912664,
+ "grad_norm": 0.028542131185531616,
+ "learning_rate": 0.0006,
+ "loss": 5.753256797790527,
+ "step": 3385
+ },
+ {
+ "epoch": 47.02794759825328,
+ "grad_norm": 0.03227119892835617,
+ "learning_rate": 0.0006,
+ "loss": 5.722246170043945,
+ "step": 3386
+ },
+ {
+ "epoch": 47.041921397379916,
+ "grad_norm": 0.031461723148822784,
+ "learning_rate": 0.0006,
+ "loss": 5.740520000457764,
+ "step": 3387
+ },
+ {
+ "epoch": 47.05589519650655,
+ "grad_norm": 0.03184220567345619,
+ "learning_rate": 0.0006,
+ "loss": 5.720890045166016,
+ "step": 3388
+ },
+ {
+ "epoch": 47.069868995633186,
+ "grad_norm": 0.030015580356121063,
+ "learning_rate": 0.0006,
+ "loss": 5.597814559936523,
+ "step": 3389
+ },
+ {
+ "epoch": 47.083842794759825,
+ "grad_norm": 0.029931779950857162,
+ "learning_rate": 0.0006,
+ "loss": 5.5412468910217285,
+ "step": 3390
+ },
+ {
+ "epoch": 47.09781659388646,
+ "grad_norm": 0.03206505626440048,
+ "learning_rate": 0.0006,
+ "loss": 5.6967597007751465,
+ "step": 3391
+ },
+ {
+ "epoch": 47.1117903930131,
+ "grad_norm": 0.03978161886334419,
+ "learning_rate": 0.0006,
+ "loss": 5.59661340713501,
+ "step": 3392
+ },
+ {
+ "epoch": 47.12576419213974,
+ "grad_norm": 0.04619080200791359,
+ "learning_rate": 0.0006,
+ "loss": 5.795788764953613,
+ "step": 3393
+ },
+ {
+ "epoch": 47.13973799126637,
+ "grad_norm": 0.049085382372140884,
+ "learning_rate": 0.0006,
+ "loss": 5.667335510253906,
+ "step": 3394
+ },
+ {
+ "epoch": 47.15371179039301,
+ "grad_norm": 0.04830858111381531,
+ "learning_rate": 0.0006,
+ "loss": 5.668806552886963,
+ "step": 3395
+ },
+ {
+ "epoch": 47.16768558951965,
+ "grad_norm": 0.0583178736269474,
+ "learning_rate": 0.0006,
+ "loss": 5.692285537719727,
+ "step": 3396
+ },
+ {
+ "epoch": 47.18165938864629,
+ "grad_norm": 0.05182253196835518,
+ "learning_rate": 0.0006,
+ "loss": 5.678645133972168,
+ "step": 3397
+ },
+ {
+ "epoch": 47.19563318777293,
+ "grad_norm": 0.04011905938386917,
+ "learning_rate": 0.0006,
+ "loss": 5.661550998687744,
+ "step": 3398
+ },
+ {
+ "epoch": 47.209606986899566,
+ "grad_norm": 0.03888486698269844,
+ "learning_rate": 0.0006,
+ "loss": 5.643001556396484,
+ "step": 3399
+ },
+ {
+ "epoch": 47.223580786026204,
+ "grad_norm": 0.030658265575766563,
+ "learning_rate": 0.0006,
+ "loss": 5.659090995788574,
+ "step": 3400
+ },
+ {
+ "epoch": 47.237554585152836,
+ "grad_norm": 0.03609256446361542,
+ "learning_rate": 0.0006,
+ "loss": 5.544501781463623,
+ "step": 3401
+ },
+ {
+ "epoch": 47.251528384279474,
+ "grad_norm": 0.033589769154787064,
+ "learning_rate": 0.0006,
+ "loss": 5.463878631591797,
+ "step": 3402
+ },
+ {
+ "epoch": 47.26550218340611,
+ "grad_norm": 0.035128287971019745,
+ "learning_rate": 0.0006,
+ "loss": 5.5680060386657715,
+ "step": 3403
+ },
+ {
+ "epoch": 47.27947598253275,
+ "grad_norm": 0.038684550672769547,
+ "learning_rate": 0.0006,
+ "loss": 5.587898254394531,
+ "step": 3404
+ },
+ {
+ "epoch": 47.29344978165939,
+ "grad_norm": 0.043566226959228516,
+ "learning_rate": 0.0006,
+ "loss": 5.580896854400635,
+ "step": 3405
+ },
+ {
+ "epoch": 47.30742358078603,
+ "grad_norm": 0.04428374394774437,
+ "learning_rate": 0.0006,
+ "loss": 5.603460311889648,
+ "step": 3406
+ },
+ {
+ "epoch": 47.32139737991266,
+ "grad_norm": 0.037368934601545334,
+ "learning_rate": 0.0006,
+ "loss": 5.564740180969238,
+ "step": 3407
+ },
+ {
+ "epoch": 47.3353711790393,
+ "grad_norm": 0.04092350974678993,
+ "learning_rate": 0.0006,
+ "loss": 5.584153652191162,
+ "step": 3408
+ },
+ {
+ "epoch": 47.34934497816594,
+ "grad_norm": 0.045031506568193436,
+ "learning_rate": 0.0006,
+ "loss": 5.601186275482178,
+ "step": 3409
+ },
+ {
+ "epoch": 47.36331877729258,
+ "grad_norm": 0.041830409318208694,
+ "learning_rate": 0.0006,
+ "loss": 5.512788772583008,
+ "step": 3410
+ },
+ {
+ "epoch": 47.377292576419215,
+ "grad_norm": 0.037633974105119705,
+ "learning_rate": 0.0006,
+ "loss": 5.503828525543213,
+ "step": 3411
+ },
+ {
+ "epoch": 47.391266375545854,
+ "grad_norm": 0.03473588079214096,
+ "learning_rate": 0.0006,
+ "loss": 5.570557594299316,
+ "step": 3412
+ },
+ {
+ "epoch": 47.40524017467249,
+ "grad_norm": 0.039325181394815445,
+ "learning_rate": 0.0006,
+ "loss": 5.427783966064453,
+ "step": 3413
+ },
+ {
+ "epoch": 47.419213973799124,
+ "grad_norm": 0.04034624621272087,
+ "learning_rate": 0.0006,
+ "loss": 5.572591304779053,
+ "step": 3414
+ },
+ {
+ "epoch": 47.43318777292576,
+ "grad_norm": 0.05245184153318405,
+ "learning_rate": 0.0006,
+ "loss": 5.505208969116211,
+ "step": 3415
+ },
+ {
+ "epoch": 47.4471615720524,
+ "grad_norm": 0.0568675734102726,
+ "learning_rate": 0.0006,
+ "loss": 5.589373588562012,
+ "step": 3416
+ },
+ {
+ "epoch": 47.46113537117904,
+ "grad_norm": 0.048336587846279144,
+ "learning_rate": 0.0006,
+ "loss": 5.522005081176758,
+ "step": 3417
+ },
+ {
+ "epoch": 47.47510917030568,
+ "grad_norm": 0.04194975271821022,
+ "learning_rate": 0.0006,
+ "loss": 5.387576103210449,
+ "step": 3418
+ },
+ {
+ "epoch": 47.48908296943232,
+ "grad_norm": 0.04019433259963989,
+ "learning_rate": 0.0006,
+ "loss": 5.522615432739258,
+ "step": 3419
+ },
+ {
+ "epoch": 47.50305676855895,
+ "grad_norm": 0.05709933117032051,
+ "learning_rate": 0.0006,
+ "loss": 5.420286178588867,
+ "step": 3420
+ },
+ {
+ "epoch": 47.51703056768559,
+ "grad_norm": 0.09130338579416275,
+ "learning_rate": 0.0006,
+ "loss": 5.484622478485107,
+ "step": 3421
+ },
+ {
+ "epoch": 47.531004366812226,
+ "grad_norm": 0.2757541835308075,
+ "learning_rate": 0.0006,
+ "loss": 5.821832656860352,
+ "step": 3422
+ },
+ {
+ "epoch": 47.544978165938865,
+ "grad_norm": 0.28292080760002136,
+ "learning_rate": 0.0006,
+ "loss": 6.579934120178223,
+ "step": 3423
+ },
+ {
+ "epoch": 47.5589519650655,
+ "grad_norm": 0.15293604135513306,
+ "learning_rate": 0.0006,
+ "loss": 6.478213310241699,
+ "step": 3424
+ },
+ {
+ "epoch": 47.57292576419214,
+ "grad_norm": 0.11284393817186356,
+ "learning_rate": 0.0006,
+ "loss": 6.4066386222839355,
+ "step": 3425
+ },
+ {
+ "epoch": 47.58689956331878,
+ "grad_norm": 0.10933073610067368,
+ "learning_rate": 0.0006,
+ "loss": 6.3306169509887695,
+ "step": 3426
+ },
+ {
+ "epoch": 47.60087336244541,
+ "grad_norm": 0.05837641656398773,
+ "learning_rate": 0.0006,
+ "loss": 6.250000953674316,
+ "step": 3427
+ },
+ {
+ "epoch": 47.61484716157205,
+ "grad_norm": 0.05987229570746422,
+ "learning_rate": 0.0006,
+ "loss": 6.152463436126709,
+ "step": 3428
+ },
+ {
+ "epoch": 47.62882096069869,
+ "grad_norm": 0.04064938426017761,
+ "learning_rate": 0.0006,
+ "loss": 6.117897033691406,
+ "step": 3429
+ },
+ {
+ "epoch": 47.64279475982533,
+ "grad_norm": 0.04565057530999184,
+ "learning_rate": 0.0006,
+ "loss": 6.042291641235352,
+ "step": 3430
+ },
+ {
+ "epoch": 47.65676855895197,
+ "grad_norm": 0.03475063294172287,
+ "learning_rate": 0.0006,
+ "loss": 6.0071868896484375,
+ "step": 3431
+ },
+ {
+ "epoch": 47.670742358078606,
+ "grad_norm": 0.03809220343828201,
+ "learning_rate": 0.0006,
+ "loss": 6.0128021240234375,
+ "step": 3432
+ },
+ {
+ "epoch": 47.68471615720524,
+ "grad_norm": 0.03307219594717026,
+ "learning_rate": 0.0006,
+ "loss": 5.894664287567139,
+ "step": 3433
+ },
+ {
+ "epoch": 47.698689956331876,
+ "grad_norm": 0.028824346140027046,
+ "learning_rate": 0.0006,
+ "loss": 5.914865493774414,
+ "step": 3434
+ },
+ {
+ "epoch": 47.712663755458514,
+ "grad_norm": 0.027923690155148506,
+ "learning_rate": 0.0006,
+ "loss": 5.851759910583496,
+ "step": 3435
+ },
+ {
+ "epoch": 47.72663755458515,
+ "grad_norm": 0.025492684915661812,
+ "learning_rate": 0.0006,
+ "loss": 5.749186038970947,
+ "step": 3436
+ },
+ {
+ "epoch": 47.74061135371179,
+ "grad_norm": 0.025105174630880356,
+ "learning_rate": 0.0006,
+ "loss": 5.803566932678223,
+ "step": 3437
+ },
+ {
+ "epoch": 47.75458515283843,
+ "grad_norm": 0.023298880085349083,
+ "learning_rate": 0.0006,
+ "loss": 5.76276159286499,
+ "step": 3438
+ },
+ {
+ "epoch": 47.76855895196506,
+ "grad_norm": 0.02468576282262802,
+ "learning_rate": 0.0006,
+ "loss": 5.833852767944336,
+ "step": 3439
+ },
+ {
+ "epoch": 47.7825327510917,
+ "grad_norm": 0.022730689495801926,
+ "learning_rate": 0.0006,
+ "loss": 5.736724376678467,
+ "step": 3440
+ },
+ {
+ "epoch": 47.79650655021834,
+ "grad_norm": 0.021335354074835777,
+ "learning_rate": 0.0006,
+ "loss": 5.644248008728027,
+ "step": 3441
+ },
+ {
+ "epoch": 47.81048034934498,
+ "grad_norm": 0.02288101427257061,
+ "learning_rate": 0.0006,
+ "loss": 5.7011942863464355,
+ "step": 3442
+ },
+ {
+ "epoch": 47.82445414847162,
+ "grad_norm": 0.01961585320532322,
+ "learning_rate": 0.0006,
+ "loss": 5.63215446472168,
+ "step": 3443
+ },
+ {
+ "epoch": 47.838427947598255,
+ "grad_norm": 0.017659205943346024,
+ "learning_rate": 0.0006,
+ "loss": 5.744845390319824,
+ "step": 3444
+ },
+ {
+ "epoch": 47.852401746724894,
+ "grad_norm": 0.02281784638762474,
+ "learning_rate": 0.0006,
+ "loss": 5.696993350982666,
+ "step": 3445
+ },
+ {
+ "epoch": 47.866375545851525,
+ "grad_norm": 0.016394242644309998,
+ "learning_rate": 0.0006,
+ "loss": 5.659870147705078,
+ "step": 3446
+ },
+ {
+ "epoch": 47.880349344978164,
+ "grad_norm": 0.01967368647456169,
+ "learning_rate": 0.0006,
+ "loss": 5.638072490692139,
+ "step": 3447
+ },
+ {
+ "epoch": 47.8943231441048,
+ "grad_norm": 0.017326267436146736,
+ "learning_rate": 0.0006,
+ "loss": 5.638836860656738,
+ "step": 3448
+ },
+ {
+ "epoch": 47.90829694323144,
+ "grad_norm": 0.017778463661670685,
+ "learning_rate": 0.0006,
+ "loss": 5.634466648101807,
+ "step": 3449
+ },
+ {
+ "epoch": 47.92227074235808,
+ "grad_norm": 0.017269758507609367,
+ "learning_rate": 0.0006,
+ "loss": 5.506277561187744,
+ "step": 3450
+ },
+ {
+ "epoch": 47.93624454148472,
+ "grad_norm": 0.01733289286494255,
+ "learning_rate": 0.0006,
+ "loss": 5.544455528259277,
+ "step": 3451
+ },
+ {
+ "epoch": 47.95021834061135,
+ "grad_norm": 0.017787795513868332,
+ "learning_rate": 0.0006,
+ "loss": 5.498943328857422,
+ "step": 3452
+ },
+ {
+ "epoch": 47.96419213973799,
+ "grad_norm": 0.015770932659506798,
+ "learning_rate": 0.0006,
+ "loss": 5.537227630615234,
+ "step": 3453
+ },
+ {
+ "epoch": 47.97816593886463,
+ "grad_norm": 0.016132809221744537,
+ "learning_rate": 0.0006,
+ "loss": 5.489081382751465,
+ "step": 3454
+ },
+ {
+ "epoch": 47.992139737991266,
+ "grad_norm": 0.01749664731323719,
+ "learning_rate": 0.0006,
+ "loss": 5.454263687133789,
+ "step": 3455
+ },
+ {
+ "epoch": 48.0,
+ "grad_norm": 0.016711296513676643,
+ "learning_rate": 0.0006,
+ "loss": 5.530783653259277,
+ "step": 3456
+ },
+ {
+ "epoch": 48.0,
+ "eval_loss": 5.596925258636475,
+ "eval_runtime": 56.9612,
+ "eval_samples_per_second": 42.871,
+ "eval_steps_per_second": 1.352,
+ "step": 3456
+ },
+ {
+ "epoch": 48.01397379912664,
+ "grad_norm": 0.01867641694843769,
+ "learning_rate": 0.0006,
+ "loss": 5.506853103637695,
+ "step": 3457
+ },
+ {
+ "epoch": 48.02794759825328,
+ "grad_norm": 0.023250792175531387,
+ "learning_rate": 0.0006,
+ "loss": 5.4074177742004395,
+ "step": 3458
+ },
+ {
+ "epoch": 48.041921397379916,
+ "grad_norm": 0.01970936730504036,
+ "learning_rate": 0.0006,
+ "loss": 5.508777618408203,
+ "step": 3459
+ },
+ {
+ "epoch": 48.05589519650655,
+ "grad_norm": 0.01879260502755642,
+ "learning_rate": 0.0006,
+ "loss": 5.412424087524414,
+ "step": 3460
+ },
+ {
+ "epoch": 48.069868995633186,
+ "grad_norm": 0.017420092597603798,
+ "learning_rate": 0.0006,
+ "loss": 5.43361759185791,
+ "step": 3461
+ },
+ {
+ "epoch": 48.083842794759825,
+ "grad_norm": 0.018707862123847008,
+ "learning_rate": 0.0006,
+ "loss": 5.476800918579102,
+ "step": 3462
+ },
+ {
+ "epoch": 48.09781659388646,
+ "grad_norm": 0.023185882717370987,
+ "learning_rate": 0.0006,
+ "loss": 5.465272426605225,
+ "step": 3463
+ },
+ {
+ "epoch": 48.1117903930131,
+ "grad_norm": 0.023570368066430092,
+ "learning_rate": 0.0006,
+ "loss": 5.391471862792969,
+ "step": 3464
+ },
+ {
+ "epoch": 48.12576419213974,
+ "grad_norm": 0.02651144750416279,
+ "learning_rate": 0.0006,
+ "loss": 5.422799110412598,
+ "step": 3465
+ },
+ {
+ "epoch": 48.13973799126637,
+ "grad_norm": 0.03316329047083855,
+ "learning_rate": 0.0006,
+ "loss": 5.3952202796936035,
+ "step": 3466
+ },
+ {
+ "epoch": 48.15371179039301,
+ "grad_norm": 0.04012278467416763,
+ "learning_rate": 0.0006,
+ "loss": 5.540853023529053,
+ "step": 3467
+ },
+ {
+ "epoch": 48.16768558951965,
+ "grad_norm": 0.04763203114271164,
+ "learning_rate": 0.0006,
+ "loss": 5.387890338897705,
+ "step": 3468
+ },
+ {
+ "epoch": 48.18165938864629,
+ "grad_norm": 0.05129165202379227,
+ "learning_rate": 0.0006,
+ "loss": 5.414722442626953,
+ "step": 3469
+ },
+ {
+ "epoch": 48.19563318777293,
+ "grad_norm": 0.04438071325421333,
+ "learning_rate": 0.0006,
+ "loss": 5.386244773864746,
+ "step": 3470
+ },
+ {
+ "epoch": 48.209606986899566,
+ "grad_norm": 0.040234919637441635,
+ "learning_rate": 0.0006,
+ "loss": 5.431617736816406,
+ "step": 3471
+ },
+ {
+ "epoch": 48.223580786026204,
+ "grad_norm": 0.03217008337378502,
+ "learning_rate": 0.0006,
+ "loss": 5.30168342590332,
+ "step": 3472
+ },
+ {
+ "epoch": 48.237554585152836,
+ "grad_norm": 0.03087974339723587,
+ "learning_rate": 0.0006,
+ "loss": 5.316389083862305,
+ "step": 3473
+ },
+ {
+ "epoch": 48.251528384279474,
+ "grad_norm": 0.02789347991347313,
+ "learning_rate": 0.0006,
+ "loss": 5.385970592498779,
+ "step": 3474
+ },
+ {
+ "epoch": 48.26550218340611,
+ "grad_norm": 0.024792097508907318,
+ "learning_rate": 0.0006,
+ "loss": 5.362329959869385,
+ "step": 3475
+ },
+ {
+ "epoch": 48.27947598253275,
+ "grad_norm": 0.025869522243738174,
+ "learning_rate": 0.0006,
+ "loss": 5.28400182723999,
+ "step": 3476
+ },
+ {
+ "epoch": 48.29344978165939,
+ "grad_norm": 0.021339882165193558,
+ "learning_rate": 0.0006,
+ "loss": 5.36669921875,
+ "step": 3477
+ },
+ {
+ "epoch": 48.30742358078603,
+ "grad_norm": 0.021462520584464073,
+ "learning_rate": 0.0006,
+ "loss": 5.311222076416016,
+ "step": 3478
+ },
+ {
+ "epoch": 48.32139737991266,
+ "grad_norm": 0.02193879708647728,
+ "learning_rate": 0.0006,
+ "loss": 5.291441917419434,
+ "step": 3479
+ },
+ {
+ "epoch": 48.3353711790393,
+ "grad_norm": 0.01963093690574169,
+ "learning_rate": 0.0006,
+ "loss": 5.342780113220215,
+ "step": 3480
+ },
+ {
+ "epoch": 48.34934497816594,
+ "grad_norm": 0.020634736865758896,
+ "learning_rate": 0.0006,
+ "loss": 5.144141674041748,
+ "step": 3481
+ },
+ {
+ "epoch": 48.36331877729258,
+ "grad_norm": 0.02126968279480934,
+ "learning_rate": 0.0006,
+ "loss": 5.232624530792236,
+ "step": 3482
+ },
+ {
+ "epoch": 48.377292576419215,
+ "grad_norm": 0.021203994750976562,
+ "learning_rate": 0.0006,
+ "loss": 5.290219783782959,
+ "step": 3483
+ },
+ {
+ "epoch": 48.391266375545854,
+ "grad_norm": 0.02193225920200348,
+ "learning_rate": 0.0006,
+ "loss": 5.29197359085083,
+ "step": 3484
+ },
+ {
+ "epoch": 48.40524017467249,
+ "grad_norm": 0.023174211382865906,
+ "learning_rate": 0.0006,
+ "loss": 5.204349517822266,
+ "step": 3485
+ },
+ {
+ "epoch": 48.419213973799124,
+ "grad_norm": 0.024161914363503456,
+ "learning_rate": 0.0006,
+ "loss": 5.339351654052734,
+ "step": 3486
+ },
+ {
+ "epoch": 48.43318777292576,
+ "grad_norm": 0.02221524342894554,
+ "learning_rate": 0.0006,
+ "loss": 5.263001441955566,
+ "step": 3487
+ },
+ {
+ "epoch": 48.4471615720524,
+ "grad_norm": 0.026632830500602722,
+ "learning_rate": 0.0006,
+ "loss": 5.221281051635742,
+ "step": 3488
+ },
+ {
+ "epoch": 48.46113537117904,
+ "grad_norm": 0.034923870116472244,
+ "learning_rate": 0.0006,
+ "loss": 5.296039581298828,
+ "step": 3489
+ },
+ {
+ "epoch": 48.47510917030568,
+ "grad_norm": 0.046184979379177094,
+ "learning_rate": 0.0006,
+ "loss": 5.212769985198975,
+ "step": 3490
+ },
+ {
+ "epoch": 48.48908296943232,
+ "grad_norm": 0.05154098570346832,
+ "learning_rate": 0.0006,
+ "loss": 5.337864875793457,
+ "step": 3491
+ },
+ {
+ "epoch": 48.50305676855895,
+ "grad_norm": 0.042967259883880615,
+ "learning_rate": 0.0006,
+ "loss": 5.267958641052246,
+ "step": 3492
+ },
+ {
+ "epoch": 48.51703056768559,
+ "grad_norm": 0.05356583371758461,
+ "learning_rate": 0.0006,
+ "loss": 5.1404218673706055,
+ "step": 3493
+ },
+ {
+ "epoch": 48.531004366812226,
+ "grad_norm": 0.08219032734632492,
+ "learning_rate": 0.0006,
+ "loss": 5.300687789916992,
+ "step": 3494
+ },
+ {
+ "epoch": 48.544978165938865,
+ "grad_norm": 0.19204089045524597,
+ "learning_rate": 0.0006,
+ "loss": 5.456025123596191,
+ "step": 3495
+ },
+ {
+ "epoch": 48.5589519650655,
+ "grad_norm": 0.25917696952819824,
+ "learning_rate": 0.0006,
+ "loss": 6.156761169433594,
+ "step": 3496
+ },
+ {
+ "epoch": 48.57292576419214,
+ "grad_norm": 0.1566283404827118,
+ "learning_rate": 0.0006,
+ "loss": 5.862346649169922,
+ "step": 3497
+ },
+ {
+ "epoch": 48.58689956331878,
+ "grad_norm": 0.1761486828327179,
+ "learning_rate": 0.0006,
+ "loss": 6.012795448303223,
+ "step": 3498
+ },
+ {
+ "epoch": 48.60087336244541,
+ "grad_norm": 0.09857621043920517,
+ "learning_rate": 0.0006,
+ "loss": 5.9248151779174805,
+ "step": 3499
+ },
+ {
+ "epoch": 48.61484716157205,
+ "grad_norm": 0.05968055501580238,
+ "learning_rate": 0.0006,
+ "loss": 5.780706882476807,
+ "step": 3500
+ },
+ {
+ "epoch": 48.62882096069869,
+ "grad_norm": 0.057433564215898514,
+ "learning_rate": 0.0006,
+ "loss": 5.645388603210449,
+ "step": 3501
+ },
+ {
+ "epoch": 48.64279475982533,
+ "grad_norm": 0.039659615606069565,
+ "learning_rate": 0.0006,
+ "loss": 5.695138931274414,
+ "step": 3502
+ },
+ {
+ "epoch": 48.65676855895197,
+ "grad_norm": 0.04019695147871971,
+ "learning_rate": 0.0006,
+ "loss": 5.672435283660889,
+ "step": 3503
+ },
+ {
+ "epoch": 48.670742358078606,
+ "grad_norm": 0.03033018670976162,
+ "learning_rate": 0.0006,
+ "loss": 5.682260513305664,
+ "step": 3504
+ },
+ {
+ "epoch": 48.68471615720524,
+ "grad_norm": 0.03215787187218666,
+ "learning_rate": 0.0006,
+ "loss": 5.604182243347168,
+ "step": 3505
+ },
+ {
+ "epoch": 48.698689956331876,
+ "grad_norm": 0.026865126565098763,
+ "learning_rate": 0.0006,
+ "loss": 5.64040470123291,
+ "step": 3506
+ },
+ {
+ "epoch": 48.712663755458514,
+ "grad_norm": 0.02831609919667244,
+ "learning_rate": 0.0006,
+ "loss": 5.552720069885254,
+ "step": 3507
+ },
+ {
+ "epoch": 48.72663755458515,
+ "grad_norm": 0.024226613342761993,
+ "learning_rate": 0.0006,
+ "loss": 5.556946277618408,
+ "step": 3508
+ },
+ {
+ "epoch": 48.74061135371179,
+ "grad_norm": 0.025393934920430183,
+ "learning_rate": 0.0006,
+ "loss": 5.476178169250488,
+ "step": 3509
+ },
+ {
+ "epoch": 48.75458515283843,
+ "grad_norm": 0.01984412968158722,
+ "learning_rate": 0.0006,
+ "loss": 5.471088409423828,
+ "step": 3510
+ },
+ {
+ "epoch": 48.76855895196506,
+ "grad_norm": 0.02035129815340042,
+ "learning_rate": 0.0006,
+ "loss": 5.438812732696533,
+ "step": 3511
+ },
+ {
+ "epoch": 48.7825327510917,
+ "grad_norm": 0.0208350270986557,
+ "learning_rate": 0.0006,
+ "loss": 5.381746292114258,
+ "step": 3512
+ },
+ {
+ "epoch": 48.79650655021834,
+ "grad_norm": 0.017783882096409798,
+ "learning_rate": 0.0006,
+ "loss": 5.351675510406494,
+ "step": 3513
+ },
+ {
+ "epoch": 48.81048034934498,
+ "grad_norm": 0.01671770215034485,
+ "learning_rate": 0.0006,
+ "loss": 5.329428672790527,
+ "step": 3514
+ },
+ {
+ "epoch": 48.82445414847162,
+ "grad_norm": 0.017891215160489082,
+ "learning_rate": 0.0006,
+ "loss": 5.212074279785156,
+ "step": 3515
+ },
+ {
+ "epoch": 48.838427947598255,
+ "grad_norm": 0.016847791150212288,
+ "learning_rate": 0.0006,
+ "loss": 5.299534797668457,
+ "step": 3516
+ },
+ {
+ "epoch": 48.852401746724894,
+ "grad_norm": 0.01593637838959694,
+ "learning_rate": 0.0006,
+ "loss": 5.30466890335083,
+ "step": 3517
+ },
+ {
+ "epoch": 48.866375545851525,
+ "grad_norm": 0.014713102020323277,
+ "learning_rate": 0.0006,
+ "loss": 5.294345378875732,
+ "step": 3518
+ },
+ {
+ "epoch": 48.880349344978164,
+ "grad_norm": 0.01433857623487711,
+ "learning_rate": 0.0006,
+ "loss": 5.2148542404174805,
+ "step": 3519
+ },
+ {
+ "epoch": 48.8943231441048,
+ "grad_norm": 0.014979379251599312,
+ "learning_rate": 0.0006,
+ "loss": 5.298582077026367,
+ "step": 3520
+ },
+ {
+ "epoch": 48.90829694323144,
+ "grad_norm": 0.013494192622601986,
+ "learning_rate": 0.0006,
+ "loss": 5.309783935546875,
+ "step": 3521
+ },
+ {
+ "epoch": 48.92227074235808,
+ "grad_norm": 0.014020012691617012,
+ "learning_rate": 0.0006,
+ "loss": 5.284879684448242,
+ "step": 3522
+ },
+ {
+ "epoch": 48.93624454148472,
+ "grad_norm": 0.01348058134317398,
+ "learning_rate": 0.0006,
+ "loss": 5.249116897583008,
+ "step": 3523
+ },
+ {
+ "epoch": 48.95021834061135,
+ "grad_norm": 0.013280218467116356,
+ "learning_rate": 0.0006,
+ "loss": 5.273489952087402,
+ "step": 3524
+ },
+ {
+ "epoch": 48.96419213973799,
+ "grad_norm": 0.012647110037505627,
+ "learning_rate": 0.0006,
+ "loss": 5.306652069091797,
+ "step": 3525
+ },
+ {
+ "epoch": 48.97816593886463,
+ "grad_norm": 0.012847724370658398,
+ "learning_rate": 0.0006,
+ "loss": 5.235745429992676,
+ "step": 3526
+ },
+ {
+ "epoch": 48.992139737991266,
+ "grad_norm": 0.011820166371762753,
+ "learning_rate": 0.0006,
+ "loss": 5.238091468811035,
+ "step": 3527
+ },
+ {
+ "epoch": 49.0,
+ "grad_norm": 0.01255585066974163,
+ "learning_rate": 0.0006,
+ "loss": 5.231295585632324,
+ "step": 3528
+ },
+ {
+ "epoch": 49.0,
+ "eval_loss": 5.394587993621826,
+ "eval_runtime": 56.3734,
+ "eval_samples_per_second": 43.318,
+ "eval_steps_per_second": 1.366,
+ "step": 3528
+ },
+ {
+ "epoch": 49.01397379912664,
+ "grad_norm": 0.011031972244381905,
+ "learning_rate": 0.0006,
+ "loss": 5.1354570388793945,
+ "step": 3529
+ },
+ {
+ "epoch": 49.02794759825328,
+ "grad_norm": 0.010631263256072998,
+ "learning_rate": 0.0006,
+ "loss": 5.177639961242676,
+ "step": 3530
+ },
+ {
+ "epoch": 49.041921397379916,
+ "grad_norm": 0.01133775431662798,
+ "learning_rate": 0.0006,
+ "loss": 5.283428192138672,
+ "step": 3531
+ },
+ {
+ "epoch": 49.05589519650655,
+ "grad_norm": 0.01154172234237194,
+ "learning_rate": 0.0006,
+ "loss": 5.206364154815674,
+ "step": 3532
+ },
+ {
+ "epoch": 49.069868995633186,
+ "grad_norm": 0.010192356072366238,
+ "learning_rate": 0.0006,
+ "loss": 5.2385735511779785,
+ "step": 3533
+ },
+ {
+ "epoch": 49.083842794759825,
+ "grad_norm": 0.012789200060069561,
+ "learning_rate": 0.0006,
+ "loss": 5.071800231933594,
+ "step": 3534
+ },
+ {
+ "epoch": 49.09781659388646,
+ "grad_norm": 0.011282820254564285,
+ "learning_rate": 0.0006,
+ "loss": 5.218020439147949,
+ "step": 3535
+ },
+ {
+ "epoch": 49.1117903930131,
+ "grad_norm": 0.012153448536992073,
+ "learning_rate": 0.0006,
+ "loss": 5.134698867797852,
+ "step": 3536
+ },
+ {
+ "epoch": 49.12576419213974,
+ "grad_norm": 0.011920543387532234,
+ "learning_rate": 0.0006,
+ "loss": 5.141620635986328,
+ "step": 3537
+ },
+ {
+ "epoch": 49.13973799126637,
+ "grad_norm": 0.010949329473078251,
+ "learning_rate": 0.0006,
+ "loss": 5.116259574890137,
+ "step": 3538
+ },
+ {
+ "epoch": 49.15371179039301,
+ "grad_norm": 0.012362448498606682,
+ "learning_rate": 0.0006,
+ "loss": 5.090177059173584,
+ "step": 3539
+ },
+ {
+ "epoch": 49.16768558951965,
+ "grad_norm": 0.011575071141123772,
+ "learning_rate": 0.0006,
+ "loss": 5.093441009521484,
+ "step": 3540
+ },
+ {
+ "epoch": 49.18165938864629,
+ "grad_norm": 0.011126236990094185,
+ "learning_rate": 0.0006,
+ "loss": 5.060235023498535,
+ "step": 3541
+ },
+ {
+ "epoch": 49.19563318777293,
+ "grad_norm": 0.0107401292771101,
+ "learning_rate": 0.0006,
+ "loss": 5.140704154968262,
+ "step": 3542
+ },
+ {
+ "epoch": 49.209606986899566,
+ "grad_norm": 0.011202906258404255,
+ "learning_rate": 0.0006,
+ "loss": 5.10855770111084,
+ "step": 3543
+ },
+ {
+ "epoch": 49.223580786026204,
+ "grad_norm": 0.011083454824984074,
+ "learning_rate": 0.0006,
+ "loss": 5.092751502990723,
+ "step": 3544
+ },
+ {
+ "epoch": 49.237554585152836,
+ "grad_norm": 0.009523800574243069,
+ "learning_rate": 0.0006,
+ "loss": 5.13733434677124,
+ "step": 3545
+ },
+ {
+ "epoch": 49.251528384279474,
+ "grad_norm": 0.011298353783786297,
+ "learning_rate": 0.0006,
+ "loss": 5.036048889160156,
+ "step": 3546
+ },
+ {
+ "epoch": 49.26550218340611,
+ "grad_norm": 0.010676483623683453,
+ "learning_rate": 0.0006,
+ "loss": 5.051901340484619,
+ "step": 3547
+ },
+ {
+ "epoch": 49.27947598253275,
+ "grad_norm": 0.011463399045169353,
+ "learning_rate": 0.0006,
+ "loss": 4.983170986175537,
+ "step": 3548
+ },
+ {
+ "epoch": 49.29344978165939,
+ "grad_norm": 0.010603402741253376,
+ "learning_rate": 0.0006,
+ "loss": 5.112311363220215,
+ "step": 3549
+ },
+ {
+ "epoch": 49.30742358078603,
+ "grad_norm": 0.010535670444369316,
+ "learning_rate": 0.0006,
+ "loss": 5.085227966308594,
+ "step": 3550
+ },
+ {
+ "epoch": 49.32139737991266,
+ "grad_norm": 0.011843928135931492,
+ "learning_rate": 0.0006,
+ "loss": 5.07575798034668,
+ "step": 3551
+ },
+ {
+ "epoch": 49.3353711790393,
+ "grad_norm": 0.010587952099740505,
+ "learning_rate": 0.0006,
+ "loss": 5.019237041473389,
+ "step": 3552
+ },
+ {
+ "epoch": 49.34934497816594,
+ "grad_norm": 0.012285659089684486,
+ "learning_rate": 0.0006,
+ "loss": 5.144603729248047,
+ "step": 3553
+ },
+ {
+ "epoch": 49.36331877729258,
+ "grad_norm": 0.013119431212544441,
+ "learning_rate": 0.0006,
+ "loss": 5.035256862640381,
+ "step": 3554
+ },
+ {
+ "epoch": 49.377292576419215,
+ "grad_norm": 0.012138537131249905,
+ "learning_rate": 0.0006,
+ "loss": 4.996667861938477,
+ "step": 3555
+ },
+ {
+ "epoch": 49.391266375545854,
+ "grad_norm": 0.012142804451286793,
+ "learning_rate": 0.0006,
+ "loss": 4.95058536529541,
+ "step": 3556
+ },
+ {
+ "epoch": 49.40524017467249,
+ "grad_norm": 0.012205676175653934,
+ "learning_rate": 0.0006,
+ "loss": 5.096931457519531,
+ "step": 3557
+ },
+ {
+ "epoch": 49.419213973799124,
+ "grad_norm": 0.012086744420230389,
+ "learning_rate": 0.0006,
+ "loss": 4.99384880065918,
+ "step": 3558
+ },
+ {
+ "epoch": 49.43318777292576,
+ "grad_norm": 0.012918103486299515,
+ "learning_rate": 0.0006,
+ "loss": 5.031294822692871,
+ "step": 3559
+ },
+ {
+ "epoch": 49.4471615720524,
+ "grad_norm": 0.012006713077425957,
+ "learning_rate": 0.0006,
+ "loss": 5.0547943115234375,
+ "step": 3560
+ },
+ {
+ "epoch": 49.46113537117904,
+ "grad_norm": 0.0120399734005332,
+ "learning_rate": 0.0006,
+ "loss": 5.014238357543945,
+ "step": 3561
+ },
+ {
+ "epoch": 49.47510917030568,
+ "grad_norm": 0.013306990265846252,
+ "learning_rate": 0.0006,
+ "loss": 4.972925186157227,
+ "step": 3562
+ },
+ {
+ "epoch": 49.48908296943232,
+ "grad_norm": 0.013878636993467808,
+ "learning_rate": 0.0006,
+ "loss": 5.079033851623535,
+ "step": 3563
+ },
+ {
+ "epoch": 49.50305676855895,
+ "grad_norm": 0.012454566545784473,
+ "learning_rate": 0.0006,
+ "loss": 4.93708610534668,
+ "step": 3564
+ },
+ {
+ "epoch": 49.51703056768559,
+ "grad_norm": 0.014075061306357384,
+ "learning_rate": 0.0006,
+ "loss": 4.965582370758057,
+ "step": 3565
+ },
+ {
+ "epoch": 49.531004366812226,
+ "grad_norm": 0.012525617145001888,
+ "learning_rate": 0.0006,
+ "loss": 4.968148708343506,
+ "step": 3566
+ },
+ {
+ "epoch": 49.544978165938865,
+ "grad_norm": 0.012418225407600403,
+ "learning_rate": 0.0006,
+ "loss": 5.06901741027832,
+ "step": 3567
+ },
+ {
+ "epoch": 49.5589519650655,
+ "grad_norm": 0.015034239739179611,
+ "learning_rate": 0.0006,
+ "loss": 4.984277725219727,
+ "step": 3568
+ },
+ {
+ "epoch": 49.57292576419214,
+ "grad_norm": 0.014329628087580204,
+ "learning_rate": 0.0006,
+ "loss": 4.951558589935303,
+ "step": 3569
+ },
+ {
+ "epoch": 49.58689956331878,
+ "grad_norm": 0.017769070342183113,
+ "learning_rate": 0.0006,
+ "loss": 4.966532230377197,
+ "step": 3570
+ },
+ {
+ "epoch": 49.60087336244541,
+ "grad_norm": 0.01909453794360161,
+ "learning_rate": 0.0006,
+ "loss": 4.919890880584717,
+ "step": 3571
+ },
+ {
+ "epoch": 49.61484716157205,
+ "grad_norm": 0.017519688233733177,
+ "learning_rate": 0.0006,
+ "loss": 5.031286716461182,
+ "step": 3572
+ },
+ {
+ "epoch": 49.62882096069869,
+ "grad_norm": 0.017143897712230682,
+ "learning_rate": 0.0006,
+ "loss": 4.969764709472656,
+ "step": 3573
+ },
+ {
+ "epoch": 49.64279475982533,
+ "grad_norm": 0.01688285917043686,
+ "learning_rate": 0.0006,
+ "loss": 5.0733442306518555,
+ "step": 3574
+ },
+ {
+ "epoch": 49.65676855895197,
+ "grad_norm": 0.018547311425209045,
+ "learning_rate": 0.0006,
+ "loss": 4.980863571166992,
+ "step": 3575
+ },
+ {
+ "epoch": 49.670742358078606,
+ "grad_norm": 0.018877137452363968,
+ "learning_rate": 0.0006,
+ "loss": 4.954646110534668,
+ "step": 3576
+ },
+ {
+ "epoch": 49.68471615720524,
+ "grad_norm": 0.018681600689888,
+ "learning_rate": 0.0006,
+ "loss": 4.959702014923096,
+ "step": 3577
+ },
+ {
+ "epoch": 49.698689956331876,
+ "grad_norm": 0.018947336822748184,
+ "learning_rate": 0.0006,
+ "loss": 5.070502758026123,
+ "step": 3578
+ },
+ {
+ "epoch": 49.712663755458514,
+ "grad_norm": 0.018817484378814697,
+ "learning_rate": 0.0006,
+ "loss": 5.056508541107178,
+ "step": 3579
+ },
+ {
+ "epoch": 49.72663755458515,
+ "grad_norm": 0.01931724324822426,
+ "learning_rate": 0.0006,
+ "loss": 4.922451972961426,
+ "step": 3580
+ },
+ {
+ "epoch": 49.74061135371179,
+ "grad_norm": 0.022047819569706917,
+ "learning_rate": 0.0006,
+ "loss": 5.053170204162598,
+ "step": 3581
+ },
+ {
+ "epoch": 49.75458515283843,
+ "grad_norm": 0.02279801294207573,
+ "learning_rate": 0.0006,
+ "loss": 4.8768439292907715,
+ "step": 3582
+ },
+ {
+ "epoch": 49.76855895196506,
+ "grad_norm": 0.020661601796746254,
+ "learning_rate": 0.0006,
+ "loss": 4.907928466796875,
+ "step": 3583
+ },
+ {
+ "epoch": 49.7825327510917,
+ "grad_norm": 0.020479509606957436,
+ "learning_rate": 0.0006,
+ "loss": 5.040694236755371,
+ "step": 3584
+ },
+ {
+ "epoch": 49.79650655021834,
+ "grad_norm": 0.021654162555933,
+ "learning_rate": 0.0006,
+ "loss": 5.018434524536133,
+ "step": 3585
+ },
+ {
+ "epoch": 49.81048034934498,
+ "grad_norm": 0.022486824542284012,
+ "learning_rate": 0.0006,
+ "loss": 5.086538314819336,
+ "step": 3586
+ },
+ {
+ "epoch": 49.82445414847162,
+ "grad_norm": 0.02141369879245758,
+ "learning_rate": 0.0006,
+ "loss": 5.035694122314453,
+ "step": 3587
+ },
+ {
+ "epoch": 49.838427947598255,
+ "grad_norm": 0.015926677733659744,
+ "learning_rate": 0.0006,
+ "loss": 4.905038833618164,
+ "step": 3588
+ },
+ {
+ "epoch": 49.852401746724894,
+ "grad_norm": 0.015775298699736595,
+ "learning_rate": 0.0006,
+ "loss": 5.003902435302734,
+ "step": 3589
+ },
+ {
+ "epoch": 49.866375545851525,
+ "grad_norm": 0.01872118003666401,
+ "learning_rate": 0.0006,
+ "loss": 4.891060829162598,
+ "step": 3590
+ },
+ {
+ "epoch": 49.880349344978164,
+ "grad_norm": 0.02066192403435707,
+ "learning_rate": 0.0006,
+ "loss": 5.002897262573242,
+ "step": 3591
+ },
+ {
+ "epoch": 49.8943231441048,
+ "grad_norm": 0.02348714880645275,
+ "learning_rate": 0.0006,
+ "loss": 4.9189958572387695,
+ "step": 3592
+ },
+ {
+ "epoch": 49.90829694323144,
+ "grad_norm": 0.020402103662490845,
+ "learning_rate": 0.0006,
+ "loss": 4.992661476135254,
+ "step": 3593
+ },
+ {
+ "epoch": 49.92227074235808,
+ "grad_norm": 0.020466072484850883,
+ "learning_rate": 0.0006,
+ "loss": 4.962968826293945,
+ "step": 3594
+ },
+ {
+ "epoch": 49.93624454148472,
+ "grad_norm": 0.01922314241528511,
+ "learning_rate": 0.0006,
+ "loss": 4.897730827331543,
+ "step": 3595
+ },
+ {
+ "epoch": 49.95021834061135,
+ "grad_norm": 0.017228754237294197,
+ "learning_rate": 0.0006,
+ "loss": 4.964842796325684,
+ "step": 3596
+ },
+ {
+ "epoch": 49.96419213973799,
+ "grad_norm": 0.016021743416786194,
+ "learning_rate": 0.0006,
+ "loss": 4.910758972167969,
+ "step": 3597
+ },
+ {
+ "epoch": 49.97816593886463,
+ "grad_norm": 0.017985103651881218,
+ "learning_rate": 0.0006,
+ "loss": 4.916072368621826,
+ "step": 3598
+ },
+ {
+ "epoch": 49.992139737991266,
+ "grad_norm": 0.018780240789055824,
+ "learning_rate": 0.0006,
+ "loss": 4.94658088684082,
+ "step": 3599
+ },
+ {
+ "epoch": 50.0,
+ "grad_norm": 0.01776314526796341,
+ "learning_rate": 0.0006,
+ "loss": 4.969357967376709,
+ "step": 3600
+ },
+ {
+ "epoch": 50.0,
+ "eval_loss": 5.141158580780029,
+ "eval_runtime": 56.2332,
+ "eval_samples_per_second": 43.426,
+ "eval_steps_per_second": 1.369,
+ "step": 3600
+ },
+ {
+ "epoch": 50.01397379912664,
+ "grad_norm": 0.016308696940541267,
+ "learning_rate": 0.0006,
+ "loss": 4.924664497375488,
+ "step": 3601
+ },
+ {
+ "epoch": 50.02794759825328,
+ "grad_norm": 0.018810199573636055,
+ "learning_rate": 0.0006,
+ "loss": 4.82293701171875,
+ "step": 3602
+ },
+ {
+ "epoch": 50.041921397379916,
+ "grad_norm": 0.02227160707116127,
+ "learning_rate": 0.0006,
+ "loss": 4.9612298011779785,
+ "step": 3603
+ },
+ {
+ "epoch": 50.05589519650655,
+ "grad_norm": 0.025266462936997414,
+ "learning_rate": 0.0006,
+ "loss": 4.963835716247559,
+ "step": 3604
+ },
+ {
+ "epoch": 50.069868995633186,
+ "grad_norm": 0.028588181361556053,
+ "learning_rate": 0.0006,
+ "loss": 4.992118835449219,
+ "step": 3605
+ },
+ {
+ "epoch": 50.083842794759825,
+ "grad_norm": 0.02739487960934639,
+ "learning_rate": 0.0006,
+ "loss": 4.915547847747803,
+ "step": 3606
+ },
+ {
+ "epoch": 50.09781659388646,
+ "grad_norm": 0.027000300586223602,
+ "learning_rate": 0.0006,
+ "loss": 4.883801460266113,
+ "step": 3607
+ },
+ {
+ "epoch": 50.1117903930131,
+ "grad_norm": 0.023468483239412308,
+ "learning_rate": 0.0006,
+ "loss": 5.006168365478516,
+ "step": 3608
+ },
+ {
+ "epoch": 50.12576419213974,
+ "grad_norm": 0.026753634214401245,
+ "learning_rate": 0.0006,
+ "loss": 4.92526912689209,
+ "step": 3609
+ },
+ {
+ "epoch": 50.13973799126637,
+ "grad_norm": 0.026816733181476593,
+ "learning_rate": 0.0006,
+ "loss": 4.847546100616455,
+ "step": 3610
+ },
+ {
+ "epoch": 50.15371179039301,
+ "grad_norm": 0.027362391352653503,
+ "learning_rate": 0.0006,
+ "loss": 4.862548351287842,
+ "step": 3611
+ },
+ {
+ "epoch": 50.16768558951965,
+ "grad_norm": 0.03450845181941986,
+ "learning_rate": 0.0006,
+ "loss": 4.963202953338623,
+ "step": 3612
+ },
+ {
+ "epoch": 50.18165938864629,
+ "grad_norm": 0.03801408037543297,
+ "learning_rate": 0.0006,
+ "loss": 4.770240306854248,
+ "step": 3613
+ },
+ {
+ "epoch": 50.19563318777293,
+ "grad_norm": 0.03579140082001686,
+ "learning_rate": 0.0006,
+ "loss": 4.797981262207031,
+ "step": 3614
+ },
+ {
+ "epoch": 50.209606986899566,
+ "grad_norm": 0.029393760487437248,
+ "learning_rate": 0.0006,
+ "loss": 4.798581600189209,
+ "step": 3615
+ },
+ {
+ "epoch": 50.223580786026204,
+ "grad_norm": 0.026111142709851265,
+ "learning_rate": 0.0006,
+ "loss": 4.990712642669678,
+ "step": 3616
+ },
+ {
+ "epoch": 50.237554585152836,
+ "grad_norm": 0.02213387005031109,
+ "learning_rate": 0.0006,
+ "loss": 4.865842342376709,
+ "step": 3617
+ },
+ {
+ "epoch": 50.251528384279474,
+ "grad_norm": 0.025450684130191803,
+ "learning_rate": 0.0006,
+ "loss": 4.903746128082275,
+ "step": 3618
+ },
+ {
+ "epoch": 50.26550218340611,
+ "grad_norm": 0.020915117114782333,
+ "learning_rate": 0.0006,
+ "loss": 4.951366424560547,
+ "step": 3619
+ },
+ {
+ "epoch": 50.27947598253275,
+ "grad_norm": 0.020972849801182747,
+ "learning_rate": 0.0006,
+ "loss": 4.783311367034912,
+ "step": 3620
+ },
+ {
+ "epoch": 50.29344978165939,
+ "grad_norm": 0.016956273466348648,
+ "learning_rate": 0.0006,
+ "loss": 4.955767631530762,
+ "step": 3621
+ },
+ {
+ "epoch": 50.30742358078603,
+ "grad_norm": 0.018045252189040184,
+ "learning_rate": 0.0006,
+ "loss": 4.869322776794434,
+ "step": 3622
+ },
+ {
+ "epoch": 50.32139737991266,
+ "grad_norm": 0.018199188634753227,
+ "learning_rate": 0.0006,
+ "loss": 4.95680046081543,
+ "step": 3623
+ },
+ {
+ "epoch": 50.3353711790393,
+ "grad_norm": 0.016688082367181778,
+ "learning_rate": 0.0006,
+ "loss": 4.783535957336426,
+ "step": 3624
+ },
+ {
+ "epoch": 50.34934497816594,
+ "grad_norm": 0.017540352419018745,
+ "learning_rate": 0.0006,
+ "loss": 4.8236083984375,
+ "step": 3625
+ },
+ {
+ "epoch": 50.36331877729258,
+ "grad_norm": 0.018054187297821045,
+ "learning_rate": 0.0006,
+ "loss": 4.764655113220215,
+ "step": 3626
+ },
+ {
+ "epoch": 50.377292576419215,
+ "grad_norm": 0.01741131953895092,
+ "learning_rate": 0.0006,
+ "loss": 4.875613689422607,
+ "step": 3627
+ },
+ {
+ "epoch": 50.391266375545854,
+ "grad_norm": 0.01682269014418125,
+ "learning_rate": 0.0006,
+ "loss": 4.818509101867676,
+ "step": 3628
+ },
+ {
+ "epoch": 50.40524017467249,
+ "grad_norm": 0.016652341932058334,
+ "learning_rate": 0.0006,
+ "loss": 4.974020481109619,
+ "step": 3629
+ },
+ {
+ "epoch": 50.419213973799124,
+ "grad_norm": 0.02035282365977764,
+ "learning_rate": 0.0006,
+ "loss": 4.893037796020508,
+ "step": 3630
+ },
+ {
+ "epoch": 50.43318777292576,
+ "grad_norm": 0.019907061010599136,
+ "learning_rate": 0.0006,
+ "loss": 4.848119735717773,
+ "step": 3631
+ },
+ {
+ "epoch": 50.4471615720524,
+ "grad_norm": 0.023861493915319443,
+ "learning_rate": 0.0006,
+ "loss": 4.819353103637695,
+ "step": 3632
+ },
+ {
+ "epoch": 50.46113537117904,
+ "grad_norm": 0.025254026055336,
+ "learning_rate": 0.0006,
+ "loss": 4.769312858581543,
+ "step": 3633
+ },
+ {
+ "epoch": 50.47510917030568,
+ "grad_norm": 0.026154937222599983,
+ "learning_rate": 0.0006,
+ "loss": 4.7586774826049805,
+ "step": 3634
+ },
+ {
+ "epoch": 50.48908296943232,
+ "grad_norm": 0.026342090219259262,
+ "learning_rate": 0.0006,
+ "loss": 4.861218452453613,
+ "step": 3635
+ },
+ {
+ "epoch": 50.50305676855895,
+ "grad_norm": 0.028169559314846992,
+ "learning_rate": 0.0006,
+ "loss": 4.818380355834961,
+ "step": 3636
+ },
+ {
+ "epoch": 50.51703056768559,
+ "grad_norm": 0.02820264920592308,
+ "learning_rate": 0.0006,
+ "loss": 4.892871856689453,
+ "step": 3637
+ },
+ {
+ "epoch": 50.531004366812226,
+ "grad_norm": 0.03345080465078354,
+ "learning_rate": 0.0006,
+ "loss": 4.811084747314453,
+ "step": 3638
+ },
+ {
+ "epoch": 50.544978165938865,
+ "grad_norm": 0.04005736857652664,
+ "learning_rate": 0.0006,
+ "loss": 4.765410900115967,
+ "step": 3639
+ },
+ {
+ "epoch": 50.5589519650655,
+ "grad_norm": 0.04333870857954025,
+ "learning_rate": 0.0006,
+ "loss": 4.956520080566406,
+ "step": 3640
+ },
+ {
+ "epoch": 50.57292576419214,
+ "grad_norm": 0.04216952994465828,
+ "learning_rate": 0.0006,
+ "loss": 4.793441295623779,
+ "step": 3641
+ },
+ {
+ "epoch": 50.58689956331878,
+ "grad_norm": 0.02936951443552971,
+ "learning_rate": 0.0006,
+ "loss": 4.8447957038879395,
+ "step": 3642
+ },
+ {
+ "epoch": 50.60087336244541,
+ "grad_norm": 0.026937026530504227,
+ "learning_rate": 0.0006,
+ "loss": 4.962215423583984,
+ "step": 3643
+ },
+ {
+ "epoch": 50.61484716157205,
+ "grad_norm": 0.02628890983760357,
+ "learning_rate": 0.0006,
+ "loss": 4.883061408996582,
+ "step": 3644
+ },
+ {
+ "epoch": 50.62882096069869,
+ "grad_norm": 0.025984834879636765,
+ "learning_rate": 0.0006,
+ "loss": 4.768662452697754,
+ "step": 3645
+ },
+ {
+ "epoch": 50.64279475982533,
+ "grad_norm": 0.023998012766242027,
+ "learning_rate": 0.0006,
+ "loss": 4.851325035095215,
+ "step": 3646
+ },
+ {
+ "epoch": 50.65676855895197,
+ "grad_norm": 0.02143184095621109,
+ "learning_rate": 0.0006,
+ "loss": 4.8274688720703125,
+ "step": 3647
+ },
+ {
+ "epoch": 50.670742358078606,
+ "grad_norm": 0.02488900162279606,
+ "learning_rate": 0.0006,
+ "loss": 4.831140518188477,
+ "step": 3648
+ },
+ {
+ "epoch": 50.68471615720524,
+ "grad_norm": 0.02330533228814602,
+ "learning_rate": 0.0006,
+ "loss": 4.850462913513184,
+ "step": 3649
+ },
+ {
+ "epoch": 50.698689956331876,
+ "grad_norm": 0.02121943049132824,
+ "learning_rate": 0.0006,
+ "loss": 4.870253086090088,
+ "step": 3650
+ },
+ {
+ "epoch": 50.712663755458514,
+ "grad_norm": 0.02080935798585415,
+ "learning_rate": 0.0006,
+ "loss": 4.798770904541016,
+ "step": 3651
+ },
+ {
+ "epoch": 50.72663755458515,
+ "grad_norm": 0.019047029316425323,
+ "learning_rate": 0.0006,
+ "loss": 4.8427300453186035,
+ "step": 3652
+ },
+ {
+ "epoch": 50.74061135371179,
+ "grad_norm": 0.01824103854596615,
+ "learning_rate": 0.0006,
+ "loss": 4.820980548858643,
+ "step": 3653
+ },
+ {
+ "epoch": 50.75458515283843,
+ "grad_norm": 0.018129967153072357,
+ "learning_rate": 0.0006,
+ "loss": 4.946732997894287,
+ "step": 3654
+ },
+ {
+ "epoch": 50.76855895196506,
+ "grad_norm": 0.01784815452992916,
+ "learning_rate": 0.0006,
+ "loss": 4.829516410827637,
+ "step": 3655
+ },
+ {
+ "epoch": 50.7825327510917,
+ "grad_norm": 0.020021555945277214,
+ "learning_rate": 0.0006,
+ "loss": 4.780428409576416,
+ "step": 3656
+ },
+ {
+ "epoch": 50.79650655021834,
+ "grad_norm": 0.01863253302872181,
+ "learning_rate": 0.0006,
+ "loss": 4.791567802429199,
+ "step": 3657
+ },
+ {
+ "epoch": 50.81048034934498,
+ "grad_norm": 0.019151683896780014,
+ "learning_rate": 0.0006,
+ "loss": 4.824892997741699,
+ "step": 3658
+ },
+ {
+ "epoch": 50.82445414847162,
+ "grad_norm": 0.01891648955643177,
+ "learning_rate": 0.0006,
+ "loss": 4.789605140686035,
+ "step": 3659
+ },
+ {
+ "epoch": 50.838427947598255,
+ "grad_norm": 0.016019510105252266,
+ "learning_rate": 0.0006,
+ "loss": 4.779555320739746,
+ "step": 3660
+ },
+ {
+ "epoch": 50.852401746724894,
+ "grad_norm": 0.016096491366624832,
+ "learning_rate": 0.0006,
+ "loss": 4.753842353820801,
+ "step": 3661
+ },
+ {
+ "epoch": 50.866375545851525,
+ "grad_norm": 0.01693476364016533,
+ "learning_rate": 0.0006,
+ "loss": 4.768477439880371,
+ "step": 3662
+ },
+ {
+ "epoch": 50.880349344978164,
+ "grad_norm": 0.01749575324356556,
+ "learning_rate": 0.0006,
+ "loss": 4.779143333435059,
+ "step": 3663
+ },
+ {
+ "epoch": 50.8943231441048,
+ "grad_norm": 0.016851043328642845,
+ "learning_rate": 0.0006,
+ "loss": 4.780387878417969,
+ "step": 3664
+ },
+ {
+ "epoch": 50.90829694323144,
+ "grad_norm": 0.019045202061533928,
+ "learning_rate": 0.0006,
+ "loss": 4.769913673400879,
+ "step": 3665
+ },
+ {
+ "epoch": 50.92227074235808,
+ "grad_norm": 0.01954123005270958,
+ "learning_rate": 0.0006,
+ "loss": 4.769009590148926,
+ "step": 3666
+ },
+ {
+ "epoch": 50.93624454148472,
+ "grad_norm": 0.022527096793055534,
+ "learning_rate": 0.0006,
+ "loss": 4.789215087890625,
+ "step": 3667
+ },
+ {
+ "epoch": 50.95021834061135,
+ "grad_norm": 0.02871634066104889,
+ "learning_rate": 0.0006,
+ "loss": 4.749727725982666,
+ "step": 3668
+ },
+ {
+ "epoch": 50.96419213973799,
+ "grad_norm": 0.04045053571462631,
+ "learning_rate": 0.0006,
+ "loss": 4.909038543701172,
+ "step": 3669
+ },
+ {
+ "epoch": 50.97816593886463,
+ "grad_norm": 0.04156879335641861,
+ "learning_rate": 0.0006,
+ "loss": 4.876831531524658,
+ "step": 3670
+ },
+ {
+ "epoch": 50.992139737991266,
+ "grad_norm": 0.02621397376060486,
+ "learning_rate": 0.0006,
+ "loss": 4.797173500061035,
+ "step": 3671
+ },
+ {
+ "epoch": 51.0,
+ "grad_norm": 0.03389597684144974,
+ "learning_rate": 0.0006,
+ "loss": 4.8428850173950195,
+ "step": 3672
+ },
+ {
+ "epoch": 51.0,
+ "eval_loss": 4.989261627197266,
+ "eval_runtime": 56.358,
+ "eval_samples_per_second": 43.33,
+ "eval_steps_per_second": 1.366,
+ "step": 3672
+ },
+ {
+ "epoch": 51.01397379912664,
+ "grad_norm": 0.03216972202062607,
+ "learning_rate": 0.0006,
+ "loss": 4.755102634429932,
+ "step": 3673
+ },
+ {
+ "epoch": 51.02794759825328,
+ "grad_norm": 0.03190374746918678,
+ "learning_rate": 0.0006,
+ "loss": 4.720424652099609,
+ "step": 3674
+ },
+ {
+ "epoch": 51.041921397379916,
+ "grad_norm": 0.031727299094200134,
+ "learning_rate": 0.0006,
+ "loss": 4.803277969360352,
+ "step": 3675
+ },
+ {
+ "epoch": 51.05589519650655,
+ "grad_norm": 0.02473079040646553,
+ "learning_rate": 0.0006,
+ "loss": 4.761087417602539,
+ "step": 3676
+ },
+ {
+ "epoch": 51.069868995633186,
+ "grad_norm": 0.020360209047794342,
+ "learning_rate": 0.0006,
+ "loss": 4.753884792327881,
+ "step": 3677
+ },
+ {
+ "epoch": 51.083842794759825,
+ "grad_norm": 0.02078923024237156,
+ "learning_rate": 0.0006,
+ "loss": 4.647130489349365,
+ "step": 3678
+ },
+ {
+ "epoch": 51.09781659388646,
+ "grad_norm": 0.019387200474739075,
+ "learning_rate": 0.0006,
+ "loss": 4.74135160446167,
+ "step": 3679
+ },
+ {
+ "epoch": 51.1117903930131,
+ "grad_norm": 0.01874949410557747,
+ "learning_rate": 0.0006,
+ "loss": 4.795329570770264,
+ "step": 3680
+ },
+ {
+ "epoch": 51.12576419213974,
+ "grad_norm": 0.019331837072968483,
+ "learning_rate": 0.0006,
+ "loss": 4.897967338562012,
+ "step": 3681
+ },
+ {
+ "epoch": 51.13973799126637,
+ "grad_norm": 0.017066050320863724,
+ "learning_rate": 0.0006,
+ "loss": 4.825351715087891,
+ "step": 3682
+ },
+ {
+ "epoch": 51.15371179039301,
+ "grad_norm": 0.01757810264825821,
+ "learning_rate": 0.0006,
+ "loss": 4.719335079193115,
+ "step": 3683
+ },
+ {
+ "epoch": 51.16768558951965,
+ "grad_norm": 0.018174266442656517,
+ "learning_rate": 0.0006,
+ "loss": 4.732680320739746,
+ "step": 3684
+ },
+ {
+ "epoch": 51.18165938864629,
+ "grad_norm": 0.018497727811336517,
+ "learning_rate": 0.0006,
+ "loss": 4.731050491333008,
+ "step": 3685
+ },
+ {
+ "epoch": 51.19563318777293,
+ "grad_norm": 0.0186338871717453,
+ "learning_rate": 0.0006,
+ "loss": 4.748035430908203,
+ "step": 3686
+ },
+ {
+ "epoch": 51.209606986899566,
+ "grad_norm": 0.018531711772084236,
+ "learning_rate": 0.0006,
+ "loss": 4.9007463455200195,
+ "step": 3687
+ },
+ {
+ "epoch": 51.223580786026204,
+ "grad_norm": 0.018221354112029076,
+ "learning_rate": 0.0006,
+ "loss": 4.84077262878418,
+ "step": 3688
+ },
+ {
+ "epoch": 51.237554585152836,
+ "grad_norm": 0.0203656405210495,
+ "learning_rate": 0.0006,
+ "loss": 4.73009729385376,
+ "step": 3689
+ },
+ {
+ "epoch": 51.251528384279474,
+ "grad_norm": 0.01964893378317356,
+ "learning_rate": 0.0006,
+ "loss": 4.772583961486816,
+ "step": 3690
+ },
+ {
+ "epoch": 51.26550218340611,
+ "grad_norm": 0.019606510177254677,
+ "learning_rate": 0.0006,
+ "loss": 4.599812030792236,
+ "step": 3691
+ },
+ {
+ "epoch": 51.27947598253275,
+ "grad_norm": 0.020501255989074707,
+ "learning_rate": 0.0006,
+ "loss": 4.809898853302002,
+ "step": 3692
+ },
+ {
+ "epoch": 51.29344978165939,
+ "grad_norm": 0.01870838552713394,
+ "learning_rate": 0.0006,
+ "loss": 4.741898059844971,
+ "step": 3693
+ },
+ {
+ "epoch": 51.30742358078603,
+ "grad_norm": 0.01802559196949005,
+ "learning_rate": 0.0006,
+ "loss": 4.746562480926514,
+ "step": 3694
+ },
+ {
+ "epoch": 51.32139737991266,
+ "grad_norm": 0.02053048275411129,
+ "learning_rate": 0.0006,
+ "loss": 4.69932746887207,
+ "step": 3695
+ },
+ {
+ "epoch": 51.3353711790393,
+ "grad_norm": 0.02661825157701969,
+ "learning_rate": 0.0006,
+ "loss": 4.68748664855957,
+ "step": 3696
+ },
+ {
+ "epoch": 51.34934497816594,
+ "grad_norm": 0.028269147500395775,
+ "learning_rate": 0.0006,
+ "loss": 4.731565475463867,
+ "step": 3697
+ },
+ {
+ "epoch": 51.36331877729258,
+ "grad_norm": 0.02650964818894863,
+ "learning_rate": 0.0006,
+ "loss": 4.733857154846191,
+ "step": 3698
+ },
+ {
+ "epoch": 51.377292576419215,
+ "grad_norm": 0.023696066811680794,
+ "learning_rate": 0.0006,
+ "loss": 4.812362194061279,
+ "step": 3699
+ },
+ {
+ "epoch": 51.391266375545854,
+ "grad_norm": 0.02365880273282528,
+ "learning_rate": 0.0006,
+ "loss": 4.733426094055176,
+ "step": 3700
+ },
+ {
+ "epoch": 51.40524017467249,
+ "grad_norm": 0.026249447837471962,
+ "learning_rate": 0.0006,
+ "loss": 4.715622901916504,
+ "step": 3701
+ },
+ {
+ "epoch": 51.419213973799124,
+ "grad_norm": 0.03301899880170822,
+ "learning_rate": 0.0006,
+ "loss": 4.7223663330078125,
+ "step": 3702
+ },
+ {
+ "epoch": 51.43318777292576,
+ "grad_norm": 0.044686660170555115,
+ "learning_rate": 0.0006,
+ "loss": 4.8121161460876465,
+ "step": 3703
+ },
+ {
+ "epoch": 51.4471615720524,
+ "grad_norm": 0.04733522981405258,
+ "learning_rate": 0.0006,
+ "loss": 4.841394424438477,
+ "step": 3704
+ },
+ {
+ "epoch": 51.46113537117904,
+ "grad_norm": 0.0395209975540638,
+ "learning_rate": 0.0006,
+ "loss": 4.773837089538574,
+ "step": 3705
+ },
+ {
+ "epoch": 51.47510917030568,
+ "grad_norm": 0.032790958881378174,
+ "learning_rate": 0.0006,
+ "loss": 4.833066463470459,
+ "step": 3706
+ },
+ {
+ "epoch": 51.48908296943232,
+ "grad_norm": 0.030578427016735077,
+ "learning_rate": 0.0006,
+ "loss": 4.685190200805664,
+ "step": 3707
+ },
+ {
+ "epoch": 51.50305676855895,
+ "grad_norm": 0.031619083136320114,
+ "learning_rate": 0.0006,
+ "loss": 4.740378379821777,
+ "step": 3708
+ },
+ {
+ "epoch": 51.51703056768559,
+ "grad_norm": 0.02385825850069523,
+ "learning_rate": 0.0006,
+ "loss": 4.812911510467529,
+ "step": 3709
+ },
+ {
+ "epoch": 51.531004366812226,
+ "grad_norm": 0.023523425683379173,
+ "learning_rate": 0.0006,
+ "loss": 4.669580459594727,
+ "step": 3710
+ },
+ {
+ "epoch": 51.544978165938865,
+ "grad_norm": 0.021316280588507652,
+ "learning_rate": 0.0006,
+ "loss": 4.76888370513916,
+ "step": 3711
+ },
+ {
+ "epoch": 51.5589519650655,
+ "grad_norm": 0.024663880467414856,
+ "learning_rate": 0.0006,
+ "loss": 4.702592372894287,
+ "step": 3712
+ },
+ {
+ "epoch": 51.57292576419214,
+ "grad_norm": 0.024173280224204063,
+ "learning_rate": 0.0006,
+ "loss": 4.710604667663574,
+ "step": 3713
+ },
+ {
+ "epoch": 51.58689956331878,
+ "grad_norm": 0.02754114381968975,
+ "learning_rate": 0.0006,
+ "loss": 4.746356964111328,
+ "step": 3714
+ },
+ {
+ "epoch": 51.60087336244541,
+ "grad_norm": 0.027903622016310692,
+ "learning_rate": 0.0006,
+ "loss": 4.7546491622924805,
+ "step": 3715
+ },
+ {
+ "epoch": 51.61484716157205,
+ "grad_norm": 0.024871617555618286,
+ "learning_rate": 0.0006,
+ "loss": 4.605746269226074,
+ "step": 3716
+ },
+ {
+ "epoch": 51.62882096069869,
+ "grad_norm": 0.021239031106233597,
+ "learning_rate": 0.0006,
+ "loss": 4.6942458152771,
+ "step": 3717
+ },
+ {
+ "epoch": 51.64279475982533,
+ "grad_norm": 0.01925622671842575,
+ "learning_rate": 0.0006,
+ "loss": 4.593558311462402,
+ "step": 3718
+ },
+ {
+ "epoch": 51.65676855895197,
+ "grad_norm": 0.01956283673644066,
+ "learning_rate": 0.0006,
+ "loss": 4.631172180175781,
+ "step": 3719
+ },
+ {
+ "epoch": 51.670742358078606,
+ "grad_norm": 0.02269871160387993,
+ "learning_rate": 0.0006,
+ "loss": 4.729236602783203,
+ "step": 3720
+ },
+ {
+ "epoch": 51.68471615720524,
+ "grad_norm": 0.02253916673362255,
+ "learning_rate": 0.0006,
+ "loss": 4.7191033363342285,
+ "step": 3721
+ },
+ {
+ "epoch": 51.698689956331876,
+ "grad_norm": 0.026231050491333008,
+ "learning_rate": 0.0006,
+ "loss": 4.811850547790527,
+ "step": 3722
+ },
+ {
+ "epoch": 51.712663755458514,
+ "grad_norm": 0.02558821812272072,
+ "learning_rate": 0.0006,
+ "loss": 4.681466102600098,
+ "step": 3723
+ },
+ {
+ "epoch": 51.72663755458515,
+ "grad_norm": 0.02124209702014923,
+ "learning_rate": 0.0006,
+ "loss": 4.693664073944092,
+ "step": 3724
+ },
+ {
+ "epoch": 51.74061135371179,
+ "grad_norm": 0.016638319939374924,
+ "learning_rate": 0.0006,
+ "loss": 4.771440029144287,
+ "step": 3725
+ },
+ {
+ "epoch": 51.75458515283843,
+ "grad_norm": 0.018048347905278206,
+ "learning_rate": 0.0006,
+ "loss": 4.643252372741699,
+ "step": 3726
+ },
+ {
+ "epoch": 51.76855895196506,
+ "grad_norm": 0.019904717803001404,
+ "learning_rate": 0.0006,
+ "loss": 4.66530704498291,
+ "step": 3727
+ },
+ {
+ "epoch": 51.7825327510917,
+ "grad_norm": 0.0194614976644516,
+ "learning_rate": 0.0006,
+ "loss": 4.711943626403809,
+ "step": 3728
+ },
+ {
+ "epoch": 51.79650655021834,
+ "grad_norm": 0.019175555557012558,
+ "learning_rate": 0.0006,
+ "loss": 4.721931457519531,
+ "step": 3729
+ },
+ {
+ "epoch": 51.81048034934498,
+ "grad_norm": 0.02093447372317314,
+ "learning_rate": 0.0006,
+ "loss": 4.633059024810791,
+ "step": 3730
+ },
+ {
+ "epoch": 51.82445414847162,
+ "grad_norm": 0.01930283196270466,
+ "learning_rate": 0.0006,
+ "loss": 4.659373760223389,
+ "step": 3731
+ },
+ {
+ "epoch": 51.838427947598255,
+ "grad_norm": 0.015350676141679287,
+ "learning_rate": 0.0006,
+ "loss": 4.763708114624023,
+ "step": 3732
+ },
+ {
+ "epoch": 51.852401746724894,
+ "grad_norm": 0.014139448292553425,
+ "learning_rate": 0.0006,
+ "loss": 4.753891468048096,
+ "step": 3733
+ },
+ {
+ "epoch": 51.866375545851525,
+ "grad_norm": 0.016215872019529343,
+ "learning_rate": 0.0006,
+ "loss": 4.658967018127441,
+ "step": 3734
+ },
+ {
+ "epoch": 51.880349344978164,
+ "grad_norm": 0.017328619956970215,
+ "learning_rate": 0.0006,
+ "loss": 4.696579933166504,
+ "step": 3735
+ },
+ {
+ "epoch": 51.8943231441048,
+ "grad_norm": 0.018103396520018578,
+ "learning_rate": 0.0006,
+ "loss": 4.627464771270752,
+ "step": 3736
+ },
+ {
+ "epoch": 51.90829694323144,
+ "grad_norm": 0.01715235412120819,
+ "learning_rate": 0.0006,
+ "loss": 4.7368011474609375,
+ "step": 3737
+ },
+ {
+ "epoch": 51.92227074235808,
+ "grad_norm": 0.014790094457566738,
+ "learning_rate": 0.0006,
+ "loss": 4.629441738128662,
+ "step": 3738
+ },
+ {
+ "epoch": 51.93624454148472,
+ "grad_norm": 0.017114296555519104,
+ "learning_rate": 0.0006,
+ "loss": 4.761256694793701,
+ "step": 3739
+ },
+ {
+ "epoch": 51.95021834061135,
+ "grad_norm": 0.019009845331311226,
+ "learning_rate": 0.0006,
+ "loss": 4.789155006408691,
+ "step": 3740
+ },
+ {
+ "epoch": 51.96419213973799,
+ "grad_norm": 0.019696692004799843,
+ "learning_rate": 0.0006,
+ "loss": 4.721269607543945,
+ "step": 3741
+ },
+ {
+ "epoch": 51.97816593886463,
+ "grad_norm": 0.020542822778224945,
+ "learning_rate": 0.0006,
+ "loss": 4.687038898468018,
+ "step": 3742
+ },
+ {
+ "epoch": 51.992139737991266,
+ "grad_norm": 0.021267160773277283,
+ "learning_rate": 0.0006,
+ "loss": 4.747860908508301,
+ "step": 3743
+ },
+ {
+ "epoch": 52.0,
+ "grad_norm": 0.025049567222595215,
+ "learning_rate": 0.0006,
+ "loss": 4.8473100662231445,
+ "step": 3744
+ },
+ {
+ "epoch": 52.0,
+ "eval_loss": 4.943043231964111,
+ "eval_runtime": 56.4929,
+ "eval_samples_per_second": 43.227,
+ "eval_steps_per_second": 1.363,
+ "step": 3744
+ },
+ {
+ "epoch": 52.01397379912664,
+ "grad_norm": 0.03054986707866192,
+ "learning_rate": 0.0006,
+ "loss": 4.856931686401367,
+ "step": 3745
+ },
+ {
+ "epoch": 52.02794759825328,
+ "grad_norm": 0.03838375583291054,
+ "learning_rate": 0.0006,
+ "loss": 4.699807167053223,
+ "step": 3746
+ },
+ {
+ "epoch": 52.041921397379916,
+ "grad_norm": 0.03383488208055496,
+ "learning_rate": 0.0006,
+ "loss": 4.6626787185668945,
+ "step": 3747
+ },
+ {
+ "epoch": 52.05589519650655,
+ "grad_norm": 0.0201561376452446,
+ "learning_rate": 0.0006,
+ "loss": 4.621831893920898,
+ "step": 3748
+ },
+ {
+ "epoch": 52.069868995633186,
+ "grad_norm": 0.028749438002705574,
+ "learning_rate": 0.0006,
+ "loss": 4.682540416717529,
+ "step": 3749
+ },
+ {
+ "epoch": 52.083842794759825,
+ "grad_norm": 0.02681174874305725,
+ "learning_rate": 0.0006,
+ "loss": 4.629997253417969,
+ "step": 3750
+ },
+ {
+ "epoch": 52.09781659388646,
+ "grad_norm": 0.024063318967819214,
+ "learning_rate": 0.0006,
+ "loss": 4.573021411895752,
+ "step": 3751
+ },
+ {
+ "epoch": 52.1117903930131,
+ "grad_norm": 0.02324330434203148,
+ "learning_rate": 0.0006,
+ "loss": 4.739465236663818,
+ "step": 3752
+ },
+ {
+ "epoch": 52.12576419213974,
+ "grad_norm": 0.0225426834076643,
+ "learning_rate": 0.0006,
+ "loss": 4.7335333824157715,
+ "step": 3753
+ },
+ {
+ "epoch": 52.13973799126637,
+ "grad_norm": 0.019593428820371628,
+ "learning_rate": 0.0006,
+ "loss": 4.704958438873291,
+ "step": 3754
+ },
+ {
+ "epoch": 52.15371179039301,
+ "grad_norm": 0.020777082070708275,
+ "learning_rate": 0.0006,
+ "loss": 4.638696670532227,
+ "step": 3755
+ },
+ {
+ "epoch": 52.16768558951965,
+ "grad_norm": 0.021525604650378227,
+ "learning_rate": 0.0006,
+ "loss": 4.643706798553467,
+ "step": 3756
+ },
+ {
+ "epoch": 52.18165938864629,
+ "grad_norm": 0.018943654373288155,
+ "learning_rate": 0.0006,
+ "loss": 4.5173211097717285,
+ "step": 3757
+ },
+ {
+ "epoch": 52.19563318777293,
+ "grad_norm": 0.01856580562889576,
+ "learning_rate": 0.0006,
+ "loss": 4.678788661956787,
+ "step": 3758
+ },
+ {
+ "epoch": 52.209606986899566,
+ "grad_norm": 0.019417259842157364,
+ "learning_rate": 0.0006,
+ "loss": 4.587244987487793,
+ "step": 3759
+ },
+ {
+ "epoch": 52.223580786026204,
+ "grad_norm": 0.01911911740899086,
+ "learning_rate": 0.0006,
+ "loss": 4.688503265380859,
+ "step": 3760
+ },
+ {
+ "epoch": 52.237554585152836,
+ "grad_norm": 0.021017879247665405,
+ "learning_rate": 0.0006,
+ "loss": 4.65829610824585,
+ "step": 3761
+ },
+ {
+ "epoch": 52.251528384279474,
+ "grad_norm": 0.02048875205218792,
+ "learning_rate": 0.0006,
+ "loss": 4.624055862426758,
+ "step": 3762
+ },
+ {
+ "epoch": 52.26550218340611,
+ "grad_norm": 0.01932772807776928,
+ "learning_rate": 0.0006,
+ "loss": 4.556562900543213,
+ "step": 3763
+ },
+ {
+ "epoch": 52.27947598253275,
+ "grad_norm": 0.01871614344418049,
+ "learning_rate": 0.0006,
+ "loss": 4.581201076507568,
+ "step": 3764
+ },
+ {
+ "epoch": 52.29344978165939,
+ "grad_norm": 0.017253106459975243,
+ "learning_rate": 0.0006,
+ "loss": 4.605219841003418,
+ "step": 3765
+ },
+ {
+ "epoch": 52.30742358078603,
+ "grad_norm": 0.01831238530576229,
+ "learning_rate": 0.0006,
+ "loss": 4.759900093078613,
+ "step": 3766
+ },
+ {
+ "epoch": 52.32139737991266,
+ "grad_norm": 0.017946207895874977,
+ "learning_rate": 0.0006,
+ "loss": 4.5598344802856445,
+ "step": 3767
+ },
+ {
+ "epoch": 52.3353711790393,
+ "grad_norm": 0.01680106669664383,
+ "learning_rate": 0.0006,
+ "loss": 4.605034828186035,
+ "step": 3768
+ },
+ {
+ "epoch": 52.34934497816594,
+ "grad_norm": 0.021482540294528008,
+ "learning_rate": 0.0006,
+ "loss": 4.620411396026611,
+ "step": 3769
+ },
+ {
+ "epoch": 52.36331877729258,
+ "grad_norm": 0.02210252732038498,
+ "learning_rate": 0.0006,
+ "loss": 4.667577743530273,
+ "step": 3770
+ },
+ {
+ "epoch": 52.377292576419215,
+ "grad_norm": 0.02237485907971859,
+ "learning_rate": 0.0006,
+ "loss": 4.628812789916992,
+ "step": 3771
+ },
+ {
+ "epoch": 52.391266375545854,
+ "grad_norm": 0.02196541801095009,
+ "learning_rate": 0.0006,
+ "loss": 4.572988510131836,
+ "step": 3772
+ },
+ {
+ "epoch": 52.40524017467249,
+ "grad_norm": 0.01930229924619198,
+ "learning_rate": 0.0006,
+ "loss": 4.588339805603027,
+ "step": 3773
+ },
+ {
+ "epoch": 52.419213973799124,
+ "grad_norm": 0.02222434990108013,
+ "learning_rate": 0.0006,
+ "loss": 4.766482353210449,
+ "step": 3774
+ },
+ {
+ "epoch": 52.43318777292576,
+ "grad_norm": 0.028806306421756744,
+ "learning_rate": 0.0006,
+ "loss": 4.617892742156982,
+ "step": 3775
+ },
+ {
+ "epoch": 52.4471615720524,
+ "grad_norm": 0.03306042402982712,
+ "learning_rate": 0.0006,
+ "loss": 4.665334701538086,
+ "step": 3776
+ },
+ {
+ "epoch": 52.46113537117904,
+ "grad_norm": 0.0358956903219223,
+ "learning_rate": 0.0006,
+ "loss": 4.566668510437012,
+ "step": 3777
+ },
+ {
+ "epoch": 52.47510917030568,
+ "grad_norm": 0.030406497418880463,
+ "learning_rate": 0.0006,
+ "loss": 4.513612270355225,
+ "step": 3778
+ },
+ {
+ "epoch": 52.48908296943232,
+ "grad_norm": 0.030295712873339653,
+ "learning_rate": 0.0006,
+ "loss": 4.640189170837402,
+ "step": 3779
+ },
+ {
+ "epoch": 52.50305676855895,
+ "grad_norm": 0.024981003254652023,
+ "learning_rate": 0.0006,
+ "loss": 4.57558012008667,
+ "step": 3780
+ },
+ {
+ "epoch": 52.51703056768559,
+ "grad_norm": 0.024990713223814964,
+ "learning_rate": 0.0006,
+ "loss": 4.708589553833008,
+ "step": 3781
+ },
+ {
+ "epoch": 52.531004366812226,
+ "grad_norm": 0.025034775957465172,
+ "learning_rate": 0.0006,
+ "loss": 4.70468282699585,
+ "step": 3782
+ },
+ {
+ "epoch": 52.544978165938865,
+ "grad_norm": 0.028243759647011757,
+ "learning_rate": 0.0006,
+ "loss": 4.712060451507568,
+ "step": 3783
+ },
+ {
+ "epoch": 52.5589519650655,
+ "grad_norm": 0.026984544470906258,
+ "learning_rate": 0.0006,
+ "loss": 4.505004405975342,
+ "step": 3784
+ },
+ {
+ "epoch": 52.57292576419214,
+ "grad_norm": 0.050843529403209686,
+ "learning_rate": 0.0006,
+ "loss": 4.593271255493164,
+ "step": 3785
+ },
+ {
+ "epoch": 52.58689956331878,
+ "grad_norm": 0.25422802567481995,
+ "learning_rate": 0.0006,
+ "loss": 4.7178802490234375,
+ "step": 3786
+ },
+ {
+ "epoch": 52.60087336244541,
+ "grad_norm": 0.6078406572341919,
+ "learning_rate": 0.0006,
+ "loss": 6.144372940063477,
+ "step": 3787
+ },
+ {
+ "epoch": 52.61484716157205,
+ "grad_norm": 0.25933101773262024,
+ "learning_rate": 0.0006,
+ "loss": 5.971083641052246,
+ "step": 3788
+ },
+ {
+ "epoch": 52.62882096069869,
+ "grad_norm": 0.294657826423645,
+ "learning_rate": 0.0006,
+ "loss": 6.204090118408203,
+ "step": 3789
+ },
+ {
+ "epoch": 52.64279475982533,
+ "grad_norm": 0.23641522228717804,
+ "learning_rate": 0.0006,
+ "loss": 6.869574546813965,
+ "step": 3790
+ },
+ {
+ "epoch": 52.65676855895197,
+ "grad_norm": 0.23717311024665833,
+ "learning_rate": 0.0006,
+ "loss": 6.739034175872803,
+ "step": 3791
+ },
+ {
+ "epoch": 52.670742358078606,
+ "grad_norm": 0.10300362855195999,
+ "learning_rate": 0.0006,
+ "loss": 6.49191951751709,
+ "step": 3792
+ },
+ {
+ "epoch": 52.68471615720524,
+ "grad_norm": 0.1484968513250351,
+ "learning_rate": 0.0006,
+ "loss": 6.6021270751953125,
+ "step": 3793
+ },
+ {
+ "epoch": 52.698689956331876,
+ "grad_norm": 0.08052115142345428,
+ "learning_rate": 0.0006,
+ "loss": 6.432796478271484,
+ "step": 3794
+ },
+ {
+ "epoch": 52.712663755458514,
+ "grad_norm": 0.06869253516197205,
+ "learning_rate": 0.0006,
+ "loss": 6.212784767150879,
+ "step": 3795
+ },
+ {
+ "epoch": 52.72663755458515,
+ "grad_norm": 0.05589582398533821,
+ "learning_rate": 0.0006,
+ "loss": 6.168288707733154,
+ "step": 3796
+ },
+ {
+ "epoch": 52.74061135371179,
+ "grad_norm": 0.04840940609574318,
+ "learning_rate": 0.0006,
+ "loss": 6.163174629211426,
+ "step": 3797
+ },
+ {
+ "epoch": 52.75458515283843,
+ "grad_norm": 0.04578601196408272,
+ "learning_rate": 0.0006,
+ "loss": 6.038390636444092,
+ "step": 3798
+ },
+ {
+ "epoch": 52.76855895196506,
+ "grad_norm": 0.04102053865790367,
+ "learning_rate": 0.0006,
+ "loss": 5.929539680480957,
+ "step": 3799
+ },
+ {
+ "epoch": 52.7825327510917,
+ "grad_norm": 0.038398485630750656,
+ "learning_rate": 0.0006,
+ "loss": 6.007378101348877,
+ "step": 3800
+ },
+ {
+ "epoch": 52.79650655021834,
+ "grad_norm": 0.03466642275452614,
+ "learning_rate": 0.0006,
+ "loss": 5.767291069030762,
+ "step": 3801
+ },
+ {
+ "epoch": 52.81048034934498,
+ "grad_norm": 0.03328905999660492,
+ "learning_rate": 0.0006,
+ "loss": 5.785032272338867,
+ "step": 3802
+ },
+ {
+ "epoch": 52.82445414847162,
+ "grad_norm": 0.03186681121587753,
+ "learning_rate": 0.0006,
+ "loss": 5.6942057609558105,
+ "step": 3803
+ },
+ {
+ "epoch": 52.838427947598255,
+ "grad_norm": 0.030553551390767097,
+ "learning_rate": 0.0006,
+ "loss": 5.680774688720703,
+ "step": 3804
+ },
+ {
+ "epoch": 52.852401746724894,
+ "grad_norm": 0.02578839473426342,
+ "learning_rate": 0.0006,
+ "loss": 5.703461647033691,
+ "step": 3805
+ },
+ {
+ "epoch": 52.866375545851525,
+ "grad_norm": 0.026261312887072563,
+ "learning_rate": 0.0006,
+ "loss": 5.642549514770508,
+ "step": 3806
+ },
+ {
+ "epoch": 52.880349344978164,
+ "grad_norm": 0.02782430872321129,
+ "learning_rate": 0.0006,
+ "loss": 5.563597679138184,
+ "step": 3807
+ },
+ {
+ "epoch": 52.8943231441048,
+ "grad_norm": 0.024467799812555313,
+ "learning_rate": 0.0006,
+ "loss": 5.583395957946777,
+ "step": 3808
+ },
+ {
+ "epoch": 52.90829694323144,
+ "grad_norm": 0.022602897137403488,
+ "learning_rate": 0.0006,
+ "loss": 5.50742769241333,
+ "step": 3809
+ },
+ {
+ "epoch": 52.92227074235808,
+ "grad_norm": 0.024292029440402985,
+ "learning_rate": 0.0006,
+ "loss": 5.367804527282715,
+ "step": 3810
+ },
+ {
+ "epoch": 52.93624454148472,
+ "grad_norm": 0.021062295883893967,
+ "learning_rate": 0.0006,
+ "loss": 5.423705101013184,
+ "step": 3811
+ },
+ {
+ "epoch": 52.95021834061135,
+ "grad_norm": 0.022226469591259956,
+ "learning_rate": 0.0006,
+ "loss": 5.344039440155029,
+ "step": 3812
+ },
+ {
+ "epoch": 52.96419213973799,
+ "grad_norm": 0.023691752925515175,
+ "learning_rate": 0.0006,
+ "loss": 5.2581892013549805,
+ "step": 3813
+ },
+ {
+ "epoch": 52.97816593886463,
+ "grad_norm": 0.02181999757885933,
+ "learning_rate": 0.0006,
+ "loss": 5.2710113525390625,
+ "step": 3814
+ },
+ {
+ "epoch": 52.992139737991266,
+ "grad_norm": 0.025295857340097427,
+ "learning_rate": 0.0006,
+ "loss": 5.278355598449707,
+ "step": 3815
+ },
+ {
+ "epoch": 53.0,
+ "grad_norm": 0.02216828241944313,
+ "learning_rate": 0.0006,
+ "loss": 5.127945899963379,
+ "step": 3816
+ },
+ {
+ "epoch": 53.0,
+ "eval_loss": 5.3411736488342285,
+ "eval_runtime": 56.2837,
+ "eval_samples_per_second": 43.387,
+ "eval_steps_per_second": 1.368,
+ "step": 3816
+ },
+ {
+ "epoch": 53.01397379912664,
+ "grad_norm": 0.022496026009321213,
+ "learning_rate": 0.0006,
+ "loss": 5.19655704498291,
+ "step": 3817
+ },
+ {
+ "epoch": 53.02794759825328,
+ "grad_norm": 0.03201168775558472,
+ "learning_rate": 0.0006,
+ "loss": 5.094621181488037,
+ "step": 3818
+ },
+ {
+ "epoch": 53.041921397379916,
+ "grad_norm": 0.02683178335428238,
+ "learning_rate": 0.0006,
+ "loss": 5.19899845123291,
+ "step": 3819
+ },
+ {
+ "epoch": 53.05589519650655,
+ "grad_norm": 0.022659169510006905,
+ "learning_rate": 0.0006,
+ "loss": 5.184103012084961,
+ "step": 3820
+ },
+ {
+ "epoch": 53.069868995633186,
+ "grad_norm": 0.024784443899989128,
+ "learning_rate": 0.0006,
+ "loss": 5.124286651611328,
+ "step": 3821
+ },
+ {
+ "epoch": 53.083842794759825,
+ "grad_norm": 0.028710370883345604,
+ "learning_rate": 0.0006,
+ "loss": 5.130426406860352,
+ "step": 3822
+ },
+ {
+ "epoch": 53.09781659388646,
+ "grad_norm": 0.02366088517010212,
+ "learning_rate": 0.0006,
+ "loss": 5.1453399658203125,
+ "step": 3823
+ },
+ {
+ "epoch": 53.1117903930131,
+ "grad_norm": 0.01813938096165657,
+ "learning_rate": 0.0006,
+ "loss": 5.05271053314209,
+ "step": 3824
+ },
+ {
+ "epoch": 53.12576419213974,
+ "grad_norm": 0.02051435597240925,
+ "learning_rate": 0.0006,
+ "loss": 5.023038864135742,
+ "step": 3825
+ },
+ {
+ "epoch": 53.13973799126637,
+ "grad_norm": 0.02185754105448723,
+ "learning_rate": 0.0006,
+ "loss": 4.8950018882751465,
+ "step": 3826
+ },
+ {
+ "epoch": 53.15371179039301,
+ "grad_norm": 0.017223743721842766,
+ "learning_rate": 0.0006,
+ "loss": 5.078621864318848,
+ "step": 3827
+ },
+ {
+ "epoch": 53.16768558951965,
+ "grad_norm": 0.01832217164337635,
+ "learning_rate": 0.0006,
+ "loss": 4.921341419219971,
+ "step": 3828
+ },
+ {
+ "epoch": 53.18165938864629,
+ "grad_norm": 0.02114005573093891,
+ "learning_rate": 0.0006,
+ "loss": 4.997865676879883,
+ "step": 3829
+ },
+ {
+ "epoch": 53.19563318777293,
+ "grad_norm": 0.021426431834697723,
+ "learning_rate": 0.0006,
+ "loss": 4.877785682678223,
+ "step": 3830
+ },
+ {
+ "epoch": 53.209606986899566,
+ "grad_norm": 0.021327560767531395,
+ "learning_rate": 0.0006,
+ "loss": 4.847512245178223,
+ "step": 3831
+ },
+ {
+ "epoch": 53.223580786026204,
+ "grad_norm": 0.027825530618429184,
+ "learning_rate": 0.0006,
+ "loss": 5.004823207855225,
+ "step": 3832
+ },
+ {
+ "epoch": 53.237554585152836,
+ "grad_norm": 0.03756481409072876,
+ "learning_rate": 0.0006,
+ "loss": 4.817290306091309,
+ "step": 3833
+ },
+ {
+ "epoch": 53.251528384279474,
+ "grad_norm": 0.04933721572160721,
+ "learning_rate": 0.0006,
+ "loss": 4.863317489624023,
+ "step": 3834
+ },
+ {
+ "epoch": 53.26550218340611,
+ "grad_norm": 0.04588694870471954,
+ "learning_rate": 0.0006,
+ "loss": 4.864742755889893,
+ "step": 3835
+ },
+ {
+ "epoch": 53.27947598253275,
+ "grad_norm": 0.02728264033794403,
+ "learning_rate": 0.0006,
+ "loss": 4.9068756103515625,
+ "step": 3836
+ },
+ {
+ "epoch": 53.29344978165939,
+ "grad_norm": 0.036240726709365845,
+ "learning_rate": 0.0006,
+ "loss": 4.902045249938965,
+ "step": 3837
+ },
+ {
+ "epoch": 53.30742358078603,
+ "grad_norm": 0.02703195810317993,
+ "learning_rate": 0.0006,
+ "loss": 4.908980369567871,
+ "step": 3838
+ },
+ {
+ "epoch": 53.32139737991266,
+ "grad_norm": 0.024894049391150475,
+ "learning_rate": 0.0006,
+ "loss": 4.794848442077637,
+ "step": 3839
+ },
+ {
+ "epoch": 53.3353711790393,
+ "grad_norm": 0.02389993891119957,
+ "learning_rate": 0.0006,
+ "loss": 4.8234734535217285,
+ "step": 3840
+ },
+ {
+ "epoch": 53.34934497816594,
+ "grad_norm": 0.02486400306224823,
+ "learning_rate": 0.0006,
+ "loss": 4.851463794708252,
+ "step": 3841
+ },
+ {
+ "epoch": 53.36331877729258,
+ "grad_norm": 0.02613813802599907,
+ "learning_rate": 0.0006,
+ "loss": 4.860444068908691,
+ "step": 3842
+ },
+ {
+ "epoch": 53.377292576419215,
+ "grad_norm": 0.0229744054377079,
+ "learning_rate": 0.0006,
+ "loss": 4.709606170654297,
+ "step": 3843
+ },
+ {
+ "epoch": 53.391266375545854,
+ "grad_norm": 0.02123042568564415,
+ "learning_rate": 0.0006,
+ "loss": 4.654613494873047,
+ "step": 3844
+ },
+ {
+ "epoch": 53.40524017467249,
+ "grad_norm": 0.017352338880300522,
+ "learning_rate": 0.0006,
+ "loss": 4.769761085510254,
+ "step": 3845
+ },
+ {
+ "epoch": 53.419213973799124,
+ "grad_norm": 0.019808724522590637,
+ "learning_rate": 0.0006,
+ "loss": 4.98514461517334,
+ "step": 3846
+ },
+ {
+ "epoch": 53.43318777292576,
+ "grad_norm": 0.018776152282953262,
+ "learning_rate": 0.0006,
+ "loss": 4.749990463256836,
+ "step": 3847
+ },
+ {
+ "epoch": 53.4471615720524,
+ "grad_norm": 0.02017521858215332,
+ "learning_rate": 0.0006,
+ "loss": 4.757381439208984,
+ "step": 3848
+ },
+ {
+ "epoch": 53.46113537117904,
+ "grad_norm": 0.019509416073560715,
+ "learning_rate": 0.0006,
+ "loss": 4.8111572265625,
+ "step": 3849
+ },
+ {
+ "epoch": 53.47510917030568,
+ "grad_norm": 0.015543580986559391,
+ "learning_rate": 0.0006,
+ "loss": 4.7501678466796875,
+ "step": 3850
+ },
+ {
+ "epoch": 53.48908296943232,
+ "grad_norm": 0.017941651865839958,
+ "learning_rate": 0.0006,
+ "loss": 4.701838493347168,
+ "step": 3851
+ },
+ {
+ "epoch": 53.50305676855895,
+ "grad_norm": 0.01678212359547615,
+ "learning_rate": 0.0006,
+ "loss": 4.787798881530762,
+ "step": 3852
+ },
+ {
+ "epoch": 53.51703056768559,
+ "grad_norm": 0.019050654023885727,
+ "learning_rate": 0.0006,
+ "loss": 4.754859924316406,
+ "step": 3853
+ },
+ {
+ "epoch": 53.531004366812226,
+ "grad_norm": 0.017406078055500984,
+ "learning_rate": 0.0006,
+ "loss": 4.752772808074951,
+ "step": 3854
+ },
+ {
+ "epoch": 53.544978165938865,
+ "grad_norm": 0.01773344725370407,
+ "learning_rate": 0.0006,
+ "loss": 4.651812553405762,
+ "step": 3855
+ },
+ {
+ "epoch": 53.5589519650655,
+ "grad_norm": 0.017221346497535706,
+ "learning_rate": 0.0006,
+ "loss": 4.792244911193848,
+ "step": 3856
+ },
+ {
+ "epoch": 53.57292576419214,
+ "grad_norm": 0.01722758449614048,
+ "learning_rate": 0.0006,
+ "loss": 4.7902960777282715,
+ "step": 3857
+ },
+ {
+ "epoch": 53.58689956331878,
+ "grad_norm": 0.020445192232728004,
+ "learning_rate": 0.0006,
+ "loss": 4.67392110824585,
+ "step": 3858
+ },
+ {
+ "epoch": 53.60087336244541,
+ "grad_norm": 0.02267114631831646,
+ "learning_rate": 0.0006,
+ "loss": 4.796089172363281,
+ "step": 3859
+ },
+ {
+ "epoch": 53.61484716157205,
+ "grad_norm": 0.027129942551255226,
+ "learning_rate": 0.0006,
+ "loss": 4.774303913116455,
+ "step": 3860
+ },
+ {
+ "epoch": 53.62882096069869,
+ "grad_norm": 0.025454839691519737,
+ "learning_rate": 0.0006,
+ "loss": 4.708813190460205,
+ "step": 3861
+ },
+ {
+ "epoch": 53.64279475982533,
+ "grad_norm": 0.019248757511377335,
+ "learning_rate": 0.0006,
+ "loss": 4.760467529296875,
+ "step": 3862
+ },
+ {
+ "epoch": 53.65676855895197,
+ "grad_norm": 0.01791176199913025,
+ "learning_rate": 0.0006,
+ "loss": 4.789203643798828,
+ "step": 3863
+ },
+ {
+ "epoch": 53.670742358078606,
+ "grad_norm": 0.016306612640619278,
+ "learning_rate": 0.0006,
+ "loss": 4.732605934143066,
+ "step": 3864
+ },
+ {
+ "epoch": 53.68471615720524,
+ "grad_norm": 0.017836976796388626,
+ "learning_rate": 0.0006,
+ "loss": 4.691871643066406,
+ "step": 3865
+ },
+ {
+ "epoch": 53.698689956331876,
+ "grad_norm": 0.015102861449122429,
+ "learning_rate": 0.0006,
+ "loss": 4.724156856536865,
+ "step": 3866
+ },
+ {
+ "epoch": 53.712663755458514,
+ "grad_norm": 0.017458142712712288,
+ "learning_rate": 0.0006,
+ "loss": 4.73524808883667,
+ "step": 3867
+ },
+ {
+ "epoch": 53.72663755458515,
+ "grad_norm": 0.01591932773590088,
+ "learning_rate": 0.0006,
+ "loss": 4.705888748168945,
+ "step": 3868
+ },
+ {
+ "epoch": 53.74061135371179,
+ "grad_norm": 0.01713593676686287,
+ "learning_rate": 0.0006,
+ "loss": 4.644516944885254,
+ "step": 3869
+ },
+ {
+ "epoch": 53.75458515283843,
+ "grad_norm": 0.01905885338783264,
+ "learning_rate": 0.0006,
+ "loss": 4.698416233062744,
+ "step": 3870
+ },
+ {
+ "epoch": 53.76855895196506,
+ "grad_norm": 0.01766437292098999,
+ "learning_rate": 0.0006,
+ "loss": 4.6796555519104,
+ "step": 3871
+ },
+ {
+ "epoch": 53.7825327510917,
+ "grad_norm": 0.01552243810147047,
+ "learning_rate": 0.0006,
+ "loss": 4.78138542175293,
+ "step": 3872
+ },
+ {
+ "epoch": 53.79650655021834,
+ "grad_norm": 0.014191331341862679,
+ "learning_rate": 0.0006,
+ "loss": 4.703588485717773,
+ "step": 3873
+ },
+ {
+ "epoch": 53.81048034934498,
+ "grad_norm": 0.015872279182076454,
+ "learning_rate": 0.0006,
+ "loss": 4.705611228942871,
+ "step": 3874
+ },
+ {
+ "epoch": 53.82445414847162,
+ "grad_norm": 0.01787560060620308,
+ "learning_rate": 0.0006,
+ "loss": 4.8227949142456055,
+ "step": 3875
+ },
+ {
+ "epoch": 53.838427947598255,
+ "grad_norm": 0.0158209390938282,
+ "learning_rate": 0.0006,
+ "loss": 4.645227432250977,
+ "step": 3876
+ },
+ {
+ "epoch": 53.852401746724894,
+ "grad_norm": 0.014362678863108158,
+ "learning_rate": 0.0006,
+ "loss": 4.7225422859191895,
+ "step": 3877
+ },
+ {
+ "epoch": 53.866375545851525,
+ "grad_norm": 0.014899161644279957,
+ "learning_rate": 0.0006,
+ "loss": 4.685914039611816,
+ "step": 3878
+ },
+ {
+ "epoch": 53.880349344978164,
+ "grad_norm": 0.01628505438566208,
+ "learning_rate": 0.0006,
+ "loss": 4.674065589904785,
+ "step": 3879
+ },
+ {
+ "epoch": 53.8943231441048,
+ "grad_norm": 0.01873331144452095,
+ "learning_rate": 0.0006,
+ "loss": 4.811744213104248,
+ "step": 3880
+ },
+ {
+ "epoch": 53.90829694323144,
+ "grad_norm": 0.02229069173336029,
+ "learning_rate": 0.0006,
+ "loss": 4.721514701843262,
+ "step": 3881
+ },
+ {
+ "epoch": 53.92227074235808,
+ "grad_norm": 0.028012916445732117,
+ "learning_rate": 0.0006,
+ "loss": 4.773603439331055,
+ "step": 3882
+ },
+ {
+ "epoch": 53.93624454148472,
+ "grad_norm": 0.036322783678770065,
+ "learning_rate": 0.0006,
+ "loss": 4.787134170532227,
+ "step": 3883
+ },
+ {
+ "epoch": 53.95021834061135,
+ "grad_norm": 0.048149142414331436,
+ "learning_rate": 0.0006,
+ "loss": 4.736196517944336,
+ "step": 3884
+ },
+ {
+ "epoch": 53.96419213973799,
+ "grad_norm": 0.051918916404247284,
+ "learning_rate": 0.0006,
+ "loss": 4.686728000640869,
+ "step": 3885
+ },
+ {
+ "epoch": 53.97816593886463,
+ "grad_norm": 0.0481342189013958,
+ "learning_rate": 0.0006,
+ "loss": 4.733730792999268,
+ "step": 3886
+ },
+ {
+ "epoch": 53.992139737991266,
+ "grad_norm": 0.045424818992614746,
+ "learning_rate": 0.0006,
+ "loss": 4.748723983764648,
+ "step": 3887
+ },
+ {
+ "epoch": 54.0,
+ "grad_norm": 0.04153316840529442,
+ "learning_rate": 0.0006,
+ "loss": 4.715220928192139,
+ "step": 3888
+ },
+ {
+ "epoch": 54.0,
+ "eval_loss": 4.964215278625488,
+ "eval_runtime": 56.7757,
+ "eval_samples_per_second": 43.011,
+ "eval_steps_per_second": 1.356,
+ "step": 3888
+ },
+ {
+ "epoch": 54.01397379912664,
+ "grad_norm": 0.03459750860929489,
+ "learning_rate": 0.0006,
+ "loss": 4.661557674407959,
+ "step": 3889
+ },
+ {
+ "epoch": 54.02794759825328,
+ "grad_norm": 0.025409208610653877,
+ "learning_rate": 0.0006,
+ "loss": 4.7065582275390625,
+ "step": 3890
+ },
+ {
+ "epoch": 54.041921397379916,
+ "grad_norm": 0.03202550485730171,
+ "learning_rate": 0.0006,
+ "loss": 4.751070976257324,
+ "step": 3891
+ },
+ {
+ "epoch": 54.05589519650655,
+ "grad_norm": 0.02968018501996994,
+ "learning_rate": 0.0006,
+ "loss": 4.635388374328613,
+ "step": 3892
+ },
+ {
+ "epoch": 54.069868995633186,
+ "grad_norm": 0.02757406234741211,
+ "learning_rate": 0.0006,
+ "loss": 4.706214427947998,
+ "step": 3893
+ },
+ {
+ "epoch": 54.083842794759825,
+ "grad_norm": 0.025223620235919952,
+ "learning_rate": 0.0006,
+ "loss": 4.703792572021484,
+ "step": 3894
+ },
+ {
+ "epoch": 54.09781659388646,
+ "grad_norm": 0.02431778982281685,
+ "learning_rate": 0.0006,
+ "loss": 4.658116340637207,
+ "step": 3895
+ },
+ {
+ "epoch": 54.1117903930131,
+ "grad_norm": 0.02792172320187092,
+ "learning_rate": 0.0006,
+ "loss": 4.638895034790039,
+ "step": 3896
+ },
+ {
+ "epoch": 54.12576419213974,
+ "grad_norm": 0.032253511250019073,
+ "learning_rate": 0.0006,
+ "loss": 4.693902492523193,
+ "step": 3897
+ },
+ {
+ "epoch": 54.13973799126637,
+ "grad_norm": 0.027044931426644325,
+ "learning_rate": 0.0006,
+ "loss": 4.575805187225342,
+ "step": 3898
+ },
+ {
+ "epoch": 54.15371179039301,
+ "grad_norm": 0.023270361125469208,
+ "learning_rate": 0.0006,
+ "loss": 4.67645263671875,
+ "step": 3899
+ },
+ {
+ "epoch": 54.16768558951965,
+ "grad_norm": 0.01929079182446003,
+ "learning_rate": 0.0006,
+ "loss": 4.68062686920166,
+ "step": 3900
+ },
+ {
+ "epoch": 54.18165938864629,
+ "grad_norm": 0.020343473181128502,
+ "learning_rate": 0.0006,
+ "loss": 4.708086967468262,
+ "step": 3901
+ },
+ {
+ "epoch": 54.19563318777293,
+ "grad_norm": 0.0211793202906847,
+ "learning_rate": 0.0006,
+ "loss": 4.677323341369629,
+ "step": 3902
+ },
+ {
+ "epoch": 54.209606986899566,
+ "grad_norm": 0.020341217517852783,
+ "learning_rate": 0.0006,
+ "loss": 4.760937690734863,
+ "step": 3903
+ },
+ {
+ "epoch": 54.223580786026204,
+ "grad_norm": 0.019421054050326347,
+ "learning_rate": 0.0006,
+ "loss": 4.508834362030029,
+ "step": 3904
+ },
+ {
+ "epoch": 54.237554585152836,
+ "grad_norm": 0.016701679676771164,
+ "learning_rate": 0.0006,
+ "loss": 4.664399147033691,
+ "step": 3905
+ },
+ {
+ "epoch": 54.251528384279474,
+ "grad_norm": 0.016561470925807953,
+ "learning_rate": 0.0006,
+ "loss": 4.742990493774414,
+ "step": 3906
+ },
+ {
+ "epoch": 54.26550218340611,
+ "grad_norm": 0.017446085810661316,
+ "learning_rate": 0.0006,
+ "loss": 4.63038444519043,
+ "step": 3907
+ },
+ {
+ "epoch": 54.27947598253275,
+ "grad_norm": 0.014340086840093136,
+ "learning_rate": 0.0006,
+ "loss": 4.688791275024414,
+ "step": 3908
+ },
+ {
+ "epoch": 54.29344978165939,
+ "grad_norm": 0.013580698519945145,
+ "learning_rate": 0.0006,
+ "loss": 4.625302314758301,
+ "step": 3909
+ },
+ {
+ "epoch": 54.30742358078603,
+ "grad_norm": 0.014931539073586464,
+ "learning_rate": 0.0006,
+ "loss": 4.6167497634887695,
+ "step": 3910
+ },
+ {
+ "epoch": 54.32139737991266,
+ "grad_norm": 0.013492150232195854,
+ "learning_rate": 0.0006,
+ "loss": 4.657521724700928,
+ "step": 3911
+ },
+ {
+ "epoch": 54.3353711790393,
+ "grad_norm": 0.015151413157582283,
+ "learning_rate": 0.0006,
+ "loss": 4.634521484375,
+ "step": 3912
+ },
+ {
+ "epoch": 54.34934497816594,
+ "grad_norm": 0.01302040833979845,
+ "learning_rate": 0.0006,
+ "loss": 4.718928337097168,
+ "step": 3913
+ },
+ {
+ "epoch": 54.36331877729258,
+ "grad_norm": 0.013433529995381832,
+ "learning_rate": 0.0006,
+ "loss": 4.584803581237793,
+ "step": 3914
+ },
+ {
+ "epoch": 54.377292576419215,
+ "grad_norm": 0.012987039051949978,
+ "learning_rate": 0.0006,
+ "loss": 4.6393585205078125,
+ "step": 3915
+ },
+ {
+ "epoch": 54.391266375545854,
+ "grad_norm": 0.013695075176656246,
+ "learning_rate": 0.0006,
+ "loss": 4.556313514709473,
+ "step": 3916
+ },
+ {
+ "epoch": 54.40524017467249,
+ "grad_norm": 0.013718255795538425,
+ "learning_rate": 0.0006,
+ "loss": 4.576223850250244,
+ "step": 3917
+ },
+ {
+ "epoch": 54.419213973799124,
+ "grad_norm": 0.013699112460017204,
+ "learning_rate": 0.0006,
+ "loss": 4.63889217376709,
+ "step": 3918
+ },
+ {
+ "epoch": 54.43318777292576,
+ "grad_norm": 0.013516264036297798,
+ "learning_rate": 0.0006,
+ "loss": 4.669532299041748,
+ "step": 3919
+ },
+ {
+ "epoch": 54.4471615720524,
+ "grad_norm": 0.013597294688224792,
+ "learning_rate": 0.0006,
+ "loss": 4.6644287109375,
+ "step": 3920
+ },
+ {
+ "epoch": 54.46113537117904,
+ "grad_norm": 0.013185849413275719,
+ "learning_rate": 0.0006,
+ "loss": 4.636900901794434,
+ "step": 3921
+ },
+ {
+ "epoch": 54.47510917030568,
+ "grad_norm": 0.014311990700662136,
+ "learning_rate": 0.0006,
+ "loss": 4.6047587394714355,
+ "step": 3922
+ },
+ {
+ "epoch": 54.48908296943232,
+ "grad_norm": 0.012426027096807957,
+ "learning_rate": 0.0006,
+ "loss": 4.6161394119262695,
+ "step": 3923
+ },
+ {
+ "epoch": 54.50305676855895,
+ "grad_norm": 0.01283493172377348,
+ "learning_rate": 0.0006,
+ "loss": 4.584190368652344,
+ "step": 3924
+ },
+ {
+ "epoch": 54.51703056768559,
+ "grad_norm": 0.01339943427592516,
+ "learning_rate": 0.0006,
+ "loss": 4.624184608459473,
+ "step": 3925
+ },
+ {
+ "epoch": 54.531004366812226,
+ "grad_norm": 0.012859724462032318,
+ "learning_rate": 0.0006,
+ "loss": 4.6174635887146,
+ "step": 3926
+ },
+ {
+ "epoch": 54.544978165938865,
+ "grad_norm": 0.014420023187994957,
+ "learning_rate": 0.0006,
+ "loss": 4.536544322967529,
+ "step": 3927
+ },
+ {
+ "epoch": 54.5589519650655,
+ "grad_norm": 0.014237185008823872,
+ "learning_rate": 0.0006,
+ "loss": 4.681514739990234,
+ "step": 3928
+ },
+ {
+ "epoch": 54.57292576419214,
+ "grad_norm": 0.013340328820049763,
+ "learning_rate": 0.0006,
+ "loss": 4.579689979553223,
+ "step": 3929
+ },
+ {
+ "epoch": 54.58689956331878,
+ "grad_norm": 0.013894530944526196,
+ "learning_rate": 0.0006,
+ "loss": 4.638958930969238,
+ "step": 3930
+ },
+ {
+ "epoch": 54.60087336244541,
+ "grad_norm": 0.016181141138076782,
+ "learning_rate": 0.0006,
+ "loss": 4.573508262634277,
+ "step": 3931
+ },
+ {
+ "epoch": 54.61484716157205,
+ "grad_norm": 0.013221992179751396,
+ "learning_rate": 0.0006,
+ "loss": 4.6229047775268555,
+ "step": 3932
+ },
+ {
+ "epoch": 54.62882096069869,
+ "grad_norm": 0.013241561129689217,
+ "learning_rate": 0.0006,
+ "loss": 4.641152858734131,
+ "step": 3933
+ },
+ {
+ "epoch": 54.64279475982533,
+ "grad_norm": 0.016237538307905197,
+ "learning_rate": 0.0006,
+ "loss": 4.6421217918396,
+ "step": 3934
+ },
+ {
+ "epoch": 54.65676855895197,
+ "grad_norm": 0.020866507664322853,
+ "learning_rate": 0.0006,
+ "loss": 4.646058082580566,
+ "step": 3935
+ },
+ {
+ "epoch": 54.670742358078606,
+ "grad_norm": 0.022182893007993698,
+ "learning_rate": 0.0006,
+ "loss": 4.554501533508301,
+ "step": 3936
+ },
+ {
+ "epoch": 54.68471615720524,
+ "grad_norm": 0.02254868485033512,
+ "learning_rate": 0.0006,
+ "loss": 4.563869953155518,
+ "step": 3937
+ },
+ {
+ "epoch": 54.698689956331876,
+ "grad_norm": 0.019730940461158752,
+ "learning_rate": 0.0006,
+ "loss": 4.592165946960449,
+ "step": 3938
+ },
+ {
+ "epoch": 54.712663755458514,
+ "grad_norm": 0.015146561898291111,
+ "learning_rate": 0.0006,
+ "loss": 4.6228742599487305,
+ "step": 3939
+ },
+ {
+ "epoch": 54.72663755458515,
+ "grad_norm": 0.01824716106057167,
+ "learning_rate": 0.0006,
+ "loss": 4.716236114501953,
+ "step": 3940
+ },
+ {
+ "epoch": 54.74061135371179,
+ "grad_norm": 0.017730550840497017,
+ "learning_rate": 0.0006,
+ "loss": 4.650232315063477,
+ "step": 3941
+ },
+ {
+ "epoch": 54.75458515283843,
+ "grad_norm": 0.014875595457851887,
+ "learning_rate": 0.0006,
+ "loss": 4.6377458572387695,
+ "step": 3942
+ },
+ {
+ "epoch": 54.76855895196506,
+ "grad_norm": 0.01613209955394268,
+ "learning_rate": 0.0006,
+ "loss": 4.5768537521362305,
+ "step": 3943
+ },
+ {
+ "epoch": 54.7825327510917,
+ "grad_norm": 0.017311280593276024,
+ "learning_rate": 0.0006,
+ "loss": 4.625744342803955,
+ "step": 3944
+ },
+ {
+ "epoch": 54.79650655021834,
+ "grad_norm": 0.018770085647702217,
+ "learning_rate": 0.0006,
+ "loss": 4.613787651062012,
+ "step": 3945
+ },
+ {
+ "epoch": 54.81048034934498,
+ "grad_norm": 0.0165257528424263,
+ "learning_rate": 0.0006,
+ "loss": 4.5991315841674805,
+ "step": 3946
+ },
+ {
+ "epoch": 54.82445414847162,
+ "grad_norm": 0.017433857545256615,
+ "learning_rate": 0.0006,
+ "loss": 4.513872146606445,
+ "step": 3947
+ },
+ {
+ "epoch": 54.838427947598255,
+ "grad_norm": 0.020126905292272568,
+ "learning_rate": 0.0006,
+ "loss": 4.616185188293457,
+ "step": 3948
+ },
+ {
+ "epoch": 54.852401746724894,
+ "grad_norm": 0.01819358579814434,
+ "learning_rate": 0.0006,
+ "loss": 4.535691261291504,
+ "step": 3949
+ },
+ {
+ "epoch": 54.866375545851525,
+ "grad_norm": 0.01899927295744419,
+ "learning_rate": 0.0006,
+ "loss": 4.595199108123779,
+ "step": 3950
+ },
+ {
+ "epoch": 54.880349344978164,
+ "grad_norm": 0.02005566656589508,
+ "learning_rate": 0.0006,
+ "loss": 4.633289337158203,
+ "step": 3951
+ },
+ {
+ "epoch": 54.8943231441048,
+ "grad_norm": 0.021570220589637756,
+ "learning_rate": 0.0006,
+ "loss": 4.592503547668457,
+ "step": 3952
+ },
+ {
+ "epoch": 54.90829694323144,
+ "grad_norm": 0.01883506216108799,
+ "learning_rate": 0.0006,
+ "loss": 4.6597795486450195,
+ "step": 3953
+ },
+ {
+ "epoch": 54.92227074235808,
+ "grad_norm": 0.018070021644234657,
+ "learning_rate": 0.0006,
+ "loss": 4.601474285125732,
+ "step": 3954
+ },
+ {
+ "epoch": 54.93624454148472,
+ "grad_norm": 0.018090378493070602,
+ "learning_rate": 0.0006,
+ "loss": 4.569846153259277,
+ "step": 3955
+ },
+ {
+ "epoch": 54.95021834061135,
+ "grad_norm": 0.018751563504338264,
+ "learning_rate": 0.0006,
+ "loss": 4.646064281463623,
+ "step": 3956
+ },
+ {
+ "epoch": 54.96419213973799,
+ "grad_norm": 0.017512254416942596,
+ "learning_rate": 0.0006,
+ "loss": 4.606670379638672,
+ "step": 3957
+ },
+ {
+ "epoch": 54.97816593886463,
+ "grad_norm": 0.018771110102534294,
+ "learning_rate": 0.0006,
+ "loss": 4.604315757751465,
+ "step": 3958
+ },
+ {
+ "epoch": 54.992139737991266,
+ "grad_norm": 0.01611141301691532,
+ "learning_rate": 0.0006,
+ "loss": 4.643245220184326,
+ "step": 3959
+ },
+ {
+ "epoch": 55.0,
+ "grad_norm": 0.0156728383153677,
+ "learning_rate": 0.0006,
+ "loss": 4.6739044189453125,
+ "step": 3960
+ },
+ {
+ "epoch": 55.0,
+ "eval_loss": 4.851032257080078,
+ "eval_runtime": 56.8699,
+ "eval_samples_per_second": 42.94,
+ "eval_steps_per_second": 1.354,
+ "step": 3960
+ },
+ {
+ "epoch": 55.01397379912664,
+ "grad_norm": 0.016214672476053238,
+ "learning_rate": 0.0006,
+ "loss": 4.571749687194824,
+ "step": 3961
+ },
+ {
+ "epoch": 55.02794759825328,
+ "grad_norm": 0.01649155467748642,
+ "learning_rate": 0.0006,
+ "loss": 4.553592205047607,
+ "step": 3962
+ },
+ {
+ "epoch": 55.041921397379916,
+ "grad_norm": 0.016793379560112953,
+ "learning_rate": 0.0006,
+ "loss": 4.650628566741943,
+ "step": 3963
+ },
+ {
+ "epoch": 55.05589519650655,
+ "grad_norm": 0.0191937405616045,
+ "learning_rate": 0.0006,
+ "loss": 4.620532989501953,
+ "step": 3964
+ },
+ {
+ "epoch": 55.069868995633186,
+ "grad_norm": 0.020976556465029716,
+ "learning_rate": 0.0006,
+ "loss": 4.6329545974731445,
+ "step": 3965
+ },
+ {
+ "epoch": 55.083842794759825,
+ "grad_norm": 0.023450477048754692,
+ "learning_rate": 0.0006,
+ "loss": 4.60828971862793,
+ "step": 3966
+ },
+ {
+ "epoch": 55.09781659388646,
+ "grad_norm": 0.026778312399983406,
+ "learning_rate": 0.0006,
+ "loss": 4.617481708526611,
+ "step": 3967
+ },
+ {
+ "epoch": 55.1117903930131,
+ "grad_norm": 0.02333568036556244,
+ "learning_rate": 0.0006,
+ "loss": 4.580722332000732,
+ "step": 3968
+ },
+ {
+ "epoch": 55.12576419213974,
+ "grad_norm": 0.018491310998797417,
+ "learning_rate": 0.0006,
+ "loss": 4.590953350067139,
+ "step": 3969
+ },
+ {
+ "epoch": 55.13973799126637,
+ "grad_norm": 0.020339803770184517,
+ "learning_rate": 0.0006,
+ "loss": 4.518374443054199,
+ "step": 3970
+ },
+ {
+ "epoch": 55.15371179039301,
+ "grad_norm": 0.01759706623852253,
+ "learning_rate": 0.0006,
+ "loss": 4.560522079467773,
+ "step": 3971
+ },
+ {
+ "epoch": 55.16768558951965,
+ "grad_norm": 0.015875980257987976,
+ "learning_rate": 0.0006,
+ "loss": 4.4795708656311035,
+ "step": 3972
+ },
+ {
+ "epoch": 55.18165938864629,
+ "grad_norm": 0.01650134287774563,
+ "learning_rate": 0.0006,
+ "loss": 4.57435941696167,
+ "step": 3973
+ },
+ {
+ "epoch": 55.19563318777293,
+ "grad_norm": 0.01734331250190735,
+ "learning_rate": 0.0006,
+ "loss": 4.661007881164551,
+ "step": 3974
+ },
+ {
+ "epoch": 55.209606986899566,
+ "grad_norm": 0.01651303470134735,
+ "learning_rate": 0.0006,
+ "loss": 4.511148452758789,
+ "step": 3975
+ },
+ {
+ "epoch": 55.223580786026204,
+ "grad_norm": 0.016189761459827423,
+ "learning_rate": 0.0006,
+ "loss": 4.592350006103516,
+ "step": 3976
+ },
+ {
+ "epoch": 55.237554585152836,
+ "grad_norm": 0.01833222061395645,
+ "learning_rate": 0.0006,
+ "loss": 4.506649971008301,
+ "step": 3977
+ },
+ {
+ "epoch": 55.251528384279474,
+ "grad_norm": 0.02433410845696926,
+ "learning_rate": 0.0006,
+ "loss": 4.49500036239624,
+ "step": 3978
+ },
+ {
+ "epoch": 55.26550218340611,
+ "grad_norm": 0.0314289927482605,
+ "learning_rate": 0.0006,
+ "loss": 4.507624626159668,
+ "step": 3979
+ },
+ {
+ "epoch": 55.27947598253275,
+ "grad_norm": 0.03013679012656212,
+ "learning_rate": 0.0006,
+ "loss": 4.533641815185547,
+ "step": 3980
+ },
+ {
+ "epoch": 55.29344978165939,
+ "grad_norm": 0.024524061009287834,
+ "learning_rate": 0.0006,
+ "loss": 4.5721116065979,
+ "step": 3981
+ },
+ {
+ "epoch": 55.30742358078603,
+ "grad_norm": 0.02640336938202381,
+ "learning_rate": 0.0006,
+ "loss": 4.652521133422852,
+ "step": 3982
+ },
+ {
+ "epoch": 55.32139737991266,
+ "grad_norm": 0.021309101954102516,
+ "learning_rate": 0.0006,
+ "loss": 4.600375175476074,
+ "step": 3983
+ },
+ {
+ "epoch": 55.3353711790393,
+ "grad_norm": 0.023008549585938454,
+ "learning_rate": 0.0006,
+ "loss": 4.53914737701416,
+ "step": 3984
+ },
+ {
+ "epoch": 55.34934497816594,
+ "grad_norm": 0.022491468116641045,
+ "learning_rate": 0.0006,
+ "loss": 4.661163330078125,
+ "step": 3985
+ },
+ {
+ "epoch": 55.36331877729258,
+ "grad_norm": 0.01904081553220749,
+ "learning_rate": 0.0006,
+ "loss": 4.5426177978515625,
+ "step": 3986
+ },
+ {
+ "epoch": 55.377292576419215,
+ "grad_norm": 0.019229518249630928,
+ "learning_rate": 0.0006,
+ "loss": 4.606066703796387,
+ "step": 3987
+ },
+ {
+ "epoch": 55.391266375545854,
+ "grad_norm": 0.01781383901834488,
+ "learning_rate": 0.0006,
+ "loss": 4.54642391204834,
+ "step": 3988
+ },
+ {
+ "epoch": 55.40524017467249,
+ "grad_norm": 0.0174600500613451,
+ "learning_rate": 0.0006,
+ "loss": 4.711891174316406,
+ "step": 3989
+ },
+ {
+ "epoch": 55.419213973799124,
+ "grad_norm": 0.018154339864850044,
+ "learning_rate": 0.0006,
+ "loss": 4.527122974395752,
+ "step": 3990
+ },
+ {
+ "epoch": 55.43318777292576,
+ "grad_norm": 0.01914648339152336,
+ "learning_rate": 0.0006,
+ "loss": 4.641975402832031,
+ "step": 3991
+ },
+ {
+ "epoch": 55.4471615720524,
+ "grad_norm": 0.017118625342845917,
+ "learning_rate": 0.0006,
+ "loss": 4.511425971984863,
+ "step": 3992
+ },
+ {
+ "epoch": 55.46113537117904,
+ "grad_norm": 0.01551489345729351,
+ "learning_rate": 0.0006,
+ "loss": 4.551141738891602,
+ "step": 3993
+ },
+ {
+ "epoch": 55.47510917030568,
+ "grad_norm": 0.015072065405547619,
+ "learning_rate": 0.0006,
+ "loss": 4.651115894317627,
+ "step": 3994
+ },
+ {
+ "epoch": 55.48908296943232,
+ "grad_norm": 0.017010986804962158,
+ "learning_rate": 0.0006,
+ "loss": 4.583282470703125,
+ "step": 3995
+ },
+ {
+ "epoch": 55.50305676855895,
+ "grad_norm": 0.01764831133186817,
+ "learning_rate": 0.0006,
+ "loss": 4.593364238739014,
+ "step": 3996
+ },
+ {
+ "epoch": 55.51703056768559,
+ "grad_norm": 0.01762884296476841,
+ "learning_rate": 0.0006,
+ "loss": 4.45671272277832,
+ "step": 3997
+ },
+ {
+ "epoch": 55.531004366812226,
+ "grad_norm": 0.01600457727909088,
+ "learning_rate": 0.0006,
+ "loss": 4.511260032653809,
+ "step": 3998
+ },
+ {
+ "epoch": 55.544978165938865,
+ "grad_norm": 0.015706021338701248,
+ "learning_rate": 0.0006,
+ "loss": 4.47593355178833,
+ "step": 3999
+ },
+ {
+ "epoch": 55.5589519650655,
+ "grad_norm": 0.017718996852636337,
+ "learning_rate": 0.0006,
+ "loss": 4.610543251037598,
+ "step": 4000
+ },
+ {
+ "epoch": 55.57292576419214,
+ "grad_norm": 0.015073757618665695,
+ "learning_rate": 0.0006,
+ "loss": 4.466620445251465,
+ "step": 4001
+ },
+ {
+ "epoch": 55.58689956331878,
+ "grad_norm": 0.01789247989654541,
+ "learning_rate": 0.0006,
+ "loss": 4.503409385681152,
+ "step": 4002
+ },
+ {
+ "epoch": 55.60087336244541,
+ "grad_norm": 0.022329598665237427,
+ "learning_rate": 0.0006,
+ "loss": 4.62099552154541,
+ "step": 4003
+ },
+ {
+ "epoch": 55.61484716157205,
+ "grad_norm": 0.02343829534947872,
+ "learning_rate": 0.0006,
+ "loss": 4.620813369750977,
+ "step": 4004
+ },
+ {
+ "epoch": 55.62882096069869,
+ "grad_norm": 0.023279687389731407,
+ "learning_rate": 0.0006,
+ "loss": 4.545765399932861,
+ "step": 4005
+ },
+ {
+ "epoch": 55.64279475982533,
+ "grad_norm": 0.020817464217543602,
+ "learning_rate": 0.0006,
+ "loss": 4.561470031738281,
+ "step": 4006
+ },
+ {
+ "epoch": 55.65676855895197,
+ "grad_norm": 0.020270058885216713,
+ "learning_rate": 0.0006,
+ "loss": 4.5875020027160645,
+ "step": 4007
+ },
+ {
+ "epoch": 55.670742358078606,
+ "grad_norm": 0.021085843443870544,
+ "learning_rate": 0.0006,
+ "loss": 4.504560470581055,
+ "step": 4008
+ },
+ {
+ "epoch": 55.68471615720524,
+ "grad_norm": 0.018507730215787888,
+ "learning_rate": 0.0006,
+ "loss": 4.520644664764404,
+ "step": 4009
+ },
+ {
+ "epoch": 55.698689956331876,
+ "grad_norm": 0.020571516826748848,
+ "learning_rate": 0.0006,
+ "loss": 4.539643287658691,
+ "step": 4010
+ },
+ {
+ "epoch": 55.712663755458514,
+ "grad_norm": 0.021484510973095894,
+ "learning_rate": 0.0006,
+ "loss": 4.628833293914795,
+ "step": 4011
+ },
+ {
+ "epoch": 55.72663755458515,
+ "grad_norm": 0.017693819478154182,
+ "learning_rate": 0.0006,
+ "loss": 4.6023783683776855,
+ "step": 4012
+ },
+ {
+ "epoch": 55.74061135371179,
+ "grad_norm": 0.020186059176921844,
+ "learning_rate": 0.0006,
+ "loss": 4.634884834289551,
+ "step": 4013
+ },
+ {
+ "epoch": 55.75458515283843,
+ "grad_norm": 0.01835780404508114,
+ "learning_rate": 0.0006,
+ "loss": 4.577791213989258,
+ "step": 4014
+ },
+ {
+ "epoch": 55.76855895196506,
+ "grad_norm": 0.018162988126277924,
+ "learning_rate": 0.0006,
+ "loss": 4.641597270965576,
+ "step": 4015
+ },
+ {
+ "epoch": 55.7825327510917,
+ "grad_norm": 0.018909504637122154,
+ "learning_rate": 0.0006,
+ "loss": 4.487447738647461,
+ "step": 4016
+ },
+ {
+ "epoch": 55.79650655021834,
+ "grad_norm": 0.018972251564264297,
+ "learning_rate": 0.0006,
+ "loss": 4.590317726135254,
+ "step": 4017
+ },
+ {
+ "epoch": 55.81048034934498,
+ "grad_norm": 0.01793607883155346,
+ "learning_rate": 0.0006,
+ "loss": 4.4994611740112305,
+ "step": 4018
+ },
+ {
+ "epoch": 55.82445414847162,
+ "grad_norm": 0.02123514749109745,
+ "learning_rate": 0.0006,
+ "loss": 4.479006767272949,
+ "step": 4019
+ },
+ {
+ "epoch": 55.838427947598255,
+ "grad_norm": 0.021246599033474922,
+ "learning_rate": 0.0006,
+ "loss": 4.4738640785217285,
+ "step": 4020
+ },
+ {
+ "epoch": 55.852401746724894,
+ "grad_norm": 0.020131602883338928,
+ "learning_rate": 0.0006,
+ "loss": 4.600308418273926,
+ "step": 4021
+ },
+ {
+ "epoch": 55.866375545851525,
+ "grad_norm": 0.019159460440278053,
+ "learning_rate": 0.0006,
+ "loss": 4.540908336639404,
+ "step": 4022
+ },
+ {
+ "epoch": 55.880349344978164,
+ "grad_norm": 0.023114705458283424,
+ "learning_rate": 0.0006,
+ "loss": 4.641247749328613,
+ "step": 4023
+ },
+ {
+ "epoch": 55.8943231441048,
+ "grad_norm": 0.02296135015785694,
+ "learning_rate": 0.0006,
+ "loss": 4.444624423980713,
+ "step": 4024
+ },
+ {
+ "epoch": 55.90829694323144,
+ "grad_norm": 0.02204655110836029,
+ "learning_rate": 0.0006,
+ "loss": 4.506349563598633,
+ "step": 4025
+ },
+ {
+ "epoch": 55.92227074235808,
+ "grad_norm": 0.019338225945830345,
+ "learning_rate": 0.0006,
+ "loss": 4.616047382354736,
+ "step": 4026
+ },
+ {
+ "epoch": 55.93624454148472,
+ "grad_norm": 0.020357884466648102,
+ "learning_rate": 0.0006,
+ "loss": 4.48170280456543,
+ "step": 4027
+ },
+ {
+ "epoch": 55.95021834061135,
+ "grad_norm": 0.02426772564649582,
+ "learning_rate": 0.0006,
+ "loss": 4.576877117156982,
+ "step": 4028
+ },
+ {
+ "epoch": 55.96419213973799,
+ "grad_norm": 0.023351816460490227,
+ "learning_rate": 0.0006,
+ "loss": 4.458885192871094,
+ "step": 4029
+ },
+ {
+ "epoch": 55.97816593886463,
+ "grad_norm": 0.019892243668437004,
+ "learning_rate": 0.0006,
+ "loss": 4.520229816436768,
+ "step": 4030
+ },
+ {
+ "epoch": 55.992139737991266,
+ "grad_norm": 0.018331089988350868,
+ "learning_rate": 0.0006,
+ "loss": 4.432201862335205,
+ "step": 4031
+ },
+ {
+ "epoch": 56.0,
+ "grad_norm": 0.021300828084349632,
+ "learning_rate": 0.0006,
+ "loss": 4.617072582244873,
+ "step": 4032
+ },
+ {
+ "epoch": 56.0,
+ "eval_loss": 4.811229705810547,
+ "eval_runtime": 56.4846,
+ "eval_samples_per_second": 43.233,
+ "eval_steps_per_second": 1.363,
+ "step": 4032
+ },
+ {
+ "epoch": 56.01397379912664,
+ "grad_norm": 0.02199401892721653,
+ "learning_rate": 0.0006,
+ "loss": 4.513965606689453,
+ "step": 4033
+ },
+ {
+ "epoch": 56.02794759825328,
+ "grad_norm": 0.022109858691692352,
+ "learning_rate": 0.0006,
+ "loss": 4.557568550109863,
+ "step": 4034
+ },
+ {
+ "epoch": 56.041921397379916,
+ "grad_norm": 0.020103694871068,
+ "learning_rate": 0.0006,
+ "loss": 4.569512844085693,
+ "step": 4035
+ },
+ {
+ "epoch": 56.05589519650655,
+ "grad_norm": 0.01898941583931446,
+ "learning_rate": 0.0006,
+ "loss": 4.623138904571533,
+ "step": 4036
+ },
+ {
+ "epoch": 56.069868995633186,
+ "grad_norm": 0.02227010577917099,
+ "learning_rate": 0.0006,
+ "loss": 4.433316230773926,
+ "step": 4037
+ },
+ {
+ "epoch": 56.083842794759825,
+ "grad_norm": 0.024683045223355293,
+ "learning_rate": 0.0006,
+ "loss": 4.458296775817871,
+ "step": 4038
+ },
+ {
+ "epoch": 56.09781659388646,
+ "grad_norm": 0.02900184690952301,
+ "learning_rate": 0.0006,
+ "loss": 4.531060218811035,
+ "step": 4039
+ },
+ {
+ "epoch": 56.1117903930131,
+ "grad_norm": 0.025557566434144974,
+ "learning_rate": 0.0006,
+ "loss": 4.56685733795166,
+ "step": 4040
+ },
+ {
+ "epoch": 56.12576419213974,
+ "grad_norm": 0.021414656192064285,
+ "learning_rate": 0.0006,
+ "loss": 4.469037055969238,
+ "step": 4041
+ },
+ {
+ "epoch": 56.13973799126637,
+ "grad_norm": 0.023262616246938705,
+ "learning_rate": 0.0006,
+ "loss": 4.493876934051514,
+ "step": 4042
+ },
+ {
+ "epoch": 56.15371179039301,
+ "grad_norm": 0.023011326789855957,
+ "learning_rate": 0.0006,
+ "loss": 4.4572906494140625,
+ "step": 4043
+ },
+ {
+ "epoch": 56.16768558951965,
+ "grad_norm": 0.02177615463733673,
+ "learning_rate": 0.0006,
+ "loss": 4.615389823913574,
+ "step": 4044
+ },
+ {
+ "epoch": 56.18165938864629,
+ "grad_norm": 0.021527713164687157,
+ "learning_rate": 0.0006,
+ "loss": 4.613762855529785,
+ "step": 4045
+ },
+ {
+ "epoch": 56.19563318777293,
+ "grad_norm": 0.01982325315475464,
+ "learning_rate": 0.0006,
+ "loss": 4.53670597076416,
+ "step": 4046
+ },
+ {
+ "epoch": 56.209606986899566,
+ "grad_norm": 0.022484654560685158,
+ "learning_rate": 0.0006,
+ "loss": 4.605565071105957,
+ "step": 4047
+ },
+ {
+ "epoch": 56.223580786026204,
+ "grad_norm": 0.02346140146255493,
+ "learning_rate": 0.0006,
+ "loss": 4.6265950202941895,
+ "step": 4048
+ },
+ {
+ "epoch": 56.237554585152836,
+ "grad_norm": 0.01876714453101158,
+ "learning_rate": 0.0006,
+ "loss": 4.481525421142578,
+ "step": 4049
+ },
+ {
+ "epoch": 56.251528384279474,
+ "grad_norm": 0.01926851086318493,
+ "learning_rate": 0.0006,
+ "loss": 4.630258083343506,
+ "step": 4050
+ },
+ {
+ "epoch": 56.26550218340611,
+ "grad_norm": 0.020858589559793472,
+ "learning_rate": 0.0006,
+ "loss": 4.5265913009643555,
+ "step": 4051
+ },
+ {
+ "epoch": 56.27947598253275,
+ "grad_norm": 0.01889280416071415,
+ "learning_rate": 0.0006,
+ "loss": 4.581275463104248,
+ "step": 4052
+ },
+ {
+ "epoch": 56.29344978165939,
+ "grad_norm": 0.01801002025604248,
+ "learning_rate": 0.0006,
+ "loss": 4.498557090759277,
+ "step": 4053
+ },
+ {
+ "epoch": 56.30742358078603,
+ "grad_norm": 0.02252180315554142,
+ "learning_rate": 0.0006,
+ "loss": 4.565267086029053,
+ "step": 4054
+ },
+ {
+ "epoch": 56.32139737991266,
+ "grad_norm": 0.02306070365011692,
+ "learning_rate": 0.0006,
+ "loss": 4.3591389656066895,
+ "step": 4055
+ },
+ {
+ "epoch": 56.3353711790393,
+ "grad_norm": 0.020292513072490692,
+ "learning_rate": 0.0006,
+ "loss": 4.407517433166504,
+ "step": 4056
+ },
+ {
+ "epoch": 56.34934497816594,
+ "grad_norm": 0.019154123961925507,
+ "learning_rate": 0.0006,
+ "loss": 4.482261657714844,
+ "step": 4057
+ },
+ {
+ "epoch": 56.36331877729258,
+ "grad_norm": 0.017338937148451805,
+ "learning_rate": 0.0006,
+ "loss": 4.4463582038879395,
+ "step": 4058
+ },
+ {
+ "epoch": 56.377292576419215,
+ "grad_norm": 0.01730559952557087,
+ "learning_rate": 0.0006,
+ "loss": 4.542474746704102,
+ "step": 4059
+ },
+ {
+ "epoch": 56.391266375545854,
+ "grad_norm": 0.017029426991939545,
+ "learning_rate": 0.0006,
+ "loss": 4.552488327026367,
+ "step": 4060
+ },
+ {
+ "epoch": 56.40524017467249,
+ "grad_norm": 0.01745498552918434,
+ "learning_rate": 0.0006,
+ "loss": 4.52928352355957,
+ "step": 4061
+ },
+ {
+ "epoch": 56.419213973799124,
+ "grad_norm": 0.0168515145778656,
+ "learning_rate": 0.0006,
+ "loss": 4.4516801834106445,
+ "step": 4062
+ },
+ {
+ "epoch": 56.43318777292576,
+ "grad_norm": 0.018545417115092278,
+ "learning_rate": 0.0006,
+ "loss": 4.569811820983887,
+ "step": 4063
+ },
+ {
+ "epoch": 56.4471615720524,
+ "grad_norm": 0.02227838523685932,
+ "learning_rate": 0.0006,
+ "loss": 4.412224769592285,
+ "step": 4064
+ },
+ {
+ "epoch": 56.46113537117904,
+ "grad_norm": 0.023338504135608673,
+ "learning_rate": 0.0006,
+ "loss": 4.490184307098389,
+ "step": 4065
+ },
+ {
+ "epoch": 56.47510917030568,
+ "grad_norm": 0.02257535606622696,
+ "learning_rate": 0.0006,
+ "loss": 4.479152679443359,
+ "step": 4066
+ },
+ {
+ "epoch": 56.48908296943232,
+ "grad_norm": 0.022737598046660423,
+ "learning_rate": 0.0006,
+ "loss": 4.418235778808594,
+ "step": 4067
+ },
+ {
+ "epoch": 56.50305676855895,
+ "grad_norm": 0.024603573605418205,
+ "learning_rate": 0.0006,
+ "loss": 4.45781135559082,
+ "step": 4068
+ },
+ {
+ "epoch": 56.51703056768559,
+ "grad_norm": 0.024953778833150864,
+ "learning_rate": 0.0006,
+ "loss": 4.593907833099365,
+ "step": 4069
+ },
+ {
+ "epoch": 56.531004366812226,
+ "grad_norm": 0.02195645309984684,
+ "learning_rate": 0.0006,
+ "loss": 4.467643737792969,
+ "step": 4070
+ },
+ {
+ "epoch": 56.544978165938865,
+ "grad_norm": 0.017126578837633133,
+ "learning_rate": 0.0006,
+ "loss": 4.555484771728516,
+ "step": 4071
+ },
+ {
+ "epoch": 56.5589519650655,
+ "grad_norm": 0.019482605159282684,
+ "learning_rate": 0.0006,
+ "loss": 4.430915832519531,
+ "step": 4072
+ },
+ {
+ "epoch": 56.57292576419214,
+ "grad_norm": 0.0219392292201519,
+ "learning_rate": 0.0006,
+ "loss": 4.593365669250488,
+ "step": 4073
+ },
+ {
+ "epoch": 56.58689956331878,
+ "grad_norm": 0.02420145645737648,
+ "learning_rate": 0.0006,
+ "loss": 4.491467475891113,
+ "step": 4074
+ },
+ {
+ "epoch": 56.60087336244541,
+ "grad_norm": 0.021523723378777504,
+ "learning_rate": 0.0006,
+ "loss": 4.463634490966797,
+ "step": 4075
+ },
+ {
+ "epoch": 56.61484716157205,
+ "grad_norm": 0.020242048427462578,
+ "learning_rate": 0.0006,
+ "loss": 4.577968597412109,
+ "step": 4076
+ },
+ {
+ "epoch": 56.62882096069869,
+ "grad_norm": 0.017771173268556595,
+ "learning_rate": 0.0006,
+ "loss": 4.568269729614258,
+ "step": 4077
+ },
+ {
+ "epoch": 56.64279475982533,
+ "grad_norm": 0.01598544418811798,
+ "learning_rate": 0.0006,
+ "loss": 4.476841926574707,
+ "step": 4078
+ },
+ {
+ "epoch": 56.65676855895197,
+ "grad_norm": 0.0168446097522974,
+ "learning_rate": 0.0006,
+ "loss": 4.480965614318848,
+ "step": 4079
+ },
+ {
+ "epoch": 56.670742358078606,
+ "grad_norm": 0.014948660507798195,
+ "learning_rate": 0.0006,
+ "loss": 4.535400390625,
+ "step": 4080
+ },
+ {
+ "epoch": 56.68471615720524,
+ "grad_norm": 0.014899153262376785,
+ "learning_rate": 0.0006,
+ "loss": 4.476048946380615,
+ "step": 4081
+ },
+ {
+ "epoch": 56.698689956331876,
+ "grad_norm": 0.014051870442926884,
+ "learning_rate": 0.0006,
+ "loss": 4.533884048461914,
+ "step": 4082
+ },
+ {
+ "epoch": 56.712663755458514,
+ "grad_norm": 0.014708572067320347,
+ "learning_rate": 0.0006,
+ "loss": 4.586690902709961,
+ "step": 4083
+ },
+ {
+ "epoch": 56.72663755458515,
+ "grad_norm": 0.014797299169003963,
+ "learning_rate": 0.0006,
+ "loss": 4.446291446685791,
+ "step": 4084
+ },
+ {
+ "epoch": 56.74061135371179,
+ "grad_norm": 0.015256541781127453,
+ "learning_rate": 0.0006,
+ "loss": 4.553572654724121,
+ "step": 4085
+ },
+ {
+ "epoch": 56.75458515283843,
+ "grad_norm": 0.016042491421103477,
+ "learning_rate": 0.0006,
+ "loss": 4.507801055908203,
+ "step": 4086
+ },
+ {
+ "epoch": 56.76855895196506,
+ "grad_norm": 0.018835173919796944,
+ "learning_rate": 0.0006,
+ "loss": 4.393911838531494,
+ "step": 4087
+ },
+ {
+ "epoch": 56.7825327510917,
+ "grad_norm": 0.021994033828377724,
+ "learning_rate": 0.0006,
+ "loss": 4.543368339538574,
+ "step": 4088
+ },
+ {
+ "epoch": 56.79650655021834,
+ "grad_norm": 0.0230990182608366,
+ "learning_rate": 0.0006,
+ "loss": 4.507420539855957,
+ "step": 4089
+ },
+ {
+ "epoch": 56.81048034934498,
+ "grad_norm": 0.018818842247128487,
+ "learning_rate": 0.0006,
+ "loss": 4.499941349029541,
+ "step": 4090
+ },
+ {
+ "epoch": 56.82445414847162,
+ "grad_norm": 0.018257278949022293,
+ "learning_rate": 0.0006,
+ "loss": 4.380057334899902,
+ "step": 4091
+ },
+ {
+ "epoch": 56.838427947598255,
+ "grad_norm": 0.020073018968105316,
+ "learning_rate": 0.0006,
+ "loss": 4.572811603546143,
+ "step": 4092
+ },
+ {
+ "epoch": 56.852401746724894,
+ "grad_norm": 0.02078658528625965,
+ "learning_rate": 0.0006,
+ "loss": 4.446835517883301,
+ "step": 4093
+ },
+ {
+ "epoch": 56.866375545851525,
+ "grad_norm": 0.01793118566274643,
+ "learning_rate": 0.0006,
+ "loss": 4.517854690551758,
+ "step": 4094
+ },
+ {
+ "epoch": 56.880349344978164,
+ "grad_norm": 0.01608281210064888,
+ "learning_rate": 0.0006,
+ "loss": 4.506482124328613,
+ "step": 4095
+ },
+ {
+ "epoch": 56.8943231441048,
+ "grad_norm": 0.017405368387699127,
+ "learning_rate": 0.0006,
+ "loss": 4.439718246459961,
+ "step": 4096
+ },
+ {
+ "epoch": 56.90829694323144,
+ "grad_norm": 0.017930863425135612,
+ "learning_rate": 0.0006,
+ "loss": 4.409365653991699,
+ "step": 4097
+ },
+ {
+ "epoch": 56.92227074235808,
+ "grad_norm": 0.015757670626044273,
+ "learning_rate": 0.0006,
+ "loss": 4.505547523498535,
+ "step": 4098
+ },
+ {
+ "epoch": 56.93624454148472,
+ "grad_norm": 0.01580345816910267,
+ "learning_rate": 0.0006,
+ "loss": 4.462774753570557,
+ "step": 4099
+ },
+ {
+ "epoch": 56.95021834061135,
+ "grad_norm": 0.01659180410206318,
+ "learning_rate": 0.0006,
+ "loss": 4.49345588684082,
+ "step": 4100
+ },
+ {
+ "epoch": 56.96419213973799,
+ "grad_norm": 0.01877259835600853,
+ "learning_rate": 0.0006,
+ "loss": 4.574443340301514,
+ "step": 4101
+ },
+ {
+ "epoch": 56.97816593886463,
+ "grad_norm": 0.01977839693427086,
+ "learning_rate": 0.0006,
+ "loss": 4.574034690856934,
+ "step": 4102
+ },
+ {
+ "epoch": 56.992139737991266,
+ "grad_norm": 0.020627597346901894,
+ "learning_rate": 0.0006,
+ "loss": 4.50969123840332,
+ "step": 4103
+ },
+ {
+ "epoch": 57.0,
+ "grad_norm": 0.025128807872533798,
+ "learning_rate": 0.0006,
+ "loss": 4.401662826538086,
+ "step": 4104
+ },
+ {
+ "epoch": 57.0,
+ "eval_loss": 4.779252052307129,
+ "eval_runtime": 56.8104,
+ "eval_samples_per_second": 42.985,
+ "eval_steps_per_second": 1.355,
+ "step": 4104
+ },
+ {
+ "epoch": 57.01397379912664,
+ "grad_norm": 0.027712570503354073,
+ "learning_rate": 0.0006,
+ "loss": 4.472955703735352,
+ "step": 4105
+ },
+ {
+ "epoch": 57.02794759825328,
+ "grad_norm": 0.02164197526872158,
+ "learning_rate": 0.0006,
+ "loss": 4.387596130371094,
+ "step": 4106
+ },
+ {
+ "epoch": 57.041921397379916,
+ "grad_norm": 0.017972752451896667,
+ "learning_rate": 0.0006,
+ "loss": 4.4326276779174805,
+ "step": 4107
+ },
+ {
+ "epoch": 57.05589519650655,
+ "grad_norm": 0.020811019465327263,
+ "learning_rate": 0.0006,
+ "loss": 4.452951431274414,
+ "step": 4108
+ },
+ {
+ "epoch": 57.069868995633186,
+ "grad_norm": 0.019913259893655777,
+ "learning_rate": 0.0006,
+ "loss": 4.448301315307617,
+ "step": 4109
+ },
+ {
+ "epoch": 57.083842794759825,
+ "grad_norm": 0.017409102991223335,
+ "learning_rate": 0.0006,
+ "loss": 4.4108357429504395,
+ "step": 4110
+ },
+ {
+ "epoch": 57.09781659388646,
+ "grad_norm": 0.019374269992113113,
+ "learning_rate": 0.0006,
+ "loss": 4.398412227630615,
+ "step": 4111
+ },
+ {
+ "epoch": 57.1117903930131,
+ "grad_norm": 0.015521776862442493,
+ "learning_rate": 0.0006,
+ "loss": 4.508594989776611,
+ "step": 4112
+ },
+ {
+ "epoch": 57.12576419213974,
+ "grad_norm": 0.017201920971274376,
+ "learning_rate": 0.0006,
+ "loss": 4.449558734893799,
+ "step": 4113
+ },
+ {
+ "epoch": 57.13973799126637,
+ "grad_norm": 0.018062541261315346,
+ "learning_rate": 0.0006,
+ "loss": 4.517874240875244,
+ "step": 4114
+ },
+ {
+ "epoch": 57.15371179039301,
+ "grad_norm": 0.015846073627471924,
+ "learning_rate": 0.0006,
+ "loss": 4.427501678466797,
+ "step": 4115
+ },
+ {
+ "epoch": 57.16768558951965,
+ "grad_norm": 0.01849362626671791,
+ "learning_rate": 0.0006,
+ "loss": 4.433181285858154,
+ "step": 4116
+ },
+ {
+ "epoch": 57.18165938864629,
+ "grad_norm": 0.01948804222047329,
+ "learning_rate": 0.0006,
+ "loss": 4.446122646331787,
+ "step": 4117
+ },
+ {
+ "epoch": 57.19563318777293,
+ "grad_norm": 0.017354389652609825,
+ "learning_rate": 0.0006,
+ "loss": 4.431153774261475,
+ "step": 4118
+ },
+ {
+ "epoch": 57.209606986899566,
+ "grad_norm": 0.015817373991012573,
+ "learning_rate": 0.0006,
+ "loss": 4.448566913604736,
+ "step": 4119
+ },
+ {
+ "epoch": 57.223580786026204,
+ "grad_norm": 0.01635785959661007,
+ "learning_rate": 0.0006,
+ "loss": 4.408031463623047,
+ "step": 4120
+ },
+ {
+ "epoch": 57.237554585152836,
+ "grad_norm": 0.014528338797390461,
+ "learning_rate": 0.0006,
+ "loss": 4.541668891906738,
+ "step": 4121
+ },
+ {
+ "epoch": 57.251528384279474,
+ "grad_norm": 0.01589174196124077,
+ "learning_rate": 0.0006,
+ "loss": 4.513199329376221,
+ "step": 4122
+ },
+ {
+ "epoch": 57.26550218340611,
+ "grad_norm": 0.016093695536255836,
+ "learning_rate": 0.0006,
+ "loss": 4.516178131103516,
+ "step": 4123
+ },
+ {
+ "epoch": 57.27947598253275,
+ "grad_norm": 0.01788361556828022,
+ "learning_rate": 0.0006,
+ "loss": 4.524880409240723,
+ "step": 4124
+ },
+ {
+ "epoch": 57.29344978165939,
+ "grad_norm": 0.01591576263308525,
+ "learning_rate": 0.0006,
+ "loss": 4.44764518737793,
+ "step": 4125
+ },
+ {
+ "epoch": 57.30742358078603,
+ "grad_norm": 0.015452570281922817,
+ "learning_rate": 0.0006,
+ "loss": 4.46844482421875,
+ "step": 4126
+ },
+ {
+ "epoch": 57.32139737991266,
+ "grad_norm": 0.016404492780566216,
+ "learning_rate": 0.0006,
+ "loss": 4.443678855895996,
+ "step": 4127
+ },
+ {
+ "epoch": 57.3353711790393,
+ "grad_norm": 0.019967148080468178,
+ "learning_rate": 0.0006,
+ "loss": 4.430230140686035,
+ "step": 4128
+ },
+ {
+ "epoch": 57.34934497816594,
+ "grad_norm": 0.02594015561044216,
+ "learning_rate": 0.0006,
+ "loss": 4.47061824798584,
+ "step": 4129
+ },
+ {
+ "epoch": 57.36331877729258,
+ "grad_norm": 0.031096890568733215,
+ "learning_rate": 0.0006,
+ "loss": 4.462275981903076,
+ "step": 4130
+ },
+ {
+ "epoch": 57.377292576419215,
+ "grad_norm": 0.02834898792207241,
+ "learning_rate": 0.0006,
+ "loss": 4.41143798828125,
+ "step": 4131
+ },
+ {
+ "epoch": 57.391266375545854,
+ "grad_norm": 0.023621153086423874,
+ "learning_rate": 0.0006,
+ "loss": 4.562548637390137,
+ "step": 4132
+ },
+ {
+ "epoch": 57.40524017467249,
+ "grad_norm": 0.02277296595275402,
+ "learning_rate": 0.0006,
+ "loss": 4.456247329711914,
+ "step": 4133
+ },
+ {
+ "epoch": 57.419213973799124,
+ "grad_norm": 0.019868768751621246,
+ "learning_rate": 0.0006,
+ "loss": 4.524300575256348,
+ "step": 4134
+ },
+ {
+ "epoch": 57.43318777292576,
+ "grad_norm": 0.01887078955769539,
+ "learning_rate": 0.0006,
+ "loss": 4.540097236633301,
+ "step": 4135
+ },
+ {
+ "epoch": 57.4471615720524,
+ "grad_norm": 0.024268802255392075,
+ "learning_rate": 0.0006,
+ "loss": 4.397233963012695,
+ "step": 4136
+ },
+ {
+ "epoch": 57.46113537117904,
+ "grad_norm": 0.025739721953868866,
+ "learning_rate": 0.0006,
+ "loss": 4.354194641113281,
+ "step": 4137
+ },
+ {
+ "epoch": 57.47510917030568,
+ "grad_norm": 0.02442290261387825,
+ "learning_rate": 0.0006,
+ "loss": 4.303822994232178,
+ "step": 4138
+ },
+ {
+ "epoch": 57.48908296943232,
+ "grad_norm": 0.022215209901332855,
+ "learning_rate": 0.0006,
+ "loss": 4.4870452880859375,
+ "step": 4139
+ },
+ {
+ "epoch": 57.50305676855895,
+ "grad_norm": 0.01789158768951893,
+ "learning_rate": 0.0006,
+ "loss": 4.343809127807617,
+ "step": 4140
+ },
+ {
+ "epoch": 57.51703056768559,
+ "grad_norm": 0.020381739363074303,
+ "learning_rate": 0.0006,
+ "loss": 4.53312873840332,
+ "step": 4141
+ },
+ {
+ "epoch": 57.531004366812226,
+ "grad_norm": 0.019585080444812775,
+ "learning_rate": 0.0006,
+ "loss": 4.510162353515625,
+ "step": 4142
+ },
+ {
+ "epoch": 57.544978165938865,
+ "grad_norm": 0.01985127478837967,
+ "learning_rate": 0.0006,
+ "loss": 4.3654022216796875,
+ "step": 4143
+ },
+ {
+ "epoch": 57.5589519650655,
+ "grad_norm": 0.01962905190885067,
+ "learning_rate": 0.0006,
+ "loss": 4.560830116271973,
+ "step": 4144
+ },
+ {
+ "epoch": 57.57292576419214,
+ "grad_norm": 0.02679629437625408,
+ "learning_rate": 0.0006,
+ "loss": 4.300804138183594,
+ "step": 4145
+ },
+ {
+ "epoch": 57.58689956331878,
+ "grad_norm": 0.039830856025218964,
+ "learning_rate": 0.0006,
+ "loss": 4.54150915145874,
+ "step": 4146
+ },
+ {
+ "epoch": 57.60087336244541,
+ "grad_norm": 0.049527477473020554,
+ "learning_rate": 0.0006,
+ "loss": 4.405792236328125,
+ "step": 4147
+ },
+ {
+ "epoch": 57.61484716157205,
+ "grad_norm": 0.042966946959495544,
+ "learning_rate": 0.0006,
+ "loss": 4.500643730163574,
+ "step": 4148
+ },
+ {
+ "epoch": 57.62882096069869,
+ "grad_norm": 0.08999177813529968,
+ "learning_rate": 0.0006,
+ "loss": 4.504790782928467,
+ "step": 4149
+ },
+ {
+ "epoch": 57.64279475982533,
+ "grad_norm": 0.44693124294281006,
+ "learning_rate": 0.0006,
+ "loss": 4.845344066619873,
+ "step": 4150
+ },
+ {
+ "epoch": 57.65676855895197,
+ "grad_norm": 0.7143144011497498,
+ "learning_rate": 0.0006,
+ "loss": 6.907845497131348,
+ "step": 4151
+ },
+ {
+ "epoch": 57.670742358078606,
+ "grad_norm": 0.25239595770835876,
+ "learning_rate": 0.0006,
+ "loss": 7.082579612731934,
+ "step": 4152
+ },
+ {
+ "epoch": 57.68471615720524,
+ "grad_norm": 0.26550790667533875,
+ "learning_rate": 0.0006,
+ "loss": 7.50180196762085,
+ "step": 4153
+ },
+ {
+ "epoch": 57.698689956331876,
+ "grad_norm": 0.14735378324985504,
+ "learning_rate": 0.0006,
+ "loss": 7.214590072631836,
+ "step": 4154
+ },
+ {
+ "epoch": 57.712663755458514,
+ "grad_norm": 0.17685334384441376,
+ "learning_rate": 0.0006,
+ "loss": 6.861320495605469,
+ "step": 4155
+ },
+ {
+ "epoch": 57.72663755458515,
+ "grad_norm": 0.18231339752674103,
+ "learning_rate": 0.0006,
+ "loss": 6.791839599609375,
+ "step": 4156
+ },
+ {
+ "epoch": 57.74061135371179,
+ "grad_norm": 0.09230359643697739,
+ "learning_rate": 0.0006,
+ "loss": 6.760223388671875,
+ "step": 4157
+ },
+ {
+ "epoch": 57.75458515283843,
+ "grad_norm": 0.06918249279260635,
+ "learning_rate": 0.0006,
+ "loss": 6.532999515533447,
+ "step": 4158
+ },
+ {
+ "epoch": 57.76855895196506,
+ "grad_norm": 0.0901007279753685,
+ "learning_rate": 0.0006,
+ "loss": 6.52907133102417,
+ "step": 4159
+ },
+ {
+ "epoch": 57.7825327510917,
+ "grad_norm": 0.05678229779005051,
+ "learning_rate": 0.0006,
+ "loss": 6.295290470123291,
+ "step": 4160
+ },
+ {
+ "epoch": 57.79650655021834,
+ "grad_norm": 0.04450210556387901,
+ "learning_rate": 0.0006,
+ "loss": 6.14900541305542,
+ "step": 4161
+ },
+ {
+ "epoch": 57.81048034934498,
+ "grad_norm": 0.049191512167453766,
+ "learning_rate": 0.0006,
+ "loss": 6.19650411605835,
+ "step": 4162
+ },
+ {
+ "epoch": 57.82445414847162,
+ "grad_norm": 0.03908967226743698,
+ "learning_rate": 0.0006,
+ "loss": 6.054948806762695,
+ "step": 4163
+ },
+ {
+ "epoch": 57.838427947598255,
+ "grad_norm": 0.04023387283086777,
+ "learning_rate": 0.0006,
+ "loss": 6.034873962402344,
+ "step": 4164
+ },
+ {
+ "epoch": 57.852401746724894,
+ "grad_norm": 0.03304268419742584,
+ "learning_rate": 0.0006,
+ "loss": 5.769852161407471,
+ "step": 4165
+ },
+ {
+ "epoch": 57.866375545851525,
+ "grad_norm": 0.03432450816035271,
+ "learning_rate": 0.0006,
+ "loss": 5.788785457611084,
+ "step": 4166
+ },
+ {
+ "epoch": 57.880349344978164,
+ "grad_norm": 0.029965840280056,
+ "learning_rate": 0.0006,
+ "loss": 5.815243721008301,
+ "step": 4167
+ },
+ {
+ "epoch": 57.8943231441048,
+ "grad_norm": 0.03637698292732239,
+ "learning_rate": 0.0006,
+ "loss": 5.7204060554504395,
+ "step": 4168
+ },
+ {
+ "epoch": 57.90829694323144,
+ "grad_norm": 0.03477517142891884,
+ "learning_rate": 0.0006,
+ "loss": 5.605930328369141,
+ "step": 4169
+ },
+ {
+ "epoch": 57.92227074235808,
+ "grad_norm": 0.05061859264969826,
+ "learning_rate": 0.0006,
+ "loss": 5.547054767608643,
+ "step": 4170
+ },
+ {
+ "epoch": 57.93624454148472,
+ "grad_norm": 0.06299655884504318,
+ "learning_rate": 0.0006,
+ "loss": 5.552166938781738,
+ "step": 4171
+ },
+ {
+ "epoch": 57.95021834061135,
+ "grad_norm": 0.04215948283672333,
+ "learning_rate": 0.0006,
+ "loss": 5.375457763671875,
+ "step": 4172
+ },
+ {
+ "epoch": 57.96419213973799,
+ "grad_norm": 0.0348566472530365,
+ "learning_rate": 0.0006,
+ "loss": 5.418606281280518,
+ "step": 4173
+ },
+ {
+ "epoch": 57.97816593886463,
+ "grad_norm": 0.03807530924677849,
+ "learning_rate": 0.0006,
+ "loss": 5.403156280517578,
+ "step": 4174
+ },
+ {
+ "epoch": 57.992139737991266,
+ "grad_norm": 0.02989993803203106,
+ "learning_rate": 0.0006,
+ "loss": 5.355951309204102,
+ "step": 4175
+ },
+ {
+ "epoch": 58.0,
+ "grad_norm": 0.034300558269023895,
+ "learning_rate": 0.0006,
+ "loss": 5.44920539855957,
+ "step": 4176
+ },
+ {
+ "epoch": 58.0,
+ "eval_loss": 5.416731834411621,
+ "eval_runtime": 56.7934,
+ "eval_samples_per_second": 42.998,
+ "eval_steps_per_second": 1.356,
+ "step": 4176
+ },
+ {
+ "epoch": 58.01397379912664,
+ "grad_norm": 0.033961448818445206,
+ "learning_rate": 0.0006,
+ "loss": 5.250707626342773,
+ "step": 4177
+ },
+ {
+ "epoch": 58.02794759825328,
+ "grad_norm": 0.027343502268195152,
+ "learning_rate": 0.0006,
+ "loss": 5.149238109588623,
+ "step": 4178
+ },
+ {
+ "epoch": 58.041921397379916,
+ "grad_norm": 0.03033650480210781,
+ "learning_rate": 0.0006,
+ "loss": 5.072463512420654,
+ "step": 4179
+ },
+ {
+ "epoch": 58.05589519650655,
+ "grad_norm": 0.028262868523597717,
+ "learning_rate": 0.0006,
+ "loss": 5.109417915344238,
+ "step": 4180
+ },
+ {
+ "epoch": 58.069868995633186,
+ "grad_norm": 0.02373034693300724,
+ "learning_rate": 0.0006,
+ "loss": 5.082464218139648,
+ "step": 4181
+ },
+ {
+ "epoch": 58.083842794759825,
+ "grad_norm": 0.025879688560962677,
+ "learning_rate": 0.0006,
+ "loss": 5.056792259216309,
+ "step": 4182
+ },
+ {
+ "epoch": 58.09781659388646,
+ "grad_norm": 0.026252275332808495,
+ "learning_rate": 0.0006,
+ "loss": 5.064239978790283,
+ "step": 4183
+ },
+ {
+ "epoch": 58.1117903930131,
+ "grad_norm": 0.028177792206406593,
+ "learning_rate": 0.0006,
+ "loss": 4.922310829162598,
+ "step": 4184
+ },
+ {
+ "epoch": 58.12576419213974,
+ "grad_norm": 0.02630820870399475,
+ "learning_rate": 0.0006,
+ "loss": 4.992884635925293,
+ "step": 4185
+ },
+ {
+ "epoch": 58.13973799126637,
+ "grad_norm": 0.024890903383493423,
+ "learning_rate": 0.0006,
+ "loss": 4.910036563873291,
+ "step": 4186
+ },
+ {
+ "epoch": 58.15371179039301,
+ "grad_norm": 0.02822992391884327,
+ "learning_rate": 0.0006,
+ "loss": 4.866640090942383,
+ "step": 4187
+ },
+ {
+ "epoch": 58.16768558951965,
+ "grad_norm": 0.03253607079386711,
+ "learning_rate": 0.0006,
+ "loss": 4.9112958908081055,
+ "step": 4188
+ },
+ {
+ "epoch": 58.18165938864629,
+ "grad_norm": 0.03765944764018059,
+ "learning_rate": 0.0006,
+ "loss": 4.774468421936035,
+ "step": 4189
+ },
+ {
+ "epoch": 58.19563318777293,
+ "grad_norm": 0.039874881505966187,
+ "learning_rate": 0.0006,
+ "loss": 4.831678867340088,
+ "step": 4190
+ },
+ {
+ "epoch": 58.209606986899566,
+ "grad_norm": 0.034595925360918045,
+ "learning_rate": 0.0006,
+ "loss": 4.8243794441223145,
+ "step": 4191
+ },
+ {
+ "epoch": 58.223580786026204,
+ "grad_norm": 0.04840189591050148,
+ "learning_rate": 0.0006,
+ "loss": 4.854156494140625,
+ "step": 4192
+ },
+ {
+ "epoch": 58.237554585152836,
+ "grad_norm": 0.06358753889799118,
+ "learning_rate": 0.0006,
+ "loss": 4.836430549621582,
+ "step": 4193
+ },
+ {
+ "epoch": 58.251528384279474,
+ "grad_norm": 0.0507555715739727,
+ "learning_rate": 0.0006,
+ "loss": 4.88357400894165,
+ "step": 4194
+ },
+ {
+ "epoch": 58.26550218340611,
+ "grad_norm": 0.04384690150618553,
+ "learning_rate": 0.0006,
+ "loss": 4.774049282073975,
+ "step": 4195
+ },
+ {
+ "epoch": 58.27947598253275,
+ "grad_norm": 0.03246928006410599,
+ "learning_rate": 0.0006,
+ "loss": 4.686605930328369,
+ "step": 4196
+ },
+ {
+ "epoch": 58.29344978165939,
+ "grad_norm": 0.028878789395093918,
+ "learning_rate": 0.0006,
+ "loss": 4.796070575714111,
+ "step": 4197
+ },
+ {
+ "epoch": 58.30742358078603,
+ "grad_norm": 0.028175361454486847,
+ "learning_rate": 0.0006,
+ "loss": 4.728490829467773,
+ "step": 4198
+ },
+ {
+ "epoch": 58.32139737991266,
+ "grad_norm": 0.029428014531731606,
+ "learning_rate": 0.0006,
+ "loss": 4.7159624099731445,
+ "step": 4199
+ },
+ {
+ "epoch": 58.3353711790393,
+ "grad_norm": 0.031227827072143555,
+ "learning_rate": 0.0006,
+ "loss": 4.702625274658203,
+ "step": 4200
+ },
+ {
+ "epoch": 58.34934497816594,
+ "grad_norm": 0.028843428939580917,
+ "learning_rate": 0.0006,
+ "loss": 4.811799049377441,
+ "step": 4201
+ },
+ {
+ "epoch": 58.36331877729258,
+ "grad_norm": 0.026780391111969948,
+ "learning_rate": 0.0006,
+ "loss": 4.81662654876709,
+ "step": 4202
+ },
+ {
+ "epoch": 58.377292576419215,
+ "grad_norm": 0.019699757918715477,
+ "learning_rate": 0.0006,
+ "loss": 4.713685989379883,
+ "step": 4203
+ },
+ {
+ "epoch": 58.391266375545854,
+ "grad_norm": 0.024389250203967094,
+ "learning_rate": 0.0006,
+ "loss": 4.72354793548584,
+ "step": 4204
+ },
+ {
+ "epoch": 58.40524017467249,
+ "grad_norm": 0.02218274027109146,
+ "learning_rate": 0.0006,
+ "loss": 4.674125671386719,
+ "step": 4205
+ },
+ {
+ "epoch": 58.419213973799124,
+ "grad_norm": 0.02041488140821457,
+ "learning_rate": 0.0006,
+ "loss": 4.7909698486328125,
+ "step": 4206
+ },
+ {
+ "epoch": 58.43318777292576,
+ "grad_norm": 0.018561935052275658,
+ "learning_rate": 0.0006,
+ "loss": 4.808770179748535,
+ "step": 4207
+ },
+ {
+ "epoch": 58.4471615720524,
+ "grad_norm": 0.01708846725523472,
+ "learning_rate": 0.0006,
+ "loss": 4.701626777648926,
+ "step": 4208
+ },
+ {
+ "epoch": 58.46113537117904,
+ "grad_norm": 0.0222734697163105,
+ "learning_rate": 0.0006,
+ "loss": 4.505410194396973,
+ "step": 4209
+ },
+ {
+ "epoch": 58.47510917030568,
+ "grad_norm": 0.03197057917714119,
+ "learning_rate": 0.0006,
+ "loss": 4.610360145568848,
+ "step": 4210
+ },
+ {
+ "epoch": 58.48908296943232,
+ "grad_norm": 0.06854520738124847,
+ "learning_rate": 0.0006,
+ "loss": 4.625649452209473,
+ "step": 4211
+ },
+ {
+ "epoch": 58.50305676855895,
+ "grad_norm": 0.09981521219015121,
+ "learning_rate": 0.0006,
+ "loss": 4.832150459289551,
+ "step": 4212
+ },
+ {
+ "epoch": 58.51703056768559,
+ "grad_norm": 0.046356040984392166,
+ "learning_rate": 0.0006,
+ "loss": 4.5641632080078125,
+ "step": 4213
+ },
+ {
+ "epoch": 58.531004366812226,
+ "grad_norm": 0.03882667422294617,
+ "learning_rate": 0.0006,
+ "loss": 4.666974067687988,
+ "step": 4214
+ },
+ {
+ "epoch": 58.544978165938865,
+ "grad_norm": 0.033203233033418655,
+ "learning_rate": 0.0006,
+ "loss": 4.629413604736328,
+ "step": 4215
+ },
+ {
+ "epoch": 58.5589519650655,
+ "grad_norm": 0.02806735597550869,
+ "learning_rate": 0.0006,
+ "loss": 4.763245582580566,
+ "step": 4216
+ },
+ {
+ "epoch": 58.57292576419214,
+ "grad_norm": 0.026728278025984764,
+ "learning_rate": 0.0006,
+ "loss": 4.629597187042236,
+ "step": 4217
+ },
+ {
+ "epoch": 58.58689956331878,
+ "grad_norm": 0.024380959570407867,
+ "learning_rate": 0.0006,
+ "loss": 4.633614540100098,
+ "step": 4218
+ },
+ {
+ "epoch": 58.60087336244541,
+ "grad_norm": 0.022312544286251068,
+ "learning_rate": 0.0006,
+ "loss": 4.671379566192627,
+ "step": 4219
+ },
+ {
+ "epoch": 58.61484716157205,
+ "grad_norm": 0.022049568593502045,
+ "learning_rate": 0.0006,
+ "loss": 4.661557197570801,
+ "step": 4220
+ },
+ {
+ "epoch": 58.62882096069869,
+ "grad_norm": 0.02119818702340126,
+ "learning_rate": 0.0006,
+ "loss": 4.688562393188477,
+ "step": 4221
+ },
+ {
+ "epoch": 58.64279475982533,
+ "grad_norm": 0.019769301638007164,
+ "learning_rate": 0.0006,
+ "loss": 4.607748031616211,
+ "step": 4222
+ },
+ {
+ "epoch": 58.65676855895197,
+ "grad_norm": 0.02124079130589962,
+ "learning_rate": 0.0006,
+ "loss": 4.570640563964844,
+ "step": 4223
+ },
+ {
+ "epoch": 58.670742358078606,
+ "grad_norm": 0.016344040632247925,
+ "learning_rate": 0.0006,
+ "loss": 4.676113128662109,
+ "step": 4224
+ },
+ {
+ "epoch": 58.68471615720524,
+ "grad_norm": 0.01735256239771843,
+ "learning_rate": 0.0006,
+ "loss": 4.62088680267334,
+ "step": 4225
+ },
+ {
+ "epoch": 58.698689956331876,
+ "grad_norm": 0.019323458895087242,
+ "learning_rate": 0.0006,
+ "loss": 4.579405307769775,
+ "step": 4226
+ },
+ {
+ "epoch": 58.712663755458514,
+ "grad_norm": 0.01760544814169407,
+ "learning_rate": 0.0006,
+ "loss": 4.625407695770264,
+ "step": 4227
+ },
+ {
+ "epoch": 58.72663755458515,
+ "grad_norm": 0.015593250282108784,
+ "learning_rate": 0.0006,
+ "loss": 4.6326117515563965,
+ "step": 4228
+ },
+ {
+ "epoch": 58.74061135371179,
+ "grad_norm": 0.014357523061335087,
+ "learning_rate": 0.0006,
+ "loss": 4.6869401931762695,
+ "step": 4229
+ },
+ {
+ "epoch": 58.75458515283843,
+ "grad_norm": 0.01604357920587063,
+ "learning_rate": 0.0006,
+ "loss": 4.690784931182861,
+ "step": 4230
+ },
+ {
+ "epoch": 58.76855895196506,
+ "grad_norm": 0.01673845760524273,
+ "learning_rate": 0.0006,
+ "loss": 4.513410568237305,
+ "step": 4231
+ },
+ {
+ "epoch": 58.7825327510917,
+ "grad_norm": 0.013657779432833195,
+ "learning_rate": 0.0006,
+ "loss": 4.637057781219482,
+ "step": 4232
+ },
+ {
+ "epoch": 58.79650655021834,
+ "grad_norm": 0.013373114168643951,
+ "learning_rate": 0.0006,
+ "loss": 4.719967842102051,
+ "step": 4233
+ },
+ {
+ "epoch": 58.81048034934498,
+ "grad_norm": 0.016345568001270294,
+ "learning_rate": 0.0006,
+ "loss": 4.604109764099121,
+ "step": 4234
+ },
+ {
+ "epoch": 58.82445414847162,
+ "grad_norm": 0.01573033444583416,
+ "learning_rate": 0.0006,
+ "loss": 4.6471428871154785,
+ "step": 4235
+ },
+ {
+ "epoch": 58.838427947598255,
+ "grad_norm": 0.01517449039965868,
+ "learning_rate": 0.0006,
+ "loss": 4.722684860229492,
+ "step": 4236
+ },
+ {
+ "epoch": 58.852401746724894,
+ "grad_norm": 0.0145102022215724,
+ "learning_rate": 0.0006,
+ "loss": 4.662420272827148,
+ "step": 4237
+ },
+ {
+ "epoch": 58.866375545851525,
+ "grad_norm": 0.01457217987626791,
+ "learning_rate": 0.0006,
+ "loss": 4.585169792175293,
+ "step": 4238
+ },
+ {
+ "epoch": 58.880349344978164,
+ "grad_norm": 0.01406773366034031,
+ "learning_rate": 0.0006,
+ "loss": 4.5413665771484375,
+ "step": 4239
+ },
+ {
+ "epoch": 58.8943231441048,
+ "grad_norm": 0.012653871439397335,
+ "learning_rate": 0.0006,
+ "loss": 4.553636074066162,
+ "step": 4240
+ },
+ {
+ "epoch": 58.90829694323144,
+ "grad_norm": 0.01261200476437807,
+ "learning_rate": 0.0006,
+ "loss": 4.534193992614746,
+ "step": 4241
+ },
+ {
+ "epoch": 58.92227074235808,
+ "grad_norm": 0.012955864891409874,
+ "learning_rate": 0.0006,
+ "loss": 4.697248458862305,
+ "step": 4242
+ },
+ {
+ "epoch": 58.93624454148472,
+ "grad_norm": 0.012134749442338943,
+ "learning_rate": 0.0006,
+ "loss": 4.557095527648926,
+ "step": 4243
+ },
+ {
+ "epoch": 58.95021834061135,
+ "grad_norm": 0.012164677493274212,
+ "learning_rate": 0.0006,
+ "loss": 4.623579025268555,
+ "step": 4244
+ },
+ {
+ "epoch": 58.96419213973799,
+ "grad_norm": 0.013802947476506233,
+ "learning_rate": 0.0006,
+ "loss": 4.562838077545166,
+ "step": 4245
+ },
+ {
+ "epoch": 58.97816593886463,
+ "grad_norm": 0.011420476250350475,
+ "learning_rate": 0.0006,
+ "loss": 4.62265682220459,
+ "step": 4246
+ },
+ {
+ "epoch": 58.992139737991266,
+ "grad_norm": 0.01169576682150364,
+ "learning_rate": 0.0006,
+ "loss": 4.727449417114258,
+ "step": 4247
+ },
+ {
+ "epoch": 59.0,
+ "grad_norm": 0.013577710837125778,
+ "learning_rate": 0.0006,
+ "loss": 4.536325454711914,
+ "step": 4248
+ },
+ {
+ "epoch": 59.0,
+ "eval_loss": 4.86527681350708,
+ "eval_runtime": 60.4463,
+ "eval_samples_per_second": 40.399,
+ "eval_steps_per_second": 1.274,
+ "step": 4248
+ },
+ {
+ "epoch": 59.01397379912664,
+ "grad_norm": 0.015251655131578445,
+ "learning_rate": 0.0006,
+ "loss": 4.634598731994629,
+ "step": 4249
+ },
+ {
+ "epoch": 59.02794759825328,
+ "grad_norm": 0.017414938658475876,
+ "learning_rate": 0.0006,
+ "loss": 4.521329879760742,
+ "step": 4250
+ },
+ {
+ "epoch": 59.041921397379916,
+ "grad_norm": 0.01656588353216648,
+ "learning_rate": 0.0006,
+ "loss": 4.446192264556885,
+ "step": 4251
+ },
+ {
+ "epoch": 59.05589519650655,
+ "grad_norm": 0.012038925662636757,
+ "learning_rate": 0.0006,
+ "loss": 4.523855209350586,
+ "step": 4252
+ },
+ {
+ "epoch": 59.069868995633186,
+ "grad_norm": 0.014403033070266247,
+ "learning_rate": 0.0006,
+ "loss": 4.561079978942871,
+ "step": 4253
+ },
+ {
+ "epoch": 59.083842794759825,
+ "grad_norm": 0.016243597492575645,
+ "learning_rate": 0.0006,
+ "loss": 4.673602104187012,
+ "step": 4254
+ },
+ {
+ "epoch": 59.09781659388646,
+ "grad_norm": 0.016492877155542374,
+ "learning_rate": 0.0006,
+ "loss": 4.564824104309082,
+ "step": 4255
+ },
+ {
+ "epoch": 59.1117903930131,
+ "grad_norm": 0.015980370342731476,
+ "learning_rate": 0.0006,
+ "loss": 4.339249134063721,
+ "step": 4256
+ },
+ {
+ "epoch": 59.12576419213974,
+ "grad_norm": 0.019687356427311897,
+ "learning_rate": 0.0006,
+ "loss": 4.481834888458252,
+ "step": 4257
+ },
+ {
+ "epoch": 59.13973799126637,
+ "grad_norm": 0.02983798086643219,
+ "learning_rate": 0.0006,
+ "loss": 4.4495697021484375,
+ "step": 4258
+ },
+ {
+ "epoch": 59.15371179039301,
+ "grad_norm": 0.04600981995463371,
+ "learning_rate": 0.0006,
+ "loss": 4.567838668823242,
+ "step": 4259
+ },
+ {
+ "epoch": 59.16768558951965,
+ "grad_norm": 0.04852001741528511,
+ "learning_rate": 0.0006,
+ "loss": 4.594701766967773,
+ "step": 4260
+ },
+ {
+ "epoch": 59.18165938864629,
+ "grad_norm": 0.03429504856467247,
+ "learning_rate": 0.0006,
+ "loss": 4.573735237121582,
+ "step": 4261
+ },
+ {
+ "epoch": 59.19563318777293,
+ "grad_norm": 0.029345160350203514,
+ "learning_rate": 0.0006,
+ "loss": 4.488258361816406,
+ "step": 4262
+ },
+ {
+ "epoch": 59.209606986899566,
+ "grad_norm": 0.01961623504757881,
+ "learning_rate": 0.0006,
+ "loss": 4.545222282409668,
+ "step": 4263
+ },
+ {
+ "epoch": 59.223580786026204,
+ "grad_norm": 0.020406130701303482,
+ "learning_rate": 0.0006,
+ "loss": 4.471898078918457,
+ "step": 4264
+ },
+ {
+ "epoch": 59.237554585152836,
+ "grad_norm": 0.020500272512435913,
+ "learning_rate": 0.0006,
+ "loss": 4.538337707519531,
+ "step": 4265
+ },
+ {
+ "epoch": 59.251528384279474,
+ "grad_norm": 0.020727017894387245,
+ "learning_rate": 0.0006,
+ "loss": 4.437284469604492,
+ "step": 4266
+ },
+ {
+ "epoch": 59.26550218340611,
+ "grad_norm": 0.03395998477935791,
+ "learning_rate": 0.0006,
+ "loss": 4.459124565124512,
+ "step": 4267
+ },
+ {
+ "epoch": 59.27947598253275,
+ "grad_norm": 0.04635605961084366,
+ "learning_rate": 0.0006,
+ "loss": 4.520228385925293,
+ "step": 4268
+ },
+ {
+ "epoch": 59.29344978165939,
+ "grad_norm": 0.04253006726503372,
+ "learning_rate": 0.0006,
+ "loss": 4.578673362731934,
+ "step": 4269
+ },
+ {
+ "epoch": 59.30742358078603,
+ "grad_norm": 0.02403509058058262,
+ "learning_rate": 0.0006,
+ "loss": 4.484639644622803,
+ "step": 4270
+ },
+ {
+ "epoch": 59.32139737991266,
+ "grad_norm": 0.02254190482199192,
+ "learning_rate": 0.0006,
+ "loss": 4.511314868927002,
+ "step": 4271
+ },
+ {
+ "epoch": 59.3353711790393,
+ "grad_norm": 0.021585190668702126,
+ "learning_rate": 0.0006,
+ "loss": 4.530239105224609,
+ "step": 4272
+ },
+ {
+ "epoch": 59.34934497816594,
+ "grad_norm": 0.021210316568613052,
+ "learning_rate": 0.0006,
+ "loss": 4.593514442443848,
+ "step": 4273
+ },
+ {
+ "epoch": 59.36331877729258,
+ "grad_norm": 0.019317107275128365,
+ "learning_rate": 0.0006,
+ "loss": 4.5591535568237305,
+ "step": 4274
+ },
+ {
+ "epoch": 59.377292576419215,
+ "grad_norm": 0.018059708178043365,
+ "learning_rate": 0.0006,
+ "loss": 4.4499592781066895,
+ "step": 4275
+ },
+ {
+ "epoch": 59.391266375545854,
+ "grad_norm": 0.017847446724772453,
+ "learning_rate": 0.0006,
+ "loss": 4.473999500274658,
+ "step": 4276
+ },
+ {
+ "epoch": 59.40524017467249,
+ "grad_norm": 0.014944756403565407,
+ "learning_rate": 0.0006,
+ "loss": 4.5934529304504395,
+ "step": 4277
+ },
+ {
+ "epoch": 59.419213973799124,
+ "grad_norm": 0.01663832925260067,
+ "learning_rate": 0.0006,
+ "loss": 4.504262924194336,
+ "step": 4278
+ },
+ {
+ "epoch": 59.43318777292576,
+ "grad_norm": 0.015571796335279942,
+ "learning_rate": 0.0006,
+ "loss": 4.512310981750488,
+ "step": 4279
+ },
+ {
+ "epoch": 59.4471615720524,
+ "grad_norm": 0.015896400436758995,
+ "learning_rate": 0.0006,
+ "loss": 4.452714920043945,
+ "step": 4280
+ },
+ {
+ "epoch": 59.46113537117904,
+ "grad_norm": 0.016010284423828125,
+ "learning_rate": 0.0006,
+ "loss": 4.6188063621521,
+ "step": 4281
+ },
+ {
+ "epoch": 59.47510917030568,
+ "grad_norm": 0.015075593255460262,
+ "learning_rate": 0.0006,
+ "loss": 4.477978229522705,
+ "step": 4282
+ },
+ {
+ "epoch": 59.48908296943232,
+ "grad_norm": 0.014698950573801994,
+ "learning_rate": 0.0006,
+ "loss": 4.544955730438232,
+ "step": 4283
+ },
+ {
+ "epoch": 59.50305676855895,
+ "grad_norm": 0.0136245833709836,
+ "learning_rate": 0.0006,
+ "loss": 4.530460357666016,
+ "step": 4284
+ },
+ {
+ "epoch": 59.51703056768559,
+ "grad_norm": 0.013118310831487179,
+ "learning_rate": 0.0006,
+ "loss": 4.614217758178711,
+ "step": 4285
+ },
+ {
+ "epoch": 59.531004366812226,
+ "grad_norm": 0.01295961532741785,
+ "learning_rate": 0.0006,
+ "loss": 4.43156099319458,
+ "step": 4286
+ },
+ {
+ "epoch": 59.544978165938865,
+ "grad_norm": 0.012429811991751194,
+ "learning_rate": 0.0006,
+ "loss": 4.517333030700684,
+ "step": 4287
+ },
+ {
+ "epoch": 59.5589519650655,
+ "grad_norm": 0.01261830423027277,
+ "learning_rate": 0.0006,
+ "loss": 4.569780349731445,
+ "step": 4288
+ },
+ {
+ "epoch": 59.57292576419214,
+ "grad_norm": 0.012394499965012074,
+ "learning_rate": 0.0006,
+ "loss": 4.5970869064331055,
+ "step": 4289
+ },
+ {
+ "epoch": 59.58689956331878,
+ "grad_norm": 0.01290759164839983,
+ "learning_rate": 0.0006,
+ "loss": 4.437100410461426,
+ "step": 4290
+ },
+ {
+ "epoch": 59.60087336244541,
+ "grad_norm": 0.012042169459164143,
+ "learning_rate": 0.0006,
+ "loss": 4.567404747009277,
+ "step": 4291
+ },
+ {
+ "epoch": 59.61484716157205,
+ "grad_norm": 0.012338520959019661,
+ "learning_rate": 0.0006,
+ "loss": 4.476193428039551,
+ "step": 4292
+ },
+ {
+ "epoch": 59.62882096069869,
+ "grad_norm": 0.01321258582174778,
+ "learning_rate": 0.0006,
+ "loss": 4.566680431365967,
+ "step": 4293
+ },
+ {
+ "epoch": 59.64279475982533,
+ "grad_norm": 0.013478122651576996,
+ "learning_rate": 0.0006,
+ "loss": 4.487398147583008,
+ "step": 4294
+ },
+ {
+ "epoch": 59.65676855895197,
+ "grad_norm": 0.01537603884935379,
+ "learning_rate": 0.0006,
+ "loss": 4.449337005615234,
+ "step": 4295
+ },
+ {
+ "epoch": 59.670742358078606,
+ "grad_norm": 0.017140565440058708,
+ "learning_rate": 0.0006,
+ "loss": 4.5174360275268555,
+ "step": 4296
+ },
+ {
+ "epoch": 59.68471615720524,
+ "grad_norm": 0.017931461334228516,
+ "learning_rate": 0.0006,
+ "loss": 4.522104740142822,
+ "step": 4297
+ },
+ {
+ "epoch": 59.698689956331876,
+ "grad_norm": 0.018037918955087662,
+ "learning_rate": 0.0006,
+ "loss": 4.496928691864014,
+ "step": 4298
+ },
+ {
+ "epoch": 59.712663755458514,
+ "grad_norm": 0.01684504561126232,
+ "learning_rate": 0.0006,
+ "loss": 4.5332417488098145,
+ "step": 4299
+ },
+ {
+ "epoch": 59.72663755458515,
+ "grad_norm": 0.016268383711576462,
+ "learning_rate": 0.0006,
+ "loss": 4.468893051147461,
+ "step": 4300
+ },
+ {
+ "epoch": 59.74061135371179,
+ "grad_norm": 0.014859385788440704,
+ "learning_rate": 0.0006,
+ "loss": 4.551092147827148,
+ "step": 4301
+ },
+ {
+ "epoch": 59.75458515283843,
+ "grad_norm": 0.013589047826826572,
+ "learning_rate": 0.0006,
+ "loss": 4.470579147338867,
+ "step": 4302
+ },
+ {
+ "epoch": 59.76855895196506,
+ "grad_norm": 0.015704551711678505,
+ "learning_rate": 0.0006,
+ "loss": 4.330068111419678,
+ "step": 4303
+ },
+ {
+ "epoch": 59.7825327510917,
+ "grad_norm": 0.017869921401143074,
+ "learning_rate": 0.0006,
+ "loss": 4.428255081176758,
+ "step": 4304
+ },
+ {
+ "epoch": 59.79650655021834,
+ "grad_norm": 0.019828537479043007,
+ "learning_rate": 0.0006,
+ "loss": 4.496023178100586,
+ "step": 4305
+ },
+ {
+ "epoch": 59.81048034934498,
+ "grad_norm": 0.015856236219406128,
+ "learning_rate": 0.0006,
+ "loss": 4.427152156829834,
+ "step": 4306
+ },
+ {
+ "epoch": 59.82445414847162,
+ "grad_norm": 0.013349304907023907,
+ "learning_rate": 0.0006,
+ "loss": 4.417333602905273,
+ "step": 4307
+ },
+ {
+ "epoch": 59.838427947598255,
+ "grad_norm": 0.014111637137830257,
+ "learning_rate": 0.0006,
+ "loss": 4.4507575035095215,
+ "step": 4308
+ },
+ {
+ "epoch": 59.852401746724894,
+ "grad_norm": 0.015186597593128681,
+ "learning_rate": 0.0006,
+ "loss": 4.375116348266602,
+ "step": 4309
+ },
+ {
+ "epoch": 59.866375545851525,
+ "grad_norm": 0.01604730449616909,
+ "learning_rate": 0.0006,
+ "loss": 4.604518890380859,
+ "step": 4310
+ },
+ {
+ "epoch": 59.880349344978164,
+ "grad_norm": 0.015308464877307415,
+ "learning_rate": 0.0006,
+ "loss": 4.507232189178467,
+ "step": 4311
+ },
+ {
+ "epoch": 59.8943231441048,
+ "grad_norm": 0.014017206616699696,
+ "learning_rate": 0.0006,
+ "loss": 4.398487091064453,
+ "step": 4312
+ },
+ {
+ "epoch": 59.90829694323144,
+ "grad_norm": 0.012911394238471985,
+ "learning_rate": 0.0006,
+ "loss": 4.567872047424316,
+ "step": 4313
+ },
+ {
+ "epoch": 59.92227074235808,
+ "grad_norm": 0.012775142677128315,
+ "learning_rate": 0.0006,
+ "loss": 4.562105655670166,
+ "step": 4314
+ },
+ {
+ "epoch": 59.93624454148472,
+ "grad_norm": 0.014750408008694649,
+ "learning_rate": 0.0006,
+ "loss": 4.456263542175293,
+ "step": 4315
+ },
+ {
+ "epoch": 59.95021834061135,
+ "grad_norm": 0.016726011410355568,
+ "learning_rate": 0.0006,
+ "loss": 4.360722064971924,
+ "step": 4316
+ },
+ {
+ "epoch": 59.96419213973799,
+ "grad_norm": 0.01734776981174946,
+ "learning_rate": 0.0006,
+ "loss": 4.572544574737549,
+ "step": 4317
+ },
+ {
+ "epoch": 59.97816593886463,
+ "grad_norm": 0.01822415366768837,
+ "learning_rate": 0.0006,
+ "loss": 4.490501880645752,
+ "step": 4318
+ },
+ {
+ "epoch": 59.992139737991266,
+ "grad_norm": 0.014638577587902546,
+ "learning_rate": 0.0006,
+ "loss": 4.440951824188232,
+ "step": 4319
+ },
+ {
+ "epoch": 60.0,
+ "grad_norm": 0.01255202479660511,
+ "learning_rate": 0.0006,
+ "loss": 4.559123516082764,
+ "step": 4320
+ },
+ {
+ "epoch": 60.0,
+ "eval_loss": 4.741089344024658,
+ "eval_runtime": 56.8997,
+ "eval_samples_per_second": 42.918,
+ "eval_steps_per_second": 1.353,
+ "step": 4320
+ },
+ {
+ "epoch": 60.01397379912664,
+ "grad_norm": 0.016548393294215202,
+ "learning_rate": 0.0006,
+ "loss": 4.442024230957031,
+ "step": 4321
+ },
+ {
+ "epoch": 60.02794759825328,
+ "grad_norm": 0.019759127870202065,
+ "learning_rate": 0.0006,
+ "loss": 4.394334316253662,
+ "step": 4322
+ },
+ {
+ "epoch": 60.041921397379916,
+ "grad_norm": 0.018325170502066612,
+ "learning_rate": 0.0006,
+ "loss": 4.318270683288574,
+ "step": 4323
+ },
+ {
+ "epoch": 60.05589519650655,
+ "grad_norm": 0.015289463102817535,
+ "learning_rate": 0.0006,
+ "loss": 4.318228721618652,
+ "step": 4324
+ },
+ {
+ "epoch": 60.069868995633186,
+ "grad_norm": 0.015424872748553753,
+ "learning_rate": 0.0006,
+ "loss": 4.3605499267578125,
+ "step": 4325
+ },
+ {
+ "epoch": 60.083842794759825,
+ "grad_norm": 0.014085669070482254,
+ "learning_rate": 0.0006,
+ "loss": 4.359585762023926,
+ "step": 4326
+ },
+ {
+ "epoch": 60.09781659388646,
+ "grad_norm": 0.014004210010170937,
+ "learning_rate": 0.0006,
+ "loss": 4.447783470153809,
+ "step": 4327
+ },
+ {
+ "epoch": 60.1117903930131,
+ "grad_norm": 0.013514582999050617,
+ "learning_rate": 0.0006,
+ "loss": 4.4850239753723145,
+ "step": 4328
+ },
+ {
+ "epoch": 60.12576419213974,
+ "grad_norm": 0.013049394823610783,
+ "learning_rate": 0.0006,
+ "loss": 4.528675556182861,
+ "step": 4329
+ },
+ {
+ "epoch": 60.13973799126637,
+ "grad_norm": 0.01378058921545744,
+ "learning_rate": 0.0006,
+ "loss": 4.444823741912842,
+ "step": 4330
+ },
+ {
+ "epoch": 60.15371179039301,
+ "grad_norm": 0.013752233237028122,
+ "learning_rate": 0.0006,
+ "loss": 4.5205206871032715,
+ "step": 4331
+ },
+ {
+ "epoch": 60.16768558951965,
+ "grad_norm": 0.013431803323328495,
+ "learning_rate": 0.0006,
+ "loss": 4.547320365905762,
+ "step": 4332
+ },
+ {
+ "epoch": 60.18165938864629,
+ "grad_norm": 0.014544196426868439,
+ "learning_rate": 0.0006,
+ "loss": 4.517143249511719,
+ "step": 4333
+ },
+ {
+ "epoch": 60.19563318777293,
+ "grad_norm": 0.01612176187336445,
+ "learning_rate": 0.0006,
+ "loss": 4.517232894897461,
+ "step": 4334
+ },
+ {
+ "epoch": 60.209606986899566,
+ "grad_norm": 0.01803239807486534,
+ "learning_rate": 0.0006,
+ "loss": 4.377574920654297,
+ "step": 4335
+ },
+ {
+ "epoch": 60.223580786026204,
+ "grad_norm": 0.014871489256620407,
+ "learning_rate": 0.0006,
+ "loss": 4.538285255432129,
+ "step": 4336
+ },
+ {
+ "epoch": 60.237554585152836,
+ "grad_norm": 0.013927377760410309,
+ "learning_rate": 0.0006,
+ "loss": 4.386175632476807,
+ "step": 4337
+ },
+ {
+ "epoch": 60.251528384279474,
+ "grad_norm": 0.014757219702005386,
+ "learning_rate": 0.0006,
+ "loss": 4.528553009033203,
+ "step": 4338
+ },
+ {
+ "epoch": 60.26550218340611,
+ "grad_norm": 0.01433873176574707,
+ "learning_rate": 0.0006,
+ "loss": 4.476929187774658,
+ "step": 4339
+ },
+ {
+ "epoch": 60.27947598253275,
+ "grad_norm": 0.015768803656101227,
+ "learning_rate": 0.0006,
+ "loss": 4.50970458984375,
+ "step": 4340
+ },
+ {
+ "epoch": 60.29344978165939,
+ "grad_norm": 0.015651429072022438,
+ "learning_rate": 0.0006,
+ "loss": 4.427133083343506,
+ "step": 4341
+ },
+ {
+ "epoch": 60.30742358078603,
+ "grad_norm": 0.01673000678420067,
+ "learning_rate": 0.0006,
+ "loss": 4.447723865509033,
+ "step": 4342
+ },
+ {
+ "epoch": 60.32139737991266,
+ "grad_norm": 0.018481194972991943,
+ "learning_rate": 0.0006,
+ "loss": 4.559832572937012,
+ "step": 4343
+ },
+ {
+ "epoch": 60.3353711790393,
+ "grad_norm": 0.016731027513742447,
+ "learning_rate": 0.0006,
+ "loss": 4.470893859863281,
+ "step": 4344
+ },
+ {
+ "epoch": 60.34934497816594,
+ "grad_norm": 0.01504011545330286,
+ "learning_rate": 0.0006,
+ "loss": 4.483014106750488,
+ "step": 4345
+ },
+ {
+ "epoch": 60.36331877729258,
+ "grad_norm": 0.01602456159889698,
+ "learning_rate": 0.0006,
+ "loss": 4.548428058624268,
+ "step": 4346
+ },
+ {
+ "epoch": 60.377292576419215,
+ "grad_norm": 0.016027364879846573,
+ "learning_rate": 0.0006,
+ "loss": 4.426779747009277,
+ "step": 4347
+ },
+ {
+ "epoch": 60.391266375545854,
+ "grad_norm": 0.01873827911913395,
+ "learning_rate": 0.0006,
+ "loss": 4.353686332702637,
+ "step": 4348
+ },
+ {
+ "epoch": 60.40524017467249,
+ "grad_norm": 0.01907406374812126,
+ "learning_rate": 0.0006,
+ "loss": 4.502945423126221,
+ "step": 4349
+ },
+ {
+ "epoch": 60.419213973799124,
+ "grad_norm": 0.016083354130387306,
+ "learning_rate": 0.0006,
+ "loss": 4.431731700897217,
+ "step": 4350
+ },
+ {
+ "epoch": 60.43318777292576,
+ "grad_norm": 0.015293709933757782,
+ "learning_rate": 0.0006,
+ "loss": 4.480112552642822,
+ "step": 4351
+ },
+ {
+ "epoch": 60.4471615720524,
+ "grad_norm": 0.01688101328909397,
+ "learning_rate": 0.0006,
+ "loss": 4.50853157043457,
+ "step": 4352
+ },
+ {
+ "epoch": 60.46113537117904,
+ "grad_norm": 0.02094118855893612,
+ "learning_rate": 0.0006,
+ "loss": 4.359930038452148,
+ "step": 4353
+ },
+ {
+ "epoch": 60.47510917030568,
+ "grad_norm": 0.018188107758760452,
+ "learning_rate": 0.0006,
+ "loss": 4.549592018127441,
+ "step": 4354
+ },
+ {
+ "epoch": 60.48908296943232,
+ "grad_norm": 0.01462319865822792,
+ "learning_rate": 0.0006,
+ "loss": 4.49467658996582,
+ "step": 4355
+ },
+ {
+ "epoch": 60.50305676855895,
+ "grad_norm": 0.017508579418063164,
+ "learning_rate": 0.0006,
+ "loss": 4.398318767547607,
+ "step": 4356
+ },
+ {
+ "epoch": 60.51703056768559,
+ "grad_norm": 0.018200315535068512,
+ "learning_rate": 0.0006,
+ "loss": 4.44169282913208,
+ "step": 4357
+ },
+ {
+ "epoch": 60.531004366812226,
+ "grad_norm": 0.01717468351125717,
+ "learning_rate": 0.0006,
+ "loss": 4.481400489807129,
+ "step": 4358
+ },
+ {
+ "epoch": 60.544978165938865,
+ "grad_norm": 0.019329151138663292,
+ "learning_rate": 0.0006,
+ "loss": 4.578839302062988,
+ "step": 4359
+ },
+ {
+ "epoch": 60.5589519650655,
+ "grad_norm": 0.016109555959701538,
+ "learning_rate": 0.0006,
+ "loss": 4.461043357849121,
+ "step": 4360
+ },
+ {
+ "epoch": 60.57292576419214,
+ "grad_norm": 0.01494457945227623,
+ "learning_rate": 0.0006,
+ "loss": 4.505096912384033,
+ "step": 4361
+ },
+ {
+ "epoch": 60.58689956331878,
+ "grad_norm": 0.01729399710893631,
+ "learning_rate": 0.0006,
+ "loss": 4.366747856140137,
+ "step": 4362
+ },
+ {
+ "epoch": 60.60087336244541,
+ "grad_norm": 0.01705821603536606,
+ "learning_rate": 0.0006,
+ "loss": 4.42534065246582,
+ "step": 4363
+ },
+ {
+ "epoch": 60.61484716157205,
+ "grad_norm": 0.018391354009509087,
+ "learning_rate": 0.0006,
+ "loss": 4.350294589996338,
+ "step": 4364
+ },
+ {
+ "epoch": 60.62882096069869,
+ "grad_norm": 0.016989829018712044,
+ "learning_rate": 0.0006,
+ "loss": 4.45905876159668,
+ "step": 4365
+ },
+ {
+ "epoch": 60.64279475982533,
+ "grad_norm": 0.017192212864756584,
+ "learning_rate": 0.0006,
+ "loss": 4.526487350463867,
+ "step": 4366
+ },
+ {
+ "epoch": 60.65676855895197,
+ "grad_norm": 0.018091315403580666,
+ "learning_rate": 0.0006,
+ "loss": 4.309333324432373,
+ "step": 4367
+ },
+ {
+ "epoch": 60.670742358078606,
+ "grad_norm": 0.014051459729671478,
+ "learning_rate": 0.0006,
+ "loss": 4.338171005249023,
+ "step": 4368
+ },
+ {
+ "epoch": 60.68471615720524,
+ "grad_norm": 0.013166418299078941,
+ "learning_rate": 0.0006,
+ "loss": 4.409758567810059,
+ "step": 4369
+ },
+ {
+ "epoch": 60.698689956331876,
+ "grad_norm": 0.014260428957641125,
+ "learning_rate": 0.0006,
+ "loss": 4.524338245391846,
+ "step": 4370
+ },
+ {
+ "epoch": 60.712663755458514,
+ "grad_norm": 0.017019255086779594,
+ "learning_rate": 0.0006,
+ "loss": 4.387436866760254,
+ "step": 4371
+ },
+ {
+ "epoch": 60.72663755458515,
+ "grad_norm": 0.017070675268769264,
+ "learning_rate": 0.0006,
+ "loss": 4.437321662902832,
+ "step": 4372
+ },
+ {
+ "epoch": 60.74061135371179,
+ "grad_norm": 0.015197164379060268,
+ "learning_rate": 0.0006,
+ "loss": 4.371205806732178,
+ "step": 4373
+ },
+ {
+ "epoch": 60.75458515283843,
+ "grad_norm": 0.014791185967624187,
+ "learning_rate": 0.0006,
+ "loss": 4.44439697265625,
+ "step": 4374
+ },
+ {
+ "epoch": 60.76855895196506,
+ "grad_norm": 0.016458261758089066,
+ "learning_rate": 0.0006,
+ "loss": 4.445584297180176,
+ "step": 4375
+ },
+ {
+ "epoch": 60.7825327510917,
+ "grad_norm": 0.017706342041492462,
+ "learning_rate": 0.0006,
+ "loss": 4.348708152770996,
+ "step": 4376
+ },
+ {
+ "epoch": 60.79650655021834,
+ "grad_norm": 0.017561476677656174,
+ "learning_rate": 0.0006,
+ "loss": 4.521929740905762,
+ "step": 4377
+ },
+ {
+ "epoch": 60.81048034934498,
+ "grad_norm": 0.016557445749640465,
+ "learning_rate": 0.0006,
+ "loss": 4.545806884765625,
+ "step": 4378
+ },
+ {
+ "epoch": 60.82445414847162,
+ "grad_norm": 0.016953222453594208,
+ "learning_rate": 0.0006,
+ "loss": 4.4230875968933105,
+ "step": 4379
+ },
+ {
+ "epoch": 60.838427947598255,
+ "grad_norm": 0.017094967886805534,
+ "learning_rate": 0.0006,
+ "loss": 4.423985481262207,
+ "step": 4380
+ },
+ {
+ "epoch": 60.852401746724894,
+ "grad_norm": 0.014787515625357628,
+ "learning_rate": 0.0006,
+ "loss": 4.39473819732666,
+ "step": 4381
+ },
+ {
+ "epoch": 60.866375545851525,
+ "grad_norm": 0.016208263114094734,
+ "learning_rate": 0.0006,
+ "loss": 4.3810038566589355,
+ "step": 4382
+ },
+ {
+ "epoch": 60.880349344978164,
+ "grad_norm": 0.017844321206212044,
+ "learning_rate": 0.0006,
+ "loss": 4.35086727142334,
+ "step": 4383
+ },
+ {
+ "epoch": 60.8943231441048,
+ "grad_norm": 0.017539294436573982,
+ "learning_rate": 0.0006,
+ "loss": 4.460260391235352,
+ "step": 4384
+ },
+ {
+ "epoch": 60.90829694323144,
+ "grad_norm": 0.01612042263150215,
+ "learning_rate": 0.0006,
+ "loss": 4.370500564575195,
+ "step": 4385
+ },
+ {
+ "epoch": 60.92227074235808,
+ "grad_norm": 0.015381601639091969,
+ "learning_rate": 0.0006,
+ "loss": 4.435868263244629,
+ "step": 4386
+ },
+ {
+ "epoch": 60.93624454148472,
+ "grad_norm": 0.01603585295379162,
+ "learning_rate": 0.0006,
+ "loss": 4.434070587158203,
+ "step": 4387
+ },
+ {
+ "epoch": 60.95021834061135,
+ "grad_norm": 0.016268130391836166,
+ "learning_rate": 0.0006,
+ "loss": 4.404343128204346,
+ "step": 4388
+ },
+ {
+ "epoch": 60.96419213973799,
+ "grad_norm": 0.014980032108724117,
+ "learning_rate": 0.0006,
+ "loss": 4.399213790893555,
+ "step": 4389
+ },
+ {
+ "epoch": 60.97816593886463,
+ "grad_norm": 0.016899267211556435,
+ "learning_rate": 0.0006,
+ "loss": 4.508069038391113,
+ "step": 4390
+ },
+ {
+ "epoch": 60.992139737991266,
+ "grad_norm": 0.013756908476352692,
+ "learning_rate": 0.0006,
+ "loss": 4.541858673095703,
+ "step": 4391
+ },
+ {
+ "epoch": 61.0,
+ "grad_norm": 0.01578413136303425,
+ "learning_rate": 0.0006,
+ "loss": 4.543513298034668,
+ "step": 4392
+ },
+ {
+ "epoch": 61.0,
+ "eval_loss": 4.756770610809326,
+ "eval_runtime": 56.4572,
+ "eval_samples_per_second": 43.254,
+ "eval_steps_per_second": 1.364,
+ "step": 4392
+ },
+ {
+ "epoch": 61.01397379912664,
+ "grad_norm": 0.013387506827712059,
+ "learning_rate": 0.0006,
+ "loss": 4.475942611694336,
+ "step": 4393
+ },
+ {
+ "epoch": 61.02794759825328,
+ "grad_norm": 0.014278904534876347,
+ "learning_rate": 0.0006,
+ "loss": 4.442403793334961,
+ "step": 4394
+ },
+ {
+ "epoch": 61.041921397379916,
+ "grad_norm": 0.014359553344547749,
+ "learning_rate": 0.0006,
+ "loss": 4.509463310241699,
+ "step": 4395
+ },
+ {
+ "epoch": 61.05589519650655,
+ "grad_norm": 0.01718050241470337,
+ "learning_rate": 0.0006,
+ "loss": 4.354458808898926,
+ "step": 4396
+ },
+ {
+ "epoch": 61.069868995633186,
+ "grad_norm": 0.015988312661647797,
+ "learning_rate": 0.0006,
+ "loss": 4.484508514404297,
+ "step": 4397
+ },
+ {
+ "epoch": 61.083842794759825,
+ "grad_norm": 0.015466735698282719,
+ "learning_rate": 0.0006,
+ "loss": 4.326908111572266,
+ "step": 4398
+ },
+ {
+ "epoch": 61.09781659388646,
+ "grad_norm": 0.014541332609951496,
+ "learning_rate": 0.0006,
+ "loss": 4.405823230743408,
+ "step": 4399
+ },
+ {
+ "epoch": 61.1117903930131,
+ "grad_norm": 0.014803987927734852,
+ "learning_rate": 0.0006,
+ "loss": 4.375386714935303,
+ "step": 4400
+ },
+ {
+ "epoch": 61.12576419213974,
+ "grad_norm": 0.013962093740701675,
+ "learning_rate": 0.0006,
+ "loss": 4.300838470458984,
+ "step": 4401
+ },
+ {
+ "epoch": 61.13973799126637,
+ "grad_norm": 0.015213954262435436,
+ "learning_rate": 0.0006,
+ "loss": 4.388306617736816,
+ "step": 4402
+ },
+ {
+ "epoch": 61.15371179039301,
+ "grad_norm": 0.018965506926178932,
+ "learning_rate": 0.0006,
+ "loss": 4.388340950012207,
+ "step": 4403
+ },
+ {
+ "epoch": 61.16768558951965,
+ "grad_norm": 0.01958410255610943,
+ "learning_rate": 0.0006,
+ "loss": 4.35012149810791,
+ "step": 4404
+ },
+ {
+ "epoch": 61.18165938864629,
+ "grad_norm": 0.01938101276755333,
+ "learning_rate": 0.0006,
+ "loss": 4.368372917175293,
+ "step": 4405
+ },
+ {
+ "epoch": 61.19563318777293,
+ "grad_norm": 0.016556516289711,
+ "learning_rate": 0.0006,
+ "loss": 4.486198425292969,
+ "step": 4406
+ },
+ {
+ "epoch": 61.209606986899566,
+ "grad_norm": 0.017031289637088776,
+ "learning_rate": 0.0006,
+ "loss": 4.421676158905029,
+ "step": 4407
+ },
+ {
+ "epoch": 61.223580786026204,
+ "grad_norm": 0.018023479729890823,
+ "learning_rate": 0.0006,
+ "loss": 4.38751220703125,
+ "step": 4408
+ },
+ {
+ "epoch": 61.237554585152836,
+ "grad_norm": 0.0178135447204113,
+ "learning_rate": 0.0006,
+ "loss": 4.336528301239014,
+ "step": 4409
+ },
+ {
+ "epoch": 61.251528384279474,
+ "grad_norm": 0.016782190650701523,
+ "learning_rate": 0.0006,
+ "loss": 4.386693954467773,
+ "step": 4410
+ },
+ {
+ "epoch": 61.26550218340611,
+ "grad_norm": 0.015494848601520061,
+ "learning_rate": 0.0006,
+ "loss": 4.308948040008545,
+ "step": 4411
+ },
+ {
+ "epoch": 61.27947598253275,
+ "grad_norm": 0.015002378262579441,
+ "learning_rate": 0.0006,
+ "loss": 4.314206123352051,
+ "step": 4412
+ },
+ {
+ "epoch": 61.29344978165939,
+ "grad_norm": 0.015798920765519142,
+ "learning_rate": 0.0006,
+ "loss": 4.502900123596191,
+ "step": 4413
+ },
+ {
+ "epoch": 61.30742358078603,
+ "grad_norm": 0.015444336459040642,
+ "learning_rate": 0.0006,
+ "loss": 4.473160743713379,
+ "step": 4414
+ },
+ {
+ "epoch": 61.32139737991266,
+ "grad_norm": 0.017025096341967583,
+ "learning_rate": 0.0006,
+ "loss": 4.460621356964111,
+ "step": 4415
+ },
+ {
+ "epoch": 61.3353711790393,
+ "grad_norm": 0.019502537325024605,
+ "learning_rate": 0.0006,
+ "loss": 4.385551452636719,
+ "step": 4416
+ },
+ {
+ "epoch": 61.34934497816594,
+ "grad_norm": 0.018529541790485382,
+ "learning_rate": 0.0006,
+ "loss": 4.425390720367432,
+ "step": 4417
+ },
+ {
+ "epoch": 61.36331877729258,
+ "grad_norm": 0.013155706226825714,
+ "learning_rate": 0.0006,
+ "loss": 4.467020034790039,
+ "step": 4418
+ },
+ {
+ "epoch": 61.377292576419215,
+ "grad_norm": 0.01518090907484293,
+ "learning_rate": 0.0006,
+ "loss": 4.369664192199707,
+ "step": 4419
+ },
+ {
+ "epoch": 61.391266375545854,
+ "grad_norm": 0.01626511849462986,
+ "learning_rate": 0.0006,
+ "loss": 4.463231086730957,
+ "step": 4420
+ },
+ {
+ "epoch": 61.40524017467249,
+ "grad_norm": 0.01701558195054531,
+ "learning_rate": 0.0006,
+ "loss": 4.43708610534668,
+ "step": 4421
+ },
+ {
+ "epoch": 61.419213973799124,
+ "grad_norm": 0.01761862449347973,
+ "learning_rate": 0.0006,
+ "loss": 4.3314208984375,
+ "step": 4422
+ },
+ {
+ "epoch": 61.43318777292576,
+ "grad_norm": 0.017084648832678795,
+ "learning_rate": 0.0006,
+ "loss": 4.39676570892334,
+ "step": 4423
+ },
+ {
+ "epoch": 61.4471615720524,
+ "grad_norm": 0.01667078025639057,
+ "learning_rate": 0.0006,
+ "loss": 4.3789873123168945,
+ "step": 4424
+ },
+ {
+ "epoch": 61.46113537117904,
+ "grad_norm": 0.01877436414361,
+ "learning_rate": 0.0006,
+ "loss": 4.415275573730469,
+ "step": 4425
+ },
+ {
+ "epoch": 61.47510917030568,
+ "grad_norm": 0.020250486209988594,
+ "learning_rate": 0.0006,
+ "loss": 4.478635787963867,
+ "step": 4426
+ },
+ {
+ "epoch": 61.48908296943232,
+ "grad_norm": 0.020244868472218513,
+ "learning_rate": 0.0006,
+ "loss": 4.296854019165039,
+ "step": 4427
+ },
+ {
+ "epoch": 61.50305676855895,
+ "grad_norm": 0.018723875284194946,
+ "learning_rate": 0.0006,
+ "loss": 4.3760881423950195,
+ "step": 4428
+ },
+ {
+ "epoch": 61.51703056768559,
+ "grad_norm": 0.015138164162635803,
+ "learning_rate": 0.0006,
+ "loss": 4.39974308013916,
+ "step": 4429
+ },
+ {
+ "epoch": 61.531004366812226,
+ "grad_norm": 0.015111138112843037,
+ "learning_rate": 0.0006,
+ "loss": 4.34659481048584,
+ "step": 4430
+ },
+ {
+ "epoch": 61.544978165938865,
+ "grad_norm": 0.0164666585624218,
+ "learning_rate": 0.0006,
+ "loss": 4.415737152099609,
+ "step": 4431
+ },
+ {
+ "epoch": 61.5589519650655,
+ "grad_norm": 0.020891312509775162,
+ "learning_rate": 0.0006,
+ "loss": 4.300319671630859,
+ "step": 4432
+ },
+ {
+ "epoch": 61.57292576419214,
+ "grad_norm": 0.02089156024158001,
+ "learning_rate": 0.0006,
+ "loss": 4.373029708862305,
+ "step": 4433
+ },
+ {
+ "epoch": 61.58689956331878,
+ "grad_norm": 0.020114559680223465,
+ "learning_rate": 0.0006,
+ "loss": 4.434499740600586,
+ "step": 4434
+ },
+ {
+ "epoch": 61.60087336244541,
+ "grad_norm": 0.01791365258395672,
+ "learning_rate": 0.0006,
+ "loss": 4.466872215270996,
+ "step": 4435
+ },
+ {
+ "epoch": 61.61484716157205,
+ "grad_norm": 0.01879458874464035,
+ "learning_rate": 0.0006,
+ "loss": 4.5176520347595215,
+ "step": 4436
+ },
+ {
+ "epoch": 61.62882096069869,
+ "grad_norm": 0.016371937468647957,
+ "learning_rate": 0.0006,
+ "loss": 4.3507585525512695,
+ "step": 4437
+ },
+ {
+ "epoch": 61.64279475982533,
+ "grad_norm": 0.01648143120110035,
+ "learning_rate": 0.0006,
+ "loss": 4.398778438568115,
+ "step": 4438
+ },
+ {
+ "epoch": 61.65676855895197,
+ "grad_norm": 0.017240576446056366,
+ "learning_rate": 0.0006,
+ "loss": 4.513933181762695,
+ "step": 4439
+ },
+ {
+ "epoch": 61.670742358078606,
+ "grad_norm": 0.01734284684062004,
+ "learning_rate": 0.0006,
+ "loss": 4.315343856811523,
+ "step": 4440
+ },
+ {
+ "epoch": 61.68471615720524,
+ "grad_norm": 0.01639772206544876,
+ "learning_rate": 0.0006,
+ "loss": 4.4641218185424805,
+ "step": 4441
+ },
+ {
+ "epoch": 61.698689956331876,
+ "grad_norm": 0.0164708960801363,
+ "learning_rate": 0.0006,
+ "loss": 4.439968585968018,
+ "step": 4442
+ },
+ {
+ "epoch": 61.712663755458514,
+ "grad_norm": 0.01757933758199215,
+ "learning_rate": 0.0006,
+ "loss": 4.408327102661133,
+ "step": 4443
+ },
+ {
+ "epoch": 61.72663755458515,
+ "grad_norm": 0.01789589412510395,
+ "learning_rate": 0.0006,
+ "loss": 4.478257179260254,
+ "step": 4444
+ },
+ {
+ "epoch": 61.74061135371179,
+ "grad_norm": 0.01822184957563877,
+ "learning_rate": 0.0006,
+ "loss": 4.412046909332275,
+ "step": 4445
+ },
+ {
+ "epoch": 61.75458515283843,
+ "grad_norm": 0.018624618649482727,
+ "learning_rate": 0.0006,
+ "loss": 4.427122116088867,
+ "step": 4446
+ },
+ {
+ "epoch": 61.76855895196506,
+ "grad_norm": 0.018674887716770172,
+ "learning_rate": 0.0006,
+ "loss": 4.429147720336914,
+ "step": 4447
+ },
+ {
+ "epoch": 61.7825327510917,
+ "grad_norm": 0.015317119657993317,
+ "learning_rate": 0.0006,
+ "loss": 4.2808451652526855,
+ "step": 4448
+ },
+ {
+ "epoch": 61.79650655021834,
+ "grad_norm": 0.015400845557451248,
+ "learning_rate": 0.0006,
+ "loss": 4.420193672180176,
+ "step": 4449
+ },
+ {
+ "epoch": 61.81048034934498,
+ "grad_norm": 0.016743650659918785,
+ "learning_rate": 0.0006,
+ "loss": 4.498254776000977,
+ "step": 4450
+ },
+ {
+ "epoch": 61.82445414847162,
+ "grad_norm": 0.014330781064927578,
+ "learning_rate": 0.0006,
+ "loss": 4.445384979248047,
+ "step": 4451
+ },
+ {
+ "epoch": 61.838427947598255,
+ "grad_norm": 0.015046479180455208,
+ "learning_rate": 0.0006,
+ "loss": 4.368041038513184,
+ "step": 4452
+ },
+ {
+ "epoch": 61.852401746724894,
+ "grad_norm": 0.01630168780684471,
+ "learning_rate": 0.0006,
+ "loss": 4.460142135620117,
+ "step": 4453
+ },
+ {
+ "epoch": 61.866375545851525,
+ "grad_norm": 0.01637885719537735,
+ "learning_rate": 0.0006,
+ "loss": 4.416606426239014,
+ "step": 4454
+ },
+ {
+ "epoch": 61.880349344978164,
+ "grad_norm": 0.016785888001322746,
+ "learning_rate": 0.0006,
+ "loss": 4.452545642852783,
+ "step": 4455
+ },
+ {
+ "epoch": 61.8943231441048,
+ "grad_norm": 0.015044555068016052,
+ "learning_rate": 0.0006,
+ "loss": 4.427334785461426,
+ "step": 4456
+ },
+ {
+ "epoch": 61.90829694323144,
+ "grad_norm": 0.013699211180210114,
+ "learning_rate": 0.0006,
+ "loss": 4.408761024475098,
+ "step": 4457
+ },
+ {
+ "epoch": 61.92227074235808,
+ "grad_norm": 0.014021803624927998,
+ "learning_rate": 0.0006,
+ "loss": 4.543607711791992,
+ "step": 4458
+ },
+ {
+ "epoch": 61.93624454148472,
+ "grad_norm": 0.015438460744917393,
+ "learning_rate": 0.0006,
+ "loss": 4.313363075256348,
+ "step": 4459
+ },
+ {
+ "epoch": 61.95021834061135,
+ "grad_norm": 0.017515428364276886,
+ "learning_rate": 0.0006,
+ "loss": 4.310412406921387,
+ "step": 4460
+ },
+ {
+ "epoch": 61.96419213973799,
+ "grad_norm": 0.016683880239725113,
+ "learning_rate": 0.0006,
+ "loss": 4.3378448486328125,
+ "step": 4461
+ },
+ {
+ "epoch": 61.97816593886463,
+ "grad_norm": 0.019384048879146576,
+ "learning_rate": 0.0006,
+ "loss": 4.450568675994873,
+ "step": 4462
+ },
+ {
+ "epoch": 61.992139737991266,
+ "grad_norm": 0.020651493221521378,
+ "learning_rate": 0.0006,
+ "loss": 4.453197479248047,
+ "step": 4463
+ },
+ {
+ "epoch": 62.0,
+ "grad_norm": 0.020991764962673187,
+ "learning_rate": 0.0006,
+ "loss": 4.441527366638184,
+ "step": 4464
+ },
+ {
+ "epoch": 62.0,
+ "eval_loss": 4.700193881988525,
+ "eval_runtime": 56.6098,
+ "eval_samples_per_second": 43.137,
+ "eval_steps_per_second": 1.36,
+ "step": 4464
+ },
+ {
+ "epoch": 62.01397379912664,
+ "grad_norm": 0.01811373233795166,
+ "learning_rate": 0.0006,
+ "loss": 4.37346076965332,
+ "step": 4465
+ },
+ {
+ "epoch": 62.02794759825328,
+ "grad_norm": 0.01667127199470997,
+ "learning_rate": 0.0006,
+ "loss": 4.237151145935059,
+ "step": 4466
+ },
+ {
+ "epoch": 62.041921397379916,
+ "grad_norm": 0.014971534721553326,
+ "learning_rate": 0.0006,
+ "loss": 4.367053031921387,
+ "step": 4467
+ },
+ {
+ "epoch": 62.05589519650655,
+ "grad_norm": 0.015655893832445145,
+ "learning_rate": 0.0006,
+ "loss": 4.437426567077637,
+ "step": 4468
+ },
+ {
+ "epoch": 62.069868995633186,
+ "grad_norm": 0.014935447834432125,
+ "learning_rate": 0.0006,
+ "loss": 4.3635125160217285,
+ "step": 4469
+ },
+ {
+ "epoch": 62.083842794759825,
+ "grad_norm": 0.015363802202045918,
+ "learning_rate": 0.0006,
+ "loss": 4.40492057800293,
+ "step": 4470
+ },
+ {
+ "epoch": 62.09781659388646,
+ "grad_norm": 0.0159201230853796,
+ "learning_rate": 0.0006,
+ "loss": 4.392058849334717,
+ "step": 4471
+ },
+ {
+ "epoch": 62.1117903930131,
+ "grad_norm": 0.014598241075873375,
+ "learning_rate": 0.0006,
+ "loss": 4.369944095611572,
+ "step": 4472
+ },
+ {
+ "epoch": 62.12576419213974,
+ "grad_norm": 0.015394841320812702,
+ "learning_rate": 0.0006,
+ "loss": 4.425097942352295,
+ "step": 4473
+ },
+ {
+ "epoch": 62.13973799126637,
+ "grad_norm": 0.015913628041744232,
+ "learning_rate": 0.0006,
+ "loss": 4.420596122741699,
+ "step": 4474
+ },
+ {
+ "epoch": 62.15371179039301,
+ "grad_norm": 0.01575258933007717,
+ "learning_rate": 0.0006,
+ "loss": 4.32290506362915,
+ "step": 4475
+ },
+ {
+ "epoch": 62.16768558951965,
+ "grad_norm": 0.0155344782397151,
+ "learning_rate": 0.0006,
+ "loss": 4.273068428039551,
+ "step": 4476
+ },
+ {
+ "epoch": 62.18165938864629,
+ "grad_norm": 0.014947040006518364,
+ "learning_rate": 0.0006,
+ "loss": 4.384868621826172,
+ "step": 4477
+ },
+ {
+ "epoch": 62.19563318777293,
+ "grad_norm": 0.01557591650635004,
+ "learning_rate": 0.0006,
+ "loss": 4.458410263061523,
+ "step": 4478
+ },
+ {
+ "epoch": 62.209606986899566,
+ "grad_norm": 0.015182198956608772,
+ "learning_rate": 0.0006,
+ "loss": 4.300738334655762,
+ "step": 4479
+ },
+ {
+ "epoch": 62.223580786026204,
+ "grad_norm": 0.01634717360138893,
+ "learning_rate": 0.0006,
+ "loss": 4.436875820159912,
+ "step": 4480
+ },
+ {
+ "epoch": 62.237554585152836,
+ "grad_norm": 0.018813718110322952,
+ "learning_rate": 0.0006,
+ "loss": 4.454665184020996,
+ "step": 4481
+ },
+ {
+ "epoch": 62.251528384279474,
+ "grad_norm": 0.020965630188584328,
+ "learning_rate": 0.0006,
+ "loss": 4.357341766357422,
+ "step": 4482
+ },
+ {
+ "epoch": 62.26550218340611,
+ "grad_norm": 0.020738353952765465,
+ "learning_rate": 0.0006,
+ "loss": 4.456644058227539,
+ "step": 4483
+ },
+ {
+ "epoch": 62.27947598253275,
+ "grad_norm": 0.018813546746969223,
+ "learning_rate": 0.0006,
+ "loss": 4.381941795349121,
+ "step": 4484
+ },
+ {
+ "epoch": 62.29344978165939,
+ "grad_norm": 0.015431761741638184,
+ "learning_rate": 0.0006,
+ "loss": 4.41126823425293,
+ "step": 4485
+ },
+ {
+ "epoch": 62.30742358078603,
+ "grad_norm": 0.016335977241396904,
+ "learning_rate": 0.0006,
+ "loss": 4.419488906860352,
+ "step": 4486
+ },
+ {
+ "epoch": 62.32139737991266,
+ "grad_norm": 0.01832406409084797,
+ "learning_rate": 0.0006,
+ "loss": 4.579026222229004,
+ "step": 4487
+ },
+ {
+ "epoch": 62.3353711790393,
+ "grad_norm": 0.01790742017328739,
+ "learning_rate": 0.0006,
+ "loss": 4.289964199066162,
+ "step": 4488
+ },
+ {
+ "epoch": 62.34934497816594,
+ "grad_norm": 0.017266202718019485,
+ "learning_rate": 0.0006,
+ "loss": 4.415959358215332,
+ "step": 4489
+ },
+ {
+ "epoch": 62.36331877729258,
+ "grad_norm": 0.015976132825016975,
+ "learning_rate": 0.0006,
+ "loss": 4.358587265014648,
+ "step": 4490
+ },
+ {
+ "epoch": 62.377292576419215,
+ "grad_norm": 0.016873881220817566,
+ "learning_rate": 0.0006,
+ "loss": 4.354607105255127,
+ "step": 4491
+ },
+ {
+ "epoch": 62.391266375545854,
+ "grad_norm": 0.01476745493710041,
+ "learning_rate": 0.0006,
+ "loss": 4.384904384613037,
+ "step": 4492
+ },
+ {
+ "epoch": 62.40524017467249,
+ "grad_norm": 0.013810782693326473,
+ "learning_rate": 0.0006,
+ "loss": 4.459230899810791,
+ "step": 4493
+ },
+ {
+ "epoch": 62.419213973799124,
+ "grad_norm": 0.015150421299040318,
+ "learning_rate": 0.0006,
+ "loss": 4.369471549987793,
+ "step": 4494
+ },
+ {
+ "epoch": 62.43318777292576,
+ "grad_norm": 0.014539610594511032,
+ "learning_rate": 0.0006,
+ "loss": 4.3564982414245605,
+ "step": 4495
+ },
+ {
+ "epoch": 62.4471615720524,
+ "grad_norm": 0.015554017387330532,
+ "learning_rate": 0.0006,
+ "loss": 4.414379119873047,
+ "step": 4496
+ },
+ {
+ "epoch": 62.46113537117904,
+ "grad_norm": 0.016608357429504395,
+ "learning_rate": 0.0006,
+ "loss": 4.311585426330566,
+ "step": 4497
+ },
+ {
+ "epoch": 62.47510917030568,
+ "grad_norm": 0.01885881833732128,
+ "learning_rate": 0.0006,
+ "loss": 4.306629180908203,
+ "step": 4498
+ },
+ {
+ "epoch": 62.48908296943232,
+ "grad_norm": 0.022128663957118988,
+ "learning_rate": 0.0006,
+ "loss": 4.345483779907227,
+ "step": 4499
+ },
+ {
+ "epoch": 62.50305676855895,
+ "grad_norm": 0.023401018232107162,
+ "learning_rate": 0.0006,
+ "loss": 4.398204326629639,
+ "step": 4500
+ },
+ {
+ "epoch": 62.51703056768559,
+ "grad_norm": 0.022080009803175926,
+ "learning_rate": 0.0006,
+ "loss": 4.375027179718018,
+ "step": 4501
+ },
+ {
+ "epoch": 62.531004366812226,
+ "grad_norm": 0.020004864782094955,
+ "learning_rate": 0.0006,
+ "loss": 4.33494758605957,
+ "step": 4502
+ },
+ {
+ "epoch": 62.544978165938865,
+ "grad_norm": 0.01786700263619423,
+ "learning_rate": 0.0006,
+ "loss": 4.4428300857543945,
+ "step": 4503
+ },
+ {
+ "epoch": 62.5589519650655,
+ "grad_norm": 0.01885388419032097,
+ "learning_rate": 0.0006,
+ "loss": 4.4138946533203125,
+ "step": 4504
+ },
+ {
+ "epoch": 62.57292576419214,
+ "grad_norm": 0.019890939816832542,
+ "learning_rate": 0.0006,
+ "loss": 4.278632164001465,
+ "step": 4505
+ },
+ {
+ "epoch": 62.58689956331878,
+ "grad_norm": 0.02109031192958355,
+ "learning_rate": 0.0006,
+ "loss": 4.3756327629089355,
+ "step": 4506
+ },
+ {
+ "epoch": 62.60087336244541,
+ "grad_norm": 0.020813899114727974,
+ "learning_rate": 0.0006,
+ "loss": 4.410213470458984,
+ "step": 4507
+ },
+ {
+ "epoch": 62.61484716157205,
+ "grad_norm": 0.021655689924955368,
+ "learning_rate": 0.0006,
+ "loss": 4.30633020401001,
+ "step": 4508
+ },
+ {
+ "epoch": 62.62882096069869,
+ "grad_norm": 0.018155649304389954,
+ "learning_rate": 0.0006,
+ "loss": 4.407284736633301,
+ "step": 4509
+ },
+ {
+ "epoch": 62.64279475982533,
+ "grad_norm": 0.017765069380402565,
+ "learning_rate": 0.0006,
+ "loss": 4.321715354919434,
+ "step": 4510
+ },
+ {
+ "epoch": 62.65676855895197,
+ "grad_norm": 0.017724554985761642,
+ "learning_rate": 0.0006,
+ "loss": 4.493371963500977,
+ "step": 4511
+ },
+ {
+ "epoch": 62.670742358078606,
+ "grad_norm": 0.01492092851549387,
+ "learning_rate": 0.0006,
+ "loss": 4.443025588989258,
+ "step": 4512
+ },
+ {
+ "epoch": 62.68471615720524,
+ "grad_norm": 0.01615324057638645,
+ "learning_rate": 0.0006,
+ "loss": 4.3289594650268555,
+ "step": 4513
+ },
+ {
+ "epoch": 62.698689956331876,
+ "grad_norm": 0.016569366678595543,
+ "learning_rate": 0.0006,
+ "loss": 4.369720458984375,
+ "step": 4514
+ },
+ {
+ "epoch": 62.712663755458514,
+ "grad_norm": 0.016509901732206345,
+ "learning_rate": 0.0006,
+ "loss": 4.292316436767578,
+ "step": 4515
+ },
+ {
+ "epoch": 62.72663755458515,
+ "grad_norm": 0.014270003885030746,
+ "learning_rate": 0.0006,
+ "loss": 4.349854469299316,
+ "step": 4516
+ },
+ {
+ "epoch": 62.74061135371179,
+ "grad_norm": 0.014235340990126133,
+ "learning_rate": 0.0006,
+ "loss": 4.473727226257324,
+ "step": 4517
+ },
+ {
+ "epoch": 62.75458515283843,
+ "grad_norm": 0.015163225121796131,
+ "learning_rate": 0.0006,
+ "loss": 4.452878475189209,
+ "step": 4518
+ },
+ {
+ "epoch": 62.76855895196506,
+ "grad_norm": 0.015748055651783943,
+ "learning_rate": 0.0006,
+ "loss": 4.521087169647217,
+ "step": 4519
+ },
+ {
+ "epoch": 62.7825327510917,
+ "grad_norm": 0.01661108434200287,
+ "learning_rate": 0.0006,
+ "loss": 4.443885326385498,
+ "step": 4520
+ },
+ {
+ "epoch": 62.79650655021834,
+ "grad_norm": 0.01610301062464714,
+ "learning_rate": 0.0006,
+ "loss": 4.312990188598633,
+ "step": 4521
+ },
+ {
+ "epoch": 62.81048034934498,
+ "grad_norm": 0.017007583752274513,
+ "learning_rate": 0.0006,
+ "loss": 4.408937454223633,
+ "step": 4522
+ },
+ {
+ "epoch": 62.82445414847162,
+ "grad_norm": 0.01749461516737938,
+ "learning_rate": 0.0006,
+ "loss": 4.346066951751709,
+ "step": 4523
+ },
+ {
+ "epoch": 62.838427947598255,
+ "grad_norm": 0.018243493512272835,
+ "learning_rate": 0.0006,
+ "loss": 4.357658386230469,
+ "step": 4524
+ },
+ {
+ "epoch": 62.852401746724894,
+ "grad_norm": 0.017499985173344612,
+ "learning_rate": 0.0006,
+ "loss": 4.374061584472656,
+ "step": 4525
+ },
+ {
+ "epoch": 62.866375545851525,
+ "grad_norm": 0.014536325819790363,
+ "learning_rate": 0.0006,
+ "loss": 4.189321994781494,
+ "step": 4526
+ },
+ {
+ "epoch": 62.880349344978164,
+ "grad_norm": 0.014728891663253307,
+ "learning_rate": 0.0006,
+ "loss": 4.425910949707031,
+ "step": 4527
+ },
+ {
+ "epoch": 62.8943231441048,
+ "grad_norm": 0.015013772062957287,
+ "learning_rate": 0.0006,
+ "loss": 4.337319850921631,
+ "step": 4528
+ },
+ {
+ "epoch": 62.90829694323144,
+ "grad_norm": 0.01626027375459671,
+ "learning_rate": 0.0006,
+ "loss": 4.330339431762695,
+ "step": 4529
+ },
+ {
+ "epoch": 62.92227074235808,
+ "grad_norm": 0.014773444272577763,
+ "learning_rate": 0.0006,
+ "loss": 4.4188737869262695,
+ "step": 4530
+ },
+ {
+ "epoch": 62.93624454148472,
+ "grad_norm": 0.01422932744026184,
+ "learning_rate": 0.0006,
+ "loss": 4.355986595153809,
+ "step": 4531
+ },
+ {
+ "epoch": 62.95021834061135,
+ "grad_norm": 0.012170110829174519,
+ "learning_rate": 0.0006,
+ "loss": 4.405722618103027,
+ "step": 4532
+ },
+ {
+ "epoch": 62.96419213973799,
+ "grad_norm": 0.013233253732323647,
+ "learning_rate": 0.0006,
+ "loss": 4.377286911010742,
+ "step": 4533
+ },
+ {
+ "epoch": 62.97816593886463,
+ "grad_norm": 0.014497299678623676,
+ "learning_rate": 0.0006,
+ "loss": 4.508796691894531,
+ "step": 4534
+ },
+ {
+ "epoch": 62.992139737991266,
+ "grad_norm": 0.014718741178512573,
+ "learning_rate": 0.0006,
+ "loss": 4.370919704437256,
+ "step": 4535
+ },
+ {
+ "epoch": 63.0,
+ "grad_norm": 0.015833502635359764,
+ "learning_rate": 0.0006,
+ "loss": 4.258753776550293,
+ "step": 4536
+ },
+ {
+ "epoch": 63.0,
+ "eval_loss": 4.714163303375244,
+ "eval_runtime": 57.0269,
+ "eval_samples_per_second": 42.822,
+ "eval_steps_per_second": 1.35,
+ "step": 4536
+ },
+ {
+ "epoch": 63.01397379912664,
+ "grad_norm": 0.014277510344982147,
+ "learning_rate": 0.0006,
+ "loss": 4.3112640380859375,
+ "step": 4537
+ },
+ {
+ "epoch": 63.02794759825328,
+ "grad_norm": 0.015377935022115707,
+ "learning_rate": 0.0006,
+ "loss": 4.342782974243164,
+ "step": 4538
+ },
+ {
+ "epoch": 63.041921397379916,
+ "grad_norm": 0.015406244434416294,
+ "learning_rate": 0.0006,
+ "loss": 4.474056720733643,
+ "step": 4539
+ },
+ {
+ "epoch": 63.05589519650655,
+ "grad_norm": 0.018121499568223953,
+ "learning_rate": 0.0006,
+ "loss": 4.397963523864746,
+ "step": 4540
+ },
+ {
+ "epoch": 63.069868995633186,
+ "grad_norm": 0.022634636610746384,
+ "learning_rate": 0.0006,
+ "loss": 4.306631088256836,
+ "step": 4541
+ },
+ {
+ "epoch": 63.083842794759825,
+ "grad_norm": 0.026448970660567284,
+ "learning_rate": 0.0006,
+ "loss": 4.283175468444824,
+ "step": 4542
+ },
+ {
+ "epoch": 63.09781659388646,
+ "grad_norm": 0.024104537442326546,
+ "learning_rate": 0.0006,
+ "loss": 4.315826416015625,
+ "step": 4543
+ },
+ {
+ "epoch": 63.1117903930131,
+ "grad_norm": 0.019040480256080627,
+ "learning_rate": 0.0006,
+ "loss": 4.44057035446167,
+ "step": 4544
+ },
+ {
+ "epoch": 63.12576419213974,
+ "grad_norm": 0.02006000280380249,
+ "learning_rate": 0.0006,
+ "loss": 4.405603408813477,
+ "step": 4545
+ },
+ {
+ "epoch": 63.13973799126637,
+ "grad_norm": 0.019980071112513542,
+ "learning_rate": 0.0006,
+ "loss": 4.393726348876953,
+ "step": 4546
+ },
+ {
+ "epoch": 63.15371179039301,
+ "grad_norm": 0.01641865260899067,
+ "learning_rate": 0.0006,
+ "loss": 4.272546768188477,
+ "step": 4547
+ },
+ {
+ "epoch": 63.16768558951965,
+ "grad_norm": 0.01890353485941887,
+ "learning_rate": 0.0006,
+ "loss": 4.382566928863525,
+ "step": 4548
+ },
+ {
+ "epoch": 63.18165938864629,
+ "grad_norm": 0.018370937556028366,
+ "learning_rate": 0.0006,
+ "loss": 4.423953533172607,
+ "step": 4549
+ },
+ {
+ "epoch": 63.19563318777293,
+ "grad_norm": 0.014985294081270695,
+ "learning_rate": 0.0006,
+ "loss": 4.308627128601074,
+ "step": 4550
+ },
+ {
+ "epoch": 63.209606986899566,
+ "grad_norm": 0.017513327300548553,
+ "learning_rate": 0.0006,
+ "loss": 4.248819351196289,
+ "step": 4551
+ },
+ {
+ "epoch": 63.223580786026204,
+ "grad_norm": 0.019464161247015,
+ "learning_rate": 0.0006,
+ "loss": 4.359729766845703,
+ "step": 4552
+ },
+ {
+ "epoch": 63.237554585152836,
+ "grad_norm": 0.018678035587072372,
+ "learning_rate": 0.0006,
+ "loss": 4.3299665451049805,
+ "step": 4553
+ },
+ {
+ "epoch": 63.251528384279474,
+ "grad_norm": 0.017909523099660873,
+ "learning_rate": 0.0006,
+ "loss": 4.246147155761719,
+ "step": 4554
+ },
+ {
+ "epoch": 63.26550218340611,
+ "grad_norm": 0.01747293211519718,
+ "learning_rate": 0.0006,
+ "loss": 4.41165018081665,
+ "step": 4555
+ },
+ {
+ "epoch": 63.27947598253275,
+ "grad_norm": 0.018670853227376938,
+ "learning_rate": 0.0006,
+ "loss": 4.409883499145508,
+ "step": 4556
+ },
+ {
+ "epoch": 63.29344978165939,
+ "grad_norm": 0.0189529862254858,
+ "learning_rate": 0.0006,
+ "loss": 4.418654918670654,
+ "step": 4557
+ },
+ {
+ "epoch": 63.30742358078603,
+ "grad_norm": 0.018394414335489273,
+ "learning_rate": 0.0006,
+ "loss": 4.38355827331543,
+ "step": 4558
+ },
+ {
+ "epoch": 63.32139737991266,
+ "grad_norm": 0.018317820504307747,
+ "learning_rate": 0.0006,
+ "loss": 4.456843376159668,
+ "step": 4559
+ },
+ {
+ "epoch": 63.3353711790393,
+ "grad_norm": 0.019225597381591797,
+ "learning_rate": 0.0006,
+ "loss": 4.232264518737793,
+ "step": 4560
+ },
+ {
+ "epoch": 63.34934497816594,
+ "grad_norm": 0.016663681715726852,
+ "learning_rate": 0.0006,
+ "loss": 4.393517017364502,
+ "step": 4561
+ },
+ {
+ "epoch": 63.36331877729258,
+ "grad_norm": 0.014975365251302719,
+ "learning_rate": 0.0006,
+ "loss": 4.375183582305908,
+ "step": 4562
+ },
+ {
+ "epoch": 63.377292576419215,
+ "grad_norm": 0.01771732047200203,
+ "learning_rate": 0.0006,
+ "loss": 4.3073410987854,
+ "step": 4563
+ },
+ {
+ "epoch": 63.391266375545854,
+ "grad_norm": 0.01663549244403839,
+ "learning_rate": 0.0006,
+ "loss": 4.311003684997559,
+ "step": 4564
+ },
+ {
+ "epoch": 63.40524017467249,
+ "grad_norm": 0.016301177442073822,
+ "learning_rate": 0.0006,
+ "loss": 4.326807022094727,
+ "step": 4565
+ },
+ {
+ "epoch": 63.419213973799124,
+ "grad_norm": 0.016781188547611237,
+ "learning_rate": 0.0006,
+ "loss": 4.306562423706055,
+ "step": 4566
+ },
+ {
+ "epoch": 63.43318777292576,
+ "grad_norm": 0.01767745055258274,
+ "learning_rate": 0.0006,
+ "loss": 4.367082595825195,
+ "step": 4567
+ },
+ {
+ "epoch": 63.4471615720524,
+ "grad_norm": 0.016728565096855164,
+ "learning_rate": 0.0006,
+ "loss": 4.327400207519531,
+ "step": 4568
+ },
+ {
+ "epoch": 63.46113537117904,
+ "grad_norm": 0.016289403662085533,
+ "learning_rate": 0.0006,
+ "loss": 4.392712116241455,
+ "step": 4569
+ },
+ {
+ "epoch": 63.47510917030568,
+ "grad_norm": 0.016233716160058975,
+ "learning_rate": 0.0006,
+ "loss": 4.354372024536133,
+ "step": 4570
+ },
+ {
+ "epoch": 63.48908296943232,
+ "grad_norm": 0.016547974199056625,
+ "learning_rate": 0.0006,
+ "loss": 4.309050559997559,
+ "step": 4571
+ },
+ {
+ "epoch": 63.50305676855895,
+ "grad_norm": 0.018466509878635406,
+ "learning_rate": 0.0006,
+ "loss": 4.318345069885254,
+ "step": 4572
+ },
+ {
+ "epoch": 63.51703056768559,
+ "grad_norm": 0.017685720697045326,
+ "learning_rate": 0.0006,
+ "loss": 4.399726867675781,
+ "step": 4573
+ },
+ {
+ "epoch": 63.531004366812226,
+ "grad_norm": 0.015801234170794487,
+ "learning_rate": 0.0006,
+ "loss": 4.328495502471924,
+ "step": 4574
+ },
+ {
+ "epoch": 63.544978165938865,
+ "grad_norm": 0.0188873540610075,
+ "learning_rate": 0.0006,
+ "loss": 4.253244400024414,
+ "step": 4575
+ },
+ {
+ "epoch": 63.5589519650655,
+ "grad_norm": 0.01868271268904209,
+ "learning_rate": 0.0006,
+ "loss": 4.463746070861816,
+ "step": 4576
+ },
+ {
+ "epoch": 63.57292576419214,
+ "grad_norm": 0.01844949834048748,
+ "learning_rate": 0.0006,
+ "loss": 4.400300025939941,
+ "step": 4577
+ },
+ {
+ "epoch": 63.58689956331878,
+ "grad_norm": 0.015256480313837528,
+ "learning_rate": 0.0006,
+ "loss": 4.445844650268555,
+ "step": 4578
+ },
+ {
+ "epoch": 63.60087336244541,
+ "grad_norm": 0.015305760316550732,
+ "learning_rate": 0.0006,
+ "loss": 4.317327499389648,
+ "step": 4579
+ },
+ {
+ "epoch": 63.61484716157205,
+ "grad_norm": 0.015200769528746605,
+ "learning_rate": 0.0006,
+ "loss": 4.25112247467041,
+ "step": 4580
+ },
+ {
+ "epoch": 63.62882096069869,
+ "grad_norm": 0.015919538214802742,
+ "learning_rate": 0.0006,
+ "loss": 4.277937889099121,
+ "step": 4581
+ },
+ {
+ "epoch": 63.64279475982533,
+ "grad_norm": 0.01572665013372898,
+ "learning_rate": 0.0006,
+ "loss": 4.302237510681152,
+ "step": 4582
+ },
+ {
+ "epoch": 63.65676855895197,
+ "grad_norm": 0.016428731381893158,
+ "learning_rate": 0.0006,
+ "loss": 4.36762809753418,
+ "step": 4583
+ },
+ {
+ "epoch": 63.670742358078606,
+ "grad_norm": 0.01706082746386528,
+ "learning_rate": 0.0006,
+ "loss": 4.236526966094971,
+ "step": 4584
+ },
+ {
+ "epoch": 63.68471615720524,
+ "grad_norm": 0.017850136384367943,
+ "learning_rate": 0.0006,
+ "loss": 4.264129638671875,
+ "step": 4585
+ },
+ {
+ "epoch": 63.698689956331876,
+ "grad_norm": 0.016079438850283623,
+ "learning_rate": 0.0006,
+ "loss": 4.449799060821533,
+ "step": 4586
+ },
+ {
+ "epoch": 63.712663755458514,
+ "grad_norm": 0.014996383339166641,
+ "learning_rate": 0.0006,
+ "loss": 4.404574394226074,
+ "step": 4587
+ },
+ {
+ "epoch": 63.72663755458515,
+ "grad_norm": 0.015910431742668152,
+ "learning_rate": 0.0006,
+ "loss": 4.348101615905762,
+ "step": 4588
+ },
+ {
+ "epoch": 63.74061135371179,
+ "grad_norm": 0.01630021072924137,
+ "learning_rate": 0.0006,
+ "loss": 4.3193817138671875,
+ "step": 4589
+ },
+ {
+ "epoch": 63.75458515283843,
+ "grad_norm": 0.014571773819625378,
+ "learning_rate": 0.0006,
+ "loss": 4.374402046203613,
+ "step": 4590
+ },
+ {
+ "epoch": 63.76855895196506,
+ "grad_norm": 0.01662345975637436,
+ "learning_rate": 0.0006,
+ "loss": 4.329075813293457,
+ "step": 4591
+ },
+ {
+ "epoch": 63.7825327510917,
+ "grad_norm": 0.016604198142886162,
+ "learning_rate": 0.0006,
+ "loss": 4.436126232147217,
+ "step": 4592
+ },
+ {
+ "epoch": 63.79650655021834,
+ "grad_norm": 0.016578884795308113,
+ "learning_rate": 0.0006,
+ "loss": 4.361468315124512,
+ "step": 4593
+ },
+ {
+ "epoch": 63.81048034934498,
+ "grad_norm": 0.016339825466275215,
+ "learning_rate": 0.0006,
+ "loss": 4.449892044067383,
+ "step": 4594
+ },
+ {
+ "epoch": 63.82445414847162,
+ "grad_norm": 0.016805455088615417,
+ "learning_rate": 0.0006,
+ "loss": 4.415866851806641,
+ "step": 4595
+ },
+ {
+ "epoch": 63.838427947598255,
+ "grad_norm": 0.017366213724017143,
+ "learning_rate": 0.0006,
+ "loss": 4.361730575561523,
+ "step": 4596
+ },
+ {
+ "epoch": 63.852401746724894,
+ "grad_norm": 0.015599003992974758,
+ "learning_rate": 0.0006,
+ "loss": 4.363637924194336,
+ "step": 4597
+ },
+ {
+ "epoch": 63.866375545851525,
+ "grad_norm": 0.015674632042646408,
+ "learning_rate": 0.0006,
+ "loss": 4.332045078277588,
+ "step": 4598
+ },
+ {
+ "epoch": 63.880349344978164,
+ "grad_norm": 0.015350443311035633,
+ "learning_rate": 0.0006,
+ "loss": 4.320063591003418,
+ "step": 4599
+ },
+ {
+ "epoch": 63.8943231441048,
+ "grad_norm": 0.015388146974146366,
+ "learning_rate": 0.0006,
+ "loss": 4.398199081420898,
+ "step": 4600
+ },
+ {
+ "epoch": 63.90829694323144,
+ "grad_norm": 0.01450140681117773,
+ "learning_rate": 0.0006,
+ "loss": 4.244076728820801,
+ "step": 4601
+ },
+ {
+ "epoch": 63.92227074235808,
+ "grad_norm": 0.01637491025030613,
+ "learning_rate": 0.0006,
+ "loss": 4.213549613952637,
+ "step": 4602
+ },
+ {
+ "epoch": 63.93624454148472,
+ "grad_norm": 0.0159814041107893,
+ "learning_rate": 0.0006,
+ "loss": 4.421775817871094,
+ "step": 4603
+ },
+ {
+ "epoch": 63.95021834061135,
+ "grad_norm": 0.017386827617883682,
+ "learning_rate": 0.0006,
+ "loss": 4.340023994445801,
+ "step": 4604
+ },
+ {
+ "epoch": 63.96419213973799,
+ "grad_norm": 0.019679656252264977,
+ "learning_rate": 0.0006,
+ "loss": 4.352534294128418,
+ "step": 4605
+ },
+ {
+ "epoch": 63.97816593886463,
+ "grad_norm": 0.022391516715288162,
+ "learning_rate": 0.0006,
+ "loss": 4.3538312911987305,
+ "step": 4606
+ },
+ {
+ "epoch": 63.992139737991266,
+ "grad_norm": 0.023281006142497063,
+ "learning_rate": 0.0006,
+ "loss": 4.432089328765869,
+ "step": 4607
+ },
+ {
+ "epoch": 64.0,
+ "grad_norm": 0.019916288554668427,
+ "learning_rate": 0.0006,
+ "loss": 4.466937065124512,
+ "step": 4608
+ },
+ {
+ "epoch": 64.0,
+ "eval_loss": 4.659008502960205,
+ "eval_runtime": 56.0019,
+ "eval_samples_per_second": 43.606,
+ "eval_steps_per_second": 1.375,
+ "step": 4608
+ },
+ {
+ "epoch": 64.01397379912663,
+ "grad_norm": 0.017351187765598297,
+ "learning_rate": 0.0006,
+ "loss": 4.294129371643066,
+ "step": 4609
+ },
+ {
+ "epoch": 64.02794759825328,
+ "grad_norm": 0.02405865117907524,
+ "learning_rate": 0.0006,
+ "loss": 4.266717910766602,
+ "step": 4610
+ },
+ {
+ "epoch": 64.04192139737991,
+ "grad_norm": 0.022158386185765266,
+ "learning_rate": 0.0006,
+ "loss": 4.395932197570801,
+ "step": 4611
+ },
+ {
+ "epoch": 64.05589519650655,
+ "grad_norm": 0.019238505512475967,
+ "learning_rate": 0.0006,
+ "loss": 4.266704082489014,
+ "step": 4612
+ },
+ {
+ "epoch": 64.06986899563319,
+ "grad_norm": 0.021670425310730934,
+ "learning_rate": 0.0006,
+ "loss": 4.3160552978515625,
+ "step": 4613
+ },
+ {
+ "epoch": 64.08384279475983,
+ "grad_norm": 0.02076641097664833,
+ "learning_rate": 0.0006,
+ "loss": 4.415626525878906,
+ "step": 4614
+ },
+ {
+ "epoch": 64.09781659388646,
+ "grad_norm": 0.018223078921437263,
+ "learning_rate": 0.0006,
+ "loss": 4.292387008666992,
+ "step": 4615
+ },
+ {
+ "epoch": 64.1117903930131,
+ "grad_norm": 0.017401937395334244,
+ "learning_rate": 0.0006,
+ "loss": 4.424524307250977,
+ "step": 4616
+ },
+ {
+ "epoch": 64.12576419213974,
+ "grad_norm": 0.01663743332028389,
+ "learning_rate": 0.0006,
+ "loss": 4.350032806396484,
+ "step": 4617
+ },
+ {
+ "epoch": 64.13973799126637,
+ "grad_norm": 0.018716629594564438,
+ "learning_rate": 0.0006,
+ "loss": 4.309157371520996,
+ "step": 4618
+ },
+ {
+ "epoch": 64.15371179039302,
+ "grad_norm": 0.019569825381040573,
+ "learning_rate": 0.0006,
+ "loss": 4.358292579650879,
+ "step": 4619
+ },
+ {
+ "epoch": 64.16768558951965,
+ "grad_norm": 0.018994422629475594,
+ "learning_rate": 0.0006,
+ "loss": 4.3338823318481445,
+ "step": 4620
+ },
+ {
+ "epoch": 64.18165938864628,
+ "grad_norm": 0.017690688371658325,
+ "learning_rate": 0.0006,
+ "loss": 4.414468765258789,
+ "step": 4621
+ },
+ {
+ "epoch": 64.19563318777293,
+ "grad_norm": 0.0170609001070261,
+ "learning_rate": 0.0006,
+ "loss": 4.4337568283081055,
+ "step": 4622
+ },
+ {
+ "epoch": 64.20960698689956,
+ "grad_norm": 0.016984131187200546,
+ "learning_rate": 0.0006,
+ "loss": 4.304839134216309,
+ "step": 4623
+ },
+ {
+ "epoch": 64.2235807860262,
+ "grad_norm": 0.01619867794215679,
+ "learning_rate": 0.0006,
+ "loss": 4.389623165130615,
+ "step": 4624
+ },
+ {
+ "epoch": 64.23755458515284,
+ "grad_norm": 0.015428110957145691,
+ "learning_rate": 0.0006,
+ "loss": 4.277515888214111,
+ "step": 4625
+ },
+ {
+ "epoch": 64.25152838427948,
+ "grad_norm": 0.015909532085061073,
+ "learning_rate": 0.0006,
+ "loss": 4.266029357910156,
+ "step": 4626
+ },
+ {
+ "epoch": 64.26550218340611,
+ "grad_norm": 0.014296560548245907,
+ "learning_rate": 0.0006,
+ "loss": 4.286067485809326,
+ "step": 4627
+ },
+ {
+ "epoch": 64.27947598253274,
+ "grad_norm": 0.014065971598029137,
+ "learning_rate": 0.0006,
+ "loss": 4.353528022766113,
+ "step": 4628
+ },
+ {
+ "epoch": 64.29344978165939,
+ "grad_norm": 0.014386293478310108,
+ "learning_rate": 0.0006,
+ "loss": 4.318715572357178,
+ "step": 4629
+ },
+ {
+ "epoch": 64.30742358078602,
+ "grad_norm": 0.014001123607158661,
+ "learning_rate": 0.0006,
+ "loss": 4.362497329711914,
+ "step": 4630
+ },
+ {
+ "epoch": 64.32139737991267,
+ "grad_norm": 0.013408979400992393,
+ "learning_rate": 0.0006,
+ "loss": 4.417397499084473,
+ "step": 4631
+ },
+ {
+ "epoch": 64.3353711790393,
+ "grad_norm": 0.014960093423724174,
+ "learning_rate": 0.0006,
+ "loss": 4.359710693359375,
+ "step": 4632
+ },
+ {
+ "epoch": 64.34934497816595,
+ "grad_norm": 0.014658777043223381,
+ "learning_rate": 0.0006,
+ "loss": 4.318822383880615,
+ "step": 4633
+ },
+ {
+ "epoch": 64.36331877729258,
+ "grad_norm": 0.013766744174063206,
+ "learning_rate": 0.0006,
+ "loss": 4.265679836273193,
+ "step": 4634
+ },
+ {
+ "epoch": 64.37729257641921,
+ "grad_norm": 0.013821255415678024,
+ "learning_rate": 0.0006,
+ "loss": 4.3452911376953125,
+ "step": 4635
+ },
+ {
+ "epoch": 64.39126637554585,
+ "grad_norm": 0.013541066087782383,
+ "learning_rate": 0.0006,
+ "loss": 4.429254531860352,
+ "step": 4636
+ },
+ {
+ "epoch": 64.40524017467249,
+ "grad_norm": 0.013985645025968552,
+ "learning_rate": 0.0006,
+ "loss": 4.304452896118164,
+ "step": 4637
+ },
+ {
+ "epoch": 64.41921397379913,
+ "grad_norm": 0.015079808421432972,
+ "learning_rate": 0.0006,
+ "loss": 4.296828746795654,
+ "step": 4638
+ },
+ {
+ "epoch": 64.43318777292576,
+ "grad_norm": 0.014896944165229797,
+ "learning_rate": 0.0006,
+ "loss": 4.299156665802002,
+ "step": 4639
+ },
+ {
+ "epoch": 64.44716157205241,
+ "grad_norm": 0.013770041987299919,
+ "learning_rate": 0.0006,
+ "loss": 4.351439952850342,
+ "step": 4640
+ },
+ {
+ "epoch": 64.46113537117904,
+ "grad_norm": 0.01531178504228592,
+ "learning_rate": 0.0006,
+ "loss": 4.373715877532959,
+ "step": 4641
+ },
+ {
+ "epoch": 64.47510917030567,
+ "grad_norm": 0.014024450443685055,
+ "learning_rate": 0.0006,
+ "loss": 4.368053436279297,
+ "step": 4642
+ },
+ {
+ "epoch": 64.48908296943232,
+ "grad_norm": 0.013864024542272091,
+ "learning_rate": 0.0006,
+ "loss": 4.378564834594727,
+ "step": 4643
+ },
+ {
+ "epoch": 64.50305676855895,
+ "grad_norm": 0.013925445266067982,
+ "learning_rate": 0.0006,
+ "loss": 4.261663436889648,
+ "step": 4644
+ },
+ {
+ "epoch": 64.5170305676856,
+ "grad_norm": 0.015011338517069817,
+ "learning_rate": 0.0006,
+ "loss": 4.337285995483398,
+ "step": 4645
+ },
+ {
+ "epoch": 64.53100436681223,
+ "grad_norm": 0.016941986978054047,
+ "learning_rate": 0.0006,
+ "loss": 4.336877822875977,
+ "step": 4646
+ },
+ {
+ "epoch": 64.54497816593886,
+ "grad_norm": 0.01719909906387329,
+ "learning_rate": 0.0006,
+ "loss": 4.149813175201416,
+ "step": 4647
+ },
+ {
+ "epoch": 64.5589519650655,
+ "grad_norm": 0.01608402095735073,
+ "learning_rate": 0.0006,
+ "loss": 4.191280364990234,
+ "step": 4648
+ },
+ {
+ "epoch": 64.57292576419214,
+ "grad_norm": 0.015284820459783077,
+ "learning_rate": 0.0006,
+ "loss": 4.35953426361084,
+ "step": 4649
+ },
+ {
+ "epoch": 64.58689956331878,
+ "grad_norm": 0.014855682849884033,
+ "learning_rate": 0.0006,
+ "loss": 4.308401107788086,
+ "step": 4650
+ },
+ {
+ "epoch": 64.60087336244541,
+ "grad_norm": 0.017553111538290977,
+ "learning_rate": 0.0006,
+ "loss": 4.279720306396484,
+ "step": 4651
+ },
+ {
+ "epoch": 64.61484716157206,
+ "grad_norm": 0.019506998360157013,
+ "learning_rate": 0.0006,
+ "loss": 4.319524765014648,
+ "step": 4652
+ },
+ {
+ "epoch": 64.62882096069869,
+ "grad_norm": 0.020293425768613815,
+ "learning_rate": 0.0006,
+ "loss": 4.340523719787598,
+ "step": 4653
+ },
+ {
+ "epoch": 64.64279475982532,
+ "grad_norm": 0.01664556749165058,
+ "learning_rate": 0.0006,
+ "loss": 4.296785831451416,
+ "step": 4654
+ },
+ {
+ "epoch": 64.65676855895197,
+ "grad_norm": 0.012592585757374763,
+ "learning_rate": 0.0006,
+ "loss": 4.464595317840576,
+ "step": 4655
+ },
+ {
+ "epoch": 64.6707423580786,
+ "grad_norm": 0.01549589168280363,
+ "learning_rate": 0.0006,
+ "loss": 4.275053024291992,
+ "step": 4656
+ },
+ {
+ "epoch": 64.68471615720524,
+ "grad_norm": 0.01705172471702099,
+ "learning_rate": 0.0006,
+ "loss": 4.388431549072266,
+ "step": 4657
+ },
+ {
+ "epoch": 64.69868995633188,
+ "grad_norm": 0.015438582748174667,
+ "learning_rate": 0.0006,
+ "loss": 4.312070846557617,
+ "step": 4658
+ },
+ {
+ "epoch": 64.71266375545852,
+ "grad_norm": 0.014448083005845547,
+ "learning_rate": 0.0006,
+ "loss": 4.237599849700928,
+ "step": 4659
+ },
+ {
+ "epoch": 64.72663755458515,
+ "grad_norm": 0.015267071314156055,
+ "learning_rate": 0.0006,
+ "loss": 4.29893684387207,
+ "step": 4660
+ },
+ {
+ "epoch": 64.74061135371178,
+ "grad_norm": 0.016615847125649452,
+ "learning_rate": 0.0006,
+ "loss": 4.432273864746094,
+ "step": 4661
+ },
+ {
+ "epoch": 64.75458515283843,
+ "grad_norm": 0.016584362834692,
+ "learning_rate": 0.0006,
+ "loss": 4.202266216278076,
+ "step": 4662
+ },
+ {
+ "epoch": 64.76855895196506,
+ "grad_norm": 0.018313994631171227,
+ "learning_rate": 0.0006,
+ "loss": 4.367966651916504,
+ "step": 4663
+ },
+ {
+ "epoch": 64.78253275109171,
+ "grad_norm": 0.017946481704711914,
+ "learning_rate": 0.0006,
+ "loss": 4.205172538757324,
+ "step": 4664
+ },
+ {
+ "epoch": 64.79650655021834,
+ "grad_norm": 0.017744354903697968,
+ "learning_rate": 0.0006,
+ "loss": 4.295661926269531,
+ "step": 4665
+ },
+ {
+ "epoch": 64.81048034934499,
+ "grad_norm": 0.016103824600577354,
+ "learning_rate": 0.0006,
+ "loss": 4.276543617248535,
+ "step": 4666
+ },
+ {
+ "epoch": 64.82445414847162,
+ "grad_norm": 0.014590012840926647,
+ "learning_rate": 0.0006,
+ "loss": 4.414247035980225,
+ "step": 4667
+ },
+ {
+ "epoch": 64.83842794759825,
+ "grad_norm": 0.01782386749982834,
+ "learning_rate": 0.0006,
+ "loss": 4.203220367431641,
+ "step": 4668
+ },
+ {
+ "epoch": 64.8524017467249,
+ "grad_norm": 0.022122733294963837,
+ "learning_rate": 0.0006,
+ "loss": 4.38078498840332,
+ "step": 4669
+ },
+ {
+ "epoch": 64.86637554585153,
+ "grad_norm": 0.0250293780118227,
+ "learning_rate": 0.0006,
+ "loss": 4.3310136795043945,
+ "step": 4670
+ },
+ {
+ "epoch": 64.88034934497817,
+ "grad_norm": 0.022438062354922295,
+ "learning_rate": 0.0006,
+ "loss": 4.428631782531738,
+ "step": 4671
+ },
+ {
+ "epoch": 64.8943231441048,
+ "grad_norm": 0.018518192693591118,
+ "learning_rate": 0.0006,
+ "loss": 4.428250312805176,
+ "step": 4672
+ },
+ {
+ "epoch": 64.90829694323143,
+ "grad_norm": 0.018105225637555122,
+ "learning_rate": 0.0006,
+ "loss": 4.336209297180176,
+ "step": 4673
+ },
+ {
+ "epoch": 64.92227074235808,
+ "grad_norm": 0.017564399167895317,
+ "learning_rate": 0.0006,
+ "loss": 4.416011810302734,
+ "step": 4674
+ },
+ {
+ "epoch": 64.93624454148471,
+ "grad_norm": 0.018747031688690186,
+ "learning_rate": 0.0006,
+ "loss": 4.419938087463379,
+ "step": 4675
+ },
+ {
+ "epoch": 64.95021834061136,
+ "grad_norm": 0.018686484545469284,
+ "learning_rate": 0.0006,
+ "loss": 4.347185134887695,
+ "step": 4676
+ },
+ {
+ "epoch": 64.96419213973799,
+ "grad_norm": 0.019329898059368134,
+ "learning_rate": 0.0006,
+ "loss": 4.343749046325684,
+ "step": 4677
+ },
+ {
+ "epoch": 64.97816593886463,
+ "grad_norm": 0.018517782911658287,
+ "learning_rate": 0.0006,
+ "loss": 4.264604568481445,
+ "step": 4678
+ },
+ {
+ "epoch": 64.99213973799127,
+ "grad_norm": 0.016579125076532364,
+ "learning_rate": 0.0006,
+ "loss": 4.303423881530762,
+ "step": 4679
+ },
+ {
+ "epoch": 65.0,
+ "grad_norm": 0.01617567613720894,
+ "learning_rate": 0.0006,
+ "loss": 4.326556205749512,
+ "step": 4680
+ },
+ {
+ "epoch": 65.0,
+ "eval_loss": 4.655026912689209,
+ "eval_runtime": 57.1299,
+ "eval_samples_per_second": 42.745,
+ "eval_steps_per_second": 1.348,
+ "step": 4680
+ },
+ {
+ "epoch": 65.01397379912663,
+ "grad_norm": 0.01607249490916729,
+ "learning_rate": 0.0006,
+ "loss": 4.276607036590576,
+ "step": 4681
+ },
+ {
+ "epoch": 65.02794759825328,
+ "grad_norm": 0.016779497265815735,
+ "learning_rate": 0.0006,
+ "loss": 4.412464618682861,
+ "step": 4682
+ },
+ {
+ "epoch": 65.04192139737991,
+ "grad_norm": 0.018161555752158165,
+ "learning_rate": 0.0006,
+ "loss": 4.199329376220703,
+ "step": 4683
+ },
+ {
+ "epoch": 65.05589519650655,
+ "grad_norm": 0.017508309334516525,
+ "learning_rate": 0.0006,
+ "loss": 4.217567443847656,
+ "step": 4684
+ },
+ {
+ "epoch": 65.06986899563319,
+ "grad_norm": 0.015716979280114174,
+ "learning_rate": 0.0006,
+ "loss": 4.322023391723633,
+ "step": 4685
+ },
+ {
+ "epoch": 65.08384279475983,
+ "grad_norm": 0.015113635919988155,
+ "learning_rate": 0.0006,
+ "loss": 4.333515167236328,
+ "step": 4686
+ },
+ {
+ "epoch": 65.09781659388646,
+ "grad_norm": 0.01437560748308897,
+ "learning_rate": 0.0006,
+ "loss": 4.295609474182129,
+ "step": 4687
+ },
+ {
+ "epoch": 65.1117903930131,
+ "grad_norm": 0.013074862770736217,
+ "learning_rate": 0.0006,
+ "loss": 4.314120769500732,
+ "step": 4688
+ },
+ {
+ "epoch": 65.12576419213974,
+ "grad_norm": 0.012908284552395344,
+ "learning_rate": 0.0006,
+ "loss": 4.216890811920166,
+ "step": 4689
+ },
+ {
+ "epoch": 65.13973799126637,
+ "grad_norm": 0.013001754879951477,
+ "learning_rate": 0.0006,
+ "loss": 4.287717342376709,
+ "step": 4690
+ },
+ {
+ "epoch": 65.15371179039302,
+ "grad_norm": 0.013694263994693756,
+ "learning_rate": 0.0006,
+ "loss": 4.333837985992432,
+ "step": 4691
+ },
+ {
+ "epoch": 65.16768558951965,
+ "grad_norm": 0.014461833983659744,
+ "learning_rate": 0.0006,
+ "loss": 4.323545455932617,
+ "step": 4692
+ },
+ {
+ "epoch": 65.18165938864628,
+ "grad_norm": 0.013670675456523895,
+ "learning_rate": 0.0006,
+ "loss": 4.3313493728637695,
+ "step": 4693
+ },
+ {
+ "epoch": 65.19563318777293,
+ "grad_norm": 0.013094350695610046,
+ "learning_rate": 0.0006,
+ "loss": 4.23858118057251,
+ "step": 4694
+ },
+ {
+ "epoch": 65.20960698689956,
+ "grad_norm": 0.013996944762766361,
+ "learning_rate": 0.0006,
+ "loss": 4.428705215454102,
+ "step": 4695
+ },
+ {
+ "epoch": 65.2235807860262,
+ "grad_norm": 0.013295048847794533,
+ "learning_rate": 0.0006,
+ "loss": 4.238950252532959,
+ "step": 4696
+ },
+ {
+ "epoch": 65.23755458515284,
+ "grad_norm": 0.013380616903305054,
+ "learning_rate": 0.0006,
+ "loss": 4.273185729980469,
+ "step": 4697
+ },
+ {
+ "epoch": 65.25152838427948,
+ "grad_norm": 0.015062890015542507,
+ "learning_rate": 0.0006,
+ "loss": 4.29472017288208,
+ "step": 4698
+ },
+ {
+ "epoch": 65.26550218340611,
+ "grad_norm": 0.017317263409495354,
+ "learning_rate": 0.0006,
+ "loss": 4.337954998016357,
+ "step": 4699
+ },
+ {
+ "epoch": 65.27947598253274,
+ "grad_norm": 0.017698992043733597,
+ "learning_rate": 0.0006,
+ "loss": 4.336437225341797,
+ "step": 4700
+ },
+ {
+ "epoch": 65.29344978165939,
+ "grad_norm": 0.01628575474023819,
+ "learning_rate": 0.0006,
+ "loss": 4.302372455596924,
+ "step": 4701
+ },
+ {
+ "epoch": 65.30742358078602,
+ "grad_norm": 0.014531854540109634,
+ "learning_rate": 0.0006,
+ "loss": 4.256450653076172,
+ "step": 4702
+ },
+ {
+ "epoch": 65.32139737991267,
+ "grad_norm": 0.014549105428159237,
+ "learning_rate": 0.0006,
+ "loss": 4.302240371704102,
+ "step": 4703
+ },
+ {
+ "epoch": 65.3353711790393,
+ "grad_norm": 0.014881663955748081,
+ "learning_rate": 0.0006,
+ "loss": 4.3316121101379395,
+ "step": 4704
+ },
+ {
+ "epoch": 65.34934497816595,
+ "grad_norm": 0.014513437636196613,
+ "learning_rate": 0.0006,
+ "loss": 4.441212177276611,
+ "step": 4705
+ },
+ {
+ "epoch": 65.36331877729258,
+ "grad_norm": 0.014122087508440018,
+ "learning_rate": 0.0006,
+ "loss": 4.343296527862549,
+ "step": 4706
+ },
+ {
+ "epoch": 65.37729257641921,
+ "grad_norm": 0.014817348681390285,
+ "learning_rate": 0.0006,
+ "loss": 4.195674896240234,
+ "step": 4707
+ },
+ {
+ "epoch": 65.39126637554585,
+ "grad_norm": 0.01787056028842926,
+ "learning_rate": 0.0006,
+ "loss": 4.3507914543151855,
+ "step": 4708
+ },
+ {
+ "epoch": 65.40524017467249,
+ "grad_norm": 0.015295136719942093,
+ "learning_rate": 0.0006,
+ "loss": 4.234372138977051,
+ "step": 4709
+ },
+ {
+ "epoch": 65.41921397379913,
+ "grad_norm": 0.014324234798550606,
+ "learning_rate": 0.0006,
+ "loss": 4.2907867431640625,
+ "step": 4710
+ },
+ {
+ "epoch": 65.43318777292576,
+ "grad_norm": 0.016090670600533485,
+ "learning_rate": 0.0006,
+ "loss": 4.177244186401367,
+ "step": 4711
+ },
+ {
+ "epoch": 65.44716157205241,
+ "grad_norm": 0.0174604132771492,
+ "learning_rate": 0.0006,
+ "loss": 4.335577964782715,
+ "step": 4712
+ },
+ {
+ "epoch": 65.46113537117904,
+ "grad_norm": 0.016731899231672287,
+ "learning_rate": 0.0006,
+ "loss": 4.30197811126709,
+ "step": 4713
+ },
+ {
+ "epoch": 65.47510917030567,
+ "grad_norm": 0.017423735931515694,
+ "learning_rate": 0.0006,
+ "loss": 4.3528852462768555,
+ "step": 4714
+ },
+ {
+ "epoch": 65.48908296943232,
+ "grad_norm": 0.01780976541340351,
+ "learning_rate": 0.0006,
+ "loss": 4.185036659240723,
+ "step": 4715
+ },
+ {
+ "epoch": 65.50305676855895,
+ "grad_norm": 0.019977200776338577,
+ "learning_rate": 0.0006,
+ "loss": 4.328766345977783,
+ "step": 4716
+ },
+ {
+ "epoch": 65.5170305676856,
+ "grad_norm": 0.020284287631511688,
+ "learning_rate": 0.0006,
+ "loss": 4.285494804382324,
+ "step": 4717
+ },
+ {
+ "epoch": 65.53100436681223,
+ "grad_norm": 0.019168002530932426,
+ "learning_rate": 0.0006,
+ "loss": 4.328560829162598,
+ "step": 4718
+ },
+ {
+ "epoch": 65.54497816593886,
+ "grad_norm": 0.019217776134610176,
+ "learning_rate": 0.0006,
+ "loss": 4.189998626708984,
+ "step": 4719
+ },
+ {
+ "epoch": 65.5589519650655,
+ "grad_norm": 0.018960488960146904,
+ "learning_rate": 0.0006,
+ "loss": 4.4995012283325195,
+ "step": 4720
+ },
+ {
+ "epoch": 65.57292576419214,
+ "grad_norm": 0.01941288262605667,
+ "learning_rate": 0.0006,
+ "loss": 4.26570987701416,
+ "step": 4721
+ },
+ {
+ "epoch": 65.58689956331878,
+ "grad_norm": 0.020263448357582092,
+ "learning_rate": 0.0006,
+ "loss": 4.411645889282227,
+ "step": 4722
+ },
+ {
+ "epoch": 65.60087336244541,
+ "grad_norm": 0.02144799567759037,
+ "learning_rate": 0.0006,
+ "loss": 4.163675308227539,
+ "step": 4723
+ },
+ {
+ "epoch": 65.61484716157206,
+ "grad_norm": 0.019790446385741234,
+ "learning_rate": 0.0006,
+ "loss": 4.2361063957214355,
+ "step": 4724
+ },
+ {
+ "epoch": 65.62882096069869,
+ "grad_norm": 0.01829805225133896,
+ "learning_rate": 0.0006,
+ "loss": 4.449810028076172,
+ "step": 4725
+ },
+ {
+ "epoch": 65.64279475982532,
+ "grad_norm": 0.01571105420589447,
+ "learning_rate": 0.0006,
+ "loss": 4.281406879425049,
+ "step": 4726
+ },
+ {
+ "epoch": 65.65676855895197,
+ "grad_norm": 0.016477147117257118,
+ "learning_rate": 0.0006,
+ "loss": 4.293408393859863,
+ "step": 4727
+ },
+ {
+ "epoch": 65.6707423580786,
+ "grad_norm": 0.016766032204031944,
+ "learning_rate": 0.0006,
+ "loss": 4.378252983093262,
+ "step": 4728
+ },
+ {
+ "epoch": 65.68471615720524,
+ "grad_norm": 0.0159933939576149,
+ "learning_rate": 0.0006,
+ "loss": 4.3065667152404785,
+ "step": 4729
+ },
+ {
+ "epoch": 65.69868995633188,
+ "grad_norm": 0.01658209227025509,
+ "learning_rate": 0.0006,
+ "loss": 4.174026966094971,
+ "step": 4730
+ },
+ {
+ "epoch": 65.71266375545852,
+ "grad_norm": 0.015081380493938923,
+ "learning_rate": 0.0006,
+ "loss": 4.360222339630127,
+ "step": 4731
+ },
+ {
+ "epoch": 65.72663755458515,
+ "grad_norm": 0.015110890381038189,
+ "learning_rate": 0.0006,
+ "loss": 4.218810558319092,
+ "step": 4732
+ },
+ {
+ "epoch": 65.74061135371178,
+ "grad_norm": 0.015300953760743141,
+ "learning_rate": 0.0006,
+ "loss": 4.331979274749756,
+ "step": 4733
+ },
+ {
+ "epoch": 65.75458515283843,
+ "grad_norm": 0.01715504191815853,
+ "learning_rate": 0.0006,
+ "loss": 4.349112033843994,
+ "step": 4734
+ },
+ {
+ "epoch": 65.76855895196506,
+ "grad_norm": 0.015870366245508194,
+ "learning_rate": 0.0006,
+ "loss": 4.329850196838379,
+ "step": 4735
+ },
+ {
+ "epoch": 65.78253275109171,
+ "grad_norm": 0.01541937980800867,
+ "learning_rate": 0.0006,
+ "loss": 4.3400163650512695,
+ "step": 4736
+ },
+ {
+ "epoch": 65.79650655021834,
+ "grad_norm": 0.01738942787051201,
+ "learning_rate": 0.0006,
+ "loss": 4.220798492431641,
+ "step": 4737
+ },
+ {
+ "epoch": 65.81048034934499,
+ "grad_norm": 0.01685990020632744,
+ "learning_rate": 0.0006,
+ "loss": 4.177238464355469,
+ "step": 4738
+ },
+ {
+ "epoch": 65.82445414847162,
+ "grad_norm": 0.016183465719223022,
+ "learning_rate": 0.0006,
+ "loss": 4.327216625213623,
+ "step": 4739
+ },
+ {
+ "epoch": 65.83842794759825,
+ "grad_norm": 0.01677694357931614,
+ "learning_rate": 0.0006,
+ "loss": 4.27766227722168,
+ "step": 4740
+ },
+ {
+ "epoch": 65.8524017467249,
+ "grad_norm": 0.01645066775381565,
+ "learning_rate": 0.0006,
+ "loss": 4.288347244262695,
+ "step": 4741
+ },
+ {
+ "epoch": 65.86637554585153,
+ "grad_norm": 0.018034635111689568,
+ "learning_rate": 0.0006,
+ "loss": 4.292458534240723,
+ "step": 4742
+ },
+ {
+ "epoch": 65.88034934497817,
+ "grad_norm": 0.017973346635699272,
+ "learning_rate": 0.0006,
+ "loss": 4.3117265701293945,
+ "step": 4743
+ },
+ {
+ "epoch": 65.8943231441048,
+ "grad_norm": 0.016605669632554054,
+ "learning_rate": 0.0006,
+ "loss": 4.280615329742432,
+ "step": 4744
+ },
+ {
+ "epoch": 65.90829694323143,
+ "grad_norm": 0.016298074275255203,
+ "learning_rate": 0.0006,
+ "loss": 4.301933288574219,
+ "step": 4745
+ },
+ {
+ "epoch": 65.92227074235808,
+ "grad_norm": 0.015383216552436352,
+ "learning_rate": 0.0006,
+ "loss": 4.352383136749268,
+ "step": 4746
+ },
+ {
+ "epoch": 65.93624454148471,
+ "grad_norm": 0.016222182661294937,
+ "learning_rate": 0.0006,
+ "loss": 4.305725574493408,
+ "step": 4747
+ },
+ {
+ "epoch": 65.95021834061136,
+ "grad_norm": 0.01691329851746559,
+ "learning_rate": 0.0006,
+ "loss": 4.29502010345459,
+ "step": 4748
+ },
+ {
+ "epoch": 65.96419213973799,
+ "grad_norm": 0.017637677490711212,
+ "learning_rate": 0.0006,
+ "loss": 4.3133039474487305,
+ "step": 4749
+ },
+ {
+ "epoch": 65.97816593886463,
+ "grad_norm": 0.021362047642469406,
+ "learning_rate": 0.0006,
+ "loss": 4.458834648132324,
+ "step": 4750
+ },
+ {
+ "epoch": 65.99213973799127,
+ "grad_norm": 0.023553011938929558,
+ "learning_rate": 0.0006,
+ "loss": 4.298727989196777,
+ "step": 4751
+ },
+ {
+ "epoch": 66.0,
+ "grad_norm": 0.02154950052499771,
+ "learning_rate": 0.0006,
+ "loss": 4.403392791748047,
+ "step": 4752
+ },
+ {
+ "epoch": 66.0,
+ "eval_loss": 4.724340438842773,
+ "eval_runtime": 56.9555,
+ "eval_samples_per_second": 42.876,
+ "eval_steps_per_second": 1.352,
+ "step": 4752
+ },
+ {
+ "epoch": 66.01397379912663,
+ "grad_norm": 0.01916784606873989,
+ "learning_rate": 0.0006,
+ "loss": 4.22931432723999,
+ "step": 4753
+ },
+ {
+ "epoch": 66.02794759825328,
+ "grad_norm": 0.017492054030299187,
+ "learning_rate": 0.0006,
+ "loss": 4.271522521972656,
+ "step": 4754
+ },
+ {
+ "epoch": 66.04192139737991,
+ "grad_norm": 0.018408095464110374,
+ "learning_rate": 0.0006,
+ "loss": 4.337705135345459,
+ "step": 4755
+ },
+ {
+ "epoch": 66.05589519650655,
+ "grad_norm": 0.019220108166337013,
+ "learning_rate": 0.0006,
+ "loss": 4.280340194702148,
+ "step": 4756
+ },
+ {
+ "epoch": 66.06986899563319,
+ "grad_norm": 0.01773735322058201,
+ "learning_rate": 0.0006,
+ "loss": 4.390201568603516,
+ "step": 4757
+ },
+ {
+ "epoch": 66.08384279475983,
+ "grad_norm": 0.01679125428199768,
+ "learning_rate": 0.0006,
+ "loss": 4.299282073974609,
+ "step": 4758
+ },
+ {
+ "epoch": 66.09781659388646,
+ "grad_norm": 0.01479059923440218,
+ "learning_rate": 0.0006,
+ "loss": 4.472630500793457,
+ "step": 4759
+ },
+ {
+ "epoch": 66.1117903930131,
+ "grad_norm": 0.015130104497075081,
+ "learning_rate": 0.0006,
+ "loss": 4.301194190979004,
+ "step": 4760
+ },
+ {
+ "epoch": 66.12576419213974,
+ "grad_norm": 0.0158979594707489,
+ "learning_rate": 0.0006,
+ "loss": 4.375751972198486,
+ "step": 4761
+ },
+ {
+ "epoch": 66.13973799126637,
+ "grad_norm": 0.014432408846914768,
+ "learning_rate": 0.0006,
+ "loss": 4.279788017272949,
+ "step": 4762
+ },
+ {
+ "epoch": 66.15371179039302,
+ "grad_norm": 0.013026784174144268,
+ "learning_rate": 0.0006,
+ "loss": 4.255035400390625,
+ "step": 4763
+ },
+ {
+ "epoch": 66.16768558951965,
+ "grad_norm": 0.014422653242945671,
+ "learning_rate": 0.0006,
+ "loss": 4.261832237243652,
+ "step": 4764
+ },
+ {
+ "epoch": 66.18165938864628,
+ "grad_norm": 0.014613019302487373,
+ "learning_rate": 0.0006,
+ "loss": 4.382184028625488,
+ "step": 4765
+ },
+ {
+ "epoch": 66.19563318777293,
+ "grad_norm": 0.014989894814789295,
+ "learning_rate": 0.0006,
+ "loss": 4.40528678894043,
+ "step": 4766
+ },
+ {
+ "epoch": 66.20960698689956,
+ "grad_norm": 0.018190350383520126,
+ "learning_rate": 0.0006,
+ "loss": 4.302414894104004,
+ "step": 4767
+ },
+ {
+ "epoch": 66.2235807860262,
+ "grad_norm": 0.020887890830636024,
+ "learning_rate": 0.0006,
+ "loss": 4.289920806884766,
+ "step": 4768
+ },
+ {
+ "epoch": 66.23755458515284,
+ "grad_norm": 0.02376876398921013,
+ "learning_rate": 0.0006,
+ "loss": 4.297855377197266,
+ "step": 4769
+ },
+ {
+ "epoch": 66.25152838427948,
+ "grad_norm": 0.02152269519865513,
+ "learning_rate": 0.0006,
+ "loss": 4.1972832679748535,
+ "step": 4770
+ },
+ {
+ "epoch": 66.26550218340611,
+ "grad_norm": 0.015572802163660526,
+ "learning_rate": 0.0006,
+ "loss": 4.307487487792969,
+ "step": 4771
+ },
+ {
+ "epoch": 66.27947598253274,
+ "grad_norm": 0.015228749252855778,
+ "learning_rate": 0.0006,
+ "loss": 4.103250503540039,
+ "step": 4772
+ },
+ {
+ "epoch": 66.29344978165939,
+ "grad_norm": 0.016925504431128502,
+ "learning_rate": 0.0006,
+ "loss": 4.247461318969727,
+ "step": 4773
+ },
+ {
+ "epoch": 66.30742358078602,
+ "grad_norm": 0.01613711006939411,
+ "learning_rate": 0.0006,
+ "loss": 4.4699273109436035,
+ "step": 4774
+ },
+ {
+ "epoch": 66.32139737991267,
+ "grad_norm": 0.018332140520215034,
+ "learning_rate": 0.0006,
+ "loss": 4.22969388961792,
+ "step": 4775
+ },
+ {
+ "epoch": 66.3353711790393,
+ "grad_norm": 0.019575074315071106,
+ "learning_rate": 0.0006,
+ "loss": 4.29397439956665,
+ "step": 4776
+ },
+ {
+ "epoch": 66.34934497816595,
+ "grad_norm": 0.018177419900894165,
+ "learning_rate": 0.0006,
+ "loss": 4.221899509429932,
+ "step": 4777
+ },
+ {
+ "epoch": 66.36331877729258,
+ "grad_norm": 0.014292889274656773,
+ "learning_rate": 0.0006,
+ "loss": 4.325457572937012,
+ "step": 4778
+ },
+ {
+ "epoch": 66.37729257641921,
+ "grad_norm": 0.015098796226084232,
+ "learning_rate": 0.0006,
+ "loss": 4.364837169647217,
+ "step": 4779
+ },
+ {
+ "epoch": 66.39126637554585,
+ "grad_norm": 0.013992635533213615,
+ "learning_rate": 0.0006,
+ "loss": 4.299082279205322,
+ "step": 4780
+ },
+ {
+ "epoch": 66.40524017467249,
+ "grad_norm": 0.01504511572420597,
+ "learning_rate": 0.0006,
+ "loss": 4.230877876281738,
+ "step": 4781
+ },
+ {
+ "epoch": 66.41921397379913,
+ "grad_norm": 0.014875639230012894,
+ "learning_rate": 0.0006,
+ "loss": 4.368405818939209,
+ "step": 4782
+ },
+ {
+ "epoch": 66.43318777292576,
+ "grad_norm": 0.014306887984275818,
+ "learning_rate": 0.0006,
+ "loss": 4.389184474945068,
+ "step": 4783
+ },
+ {
+ "epoch": 66.44716157205241,
+ "grad_norm": 0.012959563173353672,
+ "learning_rate": 0.0006,
+ "loss": 4.37934684753418,
+ "step": 4784
+ },
+ {
+ "epoch": 66.46113537117904,
+ "grad_norm": 0.013206703588366508,
+ "learning_rate": 0.0006,
+ "loss": 4.442799091339111,
+ "step": 4785
+ },
+ {
+ "epoch": 66.47510917030567,
+ "grad_norm": 0.014848909340798855,
+ "learning_rate": 0.0006,
+ "loss": 4.203481674194336,
+ "step": 4786
+ },
+ {
+ "epoch": 66.48908296943232,
+ "grad_norm": 0.015637682750821114,
+ "learning_rate": 0.0006,
+ "loss": 4.27552604675293,
+ "step": 4787
+ },
+ {
+ "epoch": 66.50305676855895,
+ "grad_norm": 0.016017816960811615,
+ "learning_rate": 0.0006,
+ "loss": 4.296402931213379,
+ "step": 4788
+ },
+ {
+ "epoch": 66.5170305676856,
+ "grad_norm": 0.01577814109623432,
+ "learning_rate": 0.0006,
+ "loss": 4.301733016967773,
+ "step": 4789
+ },
+ {
+ "epoch": 66.53100436681223,
+ "grad_norm": 0.015293175354599953,
+ "learning_rate": 0.0006,
+ "loss": 4.369877815246582,
+ "step": 4790
+ },
+ {
+ "epoch": 66.54497816593886,
+ "grad_norm": 0.014262278564274311,
+ "learning_rate": 0.0006,
+ "loss": 4.3314738273620605,
+ "step": 4791
+ },
+ {
+ "epoch": 66.5589519650655,
+ "grad_norm": 0.015123482793569565,
+ "learning_rate": 0.0006,
+ "loss": 4.313591957092285,
+ "step": 4792
+ },
+ {
+ "epoch": 66.57292576419214,
+ "grad_norm": 0.015458766371011734,
+ "learning_rate": 0.0006,
+ "loss": 4.3045854568481445,
+ "step": 4793
+ },
+ {
+ "epoch": 66.58689956331878,
+ "grad_norm": 0.017217233777046204,
+ "learning_rate": 0.0006,
+ "loss": 4.269944190979004,
+ "step": 4794
+ },
+ {
+ "epoch": 66.60087336244541,
+ "grad_norm": 0.01962282694876194,
+ "learning_rate": 0.0006,
+ "loss": 4.271735191345215,
+ "step": 4795
+ },
+ {
+ "epoch": 66.61484716157206,
+ "grad_norm": 0.01917066052556038,
+ "learning_rate": 0.0006,
+ "loss": 4.288168430328369,
+ "step": 4796
+ },
+ {
+ "epoch": 66.62882096069869,
+ "grad_norm": 0.018211524933576584,
+ "learning_rate": 0.0006,
+ "loss": 4.280660629272461,
+ "step": 4797
+ },
+ {
+ "epoch": 66.64279475982532,
+ "grad_norm": 0.01936524175107479,
+ "learning_rate": 0.0006,
+ "loss": 4.3710551261901855,
+ "step": 4798
+ },
+ {
+ "epoch": 66.65676855895197,
+ "grad_norm": 0.01882351003587246,
+ "learning_rate": 0.0006,
+ "loss": 4.342750549316406,
+ "step": 4799
+ },
+ {
+ "epoch": 66.6707423580786,
+ "grad_norm": 0.017149122431874275,
+ "learning_rate": 0.0006,
+ "loss": 4.32134485244751,
+ "step": 4800
+ },
+ {
+ "epoch": 66.68471615720524,
+ "grad_norm": 0.01699327677488327,
+ "learning_rate": 0.0006,
+ "loss": 4.248893737792969,
+ "step": 4801
+ },
+ {
+ "epoch": 66.69868995633188,
+ "grad_norm": 0.016873255372047424,
+ "learning_rate": 0.0006,
+ "loss": 4.277350425720215,
+ "step": 4802
+ },
+ {
+ "epoch": 66.71266375545852,
+ "grad_norm": 0.016592837870121002,
+ "learning_rate": 0.0006,
+ "loss": 4.221665859222412,
+ "step": 4803
+ },
+ {
+ "epoch": 66.72663755458515,
+ "grad_norm": 0.015887131914496422,
+ "learning_rate": 0.0006,
+ "loss": 4.328040599822998,
+ "step": 4804
+ },
+ {
+ "epoch": 66.74061135371178,
+ "grad_norm": 0.014057286083698273,
+ "learning_rate": 0.0006,
+ "loss": 4.3294548988342285,
+ "step": 4805
+ },
+ {
+ "epoch": 66.75458515283843,
+ "grad_norm": 0.013135905377566814,
+ "learning_rate": 0.0006,
+ "loss": 4.254383563995361,
+ "step": 4806
+ },
+ {
+ "epoch": 66.76855895196506,
+ "grad_norm": 0.013632557354867458,
+ "learning_rate": 0.0006,
+ "loss": 4.342815399169922,
+ "step": 4807
+ },
+ {
+ "epoch": 66.78253275109171,
+ "grad_norm": 0.015512688085436821,
+ "learning_rate": 0.0006,
+ "loss": 4.297835350036621,
+ "step": 4808
+ },
+ {
+ "epoch": 66.79650655021834,
+ "grad_norm": 0.01621832884848118,
+ "learning_rate": 0.0006,
+ "loss": 4.314773082733154,
+ "step": 4809
+ },
+ {
+ "epoch": 66.81048034934499,
+ "grad_norm": 0.016196032986044884,
+ "learning_rate": 0.0006,
+ "loss": 4.29209566116333,
+ "step": 4810
+ },
+ {
+ "epoch": 66.82445414847162,
+ "grad_norm": 0.015463353134691715,
+ "learning_rate": 0.0006,
+ "loss": 4.262795448303223,
+ "step": 4811
+ },
+ {
+ "epoch": 66.83842794759825,
+ "grad_norm": 0.016143787652254105,
+ "learning_rate": 0.0006,
+ "loss": 4.274284362792969,
+ "step": 4812
+ },
+ {
+ "epoch": 66.8524017467249,
+ "grad_norm": 0.01610061153769493,
+ "learning_rate": 0.0006,
+ "loss": 4.337160587310791,
+ "step": 4813
+ },
+ {
+ "epoch": 66.86637554585153,
+ "grad_norm": 0.01582838036119938,
+ "learning_rate": 0.0006,
+ "loss": 4.192723274230957,
+ "step": 4814
+ },
+ {
+ "epoch": 66.88034934497817,
+ "grad_norm": 0.015081540681421757,
+ "learning_rate": 0.0006,
+ "loss": 4.370708465576172,
+ "step": 4815
+ },
+ {
+ "epoch": 66.8943231441048,
+ "grad_norm": 0.015929723158478737,
+ "learning_rate": 0.0006,
+ "loss": 4.340244293212891,
+ "step": 4816
+ },
+ {
+ "epoch": 66.90829694323143,
+ "grad_norm": 0.01784711331129074,
+ "learning_rate": 0.0006,
+ "loss": 4.260406970977783,
+ "step": 4817
+ },
+ {
+ "epoch": 66.92227074235808,
+ "grad_norm": 0.017434703186154366,
+ "learning_rate": 0.0006,
+ "loss": 4.286900043487549,
+ "step": 4818
+ },
+ {
+ "epoch": 66.93624454148471,
+ "grad_norm": 0.016262540593743324,
+ "learning_rate": 0.0006,
+ "loss": 4.292142391204834,
+ "step": 4819
+ },
+ {
+ "epoch": 66.95021834061136,
+ "grad_norm": 0.016122829169034958,
+ "learning_rate": 0.0006,
+ "loss": 4.291485786437988,
+ "step": 4820
+ },
+ {
+ "epoch": 66.96419213973799,
+ "grad_norm": 0.014361178502440453,
+ "learning_rate": 0.0006,
+ "loss": 4.275425910949707,
+ "step": 4821
+ },
+ {
+ "epoch": 66.97816593886463,
+ "grad_norm": 0.01373987551778555,
+ "learning_rate": 0.0006,
+ "loss": 4.223120212554932,
+ "step": 4822
+ },
+ {
+ "epoch": 66.99213973799127,
+ "grad_norm": 0.015687376260757446,
+ "learning_rate": 0.0006,
+ "loss": 4.355381011962891,
+ "step": 4823
+ },
+ {
+ "epoch": 67.0,
+ "grad_norm": 0.01998465321958065,
+ "learning_rate": 0.0006,
+ "loss": 4.264053821563721,
+ "step": 4824
+ },
+ {
+ "epoch": 67.0,
+ "eval_loss": 4.662398338317871,
+ "eval_runtime": 56.7928,
+ "eval_samples_per_second": 42.998,
+ "eval_steps_per_second": 1.356,
+ "step": 4824
+ },
+ {
+ "epoch": 67.01397379912663,
+ "grad_norm": 0.021736836060881615,
+ "learning_rate": 0.0006,
+ "loss": 4.243354797363281,
+ "step": 4825
+ },
+ {
+ "epoch": 67.02794759825328,
+ "grad_norm": 0.021476037800312042,
+ "learning_rate": 0.0006,
+ "loss": 4.218207836151123,
+ "step": 4826
+ },
+ {
+ "epoch": 67.04192139737991,
+ "grad_norm": 0.017188388854265213,
+ "learning_rate": 0.0006,
+ "loss": 4.18048095703125,
+ "step": 4827
+ },
+ {
+ "epoch": 67.05589519650655,
+ "grad_norm": 0.017293719574809074,
+ "learning_rate": 0.0006,
+ "loss": 4.239145278930664,
+ "step": 4828
+ },
+ {
+ "epoch": 67.06986899563319,
+ "grad_norm": 0.017383819445967674,
+ "learning_rate": 0.0006,
+ "loss": 4.114765167236328,
+ "step": 4829
+ },
+ {
+ "epoch": 67.08384279475983,
+ "grad_norm": 0.01651529036462307,
+ "learning_rate": 0.0006,
+ "loss": 4.351023197174072,
+ "step": 4830
+ },
+ {
+ "epoch": 67.09781659388646,
+ "grad_norm": 0.015893027186393738,
+ "learning_rate": 0.0006,
+ "loss": 4.261432647705078,
+ "step": 4831
+ },
+ {
+ "epoch": 67.1117903930131,
+ "grad_norm": 0.013381517492234707,
+ "learning_rate": 0.0006,
+ "loss": 4.172064781188965,
+ "step": 4832
+ },
+ {
+ "epoch": 67.12576419213974,
+ "grad_norm": 0.014642654918134212,
+ "learning_rate": 0.0006,
+ "loss": 4.356171131134033,
+ "step": 4833
+ },
+ {
+ "epoch": 67.13973799126637,
+ "grad_norm": 0.01425282284617424,
+ "learning_rate": 0.0006,
+ "loss": 4.303495407104492,
+ "step": 4834
+ },
+ {
+ "epoch": 67.15371179039302,
+ "grad_norm": 0.014574599452316761,
+ "learning_rate": 0.0006,
+ "loss": 4.2850847244262695,
+ "step": 4835
+ },
+ {
+ "epoch": 67.16768558951965,
+ "grad_norm": 0.012676282785832882,
+ "learning_rate": 0.0006,
+ "loss": 4.297845840454102,
+ "step": 4836
+ },
+ {
+ "epoch": 67.18165938864628,
+ "grad_norm": 0.013948136940598488,
+ "learning_rate": 0.0006,
+ "loss": 4.324530601501465,
+ "step": 4837
+ },
+ {
+ "epoch": 67.19563318777293,
+ "grad_norm": 0.014195773750543594,
+ "learning_rate": 0.0006,
+ "loss": 4.31236457824707,
+ "step": 4838
+ },
+ {
+ "epoch": 67.20960698689956,
+ "grad_norm": 0.013631382025778294,
+ "learning_rate": 0.0006,
+ "loss": 4.086362838745117,
+ "step": 4839
+ },
+ {
+ "epoch": 67.2235807860262,
+ "grad_norm": 0.012798587791621685,
+ "learning_rate": 0.0006,
+ "loss": 4.328471660614014,
+ "step": 4840
+ },
+ {
+ "epoch": 67.23755458515284,
+ "grad_norm": 0.014443321153521538,
+ "learning_rate": 0.0006,
+ "loss": 4.278019905090332,
+ "step": 4841
+ },
+ {
+ "epoch": 67.25152838427948,
+ "grad_norm": 0.015355420298874378,
+ "learning_rate": 0.0006,
+ "loss": 4.333611011505127,
+ "step": 4842
+ },
+ {
+ "epoch": 67.26550218340611,
+ "grad_norm": 0.016487296670675278,
+ "learning_rate": 0.0006,
+ "loss": 4.351336479187012,
+ "step": 4843
+ },
+ {
+ "epoch": 67.27947598253274,
+ "grad_norm": 0.017422541975975037,
+ "learning_rate": 0.0006,
+ "loss": 4.386096477508545,
+ "step": 4844
+ },
+ {
+ "epoch": 67.29344978165939,
+ "grad_norm": 0.015790918841958046,
+ "learning_rate": 0.0006,
+ "loss": 4.301377773284912,
+ "step": 4845
+ },
+ {
+ "epoch": 67.30742358078602,
+ "grad_norm": 0.016427364200353622,
+ "learning_rate": 0.0006,
+ "loss": 4.2838945388793945,
+ "step": 4846
+ },
+ {
+ "epoch": 67.32139737991267,
+ "grad_norm": 0.017050761729478836,
+ "learning_rate": 0.0006,
+ "loss": 4.282417297363281,
+ "step": 4847
+ },
+ {
+ "epoch": 67.3353711790393,
+ "grad_norm": 0.015006253495812416,
+ "learning_rate": 0.0006,
+ "loss": 4.352236270904541,
+ "step": 4848
+ },
+ {
+ "epoch": 67.34934497816595,
+ "grad_norm": 0.012996627949178219,
+ "learning_rate": 0.0006,
+ "loss": 4.309242248535156,
+ "step": 4849
+ },
+ {
+ "epoch": 67.36331877729258,
+ "grad_norm": 0.01367615070194006,
+ "learning_rate": 0.0006,
+ "loss": 4.211801528930664,
+ "step": 4850
+ },
+ {
+ "epoch": 67.37729257641921,
+ "grad_norm": 0.014775963500142097,
+ "learning_rate": 0.0006,
+ "loss": 4.360859394073486,
+ "step": 4851
+ },
+ {
+ "epoch": 67.39126637554585,
+ "grad_norm": 0.015854304656386375,
+ "learning_rate": 0.0006,
+ "loss": 4.211830139160156,
+ "step": 4852
+ },
+ {
+ "epoch": 67.40524017467249,
+ "grad_norm": 0.016888894140720367,
+ "learning_rate": 0.0006,
+ "loss": 4.216549873352051,
+ "step": 4853
+ },
+ {
+ "epoch": 67.41921397379913,
+ "grad_norm": 0.0170894768089056,
+ "learning_rate": 0.0006,
+ "loss": 4.263486385345459,
+ "step": 4854
+ },
+ {
+ "epoch": 67.43318777292576,
+ "grad_norm": 0.015829868614673615,
+ "learning_rate": 0.0006,
+ "loss": 4.360196113586426,
+ "step": 4855
+ },
+ {
+ "epoch": 67.44716157205241,
+ "grad_norm": 0.015220202505588531,
+ "learning_rate": 0.0006,
+ "loss": 4.243794918060303,
+ "step": 4856
+ },
+ {
+ "epoch": 67.46113537117904,
+ "grad_norm": 0.01413388829678297,
+ "learning_rate": 0.0006,
+ "loss": 4.286624908447266,
+ "step": 4857
+ },
+ {
+ "epoch": 67.47510917030567,
+ "grad_norm": 0.01426626369357109,
+ "learning_rate": 0.0006,
+ "loss": 4.257316589355469,
+ "step": 4858
+ },
+ {
+ "epoch": 67.48908296943232,
+ "grad_norm": 0.015957845374941826,
+ "learning_rate": 0.0006,
+ "loss": 4.300443649291992,
+ "step": 4859
+ },
+ {
+ "epoch": 67.50305676855895,
+ "grad_norm": 0.015499051660299301,
+ "learning_rate": 0.0006,
+ "loss": 4.278162002563477,
+ "step": 4860
+ },
+ {
+ "epoch": 67.5170305676856,
+ "grad_norm": 0.015145624987781048,
+ "learning_rate": 0.0006,
+ "loss": 4.267950534820557,
+ "step": 4861
+ },
+ {
+ "epoch": 67.53100436681223,
+ "grad_norm": 0.014610803686082363,
+ "learning_rate": 0.0006,
+ "loss": 4.247499942779541,
+ "step": 4862
+ },
+ {
+ "epoch": 67.54497816593886,
+ "grad_norm": 0.014860909432172775,
+ "learning_rate": 0.0006,
+ "loss": 4.263200759887695,
+ "step": 4863
+ },
+ {
+ "epoch": 67.5589519650655,
+ "grad_norm": 0.01718984916806221,
+ "learning_rate": 0.0006,
+ "loss": 4.290019989013672,
+ "step": 4864
+ },
+ {
+ "epoch": 67.57292576419214,
+ "grad_norm": 0.01558963768184185,
+ "learning_rate": 0.0006,
+ "loss": 4.223816871643066,
+ "step": 4865
+ },
+ {
+ "epoch": 67.58689956331878,
+ "grad_norm": 0.014927188865840435,
+ "learning_rate": 0.0006,
+ "loss": 4.363586902618408,
+ "step": 4866
+ },
+ {
+ "epoch": 67.60087336244541,
+ "grad_norm": 0.01573595404624939,
+ "learning_rate": 0.0006,
+ "loss": 4.254765510559082,
+ "step": 4867
+ },
+ {
+ "epoch": 67.61484716157206,
+ "grad_norm": 0.015618929639458656,
+ "learning_rate": 0.0006,
+ "loss": 4.25072717666626,
+ "step": 4868
+ },
+ {
+ "epoch": 67.62882096069869,
+ "grad_norm": 0.01605590060353279,
+ "learning_rate": 0.0006,
+ "loss": 4.25208854675293,
+ "step": 4869
+ },
+ {
+ "epoch": 67.64279475982532,
+ "grad_norm": 0.01804601214826107,
+ "learning_rate": 0.0006,
+ "loss": 4.351354122161865,
+ "step": 4870
+ },
+ {
+ "epoch": 67.65676855895197,
+ "grad_norm": 0.020595047622919083,
+ "learning_rate": 0.0006,
+ "loss": 4.193732738494873,
+ "step": 4871
+ },
+ {
+ "epoch": 67.6707423580786,
+ "grad_norm": 0.023037966340780258,
+ "learning_rate": 0.0006,
+ "loss": 4.27754020690918,
+ "step": 4872
+ },
+ {
+ "epoch": 67.68471615720524,
+ "grad_norm": 0.020680895075201988,
+ "learning_rate": 0.0006,
+ "loss": 4.314730167388916,
+ "step": 4873
+ },
+ {
+ "epoch": 67.69868995633188,
+ "grad_norm": 0.01887434534728527,
+ "learning_rate": 0.0006,
+ "loss": 4.252305030822754,
+ "step": 4874
+ },
+ {
+ "epoch": 67.71266375545852,
+ "grad_norm": 0.016333596780896187,
+ "learning_rate": 0.0006,
+ "loss": 4.146123886108398,
+ "step": 4875
+ },
+ {
+ "epoch": 67.72663755458515,
+ "grad_norm": 0.015688810497522354,
+ "learning_rate": 0.0006,
+ "loss": 4.2814154624938965,
+ "step": 4876
+ },
+ {
+ "epoch": 67.74061135371178,
+ "grad_norm": 0.016270726919174194,
+ "learning_rate": 0.0006,
+ "loss": 4.323200225830078,
+ "step": 4877
+ },
+ {
+ "epoch": 67.75458515283843,
+ "grad_norm": 0.016824988648295403,
+ "learning_rate": 0.0006,
+ "loss": 4.269501686096191,
+ "step": 4878
+ },
+ {
+ "epoch": 67.76855895196506,
+ "grad_norm": 0.01730562001466751,
+ "learning_rate": 0.0006,
+ "loss": 4.187993049621582,
+ "step": 4879
+ },
+ {
+ "epoch": 67.78253275109171,
+ "grad_norm": 0.01546655036509037,
+ "learning_rate": 0.0006,
+ "loss": 4.340415000915527,
+ "step": 4880
+ },
+ {
+ "epoch": 67.79650655021834,
+ "grad_norm": 0.014173958450555801,
+ "learning_rate": 0.0006,
+ "loss": 4.217419624328613,
+ "step": 4881
+ },
+ {
+ "epoch": 67.81048034934499,
+ "grad_norm": 0.016810474917292595,
+ "learning_rate": 0.0006,
+ "loss": 4.352302074432373,
+ "step": 4882
+ },
+ {
+ "epoch": 67.82445414847162,
+ "grad_norm": 0.0175373163074255,
+ "learning_rate": 0.0006,
+ "loss": 4.20377254486084,
+ "step": 4883
+ },
+ {
+ "epoch": 67.83842794759825,
+ "grad_norm": 0.015734676271677017,
+ "learning_rate": 0.0006,
+ "loss": 4.147495746612549,
+ "step": 4884
+ },
+ {
+ "epoch": 67.8524017467249,
+ "grad_norm": 0.015091817826032639,
+ "learning_rate": 0.0006,
+ "loss": 4.272252082824707,
+ "step": 4885
+ },
+ {
+ "epoch": 67.86637554585153,
+ "grad_norm": 0.015000219456851482,
+ "learning_rate": 0.0006,
+ "loss": 4.302773952484131,
+ "step": 4886
+ },
+ {
+ "epoch": 67.88034934497817,
+ "grad_norm": 0.013827184215188026,
+ "learning_rate": 0.0006,
+ "loss": 4.337242126464844,
+ "step": 4887
+ },
+ {
+ "epoch": 67.8943231441048,
+ "grad_norm": 0.015308346599340439,
+ "learning_rate": 0.0006,
+ "loss": 4.140895366668701,
+ "step": 4888
+ },
+ {
+ "epoch": 67.90829694323143,
+ "grad_norm": 0.015356178395450115,
+ "learning_rate": 0.0006,
+ "loss": 4.312541484832764,
+ "step": 4889
+ },
+ {
+ "epoch": 67.92227074235808,
+ "grad_norm": 0.0137353939935565,
+ "learning_rate": 0.0006,
+ "loss": 4.173184394836426,
+ "step": 4890
+ },
+ {
+ "epoch": 67.93624454148471,
+ "grad_norm": 0.01332368515431881,
+ "learning_rate": 0.0006,
+ "loss": 4.191199779510498,
+ "step": 4891
+ },
+ {
+ "epoch": 67.95021834061136,
+ "grad_norm": 0.01488130260258913,
+ "learning_rate": 0.0006,
+ "loss": 4.331752300262451,
+ "step": 4892
+ },
+ {
+ "epoch": 67.96419213973799,
+ "grad_norm": 0.015449733473360538,
+ "learning_rate": 0.0006,
+ "loss": 4.2385454177856445,
+ "step": 4893
+ },
+ {
+ "epoch": 67.97816593886463,
+ "grad_norm": 0.01556103490293026,
+ "learning_rate": 0.0006,
+ "loss": 4.262353897094727,
+ "step": 4894
+ },
+ {
+ "epoch": 67.99213973799127,
+ "grad_norm": 0.014613849110901356,
+ "learning_rate": 0.0006,
+ "loss": 4.223945617675781,
+ "step": 4895
+ },
+ {
+ "epoch": 68.0,
+ "grad_norm": 0.015323462896049023,
+ "learning_rate": 0.0006,
+ "loss": 4.317168712615967,
+ "step": 4896
+ },
+ {
+ "epoch": 68.0,
+ "eval_loss": 4.671603679656982,
+ "eval_runtime": 56.7317,
+ "eval_samples_per_second": 43.045,
+ "eval_steps_per_second": 1.357,
+ "step": 4896
+ },
+ {
+ "epoch": 68.01397379912663,
+ "grad_norm": 0.01910698413848877,
+ "learning_rate": 0.0006,
+ "loss": 4.206167221069336,
+ "step": 4897
+ },
+ {
+ "epoch": 68.02794759825328,
+ "grad_norm": 0.025112485513091087,
+ "learning_rate": 0.0006,
+ "loss": 4.384706974029541,
+ "step": 4898
+ },
+ {
+ "epoch": 68.04192139737991,
+ "grad_norm": 0.02618720941245556,
+ "learning_rate": 0.0006,
+ "loss": 4.321046829223633,
+ "step": 4899
+ },
+ {
+ "epoch": 68.05589519650655,
+ "grad_norm": 0.020853538066148758,
+ "learning_rate": 0.0006,
+ "loss": 4.217638969421387,
+ "step": 4900
+ },
+ {
+ "epoch": 68.06986899563319,
+ "grad_norm": 0.020025717094540596,
+ "learning_rate": 0.0006,
+ "loss": 4.306824684143066,
+ "step": 4901
+ },
+ {
+ "epoch": 68.08384279475983,
+ "grad_norm": 0.023868650197982788,
+ "learning_rate": 0.0006,
+ "loss": 4.327455997467041,
+ "step": 4902
+ },
+ {
+ "epoch": 68.09781659388646,
+ "grad_norm": 0.022279081866145134,
+ "learning_rate": 0.0006,
+ "loss": 4.400699615478516,
+ "step": 4903
+ },
+ {
+ "epoch": 68.1117903930131,
+ "grad_norm": 0.018965687602758408,
+ "learning_rate": 0.0006,
+ "loss": 4.19097900390625,
+ "step": 4904
+ },
+ {
+ "epoch": 68.12576419213974,
+ "grad_norm": 0.01936378888785839,
+ "learning_rate": 0.0006,
+ "loss": 4.206711769104004,
+ "step": 4905
+ },
+ {
+ "epoch": 68.13973799126637,
+ "grad_norm": 0.01811872608959675,
+ "learning_rate": 0.0006,
+ "loss": 4.204622268676758,
+ "step": 4906
+ },
+ {
+ "epoch": 68.15371179039302,
+ "grad_norm": 0.018530402332544327,
+ "learning_rate": 0.0006,
+ "loss": 4.289602279663086,
+ "step": 4907
+ },
+ {
+ "epoch": 68.16768558951965,
+ "grad_norm": 0.01755962148308754,
+ "learning_rate": 0.0006,
+ "loss": 4.317145347595215,
+ "step": 4908
+ },
+ {
+ "epoch": 68.18165938864628,
+ "grad_norm": 0.020413702353835106,
+ "learning_rate": 0.0006,
+ "loss": 4.2882843017578125,
+ "step": 4909
+ },
+ {
+ "epoch": 68.19563318777293,
+ "grad_norm": 0.020064299926161766,
+ "learning_rate": 0.0006,
+ "loss": 4.248260021209717,
+ "step": 4910
+ },
+ {
+ "epoch": 68.20960698689956,
+ "grad_norm": 0.017415888607501984,
+ "learning_rate": 0.0006,
+ "loss": 4.370295524597168,
+ "step": 4911
+ },
+ {
+ "epoch": 68.2235807860262,
+ "grad_norm": 0.015005858615040779,
+ "learning_rate": 0.0006,
+ "loss": 4.319567680358887,
+ "step": 4912
+ },
+ {
+ "epoch": 68.23755458515284,
+ "grad_norm": 0.016349777579307556,
+ "learning_rate": 0.0006,
+ "loss": 4.301776885986328,
+ "step": 4913
+ },
+ {
+ "epoch": 68.25152838427948,
+ "grad_norm": 0.015779560431838036,
+ "learning_rate": 0.0006,
+ "loss": 4.271171569824219,
+ "step": 4914
+ },
+ {
+ "epoch": 68.26550218340611,
+ "grad_norm": 0.013237644918262959,
+ "learning_rate": 0.0006,
+ "loss": 4.324908256530762,
+ "step": 4915
+ },
+ {
+ "epoch": 68.27947598253274,
+ "grad_norm": 0.014076225459575653,
+ "learning_rate": 0.0006,
+ "loss": 4.177175998687744,
+ "step": 4916
+ },
+ {
+ "epoch": 68.29344978165939,
+ "grad_norm": 0.013906744308769703,
+ "learning_rate": 0.0006,
+ "loss": 4.11588191986084,
+ "step": 4917
+ },
+ {
+ "epoch": 68.30742358078602,
+ "grad_norm": 0.013939064927399158,
+ "learning_rate": 0.0006,
+ "loss": 4.183963775634766,
+ "step": 4918
+ },
+ {
+ "epoch": 68.32139737991267,
+ "grad_norm": 0.014572170563042164,
+ "learning_rate": 0.0006,
+ "loss": 4.19586181640625,
+ "step": 4919
+ },
+ {
+ "epoch": 68.3353711790393,
+ "grad_norm": 0.014085350558161736,
+ "learning_rate": 0.0006,
+ "loss": 4.170892238616943,
+ "step": 4920
+ },
+ {
+ "epoch": 68.34934497816595,
+ "grad_norm": 0.014357863925397396,
+ "learning_rate": 0.0006,
+ "loss": 4.321155548095703,
+ "step": 4921
+ },
+ {
+ "epoch": 68.36331877729258,
+ "grad_norm": 0.01461675763130188,
+ "learning_rate": 0.0006,
+ "loss": 4.293334007263184,
+ "step": 4922
+ },
+ {
+ "epoch": 68.37729257641921,
+ "grad_norm": 0.014828486368060112,
+ "learning_rate": 0.0006,
+ "loss": 4.1439714431762695,
+ "step": 4923
+ },
+ {
+ "epoch": 68.39126637554585,
+ "grad_norm": 0.013730363920331001,
+ "learning_rate": 0.0006,
+ "loss": 4.287973880767822,
+ "step": 4924
+ },
+ {
+ "epoch": 68.40524017467249,
+ "grad_norm": 0.013852203264832497,
+ "learning_rate": 0.0006,
+ "loss": 4.248824119567871,
+ "step": 4925
+ },
+ {
+ "epoch": 68.41921397379913,
+ "grad_norm": 0.014005862176418304,
+ "learning_rate": 0.0006,
+ "loss": 4.293768882751465,
+ "step": 4926
+ },
+ {
+ "epoch": 68.43318777292576,
+ "grad_norm": 0.014435851015150547,
+ "learning_rate": 0.0006,
+ "loss": 4.204813003540039,
+ "step": 4927
+ },
+ {
+ "epoch": 68.44716157205241,
+ "grad_norm": 0.015305934473872185,
+ "learning_rate": 0.0006,
+ "loss": 4.329364776611328,
+ "step": 4928
+ },
+ {
+ "epoch": 68.46113537117904,
+ "grad_norm": 0.014356814324855804,
+ "learning_rate": 0.0006,
+ "loss": 4.264206409454346,
+ "step": 4929
+ },
+ {
+ "epoch": 68.47510917030567,
+ "grad_norm": 0.01432347297668457,
+ "learning_rate": 0.0006,
+ "loss": 4.29440450668335,
+ "step": 4930
+ },
+ {
+ "epoch": 68.48908296943232,
+ "grad_norm": 0.012672432698309422,
+ "learning_rate": 0.0006,
+ "loss": 4.258024215698242,
+ "step": 4931
+ },
+ {
+ "epoch": 68.50305676855895,
+ "grad_norm": 0.014209977351129055,
+ "learning_rate": 0.0006,
+ "loss": 4.216668128967285,
+ "step": 4932
+ },
+ {
+ "epoch": 68.5170305676856,
+ "grad_norm": 0.014606663957238197,
+ "learning_rate": 0.0006,
+ "loss": 4.294184684753418,
+ "step": 4933
+ },
+ {
+ "epoch": 68.53100436681223,
+ "grad_norm": 0.014986990951001644,
+ "learning_rate": 0.0006,
+ "loss": 4.382591247558594,
+ "step": 4934
+ },
+ {
+ "epoch": 68.54497816593886,
+ "grad_norm": 0.013998783193528652,
+ "learning_rate": 0.0006,
+ "loss": 4.195143222808838,
+ "step": 4935
+ },
+ {
+ "epoch": 68.5589519650655,
+ "grad_norm": 0.015011102892458439,
+ "learning_rate": 0.0006,
+ "loss": 4.313785076141357,
+ "step": 4936
+ },
+ {
+ "epoch": 68.57292576419214,
+ "grad_norm": 0.01641209051012993,
+ "learning_rate": 0.0006,
+ "loss": 4.322803020477295,
+ "step": 4937
+ },
+ {
+ "epoch": 68.58689956331878,
+ "grad_norm": 0.016207464039325714,
+ "learning_rate": 0.0006,
+ "loss": 4.311431884765625,
+ "step": 4938
+ },
+ {
+ "epoch": 68.60087336244541,
+ "grad_norm": 0.01747811958193779,
+ "learning_rate": 0.0006,
+ "loss": 4.310845375061035,
+ "step": 4939
+ },
+ {
+ "epoch": 68.61484716157206,
+ "grad_norm": 0.019475717097520828,
+ "learning_rate": 0.0006,
+ "loss": 4.338624954223633,
+ "step": 4940
+ },
+ {
+ "epoch": 68.62882096069869,
+ "grad_norm": 0.018641719594597816,
+ "learning_rate": 0.0006,
+ "loss": 4.34913969039917,
+ "step": 4941
+ },
+ {
+ "epoch": 68.64279475982532,
+ "grad_norm": 0.018376432359218597,
+ "learning_rate": 0.0006,
+ "loss": 4.28464412689209,
+ "step": 4942
+ },
+ {
+ "epoch": 68.65676855895197,
+ "grad_norm": 0.017095010727643967,
+ "learning_rate": 0.0006,
+ "loss": 4.217957496643066,
+ "step": 4943
+ },
+ {
+ "epoch": 68.6707423580786,
+ "grad_norm": 0.016321983188390732,
+ "learning_rate": 0.0006,
+ "loss": 4.220042705535889,
+ "step": 4944
+ },
+ {
+ "epoch": 68.68471615720524,
+ "grad_norm": 0.017128000035881996,
+ "learning_rate": 0.0006,
+ "loss": 4.293500900268555,
+ "step": 4945
+ },
+ {
+ "epoch": 68.69868995633188,
+ "grad_norm": 0.015592411160469055,
+ "learning_rate": 0.0006,
+ "loss": 4.195798873901367,
+ "step": 4946
+ },
+ {
+ "epoch": 68.71266375545852,
+ "grad_norm": 0.016607079654932022,
+ "learning_rate": 0.0006,
+ "loss": 4.2654337882995605,
+ "step": 4947
+ },
+ {
+ "epoch": 68.72663755458515,
+ "grad_norm": 0.01545762736350298,
+ "learning_rate": 0.0006,
+ "loss": 4.303828239440918,
+ "step": 4948
+ },
+ {
+ "epoch": 68.74061135371178,
+ "grad_norm": 0.014328422956168652,
+ "learning_rate": 0.0006,
+ "loss": 4.29379940032959,
+ "step": 4949
+ },
+ {
+ "epoch": 68.75458515283843,
+ "grad_norm": 0.016516804695129395,
+ "learning_rate": 0.0006,
+ "loss": 4.2317214012146,
+ "step": 4950
+ },
+ {
+ "epoch": 68.76855895196506,
+ "grad_norm": 0.016324171796441078,
+ "learning_rate": 0.0006,
+ "loss": 4.278043746948242,
+ "step": 4951
+ },
+ {
+ "epoch": 68.78253275109171,
+ "grad_norm": 0.015511159785091877,
+ "learning_rate": 0.0006,
+ "loss": 4.188244342803955,
+ "step": 4952
+ },
+ {
+ "epoch": 68.79650655021834,
+ "grad_norm": 0.01676103286445141,
+ "learning_rate": 0.0006,
+ "loss": 4.2169976234436035,
+ "step": 4953
+ },
+ {
+ "epoch": 68.81048034934499,
+ "grad_norm": 0.015428312122821808,
+ "learning_rate": 0.0006,
+ "loss": 4.4004716873168945,
+ "step": 4954
+ },
+ {
+ "epoch": 68.82445414847162,
+ "grad_norm": 0.0162410456687212,
+ "learning_rate": 0.0006,
+ "loss": 4.282752513885498,
+ "step": 4955
+ },
+ {
+ "epoch": 68.83842794759825,
+ "grad_norm": 0.017578862607479095,
+ "learning_rate": 0.0006,
+ "loss": 4.098004341125488,
+ "step": 4956
+ },
+ {
+ "epoch": 68.8524017467249,
+ "grad_norm": 0.01805749163031578,
+ "learning_rate": 0.0006,
+ "loss": 4.277988433837891,
+ "step": 4957
+ },
+ {
+ "epoch": 68.86637554585153,
+ "grad_norm": 0.016728045418858528,
+ "learning_rate": 0.0006,
+ "loss": 4.29766321182251,
+ "step": 4958
+ },
+ {
+ "epoch": 68.88034934497817,
+ "grad_norm": 0.014899657107889652,
+ "learning_rate": 0.0006,
+ "loss": 4.329736709594727,
+ "step": 4959
+ },
+ {
+ "epoch": 68.8943231441048,
+ "grad_norm": 0.013441305607557297,
+ "learning_rate": 0.0006,
+ "loss": 4.2873215675354,
+ "step": 4960
+ },
+ {
+ "epoch": 68.90829694323143,
+ "grad_norm": 0.01554128248244524,
+ "learning_rate": 0.0006,
+ "loss": 4.278327941894531,
+ "step": 4961
+ },
+ {
+ "epoch": 68.92227074235808,
+ "grad_norm": 0.01806696318089962,
+ "learning_rate": 0.0006,
+ "loss": 4.38275146484375,
+ "step": 4962
+ },
+ {
+ "epoch": 68.93624454148471,
+ "grad_norm": 0.017025409266352654,
+ "learning_rate": 0.0006,
+ "loss": 4.210447311401367,
+ "step": 4963
+ },
+ {
+ "epoch": 68.95021834061136,
+ "grad_norm": 0.01597420684993267,
+ "learning_rate": 0.0006,
+ "loss": 4.309865951538086,
+ "step": 4964
+ },
+ {
+ "epoch": 68.96419213973799,
+ "grad_norm": 0.015835624188184738,
+ "learning_rate": 0.0006,
+ "loss": 4.225904941558838,
+ "step": 4965
+ },
+ {
+ "epoch": 68.97816593886463,
+ "grad_norm": 0.018170492723584175,
+ "learning_rate": 0.0006,
+ "loss": 4.248478889465332,
+ "step": 4966
+ },
+ {
+ "epoch": 68.99213973799127,
+ "grad_norm": 0.017372991889715195,
+ "learning_rate": 0.0006,
+ "loss": 4.283486843109131,
+ "step": 4967
+ },
+ {
+ "epoch": 69.0,
+ "grad_norm": 0.016305390745401382,
+ "learning_rate": 0.0006,
+ "loss": 4.273636817932129,
+ "step": 4968
+ },
+ {
+ "epoch": 69.0,
+ "eval_loss": 4.66873836517334,
+ "eval_runtime": 57.0791,
+ "eval_samples_per_second": 42.783,
+ "eval_steps_per_second": 1.349,
+ "step": 4968
+ },
+ {
+ "epoch": 69.01397379912663,
+ "grad_norm": 0.016567355021834373,
+ "learning_rate": 0.0006,
+ "loss": 4.269988059997559,
+ "step": 4969
+ },
+ {
+ "epoch": 69.02794759825328,
+ "grad_norm": 0.017476852983236313,
+ "learning_rate": 0.0006,
+ "loss": 4.2518720626831055,
+ "step": 4970
+ },
+ {
+ "epoch": 69.04192139737991,
+ "grad_norm": 0.017070338129997253,
+ "learning_rate": 0.0006,
+ "loss": 4.238024711608887,
+ "step": 4971
+ },
+ {
+ "epoch": 69.05589519650655,
+ "grad_norm": 0.01639373227953911,
+ "learning_rate": 0.0006,
+ "loss": 4.188558101654053,
+ "step": 4972
+ },
+ {
+ "epoch": 69.06986899563319,
+ "grad_norm": 0.01856113411486149,
+ "learning_rate": 0.0006,
+ "loss": 4.339556694030762,
+ "step": 4973
+ },
+ {
+ "epoch": 69.08384279475983,
+ "grad_norm": 0.01762578636407852,
+ "learning_rate": 0.0006,
+ "loss": 4.141518592834473,
+ "step": 4974
+ },
+ {
+ "epoch": 69.09781659388646,
+ "grad_norm": 0.015437949448823929,
+ "learning_rate": 0.0006,
+ "loss": 4.225461959838867,
+ "step": 4975
+ },
+ {
+ "epoch": 69.1117903930131,
+ "grad_norm": 0.015293503180146217,
+ "learning_rate": 0.0006,
+ "loss": 4.290884971618652,
+ "step": 4976
+ },
+ {
+ "epoch": 69.12576419213974,
+ "grad_norm": 0.016259880736470222,
+ "learning_rate": 0.0006,
+ "loss": 4.121919631958008,
+ "step": 4977
+ },
+ {
+ "epoch": 69.13973799126637,
+ "grad_norm": 0.017225584015250206,
+ "learning_rate": 0.0006,
+ "loss": 4.136716842651367,
+ "step": 4978
+ },
+ {
+ "epoch": 69.15371179039302,
+ "grad_norm": 0.01699289306998253,
+ "learning_rate": 0.0006,
+ "loss": 4.193601608276367,
+ "step": 4979
+ },
+ {
+ "epoch": 69.16768558951965,
+ "grad_norm": 0.015613161027431488,
+ "learning_rate": 0.0006,
+ "loss": 4.288212299346924,
+ "step": 4980
+ },
+ {
+ "epoch": 69.18165938864628,
+ "grad_norm": 0.01643933728337288,
+ "learning_rate": 0.0006,
+ "loss": 4.324707984924316,
+ "step": 4981
+ },
+ {
+ "epoch": 69.19563318777293,
+ "grad_norm": 0.01648905873298645,
+ "learning_rate": 0.0006,
+ "loss": 4.287806510925293,
+ "step": 4982
+ },
+ {
+ "epoch": 69.20960698689956,
+ "grad_norm": 0.015444950200617313,
+ "learning_rate": 0.0006,
+ "loss": 4.200622081756592,
+ "step": 4983
+ },
+ {
+ "epoch": 69.2235807860262,
+ "grad_norm": 0.016034258529543877,
+ "learning_rate": 0.0006,
+ "loss": 4.2193827629089355,
+ "step": 4984
+ },
+ {
+ "epoch": 69.23755458515284,
+ "grad_norm": 0.017984388396143913,
+ "learning_rate": 0.0006,
+ "loss": 4.18237829208374,
+ "step": 4985
+ },
+ {
+ "epoch": 69.25152838427948,
+ "grad_norm": 0.0195505041629076,
+ "learning_rate": 0.0006,
+ "loss": 4.216727256774902,
+ "step": 4986
+ },
+ {
+ "epoch": 69.26550218340611,
+ "grad_norm": 0.017779147252440453,
+ "learning_rate": 0.0006,
+ "loss": 4.278628826141357,
+ "step": 4987
+ },
+ {
+ "epoch": 69.27947598253274,
+ "grad_norm": 0.017119476571679115,
+ "learning_rate": 0.0006,
+ "loss": 4.2409138679504395,
+ "step": 4988
+ },
+ {
+ "epoch": 69.29344978165939,
+ "grad_norm": 0.016840364784002304,
+ "learning_rate": 0.0006,
+ "loss": 4.207575798034668,
+ "step": 4989
+ },
+ {
+ "epoch": 69.30742358078602,
+ "grad_norm": 0.014593573287129402,
+ "learning_rate": 0.0006,
+ "loss": 4.164905548095703,
+ "step": 4990
+ },
+ {
+ "epoch": 69.32139737991267,
+ "grad_norm": 0.013922630809247494,
+ "learning_rate": 0.0006,
+ "loss": 4.2696990966796875,
+ "step": 4991
+ },
+ {
+ "epoch": 69.3353711790393,
+ "grad_norm": 0.01577564887702465,
+ "learning_rate": 0.0006,
+ "loss": 4.201589107513428,
+ "step": 4992
+ },
+ {
+ "epoch": 69.34934497816595,
+ "grad_norm": 0.016392096877098083,
+ "learning_rate": 0.0006,
+ "loss": 4.248831748962402,
+ "step": 4993
+ },
+ {
+ "epoch": 69.36331877729258,
+ "grad_norm": 0.015961311757564545,
+ "learning_rate": 0.0006,
+ "loss": 4.181260108947754,
+ "step": 4994
+ },
+ {
+ "epoch": 69.37729257641921,
+ "grad_norm": 0.015350551344454288,
+ "learning_rate": 0.0006,
+ "loss": 4.230352878570557,
+ "step": 4995
+ },
+ {
+ "epoch": 69.39126637554585,
+ "grad_norm": 0.015870466828346252,
+ "learning_rate": 0.0006,
+ "loss": 4.298305511474609,
+ "step": 4996
+ },
+ {
+ "epoch": 69.40524017467249,
+ "grad_norm": 0.013716642744839191,
+ "learning_rate": 0.0006,
+ "loss": 4.220139503479004,
+ "step": 4997
+ },
+ {
+ "epoch": 69.41921397379913,
+ "grad_norm": 0.014431777410209179,
+ "learning_rate": 0.0006,
+ "loss": 4.288130760192871,
+ "step": 4998
+ },
+ {
+ "epoch": 69.43318777292576,
+ "grad_norm": 0.01426647324115038,
+ "learning_rate": 0.0006,
+ "loss": 4.295767784118652,
+ "step": 4999
+ },
+ {
+ "epoch": 69.44716157205241,
+ "grad_norm": 0.014286615885794163,
+ "learning_rate": 0.0006,
+ "loss": 4.38106632232666,
+ "step": 5000
+ },
+ {
+ "epoch": 69.46113537117904,
+ "grad_norm": 0.017990585416555405,
+ "learning_rate": 0.0006,
+ "loss": 4.1593017578125,
+ "step": 5001
+ },
+ {
+ "epoch": 69.47510917030567,
+ "grad_norm": 0.018846049904823303,
+ "learning_rate": 0.0006,
+ "loss": 4.216091632843018,
+ "step": 5002
+ },
+ {
+ "epoch": 69.48908296943232,
+ "grad_norm": 0.018179181963205338,
+ "learning_rate": 0.0006,
+ "loss": 4.240332126617432,
+ "step": 5003
+ },
+ {
+ "epoch": 69.50305676855895,
+ "grad_norm": 0.016693448647856712,
+ "learning_rate": 0.0006,
+ "loss": 4.256624221801758,
+ "step": 5004
+ },
+ {
+ "epoch": 69.5170305676856,
+ "grad_norm": 0.01718989387154579,
+ "learning_rate": 0.0006,
+ "loss": 4.300940990447998,
+ "step": 5005
+ },
+ {
+ "epoch": 69.53100436681223,
+ "grad_norm": 0.020236829295754433,
+ "learning_rate": 0.0006,
+ "loss": 4.289146423339844,
+ "step": 5006
+ },
+ {
+ "epoch": 69.54497816593886,
+ "grad_norm": 0.024860285222530365,
+ "learning_rate": 0.0006,
+ "loss": 4.216216087341309,
+ "step": 5007
+ },
+ {
+ "epoch": 69.5589519650655,
+ "grad_norm": 0.02559969387948513,
+ "learning_rate": 0.0006,
+ "loss": 4.262890815734863,
+ "step": 5008
+ },
+ {
+ "epoch": 69.57292576419214,
+ "grad_norm": 0.019180957227945328,
+ "learning_rate": 0.0006,
+ "loss": 4.230309009552002,
+ "step": 5009
+ },
+ {
+ "epoch": 69.58689956331878,
+ "grad_norm": 0.020442591980099678,
+ "learning_rate": 0.0006,
+ "loss": 4.291759490966797,
+ "step": 5010
+ },
+ {
+ "epoch": 69.60087336244541,
+ "grad_norm": 0.0212619137018919,
+ "learning_rate": 0.0006,
+ "loss": 4.3529839515686035,
+ "step": 5011
+ },
+ {
+ "epoch": 69.61484716157206,
+ "grad_norm": 0.02067175693809986,
+ "learning_rate": 0.0006,
+ "loss": 4.33197021484375,
+ "step": 5012
+ },
+ {
+ "epoch": 69.62882096069869,
+ "grad_norm": 0.01802872307598591,
+ "learning_rate": 0.0006,
+ "loss": 4.2101969718933105,
+ "step": 5013
+ },
+ {
+ "epoch": 69.64279475982532,
+ "grad_norm": 0.01787702925503254,
+ "learning_rate": 0.0006,
+ "loss": 4.270336627960205,
+ "step": 5014
+ },
+ {
+ "epoch": 69.65676855895197,
+ "grad_norm": 0.01586683839559555,
+ "learning_rate": 0.0006,
+ "loss": 4.222309589385986,
+ "step": 5015
+ },
+ {
+ "epoch": 69.6707423580786,
+ "grad_norm": 0.01717999391257763,
+ "learning_rate": 0.0006,
+ "loss": 4.253704071044922,
+ "step": 5016
+ },
+ {
+ "epoch": 69.68471615720524,
+ "grad_norm": 0.01594279147684574,
+ "learning_rate": 0.0006,
+ "loss": 4.344658851623535,
+ "step": 5017
+ },
+ {
+ "epoch": 69.69868995633188,
+ "grad_norm": 0.014419942162930965,
+ "learning_rate": 0.0006,
+ "loss": 4.2144880294799805,
+ "step": 5018
+ },
+ {
+ "epoch": 69.71266375545852,
+ "grad_norm": 0.015824224799871445,
+ "learning_rate": 0.0006,
+ "loss": 4.226252555847168,
+ "step": 5019
+ },
+ {
+ "epoch": 69.72663755458515,
+ "grad_norm": 0.013329979032278061,
+ "learning_rate": 0.0006,
+ "loss": 4.238587379455566,
+ "step": 5020
+ },
+ {
+ "epoch": 69.74061135371178,
+ "grad_norm": 0.013745979405939579,
+ "learning_rate": 0.0006,
+ "loss": 4.317209243774414,
+ "step": 5021
+ },
+ {
+ "epoch": 69.75458515283843,
+ "grad_norm": 0.014064167626202106,
+ "learning_rate": 0.0006,
+ "loss": 4.316982746124268,
+ "step": 5022
+ },
+ {
+ "epoch": 69.76855895196506,
+ "grad_norm": 0.014611301012337208,
+ "learning_rate": 0.0006,
+ "loss": 4.212728500366211,
+ "step": 5023
+ },
+ {
+ "epoch": 69.78253275109171,
+ "grad_norm": 0.014947416260838509,
+ "learning_rate": 0.0006,
+ "loss": 4.319550514221191,
+ "step": 5024
+ },
+ {
+ "epoch": 69.79650655021834,
+ "grad_norm": 0.014780526980757713,
+ "learning_rate": 0.0006,
+ "loss": 4.1888933181762695,
+ "step": 5025
+ },
+ {
+ "epoch": 69.81048034934499,
+ "grad_norm": 0.01611459068953991,
+ "learning_rate": 0.0006,
+ "loss": 4.239692211151123,
+ "step": 5026
+ },
+ {
+ "epoch": 69.82445414847162,
+ "grad_norm": 0.015395848080515862,
+ "learning_rate": 0.0006,
+ "loss": 4.308794021606445,
+ "step": 5027
+ },
+ {
+ "epoch": 69.83842794759825,
+ "grad_norm": 0.016113396733999252,
+ "learning_rate": 0.0006,
+ "loss": 4.328035354614258,
+ "step": 5028
+ },
+ {
+ "epoch": 69.8524017467249,
+ "grad_norm": 0.016386888921260834,
+ "learning_rate": 0.0006,
+ "loss": 4.272579193115234,
+ "step": 5029
+ },
+ {
+ "epoch": 69.86637554585153,
+ "grad_norm": 0.013675205409526825,
+ "learning_rate": 0.0006,
+ "loss": 4.226552963256836,
+ "step": 5030
+ },
+ {
+ "epoch": 69.88034934497817,
+ "grad_norm": 0.01472668256610632,
+ "learning_rate": 0.0006,
+ "loss": 4.224167346954346,
+ "step": 5031
+ },
+ {
+ "epoch": 69.8943231441048,
+ "grad_norm": 0.014958228915929794,
+ "learning_rate": 0.0006,
+ "loss": 4.232619285583496,
+ "step": 5032
+ },
+ {
+ "epoch": 69.90829694323143,
+ "grad_norm": 0.012926832772791386,
+ "learning_rate": 0.0006,
+ "loss": 4.27107048034668,
+ "step": 5033
+ },
+ {
+ "epoch": 69.92227074235808,
+ "grad_norm": 0.013138801790773869,
+ "learning_rate": 0.0006,
+ "loss": 4.288477420806885,
+ "step": 5034
+ },
+ {
+ "epoch": 69.93624454148471,
+ "grad_norm": 0.013415461406111717,
+ "learning_rate": 0.0006,
+ "loss": 4.2033257484436035,
+ "step": 5035
+ },
+ {
+ "epoch": 69.95021834061136,
+ "grad_norm": 0.013807233422994614,
+ "learning_rate": 0.0006,
+ "loss": 4.274764537811279,
+ "step": 5036
+ },
+ {
+ "epoch": 69.96419213973799,
+ "grad_norm": 0.013929193839430809,
+ "learning_rate": 0.0006,
+ "loss": 4.1854352951049805,
+ "step": 5037
+ },
+ {
+ "epoch": 69.97816593886463,
+ "grad_norm": 0.015497001819312572,
+ "learning_rate": 0.0006,
+ "loss": 4.226498126983643,
+ "step": 5038
+ },
+ {
+ "epoch": 69.99213973799127,
+ "grad_norm": 0.01874048449099064,
+ "learning_rate": 0.0006,
+ "loss": 4.304137706756592,
+ "step": 5039
+ },
+ {
+ "epoch": 70.0,
+ "grad_norm": 0.021745791658759117,
+ "learning_rate": 0.0006,
+ "loss": 4.206809997558594,
+ "step": 5040
+ },
+ {
+ "epoch": 70.0,
+ "eval_loss": 4.675539970397949,
+ "eval_runtime": 56.9034,
+ "eval_samples_per_second": 42.915,
+ "eval_steps_per_second": 1.353,
+ "step": 5040
+ },
+ {
+ "epoch": 70.01397379912663,
+ "grad_norm": 0.02124839462339878,
+ "learning_rate": 0.0006,
+ "loss": 4.198199272155762,
+ "step": 5041
+ },
+ {
+ "epoch": 70.02794759825328,
+ "grad_norm": 0.019432688131928444,
+ "learning_rate": 0.0006,
+ "loss": 4.187261581420898,
+ "step": 5042
+ },
+ {
+ "epoch": 70.04192139737991,
+ "grad_norm": 0.01769290305674076,
+ "learning_rate": 0.0006,
+ "loss": 4.155284881591797,
+ "step": 5043
+ },
+ {
+ "epoch": 70.05589519650655,
+ "grad_norm": 0.01751520484685898,
+ "learning_rate": 0.0006,
+ "loss": 4.386857986450195,
+ "step": 5044
+ },
+ {
+ "epoch": 70.06986899563319,
+ "grad_norm": 0.018138699233531952,
+ "learning_rate": 0.0006,
+ "loss": 4.117342472076416,
+ "step": 5045
+ },
+ {
+ "epoch": 70.08384279475983,
+ "grad_norm": 0.016331054270267487,
+ "learning_rate": 0.0006,
+ "loss": 4.225295066833496,
+ "step": 5046
+ },
+ {
+ "epoch": 70.09781659388646,
+ "grad_norm": 0.01658729277551174,
+ "learning_rate": 0.0006,
+ "loss": 4.143835067749023,
+ "step": 5047
+ },
+ {
+ "epoch": 70.1117903930131,
+ "grad_norm": 0.016367698088288307,
+ "learning_rate": 0.0006,
+ "loss": 4.1469268798828125,
+ "step": 5048
+ },
+ {
+ "epoch": 70.12576419213974,
+ "grad_norm": 0.015151629224419594,
+ "learning_rate": 0.0006,
+ "loss": 4.117629051208496,
+ "step": 5049
+ },
+ {
+ "epoch": 70.13973799126637,
+ "grad_norm": 0.016434503719210625,
+ "learning_rate": 0.0006,
+ "loss": 4.107848644256592,
+ "step": 5050
+ },
+ {
+ "epoch": 70.15371179039302,
+ "grad_norm": 0.017837485298514366,
+ "learning_rate": 0.0006,
+ "loss": 4.10837459564209,
+ "step": 5051
+ },
+ {
+ "epoch": 70.16768558951965,
+ "grad_norm": 0.0165961142629385,
+ "learning_rate": 0.0006,
+ "loss": 4.169414043426514,
+ "step": 5052
+ },
+ {
+ "epoch": 70.18165938864628,
+ "grad_norm": 0.0157014112919569,
+ "learning_rate": 0.0006,
+ "loss": 4.0629706382751465,
+ "step": 5053
+ },
+ {
+ "epoch": 70.19563318777293,
+ "grad_norm": 0.01581883803009987,
+ "learning_rate": 0.0006,
+ "loss": 4.220012664794922,
+ "step": 5054
+ },
+ {
+ "epoch": 70.20960698689956,
+ "grad_norm": 0.015285917557775974,
+ "learning_rate": 0.0006,
+ "loss": 4.275337219238281,
+ "step": 5055
+ },
+ {
+ "epoch": 70.2235807860262,
+ "grad_norm": 0.014984549954533577,
+ "learning_rate": 0.0006,
+ "loss": 4.284764289855957,
+ "step": 5056
+ },
+ {
+ "epoch": 70.23755458515284,
+ "grad_norm": 0.015172012150287628,
+ "learning_rate": 0.0006,
+ "loss": 4.234498023986816,
+ "step": 5057
+ },
+ {
+ "epoch": 70.25152838427948,
+ "grad_norm": 0.016846632584929466,
+ "learning_rate": 0.0006,
+ "loss": 4.181689262390137,
+ "step": 5058
+ },
+ {
+ "epoch": 70.26550218340611,
+ "grad_norm": 0.01920711249113083,
+ "learning_rate": 0.0006,
+ "loss": 4.235280513763428,
+ "step": 5059
+ },
+ {
+ "epoch": 70.27947598253274,
+ "grad_norm": 0.017765291035175323,
+ "learning_rate": 0.0006,
+ "loss": 4.238955497741699,
+ "step": 5060
+ },
+ {
+ "epoch": 70.29344978165939,
+ "grad_norm": 0.015436639077961445,
+ "learning_rate": 0.0006,
+ "loss": 4.232516288757324,
+ "step": 5061
+ },
+ {
+ "epoch": 70.30742358078602,
+ "grad_norm": 0.015660086646676064,
+ "learning_rate": 0.0006,
+ "loss": 4.260349273681641,
+ "step": 5062
+ },
+ {
+ "epoch": 70.32139737991267,
+ "grad_norm": 0.016095709055662155,
+ "learning_rate": 0.0006,
+ "loss": 4.168505668640137,
+ "step": 5063
+ },
+ {
+ "epoch": 70.3353711790393,
+ "grad_norm": 0.015001940540969372,
+ "learning_rate": 0.0006,
+ "loss": 4.161739349365234,
+ "step": 5064
+ },
+ {
+ "epoch": 70.34934497816595,
+ "grad_norm": 0.014668774791061878,
+ "learning_rate": 0.0006,
+ "loss": 4.181029796600342,
+ "step": 5065
+ },
+ {
+ "epoch": 70.36331877729258,
+ "grad_norm": 0.013266284950077534,
+ "learning_rate": 0.0006,
+ "loss": 4.3052873611450195,
+ "step": 5066
+ },
+ {
+ "epoch": 70.37729257641921,
+ "grad_norm": 0.015396138653159142,
+ "learning_rate": 0.0006,
+ "loss": 4.34413480758667,
+ "step": 5067
+ },
+ {
+ "epoch": 70.39126637554585,
+ "grad_norm": 0.015511523932218552,
+ "learning_rate": 0.0006,
+ "loss": 4.235037803649902,
+ "step": 5068
+ },
+ {
+ "epoch": 70.40524017467249,
+ "grad_norm": 0.014825738966464996,
+ "learning_rate": 0.0006,
+ "loss": 4.262448310852051,
+ "step": 5069
+ },
+ {
+ "epoch": 70.41921397379913,
+ "grad_norm": 0.015309952199459076,
+ "learning_rate": 0.0006,
+ "loss": 4.258711814880371,
+ "step": 5070
+ },
+ {
+ "epoch": 70.43318777292576,
+ "grad_norm": 0.017614055424928665,
+ "learning_rate": 0.0006,
+ "loss": 4.286028861999512,
+ "step": 5071
+ },
+ {
+ "epoch": 70.44716157205241,
+ "grad_norm": 0.0166039876639843,
+ "learning_rate": 0.0006,
+ "loss": 4.149197578430176,
+ "step": 5072
+ },
+ {
+ "epoch": 70.46113537117904,
+ "grad_norm": 0.016911419108510017,
+ "learning_rate": 0.0006,
+ "loss": 4.271981239318848,
+ "step": 5073
+ },
+ {
+ "epoch": 70.47510917030567,
+ "grad_norm": 0.016806820407509804,
+ "learning_rate": 0.0006,
+ "loss": 4.213315010070801,
+ "step": 5074
+ },
+ {
+ "epoch": 70.48908296943232,
+ "grad_norm": 0.015221742913126945,
+ "learning_rate": 0.0006,
+ "loss": 4.171091079711914,
+ "step": 5075
+ },
+ {
+ "epoch": 70.50305676855895,
+ "grad_norm": 0.014175347983837128,
+ "learning_rate": 0.0006,
+ "loss": 4.185309410095215,
+ "step": 5076
+ },
+ {
+ "epoch": 70.5170305676856,
+ "grad_norm": 0.013897798955440521,
+ "learning_rate": 0.0006,
+ "loss": 4.250119209289551,
+ "step": 5077
+ },
+ {
+ "epoch": 70.53100436681223,
+ "grad_norm": 0.014791003428399563,
+ "learning_rate": 0.0006,
+ "loss": 4.2504682540893555,
+ "step": 5078
+ },
+ {
+ "epoch": 70.54497816593886,
+ "grad_norm": 0.016700128093361855,
+ "learning_rate": 0.0006,
+ "loss": 4.235134124755859,
+ "step": 5079
+ },
+ {
+ "epoch": 70.5589519650655,
+ "grad_norm": 0.016679363325238228,
+ "learning_rate": 0.0006,
+ "loss": 4.32028865814209,
+ "step": 5080
+ },
+ {
+ "epoch": 70.57292576419214,
+ "grad_norm": 0.015117401257157326,
+ "learning_rate": 0.0006,
+ "loss": 4.210995674133301,
+ "step": 5081
+ },
+ {
+ "epoch": 70.58689956331878,
+ "grad_norm": 0.014441263861954212,
+ "learning_rate": 0.0006,
+ "loss": 4.27849006652832,
+ "step": 5082
+ },
+ {
+ "epoch": 70.60087336244541,
+ "grad_norm": 0.01523780357092619,
+ "learning_rate": 0.0006,
+ "loss": 4.197360992431641,
+ "step": 5083
+ },
+ {
+ "epoch": 70.61484716157206,
+ "grad_norm": 0.016201738268136978,
+ "learning_rate": 0.0006,
+ "loss": 4.28239107131958,
+ "step": 5084
+ },
+ {
+ "epoch": 70.62882096069869,
+ "grad_norm": 0.016003292053937912,
+ "learning_rate": 0.0006,
+ "loss": 4.209715843200684,
+ "step": 5085
+ },
+ {
+ "epoch": 70.64279475982532,
+ "grad_norm": 0.01472893450409174,
+ "learning_rate": 0.0006,
+ "loss": 4.169078826904297,
+ "step": 5086
+ },
+ {
+ "epoch": 70.65676855895197,
+ "grad_norm": 0.01604326069355011,
+ "learning_rate": 0.0006,
+ "loss": 4.090967178344727,
+ "step": 5087
+ },
+ {
+ "epoch": 70.6707423580786,
+ "grad_norm": 0.01598356105387211,
+ "learning_rate": 0.0006,
+ "loss": 4.16630220413208,
+ "step": 5088
+ },
+ {
+ "epoch": 70.68471615720524,
+ "grad_norm": 0.016988780349493027,
+ "learning_rate": 0.0006,
+ "loss": 4.117822170257568,
+ "step": 5089
+ },
+ {
+ "epoch": 70.69868995633188,
+ "grad_norm": 0.018582893535494804,
+ "learning_rate": 0.0006,
+ "loss": 4.192902088165283,
+ "step": 5090
+ },
+ {
+ "epoch": 70.71266375545852,
+ "grad_norm": 0.01793155074119568,
+ "learning_rate": 0.0006,
+ "loss": 4.314806938171387,
+ "step": 5091
+ },
+ {
+ "epoch": 70.72663755458515,
+ "grad_norm": 0.016286427155137062,
+ "learning_rate": 0.0006,
+ "loss": 4.2137556076049805,
+ "step": 5092
+ },
+ {
+ "epoch": 70.74061135371178,
+ "grad_norm": 0.014718376100063324,
+ "learning_rate": 0.0006,
+ "loss": 4.3073015213012695,
+ "step": 5093
+ },
+ {
+ "epoch": 70.75458515283843,
+ "grad_norm": 0.014675077050924301,
+ "learning_rate": 0.0006,
+ "loss": 4.160961151123047,
+ "step": 5094
+ },
+ {
+ "epoch": 70.76855895196506,
+ "grad_norm": 0.015550913289189339,
+ "learning_rate": 0.0006,
+ "loss": 4.322887420654297,
+ "step": 5095
+ },
+ {
+ "epoch": 70.78253275109171,
+ "grad_norm": 0.01453235000371933,
+ "learning_rate": 0.0006,
+ "loss": 4.2915873527526855,
+ "step": 5096
+ },
+ {
+ "epoch": 70.79650655021834,
+ "grad_norm": 0.01459164172410965,
+ "learning_rate": 0.0006,
+ "loss": 4.247288227081299,
+ "step": 5097
+ },
+ {
+ "epoch": 70.81048034934499,
+ "grad_norm": 0.015395080670714378,
+ "learning_rate": 0.0006,
+ "loss": 4.360955715179443,
+ "step": 5098
+ },
+ {
+ "epoch": 70.82445414847162,
+ "grad_norm": 0.01317683607339859,
+ "learning_rate": 0.0006,
+ "loss": 4.181969165802002,
+ "step": 5099
+ },
+ {
+ "epoch": 70.83842794759825,
+ "grad_norm": 0.013917932286858559,
+ "learning_rate": 0.0006,
+ "loss": 4.264626979827881,
+ "step": 5100
+ },
+ {
+ "epoch": 70.8524017467249,
+ "grad_norm": 0.015268695540726185,
+ "learning_rate": 0.0006,
+ "loss": 4.287857532501221,
+ "step": 5101
+ },
+ {
+ "epoch": 70.86637554585153,
+ "grad_norm": 0.015531329438090324,
+ "learning_rate": 0.0006,
+ "loss": 4.294188499450684,
+ "step": 5102
+ },
+ {
+ "epoch": 70.88034934497817,
+ "grad_norm": 0.015607129782438278,
+ "learning_rate": 0.0006,
+ "loss": 4.331040859222412,
+ "step": 5103
+ },
+ {
+ "epoch": 70.8943231441048,
+ "grad_norm": 0.015297149308025837,
+ "learning_rate": 0.0006,
+ "loss": 4.243650436401367,
+ "step": 5104
+ },
+ {
+ "epoch": 70.90829694323143,
+ "grad_norm": 0.014790025539696217,
+ "learning_rate": 0.0006,
+ "loss": 4.159859657287598,
+ "step": 5105
+ },
+ {
+ "epoch": 70.92227074235808,
+ "grad_norm": 0.013466274365782738,
+ "learning_rate": 0.0006,
+ "loss": 4.221992492675781,
+ "step": 5106
+ },
+ {
+ "epoch": 70.93624454148471,
+ "grad_norm": 0.01361898984760046,
+ "learning_rate": 0.0006,
+ "loss": 4.278751373291016,
+ "step": 5107
+ },
+ {
+ "epoch": 70.95021834061136,
+ "grad_norm": 0.01325751468539238,
+ "learning_rate": 0.0006,
+ "loss": 4.166001319885254,
+ "step": 5108
+ },
+ {
+ "epoch": 70.96419213973799,
+ "grad_norm": 0.012727087363600731,
+ "learning_rate": 0.0006,
+ "loss": 4.219997406005859,
+ "step": 5109
+ },
+ {
+ "epoch": 70.97816593886463,
+ "grad_norm": 0.013966867700219154,
+ "learning_rate": 0.0006,
+ "loss": 4.207491397857666,
+ "step": 5110
+ },
+ {
+ "epoch": 70.99213973799127,
+ "grad_norm": 0.015053587965667248,
+ "learning_rate": 0.0006,
+ "loss": 4.343245983123779,
+ "step": 5111
+ },
+ {
+ "epoch": 71.0,
+ "grad_norm": 0.017723368480801582,
+ "learning_rate": 0.0006,
+ "loss": 4.356122016906738,
+ "step": 5112
+ },
+ {
+ "epoch": 71.0,
+ "eval_loss": 4.598303318023682,
+ "eval_runtime": 56.7521,
+ "eval_samples_per_second": 43.029,
+ "eval_steps_per_second": 1.357,
+ "step": 5112
+ },
+ {
+ "epoch": 71.01397379912663,
+ "grad_norm": 0.018996279686689377,
+ "learning_rate": 0.0006,
+ "loss": 4.234772205352783,
+ "step": 5113
+ },
+ {
+ "epoch": 71.02794759825328,
+ "grad_norm": 0.02041555754840374,
+ "learning_rate": 0.0006,
+ "loss": 4.275816917419434,
+ "step": 5114
+ },
+ {
+ "epoch": 71.04192139737991,
+ "grad_norm": 0.021955285221338272,
+ "learning_rate": 0.0006,
+ "loss": 4.166038513183594,
+ "step": 5115
+ },
+ {
+ "epoch": 71.05589519650655,
+ "grad_norm": 0.021308740600943565,
+ "learning_rate": 0.0006,
+ "loss": 4.13063907623291,
+ "step": 5116
+ },
+ {
+ "epoch": 71.06986899563319,
+ "grad_norm": 0.01747054047882557,
+ "learning_rate": 0.0006,
+ "loss": 4.200522422790527,
+ "step": 5117
+ },
+ {
+ "epoch": 71.08384279475983,
+ "grad_norm": 0.018544532358646393,
+ "learning_rate": 0.0006,
+ "loss": 4.173915386199951,
+ "step": 5118
+ },
+ {
+ "epoch": 71.09781659388646,
+ "grad_norm": 0.0182361900806427,
+ "learning_rate": 0.0006,
+ "loss": 4.27392578125,
+ "step": 5119
+ },
+ {
+ "epoch": 71.1117903930131,
+ "grad_norm": 0.017062073573470116,
+ "learning_rate": 0.0006,
+ "loss": 4.219576358795166,
+ "step": 5120
+ },
+ {
+ "epoch": 71.12576419213974,
+ "grad_norm": 0.016289666295051575,
+ "learning_rate": 0.0006,
+ "loss": 4.133376598358154,
+ "step": 5121
+ },
+ {
+ "epoch": 71.13973799126637,
+ "grad_norm": 0.014864642173051834,
+ "learning_rate": 0.0006,
+ "loss": 4.168276786804199,
+ "step": 5122
+ },
+ {
+ "epoch": 71.15371179039302,
+ "grad_norm": 0.015309540554881096,
+ "learning_rate": 0.0006,
+ "loss": 4.175103187561035,
+ "step": 5123
+ },
+ {
+ "epoch": 71.16768558951965,
+ "grad_norm": 0.01556059904396534,
+ "learning_rate": 0.0006,
+ "loss": 4.289899826049805,
+ "step": 5124
+ },
+ {
+ "epoch": 71.18165938864628,
+ "grad_norm": 0.01479957066476345,
+ "learning_rate": 0.0006,
+ "loss": 4.341495037078857,
+ "step": 5125
+ },
+ {
+ "epoch": 71.19563318777293,
+ "grad_norm": 0.014762179926037788,
+ "learning_rate": 0.0006,
+ "loss": 4.252333641052246,
+ "step": 5126
+ },
+ {
+ "epoch": 71.20960698689956,
+ "grad_norm": 0.014897473156452179,
+ "learning_rate": 0.0006,
+ "loss": 4.2244181632995605,
+ "step": 5127
+ },
+ {
+ "epoch": 71.2235807860262,
+ "grad_norm": 0.012990163639187813,
+ "learning_rate": 0.0006,
+ "loss": 4.237288475036621,
+ "step": 5128
+ },
+ {
+ "epoch": 71.23755458515284,
+ "grad_norm": 0.012858416885137558,
+ "learning_rate": 0.0006,
+ "loss": 4.265738487243652,
+ "step": 5129
+ },
+ {
+ "epoch": 71.25152838427948,
+ "grad_norm": 0.013454788364470005,
+ "learning_rate": 0.0006,
+ "loss": 4.289131164550781,
+ "step": 5130
+ },
+ {
+ "epoch": 71.26550218340611,
+ "grad_norm": 0.013717238791286945,
+ "learning_rate": 0.0006,
+ "loss": 4.2844014167785645,
+ "step": 5131
+ },
+ {
+ "epoch": 71.27947598253274,
+ "grad_norm": 0.0135411461815238,
+ "learning_rate": 0.0006,
+ "loss": 4.139082908630371,
+ "step": 5132
+ },
+ {
+ "epoch": 71.29344978165939,
+ "grad_norm": 0.01406062301248312,
+ "learning_rate": 0.0006,
+ "loss": 4.226719856262207,
+ "step": 5133
+ },
+ {
+ "epoch": 71.30742358078602,
+ "grad_norm": 0.014027190394699574,
+ "learning_rate": 0.0006,
+ "loss": 4.191699981689453,
+ "step": 5134
+ },
+ {
+ "epoch": 71.32139737991267,
+ "grad_norm": 0.013809907250106335,
+ "learning_rate": 0.0006,
+ "loss": 4.242542266845703,
+ "step": 5135
+ },
+ {
+ "epoch": 71.3353711790393,
+ "grad_norm": 0.013507804833352566,
+ "learning_rate": 0.0006,
+ "loss": 4.068206310272217,
+ "step": 5136
+ },
+ {
+ "epoch": 71.34934497816595,
+ "grad_norm": 0.01432533748447895,
+ "learning_rate": 0.0006,
+ "loss": 4.378222942352295,
+ "step": 5137
+ },
+ {
+ "epoch": 71.36331877729258,
+ "grad_norm": 0.01753002591431141,
+ "learning_rate": 0.0006,
+ "loss": 4.174706935882568,
+ "step": 5138
+ },
+ {
+ "epoch": 71.37729257641921,
+ "grad_norm": 0.020048344507813454,
+ "learning_rate": 0.0006,
+ "loss": 4.19920015335083,
+ "step": 5139
+ },
+ {
+ "epoch": 71.39126637554585,
+ "grad_norm": 0.018551401793956757,
+ "learning_rate": 0.0006,
+ "loss": 4.175182342529297,
+ "step": 5140
+ },
+ {
+ "epoch": 71.40524017467249,
+ "grad_norm": 0.01598959229886532,
+ "learning_rate": 0.0006,
+ "loss": 4.2305169105529785,
+ "step": 5141
+ },
+ {
+ "epoch": 71.41921397379913,
+ "grad_norm": 0.016029560938477516,
+ "learning_rate": 0.0006,
+ "loss": 4.230090618133545,
+ "step": 5142
+ },
+ {
+ "epoch": 71.43318777292576,
+ "grad_norm": 0.01595657505095005,
+ "learning_rate": 0.0006,
+ "loss": 4.138402938842773,
+ "step": 5143
+ },
+ {
+ "epoch": 71.44716157205241,
+ "grad_norm": 0.01623230241239071,
+ "learning_rate": 0.0006,
+ "loss": 4.196106433868408,
+ "step": 5144
+ },
+ {
+ "epoch": 71.46113537117904,
+ "grad_norm": 0.014057597145438194,
+ "learning_rate": 0.0006,
+ "loss": 4.2599897384643555,
+ "step": 5145
+ },
+ {
+ "epoch": 71.47510917030567,
+ "grad_norm": 0.017161158844828606,
+ "learning_rate": 0.0006,
+ "loss": 4.183909893035889,
+ "step": 5146
+ },
+ {
+ "epoch": 71.48908296943232,
+ "grad_norm": 0.017038581892848015,
+ "learning_rate": 0.0006,
+ "loss": 4.373933792114258,
+ "step": 5147
+ },
+ {
+ "epoch": 71.50305676855895,
+ "grad_norm": 0.015851367264986038,
+ "learning_rate": 0.0006,
+ "loss": 4.2620086669921875,
+ "step": 5148
+ },
+ {
+ "epoch": 71.5170305676856,
+ "grad_norm": 0.014598303474485874,
+ "learning_rate": 0.0006,
+ "loss": 4.197151184082031,
+ "step": 5149
+ },
+ {
+ "epoch": 71.53100436681223,
+ "grad_norm": 0.01339802611619234,
+ "learning_rate": 0.0006,
+ "loss": 4.110706329345703,
+ "step": 5150
+ },
+ {
+ "epoch": 71.54497816593886,
+ "grad_norm": 0.015567775815725327,
+ "learning_rate": 0.0006,
+ "loss": 4.1210246086120605,
+ "step": 5151
+ },
+ {
+ "epoch": 71.5589519650655,
+ "grad_norm": 0.015616007149219513,
+ "learning_rate": 0.0006,
+ "loss": 4.129796028137207,
+ "step": 5152
+ },
+ {
+ "epoch": 71.57292576419214,
+ "grad_norm": 0.014198964461684227,
+ "learning_rate": 0.0006,
+ "loss": 4.193232536315918,
+ "step": 5153
+ },
+ {
+ "epoch": 71.58689956331878,
+ "grad_norm": 0.015226383693516254,
+ "learning_rate": 0.0006,
+ "loss": 4.223779678344727,
+ "step": 5154
+ },
+ {
+ "epoch": 71.60087336244541,
+ "grad_norm": 0.015610109083354473,
+ "learning_rate": 0.0006,
+ "loss": 4.223179817199707,
+ "step": 5155
+ },
+ {
+ "epoch": 71.61484716157206,
+ "grad_norm": 0.016034303233027458,
+ "learning_rate": 0.0006,
+ "loss": 4.196593284606934,
+ "step": 5156
+ },
+ {
+ "epoch": 71.62882096069869,
+ "grad_norm": 0.015648581087589264,
+ "learning_rate": 0.0006,
+ "loss": 4.341497421264648,
+ "step": 5157
+ },
+ {
+ "epoch": 71.64279475982532,
+ "grad_norm": 0.014117540791630745,
+ "learning_rate": 0.0006,
+ "loss": 4.194307327270508,
+ "step": 5158
+ },
+ {
+ "epoch": 71.65676855895197,
+ "grad_norm": 0.015612686984241009,
+ "learning_rate": 0.0006,
+ "loss": 4.290152549743652,
+ "step": 5159
+ },
+ {
+ "epoch": 71.6707423580786,
+ "grad_norm": 0.01722259446978569,
+ "learning_rate": 0.0006,
+ "loss": 4.258018493652344,
+ "step": 5160
+ },
+ {
+ "epoch": 71.68471615720524,
+ "grad_norm": 0.015923351049423218,
+ "learning_rate": 0.0006,
+ "loss": 4.21684455871582,
+ "step": 5161
+ },
+ {
+ "epoch": 71.69868995633188,
+ "grad_norm": 0.013663450255990028,
+ "learning_rate": 0.0006,
+ "loss": 4.175601959228516,
+ "step": 5162
+ },
+ {
+ "epoch": 71.71266375545852,
+ "grad_norm": 0.017075596377253532,
+ "learning_rate": 0.0006,
+ "loss": 4.205138683319092,
+ "step": 5163
+ },
+ {
+ "epoch": 71.72663755458515,
+ "grad_norm": 0.01723487675189972,
+ "learning_rate": 0.0006,
+ "loss": 4.08839225769043,
+ "step": 5164
+ },
+ {
+ "epoch": 71.74061135371178,
+ "grad_norm": 0.017355505377054214,
+ "learning_rate": 0.0006,
+ "loss": 4.165130615234375,
+ "step": 5165
+ },
+ {
+ "epoch": 71.75458515283843,
+ "grad_norm": 0.01868261583149433,
+ "learning_rate": 0.0006,
+ "loss": 4.16977071762085,
+ "step": 5166
+ },
+ {
+ "epoch": 71.76855895196506,
+ "grad_norm": 0.01897321827709675,
+ "learning_rate": 0.0006,
+ "loss": 4.173497200012207,
+ "step": 5167
+ },
+ {
+ "epoch": 71.78253275109171,
+ "grad_norm": 0.02045825682580471,
+ "learning_rate": 0.0006,
+ "loss": 4.176192283630371,
+ "step": 5168
+ },
+ {
+ "epoch": 71.79650655021834,
+ "grad_norm": 0.022895624861121178,
+ "learning_rate": 0.0006,
+ "loss": 4.163009166717529,
+ "step": 5169
+ },
+ {
+ "epoch": 71.81048034934499,
+ "grad_norm": 0.02062244527041912,
+ "learning_rate": 0.0006,
+ "loss": 4.156868934631348,
+ "step": 5170
+ },
+ {
+ "epoch": 71.82445414847162,
+ "grad_norm": 0.0176058579236269,
+ "learning_rate": 0.0006,
+ "loss": 4.154374122619629,
+ "step": 5171
+ },
+ {
+ "epoch": 71.83842794759825,
+ "grad_norm": 0.018049508333206177,
+ "learning_rate": 0.0006,
+ "loss": 4.296828269958496,
+ "step": 5172
+ },
+ {
+ "epoch": 71.8524017467249,
+ "grad_norm": 0.019466130062937737,
+ "learning_rate": 0.0006,
+ "loss": 4.19450569152832,
+ "step": 5173
+ },
+ {
+ "epoch": 71.86637554585153,
+ "grad_norm": 0.01838165521621704,
+ "learning_rate": 0.0006,
+ "loss": 4.11812686920166,
+ "step": 5174
+ },
+ {
+ "epoch": 71.88034934497817,
+ "grad_norm": 0.01608051173388958,
+ "learning_rate": 0.0006,
+ "loss": 4.240464210510254,
+ "step": 5175
+ },
+ {
+ "epoch": 71.8943231441048,
+ "grad_norm": 0.015807967633008957,
+ "learning_rate": 0.0006,
+ "loss": 4.275581359863281,
+ "step": 5176
+ },
+ {
+ "epoch": 71.90829694323143,
+ "grad_norm": 0.017366575077176094,
+ "learning_rate": 0.0006,
+ "loss": 4.064750671386719,
+ "step": 5177
+ },
+ {
+ "epoch": 71.92227074235808,
+ "grad_norm": 0.016241202130913734,
+ "learning_rate": 0.0006,
+ "loss": 4.203697681427002,
+ "step": 5178
+ },
+ {
+ "epoch": 71.93624454148471,
+ "grad_norm": 0.019105732440948486,
+ "learning_rate": 0.0006,
+ "loss": 4.140021800994873,
+ "step": 5179
+ },
+ {
+ "epoch": 71.95021834061136,
+ "grad_norm": 0.018492674455046654,
+ "learning_rate": 0.0006,
+ "loss": 4.2534074783325195,
+ "step": 5180
+ },
+ {
+ "epoch": 71.96419213973799,
+ "grad_norm": 0.019301170483231544,
+ "learning_rate": 0.0006,
+ "loss": 4.381392478942871,
+ "step": 5181
+ },
+ {
+ "epoch": 71.97816593886463,
+ "grad_norm": 0.019571911543607712,
+ "learning_rate": 0.0006,
+ "loss": 4.131056785583496,
+ "step": 5182
+ },
+ {
+ "epoch": 71.99213973799127,
+ "grad_norm": 0.01859748177230358,
+ "learning_rate": 0.0006,
+ "loss": 4.278930187225342,
+ "step": 5183
+ },
+ {
+ "epoch": 72.0,
+ "grad_norm": 0.02017994411289692,
+ "learning_rate": 0.0006,
+ "loss": 4.116464614868164,
+ "step": 5184
+ },
+ {
+ "epoch": 72.0,
+ "eval_loss": 4.60784912109375,
+ "eval_runtime": 56.3242,
+ "eval_samples_per_second": 43.356,
+ "eval_steps_per_second": 1.367,
+ "step": 5184
+ },
+ {
+ "epoch": 72.01397379912663,
+ "grad_norm": 0.030090296640992165,
+ "learning_rate": 0.0006,
+ "loss": 4.218588352203369,
+ "step": 5185
+ },
+ {
+ "epoch": 72.02794759825328,
+ "grad_norm": 1.330443024635315,
+ "learning_rate": 0.0006,
+ "loss": 4.417870044708252,
+ "step": 5186
+ },
+ {
+ "epoch": 72.04192139737991,
+ "grad_norm": 0.6991518139839172,
+ "learning_rate": 0.0006,
+ "loss": 8.414234161376953,
+ "step": 5187
+ },
+ {
+ "epoch": 72.05589519650655,
+ "grad_norm": 0.2631438374519348,
+ "learning_rate": 0.0006,
+ "loss": 7.41729736328125,
+ "step": 5188
+ },
+ {
+ "epoch": 72.06986899563319,
+ "grad_norm": 0.12391754984855652,
+ "learning_rate": 0.0006,
+ "loss": 7.209164619445801,
+ "step": 5189
+ },
+ {
+ "epoch": 72.08384279475983,
+ "grad_norm": 0.13519170880317688,
+ "learning_rate": 0.0006,
+ "loss": 6.943539619445801,
+ "step": 5190
+ },
+ {
+ "epoch": 72.09781659388646,
+ "grad_norm": 1.258439302444458,
+ "learning_rate": 0.0006,
+ "loss": 10.267714500427246,
+ "step": 5191
+ },
+ {
+ "epoch": 72.1117903930131,
+ "grad_norm": 1.0296908617019653,
+ "learning_rate": 0.0006,
+ "loss": 9.51768970489502,
+ "step": 5192
+ },
+ {
+ "epoch": 72.12576419213974,
+ "grad_norm": 0.12273016571998596,
+ "learning_rate": 0.0006,
+ "loss": 7.634477615356445,
+ "step": 5193
+ },
+ {
+ "epoch": 72.13973799126637,
+ "grad_norm": 0.06170198693871498,
+ "learning_rate": 0.0006,
+ "loss": 7.592115879058838,
+ "step": 5194
+ },
+ {
+ "epoch": 72.15371179039302,
+ "grad_norm": 0.06295108795166016,
+ "learning_rate": 0.0006,
+ "loss": 7.576314449310303,
+ "step": 5195
+ },
+ {
+ "epoch": 72.16768558951965,
+ "grad_norm": 0.0666993111371994,
+ "learning_rate": 0.0006,
+ "loss": 7.53862190246582,
+ "step": 5196
+ },
+ {
+ "epoch": 72.18165938864628,
+ "grad_norm": 0.052215222269296646,
+ "learning_rate": 0.0006,
+ "loss": 7.502930641174316,
+ "step": 5197
+ },
+ {
+ "epoch": 72.19563318777293,
+ "grad_norm": 0.04978961497545242,
+ "learning_rate": 0.0006,
+ "loss": 7.471630096435547,
+ "step": 5198
+ },
+ {
+ "epoch": 72.20960698689956,
+ "grad_norm": 0.07275046408176422,
+ "learning_rate": 0.0006,
+ "loss": 7.496924877166748,
+ "step": 5199
+ },
+ {
+ "epoch": 72.2235807860262,
+ "grad_norm": 0.04491819068789482,
+ "learning_rate": 0.0006,
+ "loss": 7.424494743347168,
+ "step": 5200
+ },
+ {
+ "epoch": 72.23755458515284,
+ "grad_norm": 0.043718695640563965,
+ "learning_rate": 0.0006,
+ "loss": 7.441090106964111,
+ "step": 5201
+ },
+ {
+ "epoch": 72.25152838427948,
+ "grad_norm": 0.04483497887849808,
+ "learning_rate": 0.0006,
+ "loss": 7.3753862380981445,
+ "step": 5202
+ },
+ {
+ "epoch": 72.26550218340611,
+ "grad_norm": 0.03463944420218468,
+ "learning_rate": 0.0006,
+ "loss": 7.375782012939453,
+ "step": 5203
+ },
+ {
+ "epoch": 72.27947598253274,
+ "grad_norm": 0.0397045724093914,
+ "learning_rate": 0.0006,
+ "loss": 7.358898639678955,
+ "step": 5204
+ },
+ {
+ "epoch": 72.29344978165939,
+ "grad_norm": 0.029339438304305077,
+ "learning_rate": 0.0006,
+ "loss": 7.3774871826171875,
+ "step": 5205
+ },
+ {
+ "epoch": 72.30742358078602,
+ "grad_norm": 0.025469699874520302,
+ "learning_rate": 0.0006,
+ "loss": 7.324203014373779,
+ "step": 5206
+ },
+ {
+ "epoch": 72.32139737991267,
+ "grad_norm": 0.030497785657644272,
+ "learning_rate": 0.0006,
+ "loss": 7.329351902008057,
+ "step": 5207
+ },
+ {
+ "epoch": 72.3353711790393,
+ "grad_norm": 0.03382078558206558,
+ "learning_rate": 0.0006,
+ "loss": 7.289940357208252,
+ "step": 5208
+ },
+ {
+ "epoch": 72.34934497816595,
+ "grad_norm": 0.019869405776262283,
+ "learning_rate": 0.0006,
+ "loss": 7.272582054138184,
+ "step": 5209
+ },
+ {
+ "epoch": 72.36331877729258,
+ "grad_norm": 0.02385753206908703,
+ "learning_rate": 0.0006,
+ "loss": 7.2654266357421875,
+ "step": 5210
+ },
+ {
+ "epoch": 72.37729257641921,
+ "grad_norm": 0.02794312685728073,
+ "learning_rate": 0.0006,
+ "loss": 7.250753402709961,
+ "step": 5211
+ },
+ {
+ "epoch": 72.39126637554585,
+ "grad_norm": 0.02445952221751213,
+ "learning_rate": 0.0006,
+ "loss": 7.293970108032227,
+ "step": 5212
+ },
+ {
+ "epoch": 72.40524017467249,
+ "grad_norm": 0.019660986959934235,
+ "learning_rate": 0.0006,
+ "loss": 7.1948771476745605,
+ "step": 5213
+ },
+ {
+ "epoch": 72.41921397379913,
+ "grad_norm": 0.024110861122608185,
+ "learning_rate": 0.0006,
+ "loss": 7.224499225616455,
+ "step": 5214
+ },
+ {
+ "epoch": 72.43318777292576,
+ "grad_norm": 0.021630609408020973,
+ "learning_rate": 0.0006,
+ "loss": 7.1736884117126465,
+ "step": 5215
+ },
+ {
+ "epoch": 72.44716157205241,
+ "grad_norm": 0.02185880020260811,
+ "learning_rate": 0.0006,
+ "loss": 7.167229652404785,
+ "step": 5216
+ },
+ {
+ "epoch": 72.46113537117904,
+ "grad_norm": 0.018913568928837776,
+ "learning_rate": 0.0006,
+ "loss": 7.196551322937012,
+ "step": 5217
+ },
+ {
+ "epoch": 72.47510917030567,
+ "grad_norm": 0.01802486553788185,
+ "learning_rate": 0.0006,
+ "loss": 7.144286632537842,
+ "step": 5218
+ },
+ {
+ "epoch": 72.48908296943232,
+ "grad_norm": 0.020386485382914543,
+ "learning_rate": 0.0006,
+ "loss": 7.196260929107666,
+ "step": 5219
+ },
+ {
+ "epoch": 72.50305676855895,
+ "grad_norm": 0.020077696070075035,
+ "learning_rate": 0.0006,
+ "loss": 7.138718605041504,
+ "step": 5220
+ },
+ {
+ "epoch": 72.5170305676856,
+ "grad_norm": 0.017224183306097984,
+ "learning_rate": 0.0006,
+ "loss": 7.119478702545166,
+ "step": 5221
+ },
+ {
+ "epoch": 72.53100436681223,
+ "grad_norm": 0.020847434177994728,
+ "learning_rate": 0.0006,
+ "loss": 7.068922996520996,
+ "step": 5222
+ },
+ {
+ "epoch": 72.54497816593886,
+ "grad_norm": 0.021764913573861122,
+ "learning_rate": 0.0006,
+ "loss": 7.071352958679199,
+ "step": 5223
+ },
+ {
+ "epoch": 72.5589519650655,
+ "grad_norm": 0.02697215974330902,
+ "learning_rate": 0.0006,
+ "loss": 7.033344268798828,
+ "step": 5224
+ },
+ {
+ "epoch": 72.57292576419214,
+ "grad_norm": 0.042392998933792114,
+ "learning_rate": 0.0006,
+ "loss": 7.0589919090271,
+ "step": 5225
+ },
+ {
+ "epoch": 72.58689956331878,
+ "grad_norm": 0.06616316735744476,
+ "learning_rate": 0.0006,
+ "loss": 7.060244083404541,
+ "step": 5226
+ },
+ {
+ "epoch": 72.60087336244541,
+ "grad_norm": 0.0828910544514656,
+ "learning_rate": 0.0006,
+ "loss": 7.057173728942871,
+ "step": 5227
+ },
+ {
+ "epoch": 72.61484716157206,
+ "grad_norm": 0.046767015010118484,
+ "learning_rate": 0.0006,
+ "loss": 7.010563850402832,
+ "step": 5228
+ },
+ {
+ "epoch": 72.62882096069869,
+ "grad_norm": 0.03831760957837105,
+ "learning_rate": 0.0006,
+ "loss": 6.9028825759887695,
+ "step": 5229
+ },
+ {
+ "epoch": 72.64279475982532,
+ "grad_norm": 0.05109567195177078,
+ "learning_rate": 0.0006,
+ "loss": 6.923257827758789,
+ "step": 5230
+ },
+ {
+ "epoch": 72.65676855895197,
+ "grad_norm": 0.02897934429347515,
+ "learning_rate": 0.0006,
+ "loss": 6.823014259338379,
+ "step": 5231
+ },
+ {
+ "epoch": 72.6707423580786,
+ "grad_norm": 0.03769192099571228,
+ "learning_rate": 0.0006,
+ "loss": 6.87416410446167,
+ "step": 5232
+ },
+ {
+ "epoch": 72.68471615720524,
+ "grad_norm": 0.03677637130022049,
+ "learning_rate": 0.0006,
+ "loss": 6.7649030685424805,
+ "step": 5233
+ },
+ {
+ "epoch": 72.69868995633188,
+ "grad_norm": 0.0371948778629303,
+ "learning_rate": 0.0006,
+ "loss": 6.749821662902832,
+ "step": 5234
+ },
+ {
+ "epoch": 72.71266375545852,
+ "grad_norm": 0.03057795763015747,
+ "learning_rate": 0.0006,
+ "loss": 6.748945236206055,
+ "step": 5235
+ },
+ {
+ "epoch": 72.72663755458515,
+ "grad_norm": 0.028109321370720863,
+ "learning_rate": 0.0006,
+ "loss": 6.663548946380615,
+ "step": 5236
+ },
+ {
+ "epoch": 72.74061135371178,
+ "grad_norm": 0.02442420832812786,
+ "learning_rate": 0.0006,
+ "loss": 6.636251449584961,
+ "step": 5237
+ },
+ {
+ "epoch": 72.75458515283843,
+ "grad_norm": 0.031841423362493515,
+ "learning_rate": 0.0006,
+ "loss": 6.606883525848389,
+ "step": 5238
+ },
+ {
+ "epoch": 72.76855895196506,
+ "grad_norm": 0.028100458905100822,
+ "learning_rate": 0.0006,
+ "loss": 6.603745937347412,
+ "step": 5239
+ },
+ {
+ "epoch": 72.78253275109171,
+ "grad_norm": 0.03222901374101639,
+ "learning_rate": 0.0006,
+ "loss": 6.463825225830078,
+ "step": 5240
+ },
+ {
+ "epoch": 72.79650655021834,
+ "grad_norm": 0.04766521975398064,
+ "learning_rate": 0.0006,
+ "loss": 6.43734073638916,
+ "step": 5241
+ },
+ {
+ "epoch": 72.81048034934499,
+ "grad_norm": 0.09058783948421478,
+ "learning_rate": 0.0006,
+ "loss": 6.448520183563232,
+ "step": 5242
+ },
+ {
+ "epoch": 72.82445414847162,
+ "grad_norm": 0.08995158225297928,
+ "learning_rate": 0.0006,
+ "loss": 6.522093296051025,
+ "step": 5243
+ },
+ {
+ "epoch": 72.83842794759825,
+ "grad_norm": 0.05491040274500847,
+ "learning_rate": 0.0006,
+ "loss": 6.37354040145874,
+ "step": 5244
+ },
+ {
+ "epoch": 72.8524017467249,
+ "grad_norm": 0.05198514088988304,
+ "learning_rate": 0.0006,
+ "loss": 6.362032413482666,
+ "step": 5245
+ },
+ {
+ "epoch": 72.86637554585153,
+ "grad_norm": 0.05464765429496765,
+ "learning_rate": 0.0006,
+ "loss": 6.326773643493652,
+ "step": 5246
+ },
+ {
+ "epoch": 72.88034934497817,
+ "grad_norm": 0.07337361574172974,
+ "learning_rate": 0.0006,
+ "loss": 6.297532081604004,
+ "step": 5247
+ },
+ {
+ "epoch": 72.8943231441048,
+ "grad_norm": 0.10061230510473251,
+ "learning_rate": 0.0006,
+ "loss": 6.359557628631592,
+ "step": 5248
+ },
+ {
+ "epoch": 72.90829694323143,
+ "grad_norm": 0.10005959123373032,
+ "learning_rate": 0.0006,
+ "loss": 6.422270774841309,
+ "step": 5249
+ },
+ {
+ "epoch": 72.92227074235808,
+ "grad_norm": 0.08989301323890686,
+ "learning_rate": 0.0006,
+ "loss": 6.289371490478516,
+ "step": 5250
+ },
+ {
+ "epoch": 72.93624454148471,
+ "grad_norm": 0.06371799111366272,
+ "learning_rate": 0.0006,
+ "loss": 6.199618339538574,
+ "step": 5251
+ },
+ {
+ "epoch": 72.95021834061136,
+ "grad_norm": 0.05932944267988205,
+ "learning_rate": 0.0006,
+ "loss": 6.111152648925781,
+ "step": 5252
+ },
+ {
+ "epoch": 72.96419213973799,
+ "grad_norm": 0.04735502973198891,
+ "learning_rate": 0.0006,
+ "loss": 6.026026248931885,
+ "step": 5253
+ },
+ {
+ "epoch": 72.97816593886463,
+ "grad_norm": 0.04270688816905022,
+ "learning_rate": 0.0006,
+ "loss": 6.01606559753418,
+ "step": 5254
+ },
+ {
+ "epoch": 72.99213973799127,
+ "grad_norm": 0.03774143010377884,
+ "learning_rate": 0.0006,
+ "loss": 5.8935441970825195,
+ "step": 5255
+ },
+ {
+ "epoch": 73.0,
+ "grad_norm": 0.04313631355762482,
+ "learning_rate": 0.0006,
+ "loss": 5.842061996459961,
+ "step": 5256
+ },
+ {
+ "epoch": 73.0,
+ "eval_loss": 5.9985785484313965,
+ "eval_runtime": 56.3891,
+ "eval_samples_per_second": 43.306,
+ "eval_steps_per_second": 1.366,
+ "step": 5256
+ },
+ {
+ "epoch": 73.01397379912663,
+ "grad_norm": 0.045038655400276184,
+ "learning_rate": 0.0006,
+ "loss": 5.8209662437438965,
+ "step": 5257
+ },
+ {
+ "epoch": 73.02794759825328,
+ "grad_norm": 0.03706910088658333,
+ "learning_rate": 0.0006,
+ "loss": 5.806189060211182,
+ "step": 5258
+ },
+ {
+ "epoch": 73.04192139737991,
+ "grad_norm": 0.04051685333251953,
+ "learning_rate": 0.0006,
+ "loss": 5.706869125366211,
+ "step": 5259
+ },
+ {
+ "epoch": 73.05589519650655,
+ "grad_norm": 0.04849153012037277,
+ "learning_rate": 0.0006,
+ "loss": 5.696385383605957,
+ "step": 5260
+ },
+ {
+ "epoch": 73.06986899563319,
+ "grad_norm": 0.06124403700232506,
+ "learning_rate": 0.0006,
+ "loss": 5.618223667144775,
+ "step": 5261
+ },
+ {
+ "epoch": 73.08384279475983,
+ "grad_norm": 0.07375937700271606,
+ "learning_rate": 0.0006,
+ "loss": 5.631045341491699,
+ "step": 5262
+ },
+ {
+ "epoch": 73.09781659388646,
+ "grad_norm": 0.07413844019174576,
+ "learning_rate": 0.0006,
+ "loss": 5.6532182693481445,
+ "step": 5263
+ },
+ {
+ "epoch": 73.1117903930131,
+ "grad_norm": 0.0633770152926445,
+ "learning_rate": 0.0006,
+ "loss": 5.57209587097168,
+ "step": 5264
+ },
+ {
+ "epoch": 73.12576419213974,
+ "grad_norm": 0.05436766892671585,
+ "learning_rate": 0.0006,
+ "loss": 5.53944730758667,
+ "step": 5265
+ },
+ {
+ "epoch": 73.13973799126637,
+ "grad_norm": 0.04639481008052826,
+ "learning_rate": 0.0006,
+ "loss": 5.502948760986328,
+ "step": 5266
+ },
+ {
+ "epoch": 73.15371179039302,
+ "grad_norm": 0.07556558400392532,
+ "learning_rate": 0.0006,
+ "loss": 5.429069995880127,
+ "step": 5267
+ },
+ {
+ "epoch": 73.16768558951965,
+ "grad_norm": 0.1127929836511612,
+ "learning_rate": 0.0006,
+ "loss": 5.545061111450195,
+ "step": 5268
+ },
+ {
+ "epoch": 73.18165938864628,
+ "grad_norm": 0.07620745897293091,
+ "learning_rate": 0.0006,
+ "loss": 5.507102966308594,
+ "step": 5269
+ },
+ {
+ "epoch": 73.19563318777293,
+ "grad_norm": 0.05076538026332855,
+ "learning_rate": 0.0006,
+ "loss": 5.329682350158691,
+ "step": 5270
+ },
+ {
+ "epoch": 73.20960698689956,
+ "grad_norm": 0.05407547950744629,
+ "learning_rate": 0.0006,
+ "loss": 5.3048553466796875,
+ "step": 5271
+ },
+ {
+ "epoch": 73.2235807860262,
+ "grad_norm": 0.059095583856105804,
+ "learning_rate": 0.0006,
+ "loss": 5.292568683624268,
+ "step": 5272
+ },
+ {
+ "epoch": 73.23755458515284,
+ "grad_norm": 0.06193305924534798,
+ "learning_rate": 0.0006,
+ "loss": 5.1870927810668945,
+ "step": 5273
+ },
+ {
+ "epoch": 73.25152838427948,
+ "grad_norm": 0.062261562794446945,
+ "learning_rate": 0.0006,
+ "loss": 5.256819248199463,
+ "step": 5274
+ },
+ {
+ "epoch": 73.26550218340611,
+ "grad_norm": 0.0732516348361969,
+ "learning_rate": 0.0006,
+ "loss": 5.256464004516602,
+ "step": 5275
+ },
+ {
+ "epoch": 73.27947598253274,
+ "grad_norm": 0.12177979946136475,
+ "learning_rate": 0.0006,
+ "loss": 5.284239768981934,
+ "step": 5276
+ },
+ {
+ "epoch": 73.29344978165939,
+ "grad_norm": 0.10459653288125992,
+ "learning_rate": 0.0006,
+ "loss": 5.269713401794434,
+ "step": 5277
+ },
+ {
+ "epoch": 73.30742358078602,
+ "grad_norm": 0.07854694873094559,
+ "learning_rate": 0.0006,
+ "loss": 5.3245391845703125,
+ "step": 5278
+ },
+ {
+ "epoch": 73.32139737991267,
+ "grad_norm": 0.06030094623565674,
+ "learning_rate": 0.0006,
+ "loss": 5.033702850341797,
+ "step": 5279
+ },
+ {
+ "epoch": 73.3353711790393,
+ "grad_norm": 0.05168461427092552,
+ "learning_rate": 0.0006,
+ "loss": 5.140530109405518,
+ "step": 5280
+ },
+ {
+ "epoch": 73.34934497816595,
+ "grad_norm": 0.041064050048589706,
+ "learning_rate": 0.0006,
+ "loss": 5.1085896492004395,
+ "step": 5281
+ },
+ {
+ "epoch": 73.36331877729258,
+ "grad_norm": 0.03788219019770622,
+ "learning_rate": 0.0006,
+ "loss": 5.045009613037109,
+ "step": 5282
+ },
+ {
+ "epoch": 73.37729257641921,
+ "grad_norm": 0.03851071000099182,
+ "learning_rate": 0.0006,
+ "loss": 5.055008888244629,
+ "step": 5283
+ },
+ {
+ "epoch": 73.39126637554585,
+ "grad_norm": 0.03448539599776268,
+ "learning_rate": 0.0006,
+ "loss": 5.10657262802124,
+ "step": 5284
+ },
+ {
+ "epoch": 73.40524017467249,
+ "grad_norm": 0.026618091389536858,
+ "learning_rate": 0.0006,
+ "loss": 5.0342912673950195,
+ "step": 5285
+ },
+ {
+ "epoch": 73.41921397379913,
+ "grad_norm": 0.03060489147901535,
+ "learning_rate": 0.0006,
+ "loss": 5.022700786590576,
+ "step": 5286
+ },
+ {
+ "epoch": 73.43318777292576,
+ "grad_norm": 0.03362000361084938,
+ "learning_rate": 0.0006,
+ "loss": 4.891584396362305,
+ "step": 5287
+ },
+ {
+ "epoch": 73.44716157205241,
+ "grad_norm": 0.03887505829334259,
+ "learning_rate": 0.0006,
+ "loss": 4.772258758544922,
+ "step": 5288
+ },
+ {
+ "epoch": 73.46113537117904,
+ "grad_norm": 0.07157225161790848,
+ "learning_rate": 0.0006,
+ "loss": 4.997142791748047,
+ "step": 5289
+ },
+ {
+ "epoch": 73.47510917030567,
+ "grad_norm": 0.10180624574422836,
+ "learning_rate": 0.0006,
+ "loss": 4.908722877502441,
+ "step": 5290
+ },
+ {
+ "epoch": 73.48908296943232,
+ "grad_norm": 0.06529122591018677,
+ "learning_rate": 0.0006,
+ "loss": 4.98001766204834,
+ "step": 5291
+ },
+ {
+ "epoch": 73.50305676855895,
+ "grad_norm": 0.055171865969896317,
+ "learning_rate": 0.0006,
+ "loss": 4.877438068389893,
+ "step": 5292
+ },
+ {
+ "epoch": 73.5170305676856,
+ "grad_norm": 0.0318770632147789,
+ "learning_rate": 0.0006,
+ "loss": 4.836934566497803,
+ "step": 5293
+ },
+ {
+ "epoch": 73.53100436681223,
+ "grad_norm": 0.042072564363479614,
+ "learning_rate": 0.0006,
+ "loss": 4.919489860534668,
+ "step": 5294
+ },
+ {
+ "epoch": 73.54497816593886,
+ "grad_norm": 0.029197754338383675,
+ "learning_rate": 0.0006,
+ "loss": 4.805579662322998,
+ "step": 5295
+ },
+ {
+ "epoch": 73.5589519650655,
+ "grad_norm": 0.03249497339129448,
+ "learning_rate": 0.0006,
+ "loss": 4.704806327819824,
+ "step": 5296
+ },
+ {
+ "epoch": 73.57292576419214,
+ "grad_norm": 0.025050945580005646,
+ "learning_rate": 0.0006,
+ "loss": 4.936512470245361,
+ "step": 5297
+ },
+ {
+ "epoch": 73.58689956331878,
+ "grad_norm": 0.029073381796479225,
+ "learning_rate": 0.0006,
+ "loss": 4.757592678070068,
+ "step": 5298
+ },
+ {
+ "epoch": 73.60087336244541,
+ "grad_norm": 0.03465264290571213,
+ "learning_rate": 0.0006,
+ "loss": 4.729528427124023,
+ "step": 5299
+ },
+ {
+ "epoch": 73.61484716157206,
+ "grad_norm": 0.054795749485492706,
+ "learning_rate": 0.0006,
+ "loss": 4.77252197265625,
+ "step": 5300
+ },
+ {
+ "epoch": 73.62882096069869,
+ "grad_norm": 0.18129196763038635,
+ "learning_rate": 0.0006,
+ "loss": 4.901700019836426,
+ "step": 5301
+ },
+ {
+ "epoch": 73.64279475982532,
+ "grad_norm": 0.23274236917495728,
+ "learning_rate": 0.0006,
+ "loss": 5.207432270050049,
+ "step": 5302
+ },
+ {
+ "epoch": 73.65676855895197,
+ "grad_norm": 0.18162772059440613,
+ "learning_rate": 0.0006,
+ "loss": 5.326632976531982,
+ "step": 5303
+ },
+ {
+ "epoch": 73.6707423580786,
+ "grad_norm": 0.11858413368463516,
+ "learning_rate": 0.0006,
+ "loss": 5.3730058670043945,
+ "step": 5304
+ },
+ {
+ "epoch": 73.68471615720524,
+ "grad_norm": 0.06031268090009689,
+ "learning_rate": 0.0006,
+ "loss": 5.330936431884766,
+ "step": 5305
+ },
+ {
+ "epoch": 73.69868995633188,
+ "grad_norm": 0.059304628521203995,
+ "learning_rate": 0.0006,
+ "loss": 5.280514717102051,
+ "step": 5306
+ },
+ {
+ "epoch": 73.71266375545852,
+ "grad_norm": 0.04702131822705269,
+ "learning_rate": 0.0006,
+ "loss": 5.288274765014648,
+ "step": 5307
+ },
+ {
+ "epoch": 73.72663755458515,
+ "grad_norm": 0.037792228162288666,
+ "learning_rate": 0.0006,
+ "loss": 5.173289775848389,
+ "step": 5308
+ },
+ {
+ "epoch": 73.74061135371178,
+ "grad_norm": 0.037763576954603195,
+ "learning_rate": 0.0006,
+ "loss": 5.084990978240967,
+ "step": 5309
+ },
+ {
+ "epoch": 73.75458515283843,
+ "grad_norm": 0.03337346389889717,
+ "learning_rate": 0.0006,
+ "loss": 5.121417045593262,
+ "step": 5310
+ },
+ {
+ "epoch": 73.76855895196506,
+ "grad_norm": 0.03184802085161209,
+ "learning_rate": 0.0006,
+ "loss": 5.208918571472168,
+ "step": 5311
+ },
+ {
+ "epoch": 73.78253275109171,
+ "grad_norm": 0.02832048572599888,
+ "learning_rate": 0.0006,
+ "loss": 4.923543453216553,
+ "step": 5312
+ },
+ {
+ "epoch": 73.79650655021834,
+ "grad_norm": 0.028297102078795433,
+ "learning_rate": 0.0006,
+ "loss": 4.944367408752441,
+ "step": 5313
+ },
+ {
+ "epoch": 73.81048034934499,
+ "grad_norm": 0.027701659128069878,
+ "learning_rate": 0.0006,
+ "loss": 4.858168601989746,
+ "step": 5314
+ },
+ {
+ "epoch": 73.82445414847162,
+ "grad_norm": 0.02214639261364937,
+ "learning_rate": 0.0006,
+ "loss": 4.887052536010742,
+ "step": 5315
+ },
+ {
+ "epoch": 73.83842794759825,
+ "grad_norm": 0.01995093561708927,
+ "learning_rate": 0.0006,
+ "loss": 4.885797500610352,
+ "step": 5316
+ },
+ {
+ "epoch": 73.8524017467249,
+ "grad_norm": 0.020473985001444817,
+ "learning_rate": 0.0006,
+ "loss": 4.917325019836426,
+ "step": 5317
+ },
+ {
+ "epoch": 73.86637554585153,
+ "grad_norm": 0.0192868635058403,
+ "learning_rate": 0.0006,
+ "loss": 4.801746845245361,
+ "step": 5318
+ },
+ {
+ "epoch": 73.88034934497817,
+ "grad_norm": 0.017108267173171043,
+ "learning_rate": 0.0006,
+ "loss": 4.893620491027832,
+ "step": 5319
+ },
+ {
+ "epoch": 73.8943231441048,
+ "grad_norm": 0.01718071475625038,
+ "learning_rate": 0.0006,
+ "loss": 4.940695285797119,
+ "step": 5320
+ },
+ {
+ "epoch": 73.90829694323143,
+ "grad_norm": 0.017821557819843292,
+ "learning_rate": 0.0006,
+ "loss": 4.767613410949707,
+ "step": 5321
+ },
+ {
+ "epoch": 73.92227074235808,
+ "grad_norm": 0.015262950211763382,
+ "learning_rate": 0.0006,
+ "loss": 4.910315036773682,
+ "step": 5322
+ },
+ {
+ "epoch": 73.93624454148471,
+ "grad_norm": 0.01769006811082363,
+ "learning_rate": 0.0006,
+ "loss": 4.7743449211120605,
+ "step": 5323
+ },
+ {
+ "epoch": 73.95021834061136,
+ "grad_norm": 0.014460788108408451,
+ "learning_rate": 0.0006,
+ "loss": 4.743494987487793,
+ "step": 5324
+ },
+ {
+ "epoch": 73.96419213973799,
+ "grad_norm": 0.01403993833810091,
+ "learning_rate": 0.0006,
+ "loss": 4.666640281677246,
+ "step": 5325
+ },
+ {
+ "epoch": 73.97816593886463,
+ "grad_norm": 0.013997809961438179,
+ "learning_rate": 0.0006,
+ "loss": 4.698747158050537,
+ "step": 5326
+ },
+ {
+ "epoch": 73.99213973799127,
+ "grad_norm": 0.014244182966649532,
+ "learning_rate": 0.0006,
+ "loss": 4.745250701904297,
+ "step": 5327
+ },
+ {
+ "epoch": 74.0,
+ "grad_norm": 0.015144186094403267,
+ "learning_rate": 0.0006,
+ "loss": 4.667946815490723,
+ "step": 5328
+ },
+ {
+ "epoch": 74.0,
+ "eval_loss": 5.005440711975098,
+ "eval_runtime": 56.7836,
+ "eval_samples_per_second": 43.005,
+ "eval_steps_per_second": 1.356,
+ "step": 5328
+ },
+ {
+ "epoch": 74.01397379912663,
+ "grad_norm": 0.020384902134537697,
+ "learning_rate": 0.0006,
+ "loss": 4.626955509185791,
+ "step": 5329
+ },
+ {
+ "epoch": 74.02794759825328,
+ "grad_norm": 0.03624720871448517,
+ "learning_rate": 0.0006,
+ "loss": 4.713998317718506,
+ "step": 5330
+ },
+ {
+ "epoch": 74.04192139737991,
+ "grad_norm": 0.05701754242181778,
+ "learning_rate": 0.0006,
+ "loss": 4.717874526977539,
+ "step": 5331
+ },
+ {
+ "epoch": 74.05589519650655,
+ "grad_norm": 0.03942621126770973,
+ "learning_rate": 0.0006,
+ "loss": 4.7026190757751465,
+ "step": 5332
+ },
+ {
+ "epoch": 74.06986899563319,
+ "grad_norm": 0.017803272232413292,
+ "learning_rate": 0.0006,
+ "loss": 4.702548027038574,
+ "step": 5333
+ },
+ {
+ "epoch": 74.08384279475983,
+ "grad_norm": 0.017024677246809006,
+ "learning_rate": 0.0006,
+ "loss": 4.69752311706543,
+ "step": 5334
+ },
+ {
+ "epoch": 74.09781659388646,
+ "grad_norm": 0.016467854380607605,
+ "learning_rate": 0.0006,
+ "loss": 4.624876976013184,
+ "step": 5335
+ },
+ {
+ "epoch": 74.1117903930131,
+ "grad_norm": 0.016569215804338455,
+ "learning_rate": 0.0006,
+ "loss": 4.645040512084961,
+ "step": 5336
+ },
+ {
+ "epoch": 74.12576419213974,
+ "grad_norm": 0.012872067280113697,
+ "learning_rate": 0.0006,
+ "loss": 4.64582633972168,
+ "step": 5337
+ },
+ {
+ "epoch": 74.13973799126637,
+ "grad_norm": 0.015689387917518616,
+ "learning_rate": 0.0006,
+ "loss": 4.596040725708008,
+ "step": 5338
+ },
+ {
+ "epoch": 74.15371179039302,
+ "grad_norm": 0.013082082383334637,
+ "learning_rate": 0.0006,
+ "loss": 4.648033142089844,
+ "step": 5339
+ },
+ {
+ "epoch": 74.16768558951965,
+ "grad_norm": 0.013728172518312931,
+ "learning_rate": 0.0006,
+ "loss": 4.657873153686523,
+ "step": 5340
+ },
+ {
+ "epoch": 74.18165938864628,
+ "grad_norm": 0.013120594434440136,
+ "learning_rate": 0.0006,
+ "loss": 4.696074485778809,
+ "step": 5341
+ },
+ {
+ "epoch": 74.19563318777293,
+ "grad_norm": 0.01431245356798172,
+ "learning_rate": 0.0006,
+ "loss": 4.5679121017456055,
+ "step": 5342
+ },
+ {
+ "epoch": 74.20960698689956,
+ "grad_norm": 0.013994511216878891,
+ "learning_rate": 0.0006,
+ "loss": 4.5236406326293945,
+ "step": 5343
+ },
+ {
+ "epoch": 74.2235807860262,
+ "grad_norm": 0.011971707455813885,
+ "learning_rate": 0.0006,
+ "loss": 4.626547813415527,
+ "step": 5344
+ },
+ {
+ "epoch": 74.23755458515284,
+ "grad_norm": 0.012805287726223469,
+ "learning_rate": 0.0006,
+ "loss": 4.556386947631836,
+ "step": 5345
+ },
+ {
+ "epoch": 74.25152838427948,
+ "grad_norm": 0.011694584973156452,
+ "learning_rate": 0.0006,
+ "loss": 4.656987190246582,
+ "step": 5346
+ },
+ {
+ "epoch": 74.26550218340611,
+ "grad_norm": 0.013396821916103363,
+ "learning_rate": 0.0006,
+ "loss": 4.526467323303223,
+ "step": 5347
+ },
+ {
+ "epoch": 74.27947598253274,
+ "grad_norm": 0.014398031868040562,
+ "learning_rate": 0.0006,
+ "loss": 4.595100402832031,
+ "step": 5348
+ },
+ {
+ "epoch": 74.29344978165939,
+ "grad_norm": 0.02165800705552101,
+ "learning_rate": 0.0006,
+ "loss": 4.460549354553223,
+ "step": 5349
+ },
+ {
+ "epoch": 74.30742358078602,
+ "grad_norm": 0.03276512026786804,
+ "learning_rate": 0.0006,
+ "loss": 4.498189449310303,
+ "step": 5350
+ },
+ {
+ "epoch": 74.32139737991267,
+ "grad_norm": 0.04616224020719528,
+ "learning_rate": 0.0006,
+ "loss": 4.5268659591674805,
+ "step": 5351
+ },
+ {
+ "epoch": 74.3353711790393,
+ "grad_norm": 0.03863786533474922,
+ "learning_rate": 0.0006,
+ "loss": 4.554747104644775,
+ "step": 5352
+ },
+ {
+ "epoch": 74.34934497816595,
+ "grad_norm": 0.01749865710735321,
+ "learning_rate": 0.0006,
+ "loss": 4.6656036376953125,
+ "step": 5353
+ },
+ {
+ "epoch": 74.36331877729258,
+ "grad_norm": 0.021814396604895592,
+ "learning_rate": 0.0006,
+ "loss": 4.607456207275391,
+ "step": 5354
+ },
+ {
+ "epoch": 74.37729257641921,
+ "grad_norm": 0.017018496990203857,
+ "learning_rate": 0.0006,
+ "loss": 4.54205322265625,
+ "step": 5355
+ },
+ {
+ "epoch": 74.39126637554585,
+ "grad_norm": 0.018297340720891953,
+ "learning_rate": 0.0006,
+ "loss": 4.608870506286621,
+ "step": 5356
+ },
+ {
+ "epoch": 74.40524017467249,
+ "grad_norm": 0.01473482046276331,
+ "learning_rate": 0.0006,
+ "loss": 4.384891510009766,
+ "step": 5357
+ },
+ {
+ "epoch": 74.41921397379913,
+ "grad_norm": 0.01579415239393711,
+ "learning_rate": 0.0006,
+ "loss": 4.588859558105469,
+ "step": 5358
+ },
+ {
+ "epoch": 74.43318777292576,
+ "grad_norm": 0.013906026259064674,
+ "learning_rate": 0.0006,
+ "loss": 4.525175094604492,
+ "step": 5359
+ },
+ {
+ "epoch": 74.44716157205241,
+ "grad_norm": 0.014163361862301826,
+ "learning_rate": 0.0006,
+ "loss": 4.495802879333496,
+ "step": 5360
+ },
+ {
+ "epoch": 74.46113537117904,
+ "grad_norm": 0.014009335078299046,
+ "learning_rate": 0.0006,
+ "loss": 4.5960869789123535,
+ "step": 5361
+ },
+ {
+ "epoch": 74.47510917030567,
+ "grad_norm": 0.012765258550643921,
+ "learning_rate": 0.0006,
+ "loss": 4.515678405761719,
+ "step": 5362
+ },
+ {
+ "epoch": 74.48908296943232,
+ "grad_norm": 0.014071918092668056,
+ "learning_rate": 0.0006,
+ "loss": 4.394025802612305,
+ "step": 5363
+ },
+ {
+ "epoch": 74.50305676855895,
+ "grad_norm": 0.01244515273720026,
+ "learning_rate": 0.0006,
+ "loss": 4.576985836029053,
+ "step": 5364
+ },
+ {
+ "epoch": 74.5170305676856,
+ "grad_norm": 0.012975298799574375,
+ "learning_rate": 0.0006,
+ "loss": 4.540657997131348,
+ "step": 5365
+ },
+ {
+ "epoch": 74.53100436681223,
+ "grad_norm": 0.012380947358906269,
+ "learning_rate": 0.0006,
+ "loss": 4.55731725692749,
+ "step": 5366
+ },
+ {
+ "epoch": 74.54497816593886,
+ "grad_norm": 0.012577617540955544,
+ "learning_rate": 0.0006,
+ "loss": 4.500924110412598,
+ "step": 5367
+ },
+ {
+ "epoch": 74.5589519650655,
+ "grad_norm": 0.012089254334568977,
+ "learning_rate": 0.0006,
+ "loss": 4.561795234680176,
+ "step": 5368
+ },
+ {
+ "epoch": 74.57292576419214,
+ "grad_norm": 0.012024011462926865,
+ "learning_rate": 0.0006,
+ "loss": 4.430635452270508,
+ "step": 5369
+ },
+ {
+ "epoch": 74.58689956331878,
+ "grad_norm": 0.012361960485577583,
+ "learning_rate": 0.0006,
+ "loss": 4.4664106369018555,
+ "step": 5370
+ },
+ {
+ "epoch": 74.60087336244541,
+ "grad_norm": 0.012149893678724766,
+ "learning_rate": 0.0006,
+ "loss": 4.522271156311035,
+ "step": 5371
+ },
+ {
+ "epoch": 74.61484716157206,
+ "grad_norm": 0.01166840922087431,
+ "learning_rate": 0.0006,
+ "loss": 4.3870463371276855,
+ "step": 5372
+ },
+ {
+ "epoch": 74.62882096069869,
+ "grad_norm": 0.01160407718271017,
+ "learning_rate": 0.0006,
+ "loss": 4.504255771636963,
+ "step": 5373
+ },
+ {
+ "epoch": 74.64279475982532,
+ "grad_norm": 0.011259177699685097,
+ "learning_rate": 0.0006,
+ "loss": 4.507267951965332,
+ "step": 5374
+ },
+ {
+ "epoch": 74.65676855895197,
+ "grad_norm": 0.012285715900361538,
+ "learning_rate": 0.0006,
+ "loss": 4.472966194152832,
+ "step": 5375
+ },
+ {
+ "epoch": 74.6707423580786,
+ "grad_norm": 0.011214440688490868,
+ "learning_rate": 0.0006,
+ "loss": 4.5840253829956055,
+ "step": 5376
+ },
+ {
+ "epoch": 74.68471615720524,
+ "grad_norm": 0.012394879944622517,
+ "learning_rate": 0.0006,
+ "loss": 4.417830944061279,
+ "step": 5377
+ },
+ {
+ "epoch": 74.69868995633188,
+ "grad_norm": 0.012360049411654472,
+ "learning_rate": 0.0006,
+ "loss": 4.535454750061035,
+ "step": 5378
+ },
+ {
+ "epoch": 74.71266375545852,
+ "grad_norm": 0.011961652897298336,
+ "learning_rate": 0.0006,
+ "loss": 4.403409957885742,
+ "step": 5379
+ },
+ {
+ "epoch": 74.72663755458515,
+ "grad_norm": 0.012078274972736835,
+ "learning_rate": 0.0006,
+ "loss": 4.448366641998291,
+ "step": 5380
+ },
+ {
+ "epoch": 74.74061135371178,
+ "grad_norm": 0.013239333406090736,
+ "learning_rate": 0.0006,
+ "loss": 4.543450355529785,
+ "step": 5381
+ },
+ {
+ "epoch": 74.75458515283843,
+ "grad_norm": 0.01476441603153944,
+ "learning_rate": 0.0006,
+ "loss": 4.488677978515625,
+ "step": 5382
+ },
+ {
+ "epoch": 74.76855895196506,
+ "grad_norm": 0.013968009501695633,
+ "learning_rate": 0.0006,
+ "loss": 4.603639602661133,
+ "step": 5383
+ },
+ {
+ "epoch": 74.78253275109171,
+ "grad_norm": 0.012322458438575268,
+ "learning_rate": 0.0006,
+ "loss": 4.637874603271484,
+ "step": 5384
+ },
+ {
+ "epoch": 74.79650655021834,
+ "grad_norm": 0.013352593407034874,
+ "learning_rate": 0.0006,
+ "loss": 4.474833965301514,
+ "step": 5385
+ },
+ {
+ "epoch": 74.81048034934499,
+ "grad_norm": 0.014671728946268559,
+ "learning_rate": 0.0006,
+ "loss": 4.516847610473633,
+ "step": 5386
+ },
+ {
+ "epoch": 74.82445414847162,
+ "grad_norm": 0.017163321375846863,
+ "learning_rate": 0.0006,
+ "loss": 4.462325096130371,
+ "step": 5387
+ },
+ {
+ "epoch": 74.83842794759825,
+ "grad_norm": 0.02012511156499386,
+ "learning_rate": 0.0006,
+ "loss": 4.355982780456543,
+ "step": 5388
+ },
+ {
+ "epoch": 74.8524017467249,
+ "grad_norm": 0.01881697215139866,
+ "learning_rate": 0.0006,
+ "loss": 4.451120376586914,
+ "step": 5389
+ },
+ {
+ "epoch": 74.86637554585153,
+ "grad_norm": 0.017279407009482384,
+ "learning_rate": 0.0006,
+ "loss": 4.461202621459961,
+ "step": 5390
+ },
+ {
+ "epoch": 74.88034934497817,
+ "grad_norm": 0.015470407903194427,
+ "learning_rate": 0.0006,
+ "loss": 4.4968671798706055,
+ "step": 5391
+ },
+ {
+ "epoch": 74.8943231441048,
+ "grad_norm": 0.01411783043295145,
+ "learning_rate": 0.0006,
+ "loss": 4.558697700500488,
+ "step": 5392
+ },
+ {
+ "epoch": 74.90829694323143,
+ "grad_norm": 0.017178812995553017,
+ "learning_rate": 0.0006,
+ "loss": 4.254153251647949,
+ "step": 5393
+ },
+ {
+ "epoch": 74.92227074235808,
+ "grad_norm": 0.019193004816770554,
+ "learning_rate": 0.0006,
+ "loss": 4.394747257232666,
+ "step": 5394
+ },
+ {
+ "epoch": 74.93624454148471,
+ "grad_norm": 0.018942881375551224,
+ "learning_rate": 0.0006,
+ "loss": 4.459539413452148,
+ "step": 5395
+ },
+ {
+ "epoch": 74.95021834061136,
+ "grad_norm": 0.01641223020851612,
+ "learning_rate": 0.0006,
+ "loss": 4.485820770263672,
+ "step": 5396
+ },
+ {
+ "epoch": 74.96419213973799,
+ "grad_norm": 0.014485753141343594,
+ "learning_rate": 0.0006,
+ "loss": 4.5068583488464355,
+ "step": 5397
+ },
+ {
+ "epoch": 74.97816593886463,
+ "grad_norm": 0.018242299556732178,
+ "learning_rate": 0.0006,
+ "loss": 4.508835792541504,
+ "step": 5398
+ },
+ {
+ "epoch": 74.99213973799127,
+ "grad_norm": 0.016491873189806938,
+ "learning_rate": 0.0006,
+ "loss": 4.4760870933532715,
+ "step": 5399
+ },
+ {
+ "epoch": 75.0,
+ "grad_norm": 0.014543353579938412,
+ "learning_rate": 0.0006,
+ "loss": 4.540217399597168,
+ "step": 5400
+ },
+ {
+ "epoch": 75.0,
+ "eval_loss": 4.7617573738098145,
+ "eval_runtime": 57.321,
+ "eval_samples_per_second": 42.602,
+ "eval_steps_per_second": 1.343,
+ "step": 5400
+ },
+ {
+ "epoch": 75.01397379912663,
+ "grad_norm": 0.014417114667594433,
+ "learning_rate": 0.0006,
+ "loss": 4.332956790924072,
+ "step": 5401
+ },
+ {
+ "epoch": 75.02794759825328,
+ "grad_norm": 0.01607639715075493,
+ "learning_rate": 0.0006,
+ "loss": 4.352453708648682,
+ "step": 5402
+ },
+ {
+ "epoch": 75.04192139737991,
+ "grad_norm": 0.015438318252563477,
+ "learning_rate": 0.0006,
+ "loss": 4.387139320373535,
+ "step": 5403
+ },
+ {
+ "epoch": 75.05589519650655,
+ "grad_norm": 0.015539007261395454,
+ "learning_rate": 0.0006,
+ "loss": 4.453171730041504,
+ "step": 5404
+ },
+ {
+ "epoch": 75.06986899563319,
+ "grad_norm": 0.017361903563141823,
+ "learning_rate": 0.0006,
+ "loss": 4.512143135070801,
+ "step": 5405
+ },
+ {
+ "epoch": 75.08384279475983,
+ "grad_norm": 0.01815599389374256,
+ "learning_rate": 0.0006,
+ "loss": 4.508997440338135,
+ "step": 5406
+ },
+ {
+ "epoch": 75.09781659388646,
+ "grad_norm": 0.01620607078075409,
+ "learning_rate": 0.0006,
+ "loss": 4.322444915771484,
+ "step": 5407
+ },
+ {
+ "epoch": 75.1117903930131,
+ "grad_norm": 0.014773265458643436,
+ "learning_rate": 0.0006,
+ "loss": 4.359251976013184,
+ "step": 5408
+ },
+ {
+ "epoch": 75.12576419213974,
+ "grad_norm": 0.014099023304879665,
+ "learning_rate": 0.0006,
+ "loss": 4.313380241394043,
+ "step": 5409
+ },
+ {
+ "epoch": 75.13973799126637,
+ "grad_norm": 0.014387735165655613,
+ "learning_rate": 0.0006,
+ "loss": 4.435023307800293,
+ "step": 5410
+ },
+ {
+ "epoch": 75.15371179039302,
+ "grad_norm": 0.014886162243783474,
+ "learning_rate": 0.0006,
+ "loss": 4.335939407348633,
+ "step": 5411
+ },
+ {
+ "epoch": 75.16768558951965,
+ "grad_norm": 0.015766102820634842,
+ "learning_rate": 0.0006,
+ "loss": 4.408674240112305,
+ "step": 5412
+ },
+ {
+ "epoch": 75.18165938864628,
+ "grad_norm": 0.016383284702897072,
+ "learning_rate": 0.0006,
+ "loss": 4.385453224182129,
+ "step": 5413
+ },
+ {
+ "epoch": 75.19563318777293,
+ "grad_norm": 0.013535384088754654,
+ "learning_rate": 0.0006,
+ "loss": 4.292785167694092,
+ "step": 5414
+ },
+ {
+ "epoch": 75.20960698689956,
+ "grad_norm": 0.01354036945849657,
+ "learning_rate": 0.0006,
+ "loss": 4.306252479553223,
+ "step": 5415
+ },
+ {
+ "epoch": 75.2235807860262,
+ "grad_norm": 0.015818078070878983,
+ "learning_rate": 0.0006,
+ "loss": 4.425487995147705,
+ "step": 5416
+ },
+ {
+ "epoch": 75.23755458515284,
+ "grad_norm": 0.014363568276166916,
+ "learning_rate": 0.0006,
+ "loss": 4.395087242126465,
+ "step": 5417
+ },
+ {
+ "epoch": 75.25152838427948,
+ "grad_norm": 0.013895470649003983,
+ "learning_rate": 0.0006,
+ "loss": 4.364090442657471,
+ "step": 5418
+ },
+ {
+ "epoch": 75.26550218340611,
+ "grad_norm": 0.014178668148815632,
+ "learning_rate": 0.0006,
+ "loss": 4.44619083404541,
+ "step": 5419
+ },
+ {
+ "epoch": 75.27947598253274,
+ "grad_norm": 0.0152514623478055,
+ "learning_rate": 0.0006,
+ "loss": 4.348445892333984,
+ "step": 5420
+ },
+ {
+ "epoch": 75.29344978165939,
+ "grad_norm": 0.015397852286696434,
+ "learning_rate": 0.0006,
+ "loss": 4.476832389831543,
+ "step": 5421
+ },
+ {
+ "epoch": 75.30742358078602,
+ "grad_norm": 0.018009863793849945,
+ "learning_rate": 0.0006,
+ "loss": 4.431162357330322,
+ "step": 5422
+ },
+ {
+ "epoch": 75.32139737991267,
+ "grad_norm": 0.017038367688655853,
+ "learning_rate": 0.0006,
+ "loss": 4.474787712097168,
+ "step": 5423
+ },
+ {
+ "epoch": 75.3353711790393,
+ "grad_norm": 0.013763288035988808,
+ "learning_rate": 0.0006,
+ "loss": 4.470419883728027,
+ "step": 5424
+ },
+ {
+ "epoch": 75.34934497816595,
+ "grad_norm": 0.013972110114991665,
+ "learning_rate": 0.0006,
+ "loss": 4.373147964477539,
+ "step": 5425
+ },
+ {
+ "epoch": 75.36331877729258,
+ "grad_norm": 0.017197875306010246,
+ "learning_rate": 0.0006,
+ "loss": 4.494963645935059,
+ "step": 5426
+ },
+ {
+ "epoch": 75.37729257641921,
+ "grad_norm": 0.01687047816812992,
+ "learning_rate": 0.0006,
+ "loss": 4.435934066772461,
+ "step": 5427
+ },
+ {
+ "epoch": 75.39126637554585,
+ "grad_norm": 0.01578272506594658,
+ "learning_rate": 0.0006,
+ "loss": 4.37918758392334,
+ "step": 5428
+ },
+ {
+ "epoch": 75.40524017467249,
+ "grad_norm": 0.015956247225403786,
+ "learning_rate": 0.0006,
+ "loss": 4.455668926239014,
+ "step": 5429
+ },
+ {
+ "epoch": 75.41921397379913,
+ "grad_norm": 0.013495155610144138,
+ "learning_rate": 0.0006,
+ "loss": 4.4271039962768555,
+ "step": 5430
+ },
+ {
+ "epoch": 75.43318777292576,
+ "grad_norm": 0.012851621024310589,
+ "learning_rate": 0.0006,
+ "loss": 4.213107109069824,
+ "step": 5431
+ },
+ {
+ "epoch": 75.44716157205241,
+ "grad_norm": 0.013351158238947392,
+ "learning_rate": 0.0006,
+ "loss": 4.3070783615112305,
+ "step": 5432
+ },
+ {
+ "epoch": 75.46113537117904,
+ "grad_norm": 0.014765217900276184,
+ "learning_rate": 0.0006,
+ "loss": 4.385021209716797,
+ "step": 5433
+ },
+ {
+ "epoch": 75.47510917030567,
+ "grad_norm": 0.016091614961624146,
+ "learning_rate": 0.0006,
+ "loss": 4.312644004821777,
+ "step": 5434
+ },
+ {
+ "epoch": 75.48908296943232,
+ "grad_norm": 0.01612013578414917,
+ "learning_rate": 0.0006,
+ "loss": 4.4482011795043945,
+ "step": 5435
+ },
+ {
+ "epoch": 75.50305676855895,
+ "grad_norm": 0.013157603330910206,
+ "learning_rate": 0.0006,
+ "loss": 4.427262306213379,
+ "step": 5436
+ },
+ {
+ "epoch": 75.5170305676856,
+ "grad_norm": 0.016039809212088585,
+ "learning_rate": 0.0006,
+ "loss": 4.396635055541992,
+ "step": 5437
+ },
+ {
+ "epoch": 75.53100436681223,
+ "grad_norm": 0.01794283092021942,
+ "learning_rate": 0.0006,
+ "loss": 4.476870059967041,
+ "step": 5438
+ },
+ {
+ "epoch": 75.54497816593886,
+ "grad_norm": 0.017030486837029457,
+ "learning_rate": 0.0006,
+ "loss": 4.361156463623047,
+ "step": 5439
+ },
+ {
+ "epoch": 75.5589519650655,
+ "grad_norm": 0.017915161326527596,
+ "learning_rate": 0.0006,
+ "loss": 4.286147117614746,
+ "step": 5440
+ },
+ {
+ "epoch": 75.57292576419214,
+ "grad_norm": 0.017514433711767197,
+ "learning_rate": 0.0006,
+ "loss": 4.394091606140137,
+ "step": 5441
+ },
+ {
+ "epoch": 75.58689956331878,
+ "grad_norm": 0.015128589235246181,
+ "learning_rate": 0.0006,
+ "loss": 4.383126735687256,
+ "step": 5442
+ },
+ {
+ "epoch": 75.60087336244541,
+ "grad_norm": 0.0202616099268198,
+ "learning_rate": 0.0006,
+ "loss": 4.336188316345215,
+ "step": 5443
+ },
+ {
+ "epoch": 75.61484716157206,
+ "grad_norm": 0.019978471100330353,
+ "learning_rate": 0.0006,
+ "loss": 4.461453437805176,
+ "step": 5444
+ },
+ {
+ "epoch": 75.62882096069869,
+ "grad_norm": 0.01738741435110569,
+ "learning_rate": 0.0006,
+ "loss": 4.371197700500488,
+ "step": 5445
+ },
+ {
+ "epoch": 75.64279475982532,
+ "grad_norm": 0.016157809644937515,
+ "learning_rate": 0.0006,
+ "loss": 4.3312835693359375,
+ "step": 5446
+ },
+ {
+ "epoch": 75.65676855895197,
+ "grad_norm": 0.016394613310694695,
+ "learning_rate": 0.0006,
+ "loss": 4.380190849304199,
+ "step": 5447
+ },
+ {
+ "epoch": 75.6707423580786,
+ "grad_norm": 0.014727243222296238,
+ "learning_rate": 0.0006,
+ "loss": 4.308028221130371,
+ "step": 5448
+ },
+ {
+ "epoch": 75.68471615720524,
+ "grad_norm": 0.01695958711206913,
+ "learning_rate": 0.0006,
+ "loss": 4.399503707885742,
+ "step": 5449
+ },
+ {
+ "epoch": 75.69868995633188,
+ "grad_norm": 0.015720795840024948,
+ "learning_rate": 0.0006,
+ "loss": 4.342249870300293,
+ "step": 5450
+ },
+ {
+ "epoch": 75.71266375545852,
+ "grad_norm": 0.01574772410094738,
+ "learning_rate": 0.0006,
+ "loss": 4.331503868103027,
+ "step": 5451
+ },
+ {
+ "epoch": 75.72663755458515,
+ "grad_norm": 0.014812648296356201,
+ "learning_rate": 0.0006,
+ "loss": 4.332413673400879,
+ "step": 5452
+ },
+ {
+ "epoch": 75.74061135371178,
+ "grad_norm": 0.012850426137447357,
+ "learning_rate": 0.0006,
+ "loss": 4.2884063720703125,
+ "step": 5453
+ },
+ {
+ "epoch": 75.75458515283843,
+ "grad_norm": 0.012876970693469048,
+ "learning_rate": 0.0006,
+ "loss": 4.399855613708496,
+ "step": 5454
+ },
+ {
+ "epoch": 75.76855895196506,
+ "grad_norm": 0.012958863750100136,
+ "learning_rate": 0.0006,
+ "loss": 4.411185264587402,
+ "step": 5455
+ },
+ {
+ "epoch": 75.78253275109171,
+ "grad_norm": 0.012197135016322136,
+ "learning_rate": 0.0006,
+ "loss": 4.272603988647461,
+ "step": 5456
+ },
+ {
+ "epoch": 75.79650655021834,
+ "grad_norm": 0.013483566232025623,
+ "learning_rate": 0.0006,
+ "loss": 4.3564348220825195,
+ "step": 5457
+ },
+ {
+ "epoch": 75.81048034934499,
+ "grad_norm": 0.012365429662168026,
+ "learning_rate": 0.0006,
+ "loss": 4.354825973510742,
+ "step": 5458
+ },
+ {
+ "epoch": 75.82445414847162,
+ "grad_norm": 0.012238141149282455,
+ "learning_rate": 0.0006,
+ "loss": 4.440438747406006,
+ "step": 5459
+ },
+ {
+ "epoch": 75.83842794759825,
+ "grad_norm": 0.013380305841565132,
+ "learning_rate": 0.0006,
+ "loss": 4.2813215255737305,
+ "step": 5460
+ },
+ {
+ "epoch": 75.8524017467249,
+ "grad_norm": 0.01443779468536377,
+ "learning_rate": 0.0006,
+ "loss": 4.289078235626221,
+ "step": 5461
+ },
+ {
+ "epoch": 75.86637554585153,
+ "grad_norm": 0.012646574527025223,
+ "learning_rate": 0.0006,
+ "loss": 4.374177932739258,
+ "step": 5462
+ },
+ {
+ "epoch": 75.88034934497817,
+ "grad_norm": 0.012403388507664204,
+ "learning_rate": 0.0006,
+ "loss": 4.343016147613525,
+ "step": 5463
+ },
+ {
+ "epoch": 75.8943231441048,
+ "grad_norm": 0.01319252047687769,
+ "learning_rate": 0.0006,
+ "loss": 4.309957027435303,
+ "step": 5464
+ },
+ {
+ "epoch": 75.90829694323143,
+ "grad_norm": 0.012597428634762764,
+ "learning_rate": 0.0006,
+ "loss": 4.403654098510742,
+ "step": 5465
+ },
+ {
+ "epoch": 75.92227074235808,
+ "grad_norm": 0.012724226340651512,
+ "learning_rate": 0.0006,
+ "loss": 4.326869010925293,
+ "step": 5466
+ },
+ {
+ "epoch": 75.93624454148471,
+ "grad_norm": 0.012698772363364697,
+ "learning_rate": 0.0006,
+ "loss": 4.268430233001709,
+ "step": 5467
+ },
+ {
+ "epoch": 75.95021834061136,
+ "grad_norm": 0.01376490481197834,
+ "learning_rate": 0.0006,
+ "loss": 4.42446231842041,
+ "step": 5468
+ },
+ {
+ "epoch": 75.96419213973799,
+ "grad_norm": 0.013175144791603088,
+ "learning_rate": 0.0006,
+ "loss": 4.374675750732422,
+ "step": 5469
+ },
+ {
+ "epoch": 75.97816593886463,
+ "grad_norm": 0.012731322087347507,
+ "learning_rate": 0.0006,
+ "loss": 4.372966766357422,
+ "step": 5470
+ },
+ {
+ "epoch": 75.99213973799127,
+ "grad_norm": 0.01360505260527134,
+ "learning_rate": 0.0006,
+ "loss": 4.3095903396606445,
+ "step": 5471
+ },
+ {
+ "epoch": 76.0,
+ "grad_norm": 0.014011901803314686,
+ "learning_rate": 0.0006,
+ "loss": 4.26954460144043,
+ "step": 5472
+ },
+ {
+ "epoch": 76.0,
+ "eval_loss": 4.689640998840332,
+ "eval_runtime": 56.9095,
+ "eval_samples_per_second": 42.91,
+ "eval_steps_per_second": 1.353,
+ "step": 5472
+ },
+ {
+ "epoch": 76.01397379912663,
+ "grad_norm": 0.015351055189967155,
+ "learning_rate": 0.0006,
+ "loss": 4.348268985748291,
+ "step": 5473
+ },
+ {
+ "epoch": 76.02794759825328,
+ "grad_norm": 0.017164217308163643,
+ "learning_rate": 0.0006,
+ "loss": 4.244204998016357,
+ "step": 5474
+ },
+ {
+ "epoch": 76.04192139737991,
+ "grad_norm": 0.021624932065606117,
+ "learning_rate": 0.0006,
+ "loss": 4.311168670654297,
+ "step": 5475
+ },
+ {
+ "epoch": 76.05589519650655,
+ "grad_norm": 0.028712468221783638,
+ "learning_rate": 0.0006,
+ "loss": 4.322287082672119,
+ "step": 5476
+ },
+ {
+ "epoch": 76.06986899563319,
+ "grad_norm": 0.027915487065911293,
+ "learning_rate": 0.0006,
+ "loss": 4.381207466125488,
+ "step": 5477
+ },
+ {
+ "epoch": 76.08384279475983,
+ "grad_norm": 0.017955303192138672,
+ "learning_rate": 0.0006,
+ "loss": 4.2635016441345215,
+ "step": 5478
+ },
+ {
+ "epoch": 76.09781659388646,
+ "grad_norm": 0.019113514572381973,
+ "learning_rate": 0.0006,
+ "loss": 4.399471282958984,
+ "step": 5479
+ },
+ {
+ "epoch": 76.1117903930131,
+ "grad_norm": 0.017502542585134506,
+ "learning_rate": 0.0006,
+ "loss": 4.321204662322998,
+ "step": 5480
+ },
+ {
+ "epoch": 76.12576419213974,
+ "grad_norm": 0.01852525770664215,
+ "learning_rate": 0.0006,
+ "loss": 4.217884063720703,
+ "step": 5481
+ },
+ {
+ "epoch": 76.13973799126637,
+ "grad_norm": 0.019836828112602234,
+ "learning_rate": 0.0006,
+ "loss": 4.278532028198242,
+ "step": 5482
+ },
+ {
+ "epoch": 76.15371179039302,
+ "grad_norm": 0.02101975679397583,
+ "learning_rate": 0.0006,
+ "loss": 4.246720314025879,
+ "step": 5483
+ },
+ {
+ "epoch": 76.16768558951965,
+ "grad_norm": 0.021202990785241127,
+ "learning_rate": 0.0006,
+ "loss": 4.517517566680908,
+ "step": 5484
+ },
+ {
+ "epoch": 76.18165938864628,
+ "grad_norm": 0.021428769454360008,
+ "learning_rate": 0.0006,
+ "loss": 4.3373517990112305,
+ "step": 5485
+ },
+ {
+ "epoch": 76.19563318777293,
+ "grad_norm": 0.018263190984725952,
+ "learning_rate": 0.0006,
+ "loss": 4.402856826782227,
+ "step": 5486
+ },
+ {
+ "epoch": 76.20960698689956,
+ "grad_norm": 0.019025536254048347,
+ "learning_rate": 0.0006,
+ "loss": 4.314227104187012,
+ "step": 5487
+ },
+ {
+ "epoch": 76.2235807860262,
+ "grad_norm": 0.020918630063533783,
+ "learning_rate": 0.0006,
+ "loss": 4.454369068145752,
+ "step": 5488
+ },
+ {
+ "epoch": 76.23755458515284,
+ "grad_norm": 0.019307617098093033,
+ "learning_rate": 0.0006,
+ "loss": 4.247509002685547,
+ "step": 5489
+ },
+ {
+ "epoch": 76.25152838427948,
+ "grad_norm": 0.014956331811845303,
+ "learning_rate": 0.0006,
+ "loss": 4.365175247192383,
+ "step": 5490
+ },
+ {
+ "epoch": 76.26550218340611,
+ "grad_norm": 0.017213059589266777,
+ "learning_rate": 0.0006,
+ "loss": 4.384005069732666,
+ "step": 5491
+ },
+ {
+ "epoch": 76.27947598253274,
+ "grad_norm": 0.017575597390532494,
+ "learning_rate": 0.0006,
+ "loss": 4.41934871673584,
+ "step": 5492
+ },
+ {
+ "epoch": 76.29344978165939,
+ "grad_norm": 0.015879256650805473,
+ "learning_rate": 0.0006,
+ "loss": 4.26675271987915,
+ "step": 5493
+ },
+ {
+ "epoch": 76.30742358078602,
+ "grad_norm": 0.01564522460103035,
+ "learning_rate": 0.0006,
+ "loss": 4.277339935302734,
+ "step": 5494
+ },
+ {
+ "epoch": 76.32139737991267,
+ "grad_norm": 0.016414586454629898,
+ "learning_rate": 0.0006,
+ "loss": 4.4138593673706055,
+ "step": 5495
+ },
+ {
+ "epoch": 76.3353711790393,
+ "grad_norm": 0.014732337556779385,
+ "learning_rate": 0.0006,
+ "loss": 4.281006336212158,
+ "step": 5496
+ },
+ {
+ "epoch": 76.34934497816595,
+ "grad_norm": 0.015461455099284649,
+ "learning_rate": 0.0006,
+ "loss": 4.263422966003418,
+ "step": 5497
+ },
+ {
+ "epoch": 76.36331877729258,
+ "grad_norm": 0.014529814012348652,
+ "learning_rate": 0.0006,
+ "loss": 4.36165189743042,
+ "step": 5498
+ },
+ {
+ "epoch": 76.37729257641921,
+ "grad_norm": 0.01469483319669962,
+ "learning_rate": 0.0006,
+ "loss": 4.309642791748047,
+ "step": 5499
+ },
+ {
+ "epoch": 76.39126637554585,
+ "grad_norm": 0.016490574926137924,
+ "learning_rate": 0.0006,
+ "loss": 4.288150787353516,
+ "step": 5500
+ },
+ {
+ "epoch": 76.40524017467249,
+ "grad_norm": 0.014593565836548805,
+ "learning_rate": 0.0006,
+ "loss": 4.271614074707031,
+ "step": 5501
+ },
+ {
+ "epoch": 76.41921397379913,
+ "grad_norm": 0.012796535156667233,
+ "learning_rate": 0.0006,
+ "loss": 4.420042037963867,
+ "step": 5502
+ },
+ {
+ "epoch": 76.43318777292576,
+ "grad_norm": 0.016086025163531303,
+ "learning_rate": 0.0006,
+ "loss": 4.289943695068359,
+ "step": 5503
+ },
+ {
+ "epoch": 76.44716157205241,
+ "grad_norm": 0.015439470298588276,
+ "learning_rate": 0.0006,
+ "loss": 4.3023271560668945,
+ "step": 5504
+ },
+ {
+ "epoch": 76.46113537117904,
+ "grad_norm": 0.015066367574036121,
+ "learning_rate": 0.0006,
+ "loss": 4.197081565856934,
+ "step": 5505
+ },
+ {
+ "epoch": 76.47510917030567,
+ "grad_norm": 0.016653263941407204,
+ "learning_rate": 0.0006,
+ "loss": 4.333785057067871,
+ "step": 5506
+ },
+ {
+ "epoch": 76.48908296943232,
+ "grad_norm": 0.014791973866522312,
+ "learning_rate": 0.0006,
+ "loss": 4.323811054229736,
+ "step": 5507
+ },
+ {
+ "epoch": 76.50305676855895,
+ "grad_norm": 0.014694827608764172,
+ "learning_rate": 0.0006,
+ "loss": 4.195852279663086,
+ "step": 5508
+ },
+ {
+ "epoch": 76.5170305676856,
+ "grad_norm": 0.015120396390557289,
+ "learning_rate": 0.0006,
+ "loss": 4.42379903793335,
+ "step": 5509
+ },
+ {
+ "epoch": 76.53100436681223,
+ "grad_norm": 0.014584025368094444,
+ "learning_rate": 0.0006,
+ "loss": 4.2702531814575195,
+ "step": 5510
+ },
+ {
+ "epoch": 76.54497816593886,
+ "grad_norm": 0.013981613330543041,
+ "learning_rate": 0.0006,
+ "loss": 4.383029937744141,
+ "step": 5511
+ },
+ {
+ "epoch": 76.5589519650655,
+ "grad_norm": 0.012930642813444138,
+ "learning_rate": 0.0006,
+ "loss": 4.296136856079102,
+ "step": 5512
+ },
+ {
+ "epoch": 76.57292576419214,
+ "grad_norm": 0.013060989789664745,
+ "learning_rate": 0.0006,
+ "loss": 4.346380233764648,
+ "step": 5513
+ },
+ {
+ "epoch": 76.58689956331878,
+ "grad_norm": 0.013209798373281956,
+ "learning_rate": 0.0006,
+ "loss": 4.316067218780518,
+ "step": 5514
+ },
+ {
+ "epoch": 76.60087336244541,
+ "grad_norm": 0.012423216365277767,
+ "learning_rate": 0.0006,
+ "loss": 4.370274543762207,
+ "step": 5515
+ },
+ {
+ "epoch": 76.61484716157206,
+ "grad_norm": 0.013205411843955517,
+ "learning_rate": 0.0006,
+ "loss": 4.429969310760498,
+ "step": 5516
+ },
+ {
+ "epoch": 76.62882096069869,
+ "grad_norm": 0.013049827888607979,
+ "learning_rate": 0.0006,
+ "loss": 4.378549098968506,
+ "step": 5517
+ },
+ {
+ "epoch": 76.64279475982532,
+ "grad_norm": 0.012075880542397499,
+ "learning_rate": 0.0006,
+ "loss": 4.348293304443359,
+ "step": 5518
+ },
+ {
+ "epoch": 76.65676855895197,
+ "grad_norm": 0.014932847581803799,
+ "learning_rate": 0.0006,
+ "loss": 4.282538890838623,
+ "step": 5519
+ },
+ {
+ "epoch": 76.6707423580786,
+ "grad_norm": 0.015615998767316341,
+ "learning_rate": 0.0006,
+ "loss": 4.274641036987305,
+ "step": 5520
+ },
+ {
+ "epoch": 76.68471615720524,
+ "grad_norm": 0.01424513477832079,
+ "learning_rate": 0.0006,
+ "loss": 4.243166923522949,
+ "step": 5521
+ },
+ {
+ "epoch": 76.69868995633188,
+ "grad_norm": 0.012775546871125698,
+ "learning_rate": 0.0006,
+ "loss": 4.335611820220947,
+ "step": 5522
+ },
+ {
+ "epoch": 76.71266375545852,
+ "grad_norm": 0.014461303129792213,
+ "learning_rate": 0.0006,
+ "loss": 4.246976375579834,
+ "step": 5523
+ },
+ {
+ "epoch": 76.72663755458515,
+ "grad_norm": 0.0145436255261302,
+ "learning_rate": 0.0006,
+ "loss": 4.275895118713379,
+ "step": 5524
+ },
+ {
+ "epoch": 76.74061135371178,
+ "grad_norm": 0.014055085368454456,
+ "learning_rate": 0.0006,
+ "loss": 4.331108093261719,
+ "step": 5525
+ },
+ {
+ "epoch": 76.75458515283843,
+ "grad_norm": 0.011933894827961922,
+ "learning_rate": 0.0006,
+ "loss": 4.293721675872803,
+ "step": 5526
+ },
+ {
+ "epoch": 76.76855895196506,
+ "grad_norm": 0.01277392078191042,
+ "learning_rate": 0.0006,
+ "loss": 4.312374114990234,
+ "step": 5527
+ },
+ {
+ "epoch": 76.78253275109171,
+ "grad_norm": 0.014227936044335365,
+ "learning_rate": 0.0006,
+ "loss": 4.238729000091553,
+ "step": 5528
+ },
+ {
+ "epoch": 76.79650655021834,
+ "grad_norm": 0.01488024927675724,
+ "learning_rate": 0.0006,
+ "loss": 4.24587345123291,
+ "step": 5529
+ },
+ {
+ "epoch": 76.81048034934499,
+ "grad_norm": 0.015341250225901604,
+ "learning_rate": 0.0006,
+ "loss": 4.257378578186035,
+ "step": 5530
+ },
+ {
+ "epoch": 76.82445414847162,
+ "grad_norm": 0.015179437585175037,
+ "learning_rate": 0.0006,
+ "loss": 4.2812910079956055,
+ "step": 5531
+ },
+ {
+ "epoch": 76.83842794759825,
+ "grad_norm": 0.012814724817872047,
+ "learning_rate": 0.0006,
+ "loss": 4.2309465408325195,
+ "step": 5532
+ },
+ {
+ "epoch": 76.8524017467249,
+ "grad_norm": 0.013810716569423676,
+ "learning_rate": 0.0006,
+ "loss": 4.325292110443115,
+ "step": 5533
+ },
+ {
+ "epoch": 76.86637554585153,
+ "grad_norm": 0.017468232661485672,
+ "learning_rate": 0.0006,
+ "loss": 4.243788719177246,
+ "step": 5534
+ },
+ {
+ "epoch": 76.88034934497817,
+ "grad_norm": 0.01739857904613018,
+ "learning_rate": 0.0006,
+ "loss": 4.345390796661377,
+ "step": 5535
+ },
+ {
+ "epoch": 76.8943231441048,
+ "grad_norm": 0.015329061076045036,
+ "learning_rate": 0.0006,
+ "loss": 4.415142059326172,
+ "step": 5536
+ },
+ {
+ "epoch": 76.90829694323143,
+ "grad_norm": 0.01598426327109337,
+ "learning_rate": 0.0006,
+ "loss": 4.299558639526367,
+ "step": 5537
+ },
+ {
+ "epoch": 76.92227074235808,
+ "grad_norm": 0.01621454581618309,
+ "learning_rate": 0.0006,
+ "loss": 4.284354209899902,
+ "step": 5538
+ },
+ {
+ "epoch": 76.93624454148471,
+ "grad_norm": 0.015956968069076538,
+ "learning_rate": 0.0006,
+ "loss": 4.293837547302246,
+ "step": 5539
+ },
+ {
+ "epoch": 76.95021834061136,
+ "grad_norm": 0.01776493526995182,
+ "learning_rate": 0.0006,
+ "loss": 4.229729175567627,
+ "step": 5540
+ },
+ {
+ "epoch": 76.96419213973799,
+ "grad_norm": 0.015317477285861969,
+ "learning_rate": 0.0006,
+ "loss": 4.3328447341918945,
+ "step": 5541
+ },
+ {
+ "epoch": 76.97816593886463,
+ "grad_norm": 0.015140251256525517,
+ "learning_rate": 0.0006,
+ "loss": 4.310840606689453,
+ "step": 5542
+ },
+ {
+ "epoch": 76.99213973799127,
+ "grad_norm": 0.015326812863349915,
+ "learning_rate": 0.0006,
+ "loss": 4.230832099914551,
+ "step": 5543
+ },
+ {
+ "epoch": 77.0,
+ "grad_norm": 0.01607763208448887,
+ "learning_rate": 0.0006,
+ "loss": 4.2236328125,
+ "step": 5544
+ },
+ {
+ "epoch": 77.0,
+ "eval_loss": 4.662221431732178,
+ "eval_runtime": 56.5194,
+ "eval_samples_per_second": 43.206,
+ "eval_steps_per_second": 1.362,
+ "step": 5544
+ },
+ {
+ "epoch": 77.01397379912663,
+ "grad_norm": 0.014839510433375835,
+ "learning_rate": 0.0006,
+ "loss": 4.254111289978027,
+ "step": 5545
+ },
+ {
+ "epoch": 77.02794759825328,
+ "grad_norm": 0.01519741676747799,
+ "learning_rate": 0.0006,
+ "loss": 4.1607184410095215,
+ "step": 5546
+ },
+ {
+ "epoch": 77.04192139737991,
+ "grad_norm": 0.015472603030502796,
+ "learning_rate": 0.0006,
+ "loss": 4.323320388793945,
+ "step": 5547
+ },
+ {
+ "epoch": 77.05589519650655,
+ "grad_norm": 0.017937948927283287,
+ "learning_rate": 0.0006,
+ "loss": 4.308490753173828,
+ "step": 5548
+ },
+ {
+ "epoch": 77.06986899563319,
+ "grad_norm": 0.020890841260552406,
+ "learning_rate": 0.0006,
+ "loss": 4.259101867675781,
+ "step": 5549
+ },
+ {
+ "epoch": 77.08384279475983,
+ "grad_norm": 0.01892857998609543,
+ "learning_rate": 0.0006,
+ "loss": 4.386970520019531,
+ "step": 5550
+ },
+ {
+ "epoch": 77.09781659388646,
+ "grad_norm": 0.016583630815148354,
+ "learning_rate": 0.0006,
+ "loss": 4.233604907989502,
+ "step": 5551
+ },
+ {
+ "epoch": 77.1117903930131,
+ "grad_norm": 0.016063468530774117,
+ "learning_rate": 0.0006,
+ "loss": 4.339544296264648,
+ "step": 5552
+ },
+ {
+ "epoch": 77.12576419213974,
+ "grad_norm": 0.018326779827475548,
+ "learning_rate": 0.0006,
+ "loss": 4.367204666137695,
+ "step": 5553
+ },
+ {
+ "epoch": 77.13973799126637,
+ "grad_norm": 0.021590933203697205,
+ "learning_rate": 0.0006,
+ "loss": 4.352685928344727,
+ "step": 5554
+ },
+ {
+ "epoch": 77.15371179039302,
+ "grad_norm": 0.019643642008304596,
+ "learning_rate": 0.0006,
+ "loss": 4.258535385131836,
+ "step": 5555
+ },
+ {
+ "epoch": 77.16768558951965,
+ "grad_norm": 0.018476149067282677,
+ "learning_rate": 0.0006,
+ "loss": 4.232568740844727,
+ "step": 5556
+ },
+ {
+ "epoch": 77.18165938864628,
+ "grad_norm": 0.01584405079483986,
+ "learning_rate": 0.0006,
+ "loss": 4.258371353149414,
+ "step": 5557
+ },
+ {
+ "epoch": 77.19563318777293,
+ "grad_norm": 0.01575201191008091,
+ "learning_rate": 0.0006,
+ "loss": 4.183218002319336,
+ "step": 5558
+ },
+ {
+ "epoch": 77.20960698689956,
+ "grad_norm": 0.016871824860572815,
+ "learning_rate": 0.0006,
+ "loss": 4.194563865661621,
+ "step": 5559
+ },
+ {
+ "epoch": 77.2235807860262,
+ "grad_norm": 0.02059333398938179,
+ "learning_rate": 0.0006,
+ "loss": 4.195281982421875,
+ "step": 5560
+ },
+ {
+ "epoch": 77.23755458515284,
+ "grad_norm": 0.022640453651547432,
+ "learning_rate": 0.0006,
+ "loss": 4.279615879058838,
+ "step": 5561
+ },
+ {
+ "epoch": 77.25152838427948,
+ "grad_norm": 0.020335959270596504,
+ "learning_rate": 0.0006,
+ "loss": 4.25040864944458,
+ "step": 5562
+ },
+ {
+ "epoch": 77.26550218340611,
+ "grad_norm": 0.020010793581604958,
+ "learning_rate": 0.0006,
+ "loss": 4.291116714477539,
+ "step": 5563
+ },
+ {
+ "epoch": 77.27947598253274,
+ "grad_norm": 0.02109045907855034,
+ "learning_rate": 0.0006,
+ "loss": 4.364088535308838,
+ "step": 5564
+ },
+ {
+ "epoch": 77.29344978165939,
+ "grad_norm": 0.019519446417689323,
+ "learning_rate": 0.0006,
+ "loss": 4.329166412353516,
+ "step": 5565
+ },
+ {
+ "epoch": 77.30742358078602,
+ "grad_norm": 0.01689804159104824,
+ "learning_rate": 0.0006,
+ "loss": 4.244176864624023,
+ "step": 5566
+ },
+ {
+ "epoch": 77.32139737991267,
+ "grad_norm": 0.01893681474030018,
+ "learning_rate": 0.0006,
+ "loss": 4.239068984985352,
+ "step": 5567
+ },
+ {
+ "epoch": 77.3353711790393,
+ "grad_norm": 0.015998193994164467,
+ "learning_rate": 0.0006,
+ "loss": 4.173398017883301,
+ "step": 5568
+ },
+ {
+ "epoch": 77.34934497816595,
+ "grad_norm": 0.013892280869185925,
+ "learning_rate": 0.0006,
+ "loss": 4.287654876708984,
+ "step": 5569
+ },
+ {
+ "epoch": 77.36331877729258,
+ "grad_norm": 0.017060643061995506,
+ "learning_rate": 0.0006,
+ "loss": 4.398447036743164,
+ "step": 5570
+ },
+ {
+ "epoch": 77.37729257641921,
+ "grad_norm": 0.015168849378824234,
+ "learning_rate": 0.0006,
+ "loss": 4.15665864944458,
+ "step": 5571
+ },
+ {
+ "epoch": 77.39126637554585,
+ "grad_norm": 0.014886523596942425,
+ "learning_rate": 0.0006,
+ "loss": 4.196673393249512,
+ "step": 5572
+ },
+ {
+ "epoch": 77.40524017467249,
+ "grad_norm": 0.016941796988248825,
+ "learning_rate": 0.0006,
+ "loss": 4.314562797546387,
+ "step": 5573
+ },
+ {
+ "epoch": 77.41921397379913,
+ "grad_norm": 0.015353812836110592,
+ "learning_rate": 0.0006,
+ "loss": 4.409579277038574,
+ "step": 5574
+ },
+ {
+ "epoch": 77.43318777292576,
+ "grad_norm": 0.015192852355539799,
+ "learning_rate": 0.0006,
+ "loss": 4.392045974731445,
+ "step": 5575
+ },
+ {
+ "epoch": 77.44716157205241,
+ "grad_norm": 0.014232569374144077,
+ "learning_rate": 0.0006,
+ "loss": 4.21639347076416,
+ "step": 5576
+ },
+ {
+ "epoch": 77.46113537117904,
+ "grad_norm": 0.014453226700425148,
+ "learning_rate": 0.0006,
+ "loss": 4.2802276611328125,
+ "step": 5577
+ },
+ {
+ "epoch": 77.47510917030567,
+ "grad_norm": 0.012881236150860786,
+ "learning_rate": 0.0006,
+ "loss": 4.375980854034424,
+ "step": 5578
+ },
+ {
+ "epoch": 77.48908296943232,
+ "grad_norm": 0.01583976298570633,
+ "learning_rate": 0.0006,
+ "loss": 4.264520645141602,
+ "step": 5579
+ },
+ {
+ "epoch": 77.50305676855895,
+ "grad_norm": 0.014171584509313107,
+ "learning_rate": 0.0006,
+ "loss": 4.225279808044434,
+ "step": 5580
+ },
+ {
+ "epoch": 77.5170305676856,
+ "grad_norm": 0.014715159311890602,
+ "learning_rate": 0.0006,
+ "loss": 4.369834899902344,
+ "step": 5581
+ },
+ {
+ "epoch": 77.53100436681223,
+ "grad_norm": 0.016554079949855804,
+ "learning_rate": 0.0006,
+ "loss": 4.2976179122924805,
+ "step": 5582
+ },
+ {
+ "epoch": 77.54497816593886,
+ "grad_norm": 0.015095747075974941,
+ "learning_rate": 0.0006,
+ "loss": 4.294407844543457,
+ "step": 5583
+ },
+ {
+ "epoch": 77.5589519650655,
+ "grad_norm": 0.012726361863315105,
+ "learning_rate": 0.0006,
+ "loss": 4.221401214599609,
+ "step": 5584
+ },
+ {
+ "epoch": 77.57292576419214,
+ "grad_norm": 0.014999309554696083,
+ "learning_rate": 0.0006,
+ "loss": 4.243929386138916,
+ "step": 5585
+ },
+ {
+ "epoch": 77.58689956331878,
+ "grad_norm": 0.016526037827134132,
+ "learning_rate": 0.0006,
+ "loss": 4.374139785766602,
+ "step": 5586
+ },
+ {
+ "epoch": 77.60087336244541,
+ "grad_norm": 0.01648339442908764,
+ "learning_rate": 0.0006,
+ "loss": 4.392425060272217,
+ "step": 5587
+ },
+ {
+ "epoch": 77.61484716157206,
+ "grad_norm": 0.015269213356077671,
+ "learning_rate": 0.0006,
+ "loss": 4.33986759185791,
+ "step": 5588
+ },
+ {
+ "epoch": 77.62882096069869,
+ "grad_norm": 0.01690507121384144,
+ "learning_rate": 0.0006,
+ "loss": 4.307476043701172,
+ "step": 5589
+ },
+ {
+ "epoch": 77.64279475982532,
+ "grad_norm": 0.016438685357570648,
+ "learning_rate": 0.0006,
+ "loss": 4.2406206130981445,
+ "step": 5590
+ },
+ {
+ "epoch": 77.65676855895197,
+ "grad_norm": 0.0157600287348032,
+ "learning_rate": 0.0006,
+ "loss": 4.314070224761963,
+ "step": 5591
+ },
+ {
+ "epoch": 77.6707423580786,
+ "grad_norm": 0.015127947553992271,
+ "learning_rate": 0.0006,
+ "loss": 4.22475528717041,
+ "step": 5592
+ },
+ {
+ "epoch": 77.68471615720524,
+ "grad_norm": 0.016632677987217903,
+ "learning_rate": 0.0006,
+ "loss": 4.26954984664917,
+ "step": 5593
+ },
+ {
+ "epoch": 77.69868995633188,
+ "grad_norm": 0.013150460086762905,
+ "learning_rate": 0.0006,
+ "loss": 4.369131088256836,
+ "step": 5594
+ },
+ {
+ "epoch": 77.71266375545852,
+ "grad_norm": 0.014513005502521992,
+ "learning_rate": 0.0006,
+ "loss": 4.3248796463012695,
+ "step": 5595
+ },
+ {
+ "epoch": 77.72663755458515,
+ "grad_norm": 0.014090991578996181,
+ "learning_rate": 0.0006,
+ "loss": 4.268381118774414,
+ "step": 5596
+ },
+ {
+ "epoch": 77.74061135371178,
+ "grad_norm": 0.01633097417652607,
+ "learning_rate": 0.0006,
+ "loss": 4.238439083099365,
+ "step": 5597
+ },
+ {
+ "epoch": 77.75458515283843,
+ "grad_norm": 0.01710933819413185,
+ "learning_rate": 0.0006,
+ "loss": 4.399697780609131,
+ "step": 5598
+ },
+ {
+ "epoch": 77.76855895196506,
+ "grad_norm": 0.019251767545938492,
+ "learning_rate": 0.0006,
+ "loss": 4.290450096130371,
+ "step": 5599
+ },
+ {
+ "epoch": 77.78253275109171,
+ "grad_norm": 0.01849171705543995,
+ "learning_rate": 0.0006,
+ "loss": 4.267734527587891,
+ "step": 5600
+ },
+ {
+ "epoch": 77.79650655021834,
+ "grad_norm": 0.014173115603625774,
+ "learning_rate": 0.0006,
+ "loss": 4.288799285888672,
+ "step": 5601
+ },
+ {
+ "epoch": 77.81048034934499,
+ "grad_norm": 0.014572718180716038,
+ "learning_rate": 0.0006,
+ "loss": 4.3790435791015625,
+ "step": 5602
+ },
+ {
+ "epoch": 77.82445414847162,
+ "grad_norm": 0.016375336796045303,
+ "learning_rate": 0.0006,
+ "loss": 4.231725692749023,
+ "step": 5603
+ },
+ {
+ "epoch": 77.83842794759825,
+ "grad_norm": 0.017326701432466507,
+ "learning_rate": 0.0006,
+ "loss": 4.211495399475098,
+ "step": 5604
+ },
+ {
+ "epoch": 77.8524017467249,
+ "grad_norm": 0.016399390995502472,
+ "learning_rate": 0.0006,
+ "loss": 4.386699199676514,
+ "step": 5605
+ },
+ {
+ "epoch": 77.86637554585153,
+ "grad_norm": 0.018269989639520645,
+ "learning_rate": 0.0006,
+ "loss": 4.254856109619141,
+ "step": 5606
+ },
+ {
+ "epoch": 77.88034934497817,
+ "grad_norm": 0.018443435430526733,
+ "learning_rate": 0.0006,
+ "loss": 4.391961097717285,
+ "step": 5607
+ },
+ {
+ "epoch": 77.8943231441048,
+ "grad_norm": 0.01784636080265045,
+ "learning_rate": 0.0006,
+ "loss": 4.273881912231445,
+ "step": 5608
+ },
+ {
+ "epoch": 77.90829694323143,
+ "grad_norm": 0.01782994158565998,
+ "learning_rate": 0.0006,
+ "loss": 4.26833963394165,
+ "step": 5609
+ },
+ {
+ "epoch": 77.92227074235808,
+ "grad_norm": 0.016260741278529167,
+ "learning_rate": 0.0006,
+ "loss": 4.306807994842529,
+ "step": 5610
+ },
+ {
+ "epoch": 77.93624454148471,
+ "grad_norm": 0.015482903458178043,
+ "learning_rate": 0.0006,
+ "loss": 4.187826633453369,
+ "step": 5611
+ },
+ {
+ "epoch": 77.95021834061136,
+ "grad_norm": 0.015193257480859756,
+ "learning_rate": 0.0006,
+ "loss": 4.286060333251953,
+ "step": 5612
+ },
+ {
+ "epoch": 77.96419213973799,
+ "grad_norm": 0.015110115520656109,
+ "learning_rate": 0.0006,
+ "loss": 4.2828192710876465,
+ "step": 5613
+ },
+ {
+ "epoch": 77.97816593886463,
+ "grad_norm": 0.015973486006259918,
+ "learning_rate": 0.0006,
+ "loss": 4.317939758300781,
+ "step": 5614
+ },
+ {
+ "epoch": 77.99213973799127,
+ "grad_norm": 0.017363594844937325,
+ "learning_rate": 0.0006,
+ "loss": 4.287721633911133,
+ "step": 5615
+ },
+ {
+ "epoch": 78.0,
+ "grad_norm": 0.015277642756700516,
+ "learning_rate": 0.0006,
+ "loss": 4.356618881225586,
+ "step": 5616
+ },
+ {
+ "epoch": 78.0,
+ "eval_loss": 4.642678737640381,
+ "eval_runtime": 56.7335,
+ "eval_samples_per_second": 43.043,
+ "eval_steps_per_second": 1.357,
+ "step": 5616
+ },
+ {
+ "epoch": 78.01397379912663,
+ "grad_norm": 0.016794051975011826,
+ "learning_rate": 0.0006,
+ "loss": 4.279930591583252,
+ "step": 5617
+ },
+ {
+ "epoch": 78.02794759825328,
+ "grad_norm": 0.01632673107087612,
+ "learning_rate": 0.0006,
+ "loss": 4.235943794250488,
+ "step": 5618
+ },
+ {
+ "epoch": 78.04192139737991,
+ "grad_norm": 0.01437876746058464,
+ "learning_rate": 0.0006,
+ "loss": 4.2542243003845215,
+ "step": 5619
+ },
+ {
+ "epoch": 78.05589519650655,
+ "grad_norm": 0.014993119053542614,
+ "learning_rate": 0.0006,
+ "loss": 4.262046813964844,
+ "step": 5620
+ },
+ {
+ "epoch": 78.06986899563319,
+ "grad_norm": 0.015721119940280914,
+ "learning_rate": 0.0006,
+ "loss": 4.173436164855957,
+ "step": 5621
+ },
+ {
+ "epoch": 78.08384279475983,
+ "grad_norm": 0.01334085687994957,
+ "learning_rate": 0.0006,
+ "loss": 4.322175979614258,
+ "step": 5622
+ },
+ {
+ "epoch": 78.09781659388646,
+ "grad_norm": 0.015561388805508614,
+ "learning_rate": 0.0006,
+ "loss": 4.242125988006592,
+ "step": 5623
+ },
+ {
+ "epoch": 78.1117903930131,
+ "grad_norm": 0.014418602921068668,
+ "learning_rate": 0.0006,
+ "loss": 4.266186714172363,
+ "step": 5624
+ },
+ {
+ "epoch": 78.12576419213974,
+ "grad_norm": 0.013739909045398235,
+ "learning_rate": 0.0006,
+ "loss": 4.231633186340332,
+ "step": 5625
+ },
+ {
+ "epoch": 78.13973799126637,
+ "grad_norm": 0.01402067206799984,
+ "learning_rate": 0.0006,
+ "loss": 4.188406944274902,
+ "step": 5626
+ },
+ {
+ "epoch": 78.15371179039302,
+ "grad_norm": 0.013639481738209724,
+ "learning_rate": 0.0006,
+ "loss": 4.1586594581604,
+ "step": 5627
+ },
+ {
+ "epoch": 78.16768558951965,
+ "grad_norm": 0.012858159840106964,
+ "learning_rate": 0.0006,
+ "loss": 4.360448360443115,
+ "step": 5628
+ },
+ {
+ "epoch": 78.18165938864628,
+ "grad_norm": 0.013089739717543125,
+ "learning_rate": 0.0006,
+ "loss": 4.344371795654297,
+ "step": 5629
+ },
+ {
+ "epoch": 78.19563318777293,
+ "grad_norm": 0.01382314320653677,
+ "learning_rate": 0.0006,
+ "loss": 4.20378303527832,
+ "step": 5630
+ },
+ {
+ "epoch": 78.20960698689956,
+ "grad_norm": 0.014206619001924992,
+ "learning_rate": 0.0006,
+ "loss": 4.239891529083252,
+ "step": 5631
+ },
+ {
+ "epoch": 78.2235807860262,
+ "grad_norm": 0.01472168043255806,
+ "learning_rate": 0.0006,
+ "loss": 4.282382488250732,
+ "step": 5632
+ },
+ {
+ "epoch": 78.23755458515284,
+ "grad_norm": 0.014237025752663612,
+ "learning_rate": 0.0006,
+ "loss": 4.241545677185059,
+ "step": 5633
+ },
+ {
+ "epoch": 78.25152838427948,
+ "grad_norm": 0.013231472112238407,
+ "learning_rate": 0.0006,
+ "loss": 4.184896469116211,
+ "step": 5634
+ },
+ {
+ "epoch": 78.26550218340611,
+ "grad_norm": 0.013538471423089504,
+ "learning_rate": 0.0006,
+ "loss": 4.325181007385254,
+ "step": 5635
+ },
+ {
+ "epoch": 78.27947598253274,
+ "grad_norm": 0.013541961088776588,
+ "learning_rate": 0.0006,
+ "loss": 4.368722915649414,
+ "step": 5636
+ },
+ {
+ "epoch": 78.29344978165939,
+ "grad_norm": 0.013446723110973835,
+ "learning_rate": 0.0006,
+ "loss": 4.301422595977783,
+ "step": 5637
+ },
+ {
+ "epoch": 78.30742358078602,
+ "grad_norm": 0.014753809198737144,
+ "learning_rate": 0.0006,
+ "loss": 4.259127616882324,
+ "step": 5638
+ },
+ {
+ "epoch": 78.32139737991267,
+ "grad_norm": 0.013256008736789227,
+ "learning_rate": 0.0006,
+ "loss": 4.284032821655273,
+ "step": 5639
+ },
+ {
+ "epoch": 78.3353711790393,
+ "grad_norm": 0.014272629283368587,
+ "learning_rate": 0.0006,
+ "loss": 4.234989166259766,
+ "step": 5640
+ },
+ {
+ "epoch": 78.34934497816595,
+ "grad_norm": 0.014756826683878899,
+ "learning_rate": 0.0006,
+ "loss": 4.143499374389648,
+ "step": 5641
+ },
+ {
+ "epoch": 78.36331877729258,
+ "grad_norm": 0.016549205407500267,
+ "learning_rate": 0.0006,
+ "loss": 4.311034202575684,
+ "step": 5642
+ },
+ {
+ "epoch": 78.37729257641921,
+ "grad_norm": 0.017735283821821213,
+ "learning_rate": 0.0006,
+ "loss": 4.246767520904541,
+ "step": 5643
+ },
+ {
+ "epoch": 78.39126637554585,
+ "grad_norm": 0.017370613291859627,
+ "learning_rate": 0.0006,
+ "loss": 4.2460832595825195,
+ "step": 5644
+ },
+ {
+ "epoch": 78.40524017467249,
+ "grad_norm": 0.014642846770584583,
+ "learning_rate": 0.0006,
+ "loss": 4.303811550140381,
+ "step": 5645
+ },
+ {
+ "epoch": 78.41921397379913,
+ "grad_norm": 0.013445726595818996,
+ "learning_rate": 0.0006,
+ "loss": 4.25921630859375,
+ "step": 5646
+ },
+ {
+ "epoch": 78.43318777292576,
+ "grad_norm": 0.014254065230488777,
+ "learning_rate": 0.0006,
+ "loss": 4.129446506500244,
+ "step": 5647
+ },
+ {
+ "epoch": 78.44716157205241,
+ "grad_norm": 0.01883019506931305,
+ "learning_rate": 0.0006,
+ "loss": 4.191716194152832,
+ "step": 5648
+ },
+ {
+ "epoch": 78.46113537117904,
+ "grad_norm": 0.0204437468200922,
+ "learning_rate": 0.0006,
+ "loss": 4.333085536956787,
+ "step": 5649
+ },
+ {
+ "epoch": 78.47510917030567,
+ "grad_norm": 0.01652318798005581,
+ "learning_rate": 0.0006,
+ "loss": 4.135462760925293,
+ "step": 5650
+ },
+ {
+ "epoch": 78.48908296943232,
+ "grad_norm": 0.012753551825881004,
+ "learning_rate": 0.0006,
+ "loss": 4.195051193237305,
+ "step": 5651
+ },
+ {
+ "epoch": 78.50305676855895,
+ "grad_norm": 0.01499118935316801,
+ "learning_rate": 0.0006,
+ "loss": 4.157092094421387,
+ "step": 5652
+ },
+ {
+ "epoch": 78.5170305676856,
+ "grad_norm": 0.015383945778012276,
+ "learning_rate": 0.0006,
+ "loss": 4.253818035125732,
+ "step": 5653
+ },
+ {
+ "epoch": 78.53100436681223,
+ "grad_norm": 0.015721332281827927,
+ "learning_rate": 0.0006,
+ "loss": 4.228775978088379,
+ "step": 5654
+ },
+ {
+ "epoch": 78.54497816593886,
+ "grad_norm": 0.014595463871955872,
+ "learning_rate": 0.0006,
+ "loss": 4.135197639465332,
+ "step": 5655
+ },
+ {
+ "epoch": 78.5589519650655,
+ "grad_norm": 0.015059644356369972,
+ "learning_rate": 0.0006,
+ "loss": 4.255614280700684,
+ "step": 5656
+ },
+ {
+ "epoch": 78.57292576419214,
+ "grad_norm": 0.013833344914019108,
+ "learning_rate": 0.0006,
+ "loss": 4.301250457763672,
+ "step": 5657
+ },
+ {
+ "epoch": 78.58689956331878,
+ "grad_norm": 0.013665376231074333,
+ "learning_rate": 0.0006,
+ "loss": 4.1734232902526855,
+ "step": 5658
+ },
+ {
+ "epoch": 78.60087336244541,
+ "grad_norm": 0.01539086177945137,
+ "learning_rate": 0.0006,
+ "loss": 4.221459865570068,
+ "step": 5659
+ },
+ {
+ "epoch": 78.61484716157206,
+ "grad_norm": 0.013313917443156242,
+ "learning_rate": 0.0006,
+ "loss": 4.255619049072266,
+ "step": 5660
+ },
+ {
+ "epoch": 78.62882096069869,
+ "grad_norm": 0.01447741687297821,
+ "learning_rate": 0.0006,
+ "loss": 4.298219680786133,
+ "step": 5661
+ },
+ {
+ "epoch": 78.64279475982532,
+ "grad_norm": 0.012796915136277676,
+ "learning_rate": 0.0006,
+ "loss": 4.2542405128479,
+ "step": 5662
+ },
+ {
+ "epoch": 78.65676855895197,
+ "grad_norm": 0.014058141969144344,
+ "learning_rate": 0.0006,
+ "loss": 4.284477233886719,
+ "step": 5663
+ },
+ {
+ "epoch": 78.6707423580786,
+ "grad_norm": 0.012705069966614246,
+ "learning_rate": 0.0006,
+ "loss": 4.2826666831970215,
+ "step": 5664
+ },
+ {
+ "epoch": 78.68471615720524,
+ "grad_norm": 0.013245890848338604,
+ "learning_rate": 0.0006,
+ "loss": 4.371612548828125,
+ "step": 5665
+ },
+ {
+ "epoch": 78.69868995633188,
+ "grad_norm": 0.014192677102982998,
+ "learning_rate": 0.0006,
+ "loss": 4.267344951629639,
+ "step": 5666
+ },
+ {
+ "epoch": 78.71266375545852,
+ "grad_norm": 0.015346777625381947,
+ "learning_rate": 0.0006,
+ "loss": 4.222797393798828,
+ "step": 5667
+ },
+ {
+ "epoch": 78.72663755458515,
+ "grad_norm": 0.014068621210753918,
+ "learning_rate": 0.0006,
+ "loss": 4.172964572906494,
+ "step": 5668
+ },
+ {
+ "epoch": 78.74061135371178,
+ "grad_norm": 0.014364945702254772,
+ "learning_rate": 0.0006,
+ "loss": 4.259319305419922,
+ "step": 5669
+ },
+ {
+ "epoch": 78.75458515283843,
+ "grad_norm": 0.017337387427687645,
+ "learning_rate": 0.0006,
+ "loss": 4.29533576965332,
+ "step": 5670
+ },
+ {
+ "epoch": 78.76855895196506,
+ "grad_norm": 0.017416715621948242,
+ "learning_rate": 0.0006,
+ "loss": 4.340364456176758,
+ "step": 5671
+ },
+ {
+ "epoch": 78.78253275109171,
+ "grad_norm": 0.015925846993923187,
+ "learning_rate": 0.0006,
+ "loss": 4.177454948425293,
+ "step": 5672
+ },
+ {
+ "epoch": 78.79650655021834,
+ "grad_norm": 0.013130443170666695,
+ "learning_rate": 0.0006,
+ "loss": 4.277454853057861,
+ "step": 5673
+ },
+ {
+ "epoch": 78.81048034934499,
+ "grad_norm": 0.014628687873482704,
+ "learning_rate": 0.0006,
+ "loss": 4.273757457733154,
+ "step": 5674
+ },
+ {
+ "epoch": 78.82445414847162,
+ "grad_norm": 0.01601659320294857,
+ "learning_rate": 0.0006,
+ "loss": 4.281774520874023,
+ "step": 5675
+ },
+ {
+ "epoch": 78.83842794759825,
+ "grad_norm": 0.015443485230207443,
+ "learning_rate": 0.0006,
+ "loss": 4.199937343597412,
+ "step": 5676
+ },
+ {
+ "epoch": 78.8524017467249,
+ "grad_norm": 0.013934246264398098,
+ "learning_rate": 0.0006,
+ "loss": 4.331369876861572,
+ "step": 5677
+ },
+ {
+ "epoch": 78.86637554585153,
+ "grad_norm": 0.01572083681821823,
+ "learning_rate": 0.0006,
+ "loss": 4.2828474044799805,
+ "step": 5678
+ },
+ {
+ "epoch": 78.88034934497817,
+ "grad_norm": 0.013444120064377785,
+ "learning_rate": 0.0006,
+ "loss": 4.183859825134277,
+ "step": 5679
+ },
+ {
+ "epoch": 78.8943231441048,
+ "grad_norm": 0.01591254211962223,
+ "learning_rate": 0.0006,
+ "loss": 4.295256614685059,
+ "step": 5680
+ },
+ {
+ "epoch": 78.90829694323143,
+ "grad_norm": 0.017556684091687202,
+ "learning_rate": 0.0006,
+ "loss": 4.132272720336914,
+ "step": 5681
+ },
+ {
+ "epoch": 78.92227074235808,
+ "grad_norm": 0.018184322863817215,
+ "learning_rate": 0.0006,
+ "loss": 4.251728057861328,
+ "step": 5682
+ },
+ {
+ "epoch": 78.93624454148471,
+ "grad_norm": 0.016207285225391388,
+ "learning_rate": 0.0006,
+ "loss": 4.205544471740723,
+ "step": 5683
+ },
+ {
+ "epoch": 78.95021834061136,
+ "grad_norm": 0.014478376135230064,
+ "learning_rate": 0.0006,
+ "loss": 4.182462215423584,
+ "step": 5684
+ },
+ {
+ "epoch": 78.96419213973799,
+ "grad_norm": 0.014255235902965069,
+ "learning_rate": 0.0006,
+ "loss": 4.253366947174072,
+ "step": 5685
+ },
+ {
+ "epoch": 78.97816593886463,
+ "grad_norm": 0.01524396613240242,
+ "learning_rate": 0.0006,
+ "loss": 4.221786022186279,
+ "step": 5686
+ },
+ {
+ "epoch": 78.99213973799127,
+ "grad_norm": 0.015057693235576153,
+ "learning_rate": 0.0006,
+ "loss": 4.208454608917236,
+ "step": 5687
+ },
+ {
+ "epoch": 79.0,
+ "grad_norm": 0.015478022396564484,
+ "learning_rate": 0.0006,
+ "loss": 4.336383819580078,
+ "step": 5688
+ },
+ {
+ "epoch": 79.0,
+ "eval_loss": 4.641749382019043,
+ "eval_runtime": 56.8167,
+ "eval_samples_per_second": 42.98,
+ "eval_steps_per_second": 1.355,
+ "step": 5688
+ },
+ {
+ "epoch": 79.01397379912663,
+ "grad_norm": 0.014143316075205803,
+ "learning_rate": 0.0006,
+ "loss": 4.260974884033203,
+ "step": 5689
+ },
+ {
+ "epoch": 79.02794759825328,
+ "grad_norm": 0.013120883144438267,
+ "learning_rate": 0.0006,
+ "loss": 4.139886856079102,
+ "step": 5690
+ },
+ {
+ "epoch": 79.04192139737991,
+ "grad_norm": 0.015019661746919155,
+ "learning_rate": 0.0006,
+ "loss": 4.267004013061523,
+ "step": 5691
+ },
+ {
+ "epoch": 79.05589519650655,
+ "grad_norm": 0.014891284517943859,
+ "learning_rate": 0.0006,
+ "loss": 4.205019950866699,
+ "step": 5692
+ },
+ {
+ "epoch": 79.06986899563319,
+ "grad_norm": 0.015527728945016861,
+ "learning_rate": 0.0006,
+ "loss": 4.1513566970825195,
+ "step": 5693
+ },
+ {
+ "epoch": 79.08384279475983,
+ "grad_norm": 0.017527181655168533,
+ "learning_rate": 0.0006,
+ "loss": 4.148828983306885,
+ "step": 5694
+ },
+ {
+ "epoch": 79.09781659388646,
+ "grad_norm": 0.01632206328213215,
+ "learning_rate": 0.0006,
+ "loss": 4.195789813995361,
+ "step": 5695
+ },
+ {
+ "epoch": 79.1117903930131,
+ "grad_norm": 0.01590503193438053,
+ "learning_rate": 0.0006,
+ "loss": 4.199155330657959,
+ "step": 5696
+ },
+ {
+ "epoch": 79.12576419213974,
+ "grad_norm": 0.016489997506141663,
+ "learning_rate": 0.0006,
+ "loss": 4.155154705047607,
+ "step": 5697
+ },
+ {
+ "epoch": 79.13973799126637,
+ "grad_norm": 0.0161368977278471,
+ "learning_rate": 0.0006,
+ "loss": 4.164384841918945,
+ "step": 5698
+ },
+ {
+ "epoch": 79.15371179039302,
+ "grad_norm": 0.018574809655547142,
+ "learning_rate": 0.0006,
+ "loss": 4.1898193359375,
+ "step": 5699
+ },
+ {
+ "epoch": 79.16768558951965,
+ "grad_norm": 0.018987983465194702,
+ "learning_rate": 0.0006,
+ "loss": 4.219719409942627,
+ "step": 5700
+ },
+ {
+ "epoch": 79.18165938864628,
+ "grad_norm": 0.01668606884777546,
+ "learning_rate": 0.0006,
+ "loss": 4.152734756469727,
+ "step": 5701
+ },
+ {
+ "epoch": 79.19563318777293,
+ "grad_norm": 0.014046299271285534,
+ "learning_rate": 0.0006,
+ "loss": 4.152964115142822,
+ "step": 5702
+ },
+ {
+ "epoch": 79.20960698689956,
+ "grad_norm": 0.01628032699227333,
+ "learning_rate": 0.0006,
+ "loss": 4.299479961395264,
+ "step": 5703
+ },
+ {
+ "epoch": 79.2235807860262,
+ "grad_norm": 0.017998795956373215,
+ "learning_rate": 0.0006,
+ "loss": 4.231203079223633,
+ "step": 5704
+ },
+ {
+ "epoch": 79.23755458515284,
+ "grad_norm": 0.014799647964537144,
+ "learning_rate": 0.0006,
+ "loss": 4.3530192375183105,
+ "step": 5705
+ },
+ {
+ "epoch": 79.25152838427948,
+ "grad_norm": 0.014640678651630878,
+ "learning_rate": 0.0006,
+ "loss": 4.287322998046875,
+ "step": 5706
+ },
+ {
+ "epoch": 79.26550218340611,
+ "grad_norm": 0.0172616895288229,
+ "learning_rate": 0.0006,
+ "loss": 4.148009777069092,
+ "step": 5707
+ },
+ {
+ "epoch": 79.27947598253274,
+ "grad_norm": 0.01811790093779564,
+ "learning_rate": 0.0006,
+ "loss": 4.269433975219727,
+ "step": 5708
+ },
+ {
+ "epoch": 79.29344978165939,
+ "grad_norm": 0.01764000579714775,
+ "learning_rate": 0.0006,
+ "loss": 4.292638778686523,
+ "step": 5709
+ },
+ {
+ "epoch": 79.30742358078602,
+ "grad_norm": 0.018030935898423195,
+ "learning_rate": 0.0006,
+ "loss": 4.267889976501465,
+ "step": 5710
+ },
+ {
+ "epoch": 79.32139737991267,
+ "grad_norm": 0.017860891297459602,
+ "learning_rate": 0.0006,
+ "loss": 4.211110591888428,
+ "step": 5711
+ },
+ {
+ "epoch": 79.3353711790393,
+ "grad_norm": 0.017295073717832565,
+ "learning_rate": 0.0006,
+ "loss": 4.224645614624023,
+ "step": 5712
+ },
+ {
+ "epoch": 79.34934497816595,
+ "grad_norm": 0.015891101211309433,
+ "learning_rate": 0.0006,
+ "loss": 4.224797248840332,
+ "step": 5713
+ },
+ {
+ "epoch": 79.36331877729258,
+ "grad_norm": 0.014680417254567146,
+ "learning_rate": 0.0006,
+ "loss": 4.182242393493652,
+ "step": 5714
+ },
+ {
+ "epoch": 79.37729257641921,
+ "grad_norm": 0.016621140763163567,
+ "learning_rate": 0.0006,
+ "loss": 4.186334609985352,
+ "step": 5715
+ },
+ {
+ "epoch": 79.39126637554585,
+ "grad_norm": 0.016207249835133553,
+ "learning_rate": 0.0006,
+ "loss": 4.277074813842773,
+ "step": 5716
+ },
+ {
+ "epoch": 79.40524017467249,
+ "grad_norm": 0.015375478193163872,
+ "learning_rate": 0.0006,
+ "loss": 4.154317855834961,
+ "step": 5717
+ },
+ {
+ "epoch": 79.41921397379913,
+ "grad_norm": 0.013453228399157524,
+ "learning_rate": 0.0006,
+ "loss": 4.265270233154297,
+ "step": 5718
+ },
+ {
+ "epoch": 79.43318777292576,
+ "grad_norm": 0.014312694780528545,
+ "learning_rate": 0.0006,
+ "loss": 4.232220649719238,
+ "step": 5719
+ },
+ {
+ "epoch": 79.44716157205241,
+ "grad_norm": 0.014759554527699947,
+ "learning_rate": 0.0006,
+ "loss": 4.272275924682617,
+ "step": 5720
+ },
+ {
+ "epoch": 79.46113537117904,
+ "grad_norm": 0.015885505825281143,
+ "learning_rate": 0.0006,
+ "loss": 4.2229719161987305,
+ "step": 5721
+ },
+ {
+ "epoch": 79.47510917030567,
+ "grad_norm": 0.016516581177711487,
+ "learning_rate": 0.0006,
+ "loss": 4.200624465942383,
+ "step": 5722
+ },
+ {
+ "epoch": 79.48908296943232,
+ "grad_norm": 0.018254555761814117,
+ "learning_rate": 0.0006,
+ "loss": 4.19648551940918,
+ "step": 5723
+ },
+ {
+ "epoch": 79.50305676855895,
+ "grad_norm": 0.01467090006917715,
+ "learning_rate": 0.0006,
+ "loss": 4.135190486907959,
+ "step": 5724
+ },
+ {
+ "epoch": 79.5170305676856,
+ "grad_norm": 0.013474797829985619,
+ "learning_rate": 0.0006,
+ "loss": 4.311361312866211,
+ "step": 5725
+ },
+ {
+ "epoch": 79.53100436681223,
+ "grad_norm": 0.01742764748632908,
+ "learning_rate": 0.0006,
+ "loss": 4.1845903396606445,
+ "step": 5726
+ },
+ {
+ "epoch": 79.54497816593886,
+ "grad_norm": 0.018786882981657982,
+ "learning_rate": 0.0006,
+ "loss": 4.163942337036133,
+ "step": 5727
+ },
+ {
+ "epoch": 79.5589519650655,
+ "grad_norm": 0.017831219360232353,
+ "learning_rate": 0.0006,
+ "loss": 4.236685276031494,
+ "step": 5728
+ },
+ {
+ "epoch": 79.57292576419214,
+ "grad_norm": 0.016334986314177513,
+ "learning_rate": 0.0006,
+ "loss": 4.255853652954102,
+ "step": 5729
+ },
+ {
+ "epoch": 79.58689956331878,
+ "grad_norm": 0.0138487434014678,
+ "learning_rate": 0.0006,
+ "loss": 4.227874279022217,
+ "step": 5730
+ },
+ {
+ "epoch": 79.60087336244541,
+ "grad_norm": 0.01531760673969984,
+ "learning_rate": 0.0006,
+ "loss": 4.221432685852051,
+ "step": 5731
+ },
+ {
+ "epoch": 79.61484716157206,
+ "grad_norm": 0.01603361964225769,
+ "learning_rate": 0.0006,
+ "loss": 4.278326511383057,
+ "step": 5732
+ },
+ {
+ "epoch": 79.62882096069869,
+ "grad_norm": 0.013149895705282688,
+ "learning_rate": 0.0006,
+ "loss": 4.2474517822265625,
+ "step": 5733
+ },
+ {
+ "epoch": 79.64279475982532,
+ "grad_norm": 0.013428786769509315,
+ "learning_rate": 0.0006,
+ "loss": 4.1701178550720215,
+ "step": 5734
+ },
+ {
+ "epoch": 79.65676855895197,
+ "grad_norm": 0.014305585995316505,
+ "learning_rate": 0.0006,
+ "loss": 4.250778675079346,
+ "step": 5735
+ },
+ {
+ "epoch": 79.6707423580786,
+ "grad_norm": 0.014544566161930561,
+ "learning_rate": 0.0006,
+ "loss": 4.243236541748047,
+ "step": 5736
+ },
+ {
+ "epoch": 79.68471615720524,
+ "grad_norm": 0.013524414040148258,
+ "learning_rate": 0.0006,
+ "loss": 4.255457878112793,
+ "step": 5737
+ },
+ {
+ "epoch": 79.69868995633188,
+ "grad_norm": 0.014750641770660877,
+ "learning_rate": 0.0006,
+ "loss": 4.202816486358643,
+ "step": 5738
+ },
+ {
+ "epoch": 79.71266375545852,
+ "grad_norm": 0.015090773813426495,
+ "learning_rate": 0.0006,
+ "loss": 4.3053178787231445,
+ "step": 5739
+ },
+ {
+ "epoch": 79.72663755458515,
+ "grad_norm": 0.01606169529259205,
+ "learning_rate": 0.0006,
+ "loss": 4.224017143249512,
+ "step": 5740
+ },
+ {
+ "epoch": 79.74061135371178,
+ "grad_norm": 0.016303520649671555,
+ "learning_rate": 0.0006,
+ "loss": 4.190760612487793,
+ "step": 5741
+ },
+ {
+ "epoch": 79.75458515283843,
+ "grad_norm": 0.01465672068297863,
+ "learning_rate": 0.0006,
+ "loss": 4.24506950378418,
+ "step": 5742
+ },
+ {
+ "epoch": 79.76855895196506,
+ "grad_norm": 0.012616301886737347,
+ "learning_rate": 0.0006,
+ "loss": 4.174039840698242,
+ "step": 5743
+ },
+ {
+ "epoch": 79.78253275109171,
+ "grad_norm": 0.01358284242451191,
+ "learning_rate": 0.0006,
+ "loss": 4.236380577087402,
+ "step": 5744
+ },
+ {
+ "epoch": 79.79650655021834,
+ "grad_norm": 0.01557177770882845,
+ "learning_rate": 0.0006,
+ "loss": 4.102962493896484,
+ "step": 5745
+ },
+ {
+ "epoch": 79.81048034934499,
+ "grad_norm": 0.015584317035973072,
+ "learning_rate": 0.0006,
+ "loss": 4.242603302001953,
+ "step": 5746
+ },
+ {
+ "epoch": 79.82445414847162,
+ "grad_norm": 0.012597334571182728,
+ "learning_rate": 0.0006,
+ "loss": 4.202791690826416,
+ "step": 5747
+ },
+ {
+ "epoch": 79.83842794759825,
+ "grad_norm": 0.013722898438572884,
+ "learning_rate": 0.0006,
+ "loss": 4.204328536987305,
+ "step": 5748
+ },
+ {
+ "epoch": 79.8524017467249,
+ "grad_norm": 0.014669225551187992,
+ "learning_rate": 0.0006,
+ "loss": 4.164331912994385,
+ "step": 5749
+ },
+ {
+ "epoch": 79.86637554585153,
+ "grad_norm": 0.015012740157544613,
+ "learning_rate": 0.0006,
+ "loss": 4.260565280914307,
+ "step": 5750
+ },
+ {
+ "epoch": 79.88034934497817,
+ "grad_norm": 0.016026340425014496,
+ "learning_rate": 0.0006,
+ "loss": 4.299723148345947,
+ "step": 5751
+ },
+ {
+ "epoch": 79.8943231441048,
+ "grad_norm": 0.01626049168407917,
+ "learning_rate": 0.0006,
+ "loss": 4.207252502441406,
+ "step": 5752
+ },
+ {
+ "epoch": 79.90829694323143,
+ "grad_norm": 0.015415354631841183,
+ "learning_rate": 0.0006,
+ "loss": 4.151934623718262,
+ "step": 5753
+ },
+ {
+ "epoch": 79.92227074235808,
+ "grad_norm": 0.014158242382109165,
+ "learning_rate": 0.0006,
+ "loss": 4.3461995124816895,
+ "step": 5754
+ },
+ {
+ "epoch": 79.93624454148471,
+ "grad_norm": 0.01585628278553486,
+ "learning_rate": 0.0006,
+ "loss": 4.334761619567871,
+ "step": 5755
+ },
+ {
+ "epoch": 79.95021834061136,
+ "grad_norm": 0.013710759580135345,
+ "learning_rate": 0.0006,
+ "loss": 4.180401802062988,
+ "step": 5756
+ },
+ {
+ "epoch": 79.96419213973799,
+ "grad_norm": 0.014159204438328743,
+ "learning_rate": 0.0006,
+ "loss": 4.322331428527832,
+ "step": 5757
+ },
+ {
+ "epoch": 79.97816593886463,
+ "grad_norm": 0.015153195708990097,
+ "learning_rate": 0.0006,
+ "loss": 4.287478923797607,
+ "step": 5758
+ },
+ {
+ "epoch": 79.99213973799127,
+ "grad_norm": 0.014100227504968643,
+ "learning_rate": 0.0006,
+ "loss": 4.231807708740234,
+ "step": 5759
+ },
+ {
+ "epoch": 80.0,
+ "grad_norm": 0.015248682349920273,
+ "learning_rate": 0.0006,
+ "loss": 4.261106014251709,
+ "step": 5760
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 28800,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 400,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": true,
+ "should_log": false,
+ "should_save": false,
+ "should_training_stop": false
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 2.455690689065779e+18,
+ "train_batch_size": 8,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-5760/training_args.bin b/runs/l2r50-i2-fulle-lm/checkpoint-5760/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..cad232831c222cc7f8b938b8bbe42f61e7d89c3a
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-5760/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:f7c40e3bfa2f5ac1450ca1243e3f2a895d44a06ed5995e6aefd97f78d229d8c8
+size 4792
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6048/chat_template.jinja b/runs/l2r50-i2-fulle-lm/checkpoint-6048/chat_template.jinja
new file mode 100644
index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-6048/chat_template.jinja
@@ -0,0 +1,85 @@
+{%- if tools %}
+ {{- '<|im_start|>system\n' }}
+ {%- if messages[0].role == 'system' %}
+ {{- messages[0].content + '\n\n' }}
+ {%- endif %}
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }}
+ {%- for tool in tools %}
+ {{- "\n" }}
+ {{- tool | tojson }}
+ {%- endfor %}
+ {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }}
+{%- else %}
+ {%- if messages[0].role == 'system' %}
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
+ {%- endif %}
+{%- endif %}
+{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
+{%- for message in messages[::-1] %}
+ {%- set index = (messages|length - 1) - loop.index0 %}
+ {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %}
+ {%- set ns.multi_step_tool = false %}
+ {%- set ns.last_query_index = index %}
+ {%- endif %}
+{%- endfor %}
+{%- for message in messages %}
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
+ {%- elif message.role == "assistant" %}
+ {%- set content = message.content %}
+ {%- set reasoning_content = '' %}
+ {%- if message.reasoning_content is defined and message.reasoning_content is not none %}
+ {%- set reasoning_content = message.reasoning_content %}
+ {%- else %}
+ {%- if '' in message.content %}
+ {%- set content = message.content.split('')[-1].lstrip('\n') %}
+ {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %}
+ {%- endif %}
+ {%- endif %}
+ {%- if loop.index0 > ns.last_query_index %}
+ {%- if loop.last or (not loop.last and reasoning_content) %}
+ {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- if message.tool_calls %}
+ {%- for tool_call in message.tool_calls %}
+ {%- if (loop.first and content) or (not loop.first) %}
+ {{- '\n' }}
+ {%- endif %}
+ {%- if tool_call.function %}
+ {%- set tool_call = tool_call.function %}
+ {%- endif %}
+ {{- '\n{"name": "' }}
+ {{- tool_call.name }}
+ {{- '", "arguments": ' }}
+ {%- if tool_call.arguments is string %}
+ {{- tool_call.arguments }}
+ {%- else %}
+ {{- tool_call.arguments | tojson }}
+ {%- endif %}
+ {{- '}\n' }}
+ {%- endfor %}
+ {%- endif %}
+ {{- '<|im_end|>\n' }}
+ {%- elif message.role == "tool" %}
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
+ {{- '<|im_start|>user' }}
+ {%- endif %}
+ {{- '\n\n' }}
+ {{- message.content }}
+ {{- '\n' }}
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
+ {{- '<|im_end|>\n' }}
+ {%- endif %}
+ {%- endif %}
+{%- endfor %}
+{%- if add_generation_prompt %}
+ {{- '<|im_start|>assistant\n' }}
+ {%- if enable_thinking is defined and enable_thinking is false %}
+ {{- '\n\n\n\n' }}
+ {%- endif %}
+{%- endif %}
\ No newline at end of file
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6048/config.json b/runs/l2r50-i2-fulle-lm/checkpoint-6048/config.json
new file mode 100644
index 0000000000000000000000000000000000000000..22f250953c6ce8fc91ee4411856dd73041a415e2
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-6048/config.json
@@ -0,0 +1,32 @@
+{
+ "architectures": [
+ "LlamaForCausalLM"
+ ],
+ "attention_bias": false,
+ "attention_dropout": 0.0,
+ "bos_token_id": null,
+ "dtype": "float32",
+ "eos_token_id": 151645,
+ "head_dim": 128,
+ "hidden_act": "silu",
+ "hidden_size": 512,
+ "initializer_range": 0.02,
+ "intermediate_size": 1536,
+ "max_position_embeddings": 2048,
+ "mlp_bias": false,
+ "model_type": "llama",
+ "num_attention_heads": 4,
+ "num_hidden_layers": 20,
+ "num_key_value_heads": 4,
+ "pad_token_id": 151645,
+ "pretraining_tp": 1,
+ "rms_norm_eps": 1e-06,
+ "rope_parameters": {
+ "rope_theta": 10000.0,
+ "rope_type": "default"
+ },
+ "tie_word_embeddings": true,
+ "transformers_version": "5.5.0",
+ "use_cache": false,
+ "vocab_size": 151673
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6048/generation_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-6048/generation_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-6048/generation_config.json
@@ -0,0 +1,11 @@
+{
+ "_from_model_config": true,
+ "eos_token_id": [
+ 151645
+ ],
+ "output_attentions": false,
+ "output_hidden_states": false,
+ "pad_token_id": 151645,
+ "transformers_version": "5.5.0",
+ "use_cache": true
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6048/model.safetensors b/runs/l2r50-i2-fulle-lm/checkpoint-6048/model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..6ad65be1b697f018928914a93b854e0496abad24
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-6048/model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:bc2cdfa23e858d40673eea21eddfd41cda0e7275b967ad1d7cc7cfc978436642
+size 583360328
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6048/optimizer.pt b/runs/l2r50-i2-fulle-lm/checkpoint-6048/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..3dcba1342bf1b37a3167694c7a637b82ea650136
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-6048/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:02943afd569dead523bc35a26b6f3fd542955ac7b567d4ef39df4b1aa8c2d59a
+size 1166833530
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6048/rng_state_0.pth b/runs/l2r50-i2-fulle-lm/checkpoint-6048/rng_state_0.pth
new file mode 100644
index 0000000000000000000000000000000000000000..312cfda1b3b5a4c030872e90729af7fda6414a64
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-6048/rng_state_0.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:1488a6864d0d399732d11ecf9366384d524ff67d5ccdb266991648ccbfd7b9bc
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6048/rng_state_1.pth b/runs/l2r50-i2-fulle-lm/checkpoint-6048/rng_state_1.pth
new file mode 100644
index 0000000000000000000000000000000000000000..015687cd7430dbe642bb8b7450a399b412a88c06
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-6048/rng_state_1.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:155a2a845fd8d4bda671f5716878c0564ba65e425b4af3efa44ce4c4b04e3988
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6048/rng_state_2.pth b/runs/l2r50-i2-fulle-lm/checkpoint-6048/rng_state_2.pth
new file mode 100644
index 0000000000000000000000000000000000000000..329ea246992b66631e9ecb4e675f05d70753f65b
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-6048/rng_state_2.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:79f726cea75304e98d91d809622dae58023fec445e7eeed8e7c7165af5871843
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6048/rng_state_3.pth b/runs/l2r50-i2-fulle-lm/checkpoint-6048/rng_state_3.pth
new file mode 100644
index 0000000000000000000000000000000000000000..63952880ff4ff8278c9056ac3378e4fd623b5e74
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-6048/rng_state_3.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:5a2073d8191537a847cb1080353e4ac586e2f8357bba6e7741b41fb002ce8e07
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6048/scheduler.pt b/runs/l2r50-i2-fulle-lm/checkpoint-6048/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..762789ce7e31cc09dc831dc70530516d1f1e01b2
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-6048/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:100b47a764841837311a60bed293b8e5dd3c3501d4d1da6c3b20120bb6630ddd
+size 1064
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6048/tokenizer.json b/runs/l2r50-i2-fulle-lm/checkpoint-6048/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..53786016f070e11184813f3c83101dc5e93e392e
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-6048/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:c3c63fda1646a555448e8f03662542e04b11d103721319e4d9b267d7da03ceb8
+size 11424463
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6048/tokenizer_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-6048/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..98e2fb9b2cc54785dc9d871b8e503f8b8f2defe7
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-6048/tokenizer_config.json
@@ -0,0 +1,20 @@
+{
+ "add_prefix_space": false,
+ "backend": "tokenizers",
+ "bos_token": null,
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "<|im_end|>",
+ "errors": "replace",
+ "extra_special_tokens": [
+ "<|l2r_pred|>",
+ "<|r2l_pred|>",
+ "<|next_1_pred|>",
+ "<|next_2_pred|>"
+ ],
+ "is_local": false,
+ "model_max_length": 131072,
+ "pad_token": "<|im_end|>",
+ "split_special_tokens": false,
+ "tokenizer_class": "Qwen2Tokenizer",
+ "unk_token": null
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6048/trainer_state.json b/runs/l2r50-i2-fulle-lm/checkpoint-6048/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..5369c2ee5a37252b276cd1b917005258493c9dcf
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-6048/trainer_state.json
@@ -0,0 +1,43034 @@
+{
+ "best_global_step": null,
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 84.0,
+ "eval_steps": 500,
+ "global_step": 6048,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.013973799126637555,
+ "grad_norm": 0.1739506870508194,
+ "learning_rate": 0.0,
+ "loss": 11.99299430847168,
+ "step": 1
+ },
+ {
+ "epoch": 0.02794759825327511,
+ "grad_norm": 0.17216597497463226,
+ "learning_rate": 5.999999999999999e-06,
+ "loss": 11.988704681396484,
+ "step": 2
+ },
+ {
+ "epoch": 0.04192139737991266,
+ "grad_norm": 0.17459715902805328,
+ "learning_rate": 1.1999999999999999e-05,
+ "loss": 11.945581436157227,
+ "step": 3
+ },
+ {
+ "epoch": 0.05589519650655022,
+ "grad_norm": 0.17577414214611053,
+ "learning_rate": 1.7999999999999997e-05,
+ "loss": 11.859772682189941,
+ "step": 4
+ },
+ {
+ "epoch": 0.06986899563318777,
+ "grad_norm": 0.16163651645183563,
+ "learning_rate": 2.3999999999999997e-05,
+ "loss": 11.744415283203125,
+ "step": 5
+ },
+ {
+ "epoch": 0.08384279475982533,
+ "grad_norm": 0.14409932494163513,
+ "learning_rate": 2.9999999999999997e-05,
+ "loss": 11.618099212646484,
+ "step": 6
+ },
+ {
+ "epoch": 0.09781659388646288,
+ "grad_norm": 0.13079848885536194,
+ "learning_rate": 3.5999999999999994e-05,
+ "loss": 11.505935668945312,
+ "step": 7
+ },
+ {
+ "epoch": 0.11179039301310044,
+ "grad_norm": 0.11964613944292068,
+ "learning_rate": 4.2e-05,
+ "loss": 11.416532516479492,
+ "step": 8
+ },
+ {
+ "epoch": 0.125764192139738,
+ "grad_norm": 0.11345142871141434,
+ "learning_rate": 4.7999999999999994e-05,
+ "loss": 11.33206558227539,
+ "step": 9
+ },
+ {
+ "epoch": 0.13973799126637554,
+ "grad_norm": 0.11033952981233597,
+ "learning_rate": 5.399999999999999e-05,
+ "loss": 11.259052276611328,
+ "step": 10
+ },
+ {
+ "epoch": 0.1537117903930131,
+ "grad_norm": 0.10776489973068237,
+ "learning_rate": 5.9999999999999995e-05,
+ "loss": 11.203384399414062,
+ "step": 11
+ },
+ {
+ "epoch": 0.16768558951965065,
+ "grad_norm": 0.10721946507692337,
+ "learning_rate": 6.599999999999999e-05,
+ "loss": 11.153295516967773,
+ "step": 12
+ },
+ {
+ "epoch": 0.18165938864628822,
+ "grad_norm": 0.10673430562019348,
+ "learning_rate": 7.199999999999999e-05,
+ "loss": 11.110200881958008,
+ "step": 13
+ },
+ {
+ "epoch": 0.19563318777292577,
+ "grad_norm": 0.10630254447460175,
+ "learning_rate": 7.8e-05,
+ "loss": 11.067477226257324,
+ "step": 14
+ },
+ {
+ "epoch": 0.2096069868995633,
+ "grad_norm": 0.10631615668535233,
+ "learning_rate": 8.4e-05,
+ "loss": 11.022941589355469,
+ "step": 15
+ },
+ {
+ "epoch": 0.22358078602620088,
+ "grad_norm": 0.10609443485736847,
+ "learning_rate": 8.999999999999999e-05,
+ "loss": 10.970746040344238,
+ "step": 16
+ },
+ {
+ "epoch": 0.23755458515283842,
+ "grad_norm": 0.10648232698440552,
+ "learning_rate": 9.599999999999999e-05,
+ "loss": 10.91267204284668,
+ "step": 17
+ },
+ {
+ "epoch": 0.251528384279476,
+ "grad_norm": 0.10584589838981628,
+ "learning_rate": 0.000102,
+ "loss": 10.854547500610352,
+ "step": 18
+ },
+ {
+ "epoch": 0.26550218340611353,
+ "grad_norm": 0.10591720044612885,
+ "learning_rate": 0.00010799999999999998,
+ "loss": 10.788917541503906,
+ "step": 19
+ },
+ {
+ "epoch": 0.2794759825327511,
+ "grad_norm": 0.10611672699451447,
+ "learning_rate": 0.00011399999999999999,
+ "loss": 10.717556953430176,
+ "step": 20
+ },
+ {
+ "epoch": 0.2934497816593886,
+ "grad_norm": 0.10455359518527985,
+ "learning_rate": 0.00011999999999999999,
+ "loss": 10.650186538696289,
+ "step": 21
+ },
+ {
+ "epoch": 0.3074235807860262,
+ "grad_norm": 0.1047525629401207,
+ "learning_rate": 0.00012599999999999997,
+ "loss": 10.56808090209961,
+ "step": 22
+ },
+ {
+ "epoch": 0.32139737991266376,
+ "grad_norm": 0.10511605441570282,
+ "learning_rate": 0.00013199999999999998,
+ "loss": 10.484028816223145,
+ "step": 23
+ },
+ {
+ "epoch": 0.3353711790393013,
+ "grad_norm": 0.10467177629470825,
+ "learning_rate": 0.000138,
+ "loss": 10.399947166442871,
+ "step": 24
+ },
+ {
+ "epoch": 0.34934497816593885,
+ "grad_norm": 0.10432136058807373,
+ "learning_rate": 0.00014399999999999998,
+ "loss": 10.312714576721191,
+ "step": 25
+ },
+ {
+ "epoch": 0.36331877729257644,
+ "grad_norm": 0.10527262091636658,
+ "learning_rate": 0.00015,
+ "loss": 10.212261199951172,
+ "step": 26
+ },
+ {
+ "epoch": 0.377292576419214,
+ "grad_norm": 0.103610560297966,
+ "learning_rate": 0.000156,
+ "loss": 10.132034301757812,
+ "step": 27
+ },
+ {
+ "epoch": 0.39126637554585153,
+ "grad_norm": 0.10553059726953506,
+ "learning_rate": 0.000162,
+ "loss": 10.020434379577637,
+ "step": 28
+ },
+ {
+ "epoch": 0.4052401746724891,
+ "grad_norm": 0.1037299782037735,
+ "learning_rate": 0.000168,
+ "loss": 9.934466361999512,
+ "step": 29
+ },
+ {
+ "epoch": 0.4192139737991266,
+ "grad_norm": 0.1042739674448967,
+ "learning_rate": 0.00017399999999999997,
+ "loss": 9.822539329528809,
+ "step": 30
+ },
+ {
+ "epoch": 0.4331877729257642,
+ "grad_norm": 0.10300865024328232,
+ "learning_rate": 0.00017999999999999998,
+ "loss": 9.734994888305664,
+ "step": 31
+ },
+ {
+ "epoch": 0.44716157205240176,
+ "grad_norm": 0.10291654616594315,
+ "learning_rate": 0.000186,
+ "loss": 9.624505996704102,
+ "step": 32
+ },
+ {
+ "epoch": 0.4611353711790393,
+ "grad_norm": 0.10210851579904556,
+ "learning_rate": 0.00019199999999999998,
+ "loss": 9.518205642700195,
+ "step": 33
+ },
+ {
+ "epoch": 0.47510917030567684,
+ "grad_norm": 0.10222529619932175,
+ "learning_rate": 0.000198,
+ "loss": 9.412222862243652,
+ "step": 34
+ },
+ {
+ "epoch": 0.4890829694323144,
+ "grad_norm": 0.09992019087076187,
+ "learning_rate": 0.000204,
+ "loss": 9.326566696166992,
+ "step": 35
+ },
+ {
+ "epoch": 0.503056768558952,
+ "grad_norm": 0.09931084513664246,
+ "learning_rate": 0.00020999999999999998,
+ "loss": 9.210111618041992,
+ "step": 36
+ },
+ {
+ "epoch": 0.5170305676855895,
+ "grad_norm": 0.09928256273269653,
+ "learning_rate": 0.00021599999999999996,
+ "loss": 9.100493431091309,
+ "step": 37
+ },
+ {
+ "epoch": 0.5310043668122271,
+ "grad_norm": 0.09843816608190536,
+ "learning_rate": 0.00022199999999999998,
+ "loss": 8.996543884277344,
+ "step": 38
+ },
+ {
+ "epoch": 0.5449781659388646,
+ "grad_norm": 0.09825374186038971,
+ "learning_rate": 0.00022799999999999999,
+ "loss": 8.888991355895996,
+ "step": 39
+ },
+ {
+ "epoch": 0.5589519650655022,
+ "grad_norm": 0.09602385759353638,
+ "learning_rate": 0.000234,
+ "loss": 8.802907943725586,
+ "step": 40
+ },
+ {
+ "epoch": 0.5729257641921397,
+ "grad_norm": 0.09414064139127731,
+ "learning_rate": 0.00023999999999999998,
+ "loss": 8.715423583984375,
+ "step": 41
+ },
+ {
+ "epoch": 0.5868995633187772,
+ "grad_norm": 0.09273893386125565,
+ "learning_rate": 0.00024599999999999996,
+ "loss": 8.61374282836914,
+ "step": 42
+ },
+ {
+ "epoch": 0.6008733624454149,
+ "grad_norm": 0.09079599380493164,
+ "learning_rate": 0.00025199999999999995,
+ "loss": 8.502904891967773,
+ "step": 43
+ },
+ {
+ "epoch": 0.6148471615720524,
+ "grad_norm": 0.08605825901031494,
+ "learning_rate": 0.000258,
+ "loss": 8.453008651733398,
+ "step": 44
+ },
+ {
+ "epoch": 0.62882096069869,
+ "grad_norm": 0.08338996022939682,
+ "learning_rate": 0.00026399999999999997,
+ "loss": 8.365163803100586,
+ "step": 45
+ },
+ {
+ "epoch": 0.6427947598253275,
+ "grad_norm": 0.08150867372751236,
+ "learning_rate": 0.00027,
+ "loss": 8.285372734069824,
+ "step": 46
+ },
+ {
+ "epoch": 0.6567685589519651,
+ "grad_norm": 0.07794807851314545,
+ "learning_rate": 0.000276,
+ "loss": 8.185811996459961,
+ "step": 47
+ },
+ {
+ "epoch": 0.6707423580786026,
+ "grad_norm": 0.07592720538377762,
+ "learning_rate": 0.00028199999999999997,
+ "loss": 8.098482131958008,
+ "step": 48
+ },
+ {
+ "epoch": 0.6847161572052402,
+ "grad_norm": 0.0709897130727768,
+ "learning_rate": 0.00028799999999999995,
+ "loss": 8.042091369628906,
+ "step": 49
+ },
+ {
+ "epoch": 0.6986899563318777,
+ "grad_norm": 0.06582305580377579,
+ "learning_rate": 0.000294,
+ "loss": 7.972844123840332,
+ "step": 50
+ },
+ {
+ "epoch": 0.7126637554585152,
+ "grad_norm": 0.059452351182699203,
+ "learning_rate": 0.0003,
+ "loss": 7.924579620361328,
+ "step": 51
+ },
+ {
+ "epoch": 0.7266375545851529,
+ "grad_norm": 0.052426718175411224,
+ "learning_rate": 0.00030599999999999996,
+ "loss": 7.876336097717285,
+ "step": 52
+ },
+ {
+ "epoch": 0.7406113537117904,
+ "grad_norm": 0.04768150672316551,
+ "learning_rate": 0.000312,
+ "loss": 7.8102707862854,
+ "step": 53
+ },
+ {
+ "epoch": 0.754585152838428,
+ "grad_norm": 0.03970075398683548,
+ "learning_rate": 0.000318,
+ "loss": 7.791913986206055,
+ "step": 54
+ },
+ {
+ "epoch": 0.7685589519650655,
+ "grad_norm": 0.03559752553701401,
+ "learning_rate": 0.000324,
+ "loss": 7.756960868835449,
+ "step": 55
+ },
+ {
+ "epoch": 0.7825327510917031,
+ "grad_norm": 0.029321353882551193,
+ "learning_rate": 0.00033,
+ "loss": 7.724878787994385,
+ "step": 56
+ },
+ {
+ "epoch": 0.7965065502183406,
+ "grad_norm": 0.023485427722334862,
+ "learning_rate": 0.000336,
+ "loss": 7.72553014755249,
+ "step": 57
+ },
+ {
+ "epoch": 0.8104803493449781,
+ "grad_norm": 0.019319474697113037,
+ "learning_rate": 0.00034199999999999996,
+ "loss": 7.709493160247803,
+ "step": 58
+ },
+ {
+ "epoch": 0.8244541484716157,
+ "grad_norm": 0.016705691814422607,
+ "learning_rate": 0.00034799999999999995,
+ "loss": 7.6873297691345215,
+ "step": 59
+ },
+ {
+ "epoch": 0.8384279475982532,
+ "grad_norm": 0.017667734995484352,
+ "learning_rate": 0.00035399999999999993,
+ "loss": 7.677059173583984,
+ "step": 60
+ },
+ {
+ "epoch": 0.8524017467248908,
+ "grad_norm": 0.01671992614865303,
+ "learning_rate": 0.00035999999999999997,
+ "loss": 7.643512725830078,
+ "step": 61
+ },
+ {
+ "epoch": 0.8663755458515284,
+ "grad_norm": 0.012288172729313374,
+ "learning_rate": 0.00036599999999999995,
+ "loss": 7.658343315124512,
+ "step": 62
+ },
+ {
+ "epoch": 0.880349344978166,
+ "grad_norm": 0.012004776857793331,
+ "learning_rate": 0.000372,
+ "loss": 7.649759292602539,
+ "step": 63
+ },
+ {
+ "epoch": 0.8943231441048035,
+ "grad_norm": 0.012487445957958698,
+ "learning_rate": 0.00037799999999999997,
+ "loss": 7.631503105163574,
+ "step": 64
+ },
+ {
+ "epoch": 0.9082969432314411,
+ "grad_norm": 0.01489241048693657,
+ "learning_rate": 0.00038399999999999996,
+ "loss": 7.647110462188721,
+ "step": 65
+ },
+ {
+ "epoch": 0.9222707423580786,
+ "grad_norm": 0.01428980939090252,
+ "learning_rate": 0.00039,
+ "loss": 7.636983871459961,
+ "step": 66
+ },
+ {
+ "epoch": 0.9362445414847161,
+ "grad_norm": 0.010711144655942917,
+ "learning_rate": 0.000396,
+ "loss": 7.621130466461182,
+ "step": 67
+ },
+ {
+ "epoch": 0.9502183406113537,
+ "grad_norm": 0.010178886353969574,
+ "learning_rate": 0.000402,
+ "loss": 7.593924522399902,
+ "step": 68
+ },
+ {
+ "epoch": 0.9641921397379912,
+ "grad_norm": 0.010604874230921268,
+ "learning_rate": 0.000408,
+ "loss": 7.6015496253967285,
+ "step": 69
+ },
+ {
+ "epoch": 0.9781659388646288,
+ "grad_norm": 0.009751426056027412,
+ "learning_rate": 0.0004139999999999999,
+ "loss": 7.571964263916016,
+ "step": 70
+ },
+ {
+ "epoch": 0.9921397379912664,
+ "grad_norm": 0.009816481731832027,
+ "learning_rate": 0.00041999999999999996,
+ "loss": 7.57344913482666,
+ "step": 71
+ },
+ {
+ "epoch": 1.0,
+ "grad_norm": 0.008665637113153934,
+ "learning_rate": 0.00042599999999999995,
+ "loss": 7.579050064086914,
+ "step": 72
+ },
+ {
+ "epoch": 1.0,
+ "eval_loss": 7.581204891204834,
+ "eval_runtime": 59.0091,
+ "eval_samples_per_second": 41.383,
+ "eval_steps_per_second": 1.305,
+ "step": 72
+ },
+ {
+ "epoch": 1.0139737991266375,
+ "grad_norm": 0.009684201329946518,
+ "learning_rate": 0.00043199999999999993,
+ "loss": 7.568594932556152,
+ "step": 73
+ },
+ {
+ "epoch": 1.027947598253275,
+ "grad_norm": 0.025783710181713104,
+ "learning_rate": 0.00043799999999999997,
+ "loss": 7.578948497772217,
+ "step": 74
+ },
+ {
+ "epoch": 1.0419213973799126,
+ "grad_norm": 0.0327819399535656,
+ "learning_rate": 0.00044399999999999995,
+ "loss": 7.548173904418945,
+ "step": 75
+ },
+ {
+ "epoch": 1.0558951965065502,
+ "grad_norm": 0.021604498848319054,
+ "learning_rate": 0.00045,
+ "loss": 7.552953720092773,
+ "step": 76
+ },
+ {
+ "epoch": 1.0698689956331877,
+ "grad_norm": 0.010230105370283127,
+ "learning_rate": 0.00045599999999999997,
+ "loss": 7.5213212966918945,
+ "step": 77
+ },
+ {
+ "epoch": 1.0838427947598253,
+ "grad_norm": 0.023315589874982834,
+ "learning_rate": 0.00046199999999999995,
+ "loss": 7.501640319824219,
+ "step": 78
+ },
+ {
+ "epoch": 1.0978165938864628,
+ "grad_norm": 0.01109944935888052,
+ "learning_rate": 0.000468,
+ "loss": 7.539519786834717,
+ "step": 79
+ },
+ {
+ "epoch": 1.1117903930131003,
+ "grad_norm": 0.026093758642673492,
+ "learning_rate": 0.000474,
+ "loss": 7.525928974151611,
+ "step": 80
+ },
+ {
+ "epoch": 1.125764192139738,
+ "grad_norm": 0.03274895250797272,
+ "learning_rate": 0.00047999999999999996,
+ "loss": 7.535602569580078,
+ "step": 81
+ },
+ {
+ "epoch": 1.1397379912663754,
+ "grad_norm": 0.01543602254241705,
+ "learning_rate": 0.000486,
+ "loss": 7.513683795928955,
+ "step": 82
+ },
+ {
+ "epoch": 1.1537117903930132,
+ "grad_norm": 0.028377249836921692,
+ "learning_rate": 0.0004919999999999999,
+ "loss": 7.49384880065918,
+ "step": 83
+ },
+ {
+ "epoch": 1.1676855895196507,
+ "grad_norm": 0.04121481627225876,
+ "learning_rate": 0.000498,
+ "loss": 7.488447189331055,
+ "step": 84
+ },
+ {
+ "epoch": 1.1816593886462883,
+ "grad_norm": 0.013282302767038345,
+ "learning_rate": 0.0005039999999999999,
+ "loss": 7.463010787963867,
+ "step": 85
+ },
+ {
+ "epoch": 1.1956331877729258,
+ "grad_norm": 0.03200364485383034,
+ "learning_rate": 0.0005099999999999999,
+ "loss": 7.473515033721924,
+ "step": 86
+ },
+ {
+ "epoch": 1.2096069868995634,
+ "grad_norm": 0.020768029615283012,
+ "learning_rate": 0.000516,
+ "loss": 7.464688777923584,
+ "step": 87
+ },
+ {
+ "epoch": 1.223580786026201,
+ "grad_norm": 0.021715788170695305,
+ "learning_rate": 0.000522,
+ "loss": 7.4835028648376465,
+ "step": 88
+ },
+ {
+ "epoch": 1.2375545851528384,
+ "grad_norm": 0.026255754753947258,
+ "learning_rate": 0.0005279999999999999,
+ "loss": 7.443392753601074,
+ "step": 89
+ },
+ {
+ "epoch": 1.251528384279476,
+ "grad_norm": 0.022325951606035233,
+ "learning_rate": 0.000534,
+ "loss": 7.457373142242432,
+ "step": 90
+ },
+ {
+ "epoch": 1.2655021834061135,
+ "grad_norm": 0.009270387701690197,
+ "learning_rate": 0.00054,
+ "loss": 7.455986976623535,
+ "step": 91
+ },
+ {
+ "epoch": 1.279475982532751,
+ "grad_norm": 0.02369544468820095,
+ "learning_rate": 0.0005459999999999999,
+ "loss": 7.442216873168945,
+ "step": 92
+ },
+ {
+ "epoch": 1.2934497816593886,
+ "grad_norm": 0.008474506437778473,
+ "learning_rate": 0.000552,
+ "loss": 7.4294915199279785,
+ "step": 93
+ },
+ {
+ "epoch": 1.3074235807860262,
+ "grad_norm": 0.017043249681591988,
+ "learning_rate": 0.000558,
+ "loss": 7.406932830810547,
+ "step": 94
+ },
+ {
+ "epoch": 1.3213973799126637,
+ "grad_norm": 0.011813412420451641,
+ "learning_rate": 0.0005639999999999999,
+ "loss": 7.415900230407715,
+ "step": 95
+ },
+ {
+ "epoch": 1.3353711790393012,
+ "grad_norm": 0.01086883433163166,
+ "learning_rate": 0.00057,
+ "loss": 7.452998638153076,
+ "step": 96
+ },
+ {
+ "epoch": 1.3493449781659388,
+ "grad_norm": 0.014700951054692268,
+ "learning_rate": 0.0005759999999999999,
+ "loss": 7.410249710083008,
+ "step": 97
+ },
+ {
+ "epoch": 1.3633187772925766,
+ "grad_norm": 0.014512976631522179,
+ "learning_rate": 0.0005819999999999999,
+ "loss": 7.429132461547852,
+ "step": 98
+ },
+ {
+ "epoch": 1.3772925764192139,
+ "grad_norm": 0.012519920244812965,
+ "learning_rate": 0.000588,
+ "loss": 7.430817604064941,
+ "step": 99
+ },
+ {
+ "epoch": 1.3912663755458516,
+ "grad_norm": 0.007132664322853088,
+ "learning_rate": 0.0005939999999999999,
+ "loss": 7.405234336853027,
+ "step": 100
+ },
+ {
+ "epoch": 1.405240174672489,
+ "grad_norm": 0.019489388912916183,
+ "learning_rate": 0.0006,
+ "loss": 7.427791595458984,
+ "step": 101
+ },
+ {
+ "epoch": 1.4192139737991267,
+ "grad_norm": 0.027252664789557457,
+ "learning_rate": 0.0006,
+ "loss": 7.41761589050293,
+ "step": 102
+ },
+ {
+ "epoch": 1.4331877729257643,
+ "grad_norm": 0.02501637488603592,
+ "learning_rate": 0.0006,
+ "loss": 7.416735649108887,
+ "step": 103
+ },
+ {
+ "epoch": 1.4471615720524018,
+ "grad_norm": 0.00989772193133831,
+ "learning_rate": 0.0006,
+ "loss": 7.3756632804870605,
+ "step": 104
+ },
+ {
+ "epoch": 1.4611353711790394,
+ "grad_norm": 0.01622949168086052,
+ "learning_rate": 0.0006,
+ "loss": 7.400811672210693,
+ "step": 105
+ },
+ {
+ "epoch": 1.475109170305677,
+ "grad_norm": 0.012407763861119747,
+ "learning_rate": 0.0006,
+ "loss": 7.37675666809082,
+ "step": 106
+ },
+ {
+ "epoch": 1.4890829694323144,
+ "grad_norm": 0.010353565216064453,
+ "learning_rate": 0.0006,
+ "loss": 7.4013142585754395,
+ "step": 107
+ },
+ {
+ "epoch": 1.503056768558952,
+ "grad_norm": 0.0120553532615304,
+ "learning_rate": 0.0006,
+ "loss": 7.379873275756836,
+ "step": 108
+ },
+ {
+ "epoch": 1.5170305676855895,
+ "grad_norm": 0.01610225811600685,
+ "learning_rate": 0.0006,
+ "loss": 7.371631145477295,
+ "step": 109
+ },
+ {
+ "epoch": 1.531004366812227,
+ "grad_norm": 0.01159879844635725,
+ "learning_rate": 0.0006,
+ "loss": 7.389007568359375,
+ "step": 110
+ },
+ {
+ "epoch": 1.5449781659388646,
+ "grad_norm": 0.008013816550374031,
+ "learning_rate": 0.0006,
+ "loss": 7.395920753479004,
+ "step": 111
+ },
+ {
+ "epoch": 1.5589519650655022,
+ "grad_norm": 0.016334809362888336,
+ "learning_rate": 0.0006,
+ "loss": 7.393121719360352,
+ "step": 112
+ },
+ {
+ "epoch": 1.5729257641921397,
+ "grad_norm": 0.018821721896529198,
+ "learning_rate": 0.0006,
+ "loss": 7.367543697357178,
+ "step": 113
+ },
+ {
+ "epoch": 1.5868995633187772,
+ "grad_norm": 0.014088994823396206,
+ "learning_rate": 0.0006,
+ "loss": 7.385537624359131,
+ "step": 114
+ },
+ {
+ "epoch": 1.600873362445415,
+ "grad_norm": 0.005864050704985857,
+ "learning_rate": 0.0006,
+ "loss": 7.365346908569336,
+ "step": 115
+ },
+ {
+ "epoch": 1.6148471615720523,
+ "grad_norm": 0.011745430529117584,
+ "learning_rate": 0.0006,
+ "loss": 7.390227317810059,
+ "step": 116
+ },
+ {
+ "epoch": 1.62882096069869,
+ "grad_norm": 0.017205573618412018,
+ "learning_rate": 0.0006,
+ "loss": 7.3558549880981445,
+ "step": 117
+ },
+ {
+ "epoch": 1.6427947598253274,
+ "grad_norm": 0.021256346255540848,
+ "learning_rate": 0.0006,
+ "loss": 7.377346992492676,
+ "step": 118
+ },
+ {
+ "epoch": 1.6567685589519652,
+ "grad_norm": 0.019335204735398293,
+ "learning_rate": 0.0006,
+ "loss": 7.346194267272949,
+ "step": 119
+ },
+ {
+ "epoch": 1.6707423580786025,
+ "grad_norm": 0.008134003728628159,
+ "learning_rate": 0.0006,
+ "loss": 7.3063578605651855,
+ "step": 120
+ },
+ {
+ "epoch": 1.6847161572052403,
+ "grad_norm": 0.017613831907510757,
+ "learning_rate": 0.0006,
+ "loss": 7.338171005249023,
+ "step": 121
+ },
+ {
+ "epoch": 1.6986899563318776,
+ "grad_norm": 0.00890734139829874,
+ "learning_rate": 0.0006,
+ "loss": 7.355384826660156,
+ "step": 122
+ },
+ {
+ "epoch": 1.7126637554585153,
+ "grad_norm": 0.013448155485093594,
+ "learning_rate": 0.0006,
+ "loss": 7.344296455383301,
+ "step": 123
+ },
+ {
+ "epoch": 1.726637554585153,
+ "grad_norm": 0.009241633117198944,
+ "learning_rate": 0.0006,
+ "loss": 7.3454060554504395,
+ "step": 124
+ },
+ {
+ "epoch": 1.7406113537117904,
+ "grad_norm": 0.018561311066150665,
+ "learning_rate": 0.0006,
+ "loss": 7.32348108291626,
+ "step": 125
+ },
+ {
+ "epoch": 1.754585152838428,
+ "grad_norm": 0.025667704641819,
+ "learning_rate": 0.0006,
+ "loss": 7.356142520904541,
+ "step": 126
+ },
+ {
+ "epoch": 1.7685589519650655,
+ "grad_norm": 0.036232151091098785,
+ "learning_rate": 0.0006,
+ "loss": 7.363539695739746,
+ "step": 127
+ },
+ {
+ "epoch": 1.782532751091703,
+ "grad_norm": 0.059483714401721954,
+ "learning_rate": 0.0006,
+ "loss": 7.402331352233887,
+ "step": 128
+ },
+ {
+ "epoch": 1.7965065502183406,
+ "grad_norm": 0.05365009233355522,
+ "learning_rate": 0.0006,
+ "loss": 7.40457820892334,
+ "step": 129
+ },
+ {
+ "epoch": 1.8104803493449781,
+ "grad_norm": 0.02738415263593197,
+ "learning_rate": 0.0006,
+ "loss": 7.3756208419799805,
+ "step": 130
+ },
+ {
+ "epoch": 1.8244541484716157,
+ "grad_norm": 0.035017091780900955,
+ "learning_rate": 0.0006,
+ "loss": 7.379263877868652,
+ "step": 131
+ },
+ {
+ "epoch": 1.8384279475982532,
+ "grad_norm": 0.02819863148033619,
+ "learning_rate": 0.0006,
+ "loss": 7.327673435211182,
+ "step": 132
+ },
+ {
+ "epoch": 1.8524017467248908,
+ "grad_norm": 0.017754238098859787,
+ "learning_rate": 0.0006,
+ "loss": 7.333420276641846,
+ "step": 133
+ },
+ {
+ "epoch": 1.8663755458515285,
+ "grad_norm": 0.024685733020305634,
+ "learning_rate": 0.0006,
+ "loss": 7.34493350982666,
+ "step": 134
+ },
+ {
+ "epoch": 1.8803493449781659,
+ "grad_norm": 0.019426532089710236,
+ "learning_rate": 0.0006,
+ "loss": 7.328448295593262,
+ "step": 135
+ },
+ {
+ "epoch": 1.8943231441048036,
+ "grad_norm": 0.02106391079723835,
+ "learning_rate": 0.0006,
+ "loss": 7.327917098999023,
+ "step": 136
+ },
+ {
+ "epoch": 1.908296943231441,
+ "grad_norm": 0.02056879736483097,
+ "learning_rate": 0.0006,
+ "loss": 7.3311991691589355,
+ "step": 137
+ },
+ {
+ "epoch": 1.9222707423580787,
+ "grad_norm": 0.016865141689777374,
+ "learning_rate": 0.0006,
+ "loss": 7.312766075134277,
+ "step": 138
+ },
+ {
+ "epoch": 1.936244541484716,
+ "grad_norm": 0.017549855634570122,
+ "learning_rate": 0.0006,
+ "loss": 7.31191349029541,
+ "step": 139
+ },
+ {
+ "epoch": 1.9502183406113538,
+ "grad_norm": 0.015930157154798508,
+ "learning_rate": 0.0006,
+ "loss": 7.330901622772217,
+ "step": 140
+ },
+ {
+ "epoch": 1.9641921397379911,
+ "grad_norm": 0.013283558189868927,
+ "learning_rate": 0.0006,
+ "loss": 7.298318386077881,
+ "step": 141
+ },
+ {
+ "epoch": 1.9781659388646289,
+ "grad_norm": 0.012023469433188438,
+ "learning_rate": 0.0006,
+ "loss": 7.282337188720703,
+ "step": 142
+ },
+ {
+ "epoch": 1.9921397379912664,
+ "grad_norm": 0.01459511648863554,
+ "learning_rate": 0.0006,
+ "loss": 7.282924175262451,
+ "step": 143
+ },
+ {
+ "epoch": 2.0,
+ "grad_norm": 0.017474235966801643,
+ "learning_rate": 0.0006,
+ "loss": 7.254333972930908,
+ "step": 144
+ },
+ {
+ "epoch": 2.0,
+ "eval_loss": 7.318708896636963,
+ "eval_runtime": 59.6555,
+ "eval_samples_per_second": 40.935,
+ "eval_steps_per_second": 1.291,
+ "step": 144
+ },
+ {
+ "epoch": 2.0139737991266378,
+ "grad_norm": 0.019757507368922234,
+ "learning_rate": 0.0006,
+ "loss": 7.30288553237915,
+ "step": 145
+ },
+ {
+ "epoch": 2.027947598253275,
+ "grad_norm": 0.014571248553693295,
+ "learning_rate": 0.0006,
+ "loss": 7.274681568145752,
+ "step": 146
+ },
+ {
+ "epoch": 2.041921397379913,
+ "grad_norm": 0.015744198113679886,
+ "learning_rate": 0.0006,
+ "loss": 7.278632640838623,
+ "step": 147
+ },
+ {
+ "epoch": 2.05589519650655,
+ "grad_norm": 0.028048858046531677,
+ "learning_rate": 0.0006,
+ "loss": 7.26558780670166,
+ "step": 148
+ },
+ {
+ "epoch": 2.069868995633188,
+ "grad_norm": 0.06547307223081589,
+ "learning_rate": 0.0006,
+ "loss": 7.32567834854126,
+ "step": 149
+ },
+ {
+ "epoch": 2.0838427947598253,
+ "grad_norm": 0.05969763547182083,
+ "learning_rate": 0.0006,
+ "loss": 7.395836353302002,
+ "step": 150
+ },
+ {
+ "epoch": 2.097816593886463,
+ "grad_norm": 0.032574281096458435,
+ "learning_rate": 0.0006,
+ "loss": 7.318150997161865,
+ "step": 151
+ },
+ {
+ "epoch": 2.1117903930131003,
+ "grad_norm": 0.029452962800860405,
+ "learning_rate": 0.0006,
+ "loss": 7.308066368103027,
+ "step": 152
+ },
+ {
+ "epoch": 2.125764192139738,
+ "grad_norm": 0.03161991387605667,
+ "learning_rate": 0.0006,
+ "loss": 7.306290626525879,
+ "step": 153
+ },
+ {
+ "epoch": 2.1397379912663754,
+ "grad_norm": 0.027096295729279518,
+ "learning_rate": 0.0006,
+ "loss": 7.297796249389648,
+ "step": 154
+ },
+ {
+ "epoch": 2.153711790393013,
+ "grad_norm": 0.028657056391239166,
+ "learning_rate": 0.0006,
+ "loss": 7.317424774169922,
+ "step": 155
+ },
+ {
+ "epoch": 2.1676855895196505,
+ "grad_norm": 0.02046995982527733,
+ "learning_rate": 0.0006,
+ "loss": 7.297986030578613,
+ "step": 156
+ },
+ {
+ "epoch": 2.1816593886462883,
+ "grad_norm": 0.02220369316637516,
+ "learning_rate": 0.0006,
+ "loss": 7.268237590789795,
+ "step": 157
+ },
+ {
+ "epoch": 2.1956331877729256,
+ "grad_norm": 0.018381357192993164,
+ "learning_rate": 0.0006,
+ "loss": 7.259415626525879,
+ "step": 158
+ },
+ {
+ "epoch": 2.2096069868995634,
+ "grad_norm": 0.0204413291066885,
+ "learning_rate": 0.0006,
+ "loss": 7.258173942565918,
+ "step": 159
+ },
+ {
+ "epoch": 2.2235807860262007,
+ "grad_norm": 0.015874288976192474,
+ "learning_rate": 0.0006,
+ "loss": 7.281834602355957,
+ "step": 160
+ },
+ {
+ "epoch": 2.2375545851528384,
+ "grad_norm": 0.01634068600833416,
+ "learning_rate": 0.0006,
+ "loss": 7.2469868659973145,
+ "step": 161
+ },
+ {
+ "epoch": 2.251528384279476,
+ "grad_norm": 0.01728747971355915,
+ "learning_rate": 0.0006,
+ "loss": 7.280995845794678,
+ "step": 162
+ },
+ {
+ "epoch": 2.2655021834061135,
+ "grad_norm": 0.016341188922524452,
+ "learning_rate": 0.0006,
+ "loss": 7.273357391357422,
+ "step": 163
+ },
+ {
+ "epoch": 2.279475982532751,
+ "grad_norm": 0.016125807538628578,
+ "learning_rate": 0.0006,
+ "loss": 7.273510932922363,
+ "step": 164
+ },
+ {
+ "epoch": 2.2934497816593886,
+ "grad_norm": 0.01629287749528885,
+ "learning_rate": 0.0006,
+ "loss": 7.267029762268066,
+ "step": 165
+ },
+ {
+ "epoch": 2.3074235807860264,
+ "grad_norm": 0.015296131372451782,
+ "learning_rate": 0.0006,
+ "loss": 7.211989402770996,
+ "step": 166
+ },
+ {
+ "epoch": 2.3213973799126637,
+ "grad_norm": 0.01709570363163948,
+ "learning_rate": 0.0006,
+ "loss": 7.242851734161377,
+ "step": 167
+ },
+ {
+ "epoch": 2.3353711790393015,
+ "grad_norm": 0.012552388943731785,
+ "learning_rate": 0.0006,
+ "loss": 7.241696834564209,
+ "step": 168
+ },
+ {
+ "epoch": 2.349344978165939,
+ "grad_norm": 0.014571291394531727,
+ "learning_rate": 0.0006,
+ "loss": 7.270681381225586,
+ "step": 169
+ },
+ {
+ "epoch": 2.3633187772925766,
+ "grad_norm": 0.013389437459409237,
+ "learning_rate": 0.0006,
+ "loss": 7.191600799560547,
+ "step": 170
+ },
+ {
+ "epoch": 2.377292576419214,
+ "grad_norm": 0.011895314790308475,
+ "learning_rate": 0.0006,
+ "loss": 7.213610649108887,
+ "step": 171
+ },
+ {
+ "epoch": 2.3912663755458516,
+ "grad_norm": 0.01345642190426588,
+ "learning_rate": 0.0006,
+ "loss": 7.193787574768066,
+ "step": 172
+ },
+ {
+ "epoch": 2.405240174672489,
+ "grad_norm": 0.0091794328764081,
+ "learning_rate": 0.0006,
+ "loss": 7.233364105224609,
+ "step": 173
+ },
+ {
+ "epoch": 2.4192139737991267,
+ "grad_norm": 0.013767086900770664,
+ "learning_rate": 0.0006,
+ "loss": 7.217272758483887,
+ "step": 174
+ },
+ {
+ "epoch": 2.433187772925764,
+ "grad_norm": 0.015134960412979126,
+ "learning_rate": 0.0006,
+ "loss": 7.192791938781738,
+ "step": 175
+ },
+ {
+ "epoch": 2.447161572052402,
+ "grad_norm": 0.01790648326277733,
+ "learning_rate": 0.0006,
+ "loss": 7.229467391967773,
+ "step": 176
+ },
+ {
+ "epoch": 2.461135371179039,
+ "grad_norm": 0.015362377278506756,
+ "learning_rate": 0.0006,
+ "loss": 7.180346488952637,
+ "step": 177
+ },
+ {
+ "epoch": 2.475109170305677,
+ "grad_norm": 0.010505498386919498,
+ "learning_rate": 0.0006,
+ "loss": 7.188833236694336,
+ "step": 178
+ },
+ {
+ "epoch": 2.489082969432314,
+ "grad_norm": 0.009129747748374939,
+ "learning_rate": 0.0006,
+ "loss": 7.203153133392334,
+ "step": 179
+ },
+ {
+ "epoch": 2.503056768558952,
+ "grad_norm": 0.01147883478552103,
+ "learning_rate": 0.0006,
+ "loss": 7.163824558258057,
+ "step": 180
+ },
+ {
+ "epoch": 2.5170305676855893,
+ "grad_norm": 0.011338942684233189,
+ "learning_rate": 0.0006,
+ "loss": 7.16287899017334,
+ "step": 181
+ },
+ {
+ "epoch": 2.531004366812227,
+ "grad_norm": 0.013277675025165081,
+ "learning_rate": 0.0006,
+ "loss": 7.155905723571777,
+ "step": 182
+ },
+ {
+ "epoch": 2.544978165938865,
+ "grad_norm": 0.015269347466528416,
+ "learning_rate": 0.0006,
+ "loss": 7.189371109008789,
+ "step": 183
+ },
+ {
+ "epoch": 2.558951965065502,
+ "grad_norm": 0.017905596643686295,
+ "learning_rate": 0.0006,
+ "loss": 7.177547931671143,
+ "step": 184
+ },
+ {
+ "epoch": 2.5729257641921395,
+ "grad_norm": 0.024431610479950905,
+ "learning_rate": 0.0006,
+ "loss": 7.192169189453125,
+ "step": 185
+ },
+ {
+ "epoch": 2.5868995633187772,
+ "grad_norm": 0.03656961768865585,
+ "learning_rate": 0.0006,
+ "loss": 7.124387741088867,
+ "step": 186
+ },
+ {
+ "epoch": 2.600873362445415,
+ "grad_norm": 0.04337048903107643,
+ "learning_rate": 0.0006,
+ "loss": 7.140231609344482,
+ "step": 187
+ },
+ {
+ "epoch": 2.6148471615720523,
+ "grad_norm": 0.04047228768467903,
+ "learning_rate": 0.0006,
+ "loss": 7.171927452087402,
+ "step": 188
+ },
+ {
+ "epoch": 2.62882096069869,
+ "grad_norm": 0.07825770974159241,
+ "learning_rate": 0.0006,
+ "loss": 7.207631587982178,
+ "step": 189
+ },
+ {
+ "epoch": 2.6427947598253274,
+ "grad_norm": 0.07118063420057297,
+ "learning_rate": 0.0006,
+ "loss": 7.209488391876221,
+ "step": 190
+ },
+ {
+ "epoch": 2.656768558951965,
+ "grad_norm": 0.05153293535113335,
+ "learning_rate": 0.0006,
+ "loss": 7.14793062210083,
+ "step": 191
+ },
+ {
+ "epoch": 2.6707423580786025,
+ "grad_norm": 0.03318839520215988,
+ "learning_rate": 0.0006,
+ "loss": 7.163010597229004,
+ "step": 192
+ },
+ {
+ "epoch": 2.6847161572052403,
+ "grad_norm": 0.029333539307117462,
+ "learning_rate": 0.0006,
+ "loss": 7.159955978393555,
+ "step": 193
+ },
+ {
+ "epoch": 2.6986899563318776,
+ "grad_norm": 0.029405880719423294,
+ "learning_rate": 0.0006,
+ "loss": 7.136178493499756,
+ "step": 194
+ },
+ {
+ "epoch": 2.7126637554585153,
+ "grad_norm": 0.037815988063812256,
+ "learning_rate": 0.0006,
+ "loss": 7.183186054229736,
+ "step": 195
+ },
+ {
+ "epoch": 2.726637554585153,
+ "grad_norm": 0.020307250320911407,
+ "learning_rate": 0.0006,
+ "loss": 7.133164882659912,
+ "step": 196
+ },
+ {
+ "epoch": 2.7406113537117904,
+ "grad_norm": 0.026048069819808006,
+ "learning_rate": 0.0006,
+ "loss": 7.124977111816406,
+ "step": 197
+ },
+ {
+ "epoch": 2.7545851528384278,
+ "grad_norm": 0.027316724881529808,
+ "learning_rate": 0.0006,
+ "loss": 7.07265567779541,
+ "step": 198
+ },
+ {
+ "epoch": 2.7685589519650655,
+ "grad_norm": 0.021629009395837784,
+ "learning_rate": 0.0006,
+ "loss": 7.150020599365234,
+ "step": 199
+ },
+ {
+ "epoch": 2.7825327510917033,
+ "grad_norm": 0.01605929434299469,
+ "learning_rate": 0.0006,
+ "loss": 7.10919189453125,
+ "step": 200
+ },
+ {
+ "epoch": 2.7965065502183406,
+ "grad_norm": 0.020800089463591576,
+ "learning_rate": 0.0006,
+ "loss": 7.132045269012451,
+ "step": 201
+ },
+ {
+ "epoch": 2.810480349344978,
+ "grad_norm": 0.019225774332880974,
+ "learning_rate": 0.0006,
+ "loss": 7.128364562988281,
+ "step": 202
+ },
+ {
+ "epoch": 2.8244541484716157,
+ "grad_norm": 0.01561372634023428,
+ "learning_rate": 0.0006,
+ "loss": 7.083669662475586,
+ "step": 203
+ },
+ {
+ "epoch": 2.8384279475982535,
+ "grad_norm": 0.021969519555568695,
+ "learning_rate": 0.0006,
+ "loss": 7.050841331481934,
+ "step": 204
+ },
+ {
+ "epoch": 2.8524017467248908,
+ "grad_norm": 0.021453695371747017,
+ "learning_rate": 0.0006,
+ "loss": 7.097085475921631,
+ "step": 205
+ },
+ {
+ "epoch": 2.8663755458515285,
+ "grad_norm": 0.022836022078990936,
+ "learning_rate": 0.0006,
+ "loss": 7.10405969619751,
+ "step": 206
+ },
+ {
+ "epoch": 2.880349344978166,
+ "grad_norm": 0.02430851384997368,
+ "learning_rate": 0.0006,
+ "loss": 7.045181751251221,
+ "step": 207
+ },
+ {
+ "epoch": 2.8943231441048036,
+ "grad_norm": 0.020166993141174316,
+ "learning_rate": 0.0006,
+ "loss": 7.068183898925781,
+ "step": 208
+ },
+ {
+ "epoch": 2.908296943231441,
+ "grad_norm": 0.0118505684658885,
+ "learning_rate": 0.0006,
+ "loss": 7.07132625579834,
+ "step": 209
+ },
+ {
+ "epoch": 2.9222707423580787,
+ "grad_norm": 0.023772427812218666,
+ "learning_rate": 0.0006,
+ "loss": 7.055238723754883,
+ "step": 210
+ },
+ {
+ "epoch": 2.936244541484716,
+ "grad_norm": 0.02350712940096855,
+ "learning_rate": 0.0006,
+ "loss": 7.055898666381836,
+ "step": 211
+ },
+ {
+ "epoch": 2.950218340611354,
+ "grad_norm": 0.017244907096028328,
+ "learning_rate": 0.0006,
+ "loss": 7.0099711418151855,
+ "step": 212
+ },
+ {
+ "epoch": 2.964192139737991,
+ "grad_norm": 0.021565575152635574,
+ "learning_rate": 0.0006,
+ "loss": 6.991751670837402,
+ "step": 213
+ },
+ {
+ "epoch": 2.978165938864629,
+ "grad_norm": 0.03241860866546631,
+ "learning_rate": 0.0006,
+ "loss": 7.005980491638184,
+ "step": 214
+ },
+ {
+ "epoch": 2.992139737991266,
+ "grad_norm": 0.04613726586103439,
+ "learning_rate": 0.0006,
+ "loss": 7.027859687805176,
+ "step": 215
+ },
+ {
+ "epoch": 3.0,
+ "grad_norm": 0.06404894590377808,
+ "learning_rate": 0.0006,
+ "loss": 7.123725891113281,
+ "step": 216
+ },
+ {
+ "epoch": 3.0,
+ "eval_loss": 7.070647716522217,
+ "eval_runtime": 58.5801,
+ "eval_samples_per_second": 41.687,
+ "eval_steps_per_second": 1.314,
+ "step": 216
+ },
+ {
+ "epoch": 3.0139737991266378,
+ "grad_norm": 0.05542432889342308,
+ "learning_rate": 0.0006,
+ "loss": 7.078299522399902,
+ "step": 217
+ },
+ {
+ "epoch": 3.027947598253275,
+ "grad_norm": 0.026484526693820953,
+ "learning_rate": 0.0006,
+ "loss": 6.961159706115723,
+ "step": 218
+ },
+ {
+ "epoch": 3.041921397379913,
+ "grad_norm": 0.028276391327381134,
+ "learning_rate": 0.0006,
+ "loss": 7.019961833953857,
+ "step": 219
+ },
+ {
+ "epoch": 3.05589519650655,
+ "grad_norm": 0.024486595764756203,
+ "learning_rate": 0.0006,
+ "loss": 7.039677619934082,
+ "step": 220
+ },
+ {
+ "epoch": 3.069868995633188,
+ "grad_norm": 0.017620893195271492,
+ "learning_rate": 0.0006,
+ "loss": 6.9270782470703125,
+ "step": 221
+ },
+ {
+ "epoch": 3.0838427947598253,
+ "grad_norm": 0.02547302283346653,
+ "learning_rate": 0.0006,
+ "loss": 6.924233436584473,
+ "step": 222
+ },
+ {
+ "epoch": 3.097816593886463,
+ "grad_norm": 0.02539009042084217,
+ "learning_rate": 0.0006,
+ "loss": 6.973979949951172,
+ "step": 223
+ },
+ {
+ "epoch": 3.1117903930131003,
+ "grad_norm": 0.03256227448582649,
+ "learning_rate": 0.0006,
+ "loss": 6.933725357055664,
+ "step": 224
+ },
+ {
+ "epoch": 3.125764192139738,
+ "grad_norm": 0.04651800915598869,
+ "learning_rate": 0.0006,
+ "loss": 6.951044082641602,
+ "step": 225
+ },
+ {
+ "epoch": 3.1397379912663754,
+ "grad_norm": 0.04822975769639015,
+ "learning_rate": 0.0006,
+ "loss": 6.9614410400390625,
+ "step": 226
+ },
+ {
+ "epoch": 3.153711790393013,
+ "grad_norm": 0.031165020540356636,
+ "learning_rate": 0.0006,
+ "loss": 6.91193151473999,
+ "step": 227
+ },
+ {
+ "epoch": 3.1676855895196505,
+ "grad_norm": 0.029754292219877243,
+ "learning_rate": 0.0006,
+ "loss": 6.91802978515625,
+ "step": 228
+ },
+ {
+ "epoch": 3.1816593886462883,
+ "grad_norm": 0.02193152718245983,
+ "learning_rate": 0.0006,
+ "loss": 6.928090572357178,
+ "step": 229
+ },
+ {
+ "epoch": 3.1956331877729256,
+ "grad_norm": 0.02428470179438591,
+ "learning_rate": 0.0006,
+ "loss": 6.90396785736084,
+ "step": 230
+ },
+ {
+ "epoch": 3.2096069868995634,
+ "grad_norm": 0.023374345153570175,
+ "learning_rate": 0.0006,
+ "loss": 6.885858535766602,
+ "step": 231
+ },
+ {
+ "epoch": 3.2235807860262007,
+ "grad_norm": 0.023405000567436218,
+ "learning_rate": 0.0006,
+ "loss": 6.8855767250061035,
+ "step": 232
+ },
+ {
+ "epoch": 3.2375545851528384,
+ "grad_norm": 0.017277134582400322,
+ "learning_rate": 0.0006,
+ "loss": 6.901449203491211,
+ "step": 233
+ },
+ {
+ "epoch": 3.251528384279476,
+ "grad_norm": 0.019496750086545944,
+ "learning_rate": 0.0006,
+ "loss": 6.885687828063965,
+ "step": 234
+ },
+ {
+ "epoch": 3.2655021834061135,
+ "grad_norm": 0.016893276944756508,
+ "learning_rate": 0.0006,
+ "loss": 6.872369766235352,
+ "step": 235
+ },
+ {
+ "epoch": 3.279475982532751,
+ "grad_norm": 0.02058715932071209,
+ "learning_rate": 0.0006,
+ "loss": 6.823280334472656,
+ "step": 236
+ },
+ {
+ "epoch": 3.2934497816593886,
+ "grad_norm": 0.015530755743384361,
+ "learning_rate": 0.0006,
+ "loss": 6.860499858856201,
+ "step": 237
+ },
+ {
+ "epoch": 3.3074235807860264,
+ "grad_norm": 0.01934926211833954,
+ "learning_rate": 0.0006,
+ "loss": 6.817448616027832,
+ "step": 238
+ },
+ {
+ "epoch": 3.3213973799126637,
+ "grad_norm": 0.017413552850484848,
+ "learning_rate": 0.0006,
+ "loss": 6.881955146789551,
+ "step": 239
+ },
+ {
+ "epoch": 3.3353711790393015,
+ "grad_norm": 0.026478875428438187,
+ "learning_rate": 0.0006,
+ "loss": 6.827404499053955,
+ "step": 240
+ },
+ {
+ "epoch": 3.349344978165939,
+ "grad_norm": 0.032611701637506485,
+ "learning_rate": 0.0006,
+ "loss": 6.78285551071167,
+ "step": 241
+ },
+ {
+ "epoch": 3.3633187772925766,
+ "grad_norm": 0.041900306940078735,
+ "learning_rate": 0.0006,
+ "loss": 6.783053874969482,
+ "step": 242
+ },
+ {
+ "epoch": 3.377292576419214,
+ "grad_norm": 0.04149497672915459,
+ "learning_rate": 0.0006,
+ "loss": 6.745220184326172,
+ "step": 243
+ },
+ {
+ "epoch": 3.3912663755458516,
+ "grad_norm": 0.023949826136231422,
+ "learning_rate": 0.0006,
+ "loss": 6.841533184051514,
+ "step": 244
+ },
+ {
+ "epoch": 3.405240174672489,
+ "grad_norm": 0.028601842001080513,
+ "learning_rate": 0.0006,
+ "loss": 6.852852821350098,
+ "step": 245
+ },
+ {
+ "epoch": 3.4192139737991267,
+ "grad_norm": 0.027365155518054962,
+ "learning_rate": 0.0006,
+ "loss": 6.74376106262207,
+ "step": 246
+ },
+ {
+ "epoch": 3.433187772925764,
+ "grad_norm": 0.02856568619608879,
+ "learning_rate": 0.0006,
+ "loss": 6.74906063079834,
+ "step": 247
+ },
+ {
+ "epoch": 3.447161572052402,
+ "grad_norm": 0.019172416999936104,
+ "learning_rate": 0.0006,
+ "loss": 6.809206008911133,
+ "step": 248
+ },
+ {
+ "epoch": 3.461135371179039,
+ "grad_norm": 0.02161746844649315,
+ "learning_rate": 0.0006,
+ "loss": 6.789888381958008,
+ "step": 249
+ },
+ {
+ "epoch": 3.475109170305677,
+ "grad_norm": 0.020273666828870773,
+ "learning_rate": 0.0006,
+ "loss": 6.73512077331543,
+ "step": 250
+ },
+ {
+ "epoch": 3.489082969432314,
+ "grad_norm": 0.022329283878207207,
+ "learning_rate": 0.0006,
+ "loss": 6.826148509979248,
+ "step": 251
+ },
+ {
+ "epoch": 3.503056768558952,
+ "grad_norm": 0.02284284122288227,
+ "learning_rate": 0.0006,
+ "loss": 6.748521327972412,
+ "step": 252
+ },
+ {
+ "epoch": 3.5170305676855893,
+ "grad_norm": 0.026368247345089912,
+ "learning_rate": 0.0006,
+ "loss": 6.645879745483398,
+ "step": 253
+ },
+ {
+ "epoch": 3.531004366812227,
+ "grad_norm": 0.03627244383096695,
+ "learning_rate": 0.0006,
+ "loss": 6.666064739227295,
+ "step": 254
+ },
+ {
+ "epoch": 3.544978165938865,
+ "grad_norm": 0.04140935093164444,
+ "learning_rate": 0.0006,
+ "loss": 6.712275981903076,
+ "step": 255
+ },
+ {
+ "epoch": 3.558951965065502,
+ "grad_norm": 0.0390462800860405,
+ "learning_rate": 0.0006,
+ "loss": 6.665900230407715,
+ "step": 256
+ },
+ {
+ "epoch": 3.5729257641921395,
+ "grad_norm": 0.05178583785891533,
+ "learning_rate": 0.0006,
+ "loss": 6.753298759460449,
+ "step": 257
+ },
+ {
+ "epoch": 3.5868995633187772,
+ "grad_norm": 0.054328061640262604,
+ "learning_rate": 0.0006,
+ "loss": 6.732633113861084,
+ "step": 258
+ },
+ {
+ "epoch": 3.600873362445415,
+ "grad_norm": 0.04590460658073425,
+ "learning_rate": 0.0006,
+ "loss": 6.65250301361084,
+ "step": 259
+ },
+ {
+ "epoch": 3.6148471615720523,
+ "grad_norm": 0.044331666082143784,
+ "learning_rate": 0.0006,
+ "loss": 6.68654203414917,
+ "step": 260
+ },
+ {
+ "epoch": 3.62882096069869,
+ "grad_norm": 0.05308730527758598,
+ "learning_rate": 0.0006,
+ "loss": 6.642482757568359,
+ "step": 261
+ },
+ {
+ "epoch": 3.6427947598253274,
+ "grad_norm": 0.03999016433954239,
+ "learning_rate": 0.0006,
+ "loss": 6.625227451324463,
+ "step": 262
+ },
+ {
+ "epoch": 3.656768558951965,
+ "grad_norm": 0.04088086634874344,
+ "learning_rate": 0.0006,
+ "loss": 6.67623233795166,
+ "step": 263
+ },
+ {
+ "epoch": 3.6707423580786025,
+ "grad_norm": 0.03615636005997658,
+ "learning_rate": 0.0006,
+ "loss": 6.6869001388549805,
+ "step": 264
+ },
+ {
+ "epoch": 3.6847161572052403,
+ "grad_norm": 0.02817360684275627,
+ "learning_rate": 0.0006,
+ "loss": 6.674601078033447,
+ "step": 265
+ },
+ {
+ "epoch": 3.6986899563318776,
+ "grad_norm": 0.032586123794317245,
+ "learning_rate": 0.0006,
+ "loss": 6.612994194030762,
+ "step": 266
+ },
+ {
+ "epoch": 3.7126637554585153,
+ "grad_norm": 0.031117867678403854,
+ "learning_rate": 0.0006,
+ "loss": 6.6244354248046875,
+ "step": 267
+ },
+ {
+ "epoch": 3.726637554585153,
+ "grad_norm": 0.028330687433481216,
+ "learning_rate": 0.0006,
+ "loss": 6.669557094573975,
+ "step": 268
+ },
+ {
+ "epoch": 3.7406113537117904,
+ "grad_norm": 0.021489202976226807,
+ "learning_rate": 0.0006,
+ "loss": 6.632035732269287,
+ "step": 269
+ },
+ {
+ "epoch": 3.7545851528384278,
+ "grad_norm": 0.024269424378871918,
+ "learning_rate": 0.0006,
+ "loss": 6.660867691040039,
+ "step": 270
+ },
+ {
+ "epoch": 3.7685589519650655,
+ "grad_norm": 0.02237873338162899,
+ "learning_rate": 0.0006,
+ "loss": 6.602587699890137,
+ "step": 271
+ },
+ {
+ "epoch": 3.7825327510917033,
+ "grad_norm": 0.026503929868340492,
+ "learning_rate": 0.0006,
+ "loss": 6.599200248718262,
+ "step": 272
+ },
+ {
+ "epoch": 3.7965065502183406,
+ "grad_norm": 0.03687124326825142,
+ "learning_rate": 0.0006,
+ "loss": 6.567296504974365,
+ "step": 273
+ },
+ {
+ "epoch": 3.810480349344978,
+ "grad_norm": 0.04708952456712723,
+ "learning_rate": 0.0006,
+ "loss": 6.621265411376953,
+ "step": 274
+ },
+ {
+ "epoch": 3.8244541484716157,
+ "grad_norm": 0.041846614331007004,
+ "learning_rate": 0.0006,
+ "loss": 6.550536155700684,
+ "step": 275
+ },
+ {
+ "epoch": 3.8384279475982535,
+ "grad_norm": 0.031249675899744034,
+ "learning_rate": 0.0006,
+ "loss": 6.54278564453125,
+ "step": 276
+ },
+ {
+ "epoch": 3.8524017467248908,
+ "grad_norm": 0.023082256317138672,
+ "learning_rate": 0.0006,
+ "loss": 6.544787406921387,
+ "step": 277
+ },
+ {
+ "epoch": 3.8663755458515285,
+ "grad_norm": 0.020995570346713066,
+ "learning_rate": 0.0006,
+ "loss": 6.518143653869629,
+ "step": 278
+ },
+ {
+ "epoch": 3.880349344978166,
+ "grad_norm": 0.020839324221014977,
+ "learning_rate": 0.0006,
+ "loss": 6.546667098999023,
+ "step": 279
+ },
+ {
+ "epoch": 3.8943231441048036,
+ "grad_norm": 0.018377432599663734,
+ "learning_rate": 0.0006,
+ "loss": 6.478307723999023,
+ "step": 280
+ },
+ {
+ "epoch": 3.908296943231441,
+ "grad_norm": 0.021713724359869957,
+ "learning_rate": 0.0006,
+ "loss": 6.485574245452881,
+ "step": 281
+ },
+ {
+ "epoch": 3.9222707423580787,
+ "grad_norm": 0.01359301246702671,
+ "learning_rate": 0.0006,
+ "loss": 6.570268630981445,
+ "step": 282
+ },
+ {
+ "epoch": 3.936244541484716,
+ "grad_norm": 0.016233332455158234,
+ "learning_rate": 0.0006,
+ "loss": 6.57150936126709,
+ "step": 283
+ },
+ {
+ "epoch": 3.950218340611354,
+ "grad_norm": 0.020687608048319817,
+ "learning_rate": 0.0006,
+ "loss": 6.527956008911133,
+ "step": 284
+ },
+ {
+ "epoch": 3.964192139737991,
+ "grad_norm": 0.02032964862883091,
+ "learning_rate": 0.0006,
+ "loss": 6.468169689178467,
+ "step": 285
+ },
+ {
+ "epoch": 3.978165938864629,
+ "grad_norm": 0.016676288098096848,
+ "learning_rate": 0.0006,
+ "loss": 6.448418617248535,
+ "step": 286
+ },
+ {
+ "epoch": 3.992139737991266,
+ "grad_norm": 0.0168539360165596,
+ "learning_rate": 0.0006,
+ "loss": 6.461180686950684,
+ "step": 287
+ },
+ {
+ "epoch": 4.0,
+ "grad_norm": 0.02666233666241169,
+ "learning_rate": 0.0006,
+ "loss": 6.344893932342529,
+ "step": 288
+ },
+ {
+ "epoch": 4.0,
+ "eval_loss": 6.5384979248046875,
+ "eval_runtime": 58.814,
+ "eval_samples_per_second": 41.521,
+ "eval_steps_per_second": 1.309,
+ "step": 288
+ },
+ {
+ "epoch": 4.013973799126638,
+ "grad_norm": 0.04154708981513977,
+ "learning_rate": 0.0006,
+ "loss": 6.41855525970459,
+ "step": 289
+ },
+ {
+ "epoch": 4.0279475982532755,
+ "grad_norm": 0.04895668104290962,
+ "learning_rate": 0.0006,
+ "loss": 6.499302864074707,
+ "step": 290
+ },
+ {
+ "epoch": 4.041921397379912,
+ "grad_norm": 0.034464482218027115,
+ "learning_rate": 0.0006,
+ "loss": 6.438611030578613,
+ "step": 291
+ },
+ {
+ "epoch": 4.05589519650655,
+ "grad_norm": 0.06845773756504059,
+ "learning_rate": 0.0006,
+ "loss": 6.538361072540283,
+ "step": 292
+ },
+ {
+ "epoch": 4.069868995633188,
+ "grad_norm": 0.053768858313560486,
+ "learning_rate": 0.0006,
+ "loss": 6.527524948120117,
+ "step": 293
+ },
+ {
+ "epoch": 4.083842794759826,
+ "grad_norm": 0.039037931710481644,
+ "learning_rate": 0.0006,
+ "loss": 6.503847122192383,
+ "step": 294
+ },
+ {
+ "epoch": 4.097816593886463,
+ "grad_norm": 0.03962196782231331,
+ "learning_rate": 0.0006,
+ "loss": 6.475178241729736,
+ "step": 295
+ },
+ {
+ "epoch": 4.1117903930131,
+ "grad_norm": 0.049101535230875015,
+ "learning_rate": 0.0006,
+ "loss": 6.474587917327881,
+ "step": 296
+ },
+ {
+ "epoch": 4.125764192139738,
+ "grad_norm": 0.02830282226204872,
+ "learning_rate": 0.0006,
+ "loss": 6.45065975189209,
+ "step": 297
+ },
+ {
+ "epoch": 4.139737991266376,
+ "grad_norm": 0.023153482005000114,
+ "learning_rate": 0.0006,
+ "loss": 6.5218706130981445,
+ "step": 298
+ },
+ {
+ "epoch": 4.153711790393013,
+ "grad_norm": 0.025064002722501755,
+ "learning_rate": 0.0006,
+ "loss": 6.456475257873535,
+ "step": 299
+ },
+ {
+ "epoch": 4.1676855895196505,
+ "grad_norm": 0.02357092685997486,
+ "learning_rate": 0.0006,
+ "loss": 6.485952377319336,
+ "step": 300
+ },
+ {
+ "epoch": 4.181659388646288,
+ "grad_norm": 0.021832305938005447,
+ "learning_rate": 0.0006,
+ "loss": 6.425506591796875,
+ "step": 301
+ },
+ {
+ "epoch": 4.195633187772926,
+ "grad_norm": 0.0227900929749012,
+ "learning_rate": 0.0006,
+ "loss": 6.439155578613281,
+ "step": 302
+ },
+ {
+ "epoch": 4.209606986899563,
+ "grad_norm": 0.015961607918143272,
+ "learning_rate": 0.0006,
+ "loss": 6.460862636566162,
+ "step": 303
+ },
+ {
+ "epoch": 4.223580786026201,
+ "grad_norm": 0.017876390367746353,
+ "learning_rate": 0.0006,
+ "loss": 6.44564151763916,
+ "step": 304
+ },
+ {
+ "epoch": 4.2375545851528384,
+ "grad_norm": 0.013703204691410065,
+ "learning_rate": 0.0006,
+ "loss": 6.418414115905762,
+ "step": 305
+ },
+ {
+ "epoch": 4.251528384279476,
+ "grad_norm": 0.015240306034684181,
+ "learning_rate": 0.0006,
+ "loss": 6.411238670349121,
+ "step": 306
+ },
+ {
+ "epoch": 4.265502183406113,
+ "grad_norm": 0.01587662845849991,
+ "learning_rate": 0.0006,
+ "loss": 6.395642280578613,
+ "step": 307
+ },
+ {
+ "epoch": 4.279475982532751,
+ "grad_norm": 0.014650700613856316,
+ "learning_rate": 0.0006,
+ "loss": 6.302793025970459,
+ "step": 308
+ },
+ {
+ "epoch": 4.293449781659389,
+ "grad_norm": 0.016533225774765015,
+ "learning_rate": 0.0006,
+ "loss": 6.380588054656982,
+ "step": 309
+ },
+ {
+ "epoch": 4.307423580786026,
+ "grad_norm": 0.019963741302490234,
+ "learning_rate": 0.0006,
+ "loss": 6.345336437225342,
+ "step": 310
+ },
+ {
+ "epoch": 4.321397379912664,
+ "grad_norm": 0.02237936295568943,
+ "learning_rate": 0.0006,
+ "loss": 6.387770652770996,
+ "step": 311
+ },
+ {
+ "epoch": 4.335371179039301,
+ "grad_norm": 0.026658328250050545,
+ "learning_rate": 0.0006,
+ "loss": 6.323662757873535,
+ "step": 312
+ },
+ {
+ "epoch": 4.349344978165939,
+ "grad_norm": 0.02852284163236618,
+ "learning_rate": 0.0006,
+ "loss": 6.382204055786133,
+ "step": 313
+ },
+ {
+ "epoch": 4.3633187772925766,
+ "grad_norm": 0.028815090656280518,
+ "learning_rate": 0.0006,
+ "loss": 6.402390480041504,
+ "step": 314
+ },
+ {
+ "epoch": 4.377292576419214,
+ "grad_norm": 0.029393529519438744,
+ "learning_rate": 0.0006,
+ "loss": 6.280278205871582,
+ "step": 315
+ },
+ {
+ "epoch": 4.391266375545851,
+ "grad_norm": 0.02186041697859764,
+ "learning_rate": 0.0006,
+ "loss": 6.302707672119141,
+ "step": 316
+ },
+ {
+ "epoch": 4.405240174672489,
+ "grad_norm": 0.02071218006312847,
+ "learning_rate": 0.0006,
+ "loss": 6.3365020751953125,
+ "step": 317
+ },
+ {
+ "epoch": 4.419213973799127,
+ "grad_norm": 0.02779117226600647,
+ "learning_rate": 0.0006,
+ "loss": 6.252157211303711,
+ "step": 318
+ },
+ {
+ "epoch": 4.4331877729257645,
+ "grad_norm": 0.03470654785633087,
+ "learning_rate": 0.0006,
+ "loss": 6.319070339202881,
+ "step": 319
+ },
+ {
+ "epoch": 4.447161572052401,
+ "grad_norm": 0.037067804485559464,
+ "learning_rate": 0.0006,
+ "loss": 6.314116477966309,
+ "step": 320
+ },
+ {
+ "epoch": 4.461135371179039,
+ "grad_norm": 0.032182756811380386,
+ "learning_rate": 0.0006,
+ "loss": 6.267298698425293,
+ "step": 321
+ },
+ {
+ "epoch": 4.475109170305677,
+ "grad_norm": 0.026305217295885086,
+ "learning_rate": 0.0006,
+ "loss": 6.331688404083252,
+ "step": 322
+ },
+ {
+ "epoch": 4.489082969432315,
+ "grad_norm": 0.027228357270359993,
+ "learning_rate": 0.0006,
+ "loss": 6.31139612197876,
+ "step": 323
+ },
+ {
+ "epoch": 4.503056768558952,
+ "grad_norm": 0.025781169533729553,
+ "learning_rate": 0.0006,
+ "loss": 6.194684982299805,
+ "step": 324
+ },
+ {
+ "epoch": 4.517030567685589,
+ "grad_norm": 0.029186977073550224,
+ "learning_rate": 0.0006,
+ "loss": 6.31306266784668,
+ "step": 325
+ },
+ {
+ "epoch": 4.531004366812227,
+ "grad_norm": 0.023608777672052383,
+ "learning_rate": 0.0006,
+ "loss": 6.285243034362793,
+ "step": 326
+ },
+ {
+ "epoch": 4.544978165938865,
+ "grad_norm": 0.02267594076693058,
+ "learning_rate": 0.0006,
+ "loss": 6.30342960357666,
+ "step": 327
+ },
+ {
+ "epoch": 4.558951965065502,
+ "grad_norm": 0.01926310732960701,
+ "learning_rate": 0.0006,
+ "loss": 6.280606269836426,
+ "step": 328
+ },
+ {
+ "epoch": 4.5729257641921395,
+ "grad_norm": 0.026484837755560875,
+ "learning_rate": 0.0006,
+ "loss": 6.237980365753174,
+ "step": 329
+ },
+ {
+ "epoch": 4.586899563318777,
+ "grad_norm": 0.026228854432702065,
+ "learning_rate": 0.0006,
+ "loss": 6.3018293380737305,
+ "step": 330
+ },
+ {
+ "epoch": 4.600873362445415,
+ "grad_norm": 0.022096967324614525,
+ "learning_rate": 0.0006,
+ "loss": 6.253863334655762,
+ "step": 331
+ },
+ {
+ "epoch": 4.614847161572053,
+ "grad_norm": 0.027223890647292137,
+ "learning_rate": 0.0006,
+ "loss": 6.166882038116455,
+ "step": 332
+ },
+ {
+ "epoch": 4.62882096069869,
+ "grad_norm": 0.03160124272108078,
+ "learning_rate": 0.0006,
+ "loss": 6.230309963226318,
+ "step": 333
+ },
+ {
+ "epoch": 4.642794759825327,
+ "grad_norm": 0.04110539332032204,
+ "learning_rate": 0.0006,
+ "loss": 6.2334818840026855,
+ "step": 334
+ },
+ {
+ "epoch": 4.656768558951965,
+ "grad_norm": 0.04355933889746666,
+ "learning_rate": 0.0006,
+ "loss": 6.160956382751465,
+ "step": 335
+ },
+ {
+ "epoch": 4.670742358078603,
+ "grad_norm": 0.03611086308956146,
+ "learning_rate": 0.0006,
+ "loss": 6.23399543762207,
+ "step": 336
+ },
+ {
+ "epoch": 4.68471615720524,
+ "grad_norm": 0.051811400800943375,
+ "learning_rate": 0.0006,
+ "loss": 6.241855144500732,
+ "step": 337
+ },
+ {
+ "epoch": 4.698689956331878,
+ "grad_norm": 0.057434193789958954,
+ "learning_rate": 0.0006,
+ "loss": 6.220993995666504,
+ "step": 338
+ },
+ {
+ "epoch": 4.712663755458515,
+ "grad_norm": 0.04420953616499901,
+ "learning_rate": 0.0006,
+ "loss": 6.300461292266846,
+ "step": 339
+ },
+ {
+ "epoch": 4.726637554585153,
+ "grad_norm": 0.03515457361936569,
+ "learning_rate": 0.0006,
+ "loss": 6.269349098205566,
+ "step": 340
+ },
+ {
+ "epoch": 4.74061135371179,
+ "grad_norm": 0.03222600743174553,
+ "learning_rate": 0.0006,
+ "loss": 6.248613357543945,
+ "step": 341
+ },
+ {
+ "epoch": 4.754585152838428,
+ "grad_norm": 0.0384046845138073,
+ "learning_rate": 0.0006,
+ "loss": 6.266979217529297,
+ "step": 342
+ },
+ {
+ "epoch": 4.7685589519650655,
+ "grad_norm": 0.0514867827296257,
+ "learning_rate": 0.0006,
+ "loss": 6.234103202819824,
+ "step": 343
+ },
+ {
+ "epoch": 4.782532751091703,
+ "grad_norm": 0.050747793167829514,
+ "learning_rate": 0.0006,
+ "loss": 6.322582244873047,
+ "step": 344
+ },
+ {
+ "epoch": 4.796506550218341,
+ "grad_norm": 0.040878988802433014,
+ "learning_rate": 0.0006,
+ "loss": 6.212509632110596,
+ "step": 345
+ },
+ {
+ "epoch": 4.810480349344978,
+ "grad_norm": 0.034655820578336716,
+ "learning_rate": 0.0006,
+ "loss": 6.265879154205322,
+ "step": 346
+ },
+ {
+ "epoch": 4.824454148471616,
+ "grad_norm": 0.04202224686741829,
+ "learning_rate": 0.0006,
+ "loss": 6.2896409034729,
+ "step": 347
+ },
+ {
+ "epoch": 4.8384279475982535,
+ "grad_norm": 0.03632277995347977,
+ "learning_rate": 0.0006,
+ "loss": 6.253917217254639,
+ "step": 348
+ },
+ {
+ "epoch": 4.85240174672489,
+ "grad_norm": 0.028645765036344528,
+ "learning_rate": 0.0006,
+ "loss": 6.207965850830078,
+ "step": 349
+ },
+ {
+ "epoch": 4.866375545851528,
+ "grad_norm": 0.025305170565843582,
+ "learning_rate": 0.0006,
+ "loss": 6.168068885803223,
+ "step": 350
+ },
+ {
+ "epoch": 4.880349344978166,
+ "grad_norm": 0.022137803956866264,
+ "learning_rate": 0.0006,
+ "loss": 6.235283374786377,
+ "step": 351
+ },
+ {
+ "epoch": 4.894323144104804,
+ "grad_norm": 0.01814538985490799,
+ "learning_rate": 0.0006,
+ "loss": 6.259641647338867,
+ "step": 352
+ },
+ {
+ "epoch": 4.908296943231441,
+ "grad_norm": 0.018465086817741394,
+ "learning_rate": 0.0006,
+ "loss": 6.111738681793213,
+ "step": 353
+ },
+ {
+ "epoch": 4.922270742358078,
+ "grad_norm": 0.018638933077454567,
+ "learning_rate": 0.0006,
+ "loss": 6.225642204284668,
+ "step": 354
+ },
+ {
+ "epoch": 4.936244541484716,
+ "grad_norm": 0.016461260616779327,
+ "learning_rate": 0.0006,
+ "loss": 6.140647888183594,
+ "step": 355
+ },
+ {
+ "epoch": 4.950218340611354,
+ "grad_norm": 0.015458385460078716,
+ "learning_rate": 0.0006,
+ "loss": 6.155224800109863,
+ "step": 356
+ },
+ {
+ "epoch": 4.964192139737992,
+ "grad_norm": 0.014934048056602478,
+ "learning_rate": 0.0006,
+ "loss": 6.101555824279785,
+ "step": 357
+ },
+ {
+ "epoch": 4.978165938864628,
+ "grad_norm": 0.013979545794427395,
+ "learning_rate": 0.0006,
+ "loss": 6.161718845367432,
+ "step": 358
+ },
+ {
+ "epoch": 4.992139737991266,
+ "grad_norm": 0.01640394888818264,
+ "learning_rate": 0.0006,
+ "loss": 6.055768966674805,
+ "step": 359
+ },
+ {
+ "epoch": 5.0,
+ "grad_norm": 0.015135680325329304,
+ "learning_rate": 0.0006,
+ "loss": 6.169695854187012,
+ "step": 360
+ },
+ {
+ "epoch": 5.0,
+ "eval_loss": 6.16408109664917,
+ "eval_runtime": 59.3649,
+ "eval_samples_per_second": 41.135,
+ "eval_steps_per_second": 1.297,
+ "step": 360
+ },
+ {
+ "epoch": 5.013973799126638,
+ "grad_norm": 0.012849048711359501,
+ "learning_rate": 0.0006,
+ "loss": 6.055830001831055,
+ "step": 361
+ },
+ {
+ "epoch": 5.0279475982532755,
+ "grad_norm": 0.018169576302170753,
+ "learning_rate": 0.0006,
+ "loss": 6.082655429840088,
+ "step": 362
+ },
+ {
+ "epoch": 5.041921397379912,
+ "grad_norm": 0.024293430149555206,
+ "learning_rate": 0.0006,
+ "loss": 6.109032154083252,
+ "step": 363
+ },
+ {
+ "epoch": 5.05589519650655,
+ "grad_norm": 0.03266787901520729,
+ "learning_rate": 0.0006,
+ "loss": 6.030613899230957,
+ "step": 364
+ },
+ {
+ "epoch": 5.069868995633188,
+ "grad_norm": 0.03572266176342964,
+ "learning_rate": 0.0006,
+ "loss": 6.050390720367432,
+ "step": 365
+ },
+ {
+ "epoch": 5.083842794759826,
+ "grad_norm": 0.032625213265419006,
+ "learning_rate": 0.0006,
+ "loss": 6.169313430786133,
+ "step": 366
+ },
+ {
+ "epoch": 5.097816593886463,
+ "grad_norm": 0.04051872715353966,
+ "learning_rate": 0.0006,
+ "loss": 6.10398006439209,
+ "step": 367
+ },
+ {
+ "epoch": 5.1117903930131,
+ "grad_norm": 0.0351213738322258,
+ "learning_rate": 0.0006,
+ "loss": 6.121973514556885,
+ "step": 368
+ },
+ {
+ "epoch": 5.125764192139738,
+ "grad_norm": 0.03523759916424751,
+ "learning_rate": 0.0006,
+ "loss": 6.0626654624938965,
+ "step": 369
+ },
+ {
+ "epoch": 5.139737991266376,
+ "grad_norm": 0.02706182189285755,
+ "learning_rate": 0.0006,
+ "loss": 6.049354076385498,
+ "step": 370
+ },
+ {
+ "epoch": 5.153711790393013,
+ "grad_norm": 0.024212589487433434,
+ "learning_rate": 0.0006,
+ "loss": 6.034826278686523,
+ "step": 371
+ },
+ {
+ "epoch": 5.1676855895196505,
+ "grad_norm": 0.024798082187771797,
+ "learning_rate": 0.0006,
+ "loss": 6.004058837890625,
+ "step": 372
+ },
+ {
+ "epoch": 5.181659388646288,
+ "grad_norm": 0.02325567416846752,
+ "learning_rate": 0.0006,
+ "loss": 5.986461639404297,
+ "step": 373
+ },
+ {
+ "epoch": 5.195633187772926,
+ "grad_norm": 0.019060570746660233,
+ "learning_rate": 0.0006,
+ "loss": 6.059736251831055,
+ "step": 374
+ },
+ {
+ "epoch": 5.209606986899563,
+ "grad_norm": 0.016822976991534233,
+ "learning_rate": 0.0006,
+ "loss": 6.072957515716553,
+ "step": 375
+ },
+ {
+ "epoch": 5.223580786026201,
+ "grad_norm": 0.017218658700585365,
+ "learning_rate": 0.0006,
+ "loss": 6.031739234924316,
+ "step": 376
+ },
+ {
+ "epoch": 5.2375545851528384,
+ "grad_norm": 0.014889383688569069,
+ "learning_rate": 0.0006,
+ "loss": 6.025674343109131,
+ "step": 377
+ },
+ {
+ "epoch": 5.251528384279476,
+ "grad_norm": 0.01708107627928257,
+ "learning_rate": 0.0006,
+ "loss": 6.090451240539551,
+ "step": 378
+ },
+ {
+ "epoch": 5.265502183406113,
+ "grad_norm": 0.020260317251086235,
+ "learning_rate": 0.0006,
+ "loss": 6.094257354736328,
+ "step": 379
+ },
+ {
+ "epoch": 5.279475982532751,
+ "grad_norm": 0.020110400393605232,
+ "learning_rate": 0.0006,
+ "loss": 6.021188735961914,
+ "step": 380
+ },
+ {
+ "epoch": 5.293449781659389,
+ "grad_norm": 0.02027830481529236,
+ "learning_rate": 0.0006,
+ "loss": 6.022156238555908,
+ "step": 381
+ },
+ {
+ "epoch": 5.307423580786026,
+ "grad_norm": 0.022747062146663666,
+ "learning_rate": 0.0006,
+ "loss": 6.011836051940918,
+ "step": 382
+ },
+ {
+ "epoch": 5.321397379912664,
+ "grad_norm": 0.01990230567753315,
+ "learning_rate": 0.0006,
+ "loss": 6.003054618835449,
+ "step": 383
+ },
+ {
+ "epoch": 5.335371179039301,
+ "grad_norm": 0.01596238650381565,
+ "learning_rate": 0.0006,
+ "loss": 5.991410732269287,
+ "step": 384
+ },
+ {
+ "epoch": 5.349344978165939,
+ "grad_norm": 0.015847105532884598,
+ "learning_rate": 0.0006,
+ "loss": 6.100622177124023,
+ "step": 385
+ },
+ {
+ "epoch": 5.3633187772925766,
+ "grad_norm": 0.016179397702217102,
+ "learning_rate": 0.0006,
+ "loss": 6.032659530639648,
+ "step": 386
+ },
+ {
+ "epoch": 5.377292576419214,
+ "grad_norm": 0.018256602808833122,
+ "learning_rate": 0.0006,
+ "loss": 5.961983680725098,
+ "step": 387
+ },
+ {
+ "epoch": 5.391266375545851,
+ "grad_norm": 0.02805912122130394,
+ "learning_rate": 0.0006,
+ "loss": 5.983541965484619,
+ "step": 388
+ },
+ {
+ "epoch": 5.405240174672489,
+ "grad_norm": 0.039082255214452744,
+ "learning_rate": 0.0006,
+ "loss": 6.021474361419678,
+ "step": 389
+ },
+ {
+ "epoch": 5.419213973799127,
+ "grad_norm": 0.036757659167051315,
+ "learning_rate": 0.0006,
+ "loss": 6.005046367645264,
+ "step": 390
+ },
+ {
+ "epoch": 5.4331877729257645,
+ "grad_norm": 0.035277366638183594,
+ "learning_rate": 0.0006,
+ "loss": 6.036296844482422,
+ "step": 391
+ },
+ {
+ "epoch": 5.447161572052401,
+ "grad_norm": 0.034404635429382324,
+ "learning_rate": 0.0006,
+ "loss": 5.994539260864258,
+ "step": 392
+ },
+ {
+ "epoch": 5.461135371179039,
+ "grad_norm": 0.0377206988632679,
+ "learning_rate": 0.0006,
+ "loss": 5.941534519195557,
+ "step": 393
+ },
+ {
+ "epoch": 5.475109170305677,
+ "grad_norm": 0.0389922633767128,
+ "learning_rate": 0.0006,
+ "loss": 5.983644485473633,
+ "step": 394
+ },
+ {
+ "epoch": 5.489082969432315,
+ "grad_norm": 0.04176778718829155,
+ "learning_rate": 0.0006,
+ "loss": 6.030922889709473,
+ "step": 395
+ },
+ {
+ "epoch": 5.503056768558952,
+ "grad_norm": 0.0337153784930706,
+ "learning_rate": 0.0006,
+ "loss": 5.961367130279541,
+ "step": 396
+ },
+ {
+ "epoch": 5.517030567685589,
+ "grad_norm": 0.036116816103458405,
+ "learning_rate": 0.0006,
+ "loss": 5.819280624389648,
+ "step": 397
+ },
+ {
+ "epoch": 5.531004366812227,
+ "grad_norm": 0.030724041163921356,
+ "learning_rate": 0.0006,
+ "loss": 5.906380653381348,
+ "step": 398
+ },
+ {
+ "epoch": 5.544978165938865,
+ "grad_norm": 0.030264202505350113,
+ "learning_rate": 0.0006,
+ "loss": 5.975833415985107,
+ "step": 399
+ },
+ {
+ "epoch": 5.558951965065502,
+ "grad_norm": 0.032096318900585175,
+ "learning_rate": 0.0006,
+ "loss": 6.001348972320557,
+ "step": 400
+ },
+ {
+ "epoch": 5.5729257641921395,
+ "grad_norm": 0.030579356476664543,
+ "learning_rate": 0.0006,
+ "loss": 5.928768634796143,
+ "step": 401
+ },
+ {
+ "epoch": 5.586899563318777,
+ "grad_norm": 0.028241973370313644,
+ "learning_rate": 0.0006,
+ "loss": 5.92582893371582,
+ "step": 402
+ },
+ {
+ "epoch": 5.600873362445415,
+ "grad_norm": 0.023086953908205032,
+ "learning_rate": 0.0006,
+ "loss": 5.892926216125488,
+ "step": 403
+ },
+ {
+ "epoch": 5.614847161572053,
+ "grad_norm": 0.02547992393374443,
+ "learning_rate": 0.0006,
+ "loss": 5.903195381164551,
+ "step": 404
+ },
+ {
+ "epoch": 5.62882096069869,
+ "grad_norm": 0.023089831694960594,
+ "learning_rate": 0.0006,
+ "loss": 5.912033557891846,
+ "step": 405
+ },
+ {
+ "epoch": 5.642794759825327,
+ "grad_norm": 0.025446368381381035,
+ "learning_rate": 0.0006,
+ "loss": 5.926138401031494,
+ "step": 406
+ },
+ {
+ "epoch": 5.656768558951965,
+ "grad_norm": 0.03748747706413269,
+ "learning_rate": 0.0006,
+ "loss": 5.913451194763184,
+ "step": 407
+ },
+ {
+ "epoch": 5.670742358078603,
+ "grad_norm": 0.03493810072541237,
+ "learning_rate": 0.0006,
+ "loss": 5.97633695602417,
+ "step": 408
+ },
+ {
+ "epoch": 5.68471615720524,
+ "grad_norm": 0.023131275549530983,
+ "learning_rate": 0.0006,
+ "loss": 5.899571418762207,
+ "step": 409
+ },
+ {
+ "epoch": 5.698689956331878,
+ "grad_norm": 0.02986014075577259,
+ "learning_rate": 0.0006,
+ "loss": 5.894903182983398,
+ "step": 410
+ },
+ {
+ "epoch": 5.712663755458515,
+ "grad_norm": 0.030171144753694534,
+ "learning_rate": 0.0006,
+ "loss": 5.876595497131348,
+ "step": 411
+ },
+ {
+ "epoch": 5.726637554585153,
+ "grad_norm": 0.029377546161413193,
+ "learning_rate": 0.0006,
+ "loss": 5.911637306213379,
+ "step": 412
+ },
+ {
+ "epoch": 5.74061135371179,
+ "grad_norm": 0.029829490929841995,
+ "learning_rate": 0.0006,
+ "loss": 5.819629669189453,
+ "step": 413
+ },
+ {
+ "epoch": 5.754585152838428,
+ "grad_norm": 0.024139299988746643,
+ "learning_rate": 0.0006,
+ "loss": 5.81728458404541,
+ "step": 414
+ },
+ {
+ "epoch": 5.7685589519650655,
+ "grad_norm": 0.021453432738780975,
+ "learning_rate": 0.0006,
+ "loss": 5.987326145172119,
+ "step": 415
+ },
+ {
+ "epoch": 5.782532751091703,
+ "grad_norm": 0.017942246049642563,
+ "learning_rate": 0.0006,
+ "loss": 5.893008232116699,
+ "step": 416
+ },
+ {
+ "epoch": 5.796506550218341,
+ "grad_norm": 0.019724637269973755,
+ "learning_rate": 0.0006,
+ "loss": 5.912441253662109,
+ "step": 417
+ },
+ {
+ "epoch": 5.810480349344978,
+ "grad_norm": 0.018597912043333054,
+ "learning_rate": 0.0006,
+ "loss": 5.860394477844238,
+ "step": 418
+ },
+ {
+ "epoch": 5.824454148471616,
+ "grad_norm": 0.016232412308454514,
+ "learning_rate": 0.0006,
+ "loss": 5.94991397857666,
+ "step": 419
+ },
+ {
+ "epoch": 5.8384279475982535,
+ "grad_norm": 0.01667204685509205,
+ "learning_rate": 0.0006,
+ "loss": 5.8506364822387695,
+ "step": 420
+ },
+ {
+ "epoch": 5.85240174672489,
+ "grad_norm": 0.014232151210308075,
+ "learning_rate": 0.0006,
+ "loss": 5.928424835205078,
+ "step": 421
+ },
+ {
+ "epoch": 5.866375545851528,
+ "grad_norm": 0.01570476032793522,
+ "learning_rate": 0.0006,
+ "loss": 5.953692436218262,
+ "step": 422
+ },
+ {
+ "epoch": 5.880349344978166,
+ "grad_norm": 0.01586179807782173,
+ "learning_rate": 0.0006,
+ "loss": 5.820014476776123,
+ "step": 423
+ },
+ {
+ "epoch": 5.894323144104804,
+ "grad_norm": 0.014729145914316177,
+ "learning_rate": 0.0006,
+ "loss": 5.860154628753662,
+ "step": 424
+ },
+ {
+ "epoch": 5.908296943231441,
+ "grad_norm": 0.013111459091305733,
+ "learning_rate": 0.0006,
+ "loss": 5.85988187789917,
+ "step": 425
+ },
+ {
+ "epoch": 5.922270742358078,
+ "grad_norm": 0.012937922962009907,
+ "learning_rate": 0.0006,
+ "loss": 5.756265640258789,
+ "step": 426
+ },
+ {
+ "epoch": 5.936244541484716,
+ "grad_norm": 0.01377453189343214,
+ "learning_rate": 0.0006,
+ "loss": 5.8333048820495605,
+ "step": 427
+ },
+ {
+ "epoch": 5.950218340611354,
+ "grad_norm": 0.014054795727133751,
+ "learning_rate": 0.0006,
+ "loss": 5.803333759307861,
+ "step": 428
+ },
+ {
+ "epoch": 5.964192139737992,
+ "grad_norm": 0.01674959622323513,
+ "learning_rate": 0.0006,
+ "loss": 5.8718485832214355,
+ "step": 429
+ },
+ {
+ "epoch": 5.978165938864628,
+ "grad_norm": 0.018733017146587372,
+ "learning_rate": 0.0006,
+ "loss": 5.875979900360107,
+ "step": 430
+ },
+ {
+ "epoch": 5.992139737991266,
+ "grad_norm": 0.02088983915746212,
+ "learning_rate": 0.0006,
+ "loss": 5.763634204864502,
+ "step": 431
+ },
+ {
+ "epoch": 6.0,
+ "grad_norm": 0.026705985888838768,
+ "learning_rate": 0.0006,
+ "loss": 5.6922478675842285,
+ "step": 432
+ },
+ {
+ "epoch": 6.0,
+ "eval_loss": 5.847958087921143,
+ "eval_runtime": 58.5451,
+ "eval_samples_per_second": 41.711,
+ "eval_steps_per_second": 1.315,
+ "step": 432
+ },
+ {
+ "epoch": 6.013973799126638,
+ "grad_norm": 0.03181544691324234,
+ "learning_rate": 0.0006,
+ "loss": 5.786368370056152,
+ "step": 433
+ },
+ {
+ "epoch": 6.0279475982532755,
+ "grad_norm": 0.03238112851977348,
+ "learning_rate": 0.0006,
+ "loss": 5.810310363769531,
+ "step": 434
+ },
+ {
+ "epoch": 6.041921397379912,
+ "grad_norm": 0.03916756808757782,
+ "learning_rate": 0.0006,
+ "loss": 5.7761993408203125,
+ "step": 435
+ },
+ {
+ "epoch": 6.05589519650655,
+ "grad_norm": 0.044009730219841,
+ "learning_rate": 0.0006,
+ "loss": 5.898112773895264,
+ "step": 436
+ },
+ {
+ "epoch": 6.069868995633188,
+ "grad_norm": 0.045235246419906616,
+ "learning_rate": 0.0006,
+ "loss": 5.876371383666992,
+ "step": 437
+ },
+ {
+ "epoch": 6.083842794759826,
+ "grad_norm": 0.050218384712934494,
+ "learning_rate": 0.0006,
+ "loss": 5.805103302001953,
+ "step": 438
+ },
+ {
+ "epoch": 6.097816593886463,
+ "grad_norm": 0.05444490164518356,
+ "learning_rate": 0.0006,
+ "loss": 5.826424598693848,
+ "step": 439
+ },
+ {
+ "epoch": 6.1117903930131,
+ "grad_norm": 0.04890982061624527,
+ "learning_rate": 0.0006,
+ "loss": 5.7735395431518555,
+ "step": 440
+ },
+ {
+ "epoch": 6.125764192139738,
+ "grad_norm": 0.05478639155626297,
+ "learning_rate": 0.0006,
+ "loss": 5.823677062988281,
+ "step": 441
+ },
+ {
+ "epoch": 6.139737991266376,
+ "grad_norm": 0.057562313973903656,
+ "learning_rate": 0.0006,
+ "loss": 5.889334678649902,
+ "step": 442
+ },
+ {
+ "epoch": 6.153711790393013,
+ "grad_norm": 0.06447703391313553,
+ "learning_rate": 0.0006,
+ "loss": 5.84621524810791,
+ "step": 443
+ },
+ {
+ "epoch": 6.1676855895196505,
+ "grad_norm": 0.048861872404813766,
+ "learning_rate": 0.0006,
+ "loss": 5.931595802307129,
+ "step": 444
+ },
+ {
+ "epoch": 6.181659388646288,
+ "grad_norm": 0.05521339178085327,
+ "learning_rate": 0.0006,
+ "loss": 5.8229780197143555,
+ "step": 445
+ },
+ {
+ "epoch": 6.195633187772926,
+ "grad_norm": 0.053666990250349045,
+ "learning_rate": 0.0006,
+ "loss": 5.879191875457764,
+ "step": 446
+ },
+ {
+ "epoch": 6.209606986899563,
+ "grad_norm": 0.0451025515794754,
+ "learning_rate": 0.0006,
+ "loss": 5.801628112792969,
+ "step": 447
+ },
+ {
+ "epoch": 6.223580786026201,
+ "grad_norm": 0.04291655868291855,
+ "learning_rate": 0.0006,
+ "loss": 5.914680480957031,
+ "step": 448
+ },
+ {
+ "epoch": 6.2375545851528384,
+ "grad_norm": 0.035102490335702896,
+ "learning_rate": 0.0006,
+ "loss": 5.843682765960693,
+ "step": 449
+ },
+ {
+ "epoch": 6.251528384279476,
+ "grad_norm": 0.03403995931148529,
+ "learning_rate": 0.0006,
+ "loss": 5.827154159545898,
+ "step": 450
+ },
+ {
+ "epoch": 6.265502183406113,
+ "grad_norm": 0.03444375470280647,
+ "learning_rate": 0.0006,
+ "loss": 5.800136566162109,
+ "step": 451
+ },
+ {
+ "epoch": 6.279475982532751,
+ "grad_norm": 0.03165159001946449,
+ "learning_rate": 0.0006,
+ "loss": 5.906252384185791,
+ "step": 452
+ },
+ {
+ "epoch": 6.293449781659389,
+ "grad_norm": 0.026153093203902245,
+ "learning_rate": 0.0006,
+ "loss": 5.7423200607299805,
+ "step": 453
+ },
+ {
+ "epoch": 6.307423580786026,
+ "grad_norm": 0.024357657879590988,
+ "learning_rate": 0.0006,
+ "loss": 5.701364517211914,
+ "step": 454
+ },
+ {
+ "epoch": 6.321397379912664,
+ "grad_norm": 0.021508049219846725,
+ "learning_rate": 0.0006,
+ "loss": 5.8328094482421875,
+ "step": 455
+ },
+ {
+ "epoch": 6.335371179039301,
+ "grad_norm": 0.017313921824097633,
+ "learning_rate": 0.0006,
+ "loss": 5.792145252227783,
+ "step": 456
+ },
+ {
+ "epoch": 6.349344978165939,
+ "grad_norm": 0.018563397228717804,
+ "learning_rate": 0.0006,
+ "loss": 5.732672214508057,
+ "step": 457
+ },
+ {
+ "epoch": 6.3633187772925766,
+ "grad_norm": 0.016568679362535477,
+ "learning_rate": 0.0006,
+ "loss": 5.81948184967041,
+ "step": 458
+ },
+ {
+ "epoch": 6.377292576419214,
+ "grad_norm": 0.014133770950138569,
+ "learning_rate": 0.0006,
+ "loss": 5.734982490539551,
+ "step": 459
+ },
+ {
+ "epoch": 6.391266375545851,
+ "grad_norm": 0.014114447869360447,
+ "learning_rate": 0.0006,
+ "loss": 5.70073127746582,
+ "step": 460
+ },
+ {
+ "epoch": 6.405240174672489,
+ "grad_norm": 0.012907393276691437,
+ "learning_rate": 0.0006,
+ "loss": 5.758626937866211,
+ "step": 461
+ },
+ {
+ "epoch": 6.419213973799127,
+ "grad_norm": 0.012673444114625454,
+ "learning_rate": 0.0006,
+ "loss": 5.775138854980469,
+ "step": 462
+ },
+ {
+ "epoch": 6.4331877729257645,
+ "grad_norm": 0.01311410591006279,
+ "learning_rate": 0.0006,
+ "loss": 5.70920991897583,
+ "step": 463
+ },
+ {
+ "epoch": 6.447161572052401,
+ "grad_norm": 0.012935544364154339,
+ "learning_rate": 0.0006,
+ "loss": 5.781479835510254,
+ "step": 464
+ },
+ {
+ "epoch": 6.461135371179039,
+ "grad_norm": 0.01439738366752863,
+ "learning_rate": 0.0006,
+ "loss": 5.691315650939941,
+ "step": 465
+ },
+ {
+ "epoch": 6.475109170305677,
+ "grad_norm": 0.01205186452716589,
+ "learning_rate": 0.0006,
+ "loss": 5.662499904632568,
+ "step": 466
+ },
+ {
+ "epoch": 6.489082969432315,
+ "grad_norm": 0.011575652286410332,
+ "learning_rate": 0.0006,
+ "loss": 5.6271772384643555,
+ "step": 467
+ },
+ {
+ "epoch": 6.503056768558952,
+ "grad_norm": 0.011672502383589745,
+ "learning_rate": 0.0006,
+ "loss": 5.761662006378174,
+ "step": 468
+ },
+ {
+ "epoch": 6.517030567685589,
+ "grad_norm": 0.011541751213371754,
+ "learning_rate": 0.0006,
+ "loss": 5.762078285217285,
+ "step": 469
+ },
+ {
+ "epoch": 6.531004366812227,
+ "grad_norm": 0.011396365240216255,
+ "learning_rate": 0.0006,
+ "loss": 5.67873477935791,
+ "step": 470
+ },
+ {
+ "epoch": 6.544978165938865,
+ "grad_norm": 0.010688499547541142,
+ "learning_rate": 0.0006,
+ "loss": 5.717051982879639,
+ "step": 471
+ },
+ {
+ "epoch": 6.558951965065502,
+ "grad_norm": 0.012224317528307438,
+ "learning_rate": 0.0006,
+ "loss": 5.707948684692383,
+ "step": 472
+ },
+ {
+ "epoch": 6.5729257641921395,
+ "grad_norm": 0.011856825090944767,
+ "learning_rate": 0.0006,
+ "loss": 5.70878791809082,
+ "step": 473
+ },
+ {
+ "epoch": 6.586899563318777,
+ "grad_norm": 0.01199064590036869,
+ "learning_rate": 0.0006,
+ "loss": 5.708697319030762,
+ "step": 474
+ },
+ {
+ "epoch": 6.600873362445415,
+ "grad_norm": 0.01219446212053299,
+ "learning_rate": 0.0006,
+ "loss": 5.6061577796936035,
+ "step": 475
+ },
+ {
+ "epoch": 6.614847161572053,
+ "grad_norm": 0.013397484086453915,
+ "learning_rate": 0.0006,
+ "loss": 5.624789714813232,
+ "step": 476
+ },
+ {
+ "epoch": 6.62882096069869,
+ "grad_norm": 0.01483136136084795,
+ "learning_rate": 0.0006,
+ "loss": 5.721141338348389,
+ "step": 477
+ },
+ {
+ "epoch": 6.642794759825327,
+ "grad_norm": 0.0194488987326622,
+ "learning_rate": 0.0006,
+ "loss": 5.604279518127441,
+ "step": 478
+ },
+ {
+ "epoch": 6.656768558951965,
+ "grad_norm": 0.02079896256327629,
+ "learning_rate": 0.0006,
+ "loss": 5.586322784423828,
+ "step": 479
+ },
+ {
+ "epoch": 6.670742358078603,
+ "grad_norm": 0.018996229395270348,
+ "learning_rate": 0.0006,
+ "loss": 5.5987772941589355,
+ "step": 480
+ },
+ {
+ "epoch": 6.68471615720524,
+ "grad_norm": 0.015779603272676468,
+ "learning_rate": 0.0006,
+ "loss": 5.588602066040039,
+ "step": 481
+ },
+ {
+ "epoch": 6.698689956331878,
+ "grad_norm": 0.015322973020374775,
+ "learning_rate": 0.0006,
+ "loss": 5.686467170715332,
+ "step": 482
+ },
+ {
+ "epoch": 6.712663755458515,
+ "grad_norm": 0.015282213687896729,
+ "learning_rate": 0.0006,
+ "loss": 5.664676666259766,
+ "step": 483
+ },
+ {
+ "epoch": 6.726637554585153,
+ "grad_norm": 0.016589272767305374,
+ "learning_rate": 0.0006,
+ "loss": 5.666739463806152,
+ "step": 484
+ },
+ {
+ "epoch": 6.74061135371179,
+ "grad_norm": 0.01811421848833561,
+ "learning_rate": 0.0006,
+ "loss": 5.647593021392822,
+ "step": 485
+ },
+ {
+ "epoch": 6.754585152838428,
+ "grad_norm": 0.019652029499411583,
+ "learning_rate": 0.0006,
+ "loss": 5.628726959228516,
+ "step": 486
+ },
+ {
+ "epoch": 6.7685589519650655,
+ "grad_norm": 0.02118665538728237,
+ "learning_rate": 0.0006,
+ "loss": 5.6580705642700195,
+ "step": 487
+ },
+ {
+ "epoch": 6.782532751091703,
+ "grad_norm": 0.02237832546234131,
+ "learning_rate": 0.0006,
+ "loss": 5.635931968688965,
+ "step": 488
+ },
+ {
+ "epoch": 6.796506550218341,
+ "grad_norm": 0.023897631093859673,
+ "learning_rate": 0.0006,
+ "loss": 5.517949104309082,
+ "step": 489
+ },
+ {
+ "epoch": 6.810480349344978,
+ "grad_norm": 0.02442227490246296,
+ "learning_rate": 0.0006,
+ "loss": 5.600021839141846,
+ "step": 490
+ },
+ {
+ "epoch": 6.824454148471616,
+ "grad_norm": 0.024675287306308746,
+ "learning_rate": 0.0006,
+ "loss": 5.6639909744262695,
+ "step": 491
+ },
+ {
+ "epoch": 6.8384279475982535,
+ "grad_norm": 0.030372392386198044,
+ "learning_rate": 0.0006,
+ "loss": 5.582888603210449,
+ "step": 492
+ },
+ {
+ "epoch": 6.85240174672489,
+ "grad_norm": 0.03237690031528473,
+ "learning_rate": 0.0006,
+ "loss": 5.573896884918213,
+ "step": 493
+ },
+ {
+ "epoch": 6.866375545851528,
+ "grad_norm": 0.03511589393019676,
+ "learning_rate": 0.0006,
+ "loss": 5.565878868103027,
+ "step": 494
+ },
+ {
+ "epoch": 6.880349344978166,
+ "grad_norm": 0.03883938118815422,
+ "learning_rate": 0.0006,
+ "loss": 5.58540678024292,
+ "step": 495
+ },
+ {
+ "epoch": 6.894323144104804,
+ "grad_norm": 0.04175502806901932,
+ "learning_rate": 0.0006,
+ "loss": 5.6095194816589355,
+ "step": 496
+ },
+ {
+ "epoch": 6.908296943231441,
+ "grad_norm": 0.04013441503047943,
+ "learning_rate": 0.0006,
+ "loss": 5.654401779174805,
+ "step": 497
+ },
+ {
+ "epoch": 6.922270742358078,
+ "grad_norm": 0.04618770629167557,
+ "learning_rate": 0.0006,
+ "loss": 5.6483473777771,
+ "step": 498
+ },
+ {
+ "epoch": 6.936244541484716,
+ "grad_norm": 0.04507308453321457,
+ "learning_rate": 0.0006,
+ "loss": 5.5856218338012695,
+ "step": 499
+ },
+ {
+ "epoch": 6.950218340611354,
+ "grad_norm": 0.03151383996009827,
+ "learning_rate": 0.0006,
+ "loss": 5.594086647033691,
+ "step": 500
+ },
+ {
+ "epoch": 6.964192139737992,
+ "grad_norm": 0.026883797720074654,
+ "learning_rate": 0.0006,
+ "loss": 5.607676029205322,
+ "step": 501
+ },
+ {
+ "epoch": 6.978165938864628,
+ "grad_norm": 0.02981836162507534,
+ "learning_rate": 0.0006,
+ "loss": 5.563666820526123,
+ "step": 502
+ },
+ {
+ "epoch": 6.992139737991266,
+ "grad_norm": 0.03233477845788002,
+ "learning_rate": 0.0006,
+ "loss": 5.667543411254883,
+ "step": 503
+ },
+ {
+ "epoch": 7.0,
+ "grad_norm": 0.029927456751465797,
+ "learning_rate": 0.0006,
+ "loss": 5.733482837677002,
+ "step": 504
+ },
+ {
+ "epoch": 7.0,
+ "eval_loss": 5.679076671600342,
+ "eval_runtime": 59.5965,
+ "eval_samples_per_second": 40.976,
+ "eval_steps_per_second": 1.292,
+ "step": 504
+ },
+ {
+ "epoch": 7.013973799126638,
+ "grad_norm": 0.027081597596406937,
+ "learning_rate": 0.0006,
+ "loss": 5.634487152099609,
+ "step": 505
+ },
+ {
+ "epoch": 7.0279475982532755,
+ "grad_norm": 0.031959421932697296,
+ "learning_rate": 0.0006,
+ "loss": 5.643294811248779,
+ "step": 506
+ },
+ {
+ "epoch": 7.041921397379912,
+ "grad_norm": 0.02802063524723053,
+ "learning_rate": 0.0006,
+ "loss": 5.573239326477051,
+ "step": 507
+ },
+ {
+ "epoch": 7.05589519650655,
+ "grad_norm": 0.032579705119132996,
+ "learning_rate": 0.0006,
+ "loss": 5.627150058746338,
+ "step": 508
+ },
+ {
+ "epoch": 7.069868995633188,
+ "grad_norm": 0.02503928542137146,
+ "learning_rate": 0.0006,
+ "loss": 5.6045308113098145,
+ "step": 509
+ },
+ {
+ "epoch": 7.083842794759826,
+ "grad_norm": 0.02421317622065544,
+ "learning_rate": 0.0006,
+ "loss": 5.522153377532959,
+ "step": 510
+ },
+ {
+ "epoch": 7.097816593886463,
+ "grad_norm": 0.021983426064252853,
+ "learning_rate": 0.0006,
+ "loss": 5.632939338684082,
+ "step": 511
+ },
+ {
+ "epoch": 7.1117903930131,
+ "grad_norm": 0.021933183073997498,
+ "learning_rate": 0.0006,
+ "loss": 5.592185974121094,
+ "step": 512
+ },
+ {
+ "epoch": 7.125764192139738,
+ "grad_norm": 0.02287031151354313,
+ "learning_rate": 0.0006,
+ "loss": 5.613700866699219,
+ "step": 513
+ },
+ {
+ "epoch": 7.139737991266376,
+ "grad_norm": 0.02281602844595909,
+ "learning_rate": 0.0006,
+ "loss": 5.551383018493652,
+ "step": 514
+ },
+ {
+ "epoch": 7.153711790393013,
+ "grad_norm": 0.021050360053777695,
+ "learning_rate": 0.0006,
+ "loss": 5.55275821685791,
+ "step": 515
+ },
+ {
+ "epoch": 7.1676855895196505,
+ "grad_norm": 0.018299926072359085,
+ "learning_rate": 0.0006,
+ "loss": 5.528225421905518,
+ "step": 516
+ },
+ {
+ "epoch": 7.181659388646288,
+ "grad_norm": 0.02024853602051735,
+ "learning_rate": 0.0006,
+ "loss": 5.522453784942627,
+ "step": 517
+ },
+ {
+ "epoch": 7.195633187772926,
+ "grad_norm": 0.02135239914059639,
+ "learning_rate": 0.0006,
+ "loss": 5.527522563934326,
+ "step": 518
+ },
+ {
+ "epoch": 7.209606986899563,
+ "grad_norm": 0.023394184187054634,
+ "learning_rate": 0.0006,
+ "loss": 5.4535017013549805,
+ "step": 519
+ },
+ {
+ "epoch": 7.223580786026201,
+ "grad_norm": 0.025952080264687538,
+ "learning_rate": 0.0006,
+ "loss": 5.5374674797058105,
+ "step": 520
+ },
+ {
+ "epoch": 7.2375545851528384,
+ "grad_norm": 0.022287718951702118,
+ "learning_rate": 0.0006,
+ "loss": 5.493753433227539,
+ "step": 521
+ },
+ {
+ "epoch": 7.251528384279476,
+ "grad_norm": 0.02018447406589985,
+ "learning_rate": 0.0006,
+ "loss": 5.510575294494629,
+ "step": 522
+ },
+ {
+ "epoch": 7.265502183406113,
+ "grad_norm": 0.02050507813692093,
+ "learning_rate": 0.0006,
+ "loss": 5.466026782989502,
+ "step": 523
+ },
+ {
+ "epoch": 7.279475982532751,
+ "grad_norm": 0.023688802495598793,
+ "learning_rate": 0.0006,
+ "loss": 5.485815525054932,
+ "step": 524
+ },
+ {
+ "epoch": 7.293449781659389,
+ "grad_norm": 0.021078843623399734,
+ "learning_rate": 0.0006,
+ "loss": 5.415581703186035,
+ "step": 525
+ },
+ {
+ "epoch": 7.307423580786026,
+ "grad_norm": 0.018266011029481888,
+ "learning_rate": 0.0006,
+ "loss": 5.402815818786621,
+ "step": 526
+ },
+ {
+ "epoch": 7.321397379912664,
+ "grad_norm": 0.019693493843078613,
+ "learning_rate": 0.0006,
+ "loss": 5.504674911499023,
+ "step": 527
+ },
+ {
+ "epoch": 7.335371179039301,
+ "grad_norm": 0.02031373605132103,
+ "learning_rate": 0.0006,
+ "loss": 5.503837585449219,
+ "step": 528
+ },
+ {
+ "epoch": 7.349344978165939,
+ "grad_norm": 0.017636626958847046,
+ "learning_rate": 0.0006,
+ "loss": 5.530580997467041,
+ "step": 529
+ },
+ {
+ "epoch": 7.3633187772925766,
+ "grad_norm": 0.01787244901061058,
+ "learning_rate": 0.0006,
+ "loss": 5.450218677520752,
+ "step": 530
+ },
+ {
+ "epoch": 7.377292576419214,
+ "grad_norm": 0.0170669574290514,
+ "learning_rate": 0.0006,
+ "loss": 5.481570243835449,
+ "step": 531
+ },
+ {
+ "epoch": 7.391266375545851,
+ "grad_norm": 0.01802165061235428,
+ "learning_rate": 0.0006,
+ "loss": 5.386394500732422,
+ "step": 532
+ },
+ {
+ "epoch": 7.405240174672489,
+ "grad_norm": 0.022949472069740295,
+ "learning_rate": 0.0006,
+ "loss": 5.424929618835449,
+ "step": 533
+ },
+ {
+ "epoch": 7.419213973799127,
+ "grad_norm": 0.034213390201330185,
+ "learning_rate": 0.0006,
+ "loss": 5.346663475036621,
+ "step": 534
+ },
+ {
+ "epoch": 7.4331877729257645,
+ "grad_norm": 0.03847593814134598,
+ "learning_rate": 0.0006,
+ "loss": 5.418482780456543,
+ "step": 535
+ },
+ {
+ "epoch": 7.447161572052401,
+ "grad_norm": 0.029393676668405533,
+ "learning_rate": 0.0006,
+ "loss": 5.456090927124023,
+ "step": 536
+ },
+ {
+ "epoch": 7.461135371179039,
+ "grad_norm": 0.03388667106628418,
+ "learning_rate": 0.0006,
+ "loss": 5.572983741760254,
+ "step": 537
+ },
+ {
+ "epoch": 7.475109170305677,
+ "grad_norm": 0.039690613746643066,
+ "learning_rate": 0.0006,
+ "loss": 5.414067268371582,
+ "step": 538
+ },
+ {
+ "epoch": 7.489082969432315,
+ "grad_norm": 0.03630776330828667,
+ "learning_rate": 0.0006,
+ "loss": 5.523739814758301,
+ "step": 539
+ },
+ {
+ "epoch": 7.503056768558952,
+ "grad_norm": 0.03356870263814926,
+ "learning_rate": 0.0006,
+ "loss": 5.444767951965332,
+ "step": 540
+ },
+ {
+ "epoch": 7.517030567685589,
+ "grad_norm": 0.030812304466962814,
+ "learning_rate": 0.0006,
+ "loss": 5.461244106292725,
+ "step": 541
+ },
+ {
+ "epoch": 7.531004366812227,
+ "grad_norm": 0.03319217637181282,
+ "learning_rate": 0.0006,
+ "loss": 5.502161026000977,
+ "step": 542
+ },
+ {
+ "epoch": 7.544978165938865,
+ "grad_norm": 0.032997481524944305,
+ "learning_rate": 0.0006,
+ "loss": 5.4162278175354,
+ "step": 543
+ },
+ {
+ "epoch": 7.558951965065502,
+ "grad_norm": 0.03364962339401245,
+ "learning_rate": 0.0006,
+ "loss": 5.415736198425293,
+ "step": 544
+ },
+ {
+ "epoch": 7.5729257641921395,
+ "grad_norm": 0.0344221405684948,
+ "learning_rate": 0.0006,
+ "loss": 5.541967868804932,
+ "step": 545
+ },
+ {
+ "epoch": 7.586899563318777,
+ "grad_norm": 0.029609503224492073,
+ "learning_rate": 0.0006,
+ "loss": 5.440614223480225,
+ "step": 546
+ },
+ {
+ "epoch": 7.600873362445415,
+ "grad_norm": 0.029055926948785782,
+ "learning_rate": 0.0006,
+ "loss": 5.462865352630615,
+ "step": 547
+ },
+ {
+ "epoch": 7.614847161572053,
+ "grad_norm": 0.02658848837018013,
+ "learning_rate": 0.0006,
+ "loss": 5.444467544555664,
+ "step": 548
+ },
+ {
+ "epoch": 7.62882096069869,
+ "grad_norm": 0.026276404038071632,
+ "learning_rate": 0.0006,
+ "loss": 5.462278842926025,
+ "step": 549
+ },
+ {
+ "epoch": 7.642794759825327,
+ "grad_norm": 0.0286889486014843,
+ "learning_rate": 0.0006,
+ "loss": 5.4849534034729,
+ "step": 550
+ },
+ {
+ "epoch": 7.656768558951965,
+ "grad_norm": 0.026849817484617233,
+ "learning_rate": 0.0006,
+ "loss": 5.431473731994629,
+ "step": 551
+ },
+ {
+ "epoch": 7.670742358078603,
+ "grad_norm": 0.02312396466732025,
+ "learning_rate": 0.0006,
+ "loss": 5.45462703704834,
+ "step": 552
+ },
+ {
+ "epoch": 7.68471615720524,
+ "grad_norm": 0.026197485625743866,
+ "learning_rate": 0.0006,
+ "loss": 5.452552318572998,
+ "step": 553
+ },
+ {
+ "epoch": 7.698689956331878,
+ "grad_norm": 0.02747255191206932,
+ "learning_rate": 0.0006,
+ "loss": 5.447169780731201,
+ "step": 554
+ },
+ {
+ "epoch": 7.712663755458515,
+ "grad_norm": 0.028123432770371437,
+ "learning_rate": 0.0006,
+ "loss": 5.394400596618652,
+ "step": 555
+ },
+ {
+ "epoch": 7.726637554585153,
+ "grad_norm": 0.02599870041012764,
+ "learning_rate": 0.0006,
+ "loss": 5.398341178894043,
+ "step": 556
+ },
+ {
+ "epoch": 7.74061135371179,
+ "grad_norm": 0.022171657532453537,
+ "learning_rate": 0.0006,
+ "loss": 5.368820667266846,
+ "step": 557
+ },
+ {
+ "epoch": 7.754585152838428,
+ "grad_norm": 0.022309480234980583,
+ "learning_rate": 0.0006,
+ "loss": 5.416772365570068,
+ "step": 558
+ },
+ {
+ "epoch": 7.7685589519650655,
+ "grad_norm": 0.024074165150523186,
+ "learning_rate": 0.0006,
+ "loss": 5.359002113342285,
+ "step": 559
+ },
+ {
+ "epoch": 7.782532751091703,
+ "grad_norm": 0.02636653557419777,
+ "learning_rate": 0.0006,
+ "loss": 5.484368324279785,
+ "step": 560
+ },
+ {
+ "epoch": 7.796506550218341,
+ "grad_norm": 0.02203752100467682,
+ "learning_rate": 0.0006,
+ "loss": 5.375498294830322,
+ "step": 561
+ },
+ {
+ "epoch": 7.810480349344978,
+ "grad_norm": 0.019924819469451904,
+ "learning_rate": 0.0006,
+ "loss": 5.334723472595215,
+ "step": 562
+ },
+ {
+ "epoch": 7.824454148471616,
+ "grad_norm": 0.018755966797471046,
+ "learning_rate": 0.0006,
+ "loss": 5.348781108856201,
+ "step": 563
+ },
+ {
+ "epoch": 7.8384279475982535,
+ "grad_norm": 0.016747845336794853,
+ "learning_rate": 0.0006,
+ "loss": 5.433079719543457,
+ "step": 564
+ },
+ {
+ "epoch": 7.85240174672489,
+ "grad_norm": 0.0170395877212286,
+ "learning_rate": 0.0006,
+ "loss": 5.346794128417969,
+ "step": 565
+ },
+ {
+ "epoch": 7.866375545851528,
+ "grad_norm": 0.016074176877737045,
+ "learning_rate": 0.0006,
+ "loss": 5.3277740478515625,
+ "step": 566
+ },
+ {
+ "epoch": 7.880349344978166,
+ "grad_norm": 0.014119806699454784,
+ "learning_rate": 0.0006,
+ "loss": 5.483551979064941,
+ "step": 567
+ },
+ {
+ "epoch": 7.894323144104804,
+ "grad_norm": 0.015270394273102283,
+ "learning_rate": 0.0006,
+ "loss": 5.378519535064697,
+ "step": 568
+ },
+ {
+ "epoch": 7.908296943231441,
+ "grad_norm": 0.015677539631724358,
+ "learning_rate": 0.0006,
+ "loss": 5.290210247039795,
+ "step": 569
+ },
+ {
+ "epoch": 7.922270742358078,
+ "grad_norm": 0.015930423513054848,
+ "learning_rate": 0.0006,
+ "loss": 5.427360534667969,
+ "step": 570
+ },
+ {
+ "epoch": 7.936244541484716,
+ "grad_norm": 0.016641564667224884,
+ "learning_rate": 0.0006,
+ "loss": 5.301599979400635,
+ "step": 571
+ },
+ {
+ "epoch": 7.950218340611354,
+ "grad_norm": 0.018293552100658417,
+ "learning_rate": 0.0006,
+ "loss": 5.277889251708984,
+ "step": 572
+ },
+ {
+ "epoch": 7.964192139737992,
+ "grad_norm": 0.017618799582123756,
+ "learning_rate": 0.0006,
+ "loss": 5.297194480895996,
+ "step": 573
+ },
+ {
+ "epoch": 7.978165938864628,
+ "grad_norm": 0.015034242533147335,
+ "learning_rate": 0.0006,
+ "loss": 5.42428731918335,
+ "step": 574
+ },
+ {
+ "epoch": 7.992139737991266,
+ "grad_norm": 0.01617511734366417,
+ "learning_rate": 0.0006,
+ "loss": 5.420044898986816,
+ "step": 575
+ },
+ {
+ "epoch": 8.0,
+ "grad_norm": 0.01837257295846939,
+ "learning_rate": 0.0006,
+ "loss": 5.240387916564941,
+ "step": 576
+ },
+ {
+ "epoch": 8.0,
+ "eval_loss": 5.423932075500488,
+ "eval_runtime": 58.0841,
+ "eval_samples_per_second": 42.042,
+ "eval_steps_per_second": 1.326,
+ "step": 576
+ },
+ {
+ "epoch": 8.013973799126637,
+ "grad_norm": 0.020953809842467308,
+ "learning_rate": 0.0006,
+ "loss": 5.4320783615112305,
+ "step": 577
+ },
+ {
+ "epoch": 8.027947598253276,
+ "grad_norm": 0.0227745920419693,
+ "learning_rate": 0.0006,
+ "loss": 5.310197830200195,
+ "step": 578
+ },
+ {
+ "epoch": 8.041921397379912,
+ "grad_norm": 0.021372603252530098,
+ "learning_rate": 0.0006,
+ "loss": 5.339468002319336,
+ "step": 579
+ },
+ {
+ "epoch": 8.055895196506551,
+ "grad_norm": 0.022401731461286545,
+ "learning_rate": 0.0006,
+ "loss": 5.279829978942871,
+ "step": 580
+ },
+ {
+ "epoch": 8.069868995633188,
+ "grad_norm": 0.023942379280924797,
+ "learning_rate": 0.0006,
+ "loss": 5.372323989868164,
+ "step": 581
+ },
+ {
+ "epoch": 8.083842794759825,
+ "grad_norm": 0.024380534887313843,
+ "learning_rate": 0.0006,
+ "loss": 5.253486633300781,
+ "step": 582
+ },
+ {
+ "epoch": 8.097816593886463,
+ "grad_norm": 0.03246625140309334,
+ "learning_rate": 0.0006,
+ "loss": 5.305774211883545,
+ "step": 583
+ },
+ {
+ "epoch": 8.1117903930131,
+ "grad_norm": 0.030399736016988754,
+ "learning_rate": 0.0006,
+ "loss": 5.370976448059082,
+ "step": 584
+ },
+ {
+ "epoch": 8.125764192139737,
+ "grad_norm": 0.027148913592100143,
+ "learning_rate": 0.0006,
+ "loss": 5.368010520935059,
+ "step": 585
+ },
+ {
+ "epoch": 8.139737991266376,
+ "grad_norm": 0.02925540879368782,
+ "learning_rate": 0.0006,
+ "loss": 5.303013801574707,
+ "step": 586
+ },
+ {
+ "epoch": 8.153711790393013,
+ "grad_norm": 0.02915453538298607,
+ "learning_rate": 0.0006,
+ "loss": 5.329667091369629,
+ "step": 587
+ },
+ {
+ "epoch": 8.167685589519651,
+ "grad_norm": 0.031732626259326935,
+ "learning_rate": 0.0006,
+ "loss": 5.3634934425354,
+ "step": 588
+ },
+ {
+ "epoch": 8.181659388646288,
+ "grad_norm": 0.036166731268167496,
+ "learning_rate": 0.0006,
+ "loss": 5.223980903625488,
+ "step": 589
+ },
+ {
+ "epoch": 8.195633187772925,
+ "grad_norm": 0.035654108971357346,
+ "learning_rate": 0.0006,
+ "loss": 5.361059188842773,
+ "step": 590
+ },
+ {
+ "epoch": 8.209606986899564,
+ "grad_norm": 0.0365324430167675,
+ "learning_rate": 0.0006,
+ "loss": 5.336151599884033,
+ "step": 591
+ },
+ {
+ "epoch": 8.2235807860262,
+ "grad_norm": 0.03578880429267883,
+ "learning_rate": 0.0006,
+ "loss": 5.434076309204102,
+ "step": 592
+ },
+ {
+ "epoch": 8.237554585152838,
+ "grad_norm": 0.035734012722969055,
+ "learning_rate": 0.0006,
+ "loss": 5.330893516540527,
+ "step": 593
+ },
+ {
+ "epoch": 8.251528384279476,
+ "grad_norm": 0.03196857124567032,
+ "learning_rate": 0.0006,
+ "loss": 5.306009769439697,
+ "step": 594
+ },
+ {
+ "epoch": 8.265502183406113,
+ "grad_norm": 0.032042086124420166,
+ "learning_rate": 0.0006,
+ "loss": 5.343267917633057,
+ "step": 595
+ },
+ {
+ "epoch": 8.279475982532752,
+ "grad_norm": 0.03160746395587921,
+ "learning_rate": 0.0006,
+ "loss": 5.2353715896606445,
+ "step": 596
+ },
+ {
+ "epoch": 8.293449781659389,
+ "grad_norm": 0.026689818128943443,
+ "learning_rate": 0.0006,
+ "loss": 5.254042625427246,
+ "step": 597
+ },
+ {
+ "epoch": 8.307423580786025,
+ "grad_norm": 0.02880188450217247,
+ "learning_rate": 0.0006,
+ "loss": 5.326833248138428,
+ "step": 598
+ },
+ {
+ "epoch": 8.321397379912664,
+ "grad_norm": 0.027516381815075874,
+ "learning_rate": 0.0006,
+ "loss": 5.342336177825928,
+ "step": 599
+ },
+ {
+ "epoch": 8.335371179039301,
+ "grad_norm": 0.027875030413269997,
+ "learning_rate": 0.0006,
+ "loss": 5.200719833374023,
+ "step": 600
+ },
+ {
+ "epoch": 8.34934497816594,
+ "grad_norm": 0.0268265288323164,
+ "learning_rate": 0.0006,
+ "loss": 5.249449729919434,
+ "step": 601
+ },
+ {
+ "epoch": 8.363318777292577,
+ "grad_norm": 0.024824608117341995,
+ "learning_rate": 0.0006,
+ "loss": 5.318952560424805,
+ "step": 602
+ },
+ {
+ "epoch": 8.377292576419213,
+ "grad_norm": 0.022990627214312553,
+ "learning_rate": 0.0006,
+ "loss": 5.244993209838867,
+ "step": 603
+ },
+ {
+ "epoch": 8.391266375545852,
+ "grad_norm": 0.022804604843258858,
+ "learning_rate": 0.0006,
+ "loss": 5.3236846923828125,
+ "step": 604
+ },
+ {
+ "epoch": 8.405240174672489,
+ "grad_norm": 0.02419872209429741,
+ "learning_rate": 0.0006,
+ "loss": 5.255486488342285,
+ "step": 605
+ },
+ {
+ "epoch": 8.419213973799126,
+ "grad_norm": 0.021952059119939804,
+ "learning_rate": 0.0006,
+ "loss": 5.3511552810668945,
+ "step": 606
+ },
+ {
+ "epoch": 8.433187772925764,
+ "grad_norm": 0.022375497967004776,
+ "learning_rate": 0.0006,
+ "loss": 5.294790267944336,
+ "step": 607
+ },
+ {
+ "epoch": 8.447161572052401,
+ "grad_norm": 0.019242815673351288,
+ "learning_rate": 0.0006,
+ "loss": 5.252618789672852,
+ "step": 608
+ },
+ {
+ "epoch": 8.46113537117904,
+ "grad_norm": 0.0173486340790987,
+ "learning_rate": 0.0006,
+ "loss": 5.30147647857666,
+ "step": 609
+ },
+ {
+ "epoch": 8.475109170305677,
+ "grad_norm": 0.01612604409456253,
+ "learning_rate": 0.0006,
+ "loss": 5.23960542678833,
+ "step": 610
+ },
+ {
+ "epoch": 8.489082969432314,
+ "grad_norm": 0.014293976128101349,
+ "learning_rate": 0.0006,
+ "loss": 5.119932651519775,
+ "step": 611
+ },
+ {
+ "epoch": 8.503056768558952,
+ "grad_norm": 0.014453536830842495,
+ "learning_rate": 0.0006,
+ "loss": 5.166794776916504,
+ "step": 612
+ },
+ {
+ "epoch": 8.51703056768559,
+ "grad_norm": 0.014685769565403461,
+ "learning_rate": 0.0006,
+ "loss": 5.264822959899902,
+ "step": 613
+ },
+ {
+ "epoch": 8.531004366812226,
+ "grad_norm": 0.01427740603685379,
+ "learning_rate": 0.0006,
+ "loss": 5.278433799743652,
+ "step": 614
+ },
+ {
+ "epoch": 8.544978165938865,
+ "grad_norm": 0.014864951372146606,
+ "learning_rate": 0.0006,
+ "loss": 5.214512348175049,
+ "step": 615
+ },
+ {
+ "epoch": 8.558951965065502,
+ "grad_norm": 0.01565164513885975,
+ "learning_rate": 0.0006,
+ "loss": 5.183066368103027,
+ "step": 616
+ },
+ {
+ "epoch": 8.57292576419214,
+ "grad_norm": 0.01779816672205925,
+ "learning_rate": 0.0006,
+ "loss": 5.163185119628906,
+ "step": 617
+ },
+ {
+ "epoch": 8.586899563318777,
+ "grad_norm": 0.017254827544093132,
+ "learning_rate": 0.0006,
+ "loss": 5.243555068969727,
+ "step": 618
+ },
+ {
+ "epoch": 8.600873362445414,
+ "grad_norm": 0.01650584116578102,
+ "learning_rate": 0.0006,
+ "loss": 5.234884738922119,
+ "step": 619
+ },
+ {
+ "epoch": 8.614847161572053,
+ "grad_norm": 0.017021115869283676,
+ "learning_rate": 0.0006,
+ "loss": 5.12271785736084,
+ "step": 620
+ },
+ {
+ "epoch": 8.62882096069869,
+ "grad_norm": 0.01773759163916111,
+ "learning_rate": 0.0006,
+ "loss": 5.242153167724609,
+ "step": 621
+ },
+ {
+ "epoch": 8.642794759825328,
+ "grad_norm": 0.015679042786359787,
+ "learning_rate": 0.0006,
+ "loss": 5.170779228210449,
+ "step": 622
+ },
+ {
+ "epoch": 8.656768558951965,
+ "grad_norm": 0.013760549947619438,
+ "learning_rate": 0.0006,
+ "loss": 5.238644599914551,
+ "step": 623
+ },
+ {
+ "epoch": 8.670742358078602,
+ "grad_norm": 0.014571522362530231,
+ "learning_rate": 0.0006,
+ "loss": 5.177056312561035,
+ "step": 624
+ },
+ {
+ "epoch": 8.68471615720524,
+ "grad_norm": 0.016209015622735023,
+ "learning_rate": 0.0006,
+ "loss": 5.163750648498535,
+ "step": 625
+ },
+ {
+ "epoch": 8.698689956331878,
+ "grad_norm": 0.016813941299915314,
+ "learning_rate": 0.0006,
+ "loss": 5.183428764343262,
+ "step": 626
+ },
+ {
+ "epoch": 8.712663755458514,
+ "grad_norm": 0.020985357463359833,
+ "learning_rate": 0.0006,
+ "loss": 5.229465007781982,
+ "step": 627
+ },
+ {
+ "epoch": 8.726637554585153,
+ "grad_norm": 0.02679632417857647,
+ "learning_rate": 0.0006,
+ "loss": 5.257368087768555,
+ "step": 628
+ },
+ {
+ "epoch": 8.74061135371179,
+ "grad_norm": 0.025756070390343666,
+ "learning_rate": 0.0006,
+ "loss": 5.253736972808838,
+ "step": 629
+ },
+ {
+ "epoch": 8.754585152838429,
+ "grad_norm": 0.02643490768969059,
+ "learning_rate": 0.0006,
+ "loss": 5.228633403778076,
+ "step": 630
+ },
+ {
+ "epoch": 8.768558951965066,
+ "grad_norm": 0.03303210437297821,
+ "learning_rate": 0.0006,
+ "loss": 5.163540363311768,
+ "step": 631
+ },
+ {
+ "epoch": 8.782532751091702,
+ "grad_norm": 0.030432431027293205,
+ "learning_rate": 0.0006,
+ "loss": 5.243169784545898,
+ "step": 632
+ },
+ {
+ "epoch": 8.796506550218341,
+ "grad_norm": 0.028631288558244705,
+ "learning_rate": 0.0006,
+ "loss": 5.166398048400879,
+ "step": 633
+ },
+ {
+ "epoch": 8.810480349344978,
+ "grad_norm": 0.026188310235738754,
+ "learning_rate": 0.0006,
+ "loss": 5.128122329711914,
+ "step": 634
+ },
+ {
+ "epoch": 8.824454148471617,
+ "grad_norm": 0.025308528915047646,
+ "learning_rate": 0.0006,
+ "loss": 5.211284637451172,
+ "step": 635
+ },
+ {
+ "epoch": 8.838427947598253,
+ "grad_norm": 0.020657729357481003,
+ "learning_rate": 0.0006,
+ "loss": 5.161575794219971,
+ "step": 636
+ },
+ {
+ "epoch": 8.85240174672489,
+ "grad_norm": 0.021901234984397888,
+ "learning_rate": 0.0006,
+ "loss": 5.20050048828125,
+ "step": 637
+ },
+ {
+ "epoch": 8.866375545851529,
+ "grad_norm": 0.021219318732619286,
+ "learning_rate": 0.0006,
+ "loss": 5.211699485778809,
+ "step": 638
+ },
+ {
+ "epoch": 8.880349344978166,
+ "grad_norm": 0.0200974028557539,
+ "learning_rate": 0.0006,
+ "loss": 5.092503547668457,
+ "step": 639
+ },
+ {
+ "epoch": 8.894323144104803,
+ "grad_norm": 0.023804882541298866,
+ "learning_rate": 0.0006,
+ "loss": 5.216068267822266,
+ "step": 640
+ },
+ {
+ "epoch": 8.908296943231441,
+ "grad_norm": 0.026088010519742966,
+ "learning_rate": 0.0006,
+ "loss": 5.155592918395996,
+ "step": 641
+ },
+ {
+ "epoch": 8.922270742358078,
+ "grad_norm": 0.02601276896893978,
+ "learning_rate": 0.0006,
+ "loss": 5.197238922119141,
+ "step": 642
+ },
+ {
+ "epoch": 8.936244541484717,
+ "grad_norm": 0.020387211814522743,
+ "learning_rate": 0.0006,
+ "loss": 5.1955437660217285,
+ "step": 643
+ },
+ {
+ "epoch": 8.950218340611354,
+ "grad_norm": 0.019214622676372528,
+ "learning_rate": 0.0006,
+ "loss": 5.206346035003662,
+ "step": 644
+ },
+ {
+ "epoch": 8.96419213973799,
+ "grad_norm": 0.019674314185976982,
+ "learning_rate": 0.0006,
+ "loss": 5.219857215881348,
+ "step": 645
+ },
+ {
+ "epoch": 8.97816593886463,
+ "grad_norm": 0.020318234339356422,
+ "learning_rate": 0.0006,
+ "loss": 5.191132545471191,
+ "step": 646
+ },
+ {
+ "epoch": 8.992139737991266,
+ "grad_norm": 0.021428676322102547,
+ "learning_rate": 0.0006,
+ "loss": 5.214695930480957,
+ "step": 647
+ },
+ {
+ "epoch": 9.0,
+ "grad_norm": 0.023235054686665535,
+ "learning_rate": 0.0006,
+ "loss": 5.229091644287109,
+ "step": 648
+ },
+ {
+ "epoch": 9.0,
+ "eval_loss": 5.289593696594238,
+ "eval_runtime": 56.9256,
+ "eval_samples_per_second": 42.898,
+ "eval_steps_per_second": 1.353,
+ "step": 648
+ },
+ {
+ "epoch": 9.013973799126637,
+ "grad_norm": 0.02320289984345436,
+ "learning_rate": 0.0006,
+ "loss": 4.983669281005859,
+ "step": 649
+ },
+ {
+ "epoch": 9.027947598253276,
+ "grad_norm": 0.018521282821893692,
+ "learning_rate": 0.0006,
+ "loss": 5.200575828552246,
+ "step": 650
+ },
+ {
+ "epoch": 9.041921397379912,
+ "grad_norm": 0.0180169939994812,
+ "learning_rate": 0.0006,
+ "loss": 5.148033618927002,
+ "step": 651
+ },
+ {
+ "epoch": 9.055895196506551,
+ "grad_norm": 0.019969860091805458,
+ "learning_rate": 0.0006,
+ "loss": 5.099125385284424,
+ "step": 652
+ },
+ {
+ "epoch": 9.069868995633188,
+ "grad_norm": 0.01643305830657482,
+ "learning_rate": 0.0006,
+ "loss": 5.092347145080566,
+ "step": 653
+ },
+ {
+ "epoch": 9.083842794759825,
+ "grad_norm": 0.016022363677620888,
+ "learning_rate": 0.0006,
+ "loss": 5.129279613494873,
+ "step": 654
+ },
+ {
+ "epoch": 9.097816593886463,
+ "grad_norm": 0.01657041162252426,
+ "learning_rate": 0.0006,
+ "loss": 5.159562587738037,
+ "step": 655
+ },
+ {
+ "epoch": 9.1117903930131,
+ "grad_norm": 0.017924228683114052,
+ "learning_rate": 0.0006,
+ "loss": 5.165497779846191,
+ "step": 656
+ },
+ {
+ "epoch": 9.125764192139737,
+ "grad_norm": 0.021592361852526665,
+ "learning_rate": 0.0006,
+ "loss": 5.204775810241699,
+ "step": 657
+ },
+ {
+ "epoch": 9.139737991266376,
+ "grad_norm": 0.026924120262265205,
+ "learning_rate": 0.0006,
+ "loss": 5.128296852111816,
+ "step": 658
+ },
+ {
+ "epoch": 9.153711790393013,
+ "grad_norm": 0.023156899958848953,
+ "learning_rate": 0.0006,
+ "loss": 5.1255340576171875,
+ "step": 659
+ },
+ {
+ "epoch": 9.167685589519651,
+ "grad_norm": 0.020670367404818535,
+ "learning_rate": 0.0006,
+ "loss": 5.128479957580566,
+ "step": 660
+ },
+ {
+ "epoch": 9.181659388646288,
+ "grad_norm": 0.025447173044085503,
+ "learning_rate": 0.0006,
+ "loss": 5.154821872711182,
+ "step": 661
+ },
+ {
+ "epoch": 9.195633187772925,
+ "grad_norm": 0.027940068393945694,
+ "learning_rate": 0.0006,
+ "loss": 5.143466472625732,
+ "step": 662
+ },
+ {
+ "epoch": 9.209606986899564,
+ "grad_norm": 0.026847844943404198,
+ "learning_rate": 0.0006,
+ "loss": 5.045665740966797,
+ "step": 663
+ },
+ {
+ "epoch": 9.2235807860262,
+ "grad_norm": 0.02340601570904255,
+ "learning_rate": 0.0006,
+ "loss": 5.21933126449585,
+ "step": 664
+ },
+ {
+ "epoch": 9.237554585152838,
+ "grad_norm": 0.02340371161699295,
+ "learning_rate": 0.0006,
+ "loss": 4.942949295043945,
+ "step": 665
+ },
+ {
+ "epoch": 9.251528384279476,
+ "grad_norm": 0.02221992425620556,
+ "learning_rate": 0.0006,
+ "loss": 5.197099208831787,
+ "step": 666
+ },
+ {
+ "epoch": 9.265502183406113,
+ "grad_norm": 0.023508677259087563,
+ "learning_rate": 0.0006,
+ "loss": 5.022897720336914,
+ "step": 667
+ },
+ {
+ "epoch": 9.279475982532752,
+ "grad_norm": 0.026120394468307495,
+ "learning_rate": 0.0006,
+ "loss": 5.087725639343262,
+ "step": 668
+ },
+ {
+ "epoch": 9.293449781659389,
+ "grad_norm": 0.027273228392004967,
+ "learning_rate": 0.0006,
+ "loss": 5.090963840484619,
+ "step": 669
+ },
+ {
+ "epoch": 9.307423580786025,
+ "grad_norm": 0.03241586685180664,
+ "learning_rate": 0.0006,
+ "loss": 5.047842979431152,
+ "step": 670
+ },
+ {
+ "epoch": 9.321397379912664,
+ "grad_norm": 0.030245667323470116,
+ "learning_rate": 0.0006,
+ "loss": 5.104803085327148,
+ "step": 671
+ },
+ {
+ "epoch": 9.335371179039301,
+ "grad_norm": 0.027698364108800888,
+ "learning_rate": 0.0006,
+ "loss": 5.0644097328186035,
+ "step": 672
+ },
+ {
+ "epoch": 9.34934497816594,
+ "grad_norm": 0.029692554846405983,
+ "learning_rate": 0.0006,
+ "loss": 5.111942291259766,
+ "step": 673
+ },
+ {
+ "epoch": 9.363318777292577,
+ "grad_norm": 0.03328656405210495,
+ "learning_rate": 0.0006,
+ "loss": 5.1574554443359375,
+ "step": 674
+ },
+ {
+ "epoch": 9.377292576419213,
+ "grad_norm": 0.031938180327415466,
+ "learning_rate": 0.0006,
+ "loss": 5.190371990203857,
+ "step": 675
+ },
+ {
+ "epoch": 9.391266375545852,
+ "grad_norm": 0.024858945980668068,
+ "learning_rate": 0.0006,
+ "loss": 5.121420860290527,
+ "step": 676
+ },
+ {
+ "epoch": 9.405240174672489,
+ "grad_norm": 0.023033203557133675,
+ "learning_rate": 0.0006,
+ "loss": 5.0974345207214355,
+ "step": 677
+ },
+ {
+ "epoch": 9.419213973799126,
+ "grad_norm": 0.022083545103669167,
+ "learning_rate": 0.0006,
+ "loss": 5.046319961547852,
+ "step": 678
+ },
+ {
+ "epoch": 9.433187772925764,
+ "grad_norm": 0.0228437427431345,
+ "learning_rate": 0.0006,
+ "loss": 5.119963645935059,
+ "step": 679
+ },
+ {
+ "epoch": 9.447161572052401,
+ "grad_norm": 0.0242351982742548,
+ "learning_rate": 0.0006,
+ "loss": 5.0444560050964355,
+ "step": 680
+ },
+ {
+ "epoch": 9.46113537117904,
+ "grad_norm": 0.02401185780763626,
+ "learning_rate": 0.0006,
+ "loss": 5.063238620758057,
+ "step": 681
+ },
+ {
+ "epoch": 9.475109170305677,
+ "grad_norm": 0.02560959942638874,
+ "learning_rate": 0.0006,
+ "loss": 5.035224914550781,
+ "step": 682
+ },
+ {
+ "epoch": 9.489082969432314,
+ "grad_norm": 0.024985626339912415,
+ "learning_rate": 0.0006,
+ "loss": 5.049284934997559,
+ "step": 683
+ },
+ {
+ "epoch": 9.503056768558952,
+ "grad_norm": 0.023178115487098694,
+ "learning_rate": 0.0006,
+ "loss": 5.2266621589660645,
+ "step": 684
+ },
+ {
+ "epoch": 9.51703056768559,
+ "grad_norm": 0.0215512253344059,
+ "learning_rate": 0.0006,
+ "loss": 5.1875715255737305,
+ "step": 685
+ },
+ {
+ "epoch": 9.531004366812226,
+ "grad_norm": 0.01662837713956833,
+ "learning_rate": 0.0006,
+ "loss": 5.137012481689453,
+ "step": 686
+ },
+ {
+ "epoch": 9.544978165938865,
+ "grad_norm": 0.01952764391899109,
+ "learning_rate": 0.0006,
+ "loss": 5.072500228881836,
+ "step": 687
+ },
+ {
+ "epoch": 9.558951965065502,
+ "grad_norm": 0.017521077767014503,
+ "learning_rate": 0.0006,
+ "loss": 5.045711994171143,
+ "step": 688
+ },
+ {
+ "epoch": 9.57292576419214,
+ "grad_norm": 0.017355265095829964,
+ "learning_rate": 0.0006,
+ "loss": 5.02881383895874,
+ "step": 689
+ },
+ {
+ "epoch": 9.586899563318777,
+ "grad_norm": 0.015587719157338142,
+ "learning_rate": 0.0006,
+ "loss": 5.132787704467773,
+ "step": 690
+ },
+ {
+ "epoch": 9.600873362445414,
+ "grad_norm": 0.017469845712184906,
+ "learning_rate": 0.0006,
+ "loss": 5.105447292327881,
+ "step": 691
+ },
+ {
+ "epoch": 9.614847161572053,
+ "grad_norm": 0.019630100578069687,
+ "learning_rate": 0.0006,
+ "loss": 5.032464981079102,
+ "step": 692
+ },
+ {
+ "epoch": 9.62882096069869,
+ "grad_norm": 0.01784994639456272,
+ "learning_rate": 0.0006,
+ "loss": 5.081345558166504,
+ "step": 693
+ },
+ {
+ "epoch": 9.642794759825328,
+ "grad_norm": 0.01872754842042923,
+ "learning_rate": 0.0006,
+ "loss": 4.9772114753723145,
+ "step": 694
+ },
+ {
+ "epoch": 9.656768558951965,
+ "grad_norm": 0.01999032311141491,
+ "learning_rate": 0.0006,
+ "loss": 5.089740753173828,
+ "step": 695
+ },
+ {
+ "epoch": 9.670742358078602,
+ "grad_norm": 0.01710767112672329,
+ "learning_rate": 0.0006,
+ "loss": 4.9869890213012695,
+ "step": 696
+ },
+ {
+ "epoch": 9.68471615720524,
+ "grad_norm": 0.015144161880016327,
+ "learning_rate": 0.0006,
+ "loss": 5.022564888000488,
+ "step": 697
+ },
+ {
+ "epoch": 9.698689956331878,
+ "grad_norm": 0.01618543453514576,
+ "learning_rate": 0.0006,
+ "loss": 5.035921573638916,
+ "step": 698
+ },
+ {
+ "epoch": 9.712663755458514,
+ "grad_norm": 0.014653387479484081,
+ "learning_rate": 0.0006,
+ "loss": 4.914456367492676,
+ "step": 699
+ },
+ {
+ "epoch": 9.726637554585153,
+ "grad_norm": 0.013434977270662785,
+ "learning_rate": 0.0006,
+ "loss": 5.063764572143555,
+ "step": 700
+ },
+ {
+ "epoch": 9.74061135371179,
+ "grad_norm": 0.015128864906728268,
+ "learning_rate": 0.0006,
+ "loss": 5.052367687225342,
+ "step": 701
+ },
+ {
+ "epoch": 9.754585152838429,
+ "grad_norm": 0.014210768043994904,
+ "learning_rate": 0.0006,
+ "loss": 5.016489028930664,
+ "step": 702
+ },
+ {
+ "epoch": 9.768558951965066,
+ "grad_norm": 0.015191680751740932,
+ "learning_rate": 0.0006,
+ "loss": 4.990176200866699,
+ "step": 703
+ },
+ {
+ "epoch": 9.782532751091702,
+ "grad_norm": 0.018285127356648445,
+ "learning_rate": 0.0006,
+ "loss": 5.015164375305176,
+ "step": 704
+ },
+ {
+ "epoch": 9.796506550218341,
+ "grad_norm": 0.0204079058021307,
+ "learning_rate": 0.0006,
+ "loss": 5.091142654418945,
+ "step": 705
+ },
+ {
+ "epoch": 9.810480349344978,
+ "grad_norm": 0.01980586163699627,
+ "learning_rate": 0.0006,
+ "loss": 5.03350830078125,
+ "step": 706
+ },
+ {
+ "epoch": 9.824454148471617,
+ "grad_norm": 0.02048729732632637,
+ "learning_rate": 0.0006,
+ "loss": 5.03611421585083,
+ "step": 707
+ },
+ {
+ "epoch": 9.838427947598253,
+ "grad_norm": 0.02058163844048977,
+ "learning_rate": 0.0006,
+ "loss": 4.9613447189331055,
+ "step": 708
+ },
+ {
+ "epoch": 9.85240174672489,
+ "grad_norm": 0.018004924058914185,
+ "learning_rate": 0.0006,
+ "loss": 4.966541767120361,
+ "step": 709
+ },
+ {
+ "epoch": 9.866375545851529,
+ "grad_norm": 0.0198152307420969,
+ "learning_rate": 0.0006,
+ "loss": 5.01750373840332,
+ "step": 710
+ },
+ {
+ "epoch": 9.880349344978166,
+ "grad_norm": 0.018532969057559967,
+ "learning_rate": 0.0006,
+ "loss": 5.097424507141113,
+ "step": 711
+ },
+ {
+ "epoch": 9.894323144104803,
+ "grad_norm": 0.018939971923828125,
+ "learning_rate": 0.0006,
+ "loss": 5.177223205566406,
+ "step": 712
+ },
+ {
+ "epoch": 9.908296943231441,
+ "grad_norm": 0.017969170585274696,
+ "learning_rate": 0.0006,
+ "loss": 5.118013858795166,
+ "step": 713
+ },
+ {
+ "epoch": 9.922270742358078,
+ "grad_norm": 0.018021270632743835,
+ "learning_rate": 0.0006,
+ "loss": 5.034787178039551,
+ "step": 714
+ },
+ {
+ "epoch": 9.936244541484717,
+ "grad_norm": 0.017009710893034935,
+ "learning_rate": 0.0006,
+ "loss": 5.003920555114746,
+ "step": 715
+ },
+ {
+ "epoch": 9.950218340611354,
+ "grad_norm": 0.017722049728035927,
+ "learning_rate": 0.0006,
+ "loss": 5.031939506530762,
+ "step": 716
+ },
+ {
+ "epoch": 9.96419213973799,
+ "grad_norm": 0.017134087160229683,
+ "learning_rate": 0.0006,
+ "loss": 5.093474388122559,
+ "step": 717
+ },
+ {
+ "epoch": 9.97816593886463,
+ "grad_norm": 0.018511613830924034,
+ "learning_rate": 0.0006,
+ "loss": 5.078604698181152,
+ "step": 718
+ },
+ {
+ "epoch": 9.992139737991266,
+ "grad_norm": 0.021206334233283997,
+ "learning_rate": 0.0006,
+ "loss": 5.052572727203369,
+ "step": 719
+ },
+ {
+ "epoch": 10.0,
+ "grad_norm": 0.02239062264561653,
+ "learning_rate": 0.0006,
+ "loss": 5.063547134399414,
+ "step": 720
+ },
+ {
+ "epoch": 10.0,
+ "eval_loss": 5.173044681549072,
+ "eval_runtime": 56.7287,
+ "eval_samples_per_second": 43.047,
+ "eval_steps_per_second": 1.357,
+ "step": 720
+ },
+ {
+ "epoch": 10.013973799126637,
+ "grad_norm": 0.020757969468832016,
+ "learning_rate": 0.0006,
+ "loss": 5.045648574829102,
+ "step": 721
+ },
+ {
+ "epoch": 10.027947598253276,
+ "grad_norm": 0.022697702050209045,
+ "learning_rate": 0.0006,
+ "loss": 4.993833065032959,
+ "step": 722
+ },
+ {
+ "epoch": 10.041921397379912,
+ "grad_norm": 0.02434331737458706,
+ "learning_rate": 0.0006,
+ "loss": 5.073219299316406,
+ "step": 723
+ },
+ {
+ "epoch": 10.055895196506551,
+ "grad_norm": 0.025499660521745682,
+ "learning_rate": 0.0006,
+ "loss": 5.019072532653809,
+ "step": 724
+ },
+ {
+ "epoch": 10.069868995633188,
+ "grad_norm": 0.029281053692102432,
+ "learning_rate": 0.0006,
+ "loss": 4.959851264953613,
+ "step": 725
+ },
+ {
+ "epoch": 10.083842794759825,
+ "grad_norm": 0.033878110349178314,
+ "learning_rate": 0.0006,
+ "loss": 5.132717609405518,
+ "step": 726
+ },
+ {
+ "epoch": 10.097816593886463,
+ "grad_norm": 0.03385764732956886,
+ "learning_rate": 0.0006,
+ "loss": 5.0482892990112305,
+ "step": 727
+ },
+ {
+ "epoch": 10.1117903930131,
+ "grad_norm": 0.03466648980975151,
+ "learning_rate": 0.0006,
+ "loss": 4.946855545043945,
+ "step": 728
+ },
+ {
+ "epoch": 10.125764192139737,
+ "grad_norm": 0.028546305373311043,
+ "learning_rate": 0.0006,
+ "loss": 5.027774333953857,
+ "step": 729
+ },
+ {
+ "epoch": 10.139737991266376,
+ "grad_norm": 0.030575869604945183,
+ "learning_rate": 0.0006,
+ "loss": 4.9768476486206055,
+ "step": 730
+ },
+ {
+ "epoch": 10.153711790393013,
+ "grad_norm": 0.03614303469657898,
+ "learning_rate": 0.0006,
+ "loss": 5.078113555908203,
+ "step": 731
+ },
+ {
+ "epoch": 10.167685589519651,
+ "grad_norm": 0.040991660207509995,
+ "learning_rate": 0.0006,
+ "loss": 5.033185958862305,
+ "step": 732
+ },
+ {
+ "epoch": 10.181659388646288,
+ "grad_norm": 0.04551699757575989,
+ "learning_rate": 0.0006,
+ "loss": 5.0820746421813965,
+ "step": 733
+ },
+ {
+ "epoch": 10.195633187772925,
+ "grad_norm": 0.040193989872932434,
+ "learning_rate": 0.0006,
+ "loss": 5.059700965881348,
+ "step": 734
+ },
+ {
+ "epoch": 10.209606986899564,
+ "grad_norm": 0.035851605236530304,
+ "learning_rate": 0.0006,
+ "loss": 5.142467498779297,
+ "step": 735
+ },
+ {
+ "epoch": 10.2235807860262,
+ "grad_norm": 0.034200169146060944,
+ "learning_rate": 0.0006,
+ "loss": 5.072312355041504,
+ "step": 736
+ },
+ {
+ "epoch": 10.237554585152838,
+ "grad_norm": 0.033621110022068024,
+ "learning_rate": 0.0006,
+ "loss": 5.08037805557251,
+ "step": 737
+ },
+ {
+ "epoch": 10.251528384279476,
+ "grad_norm": 0.037372369319200516,
+ "learning_rate": 0.0006,
+ "loss": 4.998425483703613,
+ "step": 738
+ },
+ {
+ "epoch": 10.265502183406113,
+ "grad_norm": 0.037044707685709,
+ "learning_rate": 0.0006,
+ "loss": 4.923993110656738,
+ "step": 739
+ },
+ {
+ "epoch": 10.279475982532752,
+ "grad_norm": 0.030454808846116066,
+ "learning_rate": 0.0006,
+ "loss": 5.0889997482299805,
+ "step": 740
+ },
+ {
+ "epoch": 10.293449781659389,
+ "grad_norm": 0.030969638377428055,
+ "learning_rate": 0.0006,
+ "loss": 4.984112739562988,
+ "step": 741
+ },
+ {
+ "epoch": 10.307423580786025,
+ "grad_norm": 0.0276983380317688,
+ "learning_rate": 0.0006,
+ "loss": 4.955412864685059,
+ "step": 742
+ },
+ {
+ "epoch": 10.321397379912664,
+ "grad_norm": 0.02716052532196045,
+ "learning_rate": 0.0006,
+ "loss": 5.0098876953125,
+ "step": 743
+ },
+ {
+ "epoch": 10.335371179039301,
+ "grad_norm": 0.02403552085161209,
+ "learning_rate": 0.0006,
+ "loss": 5.165160655975342,
+ "step": 744
+ },
+ {
+ "epoch": 10.34934497816594,
+ "grad_norm": 0.0257862601429224,
+ "learning_rate": 0.0006,
+ "loss": 5.001547813415527,
+ "step": 745
+ },
+ {
+ "epoch": 10.363318777292577,
+ "grad_norm": 0.02300378680229187,
+ "learning_rate": 0.0006,
+ "loss": 5.038888931274414,
+ "step": 746
+ },
+ {
+ "epoch": 10.377292576419213,
+ "grad_norm": 0.019766854122281075,
+ "learning_rate": 0.0006,
+ "loss": 4.924291610717773,
+ "step": 747
+ },
+ {
+ "epoch": 10.391266375545852,
+ "grad_norm": 0.01888395845890045,
+ "learning_rate": 0.0006,
+ "loss": 5.059971332550049,
+ "step": 748
+ },
+ {
+ "epoch": 10.405240174672489,
+ "grad_norm": 0.015577499754726887,
+ "learning_rate": 0.0006,
+ "loss": 4.987686634063721,
+ "step": 749
+ },
+ {
+ "epoch": 10.419213973799126,
+ "grad_norm": 0.016012346372008324,
+ "learning_rate": 0.0006,
+ "loss": 5.015157699584961,
+ "step": 750
+ },
+ {
+ "epoch": 10.433187772925764,
+ "grad_norm": 0.016277998685836792,
+ "learning_rate": 0.0006,
+ "loss": 5.004931449890137,
+ "step": 751
+ },
+ {
+ "epoch": 10.447161572052401,
+ "grad_norm": 0.017504561692476273,
+ "learning_rate": 0.0006,
+ "loss": 4.902827739715576,
+ "step": 752
+ },
+ {
+ "epoch": 10.46113537117904,
+ "grad_norm": 0.01569022797048092,
+ "learning_rate": 0.0006,
+ "loss": 4.954863548278809,
+ "step": 753
+ },
+ {
+ "epoch": 10.475109170305677,
+ "grad_norm": 0.014114422723650932,
+ "learning_rate": 0.0006,
+ "loss": 5.00656795501709,
+ "step": 754
+ },
+ {
+ "epoch": 10.489082969432314,
+ "grad_norm": 0.01573832333087921,
+ "learning_rate": 0.0006,
+ "loss": 4.978353023529053,
+ "step": 755
+ },
+ {
+ "epoch": 10.503056768558952,
+ "grad_norm": 0.016480036079883575,
+ "learning_rate": 0.0006,
+ "loss": 4.953327655792236,
+ "step": 756
+ },
+ {
+ "epoch": 10.51703056768559,
+ "grad_norm": 0.014008025638759136,
+ "learning_rate": 0.0006,
+ "loss": 5.050840377807617,
+ "step": 757
+ },
+ {
+ "epoch": 10.531004366812226,
+ "grad_norm": 0.013147998601198196,
+ "learning_rate": 0.0006,
+ "loss": 4.974964141845703,
+ "step": 758
+ },
+ {
+ "epoch": 10.544978165938865,
+ "grad_norm": 0.013807603158056736,
+ "learning_rate": 0.0006,
+ "loss": 4.927907943725586,
+ "step": 759
+ },
+ {
+ "epoch": 10.558951965065502,
+ "grad_norm": 0.013555224984884262,
+ "learning_rate": 0.0006,
+ "loss": 4.980690956115723,
+ "step": 760
+ },
+ {
+ "epoch": 10.57292576419214,
+ "grad_norm": 0.013045408762991428,
+ "learning_rate": 0.0006,
+ "loss": 4.9436116218566895,
+ "step": 761
+ },
+ {
+ "epoch": 10.586899563318777,
+ "grad_norm": 0.012858862057328224,
+ "learning_rate": 0.0006,
+ "loss": 4.9595770835876465,
+ "step": 762
+ },
+ {
+ "epoch": 10.600873362445414,
+ "grad_norm": 0.01468253880739212,
+ "learning_rate": 0.0006,
+ "loss": 4.90725040435791,
+ "step": 763
+ },
+ {
+ "epoch": 10.614847161572053,
+ "grad_norm": 0.013224679045379162,
+ "learning_rate": 0.0006,
+ "loss": 4.877615928649902,
+ "step": 764
+ },
+ {
+ "epoch": 10.62882096069869,
+ "grad_norm": 0.013555348850786686,
+ "learning_rate": 0.0006,
+ "loss": 4.9456610679626465,
+ "step": 765
+ },
+ {
+ "epoch": 10.642794759825328,
+ "grad_norm": 0.014387160539627075,
+ "learning_rate": 0.0006,
+ "loss": 4.9233717918396,
+ "step": 766
+ },
+ {
+ "epoch": 10.656768558951965,
+ "grad_norm": 0.014354088343679905,
+ "learning_rate": 0.0006,
+ "loss": 4.983511924743652,
+ "step": 767
+ },
+ {
+ "epoch": 10.670742358078602,
+ "grad_norm": 0.01332700252532959,
+ "learning_rate": 0.0006,
+ "loss": 4.934416770935059,
+ "step": 768
+ },
+ {
+ "epoch": 10.68471615720524,
+ "grad_norm": 0.011543313041329384,
+ "learning_rate": 0.0006,
+ "loss": 4.96388053894043,
+ "step": 769
+ },
+ {
+ "epoch": 10.698689956331878,
+ "grad_norm": 0.012330878525972366,
+ "learning_rate": 0.0006,
+ "loss": 4.946396350860596,
+ "step": 770
+ },
+ {
+ "epoch": 10.712663755458514,
+ "grad_norm": 0.01379779726266861,
+ "learning_rate": 0.0006,
+ "loss": 4.981019020080566,
+ "step": 771
+ },
+ {
+ "epoch": 10.726637554585153,
+ "grad_norm": 0.014612431637942791,
+ "learning_rate": 0.0006,
+ "loss": 4.894567966461182,
+ "step": 772
+ },
+ {
+ "epoch": 10.74061135371179,
+ "grad_norm": 0.013062899932265282,
+ "learning_rate": 0.0006,
+ "loss": 5.033038139343262,
+ "step": 773
+ },
+ {
+ "epoch": 10.754585152838429,
+ "grad_norm": 0.013898049481213093,
+ "learning_rate": 0.0006,
+ "loss": 4.823636054992676,
+ "step": 774
+ },
+ {
+ "epoch": 10.768558951965066,
+ "grad_norm": 0.01502019353210926,
+ "learning_rate": 0.0006,
+ "loss": 5.020156383514404,
+ "step": 775
+ },
+ {
+ "epoch": 10.782532751091702,
+ "grad_norm": 0.013810204342007637,
+ "learning_rate": 0.0006,
+ "loss": 4.935647010803223,
+ "step": 776
+ },
+ {
+ "epoch": 10.796506550218341,
+ "grad_norm": 0.013616513460874557,
+ "learning_rate": 0.0006,
+ "loss": 4.901486396789551,
+ "step": 777
+ },
+ {
+ "epoch": 10.810480349344978,
+ "grad_norm": 0.013601024635136127,
+ "learning_rate": 0.0006,
+ "loss": 4.931873321533203,
+ "step": 778
+ },
+ {
+ "epoch": 10.824454148471617,
+ "grad_norm": 0.013668350875377655,
+ "learning_rate": 0.0006,
+ "loss": 4.937256813049316,
+ "step": 779
+ },
+ {
+ "epoch": 10.838427947598253,
+ "grad_norm": 0.01581823080778122,
+ "learning_rate": 0.0006,
+ "loss": 4.9611406326293945,
+ "step": 780
+ },
+ {
+ "epoch": 10.85240174672489,
+ "grad_norm": 0.019478484988212585,
+ "learning_rate": 0.0006,
+ "loss": 4.9212236404418945,
+ "step": 781
+ },
+ {
+ "epoch": 10.866375545851529,
+ "grad_norm": 0.026921333745121956,
+ "learning_rate": 0.0006,
+ "loss": 4.88501501083374,
+ "step": 782
+ },
+ {
+ "epoch": 10.880349344978166,
+ "grad_norm": 0.03155914694070816,
+ "learning_rate": 0.0006,
+ "loss": 4.989967346191406,
+ "step": 783
+ },
+ {
+ "epoch": 10.894323144104803,
+ "grad_norm": 0.029689429327845573,
+ "learning_rate": 0.0006,
+ "loss": 4.936000823974609,
+ "step": 784
+ },
+ {
+ "epoch": 10.908296943231441,
+ "grad_norm": 0.026626063510775566,
+ "learning_rate": 0.0006,
+ "loss": 4.984785079956055,
+ "step": 785
+ },
+ {
+ "epoch": 10.922270742358078,
+ "grad_norm": 0.026168212294578552,
+ "learning_rate": 0.0006,
+ "loss": 4.908282279968262,
+ "step": 786
+ },
+ {
+ "epoch": 10.936244541484717,
+ "grad_norm": 0.02828158251941204,
+ "learning_rate": 0.0006,
+ "loss": 4.926390647888184,
+ "step": 787
+ },
+ {
+ "epoch": 10.950218340611354,
+ "grad_norm": 0.02649078331887722,
+ "learning_rate": 0.0006,
+ "loss": 5.00400972366333,
+ "step": 788
+ },
+ {
+ "epoch": 10.96419213973799,
+ "grad_norm": 0.0256856270134449,
+ "learning_rate": 0.0006,
+ "loss": 5.0317487716674805,
+ "step": 789
+ },
+ {
+ "epoch": 10.97816593886463,
+ "grad_norm": 0.024728883057832718,
+ "learning_rate": 0.0006,
+ "loss": 4.9299635887146,
+ "step": 790
+ },
+ {
+ "epoch": 10.992139737991266,
+ "grad_norm": 0.026241116225719452,
+ "learning_rate": 0.0006,
+ "loss": 4.947319984436035,
+ "step": 791
+ },
+ {
+ "epoch": 11.0,
+ "grad_norm": 0.023526743054389954,
+ "learning_rate": 0.0006,
+ "loss": 4.94429349899292,
+ "step": 792
+ },
+ {
+ "epoch": 11.0,
+ "eval_loss": 5.104345798492432,
+ "eval_runtime": 57.0955,
+ "eval_samples_per_second": 42.77,
+ "eval_steps_per_second": 1.349,
+ "step": 792
+ },
+ {
+ "epoch": 11.013973799126637,
+ "grad_norm": 0.0216788612306118,
+ "learning_rate": 0.0006,
+ "loss": 4.881950855255127,
+ "step": 793
+ },
+ {
+ "epoch": 11.027947598253276,
+ "grad_norm": 0.01830463856458664,
+ "learning_rate": 0.0006,
+ "loss": 4.971016883850098,
+ "step": 794
+ },
+ {
+ "epoch": 11.041921397379912,
+ "grad_norm": 0.02204521745443344,
+ "learning_rate": 0.0006,
+ "loss": 4.894623756408691,
+ "step": 795
+ },
+ {
+ "epoch": 11.055895196506551,
+ "grad_norm": 0.0210255216807127,
+ "learning_rate": 0.0006,
+ "loss": 4.834009170532227,
+ "step": 796
+ },
+ {
+ "epoch": 11.069868995633188,
+ "grad_norm": 0.019797448068857193,
+ "learning_rate": 0.0006,
+ "loss": 5.086714744567871,
+ "step": 797
+ },
+ {
+ "epoch": 11.083842794759825,
+ "grad_norm": 0.02036641724407673,
+ "learning_rate": 0.0006,
+ "loss": 4.920056343078613,
+ "step": 798
+ },
+ {
+ "epoch": 11.097816593886463,
+ "grad_norm": 0.01966066285967827,
+ "learning_rate": 0.0006,
+ "loss": 4.879859447479248,
+ "step": 799
+ },
+ {
+ "epoch": 11.1117903930131,
+ "grad_norm": 0.0202149897813797,
+ "learning_rate": 0.0006,
+ "loss": 4.933640480041504,
+ "step": 800
+ },
+ {
+ "epoch": 11.125764192139737,
+ "grad_norm": 0.01893039606511593,
+ "learning_rate": 0.0006,
+ "loss": 4.901970863342285,
+ "step": 801
+ },
+ {
+ "epoch": 11.139737991266376,
+ "grad_norm": 0.016040155664086342,
+ "learning_rate": 0.0006,
+ "loss": 4.905491828918457,
+ "step": 802
+ },
+ {
+ "epoch": 11.153711790393013,
+ "grad_norm": 0.01720350608229637,
+ "learning_rate": 0.0006,
+ "loss": 4.798816680908203,
+ "step": 803
+ },
+ {
+ "epoch": 11.167685589519651,
+ "grad_norm": 0.017002852633595467,
+ "learning_rate": 0.0006,
+ "loss": 4.851859092712402,
+ "step": 804
+ },
+ {
+ "epoch": 11.181659388646288,
+ "grad_norm": 0.015594680793583393,
+ "learning_rate": 0.0006,
+ "loss": 4.90328311920166,
+ "step": 805
+ },
+ {
+ "epoch": 11.195633187772925,
+ "grad_norm": 0.014904462732374668,
+ "learning_rate": 0.0006,
+ "loss": 4.904215335845947,
+ "step": 806
+ },
+ {
+ "epoch": 11.209606986899564,
+ "grad_norm": 0.01366228237748146,
+ "learning_rate": 0.0006,
+ "loss": 4.9813055992126465,
+ "step": 807
+ },
+ {
+ "epoch": 11.2235807860262,
+ "grad_norm": 0.014719638973474503,
+ "learning_rate": 0.0006,
+ "loss": 4.927306175231934,
+ "step": 808
+ },
+ {
+ "epoch": 11.237554585152838,
+ "grad_norm": 0.015016036108136177,
+ "learning_rate": 0.0006,
+ "loss": 4.874215126037598,
+ "step": 809
+ },
+ {
+ "epoch": 11.251528384279476,
+ "grad_norm": 0.014925424009561539,
+ "learning_rate": 0.0006,
+ "loss": 4.838257789611816,
+ "step": 810
+ },
+ {
+ "epoch": 11.265502183406113,
+ "grad_norm": 0.014335823245346546,
+ "learning_rate": 0.0006,
+ "loss": 4.940608024597168,
+ "step": 811
+ },
+ {
+ "epoch": 11.279475982532752,
+ "grad_norm": 0.014987512491643429,
+ "learning_rate": 0.0006,
+ "loss": 4.831531524658203,
+ "step": 812
+ },
+ {
+ "epoch": 11.293449781659389,
+ "grad_norm": 0.016087457537651062,
+ "learning_rate": 0.0006,
+ "loss": 4.863227367401123,
+ "step": 813
+ },
+ {
+ "epoch": 11.307423580786025,
+ "grad_norm": 0.01671300269663334,
+ "learning_rate": 0.0006,
+ "loss": 4.8535661697387695,
+ "step": 814
+ },
+ {
+ "epoch": 11.321397379912664,
+ "grad_norm": 0.015229142270982265,
+ "learning_rate": 0.0006,
+ "loss": 4.948635101318359,
+ "step": 815
+ },
+ {
+ "epoch": 11.335371179039301,
+ "grad_norm": 0.01532587967813015,
+ "learning_rate": 0.0006,
+ "loss": 4.9035797119140625,
+ "step": 816
+ },
+ {
+ "epoch": 11.34934497816594,
+ "grad_norm": 0.015715476125478745,
+ "learning_rate": 0.0006,
+ "loss": 4.894709587097168,
+ "step": 817
+ },
+ {
+ "epoch": 11.363318777292577,
+ "grad_norm": 0.014156977646052837,
+ "learning_rate": 0.0006,
+ "loss": 4.960206985473633,
+ "step": 818
+ },
+ {
+ "epoch": 11.377292576419213,
+ "grad_norm": 0.017076566815376282,
+ "learning_rate": 0.0006,
+ "loss": 5.005850791931152,
+ "step": 819
+ },
+ {
+ "epoch": 11.391266375545852,
+ "grad_norm": 0.0172084029763937,
+ "learning_rate": 0.0006,
+ "loss": 4.903355598449707,
+ "step": 820
+ },
+ {
+ "epoch": 11.405240174672489,
+ "grad_norm": 0.0180258359760046,
+ "learning_rate": 0.0006,
+ "loss": 4.842667579650879,
+ "step": 821
+ },
+ {
+ "epoch": 11.419213973799126,
+ "grad_norm": 0.022130120545625687,
+ "learning_rate": 0.0006,
+ "loss": 4.799970626831055,
+ "step": 822
+ },
+ {
+ "epoch": 11.433187772925764,
+ "grad_norm": 0.02384897693991661,
+ "learning_rate": 0.0006,
+ "loss": 4.860746383666992,
+ "step": 823
+ },
+ {
+ "epoch": 11.447161572052401,
+ "grad_norm": 0.021881572902202606,
+ "learning_rate": 0.0006,
+ "loss": 4.918107032775879,
+ "step": 824
+ },
+ {
+ "epoch": 11.46113537117904,
+ "grad_norm": 0.023085080087184906,
+ "learning_rate": 0.0006,
+ "loss": 4.878946304321289,
+ "step": 825
+ },
+ {
+ "epoch": 11.475109170305677,
+ "grad_norm": 0.02291012555360794,
+ "learning_rate": 0.0006,
+ "loss": 4.958887100219727,
+ "step": 826
+ },
+ {
+ "epoch": 11.489082969432314,
+ "grad_norm": 0.0199968870729208,
+ "learning_rate": 0.0006,
+ "loss": 4.80955696105957,
+ "step": 827
+ },
+ {
+ "epoch": 11.503056768558952,
+ "grad_norm": 0.018702229484915733,
+ "learning_rate": 0.0006,
+ "loss": 5.0276408195495605,
+ "step": 828
+ },
+ {
+ "epoch": 11.51703056768559,
+ "grad_norm": 0.017787311226129532,
+ "learning_rate": 0.0006,
+ "loss": 4.96018123626709,
+ "step": 829
+ },
+ {
+ "epoch": 11.531004366812226,
+ "grad_norm": 0.01704501546919346,
+ "learning_rate": 0.0006,
+ "loss": 4.914553165435791,
+ "step": 830
+ },
+ {
+ "epoch": 11.544978165938865,
+ "grad_norm": 0.017519677057862282,
+ "learning_rate": 0.0006,
+ "loss": 4.798098087310791,
+ "step": 831
+ },
+ {
+ "epoch": 11.558951965065502,
+ "grad_norm": 0.018511991947889328,
+ "learning_rate": 0.0006,
+ "loss": 4.817785263061523,
+ "step": 832
+ },
+ {
+ "epoch": 11.57292576419214,
+ "grad_norm": 0.01876644790172577,
+ "learning_rate": 0.0006,
+ "loss": 4.818995475769043,
+ "step": 833
+ },
+ {
+ "epoch": 11.586899563318777,
+ "grad_norm": 0.01753568835556507,
+ "learning_rate": 0.0006,
+ "loss": 4.811938285827637,
+ "step": 834
+ },
+ {
+ "epoch": 11.600873362445414,
+ "grad_norm": 0.017046531662344933,
+ "learning_rate": 0.0006,
+ "loss": 4.914674282073975,
+ "step": 835
+ },
+ {
+ "epoch": 11.614847161572053,
+ "grad_norm": 0.01802094094455242,
+ "learning_rate": 0.0006,
+ "loss": 4.993284225463867,
+ "step": 836
+ },
+ {
+ "epoch": 11.62882096069869,
+ "grad_norm": 0.019493764266371727,
+ "learning_rate": 0.0006,
+ "loss": 4.895628452301025,
+ "step": 837
+ },
+ {
+ "epoch": 11.642794759825328,
+ "grad_norm": 0.019461210817098618,
+ "learning_rate": 0.0006,
+ "loss": 4.869269371032715,
+ "step": 838
+ },
+ {
+ "epoch": 11.656768558951965,
+ "grad_norm": 0.01981567218899727,
+ "learning_rate": 0.0006,
+ "loss": 4.863052845001221,
+ "step": 839
+ },
+ {
+ "epoch": 11.670742358078602,
+ "grad_norm": 0.019894491881132126,
+ "learning_rate": 0.0006,
+ "loss": 4.8775787353515625,
+ "step": 840
+ },
+ {
+ "epoch": 11.68471615720524,
+ "grad_norm": 0.01875786855816841,
+ "learning_rate": 0.0006,
+ "loss": 4.817763328552246,
+ "step": 841
+ },
+ {
+ "epoch": 11.698689956331878,
+ "grad_norm": 0.020926695317029953,
+ "learning_rate": 0.0006,
+ "loss": 4.7470526695251465,
+ "step": 842
+ },
+ {
+ "epoch": 11.712663755458514,
+ "grad_norm": 0.021477103233337402,
+ "learning_rate": 0.0006,
+ "loss": 4.762682914733887,
+ "step": 843
+ },
+ {
+ "epoch": 11.726637554585153,
+ "grad_norm": 0.020628711208701134,
+ "learning_rate": 0.0006,
+ "loss": 4.807985782623291,
+ "step": 844
+ },
+ {
+ "epoch": 11.74061135371179,
+ "grad_norm": 0.021489612758159637,
+ "learning_rate": 0.0006,
+ "loss": 4.843091011047363,
+ "step": 845
+ },
+ {
+ "epoch": 11.754585152838429,
+ "grad_norm": 0.021194491535425186,
+ "learning_rate": 0.0006,
+ "loss": 4.8300251960754395,
+ "step": 846
+ },
+ {
+ "epoch": 11.768558951965066,
+ "grad_norm": 0.01833350397646427,
+ "learning_rate": 0.0006,
+ "loss": 5.000202178955078,
+ "step": 847
+ },
+ {
+ "epoch": 11.782532751091702,
+ "grad_norm": 0.018096931278705597,
+ "learning_rate": 0.0006,
+ "loss": 4.880648136138916,
+ "step": 848
+ },
+ {
+ "epoch": 11.796506550218341,
+ "grad_norm": 0.0197161752730608,
+ "learning_rate": 0.0006,
+ "loss": 4.788316249847412,
+ "step": 849
+ },
+ {
+ "epoch": 11.810480349344978,
+ "grad_norm": 0.0203181654214859,
+ "learning_rate": 0.0006,
+ "loss": 4.870944976806641,
+ "step": 850
+ },
+ {
+ "epoch": 11.824454148471617,
+ "grad_norm": 0.01908455230295658,
+ "learning_rate": 0.0006,
+ "loss": 4.890144348144531,
+ "step": 851
+ },
+ {
+ "epoch": 11.838427947598253,
+ "grad_norm": 0.021281801164150238,
+ "learning_rate": 0.0006,
+ "loss": 4.89578914642334,
+ "step": 852
+ },
+ {
+ "epoch": 11.85240174672489,
+ "grad_norm": 0.02124273031949997,
+ "learning_rate": 0.0006,
+ "loss": 4.883749008178711,
+ "step": 853
+ },
+ {
+ "epoch": 11.866375545851529,
+ "grad_norm": 0.020969398319721222,
+ "learning_rate": 0.0006,
+ "loss": 4.809142589569092,
+ "step": 854
+ },
+ {
+ "epoch": 11.880349344978166,
+ "grad_norm": 0.01967657171189785,
+ "learning_rate": 0.0006,
+ "loss": 4.821864604949951,
+ "step": 855
+ },
+ {
+ "epoch": 11.894323144104803,
+ "grad_norm": 0.020887333899736404,
+ "learning_rate": 0.0006,
+ "loss": 4.917985916137695,
+ "step": 856
+ },
+ {
+ "epoch": 11.908296943231441,
+ "grad_norm": 0.019203083589673042,
+ "learning_rate": 0.0006,
+ "loss": 4.824960708618164,
+ "step": 857
+ },
+ {
+ "epoch": 11.922270742358078,
+ "grad_norm": 0.020162401720881462,
+ "learning_rate": 0.0006,
+ "loss": 4.772617816925049,
+ "step": 858
+ },
+ {
+ "epoch": 11.936244541484717,
+ "grad_norm": 0.02251487225294113,
+ "learning_rate": 0.0006,
+ "loss": 4.735418319702148,
+ "step": 859
+ },
+ {
+ "epoch": 11.950218340611354,
+ "grad_norm": 0.020313527435064316,
+ "learning_rate": 0.0006,
+ "loss": 4.939233779907227,
+ "step": 860
+ },
+ {
+ "epoch": 11.96419213973799,
+ "grad_norm": 0.021635599434375763,
+ "learning_rate": 0.0006,
+ "loss": 4.861807346343994,
+ "step": 861
+ },
+ {
+ "epoch": 11.97816593886463,
+ "grad_norm": 0.02285311557352543,
+ "learning_rate": 0.0006,
+ "loss": 4.932732582092285,
+ "step": 862
+ },
+ {
+ "epoch": 11.992139737991266,
+ "grad_norm": 0.02072535641491413,
+ "learning_rate": 0.0006,
+ "loss": 4.796767711639404,
+ "step": 863
+ },
+ {
+ "epoch": 12.0,
+ "grad_norm": 0.020588304847478867,
+ "learning_rate": 0.0006,
+ "loss": 4.997537612915039,
+ "step": 864
+ },
+ {
+ "epoch": 12.0,
+ "eval_loss": 5.044651985168457,
+ "eval_runtime": 56.9145,
+ "eval_samples_per_second": 42.906,
+ "eval_steps_per_second": 1.353,
+ "step": 864
+ },
+ {
+ "epoch": 12.013973799126637,
+ "grad_norm": 0.017597166821360588,
+ "learning_rate": 0.0006,
+ "loss": 4.830524444580078,
+ "step": 865
+ },
+ {
+ "epoch": 12.027947598253276,
+ "grad_norm": 0.01716788113117218,
+ "learning_rate": 0.0006,
+ "loss": 4.894325256347656,
+ "step": 866
+ },
+ {
+ "epoch": 12.041921397379912,
+ "grad_norm": 0.017712457105517387,
+ "learning_rate": 0.0006,
+ "loss": 4.868297576904297,
+ "step": 867
+ },
+ {
+ "epoch": 12.055895196506551,
+ "grad_norm": 0.021486390382051468,
+ "learning_rate": 0.0006,
+ "loss": 4.922049522399902,
+ "step": 868
+ },
+ {
+ "epoch": 12.069868995633188,
+ "grad_norm": 0.024560654535889626,
+ "learning_rate": 0.0006,
+ "loss": 4.734241008758545,
+ "step": 869
+ },
+ {
+ "epoch": 12.083842794759825,
+ "grad_norm": 0.025870252400636673,
+ "learning_rate": 0.0006,
+ "loss": 4.868566989898682,
+ "step": 870
+ },
+ {
+ "epoch": 12.097816593886463,
+ "grad_norm": 0.0277389045804739,
+ "learning_rate": 0.0006,
+ "loss": 4.755006790161133,
+ "step": 871
+ },
+ {
+ "epoch": 12.1117903930131,
+ "grad_norm": 0.029100844636559486,
+ "learning_rate": 0.0006,
+ "loss": 4.909667015075684,
+ "step": 872
+ },
+ {
+ "epoch": 12.125764192139737,
+ "grad_norm": 0.031833067536354065,
+ "learning_rate": 0.0006,
+ "loss": 4.858674049377441,
+ "step": 873
+ },
+ {
+ "epoch": 12.139737991266376,
+ "grad_norm": 0.032709237188100815,
+ "learning_rate": 0.0006,
+ "loss": 4.823696136474609,
+ "step": 874
+ },
+ {
+ "epoch": 12.153711790393013,
+ "grad_norm": 0.02995380386710167,
+ "learning_rate": 0.0006,
+ "loss": 4.850235462188721,
+ "step": 875
+ },
+ {
+ "epoch": 12.167685589519651,
+ "grad_norm": 0.034098610281944275,
+ "learning_rate": 0.0006,
+ "loss": 4.8169989585876465,
+ "step": 876
+ },
+ {
+ "epoch": 12.181659388646288,
+ "grad_norm": 0.03377080336213112,
+ "learning_rate": 0.0006,
+ "loss": 4.90738582611084,
+ "step": 877
+ },
+ {
+ "epoch": 12.195633187772925,
+ "grad_norm": 0.029780825600028038,
+ "learning_rate": 0.0006,
+ "loss": 4.811163902282715,
+ "step": 878
+ },
+ {
+ "epoch": 12.209606986899564,
+ "grad_norm": 0.03118024580180645,
+ "learning_rate": 0.0006,
+ "loss": 4.861286640167236,
+ "step": 879
+ },
+ {
+ "epoch": 12.2235807860262,
+ "grad_norm": 0.033052004873752594,
+ "learning_rate": 0.0006,
+ "loss": 4.865021228790283,
+ "step": 880
+ },
+ {
+ "epoch": 12.237554585152838,
+ "grad_norm": 0.029070032760500908,
+ "learning_rate": 0.0006,
+ "loss": 4.947303771972656,
+ "step": 881
+ },
+ {
+ "epoch": 12.251528384279476,
+ "grad_norm": 0.028355715796351433,
+ "learning_rate": 0.0006,
+ "loss": 4.804648399353027,
+ "step": 882
+ },
+ {
+ "epoch": 12.265502183406113,
+ "grad_norm": 0.031257227063179016,
+ "learning_rate": 0.0006,
+ "loss": 4.757718086242676,
+ "step": 883
+ },
+ {
+ "epoch": 12.279475982532752,
+ "grad_norm": 0.031438373029232025,
+ "learning_rate": 0.0006,
+ "loss": 4.851268768310547,
+ "step": 884
+ },
+ {
+ "epoch": 12.293449781659389,
+ "grad_norm": 0.030492138117551804,
+ "learning_rate": 0.0006,
+ "loss": 4.7891035079956055,
+ "step": 885
+ },
+ {
+ "epoch": 12.307423580786025,
+ "grad_norm": 0.026115991175174713,
+ "learning_rate": 0.0006,
+ "loss": 4.86474609375,
+ "step": 886
+ },
+ {
+ "epoch": 12.321397379912664,
+ "grad_norm": 0.024184465408325195,
+ "learning_rate": 0.0006,
+ "loss": 4.860456943511963,
+ "step": 887
+ },
+ {
+ "epoch": 12.335371179039301,
+ "grad_norm": 0.025437982752919197,
+ "learning_rate": 0.0006,
+ "loss": 4.806251049041748,
+ "step": 888
+ },
+ {
+ "epoch": 12.34934497816594,
+ "grad_norm": 0.022390231490135193,
+ "learning_rate": 0.0006,
+ "loss": 4.807086944580078,
+ "step": 889
+ },
+ {
+ "epoch": 12.363318777292577,
+ "grad_norm": 0.020174294710159302,
+ "learning_rate": 0.0006,
+ "loss": 4.78292179107666,
+ "step": 890
+ },
+ {
+ "epoch": 12.377292576419213,
+ "grad_norm": 0.019734324887394905,
+ "learning_rate": 0.0006,
+ "loss": 4.784509658813477,
+ "step": 891
+ },
+ {
+ "epoch": 12.391266375545852,
+ "grad_norm": 0.016733523458242416,
+ "learning_rate": 0.0006,
+ "loss": 4.811755180358887,
+ "step": 892
+ },
+ {
+ "epoch": 12.405240174672489,
+ "grad_norm": 0.01570146717131138,
+ "learning_rate": 0.0006,
+ "loss": 4.720088005065918,
+ "step": 893
+ },
+ {
+ "epoch": 12.419213973799126,
+ "grad_norm": 0.01601393334567547,
+ "learning_rate": 0.0006,
+ "loss": 4.783543109893799,
+ "step": 894
+ },
+ {
+ "epoch": 12.433187772925764,
+ "grad_norm": 0.014035487547516823,
+ "learning_rate": 0.0006,
+ "loss": 4.848340034484863,
+ "step": 895
+ },
+ {
+ "epoch": 12.447161572052401,
+ "grad_norm": 0.015248487703502178,
+ "learning_rate": 0.0006,
+ "loss": 4.773287773132324,
+ "step": 896
+ },
+ {
+ "epoch": 12.46113537117904,
+ "grad_norm": 0.01532268151640892,
+ "learning_rate": 0.0006,
+ "loss": 4.799646854400635,
+ "step": 897
+ },
+ {
+ "epoch": 12.475109170305677,
+ "grad_norm": 0.013982534408569336,
+ "learning_rate": 0.0006,
+ "loss": 4.8090500831604,
+ "step": 898
+ },
+ {
+ "epoch": 12.489082969432314,
+ "grad_norm": 0.014081164263188839,
+ "learning_rate": 0.0006,
+ "loss": 4.751888751983643,
+ "step": 899
+ },
+ {
+ "epoch": 12.503056768558952,
+ "grad_norm": 0.01380133256316185,
+ "learning_rate": 0.0006,
+ "loss": 4.78144645690918,
+ "step": 900
+ },
+ {
+ "epoch": 12.51703056768559,
+ "grad_norm": 0.013555269688367844,
+ "learning_rate": 0.0006,
+ "loss": 4.779257774353027,
+ "step": 901
+ },
+ {
+ "epoch": 12.531004366812226,
+ "grad_norm": 0.013075701892375946,
+ "learning_rate": 0.0006,
+ "loss": 4.820705413818359,
+ "step": 902
+ },
+ {
+ "epoch": 12.544978165938865,
+ "grad_norm": 0.01276316400617361,
+ "learning_rate": 0.0006,
+ "loss": 4.836733818054199,
+ "step": 903
+ },
+ {
+ "epoch": 12.558951965065502,
+ "grad_norm": 0.013033892959356308,
+ "learning_rate": 0.0006,
+ "loss": 4.764178276062012,
+ "step": 904
+ },
+ {
+ "epoch": 12.57292576419214,
+ "grad_norm": 0.013996980153024197,
+ "learning_rate": 0.0006,
+ "loss": 4.767983436584473,
+ "step": 905
+ },
+ {
+ "epoch": 12.586899563318777,
+ "grad_norm": 0.015222184360027313,
+ "learning_rate": 0.0006,
+ "loss": 4.759721755981445,
+ "step": 906
+ },
+ {
+ "epoch": 12.600873362445414,
+ "grad_norm": 0.01623941957950592,
+ "learning_rate": 0.0006,
+ "loss": 4.849811553955078,
+ "step": 907
+ },
+ {
+ "epoch": 12.614847161572053,
+ "grad_norm": 0.015711182728409767,
+ "learning_rate": 0.0006,
+ "loss": 4.7829084396362305,
+ "step": 908
+ },
+ {
+ "epoch": 12.62882096069869,
+ "grad_norm": 0.014137467369437218,
+ "learning_rate": 0.0006,
+ "loss": 4.73321533203125,
+ "step": 909
+ },
+ {
+ "epoch": 12.642794759825328,
+ "grad_norm": 0.012738462537527084,
+ "learning_rate": 0.0006,
+ "loss": 4.870052337646484,
+ "step": 910
+ },
+ {
+ "epoch": 12.656768558951965,
+ "grad_norm": 0.013683130033314228,
+ "learning_rate": 0.0006,
+ "loss": 4.82658052444458,
+ "step": 911
+ },
+ {
+ "epoch": 12.670742358078602,
+ "grad_norm": 0.014409830793738365,
+ "learning_rate": 0.0006,
+ "loss": 4.707494735717773,
+ "step": 912
+ },
+ {
+ "epoch": 12.68471615720524,
+ "grad_norm": 0.013806809671223164,
+ "learning_rate": 0.0006,
+ "loss": 4.778928279876709,
+ "step": 913
+ },
+ {
+ "epoch": 12.698689956331878,
+ "grad_norm": 0.012857912108302116,
+ "learning_rate": 0.0006,
+ "loss": 4.890717506408691,
+ "step": 914
+ },
+ {
+ "epoch": 12.712663755458514,
+ "grad_norm": 0.013828758150339127,
+ "learning_rate": 0.0006,
+ "loss": 4.8357439041137695,
+ "step": 915
+ },
+ {
+ "epoch": 12.726637554585153,
+ "grad_norm": 0.01478351280093193,
+ "learning_rate": 0.0006,
+ "loss": 4.869986534118652,
+ "step": 916
+ },
+ {
+ "epoch": 12.74061135371179,
+ "grad_norm": 0.01326421182602644,
+ "learning_rate": 0.0006,
+ "loss": 4.830646514892578,
+ "step": 917
+ },
+ {
+ "epoch": 12.754585152838429,
+ "grad_norm": 0.013335433788597584,
+ "learning_rate": 0.0006,
+ "loss": 4.7786149978637695,
+ "step": 918
+ },
+ {
+ "epoch": 12.768558951965066,
+ "grad_norm": 0.014261147007346153,
+ "learning_rate": 0.0006,
+ "loss": 4.788693428039551,
+ "step": 919
+ },
+ {
+ "epoch": 12.782532751091702,
+ "grad_norm": 0.014274783432483673,
+ "learning_rate": 0.0006,
+ "loss": 4.77295446395874,
+ "step": 920
+ },
+ {
+ "epoch": 12.796506550218341,
+ "grad_norm": 0.015074674971401691,
+ "learning_rate": 0.0006,
+ "loss": 4.830754280090332,
+ "step": 921
+ },
+ {
+ "epoch": 12.810480349344978,
+ "grad_norm": 0.015599515289068222,
+ "learning_rate": 0.0006,
+ "loss": 4.837759017944336,
+ "step": 922
+ },
+ {
+ "epoch": 12.824454148471617,
+ "grad_norm": 0.014595242217183113,
+ "learning_rate": 0.0006,
+ "loss": 4.882965564727783,
+ "step": 923
+ },
+ {
+ "epoch": 12.838427947598253,
+ "grad_norm": 0.014951467514038086,
+ "learning_rate": 0.0006,
+ "loss": 4.709748268127441,
+ "step": 924
+ },
+ {
+ "epoch": 12.85240174672489,
+ "grad_norm": 0.01539295818656683,
+ "learning_rate": 0.0006,
+ "loss": 4.791160583496094,
+ "step": 925
+ },
+ {
+ "epoch": 12.866375545851529,
+ "grad_norm": 0.015987906605005264,
+ "learning_rate": 0.0006,
+ "loss": 4.661876678466797,
+ "step": 926
+ },
+ {
+ "epoch": 12.880349344978166,
+ "grad_norm": 0.016527341678738594,
+ "learning_rate": 0.0006,
+ "loss": 4.721952438354492,
+ "step": 927
+ },
+ {
+ "epoch": 12.894323144104803,
+ "grad_norm": 0.01637161523103714,
+ "learning_rate": 0.0006,
+ "loss": 4.704446792602539,
+ "step": 928
+ },
+ {
+ "epoch": 12.908296943231441,
+ "grad_norm": 0.015221050009131432,
+ "learning_rate": 0.0006,
+ "loss": 4.771907806396484,
+ "step": 929
+ },
+ {
+ "epoch": 12.922270742358078,
+ "grad_norm": 0.016242166981101036,
+ "learning_rate": 0.0006,
+ "loss": 4.806267738342285,
+ "step": 930
+ },
+ {
+ "epoch": 12.936244541484717,
+ "grad_norm": 0.01628357358276844,
+ "learning_rate": 0.0006,
+ "loss": 4.7252326011657715,
+ "step": 931
+ },
+ {
+ "epoch": 12.950218340611354,
+ "grad_norm": 0.017428133636713028,
+ "learning_rate": 0.0006,
+ "loss": 4.765689849853516,
+ "step": 932
+ },
+ {
+ "epoch": 12.96419213973799,
+ "grad_norm": 0.01870061084628105,
+ "learning_rate": 0.0006,
+ "loss": 4.814488887786865,
+ "step": 933
+ },
+ {
+ "epoch": 12.97816593886463,
+ "grad_norm": 0.019796060398221016,
+ "learning_rate": 0.0006,
+ "loss": 4.739185810089111,
+ "step": 934
+ },
+ {
+ "epoch": 12.992139737991266,
+ "grad_norm": 0.01985820196568966,
+ "learning_rate": 0.0006,
+ "loss": 4.810608863830566,
+ "step": 935
+ },
+ {
+ "epoch": 13.0,
+ "grad_norm": 0.020834477618336678,
+ "learning_rate": 0.0006,
+ "loss": 4.727858066558838,
+ "step": 936
+ },
+ {
+ "epoch": 13.0,
+ "eval_loss": 4.979116916656494,
+ "eval_runtime": 57.2877,
+ "eval_samples_per_second": 42.627,
+ "eval_steps_per_second": 1.344,
+ "step": 936
+ },
+ {
+ "epoch": 13.013973799126637,
+ "grad_norm": 0.02025519125163555,
+ "learning_rate": 0.0006,
+ "loss": 4.628875732421875,
+ "step": 937
+ },
+ {
+ "epoch": 13.027947598253276,
+ "grad_norm": 0.019596368074417114,
+ "learning_rate": 0.0006,
+ "loss": 4.694699287414551,
+ "step": 938
+ },
+ {
+ "epoch": 13.041921397379912,
+ "grad_norm": 0.020494934171438217,
+ "learning_rate": 0.0006,
+ "loss": 4.691821098327637,
+ "step": 939
+ },
+ {
+ "epoch": 13.055895196506551,
+ "grad_norm": 0.020121760666370392,
+ "learning_rate": 0.0006,
+ "loss": 4.71478271484375,
+ "step": 940
+ },
+ {
+ "epoch": 13.069868995633188,
+ "grad_norm": 0.021091405302286148,
+ "learning_rate": 0.0006,
+ "loss": 4.695915222167969,
+ "step": 941
+ },
+ {
+ "epoch": 13.083842794759825,
+ "grad_norm": 0.019787423312664032,
+ "learning_rate": 0.0006,
+ "loss": 4.8652448654174805,
+ "step": 942
+ },
+ {
+ "epoch": 13.097816593886463,
+ "grad_norm": 0.0185833927243948,
+ "learning_rate": 0.0006,
+ "loss": 4.7065887451171875,
+ "step": 943
+ },
+ {
+ "epoch": 13.1117903930131,
+ "grad_norm": 0.021220475435256958,
+ "learning_rate": 0.0006,
+ "loss": 4.741232395172119,
+ "step": 944
+ },
+ {
+ "epoch": 13.125764192139737,
+ "grad_norm": 0.02136622555553913,
+ "learning_rate": 0.0006,
+ "loss": 4.705967903137207,
+ "step": 945
+ },
+ {
+ "epoch": 13.139737991266376,
+ "grad_norm": 0.0191587433218956,
+ "learning_rate": 0.0006,
+ "loss": 4.758667945861816,
+ "step": 946
+ },
+ {
+ "epoch": 13.153711790393013,
+ "grad_norm": 0.018290022388100624,
+ "learning_rate": 0.0006,
+ "loss": 4.760666847229004,
+ "step": 947
+ },
+ {
+ "epoch": 13.167685589519651,
+ "grad_norm": 0.021704409271478653,
+ "learning_rate": 0.0006,
+ "loss": 4.750555992126465,
+ "step": 948
+ },
+ {
+ "epoch": 13.181659388646288,
+ "grad_norm": 0.021668918430805206,
+ "learning_rate": 0.0006,
+ "loss": 4.840903282165527,
+ "step": 949
+ },
+ {
+ "epoch": 13.195633187772925,
+ "grad_norm": 0.019497141242027283,
+ "learning_rate": 0.0006,
+ "loss": 4.738969326019287,
+ "step": 950
+ },
+ {
+ "epoch": 13.209606986899564,
+ "grad_norm": 0.018500596284866333,
+ "learning_rate": 0.0006,
+ "loss": 4.75154972076416,
+ "step": 951
+ },
+ {
+ "epoch": 13.2235807860262,
+ "grad_norm": 0.01872239261865616,
+ "learning_rate": 0.0006,
+ "loss": 4.811738014221191,
+ "step": 952
+ },
+ {
+ "epoch": 13.237554585152838,
+ "grad_norm": 0.020321547985076904,
+ "learning_rate": 0.0006,
+ "loss": 4.815384387969971,
+ "step": 953
+ },
+ {
+ "epoch": 13.251528384279476,
+ "grad_norm": 0.02283575013279915,
+ "learning_rate": 0.0006,
+ "loss": 4.74582052230835,
+ "step": 954
+ },
+ {
+ "epoch": 13.265502183406113,
+ "grad_norm": 0.02474398910999298,
+ "learning_rate": 0.0006,
+ "loss": 4.738375186920166,
+ "step": 955
+ },
+ {
+ "epoch": 13.279475982532752,
+ "grad_norm": 0.02065764181315899,
+ "learning_rate": 0.0006,
+ "loss": 4.746875286102295,
+ "step": 956
+ },
+ {
+ "epoch": 13.293449781659389,
+ "grad_norm": 0.020534520968794823,
+ "learning_rate": 0.0006,
+ "loss": 4.656914710998535,
+ "step": 957
+ },
+ {
+ "epoch": 13.307423580786025,
+ "grad_norm": 0.0206378772854805,
+ "learning_rate": 0.0006,
+ "loss": 4.842615604400635,
+ "step": 958
+ },
+ {
+ "epoch": 13.321397379912664,
+ "grad_norm": 0.017935654148459435,
+ "learning_rate": 0.0006,
+ "loss": 4.717827796936035,
+ "step": 959
+ },
+ {
+ "epoch": 13.335371179039301,
+ "grad_norm": 0.016958363354206085,
+ "learning_rate": 0.0006,
+ "loss": 4.681070327758789,
+ "step": 960
+ },
+ {
+ "epoch": 13.34934497816594,
+ "grad_norm": 0.015735412016510963,
+ "learning_rate": 0.0006,
+ "loss": 4.7957048416137695,
+ "step": 961
+ },
+ {
+ "epoch": 13.363318777292577,
+ "grad_norm": 0.01563699170947075,
+ "learning_rate": 0.0006,
+ "loss": 4.701192378997803,
+ "step": 962
+ },
+ {
+ "epoch": 13.377292576419213,
+ "grad_norm": 0.015207760035991669,
+ "learning_rate": 0.0006,
+ "loss": 4.73256254196167,
+ "step": 963
+ },
+ {
+ "epoch": 13.391266375545852,
+ "grad_norm": 0.015396883711218834,
+ "learning_rate": 0.0006,
+ "loss": 4.76412296295166,
+ "step": 964
+ },
+ {
+ "epoch": 13.405240174672489,
+ "grad_norm": 0.015690583735704422,
+ "learning_rate": 0.0006,
+ "loss": 4.791157245635986,
+ "step": 965
+ },
+ {
+ "epoch": 13.419213973799126,
+ "grad_norm": 0.01755724661052227,
+ "learning_rate": 0.0006,
+ "loss": 4.845666885375977,
+ "step": 966
+ },
+ {
+ "epoch": 13.433187772925764,
+ "grad_norm": 0.019443422555923462,
+ "learning_rate": 0.0006,
+ "loss": 4.5902299880981445,
+ "step": 967
+ },
+ {
+ "epoch": 13.447161572052401,
+ "grad_norm": 0.02118338644504547,
+ "learning_rate": 0.0006,
+ "loss": 4.740627288818359,
+ "step": 968
+ },
+ {
+ "epoch": 13.46113537117904,
+ "grad_norm": 0.02174839936196804,
+ "learning_rate": 0.0006,
+ "loss": 4.814560890197754,
+ "step": 969
+ },
+ {
+ "epoch": 13.475109170305677,
+ "grad_norm": 0.022860374301671982,
+ "learning_rate": 0.0006,
+ "loss": 4.66440486907959,
+ "step": 970
+ },
+ {
+ "epoch": 13.489082969432314,
+ "grad_norm": 0.024742677807807922,
+ "learning_rate": 0.0006,
+ "loss": 4.7451653480529785,
+ "step": 971
+ },
+ {
+ "epoch": 13.503056768558952,
+ "grad_norm": 0.027068907395005226,
+ "learning_rate": 0.0006,
+ "loss": 4.802584171295166,
+ "step": 972
+ },
+ {
+ "epoch": 13.51703056768559,
+ "grad_norm": 0.026316625997424126,
+ "learning_rate": 0.0006,
+ "loss": 4.842397689819336,
+ "step": 973
+ },
+ {
+ "epoch": 13.531004366812226,
+ "grad_norm": 0.024610962718725204,
+ "learning_rate": 0.0006,
+ "loss": 4.650721073150635,
+ "step": 974
+ },
+ {
+ "epoch": 13.544978165938865,
+ "grad_norm": 0.024122603237628937,
+ "learning_rate": 0.0006,
+ "loss": 4.6984357833862305,
+ "step": 975
+ },
+ {
+ "epoch": 13.558951965065502,
+ "grad_norm": 0.022699708119034767,
+ "learning_rate": 0.0006,
+ "loss": 4.793007850646973,
+ "step": 976
+ },
+ {
+ "epoch": 13.57292576419214,
+ "grad_norm": 0.021706534549593925,
+ "learning_rate": 0.0006,
+ "loss": 4.691362380981445,
+ "step": 977
+ },
+ {
+ "epoch": 13.586899563318777,
+ "grad_norm": 0.021363291889429092,
+ "learning_rate": 0.0006,
+ "loss": 4.76762580871582,
+ "step": 978
+ },
+ {
+ "epoch": 13.600873362445414,
+ "grad_norm": 0.023476677015423775,
+ "learning_rate": 0.0006,
+ "loss": 4.777202606201172,
+ "step": 979
+ },
+ {
+ "epoch": 13.614847161572053,
+ "grad_norm": 0.02457100711762905,
+ "learning_rate": 0.0006,
+ "loss": 4.590703010559082,
+ "step": 980
+ },
+ {
+ "epoch": 13.62882096069869,
+ "grad_norm": 0.02469099499285221,
+ "learning_rate": 0.0006,
+ "loss": 4.640185356140137,
+ "step": 981
+ },
+ {
+ "epoch": 13.642794759825328,
+ "grad_norm": 0.024914603680372238,
+ "learning_rate": 0.0006,
+ "loss": 4.772201061248779,
+ "step": 982
+ },
+ {
+ "epoch": 13.656768558951965,
+ "grad_norm": 0.025133173912763596,
+ "learning_rate": 0.0006,
+ "loss": 4.743337154388428,
+ "step": 983
+ },
+ {
+ "epoch": 13.670742358078602,
+ "grad_norm": 0.02034137211740017,
+ "learning_rate": 0.0006,
+ "loss": 4.752127647399902,
+ "step": 984
+ },
+ {
+ "epoch": 13.68471615720524,
+ "grad_norm": 0.019645661115646362,
+ "learning_rate": 0.0006,
+ "loss": 4.838453769683838,
+ "step": 985
+ },
+ {
+ "epoch": 13.698689956331878,
+ "grad_norm": 0.01900297962129116,
+ "learning_rate": 0.0006,
+ "loss": 4.804868221282959,
+ "step": 986
+ },
+ {
+ "epoch": 13.712663755458514,
+ "grad_norm": 0.020292341709136963,
+ "learning_rate": 0.0006,
+ "loss": 4.609743118286133,
+ "step": 987
+ },
+ {
+ "epoch": 13.726637554585153,
+ "grad_norm": 0.017014775425195694,
+ "learning_rate": 0.0006,
+ "loss": 4.740612983703613,
+ "step": 988
+ },
+ {
+ "epoch": 13.74061135371179,
+ "grad_norm": 0.015199844725430012,
+ "learning_rate": 0.0006,
+ "loss": 4.719701766967773,
+ "step": 989
+ },
+ {
+ "epoch": 13.754585152838429,
+ "grad_norm": 0.015125414356589317,
+ "learning_rate": 0.0006,
+ "loss": 4.686114311218262,
+ "step": 990
+ },
+ {
+ "epoch": 13.768558951965066,
+ "grad_norm": 0.014779540710151196,
+ "learning_rate": 0.0006,
+ "loss": 4.674347877502441,
+ "step": 991
+ },
+ {
+ "epoch": 13.782532751091702,
+ "grad_norm": 0.015255163423717022,
+ "learning_rate": 0.0006,
+ "loss": 4.6309309005737305,
+ "step": 992
+ },
+ {
+ "epoch": 13.796506550218341,
+ "grad_norm": 0.013692197389900684,
+ "learning_rate": 0.0006,
+ "loss": 4.698497295379639,
+ "step": 993
+ },
+ {
+ "epoch": 13.810480349344978,
+ "grad_norm": 0.014529623091220856,
+ "learning_rate": 0.0006,
+ "loss": 4.628443717956543,
+ "step": 994
+ },
+ {
+ "epoch": 13.824454148471617,
+ "grad_norm": 0.01576659083366394,
+ "learning_rate": 0.0006,
+ "loss": 4.788015842437744,
+ "step": 995
+ },
+ {
+ "epoch": 13.838427947598253,
+ "grad_norm": 0.015834277495741844,
+ "learning_rate": 0.0006,
+ "loss": 4.672703742980957,
+ "step": 996
+ },
+ {
+ "epoch": 13.85240174672489,
+ "grad_norm": 0.015539503656327724,
+ "learning_rate": 0.0006,
+ "loss": 4.731503486633301,
+ "step": 997
+ },
+ {
+ "epoch": 13.866375545851529,
+ "grad_norm": 0.013463746756315231,
+ "learning_rate": 0.0006,
+ "loss": 4.7006330490112305,
+ "step": 998
+ },
+ {
+ "epoch": 13.880349344978166,
+ "grad_norm": 0.016377726569771767,
+ "learning_rate": 0.0006,
+ "loss": 4.636074066162109,
+ "step": 999
+ },
+ {
+ "epoch": 13.894323144104803,
+ "grad_norm": 0.016660034656524658,
+ "learning_rate": 0.0006,
+ "loss": 4.742693901062012,
+ "step": 1000
+ },
+ {
+ "epoch": 13.908296943231441,
+ "grad_norm": 0.014902938157320023,
+ "learning_rate": 0.0006,
+ "loss": 4.769811630249023,
+ "step": 1001
+ },
+ {
+ "epoch": 13.922270742358078,
+ "grad_norm": 0.014883476309478283,
+ "learning_rate": 0.0006,
+ "loss": 4.731159210205078,
+ "step": 1002
+ },
+ {
+ "epoch": 13.936244541484717,
+ "grad_norm": 0.014494217932224274,
+ "learning_rate": 0.0006,
+ "loss": 4.659485816955566,
+ "step": 1003
+ },
+ {
+ "epoch": 13.950218340611354,
+ "grad_norm": 0.015426991507411003,
+ "learning_rate": 0.0006,
+ "loss": 4.755229949951172,
+ "step": 1004
+ },
+ {
+ "epoch": 13.96419213973799,
+ "grad_norm": 0.016648586839437485,
+ "learning_rate": 0.0006,
+ "loss": 4.74111270904541,
+ "step": 1005
+ },
+ {
+ "epoch": 13.97816593886463,
+ "grad_norm": 0.016980541869997978,
+ "learning_rate": 0.0006,
+ "loss": 4.74072265625,
+ "step": 1006
+ },
+ {
+ "epoch": 13.992139737991266,
+ "grad_norm": 0.018070152029395103,
+ "learning_rate": 0.0006,
+ "loss": 4.732975006103516,
+ "step": 1007
+ },
+ {
+ "epoch": 14.0,
+ "grad_norm": 0.019910665228962898,
+ "learning_rate": 0.0006,
+ "loss": 4.634697914123535,
+ "step": 1008
+ },
+ {
+ "epoch": 14.0,
+ "eval_loss": 4.850640296936035,
+ "eval_runtime": 57.3039,
+ "eval_samples_per_second": 42.615,
+ "eval_steps_per_second": 1.344,
+ "step": 1008
+ },
+ {
+ "epoch": 14.013973799126637,
+ "grad_norm": 0.01980053074657917,
+ "learning_rate": 0.0006,
+ "loss": 4.628744602203369,
+ "step": 1009
+ },
+ {
+ "epoch": 14.027947598253276,
+ "grad_norm": 0.02497623674571514,
+ "learning_rate": 0.0006,
+ "loss": 4.75930118560791,
+ "step": 1010
+ },
+ {
+ "epoch": 14.041921397379912,
+ "grad_norm": 0.02794097363948822,
+ "learning_rate": 0.0006,
+ "loss": 4.691340446472168,
+ "step": 1011
+ },
+ {
+ "epoch": 14.055895196506551,
+ "grad_norm": 0.023725535720586777,
+ "learning_rate": 0.0006,
+ "loss": 4.733664512634277,
+ "step": 1012
+ },
+ {
+ "epoch": 14.069868995633188,
+ "grad_norm": 0.022966474294662476,
+ "learning_rate": 0.0006,
+ "loss": 4.664808750152588,
+ "step": 1013
+ },
+ {
+ "epoch": 14.083842794759825,
+ "grad_norm": 0.02530563250184059,
+ "learning_rate": 0.0006,
+ "loss": 4.8047285079956055,
+ "step": 1014
+ },
+ {
+ "epoch": 14.097816593886463,
+ "grad_norm": 0.02491082064807415,
+ "learning_rate": 0.0006,
+ "loss": 4.667956352233887,
+ "step": 1015
+ },
+ {
+ "epoch": 14.1117903930131,
+ "grad_norm": 0.022110294550657272,
+ "learning_rate": 0.0006,
+ "loss": 4.666140556335449,
+ "step": 1016
+ },
+ {
+ "epoch": 14.125764192139737,
+ "grad_norm": 0.02065849117934704,
+ "learning_rate": 0.0006,
+ "loss": 4.581960678100586,
+ "step": 1017
+ },
+ {
+ "epoch": 14.139737991266376,
+ "grad_norm": 0.01941298507153988,
+ "learning_rate": 0.0006,
+ "loss": 4.614226341247559,
+ "step": 1018
+ },
+ {
+ "epoch": 14.153711790393013,
+ "grad_norm": 0.022441856563091278,
+ "learning_rate": 0.0006,
+ "loss": 4.638352394104004,
+ "step": 1019
+ },
+ {
+ "epoch": 14.167685589519651,
+ "grad_norm": 0.021519694477319717,
+ "learning_rate": 0.0006,
+ "loss": 4.700160980224609,
+ "step": 1020
+ },
+ {
+ "epoch": 14.181659388646288,
+ "grad_norm": 0.018956486135721207,
+ "learning_rate": 0.0006,
+ "loss": 4.649593353271484,
+ "step": 1021
+ },
+ {
+ "epoch": 14.195633187772925,
+ "grad_norm": 0.016681140288710594,
+ "learning_rate": 0.0006,
+ "loss": 4.603952407836914,
+ "step": 1022
+ },
+ {
+ "epoch": 14.209606986899564,
+ "grad_norm": 0.01788809522986412,
+ "learning_rate": 0.0006,
+ "loss": 4.669160842895508,
+ "step": 1023
+ },
+ {
+ "epoch": 14.2235807860262,
+ "grad_norm": 0.016047228127717972,
+ "learning_rate": 0.0006,
+ "loss": 4.6367950439453125,
+ "step": 1024
+ },
+ {
+ "epoch": 14.237554585152838,
+ "grad_norm": 0.01614239625632763,
+ "learning_rate": 0.0006,
+ "loss": 4.664529323577881,
+ "step": 1025
+ },
+ {
+ "epoch": 14.251528384279476,
+ "grad_norm": 0.016362549737095833,
+ "learning_rate": 0.0006,
+ "loss": 4.702373504638672,
+ "step": 1026
+ },
+ {
+ "epoch": 14.265502183406113,
+ "grad_norm": 0.015800409018993378,
+ "learning_rate": 0.0006,
+ "loss": 4.7594428062438965,
+ "step": 1027
+ },
+ {
+ "epoch": 14.279475982532752,
+ "grad_norm": 0.016730139032006264,
+ "learning_rate": 0.0006,
+ "loss": 4.677244186401367,
+ "step": 1028
+ },
+ {
+ "epoch": 14.293449781659389,
+ "grad_norm": 0.01661759801208973,
+ "learning_rate": 0.0006,
+ "loss": 4.633679389953613,
+ "step": 1029
+ },
+ {
+ "epoch": 14.307423580786025,
+ "grad_norm": 0.017673347145318985,
+ "learning_rate": 0.0006,
+ "loss": 4.649107456207275,
+ "step": 1030
+ },
+ {
+ "epoch": 14.321397379912664,
+ "grad_norm": 0.019243352115154266,
+ "learning_rate": 0.0006,
+ "loss": 4.611257553100586,
+ "step": 1031
+ },
+ {
+ "epoch": 14.335371179039301,
+ "grad_norm": 0.019817164167761803,
+ "learning_rate": 0.0006,
+ "loss": 4.588086128234863,
+ "step": 1032
+ },
+ {
+ "epoch": 14.34934497816594,
+ "grad_norm": 0.022292211651802063,
+ "learning_rate": 0.0006,
+ "loss": 4.783066749572754,
+ "step": 1033
+ },
+ {
+ "epoch": 14.363318777292577,
+ "grad_norm": 0.02140096202492714,
+ "learning_rate": 0.0006,
+ "loss": 4.604718208312988,
+ "step": 1034
+ },
+ {
+ "epoch": 14.377292576419213,
+ "grad_norm": 0.0187817495316267,
+ "learning_rate": 0.0006,
+ "loss": 4.721773147583008,
+ "step": 1035
+ },
+ {
+ "epoch": 14.391266375545852,
+ "grad_norm": 0.019879210740327835,
+ "learning_rate": 0.0006,
+ "loss": 4.789286136627197,
+ "step": 1036
+ },
+ {
+ "epoch": 14.405240174672489,
+ "grad_norm": 0.01885785162448883,
+ "learning_rate": 0.0006,
+ "loss": 4.669743537902832,
+ "step": 1037
+ },
+ {
+ "epoch": 14.419213973799126,
+ "grad_norm": 0.018324270844459534,
+ "learning_rate": 0.0006,
+ "loss": 4.65901517868042,
+ "step": 1038
+ },
+ {
+ "epoch": 14.433187772925764,
+ "grad_norm": 0.017664821818470955,
+ "learning_rate": 0.0006,
+ "loss": 4.667839527130127,
+ "step": 1039
+ },
+ {
+ "epoch": 14.447161572052401,
+ "grad_norm": 0.015702905133366585,
+ "learning_rate": 0.0006,
+ "loss": 4.64700984954834,
+ "step": 1040
+ },
+ {
+ "epoch": 14.46113537117904,
+ "grad_norm": 0.016674194484949112,
+ "learning_rate": 0.0006,
+ "loss": 4.694093227386475,
+ "step": 1041
+ },
+ {
+ "epoch": 14.475109170305677,
+ "grad_norm": 0.016869477927684784,
+ "learning_rate": 0.0006,
+ "loss": 4.619544982910156,
+ "step": 1042
+ },
+ {
+ "epoch": 14.489082969432314,
+ "grad_norm": 0.016557540744543076,
+ "learning_rate": 0.0006,
+ "loss": 4.568667411804199,
+ "step": 1043
+ },
+ {
+ "epoch": 14.503056768558952,
+ "grad_norm": 0.014500945806503296,
+ "learning_rate": 0.0006,
+ "loss": 4.561741828918457,
+ "step": 1044
+ },
+ {
+ "epoch": 14.51703056768559,
+ "grad_norm": 0.014228739775717258,
+ "learning_rate": 0.0006,
+ "loss": 4.654277324676514,
+ "step": 1045
+ },
+ {
+ "epoch": 14.531004366812226,
+ "grad_norm": 0.013894238509237766,
+ "learning_rate": 0.0006,
+ "loss": 4.5867719650268555,
+ "step": 1046
+ },
+ {
+ "epoch": 14.544978165938865,
+ "grad_norm": 0.01550068985670805,
+ "learning_rate": 0.0006,
+ "loss": 4.61168098449707,
+ "step": 1047
+ },
+ {
+ "epoch": 14.558951965065502,
+ "grad_norm": 0.015940772369503975,
+ "learning_rate": 0.0006,
+ "loss": 4.691247940063477,
+ "step": 1048
+ },
+ {
+ "epoch": 14.57292576419214,
+ "grad_norm": 0.01714535616338253,
+ "learning_rate": 0.0006,
+ "loss": 4.61183500289917,
+ "step": 1049
+ },
+ {
+ "epoch": 14.586899563318777,
+ "grad_norm": 0.017925038933753967,
+ "learning_rate": 0.0006,
+ "loss": 4.719022750854492,
+ "step": 1050
+ },
+ {
+ "epoch": 14.600873362445414,
+ "grad_norm": 0.018141552805900574,
+ "learning_rate": 0.0006,
+ "loss": 4.760281562805176,
+ "step": 1051
+ },
+ {
+ "epoch": 14.614847161572053,
+ "grad_norm": 0.017131328582763672,
+ "learning_rate": 0.0006,
+ "loss": 4.709680080413818,
+ "step": 1052
+ },
+ {
+ "epoch": 14.62882096069869,
+ "grad_norm": 0.018730048090219498,
+ "learning_rate": 0.0006,
+ "loss": 4.5994720458984375,
+ "step": 1053
+ },
+ {
+ "epoch": 14.642794759825328,
+ "grad_norm": 0.018894298002123833,
+ "learning_rate": 0.0006,
+ "loss": 4.669070720672607,
+ "step": 1054
+ },
+ {
+ "epoch": 14.656768558951965,
+ "grad_norm": 0.017643198370933533,
+ "learning_rate": 0.0006,
+ "loss": 4.641790390014648,
+ "step": 1055
+ },
+ {
+ "epoch": 14.670742358078602,
+ "grad_norm": 0.017647244036197662,
+ "learning_rate": 0.0006,
+ "loss": 4.653899192810059,
+ "step": 1056
+ },
+ {
+ "epoch": 14.68471615720524,
+ "grad_norm": 0.017304198816418648,
+ "learning_rate": 0.0006,
+ "loss": 4.7487473487854,
+ "step": 1057
+ },
+ {
+ "epoch": 14.698689956331878,
+ "grad_norm": 0.01638447679579258,
+ "learning_rate": 0.0006,
+ "loss": 4.71945858001709,
+ "step": 1058
+ },
+ {
+ "epoch": 14.712663755458514,
+ "grad_norm": 0.015892071649432182,
+ "learning_rate": 0.0006,
+ "loss": 4.744903564453125,
+ "step": 1059
+ },
+ {
+ "epoch": 14.726637554585153,
+ "grad_norm": 0.017420614138245583,
+ "learning_rate": 0.0006,
+ "loss": 4.781563758850098,
+ "step": 1060
+ },
+ {
+ "epoch": 14.74061135371179,
+ "grad_norm": 0.016498520970344543,
+ "learning_rate": 0.0006,
+ "loss": 4.622779846191406,
+ "step": 1061
+ },
+ {
+ "epoch": 14.754585152838429,
+ "grad_norm": 0.017482127994298935,
+ "learning_rate": 0.0006,
+ "loss": 4.62825870513916,
+ "step": 1062
+ },
+ {
+ "epoch": 14.768558951965066,
+ "grad_norm": 0.019018379971385002,
+ "learning_rate": 0.0006,
+ "loss": 4.693194389343262,
+ "step": 1063
+ },
+ {
+ "epoch": 14.782532751091702,
+ "grad_norm": 0.01993914321064949,
+ "learning_rate": 0.0006,
+ "loss": 4.742903709411621,
+ "step": 1064
+ },
+ {
+ "epoch": 14.796506550218341,
+ "grad_norm": 0.02301335707306862,
+ "learning_rate": 0.0006,
+ "loss": 4.613251686096191,
+ "step": 1065
+ },
+ {
+ "epoch": 14.810480349344978,
+ "grad_norm": 0.024088222533464432,
+ "learning_rate": 0.0006,
+ "loss": 4.594305992126465,
+ "step": 1066
+ },
+ {
+ "epoch": 14.824454148471617,
+ "grad_norm": 0.02493320405483246,
+ "learning_rate": 0.0006,
+ "loss": 4.700588226318359,
+ "step": 1067
+ },
+ {
+ "epoch": 14.838427947598253,
+ "grad_norm": 0.02501937560737133,
+ "learning_rate": 0.0006,
+ "loss": 4.601742744445801,
+ "step": 1068
+ },
+ {
+ "epoch": 14.85240174672489,
+ "grad_norm": 0.023490751162171364,
+ "learning_rate": 0.0006,
+ "loss": 4.589211463928223,
+ "step": 1069
+ },
+ {
+ "epoch": 14.866375545851529,
+ "grad_norm": 0.019695665687322617,
+ "learning_rate": 0.0006,
+ "loss": 4.635081768035889,
+ "step": 1070
+ },
+ {
+ "epoch": 14.880349344978166,
+ "grad_norm": 0.019276980310678482,
+ "learning_rate": 0.0006,
+ "loss": 4.737002372741699,
+ "step": 1071
+ },
+ {
+ "epoch": 14.894323144104803,
+ "grad_norm": 0.022430511191487312,
+ "learning_rate": 0.0006,
+ "loss": 4.808295726776123,
+ "step": 1072
+ },
+ {
+ "epoch": 14.908296943231441,
+ "grad_norm": 0.021543916314840317,
+ "learning_rate": 0.0006,
+ "loss": 4.559475421905518,
+ "step": 1073
+ },
+ {
+ "epoch": 14.922270742358078,
+ "grad_norm": 0.019885241985321045,
+ "learning_rate": 0.0006,
+ "loss": 4.647948741912842,
+ "step": 1074
+ },
+ {
+ "epoch": 14.936244541484717,
+ "grad_norm": 0.02115175500512123,
+ "learning_rate": 0.0006,
+ "loss": 4.680505752563477,
+ "step": 1075
+ },
+ {
+ "epoch": 14.950218340611354,
+ "grad_norm": 0.02156151458621025,
+ "learning_rate": 0.0006,
+ "loss": 4.673477649688721,
+ "step": 1076
+ },
+ {
+ "epoch": 14.96419213973799,
+ "grad_norm": 0.02184601128101349,
+ "learning_rate": 0.0006,
+ "loss": 4.644402027130127,
+ "step": 1077
+ },
+ {
+ "epoch": 14.97816593886463,
+ "grad_norm": 0.0211983360350132,
+ "learning_rate": 0.0006,
+ "loss": 4.821558475494385,
+ "step": 1078
+ },
+ {
+ "epoch": 14.992139737991266,
+ "grad_norm": 0.019636567682027817,
+ "learning_rate": 0.0006,
+ "loss": 4.578868865966797,
+ "step": 1079
+ },
+ {
+ "epoch": 15.0,
+ "grad_norm": 0.019997350871562958,
+ "learning_rate": 0.0006,
+ "loss": 4.691307544708252,
+ "step": 1080
+ },
+ {
+ "epoch": 15.0,
+ "eval_loss": 4.872478008270264,
+ "eval_runtime": 56.8513,
+ "eval_samples_per_second": 42.954,
+ "eval_steps_per_second": 1.354,
+ "step": 1080
+ },
+ {
+ "epoch": 15.013973799126637,
+ "grad_norm": 0.018830662593245506,
+ "learning_rate": 0.0006,
+ "loss": 4.761779308319092,
+ "step": 1081
+ },
+ {
+ "epoch": 15.027947598253276,
+ "grad_norm": 0.016876481473445892,
+ "learning_rate": 0.0006,
+ "loss": 4.61148738861084,
+ "step": 1082
+ },
+ {
+ "epoch": 15.041921397379912,
+ "grad_norm": 0.015550863929092884,
+ "learning_rate": 0.0006,
+ "loss": 4.676124095916748,
+ "step": 1083
+ },
+ {
+ "epoch": 15.055895196506551,
+ "grad_norm": 0.017178896814584732,
+ "learning_rate": 0.0006,
+ "loss": 4.588935852050781,
+ "step": 1084
+ },
+ {
+ "epoch": 15.069868995633188,
+ "grad_norm": 0.015465447679162025,
+ "learning_rate": 0.0006,
+ "loss": 4.605499267578125,
+ "step": 1085
+ },
+ {
+ "epoch": 15.083842794759825,
+ "grad_norm": 0.01606195978820324,
+ "learning_rate": 0.0006,
+ "loss": 4.723373889923096,
+ "step": 1086
+ },
+ {
+ "epoch": 15.097816593886463,
+ "grad_norm": 0.016521936282515526,
+ "learning_rate": 0.0006,
+ "loss": 4.646608352661133,
+ "step": 1087
+ },
+ {
+ "epoch": 15.1117903930131,
+ "grad_norm": 0.01589205674827099,
+ "learning_rate": 0.0006,
+ "loss": 4.498138427734375,
+ "step": 1088
+ },
+ {
+ "epoch": 15.125764192139737,
+ "grad_norm": 0.01524856686592102,
+ "learning_rate": 0.0006,
+ "loss": 4.594562530517578,
+ "step": 1089
+ },
+ {
+ "epoch": 15.139737991266376,
+ "grad_norm": 0.014714641496539116,
+ "learning_rate": 0.0006,
+ "loss": 4.630744934082031,
+ "step": 1090
+ },
+ {
+ "epoch": 15.153711790393013,
+ "grad_norm": 0.015333331190049648,
+ "learning_rate": 0.0006,
+ "loss": 4.682388782501221,
+ "step": 1091
+ },
+ {
+ "epoch": 15.167685589519651,
+ "grad_norm": 0.01550073828548193,
+ "learning_rate": 0.0006,
+ "loss": 4.692652702331543,
+ "step": 1092
+ },
+ {
+ "epoch": 15.181659388646288,
+ "grad_norm": 0.014783318154513836,
+ "learning_rate": 0.0006,
+ "loss": 4.6223835945129395,
+ "step": 1093
+ },
+ {
+ "epoch": 15.195633187772925,
+ "grad_norm": 0.014372751116752625,
+ "learning_rate": 0.0006,
+ "loss": 4.52687406539917,
+ "step": 1094
+ },
+ {
+ "epoch": 15.209606986899564,
+ "grad_norm": 0.014779130928218365,
+ "learning_rate": 0.0006,
+ "loss": 4.637009620666504,
+ "step": 1095
+ },
+ {
+ "epoch": 15.2235807860262,
+ "grad_norm": 0.015697073191404343,
+ "learning_rate": 0.0006,
+ "loss": 4.682981014251709,
+ "step": 1096
+ },
+ {
+ "epoch": 15.237554585152838,
+ "grad_norm": 0.016523664817214012,
+ "learning_rate": 0.0006,
+ "loss": 4.705722808837891,
+ "step": 1097
+ },
+ {
+ "epoch": 15.251528384279476,
+ "grad_norm": 0.017633996903896332,
+ "learning_rate": 0.0006,
+ "loss": 4.574686050415039,
+ "step": 1098
+ },
+ {
+ "epoch": 15.265502183406113,
+ "grad_norm": 0.019955076277256012,
+ "learning_rate": 0.0006,
+ "loss": 4.634450912475586,
+ "step": 1099
+ },
+ {
+ "epoch": 15.279475982532752,
+ "grad_norm": 0.02132618986070156,
+ "learning_rate": 0.0006,
+ "loss": 4.56110143661499,
+ "step": 1100
+ },
+ {
+ "epoch": 15.293449781659389,
+ "grad_norm": 0.0207593385130167,
+ "learning_rate": 0.0006,
+ "loss": 4.629396438598633,
+ "step": 1101
+ },
+ {
+ "epoch": 15.307423580786025,
+ "grad_norm": 0.017882540822029114,
+ "learning_rate": 0.0006,
+ "loss": 4.730963230133057,
+ "step": 1102
+ },
+ {
+ "epoch": 15.321397379912664,
+ "grad_norm": 0.01917876861989498,
+ "learning_rate": 0.0006,
+ "loss": 4.64585542678833,
+ "step": 1103
+ },
+ {
+ "epoch": 15.335371179039301,
+ "grad_norm": 0.01943974196910858,
+ "learning_rate": 0.0006,
+ "loss": 4.656181335449219,
+ "step": 1104
+ },
+ {
+ "epoch": 15.34934497816594,
+ "grad_norm": 0.019904715940356255,
+ "learning_rate": 0.0006,
+ "loss": 4.619329452514648,
+ "step": 1105
+ },
+ {
+ "epoch": 15.363318777292577,
+ "grad_norm": 0.019174346700310707,
+ "learning_rate": 0.0006,
+ "loss": 4.647892475128174,
+ "step": 1106
+ },
+ {
+ "epoch": 15.377292576419213,
+ "grad_norm": 0.018244758248329163,
+ "learning_rate": 0.0006,
+ "loss": 4.56166934967041,
+ "step": 1107
+ },
+ {
+ "epoch": 15.391266375545852,
+ "grad_norm": 0.018914205953478813,
+ "learning_rate": 0.0006,
+ "loss": 4.648063659667969,
+ "step": 1108
+ },
+ {
+ "epoch": 15.405240174672489,
+ "grad_norm": 0.01961071416735649,
+ "learning_rate": 0.0006,
+ "loss": 4.665525436401367,
+ "step": 1109
+ },
+ {
+ "epoch": 15.419213973799126,
+ "grad_norm": 0.01841311901807785,
+ "learning_rate": 0.0006,
+ "loss": 4.611342430114746,
+ "step": 1110
+ },
+ {
+ "epoch": 15.433187772925764,
+ "grad_norm": 0.01863427273929119,
+ "learning_rate": 0.0006,
+ "loss": 4.664256572723389,
+ "step": 1111
+ },
+ {
+ "epoch": 15.447161572052401,
+ "grad_norm": 0.020506877452135086,
+ "learning_rate": 0.0006,
+ "loss": 4.631237506866455,
+ "step": 1112
+ },
+ {
+ "epoch": 15.46113537117904,
+ "grad_norm": 0.02223833277821541,
+ "learning_rate": 0.0006,
+ "loss": 4.668654441833496,
+ "step": 1113
+ },
+ {
+ "epoch": 15.475109170305677,
+ "grad_norm": 0.023336702957749367,
+ "learning_rate": 0.0006,
+ "loss": 4.562833309173584,
+ "step": 1114
+ },
+ {
+ "epoch": 15.489082969432314,
+ "grad_norm": 0.023052413016557693,
+ "learning_rate": 0.0006,
+ "loss": 4.621532917022705,
+ "step": 1115
+ },
+ {
+ "epoch": 15.503056768558952,
+ "grad_norm": 0.023420924320816994,
+ "learning_rate": 0.0006,
+ "loss": 4.665465354919434,
+ "step": 1116
+ },
+ {
+ "epoch": 15.51703056768559,
+ "grad_norm": 0.024071281775832176,
+ "learning_rate": 0.0006,
+ "loss": 4.595980644226074,
+ "step": 1117
+ },
+ {
+ "epoch": 15.531004366812226,
+ "grad_norm": 0.023406682536005974,
+ "learning_rate": 0.0006,
+ "loss": 4.642635345458984,
+ "step": 1118
+ },
+ {
+ "epoch": 15.544978165938865,
+ "grad_norm": 0.02138647809624672,
+ "learning_rate": 0.0006,
+ "loss": 4.704513072967529,
+ "step": 1119
+ },
+ {
+ "epoch": 15.558951965065502,
+ "grad_norm": 0.0221384447067976,
+ "learning_rate": 0.0006,
+ "loss": 4.567503452301025,
+ "step": 1120
+ },
+ {
+ "epoch": 15.57292576419214,
+ "grad_norm": 0.020879851654171944,
+ "learning_rate": 0.0006,
+ "loss": 4.664057731628418,
+ "step": 1121
+ },
+ {
+ "epoch": 15.586899563318777,
+ "grad_norm": 0.021478967741131783,
+ "learning_rate": 0.0006,
+ "loss": 4.539627552032471,
+ "step": 1122
+ },
+ {
+ "epoch": 15.600873362445414,
+ "grad_norm": 0.022748656570911407,
+ "learning_rate": 0.0006,
+ "loss": 4.5654497146606445,
+ "step": 1123
+ },
+ {
+ "epoch": 15.614847161572053,
+ "grad_norm": 0.021870248019695282,
+ "learning_rate": 0.0006,
+ "loss": 4.56899356842041,
+ "step": 1124
+ },
+ {
+ "epoch": 15.62882096069869,
+ "grad_norm": 0.021711355075240135,
+ "learning_rate": 0.0006,
+ "loss": 4.657808303833008,
+ "step": 1125
+ },
+ {
+ "epoch": 15.642794759825328,
+ "grad_norm": 0.022810909897089005,
+ "learning_rate": 0.0006,
+ "loss": 4.6526689529418945,
+ "step": 1126
+ },
+ {
+ "epoch": 15.656768558951965,
+ "grad_norm": 0.022561343386769295,
+ "learning_rate": 0.0006,
+ "loss": 4.534887790679932,
+ "step": 1127
+ },
+ {
+ "epoch": 15.670742358078602,
+ "grad_norm": 0.019909802824258804,
+ "learning_rate": 0.0006,
+ "loss": 4.600790023803711,
+ "step": 1128
+ },
+ {
+ "epoch": 15.68471615720524,
+ "grad_norm": 0.020092325285077095,
+ "learning_rate": 0.0006,
+ "loss": 4.773830413818359,
+ "step": 1129
+ },
+ {
+ "epoch": 15.698689956331878,
+ "grad_norm": 0.018834445625543594,
+ "learning_rate": 0.0006,
+ "loss": 4.628868103027344,
+ "step": 1130
+ },
+ {
+ "epoch": 15.712663755458514,
+ "grad_norm": 0.017552798613905907,
+ "learning_rate": 0.0006,
+ "loss": 4.647500038146973,
+ "step": 1131
+ },
+ {
+ "epoch": 15.726637554585153,
+ "grad_norm": 0.01751251146197319,
+ "learning_rate": 0.0006,
+ "loss": 4.590545654296875,
+ "step": 1132
+ },
+ {
+ "epoch": 15.74061135371179,
+ "grad_norm": 0.01699099875986576,
+ "learning_rate": 0.0006,
+ "loss": 4.509699821472168,
+ "step": 1133
+ },
+ {
+ "epoch": 15.754585152838429,
+ "grad_norm": 0.015808649361133575,
+ "learning_rate": 0.0006,
+ "loss": 4.606992244720459,
+ "step": 1134
+ },
+ {
+ "epoch": 15.768558951965066,
+ "grad_norm": 0.01655443385243416,
+ "learning_rate": 0.0006,
+ "loss": 4.585219860076904,
+ "step": 1135
+ },
+ {
+ "epoch": 15.782532751091702,
+ "grad_norm": 0.01708163693547249,
+ "learning_rate": 0.0006,
+ "loss": 4.668147087097168,
+ "step": 1136
+ },
+ {
+ "epoch": 15.796506550218341,
+ "grad_norm": 0.017161913216114044,
+ "learning_rate": 0.0006,
+ "loss": 4.7170634269714355,
+ "step": 1137
+ },
+ {
+ "epoch": 15.810480349344978,
+ "grad_norm": 0.017262674868106842,
+ "learning_rate": 0.0006,
+ "loss": 4.572132110595703,
+ "step": 1138
+ },
+ {
+ "epoch": 15.824454148471617,
+ "grad_norm": 0.017713576555252075,
+ "learning_rate": 0.0006,
+ "loss": 4.702538967132568,
+ "step": 1139
+ },
+ {
+ "epoch": 15.838427947598253,
+ "grad_norm": 0.018730733543634415,
+ "learning_rate": 0.0006,
+ "loss": 4.629721641540527,
+ "step": 1140
+ },
+ {
+ "epoch": 15.85240174672489,
+ "grad_norm": 0.017960187047719955,
+ "learning_rate": 0.0006,
+ "loss": 4.661039352416992,
+ "step": 1141
+ },
+ {
+ "epoch": 15.866375545851529,
+ "grad_norm": 0.01750079356133938,
+ "learning_rate": 0.0006,
+ "loss": 4.503710746765137,
+ "step": 1142
+ },
+ {
+ "epoch": 15.880349344978166,
+ "grad_norm": 0.016630422323942184,
+ "learning_rate": 0.0006,
+ "loss": 4.6582136154174805,
+ "step": 1143
+ },
+ {
+ "epoch": 15.894323144104803,
+ "grad_norm": 0.015896832570433617,
+ "learning_rate": 0.0006,
+ "loss": 4.596693992614746,
+ "step": 1144
+ },
+ {
+ "epoch": 15.908296943231441,
+ "grad_norm": 0.01566407084465027,
+ "learning_rate": 0.0006,
+ "loss": 4.6337199211120605,
+ "step": 1145
+ },
+ {
+ "epoch": 15.922270742358078,
+ "grad_norm": 0.016823330894112587,
+ "learning_rate": 0.0006,
+ "loss": 4.686994552612305,
+ "step": 1146
+ },
+ {
+ "epoch": 15.936244541484717,
+ "grad_norm": 0.016208255663514137,
+ "learning_rate": 0.0006,
+ "loss": 4.6205549240112305,
+ "step": 1147
+ },
+ {
+ "epoch": 15.950218340611354,
+ "grad_norm": 0.0171707421541214,
+ "learning_rate": 0.0006,
+ "loss": 4.582968711853027,
+ "step": 1148
+ },
+ {
+ "epoch": 15.96419213973799,
+ "grad_norm": 0.017009446397423744,
+ "learning_rate": 0.0006,
+ "loss": 4.510274887084961,
+ "step": 1149
+ },
+ {
+ "epoch": 15.97816593886463,
+ "grad_norm": 0.01565811224281788,
+ "learning_rate": 0.0006,
+ "loss": 4.55834436416626,
+ "step": 1150
+ },
+ {
+ "epoch": 15.992139737991266,
+ "grad_norm": 0.015276037156581879,
+ "learning_rate": 0.0006,
+ "loss": 4.537870407104492,
+ "step": 1151
+ },
+ {
+ "epoch": 16.0,
+ "grad_norm": 0.017305882647633553,
+ "learning_rate": 0.0006,
+ "loss": 4.611305236816406,
+ "step": 1152
+ },
+ {
+ "epoch": 16.0,
+ "eval_loss": 4.775513172149658,
+ "eval_runtime": 56.7341,
+ "eval_samples_per_second": 43.043,
+ "eval_steps_per_second": 1.357,
+ "step": 1152
+ },
+ {
+ "epoch": 16.01397379912664,
+ "grad_norm": 0.016752436757087708,
+ "learning_rate": 0.0006,
+ "loss": 4.541312217712402,
+ "step": 1153
+ },
+ {
+ "epoch": 16.027947598253274,
+ "grad_norm": 0.01643134094774723,
+ "learning_rate": 0.0006,
+ "loss": 4.585402965545654,
+ "step": 1154
+ },
+ {
+ "epoch": 16.041921397379912,
+ "grad_norm": 0.01888720504939556,
+ "learning_rate": 0.0006,
+ "loss": 4.597825050354004,
+ "step": 1155
+ },
+ {
+ "epoch": 16.05589519650655,
+ "grad_norm": 0.022044632583856583,
+ "learning_rate": 0.0006,
+ "loss": 4.532772064208984,
+ "step": 1156
+ },
+ {
+ "epoch": 16.069868995633186,
+ "grad_norm": 0.022986987605690956,
+ "learning_rate": 0.0006,
+ "loss": 4.551351547241211,
+ "step": 1157
+ },
+ {
+ "epoch": 16.083842794759825,
+ "grad_norm": 0.020116791129112244,
+ "learning_rate": 0.0006,
+ "loss": 4.484149932861328,
+ "step": 1158
+ },
+ {
+ "epoch": 16.097816593886463,
+ "grad_norm": 0.021442247554659843,
+ "learning_rate": 0.0006,
+ "loss": 4.515316486358643,
+ "step": 1159
+ },
+ {
+ "epoch": 16.111790393013102,
+ "grad_norm": 0.020974772050976753,
+ "learning_rate": 0.0006,
+ "loss": 4.51606559753418,
+ "step": 1160
+ },
+ {
+ "epoch": 16.125764192139737,
+ "grad_norm": 0.021274864673614502,
+ "learning_rate": 0.0006,
+ "loss": 4.587329864501953,
+ "step": 1161
+ },
+ {
+ "epoch": 16.139737991266376,
+ "grad_norm": 0.022799784317612648,
+ "learning_rate": 0.0006,
+ "loss": 4.568149089813232,
+ "step": 1162
+ },
+ {
+ "epoch": 16.153711790393015,
+ "grad_norm": 0.021784933283925056,
+ "learning_rate": 0.0006,
+ "loss": 4.515336990356445,
+ "step": 1163
+ },
+ {
+ "epoch": 16.16768558951965,
+ "grad_norm": 0.024261271581053734,
+ "learning_rate": 0.0006,
+ "loss": 4.5127081871032715,
+ "step": 1164
+ },
+ {
+ "epoch": 16.18165938864629,
+ "grad_norm": 0.02248375490307808,
+ "learning_rate": 0.0006,
+ "loss": 4.754251003265381,
+ "step": 1165
+ },
+ {
+ "epoch": 16.195633187772927,
+ "grad_norm": 0.02136492356657982,
+ "learning_rate": 0.0006,
+ "loss": 4.58193826675415,
+ "step": 1166
+ },
+ {
+ "epoch": 16.209606986899562,
+ "grad_norm": 0.020727066323161125,
+ "learning_rate": 0.0006,
+ "loss": 4.564209938049316,
+ "step": 1167
+ },
+ {
+ "epoch": 16.2235807860262,
+ "grad_norm": 0.017789119854569435,
+ "learning_rate": 0.0006,
+ "loss": 4.698680877685547,
+ "step": 1168
+ },
+ {
+ "epoch": 16.23755458515284,
+ "grad_norm": 0.017603034153580666,
+ "learning_rate": 0.0006,
+ "loss": 4.501171588897705,
+ "step": 1169
+ },
+ {
+ "epoch": 16.251528384279474,
+ "grad_norm": 0.017345862463116646,
+ "learning_rate": 0.0006,
+ "loss": 4.593590259552002,
+ "step": 1170
+ },
+ {
+ "epoch": 16.265502183406113,
+ "grad_norm": 0.018047425895929337,
+ "learning_rate": 0.0006,
+ "loss": 4.597467422485352,
+ "step": 1171
+ },
+ {
+ "epoch": 16.27947598253275,
+ "grad_norm": 0.01947222836315632,
+ "learning_rate": 0.0006,
+ "loss": 4.5072221755981445,
+ "step": 1172
+ },
+ {
+ "epoch": 16.29344978165939,
+ "grad_norm": 0.019598500803112984,
+ "learning_rate": 0.0006,
+ "loss": 4.601351737976074,
+ "step": 1173
+ },
+ {
+ "epoch": 16.307423580786025,
+ "grad_norm": 0.020002514123916626,
+ "learning_rate": 0.0006,
+ "loss": 4.580431938171387,
+ "step": 1174
+ },
+ {
+ "epoch": 16.321397379912664,
+ "grad_norm": 0.0216530729085207,
+ "learning_rate": 0.0006,
+ "loss": 4.424681663513184,
+ "step": 1175
+ },
+ {
+ "epoch": 16.335371179039303,
+ "grad_norm": 0.01949235238134861,
+ "learning_rate": 0.0006,
+ "loss": 4.751883029937744,
+ "step": 1176
+ },
+ {
+ "epoch": 16.349344978165938,
+ "grad_norm": 0.017779016867280006,
+ "learning_rate": 0.0006,
+ "loss": 4.541490077972412,
+ "step": 1177
+ },
+ {
+ "epoch": 16.363318777292577,
+ "grad_norm": 0.01909121684730053,
+ "learning_rate": 0.0006,
+ "loss": 4.594141006469727,
+ "step": 1178
+ },
+ {
+ "epoch": 16.377292576419215,
+ "grad_norm": 0.019894849509000778,
+ "learning_rate": 0.0006,
+ "loss": 4.555070400238037,
+ "step": 1179
+ },
+ {
+ "epoch": 16.39126637554585,
+ "grad_norm": 0.01993032731115818,
+ "learning_rate": 0.0006,
+ "loss": 4.521401405334473,
+ "step": 1180
+ },
+ {
+ "epoch": 16.40524017467249,
+ "grad_norm": 0.0210788045078516,
+ "learning_rate": 0.0006,
+ "loss": 4.672842979431152,
+ "step": 1181
+ },
+ {
+ "epoch": 16.419213973799128,
+ "grad_norm": 0.02274598926305771,
+ "learning_rate": 0.0006,
+ "loss": 4.629504203796387,
+ "step": 1182
+ },
+ {
+ "epoch": 16.433187772925763,
+ "grad_norm": 0.020448120310902596,
+ "learning_rate": 0.0006,
+ "loss": 4.548209190368652,
+ "step": 1183
+ },
+ {
+ "epoch": 16.4471615720524,
+ "grad_norm": 0.017565643414855003,
+ "learning_rate": 0.0006,
+ "loss": 4.529054641723633,
+ "step": 1184
+ },
+ {
+ "epoch": 16.46113537117904,
+ "grad_norm": 0.019504237920045853,
+ "learning_rate": 0.0006,
+ "loss": 4.48182487487793,
+ "step": 1185
+ },
+ {
+ "epoch": 16.475109170305675,
+ "grad_norm": 0.016823742538690567,
+ "learning_rate": 0.0006,
+ "loss": 4.608696937561035,
+ "step": 1186
+ },
+ {
+ "epoch": 16.489082969432314,
+ "grad_norm": 0.01602894812822342,
+ "learning_rate": 0.0006,
+ "loss": 4.620062351226807,
+ "step": 1187
+ },
+ {
+ "epoch": 16.503056768558952,
+ "grad_norm": 0.016615642234683037,
+ "learning_rate": 0.0006,
+ "loss": 4.69378662109375,
+ "step": 1188
+ },
+ {
+ "epoch": 16.51703056768559,
+ "grad_norm": 0.018576730042696,
+ "learning_rate": 0.0006,
+ "loss": 4.537041664123535,
+ "step": 1189
+ },
+ {
+ "epoch": 16.531004366812226,
+ "grad_norm": 0.018611151725053787,
+ "learning_rate": 0.0006,
+ "loss": 4.496169090270996,
+ "step": 1190
+ },
+ {
+ "epoch": 16.544978165938865,
+ "grad_norm": 0.018622413277626038,
+ "learning_rate": 0.0006,
+ "loss": 4.557358264923096,
+ "step": 1191
+ },
+ {
+ "epoch": 16.558951965065503,
+ "grad_norm": 0.01823544315993786,
+ "learning_rate": 0.0006,
+ "loss": 4.730792999267578,
+ "step": 1192
+ },
+ {
+ "epoch": 16.57292576419214,
+ "grad_norm": 0.018796846270561218,
+ "learning_rate": 0.0006,
+ "loss": 4.781791687011719,
+ "step": 1193
+ },
+ {
+ "epoch": 16.586899563318777,
+ "grad_norm": 0.01728292554616928,
+ "learning_rate": 0.0006,
+ "loss": 4.515257835388184,
+ "step": 1194
+ },
+ {
+ "epoch": 16.600873362445416,
+ "grad_norm": 0.01699589006602764,
+ "learning_rate": 0.0006,
+ "loss": 4.5996904373168945,
+ "step": 1195
+ },
+ {
+ "epoch": 16.61484716157205,
+ "grad_norm": 0.015905970707535744,
+ "learning_rate": 0.0006,
+ "loss": 4.601301193237305,
+ "step": 1196
+ },
+ {
+ "epoch": 16.62882096069869,
+ "grad_norm": 0.01338571310043335,
+ "learning_rate": 0.0006,
+ "loss": 4.557708740234375,
+ "step": 1197
+ },
+ {
+ "epoch": 16.64279475982533,
+ "grad_norm": 0.015298855490982533,
+ "learning_rate": 0.0006,
+ "loss": 4.542287349700928,
+ "step": 1198
+ },
+ {
+ "epoch": 16.656768558951963,
+ "grad_norm": 0.014753367751836777,
+ "learning_rate": 0.0006,
+ "loss": 4.595787048339844,
+ "step": 1199
+ },
+ {
+ "epoch": 16.670742358078602,
+ "grad_norm": 0.015527483075857162,
+ "learning_rate": 0.0006,
+ "loss": 4.607529163360596,
+ "step": 1200
+ },
+ {
+ "epoch": 16.68471615720524,
+ "grad_norm": 0.015042162500321865,
+ "learning_rate": 0.0006,
+ "loss": 4.686626434326172,
+ "step": 1201
+ },
+ {
+ "epoch": 16.69868995633188,
+ "grad_norm": 0.0159393772482872,
+ "learning_rate": 0.0006,
+ "loss": 4.47656774520874,
+ "step": 1202
+ },
+ {
+ "epoch": 16.712663755458514,
+ "grad_norm": 0.01532787922769785,
+ "learning_rate": 0.0006,
+ "loss": 4.55678653717041,
+ "step": 1203
+ },
+ {
+ "epoch": 16.726637554585153,
+ "grad_norm": 0.016332164406776428,
+ "learning_rate": 0.0006,
+ "loss": 4.51441764831543,
+ "step": 1204
+ },
+ {
+ "epoch": 16.74061135371179,
+ "grad_norm": 0.018901726230978966,
+ "learning_rate": 0.0006,
+ "loss": 4.57227897644043,
+ "step": 1205
+ },
+ {
+ "epoch": 16.754585152838427,
+ "grad_norm": 0.018589433282613754,
+ "learning_rate": 0.0006,
+ "loss": 4.550070762634277,
+ "step": 1206
+ },
+ {
+ "epoch": 16.768558951965066,
+ "grad_norm": 0.01634034886956215,
+ "learning_rate": 0.0006,
+ "loss": 4.586910724639893,
+ "step": 1207
+ },
+ {
+ "epoch": 16.782532751091704,
+ "grad_norm": 0.017599433660507202,
+ "learning_rate": 0.0006,
+ "loss": 4.5891923904418945,
+ "step": 1208
+ },
+ {
+ "epoch": 16.79650655021834,
+ "grad_norm": 0.01878446340560913,
+ "learning_rate": 0.0006,
+ "loss": 4.57227087020874,
+ "step": 1209
+ },
+ {
+ "epoch": 16.810480349344978,
+ "grad_norm": 0.01803469844162464,
+ "learning_rate": 0.0006,
+ "loss": 4.583845138549805,
+ "step": 1210
+ },
+ {
+ "epoch": 16.824454148471617,
+ "grad_norm": 0.018505144864320755,
+ "learning_rate": 0.0006,
+ "loss": 4.561151504516602,
+ "step": 1211
+ },
+ {
+ "epoch": 16.83842794759825,
+ "grad_norm": 0.01920081488788128,
+ "learning_rate": 0.0006,
+ "loss": 4.506834030151367,
+ "step": 1212
+ },
+ {
+ "epoch": 16.85240174672489,
+ "grad_norm": 0.0177712831646204,
+ "learning_rate": 0.0006,
+ "loss": 4.557195663452148,
+ "step": 1213
+ },
+ {
+ "epoch": 16.86637554585153,
+ "grad_norm": 0.019113918766379356,
+ "learning_rate": 0.0006,
+ "loss": 4.651227951049805,
+ "step": 1214
+ },
+ {
+ "epoch": 16.880349344978168,
+ "grad_norm": 0.02012539468705654,
+ "learning_rate": 0.0006,
+ "loss": 4.4952239990234375,
+ "step": 1215
+ },
+ {
+ "epoch": 16.894323144104803,
+ "grad_norm": 0.019993646070361137,
+ "learning_rate": 0.0006,
+ "loss": 4.517433166503906,
+ "step": 1216
+ },
+ {
+ "epoch": 16.90829694323144,
+ "grad_norm": 0.020041441544890404,
+ "learning_rate": 0.0006,
+ "loss": 4.489407539367676,
+ "step": 1217
+ },
+ {
+ "epoch": 16.92227074235808,
+ "grad_norm": 0.0183611661195755,
+ "learning_rate": 0.0006,
+ "loss": 4.5435638427734375,
+ "step": 1218
+ },
+ {
+ "epoch": 16.936244541484715,
+ "grad_norm": 0.019138429313898087,
+ "learning_rate": 0.0006,
+ "loss": 4.545949459075928,
+ "step": 1219
+ },
+ {
+ "epoch": 16.950218340611354,
+ "grad_norm": 0.018554266542196274,
+ "learning_rate": 0.0006,
+ "loss": 4.579620361328125,
+ "step": 1220
+ },
+ {
+ "epoch": 16.964192139737992,
+ "grad_norm": 0.0193545650690794,
+ "learning_rate": 0.0006,
+ "loss": 4.588255405426025,
+ "step": 1221
+ },
+ {
+ "epoch": 16.978165938864628,
+ "grad_norm": 0.019389165565371513,
+ "learning_rate": 0.0006,
+ "loss": 4.520698547363281,
+ "step": 1222
+ },
+ {
+ "epoch": 16.992139737991266,
+ "grad_norm": 0.016601379960775375,
+ "learning_rate": 0.0006,
+ "loss": 4.4809064865112305,
+ "step": 1223
+ },
+ {
+ "epoch": 17.0,
+ "grad_norm": 0.01771303080022335,
+ "learning_rate": 0.0006,
+ "loss": 4.721122741699219,
+ "step": 1224
+ },
+ {
+ "epoch": 17.0,
+ "eval_loss": 4.7903666496276855,
+ "eval_runtime": 56.9826,
+ "eval_samples_per_second": 42.855,
+ "eval_steps_per_second": 1.351,
+ "step": 1224
+ },
+ {
+ "epoch": 17.01397379912664,
+ "grad_norm": 0.017720289528369904,
+ "learning_rate": 0.0006,
+ "loss": 4.523834705352783,
+ "step": 1225
+ },
+ {
+ "epoch": 17.027947598253274,
+ "grad_norm": 0.020137161016464233,
+ "learning_rate": 0.0006,
+ "loss": 4.574061870574951,
+ "step": 1226
+ },
+ {
+ "epoch": 17.041921397379912,
+ "grad_norm": 0.02002314105629921,
+ "learning_rate": 0.0006,
+ "loss": 4.512407302856445,
+ "step": 1227
+ },
+ {
+ "epoch": 17.05589519650655,
+ "grad_norm": 0.019628183916211128,
+ "learning_rate": 0.0006,
+ "loss": 4.502533435821533,
+ "step": 1228
+ },
+ {
+ "epoch": 17.069868995633186,
+ "grad_norm": 0.019808458164334297,
+ "learning_rate": 0.0006,
+ "loss": 4.463156700134277,
+ "step": 1229
+ },
+ {
+ "epoch": 17.083842794759825,
+ "grad_norm": 0.021250786259770393,
+ "learning_rate": 0.0006,
+ "loss": 4.622682094573975,
+ "step": 1230
+ },
+ {
+ "epoch": 17.097816593886463,
+ "grad_norm": 0.022031694650650024,
+ "learning_rate": 0.0006,
+ "loss": 4.451882839202881,
+ "step": 1231
+ },
+ {
+ "epoch": 17.111790393013102,
+ "grad_norm": 0.01898987777531147,
+ "learning_rate": 0.0006,
+ "loss": 4.529331684112549,
+ "step": 1232
+ },
+ {
+ "epoch": 17.125764192139737,
+ "grad_norm": 0.018749834969639778,
+ "learning_rate": 0.0006,
+ "loss": 4.435815334320068,
+ "step": 1233
+ },
+ {
+ "epoch": 17.139737991266376,
+ "grad_norm": 0.01896674931049347,
+ "learning_rate": 0.0006,
+ "loss": 4.484029293060303,
+ "step": 1234
+ },
+ {
+ "epoch": 17.153711790393015,
+ "grad_norm": 0.018286997452378273,
+ "learning_rate": 0.0006,
+ "loss": 4.523510932922363,
+ "step": 1235
+ },
+ {
+ "epoch": 17.16768558951965,
+ "grad_norm": 0.017273059114813805,
+ "learning_rate": 0.0006,
+ "loss": 4.451767444610596,
+ "step": 1236
+ },
+ {
+ "epoch": 17.18165938864629,
+ "grad_norm": 0.016499150544404984,
+ "learning_rate": 0.0006,
+ "loss": 4.5568647384643555,
+ "step": 1237
+ },
+ {
+ "epoch": 17.195633187772927,
+ "grad_norm": 0.015934469178318977,
+ "learning_rate": 0.0006,
+ "loss": 4.4639997482299805,
+ "step": 1238
+ },
+ {
+ "epoch": 17.209606986899562,
+ "grad_norm": 0.01714209094643593,
+ "learning_rate": 0.0006,
+ "loss": 4.547614097595215,
+ "step": 1239
+ },
+ {
+ "epoch": 17.2235807860262,
+ "grad_norm": 0.019167495891451836,
+ "learning_rate": 0.0006,
+ "loss": 4.571097373962402,
+ "step": 1240
+ },
+ {
+ "epoch": 17.23755458515284,
+ "grad_norm": 0.018276996910572052,
+ "learning_rate": 0.0006,
+ "loss": 4.489764213562012,
+ "step": 1241
+ },
+ {
+ "epoch": 17.251528384279474,
+ "grad_norm": 0.021296866238117218,
+ "learning_rate": 0.0006,
+ "loss": 4.526913642883301,
+ "step": 1242
+ },
+ {
+ "epoch": 17.265502183406113,
+ "grad_norm": 0.022606629878282547,
+ "learning_rate": 0.0006,
+ "loss": 4.521624565124512,
+ "step": 1243
+ },
+ {
+ "epoch": 17.27947598253275,
+ "grad_norm": 0.021204711869359016,
+ "learning_rate": 0.0006,
+ "loss": 4.500864028930664,
+ "step": 1244
+ },
+ {
+ "epoch": 17.29344978165939,
+ "grad_norm": 0.01988849975168705,
+ "learning_rate": 0.0006,
+ "loss": 4.45162296295166,
+ "step": 1245
+ },
+ {
+ "epoch": 17.307423580786025,
+ "grad_norm": 0.020268384367227554,
+ "learning_rate": 0.0006,
+ "loss": 4.536198616027832,
+ "step": 1246
+ },
+ {
+ "epoch": 17.321397379912664,
+ "grad_norm": 0.020185677334666252,
+ "learning_rate": 0.0006,
+ "loss": 4.440262794494629,
+ "step": 1247
+ },
+ {
+ "epoch": 17.335371179039303,
+ "grad_norm": 0.020106853917241096,
+ "learning_rate": 0.0006,
+ "loss": 4.475907325744629,
+ "step": 1248
+ },
+ {
+ "epoch": 17.349344978165938,
+ "grad_norm": 0.018883297219872475,
+ "learning_rate": 0.0006,
+ "loss": 4.640661716461182,
+ "step": 1249
+ },
+ {
+ "epoch": 17.363318777292577,
+ "grad_norm": 0.01852620206773281,
+ "learning_rate": 0.0006,
+ "loss": 4.578243255615234,
+ "step": 1250
+ },
+ {
+ "epoch": 17.377292576419215,
+ "grad_norm": 0.01886451430618763,
+ "learning_rate": 0.0006,
+ "loss": 4.546026229858398,
+ "step": 1251
+ },
+ {
+ "epoch": 17.39126637554585,
+ "grad_norm": 0.01777452416718006,
+ "learning_rate": 0.0006,
+ "loss": 4.521542549133301,
+ "step": 1252
+ },
+ {
+ "epoch": 17.40524017467249,
+ "grad_norm": 0.016626447439193726,
+ "learning_rate": 0.0006,
+ "loss": 4.52629280090332,
+ "step": 1253
+ },
+ {
+ "epoch": 17.419213973799128,
+ "grad_norm": 0.01599477231502533,
+ "learning_rate": 0.0006,
+ "loss": 4.4796037673950195,
+ "step": 1254
+ },
+ {
+ "epoch": 17.433187772925763,
+ "grad_norm": 0.017021458595991135,
+ "learning_rate": 0.0006,
+ "loss": 4.468813419342041,
+ "step": 1255
+ },
+ {
+ "epoch": 17.4471615720524,
+ "grad_norm": 0.016925333067774773,
+ "learning_rate": 0.0006,
+ "loss": 4.584898948669434,
+ "step": 1256
+ },
+ {
+ "epoch": 17.46113537117904,
+ "grad_norm": 0.01714998669922352,
+ "learning_rate": 0.0006,
+ "loss": 4.523544788360596,
+ "step": 1257
+ },
+ {
+ "epoch": 17.475109170305675,
+ "grad_norm": 0.01802288182079792,
+ "learning_rate": 0.0006,
+ "loss": 4.595581531524658,
+ "step": 1258
+ },
+ {
+ "epoch": 17.489082969432314,
+ "grad_norm": 0.0180289875715971,
+ "learning_rate": 0.0006,
+ "loss": 4.394440650939941,
+ "step": 1259
+ },
+ {
+ "epoch": 17.503056768558952,
+ "grad_norm": 0.018504245206713676,
+ "learning_rate": 0.0006,
+ "loss": 4.4794416427612305,
+ "step": 1260
+ },
+ {
+ "epoch": 17.51703056768559,
+ "grad_norm": 0.019973233342170715,
+ "learning_rate": 0.0006,
+ "loss": 4.4241437911987305,
+ "step": 1261
+ },
+ {
+ "epoch": 17.531004366812226,
+ "grad_norm": 0.020063083618879318,
+ "learning_rate": 0.0006,
+ "loss": 4.48068380355835,
+ "step": 1262
+ },
+ {
+ "epoch": 17.544978165938865,
+ "grad_norm": 0.02066640742123127,
+ "learning_rate": 0.0006,
+ "loss": 4.628993034362793,
+ "step": 1263
+ },
+ {
+ "epoch": 17.558951965065503,
+ "grad_norm": 0.02206416428089142,
+ "learning_rate": 0.0006,
+ "loss": 4.564790725708008,
+ "step": 1264
+ },
+ {
+ "epoch": 17.57292576419214,
+ "grad_norm": 0.020640335977077484,
+ "learning_rate": 0.0006,
+ "loss": 4.578705787658691,
+ "step": 1265
+ },
+ {
+ "epoch": 17.586899563318777,
+ "grad_norm": 0.019093405455350876,
+ "learning_rate": 0.0006,
+ "loss": 4.5023603439331055,
+ "step": 1266
+ },
+ {
+ "epoch": 17.600873362445416,
+ "grad_norm": 0.019459193572402,
+ "learning_rate": 0.0006,
+ "loss": 4.594531059265137,
+ "step": 1267
+ },
+ {
+ "epoch": 17.61484716157205,
+ "grad_norm": 0.02021171525120735,
+ "learning_rate": 0.0006,
+ "loss": 4.770299911499023,
+ "step": 1268
+ },
+ {
+ "epoch": 17.62882096069869,
+ "grad_norm": 0.021161191165447235,
+ "learning_rate": 0.0006,
+ "loss": 4.5380401611328125,
+ "step": 1269
+ },
+ {
+ "epoch": 17.64279475982533,
+ "grad_norm": 0.021013403311371803,
+ "learning_rate": 0.0006,
+ "loss": 4.455524921417236,
+ "step": 1270
+ },
+ {
+ "epoch": 17.656768558951963,
+ "grad_norm": 0.020336592569947243,
+ "learning_rate": 0.0006,
+ "loss": 4.533672332763672,
+ "step": 1271
+ },
+ {
+ "epoch": 17.670742358078602,
+ "grad_norm": 0.020159434527158737,
+ "learning_rate": 0.0006,
+ "loss": 4.596151351928711,
+ "step": 1272
+ },
+ {
+ "epoch": 17.68471615720524,
+ "grad_norm": 0.020432960242033005,
+ "learning_rate": 0.0006,
+ "loss": 4.55531644821167,
+ "step": 1273
+ },
+ {
+ "epoch": 17.69868995633188,
+ "grad_norm": 0.025466229766607285,
+ "learning_rate": 0.0006,
+ "loss": 4.566912651062012,
+ "step": 1274
+ },
+ {
+ "epoch": 17.712663755458514,
+ "grad_norm": 0.026537558063864708,
+ "learning_rate": 0.0006,
+ "loss": 4.526003837585449,
+ "step": 1275
+ },
+ {
+ "epoch": 17.726637554585153,
+ "grad_norm": 0.024397607892751694,
+ "learning_rate": 0.0006,
+ "loss": 4.541755676269531,
+ "step": 1276
+ },
+ {
+ "epoch": 17.74061135371179,
+ "grad_norm": 0.022062312811613083,
+ "learning_rate": 0.0006,
+ "loss": 4.6532793045043945,
+ "step": 1277
+ },
+ {
+ "epoch": 17.754585152838427,
+ "grad_norm": 0.02204282023012638,
+ "learning_rate": 0.0006,
+ "loss": 4.582825660705566,
+ "step": 1278
+ },
+ {
+ "epoch": 17.768558951965066,
+ "grad_norm": 0.020844416692852974,
+ "learning_rate": 0.0006,
+ "loss": 4.522336483001709,
+ "step": 1279
+ },
+ {
+ "epoch": 17.782532751091704,
+ "grad_norm": 0.019815703853964806,
+ "learning_rate": 0.0006,
+ "loss": 4.607398509979248,
+ "step": 1280
+ },
+ {
+ "epoch": 17.79650655021834,
+ "grad_norm": 0.016676954925060272,
+ "learning_rate": 0.0006,
+ "loss": 4.446206092834473,
+ "step": 1281
+ },
+ {
+ "epoch": 17.810480349344978,
+ "grad_norm": 0.0166452769190073,
+ "learning_rate": 0.0006,
+ "loss": 4.505931854248047,
+ "step": 1282
+ },
+ {
+ "epoch": 17.824454148471617,
+ "grad_norm": 0.015928907319903374,
+ "learning_rate": 0.0006,
+ "loss": 4.455267429351807,
+ "step": 1283
+ },
+ {
+ "epoch": 17.83842794759825,
+ "grad_norm": 0.015219560824334621,
+ "learning_rate": 0.0006,
+ "loss": 4.5623674392700195,
+ "step": 1284
+ },
+ {
+ "epoch": 17.85240174672489,
+ "grad_norm": 0.015125355683267117,
+ "learning_rate": 0.0006,
+ "loss": 4.554104804992676,
+ "step": 1285
+ },
+ {
+ "epoch": 17.86637554585153,
+ "grad_norm": 0.015612718649208546,
+ "learning_rate": 0.0006,
+ "loss": 4.484847068786621,
+ "step": 1286
+ },
+ {
+ "epoch": 17.880349344978168,
+ "grad_norm": 0.015215662308037281,
+ "learning_rate": 0.0006,
+ "loss": 4.4802093505859375,
+ "step": 1287
+ },
+ {
+ "epoch": 17.894323144104803,
+ "grad_norm": 0.01511684525758028,
+ "learning_rate": 0.0006,
+ "loss": 4.537298679351807,
+ "step": 1288
+ },
+ {
+ "epoch": 17.90829694323144,
+ "grad_norm": 0.013841322623193264,
+ "learning_rate": 0.0006,
+ "loss": 4.427488327026367,
+ "step": 1289
+ },
+ {
+ "epoch": 17.92227074235808,
+ "grad_norm": 0.015115964226424694,
+ "learning_rate": 0.0006,
+ "loss": 4.4547119140625,
+ "step": 1290
+ },
+ {
+ "epoch": 17.936244541484715,
+ "grad_norm": 0.016411006450653076,
+ "learning_rate": 0.0006,
+ "loss": 4.52994966506958,
+ "step": 1291
+ },
+ {
+ "epoch": 17.950218340611354,
+ "grad_norm": 0.017313152551651,
+ "learning_rate": 0.0006,
+ "loss": 4.5337653160095215,
+ "step": 1292
+ },
+ {
+ "epoch": 17.964192139737992,
+ "grad_norm": 0.016559796407818794,
+ "learning_rate": 0.0006,
+ "loss": 4.5277910232543945,
+ "step": 1293
+ },
+ {
+ "epoch": 17.978165938864628,
+ "grad_norm": 0.015854761004447937,
+ "learning_rate": 0.0006,
+ "loss": 4.6334309577941895,
+ "step": 1294
+ },
+ {
+ "epoch": 17.992139737991266,
+ "grad_norm": 0.015658238902688026,
+ "learning_rate": 0.0006,
+ "loss": 4.613372802734375,
+ "step": 1295
+ },
+ {
+ "epoch": 18.0,
+ "grad_norm": 0.0169155802577734,
+ "learning_rate": 0.0006,
+ "loss": 4.527356147766113,
+ "step": 1296
+ },
+ {
+ "epoch": 18.0,
+ "eval_loss": 4.764543056488037,
+ "eval_runtime": 56.5998,
+ "eval_samples_per_second": 43.145,
+ "eval_steps_per_second": 1.36,
+ "step": 1296
+ },
+ {
+ "epoch": 18.01397379912664,
+ "grad_norm": 0.015348963439464569,
+ "learning_rate": 0.0006,
+ "loss": 4.466423988342285,
+ "step": 1297
+ },
+ {
+ "epoch": 18.027947598253274,
+ "grad_norm": 0.01768588088452816,
+ "learning_rate": 0.0006,
+ "loss": 4.545933246612549,
+ "step": 1298
+ },
+ {
+ "epoch": 18.041921397379912,
+ "grad_norm": 0.017381226643919945,
+ "learning_rate": 0.0006,
+ "loss": 4.494060039520264,
+ "step": 1299
+ },
+ {
+ "epoch": 18.05589519650655,
+ "grad_norm": 0.019002335146069527,
+ "learning_rate": 0.0006,
+ "loss": 4.395200729370117,
+ "step": 1300
+ },
+ {
+ "epoch": 18.069868995633186,
+ "grad_norm": 0.01827932707965374,
+ "learning_rate": 0.0006,
+ "loss": 4.424848556518555,
+ "step": 1301
+ },
+ {
+ "epoch": 18.083842794759825,
+ "grad_norm": 0.016383031383156776,
+ "learning_rate": 0.0006,
+ "loss": 4.494722366333008,
+ "step": 1302
+ },
+ {
+ "epoch": 18.097816593886463,
+ "grad_norm": 0.016265157610177994,
+ "learning_rate": 0.0006,
+ "loss": 4.416522979736328,
+ "step": 1303
+ },
+ {
+ "epoch": 18.111790393013102,
+ "grad_norm": 0.01702386513352394,
+ "learning_rate": 0.0006,
+ "loss": 4.600103378295898,
+ "step": 1304
+ },
+ {
+ "epoch": 18.125764192139737,
+ "grad_norm": 0.018495498225092888,
+ "learning_rate": 0.0006,
+ "loss": 4.588401794433594,
+ "step": 1305
+ },
+ {
+ "epoch": 18.139737991266376,
+ "grad_norm": 0.017921684309840202,
+ "learning_rate": 0.0006,
+ "loss": 4.478075981140137,
+ "step": 1306
+ },
+ {
+ "epoch": 18.153711790393015,
+ "grad_norm": 0.01801278069615364,
+ "learning_rate": 0.0006,
+ "loss": 4.413186550140381,
+ "step": 1307
+ },
+ {
+ "epoch": 18.16768558951965,
+ "grad_norm": 0.017211275175213814,
+ "learning_rate": 0.0006,
+ "loss": 4.461341857910156,
+ "step": 1308
+ },
+ {
+ "epoch": 18.18165938864629,
+ "grad_norm": 0.017589328810572624,
+ "learning_rate": 0.0006,
+ "loss": 4.546969413757324,
+ "step": 1309
+ },
+ {
+ "epoch": 18.195633187772927,
+ "grad_norm": 0.01797698810696602,
+ "learning_rate": 0.0006,
+ "loss": 4.530797004699707,
+ "step": 1310
+ },
+ {
+ "epoch": 18.209606986899562,
+ "grad_norm": 0.020408619195222855,
+ "learning_rate": 0.0006,
+ "loss": 4.62999153137207,
+ "step": 1311
+ },
+ {
+ "epoch": 18.2235807860262,
+ "grad_norm": 0.01994038000702858,
+ "learning_rate": 0.0006,
+ "loss": 4.586811065673828,
+ "step": 1312
+ },
+ {
+ "epoch": 18.23755458515284,
+ "grad_norm": 0.019381340593099594,
+ "learning_rate": 0.0006,
+ "loss": 4.4769368171691895,
+ "step": 1313
+ },
+ {
+ "epoch": 18.251528384279474,
+ "grad_norm": 0.020484555512666702,
+ "learning_rate": 0.0006,
+ "loss": 4.524302959442139,
+ "step": 1314
+ },
+ {
+ "epoch": 18.265502183406113,
+ "grad_norm": 0.02426416426897049,
+ "learning_rate": 0.0006,
+ "loss": 4.561705589294434,
+ "step": 1315
+ },
+ {
+ "epoch": 18.27947598253275,
+ "grad_norm": 0.024748101830482483,
+ "learning_rate": 0.0006,
+ "loss": 4.4654316902160645,
+ "step": 1316
+ },
+ {
+ "epoch": 18.29344978165939,
+ "grad_norm": 0.020703328773379326,
+ "learning_rate": 0.0006,
+ "loss": 4.53033447265625,
+ "step": 1317
+ },
+ {
+ "epoch": 18.307423580786025,
+ "grad_norm": 0.01765458658337593,
+ "learning_rate": 0.0006,
+ "loss": 4.5154266357421875,
+ "step": 1318
+ },
+ {
+ "epoch": 18.321397379912664,
+ "grad_norm": 0.018646420910954475,
+ "learning_rate": 0.0006,
+ "loss": 4.5224714279174805,
+ "step": 1319
+ },
+ {
+ "epoch": 18.335371179039303,
+ "grad_norm": 0.01907474733889103,
+ "learning_rate": 0.0006,
+ "loss": 4.5055437088012695,
+ "step": 1320
+ },
+ {
+ "epoch": 18.349344978165938,
+ "grad_norm": 0.018179846927523613,
+ "learning_rate": 0.0006,
+ "loss": 4.462161064147949,
+ "step": 1321
+ },
+ {
+ "epoch": 18.363318777292577,
+ "grad_norm": 0.016697878018021584,
+ "learning_rate": 0.0006,
+ "loss": 4.473718643188477,
+ "step": 1322
+ },
+ {
+ "epoch": 18.377292576419215,
+ "grad_norm": 0.017504019662737846,
+ "learning_rate": 0.0006,
+ "loss": 4.474348068237305,
+ "step": 1323
+ },
+ {
+ "epoch": 18.39126637554585,
+ "grad_norm": 0.018889913335442543,
+ "learning_rate": 0.0006,
+ "loss": 4.4531636238098145,
+ "step": 1324
+ },
+ {
+ "epoch": 18.40524017467249,
+ "grad_norm": 0.019167251884937286,
+ "learning_rate": 0.0006,
+ "loss": 4.4278364181518555,
+ "step": 1325
+ },
+ {
+ "epoch": 18.419213973799128,
+ "grad_norm": 0.015895741060376167,
+ "learning_rate": 0.0006,
+ "loss": 4.474592208862305,
+ "step": 1326
+ },
+ {
+ "epoch": 18.433187772925763,
+ "grad_norm": 0.01531882956624031,
+ "learning_rate": 0.0006,
+ "loss": 4.403757572174072,
+ "step": 1327
+ },
+ {
+ "epoch": 18.4471615720524,
+ "grad_norm": 0.01565849967300892,
+ "learning_rate": 0.0006,
+ "loss": 4.507664680480957,
+ "step": 1328
+ },
+ {
+ "epoch": 18.46113537117904,
+ "grad_norm": 0.01602315716445446,
+ "learning_rate": 0.0006,
+ "loss": 4.5648908615112305,
+ "step": 1329
+ },
+ {
+ "epoch": 18.475109170305675,
+ "grad_norm": 0.016467366367578506,
+ "learning_rate": 0.0006,
+ "loss": 4.428762912750244,
+ "step": 1330
+ },
+ {
+ "epoch": 18.489082969432314,
+ "grad_norm": 0.01604127697646618,
+ "learning_rate": 0.0006,
+ "loss": 4.527771472930908,
+ "step": 1331
+ },
+ {
+ "epoch": 18.503056768558952,
+ "grad_norm": 0.018825586885213852,
+ "learning_rate": 0.0006,
+ "loss": 4.485386848449707,
+ "step": 1332
+ },
+ {
+ "epoch": 18.51703056768559,
+ "grad_norm": 0.019258007407188416,
+ "learning_rate": 0.0006,
+ "loss": 4.414972305297852,
+ "step": 1333
+ },
+ {
+ "epoch": 18.531004366812226,
+ "grad_norm": 0.018192021176218987,
+ "learning_rate": 0.0006,
+ "loss": 4.412166595458984,
+ "step": 1334
+ },
+ {
+ "epoch": 18.544978165938865,
+ "grad_norm": 0.016120698302984238,
+ "learning_rate": 0.0006,
+ "loss": 4.484317779541016,
+ "step": 1335
+ },
+ {
+ "epoch": 18.558951965065503,
+ "grad_norm": 0.016496622934937477,
+ "learning_rate": 0.0006,
+ "loss": 4.514986515045166,
+ "step": 1336
+ },
+ {
+ "epoch": 18.57292576419214,
+ "grad_norm": 0.01795116998255253,
+ "learning_rate": 0.0006,
+ "loss": 4.41473388671875,
+ "step": 1337
+ },
+ {
+ "epoch": 18.586899563318777,
+ "grad_norm": 0.019062276929616928,
+ "learning_rate": 0.0006,
+ "loss": 4.601996898651123,
+ "step": 1338
+ },
+ {
+ "epoch": 18.600873362445416,
+ "grad_norm": 0.017174631357192993,
+ "learning_rate": 0.0006,
+ "loss": 4.571218490600586,
+ "step": 1339
+ },
+ {
+ "epoch": 18.61484716157205,
+ "grad_norm": 0.014015606604516506,
+ "learning_rate": 0.0006,
+ "loss": 4.526606559753418,
+ "step": 1340
+ },
+ {
+ "epoch": 18.62882096069869,
+ "grad_norm": 0.015358510427176952,
+ "learning_rate": 0.0006,
+ "loss": 4.604286193847656,
+ "step": 1341
+ },
+ {
+ "epoch": 18.64279475982533,
+ "grad_norm": 0.01619391329586506,
+ "learning_rate": 0.0006,
+ "loss": 4.560883522033691,
+ "step": 1342
+ },
+ {
+ "epoch": 18.656768558951963,
+ "grad_norm": 0.015972964465618134,
+ "learning_rate": 0.0006,
+ "loss": 4.479755401611328,
+ "step": 1343
+ },
+ {
+ "epoch": 18.670742358078602,
+ "grad_norm": 0.016861025243997574,
+ "learning_rate": 0.0006,
+ "loss": 4.474471092224121,
+ "step": 1344
+ },
+ {
+ "epoch": 18.68471615720524,
+ "grad_norm": 0.01822078227996826,
+ "learning_rate": 0.0006,
+ "loss": 4.509489059448242,
+ "step": 1345
+ },
+ {
+ "epoch": 18.69868995633188,
+ "grad_norm": 0.019415007904171944,
+ "learning_rate": 0.0006,
+ "loss": 4.508941173553467,
+ "step": 1346
+ },
+ {
+ "epoch": 18.712663755458514,
+ "grad_norm": 0.019856126978993416,
+ "learning_rate": 0.0006,
+ "loss": 4.446670055389404,
+ "step": 1347
+ },
+ {
+ "epoch": 18.726637554585153,
+ "grad_norm": 0.01842520199716091,
+ "learning_rate": 0.0006,
+ "loss": 4.524422645568848,
+ "step": 1348
+ },
+ {
+ "epoch": 18.74061135371179,
+ "grad_norm": 0.017196480184793472,
+ "learning_rate": 0.0006,
+ "loss": 4.498466968536377,
+ "step": 1349
+ },
+ {
+ "epoch": 18.754585152838427,
+ "grad_norm": 0.01823010854423046,
+ "learning_rate": 0.0006,
+ "loss": 4.534684181213379,
+ "step": 1350
+ },
+ {
+ "epoch": 18.768558951965066,
+ "grad_norm": 0.020398065447807312,
+ "learning_rate": 0.0006,
+ "loss": 4.456494331359863,
+ "step": 1351
+ },
+ {
+ "epoch": 18.782532751091704,
+ "grad_norm": 0.022992262616753578,
+ "learning_rate": 0.0006,
+ "loss": 4.476374626159668,
+ "step": 1352
+ },
+ {
+ "epoch": 18.79650655021834,
+ "grad_norm": 0.02235420234501362,
+ "learning_rate": 0.0006,
+ "loss": 4.587776184082031,
+ "step": 1353
+ },
+ {
+ "epoch": 18.810480349344978,
+ "grad_norm": 0.02093626745045185,
+ "learning_rate": 0.0006,
+ "loss": 4.424172401428223,
+ "step": 1354
+ },
+ {
+ "epoch": 18.824454148471617,
+ "grad_norm": 0.01997302658855915,
+ "learning_rate": 0.0006,
+ "loss": 4.448066711425781,
+ "step": 1355
+ },
+ {
+ "epoch": 18.83842794759825,
+ "grad_norm": 0.018265459686517715,
+ "learning_rate": 0.0006,
+ "loss": 4.436610698699951,
+ "step": 1356
+ },
+ {
+ "epoch": 18.85240174672489,
+ "grad_norm": 0.01673833839595318,
+ "learning_rate": 0.0006,
+ "loss": 4.5502214431762695,
+ "step": 1357
+ },
+ {
+ "epoch": 18.86637554585153,
+ "grad_norm": 0.017908833920955658,
+ "learning_rate": 0.0006,
+ "loss": 4.550530433654785,
+ "step": 1358
+ },
+ {
+ "epoch": 18.880349344978168,
+ "grad_norm": 0.01796654611825943,
+ "learning_rate": 0.0006,
+ "loss": 4.531463146209717,
+ "step": 1359
+ },
+ {
+ "epoch": 18.894323144104803,
+ "grad_norm": 0.01646183803677559,
+ "learning_rate": 0.0006,
+ "loss": 4.498582363128662,
+ "step": 1360
+ },
+ {
+ "epoch": 18.90829694323144,
+ "grad_norm": 0.015430327504873276,
+ "learning_rate": 0.0006,
+ "loss": 4.52571439743042,
+ "step": 1361
+ },
+ {
+ "epoch": 18.92227074235808,
+ "grad_norm": 0.016527952626347542,
+ "learning_rate": 0.0006,
+ "loss": 4.547013759613037,
+ "step": 1362
+ },
+ {
+ "epoch": 18.936244541484715,
+ "grad_norm": 0.016263660043478012,
+ "learning_rate": 0.0006,
+ "loss": 4.468754291534424,
+ "step": 1363
+ },
+ {
+ "epoch": 18.950218340611354,
+ "grad_norm": 0.016347210854291916,
+ "learning_rate": 0.0006,
+ "loss": 4.450741767883301,
+ "step": 1364
+ },
+ {
+ "epoch": 18.964192139737992,
+ "grad_norm": 0.01701398566365242,
+ "learning_rate": 0.0006,
+ "loss": 4.534899711608887,
+ "step": 1365
+ },
+ {
+ "epoch": 18.978165938864628,
+ "grad_norm": 0.017693081870675087,
+ "learning_rate": 0.0006,
+ "loss": 4.460553169250488,
+ "step": 1366
+ },
+ {
+ "epoch": 18.992139737991266,
+ "grad_norm": 0.018746308982372284,
+ "learning_rate": 0.0006,
+ "loss": 4.538792610168457,
+ "step": 1367
+ },
+ {
+ "epoch": 19.0,
+ "grad_norm": 0.020099416375160217,
+ "learning_rate": 0.0006,
+ "loss": 4.449617385864258,
+ "step": 1368
+ },
+ {
+ "epoch": 19.0,
+ "eval_loss": 4.7756571769714355,
+ "eval_runtime": 57.1804,
+ "eval_samples_per_second": 42.707,
+ "eval_steps_per_second": 1.347,
+ "step": 1368
+ },
+ {
+ "epoch": 19.01397379912664,
+ "grad_norm": 0.018726933747529984,
+ "learning_rate": 0.0006,
+ "loss": 4.446059703826904,
+ "step": 1369
+ },
+ {
+ "epoch": 19.027947598253274,
+ "grad_norm": 0.01956087350845337,
+ "learning_rate": 0.0006,
+ "loss": 4.4520039558410645,
+ "step": 1370
+ },
+ {
+ "epoch": 19.041921397379912,
+ "grad_norm": 0.020092811435461044,
+ "learning_rate": 0.0006,
+ "loss": 4.543343544006348,
+ "step": 1371
+ },
+ {
+ "epoch": 19.05589519650655,
+ "grad_norm": 0.01843622885644436,
+ "learning_rate": 0.0006,
+ "loss": 4.535579204559326,
+ "step": 1372
+ },
+ {
+ "epoch": 19.069868995633186,
+ "grad_norm": 0.016989488154649734,
+ "learning_rate": 0.0006,
+ "loss": 4.52316951751709,
+ "step": 1373
+ },
+ {
+ "epoch": 19.083842794759825,
+ "grad_norm": 0.01805422455072403,
+ "learning_rate": 0.0006,
+ "loss": 4.5393266677856445,
+ "step": 1374
+ },
+ {
+ "epoch": 19.097816593886463,
+ "grad_norm": 0.01874629408121109,
+ "learning_rate": 0.0006,
+ "loss": 4.422059059143066,
+ "step": 1375
+ },
+ {
+ "epoch": 19.111790393013102,
+ "grad_norm": 0.01804221048951149,
+ "learning_rate": 0.0006,
+ "loss": 4.50508451461792,
+ "step": 1376
+ },
+ {
+ "epoch": 19.125764192139737,
+ "grad_norm": 0.01785474829375744,
+ "learning_rate": 0.0006,
+ "loss": 4.412599563598633,
+ "step": 1377
+ },
+ {
+ "epoch": 19.139737991266376,
+ "grad_norm": 0.01971791312098503,
+ "learning_rate": 0.0006,
+ "loss": 4.518318176269531,
+ "step": 1378
+ },
+ {
+ "epoch": 19.153711790393015,
+ "grad_norm": 0.02061633951961994,
+ "learning_rate": 0.0006,
+ "loss": 4.332393169403076,
+ "step": 1379
+ },
+ {
+ "epoch": 19.16768558951965,
+ "grad_norm": 0.020563429221510887,
+ "learning_rate": 0.0006,
+ "loss": 4.445863723754883,
+ "step": 1380
+ },
+ {
+ "epoch": 19.18165938864629,
+ "grad_norm": 0.019995713606476784,
+ "learning_rate": 0.0006,
+ "loss": 4.497312068939209,
+ "step": 1381
+ },
+ {
+ "epoch": 19.195633187772927,
+ "grad_norm": 0.019823070615530014,
+ "learning_rate": 0.0006,
+ "loss": 4.475085258483887,
+ "step": 1382
+ },
+ {
+ "epoch": 19.209606986899562,
+ "grad_norm": 0.019039355218410492,
+ "learning_rate": 0.0006,
+ "loss": 4.50432014465332,
+ "step": 1383
+ },
+ {
+ "epoch": 19.2235807860262,
+ "grad_norm": 0.017876973375678062,
+ "learning_rate": 0.0006,
+ "loss": 4.384788513183594,
+ "step": 1384
+ },
+ {
+ "epoch": 19.23755458515284,
+ "grad_norm": 0.018697699531912804,
+ "learning_rate": 0.0006,
+ "loss": 4.410778045654297,
+ "step": 1385
+ },
+ {
+ "epoch": 19.251528384279474,
+ "grad_norm": 0.018826806917786598,
+ "learning_rate": 0.0006,
+ "loss": 4.524034023284912,
+ "step": 1386
+ },
+ {
+ "epoch": 19.265502183406113,
+ "grad_norm": 0.01824226602911949,
+ "learning_rate": 0.0006,
+ "loss": 4.526393890380859,
+ "step": 1387
+ },
+ {
+ "epoch": 19.27947598253275,
+ "grad_norm": 0.01753208599984646,
+ "learning_rate": 0.0006,
+ "loss": 4.493703365325928,
+ "step": 1388
+ },
+ {
+ "epoch": 19.29344978165939,
+ "grad_norm": 0.01742432825267315,
+ "learning_rate": 0.0006,
+ "loss": 4.392831802368164,
+ "step": 1389
+ },
+ {
+ "epoch": 19.307423580786025,
+ "grad_norm": 0.017099356278777122,
+ "learning_rate": 0.0006,
+ "loss": 4.578668594360352,
+ "step": 1390
+ },
+ {
+ "epoch": 19.321397379912664,
+ "grad_norm": 0.017946023494005203,
+ "learning_rate": 0.0006,
+ "loss": 4.456718444824219,
+ "step": 1391
+ },
+ {
+ "epoch": 19.335371179039303,
+ "grad_norm": 0.017705943435430527,
+ "learning_rate": 0.0006,
+ "loss": 4.553682327270508,
+ "step": 1392
+ },
+ {
+ "epoch": 19.349344978165938,
+ "grad_norm": 0.016647523269057274,
+ "learning_rate": 0.0006,
+ "loss": 4.525016784667969,
+ "step": 1393
+ },
+ {
+ "epoch": 19.363318777292577,
+ "grad_norm": 0.015407416969537735,
+ "learning_rate": 0.0006,
+ "loss": 4.500231742858887,
+ "step": 1394
+ },
+ {
+ "epoch": 19.377292576419215,
+ "grad_norm": 0.017264019697904587,
+ "learning_rate": 0.0006,
+ "loss": 4.4615349769592285,
+ "step": 1395
+ },
+ {
+ "epoch": 19.39126637554585,
+ "grad_norm": 0.01871555484831333,
+ "learning_rate": 0.0006,
+ "loss": 4.281026840209961,
+ "step": 1396
+ },
+ {
+ "epoch": 19.40524017467249,
+ "grad_norm": 0.017361650243401527,
+ "learning_rate": 0.0006,
+ "loss": 4.479382514953613,
+ "step": 1397
+ },
+ {
+ "epoch": 19.419213973799128,
+ "grad_norm": 0.019781362265348434,
+ "learning_rate": 0.0006,
+ "loss": 4.402320861816406,
+ "step": 1398
+ },
+ {
+ "epoch": 19.433187772925763,
+ "grad_norm": 0.019043298438191414,
+ "learning_rate": 0.0006,
+ "loss": 4.476983070373535,
+ "step": 1399
+ },
+ {
+ "epoch": 19.4471615720524,
+ "grad_norm": 0.017137985676527023,
+ "learning_rate": 0.0006,
+ "loss": 4.418689250946045,
+ "step": 1400
+ },
+ {
+ "epoch": 19.46113537117904,
+ "grad_norm": 0.016481833532452583,
+ "learning_rate": 0.0006,
+ "loss": 4.579145431518555,
+ "step": 1401
+ },
+ {
+ "epoch": 19.475109170305675,
+ "grad_norm": 0.016507714986801147,
+ "learning_rate": 0.0006,
+ "loss": 4.519158363342285,
+ "step": 1402
+ },
+ {
+ "epoch": 19.489082969432314,
+ "grad_norm": 0.018092872574925423,
+ "learning_rate": 0.0006,
+ "loss": 4.513748645782471,
+ "step": 1403
+ },
+ {
+ "epoch": 19.503056768558952,
+ "grad_norm": 0.018153440207242966,
+ "learning_rate": 0.0006,
+ "loss": 4.645384311676025,
+ "step": 1404
+ },
+ {
+ "epoch": 19.51703056768559,
+ "grad_norm": 0.01657119207084179,
+ "learning_rate": 0.0006,
+ "loss": 4.399757385253906,
+ "step": 1405
+ },
+ {
+ "epoch": 19.531004366812226,
+ "grad_norm": 0.0156280305236578,
+ "learning_rate": 0.0006,
+ "loss": 4.467836380004883,
+ "step": 1406
+ },
+ {
+ "epoch": 19.544978165938865,
+ "grad_norm": 0.016838403418660164,
+ "learning_rate": 0.0006,
+ "loss": 4.529176235198975,
+ "step": 1407
+ },
+ {
+ "epoch": 19.558951965065503,
+ "grad_norm": 0.01743236929178238,
+ "learning_rate": 0.0006,
+ "loss": 4.440032482147217,
+ "step": 1408
+ },
+ {
+ "epoch": 19.57292576419214,
+ "grad_norm": 0.01553610060364008,
+ "learning_rate": 0.0006,
+ "loss": 4.425588607788086,
+ "step": 1409
+ },
+ {
+ "epoch": 19.586899563318777,
+ "grad_norm": 0.017259271815419197,
+ "learning_rate": 0.0006,
+ "loss": 4.405452728271484,
+ "step": 1410
+ },
+ {
+ "epoch": 19.600873362445416,
+ "grad_norm": 0.015568481758236885,
+ "learning_rate": 0.0006,
+ "loss": 4.444279193878174,
+ "step": 1411
+ },
+ {
+ "epoch": 19.61484716157205,
+ "grad_norm": 0.015331248752772808,
+ "learning_rate": 0.0006,
+ "loss": 4.451846122741699,
+ "step": 1412
+ },
+ {
+ "epoch": 19.62882096069869,
+ "grad_norm": 0.015667378902435303,
+ "learning_rate": 0.0006,
+ "loss": 4.533931255340576,
+ "step": 1413
+ },
+ {
+ "epoch": 19.64279475982533,
+ "grad_norm": 0.016177602112293243,
+ "learning_rate": 0.0006,
+ "loss": 4.467313289642334,
+ "step": 1414
+ },
+ {
+ "epoch": 19.656768558951963,
+ "grad_norm": 0.01733129471540451,
+ "learning_rate": 0.0006,
+ "loss": 4.496893882751465,
+ "step": 1415
+ },
+ {
+ "epoch": 19.670742358078602,
+ "grad_norm": 0.016539594158530235,
+ "learning_rate": 0.0006,
+ "loss": 4.401961326599121,
+ "step": 1416
+ },
+ {
+ "epoch": 19.68471615720524,
+ "grad_norm": 0.016058551147580147,
+ "learning_rate": 0.0006,
+ "loss": 4.503375053405762,
+ "step": 1417
+ },
+ {
+ "epoch": 19.69868995633188,
+ "grad_norm": 0.01715777814388275,
+ "learning_rate": 0.0006,
+ "loss": 4.3808369636535645,
+ "step": 1418
+ },
+ {
+ "epoch": 19.712663755458514,
+ "grad_norm": 0.01681119203567505,
+ "learning_rate": 0.0006,
+ "loss": 4.5469560623168945,
+ "step": 1419
+ },
+ {
+ "epoch": 19.726637554585153,
+ "grad_norm": 0.01634550653398037,
+ "learning_rate": 0.0006,
+ "loss": 4.328197956085205,
+ "step": 1420
+ },
+ {
+ "epoch": 19.74061135371179,
+ "grad_norm": 0.01549347210675478,
+ "learning_rate": 0.0006,
+ "loss": 4.534621238708496,
+ "step": 1421
+ },
+ {
+ "epoch": 19.754585152838427,
+ "grad_norm": 0.01654665172100067,
+ "learning_rate": 0.0006,
+ "loss": 4.4724225997924805,
+ "step": 1422
+ },
+ {
+ "epoch": 19.768558951965066,
+ "grad_norm": 0.017119025811553,
+ "learning_rate": 0.0006,
+ "loss": 4.3897247314453125,
+ "step": 1423
+ },
+ {
+ "epoch": 19.782532751091704,
+ "grad_norm": 0.01603223942220211,
+ "learning_rate": 0.0006,
+ "loss": 4.412731170654297,
+ "step": 1424
+ },
+ {
+ "epoch": 19.79650655021834,
+ "grad_norm": 0.016774067655205727,
+ "learning_rate": 0.0006,
+ "loss": 4.54707145690918,
+ "step": 1425
+ },
+ {
+ "epoch": 19.810480349344978,
+ "grad_norm": 0.016453025862574577,
+ "learning_rate": 0.0006,
+ "loss": 4.462501049041748,
+ "step": 1426
+ },
+ {
+ "epoch": 19.824454148471617,
+ "grad_norm": 0.01597629487514496,
+ "learning_rate": 0.0006,
+ "loss": 4.229758262634277,
+ "step": 1427
+ },
+ {
+ "epoch": 19.83842794759825,
+ "grad_norm": 0.017999034374952316,
+ "learning_rate": 0.0006,
+ "loss": 4.389263153076172,
+ "step": 1428
+ },
+ {
+ "epoch": 19.85240174672489,
+ "grad_norm": 0.02165898121893406,
+ "learning_rate": 0.0006,
+ "loss": 4.342130661010742,
+ "step": 1429
+ },
+ {
+ "epoch": 19.86637554585153,
+ "grad_norm": 0.02222822792828083,
+ "learning_rate": 0.0006,
+ "loss": 4.492661476135254,
+ "step": 1430
+ },
+ {
+ "epoch": 19.880349344978168,
+ "grad_norm": 0.019378066062927246,
+ "learning_rate": 0.0006,
+ "loss": 4.4614105224609375,
+ "step": 1431
+ },
+ {
+ "epoch": 19.894323144104803,
+ "grad_norm": 0.019908083602786064,
+ "learning_rate": 0.0006,
+ "loss": 4.370929718017578,
+ "step": 1432
+ },
+ {
+ "epoch": 19.90829694323144,
+ "grad_norm": 0.022296471521258354,
+ "learning_rate": 0.0006,
+ "loss": 4.438320159912109,
+ "step": 1433
+ },
+ {
+ "epoch": 19.92227074235808,
+ "grad_norm": 0.021810343489050865,
+ "learning_rate": 0.0006,
+ "loss": 4.462067127227783,
+ "step": 1434
+ },
+ {
+ "epoch": 19.936244541484715,
+ "grad_norm": 0.02129209227859974,
+ "learning_rate": 0.0006,
+ "loss": 4.541261196136475,
+ "step": 1435
+ },
+ {
+ "epoch": 19.950218340611354,
+ "grad_norm": 0.02079629711806774,
+ "learning_rate": 0.0006,
+ "loss": 4.36712646484375,
+ "step": 1436
+ },
+ {
+ "epoch": 19.964192139737992,
+ "grad_norm": 0.022501740604639053,
+ "learning_rate": 0.0006,
+ "loss": 4.5279083251953125,
+ "step": 1437
+ },
+ {
+ "epoch": 19.978165938864628,
+ "grad_norm": 0.019121676683425903,
+ "learning_rate": 0.0006,
+ "loss": 4.477789878845215,
+ "step": 1438
+ },
+ {
+ "epoch": 19.992139737991266,
+ "grad_norm": 0.017429588362574577,
+ "learning_rate": 0.0006,
+ "loss": 4.4023847579956055,
+ "step": 1439
+ },
+ {
+ "epoch": 20.0,
+ "grad_norm": 0.018857665359973907,
+ "learning_rate": 0.0006,
+ "loss": 4.494743347167969,
+ "step": 1440
+ },
+ {
+ "epoch": 20.0,
+ "eval_loss": 4.694555759429932,
+ "eval_runtime": 56.4146,
+ "eval_samples_per_second": 43.287,
+ "eval_steps_per_second": 1.365,
+ "step": 1440
+ },
+ {
+ "epoch": 20.01397379912664,
+ "grad_norm": 0.016505861654877663,
+ "learning_rate": 0.0006,
+ "loss": 4.506229877471924,
+ "step": 1441
+ },
+ {
+ "epoch": 20.027947598253274,
+ "grad_norm": 0.017571842297911644,
+ "learning_rate": 0.0006,
+ "loss": 4.384073257446289,
+ "step": 1442
+ },
+ {
+ "epoch": 20.041921397379912,
+ "grad_norm": 0.019398365169763565,
+ "learning_rate": 0.0006,
+ "loss": 4.482177734375,
+ "step": 1443
+ },
+ {
+ "epoch": 20.05589519650655,
+ "grad_norm": 0.02011062018573284,
+ "learning_rate": 0.0006,
+ "loss": 4.440773963928223,
+ "step": 1444
+ },
+ {
+ "epoch": 20.069868995633186,
+ "grad_norm": 0.019026953727006912,
+ "learning_rate": 0.0006,
+ "loss": 4.507168769836426,
+ "step": 1445
+ },
+ {
+ "epoch": 20.083842794759825,
+ "grad_norm": 0.020506031811237335,
+ "learning_rate": 0.0006,
+ "loss": 4.469079971313477,
+ "step": 1446
+ },
+ {
+ "epoch": 20.097816593886463,
+ "grad_norm": 0.019608162343502045,
+ "learning_rate": 0.0006,
+ "loss": 4.395927429199219,
+ "step": 1447
+ },
+ {
+ "epoch": 20.111790393013102,
+ "grad_norm": 0.019419124349951744,
+ "learning_rate": 0.0006,
+ "loss": 4.484235763549805,
+ "step": 1448
+ },
+ {
+ "epoch": 20.125764192139737,
+ "grad_norm": 0.021775128319859505,
+ "learning_rate": 0.0006,
+ "loss": 4.333173751831055,
+ "step": 1449
+ },
+ {
+ "epoch": 20.139737991266376,
+ "grad_norm": 0.025512272492051125,
+ "learning_rate": 0.0006,
+ "loss": 4.460362434387207,
+ "step": 1450
+ },
+ {
+ "epoch": 20.153711790393015,
+ "grad_norm": 0.025157401338219643,
+ "learning_rate": 0.0006,
+ "loss": 4.4361395835876465,
+ "step": 1451
+ },
+ {
+ "epoch": 20.16768558951965,
+ "grad_norm": 0.019862616434693336,
+ "learning_rate": 0.0006,
+ "loss": 4.426807403564453,
+ "step": 1452
+ },
+ {
+ "epoch": 20.18165938864629,
+ "grad_norm": 0.01932152360677719,
+ "learning_rate": 0.0006,
+ "loss": 4.375677108764648,
+ "step": 1453
+ },
+ {
+ "epoch": 20.195633187772927,
+ "grad_norm": 0.01760999858379364,
+ "learning_rate": 0.0006,
+ "loss": 4.42259407043457,
+ "step": 1454
+ },
+ {
+ "epoch": 20.209606986899562,
+ "grad_norm": 0.016998112201690674,
+ "learning_rate": 0.0006,
+ "loss": 4.481280326843262,
+ "step": 1455
+ },
+ {
+ "epoch": 20.2235807860262,
+ "grad_norm": 0.017199190333485603,
+ "learning_rate": 0.0006,
+ "loss": 4.496712684631348,
+ "step": 1456
+ },
+ {
+ "epoch": 20.23755458515284,
+ "grad_norm": 0.016833679750561714,
+ "learning_rate": 0.0006,
+ "loss": 4.40736722946167,
+ "step": 1457
+ },
+ {
+ "epoch": 20.251528384279474,
+ "grad_norm": 0.017546730116009712,
+ "learning_rate": 0.0006,
+ "loss": 4.4915266036987305,
+ "step": 1458
+ },
+ {
+ "epoch": 20.265502183406113,
+ "grad_norm": 0.016595078632235527,
+ "learning_rate": 0.0006,
+ "loss": 4.460815906524658,
+ "step": 1459
+ },
+ {
+ "epoch": 20.27947598253275,
+ "grad_norm": 0.017728568986058235,
+ "learning_rate": 0.0006,
+ "loss": 4.346048831939697,
+ "step": 1460
+ },
+ {
+ "epoch": 20.29344978165939,
+ "grad_norm": 0.01849197782576084,
+ "learning_rate": 0.0006,
+ "loss": 4.399681091308594,
+ "step": 1461
+ },
+ {
+ "epoch": 20.307423580786025,
+ "grad_norm": 0.018336588516831398,
+ "learning_rate": 0.0006,
+ "loss": 4.414183616638184,
+ "step": 1462
+ },
+ {
+ "epoch": 20.321397379912664,
+ "grad_norm": 0.018086519092321396,
+ "learning_rate": 0.0006,
+ "loss": 4.346586227416992,
+ "step": 1463
+ },
+ {
+ "epoch": 20.335371179039303,
+ "grad_norm": 0.019697299227118492,
+ "learning_rate": 0.0006,
+ "loss": 4.4420952796936035,
+ "step": 1464
+ },
+ {
+ "epoch": 20.349344978165938,
+ "grad_norm": 0.02198074571788311,
+ "learning_rate": 0.0006,
+ "loss": 4.588729381561279,
+ "step": 1465
+ },
+ {
+ "epoch": 20.363318777292577,
+ "grad_norm": 0.022813329473137856,
+ "learning_rate": 0.0006,
+ "loss": 4.49729585647583,
+ "step": 1466
+ },
+ {
+ "epoch": 20.377292576419215,
+ "grad_norm": 0.022006649523973465,
+ "learning_rate": 0.0006,
+ "loss": 4.475641250610352,
+ "step": 1467
+ },
+ {
+ "epoch": 20.39126637554585,
+ "grad_norm": 0.023188438266515732,
+ "learning_rate": 0.0006,
+ "loss": 4.404322624206543,
+ "step": 1468
+ },
+ {
+ "epoch": 20.40524017467249,
+ "grad_norm": 0.022226877510547638,
+ "learning_rate": 0.0006,
+ "loss": 4.50904655456543,
+ "step": 1469
+ },
+ {
+ "epoch": 20.419213973799128,
+ "grad_norm": 0.021082637831568718,
+ "learning_rate": 0.0006,
+ "loss": 4.339011192321777,
+ "step": 1470
+ },
+ {
+ "epoch": 20.433187772925763,
+ "grad_norm": 0.02013542130589485,
+ "learning_rate": 0.0006,
+ "loss": 4.516942024230957,
+ "step": 1471
+ },
+ {
+ "epoch": 20.4471615720524,
+ "grad_norm": 0.01898968778550625,
+ "learning_rate": 0.0006,
+ "loss": 4.315804481506348,
+ "step": 1472
+ },
+ {
+ "epoch": 20.46113537117904,
+ "grad_norm": 0.019450880587100983,
+ "learning_rate": 0.0006,
+ "loss": 4.4794921875,
+ "step": 1473
+ },
+ {
+ "epoch": 20.475109170305675,
+ "grad_norm": 0.018577802926301956,
+ "learning_rate": 0.0006,
+ "loss": 4.422323703765869,
+ "step": 1474
+ },
+ {
+ "epoch": 20.489082969432314,
+ "grad_norm": 0.01778412237763405,
+ "learning_rate": 0.0006,
+ "loss": 4.429789066314697,
+ "step": 1475
+ },
+ {
+ "epoch": 20.503056768558952,
+ "grad_norm": 0.017134476453065872,
+ "learning_rate": 0.0006,
+ "loss": 4.416459083557129,
+ "step": 1476
+ },
+ {
+ "epoch": 20.51703056768559,
+ "grad_norm": 0.017153557389974594,
+ "learning_rate": 0.0006,
+ "loss": 4.446170806884766,
+ "step": 1477
+ },
+ {
+ "epoch": 20.531004366812226,
+ "grad_norm": 0.014998532831668854,
+ "learning_rate": 0.0006,
+ "loss": 4.419447898864746,
+ "step": 1478
+ },
+ {
+ "epoch": 20.544978165938865,
+ "grad_norm": 0.015877850353717804,
+ "learning_rate": 0.0006,
+ "loss": 4.312338829040527,
+ "step": 1479
+ },
+ {
+ "epoch": 20.558951965065503,
+ "grad_norm": 0.01569489948451519,
+ "learning_rate": 0.0006,
+ "loss": 4.572905540466309,
+ "step": 1480
+ },
+ {
+ "epoch": 20.57292576419214,
+ "grad_norm": 0.01591474376618862,
+ "learning_rate": 0.0006,
+ "loss": 4.414907455444336,
+ "step": 1481
+ },
+ {
+ "epoch": 20.586899563318777,
+ "grad_norm": 0.015395903028547764,
+ "learning_rate": 0.0006,
+ "loss": 4.549131393432617,
+ "step": 1482
+ },
+ {
+ "epoch": 20.600873362445416,
+ "grad_norm": 0.015715358778834343,
+ "learning_rate": 0.0006,
+ "loss": 4.530194282531738,
+ "step": 1483
+ },
+ {
+ "epoch": 20.61484716157205,
+ "grad_norm": 0.016042975708842278,
+ "learning_rate": 0.0006,
+ "loss": 4.322149276733398,
+ "step": 1484
+ },
+ {
+ "epoch": 20.62882096069869,
+ "grad_norm": 0.01664186641573906,
+ "learning_rate": 0.0006,
+ "loss": 4.424866676330566,
+ "step": 1485
+ },
+ {
+ "epoch": 20.64279475982533,
+ "grad_norm": 0.01631534844636917,
+ "learning_rate": 0.0006,
+ "loss": 4.441752910614014,
+ "step": 1486
+ },
+ {
+ "epoch": 20.656768558951963,
+ "grad_norm": 0.014057058840990067,
+ "learning_rate": 0.0006,
+ "loss": 4.399982452392578,
+ "step": 1487
+ },
+ {
+ "epoch": 20.670742358078602,
+ "grad_norm": 0.015076273120939732,
+ "learning_rate": 0.0006,
+ "loss": 4.441324234008789,
+ "step": 1488
+ },
+ {
+ "epoch": 20.68471615720524,
+ "grad_norm": 0.014138277620077133,
+ "learning_rate": 0.0006,
+ "loss": 4.4312543869018555,
+ "step": 1489
+ },
+ {
+ "epoch": 20.69868995633188,
+ "grad_norm": 0.013762440532445908,
+ "learning_rate": 0.0006,
+ "loss": 4.387807846069336,
+ "step": 1490
+ },
+ {
+ "epoch": 20.712663755458514,
+ "grad_norm": 0.015058411285281181,
+ "learning_rate": 0.0006,
+ "loss": 4.594747543334961,
+ "step": 1491
+ },
+ {
+ "epoch": 20.726637554585153,
+ "grad_norm": 0.013827123679220676,
+ "learning_rate": 0.0006,
+ "loss": 4.434604167938232,
+ "step": 1492
+ },
+ {
+ "epoch": 20.74061135371179,
+ "grad_norm": 0.015315295197069645,
+ "learning_rate": 0.0006,
+ "loss": 4.588066101074219,
+ "step": 1493
+ },
+ {
+ "epoch": 20.754585152838427,
+ "grad_norm": 0.018678732216358185,
+ "learning_rate": 0.0006,
+ "loss": 4.3514204025268555,
+ "step": 1494
+ },
+ {
+ "epoch": 20.768558951965066,
+ "grad_norm": 0.019521350041031837,
+ "learning_rate": 0.0006,
+ "loss": 4.437658786773682,
+ "step": 1495
+ },
+ {
+ "epoch": 20.782532751091704,
+ "grad_norm": 0.018363403156399727,
+ "learning_rate": 0.0006,
+ "loss": 4.495002269744873,
+ "step": 1496
+ },
+ {
+ "epoch": 20.79650655021834,
+ "grad_norm": 0.016830725595355034,
+ "learning_rate": 0.0006,
+ "loss": 4.402471542358398,
+ "step": 1497
+ },
+ {
+ "epoch": 20.810480349344978,
+ "grad_norm": 0.01722072623670101,
+ "learning_rate": 0.0006,
+ "loss": 4.437653541564941,
+ "step": 1498
+ },
+ {
+ "epoch": 20.824454148471617,
+ "grad_norm": 0.017459599301218987,
+ "learning_rate": 0.0006,
+ "loss": 4.366603851318359,
+ "step": 1499
+ },
+ {
+ "epoch": 20.83842794759825,
+ "grad_norm": 0.018902219831943512,
+ "learning_rate": 0.0006,
+ "loss": 4.501561164855957,
+ "step": 1500
+ },
+ {
+ "epoch": 20.85240174672489,
+ "grad_norm": 0.020814096555113792,
+ "learning_rate": 0.0006,
+ "loss": 4.535295486450195,
+ "step": 1501
+ },
+ {
+ "epoch": 20.86637554585153,
+ "grad_norm": 0.018990976735949516,
+ "learning_rate": 0.0006,
+ "loss": 4.333580017089844,
+ "step": 1502
+ },
+ {
+ "epoch": 20.880349344978168,
+ "grad_norm": 0.01837880350649357,
+ "learning_rate": 0.0006,
+ "loss": 4.313388824462891,
+ "step": 1503
+ },
+ {
+ "epoch": 20.894323144104803,
+ "grad_norm": 0.019820360466837883,
+ "learning_rate": 0.0006,
+ "loss": 4.388157844543457,
+ "step": 1504
+ },
+ {
+ "epoch": 20.90829694323144,
+ "grad_norm": 0.020181143656373024,
+ "learning_rate": 0.0006,
+ "loss": 4.572224140167236,
+ "step": 1505
+ },
+ {
+ "epoch": 20.92227074235808,
+ "grad_norm": 0.021206077188253403,
+ "learning_rate": 0.0006,
+ "loss": 4.3127970695495605,
+ "step": 1506
+ },
+ {
+ "epoch": 20.936244541484715,
+ "grad_norm": 0.017962060868740082,
+ "learning_rate": 0.0006,
+ "loss": 4.384734630584717,
+ "step": 1507
+ },
+ {
+ "epoch": 20.950218340611354,
+ "grad_norm": 0.01737220585346222,
+ "learning_rate": 0.0006,
+ "loss": 4.4075751304626465,
+ "step": 1508
+ },
+ {
+ "epoch": 20.964192139737992,
+ "grad_norm": 0.01839268207550049,
+ "learning_rate": 0.0006,
+ "loss": 4.553779125213623,
+ "step": 1509
+ },
+ {
+ "epoch": 20.978165938864628,
+ "grad_norm": 0.020102640613913536,
+ "learning_rate": 0.0006,
+ "loss": 4.399398326873779,
+ "step": 1510
+ },
+ {
+ "epoch": 20.992139737991266,
+ "grad_norm": 0.018831463530659676,
+ "learning_rate": 0.0006,
+ "loss": 4.358499526977539,
+ "step": 1511
+ },
+ {
+ "epoch": 21.0,
+ "grad_norm": 0.018254734575748444,
+ "learning_rate": 0.0006,
+ "loss": 4.384119510650635,
+ "step": 1512
+ },
+ {
+ "epoch": 21.0,
+ "eval_loss": 4.7524237632751465,
+ "eval_runtime": 57.8812,
+ "eval_samples_per_second": 42.19,
+ "eval_steps_per_second": 1.33,
+ "step": 1512
+ },
+ {
+ "epoch": 21.01397379912664,
+ "grad_norm": 0.019728587940335274,
+ "learning_rate": 0.0006,
+ "loss": 4.32363224029541,
+ "step": 1513
+ },
+ {
+ "epoch": 21.027947598253274,
+ "grad_norm": 0.02216625027358532,
+ "learning_rate": 0.0006,
+ "loss": 4.436988353729248,
+ "step": 1514
+ },
+ {
+ "epoch": 21.041921397379912,
+ "grad_norm": 0.02227606810629368,
+ "learning_rate": 0.0006,
+ "loss": 4.406063556671143,
+ "step": 1515
+ },
+ {
+ "epoch": 21.05589519650655,
+ "grad_norm": 0.020187009125947952,
+ "learning_rate": 0.0006,
+ "loss": 4.324402332305908,
+ "step": 1516
+ },
+ {
+ "epoch": 21.069868995633186,
+ "grad_norm": 0.02095157466828823,
+ "learning_rate": 0.0006,
+ "loss": 4.456544876098633,
+ "step": 1517
+ },
+ {
+ "epoch": 21.083842794759825,
+ "grad_norm": 0.02146013453602791,
+ "learning_rate": 0.0006,
+ "loss": 4.368387222290039,
+ "step": 1518
+ },
+ {
+ "epoch": 21.097816593886463,
+ "grad_norm": 0.02162035182118416,
+ "learning_rate": 0.0006,
+ "loss": 4.48838996887207,
+ "step": 1519
+ },
+ {
+ "epoch": 21.111790393013102,
+ "grad_norm": 0.021745676174759865,
+ "learning_rate": 0.0006,
+ "loss": 4.30778694152832,
+ "step": 1520
+ },
+ {
+ "epoch": 21.125764192139737,
+ "grad_norm": 0.02173588052392006,
+ "learning_rate": 0.0006,
+ "loss": 4.433149814605713,
+ "step": 1521
+ },
+ {
+ "epoch": 21.139737991266376,
+ "grad_norm": 0.020950648933649063,
+ "learning_rate": 0.0006,
+ "loss": 4.520089149475098,
+ "step": 1522
+ },
+ {
+ "epoch": 21.153711790393015,
+ "grad_norm": 0.020684752613306046,
+ "learning_rate": 0.0006,
+ "loss": 4.372437477111816,
+ "step": 1523
+ },
+ {
+ "epoch": 21.16768558951965,
+ "grad_norm": 0.019472869113087654,
+ "learning_rate": 0.0006,
+ "loss": 4.437015533447266,
+ "step": 1524
+ },
+ {
+ "epoch": 21.18165938864629,
+ "grad_norm": 0.01846962980926037,
+ "learning_rate": 0.0006,
+ "loss": 4.449531555175781,
+ "step": 1525
+ },
+ {
+ "epoch": 21.195633187772927,
+ "grad_norm": 0.017377803102135658,
+ "learning_rate": 0.0006,
+ "loss": 4.458279132843018,
+ "step": 1526
+ },
+ {
+ "epoch": 21.209606986899562,
+ "grad_norm": 0.016002200543880463,
+ "learning_rate": 0.0006,
+ "loss": 4.266297817230225,
+ "step": 1527
+ },
+ {
+ "epoch": 21.2235807860262,
+ "grad_norm": 0.015575903467833996,
+ "learning_rate": 0.0006,
+ "loss": 4.342632293701172,
+ "step": 1528
+ },
+ {
+ "epoch": 21.23755458515284,
+ "grad_norm": 0.015634698793292046,
+ "learning_rate": 0.0006,
+ "loss": 4.338623523712158,
+ "step": 1529
+ },
+ {
+ "epoch": 21.251528384279474,
+ "grad_norm": 0.01449581328779459,
+ "learning_rate": 0.0006,
+ "loss": 4.354330062866211,
+ "step": 1530
+ },
+ {
+ "epoch": 21.265502183406113,
+ "grad_norm": 0.01487616915255785,
+ "learning_rate": 0.0006,
+ "loss": 4.486835479736328,
+ "step": 1531
+ },
+ {
+ "epoch": 21.27947598253275,
+ "grad_norm": 0.016346555203199387,
+ "learning_rate": 0.0006,
+ "loss": 4.407098770141602,
+ "step": 1532
+ },
+ {
+ "epoch": 21.29344978165939,
+ "grad_norm": 0.01682182401418686,
+ "learning_rate": 0.0006,
+ "loss": 4.436809539794922,
+ "step": 1533
+ },
+ {
+ "epoch": 21.307423580786025,
+ "grad_norm": 0.017704004421830177,
+ "learning_rate": 0.0006,
+ "loss": 4.3854780197143555,
+ "step": 1534
+ },
+ {
+ "epoch": 21.321397379912664,
+ "grad_norm": 0.018437420949339867,
+ "learning_rate": 0.0006,
+ "loss": 4.413283824920654,
+ "step": 1535
+ },
+ {
+ "epoch": 21.335371179039303,
+ "grad_norm": 0.016911081969738007,
+ "learning_rate": 0.0006,
+ "loss": 4.357306003570557,
+ "step": 1536
+ },
+ {
+ "epoch": 21.349344978165938,
+ "grad_norm": 0.01642223633825779,
+ "learning_rate": 0.0006,
+ "loss": 4.422322750091553,
+ "step": 1537
+ },
+ {
+ "epoch": 21.363318777292577,
+ "grad_norm": 0.01780877821147442,
+ "learning_rate": 0.0006,
+ "loss": 4.387620449066162,
+ "step": 1538
+ },
+ {
+ "epoch": 21.377292576419215,
+ "grad_norm": 0.01653975248336792,
+ "learning_rate": 0.0006,
+ "loss": 4.327267646789551,
+ "step": 1539
+ },
+ {
+ "epoch": 21.39126637554585,
+ "grad_norm": 0.01572156324982643,
+ "learning_rate": 0.0006,
+ "loss": 4.475979804992676,
+ "step": 1540
+ },
+ {
+ "epoch": 21.40524017467249,
+ "grad_norm": 0.017156066372990608,
+ "learning_rate": 0.0006,
+ "loss": 4.4011125564575195,
+ "step": 1541
+ },
+ {
+ "epoch": 21.419213973799128,
+ "grad_norm": 0.01788829080760479,
+ "learning_rate": 0.0006,
+ "loss": 4.4174089431762695,
+ "step": 1542
+ },
+ {
+ "epoch": 21.433187772925763,
+ "grad_norm": 0.01728985086083412,
+ "learning_rate": 0.0006,
+ "loss": 4.429244041442871,
+ "step": 1543
+ },
+ {
+ "epoch": 21.4471615720524,
+ "grad_norm": 0.016534466296434402,
+ "learning_rate": 0.0006,
+ "loss": 4.326068878173828,
+ "step": 1544
+ },
+ {
+ "epoch": 21.46113537117904,
+ "grad_norm": 0.016582584008574486,
+ "learning_rate": 0.0006,
+ "loss": 4.2975969314575195,
+ "step": 1545
+ },
+ {
+ "epoch": 21.475109170305675,
+ "grad_norm": 0.018034178763628006,
+ "learning_rate": 0.0006,
+ "loss": 4.501319408416748,
+ "step": 1546
+ },
+ {
+ "epoch": 21.489082969432314,
+ "grad_norm": 0.017827702686190605,
+ "learning_rate": 0.0006,
+ "loss": 4.502284526824951,
+ "step": 1547
+ },
+ {
+ "epoch": 21.503056768558952,
+ "grad_norm": 0.017182299867272377,
+ "learning_rate": 0.0006,
+ "loss": 4.380828857421875,
+ "step": 1548
+ },
+ {
+ "epoch": 21.51703056768559,
+ "grad_norm": 0.019101588055491447,
+ "learning_rate": 0.0006,
+ "loss": 4.31564998626709,
+ "step": 1549
+ },
+ {
+ "epoch": 21.531004366812226,
+ "grad_norm": 0.0204677302390337,
+ "learning_rate": 0.0006,
+ "loss": 4.3840131759643555,
+ "step": 1550
+ },
+ {
+ "epoch": 21.544978165938865,
+ "grad_norm": 0.0188993439078331,
+ "learning_rate": 0.0006,
+ "loss": 4.4257731437683105,
+ "step": 1551
+ },
+ {
+ "epoch": 21.558951965065503,
+ "grad_norm": 0.01730230636894703,
+ "learning_rate": 0.0006,
+ "loss": 4.445132255554199,
+ "step": 1552
+ },
+ {
+ "epoch": 21.57292576419214,
+ "grad_norm": 0.01801574043929577,
+ "learning_rate": 0.0006,
+ "loss": 4.427064895629883,
+ "step": 1553
+ },
+ {
+ "epoch": 21.586899563318777,
+ "grad_norm": 0.018417565152049065,
+ "learning_rate": 0.0006,
+ "loss": 4.319595813751221,
+ "step": 1554
+ },
+ {
+ "epoch": 21.600873362445416,
+ "grad_norm": 0.01683160290122032,
+ "learning_rate": 0.0006,
+ "loss": 4.349218368530273,
+ "step": 1555
+ },
+ {
+ "epoch": 21.61484716157205,
+ "grad_norm": 0.016973106190562248,
+ "learning_rate": 0.0006,
+ "loss": 4.375990390777588,
+ "step": 1556
+ },
+ {
+ "epoch": 21.62882096069869,
+ "grad_norm": 0.015688583254814148,
+ "learning_rate": 0.0006,
+ "loss": 4.408170223236084,
+ "step": 1557
+ },
+ {
+ "epoch": 21.64279475982533,
+ "grad_norm": 0.01573043316602707,
+ "learning_rate": 0.0006,
+ "loss": 4.416255950927734,
+ "step": 1558
+ },
+ {
+ "epoch": 21.656768558951963,
+ "grad_norm": 0.01744293048977852,
+ "learning_rate": 0.0006,
+ "loss": 4.477141380310059,
+ "step": 1559
+ },
+ {
+ "epoch": 21.670742358078602,
+ "grad_norm": 0.017503513023257256,
+ "learning_rate": 0.0006,
+ "loss": 4.3367486000061035,
+ "step": 1560
+ },
+ {
+ "epoch": 21.68471615720524,
+ "grad_norm": 0.016592150554060936,
+ "learning_rate": 0.0006,
+ "loss": 4.314785480499268,
+ "step": 1561
+ },
+ {
+ "epoch": 21.69868995633188,
+ "grad_norm": 0.015774084255099297,
+ "learning_rate": 0.0006,
+ "loss": 4.448209285736084,
+ "step": 1562
+ },
+ {
+ "epoch": 21.712663755458514,
+ "grad_norm": 0.015779150649905205,
+ "learning_rate": 0.0006,
+ "loss": 4.392690658569336,
+ "step": 1563
+ },
+ {
+ "epoch": 21.726637554585153,
+ "grad_norm": 0.015392648056149483,
+ "learning_rate": 0.0006,
+ "loss": 4.437741756439209,
+ "step": 1564
+ },
+ {
+ "epoch": 21.74061135371179,
+ "grad_norm": 0.01530836895108223,
+ "learning_rate": 0.0006,
+ "loss": 4.424595832824707,
+ "step": 1565
+ },
+ {
+ "epoch": 21.754585152838427,
+ "grad_norm": 0.015600779093801975,
+ "learning_rate": 0.0006,
+ "loss": 4.335417747497559,
+ "step": 1566
+ },
+ {
+ "epoch": 21.768558951965066,
+ "grad_norm": 0.014646518044173717,
+ "learning_rate": 0.0006,
+ "loss": 4.279081344604492,
+ "step": 1567
+ },
+ {
+ "epoch": 21.782532751091704,
+ "grad_norm": 0.014132864773273468,
+ "learning_rate": 0.0006,
+ "loss": 4.345950126647949,
+ "step": 1568
+ },
+ {
+ "epoch": 21.79650655021834,
+ "grad_norm": 0.01425724383443594,
+ "learning_rate": 0.0006,
+ "loss": 4.277588844299316,
+ "step": 1569
+ },
+ {
+ "epoch": 21.810480349344978,
+ "grad_norm": 0.015396572649478912,
+ "learning_rate": 0.0006,
+ "loss": 4.326196670532227,
+ "step": 1570
+ },
+ {
+ "epoch": 21.824454148471617,
+ "grad_norm": 0.01683669537305832,
+ "learning_rate": 0.0006,
+ "loss": 4.3294172286987305,
+ "step": 1571
+ },
+ {
+ "epoch": 21.83842794759825,
+ "grad_norm": 0.016365647315979004,
+ "learning_rate": 0.0006,
+ "loss": 4.523321628570557,
+ "step": 1572
+ },
+ {
+ "epoch": 21.85240174672489,
+ "grad_norm": 0.017142774537205696,
+ "learning_rate": 0.0006,
+ "loss": 4.475842475891113,
+ "step": 1573
+ },
+ {
+ "epoch": 21.86637554585153,
+ "grad_norm": 0.015950839966535568,
+ "learning_rate": 0.0006,
+ "loss": 4.397187232971191,
+ "step": 1574
+ },
+ {
+ "epoch": 21.880349344978168,
+ "grad_norm": 0.016180362552404404,
+ "learning_rate": 0.0006,
+ "loss": 4.414778709411621,
+ "step": 1575
+ },
+ {
+ "epoch": 21.894323144104803,
+ "grad_norm": 0.017615556716918945,
+ "learning_rate": 0.0006,
+ "loss": 4.348567962646484,
+ "step": 1576
+ },
+ {
+ "epoch": 21.90829694323144,
+ "grad_norm": 0.01896277628839016,
+ "learning_rate": 0.0006,
+ "loss": 4.431762218475342,
+ "step": 1577
+ },
+ {
+ "epoch": 21.92227074235808,
+ "grad_norm": 0.01990971714258194,
+ "learning_rate": 0.0006,
+ "loss": 4.431285858154297,
+ "step": 1578
+ },
+ {
+ "epoch": 21.936244541484715,
+ "grad_norm": 0.0201258547604084,
+ "learning_rate": 0.0006,
+ "loss": 4.309872627258301,
+ "step": 1579
+ },
+ {
+ "epoch": 21.950218340611354,
+ "grad_norm": 0.019620198756456375,
+ "learning_rate": 0.0006,
+ "loss": 4.484044551849365,
+ "step": 1580
+ },
+ {
+ "epoch": 21.964192139737992,
+ "grad_norm": 0.01771123707294464,
+ "learning_rate": 0.0006,
+ "loss": 4.368093013763428,
+ "step": 1581
+ },
+ {
+ "epoch": 21.978165938864628,
+ "grad_norm": 0.017501654103398323,
+ "learning_rate": 0.0006,
+ "loss": 4.372651100158691,
+ "step": 1582
+ },
+ {
+ "epoch": 21.992139737991266,
+ "grad_norm": 0.01693608984351158,
+ "learning_rate": 0.0006,
+ "loss": 4.379167556762695,
+ "step": 1583
+ },
+ {
+ "epoch": 22.0,
+ "grad_norm": 0.02002849616110325,
+ "learning_rate": 0.0006,
+ "loss": 4.329627990722656,
+ "step": 1584
+ },
+ {
+ "epoch": 22.0,
+ "eval_loss": 4.723696231842041,
+ "eval_runtime": 56.1758,
+ "eval_samples_per_second": 43.471,
+ "eval_steps_per_second": 1.371,
+ "step": 1584
+ },
+ {
+ "epoch": 22.01397379912664,
+ "grad_norm": 0.02030305378139019,
+ "learning_rate": 0.0006,
+ "loss": 4.355566024780273,
+ "step": 1585
+ },
+ {
+ "epoch": 22.027947598253274,
+ "grad_norm": 0.018638934940099716,
+ "learning_rate": 0.0006,
+ "loss": 4.314009189605713,
+ "step": 1586
+ },
+ {
+ "epoch": 22.041921397379912,
+ "grad_norm": 0.019917072728276253,
+ "learning_rate": 0.0006,
+ "loss": 4.367525577545166,
+ "step": 1587
+ },
+ {
+ "epoch": 22.05589519650655,
+ "grad_norm": 0.020785167813301086,
+ "learning_rate": 0.0006,
+ "loss": 4.375561714172363,
+ "step": 1588
+ },
+ {
+ "epoch": 22.069868995633186,
+ "grad_norm": 0.021214453503489494,
+ "learning_rate": 0.0006,
+ "loss": 4.418305397033691,
+ "step": 1589
+ },
+ {
+ "epoch": 22.083842794759825,
+ "grad_norm": 0.019051678478717804,
+ "learning_rate": 0.0006,
+ "loss": 4.215126991271973,
+ "step": 1590
+ },
+ {
+ "epoch": 22.097816593886463,
+ "grad_norm": 0.02323366142809391,
+ "learning_rate": 0.0006,
+ "loss": 4.306807994842529,
+ "step": 1591
+ },
+ {
+ "epoch": 22.111790393013102,
+ "grad_norm": 0.026186738163232803,
+ "learning_rate": 0.0006,
+ "loss": 4.408508777618408,
+ "step": 1592
+ },
+ {
+ "epoch": 22.125764192139737,
+ "grad_norm": 0.027361227199435234,
+ "learning_rate": 0.0006,
+ "loss": 4.329322814941406,
+ "step": 1593
+ },
+ {
+ "epoch": 22.139737991266376,
+ "grad_norm": 0.026564646512269974,
+ "learning_rate": 0.0006,
+ "loss": 4.346642017364502,
+ "step": 1594
+ },
+ {
+ "epoch": 22.153711790393015,
+ "grad_norm": 0.025073667988181114,
+ "learning_rate": 0.0006,
+ "loss": 4.3984198570251465,
+ "step": 1595
+ },
+ {
+ "epoch": 22.16768558951965,
+ "grad_norm": 0.02347794733941555,
+ "learning_rate": 0.0006,
+ "loss": 4.352538108825684,
+ "step": 1596
+ },
+ {
+ "epoch": 22.18165938864629,
+ "grad_norm": 0.02203468047082424,
+ "learning_rate": 0.0006,
+ "loss": 4.298148155212402,
+ "step": 1597
+ },
+ {
+ "epoch": 22.195633187772927,
+ "grad_norm": 0.020621957257390022,
+ "learning_rate": 0.0006,
+ "loss": 4.480309963226318,
+ "step": 1598
+ },
+ {
+ "epoch": 22.209606986899562,
+ "grad_norm": 0.023710055276751518,
+ "learning_rate": 0.0006,
+ "loss": 4.339317321777344,
+ "step": 1599
+ },
+ {
+ "epoch": 22.2235807860262,
+ "grad_norm": 0.0233482476323843,
+ "learning_rate": 0.0006,
+ "loss": 4.436896324157715,
+ "step": 1600
+ },
+ {
+ "epoch": 22.23755458515284,
+ "grad_norm": 0.020122403278946877,
+ "learning_rate": 0.0006,
+ "loss": 4.236002445220947,
+ "step": 1601
+ },
+ {
+ "epoch": 22.251528384279474,
+ "grad_norm": 0.01957266591489315,
+ "learning_rate": 0.0006,
+ "loss": 4.497328281402588,
+ "step": 1602
+ },
+ {
+ "epoch": 22.265502183406113,
+ "grad_norm": 0.018964895978569984,
+ "learning_rate": 0.0006,
+ "loss": 4.36360502243042,
+ "step": 1603
+ },
+ {
+ "epoch": 22.27947598253275,
+ "grad_norm": 0.018902769312262535,
+ "learning_rate": 0.0006,
+ "loss": 4.427093505859375,
+ "step": 1604
+ },
+ {
+ "epoch": 22.29344978165939,
+ "grad_norm": 0.018918119370937347,
+ "learning_rate": 0.0006,
+ "loss": 4.294950485229492,
+ "step": 1605
+ },
+ {
+ "epoch": 22.307423580786025,
+ "grad_norm": 0.019315775483846664,
+ "learning_rate": 0.0006,
+ "loss": 4.427952289581299,
+ "step": 1606
+ },
+ {
+ "epoch": 22.321397379912664,
+ "grad_norm": 0.018350522965192795,
+ "learning_rate": 0.0006,
+ "loss": 4.272945404052734,
+ "step": 1607
+ },
+ {
+ "epoch": 22.335371179039303,
+ "grad_norm": 0.01759534701704979,
+ "learning_rate": 0.0006,
+ "loss": 4.348023891448975,
+ "step": 1608
+ },
+ {
+ "epoch": 22.349344978165938,
+ "grad_norm": 0.01646057888865471,
+ "learning_rate": 0.0006,
+ "loss": 4.278518199920654,
+ "step": 1609
+ },
+ {
+ "epoch": 22.363318777292577,
+ "grad_norm": 0.015875060111284256,
+ "learning_rate": 0.0006,
+ "loss": 4.423527240753174,
+ "step": 1610
+ },
+ {
+ "epoch": 22.377292576419215,
+ "grad_norm": 0.01638529635965824,
+ "learning_rate": 0.0006,
+ "loss": 4.321723937988281,
+ "step": 1611
+ },
+ {
+ "epoch": 22.39126637554585,
+ "grad_norm": 0.016942061483860016,
+ "learning_rate": 0.0006,
+ "loss": 4.346164703369141,
+ "step": 1612
+ },
+ {
+ "epoch": 22.40524017467249,
+ "grad_norm": 0.015209621749818325,
+ "learning_rate": 0.0006,
+ "loss": 4.434340000152588,
+ "step": 1613
+ },
+ {
+ "epoch": 22.419213973799128,
+ "grad_norm": 0.014423009008169174,
+ "learning_rate": 0.0006,
+ "loss": 4.342560768127441,
+ "step": 1614
+ },
+ {
+ "epoch": 22.433187772925763,
+ "grad_norm": 0.014369525946676731,
+ "learning_rate": 0.0006,
+ "loss": 4.296586990356445,
+ "step": 1615
+ },
+ {
+ "epoch": 22.4471615720524,
+ "grad_norm": 0.01392540242522955,
+ "learning_rate": 0.0006,
+ "loss": 4.433176040649414,
+ "step": 1616
+ },
+ {
+ "epoch": 22.46113537117904,
+ "grad_norm": 0.013206899166107178,
+ "learning_rate": 0.0006,
+ "loss": 4.32962703704834,
+ "step": 1617
+ },
+ {
+ "epoch": 22.475109170305675,
+ "grad_norm": 0.013427077792584896,
+ "learning_rate": 0.0006,
+ "loss": 4.420437335968018,
+ "step": 1618
+ },
+ {
+ "epoch": 22.489082969432314,
+ "grad_norm": 0.014194196090102196,
+ "learning_rate": 0.0006,
+ "loss": 4.410478591918945,
+ "step": 1619
+ },
+ {
+ "epoch": 22.503056768558952,
+ "grad_norm": 0.01362364087253809,
+ "learning_rate": 0.0006,
+ "loss": 4.338932037353516,
+ "step": 1620
+ },
+ {
+ "epoch": 22.51703056768559,
+ "grad_norm": 0.014299589209258556,
+ "learning_rate": 0.0006,
+ "loss": 4.34089994430542,
+ "step": 1621
+ },
+ {
+ "epoch": 22.531004366812226,
+ "grad_norm": 0.014061295427381992,
+ "learning_rate": 0.0006,
+ "loss": 4.29201602935791,
+ "step": 1622
+ },
+ {
+ "epoch": 22.544978165938865,
+ "grad_norm": 0.013945229351520538,
+ "learning_rate": 0.0006,
+ "loss": 4.326052188873291,
+ "step": 1623
+ },
+ {
+ "epoch": 22.558951965065503,
+ "grad_norm": 0.01387068536132574,
+ "learning_rate": 0.0006,
+ "loss": 4.393521308898926,
+ "step": 1624
+ },
+ {
+ "epoch": 22.57292576419214,
+ "grad_norm": 0.013931898400187492,
+ "learning_rate": 0.0006,
+ "loss": 4.3739752769470215,
+ "step": 1625
+ },
+ {
+ "epoch": 22.586899563318777,
+ "grad_norm": 0.01497740764170885,
+ "learning_rate": 0.0006,
+ "loss": 4.3991618156433105,
+ "step": 1626
+ },
+ {
+ "epoch": 22.600873362445416,
+ "grad_norm": 0.015842905268073082,
+ "learning_rate": 0.0006,
+ "loss": 4.405340671539307,
+ "step": 1627
+ },
+ {
+ "epoch": 22.61484716157205,
+ "grad_norm": 0.01580633595585823,
+ "learning_rate": 0.0006,
+ "loss": 4.280102729797363,
+ "step": 1628
+ },
+ {
+ "epoch": 22.62882096069869,
+ "grad_norm": 0.016567382961511612,
+ "learning_rate": 0.0006,
+ "loss": 4.407657146453857,
+ "step": 1629
+ },
+ {
+ "epoch": 22.64279475982533,
+ "grad_norm": 0.01648622378706932,
+ "learning_rate": 0.0006,
+ "loss": 4.392590522766113,
+ "step": 1630
+ },
+ {
+ "epoch": 22.656768558951963,
+ "grad_norm": 0.017066307365894318,
+ "learning_rate": 0.0006,
+ "loss": 4.324862003326416,
+ "step": 1631
+ },
+ {
+ "epoch": 22.670742358078602,
+ "grad_norm": 0.01637062057852745,
+ "learning_rate": 0.0006,
+ "loss": 4.34598445892334,
+ "step": 1632
+ },
+ {
+ "epoch": 22.68471615720524,
+ "grad_norm": 0.018038960173726082,
+ "learning_rate": 0.0006,
+ "loss": 4.320711612701416,
+ "step": 1633
+ },
+ {
+ "epoch": 22.69868995633188,
+ "grad_norm": 0.018096428364515305,
+ "learning_rate": 0.0006,
+ "loss": 4.536527156829834,
+ "step": 1634
+ },
+ {
+ "epoch": 22.712663755458514,
+ "grad_norm": 0.017889291048049927,
+ "learning_rate": 0.0006,
+ "loss": 4.388131141662598,
+ "step": 1635
+ },
+ {
+ "epoch": 22.726637554585153,
+ "grad_norm": 0.018771521747112274,
+ "learning_rate": 0.0006,
+ "loss": 4.357668876647949,
+ "step": 1636
+ },
+ {
+ "epoch": 22.74061135371179,
+ "grad_norm": 0.02100873365998268,
+ "learning_rate": 0.0006,
+ "loss": 4.293972015380859,
+ "step": 1637
+ },
+ {
+ "epoch": 22.754585152838427,
+ "grad_norm": 0.017715785652399063,
+ "learning_rate": 0.0006,
+ "loss": 4.349849224090576,
+ "step": 1638
+ },
+ {
+ "epoch": 22.768558951965066,
+ "grad_norm": 0.015546333976089954,
+ "learning_rate": 0.0006,
+ "loss": 4.311728000640869,
+ "step": 1639
+ },
+ {
+ "epoch": 22.782532751091704,
+ "grad_norm": 0.01605990156531334,
+ "learning_rate": 0.0006,
+ "loss": 4.499123573303223,
+ "step": 1640
+ },
+ {
+ "epoch": 22.79650655021834,
+ "grad_norm": 0.016642142087221146,
+ "learning_rate": 0.0006,
+ "loss": 4.4319257736206055,
+ "step": 1641
+ },
+ {
+ "epoch": 22.810480349344978,
+ "grad_norm": 0.016120387241244316,
+ "learning_rate": 0.0006,
+ "loss": 4.4373650550842285,
+ "step": 1642
+ },
+ {
+ "epoch": 22.824454148471617,
+ "grad_norm": 0.015174890868365765,
+ "learning_rate": 0.0006,
+ "loss": 4.314082145690918,
+ "step": 1643
+ },
+ {
+ "epoch": 22.83842794759825,
+ "grad_norm": 0.01551714539527893,
+ "learning_rate": 0.0006,
+ "loss": 4.360363960266113,
+ "step": 1644
+ },
+ {
+ "epoch": 22.85240174672489,
+ "grad_norm": 0.014966742135584354,
+ "learning_rate": 0.0006,
+ "loss": 4.346253395080566,
+ "step": 1645
+ },
+ {
+ "epoch": 22.86637554585153,
+ "grad_norm": 0.014288028702139854,
+ "learning_rate": 0.0006,
+ "loss": 4.464366436004639,
+ "step": 1646
+ },
+ {
+ "epoch": 22.880349344978168,
+ "grad_norm": 0.015291682444512844,
+ "learning_rate": 0.0006,
+ "loss": 4.504101753234863,
+ "step": 1647
+ },
+ {
+ "epoch": 22.894323144104803,
+ "grad_norm": 0.014534426853060722,
+ "learning_rate": 0.0006,
+ "loss": 4.462172031402588,
+ "step": 1648
+ },
+ {
+ "epoch": 22.90829694323144,
+ "grad_norm": 0.014129819348454475,
+ "learning_rate": 0.0006,
+ "loss": 4.350126266479492,
+ "step": 1649
+ },
+ {
+ "epoch": 22.92227074235808,
+ "grad_norm": 0.014932668767869473,
+ "learning_rate": 0.0006,
+ "loss": 4.374392032623291,
+ "step": 1650
+ },
+ {
+ "epoch": 22.936244541484715,
+ "grad_norm": 0.014647853560745716,
+ "learning_rate": 0.0006,
+ "loss": 4.252264499664307,
+ "step": 1651
+ },
+ {
+ "epoch": 22.950218340611354,
+ "grad_norm": 0.013553360477089882,
+ "learning_rate": 0.0006,
+ "loss": 4.379729270935059,
+ "step": 1652
+ },
+ {
+ "epoch": 22.964192139737992,
+ "grad_norm": 0.014606194570660591,
+ "learning_rate": 0.0006,
+ "loss": 4.363372802734375,
+ "step": 1653
+ },
+ {
+ "epoch": 22.978165938864628,
+ "grad_norm": 0.015000631101429462,
+ "learning_rate": 0.0006,
+ "loss": 4.408048629760742,
+ "step": 1654
+ },
+ {
+ "epoch": 22.992139737991266,
+ "grad_norm": 0.015370816923677921,
+ "learning_rate": 0.0006,
+ "loss": 4.297451019287109,
+ "step": 1655
+ },
+ {
+ "epoch": 23.0,
+ "grad_norm": 0.01709144562482834,
+ "learning_rate": 0.0006,
+ "loss": 4.294776916503906,
+ "step": 1656
+ },
+ {
+ "epoch": 23.0,
+ "eval_loss": 4.678618907928467,
+ "eval_runtime": 56.1668,
+ "eval_samples_per_second": 43.478,
+ "eval_steps_per_second": 1.371,
+ "step": 1656
+ },
+ {
+ "epoch": 23.01397379912664,
+ "grad_norm": 0.01727290451526642,
+ "learning_rate": 0.0006,
+ "loss": 4.311701774597168,
+ "step": 1657
+ },
+ {
+ "epoch": 23.027947598253274,
+ "grad_norm": 0.020645759999752045,
+ "learning_rate": 0.0006,
+ "loss": 4.429897308349609,
+ "step": 1658
+ },
+ {
+ "epoch": 23.041921397379912,
+ "grad_norm": 0.02204226516187191,
+ "learning_rate": 0.0006,
+ "loss": 4.365151882171631,
+ "step": 1659
+ },
+ {
+ "epoch": 23.05589519650655,
+ "grad_norm": 0.023069269955158234,
+ "learning_rate": 0.0006,
+ "loss": 4.3710832595825195,
+ "step": 1660
+ },
+ {
+ "epoch": 23.069868995633186,
+ "grad_norm": 0.024151179939508438,
+ "learning_rate": 0.0006,
+ "loss": 4.403108596801758,
+ "step": 1661
+ },
+ {
+ "epoch": 23.083842794759825,
+ "grad_norm": 0.02511233650147915,
+ "learning_rate": 0.0006,
+ "loss": 4.414128303527832,
+ "step": 1662
+ },
+ {
+ "epoch": 23.097816593886463,
+ "grad_norm": 0.025802113115787506,
+ "learning_rate": 0.0006,
+ "loss": 4.388265609741211,
+ "step": 1663
+ },
+ {
+ "epoch": 23.111790393013102,
+ "grad_norm": 0.02377348765730858,
+ "learning_rate": 0.0006,
+ "loss": 4.357505798339844,
+ "step": 1664
+ },
+ {
+ "epoch": 23.125764192139737,
+ "grad_norm": 0.020876651629805565,
+ "learning_rate": 0.0006,
+ "loss": 4.23043966293335,
+ "step": 1665
+ },
+ {
+ "epoch": 23.139737991266376,
+ "grad_norm": 0.01959727331995964,
+ "learning_rate": 0.0006,
+ "loss": 4.280970096588135,
+ "step": 1666
+ },
+ {
+ "epoch": 23.153711790393015,
+ "grad_norm": 0.018678616732358932,
+ "learning_rate": 0.0006,
+ "loss": 4.346716403961182,
+ "step": 1667
+ },
+ {
+ "epoch": 23.16768558951965,
+ "grad_norm": 0.02108944021165371,
+ "learning_rate": 0.0006,
+ "loss": 4.277904510498047,
+ "step": 1668
+ },
+ {
+ "epoch": 23.18165938864629,
+ "grad_norm": 0.020221566781401634,
+ "learning_rate": 0.0006,
+ "loss": 4.367693901062012,
+ "step": 1669
+ },
+ {
+ "epoch": 23.195633187772927,
+ "grad_norm": 0.019599711522459984,
+ "learning_rate": 0.0006,
+ "loss": 4.257092475891113,
+ "step": 1670
+ },
+ {
+ "epoch": 23.209606986899562,
+ "grad_norm": 0.019833361729979515,
+ "learning_rate": 0.0006,
+ "loss": 4.540737628936768,
+ "step": 1671
+ },
+ {
+ "epoch": 23.2235807860262,
+ "grad_norm": 0.020352095365524292,
+ "learning_rate": 0.0006,
+ "loss": 4.360923767089844,
+ "step": 1672
+ },
+ {
+ "epoch": 23.23755458515284,
+ "grad_norm": 0.01939748041331768,
+ "learning_rate": 0.0006,
+ "loss": 4.393110275268555,
+ "step": 1673
+ },
+ {
+ "epoch": 23.251528384279474,
+ "grad_norm": 0.01884354278445244,
+ "learning_rate": 0.0006,
+ "loss": 4.383938789367676,
+ "step": 1674
+ },
+ {
+ "epoch": 23.265502183406113,
+ "grad_norm": 0.02008470520377159,
+ "learning_rate": 0.0006,
+ "loss": 4.346344947814941,
+ "step": 1675
+ },
+ {
+ "epoch": 23.27947598253275,
+ "grad_norm": 0.019065195694565773,
+ "learning_rate": 0.0006,
+ "loss": 4.36611795425415,
+ "step": 1676
+ },
+ {
+ "epoch": 23.29344978165939,
+ "grad_norm": 0.01938541606068611,
+ "learning_rate": 0.0006,
+ "loss": 4.4009599685668945,
+ "step": 1677
+ },
+ {
+ "epoch": 23.307423580786025,
+ "grad_norm": 0.019044723361730576,
+ "learning_rate": 0.0006,
+ "loss": 4.340169906616211,
+ "step": 1678
+ },
+ {
+ "epoch": 23.321397379912664,
+ "grad_norm": 0.018850378692150116,
+ "learning_rate": 0.0006,
+ "loss": 4.37674617767334,
+ "step": 1679
+ },
+ {
+ "epoch": 23.335371179039303,
+ "grad_norm": 0.019076095893979073,
+ "learning_rate": 0.0006,
+ "loss": 4.226874351501465,
+ "step": 1680
+ },
+ {
+ "epoch": 23.349344978165938,
+ "grad_norm": 0.017543038353323936,
+ "learning_rate": 0.0006,
+ "loss": 4.337751388549805,
+ "step": 1681
+ },
+ {
+ "epoch": 23.363318777292577,
+ "grad_norm": 0.017170218750834465,
+ "learning_rate": 0.0006,
+ "loss": 4.282512664794922,
+ "step": 1682
+ },
+ {
+ "epoch": 23.377292576419215,
+ "grad_norm": 0.01574796624481678,
+ "learning_rate": 0.0006,
+ "loss": 4.30638313293457,
+ "step": 1683
+ },
+ {
+ "epoch": 23.39126637554585,
+ "grad_norm": 0.015160846523940563,
+ "learning_rate": 0.0006,
+ "loss": 4.336716175079346,
+ "step": 1684
+ },
+ {
+ "epoch": 23.40524017467249,
+ "grad_norm": 0.01534526702016592,
+ "learning_rate": 0.0006,
+ "loss": 4.2174787521362305,
+ "step": 1685
+ },
+ {
+ "epoch": 23.419213973799128,
+ "grad_norm": 0.015289999544620514,
+ "learning_rate": 0.0006,
+ "loss": 4.272546291351318,
+ "step": 1686
+ },
+ {
+ "epoch": 23.433187772925763,
+ "grad_norm": 0.014894185587763786,
+ "learning_rate": 0.0006,
+ "loss": 4.296045303344727,
+ "step": 1687
+ },
+ {
+ "epoch": 23.4471615720524,
+ "grad_norm": 0.014869502745568752,
+ "learning_rate": 0.0006,
+ "loss": 4.425461769104004,
+ "step": 1688
+ },
+ {
+ "epoch": 23.46113537117904,
+ "grad_norm": 0.01474483497440815,
+ "learning_rate": 0.0006,
+ "loss": 4.361867427825928,
+ "step": 1689
+ },
+ {
+ "epoch": 23.475109170305675,
+ "grad_norm": 0.015299245715141296,
+ "learning_rate": 0.0006,
+ "loss": 4.364523410797119,
+ "step": 1690
+ },
+ {
+ "epoch": 23.489082969432314,
+ "grad_norm": 0.015201340429484844,
+ "learning_rate": 0.0006,
+ "loss": 4.285182952880859,
+ "step": 1691
+ },
+ {
+ "epoch": 23.503056768558952,
+ "grad_norm": 0.016057293862104416,
+ "learning_rate": 0.0006,
+ "loss": 4.284873962402344,
+ "step": 1692
+ },
+ {
+ "epoch": 23.51703056768559,
+ "grad_norm": 0.015871062874794006,
+ "learning_rate": 0.0006,
+ "loss": 4.326470851898193,
+ "step": 1693
+ },
+ {
+ "epoch": 23.531004366812226,
+ "grad_norm": 0.01586288772523403,
+ "learning_rate": 0.0006,
+ "loss": 4.324883460998535,
+ "step": 1694
+ },
+ {
+ "epoch": 23.544978165938865,
+ "grad_norm": 0.016092827543616295,
+ "learning_rate": 0.0006,
+ "loss": 4.3258867263793945,
+ "step": 1695
+ },
+ {
+ "epoch": 23.558951965065503,
+ "grad_norm": 0.01582016795873642,
+ "learning_rate": 0.0006,
+ "loss": 4.358892440795898,
+ "step": 1696
+ },
+ {
+ "epoch": 23.57292576419214,
+ "grad_norm": 0.01688452623784542,
+ "learning_rate": 0.0006,
+ "loss": 4.280842304229736,
+ "step": 1697
+ },
+ {
+ "epoch": 23.586899563318777,
+ "grad_norm": 0.015865886583924294,
+ "learning_rate": 0.0006,
+ "loss": 4.245567321777344,
+ "step": 1698
+ },
+ {
+ "epoch": 23.600873362445416,
+ "grad_norm": 0.015303662046790123,
+ "learning_rate": 0.0006,
+ "loss": 4.383112907409668,
+ "step": 1699
+ },
+ {
+ "epoch": 23.61484716157205,
+ "grad_norm": 0.015649516135454178,
+ "learning_rate": 0.0006,
+ "loss": 4.519082069396973,
+ "step": 1700
+ },
+ {
+ "epoch": 23.62882096069869,
+ "grad_norm": 0.0167120099067688,
+ "learning_rate": 0.0006,
+ "loss": 4.447847366333008,
+ "step": 1701
+ },
+ {
+ "epoch": 23.64279475982533,
+ "grad_norm": 0.016368716955184937,
+ "learning_rate": 0.0006,
+ "loss": 4.434186935424805,
+ "step": 1702
+ },
+ {
+ "epoch": 23.656768558951963,
+ "grad_norm": 0.017379503697156906,
+ "learning_rate": 0.0006,
+ "loss": 4.317594528198242,
+ "step": 1703
+ },
+ {
+ "epoch": 23.670742358078602,
+ "grad_norm": 0.018573811277747154,
+ "learning_rate": 0.0006,
+ "loss": 4.205793857574463,
+ "step": 1704
+ },
+ {
+ "epoch": 23.68471615720524,
+ "grad_norm": 0.01749090477824211,
+ "learning_rate": 0.0006,
+ "loss": 4.3985700607299805,
+ "step": 1705
+ },
+ {
+ "epoch": 23.69868995633188,
+ "grad_norm": 0.0165668074041605,
+ "learning_rate": 0.0006,
+ "loss": 4.329226970672607,
+ "step": 1706
+ },
+ {
+ "epoch": 23.712663755458514,
+ "grad_norm": 0.016913846135139465,
+ "learning_rate": 0.0006,
+ "loss": 4.388485908508301,
+ "step": 1707
+ },
+ {
+ "epoch": 23.726637554585153,
+ "grad_norm": 0.016813859343528748,
+ "learning_rate": 0.0006,
+ "loss": 4.3395915031433105,
+ "step": 1708
+ },
+ {
+ "epoch": 23.74061135371179,
+ "grad_norm": 0.01676975190639496,
+ "learning_rate": 0.0006,
+ "loss": 4.254745960235596,
+ "step": 1709
+ },
+ {
+ "epoch": 23.754585152838427,
+ "grad_norm": 0.016370350494980812,
+ "learning_rate": 0.0006,
+ "loss": 4.2612223625183105,
+ "step": 1710
+ },
+ {
+ "epoch": 23.768558951965066,
+ "grad_norm": 0.01696198247373104,
+ "learning_rate": 0.0006,
+ "loss": 4.327524662017822,
+ "step": 1711
+ },
+ {
+ "epoch": 23.782532751091704,
+ "grad_norm": 0.0189108457416296,
+ "learning_rate": 0.0006,
+ "loss": 4.325839519500732,
+ "step": 1712
+ },
+ {
+ "epoch": 23.79650655021834,
+ "grad_norm": 0.01862351782619953,
+ "learning_rate": 0.0006,
+ "loss": 4.333662509918213,
+ "step": 1713
+ },
+ {
+ "epoch": 23.810480349344978,
+ "grad_norm": 0.01655014418065548,
+ "learning_rate": 0.0006,
+ "loss": 4.39305305480957,
+ "step": 1714
+ },
+ {
+ "epoch": 23.824454148471617,
+ "grad_norm": 0.015899376943707466,
+ "learning_rate": 0.0006,
+ "loss": 4.390053749084473,
+ "step": 1715
+ },
+ {
+ "epoch": 23.83842794759825,
+ "grad_norm": 0.016519056633114815,
+ "learning_rate": 0.0006,
+ "loss": 4.255903244018555,
+ "step": 1716
+ },
+ {
+ "epoch": 23.85240174672489,
+ "grad_norm": 0.0167539119720459,
+ "learning_rate": 0.0006,
+ "loss": 4.255727767944336,
+ "step": 1717
+ },
+ {
+ "epoch": 23.86637554585153,
+ "grad_norm": 0.0171353816986084,
+ "learning_rate": 0.0006,
+ "loss": 4.376740455627441,
+ "step": 1718
+ },
+ {
+ "epoch": 23.880349344978168,
+ "grad_norm": 0.016719117760658264,
+ "learning_rate": 0.0006,
+ "loss": 4.3376688957214355,
+ "step": 1719
+ },
+ {
+ "epoch": 23.894323144104803,
+ "grad_norm": 0.01561494916677475,
+ "learning_rate": 0.0006,
+ "loss": 4.3345746994018555,
+ "step": 1720
+ },
+ {
+ "epoch": 23.90829694323144,
+ "grad_norm": 0.016303174197673798,
+ "learning_rate": 0.0006,
+ "loss": 4.332371234893799,
+ "step": 1721
+ },
+ {
+ "epoch": 23.92227074235808,
+ "grad_norm": 0.016722865402698517,
+ "learning_rate": 0.0006,
+ "loss": 4.267125129699707,
+ "step": 1722
+ },
+ {
+ "epoch": 23.936244541484715,
+ "grad_norm": 0.017072124406695366,
+ "learning_rate": 0.0006,
+ "loss": 4.322751998901367,
+ "step": 1723
+ },
+ {
+ "epoch": 23.950218340611354,
+ "grad_norm": 0.01699932850897312,
+ "learning_rate": 0.0006,
+ "loss": 4.378422737121582,
+ "step": 1724
+ },
+ {
+ "epoch": 23.964192139737992,
+ "grad_norm": 0.01742720976471901,
+ "learning_rate": 0.0006,
+ "loss": 4.345083236694336,
+ "step": 1725
+ },
+ {
+ "epoch": 23.978165938864628,
+ "grad_norm": 0.018797501921653748,
+ "learning_rate": 0.0006,
+ "loss": 4.290443420410156,
+ "step": 1726
+ },
+ {
+ "epoch": 23.992139737991266,
+ "grad_norm": 0.018090683966875076,
+ "learning_rate": 0.0006,
+ "loss": 4.377580642700195,
+ "step": 1727
+ },
+ {
+ "epoch": 24.0,
+ "grad_norm": 0.017305459827184677,
+ "learning_rate": 0.0006,
+ "loss": 4.364426612854004,
+ "step": 1728
+ },
+ {
+ "epoch": 24.0,
+ "eval_loss": 4.647762298583984,
+ "eval_runtime": 56.5402,
+ "eval_samples_per_second": 43.191,
+ "eval_steps_per_second": 1.362,
+ "step": 1728
+ },
+ {
+ "epoch": 24.01397379912664,
+ "grad_norm": 0.016493607312440872,
+ "learning_rate": 0.0006,
+ "loss": 4.252255916595459,
+ "step": 1729
+ },
+ {
+ "epoch": 24.027947598253274,
+ "grad_norm": 0.016529450193047523,
+ "learning_rate": 0.0006,
+ "loss": 4.352746963500977,
+ "step": 1730
+ },
+ {
+ "epoch": 24.041921397379912,
+ "grad_norm": 0.016940191388130188,
+ "learning_rate": 0.0006,
+ "loss": 4.301870822906494,
+ "step": 1731
+ },
+ {
+ "epoch": 24.05589519650655,
+ "grad_norm": 0.01752558909356594,
+ "learning_rate": 0.0006,
+ "loss": 4.2762451171875,
+ "step": 1732
+ },
+ {
+ "epoch": 24.069868995633186,
+ "grad_norm": 0.017112884670495987,
+ "learning_rate": 0.0006,
+ "loss": 4.2659149169921875,
+ "step": 1733
+ },
+ {
+ "epoch": 24.083842794759825,
+ "grad_norm": 0.01900586113333702,
+ "learning_rate": 0.0006,
+ "loss": 4.306058883666992,
+ "step": 1734
+ },
+ {
+ "epoch": 24.097816593886463,
+ "grad_norm": 0.019447680562734604,
+ "learning_rate": 0.0006,
+ "loss": 4.450128555297852,
+ "step": 1735
+ },
+ {
+ "epoch": 24.111790393013102,
+ "grad_norm": 0.01948186755180359,
+ "learning_rate": 0.0006,
+ "loss": 4.202869415283203,
+ "step": 1736
+ },
+ {
+ "epoch": 24.125764192139737,
+ "grad_norm": 0.01716383546590805,
+ "learning_rate": 0.0006,
+ "loss": 4.318876266479492,
+ "step": 1737
+ },
+ {
+ "epoch": 24.139737991266376,
+ "grad_norm": 0.017049988731741905,
+ "learning_rate": 0.0006,
+ "loss": 4.237238883972168,
+ "step": 1738
+ },
+ {
+ "epoch": 24.153711790393015,
+ "grad_norm": 0.015169923193752766,
+ "learning_rate": 0.0006,
+ "loss": 4.289426803588867,
+ "step": 1739
+ },
+ {
+ "epoch": 24.16768558951965,
+ "grad_norm": 0.014760667458176613,
+ "learning_rate": 0.0006,
+ "loss": 4.356924057006836,
+ "step": 1740
+ },
+ {
+ "epoch": 24.18165938864629,
+ "grad_norm": 0.015127750113606453,
+ "learning_rate": 0.0006,
+ "loss": 4.142916679382324,
+ "step": 1741
+ },
+ {
+ "epoch": 24.195633187772927,
+ "grad_norm": 0.01549555640667677,
+ "learning_rate": 0.0006,
+ "loss": 4.345563888549805,
+ "step": 1742
+ },
+ {
+ "epoch": 24.209606986899562,
+ "grad_norm": 0.016393091529607773,
+ "learning_rate": 0.0006,
+ "loss": 4.417495250701904,
+ "step": 1743
+ },
+ {
+ "epoch": 24.2235807860262,
+ "grad_norm": 0.01785266026854515,
+ "learning_rate": 0.0006,
+ "loss": 4.3553547859191895,
+ "step": 1744
+ },
+ {
+ "epoch": 24.23755458515284,
+ "grad_norm": 0.017574617639183998,
+ "learning_rate": 0.0006,
+ "loss": 4.241024494171143,
+ "step": 1745
+ },
+ {
+ "epoch": 24.251528384279474,
+ "grad_norm": 0.019060306251049042,
+ "learning_rate": 0.0006,
+ "loss": 4.3685150146484375,
+ "step": 1746
+ },
+ {
+ "epoch": 24.265502183406113,
+ "grad_norm": 0.0214633010327816,
+ "learning_rate": 0.0006,
+ "loss": 4.323866844177246,
+ "step": 1747
+ },
+ {
+ "epoch": 24.27947598253275,
+ "grad_norm": 0.021055003628134727,
+ "learning_rate": 0.0006,
+ "loss": 4.278757095336914,
+ "step": 1748
+ },
+ {
+ "epoch": 24.29344978165939,
+ "grad_norm": 0.018618572503328323,
+ "learning_rate": 0.0006,
+ "loss": 4.28849983215332,
+ "step": 1749
+ },
+ {
+ "epoch": 24.307423580786025,
+ "grad_norm": 0.018666435033082962,
+ "learning_rate": 0.0006,
+ "loss": 4.368590354919434,
+ "step": 1750
+ },
+ {
+ "epoch": 24.321397379912664,
+ "grad_norm": 0.020626146346330643,
+ "learning_rate": 0.0006,
+ "loss": 4.2951836585998535,
+ "step": 1751
+ },
+ {
+ "epoch": 24.335371179039303,
+ "grad_norm": 0.01929759792983532,
+ "learning_rate": 0.0006,
+ "loss": 4.3664703369140625,
+ "step": 1752
+ },
+ {
+ "epoch": 24.349344978165938,
+ "grad_norm": 0.020956002175807953,
+ "learning_rate": 0.0006,
+ "loss": 4.345308303833008,
+ "step": 1753
+ },
+ {
+ "epoch": 24.363318777292577,
+ "grad_norm": 0.019227957352995872,
+ "learning_rate": 0.0006,
+ "loss": 4.349849224090576,
+ "step": 1754
+ },
+ {
+ "epoch": 24.377292576419215,
+ "grad_norm": 0.015970097854733467,
+ "learning_rate": 0.0006,
+ "loss": 4.223012447357178,
+ "step": 1755
+ },
+ {
+ "epoch": 24.39126637554585,
+ "grad_norm": 0.015951646491885185,
+ "learning_rate": 0.0006,
+ "loss": 4.236420631408691,
+ "step": 1756
+ },
+ {
+ "epoch": 24.40524017467249,
+ "grad_norm": 0.016032515093684196,
+ "learning_rate": 0.0006,
+ "loss": 4.226899147033691,
+ "step": 1757
+ },
+ {
+ "epoch": 24.419213973799128,
+ "grad_norm": 0.01573132909834385,
+ "learning_rate": 0.0006,
+ "loss": 4.315916061401367,
+ "step": 1758
+ },
+ {
+ "epoch": 24.433187772925763,
+ "grad_norm": 0.015367005951702595,
+ "learning_rate": 0.0006,
+ "loss": 4.358939170837402,
+ "step": 1759
+ },
+ {
+ "epoch": 24.4471615720524,
+ "grad_norm": 0.01669907011091709,
+ "learning_rate": 0.0006,
+ "loss": 4.232147216796875,
+ "step": 1760
+ },
+ {
+ "epoch": 24.46113537117904,
+ "grad_norm": 0.018708152696490288,
+ "learning_rate": 0.0006,
+ "loss": 4.329852104187012,
+ "step": 1761
+ },
+ {
+ "epoch": 24.475109170305675,
+ "grad_norm": 0.019035987555980682,
+ "learning_rate": 0.0006,
+ "loss": 4.260340213775635,
+ "step": 1762
+ },
+ {
+ "epoch": 24.489082969432314,
+ "grad_norm": 0.01832144521176815,
+ "learning_rate": 0.0006,
+ "loss": 4.359543800354004,
+ "step": 1763
+ },
+ {
+ "epoch": 24.503056768558952,
+ "grad_norm": 0.015313294716179371,
+ "learning_rate": 0.0006,
+ "loss": 4.279088020324707,
+ "step": 1764
+ },
+ {
+ "epoch": 24.51703056768559,
+ "grad_norm": 0.015803923830389977,
+ "learning_rate": 0.0006,
+ "loss": 4.230682849884033,
+ "step": 1765
+ },
+ {
+ "epoch": 24.531004366812226,
+ "grad_norm": 0.016748374328017235,
+ "learning_rate": 0.0006,
+ "loss": 4.349661827087402,
+ "step": 1766
+ },
+ {
+ "epoch": 24.544978165938865,
+ "grad_norm": 0.015732428058981895,
+ "learning_rate": 0.0006,
+ "loss": 4.372441291809082,
+ "step": 1767
+ },
+ {
+ "epoch": 24.558951965065503,
+ "grad_norm": 0.016564829275012016,
+ "learning_rate": 0.0006,
+ "loss": 4.323090553283691,
+ "step": 1768
+ },
+ {
+ "epoch": 24.57292576419214,
+ "grad_norm": 0.01603010483086109,
+ "learning_rate": 0.0006,
+ "loss": 4.347898006439209,
+ "step": 1769
+ },
+ {
+ "epoch": 24.586899563318777,
+ "grad_norm": 0.01674809865653515,
+ "learning_rate": 0.0006,
+ "loss": 4.269165515899658,
+ "step": 1770
+ },
+ {
+ "epoch": 24.600873362445416,
+ "grad_norm": 0.017333930358290672,
+ "learning_rate": 0.0006,
+ "loss": 4.353298187255859,
+ "step": 1771
+ },
+ {
+ "epoch": 24.61484716157205,
+ "grad_norm": 0.017519580200314522,
+ "learning_rate": 0.0006,
+ "loss": 4.265199661254883,
+ "step": 1772
+ },
+ {
+ "epoch": 24.62882096069869,
+ "grad_norm": 0.017675306648015976,
+ "learning_rate": 0.0006,
+ "loss": 4.322596549987793,
+ "step": 1773
+ },
+ {
+ "epoch": 24.64279475982533,
+ "grad_norm": 0.017166534438729286,
+ "learning_rate": 0.0006,
+ "loss": 4.413900375366211,
+ "step": 1774
+ },
+ {
+ "epoch": 24.656768558951963,
+ "grad_norm": 0.016362346708774567,
+ "learning_rate": 0.0006,
+ "loss": 4.307236671447754,
+ "step": 1775
+ },
+ {
+ "epoch": 24.670742358078602,
+ "grad_norm": 0.016086872667074203,
+ "learning_rate": 0.0006,
+ "loss": 4.235316276550293,
+ "step": 1776
+ },
+ {
+ "epoch": 24.68471615720524,
+ "grad_norm": 0.01689537614583969,
+ "learning_rate": 0.0006,
+ "loss": 4.390591621398926,
+ "step": 1777
+ },
+ {
+ "epoch": 24.69868995633188,
+ "grad_norm": 0.016836725175380707,
+ "learning_rate": 0.0006,
+ "loss": 4.2990031242370605,
+ "step": 1778
+ },
+ {
+ "epoch": 24.712663755458514,
+ "grad_norm": 0.015627335757017136,
+ "learning_rate": 0.0006,
+ "loss": 4.166745185852051,
+ "step": 1779
+ },
+ {
+ "epoch": 24.726637554585153,
+ "grad_norm": 0.015007016249001026,
+ "learning_rate": 0.0006,
+ "loss": 4.295273303985596,
+ "step": 1780
+ },
+ {
+ "epoch": 24.74061135371179,
+ "grad_norm": 0.016084209084510803,
+ "learning_rate": 0.0006,
+ "loss": 4.342424392700195,
+ "step": 1781
+ },
+ {
+ "epoch": 24.754585152838427,
+ "grad_norm": 0.01595073565840721,
+ "learning_rate": 0.0006,
+ "loss": 4.3060102462768555,
+ "step": 1782
+ },
+ {
+ "epoch": 24.768558951965066,
+ "grad_norm": 0.016113461926579475,
+ "learning_rate": 0.0006,
+ "loss": 4.267908096313477,
+ "step": 1783
+ },
+ {
+ "epoch": 24.782532751091704,
+ "grad_norm": 0.01564556173980236,
+ "learning_rate": 0.0006,
+ "loss": 4.324642658233643,
+ "step": 1784
+ },
+ {
+ "epoch": 24.79650655021834,
+ "grad_norm": 0.01575268618762493,
+ "learning_rate": 0.0006,
+ "loss": 4.328181743621826,
+ "step": 1785
+ },
+ {
+ "epoch": 24.810480349344978,
+ "grad_norm": 0.016830824315547943,
+ "learning_rate": 0.0006,
+ "loss": 4.247906684875488,
+ "step": 1786
+ },
+ {
+ "epoch": 24.824454148471617,
+ "grad_norm": 0.017989547923207283,
+ "learning_rate": 0.0006,
+ "loss": 4.228281021118164,
+ "step": 1787
+ },
+ {
+ "epoch": 24.83842794759825,
+ "grad_norm": 0.016783086583018303,
+ "learning_rate": 0.0006,
+ "loss": 4.2880964279174805,
+ "step": 1788
+ },
+ {
+ "epoch": 24.85240174672489,
+ "grad_norm": 0.016467561945319176,
+ "learning_rate": 0.0006,
+ "loss": 4.36118745803833,
+ "step": 1789
+ },
+ {
+ "epoch": 24.86637554585153,
+ "grad_norm": 0.01780523732304573,
+ "learning_rate": 0.0006,
+ "loss": 4.172321319580078,
+ "step": 1790
+ },
+ {
+ "epoch": 24.880349344978168,
+ "grad_norm": 0.017102031037211418,
+ "learning_rate": 0.0006,
+ "loss": 4.352826118469238,
+ "step": 1791
+ },
+ {
+ "epoch": 24.894323144104803,
+ "grad_norm": 0.016409218311309814,
+ "learning_rate": 0.0006,
+ "loss": 4.212162017822266,
+ "step": 1792
+ },
+ {
+ "epoch": 24.90829694323144,
+ "grad_norm": 0.018660584464669228,
+ "learning_rate": 0.0006,
+ "loss": 4.346684455871582,
+ "step": 1793
+ },
+ {
+ "epoch": 24.92227074235808,
+ "grad_norm": 0.017389440909028053,
+ "learning_rate": 0.0006,
+ "loss": 4.197913646697998,
+ "step": 1794
+ },
+ {
+ "epoch": 24.936244541484715,
+ "grad_norm": 0.01596485823392868,
+ "learning_rate": 0.0006,
+ "loss": 4.370617866516113,
+ "step": 1795
+ },
+ {
+ "epoch": 24.950218340611354,
+ "grad_norm": 0.01709776371717453,
+ "learning_rate": 0.0006,
+ "loss": 4.313847064971924,
+ "step": 1796
+ },
+ {
+ "epoch": 24.964192139737992,
+ "grad_norm": 0.016571134328842163,
+ "learning_rate": 0.0006,
+ "loss": 4.1895856857299805,
+ "step": 1797
+ },
+ {
+ "epoch": 24.978165938864628,
+ "grad_norm": 0.016851048916578293,
+ "learning_rate": 0.0006,
+ "loss": 4.350772857666016,
+ "step": 1798
+ },
+ {
+ "epoch": 24.992139737991266,
+ "grad_norm": 0.017007440328598022,
+ "learning_rate": 0.0006,
+ "loss": 4.399008750915527,
+ "step": 1799
+ },
+ {
+ "epoch": 25.0,
+ "grad_norm": 0.01791212148964405,
+ "learning_rate": 0.0006,
+ "loss": 4.306121826171875,
+ "step": 1800
+ },
+ {
+ "epoch": 25.0,
+ "eval_loss": 4.636417865753174,
+ "eval_runtime": 56.8296,
+ "eval_samples_per_second": 42.971,
+ "eval_steps_per_second": 1.355,
+ "step": 1800
+ },
+ {
+ "epoch": 25.01397379912664,
+ "grad_norm": 0.01729617640376091,
+ "learning_rate": 0.0006,
+ "loss": 4.257018089294434,
+ "step": 1801
+ },
+ {
+ "epoch": 25.027947598253274,
+ "grad_norm": 0.017466643825173378,
+ "learning_rate": 0.0006,
+ "loss": 4.307605266571045,
+ "step": 1802
+ },
+ {
+ "epoch": 25.041921397379912,
+ "grad_norm": 0.018614279106259346,
+ "learning_rate": 0.0006,
+ "loss": 4.360141277313232,
+ "step": 1803
+ },
+ {
+ "epoch": 25.05589519650655,
+ "grad_norm": 0.02010771445930004,
+ "learning_rate": 0.0006,
+ "loss": 4.196428298950195,
+ "step": 1804
+ },
+ {
+ "epoch": 25.069868995633186,
+ "grad_norm": 0.02090228535234928,
+ "learning_rate": 0.0006,
+ "loss": 4.288947105407715,
+ "step": 1805
+ },
+ {
+ "epoch": 25.083842794759825,
+ "grad_norm": 0.02050255797803402,
+ "learning_rate": 0.0006,
+ "loss": 4.259582042694092,
+ "step": 1806
+ },
+ {
+ "epoch": 25.097816593886463,
+ "grad_norm": 0.021844089031219482,
+ "learning_rate": 0.0006,
+ "loss": 4.405138969421387,
+ "step": 1807
+ },
+ {
+ "epoch": 25.111790393013102,
+ "grad_norm": 0.022287018597126007,
+ "learning_rate": 0.0006,
+ "loss": 4.333793640136719,
+ "step": 1808
+ },
+ {
+ "epoch": 25.125764192139737,
+ "grad_norm": 0.02288329415023327,
+ "learning_rate": 0.0006,
+ "loss": 4.3910980224609375,
+ "step": 1809
+ },
+ {
+ "epoch": 25.139737991266376,
+ "grad_norm": 0.02373199723660946,
+ "learning_rate": 0.0006,
+ "loss": 4.32921028137207,
+ "step": 1810
+ },
+ {
+ "epoch": 25.153711790393015,
+ "grad_norm": 0.024960234761238098,
+ "learning_rate": 0.0006,
+ "loss": 4.295609474182129,
+ "step": 1811
+ },
+ {
+ "epoch": 25.16768558951965,
+ "grad_norm": 0.026743892580270767,
+ "learning_rate": 0.0006,
+ "loss": 4.315308094024658,
+ "step": 1812
+ },
+ {
+ "epoch": 25.18165938864629,
+ "grad_norm": 0.022138185799121857,
+ "learning_rate": 0.0006,
+ "loss": 4.360208511352539,
+ "step": 1813
+ },
+ {
+ "epoch": 25.195633187772927,
+ "grad_norm": 0.022120574489235878,
+ "learning_rate": 0.0006,
+ "loss": 4.1973772048950195,
+ "step": 1814
+ },
+ {
+ "epoch": 25.209606986899562,
+ "grad_norm": 0.02230502851307392,
+ "learning_rate": 0.0006,
+ "loss": 4.285558700561523,
+ "step": 1815
+ },
+ {
+ "epoch": 25.2235807860262,
+ "grad_norm": 0.022301986813545227,
+ "learning_rate": 0.0006,
+ "loss": 4.238343238830566,
+ "step": 1816
+ },
+ {
+ "epoch": 25.23755458515284,
+ "grad_norm": 0.020050447434186935,
+ "learning_rate": 0.0006,
+ "loss": 4.298235893249512,
+ "step": 1817
+ },
+ {
+ "epoch": 25.251528384279474,
+ "grad_norm": 0.01987724006175995,
+ "learning_rate": 0.0006,
+ "loss": 4.301384449005127,
+ "step": 1818
+ },
+ {
+ "epoch": 25.265502183406113,
+ "grad_norm": 0.020440855994820595,
+ "learning_rate": 0.0006,
+ "loss": 4.269218444824219,
+ "step": 1819
+ },
+ {
+ "epoch": 25.27947598253275,
+ "grad_norm": 0.01891893707215786,
+ "learning_rate": 0.0006,
+ "loss": 4.2401957511901855,
+ "step": 1820
+ },
+ {
+ "epoch": 25.29344978165939,
+ "grad_norm": 0.01903688907623291,
+ "learning_rate": 0.0006,
+ "loss": 4.139670372009277,
+ "step": 1821
+ },
+ {
+ "epoch": 25.307423580786025,
+ "grad_norm": 0.018856938928365707,
+ "learning_rate": 0.0006,
+ "loss": 4.233214378356934,
+ "step": 1822
+ },
+ {
+ "epoch": 25.321397379912664,
+ "grad_norm": 0.017791323363780975,
+ "learning_rate": 0.0006,
+ "loss": 4.2342023849487305,
+ "step": 1823
+ },
+ {
+ "epoch": 25.335371179039303,
+ "grad_norm": 0.017304565757513046,
+ "learning_rate": 0.0006,
+ "loss": 4.355198860168457,
+ "step": 1824
+ },
+ {
+ "epoch": 25.349344978165938,
+ "grad_norm": 0.01576206088066101,
+ "learning_rate": 0.0006,
+ "loss": 4.114147186279297,
+ "step": 1825
+ },
+ {
+ "epoch": 25.363318777292577,
+ "grad_norm": 0.015105154365301132,
+ "learning_rate": 0.0006,
+ "loss": 4.313795566558838,
+ "step": 1826
+ },
+ {
+ "epoch": 25.377292576419215,
+ "grad_norm": 0.01563587784767151,
+ "learning_rate": 0.0006,
+ "loss": 4.308282852172852,
+ "step": 1827
+ },
+ {
+ "epoch": 25.39126637554585,
+ "grad_norm": 0.015980906784534454,
+ "learning_rate": 0.0006,
+ "loss": 4.345149040222168,
+ "step": 1828
+ },
+ {
+ "epoch": 25.40524017467249,
+ "grad_norm": 0.01669284701347351,
+ "learning_rate": 0.0006,
+ "loss": 4.237611293792725,
+ "step": 1829
+ },
+ {
+ "epoch": 25.419213973799128,
+ "grad_norm": 0.016202064231038094,
+ "learning_rate": 0.0006,
+ "loss": 4.3072309494018555,
+ "step": 1830
+ },
+ {
+ "epoch": 25.433187772925763,
+ "grad_norm": 0.01478662807494402,
+ "learning_rate": 0.0006,
+ "loss": 4.276215076446533,
+ "step": 1831
+ },
+ {
+ "epoch": 25.4471615720524,
+ "grad_norm": 0.015168731100857258,
+ "learning_rate": 0.0006,
+ "loss": 4.37509822845459,
+ "step": 1832
+ },
+ {
+ "epoch": 25.46113537117904,
+ "grad_norm": 0.01648980937898159,
+ "learning_rate": 0.0006,
+ "loss": 4.349167346954346,
+ "step": 1833
+ },
+ {
+ "epoch": 25.475109170305675,
+ "grad_norm": 0.01708398200571537,
+ "learning_rate": 0.0006,
+ "loss": 4.346580505371094,
+ "step": 1834
+ },
+ {
+ "epoch": 25.489082969432314,
+ "grad_norm": 0.017370784655213356,
+ "learning_rate": 0.0006,
+ "loss": 4.299466133117676,
+ "step": 1835
+ },
+ {
+ "epoch": 25.503056768558952,
+ "grad_norm": 0.016248228028416634,
+ "learning_rate": 0.0006,
+ "loss": 4.184610366821289,
+ "step": 1836
+ },
+ {
+ "epoch": 25.51703056768559,
+ "grad_norm": 0.017106913030147552,
+ "learning_rate": 0.0006,
+ "loss": 4.366856098175049,
+ "step": 1837
+ },
+ {
+ "epoch": 25.531004366812226,
+ "grad_norm": 0.016071965917944908,
+ "learning_rate": 0.0006,
+ "loss": 4.297691345214844,
+ "step": 1838
+ },
+ {
+ "epoch": 25.544978165938865,
+ "grad_norm": 0.016451986506581306,
+ "learning_rate": 0.0006,
+ "loss": 4.229983329772949,
+ "step": 1839
+ },
+ {
+ "epoch": 25.558951965065503,
+ "grad_norm": 0.016195081174373627,
+ "learning_rate": 0.0006,
+ "loss": 4.259893417358398,
+ "step": 1840
+ },
+ {
+ "epoch": 25.57292576419214,
+ "grad_norm": 0.0165175162255764,
+ "learning_rate": 0.0006,
+ "loss": 4.3348236083984375,
+ "step": 1841
+ },
+ {
+ "epoch": 25.586899563318777,
+ "grad_norm": 0.01651517115533352,
+ "learning_rate": 0.0006,
+ "loss": 4.319541931152344,
+ "step": 1842
+ },
+ {
+ "epoch": 25.600873362445416,
+ "grad_norm": 0.016143690794706345,
+ "learning_rate": 0.0006,
+ "loss": 4.345652103424072,
+ "step": 1843
+ },
+ {
+ "epoch": 25.61484716157205,
+ "grad_norm": 0.015140696428716183,
+ "learning_rate": 0.0006,
+ "loss": 4.212047576904297,
+ "step": 1844
+ },
+ {
+ "epoch": 25.62882096069869,
+ "grad_norm": 0.014637443237006664,
+ "learning_rate": 0.0006,
+ "loss": 4.187453269958496,
+ "step": 1845
+ },
+ {
+ "epoch": 25.64279475982533,
+ "grad_norm": 0.014585614204406738,
+ "learning_rate": 0.0006,
+ "loss": 4.253936767578125,
+ "step": 1846
+ },
+ {
+ "epoch": 25.656768558951963,
+ "grad_norm": 0.015158289112150669,
+ "learning_rate": 0.0006,
+ "loss": 4.34612512588501,
+ "step": 1847
+ },
+ {
+ "epoch": 25.670742358078602,
+ "grad_norm": 0.015202849172055721,
+ "learning_rate": 0.0006,
+ "loss": 4.270634651184082,
+ "step": 1848
+ },
+ {
+ "epoch": 25.68471615720524,
+ "grad_norm": 0.015363575890660286,
+ "learning_rate": 0.0006,
+ "loss": 4.271080493927002,
+ "step": 1849
+ },
+ {
+ "epoch": 25.69868995633188,
+ "grad_norm": 0.014360226690769196,
+ "learning_rate": 0.0006,
+ "loss": 4.375033855438232,
+ "step": 1850
+ },
+ {
+ "epoch": 25.712663755458514,
+ "grad_norm": 0.014263181947171688,
+ "learning_rate": 0.0006,
+ "loss": 4.424138069152832,
+ "step": 1851
+ },
+ {
+ "epoch": 25.726637554585153,
+ "grad_norm": 0.014398688450455666,
+ "learning_rate": 0.0006,
+ "loss": 4.410154819488525,
+ "step": 1852
+ },
+ {
+ "epoch": 25.74061135371179,
+ "grad_norm": 0.016131488606333733,
+ "learning_rate": 0.0006,
+ "loss": 4.324014663696289,
+ "step": 1853
+ },
+ {
+ "epoch": 25.754585152838427,
+ "grad_norm": 0.017568735405802727,
+ "learning_rate": 0.0006,
+ "loss": 4.367861747741699,
+ "step": 1854
+ },
+ {
+ "epoch": 25.768558951965066,
+ "grad_norm": 0.017907511442899704,
+ "learning_rate": 0.0006,
+ "loss": 4.352965354919434,
+ "step": 1855
+ },
+ {
+ "epoch": 25.782532751091704,
+ "grad_norm": 0.016918016597628593,
+ "learning_rate": 0.0006,
+ "loss": 4.252397537231445,
+ "step": 1856
+ },
+ {
+ "epoch": 25.79650655021834,
+ "grad_norm": 0.017274610698223114,
+ "learning_rate": 0.0006,
+ "loss": 4.342309951782227,
+ "step": 1857
+ },
+ {
+ "epoch": 25.810480349344978,
+ "grad_norm": 0.017879825085401535,
+ "learning_rate": 0.0006,
+ "loss": 4.330987453460693,
+ "step": 1858
+ },
+ {
+ "epoch": 25.824454148471617,
+ "grad_norm": 0.01736099272966385,
+ "learning_rate": 0.0006,
+ "loss": 4.307255744934082,
+ "step": 1859
+ },
+ {
+ "epoch": 25.83842794759825,
+ "grad_norm": 0.016575824469327927,
+ "learning_rate": 0.0006,
+ "loss": 4.262016773223877,
+ "step": 1860
+ },
+ {
+ "epoch": 25.85240174672489,
+ "grad_norm": 0.015181140042841434,
+ "learning_rate": 0.0006,
+ "loss": 4.291863441467285,
+ "step": 1861
+ },
+ {
+ "epoch": 25.86637554585153,
+ "grad_norm": 0.015965687111020088,
+ "learning_rate": 0.0006,
+ "loss": 4.424836158752441,
+ "step": 1862
+ },
+ {
+ "epoch": 25.880349344978168,
+ "grad_norm": 0.01687219925224781,
+ "learning_rate": 0.0006,
+ "loss": 4.343531131744385,
+ "step": 1863
+ },
+ {
+ "epoch": 25.894323144104803,
+ "grad_norm": 0.015115504153072834,
+ "learning_rate": 0.0006,
+ "loss": 4.357808589935303,
+ "step": 1864
+ },
+ {
+ "epoch": 25.90829694323144,
+ "grad_norm": 0.014150998555123806,
+ "learning_rate": 0.0006,
+ "loss": 4.324550151824951,
+ "step": 1865
+ },
+ {
+ "epoch": 25.92227074235808,
+ "grad_norm": 0.014989510178565979,
+ "learning_rate": 0.0006,
+ "loss": 4.380428314208984,
+ "step": 1866
+ },
+ {
+ "epoch": 25.936244541484715,
+ "grad_norm": 0.01527960691601038,
+ "learning_rate": 0.0006,
+ "loss": 4.399374961853027,
+ "step": 1867
+ },
+ {
+ "epoch": 25.950218340611354,
+ "grad_norm": 0.015434633940458298,
+ "learning_rate": 0.0006,
+ "loss": 4.260984420776367,
+ "step": 1868
+ },
+ {
+ "epoch": 25.964192139737992,
+ "grad_norm": 0.015370228327810764,
+ "learning_rate": 0.0006,
+ "loss": 4.3340253829956055,
+ "step": 1869
+ },
+ {
+ "epoch": 25.978165938864628,
+ "grad_norm": 0.015348542481660843,
+ "learning_rate": 0.0006,
+ "loss": 4.369132995605469,
+ "step": 1870
+ },
+ {
+ "epoch": 25.992139737991266,
+ "grad_norm": 0.015354936942458153,
+ "learning_rate": 0.0006,
+ "loss": 4.381937026977539,
+ "step": 1871
+ },
+ {
+ "epoch": 26.0,
+ "grad_norm": 0.01796272210776806,
+ "learning_rate": 0.0006,
+ "loss": 4.286768436431885,
+ "step": 1872
+ },
+ {
+ "epoch": 26.0,
+ "eval_loss": 4.576984405517578,
+ "eval_runtime": 57.2091,
+ "eval_samples_per_second": 42.685,
+ "eval_steps_per_second": 1.346,
+ "step": 1872
+ },
+ {
+ "epoch": 26.01397379912664,
+ "grad_norm": 0.01748845726251602,
+ "learning_rate": 0.0006,
+ "loss": 4.337504863739014,
+ "step": 1873
+ },
+ {
+ "epoch": 26.027947598253274,
+ "grad_norm": 0.019244296476244926,
+ "learning_rate": 0.0006,
+ "loss": 4.264280796051025,
+ "step": 1874
+ },
+ {
+ "epoch": 26.041921397379912,
+ "grad_norm": 0.0215692650526762,
+ "learning_rate": 0.0006,
+ "loss": 4.350830078125,
+ "step": 1875
+ },
+ {
+ "epoch": 26.05589519650655,
+ "grad_norm": 0.022189252078533173,
+ "learning_rate": 0.0006,
+ "loss": 4.299098968505859,
+ "step": 1876
+ },
+ {
+ "epoch": 26.069868995633186,
+ "grad_norm": 0.022384043782949448,
+ "learning_rate": 0.0006,
+ "loss": 4.1185173988342285,
+ "step": 1877
+ },
+ {
+ "epoch": 26.083842794759825,
+ "grad_norm": 0.02076014317572117,
+ "learning_rate": 0.0006,
+ "loss": 4.297612190246582,
+ "step": 1878
+ },
+ {
+ "epoch": 26.097816593886463,
+ "grad_norm": 0.020843051373958588,
+ "learning_rate": 0.0006,
+ "loss": 4.294098854064941,
+ "step": 1879
+ },
+ {
+ "epoch": 26.111790393013102,
+ "grad_norm": 0.020654456689953804,
+ "learning_rate": 0.0006,
+ "loss": 4.267856597900391,
+ "step": 1880
+ },
+ {
+ "epoch": 26.125764192139737,
+ "grad_norm": 0.01860472559928894,
+ "learning_rate": 0.0006,
+ "loss": 4.284232139587402,
+ "step": 1881
+ },
+ {
+ "epoch": 26.139737991266376,
+ "grad_norm": 0.018719421699643135,
+ "learning_rate": 0.0006,
+ "loss": 4.277214527130127,
+ "step": 1882
+ },
+ {
+ "epoch": 26.153711790393015,
+ "grad_norm": 0.020903829485177994,
+ "learning_rate": 0.0006,
+ "loss": 4.254947662353516,
+ "step": 1883
+ },
+ {
+ "epoch": 26.16768558951965,
+ "grad_norm": 0.02020149491727352,
+ "learning_rate": 0.0006,
+ "loss": 4.2526936531066895,
+ "step": 1884
+ },
+ {
+ "epoch": 26.18165938864629,
+ "grad_norm": 0.0180392786860466,
+ "learning_rate": 0.0006,
+ "loss": 4.309922695159912,
+ "step": 1885
+ },
+ {
+ "epoch": 26.195633187772927,
+ "grad_norm": 0.018163125962018967,
+ "learning_rate": 0.0006,
+ "loss": 4.13860559463501,
+ "step": 1886
+ },
+ {
+ "epoch": 26.209606986899562,
+ "grad_norm": 0.01853291131556034,
+ "learning_rate": 0.0006,
+ "loss": 4.190926551818848,
+ "step": 1887
+ },
+ {
+ "epoch": 26.2235807860262,
+ "grad_norm": 0.01757677271962166,
+ "learning_rate": 0.0006,
+ "loss": 4.163999557495117,
+ "step": 1888
+ },
+ {
+ "epoch": 26.23755458515284,
+ "grad_norm": 0.01767992228269577,
+ "learning_rate": 0.0006,
+ "loss": 4.184430122375488,
+ "step": 1889
+ },
+ {
+ "epoch": 26.251528384279474,
+ "grad_norm": 0.018139973282814026,
+ "learning_rate": 0.0006,
+ "loss": 4.300766944885254,
+ "step": 1890
+ },
+ {
+ "epoch": 26.265502183406113,
+ "grad_norm": 0.020184461027383804,
+ "learning_rate": 0.0006,
+ "loss": 4.232537269592285,
+ "step": 1891
+ },
+ {
+ "epoch": 26.27947598253275,
+ "grad_norm": 0.02028856985270977,
+ "learning_rate": 0.0006,
+ "loss": 4.324883460998535,
+ "step": 1892
+ },
+ {
+ "epoch": 26.29344978165939,
+ "grad_norm": 0.017120616510510445,
+ "learning_rate": 0.0006,
+ "loss": 4.23493766784668,
+ "step": 1893
+ },
+ {
+ "epoch": 26.307423580786025,
+ "grad_norm": 0.016202978789806366,
+ "learning_rate": 0.0006,
+ "loss": 4.197602272033691,
+ "step": 1894
+ },
+ {
+ "epoch": 26.321397379912664,
+ "grad_norm": 0.017231421545147896,
+ "learning_rate": 0.0006,
+ "loss": 4.302001953125,
+ "step": 1895
+ },
+ {
+ "epoch": 26.335371179039303,
+ "grad_norm": 0.01725655607879162,
+ "learning_rate": 0.0006,
+ "loss": 4.356800079345703,
+ "step": 1896
+ },
+ {
+ "epoch": 26.349344978165938,
+ "grad_norm": 0.016657443717122078,
+ "learning_rate": 0.0006,
+ "loss": 4.266307830810547,
+ "step": 1897
+ },
+ {
+ "epoch": 26.363318777292577,
+ "grad_norm": 0.01616556942462921,
+ "learning_rate": 0.0006,
+ "loss": 4.307868003845215,
+ "step": 1898
+ },
+ {
+ "epoch": 26.377292576419215,
+ "grad_norm": 0.016621306538581848,
+ "learning_rate": 0.0006,
+ "loss": 4.172905921936035,
+ "step": 1899
+ },
+ {
+ "epoch": 26.39126637554585,
+ "grad_norm": 0.016243206337094307,
+ "learning_rate": 0.0006,
+ "loss": 4.162896156311035,
+ "step": 1900
+ },
+ {
+ "epoch": 26.40524017467249,
+ "grad_norm": 0.016663808375597,
+ "learning_rate": 0.0006,
+ "loss": 4.258232116699219,
+ "step": 1901
+ },
+ {
+ "epoch": 26.419213973799128,
+ "grad_norm": 0.01677478663623333,
+ "learning_rate": 0.0006,
+ "loss": 4.336090087890625,
+ "step": 1902
+ },
+ {
+ "epoch": 26.433187772925763,
+ "grad_norm": 0.01641303487122059,
+ "learning_rate": 0.0006,
+ "loss": 4.191476821899414,
+ "step": 1903
+ },
+ {
+ "epoch": 26.4471615720524,
+ "grad_norm": 0.016700323671102524,
+ "learning_rate": 0.0006,
+ "loss": 4.299427509307861,
+ "step": 1904
+ },
+ {
+ "epoch": 26.46113537117904,
+ "grad_norm": 0.017595035955309868,
+ "learning_rate": 0.0006,
+ "loss": 4.312828063964844,
+ "step": 1905
+ },
+ {
+ "epoch": 26.475109170305675,
+ "grad_norm": 0.020300284028053284,
+ "learning_rate": 0.0006,
+ "loss": 4.28331184387207,
+ "step": 1906
+ },
+ {
+ "epoch": 26.489082969432314,
+ "grad_norm": 0.02198321744799614,
+ "learning_rate": 0.0006,
+ "loss": 4.3280839920043945,
+ "step": 1907
+ },
+ {
+ "epoch": 26.503056768558952,
+ "grad_norm": 0.02003200724720955,
+ "learning_rate": 0.0006,
+ "loss": 4.325623989105225,
+ "step": 1908
+ },
+ {
+ "epoch": 26.51703056768559,
+ "grad_norm": 0.01870287023484707,
+ "learning_rate": 0.0006,
+ "loss": 4.251203536987305,
+ "step": 1909
+ },
+ {
+ "epoch": 26.531004366812226,
+ "grad_norm": 0.01835448481142521,
+ "learning_rate": 0.0006,
+ "loss": 4.335144996643066,
+ "step": 1910
+ },
+ {
+ "epoch": 26.544978165938865,
+ "grad_norm": 0.02046327292919159,
+ "learning_rate": 0.0006,
+ "loss": 4.238023281097412,
+ "step": 1911
+ },
+ {
+ "epoch": 26.558951965065503,
+ "grad_norm": 0.019771311432123184,
+ "learning_rate": 0.0006,
+ "loss": 4.353594779968262,
+ "step": 1912
+ },
+ {
+ "epoch": 26.57292576419214,
+ "grad_norm": 0.018090544268488884,
+ "learning_rate": 0.0006,
+ "loss": 4.304343223571777,
+ "step": 1913
+ },
+ {
+ "epoch": 26.586899563318777,
+ "grad_norm": 0.017489777877926826,
+ "learning_rate": 0.0006,
+ "loss": 4.204375267028809,
+ "step": 1914
+ },
+ {
+ "epoch": 26.600873362445416,
+ "grad_norm": 0.016785891726613045,
+ "learning_rate": 0.0006,
+ "loss": 4.286669731140137,
+ "step": 1915
+ },
+ {
+ "epoch": 26.61484716157205,
+ "grad_norm": 0.016382789239287376,
+ "learning_rate": 0.0006,
+ "loss": 4.347662448883057,
+ "step": 1916
+ },
+ {
+ "epoch": 26.62882096069869,
+ "grad_norm": 0.016581876203417778,
+ "learning_rate": 0.0006,
+ "loss": 4.2716193199157715,
+ "step": 1917
+ },
+ {
+ "epoch": 26.64279475982533,
+ "grad_norm": 0.015627184882760048,
+ "learning_rate": 0.0006,
+ "loss": 4.311489105224609,
+ "step": 1918
+ },
+ {
+ "epoch": 26.656768558951963,
+ "grad_norm": 0.015908483415842056,
+ "learning_rate": 0.0006,
+ "loss": 4.262099742889404,
+ "step": 1919
+ },
+ {
+ "epoch": 26.670742358078602,
+ "grad_norm": 0.015576236881315708,
+ "learning_rate": 0.0006,
+ "loss": 4.275871276855469,
+ "step": 1920
+ },
+ {
+ "epoch": 26.68471615720524,
+ "grad_norm": 0.014621064998209476,
+ "learning_rate": 0.0006,
+ "loss": 4.184260845184326,
+ "step": 1921
+ },
+ {
+ "epoch": 26.69868995633188,
+ "grad_norm": 0.014797811396420002,
+ "learning_rate": 0.0006,
+ "loss": 4.3412628173828125,
+ "step": 1922
+ },
+ {
+ "epoch": 26.712663755458514,
+ "grad_norm": 0.015610570088028908,
+ "learning_rate": 0.0006,
+ "loss": 4.181094646453857,
+ "step": 1923
+ },
+ {
+ "epoch": 26.726637554585153,
+ "grad_norm": 0.016048265621066093,
+ "learning_rate": 0.0006,
+ "loss": 4.365290641784668,
+ "step": 1924
+ },
+ {
+ "epoch": 26.74061135371179,
+ "grad_norm": 0.016604071483016014,
+ "learning_rate": 0.0006,
+ "loss": 4.2628912925720215,
+ "step": 1925
+ },
+ {
+ "epoch": 26.754585152838427,
+ "grad_norm": 0.017044221982359886,
+ "learning_rate": 0.0006,
+ "loss": 4.378512382507324,
+ "step": 1926
+ },
+ {
+ "epoch": 26.768558951965066,
+ "grad_norm": 0.01746384985744953,
+ "learning_rate": 0.0006,
+ "loss": 4.3888773918151855,
+ "step": 1927
+ },
+ {
+ "epoch": 26.782532751091704,
+ "grad_norm": 0.01572711206972599,
+ "learning_rate": 0.0006,
+ "loss": 4.189935684204102,
+ "step": 1928
+ },
+ {
+ "epoch": 26.79650655021834,
+ "grad_norm": 0.014976629987359047,
+ "learning_rate": 0.0006,
+ "loss": 4.280525207519531,
+ "step": 1929
+ },
+ {
+ "epoch": 26.810480349344978,
+ "grad_norm": 0.014804664999246597,
+ "learning_rate": 0.0006,
+ "loss": 4.270630836486816,
+ "step": 1930
+ },
+ {
+ "epoch": 26.824454148471617,
+ "grad_norm": 0.014641920104622841,
+ "learning_rate": 0.0006,
+ "loss": 4.319505214691162,
+ "step": 1931
+ },
+ {
+ "epoch": 26.83842794759825,
+ "grad_norm": 0.014928505755960941,
+ "learning_rate": 0.0006,
+ "loss": 4.180147647857666,
+ "step": 1932
+ },
+ {
+ "epoch": 26.85240174672489,
+ "grad_norm": 0.016130365431308746,
+ "learning_rate": 0.0006,
+ "loss": 4.260660648345947,
+ "step": 1933
+ },
+ {
+ "epoch": 26.86637554585153,
+ "grad_norm": 0.015642493963241577,
+ "learning_rate": 0.0006,
+ "loss": 4.304064750671387,
+ "step": 1934
+ },
+ {
+ "epoch": 26.880349344978168,
+ "grad_norm": 0.015988217666745186,
+ "learning_rate": 0.0006,
+ "loss": 4.2548017501831055,
+ "step": 1935
+ },
+ {
+ "epoch": 26.894323144104803,
+ "grad_norm": 0.016501398757100105,
+ "learning_rate": 0.0006,
+ "loss": 4.260058879852295,
+ "step": 1936
+ },
+ {
+ "epoch": 26.90829694323144,
+ "grad_norm": 0.01645040698349476,
+ "learning_rate": 0.0006,
+ "loss": 4.483081340789795,
+ "step": 1937
+ },
+ {
+ "epoch": 26.92227074235808,
+ "grad_norm": 0.016140863299369812,
+ "learning_rate": 0.0006,
+ "loss": 4.1776838302612305,
+ "step": 1938
+ },
+ {
+ "epoch": 26.936244541484715,
+ "grad_norm": 0.016681019216775894,
+ "learning_rate": 0.0006,
+ "loss": 4.224747657775879,
+ "step": 1939
+ },
+ {
+ "epoch": 26.950218340611354,
+ "grad_norm": 0.01650378853082657,
+ "learning_rate": 0.0006,
+ "loss": 4.305703163146973,
+ "step": 1940
+ },
+ {
+ "epoch": 26.964192139737992,
+ "grad_norm": 0.017051683738827705,
+ "learning_rate": 0.0006,
+ "loss": 4.285835266113281,
+ "step": 1941
+ },
+ {
+ "epoch": 26.978165938864628,
+ "grad_norm": 0.016894327476620674,
+ "learning_rate": 0.0006,
+ "loss": 4.318901538848877,
+ "step": 1942
+ },
+ {
+ "epoch": 26.992139737991266,
+ "grad_norm": 0.017313100397586823,
+ "learning_rate": 0.0006,
+ "loss": 4.281434535980225,
+ "step": 1943
+ },
+ {
+ "epoch": 27.0,
+ "grad_norm": 0.018450884148478508,
+ "learning_rate": 0.0006,
+ "loss": 4.335104942321777,
+ "step": 1944
+ },
+ {
+ "epoch": 27.0,
+ "eval_loss": 4.640687465667725,
+ "eval_runtime": 56.9624,
+ "eval_samples_per_second": 42.87,
+ "eval_steps_per_second": 1.352,
+ "step": 1944
+ },
+ {
+ "epoch": 27.01397379912664,
+ "grad_norm": 0.017755698412656784,
+ "learning_rate": 0.0006,
+ "loss": 4.278824329376221,
+ "step": 1945
+ },
+ {
+ "epoch": 27.027947598253274,
+ "grad_norm": 0.01724310778081417,
+ "learning_rate": 0.0006,
+ "loss": 4.311049461364746,
+ "step": 1946
+ },
+ {
+ "epoch": 27.041921397379912,
+ "grad_norm": 0.01690123789012432,
+ "learning_rate": 0.0006,
+ "loss": 4.261994361877441,
+ "step": 1947
+ },
+ {
+ "epoch": 27.05589519650655,
+ "grad_norm": 0.018406696617603302,
+ "learning_rate": 0.0006,
+ "loss": 4.1925153732299805,
+ "step": 1948
+ },
+ {
+ "epoch": 27.069868995633186,
+ "grad_norm": 0.019945867359638214,
+ "learning_rate": 0.0006,
+ "loss": 4.108537197113037,
+ "step": 1949
+ },
+ {
+ "epoch": 27.083842794759825,
+ "grad_norm": 0.020316628739237785,
+ "learning_rate": 0.0006,
+ "loss": 4.209043025970459,
+ "step": 1950
+ },
+ {
+ "epoch": 27.097816593886463,
+ "grad_norm": 0.019690370187163353,
+ "learning_rate": 0.0006,
+ "loss": 4.195789337158203,
+ "step": 1951
+ },
+ {
+ "epoch": 27.111790393013102,
+ "grad_norm": 0.01785232499241829,
+ "learning_rate": 0.0006,
+ "loss": 4.21466064453125,
+ "step": 1952
+ },
+ {
+ "epoch": 27.125764192139737,
+ "grad_norm": 0.0164767038077116,
+ "learning_rate": 0.0006,
+ "loss": 4.270247459411621,
+ "step": 1953
+ },
+ {
+ "epoch": 27.139737991266376,
+ "grad_norm": 0.018008455634117126,
+ "learning_rate": 0.0006,
+ "loss": 4.201042175292969,
+ "step": 1954
+ },
+ {
+ "epoch": 27.153711790393015,
+ "grad_norm": 0.01807340793311596,
+ "learning_rate": 0.0006,
+ "loss": 4.25289249420166,
+ "step": 1955
+ },
+ {
+ "epoch": 27.16768558951965,
+ "grad_norm": 0.016246909275650978,
+ "learning_rate": 0.0006,
+ "loss": 4.116283416748047,
+ "step": 1956
+ },
+ {
+ "epoch": 27.18165938864629,
+ "grad_norm": 0.017545776441693306,
+ "learning_rate": 0.0006,
+ "loss": 4.316999435424805,
+ "step": 1957
+ },
+ {
+ "epoch": 27.195633187772927,
+ "grad_norm": 0.017915375530719757,
+ "learning_rate": 0.0006,
+ "loss": 4.253859519958496,
+ "step": 1958
+ },
+ {
+ "epoch": 27.209606986899562,
+ "grad_norm": 0.018112223595380783,
+ "learning_rate": 0.0006,
+ "loss": 4.2541022300720215,
+ "step": 1959
+ },
+ {
+ "epoch": 27.2235807860262,
+ "grad_norm": 0.01846139505505562,
+ "learning_rate": 0.0006,
+ "loss": 4.20438289642334,
+ "step": 1960
+ },
+ {
+ "epoch": 27.23755458515284,
+ "grad_norm": 0.01943897269666195,
+ "learning_rate": 0.0006,
+ "loss": 4.2799577713012695,
+ "step": 1961
+ },
+ {
+ "epoch": 27.251528384279474,
+ "grad_norm": 0.018253348767757416,
+ "learning_rate": 0.0006,
+ "loss": 4.1720499992370605,
+ "step": 1962
+ },
+ {
+ "epoch": 27.265502183406113,
+ "grad_norm": 0.018292246386408806,
+ "learning_rate": 0.0006,
+ "loss": 4.28231143951416,
+ "step": 1963
+ },
+ {
+ "epoch": 27.27947598253275,
+ "grad_norm": 0.01943155936896801,
+ "learning_rate": 0.0006,
+ "loss": 4.314976692199707,
+ "step": 1964
+ },
+ {
+ "epoch": 27.29344978165939,
+ "grad_norm": 0.019064994528889656,
+ "learning_rate": 0.0006,
+ "loss": 4.25328254699707,
+ "step": 1965
+ },
+ {
+ "epoch": 27.307423580786025,
+ "grad_norm": 0.017423273995518684,
+ "learning_rate": 0.0006,
+ "loss": 4.195992469787598,
+ "step": 1966
+ },
+ {
+ "epoch": 27.321397379912664,
+ "grad_norm": 0.01677924580872059,
+ "learning_rate": 0.0006,
+ "loss": 4.260043144226074,
+ "step": 1967
+ },
+ {
+ "epoch": 27.335371179039303,
+ "grad_norm": 0.01720673404633999,
+ "learning_rate": 0.0006,
+ "loss": 4.222780704498291,
+ "step": 1968
+ },
+ {
+ "epoch": 27.349344978165938,
+ "grad_norm": 0.017300743609666824,
+ "learning_rate": 0.0006,
+ "loss": 4.109524250030518,
+ "step": 1969
+ },
+ {
+ "epoch": 27.363318777292577,
+ "grad_norm": 0.016065871343016624,
+ "learning_rate": 0.0006,
+ "loss": 4.36108922958374,
+ "step": 1970
+ },
+ {
+ "epoch": 27.377292576419215,
+ "grad_norm": 0.016626616939902306,
+ "learning_rate": 0.0006,
+ "loss": 4.220630645751953,
+ "step": 1971
+ },
+ {
+ "epoch": 27.39126637554585,
+ "grad_norm": 0.016482602804899216,
+ "learning_rate": 0.0006,
+ "loss": 4.344797134399414,
+ "step": 1972
+ },
+ {
+ "epoch": 27.40524017467249,
+ "grad_norm": 0.01727953553199768,
+ "learning_rate": 0.0006,
+ "loss": 4.296645164489746,
+ "step": 1973
+ },
+ {
+ "epoch": 27.419213973799128,
+ "grad_norm": 0.01522731315344572,
+ "learning_rate": 0.0006,
+ "loss": 4.223751068115234,
+ "step": 1974
+ },
+ {
+ "epoch": 27.433187772925763,
+ "grad_norm": 0.01581776700913906,
+ "learning_rate": 0.0006,
+ "loss": 4.351379871368408,
+ "step": 1975
+ },
+ {
+ "epoch": 27.4471615720524,
+ "grad_norm": 0.016394605860114098,
+ "learning_rate": 0.0006,
+ "loss": 4.2412919998168945,
+ "step": 1976
+ },
+ {
+ "epoch": 27.46113537117904,
+ "grad_norm": 0.01751169003546238,
+ "learning_rate": 0.0006,
+ "loss": 4.257606506347656,
+ "step": 1977
+ },
+ {
+ "epoch": 27.475109170305675,
+ "grad_norm": 0.017452310770750046,
+ "learning_rate": 0.0006,
+ "loss": 4.380134582519531,
+ "step": 1978
+ },
+ {
+ "epoch": 27.489082969432314,
+ "grad_norm": 0.017741898074746132,
+ "learning_rate": 0.0006,
+ "loss": 4.279632568359375,
+ "step": 1979
+ },
+ {
+ "epoch": 27.503056768558952,
+ "grad_norm": 0.017680590972304344,
+ "learning_rate": 0.0006,
+ "loss": 4.273406028747559,
+ "step": 1980
+ },
+ {
+ "epoch": 27.51703056768559,
+ "grad_norm": 0.01864718459546566,
+ "learning_rate": 0.0006,
+ "loss": 4.278583526611328,
+ "step": 1981
+ },
+ {
+ "epoch": 27.531004366812226,
+ "grad_norm": 0.019558541476726532,
+ "learning_rate": 0.0006,
+ "loss": 4.211919784545898,
+ "step": 1982
+ },
+ {
+ "epoch": 27.544978165938865,
+ "grad_norm": 0.019467400386929512,
+ "learning_rate": 0.0006,
+ "loss": 4.276822566986084,
+ "step": 1983
+ },
+ {
+ "epoch": 27.558951965065503,
+ "grad_norm": 0.01756172813475132,
+ "learning_rate": 0.0006,
+ "loss": 4.165379047393799,
+ "step": 1984
+ },
+ {
+ "epoch": 27.57292576419214,
+ "grad_norm": 0.017458263784646988,
+ "learning_rate": 0.0006,
+ "loss": 4.2992401123046875,
+ "step": 1985
+ },
+ {
+ "epoch": 27.586899563318777,
+ "grad_norm": 0.018089665099978447,
+ "learning_rate": 0.0006,
+ "loss": 4.243378639221191,
+ "step": 1986
+ },
+ {
+ "epoch": 27.600873362445416,
+ "grad_norm": 0.019535524770617485,
+ "learning_rate": 0.0006,
+ "loss": 4.321477890014648,
+ "step": 1987
+ },
+ {
+ "epoch": 27.61484716157205,
+ "grad_norm": 0.01997970975935459,
+ "learning_rate": 0.0006,
+ "loss": 4.123082637786865,
+ "step": 1988
+ },
+ {
+ "epoch": 27.62882096069869,
+ "grad_norm": 0.01913067139685154,
+ "learning_rate": 0.0006,
+ "loss": 4.289896488189697,
+ "step": 1989
+ },
+ {
+ "epoch": 27.64279475982533,
+ "grad_norm": 0.0169806070625782,
+ "learning_rate": 0.0006,
+ "loss": 4.247004985809326,
+ "step": 1990
+ },
+ {
+ "epoch": 27.656768558951963,
+ "grad_norm": 0.017315007746219635,
+ "learning_rate": 0.0006,
+ "loss": 4.2095441818237305,
+ "step": 1991
+ },
+ {
+ "epoch": 27.670742358078602,
+ "grad_norm": 0.015098759904503822,
+ "learning_rate": 0.0006,
+ "loss": 4.145560264587402,
+ "step": 1992
+ },
+ {
+ "epoch": 27.68471615720524,
+ "grad_norm": 0.016240514814853668,
+ "learning_rate": 0.0006,
+ "loss": 4.254927635192871,
+ "step": 1993
+ },
+ {
+ "epoch": 27.69868995633188,
+ "grad_norm": 0.01621108129620552,
+ "learning_rate": 0.0006,
+ "loss": 4.219725608825684,
+ "step": 1994
+ },
+ {
+ "epoch": 27.712663755458514,
+ "grad_norm": 0.015191009268164635,
+ "learning_rate": 0.0006,
+ "loss": 4.194345474243164,
+ "step": 1995
+ },
+ {
+ "epoch": 27.726637554585153,
+ "grad_norm": 0.014387983828783035,
+ "learning_rate": 0.0006,
+ "loss": 4.211467742919922,
+ "step": 1996
+ },
+ {
+ "epoch": 27.74061135371179,
+ "grad_norm": 0.015460561029613018,
+ "learning_rate": 0.0006,
+ "loss": 4.358700752258301,
+ "step": 1997
+ },
+ {
+ "epoch": 27.754585152838427,
+ "grad_norm": 0.015100076794624329,
+ "learning_rate": 0.0006,
+ "loss": 4.171223163604736,
+ "step": 1998
+ },
+ {
+ "epoch": 27.768558951965066,
+ "grad_norm": 0.015607891604304314,
+ "learning_rate": 0.0006,
+ "loss": 4.217874526977539,
+ "step": 1999
+ },
+ {
+ "epoch": 27.782532751091704,
+ "grad_norm": 0.015650689601898193,
+ "learning_rate": 0.0006,
+ "loss": 4.280780792236328,
+ "step": 2000
+ },
+ {
+ "epoch": 27.79650655021834,
+ "grad_norm": 0.01580994389951229,
+ "learning_rate": 0.0006,
+ "loss": 4.159212112426758,
+ "step": 2001
+ },
+ {
+ "epoch": 27.810480349344978,
+ "grad_norm": 0.016607413068413734,
+ "learning_rate": 0.0006,
+ "loss": 4.238853454589844,
+ "step": 2002
+ },
+ {
+ "epoch": 27.824454148471617,
+ "grad_norm": 0.017383860424160957,
+ "learning_rate": 0.0006,
+ "loss": 4.233944892883301,
+ "step": 2003
+ },
+ {
+ "epoch": 27.83842794759825,
+ "grad_norm": 0.014966659247875214,
+ "learning_rate": 0.0006,
+ "loss": 4.167392730712891,
+ "step": 2004
+ },
+ {
+ "epoch": 27.85240174672489,
+ "grad_norm": 0.014589878730475903,
+ "learning_rate": 0.0006,
+ "loss": 4.2390899658203125,
+ "step": 2005
+ },
+ {
+ "epoch": 27.86637554585153,
+ "grad_norm": 0.017179450020194054,
+ "learning_rate": 0.0006,
+ "loss": 4.365785598754883,
+ "step": 2006
+ },
+ {
+ "epoch": 27.880349344978168,
+ "grad_norm": 0.01893901452422142,
+ "learning_rate": 0.0006,
+ "loss": 4.289430141448975,
+ "step": 2007
+ },
+ {
+ "epoch": 27.894323144104803,
+ "grad_norm": 0.018619712442159653,
+ "learning_rate": 0.0006,
+ "loss": 4.310665130615234,
+ "step": 2008
+ },
+ {
+ "epoch": 27.90829694323144,
+ "grad_norm": 0.019180385395884514,
+ "learning_rate": 0.0006,
+ "loss": 4.27454137802124,
+ "step": 2009
+ },
+ {
+ "epoch": 27.92227074235808,
+ "grad_norm": 0.017811523750424385,
+ "learning_rate": 0.0006,
+ "loss": 4.312562465667725,
+ "step": 2010
+ },
+ {
+ "epoch": 27.936244541484715,
+ "grad_norm": 0.016583573073148727,
+ "learning_rate": 0.0006,
+ "loss": 4.202253341674805,
+ "step": 2011
+ },
+ {
+ "epoch": 27.950218340611354,
+ "grad_norm": 0.01658615842461586,
+ "learning_rate": 0.0006,
+ "loss": 4.183433532714844,
+ "step": 2012
+ },
+ {
+ "epoch": 27.964192139737992,
+ "grad_norm": 0.015327401459217072,
+ "learning_rate": 0.0006,
+ "loss": 4.2125139236450195,
+ "step": 2013
+ },
+ {
+ "epoch": 27.978165938864628,
+ "grad_norm": 0.015004601329565048,
+ "learning_rate": 0.0006,
+ "loss": 4.257392883300781,
+ "step": 2014
+ },
+ {
+ "epoch": 27.992139737991266,
+ "grad_norm": 0.015259161591529846,
+ "learning_rate": 0.0006,
+ "loss": 4.241818428039551,
+ "step": 2015
+ },
+ {
+ "epoch": 28.0,
+ "grad_norm": 0.016808776184916496,
+ "learning_rate": 0.0006,
+ "loss": 4.146241188049316,
+ "step": 2016
+ },
+ {
+ "epoch": 28.0,
+ "eval_loss": 4.663090705871582,
+ "eval_runtime": 56.4559,
+ "eval_samples_per_second": 43.255,
+ "eval_steps_per_second": 1.364,
+ "step": 2016
+ },
+ {
+ "epoch": 28.01397379912664,
+ "grad_norm": 0.015784382820129395,
+ "learning_rate": 0.0006,
+ "loss": 4.044191360473633,
+ "step": 2017
+ },
+ {
+ "epoch": 28.027947598253274,
+ "grad_norm": 0.015636710450053215,
+ "learning_rate": 0.0006,
+ "loss": 4.128754615783691,
+ "step": 2018
+ },
+ {
+ "epoch": 28.041921397379912,
+ "grad_norm": 0.015364975668489933,
+ "learning_rate": 0.0006,
+ "loss": 4.191835403442383,
+ "step": 2019
+ },
+ {
+ "epoch": 28.05589519650655,
+ "grad_norm": 0.015015700832009315,
+ "learning_rate": 0.0006,
+ "loss": 4.120022773742676,
+ "step": 2020
+ },
+ {
+ "epoch": 28.069868995633186,
+ "grad_norm": 0.016666799783706665,
+ "learning_rate": 0.0006,
+ "loss": 4.33906364440918,
+ "step": 2021
+ },
+ {
+ "epoch": 28.083842794759825,
+ "grad_norm": 0.01780104823410511,
+ "learning_rate": 0.0006,
+ "loss": 4.279458999633789,
+ "step": 2022
+ },
+ {
+ "epoch": 28.097816593886463,
+ "grad_norm": 0.01870882883667946,
+ "learning_rate": 0.0006,
+ "loss": 4.34728479385376,
+ "step": 2023
+ },
+ {
+ "epoch": 28.111790393013102,
+ "grad_norm": 0.018406381830573082,
+ "learning_rate": 0.0006,
+ "loss": 4.211311340332031,
+ "step": 2024
+ },
+ {
+ "epoch": 28.125764192139737,
+ "grad_norm": 0.017833448946475983,
+ "learning_rate": 0.0006,
+ "loss": 4.223254203796387,
+ "step": 2025
+ },
+ {
+ "epoch": 28.139737991266376,
+ "grad_norm": 0.017552688717842102,
+ "learning_rate": 0.0006,
+ "loss": 4.208823204040527,
+ "step": 2026
+ },
+ {
+ "epoch": 28.153711790393015,
+ "grad_norm": 0.018500229343771935,
+ "learning_rate": 0.0006,
+ "loss": 4.198025226593018,
+ "step": 2027
+ },
+ {
+ "epoch": 28.16768558951965,
+ "grad_norm": 0.01782156340777874,
+ "learning_rate": 0.0006,
+ "loss": 4.179978370666504,
+ "step": 2028
+ },
+ {
+ "epoch": 28.18165938864629,
+ "grad_norm": 0.019580967724323273,
+ "learning_rate": 0.0006,
+ "loss": 4.291114807128906,
+ "step": 2029
+ },
+ {
+ "epoch": 28.195633187772927,
+ "grad_norm": 0.020804036408662796,
+ "learning_rate": 0.0006,
+ "loss": 4.220607280731201,
+ "step": 2030
+ },
+ {
+ "epoch": 28.209606986899562,
+ "grad_norm": 0.019595693796873093,
+ "learning_rate": 0.0006,
+ "loss": 4.2905731201171875,
+ "step": 2031
+ },
+ {
+ "epoch": 28.2235807860262,
+ "grad_norm": 0.019247086718678474,
+ "learning_rate": 0.0006,
+ "loss": 4.2467360496521,
+ "step": 2032
+ },
+ {
+ "epoch": 28.23755458515284,
+ "grad_norm": 0.01958036608994007,
+ "learning_rate": 0.0006,
+ "loss": 4.286137580871582,
+ "step": 2033
+ },
+ {
+ "epoch": 28.251528384279474,
+ "grad_norm": 0.021354753524065018,
+ "learning_rate": 0.0006,
+ "loss": 4.234755039215088,
+ "step": 2034
+ },
+ {
+ "epoch": 28.265502183406113,
+ "grad_norm": 0.02206375077366829,
+ "learning_rate": 0.0006,
+ "loss": 4.149641036987305,
+ "step": 2035
+ },
+ {
+ "epoch": 28.27947598253275,
+ "grad_norm": 0.02201310358941555,
+ "learning_rate": 0.0006,
+ "loss": 4.1928629875183105,
+ "step": 2036
+ },
+ {
+ "epoch": 28.29344978165939,
+ "grad_norm": 0.01961950585246086,
+ "learning_rate": 0.0006,
+ "loss": 4.211404323577881,
+ "step": 2037
+ },
+ {
+ "epoch": 28.307423580786025,
+ "grad_norm": 0.019211795181035995,
+ "learning_rate": 0.0006,
+ "loss": 4.200976848602295,
+ "step": 2038
+ },
+ {
+ "epoch": 28.321397379912664,
+ "grad_norm": 0.01779370754957199,
+ "learning_rate": 0.0006,
+ "loss": 4.249148845672607,
+ "step": 2039
+ },
+ {
+ "epoch": 28.335371179039303,
+ "grad_norm": 0.018635908141732216,
+ "learning_rate": 0.0006,
+ "loss": 4.11806058883667,
+ "step": 2040
+ },
+ {
+ "epoch": 28.349344978165938,
+ "grad_norm": 0.019792694598436356,
+ "learning_rate": 0.0006,
+ "loss": 4.335224151611328,
+ "step": 2041
+ },
+ {
+ "epoch": 28.363318777292577,
+ "grad_norm": 0.021731717512011528,
+ "learning_rate": 0.0006,
+ "loss": 4.302778244018555,
+ "step": 2042
+ },
+ {
+ "epoch": 28.377292576419215,
+ "grad_norm": 0.020823447033762932,
+ "learning_rate": 0.0006,
+ "loss": 4.260610580444336,
+ "step": 2043
+ },
+ {
+ "epoch": 28.39126637554585,
+ "grad_norm": 0.019385412335395813,
+ "learning_rate": 0.0006,
+ "loss": 4.2247724533081055,
+ "step": 2044
+ },
+ {
+ "epoch": 28.40524017467249,
+ "grad_norm": 0.019539108499884605,
+ "learning_rate": 0.0006,
+ "loss": 4.236655235290527,
+ "step": 2045
+ },
+ {
+ "epoch": 28.419213973799128,
+ "grad_norm": 0.01942690648138523,
+ "learning_rate": 0.0006,
+ "loss": 4.251850605010986,
+ "step": 2046
+ },
+ {
+ "epoch": 28.433187772925763,
+ "grad_norm": 0.019667886197566986,
+ "learning_rate": 0.0006,
+ "loss": 4.222312927246094,
+ "step": 2047
+ },
+ {
+ "epoch": 28.4471615720524,
+ "grad_norm": 0.02029012329876423,
+ "learning_rate": 0.0006,
+ "loss": 4.246252059936523,
+ "step": 2048
+ },
+ {
+ "epoch": 28.46113537117904,
+ "grad_norm": 0.01784469373524189,
+ "learning_rate": 0.0006,
+ "loss": 4.192128658294678,
+ "step": 2049
+ },
+ {
+ "epoch": 28.475109170305675,
+ "grad_norm": 0.0160287544131279,
+ "learning_rate": 0.0006,
+ "loss": 4.208607196807861,
+ "step": 2050
+ },
+ {
+ "epoch": 28.489082969432314,
+ "grad_norm": 0.015348346903920174,
+ "learning_rate": 0.0006,
+ "loss": 4.236598014831543,
+ "step": 2051
+ },
+ {
+ "epoch": 28.503056768558952,
+ "grad_norm": 0.015198206529021263,
+ "learning_rate": 0.0006,
+ "loss": 4.235552787780762,
+ "step": 2052
+ },
+ {
+ "epoch": 28.51703056768559,
+ "grad_norm": 0.01523754745721817,
+ "learning_rate": 0.0006,
+ "loss": 4.1941728591918945,
+ "step": 2053
+ },
+ {
+ "epoch": 28.531004366812226,
+ "grad_norm": 0.015559614636003971,
+ "learning_rate": 0.0006,
+ "loss": 4.294757843017578,
+ "step": 2054
+ },
+ {
+ "epoch": 28.544978165938865,
+ "grad_norm": 0.01568358950316906,
+ "learning_rate": 0.0006,
+ "loss": 4.213065147399902,
+ "step": 2055
+ },
+ {
+ "epoch": 28.558951965065503,
+ "grad_norm": 0.015336516313254833,
+ "learning_rate": 0.0006,
+ "loss": 4.240293979644775,
+ "step": 2056
+ },
+ {
+ "epoch": 28.57292576419214,
+ "grad_norm": 0.015041496604681015,
+ "learning_rate": 0.0006,
+ "loss": 4.273397445678711,
+ "step": 2057
+ },
+ {
+ "epoch": 28.586899563318777,
+ "grad_norm": 0.01610128954052925,
+ "learning_rate": 0.0006,
+ "loss": 4.125369071960449,
+ "step": 2058
+ },
+ {
+ "epoch": 28.600873362445416,
+ "grad_norm": 0.015974976122379303,
+ "learning_rate": 0.0006,
+ "loss": 4.133459091186523,
+ "step": 2059
+ },
+ {
+ "epoch": 28.61484716157205,
+ "grad_norm": 0.015936799347400665,
+ "learning_rate": 0.0006,
+ "loss": 4.306667327880859,
+ "step": 2060
+ },
+ {
+ "epoch": 28.62882096069869,
+ "grad_norm": 0.01795247197151184,
+ "learning_rate": 0.0006,
+ "loss": 4.205156326293945,
+ "step": 2061
+ },
+ {
+ "epoch": 28.64279475982533,
+ "grad_norm": 0.019639814272522926,
+ "learning_rate": 0.0006,
+ "loss": 4.277614593505859,
+ "step": 2062
+ },
+ {
+ "epoch": 28.656768558951963,
+ "grad_norm": 0.01722300611436367,
+ "learning_rate": 0.0006,
+ "loss": 4.191032886505127,
+ "step": 2063
+ },
+ {
+ "epoch": 28.670742358078602,
+ "grad_norm": 0.01612633652985096,
+ "learning_rate": 0.0006,
+ "loss": 4.20717191696167,
+ "step": 2064
+ },
+ {
+ "epoch": 28.68471615720524,
+ "grad_norm": 0.01603279449045658,
+ "learning_rate": 0.0006,
+ "loss": 4.228884220123291,
+ "step": 2065
+ },
+ {
+ "epoch": 28.69868995633188,
+ "grad_norm": 0.016405927017331123,
+ "learning_rate": 0.0006,
+ "loss": 4.201292991638184,
+ "step": 2066
+ },
+ {
+ "epoch": 28.712663755458514,
+ "grad_norm": 0.016644183546304703,
+ "learning_rate": 0.0006,
+ "loss": 4.239282131195068,
+ "step": 2067
+ },
+ {
+ "epoch": 28.726637554585153,
+ "grad_norm": 0.0175587497651577,
+ "learning_rate": 0.0006,
+ "loss": 4.289038181304932,
+ "step": 2068
+ },
+ {
+ "epoch": 28.74061135371179,
+ "grad_norm": 0.018223397433757782,
+ "learning_rate": 0.0006,
+ "loss": 4.2266998291015625,
+ "step": 2069
+ },
+ {
+ "epoch": 28.754585152838427,
+ "grad_norm": 0.018680868670344353,
+ "learning_rate": 0.0006,
+ "loss": 4.348849296569824,
+ "step": 2070
+ },
+ {
+ "epoch": 28.768558951965066,
+ "grad_norm": 0.016726728528738022,
+ "learning_rate": 0.0006,
+ "loss": 4.292080879211426,
+ "step": 2071
+ },
+ {
+ "epoch": 28.782532751091704,
+ "grad_norm": 0.017386795952916145,
+ "learning_rate": 0.0006,
+ "loss": 4.208870887756348,
+ "step": 2072
+ },
+ {
+ "epoch": 28.79650655021834,
+ "grad_norm": 0.01865178905427456,
+ "learning_rate": 0.0006,
+ "loss": 4.054150581359863,
+ "step": 2073
+ },
+ {
+ "epoch": 28.810480349344978,
+ "grad_norm": 0.01776815392076969,
+ "learning_rate": 0.0006,
+ "loss": 4.202404499053955,
+ "step": 2074
+ },
+ {
+ "epoch": 28.824454148471617,
+ "grad_norm": 0.01643497310578823,
+ "learning_rate": 0.0006,
+ "loss": 4.131855010986328,
+ "step": 2075
+ },
+ {
+ "epoch": 28.83842794759825,
+ "grad_norm": 0.01574273779988289,
+ "learning_rate": 0.0006,
+ "loss": 4.296788692474365,
+ "step": 2076
+ },
+ {
+ "epoch": 28.85240174672489,
+ "grad_norm": 0.015613908879458904,
+ "learning_rate": 0.0006,
+ "loss": 4.210949897766113,
+ "step": 2077
+ },
+ {
+ "epoch": 28.86637554585153,
+ "grad_norm": 0.015543212182819843,
+ "learning_rate": 0.0006,
+ "loss": 4.218136310577393,
+ "step": 2078
+ },
+ {
+ "epoch": 28.880349344978168,
+ "grad_norm": 0.014681251719594002,
+ "learning_rate": 0.0006,
+ "loss": 4.19401741027832,
+ "step": 2079
+ },
+ {
+ "epoch": 28.894323144104803,
+ "grad_norm": 0.014393828809261322,
+ "learning_rate": 0.0006,
+ "loss": 4.198468208312988,
+ "step": 2080
+ },
+ {
+ "epoch": 28.90829694323144,
+ "grad_norm": 0.01580170728266239,
+ "learning_rate": 0.0006,
+ "loss": 4.224969863891602,
+ "step": 2081
+ },
+ {
+ "epoch": 28.92227074235808,
+ "grad_norm": 0.01484165620058775,
+ "learning_rate": 0.0006,
+ "loss": 4.12236213684082,
+ "step": 2082
+ },
+ {
+ "epoch": 28.936244541484715,
+ "grad_norm": 0.01497623696923256,
+ "learning_rate": 0.0006,
+ "loss": 4.24141788482666,
+ "step": 2083
+ },
+ {
+ "epoch": 28.950218340611354,
+ "grad_norm": 0.015080046840012074,
+ "learning_rate": 0.0006,
+ "loss": 4.284151077270508,
+ "step": 2084
+ },
+ {
+ "epoch": 28.964192139737992,
+ "grad_norm": 0.016207212582230568,
+ "learning_rate": 0.0006,
+ "loss": 4.298693656921387,
+ "step": 2085
+ },
+ {
+ "epoch": 28.978165938864628,
+ "grad_norm": 0.015474417246878147,
+ "learning_rate": 0.0006,
+ "loss": 4.2412261962890625,
+ "step": 2086
+ },
+ {
+ "epoch": 28.992139737991266,
+ "grad_norm": 0.01515351701527834,
+ "learning_rate": 0.0006,
+ "loss": 4.3054423332214355,
+ "step": 2087
+ },
+ {
+ "epoch": 29.0,
+ "grad_norm": 0.016411345452070236,
+ "learning_rate": 0.0006,
+ "loss": 4.104830741882324,
+ "step": 2088
+ },
+ {
+ "epoch": 29.0,
+ "eval_loss": 4.661567687988281,
+ "eval_runtime": 57.1244,
+ "eval_samples_per_second": 42.749,
+ "eval_steps_per_second": 1.348,
+ "step": 2088
+ },
+ {
+ "epoch": 29.01397379912664,
+ "grad_norm": 0.01665342226624489,
+ "learning_rate": 0.0006,
+ "loss": 4.311391830444336,
+ "step": 2089
+ },
+ {
+ "epoch": 29.027947598253274,
+ "grad_norm": 0.018053214997053146,
+ "learning_rate": 0.0006,
+ "loss": 4.135597229003906,
+ "step": 2090
+ },
+ {
+ "epoch": 29.041921397379912,
+ "grad_norm": 0.018213748931884766,
+ "learning_rate": 0.0006,
+ "loss": 4.303299903869629,
+ "step": 2091
+ },
+ {
+ "epoch": 29.05589519650655,
+ "grad_norm": 0.018460242077708244,
+ "learning_rate": 0.0006,
+ "loss": 4.20850944519043,
+ "step": 2092
+ },
+ {
+ "epoch": 29.069868995633186,
+ "grad_norm": 0.018365809693932533,
+ "learning_rate": 0.0006,
+ "loss": 4.275847434997559,
+ "step": 2093
+ },
+ {
+ "epoch": 29.083842794759825,
+ "grad_norm": 0.018346186727285385,
+ "learning_rate": 0.0006,
+ "loss": 4.202091217041016,
+ "step": 2094
+ },
+ {
+ "epoch": 29.097816593886463,
+ "grad_norm": 0.017920587211847305,
+ "learning_rate": 0.0006,
+ "loss": 4.182092666625977,
+ "step": 2095
+ },
+ {
+ "epoch": 29.111790393013102,
+ "grad_norm": 0.01812003180384636,
+ "learning_rate": 0.0006,
+ "loss": 4.259721755981445,
+ "step": 2096
+ },
+ {
+ "epoch": 29.125764192139737,
+ "grad_norm": 0.01700986735522747,
+ "learning_rate": 0.0006,
+ "loss": 4.067573070526123,
+ "step": 2097
+ },
+ {
+ "epoch": 29.139737991266376,
+ "grad_norm": 0.018069760873913765,
+ "learning_rate": 0.0006,
+ "loss": 4.234400749206543,
+ "step": 2098
+ },
+ {
+ "epoch": 29.153711790393015,
+ "grad_norm": 0.01905428245663643,
+ "learning_rate": 0.0006,
+ "loss": 4.16609001159668,
+ "step": 2099
+ },
+ {
+ "epoch": 29.16768558951965,
+ "grad_norm": 0.018711242824792862,
+ "learning_rate": 0.0006,
+ "loss": 4.249088764190674,
+ "step": 2100
+ },
+ {
+ "epoch": 29.18165938864629,
+ "grad_norm": 0.018721306696534157,
+ "learning_rate": 0.0006,
+ "loss": 4.14540958404541,
+ "step": 2101
+ },
+ {
+ "epoch": 29.195633187772927,
+ "grad_norm": 0.020139874890446663,
+ "learning_rate": 0.0006,
+ "loss": 4.049248695373535,
+ "step": 2102
+ },
+ {
+ "epoch": 29.209606986899562,
+ "grad_norm": 0.020849574357271194,
+ "learning_rate": 0.0006,
+ "loss": 4.235224723815918,
+ "step": 2103
+ },
+ {
+ "epoch": 29.2235807860262,
+ "grad_norm": 0.018693549558520317,
+ "learning_rate": 0.0006,
+ "loss": 4.232961177825928,
+ "step": 2104
+ },
+ {
+ "epoch": 29.23755458515284,
+ "grad_norm": 0.017889857292175293,
+ "learning_rate": 0.0006,
+ "loss": 4.171422004699707,
+ "step": 2105
+ },
+ {
+ "epoch": 29.251528384279474,
+ "grad_norm": 0.016690224409103394,
+ "learning_rate": 0.0006,
+ "loss": 4.244717597961426,
+ "step": 2106
+ },
+ {
+ "epoch": 29.265502183406113,
+ "grad_norm": 0.01708616502583027,
+ "learning_rate": 0.0006,
+ "loss": 4.125068664550781,
+ "step": 2107
+ },
+ {
+ "epoch": 29.27947598253275,
+ "grad_norm": 0.017928237095475197,
+ "learning_rate": 0.0006,
+ "loss": 4.208292007446289,
+ "step": 2108
+ },
+ {
+ "epoch": 29.29344978165939,
+ "grad_norm": 0.018250394612550735,
+ "learning_rate": 0.0006,
+ "loss": 4.118630886077881,
+ "step": 2109
+ },
+ {
+ "epoch": 29.307423580786025,
+ "grad_norm": 0.020160246640443802,
+ "learning_rate": 0.0006,
+ "loss": 4.206025123596191,
+ "step": 2110
+ },
+ {
+ "epoch": 29.321397379912664,
+ "grad_norm": 0.020045453682541847,
+ "learning_rate": 0.0006,
+ "loss": 4.2758684158325195,
+ "step": 2111
+ },
+ {
+ "epoch": 29.335371179039303,
+ "grad_norm": 0.019556907936930656,
+ "learning_rate": 0.0006,
+ "loss": 4.188272953033447,
+ "step": 2112
+ },
+ {
+ "epoch": 29.349344978165938,
+ "grad_norm": 0.019185200333595276,
+ "learning_rate": 0.0006,
+ "loss": 4.17451286315918,
+ "step": 2113
+ },
+ {
+ "epoch": 29.363318777292577,
+ "grad_norm": 0.01816629432141781,
+ "learning_rate": 0.0006,
+ "loss": 4.165246963500977,
+ "step": 2114
+ },
+ {
+ "epoch": 29.377292576419215,
+ "grad_norm": 0.01865146867930889,
+ "learning_rate": 0.0006,
+ "loss": 4.148184776306152,
+ "step": 2115
+ },
+ {
+ "epoch": 29.39126637554585,
+ "grad_norm": 0.018338464200496674,
+ "learning_rate": 0.0006,
+ "loss": 4.288936614990234,
+ "step": 2116
+ },
+ {
+ "epoch": 29.40524017467249,
+ "grad_norm": 0.018986187875270844,
+ "learning_rate": 0.0006,
+ "loss": 4.183379650115967,
+ "step": 2117
+ },
+ {
+ "epoch": 29.419213973799128,
+ "grad_norm": 0.018409637734293938,
+ "learning_rate": 0.0006,
+ "loss": 4.057574272155762,
+ "step": 2118
+ },
+ {
+ "epoch": 29.433187772925763,
+ "grad_norm": 0.016664542257785797,
+ "learning_rate": 0.0006,
+ "loss": 4.18517541885376,
+ "step": 2119
+ },
+ {
+ "epoch": 29.4471615720524,
+ "grad_norm": 0.016486089676618576,
+ "learning_rate": 0.0006,
+ "loss": 4.179367542266846,
+ "step": 2120
+ },
+ {
+ "epoch": 29.46113537117904,
+ "grad_norm": 0.01787523552775383,
+ "learning_rate": 0.0006,
+ "loss": 4.184126377105713,
+ "step": 2121
+ },
+ {
+ "epoch": 29.475109170305675,
+ "grad_norm": 0.016574613749980927,
+ "learning_rate": 0.0006,
+ "loss": 4.290170192718506,
+ "step": 2122
+ },
+ {
+ "epoch": 29.489082969432314,
+ "grad_norm": 0.01643376611173153,
+ "learning_rate": 0.0006,
+ "loss": 4.052913665771484,
+ "step": 2123
+ },
+ {
+ "epoch": 29.503056768558952,
+ "grad_norm": 0.016503969207406044,
+ "learning_rate": 0.0006,
+ "loss": 4.335184097290039,
+ "step": 2124
+ },
+ {
+ "epoch": 29.51703056768559,
+ "grad_norm": 0.017596479505300522,
+ "learning_rate": 0.0006,
+ "loss": 4.165539741516113,
+ "step": 2125
+ },
+ {
+ "epoch": 29.531004366812226,
+ "grad_norm": 0.017468582838773727,
+ "learning_rate": 0.0006,
+ "loss": 4.315964698791504,
+ "step": 2126
+ },
+ {
+ "epoch": 29.544978165938865,
+ "grad_norm": 0.01724618673324585,
+ "learning_rate": 0.0006,
+ "loss": 4.243681907653809,
+ "step": 2127
+ },
+ {
+ "epoch": 29.558951965065503,
+ "grad_norm": 0.018384616822004318,
+ "learning_rate": 0.0006,
+ "loss": 4.112447261810303,
+ "step": 2128
+ },
+ {
+ "epoch": 29.57292576419214,
+ "grad_norm": 0.01824074238538742,
+ "learning_rate": 0.0006,
+ "loss": 4.236065864562988,
+ "step": 2129
+ },
+ {
+ "epoch": 29.586899563318777,
+ "grad_norm": 0.01625971868634224,
+ "learning_rate": 0.0006,
+ "loss": 4.18143892288208,
+ "step": 2130
+ },
+ {
+ "epoch": 29.600873362445416,
+ "grad_norm": 0.015537494793534279,
+ "learning_rate": 0.0006,
+ "loss": 4.225147247314453,
+ "step": 2131
+ },
+ {
+ "epoch": 29.61484716157205,
+ "grad_norm": 0.015818974003195763,
+ "learning_rate": 0.0006,
+ "loss": 4.3100175857543945,
+ "step": 2132
+ },
+ {
+ "epoch": 29.62882096069869,
+ "grad_norm": 0.015909671783447266,
+ "learning_rate": 0.0006,
+ "loss": 4.169775009155273,
+ "step": 2133
+ },
+ {
+ "epoch": 29.64279475982533,
+ "grad_norm": 0.01769658550620079,
+ "learning_rate": 0.0006,
+ "loss": 4.278904914855957,
+ "step": 2134
+ },
+ {
+ "epoch": 29.656768558951963,
+ "grad_norm": 0.015641704201698303,
+ "learning_rate": 0.0006,
+ "loss": 4.2426228523254395,
+ "step": 2135
+ },
+ {
+ "epoch": 29.670742358078602,
+ "grad_norm": 0.01565658301115036,
+ "learning_rate": 0.0006,
+ "loss": 4.232687950134277,
+ "step": 2136
+ },
+ {
+ "epoch": 29.68471615720524,
+ "grad_norm": 0.016485940665006638,
+ "learning_rate": 0.0006,
+ "loss": 4.209839820861816,
+ "step": 2137
+ },
+ {
+ "epoch": 29.69868995633188,
+ "grad_norm": 0.017622729763388634,
+ "learning_rate": 0.0006,
+ "loss": 4.196017265319824,
+ "step": 2138
+ },
+ {
+ "epoch": 29.712663755458514,
+ "grad_norm": 0.018320564180612564,
+ "learning_rate": 0.0006,
+ "loss": 4.225683689117432,
+ "step": 2139
+ },
+ {
+ "epoch": 29.726637554585153,
+ "grad_norm": 0.018140794709324837,
+ "learning_rate": 0.0006,
+ "loss": 4.311944961547852,
+ "step": 2140
+ },
+ {
+ "epoch": 29.74061135371179,
+ "grad_norm": 0.016227100044488907,
+ "learning_rate": 0.0006,
+ "loss": 4.230257511138916,
+ "step": 2141
+ },
+ {
+ "epoch": 29.754585152838427,
+ "grad_norm": 0.015759488567709923,
+ "learning_rate": 0.0006,
+ "loss": 4.119174003601074,
+ "step": 2142
+ },
+ {
+ "epoch": 29.768558951965066,
+ "grad_norm": 0.01757766678929329,
+ "learning_rate": 0.0006,
+ "loss": 4.165826797485352,
+ "step": 2143
+ },
+ {
+ "epoch": 29.782532751091704,
+ "grad_norm": 0.019661923870444298,
+ "learning_rate": 0.0006,
+ "loss": 4.392203330993652,
+ "step": 2144
+ },
+ {
+ "epoch": 29.79650655021834,
+ "grad_norm": 0.019478755071759224,
+ "learning_rate": 0.0006,
+ "loss": 4.348371982574463,
+ "step": 2145
+ },
+ {
+ "epoch": 29.810480349344978,
+ "grad_norm": 0.019149569794535637,
+ "learning_rate": 0.0006,
+ "loss": 4.34104061126709,
+ "step": 2146
+ },
+ {
+ "epoch": 29.824454148471617,
+ "grad_norm": 0.017707521095871925,
+ "learning_rate": 0.0006,
+ "loss": 4.228226661682129,
+ "step": 2147
+ },
+ {
+ "epoch": 29.83842794759825,
+ "grad_norm": 0.015615378506481647,
+ "learning_rate": 0.0006,
+ "loss": 4.175541877746582,
+ "step": 2148
+ },
+ {
+ "epoch": 29.85240174672489,
+ "grad_norm": 0.016111129894852638,
+ "learning_rate": 0.0006,
+ "loss": 4.267936706542969,
+ "step": 2149
+ },
+ {
+ "epoch": 29.86637554585153,
+ "grad_norm": 0.01528779324144125,
+ "learning_rate": 0.0006,
+ "loss": 4.186014175415039,
+ "step": 2150
+ },
+ {
+ "epoch": 29.880349344978168,
+ "grad_norm": 0.014925646595656872,
+ "learning_rate": 0.0006,
+ "loss": 4.174860000610352,
+ "step": 2151
+ },
+ {
+ "epoch": 29.894323144104803,
+ "grad_norm": 0.016917673870921135,
+ "learning_rate": 0.0006,
+ "loss": 4.33908748626709,
+ "step": 2152
+ },
+ {
+ "epoch": 29.90829694323144,
+ "grad_norm": 0.016007075086236,
+ "learning_rate": 0.0006,
+ "loss": 4.266141414642334,
+ "step": 2153
+ },
+ {
+ "epoch": 29.92227074235808,
+ "grad_norm": 0.016178956255316734,
+ "learning_rate": 0.0006,
+ "loss": 4.271215915679932,
+ "step": 2154
+ },
+ {
+ "epoch": 29.936244541484715,
+ "grad_norm": 0.016180871054530144,
+ "learning_rate": 0.0006,
+ "loss": 4.278801441192627,
+ "step": 2155
+ },
+ {
+ "epoch": 29.950218340611354,
+ "grad_norm": 0.015335503034293652,
+ "learning_rate": 0.0006,
+ "loss": 4.194978713989258,
+ "step": 2156
+ },
+ {
+ "epoch": 29.964192139737992,
+ "grad_norm": 0.016580764204263687,
+ "learning_rate": 0.0006,
+ "loss": 4.290809631347656,
+ "step": 2157
+ },
+ {
+ "epoch": 29.978165938864628,
+ "grad_norm": 0.01623128168284893,
+ "learning_rate": 0.0006,
+ "loss": 4.205451488494873,
+ "step": 2158
+ },
+ {
+ "epoch": 29.992139737991266,
+ "grad_norm": 0.014904248528182507,
+ "learning_rate": 0.0006,
+ "loss": 4.24924373626709,
+ "step": 2159
+ },
+ {
+ "epoch": 30.0,
+ "grad_norm": 0.017089299857616425,
+ "learning_rate": 0.0006,
+ "loss": 4.222072601318359,
+ "step": 2160
+ },
+ {
+ "epoch": 30.0,
+ "eval_loss": 4.585411071777344,
+ "eval_runtime": 57.1164,
+ "eval_samples_per_second": 42.755,
+ "eval_steps_per_second": 1.348,
+ "step": 2160
+ },
+ {
+ "epoch": 30.01397379912664,
+ "grad_norm": 0.017226383090019226,
+ "learning_rate": 0.0006,
+ "loss": 4.19467830657959,
+ "step": 2161
+ },
+ {
+ "epoch": 30.027947598253274,
+ "grad_norm": 0.01635066606104374,
+ "learning_rate": 0.0006,
+ "loss": 4.176875591278076,
+ "step": 2162
+ },
+ {
+ "epoch": 30.041921397379912,
+ "grad_norm": 0.01735362410545349,
+ "learning_rate": 0.0006,
+ "loss": 4.13667106628418,
+ "step": 2163
+ },
+ {
+ "epoch": 30.05589519650655,
+ "grad_norm": 0.017047109082341194,
+ "learning_rate": 0.0006,
+ "loss": 4.136329650878906,
+ "step": 2164
+ },
+ {
+ "epoch": 30.069868995633186,
+ "grad_norm": 0.018116053193807602,
+ "learning_rate": 0.0006,
+ "loss": 4.228706359863281,
+ "step": 2165
+ },
+ {
+ "epoch": 30.083842794759825,
+ "grad_norm": 0.01791389286518097,
+ "learning_rate": 0.0006,
+ "loss": 4.1597747802734375,
+ "step": 2166
+ },
+ {
+ "epoch": 30.097816593886463,
+ "grad_norm": 0.017793502658605576,
+ "learning_rate": 0.0006,
+ "loss": 4.0807976722717285,
+ "step": 2167
+ },
+ {
+ "epoch": 30.111790393013102,
+ "grad_norm": 0.01827503927052021,
+ "learning_rate": 0.0006,
+ "loss": 4.170646667480469,
+ "step": 2168
+ },
+ {
+ "epoch": 30.125764192139737,
+ "grad_norm": 0.017228564247488976,
+ "learning_rate": 0.0006,
+ "loss": 4.146934509277344,
+ "step": 2169
+ },
+ {
+ "epoch": 30.139737991266376,
+ "grad_norm": 0.01757437363266945,
+ "learning_rate": 0.0006,
+ "loss": 4.1859130859375,
+ "step": 2170
+ },
+ {
+ "epoch": 30.153711790393015,
+ "grad_norm": 0.01701093092560768,
+ "learning_rate": 0.0006,
+ "loss": 4.126981258392334,
+ "step": 2171
+ },
+ {
+ "epoch": 30.16768558951965,
+ "grad_norm": 0.015220489352941513,
+ "learning_rate": 0.0006,
+ "loss": 4.164018630981445,
+ "step": 2172
+ },
+ {
+ "epoch": 30.18165938864629,
+ "grad_norm": 0.016254238784313202,
+ "learning_rate": 0.0006,
+ "loss": 4.258275508880615,
+ "step": 2173
+ },
+ {
+ "epoch": 30.195633187772927,
+ "grad_norm": 0.016609614714980125,
+ "learning_rate": 0.0006,
+ "loss": 4.222233772277832,
+ "step": 2174
+ },
+ {
+ "epoch": 30.209606986899562,
+ "grad_norm": 0.016531318426132202,
+ "learning_rate": 0.0006,
+ "loss": 4.144489288330078,
+ "step": 2175
+ },
+ {
+ "epoch": 30.2235807860262,
+ "grad_norm": 0.01567668654024601,
+ "learning_rate": 0.0006,
+ "loss": 4.181916236877441,
+ "step": 2176
+ },
+ {
+ "epoch": 30.23755458515284,
+ "grad_norm": 0.017195681110024452,
+ "learning_rate": 0.0006,
+ "loss": 4.189693927764893,
+ "step": 2177
+ },
+ {
+ "epoch": 30.251528384279474,
+ "grad_norm": 0.016882948577404022,
+ "learning_rate": 0.0006,
+ "loss": 4.21604061126709,
+ "step": 2178
+ },
+ {
+ "epoch": 30.265502183406113,
+ "grad_norm": 0.016261622309684753,
+ "learning_rate": 0.0006,
+ "loss": 4.227639198303223,
+ "step": 2179
+ },
+ {
+ "epoch": 30.27947598253275,
+ "grad_norm": 0.016411812976002693,
+ "learning_rate": 0.0006,
+ "loss": 4.130362510681152,
+ "step": 2180
+ },
+ {
+ "epoch": 30.29344978165939,
+ "grad_norm": 0.01732019893825054,
+ "learning_rate": 0.0006,
+ "loss": 4.230443000793457,
+ "step": 2181
+ },
+ {
+ "epoch": 30.307423580786025,
+ "grad_norm": 0.01825905591249466,
+ "learning_rate": 0.0006,
+ "loss": 4.175817489624023,
+ "step": 2182
+ },
+ {
+ "epoch": 30.321397379912664,
+ "grad_norm": 0.01887713558971882,
+ "learning_rate": 0.0006,
+ "loss": 4.169020652770996,
+ "step": 2183
+ },
+ {
+ "epoch": 30.335371179039303,
+ "grad_norm": 0.019391661509871483,
+ "learning_rate": 0.0006,
+ "loss": 4.176318645477295,
+ "step": 2184
+ },
+ {
+ "epoch": 30.349344978165938,
+ "grad_norm": 0.020840002223849297,
+ "learning_rate": 0.0006,
+ "loss": 4.174228668212891,
+ "step": 2185
+ },
+ {
+ "epoch": 30.363318777292577,
+ "grad_norm": 0.020193178206682205,
+ "learning_rate": 0.0006,
+ "loss": 4.127919673919678,
+ "step": 2186
+ },
+ {
+ "epoch": 30.377292576419215,
+ "grad_norm": 0.018449561670422554,
+ "learning_rate": 0.0006,
+ "loss": 4.229487895965576,
+ "step": 2187
+ },
+ {
+ "epoch": 30.39126637554585,
+ "grad_norm": 0.01711983233690262,
+ "learning_rate": 0.0006,
+ "loss": 4.201651573181152,
+ "step": 2188
+ },
+ {
+ "epoch": 30.40524017467249,
+ "grad_norm": 0.017293395474553108,
+ "learning_rate": 0.0006,
+ "loss": 4.168704032897949,
+ "step": 2189
+ },
+ {
+ "epoch": 30.419213973799128,
+ "grad_norm": 0.018526358529925346,
+ "learning_rate": 0.0006,
+ "loss": 4.171864032745361,
+ "step": 2190
+ },
+ {
+ "epoch": 30.433187772925763,
+ "grad_norm": 0.019287291914224625,
+ "learning_rate": 0.0006,
+ "loss": 4.039534568786621,
+ "step": 2191
+ },
+ {
+ "epoch": 30.4471615720524,
+ "grad_norm": 0.016902444884181023,
+ "learning_rate": 0.0006,
+ "loss": 4.282787799835205,
+ "step": 2192
+ },
+ {
+ "epoch": 30.46113537117904,
+ "grad_norm": 0.01656750962138176,
+ "learning_rate": 0.0006,
+ "loss": 4.213098526000977,
+ "step": 2193
+ },
+ {
+ "epoch": 30.475109170305675,
+ "grad_norm": 0.016805065795779228,
+ "learning_rate": 0.0006,
+ "loss": 4.1583476066589355,
+ "step": 2194
+ },
+ {
+ "epoch": 30.489082969432314,
+ "grad_norm": 0.017166638746857643,
+ "learning_rate": 0.0006,
+ "loss": 4.188634395599365,
+ "step": 2195
+ },
+ {
+ "epoch": 30.503056768558952,
+ "grad_norm": 0.01592106744647026,
+ "learning_rate": 0.0006,
+ "loss": 4.234043121337891,
+ "step": 2196
+ },
+ {
+ "epoch": 30.51703056768559,
+ "grad_norm": 0.015270989388227463,
+ "learning_rate": 0.0006,
+ "loss": 4.301420211791992,
+ "step": 2197
+ },
+ {
+ "epoch": 30.531004366812226,
+ "grad_norm": 0.015194724313914776,
+ "learning_rate": 0.0006,
+ "loss": 4.18703031539917,
+ "step": 2198
+ },
+ {
+ "epoch": 30.544978165938865,
+ "grad_norm": 0.015587746165692806,
+ "learning_rate": 0.0006,
+ "loss": 4.138928413391113,
+ "step": 2199
+ },
+ {
+ "epoch": 30.558951965065503,
+ "grad_norm": 0.016029570251703262,
+ "learning_rate": 0.0006,
+ "loss": 4.290863990783691,
+ "step": 2200
+ },
+ {
+ "epoch": 30.57292576419214,
+ "grad_norm": 0.015651309862732887,
+ "learning_rate": 0.0006,
+ "loss": 4.272062301635742,
+ "step": 2201
+ },
+ {
+ "epoch": 30.586899563318777,
+ "grad_norm": 0.015992306172847748,
+ "learning_rate": 0.0006,
+ "loss": 4.2487406730651855,
+ "step": 2202
+ },
+ {
+ "epoch": 30.600873362445416,
+ "grad_norm": 0.014969157986342907,
+ "learning_rate": 0.0006,
+ "loss": 4.115347862243652,
+ "step": 2203
+ },
+ {
+ "epoch": 30.61484716157205,
+ "grad_norm": 0.014639027416706085,
+ "learning_rate": 0.0006,
+ "loss": 4.2207489013671875,
+ "step": 2204
+ },
+ {
+ "epoch": 30.62882096069869,
+ "grad_norm": 0.015059413388371468,
+ "learning_rate": 0.0006,
+ "loss": 4.274044990539551,
+ "step": 2205
+ },
+ {
+ "epoch": 30.64279475982533,
+ "grad_norm": 0.0166641678661108,
+ "learning_rate": 0.0006,
+ "loss": 4.191373348236084,
+ "step": 2206
+ },
+ {
+ "epoch": 30.656768558951963,
+ "grad_norm": 0.017806345596909523,
+ "learning_rate": 0.0006,
+ "loss": 4.139800548553467,
+ "step": 2207
+ },
+ {
+ "epoch": 30.670742358078602,
+ "grad_norm": 0.018712421879172325,
+ "learning_rate": 0.0006,
+ "loss": 4.035965442657471,
+ "step": 2208
+ },
+ {
+ "epoch": 30.68471615720524,
+ "grad_norm": 0.018457243219017982,
+ "learning_rate": 0.0006,
+ "loss": 4.2579240798950195,
+ "step": 2209
+ },
+ {
+ "epoch": 30.69868995633188,
+ "grad_norm": 0.01920999586582184,
+ "learning_rate": 0.0006,
+ "loss": 4.131112098693848,
+ "step": 2210
+ },
+ {
+ "epoch": 30.712663755458514,
+ "grad_norm": 0.01729944534599781,
+ "learning_rate": 0.0006,
+ "loss": 4.254059314727783,
+ "step": 2211
+ },
+ {
+ "epoch": 30.726637554585153,
+ "grad_norm": 0.016914231702685356,
+ "learning_rate": 0.0006,
+ "loss": 4.233428955078125,
+ "step": 2212
+ },
+ {
+ "epoch": 30.74061135371179,
+ "grad_norm": 0.01913098618388176,
+ "learning_rate": 0.0006,
+ "loss": 4.23525333404541,
+ "step": 2213
+ },
+ {
+ "epoch": 30.754585152838427,
+ "grad_norm": 0.019150281324982643,
+ "learning_rate": 0.0006,
+ "loss": 4.213695049285889,
+ "step": 2214
+ },
+ {
+ "epoch": 30.768558951965066,
+ "grad_norm": 0.018947452306747437,
+ "learning_rate": 0.0006,
+ "loss": 4.151697158813477,
+ "step": 2215
+ },
+ {
+ "epoch": 30.782532751091704,
+ "grad_norm": 0.016058508306741714,
+ "learning_rate": 0.0006,
+ "loss": 4.220280170440674,
+ "step": 2216
+ },
+ {
+ "epoch": 30.79650655021834,
+ "grad_norm": 0.017488619312644005,
+ "learning_rate": 0.0006,
+ "loss": 4.154428482055664,
+ "step": 2217
+ },
+ {
+ "epoch": 30.810480349344978,
+ "grad_norm": 0.017355090007185936,
+ "learning_rate": 0.0006,
+ "loss": 4.256086349487305,
+ "step": 2218
+ },
+ {
+ "epoch": 30.824454148471617,
+ "grad_norm": 0.016636032611131668,
+ "learning_rate": 0.0006,
+ "loss": 4.227564334869385,
+ "step": 2219
+ },
+ {
+ "epoch": 30.83842794759825,
+ "grad_norm": 0.016273891553282738,
+ "learning_rate": 0.0006,
+ "loss": 4.200466156005859,
+ "step": 2220
+ },
+ {
+ "epoch": 30.85240174672489,
+ "grad_norm": 0.016939911991357803,
+ "learning_rate": 0.0006,
+ "loss": 4.1571221351623535,
+ "step": 2221
+ },
+ {
+ "epoch": 30.86637554585153,
+ "grad_norm": 0.018332522362470627,
+ "learning_rate": 0.0006,
+ "loss": 4.228398323059082,
+ "step": 2222
+ },
+ {
+ "epoch": 30.880349344978168,
+ "grad_norm": 0.019508114084601402,
+ "learning_rate": 0.0006,
+ "loss": 4.222960472106934,
+ "step": 2223
+ },
+ {
+ "epoch": 30.894323144104803,
+ "grad_norm": 0.018440278246998787,
+ "learning_rate": 0.0006,
+ "loss": 4.261051177978516,
+ "step": 2224
+ },
+ {
+ "epoch": 30.90829694323144,
+ "grad_norm": 0.018113229423761368,
+ "learning_rate": 0.0006,
+ "loss": 4.109235763549805,
+ "step": 2225
+ },
+ {
+ "epoch": 30.92227074235808,
+ "grad_norm": 0.015392083674669266,
+ "learning_rate": 0.0006,
+ "loss": 4.21950101852417,
+ "step": 2226
+ },
+ {
+ "epoch": 30.936244541484715,
+ "grad_norm": 0.01633290946483612,
+ "learning_rate": 0.0006,
+ "loss": 4.176053047180176,
+ "step": 2227
+ },
+ {
+ "epoch": 30.950218340611354,
+ "grad_norm": 0.016236383467912674,
+ "learning_rate": 0.0006,
+ "loss": 4.314309120178223,
+ "step": 2228
+ },
+ {
+ "epoch": 30.964192139737992,
+ "grad_norm": 0.015284373424947262,
+ "learning_rate": 0.0006,
+ "loss": 4.303133487701416,
+ "step": 2229
+ },
+ {
+ "epoch": 30.978165938864628,
+ "grad_norm": 0.015982897952198982,
+ "learning_rate": 0.0006,
+ "loss": 4.25739860534668,
+ "step": 2230
+ },
+ {
+ "epoch": 30.992139737991266,
+ "grad_norm": 0.016682198271155357,
+ "learning_rate": 0.0006,
+ "loss": 4.292210578918457,
+ "step": 2231
+ },
+ {
+ "epoch": 31.0,
+ "grad_norm": 0.017521467059850693,
+ "learning_rate": 0.0006,
+ "loss": 4.170637130737305,
+ "step": 2232
+ },
+ {
+ "epoch": 31.0,
+ "eval_loss": 4.642786979675293,
+ "eval_runtime": 57.029,
+ "eval_samples_per_second": 42.82,
+ "eval_steps_per_second": 1.35,
+ "step": 2232
+ },
+ {
+ "epoch": 31.01397379912664,
+ "grad_norm": 0.016607852652668953,
+ "learning_rate": 0.0006,
+ "loss": 4.106224536895752,
+ "step": 2233
+ },
+ {
+ "epoch": 31.027947598253274,
+ "grad_norm": 0.018448904156684875,
+ "learning_rate": 0.0006,
+ "loss": 4.224459648132324,
+ "step": 2234
+ },
+ {
+ "epoch": 31.041921397379912,
+ "grad_norm": 0.020481228828430176,
+ "learning_rate": 0.0006,
+ "loss": 4.305548667907715,
+ "step": 2235
+ },
+ {
+ "epoch": 31.05589519650655,
+ "grad_norm": 0.01999707892537117,
+ "learning_rate": 0.0006,
+ "loss": 4.253663539886475,
+ "step": 2236
+ },
+ {
+ "epoch": 31.069868995633186,
+ "grad_norm": 0.020428303629159927,
+ "learning_rate": 0.0006,
+ "loss": 4.166954040527344,
+ "step": 2237
+ },
+ {
+ "epoch": 31.083842794759825,
+ "grad_norm": 0.022608213126659393,
+ "learning_rate": 0.0006,
+ "loss": 4.172819137573242,
+ "step": 2238
+ },
+ {
+ "epoch": 31.097816593886463,
+ "grad_norm": 0.024669520556926727,
+ "learning_rate": 0.0006,
+ "loss": 4.106586456298828,
+ "step": 2239
+ },
+ {
+ "epoch": 31.111790393013102,
+ "grad_norm": 0.02247351035475731,
+ "learning_rate": 0.0006,
+ "loss": 4.160271644592285,
+ "step": 2240
+ },
+ {
+ "epoch": 31.125764192139737,
+ "grad_norm": 0.0207810141146183,
+ "learning_rate": 0.0006,
+ "loss": 4.272491931915283,
+ "step": 2241
+ },
+ {
+ "epoch": 31.139737991266376,
+ "grad_norm": 0.021576499566435814,
+ "learning_rate": 0.0006,
+ "loss": 4.171474456787109,
+ "step": 2242
+ },
+ {
+ "epoch": 31.153711790393015,
+ "grad_norm": 0.02337106689810753,
+ "learning_rate": 0.0006,
+ "loss": 4.094712734222412,
+ "step": 2243
+ },
+ {
+ "epoch": 31.16768558951965,
+ "grad_norm": 0.02219875156879425,
+ "learning_rate": 0.0006,
+ "loss": 4.1800994873046875,
+ "step": 2244
+ },
+ {
+ "epoch": 31.18165938864629,
+ "grad_norm": 0.022680504247546196,
+ "learning_rate": 0.0006,
+ "loss": 4.2054362297058105,
+ "step": 2245
+ },
+ {
+ "epoch": 31.195633187772927,
+ "grad_norm": 0.0208734143525362,
+ "learning_rate": 0.0006,
+ "loss": 4.06068229675293,
+ "step": 2246
+ },
+ {
+ "epoch": 31.209606986899562,
+ "grad_norm": 0.01811066083610058,
+ "learning_rate": 0.0006,
+ "loss": 4.2465128898620605,
+ "step": 2247
+ },
+ {
+ "epoch": 31.2235807860262,
+ "grad_norm": 0.018571637570858,
+ "learning_rate": 0.0006,
+ "loss": 4.191955089569092,
+ "step": 2248
+ },
+ {
+ "epoch": 31.23755458515284,
+ "grad_norm": 0.018598807975649834,
+ "learning_rate": 0.0006,
+ "loss": 4.200872421264648,
+ "step": 2249
+ },
+ {
+ "epoch": 31.251528384279474,
+ "grad_norm": 0.017577625811100006,
+ "learning_rate": 0.0006,
+ "loss": 4.148700714111328,
+ "step": 2250
+ },
+ {
+ "epoch": 31.265502183406113,
+ "grad_norm": 0.01817980408668518,
+ "learning_rate": 0.0006,
+ "loss": 4.256978988647461,
+ "step": 2251
+ },
+ {
+ "epoch": 31.27947598253275,
+ "grad_norm": 0.017983486875891685,
+ "learning_rate": 0.0006,
+ "loss": 4.175005912780762,
+ "step": 2252
+ },
+ {
+ "epoch": 31.29344978165939,
+ "grad_norm": 0.016751961782574654,
+ "learning_rate": 0.0006,
+ "loss": 4.237518310546875,
+ "step": 2253
+ },
+ {
+ "epoch": 31.307423580786025,
+ "grad_norm": 0.016407202929258347,
+ "learning_rate": 0.0006,
+ "loss": 4.197269439697266,
+ "step": 2254
+ },
+ {
+ "epoch": 31.321397379912664,
+ "grad_norm": 0.01578645594418049,
+ "learning_rate": 0.0006,
+ "loss": 4.179751396179199,
+ "step": 2255
+ },
+ {
+ "epoch": 31.335371179039303,
+ "grad_norm": 0.016001226380467415,
+ "learning_rate": 0.0006,
+ "loss": 4.08821964263916,
+ "step": 2256
+ },
+ {
+ "epoch": 31.349344978165938,
+ "grad_norm": 0.015122811309993267,
+ "learning_rate": 0.0006,
+ "loss": 4.047828674316406,
+ "step": 2257
+ },
+ {
+ "epoch": 31.363318777292577,
+ "grad_norm": 0.015399965457618237,
+ "learning_rate": 0.0006,
+ "loss": 4.1506524085998535,
+ "step": 2258
+ },
+ {
+ "epoch": 31.377292576419215,
+ "grad_norm": 0.01534605398774147,
+ "learning_rate": 0.0006,
+ "loss": 4.27952766418457,
+ "step": 2259
+ },
+ {
+ "epoch": 31.39126637554585,
+ "grad_norm": 0.015461381524801254,
+ "learning_rate": 0.0006,
+ "loss": 4.27102518081665,
+ "step": 2260
+ },
+ {
+ "epoch": 31.40524017467249,
+ "grad_norm": 0.016769153997302055,
+ "learning_rate": 0.0006,
+ "loss": 4.191592216491699,
+ "step": 2261
+ },
+ {
+ "epoch": 31.419213973799128,
+ "grad_norm": 0.016557959839701653,
+ "learning_rate": 0.0006,
+ "loss": 4.159705638885498,
+ "step": 2262
+ },
+ {
+ "epoch": 31.433187772925763,
+ "grad_norm": 0.01659569889307022,
+ "learning_rate": 0.0006,
+ "loss": 4.2136664390563965,
+ "step": 2263
+ },
+ {
+ "epoch": 31.4471615720524,
+ "grad_norm": 0.017101937904953957,
+ "learning_rate": 0.0006,
+ "loss": 4.2716498374938965,
+ "step": 2264
+ },
+ {
+ "epoch": 31.46113537117904,
+ "grad_norm": 0.01722985878586769,
+ "learning_rate": 0.0006,
+ "loss": 4.300534248352051,
+ "step": 2265
+ },
+ {
+ "epoch": 31.475109170305675,
+ "grad_norm": 0.016800185665488243,
+ "learning_rate": 0.0006,
+ "loss": 4.113240718841553,
+ "step": 2266
+ },
+ {
+ "epoch": 31.489082969432314,
+ "grad_norm": 0.01603875495493412,
+ "learning_rate": 0.0006,
+ "loss": 4.157105445861816,
+ "step": 2267
+ },
+ {
+ "epoch": 31.503056768558952,
+ "grad_norm": 0.015903932973742485,
+ "learning_rate": 0.0006,
+ "loss": 4.233669757843018,
+ "step": 2268
+ },
+ {
+ "epoch": 31.51703056768559,
+ "grad_norm": 0.0160983856767416,
+ "learning_rate": 0.0006,
+ "loss": 4.147704124450684,
+ "step": 2269
+ },
+ {
+ "epoch": 31.531004366812226,
+ "grad_norm": 0.01604730263352394,
+ "learning_rate": 0.0006,
+ "loss": 4.1465911865234375,
+ "step": 2270
+ },
+ {
+ "epoch": 31.544978165938865,
+ "grad_norm": 0.016205785796046257,
+ "learning_rate": 0.0006,
+ "loss": 4.106704235076904,
+ "step": 2271
+ },
+ {
+ "epoch": 31.558951965065503,
+ "grad_norm": 0.017008036375045776,
+ "learning_rate": 0.0006,
+ "loss": 4.191000938415527,
+ "step": 2272
+ },
+ {
+ "epoch": 31.57292576419214,
+ "grad_norm": 0.01699635200202465,
+ "learning_rate": 0.0006,
+ "loss": 4.187042236328125,
+ "step": 2273
+ },
+ {
+ "epoch": 31.586899563318777,
+ "grad_norm": 0.01692967116832733,
+ "learning_rate": 0.0006,
+ "loss": 4.208805561065674,
+ "step": 2274
+ },
+ {
+ "epoch": 31.600873362445416,
+ "grad_norm": 0.017608756199479103,
+ "learning_rate": 0.0006,
+ "loss": 4.200512886047363,
+ "step": 2275
+ },
+ {
+ "epoch": 31.61484716157205,
+ "grad_norm": 0.016590990126132965,
+ "learning_rate": 0.0006,
+ "loss": 4.194705009460449,
+ "step": 2276
+ },
+ {
+ "epoch": 31.62882096069869,
+ "grad_norm": 0.0170235987752676,
+ "learning_rate": 0.0006,
+ "loss": 4.210326194763184,
+ "step": 2277
+ },
+ {
+ "epoch": 31.64279475982533,
+ "grad_norm": 0.01604801043868065,
+ "learning_rate": 0.0006,
+ "loss": 4.196024417877197,
+ "step": 2278
+ },
+ {
+ "epoch": 31.656768558951963,
+ "grad_norm": 0.015694202855229378,
+ "learning_rate": 0.0006,
+ "loss": 4.178236484527588,
+ "step": 2279
+ },
+ {
+ "epoch": 31.670742358078602,
+ "grad_norm": 0.015473959036171436,
+ "learning_rate": 0.0006,
+ "loss": 4.259339809417725,
+ "step": 2280
+ },
+ {
+ "epoch": 31.68471615720524,
+ "grad_norm": 0.014689362607896328,
+ "learning_rate": 0.0006,
+ "loss": 3.9702324867248535,
+ "step": 2281
+ },
+ {
+ "epoch": 31.69868995633188,
+ "grad_norm": 0.015245123766362667,
+ "learning_rate": 0.0006,
+ "loss": 4.197025775909424,
+ "step": 2282
+ },
+ {
+ "epoch": 31.712663755458514,
+ "grad_norm": 0.016148347407579422,
+ "learning_rate": 0.0006,
+ "loss": 4.231856346130371,
+ "step": 2283
+ },
+ {
+ "epoch": 31.726637554585153,
+ "grad_norm": 0.015169818885624409,
+ "learning_rate": 0.0006,
+ "loss": 4.230169296264648,
+ "step": 2284
+ },
+ {
+ "epoch": 31.74061135371179,
+ "grad_norm": 0.014843013137578964,
+ "learning_rate": 0.0006,
+ "loss": 4.195633411407471,
+ "step": 2285
+ },
+ {
+ "epoch": 31.754585152838427,
+ "grad_norm": 0.015229340642690659,
+ "learning_rate": 0.0006,
+ "loss": 4.231280326843262,
+ "step": 2286
+ },
+ {
+ "epoch": 31.768558951965066,
+ "grad_norm": 0.01551708485931158,
+ "learning_rate": 0.0006,
+ "loss": 4.17064094543457,
+ "step": 2287
+ },
+ {
+ "epoch": 31.782532751091704,
+ "grad_norm": 0.015863575041294098,
+ "learning_rate": 0.0006,
+ "loss": 4.1579484939575195,
+ "step": 2288
+ },
+ {
+ "epoch": 31.79650655021834,
+ "grad_norm": 0.015118198469281197,
+ "learning_rate": 0.0006,
+ "loss": 4.1626787185668945,
+ "step": 2289
+ },
+ {
+ "epoch": 31.810480349344978,
+ "grad_norm": 0.015220776200294495,
+ "learning_rate": 0.0006,
+ "loss": 4.135746955871582,
+ "step": 2290
+ },
+ {
+ "epoch": 31.824454148471617,
+ "grad_norm": 0.014529784210026264,
+ "learning_rate": 0.0006,
+ "loss": 4.164931297302246,
+ "step": 2291
+ },
+ {
+ "epoch": 31.83842794759825,
+ "grad_norm": 0.015533296391367912,
+ "learning_rate": 0.0006,
+ "loss": 4.112464904785156,
+ "step": 2292
+ },
+ {
+ "epoch": 31.85240174672489,
+ "grad_norm": 0.016562335193157196,
+ "learning_rate": 0.0006,
+ "loss": 4.164898872375488,
+ "step": 2293
+ },
+ {
+ "epoch": 31.86637554585153,
+ "grad_norm": 0.016258113086223602,
+ "learning_rate": 0.0006,
+ "loss": 4.1979265213012695,
+ "step": 2294
+ },
+ {
+ "epoch": 31.880349344978168,
+ "grad_norm": 0.01558110024780035,
+ "learning_rate": 0.0006,
+ "loss": 4.174675941467285,
+ "step": 2295
+ },
+ {
+ "epoch": 31.894323144104803,
+ "grad_norm": 0.016412515193223953,
+ "learning_rate": 0.0006,
+ "loss": 4.217909336090088,
+ "step": 2296
+ },
+ {
+ "epoch": 31.90829694323144,
+ "grad_norm": 0.015118442475795746,
+ "learning_rate": 0.0006,
+ "loss": 4.137143611907959,
+ "step": 2297
+ },
+ {
+ "epoch": 31.92227074235808,
+ "grad_norm": 0.01506161317229271,
+ "learning_rate": 0.0006,
+ "loss": 4.196122169494629,
+ "step": 2298
+ },
+ {
+ "epoch": 31.936244541484715,
+ "grad_norm": 0.017165271565318108,
+ "learning_rate": 0.0006,
+ "loss": 4.14834451675415,
+ "step": 2299
+ },
+ {
+ "epoch": 31.950218340611354,
+ "grad_norm": 0.01696440577507019,
+ "learning_rate": 0.0006,
+ "loss": 4.218868732452393,
+ "step": 2300
+ },
+ {
+ "epoch": 31.964192139737992,
+ "grad_norm": 0.017349394038319588,
+ "learning_rate": 0.0006,
+ "loss": 4.2754011154174805,
+ "step": 2301
+ },
+ {
+ "epoch": 31.978165938864628,
+ "grad_norm": 0.018387921154499054,
+ "learning_rate": 0.0006,
+ "loss": 4.282623291015625,
+ "step": 2302
+ },
+ {
+ "epoch": 31.992139737991266,
+ "grad_norm": 0.018837926909327507,
+ "learning_rate": 0.0006,
+ "loss": 4.185425281524658,
+ "step": 2303
+ },
+ {
+ "epoch": 32.0,
+ "grad_norm": 0.018998464569449425,
+ "learning_rate": 0.0006,
+ "loss": 4.151291847229004,
+ "step": 2304
+ },
+ {
+ "epoch": 32.0,
+ "eval_loss": 4.522614002227783,
+ "eval_runtime": 57.0106,
+ "eval_samples_per_second": 42.834,
+ "eval_steps_per_second": 1.351,
+ "step": 2304
+ },
+ {
+ "epoch": 32.01397379912664,
+ "grad_norm": 0.019036833196878433,
+ "learning_rate": 0.0006,
+ "loss": 4.183202266693115,
+ "step": 2305
+ },
+ {
+ "epoch": 32.02794759825328,
+ "grad_norm": 0.02207397110760212,
+ "learning_rate": 0.0006,
+ "loss": 4.17875862121582,
+ "step": 2306
+ },
+ {
+ "epoch": 32.041921397379916,
+ "grad_norm": 0.022008279338479042,
+ "learning_rate": 0.0006,
+ "loss": 4.08042049407959,
+ "step": 2307
+ },
+ {
+ "epoch": 32.05589519650655,
+ "grad_norm": 0.022407017648220062,
+ "learning_rate": 0.0006,
+ "loss": 4.112189769744873,
+ "step": 2308
+ },
+ {
+ "epoch": 32.069868995633186,
+ "grad_norm": 0.021599747240543365,
+ "learning_rate": 0.0006,
+ "loss": 4.126455307006836,
+ "step": 2309
+ },
+ {
+ "epoch": 32.083842794759825,
+ "grad_norm": 0.020340966060757637,
+ "learning_rate": 0.0006,
+ "loss": 4.080538272857666,
+ "step": 2310
+ },
+ {
+ "epoch": 32.09781659388646,
+ "grad_norm": 0.020956892520189285,
+ "learning_rate": 0.0006,
+ "loss": 4.171943664550781,
+ "step": 2311
+ },
+ {
+ "epoch": 32.1117903930131,
+ "grad_norm": 0.019966667518019676,
+ "learning_rate": 0.0006,
+ "loss": 4.144007682800293,
+ "step": 2312
+ },
+ {
+ "epoch": 32.12576419213974,
+ "grad_norm": 0.018256952986121178,
+ "learning_rate": 0.0006,
+ "loss": 4.208107948303223,
+ "step": 2313
+ },
+ {
+ "epoch": 32.13973799126637,
+ "grad_norm": 0.017289170995354652,
+ "learning_rate": 0.0006,
+ "loss": 4.182024955749512,
+ "step": 2314
+ },
+ {
+ "epoch": 32.15371179039301,
+ "grad_norm": 0.017656700685620308,
+ "learning_rate": 0.0006,
+ "loss": 4.068373680114746,
+ "step": 2315
+ },
+ {
+ "epoch": 32.16768558951965,
+ "grad_norm": 0.017903361469507217,
+ "learning_rate": 0.0006,
+ "loss": 4.2345428466796875,
+ "step": 2316
+ },
+ {
+ "epoch": 32.18165938864629,
+ "grad_norm": 0.01785109005868435,
+ "learning_rate": 0.0006,
+ "loss": 4.219776153564453,
+ "step": 2317
+ },
+ {
+ "epoch": 32.19563318777293,
+ "grad_norm": 0.01887257769703865,
+ "learning_rate": 0.0006,
+ "loss": 4.139206409454346,
+ "step": 2318
+ },
+ {
+ "epoch": 32.209606986899566,
+ "grad_norm": 0.01924685761332512,
+ "learning_rate": 0.0006,
+ "loss": 4.121052265167236,
+ "step": 2319
+ },
+ {
+ "epoch": 32.223580786026204,
+ "grad_norm": 0.020574072375893593,
+ "learning_rate": 0.0006,
+ "loss": 4.13262414932251,
+ "step": 2320
+ },
+ {
+ "epoch": 32.237554585152836,
+ "grad_norm": 0.023849064484238625,
+ "learning_rate": 0.0006,
+ "loss": 4.239119529724121,
+ "step": 2321
+ },
+ {
+ "epoch": 32.251528384279474,
+ "grad_norm": 0.026239361613988876,
+ "learning_rate": 0.0006,
+ "loss": 4.248993873596191,
+ "step": 2322
+ },
+ {
+ "epoch": 32.26550218340611,
+ "grad_norm": 0.023980243131518364,
+ "learning_rate": 0.0006,
+ "loss": 4.159852504730225,
+ "step": 2323
+ },
+ {
+ "epoch": 32.27947598253275,
+ "grad_norm": 0.020442405715584755,
+ "learning_rate": 0.0006,
+ "loss": 4.133546829223633,
+ "step": 2324
+ },
+ {
+ "epoch": 32.29344978165939,
+ "grad_norm": 0.020661186426877975,
+ "learning_rate": 0.0006,
+ "loss": 4.078243732452393,
+ "step": 2325
+ },
+ {
+ "epoch": 32.30742358078603,
+ "grad_norm": 0.02046387456357479,
+ "learning_rate": 0.0006,
+ "loss": 4.172012805938721,
+ "step": 2326
+ },
+ {
+ "epoch": 32.32139737991266,
+ "grad_norm": 0.02089597098529339,
+ "learning_rate": 0.0006,
+ "loss": 4.0604705810546875,
+ "step": 2327
+ },
+ {
+ "epoch": 32.3353711790393,
+ "grad_norm": 0.01875591278076172,
+ "learning_rate": 0.0006,
+ "loss": 4.125911235809326,
+ "step": 2328
+ },
+ {
+ "epoch": 32.34934497816594,
+ "grad_norm": 0.018606796860694885,
+ "learning_rate": 0.0006,
+ "loss": 4.097362518310547,
+ "step": 2329
+ },
+ {
+ "epoch": 32.36331877729258,
+ "grad_norm": 0.01766626164317131,
+ "learning_rate": 0.0006,
+ "loss": 4.118616580963135,
+ "step": 2330
+ },
+ {
+ "epoch": 32.377292576419215,
+ "grad_norm": 0.017692960798740387,
+ "learning_rate": 0.0006,
+ "loss": 4.1457672119140625,
+ "step": 2331
+ },
+ {
+ "epoch": 32.391266375545854,
+ "grad_norm": 0.017619704827666283,
+ "learning_rate": 0.0006,
+ "loss": 4.175961494445801,
+ "step": 2332
+ },
+ {
+ "epoch": 32.40524017467249,
+ "grad_norm": 0.01784413494169712,
+ "learning_rate": 0.0006,
+ "loss": 4.051291465759277,
+ "step": 2333
+ },
+ {
+ "epoch": 32.419213973799124,
+ "grad_norm": 0.018508171662688255,
+ "learning_rate": 0.0006,
+ "loss": 4.0638556480407715,
+ "step": 2334
+ },
+ {
+ "epoch": 32.43318777292576,
+ "grad_norm": 0.01731436885893345,
+ "learning_rate": 0.0006,
+ "loss": 4.1805419921875,
+ "step": 2335
+ },
+ {
+ "epoch": 32.4471615720524,
+ "grad_norm": 0.015886498615145683,
+ "learning_rate": 0.0006,
+ "loss": 4.151026725769043,
+ "step": 2336
+ },
+ {
+ "epoch": 32.46113537117904,
+ "grad_norm": 0.016583820804953575,
+ "learning_rate": 0.0006,
+ "loss": 4.203604698181152,
+ "step": 2337
+ },
+ {
+ "epoch": 32.47510917030568,
+ "grad_norm": 0.0176097322255373,
+ "learning_rate": 0.0006,
+ "loss": 4.135159015655518,
+ "step": 2338
+ },
+ {
+ "epoch": 32.48908296943232,
+ "grad_norm": 0.016657903790473938,
+ "learning_rate": 0.0006,
+ "loss": 4.126681804656982,
+ "step": 2339
+ },
+ {
+ "epoch": 32.50305676855895,
+ "grad_norm": 0.016866130754351616,
+ "learning_rate": 0.0006,
+ "loss": 4.162341117858887,
+ "step": 2340
+ },
+ {
+ "epoch": 32.51703056768559,
+ "grad_norm": 0.016286680474877357,
+ "learning_rate": 0.0006,
+ "loss": 4.261801242828369,
+ "step": 2341
+ },
+ {
+ "epoch": 32.531004366812226,
+ "grad_norm": 0.01481365505605936,
+ "learning_rate": 0.0006,
+ "loss": 4.110016822814941,
+ "step": 2342
+ },
+ {
+ "epoch": 32.544978165938865,
+ "grad_norm": 0.016605013981461525,
+ "learning_rate": 0.0006,
+ "loss": 4.230024814605713,
+ "step": 2343
+ },
+ {
+ "epoch": 32.5589519650655,
+ "grad_norm": 0.01645175740122795,
+ "learning_rate": 0.0006,
+ "loss": 4.174612045288086,
+ "step": 2344
+ },
+ {
+ "epoch": 32.57292576419214,
+ "grad_norm": 0.014725026674568653,
+ "learning_rate": 0.0006,
+ "loss": 4.066371440887451,
+ "step": 2345
+ },
+ {
+ "epoch": 32.58689956331878,
+ "grad_norm": 0.015003837645053864,
+ "learning_rate": 0.0006,
+ "loss": 4.193343639373779,
+ "step": 2346
+ },
+ {
+ "epoch": 32.60087336244541,
+ "grad_norm": 0.015346302650868893,
+ "learning_rate": 0.0006,
+ "loss": 4.210659980773926,
+ "step": 2347
+ },
+ {
+ "epoch": 32.61484716157205,
+ "grad_norm": 0.015135962516069412,
+ "learning_rate": 0.0006,
+ "loss": 4.204618453979492,
+ "step": 2348
+ },
+ {
+ "epoch": 32.62882096069869,
+ "grad_norm": 0.01601956970989704,
+ "learning_rate": 0.0006,
+ "loss": 4.145564556121826,
+ "step": 2349
+ },
+ {
+ "epoch": 32.64279475982533,
+ "grad_norm": 0.01517851185053587,
+ "learning_rate": 0.0006,
+ "loss": 4.051673889160156,
+ "step": 2350
+ },
+ {
+ "epoch": 32.65676855895197,
+ "grad_norm": 0.014894292689859867,
+ "learning_rate": 0.0006,
+ "loss": 4.184818267822266,
+ "step": 2351
+ },
+ {
+ "epoch": 32.670742358078606,
+ "grad_norm": 0.015720834955573082,
+ "learning_rate": 0.0006,
+ "loss": 4.220036506652832,
+ "step": 2352
+ },
+ {
+ "epoch": 32.68471615720524,
+ "grad_norm": 0.01573433168232441,
+ "learning_rate": 0.0006,
+ "loss": 4.076773166656494,
+ "step": 2353
+ },
+ {
+ "epoch": 32.698689956331876,
+ "grad_norm": 0.015828052535653114,
+ "learning_rate": 0.0006,
+ "loss": 4.116329669952393,
+ "step": 2354
+ },
+ {
+ "epoch": 32.712663755458514,
+ "grad_norm": 0.015462259761989117,
+ "learning_rate": 0.0006,
+ "loss": 4.1155242919921875,
+ "step": 2355
+ },
+ {
+ "epoch": 32.72663755458515,
+ "grad_norm": 0.015763849020004272,
+ "learning_rate": 0.0006,
+ "loss": 4.169687747955322,
+ "step": 2356
+ },
+ {
+ "epoch": 32.74061135371179,
+ "grad_norm": 0.014935145154595375,
+ "learning_rate": 0.0006,
+ "loss": 4.2090911865234375,
+ "step": 2357
+ },
+ {
+ "epoch": 32.75458515283843,
+ "grad_norm": 0.015119184739887714,
+ "learning_rate": 0.0006,
+ "loss": 4.131838798522949,
+ "step": 2358
+ },
+ {
+ "epoch": 32.76855895196506,
+ "grad_norm": 0.014833525754511356,
+ "learning_rate": 0.0006,
+ "loss": 4.17618465423584,
+ "step": 2359
+ },
+ {
+ "epoch": 32.7825327510917,
+ "grad_norm": 0.01588359661400318,
+ "learning_rate": 0.0006,
+ "loss": 4.163723945617676,
+ "step": 2360
+ },
+ {
+ "epoch": 32.79650655021834,
+ "grad_norm": 0.0168188214302063,
+ "learning_rate": 0.0006,
+ "loss": 4.178189277648926,
+ "step": 2361
+ },
+ {
+ "epoch": 32.81048034934498,
+ "grad_norm": 0.01726095750927925,
+ "learning_rate": 0.0006,
+ "loss": 4.144223690032959,
+ "step": 2362
+ },
+ {
+ "epoch": 32.82445414847162,
+ "grad_norm": 0.01701410301029682,
+ "learning_rate": 0.0006,
+ "loss": 4.209896087646484,
+ "step": 2363
+ },
+ {
+ "epoch": 32.838427947598255,
+ "grad_norm": 0.01682264916598797,
+ "learning_rate": 0.0006,
+ "loss": 4.26223087310791,
+ "step": 2364
+ },
+ {
+ "epoch": 32.852401746724894,
+ "grad_norm": 0.015597456134855747,
+ "learning_rate": 0.0006,
+ "loss": 4.292664051055908,
+ "step": 2365
+ },
+ {
+ "epoch": 32.866375545851525,
+ "grad_norm": 0.015258579514920712,
+ "learning_rate": 0.0006,
+ "loss": 4.165006637573242,
+ "step": 2366
+ },
+ {
+ "epoch": 32.880349344978164,
+ "grad_norm": 0.01662200503051281,
+ "learning_rate": 0.0006,
+ "loss": 4.309786319732666,
+ "step": 2367
+ },
+ {
+ "epoch": 32.8943231441048,
+ "grad_norm": 0.015691161155700684,
+ "learning_rate": 0.0006,
+ "loss": 4.167202949523926,
+ "step": 2368
+ },
+ {
+ "epoch": 32.90829694323144,
+ "grad_norm": 0.014870637096464634,
+ "learning_rate": 0.0006,
+ "loss": 4.142415523529053,
+ "step": 2369
+ },
+ {
+ "epoch": 32.92227074235808,
+ "grad_norm": 0.01567252166569233,
+ "learning_rate": 0.0006,
+ "loss": 4.2312092781066895,
+ "step": 2370
+ },
+ {
+ "epoch": 32.93624454148472,
+ "grad_norm": 0.014948616735637188,
+ "learning_rate": 0.0006,
+ "loss": 4.1514129638671875,
+ "step": 2371
+ },
+ {
+ "epoch": 32.95021834061135,
+ "grad_norm": 0.014654500409960747,
+ "learning_rate": 0.0006,
+ "loss": 4.205010890960693,
+ "step": 2372
+ },
+ {
+ "epoch": 32.96419213973799,
+ "grad_norm": 0.015059148892760277,
+ "learning_rate": 0.0006,
+ "loss": 4.148405075073242,
+ "step": 2373
+ },
+ {
+ "epoch": 32.97816593886463,
+ "grad_norm": 0.015026576817035675,
+ "learning_rate": 0.0006,
+ "loss": 4.182535171508789,
+ "step": 2374
+ },
+ {
+ "epoch": 32.992139737991266,
+ "grad_norm": 0.016394320875406265,
+ "learning_rate": 0.0006,
+ "loss": 4.129051208496094,
+ "step": 2375
+ },
+ {
+ "epoch": 33.0,
+ "grad_norm": 0.01867043413221836,
+ "learning_rate": 0.0006,
+ "loss": 4.185556411743164,
+ "step": 2376
+ },
+ {
+ "epoch": 33.0,
+ "eval_loss": 4.640405654907227,
+ "eval_runtime": 56.5861,
+ "eval_samples_per_second": 43.155,
+ "eval_steps_per_second": 1.361,
+ "step": 2376
+ },
+ {
+ "epoch": 33.01397379912664,
+ "grad_norm": 0.018626727163791656,
+ "learning_rate": 0.0006,
+ "loss": 4.143031597137451,
+ "step": 2377
+ },
+ {
+ "epoch": 33.02794759825328,
+ "grad_norm": 0.019283363595604897,
+ "learning_rate": 0.0006,
+ "loss": 4.117845058441162,
+ "step": 2378
+ },
+ {
+ "epoch": 33.041921397379916,
+ "grad_norm": 0.018185822293162346,
+ "learning_rate": 0.0006,
+ "loss": 4.076862812042236,
+ "step": 2379
+ },
+ {
+ "epoch": 33.05589519650655,
+ "grad_norm": 0.019411196932196617,
+ "learning_rate": 0.0006,
+ "loss": 4.124156475067139,
+ "step": 2380
+ },
+ {
+ "epoch": 33.069868995633186,
+ "grad_norm": 0.019248466938734055,
+ "learning_rate": 0.0006,
+ "loss": 4.201735019683838,
+ "step": 2381
+ },
+ {
+ "epoch": 33.083842794759825,
+ "grad_norm": 0.018856720998883247,
+ "learning_rate": 0.0006,
+ "loss": 4.00296688079834,
+ "step": 2382
+ },
+ {
+ "epoch": 33.09781659388646,
+ "grad_norm": 0.018554294481873512,
+ "learning_rate": 0.0006,
+ "loss": 4.270577430725098,
+ "step": 2383
+ },
+ {
+ "epoch": 33.1117903930131,
+ "grad_norm": 0.019946224987506866,
+ "learning_rate": 0.0006,
+ "loss": 4.119589328765869,
+ "step": 2384
+ },
+ {
+ "epoch": 33.12576419213974,
+ "grad_norm": 0.021395670250058174,
+ "learning_rate": 0.0006,
+ "loss": 4.128847122192383,
+ "step": 2385
+ },
+ {
+ "epoch": 33.13973799126637,
+ "grad_norm": 0.02009674906730652,
+ "learning_rate": 0.0006,
+ "loss": 4.183228969573975,
+ "step": 2386
+ },
+ {
+ "epoch": 33.15371179039301,
+ "grad_norm": 0.01820223033428192,
+ "learning_rate": 0.0006,
+ "loss": 4.131601810455322,
+ "step": 2387
+ },
+ {
+ "epoch": 33.16768558951965,
+ "grad_norm": 0.018653156235814095,
+ "learning_rate": 0.0006,
+ "loss": 4.116430282592773,
+ "step": 2388
+ },
+ {
+ "epoch": 33.18165938864629,
+ "grad_norm": 0.016828063875436783,
+ "learning_rate": 0.0006,
+ "loss": 4.054663181304932,
+ "step": 2389
+ },
+ {
+ "epoch": 33.19563318777293,
+ "grad_norm": 0.017025409266352654,
+ "learning_rate": 0.0006,
+ "loss": 4.110044956207275,
+ "step": 2390
+ },
+ {
+ "epoch": 33.209606986899566,
+ "grad_norm": 0.016608452424407005,
+ "learning_rate": 0.0006,
+ "loss": 4.102668285369873,
+ "step": 2391
+ },
+ {
+ "epoch": 33.223580786026204,
+ "grad_norm": 0.01743577979505062,
+ "learning_rate": 0.0006,
+ "loss": 4.166791915893555,
+ "step": 2392
+ },
+ {
+ "epoch": 33.237554585152836,
+ "grad_norm": 0.017253734171390533,
+ "learning_rate": 0.0006,
+ "loss": 4.217832088470459,
+ "step": 2393
+ },
+ {
+ "epoch": 33.251528384279474,
+ "grad_norm": 0.017937595024704933,
+ "learning_rate": 0.0006,
+ "loss": 4.2173967361450195,
+ "step": 2394
+ },
+ {
+ "epoch": 33.26550218340611,
+ "grad_norm": 0.01985993981361389,
+ "learning_rate": 0.0006,
+ "loss": 4.17849588394165,
+ "step": 2395
+ },
+ {
+ "epoch": 33.27947598253275,
+ "grad_norm": 0.019243355840444565,
+ "learning_rate": 0.0006,
+ "loss": 4.086523056030273,
+ "step": 2396
+ },
+ {
+ "epoch": 33.29344978165939,
+ "grad_norm": 0.01924416422843933,
+ "learning_rate": 0.0006,
+ "loss": 4.128574371337891,
+ "step": 2397
+ },
+ {
+ "epoch": 33.30742358078603,
+ "grad_norm": 0.01983758620917797,
+ "learning_rate": 0.0006,
+ "loss": 4.061581611633301,
+ "step": 2398
+ },
+ {
+ "epoch": 33.32139737991266,
+ "grad_norm": 0.01930462196469307,
+ "learning_rate": 0.0006,
+ "loss": 4.091731071472168,
+ "step": 2399
+ },
+ {
+ "epoch": 33.3353711790393,
+ "grad_norm": 0.017988424748182297,
+ "learning_rate": 0.0006,
+ "loss": 4.155599117279053,
+ "step": 2400
+ },
+ {
+ "epoch": 33.34934497816594,
+ "grad_norm": 0.015194767154753208,
+ "learning_rate": 0.0006,
+ "loss": 4.144134044647217,
+ "step": 2401
+ },
+ {
+ "epoch": 33.36331877729258,
+ "grad_norm": 0.01536081824451685,
+ "learning_rate": 0.0006,
+ "loss": 4.127272129058838,
+ "step": 2402
+ },
+ {
+ "epoch": 33.377292576419215,
+ "grad_norm": 0.015431297942996025,
+ "learning_rate": 0.0006,
+ "loss": 4.3213019371032715,
+ "step": 2403
+ },
+ {
+ "epoch": 33.391266375545854,
+ "grad_norm": 0.015740280970931053,
+ "learning_rate": 0.0006,
+ "loss": 4.121467590332031,
+ "step": 2404
+ },
+ {
+ "epoch": 33.40524017467249,
+ "grad_norm": 0.015379384160041809,
+ "learning_rate": 0.0006,
+ "loss": 4.095216751098633,
+ "step": 2405
+ },
+ {
+ "epoch": 33.419213973799124,
+ "grad_norm": 0.01528073474764824,
+ "learning_rate": 0.0006,
+ "loss": 4.13047981262207,
+ "step": 2406
+ },
+ {
+ "epoch": 33.43318777292576,
+ "grad_norm": 0.015739979222416878,
+ "learning_rate": 0.0006,
+ "loss": 4.2533135414123535,
+ "step": 2407
+ },
+ {
+ "epoch": 33.4471615720524,
+ "grad_norm": 0.01519764307886362,
+ "learning_rate": 0.0006,
+ "loss": 4.131782054901123,
+ "step": 2408
+ },
+ {
+ "epoch": 33.46113537117904,
+ "grad_norm": 0.015425996854901314,
+ "learning_rate": 0.0006,
+ "loss": 4.02811336517334,
+ "step": 2409
+ },
+ {
+ "epoch": 33.47510917030568,
+ "grad_norm": 0.015378830023109913,
+ "learning_rate": 0.0006,
+ "loss": 4.08652925491333,
+ "step": 2410
+ },
+ {
+ "epoch": 33.48908296943232,
+ "grad_norm": 0.01473075058311224,
+ "learning_rate": 0.0006,
+ "loss": 4.214364528656006,
+ "step": 2411
+ },
+ {
+ "epoch": 33.50305676855895,
+ "grad_norm": 0.015141637064516544,
+ "learning_rate": 0.0006,
+ "loss": 4.155712127685547,
+ "step": 2412
+ },
+ {
+ "epoch": 33.51703056768559,
+ "grad_norm": 0.01641593500971794,
+ "learning_rate": 0.0006,
+ "loss": 4.097907066345215,
+ "step": 2413
+ },
+ {
+ "epoch": 33.531004366812226,
+ "grad_norm": 0.015182029455900192,
+ "learning_rate": 0.0006,
+ "loss": 4.238499641418457,
+ "step": 2414
+ },
+ {
+ "epoch": 33.544978165938865,
+ "grad_norm": 0.01643732376396656,
+ "learning_rate": 0.0006,
+ "loss": 4.0661492347717285,
+ "step": 2415
+ },
+ {
+ "epoch": 33.5589519650655,
+ "grad_norm": 0.015614228323101997,
+ "learning_rate": 0.0006,
+ "loss": 4.120795249938965,
+ "step": 2416
+ },
+ {
+ "epoch": 33.57292576419214,
+ "grad_norm": 0.016066187992691994,
+ "learning_rate": 0.0006,
+ "loss": 4.096465110778809,
+ "step": 2417
+ },
+ {
+ "epoch": 33.58689956331878,
+ "grad_norm": 0.017214342951774597,
+ "learning_rate": 0.0006,
+ "loss": 4.242628574371338,
+ "step": 2418
+ },
+ {
+ "epoch": 33.60087336244541,
+ "grad_norm": 0.01787586510181427,
+ "learning_rate": 0.0006,
+ "loss": 4.149589538574219,
+ "step": 2419
+ },
+ {
+ "epoch": 33.61484716157205,
+ "grad_norm": 0.019455241039395332,
+ "learning_rate": 0.0006,
+ "loss": 4.267796039581299,
+ "step": 2420
+ },
+ {
+ "epoch": 33.62882096069869,
+ "grad_norm": 0.01984674669802189,
+ "learning_rate": 0.0006,
+ "loss": 4.175220966339111,
+ "step": 2421
+ },
+ {
+ "epoch": 33.64279475982533,
+ "grad_norm": 0.021933717653155327,
+ "learning_rate": 0.0006,
+ "loss": 4.106647491455078,
+ "step": 2422
+ },
+ {
+ "epoch": 33.65676855895197,
+ "grad_norm": 0.02123589813709259,
+ "learning_rate": 0.0006,
+ "loss": 4.204996109008789,
+ "step": 2423
+ },
+ {
+ "epoch": 33.670742358078606,
+ "grad_norm": 0.019814589992165565,
+ "learning_rate": 0.0006,
+ "loss": 4.186171531677246,
+ "step": 2424
+ },
+ {
+ "epoch": 33.68471615720524,
+ "grad_norm": 0.02059728465974331,
+ "learning_rate": 0.0006,
+ "loss": 4.170509338378906,
+ "step": 2425
+ },
+ {
+ "epoch": 33.698689956331876,
+ "grad_norm": 0.018483608961105347,
+ "learning_rate": 0.0006,
+ "loss": 4.085585594177246,
+ "step": 2426
+ },
+ {
+ "epoch": 33.712663755458514,
+ "grad_norm": 0.018403852358460426,
+ "learning_rate": 0.0006,
+ "loss": 4.2551774978637695,
+ "step": 2427
+ },
+ {
+ "epoch": 33.72663755458515,
+ "grad_norm": 0.018653830513358116,
+ "learning_rate": 0.0006,
+ "loss": 4.091766357421875,
+ "step": 2428
+ },
+ {
+ "epoch": 33.74061135371179,
+ "grad_norm": 0.017692334949970245,
+ "learning_rate": 0.0006,
+ "loss": 4.100838661193848,
+ "step": 2429
+ },
+ {
+ "epoch": 33.75458515283843,
+ "grad_norm": 0.016029126942157745,
+ "learning_rate": 0.0006,
+ "loss": 4.1530680656433105,
+ "step": 2430
+ },
+ {
+ "epoch": 33.76855895196506,
+ "grad_norm": 0.016095684841275215,
+ "learning_rate": 0.0006,
+ "loss": 4.20544958114624,
+ "step": 2431
+ },
+ {
+ "epoch": 33.7825327510917,
+ "grad_norm": 0.016375798732042313,
+ "learning_rate": 0.0006,
+ "loss": 4.199177265167236,
+ "step": 2432
+ },
+ {
+ "epoch": 33.79650655021834,
+ "grad_norm": 0.01590786688029766,
+ "learning_rate": 0.0006,
+ "loss": 4.200799465179443,
+ "step": 2433
+ },
+ {
+ "epoch": 33.81048034934498,
+ "grad_norm": 0.014896097593009472,
+ "learning_rate": 0.0006,
+ "loss": 4.185850143432617,
+ "step": 2434
+ },
+ {
+ "epoch": 33.82445414847162,
+ "grad_norm": 0.01438513956964016,
+ "learning_rate": 0.0006,
+ "loss": 4.214962005615234,
+ "step": 2435
+ },
+ {
+ "epoch": 33.838427947598255,
+ "grad_norm": 0.014459339901804924,
+ "learning_rate": 0.0006,
+ "loss": 4.196972370147705,
+ "step": 2436
+ },
+ {
+ "epoch": 33.852401746724894,
+ "grad_norm": 0.014782169833779335,
+ "learning_rate": 0.0006,
+ "loss": 4.0885910987854,
+ "step": 2437
+ },
+ {
+ "epoch": 33.866375545851525,
+ "grad_norm": 0.01520370040088892,
+ "learning_rate": 0.0006,
+ "loss": 4.0323262214660645,
+ "step": 2438
+ },
+ {
+ "epoch": 33.880349344978164,
+ "grad_norm": 0.014056282117962837,
+ "learning_rate": 0.0006,
+ "loss": 4.1386518478393555,
+ "step": 2439
+ },
+ {
+ "epoch": 33.8943231441048,
+ "grad_norm": 0.013567990623414516,
+ "learning_rate": 0.0006,
+ "loss": 4.095754623413086,
+ "step": 2440
+ },
+ {
+ "epoch": 33.90829694323144,
+ "grad_norm": 0.013306716457009315,
+ "learning_rate": 0.0006,
+ "loss": 4.1623077392578125,
+ "step": 2441
+ },
+ {
+ "epoch": 33.92227074235808,
+ "grad_norm": 0.014047150500118732,
+ "learning_rate": 0.0006,
+ "loss": 4.142541885375977,
+ "step": 2442
+ },
+ {
+ "epoch": 33.93624454148472,
+ "grad_norm": 0.01377617847174406,
+ "learning_rate": 0.0006,
+ "loss": 4.268551826477051,
+ "step": 2443
+ },
+ {
+ "epoch": 33.95021834061135,
+ "grad_norm": 0.014671621844172478,
+ "learning_rate": 0.0006,
+ "loss": 4.198196887969971,
+ "step": 2444
+ },
+ {
+ "epoch": 33.96419213973799,
+ "grad_norm": 0.014806744642555714,
+ "learning_rate": 0.0006,
+ "loss": 4.057899475097656,
+ "step": 2445
+ },
+ {
+ "epoch": 33.97816593886463,
+ "grad_norm": 0.015734272077679634,
+ "learning_rate": 0.0006,
+ "loss": 4.163505554199219,
+ "step": 2446
+ },
+ {
+ "epoch": 33.992139737991266,
+ "grad_norm": 0.015250146389007568,
+ "learning_rate": 0.0006,
+ "loss": 4.076294898986816,
+ "step": 2447
+ },
+ {
+ "epoch": 34.0,
+ "grad_norm": 0.016733702272176743,
+ "learning_rate": 0.0006,
+ "loss": 4.138420104980469,
+ "step": 2448
+ },
+ {
+ "epoch": 34.0,
+ "eval_loss": 4.542503356933594,
+ "eval_runtime": 56.983,
+ "eval_samples_per_second": 42.855,
+ "eval_steps_per_second": 1.351,
+ "step": 2448
+ },
+ {
+ "epoch": 34.01397379912664,
+ "grad_norm": 0.015866613015532494,
+ "learning_rate": 0.0006,
+ "loss": 4.032081127166748,
+ "step": 2449
+ },
+ {
+ "epoch": 34.02794759825328,
+ "grad_norm": 0.01946570910513401,
+ "learning_rate": 0.0006,
+ "loss": 4.122320175170898,
+ "step": 2450
+ },
+ {
+ "epoch": 34.041921397379916,
+ "grad_norm": 0.020265938714146614,
+ "learning_rate": 0.0006,
+ "loss": 4.06460428237915,
+ "step": 2451
+ },
+ {
+ "epoch": 34.05589519650655,
+ "grad_norm": 0.018887171521782875,
+ "learning_rate": 0.0006,
+ "loss": 4.096906661987305,
+ "step": 2452
+ },
+ {
+ "epoch": 34.069868995633186,
+ "grad_norm": 0.017304692417383194,
+ "learning_rate": 0.0006,
+ "loss": 4.137261390686035,
+ "step": 2453
+ },
+ {
+ "epoch": 34.083842794759825,
+ "grad_norm": 0.017029082402586937,
+ "learning_rate": 0.0006,
+ "loss": 4.081112861633301,
+ "step": 2454
+ },
+ {
+ "epoch": 34.09781659388646,
+ "grad_norm": 0.016414813697338104,
+ "learning_rate": 0.0006,
+ "loss": 4.140783309936523,
+ "step": 2455
+ },
+ {
+ "epoch": 34.1117903930131,
+ "grad_norm": 0.018238281831145287,
+ "learning_rate": 0.0006,
+ "loss": 4.161116600036621,
+ "step": 2456
+ },
+ {
+ "epoch": 34.12576419213974,
+ "grad_norm": 0.019929232075810432,
+ "learning_rate": 0.0006,
+ "loss": 4.243722438812256,
+ "step": 2457
+ },
+ {
+ "epoch": 34.13973799126637,
+ "grad_norm": 0.019375786185264587,
+ "learning_rate": 0.0006,
+ "loss": 4.134873390197754,
+ "step": 2458
+ },
+ {
+ "epoch": 34.15371179039301,
+ "grad_norm": 0.018151137977838516,
+ "learning_rate": 0.0006,
+ "loss": 4.098745346069336,
+ "step": 2459
+ },
+ {
+ "epoch": 34.16768558951965,
+ "grad_norm": 0.017054934054613113,
+ "learning_rate": 0.0006,
+ "loss": 4.194662570953369,
+ "step": 2460
+ },
+ {
+ "epoch": 34.18165938864629,
+ "grad_norm": 0.018220555037260056,
+ "learning_rate": 0.0006,
+ "loss": 4.128866195678711,
+ "step": 2461
+ },
+ {
+ "epoch": 34.19563318777293,
+ "grad_norm": 0.01692879945039749,
+ "learning_rate": 0.0006,
+ "loss": 4.062477111816406,
+ "step": 2462
+ },
+ {
+ "epoch": 34.209606986899566,
+ "grad_norm": 0.01674199476838112,
+ "learning_rate": 0.0006,
+ "loss": 4.186149597167969,
+ "step": 2463
+ },
+ {
+ "epoch": 34.223580786026204,
+ "grad_norm": 0.016527455300092697,
+ "learning_rate": 0.0006,
+ "loss": 4.0734028816223145,
+ "step": 2464
+ },
+ {
+ "epoch": 34.237554585152836,
+ "grad_norm": 0.016448386013507843,
+ "learning_rate": 0.0006,
+ "loss": 4.098896026611328,
+ "step": 2465
+ },
+ {
+ "epoch": 34.251528384279474,
+ "grad_norm": 0.014644953422248363,
+ "learning_rate": 0.0006,
+ "loss": 4.1031904220581055,
+ "step": 2466
+ },
+ {
+ "epoch": 34.26550218340611,
+ "grad_norm": 0.016649341210722923,
+ "learning_rate": 0.0006,
+ "loss": 4.148791313171387,
+ "step": 2467
+ },
+ {
+ "epoch": 34.27947598253275,
+ "grad_norm": 0.017442714422941208,
+ "learning_rate": 0.0006,
+ "loss": 4.070799827575684,
+ "step": 2468
+ },
+ {
+ "epoch": 34.29344978165939,
+ "grad_norm": 0.017852013930678368,
+ "learning_rate": 0.0006,
+ "loss": 4.031924247741699,
+ "step": 2469
+ },
+ {
+ "epoch": 34.30742358078603,
+ "grad_norm": 0.016371270641684532,
+ "learning_rate": 0.0006,
+ "loss": 4.149420738220215,
+ "step": 2470
+ },
+ {
+ "epoch": 34.32139737991266,
+ "grad_norm": 0.015846814960241318,
+ "learning_rate": 0.0006,
+ "loss": 4.15427303314209,
+ "step": 2471
+ },
+ {
+ "epoch": 34.3353711790393,
+ "grad_norm": 0.017411163076758385,
+ "learning_rate": 0.0006,
+ "loss": 4.1304192543029785,
+ "step": 2472
+ },
+ {
+ "epoch": 34.34934497816594,
+ "grad_norm": 0.01780046708881855,
+ "learning_rate": 0.0006,
+ "loss": 4.191732406616211,
+ "step": 2473
+ },
+ {
+ "epoch": 34.36331877729258,
+ "grad_norm": 0.01958383060991764,
+ "learning_rate": 0.0006,
+ "loss": 4.068355083465576,
+ "step": 2474
+ },
+ {
+ "epoch": 34.377292576419215,
+ "grad_norm": 0.022173317149281502,
+ "learning_rate": 0.0006,
+ "loss": 4.182929039001465,
+ "step": 2475
+ },
+ {
+ "epoch": 34.391266375545854,
+ "grad_norm": 0.022389156743884087,
+ "learning_rate": 0.0006,
+ "loss": 4.110218048095703,
+ "step": 2476
+ },
+ {
+ "epoch": 34.40524017467249,
+ "grad_norm": 0.022478511556982994,
+ "learning_rate": 0.0006,
+ "loss": 4.095479488372803,
+ "step": 2477
+ },
+ {
+ "epoch": 34.419213973799124,
+ "grad_norm": 0.019504647701978683,
+ "learning_rate": 0.0006,
+ "loss": 4.101964950561523,
+ "step": 2478
+ },
+ {
+ "epoch": 34.43318777292576,
+ "grad_norm": 0.019378576427698135,
+ "learning_rate": 0.0006,
+ "loss": 4.14620304107666,
+ "step": 2479
+ },
+ {
+ "epoch": 34.4471615720524,
+ "grad_norm": 0.017993204295635223,
+ "learning_rate": 0.0006,
+ "loss": 4.183492660522461,
+ "step": 2480
+ },
+ {
+ "epoch": 34.46113537117904,
+ "grad_norm": 0.01814805157482624,
+ "learning_rate": 0.0006,
+ "loss": 4.151247978210449,
+ "step": 2481
+ },
+ {
+ "epoch": 34.47510917030568,
+ "grad_norm": 0.018622448667883873,
+ "learning_rate": 0.0006,
+ "loss": 4.020938396453857,
+ "step": 2482
+ },
+ {
+ "epoch": 34.48908296943232,
+ "grad_norm": 0.018360132351517677,
+ "learning_rate": 0.0006,
+ "loss": 4.064734935760498,
+ "step": 2483
+ },
+ {
+ "epoch": 34.50305676855895,
+ "grad_norm": 0.017715711146593094,
+ "learning_rate": 0.0006,
+ "loss": 4.185687065124512,
+ "step": 2484
+ },
+ {
+ "epoch": 34.51703056768559,
+ "grad_norm": 0.01714295893907547,
+ "learning_rate": 0.0006,
+ "loss": 4.1044721603393555,
+ "step": 2485
+ },
+ {
+ "epoch": 34.531004366812226,
+ "grad_norm": 0.016943803057074547,
+ "learning_rate": 0.0006,
+ "loss": 4.106837272644043,
+ "step": 2486
+ },
+ {
+ "epoch": 34.544978165938865,
+ "grad_norm": 0.0182423684746027,
+ "learning_rate": 0.0006,
+ "loss": 4.1041340827941895,
+ "step": 2487
+ },
+ {
+ "epoch": 34.5589519650655,
+ "grad_norm": 0.01636587455868721,
+ "learning_rate": 0.0006,
+ "loss": 3.9735546112060547,
+ "step": 2488
+ },
+ {
+ "epoch": 34.57292576419214,
+ "grad_norm": 0.014593546278774738,
+ "learning_rate": 0.0006,
+ "loss": 4.041980266571045,
+ "step": 2489
+ },
+ {
+ "epoch": 34.58689956331878,
+ "grad_norm": 0.015436965972185135,
+ "learning_rate": 0.0006,
+ "loss": 4.301148414611816,
+ "step": 2490
+ },
+ {
+ "epoch": 34.60087336244541,
+ "grad_norm": 0.015075269155204296,
+ "learning_rate": 0.0006,
+ "loss": 4.118527412414551,
+ "step": 2491
+ },
+ {
+ "epoch": 34.61484716157205,
+ "grad_norm": 0.01701514609158039,
+ "learning_rate": 0.0006,
+ "loss": 4.245968818664551,
+ "step": 2492
+ },
+ {
+ "epoch": 34.62882096069869,
+ "grad_norm": 0.018047887831926346,
+ "learning_rate": 0.0006,
+ "loss": 4.227261066436768,
+ "step": 2493
+ },
+ {
+ "epoch": 34.64279475982533,
+ "grad_norm": 0.018499085679650307,
+ "learning_rate": 0.0006,
+ "loss": 4.121360778808594,
+ "step": 2494
+ },
+ {
+ "epoch": 34.65676855895197,
+ "grad_norm": 0.018987711519002914,
+ "learning_rate": 0.0006,
+ "loss": 4.131564617156982,
+ "step": 2495
+ },
+ {
+ "epoch": 34.670742358078606,
+ "grad_norm": 0.01834476925432682,
+ "learning_rate": 0.0006,
+ "loss": 4.224941253662109,
+ "step": 2496
+ },
+ {
+ "epoch": 34.68471615720524,
+ "grad_norm": 0.017507942393422127,
+ "learning_rate": 0.0006,
+ "loss": 4.1394548416137695,
+ "step": 2497
+ },
+ {
+ "epoch": 34.698689956331876,
+ "grad_norm": 0.01680580899119377,
+ "learning_rate": 0.0006,
+ "loss": 4.1073431968688965,
+ "step": 2498
+ },
+ {
+ "epoch": 34.712663755458514,
+ "grad_norm": 0.015233514830470085,
+ "learning_rate": 0.0006,
+ "loss": 4.136994361877441,
+ "step": 2499
+ },
+ {
+ "epoch": 34.72663755458515,
+ "grad_norm": 0.015482917428016663,
+ "learning_rate": 0.0006,
+ "loss": 4.110719680786133,
+ "step": 2500
+ },
+ {
+ "epoch": 34.74061135371179,
+ "grad_norm": 0.015940461307764053,
+ "learning_rate": 0.0006,
+ "loss": 4.207851886749268,
+ "step": 2501
+ },
+ {
+ "epoch": 34.75458515283843,
+ "grad_norm": 0.015355597250163555,
+ "learning_rate": 0.0006,
+ "loss": 4.169469833374023,
+ "step": 2502
+ },
+ {
+ "epoch": 34.76855895196506,
+ "grad_norm": 0.015974000096321106,
+ "learning_rate": 0.0006,
+ "loss": 4.110061168670654,
+ "step": 2503
+ },
+ {
+ "epoch": 34.7825327510917,
+ "grad_norm": 0.016088927164673805,
+ "learning_rate": 0.0006,
+ "loss": 4.155452728271484,
+ "step": 2504
+ },
+ {
+ "epoch": 34.79650655021834,
+ "grad_norm": 0.016156120225787163,
+ "learning_rate": 0.0006,
+ "loss": 4.178092956542969,
+ "step": 2505
+ },
+ {
+ "epoch": 34.81048034934498,
+ "grad_norm": 0.014928647316992283,
+ "learning_rate": 0.0006,
+ "loss": 4.05457878112793,
+ "step": 2506
+ },
+ {
+ "epoch": 34.82445414847162,
+ "grad_norm": 0.015289328061044216,
+ "learning_rate": 0.0006,
+ "loss": 4.18869686126709,
+ "step": 2507
+ },
+ {
+ "epoch": 34.838427947598255,
+ "grad_norm": 0.015799039974808693,
+ "learning_rate": 0.0006,
+ "loss": 4.106073379516602,
+ "step": 2508
+ },
+ {
+ "epoch": 34.852401746724894,
+ "grad_norm": 0.015824418514966965,
+ "learning_rate": 0.0006,
+ "loss": 4.0634589195251465,
+ "step": 2509
+ },
+ {
+ "epoch": 34.866375545851525,
+ "grad_norm": 0.015321656130254269,
+ "learning_rate": 0.0006,
+ "loss": 4.093090057373047,
+ "step": 2510
+ },
+ {
+ "epoch": 34.880349344978164,
+ "grad_norm": 0.014380166307091713,
+ "learning_rate": 0.0006,
+ "loss": 4.1750640869140625,
+ "step": 2511
+ },
+ {
+ "epoch": 34.8943231441048,
+ "grad_norm": 0.014962567947804928,
+ "learning_rate": 0.0006,
+ "loss": 4.185277938842773,
+ "step": 2512
+ },
+ {
+ "epoch": 34.90829694323144,
+ "grad_norm": 0.014467543922364712,
+ "learning_rate": 0.0006,
+ "loss": 4.1600799560546875,
+ "step": 2513
+ },
+ {
+ "epoch": 34.92227074235808,
+ "grad_norm": 0.01545181218534708,
+ "learning_rate": 0.0006,
+ "loss": 4.096299648284912,
+ "step": 2514
+ },
+ {
+ "epoch": 34.93624454148472,
+ "grad_norm": 0.01605340465903282,
+ "learning_rate": 0.0006,
+ "loss": 4.203323841094971,
+ "step": 2515
+ },
+ {
+ "epoch": 34.95021834061135,
+ "grad_norm": 0.015203815884888172,
+ "learning_rate": 0.0006,
+ "loss": 4.201385021209717,
+ "step": 2516
+ },
+ {
+ "epoch": 34.96419213973799,
+ "grad_norm": 0.015149409882724285,
+ "learning_rate": 0.0006,
+ "loss": 4.088164329528809,
+ "step": 2517
+ },
+ {
+ "epoch": 34.97816593886463,
+ "grad_norm": 0.014885558746755123,
+ "learning_rate": 0.0006,
+ "loss": 4.073911666870117,
+ "step": 2518
+ },
+ {
+ "epoch": 34.992139737991266,
+ "grad_norm": 0.015003419481217861,
+ "learning_rate": 0.0006,
+ "loss": 4.157113552093506,
+ "step": 2519
+ },
+ {
+ "epoch": 35.0,
+ "grad_norm": 0.01641916297376156,
+ "learning_rate": 0.0006,
+ "loss": 3.970052719116211,
+ "step": 2520
+ },
+ {
+ "epoch": 35.0,
+ "eval_loss": 4.641692161560059,
+ "eval_runtime": 56.8836,
+ "eval_samples_per_second": 42.93,
+ "eval_steps_per_second": 1.354,
+ "step": 2520
+ },
+ {
+ "epoch": 35.01397379912664,
+ "grad_norm": 0.015894673764705658,
+ "learning_rate": 0.0006,
+ "loss": 4.104437351226807,
+ "step": 2521
+ },
+ {
+ "epoch": 35.02794759825328,
+ "grad_norm": 0.0173867866396904,
+ "learning_rate": 0.0006,
+ "loss": 4.091726303100586,
+ "step": 2522
+ },
+ {
+ "epoch": 35.041921397379916,
+ "grad_norm": 0.019272832199931145,
+ "learning_rate": 0.0006,
+ "loss": 4.217905044555664,
+ "step": 2523
+ },
+ {
+ "epoch": 35.05589519650655,
+ "grad_norm": 0.021820371970534325,
+ "learning_rate": 0.0006,
+ "loss": 4.111428260803223,
+ "step": 2524
+ },
+ {
+ "epoch": 35.069868995633186,
+ "grad_norm": 0.02340741828083992,
+ "learning_rate": 0.0006,
+ "loss": 4.115771293640137,
+ "step": 2525
+ },
+ {
+ "epoch": 35.083842794759825,
+ "grad_norm": 0.023760691285133362,
+ "learning_rate": 0.0006,
+ "loss": 4.132383346557617,
+ "step": 2526
+ },
+ {
+ "epoch": 35.09781659388646,
+ "grad_norm": 0.02073468267917633,
+ "learning_rate": 0.0006,
+ "loss": 4.272691249847412,
+ "step": 2527
+ },
+ {
+ "epoch": 35.1117903930131,
+ "grad_norm": 0.02197718247771263,
+ "learning_rate": 0.0006,
+ "loss": 4.105586528778076,
+ "step": 2528
+ },
+ {
+ "epoch": 35.12576419213974,
+ "grad_norm": 0.0200254675000906,
+ "learning_rate": 0.0006,
+ "loss": 3.968820095062256,
+ "step": 2529
+ },
+ {
+ "epoch": 35.13973799126637,
+ "grad_norm": 0.01929587684571743,
+ "learning_rate": 0.0006,
+ "loss": 4.045468807220459,
+ "step": 2530
+ },
+ {
+ "epoch": 35.15371179039301,
+ "grad_norm": 0.01886162906885147,
+ "learning_rate": 0.0006,
+ "loss": 4.185571670532227,
+ "step": 2531
+ },
+ {
+ "epoch": 35.16768558951965,
+ "grad_norm": 0.019789839163422585,
+ "learning_rate": 0.0006,
+ "loss": 3.9734551906585693,
+ "step": 2532
+ },
+ {
+ "epoch": 35.18165938864629,
+ "grad_norm": 0.0198142658919096,
+ "learning_rate": 0.0006,
+ "loss": 4.2662506103515625,
+ "step": 2533
+ },
+ {
+ "epoch": 35.19563318777293,
+ "grad_norm": 0.02087012305855751,
+ "learning_rate": 0.0006,
+ "loss": 4.175931930541992,
+ "step": 2534
+ },
+ {
+ "epoch": 35.209606986899566,
+ "grad_norm": 0.020294535905122757,
+ "learning_rate": 0.0006,
+ "loss": 4.060324668884277,
+ "step": 2535
+ },
+ {
+ "epoch": 35.223580786026204,
+ "grad_norm": 0.01908983662724495,
+ "learning_rate": 0.0006,
+ "loss": 4.217939853668213,
+ "step": 2536
+ },
+ {
+ "epoch": 35.237554585152836,
+ "grad_norm": 0.0178892333060503,
+ "learning_rate": 0.0006,
+ "loss": 4.190216541290283,
+ "step": 2537
+ },
+ {
+ "epoch": 35.251528384279474,
+ "grad_norm": 0.017892315983772278,
+ "learning_rate": 0.0006,
+ "loss": 4.067419052124023,
+ "step": 2538
+ },
+ {
+ "epoch": 35.26550218340611,
+ "grad_norm": 0.018556250259280205,
+ "learning_rate": 0.0006,
+ "loss": 4.105958461761475,
+ "step": 2539
+ },
+ {
+ "epoch": 35.27947598253275,
+ "grad_norm": 0.01944422349333763,
+ "learning_rate": 0.0006,
+ "loss": 4.155775547027588,
+ "step": 2540
+ },
+ {
+ "epoch": 35.29344978165939,
+ "grad_norm": 0.017333146184682846,
+ "learning_rate": 0.0006,
+ "loss": 4.244513511657715,
+ "step": 2541
+ },
+ {
+ "epoch": 35.30742358078603,
+ "grad_norm": 0.01639244332909584,
+ "learning_rate": 0.0006,
+ "loss": 4.167815685272217,
+ "step": 2542
+ },
+ {
+ "epoch": 35.32139737991266,
+ "grad_norm": 0.01610143668949604,
+ "learning_rate": 0.0006,
+ "loss": 4.121486663818359,
+ "step": 2543
+ },
+ {
+ "epoch": 35.3353711790393,
+ "grad_norm": 0.01656411588191986,
+ "learning_rate": 0.0006,
+ "loss": 4.1755499839782715,
+ "step": 2544
+ },
+ {
+ "epoch": 35.34934497816594,
+ "grad_norm": 0.015075297094881535,
+ "learning_rate": 0.0006,
+ "loss": 4.102992534637451,
+ "step": 2545
+ },
+ {
+ "epoch": 35.36331877729258,
+ "grad_norm": 0.01538170501589775,
+ "learning_rate": 0.0006,
+ "loss": 4.161603927612305,
+ "step": 2546
+ },
+ {
+ "epoch": 35.377292576419215,
+ "grad_norm": 0.01483312901109457,
+ "learning_rate": 0.0006,
+ "loss": 4.085355758666992,
+ "step": 2547
+ },
+ {
+ "epoch": 35.391266375545854,
+ "grad_norm": 0.014894738793373108,
+ "learning_rate": 0.0006,
+ "loss": 4.083585262298584,
+ "step": 2548
+ },
+ {
+ "epoch": 35.40524017467249,
+ "grad_norm": 0.01475780550390482,
+ "learning_rate": 0.0006,
+ "loss": 4.012645721435547,
+ "step": 2549
+ },
+ {
+ "epoch": 35.419213973799124,
+ "grad_norm": 0.014531653374433517,
+ "learning_rate": 0.0006,
+ "loss": 4.197113990783691,
+ "step": 2550
+ },
+ {
+ "epoch": 35.43318777292576,
+ "grad_norm": 0.015316493809223175,
+ "learning_rate": 0.0006,
+ "loss": 4.227970123291016,
+ "step": 2551
+ },
+ {
+ "epoch": 35.4471615720524,
+ "grad_norm": 0.014685571193695068,
+ "learning_rate": 0.0006,
+ "loss": 4.089415550231934,
+ "step": 2552
+ },
+ {
+ "epoch": 35.46113537117904,
+ "grad_norm": 0.015337035991251469,
+ "learning_rate": 0.0006,
+ "loss": 4.080583572387695,
+ "step": 2553
+ },
+ {
+ "epoch": 35.47510917030568,
+ "grad_norm": 0.016754208132624626,
+ "learning_rate": 0.0006,
+ "loss": 4.163935661315918,
+ "step": 2554
+ },
+ {
+ "epoch": 35.48908296943232,
+ "grad_norm": 0.017535798251628876,
+ "learning_rate": 0.0006,
+ "loss": 4.093027591705322,
+ "step": 2555
+ },
+ {
+ "epoch": 35.50305676855895,
+ "grad_norm": 0.017336975783109665,
+ "learning_rate": 0.0006,
+ "loss": 4.127961158752441,
+ "step": 2556
+ },
+ {
+ "epoch": 35.51703056768559,
+ "grad_norm": 0.016242152079939842,
+ "learning_rate": 0.0006,
+ "loss": 4.155593395233154,
+ "step": 2557
+ },
+ {
+ "epoch": 35.531004366812226,
+ "grad_norm": 0.015121322125196457,
+ "learning_rate": 0.0006,
+ "loss": 3.984004020690918,
+ "step": 2558
+ },
+ {
+ "epoch": 35.544978165938865,
+ "grad_norm": 0.015178884379565716,
+ "learning_rate": 0.0006,
+ "loss": 4.111564636230469,
+ "step": 2559
+ },
+ {
+ "epoch": 35.5589519650655,
+ "grad_norm": 0.014777131378650665,
+ "learning_rate": 0.0006,
+ "loss": 4.041402816772461,
+ "step": 2560
+ },
+ {
+ "epoch": 35.57292576419214,
+ "grad_norm": 0.015197676606476307,
+ "learning_rate": 0.0006,
+ "loss": 4.133706092834473,
+ "step": 2561
+ },
+ {
+ "epoch": 35.58689956331878,
+ "grad_norm": 0.015905972570180893,
+ "learning_rate": 0.0006,
+ "loss": 4.1069135665893555,
+ "step": 2562
+ },
+ {
+ "epoch": 35.60087336244541,
+ "grad_norm": 0.015942877158522606,
+ "learning_rate": 0.0006,
+ "loss": 4.0720624923706055,
+ "step": 2563
+ },
+ {
+ "epoch": 35.61484716157205,
+ "grad_norm": 0.016697583720088005,
+ "learning_rate": 0.0006,
+ "loss": 4.167448043823242,
+ "step": 2564
+ },
+ {
+ "epoch": 35.62882096069869,
+ "grad_norm": 0.015603577718138695,
+ "learning_rate": 0.0006,
+ "loss": 4.127327919006348,
+ "step": 2565
+ },
+ {
+ "epoch": 35.64279475982533,
+ "grad_norm": 0.017153460532426834,
+ "learning_rate": 0.0006,
+ "loss": 4.056543350219727,
+ "step": 2566
+ },
+ {
+ "epoch": 35.65676855895197,
+ "grad_norm": 0.016731958836317062,
+ "learning_rate": 0.0006,
+ "loss": 4.0681843757629395,
+ "step": 2567
+ },
+ {
+ "epoch": 35.670742358078606,
+ "grad_norm": 0.016167359426617622,
+ "learning_rate": 0.0006,
+ "loss": 4.174992084503174,
+ "step": 2568
+ },
+ {
+ "epoch": 35.68471615720524,
+ "grad_norm": 0.01615108922123909,
+ "learning_rate": 0.0006,
+ "loss": 4.1383161544799805,
+ "step": 2569
+ },
+ {
+ "epoch": 35.698689956331876,
+ "grad_norm": 0.015750816091895103,
+ "learning_rate": 0.0006,
+ "loss": 4.031595706939697,
+ "step": 2570
+ },
+ {
+ "epoch": 35.712663755458514,
+ "grad_norm": 0.015416189096868038,
+ "learning_rate": 0.0006,
+ "loss": 4.067498207092285,
+ "step": 2571
+ },
+ {
+ "epoch": 35.72663755458515,
+ "grad_norm": 0.01570792682468891,
+ "learning_rate": 0.0006,
+ "loss": 4.1338300704956055,
+ "step": 2572
+ },
+ {
+ "epoch": 35.74061135371179,
+ "grad_norm": 0.016367752104997635,
+ "learning_rate": 0.0006,
+ "loss": 4.152523040771484,
+ "step": 2573
+ },
+ {
+ "epoch": 35.75458515283843,
+ "grad_norm": 0.016349803656339645,
+ "learning_rate": 0.0006,
+ "loss": 4.101985454559326,
+ "step": 2574
+ },
+ {
+ "epoch": 35.76855895196506,
+ "grad_norm": 0.01663660630583763,
+ "learning_rate": 0.0006,
+ "loss": 4.190130710601807,
+ "step": 2575
+ },
+ {
+ "epoch": 35.7825327510917,
+ "grad_norm": 0.015599170699715614,
+ "learning_rate": 0.0006,
+ "loss": 4.0792436599731445,
+ "step": 2576
+ },
+ {
+ "epoch": 35.79650655021834,
+ "grad_norm": 0.015472427010536194,
+ "learning_rate": 0.0006,
+ "loss": 4.186446666717529,
+ "step": 2577
+ },
+ {
+ "epoch": 35.81048034934498,
+ "grad_norm": 0.015520439483225346,
+ "learning_rate": 0.0006,
+ "loss": 4.246635437011719,
+ "step": 2578
+ },
+ {
+ "epoch": 35.82445414847162,
+ "grad_norm": 0.016998281702399254,
+ "learning_rate": 0.0006,
+ "loss": 4.13435173034668,
+ "step": 2579
+ },
+ {
+ "epoch": 35.838427947598255,
+ "grad_norm": 0.016290361061692238,
+ "learning_rate": 0.0006,
+ "loss": 4.107428073883057,
+ "step": 2580
+ },
+ {
+ "epoch": 35.852401746724894,
+ "grad_norm": 0.0165516659617424,
+ "learning_rate": 0.0006,
+ "loss": 4.1612138748168945,
+ "step": 2581
+ },
+ {
+ "epoch": 35.866375545851525,
+ "grad_norm": 0.01736091636121273,
+ "learning_rate": 0.0006,
+ "loss": 4.1072869300842285,
+ "step": 2582
+ },
+ {
+ "epoch": 35.880349344978164,
+ "grad_norm": 0.01787031814455986,
+ "learning_rate": 0.0006,
+ "loss": 4.1076812744140625,
+ "step": 2583
+ },
+ {
+ "epoch": 35.8943231441048,
+ "grad_norm": 0.015437985770404339,
+ "learning_rate": 0.0006,
+ "loss": 4.097488880157471,
+ "step": 2584
+ },
+ {
+ "epoch": 35.90829694323144,
+ "grad_norm": 0.016306741163134575,
+ "learning_rate": 0.0006,
+ "loss": 4.105017185211182,
+ "step": 2585
+ },
+ {
+ "epoch": 35.92227074235808,
+ "grad_norm": 0.015723245218396187,
+ "learning_rate": 0.0006,
+ "loss": 4.205283164978027,
+ "step": 2586
+ },
+ {
+ "epoch": 35.93624454148472,
+ "grad_norm": 0.014626456424593925,
+ "learning_rate": 0.0006,
+ "loss": 4.086552619934082,
+ "step": 2587
+ },
+ {
+ "epoch": 35.95021834061135,
+ "grad_norm": 0.014348454773426056,
+ "learning_rate": 0.0006,
+ "loss": 4.171329975128174,
+ "step": 2588
+ },
+ {
+ "epoch": 35.96419213973799,
+ "grad_norm": 0.015598480589687824,
+ "learning_rate": 0.0006,
+ "loss": 4.140815258026123,
+ "step": 2589
+ },
+ {
+ "epoch": 35.97816593886463,
+ "grad_norm": 0.017126062884926796,
+ "learning_rate": 0.0006,
+ "loss": 4.157176971435547,
+ "step": 2590
+ },
+ {
+ "epoch": 35.992139737991266,
+ "grad_norm": 0.01860199309885502,
+ "learning_rate": 0.0006,
+ "loss": 4.252880096435547,
+ "step": 2591
+ },
+ {
+ "epoch": 36.0,
+ "grad_norm": 0.021780220791697502,
+ "learning_rate": 0.0006,
+ "loss": 4.249861717224121,
+ "step": 2592
+ },
+ {
+ "epoch": 36.0,
+ "eval_loss": 4.606466293334961,
+ "eval_runtime": 56.6217,
+ "eval_samples_per_second": 43.128,
+ "eval_steps_per_second": 1.36,
+ "step": 2592
+ },
+ {
+ "epoch": 36.01397379912664,
+ "grad_norm": 0.020892828702926636,
+ "learning_rate": 0.0006,
+ "loss": 4.102055549621582,
+ "step": 2593
+ },
+ {
+ "epoch": 36.02794759825328,
+ "grad_norm": 0.020306255668401718,
+ "learning_rate": 0.0006,
+ "loss": 4.184308052062988,
+ "step": 2594
+ },
+ {
+ "epoch": 36.041921397379916,
+ "grad_norm": 0.02025097794830799,
+ "learning_rate": 0.0006,
+ "loss": 3.9691858291625977,
+ "step": 2595
+ },
+ {
+ "epoch": 36.05589519650655,
+ "grad_norm": 0.020254382863640785,
+ "learning_rate": 0.0006,
+ "loss": 4.02957820892334,
+ "step": 2596
+ },
+ {
+ "epoch": 36.069868995633186,
+ "grad_norm": 0.019307196140289307,
+ "learning_rate": 0.0006,
+ "loss": 4.033181667327881,
+ "step": 2597
+ },
+ {
+ "epoch": 36.083842794759825,
+ "grad_norm": 0.0198476854711771,
+ "learning_rate": 0.0006,
+ "loss": 4.0816874504089355,
+ "step": 2598
+ },
+ {
+ "epoch": 36.09781659388646,
+ "grad_norm": 0.019916515797376633,
+ "learning_rate": 0.0006,
+ "loss": 4.164889812469482,
+ "step": 2599
+ },
+ {
+ "epoch": 36.1117903930131,
+ "grad_norm": 0.0189144778996706,
+ "learning_rate": 0.0006,
+ "loss": 4.038427352905273,
+ "step": 2600
+ },
+ {
+ "epoch": 36.12576419213974,
+ "grad_norm": 0.01754768192768097,
+ "learning_rate": 0.0006,
+ "loss": 4.010427474975586,
+ "step": 2601
+ },
+ {
+ "epoch": 36.13973799126637,
+ "grad_norm": 0.016862990334630013,
+ "learning_rate": 0.0006,
+ "loss": 4.035970211029053,
+ "step": 2602
+ },
+ {
+ "epoch": 36.15371179039301,
+ "grad_norm": 0.019616752862930298,
+ "learning_rate": 0.0006,
+ "loss": 4.079898834228516,
+ "step": 2603
+ },
+ {
+ "epoch": 36.16768558951965,
+ "grad_norm": 0.020131465047597885,
+ "learning_rate": 0.0006,
+ "loss": 4.144684791564941,
+ "step": 2604
+ },
+ {
+ "epoch": 36.18165938864629,
+ "grad_norm": 0.02072521299123764,
+ "learning_rate": 0.0006,
+ "loss": 4.013660430908203,
+ "step": 2605
+ },
+ {
+ "epoch": 36.19563318777293,
+ "grad_norm": 0.01924615167081356,
+ "learning_rate": 0.0006,
+ "loss": 4.071398735046387,
+ "step": 2606
+ },
+ {
+ "epoch": 36.209606986899566,
+ "grad_norm": 0.02013954520225525,
+ "learning_rate": 0.0006,
+ "loss": 4.127094268798828,
+ "step": 2607
+ },
+ {
+ "epoch": 36.223580786026204,
+ "grad_norm": 0.02042008936405182,
+ "learning_rate": 0.0006,
+ "loss": 4.154261589050293,
+ "step": 2608
+ },
+ {
+ "epoch": 36.237554585152836,
+ "grad_norm": 0.021998876705765724,
+ "learning_rate": 0.0006,
+ "loss": 4.204150199890137,
+ "step": 2609
+ },
+ {
+ "epoch": 36.251528384279474,
+ "grad_norm": 0.022946683689951897,
+ "learning_rate": 0.0006,
+ "loss": 4.013632774353027,
+ "step": 2610
+ },
+ {
+ "epoch": 36.26550218340611,
+ "grad_norm": 0.02061830274760723,
+ "learning_rate": 0.0006,
+ "loss": 4.171084403991699,
+ "step": 2611
+ },
+ {
+ "epoch": 36.27947598253275,
+ "grad_norm": 0.01883152686059475,
+ "learning_rate": 0.0006,
+ "loss": 4.149848461151123,
+ "step": 2612
+ },
+ {
+ "epoch": 36.29344978165939,
+ "grad_norm": 0.019945282489061356,
+ "learning_rate": 0.0006,
+ "loss": 4.174656867980957,
+ "step": 2613
+ },
+ {
+ "epoch": 36.30742358078603,
+ "grad_norm": 0.020751051604747772,
+ "learning_rate": 0.0006,
+ "loss": 4.080747127532959,
+ "step": 2614
+ },
+ {
+ "epoch": 36.32139737991266,
+ "grad_norm": 0.01879691146314144,
+ "learning_rate": 0.0006,
+ "loss": 4.057356357574463,
+ "step": 2615
+ },
+ {
+ "epoch": 36.3353711790393,
+ "grad_norm": 0.019037071615457535,
+ "learning_rate": 0.0006,
+ "loss": 4.204607009887695,
+ "step": 2616
+ },
+ {
+ "epoch": 36.34934497816594,
+ "grad_norm": 0.01872078701853752,
+ "learning_rate": 0.0006,
+ "loss": 4.050586700439453,
+ "step": 2617
+ },
+ {
+ "epoch": 36.36331877729258,
+ "grad_norm": 0.018092811107635498,
+ "learning_rate": 0.0006,
+ "loss": 4.102832794189453,
+ "step": 2618
+ },
+ {
+ "epoch": 36.377292576419215,
+ "grad_norm": 0.017010487616062164,
+ "learning_rate": 0.0006,
+ "loss": 4.191788673400879,
+ "step": 2619
+ },
+ {
+ "epoch": 36.391266375545854,
+ "grad_norm": 0.01691143773496151,
+ "learning_rate": 0.0006,
+ "loss": 4.160456657409668,
+ "step": 2620
+ },
+ {
+ "epoch": 36.40524017467249,
+ "grad_norm": 0.017046816647052765,
+ "learning_rate": 0.0006,
+ "loss": 4.073828220367432,
+ "step": 2621
+ },
+ {
+ "epoch": 36.419213973799124,
+ "grad_norm": 0.01707981713116169,
+ "learning_rate": 0.0006,
+ "loss": 4.0980095863342285,
+ "step": 2622
+ },
+ {
+ "epoch": 36.43318777292576,
+ "grad_norm": 0.01630452275276184,
+ "learning_rate": 0.0006,
+ "loss": 4.101321220397949,
+ "step": 2623
+ },
+ {
+ "epoch": 36.4471615720524,
+ "grad_norm": 0.01576915942132473,
+ "learning_rate": 0.0006,
+ "loss": 4.158249855041504,
+ "step": 2624
+ },
+ {
+ "epoch": 36.46113537117904,
+ "grad_norm": 0.016179712489247322,
+ "learning_rate": 0.0006,
+ "loss": 4.077852249145508,
+ "step": 2625
+ },
+ {
+ "epoch": 36.47510917030568,
+ "grad_norm": 0.01609169878065586,
+ "learning_rate": 0.0006,
+ "loss": 4.187317848205566,
+ "step": 2626
+ },
+ {
+ "epoch": 36.48908296943232,
+ "grad_norm": 0.017148254439234734,
+ "learning_rate": 0.0006,
+ "loss": 4.109071731567383,
+ "step": 2627
+ },
+ {
+ "epoch": 36.50305676855895,
+ "grad_norm": 0.015842624008655548,
+ "learning_rate": 0.0006,
+ "loss": 4.145096302032471,
+ "step": 2628
+ },
+ {
+ "epoch": 36.51703056768559,
+ "grad_norm": 0.014906194061040878,
+ "learning_rate": 0.0006,
+ "loss": 4.028440475463867,
+ "step": 2629
+ },
+ {
+ "epoch": 36.531004366812226,
+ "grad_norm": 0.01485038734972477,
+ "learning_rate": 0.0006,
+ "loss": 4.175201416015625,
+ "step": 2630
+ },
+ {
+ "epoch": 36.544978165938865,
+ "grad_norm": 0.015834150835871696,
+ "learning_rate": 0.0006,
+ "loss": 4.076434135437012,
+ "step": 2631
+ },
+ {
+ "epoch": 36.5589519650655,
+ "grad_norm": 0.015411358326673508,
+ "learning_rate": 0.0006,
+ "loss": 4.03912878036499,
+ "step": 2632
+ },
+ {
+ "epoch": 36.57292576419214,
+ "grad_norm": 0.014762846753001213,
+ "learning_rate": 0.0006,
+ "loss": 4.019827365875244,
+ "step": 2633
+ },
+ {
+ "epoch": 36.58689956331878,
+ "grad_norm": 0.015269497409462929,
+ "learning_rate": 0.0006,
+ "loss": 4.116957187652588,
+ "step": 2634
+ },
+ {
+ "epoch": 36.60087336244541,
+ "grad_norm": 0.015171617269515991,
+ "learning_rate": 0.0006,
+ "loss": 4.2069926261901855,
+ "step": 2635
+ },
+ {
+ "epoch": 36.61484716157205,
+ "grad_norm": 0.014770902693271637,
+ "learning_rate": 0.0006,
+ "loss": 4.092021942138672,
+ "step": 2636
+ },
+ {
+ "epoch": 36.62882096069869,
+ "grad_norm": 0.014957531355321407,
+ "learning_rate": 0.0006,
+ "loss": 4.16819953918457,
+ "step": 2637
+ },
+ {
+ "epoch": 36.64279475982533,
+ "grad_norm": 0.014774451963603497,
+ "learning_rate": 0.0006,
+ "loss": 4.046073913574219,
+ "step": 2638
+ },
+ {
+ "epoch": 36.65676855895197,
+ "grad_norm": 0.014320445246994495,
+ "learning_rate": 0.0006,
+ "loss": 3.977792501449585,
+ "step": 2639
+ },
+ {
+ "epoch": 36.670742358078606,
+ "grad_norm": 0.01413130946457386,
+ "learning_rate": 0.0006,
+ "loss": 4.130649566650391,
+ "step": 2640
+ },
+ {
+ "epoch": 36.68471615720524,
+ "grad_norm": 0.014336307533085346,
+ "learning_rate": 0.0006,
+ "loss": 4.161357402801514,
+ "step": 2641
+ },
+ {
+ "epoch": 36.698689956331876,
+ "grad_norm": 0.014516931027173996,
+ "learning_rate": 0.0006,
+ "loss": 4.14248514175415,
+ "step": 2642
+ },
+ {
+ "epoch": 36.712663755458514,
+ "grad_norm": 0.014377126470208168,
+ "learning_rate": 0.0006,
+ "loss": 4.062711715698242,
+ "step": 2643
+ },
+ {
+ "epoch": 36.72663755458515,
+ "grad_norm": 0.014332668855786324,
+ "learning_rate": 0.0006,
+ "loss": 4.06423807144165,
+ "step": 2644
+ },
+ {
+ "epoch": 36.74061135371179,
+ "grad_norm": 0.014030571095645428,
+ "learning_rate": 0.0006,
+ "loss": 4.178613662719727,
+ "step": 2645
+ },
+ {
+ "epoch": 36.75458515283843,
+ "grad_norm": 0.014376776292920113,
+ "learning_rate": 0.0006,
+ "loss": 4.073130130767822,
+ "step": 2646
+ },
+ {
+ "epoch": 36.76855895196506,
+ "grad_norm": 0.013827716931700706,
+ "learning_rate": 0.0006,
+ "loss": 4.043869972229004,
+ "step": 2647
+ },
+ {
+ "epoch": 36.7825327510917,
+ "grad_norm": 0.015548762865364552,
+ "learning_rate": 0.0006,
+ "loss": 4.134049415588379,
+ "step": 2648
+ },
+ {
+ "epoch": 36.79650655021834,
+ "grad_norm": 0.017041970044374466,
+ "learning_rate": 0.0006,
+ "loss": 4.205024242401123,
+ "step": 2649
+ },
+ {
+ "epoch": 36.81048034934498,
+ "grad_norm": 0.017638275399804115,
+ "learning_rate": 0.0006,
+ "loss": 4.199924468994141,
+ "step": 2650
+ },
+ {
+ "epoch": 36.82445414847162,
+ "grad_norm": 0.017771627753973007,
+ "learning_rate": 0.0006,
+ "loss": 4.15706729888916,
+ "step": 2651
+ },
+ {
+ "epoch": 36.838427947598255,
+ "grad_norm": 0.015933562070131302,
+ "learning_rate": 0.0006,
+ "loss": 4.220742225646973,
+ "step": 2652
+ },
+ {
+ "epoch": 36.852401746724894,
+ "grad_norm": 0.01668144389986992,
+ "learning_rate": 0.0006,
+ "loss": 4.161426544189453,
+ "step": 2653
+ },
+ {
+ "epoch": 36.866375545851525,
+ "grad_norm": 0.018014680594205856,
+ "learning_rate": 0.0006,
+ "loss": 4.183046817779541,
+ "step": 2654
+ },
+ {
+ "epoch": 36.880349344978164,
+ "grad_norm": 0.018808670341968536,
+ "learning_rate": 0.0006,
+ "loss": 4.105293273925781,
+ "step": 2655
+ },
+ {
+ "epoch": 36.8943231441048,
+ "grad_norm": 0.018003476783633232,
+ "learning_rate": 0.0006,
+ "loss": 4.103171348571777,
+ "step": 2656
+ },
+ {
+ "epoch": 36.90829694323144,
+ "grad_norm": 0.016865914687514305,
+ "learning_rate": 0.0006,
+ "loss": 4.125634670257568,
+ "step": 2657
+ },
+ {
+ "epoch": 36.92227074235808,
+ "grad_norm": 0.018371153622865677,
+ "learning_rate": 0.0006,
+ "loss": 4.192461013793945,
+ "step": 2658
+ },
+ {
+ "epoch": 36.93624454148472,
+ "grad_norm": 0.019045377150177956,
+ "learning_rate": 0.0006,
+ "loss": 4.198105812072754,
+ "step": 2659
+ },
+ {
+ "epoch": 36.95021834061135,
+ "grad_norm": 0.01594168320298195,
+ "learning_rate": 0.0006,
+ "loss": 4.169788360595703,
+ "step": 2660
+ },
+ {
+ "epoch": 36.96419213973799,
+ "grad_norm": 0.015645820647478104,
+ "learning_rate": 0.0006,
+ "loss": 4.047277450561523,
+ "step": 2661
+ },
+ {
+ "epoch": 36.97816593886463,
+ "grad_norm": 0.016313156113028526,
+ "learning_rate": 0.0006,
+ "loss": 4.087392807006836,
+ "step": 2662
+ },
+ {
+ "epoch": 36.992139737991266,
+ "grad_norm": 0.017050625756382942,
+ "learning_rate": 0.0006,
+ "loss": 4.0908002853393555,
+ "step": 2663
+ },
+ {
+ "epoch": 37.0,
+ "grad_norm": 0.017730968073010445,
+ "learning_rate": 0.0006,
+ "loss": 4.110957622528076,
+ "step": 2664
+ },
+ {
+ "epoch": 37.0,
+ "eval_loss": 4.563793182373047,
+ "eval_runtime": 56.5922,
+ "eval_samples_per_second": 43.151,
+ "eval_steps_per_second": 1.361,
+ "step": 2664
+ },
+ {
+ "epoch": 37.01397379912664,
+ "grad_norm": 0.015687910839915276,
+ "learning_rate": 0.0006,
+ "loss": 4.143589019775391,
+ "step": 2665
+ },
+ {
+ "epoch": 37.02794759825328,
+ "grad_norm": 0.017451366409659386,
+ "learning_rate": 0.0006,
+ "loss": 4.0326666831970215,
+ "step": 2666
+ },
+ {
+ "epoch": 37.041921397379916,
+ "grad_norm": 0.01909538544714451,
+ "learning_rate": 0.0006,
+ "loss": 3.9706249237060547,
+ "step": 2667
+ },
+ {
+ "epoch": 37.05589519650655,
+ "grad_norm": 0.01988379657268524,
+ "learning_rate": 0.0006,
+ "loss": 3.9531009197235107,
+ "step": 2668
+ },
+ {
+ "epoch": 37.069868995633186,
+ "grad_norm": 0.02035115472972393,
+ "learning_rate": 0.0006,
+ "loss": 3.962214469909668,
+ "step": 2669
+ },
+ {
+ "epoch": 37.083842794759825,
+ "grad_norm": 0.01952001079916954,
+ "learning_rate": 0.0006,
+ "loss": 4.182262897491455,
+ "step": 2670
+ },
+ {
+ "epoch": 37.09781659388646,
+ "grad_norm": 0.01854325644671917,
+ "learning_rate": 0.0006,
+ "loss": 4.127663612365723,
+ "step": 2671
+ },
+ {
+ "epoch": 37.1117903930131,
+ "grad_norm": 0.017757616937160492,
+ "learning_rate": 0.0006,
+ "loss": 4.100804805755615,
+ "step": 2672
+ },
+ {
+ "epoch": 37.12576419213974,
+ "grad_norm": 0.016712479293346405,
+ "learning_rate": 0.0006,
+ "loss": 4.026284217834473,
+ "step": 2673
+ },
+ {
+ "epoch": 37.13973799126637,
+ "grad_norm": 0.0168139711022377,
+ "learning_rate": 0.0006,
+ "loss": 4.183882713317871,
+ "step": 2674
+ },
+ {
+ "epoch": 37.15371179039301,
+ "grad_norm": 0.017060423269867897,
+ "learning_rate": 0.0006,
+ "loss": 4.131832599639893,
+ "step": 2675
+ },
+ {
+ "epoch": 37.16768558951965,
+ "grad_norm": 0.01726631447672844,
+ "learning_rate": 0.0006,
+ "loss": 3.9156057834625244,
+ "step": 2676
+ },
+ {
+ "epoch": 37.18165938864629,
+ "grad_norm": 0.016348622739315033,
+ "learning_rate": 0.0006,
+ "loss": 4.1731367111206055,
+ "step": 2677
+ },
+ {
+ "epoch": 37.19563318777293,
+ "grad_norm": 0.016640059649944305,
+ "learning_rate": 0.0006,
+ "loss": 4.10997200012207,
+ "step": 2678
+ },
+ {
+ "epoch": 37.209606986899566,
+ "grad_norm": 0.01601342298090458,
+ "learning_rate": 0.0006,
+ "loss": 4.081145286560059,
+ "step": 2679
+ },
+ {
+ "epoch": 37.223580786026204,
+ "grad_norm": 0.01735631190240383,
+ "learning_rate": 0.0006,
+ "loss": 4.202108860015869,
+ "step": 2680
+ },
+ {
+ "epoch": 37.237554585152836,
+ "grad_norm": 0.019164685159921646,
+ "learning_rate": 0.0006,
+ "loss": 4.104109764099121,
+ "step": 2681
+ },
+ {
+ "epoch": 37.251528384279474,
+ "grad_norm": 0.01881759613752365,
+ "learning_rate": 0.0006,
+ "loss": 4.145532131195068,
+ "step": 2682
+ },
+ {
+ "epoch": 37.26550218340611,
+ "grad_norm": 0.017656555399298668,
+ "learning_rate": 0.0006,
+ "loss": 4.012782096862793,
+ "step": 2683
+ },
+ {
+ "epoch": 37.27947598253275,
+ "grad_norm": 0.01624823734164238,
+ "learning_rate": 0.0006,
+ "loss": 4.001727104187012,
+ "step": 2684
+ },
+ {
+ "epoch": 37.29344978165939,
+ "grad_norm": 0.016973139718174934,
+ "learning_rate": 0.0006,
+ "loss": 4.095399856567383,
+ "step": 2685
+ },
+ {
+ "epoch": 37.30742358078603,
+ "grad_norm": 0.015503671951591969,
+ "learning_rate": 0.0006,
+ "loss": 4.034431457519531,
+ "step": 2686
+ },
+ {
+ "epoch": 37.32139737991266,
+ "grad_norm": 0.015370250679552555,
+ "learning_rate": 0.0006,
+ "loss": 4.075836181640625,
+ "step": 2687
+ },
+ {
+ "epoch": 37.3353711790393,
+ "grad_norm": 0.01564021222293377,
+ "learning_rate": 0.0006,
+ "loss": 4.125202178955078,
+ "step": 2688
+ },
+ {
+ "epoch": 37.34934497816594,
+ "grad_norm": 0.016717975959181786,
+ "learning_rate": 0.0006,
+ "loss": 4.153017044067383,
+ "step": 2689
+ },
+ {
+ "epoch": 37.36331877729258,
+ "grad_norm": 0.015263617038726807,
+ "learning_rate": 0.0006,
+ "loss": 4.021814823150635,
+ "step": 2690
+ },
+ {
+ "epoch": 37.377292576419215,
+ "grad_norm": 0.016497142612934113,
+ "learning_rate": 0.0006,
+ "loss": 4.069095611572266,
+ "step": 2691
+ },
+ {
+ "epoch": 37.391266375545854,
+ "grad_norm": 0.01669355481863022,
+ "learning_rate": 0.0006,
+ "loss": 4.174587726593018,
+ "step": 2692
+ },
+ {
+ "epoch": 37.40524017467249,
+ "grad_norm": 0.019695773720741272,
+ "learning_rate": 0.0006,
+ "loss": 4.0073041915893555,
+ "step": 2693
+ },
+ {
+ "epoch": 37.419213973799124,
+ "grad_norm": 0.023720644414424896,
+ "learning_rate": 0.0006,
+ "loss": 4.111948013305664,
+ "step": 2694
+ },
+ {
+ "epoch": 37.43318777292576,
+ "grad_norm": 0.02399350143969059,
+ "learning_rate": 0.0006,
+ "loss": 4.081164360046387,
+ "step": 2695
+ },
+ {
+ "epoch": 37.4471615720524,
+ "grad_norm": 0.019630271941423416,
+ "learning_rate": 0.0006,
+ "loss": 4.065731048583984,
+ "step": 2696
+ },
+ {
+ "epoch": 37.46113537117904,
+ "grad_norm": 0.019679008051753044,
+ "learning_rate": 0.0006,
+ "loss": 3.993852376937866,
+ "step": 2697
+ },
+ {
+ "epoch": 37.47510917030568,
+ "grad_norm": 0.020597299560904503,
+ "learning_rate": 0.0006,
+ "loss": 4.275555610656738,
+ "step": 2698
+ },
+ {
+ "epoch": 37.48908296943232,
+ "grad_norm": 0.01983889564871788,
+ "learning_rate": 0.0006,
+ "loss": 4.254708290100098,
+ "step": 2699
+ },
+ {
+ "epoch": 37.50305676855895,
+ "grad_norm": 0.019150668755173683,
+ "learning_rate": 0.0006,
+ "loss": 4.097885608673096,
+ "step": 2700
+ },
+ {
+ "epoch": 37.51703056768559,
+ "grad_norm": 0.018847715109586716,
+ "learning_rate": 0.0006,
+ "loss": 4.054081439971924,
+ "step": 2701
+ },
+ {
+ "epoch": 37.531004366812226,
+ "grad_norm": 0.019607288762927055,
+ "learning_rate": 0.0006,
+ "loss": 3.9891440868377686,
+ "step": 2702
+ },
+ {
+ "epoch": 37.544978165938865,
+ "grad_norm": 0.018863150849938393,
+ "learning_rate": 0.0006,
+ "loss": 4.084877014160156,
+ "step": 2703
+ },
+ {
+ "epoch": 37.5589519650655,
+ "grad_norm": 0.017340239137411118,
+ "learning_rate": 0.0006,
+ "loss": 4.113290786743164,
+ "step": 2704
+ },
+ {
+ "epoch": 37.57292576419214,
+ "grad_norm": 0.015628693625330925,
+ "learning_rate": 0.0006,
+ "loss": 4.1483001708984375,
+ "step": 2705
+ },
+ {
+ "epoch": 37.58689956331878,
+ "grad_norm": 0.016493283212184906,
+ "learning_rate": 0.0006,
+ "loss": 4.1040143966674805,
+ "step": 2706
+ },
+ {
+ "epoch": 37.60087336244541,
+ "grad_norm": 0.016732443124055862,
+ "learning_rate": 0.0006,
+ "loss": 4.07205057144165,
+ "step": 2707
+ },
+ {
+ "epoch": 37.61484716157205,
+ "grad_norm": 0.017176156863570213,
+ "learning_rate": 0.0006,
+ "loss": 4.1279754638671875,
+ "step": 2708
+ },
+ {
+ "epoch": 37.62882096069869,
+ "grad_norm": 0.015730608254671097,
+ "learning_rate": 0.0006,
+ "loss": 4.086658000946045,
+ "step": 2709
+ },
+ {
+ "epoch": 37.64279475982533,
+ "grad_norm": 0.016111090779304504,
+ "learning_rate": 0.0006,
+ "loss": 4.064335823059082,
+ "step": 2710
+ },
+ {
+ "epoch": 37.65676855895197,
+ "grad_norm": 0.01769379712641239,
+ "learning_rate": 0.0006,
+ "loss": 4.116047382354736,
+ "step": 2711
+ },
+ {
+ "epoch": 37.670742358078606,
+ "grad_norm": 0.01683102548122406,
+ "learning_rate": 0.0006,
+ "loss": 4.095986843109131,
+ "step": 2712
+ },
+ {
+ "epoch": 37.68471615720524,
+ "grad_norm": 0.015726450830698013,
+ "learning_rate": 0.0006,
+ "loss": 4.10775089263916,
+ "step": 2713
+ },
+ {
+ "epoch": 37.698689956331876,
+ "grad_norm": 0.014760235324501991,
+ "learning_rate": 0.0006,
+ "loss": 4.1234893798828125,
+ "step": 2714
+ },
+ {
+ "epoch": 37.712663755458514,
+ "grad_norm": 0.014933438040316105,
+ "learning_rate": 0.0006,
+ "loss": 4.079200744628906,
+ "step": 2715
+ },
+ {
+ "epoch": 37.72663755458515,
+ "grad_norm": 0.015723997727036476,
+ "learning_rate": 0.0006,
+ "loss": 4.07865047454834,
+ "step": 2716
+ },
+ {
+ "epoch": 37.74061135371179,
+ "grad_norm": 0.014539001509547234,
+ "learning_rate": 0.0006,
+ "loss": 4.036242485046387,
+ "step": 2717
+ },
+ {
+ "epoch": 37.75458515283843,
+ "grad_norm": 0.014844979159533978,
+ "learning_rate": 0.0006,
+ "loss": 4.1051483154296875,
+ "step": 2718
+ },
+ {
+ "epoch": 37.76855895196506,
+ "grad_norm": 0.015939844772219658,
+ "learning_rate": 0.0006,
+ "loss": 4.017253875732422,
+ "step": 2719
+ },
+ {
+ "epoch": 37.7825327510917,
+ "grad_norm": 0.014771555550396442,
+ "learning_rate": 0.0006,
+ "loss": 4.0694661140441895,
+ "step": 2720
+ },
+ {
+ "epoch": 37.79650655021834,
+ "grad_norm": 0.015777921304106712,
+ "learning_rate": 0.0006,
+ "loss": 3.947587490081787,
+ "step": 2721
+ },
+ {
+ "epoch": 37.81048034934498,
+ "grad_norm": 0.017451291903853416,
+ "learning_rate": 0.0006,
+ "loss": 4.093625068664551,
+ "step": 2722
+ },
+ {
+ "epoch": 37.82445414847162,
+ "grad_norm": 0.016829246655106544,
+ "learning_rate": 0.0006,
+ "loss": 4.09030818939209,
+ "step": 2723
+ },
+ {
+ "epoch": 37.838427947598255,
+ "grad_norm": 0.016984397545456886,
+ "learning_rate": 0.0006,
+ "loss": 4.140626907348633,
+ "step": 2724
+ },
+ {
+ "epoch": 37.852401746724894,
+ "grad_norm": 0.01732603646814823,
+ "learning_rate": 0.0006,
+ "loss": 4.128276824951172,
+ "step": 2725
+ },
+ {
+ "epoch": 37.866375545851525,
+ "grad_norm": 0.01744619384407997,
+ "learning_rate": 0.0006,
+ "loss": 4.103130340576172,
+ "step": 2726
+ },
+ {
+ "epoch": 37.880349344978164,
+ "grad_norm": 0.016331320628523827,
+ "learning_rate": 0.0006,
+ "loss": 4.037951469421387,
+ "step": 2727
+ },
+ {
+ "epoch": 37.8943231441048,
+ "grad_norm": 0.015919430181384087,
+ "learning_rate": 0.0006,
+ "loss": 4.121513366699219,
+ "step": 2728
+ },
+ {
+ "epoch": 37.90829694323144,
+ "grad_norm": 0.016153695061802864,
+ "learning_rate": 0.0006,
+ "loss": 4.099384307861328,
+ "step": 2729
+ },
+ {
+ "epoch": 37.92227074235808,
+ "grad_norm": 0.01628800667822361,
+ "learning_rate": 0.0006,
+ "loss": 4.092325687408447,
+ "step": 2730
+ },
+ {
+ "epoch": 37.93624454148472,
+ "grad_norm": 0.017285684123635292,
+ "learning_rate": 0.0006,
+ "loss": 3.9828906059265137,
+ "step": 2731
+ },
+ {
+ "epoch": 37.95021834061135,
+ "grad_norm": 0.016047554090619087,
+ "learning_rate": 0.0006,
+ "loss": 4.100419044494629,
+ "step": 2732
+ },
+ {
+ "epoch": 37.96419213973799,
+ "grad_norm": 0.01562986709177494,
+ "learning_rate": 0.0006,
+ "loss": 4.065832614898682,
+ "step": 2733
+ },
+ {
+ "epoch": 37.97816593886463,
+ "grad_norm": 0.01677163504064083,
+ "learning_rate": 0.0006,
+ "loss": 4.077518463134766,
+ "step": 2734
+ },
+ {
+ "epoch": 37.992139737991266,
+ "grad_norm": 0.01755492016673088,
+ "learning_rate": 0.0006,
+ "loss": 4.1076154708862305,
+ "step": 2735
+ },
+ {
+ "epoch": 38.0,
+ "grad_norm": 0.01929762214422226,
+ "learning_rate": 0.0006,
+ "loss": 4.09603214263916,
+ "step": 2736
+ },
+ {
+ "epoch": 38.0,
+ "eval_loss": 4.582442283630371,
+ "eval_runtime": 62.8649,
+ "eval_samples_per_second": 38.845,
+ "eval_steps_per_second": 1.225,
+ "step": 2736
+ },
+ {
+ "epoch": 38.01397379912664,
+ "grad_norm": 0.019129136577248573,
+ "learning_rate": 0.0006,
+ "loss": 4.095005989074707,
+ "step": 2737
+ },
+ {
+ "epoch": 38.02794759825328,
+ "grad_norm": 0.019167408347129822,
+ "learning_rate": 0.0006,
+ "loss": 4.054426670074463,
+ "step": 2738
+ },
+ {
+ "epoch": 38.041921397379916,
+ "grad_norm": 0.01789460889995098,
+ "learning_rate": 0.0006,
+ "loss": 4.0434417724609375,
+ "step": 2739
+ },
+ {
+ "epoch": 38.05589519650655,
+ "grad_norm": 0.015947027131915092,
+ "learning_rate": 0.0006,
+ "loss": 4.074516296386719,
+ "step": 2740
+ },
+ {
+ "epoch": 38.069868995633186,
+ "grad_norm": 0.016667252406477928,
+ "learning_rate": 0.0006,
+ "loss": 4.108545303344727,
+ "step": 2741
+ },
+ {
+ "epoch": 38.083842794759825,
+ "grad_norm": 0.016014399006962776,
+ "learning_rate": 0.0006,
+ "loss": 4.026925086975098,
+ "step": 2742
+ },
+ {
+ "epoch": 38.09781659388646,
+ "grad_norm": 0.015397695824503899,
+ "learning_rate": 0.0006,
+ "loss": 4.163970470428467,
+ "step": 2743
+ },
+ {
+ "epoch": 38.1117903930131,
+ "grad_norm": 0.01563289947807789,
+ "learning_rate": 0.0006,
+ "loss": 4.116578578948975,
+ "step": 2744
+ },
+ {
+ "epoch": 38.12576419213974,
+ "grad_norm": 0.015512318350374699,
+ "learning_rate": 0.0006,
+ "loss": 4.096311569213867,
+ "step": 2745
+ },
+ {
+ "epoch": 38.13973799126637,
+ "grad_norm": 0.015426642261445522,
+ "learning_rate": 0.0006,
+ "loss": 4.017065048217773,
+ "step": 2746
+ },
+ {
+ "epoch": 38.15371179039301,
+ "grad_norm": 0.015115435235202312,
+ "learning_rate": 0.0006,
+ "loss": 4.005434989929199,
+ "step": 2747
+ },
+ {
+ "epoch": 38.16768558951965,
+ "grad_norm": 0.015235894359648228,
+ "learning_rate": 0.0006,
+ "loss": 3.982743263244629,
+ "step": 2748
+ },
+ {
+ "epoch": 38.18165938864629,
+ "grad_norm": 0.016095101833343506,
+ "learning_rate": 0.0006,
+ "loss": 4.0849385261535645,
+ "step": 2749
+ },
+ {
+ "epoch": 38.19563318777293,
+ "grad_norm": 0.015201197937130928,
+ "learning_rate": 0.0006,
+ "loss": 4.103863716125488,
+ "step": 2750
+ },
+ {
+ "epoch": 38.209606986899566,
+ "grad_norm": 0.015687478706240654,
+ "learning_rate": 0.0006,
+ "loss": 4.12074089050293,
+ "step": 2751
+ },
+ {
+ "epoch": 38.223580786026204,
+ "grad_norm": 0.015201589092612267,
+ "learning_rate": 0.0006,
+ "loss": 4.056872367858887,
+ "step": 2752
+ },
+ {
+ "epoch": 38.237554585152836,
+ "grad_norm": 0.015154894441366196,
+ "learning_rate": 0.0006,
+ "loss": 4.089670181274414,
+ "step": 2753
+ },
+ {
+ "epoch": 38.251528384279474,
+ "grad_norm": 0.01586608588695526,
+ "learning_rate": 0.0006,
+ "loss": 4.155317306518555,
+ "step": 2754
+ },
+ {
+ "epoch": 38.26550218340611,
+ "grad_norm": 0.01609654724597931,
+ "learning_rate": 0.0006,
+ "loss": 3.995800495147705,
+ "step": 2755
+ },
+ {
+ "epoch": 38.27947598253275,
+ "grad_norm": 0.015495185740292072,
+ "learning_rate": 0.0006,
+ "loss": 3.995769739151001,
+ "step": 2756
+ },
+ {
+ "epoch": 38.29344978165939,
+ "grad_norm": 0.017042381688952446,
+ "learning_rate": 0.0006,
+ "loss": 4.132832050323486,
+ "step": 2757
+ },
+ {
+ "epoch": 38.30742358078603,
+ "grad_norm": 0.018706614151597023,
+ "learning_rate": 0.0006,
+ "loss": 4.038820743560791,
+ "step": 2758
+ },
+ {
+ "epoch": 38.32139737991266,
+ "grad_norm": 0.018316026777029037,
+ "learning_rate": 0.0006,
+ "loss": 3.972052574157715,
+ "step": 2759
+ },
+ {
+ "epoch": 38.3353711790393,
+ "grad_norm": 0.018375389277935028,
+ "learning_rate": 0.0006,
+ "loss": 4.152097702026367,
+ "step": 2760
+ },
+ {
+ "epoch": 38.34934497816594,
+ "grad_norm": 0.016879886388778687,
+ "learning_rate": 0.0006,
+ "loss": 4.020869731903076,
+ "step": 2761
+ },
+ {
+ "epoch": 38.36331877729258,
+ "grad_norm": 0.015728183090686798,
+ "learning_rate": 0.0006,
+ "loss": 4.008016586303711,
+ "step": 2762
+ },
+ {
+ "epoch": 38.377292576419215,
+ "grad_norm": 0.016147758811712265,
+ "learning_rate": 0.0006,
+ "loss": 4.035854339599609,
+ "step": 2763
+ },
+ {
+ "epoch": 38.391266375545854,
+ "grad_norm": 0.017774797976017,
+ "learning_rate": 0.0006,
+ "loss": 4.122506141662598,
+ "step": 2764
+ },
+ {
+ "epoch": 38.40524017467249,
+ "grad_norm": 0.016590069979429245,
+ "learning_rate": 0.0006,
+ "loss": 4.082847595214844,
+ "step": 2765
+ },
+ {
+ "epoch": 38.419213973799124,
+ "grad_norm": 0.016544528305530548,
+ "learning_rate": 0.0006,
+ "loss": 4.184270858764648,
+ "step": 2766
+ },
+ {
+ "epoch": 38.43318777292576,
+ "grad_norm": 0.017972925677895546,
+ "learning_rate": 0.0006,
+ "loss": 3.987558603286743,
+ "step": 2767
+ },
+ {
+ "epoch": 38.4471615720524,
+ "grad_norm": 0.018562227487564087,
+ "learning_rate": 0.0006,
+ "loss": 4.1193461418151855,
+ "step": 2768
+ },
+ {
+ "epoch": 38.46113537117904,
+ "grad_norm": 0.018896685913205147,
+ "learning_rate": 0.0006,
+ "loss": 4.057808876037598,
+ "step": 2769
+ },
+ {
+ "epoch": 38.47510917030568,
+ "grad_norm": 0.016793809831142426,
+ "learning_rate": 0.0006,
+ "loss": 4.069523334503174,
+ "step": 2770
+ },
+ {
+ "epoch": 38.48908296943232,
+ "grad_norm": 0.01698676310479641,
+ "learning_rate": 0.0006,
+ "loss": 3.997624158859253,
+ "step": 2771
+ },
+ {
+ "epoch": 38.50305676855895,
+ "grad_norm": 0.016784202307462692,
+ "learning_rate": 0.0006,
+ "loss": 4.174795627593994,
+ "step": 2772
+ },
+ {
+ "epoch": 38.51703056768559,
+ "grad_norm": 0.016785014420747757,
+ "learning_rate": 0.0006,
+ "loss": 3.999321222305298,
+ "step": 2773
+ },
+ {
+ "epoch": 38.531004366812226,
+ "grad_norm": 0.018615184351801872,
+ "learning_rate": 0.0006,
+ "loss": 4.191237926483154,
+ "step": 2774
+ },
+ {
+ "epoch": 38.544978165938865,
+ "grad_norm": 0.019090790301561356,
+ "learning_rate": 0.0006,
+ "loss": 4.106966018676758,
+ "step": 2775
+ },
+ {
+ "epoch": 38.5589519650655,
+ "grad_norm": 0.018298320472240448,
+ "learning_rate": 0.0006,
+ "loss": 4.063363552093506,
+ "step": 2776
+ },
+ {
+ "epoch": 38.57292576419214,
+ "grad_norm": 0.018541090190410614,
+ "learning_rate": 0.0006,
+ "loss": 4.101193428039551,
+ "step": 2777
+ },
+ {
+ "epoch": 38.58689956331878,
+ "grad_norm": 0.01849970407783985,
+ "learning_rate": 0.0006,
+ "loss": 4.071335792541504,
+ "step": 2778
+ },
+ {
+ "epoch": 38.60087336244541,
+ "grad_norm": 0.018309077247977257,
+ "learning_rate": 0.0006,
+ "loss": 4.245305061340332,
+ "step": 2779
+ },
+ {
+ "epoch": 38.61484716157205,
+ "grad_norm": 0.017681840807199478,
+ "learning_rate": 0.0006,
+ "loss": 4.081173419952393,
+ "step": 2780
+ },
+ {
+ "epoch": 38.62882096069869,
+ "grad_norm": 0.016286224126815796,
+ "learning_rate": 0.0006,
+ "loss": 4.024351596832275,
+ "step": 2781
+ },
+ {
+ "epoch": 38.64279475982533,
+ "grad_norm": 0.015837235376238823,
+ "learning_rate": 0.0006,
+ "loss": 4.028205394744873,
+ "step": 2782
+ },
+ {
+ "epoch": 38.65676855895197,
+ "grad_norm": 0.016062596812844276,
+ "learning_rate": 0.0006,
+ "loss": 4.131122589111328,
+ "step": 2783
+ },
+ {
+ "epoch": 38.670742358078606,
+ "grad_norm": 0.01711239665746689,
+ "learning_rate": 0.0006,
+ "loss": 4.124416351318359,
+ "step": 2784
+ },
+ {
+ "epoch": 38.68471615720524,
+ "grad_norm": 0.015304913744330406,
+ "learning_rate": 0.0006,
+ "loss": 4.056122779846191,
+ "step": 2785
+ },
+ {
+ "epoch": 38.698689956331876,
+ "grad_norm": 0.015971887856721878,
+ "learning_rate": 0.0006,
+ "loss": 4.132555961608887,
+ "step": 2786
+ },
+ {
+ "epoch": 38.712663755458514,
+ "grad_norm": 0.015787221491336823,
+ "learning_rate": 0.0006,
+ "loss": 4.051596164703369,
+ "step": 2787
+ },
+ {
+ "epoch": 38.72663755458515,
+ "grad_norm": 0.01605444774031639,
+ "learning_rate": 0.0006,
+ "loss": 4.191808700561523,
+ "step": 2788
+ },
+ {
+ "epoch": 38.74061135371179,
+ "grad_norm": 0.016383009031414986,
+ "learning_rate": 0.0006,
+ "loss": 4.0648417472839355,
+ "step": 2789
+ },
+ {
+ "epoch": 38.75458515283843,
+ "grad_norm": 0.016135554760694504,
+ "learning_rate": 0.0006,
+ "loss": 3.988563060760498,
+ "step": 2790
+ },
+ {
+ "epoch": 38.76855895196506,
+ "grad_norm": 0.016634158790111542,
+ "learning_rate": 0.0006,
+ "loss": 4.133493423461914,
+ "step": 2791
+ },
+ {
+ "epoch": 38.7825327510917,
+ "grad_norm": 0.01688210666179657,
+ "learning_rate": 0.0006,
+ "loss": 4.091277122497559,
+ "step": 2792
+ },
+ {
+ "epoch": 38.79650655021834,
+ "grad_norm": 0.016102122142910957,
+ "learning_rate": 0.0006,
+ "loss": 4.051219940185547,
+ "step": 2793
+ },
+ {
+ "epoch": 38.81048034934498,
+ "grad_norm": 0.016585038974881172,
+ "learning_rate": 0.0006,
+ "loss": 4.058587074279785,
+ "step": 2794
+ },
+ {
+ "epoch": 38.82445414847162,
+ "grad_norm": 0.017700951546430588,
+ "learning_rate": 0.0006,
+ "loss": 4.016030311584473,
+ "step": 2795
+ },
+ {
+ "epoch": 38.838427947598255,
+ "grad_norm": 0.017922203987836838,
+ "learning_rate": 0.0006,
+ "loss": 4.079977989196777,
+ "step": 2796
+ },
+ {
+ "epoch": 38.852401746724894,
+ "grad_norm": 0.017299890518188477,
+ "learning_rate": 0.0006,
+ "loss": 4.062984943389893,
+ "step": 2797
+ },
+ {
+ "epoch": 38.866375545851525,
+ "grad_norm": 0.018648581579327583,
+ "learning_rate": 0.0006,
+ "loss": 4.065717697143555,
+ "step": 2798
+ },
+ {
+ "epoch": 38.880349344978164,
+ "grad_norm": 0.019057920202612877,
+ "learning_rate": 0.0006,
+ "loss": 4.227892875671387,
+ "step": 2799
+ },
+ {
+ "epoch": 38.8943231441048,
+ "grad_norm": 0.018824422731995583,
+ "learning_rate": 0.0006,
+ "loss": 4.116058349609375,
+ "step": 2800
+ },
+ {
+ "epoch": 38.90829694323144,
+ "grad_norm": 0.019084978848695755,
+ "learning_rate": 0.0006,
+ "loss": 4.061677932739258,
+ "step": 2801
+ },
+ {
+ "epoch": 38.92227074235808,
+ "grad_norm": 0.017842981964349747,
+ "learning_rate": 0.0006,
+ "loss": 4.083813667297363,
+ "step": 2802
+ },
+ {
+ "epoch": 38.93624454148472,
+ "grad_norm": 0.01880858838558197,
+ "learning_rate": 0.0006,
+ "loss": 4.152561187744141,
+ "step": 2803
+ },
+ {
+ "epoch": 38.95021834061135,
+ "grad_norm": 0.017400087788701057,
+ "learning_rate": 0.0006,
+ "loss": 4.154635429382324,
+ "step": 2804
+ },
+ {
+ "epoch": 38.96419213973799,
+ "grad_norm": 0.017682507634162903,
+ "learning_rate": 0.0006,
+ "loss": 4.151128768920898,
+ "step": 2805
+ },
+ {
+ "epoch": 38.97816593886463,
+ "grad_norm": 0.015803052112460136,
+ "learning_rate": 0.0006,
+ "loss": 4.10498571395874,
+ "step": 2806
+ },
+ {
+ "epoch": 38.992139737991266,
+ "grad_norm": 0.01582511141896248,
+ "learning_rate": 0.0006,
+ "loss": 4.140159606933594,
+ "step": 2807
+ },
+ {
+ "epoch": 39.0,
+ "grad_norm": 0.01925666816532612,
+ "learning_rate": 0.0006,
+ "loss": 4.126394271850586,
+ "step": 2808
+ },
+ {
+ "epoch": 39.0,
+ "eval_loss": 4.615617275238037,
+ "eval_runtime": 56.8895,
+ "eval_samples_per_second": 42.925,
+ "eval_steps_per_second": 1.354,
+ "step": 2808
+ },
+ {
+ "epoch": 39.01397379912664,
+ "grad_norm": 0.01677853614091873,
+ "learning_rate": 0.0006,
+ "loss": 4.102395057678223,
+ "step": 2809
+ },
+ {
+ "epoch": 39.02794759825328,
+ "grad_norm": 0.017901137471199036,
+ "learning_rate": 0.0006,
+ "loss": 3.976609230041504,
+ "step": 2810
+ },
+ {
+ "epoch": 39.041921397379916,
+ "grad_norm": 0.017978668212890625,
+ "learning_rate": 0.0006,
+ "loss": 3.944416046142578,
+ "step": 2811
+ },
+ {
+ "epoch": 39.05589519650655,
+ "grad_norm": 0.019967492669820786,
+ "learning_rate": 0.0006,
+ "loss": 3.992234468460083,
+ "step": 2812
+ },
+ {
+ "epoch": 39.069868995633186,
+ "grad_norm": 0.018937544897198677,
+ "learning_rate": 0.0006,
+ "loss": 4.028573036193848,
+ "step": 2813
+ },
+ {
+ "epoch": 39.083842794759825,
+ "grad_norm": 0.016852952539920807,
+ "learning_rate": 0.0006,
+ "loss": 4.062442779541016,
+ "step": 2814
+ },
+ {
+ "epoch": 39.09781659388646,
+ "grad_norm": 0.016643647104501724,
+ "learning_rate": 0.0006,
+ "loss": 4.08871603012085,
+ "step": 2815
+ },
+ {
+ "epoch": 39.1117903930131,
+ "grad_norm": 0.016872866079211235,
+ "learning_rate": 0.0006,
+ "loss": 3.9514245986938477,
+ "step": 2816
+ },
+ {
+ "epoch": 39.12576419213974,
+ "grad_norm": 0.018045669421553612,
+ "learning_rate": 0.0006,
+ "loss": 4.038842678070068,
+ "step": 2817
+ },
+ {
+ "epoch": 39.13973799126637,
+ "grad_norm": 0.017853403463959694,
+ "learning_rate": 0.0006,
+ "loss": 3.93709135055542,
+ "step": 2818
+ },
+ {
+ "epoch": 39.15371179039301,
+ "grad_norm": 0.017507443204522133,
+ "learning_rate": 0.0006,
+ "loss": 3.98665714263916,
+ "step": 2819
+ },
+ {
+ "epoch": 39.16768558951965,
+ "grad_norm": 0.01766853965818882,
+ "learning_rate": 0.0006,
+ "loss": 3.937126636505127,
+ "step": 2820
+ },
+ {
+ "epoch": 39.18165938864629,
+ "grad_norm": 0.016660021618008614,
+ "learning_rate": 0.0006,
+ "loss": 3.8397014141082764,
+ "step": 2821
+ },
+ {
+ "epoch": 39.19563318777293,
+ "grad_norm": 0.017719309777021408,
+ "learning_rate": 0.0006,
+ "loss": 4.003579616546631,
+ "step": 2822
+ },
+ {
+ "epoch": 39.209606986899566,
+ "grad_norm": 0.018234414979815483,
+ "learning_rate": 0.0006,
+ "loss": 4.053859710693359,
+ "step": 2823
+ },
+ {
+ "epoch": 39.223580786026204,
+ "grad_norm": 0.018777024000883102,
+ "learning_rate": 0.0006,
+ "loss": 4.056182861328125,
+ "step": 2824
+ },
+ {
+ "epoch": 39.237554585152836,
+ "grad_norm": 0.017443502321839333,
+ "learning_rate": 0.0006,
+ "loss": 4.025068283081055,
+ "step": 2825
+ },
+ {
+ "epoch": 39.251528384279474,
+ "grad_norm": 0.016500331461429596,
+ "learning_rate": 0.0006,
+ "loss": 4.101631164550781,
+ "step": 2826
+ },
+ {
+ "epoch": 39.26550218340611,
+ "grad_norm": 0.016101544722914696,
+ "learning_rate": 0.0006,
+ "loss": 4.027082443237305,
+ "step": 2827
+ },
+ {
+ "epoch": 39.27947598253275,
+ "grad_norm": 0.015471171587705612,
+ "learning_rate": 0.0006,
+ "loss": 4.022909641265869,
+ "step": 2828
+ },
+ {
+ "epoch": 39.29344978165939,
+ "grad_norm": 0.015779748558998108,
+ "learning_rate": 0.0006,
+ "loss": 4.18796968460083,
+ "step": 2829
+ },
+ {
+ "epoch": 39.30742358078603,
+ "grad_norm": 0.015522679314017296,
+ "learning_rate": 0.0006,
+ "loss": 4.005777359008789,
+ "step": 2830
+ },
+ {
+ "epoch": 39.32139737991266,
+ "grad_norm": 0.015226504765450954,
+ "learning_rate": 0.0006,
+ "loss": 4.108416557312012,
+ "step": 2831
+ },
+ {
+ "epoch": 39.3353711790393,
+ "grad_norm": 0.015095778740942478,
+ "learning_rate": 0.0006,
+ "loss": 4.08029842376709,
+ "step": 2832
+ },
+ {
+ "epoch": 39.34934497816594,
+ "grad_norm": 0.015601031482219696,
+ "learning_rate": 0.0006,
+ "loss": 3.989492416381836,
+ "step": 2833
+ },
+ {
+ "epoch": 39.36331877729258,
+ "grad_norm": 0.014825052581727505,
+ "learning_rate": 0.0006,
+ "loss": 3.998520612716675,
+ "step": 2834
+ },
+ {
+ "epoch": 39.377292576419215,
+ "grad_norm": 0.014821100980043411,
+ "learning_rate": 0.0006,
+ "loss": 4.001777172088623,
+ "step": 2835
+ },
+ {
+ "epoch": 39.391266375545854,
+ "grad_norm": 0.015333754941821098,
+ "learning_rate": 0.0006,
+ "loss": 4.11985969543457,
+ "step": 2836
+ },
+ {
+ "epoch": 39.40524017467249,
+ "grad_norm": 0.015940699726343155,
+ "learning_rate": 0.0006,
+ "loss": 4.025373935699463,
+ "step": 2837
+ },
+ {
+ "epoch": 39.419213973799124,
+ "grad_norm": 0.015096287243068218,
+ "learning_rate": 0.0006,
+ "loss": 4.068551063537598,
+ "step": 2838
+ },
+ {
+ "epoch": 39.43318777292576,
+ "grad_norm": 0.01533548254519701,
+ "learning_rate": 0.0006,
+ "loss": 3.9800801277160645,
+ "step": 2839
+ },
+ {
+ "epoch": 39.4471615720524,
+ "grad_norm": 0.014648732729256153,
+ "learning_rate": 0.0006,
+ "loss": 3.9670190811157227,
+ "step": 2840
+ },
+ {
+ "epoch": 39.46113537117904,
+ "grad_norm": 0.015759671106934547,
+ "learning_rate": 0.0006,
+ "loss": 4.050443649291992,
+ "step": 2841
+ },
+ {
+ "epoch": 39.47510917030568,
+ "grad_norm": 0.016054073348641396,
+ "learning_rate": 0.0006,
+ "loss": 4.22734260559082,
+ "step": 2842
+ },
+ {
+ "epoch": 39.48908296943232,
+ "grad_norm": 0.0160264540463686,
+ "learning_rate": 0.0006,
+ "loss": 4.116174221038818,
+ "step": 2843
+ },
+ {
+ "epoch": 39.50305676855895,
+ "grad_norm": 0.015426610596477985,
+ "learning_rate": 0.0006,
+ "loss": 4.000178337097168,
+ "step": 2844
+ },
+ {
+ "epoch": 39.51703056768559,
+ "grad_norm": 0.01520642638206482,
+ "learning_rate": 0.0006,
+ "loss": 4.218410968780518,
+ "step": 2845
+ },
+ {
+ "epoch": 39.531004366812226,
+ "grad_norm": 0.015305949375033379,
+ "learning_rate": 0.0006,
+ "loss": 4.025891304016113,
+ "step": 2846
+ },
+ {
+ "epoch": 39.544978165938865,
+ "grad_norm": 0.015461564064025879,
+ "learning_rate": 0.0006,
+ "loss": 4.100950717926025,
+ "step": 2847
+ },
+ {
+ "epoch": 39.5589519650655,
+ "grad_norm": 0.01712625101208687,
+ "learning_rate": 0.0006,
+ "loss": 4.0620856285095215,
+ "step": 2848
+ },
+ {
+ "epoch": 39.57292576419214,
+ "grad_norm": 0.017133301123976707,
+ "learning_rate": 0.0006,
+ "loss": 4.097138404846191,
+ "step": 2849
+ },
+ {
+ "epoch": 39.58689956331878,
+ "grad_norm": 0.01683625392615795,
+ "learning_rate": 0.0006,
+ "loss": 4.084364891052246,
+ "step": 2850
+ },
+ {
+ "epoch": 39.60087336244541,
+ "grad_norm": 0.0165710486471653,
+ "learning_rate": 0.0006,
+ "loss": 4.042361259460449,
+ "step": 2851
+ },
+ {
+ "epoch": 39.61484716157205,
+ "grad_norm": 0.01660393364727497,
+ "learning_rate": 0.0006,
+ "loss": 4.1395487785339355,
+ "step": 2852
+ },
+ {
+ "epoch": 39.62882096069869,
+ "grad_norm": 0.017025845125317574,
+ "learning_rate": 0.0006,
+ "loss": 4.1207075119018555,
+ "step": 2853
+ },
+ {
+ "epoch": 39.64279475982533,
+ "grad_norm": 0.01660497486591339,
+ "learning_rate": 0.0006,
+ "loss": 4.103846073150635,
+ "step": 2854
+ },
+ {
+ "epoch": 39.65676855895197,
+ "grad_norm": 0.01607946865260601,
+ "learning_rate": 0.0006,
+ "loss": 4.189016342163086,
+ "step": 2855
+ },
+ {
+ "epoch": 39.670742358078606,
+ "grad_norm": 0.01614505983889103,
+ "learning_rate": 0.0006,
+ "loss": 4.079860687255859,
+ "step": 2856
+ },
+ {
+ "epoch": 39.68471615720524,
+ "grad_norm": 0.017085086554288864,
+ "learning_rate": 0.0006,
+ "loss": 3.9774293899536133,
+ "step": 2857
+ },
+ {
+ "epoch": 39.698689956331876,
+ "grad_norm": 0.01816520094871521,
+ "learning_rate": 0.0006,
+ "loss": 4.113666534423828,
+ "step": 2858
+ },
+ {
+ "epoch": 39.712663755458514,
+ "grad_norm": 0.016597526147961617,
+ "learning_rate": 0.0006,
+ "loss": 4.068242073059082,
+ "step": 2859
+ },
+ {
+ "epoch": 39.72663755458515,
+ "grad_norm": 0.01672235131263733,
+ "learning_rate": 0.0006,
+ "loss": 4.024444580078125,
+ "step": 2860
+ },
+ {
+ "epoch": 39.74061135371179,
+ "grad_norm": 0.020093556493520737,
+ "learning_rate": 0.0006,
+ "loss": 4.064124584197998,
+ "step": 2861
+ },
+ {
+ "epoch": 39.75458515283843,
+ "grad_norm": 0.023141304031014442,
+ "learning_rate": 0.0006,
+ "loss": 4.130439758300781,
+ "step": 2862
+ },
+ {
+ "epoch": 39.76855895196506,
+ "grad_norm": 0.02289372682571411,
+ "learning_rate": 0.0006,
+ "loss": 4.14794921875,
+ "step": 2863
+ },
+ {
+ "epoch": 39.7825327510917,
+ "grad_norm": 0.01919020712375641,
+ "learning_rate": 0.0006,
+ "loss": 4.123961925506592,
+ "step": 2864
+ },
+ {
+ "epoch": 39.79650655021834,
+ "grad_norm": 0.016932329162955284,
+ "learning_rate": 0.0006,
+ "loss": 4.0944132804870605,
+ "step": 2865
+ },
+ {
+ "epoch": 39.81048034934498,
+ "grad_norm": 0.016644535586237907,
+ "learning_rate": 0.0006,
+ "loss": 4.124269962310791,
+ "step": 2866
+ },
+ {
+ "epoch": 39.82445414847162,
+ "grad_norm": 0.016302216798067093,
+ "learning_rate": 0.0006,
+ "loss": 4.0267205238342285,
+ "step": 2867
+ },
+ {
+ "epoch": 39.838427947598255,
+ "grad_norm": 0.01673300378024578,
+ "learning_rate": 0.0006,
+ "loss": 4.10650634765625,
+ "step": 2868
+ },
+ {
+ "epoch": 39.852401746724894,
+ "grad_norm": 0.016581356525421143,
+ "learning_rate": 0.0006,
+ "loss": 4.1547956466674805,
+ "step": 2869
+ },
+ {
+ "epoch": 39.866375545851525,
+ "grad_norm": 0.01691797934472561,
+ "learning_rate": 0.0006,
+ "loss": 4.056936740875244,
+ "step": 2870
+ },
+ {
+ "epoch": 39.880349344978164,
+ "grad_norm": 0.01596726104617119,
+ "learning_rate": 0.0006,
+ "loss": 3.9891414642333984,
+ "step": 2871
+ },
+ {
+ "epoch": 39.8943231441048,
+ "grad_norm": 0.015615091659128666,
+ "learning_rate": 0.0006,
+ "loss": 4.137948036193848,
+ "step": 2872
+ },
+ {
+ "epoch": 39.90829694323144,
+ "grad_norm": 0.015703413635492325,
+ "learning_rate": 0.0006,
+ "loss": 3.9844930171966553,
+ "step": 2873
+ },
+ {
+ "epoch": 39.92227074235808,
+ "grad_norm": 0.015453094616532326,
+ "learning_rate": 0.0006,
+ "loss": 4.098443508148193,
+ "step": 2874
+ },
+ {
+ "epoch": 39.93624454148472,
+ "grad_norm": 0.01449158601462841,
+ "learning_rate": 0.0006,
+ "loss": 4.090852737426758,
+ "step": 2875
+ },
+ {
+ "epoch": 39.95021834061135,
+ "grad_norm": 0.01537515688687563,
+ "learning_rate": 0.0006,
+ "loss": 4.239612579345703,
+ "step": 2876
+ },
+ {
+ "epoch": 39.96419213973799,
+ "grad_norm": 0.015989921987056732,
+ "learning_rate": 0.0006,
+ "loss": 4.142073154449463,
+ "step": 2877
+ },
+ {
+ "epoch": 39.97816593886463,
+ "grad_norm": 0.015985611826181412,
+ "learning_rate": 0.0006,
+ "loss": 4.100076675415039,
+ "step": 2878
+ },
+ {
+ "epoch": 39.992139737991266,
+ "grad_norm": 0.015260833315551281,
+ "learning_rate": 0.0006,
+ "loss": 4.068576335906982,
+ "step": 2879
+ },
+ {
+ "epoch": 40.0,
+ "grad_norm": 0.018047941848635674,
+ "learning_rate": 0.0006,
+ "loss": 4.005878925323486,
+ "step": 2880
+ },
+ {
+ "epoch": 40.0,
+ "eval_loss": 4.622848987579346,
+ "eval_runtime": 56.9618,
+ "eval_samples_per_second": 42.871,
+ "eval_steps_per_second": 1.352,
+ "step": 2880
+ },
+ {
+ "epoch": 40.01397379912664,
+ "grad_norm": 0.015976838767528534,
+ "learning_rate": 0.0006,
+ "loss": 4.035689353942871,
+ "step": 2881
+ },
+ {
+ "epoch": 40.02794759825328,
+ "grad_norm": 0.015176220797002316,
+ "learning_rate": 0.0006,
+ "loss": 3.932197332382202,
+ "step": 2882
+ },
+ {
+ "epoch": 40.041921397379916,
+ "grad_norm": 0.016676228493452072,
+ "learning_rate": 0.0006,
+ "loss": 4.014171123504639,
+ "step": 2883
+ },
+ {
+ "epoch": 40.05589519650655,
+ "grad_norm": 0.01698281429708004,
+ "learning_rate": 0.0006,
+ "loss": 4.117893218994141,
+ "step": 2884
+ },
+ {
+ "epoch": 40.069868995633186,
+ "grad_norm": 0.017872486263513565,
+ "learning_rate": 0.0006,
+ "loss": 4.132119178771973,
+ "step": 2885
+ },
+ {
+ "epoch": 40.083842794759825,
+ "grad_norm": 0.017615554854273796,
+ "learning_rate": 0.0006,
+ "loss": 4.072000980377197,
+ "step": 2886
+ },
+ {
+ "epoch": 40.09781659388646,
+ "grad_norm": 0.017469394952058792,
+ "learning_rate": 0.0006,
+ "loss": 4.017726421356201,
+ "step": 2887
+ },
+ {
+ "epoch": 40.1117903930131,
+ "grad_norm": 0.01954762637615204,
+ "learning_rate": 0.0006,
+ "loss": 4.014443874359131,
+ "step": 2888
+ },
+ {
+ "epoch": 40.12576419213974,
+ "grad_norm": 0.018935853615403175,
+ "learning_rate": 0.0006,
+ "loss": 4.0042948722839355,
+ "step": 2889
+ },
+ {
+ "epoch": 40.13973799126637,
+ "grad_norm": 0.019382692873477936,
+ "learning_rate": 0.0006,
+ "loss": 4.06160831451416,
+ "step": 2890
+ },
+ {
+ "epoch": 40.15371179039301,
+ "grad_norm": 0.019677115604281425,
+ "learning_rate": 0.0006,
+ "loss": 4.13466215133667,
+ "step": 2891
+ },
+ {
+ "epoch": 40.16768558951965,
+ "grad_norm": 0.020365122705698013,
+ "learning_rate": 0.0006,
+ "loss": 4.095000267028809,
+ "step": 2892
+ },
+ {
+ "epoch": 40.18165938864629,
+ "grad_norm": 0.021842138841748238,
+ "learning_rate": 0.0006,
+ "loss": 3.9877853393554688,
+ "step": 2893
+ },
+ {
+ "epoch": 40.19563318777293,
+ "grad_norm": 0.02354881353676319,
+ "learning_rate": 0.0006,
+ "loss": 3.9430112838745117,
+ "step": 2894
+ },
+ {
+ "epoch": 40.209606986899566,
+ "grad_norm": 0.026602735742926598,
+ "learning_rate": 0.0006,
+ "loss": 4.012076377868652,
+ "step": 2895
+ },
+ {
+ "epoch": 40.223580786026204,
+ "grad_norm": 0.023976484313607216,
+ "learning_rate": 0.0006,
+ "loss": 3.9768595695495605,
+ "step": 2896
+ },
+ {
+ "epoch": 40.237554585152836,
+ "grad_norm": 0.02336616814136505,
+ "learning_rate": 0.0006,
+ "loss": 4.1096038818359375,
+ "step": 2897
+ },
+ {
+ "epoch": 40.251528384279474,
+ "grad_norm": 0.02426023595035076,
+ "learning_rate": 0.0006,
+ "loss": 4.157189846038818,
+ "step": 2898
+ },
+ {
+ "epoch": 40.26550218340611,
+ "grad_norm": 0.021572696045041084,
+ "learning_rate": 0.0006,
+ "loss": 3.979175090789795,
+ "step": 2899
+ },
+ {
+ "epoch": 40.27947598253275,
+ "grad_norm": 0.08300432562828064,
+ "learning_rate": 0.0006,
+ "loss": 4.092336654663086,
+ "step": 2900
+ },
+ {
+ "epoch": 40.29344978165939,
+ "grad_norm": 0.7323967814445496,
+ "learning_rate": 0.0006,
+ "loss": 5.337990760803223,
+ "step": 2901
+ },
+ {
+ "epoch": 40.30742358078603,
+ "grad_norm": 0.3363337814807892,
+ "learning_rate": 0.0006,
+ "loss": 6.40043830871582,
+ "step": 2902
+ },
+ {
+ "epoch": 40.32139737991266,
+ "grad_norm": 0.598686933517456,
+ "learning_rate": 0.0006,
+ "loss": 7.004765510559082,
+ "step": 2903
+ },
+ {
+ "epoch": 40.3353711790393,
+ "grad_norm": 0.5998364686965942,
+ "learning_rate": 0.0006,
+ "loss": 8.184747695922852,
+ "step": 2904
+ },
+ {
+ "epoch": 40.34934497816594,
+ "grad_norm": 1.408208966255188,
+ "learning_rate": 0.0006,
+ "loss": 8.88978385925293,
+ "step": 2905
+ },
+ {
+ "epoch": 40.36331877729258,
+ "grad_norm": 0.7829785943031311,
+ "learning_rate": 0.0006,
+ "loss": 11.13601303100586,
+ "step": 2906
+ },
+ {
+ "epoch": 40.377292576419215,
+ "grad_norm": 0.7769215106964111,
+ "learning_rate": 0.0006,
+ "loss": 9.90416145324707,
+ "step": 2907
+ },
+ {
+ "epoch": 40.391266375545854,
+ "grad_norm": 0.46508079767227173,
+ "learning_rate": 0.0006,
+ "loss": 9.647209167480469,
+ "step": 2908
+ },
+ {
+ "epoch": 40.40524017467249,
+ "grad_norm": 1.0228968858718872,
+ "learning_rate": 0.0006,
+ "loss": 11.685434341430664,
+ "step": 2909
+ },
+ {
+ "epoch": 40.419213973799124,
+ "grad_norm": 0.20553778111934662,
+ "learning_rate": 0.0006,
+ "loss": 8.621708869934082,
+ "step": 2910
+ },
+ {
+ "epoch": 40.43318777292576,
+ "grad_norm": 0.18466424942016602,
+ "learning_rate": 0.0006,
+ "loss": 8.590787887573242,
+ "step": 2911
+ },
+ {
+ "epoch": 40.4471615720524,
+ "grad_norm": 0.10458594560623169,
+ "learning_rate": 0.0006,
+ "loss": 8.324458122253418,
+ "step": 2912
+ },
+ {
+ "epoch": 40.46113537117904,
+ "grad_norm": 0.1096842959523201,
+ "learning_rate": 0.0006,
+ "loss": 8.300851821899414,
+ "step": 2913
+ },
+ {
+ "epoch": 40.47510917030568,
+ "grad_norm": 0.10501226782798767,
+ "learning_rate": 0.0006,
+ "loss": 8.21009635925293,
+ "step": 2914
+ },
+ {
+ "epoch": 40.48908296943232,
+ "grad_norm": 0.13704165816307068,
+ "learning_rate": 0.0006,
+ "loss": 8.141355514526367,
+ "step": 2915
+ },
+ {
+ "epoch": 40.50305676855895,
+ "grad_norm": 0.12129108607769012,
+ "learning_rate": 0.0006,
+ "loss": 8.074612617492676,
+ "step": 2916
+ },
+ {
+ "epoch": 40.51703056768559,
+ "grad_norm": 0.10028897225856781,
+ "learning_rate": 0.0006,
+ "loss": 8.00384521484375,
+ "step": 2917
+ },
+ {
+ "epoch": 40.531004366812226,
+ "grad_norm": 0.11764246970415115,
+ "learning_rate": 0.0006,
+ "loss": 8.078598022460938,
+ "step": 2918
+ },
+ {
+ "epoch": 40.544978165938865,
+ "grad_norm": 0.10081959515810013,
+ "learning_rate": 0.0006,
+ "loss": 8.008938789367676,
+ "step": 2919
+ },
+ {
+ "epoch": 40.5589519650655,
+ "grad_norm": 0.1003170907497406,
+ "learning_rate": 0.0006,
+ "loss": 7.962940216064453,
+ "step": 2920
+ },
+ {
+ "epoch": 40.57292576419214,
+ "grad_norm": 0.08763492107391357,
+ "learning_rate": 0.0006,
+ "loss": 7.909419536590576,
+ "step": 2921
+ },
+ {
+ "epoch": 40.58689956331878,
+ "grad_norm": 0.0636807233095169,
+ "learning_rate": 0.0006,
+ "loss": 7.878905773162842,
+ "step": 2922
+ },
+ {
+ "epoch": 40.60087336244541,
+ "grad_norm": 0.06005936488509178,
+ "learning_rate": 0.0006,
+ "loss": 7.850980758666992,
+ "step": 2923
+ },
+ {
+ "epoch": 40.61484716157205,
+ "grad_norm": 0.06415296345949173,
+ "learning_rate": 0.0006,
+ "loss": 7.830353260040283,
+ "step": 2924
+ },
+ {
+ "epoch": 40.62882096069869,
+ "grad_norm": 0.0732007846236229,
+ "learning_rate": 0.0006,
+ "loss": 7.830746650695801,
+ "step": 2925
+ },
+ {
+ "epoch": 40.64279475982533,
+ "grad_norm": 0.06628110259771347,
+ "learning_rate": 0.0006,
+ "loss": 7.82661771774292,
+ "step": 2926
+ },
+ {
+ "epoch": 40.65676855895197,
+ "grad_norm": 0.05544218048453331,
+ "learning_rate": 0.0006,
+ "loss": 7.814305782318115,
+ "step": 2927
+ },
+ {
+ "epoch": 40.670742358078606,
+ "grad_norm": 0.050327103585004807,
+ "learning_rate": 0.0006,
+ "loss": 7.764055252075195,
+ "step": 2928
+ },
+ {
+ "epoch": 40.68471615720524,
+ "grad_norm": 0.04918425530195236,
+ "learning_rate": 0.0006,
+ "loss": 7.744625091552734,
+ "step": 2929
+ },
+ {
+ "epoch": 40.698689956331876,
+ "grad_norm": 0.042862266302108765,
+ "learning_rate": 0.0006,
+ "loss": 7.764426231384277,
+ "step": 2930
+ },
+ {
+ "epoch": 40.712663755458514,
+ "grad_norm": 0.04338089004158974,
+ "learning_rate": 0.0006,
+ "loss": 7.712530136108398,
+ "step": 2931
+ },
+ {
+ "epoch": 40.72663755458515,
+ "grad_norm": 0.04127737879753113,
+ "learning_rate": 0.0006,
+ "loss": 7.742180347442627,
+ "step": 2932
+ },
+ {
+ "epoch": 40.74061135371179,
+ "grad_norm": 0.03358631581068039,
+ "learning_rate": 0.0006,
+ "loss": 7.714834213256836,
+ "step": 2933
+ },
+ {
+ "epoch": 40.75458515283843,
+ "grad_norm": 0.02845790982246399,
+ "learning_rate": 0.0006,
+ "loss": 7.7021026611328125,
+ "step": 2934
+ },
+ {
+ "epoch": 40.76855895196506,
+ "grad_norm": 0.03012419492006302,
+ "learning_rate": 0.0006,
+ "loss": 7.712377548217773,
+ "step": 2935
+ },
+ {
+ "epoch": 40.7825327510917,
+ "grad_norm": 0.032981179654598236,
+ "learning_rate": 0.0006,
+ "loss": 7.666385650634766,
+ "step": 2936
+ },
+ {
+ "epoch": 40.79650655021834,
+ "grad_norm": 0.03357074037194252,
+ "learning_rate": 0.0006,
+ "loss": 7.649097919464111,
+ "step": 2937
+ },
+ {
+ "epoch": 40.81048034934498,
+ "grad_norm": 0.030914708971977234,
+ "learning_rate": 0.0006,
+ "loss": 7.649196624755859,
+ "step": 2938
+ },
+ {
+ "epoch": 40.82445414847162,
+ "grad_norm": 0.026296131312847137,
+ "learning_rate": 0.0006,
+ "loss": 7.629964828491211,
+ "step": 2939
+ },
+ {
+ "epoch": 40.838427947598255,
+ "grad_norm": 0.026677245274186134,
+ "learning_rate": 0.0006,
+ "loss": 7.599973678588867,
+ "step": 2940
+ },
+ {
+ "epoch": 40.852401746724894,
+ "grad_norm": 0.023208219558000565,
+ "learning_rate": 0.0006,
+ "loss": 7.601624488830566,
+ "step": 2941
+ },
+ {
+ "epoch": 40.866375545851525,
+ "grad_norm": 0.020741475746035576,
+ "learning_rate": 0.0006,
+ "loss": 7.603902339935303,
+ "step": 2942
+ },
+ {
+ "epoch": 40.880349344978164,
+ "grad_norm": 0.019221751019358635,
+ "learning_rate": 0.0006,
+ "loss": 7.5535888671875,
+ "step": 2943
+ },
+ {
+ "epoch": 40.8943231441048,
+ "grad_norm": 0.019491951912641525,
+ "learning_rate": 0.0006,
+ "loss": 7.54597282409668,
+ "step": 2944
+ },
+ {
+ "epoch": 40.90829694323144,
+ "grad_norm": 0.019702572375535965,
+ "learning_rate": 0.0006,
+ "loss": 7.537567138671875,
+ "step": 2945
+ },
+ {
+ "epoch": 40.92227074235808,
+ "grad_norm": 0.01768800988793373,
+ "learning_rate": 0.0006,
+ "loss": 7.512800216674805,
+ "step": 2946
+ },
+ {
+ "epoch": 40.93624454148472,
+ "grad_norm": 0.01926138624548912,
+ "learning_rate": 0.0006,
+ "loss": 7.509932518005371,
+ "step": 2947
+ },
+ {
+ "epoch": 40.95021834061135,
+ "grad_norm": 0.019030006602406502,
+ "learning_rate": 0.0006,
+ "loss": 7.497983932495117,
+ "step": 2948
+ },
+ {
+ "epoch": 40.96419213973799,
+ "grad_norm": 0.017313841730356216,
+ "learning_rate": 0.0006,
+ "loss": 7.463518142700195,
+ "step": 2949
+ },
+ {
+ "epoch": 40.97816593886463,
+ "grad_norm": 0.01310762483626604,
+ "learning_rate": 0.0006,
+ "loss": 7.474668502807617,
+ "step": 2950
+ },
+ {
+ "epoch": 40.992139737991266,
+ "grad_norm": 0.013008282519876957,
+ "learning_rate": 0.0006,
+ "loss": 7.45982551574707,
+ "step": 2951
+ },
+ {
+ "epoch": 41.0,
+ "grad_norm": 0.01441363524645567,
+ "learning_rate": 0.0006,
+ "loss": 7.446529388427734,
+ "step": 2952
+ },
+ {
+ "epoch": 41.0,
+ "eval_loss": 7.486120223999023,
+ "eval_runtime": 70.4986,
+ "eval_samples_per_second": 34.639,
+ "eval_steps_per_second": 1.092,
+ "step": 2952
+ },
+ {
+ "epoch": 41.01397379912664,
+ "grad_norm": 0.014061855152249336,
+ "learning_rate": 0.0006,
+ "loss": 7.4692182540893555,
+ "step": 2953
+ },
+ {
+ "epoch": 41.02794759825328,
+ "grad_norm": 0.01587500050663948,
+ "learning_rate": 0.0006,
+ "loss": 7.4182024002075195,
+ "step": 2954
+ },
+ {
+ "epoch": 41.041921397379916,
+ "grad_norm": 0.014783984050154686,
+ "learning_rate": 0.0006,
+ "loss": 7.466778755187988,
+ "step": 2955
+ },
+ {
+ "epoch": 41.05589519650655,
+ "grad_norm": 0.011560854502022266,
+ "learning_rate": 0.0006,
+ "loss": 7.438002586364746,
+ "step": 2956
+ },
+ {
+ "epoch": 41.069868995633186,
+ "grad_norm": 0.00970957800745964,
+ "learning_rate": 0.0006,
+ "loss": 7.442694187164307,
+ "step": 2957
+ },
+ {
+ "epoch": 41.083842794759825,
+ "grad_norm": 0.009633459150791168,
+ "learning_rate": 0.0006,
+ "loss": 7.458884239196777,
+ "step": 2958
+ },
+ {
+ "epoch": 41.09781659388646,
+ "grad_norm": 0.010664350353181362,
+ "learning_rate": 0.0006,
+ "loss": 7.48378324508667,
+ "step": 2959
+ },
+ {
+ "epoch": 41.1117903930131,
+ "grad_norm": 0.011440591886639595,
+ "learning_rate": 0.0006,
+ "loss": 7.474484920501709,
+ "step": 2960
+ },
+ {
+ "epoch": 41.12576419213974,
+ "grad_norm": 0.01150908600538969,
+ "learning_rate": 0.0006,
+ "loss": 7.417902946472168,
+ "step": 2961
+ },
+ {
+ "epoch": 41.13973799126637,
+ "grad_norm": 0.010375697165727615,
+ "learning_rate": 0.0006,
+ "loss": 7.403279781341553,
+ "step": 2962
+ },
+ {
+ "epoch": 41.15371179039301,
+ "grad_norm": 0.009075703099370003,
+ "learning_rate": 0.0006,
+ "loss": 7.447347164154053,
+ "step": 2963
+ },
+ {
+ "epoch": 41.16768558951965,
+ "grad_norm": 0.008892177604138851,
+ "learning_rate": 0.0006,
+ "loss": 7.402681350708008,
+ "step": 2964
+ },
+ {
+ "epoch": 41.18165938864629,
+ "grad_norm": 0.009492084383964539,
+ "learning_rate": 0.0006,
+ "loss": 7.3871355056762695,
+ "step": 2965
+ },
+ {
+ "epoch": 41.19563318777293,
+ "grad_norm": 0.009139460511505604,
+ "learning_rate": 0.0006,
+ "loss": 7.419617176055908,
+ "step": 2966
+ },
+ {
+ "epoch": 41.209606986899566,
+ "grad_norm": 0.007959138602018356,
+ "learning_rate": 0.0006,
+ "loss": 7.374949932098389,
+ "step": 2967
+ },
+ {
+ "epoch": 41.223580786026204,
+ "grad_norm": 0.007571548223495483,
+ "learning_rate": 0.0006,
+ "loss": 7.4205241203308105,
+ "step": 2968
+ },
+ {
+ "epoch": 41.237554585152836,
+ "grad_norm": 0.007885400205850601,
+ "learning_rate": 0.0006,
+ "loss": 7.416543006896973,
+ "step": 2969
+ },
+ {
+ "epoch": 41.251528384279474,
+ "grad_norm": 0.008650158531963825,
+ "learning_rate": 0.0006,
+ "loss": 7.357326984405518,
+ "step": 2970
+ },
+ {
+ "epoch": 41.26550218340611,
+ "grad_norm": 0.007483262103050947,
+ "learning_rate": 0.0006,
+ "loss": 7.410891056060791,
+ "step": 2971
+ },
+ {
+ "epoch": 41.27947598253275,
+ "grad_norm": 0.00869277399033308,
+ "learning_rate": 0.0006,
+ "loss": 7.392127990722656,
+ "step": 2972
+ },
+ {
+ "epoch": 41.29344978165939,
+ "grad_norm": 0.00885078776627779,
+ "learning_rate": 0.0006,
+ "loss": 7.405449867248535,
+ "step": 2973
+ },
+ {
+ "epoch": 41.30742358078603,
+ "grad_norm": 0.00958226341754198,
+ "learning_rate": 0.0006,
+ "loss": 7.399868488311768,
+ "step": 2974
+ },
+ {
+ "epoch": 41.32139737991266,
+ "grad_norm": 0.006991599686443806,
+ "learning_rate": 0.0006,
+ "loss": 7.35369873046875,
+ "step": 2975
+ },
+ {
+ "epoch": 41.3353711790393,
+ "grad_norm": 0.007415142375975847,
+ "learning_rate": 0.0006,
+ "loss": 7.358353137969971,
+ "step": 2976
+ },
+ {
+ "epoch": 41.34934497816594,
+ "grad_norm": 0.007024371065199375,
+ "learning_rate": 0.0006,
+ "loss": 7.4070892333984375,
+ "step": 2977
+ },
+ {
+ "epoch": 41.36331877729258,
+ "grad_norm": 0.007234332151710987,
+ "learning_rate": 0.0006,
+ "loss": 7.379120349884033,
+ "step": 2978
+ },
+ {
+ "epoch": 41.377292576419215,
+ "grad_norm": 0.00798892229795456,
+ "learning_rate": 0.0006,
+ "loss": 7.365455627441406,
+ "step": 2979
+ },
+ {
+ "epoch": 41.391266375545854,
+ "grad_norm": 0.007303398102521896,
+ "learning_rate": 0.0006,
+ "loss": 7.341798782348633,
+ "step": 2980
+ },
+ {
+ "epoch": 41.40524017467249,
+ "grad_norm": 0.007119669578969479,
+ "learning_rate": 0.0006,
+ "loss": 7.363280296325684,
+ "step": 2981
+ },
+ {
+ "epoch": 41.419213973799124,
+ "grad_norm": 0.007457619532942772,
+ "learning_rate": 0.0006,
+ "loss": 7.39388370513916,
+ "step": 2982
+ },
+ {
+ "epoch": 41.43318777292576,
+ "grad_norm": 0.007905688136816025,
+ "learning_rate": 0.0006,
+ "loss": 7.380652904510498,
+ "step": 2983
+ },
+ {
+ "epoch": 41.4471615720524,
+ "grad_norm": 0.006212855689227581,
+ "learning_rate": 0.0006,
+ "loss": 7.398769378662109,
+ "step": 2984
+ },
+ {
+ "epoch": 41.46113537117904,
+ "grad_norm": 0.0070641543716192245,
+ "learning_rate": 0.0006,
+ "loss": 7.3503546714782715,
+ "step": 2985
+ },
+ {
+ "epoch": 41.47510917030568,
+ "grad_norm": 0.00787805300205946,
+ "learning_rate": 0.0006,
+ "loss": 7.347678184509277,
+ "step": 2986
+ },
+ {
+ "epoch": 41.48908296943232,
+ "grad_norm": 0.006760426796972752,
+ "learning_rate": 0.0006,
+ "loss": 7.380006790161133,
+ "step": 2987
+ },
+ {
+ "epoch": 41.50305676855895,
+ "grad_norm": 0.00772385997697711,
+ "learning_rate": 0.0006,
+ "loss": 7.363680362701416,
+ "step": 2988
+ },
+ {
+ "epoch": 41.51703056768559,
+ "grad_norm": 0.007373564876616001,
+ "learning_rate": 0.0006,
+ "loss": 7.333014488220215,
+ "step": 2989
+ },
+ {
+ "epoch": 41.531004366812226,
+ "grad_norm": 0.006717653013765812,
+ "learning_rate": 0.0006,
+ "loss": 7.360076427459717,
+ "step": 2990
+ },
+ {
+ "epoch": 41.544978165938865,
+ "grad_norm": 0.007191413082182407,
+ "learning_rate": 0.0006,
+ "loss": 7.397030830383301,
+ "step": 2991
+ },
+ {
+ "epoch": 41.5589519650655,
+ "grad_norm": 0.007831960916519165,
+ "learning_rate": 0.0006,
+ "loss": 7.369959831237793,
+ "step": 2992
+ },
+ {
+ "epoch": 41.57292576419214,
+ "grad_norm": 0.006947481073439121,
+ "learning_rate": 0.0006,
+ "loss": 7.333254337310791,
+ "step": 2993
+ },
+ {
+ "epoch": 41.58689956331878,
+ "grad_norm": 0.006243606563657522,
+ "learning_rate": 0.0006,
+ "loss": 7.327378273010254,
+ "step": 2994
+ },
+ {
+ "epoch": 41.60087336244541,
+ "grad_norm": 0.006298339460045099,
+ "learning_rate": 0.0006,
+ "loss": 7.323482036590576,
+ "step": 2995
+ },
+ {
+ "epoch": 41.61484716157205,
+ "grad_norm": 0.007587412837892771,
+ "learning_rate": 0.0006,
+ "loss": 7.313789367675781,
+ "step": 2996
+ },
+ {
+ "epoch": 41.62882096069869,
+ "grad_norm": 0.007140746805816889,
+ "learning_rate": 0.0006,
+ "loss": 7.356067657470703,
+ "step": 2997
+ },
+ {
+ "epoch": 41.64279475982533,
+ "grad_norm": 0.008315625600516796,
+ "learning_rate": 0.0006,
+ "loss": 7.3339643478393555,
+ "step": 2998
+ },
+ {
+ "epoch": 41.65676855895197,
+ "grad_norm": 0.006547102238982916,
+ "learning_rate": 0.0006,
+ "loss": 7.342191696166992,
+ "step": 2999
+ },
+ {
+ "epoch": 41.670742358078606,
+ "grad_norm": 0.006068122107535601,
+ "learning_rate": 0.0006,
+ "loss": 7.321015357971191,
+ "step": 3000
+ },
+ {
+ "epoch": 41.68471615720524,
+ "grad_norm": 0.007316991221159697,
+ "learning_rate": 0.0006,
+ "loss": 7.346490859985352,
+ "step": 3001
+ },
+ {
+ "epoch": 41.698689956331876,
+ "grad_norm": 0.0067388564348220825,
+ "learning_rate": 0.0006,
+ "loss": 7.340723991394043,
+ "step": 3002
+ },
+ {
+ "epoch": 41.712663755458514,
+ "grad_norm": 0.006839963141828775,
+ "learning_rate": 0.0006,
+ "loss": 7.329365253448486,
+ "step": 3003
+ },
+ {
+ "epoch": 41.72663755458515,
+ "grad_norm": 0.00805041566491127,
+ "learning_rate": 0.0006,
+ "loss": 7.298934459686279,
+ "step": 3004
+ },
+ {
+ "epoch": 41.74061135371179,
+ "grad_norm": 0.007197503466159105,
+ "learning_rate": 0.0006,
+ "loss": 7.331496238708496,
+ "step": 3005
+ },
+ {
+ "epoch": 41.75458515283843,
+ "grad_norm": 0.007220805156975985,
+ "learning_rate": 0.0006,
+ "loss": 7.317695140838623,
+ "step": 3006
+ },
+ {
+ "epoch": 41.76855895196506,
+ "grad_norm": 0.00874276738613844,
+ "learning_rate": 0.0006,
+ "loss": 7.272614479064941,
+ "step": 3007
+ },
+ {
+ "epoch": 41.7825327510917,
+ "grad_norm": 0.008386565372347832,
+ "learning_rate": 0.0006,
+ "loss": 7.295244216918945,
+ "step": 3008
+ },
+ {
+ "epoch": 41.79650655021834,
+ "grad_norm": 0.0077730282209813595,
+ "learning_rate": 0.0006,
+ "loss": 7.3395490646362305,
+ "step": 3009
+ },
+ {
+ "epoch": 41.81048034934498,
+ "grad_norm": 0.008441480807960033,
+ "learning_rate": 0.0006,
+ "loss": 7.319921493530273,
+ "step": 3010
+ },
+ {
+ "epoch": 41.82445414847162,
+ "grad_norm": 0.007398269604891539,
+ "learning_rate": 0.0006,
+ "loss": 7.3029465675354,
+ "step": 3011
+ },
+ {
+ "epoch": 41.838427947598255,
+ "grad_norm": 0.006878677289932966,
+ "learning_rate": 0.0006,
+ "loss": 7.318467140197754,
+ "step": 3012
+ },
+ {
+ "epoch": 41.852401746724894,
+ "grad_norm": 0.007616598624736071,
+ "learning_rate": 0.0006,
+ "loss": 7.3112077713012695,
+ "step": 3013
+ },
+ {
+ "epoch": 41.866375545851525,
+ "grad_norm": 0.011629699729382992,
+ "learning_rate": 0.0006,
+ "loss": 7.32523250579834,
+ "step": 3014
+ },
+ {
+ "epoch": 41.880349344978164,
+ "grad_norm": 0.011011249385774136,
+ "learning_rate": 0.0006,
+ "loss": 7.287247657775879,
+ "step": 3015
+ },
+ {
+ "epoch": 41.8943231441048,
+ "grad_norm": 0.010208295658230782,
+ "learning_rate": 0.0006,
+ "loss": 7.292117118835449,
+ "step": 3016
+ },
+ {
+ "epoch": 41.90829694323144,
+ "grad_norm": 0.008651095442473888,
+ "learning_rate": 0.0006,
+ "loss": 7.302201271057129,
+ "step": 3017
+ },
+ {
+ "epoch": 41.92227074235808,
+ "grad_norm": 0.006692920345813036,
+ "learning_rate": 0.0006,
+ "loss": 7.319568634033203,
+ "step": 3018
+ },
+ {
+ "epoch": 41.93624454148472,
+ "grad_norm": 0.010033348575234413,
+ "learning_rate": 0.0006,
+ "loss": 7.2738566398620605,
+ "step": 3019
+ },
+ {
+ "epoch": 41.95021834061135,
+ "grad_norm": 0.011790832504630089,
+ "learning_rate": 0.0006,
+ "loss": 7.298540115356445,
+ "step": 3020
+ },
+ {
+ "epoch": 41.96419213973799,
+ "grad_norm": 0.018622662872076035,
+ "learning_rate": 0.0006,
+ "loss": 7.267182350158691,
+ "step": 3021
+ },
+ {
+ "epoch": 41.97816593886463,
+ "grad_norm": 0.02520257607102394,
+ "learning_rate": 0.0006,
+ "loss": 7.261566162109375,
+ "step": 3022
+ },
+ {
+ "epoch": 41.992139737991266,
+ "grad_norm": 0.036332644522190094,
+ "learning_rate": 0.0006,
+ "loss": 7.270719051361084,
+ "step": 3023
+ },
+ {
+ "epoch": 42.0,
+ "grad_norm": 0.04717305675148964,
+ "learning_rate": 0.0006,
+ "loss": 7.293332099914551,
+ "step": 3024
+ },
+ {
+ "epoch": 42.0,
+ "eval_loss": 7.354280471801758,
+ "eval_runtime": 56.4353,
+ "eval_samples_per_second": 43.271,
+ "eval_steps_per_second": 1.364,
+ "step": 3024
+ },
+ {
+ "epoch": 42.01397379912664,
+ "grad_norm": 0.06661626696586609,
+ "learning_rate": 0.0006,
+ "loss": 7.338189601898193,
+ "step": 3025
+ },
+ {
+ "epoch": 42.02794759825328,
+ "grad_norm": 0.05452054366469383,
+ "learning_rate": 0.0006,
+ "loss": 7.316764831542969,
+ "step": 3026
+ },
+ {
+ "epoch": 42.041921397379916,
+ "grad_norm": 0.025638775900006294,
+ "learning_rate": 0.0006,
+ "loss": 7.34204626083374,
+ "step": 3027
+ },
+ {
+ "epoch": 42.05589519650655,
+ "grad_norm": 0.022398799657821655,
+ "learning_rate": 0.0006,
+ "loss": 7.262149333953857,
+ "step": 3028
+ },
+ {
+ "epoch": 42.069868995633186,
+ "grad_norm": 0.03713499382138252,
+ "learning_rate": 0.0006,
+ "loss": 7.27695369720459,
+ "step": 3029
+ },
+ {
+ "epoch": 42.083842794759825,
+ "grad_norm": 0.014949247241020203,
+ "learning_rate": 0.0006,
+ "loss": 7.296508312225342,
+ "step": 3030
+ },
+ {
+ "epoch": 42.09781659388646,
+ "grad_norm": 0.02703995071351528,
+ "learning_rate": 0.0006,
+ "loss": 7.258018493652344,
+ "step": 3031
+ },
+ {
+ "epoch": 42.1117903930131,
+ "grad_norm": 0.020352816209197044,
+ "learning_rate": 0.0006,
+ "loss": 7.241450309753418,
+ "step": 3032
+ },
+ {
+ "epoch": 42.12576419213974,
+ "grad_norm": 0.014578479342162609,
+ "learning_rate": 0.0006,
+ "loss": 7.257323265075684,
+ "step": 3033
+ },
+ {
+ "epoch": 42.13973799126637,
+ "grad_norm": 0.026053044945001602,
+ "learning_rate": 0.0006,
+ "loss": 7.254680633544922,
+ "step": 3034
+ },
+ {
+ "epoch": 42.15371179039301,
+ "grad_norm": 0.015120275318622589,
+ "learning_rate": 0.0006,
+ "loss": 7.261074066162109,
+ "step": 3035
+ },
+ {
+ "epoch": 42.16768558951965,
+ "grad_norm": 0.016822177916765213,
+ "learning_rate": 0.0006,
+ "loss": 7.279064178466797,
+ "step": 3036
+ },
+ {
+ "epoch": 42.18165938864629,
+ "grad_norm": 0.016367588192224503,
+ "learning_rate": 0.0006,
+ "loss": 7.251429557800293,
+ "step": 3037
+ },
+ {
+ "epoch": 42.19563318777293,
+ "grad_norm": 0.014350418001413345,
+ "learning_rate": 0.0006,
+ "loss": 7.259264945983887,
+ "step": 3038
+ },
+ {
+ "epoch": 42.209606986899566,
+ "grad_norm": 0.017078228294849396,
+ "learning_rate": 0.0006,
+ "loss": 7.227105140686035,
+ "step": 3039
+ },
+ {
+ "epoch": 42.223580786026204,
+ "grad_norm": 0.011759317480027676,
+ "learning_rate": 0.0006,
+ "loss": 7.275888442993164,
+ "step": 3040
+ },
+ {
+ "epoch": 42.237554585152836,
+ "grad_norm": 0.013019340112805367,
+ "learning_rate": 0.0006,
+ "loss": 7.2517008781433105,
+ "step": 3041
+ },
+ {
+ "epoch": 42.251528384279474,
+ "grad_norm": 0.01178441196680069,
+ "learning_rate": 0.0006,
+ "loss": 7.23576545715332,
+ "step": 3042
+ },
+ {
+ "epoch": 42.26550218340611,
+ "grad_norm": 0.014518891461193562,
+ "learning_rate": 0.0006,
+ "loss": 7.253895282745361,
+ "step": 3043
+ },
+ {
+ "epoch": 42.27947598253275,
+ "grad_norm": 0.018159950152039528,
+ "learning_rate": 0.0006,
+ "loss": 7.278058052062988,
+ "step": 3044
+ },
+ {
+ "epoch": 42.29344978165939,
+ "grad_norm": 0.007587796077132225,
+ "learning_rate": 0.0006,
+ "loss": 7.278578758239746,
+ "step": 3045
+ },
+ {
+ "epoch": 42.30742358078603,
+ "grad_norm": 0.015994979068636894,
+ "learning_rate": 0.0006,
+ "loss": 7.239841461181641,
+ "step": 3046
+ },
+ {
+ "epoch": 42.32139737991266,
+ "grad_norm": 0.017830168828368187,
+ "learning_rate": 0.0006,
+ "loss": 7.2443976402282715,
+ "step": 3047
+ },
+ {
+ "epoch": 42.3353711790393,
+ "grad_norm": 0.008738759905099869,
+ "learning_rate": 0.0006,
+ "loss": 7.232264518737793,
+ "step": 3048
+ },
+ {
+ "epoch": 42.34934497816594,
+ "grad_norm": 0.0175449438393116,
+ "learning_rate": 0.0006,
+ "loss": 7.228000640869141,
+ "step": 3049
+ },
+ {
+ "epoch": 42.36331877729258,
+ "grad_norm": 0.019611820578575134,
+ "learning_rate": 0.0006,
+ "loss": 7.234999179840088,
+ "step": 3050
+ },
+ {
+ "epoch": 42.377292576419215,
+ "grad_norm": 0.015302863903343678,
+ "learning_rate": 0.0006,
+ "loss": 7.242547988891602,
+ "step": 3051
+ },
+ {
+ "epoch": 42.391266375545854,
+ "grad_norm": 0.025919731706380844,
+ "learning_rate": 0.0006,
+ "loss": 7.216652870178223,
+ "step": 3052
+ },
+ {
+ "epoch": 42.40524017467249,
+ "grad_norm": 0.034587468951940536,
+ "learning_rate": 0.0006,
+ "loss": 7.245336532592773,
+ "step": 3053
+ },
+ {
+ "epoch": 42.419213973799124,
+ "grad_norm": 0.019198840484023094,
+ "learning_rate": 0.0006,
+ "loss": 7.256758689880371,
+ "step": 3054
+ },
+ {
+ "epoch": 42.43318777292576,
+ "grad_norm": 0.012316389009356499,
+ "learning_rate": 0.0006,
+ "loss": 7.209102630615234,
+ "step": 3055
+ },
+ {
+ "epoch": 42.4471615720524,
+ "grad_norm": 0.023512162268161774,
+ "learning_rate": 0.0006,
+ "loss": 7.250742435455322,
+ "step": 3056
+ },
+ {
+ "epoch": 42.46113537117904,
+ "grad_norm": 0.025457851588726044,
+ "learning_rate": 0.0006,
+ "loss": 7.263757705688477,
+ "step": 3057
+ },
+ {
+ "epoch": 42.47510917030568,
+ "grad_norm": 0.02192617766559124,
+ "learning_rate": 0.0006,
+ "loss": 7.246977806091309,
+ "step": 3058
+ },
+ {
+ "epoch": 42.48908296943232,
+ "grad_norm": 0.02541281282901764,
+ "learning_rate": 0.0006,
+ "loss": 7.232361316680908,
+ "step": 3059
+ },
+ {
+ "epoch": 42.50305676855895,
+ "grad_norm": 0.036353763192892075,
+ "learning_rate": 0.0006,
+ "loss": 7.242556571960449,
+ "step": 3060
+ },
+ {
+ "epoch": 42.51703056768559,
+ "grad_norm": 0.04555872827768326,
+ "learning_rate": 0.0006,
+ "loss": 7.228940963745117,
+ "step": 3061
+ },
+ {
+ "epoch": 42.531004366812226,
+ "grad_norm": 0.04563378915190697,
+ "learning_rate": 0.0006,
+ "loss": 7.235276222229004,
+ "step": 3062
+ },
+ {
+ "epoch": 42.544978165938865,
+ "grad_norm": 0.04275760427117348,
+ "learning_rate": 0.0006,
+ "loss": 7.238895416259766,
+ "step": 3063
+ },
+ {
+ "epoch": 42.5589519650655,
+ "grad_norm": 0.04535885155200958,
+ "learning_rate": 0.0006,
+ "loss": 7.205336570739746,
+ "step": 3064
+ },
+ {
+ "epoch": 42.57292576419214,
+ "grad_norm": 0.015534617006778717,
+ "learning_rate": 0.0006,
+ "loss": 7.238393783569336,
+ "step": 3065
+ },
+ {
+ "epoch": 42.58689956331878,
+ "grad_norm": 0.03286299481987953,
+ "learning_rate": 0.0006,
+ "loss": 7.216453552246094,
+ "step": 3066
+ },
+ {
+ "epoch": 42.60087336244541,
+ "grad_norm": 0.025453168898820877,
+ "learning_rate": 0.0006,
+ "loss": 7.203750133514404,
+ "step": 3067
+ },
+ {
+ "epoch": 42.61484716157205,
+ "grad_norm": 0.021501649171113968,
+ "learning_rate": 0.0006,
+ "loss": 7.241621971130371,
+ "step": 3068
+ },
+ {
+ "epoch": 42.62882096069869,
+ "grad_norm": 0.02743634767830372,
+ "learning_rate": 0.0006,
+ "loss": 7.22076940536499,
+ "step": 3069
+ },
+ {
+ "epoch": 42.64279475982533,
+ "grad_norm": 0.020890483632683754,
+ "learning_rate": 0.0006,
+ "loss": 7.218412399291992,
+ "step": 3070
+ },
+ {
+ "epoch": 42.65676855895197,
+ "grad_norm": 0.022499850019812584,
+ "learning_rate": 0.0006,
+ "loss": 7.217696189880371,
+ "step": 3071
+ },
+ {
+ "epoch": 42.670742358078606,
+ "grad_norm": 0.01653696969151497,
+ "learning_rate": 0.0006,
+ "loss": 7.156417369842529,
+ "step": 3072
+ },
+ {
+ "epoch": 42.68471615720524,
+ "grad_norm": 0.01834225095808506,
+ "learning_rate": 0.0006,
+ "loss": 7.211247444152832,
+ "step": 3073
+ },
+ {
+ "epoch": 42.698689956331876,
+ "grad_norm": 0.020137229934334755,
+ "learning_rate": 0.0006,
+ "loss": 7.232499599456787,
+ "step": 3074
+ },
+ {
+ "epoch": 42.712663755458514,
+ "grad_norm": 0.012671459466218948,
+ "learning_rate": 0.0006,
+ "loss": 7.172441482543945,
+ "step": 3075
+ },
+ {
+ "epoch": 42.72663755458515,
+ "grad_norm": 0.016803130507469177,
+ "learning_rate": 0.0006,
+ "loss": 7.152103424072266,
+ "step": 3076
+ },
+ {
+ "epoch": 42.74061135371179,
+ "grad_norm": 0.01623712107539177,
+ "learning_rate": 0.0006,
+ "loss": 7.187032699584961,
+ "step": 3077
+ },
+ {
+ "epoch": 42.75458515283843,
+ "grad_norm": 0.015938356518745422,
+ "learning_rate": 0.0006,
+ "loss": 7.171140670776367,
+ "step": 3078
+ },
+ {
+ "epoch": 42.76855895196506,
+ "grad_norm": 0.02580736204981804,
+ "learning_rate": 0.0006,
+ "loss": 7.204472541809082,
+ "step": 3079
+ },
+ {
+ "epoch": 42.7825327510917,
+ "grad_norm": 0.023372117429971695,
+ "learning_rate": 0.0006,
+ "loss": 7.165554523468018,
+ "step": 3080
+ },
+ {
+ "epoch": 42.79650655021834,
+ "grad_norm": 0.04440586268901825,
+ "learning_rate": 0.0006,
+ "loss": 7.169431209564209,
+ "step": 3081
+ },
+ {
+ "epoch": 42.81048034934498,
+ "grad_norm": 0.06797406077384949,
+ "learning_rate": 0.0006,
+ "loss": 7.168193340301514,
+ "step": 3082
+ },
+ {
+ "epoch": 42.82445414847162,
+ "grad_norm": 0.0914321169257164,
+ "learning_rate": 0.0006,
+ "loss": 7.204671859741211,
+ "step": 3083
+ },
+ {
+ "epoch": 42.838427947598255,
+ "grad_norm": 0.0506940595805645,
+ "learning_rate": 0.0006,
+ "loss": 7.15949821472168,
+ "step": 3084
+ },
+ {
+ "epoch": 42.852401746724894,
+ "grad_norm": 0.04819890484213829,
+ "learning_rate": 0.0006,
+ "loss": 7.192837238311768,
+ "step": 3085
+ },
+ {
+ "epoch": 42.866375545851525,
+ "grad_norm": 0.040044233202934265,
+ "learning_rate": 0.0006,
+ "loss": 7.1600165367126465,
+ "step": 3086
+ },
+ {
+ "epoch": 42.880349344978164,
+ "grad_norm": 0.02573299966752529,
+ "learning_rate": 0.0006,
+ "loss": 7.153427600860596,
+ "step": 3087
+ },
+ {
+ "epoch": 42.8943231441048,
+ "grad_norm": 0.046942535787820816,
+ "learning_rate": 0.0006,
+ "loss": 7.16767692565918,
+ "step": 3088
+ },
+ {
+ "epoch": 42.90829694323144,
+ "grad_norm": 0.02683161571621895,
+ "learning_rate": 0.0006,
+ "loss": 7.17144250869751,
+ "step": 3089
+ },
+ {
+ "epoch": 42.92227074235808,
+ "grad_norm": 0.05363674834370613,
+ "learning_rate": 0.0006,
+ "loss": 7.120128631591797,
+ "step": 3090
+ },
+ {
+ "epoch": 42.93624454148472,
+ "grad_norm": 0.023756247013807297,
+ "learning_rate": 0.0006,
+ "loss": 7.169703006744385,
+ "step": 3091
+ },
+ {
+ "epoch": 42.95021834061135,
+ "grad_norm": 0.04216783866286278,
+ "learning_rate": 0.0006,
+ "loss": 7.1206865310668945,
+ "step": 3092
+ },
+ {
+ "epoch": 42.96419213973799,
+ "grad_norm": 0.02219565585255623,
+ "learning_rate": 0.0006,
+ "loss": 7.157215118408203,
+ "step": 3093
+ },
+ {
+ "epoch": 42.97816593886463,
+ "grad_norm": 0.03527127206325531,
+ "learning_rate": 0.0006,
+ "loss": 7.166672706604004,
+ "step": 3094
+ },
+ {
+ "epoch": 42.992139737991266,
+ "grad_norm": 0.017748817801475525,
+ "learning_rate": 0.0006,
+ "loss": 7.154747009277344,
+ "step": 3095
+ },
+ {
+ "epoch": 43.0,
+ "grad_norm": 0.033165838569402695,
+ "learning_rate": 0.0006,
+ "loss": 7.150866508483887,
+ "step": 3096
+ },
+ {
+ "epoch": 43.0,
+ "eval_loss": 7.144403457641602,
+ "eval_runtime": 56.8853,
+ "eval_samples_per_second": 42.928,
+ "eval_steps_per_second": 1.354,
+ "step": 3096
+ },
+ {
+ "epoch": 43.01397379912664,
+ "grad_norm": 0.015199829824268818,
+ "learning_rate": 0.0006,
+ "loss": 7.122735023498535,
+ "step": 3097
+ },
+ {
+ "epoch": 43.02794759825328,
+ "grad_norm": 0.03220256045460701,
+ "learning_rate": 0.0006,
+ "loss": 7.143553733825684,
+ "step": 3098
+ },
+ {
+ "epoch": 43.041921397379916,
+ "grad_norm": 0.014687920920550823,
+ "learning_rate": 0.0006,
+ "loss": 7.080960750579834,
+ "step": 3099
+ },
+ {
+ "epoch": 43.05589519650655,
+ "grad_norm": 0.023522932082414627,
+ "learning_rate": 0.0006,
+ "loss": 7.137356758117676,
+ "step": 3100
+ },
+ {
+ "epoch": 43.069868995633186,
+ "grad_norm": 0.014808216132223606,
+ "learning_rate": 0.0006,
+ "loss": 7.09143590927124,
+ "step": 3101
+ },
+ {
+ "epoch": 43.083842794759825,
+ "grad_norm": 0.029339246451854706,
+ "learning_rate": 0.0006,
+ "loss": 7.122198104858398,
+ "step": 3102
+ },
+ {
+ "epoch": 43.09781659388646,
+ "grad_norm": 0.022953078150749207,
+ "learning_rate": 0.0006,
+ "loss": 7.092505931854248,
+ "step": 3103
+ },
+ {
+ "epoch": 43.1117903930131,
+ "grad_norm": 0.04590052738785744,
+ "learning_rate": 0.0006,
+ "loss": 7.113100051879883,
+ "step": 3104
+ },
+ {
+ "epoch": 43.12576419213974,
+ "grad_norm": 0.0524442233145237,
+ "learning_rate": 0.0006,
+ "loss": 7.147022247314453,
+ "step": 3105
+ },
+ {
+ "epoch": 43.13973799126637,
+ "grad_norm": 0.0315794013440609,
+ "learning_rate": 0.0006,
+ "loss": 7.094045162200928,
+ "step": 3106
+ },
+ {
+ "epoch": 43.15371179039301,
+ "grad_norm": 0.05112722888588905,
+ "learning_rate": 0.0006,
+ "loss": 7.054388523101807,
+ "step": 3107
+ },
+ {
+ "epoch": 43.16768558951965,
+ "grad_norm": 0.03146776929497719,
+ "learning_rate": 0.0006,
+ "loss": 7.116674423217773,
+ "step": 3108
+ },
+ {
+ "epoch": 43.18165938864629,
+ "grad_norm": 0.03391246497631073,
+ "learning_rate": 0.0006,
+ "loss": 7.150345802307129,
+ "step": 3109
+ },
+ {
+ "epoch": 43.19563318777293,
+ "grad_norm": 0.026957903057336807,
+ "learning_rate": 0.0006,
+ "loss": 7.058531284332275,
+ "step": 3110
+ },
+ {
+ "epoch": 43.209606986899566,
+ "grad_norm": 0.041532568633556366,
+ "learning_rate": 0.0006,
+ "loss": 7.085498809814453,
+ "step": 3111
+ },
+ {
+ "epoch": 43.223580786026204,
+ "grad_norm": 0.020001588389277458,
+ "learning_rate": 0.0006,
+ "loss": 7.131487846374512,
+ "step": 3112
+ },
+ {
+ "epoch": 43.237554585152836,
+ "grad_norm": 0.02268712781369686,
+ "learning_rate": 0.0006,
+ "loss": 7.0790863037109375,
+ "step": 3113
+ },
+ {
+ "epoch": 43.251528384279474,
+ "grad_norm": 0.023363925516605377,
+ "learning_rate": 0.0006,
+ "loss": 7.067816734313965,
+ "step": 3114
+ },
+ {
+ "epoch": 43.26550218340611,
+ "grad_norm": 0.019367508590221405,
+ "learning_rate": 0.0006,
+ "loss": 7.042041778564453,
+ "step": 3115
+ },
+ {
+ "epoch": 43.27947598253275,
+ "grad_norm": 0.02745090238749981,
+ "learning_rate": 0.0006,
+ "loss": 7.064733028411865,
+ "step": 3116
+ },
+ {
+ "epoch": 43.29344978165939,
+ "grad_norm": 0.02543802745640278,
+ "learning_rate": 0.0006,
+ "loss": 7.058948040008545,
+ "step": 3117
+ },
+ {
+ "epoch": 43.30742358078603,
+ "grad_norm": 0.018874092027544975,
+ "learning_rate": 0.0006,
+ "loss": 7.009946346282959,
+ "step": 3118
+ },
+ {
+ "epoch": 43.32139737991266,
+ "grad_norm": 0.024662822484970093,
+ "learning_rate": 0.0006,
+ "loss": 7.015913963317871,
+ "step": 3119
+ },
+ {
+ "epoch": 43.3353711790393,
+ "grad_norm": 0.02441416308283806,
+ "learning_rate": 0.0006,
+ "loss": 7.061622142791748,
+ "step": 3120
+ },
+ {
+ "epoch": 43.34934497816594,
+ "grad_norm": 0.031887322664260864,
+ "learning_rate": 0.0006,
+ "loss": 7.046883583068848,
+ "step": 3121
+ },
+ {
+ "epoch": 43.36331877729258,
+ "grad_norm": 0.021930739283561707,
+ "learning_rate": 0.0006,
+ "loss": 7.0309624671936035,
+ "step": 3122
+ },
+ {
+ "epoch": 43.377292576419215,
+ "grad_norm": 0.02722783200442791,
+ "learning_rate": 0.0006,
+ "loss": 7.072924613952637,
+ "step": 3123
+ },
+ {
+ "epoch": 43.391266375545854,
+ "grad_norm": 0.01483179721981287,
+ "learning_rate": 0.0006,
+ "loss": 7.028624534606934,
+ "step": 3124
+ },
+ {
+ "epoch": 43.40524017467249,
+ "grad_norm": 0.024793945252895355,
+ "learning_rate": 0.0006,
+ "loss": 7.001623153686523,
+ "step": 3125
+ },
+ {
+ "epoch": 43.419213973799124,
+ "grad_norm": 0.020148713141679764,
+ "learning_rate": 0.0006,
+ "loss": 6.9938812255859375,
+ "step": 3126
+ },
+ {
+ "epoch": 43.43318777292576,
+ "grad_norm": 0.02413850650191307,
+ "learning_rate": 0.0006,
+ "loss": 7.001243591308594,
+ "step": 3127
+ },
+ {
+ "epoch": 43.4471615720524,
+ "grad_norm": 0.028549203649163246,
+ "learning_rate": 0.0006,
+ "loss": 6.987671375274658,
+ "step": 3128
+ },
+ {
+ "epoch": 43.46113537117904,
+ "grad_norm": 0.025988785549998283,
+ "learning_rate": 0.0006,
+ "loss": 6.9551167488098145,
+ "step": 3129
+ },
+ {
+ "epoch": 43.47510917030568,
+ "grad_norm": 0.027140729129314423,
+ "learning_rate": 0.0006,
+ "loss": 6.9547810554504395,
+ "step": 3130
+ },
+ {
+ "epoch": 43.48908296943232,
+ "grad_norm": 0.030130082741379738,
+ "learning_rate": 0.0006,
+ "loss": 6.993186950683594,
+ "step": 3131
+ },
+ {
+ "epoch": 43.50305676855895,
+ "grad_norm": 0.025996994227170944,
+ "learning_rate": 0.0006,
+ "loss": 6.984624862670898,
+ "step": 3132
+ },
+ {
+ "epoch": 43.51703056768559,
+ "grad_norm": 0.02423008903861046,
+ "learning_rate": 0.0006,
+ "loss": 6.98813009262085,
+ "step": 3133
+ },
+ {
+ "epoch": 43.531004366812226,
+ "grad_norm": 0.029665837064385414,
+ "learning_rate": 0.0006,
+ "loss": 6.986610412597656,
+ "step": 3134
+ },
+ {
+ "epoch": 43.544978165938865,
+ "grad_norm": 0.028768595308065414,
+ "learning_rate": 0.0006,
+ "loss": 6.898797035217285,
+ "step": 3135
+ },
+ {
+ "epoch": 43.5589519650655,
+ "grad_norm": 0.03111208975315094,
+ "learning_rate": 0.0006,
+ "loss": 6.923741817474365,
+ "step": 3136
+ },
+ {
+ "epoch": 43.57292576419214,
+ "grad_norm": 0.0368187390267849,
+ "learning_rate": 0.0006,
+ "loss": 6.994343280792236,
+ "step": 3137
+ },
+ {
+ "epoch": 43.58689956331878,
+ "grad_norm": 0.04076763987541199,
+ "learning_rate": 0.0006,
+ "loss": 6.909458160400391,
+ "step": 3138
+ },
+ {
+ "epoch": 43.60087336244541,
+ "grad_norm": 0.03281061723828316,
+ "learning_rate": 0.0006,
+ "loss": 6.924851417541504,
+ "step": 3139
+ },
+ {
+ "epoch": 43.61484716157205,
+ "grad_norm": 0.025486988946795464,
+ "learning_rate": 0.0006,
+ "loss": 6.962926864624023,
+ "step": 3140
+ },
+ {
+ "epoch": 43.62882096069869,
+ "grad_norm": 0.04583485797047615,
+ "learning_rate": 0.0006,
+ "loss": 6.876555442810059,
+ "step": 3141
+ },
+ {
+ "epoch": 43.64279475982533,
+ "grad_norm": 0.05460785701870918,
+ "learning_rate": 0.0006,
+ "loss": 6.952932357788086,
+ "step": 3142
+ },
+ {
+ "epoch": 43.65676855895197,
+ "grad_norm": 0.040917906910181046,
+ "learning_rate": 0.0006,
+ "loss": 6.9624924659729,
+ "step": 3143
+ },
+ {
+ "epoch": 43.670742358078606,
+ "grad_norm": 0.04009668901562691,
+ "learning_rate": 0.0006,
+ "loss": 6.94033145904541,
+ "step": 3144
+ },
+ {
+ "epoch": 43.68471615720524,
+ "grad_norm": 0.04500625655055046,
+ "learning_rate": 0.0006,
+ "loss": 6.961839199066162,
+ "step": 3145
+ },
+ {
+ "epoch": 43.698689956331876,
+ "grad_norm": 0.03182258829474449,
+ "learning_rate": 0.0006,
+ "loss": 6.930580139160156,
+ "step": 3146
+ },
+ {
+ "epoch": 43.712663755458514,
+ "grad_norm": 0.053621504455804825,
+ "learning_rate": 0.0006,
+ "loss": 6.950235366821289,
+ "step": 3147
+ },
+ {
+ "epoch": 43.72663755458515,
+ "grad_norm": 0.031919289380311966,
+ "learning_rate": 0.0006,
+ "loss": 6.895111560821533,
+ "step": 3148
+ },
+ {
+ "epoch": 43.74061135371179,
+ "grad_norm": 0.042713891714811325,
+ "learning_rate": 0.0006,
+ "loss": 6.940885543823242,
+ "step": 3149
+ },
+ {
+ "epoch": 43.75458515283843,
+ "grad_norm": 0.02385653182864189,
+ "learning_rate": 0.0006,
+ "loss": 6.948801517486572,
+ "step": 3150
+ },
+ {
+ "epoch": 43.76855895196506,
+ "grad_norm": 0.04100663587450981,
+ "learning_rate": 0.0006,
+ "loss": 6.970411777496338,
+ "step": 3151
+ },
+ {
+ "epoch": 43.7825327510917,
+ "grad_norm": 0.02729583904147148,
+ "learning_rate": 0.0006,
+ "loss": 6.899379253387451,
+ "step": 3152
+ },
+ {
+ "epoch": 43.79650655021834,
+ "grad_norm": 0.03128620237112045,
+ "learning_rate": 0.0006,
+ "loss": 6.931319236755371,
+ "step": 3153
+ },
+ {
+ "epoch": 43.81048034934498,
+ "grad_norm": 0.025488659739494324,
+ "learning_rate": 0.0006,
+ "loss": 6.8486409187316895,
+ "step": 3154
+ },
+ {
+ "epoch": 43.82445414847162,
+ "grad_norm": 0.03679376840591431,
+ "learning_rate": 0.0006,
+ "loss": 6.840811252593994,
+ "step": 3155
+ },
+ {
+ "epoch": 43.838427947598255,
+ "grad_norm": 0.020578034222126007,
+ "learning_rate": 0.0006,
+ "loss": 6.9088616371154785,
+ "step": 3156
+ },
+ {
+ "epoch": 43.852401746724894,
+ "grad_norm": 0.02935841865837574,
+ "learning_rate": 0.0006,
+ "loss": 6.806676864624023,
+ "step": 3157
+ },
+ {
+ "epoch": 43.866375545851525,
+ "grad_norm": 0.023319244384765625,
+ "learning_rate": 0.0006,
+ "loss": 6.878457546234131,
+ "step": 3158
+ },
+ {
+ "epoch": 43.880349344978164,
+ "grad_norm": 0.02779589593410492,
+ "learning_rate": 0.0006,
+ "loss": 6.822447776794434,
+ "step": 3159
+ },
+ {
+ "epoch": 43.8943231441048,
+ "grad_norm": 0.01879825070500374,
+ "learning_rate": 0.0006,
+ "loss": 6.854121208190918,
+ "step": 3160
+ },
+ {
+ "epoch": 43.90829694323144,
+ "grad_norm": 0.023286139592528343,
+ "learning_rate": 0.0006,
+ "loss": 6.836827278137207,
+ "step": 3161
+ },
+ {
+ "epoch": 43.92227074235808,
+ "grad_norm": 0.030358297750353813,
+ "learning_rate": 0.0006,
+ "loss": 6.860037803649902,
+ "step": 3162
+ },
+ {
+ "epoch": 43.93624454148472,
+ "grad_norm": 0.029621105641126633,
+ "learning_rate": 0.0006,
+ "loss": 6.884025573730469,
+ "step": 3163
+ },
+ {
+ "epoch": 43.95021834061135,
+ "grad_norm": 0.029829824343323708,
+ "learning_rate": 0.0006,
+ "loss": 6.8702392578125,
+ "step": 3164
+ },
+ {
+ "epoch": 43.96419213973799,
+ "grad_norm": 0.036892473697662354,
+ "learning_rate": 0.0006,
+ "loss": 6.806652069091797,
+ "step": 3165
+ },
+ {
+ "epoch": 43.97816593886463,
+ "grad_norm": 0.03206709027290344,
+ "learning_rate": 0.0006,
+ "loss": 6.828495979309082,
+ "step": 3166
+ },
+ {
+ "epoch": 43.992139737991266,
+ "grad_norm": 0.02744000405073166,
+ "learning_rate": 0.0006,
+ "loss": 6.832002639770508,
+ "step": 3167
+ },
+ {
+ "epoch": 44.0,
+ "grad_norm": 0.02205660194158554,
+ "learning_rate": 0.0006,
+ "loss": 6.898828506469727,
+ "step": 3168
+ },
+ {
+ "epoch": 44.0,
+ "eval_loss": 6.8150553703308105,
+ "eval_runtime": 56.784,
+ "eval_samples_per_second": 43.005,
+ "eval_steps_per_second": 1.356,
+ "step": 3168
+ },
+ {
+ "epoch": 44.01397379912664,
+ "grad_norm": 0.030421260744333267,
+ "learning_rate": 0.0006,
+ "loss": 6.775057792663574,
+ "step": 3169
+ },
+ {
+ "epoch": 44.02794759825328,
+ "grad_norm": 0.0465611033141613,
+ "learning_rate": 0.0006,
+ "loss": 6.849390506744385,
+ "step": 3170
+ },
+ {
+ "epoch": 44.041921397379916,
+ "grad_norm": 0.05977972596883774,
+ "learning_rate": 0.0006,
+ "loss": 6.725576400756836,
+ "step": 3171
+ },
+ {
+ "epoch": 44.05589519650655,
+ "grad_norm": 0.05432697385549545,
+ "learning_rate": 0.0006,
+ "loss": 6.768024444580078,
+ "step": 3172
+ },
+ {
+ "epoch": 44.069868995633186,
+ "grad_norm": 0.04850967228412628,
+ "learning_rate": 0.0006,
+ "loss": 6.799457550048828,
+ "step": 3173
+ },
+ {
+ "epoch": 44.083842794759825,
+ "grad_norm": 0.06201786920428276,
+ "learning_rate": 0.0006,
+ "loss": 6.897234916687012,
+ "step": 3174
+ },
+ {
+ "epoch": 44.09781659388646,
+ "grad_norm": 0.04695591703057289,
+ "learning_rate": 0.0006,
+ "loss": 6.826470375061035,
+ "step": 3175
+ },
+ {
+ "epoch": 44.1117903930131,
+ "grad_norm": 0.03850164636969566,
+ "learning_rate": 0.0006,
+ "loss": 6.835901260375977,
+ "step": 3176
+ },
+ {
+ "epoch": 44.12576419213974,
+ "grad_norm": 0.04250155761837959,
+ "learning_rate": 0.0006,
+ "loss": 6.782310485839844,
+ "step": 3177
+ },
+ {
+ "epoch": 44.13973799126637,
+ "grad_norm": 0.041465479880571365,
+ "learning_rate": 0.0006,
+ "loss": 6.785732746124268,
+ "step": 3178
+ },
+ {
+ "epoch": 44.15371179039301,
+ "grad_norm": 0.02527807280421257,
+ "learning_rate": 0.0006,
+ "loss": 6.78061580657959,
+ "step": 3179
+ },
+ {
+ "epoch": 44.16768558951965,
+ "grad_norm": 0.03660374879837036,
+ "learning_rate": 0.0006,
+ "loss": 6.783992767333984,
+ "step": 3180
+ },
+ {
+ "epoch": 44.18165938864629,
+ "grad_norm": 0.027567964047193527,
+ "learning_rate": 0.0006,
+ "loss": 6.765413284301758,
+ "step": 3181
+ },
+ {
+ "epoch": 44.19563318777293,
+ "grad_norm": 0.024301424622535706,
+ "learning_rate": 0.0006,
+ "loss": 6.760368347167969,
+ "step": 3182
+ },
+ {
+ "epoch": 44.209606986899566,
+ "grad_norm": 0.0266821701079607,
+ "learning_rate": 0.0006,
+ "loss": 6.71417236328125,
+ "step": 3183
+ },
+ {
+ "epoch": 44.223580786026204,
+ "grad_norm": 0.028125090524554253,
+ "learning_rate": 0.0006,
+ "loss": 6.646126747131348,
+ "step": 3184
+ },
+ {
+ "epoch": 44.237554585152836,
+ "grad_norm": 0.0185737032443285,
+ "learning_rate": 0.0006,
+ "loss": 6.712944030761719,
+ "step": 3185
+ },
+ {
+ "epoch": 44.251528384279474,
+ "grad_norm": 0.025836674496531487,
+ "learning_rate": 0.0006,
+ "loss": 6.707030296325684,
+ "step": 3186
+ },
+ {
+ "epoch": 44.26550218340611,
+ "grad_norm": 0.029744870960712433,
+ "learning_rate": 0.0006,
+ "loss": 6.6271071434021,
+ "step": 3187
+ },
+ {
+ "epoch": 44.27947598253275,
+ "grad_norm": 0.03021426685154438,
+ "learning_rate": 0.0006,
+ "loss": 6.677826404571533,
+ "step": 3188
+ },
+ {
+ "epoch": 44.29344978165939,
+ "grad_norm": 0.032727014273405075,
+ "learning_rate": 0.0006,
+ "loss": 6.657764434814453,
+ "step": 3189
+ },
+ {
+ "epoch": 44.30742358078603,
+ "grad_norm": 0.03696778789162636,
+ "learning_rate": 0.0006,
+ "loss": 6.632883548736572,
+ "step": 3190
+ },
+ {
+ "epoch": 44.32139737991266,
+ "grad_norm": 0.034630827605724335,
+ "learning_rate": 0.0006,
+ "loss": 6.759583473205566,
+ "step": 3191
+ },
+ {
+ "epoch": 44.3353711790393,
+ "grad_norm": 0.030517758801579475,
+ "learning_rate": 0.0006,
+ "loss": 6.702268600463867,
+ "step": 3192
+ },
+ {
+ "epoch": 44.34934497816594,
+ "grad_norm": 0.04105190932750702,
+ "learning_rate": 0.0006,
+ "loss": 6.705792427062988,
+ "step": 3193
+ },
+ {
+ "epoch": 44.36331877729258,
+ "grad_norm": 0.06659174710512161,
+ "learning_rate": 0.0006,
+ "loss": 6.73862361907959,
+ "step": 3194
+ },
+ {
+ "epoch": 44.377292576419215,
+ "grad_norm": 0.08013619482517242,
+ "learning_rate": 0.0006,
+ "loss": 6.710282802581787,
+ "step": 3195
+ },
+ {
+ "epoch": 44.391266375545854,
+ "grad_norm": 0.049850963056087494,
+ "learning_rate": 0.0006,
+ "loss": 6.684974670410156,
+ "step": 3196
+ },
+ {
+ "epoch": 44.40524017467249,
+ "grad_norm": 0.061247408390045166,
+ "learning_rate": 0.0006,
+ "loss": 6.716034412384033,
+ "step": 3197
+ },
+ {
+ "epoch": 44.419213973799124,
+ "grad_norm": 0.05868857726454735,
+ "learning_rate": 0.0006,
+ "loss": 6.589107513427734,
+ "step": 3198
+ },
+ {
+ "epoch": 44.43318777292576,
+ "grad_norm": 0.048878561705350876,
+ "learning_rate": 0.0006,
+ "loss": 6.694855690002441,
+ "step": 3199
+ },
+ {
+ "epoch": 44.4471615720524,
+ "grad_norm": 0.028748324140906334,
+ "learning_rate": 0.0006,
+ "loss": 6.629969596862793,
+ "step": 3200
+ },
+ {
+ "epoch": 44.46113537117904,
+ "grad_norm": 0.04060306400060654,
+ "learning_rate": 0.0006,
+ "loss": 6.721431255340576,
+ "step": 3201
+ },
+ {
+ "epoch": 44.47510917030568,
+ "grad_norm": 0.03833824768662453,
+ "learning_rate": 0.0006,
+ "loss": 6.625029563903809,
+ "step": 3202
+ },
+ {
+ "epoch": 44.48908296943232,
+ "grad_norm": 0.03791997581720352,
+ "learning_rate": 0.0006,
+ "loss": 6.669187545776367,
+ "step": 3203
+ },
+ {
+ "epoch": 44.50305676855895,
+ "grad_norm": 0.03259684890508652,
+ "learning_rate": 0.0006,
+ "loss": 6.665643692016602,
+ "step": 3204
+ },
+ {
+ "epoch": 44.51703056768559,
+ "grad_norm": 0.040658626705408096,
+ "learning_rate": 0.0006,
+ "loss": 6.639796257019043,
+ "step": 3205
+ },
+ {
+ "epoch": 44.531004366812226,
+ "grad_norm": 0.03691193461418152,
+ "learning_rate": 0.0006,
+ "loss": 6.692244529724121,
+ "step": 3206
+ },
+ {
+ "epoch": 44.544978165938865,
+ "grad_norm": 0.02814812771975994,
+ "learning_rate": 0.0006,
+ "loss": 6.61978006362915,
+ "step": 3207
+ },
+ {
+ "epoch": 44.5589519650655,
+ "grad_norm": 0.022155074402689934,
+ "learning_rate": 0.0006,
+ "loss": 6.625905990600586,
+ "step": 3208
+ },
+ {
+ "epoch": 44.57292576419214,
+ "grad_norm": 0.030794885009527206,
+ "learning_rate": 0.0006,
+ "loss": 6.678500652313232,
+ "step": 3209
+ },
+ {
+ "epoch": 44.58689956331878,
+ "grad_norm": 0.03201768547296524,
+ "learning_rate": 0.0006,
+ "loss": 6.6015729904174805,
+ "step": 3210
+ },
+ {
+ "epoch": 44.60087336244541,
+ "grad_norm": 0.025138361379504204,
+ "learning_rate": 0.0006,
+ "loss": 6.568234443664551,
+ "step": 3211
+ },
+ {
+ "epoch": 44.61484716157205,
+ "grad_norm": 0.02969353087246418,
+ "learning_rate": 0.0006,
+ "loss": 6.588366508483887,
+ "step": 3212
+ },
+ {
+ "epoch": 44.62882096069869,
+ "grad_norm": 0.03136962279677391,
+ "learning_rate": 0.0006,
+ "loss": 6.479961395263672,
+ "step": 3213
+ },
+ {
+ "epoch": 44.64279475982533,
+ "grad_norm": 0.030178170651197433,
+ "learning_rate": 0.0006,
+ "loss": 6.622910976409912,
+ "step": 3214
+ },
+ {
+ "epoch": 44.65676855895197,
+ "grad_norm": 0.02682245895266533,
+ "learning_rate": 0.0006,
+ "loss": 6.597707748413086,
+ "step": 3215
+ },
+ {
+ "epoch": 44.670742358078606,
+ "grad_norm": 0.025154557079076767,
+ "learning_rate": 0.0006,
+ "loss": 6.630541801452637,
+ "step": 3216
+ },
+ {
+ "epoch": 44.68471615720524,
+ "grad_norm": 0.03564463183283806,
+ "learning_rate": 0.0006,
+ "loss": 6.542823791503906,
+ "step": 3217
+ },
+ {
+ "epoch": 44.698689956331876,
+ "grad_norm": 0.049477651715278625,
+ "learning_rate": 0.0006,
+ "loss": 6.646030426025391,
+ "step": 3218
+ },
+ {
+ "epoch": 44.712663755458514,
+ "grad_norm": 0.06536535918712616,
+ "learning_rate": 0.0006,
+ "loss": 6.595463275909424,
+ "step": 3219
+ },
+ {
+ "epoch": 44.72663755458515,
+ "grad_norm": 0.07710316777229309,
+ "learning_rate": 0.0006,
+ "loss": 6.55524206161499,
+ "step": 3220
+ },
+ {
+ "epoch": 44.74061135371179,
+ "grad_norm": 0.09571163356304169,
+ "learning_rate": 0.0006,
+ "loss": 6.659764289855957,
+ "step": 3221
+ },
+ {
+ "epoch": 44.75458515283843,
+ "grad_norm": 0.055076830089092255,
+ "learning_rate": 0.0006,
+ "loss": 6.658882141113281,
+ "step": 3222
+ },
+ {
+ "epoch": 44.76855895196506,
+ "grad_norm": 0.07671099156141281,
+ "learning_rate": 0.0006,
+ "loss": 6.605809211730957,
+ "step": 3223
+ },
+ {
+ "epoch": 44.7825327510917,
+ "grad_norm": 0.05295584350824356,
+ "learning_rate": 0.0006,
+ "loss": 6.641455173492432,
+ "step": 3224
+ },
+ {
+ "epoch": 44.79650655021834,
+ "grad_norm": 0.09128042310476303,
+ "learning_rate": 0.0006,
+ "loss": 6.691066741943359,
+ "step": 3225
+ },
+ {
+ "epoch": 44.81048034934498,
+ "grad_norm": 0.05362958461046219,
+ "learning_rate": 0.0006,
+ "loss": 6.573406219482422,
+ "step": 3226
+ },
+ {
+ "epoch": 44.82445414847162,
+ "grad_norm": 0.046910062432289124,
+ "learning_rate": 0.0006,
+ "loss": 6.550876617431641,
+ "step": 3227
+ },
+ {
+ "epoch": 44.838427947598255,
+ "grad_norm": 0.03494197130203247,
+ "learning_rate": 0.0006,
+ "loss": 6.540179252624512,
+ "step": 3228
+ },
+ {
+ "epoch": 44.852401746724894,
+ "grad_norm": 0.04716387391090393,
+ "learning_rate": 0.0006,
+ "loss": 6.490459442138672,
+ "step": 3229
+ },
+ {
+ "epoch": 44.866375545851525,
+ "grad_norm": 0.04190506786108017,
+ "learning_rate": 0.0006,
+ "loss": 6.509925365447998,
+ "step": 3230
+ },
+ {
+ "epoch": 44.880349344978164,
+ "grad_norm": 0.04749064892530441,
+ "learning_rate": 0.0006,
+ "loss": 6.478779315948486,
+ "step": 3231
+ },
+ {
+ "epoch": 44.8943231441048,
+ "grad_norm": 0.028002185747027397,
+ "learning_rate": 0.0006,
+ "loss": 6.528904914855957,
+ "step": 3232
+ },
+ {
+ "epoch": 44.90829694323144,
+ "grad_norm": 0.036473166197538376,
+ "learning_rate": 0.0006,
+ "loss": 6.549042701721191,
+ "step": 3233
+ },
+ {
+ "epoch": 44.92227074235808,
+ "grad_norm": 0.02543315477669239,
+ "learning_rate": 0.0006,
+ "loss": 6.498875141143799,
+ "step": 3234
+ },
+ {
+ "epoch": 44.93624454148472,
+ "grad_norm": 0.029906807467341423,
+ "learning_rate": 0.0006,
+ "loss": 6.453019142150879,
+ "step": 3235
+ },
+ {
+ "epoch": 44.95021834061135,
+ "grad_norm": 0.027231857180595398,
+ "learning_rate": 0.0006,
+ "loss": 6.428829193115234,
+ "step": 3236
+ },
+ {
+ "epoch": 44.96419213973799,
+ "grad_norm": 0.02506154775619507,
+ "learning_rate": 0.0006,
+ "loss": 6.420858383178711,
+ "step": 3237
+ },
+ {
+ "epoch": 44.97816593886463,
+ "grad_norm": 0.0269472673535347,
+ "learning_rate": 0.0006,
+ "loss": 6.484989643096924,
+ "step": 3238
+ },
+ {
+ "epoch": 44.992139737991266,
+ "grad_norm": 0.023410487920045853,
+ "learning_rate": 0.0006,
+ "loss": 6.454830169677734,
+ "step": 3239
+ },
+ {
+ "epoch": 45.0,
+ "grad_norm": 0.034175023436546326,
+ "learning_rate": 0.0006,
+ "loss": 6.430975437164307,
+ "step": 3240
+ },
+ {
+ "epoch": 45.0,
+ "eval_loss": 6.458896160125732,
+ "eval_runtime": 56.6934,
+ "eval_samples_per_second": 43.074,
+ "eval_steps_per_second": 1.358,
+ "step": 3240
+ },
+ {
+ "epoch": 45.01397379912664,
+ "grad_norm": 0.045031629502773285,
+ "learning_rate": 0.0006,
+ "loss": 6.4243011474609375,
+ "step": 3241
+ },
+ {
+ "epoch": 45.02794759825328,
+ "grad_norm": 0.04968734085559845,
+ "learning_rate": 0.0006,
+ "loss": 6.380091667175293,
+ "step": 3242
+ },
+ {
+ "epoch": 45.041921397379916,
+ "grad_norm": 0.047145187854766846,
+ "learning_rate": 0.0006,
+ "loss": 6.477178573608398,
+ "step": 3243
+ },
+ {
+ "epoch": 45.05589519650655,
+ "grad_norm": 0.04304948449134827,
+ "learning_rate": 0.0006,
+ "loss": 6.399726867675781,
+ "step": 3244
+ },
+ {
+ "epoch": 45.069868995633186,
+ "grad_norm": 0.04466010630130768,
+ "learning_rate": 0.0006,
+ "loss": 6.3415679931640625,
+ "step": 3245
+ },
+ {
+ "epoch": 45.083842794759825,
+ "grad_norm": 0.04305306077003479,
+ "learning_rate": 0.0006,
+ "loss": 6.3413310050964355,
+ "step": 3246
+ },
+ {
+ "epoch": 45.09781659388646,
+ "grad_norm": 0.043064575642347336,
+ "learning_rate": 0.0006,
+ "loss": 6.4838714599609375,
+ "step": 3247
+ },
+ {
+ "epoch": 45.1117903930131,
+ "grad_norm": 0.049503277987241745,
+ "learning_rate": 0.0006,
+ "loss": 6.38473653793335,
+ "step": 3248
+ },
+ {
+ "epoch": 45.12576419213974,
+ "grad_norm": 0.04613238573074341,
+ "learning_rate": 0.0006,
+ "loss": 6.370769500732422,
+ "step": 3249
+ },
+ {
+ "epoch": 45.13973799126637,
+ "grad_norm": 0.052119478583335876,
+ "learning_rate": 0.0006,
+ "loss": 6.362321853637695,
+ "step": 3250
+ },
+ {
+ "epoch": 45.15371179039301,
+ "grad_norm": 0.045110076665878296,
+ "learning_rate": 0.0006,
+ "loss": 6.369844913482666,
+ "step": 3251
+ },
+ {
+ "epoch": 45.16768558951965,
+ "grad_norm": 0.03798583149909973,
+ "learning_rate": 0.0006,
+ "loss": 6.324633598327637,
+ "step": 3252
+ },
+ {
+ "epoch": 45.18165938864629,
+ "grad_norm": 0.03744659945368767,
+ "learning_rate": 0.0006,
+ "loss": 6.349421501159668,
+ "step": 3253
+ },
+ {
+ "epoch": 45.19563318777293,
+ "grad_norm": 0.03767292574048042,
+ "learning_rate": 0.0006,
+ "loss": 6.334238052368164,
+ "step": 3254
+ },
+ {
+ "epoch": 45.209606986899566,
+ "grad_norm": 0.03577509894967079,
+ "learning_rate": 0.0006,
+ "loss": 6.3453826904296875,
+ "step": 3255
+ },
+ {
+ "epoch": 45.223580786026204,
+ "grad_norm": 0.03403240442276001,
+ "learning_rate": 0.0006,
+ "loss": 6.2836174964904785,
+ "step": 3256
+ },
+ {
+ "epoch": 45.237554585152836,
+ "grad_norm": 0.03517331928014755,
+ "learning_rate": 0.0006,
+ "loss": 6.281718730926514,
+ "step": 3257
+ },
+ {
+ "epoch": 45.251528384279474,
+ "grad_norm": 0.04409139230847359,
+ "learning_rate": 0.0006,
+ "loss": 6.320815086364746,
+ "step": 3258
+ },
+ {
+ "epoch": 45.26550218340611,
+ "grad_norm": 0.054750364273786545,
+ "learning_rate": 0.0006,
+ "loss": 6.343618392944336,
+ "step": 3259
+ },
+ {
+ "epoch": 45.27947598253275,
+ "grad_norm": 0.07612067461013794,
+ "learning_rate": 0.0006,
+ "loss": 6.310614585876465,
+ "step": 3260
+ },
+ {
+ "epoch": 45.29344978165939,
+ "grad_norm": 0.08438985049724579,
+ "learning_rate": 0.0006,
+ "loss": 6.2369184494018555,
+ "step": 3261
+ },
+ {
+ "epoch": 45.30742358078603,
+ "grad_norm": 0.0782201811671257,
+ "learning_rate": 0.0006,
+ "loss": 6.254452705383301,
+ "step": 3262
+ },
+ {
+ "epoch": 45.32139737991266,
+ "grad_norm": 0.09928355365991592,
+ "learning_rate": 0.0006,
+ "loss": 6.3112077713012695,
+ "step": 3263
+ },
+ {
+ "epoch": 45.3353711790393,
+ "grad_norm": 0.07233133912086487,
+ "learning_rate": 0.0006,
+ "loss": 6.408087253570557,
+ "step": 3264
+ },
+ {
+ "epoch": 45.34934497816594,
+ "grad_norm": 0.07478684186935425,
+ "learning_rate": 0.0006,
+ "loss": 6.3770036697387695,
+ "step": 3265
+ },
+ {
+ "epoch": 45.36331877729258,
+ "grad_norm": 0.06036846712231636,
+ "learning_rate": 0.0006,
+ "loss": 6.33865213394165,
+ "step": 3266
+ },
+ {
+ "epoch": 45.377292576419215,
+ "grad_norm": 0.05783458054065704,
+ "learning_rate": 0.0006,
+ "loss": 6.363203048706055,
+ "step": 3267
+ },
+ {
+ "epoch": 45.391266375545854,
+ "grad_norm": 0.05443909019231796,
+ "learning_rate": 0.0006,
+ "loss": 6.287069797515869,
+ "step": 3268
+ },
+ {
+ "epoch": 45.40524017467249,
+ "grad_norm": 0.0557909794151783,
+ "learning_rate": 0.0006,
+ "loss": 6.297784805297852,
+ "step": 3269
+ },
+ {
+ "epoch": 45.419213973799124,
+ "grad_norm": 0.03483372926712036,
+ "learning_rate": 0.0006,
+ "loss": 6.280734062194824,
+ "step": 3270
+ },
+ {
+ "epoch": 45.43318777292576,
+ "grad_norm": 0.0425543375313282,
+ "learning_rate": 0.0006,
+ "loss": 6.33565616607666,
+ "step": 3271
+ },
+ {
+ "epoch": 45.4471615720524,
+ "grad_norm": 0.03230159357190132,
+ "learning_rate": 0.0006,
+ "loss": 6.197698593139648,
+ "step": 3272
+ },
+ {
+ "epoch": 45.46113537117904,
+ "grad_norm": 0.036897290498018265,
+ "learning_rate": 0.0006,
+ "loss": 6.241239547729492,
+ "step": 3273
+ },
+ {
+ "epoch": 45.47510917030568,
+ "grad_norm": 0.031103838235139847,
+ "learning_rate": 0.0006,
+ "loss": 6.159435749053955,
+ "step": 3274
+ },
+ {
+ "epoch": 45.48908296943232,
+ "grad_norm": 0.027116449549794197,
+ "learning_rate": 0.0006,
+ "loss": 6.212225914001465,
+ "step": 3275
+ },
+ {
+ "epoch": 45.50305676855895,
+ "grad_norm": 0.029923981055617332,
+ "learning_rate": 0.0006,
+ "loss": 6.157269477844238,
+ "step": 3276
+ },
+ {
+ "epoch": 45.51703056768559,
+ "grad_norm": 0.02521538734436035,
+ "learning_rate": 0.0006,
+ "loss": 6.146918773651123,
+ "step": 3277
+ },
+ {
+ "epoch": 45.531004366812226,
+ "grad_norm": 0.02758920192718506,
+ "learning_rate": 0.0006,
+ "loss": 6.0446977615356445,
+ "step": 3278
+ },
+ {
+ "epoch": 45.544978165938865,
+ "grad_norm": 0.024328220635652542,
+ "learning_rate": 0.0006,
+ "loss": 6.099508285522461,
+ "step": 3279
+ },
+ {
+ "epoch": 45.5589519650655,
+ "grad_norm": 0.027425771579146385,
+ "learning_rate": 0.0006,
+ "loss": 6.084299087524414,
+ "step": 3280
+ },
+ {
+ "epoch": 45.57292576419214,
+ "grad_norm": 0.026927651837468147,
+ "learning_rate": 0.0006,
+ "loss": 6.146334171295166,
+ "step": 3281
+ },
+ {
+ "epoch": 45.58689956331878,
+ "grad_norm": 0.04054379090666771,
+ "learning_rate": 0.0006,
+ "loss": 6.1301493644714355,
+ "step": 3282
+ },
+ {
+ "epoch": 45.60087336244541,
+ "grad_norm": 0.04456290975213051,
+ "learning_rate": 0.0006,
+ "loss": 6.152511119842529,
+ "step": 3283
+ },
+ {
+ "epoch": 45.61484716157205,
+ "grad_norm": 0.052821505814790726,
+ "learning_rate": 0.0006,
+ "loss": 6.113544940948486,
+ "step": 3284
+ },
+ {
+ "epoch": 45.62882096069869,
+ "grad_norm": 0.07251739501953125,
+ "learning_rate": 0.0006,
+ "loss": 6.118298530578613,
+ "step": 3285
+ },
+ {
+ "epoch": 45.64279475982533,
+ "grad_norm": 0.09433556348085403,
+ "learning_rate": 0.0006,
+ "loss": 6.170897483825684,
+ "step": 3286
+ },
+ {
+ "epoch": 45.65676855895197,
+ "grad_norm": 0.057683493942022324,
+ "learning_rate": 0.0006,
+ "loss": 6.1747822761535645,
+ "step": 3287
+ },
+ {
+ "epoch": 45.670742358078606,
+ "grad_norm": 0.06363270431756973,
+ "learning_rate": 0.0006,
+ "loss": 6.1617889404296875,
+ "step": 3288
+ },
+ {
+ "epoch": 45.68471615720524,
+ "grad_norm": 0.07528169453144073,
+ "learning_rate": 0.0006,
+ "loss": 6.230060577392578,
+ "step": 3289
+ },
+ {
+ "epoch": 45.698689956331876,
+ "grad_norm": 0.07816170156002045,
+ "learning_rate": 0.0006,
+ "loss": 6.171257019042969,
+ "step": 3290
+ },
+ {
+ "epoch": 45.712663755458514,
+ "grad_norm": 0.05860164389014244,
+ "learning_rate": 0.0006,
+ "loss": 6.095484256744385,
+ "step": 3291
+ },
+ {
+ "epoch": 45.72663755458515,
+ "grad_norm": 0.04747429117560387,
+ "learning_rate": 0.0006,
+ "loss": 6.060067176818848,
+ "step": 3292
+ },
+ {
+ "epoch": 45.74061135371179,
+ "grad_norm": 0.04514474794268608,
+ "learning_rate": 0.0006,
+ "loss": 6.100218772888184,
+ "step": 3293
+ },
+ {
+ "epoch": 45.75458515283843,
+ "grad_norm": 0.05168681964278221,
+ "learning_rate": 0.0006,
+ "loss": 6.0621867179870605,
+ "step": 3294
+ },
+ {
+ "epoch": 45.76855895196506,
+ "grad_norm": 0.04665876924991608,
+ "learning_rate": 0.0006,
+ "loss": 6.118292808532715,
+ "step": 3295
+ },
+ {
+ "epoch": 45.7825327510917,
+ "grad_norm": 0.049374762922525406,
+ "learning_rate": 0.0006,
+ "loss": 6.098394393920898,
+ "step": 3296
+ },
+ {
+ "epoch": 45.79650655021834,
+ "grad_norm": 0.03793549910187721,
+ "learning_rate": 0.0006,
+ "loss": 6.069541931152344,
+ "step": 3297
+ },
+ {
+ "epoch": 45.81048034934498,
+ "grad_norm": 0.036362383514642715,
+ "learning_rate": 0.0006,
+ "loss": 6.1188435554504395,
+ "step": 3298
+ },
+ {
+ "epoch": 45.82445414847162,
+ "grad_norm": 0.03397097811102867,
+ "learning_rate": 0.0006,
+ "loss": 6.117589950561523,
+ "step": 3299
+ },
+ {
+ "epoch": 45.838427947598255,
+ "grad_norm": 0.030977753922343254,
+ "learning_rate": 0.0006,
+ "loss": 6.0236711502075195,
+ "step": 3300
+ },
+ {
+ "epoch": 45.852401746724894,
+ "grad_norm": 0.02800217643380165,
+ "learning_rate": 0.0006,
+ "loss": 6.034825325012207,
+ "step": 3301
+ },
+ {
+ "epoch": 45.866375545851525,
+ "grad_norm": 0.03129778802394867,
+ "learning_rate": 0.0006,
+ "loss": 5.980271339416504,
+ "step": 3302
+ },
+ {
+ "epoch": 45.880349344978164,
+ "grad_norm": 0.025886094197630882,
+ "learning_rate": 0.0006,
+ "loss": 5.966419219970703,
+ "step": 3303
+ },
+ {
+ "epoch": 45.8943231441048,
+ "grad_norm": 0.025712328031659126,
+ "learning_rate": 0.0006,
+ "loss": 5.938652515411377,
+ "step": 3304
+ },
+ {
+ "epoch": 45.90829694323144,
+ "grad_norm": 0.022508298978209496,
+ "learning_rate": 0.0006,
+ "loss": 5.982312202453613,
+ "step": 3305
+ },
+ {
+ "epoch": 45.92227074235808,
+ "grad_norm": 0.024880951270461082,
+ "learning_rate": 0.0006,
+ "loss": 5.931150436401367,
+ "step": 3306
+ },
+ {
+ "epoch": 45.93624454148472,
+ "grad_norm": 0.030007576569914818,
+ "learning_rate": 0.0006,
+ "loss": 5.936639308929443,
+ "step": 3307
+ },
+ {
+ "epoch": 45.95021834061135,
+ "grad_norm": 0.033314041793346405,
+ "learning_rate": 0.0006,
+ "loss": 5.944248676300049,
+ "step": 3308
+ },
+ {
+ "epoch": 45.96419213973799,
+ "grad_norm": 0.043078381568193436,
+ "learning_rate": 0.0006,
+ "loss": 5.971286773681641,
+ "step": 3309
+ },
+ {
+ "epoch": 45.97816593886463,
+ "grad_norm": 0.059296440333127975,
+ "learning_rate": 0.0006,
+ "loss": 5.910497665405273,
+ "step": 3310
+ },
+ {
+ "epoch": 45.992139737991266,
+ "grad_norm": 0.06232666224241257,
+ "learning_rate": 0.0006,
+ "loss": 5.929817199707031,
+ "step": 3311
+ },
+ {
+ "epoch": 46.0,
+ "grad_norm": 0.05303940549492836,
+ "learning_rate": 0.0006,
+ "loss": 5.910822868347168,
+ "step": 3312
+ },
+ {
+ "epoch": 46.0,
+ "eval_loss": 5.977851390838623,
+ "eval_runtime": 56.2435,
+ "eval_samples_per_second": 43.418,
+ "eval_steps_per_second": 1.369,
+ "step": 3312
+ },
+ {
+ "epoch": 46.01397379912664,
+ "grad_norm": 0.06943798810243607,
+ "learning_rate": 0.0006,
+ "loss": 5.854922294616699,
+ "step": 3313
+ },
+ {
+ "epoch": 46.02794759825328,
+ "grad_norm": 0.05236278846859932,
+ "learning_rate": 0.0006,
+ "loss": 5.876611232757568,
+ "step": 3314
+ },
+ {
+ "epoch": 46.041921397379916,
+ "grad_norm": 0.06240001320838928,
+ "learning_rate": 0.0006,
+ "loss": 5.878912448883057,
+ "step": 3315
+ },
+ {
+ "epoch": 46.05589519650655,
+ "grad_norm": 0.07401192933320999,
+ "learning_rate": 0.0006,
+ "loss": 5.985136032104492,
+ "step": 3316
+ },
+ {
+ "epoch": 46.069868995633186,
+ "grad_norm": 0.0826529711484909,
+ "learning_rate": 0.0006,
+ "loss": 5.939439296722412,
+ "step": 3317
+ },
+ {
+ "epoch": 46.083842794759825,
+ "grad_norm": 0.07999005168676376,
+ "learning_rate": 0.0006,
+ "loss": 5.982336044311523,
+ "step": 3318
+ },
+ {
+ "epoch": 46.09781659388646,
+ "grad_norm": 0.07014869153499603,
+ "learning_rate": 0.0006,
+ "loss": 5.896534442901611,
+ "step": 3319
+ },
+ {
+ "epoch": 46.1117903930131,
+ "grad_norm": 0.05795128643512726,
+ "learning_rate": 0.0006,
+ "loss": 5.843111991882324,
+ "step": 3320
+ },
+ {
+ "epoch": 46.12576419213974,
+ "grad_norm": 0.06374258548021317,
+ "learning_rate": 0.0006,
+ "loss": 5.78752326965332,
+ "step": 3321
+ },
+ {
+ "epoch": 46.13973799126637,
+ "grad_norm": 0.045524805784225464,
+ "learning_rate": 0.0006,
+ "loss": 5.86755895614624,
+ "step": 3322
+ },
+ {
+ "epoch": 46.15371179039301,
+ "grad_norm": 0.03773504123091698,
+ "learning_rate": 0.0006,
+ "loss": 5.83627462387085,
+ "step": 3323
+ },
+ {
+ "epoch": 46.16768558951965,
+ "grad_norm": 0.035815201699733734,
+ "learning_rate": 0.0006,
+ "loss": 5.798238277435303,
+ "step": 3324
+ },
+ {
+ "epoch": 46.18165938864629,
+ "grad_norm": 0.03756783902645111,
+ "learning_rate": 0.0006,
+ "loss": 5.845197677612305,
+ "step": 3325
+ },
+ {
+ "epoch": 46.19563318777293,
+ "grad_norm": 0.03569113463163376,
+ "learning_rate": 0.0006,
+ "loss": 5.888932704925537,
+ "step": 3326
+ },
+ {
+ "epoch": 46.209606986899566,
+ "grad_norm": 0.029467813670635223,
+ "learning_rate": 0.0006,
+ "loss": 5.792047500610352,
+ "step": 3327
+ },
+ {
+ "epoch": 46.223580786026204,
+ "grad_norm": 0.025698406621813774,
+ "learning_rate": 0.0006,
+ "loss": 5.783734321594238,
+ "step": 3328
+ },
+ {
+ "epoch": 46.237554585152836,
+ "grad_norm": 0.028426671400666237,
+ "learning_rate": 0.0006,
+ "loss": 5.762084007263184,
+ "step": 3329
+ },
+ {
+ "epoch": 46.251528384279474,
+ "grad_norm": 0.03099622018635273,
+ "learning_rate": 0.0006,
+ "loss": 5.719109058380127,
+ "step": 3330
+ },
+ {
+ "epoch": 46.26550218340611,
+ "grad_norm": 0.0310219656676054,
+ "learning_rate": 0.0006,
+ "loss": 5.737057209014893,
+ "step": 3331
+ },
+ {
+ "epoch": 46.27947598253275,
+ "grad_norm": 0.02713446132838726,
+ "learning_rate": 0.0006,
+ "loss": 5.758657455444336,
+ "step": 3332
+ },
+ {
+ "epoch": 46.29344978165939,
+ "grad_norm": 0.03099844604730606,
+ "learning_rate": 0.0006,
+ "loss": 5.742086410522461,
+ "step": 3333
+ },
+ {
+ "epoch": 46.30742358078603,
+ "grad_norm": 0.030258730053901672,
+ "learning_rate": 0.0006,
+ "loss": 5.767426490783691,
+ "step": 3334
+ },
+ {
+ "epoch": 46.32139737991266,
+ "grad_norm": 0.029034852981567383,
+ "learning_rate": 0.0006,
+ "loss": 5.651792049407959,
+ "step": 3335
+ },
+ {
+ "epoch": 46.3353711790393,
+ "grad_norm": 0.03138665854930878,
+ "learning_rate": 0.0006,
+ "loss": 5.664980888366699,
+ "step": 3336
+ },
+ {
+ "epoch": 46.34934497816594,
+ "grad_norm": 0.040546901524066925,
+ "learning_rate": 0.0006,
+ "loss": 5.665165901184082,
+ "step": 3337
+ },
+ {
+ "epoch": 46.36331877729258,
+ "grad_norm": 0.06395485252141953,
+ "learning_rate": 0.0006,
+ "loss": 5.7776713371276855,
+ "step": 3338
+ },
+ {
+ "epoch": 46.377292576419215,
+ "grad_norm": 0.10088056325912476,
+ "learning_rate": 0.0006,
+ "loss": 5.755645751953125,
+ "step": 3339
+ },
+ {
+ "epoch": 46.391266375545854,
+ "grad_norm": 0.09527938067913055,
+ "learning_rate": 0.0006,
+ "loss": 5.838809490203857,
+ "step": 3340
+ },
+ {
+ "epoch": 46.40524017467249,
+ "grad_norm": 0.0849648267030716,
+ "learning_rate": 0.0006,
+ "loss": 5.781932830810547,
+ "step": 3341
+ },
+ {
+ "epoch": 46.419213973799124,
+ "grad_norm": 0.10142666101455688,
+ "learning_rate": 0.0006,
+ "loss": 5.838970184326172,
+ "step": 3342
+ },
+ {
+ "epoch": 46.43318777292576,
+ "grad_norm": 0.42731454968452454,
+ "learning_rate": 0.0006,
+ "loss": 6.071902275085449,
+ "step": 3343
+ },
+ {
+ "epoch": 46.4471615720524,
+ "grad_norm": 0.13102488219738007,
+ "learning_rate": 0.0006,
+ "loss": 6.8635993003845215,
+ "step": 3344
+ },
+ {
+ "epoch": 46.46113537117904,
+ "grad_norm": 0.23691938817501068,
+ "learning_rate": 0.0006,
+ "loss": 7.313158988952637,
+ "step": 3345
+ },
+ {
+ "epoch": 46.47510917030568,
+ "grad_norm": 0.11670669168233871,
+ "learning_rate": 0.0006,
+ "loss": 7.240363121032715,
+ "step": 3346
+ },
+ {
+ "epoch": 46.48908296943232,
+ "grad_norm": 0.09985820949077606,
+ "learning_rate": 0.0006,
+ "loss": 7.113332271575928,
+ "step": 3347
+ },
+ {
+ "epoch": 46.50305676855895,
+ "grad_norm": 0.11406408250331879,
+ "learning_rate": 0.0006,
+ "loss": 7.051285743713379,
+ "step": 3348
+ },
+ {
+ "epoch": 46.51703056768559,
+ "grad_norm": 0.07900336384773254,
+ "learning_rate": 0.0006,
+ "loss": 7.014365196228027,
+ "step": 3349
+ },
+ {
+ "epoch": 46.531004366812226,
+ "grad_norm": 0.05380333960056305,
+ "learning_rate": 0.0006,
+ "loss": 6.8868818283081055,
+ "step": 3350
+ },
+ {
+ "epoch": 46.544978165938865,
+ "grad_norm": 0.05868449807167053,
+ "learning_rate": 0.0006,
+ "loss": 6.783599853515625,
+ "step": 3351
+ },
+ {
+ "epoch": 46.5589519650655,
+ "grad_norm": 0.04428785294294357,
+ "learning_rate": 0.0006,
+ "loss": 6.798511505126953,
+ "step": 3352
+ },
+ {
+ "epoch": 46.57292576419214,
+ "grad_norm": 0.055527277290821075,
+ "learning_rate": 0.0006,
+ "loss": 6.694507598876953,
+ "step": 3353
+ },
+ {
+ "epoch": 46.58689956331878,
+ "grad_norm": 0.03872432932257652,
+ "learning_rate": 0.0006,
+ "loss": 6.736631393432617,
+ "step": 3354
+ },
+ {
+ "epoch": 46.60087336244541,
+ "grad_norm": 0.04330779239535332,
+ "learning_rate": 0.0006,
+ "loss": 6.576806545257568,
+ "step": 3355
+ },
+ {
+ "epoch": 46.61484716157205,
+ "grad_norm": 0.036767736077308655,
+ "learning_rate": 0.0006,
+ "loss": 6.559871196746826,
+ "step": 3356
+ },
+ {
+ "epoch": 46.62882096069869,
+ "grad_norm": 0.039857108145952225,
+ "learning_rate": 0.0006,
+ "loss": 6.5749831199646,
+ "step": 3357
+ },
+ {
+ "epoch": 46.64279475982533,
+ "grad_norm": 0.048719972372055054,
+ "learning_rate": 0.0006,
+ "loss": 6.4915618896484375,
+ "step": 3358
+ },
+ {
+ "epoch": 46.65676855895197,
+ "grad_norm": 0.042849015444517136,
+ "learning_rate": 0.0006,
+ "loss": 6.413516044616699,
+ "step": 3359
+ },
+ {
+ "epoch": 46.670742358078606,
+ "grad_norm": 0.02871009334921837,
+ "learning_rate": 0.0006,
+ "loss": 6.320952415466309,
+ "step": 3360
+ },
+ {
+ "epoch": 46.68471615720524,
+ "grad_norm": 0.03193620592355728,
+ "learning_rate": 0.0006,
+ "loss": 6.4176435470581055,
+ "step": 3361
+ },
+ {
+ "epoch": 46.698689956331876,
+ "grad_norm": 0.027417024597525597,
+ "learning_rate": 0.0006,
+ "loss": 6.342319488525391,
+ "step": 3362
+ },
+ {
+ "epoch": 46.712663755458514,
+ "grad_norm": 0.03305937349796295,
+ "learning_rate": 0.0006,
+ "loss": 6.246112823486328,
+ "step": 3363
+ },
+ {
+ "epoch": 46.72663755458515,
+ "grad_norm": 0.031403183937072754,
+ "learning_rate": 0.0006,
+ "loss": 6.149503231048584,
+ "step": 3364
+ },
+ {
+ "epoch": 46.74061135371179,
+ "grad_norm": 0.031563788652420044,
+ "learning_rate": 0.0006,
+ "loss": 6.278808116912842,
+ "step": 3365
+ },
+ {
+ "epoch": 46.75458515283843,
+ "grad_norm": 0.028366589918732643,
+ "learning_rate": 0.0006,
+ "loss": 6.057036876678467,
+ "step": 3366
+ },
+ {
+ "epoch": 46.76855895196506,
+ "grad_norm": 0.023715803399682045,
+ "learning_rate": 0.0006,
+ "loss": 6.111324310302734,
+ "step": 3367
+ },
+ {
+ "epoch": 46.7825327510917,
+ "grad_norm": 0.026546811684966087,
+ "learning_rate": 0.0006,
+ "loss": 6.105669975280762,
+ "step": 3368
+ },
+ {
+ "epoch": 46.79650655021834,
+ "grad_norm": 0.023531019687652588,
+ "learning_rate": 0.0006,
+ "loss": 6.068072319030762,
+ "step": 3369
+ },
+ {
+ "epoch": 46.81048034934498,
+ "grad_norm": 0.024480128660798073,
+ "learning_rate": 0.0006,
+ "loss": 5.947896957397461,
+ "step": 3370
+ },
+ {
+ "epoch": 46.82445414847162,
+ "grad_norm": 0.02517310529947281,
+ "learning_rate": 0.0006,
+ "loss": 6.030037879943848,
+ "step": 3371
+ },
+ {
+ "epoch": 46.838427947598255,
+ "grad_norm": 0.026723403483629227,
+ "learning_rate": 0.0006,
+ "loss": 5.963776111602783,
+ "step": 3372
+ },
+ {
+ "epoch": 46.852401746724894,
+ "grad_norm": 0.03163504600524902,
+ "learning_rate": 0.0006,
+ "loss": 5.893466949462891,
+ "step": 3373
+ },
+ {
+ "epoch": 46.866375545851525,
+ "grad_norm": 0.04058683291077614,
+ "learning_rate": 0.0006,
+ "loss": 5.9202799797058105,
+ "step": 3374
+ },
+ {
+ "epoch": 46.880349344978164,
+ "grad_norm": 0.04964117333292961,
+ "learning_rate": 0.0006,
+ "loss": 5.9366559982299805,
+ "step": 3375
+ },
+ {
+ "epoch": 46.8943231441048,
+ "grad_norm": 0.055406033992767334,
+ "learning_rate": 0.0006,
+ "loss": 5.912230968475342,
+ "step": 3376
+ },
+ {
+ "epoch": 46.90829694323144,
+ "grad_norm": 0.04643625393509865,
+ "learning_rate": 0.0006,
+ "loss": 5.87260103225708,
+ "step": 3377
+ },
+ {
+ "epoch": 46.92227074235808,
+ "grad_norm": 0.029325876384973526,
+ "learning_rate": 0.0006,
+ "loss": 5.854763984680176,
+ "step": 3378
+ },
+ {
+ "epoch": 46.93624454148472,
+ "grad_norm": 0.03159382566809654,
+ "learning_rate": 0.0006,
+ "loss": 5.880882263183594,
+ "step": 3379
+ },
+ {
+ "epoch": 46.95021834061135,
+ "grad_norm": 0.030224012210965157,
+ "learning_rate": 0.0006,
+ "loss": 5.787360191345215,
+ "step": 3380
+ },
+ {
+ "epoch": 46.96419213973799,
+ "grad_norm": 0.028368130326271057,
+ "learning_rate": 0.0006,
+ "loss": 5.817195892333984,
+ "step": 3381
+ },
+ {
+ "epoch": 46.97816593886463,
+ "grad_norm": 0.029411399737000465,
+ "learning_rate": 0.0006,
+ "loss": 5.783403396606445,
+ "step": 3382
+ },
+ {
+ "epoch": 46.992139737991266,
+ "grad_norm": 0.029927462339401245,
+ "learning_rate": 0.0006,
+ "loss": 5.662444591522217,
+ "step": 3383
+ },
+ {
+ "epoch": 47.0,
+ "grad_norm": 0.02872961387038231,
+ "learning_rate": 0.0006,
+ "loss": 5.742781639099121,
+ "step": 3384
+ },
+ {
+ "epoch": 47.0,
+ "eval_loss": 5.870260238647461,
+ "eval_runtime": 57.3734,
+ "eval_samples_per_second": 42.563,
+ "eval_steps_per_second": 1.342,
+ "step": 3384
+ },
+ {
+ "epoch": 47.01397379912664,
+ "grad_norm": 0.028542131185531616,
+ "learning_rate": 0.0006,
+ "loss": 5.753256797790527,
+ "step": 3385
+ },
+ {
+ "epoch": 47.02794759825328,
+ "grad_norm": 0.03227119892835617,
+ "learning_rate": 0.0006,
+ "loss": 5.722246170043945,
+ "step": 3386
+ },
+ {
+ "epoch": 47.041921397379916,
+ "grad_norm": 0.031461723148822784,
+ "learning_rate": 0.0006,
+ "loss": 5.740520000457764,
+ "step": 3387
+ },
+ {
+ "epoch": 47.05589519650655,
+ "grad_norm": 0.03184220567345619,
+ "learning_rate": 0.0006,
+ "loss": 5.720890045166016,
+ "step": 3388
+ },
+ {
+ "epoch": 47.069868995633186,
+ "grad_norm": 0.030015580356121063,
+ "learning_rate": 0.0006,
+ "loss": 5.597814559936523,
+ "step": 3389
+ },
+ {
+ "epoch": 47.083842794759825,
+ "grad_norm": 0.029931779950857162,
+ "learning_rate": 0.0006,
+ "loss": 5.5412468910217285,
+ "step": 3390
+ },
+ {
+ "epoch": 47.09781659388646,
+ "grad_norm": 0.03206505626440048,
+ "learning_rate": 0.0006,
+ "loss": 5.6967597007751465,
+ "step": 3391
+ },
+ {
+ "epoch": 47.1117903930131,
+ "grad_norm": 0.03978161886334419,
+ "learning_rate": 0.0006,
+ "loss": 5.59661340713501,
+ "step": 3392
+ },
+ {
+ "epoch": 47.12576419213974,
+ "grad_norm": 0.04619080200791359,
+ "learning_rate": 0.0006,
+ "loss": 5.795788764953613,
+ "step": 3393
+ },
+ {
+ "epoch": 47.13973799126637,
+ "grad_norm": 0.049085382372140884,
+ "learning_rate": 0.0006,
+ "loss": 5.667335510253906,
+ "step": 3394
+ },
+ {
+ "epoch": 47.15371179039301,
+ "grad_norm": 0.04830858111381531,
+ "learning_rate": 0.0006,
+ "loss": 5.668806552886963,
+ "step": 3395
+ },
+ {
+ "epoch": 47.16768558951965,
+ "grad_norm": 0.0583178736269474,
+ "learning_rate": 0.0006,
+ "loss": 5.692285537719727,
+ "step": 3396
+ },
+ {
+ "epoch": 47.18165938864629,
+ "grad_norm": 0.05182253196835518,
+ "learning_rate": 0.0006,
+ "loss": 5.678645133972168,
+ "step": 3397
+ },
+ {
+ "epoch": 47.19563318777293,
+ "grad_norm": 0.04011905938386917,
+ "learning_rate": 0.0006,
+ "loss": 5.661550998687744,
+ "step": 3398
+ },
+ {
+ "epoch": 47.209606986899566,
+ "grad_norm": 0.03888486698269844,
+ "learning_rate": 0.0006,
+ "loss": 5.643001556396484,
+ "step": 3399
+ },
+ {
+ "epoch": 47.223580786026204,
+ "grad_norm": 0.030658265575766563,
+ "learning_rate": 0.0006,
+ "loss": 5.659090995788574,
+ "step": 3400
+ },
+ {
+ "epoch": 47.237554585152836,
+ "grad_norm": 0.03609256446361542,
+ "learning_rate": 0.0006,
+ "loss": 5.544501781463623,
+ "step": 3401
+ },
+ {
+ "epoch": 47.251528384279474,
+ "grad_norm": 0.033589769154787064,
+ "learning_rate": 0.0006,
+ "loss": 5.463878631591797,
+ "step": 3402
+ },
+ {
+ "epoch": 47.26550218340611,
+ "grad_norm": 0.035128287971019745,
+ "learning_rate": 0.0006,
+ "loss": 5.5680060386657715,
+ "step": 3403
+ },
+ {
+ "epoch": 47.27947598253275,
+ "grad_norm": 0.038684550672769547,
+ "learning_rate": 0.0006,
+ "loss": 5.587898254394531,
+ "step": 3404
+ },
+ {
+ "epoch": 47.29344978165939,
+ "grad_norm": 0.043566226959228516,
+ "learning_rate": 0.0006,
+ "loss": 5.580896854400635,
+ "step": 3405
+ },
+ {
+ "epoch": 47.30742358078603,
+ "grad_norm": 0.04428374394774437,
+ "learning_rate": 0.0006,
+ "loss": 5.603460311889648,
+ "step": 3406
+ },
+ {
+ "epoch": 47.32139737991266,
+ "grad_norm": 0.037368934601545334,
+ "learning_rate": 0.0006,
+ "loss": 5.564740180969238,
+ "step": 3407
+ },
+ {
+ "epoch": 47.3353711790393,
+ "grad_norm": 0.04092350974678993,
+ "learning_rate": 0.0006,
+ "loss": 5.584153652191162,
+ "step": 3408
+ },
+ {
+ "epoch": 47.34934497816594,
+ "grad_norm": 0.045031506568193436,
+ "learning_rate": 0.0006,
+ "loss": 5.601186275482178,
+ "step": 3409
+ },
+ {
+ "epoch": 47.36331877729258,
+ "grad_norm": 0.041830409318208694,
+ "learning_rate": 0.0006,
+ "loss": 5.512788772583008,
+ "step": 3410
+ },
+ {
+ "epoch": 47.377292576419215,
+ "grad_norm": 0.037633974105119705,
+ "learning_rate": 0.0006,
+ "loss": 5.503828525543213,
+ "step": 3411
+ },
+ {
+ "epoch": 47.391266375545854,
+ "grad_norm": 0.03473588079214096,
+ "learning_rate": 0.0006,
+ "loss": 5.570557594299316,
+ "step": 3412
+ },
+ {
+ "epoch": 47.40524017467249,
+ "grad_norm": 0.039325181394815445,
+ "learning_rate": 0.0006,
+ "loss": 5.427783966064453,
+ "step": 3413
+ },
+ {
+ "epoch": 47.419213973799124,
+ "grad_norm": 0.04034624621272087,
+ "learning_rate": 0.0006,
+ "loss": 5.572591304779053,
+ "step": 3414
+ },
+ {
+ "epoch": 47.43318777292576,
+ "grad_norm": 0.05245184153318405,
+ "learning_rate": 0.0006,
+ "loss": 5.505208969116211,
+ "step": 3415
+ },
+ {
+ "epoch": 47.4471615720524,
+ "grad_norm": 0.0568675734102726,
+ "learning_rate": 0.0006,
+ "loss": 5.589373588562012,
+ "step": 3416
+ },
+ {
+ "epoch": 47.46113537117904,
+ "grad_norm": 0.048336587846279144,
+ "learning_rate": 0.0006,
+ "loss": 5.522005081176758,
+ "step": 3417
+ },
+ {
+ "epoch": 47.47510917030568,
+ "grad_norm": 0.04194975271821022,
+ "learning_rate": 0.0006,
+ "loss": 5.387576103210449,
+ "step": 3418
+ },
+ {
+ "epoch": 47.48908296943232,
+ "grad_norm": 0.04019433259963989,
+ "learning_rate": 0.0006,
+ "loss": 5.522615432739258,
+ "step": 3419
+ },
+ {
+ "epoch": 47.50305676855895,
+ "grad_norm": 0.05709933117032051,
+ "learning_rate": 0.0006,
+ "loss": 5.420286178588867,
+ "step": 3420
+ },
+ {
+ "epoch": 47.51703056768559,
+ "grad_norm": 0.09130338579416275,
+ "learning_rate": 0.0006,
+ "loss": 5.484622478485107,
+ "step": 3421
+ },
+ {
+ "epoch": 47.531004366812226,
+ "grad_norm": 0.2757541835308075,
+ "learning_rate": 0.0006,
+ "loss": 5.821832656860352,
+ "step": 3422
+ },
+ {
+ "epoch": 47.544978165938865,
+ "grad_norm": 0.28292080760002136,
+ "learning_rate": 0.0006,
+ "loss": 6.579934120178223,
+ "step": 3423
+ },
+ {
+ "epoch": 47.5589519650655,
+ "grad_norm": 0.15293604135513306,
+ "learning_rate": 0.0006,
+ "loss": 6.478213310241699,
+ "step": 3424
+ },
+ {
+ "epoch": 47.57292576419214,
+ "grad_norm": 0.11284393817186356,
+ "learning_rate": 0.0006,
+ "loss": 6.4066386222839355,
+ "step": 3425
+ },
+ {
+ "epoch": 47.58689956331878,
+ "grad_norm": 0.10933073610067368,
+ "learning_rate": 0.0006,
+ "loss": 6.3306169509887695,
+ "step": 3426
+ },
+ {
+ "epoch": 47.60087336244541,
+ "grad_norm": 0.05837641656398773,
+ "learning_rate": 0.0006,
+ "loss": 6.250000953674316,
+ "step": 3427
+ },
+ {
+ "epoch": 47.61484716157205,
+ "grad_norm": 0.05987229570746422,
+ "learning_rate": 0.0006,
+ "loss": 6.152463436126709,
+ "step": 3428
+ },
+ {
+ "epoch": 47.62882096069869,
+ "grad_norm": 0.04064938426017761,
+ "learning_rate": 0.0006,
+ "loss": 6.117897033691406,
+ "step": 3429
+ },
+ {
+ "epoch": 47.64279475982533,
+ "grad_norm": 0.04565057530999184,
+ "learning_rate": 0.0006,
+ "loss": 6.042291641235352,
+ "step": 3430
+ },
+ {
+ "epoch": 47.65676855895197,
+ "grad_norm": 0.03475063294172287,
+ "learning_rate": 0.0006,
+ "loss": 6.0071868896484375,
+ "step": 3431
+ },
+ {
+ "epoch": 47.670742358078606,
+ "grad_norm": 0.03809220343828201,
+ "learning_rate": 0.0006,
+ "loss": 6.0128021240234375,
+ "step": 3432
+ },
+ {
+ "epoch": 47.68471615720524,
+ "grad_norm": 0.03307219594717026,
+ "learning_rate": 0.0006,
+ "loss": 5.894664287567139,
+ "step": 3433
+ },
+ {
+ "epoch": 47.698689956331876,
+ "grad_norm": 0.028824346140027046,
+ "learning_rate": 0.0006,
+ "loss": 5.914865493774414,
+ "step": 3434
+ },
+ {
+ "epoch": 47.712663755458514,
+ "grad_norm": 0.027923690155148506,
+ "learning_rate": 0.0006,
+ "loss": 5.851759910583496,
+ "step": 3435
+ },
+ {
+ "epoch": 47.72663755458515,
+ "grad_norm": 0.025492684915661812,
+ "learning_rate": 0.0006,
+ "loss": 5.749186038970947,
+ "step": 3436
+ },
+ {
+ "epoch": 47.74061135371179,
+ "grad_norm": 0.025105174630880356,
+ "learning_rate": 0.0006,
+ "loss": 5.803566932678223,
+ "step": 3437
+ },
+ {
+ "epoch": 47.75458515283843,
+ "grad_norm": 0.023298880085349083,
+ "learning_rate": 0.0006,
+ "loss": 5.76276159286499,
+ "step": 3438
+ },
+ {
+ "epoch": 47.76855895196506,
+ "grad_norm": 0.02468576282262802,
+ "learning_rate": 0.0006,
+ "loss": 5.833852767944336,
+ "step": 3439
+ },
+ {
+ "epoch": 47.7825327510917,
+ "grad_norm": 0.022730689495801926,
+ "learning_rate": 0.0006,
+ "loss": 5.736724376678467,
+ "step": 3440
+ },
+ {
+ "epoch": 47.79650655021834,
+ "grad_norm": 0.021335354074835777,
+ "learning_rate": 0.0006,
+ "loss": 5.644248008728027,
+ "step": 3441
+ },
+ {
+ "epoch": 47.81048034934498,
+ "grad_norm": 0.02288101427257061,
+ "learning_rate": 0.0006,
+ "loss": 5.7011942863464355,
+ "step": 3442
+ },
+ {
+ "epoch": 47.82445414847162,
+ "grad_norm": 0.01961585320532322,
+ "learning_rate": 0.0006,
+ "loss": 5.63215446472168,
+ "step": 3443
+ },
+ {
+ "epoch": 47.838427947598255,
+ "grad_norm": 0.017659205943346024,
+ "learning_rate": 0.0006,
+ "loss": 5.744845390319824,
+ "step": 3444
+ },
+ {
+ "epoch": 47.852401746724894,
+ "grad_norm": 0.02281784638762474,
+ "learning_rate": 0.0006,
+ "loss": 5.696993350982666,
+ "step": 3445
+ },
+ {
+ "epoch": 47.866375545851525,
+ "grad_norm": 0.016394242644309998,
+ "learning_rate": 0.0006,
+ "loss": 5.659870147705078,
+ "step": 3446
+ },
+ {
+ "epoch": 47.880349344978164,
+ "grad_norm": 0.01967368647456169,
+ "learning_rate": 0.0006,
+ "loss": 5.638072490692139,
+ "step": 3447
+ },
+ {
+ "epoch": 47.8943231441048,
+ "grad_norm": 0.017326267436146736,
+ "learning_rate": 0.0006,
+ "loss": 5.638836860656738,
+ "step": 3448
+ },
+ {
+ "epoch": 47.90829694323144,
+ "grad_norm": 0.017778463661670685,
+ "learning_rate": 0.0006,
+ "loss": 5.634466648101807,
+ "step": 3449
+ },
+ {
+ "epoch": 47.92227074235808,
+ "grad_norm": 0.017269758507609367,
+ "learning_rate": 0.0006,
+ "loss": 5.506277561187744,
+ "step": 3450
+ },
+ {
+ "epoch": 47.93624454148472,
+ "grad_norm": 0.01733289286494255,
+ "learning_rate": 0.0006,
+ "loss": 5.544455528259277,
+ "step": 3451
+ },
+ {
+ "epoch": 47.95021834061135,
+ "grad_norm": 0.017787795513868332,
+ "learning_rate": 0.0006,
+ "loss": 5.498943328857422,
+ "step": 3452
+ },
+ {
+ "epoch": 47.96419213973799,
+ "grad_norm": 0.015770932659506798,
+ "learning_rate": 0.0006,
+ "loss": 5.537227630615234,
+ "step": 3453
+ },
+ {
+ "epoch": 47.97816593886463,
+ "grad_norm": 0.016132809221744537,
+ "learning_rate": 0.0006,
+ "loss": 5.489081382751465,
+ "step": 3454
+ },
+ {
+ "epoch": 47.992139737991266,
+ "grad_norm": 0.01749664731323719,
+ "learning_rate": 0.0006,
+ "loss": 5.454263687133789,
+ "step": 3455
+ },
+ {
+ "epoch": 48.0,
+ "grad_norm": 0.016711296513676643,
+ "learning_rate": 0.0006,
+ "loss": 5.530783653259277,
+ "step": 3456
+ },
+ {
+ "epoch": 48.0,
+ "eval_loss": 5.596925258636475,
+ "eval_runtime": 56.9612,
+ "eval_samples_per_second": 42.871,
+ "eval_steps_per_second": 1.352,
+ "step": 3456
+ },
+ {
+ "epoch": 48.01397379912664,
+ "grad_norm": 0.01867641694843769,
+ "learning_rate": 0.0006,
+ "loss": 5.506853103637695,
+ "step": 3457
+ },
+ {
+ "epoch": 48.02794759825328,
+ "grad_norm": 0.023250792175531387,
+ "learning_rate": 0.0006,
+ "loss": 5.4074177742004395,
+ "step": 3458
+ },
+ {
+ "epoch": 48.041921397379916,
+ "grad_norm": 0.01970936730504036,
+ "learning_rate": 0.0006,
+ "loss": 5.508777618408203,
+ "step": 3459
+ },
+ {
+ "epoch": 48.05589519650655,
+ "grad_norm": 0.01879260502755642,
+ "learning_rate": 0.0006,
+ "loss": 5.412424087524414,
+ "step": 3460
+ },
+ {
+ "epoch": 48.069868995633186,
+ "grad_norm": 0.017420092597603798,
+ "learning_rate": 0.0006,
+ "loss": 5.43361759185791,
+ "step": 3461
+ },
+ {
+ "epoch": 48.083842794759825,
+ "grad_norm": 0.018707862123847008,
+ "learning_rate": 0.0006,
+ "loss": 5.476800918579102,
+ "step": 3462
+ },
+ {
+ "epoch": 48.09781659388646,
+ "grad_norm": 0.023185882717370987,
+ "learning_rate": 0.0006,
+ "loss": 5.465272426605225,
+ "step": 3463
+ },
+ {
+ "epoch": 48.1117903930131,
+ "grad_norm": 0.023570368066430092,
+ "learning_rate": 0.0006,
+ "loss": 5.391471862792969,
+ "step": 3464
+ },
+ {
+ "epoch": 48.12576419213974,
+ "grad_norm": 0.02651144750416279,
+ "learning_rate": 0.0006,
+ "loss": 5.422799110412598,
+ "step": 3465
+ },
+ {
+ "epoch": 48.13973799126637,
+ "grad_norm": 0.03316329047083855,
+ "learning_rate": 0.0006,
+ "loss": 5.3952202796936035,
+ "step": 3466
+ },
+ {
+ "epoch": 48.15371179039301,
+ "grad_norm": 0.04012278467416763,
+ "learning_rate": 0.0006,
+ "loss": 5.540853023529053,
+ "step": 3467
+ },
+ {
+ "epoch": 48.16768558951965,
+ "grad_norm": 0.04763203114271164,
+ "learning_rate": 0.0006,
+ "loss": 5.387890338897705,
+ "step": 3468
+ },
+ {
+ "epoch": 48.18165938864629,
+ "grad_norm": 0.05129165202379227,
+ "learning_rate": 0.0006,
+ "loss": 5.414722442626953,
+ "step": 3469
+ },
+ {
+ "epoch": 48.19563318777293,
+ "grad_norm": 0.04438071325421333,
+ "learning_rate": 0.0006,
+ "loss": 5.386244773864746,
+ "step": 3470
+ },
+ {
+ "epoch": 48.209606986899566,
+ "grad_norm": 0.040234919637441635,
+ "learning_rate": 0.0006,
+ "loss": 5.431617736816406,
+ "step": 3471
+ },
+ {
+ "epoch": 48.223580786026204,
+ "grad_norm": 0.03217008337378502,
+ "learning_rate": 0.0006,
+ "loss": 5.30168342590332,
+ "step": 3472
+ },
+ {
+ "epoch": 48.237554585152836,
+ "grad_norm": 0.03087974339723587,
+ "learning_rate": 0.0006,
+ "loss": 5.316389083862305,
+ "step": 3473
+ },
+ {
+ "epoch": 48.251528384279474,
+ "grad_norm": 0.02789347991347313,
+ "learning_rate": 0.0006,
+ "loss": 5.385970592498779,
+ "step": 3474
+ },
+ {
+ "epoch": 48.26550218340611,
+ "grad_norm": 0.024792097508907318,
+ "learning_rate": 0.0006,
+ "loss": 5.362329959869385,
+ "step": 3475
+ },
+ {
+ "epoch": 48.27947598253275,
+ "grad_norm": 0.025869522243738174,
+ "learning_rate": 0.0006,
+ "loss": 5.28400182723999,
+ "step": 3476
+ },
+ {
+ "epoch": 48.29344978165939,
+ "grad_norm": 0.021339882165193558,
+ "learning_rate": 0.0006,
+ "loss": 5.36669921875,
+ "step": 3477
+ },
+ {
+ "epoch": 48.30742358078603,
+ "grad_norm": 0.021462520584464073,
+ "learning_rate": 0.0006,
+ "loss": 5.311222076416016,
+ "step": 3478
+ },
+ {
+ "epoch": 48.32139737991266,
+ "grad_norm": 0.02193879708647728,
+ "learning_rate": 0.0006,
+ "loss": 5.291441917419434,
+ "step": 3479
+ },
+ {
+ "epoch": 48.3353711790393,
+ "grad_norm": 0.01963093690574169,
+ "learning_rate": 0.0006,
+ "loss": 5.342780113220215,
+ "step": 3480
+ },
+ {
+ "epoch": 48.34934497816594,
+ "grad_norm": 0.020634736865758896,
+ "learning_rate": 0.0006,
+ "loss": 5.144141674041748,
+ "step": 3481
+ },
+ {
+ "epoch": 48.36331877729258,
+ "grad_norm": 0.02126968279480934,
+ "learning_rate": 0.0006,
+ "loss": 5.232624530792236,
+ "step": 3482
+ },
+ {
+ "epoch": 48.377292576419215,
+ "grad_norm": 0.021203994750976562,
+ "learning_rate": 0.0006,
+ "loss": 5.290219783782959,
+ "step": 3483
+ },
+ {
+ "epoch": 48.391266375545854,
+ "grad_norm": 0.02193225920200348,
+ "learning_rate": 0.0006,
+ "loss": 5.29197359085083,
+ "step": 3484
+ },
+ {
+ "epoch": 48.40524017467249,
+ "grad_norm": 0.023174211382865906,
+ "learning_rate": 0.0006,
+ "loss": 5.204349517822266,
+ "step": 3485
+ },
+ {
+ "epoch": 48.419213973799124,
+ "grad_norm": 0.024161914363503456,
+ "learning_rate": 0.0006,
+ "loss": 5.339351654052734,
+ "step": 3486
+ },
+ {
+ "epoch": 48.43318777292576,
+ "grad_norm": 0.02221524342894554,
+ "learning_rate": 0.0006,
+ "loss": 5.263001441955566,
+ "step": 3487
+ },
+ {
+ "epoch": 48.4471615720524,
+ "grad_norm": 0.026632830500602722,
+ "learning_rate": 0.0006,
+ "loss": 5.221281051635742,
+ "step": 3488
+ },
+ {
+ "epoch": 48.46113537117904,
+ "grad_norm": 0.034923870116472244,
+ "learning_rate": 0.0006,
+ "loss": 5.296039581298828,
+ "step": 3489
+ },
+ {
+ "epoch": 48.47510917030568,
+ "grad_norm": 0.046184979379177094,
+ "learning_rate": 0.0006,
+ "loss": 5.212769985198975,
+ "step": 3490
+ },
+ {
+ "epoch": 48.48908296943232,
+ "grad_norm": 0.05154098570346832,
+ "learning_rate": 0.0006,
+ "loss": 5.337864875793457,
+ "step": 3491
+ },
+ {
+ "epoch": 48.50305676855895,
+ "grad_norm": 0.042967259883880615,
+ "learning_rate": 0.0006,
+ "loss": 5.267958641052246,
+ "step": 3492
+ },
+ {
+ "epoch": 48.51703056768559,
+ "grad_norm": 0.05356583371758461,
+ "learning_rate": 0.0006,
+ "loss": 5.1404218673706055,
+ "step": 3493
+ },
+ {
+ "epoch": 48.531004366812226,
+ "grad_norm": 0.08219032734632492,
+ "learning_rate": 0.0006,
+ "loss": 5.300687789916992,
+ "step": 3494
+ },
+ {
+ "epoch": 48.544978165938865,
+ "grad_norm": 0.19204089045524597,
+ "learning_rate": 0.0006,
+ "loss": 5.456025123596191,
+ "step": 3495
+ },
+ {
+ "epoch": 48.5589519650655,
+ "grad_norm": 0.25917696952819824,
+ "learning_rate": 0.0006,
+ "loss": 6.156761169433594,
+ "step": 3496
+ },
+ {
+ "epoch": 48.57292576419214,
+ "grad_norm": 0.1566283404827118,
+ "learning_rate": 0.0006,
+ "loss": 5.862346649169922,
+ "step": 3497
+ },
+ {
+ "epoch": 48.58689956331878,
+ "grad_norm": 0.1761486828327179,
+ "learning_rate": 0.0006,
+ "loss": 6.012795448303223,
+ "step": 3498
+ },
+ {
+ "epoch": 48.60087336244541,
+ "grad_norm": 0.09857621043920517,
+ "learning_rate": 0.0006,
+ "loss": 5.9248151779174805,
+ "step": 3499
+ },
+ {
+ "epoch": 48.61484716157205,
+ "grad_norm": 0.05968055501580238,
+ "learning_rate": 0.0006,
+ "loss": 5.780706882476807,
+ "step": 3500
+ },
+ {
+ "epoch": 48.62882096069869,
+ "grad_norm": 0.057433564215898514,
+ "learning_rate": 0.0006,
+ "loss": 5.645388603210449,
+ "step": 3501
+ },
+ {
+ "epoch": 48.64279475982533,
+ "grad_norm": 0.039659615606069565,
+ "learning_rate": 0.0006,
+ "loss": 5.695138931274414,
+ "step": 3502
+ },
+ {
+ "epoch": 48.65676855895197,
+ "grad_norm": 0.04019695147871971,
+ "learning_rate": 0.0006,
+ "loss": 5.672435283660889,
+ "step": 3503
+ },
+ {
+ "epoch": 48.670742358078606,
+ "grad_norm": 0.03033018670976162,
+ "learning_rate": 0.0006,
+ "loss": 5.682260513305664,
+ "step": 3504
+ },
+ {
+ "epoch": 48.68471615720524,
+ "grad_norm": 0.03215787187218666,
+ "learning_rate": 0.0006,
+ "loss": 5.604182243347168,
+ "step": 3505
+ },
+ {
+ "epoch": 48.698689956331876,
+ "grad_norm": 0.026865126565098763,
+ "learning_rate": 0.0006,
+ "loss": 5.64040470123291,
+ "step": 3506
+ },
+ {
+ "epoch": 48.712663755458514,
+ "grad_norm": 0.02831609919667244,
+ "learning_rate": 0.0006,
+ "loss": 5.552720069885254,
+ "step": 3507
+ },
+ {
+ "epoch": 48.72663755458515,
+ "grad_norm": 0.024226613342761993,
+ "learning_rate": 0.0006,
+ "loss": 5.556946277618408,
+ "step": 3508
+ },
+ {
+ "epoch": 48.74061135371179,
+ "grad_norm": 0.025393934920430183,
+ "learning_rate": 0.0006,
+ "loss": 5.476178169250488,
+ "step": 3509
+ },
+ {
+ "epoch": 48.75458515283843,
+ "grad_norm": 0.01984412968158722,
+ "learning_rate": 0.0006,
+ "loss": 5.471088409423828,
+ "step": 3510
+ },
+ {
+ "epoch": 48.76855895196506,
+ "grad_norm": 0.02035129815340042,
+ "learning_rate": 0.0006,
+ "loss": 5.438812732696533,
+ "step": 3511
+ },
+ {
+ "epoch": 48.7825327510917,
+ "grad_norm": 0.0208350270986557,
+ "learning_rate": 0.0006,
+ "loss": 5.381746292114258,
+ "step": 3512
+ },
+ {
+ "epoch": 48.79650655021834,
+ "grad_norm": 0.017783882096409798,
+ "learning_rate": 0.0006,
+ "loss": 5.351675510406494,
+ "step": 3513
+ },
+ {
+ "epoch": 48.81048034934498,
+ "grad_norm": 0.01671770215034485,
+ "learning_rate": 0.0006,
+ "loss": 5.329428672790527,
+ "step": 3514
+ },
+ {
+ "epoch": 48.82445414847162,
+ "grad_norm": 0.017891215160489082,
+ "learning_rate": 0.0006,
+ "loss": 5.212074279785156,
+ "step": 3515
+ },
+ {
+ "epoch": 48.838427947598255,
+ "grad_norm": 0.016847791150212288,
+ "learning_rate": 0.0006,
+ "loss": 5.299534797668457,
+ "step": 3516
+ },
+ {
+ "epoch": 48.852401746724894,
+ "grad_norm": 0.01593637838959694,
+ "learning_rate": 0.0006,
+ "loss": 5.30466890335083,
+ "step": 3517
+ },
+ {
+ "epoch": 48.866375545851525,
+ "grad_norm": 0.014713102020323277,
+ "learning_rate": 0.0006,
+ "loss": 5.294345378875732,
+ "step": 3518
+ },
+ {
+ "epoch": 48.880349344978164,
+ "grad_norm": 0.01433857623487711,
+ "learning_rate": 0.0006,
+ "loss": 5.2148542404174805,
+ "step": 3519
+ },
+ {
+ "epoch": 48.8943231441048,
+ "grad_norm": 0.014979379251599312,
+ "learning_rate": 0.0006,
+ "loss": 5.298582077026367,
+ "step": 3520
+ },
+ {
+ "epoch": 48.90829694323144,
+ "grad_norm": 0.013494192622601986,
+ "learning_rate": 0.0006,
+ "loss": 5.309783935546875,
+ "step": 3521
+ },
+ {
+ "epoch": 48.92227074235808,
+ "grad_norm": 0.014020012691617012,
+ "learning_rate": 0.0006,
+ "loss": 5.284879684448242,
+ "step": 3522
+ },
+ {
+ "epoch": 48.93624454148472,
+ "grad_norm": 0.01348058134317398,
+ "learning_rate": 0.0006,
+ "loss": 5.249116897583008,
+ "step": 3523
+ },
+ {
+ "epoch": 48.95021834061135,
+ "grad_norm": 0.013280218467116356,
+ "learning_rate": 0.0006,
+ "loss": 5.273489952087402,
+ "step": 3524
+ },
+ {
+ "epoch": 48.96419213973799,
+ "grad_norm": 0.012647110037505627,
+ "learning_rate": 0.0006,
+ "loss": 5.306652069091797,
+ "step": 3525
+ },
+ {
+ "epoch": 48.97816593886463,
+ "grad_norm": 0.012847724370658398,
+ "learning_rate": 0.0006,
+ "loss": 5.235745429992676,
+ "step": 3526
+ },
+ {
+ "epoch": 48.992139737991266,
+ "grad_norm": 0.011820166371762753,
+ "learning_rate": 0.0006,
+ "loss": 5.238091468811035,
+ "step": 3527
+ },
+ {
+ "epoch": 49.0,
+ "grad_norm": 0.01255585066974163,
+ "learning_rate": 0.0006,
+ "loss": 5.231295585632324,
+ "step": 3528
+ },
+ {
+ "epoch": 49.0,
+ "eval_loss": 5.394587993621826,
+ "eval_runtime": 56.3734,
+ "eval_samples_per_second": 43.318,
+ "eval_steps_per_second": 1.366,
+ "step": 3528
+ },
+ {
+ "epoch": 49.01397379912664,
+ "grad_norm": 0.011031972244381905,
+ "learning_rate": 0.0006,
+ "loss": 5.1354570388793945,
+ "step": 3529
+ },
+ {
+ "epoch": 49.02794759825328,
+ "grad_norm": 0.010631263256072998,
+ "learning_rate": 0.0006,
+ "loss": 5.177639961242676,
+ "step": 3530
+ },
+ {
+ "epoch": 49.041921397379916,
+ "grad_norm": 0.01133775431662798,
+ "learning_rate": 0.0006,
+ "loss": 5.283428192138672,
+ "step": 3531
+ },
+ {
+ "epoch": 49.05589519650655,
+ "grad_norm": 0.01154172234237194,
+ "learning_rate": 0.0006,
+ "loss": 5.206364154815674,
+ "step": 3532
+ },
+ {
+ "epoch": 49.069868995633186,
+ "grad_norm": 0.010192356072366238,
+ "learning_rate": 0.0006,
+ "loss": 5.2385735511779785,
+ "step": 3533
+ },
+ {
+ "epoch": 49.083842794759825,
+ "grad_norm": 0.012789200060069561,
+ "learning_rate": 0.0006,
+ "loss": 5.071800231933594,
+ "step": 3534
+ },
+ {
+ "epoch": 49.09781659388646,
+ "grad_norm": 0.011282820254564285,
+ "learning_rate": 0.0006,
+ "loss": 5.218020439147949,
+ "step": 3535
+ },
+ {
+ "epoch": 49.1117903930131,
+ "grad_norm": 0.012153448536992073,
+ "learning_rate": 0.0006,
+ "loss": 5.134698867797852,
+ "step": 3536
+ },
+ {
+ "epoch": 49.12576419213974,
+ "grad_norm": 0.011920543387532234,
+ "learning_rate": 0.0006,
+ "loss": 5.141620635986328,
+ "step": 3537
+ },
+ {
+ "epoch": 49.13973799126637,
+ "grad_norm": 0.010949329473078251,
+ "learning_rate": 0.0006,
+ "loss": 5.116259574890137,
+ "step": 3538
+ },
+ {
+ "epoch": 49.15371179039301,
+ "grad_norm": 0.012362448498606682,
+ "learning_rate": 0.0006,
+ "loss": 5.090177059173584,
+ "step": 3539
+ },
+ {
+ "epoch": 49.16768558951965,
+ "grad_norm": 0.011575071141123772,
+ "learning_rate": 0.0006,
+ "loss": 5.093441009521484,
+ "step": 3540
+ },
+ {
+ "epoch": 49.18165938864629,
+ "grad_norm": 0.011126236990094185,
+ "learning_rate": 0.0006,
+ "loss": 5.060235023498535,
+ "step": 3541
+ },
+ {
+ "epoch": 49.19563318777293,
+ "grad_norm": 0.0107401292771101,
+ "learning_rate": 0.0006,
+ "loss": 5.140704154968262,
+ "step": 3542
+ },
+ {
+ "epoch": 49.209606986899566,
+ "grad_norm": 0.011202906258404255,
+ "learning_rate": 0.0006,
+ "loss": 5.10855770111084,
+ "step": 3543
+ },
+ {
+ "epoch": 49.223580786026204,
+ "grad_norm": 0.011083454824984074,
+ "learning_rate": 0.0006,
+ "loss": 5.092751502990723,
+ "step": 3544
+ },
+ {
+ "epoch": 49.237554585152836,
+ "grad_norm": 0.009523800574243069,
+ "learning_rate": 0.0006,
+ "loss": 5.13733434677124,
+ "step": 3545
+ },
+ {
+ "epoch": 49.251528384279474,
+ "grad_norm": 0.011298353783786297,
+ "learning_rate": 0.0006,
+ "loss": 5.036048889160156,
+ "step": 3546
+ },
+ {
+ "epoch": 49.26550218340611,
+ "grad_norm": 0.010676483623683453,
+ "learning_rate": 0.0006,
+ "loss": 5.051901340484619,
+ "step": 3547
+ },
+ {
+ "epoch": 49.27947598253275,
+ "grad_norm": 0.011463399045169353,
+ "learning_rate": 0.0006,
+ "loss": 4.983170986175537,
+ "step": 3548
+ },
+ {
+ "epoch": 49.29344978165939,
+ "grad_norm": 0.010603402741253376,
+ "learning_rate": 0.0006,
+ "loss": 5.112311363220215,
+ "step": 3549
+ },
+ {
+ "epoch": 49.30742358078603,
+ "grad_norm": 0.010535670444369316,
+ "learning_rate": 0.0006,
+ "loss": 5.085227966308594,
+ "step": 3550
+ },
+ {
+ "epoch": 49.32139737991266,
+ "grad_norm": 0.011843928135931492,
+ "learning_rate": 0.0006,
+ "loss": 5.07575798034668,
+ "step": 3551
+ },
+ {
+ "epoch": 49.3353711790393,
+ "grad_norm": 0.010587952099740505,
+ "learning_rate": 0.0006,
+ "loss": 5.019237041473389,
+ "step": 3552
+ },
+ {
+ "epoch": 49.34934497816594,
+ "grad_norm": 0.012285659089684486,
+ "learning_rate": 0.0006,
+ "loss": 5.144603729248047,
+ "step": 3553
+ },
+ {
+ "epoch": 49.36331877729258,
+ "grad_norm": 0.013119431212544441,
+ "learning_rate": 0.0006,
+ "loss": 5.035256862640381,
+ "step": 3554
+ },
+ {
+ "epoch": 49.377292576419215,
+ "grad_norm": 0.012138537131249905,
+ "learning_rate": 0.0006,
+ "loss": 4.996667861938477,
+ "step": 3555
+ },
+ {
+ "epoch": 49.391266375545854,
+ "grad_norm": 0.012142804451286793,
+ "learning_rate": 0.0006,
+ "loss": 4.95058536529541,
+ "step": 3556
+ },
+ {
+ "epoch": 49.40524017467249,
+ "grad_norm": 0.012205676175653934,
+ "learning_rate": 0.0006,
+ "loss": 5.096931457519531,
+ "step": 3557
+ },
+ {
+ "epoch": 49.419213973799124,
+ "grad_norm": 0.012086744420230389,
+ "learning_rate": 0.0006,
+ "loss": 4.99384880065918,
+ "step": 3558
+ },
+ {
+ "epoch": 49.43318777292576,
+ "grad_norm": 0.012918103486299515,
+ "learning_rate": 0.0006,
+ "loss": 5.031294822692871,
+ "step": 3559
+ },
+ {
+ "epoch": 49.4471615720524,
+ "grad_norm": 0.012006713077425957,
+ "learning_rate": 0.0006,
+ "loss": 5.0547943115234375,
+ "step": 3560
+ },
+ {
+ "epoch": 49.46113537117904,
+ "grad_norm": 0.0120399734005332,
+ "learning_rate": 0.0006,
+ "loss": 5.014238357543945,
+ "step": 3561
+ },
+ {
+ "epoch": 49.47510917030568,
+ "grad_norm": 0.013306990265846252,
+ "learning_rate": 0.0006,
+ "loss": 4.972925186157227,
+ "step": 3562
+ },
+ {
+ "epoch": 49.48908296943232,
+ "grad_norm": 0.013878636993467808,
+ "learning_rate": 0.0006,
+ "loss": 5.079033851623535,
+ "step": 3563
+ },
+ {
+ "epoch": 49.50305676855895,
+ "grad_norm": 0.012454566545784473,
+ "learning_rate": 0.0006,
+ "loss": 4.93708610534668,
+ "step": 3564
+ },
+ {
+ "epoch": 49.51703056768559,
+ "grad_norm": 0.014075061306357384,
+ "learning_rate": 0.0006,
+ "loss": 4.965582370758057,
+ "step": 3565
+ },
+ {
+ "epoch": 49.531004366812226,
+ "grad_norm": 0.012525617145001888,
+ "learning_rate": 0.0006,
+ "loss": 4.968148708343506,
+ "step": 3566
+ },
+ {
+ "epoch": 49.544978165938865,
+ "grad_norm": 0.012418225407600403,
+ "learning_rate": 0.0006,
+ "loss": 5.06901741027832,
+ "step": 3567
+ },
+ {
+ "epoch": 49.5589519650655,
+ "grad_norm": 0.015034239739179611,
+ "learning_rate": 0.0006,
+ "loss": 4.984277725219727,
+ "step": 3568
+ },
+ {
+ "epoch": 49.57292576419214,
+ "grad_norm": 0.014329628087580204,
+ "learning_rate": 0.0006,
+ "loss": 4.951558589935303,
+ "step": 3569
+ },
+ {
+ "epoch": 49.58689956331878,
+ "grad_norm": 0.017769070342183113,
+ "learning_rate": 0.0006,
+ "loss": 4.966532230377197,
+ "step": 3570
+ },
+ {
+ "epoch": 49.60087336244541,
+ "grad_norm": 0.01909453794360161,
+ "learning_rate": 0.0006,
+ "loss": 4.919890880584717,
+ "step": 3571
+ },
+ {
+ "epoch": 49.61484716157205,
+ "grad_norm": 0.017519688233733177,
+ "learning_rate": 0.0006,
+ "loss": 5.031286716461182,
+ "step": 3572
+ },
+ {
+ "epoch": 49.62882096069869,
+ "grad_norm": 0.017143897712230682,
+ "learning_rate": 0.0006,
+ "loss": 4.969764709472656,
+ "step": 3573
+ },
+ {
+ "epoch": 49.64279475982533,
+ "grad_norm": 0.01688285917043686,
+ "learning_rate": 0.0006,
+ "loss": 5.0733442306518555,
+ "step": 3574
+ },
+ {
+ "epoch": 49.65676855895197,
+ "grad_norm": 0.018547311425209045,
+ "learning_rate": 0.0006,
+ "loss": 4.980863571166992,
+ "step": 3575
+ },
+ {
+ "epoch": 49.670742358078606,
+ "grad_norm": 0.018877137452363968,
+ "learning_rate": 0.0006,
+ "loss": 4.954646110534668,
+ "step": 3576
+ },
+ {
+ "epoch": 49.68471615720524,
+ "grad_norm": 0.018681600689888,
+ "learning_rate": 0.0006,
+ "loss": 4.959702014923096,
+ "step": 3577
+ },
+ {
+ "epoch": 49.698689956331876,
+ "grad_norm": 0.018947336822748184,
+ "learning_rate": 0.0006,
+ "loss": 5.070502758026123,
+ "step": 3578
+ },
+ {
+ "epoch": 49.712663755458514,
+ "grad_norm": 0.018817484378814697,
+ "learning_rate": 0.0006,
+ "loss": 5.056508541107178,
+ "step": 3579
+ },
+ {
+ "epoch": 49.72663755458515,
+ "grad_norm": 0.01931724324822426,
+ "learning_rate": 0.0006,
+ "loss": 4.922451972961426,
+ "step": 3580
+ },
+ {
+ "epoch": 49.74061135371179,
+ "grad_norm": 0.022047819569706917,
+ "learning_rate": 0.0006,
+ "loss": 5.053170204162598,
+ "step": 3581
+ },
+ {
+ "epoch": 49.75458515283843,
+ "grad_norm": 0.02279801294207573,
+ "learning_rate": 0.0006,
+ "loss": 4.8768439292907715,
+ "step": 3582
+ },
+ {
+ "epoch": 49.76855895196506,
+ "grad_norm": 0.020661601796746254,
+ "learning_rate": 0.0006,
+ "loss": 4.907928466796875,
+ "step": 3583
+ },
+ {
+ "epoch": 49.7825327510917,
+ "grad_norm": 0.020479509606957436,
+ "learning_rate": 0.0006,
+ "loss": 5.040694236755371,
+ "step": 3584
+ },
+ {
+ "epoch": 49.79650655021834,
+ "grad_norm": 0.021654162555933,
+ "learning_rate": 0.0006,
+ "loss": 5.018434524536133,
+ "step": 3585
+ },
+ {
+ "epoch": 49.81048034934498,
+ "grad_norm": 0.022486824542284012,
+ "learning_rate": 0.0006,
+ "loss": 5.086538314819336,
+ "step": 3586
+ },
+ {
+ "epoch": 49.82445414847162,
+ "grad_norm": 0.02141369879245758,
+ "learning_rate": 0.0006,
+ "loss": 5.035694122314453,
+ "step": 3587
+ },
+ {
+ "epoch": 49.838427947598255,
+ "grad_norm": 0.015926677733659744,
+ "learning_rate": 0.0006,
+ "loss": 4.905038833618164,
+ "step": 3588
+ },
+ {
+ "epoch": 49.852401746724894,
+ "grad_norm": 0.015775298699736595,
+ "learning_rate": 0.0006,
+ "loss": 5.003902435302734,
+ "step": 3589
+ },
+ {
+ "epoch": 49.866375545851525,
+ "grad_norm": 0.01872118003666401,
+ "learning_rate": 0.0006,
+ "loss": 4.891060829162598,
+ "step": 3590
+ },
+ {
+ "epoch": 49.880349344978164,
+ "grad_norm": 0.02066192403435707,
+ "learning_rate": 0.0006,
+ "loss": 5.002897262573242,
+ "step": 3591
+ },
+ {
+ "epoch": 49.8943231441048,
+ "grad_norm": 0.02348714880645275,
+ "learning_rate": 0.0006,
+ "loss": 4.9189958572387695,
+ "step": 3592
+ },
+ {
+ "epoch": 49.90829694323144,
+ "grad_norm": 0.020402103662490845,
+ "learning_rate": 0.0006,
+ "loss": 4.992661476135254,
+ "step": 3593
+ },
+ {
+ "epoch": 49.92227074235808,
+ "grad_norm": 0.020466072484850883,
+ "learning_rate": 0.0006,
+ "loss": 4.962968826293945,
+ "step": 3594
+ },
+ {
+ "epoch": 49.93624454148472,
+ "grad_norm": 0.01922314241528511,
+ "learning_rate": 0.0006,
+ "loss": 4.897730827331543,
+ "step": 3595
+ },
+ {
+ "epoch": 49.95021834061135,
+ "grad_norm": 0.017228754237294197,
+ "learning_rate": 0.0006,
+ "loss": 4.964842796325684,
+ "step": 3596
+ },
+ {
+ "epoch": 49.96419213973799,
+ "grad_norm": 0.016021743416786194,
+ "learning_rate": 0.0006,
+ "loss": 4.910758972167969,
+ "step": 3597
+ },
+ {
+ "epoch": 49.97816593886463,
+ "grad_norm": 0.017985103651881218,
+ "learning_rate": 0.0006,
+ "loss": 4.916072368621826,
+ "step": 3598
+ },
+ {
+ "epoch": 49.992139737991266,
+ "grad_norm": 0.018780240789055824,
+ "learning_rate": 0.0006,
+ "loss": 4.94658088684082,
+ "step": 3599
+ },
+ {
+ "epoch": 50.0,
+ "grad_norm": 0.01776314526796341,
+ "learning_rate": 0.0006,
+ "loss": 4.969357967376709,
+ "step": 3600
+ },
+ {
+ "epoch": 50.0,
+ "eval_loss": 5.141158580780029,
+ "eval_runtime": 56.2332,
+ "eval_samples_per_second": 43.426,
+ "eval_steps_per_second": 1.369,
+ "step": 3600
+ },
+ {
+ "epoch": 50.01397379912664,
+ "grad_norm": 0.016308696940541267,
+ "learning_rate": 0.0006,
+ "loss": 4.924664497375488,
+ "step": 3601
+ },
+ {
+ "epoch": 50.02794759825328,
+ "grad_norm": 0.018810199573636055,
+ "learning_rate": 0.0006,
+ "loss": 4.82293701171875,
+ "step": 3602
+ },
+ {
+ "epoch": 50.041921397379916,
+ "grad_norm": 0.02227160707116127,
+ "learning_rate": 0.0006,
+ "loss": 4.9612298011779785,
+ "step": 3603
+ },
+ {
+ "epoch": 50.05589519650655,
+ "grad_norm": 0.025266462936997414,
+ "learning_rate": 0.0006,
+ "loss": 4.963835716247559,
+ "step": 3604
+ },
+ {
+ "epoch": 50.069868995633186,
+ "grad_norm": 0.028588181361556053,
+ "learning_rate": 0.0006,
+ "loss": 4.992118835449219,
+ "step": 3605
+ },
+ {
+ "epoch": 50.083842794759825,
+ "grad_norm": 0.02739487960934639,
+ "learning_rate": 0.0006,
+ "loss": 4.915547847747803,
+ "step": 3606
+ },
+ {
+ "epoch": 50.09781659388646,
+ "grad_norm": 0.027000300586223602,
+ "learning_rate": 0.0006,
+ "loss": 4.883801460266113,
+ "step": 3607
+ },
+ {
+ "epoch": 50.1117903930131,
+ "grad_norm": 0.023468483239412308,
+ "learning_rate": 0.0006,
+ "loss": 5.006168365478516,
+ "step": 3608
+ },
+ {
+ "epoch": 50.12576419213974,
+ "grad_norm": 0.026753634214401245,
+ "learning_rate": 0.0006,
+ "loss": 4.92526912689209,
+ "step": 3609
+ },
+ {
+ "epoch": 50.13973799126637,
+ "grad_norm": 0.026816733181476593,
+ "learning_rate": 0.0006,
+ "loss": 4.847546100616455,
+ "step": 3610
+ },
+ {
+ "epoch": 50.15371179039301,
+ "grad_norm": 0.027362391352653503,
+ "learning_rate": 0.0006,
+ "loss": 4.862548351287842,
+ "step": 3611
+ },
+ {
+ "epoch": 50.16768558951965,
+ "grad_norm": 0.03450845181941986,
+ "learning_rate": 0.0006,
+ "loss": 4.963202953338623,
+ "step": 3612
+ },
+ {
+ "epoch": 50.18165938864629,
+ "grad_norm": 0.03801408037543297,
+ "learning_rate": 0.0006,
+ "loss": 4.770240306854248,
+ "step": 3613
+ },
+ {
+ "epoch": 50.19563318777293,
+ "grad_norm": 0.03579140082001686,
+ "learning_rate": 0.0006,
+ "loss": 4.797981262207031,
+ "step": 3614
+ },
+ {
+ "epoch": 50.209606986899566,
+ "grad_norm": 0.029393760487437248,
+ "learning_rate": 0.0006,
+ "loss": 4.798581600189209,
+ "step": 3615
+ },
+ {
+ "epoch": 50.223580786026204,
+ "grad_norm": 0.026111142709851265,
+ "learning_rate": 0.0006,
+ "loss": 4.990712642669678,
+ "step": 3616
+ },
+ {
+ "epoch": 50.237554585152836,
+ "grad_norm": 0.02213387005031109,
+ "learning_rate": 0.0006,
+ "loss": 4.865842342376709,
+ "step": 3617
+ },
+ {
+ "epoch": 50.251528384279474,
+ "grad_norm": 0.025450684130191803,
+ "learning_rate": 0.0006,
+ "loss": 4.903746128082275,
+ "step": 3618
+ },
+ {
+ "epoch": 50.26550218340611,
+ "grad_norm": 0.020915117114782333,
+ "learning_rate": 0.0006,
+ "loss": 4.951366424560547,
+ "step": 3619
+ },
+ {
+ "epoch": 50.27947598253275,
+ "grad_norm": 0.020972849801182747,
+ "learning_rate": 0.0006,
+ "loss": 4.783311367034912,
+ "step": 3620
+ },
+ {
+ "epoch": 50.29344978165939,
+ "grad_norm": 0.016956273466348648,
+ "learning_rate": 0.0006,
+ "loss": 4.955767631530762,
+ "step": 3621
+ },
+ {
+ "epoch": 50.30742358078603,
+ "grad_norm": 0.018045252189040184,
+ "learning_rate": 0.0006,
+ "loss": 4.869322776794434,
+ "step": 3622
+ },
+ {
+ "epoch": 50.32139737991266,
+ "grad_norm": 0.018199188634753227,
+ "learning_rate": 0.0006,
+ "loss": 4.95680046081543,
+ "step": 3623
+ },
+ {
+ "epoch": 50.3353711790393,
+ "grad_norm": 0.016688082367181778,
+ "learning_rate": 0.0006,
+ "loss": 4.783535957336426,
+ "step": 3624
+ },
+ {
+ "epoch": 50.34934497816594,
+ "grad_norm": 0.017540352419018745,
+ "learning_rate": 0.0006,
+ "loss": 4.8236083984375,
+ "step": 3625
+ },
+ {
+ "epoch": 50.36331877729258,
+ "grad_norm": 0.018054187297821045,
+ "learning_rate": 0.0006,
+ "loss": 4.764655113220215,
+ "step": 3626
+ },
+ {
+ "epoch": 50.377292576419215,
+ "grad_norm": 0.01741131953895092,
+ "learning_rate": 0.0006,
+ "loss": 4.875613689422607,
+ "step": 3627
+ },
+ {
+ "epoch": 50.391266375545854,
+ "grad_norm": 0.01682269014418125,
+ "learning_rate": 0.0006,
+ "loss": 4.818509101867676,
+ "step": 3628
+ },
+ {
+ "epoch": 50.40524017467249,
+ "grad_norm": 0.016652341932058334,
+ "learning_rate": 0.0006,
+ "loss": 4.974020481109619,
+ "step": 3629
+ },
+ {
+ "epoch": 50.419213973799124,
+ "grad_norm": 0.02035282365977764,
+ "learning_rate": 0.0006,
+ "loss": 4.893037796020508,
+ "step": 3630
+ },
+ {
+ "epoch": 50.43318777292576,
+ "grad_norm": 0.019907061010599136,
+ "learning_rate": 0.0006,
+ "loss": 4.848119735717773,
+ "step": 3631
+ },
+ {
+ "epoch": 50.4471615720524,
+ "grad_norm": 0.023861493915319443,
+ "learning_rate": 0.0006,
+ "loss": 4.819353103637695,
+ "step": 3632
+ },
+ {
+ "epoch": 50.46113537117904,
+ "grad_norm": 0.025254026055336,
+ "learning_rate": 0.0006,
+ "loss": 4.769312858581543,
+ "step": 3633
+ },
+ {
+ "epoch": 50.47510917030568,
+ "grad_norm": 0.026154937222599983,
+ "learning_rate": 0.0006,
+ "loss": 4.7586774826049805,
+ "step": 3634
+ },
+ {
+ "epoch": 50.48908296943232,
+ "grad_norm": 0.026342090219259262,
+ "learning_rate": 0.0006,
+ "loss": 4.861218452453613,
+ "step": 3635
+ },
+ {
+ "epoch": 50.50305676855895,
+ "grad_norm": 0.028169559314846992,
+ "learning_rate": 0.0006,
+ "loss": 4.818380355834961,
+ "step": 3636
+ },
+ {
+ "epoch": 50.51703056768559,
+ "grad_norm": 0.02820264920592308,
+ "learning_rate": 0.0006,
+ "loss": 4.892871856689453,
+ "step": 3637
+ },
+ {
+ "epoch": 50.531004366812226,
+ "grad_norm": 0.03345080465078354,
+ "learning_rate": 0.0006,
+ "loss": 4.811084747314453,
+ "step": 3638
+ },
+ {
+ "epoch": 50.544978165938865,
+ "grad_norm": 0.04005736857652664,
+ "learning_rate": 0.0006,
+ "loss": 4.765410900115967,
+ "step": 3639
+ },
+ {
+ "epoch": 50.5589519650655,
+ "grad_norm": 0.04333870857954025,
+ "learning_rate": 0.0006,
+ "loss": 4.956520080566406,
+ "step": 3640
+ },
+ {
+ "epoch": 50.57292576419214,
+ "grad_norm": 0.04216952994465828,
+ "learning_rate": 0.0006,
+ "loss": 4.793441295623779,
+ "step": 3641
+ },
+ {
+ "epoch": 50.58689956331878,
+ "grad_norm": 0.02936951443552971,
+ "learning_rate": 0.0006,
+ "loss": 4.8447957038879395,
+ "step": 3642
+ },
+ {
+ "epoch": 50.60087336244541,
+ "grad_norm": 0.026937026530504227,
+ "learning_rate": 0.0006,
+ "loss": 4.962215423583984,
+ "step": 3643
+ },
+ {
+ "epoch": 50.61484716157205,
+ "grad_norm": 0.02628890983760357,
+ "learning_rate": 0.0006,
+ "loss": 4.883061408996582,
+ "step": 3644
+ },
+ {
+ "epoch": 50.62882096069869,
+ "grad_norm": 0.025984834879636765,
+ "learning_rate": 0.0006,
+ "loss": 4.768662452697754,
+ "step": 3645
+ },
+ {
+ "epoch": 50.64279475982533,
+ "grad_norm": 0.023998012766242027,
+ "learning_rate": 0.0006,
+ "loss": 4.851325035095215,
+ "step": 3646
+ },
+ {
+ "epoch": 50.65676855895197,
+ "grad_norm": 0.02143184095621109,
+ "learning_rate": 0.0006,
+ "loss": 4.8274688720703125,
+ "step": 3647
+ },
+ {
+ "epoch": 50.670742358078606,
+ "grad_norm": 0.02488900162279606,
+ "learning_rate": 0.0006,
+ "loss": 4.831140518188477,
+ "step": 3648
+ },
+ {
+ "epoch": 50.68471615720524,
+ "grad_norm": 0.02330533228814602,
+ "learning_rate": 0.0006,
+ "loss": 4.850462913513184,
+ "step": 3649
+ },
+ {
+ "epoch": 50.698689956331876,
+ "grad_norm": 0.02121943049132824,
+ "learning_rate": 0.0006,
+ "loss": 4.870253086090088,
+ "step": 3650
+ },
+ {
+ "epoch": 50.712663755458514,
+ "grad_norm": 0.02080935798585415,
+ "learning_rate": 0.0006,
+ "loss": 4.798770904541016,
+ "step": 3651
+ },
+ {
+ "epoch": 50.72663755458515,
+ "grad_norm": 0.019047029316425323,
+ "learning_rate": 0.0006,
+ "loss": 4.8427300453186035,
+ "step": 3652
+ },
+ {
+ "epoch": 50.74061135371179,
+ "grad_norm": 0.01824103854596615,
+ "learning_rate": 0.0006,
+ "loss": 4.820980548858643,
+ "step": 3653
+ },
+ {
+ "epoch": 50.75458515283843,
+ "grad_norm": 0.018129967153072357,
+ "learning_rate": 0.0006,
+ "loss": 4.946732997894287,
+ "step": 3654
+ },
+ {
+ "epoch": 50.76855895196506,
+ "grad_norm": 0.01784815452992916,
+ "learning_rate": 0.0006,
+ "loss": 4.829516410827637,
+ "step": 3655
+ },
+ {
+ "epoch": 50.7825327510917,
+ "grad_norm": 0.020021555945277214,
+ "learning_rate": 0.0006,
+ "loss": 4.780428409576416,
+ "step": 3656
+ },
+ {
+ "epoch": 50.79650655021834,
+ "grad_norm": 0.01863253302872181,
+ "learning_rate": 0.0006,
+ "loss": 4.791567802429199,
+ "step": 3657
+ },
+ {
+ "epoch": 50.81048034934498,
+ "grad_norm": 0.019151683896780014,
+ "learning_rate": 0.0006,
+ "loss": 4.824892997741699,
+ "step": 3658
+ },
+ {
+ "epoch": 50.82445414847162,
+ "grad_norm": 0.01891648955643177,
+ "learning_rate": 0.0006,
+ "loss": 4.789605140686035,
+ "step": 3659
+ },
+ {
+ "epoch": 50.838427947598255,
+ "grad_norm": 0.016019510105252266,
+ "learning_rate": 0.0006,
+ "loss": 4.779555320739746,
+ "step": 3660
+ },
+ {
+ "epoch": 50.852401746724894,
+ "grad_norm": 0.016096491366624832,
+ "learning_rate": 0.0006,
+ "loss": 4.753842353820801,
+ "step": 3661
+ },
+ {
+ "epoch": 50.866375545851525,
+ "grad_norm": 0.01693476364016533,
+ "learning_rate": 0.0006,
+ "loss": 4.768477439880371,
+ "step": 3662
+ },
+ {
+ "epoch": 50.880349344978164,
+ "grad_norm": 0.01749575324356556,
+ "learning_rate": 0.0006,
+ "loss": 4.779143333435059,
+ "step": 3663
+ },
+ {
+ "epoch": 50.8943231441048,
+ "grad_norm": 0.016851043328642845,
+ "learning_rate": 0.0006,
+ "loss": 4.780387878417969,
+ "step": 3664
+ },
+ {
+ "epoch": 50.90829694323144,
+ "grad_norm": 0.019045202061533928,
+ "learning_rate": 0.0006,
+ "loss": 4.769913673400879,
+ "step": 3665
+ },
+ {
+ "epoch": 50.92227074235808,
+ "grad_norm": 0.01954123005270958,
+ "learning_rate": 0.0006,
+ "loss": 4.769009590148926,
+ "step": 3666
+ },
+ {
+ "epoch": 50.93624454148472,
+ "grad_norm": 0.022527096793055534,
+ "learning_rate": 0.0006,
+ "loss": 4.789215087890625,
+ "step": 3667
+ },
+ {
+ "epoch": 50.95021834061135,
+ "grad_norm": 0.02871634066104889,
+ "learning_rate": 0.0006,
+ "loss": 4.749727725982666,
+ "step": 3668
+ },
+ {
+ "epoch": 50.96419213973799,
+ "grad_norm": 0.04045053571462631,
+ "learning_rate": 0.0006,
+ "loss": 4.909038543701172,
+ "step": 3669
+ },
+ {
+ "epoch": 50.97816593886463,
+ "grad_norm": 0.04156879335641861,
+ "learning_rate": 0.0006,
+ "loss": 4.876831531524658,
+ "step": 3670
+ },
+ {
+ "epoch": 50.992139737991266,
+ "grad_norm": 0.02621397376060486,
+ "learning_rate": 0.0006,
+ "loss": 4.797173500061035,
+ "step": 3671
+ },
+ {
+ "epoch": 51.0,
+ "grad_norm": 0.03389597684144974,
+ "learning_rate": 0.0006,
+ "loss": 4.8428850173950195,
+ "step": 3672
+ },
+ {
+ "epoch": 51.0,
+ "eval_loss": 4.989261627197266,
+ "eval_runtime": 56.358,
+ "eval_samples_per_second": 43.33,
+ "eval_steps_per_second": 1.366,
+ "step": 3672
+ },
+ {
+ "epoch": 51.01397379912664,
+ "grad_norm": 0.03216972202062607,
+ "learning_rate": 0.0006,
+ "loss": 4.755102634429932,
+ "step": 3673
+ },
+ {
+ "epoch": 51.02794759825328,
+ "grad_norm": 0.03190374746918678,
+ "learning_rate": 0.0006,
+ "loss": 4.720424652099609,
+ "step": 3674
+ },
+ {
+ "epoch": 51.041921397379916,
+ "grad_norm": 0.031727299094200134,
+ "learning_rate": 0.0006,
+ "loss": 4.803277969360352,
+ "step": 3675
+ },
+ {
+ "epoch": 51.05589519650655,
+ "grad_norm": 0.02473079040646553,
+ "learning_rate": 0.0006,
+ "loss": 4.761087417602539,
+ "step": 3676
+ },
+ {
+ "epoch": 51.069868995633186,
+ "grad_norm": 0.020360209047794342,
+ "learning_rate": 0.0006,
+ "loss": 4.753884792327881,
+ "step": 3677
+ },
+ {
+ "epoch": 51.083842794759825,
+ "grad_norm": 0.02078923024237156,
+ "learning_rate": 0.0006,
+ "loss": 4.647130489349365,
+ "step": 3678
+ },
+ {
+ "epoch": 51.09781659388646,
+ "grad_norm": 0.019387200474739075,
+ "learning_rate": 0.0006,
+ "loss": 4.74135160446167,
+ "step": 3679
+ },
+ {
+ "epoch": 51.1117903930131,
+ "grad_norm": 0.01874949410557747,
+ "learning_rate": 0.0006,
+ "loss": 4.795329570770264,
+ "step": 3680
+ },
+ {
+ "epoch": 51.12576419213974,
+ "grad_norm": 0.019331837072968483,
+ "learning_rate": 0.0006,
+ "loss": 4.897967338562012,
+ "step": 3681
+ },
+ {
+ "epoch": 51.13973799126637,
+ "grad_norm": 0.017066050320863724,
+ "learning_rate": 0.0006,
+ "loss": 4.825351715087891,
+ "step": 3682
+ },
+ {
+ "epoch": 51.15371179039301,
+ "grad_norm": 0.01757810264825821,
+ "learning_rate": 0.0006,
+ "loss": 4.719335079193115,
+ "step": 3683
+ },
+ {
+ "epoch": 51.16768558951965,
+ "grad_norm": 0.018174266442656517,
+ "learning_rate": 0.0006,
+ "loss": 4.732680320739746,
+ "step": 3684
+ },
+ {
+ "epoch": 51.18165938864629,
+ "grad_norm": 0.018497727811336517,
+ "learning_rate": 0.0006,
+ "loss": 4.731050491333008,
+ "step": 3685
+ },
+ {
+ "epoch": 51.19563318777293,
+ "grad_norm": 0.0186338871717453,
+ "learning_rate": 0.0006,
+ "loss": 4.748035430908203,
+ "step": 3686
+ },
+ {
+ "epoch": 51.209606986899566,
+ "grad_norm": 0.018531711772084236,
+ "learning_rate": 0.0006,
+ "loss": 4.9007463455200195,
+ "step": 3687
+ },
+ {
+ "epoch": 51.223580786026204,
+ "grad_norm": 0.018221354112029076,
+ "learning_rate": 0.0006,
+ "loss": 4.84077262878418,
+ "step": 3688
+ },
+ {
+ "epoch": 51.237554585152836,
+ "grad_norm": 0.0203656405210495,
+ "learning_rate": 0.0006,
+ "loss": 4.73009729385376,
+ "step": 3689
+ },
+ {
+ "epoch": 51.251528384279474,
+ "grad_norm": 0.01964893378317356,
+ "learning_rate": 0.0006,
+ "loss": 4.772583961486816,
+ "step": 3690
+ },
+ {
+ "epoch": 51.26550218340611,
+ "grad_norm": 0.019606510177254677,
+ "learning_rate": 0.0006,
+ "loss": 4.599812030792236,
+ "step": 3691
+ },
+ {
+ "epoch": 51.27947598253275,
+ "grad_norm": 0.020501255989074707,
+ "learning_rate": 0.0006,
+ "loss": 4.809898853302002,
+ "step": 3692
+ },
+ {
+ "epoch": 51.29344978165939,
+ "grad_norm": 0.01870838552713394,
+ "learning_rate": 0.0006,
+ "loss": 4.741898059844971,
+ "step": 3693
+ },
+ {
+ "epoch": 51.30742358078603,
+ "grad_norm": 0.01802559196949005,
+ "learning_rate": 0.0006,
+ "loss": 4.746562480926514,
+ "step": 3694
+ },
+ {
+ "epoch": 51.32139737991266,
+ "grad_norm": 0.02053048275411129,
+ "learning_rate": 0.0006,
+ "loss": 4.69932746887207,
+ "step": 3695
+ },
+ {
+ "epoch": 51.3353711790393,
+ "grad_norm": 0.02661825157701969,
+ "learning_rate": 0.0006,
+ "loss": 4.68748664855957,
+ "step": 3696
+ },
+ {
+ "epoch": 51.34934497816594,
+ "grad_norm": 0.028269147500395775,
+ "learning_rate": 0.0006,
+ "loss": 4.731565475463867,
+ "step": 3697
+ },
+ {
+ "epoch": 51.36331877729258,
+ "grad_norm": 0.02650964818894863,
+ "learning_rate": 0.0006,
+ "loss": 4.733857154846191,
+ "step": 3698
+ },
+ {
+ "epoch": 51.377292576419215,
+ "grad_norm": 0.023696066811680794,
+ "learning_rate": 0.0006,
+ "loss": 4.812362194061279,
+ "step": 3699
+ },
+ {
+ "epoch": 51.391266375545854,
+ "grad_norm": 0.02365880273282528,
+ "learning_rate": 0.0006,
+ "loss": 4.733426094055176,
+ "step": 3700
+ },
+ {
+ "epoch": 51.40524017467249,
+ "grad_norm": 0.026249447837471962,
+ "learning_rate": 0.0006,
+ "loss": 4.715622901916504,
+ "step": 3701
+ },
+ {
+ "epoch": 51.419213973799124,
+ "grad_norm": 0.03301899880170822,
+ "learning_rate": 0.0006,
+ "loss": 4.7223663330078125,
+ "step": 3702
+ },
+ {
+ "epoch": 51.43318777292576,
+ "grad_norm": 0.044686660170555115,
+ "learning_rate": 0.0006,
+ "loss": 4.8121161460876465,
+ "step": 3703
+ },
+ {
+ "epoch": 51.4471615720524,
+ "grad_norm": 0.04733522981405258,
+ "learning_rate": 0.0006,
+ "loss": 4.841394424438477,
+ "step": 3704
+ },
+ {
+ "epoch": 51.46113537117904,
+ "grad_norm": 0.0395209975540638,
+ "learning_rate": 0.0006,
+ "loss": 4.773837089538574,
+ "step": 3705
+ },
+ {
+ "epoch": 51.47510917030568,
+ "grad_norm": 0.032790958881378174,
+ "learning_rate": 0.0006,
+ "loss": 4.833066463470459,
+ "step": 3706
+ },
+ {
+ "epoch": 51.48908296943232,
+ "grad_norm": 0.030578427016735077,
+ "learning_rate": 0.0006,
+ "loss": 4.685190200805664,
+ "step": 3707
+ },
+ {
+ "epoch": 51.50305676855895,
+ "grad_norm": 0.031619083136320114,
+ "learning_rate": 0.0006,
+ "loss": 4.740378379821777,
+ "step": 3708
+ },
+ {
+ "epoch": 51.51703056768559,
+ "grad_norm": 0.02385825850069523,
+ "learning_rate": 0.0006,
+ "loss": 4.812911510467529,
+ "step": 3709
+ },
+ {
+ "epoch": 51.531004366812226,
+ "grad_norm": 0.023523425683379173,
+ "learning_rate": 0.0006,
+ "loss": 4.669580459594727,
+ "step": 3710
+ },
+ {
+ "epoch": 51.544978165938865,
+ "grad_norm": 0.021316280588507652,
+ "learning_rate": 0.0006,
+ "loss": 4.76888370513916,
+ "step": 3711
+ },
+ {
+ "epoch": 51.5589519650655,
+ "grad_norm": 0.024663880467414856,
+ "learning_rate": 0.0006,
+ "loss": 4.702592372894287,
+ "step": 3712
+ },
+ {
+ "epoch": 51.57292576419214,
+ "grad_norm": 0.024173280224204063,
+ "learning_rate": 0.0006,
+ "loss": 4.710604667663574,
+ "step": 3713
+ },
+ {
+ "epoch": 51.58689956331878,
+ "grad_norm": 0.02754114381968975,
+ "learning_rate": 0.0006,
+ "loss": 4.746356964111328,
+ "step": 3714
+ },
+ {
+ "epoch": 51.60087336244541,
+ "grad_norm": 0.027903622016310692,
+ "learning_rate": 0.0006,
+ "loss": 4.7546491622924805,
+ "step": 3715
+ },
+ {
+ "epoch": 51.61484716157205,
+ "grad_norm": 0.024871617555618286,
+ "learning_rate": 0.0006,
+ "loss": 4.605746269226074,
+ "step": 3716
+ },
+ {
+ "epoch": 51.62882096069869,
+ "grad_norm": 0.021239031106233597,
+ "learning_rate": 0.0006,
+ "loss": 4.6942458152771,
+ "step": 3717
+ },
+ {
+ "epoch": 51.64279475982533,
+ "grad_norm": 0.01925622671842575,
+ "learning_rate": 0.0006,
+ "loss": 4.593558311462402,
+ "step": 3718
+ },
+ {
+ "epoch": 51.65676855895197,
+ "grad_norm": 0.01956283673644066,
+ "learning_rate": 0.0006,
+ "loss": 4.631172180175781,
+ "step": 3719
+ },
+ {
+ "epoch": 51.670742358078606,
+ "grad_norm": 0.02269871160387993,
+ "learning_rate": 0.0006,
+ "loss": 4.729236602783203,
+ "step": 3720
+ },
+ {
+ "epoch": 51.68471615720524,
+ "grad_norm": 0.02253916673362255,
+ "learning_rate": 0.0006,
+ "loss": 4.7191033363342285,
+ "step": 3721
+ },
+ {
+ "epoch": 51.698689956331876,
+ "grad_norm": 0.026231050491333008,
+ "learning_rate": 0.0006,
+ "loss": 4.811850547790527,
+ "step": 3722
+ },
+ {
+ "epoch": 51.712663755458514,
+ "grad_norm": 0.02558821812272072,
+ "learning_rate": 0.0006,
+ "loss": 4.681466102600098,
+ "step": 3723
+ },
+ {
+ "epoch": 51.72663755458515,
+ "grad_norm": 0.02124209702014923,
+ "learning_rate": 0.0006,
+ "loss": 4.693664073944092,
+ "step": 3724
+ },
+ {
+ "epoch": 51.74061135371179,
+ "grad_norm": 0.016638319939374924,
+ "learning_rate": 0.0006,
+ "loss": 4.771440029144287,
+ "step": 3725
+ },
+ {
+ "epoch": 51.75458515283843,
+ "grad_norm": 0.018048347905278206,
+ "learning_rate": 0.0006,
+ "loss": 4.643252372741699,
+ "step": 3726
+ },
+ {
+ "epoch": 51.76855895196506,
+ "grad_norm": 0.019904717803001404,
+ "learning_rate": 0.0006,
+ "loss": 4.66530704498291,
+ "step": 3727
+ },
+ {
+ "epoch": 51.7825327510917,
+ "grad_norm": 0.0194614976644516,
+ "learning_rate": 0.0006,
+ "loss": 4.711943626403809,
+ "step": 3728
+ },
+ {
+ "epoch": 51.79650655021834,
+ "grad_norm": 0.019175555557012558,
+ "learning_rate": 0.0006,
+ "loss": 4.721931457519531,
+ "step": 3729
+ },
+ {
+ "epoch": 51.81048034934498,
+ "grad_norm": 0.02093447372317314,
+ "learning_rate": 0.0006,
+ "loss": 4.633059024810791,
+ "step": 3730
+ },
+ {
+ "epoch": 51.82445414847162,
+ "grad_norm": 0.01930283196270466,
+ "learning_rate": 0.0006,
+ "loss": 4.659373760223389,
+ "step": 3731
+ },
+ {
+ "epoch": 51.838427947598255,
+ "grad_norm": 0.015350676141679287,
+ "learning_rate": 0.0006,
+ "loss": 4.763708114624023,
+ "step": 3732
+ },
+ {
+ "epoch": 51.852401746724894,
+ "grad_norm": 0.014139448292553425,
+ "learning_rate": 0.0006,
+ "loss": 4.753891468048096,
+ "step": 3733
+ },
+ {
+ "epoch": 51.866375545851525,
+ "grad_norm": 0.016215872019529343,
+ "learning_rate": 0.0006,
+ "loss": 4.658967018127441,
+ "step": 3734
+ },
+ {
+ "epoch": 51.880349344978164,
+ "grad_norm": 0.017328619956970215,
+ "learning_rate": 0.0006,
+ "loss": 4.696579933166504,
+ "step": 3735
+ },
+ {
+ "epoch": 51.8943231441048,
+ "grad_norm": 0.018103396520018578,
+ "learning_rate": 0.0006,
+ "loss": 4.627464771270752,
+ "step": 3736
+ },
+ {
+ "epoch": 51.90829694323144,
+ "grad_norm": 0.01715235412120819,
+ "learning_rate": 0.0006,
+ "loss": 4.7368011474609375,
+ "step": 3737
+ },
+ {
+ "epoch": 51.92227074235808,
+ "grad_norm": 0.014790094457566738,
+ "learning_rate": 0.0006,
+ "loss": 4.629441738128662,
+ "step": 3738
+ },
+ {
+ "epoch": 51.93624454148472,
+ "grad_norm": 0.017114296555519104,
+ "learning_rate": 0.0006,
+ "loss": 4.761256694793701,
+ "step": 3739
+ },
+ {
+ "epoch": 51.95021834061135,
+ "grad_norm": 0.019009845331311226,
+ "learning_rate": 0.0006,
+ "loss": 4.789155006408691,
+ "step": 3740
+ },
+ {
+ "epoch": 51.96419213973799,
+ "grad_norm": 0.019696692004799843,
+ "learning_rate": 0.0006,
+ "loss": 4.721269607543945,
+ "step": 3741
+ },
+ {
+ "epoch": 51.97816593886463,
+ "grad_norm": 0.020542822778224945,
+ "learning_rate": 0.0006,
+ "loss": 4.687038898468018,
+ "step": 3742
+ },
+ {
+ "epoch": 51.992139737991266,
+ "grad_norm": 0.021267160773277283,
+ "learning_rate": 0.0006,
+ "loss": 4.747860908508301,
+ "step": 3743
+ },
+ {
+ "epoch": 52.0,
+ "grad_norm": 0.025049567222595215,
+ "learning_rate": 0.0006,
+ "loss": 4.8473100662231445,
+ "step": 3744
+ },
+ {
+ "epoch": 52.0,
+ "eval_loss": 4.943043231964111,
+ "eval_runtime": 56.4929,
+ "eval_samples_per_second": 43.227,
+ "eval_steps_per_second": 1.363,
+ "step": 3744
+ },
+ {
+ "epoch": 52.01397379912664,
+ "grad_norm": 0.03054986707866192,
+ "learning_rate": 0.0006,
+ "loss": 4.856931686401367,
+ "step": 3745
+ },
+ {
+ "epoch": 52.02794759825328,
+ "grad_norm": 0.03838375583291054,
+ "learning_rate": 0.0006,
+ "loss": 4.699807167053223,
+ "step": 3746
+ },
+ {
+ "epoch": 52.041921397379916,
+ "grad_norm": 0.03383488208055496,
+ "learning_rate": 0.0006,
+ "loss": 4.6626787185668945,
+ "step": 3747
+ },
+ {
+ "epoch": 52.05589519650655,
+ "grad_norm": 0.0201561376452446,
+ "learning_rate": 0.0006,
+ "loss": 4.621831893920898,
+ "step": 3748
+ },
+ {
+ "epoch": 52.069868995633186,
+ "grad_norm": 0.028749438002705574,
+ "learning_rate": 0.0006,
+ "loss": 4.682540416717529,
+ "step": 3749
+ },
+ {
+ "epoch": 52.083842794759825,
+ "grad_norm": 0.02681174874305725,
+ "learning_rate": 0.0006,
+ "loss": 4.629997253417969,
+ "step": 3750
+ },
+ {
+ "epoch": 52.09781659388646,
+ "grad_norm": 0.024063318967819214,
+ "learning_rate": 0.0006,
+ "loss": 4.573021411895752,
+ "step": 3751
+ },
+ {
+ "epoch": 52.1117903930131,
+ "grad_norm": 0.02324330434203148,
+ "learning_rate": 0.0006,
+ "loss": 4.739465236663818,
+ "step": 3752
+ },
+ {
+ "epoch": 52.12576419213974,
+ "grad_norm": 0.0225426834076643,
+ "learning_rate": 0.0006,
+ "loss": 4.7335333824157715,
+ "step": 3753
+ },
+ {
+ "epoch": 52.13973799126637,
+ "grad_norm": 0.019593428820371628,
+ "learning_rate": 0.0006,
+ "loss": 4.704958438873291,
+ "step": 3754
+ },
+ {
+ "epoch": 52.15371179039301,
+ "grad_norm": 0.020777082070708275,
+ "learning_rate": 0.0006,
+ "loss": 4.638696670532227,
+ "step": 3755
+ },
+ {
+ "epoch": 52.16768558951965,
+ "grad_norm": 0.021525604650378227,
+ "learning_rate": 0.0006,
+ "loss": 4.643706798553467,
+ "step": 3756
+ },
+ {
+ "epoch": 52.18165938864629,
+ "grad_norm": 0.018943654373288155,
+ "learning_rate": 0.0006,
+ "loss": 4.5173211097717285,
+ "step": 3757
+ },
+ {
+ "epoch": 52.19563318777293,
+ "grad_norm": 0.01856580562889576,
+ "learning_rate": 0.0006,
+ "loss": 4.678788661956787,
+ "step": 3758
+ },
+ {
+ "epoch": 52.209606986899566,
+ "grad_norm": 0.019417259842157364,
+ "learning_rate": 0.0006,
+ "loss": 4.587244987487793,
+ "step": 3759
+ },
+ {
+ "epoch": 52.223580786026204,
+ "grad_norm": 0.01911911740899086,
+ "learning_rate": 0.0006,
+ "loss": 4.688503265380859,
+ "step": 3760
+ },
+ {
+ "epoch": 52.237554585152836,
+ "grad_norm": 0.021017879247665405,
+ "learning_rate": 0.0006,
+ "loss": 4.65829610824585,
+ "step": 3761
+ },
+ {
+ "epoch": 52.251528384279474,
+ "grad_norm": 0.02048875205218792,
+ "learning_rate": 0.0006,
+ "loss": 4.624055862426758,
+ "step": 3762
+ },
+ {
+ "epoch": 52.26550218340611,
+ "grad_norm": 0.01932772807776928,
+ "learning_rate": 0.0006,
+ "loss": 4.556562900543213,
+ "step": 3763
+ },
+ {
+ "epoch": 52.27947598253275,
+ "grad_norm": 0.01871614344418049,
+ "learning_rate": 0.0006,
+ "loss": 4.581201076507568,
+ "step": 3764
+ },
+ {
+ "epoch": 52.29344978165939,
+ "grad_norm": 0.017253106459975243,
+ "learning_rate": 0.0006,
+ "loss": 4.605219841003418,
+ "step": 3765
+ },
+ {
+ "epoch": 52.30742358078603,
+ "grad_norm": 0.01831238530576229,
+ "learning_rate": 0.0006,
+ "loss": 4.759900093078613,
+ "step": 3766
+ },
+ {
+ "epoch": 52.32139737991266,
+ "grad_norm": 0.017946207895874977,
+ "learning_rate": 0.0006,
+ "loss": 4.5598344802856445,
+ "step": 3767
+ },
+ {
+ "epoch": 52.3353711790393,
+ "grad_norm": 0.01680106669664383,
+ "learning_rate": 0.0006,
+ "loss": 4.605034828186035,
+ "step": 3768
+ },
+ {
+ "epoch": 52.34934497816594,
+ "grad_norm": 0.021482540294528008,
+ "learning_rate": 0.0006,
+ "loss": 4.620411396026611,
+ "step": 3769
+ },
+ {
+ "epoch": 52.36331877729258,
+ "grad_norm": 0.02210252732038498,
+ "learning_rate": 0.0006,
+ "loss": 4.667577743530273,
+ "step": 3770
+ },
+ {
+ "epoch": 52.377292576419215,
+ "grad_norm": 0.02237485907971859,
+ "learning_rate": 0.0006,
+ "loss": 4.628812789916992,
+ "step": 3771
+ },
+ {
+ "epoch": 52.391266375545854,
+ "grad_norm": 0.02196541801095009,
+ "learning_rate": 0.0006,
+ "loss": 4.572988510131836,
+ "step": 3772
+ },
+ {
+ "epoch": 52.40524017467249,
+ "grad_norm": 0.01930229924619198,
+ "learning_rate": 0.0006,
+ "loss": 4.588339805603027,
+ "step": 3773
+ },
+ {
+ "epoch": 52.419213973799124,
+ "grad_norm": 0.02222434990108013,
+ "learning_rate": 0.0006,
+ "loss": 4.766482353210449,
+ "step": 3774
+ },
+ {
+ "epoch": 52.43318777292576,
+ "grad_norm": 0.028806306421756744,
+ "learning_rate": 0.0006,
+ "loss": 4.617892742156982,
+ "step": 3775
+ },
+ {
+ "epoch": 52.4471615720524,
+ "grad_norm": 0.03306042402982712,
+ "learning_rate": 0.0006,
+ "loss": 4.665334701538086,
+ "step": 3776
+ },
+ {
+ "epoch": 52.46113537117904,
+ "grad_norm": 0.0358956903219223,
+ "learning_rate": 0.0006,
+ "loss": 4.566668510437012,
+ "step": 3777
+ },
+ {
+ "epoch": 52.47510917030568,
+ "grad_norm": 0.030406497418880463,
+ "learning_rate": 0.0006,
+ "loss": 4.513612270355225,
+ "step": 3778
+ },
+ {
+ "epoch": 52.48908296943232,
+ "grad_norm": 0.030295712873339653,
+ "learning_rate": 0.0006,
+ "loss": 4.640189170837402,
+ "step": 3779
+ },
+ {
+ "epoch": 52.50305676855895,
+ "grad_norm": 0.024981003254652023,
+ "learning_rate": 0.0006,
+ "loss": 4.57558012008667,
+ "step": 3780
+ },
+ {
+ "epoch": 52.51703056768559,
+ "grad_norm": 0.024990713223814964,
+ "learning_rate": 0.0006,
+ "loss": 4.708589553833008,
+ "step": 3781
+ },
+ {
+ "epoch": 52.531004366812226,
+ "grad_norm": 0.025034775957465172,
+ "learning_rate": 0.0006,
+ "loss": 4.70468282699585,
+ "step": 3782
+ },
+ {
+ "epoch": 52.544978165938865,
+ "grad_norm": 0.028243759647011757,
+ "learning_rate": 0.0006,
+ "loss": 4.712060451507568,
+ "step": 3783
+ },
+ {
+ "epoch": 52.5589519650655,
+ "grad_norm": 0.026984544470906258,
+ "learning_rate": 0.0006,
+ "loss": 4.505004405975342,
+ "step": 3784
+ },
+ {
+ "epoch": 52.57292576419214,
+ "grad_norm": 0.050843529403209686,
+ "learning_rate": 0.0006,
+ "loss": 4.593271255493164,
+ "step": 3785
+ },
+ {
+ "epoch": 52.58689956331878,
+ "grad_norm": 0.25422802567481995,
+ "learning_rate": 0.0006,
+ "loss": 4.7178802490234375,
+ "step": 3786
+ },
+ {
+ "epoch": 52.60087336244541,
+ "grad_norm": 0.6078406572341919,
+ "learning_rate": 0.0006,
+ "loss": 6.144372940063477,
+ "step": 3787
+ },
+ {
+ "epoch": 52.61484716157205,
+ "grad_norm": 0.25933101773262024,
+ "learning_rate": 0.0006,
+ "loss": 5.971083641052246,
+ "step": 3788
+ },
+ {
+ "epoch": 52.62882096069869,
+ "grad_norm": 0.294657826423645,
+ "learning_rate": 0.0006,
+ "loss": 6.204090118408203,
+ "step": 3789
+ },
+ {
+ "epoch": 52.64279475982533,
+ "grad_norm": 0.23641522228717804,
+ "learning_rate": 0.0006,
+ "loss": 6.869574546813965,
+ "step": 3790
+ },
+ {
+ "epoch": 52.65676855895197,
+ "grad_norm": 0.23717311024665833,
+ "learning_rate": 0.0006,
+ "loss": 6.739034175872803,
+ "step": 3791
+ },
+ {
+ "epoch": 52.670742358078606,
+ "grad_norm": 0.10300362855195999,
+ "learning_rate": 0.0006,
+ "loss": 6.49191951751709,
+ "step": 3792
+ },
+ {
+ "epoch": 52.68471615720524,
+ "grad_norm": 0.1484968513250351,
+ "learning_rate": 0.0006,
+ "loss": 6.6021270751953125,
+ "step": 3793
+ },
+ {
+ "epoch": 52.698689956331876,
+ "grad_norm": 0.08052115142345428,
+ "learning_rate": 0.0006,
+ "loss": 6.432796478271484,
+ "step": 3794
+ },
+ {
+ "epoch": 52.712663755458514,
+ "grad_norm": 0.06869253516197205,
+ "learning_rate": 0.0006,
+ "loss": 6.212784767150879,
+ "step": 3795
+ },
+ {
+ "epoch": 52.72663755458515,
+ "grad_norm": 0.05589582398533821,
+ "learning_rate": 0.0006,
+ "loss": 6.168288707733154,
+ "step": 3796
+ },
+ {
+ "epoch": 52.74061135371179,
+ "grad_norm": 0.04840940609574318,
+ "learning_rate": 0.0006,
+ "loss": 6.163174629211426,
+ "step": 3797
+ },
+ {
+ "epoch": 52.75458515283843,
+ "grad_norm": 0.04578601196408272,
+ "learning_rate": 0.0006,
+ "loss": 6.038390636444092,
+ "step": 3798
+ },
+ {
+ "epoch": 52.76855895196506,
+ "grad_norm": 0.04102053865790367,
+ "learning_rate": 0.0006,
+ "loss": 5.929539680480957,
+ "step": 3799
+ },
+ {
+ "epoch": 52.7825327510917,
+ "grad_norm": 0.038398485630750656,
+ "learning_rate": 0.0006,
+ "loss": 6.007378101348877,
+ "step": 3800
+ },
+ {
+ "epoch": 52.79650655021834,
+ "grad_norm": 0.03466642275452614,
+ "learning_rate": 0.0006,
+ "loss": 5.767291069030762,
+ "step": 3801
+ },
+ {
+ "epoch": 52.81048034934498,
+ "grad_norm": 0.03328905999660492,
+ "learning_rate": 0.0006,
+ "loss": 5.785032272338867,
+ "step": 3802
+ },
+ {
+ "epoch": 52.82445414847162,
+ "grad_norm": 0.03186681121587753,
+ "learning_rate": 0.0006,
+ "loss": 5.6942057609558105,
+ "step": 3803
+ },
+ {
+ "epoch": 52.838427947598255,
+ "grad_norm": 0.030553551390767097,
+ "learning_rate": 0.0006,
+ "loss": 5.680774688720703,
+ "step": 3804
+ },
+ {
+ "epoch": 52.852401746724894,
+ "grad_norm": 0.02578839473426342,
+ "learning_rate": 0.0006,
+ "loss": 5.703461647033691,
+ "step": 3805
+ },
+ {
+ "epoch": 52.866375545851525,
+ "grad_norm": 0.026261312887072563,
+ "learning_rate": 0.0006,
+ "loss": 5.642549514770508,
+ "step": 3806
+ },
+ {
+ "epoch": 52.880349344978164,
+ "grad_norm": 0.02782430872321129,
+ "learning_rate": 0.0006,
+ "loss": 5.563597679138184,
+ "step": 3807
+ },
+ {
+ "epoch": 52.8943231441048,
+ "grad_norm": 0.024467799812555313,
+ "learning_rate": 0.0006,
+ "loss": 5.583395957946777,
+ "step": 3808
+ },
+ {
+ "epoch": 52.90829694323144,
+ "grad_norm": 0.022602897137403488,
+ "learning_rate": 0.0006,
+ "loss": 5.50742769241333,
+ "step": 3809
+ },
+ {
+ "epoch": 52.92227074235808,
+ "grad_norm": 0.024292029440402985,
+ "learning_rate": 0.0006,
+ "loss": 5.367804527282715,
+ "step": 3810
+ },
+ {
+ "epoch": 52.93624454148472,
+ "grad_norm": 0.021062295883893967,
+ "learning_rate": 0.0006,
+ "loss": 5.423705101013184,
+ "step": 3811
+ },
+ {
+ "epoch": 52.95021834061135,
+ "grad_norm": 0.022226469591259956,
+ "learning_rate": 0.0006,
+ "loss": 5.344039440155029,
+ "step": 3812
+ },
+ {
+ "epoch": 52.96419213973799,
+ "grad_norm": 0.023691752925515175,
+ "learning_rate": 0.0006,
+ "loss": 5.2581892013549805,
+ "step": 3813
+ },
+ {
+ "epoch": 52.97816593886463,
+ "grad_norm": 0.02181999757885933,
+ "learning_rate": 0.0006,
+ "loss": 5.2710113525390625,
+ "step": 3814
+ },
+ {
+ "epoch": 52.992139737991266,
+ "grad_norm": 0.025295857340097427,
+ "learning_rate": 0.0006,
+ "loss": 5.278355598449707,
+ "step": 3815
+ },
+ {
+ "epoch": 53.0,
+ "grad_norm": 0.02216828241944313,
+ "learning_rate": 0.0006,
+ "loss": 5.127945899963379,
+ "step": 3816
+ },
+ {
+ "epoch": 53.0,
+ "eval_loss": 5.3411736488342285,
+ "eval_runtime": 56.2837,
+ "eval_samples_per_second": 43.387,
+ "eval_steps_per_second": 1.368,
+ "step": 3816
+ },
+ {
+ "epoch": 53.01397379912664,
+ "grad_norm": 0.022496026009321213,
+ "learning_rate": 0.0006,
+ "loss": 5.19655704498291,
+ "step": 3817
+ },
+ {
+ "epoch": 53.02794759825328,
+ "grad_norm": 0.03201168775558472,
+ "learning_rate": 0.0006,
+ "loss": 5.094621181488037,
+ "step": 3818
+ },
+ {
+ "epoch": 53.041921397379916,
+ "grad_norm": 0.02683178335428238,
+ "learning_rate": 0.0006,
+ "loss": 5.19899845123291,
+ "step": 3819
+ },
+ {
+ "epoch": 53.05589519650655,
+ "grad_norm": 0.022659169510006905,
+ "learning_rate": 0.0006,
+ "loss": 5.184103012084961,
+ "step": 3820
+ },
+ {
+ "epoch": 53.069868995633186,
+ "grad_norm": 0.024784443899989128,
+ "learning_rate": 0.0006,
+ "loss": 5.124286651611328,
+ "step": 3821
+ },
+ {
+ "epoch": 53.083842794759825,
+ "grad_norm": 0.028710370883345604,
+ "learning_rate": 0.0006,
+ "loss": 5.130426406860352,
+ "step": 3822
+ },
+ {
+ "epoch": 53.09781659388646,
+ "grad_norm": 0.02366088517010212,
+ "learning_rate": 0.0006,
+ "loss": 5.1453399658203125,
+ "step": 3823
+ },
+ {
+ "epoch": 53.1117903930131,
+ "grad_norm": 0.01813938096165657,
+ "learning_rate": 0.0006,
+ "loss": 5.05271053314209,
+ "step": 3824
+ },
+ {
+ "epoch": 53.12576419213974,
+ "grad_norm": 0.02051435597240925,
+ "learning_rate": 0.0006,
+ "loss": 5.023038864135742,
+ "step": 3825
+ },
+ {
+ "epoch": 53.13973799126637,
+ "grad_norm": 0.02185754105448723,
+ "learning_rate": 0.0006,
+ "loss": 4.8950018882751465,
+ "step": 3826
+ },
+ {
+ "epoch": 53.15371179039301,
+ "grad_norm": 0.017223743721842766,
+ "learning_rate": 0.0006,
+ "loss": 5.078621864318848,
+ "step": 3827
+ },
+ {
+ "epoch": 53.16768558951965,
+ "grad_norm": 0.01832217164337635,
+ "learning_rate": 0.0006,
+ "loss": 4.921341419219971,
+ "step": 3828
+ },
+ {
+ "epoch": 53.18165938864629,
+ "grad_norm": 0.02114005573093891,
+ "learning_rate": 0.0006,
+ "loss": 4.997865676879883,
+ "step": 3829
+ },
+ {
+ "epoch": 53.19563318777293,
+ "grad_norm": 0.021426431834697723,
+ "learning_rate": 0.0006,
+ "loss": 4.877785682678223,
+ "step": 3830
+ },
+ {
+ "epoch": 53.209606986899566,
+ "grad_norm": 0.021327560767531395,
+ "learning_rate": 0.0006,
+ "loss": 4.847512245178223,
+ "step": 3831
+ },
+ {
+ "epoch": 53.223580786026204,
+ "grad_norm": 0.027825530618429184,
+ "learning_rate": 0.0006,
+ "loss": 5.004823207855225,
+ "step": 3832
+ },
+ {
+ "epoch": 53.237554585152836,
+ "grad_norm": 0.03756481409072876,
+ "learning_rate": 0.0006,
+ "loss": 4.817290306091309,
+ "step": 3833
+ },
+ {
+ "epoch": 53.251528384279474,
+ "grad_norm": 0.04933721572160721,
+ "learning_rate": 0.0006,
+ "loss": 4.863317489624023,
+ "step": 3834
+ },
+ {
+ "epoch": 53.26550218340611,
+ "grad_norm": 0.04588694870471954,
+ "learning_rate": 0.0006,
+ "loss": 4.864742755889893,
+ "step": 3835
+ },
+ {
+ "epoch": 53.27947598253275,
+ "grad_norm": 0.02728264033794403,
+ "learning_rate": 0.0006,
+ "loss": 4.9068756103515625,
+ "step": 3836
+ },
+ {
+ "epoch": 53.29344978165939,
+ "grad_norm": 0.036240726709365845,
+ "learning_rate": 0.0006,
+ "loss": 4.902045249938965,
+ "step": 3837
+ },
+ {
+ "epoch": 53.30742358078603,
+ "grad_norm": 0.02703195810317993,
+ "learning_rate": 0.0006,
+ "loss": 4.908980369567871,
+ "step": 3838
+ },
+ {
+ "epoch": 53.32139737991266,
+ "grad_norm": 0.024894049391150475,
+ "learning_rate": 0.0006,
+ "loss": 4.794848442077637,
+ "step": 3839
+ },
+ {
+ "epoch": 53.3353711790393,
+ "grad_norm": 0.02389993891119957,
+ "learning_rate": 0.0006,
+ "loss": 4.8234734535217285,
+ "step": 3840
+ },
+ {
+ "epoch": 53.34934497816594,
+ "grad_norm": 0.02486400306224823,
+ "learning_rate": 0.0006,
+ "loss": 4.851463794708252,
+ "step": 3841
+ },
+ {
+ "epoch": 53.36331877729258,
+ "grad_norm": 0.02613813802599907,
+ "learning_rate": 0.0006,
+ "loss": 4.860444068908691,
+ "step": 3842
+ },
+ {
+ "epoch": 53.377292576419215,
+ "grad_norm": 0.0229744054377079,
+ "learning_rate": 0.0006,
+ "loss": 4.709606170654297,
+ "step": 3843
+ },
+ {
+ "epoch": 53.391266375545854,
+ "grad_norm": 0.02123042568564415,
+ "learning_rate": 0.0006,
+ "loss": 4.654613494873047,
+ "step": 3844
+ },
+ {
+ "epoch": 53.40524017467249,
+ "grad_norm": 0.017352338880300522,
+ "learning_rate": 0.0006,
+ "loss": 4.769761085510254,
+ "step": 3845
+ },
+ {
+ "epoch": 53.419213973799124,
+ "grad_norm": 0.019808724522590637,
+ "learning_rate": 0.0006,
+ "loss": 4.98514461517334,
+ "step": 3846
+ },
+ {
+ "epoch": 53.43318777292576,
+ "grad_norm": 0.018776152282953262,
+ "learning_rate": 0.0006,
+ "loss": 4.749990463256836,
+ "step": 3847
+ },
+ {
+ "epoch": 53.4471615720524,
+ "grad_norm": 0.02017521858215332,
+ "learning_rate": 0.0006,
+ "loss": 4.757381439208984,
+ "step": 3848
+ },
+ {
+ "epoch": 53.46113537117904,
+ "grad_norm": 0.019509416073560715,
+ "learning_rate": 0.0006,
+ "loss": 4.8111572265625,
+ "step": 3849
+ },
+ {
+ "epoch": 53.47510917030568,
+ "grad_norm": 0.015543580986559391,
+ "learning_rate": 0.0006,
+ "loss": 4.7501678466796875,
+ "step": 3850
+ },
+ {
+ "epoch": 53.48908296943232,
+ "grad_norm": 0.017941651865839958,
+ "learning_rate": 0.0006,
+ "loss": 4.701838493347168,
+ "step": 3851
+ },
+ {
+ "epoch": 53.50305676855895,
+ "grad_norm": 0.01678212359547615,
+ "learning_rate": 0.0006,
+ "loss": 4.787798881530762,
+ "step": 3852
+ },
+ {
+ "epoch": 53.51703056768559,
+ "grad_norm": 0.019050654023885727,
+ "learning_rate": 0.0006,
+ "loss": 4.754859924316406,
+ "step": 3853
+ },
+ {
+ "epoch": 53.531004366812226,
+ "grad_norm": 0.017406078055500984,
+ "learning_rate": 0.0006,
+ "loss": 4.752772808074951,
+ "step": 3854
+ },
+ {
+ "epoch": 53.544978165938865,
+ "grad_norm": 0.01773344725370407,
+ "learning_rate": 0.0006,
+ "loss": 4.651812553405762,
+ "step": 3855
+ },
+ {
+ "epoch": 53.5589519650655,
+ "grad_norm": 0.017221346497535706,
+ "learning_rate": 0.0006,
+ "loss": 4.792244911193848,
+ "step": 3856
+ },
+ {
+ "epoch": 53.57292576419214,
+ "grad_norm": 0.01722758449614048,
+ "learning_rate": 0.0006,
+ "loss": 4.7902960777282715,
+ "step": 3857
+ },
+ {
+ "epoch": 53.58689956331878,
+ "grad_norm": 0.020445192232728004,
+ "learning_rate": 0.0006,
+ "loss": 4.67392110824585,
+ "step": 3858
+ },
+ {
+ "epoch": 53.60087336244541,
+ "grad_norm": 0.02267114631831646,
+ "learning_rate": 0.0006,
+ "loss": 4.796089172363281,
+ "step": 3859
+ },
+ {
+ "epoch": 53.61484716157205,
+ "grad_norm": 0.027129942551255226,
+ "learning_rate": 0.0006,
+ "loss": 4.774303913116455,
+ "step": 3860
+ },
+ {
+ "epoch": 53.62882096069869,
+ "grad_norm": 0.025454839691519737,
+ "learning_rate": 0.0006,
+ "loss": 4.708813190460205,
+ "step": 3861
+ },
+ {
+ "epoch": 53.64279475982533,
+ "grad_norm": 0.019248757511377335,
+ "learning_rate": 0.0006,
+ "loss": 4.760467529296875,
+ "step": 3862
+ },
+ {
+ "epoch": 53.65676855895197,
+ "grad_norm": 0.01791176199913025,
+ "learning_rate": 0.0006,
+ "loss": 4.789203643798828,
+ "step": 3863
+ },
+ {
+ "epoch": 53.670742358078606,
+ "grad_norm": 0.016306612640619278,
+ "learning_rate": 0.0006,
+ "loss": 4.732605934143066,
+ "step": 3864
+ },
+ {
+ "epoch": 53.68471615720524,
+ "grad_norm": 0.017836976796388626,
+ "learning_rate": 0.0006,
+ "loss": 4.691871643066406,
+ "step": 3865
+ },
+ {
+ "epoch": 53.698689956331876,
+ "grad_norm": 0.015102861449122429,
+ "learning_rate": 0.0006,
+ "loss": 4.724156856536865,
+ "step": 3866
+ },
+ {
+ "epoch": 53.712663755458514,
+ "grad_norm": 0.017458142712712288,
+ "learning_rate": 0.0006,
+ "loss": 4.73524808883667,
+ "step": 3867
+ },
+ {
+ "epoch": 53.72663755458515,
+ "grad_norm": 0.01591932773590088,
+ "learning_rate": 0.0006,
+ "loss": 4.705888748168945,
+ "step": 3868
+ },
+ {
+ "epoch": 53.74061135371179,
+ "grad_norm": 0.01713593676686287,
+ "learning_rate": 0.0006,
+ "loss": 4.644516944885254,
+ "step": 3869
+ },
+ {
+ "epoch": 53.75458515283843,
+ "grad_norm": 0.01905885338783264,
+ "learning_rate": 0.0006,
+ "loss": 4.698416233062744,
+ "step": 3870
+ },
+ {
+ "epoch": 53.76855895196506,
+ "grad_norm": 0.01766437292098999,
+ "learning_rate": 0.0006,
+ "loss": 4.6796555519104,
+ "step": 3871
+ },
+ {
+ "epoch": 53.7825327510917,
+ "grad_norm": 0.01552243810147047,
+ "learning_rate": 0.0006,
+ "loss": 4.78138542175293,
+ "step": 3872
+ },
+ {
+ "epoch": 53.79650655021834,
+ "grad_norm": 0.014191331341862679,
+ "learning_rate": 0.0006,
+ "loss": 4.703588485717773,
+ "step": 3873
+ },
+ {
+ "epoch": 53.81048034934498,
+ "grad_norm": 0.015872279182076454,
+ "learning_rate": 0.0006,
+ "loss": 4.705611228942871,
+ "step": 3874
+ },
+ {
+ "epoch": 53.82445414847162,
+ "grad_norm": 0.01787560060620308,
+ "learning_rate": 0.0006,
+ "loss": 4.8227949142456055,
+ "step": 3875
+ },
+ {
+ "epoch": 53.838427947598255,
+ "grad_norm": 0.0158209390938282,
+ "learning_rate": 0.0006,
+ "loss": 4.645227432250977,
+ "step": 3876
+ },
+ {
+ "epoch": 53.852401746724894,
+ "grad_norm": 0.014362678863108158,
+ "learning_rate": 0.0006,
+ "loss": 4.7225422859191895,
+ "step": 3877
+ },
+ {
+ "epoch": 53.866375545851525,
+ "grad_norm": 0.014899161644279957,
+ "learning_rate": 0.0006,
+ "loss": 4.685914039611816,
+ "step": 3878
+ },
+ {
+ "epoch": 53.880349344978164,
+ "grad_norm": 0.01628505438566208,
+ "learning_rate": 0.0006,
+ "loss": 4.674065589904785,
+ "step": 3879
+ },
+ {
+ "epoch": 53.8943231441048,
+ "grad_norm": 0.01873331144452095,
+ "learning_rate": 0.0006,
+ "loss": 4.811744213104248,
+ "step": 3880
+ },
+ {
+ "epoch": 53.90829694323144,
+ "grad_norm": 0.02229069173336029,
+ "learning_rate": 0.0006,
+ "loss": 4.721514701843262,
+ "step": 3881
+ },
+ {
+ "epoch": 53.92227074235808,
+ "grad_norm": 0.028012916445732117,
+ "learning_rate": 0.0006,
+ "loss": 4.773603439331055,
+ "step": 3882
+ },
+ {
+ "epoch": 53.93624454148472,
+ "grad_norm": 0.036322783678770065,
+ "learning_rate": 0.0006,
+ "loss": 4.787134170532227,
+ "step": 3883
+ },
+ {
+ "epoch": 53.95021834061135,
+ "grad_norm": 0.048149142414331436,
+ "learning_rate": 0.0006,
+ "loss": 4.736196517944336,
+ "step": 3884
+ },
+ {
+ "epoch": 53.96419213973799,
+ "grad_norm": 0.051918916404247284,
+ "learning_rate": 0.0006,
+ "loss": 4.686728000640869,
+ "step": 3885
+ },
+ {
+ "epoch": 53.97816593886463,
+ "grad_norm": 0.0481342189013958,
+ "learning_rate": 0.0006,
+ "loss": 4.733730792999268,
+ "step": 3886
+ },
+ {
+ "epoch": 53.992139737991266,
+ "grad_norm": 0.045424818992614746,
+ "learning_rate": 0.0006,
+ "loss": 4.748723983764648,
+ "step": 3887
+ },
+ {
+ "epoch": 54.0,
+ "grad_norm": 0.04153316840529442,
+ "learning_rate": 0.0006,
+ "loss": 4.715220928192139,
+ "step": 3888
+ },
+ {
+ "epoch": 54.0,
+ "eval_loss": 4.964215278625488,
+ "eval_runtime": 56.7757,
+ "eval_samples_per_second": 43.011,
+ "eval_steps_per_second": 1.356,
+ "step": 3888
+ },
+ {
+ "epoch": 54.01397379912664,
+ "grad_norm": 0.03459750860929489,
+ "learning_rate": 0.0006,
+ "loss": 4.661557674407959,
+ "step": 3889
+ },
+ {
+ "epoch": 54.02794759825328,
+ "grad_norm": 0.025409208610653877,
+ "learning_rate": 0.0006,
+ "loss": 4.7065582275390625,
+ "step": 3890
+ },
+ {
+ "epoch": 54.041921397379916,
+ "grad_norm": 0.03202550485730171,
+ "learning_rate": 0.0006,
+ "loss": 4.751070976257324,
+ "step": 3891
+ },
+ {
+ "epoch": 54.05589519650655,
+ "grad_norm": 0.02968018501996994,
+ "learning_rate": 0.0006,
+ "loss": 4.635388374328613,
+ "step": 3892
+ },
+ {
+ "epoch": 54.069868995633186,
+ "grad_norm": 0.02757406234741211,
+ "learning_rate": 0.0006,
+ "loss": 4.706214427947998,
+ "step": 3893
+ },
+ {
+ "epoch": 54.083842794759825,
+ "grad_norm": 0.025223620235919952,
+ "learning_rate": 0.0006,
+ "loss": 4.703792572021484,
+ "step": 3894
+ },
+ {
+ "epoch": 54.09781659388646,
+ "grad_norm": 0.02431778982281685,
+ "learning_rate": 0.0006,
+ "loss": 4.658116340637207,
+ "step": 3895
+ },
+ {
+ "epoch": 54.1117903930131,
+ "grad_norm": 0.02792172320187092,
+ "learning_rate": 0.0006,
+ "loss": 4.638895034790039,
+ "step": 3896
+ },
+ {
+ "epoch": 54.12576419213974,
+ "grad_norm": 0.032253511250019073,
+ "learning_rate": 0.0006,
+ "loss": 4.693902492523193,
+ "step": 3897
+ },
+ {
+ "epoch": 54.13973799126637,
+ "grad_norm": 0.027044931426644325,
+ "learning_rate": 0.0006,
+ "loss": 4.575805187225342,
+ "step": 3898
+ },
+ {
+ "epoch": 54.15371179039301,
+ "grad_norm": 0.023270361125469208,
+ "learning_rate": 0.0006,
+ "loss": 4.67645263671875,
+ "step": 3899
+ },
+ {
+ "epoch": 54.16768558951965,
+ "grad_norm": 0.01929079182446003,
+ "learning_rate": 0.0006,
+ "loss": 4.68062686920166,
+ "step": 3900
+ },
+ {
+ "epoch": 54.18165938864629,
+ "grad_norm": 0.020343473181128502,
+ "learning_rate": 0.0006,
+ "loss": 4.708086967468262,
+ "step": 3901
+ },
+ {
+ "epoch": 54.19563318777293,
+ "grad_norm": 0.0211793202906847,
+ "learning_rate": 0.0006,
+ "loss": 4.677323341369629,
+ "step": 3902
+ },
+ {
+ "epoch": 54.209606986899566,
+ "grad_norm": 0.020341217517852783,
+ "learning_rate": 0.0006,
+ "loss": 4.760937690734863,
+ "step": 3903
+ },
+ {
+ "epoch": 54.223580786026204,
+ "grad_norm": 0.019421054050326347,
+ "learning_rate": 0.0006,
+ "loss": 4.508834362030029,
+ "step": 3904
+ },
+ {
+ "epoch": 54.237554585152836,
+ "grad_norm": 0.016701679676771164,
+ "learning_rate": 0.0006,
+ "loss": 4.664399147033691,
+ "step": 3905
+ },
+ {
+ "epoch": 54.251528384279474,
+ "grad_norm": 0.016561470925807953,
+ "learning_rate": 0.0006,
+ "loss": 4.742990493774414,
+ "step": 3906
+ },
+ {
+ "epoch": 54.26550218340611,
+ "grad_norm": 0.017446085810661316,
+ "learning_rate": 0.0006,
+ "loss": 4.63038444519043,
+ "step": 3907
+ },
+ {
+ "epoch": 54.27947598253275,
+ "grad_norm": 0.014340086840093136,
+ "learning_rate": 0.0006,
+ "loss": 4.688791275024414,
+ "step": 3908
+ },
+ {
+ "epoch": 54.29344978165939,
+ "grad_norm": 0.013580698519945145,
+ "learning_rate": 0.0006,
+ "loss": 4.625302314758301,
+ "step": 3909
+ },
+ {
+ "epoch": 54.30742358078603,
+ "grad_norm": 0.014931539073586464,
+ "learning_rate": 0.0006,
+ "loss": 4.6167497634887695,
+ "step": 3910
+ },
+ {
+ "epoch": 54.32139737991266,
+ "grad_norm": 0.013492150232195854,
+ "learning_rate": 0.0006,
+ "loss": 4.657521724700928,
+ "step": 3911
+ },
+ {
+ "epoch": 54.3353711790393,
+ "grad_norm": 0.015151413157582283,
+ "learning_rate": 0.0006,
+ "loss": 4.634521484375,
+ "step": 3912
+ },
+ {
+ "epoch": 54.34934497816594,
+ "grad_norm": 0.01302040833979845,
+ "learning_rate": 0.0006,
+ "loss": 4.718928337097168,
+ "step": 3913
+ },
+ {
+ "epoch": 54.36331877729258,
+ "grad_norm": 0.013433529995381832,
+ "learning_rate": 0.0006,
+ "loss": 4.584803581237793,
+ "step": 3914
+ },
+ {
+ "epoch": 54.377292576419215,
+ "grad_norm": 0.012987039051949978,
+ "learning_rate": 0.0006,
+ "loss": 4.6393585205078125,
+ "step": 3915
+ },
+ {
+ "epoch": 54.391266375545854,
+ "grad_norm": 0.013695075176656246,
+ "learning_rate": 0.0006,
+ "loss": 4.556313514709473,
+ "step": 3916
+ },
+ {
+ "epoch": 54.40524017467249,
+ "grad_norm": 0.013718255795538425,
+ "learning_rate": 0.0006,
+ "loss": 4.576223850250244,
+ "step": 3917
+ },
+ {
+ "epoch": 54.419213973799124,
+ "grad_norm": 0.013699112460017204,
+ "learning_rate": 0.0006,
+ "loss": 4.63889217376709,
+ "step": 3918
+ },
+ {
+ "epoch": 54.43318777292576,
+ "grad_norm": 0.013516264036297798,
+ "learning_rate": 0.0006,
+ "loss": 4.669532299041748,
+ "step": 3919
+ },
+ {
+ "epoch": 54.4471615720524,
+ "grad_norm": 0.013597294688224792,
+ "learning_rate": 0.0006,
+ "loss": 4.6644287109375,
+ "step": 3920
+ },
+ {
+ "epoch": 54.46113537117904,
+ "grad_norm": 0.013185849413275719,
+ "learning_rate": 0.0006,
+ "loss": 4.636900901794434,
+ "step": 3921
+ },
+ {
+ "epoch": 54.47510917030568,
+ "grad_norm": 0.014311990700662136,
+ "learning_rate": 0.0006,
+ "loss": 4.6047587394714355,
+ "step": 3922
+ },
+ {
+ "epoch": 54.48908296943232,
+ "grad_norm": 0.012426027096807957,
+ "learning_rate": 0.0006,
+ "loss": 4.6161394119262695,
+ "step": 3923
+ },
+ {
+ "epoch": 54.50305676855895,
+ "grad_norm": 0.01283493172377348,
+ "learning_rate": 0.0006,
+ "loss": 4.584190368652344,
+ "step": 3924
+ },
+ {
+ "epoch": 54.51703056768559,
+ "grad_norm": 0.01339943427592516,
+ "learning_rate": 0.0006,
+ "loss": 4.624184608459473,
+ "step": 3925
+ },
+ {
+ "epoch": 54.531004366812226,
+ "grad_norm": 0.012859724462032318,
+ "learning_rate": 0.0006,
+ "loss": 4.6174635887146,
+ "step": 3926
+ },
+ {
+ "epoch": 54.544978165938865,
+ "grad_norm": 0.014420023187994957,
+ "learning_rate": 0.0006,
+ "loss": 4.536544322967529,
+ "step": 3927
+ },
+ {
+ "epoch": 54.5589519650655,
+ "grad_norm": 0.014237185008823872,
+ "learning_rate": 0.0006,
+ "loss": 4.681514739990234,
+ "step": 3928
+ },
+ {
+ "epoch": 54.57292576419214,
+ "grad_norm": 0.013340328820049763,
+ "learning_rate": 0.0006,
+ "loss": 4.579689979553223,
+ "step": 3929
+ },
+ {
+ "epoch": 54.58689956331878,
+ "grad_norm": 0.013894530944526196,
+ "learning_rate": 0.0006,
+ "loss": 4.638958930969238,
+ "step": 3930
+ },
+ {
+ "epoch": 54.60087336244541,
+ "grad_norm": 0.016181141138076782,
+ "learning_rate": 0.0006,
+ "loss": 4.573508262634277,
+ "step": 3931
+ },
+ {
+ "epoch": 54.61484716157205,
+ "grad_norm": 0.013221992179751396,
+ "learning_rate": 0.0006,
+ "loss": 4.6229047775268555,
+ "step": 3932
+ },
+ {
+ "epoch": 54.62882096069869,
+ "grad_norm": 0.013241561129689217,
+ "learning_rate": 0.0006,
+ "loss": 4.641152858734131,
+ "step": 3933
+ },
+ {
+ "epoch": 54.64279475982533,
+ "grad_norm": 0.016237538307905197,
+ "learning_rate": 0.0006,
+ "loss": 4.6421217918396,
+ "step": 3934
+ },
+ {
+ "epoch": 54.65676855895197,
+ "grad_norm": 0.020866507664322853,
+ "learning_rate": 0.0006,
+ "loss": 4.646058082580566,
+ "step": 3935
+ },
+ {
+ "epoch": 54.670742358078606,
+ "grad_norm": 0.022182893007993698,
+ "learning_rate": 0.0006,
+ "loss": 4.554501533508301,
+ "step": 3936
+ },
+ {
+ "epoch": 54.68471615720524,
+ "grad_norm": 0.02254868485033512,
+ "learning_rate": 0.0006,
+ "loss": 4.563869953155518,
+ "step": 3937
+ },
+ {
+ "epoch": 54.698689956331876,
+ "grad_norm": 0.019730940461158752,
+ "learning_rate": 0.0006,
+ "loss": 4.592165946960449,
+ "step": 3938
+ },
+ {
+ "epoch": 54.712663755458514,
+ "grad_norm": 0.015146561898291111,
+ "learning_rate": 0.0006,
+ "loss": 4.6228742599487305,
+ "step": 3939
+ },
+ {
+ "epoch": 54.72663755458515,
+ "grad_norm": 0.01824716106057167,
+ "learning_rate": 0.0006,
+ "loss": 4.716236114501953,
+ "step": 3940
+ },
+ {
+ "epoch": 54.74061135371179,
+ "grad_norm": 0.017730550840497017,
+ "learning_rate": 0.0006,
+ "loss": 4.650232315063477,
+ "step": 3941
+ },
+ {
+ "epoch": 54.75458515283843,
+ "grad_norm": 0.014875595457851887,
+ "learning_rate": 0.0006,
+ "loss": 4.6377458572387695,
+ "step": 3942
+ },
+ {
+ "epoch": 54.76855895196506,
+ "grad_norm": 0.01613209955394268,
+ "learning_rate": 0.0006,
+ "loss": 4.5768537521362305,
+ "step": 3943
+ },
+ {
+ "epoch": 54.7825327510917,
+ "grad_norm": 0.017311280593276024,
+ "learning_rate": 0.0006,
+ "loss": 4.625744342803955,
+ "step": 3944
+ },
+ {
+ "epoch": 54.79650655021834,
+ "grad_norm": 0.018770085647702217,
+ "learning_rate": 0.0006,
+ "loss": 4.613787651062012,
+ "step": 3945
+ },
+ {
+ "epoch": 54.81048034934498,
+ "grad_norm": 0.0165257528424263,
+ "learning_rate": 0.0006,
+ "loss": 4.5991315841674805,
+ "step": 3946
+ },
+ {
+ "epoch": 54.82445414847162,
+ "grad_norm": 0.017433857545256615,
+ "learning_rate": 0.0006,
+ "loss": 4.513872146606445,
+ "step": 3947
+ },
+ {
+ "epoch": 54.838427947598255,
+ "grad_norm": 0.020126905292272568,
+ "learning_rate": 0.0006,
+ "loss": 4.616185188293457,
+ "step": 3948
+ },
+ {
+ "epoch": 54.852401746724894,
+ "grad_norm": 0.01819358579814434,
+ "learning_rate": 0.0006,
+ "loss": 4.535691261291504,
+ "step": 3949
+ },
+ {
+ "epoch": 54.866375545851525,
+ "grad_norm": 0.01899927295744419,
+ "learning_rate": 0.0006,
+ "loss": 4.595199108123779,
+ "step": 3950
+ },
+ {
+ "epoch": 54.880349344978164,
+ "grad_norm": 0.02005566656589508,
+ "learning_rate": 0.0006,
+ "loss": 4.633289337158203,
+ "step": 3951
+ },
+ {
+ "epoch": 54.8943231441048,
+ "grad_norm": 0.021570220589637756,
+ "learning_rate": 0.0006,
+ "loss": 4.592503547668457,
+ "step": 3952
+ },
+ {
+ "epoch": 54.90829694323144,
+ "grad_norm": 0.01883506216108799,
+ "learning_rate": 0.0006,
+ "loss": 4.6597795486450195,
+ "step": 3953
+ },
+ {
+ "epoch": 54.92227074235808,
+ "grad_norm": 0.018070021644234657,
+ "learning_rate": 0.0006,
+ "loss": 4.601474285125732,
+ "step": 3954
+ },
+ {
+ "epoch": 54.93624454148472,
+ "grad_norm": 0.018090378493070602,
+ "learning_rate": 0.0006,
+ "loss": 4.569846153259277,
+ "step": 3955
+ },
+ {
+ "epoch": 54.95021834061135,
+ "grad_norm": 0.018751563504338264,
+ "learning_rate": 0.0006,
+ "loss": 4.646064281463623,
+ "step": 3956
+ },
+ {
+ "epoch": 54.96419213973799,
+ "grad_norm": 0.017512254416942596,
+ "learning_rate": 0.0006,
+ "loss": 4.606670379638672,
+ "step": 3957
+ },
+ {
+ "epoch": 54.97816593886463,
+ "grad_norm": 0.018771110102534294,
+ "learning_rate": 0.0006,
+ "loss": 4.604315757751465,
+ "step": 3958
+ },
+ {
+ "epoch": 54.992139737991266,
+ "grad_norm": 0.01611141301691532,
+ "learning_rate": 0.0006,
+ "loss": 4.643245220184326,
+ "step": 3959
+ },
+ {
+ "epoch": 55.0,
+ "grad_norm": 0.0156728383153677,
+ "learning_rate": 0.0006,
+ "loss": 4.6739044189453125,
+ "step": 3960
+ },
+ {
+ "epoch": 55.0,
+ "eval_loss": 4.851032257080078,
+ "eval_runtime": 56.8699,
+ "eval_samples_per_second": 42.94,
+ "eval_steps_per_second": 1.354,
+ "step": 3960
+ },
+ {
+ "epoch": 55.01397379912664,
+ "grad_norm": 0.016214672476053238,
+ "learning_rate": 0.0006,
+ "loss": 4.571749687194824,
+ "step": 3961
+ },
+ {
+ "epoch": 55.02794759825328,
+ "grad_norm": 0.01649155467748642,
+ "learning_rate": 0.0006,
+ "loss": 4.553592205047607,
+ "step": 3962
+ },
+ {
+ "epoch": 55.041921397379916,
+ "grad_norm": 0.016793379560112953,
+ "learning_rate": 0.0006,
+ "loss": 4.650628566741943,
+ "step": 3963
+ },
+ {
+ "epoch": 55.05589519650655,
+ "grad_norm": 0.0191937405616045,
+ "learning_rate": 0.0006,
+ "loss": 4.620532989501953,
+ "step": 3964
+ },
+ {
+ "epoch": 55.069868995633186,
+ "grad_norm": 0.020976556465029716,
+ "learning_rate": 0.0006,
+ "loss": 4.6329545974731445,
+ "step": 3965
+ },
+ {
+ "epoch": 55.083842794759825,
+ "grad_norm": 0.023450477048754692,
+ "learning_rate": 0.0006,
+ "loss": 4.60828971862793,
+ "step": 3966
+ },
+ {
+ "epoch": 55.09781659388646,
+ "grad_norm": 0.026778312399983406,
+ "learning_rate": 0.0006,
+ "loss": 4.617481708526611,
+ "step": 3967
+ },
+ {
+ "epoch": 55.1117903930131,
+ "grad_norm": 0.02333568036556244,
+ "learning_rate": 0.0006,
+ "loss": 4.580722332000732,
+ "step": 3968
+ },
+ {
+ "epoch": 55.12576419213974,
+ "grad_norm": 0.018491310998797417,
+ "learning_rate": 0.0006,
+ "loss": 4.590953350067139,
+ "step": 3969
+ },
+ {
+ "epoch": 55.13973799126637,
+ "grad_norm": 0.020339803770184517,
+ "learning_rate": 0.0006,
+ "loss": 4.518374443054199,
+ "step": 3970
+ },
+ {
+ "epoch": 55.15371179039301,
+ "grad_norm": 0.01759706623852253,
+ "learning_rate": 0.0006,
+ "loss": 4.560522079467773,
+ "step": 3971
+ },
+ {
+ "epoch": 55.16768558951965,
+ "grad_norm": 0.015875980257987976,
+ "learning_rate": 0.0006,
+ "loss": 4.4795708656311035,
+ "step": 3972
+ },
+ {
+ "epoch": 55.18165938864629,
+ "grad_norm": 0.01650134287774563,
+ "learning_rate": 0.0006,
+ "loss": 4.57435941696167,
+ "step": 3973
+ },
+ {
+ "epoch": 55.19563318777293,
+ "grad_norm": 0.01734331250190735,
+ "learning_rate": 0.0006,
+ "loss": 4.661007881164551,
+ "step": 3974
+ },
+ {
+ "epoch": 55.209606986899566,
+ "grad_norm": 0.01651303470134735,
+ "learning_rate": 0.0006,
+ "loss": 4.511148452758789,
+ "step": 3975
+ },
+ {
+ "epoch": 55.223580786026204,
+ "grad_norm": 0.016189761459827423,
+ "learning_rate": 0.0006,
+ "loss": 4.592350006103516,
+ "step": 3976
+ },
+ {
+ "epoch": 55.237554585152836,
+ "grad_norm": 0.01833222061395645,
+ "learning_rate": 0.0006,
+ "loss": 4.506649971008301,
+ "step": 3977
+ },
+ {
+ "epoch": 55.251528384279474,
+ "grad_norm": 0.02433410845696926,
+ "learning_rate": 0.0006,
+ "loss": 4.49500036239624,
+ "step": 3978
+ },
+ {
+ "epoch": 55.26550218340611,
+ "grad_norm": 0.0314289927482605,
+ "learning_rate": 0.0006,
+ "loss": 4.507624626159668,
+ "step": 3979
+ },
+ {
+ "epoch": 55.27947598253275,
+ "grad_norm": 0.03013679012656212,
+ "learning_rate": 0.0006,
+ "loss": 4.533641815185547,
+ "step": 3980
+ },
+ {
+ "epoch": 55.29344978165939,
+ "grad_norm": 0.024524061009287834,
+ "learning_rate": 0.0006,
+ "loss": 4.5721116065979,
+ "step": 3981
+ },
+ {
+ "epoch": 55.30742358078603,
+ "grad_norm": 0.02640336938202381,
+ "learning_rate": 0.0006,
+ "loss": 4.652521133422852,
+ "step": 3982
+ },
+ {
+ "epoch": 55.32139737991266,
+ "grad_norm": 0.021309101954102516,
+ "learning_rate": 0.0006,
+ "loss": 4.600375175476074,
+ "step": 3983
+ },
+ {
+ "epoch": 55.3353711790393,
+ "grad_norm": 0.023008549585938454,
+ "learning_rate": 0.0006,
+ "loss": 4.53914737701416,
+ "step": 3984
+ },
+ {
+ "epoch": 55.34934497816594,
+ "grad_norm": 0.022491468116641045,
+ "learning_rate": 0.0006,
+ "loss": 4.661163330078125,
+ "step": 3985
+ },
+ {
+ "epoch": 55.36331877729258,
+ "grad_norm": 0.01904081553220749,
+ "learning_rate": 0.0006,
+ "loss": 4.5426177978515625,
+ "step": 3986
+ },
+ {
+ "epoch": 55.377292576419215,
+ "grad_norm": 0.019229518249630928,
+ "learning_rate": 0.0006,
+ "loss": 4.606066703796387,
+ "step": 3987
+ },
+ {
+ "epoch": 55.391266375545854,
+ "grad_norm": 0.01781383901834488,
+ "learning_rate": 0.0006,
+ "loss": 4.54642391204834,
+ "step": 3988
+ },
+ {
+ "epoch": 55.40524017467249,
+ "grad_norm": 0.0174600500613451,
+ "learning_rate": 0.0006,
+ "loss": 4.711891174316406,
+ "step": 3989
+ },
+ {
+ "epoch": 55.419213973799124,
+ "grad_norm": 0.018154339864850044,
+ "learning_rate": 0.0006,
+ "loss": 4.527122974395752,
+ "step": 3990
+ },
+ {
+ "epoch": 55.43318777292576,
+ "grad_norm": 0.01914648339152336,
+ "learning_rate": 0.0006,
+ "loss": 4.641975402832031,
+ "step": 3991
+ },
+ {
+ "epoch": 55.4471615720524,
+ "grad_norm": 0.017118625342845917,
+ "learning_rate": 0.0006,
+ "loss": 4.511425971984863,
+ "step": 3992
+ },
+ {
+ "epoch": 55.46113537117904,
+ "grad_norm": 0.01551489345729351,
+ "learning_rate": 0.0006,
+ "loss": 4.551141738891602,
+ "step": 3993
+ },
+ {
+ "epoch": 55.47510917030568,
+ "grad_norm": 0.015072065405547619,
+ "learning_rate": 0.0006,
+ "loss": 4.651115894317627,
+ "step": 3994
+ },
+ {
+ "epoch": 55.48908296943232,
+ "grad_norm": 0.017010986804962158,
+ "learning_rate": 0.0006,
+ "loss": 4.583282470703125,
+ "step": 3995
+ },
+ {
+ "epoch": 55.50305676855895,
+ "grad_norm": 0.01764831133186817,
+ "learning_rate": 0.0006,
+ "loss": 4.593364238739014,
+ "step": 3996
+ },
+ {
+ "epoch": 55.51703056768559,
+ "grad_norm": 0.01762884296476841,
+ "learning_rate": 0.0006,
+ "loss": 4.45671272277832,
+ "step": 3997
+ },
+ {
+ "epoch": 55.531004366812226,
+ "grad_norm": 0.01600457727909088,
+ "learning_rate": 0.0006,
+ "loss": 4.511260032653809,
+ "step": 3998
+ },
+ {
+ "epoch": 55.544978165938865,
+ "grad_norm": 0.015706021338701248,
+ "learning_rate": 0.0006,
+ "loss": 4.47593355178833,
+ "step": 3999
+ },
+ {
+ "epoch": 55.5589519650655,
+ "grad_norm": 0.017718996852636337,
+ "learning_rate": 0.0006,
+ "loss": 4.610543251037598,
+ "step": 4000
+ },
+ {
+ "epoch": 55.57292576419214,
+ "grad_norm": 0.015073757618665695,
+ "learning_rate": 0.0006,
+ "loss": 4.466620445251465,
+ "step": 4001
+ },
+ {
+ "epoch": 55.58689956331878,
+ "grad_norm": 0.01789247989654541,
+ "learning_rate": 0.0006,
+ "loss": 4.503409385681152,
+ "step": 4002
+ },
+ {
+ "epoch": 55.60087336244541,
+ "grad_norm": 0.022329598665237427,
+ "learning_rate": 0.0006,
+ "loss": 4.62099552154541,
+ "step": 4003
+ },
+ {
+ "epoch": 55.61484716157205,
+ "grad_norm": 0.02343829534947872,
+ "learning_rate": 0.0006,
+ "loss": 4.620813369750977,
+ "step": 4004
+ },
+ {
+ "epoch": 55.62882096069869,
+ "grad_norm": 0.023279687389731407,
+ "learning_rate": 0.0006,
+ "loss": 4.545765399932861,
+ "step": 4005
+ },
+ {
+ "epoch": 55.64279475982533,
+ "grad_norm": 0.020817464217543602,
+ "learning_rate": 0.0006,
+ "loss": 4.561470031738281,
+ "step": 4006
+ },
+ {
+ "epoch": 55.65676855895197,
+ "grad_norm": 0.020270058885216713,
+ "learning_rate": 0.0006,
+ "loss": 4.5875020027160645,
+ "step": 4007
+ },
+ {
+ "epoch": 55.670742358078606,
+ "grad_norm": 0.021085843443870544,
+ "learning_rate": 0.0006,
+ "loss": 4.504560470581055,
+ "step": 4008
+ },
+ {
+ "epoch": 55.68471615720524,
+ "grad_norm": 0.018507730215787888,
+ "learning_rate": 0.0006,
+ "loss": 4.520644664764404,
+ "step": 4009
+ },
+ {
+ "epoch": 55.698689956331876,
+ "grad_norm": 0.020571516826748848,
+ "learning_rate": 0.0006,
+ "loss": 4.539643287658691,
+ "step": 4010
+ },
+ {
+ "epoch": 55.712663755458514,
+ "grad_norm": 0.021484510973095894,
+ "learning_rate": 0.0006,
+ "loss": 4.628833293914795,
+ "step": 4011
+ },
+ {
+ "epoch": 55.72663755458515,
+ "grad_norm": 0.017693819478154182,
+ "learning_rate": 0.0006,
+ "loss": 4.6023783683776855,
+ "step": 4012
+ },
+ {
+ "epoch": 55.74061135371179,
+ "grad_norm": 0.020186059176921844,
+ "learning_rate": 0.0006,
+ "loss": 4.634884834289551,
+ "step": 4013
+ },
+ {
+ "epoch": 55.75458515283843,
+ "grad_norm": 0.01835780404508114,
+ "learning_rate": 0.0006,
+ "loss": 4.577791213989258,
+ "step": 4014
+ },
+ {
+ "epoch": 55.76855895196506,
+ "grad_norm": 0.018162988126277924,
+ "learning_rate": 0.0006,
+ "loss": 4.641597270965576,
+ "step": 4015
+ },
+ {
+ "epoch": 55.7825327510917,
+ "grad_norm": 0.018909504637122154,
+ "learning_rate": 0.0006,
+ "loss": 4.487447738647461,
+ "step": 4016
+ },
+ {
+ "epoch": 55.79650655021834,
+ "grad_norm": 0.018972251564264297,
+ "learning_rate": 0.0006,
+ "loss": 4.590317726135254,
+ "step": 4017
+ },
+ {
+ "epoch": 55.81048034934498,
+ "grad_norm": 0.01793607883155346,
+ "learning_rate": 0.0006,
+ "loss": 4.4994611740112305,
+ "step": 4018
+ },
+ {
+ "epoch": 55.82445414847162,
+ "grad_norm": 0.02123514749109745,
+ "learning_rate": 0.0006,
+ "loss": 4.479006767272949,
+ "step": 4019
+ },
+ {
+ "epoch": 55.838427947598255,
+ "grad_norm": 0.021246599033474922,
+ "learning_rate": 0.0006,
+ "loss": 4.4738640785217285,
+ "step": 4020
+ },
+ {
+ "epoch": 55.852401746724894,
+ "grad_norm": 0.020131602883338928,
+ "learning_rate": 0.0006,
+ "loss": 4.600308418273926,
+ "step": 4021
+ },
+ {
+ "epoch": 55.866375545851525,
+ "grad_norm": 0.019159460440278053,
+ "learning_rate": 0.0006,
+ "loss": 4.540908336639404,
+ "step": 4022
+ },
+ {
+ "epoch": 55.880349344978164,
+ "grad_norm": 0.023114705458283424,
+ "learning_rate": 0.0006,
+ "loss": 4.641247749328613,
+ "step": 4023
+ },
+ {
+ "epoch": 55.8943231441048,
+ "grad_norm": 0.02296135015785694,
+ "learning_rate": 0.0006,
+ "loss": 4.444624423980713,
+ "step": 4024
+ },
+ {
+ "epoch": 55.90829694323144,
+ "grad_norm": 0.02204655110836029,
+ "learning_rate": 0.0006,
+ "loss": 4.506349563598633,
+ "step": 4025
+ },
+ {
+ "epoch": 55.92227074235808,
+ "grad_norm": 0.019338225945830345,
+ "learning_rate": 0.0006,
+ "loss": 4.616047382354736,
+ "step": 4026
+ },
+ {
+ "epoch": 55.93624454148472,
+ "grad_norm": 0.020357884466648102,
+ "learning_rate": 0.0006,
+ "loss": 4.48170280456543,
+ "step": 4027
+ },
+ {
+ "epoch": 55.95021834061135,
+ "grad_norm": 0.02426772564649582,
+ "learning_rate": 0.0006,
+ "loss": 4.576877117156982,
+ "step": 4028
+ },
+ {
+ "epoch": 55.96419213973799,
+ "grad_norm": 0.023351816460490227,
+ "learning_rate": 0.0006,
+ "loss": 4.458885192871094,
+ "step": 4029
+ },
+ {
+ "epoch": 55.97816593886463,
+ "grad_norm": 0.019892243668437004,
+ "learning_rate": 0.0006,
+ "loss": 4.520229816436768,
+ "step": 4030
+ },
+ {
+ "epoch": 55.992139737991266,
+ "grad_norm": 0.018331089988350868,
+ "learning_rate": 0.0006,
+ "loss": 4.432201862335205,
+ "step": 4031
+ },
+ {
+ "epoch": 56.0,
+ "grad_norm": 0.021300828084349632,
+ "learning_rate": 0.0006,
+ "loss": 4.617072582244873,
+ "step": 4032
+ },
+ {
+ "epoch": 56.0,
+ "eval_loss": 4.811229705810547,
+ "eval_runtime": 56.4846,
+ "eval_samples_per_second": 43.233,
+ "eval_steps_per_second": 1.363,
+ "step": 4032
+ },
+ {
+ "epoch": 56.01397379912664,
+ "grad_norm": 0.02199401892721653,
+ "learning_rate": 0.0006,
+ "loss": 4.513965606689453,
+ "step": 4033
+ },
+ {
+ "epoch": 56.02794759825328,
+ "grad_norm": 0.022109858691692352,
+ "learning_rate": 0.0006,
+ "loss": 4.557568550109863,
+ "step": 4034
+ },
+ {
+ "epoch": 56.041921397379916,
+ "grad_norm": 0.020103694871068,
+ "learning_rate": 0.0006,
+ "loss": 4.569512844085693,
+ "step": 4035
+ },
+ {
+ "epoch": 56.05589519650655,
+ "grad_norm": 0.01898941583931446,
+ "learning_rate": 0.0006,
+ "loss": 4.623138904571533,
+ "step": 4036
+ },
+ {
+ "epoch": 56.069868995633186,
+ "grad_norm": 0.02227010577917099,
+ "learning_rate": 0.0006,
+ "loss": 4.433316230773926,
+ "step": 4037
+ },
+ {
+ "epoch": 56.083842794759825,
+ "grad_norm": 0.024683045223355293,
+ "learning_rate": 0.0006,
+ "loss": 4.458296775817871,
+ "step": 4038
+ },
+ {
+ "epoch": 56.09781659388646,
+ "grad_norm": 0.02900184690952301,
+ "learning_rate": 0.0006,
+ "loss": 4.531060218811035,
+ "step": 4039
+ },
+ {
+ "epoch": 56.1117903930131,
+ "grad_norm": 0.025557566434144974,
+ "learning_rate": 0.0006,
+ "loss": 4.56685733795166,
+ "step": 4040
+ },
+ {
+ "epoch": 56.12576419213974,
+ "grad_norm": 0.021414656192064285,
+ "learning_rate": 0.0006,
+ "loss": 4.469037055969238,
+ "step": 4041
+ },
+ {
+ "epoch": 56.13973799126637,
+ "grad_norm": 0.023262616246938705,
+ "learning_rate": 0.0006,
+ "loss": 4.493876934051514,
+ "step": 4042
+ },
+ {
+ "epoch": 56.15371179039301,
+ "grad_norm": 0.023011326789855957,
+ "learning_rate": 0.0006,
+ "loss": 4.4572906494140625,
+ "step": 4043
+ },
+ {
+ "epoch": 56.16768558951965,
+ "grad_norm": 0.02177615463733673,
+ "learning_rate": 0.0006,
+ "loss": 4.615389823913574,
+ "step": 4044
+ },
+ {
+ "epoch": 56.18165938864629,
+ "grad_norm": 0.021527713164687157,
+ "learning_rate": 0.0006,
+ "loss": 4.613762855529785,
+ "step": 4045
+ },
+ {
+ "epoch": 56.19563318777293,
+ "grad_norm": 0.01982325315475464,
+ "learning_rate": 0.0006,
+ "loss": 4.53670597076416,
+ "step": 4046
+ },
+ {
+ "epoch": 56.209606986899566,
+ "grad_norm": 0.022484654560685158,
+ "learning_rate": 0.0006,
+ "loss": 4.605565071105957,
+ "step": 4047
+ },
+ {
+ "epoch": 56.223580786026204,
+ "grad_norm": 0.02346140146255493,
+ "learning_rate": 0.0006,
+ "loss": 4.6265950202941895,
+ "step": 4048
+ },
+ {
+ "epoch": 56.237554585152836,
+ "grad_norm": 0.01876714453101158,
+ "learning_rate": 0.0006,
+ "loss": 4.481525421142578,
+ "step": 4049
+ },
+ {
+ "epoch": 56.251528384279474,
+ "grad_norm": 0.01926851086318493,
+ "learning_rate": 0.0006,
+ "loss": 4.630258083343506,
+ "step": 4050
+ },
+ {
+ "epoch": 56.26550218340611,
+ "grad_norm": 0.020858589559793472,
+ "learning_rate": 0.0006,
+ "loss": 4.5265913009643555,
+ "step": 4051
+ },
+ {
+ "epoch": 56.27947598253275,
+ "grad_norm": 0.01889280416071415,
+ "learning_rate": 0.0006,
+ "loss": 4.581275463104248,
+ "step": 4052
+ },
+ {
+ "epoch": 56.29344978165939,
+ "grad_norm": 0.01801002025604248,
+ "learning_rate": 0.0006,
+ "loss": 4.498557090759277,
+ "step": 4053
+ },
+ {
+ "epoch": 56.30742358078603,
+ "grad_norm": 0.02252180315554142,
+ "learning_rate": 0.0006,
+ "loss": 4.565267086029053,
+ "step": 4054
+ },
+ {
+ "epoch": 56.32139737991266,
+ "grad_norm": 0.02306070365011692,
+ "learning_rate": 0.0006,
+ "loss": 4.3591389656066895,
+ "step": 4055
+ },
+ {
+ "epoch": 56.3353711790393,
+ "grad_norm": 0.020292513072490692,
+ "learning_rate": 0.0006,
+ "loss": 4.407517433166504,
+ "step": 4056
+ },
+ {
+ "epoch": 56.34934497816594,
+ "grad_norm": 0.019154123961925507,
+ "learning_rate": 0.0006,
+ "loss": 4.482261657714844,
+ "step": 4057
+ },
+ {
+ "epoch": 56.36331877729258,
+ "grad_norm": 0.017338937148451805,
+ "learning_rate": 0.0006,
+ "loss": 4.4463582038879395,
+ "step": 4058
+ },
+ {
+ "epoch": 56.377292576419215,
+ "grad_norm": 0.01730559952557087,
+ "learning_rate": 0.0006,
+ "loss": 4.542474746704102,
+ "step": 4059
+ },
+ {
+ "epoch": 56.391266375545854,
+ "grad_norm": 0.017029426991939545,
+ "learning_rate": 0.0006,
+ "loss": 4.552488327026367,
+ "step": 4060
+ },
+ {
+ "epoch": 56.40524017467249,
+ "grad_norm": 0.01745498552918434,
+ "learning_rate": 0.0006,
+ "loss": 4.52928352355957,
+ "step": 4061
+ },
+ {
+ "epoch": 56.419213973799124,
+ "grad_norm": 0.0168515145778656,
+ "learning_rate": 0.0006,
+ "loss": 4.4516801834106445,
+ "step": 4062
+ },
+ {
+ "epoch": 56.43318777292576,
+ "grad_norm": 0.018545417115092278,
+ "learning_rate": 0.0006,
+ "loss": 4.569811820983887,
+ "step": 4063
+ },
+ {
+ "epoch": 56.4471615720524,
+ "grad_norm": 0.02227838523685932,
+ "learning_rate": 0.0006,
+ "loss": 4.412224769592285,
+ "step": 4064
+ },
+ {
+ "epoch": 56.46113537117904,
+ "grad_norm": 0.023338504135608673,
+ "learning_rate": 0.0006,
+ "loss": 4.490184307098389,
+ "step": 4065
+ },
+ {
+ "epoch": 56.47510917030568,
+ "grad_norm": 0.02257535606622696,
+ "learning_rate": 0.0006,
+ "loss": 4.479152679443359,
+ "step": 4066
+ },
+ {
+ "epoch": 56.48908296943232,
+ "grad_norm": 0.022737598046660423,
+ "learning_rate": 0.0006,
+ "loss": 4.418235778808594,
+ "step": 4067
+ },
+ {
+ "epoch": 56.50305676855895,
+ "grad_norm": 0.024603573605418205,
+ "learning_rate": 0.0006,
+ "loss": 4.45781135559082,
+ "step": 4068
+ },
+ {
+ "epoch": 56.51703056768559,
+ "grad_norm": 0.024953778833150864,
+ "learning_rate": 0.0006,
+ "loss": 4.593907833099365,
+ "step": 4069
+ },
+ {
+ "epoch": 56.531004366812226,
+ "grad_norm": 0.02195645309984684,
+ "learning_rate": 0.0006,
+ "loss": 4.467643737792969,
+ "step": 4070
+ },
+ {
+ "epoch": 56.544978165938865,
+ "grad_norm": 0.017126578837633133,
+ "learning_rate": 0.0006,
+ "loss": 4.555484771728516,
+ "step": 4071
+ },
+ {
+ "epoch": 56.5589519650655,
+ "grad_norm": 0.019482605159282684,
+ "learning_rate": 0.0006,
+ "loss": 4.430915832519531,
+ "step": 4072
+ },
+ {
+ "epoch": 56.57292576419214,
+ "grad_norm": 0.0219392292201519,
+ "learning_rate": 0.0006,
+ "loss": 4.593365669250488,
+ "step": 4073
+ },
+ {
+ "epoch": 56.58689956331878,
+ "grad_norm": 0.02420145645737648,
+ "learning_rate": 0.0006,
+ "loss": 4.491467475891113,
+ "step": 4074
+ },
+ {
+ "epoch": 56.60087336244541,
+ "grad_norm": 0.021523723378777504,
+ "learning_rate": 0.0006,
+ "loss": 4.463634490966797,
+ "step": 4075
+ },
+ {
+ "epoch": 56.61484716157205,
+ "grad_norm": 0.020242048427462578,
+ "learning_rate": 0.0006,
+ "loss": 4.577968597412109,
+ "step": 4076
+ },
+ {
+ "epoch": 56.62882096069869,
+ "grad_norm": 0.017771173268556595,
+ "learning_rate": 0.0006,
+ "loss": 4.568269729614258,
+ "step": 4077
+ },
+ {
+ "epoch": 56.64279475982533,
+ "grad_norm": 0.01598544418811798,
+ "learning_rate": 0.0006,
+ "loss": 4.476841926574707,
+ "step": 4078
+ },
+ {
+ "epoch": 56.65676855895197,
+ "grad_norm": 0.0168446097522974,
+ "learning_rate": 0.0006,
+ "loss": 4.480965614318848,
+ "step": 4079
+ },
+ {
+ "epoch": 56.670742358078606,
+ "grad_norm": 0.014948660507798195,
+ "learning_rate": 0.0006,
+ "loss": 4.535400390625,
+ "step": 4080
+ },
+ {
+ "epoch": 56.68471615720524,
+ "grad_norm": 0.014899153262376785,
+ "learning_rate": 0.0006,
+ "loss": 4.476048946380615,
+ "step": 4081
+ },
+ {
+ "epoch": 56.698689956331876,
+ "grad_norm": 0.014051870442926884,
+ "learning_rate": 0.0006,
+ "loss": 4.533884048461914,
+ "step": 4082
+ },
+ {
+ "epoch": 56.712663755458514,
+ "grad_norm": 0.014708572067320347,
+ "learning_rate": 0.0006,
+ "loss": 4.586690902709961,
+ "step": 4083
+ },
+ {
+ "epoch": 56.72663755458515,
+ "grad_norm": 0.014797299169003963,
+ "learning_rate": 0.0006,
+ "loss": 4.446291446685791,
+ "step": 4084
+ },
+ {
+ "epoch": 56.74061135371179,
+ "grad_norm": 0.015256541781127453,
+ "learning_rate": 0.0006,
+ "loss": 4.553572654724121,
+ "step": 4085
+ },
+ {
+ "epoch": 56.75458515283843,
+ "grad_norm": 0.016042491421103477,
+ "learning_rate": 0.0006,
+ "loss": 4.507801055908203,
+ "step": 4086
+ },
+ {
+ "epoch": 56.76855895196506,
+ "grad_norm": 0.018835173919796944,
+ "learning_rate": 0.0006,
+ "loss": 4.393911838531494,
+ "step": 4087
+ },
+ {
+ "epoch": 56.7825327510917,
+ "grad_norm": 0.021994033828377724,
+ "learning_rate": 0.0006,
+ "loss": 4.543368339538574,
+ "step": 4088
+ },
+ {
+ "epoch": 56.79650655021834,
+ "grad_norm": 0.0230990182608366,
+ "learning_rate": 0.0006,
+ "loss": 4.507420539855957,
+ "step": 4089
+ },
+ {
+ "epoch": 56.81048034934498,
+ "grad_norm": 0.018818842247128487,
+ "learning_rate": 0.0006,
+ "loss": 4.499941349029541,
+ "step": 4090
+ },
+ {
+ "epoch": 56.82445414847162,
+ "grad_norm": 0.018257278949022293,
+ "learning_rate": 0.0006,
+ "loss": 4.380057334899902,
+ "step": 4091
+ },
+ {
+ "epoch": 56.838427947598255,
+ "grad_norm": 0.020073018968105316,
+ "learning_rate": 0.0006,
+ "loss": 4.572811603546143,
+ "step": 4092
+ },
+ {
+ "epoch": 56.852401746724894,
+ "grad_norm": 0.02078658528625965,
+ "learning_rate": 0.0006,
+ "loss": 4.446835517883301,
+ "step": 4093
+ },
+ {
+ "epoch": 56.866375545851525,
+ "grad_norm": 0.01793118566274643,
+ "learning_rate": 0.0006,
+ "loss": 4.517854690551758,
+ "step": 4094
+ },
+ {
+ "epoch": 56.880349344978164,
+ "grad_norm": 0.01608281210064888,
+ "learning_rate": 0.0006,
+ "loss": 4.506482124328613,
+ "step": 4095
+ },
+ {
+ "epoch": 56.8943231441048,
+ "grad_norm": 0.017405368387699127,
+ "learning_rate": 0.0006,
+ "loss": 4.439718246459961,
+ "step": 4096
+ },
+ {
+ "epoch": 56.90829694323144,
+ "grad_norm": 0.017930863425135612,
+ "learning_rate": 0.0006,
+ "loss": 4.409365653991699,
+ "step": 4097
+ },
+ {
+ "epoch": 56.92227074235808,
+ "grad_norm": 0.015757670626044273,
+ "learning_rate": 0.0006,
+ "loss": 4.505547523498535,
+ "step": 4098
+ },
+ {
+ "epoch": 56.93624454148472,
+ "grad_norm": 0.01580345816910267,
+ "learning_rate": 0.0006,
+ "loss": 4.462774753570557,
+ "step": 4099
+ },
+ {
+ "epoch": 56.95021834061135,
+ "grad_norm": 0.01659180410206318,
+ "learning_rate": 0.0006,
+ "loss": 4.49345588684082,
+ "step": 4100
+ },
+ {
+ "epoch": 56.96419213973799,
+ "grad_norm": 0.01877259835600853,
+ "learning_rate": 0.0006,
+ "loss": 4.574443340301514,
+ "step": 4101
+ },
+ {
+ "epoch": 56.97816593886463,
+ "grad_norm": 0.01977839693427086,
+ "learning_rate": 0.0006,
+ "loss": 4.574034690856934,
+ "step": 4102
+ },
+ {
+ "epoch": 56.992139737991266,
+ "grad_norm": 0.020627597346901894,
+ "learning_rate": 0.0006,
+ "loss": 4.50969123840332,
+ "step": 4103
+ },
+ {
+ "epoch": 57.0,
+ "grad_norm": 0.025128807872533798,
+ "learning_rate": 0.0006,
+ "loss": 4.401662826538086,
+ "step": 4104
+ },
+ {
+ "epoch": 57.0,
+ "eval_loss": 4.779252052307129,
+ "eval_runtime": 56.8104,
+ "eval_samples_per_second": 42.985,
+ "eval_steps_per_second": 1.355,
+ "step": 4104
+ },
+ {
+ "epoch": 57.01397379912664,
+ "grad_norm": 0.027712570503354073,
+ "learning_rate": 0.0006,
+ "loss": 4.472955703735352,
+ "step": 4105
+ },
+ {
+ "epoch": 57.02794759825328,
+ "grad_norm": 0.02164197526872158,
+ "learning_rate": 0.0006,
+ "loss": 4.387596130371094,
+ "step": 4106
+ },
+ {
+ "epoch": 57.041921397379916,
+ "grad_norm": 0.017972752451896667,
+ "learning_rate": 0.0006,
+ "loss": 4.4326276779174805,
+ "step": 4107
+ },
+ {
+ "epoch": 57.05589519650655,
+ "grad_norm": 0.020811019465327263,
+ "learning_rate": 0.0006,
+ "loss": 4.452951431274414,
+ "step": 4108
+ },
+ {
+ "epoch": 57.069868995633186,
+ "grad_norm": 0.019913259893655777,
+ "learning_rate": 0.0006,
+ "loss": 4.448301315307617,
+ "step": 4109
+ },
+ {
+ "epoch": 57.083842794759825,
+ "grad_norm": 0.017409102991223335,
+ "learning_rate": 0.0006,
+ "loss": 4.4108357429504395,
+ "step": 4110
+ },
+ {
+ "epoch": 57.09781659388646,
+ "grad_norm": 0.019374269992113113,
+ "learning_rate": 0.0006,
+ "loss": 4.398412227630615,
+ "step": 4111
+ },
+ {
+ "epoch": 57.1117903930131,
+ "grad_norm": 0.015521776862442493,
+ "learning_rate": 0.0006,
+ "loss": 4.508594989776611,
+ "step": 4112
+ },
+ {
+ "epoch": 57.12576419213974,
+ "grad_norm": 0.017201920971274376,
+ "learning_rate": 0.0006,
+ "loss": 4.449558734893799,
+ "step": 4113
+ },
+ {
+ "epoch": 57.13973799126637,
+ "grad_norm": 0.018062541261315346,
+ "learning_rate": 0.0006,
+ "loss": 4.517874240875244,
+ "step": 4114
+ },
+ {
+ "epoch": 57.15371179039301,
+ "grad_norm": 0.015846073627471924,
+ "learning_rate": 0.0006,
+ "loss": 4.427501678466797,
+ "step": 4115
+ },
+ {
+ "epoch": 57.16768558951965,
+ "grad_norm": 0.01849362626671791,
+ "learning_rate": 0.0006,
+ "loss": 4.433181285858154,
+ "step": 4116
+ },
+ {
+ "epoch": 57.18165938864629,
+ "grad_norm": 0.01948804222047329,
+ "learning_rate": 0.0006,
+ "loss": 4.446122646331787,
+ "step": 4117
+ },
+ {
+ "epoch": 57.19563318777293,
+ "grad_norm": 0.017354389652609825,
+ "learning_rate": 0.0006,
+ "loss": 4.431153774261475,
+ "step": 4118
+ },
+ {
+ "epoch": 57.209606986899566,
+ "grad_norm": 0.015817373991012573,
+ "learning_rate": 0.0006,
+ "loss": 4.448566913604736,
+ "step": 4119
+ },
+ {
+ "epoch": 57.223580786026204,
+ "grad_norm": 0.01635785959661007,
+ "learning_rate": 0.0006,
+ "loss": 4.408031463623047,
+ "step": 4120
+ },
+ {
+ "epoch": 57.237554585152836,
+ "grad_norm": 0.014528338797390461,
+ "learning_rate": 0.0006,
+ "loss": 4.541668891906738,
+ "step": 4121
+ },
+ {
+ "epoch": 57.251528384279474,
+ "grad_norm": 0.01589174196124077,
+ "learning_rate": 0.0006,
+ "loss": 4.513199329376221,
+ "step": 4122
+ },
+ {
+ "epoch": 57.26550218340611,
+ "grad_norm": 0.016093695536255836,
+ "learning_rate": 0.0006,
+ "loss": 4.516178131103516,
+ "step": 4123
+ },
+ {
+ "epoch": 57.27947598253275,
+ "grad_norm": 0.01788361556828022,
+ "learning_rate": 0.0006,
+ "loss": 4.524880409240723,
+ "step": 4124
+ },
+ {
+ "epoch": 57.29344978165939,
+ "grad_norm": 0.01591576263308525,
+ "learning_rate": 0.0006,
+ "loss": 4.44764518737793,
+ "step": 4125
+ },
+ {
+ "epoch": 57.30742358078603,
+ "grad_norm": 0.015452570281922817,
+ "learning_rate": 0.0006,
+ "loss": 4.46844482421875,
+ "step": 4126
+ },
+ {
+ "epoch": 57.32139737991266,
+ "grad_norm": 0.016404492780566216,
+ "learning_rate": 0.0006,
+ "loss": 4.443678855895996,
+ "step": 4127
+ },
+ {
+ "epoch": 57.3353711790393,
+ "grad_norm": 0.019967148080468178,
+ "learning_rate": 0.0006,
+ "loss": 4.430230140686035,
+ "step": 4128
+ },
+ {
+ "epoch": 57.34934497816594,
+ "grad_norm": 0.02594015561044216,
+ "learning_rate": 0.0006,
+ "loss": 4.47061824798584,
+ "step": 4129
+ },
+ {
+ "epoch": 57.36331877729258,
+ "grad_norm": 0.031096890568733215,
+ "learning_rate": 0.0006,
+ "loss": 4.462275981903076,
+ "step": 4130
+ },
+ {
+ "epoch": 57.377292576419215,
+ "grad_norm": 0.02834898792207241,
+ "learning_rate": 0.0006,
+ "loss": 4.41143798828125,
+ "step": 4131
+ },
+ {
+ "epoch": 57.391266375545854,
+ "grad_norm": 0.023621153086423874,
+ "learning_rate": 0.0006,
+ "loss": 4.562548637390137,
+ "step": 4132
+ },
+ {
+ "epoch": 57.40524017467249,
+ "grad_norm": 0.02277296595275402,
+ "learning_rate": 0.0006,
+ "loss": 4.456247329711914,
+ "step": 4133
+ },
+ {
+ "epoch": 57.419213973799124,
+ "grad_norm": 0.019868768751621246,
+ "learning_rate": 0.0006,
+ "loss": 4.524300575256348,
+ "step": 4134
+ },
+ {
+ "epoch": 57.43318777292576,
+ "grad_norm": 0.01887078955769539,
+ "learning_rate": 0.0006,
+ "loss": 4.540097236633301,
+ "step": 4135
+ },
+ {
+ "epoch": 57.4471615720524,
+ "grad_norm": 0.024268802255392075,
+ "learning_rate": 0.0006,
+ "loss": 4.397233963012695,
+ "step": 4136
+ },
+ {
+ "epoch": 57.46113537117904,
+ "grad_norm": 0.025739721953868866,
+ "learning_rate": 0.0006,
+ "loss": 4.354194641113281,
+ "step": 4137
+ },
+ {
+ "epoch": 57.47510917030568,
+ "grad_norm": 0.02442290261387825,
+ "learning_rate": 0.0006,
+ "loss": 4.303822994232178,
+ "step": 4138
+ },
+ {
+ "epoch": 57.48908296943232,
+ "grad_norm": 0.022215209901332855,
+ "learning_rate": 0.0006,
+ "loss": 4.4870452880859375,
+ "step": 4139
+ },
+ {
+ "epoch": 57.50305676855895,
+ "grad_norm": 0.01789158768951893,
+ "learning_rate": 0.0006,
+ "loss": 4.343809127807617,
+ "step": 4140
+ },
+ {
+ "epoch": 57.51703056768559,
+ "grad_norm": 0.020381739363074303,
+ "learning_rate": 0.0006,
+ "loss": 4.53312873840332,
+ "step": 4141
+ },
+ {
+ "epoch": 57.531004366812226,
+ "grad_norm": 0.019585080444812775,
+ "learning_rate": 0.0006,
+ "loss": 4.510162353515625,
+ "step": 4142
+ },
+ {
+ "epoch": 57.544978165938865,
+ "grad_norm": 0.01985127478837967,
+ "learning_rate": 0.0006,
+ "loss": 4.3654022216796875,
+ "step": 4143
+ },
+ {
+ "epoch": 57.5589519650655,
+ "grad_norm": 0.01962905190885067,
+ "learning_rate": 0.0006,
+ "loss": 4.560830116271973,
+ "step": 4144
+ },
+ {
+ "epoch": 57.57292576419214,
+ "grad_norm": 0.02679629437625408,
+ "learning_rate": 0.0006,
+ "loss": 4.300804138183594,
+ "step": 4145
+ },
+ {
+ "epoch": 57.58689956331878,
+ "grad_norm": 0.039830856025218964,
+ "learning_rate": 0.0006,
+ "loss": 4.54150915145874,
+ "step": 4146
+ },
+ {
+ "epoch": 57.60087336244541,
+ "grad_norm": 0.049527477473020554,
+ "learning_rate": 0.0006,
+ "loss": 4.405792236328125,
+ "step": 4147
+ },
+ {
+ "epoch": 57.61484716157205,
+ "grad_norm": 0.042966946959495544,
+ "learning_rate": 0.0006,
+ "loss": 4.500643730163574,
+ "step": 4148
+ },
+ {
+ "epoch": 57.62882096069869,
+ "grad_norm": 0.08999177813529968,
+ "learning_rate": 0.0006,
+ "loss": 4.504790782928467,
+ "step": 4149
+ },
+ {
+ "epoch": 57.64279475982533,
+ "grad_norm": 0.44693124294281006,
+ "learning_rate": 0.0006,
+ "loss": 4.845344066619873,
+ "step": 4150
+ },
+ {
+ "epoch": 57.65676855895197,
+ "grad_norm": 0.7143144011497498,
+ "learning_rate": 0.0006,
+ "loss": 6.907845497131348,
+ "step": 4151
+ },
+ {
+ "epoch": 57.670742358078606,
+ "grad_norm": 0.25239595770835876,
+ "learning_rate": 0.0006,
+ "loss": 7.082579612731934,
+ "step": 4152
+ },
+ {
+ "epoch": 57.68471615720524,
+ "grad_norm": 0.26550790667533875,
+ "learning_rate": 0.0006,
+ "loss": 7.50180196762085,
+ "step": 4153
+ },
+ {
+ "epoch": 57.698689956331876,
+ "grad_norm": 0.14735378324985504,
+ "learning_rate": 0.0006,
+ "loss": 7.214590072631836,
+ "step": 4154
+ },
+ {
+ "epoch": 57.712663755458514,
+ "grad_norm": 0.17685334384441376,
+ "learning_rate": 0.0006,
+ "loss": 6.861320495605469,
+ "step": 4155
+ },
+ {
+ "epoch": 57.72663755458515,
+ "grad_norm": 0.18231339752674103,
+ "learning_rate": 0.0006,
+ "loss": 6.791839599609375,
+ "step": 4156
+ },
+ {
+ "epoch": 57.74061135371179,
+ "grad_norm": 0.09230359643697739,
+ "learning_rate": 0.0006,
+ "loss": 6.760223388671875,
+ "step": 4157
+ },
+ {
+ "epoch": 57.75458515283843,
+ "grad_norm": 0.06918249279260635,
+ "learning_rate": 0.0006,
+ "loss": 6.532999515533447,
+ "step": 4158
+ },
+ {
+ "epoch": 57.76855895196506,
+ "grad_norm": 0.0901007279753685,
+ "learning_rate": 0.0006,
+ "loss": 6.52907133102417,
+ "step": 4159
+ },
+ {
+ "epoch": 57.7825327510917,
+ "grad_norm": 0.05678229779005051,
+ "learning_rate": 0.0006,
+ "loss": 6.295290470123291,
+ "step": 4160
+ },
+ {
+ "epoch": 57.79650655021834,
+ "grad_norm": 0.04450210556387901,
+ "learning_rate": 0.0006,
+ "loss": 6.14900541305542,
+ "step": 4161
+ },
+ {
+ "epoch": 57.81048034934498,
+ "grad_norm": 0.049191512167453766,
+ "learning_rate": 0.0006,
+ "loss": 6.19650411605835,
+ "step": 4162
+ },
+ {
+ "epoch": 57.82445414847162,
+ "grad_norm": 0.03908967226743698,
+ "learning_rate": 0.0006,
+ "loss": 6.054948806762695,
+ "step": 4163
+ },
+ {
+ "epoch": 57.838427947598255,
+ "grad_norm": 0.04023387283086777,
+ "learning_rate": 0.0006,
+ "loss": 6.034873962402344,
+ "step": 4164
+ },
+ {
+ "epoch": 57.852401746724894,
+ "grad_norm": 0.03304268419742584,
+ "learning_rate": 0.0006,
+ "loss": 5.769852161407471,
+ "step": 4165
+ },
+ {
+ "epoch": 57.866375545851525,
+ "grad_norm": 0.03432450816035271,
+ "learning_rate": 0.0006,
+ "loss": 5.788785457611084,
+ "step": 4166
+ },
+ {
+ "epoch": 57.880349344978164,
+ "grad_norm": 0.029965840280056,
+ "learning_rate": 0.0006,
+ "loss": 5.815243721008301,
+ "step": 4167
+ },
+ {
+ "epoch": 57.8943231441048,
+ "grad_norm": 0.03637698292732239,
+ "learning_rate": 0.0006,
+ "loss": 5.7204060554504395,
+ "step": 4168
+ },
+ {
+ "epoch": 57.90829694323144,
+ "grad_norm": 0.03477517142891884,
+ "learning_rate": 0.0006,
+ "loss": 5.605930328369141,
+ "step": 4169
+ },
+ {
+ "epoch": 57.92227074235808,
+ "grad_norm": 0.05061859264969826,
+ "learning_rate": 0.0006,
+ "loss": 5.547054767608643,
+ "step": 4170
+ },
+ {
+ "epoch": 57.93624454148472,
+ "grad_norm": 0.06299655884504318,
+ "learning_rate": 0.0006,
+ "loss": 5.552166938781738,
+ "step": 4171
+ },
+ {
+ "epoch": 57.95021834061135,
+ "grad_norm": 0.04215948283672333,
+ "learning_rate": 0.0006,
+ "loss": 5.375457763671875,
+ "step": 4172
+ },
+ {
+ "epoch": 57.96419213973799,
+ "grad_norm": 0.0348566472530365,
+ "learning_rate": 0.0006,
+ "loss": 5.418606281280518,
+ "step": 4173
+ },
+ {
+ "epoch": 57.97816593886463,
+ "grad_norm": 0.03807530924677849,
+ "learning_rate": 0.0006,
+ "loss": 5.403156280517578,
+ "step": 4174
+ },
+ {
+ "epoch": 57.992139737991266,
+ "grad_norm": 0.02989993803203106,
+ "learning_rate": 0.0006,
+ "loss": 5.355951309204102,
+ "step": 4175
+ },
+ {
+ "epoch": 58.0,
+ "grad_norm": 0.034300558269023895,
+ "learning_rate": 0.0006,
+ "loss": 5.44920539855957,
+ "step": 4176
+ },
+ {
+ "epoch": 58.0,
+ "eval_loss": 5.416731834411621,
+ "eval_runtime": 56.7934,
+ "eval_samples_per_second": 42.998,
+ "eval_steps_per_second": 1.356,
+ "step": 4176
+ },
+ {
+ "epoch": 58.01397379912664,
+ "grad_norm": 0.033961448818445206,
+ "learning_rate": 0.0006,
+ "loss": 5.250707626342773,
+ "step": 4177
+ },
+ {
+ "epoch": 58.02794759825328,
+ "grad_norm": 0.027343502268195152,
+ "learning_rate": 0.0006,
+ "loss": 5.149238109588623,
+ "step": 4178
+ },
+ {
+ "epoch": 58.041921397379916,
+ "grad_norm": 0.03033650480210781,
+ "learning_rate": 0.0006,
+ "loss": 5.072463512420654,
+ "step": 4179
+ },
+ {
+ "epoch": 58.05589519650655,
+ "grad_norm": 0.028262868523597717,
+ "learning_rate": 0.0006,
+ "loss": 5.109417915344238,
+ "step": 4180
+ },
+ {
+ "epoch": 58.069868995633186,
+ "grad_norm": 0.02373034693300724,
+ "learning_rate": 0.0006,
+ "loss": 5.082464218139648,
+ "step": 4181
+ },
+ {
+ "epoch": 58.083842794759825,
+ "grad_norm": 0.025879688560962677,
+ "learning_rate": 0.0006,
+ "loss": 5.056792259216309,
+ "step": 4182
+ },
+ {
+ "epoch": 58.09781659388646,
+ "grad_norm": 0.026252275332808495,
+ "learning_rate": 0.0006,
+ "loss": 5.064239978790283,
+ "step": 4183
+ },
+ {
+ "epoch": 58.1117903930131,
+ "grad_norm": 0.028177792206406593,
+ "learning_rate": 0.0006,
+ "loss": 4.922310829162598,
+ "step": 4184
+ },
+ {
+ "epoch": 58.12576419213974,
+ "grad_norm": 0.02630820870399475,
+ "learning_rate": 0.0006,
+ "loss": 4.992884635925293,
+ "step": 4185
+ },
+ {
+ "epoch": 58.13973799126637,
+ "grad_norm": 0.024890903383493423,
+ "learning_rate": 0.0006,
+ "loss": 4.910036563873291,
+ "step": 4186
+ },
+ {
+ "epoch": 58.15371179039301,
+ "grad_norm": 0.02822992391884327,
+ "learning_rate": 0.0006,
+ "loss": 4.866640090942383,
+ "step": 4187
+ },
+ {
+ "epoch": 58.16768558951965,
+ "grad_norm": 0.03253607079386711,
+ "learning_rate": 0.0006,
+ "loss": 4.9112958908081055,
+ "step": 4188
+ },
+ {
+ "epoch": 58.18165938864629,
+ "grad_norm": 0.03765944764018059,
+ "learning_rate": 0.0006,
+ "loss": 4.774468421936035,
+ "step": 4189
+ },
+ {
+ "epoch": 58.19563318777293,
+ "grad_norm": 0.039874881505966187,
+ "learning_rate": 0.0006,
+ "loss": 4.831678867340088,
+ "step": 4190
+ },
+ {
+ "epoch": 58.209606986899566,
+ "grad_norm": 0.034595925360918045,
+ "learning_rate": 0.0006,
+ "loss": 4.8243794441223145,
+ "step": 4191
+ },
+ {
+ "epoch": 58.223580786026204,
+ "grad_norm": 0.04840189591050148,
+ "learning_rate": 0.0006,
+ "loss": 4.854156494140625,
+ "step": 4192
+ },
+ {
+ "epoch": 58.237554585152836,
+ "grad_norm": 0.06358753889799118,
+ "learning_rate": 0.0006,
+ "loss": 4.836430549621582,
+ "step": 4193
+ },
+ {
+ "epoch": 58.251528384279474,
+ "grad_norm": 0.0507555715739727,
+ "learning_rate": 0.0006,
+ "loss": 4.88357400894165,
+ "step": 4194
+ },
+ {
+ "epoch": 58.26550218340611,
+ "grad_norm": 0.04384690150618553,
+ "learning_rate": 0.0006,
+ "loss": 4.774049282073975,
+ "step": 4195
+ },
+ {
+ "epoch": 58.27947598253275,
+ "grad_norm": 0.03246928006410599,
+ "learning_rate": 0.0006,
+ "loss": 4.686605930328369,
+ "step": 4196
+ },
+ {
+ "epoch": 58.29344978165939,
+ "grad_norm": 0.028878789395093918,
+ "learning_rate": 0.0006,
+ "loss": 4.796070575714111,
+ "step": 4197
+ },
+ {
+ "epoch": 58.30742358078603,
+ "grad_norm": 0.028175361454486847,
+ "learning_rate": 0.0006,
+ "loss": 4.728490829467773,
+ "step": 4198
+ },
+ {
+ "epoch": 58.32139737991266,
+ "grad_norm": 0.029428014531731606,
+ "learning_rate": 0.0006,
+ "loss": 4.7159624099731445,
+ "step": 4199
+ },
+ {
+ "epoch": 58.3353711790393,
+ "grad_norm": 0.031227827072143555,
+ "learning_rate": 0.0006,
+ "loss": 4.702625274658203,
+ "step": 4200
+ },
+ {
+ "epoch": 58.34934497816594,
+ "grad_norm": 0.028843428939580917,
+ "learning_rate": 0.0006,
+ "loss": 4.811799049377441,
+ "step": 4201
+ },
+ {
+ "epoch": 58.36331877729258,
+ "grad_norm": 0.026780391111969948,
+ "learning_rate": 0.0006,
+ "loss": 4.81662654876709,
+ "step": 4202
+ },
+ {
+ "epoch": 58.377292576419215,
+ "grad_norm": 0.019699757918715477,
+ "learning_rate": 0.0006,
+ "loss": 4.713685989379883,
+ "step": 4203
+ },
+ {
+ "epoch": 58.391266375545854,
+ "grad_norm": 0.024389250203967094,
+ "learning_rate": 0.0006,
+ "loss": 4.72354793548584,
+ "step": 4204
+ },
+ {
+ "epoch": 58.40524017467249,
+ "grad_norm": 0.02218274027109146,
+ "learning_rate": 0.0006,
+ "loss": 4.674125671386719,
+ "step": 4205
+ },
+ {
+ "epoch": 58.419213973799124,
+ "grad_norm": 0.02041488140821457,
+ "learning_rate": 0.0006,
+ "loss": 4.7909698486328125,
+ "step": 4206
+ },
+ {
+ "epoch": 58.43318777292576,
+ "grad_norm": 0.018561935052275658,
+ "learning_rate": 0.0006,
+ "loss": 4.808770179748535,
+ "step": 4207
+ },
+ {
+ "epoch": 58.4471615720524,
+ "grad_norm": 0.01708846725523472,
+ "learning_rate": 0.0006,
+ "loss": 4.701626777648926,
+ "step": 4208
+ },
+ {
+ "epoch": 58.46113537117904,
+ "grad_norm": 0.0222734697163105,
+ "learning_rate": 0.0006,
+ "loss": 4.505410194396973,
+ "step": 4209
+ },
+ {
+ "epoch": 58.47510917030568,
+ "grad_norm": 0.03197057917714119,
+ "learning_rate": 0.0006,
+ "loss": 4.610360145568848,
+ "step": 4210
+ },
+ {
+ "epoch": 58.48908296943232,
+ "grad_norm": 0.06854520738124847,
+ "learning_rate": 0.0006,
+ "loss": 4.625649452209473,
+ "step": 4211
+ },
+ {
+ "epoch": 58.50305676855895,
+ "grad_norm": 0.09981521219015121,
+ "learning_rate": 0.0006,
+ "loss": 4.832150459289551,
+ "step": 4212
+ },
+ {
+ "epoch": 58.51703056768559,
+ "grad_norm": 0.046356040984392166,
+ "learning_rate": 0.0006,
+ "loss": 4.5641632080078125,
+ "step": 4213
+ },
+ {
+ "epoch": 58.531004366812226,
+ "grad_norm": 0.03882667422294617,
+ "learning_rate": 0.0006,
+ "loss": 4.666974067687988,
+ "step": 4214
+ },
+ {
+ "epoch": 58.544978165938865,
+ "grad_norm": 0.033203233033418655,
+ "learning_rate": 0.0006,
+ "loss": 4.629413604736328,
+ "step": 4215
+ },
+ {
+ "epoch": 58.5589519650655,
+ "grad_norm": 0.02806735597550869,
+ "learning_rate": 0.0006,
+ "loss": 4.763245582580566,
+ "step": 4216
+ },
+ {
+ "epoch": 58.57292576419214,
+ "grad_norm": 0.026728278025984764,
+ "learning_rate": 0.0006,
+ "loss": 4.629597187042236,
+ "step": 4217
+ },
+ {
+ "epoch": 58.58689956331878,
+ "grad_norm": 0.024380959570407867,
+ "learning_rate": 0.0006,
+ "loss": 4.633614540100098,
+ "step": 4218
+ },
+ {
+ "epoch": 58.60087336244541,
+ "grad_norm": 0.022312544286251068,
+ "learning_rate": 0.0006,
+ "loss": 4.671379566192627,
+ "step": 4219
+ },
+ {
+ "epoch": 58.61484716157205,
+ "grad_norm": 0.022049568593502045,
+ "learning_rate": 0.0006,
+ "loss": 4.661557197570801,
+ "step": 4220
+ },
+ {
+ "epoch": 58.62882096069869,
+ "grad_norm": 0.02119818702340126,
+ "learning_rate": 0.0006,
+ "loss": 4.688562393188477,
+ "step": 4221
+ },
+ {
+ "epoch": 58.64279475982533,
+ "grad_norm": 0.019769301638007164,
+ "learning_rate": 0.0006,
+ "loss": 4.607748031616211,
+ "step": 4222
+ },
+ {
+ "epoch": 58.65676855895197,
+ "grad_norm": 0.02124079130589962,
+ "learning_rate": 0.0006,
+ "loss": 4.570640563964844,
+ "step": 4223
+ },
+ {
+ "epoch": 58.670742358078606,
+ "grad_norm": 0.016344040632247925,
+ "learning_rate": 0.0006,
+ "loss": 4.676113128662109,
+ "step": 4224
+ },
+ {
+ "epoch": 58.68471615720524,
+ "grad_norm": 0.01735256239771843,
+ "learning_rate": 0.0006,
+ "loss": 4.62088680267334,
+ "step": 4225
+ },
+ {
+ "epoch": 58.698689956331876,
+ "grad_norm": 0.019323458895087242,
+ "learning_rate": 0.0006,
+ "loss": 4.579405307769775,
+ "step": 4226
+ },
+ {
+ "epoch": 58.712663755458514,
+ "grad_norm": 0.01760544814169407,
+ "learning_rate": 0.0006,
+ "loss": 4.625407695770264,
+ "step": 4227
+ },
+ {
+ "epoch": 58.72663755458515,
+ "grad_norm": 0.015593250282108784,
+ "learning_rate": 0.0006,
+ "loss": 4.6326117515563965,
+ "step": 4228
+ },
+ {
+ "epoch": 58.74061135371179,
+ "grad_norm": 0.014357523061335087,
+ "learning_rate": 0.0006,
+ "loss": 4.6869401931762695,
+ "step": 4229
+ },
+ {
+ "epoch": 58.75458515283843,
+ "grad_norm": 0.01604357920587063,
+ "learning_rate": 0.0006,
+ "loss": 4.690784931182861,
+ "step": 4230
+ },
+ {
+ "epoch": 58.76855895196506,
+ "grad_norm": 0.01673845760524273,
+ "learning_rate": 0.0006,
+ "loss": 4.513410568237305,
+ "step": 4231
+ },
+ {
+ "epoch": 58.7825327510917,
+ "grad_norm": 0.013657779432833195,
+ "learning_rate": 0.0006,
+ "loss": 4.637057781219482,
+ "step": 4232
+ },
+ {
+ "epoch": 58.79650655021834,
+ "grad_norm": 0.013373114168643951,
+ "learning_rate": 0.0006,
+ "loss": 4.719967842102051,
+ "step": 4233
+ },
+ {
+ "epoch": 58.81048034934498,
+ "grad_norm": 0.016345568001270294,
+ "learning_rate": 0.0006,
+ "loss": 4.604109764099121,
+ "step": 4234
+ },
+ {
+ "epoch": 58.82445414847162,
+ "grad_norm": 0.01573033444583416,
+ "learning_rate": 0.0006,
+ "loss": 4.6471428871154785,
+ "step": 4235
+ },
+ {
+ "epoch": 58.838427947598255,
+ "grad_norm": 0.01517449039965868,
+ "learning_rate": 0.0006,
+ "loss": 4.722684860229492,
+ "step": 4236
+ },
+ {
+ "epoch": 58.852401746724894,
+ "grad_norm": 0.0145102022215724,
+ "learning_rate": 0.0006,
+ "loss": 4.662420272827148,
+ "step": 4237
+ },
+ {
+ "epoch": 58.866375545851525,
+ "grad_norm": 0.01457217987626791,
+ "learning_rate": 0.0006,
+ "loss": 4.585169792175293,
+ "step": 4238
+ },
+ {
+ "epoch": 58.880349344978164,
+ "grad_norm": 0.01406773366034031,
+ "learning_rate": 0.0006,
+ "loss": 4.5413665771484375,
+ "step": 4239
+ },
+ {
+ "epoch": 58.8943231441048,
+ "grad_norm": 0.012653871439397335,
+ "learning_rate": 0.0006,
+ "loss": 4.553636074066162,
+ "step": 4240
+ },
+ {
+ "epoch": 58.90829694323144,
+ "grad_norm": 0.01261200476437807,
+ "learning_rate": 0.0006,
+ "loss": 4.534193992614746,
+ "step": 4241
+ },
+ {
+ "epoch": 58.92227074235808,
+ "grad_norm": 0.012955864891409874,
+ "learning_rate": 0.0006,
+ "loss": 4.697248458862305,
+ "step": 4242
+ },
+ {
+ "epoch": 58.93624454148472,
+ "grad_norm": 0.012134749442338943,
+ "learning_rate": 0.0006,
+ "loss": 4.557095527648926,
+ "step": 4243
+ },
+ {
+ "epoch": 58.95021834061135,
+ "grad_norm": 0.012164677493274212,
+ "learning_rate": 0.0006,
+ "loss": 4.623579025268555,
+ "step": 4244
+ },
+ {
+ "epoch": 58.96419213973799,
+ "grad_norm": 0.013802947476506233,
+ "learning_rate": 0.0006,
+ "loss": 4.562838077545166,
+ "step": 4245
+ },
+ {
+ "epoch": 58.97816593886463,
+ "grad_norm": 0.011420476250350475,
+ "learning_rate": 0.0006,
+ "loss": 4.62265682220459,
+ "step": 4246
+ },
+ {
+ "epoch": 58.992139737991266,
+ "grad_norm": 0.01169576682150364,
+ "learning_rate": 0.0006,
+ "loss": 4.727449417114258,
+ "step": 4247
+ },
+ {
+ "epoch": 59.0,
+ "grad_norm": 0.013577710837125778,
+ "learning_rate": 0.0006,
+ "loss": 4.536325454711914,
+ "step": 4248
+ },
+ {
+ "epoch": 59.0,
+ "eval_loss": 4.86527681350708,
+ "eval_runtime": 60.4463,
+ "eval_samples_per_second": 40.399,
+ "eval_steps_per_second": 1.274,
+ "step": 4248
+ },
+ {
+ "epoch": 59.01397379912664,
+ "grad_norm": 0.015251655131578445,
+ "learning_rate": 0.0006,
+ "loss": 4.634598731994629,
+ "step": 4249
+ },
+ {
+ "epoch": 59.02794759825328,
+ "grad_norm": 0.017414938658475876,
+ "learning_rate": 0.0006,
+ "loss": 4.521329879760742,
+ "step": 4250
+ },
+ {
+ "epoch": 59.041921397379916,
+ "grad_norm": 0.01656588353216648,
+ "learning_rate": 0.0006,
+ "loss": 4.446192264556885,
+ "step": 4251
+ },
+ {
+ "epoch": 59.05589519650655,
+ "grad_norm": 0.012038925662636757,
+ "learning_rate": 0.0006,
+ "loss": 4.523855209350586,
+ "step": 4252
+ },
+ {
+ "epoch": 59.069868995633186,
+ "grad_norm": 0.014403033070266247,
+ "learning_rate": 0.0006,
+ "loss": 4.561079978942871,
+ "step": 4253
+ },
+ {
+ "epoch": 59.083842794759825,
+ "grad_norm": 0.016243597492575645,
+ "learning_rate": 0.0006,
+ "loss": 4.673602104187012,
+ "step": 4254
+ },
+ {
+ "epoch": 59.09781659388646,
+ "grad_norm": 0.016492877155542374,
+ "learning_rate": 0.0006,
+ "loss": 4.564824104309082,
+ "step": 4255
+ },
+ {
+ "epoch": 59.1117903930131,
+ "grad_norm": 0.015980370342731476,
+ "learning_rate": 0.0006,
+ "loss": 4.339249134063721,
+ "step": 4256
+ },
+ {
+ "epoch": 59.12576419213974,
+ "grad_norm": 0.019687356427311897,
+ "learning_rate": 0.0006,
+ "loss": 4.481834888458252,
+ "step": 4257
+ },
+ {
+ "epoch": 59.13973799126637,
+ "grad_norm": 0.02983798086643219,
+ "learning_rate": 0.0006,
+ "loss": 4.4495697021484375,
+ "step": 4258
+ },
+ {
+ "epoch": 59.15371179039301,
+ "grad_norm": 0.04600981995463371,
+ "learning_rate": 0.0006,
+ "loss": 4.567838668823242,
+ "step": 4259
+ },
+ {
+ "epoch": 59.16768558951965,
+ "grad_norm": 0.04852001741528511,
+ "learning_rate": 0.0006,
+ "loss": 4.594701766967773,
+ "step": 4260
+ },
+ {
+ "epoch": 59.18165938864629,
+ "grad_norm": 0.03429504856467247,
+ "learning_rate": 0.0006,
+ "loss": 4.573735237121582,
+ "step": 4261
+ },
+ {
+ "epoch": 59.19563318777293,
+ "grad_norm": 0.029345160350203514,
+ "learning_rate": 0.0006,
+ "loss": 4.488258361816406,
+ "step": 4262
+ },
+ {
+ "epoch": 59.209606986899566,
+ "grad_norm": 0.01961623504757881,
+ "learning_rate": 0.0006,
+ "loss": 4.545222282409668,
+ "step": 4263
+ },
+ {
+ "epoch": 59.223580786026204,
+ "grad_norm": 0.020406130701303482,
+ "learning_rate": 0.0006,
+ "loss": 4.471898078918457,
+ "step": 4264
+ },
+ {
+ "epoch": 59.237554585152836,
+ "grad_norm": 0.020500272512435913,
+ "learning_rate": 0.0006,
+ "loss": 4.538337707519531,
+ "step": 4265
+ },
+ {
+ "epoch": 59.251528384279474,
+ "grad_norm": 0.020727017894387245,
+ "learning_rate": 0.0006,
+ "loss": 4.437284469604492,
+ "step": 4266
+ },
+ {
+ "epoch": 59.26550218340611,
+ "grad_norm": 0.03395998477935791,
+ "learning_rate": 0.0006,
+ "loss": 4.459124565124512,
+ "step": 4267
+ },
+ {
+ "epoch": 59.27947598253275,
+ "grad_norm": 0.04635605961084366,
+ "learning_rate": 0.0006,
+ "loss": 4.520228385925293,
+ "step": 4268
+ },
+ {
+ "epoch": 59.29344978165939,
+ "grad_norm": 0.04253006726503372,
+ "learning_rate": 0.0006,
+ "loss": 4.578673362731934,
+ "step": 4269
+ },
+ {
+ "epoch": 59.30742358078603,
+ "grad_norm": 0.02403509058058262,
+ "learning_rate": 0.0006,
+ "loss": 4.484639644622803,
+ "step": 4270
+ },
+ {
+ "epoch": 59.32139737991266,
+ "grad_norm": 0.02254190482199192,
+ "learning_rate": 0.0006,
+ "loss": 4.511314868927002,
+ "step": 4271
+ },
+ {
+ "epoch": 59.3353711790393,
+ "grad_norm": 0.021585190668702126,
+ "learning_rate": 0.0006,
+ "loss": 4.530239105224609,
+ "step": 4272
+ },
+ {
+ "epoch": 59.34934497816594,
+ "grad_norm": 0.021210316568613052,
+ "learning_rate": 0.0006,
+ "loss": 4.593514442443848,
+ "step": 4273
+ },
+ {
+ "epoch": 59.36331877729258,
+ "grad_norm": 0.019317107275128365,
+ "learning_rate": 0.0006,
+ "loss": 4.5591535568237305,
+ "step": 4274
+ },
+ {
+ "epoch": 59.377292576419215,
+ "grad_norm": 0.018059708178043365,
+ "learning_rate": 0.0006,
+ "loss": 4.4499592781066895,
+ "step": 4275
+ },
+ {
+ "epoch": 59.391266375545854,
+ "grad_norm": 0.017847446724772453,
+ "learning_rate": 0.0006,
+ "loss": 4.473999500274658,
+ "step": 4276
+ },
+ {
+ "epoch": 59.40524017467249,
+ "grad_norm": 0.014944756403565407,
+ "learning_rate": 0.0006,
+ "loss": 4.5934529304504395,
+ "step": 4277
+ },
+ {
+ "epoch": 59.419213973799124,
+ "grad_norm": 0.01663832925260067,
+ "learning_rate": 0.0006,
+ "loss": 4.504262924194336,
+ "step": 4278
+ },
+ {
+ "epoch": 59.43318777292576,
+ "grad_norm": 0.015571796335279942,
+ "learning_rate": 0.0006,
+ "loss": 4.512310981750488,
+ "step": 4279
+ },
+ {
+ "epoch": 59.4471615720524,
+ "grad_norm": 0.015896400436758995,
+ "learning_rate": 0.0006,
+ "loss": 4.452714920043945,
+ "step": 4280
+ },
+ {
+ "epoch": 59.46113537117904,
+ "grad_norm": 0.016010284423828125,
+ "learning_rate": 0.0006,
+ "loss": 4.6188063621521,
+ "step": 4281
+ },
+ {
+ "epoch": 59.47510917030568,
+ "grad_norm": 0.015075593255460262,
+ "learning_rate": 0.0006,
+ "loss": 4.477978229522705,
+ "step": 4282
+ },
+ {
+ "epoch": 59.48908296943232,
+ "grad_norm": 0.014698950573801994,
+ "learning_rate": 0.0006,
+ "loss": 4.544955730438232,
+ "step": 4283
+ },
+ {
+ "epoch": 59.50305676855895,
+ "grad_norm": 0.0136245833709836,
+ "learning_rate": 0.0006,
+ "loss": 4.530460357666016,
+ "step": 4284
+ },
+ {
+ "epoch": 59.51703056768559,
+ "grad_norm": 0.013118310831487179,
+ "learning_rate": 0.0006,
+ "loss": 4.614217758178711,
+ "step": 4285
+ },
+ {
+ "epoch": 59.531004366812226,
+ "grad_norm": 0.01295961532741785,
+ "learning_rate": 0.0006,
+ "loss": 4.43156099319458,
+ "step": 4286
+ },
+ {
+ "epoch": 59.544978165938865,
+ "grad_norm": 0.012429811991751194,
+ "learning_rate": 0.0006,
+ "loss": 4.517333030700684,
+ "step": 4287
+ },
+ {
+ "epoch": 59.5589519650655,
+ "grad_norm": 0.01261830423027277,
+ "learning_rate": 0.0006,
+ "loss": 4.569780349731445,
+ "step": 4288
+ },
+ {
+ "epoch": 59.57292576419214,
+ "grad_norm": 0.012394499965012074,
+ "learning_rate": 0.0006,
+ "loss": 4.5970869064331055,
+ "step": 4289
+ },
+ {
+ "epoch": 59.58689956331878,
+ "grad_norm": 0.01290759164839983,
+ "learning_rate": 0.0006,
+ "loss": 4.437100410461426,
+ "step": 4290
+ },
+ {
+ "epoch": 59.60087336244541,
+ "grad_norm": 0.012042169459164143,
+ "learning_rate": 0.0006,
+ "loss": 4.567404747009277,
+ "step": 4291
+ },
+ {
+ "epoch": 59.61484716157205,
+ "grad_norm": 0.012338520959019661,
+ "learning_rate": 0.0006,
+ "loss": 4.476193428039551,
+ "step": 4292
+ },
+ {
+ "epoch": 59.62882096069869,
+ "grad_norm": 0.01321258582174778,
+ "learning_rate": 0.0006,
+ "loss": 4.566680431365967,
+ "step": 4293
+ },
+ {
+ "epoch": 59.64279475982533,
+ "grad_norm": 0.013478122651576996,
+ "learning_rate": 0.0006,
+ "loss": 4.487398147583008,
+ "step": 4294
+ },
+ {
+ "epoch": 59.65676855895197,
+ "grad_norm": 0.01537603884935379,
+ "learning_rate": 0.0006,
+ "loss": 4.449337005615234,
+ "step": 4295
+ },
+ {
+ "epoch": 59.670742358078606,
+ "grad_norm": 0.017140565440058708,
+ "learning_rate": 0.0006,
+ "loss": 4.5174360275268555,
+ "step": 4296
+ },
+ {
+ "epoch": 59.68471615720524,
+ "grad_norm": 0.017931461334228516,
+ "learning_rate": 0.0006,
+ "loss": 4.522104740142822,
+ "step": 4297
+ },
+ {
+ "epoch": 59.698689956331876,
+ "grad_norm": 0.018037918955087662,
+ "learning_rate": 0.0006,
+ "loss": 4.496928691864014,
+ "step": 4298
+ },
+ {
+ "epoch": 59.712663755458514,
+ "grad_norm": 0.01684504561126232,
+ "learning_rate": 0.0006,
+ "loss": 4.5332417488098145,
+ "step": 4299
+ },
+ {
+ "epoch": 59.72663755458515,
+ "grad_norm": 0.016268383711576462,
+ "learning_rate": 0.0006,
+ "loss": 4.468893051147461,
+ "step": 4300
+ },
+ {
+ "epoch": 59.74061135371179,
+ "grad_norm": 0.014859385788440704,
+ "learning_rate": 0.0006,
+ "loss": 4.551092147827148,
+ "step": 4301
+ },
+ {
+ "epoch": 59.75458515283843,
+ "grad_norm": 0.013589047826826572,
+ "learning_rate": 0.0006,
+ "loss": 4.470579147338867,
+ "step": 4302
+ },
+ {
+ "epoch": 59.76855895196506,
+ "grad_norm": 0.015704551711678505,
+ "learning_rate": 0.0006,
+ "loss": 4.330068111419678,
+ "step": 4303
+ },
+ {
+ "epoch": 59.7825327510917,
+ "grad_norm": 0.017869921401143074,
+ "learning_rate": 0.0006,
+ "loss": 4.428255081176758,
+ "step": 4304
+ },
+ {
+ "epoch": 59.79650655021834,
+ "grad_norm": 0.019828537479043007,
+ "learning_rate": 0.0006,
+ "loss": 4.496023178100586,
+ "step": 4305
+ },
+ {
+ "epoch": 59.81048034934498,
+ "grad_norm": 0.015856236219406128,
+ "learning_rate": 0.0006,
+ "loss": 4.427152156829834,
+ "step": 4306
+ },
+ {
+ "epoch": 59.82445414847162,
+ "grad_norm": 0.013349304907023907,
+ "learning_rate": 0.0006,
+ "loss": 4.417333602905273,
+ "step": 4307
+ },
+ {
+ "epoch": 59.838427947598255,
+ "grad_norm": 0.014111637137830257,
+ "learning_rate": 0.0006,
+ "loss": 4.4507575035095215,
+ "step": 4308
+ },
+ {
+ "epoch": 59.852401746724894,
+ "grad_norm": 0.015186597593128681,
+ "learning_rate": 0.0006,
+ "loss": 4.375116348266602,
+ "step": 4309
+ },
+ {
+ "epoch": 59.866375545851525,
+ "grad_norm": 0.01604730449616909,
+ "learning_rate": 0.0006,
+ "loss": 4.604518890380859,
+ "step": 4310
+ },
+ {
+ "epoch": 59.880349344978164,
+ "grad_norm": 0.015308464877307415,
+ "learning_rate": 0.0006,
+ "loss": 4.507232189178467,
+ "step": 4311
+ },
+ {
+ "epoch": 59.8943231441048,
+ "grad_norm": 0.014017206616699696,
+ "learning_rate": 0.0006,
+ "loss": 4.398487091064453,
+ "step": 4312
+ },
+ {
+ "epoch": 59.90829694323144,
+ "grad_norm": 0.012911394238471985,
+ "learning_rate": 0.0006,
+ "loss": 4.567872047424316,
+ "step": 4313
+ },
+ {
+ "epoch": 59.92227074235808,
+ "grad_norm": 0.012775142677128315,
+ "learning_rate": 0.0006,
+ "loss": 4.562105655670166,
+ "step": 4314
+ },
+ {
+ "epoch": 59.93624454148472,
+ "grad_norm": 0.014750408008694649,
+ "learning_rate": 0.0006,
+ "loss": 4.456263542175293,
+ "step": 4315
+ },
+ {
+ "epoch": 59.95021834061135,
+ "grad_norm": 0.016726011410355568,
+ "learning_rate": 0.0006,
+ "loss": 4.360722064971924,
+ "step": 4316
+ },
+ {
+ "epoch": 59.96419213973799,
+ "grad_norm": 0.01734776981174946,
+ "learning_rate": 0.0006,
+ "loss": 4.572544574737549,
+ "step": 4317
+ },
+ {
+ "epoch": 59.97816593886463,
+ "grad_norm": 0.01822415366768837,
+ "learning_rate": 0.0006,
+ "loss": 4.490501880645752,
+ "step": 4318
+ },
+ {
+ "epoch": 59.992139737991266,
+ "grad_norm": 0.014638577587902546,
+ "learning_rate": 0.0006,
+ "loss": 4.440951824188232,
+ "step": 4319
+ },
+ {
+ "epoch": 60.0,
+ "grad_norm": 0.01255202479660511,
+ "learning_rate": 0.0006,
+ "loss": 4.559123516082764,
+ "step": 4320
+ },
+ {
+ "epoch": 60.0,
+ "eval_loss": 4.741089344024658,
+ "eval_runtime": 56.8997,
+ "eval_samples_per_second": 42.918,
+ "eval_steps_per_second": 1.353,
+ "step": 4320
+ },
+ {
+ "epoch": 60.01397379912664,
+ "grad_norm": 0.016548393294215202,
+ "learning_rate": 0.0006,
+ "loss": 4.442024230957031,
+ "step": 4321
+ },
+ {
+ "epoch": 60.02794759825328,
+ "grad_norm": 0.019759127870202065,
+ "learning_rate": 0.0006,
+ "loss": 4.394334316253662,
+ "step": 4322
+ },
+ {
+ "epoch": 60.041921397379916,
+ "grad_norm": 0.018325170502066612,
+ "learning_rate": 0.0006,
+ "loss": 4.318270683288574,
+ "step": 4323
+ },
+ {
+ "epoch": 60.05589519650655,
+ "grad_norm": 0.015289463102817535,
+ "learning_rate": 0.0006,
+ "loss": 4.318228721618652,
+ "step": 4324
+ },
+ {
+ "epoch": 60.069868995633186,
+ "grad_norm": 0.015424872748553753,
+ "learning_rate": 0.0006,
+ "loss": 4.3605499267578125,
+ "step": 4325
+ },
+ {
+ "epoch": 60.083842794759825,
+ "grad_norm": 0.014085669070482254,
+ "learning_rate": 0.0006,
+ "loss": 4.359585762023926,
+ "step": 4326
+ },
+ {
+ "epoch": 60.09781659388646,
+ "grad_norm": 0.014004210010170937,
+ "learning_rate": 0.0006,
+ "loss": 4.447783470153809,
+ "step": 4327
+ },
+ {
+ "epoch": 60.1117903930131,
+ "grad_norm": 0.013514582999050617,
+ "learning_rate": 0.0006,
+ "loss": 4.4850239753723145,
+ "step": 4328
+ },
+ {
+ "epoch": 60.12576419213974,
+ "grad_norm": 0.013049394823610783,
+ "learning_rate": 0.0006,
+ "loss": 4.528675556182861,
+ "step": 4329
+ },
+ {
+ "epoch": 60.13973799126637,
+ "grad_norm": 0.01378058921545744,
+ "learning_rate": 0.0006,
+ "loss": 4.444823741912842,
+ "step": 4330
+ },
+ {
+ "epoch": 60.15371179039301,
+ "grad_norm": 0.013752233237028122,
+ "learning_rate": 0.0006,
+ "loss": 4.5205206871032715,
+ "step": 4331
+ },
+ {
+ "epoch": 60.16768558951965,
+ "grad_norm": 0.013431803323328495,
+ "learning_rate": 0.0006,
+ "loss": 4.547320365905762,
+ "step": 4332
+ },
+ {
+ "epoch": 60.18165938864629,
+ "grad_norm": 0.014544196426868439,
+ "learning_rate": 0.0006,
+ "loss": 4.517143249511719,
+ "step": 4333
+ },
+ {
+ "epoch": 60.19563318777293,
+ "grad_norm": 0.01612176187336445,
+ "learning_rate": 0.0006,
+ "loss": 4.517232894897461,
+ "step": 4334
+ },
+ {
+ "epoch": 60.209606986899566,
+ "grad_norm": 0.01803239807486534,
+ "learning_rate": 0.0006,
+ "loss": 4.377574920654297,
+ "step": 4335
+ },
+ {
+ "epoch": 60.223580786026204,
+ "grad_norm": 0.014871489256620407,
+ "learning_rate": 0.0006,
+ "loss": 4.538285255432129,
+ "step": 4336
+ },
+ {
+ "epoch": 60.237554585152836,
+ "grad_norm": 0.013927377760410309,
+ "learning_rate": 0.0006,
+ "loss": 4.386175632476807,
+ "step": 4337
+ },
+ {
+ "epoch": 60.251528384279474,
+ "grad_norm": 0.014757219702005386,
+ "learning_rate": 0.0006,
+ "loss": 4.528553009033203,
+ "step": 4338
+ },
+ {
+ "epoch": 60.26550218340611,
+ "grad_norm": 0.01433873176574707,
+ "learning_rate": 0.0006,
+ "loss": 4.476929187774658,
+ "step": 4339
+ },
+ {
+ "epoch": 60.27947598253275,
+ "grad_norm": 0.015768803656101227,
+ "learning_rate": 0.0006,
+ "loss": 4.50970458984375,
+ "step": 4340
+ },
+ {
+ "epoch": 60.29344978165939,
+ "grad_norm": 0.015651429072022438,
+ "learning_rate": 0.0006,
+ "loss": 4.427133083343506,
+ "step": 4341
+ },
+ {
+ "epoch": 60.30742358078603,
+ "grad_norm": 0.01673000678420067,
+ "learning_rate": 0.0006,
+ "loss": 4.447723865509033,
+ "step": 4342
+ },
+ {
+ "epoch": 60.32139737991266,
+ "grad_norm": 0.018481194972991943,
+ "learning_rate": 0.0006,
+ "loss": 4.559832572937012,
+ "step": 4343
+ },
+ {
+ "epoch": 60.3353711790393,
+ "grad_norm": 0.016731027513742447,
+ "learning_rate": 0.0006,
+ "loss": 4.470893859863281,
+ "step": 4344
+ },
+ {
+ "epoch": 60.34934497816594,
+ "grad_norm": 0.01504011545330286,
+ "learning_rate": 0.0006,
+ "loss": 4.483014106750488,
+ "step": 4345
+ },
+ {
+ "epoch": 60.36331877729258,
+ "grad_norm": 0.01602456159889698,
+ "learning_rate": 0.0006,
+ "loss": 4.548428058624268,
+ "step": 4346
+ },
+ {
+ "epoch": 60.377292576419215,
+ "grad_norm": 0.016027364879846573,
+ "learning_rate": 0.0006,
+ "loss": 4.426779747009277,
+ "step": 4347
+ },
+ {
+ "epoch": 60.391266375545854,
+ "grad_norm": 0.01873827911913395,
+ "learning_rate": 0.0006,
+ "loss": 4.353686332702637,
+ "step": 4348
+ },
+ {
+ "epoch": 60.40524017467249,
+ "grad_norm": 0.01907406374812126,
+ "learning_rate": 0.0006,
+ "loss": 4.502945423126221,
+ "step": 4349
+ },
+ {
+ "epoch": 60.419213973799124,
+ "grad_norm": 0.016083354130387306,
+ "learning_rate": 0.0006,
+ "loss": 4.431731700897217,
+ "step": 4350
+ },
+ {
+ "epoch": 60.43318777292576,
+ "grad_norm": 0.015293709933757782,
+ "learning_rate": 0.0006,
+ "loss": 4.480112552642822,
+ "step": 4351
+ },
+ {
+ "epoch": 60.4471615720524,
+ "grad_norm": 0.01688101328909397,
+ "learning_rate": 0.0006,
+ "loss": 4.50853157043457,
+ "step": 4352
+ },
+ {
+ "epoch": 60.46113537117904,
+ "grad_norm": 0.02094118855893612,
+ "learning_rate": 0.0006,
+ "loss": 4.359930038452148,
+ "step": 4353
+ },
+ {
+ "epoch": 60.47510917030568,
+ "grad_norm": 0.018188107758760452,
+ "learning_rate": 0.0006,
+ "loss": 4.549592018127441,
+ "step": 4354
+ },
+ {
+ "epoch": 60.48908296943232,
+ "grad_norm": 0.01462319865822792,
+ "learning_rate": 0.0006,
+ "loss": 4.49467658996582,
+ "step": 4355
+ },
+ {
+ "epoch": 60.50305676855895,
+ "grad_norm": 0.017508579418063164,
+ "learning_rate": 0.0006,
+ "loss": 4.398318767547607,
+ "step": 4356
+ },
+ {
+ "epoch": 60.51703056768559,
+ "grad_norm": 0.018200315535068512,
+ "learning_rate": 0.0006,
+ "loss": 4.44169282913208,
+ "step": 4357
+ },
+ {
+ "epoch": 60.531004366812226,
+ "grad_norm": 0.01717468351125717,
+ "learning_rate": 0.0006,
+ "loss": 4.481400489807129,
+ "step": 4358
+ },
+ {
+ "epoch": 60.544978165938865,
+ "grad_norm": 0.019329151138663292,
+ "learning_rate": 0.0006,
+ "loss": 4.578839302062988,
+ "step": 4359
+ },
+ {
+ "epoch": 60.5589519650655,
+ "grad_norm": 0.016109555959701538,
+ "learning_rate": 0.0006,
+ "loss": 4.461043357849121,
+ "step": 4360
+ },
+ {
+ "epoch": 60.57292576419214,
+ "grad_norm": 0.01494457945227623,
+ "learning_rate": 0.0006,
+ "loss": 4.505096912384033,
+ "step": 4361
+ },
+ {
+ "epoch": 60.58689956331878,
+ "grad_norm": 0.01729399710893631,
+ "learning_rate": 0.0006,
+ "loss": 4.366747856140137,
+ "step": 4362
+ },
+ {
+ "epoch": 60.60087336244541,
+ "grad_norm": 0.01705821603536606,
+ "learning_rate": 0.0006,
+ "loss": 4.42534065246582,
+ "step": 4363
+ },
+ {
+ "epoch": 60.61484716157205,
+ "grad_norm": 0.018391354009509087,
+ "learning_rate": 0.0006,
+ "loss": 4.350294589996338,
+ "step": 4364
+ },
+ {
+ "epoch": 60.62882096069869,
+ "grad_norm": 0.016989829018712044,
+ "learning_rate": 0.0006,
+ "loss": 4.45905876159668,
+ "step": 4365
+ },
+ {
+ "epoch": 60.64279475982533,
+ "grad_norm": 0.017192212864756584,
+ "learning_rate": 0.0006,
+ "loss": 4.526487350463867,
+ "step": 4366
+ },
+ {
+ "epoch": 60.65676855895197,
+ "grad_norm": 0.018091315403580666,
+ "learning_rate": 0.0006,
+ "loss": 4.309333324432373,
+ "step": 4367
+ },
+ {
+ "epoch": 60.670742358078606,
+ "grad_norm": 0.014051459729671478,
+ "learning_rate": 0.0006,
+ "loss": 4.338171005249023,
+ "step": 4368
+ },
+ {
+ "epoch": 60.68471615720524,
+ "grad_norm": 0.013166418299078941,
+ "learning_rate": 0.0006,
+ "loss": 4.409758567810059,
+ "step": 4369
+ },
+ {
+ "epoch": 60.698689956331876,
+ "grad_norm": 0.014260428957641125,
+ "learning_rate": 0.0006,
+ "loss": 4.524338245391846,
+ "step": 4370
+ },
+ {
+ "epoch": 60.712663755458514,
+ "grad_norm": 0.017019255086779594,
+ "learning_rate": 0.0006,
+ "loss": 4.387436866760254,
+ "step": 4371
+ },
+ {
+ "epoch": 60.72663755458515,
+ "grad_norm": 0.017070675268769264,
+ "learning_rate": 0.0006,
+ "loss": 4.437321662902832,
+ "step": 4372
+ },
+ {
+ "epoch": 60.74061135371179,
+ "grad_norm": 0.015197164379060268,
+ "learning_rate": 0.0006,
+ "loss": 4.371205806732178,
+ "step": 4373
+ },
+ {
+ "epoch": 60.75458515283843,
+ "grad_norm": 0.014791185967624187,
+ "learning_rate": 0.0006,
+ "loss": 4.44439697265625,
+ "step": 4374
+ },
+ {
+ "epoch": 60.76855895196506,
+ "grad_norm": 0.016458261758089066,
+ "learning_rate": 0.0006,
+ "loss": 4.445584297180176,
+ "step": 4375
+ },
+ {
+ "epoch": 60.7825327510917,
+ "grad_norm": 0.017706342041492462,
+ "learning_rate": 0.0006,
+ "loss": 4.348708152770996,
+ "step": 4376
+ },
+ {
+ "epoch": 60.79650655021834,
+ "grad_norm": 0.017561476677656174,
+ "learning_rate": 0.0006,
+ "loss": 4.521929740905762,
+ "step": 4377
+ },
+ {
+ "epoch": 60.81048034934498,
+ "grad_norm": 0.016557445749640465,
+ "learning_rate": 0.0006,
+ "loss": 4.545806884765625,
+ "step": 4378
+ },
+ {
+ "epoch": 60.82445414847162,
+ "grad_norm": 0.016953222453594208,
+ "learning_rate": 0.0006,
+ "loss": 4.4230875968933105,
+ "step": 4379
+ },
+ {
+ "epoch": 60.838427947598255,
+ "grad_norm": 0.017094967886805534,
+ "learning_rate": 0.0006,
+ "loss": 4.423985481262207,
+ "step": 4380
+ },
+ {
+ "epoch": 60.852401746724894,
+ "grad_norm": 0.014787515625357628,
+ "learning_rate": 0.0006,
+ "loss": 4.39473819732666,
+ "step": 4381
+ },
+ {
+ "epoch": 60.866375545851525,
+ "grad_norm": 0.016208263114094734,
+ "learning_rate": 0.0006,
+ "loss": 4.3810038566589355,
+ "step": 4382
+ },
+ {
+ "epoch": 60.880349344978164,
+ "grad_norm": 0.017844321206212044,
+ "learning_rate": 0.0006,
+ "loss": 4.35086727142334,
+ "step": 4383
+ },
+ {
+ "epoch": 60.8943231441048,
+ "grad_norm": 0.017539294436573982,
+ "learning_rate": 0.0006,
+ "loss": 4.460260391235352,
+ "step": 4384
+ },
+ {
+ "epoch": 60.90829694323144,
+ "grad_norm": 0.01612042263150215,
+ "learning_rate": 0.0006,
+ "loss": 4.370500564575195,
+ "step": 4385
+ },
+ {
+ "epoch": 60.92227074235808,
+ "grad_norm": 0.015381601639091969,
+ "learning_rate": 0.0006,
+ "loss": 4.435868263244629,
+ "step": 4386
+ },
+ {
+ "epoch": 60.93624454148472,
+ "grad_norm": 0.01603585295379162,
+ "learning_rate": 0.0006,
+ "loss": 4.434070587158203,
+ "step": 4387
+ },
+ {
+ "epoch": 60.95021834061135,
+ "grad_norm": 0.016268130391836166,
+ "learning_rate": 0.0006,
+ "loss": 4.404343128204346,
+ "step": 4388
+ },
+ {
+ "epoch": 60.96419213973799,
+ "grad_norm": 0.014980032108724117,
+ "learning_rate": 0.0006,
+ "loss": 4.399213790893555,
+ "step": 4389
+ },
+ {
+ "epoch": 60.97816593886463,
+ "grad_norm": 0.016899267211556435,
+ "learning_rate": 0.0006,
+ "loss": 4.508069038391113,
+ "step": 4390
+ },
+ {
+ "epoch": 60.992139737991266,
+ "grad_norm": 0.013756908476352692,
+ "learning_rate": 0.0006,
+ "loss": 4.541858673095703,
+ "step": 4391
+ },
+ {
+ "epoch": 61.0,
+ "grad_norm": 0.01578413136303425,
+ "learning_rate": 0.0006,
+ "loss": 4.543513298034668,
+ "step": 4392
+ },
+ {
+ "epoch": 61.0,
+ "eval_loss": 4.756770610809326,
+ "eval_runtime": 56.4572,
+ "eval_samples_per_second": 43.254,
+ "eval_steps_per_second": 1.364,
+ "step": 4392
+ },
+ {
+ "epoch": 61.01397379912664,
+ "grad_norm": 0.013387506827712059,
+ "learning_rate": 0.0006,
+ "loss": 4.475942611694336,
+ "step": 4393
+ },
+ {
+ "epoch": 61.02794759825328,
+ "grad_norm": 0.014278904534876347,
+ "learning_rate": 0.0006,
+ "loss": 4.442403793334961,
+ "step": 4394
+ },
+ {
+ "epoch": 61.041921397379916,
+ "grad_norm": 0.014359553344547749,
+ "learning_rate": 0.0006,
+ "loss": 4.509463310241699,
+ "step": 4395
+ },
+ {
+ "epoch": 61.05589519650655,
+ "grad_norm": 0.01718050241470337,
+ "learning_rate": 0.0006,
+ "loss": 4.354458808898926,
+ "step": 4396
+ },
+ {
+ "epoch": 61.069868995633186,
+ "grad_norm": 0.015988312661647797,
+ "learning_rate": 0.0006,
+ "loss": 4.484508514404297,
+ "step": 4397
+ },
+ {
+ "epoch": 61.083842794759825,
+ "grad_norm": 0.015466735698282719,
+ "learning_rate": 0.0006,
+ "loss": 4.326908111572266,
+ "step": 4398
+ },
+ {
+ "epoch": 61.09781659388646,
+ "grad_norm": 0.014541332609951496,
+ "learning_rate": 0.0006,
+ "loss": 4.405823230743408,
+ "step": 4399
+ },
+ {
+ "epoch": 61.1117903930131,
+ "grad_norm": 0.014803987927734852,
+ "learning_rate": 0.0006,
+ "loss": 4.375386714935303,
+ "step": 4400
+ },
+ {
+ "epoch": 61.12576419213974,
+ "grad_norm": 0.013962093740701675,
+ "learning_rate": 0.0006,
+ "loss": 4.300838470458984,
+ "step": 4401
+ },
+ {
+ "epoch": 61.13973799126637,
+ "grad_norm": 0.015213954262435436,
+ "learning_rate": 0.0006,
+ "loss": 4.388306617736816,
+ "step": 4402
+ },
+ {
+ "epoch": 61.15371179039301,
+ "grad_norm": 0.018965506926178932,
+ "learning_rate": 0.0006,
+ "loss": 4.388340950012207,
+ "step": 4403
+ },
+ {
+ "epoch": 61.16768558951965,
+ "grad_norm": 0.01958410255610943,
+ "learning_rate": 0.0006,
+ "loss": 4.35012149810791,
+ "step": 4404
+ },
+ {
+ "epoch": 61.18165938864629,
+ "grad_norm": 0.01938101276755333,
+ "learning_rate": 0.0006,
+ "loss": 4.368372917175293,
+ "step": 4405
+ },
+ {
+ "epoch": 61.19563318777293,
+ "grad_norm": 0.016556516289711,
+ "learning_rate": 0.0006,
+ "loss": 4.486198425292969,
+ "step": 4406
+ },
+ {
+ "epoch": 61.209606986899566,
+ "grad_norm": 0.017031289637088776,
+ "learning_rate": 0.0006,
+ "loss": 4.421676158905029,
+ "step": 4407
+ },
+ {
+ "epoch": 61.223580786026204,
+ "grad_norm": 0.018023479729890823,
+ "learning_rate": 0.0006,
+ "loss": 4.38751220703125,
+ "step": 4408
+ },
+ {
+ "epoch": 61.237554585152836,
+ "grad_norm": 0.0178135447204113,
+ "learning_rate": 0.0006,
+ "loss": 4.336528301239014,
+ "step": 4409
+ },
+ {
+ "epoch": 61.251528384279474,
+ "grad_norm": 0.016782190650701523,
+ "learning_rate": 0.0006,
+ "loss": 4.386693954467773,
+ "step": 4410
+ },
+ {
+ "epoch": 61.26550218340611,
+ "grad_norm": 0.015494848601520061,
+ "learning_rate": 0.0006,
+ "loss": 4.308948040008545,
+ "step": 4411
+ },
+ {
+ "epoch": 61.27947598253275,
+ "grad_norm": 0.015002378262579441,
+ "learning_rate": 0.0006,
+ "loss": 4.314206123352051,
+ "step": 4412
+ },
+ {
+ "epoch": 61.29344978165939,
+ "grad_norm": 0.015798920765519142,
+ "learning_rate": 0.0006,
+ "loss": 4.502900123596191,
+ "step": 4413
+ },
+ {
+ "epoch": 61.30742358078603,
+ "grad_norm": 0.015444336459040642,
+ "learning_rate": 0.0006,
+ "loss": 4.473160743713379,
+ "step": 4414
+ },
+ {
+ "epoch": 61.32139737991266,
+ "grad_norm": 0.017025096341967583,
+ "learning_rate": 0.0006,
+ "loss": 4.460621356964111,
+ "step": 4415
+ },
+ {
+ "epoch": 61.3353711790393,
+ "grad_norm": 0.019502537325024605,
+ "learning_rate": 0.0006,
+ "loss": 4.385551452636719,
+ "step": 4416
+ },
+ {
+ "epoch": 61.34934497816594,
+ "grad_norm": 0.018529541790485382,
+ "learning_rate": 0.0006,
+ "loss": 4.425390720367432,
+ "step": 4417
+ },
+ {
+ "epoch": 61.36331877729258,
+ "grad_norm": 0.013155706226825714,
+ "learning_rate": 0.0006,
+ "loss": 4.467020034790039,
+ "step": 4418
+ },
+ {
+ "epoch": 61.377292576419215,
+ "grad_norm": 0.01518090907484293,
+ "learning_rate": 0.0006,
+ "loss": 4.369664192199707,
+ "step": 4419
+ },
+ {
+ "epoch": 61.391266375545854,
+ "grad_norm": 0.01626511849462986,
+ "learning_rate": 0.0006,
+ "loss": 4.463231086730957,
+ "step": 4420
+ },
+ {
+ "epoch": 61.40524017467249,
+ "grad_norm": 0.01701558195054531,
+ "learning_rate": 0.0006,
+ "loss": 4.43708610534668,
+ "step": 4421
+ },
+ {
+ "epoch": 61.419213973799124,
+ "grad_norm": 0.01761862449347973,
+ "learning_rate": 0.0006,
+ "loss": 4.3314208984375,
+ "step": 4422
+ },
+ {
+ "epoch": 61.43318777292576,
+ "grad_norm": 0.017084648832678795,
+ "learning_rate": 0.0006,
+ "loss": 4.39676570892334,
+ "step": 4423
+ },
+ {
+ "epoch": 61.4471615720524,
+ "grad_norm": 0.01667078025639057,
+ "learning_rate": 0.0006,
+ "loss": 4.3789873123168945,
+ "step": 4424
+ },
+ {
+ "epoch": 61.46113537117904,
+ "grad_norm": 0.01877436414361,
+ "learning_rate": 0.0006,
+ "loss": 4.415275573730469,
+ "step": 4425
+ },
+ {
+ "epoch": 61.47510917030568,
+ "grad_norm": 0.020250486209988594,
+ "learning_rate": 0.0006,
+ "loss": 4.478635787963867,
+ "step": 4426
+ },
+ {
+ "epoch": 61.48908296943232,
+ "grad_norm": 0.020244868472218513,
+ "learning_rate": 0.0006,
+ "loss": 4.296854019165039,
+ "step": 4427
+ },
+ {
+ "epoch": 61.50305676855895,
+ "grad_norm": 0.018723875284194946,
+ "learning_rate": 0.0006,
+ "loss": 4.3760881423950195,
+ "step": 4428
+ },
+ {
+ "epoch": 61.51703056768559,
+ "grad_norm": 0.015138164162635803,
+ "learning_rate": 0.0006,
+ "loss": 4.39974308013916,
+ "step": 4429
+ },
+ {
+ "epoch": 61.531004366812226,
+ "grad_norm": 0.015111138112843037,
+ "learning_rate": 0.0006,
+ "loss": 4.34659481048584,
+ "step": 4430
+ },
+ {
+ "epoch": 61.544978165938865,
+ "grad_norm": 0.0164666585624218,
+ "learning_rate": 0.0006,
+ "loss": 4.415737152099609,
+ "step": 4431
+ },
+ {
+ "epoch": 61.5589519650655,
+ "grad_norm": 0.020891312509775162,
+ "learning_rate": 0.0006,
+ "loss": 4.300319671630859,
+ "step": 4432
+ },
+ {
+ "epoch": 61.57292576419214,
+ "grad_norm": 0.02089156024158001,
+ "learning_rate": 0.0006,
+ "loss": 4.373029708862305,
+ "step": 4433
+ },
+ {
+ "epoch": 61.58689956331878,
+ "grad_norm": 0.020114559680223465,
+ "learning_rate": 0.0006,
+ "loss": 4.434499740600586,
+ "step": 4434
+ },
+ {
+ "epoch": 61.60087336244541,
+ "grad_norm": 0.01791365258395672,
+ "learning_rate": 0.0006,
+ "loss": 4.466872215270996,
+ "step": 4435
+ },
+ {
+ "epoch": 61.61484716157205,
+ "grad_norm": 0.01879458874464035,
+ "learning_rate": 0.0006,
+ "loss": 4.5176520347595215,
+ "step": 4436
+ },
+ {
+ "epoch": 61.62882096069869,
+ "grad_norm": 0.016371937468647957,
+ "learning_rate": 0.0006,
+ "loss": 4.3507585525512695,
+ "step": 4437
+ },
+ {
+ "epoch": 61.64279475982533,
+ "grad_norm": 0.01648143120110035,
+ "learning_rate": 0.0006,
+ "loss": 4.398778438568115,
+ "step": 4438
+ },
+ {
+ "epoch": 61.65676855895197,
+ "grad_norm": 0.017240576446056366,
+ "learning_rate": 0.0006,
+ "loss": 4.513933181762695,
+ "step": 4439
+ },
+ {
+ "epoch": 61.670742358078606,
+ "grad_norm": 0.01734284684062004,
+ "learning_rate": 0.0006,
+ "loss": 4.315343856811523,
+ "step": 4440
+ },
+ {
+ "epoch": 61.68471615720524,
+ "grad_norm": 0.01639772206544876,
+ "learning_rate": 0.0006,
+ "loss": 4.4641218185424805,
+ "step": 4441
+ },
+ {
+ "epoch": 61.698689956331876,
+ "grad_norm": 0.0164708960801363,
+ "learning_rate": 0.0006,
+ "loss": 4.439968585968018,
+ "step": 4442
+ },
+ {
+ "epoch": 61.712663755458514,
+ "grad_norm": 0.01757933758199215,
+ "learning_rate": 0.0006,
+ "loss": 4.408327102661133,
+ "step": 4443
+ },
+ {
+ "epoch": 61.72663755458515,
+ "grad_norm": 0.01789589412510395,
+ "learning_rate": 0.0006,
+ "loss": 4.478257179260254,
+ "step": 4444
+ },
+ {
+ "epoch": 61.74061135371179,
+ "grad_norm": 0.01822184957563877,
+ "learning_rate": 0.0006,
+ "loss": 4.412046909332275,
+ "step": 4445
+ },
+ {
+ "epoch": 61.75458515283843,
+ "grad_norm": 0.018624618649482727,
+ "learning_rate": 0.0006,
+ "loss": 4.427122116088867,
+ "step": 4446
+ },
+ {
+ "epoch": 61.76855895196506,
+ "grad_norm": 0.018674887716770172,
+ "learning_rate": 0.0006,
+ "loss": 4.429147720336914,
+ "step": 4447
+ },
+ {
+ "epoch": 61.7825327510917,
+ "grad_norm": 0.015317119657993317,
+ "learning_rate": 0.0006,
+ "loss": 4.2808451652526855,
+ "step": 4448
+ },
+ {
+ "epoch": 61.79650655021834,
+ "grad_norm": 0.015400845557451248,
+ "learning_rate": 0.0006,
+ "loss": 4.420193672180176,
+ "step": 4449
+ },
+ {
+ "epoch": 61.81048034934498,
+ "grad_norm": 0.016743650659918785,
+ "learning_rate": 0.0006,
+ "loss": 4.498254776000977,
+ "step": 4450
+ },
+ {
+ "epoch": 61.82445414847162,
+ "grad_norm": 0.014330781064927578,
+ "learning_rate": 0.0006,
+ "loss": 4.445384979248047,
+ "step": 4451
+ },
+ {
+ "epoch": 61.838427947598255,
+ "grad_norm": 0.015046479180455208,
+ "learning_rate": 0.0006,
+ "loss": 4.368041038513184,
+ "step": 4452
+ },
+ {
+ "epoch": 61.852401746724894,
+ "grad_norm": 0.01630168780684471,
+ "learning_rate": 0.0006,
+ "loss": 4.460142135620117,
+ "step": 4453
+ },
+ {
+ "epoch": 61.866375545851525,
+ "grad_norm": 0.01637885719537735,
+ "learning_rate": 0.0006,
+ "loss": 4.416606426239014,
+ "step": 4454
+ },
+ {
+ "epoch": 61.880349344978164,
+ "grad_norm": 0.016785888001322746,
+ "learning_rate": 0.0006,
+ "loss": 4.452545642852783,
+ "step": 4455
+ },
+ {
+ "epoch": 61.8943231441048,
+ "grad_norm": 0.015044555068016052,
+ "learning_rate": 0.0006,
+ "loss": 4.427334785461426,
+ "step": 4456
+ },
+ {
+ "epoch": 61.90829694323144,
+ "grad_norm": 0.013699211180210114,
+ "learning_rate": 0.0006,
+ "loss": 4.408761024475098,
+ "step": 4457
+ },
+ {
+ "epoch": 61.92227074235808,
+ "grad_norm": 0.014021803624927998,
+ "learning_rate": 0.0006,
+ "loss": 4.543607711791992,
+ "step": 4458
+ },
+ {
+ "epoch": 61.93624454148472,
+ "grad_norm": 0.015438460744917393,
+ "learning_rate": 0.0006,
+ "loss": 4.313363075256348,
+ "step": 4459
+ },
+ {
+ "epoch": 61.95021834061135,
+ "grad_norm": 0.017515428364276886,
+ "learning_rate": 0.0006,
+ "loss": 4.310412406921387,
+ "step": 4460
+ },
+ {
+ "epoch": 61.96419213973799,
+ "grad_norm": 0.016683880239725113,
+ "learning_rate": 0.0006,
+ "loss": 4.3378448486328125,
+ "step": 4461
+ },
+ {
+ "epoch": 61.97816593886463,
+ "grad_norm": 0.019384048879146576,
+ "learning_rate": 0.0006,
+ "loss": 4.450568675994873,
+ "step": 4462
+ },
+ {
+ "epoch": 61.992139737991266,
+ "grad_norm": 0.020651493221521378,
+ "learning_rate": 0.0006,
+ "loss": 4.453197479248047,
+ "step": 4463
+ },
+ {
+ "epoch": 62.0,
+ "grad_norm": 0.020991764962673187,
+ "learning_rate": 0.0006,
+ "loss": 4.441527366638184,
+ "step": 4464
+ },
+ {
+ "epoch": 62.0,
+ "eval_loss": 4.700193881988525,
+ "eval_runtime": 56.6098,
+ "eval_samples_per_second": 43.137,
+ "eval_steps_per_second": 1.36,
+ "step": 4464
+ },
+ {
+ "epoch": 62.01397379912664,
+ "grad_norm": 0.01811373233795166,
+ "learning_rate": 0.0006,
+ "loss": 4.37346076965332,
+ "step": 4465
+ },
+ {
+ "epoch": 62.02794759825328,
+ "grad_norm": 0.01667127199470997,
+ "learning_rate": 0.0006,
+ "loss": 4.237151145935059,
+ "step": 4466
+ },
+ {
+ "epoch": 62.041921397379916,
+ "grad_norm": 0.014971534721553326,
+ "learning_rate": 0.0006,
+ "loss": 4.367053031921387,
+ "step": 4467
+ },
+ {
+ "epoch": 62.05589519650655,
+ "grad_norm": 0.015655893832445145,
+ "learning_rate": 0.0006,
+ "loss": 4.437426567077637,
+ "step": 4468
+ },
+ {
+ "epoch": 62.069868995633186,
+ "grad_norm": 0.014935447834432125,
+ "learning_rate": 0.0006,
+ "loss": 4.3635125160217285,
+ "step": 4469
+ },
+ {
+ "epoch": 62.083842794759825,
+ "grad_norm": 0.015363802202045918,
+ "learning_rate": 0.0006,
+ "loss": 4.40492057800293,
+ "step": 4470
+ },
+ {
+ "epoch": 62.09781659388646,
+ "grad_norm": 0.0159201230853796,
+ "learning_rate": 0.0006,
+ "loss": 4.392058849334717,
+ "step": 4471
+ },
+ {
+ "epoch": 62.1117903930131,
+ "grad_norm": 0.014598241075873375,
+ "learning_rate": 0.0006,
+ "loss": 4.369944095611572,
+ "step": 4472
+ },
+ {
+ "epoch": 62.12576419213974,
+ "grad_norm": 0.015394841320812702,
+ "learning_rate": 0.0006,
+ "loss": 4.425097942352295,
+ "step": 4473
+ },
+ {
+ "epoch": 62.13973799126637,
+ "grad_norm": 0.015913628041744232,
+ "learning_rate": 0.0006,
+ "loss": 4.420596122741699,
+ "step": 4474
+ },
+ {
+ "epoch": 62.15371179039301,
+ "grad_norm": 0.01575258933007717,
+ "learning_rate": 0.0006,
+ "loss": 4.32290506362915,
+ "step": 4475
+ },
+ {
+ "epoch": 62.16768558951965,
+ "grad_norm": 0.0155344782397151,
+ "learning_rate": 0.0006,
+ "loss": 4.273068428039551,
+ "step": 4476
+ },
+ {
+ "epoch": 62.18165938864629,
+ "grad_norm": 0.014947040006518364,
+ "learning_rate": 0.0006,
+ "loss": 4.384868621826172,
+ "step": 4477
+ },
+ {
+ "epoch": 62.19563318777293,
+ "grad_norm": 0.01557591650635004,
+ "learning_rate": 0.0006,
+ "loss": 4.458410263061523,
+ "step": 4478
+ },
+ {
+ "epoch": 62.209606986899566,
+ "grad_norm": 0.015182198956608772,
+ "learning_rate": 0.0006,
+ "loss": 4.300738334655762,
+ "step": 4479
+ },
+ {
+ "epoch": 62.223580786026204,
+ "grad_norm": 0.01634717360138893,
+ "learning_rate": 0.0006,
+ "loss": 4.436875820159912,
+ "step": 4480
+ },
+ {
+ "epoch": 62.237554585152836,
+ "grad_norm": 0.018813718110322952,
+ "learning_rate": 0.0006,
+ "loss": 4.454665184020996,
+ "step": 4481
+ },
+ {
+ "epoch": 62.251528384279474,
+ "grad_norm": 0.020965630188584328,
+ "learning_rate": 0.0006,
+ "loss": 4.357341766357422,
+ "step": 4482
+ },
+ {
+ "epoch": 62.26550218340611,
+ "grad_norm": 0.020738353952765465,
+ "learning_rate": 0.0006,
+ "loss": 4.456644058227539,
+ "step": 4483
+ },
+ {
+ "epoch": 62.27947598253275,
+ "grad_norm": 0.018813546746969223,
+ "learning_rate": 0.0006,
+ "loss": 4.381941795349121,
+ "step": 4484
+ },
+ {
+ "epoch": 62.29344978165939,
+ "grad_norm": 0.015431761741638184,
+ "learning_rate": 0.0006,
+ "loss": 4.41126823425293,
+ "step": 4485
+ },
+ {
+ "epoch": 62.30742358078603,
+ "grad_norm": 0.016335977241396904,
+ "learning_rate": 0.0006,
+ "loss": 4.419488906860352,
+ "step": 4486
+ },
+ {
+ "epoch": 62.32139737991266,
+ "grad_norm": 0.01832406409084797,
+ "learning_rate": 0.0006,
+ "loss": 4.579026222229004,
+ "step": 4487
+ },
+ {
+ "epoch": 62.3353711790393,
+ "grad_norm": 0.01790742017328739,
+ "learning_rate": 0.0006,
+ "loss": 4.289964199066162,
+ "step": 4488
+ },
+ {
+ "epoch": 62.34934497816594,
+ "grad_norm": 0.017266202718019485,
+ "learning_rate": 0.0006,
+ "loss": 4.415959358215332,
+ "step": 4489
+ },
+ {
+ "epoch": 62.36331877729258,
+ "grad_norm": 0.015976132825016975,
+ "learning_rate": 0.0006,
+ "loss": 4.358587265014648,
+ "step": 4490
+ },
+ {
+ "epoch": 62.377292576419215,
+ "grad_norm": 0.016873881220817566,
+ "learning_rate": 0.0006,
+ "loss": 4.354607105255127,
+ "step": 4491
+ },
+ {
+ "epoch": 62.391266375545854,
+ "grad_norm": 0.01476745493710041,
+ "learning_rate": 0.0006,
+ "loss": 4.384904384613037,
+ "step": 4492
+ },
+ {
+ "epoch": 62.40524017467249,
+ "grad_norm": 0.013810782693326473,
+ "learning_rate": 0.0006,
+ "loss": 4.459230899810791,
+ "step": 4493
+ },
+ {
+ "epoch": 62.419213973799124,
+ "grad_norm": 0.015150421299040318,
+ "learning_rate": 0.0006,
+ "loss": 4.369471549987793,
+ "step": 4494
+ },
+ {
+ "epoch": 62.43318777292576,
+ "grad_norm": 0.014539610594511032,
+ "learning_rate": 0.0006,
+ "loss": 4.3564982414245605,
+ "step": 4495
+ },
+ {
+ "epoch": 62.4471615720524,
+ "grad_norm": 0.015554017387330532,
+ "learning_rate": 0.0006,
+ "loss": 4.414379119873047,
+ "step": 4496
+ },
+ {
+ "epoch": 62.46113537117904,
+ "grad_norm": 0.016608357429504395,
+ "learning_rate": 0.0006,
+ "loss": 4.311585426330566,
+ "step": 4497
+ },
+ {
+ "epoch": 62.47510917030568,
+ "grad_norm": 0.01885881833732128,
+ "learning_rate": 0.0006,
+ "loss": 4.306629180908203,
+ "step": 4498
+ },
+ {
+ "epoch": 62.48908296943232,
+ "grad_norm": 0.022128663957118988,
+ "learning_rate": 0.0006,
+ "loss": 4.345483779907227,
+ "step": 4499
+ },
+ {
+ "epoch": 62.50305676855895,
+ "grad_norm": 0.023401018232107162,
+ "learning_rate": 0.0006,
+ "loss": 4.398204326629639,
+ "step": 4500
+ },
+ {
+ "epoch": 62.51703056768559,
+ "grad_norm": 0.022080009803175926,
+ "learning_rate": 0.0006,
+ "loss": 4.375027179718018,
+ "step": 4501
+ },
+ {
+ "epoch": 62.531004366812226,
+ "grad_norm": 0.020004864782094955,
+ "learning_rate": 0.0006,
+ "loss": 4.33494758605957,
+ "step": 4502
+ },
+ {
+ "epoch": 62.544978165938865,
+ "grad_norm": 0.01786700263619423,
+ "learning_rate": 0.0006,
+ "loss": 4.4428300857543945,
+ "step": 4503
+ },
+ {
+ "epoch": 62.5589519650655,
+ "grad_norm": 0.01885388419032097,
+ "learning_rate": 0.0006,
+ "loss": 4.4138946533203125,
+ "step": 4504
+ },
+ {
+ "epoch": 62.57292576419214,
+ "grad_norm": 0.019890939816832542,
+ "learning_rate": 0.0006,
+ "loss": 4.278632164001465,
+ "step": 4505
+ },
+ {
+ "epoch": 62.58689956331878,
+ "grad_norm": 0.02109031192958355,
+ "learning_rate": 0.0006,
+ "loss": 4.3756327629089355,
+ "step": 4506
+ },
+ {
+ "epoch": 62.60087336244541,
+ "grad_norm": 0.020813899114727974,
+ "learning_rate": 0.0006,
+ "loss": 4.410213470458984,
+ "step": 4507
+ },
+ {
+ "epoch": 62.61484716157205,
+ "grad_norm": 0.021655689924955368,
+ "learning_rate": 0.0006,
+ "loss": 4.30633020401001,
+ "step": 4508
+ },
+ {
+ "epoch": 62.62882096069869,
+ "grad_norm": 0.018155649304389954,
+ "learning_rate": 0.0006,
+ "loss": 4.407284736633301,
+ "step": 4509
+ },
+ {
+ "epoch": 62.64279475982533,
+ "grad_norm": 0.017765069380402565,
+ "learning_rate": 0.0006,
+ "loss": 4.321715354919434,
+ "step": 4510
+ },
+ {
+ "epoch": 62.65676855895197,
+ "grad_norm": 0.017724554985761642,
+ "learning_rate": 0.0006,
+ "loss": 4.493371963500977,
+ "step": 4511
+ },
+ {
+ "epoch": 62.670742358078606,
+ "grad_norm": 0.01492092851549387,
+ "learning_rate": 0.0006,
+ "loss": 4.443025588989258,
+ "step": 4512
+ },
+ {
+ "epoch": 62.68471615720524,
+ "grad_norm": 0.01615324057638645,
+ "learning_rate": 0.0006,
+ "loss": 4.3289594650268555,
+ "step": 4513
+ },
+ {
+ "epoch": 62.698689956331876,
+ "grad_norm": 0.016569366678595543,
+ "learning_rate": 0.0006,
+ "loss": 4.369720458984375,
+ "step": 4514
+ },
+ {
+ "epoch": 62.712663755458514,
+ "grad_norm": 0.016509901732206345,
+ "learning_rate": 0.0006,
+ "loss": 4.292316436767578,
+ "step": 4515
+ },
+ {
+ "epoch": 62.72663755458515,
+ "grad_norm": 0.014270003885030746,
+ "learning_rate": 0.0006,
+ "loss": 4.349854469299316,
+ "step": 4516
+ },
+ {
+ "epoch": 62.74061135371179,
+ "grad_norm": 0.014235340990126133,
+ "learning_rate": 0.0006,
+ "loss": 4.473727226257324,
+ "step": 4517
+ },
+ {
+ "epoch": 62.75458515283843,
+ "grad_norm": 0.015163225121796131,
+ "learning_rate": 0.0006,
+ "loss": 4.452878475189209,
+ "step": 4518
+ },
+ {
+ "epoch": 62.76855895196506,
+ "grad_norm": 0.015748055651783943,
+ "learning_rate": 0.0006,
+ "loss": 4.521087169647217,
+ "step": 4519
+ },
+ {
+ "epoch": 62.7825327510917,
+ "grad_norm": 0.01661108434200287,
+ "learning_rate": 0.0006,
+ "loss": 4.443885326385498,
+ "step": 4520
+ },
+ {
+ "epoch": 62.79650655021834,
+ "grad_norm": 0.01610301062464714,
+ "learning_rate": 0.0006,
+ "loss": 4.312990188598633,
+ "step": 4521
+ },
+ {
+ "epoch": 62.81048034934498,
+ "grad_norm": 0.017007583752274513,
+ "learning_rate": 0.0006,
+ "loss": 4.408937454223633,
+ "step": 4522
+ },
+ {
+ "epoch": 62.82445414847162,
+ "grad_norm": 0.01749461516737938,
+ "learning_rate": 0.0006,
+ "loss": 4.346066951751709,
+ "step": 4523
+ },
+ {
+ "epoch": 62.838427947598255,
+ "grad_norm": 0.018243493512272835,
+ "learning_rate": 0.0006,
+ "loss": 4.357658386230469,
+ "step": 4524
+ },
+ {
+ "epoch": 62.852401746724894,
+ "grad_norm": 0.017499985173344612,
+ "learning_rate": 0.0006,
+ "loss": 4.374061584472656,
+ "step": 4525
+ },
+ {
+ "epoch": 62.866375545851525,
+ "grad_norm": 0.014536325819790363,
+ "learning_rate": 0.0006,
+ "loss": 4.189321994781494,
+ "step": 4526
+ },
+ {
+ "epoch": 62.880349344978164,
+ "grad_norm": 0.014728891663253307,
+ "learning_rate": 0.0006,
+ "loss": 4.425910949707031,
+ "step": 4527
+ },
+ {
+ "epoch": 62.8943231441048,
+ "grad_norm": 0.015013772062957287,
+ "learning_rate": 0.0006,
+ "loss": 4.337319850921631,
+ "step": 4528
+ },
+ {
+ "epoch": 62.90829694323144,
+ "grad_norm": 0.01626027375459671,
+ "learning_rate": 0.0006,
+ "loss": 4.330339431762695,
+ "step": 4529
+ },
+ {
+ "epoch": 62.92227074235808,
+ "grad_norm": 0.014773444272577763,
+ "learning_rate": 0.0006,
+ "loss": 4.4188737869262695,
+ "step": 4530
+ },
+ {
+ "epoch": 62.93624454148472,
+ "grad_norm": 0.01422932744026184,
+ "learning_rate": 0.0006,
+ "loss": 4.355986595153809,
+ "step": 4531
+ },
+ {
+ "epoch": 62.95021834061135,
+ "grad_norm": 0.012170110829174519,
+ "learning_rate": 0.0006,
+ "loss": 4.405722618103027,
+ "step": 4532
+ },
+ {
+ "epoch": 62.96419213973799,
+ "grad_norm": 0.013233253732323647,
+ "learning_rate": 0.0006,
+ "loss": 4.377286911010742,
+ "step": 4533
+ },
+ {
+ "epoch": 62.97816593886463,
+ "grad_norm": 0.014497299678623676,
+ "learning_rate": 0.0006,
+ "loss": 4.508796691894531,
+ "step": 4534
+ },
+ {
+ "epoch": 62.992139737991266,
+ "grad_norm": 0.014718741178512573,
+ "learning_rate": 0.0006,
+ "loss": 4.370919704437256,
+ "step": 4535
+ },
+ {
+ "epoch": 63.0,
+ "grad_norm": 0.015833502635359764,
+ "learning_rate": 0.0006,
+ "loss": 4.258753776550293,
+ "step": 4536
+ },
+ {
+ "epoch": 63.0,
+ "eval_loss": 4.714163303375244,
+ "eval_runtime": 57.0269,
+ "eval_samples_per_second": 42.822,
+ "eval_steps_per_second": 1.35,
+ "step": 4536
+ },
+ {
+ "epoch": 63.01397379912664,
+ "grad_norm": 0.014277510344982147,
+ "learning_rate": 0.0006,
+ "loss": 4.3112640380859375,
+ "step": 4537
+ },
+ {
+ "epoch": 63.02794759825328,
+ "grad_norm": 0.015377935022115707,
+ "learning_rate": 0.0006,
+ "loss": 4.342782974243164,
+ "step": 4538
+ },
+ {
+ "epoch": 63.041921397379916,
+ "grad_norm": 0.015406244434416294,
+ "learning_rate": 0.0006,
+ "loss": 4.474056720733643,
+ "step": 4539
+ },
+ {
+ "epoch": 63.05589519650655,
+ "grad_norm": 0.018121499568223953,
+ "learning_rate": 0.0006,
+ "loss": 4.397963523864746,
+ "step": 4540
+ },
+ {
+ "epoch": 63.069868995633186,
+ "grad_norm": 0.022634636610746384,
+ "learning_rate": 0.0006,
+ "loss": 4.306631088256836,
+ "step": 4541
+ },
+ {
+ "epoch": 63.083842794759825,
+ "grad_norm": 0.026448970660567284,
+ "learning_rate": 0.0006,
+ "loss": 4.283175468444824,
+ "step": 4542
+ },
+ {
+ "epoch": 63.09781659388646,
+ "grad_norm": 0.024104537442326546,
+ "learning_rate": 0.0006,
+ "loss": 4.315826416015625,
+ "step": 4543
+ },
+ {
+ "epoch": 63.1117903930131,
+ "grad_norm": 0.019040480256080627,
+ "learning_rate": 0.0006,
+ "loss": 4.44057035446167,
+ "step": 4544
+ },
+ {
+ "epoch": 63.12576419213974,
+ "grad_norm": 0.02006000280380249,
+ "learning_rate": 0.0006,
+ "loss": 4.405603408813477,
+ "step": 4545
+ },
+ {
+ "epoch": 63.13973799126637,
+ "grad_norm": 0.019980071112513542,
+ "learning_rate": 0.0006,
+ "loss": 4.393726348876953,
+ "step": 4546
+ },
+ {
+ "epoch": 63.15371179039301,
+ "grad_norm": 0.01641865260899067,
+ "learning_rate": 0.0006,
+ "loss": 4.272546768188477,
+ "step": 4547
+ },
+ {
+ "epoch": 63.16768558951965,
+ "grad_norm": 0.01890353485941887,
+ "learning_rate": 0.0006,
+ "loss": 4.382566928863525,
+ "step": 4548
+ },
+ {
+ "epoch": 63.18165938864629,
+ "grad_norm": 0.018370937556028366,
+ "learning_rate": 0.0006,
+ "loss": 4.423953533172607,
+ "step": 4549
+ },
+ {
+ "epoch": 63.19563318777293,
+ "grad_norm": 0.014985294081270695,
+ "learning_rate": 0.0006,
+ "loss": 4.308627128601074,
+ "step": 4550
+ },
+ {
+ "epoch": 63.209606986899566,
+ "grad_norm": 0.017513327300548553,
+ "learning_rate": 0.0006,
+ "loss": 4.248819351196289,
+ "step": 4551
+ },
+ {
+ "epoch": 63.223580786026204,
+ "grad_norm": 0.019464161247015,
+ "learning_rate": 0.0006,
+ "loss": 4.359729766845703,
+ "step": 4552
+ },
+ {
+ "epoch": 63.237554585152836,
+ "grad_norm": 0.018678035587072372,
+ "learning_rate": 0.0006,
+ "loss": 4.3299665451049805,
+ "step": 4553
+ },
+ {
+ "epoch": 63.251528384279474,
+ "grad_norm": 0.017909523099660873,
+ "learning_rate": 0.0006,
+ "loss": 4.246147155761719,
+ "step": 4554
+ },
+ {
+ "epoch": 63.26550218340611,
+ "grad_norm": 0.01747293211519718,
+ "learning_rate": 0.0006,
+ "loss": 4.41165018081665,
+ "step": 4555
+ },
+ {
+ "epoch": 63.27947598253275,
+ "grad_norm": 0.018670853227376938,
+ "learning_rate": 0.0006,
+ "loss": 4.409883499145508,
+ "step": 4556
+ },
+ {
+ "epoch": 63.29344978165939,
+ "grad_norm": 0.0189529862254858,
+ "learning_rate": 0.0006,
+ "loss": 4.418654918670654,
+ "step": 4557
+ },
+ {
+ "epoch": 63.30742358078603,
+ "grad_norm": 0.018394414335489273,
+ "learning_rate": 0.0006,
+ "loss": 4.38355827331543,
+ "step": 4558
+ },
+ {
+ "epoch": 63.32139737991266,
+ "grad_norm": 0.018317820504307747,
+ "learning_rate": 0.0006,
+ "loss": 4.456843376159668,
+ "step": 4559
+ },
+ {
+ "epoch": 63.3353711790393,
+ "grad_norm": 0.019225597381591797,
+ "learning_rate": 0.0006,
+ "loss": 4.232264518737793,
+ "step": 4560
+ },
+ {
+ "epoch": 63.34934497816594,
+ "grad_norm": 0.016663681715726852,
+ "learning_rate": 0.0006,
+ "loss": 4.393517017364502,
+ "step": 4561
+ },
+ {
+ "epoch": 63.36331877729258,
+ "grad_norm": 0.014975365251302719,
+ "learning_rate": 0.0006,
+ "loss": 4.375183582305908,
+ "step": 4562
+ },
+ {
+ "epoch": 63.377292576419215,
+ "grad_norm": 0.01771732047200203,
+ "learning_rate": 0.0006,
+ "loss": 4.3073410987854,
+ "step": 4563
+ },
+ {
+ "epoch": 63.391266375545854,
+ "grad_norm": 0.01663549244403839,
+ "learning_rate": 0.0006,
+ "loss": 4.311003684997559,
+ "step": 4564
+ },
+ {
+ "epoch": 63.40524017467249,
+ "grad_norm": 0.016301177442073822,
+ "learning_rate": 0.0006,
+ "loss": 4.326807022094727,
+ "step": 4565
+ },
+ {
+ "epoch": 63.419213973799124,
+ "grad_norm": 0.016781188547611237,
+ "learning_rate": 0.0006,
+ "loss": 4.306562423706055,
+ "step": 4566
+ },
+ {
+ "epoch": 63.43318777292576,
+ "grad_norm": 0.01767745055258274,
+ "learning_rate": 0.0006,
+ "loss": 4.367082595825195,
+ "step": 4567
+ },
+ {
+ "epoch": 63.4471615720524,
+ "grad_norm": 0.016728565096855164,
+ "learning_rate": 0.0006,
+ "loss": 4.327400207519531,
+ "step": 4568
+ },
+ {
+ "epoch": 63.46113537117904,
+ "grad_norm": 0.016289403662085533,
+ "learning_rate": 0.0006,
+ "loss": 4.392712116241455,
+ "step": 4569
+ },
+ {
+ "epoch": 63.47510917030568,
+ "grad_norm": 0.016233716160058975,
+ "learning_rate": 0.0006,
+ "loss": 4.354372024536133,
+ "step": 4570
+ },
+ {
+ "epoch": 63.48908296943232,
+ "grad_norm": 0.016547974199056625,
+ "learning_rate": 0.0006,
+ "loss": 4.309050559997559,
+ "step": 4571
+ },
+ {
+ "epoch": 63.50305676855895,
+ "grad_norm": 0.018466509878635406,
+ "learning_rate": 0.0006,
+ "loss": 4.318345069885254,
+ "step": 4572
+ },
+ {
+ "epoch": 63.51703056768559,
+ "grad_norm": 0.017685720697045326,
+ "learning_rate": 0.0006,
+ "loss": 4.399726867675781,
+ "step": 4573
+ },
+ {
+ "epoch": 63.531004366812226,
+ "grad_norm": 0.015801234170794487,
+ "learning_rate": 0.0006,
+ "loss": 4.328495502471924,
+ "step": 4574
+ },
+ {
+ "epoch": 63.544978165938865,
+ "grad_norm": 0.0188873540610075,
+ "learning_rate": 0.0006,
+ "loss": 4.253244400024414,
+ "step": 4575
+ },
+ {
+ "epoch": 63.5589519650655,
+ "grad_norm": 0.01868271268904209,
+ "learning_rate": 0.0006,
+ "loss": 4.463746070861816,
+ "step": 4576
+ },
+ {
+ "epoch": 63.57292576419214,
+ "grad_norm": 0.01844949834048748,
+ "learning_rate": 0.0006,
+ "loss": 4.400300025939941,
+ "step": 4577
+ },
+ {
+ "epoch": 63.58689956331878,
+ "grad_norm": 0.015256480313837528,
+ "learning_rate": 0.0006,
+ "loss": 4.445844650268555,
+ "step": 4578
+ },
+ {
+ "epoch": 63.60087336244541,
+ "grad_norm": 0.015305760316550732,
+ "learning_rate": 0.0006,
+ "loss": 4.317327499389648,
+ "step": 4579
+ },
+ {
+ "epoch": 63.61484716157205,
+ "grad_norm": 0.015200769528746605,
+ "learning_rate": 0.0006,
+ "loss": 4.25112247467041,
+ "step": 4580
+ },
+ {
+ "epoch": 63.62882096069869,
+ "grad_norm": 0.015919538214802742,
+ "learning_rate": 0.0006,
+ "loss": 4.277937889099121,
+ "step": 4581
+ },
+ {
+ "epoch": 63.64279475982533,
+ "grad_norm": 0.01572665013372898,
+ "learning_rate": 0.0006,
+ "loss": 4.302237510681152,
+ "step": 4582
+ },
+ {
+ "epoch": 63.65676855895197,
+ "grad_norm": 0.016428731381893158,
+ "learning_rate": 0.0006,
+ "loss": 4.36762809753418,
+ "step": 4583
+ },
+ {
+ "epoch": 63.670742358078606,
+ "grad_norm": 0.01706082746386528,
+ "learning_rate": 0.0006,
+ "loss": 4.236526966094971,
+ "step": 4584
+ },
+ {
+ "epoch": 63.68471615720524,
+ "grad_norm": 0.017850136384367943,
+ "learning_rate": 0.0006,
+ "loss": 4.264129638671875,
+ "step": 4585
+ },
+ {
+ "epoch": 63.698689956331876,
+ "grad_norm": 0.016079438850283623,
+ "learning_rate": 0.0006,
+ "loss": 4.449799060821533,
+ "step": 4586
+ },
+ {
+ "epoch": 63.712663755458514,
+ "grad_norm": 0.014996383339166641,
+ "learning_rate": 0.0006,
+ "loss": 4.404574394226074,
+ "step": 4587
+ },
+ {
+ "epoch": 63.72663755458515,
+ "grad_norm": 0.015910431742668152,
+ "learning_rate": 0.0006,
+ "loss": 4.348101615905762,
+ "step": 4588
+ },
+ {
+ "epoch": 63.74061135371179,
+ "grad_norm": 0.01630021072924137,
+ "learning_rate": 0.0006,
+ "loss": 4.3193817138671875,
+ "step": 4589
+ },
+ {
+ "epoch": 63.75458515283843,
+ "grad_norm": 0.014571773819625378,
+ "learning_rate": 0.0006,
+ "loss": 4.374402046203613,
+ "step": 4590
+ },
+ {
+ "epoch": 63.76855895196506,
+ "grad_norm": 0.01662345975637436,
+ "learning_rate": 0.0006,
+ "loss": 4.329075813293457,
+ "step": 4591
+ },
+ {
+ "epoch": 63.7825327510917,
+ "grad_norm": 0.016604198142886162,
+ "learning_rate": 0.0006,
+ "loss": 4.436126232147217,
+ "step": 4592
+ },
+ {
+ "epoch": 63.79650655021834,
+ "grad_norm": 0.016578884795308113,
+ "learning_rate": 0.0006,
+ "loss": 4.361468315124512,
+ "step": 4593
+ },
+ {
+ "epoch": 63.81048034934498,
+ "grad_norm": 0.016339825466275215,
+ "learning_rate": 0.0006,
+ "loss": 4.449892044067383,
+ "step": 4594
+ },
+ {
+ "epoch": 63.82445414847162,
+ "grad_norm": 0.016805455088615417,
+ "learning_rate": 0.0006,
+ "loss": 4.415866851806641,
+ "step": 4595
+ },
+ {
+ "epoch": 63.838427947598255,
+ "grad_norm": 0.017366213724017143,
+ "learning_rate": 0.0006,
+ "loss": 4.361730575561523,
+ "step": 4596
+ },
+ {
+ "epoch": 63.852401746724894,
+ "grad_norm": 0.015599003992974758,
+ "learning_rate": 0.0006,
+ "loss": 4.363637924194336,
+ "step": 4597
+ },
+ {
+ "epoch": 63.866375545851525,
+ "grad_norm": 0.015674632042646408,
+ "learning_rate": 0.0006,
+ "loss": 4.332045078277588,
+ "step": 4598
+ },
+ {
+ "epoch": 63.880349344978164,
+ "grad_norm": 0.015350443311035633,
+ "learning_rate": 0.0006,
+ "loss": 4.320063591003418,
+ "step": 4599
+ },
+ {
+ "epoch": 63.8943231441048,
+ "grad_norm": 0.015388146974146366,
+ "learning_rate": 0.0006,
+ "loss": 4.398199081420898,
+ "step": 4600
+ },
+ {
+ "epoch": 63.90829694323144,
+ "grad_norm": 0.01450140681117773,
+ "learning_rate": 0.0006,
+ "loss": 4.244076728820801,
+ "step": 4601
+ },
+ {
+ "epoch": 63.92227074235808,
+ "grad_norm": 0.01637491025030613,
+ "learning_rate": 0.0006,
+ "loss": 4.213549613952637,
+ "step": 4602
+ },
+ {
+ "epoch": 63.93624454148472,
+ "grad_norm": 0.0159814041107893,
+ "learning_rate": 0.0006,
+ "loss": 4.421775817871094,
+ "step": 4603
+ },
+ {
+ "epoch": 63.95021834061135,
+ "grad_norm": 0.017386827617883682,
+ "learning_rate": 0.0006,
+ "loss": 4.340023994445801,
+ "step": 4604
+ },
+ {
+ "epoch": 63.96419213973799,
+ "grad_norm": 0.019679656252264977,
+ "learning_rate": 0.0006,
+ "loss": 4.352534294128418,
+ "step": 4605
+ },
+ {
+ "epoch": 63.97816593886463,
+ "grad_norm": 0.022391516715288162,
+ "learning_rate": 0.0006,
+ "loss": 4.3538312911987305,
+ "step": 4606
+ },
+ {
+ "epoch": 63.992139737991266,
+ "grad_norm": 0.023281006142497063,
+ "learning_rate": 0.0006,
+ "loss": 4.432089328765869,
+ "step": 4607
+ },
+ {
+ "epoch": 64.0,
+ "grad_norm": 0.019916288554668427,
+ "learning_rate": 0.0006,
+ "loss": 4.466937065124512,
+ "step": 4608
+ },
+ {
+ "epoch": 64.0,
+ "eval_loss": 4.659008502960205,
+ "eval_runtime": 56.0019,
+ "eval_samples_per_second": 43.606,
+ "eval_steps_per_second": 1.375,
+ "step": 4608
+ },
+ {
+ "epoch": 64.01397379912663,
+ "grad_norm": 0.017351187765598297,
+ "learning_rate": 0.0006,
+ "loss": 4.294129371643066,
+ "step": 4609
+ },
+ {
+ "epoch": 64.02794759825328,
+ "grad_norm": 0.02405865117907524,
+ "learning_rate": 0.0006,
+ "loss": 4.266717910766602,
+ "step": 4610
+ },
+ {
+ "epoch": 64.04192139737991,
+ "grad_norm": 0.022158386185765266,
+ "learning_rate": 0.0006,
+ "loss": 4.395932197570801,
+ "step": 4611
+ },
+ {
+ "epoch": 64.05589519650655,
+ "grad_norm": 0.019238505512475967,
+ "learning_rate": 0.0006,
+ "loss": 4.266704082489014,
+ "step": 4612
+ },
+ {
+ "epoch": 64.06986899563319,
+ "grad_norm": 0.021670425310730934,
+ "learning_rate": 0.0006,
+ "loss": 4.3160552978515625,
+ "step": 4613
+ },
+ {
+ "epoch": 64.08384279475983,
+ "grad_norm": 0.02076641097664833,
+ "learning_rate": 0.0006,
+ "loss": 4.415626525878906,
+ "step": 4614
+ },
+ {
+ "epoch": 64.09781659388646,
+ "grad_norm": 0.018223078921437263,
+ "learning_rate": 0.0006,
+ "loss": 4.292387008666992,
+ "step": 4615
+ },
+ {
+ "epoch": 64.1117903930131,
+ "grad_norm": 0.017401937395334244,
+ "learning_rate": 0.0006,
+ "loss": 4.424524307250977,
+ "step": 4616
+ },
+ {
+ "epoch": 64.12576419213974,
+ "grad_norm": 0.01663743332028389,
+ "learning_rate": 0.0006,
+ "loss": 4.350032806396484,
+ "step": 4617
+ },
+ {
+ "epoch": 64.13973799126637,
+ "grad_norm": 0.018716629594564438,
+ "learning_rate": 0.0006,
+ "loss": 4.309157371520996,
+ "step": 4618
+ },
+ {
+ "epoch": 64.15371179039302,
+ "grad_norm": 0.019569825381040573,
+ "learning_rate": 0.0006,
+ "loss": 4.358292579650879,
+ "step": 4619
+ },
+ {
+ "epoch": 64.16768558951965,
+ "grad_norm": 0.018994422629475594,
+ "learning_rate": 0.0006,
+ "loss": 4.3338823318481445,
+ "step": 4620
+ },
+ {
+ "epoch": 64.18165938864628,
+ "grad_norm": 0.017690688371658325,
+ "learning_rate": 0.0006,
+ "loss": 4.414468765258789,
+ "step": 4621
+ },
+ {
+ "epoch": 64.19563318777293,
+ "grad_norm": 0.0170609001070261,
+ "learning_rate": 0.0006,
+ "loss": 4.4337568283081055,
+ "step": 4622
+ },
+ {
+ "epoch": 64.20960698689956,
+ "grad_norm": 0.016984131187200546,
+ "learning_rate": 0.0006,
+ "loss": 4.304839134216309,
+ "step": 4623
+ },
+ {
+ "epoch": 64.2235807860262,
+ "grad_norm": 0.01619867794215679,
+ "learning_rate": 0.0006,
+ "loss": 4.389623165130615,
+ "step": 4624
+ },
+ {
+ "epoch": 64.23755458515284,
+ "grad_norm": 0.015428110957145691,
+ "learning_rate": 0.0006,
+ "loss": 4.277515888214111,
+ "step": 4625
+ },
+ {
+ "epoch": 64.25152838427948,
+ "grad_norm": 0.015909532085061073,
+ "learning_rate": 0.0006,
+ "loss": 4.266029357910156,
+ "step": 4626
+ },
+ {
+ "epoch": 64.26550218340611,
+ "grad_norm": 0.014296560548245907,
+ "learning_rate": 0.0006,
+ "loss": 4.286067485809326,
+ "step": 4627
+ },
+ {
+ "epoch": 64.27947598253274,
+ "grad_norm": 0.014065971598029137,
+ "learning_rate": 0.0006,
+ "loss": 4.353528022766113,
+ "step": 4628
+ },
+ {
+ "epoch": 64.29344978165939,
+ "grad_norm": 0.014386293478310108,
+ "learning_rate": 0.0006,
+ "loss": 4.318715572357178,
+ "step": 4629
+ },
+ {
+ "epoch": 64.30742358078602,
+ "grad_norm": 0.014001123607158661,
+ "learning_rate": 0.0006,
+ "loss": 4.362497329711914,
+ "step": 4630
+ },
+ {
+ "epoch": 64.32139737991267,
+ "grad_norm": 0.013408979400992393,
+ "learning_rate": 0.0006,
+ "loss": 4.417397499084473,
+ "step": 4631
+ },
+ {
+ "epoch": 64.3353711790393,
+ "grad_norm": 0.014960093423724174,
+ "learning_rate": 0.0006,
+ "loss": 4.359710693359375,
+ "step": 4632
+ },
+ {
+ "epoch": 64.34934497816595,
+ "grad_norm": 0.014658777043223381,
+ "learning_rate": 0.0006,
+ "loss": 4.318822383880615,
+ "step": 4633
+ },
+ {
+ "epoch": 64.36331877729258,
+ "grad_norm": 0.013766744174063206,
+ "learning_rate": 0.0006,
+ "loss": 4.265679836273193,
+ "step": 4634
+ },
+ {
+ "epoch": 64.37729257641921,
+ "grad_norm": 0.013821255415678024,
+ "learning_rate": 0.0006,
+ "loss": 4.3452911376953125,
+ "step": 4635
+ },
+ {
+ "epoch": 64.39126637554585,
+ "grad_norm": 0.013541066087782383,
+ "learning_rate": 0.0006,
+ "loss": 4.429254531860352,
+ "step": 4636
+ },
+ {
+ "epoch": 64.40524017467249,
+ "grad_norm": 0.013985645025968552,
+ "learning_rate": 0.0006,
+ "loss": 4.304452896118164,
+ "step": 4637
+ },
+ {
+ "epoch": 64.41921397379913,
+ "grad_norm": 0.015079808421432972,
+ "learning_rate": 0.0006,
+ "loss": 4.296828746795654,
+ "step": 4638
+ },
+ {
+ "epoch": 64.43318777292576,
+ "grad_norm": 0.014896944165229797,
+ "learning_rate": 0.0006,
+ "loss": 4.299156665802002,
+ "step": 4639
+ },
+ {
+ "epoch": 64.44716157205241,
+ "grad_norm": 0.013770041987299919,
+ "learning_rate": 0.0006,
+ "loss": 4.351439952850342,
+ "step": 4640
+ },
+ {
+ "epoch": 64.46113537117904,
+ "grad_norm": 0.01531178504228592,
+ "learning_rate": 0.0006,
+ "loss": 4.373715877532959,
+ "step": 4641
+ },
+ {
+ "epoch": 64.47510917030567,
+ "grad_norm": 0.014024450443685055,
+ "learning_rate": 0.0006,
+ "loss": 4.368053436279297,
+ "step": 4642
+ },
+ {
+ "epoch": 64.48908296943232,
+ "grad_norm": 0.013864024542272091,
+ "learning_rate": 0.0006,
+ "loss": 4.378564834594727,
+ "step": 4643
+ },
+ {
+ "epoch": 64.50305676855895,
+ "grad_norm": 0.013925445266067982,
+ "learning_rate": 0.0006,
+ "loss": 4.261663436889648,
+ "step": 4644
+ },
+ {
+ "epoch": 64.5170305676856,
+ "grad_norm": 0.015011338517069817,
+ "learning_rate": 0.0006,
+ "loss": 4.337285995483398,
+ "step": 4645
+ },
+ {
+ "epoch": 64.53100436681223,
+ "grad_norm": 0.016941986978054047,
+ "learning_rate": 0.0006,
+ "loss": 4.336877822875977,
+ "step": 4646
+ },
+ {
+ "epoch": 64.54497816593886,
+ "grad_norm": 0.01719909906387329,
+ "learning_rate": 0.0006,
+ "loss": 4.149813175201416,
+ "step": 4647
+ },
+ {
+ "epoch": 64.5589519650655,
+ "grad_norm": 0.01608402095735073,
+ "learning_rate": 0.0006,
+ "loss": 4.191280364990234,
+ "step": 4648
+ },
+ {
+ "epoch": 64.57292576419214,
+ "grad_norm": 0.015284820459783077,
+ "learning_rate": 0.0006,
+ "loss": 4.35953426361084,
+ "step": 4649
+ },
+ {
+ "epoch": 64.58689956331878,
+ "grad_norm": 0.014855682849884033,
+ "learning_rate": 0.0006,
+ "loss": 4.308401107788086,
+ "step": 4650
+ },
+ {
+ "epoch": 64.60087336244541,
+ "grad_norm": 0.017553111538290977,
+ "learning_rate": 0.0006,
+ "loss": 4.279720306396484,
+ "step": 4651
+ },
+ {
+ "epoch": 64.61484716157206,
+ "grad_norm": 0.019506998360157013,
+ "learning_rate": 0.0006,
+ "loss": 4.319524765014648,
+ "step": 4652
+ },
+ {
+ "epoch": 64.62882096069869,
+ "grad_norm": 0.020293425768613815,
+ "learning_rate": 0.0006,
+ "loss": 4.340523719787598,
+ "step": 4653
+ },
+ {
+ "epoch": 64.64279475982532,
+ "grad_norm": 0.01664556749165058,
+ "learning_rate": 0.0006,
+ "loss": 4.296785831451416,
+ "step": 4654
+ },
+ {
+ "epoch": 64.65676855895197,
+ "grad_norm": 0.012592585757374763,
+ "learning_rate": 0.0006,
+ "loss": 4.464595317840576,
+ "step": 4655
+ },
+ {
+ "epoch": 64.6707423580786,
+ "grad_norm": 0.01549589168280363,
+ "learning_rate": 0.0006,
+ "loss": 4.275053024291992,
+ "step": 4656
+ },
+ {
+ "epoch": 64.68471615720524,
+ "grad_norm": 0.01705172471702099,
+ "learning_rate": 0.0006,
+ "loss": 4.388431549072266,
+ "step": 4657
+ },
+ {
+ "epoch": 64.69868995633188,
+ "grad_norm": 0.015438582748174667,
+ "learning_rate": 0.0006,
+ "loss": 4.312070846557617,
+ "step": 4658
+ },
+ {
+ "epoch": 64.71266375545852,
+ "grad_norm": 0.014448083005845547,
+ "learning_rate": 0.0006,
+ "loss": 4.237599849700928,
+ "step": 4659
+ },
+ {
+ "epoch": 64.72663755458515,
+ "grad_norm": 0.015267071314156055,
+ "learning_rate": 0.0006,
+ "loss": 4.29893684387207,
+ "step": 4660
+ },
+ {
+ "epoch": 64.74061135371178,
+ "grad_norm": 0.016615847125649452,
+ "learning_rate": 0.0006,
+ "loss": 4.432273864746094,
+ "step": 4661
+ },
+ {
+ "epoch": 64.75458515283843,
+ "grad_norm": 0.016584362834692,
+ "learning_rate": 0.0006,
+ "loss": 4.202266216278076,
+ "step": 4662
+ },
+ {
+ "epoch": 64.76855895196506,
+ "grad_norm": 0.018313994631171227,
+ "learning_rate": 0.0006,
+ "loss": 4.367966651916504,
+ "step": 4663
+ },
+ {
+ "epoch": 64.78253275109171,
+ "grad_norm": 0.017946481704711914,
+ "learning_rate": 0.0006,
+ "loss": 4.205172538757324,
+ "step": 4664
+ },
+ {
+ "epoch": 64.79650655021834,
+ "grad_norm": 0.017744354903697968,
+ "learning_rate": 0.0006,
+ "loss": 4.295661926269531,
+ "step": 4665
+ },
+ {
+ "epoch": 64.81048034934499,
+ "grad_norm": 0.016103824600577354,
+ "learning_rate": 0.0006,
+ "loss": 4.276543617248535,
+ "step": 4666
+ },
+ {
+ "epoch": 64.82445414847162,
+ "grad_norm": 0.014590012840926647,
+ "learning_rate": 0.0006,
+ "loss": 4.414247035980225,
+ "step": 4667
+ },
+ {
+ "epoch": 64.83842794759825,
+ "grad_norm": 0.01782386749982834,
+ "learning_rate": 0.0006,
+ "loss": 4.203220367431641,
+ "step": 4668
+ },
+ {
+ "epoch": 64.8524017467249,
+ "grad_norm": 0.022122733294963837,
+ "learning_rate": 0.0006,
+ "loss": 4.38078498840332,
+ "step": 4669
+ },
+ {
+ "epoch": 64.86637554585153,
+ "grad_norm": 0.0250293780118227,
+ "learning_rate": 0.0006,
+ "loss": 4.3310136795043945,
+ "step": 4670
+ },
+ {
+ "epoch": 64.88034934497817,
+ "grad_norm": 0.022438062354922295,
+ "learning_rate": 0.0006,
+ "loss": 4.428631782531738,
+ "step": 4671
+ },
+ {
+ "epoch": 64.8943231441048,
+ "grad_norm": 0.018518192693591118,
+ "learning_rate": 0.0006,
+ "loss": 4.428250312805176,
+ "step": 4672
+ },
+ {
+ "epoch": 64.90829694323143,
+ "grad_norm": 0.018105225637555122,
+ "learning_rate": 0.0006,
+ "loss": 4.336209297180176,
+ "step": 4673
+ },
+ {
+ "epoch": 64.92227074235808,
+ "grad_norm": 0.017564399167895317,
+ "learning_rate": 0.0006,
+ "loss": 4.416011810302734,
+ "step": 4674
+ },
+ {
+ "epoch": 64.93624454148471,
+ "grad_norm": 0.018747031688690186,
+ "learning_rate": 0.0006,
+ "loss": 4.419938087463379,
+ "step": 4675
+ },
+ {
+ "epoch": 64.95021834061136,
+ "grad_norm": 0.018686484545469284,
+ "learning_rate": 0.0006,
+ "loss": 4.347185134887695,
+ "step": 4676
+ },
+ {
+ "epoch": 64.96419213973799,
+ "grad_norm": 0.019329898059368134,
+ "learning_rate": 0.0006,
+ "loss": 4.343749046325684,
+ "step": 4677
+ },
+ {
+ "epoch": 64.97816593886463,
+ "grad_norm": 0.018517782911658287,
+ "learning_rate": 0.0006,
+ "loss": 4.264604568481445,
+ "step": 4678
+ },
+ {
+ "epoch": 64.99213973799127,
+ "grad_norm": 0.016579125076532364,
+ "learning_rate": 0.0006,
+ "loss": 4.303423881530762,
+ "step": 4679
+ },
+ {
+ "epoch": 65.0,
+ "grad_norm": 0.01617567613720894,
+ "learning_rate": 0.0006,
+ "loss": 4.326556205749512,
+ "step": 4680
+ },
+ {
+ "epoch": 65.0,
+ "eval_loss": 4.655026912689209,
+ "eval_runtime": 57.1299,
+ "eval_samples_per_second": 42.745,
+ "eval_steps_per_second": 1.348,
+ "step": 4680
+ },
+ {
+ "epoch": 65.01397379912663,
+ "grad_norm": 0.01607249490916729,
+ "learning_rate": 0.0006,
+ "loss": 4.276607036590576,
+ "step": 4681
+ },
+ {
+ "epoch": 65.02794759825328,
+ "grad_norm": 0.016779497265815735,
+ "learning_rate": 0.0006,
+ "loss": 4.412464618682861,
+ "step": 4682
+ },
+ {
+ "epoch": 65.04192139737991,
+ "grad_norm": 0.018161555752158165,
+ "learning_rate": 0.0006,
+ "loss": 4.199329376220703,
+ "step": 4683
+ },
+ {
+ "epoch": 65.05589519650655,
+ "grad_norm": 0.017508309334516525,
+ "learning_rate": 0.0006,
+ "loss": 4.217567443847656,
+ "step": 4684
+ },
+ {
+ "epoch": 65.06986899563319,
+ "grad_norm": 0.015716979280114174,
+ "learning_rate": 0.0006,
+ "loss": 4.322023391723633,
+ "step": 4685
+ },
+ {
+ "epoch": 65.08384279475983,
+ "grad_norm": 0.015113635919988155,
+ "learning_rate": 0.0006,
+ "loss": 4.333515167236328,
+ "step": 4686
+ },
+ {
+ "epoch": 65.09781659388646,
+ "grad_norm": 0.01437560748308897,
+ "learning_rate": 0.0006,
+ "loss": 4.295609474182129,
+ "step": 4687
+ },
+ {
+ "epoch": 65.1117903930131,
+ "grad_norm": 0.013074862770736217,
+ "learning_rate": 0.0006,
+ "loss": 4.314120769500732,
+ "step": 4688
+ },
+ {
+ "epoch": 65.12576419213974,
+ "grad_norm": 0.012908284552395344,
+ "learning_rate": 0.0006,
+ "loss": 4.216890811920166,
+ "step": 4689
+ },
+ {
+ "epoch": 65.13973799126637,
+ "grad_norm": 0.013001754879951477,
+ "learning_rate": 0.0006,
+ "loss": 4.287717342376709,
+ "step": 4690
+ },
+ {
+ "epoch": 65.15371179039302,
+ "grad_norm": 0.013694263994693756,
+ "learning_rate": 0.0006,
+ "loss": 4.333837985992432,
+ "step": 4691
+ },
+ {
+ "epoch": 65.16768558951965,
+ "grad_norm": 0.014461833983659744,
+ "learning_rate": 0.0006,
+ "loss": 4.323545455932617,
+ "step": 4692
+ },
+ {
+ "epoch": 65.18165938864628,
+ "grad_norm": 0.013670675456523895,
+ "learning_rate": 0.0006,
+ "loss": 4.3313493728637695,
+ "step": 4693
+ },
+ {
+ "epoch": 65.19563318777293,
+ "grad_norm": 0.013094350695610046,
+ "learning_rate": 0.0006,
+ "loss": 4.23858118057251,
+ "step": 4694
+ },
+ {
+ "epoch": 65.20960698689956,
+ "grad_norm": 0.013996944762766361,
+ "learning_rate": 0.0006,
+ "loss": 4.428705215454102,
+ "step": 4695
+ },
+ {
+ "epoch": 65.2235807860262,
+ "grad_norm": 0.013295048847794533,
+ "learning_rate": 0.0006,
+ "loss": 4.238950252532959,
+ "step": 4696
+ },
+ {
+ "epoch": 65.23755458515284,
+ "grad_norm": 0.013380616903305054,
+ "learning_rate": 0.0006,
+ "loss": 4.273185729980469,
+ "step": 4697
+ },
+ {
+ "epoch": 65.25152838427948,
+ "grad_norm": 0.015062890015542507,
+ "learning_rate": 0.0006,
+ "loss": 4.29472017288208,
+ "step": 4698
+ },
+ {
+ "epoch": 65.26550218340611,
+ "grad_norm": 0.017317263409495354,
+ "learning_rate": 0.0006,
+ "loss": 4.337954998016357,
+ "step": 4699
+ },
+ {
+ "epoch": 65.27947598253274,
+ "grad_norm": 0.017698992043733597,
+ "learning_rate": 0.0006,
+ "loss": 4.336437225341797,
+ "step": 4700
+ },
+ {
+ "epoch": 65.29344978165939,
+ "grad_norm": 0.01628575474023819,
+ "learning_rate": 0.0006,
+ "loss": 4.302372455596924,
+ "step": 4701
+ },
+ {
+ "epoch": 65.30742358078602,
+ "grad_norm": 0.014531854540109634,
+ "learning_rate": 0.0006,
+ "loss": 4.256450653076172,
+ "step": 4702
+ },
+ {
+ "epoch": 65.32139737991267,
+ "grad_norm": 0.014549105428159237,
+ "learning_rate": 0.0006,
+ "loss": 4.302240371704102,
+ "step": 4703
+ },
+ {
+ "epoch": 65.3353711790393,
+ "grad_norm": 0.014881663955748081,
+ "learning_rate": 0.0006,
+ "loss": 4.3316121101379395,
+ "step": 4704
+ },
+ {
+ "epoch": 65.34934497816595,
+ "grad_norm": 0.014513437636196613,
+ "learning_rate": 0.0006,
+ "loss": 4.441212177276611,
+ "step": 4705
+ },
+ {
+ "epoch": 65.36331877729258,
+ "grad_norm": 0.014122087508440018,
+ "learning_rate": 0.0006,
+ "loss": 4.343296527862549,
+ "step": 4706
+ },
+ {
+ "epoch": 65.37729257641921,
+ "grad_norm": 0.014817348681390285,
+ "learning_rate": 0.0006,
+ "loss": 4.195674896240234,
+ "step": 4707
+ },
+ {
+ "epoch": 65.39126637554585,
+ "grad_norm": 0.01787056028842926,
+ "learning_rate": 0.0006,
+ "loss": 4.3507914543151855,
+ "step": 4708
+ },
+ {
+ "epoch": 65.40524017467249,
+ "grad_norm": 0.015295136719942093,
+ "learning_rate": 0.0006,
+ "loss": 4.234372138977051,
+ "step": 4709
+ },
+ {
+ "epoch": 65.41921397379913,
+ "grad_norm": 0.014324234798550606,
+ "learning_rate": 0.0006,
+ "loss": 4.2907867431640625,
+ "step": 4710
+ },
+ {
+ "epoch": 65.43318777292576,
+ "grad_norm": 0.016090670600533485,
+ "learning_rate": 0.0006,
+ "loss": 4.177244186401367,
+ "step": 4711
+ },
+ {
+ "epoch": 65.44716157205241,
+ "grad_norm": 0.0174604132771492,
+ "learning_rate": 0.0006,
+ "loss": 4.335577964782715,
+ "step": 4712
+ },
+ {
+ "epoch": 65.46113537117904,
+ "grad_norm": 0.016731899231672287,
+ "learning_rate": 0.0006,
+ "loss": 4.30197811126709,
+ "step": 4713
+ },
+ {
+ "epoch": 65.47510917030567,
+ "grad_norm": 0.017423735931515694,
+ "learning_rate": 0.0006,
+ "loss": 4.3528852462768555,
+ "step": 4714
+ },
+ {
+ "epoch": 65.48908296943232,
+ "grad_norm": 0.01780976541340351,
+ "learning_rate": 0.0006,
+ "loss": 4.185036659240723,
+ "step": 4715
+ },
+ {
+ "epoch": 65.50305676855895,
+ "grad_norm": 0.019977200776338577,
+ "learning_rate": 0.0006,
+ "loss": 4.328766345977783,
+ "step": 4716
+ },
+ {
+ "epoch": 65.5170305676856,
+ "grad_norm": 0.020284287631511688,
+ "learning_rate": 0.0006,
+ "loss": 4.285494804382324,
+ "step": 4717
+ },
+ {
+ "epoch": 65.53100436681223,
+ "grad_norm": 0.019168002530932426,
+ "learning_rate": 0.0006,
+ "loss": 4.328560829162598,
+ "step": 4718
+ },
+ {
+ "epoch": 65.54497816593886,
+ "grad_norm": 0.019217776134610176,
+ "learning_rate": 0.0006,
+ "loss": 4.189998626708984,
+ "step": 4719
+ },
+ {
+ "epoch": 65.5589519650655,
+ "grad_norm": 0.018960488960146904,
+ "learning_rate": 0.0006,
+ "loss": 4.4995012283325195,
+ "step": 4720
+ },
+ {
+ "epoch": 65.57292576419214,
+ "grad_norm": 0.01941288262605667,
+ "learning_rate": 0.0006,
+ "loss": 4.26570987701416,
+ "step": 4721
+ },
+ {
+ "epoch": 65.58689956331878,
+ "grad_norm": 0.020263448357582092,
+ "learning_rate": 0.0006,
+ "loss": 4.411645889282227,
+ "step": 4722
+ },
+ {
+ "epoch": 65.60087336244541,
+ "grad_norm": 0.02144799567759037,
+ "learning_rate": 0.0006,
+ "loss": 4.163675308227539,
+ "step": 4723
+ },
+ {
+ "epoch": 65.61484716157206,
+ "grad_norm": 0.019790446385741234,
+ "learning_rate": 0.0006,
+ "loss": 4.2361063957214355,
+ "step": 4724
+ },
+ {
+ "epoch": 65.62882096069869,
+ "grad_norm": 0.01829805225133896,
+ "learning_rate": 0.0006,
+ "loss": 4.449810028076172,
+ "step": 4725
+ },
+ {
+ "epoch": 65.64279475982532,
+ "grad_norm": 0.01571105420589447,
+ "learning_rate": 0.0006,
+ "loss": 4.281406879425049,
+ "step": 4726
+ },
+ {
+ "epoch": 65.65676855895197,
+ "grad_norm": 0.016477147117257118,
+ "learning_rate": 0.0006,
+ "loss": 4.293408393859863,
+ "step": 4727
+ },
+ {
+ "epoch": 65.6707423580786,
+ "grad_norm": 0.016766032204031944,
+ "learning_rate": 0.0006,
+ "loss": 4.378252983093262,
+ "step": 4728
+ },
+ {
+ "epoch": 65.68471615720524,
+ "grad_norm": 0.0159933939576149,
+ "learning_rate": 0.0006,
+ "loss": 4.3065667152404785,
+ "step": 4729
+ },
+ {
+ "epoch": 65.69868995633188,
+ "grad_norm": 0.01658209227025509,
+ "learning_rate": 0.0006,
+ "loss": 4.174026966094971,
+ "step": 4730
+ },
+ {
+ "epoch": 65.71266375545852,
+ "grad_norm": 0.015081380493938923,
+ "learning_rate": 0.0006,
+ "loss": 4.360222339630127,
+ "step": 4731
+ },
+ {
+ "epoch": 65.72663755458515,
+ "grad_norm": 0.015110890381038189,
+ "learning_rate": 0.0006,
+ "loss": 4.218810558319092,
+ "step": 4732
+ },
+ {
+ "epoch": 65.74061135371178,
+ "grad_norm": 0.015300953760743141,
+ "learning_rate": 0.0006,
+ "loss": 4.331979274749756,
+ "step": 4733
+ },
+ {
+ "epoch": 65.75458515283843,
+ "grad_norm": 0.01715504191815853,
+ "learning_rate": 0.0006,
+ "loss": 4.349112033843994,
+ "step": 4734
+ },
+ {
+ "epoch": 65.76855895196506,
+ "grad_norm": 0.015870366245508194,
+ "learning_rate": 0.0006,
+ "loss": 4.329850196838379,
+ "step": 4735
+ },
+ {
+ "epoch": 65.78253275109171,
+ "grad_norm": 0.01541937980800867,
+ "learning_rate": 0.0006,
+ "loss": 4.3400163650512695,
+ "step": 4736
+ },
+ {
+ "epoch": 65.79650655021834,
+ "grad_norm": 0.01738942787051201,
+ "learning_rate": 0.0006,
+ "loss": 4.220798492431641,
+ "step": 4737
+ },
+ {
+ "epoch": 65.81048034934499,
+ "grad_norm": 0.01685990020632744,
+ "learning_rate": 0.0006,
+ "loss": 4.177238464355469,
+ "step": 4738
+ },
+ {
+ "epoch": 65.82445414847162,
+ "grad_norm": 0.016183465719223022,
+ "learning_rate": 0.0006,
+ "loss": 4.327216625213623,
+ "step": 4739
+ },
+ {
+ "epoch": 65.83842794759825,
+ "grad_norm": 0.01677694357931614,
+ "learning_rate": 0.0006,
+ "loss": 4.27766227722168,
+ "step": 4740
+ },
+ {
+ "epoch": 65.8524017467249,
+ "grad_norm": 0.01645066775381565,
+ "learning_rate": 0.0006,
+ "loss": 4.288347244262695,
+ "step": 4741
+ },
+ {
+ "epoch": 65.86637554585153,
+ "grad_norm": 0.018034635111689568,
+ "learning_rate": 0.0006,
+ "loss": 4.292458534240723,
+ "step": 4742
+ },
+ {
+ "epoch": 65.88034934497817,
+ "grad_norm": 0.017973346635699272,
+ "learning_rate": 0.0006,
+ "loss": 4.3117265701293945,
+ "step": 4743
+ },
+ {
+ "epoch": 65.8943231441048,
+ "grad_norm": 0.016605669632554054,
+ "learning_rate": 0.0006,
+ "loss": 4.280615329742432,
+ "step": 4744
+ },
+ {
+ "epoch": 65.90829694323143,
+ "grad_norm": 0.016298074275255203,
+ "learning_rate": 0.0006,
+ "loss": 4.301933288574219,
+ "step": 4745
+ },
+ {
+ "epoch": 65.92227074235808,
+ "grad_norm": 0.015383216552436352,
+ "learning_rate": 0.0006,
+ "loss": 4.352383136749268,
+ "step": 4746
+ },
+ {
+ "epoch": 65.93624454148471,
+ "grad_norm": 0.016222182661294937,
+ "learning_rate": 0.0006,
+ "loss": 4.305725574493408,
+ "step": 4747
+ },
+ {
+ "epoch": 65.95021834061136,
+ "grad_norm": 0.01691329851746559,
+ "learning_rate": 0.0006,
+ "loss": 4.29502010345459,
+ "step": 4748
+ },
+ {
+ "epoch": 65.96419213973799,
+ "grad_norm": 0.017637677490711212,
+ "learning_rate": 0.0006,
+ "loss": 4.3133039474487305,
+ "step": 4749
+ },
+ {
+ "epoch": 65.97816593886463,
+ "grad_norm": 0.021362047642469406,
+ "learning_rate": 0.0006,
+ "loss": 4.458834648132324,
+ "step": 4750
+ },
+ {
+ "epoch": 65.99213973799127,
+ "grad_norm": 0.023553011938929558,
+ "learning_rate": 0.0006,
+ "loss": 4.298727989196777,
+ "step": 4751
+ },
+ {
+ "epoch": 66.0,
+ "grad_norm": 0.02154950052499771,
+ "learning_rate": 0.0006,
+ "loss": 4.403392791748047,
+ "step": 4752
+ },
+ {
+ "epoch": 66.0,
+ "eval_loss": 4.724340438842773,
+ "eval_runtime": 56.9555,
+ "eval_samples_per_second": 42.876,
+ "eval_steps_per_second": 1.352,
+ "step": 4752
+ },
+ {
+ "epoch": 66.01397379912663,
+ "grad_norm": 0.01916784606873989,
+ "learning_rate": 0.0006,
+ "loss": 4.22931432723999,
+ "step": 4753
+ },
+ {
+ "epoch": 66.02794759825328,
+ "grad_norm": 0.017492054030299187,
+ "learning_rate": 0.0006,
+ "loss": 4.271522521972656,
+ "step": 4754
+ },
+ {
+ "epoch": 66.04192139737991,
+ "grad_norm": 0.018408095464110374,
+ "learning_rate": 0.0006,
+ "loss": 4.337705135345459,
+ "step": 4755
+ },
+ {
+ "epoch": 66.05589519650655,
+ "grad_norm": 0.019220108166337013,
+ "learning_rate": 0.0006,
+ "loss": 4.280340194702148,
+ "step": 4756
+ },
+ {
+ "epoch": 66.06986899563319,
+ "grad_norm": 0.01773735322058201,
+ "learning_rate": 0.0006,
+ "loss": 4.390201568603516,
+ "step": 4757
+ },
+ {
+ "epoch": 66.08384279475983,
+ "grad_norm": 0.01679125428199768,
+ "learning_rate": 0.0006,
+ "loss": 4.299282073974609,
+ "step": 4758
+ },
+ {
+ "epoch": 66.09781659388646,
+ "grad_norm": 0.01479059923440218,
+ "learning_rate": 0.0006,
+ "loss": 4.472630500793457,
+ "step": 4759
+ },
+ {
+ "epoch": 66.1117903930131,
+ "grad_norm": 0.015130104497075081,
+ "learning_rate": 0.0006,
+ "loss": 4.301194190979004,
+ "step": 4760
+ },
+ {
+ "epoch": 66.12576419213974,
+ "grad_norm": 0.0158979594707489,
+ "learning_rate": 0.0006,
+ "loss": 4.375751972198486,
+ "step": 4761
+ },
+ {
+ "epoch": 66.13973799126637,
+ "grad_norm": 0.014432408846914768,
+ "learning_rate": 0.0006,
+ "loss": 4.279788017272949,
+ "step": 4762
+ },
+ {
+ "epoch": 66.15371179039302,
+ "grad_norm": 0.013026784174144268,
+ "learning_rate": 0.0006,
+ "loss": 4.255035400390625,
+ "step": 4763
+ },
+ {
+ "epoch": 66.16768558951965,
+ "grad_norm": 0.014422653242945671,
+ "learning_rate": 0.0006,
+ "loss": 4.261832237243652,
+ "step": 4764
+ },
+ {
+ "epoch": 66.18165938864628,
+ "grad_norm": 0.014613019302487373,
+ "learning_rate": 0.0006,
+ "loss": 4.382184028625488,
+ "step": 4765
+ },
+ {
+ "epoch": 66.19563318777293,
+ "grad_norm": 0.014989894814789295,
+ "learning_rate": 0.0006,
+ "loss": 4.40528678894043,
+ "step": 4766
+ },
+ {
+ "epoch": 66.20960698689956,
+ "grad_norm": 0.018190350383520126,
+ "learning_rate": 0.0006,
+ "loss": 4.302414894104004,
+ "step": 4767
+ },
+ {
+ "epoch": 66.2235807860262,
+ "grad_norm": 0.020887890830636024,
+ "learning_rate": 0.0006,
+ "loss": 4.289920806884766,
+ "step": 4768
+ },
+ {
+ "epoch": 66.23755458515284,
+ "grad_norm": 0.02376876398921013,
+ "learning_rate": 0.0006,
+ "loss": 4.297855377197266,
+ "step": 4769
+ },
+ {
+ "epoch": 66.25152838427948,
+ "grad_norm": 0.02152269519865513,
+ "learning_rate": 0.0006,
+ "loss": 4.1972832679748535,
+ "step": 4770
+ },
+ {
+ "epoch": 66.26550218340611,
+ "grad_norm": 0.015572802163660526,
+ "learning_rate": 0.0006,
+ "loss": 4.307487487792969,
+ "step": 4771
+ },
+ {
+ "epoch": 66.27947598253274,
+ "grad_norm": 0.015228749252855778,
+ "learning_rate": 0.0006,
+ "loss": 4.103250503540039,
+ "step": 4772
+ },
+ {
+ "epoch": 66.29344978165939,
+ "grad_norm": 0.016925504431128502,
+ "learning_rate": 0.0006,
+ "loss": 4.247461318969727,
+ "step": 4773
+ },
+ {
+ "epoch": 66.30742358078602,
+ "grad_norm": 0.01613711006939411,
+ "learning_rate": 0.0006,
+ "loss": 4.4699273109436035,
+ "step": 4774
+ },
+ {
+ "epoch": 66.32139737991267,
+ "grad_norm": 0.018332140520215034,
+ "learning_rate": 0.0006,
+ "loss": 4.22969388961792,
+ "step": 4775
+ },
+ {
+ "epoch": 66.3353711790393,
+ "grad_norm": 0.019575074315071106,
+ "learning_rate": 0.0006,
+ "loss": 4.29397439956665,
+ "step": 4776
+ },
+ {
+ "epoch": 66.34934497816595,
+ "grad_norm": 0.018177419900894165,
+ "learning_rate": 0.0006,
+ "loss": 4.221899509429932,
+ "step": 4777
+ },
+ {
+ "epoch": 66.36331877729258,
+ "grad_norm": 0.014292889274656773,
+ "learning_rate": 0.0006,
+ "loss": 4.325457572937012,
+ "step": 4778
+ },
+ {
+ "epoch": 66.37729257641921,
+ "grad_norm": 0.015098796226084232,
+ "learning_rate": 0.0006,
+ "loss": 4.364837169647217,
+ "step": 4779
+ },
+ {
+ "epoch": 66.39126637554585,
+ "grad_norm": 0.013992635533213615,
+ "learning_rate": 0.0006,
+ "loss": 4.299082279205322,
+ "step": 4780
+ },
+ {
+ "epoch": 66.40524017467249,
+ "grad_norm": 0.01504511572420597,
+ "learning_rate": 0.0006,
+ "loss": 4.230877876281738,
+ "step": 4781
+ },
+ {
+ "epoch": 66.41921397379913,
+ "grad_norm": 0.014875639230012894,
+ "learning_rate": 0.0006,
+ "loss": 4.368405818939209,
+ "step": 4782
+ },
+ {
+ "epoch": 66.43318777292576,
+ "grad_norm": 0.014306887984275818,
+ "learning_rate": 0.0006,
+ "loss": 4.389184474945068,
+ "step": 4783
+ },
+ {
+ "epoch": 66.44716157205241,
+ "grad_norm": 0.012959563173353672,
+ "learning_rate": 0.0006,
+ "loss": 4.37934684753418,
+ "step": 4784
+ },
+ {
+ "epoch": 66.46113537117904,
+ "grad_norm": 0.013206703588366508,
+ "learning_rate": 0.0006,
+ "loss": 4.442799091339111,
+ "step": 4785
+ },
+ {
+ "epoch": 66.47510917030567,
+ "grad_norm": 0.014848909340798855,
+ "learning_rate": 0.0006,
+ "loss": 4.203481674194336,
+ "step": 4786
+ },
+ {
+ "epoch": 66.48908296943232,
+ "grad_norm": 0.015637682750821114,
+ "learning_rate": 0.0006,
+ "loss": 4.27552604675293,
+ "step": 4787
+ },
+ {
+ "epoch": 66.50305676855895,
+ "grad_norm": 0.016017816960811615,
+ "learning_rate": 0.0006,
+ "loss": 4.296402931213379,
+ "step": 4788
+ },
+ {
+ "epoch": 66.5170305676856,
+ "grad_norm": 0.01577814109623432,
+ "learning_rate": 0.0006,
+ "loss": 4.301733016967773,
+ "step": 4789
+ },
+ {
+ "epoch": 66.53100436681223,
+ "grad_norm": 0.015293175354599953,
+ "learning_rate": 0.0006,
+ "loss": 4.369877815246582,
+ "step": 4790
+ },
+ {
+ "epoch": 66.54497816593886,
+ "grad_norm": 0.014262278564274311,
+ "learning_rate": 0.0006,
+ "loss": 4.3314738273620605,
+ "step": 4791
+ },
+ {
+ "epoch": 66.5589519650655,
+ "grad_norm": 0.015123482793569565,
+ "learning_rate": 0.0006,
+ "loss": 4.313591957092285,
+ "step": 4792
+ },
+ {
+ "epoch": 66.57292576419214,
+ "grad_norm": 0.015458766371011734,
+ "learning_rate": 0.0006,
+ "loss": 4.3045854568481445,
+ "step": 4793
+ },
+ {
+ "epoch": 66.58689956331878,
+ "grad_norm": 0.017217233777046204,
+ "learning_rate": 0.0006,
+ "loss": 4.269944190979004,
+ "step": 4794
+ },
+ {
+ "epoch": 66.60087336244541,
+ "grad_norm": 0.01962282694876194,
+ "learning_rate": 0.0006,
+ "loss": 4.271735191345215,
+ "step": 4795
+ },
+ {
+ "epoch": 66.61484716157206,
+ "grad_norm": 0.01917066052556038,
+ "learning_rate": 0.0006,
+ "loss": 4.288168430328369,
+ "step": 4796
+ },
+ {
+ "epoch": 66.62882096069869,
+ "grad_norm": 0.018211524933576584,
+ "learning_rate": 0.0006,
+ "loss": 4.280660629272461,
+ "step": 4797
+ },
+ {
+ "epoch": 66.64279475982532,
+ "grad_norm": 0.01936524175107479,
+ "learning_rate": 0.0006,
+ "loss": 4.3710551261901855,
+ "step": 4798
+ },
+ {
+ "epoch": 66.65676855895197,
+ "grad_norm": 0.01882351003587246,
+ "learning_rate": 0.0006,
+ "loss": 4.342750549316406,
+ "step": 4799
+ },
+ {
+ "epoch": 66.6707423580786,
+ "grad_norm": 0.017149122431874275,
+ "learning_rate": 0.0006,
+ "loss": 4.32134485244751,
+ "step": 4800
+ },
+ {
+ "epoch": 66.68471615720524,
+ "grad_norm": 0.01699327677488327,
+ "learning_rate": 0.0006,
+ "loss": 4.248893737792969,
+ "step": 4801
+ },
+ {
+ "epoch": 66.69868995633188,
+ "grad_norm": 0.016873255372047424,
+ "learning_rate": 0.0006,
+ "loss": 4.277350425720215,
+ "step": 4802
+ },
+ {
+ "epoch": 66.71266375545852,
+ "grad_norm": 0.016592837870121002,
+ "learning_rate": 0.0006,
+ "loss": 4.221665859222412,
+ "step": 4803
+ },
+ {
+ "epoch": 66.72663755458515,
+ "grad_norm": 0.015887131914496422,
+ "learning_rate": 0.0006,
+ "loss": 4.328040599822998,
+ "step": 4804
+ },
+ {
+ "epoch": 66.74061135371178,
+ "grad_norm": 0.014057286083698273,
+ "learning_rate": 0.0006,
+ "loss": 4.3294548988342285,
+ "step": 4805
+ },
+ {
+ "epoch": 66.75458515283843,
+ "grad_norm": 0.013135905377566814,
+ "learning_rate": 0.0006,
+ "loss": 4.254383563995361,
+ "step": 4806
+ },
+ {
+ "epoch": 66.76855895196506,
+ "grad_norm": 0.013632557354867458,
+ "learning_rate": 0.0006,
+ "loss": 4.342815399169922,
+ "step": 4807
+ },
+ {
+ "epoch": 66.78253275109171,
+ "grad_norm": 0.015512688085436821,
+ "learning_rate": 0.0006,
+ "loss": 4.297835350036621,
+ "step": 4808
+ },
+ {
+ "epoch": 66.79650655021834,
+ "grad_norm": 0.01621832884848118,
+ "learning_rate": 0.0006,
+ "loss": 4.314773082733154,
+ "step": 4809
+ },
+ {
+ "epoch": 66.81048034934499,
+ "grad_norm": 0.016196032986044884,
+ "learning_rate": 0.0006,
+ "loss": 4.29209566116333,
+ "step": 4810
+ },
+ {
+ "epoch": 66.82445414847162,
+ "grad_norm": 0.015463353134691715,
+ "learning_rate": 0.0006,
+ "loss": 4.262795448303223,
+ "step": 4811
+ },
+ {
+ "epoch": 66.83842794759825,
+ "grad_norm": 0.016143787652254105,
+ "learning_rate": 0.0006,
+ "loss": 4.274284362792969,
+ "step": 4812
+ },
+ {
+ "epoch": 66.8524017467249,
+ "grad_norm": 0.01610061153769493,
+ "learning_rate": 0.0006,
+ "loss": 4.337160587310791,
+ "step": 4813
+ },
+ {
+ "epoch": 66.86637554585153,
+ "grad_norm": 0.01582838036119938,
+ "learning_rate": 0.0006,
+ "loss": 4.192723274230957,
+ "step": 4814
+ },
+ {
+ "epoch": 66.88034934497817,
+ "grad_norm": 0.015081540681421757,
+ "learning_rate": 0.0006,
+ "loss": 4.370708465576172,
+ "step": 4815
+ },
+ {
+ "epoch": 66.8943231441048,
+ "grad_norm": 0.015929723158478737,
+ "learning_rate": 0.0006,
+ "loss": 4.340244293212891,
+ "step": 4816
+ },
+ {
+ "epoch": 66.90829694323143,
+ "grad_norm": 0.01784711331129074,
+ "learning_rate": 0.0006,
+ "loss": 4.260406970977783,
+ "step": 4817
+ },
+ {
+ "epoch": 66.92227074235808,
+ "grad_norm": 0.017434703186154366,
+ "learning_rate": 0.0006,
+ "loss": 4.286900043487549,
+ "step": 4818
+ },
+ {
+ "epoch": 66.93624454148471,
+ "grad_norm": 0.016262540593743324,
+ "learning_rate": 0.0006,
+ "loss": 4.292142391204834,
+ "step": 4819
+ },
+ {
+ "epoch": 66.95021834061136,
+ "grad_norm": 0.016122829169034958,
+ "learning_rate": 0.0006,
+ "loss": 4.291485786437988,
+ "step": 4820
+ },
+ {
+ "epoch": 66.96419213973799,
+ "grad_norm": 0.014361178502440453,
+ "learning_rate": 0.0006,
+ "loss": 4.275425910949707,
+ "step": 4821
+ },
+ {
+ "epoch": 66.97816593886463,
+ "grad_norm": 0.01373987551778555,
+ "learning_rate": 0.0006,
+ "loss": 4.223120212554932,
+ "step": 4822
+ },
+ {
+ "epoch": 66.99213973799127,
+ "grad_norm": 0.015687376260757446,
+ "learning_rate": 0.0006,
+ "loss": 4.355381011962891,
+ "step": 4823
+ },
+ {
+ "epoch": 67.0,
+ "grad_norm": 0.01998465321958065,
+ "learning_rate": 0.0006,
+ "loss": 4.264053821563721,
+ "step": 4824
+ },
+ {
+ "epoch": 67.0,
+ "eval_loss": 4.662398338317871,
+ "eval_runtime": 56.7928,
+ "eval_samples_per_second": 42.998,
+ "eval_steps_per_second": 1.356,
+ "step": 4824
+ },
+ {
+ "epoch": 67.01397379912663,
+ "grad_norm": 0.021736836060881615,
+ "learning_rate": 0.0006,
+ "loss": 4.243354797363281,
+ "step": 4825
+ },
+ {
+ "epoch": 67.02794759825328,
+ "grad_norm": 0.021476037800312042,
+ "learning_rate": 0.0006,
+ "loss": 4.218207836151123,
+ "step": 4826
+ },
+ {
+ "epoch": 67.04192139737991,
+ "grad_norm": 0.017188388854265213,
+ "learning_rate": 0.0006,
+ "loss": 4.18048095703125,
+ "step": 4827
+ },
+ {
+ "epoch": 67.05589519650655,
+ "grad_norm": 0.017293719574809074,
+ "learning_rate": 0.0006,
+ "loss": 4.239145278930664,
+ "step": 4828
+ },
+ {
+ "epoch": 67.06986899563319,
+ "grad_norm": 0.017383819445967674,
+ "learning_rate": 0.0006,
+ "loss": 4.114765167236328,
+ "step": 4829
+ },
+ {
+ "epoch": 67.08384279475983,
+ "grad_norm": 0.01651529036462307,
+ "learning_rate": 0.0006,
+ "loss": 4.351023197174072,
+ "step": 4830
+ },
+ {
+ "epoch": 67.09781659388646,
+ "grad_norm": 0.015893027186393738,
+ "learning_rate": 0.0006,
+ "loss": 4.261432647705078,
+ "step": 4831
+ },
+ {
+ "epoch": 67.1117903930131,
+ "grad_norm": 0.013381517492234707,
+ "learning_rate": 0.0006,
+ "loss": 4.172064781188965,
+ "step": 4832
+ },
+ {
+ "epoch": 67.12576419213974,
+ "grad_norm": 0.014642654918134212,
+ "learning_rate": 0.0006,
+ "loss": 4.356171131134033,
+ "step": 4833
+ },
+ {
+ "epoch": 67.13973799126637,
+ "grad_norm": 0.01425282284617424,
+ "learning_rate": 0.0006,
+ "loss": 4.303495407104492,
+ "step": 4834
+ },
+ {
+ "epoch": 67.15371179039302,
+ "grad_norm": 0.014574599452316761,
+ "learning_rate": 0.0006,
+ "loss": 4.2850847244262695,
+ "step": 4835
+ },
+ {
+ "epoch": 67.16768558951965,
+ "grad_norm": 0.012676282785832882,
+ "learning_rate": 0.0006,
+ "loss": 4.297845840454102,
+ "step": 4836
+ },
+ {
+ "epoch": 67.18165938864628,
+ "grad_norm": 0.013948136940598488,
+ "learning_rate": 0.0006,
+ "loss": 4.324530601501465,
+ "step": 4837
+ },
+ {
+ "epoch": 67.19563318777293,
+ "grad_norm": 0.014195773750543594,
+ "learning_rate": 0.0006,
+ "loss": 4.31236457824707,
+ "step": 4838
+ },
+ {
+ "epoch": 67.20960698689956,
+ "grad_norm": 0.013631382025778294,
+ "learning_rate": 0.0006,
+ "loss": 4.086362838745117,
+ "step": 4839
+ },
+ {
+ "epoch": 67.2235807860262,
+ "grad_norm": 0.012798587791621685,
+ "learning_rate": 0.0006,
+ "loss": 4.328471660614014,
+ "step": 4840
+ },
+ {
+ "epoch": 67.23755458515284,
+ "grad_norm": 0.014443321153521538,
+ "learning_rate": 0.0006,
+ "loss": 4.278019905090332,
+ "step": 4841
+ },
+ {
+ "epoch": 67.25152838427948,
+ "grad_norm": 0.015355420298874378,
+ "learning_rate": 0.0006,
+ "loss": 4.333611011505127,
+ "step": 4842
+ },
+ {
+ "epoch": 67.26550218340611,
+ "grad_norm": 0.016487296670675278,
+ "learning_rate": 0.0006,
+ "loss": 4.351336479187012,
+ "step": 4843
+ },
+ {
+ "epoch": 67.27947598253274,
+ "grad_norm": 0.017422541975975037,
+ "learning_rate": 0.0006,
+ "loss": 4.386096477508545,
+ "step": 4844
+ },
+ {
+ "epoch": 67.29344978165939,
+ "grad_norm": 0.015790918841958046,
+ "learning_rate": 0.0006,
+ "loss": 4.301377773284912,
+ "step": 4845
+ },
+ {
+ "epoch": 67.30742358078602,
+ "grad_norm": 0.016427364200353622,
+ "learning_rate": 0.0006,
+ "loss": 4.2838945388793945,
+ "step": 4846
+ },
+ {
+ "epoch": 67.32139737991267,
+ "grad_norm": 0.017050761729478836,
+ "learning_rate": 0.0006,
+ "loss": 4.282417297363281,
+ "step": 4847
+ },
+ {
+ "epoch": 67.3353711790393,
+ "grad_norm": 0.015006253495812416,
+ "learning_rate": 0.0006,
+ "loss": 4.352236270904541,
+ "step": 4848
+ },
+ {
+ "epoch": 67.34934497816595,
+ "grad_norm": 0.012996627949178219,
+ "learning_rate": 0.0006,
+ "loss": 4.309242248535156,
+ "step": 4849
+ },
+ {
+ "epoch": 67.36331877729258,
+ "grad_norm": 0.01367615070194006,
+ "learning_rate": 0.0006,
+ "loss": 4.211801528930664,
+ "step": 4850
+ },
+ {
+ "epoch": 67.37729257641921,
+ "grad_norm": 0.014775963500142097,
+ "learning_rate": 0.0006,
+ "loss": 4.360859394073486,
+ "step": 4851
+ },
+ {
+ "epoch": 67.39126637554585,
+ "grad_norm": 0.015854304656386375,
+ "learning_rate": 0.0006,
+ "loss": 4.211830139160156,
+ "step": 4852
+ },
+ {
+ "epoch": 67.40524017467249,
+ "grad_norm": 0.016888894140720367,
+ "learning_rate": 0.0006,
+ "loss": 4.216549873352051,
+ "step": 4853
+ },
+ {
+ "epoch": 67.41921397379913,
+ "grad_norm": 0.0170894768089056,
+ "learning_rate": 0.0006,
+ "loss": 4.263486385345459,
+ "step": 4854
+ },
+ {
+ "epoch": 67.43318777292576,
+ "grad_norm": 0.015829868614673615,
+ "learning_rate": 0.0006,
+ "loss": 4.360196113586426,
+ "step": 4855
+ },
+ {
+ "epoch": 67.44716157205241,
+ "grad_norm": 0.015220202505588531,
+ "learning_rate": 0.0006,
+ "loss": 4.243794918060303,
+ "step": 4856
+ },
+ {
+ "epoch": 67.46113537117904,
+ "grad_norm": 0.01413388829678297,
+ "learning_rate": 0.0006,
+ "loss": 4.286624908447266,
+ "step": 4857
+ },
+ {
+ "epoch": 67.47510917030567,
+ "grad_norm": 0.01426626369357109,
+ "learning_rate": 0.0006,
+ "loss": 4.257316589355469,
+ "step": 4858
+ },
+ {
+ "epoch": 67.48908296943232,
+ "grad_norm": 0.015957845374941826,
+ "learning_rate": 0.0006,
+ "loss": 4.300443649291992,
+ "step": 4859
+ },
+ {
+ "epoch": 67.50305676855895,
+ "grad_norm": 0.015499051660299301,
+ "learning_rate": 0.0006,
+ "loss": 4.278162002563477,
+ "step": 4860
+ },
+ {
+ "epoch": 67.5170305676856,
+ "grad_norm": 0.015145624987781048,
+ "learning_rate": 0.0006,
+ "loss": 4.267950534820557,
+ "step": 4861
+ },
+ {
+ "epoch": 67.53100436681223,
+ "grad_norm": 0.014610803686082363,
+ "learning_rate": 0.0006,
+ "loss": 4.247499942779541,
+ "step": 4862
+ },
+ {
+ "epoch": 67.54497816593886,
+ "grad_norm": 0.014860909432172775,
+ "learning_rate": 0.0006,
+ "loss": 4.263200759887695,
+ "step": 4863
+ },
+ {
+ "epoch": 67.5589519650655,
+ "grad_norm": 0.01718984916806221,
+ "learning_rate": 0.0006,
+ "loss": 4.290019989013672,
+ "step": 4864
+ },
+ {
+ "epoch": 67.57292576419214,
+ "grad_norm": 0.01558963768184185,
+ "learning_rate": 0.0006,
+ "loss": 4.223816871643066,
+ "step": 4865
+ },
+ {
+ "epoch": 67.58689956331878,
+ "grad_norm": 0.014927188865840435,
+ "learning_rate": 0.0006,
+ "loss": 4.363586902618408,
+ "step": 4866
+ },
+ {
+ "epoch": 67.60087336244541,
+ "grad_norm": 0.01573595404624939,
+ "learning_rate": 0.0006,
+ "loss": 4.254765510559082,
+ "step": 4867
+ },
+ {
+ "epoch": 67.61484716157206,
+ "grad_norm": 0.015618929639458656,
+ "learning_rate": 0.0006,
+ "loss": 4.25072717666626,
+ "step": 4868
+ },
+ {
+ "epoch": 67.62882096069869,
+ "grad_norm": 0.01605590060353279,
+ "learning_rate": 0.0006,
+ "loss": 4.25208854675293,
+ "step": 4869
+ },
+ {
+ "epoch": 67.64279475982532,
+ "grad_norm": 0.01804601214826107,
+ "learning_rate": 0.0006,
+ "loss": 4.351354122161865,
+ "step": 4870
+ },
+ {
+ "epoch": 67.65676855895197,
+ "grad_norm": 0.020595047622919083,
+ "learning_rate": 0.0006,
+ "loss": 4.193732738494873,
+ "step": 4871
+ },
+ {
+ "epoch": 67.6707423580786,
+ "grad_norm": 0.023037966340780258,
+ "learning_rate": 0.0006,
+ "loss": 4.27754020690918,
+ "step": 4872
+ },
+ {
+ "epoch": 67.68471615720524,
+ "grad_norm": 0.020680895075201988,
+ "learning_rate": 0.0006,
+ "loss": 4.314730167388916,
+ "step": 4873
+ },
+ {
+ "epoch": 67.69868995633188,
+ "grad_norm": 0.01887434534728527,
+ "learning_rate": 0.0006,
+ "loss": 4.252305030822754,
+ "step": 4874
+ },
+ {
+ "epoch": 67.71266375545852,
+ "grad_norm": 0.016333596780896187,
+ "learning_rate": 0.0006,
+ "loss": 4.146123886108398,
+ "step": 4875
+ },
+ {
+ "epoch": 67.72663755458515,
+ "grad_norm": 0.015688810497522354,
+ "learning_rate": 0.0006,
+ "loss": 4.2814154624938965,
+ "step": 4876
+ },
+ {
+ "epoch": 67.74061135371178,
+ "grad_norm": 0.016270726919174194,
+ "learning_rate": 0.0006,
+ "loss": 4.323200225830078,
+ "step": 4877
+ },
+ {
+ "epoch": 67.75458515283843,
+ "grad_norm": 0.016824988648295403,
+ "learning_rate": 0.0006,
+ "loss": 4.269501686096191,
+ "step": 4878
+ },
+ {
+ "epoch": 67.76855895196506,
+ "grad_norm": 0.01730562001466751,
+ "learning_rate": 0.0006,
+ "loss": 4.187993049621582,
+ "step": 4879
+ },
+ {
+ "epoch": 67.78253275109171,
+ "grad_norm": 0.01546655036509037,
+ "learning_rate": 0.0006,
+ "loss": 4.340415000915527,
+ "step": 4880
+ },
+ {
+ "epoch": 67.79650655021834,
+ "grad_norm": 0.014173958450555801,
+ "learning_rate": 0.0006,
+ "loss": 4.217419624328613,
+ "step": 4881
+ },
+ {
+ "epoch": 67.81048034934499,
+ "grad_norm": 0.016810474917292595,
+ "learning_rate": 0.0006,
+ "loss": 4.352302074432373,
+ "step": 4882
+ },
+ {
+ "epoch": 67.82445414847162,
+ "grad_norm": 0.0175373163074255,
+ "learning_rate": 0.0006,
+ "loss": 4.20377254486084,
+ "step": 4883
+ },
+ {
+ "epoch": 67.83842794759825,
+ "grad_norm": 0.015734676271677017,
+ "learning_rate": 0.0006,
+ "loss": 4.147495746612549,
+ "step": 4884
+ },
+ {
+ "epoch": 67.8524017467249,
+ "grad_norm": 0.015091817826032639,
+ "learning_rate": 0.0006,
+ "loss": 4.272252082824707,
+ "step": 4885
+ },
+ {
+ "epoch": 67.86637554585153,
+ "grad_norm": 0.015000219456851482,
+ "learning_rate": 0.0006,
+ "loss": 4.302773952484131,
+ "step": 4886
+ },
+ {
+ "epoch": 67.88034934497817,
+ "grad_norm": 0.013827184215188026,
+ "learning_rate": 0.0006,
+ "loss": 4.337242126464844,
+ "step": 4887
+ },
+ {
+ "epoch": 67.8943231441048,
+ "grad_norm": 0.015308346599340439,
+ "learning_rate": 0.0006,
+ "loss": 4.140895366668701,
+ "step": 4888
+ },
+ {
+ "epoch": 67.90829694323143,
+ "grad_norm": 0.015356178395450115,
+ "learning_rate": 0.0006,
+ "loss": 4.312541484832764,
+ "step": 4889
+ },
+ {
+ "epoch": 67.92227074235808,
+ "grad_norm": 0.0137353939935565,
+ "learning_rate": 0.0006,
+ "loss": 4.173184394836426,
+ "step": 4890
+ },
+ {
+ "epoch": 67.93624454148471,
+ "grad_norm": 0.01332368515431881,
+ "learning_rate": 0.0006,
+ "loss": 4.191199779510498,
+ "step": 4891
+ },
+ {
+ "epoch": 67.95021834061136,
+ "grad_norm": 0.01488130260258913,
+ "learning_rate": 0.0006,
+ "loss": 4.331752300262451,
+ "step": 4892
+ },
+ {
+ "epoch": 67.96419213973799,
+ "grad_norm": 0.015449733473360538,
+ "learning_rate": 0.0006,
+ "loss": 4.2385454177856445,
+ "step": 4893
+ },
+ {
+ "epoch": 67.97816593886463,
+ "grad_norm": 0.01556103490293026,
+ "learning_rate": 0.0006,
+ "loss": 4.262353897094727,
+ "step": 4894
+ },
+ {
+ "epoch": 67.99213973799127,
+ "grad_norm": 0.014613849110901356,
+ "learning_rate": 0.0006,
+ "loss": 4.223945617675781,
+ "step": 4895
+ },
+ {
+ "epoch": 68.0,
+ "grad_norm": 0.015323462896049023,
+ "learning_rate": 0.0006,
+ "loss": 4.317168712615967,
+ "step": 4896
+ },
+ {
+ "epoch": 68.0,
+ "eval_loss": 4.671603679656982,
+ "eval_runtime": 56.7317,
+ "eval_samples_per_second": 43.045,
+ "eval_steps_per_second": 1.357,
+ "step": 4896
+ },
+ {
+ "epoch": 68.01397379912663,
+ "grad_norm": 0.01910698413848877,
+ "learning_rate": 0.0006,
+ "loss": 4.206167221069336,
+ "step": 4897
+ },
+ {
+ "epoch": 68.02794759825328,
+ "grad_norm": 0.025112485513091087,
+ "learning_rate": 0.0006,
+ "loss": 4.384706974029541,
+ "step": 4898
+ },
+ {
+ "epoch": 68.04192139737991,
+ "grad_norm": 0.02618720941245556,
+ "learning_rate": 0.0006,
+ "loss": 4.321046829223633,
+ "step": 4899
+ },
+ {
+ "epoch": 68.05589519650655,
+ "grad_norm": 0.020853538066148758,
+ "learning_rate": 0.0006,
+ "loss": 4.217638969421387,
+ "step": 4900
+ },
+ {
+ "epoch": 68.06986899563319,
+ "grad_norm": 0.020025717094540596,
+ "learning_rate": 0.0006,
+ "loss": 4.306824684143066,
+ "step": 4901
+ },
+ {
+ "epoch": 68.08384279475983,
+ "grad_norm": 0.023868650197982788,
+ "learning_rate": 0.0006,
+ "loss": 4.327455997467041,
+ "step": 4902
+ },
+ {
+ "epoch": 68.09781659388646,
+ "grad_norm": 0.022279081866145134,
+ "learning_rate": 0.0006,
+ "loss": 4.400699615478516,
+ "step": 4903
+ },
+ {
+ "epoch": 68.1117903930131,
+ "grad_norm": 0.018965687602758408,
+ "learning_rate": 0.0006,
+ "loss": 4.19097900390625,
+ "step": 4904
+ },
+ {
+ "epoch": 68.12576419213974,
+ "grad_norm": 0.01936378888785839,
+ "learning_rate": 0.0006,
+ "loss": 4.206711769104004,
+ "step": 4905
+ },
+ {
+ "epoch": 68.13973799126637,
+ "grad_norm": 0.01811872608959675,
+ "learning_rate": 0.0006,
+ "loss": 4.204622268676758,
+ "step": 4906
+ },
+ {
+ "epoch": 68.15371179039302,
+ "grad_norm": 0.018530402332544327,
+ "learning_rate": 0.0006,
+ "loss": 4.289602279663086,
+ "step": 4907
+ },
+ {
+ "epoch": 68.16768558951965,
+ "grad_norm": 0.01755962148308754,
+ "learning_rate": 0.0006,
+ "loss": 4.317145347595215,
+ "step": 4908
+ },
+ {
+ "epoch": 68.18165938864628,
+ "grad_norm": 0.020413702353835106,
+ "learning_rate": 0.0006,
+ "loss": 4.2882843017578125,
+ "step": 4909
+ },
+ {
+ "epoch": 68.19563318777293,
+ "grad_norm": 0.020064299926161766,
+ "learning_rate": 0.0006,
+ "loss": 4.248260021209717,
+ "step": 4910
+ },
+ {
+ "epoch": 68.20960698689956,
+ "grad_norm": 0.017415888607501984,
+ "learning_rate": 0.0006,
+ "loss": 4.370295524597168,
+ "step": 4911
+ },
+ {
+ "epoch": 68.2235807860262,
+ "grad_norm": 0.015005858615040779,
+ "learning_rate": 0.0006,
+ "loss": 4.319567680358887,
+ "step": 4912
+ },
+ {
+ "epoch": 68.23755458515284,
+ "grad_norm": 0.016349777579307556,
+ "learning_rate": 0.0006,
+ "loss": 4.301776885986328,
+ "step": 4913
+ },
+ {
+ "epoch": 68.25152838427948,
+ "grad_norm": 0.015779560431838036,
+ "learning_rate": 0.0006,
+ "loss": 4.271171569824219,
+ "step": 4914
+ },
+ {
+ "epoch": 68.26550218340611,
+ "grad_norm": 0.013237644918262959,
+ "learning_rate": 0.0006,
+ "loss": 4.324908256530762,
+ "step": 4915
+ },
+ {
+ "epoch": 68.27947598253274,
+ "grad_norm": 0.014076225459575653,
+ "learning_rate": 0.0006,
+ "loss": 4.177175998687744,
+ "step": 4916
+ },
+ {
+ "epoch": 68.29344978165939,
+ "grad_norm": 0.013906744308769703,
+ "learning_rate": 0.0006,
+ "loss": 4.11588191986084,
+ "step": 4917
+ },
+ {
+ "epoch": 68.30742358078602,
+ "grad_norm": 0.013939064927399158,
+ "learning_rate": 0.0006,
+ "loss": 4.183963775634766,
+ "step": 4918
+ },
+ {
+ "epoch": 68.32139737991267,
+ "grad_norm": 0.014572170563042164,
+ "learning_rate": 0.0006,
+ "loss": 4.19586181640625,
+ "step": 4919
+ },
+ {
+ "epoch": 68.3353711790393,
+ "grad_norm": 0.014085350558161736,
+ "learning_rate": 0.0006,
+ "loss": 4.170892238616943,
+ "step": 4920
+ },
+ {
+ "epoch": 68.34934497816595,
+ "grad_norm": 0.014357863925397396,
+ "learning_rate": 0.0006,
+ "loss": 4.321155548095703,
+ "step": 4921
+ },
+ {
+ "epoch": 68.36331877729258,
+ "grad_norm": 0.01461675763130188,
+ "learning_rate": 0.0006,
+ "loss": 4.293334007263184,
+ "step": 4922
+ },
+ {
+ "epoch": 68.37729257641921,
+ "grad_norm": 0.014828486368060112,
+ "learning_rate": 0.0006,
+ "loss": 4.1439714431762695,
+ "step": 4923
+ },
+ {
+ "epoch": 68.39126637554585,
+ "grad_norm": 0.013730363920331001,
+ "learning_rate": 0.0006,
+ "loss": 4.287973880767822,
+ "step": 4924
+ },
+ {
+ "epoch": 68.40524017467249,
+ "grad_norm": 0.013852203264832497,
+ "learning_rate": 0.0006,
+ "loss": 4.248824119567871,
+ "step": 4925
+ },
+ {
+ "epoch": 68.41921397379913,
+ "grad_norm": 0.014005862176418304,
+ "learning_rate": 0.0006,
+ "loss": 4.293768882751465,
+ "step": 4926
+ },
+ {
+ "epoch": 68.43318777292576,
+ "grad_norm": 0.014435851015150547,
+ "learning_rate": 0.0006,
+ "loss": 4.204813003540039,
+ "step": 4927
+ },
+ {
+ "epoch": 68.44716157205241,
+ "grad_norm": 0.015305934473872185,
+ "learning_rate": 0.0006,
+ "loss": 4.329364776611328,
+ "step": 4928
+ },
+ {
+ "epoch": 68.46113537117904,
+ "grad_norm": 0.014356814324855804,
+ "learning_rate": 0.0006,
+ "loss": 4.264206409454346,
+ "step": 4929
+ },
+ {
+ "epoch": 68.47510917030567,
+ "grad_norm": 0.01432347297668457,
+ "learning_rate": 0.0006,
+ "loss": 4.29440450668335,
+ "step": 4930
+ },
+ {
+ "epoch": 68.48908296943232,
+ "grad_norm": 0.012672432698309422,
+ "learning_rate": 0.0006,
+ "loss": 4.258024215698242,
+ "step": 4931
+ },
+ {
+ "epoch": 68.50305676855895,
+ "grad_norm": 0.014209977351129055,
+ "learning_rate": 0.0006,
+ "loss": 4.216668128967285,
+ "step": 4932
+ },
+ {
+ "epoch": 68.5170305676856,
+ "grad_norm": 0.014606663957238197,
+ "learning_rate": 0.0006,
+ "loss": 4.294184684753418,
+ "step": 4933
+ },
+ {
+ "epoch": 68.53100436681223,
+ "grad_norm": 0.014986990951001644,
+ "learning_rate": 0.0006,
+ "loss": 4.382591247558594,
+ "step": 4934
+ },
+ {
+ "epoch": 68.54497816593886,
+ "grad_norm": 0.013998783193528652,
+ "learning_rate": 0.0006,
+ "loss": 4.195143222808838,
+ "step": 4935
+ },
+ {
+ "epoch": 68.5589519650655,
+ "grad_norm": 0.015011102892458439,
+ "learning_rate": 0.0006,
+ "loss": 4.313785076141357,
+ "step": 4936
+ },
+ {
+ "epoch": 68.57292576419214,
+ "grad_norm": 0.01641209051012993,
+ "learning_rate": 0.0006,
+ "loss": 4.322803020477295,
+ "step": 4937
+ },
+ {
+ "epoch": 68.58689956331878,
+ "grad_norm": 0.016207464039325714,
+ "learning_rate": 0.0006,
+ "loss": 4.311431884765625,
+ "step": 4938
+ },
+ {
+ "epoch": 68.60087336244541,
+ "grad_norm": 0.01747811958193779,
+ "learning_rate": 0.0006,
+ "loss": 4.310845375061035,
+ "step": 4939
+ },
+ {
+ "epoch": 68.61484716157206,
+ "grad_norm": 0.019475717097520828,
+ "learning_rate": 0.0006,
+ "loss": 4.338624954223633,
+ "step": 4940
+ },
+ {
+ "epoch": 68.62882096069869,
+ "grad_norm": 0.018641719594597816,
+ "learning_rate": 0.0006,
+ "loss": 4.34913969039917,
+ "step": 4941
+ },
+ {
+ "epoch": 68.64279475982532,
+ "grad_norm": 0.018376432359218597,
+ "learning_rate": 0.0006,
+ "loss": 4.28464412689209,
+ "step": 4942
+ },
+ {
+ "epoch": 68.65676855895197,
+ "grad_norm": 0.017095010727643967,
+ "learning_rate": 0.0006,
+ "loss": 4.217957496643066,
+ "step": 4943
+ },
+ {
+ "epoch": 68.6707423580786,
+ "grad_norm": 0.016321983188390732,
+ "learning_rate": 0.0006,
+ "loss": 4.220042705535889,
+ "step": 4944
+ },
+ {
+ "epoch": 68.68471615720524,
+ "grad_norm": 0.017128000035881996,
+ "learning_rate": 0.0006,
+ "loss": 4.293500900268555,
+ "step": 4945
+ },
+ {
+ "epoch": 68.69868995633188,
+ "grad_norm": 0.015592411160469055,
+ "learning_rate": 0.0006,
+ "loss": 4.195798873901367,
+ "step": 4946
+ },
+ {
+ "epoch": 68.71266375545852,
+ "grad_norm": 0.016607079654932022,
+ "learning_rate": 0.0006,
+ "loss": 4.2654337882995605,
+ "step": 4947
+ },
+ {
+ "epoch": 68.72663755458515,
+ "grad_norm": 0.01545762736350298,
+ "learning_rate": 0.0006,
+ "loss": 4.303828239440918,
+ "step": 4948
+ },
+ {
+ "epoch": 68.74061135371178,
+ "grad_norm": 0.014328422956168652,
+ "learning_rate": 0.0006,
+ "loss": 4.29379940032959,
+ "step": 4949
+ },
+ {
+ "epoch": 68.75458515283843,
+ "grad_norm": 0.016516804695129395,
+ "learning_rate": 0.0006,
+ "loss": 4.2317214012146,
+ "step": 4950
+ },
+ {
+ "epoch": 68.76855895196506,
+ "grad_norm": 0.016324171796441078,
+ "learning_rate": 0.0006,
+ "loss": 4.278043746948242,
+ "step": 4951
+ },
+ {
+ "epoch": 68.78253275109171,
+ "grad_norm": 0.015511159785091877,
+ "learning_rate": 0.0006,
+ "loss": 4.188244342803955,
+ "step": 4952
+ },
+ {
+ "epoch": 68.79650655021834,
+ "grad_norm": 0.01676103286445141,
+ "learning_rate": 0.0006,
+ "loss": 4.2169976234436035,
+ "step": 4953
+ },
+ {
+ "epoch": 68.81048034934499,
+ "grad_norm": 0.015428312122821808,
+ "learning_rate": 0.0006,
+ "loss": 4.4004716873168945,
+ "step": 4954
+ },
+ {
+ "epoch": 68.82445414847162,
+ "grad_norm": 0.0162410456687212,
+ "learning_rate": 0.0006,
+ "loss": 4.282752513885498,
+ "step": 4955
+ },
+ {
+ "epoch": 68.83842794759825,
+ "grad_norm": 0.017578862607479095,
+ "learning_rate": 0.0006,
+ "loss": 4.098004341125488,
+ "step": 4956
+ },
+ {
+ "epoch": 68.8524017467249,
+ "grad_norm": 0.01805749163031578,
+ "learning_rate": 0.0006,
+ "loss": 4.277988433837891,
+ "step": 4957
+ },
+ {
+ "epoch": 68.86637554585153,
+ "grad_norm": 0.016728045418858528,
+ "learning_rate": 0.0006,
+ "loss": 4.29766321182251,
+ "step": 4958
+ },
+ {
+ "epoch": 68.88034934497817,
+ "grad_norm": 0.014899657107889652,
+ "learning_rate": 0.0006,
+ "loss": 4.329736709594727,
+ "step": 4959
+ },
+ {
+ "epoch": 68.8943231441048,
+ "grad_norm": 0.013441305607557297,
+ "learning_rate": 0.0006,
+ "loss": 4.2873215675354,
+ "step": 4960
+ },
+ {
+ "epoch": 68.90829694323143,
+ "grad_norm": 0.01554128248244524,
+ "learning_rate": 0.0006,
+ "loss": 4.278327941894531,
+ "step": 4961
+ },
+ {
+ "epoch": 68.92227074235808,
+ "grad_norm": 0.01806696318089962,
+ "learning_rate": 0.0006,
+ "loss": 4.38275146484375,
+ "step": 4962
+ },
+ {
+ "epoch": 68.93624454148471,
+ "grad_norm": 0.017025409266352654,
+ "learning_rate": 0.0006,
+ "loss": 4.210447311401367,
+ "step": 4963
+ },
+ {
+ "epoch": 68.95021834061136,
+ "grad_norm": 0.01597420684993267,
+ "learning_rate": 0.0006,
+ "loss": 4.309865951538086,
+ "step": 4964
+ },
+ {
+ "epoch": 68.96419213973799,
+ "grad_norm": 0.015835624188184738,
+ "learning_rate": 0.0006,
+ "loss": 4.225904941558838,
+ "step": 4965
+ },
+ {
+ "epoch": 68.97816593886463,
+ "grad_norm": 0.018170492723584175,
+ "learning_rate": 0.0006,
+ "loss": 4.248478889465332,
+ "step": 4966
+ },
+ {
+ "epoch": 68.99213973799127,
+ "grad_norm": 0.017372991889715195,
+ "learning_rate": 0.0006,
+ "loss": 4.283486843109131,
+ "step": 4967
+ },
+ {
+ "epoch": 69.0,
+ "grad_norm": 0.016305390745401382,
+ "learning_rate": 0.0006,
+ "loss": 4.273636817932129,
+ "step": 4968
+ },
+ {
+ "epoch": 69.0,
+ "eval_loss": 4.66873836517334,
+ "eval_runtime": 57.0791,
+ "eval_samples_per_second": 42.783,
+ "eval_steps_per_second": 1.349,
+ "step": 4968
+ },
+ {
+ "epoch": 69.01397379912663,
+ "grad_norm": 0.016567355021834373,
+ "learning_rate": 0.0006,
+ "loss": 4.269988059997559,
+ "step": 4969
+ },
+ {
+ "epoch": 69.02794759825328,
+ "grad_norm": 0.017476852983236313,
+ "learning_rate": 0.0006,
+ "loss": 4.2518720626831055,
+ "step": 4970
+ },
+ {
+ "epoch": 69.04192139737991,
+ "grad_norm": 0.017070338129997253,
+ "learning_rate": 0.0006,
+ "loss": 4.238024711608887,
+ "step": 4971
+ },
+ {
+ "epoch": 69.05589519650655,
+ "grad_norm": 0.01639373227953911,
+ "learning_rate": 0.0006,
+ "loss": 4.188558101654053,
+ "step": 4972
+ },
+ {
+ "epoch": 69.06986899563319,
+ "grad_norm": 0.01856113411486149,
+ "learning_rate": 0.0006,
+ "loss": 4.339556694030762,
+ "step": 4973
+ },
+ {
+ "epoch": 69.08384279475983,
+ "grad_norm": 0.01762578636407852,
+ "learning_rate": 0.0006,
+ "loss": 4.141518592834473,
+ "step": 4974
+ },
+ {
+ "epoch": 69.09781659388646,
+ "grad_norm": 0.015437949448823929,
+ "learning_rate": 0.0006,
+ "loss": 4.225461959838867,
+ "step": 4975
+ },
+ {
+ "epoch": 69.1117903930131,
+ "grad_norm": 0.015293503180146217,
+ "learning_rate": 0.0006,
+ "loss": 4.290884971618652,
+ "step": 4976
+ },
+ {
+ "epoch": 69.12576419213974,
+ "grad_norm": 0.016259880736470222,
+ "learning_rate": 0.0006,
+ "loss": 4.121919631958008,
+ "step": 4977
+ },
+ {
+ "epoch": 69.13973799126637,
+ "grad_norm": 0.017225584015250206,
+ "learning_rate": 0.0006,
+ "loss": 4.136716842651367,
+ "step": 4978
+ },
+ {
+ "epoch": 69.15371179039302,
+ "grad_norm": 0.01699289306998253,
+ "learning_rate": 0.0006,
+ "loss": 4.193601608276367,
+ "step": 4979
+ },
+ {
+ "epoch": 69.16768558951965,
+ "grad_norm": 0.015613161027431488,
+ "learning_rate": 0.0006,
+ "loss": 4.288212299346924,
+ "step": 4980
+ },
+ {
+ "epoch": 69.18165938864628,
+ "grad_norm": 0.01643933728337288,
+ "learning_rate": 0.0006,
+ "loss": 4.324707984924316,
+ "step": 4981
+ },
+ {
+ "epoch": 69.19563318777293,
+ "grad_norm": 0.01648905873298645,
+ "learning_rate": 0.0006,
+ "loss": 4.287806510925293,
+ "step": 4982
+ },
+ {
+ "epoch": 69.20960698689956,
+ "grad_norm": 0.015444950200617313,
+ "learning_rate": 0.0006,
+ "loss": 4.200622081756592,
+ "step": 4983
+ },
+ {
+ "epoch": 69.2235807860262,
+ "grad_norm": 0.016034258529543877,
+ "learning_rate": 0.0006,
+ "loss": 4.2193827629089355,
+ "step": 4984
+ },
+ {
+ "epoch": 69.23755458515284,
+ "grad_norm": 0.017984388396143913,
+ "learning_rate": 0.0006,
+ "loss": 4.18237829208374,
+ "step": 4985
+ },
+ {
+ "epoch": 69.25152838427948,
+ "grad_norm": 0.0195505041629076,
+ "learning_rate": 0.0006,
+ "loss": 4.216727256774902,
+ "step": 4986
+ },
+ {
+ "epoch": 69.26550218340611,
+ "grad_norm": 0.017779147252440453,
+ "learning_rate": 0.0006,
+ "loss": 4.278628826141357,
+ "step": 4987
+ },
+ {
+ "epoch": 69.27947598253274,
+ "grad_norm": 0.017119476571679115,
+ "learning_rate": 0.0006,
+ "loss": 4.2409138679504395,
+ "step": 4988
+ },
+ {
+ "epoch": 69.29344978165939,
+ "grad_norm": 0.016840364784002304,
+ "learning_rate": 0.0006,
+ "loss": 4.207575798034668,
+ "step": 4989
+ },
+ {
+ "epoch": 69.30742358078602,
+ "grad_norm": 0.014593573287129402,
+ "learning_rate": 0.0006,
+ "loss": 4.164905548095703,
+ "step": 4990
+ },
+ {
+ "epoch": 69.32139737991267,
+ "grad_norm": 0.013922630809247494,
+ "learning_rate": 0.0006,
+ "loss": 4.2696990966796875,
+ "step": 4991
+ },
+ {
+ "epoch": 69.3353711790393,
+ "grad_norm": 0.01577564887702465,
+ "learning_rate": 0.0006,
+ "loss": 4.201589107513428,
+ "step": 4992
+ },
+ {
+ "epoch": 69.34934497816595,
+ "grad_norm": 0.016392096877098083,
+ "learning_rate": 0.0006,
+ "loss": 4.248831748962402,
+ "step": 4993
+ },
+ {
+ "epoch": 69.36331877729258,
+ "grad_norm": 0.015961311757564545,
+ "learning_rate": 0.0006,
+ "loss": 4.181260108947754,
+ "step": 4994
+ },
+ {
+ "epoch": 69.37729257641921,
+ "grad_norm": 0.015350551344454288,
+ "learning_rate": 0.0006,
+ "loss": 4.230352878570557,
+ "step": 4995
+ },
+ {
+ "epoch": 69.39126637554585,
+ "grad_norm": 0.015870466828346252,
+ "learning_rate": 0.0006,
+ "loss": 4.298305511474609,
+ "step": 4996
+ },
+ {
+ "epoch": 69.40524017467249,
+ "grad_norm": 0.013716642744839191,
+ "learning_rate": 0.0006,
+ "loss": 4.220139503479004,
+ "step": 4997
+ },
+ {
+ "epoch": 69.41921397379913,
+ "grad_norm": 0.014431777410209179,
+ "learning_rate": 0.0006,
+ "loss": 4.288130760192871,
+ "step": 4998
+ },
+ {
+ "epoch": 69.43318777292576,
+ "grad_norm": 0.01426647324115038,
+ "learning_rate": 0.0006,
+ "loss": 4.295767784118652,
+ "step": 4999
+ },
+ {
+ "epoch": 69.44716157205241,
+ "grad_norm": 0.014286615885794163,
+ "learning_rate": 0.0006,
+ "loss": 4.38106632232666,
+ "step": 5000
+ },
+ {
+ "epoch": 69.46113537117904,
+ "grad_norm": 0.017990585416555405,
+ "learning_rate": 0.0006,
+ "loss": 4.1593017578125,
+ "step": 5001
+ },
+ {
+ "epoch": 69.47510917030567,
+ "grad_norm": 0.018846049904823303,
+ "learning_rate": 0.0006,
+ "loss": 4.216091632843018,
+ "step": 5002
+ },
+ {
+ "epoch": 69.48908296943232,
+ "grad_norm": 0.018179181963205338,
+ "learning_rate": 0.0006,
+ "loss": 4.240332126617432,
+ "step": 5003
+ },
+ {
+ "epoch": 69.50305676855895,
+ "grad_norm": 0.016693448647856712,
+ "learning_rate": 0.0006,
+ "loss": 4.256624221801758,
+ "step": 5004
+ },
+ {
+ "epoch": 69.5170305676856,
+ "grad_norm": 0.01718989387154579,
+ "learning_rate": 0.0006,
+ "loss": 4.300940990447998,
+ "step": 5005
+ },
+ {
+ "epoch": 69.53100436681223,
+ "grad_norm": 0.020236829295754433,
+ "learning_rate": 0.0006,
+ "loss": 4.289146423339844,
+ "step": 5006
+ },
+ {
+ "epoch": 69.54497816593886,
+ "grad_norm": 0.024860285222530365,
+ "learning_rate": 0.0006,
+ "loss": 4.216216087341309,
+ "step": 5007
+ },
+ {
+ "epoch": 69.5589519650655,
+ "grad_norm": 0.02559969387948513,
+ "learning_rate": 0.0006,
+ "loss": 4.262890815734863,
+ "step": 5008
+ },
+ {
+ "epoch": 69.57292576419214,
+ "grad_norm": 0.019180957227945328,
+ "learning_rate": 0.0006,
+ "loss": 4.230309009552002,
+ "step": 5009
+ },
+ {
+ "epoch": 69.58689956331878,
+ "grad_norm": 0.020442591980099678,
+ "learning_rate": 0.0006,
+ "loss": 4.291759490966797,
+ "step": 5010
+ },
+ {
+ "epoch": 69.60087336244541,
+ "grad_norm": 0.0212619137018919,
+ "learning_rate": 0.0006,
+ "loss": 4.3529839515686035,
+ "step": 5011
+ },
+ {
+ "epoch": 69.61484716157206,
+ "grad_norm": 0.02067175693809986,
+ "learning_rate": 0.0006,
+ "loss": 4.33197021484375,
+ "step": 5012
+ },
+ {
+ "epoch": 69.62882096069869,
+ "grad_norm": 0.01802872307598591,
+ "learning_rate": 0.0006,
+ "loss": 4.2101969718933105,
+ "step": 5013
+ },
+ {
+ "epoch": 69.64279475982532,
+ "grad_norm": 0.01787702925503254,
+ "learning_rate": 0.0006,
+ "loss": 4.270336627960205,
+ "step": 5014
+ },
+ {
+ "epoch": 69.65676855895197,
+ "grad_norm": 0.01586683839559555,
+ "learning_rate": 0.0006,
+ "loss": 4.222309589385986,
+ "step": 5015
+ },
+ {
+ "epoch": 69.6707423580786,
+ "grad_norm": 0.01717999391257763,
+ "learning_rate": 0.0006,
+ "loss": 4.253704071044922,
+ "step": 5016
+ },
+ {
+ "epoch": 69.68471615720524,
+ "grad_norm": 0.01594279147684574,
+ "learning_rate": 0.0006,
+ "loss": 4.344658851623535,
+ "step": 5017
+ },
+ {
+ "epoch": 69.69868995633188,
+ "grad_norm": 0.014419942162930965,
+ "learning_rate": 0.0006,
+ "loss": 4.2144880294799805,
+ "step": 5018
+ },
+ {
+ "epoch": 69.71266375545852,
+ "grad_norm": 0.015824224799871445,
+ "learning_rate": 0.0006,
+ "loss": 4.226252555847168,
+ "step": 5019
+ },
+ {
+ "epoch": 69.72663755458515,
+ "grad_norm": 0.013329979032278061,
+ "learning_rate": 0.0006,
+ "loss": 4.238587379455566,
+ "step": 5020
+ },
+ {
+ "epoch": 69.74061135371178,
+ "grad_norm": 0.013745979405939579,
+ "learning_rate": 0.0006,
+ "loss": 4.317209243774414,
+ "step": 5021
+ },
+ {
+ "epoch": 69.75458515283843,
+ "grad_norm": 0.014064167626202106,
+ "learning_rate": 0.0006,
+ "loss": 4.316982746124268,
+ "step": 5022
+ },
+ {
+ "epoch": 69.76855895196506,
+ "grad_norm": 0.014611301012337208,
+ "learning_rate": 0.0006,
+ "loss": 4.212728500366211,
+ "step": 5023
+ },
+ {
+ "epoch": 69.78253275109171,
+ "grad_norm": 0.014947416260838509,
+ "learning_rate": 0.0006,
+ "loss": 4.319550514221191,
+ "step": 5024
+ },
+ {
+ "epoch": 69.79650655021834,
+ "grad_norm": 0.014780526980757713,
+ "learning_rate": 0.0006,
+ "loss": 4.1888933181762695,
+ "step": 5025
+ },
+ {
+ "epoch": 69.81048034934499,
+ "grad_norm": 0.01611459068953991,
+ "learning_rate": 0.0006,
+ "loss": 4.239692211151123,
+ "step": 5026
+ },
+ {
+ "epoch": 69.82445414847162,
+ "grad_norm": 0.015395848080515862,
+ "learning_rate": 0.0006,
+ "loss": 4.308794021606445,
+ "step": 5027
+ },
+ {
+ "epoch": 69.83842794759825,
+ "grad_norm": 0.016113396733999252,
+ "learning_rate": 0.0006,
+ "loss": 4.328035354614258,
+ "step": 5028
+ },
+ {
+ "epoch": 69.8524017467249,
+ "grad_norm": 0.016386888921260834,
+ "learning_rate": 0.0006,
+ "loss": 4.272579193115234,
+ "step": 5029
+ },
+ {
+ "epoch": 69.86637554585153,
+ "grad_norm": 0.013675205409526825,
+ "learning_rate": 0.0006,
+ "loss": 4.226552963256836,
+ "step": 5030
+ },
+ {
+ "epoch": 69.88034934497817,
+ "grad_norm": 0.01472668256610632,
+ "learning_rate": 0.0006,
+ "loss": 4.224167346954346,
+ "step": 5031
+ },
+ {
+ "epoch": 69.8943231441048,
+ "grad_norm": 0.014958228915929794,
+ "learning_rate": 0.0006,
+ "loss": 4.232619285583496,
+ "step": 5032
+ },
+ {
+ "epoch": 69.90829694323143,
+ "grad_norm": 0.012926832772791386,
+ "learning_rate": 0.0006,
+ "loss": 4.27107048034668,
+ "step": 5033
+ },
+ {
+ "epoch": 69.92227074235808,
+ "grad_norm": 0.013138801790773869,
+ "learning_rate": 0.0006,
+ "loss": 4.288477420806885,
+ "step": 5034
+ },
+ {
+ "epoch": 69.93624454148471,
+ "grad_norm": 0.013415461406111717,
+ "learning_rate": 0.0006,
+ "loss": 4.2033257484436035,
+ "step": 5035
+ },
+ {
+ "epoch": 69.95021834061136,
+ "grad_norm": 0.013807233422994614,
+ "learning_rate": 0.0006,
+ "loss": 4.274764537811279,
+ "step": 5036
+ },
+ {
+ "epoch": 69.96419213973799,
+ "grad_norm": 0.013929193839430809,
+ "learning_rate": 0.0006,
+ "loss": 4.1854352951049805,
+ "step": 5037
+ },
+ {
+ "epoch": 69.97816593886463,
+ "grad_norm": 0.015497001819312572,
+ "learning_rate": 0.0006,
+ "loss": 4.226498126983643,
+ "step": 5038
+ },
+ {
+ "epoch": 69.99213973799127,
+ "grad_norm": 0.01874048449099064,
+ "learning_rate": 0.0006,
+ "loss": 4.304137706756592,
+ "step": 5039
+ },
+ {
+ "epoch": 70.0,
+ "grad_norm": 0.021745791658759117,
+ "learning_rate": 0.0006,
+ "loss": 4.206809997558594,
+ "step": 5040
+ },
+ {
+ "epoch": 70.0,
+ "eval_loss": 4.675539970397949,
+ "eval_runtime": 56.9034,
+ "eval_samples_per_second": 42.915,
+ "eval_steps_per_second": 1.353,
+ "step": 5040
+ },
+ {
+ "epoch": 70.01397379912663,
+ "grad_norm": 0.02124839462339878,
+ "learning_rate": 0.0006,
+ "loss": 4.198199272155762,
+ "step": 5041
+ },
+ {
+ "epoch": 70.02794759825328,
+ "grad_norm": 0.019432688131928444,
+ "learning_rate": 0.0006,
+ "loss": 4.187261581420898,
+ "step": 5042
+ },
+ {
+ "epoch": 70.04192139737991,
+ "grad_norm": 0.01769290305674076,
+ "learning_rate": 0.0006,
+ "loss": 4.155284881591797,
+ "step": 5043
+ },
+ {
+ "epoch": 70.05589519650655,
+ "grad_norm": 0.01751520484685898,
+ "learning_rate": 0.0006,
+ "loss": 4.386857986450195,
+ "step": 5044
+ },
+ {
+ "epoch": 70.06986899563319,
+ "grad_norm": 0.018138699233531952,
+ "learning_rate": 0.0006,
+ "loss": 4.117342472076416,
+ "step": 5045
+ },
+ {
+ "epoch": 70.08384279475983,
+ "grad_norm": 0.016331054270267487,
+ "learning_rate": 0.0006,
+ "loss": 4.225295066833496,
+ "step": 5046
+ },
+ {
+ "epoch": 70.09781659388646,
+ "grad_norm": 0.01658729277551174,
+ "learning_rate": 0.0006,
+ "loss": 4.143835067749023,
+ "step": 5047
+ },
+ {
+ "epoch": 70.1117903930131,
+ "grad_norm": 0.016367698088288307,
+ "learning_rate": 0.0006,
+ "loss": 4.1469268798828125,
+ "step": 5048
+ },
+ {
+ "epoch": 70.12576419213974,
+ "grad_norm": 0.015151629224419594,
+ "learning_rate": 0.0006,
+ "loss": 4.117629051208496,
+ "step": 5049
+ },
+ {
+ "epoch": 70.13973799126637,
+ "grad_norm": 0.016434503719210625,
+ "learning_rate": 0.0006,
+ "loss": 4.107848644256592,
+ "step": 5050
+ },
+ {
+ "epoch": 70.15371179039302,
+ "grad_norm": 0.017837485298514366,
+ "learning_rate": 0.0006,
+ "loss": 4.10837459564209,
+ "step": 5051
+ },
+ {
+ "epoch": 70.16768558951965,
+ "grad_norm": 0.0165961142629385,
+ "learning_rate": 0.0006,
+ "loss": 4.169414043426514,
+ "step": 5052
+ },
+ {
+ "epoch": 70.18165938864628,
+ "grad_norm": 0.0157014112919569,
+ "learning_rate": 0.0006,
+ "loss": 4.0629706382751465,
+ "step": 5053
+ },
+ {
+ "epoch": 70.19563318777293,
+ "grad_norm": 0.01581883803009987,
+ "learning_rate": 0.0006,
+ "loss": 4.220012664794922,
+ "step": 5054
+ },
+ {
+ "epoch": 70.20960698689956,
+ "grad_norm": 0.015285917557775974,
+ "learning_rate": 0.0006,
+ "loss": 4.275337219238281,
+ "step": 5055
+ },
+ {
+ "epoch": 70.2235807860262,
+ "grad_norm": 0.014984549954533577,
+ "learning_rate": 0.0006,
+ "loss": 4.284764289855957,
+ "step": 5056
+ },
+ {
+ "epoch": 70.23755458515284,
+ "grad_norm": 0.015172012150287628,
+ "learning_rate": 0.0006,
+ "loss": 4.234498023986816,
+ "step": 5057
+ },
+ {
+ "epoch": 70.25152838427948,
+ "grad_norm": 0.016846632584929466,
+ "learning_rate": 0.0006,
+ "loss": 4.181689262390137,
+ "step": 5058
+ },
+ {
+ "epoch": 70.26550218340611,
+ "grad_norm": 0.01920711249113083,
+ "learning_rate": 0.0006,
+ "loss": 4.235280513763428,
+ "step": 5059
+ },
+ {
+ "epoch": 70.27947598253274,
+ "grad_norm": 0.017765291035175323,
+ "learning_rate": 0.0006,
+ "loss": 4.238955497741699,
+ "step": 5060
+ },
+ {
+ "epoch": 70.29344978165939,
+ "grad_norm": 0.015436639077961445,
+ "learning_rate": 0.0006,
+ "loss": 4.232516288757324,
+ "step": 5061
+ },
+ {
+ "epoch": 70.30742358078602,
+ "grad_norm": 0.015660086646676064,
+ "learning_rate": 0.0006,
+ "loss": 4.260349273681641,
+ "step": 5062
+ },
+ {
+ "epoch": 70.32139737991267,
+ "grad_norm": 0.016095709055662155,
+ "learning_rate": 0.0006,
+ "loss": 4.168505668640137,
+ "step": 5063
+ },
+ {
+ "epoch": 70.3353711790393,
+ "grad_norm": 0.015001940540969372,
+ "learning_rate": 0.0006,
+ "loss": 4.161739349365234,
+ "step": 5064
+ },
+ {
+ "epoch": 70.34934497816595,
+ "grad_norm": 0.014668774791061878,
+ "learning_rate": 0.0006,
+ "loss": 4.181029796600342,
+ "step": 5065
+ },
+ {
+ "epoch": 70.36331877729258,
+ "grad_norm": 0.013266284950077534,
+ "learning_rate": 0.0006,
+ "loss": 4.3052873611450195,
+ "step": 5066
+ },
+ {
+ "epoch": 70.37729257641921,
+ "grad_norm": 0.015396138653159142,
+ "learning_rate": 0.0006,
+ "loss": 4.34413480758667,
+ "step": 5067
+ },
+ {
+ "epoch": 70.39126637554585,
+ "grad_norm": 0.015511523932218552,
+ "learning_rate": 0.0006,
+ "loss": 4.235037803649902,
+ "step": 5068
+ },
+ {
+ "epoch": 70.40524017467249,
+ "grad_norm": 0.014825738966464996,
+ "learning_rate": 0.0006,
+ "loss": 4.262448310852051,
+ "step": 5069
+ },
+ {
+ "epoch": 70.41921397379913,
+ "grad_norm": 0.015309952199459076,
+ "learning_rate": 0.0006,
+ "loss": 4.258711814880371,
+ "step": 5070
+ },
+ {
+ "epoch": 70.43318777292576,
+ "grad_norm": 0.017614055424928665,
+ "learning_rate": 0.0006,
+ "loss": 4.286028861999512,
+ "step": 5071
+ },
+ {
+ "epoch": 70.44716157205241,
+ "grad_norm": 0.0166039876639843,
+ "learning_rate": 0.0006,
+ "loss": 4.149197578430176,
+ "step": 5072
+ },
+ {
+ "epoch": 70.46113537117904,
+ "grad_norm": 0.016911419108510017,
+ "learning_rate": 0.0006,
+ "loss": 4.271981239318848,
+ "step": 5073
+ },
+ {
+ "epoch": 70.47510917030567,
+ "grad_norm": 0.016806820407509804,
+ "learning_rate": 0.0006,
+ "loss": 4.213315010070801,
+ "step": 5074
+ },
+ {
+ "epoch": 70.48908296943232,
+ "grad_norm": 0.015221742913126945,
+ "learning_rate": 0.0006,
+ "loss": 4.171091079711914,
+ "step": 5075
+ },
+ {
+ "epoch": 70.50305676855895,
+ "grad_norm": 0.014175347983837128,
+ "learning_rate": 0.0006,
+ "loss": 4.185309410095215,
+ "step": 5076
+ },
+ {
+ "epoch": 70.5170305676856,
+ "grad_norm": 0.013897798955440521,
+ "learning_rate": 0.0006,
+ "loss": 4.250119209289551,
+ "step": 5077
+ },
+ {
+ "epoch": 70.53100436681223,
+ "grad_norm": 0.014791003428399563,
+ "learning_rate": 0.0006,
+ "loss": 4.2504682540893555,
+ "step": 5078
+ },
+ {
+ "epoch": 70.54497816593886,
+ "grad_norm": 0.016700128093361855,
+ "learning_rate": 0.0006,
+ "loss": 4.235134124755859,
+ "step": 5079
+ },
+ {
+ "epoch": 70.5589519650655,
+ "grad_norm": 0.016679363325238228,
+ "learning_rate": 0.0006,
+ "loss": 4.32028865814209,
+ "step": 5080
+ },
+ {
+ "epoch": 70.57292576419214,
+ "grad_norm": 0.015117401257157326,
+ "learning_rate": 0.0006,
+ "loss": 4.210995674133301,
+ "step": 5081
+ },
+ {
+ "epoch": 70.58689956331878,
+ "grad_norm": 0.014441263861954212,
+ "learning_rate": 0.0006,
+ "loss": 4.27849006652832,
+ "step": 5082
+ },
+ {
+ "epoch": 70.60087336244541,
+ "grad_norm": 0.01523780357092619,
+ "learning_rate": 0.0006,
+ "loss": 4.197360992431641,
+ "step": 5083
+ },
+ {
+ "epoch": 70.61484716157206,
+ "grad_norm": 0.016201738268136978,
+ "learning_rate": 0.0006,
+ "loss": 4.28239107131958,
+ "step": 5084
+ },
+ {
+ "epoch": 70.62882096069869,
+ "grad_norm": 0.016003292053937912,
+ "learning_rate": 0.0006,
+ "loss": 4.209715843200684,
+ "step": 5085
+ },
+ {
+ "epoch": 70.64279475982532,
+ "grad_norm": 0.01472893450409174,
+ "learning_rate": 0.0006,
+ "loss": 4.169078826904297,
+ "step": 5086
+ },
+ {
+ "epoch": 70.65676855895197,
+ "grad_norm": 0.01604326069355011,
+ "learning_rate": 0.0006,
+ "loss": 4.090967178344727,
+ "step": 5087
+ },
+ {
+ "epoch": 70.6707423580786,
+ "grad_norm": 0.01598356105387211,
+ "learning_rate": 0.0006,
+ "loss": 4.16630220413208,
+ "step": 5088
+ },
+ {
+ "epoch": 70.68471615720524,
+ "grad_norm": 0.016988780349493027,
+ "learning_rate": 0.0006,
+ "loss": 4.117822170257568,
+ "step": 5089
+ },
+ {
+ "epoch": 70.69868995633188,
+ "grad_norm": 0.018582893535494804,
+ "learning_rate": 0.0006,
+ "loss": 4.192902088165283,
+ "step": 5090
+ },
+ {
+ "epoch": 70.71266375545852,
+ "grad_norm": 0.01793155074119568,
+ "learning_rate": 0.0006,
+ "loss": 4.314806938171387,
+ "step": 5091
+ },
+ {
+ "epoch": 70.72663755458515,
+ "grad_norm": 0.016286427155137062,
+ "learning_rate": 0.0006,
+ "loss": 4.2137556076049805,
+ "step": 5092
+ },
+ {
+ "epoch": 70.74061135371178,
+ "grad_norm": 0.014718376100063324,
+ "learning_rate": 0.0006,
+ "loss": 4.3073015213012695,
+ "step": 5093
+ },
+ {
+ "epoch": 70.75458515283843,
+ "grad_norm": 0.014675077050924301,
+ "learning_rate": 0.0006,
+ "loss": 4.160961151123047,
+ "step": 5094
+ },
+ {
+ "epoch": 70.76855895196506,
+ "grad_norm": 0.015550913289189339,
+ "learning_rate": 0.0006,
+ "loss": 4.322887420654297,
+ "step": 5095
+ },
+ {
+ "epoch": 70.78253275109171,
+ "grad_norm": 0.01453235000371933,
+ "learning_rate": 0.0006,
+ "loss": 4.2915873527526855,
+ "step": 5096
+ },
+ {
+ "epoch": 70.79650655021834,
+ "grad_norm": 0.01459164172410965,
+ "learning_rate": 0.0006,
+ "loss": 4.247288227081299,
+ "step": 5097
+ },
+ {
+ "epoch": 70.81048034934499,
+ "grad_norm": 0.015395080670714378,
+ "learning_rate": 0.0006,
+ "loss": 4.360955715179443,
+ "step": 5098
+ },
+ {
+ "epoch": 70.82445414847162,
+ "grad_norm": 0.01317683607339859,
+ "learning_rate": 0.0006,
+ "loss": 4.181969165802002,
+ "step": 5099
+ },
+ {
+ "epoch": 70.83842794759825,
+ "grad_norm": 0.013917932286858559,
+ "learning_rate": 0.0006,
+ "loss": 4.264626979827881,
+ "step": 5100
+ },
+ {
+ "epoch": 70.8524017467249,
+ "grad_norm": 0.015268695540726185,
+ "learning_rate": 0.0006,
+ "loss": 4.287857532501221,
+ "step": 5101
+ },
+ {
+ "epoch": 70.86637554585153,
+ "grad_norm": 0.015531329438090324,
+ "learning_rate": 0.0006,
+ "loss": 4.294188499450684,
+ "step": 5102
+ },
+ {
+ "epoch": 70.88034934497817,
+ "grad_norm": 0.015607129782438278,
+ "learning_rate": 0.0006,
+ "loss": 4.331040859222412,
+ "step": 5103
+ },
+ {
+ "epoch": 70.8943231441048,
+ "grad_norm": 0.015297149308025837,
+ "learning_rate": 0.0006,
+ "loss": 4.243650436401367,
+ "step": 5104
+ },
+ {
+ "epoch": 70.90829694323143,
+ "grad_norm": 0.014790025539696217,
+ "learning_rate": 0.0006,
+ "loss": 4.159859657287598,
+ "step": 5105
+ },
+ {
+ "epoch": 70.92227074235808,
+ "grad_norm": 0.013466274365782738,
+ "learning_rate": 0.0006,
+ "loss": 4.221992492675781,
+ "step": 5106
+ },
+ {
+ "epoch": 70.93624454148471,
+ "grad_norm": 0.01361898984760046,
+ "learning_rate": 0.0006,
+ "loss": 4.278751373291016,
+ "step": 5107
+ },
+ {
+ "epoch": 70.95021834061136,
+ "grad_norm": 0.01325751468539238,
+ "learning_rate": 0.0006,
+ "loss": 4.166001319885254,
+ "step": 5108
+ },
+ {
+ "epoch": 70.96419213973799,
+ "grad_norm": 0.012727087363600731,
+ "learning_rate": 0.0006,
+ "loss": 4.219997406005859,
+ "step": 5109
+ },
+ {
+ "epoch": 70.97816593886463,
+ "grad_norm": 0.013966867700219154,
+ "learning_rate": 0.0006,
+ "loss": 4.207491397857666,
+ "step": 5110
+ },
+ {
+ "epoch": 70.99213973799127,
+ "grad_norm": 0.015053587965667248,
+ "learning_rate": 0.0006,
+ "loss": 4.343245983123779,
+ "step": 5111
+ },
+ {
+ "epoch": 71.0,
+ "grad_norm": 0.017723368480801582,
+ "learning_rate": 0.0006,
+ "loss": 4.356122016906738,
+ "step": 5112
+ },
+ {
+ "epoch": 71.0,
+ "eval_loss": 4.598303318023682,
+ "eval_runtime": 56.7521,
+ "eval_samples_per_second": 43.029,
+ "eval_steps_per_second": 1.357,
+ "step": 5112
+ },
+ {
+ "epoch": 71.01397379912663,
+ "grad_norm": 0.018996279686689377,
+ "learning_rate": 0.0006,
+ "loss": 4.234772205352783,
+ "step": 5113
+ },
+ {
+ "epoch": 71.02794759825328,
+ "grad_norm": 0.02041555754840374,
+ "learning_rate": 0.0006,
+ "loss": 4.275816917419434,
+ "step": 5114
+ },
+ {
+ "epoch": 71.04192139737991,
+ "grad_norm": 0.021955285221338272,
+ "learning_rate": 0.0006,
+ "loss": 4.166038513183594,
+ "step": 5115
+ },
+ {
+ "epoch": 71.05589519650655,
+ "grad_norm": 0.021308740600943565,
+ "learning_rate": 0.0006,
+ "loss": 4.13063907623291,
+ "step": 5116
+ },
+ {
+ "epoch": 71.06986899563319,
+ "grad_norm": 0.01747054047882557,
+ "learning_rate": 0.0006,
+ "loss": 4.200522422790527,
+ "step": 5117
+ },
+ {
+ "epoch": 71.08384279475983,
+ "grad_norm": 0.018544532358646393,
+ "learning_rate": 0.0006,
+ "loss": 4.173915386199951,
+ "step": 5118
+ },
+ {
+ "epoch": 71.09781659388646,
+ "grad_norm": 0.0182361900806427,
+ "learning_rate": 0.0006,
+ "loss": 4.27392578125,
+ "step": 5119
+ },
+ {
+ "epoch": 71.1117903930131,
+ "grad_norm": 0.017062073573470116,
+ "learning_rate": 0.0006,
+ "loss": 4.219576358795166,
+ "step": 5120
+ },
+ {
+ "epoch": 71.12576419213974,
+ "grad_norm": 0.016289666295051575,
+ "learning_rate": 0.0006,
+ "loss": 4.133376598358154,
+ "step": 5121
+ },
+ {
+ "epoch": 71.13973799126637,
+ "grad_norm": 0.014864642173051834,
+ "learning_rate": 0.0006,
+ "loss": 4.168276786804199,
+ "step": 5122
+ },
+ {
+ "epoch": 71.15371179039302,
+ "grad_norm": 0.015309540554881096,
+ "learning_rate": 0.0006,
+ "loss": 4.175103187561035,
+ "step": 5123
+ },
+ {
+ "epoch": 71.16768558951965,
+ "grad_norm": 0.01556059904396534,
+ "learning_rate": 0.0006,
+ "loss": 4.289899826049805,
+ "step": 5124
+ },
+ {
+ "epoch": 71.18165938864628,
+ "grad_norm": 0.01479957066476345,
+ "learning_rate": 0.0006,
+ "loss": 4.341495037078857,
+ "step": 5125
+ },
+ {
+ "epoch": 71.19563318777293,
+ "grad_norm": 0.014762179926037788,
+ "learning_rate": 0.0006,
+ "loss": 4.252333641052246,
+ "step": 5126
+ },
+ {
+ "epoch": 71.20960698689956,
+ "grad_norm": 0.014897473156452179,
+ "learning_rate": 0.0006,
+ "loss": 4.2244181632995605,
+ "step": 5127
+ },
+ {
+ "epoch": 71.2235807860262,
+ "grad_norm": 0.012990163639187813,
+ "learning_rate": 0.0006,
+ "loss": 4.237288475036621,
+ "step": 5128
+ },
+ {
+ "epoch": 71.23755458515284,
+ "grad_norm": 0.012858416885137558,
+ "learning_rate": 0.0006,
+ "loss": 4.265738487243652,
+ "step": 5129
+ },
+ {
+ "epoch": 71.25152838427948,
+ "grad_norm": 0.013454788364470005,
+ "learning_rate": 0.0006,
+ "loss": 4.289131164550781,
+ "step": 5130
+ },
+ {
+ "epoch": 71.26550218340611,
+ "grad_norm": 0.013717238791286945,
+ "learning_rate": 0.0006,
+ "loss": 4.2844014167785645,
+ "step": 5131
+ },
+ {
+ "epoch": 71.27947598253274,
+ "grad_norm": 0.0135411461815238,
+ "learning_rate": 0.0006,
+ "loss": 4.139082908630371,
+ "step": 5132
+ },
+ {
+ "epoch": 71.29344978165939,
+ "grad_norm": 0.01406062301248312,
+ "learning_rate": 0.0006,
+ "loss": 4.226719856262207,
+ "step": 5133
+ },
+ {
+ "epoch": 71.30742358078602,
+ "grad_norm": 0.014027190394699574,
+ "learning_rate": 0.0006,
+ "loss": 4.191699981689453,
+ "step": 5134
+ },
+ {
+ "epoch": 71.32139737991267,
+ "grad_norm": 0.013809907250106335,
+ "learning_rate": 0.0006,
+ "loss": 4.242542266845703,
+ "step": 5135
+ },
+ {
+ "epoch": 71.3353711790393,
+ "grad_norm": 0.013507804833352566,
+ "learning_rate": 0.0006,
+ "loss": 4.068206310272217,
+ "step": 5136
+ },
+ {
+ "epoch": 71.34934497816595,
+ "grad_norm": 0.01432533748447895,
+ "learning_rate": 0.0006,
+ "loss": 4.378222942352295,
+ "step": 5137
+ },
+ {
+ "epoch": 71.36331877729258,
+ "grad_norm": 0.01753002591431141,
+ "learning_rate": 0.0006,
+ "loss": 4.174706935882568,
+ "step": 5138
+ },
+ {
+ "epoch": 71.37729257641921,
+ "grad_norm": 0.020048344507813454,
+ "learning_rate": 0.0006,
+ "loss": 4.19920015335083,
+ "step": 5139
+ },
+ {
+ "epoch": 71.39126637554585,
+ "grad_norm": 0.018551401793956757,
+ "learning_rate": 0.0006,
+ "loss": 4.175182342529297,
+ "step": 5140
+ },
+ {
+ "epoch": 71.40524017467249,
+ "grad_norm": 0.01598959229886532,
+ "learning_rate": 0.0006,
+ "loss": 4.2305169105529785,
+ "step": 5141
+ },
+ {
+ "epoch": 71.41921397379913,
+ "grad_norm": 0.016029560938477516,
+ "learning_rate": 0.0006,
+ "loss": 4.230090618133545,
+ "step": 5142
+ },
+ {
+ "epoch": 71.43318777292576,
+ "grad_norm": 0.01595657505095005,
+ "learning_rate": 0.0006,
+ "loss": 4.138402938842773,
+ "step": 5143
+ },
+ {
+ "epoch": 71.44716157205241,
+ "grad_norm": 0.01623230241239071,
+ "learning_rate": 0.0006,
+ "loss": 4.196106433868408,
+ "step": 5144
+ },
+ {
+ "epoch": 71.46113537117904,
+ "grad_norm": 0.014057597145438194,
+ "learning_rate": 0.0006,
+ "loss": 4.2599897384643555,
+ "step": 5145
+ },
+ {
+ "epoch": 71.47510917030567,
+ "grad_norm": 0.017161158844828606,
+ "learning_rate": 0.0006,
+ "loss": 4.183909893035889,
+ "step": 5146
+ },
+ {
+ "epoch": 71.48908296943232,
+ "grad_norm": 0.017038581892848015,
+ "learning_rate": 0.0006,
+ "loss": 4.373933792114258,
+ "step": 5147
+ },
+ {
+ "epoch": 71.50305676855895,
+ "grad_norm": 0.015851367264986038,
+ "learning_rate": 0.0006,
+ "loss": 4.2620086669921875,
+ "step": 5148
+ },
+ {
+ "epoch": 71.5170305676856,
+ "grad_norm": 0.014598303474485874,
+ "learning_rate": 0.0006,
+ "loss": 4.197151184082031,
+ "step": 5149
+ },
+ {
+ "epoch": 71.53100436681223,
+ "grad_norm": 0.01339802611619234,
+ "learning_rate": 0.0006,
+ "loss": 4.110706329345703,
+ "step": 5150
+ },
+ {
+ "epoch": 71.54497816593886,
+ "grad_norm": 0.015567775815725327,
+ "learning_rate": 0.0006,
+ "loss": 4.1210246086120605,
+ "step": 5151
+ },
+ {
+ "epoch": 71.5589519650655,
+ "grad_norm": 0.015616007149219513,
+ "learning_rate": 0.0006,
+ "loss": 4.129796028137207,
+ "step": 5152
+ },
+ {
+ "epoch": 71.57292576419214,
+ "grad_norm": 0.014198964461684227,
+ "learning_rate": 0.0006,
+ "loss": 4.193232536315918,
+ "step": 5153
+ },
+ {
+ "epoch": 71.58689956331878,
+ "grad_norm": 0.015226383693516254,
+ "learning_rate": 0.0006,
+ "loss": 4.223779678344727,
+ "step": 5154
+ },
+ {
+ "epoch": 71.60087336244541,
+ "grad_norm": 0.015610109083354473,
+ "learning_rate": 0.0006,
+ "loss": 4.223179817199707,
+ "step": 5155
+ },
+ {
+ "epoch": 71.61484716157206,
+ "grad_norm": 0.016034303233027458,
+ "learning_rate": 0.0006,
+ "loss": 4.196593284606934,
+ "step": 5156
+ },
+ {
+ "epoch": 71.62882096069869,
+ "grad_norm": 0.015648581087589264,
+ "learning_rate": 0.0006,
+ "loss": 4.341497421264648,
+ "step": 5157
+ },
+ {
+ "epoch": 71.64279475982532,
+ "grad_norm": 0.014117540791630745,
+ "learning_rate": 0.0006,
+ "loss": 4.194307327270508,
+ "step": 5158
+ },
+ {
+ "epoch": 71.65676855895197,
+ "grad_norm": 0.015612686984241009,
+ "learning_rate": 0.0006,
+ "loss": 4.290152549743652,
+ "step": 5159
+ },
+ {
+ "epoch": 71.6707423580786,
+ "grad_norm": 0.01722259446978569,
+ "learning_rate": 0.0006,
+ "loss": 4.258018493652344,
+ "step": 5160
+ },
+ {
+ "epoch": 71.68471615720524,
+ "grad_norm": 0.015923351049423218,
+ "learning_rate": 0.0006,
+ "loss": 4.21684455871582,
+ "step": 5161
+ },
+ {
+ "epoch": 71.69868995633188,
+ "grad_norm": 0.013663450255990028,
+ "learning_rate": 0.0006,
+ "loss": 4.175601959228516,
+ "step": 5162
+ },
+ {
+ "epoch": 71.71266375545852,
+ "grad_norm": 0.017075596377253532,
+ "learning_rate": 0.0006,
+ "loss": 4.205138683319092,
+ "step": 5163
+ },
+ {
+ "epoch": 71.72663755458515,
+ "grad_norm": 0.01723487675189972,
+ "learning_rate": 0.0006,
+ "loss": 4.08839225769043,
+ "step": 5164
+ },
+ {
+ "epoch": 71.74061135371178,
+ "grad_norm": 0.017355505377054214,
+ "learning_rate": 0.0006,
+ "loss": 4.165130615234375,
+ "step": 5165
+ },
+ {
+ "epoch": 71.75458515283843,
+ "grad_norm": 0.01868261583149433,
+ "learning_rate": 0.0006,
+ "loss": 4.16977071762085,
+ "step": 5166
+ },
+ {
+ "epoch": 71.76855895196506,
+ "grad_norm": 0.01897321827709675,
+ "learning_rate": 0.0006,
+ "loss": 4.173497200012207,
+ "step": 5167
+ },
+ {
+ "epoch": 71.78253275109171,
+ "grad_norm": 0.02045825682580471,
+ "learning_rate": 0.0006,
+ "loss": 4.176192283630371,
+ "step": 5168
+ },
+ {
+ "epoch": 71.79650655021834,
+ "grad_norm": 0.022895624861121178,
+ "learning_rate": 0.0006,
+ "loss": 4.163009166717529,
+ "step": 5169
+ },
+ {
+ "epoch": 71.81048034934499,
+ "grad_norm": 0.02062244527041912,
+ "learning_rate": 0.0006,
+ "loss": 4.156868934631348,
+ "step": 5170
+ },
+ {
+ "epoch": 71.82445414847162,
+ "grad_norm": 0.0176058579236269,
+ "learning_rate": 0.0006,
+ "loss": 4.154374122619629,
+ "step": 5171
+ },
+ {
+ "epoch": 71.83842794759825,
+ "grad_norm": 0.018049508333206177,
+ "learning_rate": 0.0006,
+ "loss": 4.296828269958496,
+ "step": 5172
+ },
+ {
+ "epoch": 71.8524017467249,
+ "grad_norm": 0.019466130062937737,
+ "learning_rate": 0.0006,
+ "loss": 4.19450569152832,
+ "step": 5173
+ },
+ {
+ "epoch": 71.86637554585153,
+ "grad_norm": 0.01838165521621704,
+ "learning_rate": 0.0006,
+ "loss": 4.11812686920166,
+ "step": 5174
+ },
+ {
+ "epoch": 71.88034934497817,
+ "grad_norm": 0.01608051173388958,
+ "learning_rate": 0.0006,
+ "loss": 4.240464210510254,
+ "step": 5175
+ },
+ {
+ "epoch": 71.8943231441048,
+ "grad_norm": 0.015807967633008957,
+ "learning_rate": 0.0006,
+ "loss": 4.275581359863281,
+ "step": 5176
+ },
+ {
+ "epoch": 71.90829694323143,
+ "grad_norm": 0.017366575077176094,
+ "learning_rate": 0.0006,
+ "loss": 4.064750671386719,
+ "step": 5177
+ },
+ {
+ "epoch": 71.92227074235808,
+ "grad_norm": 0.016241202130913734,
+ "learning_rate": 0.0006,
+ "loss": 4.203697681427002,
+ "step": 5178
+ },
+ {
+ "epoch": 71.93624454148471,
+ "grad_norm": 0.019105732440948486,
+ "learning_rate": 0.0006,
+ "loss": 4.140021800994873,
+ "step": 5179
+ },
+ {
+ "epoch": 71.95021834061136,
+ "grad_norm": 0.018492674455046654,
+ "learning_rate": 0.0006,
+ "loss": 4.2534074783325195,
+ "step": 5180
+ },
+ {
+ "epoch": 71.96419213973799,
+ "grad_norm": 0.019301170483231544,
+ "learning_rate": 0.0006,
+ "loss": 4.381392478942871,
+ "step": 5181
+ },
+ {
+ "epoch": 71.97816593886463,
+ "grad_norm": 0.019571911543607712,
+ "learning_rate": 0.0006,
+ "loss": 4.131056785583496,
+ "step": 5182
+ },
+ {
+ "epoch": 71.99213973799127,
+ "grad_norm": 0.01859748177230358,
+ "learning_rate": 0.0006,
+ "loss": 4.278930187225342,
+ "step": 5183
+ },
+ {
+ "epoch": 72.0,
+ "grad_norm": 0.02017994411289692,
+ "learning_rate": 0.0006,
+ "loss": 4.116464614868164,
+ "step": 5184
+ },
+ {
+ "epoch": 72.0,
+ "eval_loss": 4.60784912109375,
+ "eval_runtime": 56.3242,
+ "eval_samples_per_second": 43.356,
+ "eval_steps_per_second": 1.367,
+ "step": 5184
+ },
+ {
+ "epoch": 72.01397379912663,
+ "grad_norm": 0.030090296640992165,
+ "learning_rate": 0.0006,
+ "loss": 4.218588352203369,
+ "step": 5185
+ },
+ {
+ "epoch": 72.02794759825328,
+ "grad_norm": 1.330443024635315,
+ "learning_rate": 0.0006,
+ "loss": 4.417870044708252,
+ "step": 5186
+ },
+ {
+ "epoch": 72.04192139737991,
+ "grad_norm": 0.6991518139839172,
+ "learning_rate": 0.0006,
+ "loss": 8.414234161376953,
+ "step": 5187
+ },
+ {
+ "epoch": 72.05589519650655,
+ "grad_norm": 0.2631438374519348,
+ "learning_rate": 0.0006,
+ "loss": 7.41729736328125,
+ "step": 5188
+ },
+ {
+ "epoch": 72.06986899563319,
+ "grad_norm": 0.12391754984855652,
+ "learning_rate": 0.0006,
+ "loss": 7.209164619445801,
+ "step": 5189
+ },
+ {
+ "epoch": 72.08384279475983,
+ "grad_norm": 0.13519170880317688,
+ "learning_rate": 0.0006,
+ "loss": 6.943539619445801,
+ "step": 5190
+ },
+ {
+ "epoch": 72.09781659388646,
+ "grad_norm": 1.258439302444458,
+ "learning_rate": 0.0006,
+ "loss": 10.267714500427246,
+ "step": 5191
+ },
+ {
+ "epoch": 72.1117903930131,
+ "grad_norm": 1.0296908617019653,
+ "learning_rate": 0.0006,
+ "loss": 9.51768970489502,
+ "step": 5192
+ },
+ {
+ "epoch": 72.12576419213974,
+ "grad_norm": 0.12273016571998596,
+ "learning_rate": 0.0006,
+ "loss": 7.634477615356445,
+ "step": 5193
+ },
+ {
+ "epoch": 72.13973799126637,
+ "grad_norm": 0.06170198693871498,
+ "learning_rate": 0.0006,
+ "loss": 7.592115879058838,
+ "step": 5194
+ },
+ {
+ "epoch": 72.15371179039302,
+ "grad_norm": 0.06295108795166016,
+ "learning_rate": 0.0006,
+ "loss": 7.576314449310303,
+ "step": 5195
+ },
+ {
+ "epoch": 72.16768558951965,
+ "grad_norm": 0.0666993111371994,
+ "learning_rate": 0.0006,
+ "loss": 7.53862190246582,
+ "step": 5196
+ },
+ {
+ "epoch": 72.18165938864628,
+ "grad_norm": 0.052215222269296646,
+ "learning_rate": 0.0006,
+ "loss": 7.502930641174316,
+ "step": 5197
+ },
+ {
+ "epoch": 72.19563318777293,
+ "grad_norm": 0.04978961497545242,
+ "learning_rate": 0.0006,
+ "loss": 7.471630096435547,
+ "step": 5198
+ },
+ {
+ "epoch": 72.20960698689956,
+ "grad_norm": 0.07275046408176422,
+ "learning_rate": 0.0006,
+ "loss": 7.496924877166748,
+ "step": 5199
+ },
+ {
+ "epoch": 72.2235807860262,
+ "grad_norm": 0.04491819068789482,
+ "learning_rate": 0.0006,
+ "loss": 7.424494743347168,
+ "step": 5200
+ },
+ {
+ "epoch": 72.23755458515284,
+ "grad_norm": 0.043718695640563965,
+ "learning_rate": 0.0006,
+ "loss": 7.441090106964111,
+ "step": 5201
+ },
+ {
+ "epoch": 72.25152838427948,
+ "grad_norm": 0.04483497887849808,
+ "learning_rate": 0.0006,
+ "loss": 7.3753862380981445,
+ "step": 5202
+ },
+ {
+ "epoch": 72.26550218340611,
+ "grad_norm": 0.03463944420218468,
+ "learning_rate": 0.0006,
+ "loss": 7.375782012939453,
+ "step": 5203
+ },
+ {
+ "epoch": 72.27947598253274,
+ "grad_norm": 0.0397045724093914,
+ "learning_rate": 0.0006,
+ "loss": 7.358898639678955,
+ "step": 5204
+ },
+ {
+ "epoch": 72.29344978165939,
+ "grad_norm": 0.029339438304305077,
+ "learning_rate": 0.0006,
+ "loss": 7.3774871826171875,
+ "step": 5205
+ },
+ {
+ "epoch": 72.30742358078602,
+ "grad_norm": 0.025469699874520302,
+ "learning_rate": 0.0006,
+ "loss": 7.324203014373779,
+ "step": 5206
+ },
+ {
+ "epoch": 72.32139737991267,
+ "grad_norm": 0.030497785657644272,
+ "learning_rate": 0.0006,
+ "loss": 7.329351902008057,
+ "step": 5207
+ },
+ {
+ "epoch": 72.3353711790393,
+ "grad_norm": 0.03382078558206558,
+ "learning_rate": 0.0006,
+ "loss": 7.289940357208252,
+ "step": 5208
+ },
+ {
+ "epoch": 72.34934497816595,
+ "grad_norm": 0.019869405776262283,
+ "learning_rate": 0.0006,
+ "loss": 7.272582054138184,
+ "step": 5209
+ },
+ {
+ "epoch": 72.36331877729258,
+ "grad_norm": 0.02385753206908703,
+ "learning_rate": 0.0006,
+ "loss": 7.2654266357421875,
+ "step": 5210
+ },
+ {
+ "epoch": 72.37729257641921,
+ "grad_norm": 0.02794312685728073,
+ "learning_rate": 0.0006,
+ "loss": 7.250753402709961,
+ "step": 5211
+ },
+ {
+ "epoch": 72.39126637554585,
+ "grad_norm": 0.02445952221751213,
+ "learning_rate": 0.0006,
+ "loss": 7.293970108032227,
+ "step": 5212
+ },
+ {
+ "epoch": 72.40524017467249,
+ "grad_norm": 0.019660986959934235,
+ "learning_rate": 0.0006,
+ "loss": 7.1948771476745605,
+ "step": 5213
+ },
+ {
+ "epoch": 72.41921397379913,
+ "grad_norm": 0.024110861122608185,
+ "learning_rate": 0.0006,
+ "loss": 7.224499225616455,
+ "step": 5214
+ },
+ {
+ "epoch": 72.43318777292576,
+ "grad_norm": 0.021630609408020973,
+ "learning_rate": 0.0006,
+ "loss": 7.1736884117126465,
+ "step": 5215
+ },
+ {
+ "epoch": 72.44716157205241,
+ "grad_norm": 0.02185880020260811,
+ "learning_rate": 0.0006,
+ "loss": 7.167229652404785,
+ "step": 5216
+ },
+ {
+ "epoch": 72.46113537117904,
+ "grad_norm": 0.018913568928837776,
+ "learning_rate": 0.0006,
+ "loss": 7.196551322937012,
+ "step": 5217
+ },
+ {
+ "epoch": 72.47510917030567,
+ "grad_norm": 0.01802486553788185,
+ "learning_rate": 0.0006,
+ "loss": 7.144286632537842,
+ "step": 5218
+ },
+ {
+ "epoch": 72.48908296943232,
+ "grad_norm": 0.020386485382914543,
+ "learning_rate": 0.0006,
+ "loss": 7.196260929107666,
+ "step": 5219
+ },
+ {
+ "epoch": 72.50305676855895,
+ "grad_norm": 0.020077696070075035,
+ "learning_rate": 0.0006,
+ "loss": 7.138718605041504,
+ "step": 5220
+ },
+ {
+ "epoch": 72.5170305676856,
+ "grad_norm": 0.017224183306097984,
+ "learning_rate": 0.0006,
+ "loss": 7.119478702545166,
+ "step": 5221
+ },
+ {
+ "epoch": 72.53100436681223,
+ "grad_norm": 0.020847434177994728,
+ "learning_rate": 0.0006,
+ "loss": 7.068922996520996,
+ "step": 5222
+ },
+ {
+ "epoch": 72.54497816593886,
+ "grad_norm": 0.021764913573861122,
+ "learning_rate": 0.0006,
+ "loss": 7.071352958679199,
+ "step": 5223
+ },
+ {
+ "epoch": 72.5589519650655,
+ "grad_norm": 0.02697215974330902,
+ "learning_rate": 0.0006,
+ "loss": 7.033344268798828,
+ "step": 5224
+ },
+ {
+ "epoch": 72.57292576419214,
+ "grad_norm": 0.042392998933792114,
+ "learning_rate": 0.0006,
+ "loss": 7.0589919090271,
+ "step": 5225
+ },
+ {
+ "epoch": 72.58689956331878,
+ "grad_norm": 0.06616316735744476,
+ "learning_rate": 0.0006,
+ "loss": 7.060244083404541,
+ "step": 5226
+ },
+ {
+ "epoch": 72.60087336244541,
+ "grad_norm": 0.0828910544514656,
+ "learning_rate": 0.0006,
+ "loss": 7.057173728942871,
+ "step": 5227
+ },
+ {
+ "epoch": 72.61484716157206,
+ "grad_norm": 0.046767015010118484,
+ "learning_rate": 0.0006,
+ "loss": 7.010563850402832,
+ "step": 5228
+ },
+ {
+ "epoch": 72.62882096069869,
+ "grad_norm": 0.03831760957837105,
+ "learning_rate": 0.0006,
+ "loss": 6.9028825759887695,
+ "step": 5229
+ },
+ {
+ "epoch": 72.64279475982532,
+ "grad_norm": 0.05109567195177078,
+ "learning_rate": 0.0006,
+ "loss": 6.923257827758789,
+ "step": 5230
+ },
+ {
+ "epoch": 72.65676855895197,
+ "grad_norm": 0.02897934429347515,
+ "learning_rate": 0.0006,
+ "loss": 6.823014259338379,
+ "step": 5231
+ },
+ {
+ "epoch": 72.6707423580786,
+ "grad_norm": 0.03769192099571228,
+ "learning_rate": 0.0006,
+ "loss": 6.87416410446167,
+ "step": 5232
+ },
+ {
+ "epoch": 72.68471615720524,
+ "grad_norm": 0.03677637130022049,
+ "learning_rate": 0.0006,
+ "loss": 6.7649030685424805,
+ "step": 5233
+ },
+ {
+ "epoch": 72.69868995633188,
+ "grad_norm": 0.0371948778629303,
+ "learning_rate": 0.0006,
+ "loss": 6.749821662902832,
+ "step": 5234
+ },
+ {
+ "epoch": 72.71266375545852,
+ "grad_norm": 0.03057795763015747,
+ "learning_rate": 0.0006,
+ "loss": 6.748945236206055,
+ "step": 5235
+ },
+ {
+ "epoch": 72.72663755458515,
+ "grad_norm": 0.028109321370720863,
+ "learning_rate": 0.0006,
+ "loss": 6.663548946380615,
+ "step": 5236
+ },
+ {
+ "epoch": 72.74061135371178,
+ "grad_norm": 0.02442420832812786,
+ "learning_rate": 0.0006,
+ "loss": 6.636251449584961,
+ "step": 5237
+ },
+ {
+ "epoch": 72.75458515283843,
+ "grad_norm": 0.031841423362493515,
+ "learning_rate": 0.0006,
+ "loss": 6.606883525848389,
+ "step": 5238
+ },
+ {
+ "epoch": 72.76855895196506,
+ "grad_norm": 0.028100458905100822,
+ "learning_rate": 0.0006,
+ "loss": 6.603745937347412,
+ "step": 5239
+ },
+ {
+ "epoch": 72.78253275109171,
+ "grad_norm": 0.03222901374101639,
+ "learning_rate": 0.0006,
+ "loss": 6.463825225830078,
+ "step": 5240
+ },
+ {
+ "epoch": 72.79650655021834,
+ "grad_norm": 0.04766521975398064,
+ "learning_rate": 0.0006,
+ "loss": 6.43734073638916,
+ "step": 5241
+ },
+ {
+ "epoch": 72.81048034934499,
+ "grad_norm": 0.09058783948421478,
+ "learning_rate": 0.0006,
+ "loss": 6.448520183563232,
+ "step": 5242
+ },
+ {
+ "epoch": 72.82445414847162,
+ "grad_norm": 0.08995158225297928,
+ "learning_rate": 0.0006,
+ "loss": 6.522093296051025,
+ "step": 5243
+ },
+ {
+ "epoch": 72.83842794759825,
+ "grad_norm": 0.05491040274500847,
+ "learning_rate": 0.0006,
+ "loss": 6.37354040145874,
+ "step": 5244
+ },
+ {
+ "epoch": 72.8524017467249,
+ "grad_norm": 0.05198514088988304,
+ "learning_rate": 0.0006,
+ "loss": 6.362032413482666,
+ "step": 5245
+ },
+ {
+ "epoch": 72.86637554585153,
+ "grad_norm": 0.05464765429496765,
+ "learning_rate": 0.0006,
+ "loss": 6.326773643493652,
+ "step": 5246
+ },
+ {
+ "epoch": 72.88034934497817,
+ "grad_norm": 0.07337361574172974,
+ "learning_rate": 0.0006,
+ "loss": 6.297532081604004,
+ "step": 5247
+ },
+ {
+ "epoch": 72.8943231441048,
+ "grad_norm": 0.10061230510473251,
+ "learning_rate": 0.0006,
+ "loss": 6.359557628631592,
+ "step": 5248
+ },
+ {
+ "epoch": 72.90829694323143,
+ "grad_norm": 0.10005959123373032,
+ "learning_rate": 0.0006,
+ "loss": 6.422270774841309,
+ "step": 5249
+ },
+ {
+ "epoch": 72.92227074235808,
+ "grad_norm": 0.08989301323890686,
+ "learning_rate": 0.0006,
+ "loss": 6.289371490478516,
+ "step": 5250
+ },
+ {
+ "epoch": 72.93624454148471,
+ "grad_norm": 0.06371799111366272,
+ "learning_rate": 0.0006,
+ "loss": 6.199618339538574,
+ "step": 5251
+ },
+ {
+ "epoch": 72.95021834061136,
+ "grad_norm": 0.05932944267988205,
+ "learning_rate": 0.0006,
+ "loss": 6.111152648925781,
+ "step": 5252
+ },
+ {
+ "epoch": 72.96419213973799,
+ "grad_norm": 0.04735502973198891,
+ "learning_rate": 0.0006,
+ "loss": 6.026026248931885,
+ "step": 5253
+ },
+ {
+ "epoch": 72.97816593886463,
+ "grad_norm": 0.04270688816905022,
+ "learning_rate": 0.0006,
+ "loss": 6.01606559753418,
+ "step": 5254
+ },
+ {
+ "epoch": 72.99213973799127,
+ "grad_norm": 0.03774143010377884,
+ "learning_rate": 0.0006,
+ "loss": 5.8935441970825195,
+ "step": 5255
+ },
+ {
+ "epoch": 73.0,
+ "grad_norm": 0.04313631355762482,
+ "learning_rate": 0.0006,
+ "loss": 5.842061996459961,
+ "step": 5256
+ },
+ {
+ "epoch": 73.0,
+ "eval_loss": 5.9985785484313965,
+ "eval_runtime": 56.3891,
+ "eval_samples_per_second": 43.306,
+ "eval_steps_per_second": 1.366,
+ "step": 5256
+ },
+ {
+ "epoch": 73.01397379912663,
+ "grad_norm": 0.045038655400276184,
+ "learning_rate": 0.0006,
+ "loss": 5.8209662437438965,
+ "step": 5257
+ },
+ {
+ "epoch": 73.02794759825328,
+ "grad_norm": 0.03706910088658333,
+ "learning_rate": 0.0006,
+ "loss": 5.806189060211182,
+ "step": 5258
+ },
+ {
+ "epoch": 73.04192139737991,
+ "grad_norm": 0.04051685333251953,
+ "learning_rate": 0.0006,
+ "loss": 5.706869125366211,
+ "step": 5259
+ },
+ {
+ "epoch": 73.05589519650655,
+ "grad_norm": 0.04849153012037277,
+ "learning_rate": 0.0006,
+ "loss": 5.696385383605957,
+ "step": 5260
+ },
+ {
+ "epoch": 73.06986899563319,
+ "grad_norm": 0.06124403700232506,
+ "learning_rate": 0.0006,
+ "loss": 5.618223667144775,
+ "step": 5261
+ },
+ {
+ "epoch": 73.08384279475983,
+ "grad_norm": 0.07375937700271606,
+ "learning_rate": 0.0006,
+ "loss": 5.631045341491699,
+ "step": 5262
+ },
+ {
+ "epoch": 73.09781659388646,
+ "grad_norm": 0.07413844019174576,
+ "learning_rate": 0.0006,
+ "loss": 5.6532182693481445,
+ "step": 5263
+ },
+ {
+ "epoch": 73.1117903930131,
+ "grad_norm": 0.0633770152926445,
+ "learning_rate": 0.0006,
+ "loss": 5.57209587097168,
+ "step": 5264
+ },
+ {
+ "epoch": 73.12576419213974,
+ "grad_norm": 0.05436766892671585,
+ "learning_rate": 0.0006,
+ "loss": 5.53944730758667,
+ "step": 5265
+ },
+ {
+ "epoch": 73.13973799126637,
+ "grad_norm": 0.04639481008052826,
+ "learning_rate": 0.0006,
+ "loss": 5.502948760986328,
+ "step": 5266
+ },
+ {
+ "epoch": 73.15371179039302,
+ "grad_norm": 0.07556558400392532,
+ "learning_rate": 0.0006,
+ "loss": 5.429069995880127,
+ "step": 5267
+ },
+ {
+ "epoch": 73.16768558951965,
+ "grad_norm": 0.1127929836511612,
+ "learning_rate": 0.0006,
+ "loss": 5.545061111450195,
+ "step": 5268
+ },
+ {
+ "epoch": 73.18165938864628,
+ "grad_norm": 0.07620745897293091,
+ "learning_rate": 0.0006,
+ "loss": 5.507102966308594,
+ "step": 5269
+ },
+ {
+ "epoch": 73.19563318777293,
+ "grad_norm": 0.05076538026332855,
+ "learning_rate": 0.0006,
+ "loss": 5.329682350158691,
+ "step": 5270
+ },
+ {
+ "epoch": 73.20960698689956,
+ "grad_norm": 0.05407547950744629,
+ "learning_rate": 0.0006,
+ "loss": 5.3048553466796875,
+ "step": 5271
+ },
+ {
+ "epoch": 73.2235807860262,
+ "grad_norm": 0.059095583856105804,
+ "learning_rate": 0.0006,
+ "loss": 5.292568683624268,
+ "step": 5272
+ },
+ {
+ "epoch": 73.23755458515284,
+ "grad_norm": 0.06193305924534798,
+ "learning_rate": 0.0006,
+ "loss": 5.1870927810668945,
+ "step": 5273
+ },
+ {
+ "epoch": 73.25152838427948,
+ "grad_norm": 0.062261562794446945,
+ "learning_rate": 0.0006,
+ "loss": 5.256819248199463,
+ "step": 5274
+ },
+ {
+ "epoch": 73.26550218340611,
+ "grad_norm": 0.0732516348361969,
+ "learning_rate": 0.0006,
+ "loss": 5.256464004516602,
+ "step": 5275
+ },
+ {
+ "epoch": 73.27947598253274,
+ "grad_norm": 0.12177979946136475,
+ "learning_rate": 0.0006,
+ "loss": 5.284239768981934,
+ "step": 5276
+ },
+ {
+ "epoch": 73.29344978165939,
+ "grad_norm": 0.10459653288125992,
+ "learning_rate": 0.0006,
+ "loss": 5.269713401794434,
+ "step": 5277
+ },
+ {
+ "epoch": 73.30742358078602,
+ "grad_norm": 0.07854694873094559,
+ "learning_rate": 0.0006,
+ "loss": 5.3245391845703125,
+ "step": 5278
+ },
+ {
+ "epoch": 73.32139737991267,
+ "grad_norm": 0.06030094623565674,
+ "learning_rate": 0.0006,
+ "loss": 5.033702850341797,
+ "step": 5279
+ },
+ {
+ "epoch": 73.3353711790393,
+ "grad_norm": 0.05168461427092552,
+ "learning_rate": 0.0006,
+ "loss": 5.140530109405518,
+ "step": 5280
+ },
+ {
+ "epoch": 73.34934497816595,
+ "grad_norm": 0.041064050048589706,
+ "learning_rate": 0.0006,
+ "loss": 5.1085896492004395,
+ "step": 5281
+ },
+ {
+ "epoch": 73.36331877729258,
+ "grad_norm": 0.03788219019770622,
+ "learning_rate": 0.0006,
+ "loss": 5.045009613037109,
+ "step": 5282
+ },
+ {
+ "epoch": 73.37729257641921,
+ "grad_norm": 0.03851071000099182,
+ "learning_rate": 0.0006,
+ "loss": 5.055008888244629,
+ "step": 5283
+ },
+ {
+ "epoch": 73.39126637554585,
+ "grad_norm": 0.03448539599776268,
+ "learning_rate": 0.0006,
+ "loss": 5.10657262802124,
+ "step": 5284
+ },
+ {
+ "epoch": 73.40524017467249,
+ "grad_norm": 0.026618091389536858,
+ "learning_rate": 0.0006,
+ "loss": 5.0342912673950195,
+ "step": 5285
+ },
+ {
+ "epoch": 73.41921397379913,
+ "grad_norm": 0.03060489147901535,
+ "learning_rate": 0.0006,
+ "loss": 5.022700786590576,
+ "step": 5286
+ },
+ {
+ "epoch": 73.43318777292576,
+ "grad_norm": 0.03362000361084938,
+ "learning_rate": 0.0006,
+ "loss": 4.891584396362305,
+ "step": 5287
+ },
+ {
+ "epoch": 73.44716157205241,
+ "grad_norm": 0.03887505829334259,
+ "learning_rate": 0.0006,
+ "loss": 4.772258758544922,
+ "step": 5288
+ },
+ {
+ "epoch": 73.46113537117904,
+ "grad_norm": 0.07157225161790848,
+ "learning_rate": 0.0006,
+ "loss": 4.997142791748047,
+ "step": 5289
+ },
+ {
+ "epoch": 73.47510917030567,
+ "grad_norm": 0.10180624574422836,
+ "learning_rate": 0.0006,
+ "loss": 4.908722877502441,
+ "step": 5290
+ },
+ {
+ "epoch": 73.48908296943232,
+ "grad_norm": 0.06529122591018677,
+ "learning_rate": 0.0006,
+ "loss": 4.98001766204834,
+ "step": 5291
+ },
+ {
+ "epoch": 73.50305676855895,
+ "grad_norm": 0.055171865969896317,
+ "learning_rate": 0.0006,
+ "loss": 4.877438068389893,
+ "step": 5292
+ },
+ {
+ "epoch": 73.5170305676856,
+ "grad_norm": 0.0318770632147789,
+ "learning_rate": 0.0006,
+ "loss": 4.836934566497803,
+ "step": 5293
+ },
+ {
+ "epoch": 73.53100436681223,
+ "grad_norm": 0.042072564363479614,
+ "learning_rate": 0.0006,
+ "loss": 4.919489860534668,
+ "step": 5294
+ },
+ {
+ "epoch": 73.54497816593886,
+ "grad_norm": 0.029197754338383675,
+ "learning_rate": 0.0006,
+ "loss": 4.805579662322998,
+ "step": 5295
+ },
+ {
+ "epoch": 73.5589519650655,
+ "grad_norm": 0.03249497339129448,
+ "learning_rate": 0.0006,
+ "loss": 4.704806327819824,
+ "step": 5296
+ },
+ {
+ "epoch": 73.57292576419214,
+ "grad_norm": 0.025050945580005646,
+ "learning_rate": 0.0006,
+ "loss": 4.936512470245361,
+ "step": 5297
+ },
+ {
+ "epoch": 73.58689956331878,
+ "grad_norm": 0.029073381796479225,
+ "learning_rate": 0.0006,
+ "loss": 4.757592678070068,
+ "step": 5298
+ },
+ {
+ "epoch": 73.60087336244541,
+ "grad_norm": 0.03465264290571213,
+ "learning_rate": 0.0006,
+ "loss": 4.729528427124023,
+ "step": 5299
+ },
+ {
+ "epoch": 73.61484716157206,
+ "grad_norm": 0.054795749485492706,
+ "learning_rate": 0.0006,
+ "loss": 4.77252197265625,
+ "step": 5300
+ },
+ {
+ "epoch": 73.62882096069869,
+ "grad_norm": 0.18129196763038635,
+ "learning_rate": 0.0006,
+ "loss": 4.901700019836426,
+ "step": 5301
+ },
+ {
+ "epoch": 73.64279475982532,
+ "grad_norm": 0.23274236917495728,
+ "learning_rate": 0.0006,
+ "loss": 5.207432270050049,
+ "step": 5302
+ },
+ {
+ "epoch": 73.65676855895197,
+ "grad_norm": 0.18162772059440613,
+ "learning_rate": 0.0006,
+ "loss": 5.326632976531982,
+ "step": 5303
+ },
+ {
+ "epoch": 73.6707423580786,
+ "grad_norm": 0.11858413368463516,
+ "learning_rate": 0.0006,
+ "loss": 5.3730058670043945,
+ "step": 5304
+ },
+ {
+ "epoch": 73.68471615720524,
+ "grad_norm": 0.06031268090009689,
+ "learning_rate": 0.0006,
+ "loss": 5.330936431884766,
+ "step": 5305
+ },
+ {
+ "epoch": 73.69868995633188,
+ "grad_norm": 0.059304628521203995,
+ "learning_rate": 0.0006,
+ "loss": 5.280514717102051,
+ "step": 5306
+ },
+ {
+ "epoch": 73.71266375545852,
+ "grad_norm": 0.04702131822705269,
+ "learning_rate": 0.0006,
+ "loss": 5.288274765014648,
+ "step": 5307
+ },
+ {
+ "epoch": 73.72663755458515,
+ "grad_norm": 0.037792228162288666,
+ "learning_rate": 0.0006,
+ "loss": 5.173289775848389,
+ "step": 5308
+ },
+ {
+ "epoch": 73.74061135371178,
+ "grad_norm": 0.037763576954603195,
+ "learning_rate": 0.0006,
+ "loss": 5.084990978240967,
+ "step": 5309
+ },
+ {
+ "epoch": 73.75458515283843,
+ "grad_norm": 0.03337346389889717,
+ "learning_rate": 0.0006,
+ "loss": 5.121417045593262,
+ "step": 5310
+ },
+ {
+ "epoch": 73.76855895196506,
+ "grad_norm": 0.03184802085161209,
+ "learning_rate": 0.0006,
+ "loss": 5.208918571472168,
+ "step": 5311
+ },
+ {
+ "epoch": 73.78253275109171,
+ "grad_norm": 0.02832048572599888,
+ "learning_rate": 0.0006,
+ "loss": 4.923543453216553,
+ "step": 5312
+ },
+ {
+ "epoch": 73.79650655021834,
+ "grad_norm": 0.028297102078795433,
+ "learning_rate": 0.0006,
+ "loss": 4.944367408752441,
+ "step": 5313
+ },
+ {
+ "epoch": 73.81048034934499,
+ "grad_norm": 0.027701659128069878,
+ "learning_rate": 0.0006,
+ "loss": 4.858168601989746,
+ "step": 5314
+ },
+ {
+ "epoch": 73.82445414847162,
+ "grad_norm": 0.02214639261364937,
+ "learning_rate": 0.0006,
+ "loss": 4.887052536010742,
+ "step": 5315
+ },
+ {
+ "epoch": 73.83842794759825,
+ "grad_norm": 0.01995093561708927,
+ "learning_rate": 0.0006,
+ "loss": 4.885797500610352,
+ "step": 5316
+ },
+ {
+ "epoch": 73.8524017467249,
+ "grad_norm": 0.020473985001444817,
+ "learning_rate": 0.0006,
+ "loss": 4.917325019836426,
+ "step": 5317
+ },
+ {
+ "epoch": 73.86637554585153,
+ "grad_norm": 0.0192868635058403,
+ "learning_rate": 0.0006,
+ "loss": 4.801746845245361,
+ "step": 5318
+ },
+ {
+ "epoch": 73.88034934497817,
+ "grad_norm": 0.017108267173171043,
+ "learning_rate": 0.0006,
+ "loss": 4.893620491027832,
+ "step": 5319
+ },
+ {
+ "epoch": 73.8943231441048,
+ "grad_norm": 0.01718071475625038,
+ "learning_rate": 0.0006,
+ "loss": 4.940695285797119,
+ "step": 5320
+ },
+ {
+ "epoch": 73.90829694323143,
+ "grad_norm": 0.017821557819843292,
+ "learning_rate": 0.0006,
+ "loss": 4.767613410949707,
+ "step": 5321
+ },
+ {
+ "epoch": 73.92227074235808,
+ "grad_norm": 0.015262950211763382,
+ "learning_rate": 0.0006,
+ "loss": 4.910315036773682,
+ "step": 5322
+ },
+ {
+ "epoch": 73.93624454148471,
+ "grad_norm": 0.01769006811082363,
+ "learning_rate": 0.0006,
+ "loss": 4.7743449211120605,
+ "step": 5323
+ },
+ {
+ "epoch": 73.95021834061136,
+ "grad_norm": 0.014460788108408451,
+ "learning_rate": 0.0006,
+ "loss": 4.743494987487793,
+ "step": 5324
+ },
+ {
+ "epoch": 73.96419213973799,
+ "grad_norm": 0.01403993833810091,
+ "learning_rate": 0.0006,
+ "loss": 4.666640281677246,
+ "step": 5325
+ },
+ {
+ "epoch": 73.97816593886463,
+ "grad_norm": 0.013997809961438179,
+ "learning_rate": 0.0006,
+ "loss": 4.698747158050537,
+ "step": 5326
+ },
+ {
+ "epoch": 73.99213973799127,
+ "grad_norm": 0.014244182966649532,
+ "learning_rate": 0.0006,
+ "loss": 4.745250701904297,
+ "step": 5327
+ },
+ {
+ "epoch": 74.0,
+ "grad_norm": 0.015144186094403267,
+ "learning_rate": 0.0006,
+ "loss": 4.667946815490723,
+ "step": 5328
+ },
+ {
+ "epoch": 74.0,
+ "eval_loss": 5.005440711975098,
+ "eval_runtime": 56.7836,
+ "eval_samples_per_second": 43.005,
+ "eval_steps_per_second": 1.356,
+ "step": 5328
+ },
+ {
+ "epoch": 74.01397379912663,
+ "grad_norm": 0.020384902134537697,
+ "learning_rate": 0.0006,
+ "loss": 4.626955509185791,
+ "step": 5329
+ },
+ {
+ "epoch": 74.02794759825328,
+ "grad_norm": 0.03624720871448517,
+ "learning_rate": 0.0006,
+ "loss": 4.713998317718506,
+ "step": 5330
+ },
+ {
+ "epoch": 74.04192139737991,
+ "grad_norm": 0.05701754242181778,
+ "learning_rate": 0.0006,
+ "loss": 4.717874526977539,
+ "step": 5331
+ },
+ {
+ "epoch": 74.05589519650655,
+ "grad_norm": 0.03942621126770973,
+ "learning_rate": 0.0006,
+ "loss": 4.7026190757751465,
+ "step": 5332
+ },
+ {
+ "epoch": 74.06986899563319,
+ "grad_norm": 0.017803272232413292,
+ "learning_rate": 0.0006,
+ "loss": 4.702548027038574,
+ "step": 5333
+ },
+ {
+ "epoch": 74.08384279475983,
+ "grad_norm": 0.017024677246809006,
+ "learning_rate": 0.0006,
+ "loss": 4.69752311706543,
+ "step": 5334
+ },
+ {
+ "epoch": 74.09781659388646,
+ "grad_norm": 0.016467854380607605,
+ "learning_rate": 0.0006,
+ "loss": 4.624876976013184,
+ "step": 5335
+ },
+ {
+ "epoch": 74.1117903930131,
+ "grad_norm": 0.016569215804338455,
+ "learning_rate": 0.0006,
+ "loss": 4.645040512084961,
+ "step": 5336
+ },
+ {
+ "epoch": 74.12576419213974,
+ "grad_norm": 0.012872067280113697,
+ "learning_rate": 0.0006,
+ "loss": 4.64582633972168,
+ "step": 5337
+ },
+ {
+ "epoch": 74.13973799126637,
+ "grad_norm": 0.015689387917518616,
+ "learning_rate": 0.0006,
+ "loss": 4.596040725708008,
+ "step": 5338
+ },
+ {
+ "epoch": 74.15371179039302,
+ "grad_norm": 0.013082082383334637,
+ "learning_rate": 0.0006,
+ "loss": 4.648033142089844,
+ "step": 5339
+ },
+ {
+ "epoch": 74.16768558951965,
+ "grad_norm": 0.013728172518312931,
+ "learning_rate": 0.0006,
+ "loss": 4.657873153686523,
+ "step": 5340
+ },
+ {
+ "epoch": 74.18165938864628,
+ "grad_norm": 0.013120594434440136,
+ "learning_rate": 0.0006,
+ "loss": 4.696074485778809,
+ "step": 5341
+ },
+ {
+ "epoch": 74.19563318777293,
+ "grad_norm": 0.01431245356798172,
+ "learning_rate": 0.0006,
+ "loss": 4.5679121017456055,
+ "step": 5342
+ },
+ {
+ "epoch": 74.20960698689956,
+ "grad_norm": 0.013994511216878891,
+ "learning_rate": 0.0006,
+ "loss": 4.5236406326293945,
+ "step": 5343
+ },
+ {
+ "epoch": 74.2235807860262,
+ "grad_norm": 0.011971707455813885,
+ "learning_rate": 0.0006,
+ "loss": 4.626547813415527,
+ "step": 5344
+ },
+ {
+ "epoch": 74.23755458515284,
+ "grad_norm": 0.012805287726223469,
+ "learning_rate": 0.0006,
+ "loss": 4.556386947631836,
+ "step": 5345
+ },
+ {
+ "epoch": 74.25152838427948,
+ "grad_norm": 0.011694584973156452,
+ "learning_rate": 0.0006,
+ "loss": 4.656987190246582,
+ "step": 5346
+ },
+ {
+ "epoch": 74.26550218340611,
+ "grad_norm": 0.013396821916103363,
+ "learning_rate": 0.0006,
+ "loss": 4.526467323303223,
+ "step": 5347
+ },
+ {
+ "epoch": 74.27947598253274,
+ "grad_norm": 0.014398031868040562,
+ "learning_rate": 0.0006,
+ "loss": 4.595100402832031,
+ "step": 5348
+ },
+ {
+ "epoch": 74.29344978165939,
+ "grad_norm": 0.02165800705552101,
+ "learning_rate": 0.0006,
+ "loss": 4.460549354553223,
+ "step": 5349
+ },
+ {
+ "epoch": 74.30742358078602,
+ "grad_norm": 0.03276512026786804,
+ "learning_rate": 0.0006,
+ "loss": 4.498189449310303,
+ "step": 5350
+ },
+ {
+ "epoch": 74.32139737991267,
+ "grad_norm": 0.04616224020719528,
+ "learning_rate": 0.0006,
+ "loss": 4.5268659591674805,
+ "step": 5351
+ },
+ {
+ "epoch": 74.3353711790393,
+ "grad_norm": 0.03863786533474922,
+ "learning_rate": 0.0006,
+ "loss": 4.554747104644775,
+ "step": 5352
+ },
+ {
+ "epoch": 74.34934497816595,
+ "grad_norm": 0.01749865710735321,
+ "learning_rate": 0.0006,
+ "loss": 4.6656036376953125,
+ "step": 5353
+ },
+ {
+ "epoch": 74.36331877729258,
+ "grad_norm": 0.021814396604895592,
+ "learning_rate": 0.0006,
+ "loss": 4.607456207275391,
+ "step": 5354
+ },
+ {
+ "epoch": 74.37729257641921,
+ "grad_norm": 0.017018496990203857,
+ "learning_rate": 0.0006,
+ "loss": 4.54205322265625,
+ "step": 5355
+ },
+ {
+ "epoch": 74.39126637554585,
+ "grad_norm": 0.018297340720891953,
+ "learning_rate": 0.0006,
+ "loss": 4.608870506286621,
+ "step": 5356
+ },
+ {
+ "epoch": 74.40524017467249,
+ "grad_norm": 0.01473482046276331,
+ "learning_rate": 0.0006,
+ "loss": 4.384891510009766,
+ "step": 5357
+ },
+ {
+ "epoch": 74.41921397379913,
+ "grad_norm": 0.01579415239393711,
+ "learning_rate": 0.0006,
+ "loss": 4.588859558105469,
+ "step": 5358
+ },
+ {
+ "epoch": 74.43318777292576,
+ "grad_norm": 0.013906026259064674,
+ "learning_rate": 0.0006,
+ "loss": 4.525175094604492,
+ "step": 5359
+ },
+ {
+ "epoch": 74.44716157205241,
+ "grad_norm": 0.014163361862301826,
+ "learning_rate": 0.0006,
+ "loss": 4.495802879333496,
+ "step": 5360
+ },
+ {
+ "epoch": 74.46113537117904,
+ "grad_norm": 0.014009335078299046,
+ "learning_rate": 0.0006,
+ "loss": 4.5960869789123535,
+ "step": 5361
+ },
+ {
+ "epoch": 74.47510917030567,
+ "grad_norm": 0.012765258550643921,
+ "learning_rate": 0.0006,
+ "loss": 4.515678405761719,
+ "step": 5362
+ },
+ {
+ "epoch": 74.48908296943232,
+ "grad_norm": 0.014071918092668056,
+ "learning_rate": 0.0006,
+ "loss": 4.394025802612305,
+ "step": 5363
+ },
+ {
+ "epoch": 74.50305676855895,
+ "grad_norm": 0.01244515273720026,
+ "learning_rate": 0.0006,
+ "loss": 4.576985836029053,
+ "step": 5364
+ },
+ {
+ "epoch": 74.5170305676856,
+ "grad_norm": 0.012975298799574375,
+ "learning_rate": 0.0006,
+ "loss": 4.540657997131348,
+ "step": 5365
+ },
+ {
+ "epoch": 74.53100436681223,
+ "grad_norm": 0.012380947358906269,
+ "learning_rate": 0.0006,
+ "loss": 4.55731725692749,
+ "step": 5366
+ },
+ {
+ "epoch": 74.54497816593886,
+ "grad_norm": 0.012577617540955544,
+ "learning_rate": 0.0006,
+ "loss": 4.500924110412598,
+ "step": 5367
+ },
+ {
+ "epoch": 74.5589519650655,
+ "grad_norm": 0.012089254334568977,
+ "learning_rate": 0.0006,
+ "loss": 4.561795234680176,
+ "step": 5368
+ },
+ {
+ "epoch": 74.57292576419214,
+ "grad_norm": 0.012024011462926865,
+ "learning_rate": 0.0006,
+ "loss": 4.430635452270508,
+ "step": 5369
+ },
+ {
+ "epoch": 74.58689956331878,
+ "grad_norm": 0.012361960485577583,
+ "learning_rate": 0.0006,
+ "loss": 4.4664106369018555,
+ "step": 5370
+ },
+ {
+ "epoch": 74.60087336244541,
+ "grad_norm": 0.012149893678724766,
+ "learning_rate": 0.0006,
+ "loss": 4.522271156311035,
+ "step": 5371
+ },
+ {
+ "epoch": 74.61484716157206,
+ "grad_norm": 0.01166840922087431,
+ "learning_rate": 0.0006,
+ "loss": 4.3870463371276855,
+ "step": 5372
+ },
+ {
+ "epoch": 74.62882096069869,
+ "grad_norm": 0.01160407718271017,
+ "learning_rate": 0.0006,
+ "loss": 4.504255771636963,
+ "step": 5373
+ },
+ {
+ "epoch": 74.64279475982532,
+ "grad_norm": 0.011259177699685097,
+ "learning_rate": 0.0006,
+ "loss": 4.507267951965332,
+ "step": 5374
+ },
+ {
+ "epoch": 74.65676855895197,
+ "grad_norm": 0.012285715900361538,
+ "learning_rate": 0.0006,
+ "loss": 4.472966194152832,
+ "step": 5375
+ },
+ {
+ "epoch": 74.6707423580786,
+ "grad_norm": 0.011214440688490868,
+ "learning_rate": 0.0006,
+ "loss": 4.5840253829956055,
+ "step": 5376
+ },
+ {
+ "epoch": 74.68471615720524,
+ "grad_norm": 0.012394879944622517,
+ "learning_rate": 0.0006,
+ "loss": 4.417830944061279,
+ "step": 5377
+ },
+ {
+ "epoch": 74.69868995633188,
+ "grad_norm": 0.012360049411654472,
+ "learning_rate": 0.0006,
+ "loss": 4.535454750061035,
+ "step": 5378
+ },
+ {
+ "epoch": 74.71266375545852,
+ "grad_norm": 0.011961652897298336,
+ "learning_rate": 0.0006,
+ "loss": 4.403409957885742,
+ "step": 5379
+ },
+ {
+ "epoch": 74.72663755458515,
+ "grad_norm": 0.012078274972736835,
+ "learning_rate": 0.0006,
+ "loss": 4.448366641998291,
+ "step": 5380
+ },
+ {
+ "epoch": 74.74061135371178,
+ "grad_norm": 0.013239333406090736,
+ "learning_rate": 0.0006,
+ "loss": 4.543450355529785,
+ "step": 5381
+ },
+ {
+ "epoch": 74.75458515283843,
+ "grad_norm": 0.01476441603153944,
+ "learning_rate": 0.0006,
+ "loss": 4.488677978515625,
+ "step": 5382
+ },
+ {
+ "epoch": 74.76855895196506,
+ "grad_norm": 0.013968009501695633,
+ "learning_rate": 0.0006,
+ "loss": 4.603639602661133,
+ "step": 5383
+ },
+ {
+ "epoch": 74.78253275109171,
+ "grad_norm": 0.012322458438575268,
+ "learning_rate": 0.0006,
+ "loss": 4.637874603271484,
+ "step": 5384
+ },
+ {
+ "epoch": 74.79650655021834,
+ "grad_norm": 0.013352593407034874,
+ "learning_rate": 0.0006,
+ "loss": 4.474833965301514,
+ "step": 5385
+ },
+ {
+ "epoch": 74.81048034934499,
+ "grad_norm": 0.014671728946268559,
+ "learning_rate": 0.0006,
+ "loss": 4.516847610473633,
+ "step": 5386
+ },
+ {
+ "epoch": 74.82445414847162,
+ "grad_norm": 0.017163321375846863,
+ "learning_rate": 0.0006,
+ "loss": 4.462325096130371,
+ "step": 5387
+ },
+ {
+ "epoch": 74.83842794759825,
+ "grad_norm": 0.02012511156499386,
+ "learning_rate": 0.0006,
+ "loss": 4.355982780456543,
+ "step": 5388
+ },
+ {
+ "epoch": 74.8524017467249,
+ "grad_norm": 0.01881697215139866,
+ "learning_rate": 0.0006,
+ "loss": 4.451120376586914,
+ "step": 5389
+ },
+ {
+ "epoch": 74.86637554585153,
+ "grad_norm": 0.017279407009482384,
+ "learning_rate": 0.0006,
+ "loss": 4.461202621459961,
+ "step": 5390
+ },
+ {
+ "epoch": 74.88034934497817,
+ "grad_norm": 0.015470407903194427,
+ "learning_rate": 0.0006,
+ "loss": 4.4968671798706055,
+ "step": 5391
+ },
+ {
+ "epoch": 74.8943231441048,
+ "grad_norm": 0.01411783043295145,
+ "learning_rate": 0.0006,
+ "loss": 4.558697700500488,
+ "step": 5392
+ },
+ {
+ "epoch": 74.90829694323143,
+ "grad_norm": 0.017178812995553017,
+ "learning_rate": 0.0006,
+ "loss": 4.254153251647949,
+ "step": 5393
+ },
+ {
+ "epoch": 74.92227074235808,
+ "grad_norm": 0.019193004816770554,
+ "learning_rate": 0.0006,
+ "loss": 4.394747257232666,
+ "step": 5394
+ },
+ {
+ "epoch": 74.93624454148471,
+ "grad_norm": 0.018942881375551224,
+ "learning_rate": 0.0006,
+ "loss": 4.459539413452148,
+ "step": 5395
+ },
+ {
+ "epoch": 74.95021834061136,
+ "grad_norm": 0.01641223020851612,
+ "learning_rate": 0.0006,
+ "loss": 4.485820770263672,
+ "step": 5396
+ },
+ {
+ "epoch": 74.96419213973799,
+ "grad_norm": 0.014485753141343594,
+ "learning_rate": 0.0006,
+ "loss": 4.5068583488464355,
+ "step": 5397
+ },
+ {
+ "epoch": 74.97816593886463,
+ "grad_norm": 0.018242299556732178,
+ "learning_rate": 0.0006,
+ "loss": 4.508835792541504,
+ "step": 5398
+ },
+ {
+ "epoch": 74.99213973799127,
+ "grad_norm": 0.016491873189806938,
+ "learning_rate": 0.0006,
+ "loss": 4.4760870933532715,
+ "step": 5399
+ },
+ {
+ "epoch": 75.0,
+ "grad_norm": 0.014543353579938412,
+ "learning_rate": 0.0006,
+ "loss": 4.540217399597168,
+ "step": 5400
+ },
+ {
+ "epoch": 75.0,
+ "eval_loss": 4.7617573738098145,
+ "eval_runtime": 57.321,
+ "eval_samples_per_second": 42.602,
+ "eval_steps_per_second": 1.343,
+ "step": 5400
+ },
+ {
+ "epoch": 75.01397379912663,
+ "grad_norm": 0.014417114667594433,
+ "learning_rate": 0.0006,
+ "loss": 4.332956790924072,
+ "step": 5401
+ },
+ {
+ "epoch": 75.02794759825328,
+ "grad_norm": 0.01607639715075493,
+ "learning_rate": 0.0006,
+ "loss": 4.352453708648682,
+ "step": 5402
+ },
+ {
+ "epoch": 75.04192139737991,
+ "grad_norm": 0.015438318252563477,
+ "learning_rate": 0.0006,
+ "loss": 4.387139320373535,
+ "step": 5403
+ },
+ {
+ "epoch": 75.05589519650655,
+ "grad_norm": 0.015539007261395454,
+ "learning_rate": 0.0006,
+ "loss": 4.453171730041504,
+ "step": 5404
+ },
+ {
+ "epoch": 75.06986899563319,
+ "grad_norm": 0.017361903563141823,
+ "learning_rate": 0.0006,
+ "loss": 4.512143135070801,
+ "step": 5405
+ },
+ {
+ "epoch": 75.08384279475983,
+ "grad_norm": 0.01815599389374256,
+ "learning_rate": 0.0006,
+ "loss": 4.508997440338135,
+ "step": 5406
+ },
+ {
+ "epoch": 75.09781659388646,
+ "grad_norm": 0.01620607078075409,
+ "learning_rate": 0.0006,
+ "loss": 4.322444915771484,
+ "step": 5407
+ },
+ {
+ "epoch": 75.1117903930131,
+ "grad_norm": 0.014773265458643436,
+ "learning_rate": 0.0006,
+ "loss": 4.359251976013184,
+ "step": 5408
+ },
+ {
+ "epoch": 75.12576419213974,
+ "grad_norm": 0.014099023304879665,
+ "learning_rate": 0.0006,
+ "loss": 4.313380241394043,
+ "step": 5409
+ },
+ {
+ "epoch": 75.13973799126637,
+ "grad_norm": 0.014387735165655613,
+ "learning_rate": 0.0006,
+ "loss": 4.435023307800293,
+ "step": 5410
+ },
+ {
+ "epoch": 75.15371179039302,
+ "grad_norm": 0.014886162243783474,
+ "learning_rate": 0.0006,
+ "loss": 4.335939407348633,
+ "step": 5411
+ },
+ {
+ "epoch": 75.16768558951965,
+ "grad_norm": 0.015766102820634842,
+ "learning_rate": 0.0006,
+ "loss": 4.408674240112305,
+ "step": 5412
+ },
+ {
+ "epoch": 75.18165938864628,
+ "grad_norm": 0.016383284702897072,
+ "learning_rate": 0.0006,
+ "loss": 4.385453224182129,
+ "step": 5413
+ },
+ {
+ "epoch": 75.19563318777293,
+ "grad_norm": 0.013535384088754654,
+ "learning_rate": 0.0006,
+ "loss": 4.292785167694092,
+ "step": 5414
+ },
+ {
+ "epoch": 75.20960698689956,
+ "grad_norm": 0.01354036945849657,
+ "learning_rate": 0.0006,
+ "loss": 4.306252479553223,
+ "step": 5415
+ },
+ {
+ "epoch": 75.2235807860262,
+ "grad_norm": 0.015818078070878983,
+ "learning_rate": 0.0006,
+ "loss": 4.425487995147705,
+ "step": 5416
+ },
+ {
+ "epoch": 75.23755458515284,
+ "grad_norm": 0.014363568276166916,
+ "learning_rate": 0.0006,
+ "loss": 4.395087242126465,
+ "step": 5417
+ },
+ {
+ "epoch": 75.25152838427948,
+ "grad_norm": 0.013895470649003983,
+ "learning_rate": 0.0006,
+ "loss": 4.364090442657471,
+ "step": 5418
+ },
+ {
+ "epoch": 75.26550218340611,
+ "grad_norm": 0.014178668148815632,
+ "learning_rate": 0.0006,
+ "loss": 4.44619083404541,
+ "step": 5419
+ },
+ {
+ "epoch": 75.27947598253274,
+ "grad_norm": 0.0152514623478055,
+ "learning_rate": 0.0006,
+ "loss": 4.348445892333984,
+ "step": 5420
+ },
+ {
+ "epoch": 75.29344978165939,
+ "grad_norm": 0.015397852286696434,
+ "learning_rate": 0.0006,
+ "loss": 4.476832389831543,
+ "step": 5421
+ },
+ {
+ "epoch": 75.30742358078602,
+ "grad_norm": 0.018009863793849945,
+ "learning_rate": 0.0006,
+ "loss": 4.431162357330322,
+ "step": 5422
+ },
+ {
+ "epoch": 75.32139737991267,
+ "grad_norm": 0.017038367688655853,
+ "learning_rate": 0.0006,
+ "loss": 4.474787712097168,
+ "step": 5423
+ },
+ {
+ "epoch": 75.3353711790393,
+ "grad_norm": 0.013763288035988808,
+ "learning_rate": 0.0006,
+ "loss": 4.470419883728027,
+ "step": 5424
+ },
+ {
+ "epoch": 75.34934497816595,
+ "grad_norm": 0.013972110114991665,
+ "learning_rate": 0.0006,
+ "loss": 4.373147964477539,
+ "step": 5425
+ },
+ {
+ "epoch": 75.36331877729258,
+ "grad_norm": 0.017197875306010246,
+ "learning_rate": 0.0006,
+ "loss": 4.494963645935059,
+ "step": 5426
+ },
+ {
+ "epoch": 75.37729257641921,
+ "grad_norm": 0.01687047816812992,
+ "learning_rate": 0.0006,
+ "loss": 4.435934066772461,
+ "step": 5427
+ },
+ {
+ "epoch": 75.39126637554585,
+ "grad_norm": 0.01578272506594658,
+ "learning_rate": 0.0006,
+ "loss": 4.37918758392334,
+ "step": 5428
+ },
+ {
+ "epoch": 75.40524017467249,
+ "grad_norm": 0.015956247225403786,
+ "learning_rate": 0.0006,
+ "loss": 4.455668926239014,
+ "step": 5429
+ },
+ {
+ "epoch": 75.41921397379913,
+ "grad_norm": 0.013495155610144138,
+ "learning_rate": 0.0006,
+ "loss": 4.4271039962768555,
+ "step": 5430
+ },
+ {
+ "epoch": 75.43318777292576,
+ "grad_norm": 0.012851621024310589,
+ "learning_rate": 0.0006,
+ "loss": 4.213107109069824,
+ "step": 5431
+ },
+ {
+ "epoch": 75.44716157205241,
+ "grad_norm": 0.013351158238947392,
+ "learning_rate": 0.0006,
+ "loss": 4.3070783615112305,
+ "step": 5432
+ },
+ {
+ "epoch": 75.46113537117904,
+ "grad_norm": 0.014765217900276184,
+ "learning_rate": 0.0006,
+ "loss": 4.385021209716797,
+ "step": 5433
+ },
+ {
+ "epoch": 75.47510917030567,
+ "grad_norm": 0.016091614961624146,
+ "learning_rate": 0.0006,
+ "loss": 4.312644004821777,
+ "step": 5434
+ },
+ {
+ "epoch": 75.48908296943232,
+ "grad_norm": 0.01612013578414917,
+ "learning_rate": 0.0006,
+ "loss": 4.4482011795043945,
+ "step": 5435
+ },
+ {
+ "epoch": 75.50305676855895,
+ "grad_norm": 0.013157603330910206,
+ "learning_rate": 0.0006,
+ "loss": 4.427262306213379,
+ "step": 5436
+ },
+ {
+ "epoch": 75.5170305676856,
+ "grad_norm": 0.016039809212088585,
+ "learning_rate": 0.0006,
+ "loss": 4.396635055541992,
+ "step": 5437
+ },
+ {
+ "epoch": 75.53100436681223,
+ "grad_norm": 0.01794283092021942,
+ "learning_rate": 0.0006,
+ "loss": 4.476870059967041,
+ "step": 5438
+ },
+ {
+ "epoch": 75.54497816593886,
+ "grad_norm": 0.017030486837029457,
+ "learning_rate": 0.0006,
+ "loss": 4.361156463623047,
+ "step": 5439
+ },
+ {
+ "epoch": 75.5589519650655,
+ "grad_norm": 0.017915161326527596,
+ "learning_rate": 0.0006,
+ "loss": 4.286147117614746,
+ "step": 5440
+ },
+ {
+ "epoch": 75.57292576419214,
+ "grad_norm": 0.017514433711767197,
+ "learning_rate": 0.0006,
+ "loss": 4.394091606140137,
+ "step": 5441
+ },
+ {
+ "epoch": 75.58689956331878,
+ "grad_norm": 0.015128589235246181,
+ "learning_rate": 0.0006,
+ "loss": 4.383126735687256,
+ "step": 5442
+ },
+ {
+ "epoch": 75.60087336244541,
+ "grad_norm": 0.0202616099268198,
+ "learning_rate": 0.0006,
+ "loss": 4.336188316345215,
+ "step": 5443
+ },
+ {
+ "epoch": 75.61484716157206,
+ "grad_norm": 0.019978471100330353,
+ "learning_rate": 0.0006,
+ "loss": 4.461453437805176,
+ "step": 5444
+ },
+ {
+ "epoch": 75.62882096069869,
+ "grad_norm": 0.01738741435110569,
+ "learning_rate": 0.0006,
+ "loss": 4.371197700500488,
+ "step": 5445
+ },
+ {
+ "epoch": 75.64279475982532,
+ "grad_norm": 0.016157809644937515,
+ "learning_rate": 0.0006,
+ "loss": 4.3312835693359375,
+ "step": 5446
+ },
+ {
+ "epoch": 75.65676855895197,
+ "grad_norm": 0.016394613310694695,
+ "learning_rate": 0.0006,
+ "loss": 4.380190849304199,
+ "step": 5447
+ },
+ {
+ "epoch": 75.6707423580786,
+ "grad_norm": 0.014727243222296238,
+ "learning_rate": 0.0006,
+ "loss": 4.308028221130371,
+ "step": 5448
+ },
+ {
+ "epoch": 75.68471615720524,
+ "grad_norm": 0.01695958711206913,
+ "learning_rate": 0.0006,
+ "loss": 4.399503707885742,
+ "step": 5449
+ },
+ {
+ "epoch": 75.69868995633188,
+ "grad_norm": 0.015720795840024948,
+ "learning_rate": 0.0006,
+ "loss": 4.342249870300293,
+ "step": 5450
+ },
+ {
+ "epoch": 75.71266375545852,
+ "grad_norm": 0.01574772410094738,
+ "learning_rate": 0.0006,
+ "loss": 4.331503868103027,
+ "step": 5451
+ },
+ {
+ "epoch": 75.72663755458515,
+ "grad_norm": 0.014812648296356201,
+ "learning_rate": 0.0006,
+ "loss": 4.332413673400879,
+ "step": 5452
+ },
+ {
+ "epoch": 75.74061135371178,
+ "grad_norm": 0.012850426137447357,
+ "learning_rate": 0.0006,
+ "loss": 4.2884063720703125,
+ "step": 5453
+ },
+ {
+ "epoch": 75.75458515283843,
+ "grad_norm": 0.012876970693469048,
+ "learning_rate": 0.0006,
+ "loss": 4.399855613708496,
+ "step": 5454
+ },
+ {
+ "epoch": 75.76855895196506,
+ "grad_norm": 0.012958863750100136,
+ "learning_rate": 0.0006,
+ "loss": 4.411185264587402,
+ "step": 5455
+ },
+ {
+ "epoch": 75.78253275109171,
+ "grad_norm": 0.012197135016322136,
+ "learning_rate": 0.0006,
+ "loss": 4.272603988647461,
+ "step": 5456
+ },
+ {
+ "epoch": 75.79650655021834,
+ "grad_norm": 0.013483566232025623,
+ "learning_rate": 0.0006,
+ "loss": 4.3564348220825195,
+ "step": 5457
+ },
+ {
+ "epoch": 75.81048034934499,
+ "grad_norm": 0.012365429662168026,
+ "learning_rate": 0.0006,
+ "loss": 4.354825973510742,
+ "step": 5458
+ },
+ {
+ "epoch": 75.82445414847162,
+ "grad_norm": 0.012238141149282455,
+ "learning_rate": 0.0006,
+ "loss": 4.440438747406006,
+ "step": 5459
+ },
+ {
+ "epoch": 75.83842794759825,
+ "grad_norm": 0.013380305841565132,
+ "learning_rate": 0.0006,
+ "loss": 4.2813215255737305,
+ "step": 5460
+ },
+ {
+ "epoch": 75.8524017467249,
+ "grad_norm": 0.01443779468536377,
+ "learning_rate": 0.0006,
+ "loss": 4.289078235626221,
+ "step": 5461
+ },
+ {
+ "epoch": 75.86637554585153,
+ "grad_norm": 0.012646574527025223,
+ "learning_rate": 0.0006,
+ "loss": 4.374177932739258,
+ "step": 5462
+ },
+ {
+ "epoch": 75.88034934497817,
+ "grad_norm": 0.012403388507664204,
+ "learning_rate": 0.0006,
+ "loss": 4.343016147613525,
+ "step": 5463
+ },
+ {
+ "epoch": 75.8943231441048,
+ "grad_norm": 0.01319252047687769,
+ "learning_rate": 0.0006,
+ "loss": 4.309957027435303,
+ "step": 5464
+ },
+ {
+ "epoch": 75.90829694323143,
+ "grad_norm": 0.012597428634762764,
+ "learning_rate": 0.0006,
+ "loss": 4.403654098510742,
+ "step": 5465
+ },
+ {
+ "epoch": 75.92227074235808,
+ "grad_norm": 0.012724226340651512,
+ "learning_rate": 0.0006,
+ "loss": 4.326869010925293,
+ "step": 5466
+ },
+ {
+ "epoch": 75.93624454148471,
+ "grad_norm": 0.012698772363364697,
+ "learning_rate": 0.0006,
+ "loss": 4.268430233001709,
+ "step": 5467
+ },
+ {
+ "epoch": 75.95021834061136,
+ "grad_norm": 0.01376490481197834,
+ "learning_rate": 0.0006,
+ "loss": 4.42446231842041,
+ "step": 5468
+ },
+ {
+ "epoch": 75.96419213973799,
+ "grad_norm": 0.013175144791603088,
+ "learning_rate": 0.0006,
+ "loss": 4.374675750732422,
+ "step": 5469
+ },
+ {
+ "epoch": 75.97816593886463,
+ "grad_norm": 0.012731322087347507,
+ "learning_rate": 0.0006,
+ "loss": 4.372966766357422,
+ "step": 5470
+ },
+ {
+ "epoch": 75.99213973799127,
+ "grad_norm": 0.01360505260527134,
+ "learning_rate": 0.0006,
+ "loss": 4.3095903396606445,
+ "step": 5471
+ },
+ {
+ "epoch": 76.0,
+ "grad_norm": 0.014011901803314686,
+ "learning_rate": 0.0006,
+ "loss": 4.26954460144043,
+ "step": 5472
+ },
+ {
+ "epoch": 76.0,
+ "eval_loss": 4.689640998840332,
+ "eval_runtime": 56.9095,
+ "eval_samples_per_second": 42.91,
+ "eval_steps_per_second": 1.353,
+ "step": 5472
+ },
+ {
+ "epoch": 76.01397379912663,
+ "grad_norm": 0.015351055189967155,
+ "learning_rate": 0.0006,
+ "loss": 4.348268985748291,
+ "step": 5473
+ },
+ {
+ "epoch": 76.02794759825328,
+ "grad_norm": 0.017164217308163643,
+ "learning_rate": 0.0006,
+ "loss": 4.244204998016357,
+ "step": 5474
+ },
+ {
+ "epoch": 76.04192139737991,
+ "grad_norm": 0.021624932065606117,
+ "learning_rate": 0.0006,
+ "loss": 4.311168670654297,
+ "step": 5475
+ },
+ {
+ "epoch": 76.05589519650655,
+ "grad_norm": 0.028712468221783638,
+ "learning_rate": 0.0006,
+ "loss": 4.322287082672119,
+ "step": 5476
+ },
+ {
+ "epoch": 76.06986899563319,
+ "grad_norm": 0.027915487065911293,
+ "learning_rate": 0.0006,
+ "loss": 4.381207466125488,
+ "step": 5477
+ },
+ {
+ "epoch": 76.08384279475983,
+ "grad_norm": 0.017955303192138672,
+ "learning_rate": 0.0006,
+ "loss": 4.2635016441345215,
+ "step": 5478
+ },
+ {
+ "epoch": 76.09781659388646,
+ "grad_norm": 0.019113514572381973,
+ "learning_rate": 0.0006,
+ "loss": 4.399471282958984,
+ "step": 5479
+ },
+ {
+ "epoch": 76.1117903930131,
+ "grad_norm": 0.017502542585134506,
+ "learning_rate": 0.0006,
+ "loss": 4.321204662322998,
+ "step": 5480
+ },
+ {
+ "epoch": 76.12576419213974,
+ "grad_norm": 0.01852525770664215,
+ "learning_rate": 0.0006,
+ "loss": 4.217884063720703,
+ "step": 5481
+ },
+ {
+ "epoch": 76.13973799126637,
+ "grad_norm": 0.019836828112602234,
+ "learning_rate": 0.0006,
+ "loss": 4.278532028198242,
+ "step": 5482
+ },
+ {
+ "epoch": 76.15371179039302,
+ "grad_norm": 0.02101975679397583,
+ "learning_rate": 0.0006,
+ "loss": 4.246720314025879,
+ "step": 5483
+ },
+ {
+ "epoch": 76.16768558951965,
+ "grad_norm": 0.021202990785241127,
+ "learning_rate": 0.0006,
+ "loss": 4.517517566680908,
+ "step": 5484
+ },
+ {
+ "epoch": 76.18165938864628,
+ "grad_norm": 0.021428769454360008,
+ "learning_rate": 0.0006,
+ "loss": 4.3373517990112305,
+ "step": 5485
+ },
+ {
+ "epoch": 76.19563318777293,
+ "grad_norm": 0.018263190984725952,
+ "learning_rate": 0.0006,
+ "loss": 4.402856826782227,
+ "step": 5486
+ },
+ {
+ "epoch": 76.20960698689956,
+ "grad_norm": 0.019025536254048347,
+ "learning_rate": 0.0006,
+ "loss": 4.314227104187012,
+ "step": 5487
+ },
+ {
+ "epoch": 76.2235807860262,
+ "grad_norm": 0.020918630063533783,
+ "learning_rate": 0.0006,
+ "loss": 4.454369068145752,
+ "step": 5488
+ },
+ {
+ "epoch": 76.23755458515284,
+ "grad_norm": 0.019307617098093033,
+ "learning_rate": 0.0006,
+ "loss": 4.247509002685547,
+ "step": 5489
+ },
+ {
+ "epoch": 76.25152838427948,
+ "grad_norm": 0.014956331811845303,
+ "learning_rate": 0.0006,
+ "loss": 4.365175247192383,
+ "step": 5490
+ },
+ {
+ "epoch": 76.26550218340611,
+ "grad_norm": 0.017213059589266777,
+ "learning_rate": 0.0006,
+ "loss": 4.384005069732666,
+ "step": 5491
+ },
+ {
+ "epoch": 76.27947598253274,
+ "grad_norm": 0.017575597390532494,
+ "learning_rate": 0.0006,
+ "loss": 4.41934871673584,
+ "step": 5492
+ },
+ {
+ "epoch": 76.29344978165939,
+ "grad_norm": 0.015879256650805473,
+ "learning_rate": 0.0006,
+ "loss": 4.26675271987915,
+ "step": 5493
+ },
+ {
+ "epoch": 76.30742358078602,
+ "grad_norm": 0.01564522460103035,
+ "learning_rate": 0.0006,
+ "loss": 4.277339935302734,
+ "step": 5494
+ },
+ {
+ "epoch": 76.32139737991267,
+ "grad_norm": 0.016414586454629898,
+ "learning_rate": 0.0006,
+ "loss": 4.4138593673706055,
+ "step": 5495
+ },
+ {
+ "epoch": 76.3353711790393,
+ "grad_norm": 0.014732337556779385,
+ "learning_rate": 0.0006,
+ "loss": 4.281006336212158,
+ "step": 5496
+ },
+ {
+ "epoch": 76.34934497816595,
+ "grad_norm": 0.015461455099284649,
+ "learning_rate": 0.0006,
+ "loss": 4.263422966003418,
+ "step": 5497
+ },
+ {
+ "epoch": 76.36331877729258,
+ "grad_norm": 0.014529814012348652,
+ "learning_rate": 0.0006,
+ "loss": 4.36165189743042,
+ "step": 5498
+ },
+ {
+ "epoch": 76.37729257641921,
+ "grad_norm": 0.01469483319669962,
+ "learning_rate": 0.0006,
+ "loss": 4.309642791748047,
+ "step": 5499
+ },
+ {
+ "epoch": 76.39126637554585,
+ "grad_norm": 0.016490574926137924,
+ "learning_rate": 0.0006,
+ "loss": 4.288150787353516,
+ "step": 5500
+ },
+ {
+ "epoch": 76.40524017467249,
+ "grad_norm": 0.014593565836548805,
+ "learning_rate": 0.0006,
+ "loss": 4.271614074707031,
+ "step": 5501
+ },
+ {
+ "epoch": 76.41921397379913,
+ "grad_norm": 0.012796535156667233,
+ "learning_rate": 0.0006,
+ "loss": 4.420042037963867,
+ "step": 5502
+ },
+ {
+ "epoch": 76.43318777292576,
+ "grad_norm": 0.016086025163531303,
+ "learning_rate": 0.0006,
+ "loss": 4.289943695068359,
+ "step": 5503
+ },
+ {
+ "epoch": 76.44716157205241,
+ "grad_norm": 0.015439470298588276,
+ "learning_rate": 0.0006,
+ "loss": 4.3023271560668945,
+ "step": 5504
+ },
+ {
+ "epoch": 76.46113537117904,
+ "grad_norm": 0.015066367574036121,
+ "learning_rate": 0.0006,
+ "loss": 4.197081565856934,
+ "step": 5505
+ },
+ {
+ "epoch": 76.47510917030567,
+ "grad_norm": 0.016653263941407204,
+ "learning_rate": 0.0006,
+ "loss": 4.333785057067871,
+ "step": 5506
+ },
+ {
+ "epoch": 76.48908296943232,
+ "grad_norm": 0.014791973866522312,
+ "learning_rate": 0.0006,
+ "loss": 4.323811054229736,
+ "step": 5507
+ },
+ {
+ "epoch": 76.50305676855895,
+ "grad_norm": 0.014694827608764172,
+ "learning_rate": 0.0006,
+ "loss": 4.195852279663086,
+ "step": 5508
+ },
+ {
+ "epoch": 76.5170305676856,
+ "grad_norm": 0.015120396390557289,
+ "learning_rate": 0.0006,
+ "loss": 4.42379903793335,
+ "step": 5509
+ },
+ {
+ "epoch": 76.53100436681223,
+ "grad_norm": 0.014584025368094444,
+ "learning_rate": 0.0006,
+ "loss": 4.2702531814575195,
+ "step": 5510
+ },
+ {
+ "epoch": 76.54497816593886,
+ "grad_norm": 0.013981613330543041,
+ "learning_rate": 0.0006,
+ "loss": 4.383029937744141,
+ "step": 5511
+ },
+ {
+ "epoch": 76.5589519650655,
+ "grad_norm": 0.012930642813444138,
+ "learning_rate": 0.0006,
+ "loss": 4.296136856079102,
+ "step": 5512
+ },
+ {
+ "epoch": 76.57292576419214,
+ "grad_norm": 0.013060989789664745,
+ "learning_rate": 0.0006,
+ "loss": 4.346380233764648,
+ "step": 5513
+ },
+ {
+ "epoch": 76.58689956331878,
+ "grad_norm": 0.013209798373281956,
+ "learning_rate": 0.0006,
+ "loss": 4.316067218780518,
+ "step": 5514
+ },
+ {
+ "epoch": 76.60087336244541,
+ "grad_norm": 0.012423216365277767,
+ "learning_rate": 0.0006,
+ "loss": 4.370274543762207,
+ "step": 5515
+ },
+ {
+ "epoch": 76.61484716157206,
+ "grad_norm": 0.013205411843955517,
+ "learning_rate": 0.0006,
+ "loss": 4.429969310760498,
+ "step": 5516
+ },
+ {
+ "epoch": 76.62882096069869,
+ "grad_norm": 0.013049827888607979,
+ "learning_rate": 0.0006,
+ "loss": 4.378549098968506,
+ "step": 5517
+ },
+ {
+ "epoch": 76.64279475982532,
+ "grad_norm": 0.012075880542397499,
+ "learning_rate": 0.0006,
+ "loss": 4.348293304443359,
+ "step": 5518
+ },
+ {
+ "epoch": 76.65676855895197,
+ "grad_norm": 0.014932847581803799,
+ "learning_rate": 0.0006,
+ "loss": 4.282538890838623,
+ "step": 5519
+ },
+ {
+ "epoch": 76.6707423580786,
+ "grad_norm": 0.015615998767316341,
+ "learning_rate": 0.0006,
+ "loss": 4.274641036987305,
+ "step": 5520
+ },
+ {
+ "epoch": 76.68471615720524,
+ "grad_norm": 0.01424513477832079,
+ "learning_rate": 0.0006,
+ "loss": 4.243166923522949,
+ "step": 5521
+ },
+ {
+ "epoch": 76.69868995633188,
+ "grad_norm": 0.012775546871125698,
+ "learning_rate": 0.0006,
+ "loss": 4.335611820220947,
+ "step": 5522
+ },
+ {
+ "epoch": 76.71266375545852,
+ "grad_norm": 0.014461303129792213,
+ "learning_rate": 0.0006,
+ "loss": 4.246976375579834,
+ "step": 5523
+ },
+ {
+ "epoch": 76.72663755458515,
+ "grad_norm": 0.0145436255261302,
+ "learning_rate": 0.0006,
+ "loss": 4.275895118713379,
+ "step": 5524
+ },
+ {
+ "epoch": 76.74061135371178,
+ "grad_norm": 0.014055085368454456,
+ "learning_rate": 0.0006,
+ "loss": 4.331108093261719,
+ "step": 5525
+ },
+ {
+ "epoch": 76.75458515283843,
+ "grad_norm": 0.011933894827961922,
+ "learning_rate": 0.0006,
+ "loss": 4.293721675872803,
+ "step": 5526
+ },
+ {
+ "epoch": 76.76855895196506,
+ "grad_norm": 0.01277392078191042,
+ "learning_rate": 0.0006,
+ "loss": 4.312374114990234,
+ "step": 5527
+ },
+ {
+ "epoch": 76.78253275109171,
+ "grad_norm": 0.014227936044335365,
+ "learning_rate": 0.0006,
+ "loss": 4.238729000091553,
+ "step": 5528
+ },
+ {
+ "epoch": 76.79650655021834,
+ "grad_norm": 0.01488024927675724,
+ "learning_rate": 0.0006,
+ "loss": 4.24587345123291,
+ "step": 5529
+ },
+ {
+ "epoch": 76.81048034934499,
+ "grad_norm": 0.015341250225901604,
+ "learning_rate": 0.0006,
+ "loss": 4.257378578186035,
+ "step": 5530
+ },
+ {
+ "epoch": 76.82445414847162,
+ "grad_norm": 0.015179437585175037,
+ "learning_rate": 0.0006,
+ "loss": 4.2812910079956055,
+ "step": 5531
+ },
+ {
+ "epoch": 76.83842794759825,
+ "grad_norm": 0.012814724817872047,
+ "learning_rate": 0.0006,
+ "loss": 4.2309465408325195,
+ "step": 5532
+ },
+ {
+ "epoch": 76.8524017467249,
+ "grad_norm": 0.013810716569423676,
+ "learning_rate": 0.0006,
+ "loss": 4.325292110443115,
+ "step": 5533
+ },
+ {
+ "epoch": 76.86637554585153,
+ "grad_norm": 0.017468232661485672,
+ "learning_rate": 0.0006,
+ "loss": 4.243788719177246,
+ "step": 5534
+ },
+ {
+ "epoch": 76.88034934497817,
+ "grad_norm": 0.01739857904613018,
+ "learning_rate": 0.0006,
+ "loss": 4.345390796661377,
+ "step": 5535
+ },
+ {
+ "epoch": 76.8943231441048,
+ "grad_norm": 0.015329061076045036,
+ "learning_rate": 0.0006,
+ "loss": 4.415142059326172,
+ "step": 5536
+ },
+ {
+ "epoch": 76.90829694323143,
+ "grad_norm": 0.01598426327109337,
+ "learning_rate": 0.0006,
+ "loss": 4.299558639526367,
+ "step": 5537
+ },
+ {
+ "epoch": 76.92227074235808,
+ "grad_norm": 0.01621454581618309,
+ "learning_rate": 0.0006,
+ "loss": 4.284354209899902,
+ "step": 5538
+ },
+ {
+ "epoch": 76.93624454148471,
+ "grad_norm": 0.015956968069076538,
+ "learning_rate": 0.0006,
+ "loss": 4.293837547302246,
+ "step": 5539
+ },
+ {
+ "epoch": 76.95021834061136,
+ "grad_norm": 0.01776493526995182,
+ "learning_rate": 0.0006,
+ "loss": 4.229729175567627,
+ "step": 5540
+ },
+ {
+ "epoch": 76.96419213973799,
+ "grad_norm": 0.015317477285861969,
+ "learning_rate": 0.0006,
+ "loss": 4.3328447341918945,
+ "step": 5541
+ },
+ {
+ "epoch": 76.97816593886463,
+ "grad_norm": 0.015140251256525517,
+ "learning_rate": 0.0006,
+ "loss": 4.310840606689453,
+ "step": 5542
+ },
+ {
+ "epoch": 76.99213973799127,
+ "grad_norm": 0.015326812863349915,
+ "learning_rate": 0.0006,
+ "loss": 4.230832099914551,
+ "step": 5543
+ },
+ {
+ "epoch": 77.0,
+ "grad_norm": 0.01607763208448887,
+ "learning_rate": 0.0006,
+ "loss": 4.2236328125,
+ "step": 5544
+ },
+ {
+ "epoch": 77.0,
+ "eval_loss": 4.662221431732178,
+ "eval_runtime": 56.5194,
+ "eval_samples_per_second": 43.206,
+ "eval_steps_per_second": 1.362,
+ "step": 5544
+ },
+ {
+ "epoch": 77.01397379912663,
+ "grad_norm": 0.014839510433375835,
+ "learning_rate": 0.0006,
+ "loss": 4.254111289978027,
+ "step": 5545
+ },
+ {
+ "epoch": 77.02794759825328,
+ "grad_norm": 0.01519741676747799,
+ "learning_rate": 0.0006,
+ "loss": 4.1607184410095215,
+ "step": 5546
+ },
+ {
+ "epoch": 77.04192139737991,
+ "grad_norm": 0.015472603030502796,
+ "learning_rate": 0.0006,
+ "loss": 4.323320388793945,
+ "step": 5547
+ },
+ {
+ "epoch": 77.05589519650655,
+ "grad_norm": 0.017937948927283287,
+ "learning_rate": 0.0006,
+ "loss": 4.308490753173828,
+ "step": 5548
+ },
+ {
+ "epoch": 77.06986899563319,
+ "grad_norm": 0.020890841260552406,
+ "learning_rate": 0.0006,
+ "loss": 4.259101867675781,
+ "step": 5549
+ },
+ {
+ "epoch": 77.08384279475983,
+ "grad_norm": 0.01892857998609543,
+ "learning_rate": 0.0006,
+ "loss": 4.386970520019531,
+ "step": 5550
+ },
+ {
+ "epoch": 77.09781659388646,
+ "grad_norm": 0.016583630815148354,
+ "learning_rate": 0.0006,
+ "loss": 4.233604907989502,
+ "step": 5551
+ },
+ {
+ "epoch": 77.1117903930131,
+ "grad_norm": 0.016063468530774117,
+ "learning_rate": 0.0006,
+ "loss": 4.339544296264648,
+ "step": 5552
+ },
+ {
+ "epoch": 77.12576419213974,
+ "grad_norm": 0.018326779827475548,
+ "learning_rate": 0.0006,
+ "loss": 4.367204666137695,
+ "step": 5553
+ },
+ {
+ "epoch": 77.13973799126637,
+ "grad_norm": 0.021590933203697205,
+ "learning_rate": 0.0006,
+ "loss": 4.352685928344727,
+ "step": 5554
+ },
+ {
+ "epoch": 77.15371179039302,
+ "grad_norm": 0.019643642008304596,
+ "learning_rate": 0.0006,
+ "loss": 4.258535385131836,
+ "step": 5555
+ },
+ {
+ "epoch": 77.16768558951965,
+ "grad_norm": 0.018476149067282677,
+ "learning_rate": 0.0006,
+ "loss": 4.232568740844727,
+ "step": 5556
+ },
+ {
+ "epoch": 77.18165938864628,
+ "grad_norm": 0.01584405079483986,
+ "learning_rate": 0.0006,
+ "loss": 4.258371353149414,
+ "step": 5557
+ },
+ {
+ "epoch": 77.19563318777293,
+ "grad_norm": 0.01575201191008091,
+ "learning_rate": 0.0006,
+ "loss": 4.183218002319336,
+ "step": 5558
+ },
+ {
+ "epoch": 77.20960698689956,
+ "grad_norm": 0.016871824860572815,
+ "learning_rate": 0.0006,
+ "loss": 4.194563865661621,
+ "step": 5559
+ },
+ {
+ "epoch": 77.2235807860262,
+ "grad_norm": 0.02059333398938179,
+ "learning_rate": 0.0006,
+ "loss": 4.195281982421875,
+ "step": 5560
+ },
+ {
+ "epoch": 77.23755458515284,
+ "grad_norm": 0.022640453651547432,
+ "learning_rate": 0.0006,
+ "loss": 4.279615879058838,
+ "step": 5561
+ },
+ {
+ "epoch": 77.25152838427948,
+ "grad_norm": 0.020335959270596504,
+ "learning_rate": 0.0006,
+ "loss": 4.25040864944458,
+ "step": 5562
+ },
+ {
+ "epoch": 77.26550218340611,
+ "grad_norm": 0.020010793581604958,
+ "learning_rate": 0.0006,
+ "loss": 4.291116714477539,
+ "step": 5563
+ },
+ {
+ "epoch": 77.27947598253274,
+ "grad_norm": 0.02109045907855034,
+ "learning_rate": 0.0006,
+ "loss": 4.364088535308838,
+ "step": 5564
+ },
+ {
+ "epoch": 77.29344978165939,
+ "grad_norm": 0.019519446417689323,
+ "learning_rate": 0.0006,
+ "loss": 4.329166412353516,
+ "step": 5565
+ },
+ {
+ "epoch": 77.30742358078602,
+ "grad_norm": 0.01689804159104824,
+ "learning_rate": 0.0006,
+ "loss": 4.244176864624023,
+ "step": 5566
+ },
+ {
+ "epoch": 77.32139737991267,
+ "grad_norm": 0.01893681474030018,
+ "learning_rate": 0.0006,
+ "loss": 4.239068984985352,
+ "step": 5567
+ },
+ {
+ "epoch": 77.3353711790393,
+ "grad_norm": 0.015998193994164467,
+ "learning_rate": 0.0006,
+ "loss": 4.173398017883301,
+ "step": 5568
+ },
+ {
+ "epoch": 77.34934497816595,
+ "grad_norm": 0.013892280869185925,
+ "learning_rate": 0.0006,
+ "loss": 4.287654876708984,
+ "step": 5569
+ },
+ {
+ "epoch": 77.36331877729258,
+ "grad_norm": 0.017060643061995506,
+ "learning_rate": 0.0006,
+ "loss": 4.398447036743164,
+ "step": 5570
+ },
+ {
+ "epoch": 77.37729257641921,
+ "grad_norm": 0.015168849378824234,
+ "learning_rate": 0.0006,
+ "loss": 4.15665864944458,
+ "step": 5571
+ },
+ {
+ "epoch": 77.39126637554585,
+ "grad_norm": 0.014886523596942425,
+ "learning_rate": 0.0006,
+ "loss": 4.196673393249512,
+ "step": 5572
+ },
+ {
+ "epoch": 77.40524017467249,
+ "grad_norm": 0.016941796988248825,
+ "learning_rate": 0.0006,
+ "loss": 4.314562797546387,
+ "step": 5573
+ },
+ {
+ "epoch": 77.41921397379913,
+ "grad_norm": 0.015353812836110592,
+ "learning_rate": 0.0006,
+ "loss": 4.409579277038574,
+ "step": 5574
+ },
+ {
+ "epoch": 77.43318777292576,
+ "grad_norm": 0.015192852355539799,
+ "learning_rate": 0.0006,
+ "loss": 4.392045974731445,
+ "step": 5575
+ },
+ {
+ "epoch": 77.44716157205241,
+ "grad_norm": 0.014232569374144077,
+ "learning_rate": 0.0006,
+ "loss": 4.21639347076416,
+ "step": 5576
+ },
+ {
+ "epoch": 77.46113537117904,
+ "grad_norm": 0.014453226700425148,
+ "learning_rate": 0.0006,
+ "loss": 4.2802276611328125,
+ "step": 5577
+ },
+ {
+ "epoch": 77.47510917030567,
+ "grad_norm": 0.012881236150860786,
+ "learning_rate": 0.0006,
+ "loss": 4.375980854034424,
+ "step": 5578
+ },
+ {
+ "epoch": 77.48908296943232,
+ "grad_norm": 0.01583976298570633,
+ "learning_rate": 0.0006,
+ "loss": 4.264520645141602,
+ "step": 5579
+ },
+ {
+ "epoch": 77.50305676855895,
+ "grad_norm": 0.014171584509313107,
+ "learning_rate": 0.0006,
+ "loss": 4.225279808044434,
+ "step": 5580
+ },
+ {
+ "epoch": 77.5170305676856,
+ "grad_norm": 0.014715159311890602,
+ "learning_rate": 0.0006,
+ "loss": 4.369834899902344,
+ "step": 5581
+ },
+ {
+ "epoch": 77.53100436681223,
+ "grad_norm": 0.016554079949855804,
+ "learning_rate": 0.0006,
+ "loss": 4.2976179122924805,
+ "step": 5582
+ },
+ {
+ "epoch": 77.54497816593886,
+ "grad_norm": 0.015095747075974941,
+ "learning_rate": 0.0006,
+ "loss": 4.294407844543457,
+ "step": 5583
+ },
+ {
+ "epoch": 77.5589519650655,
+ "grad_norm": 0.012726361863315105,
+ "learning_rate": 0.0006,
+ "loss": 4.221401214599609,
+ "step": 5584
+ },
+ {
+ "epoch": 77.57292576419214,
+ "grad_norm": 0.014999309554696083,
+ "learning_rate": 0.0006,
+ "loss": 4.243929386138916,
+ "step": 5585
+ },
+ {
+ "epoch": 77.58689956331878,
+ "grad_norm": 0.016526037827134132,
+ "learning_rate": 0.0006,
+ "loss": 4.374139785766602,
+ "step": 5586
+ },
+ {
+ "epoch": 77.60087336244541,
+ "grad_norm": 0.01648339442908764,
+ "learning_rate": 0.0006,
+ "loss": 4.392425060272217,
+ "step": 5587
+ },
+ {
+ "epoch": 77.61484716157206,
+ "grad_norm": 0.015269213356077671,
+ "learning_rate": 0.0006,
+ "loss": 4.33986759185791,
+ "step": 5588
+ },
+ {
+ "epoch": 77.62882096069869,
+ "grad_norm": 0.01690507121384144,
+ "learning_rate": 0.0006,
+ "loss": 4.307476043701172,
+ "step": 5589
+ },
+ {
+ "epoch": 77.64279475982532,
+ "grad_norm": 0.016438685357570648,
+ "learning_rate": 0.0006,
+ "loss": 4.2406206130981445,
+ "step": 5590
+ },
+ {
+ "epoch": 77.65676855895197,
+ "grad_norm": 0.0157600287348032,
+ "learning_rate": 0.0006,
+ "loss": 4.314070224761963,
+ "step": 5591
+ },
+ {
+ "epoch": 77.6707423580786,
+ "grad_norm": 0.015127947553992271,
+ "learning_rate": 0.0006,
+ "loss": 4.22475528717041,
+ "step": 5592
+ },
+ {
+ "epoch": 77.68471615720524,
+ "grad_norm": 0.016632677987217903,
+ "learning_rate": 0.0006,
+ "loss": 4.26954984664917,
+ "step": 5593
+ },
+ {
+ "epoch": 77.69868995633188,
+ "grad_norm": 0.013150460086762905,
+ "learning_rate": 0.0006,
+ "loss": 4.369131088256836,
+ "step": 5594
+ },
+ {
+ "epoch": 77.71266375545852,
+ "grad_norm": 0.014513005502521992,
+ "learning_rate": 0.0006,
+ "loss": 4.3248796463012695,
+ "step": 5595
+ },
+ {
+ "epoch": 77.72663755458515,
+ "grad_norm": 0.014090991578996181,
+ "learning_rate": 0.0006,
+ "loss": 4.268381118774414,
+ "step": 5596
+ },
+ {
+ "epoch": 77.74061135371178,
+ "grad_norm": 0.01633097417652607,
+ "learning_rate": 0.0006,
+ "loss": 4.238439083099365,
+ "step": 5597
+ },
+ {
+ "epoch": 77.75458515283843,
+ "grad_norm": 0.01710933819413185,
+ "learning_rate": 0.0006,
+ "loss": 4.399697780609131,
+ "step": 5598
+ },
+ {
+ "epoch": 77.76855895196506,
+ "grad_norm": 0.019251767545938492,
+ "learning_rate": 0.0006,
+ "loss": 4.290450096130371,
+ "step": 5599
+ },
+ {
+ "epoch": 77.78253275109171,
+ "grad_norm": 0.01849171705543995,
+ "learning_rate": 0.0006,
+ "loss": 4.267734527587891,
+ "step": 5600
+ },
+ {
+ "epoch": 77.79650655021834,
+ "grad_norm": 0.014173115603625774,
+ "learning_rate": 0.0006,
+ "loss": 4.288799285888672,
+ "step": 5601
+ },
+ {
+ "epoch": 77.81048034934499,
+ "grad_norm": 0.014572718180716038,
+ "learning_rate": 0.0006,
+ "loss": 4.3790435791015625,
+ "step": 5602
+ },
+ {
+ "epoch": 77.82445414847162,
+ "grad_norm": 0.016375336796045303,
+ "learning_rate": 0.0006,
+ "loss": 4.231725692749023,
+ "step": 5603
+ },
+ {
+ "epoch": 77.83842794759825,
+ "grad_norm": 0.017326701432466507,
+ "learning_rate": 0.0006,
+ "loss": 4.211495399475098,
+ "step": 5604
+ },
+ {
+ "epoch": 77.8524017467249,
+ "grad_norm": 0.016399390995502472,
+ "learning_rate": 0.0006,
+ "loss": 4.386699199676514,
+ "step": 5605
+ },
+ {
+ "epoch": 77.86637554585153,
+ "grad_norm": 0.018269989639520645,
+ "learning_rate": 0.0006,
+ "loss": 4.254856109619141,
+ "step": 5606
+ },
+ {
+ "epoch": 77.88034934497817,
+ "grad_norm": 0.018443435430526733,
+ "learning_rate": 0.0006,
+ "loss": 4.391961097717285,
+ "step": 5607
+ },
+ {
+ "epoch": 77.8943231441048,
+ "grad_norm": 0.01784636080265045,
+ "learning_rate": 0.0006,
+ "loss": 4.273881912231445,
+ "step": 5608
+ },
+ {
+ "epoch": 77.90829694323143,
+ "grad_norm": 0.01782994158565998,
+ "learning_rate": 0.0006,
+ "loss": 4.26833963394165,
+ "step": 5609
+ },
+ {
+ "epoch": 77.92227074235808,
+ "grad_norm": 0.016260741278529167,
+ "learning_rate": 0.0006,
+ "loss": 4.306807994842529,
+ "step": 5610
+ },
+ {
+ "epoch": 77.93624454148471,
+ "grad_norm": 0.015482903458178043,
+ "learning_rate": 0.0006,
+ "loss": 4.187826633453369,
+ "step": 5611
+ },
+ {
+ "epoch": 77.95021834061136,
+ "grad_norm": 0.015193257480859756,
+ "learning_rate": 0.0006,
+ "loss": 4.286060333251953,
+ "step": 5612
+ },
+ {
+ "epoch": 77.96419213973799,
+ "grad_norm": 0.015110115520656109,
+ "learning_rate": 0.0006,
+ "loss": 4.2828192710876465,
+ "step": 5613
+ },
+ {
+ "epoch": 77.97816593886463,
+ "grad_norm": 0.015973486006259918,
+ "learning_rate": 0.0006,
+ "loss": 4.317939758300781,
+ "step": 5614
+ },
+ {
+ "epoch": 77.99213973799127,
+ "grad_norm": 0.017363594844937325,
+ "learning_rate": 0.0006,
+ "loss": 4.287721633911133,
+ "step": 5615
+ },
+ {
+ "epoch": 78.0,
+ "grad_norm": 0.015277642756700516,
+ "learning_rate": 0.0006,
+ "loss": 4.356618881225586,
+ "step": 5616
+ },
+ {
+ "epoch": 78.0,
+ "eval_loss": 4.642678737640381,
+ "eval_runtime": 56.7335,
+ "eval_samples_per_second": 43.043,
+ "eval_steps_per_second": 1.357,
+ "step": 5616
+ },
+ {
+ "epoch": 78.01397379912663,
+ "grad_norm": 0.016794051975011826,
+ "learning_rate": 0.0006,
+ "loss": 4.279930591583252,
+ "step": 5617
+ },
+ {
+ "epoch": 78.02794759825328,
+ "grad_norm": 0.01632673107087612,
+ "learning_rate": 0.0006,
+ "loss": 4.235943794250488,
+ "step": 5618
+ },
+ {
+ "epoch": 78.04192139737991,
+ "grad_norm": 0.01437876746058464,
+ "learning_rate": 0.0006,
+ "loss": 4.2542243003845215,
+ "step": 5619
+ },
+ {
+ "epoch": 78.05589519650655,
+ "grad_norm": 0.014993119053542614,
+ "learning_rate": 0.0006,
+ "loss": 4.262046813964844,
+ "step": 5620
+ },
+ {
+ "epoch": 78.06986899563319,
+ "grad_norm": 0.015721119940280914,
+ "learning_rate": 0.0006,
+ "loss": 4.173436164855957,
+ "step": 5621
+ },
+ {
+ "epoch": 78.08384279475983,
+ "grad_norm": 0.01334085687994957,
+ "learning_rate": 0.0006,
+ "loss": 4.322175979614258,
+ "step": 5622
+ },
+ {
+ "epoch": 78.09781659388646,
+ "grad_norm": 0.015561388805508614,
+ "learning_rate": 0.0006,
+ "loss": 4.242125988006592,
+ "step": 5623
+ },
+ {
+ "epoch": 78.1117903930131,
+ "grad_norm": 0.014418602921068668,
+ "learning_rate": 0.0006,
+ "loss": 4.266186714172363,
+ "step": 5624
+ },
+ {
+ "epoch": 78.12576419213974,
+ "grad_norm": 0.013739909045398235,
+ "learning_rate": 0.0006,
+ "loss": 4.231633186340332,
+ "step": 5625
+ },
+ {
+ "epoch": 78.13973799126637,
+ "grad_norm": 0.01402067206799984,
+ "learning_rate": 0.0006,
+ "loss": 4.188406944274902,
+ "step": 5626
+ },
+ {
+ "epoch": 78.15371179039302,
+ "grad_norm": 0.013639481738209724,
+ "learning_rate": 0.0006,
+ "loss": 4.1586594581604,
+ "step": 5627
+ },
+ {
+ "epoch": 78.16768558951965,
+ "grad_norm": 0.012858159840106964,
+ "learning_rate": 0.0006,
+ "loss": 4.360448360443115,
+ "step": 5628
+ },
+ {
+ "epoch": 78.18165938864628,
+ "grad_norm": 0.013089739717543125,
+ "learning_rate": 0.0006,
+ "loss": 4.344371795654297,
+ "step": 5629
+ },
+ {
+ "epoch": 78.19563318777293,
+ "grad_norm": 0.01382314320653677,
+ "learning_rate": 0.0006,
+ "loss": 4.20378303527832,
+ "step": 5630
+ },
+ {
+ "epoch": 78.20960698689956,
+ "grad_norm": 0.014206619001924992,
+ "learning_rate": 0.0006,
+ "loss": 4.239891529083252,
+ "step": 5631
+ },
+ {
+ "epoch": 78.2235807860262,
+ "grad_norm": 0.01472168043255806,
+ "learning_rate": 0.0006,
+ "loss": 4.282382488250732,
+ "step": 5632
+ },
+ {
+ "epoch": 78.23755458515284,
+ "grad_norm": 0.014237025752663612,
+ "learning_rate": 0.0006,
+ "loss": 4.241545677185059,
+ "step": 5633
+ },
+ {
+ "epoch": 78.25152838427948,
+ "grad_norm": 0.013231472112238407,
+ "learning_rate": 0.0006,
+ "loss": 4.184896469116211,
+ "step": 5634
+ },
+ {
+ "epoch": 78.26550218340611,
+ "grad_norm": 0.013538471423089504,
+ "learning_rate": 0.0006,
+ "loss": 4.325181007385254,
+ "step": 5635
+ },
+ {
+ "epoch": 78.27947598253274,
+ "grad_norm": 0.013541961088776588,
+ "learning_rate": 0.0006,
+ "loss": 4.368722915649414,
+ "step": 5636
+ },
+ {
+ "epoch": 78.29344978165939,
+ "grad_norm": 0.013446723110973835,
+ "learning_rate": 0.0006,
+ "loss": 4.301422595977783,
+ "step": 5637
+ },
+ {
+ "epoch": 78.30742358078602,
+ "grad_norm": 0.014753809198737144,
+ "learning_rate": 0.0006,
+ "loss": 4.259127616882324,
+ "step": 5638
+ },
+ {
+ "epoch": 78.32139737991267,
+ "grad_norm": 0.013256008736789227,
+ "learning_rate": 0.0006,
+ "loss": 4.284032821655273,
+ "step": 5639
+ },
+ {
+ "epoch": 78.3353711790393,
+ "grad_norm": 0.014272629283368587,
+ "learning_rate": 0.0006,
+ "loss": 4.234989166259766,
+ "step": 5640
+ },
+ {
+ "epoch": 78.34934497816595,
+ "grad_norm": 0.014756826683878899,
+ "learning_rate": 0.0006,
+ "loss": 4.143499374389648,
+ "step": 5641
+ },
+ {
+ "epoch": 78.36331877729258,
+ "grad_norm": 0.016549205407500267,
+ "learning_rate": 0.0006,
+ "loss": 4.311034202575684,
+ "step": 5642
+ },
+ {
+ "epoch": 78.37729257641921,
+ "grad_norm": 0.017735283821821213,
+ "learning_rate": 0.0006,
+ "loss": 4.246767520904541,
+ "step": 5643
+ },
+ {
+ "epoch": 78.39126637554585,
+ "grad_norm": 0.017370613291859627,
+ "learning_rate": 0.0006,
+ "loss": 4.2460832595825195,
+ "step": 5644
+ },
+ {
+ "epoch": 78.40524017467249,
+ "grad_norm": 0.014642846770584583,
+ "learning_rate": 0.0006,
+ "loss": 4.303811550140381,
+ "step": 5645
+ },
+ {
+ "epoch": 78.41921397379913,
+ "grad_norm": 0.013445726595818996,
+ "learning_rate": 0.0006,
+ "loss": 4.25921630859375,
+ "step": 5646
+ },
+ {
+ "epoch": 78.43318777292576,
+ "grad_norm": 0.014254065230488777,
+ "learning_rate": 0.0006,
+ "loss": 4.129446506500244,
+ "step": 5647
+ },
+ {
+ "epoch": 78.44716157205241,
+ "grad_norm": 0.01883019506931305,
+ "learning_rate": 0.0006,
+ "loss": 4.191716194152832,
+ "step": 5648
+ },
+ {
+ "epoch": 78.46113537117904,
+ "grad_norm": 0.0204437468200922,
+ "learning_rate": 0.0006,
+ "loss": 4.333085536956787,
+ "step": 5649
+ },
+ {
+ "epoch": 78.47510917030567,
+ "grad_norm": 0.01652318798005581,
+ "learning_rate": 0.0006,
+ "loss": 4.135462760925293,
+ "step": 5650
+ },
+ {
+ "epoch": 78.48908296943232,
+ "grad_norm": 0.012753551825881004,
+ "learning_rate": 0.0006,
+ "loss": 4.195051193237305,
+ "step": 5651
+ },
+ {
+ "epoch": 78.50305676855895,
+ "grad_norm": 0.01499118935316801,
+ "learning_rate": 0.0006,
+ "loss": 4.157092094421387,
+ "step": 5652
+ },
+ {
+ "epoch": 78.5170305676856,
+ "grad_norm": 0.015383945778012276,
+ "learning_rate": 0.0006,
+ "loss": 4.253818035125732,
+ "step": 5653
+ },
+ {
+ "epoch": 78.53100436681223,
+ "grad_norm": 0.015721332281827927,
+ "learning_rate": 0.0006,
+ "loss": 4.228775978088379,
+ "step": 5654
+ },
+ {
+ "epoch": 78.54497816593886,
+ "grad_norm": 0.014595463871955872,
+ "learning_rate": 0.0006,
+ "loss": 4.135197639465332,
+ "step": 5655
+ },
+ {
+ "epoch": 78.5589519650655,
+ "grad_norm": 0.015059644356369972,
+ "learning_rate": 0.0006,
+ "loss": 4.255614280700684,
+ "step": 5656
+ },
+ {
+ "epoch": 78.57292576419214,
+ "grad_norm": 0.013833344914019108,
+ "learning_rate": 0.0006,
+ "loss": 4.301250457763672,
+ "step": 5657
+ },
+ {
+ "epoch": 78.58689956331878,
+ "grad_norm": 0.013665376231074333,
+ "learning_rate": 0.0006,
+ "loss": 4.1734232902526855,
+ "step": 5658
+ },
+ {
+ "epoch": 78.60087336244541,
+ "grad_norm": 0.01539086177945137,
+ "learning_rate": 0.0006,
+ "loss": 4.221459865570068,
+ "step": 5659
+ },
+ {
+ "epoch": 78.61484716157206,
+ "grad_norm": 0.013313917443156242,
+ "learning_rate": 0.0006,
+ "loss": 4.255619049072266,
+ "step": 5660
+ },
+ {
+ "epoch": 78.62882096069869,
+ "grad_norm": 0.01447741687297821,
+ "learning_rate": 0.0006,
+ "loss": 4.298219680786133,
+ "step": 5661
+ },
+ {
+ "epoch": 78.64279475982532,
+ "grad_norm": 0.012796915136277676,
+ "learning_rate": 0.0006,
+ "loss": 4.2542405128479,
+ "step": 5662
+ },
+ {
+ "epoch": 78.65676855895197,
+ "grad_norm": 0.014058141969144344,
+ "learning_rate": 0.0006,
+ "loss": 4.284477233886719,
+ "step": 5663
+ },
+ {
+ "epoch": 78.6707423580786,
+ "grad_norm": 0.012705069966614246,
+ "learning_rate": 0.0006,
+ "loss": 4.2826666831970215,
+ "step": 5664
+ },
+ {
+ "epoch": 78.68471615720524,
+ "grad_norm": 0.013245890848338604,
+ "learning_rate": 0.0006,
+ "loss": 4.371612548828125,
+ "step": 5665
+ },
+ {
+ "epoch": 78.69868995633188,
+ "grad_norm": 0.014192677102982998,
+ "learning_rate": 0.0006,
+ "loss": 4.267344951629639,
+ "step": 5666
+ },
+ {
+ "epoch": 78.71266375545852,
+ "grad_norm": 0.015346777625381947,
+ "learning_rate": 0.0006,
+ "loss": 4.222797393798828,
+ "step": 5667
+ },
+ {
+ "epoch": 78.72663755458515,
+ "grad_norm": 0.014068621210753918,
+ "learning_rate": 0.0006,
+ "loss": 4.172964572906494,
+ "step": 5668
+ },
+ {
+ "epoch": 78.74061135371178,
+ "grad_norm": 0.014364945702254772,
+ "learning_rate": 0.0006,
+ "loss": 4.259319305419922,
+ "step": 5669
+ },
+ {
+ "epoch": 78.75458515283843,
+ "grad_norm": 0.017337387427687645,
+ "learning_rate": 0.0006,
+ "loss": 4.29533576965332,
+ "step": 5670
+ },
+ {
+ "epoch": 78.76855895196506,
+ "grad_norm": 0.017416715621948242,
+ "learning_rate": 0.0006,
+ "loss": 4.340364456176758,
+ "step": 5671
+ },
+ {
+ "epoch": 78.78253275109171,
+ "grad_norm": 0.015925846993923187,
+ "learning_rate": 0.0006,
+ "loss": 4.177454948425293,
+ "step": 5672
+ },
+ {
+ "epoch": 78.79650655021834,
+ "grad_norm": 0.013130443170666695,
+ "learning_rate": 0.0006,
+ "loss": 4.277454853057861,
+ "step": 5673
+ },
+ {
+ "epoch": 78.81048034934499,
+ "grad_norm": 0.014628687873482704,
+ "learning_rate": 0.0006,
+ "loss": 4.273757457733154,
+ "step": 5674
+ },
+ {
+ "epoch": 78.82445414847162,
+ "grad_norm": 0.01601659320294857,
+ "learning_rate": 0.0006,
+ "loss": 4.281774520874023,
+ "step": 5675
+ },
+ {
+ "epoch": 78.83842794759825,
+ "grad_norm": 0.015443485230207443,
+ "learning_rate": 0.0006,
+ "loss": 4.199937343597412,
+ "step": 5676
+ },
+ {
+ "epoch": 78.8524017467249,
+ "grad_norm": 0.013934246264398098,
+ "learning_rate": 0.0006,
+ "loss": 4.331369876861572,
+ "step": 5677
+ },
+ {
+ "epoch": 78.86637554585153,
+ "grad_norm": 0.01572083681821823,
+ "learning_rate": 0.0006,
+ "loss": 4.2828474044799805,
+ "step": 5678
+ },
+ {
+ "epoch": 78.88034934497817,
+ "grad_norm": 0.013444120064377785,
+ "learning_rate": 0.0006,
+ "loss": 4.183859825134277,
+ "step": 5679
+ },
+ {
+ "epoch": 78.8943231441048,
+ "grad_norm": 0.01591254211962223,
+ "learning_rate": 0.0006,
+ "loss": 4.295256614685059,
+ "step": 5680
+ },
+ {
+ "epoch": 78.90829694323143,
+ "grad_norm": 0.017556684091687202,
+ "learning_rate": 0.0006,
+ "loss": 4.132272720336914,
+ "step": 5681
+ },
+ {
+ "epoch": 78.92227074235808,
+ "grad_norm": 0.018184322863817215,
+ "learning_rate": 0.0006,
+ "loss": 4.251728057861328,
+ "step": 5682
+ },
+ {
+ "epoch": 78.93624454148471,
+ "grad_norm": 0.016207285225391388,
+ "learning_rate": 0.0006,
+ "loss": 4.205544471740723,
+ "step": 5683
+ },
+ {
+ "epoch": 78.95021834061136,
+ "grad_norm": 0.014478376135230064,
+ "learning_rate": 0.0006,
+ "loss": 4.182462215423584,
+ "step": 5684
+ },
+ {
+ "epoch": 78.96419213973799,
+ "grad_norm": 0.014255235902965069,
+ "learning_rate": 0.0006,
+ "loss": 4.253366947174072,
+ "step": 5685
+ },
+ {
+ "epoch": 78.97816593886463,
+ "grad_norm": 0.01524396613240242,
+ "learning_rate": 0.0006,
+ "loss": 4.221786022186279,
+ "step": 5686
+ },
+ {
+ "epoch": 78.99213973799127,
+ "grad_norm": 0.015057693235576153,
+ "learning_rate": 0.0006,
+ "loss": 4.208454608917236,
+ "step": 5687
+ },
+ {
+ "epoch": 79.0,
+ "grad_norm": 0.015478022396564484,
+ "learning_rate": 0.0006,
+ "loss": 4.336383819580078,
+ "step": 5688
+ },
+ {
+ "epoch": 79.0,
+ "eval_loss": 4.641749382019043,
+ "eval_runtime": 56.8167,
+ "eval_samples_per_second": 42.98,
+ "eval_steps_per_second": 1.355,
+ "step": 5688
+ },
+ {
+ "epoch": 79.01397379912663,
+ "grad_norm": 0.014143316075205803,
+ "learning_rate": 0.0006,
+ "loss": 4.260974884033203,
+ "step": 5689
+ },
+ {
+ "epoch": 79.02794759825328,
+ "grad_norm": 0.013120883144438267,
+ "learning_rate": 0.0006,
+ "loss": 4.139886856079102,
+ "step": 5690
+ },
+ {
+ "epoch": 79.04192139737991,
+ "grad_norm": 0.015019661746919155,
+ "learning_rate": 0.0006,
+ "loss": 4.267004013061523,
+ "step": 5691
+ },
+ {
+ "epoch": 79.05589519650655,
+ "grad_norm": 0.014891284517943859,
+ "learning_rate": 0.0006,
+ "loss": 4.205019950866699,
+ "step": 5692
+ },
+ {
+ "epoch": 79.06986899563319,
+ "grad_norm": 0.015527728945016861,
+ "learning_rate": 0.0006,
+ "loss": 4.1513566970825195,
+ "step": 5693
+ },
+ {
+ "epoch": 79.08384279475983,
+ "grad_norm": 0.017527181655168533,
+ "learning_rate": 0.0006,
+ "loss": 4.148828983306885,
+ "step": 5694
+ },
+ {
+ "epoch": 79.09781659388646,
+ "grad_norm": 0.01632206328213215,
+ "learning_rate": 0.0006,
+ "loss": 4.195789813995361,
+ "step": 5695
+ },
+ {
+ "epoch": 79.1117903930131,
+ "grad_norm": 0.01590503193438053,
+ "learning_rate": 0.0006,
+ "loss": 4.199155330657959,
+ "step": 5696
+ },
+ {
+ "epoch": 79.12576419213974,
+ "grad_norm": 0.016489997506141663,
+ "learning_rate": 0.0006,
+ "loss": 4.155154705047607,
+ "step": 5697
+ },
+ {
+ "epoch": 79.13973799126637,
+ "grad_norm": 0.0161368977278471,
+ "learning_rate": 0.0006,
+ "loss": 4.164384841918945,
+ "step": 5698
+ },
+ {
+ "epoch": 79.15371179039302,
+ "grad_norm": 0.018574809655547142,
+ "learning_rate": 0.0006,
+ "loss": 4.1898193359375,
+ "step": 5699
+ },
+ {
+ "epoch": 79.16768558951965,
+ "grad_norm": 0.018987983465194702,
+ "learning_rate": 0.0006,
+ "loss": 4.219719409942627,
+ "step": 5700
+ },
+ {
+ "epoch": 79.18165938864628,
+ "grad_norm": 0.01668606884777546,
+ "learning_rate": 0.0006,
+ "loss": 4.152734756469727,
+ "step": 5701
+ },
+ {
+ "epoch": 79.19563318777293,
+ "grad_norm": 0.014046299271285534,
+ "learning_rate": 0.0006,
+ "loss": 4.152964115142822,
+ "step": 5702
+ },
+ {
+ "epoch": 79.20960698689956,
+ "grad_norm": 0.01628032699227333,
+ "learning_rate": 0.0006,
+ "loss": 4.299479961395264,
+ "step": 5703
+ },
+ {
+ "epoch": 79.2235807860262,
+ "grad_norm": 0.017998795956373215,
+ "learning_rate": 0.0006,
+ "loss": 4.231203079223633,
+ "step": 5704
+ },
+ {
+ "epoch": 79.23755458515284,
+ "grad_norm": 0.014799647964537144,
+ "learning_rate": 0.0006,
+ "loss": 4.3530192375183105,
+ "step": 5705
+ },
+ {
+ "epoch": 79.25152838427948,
+ "grad_norm": 0.014640678651630878,
+ "learning_rate": 0.0006,
+ "loss": 4.287322998046875,
+ "step": 5706
+ },
+ {
+ "epoch": 79.26550218340611,
+ "grad_norm": 0.0172616895288229,
+ "learning_rate": 0.0006,
+ "loss": 4.148009777069092,
+ "step": 5707
+ },
+ {
+ "epoch": 79.27947598253274,
+ "grad_norm": 0.01811790093779564,
+ "learning_rate": 0.0006,
+ "loss": 4.269433975219727,
+ "step": 5708
+ },
+ {
+ "epoch": 79.29344978165939,
+ "grad_norm": 0.01764000579714775,
+ "learning_rate": 0.0006,
+ "loss": 4.292638778686523,
+ "step": 5709
+ },
+ {
+ "epoch": 79.30742358078602,
+ "grad_norm": 0.018030935898423195,
+ "learning_rate": 0.0006,
+ "loss": 4.267889976501465,
+ "step": 5710
+ },
+ {
+ "epoch": 79.32139737991267,
+ "grad_norm": 0.017860891297459602,
+ "learning_rate": 0.0006,
+ "loss": 4.211110591888428,
+ "step": 5711
+ },
+ {
+ "epoch": 79.3353711790393,
+ "grad_norm": 0.017295073717832565,
+ "learning_rate": 0.0006,
+ "loss": 4.224645614624023,
+ "step": 5712
+ },
+ {
+ "epoch": 79.34934497816595,
+ "grad_norm": 0.015891101211309433,
+ "learning_rate": 0.0006,
+ "loss": 4.224797248840332,
+ "step": 5713
+ },
+ {
+ "epoch": 79.36331877729258,
+ "grad_norm": 0.014680417254567146,
+ "learning_rate": 0.0006,
+ "loss": 4.182242393493652,
+ "step": 5714
+ },
+ {
+ "epoch": 79.37729257641921,
+ "grad_norm": 0.016621140763163567,
+ "learning_rate": 0.0006,
+ "loss": 4.186334609985352,
+ "step": 5715
+ },
+ {
+ "epoch": 79.39126637554585,
+ "grad_norm": 0.016207249835133553,
+ "learning_rate": 0.0006,
+ "loss": 4.277074813842773,
+ "step": 5716
+ },
+ {
+ "epoch": 79.40524017467249,
+ "grad_norm": 0.015375478193163872,
+ "learning_rate": 0.0006,
+ "loss": 4.154317855834961,
+ "step": 5717
+ },
+ {
+ "epoch": 79.41921397379913,
+ "grad_norm": 0.013453228399157524,
+ "learning_rate": 0.0006,
+ "loss": 4.265270233154297,
+ "step": 5718
+ },
+ {
+ "epoch": 79.43318777292576,
+ "grad_norm": 0.014312694780528545,
+ "learning_rate": 0.0006,
+ "loss": 4.232220649719238,
+ "step": 5719
+ },
+ {
+ "epoch": 79.44716157205241,
+ "grad_norm": 0.014759554527699947,
+ "learning_rate": 0.0006,
+ "loss": 4.272275924682617,
+ "step": 5720
+ },
+ {
+ "epoch": 79.46113537117904,
+ "grad_norm": 0.015885505825281143,
+ "learning_rate": 0.0006,
+ "loss": 4.2229719161987305,
+ "step": 5721
+ },
+ {
+ "epoch": 79.47510917030567,
+ "grad_norm": 0.016516581177711487,
+ "learning_rate": 0.0006,
+ "loss": 4.200624465942383,
+ "step": 5722
+ },
+ {
+ "epoch": 79.48908296943232,
+ "grad_norm": 0.018254555761814117,
+ "learning_rate": 0.0006,
+ "loss": 4.19648551940918,
+ "step": 5723
+ },
+ {
+ "epoch": 79.50305676855895,
+ "grad_norm": 0.01467090006917715,
+ "learning_rate": 0.0006,
+ "loss": 4.135190486907959,
+ "step": 5724
+ },
+ {
+ "epoch": 79.5170305676856,
+ "grad_norm": 0.013474797829985619,
+ "learning_rate": 0.0006,
+ "loss": 4.311361312866211,
+ "step": 5725
+ },
+ {
+ "epoch": 79.53100436681223,
+ "grad_norm": 0.01742764748632908,
+ "learning_rate": 0.0006,
+ "loss": 4.1845903396606445,
+ "step": 5726
+ },
+ {
+ "epoch": 79.54497816593886,
+ "grad_norm": 0.018786882981657982,
+ "learning_rate": 0.0006,
+ "loss": 4.163942337036133,
+ "step": 5727
+ },
+ {
+ "epoch": 79.5589519650655,
+ "grad_norm": 0.017831219360232353,
+ "learning_rate": 0.0006,
+ "loss": 4.236685276031494,
+ "step": 5728
+ },
+ {
+ "epoch": 79.57292576419214,
+ "grad_norm": 0.016334986314177513,
+ "learning_rate": 0.0006,
+ "loss": 4.255853652954102,
+ "step": 5729
+ },
+ {
+ "epoch": 79.58689956331878,
+ "grad_norm": 0.0138487434014678,
+ "learning_rate": 0.0006,
+ "loss": 4.227874279022217,
+ "step": 5730
+ },
+ {
+ "epoch": 79.60087336244541,
+ "grad_norm": 0.01531760673969984,
+ "learning_rate": 0.0006,
+ "loss": 4.221432685852051,
+ "step": 5731
+ },
+ {
+ "epoch": 79.61484716157206,
+ "grad_norm": 0.01603361964225769,
+ "learning_rate": 0.0006,
+ "loss": 4.278326511383057,
+ "step": 5732
+ },
+ {
+ "epoch": 79.62882096069869,
+ "grad_norm": 0.013149895705282688,
+ "learning_rate": 0.0006,
+ "loss": 4.2474517822265625,
+ "step": 5733
+ },
+ {
+ "epoch": 79.64279475982532,
+ "grad_norm": 0.013428786769509315,
+ "learning_rate": 0.0006,
+ "loss": 4.1701178550720215,
+ "step": 5734
+ },
+ {
+ "epoch": 79.65676855895197,
+ "grad_norm": 0.014305585995316505,
+ "learning_rate": 0.0006,
+ "loss": 4.250778675079346,
+ "step": 5735
+ },
+ {
+ "epoch": 79.6707423580786,
+ "grad_norm": 0.014544566161930561,
+ "learning_rate": 0.0006,
+ "loss": 4.243236541748047,
+ "step": 5736
+ },
+ {
+ "epoch": 79.68471615720524,
+ "grad_norm": 0.013524414040148258,
+ "learning_rate": 0.0006,
+ "loss": 4.255457878112793,
+ "step": 5737
+ },
+ {
+ "epoch": 79.69868995633188,
+ "grad_norm": 0.014750641770660877,
+ "learning_rate": 0.0006,
+ "loss": 4.202816486358643,
+ "step": 5738
+ },
+ {
+ "epoch": 79.71266375545852,
+ "grad_norm": 0.015090773813426495,
+ "learning_rate": 0.0006,
+ "loss": 4.3053178787231445,
+ "step": 5739
+ },
+ {
+ "epoch": 79.72663755458515,
+ "grad_norm": 0.01606169529259205,
+ "learning_rate": 0.0006,
+ "loss": 4.224017143249512,
+ "step": 5740
+ },
+ {
+ "epoch": 79.74061135371178,
+ "grad_norm": 0.016303520649671555,
+ "learning_rate": 0.0006,
+ "loss": 4.190760612487793,
+ "step": 5741
+ },
+ {
+ "epoch": 79.75458515283843,
+ "grad_norm": 0.01465672068297863,
+ "learning_rate": 0.0006,
+ "loss": 4.24506950378418,
+ "step": 5742
+ },
+ {
+ "epoch": 79.76855895196506,
+ "grad_norm": 0.012616301886737347,
+ "learning_rate": 0.0006,
+ "loss": 4.174039840698242,
+ "step": 5743
+ },
+ {
+ "epoch": 79.78253275109171,
+ "grad_norm": 0.01358284242451191,
+ "learning_rate": 0.0006,
+ "loss": 4.236380577087402,
+ "step": 5744
+ },
+ {
+ "epoch": 79.79650655021834,
+ "grad_norm": 0.01557177770882845,
+ "learning_rate": 0.0006,
+ "loss": 4.102962493896484,
+ "step": 5745
+ },
+ {
+ "epoch": 79.81048034934499,
+ "grad_norm": 0.015584317035973072,
+ "learning_rate": 0.0006,
+ "loss": 4.242603302001953,
+ "step": 5746
+ },
+ {
+ "epoch": 79.82445414847162,
+ "grad_norm": 0.012597334571182728,
+ "learning_rate": 0.0006,
+ "loss": 4.202791690826416,
+ "step": 5747
+ },
+ {
+ "epoch": 79.83842794759825,
+ "grad_norm": 0.013722898438572884,
+ "learning_rate": 0.0006,
+ "loss": 4.204328536987305,
+ "step": 5748
+ },
+ {
+ "epoch": 79.8524017467249,
+ "grad_norm": 0.014669225551187992,
+ "learning_rate": 0.0006,
+ "loss": 4.164331912994385,
+ "step": 5749
+ },
+ {
+ "epoch": 79.86637554585153,
+ "grad_norm": 0.015012740157544613,
+ "learning_rate": 0.0006,
+ "loss": 4.260565280914307,
+ "step": 5750
+ },
+ {
+ "epoch": 79.88034934497817,
+ "grad_norm": 0.016026340425014496,
+ "learning_rate": 0.0006,
+ "loss": 4.299723148345947,
+ "step": 5751
+ },
+ {
+ "epoch": 79.8943231441048,
+ "grad_norm": 0.01626049168407917,
+ "learning_rate": 0.0006,
+ "loss": 4.207252502441406,
+ "step": 5752
+ },
+ {
+ "epoch": 79.90829694323143,
+ "grad_norm": 0.015415354631841183,
+ "learning_rate": 0.0006,
+ "loss": 4.151934623718262,
+ "step": 5753
+ },
+ {
+ "epoch": 79.92227074235808,
+ "grad_norm": 0.014158242382109165,
+ "learning_rate": 0.0006,
+ "loss": 4.3461995124816895,
+ "step": 5754
+ },
+ {
+ "epoch": 79.93624454148471,
+ "grad_norm": 0.01585628278553486,
+ "learning_rate": 0.0006,
+ "loss": 4.334761619567871,
+ "step": 5755
+ },
+ {
+ "epoch": 79.95021834061136,
+ "grad_norm": 0.013710759580135345,
+ "learning_rate": 0.0006,
+ "loss": 4.180401802062988,
+ "step": 5756
+ },
+ {
+ "epoch": 79.96419213973799,
+ "grad_norm": 0.014159204438328743,
+ "learning_rate": 0.0006,
+ "loss": 4.322331428527832,
+ "step": 5757
+ },
+ {
+ "epoch": 79.97816593886463,
+ "grad_norm": 0.015153195708990097,
+ "learning_rate": 0.0006,
+ "loss": 4.287478923797607,
+ "step": 5758
+ },
+ {
+ "epoch": 79.99213973799127,
+ "grad_norm": 0.014100227504968643,
+ "learning_rate": 0.0006,
+ "loss": 4.231807708740234,
+ "step": 5759
+ },
+ {
+ "epoch": 80.0,
+ "grad_norm": 0.015248682349920273,
+ "learning_rate": 0.0006,
+ "loss": 4.261106014251709,
+ "step": 5760
+ },
+ {
+ "epoch": 80.0,
+ "eval_loss": 4.610264301300049,
+ "eval_runtime": 56.514,
+ "eval_samples_per_second": 43.211,
+ "eval_steps_per_second": 1.362,
+ "step": 5760
+ },
+ {
+ "epoch": 80.01397379912663,
+ "grad_norm": 0.014730271883308887,
+ "learning_rate": 0.0006,
+ "loss": 4.108556747436523,
+ "step": 5761
+ },
+ {
+ "epoch": 80.02794759825328,
+ "grad_norm": 0.014575716108083725,
+ "learning_rate": 0.0006,
+ "loss": 4.169658660888672,
+ "step": 5762
+ },
+ {
+ "epoch": 80.04192139737991,
+ "grad_norm": 0.014706958085298538,
+ "learning_rate": 0.0006,
+ "loss": 4.252662658691406,
+ "step": 5763
+ },
+ {
+ "epoch": 80.05589519650655,
+ "grad_norm": 0.01561224739998579,
+ "learning_rate": 0.0006,
+ "loss": 4.193892955780029,
+ "step": 5764
+ },
+ {
+ "epoch": 80.06986899563319,
+ "grad_norm": 0.014702324755489826,
+ "learning_rate": 0.0006,
+ "loss": 4.1889238357543945,
+ "step": 5765
+ },
+ {
+ "epoch": 80.08384279475983,
+ "grad_norm": 0.015565741807222366,
+ "learning_rate": 0.0006,
+ "loss": 4.243549823760986,
+ "step": 5766
+ },
+ {
+ "epoch": 80.09781659388646,
+ "grad_norm": 0.01647181063890457,
+ "learning_rate": 0.0006,
+ "loss": 4.275446891784668,
+ "step": 5767
+ },
+ {
+ "epoch": 80.1117903930131,
+ "grad_norm": 0.018177716061472893,
+ "learning_rate": 0.0006,
+ "loss": 4.12040376663208,
+ "step": 5768
+ },
+ {
+ "epoch": 80.12576419213974,
+ "grad_norm": 0.01745191030204296,
+ "learning_rate": 0.0006,
+ "loss": 4.140371322631836,
+ "step": 5769
+ },
+ {
+ "epoch": 80.13973799126637,
+ "grad_norm": 0.01568993367254734,
+ "learning_rate": 0.0006,
+ "loss": 4.258280277252197,
+ "step": 5770
+ },
+ {
+ "epoch": 80.15371179039302,
+ "grad_norm": 0.01636597141623497,
+ "learning_rate": 0.0006,
+ "loss": 4.179052352905273,
+ "step": 5771
+ },
+ {
+ "epoch": 80.16768558951965,
+ "grad_norm": 0.016839100047945976,
+ "learning_rate": 0.0006,
+ "loss": 4.127636432647705,
+ "step": 5772
+ },
+ {
+ "epoch": 80.18165938864628,
+ "grad_norm": 0.015471016988158226,
+ "learning_rate": 0.0006,
+ "loss": 4.181325912475586,
+ "step": 5773
+ },
+ {
+ "epoch": 80.19563318777293,
+ "grad_norm": 0.015836089849472046,
+ "learning_rate": 0.0006,
+ "loss": 4.142168998718262,
+ "step": 5774
+ },
+ {
+ "epoch": 80.20960698689956,
+ "grad_norm": 0.014782954007387161,
+ "learning_rate": 0.0006,
+ "loss": 4.308799743652344,
+ "step": 5775
+ },
+ {
+ "epoch": 80.2235807860262,
+ "grad_norm": 0.014040893875062466,
+ "learning_rate": 0.0006,
+ "loss": 4.216936111450195,
+ "step": 5776
+ },
+ {
+ "epoch": 80.23755458515284,
+ "grad_norm": 0.01420280709862709,
+ "learning_rate": 0.0006,
+ "loss": 4.235230445861816,
+ "step": 5777
+ },
+ {
+ "epoch": 80.25152838427948,
+ "grad_norm": 0.013341007754206657,
+ "learning_rate": 0.0006,
+ "loss": 4.369109153747559,
+ "step": 5778
+ },
+ {
+ "epoch": 80.26550218340611,
+ "grad_norm": 0.014208327978849411,
+ "learning_rate": 0.0006,
+ "loss": 4.219112396240234,
+ "step": 5779
+ },
+ {
+ "epoch": 80.27947598253274,
+ "grad_norm": 0.013825052417814732,
+ "learning_rate": 0.0006,
+ "loss": 4.1983489990234375,
+ "step": 5780
+ },
+ {
+ "epoch": 80.29344978165939,
+ "grad_norm": 0.012916515581309795,
+ "learning_rate": 0.0006,
+ "loss": 4.204582691192627,
+ "step": 5781
+ },
+ {
+ "epoch": 80.30742358078602,
+ "grad_norm": 0.013071243651211262,
+ "learning_rate": 0.0006,
+ "loss": 4.17643928527832,
+ "step": 5782
+ },
+ {
+ "epoch": 80.32139737991267,
+ "grad_norm": 0.015507174655795097,
+ "learning_rate": 0.0006,
+ "loss": 4.172976016998291,
+ "step": 5783
+ },
+ {
+ "epoch": 80.3353711790393,
+ "grad_norm": 0.014967243187129498,
+ "learning_rate": 0.0006,
+ "loss": 4.141168594360352,
+ "step": 5784
+ },
+ {
+ "epoch": 80.34934497816595,
+ "grad_norm": 0.015629924833774567,
+ "learning_rate": 0.0006,
+ "loss": 4.253526210784912,
+ "step": 5785
+ },
+ {
+ "epoch": 80.36331877729258,
+ "grad_norm": 0.01607954502105713,
+ "learning_rate": 0.0006,
+ "loss": 4.249395370483398,
+ "step": 5786
+ },
+ {
+ "epoch": 80.37729257641921,
+ "grad_norm": 0.01500706560909748,
+ "learning_rate": 0.0006,
+ "loss": 4.279012680053711,
+ "step": 5787
+ },
+ {
+ "epoch": 80.39126637554585,
+ "grad_norm": 0.014343900606036186,
+ "learning_rate": 0.0006,
+ "loss": 4.268509864807129,
+ "step": 5788
+ },
+ {
+ "epoch": 80.40524017467249,
+ "grad_norm": 0.014682197012007236,
+ "learning_rate": 0.0006,
+ "loss": 4.206334114074707,
+ "step": 5789
+ },
+ {
+ "epoch": 80.41921397379913,
+ "grad_norm": 0.014989163726568222,
+ "learning_rate": 0.0006,
+ "loss": 4.143601417541504,
+ "step": 5790
+ },
+ {
+ "epoch": 80.43318777292576,
+ "grad_norm": 0.015001608058810234,
+ "learning_rate": 0.0006,
+ "loss": 4.164111137390137,
+ "step": 5791
+ },
+ {
+ "epoch": 80.44716157205241,
+ "grad_norm": 0.015222722664475441,
+ "learning_rate": 0.0006,
+ "loss": 4.2369914054870605,
+ "step": 5792
+ },
+ {
+ "epoch": 80.46113537117904,
+ "grad_norm": 0.014126183465123177,
+ "learning_rate": 0.0006,
+ "loss": 4.2209930419921875,
+ "step": 5793
+ },
+ {
+ "epoch": 80.47510917030567,
+ "grad_norm": 0.013065959326922894,
+ "learning_rate": 0.0006,
+ "loss": 4.231184959411621,
+ "step": 5794
+ },
+ {
+ "epoch": 80.48908296943232,
+ "grad_norm": 0.013833962380886078,
+ "learning_rate": 0.0006,
+ "loss": 4.226434707641602,
+ "step": 5795
+ },
+ {
+ "epoch": 80.50305676855895,
+ "grad_norm": 0.01387125626206398,
+ "learning_rate": 0.0006,
+ "loss": 4.241328716278076,
+ "step": 5796
+ },
+ {
+ "epoch": 80.5170305676856,
+ "grad_norm": 0.014145063236355782,
+ "learning_rate": 0.0006,
+ "loss": 4.2268805503845215,
+ "step": 5797
+ },
+ {
+ "epoch": 80.53100436681223,
+ "grad_norm": 0.012990240007638931,
+ "learning_rate": 0.0006,
+ "loss": 4.1718339920043945,
+ "step": 5798
+ },
+ {
+ "epoch": 80.54497816593886,
+ "grad_norm": 0.014185411855578423,
+ "learning_rate": 0.0006,
+ "loss": 4.232397556304932,
+ "step": 5799
+ },
+ {
+ "epoch": 80.5589519650655,
+ "grad_norm": 0.014260523952543736,
+ "learning_rate": 0.0006,
+ "loss": 4.282734394073486,
+ "step": 5800
+ },
+ {
+ "epoch": 80.57292576419214,
+ "grad_norm": 0.012542120181024075,
+ "learning_rate": 0.0006,
+ "loss": 4.134551525115967,
+ "step": 5801
+ },
+ {
+ "epoch": 80.58689956331878,
+ "grad_norm": 0.013078657910227776,
+ "learning_rate": 0.0006,
+ "loss": 4.229815483093262,
+ "step": 5802
+ },
+ {
+ "epoch": 80.60087336244541,
+ "grad_norm": 0.012821056880056858,
+ "learning_rate": 0.0006,
+ "loss": 4.208278179168701,
+ "step": 5803
+ },
+ {
+ "epoch": 80.61484716157206,
+ "grad_norm": 0.012964180670678616,
+ "learning_rate": 0.0006,
+ "loss": 4.188477039337158,
+ "step": 5804
+ },
+ {
+ "epoch": 80.62882096069869,
+ "grad_norm": 0.01631746254861355,
+ "learning_rate": 0.0006,
+ "loss": 4.3214497566223145,
+ "step": 5805
+ },
+ {
+ "epoch": 80.64279475982532,
+ "grad_norm": 0.01726640947163105,
+ "learning_rate": 0.0006,
+ "loss": 4.279359817504883,
+ "step": 5806
+ },
+ {
+ "epoch": 80.65676855895197,
+ "grad_norm": 0.018778258934617043,
+ "learning_rate": 0.0006,
+ "loss": 4.294047832489014,
+ "step": 5807
+ },
+ {
+ "epoch": 80.6707423580786,
+ "grad_norm": 0.01796831749379635,
+ "learning_rate": 0.0006,
+ "loss": 4.314162254333496,
+ "step": 5808
+ },
+ {
+ "epoch": 80.68471615720524,
+ "grad_norm": 0.016495538875460625,
+ "learning_rate": 0.0006,
+ "loss": 4.273871421813965,
+ "step": 5809
+ },
+ {
+ "epoch": 80.69868995633188,
+ "grad_norm": 0.013590306043624878,
+ "learning_rate": 0.0006,
+ "loss": 4.14961051940918,
+ "step": 5810
+ },
+ {
+ "epoch": 80.71266375545852,
+ "grad_norm": 0.014670426957309246,
+ "learning_rate": 0.0006,
+ "loss": 4.211853981018066,
+ "step": 5811
+ },
+ {
+ "epoch": 80.72663755458515,
+ "grad_norm": 0.01641341857612133,
+ "learning_rate": 0.0006,
+ "loss": 4.211948871612549,
+ "step": 5812
+ },
+ {
+ "epoch": 80.74061135371178,
+ "grad_norm": 0.016365470364689827,
+ "learning_rate": 0.0006,
+ "loss": 4.200774192810059,
+ "step": 5813
+ },
+ {
+ "epoch": 80.75458515283843,
+ "grad_norm": 0.01613970287144184,
+ "learning_rate": 0.0006,
+ "loss": 4.288322448730469,
+ "step": 5814
+ },
+ {
+ "epoch": 80.76855895196506,
+ "grad_norm": 0.016742777079343796,
+ "learning_rate": 0.0006,
+ "loss": 4.23801326751709,
+ "step": 5815
+ },
+ {
+ "epoch": 80.78253275109171,
+ "grad_norm": 0.015351069159805775,
+ "learning_rate": 0.0006,
+ "loss": 4.263832092285156,
+ "step": 5816
+ },
+ {
+ "epoch": 80.79650655021834,
+ "grad_norm": 0.013435336761176586,
+ "learning_rate": 0.0006,
+ "loss": 4.160857200622559,
+ "step": 5817
+ },
+ {
+ "epoch": 80.81048034934499,
+ "grad_norm": 0.01474886853247881,
+ "learning_rate": 0.0006,
+ "loss": 4.216025352478027,
+ "step": 5818
+ },
+ {
+ "epoch": 80.82445414847162,
+ "grad_norm": 0.01586935855448246,
+ "learning_rate": 0.0006,
+ "loss": 4.256402969360352,
+ "step": 5819
+ },
+ {
+ "epoch": 80.83842794759825,
+ "grad_norm": 0.014596572145819664,
+ "learning_rate": 0.0006,
+ "loss": 4.142584800720215,
+ "step": 5820
+ },
+ {
+ "epoch": 80.8524017467249,
+ "grad_norm": 0.01501593366265297,
+ "learning_rate": 0.0006,
+ "loss": 4.250462055206299,
+ "step": 5821
+ },
+ {
+ "epoch": 80.86637554585153,
+ "grad_norm": 0.014773511327803135,
+ "learning_rate": 0.0006,
+ "loss": 4.253305912017822,
+ "step": 5822
+ },
+ {
+ "epoch": 80.88034934497817,
+ "grad_norm": 0.013246326707303524,
+ "learning_rate": 0.0006,
+ "loss": 4.170881271362305,
+ "step": 5823
+ },
+ {
+ "epoch": 80.8943231441048,
+ "grad_norm": 0.013013756833970547,
+ "learning_rate": 0.0006,
+ "loss": 4.264721870422363,
+ "step": 5824
+ },
+ {
+ "epoch": 80.90829694323143,
+ "grad_norm": 0.013155903667211533,
+ "learning_rate": 0.0006,
+ "loss": 4.274667739868164,
+ "step": 5825
+ },
+ {
+ "epoch": 80.92227074235808,
+ "grad_norm": 0.01448952779173851,
+ "learning_rate": 0.0006,
+ "loss": 4.24777889251709,
+ "step": 5826
+ },
+ {
+ "epoch": 80.93624454148471,
+ "grad_norm": 0.01415231078863144,
+ "learning_rate": 0.0006,
+ "loss": 4.22014856338501,
+ "step": 5827
+ },
+ {
+ "epoch": 80.95021834061136,
+ "grad_norm": 0.014796273782849312,
+ "learning_rate": 0.0006,
+ "loss": 4.22036600112915,
+ "step": 5828
+ },
+ {
+ "epoch": 80.96419213973799,
+ "grad_norm": 0.013005326502025127,
+ "learning_rate": 0.0006,
+ "loss": 4.276123046875,
+ "step": 5829
+ },
+ {
+ "epoch": 80.97816593886463,
+ "grad_norm": 0.012851177714765072,
+ "learning_rate": 0.0006,
+ "loss": 4.248575687408447,
+ "step": 5830
+ },
+ {
+ "epoch": 80.99213973799127,
+ "grad_norm": 0.013740002177655697,
+ "learning_rate": 0.0006,
+ "loss": 4.051140308380127,
+ "step": 5831
+ },
+ {
+ "epoch": 81.0,
+ "grad_norm": 0.015296104364097118,
+ "learning_rate": 0.0006,
+ "loss": 4.251155853271484,
+ "step": 5832
+ },
+ {
+ "epoch": 81.0,
+ "eval_loss": 4.594127655029297,
+ "eval_runtime": 56.0959,
+ "eval_samples_per_second": 43.533,
+ "eval_steps_per_second": 1.373,
+ "step": 5832
+ },
+ {
+ "epoch": 81.01397379912663,
+ "grad_norm": 0.015804558992385864,
+ "learning_rate": 0.0006,
+ "loss": 4.152894020080566,
+ "step": 5833
+ },
+ {
+ "epoch": 81.02794759825328,
+ "grad_norm": 0.017000854015350342,
+ "learning_rate": 0.0006,
+ "loss": 4.266067981719971,
+ "step": 5834
+ },
+ {
+ "epoch": 81.04192139737991,
+ "grad_norm": 0.016351008787751198,
+ "learning_rate": 0.0006,
+ "loss": 4.192703723907471,
+ "step": 5835
+ },
+ {
+ "epoch": 81.05589519650655,
+ "grad_norm": 0.014703895896673203,
+ "learning_rate": 0.0006,
+ "loss": 4.23554801940918,
+ "step": 5836
+ },
+ {
+ "epoch": 81.06986899563319,
+ "grad_norm": 0.01619596965610981,
+ "learning_rate": 0.0006,
+ "loss": 4.233000755310059,
+ "step": 5837
+ },
+ {
+ "epoch": 81.08384279475983,
+ "grad_norm": 0.01815493032336235,
+ "learning_rate": 0.0006,
+ "loss": 4.200490951538086,
+ "step": 5838
+ },
+ {
+ "epoch": 81.09781659388646,
+ "grad_norm": 0.018305106088519096,
+ "learning_rate": 0.0006,
+ "loss": 4.154242515563965,
+ "step": 5839
+ },
+ {
+ "epoch": 81.1117903930131,
+ "grad_norm": 0.0200856514275074,
+ "learning_rate": 0.0006,
+ "loss": 4.25820779800415,
+ "step": 5840
+ },
+ {
+ "epoch": 81.12576419213974,
+ "grad_norm": 0.02101440541446209,
+ "learning_rate": 0.0006,
+ "loss": 4.168209075927734,
+ "step": 5841
+ },
+ {
+ "epoch": 81.13973799126637,
+ "grad_norm": 0.022829681634902954,
+ "learning_rate": 0.0006,
+ "loss": 4.25560998916626,
+ "step": 5842
+ },
+ {
+ "epoch": 81.15371179039302,
+ "grad_norm": 0.0195082426071167,
+ "learning_rate": 0.0006,
+ "loss": 4.2035627365112305,
+ "step": 5843
+ },
+ {
+ "epoch": 81.16768558951965,
+ "grad_norm": 0.01716248318552971,
+ "learning_rate": 0.0006,
+ "loss": 4.11322021484375,
+ "step": 5844
+ },
+ {
+ "epoch": 81.18165938864628,
+ "grad_norm": 0.02034664712846279,
+ "learning_rate": 0.0006,
+ "loss": 4.179389953613281,
+ "step": 5845
+ },
+ {
+ "epoch": 81.19563318777293,
+ "grad_norm": 0.023578710854053497,
+ "learning_rate": 0.0006,
+ "loss": 4.265695571899414,
+ "step": 5846
+ },
+ {
+ "epoch": 81.20960698689956,
+ "grad_norm": 0.019244182854890823,
+ "learning_rate": 0.0006,
+ "loss": 4.292043209075928,
+ "step": 5847
+ },
+ {
+ "epoch": 81.2235807860262,
+ "grad_norm": 0.01577291637659073,
+ "learning_rate": 0.0006,
+ "loss": 4.250776290893555,
+ "step": 5848
+ },
+ {
+ "epoch": 81.23755458515284,
+ "grad_norm": 0.01617090217769146,
+ "learning_rate": 0.0006,
+ "loss": 4.193098068237305,
+ "step": 5849
+ },
+ {
+ "epoch": 81.25152838427948,
+ "grad_norm": 0.016090184450149536,
+ "learning_rate": 0.0006,
+ "loss": 4.112357139587402,
+ "step": 5850
+ },
+ {
+ "epoch": 81.26550218340611,
+ "grad_norm": 0.01441365759819746,
+ "learning_rate": 0.0006,
+ "loss": 4.114147186279297,
+ "step": 5851
+ },
+ {
+ "epoch": 81.27947598253274,
+ "grad_norm": 0.014122162945568562,
+ "learning_rate": 0.0006,
+ "loss": 4.276175022125244,
+ "step": 5852
+ },
+ {
+ "epoch": 81.29344978165939,
+ "grad_norm": 0.014859487302601337,
+ "learning_rate": 0.0006,
+ "loss": 4.094699859619141,
+ "step": 5853
+ },
+ {
+ "epoch": 81.30742358078602,
+ "grad_norm": 0.013624267652630806,
+ "learning_rate": 0.0006,
+ "loss": 4.2648773193359375,
+ "step": 5854
+ },
+ {
+ "epoch": 81.32139737991267,
+ "grad_norm": 0.01417913381010294,
+ "learning_rate": 0.0006,
+ "loss": 4.228567123413086,
+ "step": 5855
+ },
+ {
+ "epoch": 81.3353711790393,
+ "grad_norm": 0.014560805633664131,
+ "learning_rate": 0.0006,
+ "loss": 4.148930549621582,
+ "step": 5856
+ },
+ {
+ "epoch": 81.34934497816595,
+ "grad_norm": 0.013662154786288738,
+ "learning_rate": 0.0006,
+ "loss": 4.192107677459717,
+ "step": 5857
+ },
+ {
+ "epoch": 81.36331877729258,
+ "grad_norm": 0.013434980995953083,
+ "learning_rate": 0.0006,
+ "loss": 4.162052154541016,
+ "step": 5858
+ },
+ {
+ "epoch": 81.37729257641921,
+ "grad_norm": 0.01300742756575346,
+ "learning_rate": 0.0006,
+ "loss": 4.262385368347168,
+ "step": 5859
+ },
+ {
+ "epoch": 81.39126637554585,
+ "grad_norm": 0.012544351629912853,
+ "learning_rate": 0.0006,
+ "loss": 4.215371131896973,
+ "step": 5860
+ },
+ {
+ "epoch": 81.40524017467249,
+ "grad_norm": 0.013663645833730698,
+ "learning_rate": 0.0006,
+ "loss": 4.247312545776367,
+ "step": 5861
+ },
+ {
+ "epoch": 81.41921397379913,
+ "grad_norm": 0.013127296231687069,
+ "learning_rate": 0.0006,
+ "loss": 4.087685585021973,
+ "step": 5862
+ },
+ {
+ "epoch": 81.43318777292576,
+ "grad_norm": 0.014081938192248344,
+ "learning_rate": 0.0006,
+ "loss": 4.272648811340332,
+ "step": 5863
+ },
+ {
+ "epoch": 81.44716157205241,
+ "grad_norm": 0.015470641665160656,
+ "learning_rate": 0.0006,
+ "loss": 4.278578758239746,
+ "step": 5864
+ },
+ {
+ "epoch": 81.46113537117904,
+ "grad_norm": 0.016524532809853554,
+ "learning_rate": 0.0006,
+ "loss": 4.118268013000488,
+ "step": 5865
+ },
+ {
+ "epoch": 81.47510917030567,
+ "grad_norm": 0.016242722049355507,
+ "learning_rate": 0.0006,
+ "loss": 4.182405471801758,
+ "step": 5866
+ },
+ {
+ "epoch": 81.48908296943232,
+ "grad_norm": 0.01557017583400011,
+ "learning_rate": 0.0006,
+ "loss": 4.229862213134766,
+ "step": 5867
+ },
+ {
+ "epoch": 81.50305676855895,
+ "grad_norm": 0.01562870666384697,
+ "learning_rate": 0.0006,
+ "loss": 4.27754020690918,
+ "step": 5868
+ },
+ {
+ "epoch": 81.5170305676856,
+ "grad_norm": 0.018606074154376984,
+ "learning_rate": 0.0006,
+ "loss": 4.148681640625,
+ "step": 5869
+ },
+ {
+ "epoch": 81.53100436681223,
+ "grad_norm": 0.017740730196237564,
+ "learning_rate": 0.0006,
+ "loss": 4.321932792663574,
+ "step": 5870
+ },
+ {
+ "epoch": 81.54497816593886,
+ "grad_norm": 0.014517677947878838,
+ "learning_rate": 0.0006,
+ "loss": 4.22862434387207,
+ "step": 5871
+ },
+ {
+ "epoch": 81.5589519650655,
+ "grad_norm": 0.01389714702963829,
+ "learning_rate": 0.0006,
+ "loss": 4.198000907897949,
+ "step": 5872
+ },
+ {
+ "epoch": 81.57292576419214,
+ "grad_norm": 0.015348682180047035,
+ "learning_rate": 0.0006,
+ "loss": 4.293485164642334,
+ "step": 5873
+ },
+ {
+ "epoch": 81.58689956331878,
+ "grad_norm": 0.015965791419148445,
+ "learning_rate": 0.0006,
+ "loss": 4.23481559753418,
+ "step": 5874
+ },
+ {
+ "epoch": 81.60087336244541,
+ "grad_norm": 0.015662726014852524,
+ "learning_rate": 0.0006,
+ "loss": 4.158148765563965,
+ "step": 5875
+ },
+ {
+ "epoch": 81.61484716157206,
+ "grad_norm": 0.014402080327272415,
+ "learning_rate": 0.0006,
+ "loss": 4.132723808288574,
+ "step": 5876
+ },
+ {
+ "epoch": 81.62882096069869,
+ "grad_norm": 0.016846125945448875,
+ "learning_rate": 0.0006,
+ "loss": 4.1654462814331055,
+ "step": 5877
+ },
+ {
+ "epoch": 81.64279475982532,
+ "grad_norm": 0.0196097232401371,
+ "learning_rate": 0.0006,
+ "loss": 4.228450298309326,
+ "step": 5878
+ },
+ {
+ "epoch": 81.65676855895197,
+ "grad_norm": 0.016005294397473335,
+ "learning_rate": 0.0006,
+ "loss": 4.230222225189209,
+ "step": 5879
+ },
+ {
+ "epoch": 81.6707423580786,
+ "grad_norm": 0.014685059897601604,
+ "learning_rate": 0.0006,
+ "loss": 4.1367387771606445,
+ "step": 5880
+ },
+ {
+ "epoch": 81.68471615720524,
+ "grad_norm": 0.01647144928574562,
+ "learning_rate": 0.0006,
+ "loss": 4.096675395965576,
+ "step": 5881
+ },
+ {
+ "epoch": 81.69868995633188,
+ "grad_norm": 0.015978610143065453,
+ "learning_rate": 0.0006,
+ "loss": 4.232292175292969,
+ "step": 5882
+ },
+ {
+ "epoch": 81.71266375545852,
+ "grad_norm": 0.01510288193821907,
+ "learning_rate": 0.0006,
+ "loss": 4.274591445922852,
+ "step": 5883
+ },
+ {
+ "epoch": 81.72663755458515,
+ "grad_norm": 0.014942476525902748,
+ "learning_rate": 0.0006,
+ "loss": 4.161482810974121,
+ "step": 5884
+ },
+ {
+ "epoch": 81.74061135371178,
+ "grad_norm": 0.014595299027860165,
+ "learning_rate": 0.0006,
+ "loss": 4.204291343688965,
+ "step": 5885
+ },
+ {
+ "epoch": 81.75458515283843,
+ "grad_norm": 0.014952429570257664,
+ "learning_rate": 0.0006,
+ "loss": 4.242499351501465,
+ "step": 5886
+ },
+ {
+ "epoch": 81.76855895196506,
+ "grad_norm": 0.014761554077267647,
+ "learning_rate": 0.0006,
+ "loss": 4.20393180847168,
+ "step": 5887
+ },
+ {
+ "epoch": 81.78253275109171,
+ "grad_norm": 0.014749647118151188,
+ "learning_rate": 0.0006,
+ "loss": 4.227565765380859,
+ "step": 5888
+ },
+ {
+ "epoch": 81.79650655021834,
+ "grad_norm": 0.014384493231773376,
+ "learning_rate": 0.0006,
+ "loss": 4.282228946685791,
+ "step": 5889
+ },
+ {
+ "epoch": 81.81048034934499,
+ "grad_norm": 0.01260330155491829,
+ "learning_rate": 0.0006,
+ "loss": 4.2882866859436035,
+ "step": 5890
+ },
+ {
+ "epoch": 81.82445414847162,
+ "grad_norm": 0.014055311679840088,
+ "learning_rate": 0.0006,
+ "loss": 4.22351598739624,
+ "step": 5891
+ },
+ {
+ "epoch": 81.83842794759825,
+ "grad_norm": 0.013674065470695496,
+ "learning_rate": 0.0006,
+ "loss": 4.240716934204102,
+ "step": 5892
+ },
+ {
+ "epoch": 81.8524017467249,
+ "grad_norm": 0.01225196197628975,
+ "learning_rate": 0.0006,
+ "loss": 4.2958598136901855,
+ "step": 5893
+ },
+ {
+ "epoch": 81.86637554585153,
+ "grad_norm": 0.013538680970668793,
+ "learning_rate": 0.0006,
+ "loss": 4.24342679977417,
+ "step": 5894
+ },
+ {
+ "epoch": 81.88034934497817,
+ "grad_norm": 0.013151394203305244,
+ "learning_rate": 0.0006,
+ "loss": 4.081999778747559,
+ "step": 5895
+ },
+ {
+ "epoch": 81.8943231441048,
+ "grad_norm": 0.014260678552091122,
+ "learning_rate": 0.0006,
+ "loss": 4.156423568725586,
+ "step": 5896
+ },
+ {
+ "epoch": 81.90829694323143,
+ "grad_norm": 0.01329776830971241,
+ "learning_rate": 0.0006,
+ "loss": 4.1713995933532715,
+ "step": 5897
+ },
+ {
+ "epoch": 81.92227074235808,
+ "grad_norm": 0.014623276889324188,
+ "learning_rate": 0.0006,
+ "loss": 4.244891166687012,
+ "step": 5898
+ },
+ {
+ "epoch": 81.93624454148471,
+ "grad_norm": 0.015760324895381927,
+ "learning_rate": 0.0006,
+ "loss": 4.076528549194336,
+ "step": 5899
+ },
+ {
+ "epoch": 81.95021834061136,
+ "grad_norm": 0.014331766404211521,
+ "learning_rate": 0.0006,
+ "loss": 4.283037185668945,
+ "step": 5900
+ },
+ {
+ "epoch": 81.96419213973799,
+ "grad_norm": 0.012383842840790749,
+ "learning_rate": 0.0006,
+ "loss": 4.117308616638184,
+ "step": 5901
+ },
+ {
+ "epoch": 81.97816593886463,
+ "grad_norm": 0.01250660140067339,
+ "learning_rate": 0.0006,
+ "loss": 4.241250991821289,
+ "step": 5902
+ },
+ {
+ "epoch": 81.99213973799127,
+ "grad_norm": 0.012276900932192802,
+ "learning_rate": 0.0006,
+ "loss": 4.141091823577881,
+ "step": 5903
+ },
+ {
+ "epoch": 82.0,
+ "grad_norm": 0.01588737778365612,
+ "learning_rate": 0.0006,
+ "loss": 4.065532684326172,
+ "step": 5904
+ },
+ {
+ "epoch": 82.0,
+ "eval_loss": 4.612304210662842,
+ "eval_runtime": 57.5411,
+ "eval_samples_per_second": 42.439,
+ "eval_steps_per_second": 1.338,
+ "step": 5904
+ },
+ {
+ "epoch": 82.01397379912663,
+ "grad_norm": 0.017433544620871544,
+ "learning_rate": 0.0006,
+ "loss": 4.202497482299805,
+ "step": 5905
+ },
+ {
+ "epoch": 82.02794759825328,
+ "grad_norm": 0.017260415479540825,
+ "learning_rate": 0.0006,
+ "loss": 4.099949359893799,
+ "step": 5906
+ },
+ {
+ "epoch": 82.04192139737991,
+ "grad_norm": 0.01577908545732498,
+ "learning_rate": 0.0006,
+ "loss": 4.17806339263916,
+ "step": 5907
+ },
+ {
+ "epoch": 82.05589519650655,
+ "grad_norm": 0.016093818470835686,
+ "learning_rate": 0.0006,
+ "loss": 4.219491004943848,
+ "step": 5908
+ },
+ {
+ "epoch": 82.06986899563319,
+ "grad_norm": 0.015426302328705788,
+ "learning_rate": 0.0006,
+ "loss": 4.207982063293457,
+ "step": 5909
+ },
+ {
+ "epoch": 82.08384279475983,
+ "grad_norm": 0.01605442352592945,
+ "learning_rate": 0.0006,
+ "loss": 4.265942573547363,
+ "step": 5910
+ },
+ {
+ "epoch": 82.09781659388646,
+ "grad_norm": 0.01787225715816021,
+ "learning_rate": 0.0006,
+ "loss": 4.151535511016846,
+ "step": 5911
+ },
+ {
+ "epoch": 82.1117903930131,
+ "grad_norm": 0.01750164106488228,
+ "learning_rate": 0.0006,
+ "loss": 4.133655071258545,
+ "step": 5912
+ },
+ {
+ "epoch": 82.12576419213974,
+ "grad_norm": 0.01602957211434841,
+ "learning_rate": 0.0006,
+ "loss": 4.176972389221191,
+ "step": 5913
+ },
+ {
+ "epoch": 82.13973799126637,
+ "grad_norm": 0.01553812064230442,
+ "learning_rate": 0.0006,
+ "loss": 4.125692367553711,
+ "step": 5914
+ },
+ {
+ "epoch": 82.15371179039302,
+ "grad_norm": 0.013978508301079273,
+ "learning_rate": 0.0006,
+ "loss": 4.207271575927734,
+ "step": 5915
+ },
+ {
+ "epoch": 82.16768558951965,
+ "grad_norm": 0.013561118394136429,
+ "learning_rate": 0.0006,
+ "loss": 4.238930702209473,
+ "step": 5916
+ },
+ {
+ "epoch": 82.18165938864628,
+ "grad_norm": 0.013285507448017597,
+ "learning_rate": 0.0006,
+ "loss": 4.129464626312256,
+ "step": 5917
+ },
+ {
+ "epoch": 82.19563318777293,
+ "grad_norm": 0.013903839513659477,
+ "learning_rate": 0.0006,
+ "loss": 4.214109420776367,
+ "step": 5918
+ },
+ {
+ "epoch": 82.20960698689956,
+ "grad_norm": 0.014426837675273418,
+ "learning_rate": 0.0006,
+ "loss": 4.143196105957031,
+ "step": 5919
+ },
+ {
+ "epoch": 82.2235807860262,
+ "grad_norm": 0.015347165986895561,
+ "learning_rate": 0.0006,
+ "loss": 4.235439300537109,
+ "step": 5920
+ },
+ {
+ "epoch": 82.23755458515284,
+ "grad_norm": 0.014266039244830608,
+ "learning_rate": 0.0006,
+ "loss": 3.991617202758789,
+ "step": 5921
+ },
+ {
+ "epoch": 82.25152838427948,
+ "grad_norm": 0.013101593591272831,
+ "learning_rate": 0.0006,
+ "loss": 4.266017913818359,
+ "step": 5922
+ },
+ {
+ "epoch": 82.26550218340611,
+ "grad_norm": 0.013741881586611271,
+ "learning_rate": 0.0006,
+ "loss": 4.224879741668701,
+ "step": 5923
+ },
+ {
+ "epoch": 82.27947598253274,
+ "grad_norm": 0.014206945896148682,
+ "learning_rate": 0.0006,
+ "loss": 4.102272987365723,
+ "step": 5924
+ },
+ {
+ "epoch": 82.29344978165939,
+ "grad_norm": 0.013798723928630352,
+ "learning_rate": 0.0006,
+ "loss": 4.30418586730957,
+ "step": 5925
+ },
+ {
+ "epoch": 82.30742358078602,
+ "grad_norm": 0.015470122918486595,
+ "learning_rate": 0.0006,
+ "loss": 4.2513322830200195,
+ "step": 5926
+ },
+ {
+ "epoch": 82.32139737991267,
+ "grad_norm": 0.01655442826449871,
+ "learning_rate": 0.0006,
+ "loss": 4.223002910614014,
+ "step": 5927
+ },
+ {
+ "epoch": 82.3353711790393,
+ "grad_norm": 0.017170004546642303,
+ "learning_rate": 0.0006,
+ "loss": 4.253904819488525,
+ "step": 5928
+ },
+ {
+ "epoch": 82.34934497816595,
+ "grad_norm": 0.017595812678337097,
+ "learning_rate": 0.0006,
+ "loss": 4.184927940368652,
+ "step": 5929
+ },
+ {
+ "epoch": 82.36331877729258,
+ "grad_norm": 0.015073256567120552,
+ "learning_rate": 0.0006,
+ "loss": 4.202239036560059,
+ "step": 5930
+ },
+ {
+ "epoch": 82.37729257641921,
+ "grad_norm": 0.014310941100120544,
+ "learning_rate": 0.0006,
+ "loss": 4.124934673309326,
+ "step": 5931
+ },
+ {
+ "epoch": 82.39126637554585,
+ "grad_norm": 0.013671821914613247,
+ "learning_rate": 0.0006,
+ "loss": 4.138919353485107,
+ "step": 5932
+ },
+ {
+ "epoch": 82.40524017467249,
+ "grad_norm": 0.014103314839303493,
+ "learning_rate": 0.0006,
+ "loss": 4.211289405822754,
+ "step": 5933
+ },
+ {
+ "epoch": 82.41921397379913,
+ "grad_norm": 0.014272668398916721,
+ "learning_rate": 0.0006,
+ "loss": 4.06290340423584,
+ "step": 5934
+ },
+ {
+ "epoch": 82.43318777292576,
+ "grad_norm": 0.01387016475200653,
+ "learning_rate": 0.0006,
+ "loss": 4.101334571838379,
+ "step": 5935
+ },
+ {
+ "epoch": 82.44716157205241,
+ "grad_norm": 0.013318683952093124,
+ "learning_rate": 0.0006,
+ "loss": 4.206409931182861,
+ "step": 5936
+ },
+ {
+ "epoch": 82.46113537117904,
+ "grad_norm": 0.013781268149614334,
+ "learning_rate": 0.0006,
+ "loss": 4.189474105834961,
+ "step": 5937
+ },
+ {
+ "epoch": 82.47510917030567,
+ "grad_norm": 0.013195851817727089,
+ "learning_rate": 0.0006,
+ "loss": 4.276607036590576,
+ "step": 5938
+ },
+ {
+ "epoch": 82.48908296943232,
+ "grad_norm": 0.014358163811266422,
+ "learning_rate": 0.0006,
+ "loss": 4.203262805938721,
+ "step": 5939
+ },
+ {
+ "epoch": 82.50305676855895,
+ "grad_norm": 0.01403636671602726,
+ "learning_rate": 0.0006,
+ "loss": 4.079258441925049,
+ "step": 5940
+ },
+ {
+ "epoch": 82.5170305676856,
+ "grad_norm": 0.013272535055875778,
+ "learning_rate": 0.0006,
+ "loss": 4.19197940826416,
+ "step": 5941
+ },
+ {
+ "epoch": 82.53100436681223,
+ "grad_norm": 0.013787264935672283,
+ "learning_rate": 0.0006,
+ "loss": 4.12995719909668,
+ "step": 5942
+ },
+ {
+ "epoch": 82.54497816593886,
+ "grad_norm": 0.015020057559013367,
+ "learning_rate": 0.0006,
+ "loss": 4.148890495300293,
+ "step": 5943
+ },
+ {
+ "epoch": 82.5589519650655,
+ "grad_norm": 0.014199157245457172,
+ "learning_rate": 0.0006,
+ "loss": 4.201516151428223,
+ "step": 5944
+ },
+ {
+ "epoch": 82.57292576419214,
+ "grad_norm": 0.013827123679220676,
+ "learning_rate": 0.0006,
+ "loss": 4.240894317626953,
+ "step": 5945
+ },
+ {
+ "epoch": 82.58689956331878,
+ "grad_norm": 0.014767560176551342,
+ "learning_rate": 0.0006,
+ "loss": 4.173320770263672,
+ "step": 5946
+ },
+ {
+ "epoch": 82.60087336244541,
+ "grad_norm": 0.01612078957259655,
+ "learning_rate": 0.0006,
+ "loss": 4.266448020935059,
+ "step": 5947
+ },
+ {
+ "epoch": 82.61484716157206,
+ "grad_norm": 0.016706543043255806,
+ "learning_rate": 0.0006,
+ "loss": 4.229398250579834,
+ "step": 5948
+ },
+ {
+ "epoch": 82.62882096069869,
+ "grad_norm": 0.015630796551704407,
+ "learning_rate": 0.0006,
+ "loss": 4.181282997131348,
+ "step": 5949
+ },
+ {
+ "epoch": 82.64279475982532,
+ "grad_norm": 0.013103844597935677,
+ "learning_rate": 0.0006,
+ "loss": 4.309617042541504,
+ "step": 5950
+ },
+ {
+ "epoch": 82.65676855895197,
+ "grad_norm": 0.016136541962623596,
+ "learning_rate": 0.0006,
+ "loss": 4.224159240722656,
+ "step": 5951
+ },
+ {
+ "epoch": 82.6707423580786,
+ "grad_norm": 0.016160814091563225,
+ "learning_rate": 0.0006,
+ "loss": 4.267970085144043,
+ "step": 5952
+ },
+ {
+ "epoch": 82.68471615720524,
+ "grad_norm": 0.01672680303454399,
+ "learning_rate": 0.0006,
+ "loss": 4.230506896972656,
+ "step": 5953
+ },
+ {
+ "epoch": 82.69868995633188,
+ "grad_norm": 0.0159586351364851,
+ "learning_rate": 0.0006,
+ "loss": 4.159542083740234,
+ "step": 5954
+ },
+ {
+ "epoch": 82.71266375545852,
+ "grad_norm": 0.01461219135671854,
+ "learning_rate": 0.0006,
+ "loss": 4.27863073348999,
+ "step": 5955
+ },
+ {
+ "epoch": 82.72663755458515,
+ "grad_norm": 0.015088613145053387,
+ "learning_rate": 0.0006,
+ "loss": 4.272418975830078,
+ "step": 5956
+ },
+ {
+ "epoch": 82.74061135371178,
+ "grad_norm": 0.01594499684870243,
+ "learning_rate": 0.0006,
+ "loss": 4.271326065063477,
+ "step": 5957
+ },
+ {
+ "epoch": 82.75458515283843,
+ "grad_norm": 0.016648447141051292,
+ "learning_rate": 0.0006,
+ "loss": 4.141478538513184,
+ "step": 5958
+ },
+ {
+ "epoch": 82.76855895196506,
+ "grad_norm": 0.01540316641330719,
+ "learning_rate": 0.0006,
+ "loss": 4.111814498901367,
+ "step": 5959
+ },
+ {
+ "epoch": 82.78253275109171,
+ "grad_norm": 0.01634364202618599,
+ "learning_rate": 0.0006,
+ "loss": 4.3113112449646,
+ "step": 5960
+ },
+ {
+ "epoch": 82.79650655021834,
+ "grad_norm": 0.014768977649509907,
+ "learning_rate": 0.0006,
+ "loss": 4.184217929840088,
+ "step": 5961
+ },
+ {
+ "epoch": 82.81048034934499,
+ "grad_norm": 0.013453755527734756,
+ "learning_rate": 0.0006,
+ "loss": 4.247729301452637,
+ "step": 5962
+ },
+ {
+ "epoch": 82.82445414847162,
+ "grad_norm": 0.014731907285749912,
+ "learning_rate": 0.0006,
+ "loss": 4.352839469909668,
+ "step": 5963
+ },
+ {
+ "epoch": 82.83842794759825,
+ "grad_norm": 0.016047345474362373,
+ "learning_rate": 0.0006,
+ "loss": 4.257283687591553,
+ "step": 5964
+ },
+ {
+ "epoch": 82.8524017467249,
+ "grad_norm": 0.017141170799732208,
+ "learning_rate": 0.0006,
+ "loss": 4.22850227355957,
+ "step": 5965
+ },
+ {
+ "epoch": 82.86637554585153,
+ "grad_norm": 0.01370396837592125,
+ "learning_rate": 0.0006,
+ "loss": 4.1169257164001465,
+ "step": 5966
+ },
+ {
+ "epoch": 82.88034934497817,
+ "grad_norm": 0.012060081586241722,
+ "learning_rate": 0.0006,
+ "loss": 4.195242404937744,
+ "step": 5967
+ },
+ {
+ "epoch": 82.8943231441048,
+ "grad_norm": 0.015528378076851368,
+ "learning_rate": 0.0006,
+ "loss": 4.294939994812012,
+ "step": 5968
+ },
+ {
+ "epoch": 82.90829694323143,
+ "grad_norm": 0.018151041120290756,
+ "learning_rate": 0.0006,
+ "loss": 4.137192726135254,
+ "step": 5969
+ },
+ {
+ "epoch": 82.92227074235808,
+ "grad_norm": 0.017895735800266266,
+ "learning_rate": 0.0006,
+ "loss": 4.174036026000977,
+ "step": 5970
+ },
+ {
+ "epoch": 82.93624454148471,
+ "grad_norm": 0.015610386617481709,
+ "learning_rate": 0.0006,
+ "loss": 4.230136871337891,
+ "step": 5971
+ },
+ {
+ "epoch": 82.95021834061136,
+ "grad_norm": 0.01339928898960352,
+ "learning_rate": 0.0006,
+ "loss": 4.199074745178223,
+ "step": 5972
+ },
+ {
+ "epoch": 82.96419213973799,
+ "grad_norm": 0.015454214997589588,
+ "learning_rate": 0.0006,
+ "loss": 4.216109752655029,
+ "step": 5973
+ },
+ {
+ "epoch": 82.97816593886463,
+ "grad_norm": 0.014978132210671902,
+ "learning_rate": 0.0006,
+ "loss": 4.131511688232422,
+ "step": 5974
+ },
+ {
+ "epoch": 82.99213973799127,
+ "grad_norm": 0.015611122362315655,
+ "learning_rate": 0.0006,
+ "loss": 4.246518135070801,
+ "step": 5975
+ },
+ {
+ "epoch": 83.0,
+ "grad_norm": 0.016682764515280724,
+ "learning_rate": 0.0006,
+ "loss": 4.110680103302002,
+ "step": 5976
+ },
+ {
+ "epoch": 83.0,
+ "eval_loss": 4.641707897186279,
+ "eval_runtime": 57.2799,
+ "eval_samples_per_second": 42.633,
+ "eval_steps_per_second": 1.344,
+ "step": 5976
+ },
+ {
+ "epoch": 83.01397379912663,
+ "grad_norm": 0.015007982030510902,
+ "learning_rate": 0.0006,
+ "loss": 4.100912094116211,
+ "step": 5977
+ },
+ {
+ "epoch": 83.02794759825328,
+ "grad_norm": 0.017609436064958572,
+ "learning_rate": 0.0006,
+ "loss": 4.233399868011475,
+ "step": 5978
+ },
+ {
+ "epoch": 83.04192139737991,
+ "grad_norm": 0.019817272201180458,
+ "learning_rate": 0.0006,
+ "loss": 4.191173076629639,
+ "step": 5979
+ },
+ {
+ "epoch": 83.05589519650655,
+ "grad_norm": 0.017752140760421753,
+ "learning_rate": 0.0006,
+ "loss": 4.044437885284424,
+ "step": 5980
+ },
+ {
+ "epoch": 83.06986899563319,
+ "grad_norm": 0.015482760034501553,
+ "learning_rate": 0.0006,
+ "loss": 4.319931983947754,
+ "step": 5981
+ },
+ {
+ "epoch": 83.08384279475983,
+ "grad_norm": 0.015153988264501095,
+ "learning_rate": 0.0006,
+ "loss": 4.095277309417725,
+ "step": 5982
+ },
+ {
+ "epoch": 83.09781659388646,
+ "grad_norm": 0.017025282606482506,
+ "learning_rate": 0.0006,
+ "loss": 4.235110759735107,
+ "step": 5983
+ },
+ {
+ "epoch": 83.1117903930131,
+ "grad_norm": 0.01566586270928383,
+ "learning_rate": 0.0006,
+ "loss": 4.1934685707092285,
+ "step": 5984
+ },
+ {
+ "epoch": 83.12576419213974,
+ "grad_norm": 0.016419263556599617,
+ "learning_rate": 0.0006,
+ "loss": 4.234419822692871,
+ "step": 5985
+ },
+ {
+ "epoch": 83.13973799126637,
+ "grad_norm": 0.017132675275206566,
+ "learning_rate": 0.0006,
+ "loss": 4.221447944641113,
+ "step": 5986
+ },
+ {
+ "epoch": 83.15371179039302,
+ "grad_norm": 0.016138629987835884,
+ "learning_rate": 0.0006,
+ "loss": 4.199830055236816,
+ "step": 5987
+ },
+ {
+ "epoch": 83.16768558951965,
+ "grad_norm": 0.016210174188017845,
+ "learning_rate": 0.0006,
+ "loss": 4.081717491149902,
+ "step": 5988
+ },
+ {
+ "epoch": 83.18165938864628,
+ "grad_norm": 0.016055308282375336,
+ "learning_rate": 0.0006,
+ "loss": 4.247655391693115,
+ "step": 5989
+ },
+ {
+ "epoch": 83.19563318777293,
+ "grad_norm": 0.016132952645421028,
+ "learning_rate": 0.0006,
+ "loss": 4.262535572052002,
+ "step": 5990
+ },
+ {
+ "epoch": 83.20960698689956,
+ "grad_norm": 0.01779935695230961,
+ "learning_rate": 0.0006,
+ "loss": 4.153120040893555,
+ "step": 5991
+ },
+ {
+ "epoch": 83.2235807860262,
+ "grad_norm": 0.018360646441578865,
+ "learning_rate": 0.0006,
+ "loss": 4.232644081115723,
+ "step": 5992
+ },
+ {
+ "epoch": 83.23755458515284,
+ "grad_norm": 0.019629675894975662,
+ "learning_rate": 0.0006,
+ "loss": 3.9758734703063965,
+ "step": 5993
+ },
+ {
+ "epoch": 83.25152838427948,
+ "grad_norm": 0.017819996923208237,
+ "learning_rate": 0.0006,
+ "loss": 4.196932315826416,
+ "step": 5994
+ },
+ {
+ "epoch": 83.26550218340611,
+ "grad_norm": 0.01557681243866682,
+ "learning_rate": 0.0006,
+ "loss": 4.122654914855957,
+ "step": 5995
+ },
+ {
+ "epoch": 83.27947598253274,
+ "grad_norm": 0.017858853563666344,
+ "learning_rate": 0.0006,
+ "loss": 4.185330867767334,
+ "step": 5996
+ },
+ {
+ "epoch": 83.29344978165939,
+ "grad_norm": 0.018065953627228737,
+ "learning_rate": 0.0006,
+ "loss": 4.218471050262451,
+ "step": 5997
+ },
+ {
+ "epoch": 83.30742358078602,
+ "grad_norm": 0.017247939482331276,
+ "learning_rate": 0.0006,
+ "loss": 4.127076148986816,
+ "step": 5998
+ },
+ {
+ "epoch": 83.32139737991267,
+ "grad_norm": 0.017047952860593796,
+ "learning_rate": 0.0006,
+ "loss": 4.131356716156006,
+ "step": 5999
+ },
+ {
+ "epoch": 83.3353711790393,
+ "grad_norm": 0.015905629843473434,
+ "learning_rate": 0.0006,
+ "loss": 4.2239789962768555,
+ "step": 6000
+ },
+ {
+ "epoch": 83.34934497816595,
+ "grad_norm": 0.016522347927093506,
+ "learning_rate": 0.0006,
+ "loss": 4.054897785186768,
+ "step": 6001
+ },
+ {
+ "epoch": 83.36331877729258,
+ "grad_norm": 0.01622571051120758,
+ "learning_rate": 0.0006,
+ "loss": 4.284210681915283,
+ "step": 6002
+ },
+ {
+ "epoch": 83.37729257641921,
+ "grad_norm": 0.01640467159450054,
+ "learning_rate": 0.0006,
+ "loss": 4.268840789794922,
+ "step": 6003
+ },
+ {
+ "epoch": 83.39126637554585,
+ "grad_norm": 0.015698859468102455,
+ "learning_rate": 0.0006,
+ "loss": 4.179793357849121,
+ "step": 6004
+ },
+ {
+ "epoch": 83.40524017467249,
+ "grad_norm": 0.016694528982043266,
+ "learning_rate": 0.0006,
+ "loss": 4.132297515869141,
+ "step": 6005
+ },
+ {
+ "epoch": 83.41921397379913,
+ "grad_norm": 0.015345900319516659,
+ "learning_rate": 0.0006,
+ "loss": 4.090163230895996,
+ "step": 6006
+ },
+ {
+ "epoch": 83.43318777292576,
+ "grad_norm": 0.014722758904099464,
+ "learning_rate": 0.0006,
+ "loss": 4.179166793823242,
+ "step": 6007
+ },
+ {
+ "epoch": 83.44716157205241,
+ "grad_norm": 0.014884887263178825,
+ "learning_rate": 0.0006,
+ "loss": 4.137659072875977,
+ "step": 6008
+ },
+ {
+ "epoch": 83.46113537117904,
+ "grad_norm": 0.014041902497410774,
+ "learning_rate": 0.0006,
+ "loss": 4.1202073097229,
+ "step": 6009
+ },
+ {
+ "epoch": 83.47510917030567,
+ "grad_norm": 0.012799333781003952,
+ "learning_rate": 0.0006,
+ "loss": 4.092607498168945,
+ "step": 6010
+ },
+ {
+ "epoch": 83.48908296943232,
+ "grad_norm": 0.013272707350552082,
+ "learning_rate": 0.0006,
+ "loss": 4.132715225219727,
+ "step": 6011
+ },
+ {
+ "epoch": 83.50305676855895,
+ "grad_norm": 0.012934371829032898,
+ "learning_rate": 0.0006,
+ "loss": 4.230278968811035,
+ "step": 6012
+ },
+ {
+ "epoch": 83.5170305676856,
+ "grad_norm": 0.0120006762444973,
+ "learning_rate": 0.0006,
+ "loss": 4.140068531036377,
+ "step": 6013
+ },
+ {
+ "epoch": 83.53100436681223,
+ "grad_norm": 0.013616700656712055,
+ "learning_rate": 0.0006,
+ "loss": 4.20107364654541,
+ "step": 6014
+ },
+ {
+ "epoch": 83.54497816593886,
+ "grad_norm": 0.013957851566374302,
+ "learning_rate": 0.0006,
+ "loss": 4.124945163726807,
+ "step": 6015
+ },
+ {
+ "epoch": 83.5589519650655,
+ "grad_norm": 0.012762055732309818,
+ "learning_rate": 0.0006,
+ "loss": 4.022052764892578,
+ "step": 6016
+ },
+ {
+ "epoch": 83.57292576419214,
+ "grad_norm": 0.013080576434731483,
+ "learning_rate": 0.0006,
+ "loss": 4.103252410888672,
+ "step": 6017
+ },
+ {
+ "epoch": 83.58689956331878,
+ "grad_norm": 0.013957500457763672,
+ "learning_rate": 0.0006,
+ "loss": 4.235089302062988,
+ "step": 6018
+ },
+ {
+ "epoch": 83.60087336244541,
+ "grad_norm": 0.014707164838910103,
+ "learning_rate": 0.0006,
+ "loss": 4.191570281982422,
+ "step": 6019
+ },
+ {
+ "epoch": 83.61484716157206,
+ "grad_norm": 0.012807869352400303,
+ "learning_rate": 0.0006,
+ "loss": 4.294543266296387,
+ "step": 6020
+ },
+ {
+ "epoch": 83.62882096069869,
+ "grad_norm": 0.012027994729578495,
+ "learning_rate": 0.0006,
+ "loss": 4.13849401473999,
+ "step": 6021
+ },
+ {
+ "epoch": 83.64279475982532,
+ "grad_norm": 0.012147724628448486,
+ "learning_rate": 0.0006,
+ "loss": 4.127481937408447,
+ "step": 6022
+ },
+ {
+ "epoch": 83.65676855895197,
+ "grad_norm": 0.013241392560303211,
+ "learning_rate": 0.0006,
+ "loss": 4.218972682952881,
+ "step": 6023
+ },
+ {
+ "epoch": 83.6707423580786,
+ "grad_norm": 0.014071648940443993,
+ "learning_rate": 0.0006,
+ "loss": 4.096724033355713,
+ "step": 6024
+ },
+ {
+ "epoch": 83.68471615720524,
+ "grad_norm": 0.015212797559797764,
+ "learning_rate": 0.0006,
+ "loss": 4.27097225189209,
+ "step": 6025
+ },
+ {
+ "epoch": 83.69868995633188,
+ "grad_norm": 0.017647385597229004,
+ "learning_rate": 0.0006,
+ "loss": 4.209835529327393,
+ "step": 6026
+ },
+ {
+ "epoch": 83.71266375545852,
+ "grad_norm": 0.01709647662937641,
+ "learning_rate": 0.0006,
+ "loss": 4.042106628417969,
+ "step": 6027
+ },
+ {
+ "epoch": 83.72663755458515,
+ "grad_norm": 0.014530927874147892,
+ "learning_rate": 0.0006,
+ "loss": 4.181731700897217,
+ "step": 6028
+ },
+ {
+ "epoch": 83.74061135371178,
+ "grad_norm": 0.015356711111962795,
+ "learning_rate": 0.0006,
+ "loss": 4.314183235168457,
+ "step": 6029
+ },
+ {
+ "epoch": 83.75458515283843,
+ "grad_norm": 0.015500029549002647,
+ "learning_rate": 0.0006,
+ "loss": 4.06776762008667,
+ "step": 6030
+ },
+ {
+ "epoch": 83.76855895196506,
+ "grad_norm": 0.015351332724094391,
+ "learning_rate": 0.0006,
+ "loss": 4.185141563415527,
+ "step": 6031
+ },
+ {
+ "epoch": 83.78253275109171,
+ "grad_norm": 0.016678154468536377,
+ "learning_rate": 0.0006,
+ "loss": 4.179529666900635,
+ "step": 6032
+ },
+ {
+ "epoch": 83.79650655021834,
+ "grad_norm": 0.0156265702098608,
+ "learning_rate": 0.0006,
+ "loss": 4.238584041595459,
+ "step": 6033
+ },
+ {
+ "epoch": 83.81048034934499,
+ "grad_norm": 0.016418596729636192,
+ "learning_rate": 0.0006,
+ "loss": 4.1593523025512695,
+ "step": 6034
+ },
+ {
+ "epoch": 83.82445414847162,
+ "grad_norm": 0.016113586723804474,
+ "learning_rate": 0.0006,
+ "loss": 4.157470226287842,
+ "step": 6035
+ },
+ {
+ "epoch": 83.83842794759825,
+ "grad_norm": 0.014434042386710644,
+ "learning_rate": 0.0006,
+ "loss": 4.196226119995117,
+ "step": 6036
+ },
+ {
+ "epoch": 83.8524017467249,
+ "grad_norm": 0.012647042982280254,
+ "learning_rate": 0.0006,
+ "loss": 4.130589485168457,
+ "step": 6037
+ },
+ {
+ "epoch": 83.86637554585153,
+ "grad_norm": 0.014412877149879932,
+ "learning_rate": 0.0006,
+ "loss": 4.154845237731934,
+ "step": 6038
+ },
+ {
+ "epoch": 83.88034934497817,
+ "grad_norm": 0.014222027733922005,
+ "learning_rate": 0.0006,
+ "loss": 4.189724922180176,
+ "step": 6039
+ },
+ {
+ "epoch": 83.8943231441048,
+ "grad_norm": 0.013084255158901215,
+ "learning_rate": 0.0006,
+ "loss": 4.303338527679443,
+ "step": 6040
+ },
+ {
+ "epoch": 83.90829694323143,
+ "grad_norm": 0.012659232132136822,
+ "learning_rate": 0.0006,
+ "loss": 4.08192253112793,
+ "step": 6041
+ },
+ {
+ "epoch": 83.92227074235808,
+ "grad_norm": 0.01246276218444109,
+ "learning_rate": 0.0006,
+ "loss": 4.283324718475342,
+ "step": 6042
+ },
+ {
+ "epoch": 83.93624454148471,
+ "grad_norm": 0.01340118795633316,
+ "learning_rate": 0.0006,
+ "loss": 4.117908477783203,
+ "step": 6043
+ },
+ {
+ "epoch": 83.95021834061136,
+ "grad_norm": 0.014496722258627415,
+ "learning_rate": 0.0006,
+ "loss": 4.110318660736084,
+ "step": 6044
+ },
+ {
+ "epoch": 83.96419213973799,
+ "grad_norm": 0.013348298147320747,
+ "learning_rate": 0.0006,
+ "loss": 4.185139179229736,
+ "step": 6045
+ },
+ {
+ "epoch": 83.97816593886463,
+ "grad_norm": 0.013294277712702751,
+ "learning_rate": 0.0006,
+ "loss": 4.271645545959473,
+ "step": 6046
+ },
+ {
+ "epoch": 83.99213973799127,
+ "grad_norm": 0.014431594870984554,
+ "learning_rate": 0.0006,
+ "loss": 4.161383152008057,
+ "step": 6047
+ },
+ {
+ "epoch": 84.0,
+ "grad_norm": 0.016268927603960037,
+ "learning_rate": 0.0006,
+ "loss": 4.178324222564697,
+ "step": 6048
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 28800,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 400,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": true,
+ "should_log": false,
+ "should_save": false,
+ "should_training_stop": false
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 2.578475223519068e+18,
+ "train_batch_size": 8,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6048/training_args.bin b/runs/l2r50-i2-fulle-lm/checkpoint-6048/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..cad232831c222cc7f8b938b8bbe42f61e7d89c3a
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-6048/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:f7c40e3bfa2f5ac1450ca1243e3f2a895d44a06ed5995e6aefd97f78d229d8c8
+size 4792
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6336/chat_template.jinja b/runs/l2r50-i2-fulle-lm/checkpoint-6336/chat_template.jinja
new file mode 100644
index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-6336/chat_template.jinja
@@ -0,0 +1,85 @@
+{%- if tools %}
+ {{- '<|im_start|>system\n' }}
+ {%- if messages[0].role == 'system' %}
+ {{- messages[0].content + '\n\n' }}
+ {%- endif %}
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }}
+ {%- for tool in tools %}
+ {{- "\n" }}
+ {{- tool | tojson }}
+ {%- endfor %}
+ {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }}
+{%- else %}
+ {%- if messages[0].role == 'system' %}
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
+ {%- endif %}
+{%- endif %}
+{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
+{%- for message in messages[::-1] %}
+ {%- set index = (messages|length - 1) - loop.index0 %}
+ {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %}
+ {%- set ns.multi_step_tool = false %}
+ {%- set ns.last_query_index = index %}
+ {%- endif %}
+{%- endfor %}
+{%- for message in messages %}
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
+ {%- elif message.role == "assistant" %}
+ {%- set content = message.content %}
+ {%- set reasoning_content = '' %}
+ {%- if message.reasoning_content is defined and message.reasoning_content is not none %}
+ {%- set reasoning_content = message.reasoning_content %}
+ {%- else %}
+ {%- if '' in message.content %}
+ {%- set content = message.content.split('')[-1].lstrip('\n') %}
+ {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %}
+ {%- endif %}
+ {%- endif %}
+ {%- if loop.index0 > ns.last_query_index %}
+ {%- if loop.last or (not loop.last and reasoning_content) %}
+ {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- if message.tool_calls %}
+ {%- for tool_call in message.tool_calls %}
+ {%- if (loop.first and content) or (not loop.first) %}
+ {{- '\n' }}
+ {%- endif %}
+ {%- if tool_call.function %}
+ {%- set tool_call = tool_call.function %}
+ {%- endif %}
+ {{- '\n{"name": "' }}
+ {{- tool_call.name }}
+ {{- '", "arguments": ' }}
+ {%- if tool_call.arguments is string %}
+ {{- tool_call.arguments }}
+ {%- else %}
+ {{- tool_call.arguments | tojson }}
+ {%- endif %}
+ {{- '}\n' }}
+ {%- endfor %}
+ {%- endif %}
+ {{- '<|im_end|>\n' }}
+ {%- elif message.role == "tool" %}
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
+ {{- '<|im_start|>user' }}
+ {%- endif %}
+ {{- '\n\n' }}
+ {{- message.content }}
+ {{- '\n' }}
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
+ {{- '<|im_end|>\n' }}
+ {%- endif %}
+ {%- endif %}
+{%- endfor %}
+{%- if add_generation_prompt %}
+ {{- '<|im_start|>assistant\n' }}
+ {%- if enable_thinking is defined and enable_thinking is false %}
+ {{- '\n\n\n\n' }}
+ {%- endif %}
+{%- endif %}
\ No newline at end of file
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6336/config.json b/runs/l2r50-i2-fulle-lm/checkpoint-6336/config.json
new file mode 100644
index 0000000000000000000000000000000000000000..22f250953c6ce8fc91ee4411856dd73041a415e2
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-6336/config.json
@@ -0,0 +1,32 @@
+{
+ "architectures": [
+ "LlamaForCausalLM"
+ ],
+ "attention_bias": false,
+ "attention_dropout": 0.0,
+ "bos_token_id": null,
+ "dtype": "float32",
+ "eos_token_id": 151645,
+ "head_dim": 128,
+ "hidden_act": "silu",
+ "hidden_size": 512,
+ "initializer_range": 0.02,
+ "intermediate_size": 1536,
+ "max_position_embeddings": 2048,
+ "mlp_bias": false,
+ "model_type": "llama",
+ "num_attention_heads": 4,
+ "num_hidden_layers": 20,
+ "num_key_value_heads": 4,
+ "pad_token_id": 151645,
+ "pretraining_tp": 1,
+ "rms_norm_eps": 1e-06,
+ "rope_parameters": {
+ "rope_theta": 10000.0,
+ "rope_type": "default"
+ },
+ "tie_word_embeddings": true,
+ "transformers_version": "5.5.0",
+ "use_cache": false,
+ "vocab_size": 151673
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6336/generation_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-6336/generation_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-6336/generation_config.json
@@ -0,0 +1,11 @@
+{
+ "_from_model_config": true,
+ "eos_token_id": [
+ 151645
+ ],
+ "output_attentions": false,
+ "output_hidden_states": false,
+ "pad_token_id": 151645,
+ "transformers_version": "5.5.0",
+ "use_cache": true
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6336/model.safetensors b/runs/l2r50-i2-fulle-lm/checkpoint-6336/model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..a2ea71b3237632ffc28130e245c13a34ee03a134
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-6336/model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:6bcfdc4bd44a4657ef33e702c02341b87b4a18ccaac135c4d8c6ff5ff8e3e5e3
+size 583360328
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6336/optimizer.pt b/runs/l2r50-i2-fulle-lm/checkpoint-6336/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..15000b58fd2fa4486c6968e8a39b11756d6a5e15
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-6336/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:3ed1db5379aaee2d0aeeaea8df19fe785ad63c518130165e8d0d5613ea85bf89
+size 1166833530
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6336/rng_state_0.pth b/runs/l2r50-i2-fulle-lm/checkpoint-6336/rng_state_0.pth
new file mode 100644
index 0000000000000000000000000000000000000000..9bfd8fd41317ce97ec5426ef00a336bb43093975
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-6336/rng_state_0.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:5669c00dd9e0c288cf3502acb77b3f67f83fa66af28dd688459b2b61b04597d5
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6336/rng_state_1.pth b/runs/l2r50-i2-fulle-lm/checkpoint-6336/rng_state_1.pth
new file mode 100644
index 0000000000000000000000000000000000000000..166c8af41498ffc28a7641d3b21dd7a6b15c17b7
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-6336/rng_state_1.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:7e1e292f0fe3c7fb524391e6995752fb45102689b83575bb8bf380e55677c963
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6336/rng_state_2.pth b/runs/l2r50-i2-fulle-lm/checkpoint-6336/rng_state_2.pth
new file mode 100644
index 0000000000000000000000000000000000000000..3b04d3b8510e2cb426c14bf493785c01f9b3953a
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-6336/rng_state_2.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:62c7b95ac7104656fd3a5a8a50a70e41f051ca80db9ee60b624c946b282b9466
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6336/rng_state_3.pth b/runs/l2r50-i2-fulle-lm/checkpoint-6336/rng_state_3.pth
new file mode 100644
index 0000000000000000000000000000000000000000..b7ea885e549895088ffc7c40f579739debda147d
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-6336/rng_state_3.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:64fde3e7147a7a919872654aaa3b6ff513c5a49eeb3b65f8fdeea1def70174ae
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6336/scheduler.pt b/runs/l2r50-i2-fulle-lm/checkpoint-6336/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..806ff31f0128f007291ee83361dda9fba5eeb32b
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-6336/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:79947657782b5ee38d1d49b023c1983388f9e08fc318792128b0f35b4e555422
+size 1064
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6336/tokenizer.json b/runs/l2r50-i2-fulle-lm/checkpoint-6336/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..53786016f070e11184813f3c83101dc5e93e392e
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-6336/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:c3c63fda1646a555448e8f03662542e04b11d103721319e4d9b267d7da03ceb8
+size 11424463
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6336/tokenizer_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-6336/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..98e2fb9b2cc54785dc9d871b8e503f8b8f2defe7
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-6336/tokenizer_config.json
@@ -0,0 +1,20 @@
+{
+ "add_prefix_space": false,
+ "backend": "tokenizers",
+ "bos_token": null,
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "<|im_end|>",
+ "errors": "replace",
+ "extra_special_tokens": [
+ "<|l2r_pred|>",
+ "<|r2l_pred|>",
+ "<|next_1_pred|>",
+ "<|next_2_pred|>"
+ ],
+ "is_local": false,
+ "model_max_length": 131072,
+ "pad_token": "<|im_end|>",
+ "split_special_tokens": false,
+ "tokenizer_class": "Qwen2Tokenizer",
+ "unk_token": null
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6336/trainer_state.json b/runs/l2r50-i2-fulle-lm/checkpoint-6336/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..7cceab1151587e723d3d114ba68783317d29dbd1
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-6336/trainer_state.json
@@ -0,0 +1,45082 @@
+{
+ "best_global_step": null,
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 88.0,
+ "eval_steps": 500,
+ "global_step": 6336,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.013973799126637555,
+ "grad_norm": 0.1739506870508194,
+ "learning_rate": 0.0,
+ "loss": 11.99299430847168,
+ "step": 1
+ },
+ {
+ "epoch": 0.02794759825327511,
+ "grad_norm": 0.17216597497463226,
+ "learning_rate": 5.999999999999999e-06,
+ "loss": 11.988704681396484,
+ "step": 2
+ },
+ {
+ "epoch": 0.04192139737991266,
+ "grad_norm": 0.17459715902805328,
+ "learning_rate": 1.1999999999999999e-05,
+ "loss": 11.945581436157227,
+ "step": 3
+ },
+ {
+ "epoch": 0.05589519650655022,
+ "grad_norm": 0.17577414214611053,
+ "learning_rate": 1.7999999999999997e-05,
+ "loss": 11.859772682189941,
+ "step": 4
+ },
+ {
+ "epoch": 0.06986899563318777,
+ "grad_norm": 0.16163651645183563,
+ "learning_rate": 2.3999999999999997e-05,
+ "loss": 11.744415283203125,
+ "step": 5
+ },
+ {
+ "epoch": 0.08384279475982533,
+ "grad_norm": 0.14409932494163513,
+ "learning_rate": 2.9999999999999997e-05,
+ "loss": 11.618099212646484,
+ "step": 6
+ },
+ {
+ "epoch": 0.09781659388646288,
+ "grad_norm": 0.13079848885536194,
+ "learning_rate": 3.5999999999999994e-05,
+ "loss": 11.505935668945312,
+ "step": 7
+ },
+ {
+ "epoch": 0.11179039301310044,
+ "grad_norm": 0.11964613944292068,
+ "learning_rate": 4.2e-05,
+ "loss": 11.416532516479492,
+ "step": 8
+ },
+ {
+ "epoch": 0.125764192139738,
+ "grad_norm": 0.11345142871141434,
+ "learning_rate": 4.7999999999999994e-05,
+ "loss": 11.33206558227539,
+ "step": 9
+ },
+ {
+ "epoch": 0.13973799126637554,
+ "grad_norm": 0.11033952981233597,
+ "learning_rate": 5.399999999999999e-05,
+ "loss": 11.259052276611328,
+ "step": 10
+ },
+ {
+ "epoch": 0.1537117903930131,
+ "grad_norm": 0.10776489973068237,
+ "learning_rate": 5.9999999999999995e-05,
+ "loss": 11.203384399414062,
+ "step": 11
+ },
+ {
+ "epoch": 0.16768558951965065,
+ "grad_norm": 0.10721946507692337,
+ "learning_rate": 6.599999999999999e-05,
+ "loss": 11.153295516967773,
+ "step": 12
+ },
+ {
+ "epoch": 0.18165938864628822,
+ "grad_norm": 0.10673430562019348,
+ "learning_rate": 7.199999999999999e-05,
+ "loss": 11.110200881958008,
+ "step": 13
+ },
+ {
+ "epoch": 0.19563318777292577,
+ "grad_norm": 0.10630254447460175,
+ "learning_rate": 7.8e-05,
+ "loss": 11.067477226257324,
+ "step": 14
+ },
+ {
+ "epoch": 0.2096069868995633,
+ "grad_norm": 0.10631615668535233,
+ "learning_rate": 8.4e-05,
+ "loss": 11.022941589355469,
+ "step": 15
+ },
+ {
+ "epoch": 0.22358078602620088,
+ "grad_norm": 0.10609443485736847,
+ "learning_rate": 8.999999999999999e-05,
+ "loss": 10.970746040344238,
+ "step": 16
+ },
+ {
+ "epoch": 0.23755458515283842,
+ "grad_norm": 0.10648232698440552,
+ "learning_rate": 9.599999999999999e-05,
+ "loss": 10.91267204284668,
+ "step": 17
+ },
+ {
+ "epoch": 0.251528384279476,
+ "grad_norm": 0.10584589838981628,
+ "learning_rate": 0.000102,
+ "loss": 10.854547500610352,
+ "step": 18
+ },
+ {
+ "epoch": 0.26550218340611353,
+ "grad_norm": 0.10591720044612885,
+ "learning_rate": 0.00010799999999999998,
+ "loss": 10.788917541503906,
+ "step": 19
+ },
+ {
+ "epoch": 0.2794759825327511,
+ "grad_norm": 0.10611672699451447,
+ "learning_rate": 0.00011399999999999999,
+ "loss": 10.717556953430176,
+ "step": 20
+ },
+ {
+ "epoch": 0.2934497816593886,
+ "grad_norm": 0.10455359518527985,
+ "learning_rate": 0.00011999999999999999,
+ "loss": 10.650186538696289,
+ "step": 21
+ },
+ {
+ "epoch": 0.3074235807860262,
+ "grad_norm": 0.1047525629401207,
+ "learning_rate": 0.00012599999999999997,
+ "loss": 10.56808090209961,
+ "step": 22
+ },
+ {
+ "epoch": 0.32139737991266376,
+ "grad_norm": 0.10511605441570282,
+ "learning_rate": 0.00013199999999999998,
+ "loss": 10.484028816223145,
+ "step": 23
+ },
+ {
+ "epoch": 0.3353711790393013,
+ "grad_norm": 0.10467177629470825,
+ "learning_rate": 0.000138,
+ "loss": 10.399947166442871,
+ "step": 24
+ },
+ {
+ "epoch": 0.34934497816593885,
+ "grad_norm": 0.10432136058807373,
+ "learning_rate": 0.00014399999999999998,
+ "loss": 10.312714576721191,
+ "step": 25
+ },
+ {
+ "epoch": 0.36331877729257644,
+ "grad_norm": 0.10527262091636658,
+ "learning_rate": 0.00015,
+ "loss": 10.212261199951172,
+ "step": 26
+ },
+ {
+ "epoch": 0.377292576419214,
+ "grad_norm": 0.103610560297966,
+ "learning_rate": 0.000156,
+ "loss": 10.132034301757812,
+ "step": 27
+ },
+ {
+ "epoch": 0.39126637554585153,
+ "grad_norm": 0.10553059726953506,
+ "learning_rate": 0.000162,
+ "loss": 10.020434379577637,
+ "step": 28
+ },
+ {
+ "epoch": 0.4052401746724891,
+ "grad_norm": 0.1037299782037735,
+ "learning_rate": 0.000168,
+ "loss": 9.934466361999512,
+ "step": 29
+ },
+ {
+ "epoch": 0.4192139737991266,
+ "grad_norm": 0.1042739674448967,
+ "learning_rate": 0.00017399999999999997,
+ "loss": 9.822539329528809,
+ "step": 30
+ },
+ {
+ "epoch": 0.4331877729257642,
+ "grad_norm": 0.10300865024328232,
+ "learning_rate": 0.00017999999999999998,
+ "loss": 9.734994888305664,
+ "step": 31
+ },
+ {
+ "epoch": 0.44716157205240176,
+ "grad_norm": 0.10291654616594315,
+ "learning_rate": 0.000186,
+ "loss": 9.624505996704102,
+ "step": 32
+ },
+ {
+ "epoch": 0.4611353711790393,
+ "grad_norm": 0.10210851579904556,
+ "learning_rate": 0.00019199999999999998,
+ "loss": 9.518205642700195,
+ "step": 33
+ },
+ {
+ "epoch": 0.47510917030567684,
+ "grad_norm": 0.10222529619932175,
+ "learning_rate": 0.000198,
+ "loss": 9.412222862243652,
+ "step": 34
+ },
+ {
+ "epoch": 0.4890829694323144,
+ "grad_norm": 0.09992019087076187,
+ "learning_rate": 0.000204,
+ "loss": 9.326566696166992,
+ "step": 35
+ },
+ {
+ "epoch": 0.503056768558952,
+ "grad_norm": 0.09931084513664246,
+ "learning_rate": 0.00020999999999999998,
+ "loss": 9.210111618041992,
+ "step": 36
+ },
+ {
+ "epoch": 0.5170305676855895,
+ "grad_norm": 0.09928256273269653,
+ "learning_rate": 0.00021599999999999996,
+ "loss": 9.100493431091309,
+ "step": 37
+ },
+ {
+ "epoch": 0.5310043668122271,
+ "grad_norm": 0.09843816608190536,
+ "learning_rate": 0.00022199999999999998,
+ "loss": 8.996543884277344,
+ "step": 38
+ },
+ {
+ "epoch": 0.5449781659388646,
+ "grad_norm": 0.09825374186038971,
+ "learning_rate": 0.00022799999999999999,
+ "loss": 8.888991355895996,
+ "step": 39
+ },
+ {
+ "epoch": 0.5589519650655022,
+ "grad_norm": 0.09602385759353638,
+ "learning_rate": 0.000234,
+ "loss": 8.802907943725586,
+ "step": 40
+ },
+ {
+ "epoch": 0.5729257641921397,
+ "grad_norm": 0.09414064139127731,
+ "learning_rate": 0.00023999999999999998,
+ "loss": 8.715423583984375,
+ "step": 41
+ },
+ {
+ "epoch": 0.5868995633187772,
+ "grad_norm": 0.09273893386125565,
+ "learning_rate": 0.00024599999999999996,
+ "loss": 8.61374282836914,
+ "step": 42
+ },
+ {
+ "epoch": 0.6008733624454149,
+ "grad_norm": 0.09079599380493164,
+ "learning_rate": 0.00025199999999999995,
+ "loss": 8.502904891967773,
+ "step": 43
+ },
+ {
+ "epoch": 0.6148471615720524,
+ "grad_norm": 0.08605825901031494,
+ "learning_rate": 0.000258,
+ "loss": 8.453008651733398,
+ "step": 44
+ },
+ {
+ "epoch": 0.62882096069869,
+ "grad_norm": 0.08338996022939682,
+ "learning_rate": 0.00026399999999999997,
+ "loss": 8.365163803100586,
+ "step": 45
+ },
+ {
+ "epoch": 0.6427947598253275,
+ "grad_norm": 0.08150867372751236,
+ "learning_rate": 0.00027,
+ "loss": 8.285372734069824,
+ "step": 46
+ },
+ {
+ "epoch": 0.6567685589519651,
+ "grad_norm": 0.07794807851314545,
+ "learning_rate": 0.000276,
+ "loss": 8.185811996459961,
+ "step": 47
+ },
+ {
+ "epoch": 0.6707423580786026,
+ "grad_norm": 0.07592720538377762,
+ "learning_rate": 0.00028199999999999997,
+ "loss": 8.098482131958008,
+ "step": 48
+ },
+ {
+ "epoch": 0.6847161572052402,
+ "grad_norm": 0.0709897130727768,
+ "learning_rate": 0.00028799999999999995,
+ "loss": 8.042091369628906,
+ "step": 49
+ },
+ {
+ "epoch": 0.6986899563318777,
+ "grad_norm": 0.06582305580377579,
+ "learning_rate": 0.000294,
+ "loss": 7.972844123840332,
+ "step": 50
+ },
+ {
+ "epoch": 0.7126637554585152,
+ "grad_norm": 0.059452351182699203,
+ "learning_rate": 0.0003,
+ "loss": 7.924579620361328,
+ "step": 51
+ },
+ {
+ "epoch": 0.7266375545851529,
+ "grad_norm": 0.052426718175411224,
+ "learning_rate": 0.00030599999999999996,
+ "loss": 7.876336097717285,
+ "step": 52
+ },
+ {
+ "epoch": 0.7406113537117904,
+ "grad_norm": 0.04768150672316551,
+ "learning_rate": 0.000312,
+ "loss": 7.8102707862854,
+ "step": 53
+ },
+ {
+ "epoch": 0.754585152838428,
+ "grad_norm": 0.03970075398683548,
+ "learning_rate": 0.000318,
+ "loss": 7.791913986206055,
+ "step": 54
+ },
+ {
+ "epoch": 0.7685589519650655,
+ "grad_norm": 0.03559752553701401,
+ "learning_rate": 0.000324,
+ "loss": 7.756960868835449,
+ "step": 55
+ },
+ {
+ "epoch": 0.7825327510917031,
+ "grad_norm": 0.029321353882551193,
+ "learning_rate": 0.00033,
+ "loss": 7.724878787994385,
+ "step": 56
+ },
+ {
+ "epoch": 0.7965065502183406,
+ "grad_norm": 0.023485427722334862,
+ "learning_rate": 0.000336,
+ "loss": 7.72553014755249,
+ "step": 57
+ },
+ {
+ "epoch": 0.8104803493449781,
+ "grad_norm": 0.019319474697113037,
+ "learning_rate": 0.00034199999999999996,
+ "loss": 7.709493160247803,
+ "step": 58
+ },
+ {
+ "epoch": 0.8244541484716157,
+ "grad_norm": 0.016705691814422607,
+ "learning_rate": 0.00034799999999999995,
+ "loss": 7.6873297691345215,
+ "step": 59
+ },
+ {
+ "epoch": 0.8384279475982532,
+ "grad_norm": 0.017667734995484352,
+ "learning_rate": 0.00035399999999999993,
+ "loss": 7.677059173583984,
+ "step": 60
+ },
+ {
+ "epoch": 0.8524017467248908,
+ "grad_norm": 0.01671992614865303,
+ "learning_rate": 0.00035999999999999997,
+ "loss": 7.643512725830078,
+ "step": 61
+ },
+ {
+ "epoch": 0.8663755458515284,
+ "grad_norm": 0.012288172729313374,
+ "learning_rate": 0.00036599999999999995,
+ "loss": 7.658343315124512,
+ "step": 62
+ },
+ {
+ "epoch": 0.880349344978166,
+ "grad_norm": 0.012004776857793331,
+ "learning_rate": 0.000372,
+ "loss": 7.649759292602539,
+ "step": 63
+ },
+ {
+ "epoch": 0.8943231441048035,
+ "grad_norm": 0.012487445957958698,
+ "learning_rate": 0.00037799999999999997,
+ "loss": 7.631503105163574,
+ "step": 64
+ },
+ {
+ "epoch": 0.9082969432314411,
+ "grad_norm": 0.01489241048693657,
+ "learning_rate": 0.00038399999999999996,
+ "loss": 7.647110462188721,
+ "step": 65
+ },
+ {
+ "epoch": 0.9222707423580786,
+ "grad_norm": 0.01428980939090252,
+ "learning_rate": 0.00039,
+ "loss": 7.636983871459961,
+ "step": 66
+ },
+ {
+ "epoch": 0.9362445414847161,
+ "grad_norm": 0.010711144655942917,
+ "learning_rate": 0.000396,
+ "loss": 7.621130466461182,
+ "step": 67
+ },
+ {
+ "epoch": 0.9502183406113537,
+ "grad_norm": 0.010178886353969574,
+ "learning_rate": 0.000402,
+ "loss": 7.593924522399902,
+ "step": 68
+ },
+ {
+ "epoch": 0.9641921397379912,
+ "grad_norm": 0.010604874230921268,
+ "learning_rate": 0.000408,
+ "loss": 7.6015496253967285,
+ "step": 69
+ },
+ {
+ "epoch": 0.9781659388646288,
+ "grad_norm": 0.009751426056027412,
+ "learning_rate": 0.0004139999999999999,
+ "loss": 7.571964263916016,
+ "step": 70
+ },
+ {
+ "epoch": 0.9921397379912664,
+ "grad_norm": 0.009816481731832027,
+ "learning_rate": 0.00041999999999999996,
+ "loss": 7.57344913482666,
+ "step": 71
+ },
+ {
+ "epoch": 1.0,
+ "grad_norm": 0.008665637113153934,
+ "learning_rate": 0.00042599999999999995,
+ "loss": 7.579050064086914,
+ "step": 72
+ },
+ {
+ "epoch": 1.0,
+ "eval_loss": 7.581204891204834,
+ "eval_runtime": 59.0091,
+ "eval_samples_per_second": 41.383,
+ "eval_steps_per_second": 1.305,
+ "step": 72
+ },
+ {
+ "epoch": 1.0139737991266375,
+ "grad_norm": 0.009684201329946518,
+ "learning_rate": 0.00043199999999999993,
+ "loss": 7.568594932556152,
+ "step": 73
+ },
+ {
+ "epoch": 1.027947598253275,
+ "grad_norm": 0.025783710181713104,
+ "learning_rate": 0.00043799999999999997,
+ "loss": 7.578948497772217,
+ "step": 74
+ },
+ {
+ "epoch": 1.0419213973799126,
+ "grad_norm": 0.0327819399535656,
+ "learning_rate": 0.00044399999999999995,
+ "loss": 7.548173904418945,
+ "step": 75
+ },
+ {
+ "epoch": 1.0558951965065502,
+ "grad_norm": 0.021604498848319054,
+ "learning_rate": 0.00045,
+ "loss": 7.552953720092773,
+ "step": 76
+ },
+ {
+ "epoch": 1.0698689956331877,
+ "grad_norm": 0.010230105370283127,
+ "learning_rate": 0.00045599999999999997,
+ "loss": 7.5213212966918945,
+ "step": 77
+ },
+ {
+ "epoch": 1.0838427947598253,
+ "grad_norm": 0.023315589874982834,
+ "learning_rate": 0.00046199999999999995,
+ "loss": 7.501640319824219,
+ "step": 78
+ },
+ {
+ "epoch": 1.0978165938864628,
+ "grad_norm": 0.01109944935888052,
+ "learning_rate": 0.000468,
+ "loss": 7.539519786834717,
+ "step": 79
+ },
+ {
+ "epoch": 1.1117903930131003,
+ "grad_norm": 0.026093758642673492,
+ "learning_rate": 0.000474,
+ "loss": 7.525928974151611,
+ "step": 80
+ },
+ {
+ "epoch": 1.125764192139738,
+ "grad_norm": 0.03274895250797272,
+ "learning_rate": 0.00047999999999999996,
+ "loss": 7.535602569580078,
+ "step": 81
+ },
+ {
+ "epoch": 1.1397379912663754,
+ "grad_norm": 0.01543602254241705,
+ "learning_rate": 0.000486,
+ "loss": 7.513683795928955,
+ "step": 82
+ },
+ {
+ "epoch": 1.1537117903930132,
+ "grad_norm": 0.028377249836921692,
+ "learning_rate": 0.0004919999999999999,
+ "loss": 7.49384880065918,
+ "step": 83
+ },
+ {
+ "epoch": 1.1676855895196507,
+ "grad_norm": 0.04121481627225876,
+ "learning_rate": 0.000498,
+ "loss": 7.488447189331055,
+ "step": 84
+ },
+ {
+ "epoch": 1.1816593886462883,
+ "grad_norm": 0.013282302767038345,
+ "learning_rate": 0.0005039999999999999,
+ "loss": 7.463010787963867,
+ "step": 85
+ },
+ {
+ "epoch": 1.1956331877729258,
+ "grad_norm": 0.03200364485383034,
+ "learning_rate": 0.0005099999999999999,
+ "loss": 7.473515033721924,
+ "step": 86
+ },
+ {
+ "epoch": 1.2096069868995634,
+ "grad_norm": 0.020768029615283012,
+ "learning_rate": 0.000516,
+ "loss": 7.464688777923584,
+ "step": 87
+ },
+ {
+ "epoch": 1.223580786026201,
+ "grad_norm": 0.021715788170695305,
+ "learning_rate": 0.000522,
+ "loss": 7.4835028648376465,
+ "step": 88
+ },
+ {
+ "epoch": 1.2375545851528384,
+ "grad_norm": 0.026255754753947258,
+ "learning_rate": 0.0005279999999999999,
+ "loss": 7.443392753601074,
+ "step": 89
+ },
+ {
+ "epoch": 1.251528384279476,
+ "grad_norm": 0.022325951606035233,
+ "learning_rate": 0.000534,
+ "loss": 7.457373142242432,
+ "step": 90
+ },
+ {
+ "epoch": 1.2655021834061135,
+ "grad_norm": 0.009270387701690197,
+ "learning_rate": 0.00054,
+ "loss": 7.455986976623535,
+ "step": 91
+ },
+ {
+ "epoch": 1.279475982532751,
+ "grad_norm": 0.02369544468820095,
+ "learning_rate": 0.0005459999999999999,
+ "loss": 7.442216873168945,
+ "step": 92
+ },
+ {
+ "epoch": 1.2934497816593886,
+ "grad_norm": 0.008474506437778473,
+ "learning_rate": 0.000552,
+ "loss": 7.4294915199279785,
+ "step": 93
+ },
+ {
+ "epoch": 1.3074235807860262,
+ "grad_norm": 0.017043249681591988,
+ "learning_rate": 0.000558,
+ "loss": 7.406932830810547,
+ "step": 94
+ },
+ {
+ "epoch": 1.3213973799126637,
+ "grad_norm": 0.011813412420451641,
+ "learning_rate": 0.0005639999999999999,
+ "loss": 7.415900230407715,
+ "step": 95
+ },
+ {
+ "epoch": 1.3353711790393012,
+ "grad_norm": 0.01086883433163166,
+ "learning_rate": 0.00057,
+ "loss": 7.452998638153076,
+ "step": 96
+ },
+ {
+ "epoch": 1.3493449781659388,
+ "grad_norm": 0.014700951054692268,
+ "learning_rate": 0.0005759999999999999,
+ "loss": 7.410249710083008,
+ "step": 97
+ },
+ {
+ "epoch": 1.3633187772925766,
+ "grad_norm": 0.014512976631522179,
+ "learning_rate": 0.0005819999999999999,
+ "loss": 7.429132461547852,
+ "step": 98
+ },
+ {
+ "epoch": 1.3772925764192139,
+ "grad_norm": 0.012519920244812965,
+ "learning_rate": 0.000588,
+ "loss": 7.430817604064941,
+ "step": 99
+ },
+ {
+ "epoch": 1.3912663755458516,
+ "grad_norm": 0.007132664322853088,
+ "learning_rate": 0.0005939999999999999,
+ "loss": 7.405234336853027,
+ "step": 100
+ },
+ {
+ "epoch": 1.405240174672489,
+ "grad_norm": 0.019489388912916183,
+ "learning_rate": 0.0006,
+ "loss": 7.427791595458984,
+ "step": 101
+ },
+ {
+ "epoch": 1.4192139737991267,
+ "grad_norm": 0.027252664789557457,
+ "learning_rate": 0.0006,
+ "loss": 7.41761589050293,
+ "step": 102
+ },
+ {
+ "epoch": 1.4331877729257643,
+ "grad_norm": 0.02501637488603592,
+ "learning_rate": 0.0006,
+ "loss": 7.416735649108887,
+ "step": 103
+ },
+ {
+ "epoch": 1.4471615720524018,
+ "grad_norm": 0.00989772193133831,
+ "learning_rate": 0.0006,
+ "loss": 7.3756632804870605,
+ "step": 104
+ },
+ {
+ "epoch": 1.4611353711790394,
+ "grad_norm": 0.01622949168086052,
+ "learning_rate": 0.0006,
+ "loss": 7.400811672210693,
+ "step": 105
+ },
+ {
+ "epoch": 1.475109170305677,
+ "grad_norm": 0.012407763861119747,
+ "learning_rate": 0.0006,
+ "loss": 7.37675666809082,
+ "step": 106
+ },
+ {
+ "epoch": 1.4890829694323144,
+ "grad_norm": 0.010353565216064453,
+ "learning_rate": 0.0006,
+ "loss": 7.4013142585754395,
+ "step": 107
+ },
+ {
+ "epoch": 1.503056768558952,
+ "grad_norm": 0.0120553532615304,
+ "learning_rate": 0.0006,
+ "loss": 7.379873275756836,
+ "step": 108
+ },
+ {
+ "epoch": 1.5170305676855895,
+ "grad_norm": 0.01610225811600685,
+ "learning_rate": 0.0006,
+ "loss": 7.371631145477295,
+ "step": 109
+ },
+ {
+ "epoch": 1.531004366812227,
+ "grad_norm": 0.01159879844635725,
+ "learning_rate": 0.0006,
+ "loss": 7.389007568359375,
+ "step": 110
+ },
+ {
+ "epoch": 1.5449781659388646,
+ "grad_norm": 0.008013816550374031,
+ "learning_rate": 0.0006,
+ "loss": 7.395920753479004,
+ "step": 111
+ },
+ {
+ "epoch": 1.5589519650655022,
+ "grad_norm": 0.016334809362888336,
+ "learning_rate": 0.0006,
+ "loss": 7.393121719360352,
+ "step": 112
+ },
+ {
+ "epoch": 1.5729257641921397,
+ "grad_norm": 0.018821721896529198,
+ "learning_rate": 0.0006,
+ "loss": 7.367543697357178,
+ "step": 113
+ },
+ {
+ "epoch": 1.5868995633187772,
+ "grad_norm": 0.014088994823396206,
+ "learning_rate": 0.0006,
+ "loss": 7.385537624359131,
+ "step": 114
+ },
+ {
+ "epoch": 1.600873362445415,
+ "grad_norm": 0.005864050704985857,
+ "learning_rate": 0.0006,
+ "loss": 7.365346908569336,
+ "step": 115
+ },
+ {
+ "epoch": 1.6148471615720523,
+ "grad_norm": 0.011745430529117584,
+ "learning_rate": 0.0006,
+ "loss": 7.390227317810059,
+ "step": 116
+ },
+ {
+ "epoch": 1.62882096069869,
+ "grad_norm": 0.017205573618412018,
+ "learning_rate": 0.0006,
+ "loss": 7.3558549880981445,
+ "step": 117
+ },
+ {
+ "epoch": 1.6427947598253274,
+ "grad_norm": 0.021256346255540848,
+ "learning_rate": 0.0006,
+ "loss": 7.377346992492676,
+ "step": 118
+ },
+ {
+ "epoch": 1.6567685589519652,
+ "grad_norm": 0.019335204735398293,
+ "learning_rate": 0.0006,
+ "loss": 7.346194267272949,
+ "step": 119
+ },
+ {
+ "epoch": 1.6707423580786025,
+ "grad_norm": 0.008134003728628159,
+ "learning_rate": 0.0006,
+ "loss": 7.3063578605651855,
+ "step": 120
+ },
+ {
+ "epoch": 1.6847161572052403,
+ "grad_norm": 0.017613831907510757,
+ "learning_rate": 0.0006,
+ "loss": 7.338171005249023,
+ "step": 121
+ },
+ {
+ "epoch": 1.6986899563318776,
+ "grad_norm": 0.00890734139829874,
+ "learning_rate": 0.0006,
+ "loss": 7.355384826660156,
+ "step": 122
+ },
+ {
+ "epoch": 1.7126637554585153,
+ "grad_norm": 0.013448155485093594,
+ "learning_rate": 0.0006,
+ "loss": 7.344296455383301,
+ "step": 123
+ },
+ {
+ "epoch": 1.726637554585153,
+ "grad_norm": 0.009241633117198944,
+ "learning_rate": 0.0006,
+ "loss": 7.3454060554504395,
+ "step": 124
+ },
+ {
+ "epoch": 1.7406113537117904,
+ "grad_norm": 0.018561311066150665,
+ "learning_rate": 0.0006,
+ "loss": 7.32348108291626,
+ "step": 125
+ },
+ {
+ "epoch": 1.754585152838428,
+ "grad_norm": 0.025667704641819,
+ "learning_rate": 0.0006,
+ "loss": 7.356142520904541,
+ "step": 126
+ },
+ {
+ "epoch": 1.7685589519650655,
+ "grad_norm": 0.036232151091098785,
+ "learning_rate": 0.0006,
+ "loss": 7.363539695739746,
+ "step": 127
+ },
+ {
+ "epoch": 1.782532751091703,
+ "grad_norm": 0.059483714401721954,
+ "learning_rate": 0.0006,
+ "loss": 7.402331352233887,
+ "step": 128
+ },
+ {
+ "epoch": 1.7965065502183406,
+ "grad_norm": 0.05365009233355522,
+ "learning_rate": 0.0006,
+ "loss": 7.40457820892334,
+ "step": 129
+ },
+ {
+ "epoch": 1.8104803493449781,
+ "grad_norm": 0.02738415263593197,
+ "learning_rate": 0.0006,
+ "loss": 7.3756208419799805,
+ "step": 130
+ },
+ {
+ "epoch": 1.8244541484716157,
+ "grad_norm": 0.035017091780900955,
+ "learning_rate": 0.0006,
+ "loss": 7.379263877868652,
+ "step": 131
+ },
+ {
+ "epoch": 1.8384279475982532,
+ "grad_norm": 0.02819863148033619,
+ "learning_rate": 0.0006,
+ "loss": 7.327673435211182,
+ "step": 132
+ },
+ {
+ "epoch": 1.8524017467248908,
+ "grad_norm": 0.017754238098859787,
+ "learning_rate": 0.0006,
+ "loss": 7.333420276641846,
+ "step": 133
+ },
+ {
+ "epoch": 1.8663755458515285,
+ "grad_norm": 0.024685733020305634,
+ "learning_rate": 0.0006,
+ "loss": 7.34493350982666,
+ "step": 134
+ },
+ {
+ "epoch": 1.8803493449781659,
+ "grad_norm": 0.019426532089710236,
+ "learning_rate": 0.0006,
+ "loss": 7.328448295593262,
+ "step": 135
+ },
+ {
+ "epoch": 1.8943231441048036,
+ "grad_norm": 0.02106391079723835,
+ "learning_rate": 0.0006,
+ "loss": 7.327917098999023,
+ "step": 136
+ },
+ {
+ "epoch": 1.908296943231441,
+ "grad_norm": 0.02056879736483097,
+ "learning_rate": 0.0006,
+ "loss": 7.3311991691589355,
+ "step": 137
+ },
+ {
+ "epoch": 1.9222707423580787,
+ "grad_norm": 0.016865141689777374,
+ "learning_rate": 0.0006,
+ "loss": 7.312766075134277,
+ "step": 138
+ },
+ {
+ "epoch": 1.936244541484716,
+ "grad_norm": 0.017549855634570122,
+ "learning_rate": 0.0006,
+ "loss": 7.31191349029541,
+ "step": 139
+ },
+ {
+ "epoch": 1.9502183406113538,
+ "grad_norm": 0.015930157154798508,
+ "learning_rate": 0.0006,
+ "loss": 7.330901622772217,
+ "step": 140
+ },
+ {
+ "epoch": 1.9641921397379911,
+ "grad_norm": 0.013283558189868927,
+ "learning_rate": 0.0006,
+ "loss": 7.298318386077881,
+ "step": 141
+ },
+ {
+ "epoch": 1.9781659388646289,
+ "grad_norm": 0.012023469433188438,
+ "learning_rate": 0.0006,
+ "loss": 7.282337188720703,
+ "step": 142
+ },
+ {
+ "epoch": 1.9921397379912664,
+ "grad_norm": 0.01459511648863554,
+ "learning_rate": 0.0006,
+ "loss": 7.282924175262451,
+ "step": 143
+ },
+ {
+ "epoch": 2.0,
+ "grad_norm": 0.017474235966801643,
+ "learning_rate": 0.0006,
+ "loss": 7.254333972930908,
+ "step": 144
+ },
+ {
+ "epoch": 2.0,
+ "eval_loss": 7.318708896636963,
+ "eval_runtime": 59.6555,
+ "eval_samples_per_second": 40.935,
+ "eval_steps_per_second": 1.291,
+ "step": 144
+ },
+ {
+ "epoch": 2.0139737991266378,
+ "grad_norm": 0.019757507368922234,
+ "learning_rate": 0.0006,
+ "loss": 7.30288553237915,
+ "step": 145
+ },
+ {
+ "epoch": 2.027947598253275,
+ "grad_norm": 0.014571248553693295,
+ "learning_rate": 0.0006,
+ "loss": 7.274681568145752,
+ "step": 146
+ },
+ {
+ "epoch": 2.041921397379913,
+ "grad_norm": 0.015744198113679886,
+ "learning_rate": 0.0006,
+ "loss": 7.278632640838623,
+ "step": 147
+ },
+ {
+ "epoch": 2.05589519650655,
+ "grad_norm": 0.028048858046531677,
+ "learning_rate": 0.0006,
+ "loss": 7.26558780670166,
+ "step": 148
+ },
+ {
+ "epoch": 2.069868995633188,
+ "grad_norm": 0.06547307223081589,
+ "learning_rate": 0.0006,
+ "loss": 7.32567834854126,
+ "step": 149
+ },
+ {
+ "epoch": 2.0838427947598253,
+ "grad_norm": 0.05969763547182083,
+ "learning_rate": 0.0006,
+ "loss": 7.395836353302002,
+ "step": 150
+ },
+ {
+ "epoch": 2.097816593886463,
+ "grad_norm": 0.032574281096458435,
+ "learning_rate": 0.0006,
+ "loss": 7.318150997161865,
+ "step": 151
+ },
+ {
+ "epoch": 2.1117903930131003,
+ "grad_norm": 0.029452962800860405,
+ "learning_rate": 0.0006,
+ "loss": 7.308066368103027,
+ "step": 152
+ },
+ {
+ "epoch": 2.125764192139738,
+ "grad_norm": 0.03161991387605667,
+ "learning_rate": 0.0006,
+ "loss": 7.306290626525879,
+ "step": 153
+ },
+ {
+ "epoch": 2.1397379912663754,
+ "grad_norm": 0.027096295729279518,
+ "learning_rate": 0.0006,
+ "loss": 7.297796249389648,
+ "step": 154
+ },
+ {
+ "epoch": 2.153711790393013,
+ "grad_norm": 0.028657056391239166,
+ "learning_rate": 0.0006,
+ "loss": 7.317424774169922,
+ "step": 155
+ },
+ {
+ "epoch": 2.1676855895196505,
+ "grad_norm": 0.02046995982527733,
+ "learning_rate": 0.0006,
+ "loss": 7.297986030578613,
+ "step": 156
+ },
+ {
+ "epoch": 2.1816593886462883,
+ "grad_norm": 0.02220369316637516,
+ "learning_rate": 0.0006,
+ "loss": 7.268237590789795,
+ "step": 157
+ },
+ {
+ "epoch": 2.1956331877729256,
+ "grad_norm": 0.018381357192993164,
+ "learning_rate": 0.0006,
+ "loss": 7.259415626525879,
+ "step": 158
+ },
+ {
+ "epoch": 2.2096069868995634,
+ "grad_norm": 0.0204413291066885,
+ "learning_rate": 0.0006,
+ "loss": 7.258173942565918,
+ "step": 159
+ },
+ {
+ "epoch": 2.2235807860262007,
+ "grad_norm": 0.015874288976192474,
+ "learning_rate": 0.0006,
+ "loss": 7.281834602355957,
+ "step": 160
+ },
+ {
+ "epoch": 2.2375545851528384,
+ "grad_norm": 0.01634068600833416,
+ "learning_rate": 0.0006,
+ "loss": 7.2469868659973145,
+ "step": 161
+ },
+ {
+ "epoch": 2.251528384279476,
+ "grad_norm": 0.01728747971355915,
+ "learning_rate": 0.0006,
+ "loss": 7.280995845794678,
+ "step": 162
+ },
+ {
+ "epoch": 2.2655021834061135,
+ "grad_norm": 0.016341188922524452,
+ "learning_rate": 0.0006,
+ "loss": 7.273357391357422,
+ "step": 163
+ },
+ {
+ "epoch": 2.279475982532751,
+ "grad_norm": 0.016125807538628578,
+ "learning_rate": 0.0006,
+ "loss": 7.273510932922363,
+ "step": 164
+ },
+ {
+ "epoch": 2.2934497816593886,
+ "grad_norm": 0.01629287749528885,
+ "learning_rate": 0.0006,
+ "loss": 7.267029762268066,
+ "step": 165
+ },
+ {
+ "epoch": 2.3074235807860264,
+ "grad_norm": 0.015296131372451782,
+ "learning_rate": 0.0006,
+ "loss": 7.211989402770996,
+ "step": 166
+ },
+ {
+ "epoch": 2.3213973799126637,
+ "grad_norm": 0.01709570363163948,
+ "learning_rate": 0.0006,
+ "loss": 7.242851734161377,
+ "step": 167
+ },
+ {
+ "epoch": 2.3353711790393015,
+ "grad_norm": 0.012552388943731785,
+ "learning_rate": 0.0006,
+ "loss": 7.241696834564209,
+ "step": 168
+ },
+ {
+ "epoch": 2.349344978165939,
+ "grad_norm": 0.014571291394531727,
+ "learning_rate": 0.0006,
+ "loss": 7.270681381225586,
+ "step": 169
+ },
+ {
+ "epoch": 2.3633187772925766,
+ "grad_norm": 0.013389437459409237,
+ "learning_rate": 0.0006,
+ "loss": 7.191600799560547,
+ "step": 170
+ },
+ {
+ "epoch": 2.377292576419214,
+ "grad_norm": 0.011895314790308475,
+ "learning_rate": 0.0006,
+ "loss": 7.213610649108887,
+ "step": 171
+ },
+ {
+ "epoch": 2.3912663755458516,
+ "grad_norm": 0.01345642190426588,
+ "learning_rate": 0.0006,
+ "loss": 7.193787574768066,
+ "step": 172
+ },
+ {
+ "epoch": 2.405240174672489,
+ "grad_norm": 0.0091794328764081,
+ "learning_rate": 0.0006,
+ "loss": 7.233364105224609,
+ "step": 173
+ },
+ {
+ "epoch": 2.4192139737991267,
+ "grad_norm": 0.013767086900770664,
+ "learning_rate": 0.0006,
+ "loss": 7.217272758483887,
+ "step": 174
+ },
+ {
+ "epoch": 2.433187772925764,
+ "grad_norm": 0.015134960412979126,
+ "learning_rate": 0.0006,
+ "loss": 7.192791938781738,
+ "step": 175
+ },
+ {
+ "epoch": 2.447161572052402,
+ "grad_norm": 0.01790648326277733,
+ "learning_rate": 0.0006,
+ "loss": 7.229467391967773,
+ "step": 176
+ },
+ {
+ "epoch": 2.461135371179039,
+ "grad_norm": 0.015362377278506756,
+ "learning_rate": 0.0006,
+ "loss": 7.180346488952637,
+ "step": 177
+ },
+ {
+ "epoch": 2.475109170305677,
+ "grad_norm": 0.010505498386919498,
+ "learning_rate": 0.0006,
+ "loss": 7.188833236694336,
+ "step": 178
+ },
+ {
+ "epoch": 2.489082969432314,
+ "grad_norm": 0.009129747748374939,
+ "learning_rate": 0.0006,
+ "loss": 7.203153133392334,
+ "step": 179
+ },
+ {
+ "epoch": 2.503056768558952,
+ "grad_norm": 0.01147883478552103,
+ "learning_rate": 0.0006,
+ "loss": 7.163824558258057,
+ "step": 180
+ },
+ {
+ "epoch": 2.5170305676855893,
+ "grad_norm": 0.011338942684233189,
+ "learning_rate": 0.0006,
+ "loss": 7.16287899017334,
+ "step": 181
+ },
+ {
+ "epoch": 2.531004366812227,
+ "grad_norm": 0.013277675025165081,
+ "learning_rate": 0.0006,
+ "loss": 7.155905723571777,
+ "step": 182
+ },
+ {
+ "epoch": 2.544978165938865,
+ "grad_norm": 0.015269347466528416,
+ "learning_rate": 0.0006,
+ "loss": 7.189371109008789,
+ "step": 183
+ },
+ {
+ "epoch": 2.558951965065502,
+ "grad_norm": 0.017905596643686295,
+ "learning_rate": 0.0006,
+ "loss": 7.177547931671143,
+ "step": 184
+ },
+ {
+ "epoch": 2.5729257641921395,
+ "grad_norm": 0.024431610479950905,
+ "learning_rate": 0.0006,
+ "loss": 7.192169189453125,
+ "step": 185
+ },
+ {
+ "epoch": 2.5868995633187772,
+ "grad_norm": 0.03656961768865585,
+ "learning_rate": 0.0006,
+ "loss": 7.124387741088867,
+ "step": 186
+ },
+ {
+ "epoch": 2.600873362445415,
+ "grad_norm": 0.04337048903107643,
+ "learning_rate": 0.0006,
+ "loss": 7.140231609344482,
+ "step": 187
+ },
+ {
+ "epoch": 2.6148471615720523,
+ "grad_norm": 0.04047228768467903,
+ "learning_rate": 0.0006,
+ "loss": 7.171927452087402,
+ "step": 188
+ },
+ {
+ "epoch": 2.62882096069869,
+ "grad_norm": 0.07825770974159241,
+ "learning_rate": 0.0006,
+ "loss": 7.207631587982178,
+ "step": 189
+ },
+ {
+ "epoch": 2.6427947598253274,
+ "grad_norm": 0.07118063420057297,
+ "learning_rate": 0.0006,
+ "loss": 7.209488391876221,
+ "step": 190
+ },
+ {
+ "epoch": 2.656768558951965,
+ "grad_norm": 0.05153293535113335,
+ "learning_rate": 0.0006,
+ "loss": 7.14793062210083,
+ "step": 191
+ },
+ {
+ "epoch": 2.6707423580786025,
+ "grad_norm": 0.03318839520215988,
+ "learning_rate": 0.0006,
+ "loss": 7.163010597229004,
+ "step": 192
+ },
+ {
+ "epoch": 2.6847161572052403,
+ "grad_norm": 0.029333539307117462,
+ "learning_rate": 0.0006,
+ "loss": 7.159955978393555,
+ "step": 193
+ },
+ {
+ "epoch": 2.6986899563318776,
+ "grad_norm": 0.029405880719423294,
+ "learning_rate": 0.0006,
+ "loss": 7.136178493499756,
+ "step": 194
+ },
+ {
+ "epoch": 2.7126637554585153,
+ "grad_norm": 0.037815988063812256,
+ "learning_rate": 0.0006,
+ "loss": 7.183186054229736,
+ "step": 195
+ },
+ {
+ "epoch": 2.726637554585153,
+ "grad_norm": 0.020307250320911407,
+ "learning_rate": 0.0006,
+ "loss": 7.133164882659912,
+ "step": 196
+ },
+ {
+ "epoch": 2.7406113537117904,
+ "grad_norm": 0.026048069819808006,
+ "learning_rate": 0.0006,
+ "loss": 7.124977111816406,
+ "step": 197
+ },
+ {
+ "epoch": 2.7545851528384278,
+ "grad_norm": 0.027316724881529808,
+ "learning_rate": 0.0006,
+ "loss": 7.07265567779541,
+ "step": 198
+ },
+ {
+ "epoch": 2.7685589519650655,
+ "grad_norm": 0.021629009395837784,
+ "learning_rate": 0.0006,
+ "loss": 7.150020599365234,
+ "step": 199
+ },
+ {
+ "epoch": 2.7825327510917033,
+ "grad_norm": 0.01605929434299469,
+ "learning_rate": 0.0006,
+ "loss": 7.10919189453125,
+ "step": 200
+ },
+ {
+ "epoch": 2.7965065502183406,
+ "grad_norm": 0.020800089463591576,
+ "learning_rate": 0.0006,
+ "loss": 7.132045269012451,
+ "step": 201
+ },
+ {
+ "epoch": 2.810480349344978,
+ "grad_norm": 0.019225774332880974,
+ "learning_rate": 0.0006,
+ "loss": 7.128364562988281,
+ "step": 202
+ },
+ {
+ "epoch": 2.8244541484716157,
+ "grad_norm": 0.01561372634023428,
+ "learning_rate": 0.0006,
+ "loss": 7.083669662475586,
+ "step": 203
+ },
+ {
+ "epoch": 2.8384279475982535,
+ "grad_norm": 0.021969519555568695,
+ "learning_rate": 0.0006,
+ "loss": 7.050841331481934,
+ "step": 204
+ },
+ {
+ "epoch": 2.8524017467248908,
+ "grad_norm": 0.021453695371747017,
+ "learning_rate": 0.0006,
+ "loss": 7.097085475921631,
+ "step": 205
+ },
+ {
+ "epoch": 2.8663755458515285,
+ "grad_norm": 0.022836022078990936,
+ "learning_rate": 0.0006,
+ "loss": 7.10405969619751,
+ "step": 206
+ },
+ {
+ "epoch": 2.880349344978166,
+ "grad_norm": 0.02430851384997368,
+ "learning_rate": 0.0006,
+ "loss": 7.045181751251221,
+ "step": 207
+ },
+ {
+ "epoch": 2.8943231441048036,
+ "grad_norm": 0.020166993141174316,
+ "learning_rate": 0.0006,
+ "loss": 7.068183898925781,
+ "step": 208
+ },
+ {
+ "epoch": 2.908296943231441,
+ "grad_norm": 0.0118505684658885,
+ "learning_rate": 0.0006,
+ "loss": 7.07132625579834,
+ "step": 209
+ },
+ {
+ "epoch": 2.9222707423580787,
+ "grad_norm": 0.023772427812218666,
+ "learning_rate": 0.0006,
+ "loss": 7.055238723754883,
+ "step": 210
+ },
+ {
+ "epoch": 2.936244541484716,
+ "grad_norm": 0.02350712940096855,
+ "learning_rate": 0.0006,
+ "loss": 7.055898666381836,
+ "step": 211
+ },
+ {
+ "epoch": 2.950218340611354,
+ "grad_norm": 0.017244907096028328,
+ "learning_rate": 0.0006,
+ "loss": 7.0099711418151855,
+ "step": 212
+ },
+ {
+ "epoch": 2.964192139737991,
+ "grad_norm": 0.021565575152635574,
+ "learning_rate": 0.0006,
+ "loss": 6.991751670837402,
+ "step": 213
+ },
+ {
+ "epoch": 2.978165938864629,
+ "grad_norm": 0.03241860866546631,
+ "learning_rate": 0.0006,
+ "loss": 7.005980491638184,
+ "step": 214
+ },
+ {
+ "epoch": 2.992139737991266,
+ "grad_norm": 0.04613726586103439,
+ "learning_rate": 0.0006,
+ "loss": 7.027859687805176,
+ "step": 215
+ },
+ {
+ "epoch": 3.0,
+ "grad_norm": 0.06404894590377808,
+ "learning_rate": 0.0006,
+ "loss": 7.123725891113281,
+ "step": 216
+ },
+ {
+ "epoch": 3.0,
+ "eval_loss": 7.070647716522217,
+ "eval_runtime": 58.5801,
+ "eval_samples_per_second": 41.687,
+ "eval_steps_per_second": 1.314,
+ "step": 216
+ },
+ {
+ "epoch": 3.0139737991266378,
+ "grad_norm": 0.05542432889342308,
+ "learning_rate": 0.0006,
+ "loss": 7.078299522399902,
+ "step": 217
+ },
+ {
+ "epoch": 3.027947598253275,
+ "grad_norm": 0.026484526693820953,
+ "learning_rate": 0.0006,
+ "loss": 6.961159706115723,
+ "step": 218
+ },
+ {
+ "epoch": 3.041921397379913,
+ "grad_norm": 0.028276391327381134,
+ "learning_rate": 0.0006,
+ "loss": 7.019961833953857,
+ "step": 219
+ },
+ {
+ "epoch": 3.05589519650655,
+ "grad_norm": 0.024486595764756203,
+ "learning_rate": 0.0006,
+ "loss": 7.039677619934082,
+ "step": 220
+ },
+ {
+ "epoch": 3.069868995633188,
+ "grad_norm": 0.017620893195271492,
+ "learning_rate": 0.0006,
+ "loss": 6.9270782470703125,
+ "step": 221
+ },
+ {
+ "epoch": 3.0838427947598253,
+ "grad_norm": 0.02547302283346653,
+ "learning_rate": 0.0006,
+ "loss": 6.924233436584473,
+ "step": 222
+ },
+ {
+ "epoch": 3.097816593886463,
+ "grad_norm": 0.02539009042084217,
+ "learning_rate": 0.0006,
+ "loss": 6.973979949951172,
+ "step": 223
+ },
+ {
+ "epoch": 3.1117903930131003,
+ "grad_norm": 0.03256227448582649,
+ "learning_rate": 0.0006,
+ "loss": 6.933725357055664,
+ "step": 224
+ },
+ {
+ "epoch": 3.125764192139738,
+ "grad_norm": 0.04651800915598869,
+ "learning_rate": 0.0006,
+ "loss": 6.951044082641602,
+ "step": 225
+ },
+ {
+ "epoch": 3.1397379912663754,
+ "grad_norm": 0.04822975769639015,
+ "learning_rate": 0.0006,
+ "loss": 6.9614410400390625,
+ "step": 226
+ },
+ {
+ "epoch": 3.153711790393013,
+ "grad_norm": 0.031165020540356636,
+ "learning_rate": 0.0006,
+ "loss": 6.91193151473999,
+ "step": 227
+ },
+ {
+ "epoch": 3.1676855895196505,
+ "grad_norm": 0.029754292219877243,
+ "learning_rate": 0.0006,
+ "loss": 6.91802978515625,
+ "step": 228
+ },
+ {
+ "epoch": 3.1816593886462883,
+ "grad_norm": 0.02193152718245983,
+ "learning_rate": 0.0006,
+ "loss": 6.928090572357178,
+ "step": 229
+ },
+ {
+ "epoch": 3.1956331877729256,
+ "grad_norm": 0.02428470179438591,
+ "learning_rate": 0.0006,
+ "loss": 6.90396785736084,
+ "step": 230
+ },
+ {
+ "epoch": 3.2096069868995634,
+ "grad_norm": 0.023374345153570175,
+ "learning_rate": 0.0006,
+ "loss": 6.885858535766602,
+ "step": 231
+ },
+ {
+ "epoch": 3.2235807860262007,
+ "grad_norm": 0.023405000567436218,
+ "learning_rate": 0.0006,
+ "loss": 6.8855767250061035,
+ "step": 232
+ },
+ {
+ "epoch": 3.2375545851528384,
+ "grad_norm": 0.017277134582400322,
+ "learning_rate": 0.0006,
+ "loss": 6.901449203491211,
+ "step": 233
+ },
+ {
+ "epoch": 3.251528384279476,
+ "grad_norm": 0.019496750086545944,
+ "learning_rate": 0.0006,
+ "loss": 6.885687828063965,
+ "step": 234
+ },
+ {
+ "epoch": 3.2655021834061135,
+ "grad_norm": 0.016893276944756508,
+ "learning_rate": 0.0006,
+ "loss": 6.872369766235352,
+ "step": 235
+ },
+ {
+ "epoch": 3.279475982532751,
+ "grad_norm": 0.02058715932071209,
+ "learning_rate": 0.0006,
+ "loss": 6.823280334472656,
+ "step": 236
+ },
+ {
+ "epoch": 3.2934497816593886,
+ "grad_norm": 0.015530755743384361,
+ "learning_rate": 0.0006,
+ "loss": 6.860499858856201,
+ "step": 237
+ },
+ {
+ "epoch": 3.3074235807860264,
+ "grad_norm": 0.01934926211833954,
+ "learning_rate": 0.0006,
+ "loss": 6.817448616027832,
+ "step": 238
+ },
+ {
+ "epoch": 3.3213973799126637,
+ "grad_norm": 0.017413552850484848,
+ "learning_rate": 0.0006,
+ "loss": 6.881955146789551,
+ "step": 239
+ },
+ {
+ "epoch": 3.3353711790393015,
+ "grad_norm": 0.026478875428438187,
+ "learning_rate": 0.0006,
+ "loss": 6.827404499053955,
+ "step": 240
+ },
+ {
+ "epoch": 3.349344978165939,
+ "grad_norm": 0.032611701637506485,
+ "learning_rate": 0.0006,
+ "loss": 6.78285551071167,
+ "step": 241
+ },
+ {
+ "epoch": 3.3633187772925766,
+ "grad_norm": 0.041900306940078735,
+ "learning_rate": 0.0006,
+ "loss": 6.783053874969482,
+ "step": 242
+ },
+ {
+ "epoch": 3.377292576419214,
+ "grad_norm": 0.04149497672915459,
+ "learning_rate": 0.0006,
+ "loss": 6.745220184326172,
+ "step": 243
+ },
+ {
+ "epoch": 3.3912663755458516,
+ "grad_norm": 0.023949826136231422,
+ "learning_rate": 0.0006,
+ "loss": 6.841533184051514,
+ "step": 244
+ },
+ {
+ "epoch": 3.405240174672489,
+ "grad_norm": 0.028601842001080513,
+ "learning_rate": 0.0006,
+ "loss": 6.852852821350098,
+ "step": 245
+ },
+ {
+ "epoch": 3.4192139737991267,
+ "grad_norm": 0.027365155518054962,
+ "learning_rate": 0.0006,
+ "loss": 6.74376106262207,
+ "step": 246
+ },
+ {
+ "epoch": 3.433187772925764,
+ "grad_norm": 0.02856568619608879,
+ "learning_rate": 0.0006,
+ "loss": 6.74906063079834,
+ "step": 247
+ },
+ {
+ "epoch": 3.447161572052402,
+ "grad_norm": 0.019172416999936104,
+ "learning_rate": 0.0006,
+ "loss": 6.809206008911133,
+ "step": 248
+ },
+ {
+ "epoch": 3.461135371179039,
+ "grad_norm": 0.02161746844649315,
+ "learning_rate": 0.0006,
+ "loss": 6.789888381958008,
+ "step": 249
+ },
+ {
+ "epoch": 3.475109170305677,
+ "grad_norm": 0.020273666828870773,
+ "learning_rate": 0.0006,
+ "loss": 6.73512077331543,
+ "step": 250
+ },
+ {
+ "epoch": 3.489082969432314,
+ "grad_norm": 0.022329283878207207,
+ "learning_rate": 0.0006,
+ "loss": 6.826148509979248,
+ "step": 251
+ },
+ {
+ "epoch": 3.503056768558952,
+ "grad_norm": 0.02284284122288227,
+ "learning_rate": 0.0006,
+ "loss": 6.748521327972412,
+ "step": 252
+ },
+ {
+ "epoch": 3.5170305676855893,
+ "grad_norm": 0.026368247345089912,
+ "learning_rate": 0.0006,
+ "loss": 6.645879745483398,
+ "step": 253
+ },
+ {
+ "epoch": 3.531004366812227,
+ "grad_norm": 0.03627244383096695,
+ "learning_rate": 0.0006,
+ "loss": 6.666064739227295,
+ "step": 254
+ },
+ {
+ "epoch": 3.544978165938865,
+ "grad_norm": 0.04140935093164444,
+ "learning_rate": 0.0006,
+ "loss": 6.712275981903076,
+ "step": 255
+ },
+ {
+ "epoch": 3.558951965065502,
+ "grad_norm": 0.0390462800860405,
+ "learning_rate": 0.0006,
+ "loss": 6.665900230407715,
+ "step": 256
+ },
+ {
+ "epoch": 3.5729257641921395,
+ "grad_norm": 0.05178583785891533,
+ "learning_rate": 0.0006,
+ "loss": 6.753298759460449,
+ "step": 257
+ },
+ {
+ "epoch": 3.5868995633187772,
+ "grad_norm": 0.054328061640262604,
+ "learning_rate": 0.0006,
+ "loss": 6.732633113861084,
+ "step": 258
+ },
+ {
+ "epoch": 3.600873362445415,
+ "grad_norm": 0.04590460658073425,
+ "learning_rate": 0.0006,
+ "loss": 6.65250301361084,
+ "step": 259
+ },
+ {
+ "epoch": 3.6148471615720523,
+ "grad_norm": 0.044331666082143784,
+ "learning_rate": 0.0006,
+ "loss": 6.68654203414917,
+ "step": 260
+ },
+ {
+ "epoch": 3.62882096069869,
+ "grad_norm": 0.05308730527758598,
+ "learning_rate": 0.0006,
+ "loss": 6.642482757568359,
+ "step": 261
+ },
+ {
+ "epoch": 3.6427947598253274,
+ "grad_norm": 0.03999016433954239,
+ "learning_rate": 0.0006,
+ "loss": 6.625227451324463,
+ "step": 262
+ },
+ {
+ "epoch": 3.656768558951965,
+ "grad_norm": 0.04088086634874344,
+ "learning_rate": 0.0006,
+ "loss": 6.67623233795166,
+ "step": 263
+ },
+ {
+ "epoch": 3.6707423580786025,
+ "grad_norm": 0.03615636005997658,
+ "learning_rate": 0.0006,
+ "loss": 6.6869001388549805,
+ "step": 264
+ },
+ {
+ "epoch": 3.6847161572052403,
+ "grad_norm": 0.02817360684275627,
+ "learning_rate": 0.0006,
+ "loss": 6.674601078033447,
+ "step": 265
+ },
+ {
+ "epoch": 3.6986899563318776,
+ "grad_norm": 0.032586123794317245,
+ "learning_rate": 0.0006,
+ "loss": 6.612994194030762,
+ "step": 266
+ },
+ {
+ "epoch": 3.7126637554585153,
+ "grad_norm": 0.031117867678403854,
+ "learning_rate": 0.0006,
+ "loss": 6.6244354248046875,
+ "step": 267
+ },
+ {
+ "epoch": 3.726637554585153,
+ "grad_norm": 0.028330687433481216,
+ "learning_rate": 0.0006,
+ "loss": 6.669557094573975,
+ "step": 268
+ },
+ {
+ "epoch": 3.7406113537117904,
+ "grad_norm": 0.021489202976226807,
+ "learning_rate": 0.0006,
+ "loss": 6.632035732269287,
+ "step": 269
+ },
+ {
+ "epoch": 3.7545851528384278,
+ "grad_norm": 0.024269424378871918,
+ "learning_rate": 0.0006,
+ "loss": 6.660867691040039,
+ "step": 270
+ },
+ {
+ "epoch": 3.7685589519650655,
+ "grad_norm": 0.02237873338162899,
+ "learning_rate": 0.0006,
+ "loss": 6.602587699890137,
+ "step": 271
+ },
+ {
+ "epoch": 3.7825327510917033,
+ "grad_norm": 0.026503929868340492,
+ "learning_rate": 0.0006,
+ "loss": 6.599200248718262,
+ "step": 272
+ },
+ {
+ "epoch": 3.7965065502183406,
+ "grad_norm": 0.03687124326825142,
+ "learning_rate": 0.0006,
+ "loss": 6.567296504974365,
+ "step": 273
+ },
+ {
+ "epoch": 3.810480349344978,
+ "grad_norm": 0.04708952456712723,
+ "learning_rate": 0.0006,
+ "loss": 6.621265411376953,
+ "step": 274
+ },
+ {
+ "epoch": 3.8244541484716157,
+ "grad_norm": 0.041846614331007004,
+ "learning_rate": 0.0006,
+ "loss": 6.550536155700684,
+ "step": 275
+ },
+ {
+ "epoch": 3.8384279475982535,
+ "grad_norm": 0.031249675899744034,
+ "learning_rate": 0.0006,
+ "loss": 6.54278564453125,
+ "step": 276
+ },
+ {
+ "epoch": 3.8524017467248908,
+ "grad_norm": 0.023082256317138672,
+ "learning_rate": 0.0006,
+ "loss": 6.544787406921387,
+ "step": 277
+ },
+ {
+ "epoch": 3.8663755458515285,
+ "grad_norm": 0.020995570346713066,
+ "learning_rate": 0.0006,
+ "loss": 6.518143653869629,
+ "step": 278
+ },
+ {
+ "epoch": 3.880349344978166,
+ "grad_norm": 0.020839324221014977,
+ "learning_rate": 0.0006,
+ "loss": 6.546667098999023,
+ "step": 279
+ },
+ {
+ "epoch": 3.8943231441048036,
+ "grad_norm": 0.018377432599663734,
+ "learning_rate": 0.0006,
+ "loss": 6.478307723999023,
+ "step": 280
+ },
+ {
+ "epoch": 3.908296943231441,
+ "grad_norm": 0.021713724359869957,
+ "learning_rate": 0.0006,
+ "loss": 6.485574245452881,
+ "step": 281
+ },
+ {
+ "epoch": 3.9222707423580787,
+ "grad_norm": 0.01359301246702671,
+ "learning_rate": 0.0006,
+ "loss": 6.570268630981445,
+ "step": 282
+ },
+ {
+ "epoch": 3.936244541484716,
+ "grad_norm": 0.016233332455158234,
+ "learning_rate": 0.0006,
+ "loss": 6.57150936126709,
+ "step": 283
+ },
+ {
+ "epoch": 3.950218340611354,
+ "grad_norm": 0.020687608048319817,
+ "learning_rate": 0.0006,
+ "loss": 6.527956008911133,
+ "step": 284
+ },
+ {
+ "epoch": 3.964192139737991,
+ "grad_norm": 0.02032964862883091,
+ "learning_rate": 0.0006,
+ "loss": 6.468169689178467,
+ "step": 285
+ },
+ {
+ "epoch": 3.978165938864629,
+ "grad_norm": 0.016676288098096848,
+ "learning_rate": 0.0006,
+ "loss": 6.448418617248535,
+ "step": 286
+ },
+ {
+ "epoch": 3.992139737991266,
+ "grad_norm": 0.0168539360165596,
+ "learning_rate": 0.0006,
+ "loss": 6.461180686950684,
+ "step": 287
+ },
+ {
+ "epoch": 4.0,
+ "grad_norm": 0.02666233666241169,
+ "learning_rate": 0.0006,
+ "loss": 6.344893932342529,
+ "step": 288
+ },
+ {
+ "epoch": 4.0,
+ "eval_loss": 6.5384979248046875,
+ "eval_runtime": 58.814,
+ "eval_samples_per_second": 41.521,
+ "eval_steps_per_second": 1.309,
+ "step": 288
+ },
+ {
+ "epoch": 4.013973799126638,
+ "grad_norm": 0.04154708981513977,
+ "learning_rate": 0.0006,
+ "loss": 6.41855525970459,
+ "step": 289
+ },
+ {
+ "epoch": 4.0279475982532755,
+ "grad_norm": 0.04895668104290962,
+ "learning_rate": 0.0006,
+ "loss": 6.499302864074707,
+ "step": 290
+ },
+ {
+ "epoch": 4.041921397379912,
+ "grad_norm": 0.034464482218027115,
+ "learning_rate": 0.0006,
+ "loss": 6.438611030578613,
+ "step": 291
+ },
+ {
+ "epoch": 4.05589519650655,
+ "grad_norm": 0.06845773756504059,
+ "learning_rate": 0.0006,
+ "loss": 6.538361072540283,
+ "step": 292
+ },
+ {
+ "epoch": 4.069868995633188,
+ "grad_norm": 0.053768858313560486,
+ "learning_rate": 0.0006,
+ "loss": 6.527524948120117,
+ "step": 293
+ },
+ {
+ "epoch": 4.083842794759826,
+ "grad_norm": 0.039037931710481644,
+ "learning_rate": 0.0006,
+ "loss": 6.503847122192383,
+ "step": 294
+ },
+ {
+ "epoch": 4.097816593886463,
+ "grad_norm": 0.03962196782231331,
+ "learning_rate": 0.0006,
+ "loss": 6.475178241729736,
+ "step": 295
+ },
+ {
+ "epoch": 4.1117903930131,
+ "grad_norm": 0.049101535230875015,
+ "learning_rate": 0.0006,
+ "loss": 6.474587917327881,
+ "step": 296
+ },
+ {
+ "epoch": 4.125764192139738,
+ "grad_norm": 0.02830282226204872,
+ "learning_rate": 0.0006,
+ "loss": 6.45065975189209,
+ "step": 297
+ },
+ {
+ "epoch": 4.139737991266376,
+ "grad_norm": 0.023153482005000114,
+ "learning_rate": 0.0006,
+ "loss": 6.5218706130981445,
+ "step": 298
+ },
+ {
+ "epoch": 4.153711790393013,
+ "grad_norm": 0.025064002722501755,
+ "learning_rate": 0.0006,
+ "loss": 6.456475257873535,
+ "step": 299
+ },
+ {
+ "epoch": 4.1676855895196505,
+ "grad_norm": 0.02357092685997486,
+ "learning_rate": 0.0006,
+ "loss": 6.485952377319336,
+ "step": 300
+ },
+ {
+ "epoch": 4.181659388646288,
+ "grad_norm": 0.021832305938005447,
+ "learning_rate": 0.0006,
+ "loss": 6.425506591796875,
+ "step": 301
+ },
+ {
+ "epoch": 4.195633187772926,
+ "grad_norm": 0.0227900929749012,
+ "learning_rate": 0.0006,
+ "loss": 6.439155578613281,
+ "step": 302
+ },
+ {
+ "epoch": 4.209606986899563,
+ "grad_norm": 0.015961607918143272,
+ "learning_rate": 0.0006,
+ "loss": 6.460862636566162,
+ "step": 303
+ },
+ {
+ "epoch": 4.223580786026201,
+ "grad_norm": 0.017876390367746353,
+ "learning_rate": 0.0006,
+ "loss": 6.44564151763916,
+ "step": 304
+ },
+ {
+ "epoch": 4.2375545851528384,
+ "grad_norm": 0.013703204691410065,
+ "learning_rate": 0.0006,
+ "loss": 6.418414115905762,
+ "step": 305
+ },
+ {
+ "epoch": 4.251528384279476,
+ "grad_norm": 0.015240306034684181,
+ "learning_rate": 0.0006,
+ "loss": 6.411238670349121,
+ "step": 306
+ },
+ {
+ "epoch": 4.265502183406113,
+ "grad_norm": 0.01587662845849991,
+ "learning_rate": 0.0006,
+ "loss": 6.395642280578613,
+ "step": 307
+ },
+ {
+ "epoch": 4.279475982532751,
+ "grad_norm": 0.014650700613856316,
+ "learning_rate": 0.0006,
+ "loss": 6.302793025970459,
+ "step": 308
+ },
+ {
+ "epoch": 4.293449781659389,
+ "grad_norm": 0.016533225774765015,
+ "learning_rate": 0.0006,
+ "loss": 6.380588054656982,
+ "step": 309
+ },
+ {
+ "epoch": 4.307423580786026,
+ "grad_norm": 0.019963741302490234,
+ "learning_rate": 0.0006,
+ "loss": 6.345336437225342,
+ "step": 310
+ },
+ {
+ "epoch": 4.321397379912664,
+ "grad_norm": 0.02237936295568943,
+ "learning_rate": 0.0006,
+ "loss": 6.387770652770996,
+ "step": 311
+ },
+ {
+ "epoch": 4.335371179039301,
+ "grad_norm": 0.026658328250050545,
+ "learning_rate": 0.0006,
+ "loss": 6.323662757873535,
+ "step": 312
+ },
+ {
+ "epoch": 4.349344978165939,
+ "grad_norm": 0.02852284163236618,
+ "learning_rate": 0.0006,
+ "loss": 6.382204055786133,
+ "step": 313
+ },
+ {
+ "epoch": 4.3633187772925766,
+ "grad_norm": 0.028815090656280518,
+ "learning_rate": 0.0006,
+ "loss": 6.402390480041504,
+ "step": 314
+ },
+ {
+ "epoch": 4.377292576419214,
+ "grad_norm": 0.029393529519438744,
+ "learning_rate": 0.0006,
+ "loss": 6.280278205871582,
+ "step": 315
+ },
+ {
+ "epoch": 4.391266375545851,
+ "grad_norm": 0.02186041697859764,
+ "learning_rate": 0.0006,
+ "loss": 6.302707672119141,
+ "step": 316
+ },
+ {
+ "epoch": 4.405240174672489,
+ "grad_norm": 0.02071218006312847,
+ "learning_rate": 0.0006,
+ "loss": 6.3365020751953125,
+ "step": 317
+ },
+ {
+ "epoch": 4.419213973799127,
+ "grad_norm": 0.02779117226600647,
+ "learning_rate": 0.0006,
+ "loss": 6.252157211303711,
+ "step": 318
+ },
+ {
+ "epoch": 4.4331877729257645,
+ "grad_norm": 0.03470654785633087,
+ "learning_rate": 0.0006,
+ "loss": 6.319070339202881,
+ "step": 319
+ },
+ {
+ "epoch": 4.447161572052401,
+ "grad_norm": 0.037067804485559464,
+ "learning_rate": 0.0006,
+ "loss": 6.314116477966309,
+ "step": 320
+ },
+ {
+ "epoch": 4.461135371179039,
+ "grad_norm": 0.032182756811380386,
+ "learning_rate": 0.0006,
+ "loss": 6.267298698425293,
+ "step": 321
+ },
+ {
+ "epoch": 4.475109170305677,
+ "grad_norm": 0.026305217295885086,
+ "learning_rate": 0.0006,
+ "loss": 6.331688404083252,
+ "step": 322
+ },
+ {
+ "epoch": 4.489082969432315,
+ "grad_norm": 0.027228357270359993,
+ "learning_rate": 0.0006,
+ "loss": 6.31139612197876,
+ "step": 323
+ },
+ {
+ "epoch": 4.503056768558952,
+ "grad_norm": 0.025781169533729553,
+ "learning_rate": 0.0006,
+ "loss": 6.194684982299805,
+ "step": 324
+ },
+ {
+ "epoch": 4.517030567685589,
+ "grad_norm": 0.029186977073550224,
+ "learning_rate": 0.0006,
+ "loss": 6.31306266784668,
+ "step": 325
+ },
+ {
+ "epoch": 4.531004366812227,
+ "grad_norm": 0.023608777672052383,
+ "learning_rate": 0.0006,
+ "loss": 6.285243034362793,
+ "step": 326
+ },
+ {
+ "epoch": 4.544978165938865,
+ "grad_norm": 0.02267594076693058,
+ "learning_rate": 0.0006,
+ "loss": 6.30342960357666,
+ "step": 327
+ },
+ {
+ "epoch": 4.558951965065502,
+ "grad_norm": 0.01926310732960701,
+ "learning_rate": 0.0006,
+ "loss": 6.280606269836426,
+ "step": 328
+ },
+ {
+ "epoch": 4.5729257641921395,
+ "grad_norm": 0.026484837755560875,
+ "learning_rate": 0.0006,
+ "loss": 6.237980365753174,
+ "step": 329
+ },
+ {
+ "epoch": 4.586899563318777,
+ "grad_norm": 0.026228854432702065,
+ "learning_rate": 0.0006,
+ "loss": 6.3018293380737305,
+ "step": 330
+ },
+ {
+ "epoch": 4.600873362445415,
+ "grad_norm": 0.022096967324614525,
+ "learning_rate": 0.0006,
+ "loss": 6.253863334655762,
+ "step": 331
+ },
+ {
+ "epoch": 4.614847161572053,
+ "grad_norm": 0.027223890647292137,
+ "learning_rate": 0.0006,
+ "loss": 6.166882038116455,
+ "step": 332
+ },
+ {
+ "epoch": 4.62882096069869,
+ "grad_norm": 0.03160124272108078,
+ "learning_rate": 0.0006,
+ "loss": 6.230309963226318,
+ "step": 333
+ },
+ {
+ "epoch": 4.642794759825327,
+ "grad_norm": 0.04110539332032204,
+ "learning_rate": 0.0006,
+ "loss": 6.2334818840026855,
+ "step": 334
+ },
+ {
+ "epoch": 4.656768558951965,
+ "grad_norm": 0.04355933889746666,
+ "learning_rate": 0.0006,
+ "loss": 6.160956382751465,
+ "step": 335
+ },
+ {
+ "epoch": 4.670742358078603,
+ "grad_norm": 0.03611086308956146,
+ "learning_rate": 0.0006,
+ "loss": 6.23399543762207,
+ "step": 336
+ },
+ {
+ "epoch": 4.68471615720524,
+ "grad_norm": 0.051811400800943375,
+ "learning_rate": 0.0006,
+ "loss": 6.241855144500732,
+ "step": 337
+ },
+ {
+ "epoch": 4.698689956331878,
+ "grad_norm": 0.057434193789958954,
+ "learning_rate": 0.0006,
+ "loss": 6.220993995666504,
+ "step": 338
+ },
+ {
+ "epoch": 4.712663755458515,
+ "grad_norm": 0.04420953616499901,
+ "learning_rate": 0.0006,
+ "loss": 6.300461292266846,
+ "step": 339
+ },
+ {
+ "epoch": 4.726637554585153,
+ "grad_norm": 0.03515457361936569,
+ "learning_rate": 0.0006,
+ "loss": 6.269349098205566,
+ "step": 340
+ },
+ {
+ "epoch": 4.74061135371179,
+ "grad_norm": 0.03222600743174553,
+ "learning_rate": 0.0006,
+ "loss": 6.248613357543945,
+ "step": 341
+ },
+ {
+ "epoch": 4.754585152838428,
+ "grad_norm": 0.0384046845138073,
+ "learning_rate": 0.0006,
+ "loss": 6.266979217529297,
+ "step": 342
+ },
+ {
+ "epoch": 4.7685589519650655,
+ "grad_norm": 0.0514867827296257,
+ "learning_rate": 0.0006,
+ "loss": 6.234103202819824,
+ "step": 343
+ },
+ {
+ "epoch": 4.782532751091703,
+ "grad_norm": 0.050747793167829514,
+ "learning_rate": 0.0006,
+ "loss": 6.322582244873047,
+ "step": 344
+ },
+ {
+ "epoch": 4.796506550218341,
+ "grad_norm": 0.040878988802433014,
+ "learning_rate": 0.0006,
+ "loss": 6.212509632110596,
+ "step": 345
+ },
+ {
+ "epoch": 4.810480349344978,
+ "grad_norm": 0.034655820578336716,
+ "learning_rate": 0.0006,
+ "loss": 6.265879154205322,
+ "step": 346
+ },
+ {
+ "epoch": 4.824454148471616,
+ "grad_norm": 0.04202224686741829,
+ "learning_rate": 0.0006,
+ "loss": 6.2896409034729,
+ "step": 347
+ },
+ {
+ "epoch": 4.8384279475982535,
+ "grad_norm": 0.03632277995347977,
+ "learning_rate": 0.0006,
+ "loss": 6.253917217254639,
+ "step": 348
+ },
+ {
+ "epoch": 4.85240174672489,
+ "grad_norm": 0.028645765036344528,
+ "learning_rate": 0.0006,
+ "loss": 6.207965850830078,
+ "step": 349
+ },
+ {
+ "epoch": 4.866375545851528,
+ "grad_norm": 0.025305170565843582,
+ "learning_rate": 0.0006,
+ "loss": 6.168068885803223,
+ "step": 350
+ },
+ {
+ "epoch": 4.880349344978166,
+ "grad_norm": 0.022137803956866264,
+ "learning_rate": 0.0006,
+ "loss": 6.235283374786377,
+ "step": 351
+ },
+ {
+ "epoch": 4.894323144104804,
+ "grad_norm": 0.01814538985490799,
+ "learning_rate": 0.0006,
+ "loss": 6.259641647338867,
+ "step": 352
+ },
+ {
+ "epoch": 4.908296943231441,
+ "grad_norm": 0.018465086817741394,
+ "learning_rate": 0.0006,
+ "loss": 6.111738681793213,
+ "step": 353
+ },
+ {
+ "epoch": 4.922270742358078,
+ "grad_norm": 0.018638933077454567,
+ "learning_rate": 0.0006,
+ "loss": 6.225642204284668,
+ "step": 354
+ },
+ {
+ "epoch": 4.936244541484716,
+ "grad_norm": 0.016461260616779327,
+ "learning_rate": 0.0006,
+ "loss": 6.140647888183594,
+ "step": 355
+ },
+ {
+ "epoch": 4.950218340611354,
+ "grad_norm": 0.015458385460078716,
+ "learning_rate": 0.0006,
+ "loss": 6.155224800109863,
+ "step": 356
+ },
+ {
+ "epoch": 4.964192139737992,
+ "grad_norm": 0.014934048056602478,
+ "learning_rate": 0.0006,
+ "loss": 6.101555824279785,
+ "step": 357
+ },
+ {
+ "epoch": 4.978165938864628,
+ "grad_norm": 0.013979545794427395,
+ "learning_rate": 0.0006,
+ "loss": 6.161718845367432,
+ "step": 358
+ },
+ {
+ "epoch": 4.992139737991266,
+ "grad_norm": 0.01640394888818264,
+ "learning_rate": 0.0006,
+ "loss": 6.055768966674805,
+ "step": 359
+ },
+ {
+ "epoch": 5.0,
+ "grad_norm": 0.015135680325329304,
+ "learning_rate": 0.0006,
+ "loss": 6.169695854187012,
+ "step": 360
+ },
+ {
+ "epoch": 5.0,
+ "eval_loss": 6.16408109664917,
+ "eval_runtime": 59.3649,
+ "eval_samples_per_second": 41.135,
+ "eval_steps_per_second": 1.297,
+ "step": 360
+ },
+ {
+ "epoch": 5.013973799126638,
+ "grad_norm": 0.012849048711359501,
+ "learning_rate": 0.0006,
+ "loss": 6.055830001831055,
+ "step": 361
+ },
+ {
+ "epoch": 5.0279475982532755,
+ "grad_norm": 0.018169576302170753,
+ "learning_rate": 0.0006,
+ "loss": 6.082655429840088,
+ "step": 362
+ },
+ {
+ "epoch": 5.041921397379912,
+ "grad_norm": 0.024293430149555206,
+ "learning_rate": 0.0006,
+ "loss": 6.109032154083252,
+ "step": 363
+ },
+ {
+ "epoch": 5.05589519650655,
+ "grad_norm": 0.03266787901520729,
+ "learning_rate": 0.0006,
+ "loss": 6.030613899230957,
+ "step": 364
+ },
+ {
+ "epoch": 5.069868995633188,
+ "grad_norm": 0.03572266176342964,
+ "learning_rate": 0.0006,
+ "loss": 6.050390720367432,
+ "step": 365
+ },
+ {
+ "epoch": 5.083842794759826,
+ "grad_norm": 0.032625213265419006,
+ "learning_rate": 0.0006,
+ "loss": 6.169313430786133,
+ "step": 366
+ },
+ {
+ "epoch": 5.097816593886463,
+ "grad_norm": 0.04051872715353966,
+ "learning_rate": 0.0006,
+ "loss": 6.10398006439209,
+ "step": 367
+ },
+ {
+ "epoch": 5.1117903930131,
+ "grad_norm": 0.0351213738322258,
+ "learning_rate": 0.0006,
+ "loss": 6.121973514556885,
+ "step": 368
+ },
+ {
+ "epoch": 5.125764192139738,
+ "grad_norm": 0.03523759916424751,
+ "learning_rate": 0.0006,
+ "loss": 6.0626654624938965,
+ "step": 369
+ },
+ {
+ "epoch": 5.139737991266376,
+ "grad_norm": 0.02706182189285755,
+ "learning_rate": 0.0006,
+ "loss": 6.049354076385498,
+ "step": 370
+ },
+ {
+ "epoch": 5.153711790393013,
+ "grad_norm": 0.024212589487433434,
+ "learning_rate": 0.0006,
+ "loss": 6.034826278686523,
+ "step": 371
+ },
+ {
+ "epoch": 5.1676855895196505,
+ "grad_norm": 0.024798082187771797,
+ "learning_rate": 0.0006,
+ "loss": 6.004058837890625,
+ "step": 372
+ },
+ {
+ "epoch": 5.181659388646288,
+ "grad_norm": 0.02325567416846752,
+ "learning_rate": 0.0006,
+ "loss": 5.986461639404297,
+ "step": 373
+ },
+ {
+ "epoch": 5.195633187772926,
+ "grad_norm": 0.019060570746660233,
+ "learning_rate": 0.0006,
+ "loss": 6.059736251831055,
+ "step": 374
+ },
+ {
+ "epoch": 5.209606986899563,
+ "grad_norm": 0.016822976991534233,
+ "learning_rate": 0.0006,
+ "loss": 6.072957515716553,
+ "step": 375
+ },
+ {
+ "epoch": 5.223580786026201,
+ "grad_norm": 0.017218658700585365,
+ "learning_rate": 0.0006,
+ "loss": 6.031739234924316,
+ "step": 376
+ },
+ {
+ "epoch": 5.2375545851528384,
+ "grad_norm": 0.014889383688569069,
+ "learning_rate": 0.0006,
+ "loss": 6.025674343109131,
+ "step": 377
+ },
+ {
+ "epoch": 5.251528384279476,
+ "grad_norm": 0.01708107627928257,
+ "learning_rate": 0.0006,
+ "loss": 6.090451240539551,
+ "step": 378
+ },
+ {
+ "epoch": 5.265502183406113,
+ "grad_norm": 0.020260317251086235,
+ "learning_rate": 0.0006,
+ "loss": 6.094257354736328,
+ "step": 379
+ },
+ {
+ "epoch": 5.279475982532751,
+ "grad_norm": 0.020110400393605232,
+ "learning_rate": 0.0006,
+ "loss": 6.021188735961914,
+ "step": 380
+ },
+ {
+ "epoch": 5.293449781659389,
+ "grad_norm": 0.02027830481529236,
+ "learning_rate": 0.0006,
+ "loss": 6.022156238555908,
+ "step": 381
+ },
+ {
+ "epoch": 5.307423580786026,
+ "grad_norm": 0.022747062146663666,
+ "learning_rate": 0.0006,
+ "loss": 6.011836051940918,
+ "step": 382
+ },
+ {
+ "epoch": 5.321397379912664,
+ "grad_norm": 0.01990230567753315,
+ "learning_rate": 0.0006,
+ "loss": 6.003054618835449,
+ "step": 383
+ },
+ {
+ "epoch": 5.335371179039301,
+ "grad_norm": 0.01596238650381565,
+ "learning_rate": 0.0006,
+ "loss": 5.991410732269287,
+ "step": 384
+ },
+ {
+ "epoch": 5.349344978165939,
+ "grad_norm": 0.015847105532884598,
+ "learning_rate": 0.0006,
+ "loss": 6.100622177124023,
+ "step": 385
+ },
+ {
+ "epoch": 5.3633187772925766,
+ "grad_norm": 0.016179397702217102,
+ "learning_rate": 0.0006,
+ "loss": 6.032659530639648,
+ "step": 386
+ },
+ {
+ "epoch": 5.377292576419214,
+ "grad_norm": 0.018256602808833122,
+ "learning_rate": 0.0006,
+ "loss": 5.961983680725098,
+ "step": 387
+ },
+ {
+ "epoch": 5.391266375545851,
+ "grad_norm": 0.02805912122130394,
+ "learning_rate": 0.0006,
+ "loss": 5.983541965484619,
+ "step": 388
+ },
+ {
+ "epoch": 5.405240174672489,
+ "grad_norm": 0.039082255214452744,
+ "learning_rate": 0.0006,
+ "loss": 6.021474361419678,
+ "step": 389
+ },
+ {
+ "epoch": 5.419213973799127,
+ "grad_norm": 0.036757659167051315,
+ "learning_rate": 0.0006,
+ "loss": 6.005046367645264,
+ "step": 390
+ },
+ {
+ "epoch": 5.4331877729257645,
+ "grad_norm": 0.035277366638183594,
+ "learning_rate": 0.0006,
+ "loss": 6.036296844482422,
+ "step": 391
+ },
+ {
+ "epoch": 5.447161572052401,
+ "grad_norm": 0.034404635429382324,
+ "learning_rate": 0.0006,
+ "loss": 5.994539260864258,
+ "step": 392
+ },
+ {
+ "epoch": 5.461135371179039,
+ "grad_norm": 0.0377206988632679,
+ "learning_rate": 0.0006,
+ "loss": 5.941534519195557,
+ "step": 393
+ },
+ {
+ "epoch": 5.475109170305677,
+ "grad_norm": 0.0389922633767128,
+ "learning_rate": 0.0006,
+ "loss": 5.983644485473633,
+ "step": 394
+ },
+ {
+ "epoch": 5.489082969432315,
+ "grad_norm": 0.04176778718829155,
+ "learning_rate": 0.0006,
+ "loss": 6.030922889709473,
+ "step": 395
+ },
+ {
+ "epoch": 5.503056768558952,
+ "grad_norm": 0.0337153784930706,
+ "learning_rate": 0.0006,
+ "loss": 5.961367130279541,
+ "step": 396
+ },
+ {
+ "epoch": 5.517030567685589,
+ "grad_norm": 0.036116816103458405,
+ "learning_rate": 0.0006,
+ "loss": 5.819280624389648,
+ "step": 397
+ },
+ {
+ "epoch": 5.531004366812227,
+ "grad_norm": 0.030724041163921356,
+ "learning_rate": 0.0006,
+ "loss": 5.906380653381348,
+ "step": 398
+ },
+ {
+ "epoch": 5.544978165938865,
+ "grad_norm": 0.030264202505350113,
+ "learning_rate": 0.0006,
+ "loss": 5.975833415985107,
+ "step": 399
+ },
+ {
+ "epoch": 5.558951965065502,
+ "grad_norm": 0.032096318900585175,
+ "learning_rate": 0.0006,
+ "loss": 6.001348972320557,
+ "step": 400
+ },
+ {
+ "epoch": 5.5729257641921395,
+ "grad_norm": 0.030579356476664543,
+ "learning_rate": 0.0006,
+ "loss": 5.928768634796143,
+ "step": 401
+ },
+ {
+ "epoch": 5.586899563318777,
+ "grad_norm": 0.028241973370313644,
+ "learning_rate": 0.0006,
+ "loss": 5.92582893371582,
+ "step": 402
+ },
+ {
+ "epoch": 5.600873362445415,
+ "grad_norm": 0.023086953908205032,
+ "learning_rate": 0.0006,
+ "loss": 5.892926216125488,
+ "step": 403
+ },
+ {
+ "epoch": 5.614847161572053,
+ "grad_norm": 0.02547992393374443,
+ "learning_rate": 0.0006,
+ "loss": 5.903195381164551,
+ "step": 404
+ },
+ {
+ "epoch": 5.62882096069869,
+ "grad_norm": 0.023089831694960594,
+ "learning_rate": 0.0006,
+ "loss": 5.912033557891846,
+ "step": 405
+ },
+ {
+ "epoch": 5.642794759825327,
+ "grad_norm": 0.025446368381381035,
+ "learning_rate": 0.0006,
+ "loss": 5.926138401031494,
+ "step": 406
+ },
+ {
+ "epoch": 5.656768558951965,
+ "grad_norm": 0.03748747706413269,
+ "learning_rate": 0.0006,
+ "loss": 5.913451194763184,
+ "step": 407
+ },
+ {
+ "epoch": 5.670742358078603,
+ "grad_norm": 0.03493810072541237,
+ "learning_rate": 0.0006,
+ "loss": 5.97633695602417,
+ "step": 408
+ },
+ {
+ "epoch": 5.68471615720524,
+ "grad_norm": 0.023131275549530983,
+ "learning_rate": 0.0006,
+ "loss": 5.899571418762207,
+ "step": 409
+ },
+ {
+ "epoch": 5.698689956331878,
+ "grad_norm": 0.02986014075577259,
+ "learning_rate": 0.0006,
+ "loss": 5.894903182983398,
+ "step": 410
+ },
+ {
+ "epoch": 5.712663755458515,
+ "grad_norm": 0.030171144753694534,
+ "learning_rate": 0.0006,
+ "loss": 5.876595497131348,
+ "step": 411
+ },
+ {
+ "epoch": 5.726637554585153,
+ "grad_norm": 0.029377546161413193,
+ "learning_rate": 0.0006,
+ "loss": 5.911637306213379,
+ "step": 412
+ },
+ {
+ "epoch": 5.74061135371179,
+ "grad_norm": 0.029829490929841995,
+ "learning_rate": 0.0006,
+ "loss": 5.819629669189453,
+ "step": 413
+ },
+ {
+ "epoch": 5.754585152838428,
+ "grad_norm": 0.024139299988746643,
+ "learning_rate": 0.0006,
+ "loss": 5.81728458404541,
+ "step": 414
+ },
+ {
+ "epoch": 5.7685589519650655,
+ "grad_norm": 0.021453432738780975,
+ "learning_rate": 0.0006,
+ "loss": 5.987326145172119,
+ "step": 415
+ },
+ {
+ "epoch": 5.782532751091703,
+ "grad_norm": 0.017942246049642563,
+ "learning_rate": 0.0006,
+ "loss": 5.893008232116699,
+ "step": 416
+ },
+ {
+ "epoch": 5.796506550218341,
+ "grad_norm": 0.019724637269973755,
+ "learning_rate": 0.0006,
+ "loss": 5.912441253662109,
+ "step": 417
+ },
+ {
+ "epoch": 5.810480349344978,
+ "grad_norm": 0.018597912043333054,
+ "learning_rate": 0.0006,
+ "loss": 5.860394477844238,
+ "step": 418
+ },
+ {
+ "epoch": 5.824454148471616,
+ "grad_norm": 0.016232412308454514,
+ "learning_rate": 0.0006,
+ "loss": 5.94991397857666,
+ "step": 419
+ },
+ {
+ "epoch": 5.8384279475982535,
+ "grad_norm": 0.01667204685509205,
+ "learning_rate": 0.0006,
+ "loss": 5.8506364822387695,
+ "step": 420
+ },
+ {
+ "epoch": 5.85240174672489,
+ "grad_norm": 0.014232151210308075,
+ "learning_rate": 0.0006,
+ "loss": 5.928424835205078,
+ "step": 421
+ },
+ {
+ "epoch": 5.866375545851528,
+ "grad_norm": 0.01570476032793522,
+ "learning_rate": 0.0006,
+ "loss": 5.953692436218262,
+ "step": 422
+ },
+ {
+ "epoch": 5.880349344978166,
+ "grad_norm": 0.01586179807782173,
+ "learning_rate": 0.0006,
+ "loss": 5.820014476776123,
+ "step": 423
+ },
+ {
+ "epoch": 5.894323144104804,
+ "grad_norm": 0.014729145914316177,
+ "learning_rate": 0.0006,
+ "loss": 5.860154628753662,
+ "step": 424
+ },
+ {
+ "epoch": 5.908296943231441,
+ "grad_norm": 0.013111459091305733,
+ "learning_rate": 0.0006,
+ "loss": 5.85988187789917,
+ "step": 425
+ },
+ {
+ "epoch": 5.922270742358078,
+ "grad_norm": 0.012937922962009907,
+ "learning_rate": 0.0006,
+ "loss": 5.756265640258789,
+ "step": 426
+ },
+ {
+ "epoch": 5.936244541484716,
+ "grad_norm": 0.01377453189343214,
+ "learning_rate": 0.0006,
+ "loss": 5.8333048820495605,
+ "step": 427
+ },
+ {
+ "epoch": 5.950218340611354,
+ "grad_norm": 0.014054795727133751,
+ "learning_rate": 0.0006,
+ "loss": 5.803333759307861,
+ "step": 428
+ },
+ {
+ "epoch": 5.964192139737992,
+ "grad_norm": 0.01674959622323513,
+ "learning_rate": 0.0006,
+ "loss": 5.8718485832214355,
+ "step": 429
+ },
+ {
+ "epoch": 5.978165938864628,
+ "grad_norm": 0.018733017146587372,
+ "learning_rate": 0.0006,
+ "loss": 5.875979900360107,
+ "step": 430
+ },
+ {
+ "epoch": 5.992139737991266,
+ "grad_norm": 0.02088983915746212,
+ "learning_rate": 0.0006,
+ "loss": 5.763634204864502,
+ "step": 431
+ },
+ {
+ "epoch": 6.0,
+ "grad_norm": 0.026705985888838768,
+ "learning_rate": 0.0006,
+ "loss": 5.6922478675842285,
+ "step": 432
+ },
+ {
+ "epoch": 6.0,
+ "eval_loss": 5.847958087921143,
+ "eval_runtime": 58.5451,
+ "eval_samples_per_second": 41.711,
+ "eval_steps_per_second": 1.315,
+ "step": 432
+ },
+ {
+ "epoch": 6.013973799126638,
+ "grad_norm": 0.03181544691324234,
+ "learning_rate": 0.0006,
+ "loss": 5.786368370056152,
+ "step": 433
+ },
+ {
+ "epoch": 6.0279475982532755,
+ "grad_norm": 0.03238112851977348,
+ "learning_rate": 0.0006,
+ "loss": 5.810310363769531,
+ "step": 434
+ },
+ {
+ "epoch": 6.041921397379912,
+ "grad_norm": 0.03916756808757782,
+ "learning_rate": 0.0006,
+ "loss": 5.7761993408203125,
+ "step": 435
+ },
+ {
+ "epoch": 6.05589519650655,
+ "grad_norm": 0.044009730219841,
+ "learning_rate": 0.0006,
+ "loss": 5.898112773895264,
+ "step": 436
+ },
+ {
+ "epoch": 6.069868995633188,
+ "grad_norm": 0.045235246419906616,
+ "learning_rate": 0.0006,
+ "loss": 5.876371383666992,
+ "step": 437
+ },
+ {
+ "epoch": 6.083842794759826,
+ "grad_norm": 0.050218384712934494,
+ "learning_rate": 0.0006,
+ "loss": 5.805103302001953,
+ "step": 438
+ },
+ {
+ "epoch": 6.097816593886463,
+ "grad_norm": 0.05444490164518356,
+ "learning_rate": 0.0006,
+ "loss": 5.826424598693848,
+ "step": 439
+ },
+ {
+ "epoch": 6.1117903930131,
+ "grad_norm": 0.04890982061624527,
+ "learning_rate": 0.0006,
+ "loss": 5.7735395431518555,
+ "step": 440
+ },
+ {
+ "epoch": 6.125764192139738,
+ "grad_norm": 0.05478639155626297,
+ "learning_rate": 0.0006,
+ "loss": 5.823677062988281,
+ "step": 441
+ },
+ {
+ "epoch": 6.139737991266376,
+ "grad_norm": 0.057562313973903656,
+ "learning_rate": 0.0006,
+ "loss": 5.889334678649902,
+ "step": 442
+ },
+ {
+ "epoch": 6.153711790393013,
+ "grad_norm": 0.06447703391313553,
+ "learning_rate": 0.0006,
+ "loss": 5.84621524810791,
+ "step": 443
+ },
+ {
+ "epoch": 6.1676855895196505,
+ "grad_norm": 0.048861872404813766,
+ "learning_rate": 0.0006,
+ "loss": 5.931595802307129,
+ "step": 444
+ },
+ {
+ "epoch": 6.181659388646288,
+ "grad_norm": 0.05521339178085327,
+ "learning_rate": 0.0006,
+ "loss": 5.8229780197143555,
+ "step": 445
+ },
+ {
+ "epoch": 6.195633187772926,
+ "grad_norm": 0.053666990250349045,
+ "learning_rate": 0.0006,
+ "loss": 5.879191875457764,
+ "step": 446
+ },
+ {
+ "epoch": 6.209606986899563,
+ "grad_norm": 0.0451025515794754,
+ "learning_rate": 0.0006,
+ "loss": 5.801628112792969,
+ "step": 447
+ },
+ {
+ "epoch": 6.223580786026201,
+ "grad_norm": 0.04291655868291855,
+ "learning_rate": 0.0006,
+ "loss": 5.914680480957031,
+ "step": 448
+ },
+ {
+ "epoch": 6.2375545851528384,
+ "grad_norm": 0.035102490335702896,
+ "learning_rate": 0.0006,
+ "loss": 5.843682765960693,
+ "step": 449
+ },
+ {
+ "epoch": 6.251528384279476,
+ "grad_norm": 0.03403995931148529,
+ "learning_rate": 0.0006,
+ "loss": 5.827154159545898,
+ "step": 450
+ },
+ {
+ "epoch": 6.265502183406113,
+ "grad_norm": 0.03444375470280647,
+ "learning_rate": 0.0006,
+ "loss": 5.800136566162109,
+ "step": 451
+ },
+ {
+ "epoch": 6.279475982532751,
+ "grad_norm": 0.03165159001946449,
+ "learning_rate": 0.0006,
+ "loss": 5.906252384185791,
+ "step": 452
+ },
+ {
+ "epoch": 6.293449781659389,
+ "grad_norm": 0.026153093203902245,
+ "learning_rate": 0.0006,
+ "loss": 5.7423200607299805,
+ "step": 453
+ },
+ {
+ "epoch": 6.307423580786026,
+ "grad_norm": 0.024357657879590988,
+ "learning_rate": 0.0006,
+ "loss": 5.701364517211914,
+ "step": 454
+ },
+ {
+ "epoch": 6.321397379912664,
+ "grad_norm": 0.021508049219846725,
+ "learning_rate": 0.0006,
+ "loss": 5.8328094482421875,
+ "step": 455
+ },
+ {
+ "epoch": 6.335371179039301,
+ "grad_norm": 0.017313921824097633,
+ "learning_rate": 0.0006,
+ "loss": 5.792145252227783,
+ "step": 456
+ },
+ {
+ "epoch": 6.349344978165939,
+ "grad_norm": 0.018563397228717804,
+ "learning_rate": 0.0006,
+ "loss": 5.732672214508057,
+ "step": 457
+ },
+ {
+ "epoch": 6.3633187772925766,
+ "grad_norm": 0.016568679362535477,
+ "learning_rate": 0.0006,
+ "loss": 5.81948184967041,
+ "step": 458
+ },
+ {
+ "epoch": 6.377292576419214,
+ "grad_norm": 0.014133770950138569,
+ "learning_rate": 0.0006,
+ "loss": 5.734982490539551,
+ "step": 459
+ },
+ {
+ "epoch": 6.391266375545851,
+ "grad_norm": 0.014114447869360447,
+ "learning_rate": 0.0006,
+ "loss": 5.70073127746582,
+ "step": 460
+ },
+ {
+ "epoch": 6.405240174672489,
+ "grad_norm": 0.012907393276691437,
+ "learning_rate": 0.0006,
+ "loss": 5.758626937866211,
+ "step": 461
+ },
+ {
+ "epoch": 6.419213973799127,
+ "grad_norm": 0.012673444114625454,
+ "learning_rate": 0.0006,
+ "loss": 5.775138854980469,
+ "step": 462
+ },
+ {
+ "epoch": 6.4331877729257645,
+ "grad_norm": 0.01311410591006279,
+ "learning_rate": 0.0006,
+ "loss": 5.70920991897583,
+ "step": 463
+ },
+ {
+ "epoch": 6.447161572052401,
+ "grad_norm": 0.012935544364154339,
+ "learning_rate": 0.0006,
+ "loss": 5.781479835510254,
+ "step": 464
+ },
+ {
+ "epoch": 6.461135371179039,
+ "grad_norm": 0.01439738366752863,
+ "learning_rate": 0.0006,
+ "loss": 5.691315650939941,
+ "step": 465
+ },
+ {
+ "epoch": 6.475109170305677,
+ "grad_norm": 0.01205186452716589,
+ "learning_rate": 0.0006,
+ "loss": 5.662499904632568,
+ "step": 466
+ },
+ {
+ "epoch": 6.489082969432315,
+ "grad_norm": 0.011575652286410332,
+ "learning_rate": 0.0006,
+ "loss": 5.6271772384643555,
+ "step": 467
+ },
+ {
+ "epoch": 6.503056768558952,
+ "grad_norm": 0.011672502383589745,
+ "learning_rate": 0.0006,
+ "loss": 5.761662006378174,
+ "step": 468
+ },
+ {
+ "epoch": 6.517030567685589,
+ "grad_norm": 0.011541751213371754,
+ "learning_rate": 0.0006,
+ "loss": 5.762078285217285,
+ "step": 469
+ },
+ {
+ "epoch": 6.531004366812227,
+ "grad_norm": 0.011396365240216255,
+ "learning_rate": 0.0006,
+ "loss": 5.67873477935791,
+ "step": 470
+ },
+ {
+ "epoch": 6.544978165938865,
+ "grad_norm": 0.010688499547541142,
+ "learning_rate": 0.0006,
+ "loss": 5.717051982879639,
+ "step": 471
+ },
+ {
+ "epoch": 6.558951965065502,
+ "grad_norm": 0.012224317528307438,
+ "learning_rate": 0.0006,
+ "loss": 5.707948684692383,
+ "step": 472
+ },
+ {
+ "epoch": 6.5729257641921395,
+ "grad_norm": 0.011856825090944767,
+ "learning_rate": 0.0006,
+ "loss": 5.70878791809082,
+ "step": 473
+ },
+ {
+ "epoch": 6.586899563318777,
+ "grad_norm": 0.01199064590036869,
+ "learning_rate": 0.0006,
+ "loss": 5.708697319030762,
+ "step": 474
+ },
+ {
+ "epoch": 6.600873362445415,
+ "grad_norm": 0.01219446212053299,
+ "learning_rate": 0.0006,
+ "loss": 5.6061577796936035,
+ "step": 475
+ },
+ {
+ "epoch": 6.614847161572053,
+ "grad_norm": 0.013397484086453915,
+ "learning_rate": 0.0006,
+ "loss": 5.624789714813232,
+ "step": 476
+ },
+ {
+ "epoch": 6.62882096069869,
+ "grad_norm": 0.01483136136084795,
+ "learning_rate": 0.0006,
+ "loss": 5.721141338348389,
+ "step": 477
+ },
+ {
+ "epoch": 6.642794759825327,
+ "grad_norm": 0.0194488987326622,
+ "learning_rate": 0.0006,
+ "loss": 5.604279518127441,
+ "step": 478
+ },
+ {
+ "epoch": 6.656768558951965,
+ "grad_norm": 0.02079896256327629,
+ "learning_rate": 0.0006,
+ "loss": 5.586322784423828,
+ "step": 479
+ },
+ {
+ "epoch": 6.670742358078603,
+ "grad_norm": 0.018996229395270348,
+ "learning_rate": 0.0006,
+ "loss": 5.5987772941589355,
+ "step": 480
+ },
+ {
+ "epoch": 6.68471615720524,
+ "grad_norm": 0.015779603272676468,
+ "learning_rate": 0.0006,
+ "loss": 5.588602066040039,
+ "step": 481
+ },
+ {
+ "epoch": 6.698689956331878,
+ "grad_norm": 0.015322973020374775,
+ "learning_rate": 0.0006,
+ "loss": 5.686467170715332,
+ "step": 482
+ },
+ {
+ "epoch": 6.712663755458515,
+ "grad_norm": 0.015282213687896729,
+ "learning_rate": 0.0006,
+ "loss": 5.664676666259766,
+ "step": 483
+ },
+ {
+ "epoch": 6.726637554585153,
+ "grad_norm": 0.016589272767305374,
+ "learning_rate": 0.0006,
+ "loss": 5.666739463806152,
+ "step": 484
+ },
+ {
+ "epoch": 6.74061135371179,
+ "grad_norm": 0.01811421848833561,
+ "learning_rate": 0.0006,
+ "loss": 5.647593021392822,
+ "step": 485
+ },
+ {
+ "epoch": 6.754585152838428,
+ "grad_norm": 0.019652029499411583,
+ "learning_rate": 0.0006,
+ "loss": 5.628726959228516,
+ "step": 486
+ },
+ {
+ "epoch": 6.7685589519650655,
+ "grad_norm": 0.02118665538728237,
+ "learning_rate": 0.0006,
+ "loss": 5.6580705642700195,
+ "step": 487
+ },
+ {
+ "epoch": 6.782532751091703,
+ "grad_norm": 0.02237832546234131,
+ "learning_rate": 0.0006,
+ "loss": 5.635931968688965,
+ "step": 488
+ },
+ {
+ "epoch": 6.796506550218341,
+ "grad_norm": 0.023897631093859673,
+ "learning_rate": 0.0006,
+ "loss": 5.517949104309082,
+ "step": 489
+ },
+ {
+ "epoch": 6.810480349344978,
+ "grad_norm": 0.02442227490246296,
+ "learning_rate": 0.0006,
+ "loss": 5.600021839141846,
+ "step": 490
+ },
+ {
+ "epoch": 6.824454148471616,
+ "grad_norm": 0.024675287306308746,
+ "learning_rate": 0.0006,
+ "loss": 5.6639909744262695,
+ "step": 491
+ },
+ {
+ "epoch": 6.8384279475982535,
+ "grad_norm": 0.030372392386198044,
+ "learning_rate": 0.0006,
+ "loss": 5.582888603210449,
+ "step": 492
+ },
+ {
+ "epoch": 6.85240174672489,
+ "grad_norm": 0.03237690031528473,
+ "learning_rate": 0.0006,
+ "loss": 5.573896884918213,
+ "step": 493
+ },
+ {
+ "epoch": 6.866375545851528,
+ "grad_norm": 0.03511589393019676,
+ "learning_rate": 0.0006,
+ "loss": 5.565878868103027,
+ "step": 494
+ },
+ {
+ "epoch": 6.880349344978166,
+ "grad_norm": 0.03883938118815422,
+ "learning_rate": 0.0006,
+ "loss": 5.58540678024292,
+ "step": 495
+ },
+ {
+ "epoch": 6.894323144104804,
+ "grad_norm": 0.04175502806901932,
+ "learning_rate": 0.0006,
+ "loss": 5.6095194816589355,
+ "step": 496
+ },
+ {
+ "epoch": 6.908296943231441,
+ "grad_norm": 0.04013441503047943,
+ "learning_rate": 0.0006,
+ "loss": 5.654401779174805,
+ "step": 497
+ },
+ {
+ "epoch": 6.922270742358078,
+ "grad_norm": 0.04618770629167557,
+ "learning_rate": 0.0006,
+ "loss": 5.6483473777771,
+ "step": 498
+ },
+ {
+ "epoch": 6.936244541484716,
+ "grad_norm": 0.04507308453321457,
+ "learning_rate": 0.0006,
+ "loss": 5.5856218338012695,
+ "step": 499
+ },
+ {
+ "epoch": 6.950218340611354,
+ "grad_norm": 0.03151383996009827,
+ "learning_rate": 0.0006,
+ "loss": 5.594086647033691,
+ "step": 500
+ },
+ {
+ "epoch": 6.964192139737992,
+ "grad_norm": 0.026883797720074654,
+ "learning_rate": 0.0006,
+ "loss": 5.607676029205322,
+ "step": 501
+ },
+ {
+ "epoch": 6.978165938864628,
+ "grad_norm": 0.02981836162507534,
+ "learning_rate": 0.0006,
+ "loss": 5.563666820526123,
+ "step": 502
+ },
+ {
+ "epoch": 6.992139737991266,
+ "grad_norm": 0.03233477845788002,
+ "learning_rate": 0.0006,
+ "loss": 5.667543411254883,
+ "step": 503
+ },
+ {
+ "epoch": 7.0,
+ "grad_norm": 0.029927456751465797,
+ "learning_rate": 0.0006,
+ "loss": 5.733482837677002,
+ "step": 504
+ },
+ {
+ "epoch": 7.0,
+ "eval_loss": 5.679076671600342,
+ "eval_runtime": 59.5965,
+ "eval_samples_per_second": 40.976,
+ "eval_steps_per_second": 1.292,
+ "step": 504
+ },
+ {
+ "epoch": 7.013973799126638,
+ "grad_norm": 0.027081597596406937,
+ "learning_rate": 0.0006,
+ "loss": 5.634487152099609,
+ "step": 505
+ },
+ {
+ "epoch": 7.0279475982532755,
+ "grad_norm": 0.031959421932697296,
+ "learning_rate": 0.0006,
+ "loss": 5.643294811248779,
+ "step": 506
+ },
+ {
+ "epoch": 7.041921397379912,
+ "grad_norm": 0.02802063524723053,
+ "learning_rate": 0.0006,
+ "loss": 5.573239326477051,
+ "step": 507
+ },
+ {
+ "epoch": 7.05589519650655,
+ "grad_norm": 0.032579705119132996,
+ "learning_rate": 0.0006,
+ "loss": 5.627150058746338,
+ "step": 508
+ },
+ {
+ "epoch": 7.069868995633188,
+ "grad_norm": 0.02503928542137146,
+ "learning_rate": 0.0006,
+ "loss": 5.6045308113098145,
+ "step": 509
+ },
+ {
+ "epoch": 7.083842794759826,
+ "grad_norm": 0.02421317622065544,
+ "learning_rate": 0.0006,
+ "loss": 5.522153377532959,
+ "step": 510
+ },
+ {
+ "epoch": 7.097816593886463,
+ "grad_norm": 0.021983426064252853,
+ "learning_rate": 0.0006,
+ "loss": 5.632939338684082,
+ "step": 511
+ },
+ {
+ "epoch": 7.1117903930131,
+ "grad_norm": 0.021933183073997498,
+ "learning_rate": 0.0006,
+ "loss": 5.592185974121094,
+ "step": 512
+ },
+ {
+ "epoch": 7.125764192139738,
+ "grad_norm": 0.02287031151354313,
+ "learning_rate": 0.0006,
+ "loss": 5.613700866699219,
+ "step": 513
+ },
+ {
+ "epoch": 7.139737991266376,
+ "grad_norm": 0.02281602844595909,
+ "learning_rate": 0.0006,
+ "loss": 5.551383018493652,
+ "step": 514
+ },
+ {
+ "epoch": 7.153711790393013,
+ "grad_norm": 0.021050360053777695,
+ "learning_rate": 0.0006,
+ "loss": 5.55275821685791,
+ "step": 515
+ },
+ {
+ "epoch": 7.1676855895196505,
+ "grad_norm": 0.018299926072359085,
+ "learning_rate": 0.0006,
+ "loss": 5.528225421905518,
+ "step": 516
+ },
+ {
+ "epoch": 7.181659388646288,
+ "grad_norm": 0.02024853602051735,
+ "learning_rate": 0.0006,
+ "loss": 5.522453784942627,
+ "step": 517
+ },
+ {
+ "epoch": 7.195633187772926,
+ "grad_norm": 0.02135239914059639,
+ "learning_rate": 0.0006,
+ "loss": 5.527522563934326,
+ "step": 518
+ },
+ {
+ "epoch": 7.209606986899563,
+ "grad_norm": 0.023394184187054634,
+ "learning_rate": 0.0006,
+ "loss": 5.4535017013549805,
+ "step": 519
+ },
+ {
+ "epoch": 7.223580786026201,
+ "grad_norm": 0.025952080264687538,
+ "learning_rate": 0.0006,
+ "loss": 5.5374674797058105,
+ "step": 520
+ },
+ {
+ "epoch": 7.2375545851528384,
+ "grad_norm": 0.022287718951702118,
+ "learning_rate": 0.0006,
+ "loss": 5.493753433227539,
+ "step": 521
+ },
+ {
+ "epoch": 7.251528384279476,
+ "grad_norm": 0.02018447406589985,
+ "learning_rate": 0.0006,
+ "loss": 5.510575294494629,
+ "step": 522
+ },
+ {
+ "epoch": 7.265502183406113,
+ "grad_norm": 0.02050507813692093,
+ "learning_rate": 0.0006,
+ "loss": 5.466026782989502,
+ "step": 523
+ },
+ {
+ "epoch": 7.279475982532751,
+ "grad_norm": 0.023688802495598793,
+ "learning_rate": 0.0006,
+ "loss": 5.485815525054932,
+ "step": 524
+ },
+ {
+ "epoch": 7.293449781659389,
+ "grad_norm": 0.021078843623399734,
+ "learning_rate": 0.0006,
+ "loss": 5.415581703186035,
+ "step": 525
+ },
+ {
+ "epoch": 7.307423580786026,
+ "grad_norm": 0.018266011029481888,
+ "learning_rate": 0.0006,
+ "loss": 5.402815818786621,
+ "step": 526
+ },
+ {
+ "epoch": 7.321397379912664,
+ "grad_norm": 0.019693493843078613,
+ "learning_rate": 0.0006,
+ "loss": 5.504674911499023,
+ "step": 527
+ },
+ {
+ "epoch": 7.335371179039301,
+ "grad_norm": 0.02031373605132103,
+ "learning_rate": 0.0006,
+ "loss": 5.503837585449219,
+ "step": 528
+ },
+ {
+ "epoch": 7.349344978165939,
+ "grad_norm": 0.017636626958847046,
+ "learning_rate": 0.0006,
+ "loss": 5.530580997467041,
+ "step": 529
+ },
+ {
+ "epoch": 7.3633187772925766,
+ "grad_norm": 0.01787244901061058,
+ "learning_rate": 0.0006,
+ "loss": 5.450218677520752,
+ "step": 530
+ },
+ {
+ "epoch": 7.377292576419214,
+ "grad_norm": 0.0170669574290514,
+ "learning_rate": 0.0006,
+ "loss": 5.481570243835449,
+ "step": 531
+ },
+ {
+ "epoch": 7.391266375545851,
+ "grad_norm": 0.01802165061235428,
+ "learning_rate": 0.0006,
+ "loss": 5.386394500732422,
+ "step": 532
+ },
+ {
+ "epoch": 7.405240174672489,
+ "grad_norm": 0.022949472069740295,
+ "learning_rate": 0.0006,
+ "loss": 5.424929618835449,
+ "step": 533
+ },
+ {
+ "epoch": 7.419213973799127,
+ "grad_norm": 0.034213390201330185,
+ "learning_rate": 0.0006,
+ "loss": 5.346663475036621,
+ "step": 534
+ },
+ {
+ "epoch": 7.4331877729257645,
+ "grad_norm": 0.03847593814134598,
+ "learning_rate": 0.0006,
+ "loss": 5.418482780456543,
+ "step": 535
+ },
+ {
+ "epoch": 7.447161572052401,
+ "grad_norm": 0.029393676668405533,
+ "learning_rate": 0.0006,
+ "loss": 5.456090927124023,
+ "step": 536
+ },
+ {
+ "epoch": 7.461135371179039,
+ "grad_norm": 0.03388667106628418,
+ "learning_rate": 0.0006,
+ "loss": 5.572983741760254,
+ "step": 537
+ },
+ {
+ "epoch": 7.475109170305677,
+ "grad_norm": 0.039690613746643066,
+ "learning_rate": 0.0006,
+ "loss": 5.414067268371582,
+ "step": 538
+ },
+ {
+ "epoch": 7.489082969432315,
+ "grad_norm": 0.03630776330828667,
+ "learning_rate": 0.0006,
+ "loss": 5.523739814758301,
+ "step": 539
+ },
+ {
+ "epoch": 7.503056768558952,
+ "grad_norm": 0.03356870263814926,
+ "learning_rate": 0.0006,
+ "loss": 5.444767951965332,
+ "step": 540
+ },
+ {
+ "epoch": 7.517030567685589,
+ "grad_norm": 0.030812304466962814,
+ "learning_rate": 0.0006,
+ "loss": 5.461244106292725,
+ "step": 541
+ },
+ {
+ "epoch": 7.531004366812227,
+ "grad_norm": 0.03319217637181282,
+ "learning_rate": 0.0006,
+ "loss": 5.502161026000977,
+ "step": 542
+ },
+ {
+ "epoch": 7.544978165938865,
+ "grad_norm": 0.032997481524944305,
+ "learning_rate": 0.0006,
+ "loss": 5.4162278175354,
+ "step": 543
+ },
+ {
+ "epoch": 7.558951965065502,
+ "grad_norm": 0.03364962339401245,
+ "learning_rate": 0.0006,
+ "loss": 5.415736198425293,
+ "step": 544
+ },
+ {
+ "epoch": 7.5729257641921395,
+ "grad_norm": 0.0344221405684948,
+ "learning_rate": 0.0006,
+ "loss": 5.541967868804932,
+ "step": 545
+ },
+ {
+ "epoch": 7.586899563318777,
+ "grad_norm": 0.029609503224492073,
+ "learning_rate": 0.0006,
+ "loss": 5.440614223480225,
+ "step": 546
+ },
+ {
+ "epoch": 7.600873362445415,
+ "grad_norm": 0.029055926948785782,
+ "learning_rate": 0.0006,
+ "loss": 5.462865352630615,
+ "step": 547
+ },
+ {
+ "epoch": 7.614847161572053,
+ "grad_norm": 0.02658848837018013,
+ "learning_rate": 0.0006,
+ "loss": 5.444467544555664,
+ "step": 548
+ },
+ {
+ "epoch": 7.62882096069869,
+ "grad_norm": 0.026276404038071632,
+ "learning_rate": 0.0006,
+ "loss": 5.462278842926025,
+ "step": 549
+ },
+ {
+ "epoch": 7.642794759825327,
+ "grad_norm": 0.0286889486014843,
+ "learning_rate": 0.0006,
+ "loss": 5.4849534034729,
+ "step": 550
+ },
+ {
+ "epoch": 7.656768558951965,
+ "grad_norm": 0.026849817484617233,
+ "learning_rate": 0.0006,
+ "loss": 5.431473731994629,
+ "step": 551
+ },
+ {
+ "epoch": 7.670742358078603,
+ "grad_norm": 0.02312396466732025,
+ "learning_rate": 0.0006,
+ "loss": 5.45462703704834,
+ "step": 552
+ },
+ {
+ "epoch": 7.68471615720524,
+ "grad_norm": 0.026197485625743866,
+ "learning_rate": 0.0006,
+ "loss": 5.452552318572998,
+ "step": 553
+ },
+ {
+ "epoch": 7.698689956331878,
+ "grad_norm": 0.02747255191206932,
+ "learning_rate": 0.0006,
+ "loss": 5.447169780731201,
+ "step": 554
+ },
+ {
+ "epoch": 7.712663755458515,
+ "grad_norm": 0.028123432770371437,
+ "learning_rate": 0.0006,
+ "loss": 5.394400596618652,
+ "step": 555
+ },
+ {
+ "epoch": 7.726637554585153,
+ "grad_norm": 0.02599870041012764,
+ "learning_rate": 0.0006,
+ "loss": 5.398341178894043,
+ "step": 556
+ },
+ {
+ "epoch": 7.74061135371179,
+ "grad_norm": 0.022171657532453537,
+ "learning_rate": 0.0006,
+ "loss": 5.368820667266846,
+ "step": 557
+ },
+ {
+ "epoch": 7.754585152838428,
+ "grad_norm": 0.022309480234980583,
+ "learning_rate": 0.0006,
+ "loss": 5.416772365570068,
+ "step": 558
+ },
+ {
+ "epoch": 7.7685589519650655,
+ "grad_norm": 0.024074165150523186,
+ "learning_rate": 0.0006,
+ "loss": 5.359002113342285,
+ "step": 559
+ },
+ {
+ "epoch": 7.782532751091703,
+ "grad_norm": 0.02636653557419777,
+ "learning_rate": 0.0006,
+ "loss": 5.484368324279785,
+ "step": 560
+ },
+ {
+ "epoch": 7.796506550218341,
+ "grad_norm": 0.02203752100467682,
+ "learning_rate": 0.0006,
+ "loss": 5.375498294830322,
+ "step": 561
+ },
+ {
+ "epoch": 7.810480349344978,
+ "grad_norm": 0.019924819469451904,
+ "learning_rate": 0.0006,
+ "loss": 5.334723472595215,
+ "step": 562
+ },
+ {
+ "epoch": 7.824454148471616,
+ "grad_norm": 0.018755966797471046,
+ "learning_rate": 0.0006,
+ "loss": 5.348781108856201,
+ "step": 563
+ },
+ {
+ "epoch": 7.8384279475982535,
+ "grad_norm": 0.016747845336794853,
+ "learning_rate": 0.0006,
+ "loss": 5.433079719543457,
+ "step": 564
+ },
+ {
+ "epoch": 7.85240174672489,
+ "grad_norm": 0.0170395877212286,
+ "learning_rate": 0.0006,
+ "loss": 5.346794128417969,
+ "step": 565
+ },
+ {
+ "epoch": 7.866375545851528,
+ "grad_norm": 0.016074176877737045,
+ "learning_rate": 0.0006,
+ "loss": 5.3277740478515625,
+ "step": 566
+ },
+ {
+ "epoch": 7.880349344978166,
+ "grad_norm": 0.014119806699454784,
+ "learning_rate": 0.0006,
+ "loss": 5.483551979064941,
+ "step": 567
+ },
+ {
+ "epoch": 7.894323144104804,
+ "grad_norm": 0.015270394273102283,
+ "learning_rate": 0.0006,
+ "loss": 5.378519535064697,
+ "step": 568
+ },
+ {
+ "epoch": 7.908296943231441,
+ "grad_norm": 0.015677539631724358,
+ "learning_rate": 0.0006,
+ "loss": 5.290210247039795,
+ "step": 569
+ },
+ {
+ "epoch": 7.922270742358078,
+ "grad_norm": 0.015930423513054848,
+ "learning_rate": 0.0006,
+ "loss": 5.427360534667969,
+ "step": 570
+ },
+ {
+ "epoch": 7.936244541484716,
+ "grad_norm": 0.016641564667224884,
+ "learning_rate": 0.0006,
+ "loss": 5.301599979400635,
+ "step": 571
+ },
+ {
+ "epoch": 7.950218340611354,
+ "grad_norm": 0.018293552100658417,
+ "learning_rate": 0.0006,
+ "loss": 5.277889251708984,
+ "step": 572
+ },
+ {
+ "epoch": 7.964192139737992,
+ "grad_norm": 0.017618799582123756,
+ "learning_rate": 0.0006,
+ "loss": 5.297194480895996,
+ "step": 573
+ },
+ {
+ "epoch": 7.978165938864628,
+ "grad_norm": 0.015034242533147335,
+ "learning_rate": 0.0006,
+ "loss": 5.42428731918335,
+ "step": 574
+ },
+ {
+ "epoch": 7.992139737991266,
+ "grad_norm": 0.01617511734366417,
+ "learning_rate": 0.0006,
+ "loss": 5.420044898986816,
+ "step": 575
+ },
+ {
+ "epoch": 8.0,
+ "grad_norm": 0.01837257295846939,
+ "learning_rate": 0.0006,
+ "loss": 5.240387916564941,
+ "step": 576
+ },
+ {
+ "epoch": 8.0,
+ "eval_loss": 5.423932075500488,
+ "eval_runtime": 58.0841,
+ "eval_samples_per_second": 42.042,
+ "eval_steps_per_second": 1.326,
+ "step": 576
+ },
+ {
+ "epoch": 8.013973799126637,
+ "grad_norm": 0.020953809842467308,
+ "learning_rate": 0.0006,
+ "loss": 5.4320783615112305,
+ "step": 577
+ },
+ {
+ "epoch": 8.027947598253276,
+ "grad_norm": 0.0227745920419693,
+ "learning_rate": 0.0006,
+ "loss": 5.310197830200195,
+ "step": 578
+ },
+ {
+ "epoch": 8.041921397379912,
+ "grad_norm": 0.021372603252530098,
+ "learning_rate": 0.0006,
+ "loss": 5.339468002319336,
+ "step": 579
+ },
+ {
+ "epoch": 8.055895196506551,
+ "grad_norm": 0.022401731461286545,
+ "learning_rate": 0.0006,
+ "loss": 5.279829978942871,
+ "step": 580
+ },
+ {
+ "epoch": 8.069868995633188,
+ "grad_norm": 0.023942379280924797,
+ "learning_rate": 0.0006,
+ "loss": 5.372323989868164,
+ "step": 581
+ },
+ {
+ "epoch": 8.083842794759825,
+ "grad_norm": 0.024380534887313843,
+ "learning_rate": 0.0006,
+ "loss": 5.253486633300781,
+ "step": 582
+ },
+ {
+ "epoch": 8.097816593886463,
+ "grad_norm": 0.03246625140309334,
+ "learning_rate": 0.0006,
+ "loss": 5.305774211883545,
+ "step": 583
+ },
+ {
+ "epoch": 8.1117903930131,
+ "grad_norm": 0.030399736016988754,
+ "learning_rate": 0.0006,
+ "loss": 5.370976448059082,
+ "step": 584
+ },
+ {
+ "epoch": 8.125764192139737,
+ "grad_norm": 0.027148913592100143,
+ "learning_rate": 0.0006,
+ "loss": 5.368010520935059,
+ "step": 585
+ },
+ {
+ "epoch": 8.139737991266376,
+ "grad_norm": 0.02925540879368782,
+ "learning_rate": 0.0006,
+ "loss": 5.303013801574707,
+ "step": 586
+ },
+ {
+ "epoch": 8.153711790393013,
+ "grad_norm": 0.02915453538298607,
+ "learning_rate": 0.0006,
+ "loss": 5.329667091369629,
+ "step": 587
+ },
+ {
+ "epoch": 8.167685589519651,
+ "grad_norm": 0.031732626259326935,
+ "learning_rate": 0.0006,
+ "loss": 5.3634934425354,
+ "step": 588
+ },
+ {
+ "epoch": 8.181659388646288,
+ "grad_norm": 0.036166731268167496,
+ "learning_rate": 0.0006,
+ "loss": 5.223980903625488,
+ "step": 589
+ },
+ {
+ "epoch": 8.195633187772925,
+ "grad_norm": 0.035654108971357346,
+ "learning_rate": 0.0006,
+ "loss": 5.361059188842773,
+ "step": 590
+ },
+ {
+ "epoch": 8.209606986899564,
+ "grad_norm": 0.0365324430167675,
+ "learning_rate": 0.0006,
+ "loss": 5.336151599884033,
+ "step": 591
+ },
+ {
+ "epoch": 8.2235807860262,
+ "grad_norm": 0.03578880429267883,
+ "learning_rate": 0.0006,
+ "loss": 5.434076309204102,
+ "step": 592
+ },
+ {
+ "epoch": 8.237554585152838,
+ "grad_norm": 0.035734012722969055,
+ "learning_rate": 0.0006,
+ "loss": 5.330893516540527,
+ "step": 593
+ },
+ {
+ "epoch": 8.251528384279476,
+ "grad_norm": 0.03196857124567032,
+ "learning_rate": 0.0006,
+ "loss": 5.306009769439697,
+ "step": 594
+ },
+ {
+ "epoch": 8.265502183406113,
+ "grad_norm": 0.032042086124420166,
+ "learning_rate": 0.0006,
+ "loss": 5.343267917633057,
+ "step": 595
+ },
+ {
+ "epoch": 8.279475982532752,
+ "grad_norm": 0.03160746395587921,
+ "learning_rate": 0.0006,
+ "loss": 5.2353715896606445,
+ "step": 596
+ },
+ {
+ "epoch": 8.293449781659389,
+ "grad_norm": 0.026689818128943443,
+ "learning_rate": 0.0006,
+ "loss": 5.254042625427246,
+ "step": 597
+ },
+ {
+ "epoch": 8.307423580786025,
+ "grad_norm": 0.02880188450217247,
+ "learning_rate": 0.0006,
+ "loss": 5.326833248138428,
+ "step": 598
+ },
+ {
+ "epoch": 8.321397379912664,
+ "grad_norm": 0.027516381815075874,
+ "learning_rate": 0.0006,
+ "loss": 5.342336177825928,
+ "step": 599
+ },
+ {
+ "epoch": 8.335371179039301,
+ "grad_norm": 0.027875030413269997,
+ "learning_rate": 0.0006,
+ "loss": 5.200719833374023,
+ "step": 600
+ },
+ {
+ "epoch": 8.34934497816594,
+ "grad_norm": 0.0268265288323164,
+ "learning_rate": 0.0006,
+ "loss": 5.249449729919434,
+ "step": 601
+ },
+ {
+ "epoch": 8.363318777292577,
+ "grad_norm": 0.024824608117341995,
+ "learning_rate": 0.0006,
+ "loss": 5.318952560424805,
+ "step": 602
+ },
+ {
+ "epoch": 8.377292576419213,
+ "grad_norm": 0.022990627214312553,
+ "learning_rate": 0.0006,
+ "loss": 5.244993209838867,
+ "step": 603
+ },
+ {
+ "epoch": 8.391266375545852,
+ "grad_norm": 0.022804604843258858,
+ "learning_rate": 0.0006,
+ "loss": 5.3236846923828125,
+ "step": 604
+ },
+ {
+ "epoch": 8.405240174672489,
+ "grad_norm": 0.02419872209429741,
+ "learning_rate": 0.0006,
+ "loss": 5.255486488342285,
+ "step": 605
+ },
+ {
+ "epoch": 8.419213973799126,
+ "grad_norm": 0.021952059119939804,
+ "learning_rate": 0.0006,
+ "loss": 5.3511552810668945,
+ "step": 606
+ },
+ {
+ "epoch": 8.433187772925764,
+ "grad_norm": 0.022375497967004776,
+ "learning_rate": 0.0006,
+ "loss": 5.294790267944336,
+ "step": 607
+ },
+ {
+ "epoch": 8.447161572052401,
+ "grad_norm": 0.019242815673351288,
+ "learning_rate": 0.0006,
+ "loss": 5.252618789672852,
+ "step": 608
+ },
+ {
+ "epoch": 8.46113537117904,
+ "grad_norm": 0.0173486340790987,
+ "learning_rate": 0.0006,
+ "loss": 5.30147647857666,
+ "step": 609
+ },
+ {
+ "epoch": 8.475109170305677,
+ "grad_norm": 0.01612604409456253,
+ "learning_rate": 0.0006,
+ "loss": 5.23960542678833,
+ "step": 610
+ },
+ {
+ "epoch": 8.489082969432314,
+ "grad_norm": 0.014293976128101349,
+ "learning_rate": 0.0006,
+ "loss": 5.119932651519775,
+ "step": 611
+ },
+ {
+ "epoch": 8.503056768558952,
+ "grad_norm": 0.014453536830842495,
+ "learning_rate": 0.0006,
+ "loss": 5.166794776916504,
+ "step": 612
+ },
+ {
+ "epoch": 8.51703056768559,
+ "grad_norm": 0.014685769565403461,
+ "learning_rate": 0.0006,
+ "loss": 5.264822959899902,
+ "step": 613
+ },
+ {
+ "epoch": 8.531004366812226,
+ "grad_norm": 0.01427740603685379,
+ "learning_rate": 0.0006,
+ "loss": 5.278433799743652,
+ "step": 614
+ },
+ {
+ "epoch": 8.544978165938865,
+ "grad_norm": 0.014864951372146606,
+ "learning_rate": 0.0006,
+ "loss": 5.214512348175049,
+ "step": 615
+ },
+ {
+ "epoch": 8.558951965065502,
+ "grad_norm": 0.01565164513885975,
+ "learning_rate": 0.0006,
+ "loss": 5.183066368103027,
+ "step": 616
+ },
+ {
+ "epoch": 8.57292576419214,
+ "grad_norm": 0.01779816672205925,
+ "learning_rate": 0.0006,
+ "loss": 5.163185119628906,
+ "step": 617
+ },
+ {
+ "epoch": 8.586899563318777,
+ "grad_norm": 0.017254827544093132,
+ "learning_rate": 0.0006,
+ "loss": 5.243555068969727,
+ "step": 618
+ },
+ {
+ "epoch": 8.600873362445414,
+ "grad_norm": 0.01650584116578102,
+ "learning_rate": 0.0006,
+ "loss": 5.234884738922119,
+ "step": 619
+ },
+ {
+ "epoch": 8.614847161572053,
+ "grad_norm": 0.017021115869283676,
+ "learning_rate": 0.0006,
+ "loss": 5.12271785736084,
+ "step": 620
+ },
+ {
+ "epoch": 8.62882096069869,
+ "grad_norm": 0.01773759163916111,
+ "learning_rate": 0.0006,
+ "loss": 5.242153167724609,
+ "step": 621
+ },
+ {
+ "epoch": 8.642794759825328,
+ "grad_norm": 0.015679042786359787,
+ "learning_rate": 0.0006,
+ "loss": 5.170779228210449,
+ "step": 622
+ },
+ {
+ "epoch": 8.656768558951965,
+ "grad_norm": 0.013760549947619438,
+ "learning_rate": 0.0006,
+ "loss": 5.238644599914551,
+ "step": 623
+ },
+ {
+ "epoch": 8.670742358078602,
+ "grad_norm": 0.014571522362530231,
+ "learning_rate": 0.0006,
+ "loss": 5.177056312561035,
+ "step": 624
+ },
+ {
+ "epoch": 8.68471615720524,
+ "grad_norm": 0.016209015622735023,
+ "learning_rate": 0.0006,
+ "loss": 5.163750648498535,
+ "step": 625
+ },
+ {
+ "epoch": 8.698689956331878,
+ "grad_norm": 0.016813941299915314,
+ "learning_rate": 0.0006,
+ "loss": 5.183428764343262,
+ "step": 626
+ },
+ {
+ "epoch": 8.712663755458514,
+ "grad_norm": 0.020985357463359833,
+ "learning_rate": 0.0006,
+ "loss": 5.229465007781982,
+ "step": 627
+ },
+ {
+ "epoch": 8.726637554585153,
+ "grad_norm": 0.02679632417857647,
+ "learning_rate": 0.0006,
+ "loss": 5.257368087768555,
+ "step": 628
+ },
+ {
+ "epoch": 8.74061135371179,
+ "grad_norm": 0.025756070390343666,
+ "learning_rate": 0.0006,
+ "loss": 5.253736972808838,
+ "step": 629
+ },
+ {
+ "epoch": 8.754585152838429,
+ "grad_norm": 0.02643490768969059,
+ "learning_rate": 0.0006,
+ "loss": 5.228633403778076,
+ "step": 630
+ },
+ {
+ "epoch": 8.768558951965066,
+ "grad_norm": 0.03303210437297821,
+ "learning_rate": 0.0006,
+ "loss": 5.163540363311768,
+ "step": 631
+ },
+ {
+ "epoch": 8.782532751091702,
+ "grad_norm": 0.030432431027293205,
+ "learning_rate": 0.0006,
+ "loss": 5.243169784545898,
+ "step": 632
+ },
+ {
+ "epoch": 8.796506550218341,
+ "grad_norm": 0.028631288558244705,
+ "learning_rate": 0.0006,
+ "loss": 5.166398048400879,
+ "step": 633
+ },
+ {
+ "epoch": 8.810480349344978,
+ "grad_norm": 0.026188310235738754,
+ "learning_rate": 0.0006,
+ "loss": 5.128122329711914,
+ "step": 634
+ },
+ {
+ "epoch": 8.824454148471617,
+ "grad_norm": 0.025308528915047646,
+ "learning_rate": 0.0006,
+ "loss": 5.211284637451172,
+ "step": 635
+ },
+ {
+ "epoch": 8.838427947598253,
+ "grad_norm": 0.020657729357481003,
+ "learning_rate": 0.0006,
+ "loss": 5.161575794219971,
+ "step": 636
+ },
+ {
+ "epoch": 8.85240174672489,
+ "grad_norm": 0.021901234984397888,
+ "learning_rate": 0.0006,
+ "loss": 5.20050048828125,
+ "step": 637
+ },
+ {
+ "epoch": 8.866375545851529,
+ "grad_norm": 0.021219318732619286,
+ "learning_rate": 0.0006,
+ "loss": 5.211699485778809,
+ "step": 638
+ },
+ {
+ "epoch": 8.880349344978166,
+ "grad_norm": 0.0200974028557539,
+ "learning_rate": 0.0006,
+ "loss": 5.092503547668457,
+ "step": 639
+ },
+ {
+ "epoch": 8.894323144104803,
+ "grad_norm": 0.023804882541298866,
+ "learning_rate": 0.0006,
+ "loss": 5.216068267822266,
+ "step": 640
+ },
+ {
+ "epoch": 8.908296943231441,
+ "grad_norm": 0.026088010519742966,
+ "learning_rate": 0.0006,
+ "loss": 5.155592918395996,
+ "step": 641
+ },
+ {
+ "epoch": 8.922270742358078,
+ "grad_norm": 0.02601276896893978,
+ "learning_rate": 0.0006,
+ "loss": 5.197238922119141,
+ "step": 642
+ },
+ {
+ "epoch": 8.936244541484717,
+ "grad_norm": 0.020387211814522743,
+ "learning_rate": 0.0006,
+ "loss": 5.1955437660217285,
+ "step": 643
+ },
+ {
+ "epoch": 8.950218340611354,
+ "grad_norm": 0.019214622676372528,
+ "learning_rate": 0.0006,
+ "loss": 5.206346035003662,
+ "step": 644
+ },
+ {
+ "epoch": 8.96419213973799,
+ "grad_norm": 0.019674314185976982,
+ "learning_rate": 0.0006,
+ "loss": 5.219857215881348,
+ "step": 645
+ },
+ {
+ "epoch": 8.97816593886463,
+ "grad_norm": 0.020318234339356422,
+ "learning_rate": 0.0006,
+ "loss": 5.191132545471191,
+ "step": 646
+ },
+ {
+ "epoch": 8.992139737991266,
+ "grad_norm": 0.021428676322102547,
+ "learning_rate": 0.0006,
+ "loss": 5.214695930480957,
+ "step": 647
+ },
+ {
+ "epoch": 9.0,
+ "grad_norm": 0.023235054686665535,
+ "learning_rate": 0.0006,
+ "loss": 5.229091644287109,
+ "step": 648
+ },
+ {
+ "epoch": 9.0,
+ "eval_loss": 5.289593696594238,
+ "eval_runtime": 56.9256,
+ "eval_samples_per_second": 42.898,
+ "eval_steps_per_second": 1.353,
+ "step": 648
+ },
+ {
+ "epoch": 9.013973799126637,
+ "grad_norm": 0.02320289984345436,
+ "learning_rate": 0.0006,
+ "loss": 4.983669281005859,
+ "step": 649
+ },
+ {
+ "epoch": 9.027947598253276,
+ "grad_norm": 0.018521282821893692,
+ "learning_rate": 0.0006,
+ "loss": 5.200575828552246,
+ "step": 650
+ },
+ {
+ "epoch": 9.041921397379912,
+ "grad_norm": 0.0180169939994812,
+ "learning_rate": 0.0006,
+ "loss": 5.148033618927002,
+ "step": 651
+ },
+ {
+ "epoch": 9.055895196506551,
+ "grad_norm": 0.019969860091805458,
+ "learning_rate": 0.0006,
+ "loss": 5.099125385284424,
+ "step": 652
+ },
+ {
+ "epoch": 9.069868995633188,
+ "grad_norm": 0.01643305830657482,
+ "learning_rate": 0.0006,
+ "loss": 5.092347145080566,
+ "step": 653
+ },
+ {
+ "epoch": 9.083842794759825,
+ "grad_norm": 0.016022363677620888,
+ "learning_rate": 0.0006,
+ "loss": 5.129279613494873,
+ "step": 654
+ },
+ {
+ "epoch": 9.097816593886463,
+ "grad_norm": 0.01657041162252426,
+ "learning_rate": 0.0006,
+ "loss": 5.159562587738037,
+ "step": 655
+ },
+ {
+ "epoch": 9.1117903930131,
+ "grad_norm": 0.017924228683114052,
+ "learning_rate": 0.0006,
+ "loss": 5.165497779846191,
+ "step": 656
+ },
+ {
+ "epoch": 9.125764192139737,
+ "grad_norm": 0.021592361852526665,
+ "learning_rate": 0.0006,
+ "loss": 5.204775810241699,
+ "step": 657
+ },
+ {
+ "epoch": 9.139737991266376,
+ "grad_norm": 0.026924120262265205,
+ "learning_rate": 0.0006,
+ "loss": 5.128296852111816,
+ "step": 658
+ },
+ {
+ "epoch": 9.153711790393013,
+ "grad_norm": 0.023156899958848953,
+ "learning_rate": 0.0006,
+ "loss": 5.1255340576171875,
+ "step": 659
+ },
+ {
+ "epoch": 9.167685589519651,
+ "grad_norm": 0.020670367404818535,
+ "learning_rate": 0.0006,
+ "loss": 5.128479957580566,
+ "step": 660
+ },
+ {
+ "epoch": 9.181659388646288,
+ "grad_norm": 0.025447173044085503,
+ "learning_rate": 0.0006,
+ "loss": 5.154821872711182,
+ "step": 661
+ },
+ {
+ "epoch": 9.195633187772925,
+ "grad_norm": 0.027940068393945694,
+ "learning_rate": 0.0006,
+ "loss": 5.143466472625732,
+ "step": 662
+ },
+ {
+ "epoch": 9.209606986899564,
+ "grad_norm": 0.026847844943404198,
+ "learning_rate": 0.0006,
+ "loss": 5.045665740966797,
+ "step": 663
+ },
+ {
+ "epoch": 9.2235807860262,
+ "grad_norm": 0.02340601570904255,
+ "learning_rate": 0.0006,
+ "loss": 5.21933126449585,
+ "step": 664
+ },
+ {
+ "epoch": 9.237554585152838,
+ "grad_norm": 0.02340371161699295,
+ "learning_rate": 0.0006,
+ "loss": 4.942949295043945,
+ "step": 665
+ },
+ {
+ "epoch": 9.251528384279476,
+ "grad_norm": 0.02221992425620556,
+ "learning_rate": 0.0006,
+ "loss": 5.197099208831787,
+ "step": 666
+ },
+ {
+ "epoch": 9.265502183406113,
+ "grad_norm": 0.023508677259087563,
+ "learning_rate": 0.0006,
+ "loss": 5.022897720336914,
+ "step": 667
+ },
+ {
+ "epoch": 9.279475982532752,
+ "grad_norm": 0.026120394468307495,
+ "learning_rate": 0.0006,
+ "loss": 5.087725639343262,
+ "step": 668
+ },
+ {
+ "epoch": 9.293449781659389,
+ "grad_norm": 0.027273228392004967,
+ "learning_rate": 0.0006,
+ "loss": 5.090963840484619,
+ "step": 669
+ },
+ {
+ "epoch": 9.307423580786025,
+ "grad_norm": 0.03241586685180664,
+ "learning_rate": 0.0006,
+ "loss": 5.047842979431152,
+ "step": 670
+ },
+ {
+ "epoch": 9.321397379912664,
+ "grad_norm": 0.030245667323470116,
+ "learning_rate": 0.0006,
+ "loss": 5.104803085327148,
+ "step": 671
+ },
+ {
+ "epoch": 9.335371179039301,
+ "grad_norm": 0.027698364108800888,
+ "learning_rate": 0.0006,
+ "loss": 5.0644097328186035,
+ "step": 672
+ },
+ {
+ "epoch": 9.34934497816594,
+ "grad_norm": 0.029692554846405983,
+ "learning_rate": 0.0006,
+ "loss": 5.111942291259766,
+ "step": 673
+ },
+ {
+ "epoch": 9.363318777292577,
+ "grad_norm": 0.03328656405210495,
+ "learning_rate": 0.0006,
+ "loss": 5.1574554443359375,
+ "step": 674
+ },
+ {
+ "epoch": 9.377292576419213,
+ "grad_norm": 0.031938180327415466,
+ "learning_rate": 0.0006,
+ "loss": 5.190371990203857,
+ "step": 675
+ },
+ {
+ "epoch": 9.391266375545852,
+ "grad_norm": 0.024858945980668068,
+ "learning_rate": 0.0006,
+ "loss": 5.121420860290527,
+ "step": 676
+ },
+ {
+ "epoch": 9.405240174672489,
+ "grad_norm": 0.023033203557133675,
+ "learning_rate": 0.0006,
+ "loss": 5.0974345207214355,
+ "step": 677
+ },
+ {
+ "epoch": 9.419213973799126,
+ "grad_norm": 0.022083545103669167,
+ "learning_rate": 0.0006,
+ "loss": 5.046319961547852,
+ "step": 678
+ },
+ {
+ "epoch": 9.433187772925764,
+ "grad_norm": 0.0228437427431345,
+ "learning_rate": 0.0006,
+ "loss": 5.119963645935059,
+ "step": 679
+ },
+ {
+ "epoch": 9.447161572052401,
+ "grad_norm": 0.0242351982742548,
+ "learning_rate": 0.0006,
+ "loss": 5.0444560050964355,
+ "step": 680
+ },
+ {
+ "epoch": 9.46113537117904,
+ "grad_norm": 0.02401185780763626,
+ "learning_rate": 0.0006,
+ "loss": 5.063238620758057,
+ "step": 681
+ },
+ {
+ "epoch": 9.475109170305677,
+ "grad_norm": 0.02560959942638874,
+ "learning_rate": 0.0006,
+ "loss": 5.035224914550781,
+ "step": 682
+ },
+ {
+ "epoch": 9.489082969432314,
+ "grad_norm": 0.024985626339912415,
+ "learning_rate": 0.0006,
+ "loss": 5.049284934997559,
+ "step": 683
+ },
+ {
+ "epoch": 9.503056768558952,
+ "grad_norm": 0.023178115487098694,
+ "learning_rate": 0.0006,
+ "loss": 5.2266621589660645,
+ "step": 684
+ },
+ {
+ "epoch": 9.51703056768559,
+ "grad_norm": 0.0215512253344059,
+ "learning_rate": 0.0006,
+ "loss": 5.1875715255737305,
+ "step": 685
+ },
+ {
+ "epoch": 9.531004366812226,
+ "grad_norm": 0.01662837713956833,
+ "learning_rate": 0.0006,
+ "loss": 5.137012481689453,
+ "step": 686
+ },
+ {
+ "epoch": 9.544978165938865,
+ "grad_norm": 0.01952764391899109,
+ "learning_rate": 0.0006,
+ "loss": 5.072500228881836,
+ "step": 687
+ },
+ {
+ "epoch": 9.558951965065502,
+ "grad_norm": 0.017521077767014503,
+ "learning_rate": 0.0006,
+ "loss": 5.045711994171143,
+ "step": 688
+ },
+ {
+ "epoch": 9.57292576419214,
+ "grad_norm": 0.017355265095829964,
+ "learning_rate": 0.0006,
+ "loss": 5.02881383895874,
+ "step": 689
+ },
+ {
+ "epoch": 9.586899563318777,
+ "grad_norm": 0.015587719157338142,
+ "learning_rate": 0.0006,
+ "loss": 5.132787704467773,
+ "step": 690
+ },
+ {
+ "epoch": 9.600873362445414,
+ "grad_norm": 0.017469845712184906,
+ "learning_rate": 0.0006,
+ "loss": 5.105447292327881,
+ "step": 691
+ },
+ {
+ "epoch": 9.614847161572053,
+ "grad_norm": 0.019630100578069687,
+ "learning_rate": 0.0006,
+ "loss": 5.032464981079102,
+ "step": 692
+ },
+ {
+ "epoch": 9.62882096069869,
+ "grad_norm": 0.01784994639456272,
+ "learning_rate": 0.0006,
+ "loss": 5.081345558166504,
+ "step": 693
+ },
+ {
+ "epoch": 9.642794759825328,
+ "grad_norm": 0.01872754842042923,
+ "learning_rate": 0.0006,
+ "loss": 4.9772114753723145,
+ "step": 694
+ },
+ {
+ "epoch": 9.656768558951965,
+ "grad_norm": 0.01999032311141491,
+ "learning_rate": 0.0006,
+ "loss": 5.089740753173828,
+ "step": 695
+ },
+ {
+ "epoch": 9.670742358078602,
+ "grad_norm": 0.01710767112672329,
+ "learning_rate": 0.0006,
+ "loss": 4.9869890213012695,
+ "step": 696
+ },
+ {
+ "epoch": 9.68471615720524,
+ "grad_norm": 0.015144161880016327,
+ "learning_rate": 0.0006,
+ "loss": 5.022564888000488,
+ "step": 697
+ },
+ {
+ "epoch": 9.698689956331878,
+ "grad_norm": 0.01618543453514576,
+ "learning_rate": 0.0006,
+ "loss": 5.035921573638916,
+ "step": 698
+ },
+ {
+ "epoch": 9.712663755458514,
+ "grad_norm": 0.014653387479484081,
+ "learning_rate": 0.0006,
+ "loss": 4.914456367492676,
+ "step": 699
+ },
+ {
+ "epoch": 9.726637554585153,
+ "grad_norm": 0.013434977270662785,
+ "learning_rate": 0.0006,
+ "loss": 5.063764572143555,
+ "step": 700
+ },
+ {
+ "epoch": 9.74061135371179,
+ "grad_norm": 0.015128864906728268,
+ "learning_rate": 0.0006,
+ "loss": 5.052367687225342,
+ "step": 701
+ },
+ {
+ "epoch": 9.754585152838429,
+ "grad_norm": 0.014210768043994904,
+ "learning_rate": 0.0006,
+ "loss": 5.016489028930664,
+ "step": 702
+ },
+ {
+ "epoch": 9.768558951965066,
+ "grad_norm": 0.015191680751740932,
+ "learning_rate": 0.0006,
+ "loss": 4.990176200866699,
+ "step": 703
+ },
+ {
+ "epoch": 9.782532751091702,
+ "grad_norm": 0.018285127356648445,
+ "learning_rate": 0.0006,
+ "loss": 5.015164375305176,
+ "step": 704
+ },
+ {
+ "epoch": 9.796506550218341,
+ "grad_norm": 0.0204079058021307,
+ "learning_rate": 0.0006,
+ "loss": 5.091142654418945,
+ "step": 705
+ },
+ {
+ "epoch": 9.810480349344978,
+ "grad_norm": 0.01980586163699627,
+ "learning_rate": 0.0006,
+ "loss": 5.03350830078125,
+ "step": 706
+ },
+ {
+ "epoch": 9.824454148471617,
+ "grad_norm": 0.02048729732632637,
+ "learning_rate": 0.0006,
+ "loss": 5.03611421585083,
+ "step": 707
+ },
+ {
+ "epoch": 9.838427947598253,
+ "grad_norm": 0.02058163844048977,
+ "learning_rate": 0.0006,
+ "loss": 4.9613447189331055,
+ "step": 708
+ },
+ {
+ "epoch": 9.85240174672489,
+ "grad_norm": 0.018004924058914185,
+ "learning_rate": 0.0006,
+ "loss": 4.966541767120361,
+ "step": 709
+ },
+ {
+ "epoch": 9.866375545851529,
+ "grad_norm": 0.0198152307420969,
+ "learning_rate": 0.0006,
+ "loss": 5.01750373840332,
+ "step": 710
+ },
+ {
+ "epoch": 9.880349344978166,
+ "grad_norm": 0.018532969057559967,
+ "learning_rate": 0.0006,
+ "loss": 5.097424507141113,
+ "step": 711
+ },
+ {
+ "epoch": 9.894323144104803,
+ "grad_norm": 0.018939971923828125,
+ "learning_rate": 0.0006,
+ "loss": 5.177223205566406,
+ "step": 712
+ },
+ {
+ "epoch": 9.908296943231441,
+ "grad_norm": 0.017969170585274696,
+ "learning_rate": 0.0006,
+ "loss": 5.118013858795166,
+ "step": 713
+ },
+ {
+ "epoch": 9.922270742358078,
+ "grad_norm": 0.018021270632743835,
+ "learning_rate": 0.0006,
+ "loss": 5.034787178039551,
+ "step": 714
+ },
+ {
+ "epoch": 9.936244541484717,
+ "grad_norm": 0.017009710893034935,
+ "learning_rate": 0.0006,
+ "loss": 5.003920555114746,
+ "step": 715
+ },
+ {
+ "epoch": 9.950218340611354,
+ "grad_norm": 0.017722049728035927,
+ "learning_rate": 0.0006,
+ "loss": 5.031939506530762,
+ "step": 716
+ },
+ {
+ "epoch": 9.96419213973799,
+ "grad_norm": 0.017134087160229683,
+ "learning_rate": 0.0006,
+ "loss": 5.093474388122559,
+ "step": 717
+ },
+ {
+ "epoch": 9.97816593886463,
+ "grad_norm": 0.018511613830924034,
+ "learning_rate": 0.0006,
+ "loss": 5.078604698181152,
+ "step": 718
+ },
+ {
+ "epoch": 9.992139737991266,
+ "grad_norm": 0.021206334233283997,
+ "learning_rate": 0.0006,
+ "loss": 5.052572727203369,
+ "step": 719
+ },
+ {
+ "epoch": 10.0,
+ "grad_norm": 0.02239062264561653,
+ "learning_rate": 0.0006,
+ "loss": 5.063547134399414,
+ "step": 720
+ },
+ {
+ "epoch": 10.0,
+ "eval_loss": 5.173044681549072,
+ "eval_runtime": 56.7287,
+ "eval_samples_per_second": 43.047,
+ "eval_steps_per_second": 1.357,
+ "step": 720
+ },
+ {
+ "epoch": 10.013973799126637,
+ "grad_norm": 0.020757969468832016,
+ "learning_rate": 0.0006,
+ "loss": 5.045648574829102,
+ "step": 721
+ },
+ {
+ "epoch": 10.027947598253276,
+ "grad_norm": 0.022697702050209045,
+ "learning_rate": 0.0006,
+ "loss": 4.993833065032959,
+ "step": 722
+ },
+ {
+ "epoch": 10.041921397379912,
+ "grad_norm": 0.02434331737458706,
+ "learning_rate": 0.0006,
+ "loss": 5.073219299316406,
+ "step": 723
+ },
+ {
+ "epoch": 10.055895196506551,
+ "grad_norm": 0.025499660521745682,
+ "learning_rate": 0.0006,
+ "loss": 5.019072532653809,
+ "step": 724
+ },
+ {
+ "epoch": 10.069868995633188,
+ "grad_norm": 0.029281053692102432,
+ "learning_rate": 0.0006,
+ "loss": 4.959851264953613,
+ "step": 725
+ },
+ {
+ "epoch": 10.083842794759825,
+ "grad_norm": 0.033878110349178314,
+ "learning_rate": 0.0006,
+ "loss": 5.132717609405518,
+ "step": 726
+ },
+ {
+ "epoch": 10.097816593886463,
+ "grad_norm": 0.03385764732956886,
+ "learning_rate": 0.0006,
+ "loss": 5.0482892990112305,
+ "step": 727
+ },
+ {
+ "epoch": 10.1117903930131,
+ "grad_norm": 0.03466648980975151,
+ "learning_rate": 0.0006,
+ "loss": 4.946855545043945,
+ "step": 728
+ },
+ {
+ "epoch": 10.125764192139737,
+ "grad_norm": 0.028546305373311043,
+ "learning_rate": 0.0006,
+ "loss": 5.027774333953857,
+ "step": 729
+ },
+ {
+ "epoch": 10.139737991266376,
+ "grad_norm": 0.030575869604945183,
+ "learning_rate": 0.0006,
+ "loss": 4.9768476486206055,
+ "step": 730
+ },
+ {
+ "epoch": 10.153711790393013,
+ "grad_norm": 0.03614303469657898,
+ "learning_rate": 0.0006,
+ "loss": 5.078113555908203,
+ "step": 731
+ },
+ {
+ "epoch": 10.167685589519651,
+ "grad_norm": 0.040991660207509995,
+ "learning_rate": 0.0006,
+ "loss": 5.033185958862305,
+ "step": 732
+ },
+ {
+ "epoch": 10.181659388646288,
+ "grad_norm": 0.04551699757575989,
+ "learning_rate": 0.0006,
+ "loss": 5.0820746421813965,
+ "step": 733
+ },
+ {
+ "epoch": 10.195633187772925,
+ "grad_norm": 0.040193989872932434,
+ "learning_rate": 0.0006,
+ "loss": 5.059700965881348,
+ "step": 734
+ },
+ {
+ "epoch": 10.209606986899564,
+ "grad_norm": 0.035851605236530304,
+ "learning_rate": 0.0006,
+ "loss": 5.142467498779297,
+ "step": 735
+ },
+ {
+ "epoch": 10.2235807860262,
+ "grad_norm": 0.034200169146060944,
+ "learning_rate": 0.0006,
+ "loss": 5.072312355041504,
+ "step": 736
+ },
+ {
+ "epoch": 10.237554585152838,
+ "grad_norm": 0.033621110022068024,
+ "learning_rate": 0.0006,
+ "loss": 5.08037805557251,
+ "step": 737
+ },
+ {
+ "epoch": 10.251528384279476,
+ "grad_norm": 0.037372369319200516,
+ "learning_rate": 0.0006,
+ "loss": 4.998425483703613,
+ "step": 738
+ },
+ {
+ "epoch": 10.265502183406113,
+ "grad_norm": 0.037044707685709,
+ "learning_rate": 0.0006,
+ "loss": 4.923993110656738,
+ "step": 739
+ },
+ {
+ "epoch": 10.279475982532752,
+ "grad_norm": 0.030454808846116066,
+ "learning_rate": 0.0006,
+ "loss": 5.0889997482299805,
+ "step": 740
+ },
+ {
+ "epoch": 10.293449781659389,
+ "grad_norm": 0.030969638377428055,
+ "learning_rate": 0.0006,
+ "loss": 4.984112739562988,
+ "step": 741
+ },
+ {
+ "epoch": 10.307423580786025,
+ "grad_norm": 0.0276983380317688,
+ "learning_rate": 0.0006,
+ "loss": 4.955412864685059,
+ "step": 742
+ },
+ {
+ "epoch": 10.321397379912664,
+ "grad_norm": 0.02716052532196045,
+ "learning_rate": 0.0006,
+ "loss": 5.0098876953125,
+ "step": 743
+ },
+ {
+ "epoch": 10.335371179039301,
+ "grad_norm": 0.02403552085161209,
+ "learning_rate": 0.0006,
+ "loss": 5.165160655975342,
+ "step": 744
+ },
+ {
+ "epoch": 10.34934497816594,
+ "grad_norm": 0.0257862601429224,
+ "learning_rate": 0.0006,
+ "loss": 5.001547813415527,
+ "step": 745
+ },
+ {
+ "epoch": 10.363318777292577,
+ "grad_norm": 0.02300378680229187,
+ "learning_rate": 0.0006,
+ "loss": 5.038888931274414,
+ "step": 746
+ },
+ {
+ "epoch": 10.377292576419213,
+ "grad_norm": 0.019766854122281075,
+ "learning_rate": 0.0006,
+ "loss": 4.924291610717773,
+ "step": 747
+ },
+ {
+ "epoch": 10.391266375545852,
+ "grad_norm": 0.01888395845890045,
+ "learning_rate": 0.0006,
+ "loss": 5.059971332550049,
+ "step": 748
+ },
+ {
+ "epoch": 10.405240174672489,
+ "grad_norm": 0.015577499754726887,
+ "learning_rate": 0.0006,
+ "loss": 4.987686634063721,
+ "step": 749
+ },
+ {
+ "epoch": 10.419213973799126,
+ "grad_norm": 0.016012346372008324,
+ "learning_rate": 0.0006,
+ "loss": 5.015157699584961,
+ "step": 750
+ },
+ {
+ "epoch": 10.433187772925764,
+ "grad_norm": 0.016277998685836792,
+ "learning_rate": 0.0006,
+ "loss": 5.004931449890137,
+ "step": 751
+ },
+ {
+ "epoch": 10.447161572052401,
+ "grad_norm": 0.017504561692476273,
+ "learning_rate": 0.0006,
+ "loss": 4.902827739715576,
+ "step": 752
+ },
+ {
+ "epoch": 10.46113537117904,
+ "grad_norm": 0.01569022797048092,
+ "learning_rate": 0.0006,
+ "loss": 4.954863548278809,
+ "step": 753
+ },
+ {
+ "epoch": 10.475109170305677,
+ "grad_norm": 0.014114422723650932,
+ "learning_rate": 0.0006,
+ "loss": 5.00656795501709,
+ "step": 754
+ },
+ {
+ "epoch": 10.489082969432314,
+ "grad_norm": 0.01573832333087921,
+ "learning_rate": 0.0006,
+ "loss": 4.978353023529053,
+ "step": 755
+ },
+ {
+ "epoch": 10.503056768558952,
+ "grad_norm": 0.016480036079883575,
+ "learning_rate": 0.0006,
+ "loss": 4.953327655792236,
+ "step": 756
+ },
+ {
+ "epoch": 10.51703056768559,
+ "grad_norm": 0.014008025638759136,
+ "learning_rate": 0.0006,
+ "loss": 5.050840377807617,
+ "step": 757
+ },
+ {
+ "epoch": 10.531004366812226,
+ "grad_norm": 0.013147998601198196,
+ "learning_rate": 0.0006,
+ "loss": 4.974964141845703,
+ "step": 758
+ },
+ {
+ "epoch": 10.544978165938865,
+ "grad_norm": 0.013807603158056736,
+ "learning_rate": 0.0006,
+ "loss": 4.927907943725586,
+ "step": 759
+ },
+ {
+ "epoch": 10.558951965065502,
+ "grad_norm": 0.013555224984884262,
+ "learning_rate": 0.0006,
+ "loss": 4.980690956115723,
+ "step": 760
+ },
+ {
+ "epoch": 10.57292576419214,
+ "grad_norm": 0.013045408762991428,
+ "learning_rate": 0.0006,
+ "loss": 4.9436116218566895,
+ "step": 761
+ },
+ {
+ "epoch": 10.586899563318777,
+ "grad_norm": 0.012858862057328224,
+ "learning_rate": 0.0006,
+ "loss": 4.9595770835876465,
+ "step": 762
+ },
+ {
+ "epoch": 10.600873362445414,
+ "grad_norm": 0.01468253880739212,
+ "learning_rate": 0.0006,
+ "loss": 4.90725040435791,
+ "step": 763
+ },
+ {
+ "epoch": 10.614847161572053,
+ "grad_norm": 0.013224679045379162,
+ "learning_rate": 0.0006,
+ "loss": 4.877615928649902,
+ "step": 764
+ },
+ {
+ "epoch": 10.62882096069869,
+ "grad_norm": 0.013555348850786686,
+ "learning_rate": 0.0006,
+ "loss": 4.9456610679626465,
+ "step": 765
+ },
+ {
+ "epoch": 10.642794759825328,
+ "grad_norm": 0.014387160539627075,
+ "learning_rate": 0.0006,
+ "loss": 4.9233717918396,
+ "step": 766
+ },
+ {
+ "epoch": 10.656768558951965,
+ "grad_norm": 0.014354088343679905,
+ "learning_rate": 0.0006,
+ "loss": 4.983511924743652,
+ "step": 767
+ },
+ {
+ "epoch": 10.670742358078602,
+ "grad_norm": 0.01332700252532959,
+ "learning_rate": 0.0006,
+ "loss": 4.934416770935059,
+ "step": 768
+ },
+ {
+ "epoch": 10.68471615720524,
+ "grad_norm": 0.011543313041329384,
+ "learning_rate": 0.0006,
+ "loss": 4.96388053894043,
+ "step": 769
+ },
+ {
+ "epoch": 10.698689956331878,
+ "grad_norm": 0.012330878525972366,
+ "learning_rate": 0.0006,
+ "loss": 4.946396350860596,
+ "step": 770
+ },
+ {
+ "epoch": 10.712663755458514,
+ "grad_norm": 0.01379779726266861,
+ "learning_rate": 0.0006,
+ "loss": 4.981019020080566,
+ "step": 771
+ },
+ {
+ "epoch": 10.726637554585153,
+ "grad_norm": 0.014612431637942791,
+ "learning_rate": 0.0006,
+ "loss": 4.894567966461182,
+ "step": 772
+ },
+ {
+ "epoch": 10.74061135371179,
+ "grad_norm": 0.013062899932265282,
+ "learning_rate": 0.0006,
+ "loss": 5.033038139343262,
+ "step": 773
+ },
+ {
+ "epoch": 10.754585152838429,
+ "grad_norm": 0.013898049481213093,
+ "learning_rate": 0.0006,
+ "loss": 4.823636054992676,
+ "step": 774
+ },
+ {
+ "epoch": 10.768558951965066,
+ "grad_norm": 0.01502019353210926,
+ "learning_rate": 0.0006,
+ "loss": 5.020156383514404,
+ "step": 775
+ },
+ {
+ "epoch": 10.782532751091702,
+ "grad_norm": 0.013810204342007637,
+ "learning_rate": 0.0006,
+ "loss": 4.935647010803223,
+ "step": 776
+ },
+ {
+ "epoch": 10.796506550218341,
+ "grad_norm": 0.013616513460874557,
+ "learning_rate": 0.0006,
+ "loss": 4.901486396789551,
+ "step": 777
+ },
+ {
+ "epoch": 10.810480349344978,
+ "grad_norm": 0.013601024635136127,
+ "learning_rate": 0.0006,
+ "loss": 4.931873321533203,
+ "step": 778
+ },
+ {
+ "epoch": 10.824454148471617,
+ "grad_norm": 0.013668350875377655,
+ "learning_rate": 0.0006,
+ "loss": 4.937256813049316,
+ "step": 779
+ },
+ {
+ "epoch": 10.838427947598253,
+ "grad_norm": 0.01581823080778122,
+ "learning_rate": 0.0006,
+ "loss": 4.9611406326293945,
+ "step": 780
+ },
+ {
+ "epoch": 10.85240174672489,
+ "grad_norm": 0.019478484988212585,
+ "learning_rate": 0.0006,
+ "loss": 4.9212236404418945,
+ "step": 781
+ },
+ {
+ "epoch": 10.866375545851529,
+ "grad_norm": 0.026921333745121956,
+ "learning_rate": 0.0006,
+ "loss": 4.88501501083374,
+ "step": 782
+ },
+ {
+ "epoch": 10.880349344978166,
+ "grad_norm": 0.03155914694070816,
+ "learning_rate": 0.0006,
+ "loss": 4.989967346191406,
+ "step": 783
+ },
+ {
+ "epoch": 10.894323144104803,
+ "grad_norm": 0.029689429327845573,
+ "learning_rate": 0.0006,
+ "loss": 4.936000823974609,
+ "step": 784
+ },
+ {
+ "epoch": 10.908296943231441,
+ "grad_norm": 0.026626063510775566,
+ "learning_rate": 0.0006,
+ "loss": 4.984785079956055,
+ "step": 785
+ },
+ {
+ "epoch": 10.922270742358078,
+ "grad_norm": 0.026168212294578552,
+ "learning_rate": 0.0006,
+ "loss": 4.908282279968262,
+ "step": 786
+ },
+ {
+ "epoch": 10.936244541484717,
+ "grad_norm": 0.02828158251941204,
+ "learning_rate": 0.0006,
+ "loss": 4.926390647888184,
+ "step": 787
+ },
+ {
+ "epoch": 10.950218340611354,
+ "grad_norm": 0.02649078331887722,
+ "learning_rate": 0.0006,
+ "loss": 5.00400972366333,
+ "step": 788
+ },
+ {
+ "epoch": 10.96419213973799,
+ "grad_norm": 0.0256856270134449,
+ "learning_rate": 0.0006,
+ "loss": 5.0317487716674805,
+ "step": 789
+ },
+ {
+ "epoch": 10.97816593886463,
+ "grad_norm": 0.024728883057832718,
+ "learning_rate": 0.0006,
+ "loss": 4.9299635887146,
+ "step": 790
+ },
+ {
+ "epoch": 10.992139737991266,
+ "grad_norm": 0.026241116225719452,
+ "learning_rate": 0.0006,
+ "loss": 4.947319984436035,
+ "step": 791
+ },
+ {
+ "epoch": 11.0,
+ "grad_norm": 0.023526743054389954,
+ "learning_rate": 0.0006,
+ "loss": 4.94429349899292,
+ "step": 792
+ },
+ {
+ "epoch": 11.0,
+ "eval_loss": 5.104345798492432,
+ "eval_runtime": 57.0955,
+ "eval_samples_per_second": 42.77,
+ "eval_steps_per_second": 1.349,
+ "step": 792
+ },
+ {
+ "epoch": 11.013973799126637,
+ "grad_norm": 0.0216788612306118,
+ "learning_rate": 0.0006,
+ "loss": 4.881950855255127,
+ "step": 793
+ },
+ {
+ "epoch": 11.027947598253276,
+ "grad_norm": 0.01830463856458664,
+ "learning_rate": 0.0006,
+ "loss": 4.971016883850098,
+ "step": 794
+ },
+ {
+ "epoch": 11.041921397379912,
+ "grad_norm": 0.02204521745443344,
+ "learning_rate": 0.0006,
+ "loss": 4.894623756408691,
+ "step": 795
+ },
+ {
+ "epoch": 11.055895196506551,
+ "grad_norm": 0.0210255216807127,
+ "learning_rate": 0.0006,
+ "loss": 4.834009170532227,
+ "step": 796
+ },
+ {
+ "epoch": 11.069868995633188,
+ "grad_norm": 0.019797448068857193,
+ "learning_rate": 0.0006,
+ "loss": 5.086714744567871,
+ "step": 797
+ },
+ {
+ "epoch": 11.083842794759825,
+ "grad_norm": 0.02036641724407673,
+ "learning_rate": 0.0006,
+ "loss": 4.920056343078613,
+ "step": 798
+ },
+ {
+ "epoch": 11.097816593886463,
+ "grad_norm": 0.01966066285967827,
+ "learning_rate": 0.0006,
+ "loss": 4.879859447479248,
+ "step": 799
+ },
+ {
+ "epoch": 11.1117903930131,
+ "grad_norm": 0.0202149897813797,
+ "learning_rate": 0.0006,
+ "loss": 4.933640480041504,
+ "step": 800
+ },
+ {
+ "epoch": 11.125764192139737,
+ "grad_norm": 0.01893039606511593,
+ "learning_rate": 0.0006,
+ "loss": 4.901970863342285,
+ "step": 801
+ },
+ {
+ "epoch": 11.139737991266376,
+ "grad_norm": 0.016040155664086342,
+ "learning_rate": 0.0006,
+ "loss": 4.905491828918457,
+ "step": 802
+ },
+ {
+ "epoch": 11.153711790393013,
+ "grad_norm": 0.01720350608229637,
+ "learning_rate": 0.0006,
+ "loss": 4.798816680908203,
+ "step": 803
+ },
+ {
+ "epoch": 11.167685589519651,
+ "grad_norm": 0.017002852633595467,
+ "learning_rate": 0.0006,
+ "loss": 4.851859092712402,
+ "step": 804
+ },
+ {
+ "epoch": 11.181659388646288,
+ "grad_norm": 0.015594680793583393,
+ "learning_rate": 0.0006,
+ "loss": 4.90328311920166,
+ "step": 805
+ },
+ {
+ "epoch": 11.195633187772925,
+ "grad_norm": 0.014904462732374668,
+ "learning_rate": 0.0006,
+ "loss": 4.904215335845947,
+ "step": 806
+ },
+ {
+ "epoch": 11.209606986899564,
+ "grad_norm": 0.01366228237748146,
+ "learning_rate": 0.0006,
+ "loss": 4.9813055992126465,
+ "step": 807
+ },
+ {
+ "epoch": 11.2235807860262,
+ "grad_norm": 0.014719638973474503,
+ "learning_rate": 0.0006,
+ "loss": 4.927306175231934,
+ "step": 808
+ },
+ {
+ "epoch": 11.237554585152838,
+ "grad_norm": 0.015016036108136177,
+ "learning_rate": 0.0006,
+ "loss": 4.874215126037598,
+ "step": 809
+ },
+ {
+ "epoch": 11.251528384279476,
+ "grad_norm": 0.014925424009561539,
+ "learning_rate": 0.0006,
+ "loss": 4.838257789611816,
+ "step": 810
+ },
+ {
+ "epoch": 11.265502183406113,
+ "grad_norm": 0.014335823245346546,
+ "learning_rate": 0.0006,
+ "loss": 4.940608024597168,
+ "step": 811
+ },
+ {
+ "epoch": 11.279475982532752,
+ "grad_norm": 0.014987512491643429,
+ "learning_rate": 0.0006,
+ "loss": 4.831531524658203,
+ "step": 812
+ },
+ {
+ "epoch": 11.293449781659389,
+ "grad_norm": 0.016087457537651062,
+ "learning_rate": 0.0006,
+ "loss": 4.863227367401123,
+ "step": 813
+ },
+ {
+ "epoch": 11.307423580786025,
+ "grad_norm": 0.01671300269663334,
+ "learning_rate": 0.0006,
+ "loss": 4.8535661697387695,
+ "step": 814
+ },
+ {
+ "epoch": 11.321397379912664,
+ "grad_norm": 0.015229142270982265,
+ "learning_rate": 0.0006,
+ "loss": 4.948635101318359,
+ "step": 815
+ },
+ {
+ "epoch": 11.335371179039301,
+ "grad_norm": 0.01532587967813015,
+ "learning_rate": 0.0006,
+ "loss": 4.9035797119140625,
+ "step": 816
+ },
+ {
+ "epoch": 11.34934497816594,
+ "grad_norm": 0.015715476125478745,
+ "learning_rate": 0.0006,
+ "loss": 4.894709587097168,
+ "step": 817
+ },
+ {
+ "epoch": 11.363318777292577,
+ "grad_norm": 0.014156977646052837,
+ "learning_rate": 0.0006,
+ "loss": 4.960206985473633,
+ "step": 818
+ },
+ {
+ "epoch": 11.377292576419213,
+ "grad_norm": 0.017076566815376282,
+ "learning_rate": 0.0006,
+ "loss": 5.005850791931152,
+ "step": 819
+ },
+ {
+ "epoch": 11.391266375545852,
+ "grad_norm": 0.0172084029763937,
+ "learning_rate": 0.0006,
+ "loss": 4.903355598449707,
+ "step": 820
+ },
+ {
+ "epoch": 11.405240174672489,
+ "grad_norm": 0.0180258359760046,
+ "learning_rate": 0.0006,
+ "loss": 4.842667579650879,
+ "step": 821
+ },
+ {
+ "epoch": 11.419213973799126,
+ "grad_norm": 0.022130120545625687,
+ "learning_rate": 0.0006,
+ "loss": 4.799970626831055,
+ "step": 822
+ },
+ {
+ "epoch": 11.433187772925764,
+ "grad_norm": 0.02384897693991661,
+ "learning_rate": 0.0006,
+ "loss": 4.860746383666992,
+ "step": 823
+ },
+ {
+ "epoch": 11.447161572052401,
+ "grad_norm": 0.021881572902202606,
+ "learning_rate": 0.0006,
+ "loss": 4.918107032775879,
+ "step": 824
+ },
+ {
+ "epoch": 11.46113537117904,
+ "grad_norm": 0.023085080087184906,
+ "learning_rate": 0.0006,
+ "loss": 4.878946304321289,
+ "step": 825
+ },
+ {
+ "epoch": 11.475109170305677,
+ "grad_norm": 0.02291012555360794,
+ "learning_rate": 0.0006,
+ "loss": 4.958887100219727,
+ "step": 826
+ },
+ {
+ "epoch": 11.489082969432314,
+ "grad_norm": 0.0199968870729208,
+ "learning_rate": 0.0006,
+ "loss": 4.80955696105957,
+ "step": 827
+ },
+ {
+ "epoch": 11.503056768558952,
+ "grad_norm": 0.018702229484915733,
+ "learning_rate": 0.0006,
+ "loss": 5.0276408195495605,
+ "step": 828
+ },
+ {
+ "epoch": 11.51703056768559,
+ "grad_norm": 0.017787311226129532,
+ "learning_rate": 0.0006,
+ "loss": 4.96018123626709,
+ "step": 829
+ },
+ {
+ "epoch": 11.531004366812226,
+ "grad_norm": 0.01704501546919346,
+ "learning_rate": 0.0006,
+ "loss": 4.914553165435791,
+ "step": 830
+ },
+ {
+ "epoch": 11.544978165938865,
+ "grad_norm": 0.017519677057862282,
+ "learning_rate": 0.0006,
+ "loss": 4.798098087310791,
+ "step": 831
+ },
+ {
+ "epoch": 11.558951965065502,
+ "grad_norm": 0.018511991947889328,
+ "learning_rate": 0.0006,
+ "loss": 4.817785263061523,
+ "step": 832
+ },
+ {
+ "epoch": 11.57292576419214,
+ "grad_norm": 0.01876644790172577,
+ "learning_rate": 0.0006,
+ "loss": 4.818995475769043,
+ "step": 833
+ },
+ {
+ "epoch": 11.586899563318777,
+ "grad_norm": 0.01753568835556507,
+ "learning_rate": 0.0006,
+ "loss": 4.811938285827637,
+ "step": 834
+ },
+ {
+ "epoch": 11.600873362445414,
+ "grad_norm": 0.017046531662344933,
+ "learning_rate": 0.0006,
+ "loss": 4.914674282073975,
+ "step": 835
+ },
+ {
+ "epoch": 11.614847161572053,
+ "grad_norm": 0.01802094094455242,
+ "learning_rate": 0.0006,
+ "loss": 4.993284225463867,
+ "step": 836
+ },
+ {
+ "epoch": 11.62882096069869,
+ "grad_norm": 0.019493764266371727,
+ "learning_rate": 0.0006,
+ "loss": 4.895628452301025,
+ "step": 837
+ },
+ {
+ "epoch": 11.642794759825328,
+ "grad_norm": 0.019461210817098618,
+ "learning_rate": 0.0006,
+ "loss": 4.869269371032715,
+ "step": 838
+ },
+ {
+ "epoch": 11.656768558951965,
+ "grad_norm": 0.01981567218899727,
+ "learning_rate": 0.0006,
+ "loss": 4.863052845001221,
+ "step": 839
+ },
+ {
+ "epoch": 11.670742358078602,
+ "grad_norm": 0.019894491881132126,
+ "learning_rate": 0.0006,
+ "loss": 4.8775787353515625,
+ "step": 840
+ },
+ {
+ "epoch": 11.68471615720524,
+ "grad_norm": 0.01875786855816841,
+ "learning_rate": 0.0006,
+ "loss": 4.817763328552246,
+ "step": 841
+ },
+ {
+ "epoch": 11.698689956331878,
+ "grad_norm": 0.020926695317029953,
+ "learning_rate": 0.0006,
+ "loss": 4.7470526695251465,
+ "step": 842
+ },
+ {
+ "epoch": 11.712663755458514,
+ "grad_norm": 0.021477103233337402,
+ "learning_rate": 0.0006,
+ "loss": 4.762682914733887,
+ "step": 843
+ },
+ {
+ "epoch": 11.726637554585153,
+ "grad_norm": 0.020628711208701134,
+ "learning_rate": 0.0006,
+ "loss": 4.807985782623291,
+ "step": 844
+ },
+ {
+ "epoch": 11.74061135371179,
+ "grad_norm": 0.021489612758159637,
+ "learning_rate": 0.0006,
+ "loss": 4.843091011047363,
+ "step": 845
+ },
+ {
+ "epoch": 11.754585152838429,
+ "grad_norm": 0.021194491535425186,
+ "learning_rate": 0.0006,
+ "loss": 4.8300251960754395,
+ "step": 846
+ },
+ {
+ "epoch": 11.768558951965066,
+ "grad_norm": 0.01833350397646427,
+ "learning_rate": 0.0006,
+ "loss": 5.000202178955078,
+ "step": 847
+ },
+ {
+ "epoch": 11.782532751091702,
+ "grad_norm": 0.018096931278705597,
+ "learning_rate": 0.0006,
+ "loss": 4.880648136138916,
+ "step": 848
+ },
+ {
+ "epoch": 11.796506550218341,
+ "grad_norm": 0.0197161752730608,
+ "learning_rate": 0.0006,
+ "loss": 4.788316249847412,
+ "step": 849
+ },
+ {
+ "epoch": 11.810480349344978,
+ "grad_norm": 0.0203181654214859,
+ "learning_rate": 0.0006,
+ "loss": 4.870944976806641,
+ "step": 850
+ },
+ {
+ "epoch": 11.824454148471617,
+ "grad_norm": 0.01908455230295658,
+ "learning_rate": 0.0006,
+ "loss": 4.890144348144531,
+ "step": 851
+ },
+ {
+ "epoch": 11.838427947598253,
+ "grad_norm": 0.021281801164150238,
+ "learning_rate": 0.0006,
+ "loss": 4.89578914642334,
+ "step": 852
+ },
+ {
+ "epoch": 11.85240174672489,
+ "grad_norm": 0.02124273031949997,
+ "learning_rate": 0.0006,
+ "loss": 4.883749008178711,
+ "step": 853
+ },
+ {
+ "epoch": 11.866375545851529,
+ "grad_norm": 0.020969398319721222,
+ "learning_rate": 0.0006,
+ "loss": 4.809142589569092,
+ "step": 854
+ },
+ {
+ "epoch": 11.880349344978166,
+ "grad_norm": 0.01967657171189785,
+ "learning_rate": 0.0006,
+ "loss": 4.821864604949951,
+ "step": 855
+ },
+ {
+ "epoch": 11.894323144104803,
+ "grad_norm": 0.020887333899736404,
+ "learning_rate": 0.0006,
+ "loss": 4.917985916137695,
+ "step": 856
+ },
+ {
+ "epoch": 11.908296943231441,
+ "grad_norm": 0.019203083589673042,
+ "learning_rate": 0.0006,
+ "loss": 4.824960708618164,
+ "step": 857
+ },
+ {
+ "epoch": 11.922270742358078,
+ "grad_norm": 0.020162401720881462,
+ "learning_rate": 0.0006,
+ "loss": 4.772617816925049,
+ "step": 858
+ },
+ {
+ "epoch": 11.936244541484717,
+ "grad_norm": 0.02251487225294113,
+ "learning_rate": 0.0006,
+ "loss": 4.735418319702148,
+ "step": 859
+ },
+ {
+ "epoch": 11.950218340611354,
+ "grad_norm": 0.020313527435064316,
+ "learning_rate": 0.0006,
+ "loss": 4.939233779907227,
+ "step": 860
+ },
+ {
+ "epoch": 11.96419213973799,
+ "grad_norm": 0.021635599434375763,
+ "learning_rate": 0.0006,
+ "loss": 4.861807346343994,
+ "step": 861
+ },
+ {
+ "epoch": 11.97816593886463,
+ "grad_norm": 0.02285311557352543,
+ "learning_rate": 0.0006,
+ "loss": 4.932732582092285,
+ "step": 862
+ },
+ {
+ "epoch": 11.992139737991266,
+ "grad_norm": 0.02072535641491413,
+ "learning_rate": 0.0006,
+ "loss": 4.796767711639404,
+ "step": 863
+ },
+ {
+ "epoch": 12.0,
+ "grad_norm": 0.020588304847478867,
+ "learning_rate": 0.0006,
+ "loss": 4.997537612915039,
+ "step": 864
+ },
+ {
+ "epoch": 12.0,
+ "eval_loss": 5.044651985168457,
+ "eval_runtime": 56.9145,
+ "eval_samples_per_second": 42.906,
+ "eval_steps_per_second": 1.353,
+ "step": 864
+ },
+ {
+ "epoch": 12.013973799126637,
+ "grad_norm": 0.017597166821360588,
+ "learning_rate": 0.0006,
+ "loss": 4.830524444580078,
+ "step": 865
+ },
+ {
+ "epoch": 12.027947598253276,
+ "grad_norm": 0.01716788113117218,
+ "learning_rate": 0.0006,
+ "loss": 4.894325256347656,
+ "step": 866
+ },
+ {
+ "epoch": 12.041921397379912,
+ "grad_norm": 0.017712457105517387,
+ "learning_rate": 0.0006,
+ "loss": 4.868297576904297,
+ "step": 867
+ },
+ {
+ "epoch": 12.055895196506551,
+ "grad_norm": 0.021486390382051468,
+ "learning_rate": 0.0006,
+ "loss": 4.922049522399902,
+ "step": 868
+ },
+ {
+ "epoch": 12.069868995633188,
+ "grad_norm": 0.024560654535889626,
+ "learning_rate": 0.0006,
+ "loss": 4.734241008758545,
+ "step": 869
+ },
+ {
+ "epoch": 12.083842794759825,
+ "grad_norm": 0.025870252400636673,
+ "learning_rate": 0.0006,
+ "loss": 4.868566989898682,
+ "step": 870
+ },
+ {
+ "epoch": 12.097816593886463,
+ "grad_norm": 0.0277389045804739,
+ "learning_rate": 0.0006,
+ "loss": 4.755006790161133,
+ "step": 871
+ },
+ {
+ "epoch": 12.1117903930131,
+ "grad_norm": 0.029100844636559486,
+ "learning_rate": 0.0006,
+ "loss": 4.909667015075684,
+ "step": 872
+ },
+ {
+ "epoch": 12.125764192139737,
+ "grad_norm": 0.031833067536354065,
+ "learning_rate": 0.0006,
+ "loss": 4.858674049377441,
+ "step": 873
+ },
+ {
+ "epoch": 12.139737991266376,
+ "grad_norm": 0.032709237188100815,
+ "learning_rate": 0.0006,
+ "loss": 4.823696136474609,
+ "step": 874
+ },
+ {
+ "epoch": 12.153711790393013,
+ "grad_norm": 0.02995380386710167,
+ "learning_rate": 0.0006,
+ "loss": 4.850235462188721,
+ "step": 875
+ },
+ {
+ "epoch": 12.167685589519651,
+ "grad_norm": 0.034098610281944275,
+ "learning_rate": 0.0006,
+ "loss": 4.8169989585876465,
+ "step": 876
+ },
+ {
+ "epoch": 12.181659388646288,
+ "grad_norm": 0.03377080336213112,
+ "learning_rate": 0.0006,
+ "loss": 4.90738582611084,
+ "step": 877
+ },
+ {
+ "epoch": 12.195633187772925,
+ "grad_norm": 0.029780825600028038,
+ "learning_rate": 0.0006,
+ "loss": 4.811163902282715,
+ "step": 878
+ },
+ {
+ "epoch": 12.209606986899564,
+ "grad_norm": 0.03118024580180645,
+ "learning_rate": 0.0006,
+ "loss": 4.861286640167236,
+ "step": 879
+ },
+ {
+ "epoch": 12.2235807860262,
+ "grad_norm": 0.033052004873752594,
+ "learning_rate": 0.0006,
+ "loss": 4.865021228790283,
+ "step": 880
+ },
+ {
+ "epoch": 12.237554585152838,
+ "grad_norm": 0.029070032760500908,
+ "learning_rate": 0.0006,
+ "loss": 4.947303771972656,
+ "step": 881
+ },
+ {
+ "epoch": 12.251528384279476,
+ "grad_norm": 0.028355715796351433,
+ "learning_rate": 0.0006,
+ "loss": 4.804648399353027,
+ "step": 882
+ },
+ {
+ "epoch": 12.265502183406113,
+ "grad_norm": 0.031257227063179016,
+ "learning_rate": 0.0006,
+ "loss": 4.757718086242676,
+ "step": 883
+ },
+ {
+ "epoch": 12.279475982532752,
+ "grad_norm": 0.031438373029232025,
+ "learning_rate": 0.0006,
+ "loss": 4.851268768310547,
+ "step": 884
+ },
+ {
+ "epoch": 12.293449781659389,
+ "grad_norm": 0.030492138117551804,
+ "learning_rate": 0.0006,
+ "loss": 4.7891035079956055,
+ "step": 885
+ },
+ {
+ "epoch": 12.307423580786025,
+ "grad_norm": 0.026115991175174713,
+ "learning_rate": 0.0006,
+ "loss": 4.86474609375,
+ "step": 886
+ },
+ {
+ "epoch": 12.321397379912664,
+ "grad_norm": 0.024184465408325195,
+ "learning_rate": 0.0006,
+ "loss": 4.860456943511963,
+ "step": 887
+ },
+ {
+ "epoch": 12.335371179039301,
+ "grad_norm": 0.025437982752919197,
+ "learning_rate": 0.0006,
+ "loss": 4.806251049041748,
+ "step": 888
+ },
+ {
+ "epoch": 12.34934497816594,
+ "grad_norm": 0.022390231490135193,
+ "learning_rate": 0.0006,
+ "loss": 4.807086944580078,
+ "step": 889
+ },
+ {
+ "epoch": 12.363318777292577,
+ "grad_norm": 0.020174294710159302,
+ "learning_rate": 0.0006,
+ "loss": 4.78292179107666,
+ "step": 890
+ },
+ {
+ "epoch": 12.377292576419213,
+ "grad_norm": 0.019734324887394905,
+ "learning_rate": 0.0006,
+ "loss": 4.784509658813477,
+ "step": 891
+ },
+ {
+ "epoch": 12.391266375545852,
+ "grad_norm": 0.016733523458242416,
+ "learning_rate": 0.0006,
+ "loss": 4.811755180358887,
+ "step": 892
+ },
+ {
+ "epoch": 12.405240174672489,
+ "grad_norm": 0.01570146717131138,
+ "learning_rate": 0.0006,
+ "loss": 4.720088005065918,
+ "step": 893
+ },
+ {
+ "epoch": 12.419213973799126,
+ "grad_norm": 0.01601393334567547,
+ "learning_rate": 0.0006,
+ "loss": 4.783543109893799,
+ "step": 894
+ },
+ {
+ "epoch": 12.433187772925764,
+ "grad_norm": 0.014035487547516823,
+ "learning_rate": 0.0006,
+ "loss": 4.848340034484863,
+ "step": 895
+ },
+ {
+ "epoch": 12.447161572052401,
+ "grad_norm": 0.015248487703502178,
+ "learning_rate": 0.0006,
+ "loss": 4.773287773132324,
+ "step": 896
+ },
+ {
+ "epoch": 12.46113537117904,
+ "grad_norm": 0.01532268151640892,
+ "learning_rate": 0.0006,
+ "loss": 4.799646854400635,
+ "step": 897
+ },
+ {
+ "epoch": 12.475109170305677,
+ "grad_norm": 0.013982534408569336,
+ "learning_rate": 0.0006,
+ "loss": 4.8090500831604,
+ "step": 898
+ },
+ {
+ "epoch": 12.489082969432314,
+ "grad_norm": 0.014081164263188839,
+ "learning_rate": 0.0006,
+ "loss": 4.751888751983643,
+ "step": 899
+ },
+ {
+ "epoch": 12.503056768558952,
+ "grad_norm": 0.01380133256316185,
+ "learning_rate": 0.0006,
+ "loss": 4.78144645690918,
+ "step": 900
+ },
+ {
+ "epoch": 12.51703056768559,
+ "grad_norm": 0.013555269688367844,
+ "learning_rate": 0.0006,
+ "loss": 4.779257774353027,
+ "step": 901
+ },
+ {
+ "epoch": 12.531004366812226,
+ "grad_norm": 0.013075701892375946,
+ "learning_rate": 0.0006,
+ "loss": 4.820705413818359,
+ "step": 902
+ },
+ {
+ "epoch": 12.544978165938865,
+ "grad_norm": 0.01276316400617361,
+ "learning_rate": 0.0006,
+ "loss": 4.836733818054199,
+ "step": 903
+ },
+ {
+ "epoch": 12.558951965065502,
+ "grad_norm": 0.013033892959356308,
+ "learning_rate": 0.0006,
+ "loss": 4.764178276062012,
+ "step": 904
+ },
+ {
+ "epoch": 12.57292576419214,
+ "grad_norm": 0.013996980153024197,
+ "learning_rate": 0.0006,
+ "loss": 4.767983436584473,
+ "step": 905
+ },
+ {
+ "epoch": 12.586899563318777,
+ "grad_norm": 0.015222184360027313,
+ "learning_rate": 0.0006,
+ "loss": 4.759721755981445,
+ "step": 906
+ },
+ {
+ "epoch": 12.600873362445414,
+ "grad_norm": 0.01623941957950592,
+ "learning_rate": 0.0006,
+ "loss": 4.849811553955078,
+ "step": 907
+ },
+ {
+ "epoch": 12.614847161572053,
+ "grad_norm": 0.015711182728409767,
+ "learning_rate": 0.0006,
+ "loss": 4.7829084396362305,
+ "step": 908
+ },
+ {
+ "epoch": 12.62882096069869,
+ "grad_norm": 0.014137467369437218,
+ "learning_rate": 0.0006,
+ "loss": 4.73321533203125,
+ "step": 909
+ },
+ {
+ "epoch": 12.642794759825328,
+ "grad_norm": 0.012738462537527084,
+ "learning_rate": 0.0006,
+ "loss": 4.870052337646484,
+ "step": 910
+ },
+ {
+ "epoch": 12.656768558951965,
+ "grad_norm": 0.013683130033314228,
+ "learning_rate": 0.0006,
+ "loss": 4.82658052444458,
+ "step": 911
+ },
+ {
+ "epoch": 12.670742358078602,
+ "grad_norm": 0.014409830793738365,
+ "learning_rate": 0.0006,
+ "loss": 4.707494735717773,
+ "step": 912
+ },
+ {
+ "epoch": 12.68471615720524,
+ "grad_norm": 0.013806809671223164,
+ "learning_rate": 0.0006,
+ "loss": 4.778928279876709,
+ "step": 913
+ },
+ {
+ "epoch": 12.698689956331878,
+ "grad_norm": 0.012857912108302116,
+ "learning_rate": 0.0006,
+ "loss": 4.890717506408691,
+ "step": 914
+ },
+ {
+ "epoch": 12.712663755458514,
+ "grad_norm": 0.013828758150339127,
+ "learning_rate": 0.0006,
+ "loss": 4.8357439041137695,
+ "step": 915
+ },
+ {
+ "epoch": 12.726637554585153,
+ "grad_norm": 0.01478351280093193,
+ "learning_rate": 0.0006,
+ "loss": 4.869986534118652,
+ "step": 916
+ },
+ {
+ "epoch": 12.74061135371179,
+ "grad_norm": 0.01326421182602644,
+ "learning_rate": 0.0006,
+ "loss": 4.830646514892578,
+ "step": 917
+ },
+ {
+ "epoch": 12.754585152838429,
+ "grad_norm": 0.013335433788597584,
+ "learning_rate": 0.0006,
+ "loss": 4.7786149978637695,
+ "step": 918
+ },
+ {
+ "epoch": 12.768558951965066,
+ "grad_norm": 0.014261147007346153,
+ "learning_rate": 0.0006,
+ "loss": 4.788693428039551,
+ "step": 919
+ },
+ {
+ "epoch": 12.782532751091702,
+ "grad_norm": 0.014274783432483673,
+ "learning_rate": 0.0006,
+ "loss": 4.77295446395874,
+ "step": 920
+ },
+ {
+ "epoch": 12.796506550218341,
+ "grad_norm": 0.015074674971401691,
+ "learning_rate": 0.0006,
+ "loss": 4.830754280090332,
+ "step": 921
+ },
+ {
+ "epoch": 12.810480349344978,
+ "grad_norm": 0.015599515289068222,
+ "learning_rate": 0.0006,
+ "loss": 4.837759017944336,
+ "step": 922
+ },
+ {
+ "epoch": 12.824454148471617,
+ "grad_norm": 0.014595242217183113,
+ "learning_rate": 0.0006,
+ "loss": 4.882965564727783,
+ "step": 923
+ },
+ {
+ "epoch": 12.838427947598253,
+ "grad_norm": 0.014951467514038086,
+ "learning_rate": 0.0006,
+ "loss": 4.709748268127441,
+ "step": 924
+ },
+ {
+ "epoch": 12.85240174672489,
+ "grad_norm": 0.01539295818656683,
+ "learning_rate": 0.0006,
+ "loss": 4.791160583496094,
+ "step": 925
+ },
+ {
+ "epoch": 12.866375545851529,
+ "grad_norm": 0.015987906605005264,
+ "learning_rate": 0.0006,
+ "loss": 4.661876678466797,
+ "step": 926
+ },
+ {
+ "epoch": 12.880349344978166,
+ "grad_norm": 0.016527341678738594,
+ "learning_rate": 0.0006,
+ "loss": 4.721952438354492,
+ "step": 927
+ },
+ {
+ "epoch": 12.894323144104803,
+ "grad_norm": 0.01637161523103714,
+ "learning_rate": 0.0006,
+ "loss": 4.704446792602539,
+ "step": 928
+ },
+ {
+ "epoch": 12.908296943231441,
+ "grad_norm": 0.015221050009131432,
+ "learning_rate": 0.0006,
+ "loss": 4.771907806396484,
+ "step": 929
+ },
+ {
+ "epoch": 12.922270742358078,
+ "grad_norm": 0.016242166981101036,
+ "learning_rate": 0.0006,
+ "loss": 4.806267738342285,
+ "step": 930
+ },
+ {
+ "epoch": 12.936244541484717,
+ "grad_norm": 0.01628357358276844,
+ "learning_rate": 0.0006,
+ "loss": 4.7252326011657715,
+ "step": 931
+ },
+ {
+ "epoch": 12.950218340611354,
+ "grad_norm": 0.017428133636713028,
+ "learning_rate": 0.0006,
+ "loss": 4.765689849853516,
+ "step": 932
+ },
+ {
+ "epoch": 12.96419213973799,
+ "grad_norm": 0.01870061084628105,
+ "learning_rate": 0.0006,
+ "loss": 4.814488887786865,
+ "step": 933
+ },
+ {
+ "epoch": 12.97816593886463,
+ "grad_norm": 0.019796060398221016,
+ "learning_rate": 0.0006,
+ "loss": 4.739185810089111,
+ "step": 934
+ },
+ {
+ "epoch": 12.992139737991266,
+ "grad_norm": 0.01985820196568966,
+ "learning_rate": 0.0006,
+ "loss": 4.810608863830566,
+ "step": 935
+ },
+ {
+ "epoch": 13.0,
+ "grad_norm": 0.020834477618336678,
+ "learning_rate": 0.0006,
+ "loss": 4.727858066558838,
+ "step": 936
+ },
+ {
+ "epoch": 13.0,
+ "eval_loss": 4.979116916656494,
+ "eval_runtime": 57.2877,
+ "eval_samples_per_second": 42.627,
+ "eval_steps_per_second": 1.344,
+ "step": 936
+ },
+ {
+ "epoch": 13.013973799126637,
+ "grad_norm": 0.02025519125163555,
+ "learning_rate": 0.0006,
+ "loss": 4.628875732421875,
+ "step": 937
+ },
+ {
+ "epoch": 13.027947598253276,
+ "grad_norm": 0.019596368074417114,
+ "learning_rate": 0.0006,
+ "loss": 4.694699287414551,
+ "step": 938
+ },
+ {
+ "epoch": 13.041921397379912,
+ "grad_norm": 0.020494934171438217,
+ "learning_rate": 0.0006,
+ "loss": 4.691821098327637,
+ "step": 939
+ },
+ {
+ "epoch": 13.055895196506551,
+ "grad_norm": 0.020121760666370392,
+ "learning_rate": 0.0006,
+ "loss": 4.71478271484375,
+ "step": 940
+ },
+ {
+ "epoch": 13.069868995633188,
+ "grad_norm": 0.021091405302286148,
+ "learning_rate": 0.0006,
+ "loss": 4.695915222167969,
+ "step": 941
+ },
+ {
+ "epoch": 13.083842794759825,
+ "grad_norm": 0.019787423312664032,
+ "learning_rate": 0.0006,
+ "loss": 4.8652448654174805,
+ "step": 942
+ },
+ {
+ "epoch": 13.097816593886463,
+ "grad_norm": 0.0185833927243948,
+ "learning_rate": 0.0006,
+ "loss": 4.7065887451171875,
+ "step": 943
+ },
+ {
+ "epoch": 13.1117903930131,
+ "grad_norm": 0.021220475435256958,
+ "learning_rate": 0.0006,
+ "loss": 4.741232395172119,
+ "step": 944
+ },
+ {
+ "epoch": 13.125764192139737,
+ "grad_norm": 0.02136622555553913,
+ "learning_rate": 0.0006,
+ "loss": 4.705967903137207,
+ "step": 945
+ },
+ {
+ "epoch": 13.139737991266376,
+ "grad_norm": 0.0191587433218956,
+ "learning_rate": 0.0006,
+ "loss": 4.758667945861816,
+ "step": 946
+ },
+ {
+ "epoch": 13.153711790393013,
+ "grad_norm": 0.018290022388100624,
+ "learning_rate": 0.0006,
+ "loss": 4.760666847229004,
+ "step": 947
+ },
+ {
+ "epoch": 13.167685589519651,
+ "grad_norm": 0.021704409271478653,
+ "learning_rate": 0.0006,
+ "loss": 4.750555992126465,
+ "step": 948
+ },
+ {
+ "epoch": 13.181659388646288,
+ "grad_norm": 0.021668918430805206,
+ "learning_rate": 0.0006,
+ "loss": 4.840903282165527,
+ "step": 949
+ },
+ {
+ "epoch": 13.195633187772925,
+ "grad_norm": 0.019497141242027283,
+ "learning_rate": 0.0006,
+ "loss": 4.738969326019287,
+ "step": 950
+ },
+ {
+ "epoch": 13.209606986899564,
+ "grad_norm": 0.018500596284866333,
+ "learning_rate": 0.0006,
+ "loss": 4.75154972076416,
+ "step": 951
+ },
+ {
+ "epoch": 13.2235807860262,
+ "grad_norm": 0.01872239261865616,
+ "learning_rate": 0.0006,
+ "loss": 4.811738014221191,
+ "step": 952
+ },
+ {
+ "epoch": 13.237554585152838,
+ "grad_norm": 0.020321547985076904,
+ "learning_rate": 0.0006,
+ "loss": 4.815384387969971,
+ "step": 953
+ },
+ {
+ "epoch": 13.251528384279476,
+ "grad_norm": 0.02283575013279915,
+ "learning_rate": 0.0006,
+ "loss": 4.74582052230835,
+ "step": 954
+ },
+ {
+ "epoch": 13.265502183406113,
+ "grad_norm": 0.02474398910999298,
+ "learning_rate": 0.0006,
+ "loss": 4.738375186920166,
+ "step": 955
+ },
+ {
+ "epoch": 13.279475982532752,
+ "grad_norm": 0.02065764181315899,
+ "learning_rate": 0.0006,
+ "loss": 4.746875286102295,
+ "step": 956
+ },
+ {
+ "epoch": 13.293449781659389,
+ "grad_norm": 0.020534520968794823,
+ "learning_rate": 0.0006,
+ "loss": 4.656914710998535,
+ "step": 957
+ },
+ {
+ "epoch": 13.307423580786025,
+ "grad_norm": 0.0206378772854805,
+ "learning_rate": 0.0006,
+ "loss": 4.842615604400635,
+ "step": 958
+ },
+ {
+ "epoch": 13.321397379912664,
+ "grad_norm": 0.017935654148459435,
+ "learning_rate": 0.0006,
+ "loss": 4.717827796936035,
+ "step": 959
+ },
+ {
+ "epoch": 13.335371179039301,
+ "grad_norm": 0.016958363354206085,
+ "learning_rate": 0.0006,
+ "loss": 4.681070327758789,
+ "step": 960
+ },
+ {
+ "epoch": 13.34934497816594,
+ "grad_norm": 0.015735412016510963,
+ "learning_rate": 0.0006,
+ "loss": 4.7957048416137695,
+ "step": 961
+ },
+ {
+ "epoch": 13.363318777292577,
+ "grad_norm": 0.01563699170947075,
+ "learning_rate": 0.0006,
+ "loss": 4.701192378997803,
+ "step": 962
+ },
+ {
+ "epoch": 13.377292576419213,
+ "grad_norm": 0.015207760035991669,
+ "learning_rate": 0.0006,
+ "loss": 4.73256254196167,
+ "step": 963
+ },
+ {
+ "epoch": 13.391266375545852,
+ "grad_norm": 0.015396883711218834,
+ "learning_rate": 0.0006,
+ "loss": 4.76412296295166,
+ "step": 964
+ },
+ {
+ "epoch": 13.405240174672489,
+ "grad_norm": 0.015690583735704422,
+ "learning_rate": 0.0006,
+ "loss": 4.791157245635986,
+ "step": 965
+ },
+ {
+ "epoch": 13.419213973799126,
+ "grad_norm": 0.01755724661052227,
+ "learning_rate": 0.0006,
+ "loss": 4.845666885375977,
+ "step": 966
+ },
+ {
+ "epoch": 13.433187772925764,
+ "grad_norm": 0.019443422555923462,
+ "learning_rate": 0.0006,
+ "loss": 4.5902299880981445,
+ "step": 967
+ },
+ {
+ "epoch": 13.447161572052401,
+ "grad_norm": 0.02118338644504547,
+ "learning_rate": 0.0006,
+ "loss": 4.740627288818359,
+ "step": 968
+ },
+ {
+ "epoch": 13.46113537117904,
+ "grad_norm": 0.02174839936196804,
+ "learning_rate": 0.0006,
+ "loss": 4.814560890197754,
+ "step": 969
+ },
+ {
+ "epoch": 13.475109170305677,
+ "grad_norm": 0.022860374301671982,
+ "learning_rate": 0.0006,
+ "loss": 4.66440486907959,
+ "step": 970
+ },
+ {
+ "epoch": 13.489082969432314,
+ "grad_norm": 0.024742677807807922,
+ "learning_rate": 0.0006,
+ "loss": 4.7451653480529785,
+ "step": 971
+ },
+ {
+ "epoch": 13.503056768558952,
+ "grad_norm": 0.027068907395005226,
+ "learning_rate": 0.0006,
+ "loss": 4.802584171295166,
+ "step": 972
+ },
+ {
+ "epoch": 13.51703056768559,
+ "grad_norm": 0.026316625997424126,
+ "learning_rate": 0.0006,
+ "loss": 4.842397689819336,
+ "step": 973
+ },
+ {
+ "epoch": 13.531004366812226,
+ "grad_norm": 0.024610962718725204,
+ "learning_rate": 0.0006,
+ "loss": 4.650721073150635,
+ "step": 974
+ },
+ {
+ "epoch": 13.544978165938865,
+ "grad_norm": 0.024122603237628937,
+ "learning_rate": 0.0006,
+ "loss": 4.6984357833862305,
+ "step": 975
+ },
+ {
+ "epoch": 13.558951965065502,
+ "grad_norm": 0.022699708119034767,
+ "learning_rate": 0.0006,
+ "loss": 4.793007850646973,
+ "step": 976
+ },
+ {
+ "epoch": 13.57292576419214,
+ "grad_norm": 0.021706534549593925,
+ "learning_rate": 0.0006,
+ "loss": 4.691362380981445,
+ "step": 977
+ },
+ {
+ "epoch": 13.586899563318777,
+ "grad_norm": 0.021363291889429092,
+ "learning_rate": 0.0006,
+ "loss": 4.76762580871582,
+ "step": 978
+ },
+ {
+ "epoch": 13.600873362445414,
+ "grad_norm": 0.023476677015423775,
+ "learning_rate": 0.0006,
+ "loss": 4.777202606201172,
+ "step": 979
+ },
+ {
+ "epoch": 13.614847161572053,
+ "grad_norm": 0.02457100711762905,
+ "learning_rate": 0.0006,
+ "loss": 4.590703010559082,
+ "step": 980
+ },
+ {
+ "epoch": 13.62882096069869,
+ "grad_norm": 0.02469099499285221,
+ "learning_rate": 0.0006,
+ "loss": 4.640185356140137,
+ "step": 981
+ },
+ {
+ "epoch": 13.642794759825328,
+ "grad_norm": 0.024914603680372238,
+ "learning_rate": 0.0006,
+ "loss": 4.772201061248779,
+ "step": 982
+ },
+ {
+ "epoch": 13.656768558951965,
+ "grad_norm": 0.025133173912763596,
+ "learning_rate": 0.0006,
+ "loss": 4.743337154388428,
+ "step": 983
+ },
+ {
+ "epoch": 13.670742358078602,
+ "grad_norm": 0.02034137211740017,
+ "learning_rate": 0.0006,
+ "loss": 4.752127647399902,
+ "step": 984
+ },
+ {
+ "epoch": 13.68471615720524,
+ "grad_norm": 0.019645661115646362,
+ "learning_rate": 0.0006,
+ "loss": 4.838453769683838,
+ "step": 985
+ },
+ {
+ "epoch": 13.698689956331878,
+ "grad_norm": 0.01900297962129116,
+ "learning_rate": 0.0006,
+ "loss": 4.804868221282959,
+ "step": 986
+ },
+ {
+ "epoch": 13.712663755458514,
+ "grad_norm": 0.020292341709136963,
+ "learning_rate": 0.0006,
+ "loss": 4.609743118286133,
+ "step": 987
+ },
+ {
+ "epoch": 13.726637554585153,
+ "grad_norm": 0.017014775425195694,
+ "learning_rate": 0.0006,
+ "loss": 4.740612983703613,
+ "step": 988
+ },
+ {
+ "epoch": 13.74061135371179,
+ "grad_norm": 0.015199844725430012,
+ "learning_rate": 0.0006,
+ "loss": 4.719701766967773,
+ "step": 989
+ },
+ {
+ "epoch": 13.754585152838429,
+ "grad_norm": 0.015125414356589317,
+ "learning_rate": 0.0006,
+ "loss": 4.686114311218262,
+ "step": 990
+ },
+ {
+ "epoch": 13.768558951965066,
+ "grad_norm": 0.014779540710151196,
+ "learning_rate": 0.0006,
+ "loss": 4.674347877502441,
+ "step": 991
+ },
+ {
+ "epoch": 13.782532751091702,
+ "grad_norm": 0.015255163423717022,
+ "learning_rate": 0.0006,
+ "loss": 4.6309309005737305,
+ "step": 992
+ },
+ {
+ "epoch": 13.796506550218341,
+ "grad_norm": 0.013692197389900684,
+ "learning_rate": 0.0006,
+ "loss": 4.698497295379639,
+ "step": 993
+ },
+ {
+ "epoch": 13.810480349344978,
+ "grad_norm": 0.014529623091220856,
+ "learning_rate": 0.0006,
+ "loss": 4.628443717956543,
+ "step": 994
+ },
+ {
+ "epoch": 13.824454148471617,
+ "grad_norm": 0.01576659083366394,
+ "learning_rate": 0.0006,
+ "loss": 4.788015842437744,
+ "step": 995
+ },
+ {
+ "epoch": 13.838427947598253,
+ "grad_norm": 0.015834277495741844,
+ "learning_rate": 0.0006,
+ "loss": 4.672703742980957,
+ "step": 996
+ },
+ {
+ "epoch": 13.85240174672489,
+ "grad_norm": 0.015539503656327724,
+ "learning_rate": 0.0006,
+ "loss": 4.731503486633301,
+ "step": 997
+ },
+ {
+ "epoch": 13.866375545851529,
+ "grad_norm": 0.013463746756315231,
+ "learning_rate": 0.0006,
+ "loss": 4.7006330490112305,
+ "step": 998
+ },
+ {
+ "epoch": 13.880349344978166,
+ "grad_norm": 0.016377726569771767,
+ "learning_rate": 0.0006,
+ "loss": 4.636074066162109,
+ "step": 999
+ },
+ {
+ "epoch": 13.894323144104803,
+ "grad_norm": 0.016660034656524658,
+ "learning_rate": 0.0006,
+ "loss": 4.742693901062012,
+ "step": 1000
+ },
+ {
+ "epoch": 13.908296943231441,
+ "grad_norm": 0.014902938157320023,
+ "learning_rate": 0.0006,
+ "loss": 4.769811630249023,
+ "step": 1001
+ },
+ {
+ "epoch": 13.922270742358078,
+ "grad_norm": 0.014883476309478283,
+ "learning_rate": 0.0006,
+ "loss": 4.731159210205078,
+ "step": 1002
+ },
+ {
+ "epoch": 13.936244541484717,
+ "grad_norm": 0.014494217932224274,
+ "learning_rate": 0.0006,
+ "loss": 4.659485816955566,
+ "step": 1003
+ },
+ {
+ "epoch": 13.950218340611354,
+ "grad_norm": 0.015426991507411003,
+ "learning_rate": 0.0006,
+ "loss": 4.755229949951172,
+ "step": 1004
+ },
+ {
+ "epoch": 13.96419213973799,
+ "grad_norm": 0.016648586839437485,
+ "learning_rate": 0.0006,
+ "loss": 4.74111270904541,
+ "step": 1005
+ },
+ {
+ "epoch": 13.97816593886463,
+ "grad_norm": 0.016980541869997978,
+ "learning_rate": 0.0006,
+ "loss": 4.74072265625,
+ "step": 1006
+ },
+ {
+ "epoch": 13.992139737991266,
+ "grad_norm": 0.018070152029395103,
+ "learning_rate": 0.0006,
+ "loss": 4.732975006103516,
+ "step": 1007
+ },
+ {
+ "epoch": 14.0,
+ "grad_norm": 0.019910665228962898,
+ "learning_rate": 0.0006,
+ "loss": 4.634697914123535,
+ "step": 1008
+ },
+ {
+ "epoch": 14.0,
+ "eval_loss": 4.850640296936035,
+ "eval_runtime": 57.3039,
+ "eval_samples_per_second": 42.615,
+ "eval_steps_per_second": 1.344,
+ "step": 1008
+ },
+ {
+ "epoch": 14.013973799126637,
+ "grad_norm": 0.01980053074657917,
+ "learning_rate": 0.0006,
+ "loss": 4.628744602203369,
+ "step": 1009
+ },
+ {
+ "epoch": 14.027947598253276,
+ "grad_norm": 0.02497623674571514,
+ "learning_rate": 0.0006,
+ "loss": 4.75930118560791,
+ "step": 1010
+ },
+ {
+ "epoch": 14.041921397379912,
+ "grad_norm": 0.02794097363948822,
+ "learning_rate": 0.0006,
+ "loss": 4.691340446472168,
+ "step": 1011
+ },
+ {
+ "epoch": 14.055895196506551,
+ "grad_norm": 0.023725535720586777,
+ "learning_rate": 0.0006,
+ "loss": 4.733664512634277,
+ "step": 1012
+ },
+ {
+ "epoch": 14.069868995633188,
+ "grad_norm": 0.022966474294662476,
+ "learning_rate": 0.0006,
+ "loss": 4.664808750152588,
+ "step": 1013
+ },
+ {
+ "epoch": 14.083842794759825,
+ "grad_norm": 0.02530563250184059,
+ "learning_rate": 0.0006,
+ "loss": 4.8047285079956055,
+ "step": 1014
+ },
+ {
+ "epoch": 14.097816593886463,
+ "grad_norm": 0.02491082064807415,
+ "learning_rate": 0.0006,
+ "loss": 4.667956352233887,
+ "step": 1015
+ },
+ {
+ "epoch": 14.1117903930131,
+ "grad_norm": 0.022110294550657272,
+ "learning_rate": 0.0006,
+ "loss": 4.666140556335449,
+ "step": 1016
+ },
+ {
+ "epoch": 14.125764192139737,
+ "grad_norm": 0.02065849117934704,
+ "learning_rate": 0.0006,
+ "loss": 4.581960678100586,
+ "step": 1017
+ },
+ {
+ "epoch": 14.139737991266376,
+ "grad_norm": 0.01941298507153988,
+ "learning_rate": 0.0006,
+ "loss": 4.614226341247559,
+ "step": 1018
+ },
+ {
+ "epoch": 14.153711790393013,
+ "grad_norm": 0.022441856563091278,
+ "learning_rate": 0.0006,
+ "loss": 4.638352394104004,
+ "step": 1019
+ },
+ {
+ "epoch": 14.167685589519651,
+ "grad_norm": 0.021519694477319717,
+ "learning_rate": 0.0006,
+ "loss": 4.700160980224609,
+ "step": 1020
+ },
+ {
+ "epoch": 14.181659388646288,
+ "grad_norm": 0.018956486135721207,
+ "learning_rate": 0.0006,
+ "loss": 4.649593353271484,
+ "step": 1021
+ },
+ {
+ "epoch": 14.195633187772925,
+ "grad_norm": 0.016681140288710594,
+ "learning_rate": 0.0006,
+ "loss": 4.603952407836914,
+ "step": 1022
+ },
+ {
+ "epoch": 14.209606986899564,
+ "grad_norm": 0.01788809522986412,
+ "learning_rate": 0.0006,
+ "loss": 4.669160842895508,
+ "step": 1023
+ },
+ {
+ "epoch": 14.2235807860262,
+ "grad_norm": 0.016047228127717972,
+ "learning_rate": 0.0006,
+ "loss": 4.6367950439453125,
+ "step": 1024
+ },
+ {
+ "epoch": 14.237554585152838,
+ "grad_norm": 0.01614239625632763,
+ "learning_rate": 0.0006,
+ "loss": 4.664529323577881,
+ "step": 1025
+ },
+ {
+ "epoch": 14.251528384279476,
+ "grad_norm": 0.016362549737095833,
+ "learning_rate": 0.0006,
+ "loss": 4.702373504638672,
+ "step": 1026
+ },
+ {
+ "epoch": 14.265502183406113,
+ "grad_norm": 0.015800409018993378,
+ "learning_rate": 0.0006,
+ "loss": 4.7594428062438965,
+ "step": 1027
+ },
+ {
+ "epoch": 14.279475982532752,
+ "grad_norm": 0.016730139032006264,
+ "learning_rate": 0.0006,
+ "loss": 4.677244186401367,
+ "step": 1028
+ },
+ {
+ "epoch": 14.293449781659389,
+ "grad_norm": 0.01661759801208973,
+ "learning_rate": 0.0006,
+ "loss": 4.633679389953613,
+ "step": 1029
+ },
+ {
+ "epoch": 14.307423580786025,
+ "grad_norm": 0.017673347145318985,
+ "learning_rate": 0.0006,
+ "loss": 4.649107456207275,
+ "step": 1030
+ },
+ {
+ "epoch": 14.321397379912664,
+ "grad_norm": 0.019243352115154266,
+ "learning_rate": 0.0006,
+ "loss": 4.611257553100586,
+ "step": 1031
+ },
+ {
+ "epoch": 14.335371179039301,
+ "grad_norm": 0.019817164167761803,
+ "learning_rate": 0.0006,
+ "loss": 4.588086128234863,
+ "step": 1032
+ },
+ {
+ "epoch": 14.34934497816594,
+ "grad_norm": 0.022292211651802063,
+ "learning_rate": 0.0006,
+ "loss": 4.783066749572754,
+ "step": 1033
+ },
+ {
+ "epoch": 14.363318777292577,
+ "grad_norm": 0.02140096202492714,
+ "learning_rate": 0.0006,
+ "loss": 4.604718208312988,
+ "step": 1034
+ },
+ {
+ "epoch": 14.377292576419213,
+ "grad_norm": 0.0187817495316267,
+ "learning_rate": 0.0006,
+ "loss": 4.721773147583008,
+ "step": 1035
+ },
+ {
+ "epoch": 14.391266375545852,
+ "grad_norm": 0.019879210740327835,
+ "learning_rate": 0.0006,
+ "loss": 4.789286136627197,
+ "step": 1036
+ },
+ {
+ "epoch": 14.405240174672489,
+ "grad_norm": 0.01885785162448883,
+ "learning_rate": 0.0006,
+ "loss": 4.669743537902832,
+ "step": 1037
+ },
+ {
+ "epoch": 14.419213973799126,
+ "grad_norm": 0.018324270844459534,
+ "learning_rate": 0.0006,
+ "loss": 4.65901517868042,
+ "step": 1038
+ },
+ {
+ "epoch": 14.433187772925764,
+ "grad_norm": 0.017664821818470955,
+ "learning_rate": 0.0006,
+ "loss": 4.667839527130127,
+ "step": 1039
+ },
+ {
+ "epoch": 14.447161572052401,
+ "grad_norm": 0.015702905133366585,
+ "learning_rate": 0.0006,
+ "loss": 4.64700984954834,
+ "step": 1040
+ },
+ {
+ "epoch": 14.46113537117904,
+ "grad_norm": 0.016674194484949112,
+ "learning_rate": 0.0006,
+ "loss": 4.694093227386475,
+ "step": 1041
+ },
+ {
+ "epoch": 14.475109170305677,
+ "grad_norm": 0.016869477927684784,
+ "learning_rate": 0.0006,
+ "loss": 4.619544982910156,
+ "step": 1042
+ },
+ {
+ "epoch": 14.489082969432314,
+ "grad_norm": 0.016557540744543076,
+ "learning_rate": 0.0006,
+ "loss": 4.568667411804199,
+ "step": 1043
+ },
+ {
+ "epoch": 14.503056768558952,
+ "grad_norm": 0.014500945806503296,
+ "learning_rate": 0.0006,
+ "loss": 4.561741828918457,
+ "step": 1044
+ },
+ {
+ "epoch": 14.51703056768559,
+ "grad_norm": 0.014228739775717258,
+ "learning_rate": 0.0006,
+ "loss": 4.654277324676514,
+ "step": 1045
+ },
+ {
+ "epoch": 14.531004366812226,
+ "grad_norm": 0.013894238509237766,
+ "learning_rate": 0.0006,
+ "loss": 4.5867719650268555,
+ "step": 1046
+ },
+ {
+ "epoch": 14.544978165938865,
+ "grad_norm": 0.01550068985670805,
+ "learning_rate": 0.0006,
+ "loss": 4.61168098449707,
+ "step": 1047
+ },
+ {
+ "epoch": 14.558951965065502,
+ "grad_norm": 0.015940772369503975,
+ "learning_rate": 0.0006,
+ "loss": 4.691247940063477,
+ "step": 1048
+ },
+ {
+ "epoch": 14.57292576419214,
+ "grad_norm": 0.01714535616338253,
+ "learning_rate": 0.0006,
+ "loss": 4.61183500289917,
+ "step": 1049
+ },
+ {
+ "epoch": 14.586899563318777,
+ "grad_norm": 0.017925038933753967,
+ "learning_rate": 0.0006,
+ "loss": 4.719022750854492,
+ "step": 1050
+ },
+ {
+ "epoch": 14.600873362445414,
+ "grad_norm": 0.018141552805900574,
+ "learning_rate": 0.0006,
+ "loss": 4.760281562805176,
+ "step": 1051
+ },
+ {
+ "epoch": 14.614847161572053,
+ "grad_norm": 0.017131328582763672,
+ "learning_rate": 0.0006,
+ "loss": 4.709680080413818,
+ "step": 1052
+ },
+ {
+ "epoch": 14.62882096069869,
+ "grad_norm": 0.018730048090219498,
+ "learning_rate": 0.0006,
+ "loss": 4.5994720458984375,
+ "step": 1053
+ },
+ {
+ "epoch": 14.642794759825328,
+ "grad_norm": 0.018894298002123833,
+ "learning_rate": 0.0006,
+ "loss": 4.669070720672607,
+ "step": 1054
+ },
+ {
+ "epoch": 14.656768558951965,
+ "grad_norm": 0.017643198370933533,
+ "learning_rate": 0.0006,
+ "loss": 4.641790390014648,
+ "step": 1055
+ },
+ {
+ "epoch": 14.670742358078602,
+ "grad_norm": 0.017647244036197662,
+ "learning_rate": 0.0006,
+ "loss": 4.653899192810059,
+ "step": 1056
+ },
+ {
+ "epoch": 14.68471615720524,
+ "grad_norm": 0.017304198816418648,
+ "learning_rate": 0.0006,
+ "loss": 4.7487473487854,
+ "step": 1057
+ },
+ {
+ "epoch": 14.698689956331878,
+ "grad_norm": 0.01638447679579258,
+ "learning_rate": 0.0006,
+ "loss": 4.71945858001709,
+ "step": 1058
+ },
+ {
+ "epoch": 14.712663755458514,
+ "grad_norm": 0.015892071649432182,
+ "learning_rate": 0.0006,
+ "loss": 4.744903564453125,
+ "step": 1059
+ },
+ {
+ "epoch": 14.726637554585153,
+ "grad_norm": 0.017420614138245583,
+ "learning_rate": 0.0006,
+ "loss": 4.781563758850098,
+ "step": 1060
+ },
+ {
+ "epoch": 14.74061135371179,
+ "grad_norm": 0.016498520970344543,
+ "learning_rate": 0.0006,
+ "loss": 4.622779846191406,
+ "step": 1061
+ },
+ {
+ "epoch": 14.754585152838429,
+ "grad_norm": 0.017482127994298935,
+ "learning_rate": 0.0006,
+ "loss": 4.62825870513916,
+ "step": 1062
+ },
+ {
+ "epoch": 14.768558951965066,
+ "grad_norm": 0.019018379971385002,
+ "learning_rate": 0.0006,
+ "loss": 4.693194389343262,
+ "step": 1063
+ },
+ {
+ "epoch": 14.782532751091702,
+ "grad_norm": 0.01993914321064949,
+ "learning_rate": 0.0006,
+ "loss": 4.742903709411621,
+ "step": 1064
+ },
+ {
+ "epoch": 14.796506550218341,
+ "grad_norm": 0.02301335707306862,
+ "learning_rate": 0.0006,
+ "loss": 4.613251686096191,
+ "step": 1065
+ },
+ {
+ "epoch": 14.810480349344978,
+ "grad_norm": 0.024088222533464432,
+ "learning_rate": 0.0006,
+ "loss": 4.594305992126465,
+ "step": 1066
+ },
+ {
+ "epoch": 14.824454148471617,
+ "grad_norm": 0.02493320405483246,
+ "learning_rate": 0.0006,
+ "loss": 4.700588226318359,
+ "step": 1067
+ },
+ {
+ "epoch": 14.838427947598253,
+ "grad_norm": 0.02501937560737133,
+ "learning_rate": 0.0006,
+ "loss": 4.601742744445801,
+ "step": 1068
+ },
+ {
+ "epoch": 14.85240174672489,
+ "grad_norm": 0.023490751162171364,
+ "learning_rate": 0.0006,
+ "loss": 4.589211463928223,
+ "step": 1069
+ },
+ {
+ "epoch": 14.866375545851529,
+ "grad_norm": 0.019695665687322617,
+ "learning_rate": 0.0006,
+ "loss": 4.635081768035889,
+ "step": 1070
+ },
+ {
+ "epoch": 14.880349344978166,
+ "grad_norm": 0.019276980310678482,
+ "learning_rate": 0.0006,
+ "loss": 4.737002372741699,
+ "step": 1071
+ },
+ {
+ "epoch": 14.894323144104803,
+ "grad_norm": 0.022430511191487312,
+ "learning_rate": 0.0006,
+ "loss": 4.808295726776123,
+ "step": 1072
+ },
+ {
+ "epoch": 14.908296943231441,
+ "grad_norm": 0.021543916314840317,
+ "learning_rate": 0.0006,
+ "loss": 4.559475421905518,
+ "step": 1073
+ },
+ {
+ "epoch": 14.922270742358078,
+ "grad_norm": 0.019885241985321045,
+ "learning_rate": 0.0006,
+ "loss": 4.647948741912842,
+ "step": 1074
+ },
+ {
+ "epoch": 14.936244541484717,
+ "grad_norm": 0.02115175500512123,
+ "learning_rate": 0.0006,
+ "loss": 4.680505752563477,
+ "step": 1075
+ },
+ {
+ "epoch": 14.950218340611354,
+ "grad_norm": 0.02156151458621025,
+ "learning_rate": 0.0006,
+ "loss": 4.673477649688721,
+ "step": 1076
+ },
+ {
+ "epoch": 14.96419213973799,
+ "grad_norm": 0.02184601128101349,
+ "learning_rate": 0.0006,
+ "loss": 4.644402027130127,
+ "step": 1077
+ },
+ {
+ "epoch": 14.97816593886463,
+ "grad_norm": 0.0211983360350132,
+ "learning_rate": 0.0006,
+ "loss": 4.821558475494385,
+ "step": 1078
+ },
+ {
+ "epoch": 14.992139737991266,
+ "grad_norm": 0.019636567682027817,
+ "learning_rate": 0.0006,
+ "loss": 4.578868865966797,
+ "step": 1079
+ },
+ {
+ "epoch": 15.0,
+ "grad_norm": 0.019997350871562958,
+ "learning_rate": 0.0006,
+ "loss": 4.691307544708252,
+ "step": 1080
+ },
+ {
+ "epoch": 15.0,
+ "eval_loss": 4.872478008270264,
+ "eval_runtime": 56.8513,
+ "eval_samples_per_second": 42.954,
+ "eval_steps_per_second": 1.354,
+ "step": 1080
+ },
+ {
+ "epoch": 15.013973799126637,
+ "grad_norm": 0.018830662593245506,
+ "learning_rate": 0.0006,
+ "loss": 4.761779308319092,
+ "step": 1081
+ },
+ {
+ "epoch": 15.027947598253276,
+ "grad_norm": 0.016876481473445892,
+ "learning_rate": 0.0006,
+ "loss": 4.61148738861084,
+ "step": 1082
+ },
+ {
+ "epoch": 15.041921397379912,
+ "grad_norm": 0.015550863929092884,
+ "learning_rate": 0.0006,
+ "loss": 4.676124095916748,
+ "step": 1083
+ },
+ {
+ "epoch": 15.055895196506551,
+ "grad_norm": 0.017178896814584732,
+ "learning_rate": 0.0006,
+ "loss": 4.588935852050781,
+ "step": 1084
+ },
+ {
+ "epoch": 15.069868995633188,
+ "grad_norm": 0.015465447679162025,
+ "learning_rate": 0.0006,
+ "loss": 4.605499267578125,
+ "step": 1085
+ },
+ {
+ "epoch": 15.083842794759825,
+ "grad_norm": 0.01606195978820324,
+ "learning_rate": 0.0006,
+ "loss": 4.723373889923096,
+ "step": 1086
+ },
+ {
+ "epoch": 15.097816593886463,
+ "grad_norm": 0.016521936282515526,
+ "learning_rate": 0.0006,
+ "loss": 4.646608352661133,
+ "step": 1087
+ },
+ {
+ "epoch": 15.1117903930131,
+ "grad_norm": 0.01589205674827099,
+ "learning_rate": 0.0006,
+ "loss": 4.498138427734375,
+ "step": 1088
+ },
+ {
+ "epoch": 15.125764192139737,
+ "grad_norm": 0.01524856686592102,
+ "learning_rate": 0.0006,
+ "loss": 4.594562530517578,
+ "step": 1089
+ },
+ {
+ "epoch": 15.139737991266376,
+ "grad_norm": 0.014714641496539116,
+ "learning_rate": 0.0006,
+ "loss": 4.630744934082031,
+ "step": 1090
+ },
+ {
+ "epoch": 15.153711790393013,
+ "grad_norm": 0.015333331190049648,
+ "learning_rate": 0.0006,
+ "loss": 4.682388782501221,
+ "step": 1091
+ },
+ {
+ "epoch": 15.167685589519651,
+ "grad_norm": 0.01550073828548193,
+ "learning_rate": 0.0006,
+ "loss": 4.692652702331543,
+ "step": 1092
+ },
+ {
+ "epoch": 15.181659388646288,
+ "grad_norm": 0.014783318154513836,
+ "learning_rate": 0.0006,
+ "loss": 4.6223835945129395,
+ "step": 1093
+ },
+ {
+ "epoch": 15.195633187772925,
+ "grad_norm": 0.014372751116752625,
+ "learning_rate": 0.0006,
+ "loss": 4.52687406539917,
+ "step": 1094
+ },
+ {
+ "epoch": 15.209606986899564,
+ "grad_norm": 0.014779130928218365,
+ "learning_rate": 0.0006,
+ "loss": 4.637009620666504,
+ "step": 1095
+ },
+ {
+ "epoch": 15.2235807860262,
+ "grad_norm": 0.015697073191404343,
+ "learning_rate": 0.0006,
+ "loss": 4.682981014251709,
+ "step": 1096
+ },
+ {
+ "epoch": 15.237554585152838,
+ "grad_norm": 0.016523664817214012,
+ "learning_rate": 0.0006,
+ "loss": 4.705722808837891,
+ "step": 1097
+ },
+ {
+ "epoch": 15.251528384279476,
+ "grad_norm": 0.017633996903896332,
+ "learning_rate": 0.0006,
+ "loss": 4.574686050415039,
+ "step": 1098
+ },
+ {
+ "epoch": 15.265502183406113,
+ "grad_norm": 0.019955076277256012,
+ "learning_rate": 0.0006,
+ "loss": 4.634450912475586,
+ "step": 1099
+ },
+ {
+ "epoch": 15.279475982532752,
+ "grad_norm": 0.02132618986070156,
+ "learning_rate": 0.0006,
+ "loss": 4.56110143661499,
+ "step": 1100
+ },
+ {
+ "epoch": 15.293449781659389,
+ "grad_norm": 0.0207593385130167,
+ "learning_rate": 0.0006,
+ "loss": 4.629396438598633,
+ "step": 1101
+ },
+ {
+ "epoch": 15.307423580786025,
+ "grad_norm": 0.017882540822029114,
+ "learning_rate": 0.0006,
+ "loss": 4.730963230133057,
+ "step": 1102
+ },
+ {
+ "epoch": 15.321397379912664,
+ "grad_norm": 0.01917876861989498,
+ "learning_rate": 0.0006,
+ "loss": 4.64585542678833,
+ "step": 1103
+ },
+ {
+ "epoch": 15.335371179039301,
+ "grad_norm": 0.01943974196910858,
+ "learning_rate": 0.0006,
+ "loss": 4.656181335449219,
+ "step": 1104
+ },
+ {
+ "epoch": 15.34934497816594,
+ "grad_norm": 0.019904715940356255,
+ "learning_rate": 0.0006,
+ "loss": 4.619329452514648,
+ "step": 1105
+ },
+ {
+ "epoch": 15.363318777292577,
+ "grad_norm": 0.019174346700310707,
+ "learning_rate": 0.0006,
+ "loss": 4.647892475128174,
+ "step": 1106
+ },
+ {
+ "epoch": 15.377292576419213,
+ "grad_norm": 0.018244758248329163,
+ "learning_rate": 0.0006,
+ "loss": 4.56166934967041,
+ "step": 1107
+ },
+ {
+ "epoch": 15.391266375545852,
+ "grad_norm": 0.018914205953478813,
+ "learning_rate": 0.0006,
+ "loss": 4.648063659667969,
+ "step": 1108
+ },
+ {
+ "epoch": 15.405240174672489,
+ "grad_norm": 0.01961071416735649,
+ "learning_rate": 0.0006,
+ "loss": 4.665525436401367,
+ "step": 1109
+ },
+ {
+ "epoch": 15.419213973799126,
+ "grad_norm": 0.01841311901807785,
+ "learning_rate": 0.0006,
+ "loss": 4.611342430114746,
+ "step": 1110
+ },
+ {
+ "epoch": 15.433187772925764,
+ "grad_norm": 0.01863427273929119,
+ "learning_rate": 0.0006,
+ "loss": 4.664256572723389,
+ "step": 1111
+ },
+ {
+ "epoch": 15.447161572052401,
+ "grad_norm": 0.020506877452135086,
+ "learning_rate": 0.0006,
+ "loss": 4.631237506866455,
+ "step": 1112
+ },
+ {
+ "epoch": 15.46113537117904,
+ "grad_norm": 0.02223833277821541,
+ "learning_rate": 0.0006,
+ "loss": 4.668654441833496,
+ "step": 1113
+ },
+ {
+ "epoch": 15.475109170305677,
+ "grad_norm": 0.023336702957749367,
+ "learning_rate": 0.0006,
+ "loss": 4.562833309173584,
+ "step": 1114
+ },
+ {
+ "epoch": 15.489082969432314,
+ "grad_norm": 0.023052413016557693,
+ "learning_rate": 0.0006,
+ "loss": 4.621532917022705,
+ "step": 1115
+ },
+ {
+ "epoch": 15.503056768558952,
+ "grad_norm": 0.023420924320816994,
+ "learning_rate": 0.0006,
+ "loss": 4.665465354919434,
+ "step": 1116
+ },
+ {
+ "epoch": 15.51703056768559,
+ "grad_norm": 0.024071281775832176,
+ "learning_rate": 0.0006,
+ "loss": 4.595980644226074,
+ "step": 1117
+ },
+ {
+ "epoch": 15.531004366812226,
+ "grad_norm": 0.023406682536005974,
+ "learning_rate": 0.0006,
+ "loss": 4.642635345458984,
+ "step": 1118
+ },
+ {
+ "epoch": 15.544978165938865,
+ "grad_norm": 0.02138647809624672,
+ "learning_rate": 0.0006,
+ "loss": 4.704513072967529,
+ "step": 1119
+ },
+ {
+ "epoch": 15.558951965065502,
+ "grad_norm": 0.0221384447067976,
+ "learning_rate": 0.0006,
+ "loss": 4.567503452301025,
+ "step": 1120
+ },
+ {
+ "epoch": 15.57292576419214,
+ "grad_norm": 0.020879851654171944,
+ "learning_rate": 0.0006,
+ "loss": 4.664057731628418,
+ "step": 1121
+ },
+ {
+ "epoch": 15.586899563318777,
+ "grad_norm": 0.021478967741131783,
+ "learning_rate": 0.0006,
+ "loss": 4.539627552032471,
+ "step": 1122
+ },
+ {
+ "epoch": 15.600873362445414,
+ "grad_norm": 0.022748656570911407,
+ "learning_rate": 0.0006,
+ "loss": 4.5654497146606445,
+ "step": 1123
+ },
+ {
+ "epoch": 15.614847161572053,
+ "grad_norm": 0.021870248019695282,
+ "learning_rate": 0.0006,
+ "loss": 4.56899356842041,
+ "step": 1124
+ },
+ {
+ "epoch": 15.62882096069869,
+ "grad_norm": 0.021711355075240135,
+ "learning_rate": 0.0006,
+ "loss": 4.657808303833008,
+ "step": 1125
+ },
+ {
+ "epoch": 15.642794759825328,
+ "grad_norm": 0.022810909897089005,
+ "learning_rate": 0.0006,
+ "loss": 4.6526689529418945,
+ "step": 1126
+ },
+ {
+ "epoch": 15.656768558951965,
+ "grad_norm": 0.022561343386769295,
+ "learning_rate": 0.0006,
+ "loss": 4.534887790679932,
+ "step": 1127
+ },
+ {
+ "epoch": 15.670742358078602,
+ "grad_norm": 0.019909802824258804,
+ "learning_rate": 0.0006,
+ "loss": 4.600790023803711,
+ "step": 1128
+ },
+ {
+ "epoch": 15.68471615720524,
+ "grad_norm": 0.020092325285077095,
+ "learning_rate": 0.0006,
+ "loss": 4.773830413818359,
+ "step": 1129
+ },
+ {
+ "epoch": 15.698689956331878,
+ "grad_norm": 0.018834445625543594,
+ "learning_rate": 0.0006,
+ "loss": 4.628868103027344,
+ "step": 1130
+ },
+ {
+ "epoch": 15.712663755458514,
+ "grad_norm": 0.017552798613905907,
+ "learning_rate": 0.0006,
+ "loss": 4.647500038146973,
+ "step": 1131
+ },
+ {
+ "epoch": 15.726637554585153,
+ "grad_norm": 0.01751251146197319,
+ "learning_rate": 0.0006,
+ "loss": 4.590545654296875,
+ "step": 1132
+ },
+ {
+ "epoch": 15.74061135371179,
+ "grad_norm": 0.01699099875986576,
+ "learning_rate": 0.0006,
+ "loss": 4.509699821472168,
+ "step": 1133
+ },
+ {
+ "epoch": 15.754585152838429,
+ "grad_norm": 0.015808649361133575,
+ "learning_rate": 0.0006,
+ "loss": 4.606992244720459,
+ "step": 1134
+ },
+ {
+ "epoch": 15.768558951965066,
+ "grad_norm": 0.01655443385243416,
+ "learning_rate": 0.0006,
+ "loss": 4.585219860076904,
+ "step": 1135
+ },
+ {
+ "epoch": 15.782532751091702,
+ "grad_norm": 0.01708163693547249,
+ "learning_rate": 0.0006,
+ "loss": 4.668147087097168,
+ "step": 1136
+ },
+ {
+ "epoch": 15.796506550218341,
+ "grad_norm": 0.017161913216114044,
+ "learning_rate": 0.0006,
+ "loss": 4.7170634269714355,
+ "step": 1137
+ },
+ {
+ "epoch": 15.810480349344978,
+ "grad_norm": 0.017262674868106842,
+ "learning_rate": 0.0006,
+ "loss": 4.572132110595703,
+ "step": 1138
+ },
+ {
+ "epoch": 15.824454148471617,
+ "grad_norm": 0.017713576555252075,
+ "learning_rate": 0.0006,
+ "loss": 4.702538967132568,
+ "step": 1139
+ },
+ {
+ "epoch": 15.838427947598253,
+ "grad_norm": 0.018730733543634415,
+ "learning_rate": 0.0006,
+ "loss": 4.629721641540527,
+ "step": 1140
+ },
+ {
+ "epoch": 15.85240174672489,
+ "grad_norm": 0.017960187047719955,
+ "learning_rate": 0.0006,
+ "loss": 4.661039352416992,
+ "step": 1141
+ },
+ {
+ "epoch": 15.866375545851529,
+ "grad_norm": 0.01750079356133938,
+ "learning_rate": 0.0006,
+ "loss": 4.503710746765137,
+ "step": 1142
+ },
+ {
+ "epoch": 15.880349344978166,
+ "grad_norm": 0.016630422323942184,
+ "learning_rate": 0.0006,
+ "loss": 4.6582136154174805,
+ "step": 1143
+ },
+ {
+ "epoch": 15.894323144104803,
+ "grad_norm": 0.015896832570433617,
+ "learning_rate": 0.0006,
+ "loss": 4.596693992614746,
+ "step": 1144
+ },
+ {
+ "epoch": 15.908296943231441,
+ "grad_norm": 0.01566407084465027,
+ "learning_rate": 0.0006,
+ "loss": 4.6337199211120605,
+ "step": 1145
+ },
+ {
+ "epoch": 15.922270742358078,
+ "grad_norm": 0.016823330894112587,
+ "learning_rate": 0.0006,
+ "loss": 4.686994552612305,
+ "step": 1146
+ },
+ {
+ "epoch": 15.936244541484717,
+ "grad_norm": 0.016208255663514137,
+ "learning_rate": 0.0006,
+ "loss": 4.6205549240112305,
+ "step": 1147
+ },
+ {
+ "epoch": 15.950218340611354,
+ "grad_norm": 0.0171707421541214,
+ "learning_rate": 0.0006,
+ "loss": 4.582968711853027,
+ "step": 1148
+ },
+ {
+ "epoch": 15.96419213973799,
+ "grad_norm": 0.017009446397423744,
+ "learning_rate": 0.0006,
+ "loss": 4.510274887084961,
+ "step": 1149
+ },
+ {
+ "epoch": 15.97816593886463,
+ "grad_norm": 0.01565811224281788,
+ "learning_rate": 0.0006,
+ "loss": 4.55834436416626,
+ "step": 1150
+ },
+ {
+ "epoch": 15.992139737991266,
+ "grad_norm": 0.015276037156581879,
+ "learning_rate": 0.0006,
+ "loss": 4.537870407104492,
+ "step": 1151
+ },
+ {
+ "epoch": 16.0,
+ "grad_norm": 0.017305882647633553,
+ "learning_rate": 0.0006,
+ "loss": 4.611305236816406,
+ "step": 1152
+ },
+ {
+ "epoch": 16.0,
+ "eval_loss": 4.775513172149658,
+ "eval_runtime": 56.7341,
+ "eval_samples_per_second": 43.043,
+ "eval_steps_per_second": 1.357,
+ "step": 1152
+ },
+ {
+ "epoch": 16.01397379912664,
+ "grad_norm": 0.016752436757087708,
+ "learning_rate": 0.0006,
+ "loss": 4.541312217712402,
+ "step": 1153
+ },
+ {
+ "epoch": 16.027947598253274,
+ "grad_norm": 0.01643134094774723,
+ "learning_rate": 0.0006,
+ "loss": 4.585402965545654,
+ "step": 1154
+ },
+ {
+ "epoch": 16.041921397379912,
+ "grad_norm": 0.01888720504939556,
+ "learning_rate": 0.0006,
+ "loss": 4.597825050354004,
+ "step": 1155
+ },
+ {
+ "epoch": 16.05589519650655,
+ "grad_norm": 0.022044632583856583,
+ "learning_rate": 0.0006,
+ "loss": 4.532772064208984,
+ "step": 1156
+ },
+ {
+ "epoch": 16.069868995633186,
+ "grad_norm": 0.022986987605690956,
+ "learning_rate": 0.0006,
+ "loss": 4.551351547241211,
+ "step": 1157
+ },
+ {
+ "epoch": 16.083842794759825,
+ "grad_norm": 0.020116791129112244,
+ "learning_rate": 0.0006,
+ "loss": 4.484149932861328,
+ "step": 1158
+ },
+ {
+ "epoch": 16.097816593886463,
+ "grad_norm": 0.021442247554659843,
+ "learning_rate": 0.0006,
+ "loss": 4.515316486358643,
+ "step": 1159
+ },
+ {
+ "epoch": 16.111790393013102,
+ "grad_norm": 0.020974772050976753,
+ "learning_rate": 0.0006,
+ "loss": 4.51606559753418,
+ "step": 1160
+ },
+ {
+ "epoch": 16.125764192139737,
+ "grad_norm": 0.021274864673614502,
+ "learning_rate": 0.0006,
+ "loss": 4.587329864501953,
+ "step": 1161
+ },
+ {
+ "epoch": 16.139737991266376,
+ "grad_norm": 0.022799784317612648,
+ "learning_rate": 0.0006,
+ "loss": 4.568149089813232,
+ "step": 1162
+ },
+ {
+ "epoch": 16.153711790393015,
+ "grad_norm": 0.021784933283925056,
+ "learning_rate": 0.0006,
+ "loss": 4.515336990356445,
+ "step": 1163
+ },
+ {
+ "epoch": 16.16768558951965,
+ "grad_norm": 0.024261271581053734,
+ "learning_rate": 0.0006,
+ "loss": 4.5127081871032715,
+ "step": 1164
+ },
+ {
+ "epoch": 16.18165938864629,
+ "grad_norm": 0.02248375490307808,
+ "learning_rate": 0.0006,
+ "loss": 4.754251003265381,
+ "step": 1165
+ },
+ {
+ "epoch": 16.195633187772927,
+ "grad_norm": 0.02136492356657982,
+ "learning_rate": 0.0006,
+ "loss": 4.58193826675415,
+ "step": 1166
+ },
+ {
+ "epoch": 16.209606986899562,
+ "grad_norm": 0.020727066323161125,
+ "learning_rate": 0.0006,
+ "loss": 4.564209938049316,
+ "step": 1167
+ },
+ {
+ "epoch": 16.2235807860262,
+ "grad_norm": 0.017789119854569435,
+ "learning_rate": 0.0006,
+ "loss": 4.698680877685547,
+ "step": 1168
+ },
+ {
+ "epoch": 16.23755458515284,
+ "grad_norm": 0.017603034153580666,
+ "learning_rate": 0.0006,
+ "loss": 4.501171588897705,
+ "step": 1169
+ },
+ {
+ "epoch": 16.251528384279474,
+ "grad_norm": 0.017345862463116646,
+ "learning_rate": 0.0006,
+ "loss": 4.593590259552002,
+ "step": 1170
+ },
+ {
+ "epoch": 16.265502183406113,
+ "grad_norm": 0.018047425895929337,
+ "learning_rate": 0.0006,
+ "loss": 4.597467422485352,
+ "step": 1171
+ },
+ {
+ "epoch": 16.27947598253275,
+ "grad_norm": 0.01947222836315632,
+ "learning_rate": 0.0006,
+ "loss": 4.5072221755981445,
+ "step": 1172
+ },
+ {
+ "epoch": 16.29344978165939,
+ "grad_norm": 0.019598500803112984,
+ "learning_rate": 0.0006,
+ "loss": 4.601351737976074,
+ "step": 1173
+ },
+ {
+ "epoch": 16.307423580786025,
+ "grad_norm": 0.020002514123916626,
+ "learning_rate": 0.0006,
+ "loss": 4.580431938171387,
+ "step": 1174
+ },
+ {
+ "epoch": 16.321397379912664,
+ "grad_norm": 0.0216530729085207,
+ "learning_rate": 0.0006,
+ "loss": 4.424681663513184,
+ "step": 1175
+ },
+ {
+ "epoch": 16.335371179039303,
+ "grad_norm": 0.01949235238134861,
+ "learning_rate": 0.0006,
+ "loss": 4.751883029937744,
+ "step": 1176
+ },
+ {
+ "epoch": 16.349344978165938,
+ "grad_norm": 0.017779016867280006,
+ "learning_rate": 0.0006,
+ "loss": 4.541490077972412,
+ "step": 1177
+ },
+ {
+ "epoch": 16.363318777292577,
+ "grad_norm": 0.01909121684730053,
+ "learning_rate": 0.0006,
+ "loss": 4.594141006469727,
+ "step": 1178
+ },
+ {
+ "epoch": 16.377292576419215,
+ "grad_norm": 0.019894849509000778,
+ "learning_rate": 0.0006,
+ "loss": 4.555070400238037,
+ "step": 1179
+ },
+ {
+ "epoch": 16.39126637554585,
+ "grad_norm": 0.01993032731115818,
+ "learning_rate": 0.0006,
+ "loss": 4.521401405334473,
+ "step": 1180
+ },
+ {
+ "epoch": 16.40524017467249,
+ "grad_norm": 0.0210788045078516,
+ "learning_rate": 0.0006,
+ "loss": 4.672842979431152,
+ "step": 1181
+ },
+ {
+ "epoch": 16.419213973799128,
+ "grad_norm": 0.02274598926305771,
+ "learning_rate": 0.0006,
+ "loss": 4.629504203796387,
+ "step": 1182
+ },
+ {
+ "epoch": 16.433187772925763,
+ "grad_norm": 0.020448120310902596,
+ "learning_rate": 0.0006,
+ "loss": 4.548209190368652,
+ "step": 1183
+ },
+ {
+ "epoch": 16.4471615720524,
+ "grad_norm": 0.017565643414855003,
+ "learning_rate": 0.0006,
+ "loss": 4.529054641723633,
+ "step": 1184
+ },
+ {
+ "epoch": 16.46113537117904,
+ "grad_norm": 0.019504237920045853,
+ "learning_rate": 0.0006,
+ "loss": 4.48182487487793,
+ "step": 1185
+ },
+ {
+ "epoch": 16.475109170305675,
+ "grad_norm": 0.016823742538690567,
+ "learning_rate": 0.0006,
+ "loss": 4.608696937561035,
+ "step": 1186
+ },
+ {
+ "epoch": 16.489082969432314,
+ "grad_norm": 0.01602894812822342,
+ "learning_rate": 0.0006,
+ "loss": 4.620062351226807,
+ "step": 1187
+ },
+ {
+ "epoch": 16.503056768558952,
+ "grad_norm": 0.016615642234683037,
+ "learning_rate": 0.0006,
+ "loss": 4.69378662109375,
+ "step": 1188
+ },
+ {
+ "epoch": 16.51703056768559,
+ "grad_norm": 0.018576730042696,
+ "learning_rate": 0.0006,
+ "loss": 4.537041664123535,
+ "step": 1189
+ },
+ {
+ "epoch": 16.531004366812226,
+ "grad_norm": 0.018611151725053787,
+ "learning_rate": 0.0006,
+ "loss": 4.496169090270996,
+ "step": 1190
+ },
+ {
+ "epoch": 16.544978165938865,
+ "grad_norm": 0.018622413277626038,
+ "learning_rate": 0.0006,
+ "loss": 4.557358264923096,
+ "step": 1191
+ },
+ {
+ "epoch": 16.558951965065503,
+ "grad_norm": 0.01823544315993786,
+ "learning_rate": 0.0006,
+ "loss": 4.730792999267578,
+ "step": 1192
+ },
+ {
+ "epoch": 16.57292576419214,
+ "grad_norm": 0.018796846270561218,
+ "learning_rate": 0.0006,
+ "loss": 4.781791687011719,
+ "step": 1193
+ },
+ {
+ "epoch": 16.586899563318777,
+ "grad_norm": 0.01728292554616928,
+ "learning_rate": 0.0006,
+ "loss": 4.515257835388184,
+ "step": 1194
+ },
+ {
+ "epoch": 16.600873362445416,
+ "grad_norm": 0.01699589006602764,
+ "learning_rate": 0.0006,
+ "loss": 4.5996904373168945,
+ "step": 1195
+ },
+ {
+ "epoch": 16.61484716157205,
+ "grad_norm": 0.015905970707535744,
+ "learning_rate": 0.0006,
+ "loss": 4.601301193237305,
+ "step": 1196
+ },
+ {
+ "epoch": 16.62882096069869,
+ "grad_norm": 0.01338571310043335,
+ "learning_rate": 0.0006,
+ "loss": 4.557708740234375,
+ "step": 1197
+ },
+ {
+ "epoch": 16.64279475982533,
+ "grad_norm": 0.015298855490982533,
+ "learning_rate": 0.0006,
+ "loss": 4.542287349700928,
+ "step": 1198
+ },
+ {
+ "epoch": 16.656768558951963,
+ "grad_norm": 0.014753367751836777,
+ "learning_rate": 0.0006,
+ "loss": 4.595787048339844,
+ "step": 1199
+ },
+ {
+ "epoch": 16.670742358078602,
+ "grad_norm": 0.015527483075857162,
+ "learning_rate": 0.0006,
+ "loss": 4.607529163360596,
+ "step": 1200
+ },
+ {
+ "epoch": 16.68471615720524,
+ "grad_norm": 0.015042162500321865,
+ "learning_rate": 0.0006,
+ "loss": 4.686626434326172,
+ "step": 1201
+ },
+ {
+ "epoch": 16.69868995633188,
+ "grad_norm": 0.0159393772482872,
+ "learning_rate": 0.0006,
+ "loss": 4.47656774520874,
+ "step": 1202
+ },
+ {
+ "epoch": 16.712663755458514,
+ "grad_norm": 0.01532787922769785,
+ "learning_rate": 0.0006,
+ "loss": 4.55678653717041,
+ "step": 1203
+ },
+ {
+ "epoch": 16.726637554585153,
+ "grad_norm": 0.016332164406776428,
+ "learning_rate": 0.0006,
+ "loss": 4.51441764831543,
+ "step": 1204
+ },
+ {
+ "epoch": 16.74061135371179,
+ "grad_norm": 0.018901726230978966,
+ "learning_rate": 0.0006,
+ "loss": 4.57227897644043,
+ "step": 1205
+ },
+ {
+ "epoch": 16.754585152838427,
+ "grad_norm": 0.018589433282613754,
+ "learning_rate": 0.0006,
+ "loss": 4.550070762634277,
+ "step": 1206
+ },
+ {
+ "epoch": 16.768558951965066,
+ "grad_norm": 0.01634034886956215,
+ "learning_rate": 0.0006,
+ "loss": 4.586910724639893,
+ "step": 1207
+ },
+ {
+ "epoch": 16.782532751091704,
+ "grad_norm": 0.017599433660507202,
+ "learning_rate": 0.0006,
+ "loss": 4.5891923904418945,
+ "step": 1208
+ },
+ {
+ "epoch": 16.79650655021834,
+ "grad_norm": 0.01878446340560913,
+ "learning_rate": 0.0006,
+ "loss": 4.57227087020874,
+ "step": 1209
+ },
+ {
+ "epoch": 16.810480349344978,
+ "grad_norm": 0.01803469844162464,
+ "learning_rate": 0.0006,
+ "loss": 4.583845138549805,
+ "step": 1210
+ },
+ {
+ "epoch": 16.824454148471617,
+ "grad_norm": 0.018505144864320755,
+ "learning_rate": 0.0006,
+ "loss": 4.561151504516602,
+ "step": 1211
+ },
+ {
+ "epoch": 16.83842794759825,
+ "grad_norm": 0.01920081488788128,
+ "learning_rate": 0.0006,
+ "loss": 4.506834030151367,
+ "step": 1212
+ },
+ {
+ "epoch": 16.85240174672489,
+ "grad_norm": 0.0177712831646204,
+ "learning_rate": 0.0006,
+ "loss": 4.557195663452148,
+ "step": 1213
+ },
+ {
+ "epoch": 16.86637554585153,
+ "grad_norm": 0.019113918766379356,
+ "learning_rate": 0.0006,
+ "loss": 4.651227951049805,
+ "step": 1214
+ },
+ {
+ "epoch": 16.880349344978168,
+ "grad_norm": 0.02012539468705654,
+ "learning_rate": 0.0006,
+ "loss": 4.4952239990234375,
+ "step": 1215
+ },
+ {
+ "epoch": 16.894323144104803,
+ "grad_norm": 0.019993646070361137,
+ "learning_rate": 0.0006,
+ "loss": 4.517433166503906,
+ "step": 1216
+ },
+ {
+ "epoch": 16.90829694323144,
+ "grad_norm": 0.020041441544890404,
+ "learning_rate": 0.0006,
+ "loss": 4.489407539367676,
+ "step": 1217
+ },
+ {
+ "epoch": 16.92227074235808,
+ "grad_norm": 0.0183611661195755,
+ "learning_rate": 0.0006,
+ "loss": 4.5435638427734375,
+ "step": 1218
+ },
+ {
+ "epoch": 16.936244541484715,
+ "grad_norm": 0.019138429313898087,
+ "learning_rate": 0.0006,
+ "loss": 4.545949459075928,
+ "step": 1219
+ },
+ {
+ "epoch": 16.950218340611354,
+ "grad_norm": 0.018554266542196274,
+ "learning_rate": 0.0006,
+ "loss": 4.579620361328125,
+ "step": 1220
+ },
+ {
+ "epoch": 16.964192139737992,
+ "grad_norm": 0.0193545650690794,
+ "learning_rate": 0.0006,
+ "loss": 4.588255405426025,
+ "step": 1221
+ },
+ {
+ "epoch": 16.978165938864628,
+ "grad_norm": 0.019389165565371513,
+ "learning_rate": 0.0006,
+ "loss": 4.520698547363281,
+ "step": 1222
+ },
+ {
+ "epoch": 16.992139737991266,
+ "grad_norm": 0.016601379960775375,
+ "learning_rate": 0.0006,
+ "loss": 4.4809064865112305,
+ "step": 1223
+ },
+ {
+ "epoch": 17.0,
+ "grad_norm": 0.01771303080022335,
+ "learning_rate": 0.0006,
+ "loss": 4.721122741699219,
+ "step": 1224
+ },
+ {
+ "epoch": 17.0,
+ "eval_loss": 4.7903666496276855,
+ "eval_runtime": 56.9826,
+ "eval_samples_per_second": 42.855,
+ "eval_steps_per_second": 1.351,
+ "step": 1224
+ },
+ {
+ "epoch": 17.01397379912664,
+ "grad_norm": 0.017720289528369904,
+ "learning_rate": 0.0006,
+ "loss": 4.523834705352783,
+ "step": 1225
+ },
+ {
+ "epoch": 17.027947598253274,
+ "grad_norm": 0.020137161016464233,
+ "learning_rate": 0.0006,
+ "loss": 4.574061870574951,
+ "step": 1226
+ },
+ {
+ "epoch": 17.041921397379912,
+ "grad_norm": 0.02002314105629921,
+ "learning_rate": 0.0006,
+ "loss": 4.512407302856445,
+ "step": 1227
+ },
+ {
+ "epoch": 17.05589519650655,
+ "grad_norm": 0.019628183916211128,
+ "learning_rate": 0.0006,
+ "loss": 4.502533435821533,
+ "step": 1228
+ },
+ {
+ "epoch": 17.069868995633186,
+ "grad_norm": 0.019808458164334297,
+ "learning_rate": 0.0006,
+ "loss": 4.463156700134277,
+ "step": 1229
+ },
+ {
+ "epoch": 17.083842794759825,
+ "grad_norm": 0.021250786259770393,
+ "learning_rate": 0.0006,
+ "loss": 4.622682094573975,
+ "step": 1230
+ },
+ {
+ "epoch": 17.097816593886463,
+ "grad_norm": 0.022031694650650024,
+ "learning_rate": 0.0006,
+ "loss": 4.451882839202881,
+ "step": 1231
+ },
+ {
+ "epoch": 17.111790393013102,
+ "grad_norm": 0.01898987777531147,
+ "learning_rate": 0.0006,
+ "loss": 4.529331684112549,
+ "step": 1232
+ },
+ {
+ "epoch": 17.125764192139737,
+ "grad_norm": 0.018749834969639778,
+ "learning_rate": 0.0006,
+ "loss": 4.435815334320068,
+ "step": 1233
+ },
+ {
+ "epoch": 17.139737991266376,
+ "grad_norm": 0.01896674931049347,
+ "learning_rate": 0.0006,
+ "loss": 4.484029293060303,
+ "step": 1234
+ },
+ {
+ "epoch": 17.153711790393015,
+ "grad_norm": 0.018286997452378273,
+ "learning_rate": 0.0006,
+ "loss": 4.523510932922363,
+ "step": 1235
+ },
+ {
+ "epoch": 17.16768558951965,
+ "grad_norm": 0.017273059114813805,
+ "learning_rate": 0.0006,
+ "loss": 4.451767444610596,
+ "step": 1236
+ },
+ {
+ "epoch": 17.18165938864629,
+ "grad_norm": 0.016499150544404984,
+ "learning_rate": 0.0006,
+ "loss": 4.5568647384643555,
+ "step": 1237
+ },
+ {
+ "epoch": 17.195633187772927,
+ "grad_norm": 0.015934469178318977,
+ "learning_rate": 0.0006,
+ "loss": 4.4639997482299805,
+ "step": 1238
+ },
+ {
+ "epoch": 17.209606986899562,
+ "grad_norm": 0.01714209094643593,
+ "learning_rate": 0.0006,
+ "loss": 4.547614097595215,
+ "step": 1239
+ },
+ {
+ "epoch": 17.2235807860262,
+ "grad_norm": 0.019167495891451836,
+ "learning_rate": 0.0006,
+ "loss": 4.571097373962402,
+ "step": 1240
+ },
+ {
+ "epoch": 17.23755458515284,
+ "grad_norm": 0.018276996910572052,
+ "learning_rate": 0.0006,
+ "loss": 4.489764213562012,
+ "step": 1241
+ },
+ {
+ "epoch": 17.251528384279474,
+ "grad_norm": 0.021296866238117218,
+ "learning_rate": 0.0006,
+ "loss": 4.526913642883301,
+ "step": 1242
+ },
+ {
+ "epoch": 17.265502183406113,
+ "grad_norm": 0.022606629878282547,
+ "learning_rate": 0.0006,
+ "loss": 4.521624565124512,
+ "step": 1243
+ },
+ {
+ "epoch": 17.27947598253275,
+ "grad_norm": 0.021204711869359016,
+ "learning_rate": 0.0006,
+ "loss": 4.500864028930664,
+ "step": 1244
+ },
+ {
+ "epoch": 17.29344978165939,
+ "grad_norm": 0.01988849975168705,
+ "learning_rate": 0.0006,
+ "loss": 4.45162296295166,
+ "step": 1245
+ },
+ {
+ "epoch": 17.307423580786025,
+ "grad_norm": 0.020268384367227554,
+ "learning_rate": 0.0006,
+ "loss": 4.536198616027832,
+ "step": 1246
+ },
+ {
+ "epoch": 17.321397379912664,
+ "grad_norm": 0.020185677334666252,
+ "learning_rate": 0.0006,
+ "loss": 4.440262794494629,
+ "step": 1247
+ },
+ {
+ "epoch": 17.335371179039303,
+ "grad_norm": 0.020106853917241096,
+ "learning_rate": 0.0006,
+ "loss": 4.475907325744629,
+ "step": 1248
+ },
+ {
+ "epoch": 17.349344978165938,
+ "grad_norm": 0.018883297219872475,
+ "learning_rate": 0.0006,
+ "loss": 4.640661716461182,
+ "step": 1249
+ },
+ {
+ "epoch": 17.363318777292577,
+ "grad_norm": 0.01852620206773281,
+ "learning_rate": 0.0006,
+ "loss": 4.578243255615234,
+ "step": 1250
+ },
+ {
+ "epoch": 17.377292576419215,
+ "grad_norm": 0.01886451430618763,
+ "learning_rate": 0.0006,
+ "loss": 4.546026229858398,
+ "step": 1251
+ },
+ {
+ "epoch": 17.39126637554585,
+ "grad_norm": 0.01777452416718006,
+ "learning_rate": 0.0006,
+ "loss": 4.521542549133301,
+ "step": 1252
+ },
+ {
+ "epoch": 17.40524017467249,
+ "grad_norm": 0.016626447439193726,
+ "learning_rate": 0.0006,
+ "loss": 4.52629280090332,
+ "step": 1253
+ },
+ {
+ "epoch": 17.419213973799128,
+ "grad_norm": 0.01599477231502533,
+ "learning_rate": 0.0006,
+ "loss": 4.4796037673950195,
+ "step": 1254
+ },
+ {
+ "epoch": 17.433187772925763,
+ "grad_norm": 0.017021458595991135,
+ "learning_rate": 0.0006,
+ "loss": 4.468813419342041,
+ "step": 1255
+ },
+ {
+ "epoch": 17.4471615720524,
+ "grad_norm": 0.016925333067774773,
+ "learning_rate": 0.0006,
+ "loss": 4.584898948669434,
+ "step": 1256
+ },
+ {
+ "epoch": 17.46113537117904,
+ "grad_norm": 0.01714998669922352,
+ "learning_rate": 0.0006,
+ "loss": 4.523544788360596,
+ "step": 1257
+ },
+ {
+ "epoch": 17.475109170305675,
+ "grad_norm": 0.01802288182079792,
+ "learning_rate": 0.0006,
+ "loss": 4.595581531524658,
+ "step": 1258
+ },
+ {
+ "epoch": 17.489082969432314,
+ "grad_norm": 0.0180289875715971,
+ "learning_rate": 0.0006,
+ "loss": 4.394440650939941,
+ "step": 1259
+ },
+ {
+ "epoch": 17.503056768558952,
+ "grad_norm": 0.018504245206713676,
+ "learning_rate": 0.0006,
+ "loss": 4.4794416427612305,
+ "step": 1260
+ },
+ {
+ "epoch": 17.51703056768559,
+ "grad_norm": 0.019973233342170715,
+ "learning_rate": 0.0006,
+ "loss": 4.4241437911987305,
+ "step": 1261
+ },
+ {
+ "epoch": 17.531004366812226,
+ "grad_norm": 0.020063083618879318,
+ "learning_rate": 0.0006,
+ "loss": 4.48068380355835,
+ "step": 1262
+ },
+ {
+ "epoch": 17.544978165938865,
+ "grad_norm": 0.02066640742123127,
+ "learning_rate": 0.0006,
+ "loss": 4.628993034362793,
+ "step": 1263
+ },
+ {
+ "epoch": 17.558951965065503,
+ "grad_norm": 0.02206416428089142,
+ "learning_rate": 0.0006,
+ "loss": 4.564790725708008,
+ "step": 1264
+ },
+ {
+ "epoch": 17.57292576419214,
+ "grad_norm": 0.020640335977077484,
+ "learning_rate": 0.0006,
+ "loss": 4.578705787658691,
+ "step": 1265
+ },
+ {
+ "epoch": 17.586899563318777,
+ "grad_norm": 0.019093405455350876,
+ "learning_rate": 0.0006,
+ "loss": 4.5023603439331055,
+ "step": 1266
+ },
+ {
+ "epoch": 17.600873362445416,
+ "grad_norm": 0.019459193572402,
+ "learning_rate": 0.0006,
+ "loss": 4.594531059265137,
+ "step": 1267
+ },
+ {
+ "epoch": 17.61484716157205,
+ "grad_norm": 0.02021171525120735,
+ "learning_rate": 0.0006,
+ "loss": 4.770299911499023,
+ "step": 1268
+ },
+ {
+ "epoch": 17.62882096069869,
+ "grad_norm": 0.021161191165447235,
+ "learning_rate": 0.0006,
+ "loss": 4.5380401611328125,
+ "step": 1269
+ },
+ {
+ "epoch": 17.64279475982533,
+ "grad_norm": 0.021013403311371803,
+ "learning_rate": 0.0006,
+ "loss": 4.455524921417236,
+ "step": 1270
+ },
+ {
+ "epoch": 17.656768558951963,
+ "grad_norm": 0.020336592569947243,
+ "learning_rate": 0.0006,
+ "loss": 4.533672332763672,
+ "step": 1271
+ },
+ {
+ "epoch": 17.670742358078602,
+ "grad_norm": 0.020159434527158737,
+ "learning_rate": 0.0006,
+ "loss": 4.596151351928711,
+ "step": 1272
+ },
+ {
+ "epoch": 17.68471615720524,
+ "grad_norm": 0.020432960242033005,
+ "learning_rate": 0.0006,
+ "loss": 4.55531644821167,
+ "step": 1273
+ },
+ {
+ "epoch": 17.69868995633188,
+ "grad_norm": 0.025466229766607285,
+ "learning_rate": 0.0006,
+ "loss": 4.566912651062012,
+ "step": 1274
+ },
+ {
+ "epoch": 17.712663755458514,
+ "grad_norm": 0.026537558063864708,
+ "learning_rate": 0.0006,
+ "loss": 4.526003837585449,
+ "step": 1275
+ },
+ {
+ "epoch": 17.726637554585153,
+ "grad_norm": 0.024397607892751694,
+ "learning_rate": 0.0006,
+ "loss": 4.541755676269531,
+ "step": 1276
+ },
+ {
+ "epoch": 17.74061135371179,
+ "grad_norm": 0.022062312811613083,
+ "learning_rate": 0.0006,
+ "loss": 4.6532793045043945,
+ "step": 1277
+ },
+ {
+ "epoch": 17.754585152838427,
+ "grad_norm": 0.02204282023012638,
+ "learning_rate": 0.0006,
+ "loss": 4.582825660705566,
+ "step": 1278
+ },
+ {
+ "epoch": 17.768558951965066,
+ "grad_norm": 0.020844416692852974,
+ "learning_rate": 0.0006,
+ "loss": 4.522336483001709,
+ "step": 1279
+ },
+ {
+ "epoch": 17.782532751091704,
+ "grad_norm": 0.019815703853964806,
+ "learning_rate": 0.0006,
+ "loss": 4.607398509979248,
+ "step": 1280
+ },
+ {
+ "epoch": 17.79650655021834,
+ "grad_norm": 0.016676954925060272,
+ "learning_rate": 0.0006,
+ "loss": 4.446206092834473,
+ "step": 1281
+ },
+ {
+ "epoch": 17.810480349344978,
+ "grad_norm": 0.0166452769190073,
+ "learning_rate": 0.0006,
+ "loss": 4.505931854248047,
+ "step": 1282
+ },
+ {
+ "epoch": 17.824454148471617,
+ "grad_norm": 0.015928907319903374,
+ "learning_rate": 0.0006,
+ "loss": 4.455267429351807,
+ "step": 1283
+ },
+ {
+ "epoch": 17.83842794759825,
+ "grad_norm": 0.015219560824334621,
+ "learning_rate": 0.0006,
+ "loss": 4.5623674392700195,
+ "step": 1284
+ },
+ {
+ "epoch": 17.85240174672489,
+ "grad_norm": 0.015125355683267117,
+ "learning_rate": 0.0006,
+ "loss": 4.554104804992676,
+ "step": 1285
+ },
+ {
+ "epoch": 17.86637554585153,
+ "grad_norm": 0.015612718649208546,
+ "learning_rate": 0.0006,
+ "loss": 4.484847068786621,
+ "step": 1286
+ },
+ {
+ "epoch": 17.880349344978168,
+ "grad_norm": 0.015215662308037281,
+ "learning_rate": 0.0006,
+ "loss": 4.4802093505859375,
+ "step": 1287
+ },
+ {
+ "epoch": 17.894323144104803,
+ "grad_norm": 0.01511684525758028,
+ "learning_rate": 0.0006,
+ "loss": 4.537298679351807,
+ "step": 1288
+ },
+ {
+ "epoch": 17.90829694323144,
+ "grad_norm": 0.013841322623193264,
+ "learning_rate": 0.0006,
+ "loss": 4.427488327026367,
+ "step": 1289
+ },
+ {
+ "epoch": 17.92227074235808,
+ "grad_norm": 0.015115964226424694,
+ "learning_rate": 0.0006,
+ "loss": 4.4547119140625,
+ "step": 1290
+ },
+ {
+ "epoch": 17.936244541484715,
+ "grad_norm": 0.016411006450653076,
+ "learning_rate": 0.0006,
+ "loss": 4.52994966506958,
+ "step": 1291
+ },
+ {
+ "epoch": 17.950218340611354,
+ "grad_norm": 0.017313152551651,
+ "learning_rate": 0.0006,
+ "loss": 4.5337653160095215,
+ "step": 1292
+ },
+ {
+ "epoch": 17.964192139737992,
+ "grad_norm": 0.016559796407818794,
+ "learning_rate": 0.0006,
+ "loss": 4.5277910232543945,
+ "step": 1293
+ },
+ {
+ "epoch": 17.978165938864628,
+ "grad_norm": 0.015854761004447937,
+ "learning_rate": 0.0006,
+ "loss": 4.6334309577941895,
+ "step": 1294
+ },
+ {
+ "epoch": 17.992139737991266,
+ "grad_norm": 0.015658238902688026,
+ "learning_rate": 0.0006,
+ "loss": 4.613372802734375,
+ "step": 1295
+ },
+ {
+ "epoch": 18.0,
+ "grad_norm": 0.0169155802577734,
+ "learning_rate": 0.0006,
+ "loss": 4.527356147766113,
+ "step": 1296
+ },
+ {
+ "epoch": 18.0,
+ "eval_loss": 4.764543056488037,
+ "eval_runtime": 56.5998,
+ "eval_samples_per_second": 43.145,
+ "eval_steps_per_second": 1.36,
+ "step": 1296
+ },
+ {
+ "epoch": 18.01397379912664,
+ "grad_norm": 0.015348963439464569,
+ "learning_rate": 0.0006,
+ "loss": 4.466423988342285,
+ "step": 1297
+ },
+ {
+ "epoch": 18.027947598253274,
+ "grad_norm": 0.01768588088452816,
+ "learning_rate": 0.0006,
+ "loss": 4.545933246612549,
+ "step": 1298
+ },
+ {
+ "epoch": 18.041921397379912,
+ "grad_norm": 0.017381226643919945,
+ "learning_rate": 0.0006,
+ "loss": 4.494060039520264,
+ "step": 1299
+ },
+ {
+ "epoch": 18.05589519650655,
+ "grad_norm": 0.019002335146069527,
+ "learning_rate": 0.0006,
+ "loss": 4.395200729370117,
+ "step": 1300
+ },
+ {
+ "epoch": 18.069868995633186,
+ "grad_norm": 0.01827932707965374,
+ "learning_rate": 0.0006,
+ "loss": 4.424848556518555,
+ "step": 1301
+ },
+ {
+ "epoch": 18.083842794759825,
+ "grad_norm": 0.016383031383156776,
+ "learning_rate": 0.0006,
+ "loss": 4.494722366333008,
+ "step": 1302
+ },
+ {
+ "epoch": 18.097816593886463,
+ "grad_norm": 0.016265157610177994,
+ "learning_rate": 0.0006,
+ "loss": 4.416522979736328,
+ "step": 1303
+ },
+ {
+ "epoch": 18.111790393013102,
+ "grad_norm": 0.01702386513352394,
+ "learning_rate": 0.0006,
+ "loss": 4.600103378295898,
+ "step": 1304
+ },
+ {
+ "epoch": 18.125764192139737,
+ "grad_norm": 0.018495498225092888,
+ "learning_rate": 0.0006,
+ "loss": 4.588401794433594,
+ "step": 1305
+ },
+ {
+ "epoch": 18.139737991266376,
+ "grad_norm": 0.017921684309840202,
+ "learning_rate": 0.0006,
+ "loss": 4.478075981140137,
+ "step": 1306
+ },
+ {
+ "epoch": 18.153711790393015,
+ "grad_norm": 0.01801278069615364,
+ "learning_rate": 0.0006,
+ "loss": 4.413186550140381,
+ "step": 1307
+ },
+ {
+ "epoch": 18.16768558951965,
+ "grad_norm": 0.017211275175213814,
+ "learning_rate": 0.0006,
+ "loss": 4.461341857910156,
+ "step": 1308
+ },
+ {
+ "epoch": 18.18165938864629,
+ "grad_norm": 0.017589328810572624,
+ "learning_rate": 0.0006,
+ "loss": 4.546969413757324,
+ "step": 1309
+ },
+ {
+ "epoch": 18.195633187772927,
+ "grad_norm": 0.01797698810696602,
+ "learning_rate": 0.0006,
+ "loss": 4.530797004699707,
+ "step": 1310
+ },
+ {
+ "epoch": 18.209606986899562,
+ "grad_norm": 0.020408619195222855,
+ "learning_rate": 0.0006,
+ "loss": 4.62999153137207,
+ "step": 1311
+ },
+ {
+ "epoch": 18.2235807860262,
+ "grad_norm": 0.01994038000702858,
+ "learning_rate": 0.0006,
+ "loss": 4.586811065673828,
+ "step": 1312
+ },
+ {
+ "epoch": 18.23755458515284,
+ "grad_norm": 0.019381340593099594,
+ "learning_rate": 0.0006,
+ "loss": 4.4769368171691895,
+ "step": 1313
+ },
+ {
+ "epoch": 18.251528384279474,
+ "grad_norm": 0.020484555512666702,
+ "learning_rate": 0.0006,
+ "loss": 4.524302959442139,
+ "step": 1314
+ },
+ {
+ "epoch": 18.265502183406113,
+ "grad_norm": 0.02426416426897049,
+ "learning_rate": 0.0006,
+ "loss": 4.561705589294434,
+ "step": 1315
+ },
+ {
+ "epoch": 18.27947598253275,
+ "grad_norm": 0.024748101830482483,
+ "learning_rate": 0.0006,
+ "loss": 4.4654316902160645,
+ "step": 1316
+ },
+ {
+ "epoch": 18.29344978165939,
+ "grad_norm": 0.020703328773379326,
+ "learning_rate": 0.0006,
+ "loss": 4.53033447265625,
+ "step": 1317
+ },
+ {
+ "epoch": 18.307423580786025,
+ "grad_norm": 0.01765458658337593,
+ "learning_rate": 0.0006,
+ "loss": 4.5154266357421875,
+ "step": 1318
+ },
+ {
+ "epoch": 18.321397379912664,
+ "grad_norm": 0.018646420910954475,
+ "learning_rate": 0.0006,
+ "loss": 4.5224714279174805,
+ "step": 1319
+ },
+ {
+ "epoch": 18.335371179039303,
+ "grad_norm": 0.01907474733889103,
+ "learning_rate": 0.0006,
+ "loss": 4.5055437088012695,
+ "step": 1320
+ },
+ {
+ "epoch": 18.349344978165938,
+ "grad_norm": 0.018179846927523613,
+ "learning_rate": 0.0006,
+ "loss": 4.462161064147949,
+ "step": 1321
+ },
+ {
+ "epoch": 18.363318777292577,
+ "grad_norm": 0.016697878018021584,
+ "learning_rate": 0.0006,
+ "loss": 4.473718643188477,
+ "step": 1322
+ },
+ {
+ "epoch": 18.377292576419215,
+ "grad_norm": 0.017504019662737846,
+ "learning_rate": 0.0006,
+ "loss": 4.474348068237305,
+ "step": 1323
+ },
+ {
+ "epoch": 18.39126637554585,
+ "grad_norm": 0.018889913335442543,
+ "learning_rate": 0.0006,
+ "loss": 4.4531636238098145,
+ "step": 1324
+ },
+ {
+ "epoch": 18.40524017467249,
+ "grad_norm": 0.019167251884937286,
+ "learning_rate": 0.0006,
+ "loss": 4.4278364181518555,
+ "step": 1325
+ },
+ {
+ "epoch": 18.419213973799128,
+ "grad_norm": 0.015895741060376167,
+ "learning_rate": 0.0006,
+ "loss": 4.474592208862305,
+ "step": 1326
+ },
+ {
+ "epoch": 18.433187772925763,
+ "grad_norm": 0.01531882956624031,
+ "learning_rate": 0.0006,
+ "loss": 4.403757572174072,
+ "step": 1327
+ },
+ {
+ "epoch": 18.4471615720524,
+ "grad_norm": 0.01565849967300892,
+ "learning_rate": 0.0006,
+ "loss": 4.507664680480957,
+ "step": 1328
+ },
+ {
+ "epoch": 18.46113537117904,
+ "grad_norm": 0.01602315716445446,
+ "learning_rate": 0.0006,
+ "loss": 4.5648908615112305,
+ "step": 1329
+ },
+ {
+ "epoch": 18.475109170305675,
+ "grad_norm": 0.016467366367578506,
+ "learning_rate": 0.0006,
+ "loss": 4.428762912750244,
+ "step": 1330
+ },
+ {
+ "epoch": 18.489082969432314,
+ "grad_norm": 0.01604127697646618,
+ "learning_rate": 0.0006,
+ "loss": 4.527771472930908,
+ "step": 1331
+ },
+ {
+ "epoch": 18.503056768558952,
+ "grad_norm": 0.018825586885213852,
+ "learning_rate": 0.0006,
+ "loss": 4.485386848449707,
+ "step": 1332
+ },
+ {
+ "epoch": 18.51703056768559,
+ "grad_norm": 0.019258007407188416,
+ "learning_rate": 0.0006,
+ "loss": 4.414972305297852,
+ "step": 1333
+ },
+ {
+ "epoch": 18.531004366812226,
+ "grad_norm": 0.018192021176218987,
+ "learning_rate": 0.0006,
+ "loss": 4.412166595458984,
+ "step": 1334
+ },
+ {
+ "epoch": 18.544978165938865,
+ "grad_norm": 0.016120698302984238,
+ "learning_rate": 0.0006,
+ "loss": 4.484317779541016,
+ "step": 1335
+ },
+ {
+ "epoch": 18.558951965065503,
+ "grad_norm": 0.016496622934937477,
+ "learning_rate": 0.0006,
+ "loss": 4.514986515045166,
+ "step": 1336
+ },
+ {
+ "epoch": 18.57292576419214,
+ "grad_norm": 0.01795116998255253,
+ "learning_rate": 0.0006,
+ "loss": 4.41473388671875,
+ "step": 1337
+ },
+ {
+ "epoch": 18.586899563318777,
+ "grad_norm": 0.019062276929616928,
+ "learning_rate": 0.0006,
+ "loss": 4.601996898651123,
+ "step": 1338
+ },
+ {
+ "epoch": 18.600873362445416,
+ "grad_norm": 0.017174631357192993,
+ "learning_rate": 0.0006,
+ "loss": 4.571218490600586,
+ "step": 1339
+ },
+ {
+ "epoch": 18.61484716157205,
+ "grad_norm": 0.014015606604516506,
+ "learning_rate": 0.0006,
+ "loss": 4.526606559753418,
+ "step": 1340
+ },
+ {
+ "epoch": 18.62882096069869,
+ "grad_norm": 0.015358510427176952,
+ "learning_rate": 0.0006,
+ "loss": 4.604286193847656,
+ "step": 1341
+ },
+ {
+ "epoch": 18.64279475982533,
+ "grad_norm": 0.01619391329586506,
+ "learning_rate": 0.0006,
+ "loss": 4.560883522033691,
+ "step": 1342
+ },
+ {
+ "epoch": 18.656768558951963,
+ "grad_norm": 0.015972964465618134,
+ "learning_rate": 0.0006,
+ "loss": 4.479755401611328,
+ "step": 1343
+ },
+ {
+ "epoch": 18.670742358078602,
+ "grad_norm": 0.016861025243997574,
+ "learning_rate": 0.0006,
+ "loss": 4.474471092224121,
+ "step": 1344
+ },
+ {
+ "epoch": 18.68471615720524,
+ "grad_norm": 0.01822078227996826,
+ "learning_rate": 0.0006,
+ "loss": 4.509489059448242,
+ "step": 1345
+ },
+ {
+ "epoch": 18.69868995633188,
+ "grad_norm": 0.019415007904171944,
+ "learning_rate": 0.0006,
+ "loss": 4.508941173553467,
+ "step": 1346
+ },
+ {
+ "epoch": 18.712663755458514,
+ "grad_norm": 0.019856126978993416,
+ "learning_rate": 0.0006,
+ "loss": 4.446670055389404,
+ "step": 1347
+ },
+ {
+ "epoch": 18.726637554585153,
+ "grad_norm": 0.01842520199716091,
+ "learning_rate": 0.0006,
+ "loss": 4.524422645568848,
+ "step": 1348
+ },
+ {
+ "epoch": 18.74061135371179,
+ "grad_norm": 0.017196480184793472,
+ "learning_rate": 0.0006,
+ "loss": 4.498466968536377,
+ "step": 1349
+ },
+ {
+ "epoch": 18.754585152838427,
+ "grad_norm": 0.01823010854423046,
+ "learning_rate": 0.0006,
+ "loss": 4.534684181213379,
+ "step": 1350
+ },
+ {
+ "epoch": 18.768558951965066,
+ "grad_norm": 0.020398065447807312,
+ "learning_rate": 0.0006,
+ "loss": 4.456494331359863,
+ "step": 1351
+ },
+ {
+ "epoch": 18.782532751091704,
+ "grad_norm": 0.022992262616753578,
+ "learning_rate": 0.0006,
+ "loss": 4.476374626159668,
+ "step": 1352
+ },
+ {
+ "epoch": 18.79650655021834,
+ "grad_norm": 0.02235420234501362,
+ "learning_rate": 0.0006,
+ "loss": 4.587776184082031,
+ "step": 1353
+ },
+ {
+ "epoch": 18.810480349344978,
+ "grad_norm": 0.02093626745045185,
+ "learning_rate": 0.0006,
+ "loss": 4.424172401428223,
+ "step": 1354
+ },
+ {
+ "epoch": 18.824454148471617,
+ "grad_norm": 0.01997302658855915,
+ "learning_rate": 0.0006,
+ "loss": 4.448066711425781,
+ "step": 1355
+ },
+ {
+ "epoch": 18.83842794759825,
+ "grad_norm": 0.018265459686517715,
+ "learning_rate": 0.0006,
+ "loss": 4.436610698699951,
+ "step": 1356
+ },
+ {
+ "epoch": 18.85240174672489,
+ "grad_norm": 0.01673833839595318,
+ "learning_rate": 0.0006,
+ "loss": 4.5502214431762695,
+ "step": 1357
+ },
+ {
+ "epoch": 18.86637554585153,
+ "grad_norm": 0.017908833920955658,
+ "learning_rate": 0.0006,
+ "loss": 4.550530433654785,
+ "step": 1358
+ },
+ {
+ "epoch": 18.880349344978168,
+ "grad_norm": 0.01796654611825943,
+ "learning_rate": 0.0006,
+ "loss": 4.531463146209717,
+ "step": 1359
+ },
+ {
+ "epoch": 18.894323144104803,
+ "grad_norm": 0.01646183803677559,
+ "learning_rate": 0.0006,
+ "loss": 4.498582363128662,
+ "step": 1360
+ },
+ {
+ "epoch": 18.90829694323144,
+ "grad_norm": 0.015430327504873276,
+ "learning_rate": 0.0006,
+ "loss": 4.52571439743042,
+ "step": 1361
+ },
+ {
+ "epoch": 18.92227074235808,
+ "grad_norm": 0.016527952626347542,
+ "learning_rate": 0.0006,
+ "loss": 4.547013759613037,
+ "step": 1362
+ },
+ {
+ "epoch": 18.936244541484715,
+ "grad_norm": 0.016263660043478012,
+ "learning_rate": 0.0006,
+ "loss": 4.468754291534424,
+ "step": 1363
+ },
+ {
+ "epoch": 18.950218340611354,
+ "grad_norm": 0.016347210854291916,
+ "learning_rate": 0.0006,
+ "loss": 4.450741767883301,
+ "step": 1364
+ },
+ {
+ "epoch": 18.964192139737992,
+ "grad_norm": 0.01701398566365242,
+ "learning_rate": 0.0006,
+ "loss": 4.534899711608887,
+ "step": 1365
+ },
+ {
+ "epoch": 18.978165938864628,
+ "grad_norm": 0.017693081870675087,
+ "learning_rate": 0.0006,
+ "loss": 4.460553169250488,
+ "step": 1366
+ },
+ {
+ "epoch": 18.992139737991266,
+ "grad_norm": 0.018746308982372284,
+ "learning_rate": 0.0006,
+ "loss": 4.538792610168457,
+ "step": 1367
+ },
+ {
+ "epoch": 19.0,
+ "grad_norm": 0.020099416375160217,
+ "learning_rate": 0.0006,
+ "loss": 4.449617385864258,
+ "step": 1368
+ },
+ {
+ "epoch": 19.0,
+ "eval_loss": 4.7756571769714355,
+ "eval_runtime": 57.1804,
+ "eval_samples_per_second": 42.707,
+ "eval_steps_per_second": 1.347,
+ "step": 1368
+ },
+ {
+ "epoch": 19.01397379912664,
+ "grad_norm": 0.018726933747529984,
+ "learning_rate": 0.0006,
+ "loss": 4.446059703826904,
+ "step": 1369
+ },
+ {
+ "epoch": 19.027947598253274,
+ "grad_norm": 0.01956087350845337,
+ "learning_rate": 0.0006,
+ "loss": 4.4520039558410645,
+ "step": 1370
+ },
+ {
+ "epoch": 19.041921397379912,
+ "grad_norm": 0.020092811435461044,
+ "learning_rate": 0.0006,
+ "loss": 4.543343544006348,
+ "step": 1371
+ },
+ {
+ "epoch": 19.05589519650655,
+ "grad_norm": 0.01843622885644436,
+ "learning_rate": 0.0006,
+ "loss": 4.535579204559326,
+ "step": 1372
+ },
+ {
+ "epoch": 19.069868995633186,
+ "grad_norm": 0.016989488154649734,
+ "learning_rate": 0.0006,
+ "loss": 4.52316951751709,
+ "step": 1373
+ },
+ {
+ "epoch": 19.083842794759825,
+ "grad_norm": 0.01805422455072403,
+ "learning_rate": 0.0006,
+ "loss": 4.5393266677856445,
+ "step": 1374
+ },
+ {
+ "epoch": 19.097816593886463,
+ "grad_norm": 0.01874629408121109,
+ "learning_rate": 0.0006,
+ "loss": 4.422059059143066,
+ "step": 1375
+ },
+ {
+ "epoch": 19.111790393013102,
+ "grad_norm": 0.01804221048951149,
+ "learning_rate": 0.0006,
+ "loss": 4.50508451461792,
+ "step": 1376
+ },
+ {
+ "epoch": 19.125764192139737,
+ "grad_norm": 0.01785474829375744,
+ "learning_rate": 0.0006,
+ "loss": 4.412599563598633,
+ "step": 1377
+ },
+ {
+ "epoch": 19.139737991266376,
+ "grad_norm": 0.01971791312098503,
+ "learning_rate": 0.0006,
+ "loss": 4.518318176269531,
+ "step": 1378
+ },
+ {
+ "epoch": 19.153711790393015,
+ "grad_norm": 0.02061633951961994,
+ "learning_rate": 0.0006,
+ "loss": 4.332393169403076,
+ "step": 1379
+ },
+ {
+ "epoch": 19.16768558951965,
+ "grad_norm": 0.020563429221510887,
+ "learning_rate": 0.0006,
+ "loss": 4.445863723754883,
+ "step": 1380
+ },
+ {
+ "epoch": 19.18165938864629,
+ "grad_norm": 0.019995713606476784,
+ "learning_rate": 0.0006,
+ "loss": 4.497312068939209,
+ "step": 1381
+ },
+ {
+ "epoch": 19.195633187772927,
+ "grad_norm": 0.019823070615530014,
+ "learning_rate": 0.0006,
+ "loss": 4.475085258483887,
+ "step": 1382
+ },
+ {
+ "epoch": 19.209606986899562,
+ "grad_norm": 0.019039355218410492,
+ "learning_rate": 0.0006,
+ "loss": 4.50432014465332,
+ "step": 1383
+ },
+ {
+ "epoch": 19.2235807860262,
+ "grad_norm": 0.017876973375678062,
+ "learning_rate": 0.0006,
+ "loss": 4.384788513183594,
+ "step": 1384
+ },
+ {
+ "epoch": 19.23755458515284,
+ "grad_norm": 0.018697699531912804,
+ "learning_rate": 0.0006,
+ "loss": 4.410778045654297,
+ "step": 1385
+ },
+ {
+ "epoch": 19.251528384279474,
+ "grad_norm": 0.018826806917786598,
+ "learning_rate": 0.0006,
+ "loss": 4.524034023284912,
+ "step": 1386
+ },
+ {
+ "epoch": 19.265502183406113,
+ "grad_norm": 0.01824226602911949,
+ "learning_rate": 0.0006,
+ "loss": 4.526393890380859,
+ "step": 1387
+ },
+ {
+ "epoch": 19.27947598253275,
+ "grad_norm": 0.01753208599984646,
+ "learning_rate": 0.0006,
+ "loss": 4.493703365325928,
+ "step": 1388
+ },
+ {
+ "epoch": 19.29344978165939,
+ "grad_norm": 0.01742432825267315,
+ "learning_rate": 0.0006,
+ "loss": 4.392831802368164,
+ "step": 1389
+ },
+ {
+ "epoch": 19.307423580786025,
+ "grad_norm": 0.017099356278777122,
+ "learning_rate": 0.0006,
+ "loss": 4.578668594360352,
+ "step": 1390
+ },
+ {
+ "epoch": 19.321397379912664,
+ "grad_norm": 0.017946023494005203,
+ "learning_rate": 0.0006,
+ "loss": 4.456718444824219,
+ "step": 1391
+ },
+ {
+ "epoch": 19.335371179039303,
+ "grad_norm": 0.017705943435430527,
+ "learning_rate": 0.0006,
+ "loss": 4.553682327270508,
+ "step": 1392
+ },
+ {
+ "epoch": 19.349344978165938,
+ "grad_norm": 0.016647523269057274,
+ "learning_rate": 0.0006,
+ "loss": 4.525016784667969,
+ "step": 1393
+ },
+ {
+ "epoch": 19.363318777292577,
+ "grad_norm": 0.015407416969537735,
+ "learning_rate": 0.0006,
+ "loss": 4.500231742858887,
+ "step": 1394
+ },
+ {
+ "epoch": 19.377292576419215,
+ "grad_norm": 0.017264019697904587,
+ "learning_rate": 0.0006,
+ "loss": 4.4615349769592285,
+ "step": 1395
+ },
+ {
+ "epoch": 19.39126637554585,
+ "grad_norm": 0.01871555484831333,
+ "learning_rate": 0.0006,
+ "loss": 4.281026840209961,
+ "step": 1396
+ },
+ {
+ "epoch": 19.40524017467249,
+ "grad_norm": 0.017361650243401527,
+ "learning_rate": 0.0006,
+ "loss": 4.479382514953613,
+ "step": 1397
+ },
+ {
+ "epoch": 19.419213973799128,
+ "grad_norm": 0.019781362265348434,
+ "learning_rate": 0.0006,
+ "loss": 4.402320861816406,
+ "step": 1398
+ },
+ {
+ "epoch": 19.433187772925763,
+ "grad_norm": 0.019043298438191414,
+ "learning_rate": 0.0006,
+ "loss": 4.476983070373535,
+ "step": 1399
+ },
+ {
+ "epoch": 19.4471615720524,
+ "grad_norm": 0.017137985676527023,
+ "learning_rate": 0.0006,
+ "loss": 4.418689250946045,
+ "step": 1400
+ },
+ {
+ "epoch": 19.46113537117904,
+ "grad_norm": 0.016481833532452583,
+ "learning_rate": 0.0006,
+ "loss": 4.579145431518555,
+ "step": 1401
+ },
+ {
+ "epoch": 19.475109170305675,
+ "grad_norm": 0.016507714986801147,
+ "learning_rate": 0.0006,
+ "loss": 4.519158363342285,
+ "step": 1402
+ },
+ {
+ "epoch": 19.489082969432314,
+ "grad_norm": 0.018092872574925423,
+ "learning_rate": 0.0006,
+ "loss": 4.513748645782471,
+ "step": 1403
+ },
+ {
+ "epoch": 19.503056768558952,
+ "grad_norm": 0.018153440207242966,
+ "learning_rate": 0.0006,
+ "loss": 4.645384311676025,
+ "step": 1404
+ },
+ {
+ "epoch": 19.51703056768559,
+ "grad_norm": 0.01657119207084179,
+ "learning_rate": 0.0006,
+ "loss": 4.399757385253906,
+ "step": 1405
+ },
+ {
+ "epoch": 19.531004366812226,
+ "grad_norm": 0.0156280305236578,
+ "learning_rate": 0.0006,
+ "loss": 4.467836380004883,
+ "step": 1406
+ },
+ {
+ "epoch": 19.544978165938865,
+ "grad_norm": 0.016838403418660164,
+ "learning_rate": 0.0006,
+ "loss": 4.529176235198975,
+ "step": 1407
+ },
+ {
+ "epoch": 19.558951965065503,
+ "grad_norm": 0.01743236929178238,
+ "learning_rate": 0.0006,
+ "loss": 4.440032482147217,
+ "step": 1408
+ },
+ {
+ "epoch": 19.57292576419214,
+ "grad_norm": 0.01553610060364008,
+ "learning_rate": 0.0006,
+ "loss": 4.425588607788086,
+ "step": 1409
+ },
+ {
+ "epoch": 19.586899563318777,
+ "grad_norm": 0.017259271815419197,
+ "learning_rate": 0.0006,
+ "loss": 4.405452728271484,
+ "step": 1410
+ },
+ {
+ "epoch": 19.600873362445416,
+ "grad_norm": 0.015568481758236885,
+ "learning_rate": 0.0006,
+ "loss": 4.444279193878174,
+ "step": 1411
+ },
+ {
+ "epoch": 19.61484716157205,
+ "grad_norm": 0.015331248752772808,
+ "learning_rate": 0.0006,
+ "loss": 4.451846122741699,
+ "step": 1412
+ },
+ {
+ "epoch": 19.62882096069869,
+ "grad_norm": 0.015667378902435303,
+ "learning_rate": 0.0006,
+ "loss": 4.533931255340576,
+ "step": 1413
+ },
+ {
+ "epoch": 19.64279475982533,
+ "grad_norm": 0.016177602112293243,
+ "learning_rate": 0.0006,
+ "loss": 4.467313289642334,
+ "step": 1414
+ },
+ {
+ "epoch": 19.656768558951963,
+ "grad_norm": 0.01733129471540451,
+ "learning_rate": 0.0006,
+ "loss": 4.496893882751465,
+ "step": 1415
+ },
+ {
+ "epoch": 19.670742358078602,
+ "grad_norm": 0.016539594158530235,
+ "learning_rate": 0.0006,
+ "loss": 4.401961326599121,
+ "step": 1416
+ },
+ {
+ "epoch": 19.68471615720524,
+ "grad_norm": 0.016058551147580147,
+ "learning_rate": 0.0006,
+ "loss": 4.503375053405762,
+ "step": 1417
+ },
+ {
+ "epoch": 19.69868995633188,
+ "grad_norm": 0.01715777814388275,
+ "learning_rate": 0.0006,
+ "loss": 4.3808369636535645,
+ "step": 1418
+ },
+ {
+ "epoch": 19.712663755458514,
+ "grad_norm": 0.01681119203567505,
+ "learning_rate": 0.0006,
+ "loss": 4.5469560623168945,
+ "step": 1419
+ },
+ {
+ "epoch": 19.726637554585153,
+ "grad_norm": 0.01634550653398037,
+ "learning_rate": 0.0006,
+ "loss": 4.328197956085205,
+ "step": 1420
+ },
+ {
+ "epoch": 19.74061135371179,
+ "grad_norm": 0.01549347210675478,
+ "learning_rate": 0.0006,
+ "loss": 4.534621238708496,
+ "step": 1421
+ },
+ {
+ "epoch": 19.754585152838427,
+ "grad_norm": 0.01654665172100067,
+ "learning_rate": 0.0006,
+ "loss": 4.4724225997924805,
+ "step": 1422
+ },
+ {
+ "epoch": 19.768558951965066,
+ "grad_norm": 0.017119025811553,
+ "learning_rate": 0.0006,
+ "loss": 4.3897247314453125,
+ "step": 1423
+ },
+ {
+ "epoch": 19.782532751091704,
+ "grad_norm": 0.01603223942220211,
+ "learning_rate": 0.0006,
+ "loss": 4.412731170654297,
+ "step": 1424
+ },
+ {
+ "epoch": 19.79650655021834,
+ "grad_norm": 0.016774067655205727,
+ "learning_rate": 0.0006,
+ "loss": 4.54707145690918,
+ "step": 1425
+ },
+ {
+ "epoch": 19.810480349344978,
+ "grad_norm": 0.016453025862574577,
+ "learning_rate": 0.0006,
+ "loss": 4.462501049041748,
+ "step": 1426
+ },
+ {
+ "epoch": 19.824454148471617,
+ "grad_norm": 0.01597629487514496,
+ "learning_rate": 0.0006,
+ "loss": 4.229758262634277,
+ "step": 1427
+ },
+ {
+ "epoch": 19.83842794759825,
+ "grad_norm": 0.017999034374952316,
+ "learning_rate": 0.0006,
+ "loss": 4.389263153076172,
+ "step": 1428
+ },
+ {
+ "epoch": 19.85240174672489,
+ "grad_norm": 0.02165898121893406,
+ "learning_rate": 0.0006,
+ "loss": 4.342130661010742,
+ "step": 1429
+ },
+ {
+ "epoch": 19.86637554585153,
+ "grad_norm": 0.02222822792828083,
+ "learning_rate": 0.0006,
+ "loss": 4.492661476135254,
+ "step": 1430
+ },
+ {
+ "epoch": 19.880349344978168,
+ "grad_norm": 0.019378066062927246,
+ "learning_rate": 0.0006,
+ "loss": 4.4614105224609375,
+ "step": 1431
+ },
+ {
+ "epoch": 19.894323144104803,
+ "grad_norm": 0.019908083602786064,
+ "learning_rate": 0.0006,
+ "loss": 4.370929718017578,
+ "step": 1432
+ },
+ {
+ "epoch": 19.90829694323144,
+ "grad_norm": 0.022296471521258354,
+ "learning_rate": 0.0006,
+ "loss": 4.438320159912109,
+ "step": 1433
+ },
+ {
+ "epoch": 19.92227074235808,
+ "grad_norm": 0.021810343489050865,
+ "learning_rate": 0.0006,
+ "loss": 4.462067127227783,
+ "step": 1434
+ },
+ {
+ "epoch": 19.936244541484715,
+ "grad_norm": 0.02129209227859974,
+ "learning_rate": 0.0006,
+ "loss": 4.541261196136475,
+ "step": 1435
+ },
+ {
+ "epoch": 19.950218340611354,
+ "grad_norm": 0.02079629711806774,
+ "learning_rate": 0.0006,
+ "loss": 4.36712646484375,
+ "step": 1436
+ },
+ {
+ "epoch": 19.964192139737992,
+ "grad_norm": 0.022501740604639053,
+ "learning_rate": 0.0006,
+ "loss": 4.5279083251953125,
+ "step": 1437
+ },
+ {
+ "epoch": 19.978165938864628,
+ "grad_norm": 0.019121676683425903,
+ "learning_rate": 0.0006,
+ "loss": 4.477789878845215,
+ "step": 1438
+ },
+ {
+ "epoch": 19.992139737991266,
+ "grad_norm": 0.017429588362574577,
+ "learning_rate": 0.0006,
+ "loss": 4.4023847579956055,
+ "step": 1439
+ },
+ {
+ "epoch": 20.0,
+ "grad_norm": 0.018857665359973907,
+ "learning_rate": 0.0006,
+ "loss": 4.494743347167969,
+ "step": 1440
+ },
+ {
+ "epoch": 20.0,
+ "eval_loss": 4.694555759429932,
+ "eval_runtime": 56.4146,
+ "eval_samples_per_second": 43.287,
+ "eval_steps_per_second": 1.365,
+ "step": 1440
+ },
+ {
+ "epoch": 20.01397379912664,
+ "grad_norm": 0.016505861654877663,
+ "learning_rate": 0.0006,
+ "loss": 4.506229877471924,
+ "step": 1441
+ },
+ {
+ "epoch": 20.027947598253274,
+ "grad_norm": 0.017571842297911644,
+ "learning_rate": 0.0006,
+ "loss": 4.384073257446289,
+ "step": 1442
+ },
+ {
+ "epoch": 20.041921397379912,
+ "grad_norm": 0.019398365169763565,
+ "learning_rate": 0.0006,
+ "loss": 4.482177734375,
+ "step": 1443
+ },
+ {
+ "epoch": 20.05589519650655,
+ "grad_norm": 0.02011062018573284,
+ "learning_rate": 0.0006,
+ "loss": 4.440773963928223,
+ "step": 1444
+ },
+ {
+ "epoch": 20.069868995633186,
+ "grad_norm": 0.019026953727006912,
+ "learning_rate": 0.0006,
+ "loss": 4.507168769836426,
+ "step": 1445
+ },
+ {
+ "epoch": 20.083842794759825,
+ "grad_norm": 0.020506031811237335,
+ "learning_rate": 0.0006,
+ "loss": 4.469079971313477,
+ "step": 1446
+ },
+ {
+ "epoch": 20.097816593886463,
+ "grad_norm": 0.019608162343502045,
+ "learning_rate": 0.0006,
+ "loss": 4.395927429199219,
+ "step": 1447
+ },
+ {
+ "epoch": 20.111790393013102,
+ "grad_norm": 0.019419124349951744,
+ "learning_rate": 0.0006,
+ "loss": 4.484235763549805,
+ "step": 1448
+ },
+ {
+ "epoch": 20.125764192139737,
+ "grad_norm": 0.021775128319859505,
+ "learning_rate": 0.0006,
+ "loss": 4.333173751831055,
+ "step": 1449
+ },
+ {
+ "epoch": 20.139737991266376,
+ "grad_norm": 0.025512272492051125,
+ "learning_rate": 0.0006,
+ "loss": 4.460362434387207,
+ "step": 1450
+ },
+ {
+ "epoch": 20.153711790393015,
+ "grad_norm": 0.025157401338219643,
+ "learning_rate": 0.0006,
+ "loss": 4.4361395835876465,
+ "step": 1451
+ },
+ {
+ "epoch": 20.16768558951965,
+ "grad_norm": 0.019862616434693336,
+ "learning_rate": 0.0006,
+ "loss": 4.426807403564453,
+ "step": 1452
+ },
+ {
+ "epoch": 20.18165938864629,
+ "grad_norm": 0.01932152360677719,
+ "learning_rate": 0.0006,
+ "loss": 4.375677108764648,
+ "step": 1453
+ },
+ {
+ "epoch": 20.195633187772927,
+ "grad_norm": 0.01760999858379364,
+ "learning_rate": 0.0006,
+ "loss": 4.42259407043457,
+ "step": 1454
+ },
+ {
+ "epoch": 20.209606986899562,
+ "grad_norm": 0.016998112201690674,
+ "learning_rate": 0.0006,
+ "loss": 4.481280326843262,
+ "step": 1455
+ },
+ {
+ "epoch": 20.2235807860262,
+ "grad_norm": 0.017199190333485603,
+ "learning_rate": 0.0006,
+ "loss": 4.496712684631348,
+ "step": 1456
+ },
+ {
+ "epoch": 20.23755458515284,
+ "grad_norm": 0.016833679750561714,
+ "learning_rate": 0.0006,
+ "loss": 4.40736722946167,
+ "step": 1457
+ },
+ {
+ "epoch": 20.251528384279474,
+ "grad_norm": 0.017546730116009712,
+ "learning_rate": 0.0006,
+ "loss": 4.4915266036987305,
+ "step": 1458
+ },
+ {
+ "epoch": 20.265502183406113,
+ "grad_norm": 0.016595078632235527,
+ "learning_rate": 0.0006,
+ "loss": 4.460815906524658,
+ "step": 1459
+ },
+ {
+ "epoch": 20.27947598253275,
+ "grad_norm": 0.017728568986058235,
+ "learning_rate": 0.0006,
+ "loss": 4.346048831939697,
+ "step": 1460
+ },
+ {
+ "epoch": 20.29344978165939,
+ "grad_norm": 0.01849197782576084,
+ "learning_rate": 0.0006,
+ "loss": 4.399681091308594,
+ "step": 1461
+ },
+ {
+ "epoch": 20.307423580786025,
+ "grad_norm": 0.018336588516831398,
+ "learning_rate": 0.0006,
+ "loss": 4.414183616638184,
+ "step": 1462
+ },
+ {
+ "epoch": 20.321397379912664,
+ "grad_norm": 0.018086519092321396,
+ "learning_rate": 0.0006,
+ "loss": 4.346586227416992,
+ "step": 1463
+ },
+ {
+ "epoch": 20.335371179039303,
+ "grad_norm": 0.019697299227118492,
+ "learning_rate": 0.0006,
+ "loss": 4.4420952796936035,
+ "step": 1464
+ },
+ {
+ "epoch": 20.349344978165938,
+ "grad_norm": 0.02198074571788311,
+ "learning_rate": 0.0006,
+ "loss": 4.588729381561279,
+ "step": 1465
+ },
+ {
+ "epoch": 20.363318777292577,
+ "grad_norm": 0.022813329473137856,
+ "learning_rate": 0.0006,
+ "loss": 4.49729585647583,
+ "step": 1466
+ },
+ {
+ "epoch": 20.377292576419215,
+ "grad_norm": 0.022006649523973465,
+ "learning_rate": 0.0006,
+ "loss": 4.475641250610352,
+ "step": 1467
+ },
+ {
+ "epoch": 20.39126637554585,
+ "grad_norm": 0.023188438266515732,
+ "learning_rate": 0.0006,
+ "loss": 4.404322624206543,
+ "step": 1468
+ },
+ {
+ "epoch": 20.40524017467249,
+ "grad_norm": 0.022226877510547638,
+ "learning_rate": 0.0006,
+ "loss": 4.50904655456543,
+ "step": 1469
+ },
+ {
+ "epoch": 20.419213973799128,
+ "grad_norm": 0.021082637831568718,
+ "learning_rate": 0.0006,
+ "loss": 4.339011192321777,
+ "step": 1470
+ },
+ {
+ "epoch": 20.433187772925763,
+ "grad_norm": 0.02013542130589485,
+ "learning_rate": 0.0006,
+ "loss": 4.516942024230957,
+ "step": 1471
+ },
+ {
+ "epoch": 20.4471615720524,
+ "grad_norm": 0.01898968778550625,
+ "learning_rate": 0.0006,
+ "loss": 4.315804481506348,
+ "step": 1472
+ },
+ {
+ "epoch": 20.46113537117904,
+ "grad_norm": 0.019450880587100983,
+ "learning_rate": 0.0006,
+ "loss": 4.4794921875,
+ "step": 1473
+ },
+ {
+ "epoch": 20.475109170305675,
+ "grad_norm": 0.018577802926301956,
+ "learning_rate": 0.0006,
+ "loss": 4.422323703765869,
+ "step": 1474
+ },
+ {
+ "epoch": 20.489082969432314,
+ "grad_norm": 0.01778412237763405,
+ "learning_rate": 0.0006,
+ "loss": 4.429789066314697,
+ "step": 1475
+ },
+ {
+ "epoch": 20.503056768558952,
+ "grad_norm": 0.017134476453065872,
+ "learning_rate": 0.0006,
+ "loss": 4.416459083557129,
+ "step": 1476
+ },
+ {
+ "epoch": 20.51703056768559,
+ "grad_norm": 0.017153557389974594,
+ "learning_rate": 0.0006,
+ "loss": 4.446170806884766,
+ "step": 1477
+ },
+ {
+ "epoch": 20.531004366812226,
+ "grad_norm": 0.014998532831668854,
+ "learning_rate": 0.0006,
+ "loss": 4.419447898864746,
+ "step": 1478
+ },
+ {
+ "epoch": 20.544978165938865,
+ "grad_norm": 0.015877850353717804,
+ "learning_rate": 0.0006,
+ "loss": 4.312338829040527,
+ "step": 1479
+ },
+ {
+ "epoch": 20.558951965065503,
+ "grad_norm": 0.01569489948451519,
+ "learning_rate": 0.0006,
+ "loss": 4.572905540466309,
+ "step": 1480
+ },
+ {
+ "epoch": 20.57292576419214,
+ "grad_norm": 0.01591474376618862,
+ "learning_rate": 0.0006,
+ "loss": 4.414907455444336,
+ "step": 1481
+ },
+ {
+ "epoch": 20.586899563318777,
+ "grad_norm": 0.015395903028547764,
+ "learning_rate": 0.0006,
+ "loss": 4.549131393432617,
+ "step": 1482
+ },
+ {
+ "epoch": 20.600873362445416,
+ "grad_norm": 0.015715358778834343,
+ "learning_rate": 0.0006,
+ "loss": 4.530194282531738,
+ "step": 1483
+ },
+ {
+ "epoch": 20.61484716157205,
+ "grad_norm": 0.016042975708842278,
+ "learning_rate": 0.0006,
+ "loss": 4.322149276733398,
+ "step": 1484
+ },
+ {
+ "epoch": 20.62882096069869,
+ "grad_norm": 0.01664186641573906,
+ "learning_rate": 0.0006,
+ "loss": 4.424866676330566,
+ "step": 1485
+ },
+ {
+ "epoch": 20.64279475982533,
+ "grad_norm": 0.01631534844636917,
+ "learning_rate": 0.0006,
+ "loss": 4.441752910614014,
+ "step": 1486
+ },
+ {
+ "epoch": 20.656768558951963,
+ "grad_norm": 0.014057058840990067,
+ "learning_rate": 0.0006,
+ "loss": 4.399982452392578,
+ "step": 1487
+ },
+ {
+ "epoch": 20.670742358078602,
+ "grad_norm": 0.015076273120939732,
+ "learning_rate": 0.0006,
+ "loss": 4.441324234008789,
+ "step": 1488
+ },
+ {
+ "epoch": 20.68471615720524,
+ "grad_norm": 0.014138277620077133,
+ "learning_rate": 0.0006,
+ "loss": 4.4312543869018555,
+ "step": 1489
+ },
+ {
+ "epoch": 20.69868995633188,
+ "grad_norm": 0.013762440532445908,
+ "learning_rate": 0.0006,
+ "loss": 4.387807846069336,
+ "step": 1490
+ },
+ {
+ "epoch": 20.712663755458514,
+ "grad_norm": 0.015058411285281181,
+ "learning_rate": 0.0006,
+ "loss": 4.594747543334961,
+ "step": 1491
+ },
+ {
+ "epoch": 20.726637554585153,
+ "grad_norm": 0.013827123679220676,
+ "learning_rate": 0.0006,
+ "loss": 4.434604167938232,
+ "step": 1492
+ },
+ {
+ "epoch": 20.74061135371179,
+ "grad_norm": 0.015315295197069645,
+ "learning_rate": 0.0006,
+ "loss": 4.588066101074219,
+ "step": 1493
+ },
+ {
+ "epoch": 20.754585152838427,
+ "grad_norm": 0.018678732216358185,
+ "learning_rate": 0.0006,
+ "loss": 4.3514204025268555,
+ "step": 1494
+ },
+ {
+ "epoch": 20.768558951965066,
+ "grad_norm": 0.019521350041031837,
+ "learning_rate": 0.0006,
+ "loss": 4.437658786773682,
+ "step": 1495
+ },
+ {
+ "epoch": 20.782532751091704,
+ "grad_norm": 0.018363403156399727,
+ "learning_rate": 0.0006,
+ "loss": 4.495002269744873,
+ "step": 1496
+ },
+ {
+ "epoch": 20.79650655021834,
+ "grad_norm": 0.016830725595355034,
+ "learning_rate": 0.0006,
+ "loss": 4.402471542358398,
+ "step": 1497
+ },
+ {
+ "epoch": 20.810480349344978,
+ "grad_norm": 0.01722072623670101,
+ "learning_rate": 0.0006,
+ "loss": 4.437653541564941,
+ "step": 1498
+ },
+ {
+ "epoch": 20.824454148471617,
+ "grad_norm": 0.017459599301218987,
+ "learning_rate": 0.0006,
+ "loss": 4.366603851318359,
+ "step": 1499
+ },
+ {
+ "epoch": 20.83842794759825,
+ "grad_norm": 0.018902219831943512,
+ "learning_rate": 0.0006,
+ "loss": 4.501561164855957,
+ "step": 1500
+ },
+ {
+ "epoch": 20.85240174672489,
+ "grad_norm": 0.020814096555113792,
+ "learning_rate": 0.0006,
+ "loss": 4.535295486450195,
+ "step": 1501
+ },
+ {
+ "epoch": 20.86637554585153,
+ "grad_norm": 0.018990976735949516,
+ "learning_rate": 0.0006,
+ "loss": 4.333580017089844,
+ "step": 1502
+ },
+ {
+ "epoch": 20.880349344978168,
+ "grad_norm": 0.01837880350649357,
+ "learning_rate": 0.0006,
+ "loss": 4.313388824462891,
+ "step": 1503
+ },
+ {
+ "epoch": 20.894323144104803,
+ "grad_norm": 0.019820360466837883,
+ "learning_rate": 0.0006,
+ "loss": 4.388157844543457,
+ "step": 1504
+ },
+ {
+ "epoch": 20.90829694323144,
+ "grad_norm": 0.020181143656373024,
+ "learning_rate": 0.0006,
+ "loss": 4.572224140167236,
+ "step": 1505
+ },
+ {
+ "epoch": 20.92227074235808,
+ "grad_norm": 0.021206077188253403,
+ "learning_rate": 0.0006,
+ "loss": 4.3127970695495605,
+ "step": 1506
+ },
+ {
+ "epoch": 20.936244541484715,
+ "grad_norm": 0.017962060868740082,
+ "learning_rate": 0.0006,
+ "loss": 4.384734630584717,
+ "step": 1507
+ },
+ {
+ "epoch": 20.950218340611354,
+ "grad_norm": 0.01737220585346222,
+ "learning_rate": 0.0006,
+ "loss": 4.4075751304626465,
+ "step": 1508
+ },
+ {
+ "epoch": 20.964192139737992,
+ "grad_norm": 0.01839268207550049,
+ "learning_rate": 0.0006,
+ "loss": 4.553779125213623,
+ "step": 1509
+ },
+ {
+ "epoch": 20.978165938864628,
+ "grad_norm": 0.020102640613913536,
+ "learning_rate": 0.0006,
+ "loss": 4.399398326873779,
+ "step": 1510
+ },
+ {
+ "epoch": 20.992139737991266,
+ "grad_norm": 0.018831463530659676,
+ "learning_rate": 0.0006,
+ "loss": 4.358499526977539,
+ "step": 1511
+ },
+ {
+ "epoch": 21.0,
+ "grad_norm": 0.018254734575748444,
+ "learning_rate": 0.0006,
+ "loss": 4.384119510650635,
+ "step": 1512
+ },
+ {
+ "epoch": 21.0,
+ "eval_loss": 4.7524237632751465,
+ "eval_runtime": 57.8812,
+ "eval_samples_per_second": 42.19,
+ "eval_steps_per_second": 1.33,
+ "step": 1512
+ },
+ {
+ "epoch": 21.01397379912664,
+ "grad_norm": 0.019728587940335274,
+ "learning_rate": 0.0006,
+ "loss": 4.32363224029541,
+ "step": 1513
+ },
+ {
+ "epoch": 21.027947598253274,
+ "grad_norm": 0.02216625027358532,
+ "learning_rate": 0.0006,
+ "loss": 4.436988353729248,
+ "step": 1514
+ },
+ {
+ "epoch": 21.041921397379912,
+ "grad_norm": 0.02227606810629368,
+ "learning_rate": 0.0006,
+ "loss": 4.406063556671143,
+ "step": 1515
+ },
+ {
+ "epoch": 21.05589519650655,
+ "grad_norm": 0.020187009125947952,
+ "learning_rate": 0.0006,
+ "loss": 4.324402332305908,
+ "step": 1516
+ },
+ {
+ "epoch": 21.069868995633186,
+ "grad_norm": 0.02095157466828823,
+ "learning_rate": 0.0006,
+ "loss": 4.456544876098633,
+ "step": 1517
+ },
+ {
+ "epoch": 21.083842794759825,
+ "grad_norm": 0.02146013453602791,
+ "learning_rate": 0.0006,
+ "loss": 4.368387222290039,
+ "step": 1518
+ },
+ {
+ "epoch": 21.097816593886463,
+ "grad_norm": 0.02162035182118416,
+ "learning_rate": 0.0006,
+ "loss": 4.48838996887207,
+ "step": 1519
+ },
+ {
+ "epoch": 21.111790393013102,
+ "grad_norm": 0.021745676174759865,
+ "learning_rate": 0.0006,
+ "loss": 4.30778694152832,
+ "step": 1520
+ },
+ {
+ "epoch": 21.125764192139737,
+ "grad_norm": 0.02173588052392006,
+ "learning_rate": 0.0006,
+ "loss": 4.433149814605713,
+ "step": 1521
+ },
+ {
+ "epoch": 21.139737991266376,
+ "grad_norm": 0.020950648933649063,
+ "learning_rate": 0.0006,
+ "loss": 4.520089149475098,
+ "step": 1522
+ },
+ {
+ "epoch": 21.153711790393015,
+ "grad_norm": 0.020684752613306046,
+ "learning_rate": 0.0006,
+ "loss": 4.372437477111816,
+ "step": 1523
+ },
+ {
+ "epoch": 21.16768558951965,
+ "grad_norm": 0.019472869113087654,
+ "learning_rate": 0.0006,
+ "loss": 4.437015533447266,
+ "step": 1524
+ },
+ {
+ "epoch": 21.18165938864629,
+ "grad_norm": 0.01846962980926037,
+ "learning_rate": 0.0006,
+ "loss": 4.449531555175781,
+ "step": 1525
+ },
+ {
+ "epoch": 21.195633187772927,
+ "grad_norm": 0.017377803102135658,
+ "learning_rate": 0.0006,
+ "loss": 4.458279132843018,
+ "step": 1526
+ },
+ {
+ "epoch": 21.209606986899562,
+ "grad_norm": 0.016002200543880463,
+ "learning_rate": 0.0006,
+ "loss": 4.266297817230225,
+ "step": 1527
+ },
+ {
+ "epoch": 21.2235807860262,
+ "grad_norm": 0.015575903467833996,
+ "learning_rate": 0.0006,
+ "loss": 4.342632293701172,
+ "step": 1528
+ },
+ {
+ "epoch": 21.23755458515284,
+ "grad_norm": 0.015634698793292046,
+ "learning_rate": 0.0006,
+ "loss": 4.338623523712158,
+ "step": 1529
+ },
+ {
+ "epoch": 21.251528384279474,
+ "grad_norm": 0.01449581328779459,
+ "learning_rate": 0.0006,
+ "loss": 4.354330062866211,
+ "step": 1530
+ },
+ {
+ "epoch": 21.265502183406113,
+ "grad_norm": 0.01487616915255785,
+ "learning_rate": 0.0006,
+ "loss": 4.486835479736328,
+ "step": 1531
+ },
+ {
+ "epoch": 21.27947598253275,
+ "grad_norm": 0.016346555203199387,
+ "learning_rate": 0.0006,
+ "loss": 4.407098770141602,
+ "step": 1532
+ },
+ {
+ "epoch": 21.29344978165939,
+ "grad_norm": 0.01682182401418686,
+ "learning_rate": 0.0006,
+ "loss": 4.436809539794922,
+ "step": 1533
+ },
+ {
+ "epoch": 21.307423580786025,
+ "grad_norm": 0.017704004421830177,
+ "learning_rate": 0.0006,
+ "loss": 4.3854780197143555,
+ "step": 1534
+ },
+ {
+ "epoch": 21.321397379912664,
+ "grad_norm": 0.018437420949339867,
+ "learning_rate": 0.0006,
+ "loss": 4.413283824920654,
+ "step": 1535
+ },
+ {
+ "epoch": 21.335371179039303,
+ "grad_norm": 0.016911081969738007,
+ "learning_rate": 0.0006,
+ "loss": 4.357306003570557,
+ "step": 1536
+ },
+ {
+ "epoch": 21.349344978165938,
+ "grad_norm": 0.01642223633825779,
+ "learning_rate": 0.0006,
+ "loss": 4.422322750091553,
+ "step": 1537
+ },
+ {
+ "epoch": 21.363318777292577,
+ "grad_norm": 0.01780877821147442,
+ "learning_rate": 0.0006,
+ "loss": 4.387620449066162,
+ "step": 1538
+ },
+ {
+ "epoch": 21.377292576419215,
+ "grad_norm": 0.01653975248336792,
+ "learning_rate": 0.0006,
+ "loss": 4.327267646789551,
+ "step": 1539
+ },
+ {
+ "epoch": 21.39126637554585,
+ "grad_norm": 0.01572156324982643,
+ "learning_rate": 0.0006,
+ "loss": 4.475979804992676,
+ "step": 1540
+ },
+ {
+ "epoch": 21.40524017467249,
+ "grad_norm": 0.017156066372990608,
+ "learning_rate": 0.0006,
+ "loss": 4.4011125564575195,
+ "step": 1541
+ },
+ {
+ "epoch": 21.419213973799128,
+ "grad_norm": 0.01788829080760479,
+ "learning_rate": 0.0006,
+ "loss": 4.4174089431762695,
+ "step": 1542
+ },
+ {
+ "epoch": 21.433187772925763,
+ "grad_norm": 0.01728985086083412,
+ "learning_rate": 0.0006,
+ "loss": 4.429244041442871,
+ "step": 1543
+ },
+ {
+ "epoch": 21.4471615720524,
+ "grad_norm": 0.016534466296434402,
+ "learning_rate": 0.0006,
+ "loss": 4.326068878173828,
+ "step": 1544
+ },
+ {
+ "epoch": 21.46113537117904,
+ "grad_norm": 0.016582584008574486,
+ "learning_rate": 0.0006,
+ "loss": 4.2975969314575195,
+ "step": 1545
+ },
+ {
+ "epoch": 21.475109170305675,
+ "grad_norm": 0.018034178763628006,
+ "learning_rate": 0.0006,
+ "loss": 4.501319408416748,
+ "step": 1546
+ },
+ {
+ "epoch": 21.489082969432314,
+ "grad_norm": 0.017827702686190605,
+ "learning_rate": 0.0006,
+ "loss": 4.502284526824951,
+ "step": 1547
+ },
+ {
+ "epoch": 21.503056768558952,
+ "grad_norm": 0.017182299867272377,
+ "learning_rate": 0.0006,
+ "loss": 4.380828857421875,
+ "step": 1548
+ },
+ {
+ "epoch": 21.51703056768559,
+ "grad_norm": 0.019101588055491447,
+ "learning_rate": 0.0006,
+ "loss": 4.31564998626709,
+ "step": 1549
+ },
+ {
+ "epoch": 21.531004366812226,
+ "grad_norm": 0.0204677302390337,
+ "learning_rate": 0.0006,
+ "loss": 4.3840131759643555,
+ "step": 1550
+ },
+ {
+ "epoch": 21.544978165938865,
+ "grad_norm": 0.0188993439078331,
+ "learning_rate": 0.0006,
+ "loss": 4.4257731437683105,
+ "step": 1551
+ },
+ {
+ "epoch": 21.558951965065503,
+ "grad_norm": 0.01730230636894703,
+ "learning_rate": 0.0006,
+ "loss": 4.445132255554199,
+ "step": 1552
+ },
+ {
+ "epoch": 21.57292576419214,
+ "grad_norm": 0.01801574043929577,
+ "learning_rate": 0.0006,
+ "loss": 4.427064895629883,
+ "step": 1553
+ },
+ {
+ "epoch": 21.586899563318777,
+ "grad_norm": 0.018417565152049065,
+ "learning_rate": 0.0006,
+ "loss": 4.319595813751221,
+ "step": 1554
+ },
+ {
+ "epoch": 21.600873362445416,
+ "grad_norm": 0.01683160290122032,
+ "learning_rate": 0.0006,
+ "loss": 4.349218368530273,
+ "step": 1555
+ },
+ {
+ "epoch": 21.61484716157205,
+ "grad_norm": 0.016973106190562248,
+ "learning_rate": 0.0006,
+ "loss": 4.375990390777588,
+ "step": 1556
+ },
+ {
+ "epoch": 21.62882096069869,
+ "grad_norm": 0.015688583254814148,
+ "learning_rate": 0.0006,
+ "loss": 4.408170223236084,
+ "step": 1557
+ },
+ {
+ "epoch": 21.64279475982533,
+ "grad_norm": 0.01573043316602707,
+ "learning_rate": 0.0006,
+ "loss": 4.416255950927734,
+ "step": 1558
+ },
+ {
+ "epoch": 21.656768558951963,
+ "grad_norm": 0.01744293048977852,
+ "learning_rate": 0.0006,
+ "loss": 4.477141380310059,
+ "step": 1559
+ },
+ {
+ "epoch": 21.670742358078602,
+ "grad_norm": 0.017503513023257256,
+ "learning_rate": 0.0006,
+ "loss": 4.3367486000061035,
+ "step": 1560
+ },
+ {
+ "epoch": 21.68471615720524,
+ "grad_norm": 0.016592150554060936,
+ "learning_rate": 0.0006,
+ "loss": 4.314785480499268,
+ "step": 1561
+ },
+ {
+ "epoch": 21.69868995633188,
+ "grad_norm": 0.015774084255099297,
+ "learning_rate": 0.0006,
+ "loss": 4.448209285736084,
+ "step": 1562
+ },
+ {
+ "epoch": 21.712663755458514,
+ "grad_norm": 0.015779150649905205,
+ "learning_rate": 0.0006,
+ "loss": 4.392690658569336,
+ "step": 1563
+ },
+ {
+ "epoch": 21.726637554585153,
+ "grad_norm": 0.015392648056149483,
+ "learning_rate": 0.0006,
+ "loss": 4.437741756439209,
+ "step": 1564
+ },
+ {
+ "epoch": 21.74061135371179,
+ "grad_norm": 0.01530836895108223,
+ "learning_rate": 0.0006,
+ "loss": 4.424595832824707,
+ "step": 1565
+ },
+ {
+ "epoch": 21.754585152838427,
+ "grad_norm": 0.015600779093801975,
+ "learning_rate": 0.0006,
+ "loss": 4.335417747497559,
+ "step": 1566
+ },
+ {
+ "epoch": 21.768558951965066,
+ "grad_norm": 0.014646518044173717,
+ "learning_rate": 0.0006,
+ "loss": 4.279081344604492,
+ "step": 1567
+ },
+ {
+ "epoch": 21.782532751091704,
+ "grad_norm": 0.014132864773273468,
+ "learning_rate": 0.0006,
+ "loss": 4.345950126647949,
+ "step": 1568
+ },
+ {
+ "epoch": 21.79650655021834,
+ "grad_norm": 0.01425724383443594,
+ "learning_rate": 0.0006,
+ "loss": 4.277588844299316,
+ "step": 1569
+ },
+ {
+ "epoch": 21.810480349344978,
+ "grad_norm": 0.015396572649478912,
+ "learning_rate": 0.0006,
+ "loss": 4.326196670532227,
+ "step": 1570
+ },
+ {
+ "epoch": 21.824454148471617,
+ "grad_norm": 0.01683669537305832,
+ "learning_rate": 0.0006,
+ "loss": 4.3294172286987305,
+ "step": 1571
+ },
+ {
+ "epoch": 21.83842794759825,
+ "grad_norm": 0.016365647315979004,
+ "learning_rate": 0.0006,
+ "loss": 4.523321628570557,
+ "step": 1572
+ },
+ {
+ "epoch": 21.85240174672489,
+ "grad_norm": 0.017142774537205696,
+ "learning_rate": 0.0006,
+ "loss": 4.475842475891113,
+ "step": 1573
+ },
+ {
+ "epoch": 21.86637554585153,
+ "grad_norm": 0.015950839966535568,
+ "learning_rate": 0.0006,
+ "loss": 4.397187232971191,
+ "step": 1574
+ },
+ {
+ "epoch": 21.880349344978168,
+ "grad_norm": 0.016180362552404404,
+ "learning_rate": 0.0006,
+ "loss": 4.414778709411621,
+ "step": 1575
+ },
+ {
+ "epoch": 21.894323144104803,
+ "grad_norm": 0.017615556716918945,
+ "learning_rate": 0.0006,
+ "loss": 4.348567962646484,
+ "step": 1576
+ },
+ {
+ "epoch": 21.90829694323144,
+ "grad_norm": 0.01896277628839016,
+ "learning_rate": 0.0006,
+ "loss": 4.431762218475342,
+ "step": 1577
+ },
+ {
+ "epoch": 21.92227074235808,
+ "grad_norm": 0.01990971714258194,
+ "learning_rate": 0.0006,
+ "loss": 4.431285858154297,
+ "step": 1578
+ },
+ {
+ "epoch": 21.936244541484715,
+ "grad_norm": 0.0201258547604084,
+ "learning_rate": 0.0006,
+ "loss": 4.309872627258301,
+ "step": 1579
+ },
+ {
+ "epoch": 21.950218340611354,
+ "grad_norm": 0.019620198756456375,
+ "learning_rate": 0.0006,
+ "loss": 4.484044551849365,
+ "step": 1580
+ },
+ {
+ "epoch": 21.964192139737992,
+ "grad_norm": 0.01771123707294464,
+ "learning_rate": 0.0006,
+ "loss": 4.368093013763428,
+ "step": 1581
+ },
+ {
+ "epoch": 21.978165938864628,
+ "grad_norm": 0.017501654103398323,
+ "learning_rate": 0.0006,
+ "loss": 4.372651100158691,
+ "step": 1582
+ },
+ {
+ "epoch": 21.992139737991266,
+ "grad_norm": 0.01693608984351158,
+ "learning_rate": 0.0006,
+ "loss": 4.379167556762695,
+ "step": 1583
+ },
+ {
+ "epoch": 22.0,
+ "grad_norm": 0.02002849616110325,
+ "learning_rate": 0.0006,
+ "loss": 4.329627990722656,
+ "step": 1584
+ },
+ {
+ "epoch": 22.0,
+ "eval_loss": 4.723696231842041,
+ "eval_runtime": 56.1758,
+ "eval_samples_per_second": 43.471,
+ "eval_steps_per_second": 1.371,
+ "step": 1584
+ },
+ {
+ "epoch": 22.01397379912664,
+ "grad_norm": 0.02030305378139019,
+ "learning_rate": 0.0006,
+ "loss": 4.355566024780273,
+ "step": 1585
+ },
+ {
+ "epoch": 22.027947598253274,
+ "grad_norm": 0.018638934940099716,
+ "learning_rate": 0.0006,
+ "loss": 4.314009189605713,
+ "step": 1586
+ },
+ {
+ "epoch": 22.041921397379912,
+ "grad_norm": 0.019917072728276253,
+ "learning_rate": 0.0006,
+ "loss": 4.367525577545166,
+ "step": 1587
+ },
+ {
+ "epoch": 22.05589519650655,
+ "grad_norm": 0.020785167813301086,
+ "learning_rate": 0.0006,
+ "loss": 4.375561714172363,
+ "step": 1588
+ },
+ {
+ "epoch": 22.069868995633186,
+ "grad_norm": 0.021214453503489494,
+ "learning_rate": 0.0006,
+ "loss": 4.418305397033691,
+ "step": 1589
+ },
+ {
+ "epoch": 22.083842794759825,
+ "grad_norm": 0.019051678478717804,
+ "learning_rate": 0.0006,
+ "loss": 4.215126991271973,
+ "step": 1590
+ },
+ {
+ "epoch": 22.097816593886463,
+ "grad_norm": 0.02323366142809391,
+ "learning_rate": 0.0006,
+ "loss": 4.306807994842529,
+ "step": 1591
+ },
+ {
+ "epoch": 22.111790393013102,
+ "grad_norm": 0.026186738163232803,
+ "learning_rate": 0.0006,
+ "loss": 4.408508777618408,
+ "step": 1592
+ },
+ {
+ "epoch": 22.125764192139737,
+ "grad_norm": 0.027361227199435234,
+ "learning_rate": 0.0006,
+ "loss": 4.329322814941406,
+ "step": 1593
+ },
+ {
+ "epoch": 22.139737991266376,
+ "grad_norm": 0.026564646512269974,
+ "learning_rate": 0.0006,
+ "loss": 4.346642017364502,
+ "step": 1594
+ },
+ {
+ "epoch": 22.153711790393015,
+ "grad_norm": 0.025073667988181114,
+ "learning_rate": 0.0006,
+ "loss": 4.3984198570251465,
+ "step": 1595
+ },
+ {
+ "epoch": 22.16768558951965,
+ "grad_norm": 0.02347794733941555,
+ "learning_rate": 0.0006,
+ "loss": 4.352538108825684,
+ "step": 1596
+ },
+ {
+ "epoch": 22.18165938864629,
+ "grad_norm": 0.02203468047082424,
+ "learning_rate": 0.0006,
+ "loss": 4.298148155212402,
+ "step": 1597
+ },
+ {
+ "epoch": 22.195633187772927,
+ "grad_norm": 0.020621957257390022,
+ "learning_rate": 0.0006,
+ "loss": 4.480309963226318,
+ "step": 1598
+ },
+ {
+ "epoch": 22.209606986899562,
+ "grad_norm": 0.023710055276751518,
+ "learning_rate": 0.0006,
+ "loss": 4.339317321777344,
+ "step": 1599
+ },
+ {
+ "epoch": 22.2235807860262,
+ "grad_norm": 0.0233482476323843,
+ "learning_rate": 0.0006,
+ "loss": 4.436896324157715,
+ "step": 1600
+ },
+ {
+ "epoch": 22.23755458515284,
+ "grad_norm": 0.020122403278946877,
+ "learning_rate": 0.0006,
+ "loss": 4.236002445220947,
+ "step": 1601
+ },
+ {
+ "epoch": 22.251528384279474,
+ "grad_norm": 0.01957266591489315,
+ "learning_rate": 0.0006,
+ "loss": 4.497328281402588,
+ "step": 1602
+ },
+ {
+ "epoch": 22.265502183406113,
+ "grad_norm": 0.018964895978569984,
+ "learning_rate": 0.0006,
+ "loss": 4.36360502243042,
+ "step": 1603
+ },
+ {
+ "epoch": 22.27947598253275,
+ "grad_norm": 0.018902769312262535,
+ "learning_rate": 0.0006,
+ "loss": 4.427093505859375,
+ "step": 1604
+ },
+ {
+ "epoch": 22.29344978165939,
+ "grad_norm": 0.018918119370937347,
+ "learning_rate": 0.0006,
+ "loss": 4.294950485229492,
+ "step": 1605
+ },
+ {
+ "epoch": 22.307423580786025,
+ "grad_norm": 0.019315775483846664,
+ "learning_rate": 0.0006,
+ "loss": 4.427952289581299,
+ "step": 1606
+ },
+ {
+ "epoch": 22.321397379912664,
+ "grad_norm": 0.018350522965192795,
+ "learning_rate": 0.0006,
+ "loss": 4.272945404052734,
+ "step": 1607
+ },
+ {
+ "epoch": 22.335371179039303,
+ "grad_norm": 0.01759534701704979,
+ "learning_rate": 0.0006,
+ "loss": 4.348023891448975,
+ "step": 1608
+ },
+ {
+ "epoch": 22.349344978165938,
+ "grad_norm": 0.01646057888865471,
+ "learning_rate": 0.0006,
+ "loss": 4.278518199920654,
+ "step": 1609
+ },
+ {
+ "epoch": 22.363318777292577,
+ "grad_norm": 0.015875060111284256,
+ "learning_rate": 0.0006,
+ "loss": 4.423527240753174,
+ "step": 1610
+ },
+ {
+ "epoch": 22.377292576419215,
+ "grad_norm": 0.01638529635965824,
+ "learning_rate": 0.0006,
+ "loss": 4.321723937988281,
+ "step": 1611
+ },
+ {
+ "epoch": 22.39126637554585,
+ "grad_norm": 0.016942061483860016,
+ "learning_rate": 0.0006,
+ "loss": 4.346164703369141,
+ "step": 1612
+ },
+ {
+ "epoch": 22.40524017467249,
+ "grad_norm": 0.015209621749818325,
+ "learning_rate": 0.0006,
+ "loss": 4.434340000152588,
+ "step": 1613
+ },
+ {
+ "epoch": 22.419213973799128,
+ "grad_norm": 0.014423009008169174,
+ "learning_rate": 0.0006,
+ "loss": 4.342560768127441,
+ "step": 1614
+ },
+ {
+ "epoch": 22.433187772925763,
+ "grad_norm": 0.014369525946676731,
+ "learning_rate": 0.0006,
+ "loss": 4.296586990356445,
+ "step": 1615
+ },
+ {
+ "epoch": 22.4471615720524,
+ "grad_norm": 0.01392540242522955,
+ "learning_rate": 0.0006,
+ "loss": 4.433176040649414,
+ "step": 1616
+ },
+ {
+ "epoch": 22.46113537117904,
+ "grad_norm": 0.013206899166107178,
+ "learning_rate": 0.0006,
+ "loss": 4.32962703704834,
+ "step": 1617
+ },
+ {
+ "epoch": 22.475109170305675,
+ "grad_norm": 0.013427077792584896,
+ "learning_rate": 0.0006,
+ "loss": 4.420437335968018,
+ "step": 1618
+ },
+ {
+ "epoch": 22.489082969432314,
+ "grad_norm": 0.014194196090102196,
+ "learning_rate": 0.0006,
+ "loss": 4.410478591918945,
+ "step": 1619
+ },
+ {
+ "epoch": 22.503056768558952,
+ "grad_norm": 0.01362364087253809,
+ "learning_rate": 0.0006,
+ "loss": 4.338932037353516,
+ "step": 1620
+ },
+ {
+ "epoch": 22.51703056768559,
+ "grad_norm": 0.014299589209258556,
+ "learning_rate": 0.0006,
+ "loss": 4.34089994430542,
+ "step": 1621
+ },
+ {
+ "epoch": 22.531004366812226,
+ "grad_norm": 0.014061295427381992,
+ "learning_rate": 0.0006,
+ "loss": 4.29201602935791,
+ "step": 1622
+ },
+ {
+ "epoch": 22.544978165938865,
+ "grad_norm": 0.013945229351520538,
+ "learning_rate": 0.0006,
+ "loss": 4.326052188873291,
+ "step": 1623
+ },
+ {
+ "epoch": 22.558951965065503,
+ "grad_norm": 0.01387068536132574,
+ "learning_rate": 0.0006,
+ "loss": 4.393521308898926,
+ "step": 1624
+ },
+ {
+ "epoch": 22.57292576419214,
+ "grad_norm": 0.013931898400187492,
+ "learning_rate": 0.0006,
+ "loss": 4.3739752769470215,
+ "step": 1625
+ },
+ {
+ "epoch": 22.586899563318777,
+ "grad_norm": 0.01497740764170885,
+ "learning_rate": 0.0006,
+ "loss": 4.3991618156433105,
+ "step": 1626
+ },
+ {
+ "epoch": 22.600873362445416,
+ "grad_norm": 0.015842905268073082,
+ "learning_rate": 0.0006,
+ "loss": 4.405340671539307,
+ "step": 1627
+ },
+ {
+ "epoch": 22.61484716157205,
+ "grad_norm": 0.01580633595585823,
+ "learning_rate": 0.0006,
+ "loss": 4.280102729797363,
+ "step": 1628
+ },
+ {
+ "epoch": 22.62882096069869,
+ "grad_norm": 0.016567382961511612,
+ "learning_rate": 0.0006,
+ "loss": 4.407657146453857,
+ "step": 1629
+ },
+ {
+ "epoch": 22.64279475982533,
+ "grad_norm": 0.01648622378706932,
+ "learning_rate": 0.0006,
+ "loss": 4.392590522766113,
+ "step": 1630
+ },
+ {
+ "epoch": 22.656768558951963,
+ "grad_norm": 0.017066307365894318,
+ "learning_rate": 0.0006,
+ "loss": 4.324862003326416,
+ "step": 1631
+ },
+ {
+ "epoch": 22.670742358078602,
+ "grad_norm": 0.01637062057852745,
+ "learning_rate": 0.0006,
+ "loss": 4.34598445892334,
+ "step": 1632
+ },
+ {
+ "epoch": 22.68471615720524,
+ "grad_norm": 0.018038960173726082,
+ "learning_rate": 0.0006,
+ "loss": 4.320711612701416,
+ "step": 1633
+ },
+ {
+ "epoch": 22.69868995633188,
+ "grad_norm": 0.018096428364515305,
+ "learning_rate": 0.0006,
+ "loss": 4.536527156829834,
+ "step": 1634
+ },
+ {
+ "epoch": 22.712663755458514,
+ "grad_norm": 0.017889291048049927,
+ "learning_rate": 0.0006,
+ "loss": 4.388131141662598,
+ "step": 1635
+ },
+ {
+ "epoch": 22.726637554585153,
+ "grad_norm": 0.018771521747112274,
+ "learning_rate": 0.0006,
+ "loss": 4.357668876647949,
+ "step": 1636
+ },
+ {
+ "epoch": 22.74061135371179,
+ "grad_norm": 0.02100873365998268,
+ "learning_rate": 0.0006,
+ "loss": 4.293972015380859,
+ "step": 1637
+ },
+ {
+ "epoch": 22.754585152838427,
+ "grad_norm": 0.017715785652399063,
+ "learning_rate": 0.0006,
+ "loss": 4.349849224090576,
+ "step": 1638
+ },
+ {
+ "epoch": 22.768558951965066,
+ "grad_norm": 0.015546333976089954,
+ "learning_rate": 0.0006,
+ "loss": 4.311728000640869,
+ "step": 1639
+ },
+ {
+ "epoch": 22.782532751091704,
+ "grad_norm": 0.01605990156531334,
+ "learning_rate": 0.0006,
+ "loss": 4.499123573303223,
+ "step": 1640
+ },
+ {
+ "epoch": 22.79650655021834,
+ "grad_norm": 0.016642142087221146,
+ "learning_rate": 0.0006,
+ "loss": 4.4319257736206055,
+ "step": 1641
+ },
+ {
+ "epoch": 22.810480349344978,
+ "grad_norm": 0.016120387241244316,
+ "learning_rate": 0.0006,
+ "loss": 4.4373650550842285,
+ "step": 1642
+ },
+ {
+ "epoch": 22.824454148471617,
+ "grad_norm": 0.015174890868365765,
+ "learning_rate": 0.0006,
+ "loss": 4.314082145690918,
+ "step": 1643
+ },
+ {
+ "epoch": 22.83842794759825,
+ "grad_norm": 0.01551714539527893,
+ "learning_rate": 0.0006,
+ "loss": 4.360363960266113,
+ "step": 1644
+ },
+ {
+ "epoch": 22.85240174672489,
+ "grad_norm": 0.014966742135584354,
+ "learning_rate": 0.0006,
+ "loss": 4.346253395080566,
+ "step": 1645
+ },
+ {
+ "epoch": 22.86637554585153,
+ "grad_norm": 0.014288028702139854,
+ "learning_rate": 0.0006,
+ "loss": 4.464366436004639,
+ "step": 1646
+ },
+ {
+ "epoch": 22.880349344978168,
+ "grad_norm": 0.015291682444512844,
+ "learning_rate": 0.0006,
+ "loss": 4.504101753234863,
+ "step": 1647
+ },
+ {
+ "epoch": 22.894323144104803,
+ "grad_norm": 0.014534426853060722,
+ "learning_rate": 0.0006,
+ "loss": 4.462172031402588,
+ "step": 1648
+ },
+ {
+ "epoch": 22.90829694323144,
+ "grad_norm": 0.014129819348454475,
+ "learning_rate": 0.0006,
+ "loss": 4.350126266479492,
+ "step": 1649
+ },
+ {
+ "epoch": 22.92227074235808,
+ "grad_norm": 0.014932668767869473,
+ "learning_rate": 0.0006,
+ "loss": 4.374392032623291,
+ "step": 1650
+ },
+ {
+ "epoch": 22.936244541484715,
+ "grad_norm": 0.014647853560745716,
+ "learning_rate": 0.0006,
+ "loss": 4.252264499664307,
+ "step": 1651
+ },
+ {
+ "epoch": 22.950218340611354,
+ "grad_norm": 0.013553360477089882,
+ "learning_rate": 0.0006,
+ "loss": 4.379729270935059,
+ "step": 1652
+ },
+ {
+ "epoch": 22.964192139737992,
+ "grad_norm": 0.014606194570660591,
+ "learning_rate": 0.0006,
+ "loss": 4.363372802734375,
+ "step": 1653
+ },
+ {
+ "epoch": 22.978165938864628,
+ "grad_norm": 0.015000631101429462,
+ "learning_rate": 0.0006,
+ "loss": 4.408048629760742,
+ "step": 1654
+ },
+ {
+ "epoch": 22.992139737991266,
+ "grad_norm": 0.015370816923677921,
+ "learning_rate": 0.0006,
+ "loss": 4.297451019287109,
+ "step": 1655
+ },
+ {
+ "epoch": 23.0,
+ "grad_norm": 0.01709144562482834,
+ "learning_rate": 0.0006,
+ "loss": 4.294776916503906,
+ "step": 1656
+ },
+ {
+ "epoch": 23.0,
+ "eval_loss": 4.678618907928467,
+ "eval_runtime": 56.1668,
+ "eval_samples_per_second": 43.478,
+ "eval_steps_per_second": 1.371,
+ "step": 1656
+ },
+ {
+ "epoch": 23.01397379912664,
+ "grad_norm": 0.01727290451526642,
+ "learning_rate": 0.0006,
+ "loss": 4.311701774597168,
+ "step": 1657
+ },
+ {
+ "epoch": 23.027947598253274,
+ "grad_norm": 0.020645759999752045,
+ "learning_rate": 0.0006,
+ "loss": 4.429897308349609,
+ "step": 1658
+ },
+ {
+ "epoch": 23.041921397379912,
+ "grad_norm": 0.02204226516187191,
+ "learning_rate": 0.0006,
+ "loss": 4.365151882171631,
+ "step": 1659
+ },
+ {
+ "epoch": 23.05589519650655,
+ "grad_norm": 0.023069269955158234,
+ "learning_rate": 0.0006,
+ "loss": 4.3710832595825195,
+ "step": 1660
+ },
+ {
+ "epoch": 23.069868995633186,
+ "grad_norm": 0.024151179939508438,
+ "learning_rate": 0.0006,
+ "loss": 4.403108596801758,
+ "step": 1661
+ },
+ {
+ "epoch": 23.083842794759825,
+ "grad_norm": 0.02511233650147915,
+ "learning_rate": 0.0006,
+ "loss": 4.414128303527832,
+ "step": 1662
+ },
+ {
+ "epoch": 23.097816593886463,
+ "grad_norm": 0.025802113115787506,
+ "learning_rate": 0.0006,
+ "loss": 4.388265609741211,
+ "step": 1663
+ },
+ {
+ "epoch": 23.111790393013102,
+ "grad_norm": 0.02377348765730858,
+ "learning_rate": 0.0006,
+ "loss": 4.357505798339844,
+ "step": 1664
+ },
+ {
+ "epoch": 23.125764192139737,
+ "grad_norm": 0.020876651629805565,
+ "learning_rate": 0.0006,
+ "loss": 4.23043966293335,
+ "step": 1665
+ },
+ {
+ "epoch": 23.139737991266376,
+ "grad_norm": 0.01959727331995964,
+ "learning_rate": 0.0006,
+ "loss": 4.280970096588135,
+ "step": 1666
+ },
+ {
+ "epoch": 23.153711790393015,
+ "grad_norm": 0.018678616732358932,
+ "learning_rate": 0.0006,
+ "loss": 4.346716403961182,
+ "step": 1667
+ },
+ {
+ "epoch": 23.16768558951965,
+ "grad_norm": 0.02108944021165371,
+ "learning_rate": 0.0006,
+ "loss": 4.277904510498047,
+ "step": 1668
+ },
+ {
+ "epoch": 23.18165938864629,
+ "grad_norm": 0.020221566781401634,
+ "learning_rate": 0.0006,
+ "loss": 4.367693901062012,
+ "step": 1669
+ },
+ {
+ "epoch": 23.195633187772927,
+ "grad_norm": 0.019599711522459984,
+ "learning_rate": 0.0006,
+ "loss": 4.257092475891113,
+ "step": 1670
+ },
+ {
+ "epoch": 23.209606986899562,
+ "grad_norm": 0.019833361729979515,
+ "learning_rate": 0.0006,
+ "loss": 4.540737628936768,
+ "step": 1671
+ },
+ {
+ "epoch": 23.2235807860262,
+ "grad_norm": 0.020352095365524292,
+ "learning_rate": 0.0006,
+ "loss": 4.360923767089844,
+ "step": 1672
+ },
+ {
+ "epoch": 23.23755458515284,
+ "grad_norm": 0.01939748041331768,
+ "learning_rate": 0.0006,
+ "loss": 4.393110275268555,
+ "step": 1673
+ },
+ {
+ "epoch": 23.251528384279474,
+ "grad_norm": 0.01884354278445244,
+ "learning_rate": 0.0006,
+ "loss": 4.383938789367676,
+ "step": 1674
+ },
+ {
+ "epoch": 23.265502183406113,
+ "grad_norm": 0.02008470520377159,
+ "learning_rate": 0.0006,
+ "loss": 4.346344947814941,
+ "step": 1675
+ },
+ {
+ "epoch": 23.27947598253275,
+ "grad_norm": 0.019065195694565773,
+ "learning_rate": 0.0006,
+ "loss": 4.36611795425415,
+ "step": 1676
+ },
+ {
+ "epoch": 23.29344978165939,
+ "grad_norm": 0.01938541606068611,
+ "learning_rate": 0.0006,
+ "loss": 4.4009599685668945,
+ "step": 1677
+ },
+ {
+ "epoch": 23.307423580786025,
+ "grad_norm": 0.019044723361730576,
+ "learning_rate": 0.0006,
+ "loss": 4.340169906616211,
+ "step": 1678
+ },
+ {
+ "epoch": 23.321397379912664,
+ "grad_norm": 0.018850378692150116,
+ "learning_rate": 0.0006,
+ "loss": 4.37674617767334,
+ "step": 1679
+ },
+ {
+ "epoch": 23.335371179039303,
+ "grad_norm": 0.019076095893979073,
+ "learning_rate": 0.0006,
+ "loss": 4.226874351501465,
+ "step": 1680
+ },
+ {
+ "epoch": 23.349344978165938,
+ "grad_norm": 0.017543038353323936,
+ "learning_rate": 0.0006,
+ "loss": 4.337751388549805,
+ "step": 1681
+ },
+ {
+ "epoch": 23.363318777292577,
+ "grad_norm": 0.017170218750834465,
+ "learning_rate": 0.0006,
+ "loss": 4.282512664794922,
+ "step": 1682
+ },
+ {
+ "epoch": 23.377292576419215,
+ "grad_norm": 0.01574796624481678,
+ "learning_rate": 0.0006,
+ "loss": 4.30638313293457,
+ "step": 1683
+ },
+ {
+ "epoch": 23.39126637554585,
+ "grad_norm": 0.015160846523940563,
+ "learning_rate": 0.0006,
+ "loss": 4.336716175079346,
+ "step": 1684
+ },
+ {
+ "epoch": 23.40524017467249,
+ "grad_norm": 0.01534526702016592,
+ "learning_rate": 0.0006,
+ "loss": 4.2174787521362305,
+ "step": 1685
+ },
+ {
+ "epoch": 23.419213973799128,
+ "grad_norm": 0.015289999544620514,
+ "learning_rate": 0.0006,
+ "loss": 4.272546291351318,
+ "step": 1686
+ },
+ {
+ "epoch": 23.433187772925763,
+ "grad_norm": 0.014894185587763786,
+ "learning_rate": 0.0006,
+ "loss": 4.296045303344727,
+ "step": 1687
+ },
+ {
+ "epoch": 23.4471615720524,
+ "grad_norm": 0.014869502745568752,
+ "learning_rate": 0.0006,
+ "loss": 4.425461769104004,
+ "step": 1688
+ },
+ {
+ "epoch": 23.46113537117904,
+ "grad_norm": 0.01474483497440815,
+ "learning_rate": 0.0006,
+ "loss": 4.361867427825928,
+ "step": 1689
+ },
+ {
+ "epoch": 23.475109170305675,
+ "grad_norm": 0.015299245715141296,
+ "learning_rate": 0.0006,
+ "loss": 4.364523410797119,
+ "step": 1690
+ },
+ {
+ "epoch": 23.489082969432314,
+ "grad_norm": 0.015201340429484844,
+ "learning_rate": 0.0006,
+ "loss": 4.285182952880859,
+ "step": 1691
+ },
+ {
+ "epoch": 23.503056768558952,
+ "grad_norm": 0.016057293862104416,
+ "learning_rate": 0.0006,
+ "loss": 4.284873962402344,
+ "step": 1692
+ },
+ {
+ "epoch": 23.51703056768559,
+ "grad_norm": 0.015871062874794006,
+ "learning_rate": 0.0006,
+ "loss": 4.326470851898193,
+ "step": 1693
+ },
+ {
+ "epoch": 23.531004366812226,
+ "grad_norm": 0.01586288772523403,
+ "learning_rate": 0.0006,
+ "loss": 4.324883460998535,
+ "step": 1694
+ },
+ {
+ "epoch": 23.544978165938865,
+ "grad_norm": 0.016092827543616295,
+ "learning_rate": 0.0006,
+ "loss": 4.3258867263793945,
+ "step": 1695
+ },
+ {
+ "epoch": 23.558951965065503,
+ "grad_norm": 0.01582016795873642,
+ "learning_rate": 0.0006,
+ "loss": 4.358892440795898,
+ "step": 1696
+ },
+ {
+ "epoch": 23.57292576419214,
+ "grad_norm": 0.01688452623784542,
+ "learning_rate": 0.0006,
+ "loss": 4.280842304229736,
+ "step": 1697
+ },
+ {
+ "epoch": 23.586899563318777,
+ "grad_norm": 0.015865886583924294,
+ "learning_rate": 0.0006,
+ "loss": 4.245567321777344,
+ "step": 1698
+ },
+ {
+ "epoch": 23.600873362445416,
+ "grad_norm": 0.015303662046790123,
+ "learning_rate": 0.0006,
+ "loss": 4.383112907409668,
+ "step": 1699
+ },
+ {
+ "epoch": 23.61484716157205,
+ "grad_norm": 0.015649516135454178,
+ "learning_rate": 0.0006,
+ "loss": 4.519082069396973,
+ "step": 1700
+ },
+ {
+ "epoch": 23.62882096069869,
+ "grad_norm": 0.0167120099067688,
+ "learning_rate": 0.0006,
+ "loss": 4.447847366333008,
+ "step": 1701
+ },
+ {
+ "epoch": 23.64279475982533,
+ "grad_norm": 0.016368716955184937,
+ "learning_rate": 0.0006,
+ "loss": 4.434186935424805,
+ "step": 1702
+ },
+ {
+ "epoch": 23.656768558951963,
+ "grad_norm": 0.017379503697156906,
+ "learning_rate": 0.0006,
+ "loss": 4.317594528198242,
+ "step": 1703
+ },
+ {
+ "epoch": 23.670742358078602,
+ "grad_norm": 0.018573811277747154,
+ "learning_rate": 0.0006,
+ "loss": 4.205793857574463,
+ "step": 1704
+ },
+ {
+ "epoch": 23.68471615720524,
+ "grad_norm": 0.01749090477824211,
+ "learning_rate": 0.0006,
+ "loss": 4.3985700607299805,
+ "step": 1705
+ },
+ {
+ "epoch": 23.69868995633188,
+ "grad_norm": 0.0165668074041605,
+ "learning_rate": 0.0006,
+ "loss": 4.329226970672607,
+ "step": 1706
+ },
+ {
+ "epoch": 23.712663755458514,
+ "grad_norm": 0.016913846135139465,
+ "learning_rate": 0.0006,
+ "loss": 4.388485908508301,
+ "step": 1707
+ },
+ {
+ "epoch": 23.726637554585153,
+ "grad_norm": 0.016813859343528748,
+ "learning_rate": 0.0006,
+ "loss": 4.3395915031433105,
+ "step": 1708
+ },
+ {
+ "epoch": 23.74061135371179,
+ "grad_norm": 0.01676975190639496,
+ "learning_rate": 0.0006,
+ "loss": 4.254745960235596,
+ "step": 1709
+ },
+ {
+ "epoch": 23.754585152838427,
+ "grad_norm": 0.016370350494980812,
+ "learning_rate": 0.0006,
+ "loss": 4.2612223625183105,
+ "step": 1710
+ },
+ {
+ "epoch": 23.768558951965066,
+ "grad_norm": 0.01696198247373104,
+ "learning_rate": 0.0006,
+ "loss": 4.327524662017822,
+ "step": 1711
+ },
+ {
+ "epoch": 23.782532751091704,
+ "grad_norm": 0.0189108457416296,
+ "learning_rate": 0.0006,
+ "loss": 4.325839519500732,
+ "step": 1712
+ },
+ {
+ "epoch": 23.79650655021834,
+ "grad_norm": 0.01862351782619953,
+ "learning_rate": 0.0006,
+ "loss": 4.333662509918213,
+ "step": 1713
+ },
+ {
+ "epoch": 23.810480349344978,
+ "grad_norm": 0.01655014418065548,
+ "learning_rate": 0.0006,
+ "loss": 4.39305305480957,
+ "step": 1714
+ },
+ {
+ "epoch": 23.824454148471617,
+ "grad_norm": 0.015899376943707466,
+ "learning_rate": 0.0006,
+ "loss": 4.390053749084473,
+ "step": 1715
+ },
+ {
+ "epoch": 23.83842794759825,
+ "grad_norm": 0.016519056633114815,
+ "learning_rate": 0.0006,
+ "loss": 4.255903244018555,
+ "step": 1716
+ },
+ {
+ "epoch": 23.85240174672489,
+ "grad_norm": 0.0167539119720459,
+ "learning_rate": 0.0006,
+ "loss": 4.255727767944336,
+ "step": 1717
+ },
+ {
+ "epoch": 23.86637554585153,
+ "grad_norm": 0.0171353816986084,
+ "learning_rate": 0.0006,
+ "loss": 4.376740455627441,
+ "step": 1718
+ },
+ {
+ "epoch": 23.880349344978168,
+ "grad_norm": 0.016719117760658264,
+ "learning_rate": 0.0006,
+ "loss": 4.3376688957214355,
+ "step": 1719
+ },
+ {
+ "epoch": 23.894323144104803,
+ "grad_norm": 0.01561494916677475,
+ "learning_rate": 0.0006,
+ "loss": 4.3345746994018555,
+ "step": 1720
+ },
+ {
+ "epoch": 23.90829694323144,
+ "grad_norm": 0.016303174197673798,
+ "learning_rate": 0.0006,
+ "loss": 4.332371234893799,
+ "step": 1721
+ },
+ {
+ "epoch": 23.92227074235808,
+ "grad_norm": 0.016722865402698517,
+ "learning_rate": 0.0006,
+ "loss": 4.267125129699707,
+ "step": 1722
+ },
+ {
+ "epoch": 23.936244541484715,
+ "grad_norm": 0.017072124406695366,
+ "learning_rate": 0.0006,
+ "loss": 4.322751998901367,
+ "step": 1723
+ },
+ {
+ "epoch": 23.950218340611354,
+ "grad_norm": 0.01699932850897312,
+ "learning_rate": 0.0006,
+ "loss": 4.378422737121582,
+ "step": 1724
+ },
+ {
+ "epoch": 23.964192139737992,
+ "grad_norm": 0.01742720976471901,
+ "learning_rate": 0.0006,
+ "loss": 4.345083236694336,
+ "step": 1725
+ },
+ {
+ "epoch": 23.978165938864628,
+ "grad_norm": 0.018797501921653748,
+ "learning_rate": 0.0006,
+ "loss": 4.290443420410156,
+ "step": 1726
+ },
+ {
+ "epoch": 23.992139737991266,
+ "grad_norm": 0.018090683966875076,
+ "learning_rate": 0.0006,
+ "loss": 4.377580642700195,
+ "step": 1727
+ },
+ {
+ "epoch": 24.0,
+ "grad_norm": 0.017305459827184677,
+ "learning_rate": 0.0006,
+ "loss": 4.364426612854004,
+ "step": 1728
+ },
+ {
+ "epoch": 24.0,
+ "eval_loss": 4.647762298583984,
+ "eval_runtime": 56.5402,
+ "eval_samples_per_second": 43.191,
+ "eval_steps_per_second": 1.362,
+ "step": 1728
+ },
+ {
+ "epoch": 24.01397379912664,
+ "grad_norm": 0.016493607312440872,
+ "learning_rate": 0.0006,
+ "loss": 4.252255916595459,
+ "step": 1729
+ },
+ {
+ "epoch": 24.027947598253274,
+ "grad_norm": 0.016529450193047523,
+ "learning_rate": 0.0006,
+ "loss": 4.352746963500977,
+ "step": 1730
+ },
+ {
+ "epoch": 24.041921397379912,
+ "grad_norm": 0.016940191388130188,
+ "learning_rate": 0.0006,
+ "loss": 4.301870822906494,
+ "step": 1731
+ },
+ {
+ "epoch": 24.05589519650655,
+ "grad_norm": 0.01752558909356594,
+ "learning_rate": 0.0006,
+ "loss": 4.2762451171875,
+ "step": 1732
+ },
+ {
+ "epoch": 24.069868995633186,
+ "grad_norm": 0.017112884670495987,
+ "learning_rate": 0.0006,
+ "loss": 4.2659149169921875,
+ "step": 1733
+ },
+ {
+ "epoch": 24.083842794759825,
+ "grad_norm": 0.01900586113333702,
+ "learning_rate": 0.0006,
+ "loss": 4.306058883666992,
+ "step": 1734
+ },
+ {
+ "epoch": 24.097816593886463,
+ "grad_norm": 0.019447680562734604,
+ "learning_rate": 0.0006,
+ "loss": 4.450128555297852,
+ "step": 1735
+ },
+ {
+ "epoch": 24.111790393013102,
+ "grad_norm": 0.01948186755180359,
+ "learning_rate": 0.0006,
+ "loss": 4.202869415283203,
+ "step": 1736
+ },
+ {
+ "epoch": 24.125764192139737,
+ "grad_norm": 0.01716383546590805,
+ "learning_rate": 0.0006,
+ "loss": 4.318876266479492,
+ "step": 1737
+ },
+ {
+ "epoch": 24.139737991266376,
+ "grad_norm": 0.017049988731741905,
+ "learning_rate": 0.0006,
+ "loss": 4.237238883972168,
+ "step": 1738
+ },
+ {
+ "epoch": 24.153711790393015,
+ "grad_norm": 0.015169923193752766,
+ "learning_rate": 0.0006,
+ "loss": 4.289426803588867,
+ "step": 1739
+ },
+ {
+ "epoch": 24.16768558951965,
+ "grad_norm": 0.014760667458176613,
+ "learning_rate": 0.0006,
+ "loss": 4.356924057006836,
+ "step": 1740
+ },
+ {
+ "epoch": 24.18165938864629,
+ "grad_norm": 0.015127750113606453,
+ "learning_rate": 0.0006,
+ "loss": 4.142916679382324,
+ "step": 1741
+ },
+ {
+ "epoch": 24.195633187772927,
+ "grad_norm": 0.01549555640667677,
+ "learning_rate": 0.0006,
+ "loss": 4.345563888549805,
+ "step": 1742
+ },
+ {
+ "epoch": 24.209606986899562,
+ "grad_norm": 0.016393091529607773,
+ "learning_rate": 0.0006,
+ "loss": 4.417495250701904,
+ "step": 1743
+ },
+ {
+ "epoch": 24.2235807860262,
+ "grad_norm": 0.01785266026854515,
+ "learning_rate": 0.0006,
+ "loss": 4.3553547859191895,
+ "step": 1744
+ },
+ {
+ "epoch": 24.23755458515284,
+ "grad_norm": 0.017574617639183998,
+ "learning_rate": 0.0006,
+ "loss": 4.241024494171143,
+ "step": 1745
+ },
+ {
+ "epoch": 24.251528384279474,
+ "grad_norm": 0.019060306251049042,
+ "learning_rate": 0.0006,
+ "loss": 4.3685150146484375,
+ "step": 1746
+ },
+ {
+ "epoch": 24.265502183406113,
+ "grad_norm": 0.0214633010327816,
+ "learning_rate": 0.0006,
+ "loss": 4.323866844177246,
+ "step": 1747
+ },
+ {
+ "epoch": 24.27947598253275,
+ "grad_norm": 0.021055003628134727,
+ "learning_rate": 0.0006,
+ "loss": 4.278757095336914,
+ "step": 1748
+ },
+ {
+ "epoch": 24.29344978165939,
+ "grad_norm": 0.018618572503328323,
+ "learning_rate": 0.0006,
+ "loss": 4.28849983215332,
+ "step": 1749
+ },
+ {
+ "epoch": 24.307423580786025,
+ "grad_norm": 0.018666435033082962,
+ "learning_rate": 0.0006,
+ "loss": 4.368590354919434,
+ "step": 1750
+ },
+ {
+ "epoch": 24.321397379912664,
+ "grad_norm": 0.020626146346330643,
+ "learning_rate": 0.0006,
+ "loss": 4.2951836585998535,
+ "step": 1751
+ },
+ {
+ "epoch": 24.335371179039303,
+ "grad_norm": 0.01929759792983532,
+ "learning_rate": 0.0006,
+ "loss": 4.3664703369140625,
+ "step": 1752
+ },
+ {
+ "epoch": 24.349344978165938,
+ "grad_norm": 0.020956002175807953,
+ "learning_rate": 0.0006,
+ "loss": 4.345308303833008,
+ "step": 1753
+ },
+ {
+ "epoch": 24.363318777292577,
+ "grad_norm": 0.019227957352995872,
+ "learning_rate": 0.0006,
+ "loss": 4.349849224090576,
+ "step": 1754
+ },
+ {
+ "epoch": 24.377292576419215,
+ "grad_norm": 0.015970097854733467,
+ "learning_rate": 0.0006,
+ "loss": 4.223012447357178,
+ "step": 1755
+ },
+ {
+ "epoch": 24.39126637554585,
+ "grad_norm": 0.015951646491885185,
+ "learning_rate": 0.0006,
+ "loss": 4.236420631408691,
+ "step": 1756
+ },
+ {
+ "epoch": 24.40524017467249,
+ "grad_norm": 0.016032515093684196,
+ "learning_rate": 0.0006,
+ "loss": 4.226899147033691,
+ "step": 1757
+ },
+ {
+ "epoch": 24.419213973799128,
+ "grad_norm": 0.01573132909834385,
+ "learning_rate": 0.0006,
+ "loss": 4.315916061401367,
+ "step": 1758
+ },
+ {
+ "epoch": 24.433187772925763,
+ "grad_norm": 0.015367005951702595,
+ "learning_rate": 0.0006,
+ "loss": 4.358939170837402,
+ "step": 1759
+ },
+ {
+ "epoch": 24.4471615720524,
+ "grad_norm": 0.01669907011091709,
+ "learning_rate": 0.0006,
+ "loss": 4.232147216796875,
+ "step": 1760
+ },
+ {
+ "epoch": 24.46113537117904,
+ "grad_norm": 0.018708152696490288,
+ "learning_rate": 0.0006,
+ "loss": 4.329852104187012,
+ "step": 1761
+ },
+ {
+ "epoch": 24.475109170305675,
+ "grad_norm": 0.019035987555980682,
+ "learning_rate": 0.0006,
+ "loss": 4.260340213775635,
+ "step": 1762
+ },
+ {
+ "epoch": 24.489082969432314,
+ "grad_norm": 0.01832144521176815,
+ "learning_rate": 0.0006,
+ "loss": 4.359543800354004,
+ "step": 1763
+ },
+ {
+ "epoch": 24.503056768558952,
+ "grad_norm": 0.015313294716179371,
+ "learning_rate": 0.0006,
+ "loss": 4.279088020324707,
+ "step": 1764
+ },
+ {
+ "epoch": 24.51703056768559,
+ "grad_norm": 0.015803923830389977,
+ "learning_rate": 0.0006,
+ "loss": 4.230682849884033,
+ "step": 1765
+ },
+ {
+ "epoch": 24.531004366812226,
+ "grad_norm": 0.016748374328017235,
+ "learning_rate": 0.0006,
+ "loss": 4.349661827087402,
+ "step": 1766
+ },
+ {
+ "epoch": 24.544978165938865,
+ "grad_norm": 0.015732428058981895,
+ "learning_rate": 0.0006,
+ "loss": 4.372441291809082,
+ "step": 1767
+ },
+ {
+ "epoch": 24.558951965065503,
+ "grad_norm": 0.016564829275012016,
+ "learning_rate": 0.0006,
+ "loss": 4.323090553283691,
+ "step": 1768
+ },
+ {
+ "epoch": 24.57292576419214,
+ "grad_norm": 0.01603010483086109,
+ "learning_rate": 0.0006,
+ "loss": 4.347898006439209,
+ "step": 1769
+ },
+ {
+ "epoch": 24.586899563318777,
+ "grad_norm": 0.01674809865653515,
+ "learning_rate": 0.0006,
+ "loss": 4.269165515899658,
+ "step": 1770
+ },
+ {
+ "epoch": 24.600873362445416,
+ "grad_norm": 0.017333930358290672,
+ "learning_rate": 0.0006,
+ "loss": 4.353298187255859,
+ "step": 1771
+ },
+ {
+ "epoch": 24.61484716157205,
+ "grad_norm": 0.017519580200314522,
+ "learning_rate": 0.0006,
+ "loss": 4.265199661254883,
+ "step": 1772
+ },
+ {
+ "epoch": 24.62882096069869,
+ "grad_norm": 0.017675306648015976,
+ "learning_rate": 0.0006,
+ "loss": 4.322596549987793,
+ "step": 1773
+ },
+ {
+ "epoch": 24.64279475982533,
+ "grad_norm": 0.017166534438729286,
+ "learning_rate": 0.0006,
+ "loss": 4.413900375366211,
+ "step": 1774
+ },
+ {
+ "epoch": 24.656768558951963,
+ "grad_norm": 0.016362346708774567,
+ "learning_rate": 0.0006,
+ "loss": 4.307236671447754,
+ "step": 1775
+ },
+ {
+ "epoch": 24.670742358078602,
+ "grad_norm": 0.016086872667074203,
+ "learning_rate": 0.0006,
+ "loss": 4.235316276550293,
+ "step": 1776
+ },
+ {
+ "epoch": 24.68471615720524,
+ "grad_norm": 0.01689537614583969,
+ "learning_rate": 0.0006,
+ "loss": 4.390591621398926,
+ "step": 1777
+ },
+ {
+ "epoch": 24.69868995633188,
+ "grad_norm": 0.016836725175380707,
+ "learning_rate": 0.0006,
+ "loss": 4.2990031242370605,
+ "step": 1778
+ },
+ {
+ "epoch": 24.712663755458514,
+ "grad_norm": 0.015627335757017136,
+ "learning_rate": 0.0006,
+ "loss": 4.166745185852051,
+ "step": 1779
+ },
+ {
+ "epoch": 24.726637554585153,
+ "grad_norm": 0.015007016249001026,
+ "learning_rate": 0.0006,
+ "loss": 4.295273303985596,
+ "step": 1780
+ },
+ {
+ "epoch": 24.74061135371179,
+ "grad_norm": 0.016084209084510803,
+ "learning_rate": 0.0006,
+ "loss": 4.342424392700195,
+ "step": 1781
+ },
+ {
+ "epoch": 24.754585152838427,
+ "grad_norm": 0.01595073565840721,
+ "learning_rate": 0.0006,
+ "loss": 4.3060102462768555,
+ "step": 1782
+ },
+ {
+ "epoch": 24.768558951965066,
+ "grad_norm": 0.016113461926579475,
+ "learning_rate": 0.0006,
+ "loss": 4.267908096313477,
+ "step": 1783
+ },
+ {
+ "epoch": 24.782532751091704,
+ "grad_norm": 0.01564556173980236,
+ "learning_rate": 0.0006,
+ "loss": 4.324642658233643,
+ "step": 1784
+ },
+ {
+ "epoch": 24.79650655021834,
+ "grad_norm": 0.01575268618762493,
+ "learning_rate": 0.0006,
+ "loss": 4.328181743621826,
+ "step": 1785
+ },
+ {
+ "epoch": 24.810480349344978,
+ "grad_norm": 0.016830824315547943,
+ "learning_rate": 0.0006,
+ "loss": 4.247906684875488,
+ "step": 1786
+ },
+ {
+ "epoch": 24.824454148471617,
+ "grad_norm": 0.017989547923207283,
+ "learning_rate": 0.0006,
+ "loss": 4.228281021118164,
+ "step": 1787
+ },
+ {
+ "epoch": 24.83842794759825,
+ "grad_norm": 0.016783086583018303,
+ "learning_rate": 0.0006,
+ "loss": 4.2880964279174805,
+ "step": 1788
+ },
+ {
+ "epoch": 24.85240174672489,
+ "grad_norm": 0.016467561945319176,
+ "learning_rate": 0.0006,
+ "loss": 4.36118745803833,
+ "step": 1789
+ },
+ {
+ "epoch": 24.86637554585153,
+ "grad_norm": 0.01780523732304573,
+ "learning_rate": 0.0006,
+ "loss": 4.172321319580078,
+ "step": 1790
+ },
+ {
+ "epoch": 24.880349344978168,
+ "grad_norm": 0.017102031037211418,
+ "learning_rate": 0.0006,
+ "loss": 4.352826118469238,
+ "step": 1791
+ },
+ {
+ "epoch": 24.894323144104803,
+ "grad_norm": 0.016409218311309814,
+ "learning_rate": 0.0006,
+ "loss": 4.212162017822266,
+ "step": 1792
+ },
+ {
+ "epoch": 24.90829694323144,
+ "grad_norm": 0.018660584464669228,
+ "learning_rate": 0.0006,
+ "loss": 4.346684455871582,
+ "step": 1793
+ },
+ {
+ "epoch": 24.92227074235808,
+ "grad_norm": 0.017389440909028053,
+ "learning_rate": 0.0006,
+ "loss": 4.197913646697998,
+ "step": 1794
+ },
+ {
+ "epoch": 24.936244541484715,
+ "grad_norm": 0.01596485823392868,
+ "learning_rate": 0.0006,
+ "loss": 4.370617866516113,
+ "step": 1795
+ },
+ {
+ "epoch": 24.950218340611354,
+ "grad_norm": 0.01709776371717453,
+ "learning_rate": 0.0006,
+ "loss": 4.313847064971924,
+ "step": 1796
+ },
+ {
+ "epoch": 24.964192139737992,
+ "grad_norm": 0.016571134328842163,
+ "learning_rate": 0.0006,
+ "loss": 4.1895856857299805,
+ "step": 1797
+ },
+ {
+ "epoch": 24.978165938864628,
+ "grad_norm": 0.016851048916578293,
+ "learning_rate": 0.0006,
+ "loss": 4.350772857666016,
+ "step": 1798
+ },
+ {
+ "epoch": 24.992139737991266,
+ "grad_norm": 0.017007440328598022,
+ "learning_rate": 0.0006,
+ "loss": 4.399008750915527,
+ "step": 1799
+ },
+ {
+ "epoch": 25.0,
+ "grad_norm": 0.01791212148964405,
+ "learning_rate": 0.0006,
+ "loss": 4.306121826171875,
+ "step": 1800
+ },
+ {
+ "epoch": 25.0,
+ "eval_loss": 4.636417865753174,
+ "eval_runtime": 56.8296,
+ "eval_samples_per_second": 42.971,
+ "eval_steps_per_second": 1.355,
+ "step": 1800
+ },
+ {
+ "epoch": 25.01397379912664,
+ "grad_norm": 0.01729617640376091,
+ "learning_rate": 0.0006,
+ "loss": 4.257018089294434,
+ "step": 1801
+ },
+ {
+ "epoch": 25.027947598253274,
+ "grad_norm": 0.017466643825173378,
+ "learning_rate": 0.0006,
+ "loss": 4.307605266571045,
+ "step": 1802
+ },
+ {
+ "epoch": 25.041921397379912,
+ "grad_norm": 0.018614279106259346,
+ "learning_rate": 0.0006,
+ "loss": 4.360141277313232,
+ "step": 1803
+ },
+ {
+ "epoch": 25.05589519650655,
+ "grad_norm": 0.02010771445930004,
+ "learning_rate": 0.0006,
+ "loss": 4.196428298950195,
+ "step": 1804
+ },
+ {
+ "epoch": 25.069868995633186,
+ "grad_norm": 0.02090228535234928,
+ "learning_rate": 0.0006,
+ "loss": 4.288947105407715,
+ "step": 1805
+ },
+ {
+ "epoch": 25.083842794759825,
+ "grad_norm": 0.02050255797803402,
+ "learning_rate": 0.0006,
+ "loss": 4.259582042694092,
+ "step": 1806
+ },
+ {
+ "epoch": 25.097816593886463,
+ "grad_norm": 0.021844089031219482,
+ "learning_rate": 0.0006,
+ "loss": 4.405138969421387,
+ "step": 1807
+ },
+ {
+ "epoch": 25.111790393013102,
+ "grad_norm": 0.022287018597126007,
+ "learning_rate": 0.0006,
+ "loss": 4.333793640136719,
+ "step": 1808
+ },
+ {
+ "epoch": 25.125764192139737,
+ "grad_norm": 0.02288329415023327,
+ "learning_rate": 0.0006,
+ "loss": 4.3910980224609375,
+ "step": 1809
+ },
+ {
+ "epoch": 25.139737991266376,
+ "grad_norm": 0.02373199723660946,
+ "learning_rate": 0.0006,
+ "loss": 4.32921028137207,
+ "step": 1810
+ },
+ {
+ "epoch": 25.153711790393015,
+ "grad_norm": 0.024960234761238098,
+ "learning_rate": 0.0006,
+ "loss": 4.295609474182129,
+ "step": 1811
+ },
+ {
+ "epoch": 25.16768558951965,
+ "grad_norm": 0.026743892580270767,
+ "learning_rate": 0.0006,
+ "loss": 4.315308094024658,
+ "step": 1812
+ },
+ {
+ "epoch": 25.18165938864629,
+ "grad_norm": 0.022138185799121857,
+ "learning_rate": 0.0006,
+ "loss": 4.360208511352539,
+ "step": 1813
+ },
+ {
+ "epoch": 25.195633187772927,
+ "grad_norm": 0.022120574489235878,
+ "learning_rate": 0.0006,
+ "loss": 4.1973772048950195,
+ "step": 1814
+ },
+ {
+ "epoch": 25.209606986899562,
+ "grad_norm": 0.02230502851307392,
+ "learning_rate": 0.0006,
+ "loss": 4.285558700561523,
+ "step": 1815
+ },
+ {
+ "epoch": 25.2235807860262,
+ "grad_norm": 0.022301986813545227,
+ "learning_rate": 0.0006,
+ "loss": 4.238343238830566,
+ "step": 1816
+ },
+ {
+ "epoch": 25.23755458515284,
+ "grad_norm": 0.020050447434186935,
+ "learning_rate": 0.0006,
+ "loss": 4.298235893249512,
+ "step": 1817
+ },
+ {
+ "epoch": 25.251528384279474,
+ "grad_norm": 0.01987724006175995,
+ "learning_rate": 0.0006,
+ "loss": 4.301384449005127,
+ "step": 1818
+ },
+ {
+ "epoch": 25.265502183406113,
+ "grad_norm": 0.020440855994820595,
+ "learning_rate": 0.0006,
+ "loss": 4.269218444824219,
+ "step": 1819
+ },
+ {
+ "epoch": 25.27947598253275,
+ "grad_norm": 0.01891893707215786,
+ "learning_rate": 0.0006,
+ "loss": 4.2401957511901855,
+ "step": 1820
+ },
+ {
+ "epoch": 25.29344978165939,
+ "grad_norm": 0.01903688907623291,
+ "learning_rate": 0.0006,
+ "loss": 4.139670372009277,
+ "step": 1821
+ },
+ {
+ "epoch": 25.307423580786025,
+ "grad_norm": 0.018856938928365707,
+ "learning_rate": 0.0006,
+ "loss": 4.233214378356934,
+ "step": 1822
+ },
+ {
+ "epoch": 25.321397379912664,
+ "grad_norm": 0.017791323363780975,
+ "learning_rate": 0.0006,
+ "loss": 4.2342023849487305,
+ "step": 1823
+ },
+ {
+ "epoch": 25.335371179039303,
+ "grad_norm": 0.017304565757513046,
+ "learning_rate": 0.0006,
+ "loss": 4.355198860168457,
+ "step": 1824
+ },
+ {
+ "epoch": 25.349344978165938,
+ "grad_norm": 0.01576206088066101,
+ "learning_rate": 0.0006,
+ "loss": 4.114147186279297,
+ "step": 1825
+ },
+ {
+ "epoch": 25.363318777292577,
+ "grad_norm": 0.015105154365301132,
+ "learning_rate": 0.0006,
+ "loss": 4.313795566558838,
+ "step": 1826
+ },
+ {
+ "epoch": 25.377292576419215,
+ "grad_norm": 0.01563587784767151,
+ "learning_rate": 0.0006,
+ "loss": 4.308282852172852,
+ "step": 1827
+ },
+ {
+ "epoch": 25.39126637554585,
+ "grad_norm": 0.015980906784534454,
+ "learning_rate": 0.0006,
+ "loss": 4.345149040222168,
+ "step": 1828
+ },
+ {
+ "epoch": 25.40524017467249,
+ "grad_norm": 0.01669284701347351,
+ "learning_rate": 0.0006,
+ "loss": 4.237611293792725,
+ "step": 1829
+ },
+ {
+ "epoch": 25.419213973799128,
+ "grad_norm": 0.016202064231038094,
+ "learning_rate": 0.0006,
+ "loss": 4.3072309494018555,
+ "step": 1830
+ },
+ {
+ "epoch": 25.433187772925763,
+ "grad_norm": 0.01478662807494402,
+ "learning_rate": 0.0006,
+ "loss": 4.276215076446533,
+ "step": 1831
+ },
+ {
+ "epoch": 25.4471615720524,
+ "grad_norm": 0.015168731100857258,
+ "learning_rate": 0.0006,
+ "loss": 4.37509822845459,
+ "step": 1832
+ },
+ {
+ "epoch": 25.46113537117904,
+ "grad_norm": 0.01648980937898159,
+ "learning_rate": 0.0006,
+ "loss": 4.349167346954346,
+ "step": 1833
+ },
+ {
+ "epoch": 25.475109170305675,
+ "grad_norm": 0.01708398200571537,
+ "learning_rate": 0.0006,
+ "loss": 4.346580505371094,
+ "step": 1834
+ },
+ {
+ "epoch": 25.489082969432314,
+ "grad_norm": 0.017370784655213356,
+ "learning_rate": 0.0006,
+ "loss": 4.299466133117676,
+ "step": 1835
+ },
+ {
+ "epoch": 25.503056768558952,
+ "grad_norm": 0.016248228028416634,
+ "learning_rate": 0.0006,
+ "loss": 4.184610366821289,
+ "step": 1836
+ },
+ {
+ "epoch": 25.51703056768559,
+ "grad_norm": 0.017106913030147552,
+ "learning_rate": 0.0006,
+ "loss": 4.366856098175049,
+ "step": 1837
+ },
+ {
+ "epoch": 25.531004366812226,
+ "grad_norm": 0.016071965917944908,
+ "learning_rate": 0.0006,
+ "loss": 4.297691345214844,
+ "step": 1838
+ },
+ {
+ "epoch": 25.544978165938865,
+ "grad_norm": 0.016451986506581306,
+ "learning_rate": 0.0006,
+ "loss": 4.229983329772949,
+ "step": 1839
+ },
+ {
+ "epoch": 25.558951965065503,
+ "grad_norm": 0.016195081174373627,
+ "learning_rate": 0.0006,
+ "loss": 4.259893417358398,
+ "step": 1840
+ },
+ {
+ "epoch": 25.57292576419214,
+ "grad_norm": 0.0165175162255764,
+ "learning_rate": 0.0006,
+ "loss": 4.3348236083984375,
+ "step": 1841
+ },
+ {
+ "epoch": 25.586899563318777,
+ "grad_norm": 0.01651517115533352,
+ "learning_rate": 0.0006,
+ "loss": 4.319541931152344,
+ "step": 1842
+ },
+ {
+ "epoch": 25.600873362445416,
+ "grad_norm": 0.016143690794706345,
+ "learning_rate": 0.0006,
+ "loss": 4.345652103424072,
+ "step": 1843
+ },
+ {
+ "epoch": 25.61484716157205,
+ "grad_norm": 0.015140696428716183,
+ "learning_rate": 0.0006,
+ "loss": 4.212047576904297,
+ "step": 1844
+ },
+ {
+ "epoch": 25.62882096069869,
+ "grad_norm": 0.014637443237006664,
+ "learning_rate": 0.0006,
+ "loss": 4.187453269958496,
+ "step": 1845
+ },
+ {
+ "epoch": 25.64279475982533,
+ "grad_norm": 0.014585614204406738,
+ "learning_rate": 0.0006,
+ "loss": 4.253936767578125,
+ "step": 1846
+ },
+ {
+ "epoch": 25.656768558951963,
+ "grad_norm": 0.015158289112150669,
+ "learning_rate": 0.0006,
+ "loss": 4.34612512588501,
+ "step": 1847
+ },
+ {
+ "epoch": 25.670742358078602,
+ "grad_norm": 0.015202849172055721,
+ "learning_rate": 0.0006,
+ "loss": 4.270634651184082,
+ "step": 1848
+ },
+ {
+ "epoch": 25.68471615720524,
+ "grad_norm": 0.015363575890660286,
+ "learning_rate": 0.0006,
+ "loss": 4.271080493927002,
+ "step": 1849
+ },
+ {
+ "epoch": 25.69868995633188,
+ "grad_norm": 0.014360226690769196,
+ "learning_rate": 0.0006,
+ "loss": 4.375033855438232,
+ "step": 1850
+ },
+ {
+ "epoch": 25.712663755458514,
+ "grad_norm": 0.014263181947171688,
+ "learning_rate": 0.0006,
+ "loss": 4.424138069152832,
+ "step": 1851
+ },
+ {
+ "epoch": 25.726637554585153,
+ "grad_norm": 0.014398688450455666,
+ "learning_rate": 0.0006,
+ "loss": 4.410154819488525,
+ "step": 1852
+ },
+ {
+ "epoch": 25.74061135371179,
+ "grad_norm": 0.016131488606333733,
+ "learning_rate": 0.0006,
+ "loss": 4.324014663696289,
+ "step": 1853
+ },
+ {
+ "epoch": 25.754585152838427,
+ "grad_norm": 0.017568735405802727,
+ "learning_rate": 0.0006,
+ "loss": 4.367861747741699,
+ "step": 1854
+ },
+ {
+ "epoch": 25.768558951965066,
+ "grad_norm": 0.017907511442899704,
+ "learning_rate": 0.0006,
+ "loss": 4.352965354919434,
+ "step": 1855
+ },
+ {
+ "epoch": 25.782532751091704,
+ "grad_norm": 0.016918016597628593,
+ "learning_rate": 0.0006,
+ "loss": 4.252397537231445,
+ "step": 1856
+ },
+ {
+ "epoch": 25.79650655021834,
+ "grad_norm": 0.017274610698223114,
+ "learning_rate": 0.0006,
+ "loss": 4.342309951782227,
+ "step": 1857
+ },
+ {
+ "epoch": 25.810480349344978,
+ "grad_norm": 0.017879825085401535,
+ "learning_rate": 0.0006,
+ "loss": 4.330987453460693,
+ "step": 1858
+ },
+ {
+ "epoch": 25.824454148471617,
+ "grad_norm": 0.01736099272966385,
+ "learning_rate": 0.0006,
+ "loss": 4.307255744934082,
+ "step": 1859
+ },
+ {
+ "epoch": 25.83842794759825,
+ "grad_norm": 0.016575824469327927,
+ "learning_rate": 0.0006,
+ "loss": 4.262016773223877,
+ "step": 1860
+ },
+ {
+ "epoch": 25.85240174672489,
+ "grad_norm": 0.015181140042841434,
+ "learning_rate": 0.0006,
+ "loss": 4.291863441467285,
+ "step": 1861
+ },
+ {
+ "epoch": 25.86637554585153,
+ "grad_norm": 0.015965687111020088,
+ "learning_rate": 0.0006,
+ "loss": 4.424836158752441,
+ "step": 1862
+ },
+ {
+ "epoch": 25.880349344978168,
+ "grad_norm": 0.01687219925224781,
+ "learning_rate": 0.0006,
+ "loss": 4.343531131744385,
+ "step": 1863
+ },
+ {
+ "epoch": 25.894323144104803,
+ "grad_norm": 0.015115504153072834,
+ "learning_rate": 0.0006,
+ "loss": 4.357808589935303,
+ "step": 1864
+ },
+ {
+ "epoch": 25.90829694323144,
+ "grad_norm": 0.014150998555123806,
+ "learning_rate": 0.0006,
+ "loss": 4.324550151824951,
+ "step": 1865
+ },
+ {
+ "epoch": 25.92227074235808,
+ "grad_norm": 0.014989510178565979,
+ "learning_rate": 0.0006,
+ "loss": 4.380428314208984,
+ "step": 1866
+ },
+ {
+ "epoch": 25.936244541484715,
+ "grad_norm": 0.01527960691601038,
+ "learning_rate": 0.0006,
+ "loss": 4.399374961853027,
+ "step": 1867
+ },
+ {
+ "epoch": 25.950218340611354,
+ "grad_norm": 0.015434633940458298,
+ "learning_rate": 0.0006,
+ "loss": 4.260984420776367,
+ "step": 1868
+ },
+ {
+ "epoch": 25.964192139737992,
+ "grad_norm": 0.015370228327810764,
+ "learning_rate": 0.0006,
+ "loss": 4.3340253829956055,
+ "step": 1869
+ },
+ {
+ "epoch": 25.978165938864628,
+ "grad_norm": 0.015348542481660843,
+ "learning_rate": 0.0006,
+ "loss": 4.369132995605469,
+ "step": 1870
+ },
+ {
+ "epoch": 25.992139737991266,
+ "grad_norm": 0.015354936942458153,
+ "learning_rate": 0.0006,
+ "loss": 4.381937026977539,
+ "step": 1871
+ },
+ {
+ "epoch": 26.0,
+ "grad_norm": 0.01796272210776806,
+ "learning_rate": 0.0006,
+ "loss": 4.286768436431885,
+ "step": 1872
+ },
+ {
+ "epoch": 26.0,
+ "eval_loss": 4.576984405517578,
+ "eval_runtime": 57.2091,
+ "eval_samples_per_second": 42.685,
+ "eval_steps_per_second": 1.346,
+ "step": 1872
+ },
+ {
+ "epoch": 26.01397379912664,
+ "grad_norm": 0.01748845726251602,
+ "learning_rate": 0.0006,
+ "loss": 4.337504863739014,
+ "step": 1873
+ },
+ {
+ "epoch": 26.027947598253274,
+ "grad_norm": 0.019244296476244926,
+ "learning_rate": 0.0006,
+ "loss": 4.264280796051025,
+ "step": 1874
+ },
+ {
+ "epoch": 26.041921397379912,
+ "grad_norm": 0.0215692650526762,
+ "learning_rate": 0.0006,
+ "loss": 4.350830078125,
+ "step": 1875
+ },
+ {
+ "epoch": 26.05589519650655,
+ "grad_norm": 0.022189252078533173,
+ "learning_rate": 0.0006,
+ "loss": 4.299098968505859,
+ "step": 1876
+ },
+ {
+ "epoch": 26.069868995633186,
+ "grad_norm": 0.022384043782949448,
+ "learning_rate": 0.0006,
+ "loss": 4.1185173988342285,
+ "step": 1877
+ },
+ {
+ "epoch": 26.083842794759825,
+ "grad_norm": 0.02076014317572117,
+ "learning_rate": 0.0006,
+ "loss": 4.297612190246582,
+ "step": 1878
+ },
+ {
+ "epoch": 26.097816593886463,
+ "grad_norm": 0.020843051373958588,
+ "learning_rate": 0.0006,
+ "loss": 4.294098854064941,
+ "step": 1879
+ },
+ {
+ "epoch": 26.111790393013102,
+ "grad_norm": 0.020654456689953804,
+ "learning_rate": 0.0006,
+ "loss": 4.267856597900391,
+ "step": 1880
+ },
+ {
+ "epoch": 26.125764192139737,
+ "grad_norm": 0.01860472559928894,
+ "learning_rate": 0.0006,
+ "loss": 4.284232139587402,
+ "step": 1881
+ },
+ {
+ "epoch": 26.139737991266376,
+ "grad_norm": 0.018719421699643135,
+ "learning_rate": 0.0006,
+ "loss": 4.277214527130127,
+ "step": 1882
+ },
+ {
+ "epoch": 26.153711790393015,
+ "grad_norm": 0.020903829485177994,
+ "learning_rate": 0.0006,
+ "loss": 4.254947662353516,
+ "step": 1883
+ },
+ {
+ "epoch": 26.16768558951965,
+ "grad_norm": 0.02020149491727352,
+ "learning_rate": 0.0006,
+ "loss": 4.2526936531066895,
+ "step": 1884
+ },
+ {
+ "epoch": 26.18165938864629,
+ "grad_norm": 0.0180392786860466,
+ "learning_rate": 0.0006,
+ "loss": 4.309922695159912,
+ "step": 1885
+ },
+ {
+ "epoch": 26.195633187772927,
+ "grad_norm": 0.018163125962018967,
+ "learning_rate": 0.0006,
+ "loss": 4.13860559463501,
+ "step": 1886
+ },
+ {
+ "epoch": 26.209606986899562,
+ "grad_norm": 0.01853291131556034,
+ "learning_rate": 0.0006,
+ "loss": 4.190926551818848,
+ "step": 1887
+ },
+ {
+ "epoch": 26.2235807860262,
+ "grad_norm": 0.01757677271962166,
+ "learning_rate": 0.0006,
+ "loss": 4.163999557495117,
+ "step": 1888
+ },
+ {
+ "epoch": 26.23755458515284,
+ "grad_norm": 0.01767992228269577,
+ "learning_rate": 0.0006,
+ "loss": 4.184430122375488,
+ "step": 1889
+ },
+ {
+ "epoch": 26.251528384279474,
+ "grad_norm": 0.018139973282814026,
+ "learning_rate": 0.0006,
+ "loss": 4.300766944885254,
+ "step": 1890
+ },
+ {
+ "epoch": 26.265502183406113,
+ "grad_norm": 0.020184461027383804,
+ "learning_rate": 0.0006,
+ "loss": 4.232537269592285,
+ "step": 1891
+ },
+ {
+ "epoch": 26.27947598253275,
+ "grad_norm": 0.02028856985270977,
+ "learning_rate": 0.0006,
+ "loss": 4.324883460998535,
+ "step": 1892
+ },
+ {
+ "epoch": 26.29344978165939,
+ "grad_norm": 0.017120616510510445,
+ "learning_rate": 0.0006,
+ "loss": 4.23493766784668,
+ "step": 1893
+ },
+ {
+ "epoch": 26.307423580786025,
+ "grad_norm": 0.016202978789806366,
+ "learning_rate": 0.0006,
+ "loss": 4.197602272033691,
+ "step": 1894
+ },
+ {
+ "epoch": 26.321397379912664,
+ "grad_norm": 0.017231421545147896,
+ "learning_rate": 0.0006,
+ "loss": 4.302001953125,
+ "step": 1895
+ },
+ {
+ "epoch": 26.335371179039303,
+ "grad_norm": 0.01725655607879162,
+ "learning_rate": 0.0006,
+ "loss": 4.356800079345703,
+ "step": 1896
+ },
+ {
+ "epoch": 26.349344978165938,
+ "grad_norm": 0.016657443717122078,
+ "learning_rate": 0.0006,
+ "loss": 4.266307830810547,
+ "step": 1897
+ },
+ {
+ "epoch": 26.363318777292577,
+ "grad_norm": 0.01616556942462921,
+ "learning_rate": 0.0006,
+ "loss": 4.307868003845215,
+ "step": 1898
+ },
+ {
+ "epoch": 26.377292576419215,
+ "grad_norm": 0.016621306538581848,
+ "learning_rate": 0.0006,
+ "loss": 4.172905921936035,
+ "step": 1899
+ },
+ {
+ "epoch": 26.39126637554585,
+ "grad_norm": 0.016243206337094307,
+ "learning_rate": 0.0006,
+ "loss": 4.162896156311035,
+ "step": 1900
+ },
+ {
+ "epoch": 26.40524017467249,
+ "grad_norm": 0.016663808375597,
+ "learning_rate": 0.0006,
+ "loss": 4.258232116699219,
+ "step": 1901
+ },
+ {
+ "epoch": 26.419213973799128,
+ "grad_norm": 0.01677478663623333,
+ "learning_rate": 0.0006,
+ "loss": 4.336090087890625,
+ "step": 1902
+ },
+ {
+ "epoch": 26.433187772925763,
+ "grad_norm": 0.01641303487122059,
+ "learning_rate": 0.0006,
+ "loss": 4.191476821899414,
+ "step": 1903
+ },
+ {
+ "epoch": 26.4471615720524,
+ "grad_norm": 0.016700323671102524,
+ "learning_rate": 0.0006,
+ "loss": 4.299427509307861,
+ "step": 1904
+ },
+ {
+ "epoch": 26.46113537117904,
+ "grad_norm": 0.017595035955309868,
+ "learning_rate": 0.0006,
+ "loss": 4.312828063964844,
+ "step": 1905
+ },
+ {
+ "epoch": 26.475109170305675,
+ "grad_norm": 0.020300284028053284,
+ "learning_rate": 0.0006,
+ "loss": 4.28331184387207,
+ "step": 1906
+ },
+ {
+ "epoch": 26.489082969432314,
+ "grad_norm": 0.02198321744799614,
+ "learning_rate": 0.0006,
+ "loss": 4.3280839920043945,
+ "step": 1907
+ },
+ {
+ "epoch": 26.503056768558952,
+ "grad_norm": 0.02003200724720955,
+ "learning_rate": 0.0006,
+ "loss": 4.325623989105225,
+ "step": 1908
+ },
+ {
+ "epoch": 26.51703056768559,
+ "grad_norm": 0.01870287023484707,
+ "learning_rate": 0.0006,
+ "loss": 4.251203536987305,
+ "step": 1909
+ },
+ {
+ "epoch": 26.531004366812226,
+ "grad_norm": 0.01835448481142521,
+ "learning_rate": 0.0006,
+ "loss": 4.335144996643066,
+ "step": 1910
+ },
+ {
+ "epoch": 26.544978165938865,
+ "grad_norm": 0.02046327292919159,
+ "learning_rate": 0.0006,
+ "loss": 4.238023281097412,
+ "step": 1911
+ },
+ {
+ "epoch": 26.558951965065503,
+ "grad_norm": 0.019771311432123184,
+ "learning_rate": 0.0006,
+ "loss": 4.353594779968262,
+ "step": 1912
+ },
+ {
+ "epoch": 26.57292576419214,
+ "grad_norm": 0.018090544268488884,
+ "learning_rate": 0.0006,
+ "loss": 4.304343223571777,
+ "step": 1913
+ },
+ {
+ "epoch": 26.586899563318777,
+ "grad_norm": 0.017489777877926826,
+ "learning_rate": 0.0006,
+ "loss": 4.204375267028809,
+ "step": 1914
+ },
+ {
+ "epoch": 26.600873362445416,
+ "grad_norm": 0.016785891726613045,
+ "learning_rate": 0.0006,
+ "loss": 4.286669731140137,
+ "step": 1915
+ },
+ {
+ "epoch": 26.61484716157205,
+ "grad_norm": 0.016382789239287376,
+ "learning_rate": 0.0006,
+ "loss": 4.347662448883057,
+ "step": 1916
+ },
+ {
+ "epoch": 26.62882096069869,
+ "grad_norm": 0.016581876203417778,
+ "learning_rate": 0.0006,
+ "loss": 4.2716193199157715,
+ "step": 1917
+ },
+ {
+ "epoch": 26.64279475982533,
+ "grad_norm": 0.015627184882760048,
+ "learning_rate": 0.0006,
+ "loss": 4.311489105224609,
+ "step": 1918
+ },
+ {
+ "epoch": 26.656768558951963,
+ "grad_norm": 0.015908483415842056,
+ "learning_rate": 0.0006,
+ "loss": 4.262099742889404,
+ "step": 1919
+ },
+ {
+ "epoch": 26.670742358078602,
+ "grad_norm": 0.015576236881315708,
+ "learning_rate": 0.0006,
+ "loss": 4.275871276855469,
+ "step": 1920
+ },
+ {
+ "epoch": 26.68471615720524,
+ "grad_norm": 0.014621064998209476,
+ "learning_rate": 0.0006,
+ "loss": 4.184260845184326,
+ "step": 1921
+ },
+ {
+ "epoch": 26.69868995633188,
+ "grad_norm": 0.014797811396420002,
+ "learning_rate": 0.0006,
+ "loss": 4.3412628173828125,
+ "step": 1922
+ },
+ {
+ "epoch": 26.712663755458514,
+ "grad_norm": 0.015610570088028908,
+ "learning_rate": 0.0006,
+ "loss": 4.181094646453857,
+ "step": 1923
+ },
+ {
+ "epoch": 26.726637554585153,
+ "grad_norm": 0.016048265621066093,
+ "learning_rate": 0.0006,
+ "loss": 4.365290641784668,
+ "step": 1924
+ },
+ {
+ "epoch": 26.74061135371179,
+ "grad_norm": 0.016604071483016014,
+ "learning_rate": 0.0006,
+ "loss": 4.2628912925720215,
+ "step": 1925
+ },
+ {
+ "epoch": 26.754585152838427,
+ "grad_norm": 0.017044221982359886,
+ "learning_rate": 0.0006,
+ "loss": 4.378512382507324,
+ "step": 1926
+ },
+ {
+ "epoch": 26.768558951965066,
+ "grad_norm": 0.01746384985744953,
+ "learning_rate": 0.0006,
+ "loss": 4.3888773918151855,
+ "step": 1927
+ },
+ {
+ "epoch": 26.782532751091704,
+ "grad_norm": 0.01572711206972599,
+ "learning_rate": 0.0006,
+ "loss": 4.189935684204102,
+ "step": 1928
+ },
+ {
+ "epoch": 26.79650655021834,
+ "grad_norm": 0.014976629987359047,
+ "learning_rate": 0.0006,
+ "loss": 4.280525207519531,
+ "step": 1929
+ },
+ {
+ "epoch": 26.810480349344978,
+ "grad_norm": 0.014804664999246597,
+ "learning_rate": 0.0006,
+ "loss": 4.270630836486816,
+ "step": 1930
+ },
+ {
+ "epoch": 26.824454148471617,
+ "grad_norm": 0.014641920104622841,
+ "learning_rate": 0.0006,
+ "loss": 4.319505214691162,
+ "step": 1931
+ },
+ {
+ "epoch": 26.83842794759825,
+ "grad_norm": 0.014928505755960941,
+ "learning_rate": 0.0006,
+ "loss": 4.180147647857666,
+ "step": 1932
+ },
+ {
+ "epoch": 26.85240174672489,
+ "grad_norm": 0.016130365431308746,
+ "learning_rate": 0.0006,
+ "loss": 4.260660648345947,
+ "step": 1933
+ },
+ {
+ "epoch": 26.86637554585153,
+ "grad_norm": 0.015642493963241577,
+ "learning_rate": 0.0006,
+ "loss": 4.304064750671387,
+ "step": 1934
+ },
+ {
+ "epoch": 26.880349344978168,
+ "grad_norm": 0.015988217666745186,
+ "learning_rate": 0.0006,
+ "loss": 4.2548017501831055,
+ "step": 1935
+ },
+ {
+ "epoch": 26.894323144104803,
+ "grad_norm": 0.016501398757100105,
+ "learning_rate": 0.0006,
+ "loss": 4.260058879852295,
+ "step": 1936
+ },
+ {
+ "epoch": 26.90829694323144,
+ "grad_norm": 0.01645040698349476,
+ "learning_rate": 0.0006,
+ "loss": 4.483081340789795,
+ "step": 1937
+ },
+ {
+ "epoch": 26.92227074235808,
+ "grad_norm": 0.016140863299369812,
+ "learning_rate": 0.0006,
+ "loss": 4.1776838302612305,
+ "step": 1938
+ },
+ {
+ "epoch": 26.936244541484715,
+ "grad_norm": 0.016681019216775894,
+ "learning_rate": 0.0006,
+ "loss": 4.224747657775879,
+ "step": 1939
+ },
+ {
+ "epoch": 26.950218340611354,
+ "grad_norm": 0.01650378853082657,
+ "learning_rate": 0.0006,
+ "loss": 4.305703163146973,
+ "step": 1940
+ },
+ {
+ "epoch": 26.964192139737992,
+ "grad_norm": 0.017051683738827705,
+ "learning_rate": 0.0006,
+ "loss": 4.285835266113281,
+ "step": 1941
+ },
+ {
+ "epoch": 26.978165938864628,
+ "grad_norm": 0.016894327476620674,
+ "learning_rate": 0.0006,
+ "loss": 4.318901538848877,
+ "step": 1942
+ },
+ {
+ "epoch": 26.992139737991266,
+ "grad_norm": 0.017313100397586823,
+ "learning_rate": 0.0006,
+ "loss": 4.281434535980225,
+ "step": 1943
+ },
+ {
+ "epoch": 27.0,
+ "grad_norm": 0.018450884148478508,
+ "learning_rate": 0.0006,
+ "loss": 4.335104942321777,
+ "step": 1944
+ },
+ {
+ "epoch": 27.0,
+ "eval_loss": 4.640687465667725,
+ "eval_runtime": 56.9624,
+ "eval_samples_per_second": 42.87,
+ "eval_steps_per_second": 1.352,
+ "step": 1944
+ },
+ {
+ "epoch": 27.01397379912664,
+ "grad_norm": 0.017755698412656784,
+ "learning_rate": 0.0006,
+ "loss": 4.278824329376221,
+ "step": 1945
+ },
+ {
+ "epoch": 27.027947598253274,
+ "grad_norm": 0.01724310778081417,
+ "learning_rate": 0.0006,
+ "loss": 4.311049461364746,
+ "step": 1946
+ },
+ {
+ "epoch": 27.041921397379912,
+ "grad_norm": 0.01690123789012432,
+ "learning_rate": 0.0006,
+ "loss": 4.261994361877441,
+ "step": 1947
+ },
+ {
+ "epoch": 27.05589519650655,
+ "grad_norm": 0.018406696617603302,
+ "learning_rate": 0.0006,
+ "loss": 4.1925153732299805,
+ "step": 1948
+ },
+ {
+ "epoch": 27.069868995633186,
+ "grad_norm": 0.019945867359638214,
+ "learning_rate": 0.0006,
+ "loss": 4.108537197113037,
+ "step": 1949
+ },
+ {
+ "epoch": 27.083842794759825,
+ "grad_norm": 0.020316628739237785,
+ "learning_rate": 0.0006,
+ "loss": 4.209043025970459,
+ "step": 1950
+ },
+ {
+ "epoch": 27.097816593886463,
+ "grad_norm": 0.019690370187163353,
+ "learning_rate": 0.0006,
+ "loss": 4.195789337158203,
+ "step": 1951
+ },
+ {
+ "epoch": 27.111790393013102,
+ "grad_norm": 0.01785232499241829,
+ "learning_rate": 0.0006,
+ "loss": 4.21466064453125,
+ "step": 1952
+ },
+ {
+ "epoch": 27.125764192139737,
+ "grad_norm": 0.0164767038077116,
+ "learning_rate": 0.0006,
+ "loss": 4.270247459411621,
+ "step": 1953
+ },
+ {
+ "epoch": 27.139737991266376,
+ "grad_norm": 0.018008455634117126,
+ "learning_rate": 0.0006,
+ "loss": 4.201042175292969,
+ "step": 1954
+ },
+ {
+ "epoch": 27.153711790393015,
+ "grad_norm": 0.01807340793311596,
+ "learning_rate": 0.0006,
+ "loss": 4.25289249420166,
+ "step": 1955
+ },
+ {
+ "epoch": 27.16768558951965,
+ "grad_norm": 0.016246909275650978,
+ "learning_rate": 0.0006,
+ "loss": 4.116283416748047,
+ "step": 1956
+ },
+ {
+ "epoch": 27.18165938864629,
+ "grad_norm": 0.017545776441693306,
+ "learning_rate": 0.0006,
+ "loss": 4.316999435424805,
+ "step": 1957
+ },
+ {
+ "epoch": 27.195633187772927,
+ "grad_norm": 0.017915375530719757,
+ "learning_rate": 0.0006,
+ "loss": 4.253859519958496,
+ "step": 1958
+ },
+ {
+ "epoch": 27.209606986899562,
+ "grad_norm": 0.018112223595380783,
+ "learning_rate": 0.0006,
+ "loss": 4.2541022300720215,
+ "step": 1959
+ },
+ {
+ "epoch": 27.2235807860262,
+ "grad_norm": 0.01846139505505562,
+ "learning_rate": 0.0006,
+ "loss": 4.20438289642334,
+ "step": 1960
+ },
+ {
+ "epoch": 27.23755458515284,
+ "grad_norm": 0.01943897269666195,
+ "learning_rate": 0.0006,
+ "loss": 4.2799577713012695,
+ "step": 1961
+ },
+ {
+ "epoch": 27.251528384279474,
+ "grad_norm": 0.018253348767757416,
+ "learning_rate": 0.0006,
+ "loss": 4.1720499992370605,
+ "step": 1962
+ },
+ {
+ "epoch": 27.265502183406113,
+ "grad_norm": 0.018292246386408806,
+ "learning_rate": 0.0006,
+ "loss": 4.28231143951416,
+ "step": 1963
+ },
+ {
+ "epoch": 27.27947598253275,
+ "grad_norm": 0.01943155936896801,
+ "learning_rate": 0.0006,
+ "loss": 4.314976692199707,
+ "step": 1964
+ },
+ {
+ "epoch": 27.29344978165939,
+ "grad_norm": 0.019064994528889656,
+ "learning_rate": 0.0006,
+ "loss": 4.25328254699707,
+ "step": 1965
+ },
+ {
+ "epoch": 27.307423580786025,
+ "grad_norm": 0.017423273995518684,
+ "learning_rate": 0.0006,
+ "loss": 4.195992469787598,
+ "step": 1966
+ },
+ {
+ "epoch": 27.321397379912664,
+ "grad_norm": 0.01677924580872059,
+ "learning_rate": 0.0006,
+ "loss": 4.260043144226074,
+ "step": 1967
+ },
+ {
+ "epoch": 27.335371179039303,
+ "grad_norm": 0.01720673404633999,
+ "learning_rate": 0.0006,
+ "loss": 4.222780704498291,
+ "step": 1968
+ },
+ {
+ "epoch": 27.349344978165938,
+ "grad_norm": 0.017300743609666824,
+ "learning_rate": 0.0006,
+ "loss": 4.109524250030518,
+ "step": 1969
+ },
+ {
+ "epoch": 27.363318777292577,
+ "grad_norm": 0.016065871343016624,
+ "learning_rate": 0.0006,
+ "loss": 4.36108922958374,
+ "step": 1970
+ },
+ {
+ "epoch": 27.377292576419215,
+ "grad_norm": 0.016626616939902306,
+ "learning_rate": 0.0006,
+ "loss": 4.220630645751953,
+ "step": 1971
+ },
+ {
+ "epoch": 27.39126637554585,
+ "grad_norm": 0.016482602804899216,
+ "learning_rate": 0.0006,
+ "loss": 4.344797134399414,
+ "step": 1972
+ },
+ {
+ "epoch": 27.40524017467249,
+ "grad_norm": 0.01727953553199768,
+ "learning_rate": 0.0006,
+ "loss": 4.296645164489746,
+ "step": 1973
+ },
+ {
+ "epoch": 27.419213973799128,
+ "grad_norm": 0.01522731315344572,
+ "learning_rate": 0.0006,
+ "loss": 4.223751068115234,
+ "step": 1974
+ },
+ {
+ "epoch": 27.433187772925763,
+ "grad_norm": 0.01581776700913906,
+ "learning_rate": 0.0006,
+ "loss": 4.351379871368408,
+ "step": 1975
+ },
+ {
+ "epoch": 27.4471615720524,
+ "grad_norm": 0.016394605860114098,
+ "learning_rate": 0.0006,
+ "loss": 4.2412919998168945,
+ "step": 1976
+ },
+ {
+ "epoch": 27.46113537117904,
+ "grad_norm": 0.01751169003546238,
+ "learning_rate": 0.0006,
+ "loss": 4.257606506347656,
+ "step": 1977
+ },
+ {
+ "epoch": 27.475109170305675,
+ "grad_norm": 0.017452310770750046,
+ "learning_rate": 0.0006,
+ "loss": 4.380134582519531,
+ "step": 1978
+ },
+ {
+ "epoch": 27.489082969432314,
+ "grad_norm": 0.017741898074746132,
+ "learning_rate": 0.0006,
+ "loss": 4.279632568359375,
+ "step": 1979
+ },
+ {
+ "epoch": 27.503056768558952,
+ "grad_norm": 0.017680590972304344,
+ "learning_rate": 0.0006,
+ "loss": 4.273406028747559,
+ "step": 1980
+ },
+ {
+ "epoch": 27.51703056768559,
+ "grad_norm": 0.01864718459546566,
+ "learning_rate": 0.0006,
+ "loss": 4.278583526611328,
+ "step": 1981
+ },
+ {
+ "epoch": 27.531004366812226,
+ "grad_norm": 0.019558541476726532,
+ "learning_rate": 0.0006,
+ "loss": 4.211919784545898,
+ "step": 1982
+ },
+ {
+ "epoch": 27.544978165938865,
+ "grad_norm": 0.019467400386929512,
+ "learning_rate": 0.0006,
+ "loss": 4.276822566986084,
+ "step": 1983
+ },
+ {
+ "epoch": 27.558951965065503,
+ "grad_norm": 0.01756172813475132,
+ "learning_rate": 0.0006,
+ "loss": 4.165379047393799,
+ "step": 1984
+ },
+ {
+ "epoch": 27.57292576419214,
+ "grad_norm": 0.017458263784646988,
+ "learning_rate": 0.0006,
+ "loss": 4.2992401123046875,
+ "step": 1985
+ },
+ {
+ "epoch": 27.586899563318777,
+ "grad_norm": 0.018089665099978447,
+ "learning_rate": 0.0006,
+ "loss": 4.243378639221191,
+ "step": 1986
+ },
+ {
+ "epoch": 27.600873362445416,
+ "grad_norm": 0.019535524770617485,
+ "learning_rate": 0.0006,
+ "loss": 4.321477890014648,
+ "step": 1987
+ },
+ {
+ "epoch": 27.61484716157205,
+ "grad_norm": 0.01997970975935459,
+ "learning_rate": 0.0006,
+ "loss": 4.123082637786865,
+ "step": 1988
+ },
+ {
+ "epoch": 27.62882096069869,
+ "grad_norm": 0.01913067139685154,
+ "learning_rate": 0.0006,
+ "loss": 4.289896488189697,
+ "step": 1989
+ },
+ {
+ "epoch": 27.64279475982533,
+ "grad_norm": 0.0169806070625782,
+ "learning_rate": 0.0006,
+ "loss": 4.247004985809326,
+ "step": 1990
+ },
+ {
+ "epoch": 27.656768558951963,
+ "grad_norm": 0.017315007746219635,
+ "learning_rate": 0.0006,
+ "loss": 4.2095441818237305,
+ "step": 1991
+ },
+ {
+ "epoch": 27.670742358078602,
+ "grad_norm": 0.015098759904503822,
+ "learning_rate": 0.0006,
+ "loss": 4.145560264587402,
+ "step": 1992
+ },
+ {
+ "epoch": 27.68471615720524,
+ "grad_norm": 0.016240514814853668,
+ "learning_rate": 0.0006,
+ "loss": 4.254927635192871,
+ "step": 1993
+ },
+ {
+ "epoch": 27.69868995633188,
+ "grad_norm": 0.01621108129620552,
+ "learning_rate": 0.0006,
+ "loss": 4.219725608825684,
+ "step": 1994
+ },
+ {
+ "epoch": 27.712663755458514,
+ "grad_norm": 0.015191009268164635,
+ "learning_rate": 0.0006,
+ "loss": 4.194345474243164,
+ "step": 1995
+ },
+ {
+ "epoch": 27.726637554585153,
+ "grad_norm": 0.014387983828783035,
+ "learning_rate": 0.0006,
+ "loss": 4.211467742919922,
+ "step": 1996
+ },
+ {
+ "epoch": 27.74061135371179,
+ "grad_norm": 0.015460561029613018,
+ "learning_rate": 0.0006,
+ "loss": 4.358700752258301,
+ "step": 1997
+ },
+ {
+ "epoch": 27.754585152838427,
+ "grad_norm": 0.015100076794624329,
+ "learning_rate": 0.0006,
+ "loss": 4.171223163604736,
+ "step": 1998
+ },
+ {
+ "epoch": 27.768558951965066,
+ "grad_norm": 0.015607891604304314,
+ "learning_rate": 0.0006,
+ "loss": 4.217874526977539,
+ "step": 1999
+ },
+ {
+ "epoch": 27.782532751091704,
+ "grad_norm": 0.015650689601898193,
+ "learning_rate": 0.0006,
+ "loss": 4.280780792236328,
+ "step": 2000
+ },
+ {
+ "epoch": 27.79650655021834,
+ "grad_norm": 0.01580994389951229,
+ "learning_rate": 0.0006,
+ "loss": 4.159212112426758,
+ "step": 2001
+ },
+ {
+ "epoch": 27.810480349344978,
+ "grad_norm": 0.016607413068413734,
+ "learning_rate": 0.0006,
+ "loss": 4.238853454589844,
+ "step": 2002
+ },
+ {
+ "epoch": 27.824454148471617,
+ "grad_norm": 0.017383860424160957,
+ "learning_rate": 0.0006,
+ "loss": 4.233944892883301,
+ "step": 2003
+ },
+ {
+ "epoch": 27.83842794759825,
+ "grad_norm": 0.014966659247875214,
+ "learning_rate": 0.0006,
+ "loss": 4.167392730712891,
+ "step": 2004
+ },
+ {
+ "epoch": 27.85240174672489,
+ "grad_norm": 0.014589878730475903,
+ "learning_rate": 0.0006,
+ "loss": 4.2390899658203125,
+ "step": 2005
+ },
+ {
+ "epoch": 27.86637554585153,
+ "grad_norm": 0.017179450020194054,
+ "learning_rate": 0.0006,
+ "loss": 4.365785598754883,
+ "step": 2006
+ },
+ {
+ "epoch": 27.880349344978168,
+ "grad_norm": 0.01893901452422142,
+ "learning_rate": 0.0006,
+ "loss": 4.289430141448975,
+ "step": 2007
+ },
+ {
+ "epoch": 27.894323144104803,
+ "grad_norm": 0.018619712442159653,
+ "learning_rate": 0.0006,
+ "loss": 4.310665130615234,
+ "step": 2008
+ },
+ {
+ "epoch": 27.90829694323144,
+ "grad_norm": 0.019180385395884514,
+ "learning_rate": 0.0006,
+ "loss": 4.27454137802124,
+ "step": 2009
+ },
+ {
+ "epoch": 27.92227074235808,
+ "grad_norm": 0.017811523750424385,
+ "learning_rate": 0.0006,
+ "loss": 4.312562465667725,
+ "step": 2010
+ },
+ {
+ "epoch": 27.936244541484715,
+ "grad_norm": 0.016583573073148727,
+ "learning_rate": 0.0006,
+ "loss": 4.202253341674805,
+ "step": 2011
+ },
+ {
+ "epoch": 27.950218340611354,
+ "grad_norm": 0.01658615842461586,
+ "learning_rate": 0.0006,
+ "loss": 4.183433532714844,
+ "step": 2012
+ },
+ {
+ "epoch": 27.964192139737992,
+ "grad_norm": 0.015327401459217072,
+ "learning_rate": 0.0006,
+ "loss": 4.2125139236450195,
+ "step": 2013
+ },
+ {
+ "epoch": 27.978165938864628,
+ "grad_norm": 0.015004601329565048,
+ "learning_rate": 0.0006,
+ "loss": 4.257392883300781,
+ "step": 2014
+ },
+ {
+ "epoch": 27.992139737991266,
+ "grad_norm": 0.015259161591529846,
+ "learning_rate": 0.0006,
+ "loss": 4.241818428039551,
+ "step": 2015
+ },
+ {
+ "epoch": 28.0,
+ "grad_norm": 0.016808776184916496,
+ "learning_rate": 0.0006,
+ "loss": 4.146241188049316,
+ "step": 2016
+ },
+ {
+ "epoch": 28.0,
+ "eval_loss": 4.663090705871582,
+ "eval_runtime": 56.4559,
+ "eval_samples_per_second": 43.255,
+ "eval_steps_per_second": 1.364,
+ "step": 2016
+ },
+ {
+ "epoch": 28.01397379912664,
+ "grad_norm": 0.015784382820129395,
+ "learning_rate": 0.0006,
+ "loss": 4.044191360473633,
+ "step": 2017
+ },
+ {
+ "epoch": 28.027947598253274,
+ "grad_norm": 0.015636710450053215,
+ "learning_rate": 0.0006,
+ "loss": 4.128754615783691,
+ "step": 2018
+ },
+ {
+ "epoch": 28.041921397379912,
+ "grad_norm": 0.015364975668489933,
+ "learning_rate": 0.0006,
+ "loss": 4.191835403442383,
+ "step": 2019
+ },
+ {
+ "epoch": 28.05589519650655,
+ "grad_norm": 0.015015700832009315,
+ "learning_rate": 0.0006,
+ "loss": 4.120022773742676,
+ "step": 2020
+ },
+ {
+ "epoch": 28.069868995633186,
+ "grad_norm": 0.016666799783706665,
+ "learning_rate": 0.0006,
+ "loss": 4.33906364440918,
+ "step": 2021
+ },
+ {
+ "epoch": 28.083842794759825,
+ "grad_norm": 0.01780104823410511,
+ "learning_rate": 0.0006,
+ "loss": 4.279458999633789,
+ "step": 2022
+ },
+ {
+ "epoch": 28.097816593886463,
+ "grad_norm": 0.01870882883667946,
+ "learning_rate": 0.0006,
+ "loss": 4.34728479385376,
+ "step": 2023
+ },
+ {
+ "epoch": 28.111790393013102,
+ "grad_norm": 0.018406381830573082,
+ "learning_rate": 0.0006,
+ "loss": 4.211311340332031,
+ "step": 2024
+ },
+ {
+ "epoch": 28.125764192139737,
+ "grad_norm": 0.017833448946475983,
+ "learning_rate": 0.0006,
+ "loss": 4.223254203796387,
+ "step": 2025
+ },
+ {
+ "epoch": 28.139737991266376,
+ "grad_norm": 0.017552688717842102,
+ "learning_rate": 0.0006,
+ "loss": 4.208823204040527,
+ "step": 2026
+ },
+ {
+ "epoch": 28.153711790393015,
+ "grad_norm": 0.018500229343771935,
+ "learning_rate": 0.0006,
+ "loss": 4.198025226593018,
+ "step": 2027
+ },
+ {
+ "epoch": 28.16768558951965,
+ "grad_norm": 0.01782156340777874,
+ "learning_rate": 0.0006,
+ "loss": 4.179978370666504,
+ "step": 2028
+ },
+ {
+ "epoch": 28.18165938864629,
+ "grad_norm": 0.019580967724323273,
+ "learning_rate": 0.0006,
+ "loss": 4.291114807128906,
+ "step": 2029
+ },
+ {
+ "epoch": 28.195633187772927,
+ "grad_norm": 0.020804036408662796,
+ "learning_rate": 0.0006,
+ "loss": 4.220607280731201,
+ "step": 2030
+ },
+ {
+ "epoch": 28.209606986899562,
+ "grad_norm": 0.019595693796873093,
+ "learning_rate": 0.0006,
+ "loss": 4.2905731201171875,
+ "step": 2031
+ },
+ {
+ "epoch": 28.2235807860262,
+ "grad_norm": 0.019247086718678474,
+ "learning_rate": 0.0006,
+ "loss": 4.2467360496521,
+ "step": 2032
+ },
+ {
+ "epoch": 28.23755458515284,
+ "grad_norm": 0.01958036608994007,
+ "learning_rate": 0.0006,
+ "loss": 4.286137580871582,
+ "step": 2033
+ },
+ {
+ "epoch": 28.251528384279474,
+ "grad_norm": 0.021354753524065018,
+ "learning_rate": 0.0006,
+ "loss": 4.234755039215088,
+ "step": 2034
+ },
+ {
+ "epoch": 28.265502183406113,
+ "grad_norm": 0.02206375077366829,
+ "learning_rate": 0.0006,
+ "loss": 4.149641036987305,
+ "step": 2035
+ },
+ {
+ "epoch": 28.27947598253275,
+ "grad_norm": 0.02201310358941555,
+ "learning_rate": 0.0006,
+ "loss": 4.1928629875183105,
+ "step": 2036
+ },
+ {
+ "epoch": 28.29344978165939,
+ "grad_norm": 0.01961950585246086,
+ "learning_rate": 0.0006,
+ "loss": 4.211404323577881,
+ "step": 2037
+ },
+ {
+ "epoch": 28.307423580786025,
+ "grad_norm": 0.019211795181035995,
+ "learning_rate": 0.0006,
+ "loss": 4.200976848602295,
+ "step": 2038
+ },
+ {
+ "epoch": 28.321397379912664,
+ "grad_norm": 0.01779370754957199,
+ "learning_rate": 0.0006,
+ "loss": 4.249148845672607,
+ "step": 2039
+ },
+ {
+ "epoch": 28.335371179039303,
+ "grad_norm": 0.018635908141732216,
+ "learning_rate": 0.0006,
+ "loss": 4.11806058883667,
+ "step": 2040
+ },
+ {
+ "epoch": 28.349344978165938,
+ "grad_norm": 0.019792694598436356,
+ "learning_rate": 0.0006,
+ "loss": 4.335224151611328,
+ "step": 2041
+ },
+ {
+ "epoch": 28.363318777292577,
+ "grad_norm": 0.021731717512011528,
+ "learning_rate": 0.0006,
+ "loss": 4.302778244018555,
+ "step": 2042
+ },
+ {
+ "epoch": 28.377292576419215,
+ "grad_norm": 0.020823447033762932,
+ "learning_rate": 0.0006,
+ "loss": 4.260610580444336,
+ "step": 2043
+ },
+ {
+ "epoch": 28.39126637554585,
+ "grad_norm": 0.019385412335395813,
+ "learning_rate": 0.0006,
+ "loss": 4.2247724533081055,
+ "step": 2044
+ },
+ {
+ "epoch": 28.40524017467249,
+ "grad_norm": 0.019539108499884605,
+ "learning_rate": 0.0006,
+ "loss": 4.236655235290527,
+ "step": 2045
+ },
+ {
+ "epoch": 28.419213973799128,
+ "grad_norm": 0.01942690648138523,
+ "learning_rate": 0.0006,
+ "loss": 4.251850605010986,
+ "step": 2046
+ },
+ {
+ "epoch": 28.433187772925763,
+ "grad_norm": 0.019667886197566986,
+ "learning_rate": 0.0006,
+ "loss": 4.222312927246094,
+ "step": 2047
+ },
+ {
+ "epoch": 28.4471615720524,
+ "grad_norm": 0.02029012329876423,
+ "learning_rate": 0.0006,
+ "loss": 4.246252059936523,
+ "step": 2048
+ },
+ {
+ "epoch": 28.46113537117904,
+ "grad_norm": 0.01784469373524189,
+ "learning_rate": 0.0006,
+ "loss": 4.192128658294678,
+ "step": 2049
+ },
+ {
+ "epoch": 28.475109170305675,
+ "grad_norm": 0.0160287544131279,
+ "learning_rate": 0.0006,
+ "loss": 4.208607196807861,
+ "step": 2050
+ },
+ {
+ "epoch": 28.489082969432314,
+ "grad_norm": 0.015348346903920174,
+ "learning_rate": 0.0006,
+ "loss": 4.236598014831543,
+ "step": 2051
+ },
+ {
+ "epoch": 28.503056768558952,
+ "grad_norm": 0.015198206529021263,
+ "learning_rate": 0.0006,
+ "loss": 4.235552787780762,
+ "step": 2052
+ },
+ {
+ "epoch": 28.51703056768559,
+ "grad_norm": 0.01523754745721817,
+ "learning_rate": 0.0006,
+ "loss": 4.1941728591918945,
+ "step": 2053
+ },
+ {
+ "epoch": 28.531004366812226,
+ "grad_norm": 0.015559614636003971,
+ "learning_rate": 0.0006,
+ "loss": 4.294757843017578,
+ "step": 2054
+ },
+ {
+ "epoch": 28.544978165938865,
+ "grad_norm": 0.01568358950316906,
+ "learning_rate": 0.0006,
+ "loss": 4.213065147399902,
+ "step": 2055
+ },
+ {
+ "epoch": 28.558951965065503,
+ "grad_norm": 0.015336516313254833,
+ "learning_rate": 0.0006,
+ "loss": 4.240293979644775,
+ "step": 2056
+ },
+ {
+ "epoch": 28.57292576419214,
+ "grad_norm": 0.015041496604681015,
+ "learning_rate": 0.0006,
+ "loss": 4.273397445678711,
+ "step": 2057
+ },
+ {
+ "epoch": 28.586899563318777,
+ "grad_norm": 0.01610128954052925,
+ "learning_rate": 0.0006,
+ "loss": 4.125369071960449,
+ "step": 2058
+ },
+ {
+ "epoch": 28.600873362445416,
+ "grad_norm": 0.015974976122379303,
+ "learning_rate": 0.0006,
+ "loss": 4.133459091186523,
+ "step": 2059
+ },
+ {
+ "epoch": 28.61484716157205,
+ "grad_norm": 0.015936799347400665,
+ "learning_rate": 0.0006,
+ "loss": 4.306667327880859,
+ "step": 2060
+ },
+ {
+ "epoch": 28.62882096069869,
+ "grad_norm": 0.01795247197151184,
+ "learning_rate": 0.0006,
+ "loss": 4.205156326293945,
+ "step": 2061
+ },
+ {
+ "epoch": 28.64279475982533,
+ "grad_norm": 0.019639814272522926,
+ "learning_rate": 0.0006,
+ "loss": 4.277614593505859,
+ "step": 2062
+ },
+ {
+ "epoch": 28.656768558951963,
+ "grad_norm": 0.01722300611436367,
+ "learning_rate": 0.0006,
+ "loss": 4.191032886505127,
+ "step": 2063
+ },
+ {
+ "epoch": 28.670742358078602,
+ "grad_norm": 0.01612633652985096,
+ "learning_rate": 0.0006,
+ "loss": 4.20717191696167,
+ "step": 2064
+ },
+ {
+ "epoch": 28.68471615720524,
+ "grad_norm": 0.01603279449045658,
+ "learning_rate": 0.0006,
+ "loss": 4.228884220123291,
+ "step": 2065
+ },
+ {
+ "epoch": 28.69868995633188,
+ "grad_norm": 0.016405927017331123,
+ "learning_rate": 0.0006,
+ "loss": 4.201292991638184,
+ "step": 2066
+ },
+ {
+ "epoch": 28.712663755458514,
+ "grad_norm": 0.016644183546304703,
+ "learning_rate": 0.0006,
+ "loss": 4.239282131195068,
+ "step": 2067
+ },
+ {
+ "epoch": 28.726637554585153,
+ "grad_norm": 0.0175587497651577,
+ "learning_rate": 0.0006,
+ "loss": 4.289038181304932,
+ "step": 2068
+ },
+ {
+ "epoch": 28.74061135371179,
+ "grad_norm": 0.018223397433757782,
+ "learning_rate": 0.0006,
+ "loss": 4.2266998291015625,
+ "step": 2069
+ },
+ {
+ "epoch": 28.754585152838427,
+ "grad_norm": 0.018680868670344353,
+ "learning_rate": 0.0006,
+ "loss": 4.348849296569824,
+ "step": 2070
+ },
+ {
+ "epoch": 28.768558951965066,
+ "grad_norm": 0.016726728528738022,
+ "learning_rate": 0.0006,
+ "loss": 4.292080879211426,
+ "step": 2071
+ },
+ {
+ "epoch": 28.782532751091704,
+ "grad_norm": 0.017386795952916145,
+ "learning_rate": 0.0006,
+ "loss": 4.208870887756348,
+ "step": 2072
+ },
+ {
+ "epoch": 28.79650655021834,
+ "grad_norm": 0.01865178905427456,
+ "learning_rate": 0.0006,
+ "loss": 4.054150581359863,
+ "step": 2073
+ },
+ {
+ "epoch": 28.810480349344978,
+ "grad_norm": 0.01776815392076969,
+ "learning_rate": 0.0006,
+ "loss": 4.202404499053955,
+ "step": 2074
+ },
+ {
+ "epoch": 28.824454148471617,
+ "grad_norm": 0.01643497310578823,
+ "learning_rate": 0.0006,
+ "loss": 4.131855010986328,
+ "step": 2075
+ },
+ {
+ "epoch": 28.83842794759825,
+ "grad_norm": 0.01574273779988289,
+ "learning_rate": 0.0006,
+ "loss": 4.296788692474365,
+ "step": 2076
+ },
+ {
+ "epoch": 28.85240174672489,
+ "grad_norm": 0.015613908879458904,
+ "learning_rate": 0.0006,
+ "loss": 4.210949897766113,
+ "step": 2077
+ },
+ {
+ "epoch": 28.86637554585153,
+ "grad_norm": 0.015543212182819843,
+ "learning_rate": 0.0006,
+ "loss": 4.218136310577393,
+ "step": 2078
+ },
+ {
+ "epoch": 28.880349344978168,
+ "grad_norm": 0.014681251719594002,
+ "learning_rate": 0.0006,
+ "loss": 4.19401741027832,
+ "step": 2079
+ },
+ {
+ "epoch": 28.894323144104803,
+ "grad_norm": 0.014393828809261322,
+ "learning_rate": 0.0006,
+ "loss": 4.198468208312988,
+ "step": 2080
+ },
+ {
+ "epoch": 28.90829694323144,
+ "grad_norm": 0.01580170728266239,
+ "learning_rate": 0.0006,
+ "loss": 4.224969863891602,
+ "step": 2081
+ },
+ {
+ "epoch": 28.92227074235808,
+ "grad_norm": 0.01484165620058775,
+ "learning_rate": 0.0006,
+ "loss": 4.12236213684082,
+ "step": 2082
+ },
+ {
+ "epoch": 28.936244541484715,
+ "grad_norm": 0.01497623696923256,
+ "learning_rate": 0.0006,
+ "loss": 4.24141788482666,
+ "step": 2083
+ },
+ {
+ "epoch": 28.950218340611354,
+ "grad_norm": 0.015080046840012074,
+ "learning_rate": 0.0006,
+ "loss": 4.284151077270508,
+ "step": 2084
+ },
+ {
+ "epoch": 28.964192139737992,
+ "grad_norm": 0.016207212582230568,
+ "learning_rate": 0.0006,
+ "loss": 4.298693656921387,
+ "step": 2085
+ },
+ {
+ "epoch": 28.978165938864628,
+ "grad_norm": 0.015474417246878147,
+ "learning_rate": 0.0006,
+ "loss": 4.2412261962890625,
+ "step": 2086
+ },
+ {
+ "epoch": 28.992139737991266,
+ "grad_norm": 0.01515351701527834,
+ "learning_rate": 0.0006,
+ "loss": 4.3054423332214355,
+ "step": 2087
+ },
+ {
+ "epoch": 29.0,
+ "grad_norm": 0.016411345452070236,
+ "learning_rate": 0.0006,
+ "loss": 4.104830741882324,
+ "step": 2088
+ },
+ {
+ "epoch": 29.0,
+ "eval_loss": 4.661567687988281,
+ "eval_runtime": 57.1244,
+ "eval_samples_per_second": 42.749,
+ "eval_steps_per_second": 1.348,
+ "step": 2088
+ },
+ {
+ "epoch": 29.01397379912664,
+ "grad_norm": 0.01665342226624489,
+ "learning_rate": 0.0006,
+ "loss": 4.311391830444336,
+ "step": 2089
+ },
+ {
+ "epoch": 29.027947598253274,
+ "grad_norm": 0.018053214997053146,
+ "learning_rate": 0.0006,
+ "loss": 4.135597229003906,
+ "step": 2090
+ },
+ {
+ "epoch": 29.041921397379912,
+ "grad_norm": 0.018213748931884766,
+ "learning_rate": 0.0006,
+ "loss": 4.303299903869629,
+ "step": 2091
+ },
+ {
+ "epoch": 29.05589519650655,
+ "grad_norm": 0.018460242077708244,
+ "learning_rate": 0.0006,
+ "loss": 4.20850944519043,
+ "step": 2092
+ },
+ {
+ "epoch": 29.069868995633186,
+ "grad_norm": 0.018365809693932533,
+ "learning_rate": 0.0006,
+ "loss": 4.275847434997559,
+ "step": 2093
+ },
+ {
+ "epoch": 29.083842794759825,
+ "grad_norm": 0.018346186727285385,
+ "learning_rate": 0.0006,
+ "loss": 4.202091217041016,
+ "step": 2094
+ },
+ {
+ "epoch": 29.097816593886463,
+ "grad_norm": 0.017920587211847305,
+ "learning_rate": 0.0006,
+ "loss": 4.182092666625977,
+ "step": 2095
+ },
+ {
+ "epoch": 29.111790393013102,
+ "grad_norm": 0.01812003180384636,
+ "learning_rate": 0.0006,
+ "loss": 4.259721755981445,
+ "step": 2096
+ },
+ {
+ "epoch": 29.125764192139737,
+ "grad_norm": 0.01700986735522747,
+ "learning_rate": 0.0006,
+ "loss": 4.067573070526123,
+ "step": 2097
+ },
+ {
+ "epoch": 29.139737991266376,
+ "grad_norm": 0.018069760873913765,
+ "learning_rate": 0.0006,
+ "loss": 4.234400749206543,
+ "step": 2098
+ },
+ {
+ "epoch": 29.153711790393015,
+ "grad_norm": 0.01905428245663643,
+ "learning_rate": 0.0006,
+ "loss": 4.16609001159668,
+ "step": 2099
+ },
+ {
+ "epoch": 29.16768558951965,
+ "grad_norm": 0.018711242824792862,
+ "learning_rate": 0.0006,
+ "loss": 4.249088764190674,
+ "step": 2100
+ },
+ {
+ "epoch": 29.18165938864629,
+ "grad_norm": 0.018721306696534157,
+ "learning_rate": 0.0006,
+ "loss": 4.14540958404541,
+ "step": 2101
+ },
+ {
+ "epoch": 29.195633187772927,
+ "grad_norm": 0.020139874890446663,
+ "learning_rate": 0.0006,
+ "loss": 4.049248695373535,
+ "step": 2102
+ },
+ {
+ "epoch": 29.209606986899562,
+ "grad_norm": 0.020849574357271194,
+ "learning_rate": 0.0006,
+ "loss": 4.235224723815918,
+ "step": 2103
+ },
+ {
+ "epoch": 29.2235807860262,
+ "grad_norm": 0.018693549558520317,
+ "learning_rate": 0.0006,
+ "loss": 4.232961177825928,
+ "step": 2104
+ },
+ {
+ "epoch": 29.23755458515284,
+ "grad_norm": 0.017889857292175293,
+ "learning_rate": 0.0006,
+ "loss": 4.171422004699707,
+ "step": 2105
+ },
+ {
+ "epoch": 29.251528384279474,
+ "grad_norm": 0.016690224409103394,
+ "learning_rate": 0.0006,
+ "loss": 4.244717597961426,
+ "step": 2106
+ },
+ {
+ "epoch": 29.265502183406113,
+ "grad_norm": 0.01708616502583027,
+ "learning_rate": 0.0006,
+ "loss": 4.125068664550781,
+ "step": 2107
+ },
+ {
+ "epoch": 29.27947598253275,
+ "grad_norm": 0.017928237095475197,
+ "learning_rate": 0.0006,
+ "loss": 4.208292007446289,
+ "step": 2108
+ },
+ {
+ "epoch": 29.29344978165939,
+ "grad_norm": 0.018250394612550735,
+ "learning_rate": 0.0006,
+ "loss": 4.118630886077881,
+ "step": 2109
+ },
+ {
+ "epoch": 29.307423580786025,
+ "grad_norm": 0.020160246640443802,
+ "learning_rate": 0.0006,
+ "loss": 4.206025123596191,
+ "step": 2110
+ },
+ {
+ "epoch": 29.321397379912664,
+ "grad_norm": 0.020045453682541847,
+ "learning_rate": 0.0006,
+ "loss": 4.2758684158325195,
+ "step": 2111
+ },
+ {
+ "epoch": 29.335371179039303,
+ "grad_norm": 0.019556907936930656,
+ "learning_rate": 0.0006,
+ "loss": 4.188272953033447,
+ "step": 2112
+ },
+ {
+ "epoch": 29.349344978165938,
+ "grad_norm": 0.019185200333595276,
+ "learning_rate": 0.0006,
+ "loss": 4.17451286315918,
+ "step": 2113
+ },
+ {
+ "epoch": 29.363318777292577,
+ "grad_norm": 0.01816629432141781,
+ "learning_rate": 0.0006,
+ "loss": 4.165246963500977,
+ "step": 2114
+ },
+ {
+ "epoch": 29.377292576419215,
+ "grad_norm": 0.01865146867930889,
+ "learning_rate": 0.0006,
+ "loss": 4.148184776306152,
+ "step": 2115
+ },
+ {
+ "epoch": 29.39126637554585,
+ "grad_norm": 0.018338464200496674,
+ "learning_rate": 0.0006,
+ "loss": 4.288936614990234,
+ "step": 2116
+ },
+ {
+ "epoch": 29.40524017467249,
+ "grad_norm": 0.018986187875270844,
+ "learning_rate": 0.0006,
+ "loss": 4.183379650115967,
+ "step": 2117
+ },
+ {
+ "epoch": 29.419213973799128,
+ "grad_norm": 0.018409637734293938,
+ "learning_rate": 0.0006,
+ "loss": 4.057574272155762,
+ "step": 2118
+ },
+ {
+ "epoch": 29.433187772925763,
+ "grad_norm": 0.016664542257785797,
+ "learning_rate": 0.0006,
+ "loss": 4.18517541885376,
+ "step": 2119
+ },
+ {
+ "epoch": 29.4471615720524,
+ "grad_norm": 0.016486089676618576,
+ "learning_rate": 0.0006,
+ "loss": 4.179367542266846,
+ "step": 2120
+ },
+ {
+ "epoch": 29.46113537117904,
+ "grad_norm": 0.01787523552775383,
+ "learning_rate": 0.0006,
+ "loss": 4.184126377105713,
+ "step": 2121
+ },
+ {
+ "epoch": 29.475109170305675,
+ "grad_norm": 0.016574613749980927,
+ "learning_rate": 0.0006,
+ "loss": 4.290170192718506,
+ "step": 2122
+ },
+ {
+ "epoch": 29.489082969432314,
+ "grad_norm": 0.01643376611173153,
+ "learning_rate": 0.0006,
+ "loss": 4.052913665771484,
+ "step": 2123
+ },
+ {
+ "epoch": 29.503056768558952,
+ "grad_norm": 0.016503969207406044,
+ "learning_rate": 0.0006,
+ "loss": 4.335184097290039,
+ "step": 2124
+ },
+ {
+ "epoch": 29.51703056768559,
+ "grad_norm": 0.017596479505300522,
+ "learning_rate": 0.0006,
+ "loss": 4.165539741516113,
+ "step": 2125
+ },
+ {
+ "epoch": 29.531004366812226,
+ "grad_norm": 0.017468582838773727,
+ "learning_rate": 0.0006,
+ "loss": 4.315964698791504,
+ "step": 2126
+ },
+ {
+ "epoch": 29.544978165938865,
+ "grad_norm": 0.01724618673324585,
+ "learning_rate": 0.0006,
+ "loss": 4.243681907653809,
+ "step": 2127
+ },
+ {
+ "epoch": 29.558951965065503,
+ "grad_norm": 0.018384616822004318,
+ "learning_rate": 0.0006,
+ "loss": 4.112447261810303,
+ "step": 2128
+ },
+ {
+ "epoch": 29.57292576419214,
+ "grad_norm": 0.01824074238538742,
+ "learning_rate": 0.0006,
+ "loss": 4.236065864562988,
+ "step": 2129
+ },
+ {
+ "epoch": 29.586899563318777,
+ "grad_norm": 0.01625971868634224,
+ "learning_rate": 0.0006,
+ "loss": 4.18143892288208,
+ "step": 2130
+ },
+ {
+ "epoch": 29.600873362445416,
+ "grad_norm": 0.015537494793534279,
+ "learning_rate": 0.0006,
+ "loss": 4.225147247314453,
+ "step": 2131
+ },
+ {
+ "epoch": 29.61484716157205,
+ "grad_norm": 0.015818974003195763,
+ "learning_rate": 0.0006,
+ "loss": 4.3100175857543945,
+ "step": 2132
+ },
+ {
+ "epoch": 29.62882096069869,
+ "grad_norm": 0.015909671783447266,
+ "learning_rate": 0.0006,
+ "loss": 4.169775009155273,
+ "step": 2133
+ },
+ {
+ "epoch": 29.64279475982533,
+ "grad_norm": 0.01769658550620079,
+ "learning_rate": 0.0006,
+ "loss": 4.278904914855957,
+ "step": 2134
+ },
+ {
+ "epoch": 29.656768558951963,
+ "grad_norm": 0.015641704201698303,
+ "learning_rate": 0.0006,
+ "loss": 4.2426228523254395,
+ "step": 2135
+ },
+ {
+ "epoch": 29.670742358078602,
+ "grad_norm": 0.01565658301115036,
+ "learning_rate": 0.0006,
+ "loss": 4.232687950134277,
+ "step": 2136
+ },
+ {
+ "epoch": 29.68471615720524,
+ "grad_norm": 0.016485940665006638,
+ "learning_rate": 0.0006,
+ "loss": 4.209839820861816,
+ "step": 2137
+ },
+ {
+ "epoch": 29.69868995633188,
+ "grad_norm": 0.017622729763388634,
+ "learning_rate": 0.0006,
+ "loss": 4.196017265319824,
+ "step": 2138
+ },
+ {
+ "epoch": 29.712663755458514,
+ "grad_norm": 0.018320564180612564,
+ "learning_rate": 0.0006,
+ "loss": 4.225683689117432,
+ "step": 2139
+ },
+ {
+ "epoch": 29.726637554585153,
+ "grad_norm": 0.018140794709324837,
+ "learning_rate": 0.0006,
+ "loss": 4.311944961547852,
+ "step": 2140
+ },
+ {
+ "epoch": 29.74061135371179,
+ "grad_norm": 0.016227100044488907,
+ "learning_rate": 0.0006,
+ "loss": 4.230257511138916,
+ "step": 2141
+ },
+ {
+ "epoch": 29.754585152838427,
+ "grad_norm": 0.015759488567709923,
+ "learning_rate": 0.0006,
+ "loss": 4.119174003601074,
+ "step": 2142
+ },
+ {
+ "epoch": 29.768558951965066,
+ "grad_norm": 0.01757766678929329,
+ "learning_rate": 0.0006,
+ "loss": 4.165826797485352,
+ "step": 2143
+ },
+ {
+ "epoch": 29.782532751091704,
+ "grad_norm": 0.019661923870444298,
+ "learning_rate": 0.0006,
+ "loss": 4.392203330993652,
+ "step": 2144
+ },
+ {
+ "epoch": 29.79650655021834,
+ "grad_norm": 0.019478755071759224,
+ "learning_rate": 0.0006,
+ "loss": 4.348371982574463,
+ "step": 2145
+ },
+ {
+ "epoch": 29.810480349344978,
+ "grad_norm": 0.019149569794535637,
+ "learning_rate": 0.0006,
+ "loss": 4.34104061126709,
+ "step": 2146
+ },
+ {
+ "epoch": 29.824454148471617,
+ "grad_norm": 0.017707521095871925,
+ "learning_rate": 0.0006,
+ "loss": 4.228226661682129,
+ "step": 2147
+ },
+ {
+ "epoch": 29.83842794759825,
+ "grad_norm": 0.015615378506481647,
+ "learning_rate": 0.0006,
+ "loss": 4.175541877746582,
+ "step": 2148
+ },
+ {
+ "epoch": 29.85240174672489,
+ "grad_norm": 0.016111129894852638,
+ "learning_rate": 0.0006,
+ "loss": 4.267936706542969,
+ "step": 2149
+ },
+ {
+ "epoch": 29.86637554585153,
+ "grad_norm": 0.01528779324144125,
+ "learning_rate": 0.0006,
+ "loss": 4.186014175415039,
+ "step": 2150
+ },
+ {
+ "epoch": 29.880349344978168,
+ "grad_norm": 0.014925646595656872,
+ "learning_rate": 0.0006,
+ "loss": 4.174860000610352,
+ "step": 2151
+ },
+ {
+ "epoch": 29.894323144104803,
+ "grad_norm": 0.016917673870921135,
+ "learning_rate": 0.0006,
+ "loss": 4.33908748626709,
+ "step": 2152
+ },
+ {
+ "epoch": 29.90829694323144,
+ "grad_norm": 0.016007075086236,
+ "learning_rate": 0.0006,
+ "loss": 4.266141414642334,
+ "step": 2153
+ },
+ {
+ "epoch": 29.92227074235808,
+ "grad_norm": 0.016178956255316734,
+ "learning_rate": 0.0006,
+ "loss": 4.271215915679932,
+ "step": 2154
+ },
+ {
+ "epoch": 29.936244541484715,
+ "grad_norm": 0.016180871054530144,
+ "learning_rate": 0.0006,
+ "loss": 4.278801441192627,
+ "step": 2155
+ },
+ {
+ "epoch": 29.950218340611354,
+ "grad_norm": 0.015335503034293652,
+ "learning_rate": 0.0006,
+ "loss": 4.194978713989258,
+ "step": 2156
+ },
+ {
+ "epoch": 29.964192139737992,
+ "grad_norm": 0.016580764204263687,
+ "learning_rate": 0.0006,
+ "loss": 4.290809631347656,
+ "step": 2157
+ },
+ {
+ "epoch": 29.978165938864628,
+ "grad_norm": 0.01623128168284893,
+ "learning_rate": 0.0006,
+ "loss": 4.205451488494873,
+ "step": 2158
+ },
+ {
+ "epoch": 29.992139737991266,
+ "grad_norm": 0.014904248528182507,
+ "learning_rate": 0.0006,
+ "loss": 4.24924373626709,
+ "step": 2159
+ },
+ {
+ "epoch": 30.0,
+ "grad_norm": 0.017089299857616425,
+ "learning_rate": 0.0006,
+ "loss": 4.222072601318359,
+ "step": 2160
+ },
+ {
+ "epoch": 30.0,
+ "eval_loss": 4.585411071777344,
+ "eval_runtime": 57.1164,
+ "eval_samples_per_second": 42.755,
+ "eval_steps_per_second": 1.348,
+ "step": 2160
+ },
+ {
+ "epoch": 30.01397379912664,
+ "grad_norm": 0.017226383090019226,
+ "learning_rate": 0.0006,
+ "loss": 4.19467830657959,
+ "step": 2161
+ },
+ {
+ "epoch": 30.027947598253274,
+ "grad_norm": 0.01635066606104374,
+ "learning_rate": 0.0006,
+ "loss": 4.176875591278076,
+ "step": 2162
+ },
+ {
+ "epoch": 30.041921397379912,
+ "grad_norm": 0.01735362410545349,
+ "learning_rate": 0.0006,
+ "loss": 4.13667106628418,
+ "step": 2163
+ },
+ {
+ "epoch": 30.05589519650655,
+ "grad_norm": 0.017047109082341194,
+ "learning_rate": 0.0006,
+ "loss": 4.136329650878906,
+ "step": 2164
+ },
+ {
+ "epoch": 30.069868995633186,
+ "grad_norm": 0.018116053193807602,
+ "learning_rate": 0.0006,
+ "loss": 4.228706359863281,
+ "step": 2165
+ },
+ {
+ "epoch": 30.083842794759825,
+ "grad_norm": 0.01791389286518097,
+ "learning_rate": 0.0006,
+ "loss": 4.1597747802734375,
+ "step": 2166
+ },
+ {
+ "epoch": 30.097816593886463,
+ "grad_norm": 0.017793502658605576,
+ "learning_rate": 0.0006,
+ "loss": 4.0807976722717285,
+ "step": 2167
+ },
+ {
+ "epoch": 30.111790393013102,
+ "grad_norm": 0.01827503927052021,
+ "learning_rate": 0.0006,
+ "loss": 4.170646667480469,
+ "step": 2168
+ },
+ {
+ "epoch": 30.125764192139737,
+ "grad_norm": 0.017228564247488976,
+ "learning_rate": 0.0006,
+ "loss": 4.146934509277344,
+ "step": 2169
+ },
+ {
+ "epoch": 30.139737991266376,
+ "grad_norm": 0.01757437363266945,
+ "learning_rate": 0.0006,
+ "loss": 4.1859130859375,
+ "step": 2170
+ },
+ {
+ "epoch": 30.153711790393015,
+ "grad_norm": 0.01701093092560768,
+ "learning_rate": 0.0006,
+ "loss": 4.126981258392334,
+ "step": 2171
+ },
+ {
+ "epoch": 30.16768558951965,
+ "grad_norm": 0.015220489352941513,
+ "learning_rate": 0.0006,
+ "loss": 4.164018630981445,
+ "step": 2172
+ },
+ {
+ "epoch": 30.18165938864629,
+ "grad_norm": 0.016254238784313202,
+ "learning_rate": 0.0006,
+ "loss": 4.258275508880615,
+ "step": 2173
+ },
+ {
+ "epoch": 30.195633187772927,
+ "grad_norm": 0.016609614714980125,
+ "learning_rate": 0.0006,
+ "loss": 4.222233772277832,
+ "step": 2174
+ },
+ {
+ "epoch": 30.209606986899562,
+ "grad_norm": 0.016531318426132202,
+ "learning_rate": 0.0006,
+ "loss": 4.144489288330078,
+ "step": 2175
+ },
+ {
+ "epoch": 30.2235807860262,
+ "grad_norm": 0.01567668654024601,
+ "learning_rate": 0.0006,
+ "loss": 4.181916236877441,
+ "step": 2176
+ },
+ {
+ "epoch": 30.23755458515284,
+ "grad_norm": 0.017195681110024452,
+ "learning_rate": 0.0006,
+ "loss": 4.189693927764893,
+ "step": 2177
+ },
+ {
+ "epoch": 30.251528384279474,
+ "grad_norm": 0.016882948577404022,
+ "learning_rate": 0.0006,
+ "loss": 4.21604061126709,
+ "step": 2178
+ },
+ {
+ "epoch": 30.265502183406113,
+ "grad_norm": 0.016261622309684753,
+ "learning_rate": 0.0006,
+ "loss": 4.227639198303223,
+ "step": 2179
+ },
+ {
+ "epoch": 30.27947598253275,
+ "grad_norm": 0.016411812976002693,
+ "learning_rate": 0.0006,
+ "loss": 4.130362510681152,
+ "step": 2180
+ },
+ {
+ "epoch": 30.29344978165939,
+ "grad_norm": 0.01732019893825054,
+ "learning_rate": 0.0006,
+ "loss": 4.230443000793457,
+ "step": 2181
+ },
+ {
+ "epoch": 30.307423580786025,
+ "grad_norm": 0.01825905591249466,
+ "learning_rate": 0.0006,
+ "loss": 4.175817489624023,
+ "step": 2182
+ },
+ {
+ "epoch": 30.321397379912664,
+ "grad_norm": 0.01887713558971882,
+ "learning_rate": 0.0006,
+ "loss": 4.169020652770996,
+ "step": 2183
+ },
+ {
+ "epoch": 30.335371179039303,
+ "grad_norm": 0.019391661509871483,
+ "learning_rate": 0.0006,
+ "loss": 4.176318645477295,
+ "step": 2184
+ },
+ {
+ "epoch": 30.349344978165938,
+ "grad_norm": 0.020840002223849297,
+ "learning_rate": 0.0006,
+ "loss": 4.174228668212891,
+ "step": 2185
+ },
+ {
+ "epoch": 30.363318777292577,
+ "grad_norm": 0.020193178206682205,
+ "learning_rate": 0.0006,
+ "loss": 4.127919673919678,
+ "step": 2186
+ },
+ {
+ "epoch": 30.377292576419215,
+ "grad_norm": 0.018449561670422554,
+ "learning_rate": 0.0006,
+ "loss": 4.229487895965576,
+ "step": 2187
+ },
+ {
+ "epoch": 30.39126637554585,
+ "grad_norm": 0.01711983233690262,
+ "learning_rate": 0.0006,
+ "loss": 4.201651573181152,
+ "step": 2188
+ },
+ {
+ "epoch": 30.40524017467249,
+ "grad_norm": 0.017293395474553108,
+ "learning_rate": 0.0006,
+ "loss": 4.168704032897949,
+ "step": 2189
+ },
+ {
+ "epoch": 30.419213973799128,
+ "grad_norm": 0.018526358529925346,
+ "learning_rate": 0.0006,
+ "loss": 4.171864032745361,
+ "step": 2190
+ },
+ {
+ "epoch": 30.433187772925763,
+ "grad_norm": 0.019287291914224625,
+ "learning_rate": 0.0006,
+ "loss": 4.039534568786621,
+ "step": 2191
+ },
+ {
+ "epoch": 30.4471615720524,
+ "grad_norm": 0.016902444884181023,
+ "learning_rate": 0.0006,
+ "loss": 4.282787799835205,
+ "step": 2192
+ },
+ {
+ "epoch": 30.46113537117904,
+ "grad_norm": 0.01656750962138176,
+ "learning_rate": 0.0006,
+ "loss": 4.213098526000977,
+ "step": 2193
+ },
+ {
+ "epoch": 30.475109170305675,
+ "grad_norm": 0.016805065795779228,
+ "learning_rate": 0.0006,
+ "loss": 4.1583476066589355,
+ "step": 2194
+ },
+ {
+ "epoch": 30.489082969432314,
+ "grad_norm": 0.017166638746857643,
+ "learning_rate": 0.0006,
+ "loss": 4.188634395599365,
+ "step": 2195
+ },
+ {
+ "epoch": 30.503056768558952,
+ "grad_norm": 0.01592106744647026,
+ "learning_rate": 0.0006,
+ "loss": 4.234043121337891,
+ "step": 2196
+ },
+ {
+ "epoch": 30.51703056768559,
+ "grad_norm": 0.015270989388227463,
+ "learning_rate": 0.0006,
+ "loss": 4.301420211791992,
+ "step": 2197
+ },
+ {
+ "epoch": 30.531004366812226,
+ "grad_norm": 0.015194724313914776,
+ "learning_rate": 0.0006,
+ "loss": 4.18703031539917,
+ "step": 2198
+ },
+ {
+ "epoch": 30.544978165938865,
+ "grad_norm": 0.015587746165692806,
+ "learning_rate": 0.0006,
+ "loss": 4.138928413391113,
+ "step": 2199
+ },
+ {
+ "epoch": 30.558951965065503,
+ "grad_norm": 0.016029570251703262,
+ "learning_rate": 0.0006,
+ "loss": 4.290863990783691,
+ "step": 2200
+ },
+ {
+ "epoch": 30.57292576419214,
+ "grad_norm": 0.015651309862732887,
+ "learning_rate": 0.0006,
+ "loss": 4.272062301635742,
+ "step": 2201
+ },
+ {
+ "epoch": 30.586899563318777,
+ "grad_norm": 0.015992306172847748,
+ "learning_rate": 0.0006,
+ "loss": 4.2487406730651855,
+ "step": 2202
+ },
+ {
+ "epoch": 30.600873362445416,
+ "grad_norm": 0.014969157986342907,
+ "learning_rate": 0.0006,
+ "loss": 4.115347862243652,
+ "step": 2203
+ },
+ {
+ "epoch": 30.61484716157205,
+ "grad_norm": 0.014639027416706085,
+ "learning_rate": 0.0006,
+ "loss": 4.2207489013671875,
+ "step": 2204
+ },
+ {
+ "epoch": 30.62882096069869,
+ "grad_norm": 0.015059413388371468,
+ "learning_rate": 0.0006,
+ "loss": 4.274044990539551,
+ "step": 2205
+ },
+ {
+ "epoch": 30.64279475982533,
+ "grad_norm": 0.0166641678661108,
+ "learning_rate": 0.0006,
+ "loss": 4.191373348236084,
+ "step": 2206
+ },
+ {
+ "epoch": 30.656768558951963,
+ "grad_norm": 0.017806345596909523,
+ "learning_rate": 0.0006,
+ "loss": 4.139800548553467,
+ "step": 2207
+ },
+ {
+ "epoch": 30.670742358078602,
+ "grad_norm": 0.018712421879172325,
+ "learning_rate": 0.0006,
+ "loss": 4.035965442657471,
+ "step": 2208
+ },
+ {
+ "epoch": 30.68471615720524,
+ "grad_norm": 0.018457243219017982,
+ "learning_rate": 0.0006,
+ "loss": 4.2579240798950195,
+ "step": 2209
+ },
+ {
+ "epoch": 30.69868995633188,
+ "grad_norm": 0.01920999586582184,
+ "learning_rate": 0.0006,
+ "loss": 4.131112098693848,
+ "step": 2210
+ },
+ {
+ "epoch": 30.712663755458514,
+ "grad_norm": 0.01729944534599781,
+ "learning_rate": 0.0006,
+ "loss": 4.254059314727783,
+ "step": 2211
+ },
+ {
+ "epoch": 30.726637554585153,
+ "grad_norm": 0.016914231702685356,
+ "learning_rate": 0.0006,
+ "loss": 4.233428955078125,
+ "step": 2212
+ },
+ {
+ "epoch": 30.74061135371179,
+ "grad_norm": 0.01913098618388176,
+ "learning_rate": 0.0006,
+ "loss": 4.23525333404541,
+ "step": 2213
+ },
+ {
+ "epoch": 30.754585152838427,
+ "grad_norm": 0.019150281324982643,
+ "learning_rate": 0.0006,
+ "loss": 4.213695049285889,
+ "step": 2214
+ },
+ {
+ "epoch": 30.768558951965066,
+ "grad_norm": 0.018947452306747437,
+ "learning_rate": 0.0006,
+ "loss": 4.151697158813477,
+ "step": 2215
+ },
+ {
+ "epoch": 30.782532751091704,
+ "grad_norm": 0.016058508306741714,
+ "learning_rate": 0.0006,
+ "loss": 4.220280170440674,
+ "step": 2216
+ },
+ {
+ "epoch": 30.79650655021834,
+ "grad_norm": 0.017488619312644005,
+ "learning_rate": 0.0006,
+ "loss": 4.154428482055664,
+ "step": 2217
+ },
+ {
+ "epoch": 30.810480349344978,
+ "grad_norm": 0.017355090007185936,
+ "learning_rate": 0.0006,
+ "loss": 4.256086349487305,
+ "step": 2218
+ },
+ {
+ "epoch": 30.824454148471617,
+ "grad_norm": 0.016636032611131668,
+ "learning_rate": 0.0006,
+ "loss": 4.227564334869385,
+ "step": 2219
+ },
+ {
+ "epoch": 30.83842794759825,
+ "grad_norm": 0.016273891553282738,
+ "learning_rate": 0.0006,
+ "loss": 4.200466156005859,
+ "step": 2220
+ },
+ {
+ "epoch": 30.85240174672489,
+ "grad_norm": 0.016939911991357803,
+ "learning_rate": 0.0006,
+ "loss": 4.1571221351623535,
+ "step": 2221
+ },
+ {
+ "epoch": 30.86637554585153,
+ "grad_norm": 0.018332522362470627,
+ "learning_rate": 0.0006,
+ "loss": 4.228398323059082,
+ "step": 2222
+ },
+ {
+ "epoch": 30.880349344978168,
+ "grad_norm": 0.019508114084601402,
+ "learning_rate": 0.0006,
+ "loss": 4.222960472106934,
+ "step": 2223
+ },
+ {
+ "epoch": 30.894323144104803,
+ "grad_norm": 0.018440278246998787,
+ "learning_rate": 0.0006,
+ "loss": 4.261051177978516,
+ "step": 2224
+ },
+ {
+ "epoch": 30.90829694323144,
+ "grad_norm": 0.018113229423761368,
+ "learning_rate": 0.0006,
+ "loss": 4.109235763549805,
+ "step": 2225
+ },
+ {
+ "epoch": 30.92227074235808,
+ "grad_norm": 0.015392083674669266,
+ "learning_rate": 0.0006,
+ "loss": 4.21950101852417,
+ "step": 2226
+ },
+ {
+ "epoch": 30.936244541484715,
+ "grad_norm": 0.01633290946483612,
+ "learning_rate": 0.0006,
+ "loss": 4.176053047180176,
+ "step": 2227
+ },
+ {
+ "epoch": 30.950218340611354,
+ "grad_norm": 0.016236383467912674,
+ "learning_rate": 0.0006,
+ "loss": 4.314309120178223,
+ "step": 2228
+ },
+ {
+ "epoch": 30.964192139737992,
+ "grad_norm": 0.015284373424947262,
+ "learning_rate": 0.0006,
+ "loss": 4.303133487701416,
+ "step": 2229
+ },
+ {
+ "epoch": 30.978165938864628,
+ "grad_norm": 0.015982897952198982,
+ "learning_rate": 0.0006,
+ "loss": 4.25739860534668,
+ "step": 2230
+ },
+ {
+ "epoch": 30.992139737991266,
+ "grad_norm": 0.016682198271155357,
+ "learning_rate": 0.0006,
+ "loss": 4.292210578918457,
+ "step": 2231
+ },
+ {
+ "epoch": 31.0,
+ "grad_norm": 0.017521467059850693,
+ "learning_rate": 0.0006,
+ "loss": 4.170637130737305,
+ "step": 2232
+ },
+ {
+ "epoch": 31.0,
+ "eval_loss": 4.642786979675293,
+ "eval_runtime": 57.029,
+ "eval_samples_per_second": 42.82,
+ "eval_steps_per_second": 1.35,
+ "step": 2232
+ },
+ {
+ "epoch": 31.01397379912664,
+ "grad_norm": 0.016607852652668953,
+ "learning_rate": 0.0006,
+ "loss": 4.106224536895752,
+ "step": 2233
+ },
+ {
+ "epoch": 31.027947598253274,
+ "grad_norm": 0.018448904156684875,
+ "learning_rate": 0.0006,
+ "loss": 4.224459648132324,
+ "step": 2234
+ },
+ {
+ "epoch": 31.041921397379912,
+ "grad_norm": 0.020481228828430176,
+ "learning_rate": 0.0006,
+ "loss": 4.305548667907715,
+ "step": 2235
+ },
+ {
+ "epoch": 31.05589519650655,
+ "grad_norm": 0.01999707892537117,
+ "learning_rate": 0.0006,
+ "loss": 4.253663539886475,
+ "step": 2236
+ },
+ {
+ "epoch": 31.069868995633186,
+ "grad_norm": 0.020428303629159927,
+ "learning_rate": 0.0006,
+ "loss": 4.166954040527344,
+ "step": 2237
+ },
+ {
+ "epoch": 31.083842794759825,
+ "grad_norm": 0.022608213126659393,
+ "learning_rate": 0.0006,
+ "loss": 4.172819137573242,
+ "step": 2238
+ },
+ {
+ "epoch": 31.097816593886463,
+ "grad_norm": 0.024669520556926727,
+ "learning_rate": 0.0006,
+ "loss": 4.106586456298828,
+ "step": 2239
+ },
+ {
+ "epoch": 31.111790393013102,
+ "grad_norm": 0.02247351035475731,
+ "learning_rate": 0.0006,
+ "loss": 4.160271644592285,
+ "step": 2240
+ },
+ {
+ "epoch": 31.125764192139737,
+ "grad_norm": 0.0207810141146183,
+ "learning_rate": 0.0006,
+ "loss": 4.272491931915283,
+ "step": 2241
+ },
+ {
+ "epoch": 31.139737991266376,
+ "grad_norm": 0.021576499566435814,
+ "learning_rate": 0.0006,
+ "loss": 4.171474456787109,
+ "step": 2242
+ },
+ {
+ "epoch": 31.153711790393015,
+ "grad_norm": 0.02337106689810753,
+ "learning_rate": 0.0006,
+ "loss": 4.094712734222412,
+ "step": 2243
+ },
+ {
+ "epoch": 31.16768558951965,
+ "grad_norm": 0.02219875156879425,
+ "learning_rate": 0.0006,
+ "loss": 4.1800994873046875,
+ "step": 2244
+ },
+ {
+ "epoch": 31.18165938864629,
+ "grad_norm": 0.022680504247546196,
+ "learning_rate": 0.0006,
+ "loss": 4.2054362297058105,
+ "step": 2245
+ },
+ {
+ "epoch": 31.195633187772927,
+ "grad_norm": 0.0208734143525362,
+ "learning_rate": 0.0006,
+ "loss": 4.06068229675293,
+ "step": 2246
+ },
+ {
+ "epoch": 31.209606986899562,
+ "grad_norm": 0.01811066083610058,
+ "learning_rate": 0.0006,
+ "loss": 4.2465128898620605,
+ "step": 2247
+ },
+ {
+ "epoch": 31.2235807860262,
+ "grad_norm": 0.018571637570858,
+ "learning_rate": 0.0006,
+ "loss": 4.191955089569092,
+ "step": 2248
+ },
+ {
+ "epoch": 31.23755458515284,
+ "grad_norm": 0.018598807975649834,
+ "learning_rate": 0.0006,
+ "loss": 4.200872421264648,
+ "step": 2249
+ },
+ {
+ "epoch": 31.251528384279474,
+ "grad_norm": 0.017577625811100006,
+ "learning_rate": 0.0006,
+ "loss": 4.148700714111328,
+ "step": 2250
+ },
+ {
+ "epoch": 31.265502183406113,
+ "grad_norm": 0.01817980408668518,
+ "learning_rate": 0.0006,
+ "loss": 4.256978988647461,
+ "step": 2251
+ },
+ {
+ "epoch": 31.27947598253275,
+ "grad_norm": 0.017983486875891685,
+ "learning_rate": 0.0006,
+ "loss": 4.175005912780762,
+ "step": 2252
+ },
+ {
+ "epoch": 31.29344978165939,
+ "grad_norm": 0.016751961782574654,
+ "learning_rate": 0.0006,
+ "loss": 4.237518310546875,
+ "step": 2253
+ },
+ {
+ "epoch": 31.307423580786025,
+ "grad_norm": 0.016407202929258347,
+ "learning_rate": 0.0006,
+ "loss": 4.197269439697266,
+ "step": 2254
+ },
+ {
+ "epoch": 31.321397379912664,
+ "grad_norm": 0.01578645594418049,
+ "learning_rate": 0.0006,
+ "loss": 4.179751396179199,
+ "step": 2255
+ },
+ {
+ "epoch": 31.335371179039303,
+ "grad_norm": 0.016001226380467415,
+ "learning_rate": 0.0006,
+ "loss": 4.08821964263916,
+ "step": 2256
+ },
+ {
+ "epoch": 31.349344978165938,
+ "grad_norm": 0.015122811309993267,
+ "learning_rate": 0.0006,
+ "loss": 4.047828674316406,
+ "step": 2257
+ },
+ {
+ "epoch": 31.363318777292577,
+ "grad_norm": 0.015399965457618237,
+ "learning_rate": 0.0006,
+ "loss": 4.1506524085998535,
+ "step": 2258
+ },
+ {
+ "epoch": 31.377292576419215,
+ "grad_norm": 0.01534605398774147,
+ "learning_rate": 0.0006,
+ "loss": 4.27952766418457,
+ "step": 2259
+ },
+ {
+ "epoch": 31.39126637554585,
+ "grad_norm": 0.015461381524801254,
+ "learning_rate": 0.0006,
+ "loss": 4.27102518081665,
+ "step": 2260
+ },
+ {
+ "epoch": 31.40524017467249,
+ "grad_norm": 0.016769153997302055,
+ "learning_rate": 0.0006,
+ "loss": 4.191592216491699,
+ "step": 2261
+ },
+ {
+ "epoch": 31.419213973799128,
+ "grad_norm": 0.016557959839701653,
+ "learning_rate": 0.0006,
+ "loss": 4.159705638885498,
+ "step": 2262
+ },
+ {
+ "epoch": 31.433187772925763,
+ "grad_norm": 0.01659569889307022,
+ "learning_rate": 0.0006,
+ "loss": 4.2136664390563965,
+ "step": 2263
+ },
+ {
+ "epoch": 31.4471615720524,
+ "grad_norm": 0.017101937904953957,
+ "learning_rate": 0.0006,
+ "loss": 4.2716498374938965,
+ "step": 2264
+ },
+ {
+ "epoch": 31.46113537117904,
+ "grad_norm": 0.01722985878586769,
+ "learning_rate": 0.0006,
+ "loss": 4.300534248352051,
+ "step": 2265
+ },
+ {
+ "epoch": 31.475109170305675,
+ "grad_norm": 0.016800185665488243,
+ "learning_rate": 0.0006,
+ "loss": 4.113240718841553,
+ "step": 2266
+ },
+ {
+ "epoch": 31.489082969432314,
+ "grad_norm": 0.01603875495493412,
+ "learning_rate": 0.0006,
+ "loss": 4.157105445861816,
+ "step": 2267
+ },
+ {
+ "epoch": 31.503056768558952,
+ "grad_norm": 0.015903932973742485,
+ "learning_rate": 0.0006,
+ "loss": 4.233669757843018,
+ "step": 2268
+ },
+ {
+ "epoch": 31.51703056768559,
+ "grad_norm": 0.0160983856767416,
+ "learning_rate": 0.0006,
+ "loss": 4.147704124450684,
+ "step": 2269
+ },
+ {
+ "epoch": 31.531004366812226,
+ "grad_norm": 0.01604730263352394,
+ "learning_rate": 0.0006,
+ "loss": 4.1465911865234375,
+ "step": 2270
+ },
+ {
+ "epoch": 31.544978165938865,
+ "grad_norm": 0.016205785796046257,
+ "learning_rate": 0.0006,
+ "loss": 4.106704235076904,
+ "step": 2271
+ },
+ {
+ "epoch": 31.558951965065503,
+ "grad_norm": 0.017008036375045776,
+ "learning_rate": 0.0006,
+ "loss": 4.191000938415527,
+ "step": 2272
+ },
+ {
+ "epoch": 31.57292576419214,
+ "grad_norm": 0.01699635200202465,
+ "learning_rate": 0.0006,
+ "loss": 4.187042236328125,
+ "step": 2273
+ },
+ {
+ "epoch": 31.586899563318777,
+ "grad_norm": 0.01692967116832733,
+ "learning_rate": 0.0006,
+ "loss": 4.208805561065674,
+ "step": 2274
+ },
+ {
+ "epoch": 31.600873362445416,
+ "grad_norm": 0.017608756199479103,
+ "learning_rate": 0.0006,
+ "loss": 4.200512886047363,
+ "step": 2275
+ },
+ {
+ "epoch": 31.61484716157205,
+ "grad_norm": 0.016590990126132965,
+ "learning_rate": 0.0006,
+ "loss": 4.194705009460449,
+ "step": 2276
+ },
+ {
+ "epoch": 31.62882096069869,
+ "grad_norm": 0.0170235987752676,
+ "learning_rate": 0.0006,
+ "loss": 4.210326194763184,
+ "step": 2277
+ },
+ {
+ "epoch": 31.64279475982533,
+ "grad_norm": 0.01604801043868065,
+ "learning_rate": 0.0006,
+ "loss": 4.196024417877197,
+ "step": 2278
+ },
+ {
+ "epoch": 31.656768558951963,
+ "grad_norm": 0.015694202855229378,
+ "learning_rate": 0.0006,
+ "loss": 4.178236484527588,
+ "step": 2279
+ },
+ {
+ "epoch": 31.670742358078602,
+ "grad_norm": 0.015473959036171436,
+ "learning_rate": 0.0006,
+ "loss": 4.259339809417725,
+ "step": 2280
+ },
+ {
+ "epoch": 31.68471615720524,
+ "grad_norm": 0.014689362607896328,
+ "learning_rate": 0.0006,
+ "loss": 3.9702324867248535,
+ "step": 2281
+ },
+ {
+ "epoch": 31.69868995633188,
+ "grad_norm": 0.015245123766362667,
+ "learning_rate": 0.0006,
+ "loss": 4.197025775909424,
+ "step": 2282
+ },
+ {
+ "epoch": 31.712663755458514,
+ "grad_norm": 0.016148347407579422,
+ "learning_rate": 0.0006,
+ "loss": 4.231856346130371,
+ "step": 2283
+ },
+ {
+ "epoch": 31.726637554585153,
+ "grad_norm": 0.015169818885624409,
+ "learning_rate": 0.0006,
+ "loss": 4.230169296264648,
+ "step": 2284
+ },
+ {
+ "epoch": 31.74061135371179,
+ "grad_norm": 0.014843013137578964,
+ "learning_rate": 0.0006,
+ "loss": 4.195633411407471,
+ "step": 2285
+ },
+ {
+ "epoch": 31.754585152838427,
+ "grad_norm": 0.015229340642690659,
+ "learning_rate": 0.0006,
+ "loss": 4.231280326843262,
+ "step": 2286
+ },
+ {
+ "epoch": 31.768558951965066,
+ "grad_norm": 0.01551708485931158,
+ "learning_rate": 0.0006,
+ "loss": 4.17064094543457,
+ "step": 2287
+ },
+ {
+ "epoch": 31.782532751091704,
+ "grad_norm": 0.015863575041294098,
+ "learning_rate": 0.0006,
+ "loss": 4.1579484939575195,
+ "step": 2288
+ },
+ {
+ "epoch": 31.79650655021834,
+ "grad_norm": 0.015118198469281197,
+ "learning_rate": 0.0006,
+ "loss": 4.1626787185668945,
+ "step": 2289
+ },
+ {
+ "epoch": 31.810480349344978,
+ "grad_norm": 0.015220776200294495,
+ "learning_rate": 0.0006,
+ "loss": 4.135746955871582,
+ "step": 2290
+ },
+ {
+ "epoch": 31.824454148471617,
+ "grad_norm": 0.014529784210026264,
+ "learning_rate": 0.0006,
+ "loss": 4.164931297302246,
+ "step": 2291
+ },
+ {
+ "epoch": 31.83842794759825,
+ "grad_norm": 0.015533296391367912,
+ "learning_rate": 0.0006,
+ "loss": 4.112464904785156,
+ "step": 2292
+ },
+ {
+ "epoch": 31.85240174672489,
+ "grad_norm": 0.016562335193157196,
+ "learning_rate": 0.0006,
+ "loss": 4.164898872375488,
+ "step": 2293
+ },
+ {
+ "epoch": 31.86637554585153,
+ "grad_norm": 0.016258113086223602,
+ "learning_rate": 0.0006,
+ "loss": 4.1979265213012695,
+ "step": 2294
+ },
+ {
+ "epoch": 31.880349344978168,
+ "grad_norm": 0.01558110024780035,
+ "learning_rate": 0.0006,
+ "loss": 4.174675941467285,
+ "step": 2295
+ },
+ {
+ "epoch": 31.894323144104803,
+ "grad_norm": 0.016412515193223953,
+ "learning_rate": 0.0006,
+ "loss": 4.217909336090088,
+ "step": 2296
+ },
+ {
+ "epoch": 31.90829694323144,
+ "grad_norm": 0.015118442475795746,
+ "learning_rate": 0.0006,
+ "loss": 4.137143611907959,
+ "step": 2297
+ },
+ {
+ "epoch": 31.92227074235808,
+ "grad_norm": 0.01506161317229271,
+ "learning_rate": 0.0006,
+ "loss": 4.196122169494629,
+ "step": 2298
+ },
+ {
+ "epoch": 31.936244541484715,
+ "grad_norm": 0.017165271565318108,
+ "learning_rate": 0.0006,
+ "loss": 4.14834451675415,
+ "step": 2299
+ },
+ {
+ "epoch": 31.950218340611354,
+ "grad_norm": 0.01696440577507019,
+ "learning_rate": 0.0006,
+ "loss": 4.218868732452393,
+ "step": 2300
+ },
+ {
+ "epoch": 31.964192139737992,
+ "grad_norm": 0.017349394038319588,
+ "learning_rate": 0.0006,
+ "loss": 4.2754011154174805,
+ "step": 2301
+ },
+ {
+ "epoch": 31.978165938864628,
+ "grad_norm": 0.018387921154499054,
+ "learning_rate": 0.0006,
+ "loss": 4.282623291015625,
+ "step": 2302
+ },
+ {
+ "epoch": 31.992139737991266,
+ "grad_norm": 0.018837926909327507,
+ "learning_rate": 0.0006,
+ "loss": 4.185425281524658,
+ "step": 2303
+ },
+ {
+ "epoch": 32.0,
+ "grad_norm": 0.018998464569449425,
+ "learning_rate": 0.0006,
+ "loss": 4.151291847229004,
+ "step": 2304
+ },
+ {
+ "epoch": 32.0,
+ "eval_loss": 4.522614002227783,
+ "eval_runtime": 57.0106,
+ "eval_samples_per_second": 42.834,
+ "eval_steps_per_second": 1.351,
+ "step": 2304
+ },
+ {
+ "epoch": 32.01397379912664,
+ "grad_norm": 0.019036833196878433,
+ "learning_rate": 0.0006,
+ "loss": 4.183202266693115,
+ "step": 2305
+ },
+ {
+ "epoch": 32.02794759825328,
+ "grad_norm": 0.02207397110760212,
+ "learning_rate": 0.0006,
+ "loss": 4.17875862121582,
+ "step": 2306
+ },
+ {
+ "epoch": 32.041921397379916,
+ "grad_norm": 0.022008279338479042,
+ "learning_rate": 0.0006,
+ "loss": 4.08042049407959,
+ "step": 2307
+ },
+ {
+ "epoch": 32.05589519650655,
+ "grad_norm": 0.022407017648220062,
+ "learning_rate": 0.0006,
+ "loss": 4.112189769744873,
+ "step": 2308
+ },
+ {
+ "epoch": 32.069868995633186,
+ "grad_norm": 0.021599747240543365,
+ "learning_rate": 0.0006,
+ "loss": 4.126455307006836,
+ "step": 2309
+ },
+ {
+ "epoch": 32.083842794759825,
+ "grad_norm": 0.020340966060757637,
+ "learning_rate": 0.0006,
+ "loss": 4.080538272857666,
+ "step": 2310
+ },
+ {
+ "epoch": 32.09781659388646,
+ "grad_norm": 0.020956892520189285,
+ "learning_rate": 0.0006,
+ "loss": 4.171943664550781,
+ "step": 2311
+ },
+ {
+ "epoch": 32.1117903930131,
+ "grad_norm": 0.019966667518019676,
+ "learning_rate": 0.0006,
+ "loss": 4.144007682800293,
+ "step": 2312
+ },
+ {
+ "epoch": 32.12576419213974,
+ "grad_norm": 0.018256952986121178,
+ "learning_rate": 0.0006,
+ "loss": 4.208107948303223,
+ "step": 2313
+ },
+ {
+ "epoch": 32.13973799126637,
+ "grad_norm": 0.017289170995354652,
+ "learning_rate": 0.0006,
+ "loss": 4.182024955749512,
+ "step": 2314
+ },
+ {
+ "epoch": 32.15371179039301,
+ "grad_norm": 0.017656700685620308,
+ "learning_rate": 0.0006,
+ "loss": 4.068373680114746,
+ "step": 2315
+ },
+ {
+ "epoch": 32.16768558951965,
+ "grad_norm": 0.017903361469507217,
+ "learning_rate": 0.0006,
+ "loss": 4.2345428466796875,
+ "step": 2316
+ },
+ {
+ "epoch": 32.18165938864629,
+ "grad_norm": 0.01785109005868435,
+ "learning_rate": 0.0006,
+ "loss": 4.219776153564453,
+ "step": 2317
+ },
+ {
+ "epoch": 32.19563318777293,
+ "grad_norm": 0.01887257769703865,
+ "learning_rate": 0.0006,
+ "loss": 4.139206409454346,
+ "step": 2318
+ },
+ {
+ "epoch": 32.209606986899566,
+ "grad_norm": 0.01924685761332512,
+ "learning_rate": 0.0006,
+ "loss": 4.121052265167236,
+ "step": 2319
+ },
+ {
+ "epoch": 32.223580786026204,
+ "grad_norm": 0.020574072375893593,
+ "learning_rate": 0.0006,
+ "loss": 4.13262414932251,
+ "step": 2320
+ },
+ {
+ "epoch": 32.237554585152836,
+ "grad_norm": 0.023849064484238625,
+ "learning_rate": 0.0006,
+ "loss": 4.239119529724121,
+ "step": 2321
+ },
+ {
+ "epoch": 32.251528384279474,
+ "grad_norm": 0.026239361613988876,
+ "learning_rate": 0.0006,
+ "loss": 4.248993873596191,
+ "step": 2322
+ },
+ {
+ "epoch": 32.26550218340611,
+ "grad_norm": 0.023980243131518364,
+ "learning_rate": 0.0006,
+ "loss": 4.159852504730225,
+ "step": 2323
+ },
+ {
+ "epoch": 32.27947598253275,
+ "grad_norm": 0.020442405715584755,
+ "learning_rate": 0.0006,
+ "loss": 4.133546829223633,
+ "step": 2324
+ },
+ {
+ "epoch": 32.29344978165939,
+ "grad_norm": 0.020661186426877975,
+ "learning_rate": 0.0006,
+ "loss": 4.078243732452393,
+ "step": 2325
+ },
+ {
+ "epoch": 32.30742358078603,
+ "grad_norm": 0.02046387456357479,
+ "learning_rate": 0.0006,
+ "loss": 4.172012805938721,
+ "step": 2326
+ },
+ {
+ "epoch": 32.32139737991266,
+ "grad_norm": 0.02089597098529339,
+ "learning_rate": 0.0006,
+ "loss": 4.0604705810546875,
+ "step": 2327
+ },
+ {
+ "epoch": 32.3353711790393,
+ "grad_norm": 0.01875591278076172,
+ "learning_rate": 0.0006,
+ "loss": 4.125911235809326,
+ "step": 2328
+ },
+ {
+ "epoch": 32.34934497816594,
+ "grad_norm": 0.018606796860694885,
+ "learning_rate": 0.0006,
+ "loss": 4.097362518310547,
+ "step": 2329
+ },
+ {
+ "epoch": 32.36331877729258,
+ "grad_norm": 0.01766626164317131,
+ "learning_rate": 0.0006,
+ "loss": 4.118616580963135,
+ "step": 2330
+ },
+ {
+ "epoch": 32.377292576419215,
+ "grad_norm": 0.017692960798740387,
+ "learning_rate": 0.0006,
+ "loss": 4.1457672119140625,
+ "step": 2331
+ },
+ {
+ "epoch": 32.391266375545854,
+ "grad_norm": 0.017619704827666283,
+ "learning_rate": 0.0006,
+ "loss": 4.175961494445801,
+ "step": 2332
+ },
+ {
+ "epoch": 32.40524017467249,
+ "grad_norm": 0.01784413494169712,
+ "learning_rate": 0.0006,
+ "loss": 4.051291465759277,
+ "step": 2333
+ },
+ {
+ "epoch": 32.419213973799124,
+ "grad_norm": 0.018508171662688255,
+ "learning_rate": 0.0006,
+ "loss": 4.0638556480407715,
+ "step": 2334
+ },
+ {
+ "epoch": 32.43318777292576,
+ "grad_norm": 0.01731436885893345,
+ "learning_rate": 0.0006,
+ "loss": 4.1805419921875,
+ "step": 2335
+ },
+ {
+ "epoch": 32.4471615720524,
+ "grad_norm": 0.015886498615145683,
+ "learning_rate": 0.0006,
+ "loss": 4.151026725769043,
+ "step": 2336
+ },
+ {
+ "epoch": 32.46113537117904,
+ "grad_norm": 0.016583820804953575,
+ "learning_rate": 0.0006,
+ "loss": 4.203604698181152,
+ "step": 2337
+ },
+ {
+ "epoch": 32.47510917030568,
+ "grad_norm": 0.0176097322255373,
+ "learning_rate": 0.0006,
+ "loss": 4.135159015655518,
+ "step": 2338
+ },
+ {
+ "epoch": 32.48908296943232,
+ "grad_norm": 0.016657903790473938,
+ "learning_rate": 0.0006,
+ "loss": 4.126681804656982,
+ "step": 2339
+ },
+ {
+ "epoch": 32.50305676855895,
+ "grad_norm": 0.016866130754351616,
+ "learning_rate": 0.0006,
+ "loss": 4.162341117858887,
+ "step": 2340
+ },
+ {
+ "epoch": 32.51703056768559,
+ "grad_norm": 0.016286680474877357,
+ "learning_rate": 0.0006,
+ "loss": 4.261801242828369,
+ "step": 2341
+ },
+ {
+ "epoch": 32.531004366812226,
+ "grad_norm": 0.01481365505605936,
+ "learning_rate": 0.0006,
+ "loss": 4.110016822814941,
+ "step": 2342
+ },
+ {
+ "epoch": 32.544978165938865,
+ "grad_norm": 0.016605013981461525,
+ "learning_rate": 0.0006,
+ "loss": 4.230024814605713,
+ "step": 2343
+ },
+ {
+ "epoch": 32.5589519650655,
+ "grad_norm": 0.01645175740122795,
+ "learning_rate": 0.0006,
+ "loss": 4.174612045288086,
+ "step": 2344
+ },
+ {
+ "epoch": 32.57292576419214,
+ "grad_norm": 0.014725026674568653,
+ "learning_rate": 0.0006,
+ "loss": 4.066371440887451,
+ "step": 2345
+ },
+ {
+ "epoch": 32.58689956331878,
+ "grad_norm": 0.015003837645053864,
+ "learning_rate": 0.0006,
+ "loss": 4.193343639373779,
+ "step": 2346
+ },
+ {
+ "epoch": 32.60087336244541,
+ "grad_norm": 0.015346302650868893,
+ "learning_rate": 0.0006,
+ "loss": 4.210659980773926,
+ "step": 2347
+ },
+ {
+ "epoch": 32.61484716157205,
+ "grad_norm": 0.015135962516069412,
+ "learning_rate": 0.0006,
+ "loss": 4.204618453979492,
+ "step": 2348
+ },
+ {
+ "epoch": 32.62882096069869,
+ "grad_norm": 0.01601956970989704,
+ "learning_rate": 0.0006,
+ "loss": 4.145564556121826,
+ "step": 2349
+ },
+ {
+ "epoch": 32.64279475982533,
+ "grad_norm": 0.01517851185053587,
+ "learning_rate": 0.0006,
+ "loss": 4.051673889160156,
+ "step": 2350
+ },
+ {
+ "epoch": 32.65676855895197,
+ "grad_norm": 0.014894292689859867,
+ "learning_rate": 0.0006,
+ "loss": 4.184818267822266,
+ "step": 2351
+ },
+ {
+ "epoch": 32.670742358078606,
+ "grad_norm": 0.015720834955573082,
+ "learning_rate": 0.0006,
+ "loss": 4.220036506652832,
+ "step": 2352
+ },
+ {
+ "epoch": 32.68471615720524,
+ "grad_norm": 0.01573433168232441,
+ "learning_rate": 0.0006,
+ "loss": 4.076773166656494,
+ "step": 2353
+ },
+ {
+ "epoch": 32.698689956331876,
+ "grad_norm": 0.015828052535653114,
+ "learning_rate": 0.0006,
+ "loss": 4.116329669952393,
+ "step": 2354
+ },
+ {
+ "epoch": 32.712663755458514,
+ "grad_norm": 0.015462259761989117,
+ "learning_rate": 0.0006,
+ "loss": 4.1155242919921875,
+ "step": 2355
+ },
+ {
+ "epoch": 32.72663755458515,
+ "grad_norm": 0.015763849020004272,
+ "learning_rate": 0.0006,
+ "loss": 4.169687747955322,
+ "step": 2356
+ },
+ {
+ "epoch": 32.74061135371179,
+ "grad_norm": 0.014935145154595375,
+ "learning_rate": 0.0006,
+ "loss": 4.2090911865234375,
+ "step": 2357
+ },
+ {
+ "epoch": 32.75458515283843,
+ "grad_norm": 0.015119184739887714,
+ "learning_rate": 0.0006,
+ "loss": 4.131838798522949,
+ "step": 2358
+ },
+ {
+ "epoch": 32.76855895196506,
+ "grad_norm": 0.014833525754511356,
+ "learning_rate": 0.0006,
+ "loss": 4.17618465423584,
+ "step": 2359
+ },
+ {
+ "epoch": 32.7825327510917,
+ "grad_norm": 0.01588359661400318,
+ "learning_rate": 0.0006,
+ "loss": 4.163723945617676,
+ "step": 2360
+ },
+ {
+ "epoch": 32.79650655021834,
+ "grad_norm": 0.0168188214302063,
+ "learning_rate": 0.0006,
+ "loss": 4.178189277648926,
+ "step": 2361
+ },
+ {
+ "epoch": 32.81048034934498,
+ "grad_norm": 0.01726095750927925,
+ "learning_rate": 0.0006,
+ "loss": 4.144223690032959,
+ "step": 2362
+ },
+ {
+ "epoch": 32.82445414847162,
+ "grad_norm": 0.01701410301029682,
+ "learning_rate": 0.0006,
+ "loss": 4.209896087646484,
+ "step": 2363
+ },
+ {
+ "epoch": 32.838427947598255,
+ "grad_norm": 0.01682264916598797,
+ "learning_rate": 0.0006,
+ "loss": 4.26223087310791,
+ "step": 2364
+ },
+ {
+ "epoch": 32.852401746724894,
+ "grad_norm": 0.015597456134855747,
+ "learning_rate": 0.0006,
+ "loss": 4.292664051055908,
+ "step": 2365
+ },
+ {
+ "epoch": 32.866375545851525,
+ "grad_norm": 0.015258579514920712,
+ "learning_rate": 0.0006,
+ "loss": 4.165006637573242,
+ "step": 2366
+ },
+ {
+ "epoch": 32.880349344978164,
+ "grad_norm": 0.01662200503051281,
+ "learning_rate": 0.0006,
+ "loss": 4.309786319732666,
+ "step": 2367
+ },
+ {
+ "epoch": 32.8943231441048,
+ "grad_norm": 0.015691161155700684,
+ "learning_rate": 0.0006,
+ "loss": 4.167202949523926,
+ "step": 2368
+ },
+ {
+ "epoch": 32.90829694323144,
+ "grad_norm": 0.014870637096464634,
+ "learning_rate": 0.0006,
+ "loss": 4.142415523529053,
+ "step": 2369
+ },
+ {
+ "epoch": 32.92227074235808,
+ "grad_norm": 0.01567252166569233,
+ "learning_rate": 0.0006,
+ "loss": 4.2312092781066895,
+ "step": 2370
+ },
+ {
+ "epoch": 32.93624454148472,
+ "grad_norm": 0.014948616735637188,
+ "learning_rate": 0.0006,
+ "loss": 4.1514129638671875,
+ "step": 2371
+ },
+ {
+ "epoch": 32.95021834061135,
+ "grad_norm": 0.014654500409960747,
+ "learning_rate": 0.0006,
+ "loss": 4.205010890960693,
+ "step": 2372
+ },
+ {
+ "epoch": 32.96419213973799,
+ "grad_norm": 0.015059148892760277,
+ "learning_rate": 0.0006,
+ "loss": 4.148405075073242,
+ "step": 2373
+ },
+ {
+ "epoch": 32.97816593886463,
+ "grad_norm": 0.015026576817035675,
+ "learning_rate": 0.0006,
+ "loss": 4.182535171508789,
+ "step": 2374
+ },
+ {
+ "epoch": 32.992139737991266,
+ "grad_norm": 0.016394320875406265,
+ "learning_rate": 0.0006,
+ "loss": 4.129051208496094,
+ "step": 2375
+ },
+ {
+ "epoch": 33.0,
+ "grad_norm": 0.01867043413221836,
+ "learning_rate": 0.0006,
+ "loss": 4.185556411743164,
+ "step": 2376
+ },
+ {
+ "epoch": 33.0,
+ "eval_loss": 4.640405654907227,
+ "eval_runtime": 56.5861,
+ "eval_samples_per_second": 43.155,
+ "eval_steps_per_second": 1.361,
+ "step": 2376
+ },
+ {
+ "epoch": 33.01397379912664,
+ "grad_norm": 0.018626727163791656,
+ "learning_rate": 0.0006,
+ "loss": 4.143031597137451,
+ "step": 2377
+ },
+ {
+ "epoch": 33.02794759825328,
+ "grad_norm": 0.019283363595604897,
+ "learning_rate": 0.0006,
+ "loss": 4.117845058441162,
+ "step": 2378
+ },
+ {
+ "epoch": 33.041921397379916,
+ "grad_norm": 0.018185822293162346,
+ "learning_rate": 0.0006,
+ "loss": 4.076862812042236,
+ "step": 2379
+ },
+ {
+ "epoch": 33.05589519650655,
+ "grad_norm": 0.019411196932196617,
+ "learning_rate": 0.0006,
+ "loss": 4.124156475067139,
+ "step": 2380
+ },
+ {
+ "epoch": 33.069868995633186,
+ "grad_norm": 0.019248466938734055,
+ "learning_rate": 0.0006,
+ "loss": 4.201735019683838,
+ "step": 2381
+ },
+ {
+ "epoch": 33.083842794759825,
+ "grad_norm": 0.018856720998883247,
+ "learning_rate": 0.0006,
+ "loss": 4.00296688079834,
+ "step": 2382
+ },
+ {
+ "epoch": 33.09781659388646,
+ "grad_norm": 0.018554294481873512,
+ "learning_rate": 0.0006,
+ "loss": 4.270577430725098,
+ "step": 2383
+ },
+ {
+ "epoch": 33.1117903930131,
+ "grad_norm": 0.019946224987506866,
+ "learning_rate": 0.0006,
+ "loss": 4.119589328765869,
+ "step": 2384
+ },
+ {
+ "epoch": 33.12576419213974,
+ "grad_norm": 0.021395670250058174,
+ "learning_rate": 0.0006,
+ "loss": 4.128847122192383,
+ "step": 2385
+ },
+ {
+ "epoch": 33.13973799126637,
+ "grad_norm": 0.02009674906730652,
+ "learning_rate": 0.0006,
+ "loss": 4.183228969573975,
+ "step": 2386
+ },
+ {
+ "epoch": 33.15371179039301,
+ "grad_norm": 0.01820223033428192,
+ "learning_rate": 0.0006,
+ "loss": 4.131601810455322,
+ "step": 2387
+ },
+ {
+ "epoch": 33.16768558951965,
+ "grad_norm": 0.018653156235814095,
+ "learning_rate": 0.0006,
+ "loss": 4.116430282592773,
+ "step": 2388
+ },
+ {
+ "epoch": 33.18165938864629,
+ "grad_norm": 0.016828063875436783,
+ "learning_rate": 0.0006,
+ "loss": 4.054663181304932,
+ "step": 2389
+ },
+ {
+ "epoch": 33.19563318777293,
+ "grad_norm": 0.017025409266352654,
+ "learning_rate": 0.0006,
+ "loss": 4.110044956207275,
+ "step": 2390
+ },
+ {
+ "epoch": 33.209606986899566,
+ "grad_norm": 0.016608452424407005,
+ "learning_rate": 0.0006,
+ "loss": 4.102668285369873,
+ "step": 2391
+ },
+ {
+ "epoch": 33.223580786026204,
+ "grad_norm": 0.01743577979505062,
+ "learning_rate": 0.0006,
+ "loss": 4.166791915893555,
+ "step": 2392
+ },
+ {
+ "epoch": 33.237554585152836,
+ "grad_norm": 0.017253734171390533,
+ "learning_rate": 0.0006,
+ "loss": 4.217832088470459,
+ "step": 2393
+ },
+ {
+ "epoch": 33.251528384279474,
+ "grad_norm": 0.017937595024704933,
+ "learning_rate": 0.0006,
+ "loss": 4.2173967361450195,
+ "step": 2394
+ },
+ {
+ "epoch": 33.26550218340611,
+ "grad_norm": 0.01985993981361389,
+ "learning_rate": 0.0006,
+ "loss": 4.17849588394165,
+ "step": 2395
+ },
+ {
+ "epoch": 33.27947598253275,
+ "grad_norm": 0.019243355840444565,
+ "learning_rate": 0.0006,
+ "loss": 4.086523056030273,
+ "step": 2396
+ },
+ {
+ "epoch": 33.29344978165939,
+ "grad_norm": 0.01924416422843933,
+ "learning_rate": 0.0006,
+ "loss": 4.128574371337891,
+ "step": 2397
+ },
+ {
+ "epoch": 33.30742358078603,
+ "grad_norm": 0.01983758620917797,
+ "learning_rate": 0.0006,
+ "loss": 4.061581611633301,
+ "step": 2398
+ },
+ {
+ "epoch": 33.32139737991266,
+ "grad_norm": 0.01930462196469307,
+ "learning_rate": 0.0006,
+ "loss": 4.091731071472168,
+ "step": 2399
+ },
+ {
+ "epoch": 33.3353711790393,
+ "grad_norm": 0.017988424748182297,
+ "learning_rate": 0.0006,
+ "loss": 4.155599117279053,
+ "step": 2400
+ },
+ {
+ "epoch": 33.34934497816594,
+ "grad_norm": 0.015194767154753208,
+ "learning_rate": 0.0006,
+ "loss": 4.144134044647217,
+ "step": 2401
+ },
+ {
+ "epoch": 33.36331877729258,
+ "grad_norm": 0.01536081824451685,
+ "learning_rate": 0.0006,
+ "loss": 4.127272129058838,
+ "step": 2402
+ },
+ {
+ "epoch": 33.377292576419215,
+ "grad_norm": 0.015431297942996025,
+ "learning_rate": 0.0006,
+ "loss": 4.3213019371032715,
+ "step": 2403
+ },
+ {
+ "epoch": 33.391266375545854,
+ "grad_norm": 0.015740280970931053,
+ "learning_rate": 0.0006,
+ "loss": 4.121467590332031,
+ "step": 2404
+ },
+ {
+ "epoch": 33.40524017467249,
+ "grad_norm": 0.015379384160041809,
+ "learning_rate": 0.0006,
+ "loss": 4.095216751098633,
+ "step": 2405
+ },
+ {
+ "epoch": 33.419213973799124,
+ "grad_norm": 0.01528073474764824,
+ "learning_rate": 0.0006,
+ "loss": 4.13047981262207,
+ "step": 2406
+ },
+ {
+ "epoch": 33.43318777292576,
+ "grad_norm": 0.015739979222416878,
+ "learning_rate": 0.0006,
+ "loss": 4.2533135414123535,
+ "step": 2407
+ },
+ {
+ "epoch": 33.4471615720524,
+ "grad_norm": 0.01519764307886362,
+ "learning_rate": 0.0006,
+ "loss": 4.131782054901123,
+ "step": 2408
+ },
+ {
+ "epoch": 33.46113537117904,
+ "grad_norm": 0.015425996854901314,
+ "learning_rate": 0.0006,
+ "loss": 4.02811336517334,
+ "step": 2409
+ },
+ {
+ "epoch": 33.47510917030568,
+ "grad_norm": 0.015378830023109913,
+ "learning_rate": 0.0006,
+ "loss": 4.08652925491333,
+ "step": 2410
+ },
+ {
+ "epoch": 33.48908296943232,
+ "grad_norm": 0.01473075058311224,
+ "learning_rate": 0.0006,
+ "loss": 4.214364528656006,
+ "step": 2411
+ },
+ {
+ "epoch": 33.50305676855895,
+ "grad_norm": 0.015141637064516544,
+ "learning_rate": 0.0006,
+ "loss": 4.155712127685547,
+ "step": 2412
+ },
+ {
+ "epoch": 33.51703056768559,
+ "grad_norm": 0.01641593500971794,
+ "learning_rate": 0.0006,
+ "loss": 4.097907066345215,
+ "step": 2413
+ },
+ {
+ "epoch": 33.531004366812226,
+ "grad_norm": 0.015182029455900192,
+ "learning_rate": 0.0006,
+ "loss": 4.238499641418457,
+ "step": 2414
+ },
+ {
+ "epoch": 33.544978165938865,
+ "grad_norm": 0.01643732376396656,
+ "learning_rate": 0.0006,
+ "loss": 4.0661492347717285,
+ "step": 2415
+ },
+ {
+ "epoch": 33.5589519650655,
+ "grad_norm": 0.015614228323101997,
+ "learning_rate": 0.0006,
+ "loss": 4.120795249938965,
+ "step": 2416
+ },
+ {
+ "epoch": 33.57292576419214,
+ "grad_norm": 0.016066187992691994,
+ "learning_rate": 0.0006,
+ "loss": 4.096465110778809,
+ "step": 2417
+ },
+ {
+ "epoch": 33.58689956331878,
+ "grad_norm": 0.017214342951774597,
+ "learning_rate": 0.0006,
+ "loss": 4.242628574371338,
+ "step": 2418
+ },
+ {
+ "epoch": 33.60087336244541,
+ "grad_norm": 0.01787586510181427,
+ "learning_rate": 0.0006,
+ "loss": 4.149589538574219,
+ "step": 2419
+ },
+ {
+ "epoch": 33.61484716157205,
+ "grad_norm": 0.019455241039395332,
+ "learning_rate": 0.0006,
+ "loss": 4.267796039581299,
+ "step": 2420
+ },
+ {
+ "epoch": 33.62882096069869,
+ "grad_norm": 0.01984674669802189,
+ "learning_rate": 0.0006,
+ "loss": 4.175220966339111,
+ "step": 2421
+ },
+ {
+ "epoch": 33.64279475982533,
+ "grad_norm": 0.021933717653155327,
+ "learning_rate": 0.0006,
+ "loss": 4.106647491455078,
+ "step": 2422
+ },
+ {
+ "epoch": 33.65676855895197,
+ "grad_norm": 0.02123589813709259,
+ "learning_rate": 0.0006,
+ "loss": 4.204996109008789,
+ "step": 2423
+ },
+ {
+ "epoch": 33.670742358078606,
+ "grad_norm": 0.019814589992165565,
+ "learning_rate": 0.0006,
+ "loss": 4.186171531677246,
+ "step": 2424
+ },
+ {
+ "epoch": 33.68471615720524,
+ "grad_norm": 0.02059728465974331,
+ "learning_rate": 0.0006,
+ "loss": 4.170509338378906,
+ "step": 2425
+ },
+ {
+ "epoch": 33.698689956331876,
+ "grad_norm": 0.018483608961105347,
+ "learning_rate": 0.0006,
+ "loss": 4.085585594177246,
+ "step": 2426
+ },
+ {
+ "epoch": 33.712663755458514,
+ "grad_norm": 0.018403852358460426,
+ "learning_rate": 0.0006,
+ "loss": 4.2551774978637695,
+ "step": 2427
+ },
+ {
+ "epoch": 33.72663755458515,
+ "grad_norm": 0.018653830513358116,
+ "learning_rate": 0.0006,
+ "loss": 4.091766357421875,
+ "step": 2428
+ },
+ {
+ "epoch": 33.74061135371179,
+ "grad_norm": 0.017692334949970245,
+ "learning_rate": 0.0006,
+ "loss": 4.100838661193848,
+ "step": 2429
+ },
+ {
+ "epoch": 33.75458515283843,
+ "grad_norm": 0.016029126942157745,
+ "learning_rate": 0.0006,
+ "loss": 4.1530680656433105,
+ "step": 2430
+ },
+ {
+ "epoch": 33.76855895196506,
+ "grad_norm": 0.016095684841275215,
+ "learning_rate": 0.0006,
+ "loss": 4.20544958114624,
+ "step": 2431
+ },
+ {
+ "epoch": 33.7825327510917,
+ "grad_norm": 0.016375798732042313,
+ "learning_rate": 0.0006,
+ "loss": 4.199177265167236,
+ "step": 2432
+ },
+ {
+ "epoch": 33.79650655021834,
+ "grad_norm": 0.01590786688029766,
+ "learning_rate": 0.0006,
+ "loss": 4.200799465179443,
+ "step": 2433
+ },
+ {
+ "epoch": 33.81048034934498,
+ "grad_norm": 0.014896097593009472,
+ "learning_rate": 0.0006,
+ "loss": 4.185850143432617,
+ "step": 2434
+ },
+ {
+ "epoch": 33.82445414847162,
+ "grad_norm": 0.01438513956964016,
+ "learning_rate": 0.0006,
+ "loss": 4.214962005615234,
+ "step": 2435
+ },
+ {
+ "epoch": 33.838427947598255,
+ "grad_norm": 0.014459339901804924,
+ "learning_rate": 0.0006,
+ "loss": 4.196972370147705,
+ "step": 2436
+ },
+ {
+ "epoch": 33.852401746724894,
+ "grad_norm": 0.014782169833779335,
+ "learning_rate": 0.0006,
+ "loss": 4.0885910987854,
+ "step": 2437
+ },
+ {
+ "epoch": 33.866375545851525,
+ "grad_norm": 0.01520370040088892,
+ "learning_rate": 0.0006,
+ "loss": 4.0323262214660645,
+ "step": 2438
+ },
+ {
+ "epoch": 33.880349344978164,
+ "grad_norm": 0.014056282117962837,
+ "learning_rate": 0.0006,
+ "loss": 4.1386518478393555,
+ "step": 2439
+ },
+ {
+ "epoch": 33.8943231441048,
+ "grad_norm": 0.013567990623414516,
+ "learning_rate": 0.0006,
+ "loss": 4.095754623413086,
+ "step": 2440
+ },
+ {
+ "epoch": 33.90829694323144,
+ "grad_norm": 0.013306716457009315,
+ "learning_rate": 0.0006,
+ "loss": 4.1623077392578125,
+ "step": 2441
+ },
+ {
+ "epoch": 33.92227074235808,
+ "grad_norm": 0.014047150500118732,
+ "learning_rate": 0.0006,
+ "loss": 4.142541885375977,
+ "step": 2442
+ },
+ {
+ "epoch": 33.93624454148472,
+ "grad_norm": 0.01377617847174406,
+ "learning_rate": 0.0006,
+ "loss": 4.268551826477051,
+ "step": 2443
+ },
+ {
+ "epoch": 33.95021834061135,
+ "grad_norm": 0.014671621844172478,
+ "learning_rate": 0.0006,
+ "loss": 4.198196887969971,
+ "step": 2444
+ },
+ {
+ "epoch": 33.96419213973799,
+ "grad_norm": 0.014806744642555714,
+ "learning_rate": 0.0006,
+ "loss": 4.057899475097656,
+ "step": 2445
+ },
+ {
+ "epoch": 33.97816593886463,
+ "grad_norm": 0.015734272077679634,
+ "learning_rate": 0.0006,
+ "loss": 4.163505554199219,
+ "step": 2446
+ },
+ {
+ "epoch": 33.992139737991266,
+ "grad_norm": 0.015250146389007568,
+ "learning_rate": 0.0006,
+ "loss": 4.076294898986816,
+ "step": 2447
+ },
+ {
+ "epoch": 34.0,
+ "grad_norm": 0.016733702272176743,
+ "learning_rate": 0.0006,
+ "loss": 4.138420104980469,
+ "step": 2448
+ },
+ {
+ "epoch": 34.0,
+ "eval_loss": 4.542503356933594,
+ "eval_runtime": 56.983,
+ "eval_samples_per_second": 42.855,
+ "eval_steps_per_second": 1.351,
+ "step": 2448
+ },
+ {
+ "epoch": 34.01397379912664,
+ "grad_norm": 0.015866613015532494,
+ "learning_rate": 0.0006,
+ "loss": 4.032081127166748,
+ "step": 2449
+ },
+ {
+ "epoch": 34.02794759825328,
+ "grad_norm": 0.01946570910513401,
+ "learning_rate": 0.0006,
+ "loss": 4.122320175170898,
+ "step": 2450
+ },
+ {
+ "epoch": 34.041921397379916,
+ "grad_norm": 0.020265938714146614,
+ "learning_rate": 0.0006,
+ "loss": 4.06460428237915,
+ "step": 2451
+ },
+ {
+ "epoch": 34.05589519650655,
+ "grad_norm": 0.018887171521782875,
+ "learning_rate": 0.0006,
+ "loss": 4.096906661987305,
+ "step": 2452
+ },
+ {
+ "epoch": 34.069868995633186,
+ "grad_norm": 0.017304692417383194,
+ "learning_rate": 0.0006,
+ "loss": 4.137261390686035,
+ "step": 2453
+ },
+ {
+ "epoch": 34.083842794759825,
+ "grad_norm": 0.017029082402586937,
+ "learning_rate": 0.0006,
+ "loss": 4.081112861633301,
+ "step": 2454
+ },
+ {
+ "epoch": 34.09781659388646,
+ "grad_norm": 0.016414813697338104,
+ "learning_rate": 0.0006,
+ "loss": 4.140783309936523,
+ "step": 2455
+ },
+ {
+ "epoch": 34.1117903930131,
+ "grad_norm": 0.018238281831145287,
+ "learning_rate": 0.0006,
+ "loss": 4.161116600036621,
+ "step": 2456
+ },
+ {
+ "epoch": 34.12576419213974,
+ "grad_norm": 0.019929232075810432,
+ "learning_rate": 0.0006,
+ "loss": 4.243722438812256,
+ "step": 2457
+ },
+ {
+ "epoch": 34.13973799126637,
+ "grad_norm": 0.019375786185264587,
+ "learning_rate": 0.0006,
+ "loss": 4.134873390197754,
+ "step": 2458
+ },
+ {
+ "epoch": 34.15371179039301,
+ "grad_norm": 0.018151137977838516,
+ "learning_rate": 0.0006,
+ "loss": 4.098745346069336,
+ "step": 2459
+ },
+ {
+ "epoch": 34.16768558951965,
+ "grad_norm": 0.017054934054613113,
+ "learning_rate": 0.0006,
+ "loss": 4.194662570953369,
+ "step": 2460
+ },
+ {
+ "epoch": 34.18165938864629,
+ "grad_norm": 0.018220555037260056,
+ "learning_rate": 0.0006,
+ "loss": 4.128866195678711,
+ "step": 2461
+ },
+ {
+ "epoch": 34.19563318777293,
+ "grad_norm": 0.01692879945039749,
+ "learning_rate": 0.0006,
+ "loss": 4.062477111816406,
+ "step": 2462
+ },
+ {
+ "epoch": 34.209606986899566,
+ "grad_norm": 0.01674199476838112,
+ "learning_rate": 0.0006,
+ "loss": 4.186149597167969,
+ "step": 2463
+ },
+ {
+ "epoch": 34.223580786026204,
+ "grad_norm": 0.016527455300092697,
+ "learning_rate": 0.0006,
+ "loss": 4.0734028816223145,
+ "step": 2464
+ },
+ {
+ "epoch": 34.237554585152836,
+ "grad_norm": 0.016448386013507843,
+ "learning_rate": 0.0006,
+ "loss": 4.098896026611328,
+ "step": 2465
+ },
+ {
+ "epoch": 34.251528384279474,
+ "grad_norm": 0.014644953422248363,
+ "learning_rate": 0.0006,
+ "loss": 4.1031904220581055,
+ "step": 2466
+ },
+ {
+ "epoch": 34.26550218340611,
+ "grad_norm": 0.016649341210722923,
+ "learning_rate": 0.0006,
+ "loss": 4.148791313171387,
+ "step": 2467
+ },
+ {
+ "epoch": 34.27947598253275,
+ "grad_norm": 0.017442714422941208,
+ "learning_rate": 0.0006,
+ "loss": 4.070799827575684,
+ "step": 2468
+ },
+ {
+ "epoch": 34.29344978165939,
+ "grad_norm": 0.017852013930678368,
+ "learning_rate": 0.0006,
+ "loss": 4.031924247741699,
+ "step": 2469
+ },
+ {
+ "epoch": 34.30742358078603,
+ "grad_norm": 0.016371270641684532,
+ "learning_rate": 0.0006,
+ "loss": 4.149420738220215,
+ "step": 2470
+ },
+ {
+ "epoch": 34.32139737991266,
+ "grad_norm": 0.015846814960241318,
+ "learning_rate": 0.0006,
+ "loss": 4.15427303314209,
+ "step": 2471
+ },
+ {
+ "epoch": 34.3353711790393,
+ "grad_norm": 0.017411163076758385,
+ "learning_rate": 0.0006,
+ "loss": 4.1304192543029785,
+ "step": 2472
+ },
+ {
+ "epoch": 34.34934497816594,
+ "grad_norm": 0.01780046708881855,
+ "learning_rate": 0.0006,
+ "loss": 4.191732406616211,
+ "step": 2473
+ },
+ {
+ "epoch": 34.36331877729258,
+ "grad_norm": 0.01958383060991764,
+ "learning_rate": 0.0006,
+ "loss": 4.068355083465576,
+ "step": 2474
+ },
+ {
+ "epoch": 34.377292576419215,
+ "grad_norm": 0.022173317149281502,
+ "learning_rate": 0.0006,
+ "loss": 4.182929039001465,
+ "step": 2475
+ },
+ {
+ "epoch": 34.391266375545854,
+ "grad_norm": 0.022389156743884087,
+ "learning_rate": 0.0006,
+ "loss": 4.110218048095703,
+ "step": 2476
+ },
+ {
+ "epoch": 34.40524017467249,
+ "grad_norm": 0.022478511556982994,
+ "learning_rate": 0.0006,
+ "loss": 4.095479488372803,
+ "step": 2477
+ },
+ {
+ "epoch": 34.419213973799124,
+ "grad_norm": 0.019504647701978683,
+ "learning_rate": 0.0006,
+ "loss": 4.101964950561523,
+ "step": 2478
+ },
+ {
+ "epoch": 34.43318777292576,
+ "grad_norm": 0.019378576427698135,
+ "learning_rate": 0.0006,
+ "loss": 4.14620304107666,
+ "step": 2479
+ },
+ {
+ "epoch": 34.4471615720524,
+ "grad_norm": 0.017993204295635223,
+ "learning_rate": 0.0006,
+ "loss": 4.183492660522461,
+ "step": 2480
+ },
+ {
+ "epoch": 34.46113537117904,
+ "grad_norm": 0.01814805157482624,
+ "learning_rate": 0.0006,
+ "loss": 4.151247978210449,
+ "step": 2481
+ },
+ {
+ "epoch": 34.47510917030568,
+ "grad_norm": 0.018622448667883873,
+ "learning_rate": 0.0006,
+ "loss": 4.020938396453857,
+ "step": 2482
+ },
+ {
+ "epoch": 34.48908296943232,
+ "grad_norm": 0.018360132351517677,
+ "learning_rate": 0.0006,
+ "loss": 4.064734935760498,
+ "step": 2483
+ },
+ {
+ "epoch": 34.50305676855895,
+ "grad_norm": 0.017715711146593094,
+ "learning_rate": 0.0006,
+ "loss": 4.185687065124512,
+ "step": 2484
+ },
+ {
+ "epoch": 34.51703056768559,
+ "grad_norm": 0.01714295893907547,
+ "learning_rate": 0.0006,
+ "loss": 4.1044721603393555,
+ "step": 2485
+ },
+ {
+ "epoch": 34.531004366812226,
+ "grad_norm": 0.016943803057074547,
+ "learning_rate": 0.0006,
+ "loss": 4.106837272644043,
+ "step": 2486
+ },
+ {
+ "epoch": 34.544978165938865,
+ "grad_norm": 0.0182423684746027,
+ "learning_rate": 0.0006,
+ "loss": 4.1041340827941895,
+ "step": 2487
+ },
+ {
+ "epoch": 34.5589519650655,
+ "grad_norm": 0.01636587455868721,
+ "learning_rate": 0.0006,
+ "loss": 3.9735546112060547,
+ "step": 2488
+ },
+ {
+ "epoch": 34.57292576419214,
+ "grad_norm": 0.014593546278774738,
+ "learning_rate": 0.0006,
+ "loss": 4.041980266571045,
+ "step": 2489
+ },
+ {
+ "epoch": 34.58689956331878,
+ "grad_norm": 0.015436965972185135,
+ "learning_rate": 0.0006,
+ "loss": 4.301148414611816,
+ "step": 2490
+ },
+ {
+ "epoch": 34.60087336244541,
+ "grad_norm": 0.015075269155204296,
+ "learning_rate": 0.0006,
+ "loss": 4.118527412414551,
+ "step": 2491
+ },
+ {
+ "epoch": 34.61484716157205,
+ "grad_norm": 0.01701514609158039,
+ "learning_rate": 0.0006,
+ "loss": 4.245968818664551,
+ "step": 2492
+ },
+ {
+ "epoch": 34.62882096069869,
+ "grad_norm": 0.018047887831926346,
+ "learning_rate": 0.0006,
+ "loss": 4.227261066436768,
+ "step": 2493
+ },
+ {
+ "epoch": 34.64279475982533,
+ "grad_norm": 0.018499085679650307,
+ "learning_rate": 0.0006,
+ "loss": 4.121360778808594,
+ "step": 2494
+ },
+ {
+ "epoch": 34.65676855895197,
+ "grad_norm": 0.018987711519002914,
+ "learning_rate": 0.0006,
+ "loss": 4.131564617156982,
+ "step": 2495
+ },
+ {
+ "epoch": 34.670742358078606,
+ "grad_norm": 0.01834476925432682,
+ "learning_rate": 0.0006,
+ "loss": 4.224941253662109,
+ "step": 2496
+ },
+ {
+ "epoch": 34.68471615720524,
+ "grad_norm": 0.017507942393422127,
+ "learning_rate": 0.0006,
+ "loss": 4.1394548416137695,
+ "step": 2497
+ },
+ {
+ "epoch": 34.698689956331876,
+ "grad_norm": 0.01680580899119377,
+ "learning_rate": 0.0006,
+ "loss": 4.1073431968688965,
+ "step": 2498
+ },
+ {
+ "epoch": 34.712663755458514,
+ "grad_norm": 0.015233514830470085,
+ "learning_rate": 0.0006,
+ "loss": 4.136994361877441,
+ "step": 2499
+ },
+ {
+ "epoch": 34.72663755458515,
+ "grad_norm": 0.015482917428016663,
+ "learning_rate": 0.0006,
+ "loss": 4.110719680786133,
+ "step": 2500
+ },
+ {
+ "epoch": 34.74061135371179,
+ "grad_norm": 0.015940461307764053,
+ "learning_rate": 0.0006,
+ "loss": 4.207851886749268,
+ "step": 2501
+ },
+ {
+ "epoch": 34.75458515283843,
+ "grad_norm": 0.015355597250163555,
+ "learning_rate": 0.0006,
+ "loss": 4.169469833374023,
+ "step": 2502
+ },
+ {
+ "epoch": 34.76855895196506,
+ "grad_norm": 0.015974000096321106,
+ "learning_rate": 0.0006,
+ "loss": 4.110061168670654,
+ "step": 2503
+ },
+ {
+ "epoch": 34.7825327510917,
+ "grad_norm": 0.016088927164673805,
+ "learning_rate": 0.0006,
+ "loss": 4.155452728271484,
+ "step": 2504
+ },
+ {
+ "epoch": 34.79650655021834,
+ "grad_norm": 0.016156120225787163,
+ "learning_rate": 0.0006,
+ "loss": 4.178092956542969,
+ "step": 2505
+ },
+ {
+ "epoch": 34.81048034934498,
+ "grad_norm": 0.014928647316992283,
+ "learning_rate": 0.0006,
+ "loss": 4.05457878112793,
+ "step": 2506
+ },
+ {
+ "epoch": 34.82445414847162,
+ "grad_norm": 0.015289328061044216,
+ "learning_rate": 0.0006,
+ "loss": 4.18869686126709,
+ "step": 2507
+ },
+ {
+ "epoch": 34.838427947598255,
+ "grad_norm": 0.015799039974808693,
+ "learning_rate": 0.0006,
+ "loss": 4.106073379516602,
+ "step": 2508
+ },
+ {
+ "epoch": 34.852401746724894,
+ "grad_norm": 0.015824418514966965,
+ "learning_rate": 0.0006,
+ "loss": 4.0634589195251465,
+ "step": 2509
+ },
+ {
+ "epoch": 34.866375545851525,
+ "grad_norm": 0.015321656130254269,
+ "learning_rate": 0.0006,
+ "loss": 4.093090057373047,
+ "step": 2510
+ },
+ {
+ "epoch": 34.880349344978164,
+ "grad_norm": 0.014380166307091713,
+ "learning_rate": 0.0006,
+ "loss": 4.1750640869140625,
+ "step": 2511
+ },
+ {
+ "epoch": 34.8943231441048,
+ "grad_norm": 0.014962567947804928,
+ "learning_rate": 0.0006,
+ "loss": 4.185277938842773,
+ "step": 2512
+ },
+ {
+ "epoch": 34.90829694323144,
+ "grad_norm": 0.014467543922364712,
+ "learning_rate": 0.0006,
+ "loss": 4.1600799560546875,
+ "step": 2513
+ },
+ {
+ "epoch": 34.92227074235808,
+ "grad_norm": 0.01545181218534708,
+ "learning_rate": 0.0006,
+ "loss": 4.096299648284912,
+ "step": 2514
+ },
+ {
+ "epoch": 34.93624454148472,
+ "grad_norm": 0.01605340465903282,
+ "learning_rate": 0.0006,
+ "loss": 4.203323841094971,
+ "step": 2515
+ },
+ {
+ "epoch": 34.95021834061135,
+ "grad_norm": 0.015203815884888172,
+ "learning_rate": 0.0006,
+ "loss": 4.201385021209717,
+ "step": 2516
+ },
+ {
+ "epoch": 34.96419213973799,
+ "grad_norm": 0.015149409882724285,
+ "learning_rate": 0.0006,
+ "loss": 4.088164329528809,
+ "step": 2517
+ },
+ {
+ "epoch": 34.97816593886463,
+ "grad_norm": 0.014885558746755123,
+ "learning_rate": 0.0006,
+ "loss": 4.073911666870117,
+ "step": 2518
+ },
+ {
+ "epoch": 34.992139737991266,
+ "grad_norm": 0.015003419481217861,
+ "learning_rate": 0.0006,
+ "loss": 4.157113552093506,
+ "step": 2519
+ },
+ {
+ "epoch": 35.0,
+ "grad_norm": 0.01641916297376156,
+ "learning_rate": 0.0006,
+ "loss": 3.970052719116211,
+ "step": 2520
+ },
+ {
+ "epoch": 35.0,
+ "eval_loss": 4.641692161560059,
+ "eval_runtime": 56.8836,
+ "eval_samples_per_second": 42.93,
+ "eval_steps_per_second": 1.354,
+ "step": 2520
+ },
+ {
+ "epoch": 35.01397379912664,
+ "grad_norm": 0.015894673764705658,
+ "learning_rate": 0.0006,
+ "loss": 4.104437351226807,
+ "step": 2521
+ },
+ {
+ "epoch": 35.02794759825328,
+ "grad_norm": 0.0173867866396904,
+ "learning_rate": 0.0006,
+ "loss": 4.091726303100586,
+ "step": 2522
+ },
+ {
+ "epoch": 35.041921397379916,
+ "grad_norm": 0.019272832199931145,
+ "learning_rate": 0.0006,
+ "loss": 4.217905044555664,
+ "step": 2523
+ },
+ {
+ "epoch": 35.05589519650655,
+ "grad_norm": 0.021820371970534325,
+ "learning_rate": 0.0006,
+ "loss": 4.111428260803223,
+ "step": 2524
+ },
+ {
+ "epoch": 35.069868995633186,
+ "grad_norm": 0.02340741828083992,
+ "learning_rate": 0.0006,
+ "loss": 4.115771293640137,
+ "step": 2525
+ },
+ {
+ "epoch": 35.083842794759825,
+ "grad_norm": 0.023760691285133362,
+ "learning_rate": 0.0006,
+ "loss": 4.132383346557617,
+ "step": 2526
+ },
+ {
+ "epoch": 35.09781659388646,
+ "grad_norm": 0.02073468267917633,
+ "learning_rate": 0.0006,
+ "loss": 4.272691249847412,
+ "step": 2527
+ },
+ {
+ "epoch": 35.1117903930131,
+ "grad_norm": 0.02197718247771263,
+ "learning_rate": 0.0006,
+ "loss": 4.105586528778076,
+ "step": 2528
+ },
+ {
+ "epoch": 35.12576419213974,
+ "grad_norm": 0.0200254675000906,
+ "learning_rate": 0.0006,
+ "loss": 3.968820095062256,
+ "step": 2529
+ },
+ {
+ "epoch": 35.13973799126637,
+ "grad_norm": 0.01929587684571743,
+ "learning_rate": 0.0006,
+ "loss": 4.045468807220459,
+ "step": 2530
+ },
+ {
+ "epoch": 35.15371179039301,
+ "grad_norm": 0.01886162906885147,
+ "learning_rate": 0.0006,
+ "loss": 4.185571670532227,
+ "step": 2531
+ },
+ {
+ "epoch": 35.16768558951965,
+ "grad_norm": 0.019789839163422585,
+ "learning_rate": 0.0006,
+ "loss": 3.9734551906585693,
+ "step": 2532
+ },
+ {
+ "epoch": 35.18165938864629,
+ "grad_norm": 0.0198142658919096,
+ "learning_rate": 0.0006,
+ "loss": 4.2662506103515625,
+ "step": 2533
+ },
+ {
+ "epoch": 35.19563318777293,
+ "grad_norm": 0.02087012305855751,
+ "learning_rate": 0.0006,
+ "loss": 4.175931930541992,
+ "step": 2534
+ },
+ {
+ "epoch": 35.209606986899566,
+ "grad_norm": 0.020294535905122757,
+ "learning_rate": 0.0006,
+ "loss": 4.060324668884277,
+ "step": 2535
+ },
+ {
+ "epoch": 35.223580786026204,
+ "grad_norm": 0.01908983662724495,
+ "learning_rate": 0.0006,
+ "loss": 4.217939853668213,
+ "step": 2536
+ },
+ {
+ "epoch": 35.237554585152836,
+ "grad_norm": 0.0178892333060503,
+ "learning_rate": 0.0006,
+ "loss": 4.190216541290283,
+ "step": 2537
+ },
+ {
+ "epoch": 35.251528384279474,
+ "grad_norm": 0.017892315983772278,
+ "learning_rate": 0.0006,
+ "loss": 4.067419052124023,
+ "step": 2538
+ },
+ {
+ "epoch": 35.26550218340611,
+ "grad_norm": 0.018556250259280205,
+ "learning_rate": 0.0006,
+ "loss": 4.105958461761475,
+ "step": 2539
+ },
+ {
+ "epoch": 35.27947598253275,
+ "grad_norm": 0.01944422349333763,
+ "learning_rate": 0.0006,
+ "loss": 4.155775547027588,
+ "step": 2540
+ },
+ {
+ "epoch": 35.29344978165939,
+ "grad_norm": 0.017333146184682846,
+ "learning_rate": 0.0006,
+ "loss": 4.244513511657715,
+ "step": 2541
+ },
+ {
+ "epoch": 35.30742358078603,
+ "grad_norm": 0.01639244332909584,
+ "learning_rate": 0.0006,
+ "loss": 4.167815685272217,
+ "step": 2542
+ },
+ {
+ "epoch": 35.32139737991266,
+ "grad_norm": 0.01610143668949604,
+ "learning_rate": 0.0006,
+ "loss": 4.121486663818359,
+ "step": 2543
+ },
+ {
+ "epoch": 35.3353711790393,
+ "grad_norm": 0.01656411588191986,
+ "learning_rate": 0.0006,
+ "loss": 4.1755499839782715,
+ "step": 2544
+ },
+ {
+ "epoch": 35.34934497816594,
+ "grad_norm": 0.015075297094881535,
+ "learning_rate": 0.0006,
+ "loss": 4.102992534637451,
+ "step": 2545
+ },
+ {
+ "epoch": 35.36331877729258,
+ "grad_norm": 0.01538170501589775,
+ "learning_rate": 0.0006,
+ "loss": 4.161603927612305,
+ "step": 2546
+ },
+ {
+ "epoch": 35.377292576419215,
+ "grad_norm": 0.01483312901109457,
+ "learning_rate": 0.0006,
+ "loss": 4.085355758666992,
+ "step": 2547
+ },
+ {
+ "epoch": 35.391266375545854,
+ "grad_norm": 0.014894738793373108,
+ "learning_rate": 0.0006,
+ "loss": 4.083585262298584,
+ "step": 2548
+ },
+ {
+ "epoch": 35.40524017467249,
+ "grad_norm": 0.01475780550390482,
+ "learning_rate": 0.0006,
+ "loss": 4.012645721435547,
+ "step": 2549
+ },
+ {
+ "epoch": 35.419213973799124,
+ "grad_norm": 0.014531653374433517,
+ "learning_rate": 0.0006,
+ "loss": 4.197113990783691,
+ "step": 2550
+ },
+ {
+ "epoch": 35.43318777292576,
+ "grad_norm": 0.015316493809223175,
+ "learning_rate": 0.0006,
+ "loss": 4.227970123291016,
+ "step": 2551
+ },
+ {
+ "epoch": 35.4471615720524,
+ "grad_norm": 0.014685571193695068,
+ "learning_rate": 0.0006,
+ "loss": 4.089415550231934,
+ "step": 2552
+ },
+ {
+ "epoch": 35.46113537117904,
+ "grad_norm": 0.015337035991251469,
+ "learning_rate": 0.0006,
+ "loss": 4.080583572387695,
+ "step": 2553
+ },
+ {
+ "epoch": 35.47510917030568,
+ "grad_norm": 0.016754208132624626,
+ "learning_rate": 0.0006,
+ "loss": 4.163935661315918,
+ "step": 2554
+ },
+ {
+ "epoch": 35.48908296943232,
+ "grad_norm": 0.017535798251628876,
+ "learning_rate": 0.0006,
+ "loss": 4.093027591705322,
+ "step": 2555
+ },
+ {
+ "epoch": 35.50305676855895,
+ "grad_norm": 0.017336975783109665,
+ "learning_rate": 0.0006,
+ "loss": 4.127961158752441,
+ "step": 2556
+ },
+ {
+ "epoch": 35.51703056768559,
+ "grad_norm": 0.016242152079939842,
+ "learning_rate": 0.0006,
+ "loss": 4.155593395233154,
+ "step": 2557
+ },
+ {
+ "epoch": 35.531004366812226,
+ "grad_norm": 0.015121322125196457,
+ "learning_rate": 0.0006,
+ "loss": 3.984004020690918,
+ "step": 2558
+ },
+ {
+ "epoch": 35.544978165938865,
+ "grad_norm": 0.015178884379565716,
+ "learning_rate": 0.0006,
+ "loss": 4.111564636230469,
+ "step": 2559
+ },
+ {
+ "epoch": 35.5589519650655,
+ "grad_norm": 0.014777131378650665,
+ "learning_rate": 0.0006,
+ "loss": 4.041402816772461,
+ "step": 2560
+ },
+ {
+ "epoch": 35.57292576419214,
+ "grad_norm": 0.015197676606476307,
+ "learning_rate": 0.0006,
+ "loss": 4.133706092834473,
+ "step": 2561
+ },
+ {
+ "epoch": 35.58689956331878,
+ "grad_norm": 0.015905972570180893,
+ "learning_rate": 0.0006,
+ "loss": 4.1069135665893555,
+ "step": 2562
+ },
+ {
+ "epoch": 35.60087336244541,
+ "grad_norm": 0.015942877158522606,
+ "learning_rate": 0.0006,
+ "loss": 4.0720624923706055,
+ "step": 2563
+ },
+ {
+ "epoch": 35.61484716157205,
+ "grad_norm": 0.016697583720088005,
+ "learning_rate": 0.0006,
+ "loss": 4.167448043823242,
+ "step": 2564
+ },
+ {
+ "epoch": 35.62882096069869,
+ "grad_norm": 0.015603577718138695,
+ "learning_rate": 0.0006,
+ "loss": 4.127327919006348,
+ "step": 2565
+ },
+ {
+ "epoch": 35.64279475982533,
+ "grad_norm": 0.017153460532426834,
+ "learning_rate": 0.0006,
+ "loss": 4.056543350219727,
+ "step": 2566
+ },
+ {
+ "epoch": 35.65676855895197,
+ "grad_norm": 0.016731958836317062,
+ "learning_rate": 0.0006,
+ "loss": 4.0681843757629395,
+ "step": 2567
+ },
+ {
+ "epoch": 35.670742358078606,
+ "grad_norm": 0.016167359426617622,
+ "learning_rate": 0.0006,
+ "loss": 4.174992084503174,
+ "step": 2568
+ },
+ {
+ "epoch": 35.68471615720524,
+ "grad_norm": 0.01615108922123909,
+ "learning_rate": 0.0006,
+ "loss": 4.1383161544799805,
+ "step": 2569
+ },
+ {
+ "epoch": 35.698689956331876,
+ "grad_norm": 0.015750816091895103,
+ "learning_rate": 0.0006,
+ "loss": 4.031595706939697,
+ "step": 2570
+ },
+ {
+ "epoch": 35.712663755458514,
+ "grad_norm": 0.015416189096868038,
+ "learning_rate": 0.0006,
+ "loss": 4.067498207092285,
+ "step": 2571
+ },
+ {
+ "epoch": 35.72663755458515,
+ "grad_norm": 0.01570792682468891,
+ "learning_rate": 0.0006,
+ "loss": 4.1338300704956055,
+ "step": 2572
+ },
+ {
+ "epoch": 35.74061135371179,
+ "grad_norm": 0.016367752104997635,
+ "learning_rate": 0.0006,
+ "loss": 4.152523040771484,
+ "step": 2573
+ },
+ {
+ "epoch": 35.75458515283843,
+ "grad_norm": 0.016349803656339645,
+ "learning_rate": 0.0006,
+ "loss": 4.101985454559326,
+ "step": 2574
+ },
+ {
+ "epoch": 35.76855895196506,
+ "grad_norm": 0.01663660630583763,
+ "learning_rate": 0.0006,
+ "loss": 4.190130710601807,
+ "step": 2575
+ },
+ {
+ "epoch": 35.7825327510917,
+ "grad_norm": 0.015599170699715614,
+ "learning_rate": 0.0006,
+ "loss": 4.0792436599731445,
+ "step": 2576
+ },
+ {
+ "epoch": 35.79650655021834,
+ "grad_norm": 0.015472427010536194,
+ "learning_rate": 0.0006,
+ "loss": 4.186446666717529,
+ "step": 2577
+ },
+ {
+ "epoch": 35.81048034934498,
+ "grad_norm": 0.015520439483225346,
+ "learning_rate": 0.0006,
+ "loss": 4.246635437011719,
+ "step": 2578
+ },
+ {
+ "epoch": 35.82445414847162,
+ "grad_norm": 0.016998281702399254,
+ "learning_rate": 0.0006,
+ "loss": 4.13435173034668,
+ "step": 2579
+ },
+ {
+ "epoch": 35.838427947598255,
+ "grad_norm": 0.016290361061692238,
+ "learning_rate": 0.0006,
+ "loss": 4.107428073883057,
+ "step": 2580
+ },
+ {
+ "epoch": 35.852401746724894,
+ "grad_norm": 0.0165516659617424,
+ "learning_rate": 0.0006,
+ "loss": 4.1612138748168945,
+ "step": 2581
+ },
+ {
+ "epoch": 35.866375545851525,
+ "grad_norm": 0.01736091636121273,
+ "learning_rate": 0.0006,
+ "loss": 4.1072869300842285,
+ "step": 2582
+ },
+ {
+ "epoch": 35.880349344978164,
+ "grad_norm": 0.01787031814455986,
+ "learning_rate": 0.0006,
+ "loss": 4.1076812744140625,
+ "step": 2583
+ },
+ {
+ "epoch": 35.8943231441048,
+ "grad_norm": 0.015437985770404339,
+ "learning_rate": 0.0006,
+ "loss": 4.097488880157471,
+ "step": 2584
+ },
+ {
+ "epoch": 35.90829694323144,
+ "grad_norm": 0.016306741163134575,
+ "learning_rate": 0.0006,
+ "loss": 4.105017185211182,
+ "step": 2585
+ },
+ {
+ "epoch": 35.92227074235808,
+ "grad_norm": 0.015723245218396187,
+ "learning_rate": 0.0006,
+ "loss": 4.205283164978027,
+ "step": 2586
+ },
+ {
+ "epoch": 35.93624454148472,
+ "grad_norm": 0.014626456424593925,
+ "learning_rate": 0.0006,
+ "loss": 4.086552619934082,
+ "step": 2587
+ },
+ {
+ "epoch": 35.95021834061135,
+ "grad_norm": 0.014348454773426056,
+ "learning_rate": 0.0006,
+ "loss": 4.171329975128174,
+ "step": 2588
+ },
+ {
+ "epoch": 35.96419213973799,
+ "grad_norm": 0.015598480589687824,
+ "learning_rate": 0.0006,
+ "loss": 4.140815258026123,
+ "step": 2589
+ },
+ {
+ "epoch": 35.97816593886463,
+ "grad_norm": 0.017126062884926796,
+ "learning_rate": 0.0006,
+ "loss": 4.157176971435547,
+ "step": 2590
+ },
+ {
+ "epoch": 35.992139737991266,
+ "grad_norm": 0.01860199309885502,
+ "learning_rate": 0.0006,
+ "loss": 4.252880096435547,
+ "step": 2591
+ },
+ {
+ "epoch": 36.0,
+ "grad_norm": 0.021780220791697502,
+ "learning_rate": 0.0006,
+ "loss": 4.249861717224121,
+ "step": 2592
+ },
+ {
+ "epoch": 36.0,
+ "eval_loss": 4.606466293334961,
+ "eval_runtime": 56.6217,
+ "eval_samples_per_second": 43.128,
+ "eval_steps_per_second": 1.36,
+ "step": 2592
+ },
+ {
+ "epoch": 36.01397379912664,
+ "grad_norm": 0.020892828702926636,
+ "learning_rate": 0.0006,
+ "loss": 4.102055549621582,
+ "step": 2593
+ },
+ {
+ "epoch": 36.02794759825328,
+ "grad_norm": 0.020306255668401718,
+ "learning_rate": 0.0006,
+ "loss": 4.184308052062988,
+ "step": 2594
+ },
+ {
+ "epoch": 36.041921397379916,
+ "grad_norm": 0.02025097794830799,
+ "learning_rate": 0.0006,
+ "loss": 3.9691858291625977,
+ "step": 2595
+ },
+ {
+ "epoch": 36.05589519650655,
+ "grad_norm": 0.020254382863640785,
+ "learning_rate": 0.0006,
+ "loss": 4.02957820892334,
+ "step": 2596
+ },
+ {
+ "epoch": 36.069868995633186,
+ "grad_norm": 0.019307196140289307,
+ "learning_rate": 0.0006,
+ "loss": 4.033181667327881,
+ "step": 2597
+ },
+ {
+ "epoch": 36.083842794759825,
+ "grad_norm": 0.0198476854711771,
+ "learning_rate": 0.0006,
+ "loss": 4.0816874504089355,
+ "step": 2598
+ },
+ {
+ "epoch": 36.09781659388646,
+ "grad_norm": 0.019916515797376633,
+ "learning_rate": 0.0006,
+ "loss": 4.164889812469482,
+ "step": 2599
+ },
+ {
+ "epoch": 36.1117903930131,
+ "grad_norm": 0.0189144778996706,
+ "learning_rate": 0.0006,
+ "loss": 4.038427352905273,
+ "step": 2600
+ },
+ {
+ "epoch": 36.12576419213974,
+ "grad_norm": 0.01754768192768097,
+ "learning_rate": 0.0006,
+ "loss": 4.010427474975586,
+ "step": 2601
+ },
+ {
+ "epoch": 36.13973799126637,
+ "grad_norm": 0.016862990334630013,
+ "learning_rate": 0.0006,
+ "loss": 4.035970211029053,
+ "step": 2602
+ },
+ {
+ "epoch": 36.15371179039301,
+ "grad_norm": 0.019616752862930298,
+ "learning_rate": 0.0006,
+ "loss": 4.079898834228516,
+ "step": 2603
+ },
+ {
+ "epoch": 36.16768558951965,
+ "grad_norm": 0.020131465047597885,
+ "learning_rate": 0.0006,
+ "loss": 4.144684791564941,
+ "step": 2604
+ },
+ {
+ "epoch": 36.18165938864629,
+ "grad_norm": 0.02072521299123764,
+ "learning_rate": 0.0006,
+ "loss": 4.013660430908203,
+ "step": 2605
+ },
+ {
+ "epoch": 36.19563318777293,
+ "grad_norm": 0.01924615167081356,
+ "learning_rate": 0.0006,
+ "loss": 4.071398735046387,
+ "step": 2606
+ },
+ {
+ "epoch": 36.209606986899566,
+ "grad_norm": 0.02013954520225525,
+ "learning_rate": 0.0006,
+ "loss": 4.127094268798828,
+ "step": 2607
+ },
+ {
+ "epoch": 36.223580786026204,
+ "grad_norm": 0.02042008936405182,
+ "learning_rate": 0.0006,
+ "loss": 4.154261589050293,
+ "step": 2608
+ },
+ {
+ "epoch": 36.237554585152836,
+ "grad_norm": 0.021998876705765724,
+ "learning_rate": 0.0006,
+ "loss": 4.204150199890137,
+ "step": 2609
+ },
+ {
+ "epoch": 36.251528384279474,
+ "grad_norm": 0.022946683689951897,
+ "learning_rate": 0.0006,
+ "loss": 4.013632774353027,
+ "step": 2610
+ },
+ {
+ "epoch": 36.26550218340611,
+ "grad_norm": 0.02061830274760723,
+ "learning_rate": 0.0006,
+ "loss": 4.171084403991699,
+ "step": 2611
+ },
+ {
+ "epoch": 36.27947598253275,
+ "grad_norm": 0.01883152686059475,
+ "learning_rate": 0.0006,
+ "loss": 4.149848461151123,
+ "step": 2612
+ },
+ {
+ "epoch": 36.29344978165939,
+ "grad_norm": 0.019945282489061356,
+ "learning_rate": 0.0006,
+ "loss": 4.174656867980957,
+ "step": 2613
+ },
+ {
+ "epoch": 36.30742358078603,
+ "grad_norm": 0.020751051604747772,
+ "learning_rate": 0.0006,
+ "loss": 4.080747127532959,
+ "step": 2614
+ },
+ {
+ "epoch": 36.32139737991266,
+ "grad_norm": 0.01879691146314144,
+ "learning_rate": 0.0006,
+ "loss": 4.057356357574463,
+ "step": 2615
+ },
+ {
+ "epoch": 36.3353711790393,
+ "grad_norm": 0.019037071615457535,
+ "learning_rate": 0.0006,
+ "loss": 4.204607009887695,
+ "step": 2616
+ },
+ {
+ "epoch": 36.34934497816594,
+ "grad_norm": 0.01872078701853752,
+ "learning_rate": 0.0006,
+ "loss": 4.050586700439453,
+ "step": 2617
+ },
+ {
+ "epoch": 36.36331877729258,
+ "grad_norm": 0.018092811107635498,
+ "learning_rate": 0.0006,
+ "loss": 4.102832794189453,
+ "step": 2618
+ },
+ {
+ "epoch": 36.377292576419215,
+ "grad_norm": 0.017010487616062164,
+ "learning_rate": 0.0006,
+ "loss": 4.191788673400879,
+ "step": 2619
+ },
+ {
+ "epoch": 36.391266375545854,
+ "grad_norm": 0.01691143773496151,
+ "learning_rate": 0.0006,
+ "loss": 4.160456657409668,
+ "step": 2620
+ },
+ {
+ "epoch": 36.40524017467249,
+ "grad_norm": 0.017046816647052765,
+ "learning_rate": 0.0006,
+ "loss": 4.073828220367432,
+ "step": 2621
+ },
+ {
+ "epoch": 36.419213973799124,
+ "grad_norm": 0.01707981713116169,
+ "learning_rate": 0.0006,
+ "loss": 4.0980095863342285,
+ "step": 2622
+ },
+ {
+ "epoch": 36.43318777292576,
+ "grad_norm": 0.01630452275276184,
+ "learning_rate": 0.0006,
+ "loss": 4.101321220397949,
+ "step": 2623
+ },
+ {
+ "epoch": 36.4471615720524,
+ "grad_norm": 0.01576915942132473,
+ "learning_rate": 0.0006,
+ "loss": 4.158249855041504,
+ "step": 2624
+ },
+ {
+ "epoch": 36.46113537117904,
+ "grad_norm": 0.016179712489247322,
+ "learning_rate": 0.0006,
+ "loss": 4.077852249145508,
+ "step": 2625
+ },
+ {
+ "epoch": 36.47510917030568,
+ "grad_norm": 0.01609169878065586,
+ "learning_rate": 0.0006,
+ "loss": 4.187317848205566,
+ "step": 2626
+ },
+ {
+ "epoch": 36.48908296943232,
+ "grad_norm": 0.017148254439234734,
+ "learning_rate": 0.0006,
+ "loss": 4.109071731567383,
+ "step": 2627
+ },
+ {
+ "epoch": 36.50305676855895,
+ "grad_norm": 0.015842624008655548,
+ "learning_rate": 0.0006,
+ "loss": 4.145096302032471,
+ "step": 2628
+ },
+ {
+ "epoch": 36.51703056768559,
+ "grad_norm": 0.014906194061040878,
+ "learning_rate": 0.0006,
+ "loss": 4.028440475463867,
+ "step": 2629
+ },
+ {
+ "epoch": 36.531004366812226,
+ "grad_norm": 0.01485038734972477,
+ "learning_rate": 0.0006,
+ "loss": 4.175201416015625,
+ "step": 2630
+ },
+ {
+ "epoch": 36.544978165938865,
+ "grad_norm": 0.015834150835871696,
+ "learning_rate": 0.0006,
+ "loss": 4.076434135437012,
+ "step": 2631
+ },
+ {
+ "epoch": 36.5589519650655,
+ "grad_norm": 0.015411358326673508,
+ "learning_rate": 0.0006,
+ "loss": 4.03912878036499,
+ "step": 2632
+ },
+ {
+ "epoch": 36.57292576419214,
+ "grad_norm": 0.014762846753001213,
+ "learning_rate": 0.0006,
+ "loss": 4.019827365875244,
+ "step": 2633
+ },
+ {
+ "epoch": 36.58689956331878,
+ "grad_norm": 0.015269497409462929,
+ "learning_rate": 0.0006,
+ "loss": 4.116957187652588,
+ "step": 2634
+ },
+ {
+ "epoch": 36.60087336244541,
+ "grad_norm": 0.015171617269515991,
+ "learning_rate": 0.0006,
+ "loss": 4.2069926261901855,
+ "step": 2635
+ },
+ {
+ "epoch": 36.61484716157205,
+ "grad_norm": 0.014770902693271637,
+ "learning_rate": 0.0006,
+ "loss": 4.092021942138672,
+ "step": 2636
+ },
+ {
+ "epoch": 36.62882096069869,
+ "grad_norm": 0.014957531355321407,
+ "learning_rate": 0.0006,
+ "loss": 4.16819953918457,
+ "step": 2637
+ },
+ {
+ "epoch": 36.64279475982533,
+ "grad_norm": 0.014774451963603497,
+ "learning_rate": 0.0006,
+ "loss": 4.046073913574219,
+ "step": 2638
+ },
+ {
+ "epoch": 36.65676855895197,
+ "grad_norm": 0.014320445246994495,
+ "learning_rate": 0.0006,
+ "loss": 3.977792501449585,
+ "step": 2639
+ },
+ {
+ "epoch": 36.670742358078606,
+ "grad_norm": 0.01413130946457386,
+ "learning_rate": 0.0006,
+ "loss": 4.130649566650391,
+ "step": 2640
+ },
+ {
+ "epoch": 36.68471615720524,
+ "grad_norm": 0.014336307533085346,
+ "learning_rate": 0.0006,
+ "loss": 4.161357402801514,
+ "step": 2641
+ },
+ {
+ "epoch": 36.698689956331876,
+ "grad_norm": 0.014516931027173996,
+ "learning_rate": 0.0006,
+ "loss": 4.14248514175415,
+ "step": 2642
+ },
+ {
+ "epoch": 36.712663755458514,
+ "grad_norm": 0.014377126470208168,
+ "learning_rate": 0.0006,
+ "loss": 4.062711715698242,
+ "step": 2643
+ },
+ {
+ "epoch": 36.72663755458515,
+ "grad_norm": 0.014332668855786324,
+ "learning_rate": 0.0006,
+ "loss": 4.06423807144165,
+ "step": 2644
+ },
+ {
+ "epoch": 36.74061135371179,
+ "grad_norm": 0.014030571095645428,
+ "learning_rate": 0.0006,
+ "loss": 4.178613662719727,
+ "step": 2645
+ },
+ {
+ "epoch": 36.75458515283843,
+ "grad_norm": 0.014376776292920113,
+ "learning_rate": 0.0006,
+ "loss": 4.073130130767822,
+ "step": 2646
+ },
+ {
+ "epoch": 36.76855895196506,
+ "grad_norm": 0.013827716931700706,
+ "learning_rate": 0.0006,
+ "loss": 4.043869972229004,
+ "step": 2647
+ },
+ {
+ "epoch": 36.7825327510917,
+ "grad_norm": 0.015548762865364552,
+ "learning_rate": 0.0006,
+ "loss": 4.134049415588379,
+ "step": 2648
+ },
+ {
+ "epoch": 36.79650655021834,
+ "grad_norm": 0.017041970044374466,
+ "learning_rate": 0.0006,
+ "loss": 4.205024242401123,
+ "step": 2649
+ },
+ {
+ "epoch": 36.81048034934498,
+ "grad_norm": 0.017638275399804115,
+ "learning_rate": 0.0006,
+ "loss": 4.199924468994141,
+ "step": 2650
+ },
+ {
+ "epoch": 36.82445414847162,
+ "grad_norm": 0.017771627753973007,
+ "learning_rate": 0.0006,
+ "loss": 4.15706729888916,
+ "step": 2651
+ },
+ {
+ "epoch": 36.838427947598255,
+ "grad_norm": 0.015933562070131302,
+ "learning_rate": 0.0006,
+ "loss": 4.220742225646973,
+ "step": 2652
+ },
+ {
+ "epoch": 36.852401746724894,
+ "grad_norm": 0.01668144389986992,
+ "learning_rate": 0.0006,
+ "loss": 4.161426544189453,
+ "step": 2653
+ },
+ {
+ "epoch": 36.866375545851525,
+ "grad_norm": 0.018014680594205856,
+ "learning_rate": 0.0006,
+ "loss": 4.183046817779541,
+ "step": 2654
+ },
+ {
+ "epoch": 36.880349344978164,
+ "grad_norm": 0.018808670341968536,
+ "learning_rate": 0.0006,
+ "loss": 4.105293273925781,
+ "step": 2655
+ },
+ {
+ "epoch": 36.8943231441048,
+ "grad_norm": 0.018003476783633232,
+ "learning_rate": 0.0006,
+ "loss": 4.103171348571777,
+ "step": 2656
+ },
+ {
+ "epoch": 36.90829694323144,
+ "grad_norm": 0.016865914687514305,
+ "learning_rate": 0.0006,
+ "loss": 4.125634670257568,
+ "step": 2657
+ },
+ {
+ "epoch": 36.92227074235808,
+ "grad_norm": 0.018371153622865677,
+ "learning_rate": 0.0006,
+ "loss": 4.192461013793945,
+ "step": 2658
+ },
+ {
+ "epoch": 36.93624454148472,
+ "grad_norm": 0.019045377150177956,
+ "learning_rate": 0.0006,
+ "loss": 4.198105812072754,
+ "step": 2659
+ },
+ {
+ "epoch": 36.95021834061135,
+ "grad_norm": 0.01594168320298195,
+ "learning_rate": 0.0006,
+ "loss": 4.169788360595703,
+ "step": 2660
+ },
+ {
+ "epoch": 36.96419213973799,
+ "grad_norm": 0.015645820647478104,
+ "learning_rate": 0.0006,
+ "loss": 4.047277450561523,
+ "step": 2661
+ },
+ {
+ "epoch": 36.97816593886463,
+ "grad_norm": 0.016313156113028526,
+ "learning_rate": 0.0006,
+ "loss": 4.087392807006836,
+ "step": 2662
+ },
+ {
+ "epoch": 36.992139737991266,
+ "grad_norm": 0.017050625756382942,
+ "learning_rate": 0.0006,
+ "loss": 4.0908002853393555,
+ "step": 2663
+ },
+ {
+ "epoch": 37.0,
+ "grad_norm": 0.017730968073010445,
+ "learning_rate": 0.0006,
+ "loss": 4.110957622528076,
+ "step": 2664
+ },
+ {
+ "epoch": 37.0,
+ "eval_loss": 4.563793182373047,
+ "eval_runtime": 56.5922,
+ "eval_samples_per_second": 43.151,
+ "eval_steps_per_second": 1.361,
+ "step": 2664
+ },
+ {
+ "epoch": 37.01397379912664,
+ "grad_norm": 0.015687910839915276,
+ "learning_rate": 0.0006,
+ "loss": 4.143589019775391,
+ "step": 2665
+ },
+ {
+ "epoch": 37.02794759825328,
+ "grad_norm": 0.017451366409659386,
+ "learning_rate": 0.0006,
+ "loss": 4.0326666831970215,
+ "step": 2666
+ },
+ {
+ "epoch": 37.041921397379916,
+ "grad_norm": 0.01909538544714451,
+ "learning_rate": 0.0006,
+ "loss": 3.9706249237060547,
+ "step": 2667
+ },
+ {
+ "epoch": 37.05589519650655,
+ "grad_norm": 0.01988379657268524,
+ "learning_rate": 0.0006,
+ "loss": 3.9531009197235107,
+ "step": 2668
+ },
+ {
+ "epoch": 37.069868995633186,
+ "grad_norm": 0.02035115472972393,
+ "learning_rate": 0.0006,
+ "loss": 3.962214469909668,
+ "step": 2669
+ },
+ {
+ "epoch": 37.083842794759825,
+ "grad_norm": 0.01952001079916954,
+ "learning_rate": 0.0006,
+ "loss": 4.182262897491455,
+ "step": 2670
+ },
+ {
+ "epoch": 37.09781659388646,
+ "grad_norm": 0.01854325644671917,
+ "learning_rate": 0.0006,
+ "loss": 4.127663612365723,
+ "step": 2671
+ },
+ {
+ "epoch": 37.1117903930131,
+ "grad_norm": 0.017757616937160492,
+ "learning_rate": 0.0006,
+ "loss": 4.100804805755615,
+ "step": 2672
+ },
+ {
+ "epoch": 37.12576419213974,
+ "grad_norm": 0.016712479293346405,
+ "learning_rate": 0.0006,
+ "loss": 4.026284217834473,
+ "step": 2673
+ },
+ {
+ "epoch": 37.13973799126637,
+ "grad_norm": 0.0168139711022377,
+ "learning_rate": 0.0006,
+ "loss": 4.183882713317871,
+ "step": 2674
+ },
+ {
+ "epoch": 37.15371179039301,
+ "grad_norm": 0.017060423269867897,
+ "learning_rate": 0.0006,
+ "loss": 4.131832599639893,
+ "step": 2675
+ },
+ {
+ "epoch": 37.16768558951965,
+ "grad_norm": 0.01726631447672844,
+ "learning_rate": 0.0006,
+ "loss": 3.9156057834625244,
+ "step": 2676
+ },
+ {
+ "epoch": 37.18165938864629,
+ "grad_norm": 0.016348622739315033,
+ "learning_rate": 0.0006,
+ "loss": 4.1731367111206055,
+ "step": 2677
+ },
+ {
+ "epoch": 37.19563318777293,
+ "grad_norm": 0.016640059649944305,
+ "learning_rate": 0.0006,
+ "loss": 4.10997200012207,
+ "step": 2678
+ },
+ {
+ "epoch": 37.209606986899566,
+ "grad_norm": 0.01601342298090458,
+ "learning_rate": 0.0006,
+ "loss": 4.081145286560059,
+ "step": 2679
+ },
+ {
+ "epoch": 37.223580786026204,
+ "grad_norm": 0.01735631190240383,
+ "learning_rate": 0.0006,
+ "loss": 4.202108860015869,
+ "step": 2680
+ },
+ {
+ "epoch": 37.237554585152836,
+ "grad_norm": 0.019164685159921646,
+ "learning_rate": 0.0006,
+ "loss": 4.104109764099121,
+ "step": 2681
+ },
+ {
+ "epoch": 37.251528384279474,
+ "grad_norm": 0.01881759613752365,
+ "learning_rate": 0.0006,
+ "loss": 4.145532131195068,
+ "step": 2682
+ },
+ {
+ "epoch": 37.26550218340611,
+ "grad_norm": 0.017656555399298668,
+ "learning_rate": 0.0006,
+ "loss": 4.012782096862793,
+ "step": 2683
+ },
+ {
+ "epoch": 37.27947598253275,
+ "grad_norm": 0.01624823734164238,
+ "learning_rate": 0.0006,
+ "loss": 4.001727104187012,
+ "step": 2684
+ },
+ {
+ "epoch": 37.29344978165939,
+ "grad_norm": 0.016973139718174934,
+ "learning_rate": 0.0006,
+ "loss": 4.095399856567383,
+ "step": 2685
+ },
+ {
+ "epoch": 37.30742358078603,
+ "grad_norm": 0.015503671951591969,
+ "learning_rate": 0.0006,
+ "loss": 4.034431457519531,
+ "step": 2686
+ },
+ {
+ "epoch": 37.32139737991266,
+ "grad_norm": 0.015370250679552555,
+ "learning_rate": 0.0006,
+ "loss": 4.075836181640625,
+ "step": 2687
+ },
+ {
+ "epoch": 37.3353711790393,
+ "grad_norm": 0.01564021222293377,
+ "learning_rate": 0.0006,
+ "loss": 4.125202178955078,
+ "step": 2688
+ },
+ {
+ "epoch": 37.34934497816594,
+ "grad_norm": 0.016717975959181786,
+ "learning_rate": 0.0006,
+ "loss": 4.153017044067383,
+ "step": 2689
+ },
+ {
+ "epoch": 37.36331877729258,
+ "grad_norm": 0.015263617038726807,
+ "learning_rate": 0.0006,
+ "loss": 4.021814823150635,
+ "step": 2690
+ },
+ {
+ "epoch": 37.377292576419215,
+ "grad_norm": 0.016497142612934113,
+ "learning_rate": 0.0006,
+ "loss": 4.069095611572266,
+ "step": 2691
+ },
+ {
+ "epoch": 37.391266375545854,
+ "grad_norm": 0.01669355481863022,
+ "learning_rate": 0.0006,
+ "loss": 4.174587726593018,
+ "step": 2692
+ },
+ {
+ "epoch": 37.40524017467249,
+ "grad_norm": 0.019695773720741272,
+ "learning_rate": 0.0006,
+ "loss": 4.0073041915893555,
+ "step": 2693
+ },
+ {
+ "epoch": 37.419213973799124,
+ "grad_norm": 0.023720644414424896,
+ "learning_rate": 0.0006,
+ "loss": 4.111948013305664,
+ "step": 2694
+ },
+ {
+ "epoch": 37.43318777292576,
+ "grad_norm": 0.02399350143969059,
+ "learning_rate": 0.0006,
+ "loss": 4.081164360046387,
+ "step": 2695
+ },
+ {
+ "epoch": 37.4471615720524,
+ "grad_norm": 0.019630271941423416,
+ "learning_rate": 0.0006,
+ "loss": 4.065731048583984,
+ "step": 2696
+ },
+ {
+ "epoch": 37.46113537117904,
+ "grad_norm": 0.019679008051753044,
+ "learning_rate": 0.0006,
+ "loss": 3.993852376937866,
+ "step": 2697
+ },
+ {
+ "epoch": 37.47510917030568,
+ "grad_norm": 0.020597299560904503,
+ "learning_rate": 0.0006,
+ "loss": 4.275555610656738,
+ "step": 2698
+ },
+ {
+ "epoch": 37.48908296943232,
+ "grad_norm": 0.01983889564871788,
+ "learning_rate": 0.0006,
+ "loss": 4.254708290100098,
+ "step": 2699
+ },
+ {
+ "epoch": 37.50305676855895,
+ "grad_norm": 0.019150668755173683,
+ "learning_rate": 0.0006,
+ "loss": 4.097885608673096,
+ "step": 2700
+ },
+ {
+ "epoch": 37.51703056768559,
+ "grad_norm": 0.018847715109586716,
+ "learning_rate": 0.0006,
+ "loss": 4.054081439971924,
+ "step": 2701
+ },
+ {
+ "epoch": 37.531004366812226,
+ "grad_norm": 0.019607288762927055,
+ "learning_rate": 0.0006,
+ "loss": 3.9891440868377686,
+ "step": 2702
+ },
+ {
+ "epoch": 37.544978165938865,
+ "grad_norm": 0.018863150849938393,
+ "learning_rate": 0.0006,
+ "loss": 4.084877014160156,
+ "step": 2703
+ },
+ {
+ "epoch": 37.5589519650655,
+ "grad_norm": 0.017340239137411118,
+ "learning_rate": 0.0006,
+ "loss": 4.113290786743164,
+ "step": 2704
+ },
+ {
+ "epoch": 37.57292576419214,
+ "grad_norm": 0.015628693625330925,
+ "learning_rate": 0.0006,
+ "loss": 4.1483001708984375,
+ "step": 2705
+ },
+ {
+ "epoch": 37.58689956331878,
+ "grad_norm": 0.016493283212184906,
+ "learning_rate": 0.0006,
+ "loss": 4.1040143966674805,
+ "step": 2706
+ },
+ {
+ "epoch": 37.60087336244541,
+ "grad_norm": 0.016732443124055862,
+ "learning_rate": 0.0006,
+ "loss": 4.07205057144165,
+ "step": 2707
+ },
+ {
+ "epoch": 37.61484716157205,
+ "grad_norm": 0.017176156863570213,
+ "learning_rate": 0.0006,
+ "loss": 4.1279754638671875,
+ "step": 2708
+ },
+ {
+ "epoch": 37.62882096069869,
+ "grad_norm": 0.015730608254671097,
+ "learning_rate": 0.0006,
+ "loss": 4.086658000946045,
+ "step": 2709
+ },
+ {
+ "epoch": 37.64279475982533,
+ "grad_norm": 0.016111090779304504,
+ "learning_rate": 0.0006,
+ "loss": 4.064335823059082,
+ "step": 2710
+ },
+ {
+ "epoch": 37.65676855895197,
+ "grad_norm": 0.01769379712641239,
+ "learning_rate": 0.0006,
+ "loss": 4.116047382354736,
+ "step": 2711
+ },
+ {
+ "epoch": 37.670742358078606,
+ "grad_norm": 0.01683102548122406,
+ "learning_rate": 0.0006,
+ "loss": 4.095986843109131,
+ "step": 2712
+ },
+ {
+ "epoch": 37.68471615720524,
+ "grad_norm": 0.015726450830698013,
+ "learning_rate": 0.0006,
+ "loss": 4.10775089263916,
+ "step": 2713
+ },
+ {
+ "epoch": 37.698689956331876,
+ "grad_norm": 0.014760235324501991,
+ "learning_rate": 0.0006,
+ "loss": 4.1234893798828125,
+ "step": 2714
+ },
+ {
+ "epoch": 37.712663755458514,
+ "grad_norm": 0.014933438040316105,
+ "learning_rate": 0.0006,
+ "loss": 4.079200744628906,
+ "step": 2715
+ },
+ {
+ "epoch": 37.72663755458515,
+ "grad_norm": 0.015723997727036476,
+ "learning_rate": 0.0006,
+ "loss": 4.07865047454834,
+ "step": 2716
+ },
+ {
+ "epoch": 37.74061135371179,
+ "grad_norm": 0.014539001509547234,
+ "learning_rate": 0.0006,
+ "loss": 4.036242485046387,
+ "step": 2717
+ },
+ {
+ "epoch": 37.75458515283843,
+ "grad_norm": 0.014844979159533978,
+ "learning_rate": 0.0006,
+ "loss": 4.1051483154296875,
+ "step": 2718
+ },
+ {
+ "epoch": 37.76855895196506,
+ "grad_norm": 0.015939844772219658,
+ "learning_rate": 0.0006,
+ "loss": 4.017253875732422,
+ "step": 2719
+ },
+ {
+ "epoch": 37.7825327510917,
+ "grad_norm": 0.014771555550396442,
+ "learning_rate": 0.0006,
+ "loss": 4.0694661140441895,
+ "step": 2720
+ },
+ {
+ "epoch": 37.79650655021834,
+ "grad_norm": 0.015777921304106712,
+ "learning_rate": 0.0006,
+ "loss": 3.947587490081787,
+ "step": 2721
+ },
+ {
+ "epoch": 37.81048034934498,
+ "grad_norm": 0.017451291903853416,
+ "learning_rate": 0.0006,
+ "loss": 4.093625068664551,
+ "step": 2722
+ },
+ {
+ "epoch": 37.82445414847162,
+ "grad_norm": 0.016829246655106544,
+ "learning_rate": 0.0006,
+ "loss": 4.09030818939209,
+ "step": 2723
+ },
+ {
+ "epoch": 37.838427947598255,
+ "grad_norm": 0.016984397545456886,
+ "learning_rate": 0.0006,
+ "loss": 4.140626907348633,
+ "step": 2724
+ },
+ {
+ "epoch": 37.852401746724894,
+ "grad_norm": 0.01732603646814823,
+ "learning_rate": 0.0006,
+ "loss": 4.128276824951172,
+ "step": 2725
+ },
+ {
+ "epoch": 37.866375545851525,
+ "grad_norm": 0.01744619384407997,
+ "learning_rate": 0.0006,
+ "loss": 4.103130340576172,
+ "step": 2726
+ },
+ {
+ "epoch": 37.880349344978164,
+ "grad_norm": 0.016331320628523827,
+ "learning_rate": 0.0006,
+ "loss": 4.037951469421387,
+ "step": 2727
+ },
+ {
+ "epoch": 37.8943231441048,
+ "grad_norm": 0.015919430181384087,
+ "learning_rate": 0.0006,
+ "loss": 4.121513366699219,
+ "step": 2728
+ },
+ {
+ "epoch": 37.90829694323144,
+ "grad_norm": 0.016153695061802864,
+ "learning_rate": 0.0006,
+ "loss": 4.099384307861328,
+ "step": 2729
+ },
+ {
+ "epoch": 37.92227074235808,
+ "grad_norm": 0.01628800667822361,
+ "learning_rate": 0.0006,
+ "loss": 4.092325687408447,
+ "step": 2730
+ },
+ {
+ "epoch": 37.93624454148472,
+ "grad_norm": 0.017285684123635292,
+ "learning_rate": 0.0006,
+ "loss": 3.9828906059265137,
+ "step": 2731
+ },
+ {
+ "epoch": 37.95021834061135,
+ "grad_norm": 0.016047554090619087,
+ "learning_rate": 0.0006,
+ "loss": 4.100419044494629,
+ "step": 2732
+ },
+ {
+ "epoch": 37.96419213973799,
+ "grad_norm": 0.01562986709177494,
+ "learning_rate": 0.0006,
+ "loss": 4.065832614898682,
+ "step": 2733
+ },
+ {
+ "epoch": 37.97816593886463,
+ "grad_norm": 0.01677163504064083,
+ "learning_rate": 0.0006,
+ "loss": 4.077518463134766,
+ "step": 2734
+ },
+ {
+ "epoch": 37.992139737991266,
+ "grad_norm": 0.01755492016673088,
+ "learning_rate": 0.0006,
+ "loss": 4.1076154708862305,
+ "step": 2735
+ },
+ {
+ "epoch": 38.0,
+ "grad_norm": 0.01929762214422226,
+ "learning_rate": 0.0006,
+ "loss": 4.09603214263916,
+ "step": 2736
+ },
+ {
+ "epoch": 38.0,
+ "eval_loss": 4.582442283630371,
+ "eval_runtime": 62.8649,
+ "eval_samples_per_second": 38.845,
+ "eval_steps_per_second": 1.225,
+ "step": 2736
+ },
+ {
+ "epoch": 38.01397379912664,
+ "grad_norm": 0.019129136577248573,
+ "learning_rate": 0.0006,
+ "loss": 4.095005989074707,
+ "step": 2737
+ },
+ {
+ "epoch": 38.02794759825328,
+ "grad_norm": 0.019167408347129822,
+ "learning_rate": 0.0006,
+ "loss": 4.054426670074463,
+ "step": 2738
+ },
+ {
+ "epoch": 38.041921397379916,
+ "grad_norm": 0.01789460889995098,
+ "learning_rate": 0.0006,
+ "loss": 4.0434417724609375,
+ "step": 2739
+ },
+ {
+ "epoch": 38.05589519650655,
+ "grad_norm": 0.015947027131915092,
+ "learning_rate": 0.0006,
+ "loss": 4.074516296386719,
+ "step": 2740
+ },
+ {
+ "epoch": 38.069868995633186,
+ "grad_norm": 0.016667252406477928,
+ "learning_rate": 0.0006,
+ "loss": 4.108545303344727,
+ "step": 2741
+ },
+ {
+ "epoch": 38.083842794759825,
+ "grad_norm": 0.016014399006962776,
+ "learning_rate": 0.0006,
+ "loss": 4.026925086975098,
+ "step": 2742
+ },
+ {
+ "epoch": 38.09781659388646,
+ "grad_norm": 0.015397695824503899,
+ "learning_rate": 0.0006,
+ "loss": 4.163970470428467,
+ "step": 2743
+ },
+ {
+ "epoch": 38.1117903930131,
+ "grad_norm": 0.01563289947807789,
+ "learning_rate": 0.0006,
+ "loss": 4.116578578948975,
+ "step": 2744
+ },
+ {
+ "epoch": 38.12576419213974,
+ "grad_norm": 0.015512318350374699,
+ "learning_rate": 0.0006,
+ "loss": 4.096311569213867,
+ "step": 2745
+ },
+ {
+ "epoch": 38.13973799126637,
+ "grad_norm": 0.015426642261445522,
+ "learning_rate": 0.0006,
+ "loss": 4.017065048217773,
+ "step": 2746
+ },
+ {
+ "epoch": 38.15371179039301,
+ "grad_norm": 0.015115435235202312,
+ "learning_rate": 0.0006,
+ "loss": 4.005434989929199,
+ "step": 2747
+ },
+ {
+ "epoch": 38.16768558951965,
+ "grad_norm": 0.015235894359648228,
+ "learning_rate": 0.0006,
+ "loss": 3.982743263244629,
+ "step": 2748
+ },
+ {
+ "epoch": 38.18165938864629,
+ "grad_norm": 0.016095101833343506,
+ "learning_rate": 0.0006,
+ "loss": 4.0849385261535645,
+ "step": 2749
+ },
+ {
+ "epoch": 38.19563318777293,
+ "grad_norm": 0.015201197937130928,
+ "learning_rate": 0.0006,
+ "loss": 4.103863716125488,
+ "step": 2750
+ },
+ {
+ "epoch": 38.209606986899566,
+ "grad_norm": 0.015687478706240654,
+ "learning_rate": 0.0006,
+ "loss": 4.12074089050293,
+ "step": 2751
+ },
+ {
+ "epoch": 38.223580786026204,
+ "grad_norm": 0.015201589092612267,
+ "learning_rate": 0.0006,
+ "loss": 4.056872367858887,
+ "step": 2752
+ },
+ {
+ "epoch": 38.237554585152836,
+ "grad_norm": 0.015154894441366196,
+ "learning_rate": 0.0006,
+ "loss": 4.089670181274414,
+ "step": 2753
+ },
+ {
+ "epoch": 38.251528384279474,
+ "grad_norm": 0.01586608588695526,
+ "learning_rate": 0.0006,
+ "loss": 4.155317306518555,
+ "step": 2754
+ },
+ {
+ "epoch": 38.26550218340611,
+ "grad_norm": 0.01609654724597931,
+ "learning_rate": 0.0006,
+ "loss": 3.995800495147705,
+ "step": 2755
+ },
+ {
+ "epoch": 38.27947598253275,
+ "grad_norm": 0.015495185740292072,
+ "learning_rate": 0.0006,
+ "loss": 3.995769739151001,
+ "step": 2756
+ },
+ {
+ "epoch": 38.29344978165939,
+ "grad_norm": 0.017042381688952446,
+ "learning_rate": 0.0006,
+ "loss": 4.132832050323486,
+ "step": 2757
+ },
+ {
+ "epoch": 38.30742358078603,
+ "grad_norm": 0.018706614151597023,
+ "learning_rate": 0.0006,
+ "loss": 4.038820743560791,
+ "step": 2758
+ },
+ {
+ "epoch": 38.32139737991266,
+ "grad_norm": 0.018316026777029037,
+ "learning_rate": 0.0006,
+ "loss": 3.972052574157715,
+ "step": 2759
+ },
+ {
+ "epoch": 38.3353711790393,
+ "grad_norm": 0.018375389277935028,
+ "learning_rate": 0.0006,
+ "loss": 4.152097702026367,
+ "step": 2760
+ },
+ {
+ "epoch": 38.34934497816594,
+ "grad_norm": 0.016879886388778687,
+ "learning_rate": 0.0006,
+ "loss": 4.020869731903076,
+ "step": 2761
+ },
+ {
+ "epoch": 38.36331877729258,
+ "grad_norm": 0.015728183090686798,
+ "learning_rate": 0.0006,
+ "loss": 4.008016586303711,
+ "step": 2762
+ },
+ {
+ "epoch": 38.377292576419215,
+ "grad_norm": 0.016147758811712265,
+ "learning_rate": 0.0006,
+ "loss": 4.035854339599609,
+ "step": 2763
+ },
+ {
+ "epoch": 38.391266375545854,
+ "grad_norm": 0.017774797976017,
+ "learning_rate": 0.0006,
+ "loss": 4.122506141662598,
+ "step": 2764
+ },
+ {
+ "epoch": 38.40524017467249,
+ "grad_norm": 0.016590069979429245,
+ "learning_rate": 0.0006,
+ "loss": 4.082847595214844,
+ "step": 2765
+ },
+ {
+ "epoch": 38.419213973799124,
+ "grad_norm": 0.016544528305530548,
+ "learning_rate": 0.0006,
+ "loss": 4.184270858764648,
+ "step": 2766
+ },
+ {
+ "epoch": 38.43318777292576,
+ "grad_norm": 0.017972925677895546,
+ "learning_rate": 0.0006,
+ "loss": 3.987558603286743,
+ "step": 2767
+ },
+ {
+ "epoch": 38.4471615720524,
+ "grad_norm": 0.018562227487564087,
+ "learning_rate": 0.0006,
+ "loss": 4.1193461418151855,
+ "step": 2768
+ },
+ {
+ "epoch": 38.46113537117904,
+ "grad_norm": 0.018896685913205147,
+ "learning_rate": 0.0006,
+ "loss": 4.057808876037598,
+ "step": 2769
+ },
+ {
+ "epoch": 38.47510917030568,
+ "grad_norm": 0.016793809831142426,
+ "learning_rate": 0.0006,
+ "loss": 4.069523334503174,
+ "step": 2770
+ },
+ {
+ "epoch": 38.48908296943232,
+ "grad_norm": 0.01698676310479641,
+ "learning_rate": 0.0006,
+ "loss": 3.997624158859253,
+ "step": 2771
+ },
+ {
+ "epoch": 38.50305676855895,
+ "grad_norm": 0.016784202307462692,
+ "learning_rate": 0.0006,
+ "loss": 4.174795627593994,
+ "step": 2772
+ },
+ {
+ "epoch": 38.51703056768559,
+ "grad_norm": 0.016785014420747757,
+ "learning_rate": 0.0006,
+ "loss": 3.999321222305298,
+ "step": 2773
+ },
+ {
+ "epoch": 38.531004366812226,
+ "grad_norm": 0.018615184351801872,
+ "learning_rate": 0.0006,
+ "loss": 4.191237926483154,
+ "step": 2774
+ },
+ {
+ "epoch": 38.544978165938865,
+ "grad_norm": 0.019090790301561356,
+ "learning_rate": 0.0006,
+ "loss": 4.106966018676758,
+ "step": 2775
+ },
+ {
+ "epoch": 38.5589519650655,
+ "grad_norm": 0.018298320472240448,
+ "learning_rate": 0.0006,
+ "loss": 4.063363552093506,
+ "step": 2776
+ },
+ {
+ "epoch": 38.57292576419214,
+ "grad_norm": 0.018541090190410614,
+ "learning_rate": 0.0006,
+ "loss": 4.101193428039551,
+ "step": 2777
+ },
+ {
+ "epoch": 38.58689956331878,
+ "grad_norm": 0.01849970407783985,
+ "learning_rate": 0.0006,
+ "loss": 4.071335792541504,
+ "step": 2778
+ },
+ {
+ "epoch": 38.60087336244541,
+ "grad_norm": 0.018309077247977257,
+ "learning_rate": 0.0006,
+ "loss": 4.245305061340332,
+ "step": 2779
+ },
+ {
+ "epoch": 38.61484716157205,
+ "grad_norm": 0.017681840807199478,
+ "learning_rate": 0.0006,
+ "loss": 4.081173419952393,
+ "step": 2780
+ },
+ {
+ "epoch": 38.62882096069869,
+ "grad_norm": 0.016286224126815796,
+ "learning_rate": 0.0006,
+ "loss": 4.024351596832275,
+ "step": 2781
+ },
+ {
+ "epoch": 38.64279475982533,
+ "grad_norm": 0.015837235376238823,
+ "learning_rate": 0.0006,
+ "loss": 4.028205394744873,
+ "step": 2782
+ },
+ {
+ "epoch": 38.65676855895197,
+ "grad_norm": 0.016062596812844276,
+ "learning_rate": 0.0006,
+ "loss": 4.131122589111328,
+ "step": 2783
+ },
+ {
+ "epoch": 38.670742358078606,
+ "grad_norm": 0.01711239665746689,
+ "learning_rate": 0.0006,
+ "loss": 4.124416351318359,
+ "step": 2784
+ },
+ {
+ "epoch": 38.68471615720524,
+ "grad_norm": 0.015304913744330406,
+ "learning_rate": 0.0006,
+ "loss": 4.056122779846191,
+ "step": 2785
+ },
+ {
+ "epoch": 38.698689956331876,
+ "grad_norm": 0.015971887856721878,
+ "learning_rate": 0.0006,
+ "loss": 4.132555961608887,
+ "step": 2786
+ },
+ {
+ "epoch": 38.712663755458514,
+ "grad_norm": 0.015787221491336823,
+ "learning_rate": 0.0006,
+ "loss": 4.051596164703369,
+ "step": 2787
+ },
+ {
+ "epoch": 38.72663755458515,
+ "grad_norm": 0.01605444774031639,
+ "learning_rate": 0.0006,
+ "loss": 4.191808700561523,
+ "step": 2788
+ },
+ {
+ "epoch": 38.74061135371179,
+ "grad_norm": 0.016383009031414986,
+ "learning_rate": 0.0006,
+ "loss": 4.0648417472839355,
+ "step": 2789
+ },
+ {
+ "epoch": 38.75458515283843,
+ "grad_norm": 0.016135554760694504,
+ "learning_rate": 0.0006,
+ "loss": 3.988563060760498,
+ "step": 2790
+ },
+ {
+ "epoch": 38.76855895196506,
+ "grad_norm": 0.016634158790111542,
+ "learning_rate": 0.0006,
+ "loss": 4.133493423461914,
+ "step": 2791
+ },
+ {
+ "epoch": 38.7825327510917,
+ "grad_norm": 0.01688210666179657,
+ "learning_rate": 0.0006,
+ "loss": 4.091277122497559,
+ "step": 2792
+ },
+ {
+ "epoch": 38.79650655021834,
+ "grad_norm": 0.016102122142910957,
+ "learning_rate": 0.0006,
+ "loss": 4.051219940185547,
+ "step": 2793
+ },
+ {
+ "epoch": 38.81048034934498,
+ "grad_norm": 0.016585038974881172,
+ "learning_rate": 0.0006,
+ "loss": 4.058587074279785,
+ "step": 2794
+ },
+ {
+ "epoch": 38.82445414847162,
+ "grad_norm": 0.017700951546430588,
+ "learning_rate": 0.0006,
+ "loss": 4.016030311584473,
+ "step": 2795
+ },
+ {
+ "epoch": 38.838427947598255,
+ "grad_norm": 0.017922203987836838,
+ "learning_rate": 0.0006,
+ "loss": 4.079977989196777,
+ "step": 2796
+ },
+ {
+ "epoch": 38.852401746724894,
+ "grad_norm": 0.017299890518188477,
+ "learning_rate": 0.0006,
+ "loss": 4.062984943389893,
+ "step": 2797
+ },
+ {
+ "epoch": 38.866375545851525,
+ "grad_norm": 0.018648581579327583,
+ "learning_rate": 0.0006,
+ "loss": 4.065717697143555,
+ "step": 2798
+ },
+ {
+ "epoch": 38.880349344978164,
+ "grad_norm": 0.019057920202612877,
+ "learning_rate": 0.0006,
+ "loss": 4.227892875671387,
+ "step": 2799
+ },
+ {
+ "epoch": 38.8943231441048,
+ "grad_norm": 0.018824422731995583,
+ "learning_rate": 0.0006,
+ "loss": 4.116058349609375,
+ "step": 2800
+ },
+ {
+ "epoch": 38.90829694323144,
+ "grad_norm": 0.019084978848695755,
+ "learning_rate": 0.0006,
+ "loss": 4.061677932739258,
+ "step": 2801
+ },
+ {
+ "epoch": 38.92227074235808,
+ "grad_norm": 0.017842981964349747,
+ "learning_rate": 0.0006,
+ "loss": 4.083813667297363,
+ "step": 2802
+ },
+ {
+ "epoch": 38.93624454148472,
+ "grad_norm": 0.01880858838558197,
+ "learning_rate": 0.0006,
+ "loss": 4.152561187744141,
+ "step": 2803
+ },
+ {
+ "epoch": 38.95021834061135,
+ "grad_norm": 0.017400087788701057,
+ "learning_rate": 0.0006,
+ "loss": 4.154635429382324,
+ "step": 2804
+ },
+ {
+ "epoch": 38.96419213973799,
+ "grad_norm": 0.017682507634162903,
+ "learning_rate": 0.0006,
+ "loss": 4.151128768920898,
+ "step": 2805
+ },
+ {
+ "epoch": 38.97816593886463,
+ "grad_norm": 0.015803052112460136,
+ "learning_rate": 0.0006,
+ "loss": 4.10498571395874,
+ "step": 2806
+ },
+ {
+ "epoch": 38.992139737991266,
+ "grad_norm": 0.01582511141896248,
+ "learning_rate": 0.0006,
+ "loss": 4.140159606933594,
+ "step": 2807
+ },
+ {
+ "epoch": 39.0,
+ "grad_norm": 0.01925666816532612,
+ "learning_rate": 0.0006,
+ "loss": 4.126394271850586,
+ "step": 2808
+ },
+ {
+ "epoch": 39.0,
+ "eval_loss": 4.615617275238037,
+ "eval_runtime": 56.8895,
+ "eval_samples_per_second": 42.925,
+ "eval_steps_per_second": 1.354,
+ "step": 2808
+ },
+ {
+ "epoch": 39.01397379912664,
+ "grad_norm": 0.01677853614091873,
+ "learning_rate": 0.0006,
+ "loss": 4.102395057678223,
+ "step": 2809
+ },
+ {
+ "epoch": 39.02794759825328,
+ "grad_norm": 0.017901137471199036,
+ "learning_rate": 0.0006,
+ "loss": 3.976609230041504,
+ "step": 2810
+ },
+ {
+ "epoch": 39.041921397379916,
+ "grad_norm": 0.017978668212890625,
+ "learning_rate": 0.0006,
+ "loss": 3.944416046142578,
+ "step": 2811
+ },
+ {
+ "epoch": 39.05589519650655,
+ "grad_norm": 0.019967492669820786,
+ "learning_rate": 0.0006,
+ "loss": 3.992234468460083,
+ "step": 2812
+ },
+ {
+ "epoch": 39.069868995633186,
+ "grad_norm": 0.018937544897198677,
+ "learning_rate": 0.0006,
+ "loss": 4.028573036193848,
+ "step": 2813
+ },
+ {
+ "epoch": 39.083842794759825,
+ "grad_norm": 0.016852952539920807,
+ "learning_rate": 0.0006,
+ "loss": 4.062442779541016,
+ "step": 2814
+ },
+ {
+ "epoch": 39.09781659388646,
+ "grad_norm": 0.016643647104501724,
+ "learning_rate": 0.0006,
+ "loss": 4.08871603012085,
+ "step": 2815
+ },
+ {
+ "epoch": 39.1117903930131,
+ "grad_norm": 0.016872866079211235,
+ "learning_rate": 0.0006,
+ "loss": 3.9514245986938477,
+ "step": 2816
+ },
+ {
+ "epoch": 39.12576419213974,
+ "grad_norm": 0.018045669421553612,
+ "learning_rate": 0.0006,
+ "loss": 4.038842678070068,
+ "step": 2817
+ },
+ {
+ "epoch": 39.13973799126637,
+ "grad_norm": 0.017853403463959694,
+ "learning_rate": 0.0006,
+ "loss": 3.93709135055542,
+ "step": 2818
+ },
+ {
+ "epoch": 39.15371179039301,
+ "grad_norm": 0.017507443204522133,
+ "learning_rate": 0.0006,
+ "loss": 3.98665714263916,
+ "step": 2819
+ },
+ {
+ "epoch": 39.16768558951965,
+ "grad_norm": 0.01766853965818882,
+ "learning_rate": 0.0006,
+ "loss": 3.937126636505127,
+ "step": 2820
+ },
+ {
+ "epoch": 39.18165938864629,
+ "grad_norm": 0.016660021618008614,
+ "learning_rate": 0.0006,
+ "loss": 3.8397014141082764,
+ "step": 2821
+ },
+ {
+ "epoch": 39.19563318777293,
+ "grad_norm": 0.017719309777021408,
+ "learning_rate": 0.0006,
+ "loss": 4.003579616546631,
+ "step": 2822
+ },
+ {
+ "epoch": 39.209606986899566,
+ "grad_norm": 0.018234414979815483,
+ "learning_rate": 0.0006,
+ "loss": 4.053859710693359,
+ "step": 2823
+ },
+ {
+ "epoch": 39.223580786026204,
+ "grad_norm": 0.018777024000883102,
+ "learning_rate": 0.0006,
+ "loss": 4.056182861328125,
+ "step": 2824
+ },
+ {
+ "epoch": 39.237554585152836,
+ "grad_norm": 0.017443502321839333,
+ "learning_rate": 0.0006,
+ "loss": 4.025068283081055,
+ "step": 2825
+ },
+ {
+ "epoch": 39.251528384279474,
+ "grad_norm": 0.016500331461429596,
+ "learning_rate": 0.0006,
+ "loss": 4.101631164550781,
+ "step": 2826
+ },
+ {
+ "epoch": 39.26550218340611,
+ "grad_norm": 0.016101544722914696,
+ "learning_rate": 0.0006,
+ "loss": 4.027082443237305,
+ "step": 2827
+ },
+ {
+ "epoch": 39.27947598253275,
+ "grad_norm": 0.015471171587705612,
+ "learning_rate": 0.0006,
+ "loss": 4.022909641265869,
+ "step": 2828
+ },
+ {
+ "epoch": 39.29344978165939,
+ "grad_norm": 0.015779748558998108,
+ "learning_rate": 0.0006,
+ "loss": 4.18796968460083,
+ "step": 2829
+ },
+ {
+ "epoch": 39.30742358078603,
+ "grad_norm": 0.015522679314017296,
+ "learning_rate": 0.0006,
+ "loss": 4.005777359008789,
+ "step": 2830
+ },
+ {
+ "epoch": 39.32139737991266,
+ "grad_norm": 0.015226504765450954,
+ "learning_rate": 0.0006,
+ "loss": 4.108416557312012,
+ "step": 2831
+ },
+ {
+ "epoch": 39.3353711790393,
+ "grad_norm": 0.015095778740942478,
+ "learning_rate": 0.0006,
+ "loss": 4.08029842376709,
+ "step": 2832
+ },
+ {
+ "epoch": 39.34934497816594,
+ "grad_norm": 0.015601031482219696,
+ "learning_rate": 0.0006,
+ "loss": 3.989492416381836,
+ "step": 2833
+ },
+ {
+ "epoch": 39.36331877729258,
+ "grad_norm": 0.014825052581727505,
+ "learning_rate": 0.0006,
+ "loss": 3.998520612716675,
+ "step": 2834
+ },
+ {
+ "epoch": 39.377292576419215,
+ "grad_norm": 0.014821100980043411,
+ "learning_rate": 0.0006,
+ "loss": 4.001777172088623,
+ "step": 2835
+ },
+ {
+ "epoch": 39.391266375545854,
+ "grad_norm": 0.015333754941821098,
+ "learning_rate": 0.0006,
+ "loss": 4.11985969543457,
+ "step": 2836
+ },
+ {
+ "epoch": 39.40524017467249,
+ "grad_norm": 0.015940699726343155,
+ "learning_rate": 0.0006,
+ "loss": 4.025373935699463,
+ "step": 2837
+ },
+ {
+ "epoch": 39.419213973799124,
+ "grad_norm": 0.015096287243068218,
+ "learning_rate": 0.0006,
+ "loss": 4.068551063537598,
+ "step": 2838
+ },
+ {
+ "epoch": 39.43318777292576,
+ "grad_norm": 0.01533548254519701,
+ "learning_rate": 0.0006,
+ "loss": 3.9800801277160645,
+ "step": 2839
+ },
+ {
+ "epoch": 39.4471615720524,
+ "grad_norm": 0.014648732729256153,
+ "learning_rate": 0.0006,
+ "loss": 3.9670190811157227,
+ "step": 2840
+ },
+ {
+ "epoch": 39.46113537117904,
+ "grad_norm": 0.015759671106934547,
+ "learning_rate": 0.0006,
+ "loss": 4.050443649291992,
+ "step": 2841
+ },
+ {
+ "epoch": 39.47510917030568,
+ "grad_norm": 0.016054073348641396,
+ "learning_rate": 0.0006,
+ "loss": 4.22734260559082,
+ "step": 2842
+ },
+ {
+ "epoch": 39.48908296943232,
+ "grad_norm": 0.0160264540463686,
+ "learning_rate": 0.0006,
+ "loss": 4.116174221038818,
+ "step": 2843
+ },
+ {
+ "epoch": 39.50305676855895,
+ "grad_norm": 0.015426610596477985,
+ "learning_rate": 0.0006,
+ "loss": 4.000178337097168,
+ "step": 2844
+ },
+ {
+ "epoch": 39.51703056768559,
+ "grad_norm": 0.01520642638206482,
+ "learning_rate": 0.0006,
+ "loss": 4.218410968780518,
+ "step": 2845
+ },
+ {
+ "epoch": 39.531004366812226,
+ "grad_norm": 0.015305949375033379,
+ "learning_rate": 0.0006,
+ "loss": 4.025891304016113,
+ "step": 2846
+ },
+ {
+ "epoch": 39.544978165938865,
+ "grad_norm": 0.015461564064025879,
+ "learning_rate": 0.0006,
+ "loss": 4.100950717926025,
+ "step": 2847
+ },
+ {
+ "epoch": 39.5589519650655,
+ "grad_norm": 0.01712625101208687,
+ "learning_rate": 0.0006,
+ "loss": 4.0620856285095215,
+ "step": 2848
+ },
+ {
+ "epoch": 39.57292576419214,
+ "grad_norm": 0.017133301123976707,
+ "learning_rate": 0.0006,
+ "loss": 4.097138404846191,
+ "step": 2849
+ },
+ {
+ "epoch": 39.58689956331878,
+ "grad_norm": 0.01683625392615795,
+ "learning_rate": 0.0006,
+ "loss": 4.084364891052246,
+ "step": 2850
+ },
+ {
+ "epoch": 39.60087336244541,
+ "grad_norm": 0.0165710486471653,
+ "learning_rate": 0.0006,
+ "loss": 4.042361259460449,
+ "step": 2851
+ },
+ {
+ "epoch": 39.61484716157205,
+ "grad_norm": 0.01660393364727497,
+ "learning_rate": 0.0006,
+ "loss": 4.1395487785339355,
+ "step": 2852
+ },
+ {
+ "epoch": 39.62882096069869,
+ "grad_norm": 0.017025845125317574,
+ "learning_rate": 0.0006,
+ "loss": 4.1207075119018555,
+ "step": 2853
+ },
+ {
+ "epoch": 39.64279475982533,
+ "grad_norm": 0.01660497486591339,
+ "learning_rate": 0.0006,
+ "loss": 4.103846073150635,
+ "step": 2854
+ },
+ {
+ "epoch": 39.65676855895197,
+ "grad_norm": 0.01607946865260601,
+ "learning_rate": 0.0006,
+ "loss": 4.189016342163086,
+ "step": 2855
+ },
+ {
+ "epoch": 39.670742358078606,
+ "grad_norm": 0.01614505983889103,
+ "learning_rate": 0.0006,
+ "loss": 4.079860687255859,
+ "step": 2856
+ },
+ {
+ "epoch": 39.68471615720524,
+ "grad_norm": 0.017085086554288864,
+ "learning_rate": 0.0006,
+ "loss": 3.9774293899536133,
+ "step": 2857
+ },
+ {
+ "epoch": 39.698689956331876,
+ "grad_norm": 0.01816520094871521,
+ "learning_rate": 0.0006,
+ "loss": 4.113666534423828,
+ "step": 2858
+ },
+ {
+ "epoch": 39.712663755458514,
+ "grad_norm": 0.016597526147961617,
+ "learning_rate": 0.0006,
+ "loss": 4.068242073059082,
+ "step": 2859
+ },
+ {
+ "epoch": 39.72663755458515,
+ "grad_norm": 0.01672235131263733,
+ "learning_rate": 0.0006,
+ "loss": 4.024444580078125,
+ "step": 2860
+ },
+ {
+ "epoch": 39.74061135371179,
+ "grad_norm": 0.020093556493520737,
+ "learning_rate": 0.0006,
+ "loss": 4.064124584197998,
+ "step": 2861
+ },
+ {
+ "epoch": 39.75458515283843,
+ "grad_norm": 0.023141304031014442,
+ "learning_rate": 0.0006,
+ "loss": 4.130439758300781,
+ "step": 2862
+ },
+ {
+ "epoch": 39.76855895196506,
+ "grad_norm": 0.02289372682571411,
+ "learning_rate": 0.0006,
+ "loss": 4.14794921875,
+ "step": 2863
+ },
+ {
+ "epoch": 39.7825327510917,
+ "grad_norm": 0.01919020712375641,
+ "learning_rate": 0.0006,
+ "loss": 4.123961925506592,
+ "step": 2864
+ },
+ {
+ "epoch": 39.79650655021834,
+ "grad_norm": 0.016932329162955284,
+ "learning_rate": 0.0006,
+ "loss": 4.0944132804870605,
+ "step": 2865
+ },
+ {
+ "epoch": 39.81048034934498,
+ "grad_norm": 0.016644535586237907,
+ "learning_rate": 0.0006,
+ "loss": 4.124269962310791,
+ "step": 2866
+ },
+ {
+ "epoch": 39.82445414847162,
+ "grad_norm": 0.016302216798067093,
+ "learning_rate": 0.0006,
+ "loss": 4.0267205238342285,
+ "step": 2867
+ },
+ {
+ "epoch": 39.838427947598255,
+ "grad_norm": 0.01673300378024578,
+ "learning_rate": 0.0006,
+ "loss": 4.10650634765625,
+ "step": 2868
+ },
+ {
+ "epoch": 39.852401746724894,
+ "grad_norm": 0.016581356525421143,
+ "learning_rate": 0.0006,
+ "loss": 4.1547956466674805,
+ "step": 2869
+ },
+ {
+ "epoch": 39.866375545851525,
+ "grad_norm": 0.01691797934472561,
+ "learning_rate": 0.0006,
+ "loss": 4.056936740875244,
+ "step": 2870
+ },
+ {
+ "epoch": 39.880349344978164,
+ "grad_norm": 0.01596726104617119,
+ "learning_rate": 0.0006,
+ "loss": 3.9891414642333984,
+ "step": 2871
+ },
+ {
+ "epoch": 39.8943231441048,
+ "grad_norm": 0.015615091659128666,
+ "learning_rate": 0.0006,
+ "loss": 4.137948036193848,
+ "step": 2872
+ },
+ {
+ "epoch": 39.90829694323144,
+ "grad_norm": 0.015703413635492325,
+ "learning_rate": 0.0006,
+ "loss": 3.9844930171966553,
+ "step": 2873
+ },
+ {
+ "epoch": 39.92227074235808,
+ "grad_norm": 0.015453094616532326,
+ "learning_rate": 0.0006,
+ "loss": 4.098443508148193,
+ "step": 2874
+ },
+ {
+ "epoch": 39.93624454148472,
+ "grad_norm": 0.01449158601462841,
+ "learning_rate": 0.0006,
+ "loss": 4.090852737426758,
+ "step": 2875
+ },
+ {
+ "epoch": 39.95021834061135,
+ "grad_norm": 0.01537515688687563,
+ "learning_rate": 0.0006,
+ "loss": 4.239612579345703,
+ "step": 2876
+ },
+ {
+ "epoch": 39.96419213973799,
+ "grad_norm": 0.015989921987056732,
+ "learning_rate": 0.0006,
+ "loss": 4.142073154449463,
+ "step": 2877
+ },
+ {
+ "epoch": 39.97816593886463,
+ "grad_norm": 0.015985611826181412,
+ "learning_rate": 0.0006,
+ "loss": 4.100076675415039,
+ "step": 2878
+ },
+ {
+ "epoch": 39.992139737991266,
+ "grad_norm": 0.015260833315551281,
+ "learning_rate": 0.0006,
+ "loss": 4.068576335906982,
+ "step": 2879
+ },
+ {
+ "epoch": 40.0,
+ "grad_norm": 0.018047941848635674,
+ "learning_rate": 0.0006,
+ "loss": 4.005878925323486,
+ "step": 2880
+ },
+ {
+ "epoch": 40.0,
+ "eval_loss": 4.622848987579346,
+ "eval_runtime": 56.9618,
+ "eval_samples_per_second": 42.871,
+ "eval_steps_per_second": 1.352,
+ "step": 2880
+ },
+ {
+ "epoch": 40.01397379912664,
+ "grad_norm": 0.015976838767528534,
+ "learning_rate": 0.0006,
+ "loss": 4.035689353942871,
+ "step": 2881
+ },
+ {
+ "epoch": 40.02794759825328,
+ "grad_norm": 0.015176220797002316,
+ "learning_rate": 0.0006,
+ "loss": 3.932197332382202,
+ "step": 2882
+ },
+ {
+ "epoch": 40.041921397379916,
+ "grad_norm": 0.016676228493452072,
+ "learning_rate": 0.0006,
+ "loss": 4.014171123504639,
+ "step": 2883
+ },
+ {
+ "epoch": 40.05589519650655,
+ "grad_norm": 0.01698281429708004,
+ "learning_rate": 0.0006,
+ "loss": 4.117893218994141,
+ "step": 2884
+ },
+ {
+ "epoch": 40.069868995633186,
+ "grad_norm": 0.017872486263513565,
+ "learning_rate": 0.0006,
+ "loss": 4.132119178771973,
+ "step": 2885
+ },
+ {
+ "epoch": 40.083842794759825,
+ "grad_norm": 0.017615554854273796,
+ "learning_rate": 0.0006,
+ "loss": 4.072000980377197,
+ "step": 2886
+ },
+ {
+ "epoch": 40.09781659388646,
+ "grad_norm": 0.017469394952058792,
+ "learning_rate": 0.0006,
+ "loss": 4.017726421356201,
+ "step": 2887
+ },
+ {
+ "epoch": 40.1117903930131,
+ "grad_norm": 0.01954762637615204,
+ "learning_rate": 0.0006,
+ "loss": 4.014443874359131,
+ "step": 2888
+ },
+ {
+ "epoch": 40.12576419213974,
+ "grad_norm": 0.018935853615403175,
+ "learning_rate": 0.0006,
+ "loss": 4.0042948722839355,
+ "step": 2889
+ },
+ {
+ "epoch": 40.13973799126637,
+ "grad_norm": 0.019382692873477936,
+ "learning_rate": 0.0006,
+ "loss": 4.06160831451416,
+ "step": 2890
+ },
+ {
+ "epoch": 40.15371179039301,
+ "grad_norm": 0.019677115604281425,
+ "learning_rate": 0.0006,
+ "loss": 4.13466215133667,
+ "step": 2891
+ },
+ {
+ "epoch": 40.16768558951965,
+ "grad_norm": 0.020365122705698013,
+ "learning_rate": 0.0006,
+ "loss": 4.095000267028809,
+ "step": 2892
+ },
+ {
+ "epoch": 40.18165938864629,
+ "grad_norm": 0.021842138841748238,
+ "learning_rate": 0.0006,
+ "loss": 3.9877853393554688,
+ "step": 2893
+ },
+ {
+ "epoch": 40.19563318777293,
+ "grad_norm": 0.02354881353676319,
+ "learning_rate": 0.0006,
+ "loss": 3.9430112838745117,
+ "step": 2894
+ },
+ {
+ "epoch": 40.209606986899566,
+ "grad_norm": 0.026602735742926598,
+ "learning_rate": 0.0006,
+ "loss": 4.012076377868652,
+ "step": 2895
+ },
+ {
+ "epoch": 40.223580786026204,
+ "grad_norm": 0.023976484313607216,
+ "learning_rate": 0.0006,
+ "loss": 3.9768595695495605,
+ "step": 2896
+ },
+ {
+ "epoch": 40.237554585152836,
+ "grad_norm": 0.02336616814136505,
+ "learning_rate": 0.0006,
+ "loss": 4.1096038818359375,
+ "step": 2897
+ },
+ {
+ "epoch": 40.251528384279474,
+ "grad_norm": 0.02426023595035076,
+ "learning_rate": 0.0006,
+ "loss": 4.157189846038818,
+ "step": 2898
+ },
+ {
+ "epoch": 40.26550218340611,
+ "grad_norm": 0.021572696045041084,
+ "learning_rate": 0.0006,
+ "loss": 3.979175090789795,
+ "step": 2899
+ },
+ {
+ "epoch": 40.27947598253275,
+ "grad_norm": 0.08300432562828064,
+ "learning_rate": 0.0006,
+ "loss": 4.092336654663086,
+ "step": 2900
+ },
+ {
+ "epoch": 40.29344978165939,
+ "grad_norm": 0.7323967814445496,
+ "learning_rate": 0.0006,
+ "loss": 5.337990760803223,
+ "step": 2901
+ },
+ {
+ "epoch": 40.30742358078603,
+ "grad_norm": 0.3363337814807892,
+ "learning_rate": 0.0006,
+ "loss": 6.40043830871582,
+ "step": 2902
+ },
+ {
+ "epoch": 40.32139737991266,
+ "grad_norm": 0.598686933517456,
+ "learning_rate": 0.0006,
+ "loss": 7.004765510559082,
+ "step": 2903
+ },
+ {
+ "epoch": 40.3353711790393,
+ "grad_norm": 0.5998364686965942,
+ "learning_rate": 0.0006,
+ "loss": 8.184747695922852,
+ "step": 2904
+ },
+ {
+ "epoch": 40.34934497816594,
+ "grad_norm": 1.408208966255188,
+ "learning_rate": 0.0006,
+ "loss": 8.88978385925293,
+ "step": 2905
+ },
+ {
+ "epoch": 40.36331877729258,
+ "grad_norm": 0.7829785943031311,
+ "learning_rate": 0.0006,
+ "loss": 11.13601303100586,
+ "step": 2906
+ },
+ {
+ "epoch": 40.377292576419215,
+ "grad_norm": 0.7769215106964111,
+ "learning_rate": 0.0006,
+ "loss": 9.90416145324707,
+ "step": 2907
+ },
+ {
+ "epoch": 40.391266375545854,
+ "grad_norm": 0.46508079767227173,
+ "learning_rate": 0.0006,
+ "loss": 9.647209167480469,
+ "step": 2908
+ },
+ {
+ "epoch": 40.40524017467249,
+ "grad_norm": 1.0228968858718872,
+ "learning_rate": 0.0006,
+ "loss": 11.685434341430664,
+ "step": 2909
+ },
+ {
+ "epoch": 40.419213973799124,
+ "grad_norm": 0.20553778111934662,
+ "learning_rate": 0.0006,
+ "loss": 8.621708869934082,
+ "step": 2910
+ },
+ {
+ "epoch": 40.43318777292576,
+ "grad_norm": 0.18466424942016602,
+ "learning_rate": 0.0006,
+ "loss": 8.590787887573242,
+ "step": 2911
+ },
+ {
+ "epoch": 40.4471615720524,
+ "grad_norm": 0.10458594560623169,
+ "learning_rate": 0.0006,
+ "loss": 8.324458122253418,
+ "step": 2912
+ },
+ {
+ "epoch": 40.46113537117904,
+ "grad_norm": 0.1096842959523201,
+ "learning_rate": 0.0006,
+ "loss": 8.300851821899414,
+ "step": 2913
+ },
+ {
+ "epoch": 40.47510917030568,
+ "grad_norm": 0.10501226782798767,
+ "learning_rate": 0.0006,
+ "loss": 8.21009635925293,
+ "step": 2914
+ },
+ {
+ "epoch": 40.48908296943232,
+ "grad_norm": 0.13704165816307068,
+ "learning_rate": 0.0006,
+ "loss": 8.141355514526367,
+ "step": 2915
+ },
+ {
+ "epoch": 40.50305676855895,
+ "grad_norm": 0.12129108607769012,
+ "learning_rate": 0.0006,
+ "loss": 8.074612617492676,
+ "step": 2916
+ },
+ {
+ "epoch": 40.51703056768559,
+ "grad_norm": 0.10028897225856781,
+ "learning_rate": 0.0006,
+ "loss": 8.00384521484375,
+ "step": 2917
+ },
+ {
+ "epoch": 40.531004366812226,
+ "grad_norm": 0.11764246970415115,
+ "learning_rate": 0.0006,
+ "loss": 8.078598022460938,
+ "step": 2918
+ },
+ {
+ "epoch": 40.544978165938865,
+ "grad_norm": 0.10081959515810013,
+ "learning_rate": 0.0006,
+ "loss": 8.008938789367676,
+ "step": 2919
+ },
+ {
+ "epoch": 40.5589519650655,
+ "grad_norm": 0.1003170907497406,
+ "learning_rate": 0.0006,
+ "loss": 7.962940216064453,
+ "step": 2920
+ },
+ {
+ "epoch": 40.57292576419214,
+ "grad_norm": 0.08763492107391357,
+ "learning_rate": 0.0006,
+ "loss": 7.909419536590576,
+ "step": 2921
+ },
+ {
+ "epoch": 40.58689956331878,
+ "grad_norm": 0.0636807233095169,
+ "learning_rate": 0.0006,
+ "loss": 7.878905773162842,
+ "step": 2922
+ },
+ {
+ "epoch": 40.60087336244541,
+ "grad_norm": 0.06005936488509178,
+ "learning_rate": 0.0006,
+ "loss": 7.850980758666992,
+ "step": 2923
+ },
+ {
+ "epoch": 40.61484716157205,
+ "grad_norm": 0.06415296345949173,
+ "learning_rate": 0.0006,
+ "loss": 7.830353260040283,
+ "step": 2924
+ },
+ {
+ "epoch": 40.62882096069869,
+ "grad_norm": 0.0732007846236229,
+ "learning_rate": 0.0006,
+ "loss": 7.830746650695801,
+ "step": 2925
+ },
+ {
+ "epoch": 40.64279475982533,
+ "grad_norm": 0.06628110259771347,
+ "learning_rate": 0.0006,
+ "loss": 7.82661771774292,
+ "step": 2926
+ },
+ {
+ "epoch": 40.65676855895197,
+ "grad_norm": 0.05544218048453331,
+ "learning_rate": 0.0006,
+ "loss": 7.814305782318115,
+ "step": 2927
+ },
+ {
+ "epoch": 40.670742358078606,
+ "grad_norm": 0.050327103585004807,
+ "learning_rate": 0.0006,
+ "loss": 7.764055252075195,
+ "step": 2928
+ },
+ {
+ "epoch": 40.68471615720524,
+ "grad_norm": 0.04918425530195236,
+ "learning_rate": 0.0006,
+ "loss": 7.744625091552734,
+ "step": 2929
+ },
+ {
+ "epoch": 40.698689956331876,
+ "grad_norm": 0.042862266302108765,
+ "learning_rate": 0.0006,
+ "loss": 7.764426231384277,
+ "step": 2930
+ },
+ {
+ "epoch": 40.712663755458514,
+ "grad_norm": 0.04338089004158974,
+ "learning_rate": 0.0006,
+ "loss": 7.712530136108398,
+ "step": 2931
+ },
+ {
+ "epoch": 40.72663755458515,
+ "grad_norm": 0.04127737879753113,
+ "learning_rate": 0.0006,
+ "loss": 7.742180347442627,
+ "step": 2932
+ },
+ {
+ "epoch": 40.74061135371179,
+ "grad_norm": 0.03358631581068039,
+ "learning_rate": 0.0006,
+ "loss": 7.714834213256836,
+ "step": 2933
+ },
+ {
+ "epoch": 40.75458515283843,
+ "grad_norm": 0.02845790982246399,
+ "learning_rate": 0.0006,
+ "loss": 7.7021026611328125,
+ "step": 2934
+ },
+ {
+ "epoch": 40.76855895196506,
+ "grad_norm": 0.03012419492006302,
+ "learning_rate": 0.0006,
+ "loss": 7.712377548217773,
+ "step": 2935
+ },
+ {
+ "epoch": 40.7825327510917,
+ "grad_norm": 0.032981179654598236,
+ "learning_rate": 0.0006,
+ "loss": 7.666385650634766,
+ "step": 2936
+ },
+ {
+ "epoch": 40.79650655021834,
+ "grad_norm": 0.03357074037194252,
+ "learning_rate": 0.0006,
+ "loss": 7.649097919464111,
+ "step": 2937
+ },
+ {
+ "epoch": 40.81048034934498,
+ "grad_norm": 0.030914708971977234,
+ "learning_rate": 0.0006,
+ "loss": 7.649196624755859,
+ "step": 2938
+ },
+ {
+ "epoch": 40.82445414847162,
+ "grad_norm": 0.026296131312847137,
+ "learning_rate": 0.0006,
+ "loss": 7.629964828491211,
+ "step": 2939
+ },
+ {
+ "epoch": 40.838427947598255,
+ "grad_norm": 0.026677245274186134,
+ "learning_rate": 0.0006,
+ "loss": 7.599973678588867,
+ "step": 2940
+ },
+ {
+ "epoch": 40.852401746724894,
+ "grad_norm": 0.023208219558000565,
+ "learning_rate": 0.0006,
+ "loss": 7.601624488830566,
+ "step": 2941
+ },
+ {
+ "epoch": 40.866375545851525,
+ "grad_norm": 0.020741475746035576,
+ "learning_rate": 0.0006,
+ "loss": 7.603902339935303,
+ "step": 2942
+ },
+ {
+ "epoch": 40.880349344978164,
+ "grad_norm": 0.019221751019358635,
+ "learning_rate": 0.0006,
+ "loss": 7.5535888671875,
+ "step": 2943
+ },
+ {
+ "epoch": 40.8943231441048,
+ "grad_norm": 0.019491951912641525,
+ "learning_rate": 0.0006,
+ "loss": 7.54597282409668,
+ "step": 2944
+ },
+ {
+ "epoch": 40.90829694323144,
+ "grad_norm": 0.019702572375535965,
+ "learning_rate": 0.0006,
+ "loss": 7.537567138671875,
+ "step": 2945
+ },
+ {
+ "epoch": 40.92227074235808,
+ "grad_norm": 0.01768800988793373,
+ "learning_rate": 0.0006,
+ "loss": 7.512800216674805,
+ "step": 2946
+ },
+ {
+ "epoch": 40.93624454148472,
+ "grad_norm": 0.01926138624548912,
+ "learning_rate": 0.0006,
+ "loss": 7.509932518005371,
+ "step": 2947
+ },
+ {
+ "epoch": 40.95021834061135,
+ "grad_norm": 0.019030006602406502,
+ "learning_rate": 0.0006,
+ "loss": 7.497983932495117,
+ "step": 2948
+ },
+ {
+ "epoch": 40.96419213973799,
+ "grad_norm": 0.017313841730356216,
+ "learning_rate": 0.0006,
+ "loss": 7.463518142700195,
+ "step": 2949
+ },
+ {
+ "epoch": 40.97816593886463,
+ "grad_norm": 0.01310762483626604,
+ "learning_rate": 0.0006,
+ "loss": 7.474668502807617,
+ "step": 2950
+ },
+ {
+ "epoch": 40.992139737991266,
+ "grad_norm": 0.013008282519876957,
+ "learning_rate": 0.0006,
+ "loss": 7.45982551574707,
+ "step": 2951
+ },
+ {
+ "epoch": 41.0,
+ "grad_norm": 0.01441363524645567,
+ "learning_rate": 0.0006,
+ "loss": 7.446529388427734,
+ "step": 2952
+ },
+ {
+ "epoch": 41.0,
+ "eval_loss": 7.486120223999023,
+ "eval_runtime": 70.4986,
+ "eval_samples_per_second": 34.639,
+ "eval_steps_per_second": 1.092,
+ "step": 2952
+ },
+ {
+ "epoch": 41.01397379912664,
+ "grad_norm": 0.014061855152249336,
+ "learning_rate": 0.0006,
+ "loss": 7.4692182540893555,
+ "step": 2953
+ },
+ {
+ "epoch": 41.02794759825328,
+ "grad_norm": 0.01587500050663948,
+ "learning_rate": 0.0006,
+ "loss": 7.4182024002075195,
+ "step": 2954
+ },
+ {
+ "epoch": 41.041921397379916,
+ "grad_norm": 0.014783984050154686,
+ "learning_rate": 0.0006,
+ "loss": 7.466778755187988,
+ "step": 2955
+ },
+ {
+ "epoch": 41.05589519650655,
+ "grad_norm": 0.011560854502022266,
+ "learning_rate": 0.0006,
+ "loss": 7.438002586364746,
+ "step": 2956
+ },
+ {
+ "epoch": 41.069868995633186,
+ "grad_norm": 0.00970957800745964,
+ "learning_rate": 0.0006,
+ "loss": 7.442694187164307,
+ "step": 2957
+ },
+ {
+ "epoch": 41.083842794759825,
+ "grad_norm": 0.009633459150791168,
+ "learning_rate": 0.0006,
+ "loss": 7.458884239196777,
+ "step": 2958
+ },
+ {
+ "epoch": 41.09781659388646,
+ "grad_norm": 0.010664350353181362,
+ "learning_rate": 0.0006,
+ "loss": 7.48378324508667,
+ "step": 2959
+ },
+ {
+ "epoch": 41.1117903930131,
+ "grad_norm": 0.011440591886639595,
+ "learning_rate": 0.0006,
+ "loss": 7.474484920501709,
+ "step": 2960
+ },
+ {
+ "epoch": 41.12576419213974,
+ "grad_norm": 0.01150908600538969,
+ "learning_rate": 0.0006,
+ "loss": 7.417902946472168,
+ "step": 2961
+ },
+ {
+ "epoch": 41.13973799126637,
+ "grad_norm": 0.010375697165727615,
+ "learning_rate": 0.0006,
+ "loss": 7.403279781341553,
+ "step": 2962
+ },
+ {
+ "epoch": 41.15371179039301,
+ "grad_norm": 0.009075703099370003,
+ "learning_rate": 0.0006,
+ "loss": 7.447347164154053,
+ "step": 2963
+ },
+ {
+ "epoch": 41.16768558951965,
+ "grad_norm": 0.008892177604138851,
+ "learning_rate": 0.0006,
+ "loss": 7.402681350708008,
+ "step": 2964
+ },
+ {
+ "epoch": 41.18165938864629,
+ "grad_norm": 0.009492084383964539,
+ "learning_rate": 0.0006,
+ "loss": 7.3871355056762695,
+ "step": 2965
+ },
+ {
+ "epoch": 41.19563318777293,
+ "grad_norm": 0.009139460511505604,
+ "learning_rate": 0.0006,
+ "loss": 7.419617176055908,
+ "step": 2966
+ },
+ {
+ "epoch": 41.209606986899566,
+ "grad_norm": 0.007959138602018356,
+ "learning_rate": 0.0006,
+ "loss": 7.374949932098389,
+ "step": 2967
+ },
+ {
+ "epoch": 41.223580786026204,
+ "grad_norm": 0.007571548223495483,
+ "learning_rate": 0.0006,
+ "loss": 7.4205241203308105,
+ "step": 2968
+ },
+ {
+ "epoch": 41.237554585152836,
+ "grad_norm": 0.007885400205850601,
+ "learning_rate": 0.0006,
+ "loss": 7.416543006896973,
+ "step": 2969
+ },
+ {
+ "epoch": 41.251528384279474,
+ "grad_norm": 0.008650158531963825,
+ "learning_rate": 0.0006,
+ "loss": 7.357326984405518,
+ "step": 2970
+ },
+ {
+ "epoch": 41.26550218340611,
+ "grad_norm": 0.007483262103050947,
+ "learning_rate": 0.0006,
+ "loss": 7.410891056060791,
+ "step": 2971
+ },
+ {
+ "epoch": 41.27947598253275,
+ "grad_norm": 0.00869277399033308,
+ "learning_rate": 0.0006,
+ "loss": 7.392127990722656,
+ "step": 2972
+ },
+ {
+ "epoch": 41.29344978165939,
+ "grad_norm": 0.00885078776627779,
+ "learning_rate": 0.0006,
+ "loss": 7.405449867248535,
+ "step": 2973
+ },
+ {
+ "epoch": 41.30742358078603,
+ "grad_norm": 0.00958226341754198,
+ "learning_rate": 0.0006,
+ "loss": 7.399868488311768,
+ "step": 2974
+ },
+ {
+ "epoch": 41.32139737991266,
+ "grad_norm": 0.006991599686443806,
+ "learning_rate": 0.0006,
+ "loss": 7.35369873046875,
+ "step": 2975
+ },
+ {
+ "epoch": 41.3353711790393,
+ "grad_norm": 0.007415142375975847,
+ "learning_rate": 0.0006,
+ "loss": 7.358353137969971,
+ "step": 2976
+ },
+ {
+ "epoch": 41.34934497816594,
+ "grad_norm": 0.007024371065199375,
+ "learning_rate": 0.0006,
+ "loss": 7.4070892333984375,
+ "step": 2977
+ },
+ {
+ "epoch": 41.36331877729258,
+ "grad_norm": 0.007234332151710987,
+ "learning_rate": 0.0006,
+ "loss": 7.379120349884033,
+ "step": 2978
+ },
+ {
+ "epoch": 41.377292576419215,
+ "grad_norm": 0.00798892229795456,
+ "learning_rate": 0.0006,
+ "loss": 7.365455627441406,
+ "step": 2979
+ },
+ {
+ "epoch": 41.391266375545854,
+ "grad_norm": 0.007303398102521896,
+ "learning_rate": 0.0006,
+ "loss": 7.341798782348633,
+ "step": 2980
+ },
+ {
+ "epoch": 41.40524017467249,
+ "grad_norm": 0.007119669578969479,
+ "learning_rate": 0.0006,
+ "loss": 7.363280296325684,
+ "step": 2981
+ },
+ {
+ "epoch": 41.419213973799124,
+ "grad_norm": 0.007457619532942772,
+ "learning_rate": 0.0006,
+ "loss": 7.39388370513916,
+ "step": 2982
+ },
+ {
+ "epoch": 41.43318777292576,
+ "grad_norm": 0.007905688136816025,
+ "learning_rate": 0.0006,
+ "loss": 7.380652904510498,
+ "step": 2983
+ },
+ {
+ "epoch": 41.4471615720524,
+ "grad_norm": 0.006212855689227581,
+ "learning_rate": 0.0006,
+ "loss": 7.398769378662109,
+ "step": 2984
+ },
+ {
+ "epoch": 41.46113537117904,
+ "grad_norm": 0.0070641543716192245,
+ "learning_rate": 0.0006,
+ "loss": 7.3503546714782715,
+ "step": 2985
+ },
+ {
+ "epoch": 41.47510917030568,
+ "grad_norm": 0.00787805300205946,
+ "learning_rate": 0.0006,
+ "loss": 7.347678184509277,
+ "step": 2986
+ },
+ {
+ "epoch": 41.48908296943232,
+ "grad_norm": 0.006760426796972752,
+ "learning_rate": 0.0006,
+ "loss": 7.380006790161133,
+ "step": 2987
+ },
+ {
+ "epoch": 41.50305676855895,
+ "grad_norm": 0.00772385997697711,
+ "learning_rate": 0.0006,
+ "loss": 7.363680362701416,
+ "step": 2988
+ },
+ {
+ "epoch": 41.51703056768559,
+ "grad_norm": 0.007373564876616001,
+ "learning_rate": 0.0006,
+ "loss": 7.333014488220215,
+ "step": 2989
+ },
+ {
+ "epoch": 41.531004366812226,
+ "grad_norm": 0.006717653013765812,
+ "learning_rate": 0.0006,
+ "loss": 7.360076427459717,
+ "step": 2990
+ },
+ {
+ "epoch": 41.544978165938865,
+ "grad_norm": 0.007191413082182407,
+ "learning_rate": 0.0006,
+ "loss": 7.397030830383301,
+ "step": 2991
+ },
+ {
+ "epoch": 41.5589519650655,
+ "grad_norm": 0.007831960916519165,
+ "learning_rate": 0.0006,
+ "loss": 7.369959831237793,
+ "step": 2992
+ },
+ {
+ "epoch": 41.57292576419214,
+ "grad_norm": 0.006947481073439121,
+ "learning_rate": 0.0006,
+ "loss": 7.333254337310791,
+ "step": 2993
+ },
+ {
+ "epoch": 41.58689956331878,
+ "grad_norm": 0.006243606563657522,
+ "learning_rate": 0.0006,
+ "loss": 7.327378273010254,
+ "step": 2994
+ },
+ {
+ "epoch": 41.60087336244541,
+ "grad_norm": 0.006298339460045099,
+ "learning_rate": 0.0006,
+ "loss": 7.323482036590576,
+ "step": 2995
+ },
+ {
+ "epoch": 41.61484716157205,
+ "grad_norm": 0.007587412837892771,
+ "learning_rate": 0.0006,
+ "loss": 7.313789367675781,
+ "step": 2996
+ },
+ {
+ "epoch": 41.62882096069869,
+ "grad_norm": 0.007140746805816889,
+ "learning_rate": 0.0006,
+ "loss": 7.356067657470703,
+ "step": 2997
+ },
+ {
+ "epoch": 41.64279475982533,
+ "grad_norm": 0.008315625600516796,
+ "learning_rate": 0.0006,
+ "loss": 7.3339643478393555,
+ "step": 2998
+ },
+ {
+ "epoch": 41.65676855895197,
+ "grad_norm": 0.006547102238982916,
+ "learning_rate": 0.0006,
+ "loss": 7.342191696166992,
+ "step": 2999
+ },
+ {
+ "epoch": 41.670742358078606,
+ "grad_norm": 0.006068122107535601,
+ "learning_rate": 0.0006,
+ "loss": 7.321015357971191,
+ "step": 3000
+ },
+ {
+ "epoch": 41.68471615720524,
+ "grad_norm": 0.007316991221159697,
+ "learning_rate": 0.0006,
+ "loss": 7.346490859985352,
+ "step": 3001
+ },
+ {
+ "epoch": 41.698689956331876,
+ "grad_norm": 0.0067388564348220825,
+ "learning_rate": 0.0006,
+ "loss": 7.340723991394043,
+ "step": 3002
+ },
+ {
+ "epoch": 41.712663755458514,
+ "grad_norm": 0.006839963141828775,
+ "learning_rate": 0.0006,
+ "loss": 7.329365253448486,
+ "step": 3003
+ },
+ {
+ "epoch": 41.72663755458515,
+ "grad_norm": 0.00805041566491127,
+ "learning_rate": 0.0006,
+ "loss": 7.298934459686279,
+ "step": 3004
+ },
+ {
+ "epoch": 41.74061135371179,
+ "grad_norm": 0.007197503466159105,
+ "learning_rate": 0.0006,
+ "loss": 7.331496238708496,
+ "step": 3005
+ },
+ {
+ "epoch": 41.75458515283843,
+ "grad_norm": 0.007220805156975985,
+ "learning_rate": 0.0006,
+ "loss": 7.317695140838623,
+ "step": 3006
+ },
+ {
+ "epoch": 41.76855895196506,
+ "grad_norm": 0.00874276738613844,
+ "learning_rate": 0.0006,
+ "loss": 7.272614479064941,
+ "step": 3007
+ },
+ {
+ "epoch": 41.7825327510917,
+ "grad_norm": 0.008386565372347832,
+ "learning_rate": 0.0006,
+ "loss": 7.295244216918945,
+ "step": 3008
+ },
+ {
+ "epoch": 41.79650655021834,
+ "grad_norm": 0.0077730282209813595,
+ "learning_rate": 0.0006,
+ "loss": 7.3395490646362305,
+ "step": 3009
+ },
+ {
+ "epoch": 41.81048034934498,
+ "grad_norm": 0.008441480807960033,
+ "learning_rate": 0.0006,
+ "loss": 7.319921493530273,
+ "step": 3010
+ },
+ {
+ "epoch": 41.82445414847162,
+ "grad_norm": 0.007398269604891539,
+ "learning_rate": 0.0006,
+ "loss": 7.3029465675354,
+ "step": 3011
+ },
+ {
+ "epoch": 41.838427947598255,
+ "grad_norm": 0.006878677289932966,
+ "learning_rate": 0.0006,
+ "loss": 7.318467140197754,
+ "step": 3012
+ },
+ {
+ "epoch": 41.852401746724894,
+ "grad_norm": 0.007616598624736071,
+ "learning_rate": 0.0006,
+ "loss": 7.3112077713012695,
+ "step": 3013
+ },
+ {
+ "epoch": 41.866375545851525,
+ "grad_norm": 0.011629699729382992,
+ "learning_rate": 0.0006,
+ "loss": 7.32523250579834,
+ "step": 3014
+ },
+ {
+ "epoch": 41.880349344978164,
+ "grad_norm": 0.011011249385774136,
+ "learning_rate": 0.0006,
+ "loss": 7.287247657775879,
+ "step": 3015
+ },
+ {
+ "epoch": 41.8943231441048,
+ "grad_norm": 0.010208295658230782,
+ "learning_rate": 0.0006,
+ "loss": 7.292117118835449,
+ "step": 3016
+ },
+ {
+ "epoch": 41.90829694323144,
+ "grad_norm": 0.008651095442473888,
+ "learning_rate": 0.0006,
+ "loss": 7.302201271057129,
+ "step": 3017
+ },
+ {
+ "epoch": 41.92227074235808,
+ "grad_norm": 0.006692920345813036,
+ "learning_rate": 0.0006,
+ "loss": 7.319568634033203,
+ "step": 3018
+ },
+ {
+ "epoch": 41.93624454148472,
+ "grad_norm": 0.010033348575234413,
+ "learning_rate": 0.0006,
+ "loss": 7.2738566398620605,
+ "step": 3019
+ },
+ {
+ "epoch": 41.95021834061135,
+ "grad_norm": 0.011790832504630089,
+ "learning_rate": 0.0006,
+ "loss": 7.298540115356445,
+ "step": 3020
+ },
+ {
+ "epoch": 41.96419213973799,
+ "grad_norm": 0.018622662872076035,
+ "learning_rate": 0.0006,
+ "loss": 7.267182350158691,
+ "step": 3021
+ },
+ {
+ "epoch": 41.97816593886463,
+ "grad_norm": 0.02520257607102394,
+ "learning_rate": 0.0006,
+ "loss": 7.261566162109375,
+ "step": 3022
+ },
+ {
+ "epoch": 41.992139737991266,
+ "grad_norm": 0.036332644522190094,
+ "learning_rate": 0.0006,
+ "loss": 7.270719051361084,
+ "step": 3023
+ },
+ {
+ "epoch": 42.0,
+ "grad_norm": 0.04717305675148964,
+ "learning_rate": 0.0006,
+ "loss": 7.293332099914551,
+ "step": 3024
+ },
+ {
+ "epoch": 42.0,
+ "eval_loss": 7.354280471801758,
+ "eval_runtime": 56.4353,
+ "eval_samples_per_second": 43.271,
+ "eval_steps_per_second": 1.364,
+ "step": 3024
+ },
+ {
+ "epoch": 42.01397379912664,
+ "grad_norm": 0.06661626696586609,
+ "learning_rate": 0.0006,
+ "loss": 7.338189601898193,
+ "step": 3025
+ },
+ {
+ "epoch": 42.02794759825328,
+ "grad_norm": 0.05452054366469383,
+ "learning_rate": 0.0006,
+ "loss": 7.316764831542969,
+ "step": 3026
+ },
+ {
+ "epoch": 42.041921397379916,
+ "grad_norm": 0.025638775900006294,
+ "learning_rate": 0.0006,
+ "loss": 7.34204626083374,
+ "step": 3027
+ },
+ {
+ "epoch": 42.05589519650655,
+ "grad_norm": 0.022398799657821655,
+ "learning_rate": 0.0006,
+ "loss": 7.262149333953857,
+ "step": 3028
+ },
+ {
+ "epoch": 42.069868995633186,
+ "grad_norm": 0.03713499382138252,
+ "learning_rate": 0.0006,
+ "loss": 7.27695369720459,
+ "step": 3029
+ },
+ {
+ "epoch": 42.083842794759825,
+ "grad_norm": 0.014949247241020203,
+ "learning_rate": 0.0006,
+ "loss": 7.296508312225342,
+ "step": 3030
+ },
+ {
+ "epoch": 42.09781659388646,
+ "grad_norm": 0.02703995071351528,
+ "learning_rate": 0.0006,
+ "loss": 7.258018493652344,
+ "step": 3031
+ },
+ {
+ "epoch": 42.1117903930131,
+ "grad_norm": 0.020352816209197044,
+ "learning_rate": 0.0006,
+ "loss": 7.241450309753418,
+ "step": 3032
+ },
+ {
+ "epoch": 42.12576419213974,
+ "grad_norm": 0.014578479342162609,
+ "learning_rate": 0.0006,
+ "loss": 7.257323265075684,
+ "step": 3033
+ },
+ {
+ "epoch": 42.13973799126637,
+ "grad_norm": 0.026053044945001602,
+ "learning_rate": 0.0006,
+ "loss": 7.254680633544922,
+ "step": 3034
+ },
+ {
+ "epoch": 42.15371179039301,
+ "grad_norm": 0.015120275318622589,
+ "learning_rate": 0.0006,
+ "loss": 7.261074066162109,
+ "step": 3035
+ },
+ {
+ "epoch": 42.16768558951965,
+ "grad_norm": 0.016822177916765213,
+ "learning_rate": 0.0006,
+ "loss": 7.279064178466797,
+ "step": 3036
+ },
+ {
+ "epoch": 42.18165938864629,
+ "grad_norm": 0.016367588192224503,
+ "learning_rate": 0.0006,
+ "loss": 7.251429557800293,
+ "step": 3037
+ },
+ {
+ "epoch": 42.19563318777293,
+ "grad_norm": 0.014350418001413345,
+ "learning_rate": 0.0006,
+ "loss": 7.259264945983887,
+ "step": 3038
+ },
+ {
+ "epoch": 42.209606986899566,
+ "grad_norm": 0.017078228294849396,
+ "learning_rate": 0.0006,
+ "loss": 7.227105140686035,
+ "step": 3039
+ },
+ {
+ "epoch": 42.223580786026204,
+ "grad_norm": 0.011759317480027676,
+ "learning_rate": 0.0006,
+ "loss": 7.275888442993164,
+ "step": 3040
+ },
+ {
+ "epoch": 42.237554585152836,
+ "grad_norm": 0.013019340112805367,
+ "learning_rate": 0.0006,
+ "loss": 7.2517008781433105,
+ "step": 3041
+ },
+ {
+ "epoch": 42.251528384279474,
+ "grad_norm": 0.01178441196680069,
+ "learning_rate": 0.0006,
+ "loss": 7.23576545715332,
+ "step": 3042
+ },
+ {
+ "epoch": 42.26550218340611,
+ "grad_norm": 0.014518891461193562,
+ "learning_rate": 0.0006,
+ "loss": 7.253895282745361,
+ "step": 3043
+ },
+ {
+ "epoch": 42.27947598253275,
+ "grad_norm": 0.018159950152039528,
+ "learning_rate": 0.0006,
+ "loss": 7.278058052062988,
+ "step": 3044
+ },
+ {
+ "epoch": 42.29344978165939,
+ "grad_norm": 0.007587796077132225,
+ "learning_rate": 0.0006,
+ "loss": 7.278578758239746,
+ "step": 3045
+ },
+ {
+ "epoch": 42.30742358078603,
+ "grad_norm": 0.015994979068636894,
+ "learning_rate": 0.0006,
+ "loss": 7.239841461181641,
+ "step": 3046
+ },
+ {
+ "epoch": 42.32139737991266,
+ "grad_norm": 0.017830168828368187,
+ "learning_rate": 0.0006,
+ "loss": 7.2443976402282715,
+ "step": 3047
+ },
+ {
+ "epoch": 42.3353711790393,
+ "grad_norm": 0.008738759905099869,
+ "learning_rate": 0.0006,
+ "loss": 7.232264518737793,
+ "step": 3048
+ },
+ {
+ "epoch": 42.34934497816594,
+ "grad_norm": 0.0175449438393116,
+ "learning_rate": 0.0006,
+ "loss": 7.228000640869141,
+ "step": 3049
+ },
+ {
+ "epoch": 42.36331877729258,
+ "grad_norm": 0.019611820578575134,
+ "learning_rate": 0.0006,
+ "loss": 7.234999179840088,
+ "step": 3050
+ },
+ {
+ "epoch": 42.377292576419215,
+ "grad_norm": 0.015302863903343678,
+ "learning_rate": 0.0006,
+ "loss": 7.242547988891602,
+ "step": 3051
+ },
+ {
+ "epoch": 42.391266375545854,
+ "grad_norm": 0.025919731706380844,
+ "learning_rate": 0.0006,
+ "loss": 7.216652870178223,
+ "step": 3052
+ },
+ {
+ "epoch": 42.40524017467249,
+ "grad_norm": 0.034587468951940536,
+ "learning_rate": 0.0006,
+ "loss": 7.245336532592773,
+ "step": 3053
+ },
+ {
+ "epoch": 42.419213973799124,
+ "grad_norm": 0.019198840484023094,
+ "learning_rate": 0.0006,
+ "loss": 7.256758689880371,
+ "step": 3054
+ },
+ {
+ "epoch": 42.43318777292576,
+ "grad_norm": 0.012316389009356499,
+ "learning_rate": 0.0006,
+ "loss": 7.209102630615234,
+ "step": 3055
+ },
+ {
+ "epoch": 42.4471615720524,
+ "grad_norm": 0.023512162268161774,
+ "learning_rate": 0.0006,
+ "loss": 7.250742435455322,
+ "step": 3056
+ },
+ {
+ "epoch": 42.46113537117904,
+ "grad_norm": 0.025457851588726044,
+ "learning_rate": 0.0006,
+ "loss": 7.263757705688477,
+ "step": 3057
+ },
+ {
+ "epoch": 42.47510917030568,
+ "grad_norm": 0.02192617766559124,
+ "learning_rate": 0.0006,
+ "loss": 7.246977806091309,
+ "step": 3058
+ },
+ {
+ "epoch": 42.48908296943232,
+ "grad_norm": 0.02541281282901764,
+ "learning_rate": 0.0006,
+ "loss": 7.232361316680908,
+ "step": 3059
+ },
+ {
+ "epoch": 42.50305676855895,
+ "grad_norm": 0.036353763192892075,
+ "learning_rate": 0.0006,
+ "loss": 7.242556571960449,
+ "step": 3060
+ },
+ {
+ "epoch": 42.51703056768559,
+ "grad_norm": 0.04555872827768326,
+ "learning_rate": 0.0006,
+ "loss": 7.228940963745117,
+ "step": 3061
+ },
+ {
+ "epoch": 42.531004366812226,
+ "grad_norm": 0.04563378915190697,
+ "learning_rate": 0.0006,
+ "loss": 7.235276222229004,
+ "step": 3062
+ },
+ {
+ "epoch": 42.544978165938865,
+ "grad_norm": 0.04275760427117348,
+ "learning_rate": 0.0006,
+ "loss": 7.238895416259766,
+ "step": 3063
+ },
+ {
+ "epoch": 42.5589519650655,
+ "grad_norm": 0.04535885155200958,
+ "learning_rate": 0.0006,
+ "loss": 7.205336570739746,
+ "step": 3064
+ },
+ {
+ "epoch": 42.57292576419214,
+ "grad_norm": 0.015534617006778717,
+ "learning_rate": 0.0006,
+ "loss": 7.238393783569336,
+ "step": 3065
+ },
+ {
+ "epoch": 42.58689956331878,
+ "grad_norm": 0.03286299481987953,
+ "learning_rate": 0.0006,
+ "loss": 7.216453552246094,
+ "step": 3066
+ },
+ {
+ "epoch": 42.60087336244541,
+ "grad_norm": 0.025453168898820877,
+ "learning_rate": 0.0006,
+ "loss": 7.203750133514404,
+ "step": 3067
+ },
+ {
+ "epoch": 42.61484716157205,
+ "grad_norm": 0.021501649171113968,
+ "learning_rate": 0.0006,
+ "loss": 7.241621971130371,
+ "step": 3068
+ },
+ {
+ "epoch": 42.62882096069869,
+ "grad_norm": 0.02743634767830372,
+ "learning_rate": 0.0006,
+ "loss": 7.22076940536499,
+ "step": 3069
+ },
+ {
+ "epoch": 42.64279475982533,
+ "grad_norm": 0.020890483632683754,
+ "learning_rate": 0.0006,
+ "loss": 7.218412399291992,
+ "step": 3070
+ },
+ {
+ "epoch": 42.65676855895197,
+ "grad_norm": 0.022499850019812584,
+ "learning_rate": 0.0006,
+ "loss": 7.217696189880371,
+ "step": 3071
+ },
+ {
+ "epoch": 42.670742358078606,
+ "grad_norm": 0.01653696969151497,
+ "learning_rate": 0.0006,
+ "loss": 7.156417369842529,
+ "step": 3072
+ },
+ {
+ "epoch": 42.68471615720524,
+ "grad_norm": 0.01834225095808506,
+ "learning_rate": 0.0006,
+ "loss": 7.211247444152832,
+ "step": 3073
+ },
+ {
+ "epoch": 42.698689956331876,
+ "grad_norm": 0.020137229934334755,
+ "learning_rate": 0.0006,
+ "loss": 7.232499599456787,
+ "step": 3074
+ },
+ {
+ "epoch": 42.712663755458514,
+ "grad_norm": 0.012671459466218948,
+ "learning_rate": 0.0006,
+ "loss": 7.172441482543945,
+ "step": 3075
+ },
+ {
+ "epoch": 42.72663755458515,
+ "grad_norm": 0.016803130507469177,
+ "learning_rate": 0.0006,
+ "loss": 7.152103424072266,
+ "step": 3076
+ },
+ {
+ "epoch": 42.74061135371179,
+ "grad_norm": 0.01623712107539177,
+ "learning_rate": 0.0006,
+ "loss": 7.187032699584961,
+ "step": 3077
+ },
+ {
+ "epoch": 42.75458515283843,
+ "grad_norm": 0.015938356518745422,
+ "learning_rate": 0.0006,
+ "loss": 7.171140670776367,
+ "step": 3078
+ },
+ {
+ "epoch": 42.76855895196506,
+ "grad_norm": 0.02580736204981804,
+ "learning_rate": 0.0006,
+ "loss": 7.204472541809082,
+ "step": 3079
+ },
+ {
+ "epoch": 42.7825327510917,
+ "grad_norm": 0.023372117429971695,
+ "learning_rate": 0.0006,
+ "loss": 7.165554523468018,
+ "step": 3080
+ },
+ {
+ "epoch": 42.79650655021834,
+ "grad_norm": 0.04440586268901825,
+ "learning_rate": 0.0006,
+ "loss": 7.169431209564209,
+ "step": 3081
+ },
+ {
+ "epoch": 42.81048034934498,
+ "grad_norm": 0.06797406077384949,
+ "learning_rate": 0.0006,
+ "loss": 7.168193340301514,
+ "step": 3082
+ },
+ {
+ "epoch": 42.82445414847162,
+ "grad_norm": 0.0914321169257164,
+ "learning_rate": 0.0006,
+ "loss": 7.204671859741211,
+ "step": 3083
+ },
+ {
+ "epoch": 42.838427947598255,
+ "grad_norm": 0.0506940595805645,
+ "learning_rate": 0.0006,
+ "loss": 7.15949821472168,
+ "step": 3084
+ },
+ {
+ "epoch": 42.852401746724894,
+ "grad_norm": 0.04819890484213829,
+ "learning_rate": 0.0006,
+ "loss": 7.192837238311768,
+ "step": 3085
+ },
+ {
+ "epoch": 42.866375545851525,
+ "grad_norm": 0.040044233202934265,
+ "learning_rate": 0.0006,
+ "loss": 7.1600165367126465,
+ "step": 3086
+ },
+ {
+ "epoch": 42.880349344978164,
+ "grad_norm": 0.02573299966752529,
+ "learning_rate": 0.0006,
+ "loss": 7.153427600860596,
+ "step": 3087
+ },
+ {
+ "epoch": 42.8943231441048,
+ "grad_norm": 0.046942535787820816,
+ "learning_rate": 0.0006,
+ "loss": 7.16767692565918,
+ "step": 3088
+ },
+ {
+ "epoch": 42.90829694323144,
+ "grad_norm": 0.02683161571621895,
+ "learning_rate": 0.0006,
+ "loss": 7.17144250869751,
+ "step": 3089
+ },
+ {
+ "epoch": 42.92227074235808,
+ "grad_norm": 0.05363674834370613,
+ "learning_rate": 0.0006,
+ "loss": 7.120128631591797,
+ "step": 3090
+ },
+ {
+ "epoch": 42.93624454148472,
+ "grad_norm": 0.023756247013807297,
+ "learning_rate": 0.0006,
+ "loss": 7.169703006744385,
+ "step": 3091
+ },
+ {
+ "epoch": 42.95021834061135,
+ "grad_norm": 0.04216783866286278,
+ "learning_rate": 0.0006,
+ "loss": 7.1206865310668945,
+ "step": 3092
+ },
+ {
+ "epoch": 42.96419213973799,
+ "grad_norm": 0.02219565585255623,
+ "learning_rate": 0.0006,
+ "loss": 7.157215118408203,
+ "step": 3093
+ },
+ {
+ "epoch": 42.97816593886463,
+ "grad_norm": 0.03527127206325531,
+ "learning_rate": 0.0006,
+ "loss": 7.166672706604004,
+ "step": 3094
+ },
+ {
+ "epoch": 42.992139737991266,
+ "grad_norm": 0.017748817801475525,
+ "learning_rate": 0.0006,
+ "loss": 7.154747009277344,
+ "step": 3095
+ },
+ {
+ "epoch": 43.0,
+ "grad_norm": 0.033165838569402695,
+ "learning_rate": 0.0006,
+ "loss": 7.150866508483887,
+ "step": 3096
+ },
+ {
+ "epoch": 43.0,
+ "eval_loss": 7.144403457641602,
+ "eval_runtime": 56.8853,
+ "eval_samples_per_second": 42.928,
+ "eval_steps_per_second": 1.354,
+ "step": 3096
+ },
+ {
+ "epoch": 43.01397379912664,
+ "grad_norm": 0.015199829824268818,
+ "learning_rate": 0.0006,
+ "loss": 7.122735023498535,
+ "step": 3097
+ },
+ {
+ "epoch": 43.02794759825328,
+ "grad_norm": 0.03220256045460701,
+ "learning_rate": 0.0006,
+ "loss": 7.143553733825684,
+ "step": 3098
+ },
+ {
+ "epoch": 43.041921397379916,
+ "grad_norm": 0.014687920920550823,
+ "learning_rate": 0.0006,
+ "loss": 7.080960750579834,
+ "step": 3099
+ },
+ {
+ "epoch": 43.05589519650655,
+ "grad_norm": 0.023522932082414627,
+ "learning_rate": 0.0006,
+ "loss": 7.137356758117676,
+ "step": 3100
+ },
+ {
+ "epoch": 43.069868995633186,
+ "grad_norm": 0.014808216132223606,
+ "learning_rate": 0.0006,
+ "loss": 7.09143590927124,
+ "step": 3101
+ },
+ {
+ "epoch": 43.083842794759825,
+ "grad_norm": 0.029339246451854706,
+ "learning_rate": 0.0006,
+ "loss": 7.122198104858398,
+ "step": 3102
+ },
+ {
+ "epoch": 43.09781659388646,
+ "grad_norm": 0.022953078150749207,
+ "learning_rate": 0.0006,
+ "loss": 7.092505931854248,
+ "step": 3103
+ },
+ {
+ "epoch": 43.1117903930131,
+ "grad_norm": 0.04590052738785744,
+ "learning_rate": 0.0006,
+ "loss": 7.113100051879883,
+ "step": 3104
+ },
+ {
+ "epoch": 43.12576419213974,
+ "grad_norm": 0.0524442233145237,
+ "learning_rate": 0.0006,
+ "loss": 7.147022247314453,
+ "step": 3105
+ },
+ {
+ "epoch": 43.13973799126637,
+ "grad_norm": 0.0315794013440609,
+ "learning_rate": 0.0006,
+ "loss": 7.094045162200928,
+ "step": 3106
+ },
+ {
+ "epoch": 43.15371179039301,
+ "grad_norm": 0.05112722888588905,
+ "learning_rate": 0.0006,
+ "loss": 7.054388523101807,
+ "step": 3107
+ },
+ {
+ "epoch": 43.16768558951965,
+ "grad_norm": 0.03146776929497719,
+ "learning_rate": 0.0006,
+ "loss": 7.116674423217773,
+ "step": 3108
+ },
+ {
+ "epoch": 43.18165938864629,
+ "grad_norm": 0.03391246497631073,
+ "learning_rate": 0.0006,
+ "loss": 7.150345802307129,
+ "step": 3109
+ },
+ {
+ "epoch": 43.19563318777293,
+ "grad_norm": 0.026957903057336807,
+ "learning_rate": 0.0006,
+ "loss": 7.058531284332275,
+ "step": 3110
+ },
+ {
+ "epoch": 43.209606986899566,
+ "grad_norm": 0.041532568633556366,
+ "learning_rate": 0.0006,
+ "loss": 7.085498809814453,
+ "step": 3111
+ },
+ {
+ "epoch": 43.223580786026204,
+ "grad_norm": 0.020001588389277458,
+ "learning_rate": 0.0006,
+ "loss": 7.131487846374512,
+ "step": 3112
+ },
+ {
+ "epoch": 43.237554585152836,
+ "grad_norm": 0.02268712781369686,
+ "learning_rate": 0.0006,
+ "loss": 7.0790863037109375,
+ "step": 3113
+ },
+ {
+ "epoch": 43.251528384279474,
+ "grad_norm": 0.023363925516605377,
+ "learning_rate": 0.0006,
+ "loss": 7.067816734313965,
+ "step": 3114
+ },
+ {
+ "epoch": 43.26550218340611,
+ "grad_norm": 0.019367508590221405,
+ "learning_rate": 0.0006,
+ "loss": 7.042041778564453,
+ "step": 3115
+ },
+ {
+ "epoch": 43.27947598253275,
+ "grad_norm": 0.02745090238749981,
+ "learning_rate": 0.0006,
+ "loss": 7.064733028411865,
+ "step": 3116
+ },
+ {
+ "epoch": 43.29344978165939,
+ "grad_norm": 0.02543802745640278,
+ "learning_rate": 0.0006,
+ "loss": 7.058948040008545,
+ "step": 3117
+ },
+ {
+ "epoch": 43.30742358078603,
+ "grad_norm": 0.018874092027544975,
+ "learning_rate": 0.0006,
+ "loss": 7.009946346282959,
+ "step": 3118
+ },
+ {
+ "epoch": 43.32139737991266,
+ "grad_norm": 0.024662822484970093,
+ "learning_rate": 0.0006,
+ "loss": 7.015913963317871,
+ "step": 3119
+ },
+ {
+ "epoch": 43.3353711790393,
+ "grad_norm": 0.02441416308283806,
+ "learning_rate": 0.0006,
+ "loss": 7.061622142791748,
+ "step": 3120
+ },
+ {
+ "epoch": 43.34934497816594,
+ "grad_norm": 0.031887322664260864,
+ "learning_rate": 0.0006,
+ "loss": 7.046883583068848,
+ "step": 3121
+ },
+ {
+ "epoch": 43.36331877729258,
+ "grad_norm": 0.021930739283561707,
+ "learning_rate": 0.0006,
+ "loss": 7.0309624671936035,
+ "step": 3122
+ },
+ {
+ "epoch": 43.377292576419215,
+ "grad_norm": 0.02722783200442791,
+ "learning_rate": 0.0006,
+ "loss": 7.072924613952637,
+ "step": 3123
+ },
+ {
+ "epoch": 43.391266375545854,
+ "grad_norm": 0.01483179721981287,
+ "learning_rate": 0.0006,
+ "loss": 7.028624534606934,
+ "step": 3124
+ },
+ {
+ "epoch": 43.40524017467249,
+ "grad_norm": 0.024793945252895355,
+ "learning_rate": 0.0006,
+ "loss": 7.001623153686523,
+ "step": 3125
+ },
+ {
+ "epoch": 43.419213973799124,
+ "grad_norm": 0.020148713141679764,
+ "learning_rate": 0.0006,
+ "loss": 6.9938812255859375,
+ "step": 3126
+ },
+ {
+ "epoch": 43.43318777292576,
+ "grad_norm": 0.02413850650191307,
+ "learning_rate": 0.0006,
+ "loss": 7.001243591308594,
+ "step": 3127
+ },
+ {
+ "epoch": 43.4471615720524,
+ "grad_norm": 0.028549203649163246,
+ "learning_rate": 0.0006,
+ "loss": 6.987671375274658,
+ "step": 3128
+ },
+ {
+ "epoch": 43.46113537117904,
+ "grad_norm": 0.025988785549998283,
+ "learning_rate": 0.0006,
+ "loss": 6.9551167488098145,
+ "step": 3129
+ },
+ {
+ "epoch": 43.47510917030568,
+ "grad_norm": 0.027140729129314423,
+ "learning_rate": 0.0006,
+ "loss": 6.9547810554504395,
+ "step": 3130
+ },
+ {
+ "epoch": 43.48908296943232,
+ "grad_norm": 0.030130082741379738,
+ "learning_rate": 0.0006,
+ "loss": 6.993186950683594,
+ "step": 3131
+ },
+ {
+ "epoch": 43.50305676855895,
+ "grad_norm": 0.025996994227170944,
+ "learning_rate": 0.0006,
+ "loss": 6.984624862670898,
+ "step": 3132
+ },
+ {
+ "epoch": 43.51703056768559,
+ "grad_norm": 0.02423008903861046,
+ "learning_rate": 0.0006,
+ "loss": 6.98813009262085,
+ "step": 3133
+ },
+ {
+ "epoch": 43.531004366812226,
+ "grad_norm": 0.029665837064385414,
+ "learning_rate": 0.0006,
+ "loss": 6.986610412597656,
+ "step": 3134
+ },
+ {
+ "epoch": 43.544978165938865,
+ "grad_norm": 0.028768595308065414,
+ "learning_rate": 0.0006,
+ "loss": 6.898797035217285,
+ "step": 3135
+ },
+ {
+ "epoch": 43.5589519650655,
+ "grad_norm": 0.03111208975315094,
+ "learning_rate": 0.0006,
+ "loss": 6.923741817474365,
+ "step": 3136
+ },
+ {
+ "epoch": 43.57292576419214,
+ "grad_norm": 0.0368187390267849,
+ "learning_rate": 0.0006,
+ "loss": 6.994343280792236,
+ "step": 3137
+ },
+ {
+ "epoch": 43.58689956331878,
+ "grad_norm": 0.04076763987541199,
+ "learning_rate": 0.0006,
+ "loss": 6.909458160400391,
+ "step": 3138
+ },
+ {
+ "epoch": 43.60087336244541,
+ "grad_norm": 0.03281061723828316,
+ "learning_rate": 0.0006,
+ "loss": 6.924851417541504,
+ "step": 3139
+ },
+ {
+ "epoch": 43.61484716157205,
+ "grad_norm": 0.025486988946795464,
+ "learning_rate": 0.0006,
+ "loss": 6.962926864624023,
+ "step": 3140
+ },
+ {
+ "epoch": 43.62882096069869,
+ "grad_norm": 0.04583485797047615,
+ "learning_rate": 0.0006,
+ "loss": 6.876555442810059,
+ "step": 3141
+ },
+ {
+ "epoch": 43.64279475982533,
+ "grad_norm": 0.05460785701870918,
+ "learning_rate": 0.0006,
+ "loss": 6.952932357788086,
+ "step": 3142
+ },
+ {
+ "epoch": 43.65676855895197,
+ "grad_norm": 0.040917906910181046,
+ "learning_rate": 0.0006,
+ "loss": 6.9624924659729,
+ "step": 3143
+ },
+ {
+ "epoch": 43.670742358078606,
+ "grad_norm": 0.04009668901562691,
+ "learning_rate": 0.0006,
+ "loss": 6.94033145904541,
+ "step": 3144
+ },
+ {
+ "epoch": 43.68471615720524,
+ "grad_norm": 0.04500625655055046,
+ "learning_rate": 0.0006,
+ "loss": 6.961839199066162,
+ "step": 3145
+ },
+ {
+ "epoch": 43.698689956331876,
+ "grad_norm": 0.03182258829474449,
+ "learning_rate": 0.0006,
+ "loss": 6.930580139160156,
+ "step": 3146
+ },
+ {
+ "epoch": 43.712663755458514,
+ "grad_norm": 0.053621504455804825,
+ "learning_rate": 0.0006,
+ "loss": 6.950235366821289,
+ "step": 3147
+ },
+ {
+ "epoch": 43.72663755458515,
+ "grad_norm": 0.031919289380311966,
+ "learning_rate": 0.0006,
+ "loss": 6.895111560821533,
+ "step": 3148
+ },
+ {
+ "epoch": 43.74061135371179,
+ "grad_norm": 0.042713891714811325,
+ "learning_rate": 0.0006,
+ "loss": 6.940885543823242,
+ "step": 3149
+ },
+ {
+ "epoch": 43.75458515283843,
+ "grad_norm": 0.02385653182864189,
+ "learning_rate": 0.0006,
+ "loss": 6.948801517486572,
+ "step": 3150
+ },
+ {
+ "epoch": 43.76855895196506,
+ "grad_norm": 0.04100663587450981,
+ "learning_rate": 0.0006,
+ "loss": 6.970411777496338,
+ "step": 3151
+ },
+ {
+ "epoch": 43.7825327510917,
+ "grad_norm": 0.02729583904147148,
+ "learning_rate": 0.0006,
+ "loss": 6.899379253387451,
+ "step": 3152
+ },
+ {
+ "epoch": 43.79650655021834,
+ "grad_norm": 0.03128620237112045,
+ "learning_rate": 0.0006,
+ "loss": 6.931319236755371,
+ "step": 3153
+ },
+ {
+ "epoch": 43.81048034934498,
+ "grad_norm": 0.025488659739494324,
+ "learning_rate": 0.0006,
+ "loss": 6.8486409187316895,
+ "step": 3154
+ },
+ {
+ "epoch": 43.82445414847162,
+ "grad_norm": 0.03679376840591431,
+ "learning_rate": 0.0006,
+ "loss": 6.840811252593994,
+ "step": 3155
+ },
+ {
+ "epoch": 43.838427947598255,
+ "grad_norm": 0.020578034222126007,
+ "learning_rate": 0.0006,
+ "loss": 6.9088616371154785,
+ "step": 3156
+ },
+ {
+ "epoch": 43.852401746724894,
+ "grad_norm": 0.02935841865837574,
+ "learning_rate": 0.0006,
+ "loss": 6.806676864624023,
+ "step": 3157
+ },
+ {
+ "epoch": 43.866375545851525,
+ "grad_norm": 0.023319244384765625,
+ "learning_rate": 0.0006,
+ "loss": 6.878457546234131,
+ "step": 3158
+ },
+ {
+ "epoch": 43.880349344978164,
+ "grad_norm": 0.02779589593410492,
+ "learning_rate": 0.0006,
+ "loss": 6.822447776794434,
+ "step": 3159
+ },
+ {
+ "epoch": 43.8943231441048,
+ "grad_norm": 0.01879825070500374,
+ "learning_rate": 0.0006,
+ "loss": 6.854121208190918,
+ "step": 3160
+ },
+ {
+ "epoch": 43.90829694323144,
+ "grad_norm": 0.023286139592528343,
+ "learning_rate": 0.0006,
+ "loss": 6.836827278137207,
+ "step": 3161
+ },
+ {
+ "epoch": 43.92227074235808,
+ "grad_norm": 0.030358297750353813,
+ "learning_rate": 0.0006,
+ "loss": 6.860037803649902,
+ "step": 3162
+ },
+ {
+ "epoch": 43.93624454148472,
+ "grad_norm": 0.029621105641126633,
+ "learning_rate": 0.0006,
+ "loss": 6.884025573730469,
+ "step": 3163
+ },
+ {
+ "epoch": 43.95021834061135,
+ "grad_norm": 0.029829824343323708,
+ "learning_rate": 0.0006,
+ "loss": 6.8702392578125,
+ "step": 3164
+ },
+ {
+ "epoch": 43.96419213973799,
+ "grad_norm": 0.036892473697662354,
+ "learning_rate": 0.0006,
+ "loss": 6.806652069091797,
+ "step": 3165
+ },
+ {
+ "epoch": 43.97816593886463,
+ "grad_norm": 0.03206709027290344,
+ "learning_rate": 0.0006,
+ "loss": 6.828495979309082,
+ "step": 3166
+ },
+ {
+ "epoch": 43.992139737991266,
+ "grad_norm": 0.02744000405073166,
+ "learning_rate": 0.0006,
+ "loss": 6.832002639770508,
+ "step": 3167
+ },
+ {
+ "epoch": 44.0,
+ "grad_norm": 0.02205660194158554,
+ "learning_rate": 0.0006,
+ "loss": 6.898828506469727,
+ "step": 3168
+ },
+ {
+ "epoch": 44.0,
+ "eval_loss": 6.8150553703308105,
+ "eval_runtime": 56.784,
+ "eval_samples_per_second": 43.005,
+ "eval_steps_per_second": 1.356,
+ "step": 3168
+ },
+ {
+ "epoch": 44.01397379912664,
+ "grad_norm": 0.030421260744333267,
+ "learning_rate": 0.0006,
+ "loss": 6.775057792663574,
+ "step": 3169
+ },
+ {
+ "epoch": 44.02794759825328,
+ "grad_norm": 0.0465611033141613,
+ "learning_rate": 0.0006,
+ "loss": 6.849390506744385,
+ "step": 3170
+ },
+ {
+ "epoch": 44.041921397379916,
+ "grad_norm": 0.05977972596883774,
+ "learning_rate": 0.0006,
+ "loss": 6.725576400756836,
+ "step": 3171
+ },
+ {
+ "epoch": 44.05589519650655,
+ "grad_norm": 0.05432697385549545,
+ "learning_rate": 0.0006,
+ "loss": 6.768024444580078,
+ "step": 3172
+ },
+ {
+ "epoch": 44.069868995633186,
+ "grad_norm": 0.04850967228412628,
+ "learning_rate": 0.0006,
+ "loss": 6.799457550048828,
+ "step": 3173
+ },
+ {
+ "epoch": 44.083842794759825,
+ "grad_norm": 0.06201786920428276,
+ "learning_rate": 0.0006,
+ "loss": 6.897234916687012,
+ "step": 3174
+ },
+ {
+ "epoch": 44.09781659388646,
+ "grad_norm": 0.04695591703057289,
+ "learning_rate": 0.0006,
+ "loss": 6.826470375061035,
+ "step": 3175
+ },
+ {
+ "epoch": 44.1117903930131,
+ "grad_norm": 0.03850164636969566,
+ "learning_rate": 0.0006,
+ "loss": 6.835901260375977,
+ "step": 3176
+ },
+ {
+ "epoch": 44.12576419213974,
+ "grad_norm": 0.04250155761837959,
+ "learning_rate": 0.0006,
+ "loss": 6.782310485839844,
+ "step": 3177
+ },
+ {
+ "epoch": 44.13973799126637,
+ "grad_norm": 0.041465479880571365,
+ "learning_rate": 0.0006,
+ "loss": 6.785732746124268,
+ "step": 3178
+ },
+ {
+ "epoch": 44.15371179039301,
+ "grad_norm": 0.02527807280421257,
+ "learning_rate": 0.0006,
+ "loss": 6.78061580657959,
+ "step": 3179
+ },
+ {
+ "epoch": 44.16768558951965,
+ "grad_norm": 0.03660374879837036,
+ "learning_rate": 0.0006,
+ "loss": 6.783992767333984,
+ "step": 3180
+ },
+ {
+ "epoch": 44.18165938864629,
+ "grad_norm": 0.027567964047193527,
+ "learning_rate": 0.0006,
+ "loss": 6.765413284301758,
+ "step": 3181
+ },
+ {
+ "epoch": 44.19563318777293,
+ "grad_norm": 0.024301424622535706,
+ "learning_rate": 0.0006,
+ "loss": 6.760368347167969,
+ "step": 3182
+ },
+ {
+ "epoch": 44.209606986899566,
+ "grad_norm": 0.0266821701079607,
+ "learning_rate": 0.0006,
+ "loss": 6.71417236328125,
+ "step": 3183
+ },
+ {
+ "epoch": 44.223580786026204,
+ "grad_norm": 0.028125090524554253,
+ "learning_rate": 0.0006,
+ "loss": 6.646126747131348,
+ "step": 3184
+ },
+ {
+ "epoch": 44.237554585152836,
+ "grad_norm": 0.0185737032443285,
+ "learning_rate": 0.0006,
+ "loss": 6.712944030761719,
+ "step": 3185
+ },
+ {
+ "epoch": 44.251528384279474,
+ "grad_norm": 0.025836674496531487,
+ "learning_rate": 0.0006,
+ "loss": 6.707030296325684,
+ "step": 3186
+ },
+ {
+ "epoch": 44.26550218340611,
+ "grad_norm": 0.029744870960712433,
+ "learning_rate": 0.0006,
+ "loss": 6.6271071434021,
+ "step": 3187
+ },
+ {
+ "epoch": 44.27947598253275,
+ "grad_norm": 0.03021426685154438,
+ "learning_rate": 0.0006,
+ "loss": 6.677826404571533,
+ "step": 3188
+ },
+ {
+ "epoch": 44.29344978165939,
+ "grad_norm": 0.032727014273405075,
+ "learning_rate": 0.0006,
+ "loss": 6.657764434814453,
+ "step": 3189
+ },
+ {
+ "epoch": 44.30742358078603,
+ "grad_norm": 0.03696778789162636,
+ "learning_rate": 0.0006,
+ "loss": 6.632883548736572,
+ "step": 3190
+ },
+ {
+ "epoch": 44.32139737991266,
+ "grad_norm": 0.034630827605724335,
+ "learning_rate": 0.0006,
+ "loss": 6.759583473205566,
+ "step": 3191
+ },
+ {
+ "epoch": 44.3353711790393,
+ "grad_norm": 0.030517758801579475,
+ "learning_rate": 0.0006,
+ "loss": 6.702268600463867,
+ "step": 3192
+ },
+ {
+ "epoch": 44.34934497816594,
+ "grad_norm": 0.04105190932750702,
+ "learning_rate": 0.0006,
+ "loss": 6.705792427062988,
+ "step": 3193
+ },
+ {
+ "epoch": 44.36331877729258,
+ "grad_norm": 0.06659174710512161,
+ "learning_rate": 0.0006,
+ "loss": 6.73862361907959,
+ "step": 3194
+ },
+ {
+ "epoch": 44.377292576419215,
+ "grad_norm": 0.08013619482517242,
+ "learning_rate": 0.0006,
+ "loss": 6.710282802581787,
+ "step": 3195
+ },
+ {
+ "epoch": 44.391266375545854,
+ "grad_norm": 0.049850963056087494,
+ "learning_rate": 0.0006,
+ "loss": 6.684974670410156,
+ "step": 3196
+ },
+ {
+ "epoch": 44.40524017467249,
+ "grad_norm": 0.061247408390045166,
+ "learning_rate": 0.0006,
+ "loss": 6.716034412384033,
+ "step": 3197
+ },
+ {
+ "epoch": 44.419213973799124,
+ "grad_norm": 0.05868857726454735,
+ "learning_rate": 0.0006,
+ "loss": 6.589107513427734,
+ "step": 3198
+ },
+ {
+ "epoch": 44.43318777292576,
+ "grad_norm": 0.048878561705350876,
+ "learning_rate": 0.0006,
+ "loss": 6.694855690002441,
+ "step": 3199
+ },
+ {
+ "epoch": 44.4471615720524,
+ "grad_norm": 0.028748324140906334,
+ "learning_rate": 0.0006,
+ "loss": 6.629969596862793,
+ "step": 3200
+ },
+ {
+ "epoch": 44.46113537117904,
+ "grad_norm": 0.04060306400060654,
+ "learning_rate": 0.0006,
+ "loss": 6.721431255340576,
+ "step": 3201
+ },
+ {
+ "epoch": 44.47510917030568,
+ "grad_norm": 0.03833824768662453,
+ "learning_rate": 0.0006,
+ "loss": 6.625029563903809,
+ "step": 3202
+ },
+ {
+ "epoch": 44.48908296943232,
+ "grad_norm": 0.03791997581720352,
+ "learning_rate": 0.0006,
+ "loss": 6.669187545776367,
+ "step": 3203
+ },
+ {
+ "epoch": 44.50305676855895,
+ "grad_norm": 0.03259684890508652,
+ "learning_rate": 0.0006,
+ "loss": 6.665643692016602,
+ "step": 3204
+ },
+ {
+ "epoch": 44.51703056768559,
+ "grad_norm": 0.040658626705408096,
+ "learning_rate": 0.0006,
+ "loss": 6.639796257019043,
+ "step": 3205
+ },
+ {
+ "epoch": 44.531004366812226,
+ "grad_norm": 0.03691193461418152,
+ "learning_rate": 0.0006,
+ "loss": 6.692244529724121,
+ "step": 3206
+ },
+ {
+ "epoch": 44.544978165938865,
+ "grad_norm": 0.02814812771975994,
+ "learning_rate": 0.0006,
+ "loss": 6.61978006362915,
+ "step": 3207
+ },
+ {
+ "epoch": 44.5589519650655,
+ "grad_norm": 0.022155074402689934,
+ "learning_rate": 0.0006,
+ "loss": 6.625905990600586,
+ "step": 3208
+ },
+ {
+ "epoch": 44.57292576419214,
+ "grad_norm": 0.030794885009527206,
+ "learning_rate": 0.0006,
+ "loss": 6.678500652313232,
+ "step": 3209
+ },
+ {
+ "epoch": 44.58689956331878,
+ "grad_norm": 0.03201768547296524,
+ "learning_rate": 0.0006,
+ "loss": 6.6015729904174805,
+ "step": 3210
+ },
+ {
+ "epoch": 44.60087336244541,
+ "grad_norm": 0.025138361379504204,
+ "learning_rate": 0.0006,
+ "loss": 6.568234443664551,
+ "step": 3211
+ },
+ {
+ "epoch": 44.61484716157205,
+ "grad_norm": 0.02969353087246418,
+ "learning_rate": 0.0006,
+ "loss": 6.588366508483887,
+ "step": 3212
+ },
+ {
+ "epoch": 44.62882096069869,
+ "grad_norm": 0.03136962279677391,
+ "learning_rate": 0.0006,
+ "loss": 6.479961395263672,
+ "step": 3213
+ },
+ {
+ "epoch": 44.64279475982533,
+ "grad_norm": 0.030178170651197433,
+ "learning_rate": 0.0006,
+ "loss": 6.622910976409912,
+ "step": 3214
+ },
+ {
+ "epoch": 44.65676855895197,
+ "grad_norm": 0.02682245895266533,
+ "learning_rate": 0.0006,
+ "loss": 6.597707748413086,
+ "step": 3215
+ },
+ {
+ "epoch": 44.670742358078606,
+ "grad_norm": 0.025154557079076767,
+ "learning_rate": 0.0006,
+ "loss": 6.630541801452637,
+ "step": 3216
+ },
+ {
+ "epoch": 44.68471615720524,
+ "grad_norm": 0.03564463183283806,
+ "learning_rate": 0.0006,
+ "loss": 6.542823791503906,
+ "step": 3217
+ },
+ {
+ "epoch": 44.698689956331876,
+ "grad_norm": 0.049477651715278625,
+ "learning_rate": 0.0006,
+ "loss": 6.646030426025391,
+ "step": 3218
+ },
+ {
+ "epoch": 44.712663755458514,
+ "grad_norm": 0.06536535918712616,
+ "learning_rate": 0.0006,
+ "loss": 6.595463275909424,
+ "step": 3219
+ },
+ {
+ "epoch": 44.72663755458515,
+ "grad_norm": 0.07710316777229309,
+ "learning_rate": 0.0006,
+ "loss": 6.55524206161499,
+ "step": 3220
+ },
+ {
+ "epoch": 44.74061135371179,
+ "grad_norm": 0.09571163356304169,
+ "learning_rate": 0.0006,
+ "loss": 6.659764289855957,
+ "step": 3221
+ },
+ {
+ "epoch": 44.75458515283843,
+ "grad_norm": 0.055076830089092255,
+ "learning_rate": 0.0006,
+ "loss": 6.658882141113281,
+ "step": 3222
+ },
+ {
+ "epoch": 44.76855895196506,
+ "grad_norm": 0.07671099156141281,
+ "learning_rate": 0.0006,
+ "loss": 6.605809211730957,
+ "step": 3223
+ },
+ {
+ "epoch": 44.7825327510917,
+ "grad_norm": 0.05295584350824356,
+ "learning_rate": 0.0006,
+ "loss": 6.641455173492432,
+ "step": 3224
+ },
+ {
+ "epoch": 44.79650655021834,
+ "grad_norm": 0.09128042310476303,
+ "learning_rate": 0.0006,
+ "loss": 6.691066741943359,
+ "step": 3225
+ },
+ {
+ "epoch": 44.81048034934498,
+ "grad_norm": 0.05362958461046219,
+ "learning_rate": 0.0006,
+ "loss": 6.573406219482422,
+ "step": 3226
+ },
+ {
+ "epoch": 44.82445414847162,
+ "grad_norm": 0.046910062432289124,
+ "learning_rate": 0.0006,
+ "loss": 6.550876617431641,
+ "step": 3227
+ },
+ {
+ "epoch": 44.838427947598255,
+ "grad_norm": 0.03494197130203247,
+ "learning_rate": 0.0006,
+ "loss": 6.540179252624512,
+ "step": 3228
+ },
+ {
+ "epoch": 44.852401746724894,
+ "grad_norm": 0.04716387391090393,
+ "learning_rate": 0.0006,
+ "loss": 6.490459442138672,
+ "step": 3229
+ },
+ {
+ "epoch": 44.866375545851525,
+ "grad_norm": 0.04190506786108017,
+ "learning_rate": 0.0006,
+ "loss": 6.509925365447998,
+ "step": 3230
+ },
+ {
+ "epoch": 44.880349344978164,
+ "grad_norm": 0.04749064892530441,
+ "learning_rate": 0.0006,
+ "loss": 6.478779315948486,
+ "step": 3231
+ },
+ {
+ "epoch": 44.8943231441048,
+ "grad_norm": 0.028002185747027397,
+ "learning_rate": 0.0006,
+ "loss": 6.528904914855957,
+ "step": 3232
+ },
+ {
+ "epoch": 44.90829694323144,
+ "grad_norm": 0.036473166197538376,
+ "learning_rate": 0.0006,
+ "loss": 6.549042701721191,
+ "step": 3233
+ },
+ {
+ "epoch": 44.92227074235808,
+ "grad_norm": 0.02543315477669239,
+ "learning_rate": 0.0006,
+ "loss": 6.498875141143799,
+ "step": 3234
+ },
+ {
+ "epoch": 44.93624454148472,
+ "grad_norm": 0.029906807467341423,
+ "learning_rate": 0.0006,
+ "loss": 6.453019142150879,
+ "step": 3235
+ },
+ {
+ "epoch": 44.95021834061135,
+ "grad_norm": 0.027231857180595398,
+ "learning_rate": 0.0006,
+ "loss": 6.428829193115234,
+ "step": 3236
+ },
+ {
+ "epoch": 44.96419213973799,
+ "grad_norm": 0.02506154775619507,
+ "learning_rate": 0.0006,
+ "loss": 6.420858383178711,
+ "step": 3237
+ },
+ {
+ "epoch": 44.97816593886463,
+ "grad_norm": 0.0269472673535347,
+ "learning_rate": 0.0006,
+ "loss": 6.484989643096924,
+ "step": 3238
+ },
+ {
+ "epoch": 44.992139737991266,
+ "grad_norm": 0.023410487920045853,
+ "learning_rate": 0.0006,
+ "loss": 6.454830169677734,
+ "step": 3239
+ },
+ {
+ "epoch": 45.0,
+ "grad_norm": 0.034175023436546326,
+ "learning_rate": 0.0006,
+ "loss": 6.430975437164307,
+ "step": 3240
+ },
+ {
+ "epoch": 45.0,
+ "eval_loss": 6.458896160125732,
+ "eval_runtime": 56.6934,
+ "eval_samples_per_second": 43.074,
+ "eval_steps_per_second": 1.358,
+ "step": 3240
+ },
+ {
+ "epoch": 45.01397379912664,
+ "grad_norm": 0.045031629502773285,
+ "learning_rate": 0.0006,
+ "loss": 6.4243011474609375,
+ "step": 3241
+ },
+ {
+ "epoch": 45.02794759825328,
+ "grad_norm": 0.04968734085559845,
+ "learning_rate": 0.0006,
+ "loss": 6.380091667175293,
+ "step": 3242
+ },
+ {
+ "epoch": 45.041921397379916,
+ "grad_norm": 0.047145187854766846,
+ "learning_rate": 0.0006,
+ "loss": 6.477178573608398,
+ "step": 3243
+ },
+ {
+ "epoch": 45.05589519650655,
+ "grad_norm": 0.04304948449134827,
+ "learning_rate": 0.0006,
+ "loss": 6.399726867675781,
+ "step": 3244
+ },
+ {
+ "epoch": 45.069868995633186,
+ "grad_norm": 0.04466010630130768,
+ "learning_rate": 0.0006,
+ "loss": 6.3415679931640625,
+ "step": 3245
+ },
+ {
+ "epoch": 45.083842794759825,
+ "grad_norm": 0.04305306077003479,
+ "learning_rate": 0.0006,
+ "loss": 6.3413310050964355,
+ "step": 3246
+ },
+ {
+ "epoch": 45.09781659388646,
+ "grad_norm": 0.043064575642347336,
+ "learning_rate": 0.0006,
+ "loss": 6.4838714599609375,
+ "step": 3247
+ },
+ {
+ "epoch": 45.1117903930131,
+ "grad_norm": 0.049503277987241745,
+ "learning_rate": 0.0006,
+ "loss": 6.38473653793335,
+ "step": 3248
+ },
+ {
+ "epoch": 45.12576419213974,
+ "grad_norm": 0.04613238573074341,
+ "learning_rate": 0.0006,
+ "loss": 6.370769500732422,
+ "step": 3249
+ },
+ {
+ "epoch": 45.13973799126637,
+ "grad_norm": 0.052119478583335876,
+ "learning_rate": 0.0006,
+ "loss": 6.362321853637695,
+ "step": 3250
+ },
+ {
+ "epoch": 45.15371179039301,
+ "grad_norm": 0.045110076665878296,
+ "learning_rate": 0.0006,
+ "loss": 6.369844913482666,
+ "step": 3251
+ },
+ {
+ "epoch": 45.16768558951965,
+ "grad_norm": 0.03798583149909973,
+ "learning_rate": 0.0006,
+ "loss": 6.324633598327637,
+ "step": 3252
+ },
+ {
+ "epoch": 45.18165938864629,
+ "grad_norm": 0.03744659945368767,
+ "learning_rate": 0.0006,
+ "loss": 6.349421501159668,
+ "step": 3253
+ },
+ {
+ "epoch": 45.19563318777293,
+ "grad_norm": 0.03767292574048042,
+ "learning_rate": 0.0006,
+ "loss": 6.334238052368164,
+ "step": 3254
+ },
+ {
+ "epoch": 45.209606986899566,
+ "grad_norm": 0.03577509894967079,
+ "learning_rate": 0.0006,
+ "loss": 6.3453826904296875,
+ "step": 3255
+ },
+ {
+ "epoch": 45.223580786026204,
+ "grad_norm": 0.03403240442276001,
+ "learning_rate": 0.0006,
+ "loss": 6.2836174964904785,
+ "step": 3256
+ },
+ {
+ "epoch": 45.237554585152836,
+ "grad_norm": 0.03517331928014755,
+ "learning_rate": 0.0006,
+ "loss": 6.281718730926514,
+ "step": 3257
+ },
+ {
+ "epoch": 45.251528384279474,
+ "grad_norm": 0.04409139230847359,
+ "learning_rate": 0.0006,
+ "loss": 6.320815086364746,
+ "step": 3258
+ },
+ {
+ "epoch": 45.26550218340611,
+ "grad_norm": 0.054750364273786545,
+ "learning_rate": 0.0006,
+ "loss": 6.343618392944336,
+ "step": 3259
+ },
+ {
+ "epoch": 45.27947598253275,
+ "grad_norm": 0.07612067461013794,
+ "learning_rate": 0.0006,
+ "loss": 6.310614585876465,
+ "step": 3260
+ },
+ {
+ "epoch": 45.29344978165939,
+ "grad_norm": 0.08438985049724579,
+ "learning_rate": 0.0006,
+ "loss": 6.2369184494018555,
+ "step": 3261
+ },
+ {
+ "epoch": 45.30742358078603,
+ "grad_norm": 0.0782201811671257,
+ "learning_rate": 0.0006,
+ "loss": 6.254452705383301,
+ "step": 3262
+ },
+ {
+ "epoch": 45.32139737991266,
+ "grad_norm": 0.09928355365991592,
+ "learning_rate": 0.0006,
+ "loss": 6.3112077713012695,
+ "step": 3263
+ },
+ {
+ "epoch": 45.3353711790393,
+ "grad_norm": 0.07233133912086487,
+ "learning_rate": 0.0006,
+ "loss": 6.408087253570557,
+ "step": 3264
+ },
+ {
+ "epoch": 45.34934497816594,
+ "grad_norm": 0.07478684186935425,
+ "learning_rate": 0.0006,
+ "loss": 6.3770036697387695,
+ "step": 3265
+ },
+ {
+ "epoch": 45.36331877729258,
+ "grad_norm": 0.06036846712231636,
+ "learning_rate": 0.0006,
+ "loss": 6.33865213394165,
+ "step": 3266
+ },
+ {
+ "epoch": 45.377292576419215,
+ "grad_norm": 0.05783458054065704,
+ "learning_rate": 0.0006,
+ "loss": 6.363203048706055,
+ "step": 3267
+ },
+ {
+ "epoch": 45.391266375545854,
+ "grad_norm": 0.05443909019231796,
+ "learning_rate": 0.0006,
+ "loss": 6.287069797515869,
+ "step": 3268
+ },
+ {
+ "epoch": 45.40524017467249,
+ "grad_norm": 0.0557909794151783,
+ "learning_rate": 0.0006,
+ "loss": 6.297784805297852,
+ "step": 3269
+ },
+ {
+ "epoch": 45.419213973799124,
+ "grad_norm": 0.03483372926712036,
+ "learning_rate": 0.0006,
+ "loss": 6.280734062194824,
+ "step": 3270
+ },
+ {
+ "epoch": 45.43318777292576,
+ "grad_norm": 0.0425543375313282,
+ "learning_rate": 0.0006,
+ "loss": 6.33565616607666,
+ "step": 3271
+ },
+ {
+ "epoch": 45.4471615720524,
+ "grad_norm": 0.03230159357190132,
+ "learning_rate": 0.0006,
+ "loss": 6.197698593139648,
+ "step": 3272
+ },
+ {
+ "epoch": 45.46113537117904,
+ "grad_norm": 0.036897290498018265,
+ "learning_rate": 0.0006,
+ "loss": 6.241239547729492,
+ "step": 3273
+ },
+ {
+ "epoch": 45.47510917030568,
+ "grad_norm": 0.031103838235139847,
+ "learning_rate": 0.0006,
+ "loss": 6.159435749053955,
+ "step": 3274
+ },
+ {
+ "epoch": 45.48908296943232,
+ "grad_norm": 0.027116449549794197,
+ "learning_rate": 0.0006,
+ "loss": 6.212225914001465,
+ "step": 3275
+ },
+ {
+ "epoch": 45.50305676855895,
+ "grad_norm": 0.029923981055617332,
+ "learning_rate": 0.0006,
+ "loss": 6.157269477844238,
+ "step": 3276
+ },
+ {
+ "epoch": 45.51703056768559,
+ "grad_norm": 0.02521538734436035,
+ "learning_rate": 0.0006,
+ "loss": 6.146918773651123,
+ "step": 3277
+ },
+ {
+ "epoch": 45.531004366812226,
+ "grad_norm": 0.02758920192718506,
+ "learning_rate": 0.0006,
+ "loss": 6.0446977615356445,
+ "step": 3278
+ },
+ {
+ "epoch": 45.544978165938865,
+ "grad_norm": 0.024328220635652542,
+ "learning_rate": 0.0006,
+ "loss": 6.099508285522461,
+ "step": 3279
+ },
+ {
+ "epoch": 45.5589519650655,
+ "grad_norm": 0.027425771579146385,
+ "learning_rate": 0.0006,
+ "loss": 6.084299087524414,
+ "step": 3280
+ },
+ {
+ "epoch": 45.57292576419214,
+ "grad_norm": 0.026927651837468147,
+ "learning_rate": 0.0006,
+ "loss": 6.146334171295166,
+ "step": 3281
+ },
+ {
+ "epoch": 45.58689956331878,
+ "grad_norm": 0.04054379090666771,
+ "learning_rate": 0.0006,
+ "loss": 6.1301493644714355,
+ "step": 3282
+ },
+ {
+ "epoch": 45.60087336244541,
+ "grad_norm": 0.04456290975213051,
+ "learning_rate": 0.0006,
+ "loss": 6.152511119842529,
+ "step": 3283
+ },
+ {
+ "epoch": 45.61484716157205,
+ "grad_norm": 0.052821505814790726,
+ "learning_rate": 0.0006,
+ "loss": 6.113544940948486,
+ "step": 3284
+ },
+ {
+ "epoch": 45.62882096069869,
+ "grad_norm": 0.07251739501953125,
+ "learning_rate": 0.0006,
+ "loss": 6.118298530578613,
+ "step": 3285
+ },
+ {
+ "epoch": 45.64279475982533,
+ "grad_norm": 0.09433556348085403,
+ "learning_rate": 0.0006,
+ "loss": 6.170897483825684,
+ "step": 3286
+ },
+ {
+ "epoch": 45.65676855895197,
+ "grad_norm": 0.057683493942022324,
+ "learning_rate": 0.0006,
+ "loss": 6.1747822761535645,
+ "step": 3287
+ },
+ {
+ "epoch": 45.670742358078606,
+ "grad_norm": 0.06363270431756973,
+ "learning_rate": 0.0006,
+ "loss": 6.1617889404296875,
+ "step": 3288
+ },
+ {
+ "epoch": 45.68471615720524,
+ "grad_norm": 0.07528169453144073,
+ "learning_rate": 0.0006,
+ "loss": 6.230060577392578,
+ "step": 3289
+ },
+ {
+ "epoch": 45.698689956331876,
+ "grad_norm": 0.07816170156002045,
+ "learning_rate": 0.0006,
+ "loss": 6.171257019042969,
+ "step": 3290
+ },
+ {
+ "epoch": 45.712663755458514,
+ "grad_norm": 0.05860164389014244,
+ "learning_rate": 0.0006,
+ "loss": 6.095484256744385,
+ "step": 3291
+ },
+ {
+ "epoch": 45.72663755458515,
+ "grad_norm": 0.04747429117560387,
+ "learning_rate": 0.0006,
+ "loss": 6.060067176818848,
+ "step": 3292
+ },
+ {
+ "epoch": 45.74061135371179,
+ "grad_norm": 0.04514474794268608,
+ "learning_rate": 0.0006,
+ "loss": 6.100218772888184,
+ "step": 3293
+ },
+ {
+ "epoch": 45.75458515283843,
+ "grad_norm": 0.05168681964278221,
+ "learning_rate": 0.0006,
+ "loss": 6.0621867179870605,
+ "step": 3294
+ },
+ {
+ "epoch": 45.76855895196506,
+ "grad_norm": 0.04665876924991608,
+ "learning_rate": 0.0006,
+ "loss": 6.118292808532715,
+ "step": 3295
+ },
+ {
+ "epoch": 45.7825327510917,
+ "grad_norm": 0.049374762922525406,
+ "learning_rate": 0.0006,
+ "loss": 6.098394393920898,
+ "step": 3296
+ },
+ {
+ "epoch": 45.79650655021834,
+ "grad_norm": 0.03793549910187721,
+ "learning_rate": 0.0006,
+ "loss": 6.069541931152344,
+ "step": 3297
+ },
+ {
+ "epoch": 45.81048034934498,
+ "grad_norm": 0.036362383514642715,
+ "learning_rate": 0.0006,
+ "loss": 6.1188435554504395,
+ "step": 3298
+ },
+ {
+ "epoch": 45.82445414847162,
+ "grad_norm": 0.03397097811102867,
+ "learning_rate": 0.0006,
+ "loss": 6.117589950561523,
+ "step": 3299
+ },
+ {
+ "epoch": 45.838427947598255,
+ "grad_norm": 0.030977753922343254,
+ "learning_rate": 0.0006,
+ "loss": 6.0236711502075195,
+ "step": 3300
+ },
+ {
+ "epoch": 45.852401746724894,
+ "grad_norm": 0.02800217643380165,
+ "learning_rate": 0.0006,
+ "loss": 6.034825325012207,
+ "step": 3301
+ },
+ {
+ "epoch": 45.866375545851525,
+ "grad_norm": 0.03129778802394867,
+ "learning_rate": 0.0006,
+ "loss": 5.980271339416504,
+ "step": 3302
+ },
+ {
+ "epoch": 45.880349344978164,
+ "grad_norm": 0.025886094197630882,
+ "learning_rate": 0.0006,
+ "loss": 5.966419219970703,
+ "step": 3303
+ },
+ {
+ "epoch": 45.8943231441048,
+ "grad_norm": 0.025712328031659126,
+ "learning_rate": 0.0006,
+ "loss": 5.938652515411377,
+ "step": 3304
+ },
+ {
+ "epoch": 45.90829694323144,
+ "grad_norm": 0.022508298978209496,
+ "learning_rate": 0.0006,
+ "loss": 5.982312202453613,
+ "step": 3305
+ },
+ {
+ "epoch": 45.92227074235808,
+ "grad_norm": 0.024880951270461082,
+ "learning_rate": 0.0006,
+ "loss": 5.931150436401367,
+ "step": 3306
+ },
+ {
+ "epoch": 45.93624454148472,
+ "grad_norm": 0.030007576569914818,
+ "learning_rate": 0.0006,
+ "loss": 5.936639308929443,
+ "step": 3307
+ },
+ {
+ "epoch": 45.95021834061135,
+ "grad_norm": 0.033314041793346405,
+ "learning_rate": 0.0006,
+ "loss": 5.944248676300049,
+ "step": 3308
+ },
+ {
+ "epoch": 45.96419213973799,
+ "grad_norm": 0.043078381568193436,
+ "learning_rate": 0.0006,
+ "loss": 5.971286773681641,
+ "step": 3309
+ },
+ {
+ "epoch": 45.97816593886463,
+ "grad_norm": 0.059296440333127975,
+ "learning_rate": 0.0006,
+ "loss": 5.910497665405273,
+ "step": 3310
+ },
+ {
+ "epoch": 45.992139737991266,
+ "grad_norm": 0.06232666224241257,
+ "learning_rate": 0.0006,
+ "loss": 5.929817199707031,
+ "step": 3311
+ },
+ {
+ "epoch": 46.0,
+ "grad_norm": 0.05303940549492836,
+ "learning_rate": 0.0006,
+ "loss": 5.910822868347168,
+ "step": 3312
+ },
+ {
+ "epoch": 46.0,
+ "eval_loss": 5.977851390838623,
+ "eval_runtime": 56.2435,
+ "eval_samples_per_second": 43.418,
+ "eval_steps_per_second": 1.369,
+ "step": 3312
+ },
+ {
+ "epoch": 46.01397379912664,
+ "grad_norm": 0.06943798810243607,
+ "learning_rate": 0.0006,
+ "loss": 5.854922294616699,
+ "step": 3313
+ },
+ {
+ "epoch": 46.02794759825328,
+ "grad_norm": 0.05236278846859932,
+ "learning_rate": 0.0006,
+ "loss": 5.876611232757568,
+ "step": 3314
+ },
+ {
+ "epoch": 46.041921397379916,
+ "grad_norm": 0.06240001320838928,
+ "learning_rate": 0.0006,
+ "loss": 5.878912448883057,
+ "step": 3315
+ },
+ {
+ "epoch": 46.05589519650655,
+ "grad_norm": 0.07401192933320999,
+ "learning_rate": 0.0006,
+ "loss": 5.985136032104492,
+ "step": 3316
+ },
+ {
+ "epoch": 46.069868995633186,
+ "grad_norm": 0.0826529711484909,
+ "learning_rate": 0.0006,
+ "loss": 5.939439296722412,
+ "step": 3317
+ },
+ {
+ "epoch": 46.083842794759825,
+ "grad_norm": 0.07999005168676376,
+ "learning_rate": 0.0006,
+ "loss": 5.982336044311523,
+ "step": 3318
+ },
+ {
+ "epoch": 46.09781659388646,
+ "grad_norm": 0.07014869153499603,
+ "learning_rate": 0.0006,
+ "loss": 5.896534442901611,
+ "step": 3319
+ },
+ {
+ "epoch": 46.1117903930131,
+ "grad_norm": 0.05795128643512726,
+ "learning_rate": 0.0006,
+ "loss": 5.843111991882324,
+ "step": 3320
+ },
+ {
+ "epoch": 46.12576419213974,
+ "grad_norm": 0.06374258548021317,
+ "learning_rate": 0.0006,
+ "loss": 5.78752326965332,
+ "step": 3321
+ },
+ {
+ "epoch": 46.13973799126637,
+ "grad_norm": 0.045524805784225464,
+ "learning_rate": 0.0006,
+ "loss": 5.86755895614624,
+ "step": 3322
+ },
+ {
+ "epoch": 46.15371179039301,
+ "grad_norm": 0.03773504123091698,
+ "learning_rate": 0.0006,
+ "loss": 5.83627462387085,
+ "step": 3323
+ },
+ {
+ "epoch": 46.16768558951965,
+ "grad_norm": 0.035815201699733734,
+ "learning_rate": 0.0006,
+ "loss": 5.798238277435303,
+ "step": 3324
+ },
+ {
+ "epoch": 46.18165938864629,
+ "grad_norm": 0.03756783902645111,
+ "learning_rate": 0.0006,
+ "loss": 5.845197677612305,
+ "step": 3325
+ },
+ {
+ "epoch": 46.19563318777293,
+ "grad_norm": 0.03569113463163376,
+ "learning_rate": 0.0006,
+ "loss": 5.888932704925537,
+ "step": 3326
+ },
+ {
+ "epoch": 46.209606986899566,
+ "grad_norm": 0.029467813670635223,
+ "learning_rate": 0.0006,
+ "loss": 5.792047500610352,
+ "step": 3327
+ },
+ {
+ "epoch": 46.223580786026204,
+ "grad_norm": 0.025698406621813774,
+ "learning_rate": 0.0006,
+ "loss": 5.783734321594238,
+ "step": 3328
+ },
+ {
+ "epoch": 46.237554585152836,
+ "grad_norm": 0.028426671400666237,
+ "learning_rate": 0.0006,
+ "loss": 5.762084007263184,
+ "step": 3329
+ },
+ {
+ "epoch": 46.251528384279474,
+ "grad_norm": 0.03099622018635273,
+ "learning_rate": 0.0006,
+ "loss": 5.719109058380127,
+ "step": 3330
+ },
+ {
+ "epoch": 46.26550218340611,
+ "grad_norm": 0.0310219656676054,
+ "learning_rate": 0.0006,
+ "loss": 5.737057209014893,
+ "step": 3331
+ },
+ {
+ "epoch": 46.27947598253275,
+ "grad_norm": 0.02713446132838726,
+ "learning_rate": 0.0006,
+ "loss": 5.758657455444336,
+ "step": 3332
+ },
+ {
+ "epoch": 46.29344978165939,
+ "grad_norm": 0.03099844604730606,
+ "learning_rate": 0.0006,
+ "loss": 5.742086410522461,
+ "step": 3333
+ },
+ {
+ "epoch": 46.30742358078603,
+ "grad_norm": 0.030258730053901672,
+ "learning_rate": 0.0006,
+ "loss": 5.767426490783691,
+ "step": 3334
+ },
+ {
+ "epoch": 46.32139737991266,
+ "grad_norm": 0.029034852981567383,
+ "learning_rate": 0.0006,
+ "loss": 5.651792049407959,
+ "step": 3335
+ },
+ {
+ "epoch": 46.3353711790393,
+ "grad_norm": 0.03138665854930878,
+ "learning_rate": 0.0006,
+ "loss": 5.664980888366699,
+ "step": 3336
+ },
+ {
+ "epoch": 46.34934497816594,
+ "grad_norm": 0.040546901524066925,
+ "learning_rate": 0.0006,
+ "loss": 5.665165901184082,
+ "step": 3337
+ },
+ {
+ "epoch": 46.36331877729258,
+ "grad_norm": 0.06395485252141953,
+ "learning_rate": 0.0006,
+ "loss": 5.7776713371276855,
+ "step": 3338
+ },
+ {
+ "epoch": 46.377292576419215,
+ "grad_norm": 0.10088056325912476,
+ "learning_rate": 0.0006,
+ "loss": 5.755645751953125,
+ "step": 3339
+ },
+ {
+ "epoch": 46.391266375545854,
+ "grad_norm": 0.09527938067913055,
+ "learning_rate": 0.0006,
+ "loss": 5.838809490203857,
+ "step": 3340
+ },
+ {
+ "epoch": 46.40524017467249,
+ "grad_norm": 0.0849648267030716,
+ "learning_rate": 0.0006,
+ "loss": 5.781932830810547,
+ "step": 3341
+ },
+ {
+ "epoch": 46.419213973799124,
+ "grad_norm": 0.10142666101455688,
+ "learning_rate": 0.0006,
+ "loss": 5.838970184326172,
+ "step": 3342
+ },
+ {
+ "epoch": 46.43318777292576,
+ "grad_norm": 0.42731454968452454,
+ "learning_rate": 0.0006,
+ "loss": 6.071902275085449,
+ "step": 3343
+ },
+ {
+ "epoch": 46.4471615720524,
+ "grad_norm": 0.13102488219738007,
+ "learning_rate": 0.0006,
+ "loss": 6.8635993003845215,
+ "step": 3344
+ },
+ {
+ "epoch": 46.46113537117904,
+ "grad_norm": 0.23691938817501068,
+ "learning_rate": 0.0006,
+ "loss": 7.313158988952637,
+ "step": 3345
+ },
+ {
+ "epoch": 46.47510917030568,
+ "grad_norm": 0.11670669168233871,
+ "learning_rate": 0.0006,
+ "loss": 7.240363121032715,
+ "step": 3346
+ },
+ {
+ "epoch": 46.48908296943232,
+ "grad_norm": 0.09985820949077606,
+ "learning_rate": 0.0006,
+ "loss": 7.113332271575928,
+ "step": 3347
+ },
+ {
+ "epoch": 46.50305676855895,
+ "grad_norm": 0.11406408250331879,
+ "learning_rate": 0.0006,
+ "loss": 7.051285743713379,
+ "step": 3348
+ },
+ {
+ "epoch": 46.51703056768559,
+ "grad_norm": 0.07900336384773254,
+ "learning_rate": 0.0006,
+ "loss": 7.014365196228027,
+ "step": 3349
+ },
+ {
+ "epoch": 46.531004366812226,
+ "grad_norm": 0.05380333960056305,
+ "learning_rate": 0.0006,
+ "loss": 6.8868818283081055,
+ "step": 3350
+ },
+ {
+ "epoch": 46.544978165938865,
+ "grad_norm": 0.05868449807167053,
+ "learning_rate": 0.0006,
+ "loss": 6.783599853515625,
+ "step": 3351
+ },
+ {
+ "epoch": 46.5589519650655,
+ "grad_norm": 0.04428785294294357,
+ "learning_rate": 0.0006,
+ "loss": 6.798511505126953,
+ "step": 3352
+ },
+ {
+ "epoch": 46.57292576419214,
+ "grad_norm": 0.055527277290821075,
+ "learning_rate": 0.0006,
+ "loss": 6.694507598876953,
+ "step": 3353
+ },
+ {
+ "epoch": 46.58689956331878,
+ "grad_norm": 0.03872432932257652,
+ "learning_rate": 0.0006,
+ "loss": 6.736631393432617,
+ "step": 3354
+ },
+ {
+ "epoch": 46.60087336244541,
+ "grad_norm": 0.04330779239535332,
+ "learning_rate": 0.0006,
+ "loss": 6.576806545257568,
+ "step": 3355
+ },
+ {
+ "epoch": 46.61484716157205,
+ "grad_norm": 0.036767736077308655,
+ "learning_rate": 0.0006,
+ "loss": 6.559871196746826,
+ "step": 3356
+ },
+ {
+ "epoch": 46.62882096069869,
+ "grad_norm": 0.039857108145952225,
+ "learning_rate": 0.0006,
+ "loss": 6.5749831199646,
+ "step": 3357
+ },
+ {
+ "epoch": 46.64279475982533,
+ "grad_norm": 0.048719972372055054,
+ "learning_rate": 0.0006,
+ "loss": 6.4915618896484375,
+ "step": 3358
+ },
+ {
+ "epoch": 46.65676855895197,
+ "grad_norm": 0.042849015444517136,
+ "learning_rate": 0.0006,
+ "loss": 6.413516044616699,
+ "step": 3359
+ },
+ {
+ "epoch": 46.670742358078606,
+ "grad_norm": 0.02871009334921837,
+ "learning_rate": 0.0006,
+ "loss": 6.320952415466309,
+ "step": 3360
+ },
+ {
+ "epoch": 46.68471615720524,
+ "grad_norm": 0.03193620592355728,
+ "learning_rate": 0.0006,
+ "loss": 6.4176435470581055,
+ "step": 3361
+ },
+ {
+ "epoch": 46.698689956331876,
+ "grad_norm": 0.027417024597525597,
+ "learning_rate": 0.0006,
+ "loss": 6.342319488525391,
+ "step": 3362
+ },
+ {
+ "epoch": 46.712663755458514,
+ "grad_norm": 0.03305937349796295,
+ "learning_rate": 0.0006,
+ "loss": 6.246112823486328,
+ "step": 3363
+ },
+ {
+ "epoch": 46.72663755458515,
+ "grad_norm": 0.031403183937072754,
+ "learning_rate": 0.0006,
+ "loss": 6.149503231048584,
+ "step": 3364
+ },
+ {
+ "epoch": 46.74061135371179,
+ "grad_norm": 0.031563788652420044,
+ "learning_rate": 0.0006,
+ "loss": 6.278808116912842,
+ "step": 3365
+ },
+ {
+ "epoch": 46.75458515283843,
+ "grad_norm": 0.028366589918732643,
+ "learning_rate": 0.0006,
+ "loss": 6.057036876678467,
+ "step": 3366
+ },
+ {
+ "epoch": 46.76855895196506,
+ "grad_norm": 0.023715803399682045,
+ "learning_rate": 0.0006,
+ "loss": 6.111324310302734,
+ "step": 3367
+ },
+ {
+ "epoch": 46.7825327510917,
+ "grad_norm": 0.026546811684966087,
+ "learning_rate": 0.0006,
+ "loss": 6.105669975280762,
+ "step": 3368
+ },
+ {
+ "epoch": 46.79650655021834,
+ "grad_norm": 0.023531019687652588,
+ "learning_rate": 0.0006,
+ "loss": 6.068072319030762,
+ "step": 3369
+ },
+ {
+ "epoch": 46.81048034934498,
+ "grad_norm": 0.024480128660798073,
+ "learning_rate": 0.0006,
+ "loss": 5.947896957397461,
+ "step": 3370
+ },
+ {
+ "epoch": 46.82445414847162,
+ "grad_norm": 0.02517310529947281,
+ "learning_rate": 0.0006,
+ "loss": 6.030037879943848,
+ "step": 3371
+ },
+ {
+ "epoch": 46.838427947598255,
+ "grad_norm": 0.026723403483629227,
+ "learning_rate": 0.0006,
+ "loss": 5.963776111602783,
+ "step": 3372
+ },
+ {
+ "epoch": 46.852401746724894,
+ "grad_norm": 0.03163504600524902,
+ "learning_rate": 0.0006,
+ "loss": 5.893466949462891,
+ "step": 3373
+ },
+ {
+ "epoch": 46.866375545851525,
+ "grad_norm": 0.04058683291077614,
+ "learning_rate": 0.0006,
+ "loss": 5.9202799797058105,
+ "step": 3374
+ },
+ {
+ "epoch": 46.880349344978164,
+ "grad_norm": 0.04964117333292961,
+ "learning_rate": 0.0006,
+ "loss": 5.9366559982299805,
+ "step": 3375
+ },
+ {
+ "epoch": 46.8943231441048,
+ "grad_norm": 0.055406033992767334,
+ "learning_rate": 0.0006,
+ "loss": 5.912230968475342,
+ "step": 3376
+ },
+ {
+ "epoch": 46.90829694323144,
+ "grad_norm": 0.04643625393509865,
+ "learning_rate": 0.0006,
+ "loss": 5.87260103225708,
+ "step": 3377
+ },
+ {
+ "epoch": 46.92227074235808,
+ "grad_norm": 0.029325876384973526,
+ "learning_rate": 0.0006,
+ "loss": 5.854763984680176,
+ "step": 3378
+ },
+ {
+ "epoch": 46.93624454148472,
+ "grad_norm": 0.03159382566809654,
+ "learning_rate": 0.0006,
+ "loss": 5.880882263183594,
+ "step": 3379
+ },
+ {
+ "epoch": 46.95021834061135,
+ "grad_norm": 0.030224012210965157,
+ "learning_rate": 0.0006,
+ "loss": 5.787360191345215,
+ "step": 3380
+ },
+ {
+ "epoch": 46.96419213973799,
+ "grad_norm": 0.028368130326271057,
+ "learning_rate": 0.0006,
+ "loss": 5.817195892333984,
+ "step": 3381
+ },
+ {
+ "epoch": 46.97816593886463,
+ "grad_norm": 0.029411399737000465,
+ "learning_rate": 0.0006,
+ "loss": 5.783403396606445,
+ "step": 3382
+ },
+ {
+ "epoch": 46.992139737991266,
+ "grad_norm": 0.029927462339401245,
+ "learning_rate": 0.0006,
+ "loss": 5.662444591522217,
+ "step": 3383
+ },
+ {
+ "epoch": 47.0,
+ "grad_norm": 0.02872961387038231,
+ "learning_rate": 0.0006,
+ "loss": 5.742781639099121,
+ "step": 3384
+ },
+ {
+ "epoch": 47.0,
+ "eval_loss": 5.870260238647461,
+ "eval_runtime": 57.3734,
+ "eval_samples_per_second": 42.563,
+ "eval_steps_per_second": 1.342,
+ "step": 3384
+ },
+ {
+ "epoch": 47.01397379912664,
+ "grad_norm": 0.028542131185531616,
+ "learning_rate": 0.0006,
+ "loss": 5.753256797790527,
+ "step": 3385
+ },
+ {
+ "epoch": 47.02794759825328,
+ "grad_norm": 0.03227119892835617,
+ "learning_rate": 0.0006,
+ "loss": 5.722246170043945,
+ "step": 3386
+ },
+ {
+ "epoch": 47.041921397379916,
+ "grad_norm": 0.031461723148822784,
+ "learning_rate": 0.0006,
+ "loss": 5.740520000457764,
+ "step": 3387
+ },
+ {
+ "epoch": 47.05589519650655,
+ "grad_norm": 0.03184220567345619,
+ "learning_rate": 0.0006,
+ "loss": 5.720890045166016,
+ "step": 3388
+ },
+ {
+ "epoch": 47.069868995633186,
+ "grad_norm": 0.030015580356121063,
+ "learning_rate": 0.0006,
+ "loss": 5.597814559936523,
+ "step": 3389
+ },
+ {
+ "epoch": 47.083842794759825,
+ "grad_norm": 0.029931779950857162,
+ "learning_rate": 0.0006,
+ "loss": 5.5412468910217285,
+ "step": 3390
+ },
+ {
+ "epoch": 47.09781659388646,
+ "grad_norm": 0.03206505626440048,
+ "learning_rate": 0.0006,
+ "loss": 5.6967597007751465,
+ "step": 3391
+ },
+ {
+ "epoch": 47.1117903930131,
+ "grad_norm": 0.03978161886334419,
+ "learning_rate": 0.0006,
+ "loss": 5.59661340713501,
+ "step": 3392
+ },
+ {
+ "epoch": 47.12576419213974,
+ "grad_norm": 0.04619080200791359,
+ "learning_rate": 0.0006,
+ "loss": 5.795788764953613,
+ "step": 3393
+ },
+ {
+ "epoch": 47.13973799126637,
+ "grad_norm": 0.049085382372140884,
+ "learning_rate": 0.0006,
+ "loss": 5.667335510253906,
+ "step": 3394
+ },
+ {
+ "epoch": 47.15371179039301,
+ "grad_norm": 0.04830858111381531,
+ "learning_rate": 0.0006,
+ "loss": 5.668806552886963,
+ "step": 3395
+ },
+ {
+ "epoch": 47.16768558951965,
+ "grad_norm": 0.0583178736269474,
+ "learning_rate": 0.0006,
+ "loss": 5.692285537719727,
+ "step": 3396
+ },
+ {
+ "epoch": 47.18165938864629,
+ "grad_norm": 0.05182253196835518,
+ "learning_rate": 0.0006,
+ "loss": 5.678645133972168,
+ "step": 3397
+ },
+ {
+ "epoch": 47.19563318777293,
+ "grad_norm": 0.04011905938386917,
+ "learning_rate": 0.0006,
+ "loss": 5.661550998687744,
+ "step": 3398
+ },
+ {
+ "epoch": 47.209606986899566,
+ "grad_norm": 0.03888486698269844,
+ "learning_rate": 0.0006,
+ "loss": 5.643001556396484,
+ "step": 3399
+ },
+ {
+ "epoch": 47.223580786026204,
+ "grad_norm": 0.030658265575766563,
+ "learning_rate": 0.0006,
+ "loss": 5.659090995788574,
+ "step": 3400
+ },
+ {
+ "epoch": 47.237554585152836,
+ "grad_norm": 0.03609256446361542,
+ "learning_rate": 0.0006,
+ "loss": 5.544501781463623,
+ "step": 3401
+ },
+ {
+ "epoch": 47.251528384279474,
+ "grad_norm": 0.033589769154787064,
+ "learning_rate": 0.0006,
+ "loss": 5.463878631591797,
+ "step": 3402
+ },
+ {
+ "epoch": 47.26550218340611,
+ "grad_norm": 0.035128287971019745,
+ "learning_rate": 0.0006,
+ "loss": 5.5680060386657715,
+ "step": 3403
+ },
+ {
+ "epoch": 47.27947598253275,
+ "grad_norm": 0.038684550672769547,
+ "learning_rate": 0.0006,
+ "loss": 5.587898254394531,
+ "step": 3404
+ },
+ {
+ "epoch": 47.29344978165939,
+ "grad_norm": 0.043566226959228516,
+ "learning_rate": 0.0006,
+ "loss": 5.580896854400635,
+ "step": 3405
+ },
+ {
+ "epoch": 47.30742358078603,
+ "grad_norm": 0.04428374394774437,
+ "learning_rate": 0.0006,
+ "loss": 5.603460311889648,
+ "step": 3406
+ },
+ {
+ "epoch": 47.32139737991266,
+ "grad_norm": 0.037368934601545334,
+ "learning_rate": 0.0006,
+ "loss": 5.564740180969238,
+ "step": 3407
+ },
+ {
+ "epoch": 47.3353711790393,
+ "grad_norm": 0.04092350974678993,
+ "learning_rate": 0.0006,
+ "loss": 5.584153652191162,
+ "step": 3408
+ },
+ {
+ "epoch": 47.34934497816594,
+ "grad_norm": 0.045031506568193436,
+ "learning_rate": 0.0006,
+ "loss": 5.601186275482178,
+ "step": 3409
+ },
+ {
+ "epoch": 47.36331877729258,
+ "grad_norm": 0.041830409318208694,
+ "learning_rate": 0.0006,
+ "loss": 5.512788772583008,
+ "step": 3410
+ },
+ {
+ "epoch": 47.377292576419215,
+ "grad_norm": 0.037633974105119705,
+ "learning_rate": 0.0006,
+ "loss": 5.503828525543213,
+ "step": 3411
+ },
+ {
+ "epoch": 47.391266375545854,
+ "grad_norm": 0.03473588079214096,
+ "learning_rate": 0.0006,
+ "loss": 5.570557594299316,
+ "step": 3412
+ },
+ {
+ "epoch": 47.40524017467249,
+ "grad_norm": 0.039325181394815445,
+ "learning_rate": 0.0006,
+ "loss": 5.427783966064453,
+ "step": 3413
+ },
+ {
+ "epoch": 47.419213973799124,
+ "grad_norm": 0.04034624621272087,
+ "learning_rate": 0.0006,
+ "loss": 5.572591304779053,
+ "step": 3414
+ },
+ {
+ "epoch": 47.43318777292576,
+ "grad_norm": 0.05245184153318405,
+ "learning_rate": 0.0006,
+ "loss": 5.505208969116211,
+ "step": 3415
+ },
+ {
+ "epoch": 47.4471615720524,
+ "grad_norm": 0.0568675734102726,
+ "learning_rate": 0.0006,
+ "loss": 5.589373588562012,
+ "step": 3416
+ },
+ {
+ "epoch": 47.46113537117904,
+ "grad_norm": 0.048336587846279144,
+ "learning_rate": 0.0006,
+ "loss": 5.522005081176758,
+ "step": 3417
+ },
+ {
+ "epoch": 47.47510917030568,
+ "grad_norm": 0.04194975271821022,
+ "learning_rate": 0.0006,
+ "loss": 5.387576103210449,
+ "step": 3418
+ },
+ {
+ "epoch": 47.48908296943232,
+ "grad_norm": 0.04019433259963989,
+ "learning_rate": 0.0006,
+ "loss": 5.522615432739258,
+ "step": 3419
+ },
+ {
+ "epoch": 47.50305676855895,
+ "grad_norm": 0.05709933117032051,
+ "learning_rate": 0.0006,
+ "loss": 5.420286178588867,
+ "step": 3420
+ },
+ {
+ "epoch": 47.51703056768559,
+ "grad_norm": 0.09130338579416275,
+ "learning_rate": 0.0006,
+ "loss": 5.484622478485107,
+ "step": 3421
+ },
+ {
+ "epoch": 47.531004366812226,
+ "grad_norm": 0.2757541835308075,
+ "learning_rate": 0.0006,
+ "loss": 5.821832656860352,
+ "step": 3422
+ },
+ {
+ "epoch": 47.544978165938865,
+ "grad_norm": 0.28292080760002136,
+ "learning_rate": 0.0006,
+ "loss": 6.579934120178223,
+ "step": 3423
+ },
+ {
+ "epoch": 47.5589519650655,
+ "grad_norm": 0.15293604135513306,
+ "learning_rate": 0.0006,
+ "loss": 6.478213310241699,
+ "step": 3424
+ },
+ {
+ "epoch": 47.57292576419214,
+ "grad_norm": 0.11284393817186356,
+ "learning_rate": 0.0006,
+ "loss": 6.4066386222839355,
+ "step": 3425
+ },
+ {
+ "epoch": 47.58689956331878,
+ "grad_norm": 0.10933073610067368,
+ "learning_rate": 0.0006,
+ "loss": 6.3306169509887695,
+ "step": 3426
+ },
+ {
+ "epoch": 47.60087336244541,
+ "grad_norm": 0.05837641656398773,
+ "learning_rate": 0.0006,
+ "loss": 6.250000953674316,
+ "step": 3427
+ },
+ {
+ "epoch": 47.61484716157205,
+ "grad_norm": 0.05987229570746422,
+ "learning_rate": 0.0006,
+ "loss": 6.152463436126709,
+ "step": 3428
+ },
+ {
+ "epoch": 47.62882096069869,
+ "grad_norm": 0.04064938426017761,
+ "learning_rate": 0.0006,
+ "loss": 6.117897033691406,
+ "step": 3429
+ },
+ {
+ "epoch": 47.64279475982533,
+ "grad_norm": 0.04565057530999184,
+ "learning_rate": 0.0006,
+ "loss": 6.042291641235352,
+ "step": 3430
+ },
+ {
+ "epoch": 47.65676855895197,
+ "grad_norm": 0.03475063294172287,
+ "learning_rate": 0.0006,
+ "loss": 6.0071868896484375,
+ "step": 3431
+ },
+ {
+ "epoch": 47.670742358078606,
+ "grad_norm": 0.03809220343828201,
+ "learning_rate": 0.0006,
+ "loss": 6.0128021240234375,
+ "step": 3432
+ },
+ {
+ "epoch": 47.68471615720524,
+ "grad_norm": 0.03307219594717026,
+ "learning_rate": 0.0006,
+ "loss": 5.894664287567139,
+ "step": 3433
+ },
+ {
+ "epoch": 47.698689956331876,
+ "grad_norm": 0.028824346140027046,
+ "learning_rate": 0.0006,
+ "loss": 5.914865493774414,
+ "step": 3434
+ },
+ {
+ "epoch": 47.712663755458514,
+ "grad_norm": 0.027923690155148506,
+ "learning_rate": 0.0006,
+ "loss": 5.851759910583496,
+ "step": 3435
+ },
+ {
+ "epoch": 47.72663755458515,
+ "grad_norm": 0.025492684915661812,
+ "learning_rate": 0.0006,
+ "loss": 5.749186038970947,
+ "step": 3436
+ },
+ {
+ "epoch": 47.74061135371179,
+ "grad_norm": 0.025105174630880356,
+ "learning_rate": 0.0006,
+ "loss": 5.803566932678223,
+ "step": 3437
+ },
+ {
+ "epoch": 47.75458515283843,
+ "grad_norm": 0.023298880085349083,
+ "learning_rate": 0.0006,
+ "loss": 5.76276159286499,
+ "step": 3438
+ },
+ {
+ "epoch": 47.76855895196506,
+ "grad_norm": 0.02468576282262802,
+ "learning_rate": 0.0006,
+ "loss": 5.833852767944336,
+ "step": 3439
+ },
+ {
+ "epoch": 47.7825327510917,
+ "grad_norm": 0.022730689495801926,
+ "learning_rate": 0.0006,
+ "loss": 5.736724376678467,
+ "step": 3440
+ },
+ {
+ "epoch": 47.79650655021834,
+ "grad_norm": 0.021335354074835777,
+ "learning_rate": 0.0006,
+ "loss": 5.644248008728027,
+ "step": 3441
+ },
+ {
+ "epoch": 47.81048034934498,
+ "grad_norm": 0.02288101427257061,
+ "learning_rate": 0.0006,
+ "loss": 5.7011942863464355,
+ "step": 3442
+ },
+ {
+ "epoch": 47.82445414847162,
+ "grad_norm": 0.01961585320532322,
+ "learning_rate": 0.0006,
+ "loss": 5.63215446472168,
+ "step": 3443
+ },
+ {
+ "epoch": 47.838427947598255,
+ "grad_norm": 0.017659205943346024,
+ "learning_rate": 0.0006,
+ "loss": 5.744845390319824,
+ "step": 3444
+ },
+ {
+ "epoch": 47.852401746724894,
+ "grad_norm": 0.02281784638762474,
+ "learning_rate": 0.0006,
+ "loss": 5.696993350982666,
+ "step": 3445
+ },
+ {
+ "epoch": 47.866375545851525,
+ "grad_norm": 0.016394242644309998,
+ "learning_rate": 0.0006,
+ "loss": 5.659870147705078,
+ "step": 3446
+ },
+ {
+ "epoch": 47.880349344978164,
+ "grad_norm": 0.01967368647456169,
+ "learning_rate": 0.0006,
+ "loss": 5.638072490692139,
+ "step": 3447
+ },
+ {
+ "epoch": 47.8943231441048,
+ "grad_norm": 0.017326267436146736,
+ "learning_rate": 0.0006,
+ "loss": 5.638836860656738,
+ "step": 3448
+ },
+ {
+ "epoch": 47.90829694323144,
+ "grad_norm": 0.017778463661670685,
+ "learning_rate": 0.0006,
+ "loss": 5.634466648101807,
+ "step": 3449
+ },
+ {
+ "epoch": 47.92227074235808,
+ "grad_norm": 0.017269758507609367,
+ "learning_rate": 0.0006,
+ "loss": 5.506277561187744,
+ "step": 3450
+ },
+ {
+ "epoch": 47.93624454148472,
+ "grad_norm": 0.01733289286494255,
+ "learning_rate": 0.0006,
+ "loss": 5.544455528259277,
+ "step": 3451
+ },
+ {
+ "epoch": 47.95021834061135,
+ "grad_norm": 0.017787795513868332,
+ "learning_rate": 0.0006,
+ "loss": 5.498943328857422,
+ "step": 3452
+ },
+ {
+ "epoch": 47.96419213973799,
+ "grad_norm": 0.015770932659506798,
+ "learning_rate": 0.0006,
+ "loss": 5.537227630615234,
+ "step": 3453
+ },
+ {
+ "epoch": 47.97816593886463,
+ "grad_norm": 0.016132809221744537,
+ "learning_rate": 0.0006,
+ "loss": 5.489081382751465,
+ "step": 3454
+ },
+ {
+ "epoch": 47.992139737991266,
+ "grad_norm": 0.01749664731323719,
+ "learning_rate": 0.0006,
+ "loss": 5.454263687133789,
+ "step": 3455
+ },
+ {
+ "epoch": 48.0,
+ "grad_norm": 0.016711296513676643,
+ "learning_rate": 0.0006,
+ "loss": 5.530783653259277,
+ "step": 3456
+ },
+ {
+ "epoch": 48.0,
+ "eval_loss": 5.596925258636475,
+ "eval_runtime": 56.9612,
+ "eval_samples_per_second": 42.871,
+ "eval_steps_per_second": 1.352,
+ "step": 3456
+ },
+ {
+ "epoch": 48.01397379912664,
+ "grad_norm": 0.01867641694843769,
+ "learning_rate": 0.0006,
+ "loss": 5.506853103637695,
+ "step": 3457
+ },
+ {
+ "epoch": 48.02794759825328,
+ "grad_norm": 0.023250792175531387,
+ "learning_rate": 0.0006,
+ "loss": 5.4074177742004395,
+ "step": 3458
+ },
+ {
+ "epoch": 48.041921397379916,
+ "grad_norm": 0.01970936730504036,
+ "learning_rate": 0.0006,
+ "loss": 5.508777618408203,
+ "step": 3459
+ },
+ {
+ "epoch": 48.05589519650655,
+ "grad_norm": 0.01879260502755642,
+ "learning_rate": 0.0006,
+ "loss": 5.412424087524414,
+ "step": 3460
+ },
+ {
+ "epoch": 48.069868995633186,
+ "grad_norm": 0.017420092597603798,
+ "learning_rate": 0.0006,
+ "loss": 5.43361759185791,
+ "step": 3461
+ },
+ {
+ "epoch": 48.083842794759825,
+ "grad_norm": 0.018707862123847008,
+ "learning_rate": 0.0006,
+ "loss": 5.476800918579102,
+ "step": 3462
+ },
+ {
+ "epoch": 48.09781659388646,
+ "grad_norm": 0.023185882717370987,
+ "learning_rate": 0.0006,
+ "loss": 5.465272426605225,
+ "step": 3463
+ },
+ {
+ "epoch": 48.1117903930131,
+ "grad_norm": 0.023570368066430092,
+ "learning_rate": 0.0006,
+ "loss": 5.391471862792969,
+ "step": 3464
+ },
+ {
+ "epoch": 48.12576419213974,
+ "grad_norm": 0.02651144750416279,
+ "learning_rate": 0.0006,
+ "loss": 5.422799110412598,
+ "step": 3465
+ },
+ {
+ "epoch": 48.13973799126637,
+ "grad_norm": 0.03316329047083855,
+ "learning_rate": 0.0006,
+ "loss": 5.3952202796936035,
+ "step": 3466
+ },
+ {
+ "epoch": 48.15371179039301,
+ "grad_norm": 0.04012278467416763,
+ "learning_rate": 0.0006,
+ "loss": 5.540853023529053,
+ "step": 3467
+ },
+ {
+ "epoch": 48.16768558951965,
+ "grad_norm": 0.04763203114271164,
+ "learning_rate": 0.0006,
+ "loss": 5.387890338897705,
+ "step": 3468
+ },
+ {
+ "epoch": 48.18165938864629,
+ "grad_norm": 0.05129165202379227,
+ "learning_rate": 0.0006,
+ "loss": 5.414722442626953,
+ "step": 3469
+ },
+ {
+ "epoch": 48.19563318777293,
+ "grad_norm": 0.04438071325421333,
+ "learning_rate": 0.0006,
+ "loss": 5.386244773864746,
+ "step": 3470
+ },
+ {
+ "epoch": 48.209606986899566,
+ "grad_norm": 0.040234919637441635,
+ "learning_rate": 0.0006,
+ "loss": 5.431617736816406,
+ "step": 3471
+ },
+ {
+ "epoch": 48.223580786026204,
+ "grad_norm": 0.03217008337378502,
+ "learning_rate": 0.0006,
+ "loss": 5.30168342590332,
+ "step": 3472
+ },
+ {
+ "epoch": 48.237554585152836,
+ "grad_norm": 0.03087974339723587,
+ "learning_rate": 0.0006,
+ "loss": 5.316389083862305,
+ "step": 3473
+ },
+ {
+ "epoch": 48.251528384279474,
+ "grad_norm": 0.02789347991347313,
+ "learning_rate": 0.0006,
+ "loss": 5.385970592498779,
+ "step": 3474
+ },
+ {
+ "epoch": 48.26550218340611,
+ "grad_norm": 0.024792097508907318,
+ "learning_rate": 0.0006,
+ "loss": 5.362329959869385,
+ "step": 3475
+ },
+ {
+ "epoch": 48.27947598253275,
+ "grad_norm": 0.025869522243738174,
+ "learning_rate": 0.0006,
+ "loss": 5.28400182723999,
+ "step": 3476
+ },
+ {
+ "epoch": 48.29344978165939,
+ "grad_norm": 0.021339882165193558,
+ "learning_rate": 0.0006,
+ "loss": 5.36669921875,
+ "step": 3477
+ },
+ {
+ "epoch": 48.30742358078603,
+ "grad_norm": 0.021462520584464073,
+ "learning_rate": 0.0006,
+ "loss": 5.311222076416016,
+ "step": 3478
+ },
+ {
+ "epoch": 48.32139737991266,
+ "grad_norm": 0.02193879708647728,
+ "learning_rate": 0.0006,
+ "loss": 5.291441917419434,
+ "step": 3479
+ },
+ {
+ "epoch": 48.3353711790393,
+ "grad_norm": 0.01963093690574169,
+ "learning_rate": 0.0006,
+ "loss": 5.342780113220215,
+ "step": 3480
+ },
+ {
+ "epoch": 48.34934497816594,
+ "grad_norm": 0.020634736865758896,
+ "learning_rate": 0.0006,
+ "loss": 5.144141674041748,
+ "step": 3481
+ },
+ {
+ "epoch": 48.36331877729258,
+ "grad_norm": 0.02126968279480934,
+ "learning_rate": 0.0006,
+ "loss": 5.232624530792236,
+ "step": 3482
+ },
+ {
+ "epoch": 48.377292576419215,
+ "grad_norm": 0.021203994750976562,
+ "learning_rate": 0.0006,
+ "loss": 5.290219783782959,
+ "step": 3483
+ },
+ {
+ "epoch": 48.391266375545854,
+ "grad_norm": 0.02193225920200348,
+ "learning_rate": 0.0006,
+ "loss": 5.29197359085083,
+ "step": 3484
+ },
+ {
+ "epoch": 48.40524017467249,
+ "grad_norm": 0.023174211382865906,
+ "learning_rate": 0.0006,
+ "loss": 5.204349517822266,
+ "step": 3485
+ },
+ {
+ "epoch": 48.419213973799124,
+ "grad_norm": 0.024161914363503456,
+ "learning_rate": 0.0006,
+ "loss": 5.339351654052734,
+ "step": 3486
+ },
+ {
+ "epoch": 48.43318777292576,
+ "grad_norm": 0.02221524342894554,
+ "learning_rate": 0.0006,
+ "loss": 5.263001441955566,
+ "step": 3487
+ },
+ {
+ "epoch": 48.4471615720524,
+ "grad_norm": 0.026632830500602722,
+ "learning_rate": 0.0006,
+ "loss": 5.221281051635742,
+ "step": 3488
+ },
+ {
+ "epoch": 48.46113537117904,
+ "grad_norm": 0.034923870116472244,
+ "learning_rate": 0.0006,
+ "loss": 5.296039581298828,
+ "step": 3489
+ },
+ {
+ "epoch": 48.47510917030568,
+ "grad_norm": 0.046184979379177094,
+ "learning_rate": 0.0006,
+ "loss": 5.212769985198975,
+ "step": 3490
+ },
+ {
+ "epoch": 48.48908296943232,
+ "grad_norm": 0.05154098570346832,
+ "learning_rate": 0.0006,
+ "loss": 5.337864875793457,
+ "step": 3491
+ },
+ {
+ "epoch": 48.50305676855895,
+ "grad_norm": 0.042967259883880615,
+ "learning_rate": 0.0006,
+ "loss": 5.267958641052246,
+ "step": 3492
+ },
+ {
+ "epoch": 48.51703056768559,
+ "grad_norm": 0.05356583371758461,
+ "learning_rate": 0.0006,
+ "loss": 5.1404218673706055,
+ "step": 3493
+ },
+ {
+ "epoch": 48.531004366812226,
+ "grad_norm": 0.08219032734632492,
+ "learning_rate": 0.0006,
+ "loss": 5.300687789916992,
+ "step": 3494
+ },
+ {
+ "epoch": 48.544978165938865,
+ "grad_norm": 0.19204089045524597,
+ "learning_rate": 0.0006,
+ "loss": 5.456025123596191,
+ "step": 3495
+ },
+ {
+ "epoch": 48.5589519650655,
+ "grad_norm": 0.25917696952819824,
+ "learning_rate": 0.0006,
+ "loss": 6.156761169433594,
+ "step": 3496
+ },
+ {
+ "epoch": 48.57292576419214,
+ "grad_norm": 0.1566283404827118,
+ "learning_rate": 0.0006,
+ "loss": 5.862346649169922,
+ "step": 3497
+ },
+ {
+ "epoch": 48.58689956331878,
+ "grad_norm": 0.1761486828327179,
+ "learning_rate": 0.0006,
+ "loss": 6.012795448303223,
+ "step": 3498
+ },
+ {
+ "epoch": 48.60087336244541,
+ "grad_norm": 0.09857621043920517,
+ "learning_rate": 0.0006,
+ "loss": 5.9248151779174805,
+ "step": 3499
+ },
+ {
+ "epoch": 48.61484716157205,
+ "grad_norm": 0.05968055501580238,
+ "learning_rate": 0.0006,
+ "loss": 5.780706882476807,
+ "step": 3500
+ },
+ {
+ "epoch": 48.62882096069869,
+ "grad_norm": 0.057433564215898514,
+ "learning_rate": 0.0006,
+ "loss": 5.645388603210449,
+ "step": 3501
+ },
+ {
+ "epoch": 48.64279475982533,
+ "grad_norm": 0.039659615606069565,
+ "learning_rate": 0.0006,
+ "loss": 5.695138931274414,
+ "step": 3502
+ },
+ {
+ "epoch": 48.65676855895197,
+ "grad_norm": 0.04019695147871971,
+ "learning_rate": 0.0006,
+ "loss": 5.672435283660889,
+ "step": 3503
+ },
+ {
+ "epoch": 48.670742358078606,
+ "grad_norm": 0.03033018670976162,
+ "learning_rate": 0.0006,
+ "loss": 5.682260513305664,
+ "step": 3504
+ },
+ {
+ "epoch": 48.68471615720524,
+ "grad_norm": 0.03215787187218666,
+ "learning_rate": 0.0006,
+ "loss": 5.604182243347168,
+ "step": 3505
+ },
+ {
+ "epoch": 48.698689956331876,
+ "grad_norm": 0.026865126565098763,
+ "learning_rate": 0.0006,
+ "loss": 5.64040470123291,
+ "step": 3506
+ },
+ {
+ "epoch": 48.712663755458514,
+ "grad_norm": 0.02831609919667244,
+ "learning_rate": 0.0006,
+ "loss": 5.552720069885254,
+ "step": 3507
+ },
+ {
+ "epoch": 48.72663755458515,
+ "grad_norm": 0.024226613342761993,
+ "learning_rate": 0.0006,
+ "loss": 5.556946277618408,
+ "step": 3508
+ },
+ {
+ "epoch": 48.74061135371179,
+ "grad_norm": 0.025393934920430183,
+ "learning_rate": 0.0006,
+ "loss": 5.476178169250488,
+ "step": 3509
+ },
+ {
+ "epoch": 48.75458515283843,
+ "grad_norm": 0.01984412968158722,
+ "learning_rate": 0.0006,
+ "loss": 5.471088409423828,
+ "step": 3510
+ },
+ {
+ "epoch": 48.76855895196506,
+ "grad_norm": 0.02035129815340042,
+ "learning_rate": 0.0006,
+ "loss": 5.438812732696533,
+ "step": 3511
+ },
+ {
+ "epoch": 48.7825327510917,
+ "grad_norm": 0.0208350270986557,
+ "learning_rate": 0.0006,
+ "loss": 5.381746292114258,
+ "step": 3512
+ },
+ {
+ "epoch": 48.79650655021834,
+ "grad_norm": 0.017783882096409798,
+ "learning_rate": 0.0006,
+ "loss": 5.351675510406494,
+ "step": 3513
+ },
+ {
+ "epoch": 48.81048034934498,
+ "grad_norm": 0.01671770215034485,
+ "learning_rate": 0.0006,
+ "loss": 5.329428672790527,
+ "step": 3514
+ },
+ {
+ "epoch": 48.82445414847162,
+ "grad_norm": 0.017891215160489082,
+ "learning_rate": 0.0006,
+ "loss": 5.212074279785156,
+ "step": 3515
+ },
+ {
+ "epoch": 48.838427947598255,
+ "grad_norm": 0.016847791150212288,
+ "learning_rate": 0.0006,
+ "loss": 5.299534797668457,
+ "step": 3516
+ },
+ {
+ "epoch": 48.852401746724894,
+ "grad_norm": 0.01593637838959694,
+ "learning_rate": 0.0006,
+ "loss": 5.30466890335083,
+ "step": 3517
+ },
+ {
+ "epoch": 48.866375545851525,
+ "grad_norm": 0.014713102020323277,
+ "learning_rate": 0.0006,
+ "loss": 5.294345378875732,
+ "step": 3518
+ },
+ {
+ "epoch": 48.880349344978164,
+ "grad_norm": 0.01433857623487711,
+ "learning_rate": 0.0006,
+ "loss": 5.2148542404174805,
+ "step": 3519
+ },
+ {
+ "epoch": 48.8943231441048,
+ "grad_norm": 0.014979379251599312,
+ "learning_rate": 0.0006,
+ "loss": 5.298582077026367,
+ "step": 3520
+ },
+ {
+ "epoch": 48.90829694323144,
+ "grad_norm": 0.013494192622601986,
+ "learning_rate": 0.0006,
+ "loss": 5.309783935546875,
+ "step": 3521
+ },
+ {
+ "epoch": 48.92227074235808,
+ "grad_norm": 0.014020012691617012,
+ "learning_rate": 0.0006,
+ "loss": 5.284879684448242,
+ "step": 3522
+ },
+ {
+ "epoch": 48.93624454148472,
+ "grad_norm": 0.01348058134317398,
+ "learning_rate": 0.0006,
+ "loss": 5.249116897583008,
+ "step": 3523
+ },
+ {
+ "epoch": 48.95021834061135,
+ "grad_norm": 0.013280218467116356,
+ "learning_rate": 0.0006,
+ "loss": 5.273489952087402,
+ "step": 3524
+ },
+ {
+ "epoch": 48.96419213973799,
+ "grad_norm": 0.012647110037505627,
+ "learning_rate": 0.0006,
+ "loss": 5.306652069091797,
+ "step": 3525
+ },
+ {
+ "epoch": 48.97816593886463,
+ "grad_norm": 0.012847724370658398,
+ "learning_rate": 0.0006,
+ "loss": 5.235745429992676,
+ "step": 3526
+ },
+ {
+ "epoch": 48.992139737991266,
+ "grad_norm": 0.011820166371762753,
+ "learning_rate": 0.0006,
+ "loss": 5.238091468811035,
+ "step": 3527
+ },
+ {
+ "epoch": 49.0,
+ "grad_norm": 0.01255585066974163,
+ "learning_rate": 0.0006,
+ "loss": 5.231295585632324,
+ "step": 3528
+ },
+ {
+ "epoch": 49.0,
+ "eval_loss": 5.394587993621826,
+ "eval_runtime": 56.3734,
+ "eval_samples_per_second": 43.318,
+ "eval_steps_per_second": 1.366,
+ "step": 3528
+ },
+ {
+ "epoch": 49.01397379912664,
+ "grad_norm": 0.011031972244381905,
+ "learning_rate": 0.0006,
+ "loss": 5.1354570388793945,
+ "step": 3529
+ },
+ {
+ "epoch": 49.02794759825328,
+ "grad_norm": 0.010631263256072998,
+ "learning_rate": 0.0006,
+ "loss": 5.177639961242676,
+ "step": 3530
+ },
+ {
+ "epoch": 49.041921397379916,
+ "grad_norm": 0.01133775431662798,
+ "learning_rate": 0.0006,
+ "loss": 5.283428192138672,
+ "step": 3531
+ },
+ {
+ "epoch": 49.05589519650655,
+ "grad_norm": 0.01154172234237194,
+ "learning_rate": 0.0006,
+ "loss": 5.206364154815674,
+ "step": 3532
+ },
+ {
+ "epoch": 49.069868995633186,
+ "grad_norm": 0.010192356072366238,
+ "learning_rate": 0.0006,
+ "loss": 5.2385735511779785,
+ "step": 3533
+ },
+ {
+ "epoch": 49.083842794759825,
+ "grad_norm": 0.012789200060069561,
+ "learning_rate": 0.0006,
+ "loss": 5.071800231933594,
+ "step": 3534
+ },
+ {
+ "epoch": 49.09781659388646,
+ "grad_norm": 0.011282820254564285,
+ "learning_rate": 0.0006,
+ "loss": 5.218020439147949,
+ "step": 3535
+ },
+ {
+ "epoch": 49.1117903930131,
+ "grad_norm": 0.012153448536992073,
+ "learning_rate": 0.0006,
+ "loss": 5.134698867797852,
+ "step": 3536
+ },
+ {
+ "epoch": 49.12576419213974,
+ "grad_norm": 0.011920543387532234,
+ "learning_rate": 0.0006,
+ "loss": 5.141620635986328,
+ "step": 3537
+ },
+ {
+ "epoch": 49.13973799126637,
+ "grad_norm": 0.010949329473078251,
+ "learning_rate": 0.0006,
+ "loss": 5.116259574890137,
+ "step": 3538
+ },
+ {
+ "epoch": 49.15371179039301,
+ "grad_norm": 0.012362448498606682,
+ "learning_rate": 0.0006,
+ "loss": 5.090177059173584,
+ "step": 3539
+ },
+ {
+ "epoch": 49.16768558951965,
+ "grad_norm": 0.011575071141123772,
+ "learning_rate": 0.0006,
+ "loss": 5.093441009521484,
+ "step": 3540
+ },
+ {
+ "epoch": 49.18165938864629,
+ "grad_norm": 0.011126236990094185,
+ "learning_rate": 0.0006,
+ "loss": 5.060235023498535,
+ "step": 3541
+ },
+ {
+ "epoch": 49.19563318777293,
+ "grad_norm": 0.0107401292771101,
+ "learning_rate": 0.0006,
+ "loss": 5.140704154968262,
+ "step": 3542
+ },
+ {
+ "epoch": 49.209606986899566,
+ "grad_norm": 0.011202906258404255,
+ "learning_rate": 0.0006,
+ "loss": 5.10855770111084,
+ "step": 3543
+ },
+ {
+ "epoch": 49.223580786026204,
+ "grad_norm": 0.011083454824984074,
+ "learning_rate": 0.0006,
+ "loss": 5.092751502990723,
+ "step": 3544
+ },
+ {
+ "epoch": 49.237554585152836,
+ "grad_norm": 0.009523800574243069,
+ "learning_rate": 0.0006,
+ "loss": 5.13733434677124,
+ "step": 3545
+ },
+ {
+ "epoch": 49.251528384279474,
+ "grad_norm": 0.011298353783786297,
+ "learning_rate": 0.0006,
+ "loss": 5.036048889160156,
+ "step": 3546
+ },
+ {
+ "epoch": 49.26550218340611,
+ "grad_norm": 0.010676483623683453,
+ "learning_rate": 0.0006,
+ "loss": 5.051901340484619,
+ "step": 3547
+ },
+ {
+ "epoch": 49.27947598253275,
+ "grad_norm": 0.011463399045169353,
+ "learning_rate": 0.0006,
+ "loss": 4.983170986175537,
+ "step": 3548
+ },
+ {
+ "epoch": 49.29344978165939,
+ "grad_norm": 0.010603402741253376,
+ "learning_rate": 0.0006,
+ "loss": 5.112311363220215,
+ "step": 3549
+ },
+ {
+ "epoch": 49.30742358078603,
+ "grad_norm": 0.010535670444369316,
+ "learning_rate": 0.0006,
+ "loss": 5.085227966308594,
+ "step": 3550
+ },
+ {
+ "epoch": 49.32139737991266,
+ "grad_norm": 0.011843928135931492,
+ "learning_rate": 0.0006,
+ "loss": 5.07575798034668,
+ "step": 3551
+ },
+ {
+ "epoch": 49.3353711790393,
+ "grad_norm": 0.010587952099740505,
+ "learning_rate": 0.0006,
+ "loss": 5.019237041473389,
+ "step": 3552
+ },
+ {
+ "epoch": 49.34934497816594,
+ "grad_norm": 0.012285659089684486,
+ "learning_rate": 0.0006,
+ "loss": 5.144603729248047,
+ "step": 3553
+ },
+ {
+ "epoch": 49.36331877729258,
+ "grad_norm": 0.013119431212544441,
+ "learning_rate": 0.0006,
+ "loss": 5.035256862640381,
+ "step": 3554
+ },
+ {
+ "epoch": 49.377292576419215,
+ "grad_norm": 0.012138537131249905,
+ "learning_rate": 0.0006,
+ "loss": 4.996667861938477,
+ "step": 3555
+ },
+ {
+ "epoch": 49.391266375545854,
+ "grad_norm": 0.012142804451286793,
+ "learning_rate": 0.0006,
+ "loss": 4.95058536529541,
+ "step": 3556
+ },
+ {
+ "epoch": 49.40524017467249,
+ "grad_norm": 0.012205676175653934,
+ "learning_rate": 0.0006,
+ "loss": 5.096931457519531,
+ "step": 3557
+ },
+ {
+ "epoch": 49.419213973799124,
+ "grad_norm": 0.012086744420230389,
+ "learning_rate": 0.0006,
+ "loss": 4.99384880065918,
+ "step": 3558
+ },
+ {
+ "epoch": 49.43318777292576,
+ "grad_norm": 0.012918103486299515,
+ "learning_rate": 0.0006,
+ "loss": 5.031294822692871,
+ "step": 3559
+ },
+ {
+ "epoch": 49.4471615720524,
+ "grad_norm": 0.012006713077425957,
+ "learning_rate": 0.0006,
+ "loss": 5.0547943115234375,
+ "step": 3560
+ },
+ {
+ "epoch": 49.46113537117904,
+ "grad_norm": 0.0120399734005332,
+ "learning_rate": 0.0006,
+ "loss": 5.014238357543945,
+ "step": 3561
+ },
+ {
+ "epoch": 49.47510917030568,
+ "grad_norm": 0.013306990265846252,
+ "learning_rate": 0.0006,
+ "loss": 4.972925186157227,
+ "step": 3562
+ },
+ {
+ "epoch": 49.48908296943232,
+ "grad_norm": 0.013878636993467808,
+ "learning_rate": 0.0006,
+ "loss": 5.079033851623535,
+ "step": 3563
+ },
+ {
+ "epoch": 49.50305676855895,
+ "grad_norm": 0.012454566545784473,
+ "learning_rate": 0.0006,
+ "loss": 4.93708610534668,
+ "step": 3564
+ },
+ {
+ "epoch": 49.51703056768559,
+ "grad_norm": 0.014075061306357384,
+ "learning_rate": 0.0006,
+ "loss": 4.965582370758057,
+ "step": 3565
+ },
+ {
+ "epoch": 49.531004366812226,
+ "grad_norm": 0.012525617145001888,
+ "learning_rate": 0.0006,
+ "loss": 4.968148708343506,
+ "step": 3566
+ },
+ {
+ "epoch": 49.544978165938865,
+ "grad_norm": 0.012418225407600403,
+ "learning_rate": 0.0006,
+ "loss": 5.06901741027832,
+ "step": 3567
+ },
+ {
+ "epoch": 49.5589519650655,
+ "grad_norm": 0.015034239739179611,
+ "learning_rate": 0.0006,
+ "loss": 4.984277725219727,
+ "step": 3568
+ },
+ {
+ "epoch": 49.57292576419214,
+ "grad_norm": 0.014329628087580204,
+ "learning_rate": 0.0006,
+ "loss": 4.951558589935303,
+ "step": 3569
+ },
+ {
+ "epoch": 49.58689956331878,
+ "grad_norm": 0.017769070342183113,
+ "learning_rate": 0.0006,
+ "loss": 4.966532230377197,
+ "step": 3570
+ },
+ {
+ "epoch": 49.60087336244541,
+ "grad_norm": 0.01909453794360161,
+ "learning_rate": 0.0006,
+ "loss": 4.919890880584717,
+ "step": 3571
+ },
+ {
+ "epoch": 49.61484716157205,
+ "grad_norm": 0.017519688233733177,
+ "learning_rate": 0.0006,
+ "loss": 5.031286716461182,
+ "step": 3572
+ },
+ {
+ "epoch": 49.62882096069869,
+ "grad_norm": 0.017143897712230682,
+ "learning_rate": 0.0006,
+ "loss": 4.969764709472656,
+ "step": 3573
+ },
+ {
+ "epoch": 49.64279475982533,
+ "grad_norm": 0.01688285917043686,
+ "learning_rate": 0.0006,
+ "loss": 5.0733442306518555,
+ "step": 3574
+ },
+ {
+ "epoch": 49.65676855895197,
+ "grad_norm": 0.018547311425209045,
+ "learning_rate": 0.0006,
+ "loss": 4.980863571166992,
+ "step": 3575
+ },
+ {
+ "epoch": 49.670742358078606,
+ "grad_norm": 0.018877137452363968,
+ "learning_rate": 0.0006,
+ "loss": 4.954646110534668,
+ "step": 3576
+ },
+ {
+ "epoch": 49.68471615720524,
+ "grad_norm": 0.018681600689888,
+ "learning_rate": 0.0006,
+ "loss": 4.959702014923096,
+ "step": 3577
+ },
+ {
+ "epoch": 49.698689956331876,
+ "grad_norm": 0.018947336822748184,
+ "learning_rate": 0.0006,
+ "loss": 5.070502758026123,
+ "step": 3578
+ },
+ {
+ "epoch": 49.712663755458514,
+ "grad_norm": 0.018817484378814697,
+ "learning_rate": 0.0006,
+ "loss": 5.056508541107178,
+ "step": 3579
+ },
+ {
+ "epoch": 49.72663755458515,
+ "grad_norm": 0.01931724324822426,
+ "learning_rate": 0.0006,
+ "loss": 4.922451972961426,
+ "step": 3580
+ },
+ {
+ "epoch": 49.74061135371179,
+ "grad_norm": 0.022047819569706917,
+ "learning_rate": 0.0006,
+ "loss": 5.053170204162598,
+ "step": 3581
+ },
+ {
+ "epoch": 49.75458515283843,
+ "grad_norm": 0.02279801294207573,
+ "learning_rate": 0.0006,
+ "loss": 4.8768439292907715,
+ "step": 3582
+ },
+ {
+ "epoch": 49.76855895196506,
+ "grad_norm": 0.020661601796746254,
+ "learning_rate": 0.0006,
+ "loss": 4.907928466796875,
+ "step": 3583
+ },
+ {
+ "epoch": 49.7825327510917,
+ "grad_norm": 0.020479509606957436,
+ "learning_rate": 0.0006,
+ "loss": 5.040694236755371,
+ "step": 3584
+ },
+ {
+ "epoch": 49.79650655021834,
+ "grad_norm": 0.021654162555933,
+ "learning_rate": 0.0006,
+ "loss": 5.018434524536133,
+ "step": 3585
+ },
+ {
+ "epoch": 49.81048034934498,
+ "grad_norm": 0.022486824542284012,
+ "learning_rate": 0.0006,
+ "loss": 5.086538314819336,
+ "step": 3586
+ },
+ {
+ "epoch": 49.82445414847162,
+ "grad_norm": 0.02141369879245758,
+ "learning_rate": 0.0006,
+ "loss": 5.035694122314453,
+ "step": 3587
+ },
+ {
+ "epoch": 49.838427947598255,
+ "grad_norm": 0.015926677733659744,
+ "learning_rate": 0.0006,
+ "loss": 4.905038833618164,
+ "step": 3588
+ },
+ {
+ "epoch": 49.852401746724894,
+ "grad_norm": 0.015775298699736595,
+ "learning_rate": 0.0006,
+ "loss": 5.003902435302734,
+ "step": 3589
+ },
+ {
+ "epoch": 49.866375545851525,
+ "grad_norm": 0.01872118003666401,
+ "learning_rate": 0.0006,
+ "loss": 4.891060829162598,
+ "step": 3590
+ },
+ {
+ "epoch": 49.880349344978164,
+ "grad_norm": 0.02066192403435707,
+ "learning_rate": 0.0006,
+ "loss": 5.002897262573242,
+ "step": 3591
+ },
+ {
+ "epoch": 49.8943231441048,
+ "grad_norm": 0.02348714880645275,
+ "learning_rate": 0.0006,
+ "loss": 4.9189958572387695,
+ "step": 3592
+ },
+ {
+ "epoch": 49.90829694323144,
+ "grad_norm": 0.020402103662490845,
+ "learning_rate": 0.0006,
+ "loss": 4.992661476135254,
+ "step": 3593
+ },
+ {
+ "epoch": 49.92227074235808,
+ "grad_norm": 0.020466072484850883,
+ "learning_rate": 0.0006,
+ "loss": 4.962968826293945,
+ "step": 3594
+ },
+ {
+ "epoch": 49.93624454148472,
+ "grad_norm": 0.01922314241528511,
+ "learning_rate": 0.0006,
+ "loss": 4.897730827331543,
+ "step": 3595
+ },
+ {
+ "epoch": 49.95021834061135,
+ "grad_norm": 0.017228754237294197,
+ "learning_rate": 0.0006,
+ "loss": 4.964842796325684,
+ "step": 3596
+ },
+ {
+ "epoch": 49.96419213973799,
+ "grad_norm": 0.016021743416786194,
+ "learning_rate": 0.0006,
+ "loss": 4.910758972167969,
+ "step": 3597
+ },
+ {
+ "epoch": 49.97816593886463,
+ "grad_norm": 0.017985103651881218,
+ "learning_rate": 0.0006,
+ "loss": 4.916072368621826,
+ "step": 3598
+ },
+ {
+ "epoch": 49.992139737991266,
+ "grad_norm": 0.018780240789055824,
+ "learning_rate": 0.0006,
+ "loss": 4.94658088684082,
+ "step": 3599
+ },
+ {
+ "epoch": 50.0,
+ "grad_norm": 0.01776314526796341,
+ "learning_rate": 0.0006,
+ "loss": 4.969357967376709,
+ "step": 3600
+ },
+ {
+ "epoch": 50.0,
+ "eval_loss": 5.141158580780029,
+ "eval_runtime": 56.2332,
+ "eval_samples_per_second": 43.426,
+ "eval_steps_per_second": 1.369,
+ "step": 3600
+ },
+ {
+ "epoch": 50.01397379912664,
+ "grad_norm": 0.016308696940541267,
+ "learning_rate": 0.0006,
+ "loss": 4.924664497375488,
+ "step": 3601
+ },
+ {
+ "epoch": 50.02794759825328,
+ "grad_norm": 0.018810199573636055,
+ "learning_rate": 0.0006,
+ "loss": 4.82293701171875,
+ "step": 3602
+ },
+ {
+ "epoch": 50.041921397379916,
+ "grad_norm": 0.02227160707116127,
+ "learning_rate": 0.0006,
+ "loss": 4.9612298011779785,
+ "step": 3603
+ },
+ {
+ "epoch": 50.05589519650655,
+ "grad_norm": 0.025266462936997414,
+ "learning_rate": 0.0006,
+ "loss": 4.963835716247559,
+ "step": 3604
+ },
+ {
+ "epoch": 50.069868995633186,
+ "grad_norm": 0.028588181361556053,
+ "learning_rate": 0.0006,
+ "loss": 4.992118835449219,
+ "step": 3605
+ },
+ {
+ "epoch": 50.083842794759825,
+ "grad_norm": 0.02739487960934639,
+ "learning_rate": 0.0006,
+ "loss": 4.915547847747803,
+ "step": 3606
+ },
+ {
+ "epoch": 50.09781659388646,
+ "grad_norm": 0.027000300586223602,
+ "learning_rate": 0.0006,
+ "loss": 4.883801460266113,
+ "step": 3607
+ },
+ {
+ "epoch": 50.1117903930131,
+ "grad_norm": 0.023468483239412308,
+ "learning_rate": 0.0006,
+ "loss": 5.006168365478516,
+ "step": 3608
+ },
+ {
+ "epoch": 50.12576419213974,
+ "grad_norm": 0.026753634214401245,
+ "learning_rate": 0.0006,
+ "loss": 4.92526912689209,
+ "step": 3609
+ },
+ {
+ "epoch": 50.13973799126637,
+ "grad_norm": 0.026816733181476593,
+ "learning_rate": 0.0006,
+ "loss": 4.847546100616455,
+ "step": 3610
+ },
+ {
+ "epoch": 50.15371179039301,
+ "grad_norm": 0.027362391352653503,
+ "learning_rate": 0.0006,
+ "loss": 4.862548351287842,
+ "step": 3611
+ },
+ {
+ "epoch": 50.16768558951965,
+ "grad_norm": 0.03450845181941986,
+ "learning_rate": 0.0006,
+ "loss": 4.963202953338623,
+ "step": 3612
+ },
+ {
+ "epoch": 50.18165938864629,
+ "grad_norm": 0.03801408037543297,
+ "learning_rate": 0.0006,
+ "loss": 4.770240306854248,
+ "step": 3613
+ },
+ {
+ "epoch": 50.19563318777293,
+ "grad_norm": 0.03579140082001686,
+ "learning_rate": 0.0006,
+ "loss": 4.797981262207031,
+ "step": 3614
+ },
+ {
+ "epoch": 50.209606986899566,
+ "grad_norm": 0.029393760487437248,
+ "learning_rate": 0.0006,
+ "loss": 4.798581600189209,
+ "step": 3615
+ },
+ {
+ "epoch": 50.223580786026204,
+ "grad_norm": 0.026111142709851265,
+ "learning_rate": 0.0006,
+ "loss": 4.990712642669678,
+ "step": 3616
+ },
+ {
+ "epoch": 50.237554585152836,
+ "grad_norm": 0.02213387005031109,
+ "learning_rate": 0.0006,
+ "loss": 4.865842342376709,
+ "step": 3617
+ },
+ {
+ "epoch": 50.251528384279474,
+ "grad_norm": 0.025450684130191803,
+ "learning_rate": 0.0006,
+ "loss": 4.903746128082275,
+ "step": 3618
+ },
+ {
+ "epoch": 50.26550218340611,
+ "grad_norm": 0.020915117114782333,
+ "learning_rate": 0.0006,
+ "loss": 4.951366424560547,
+ "step": 3619
+ },
+ {
+ "epoch": 50.27947598253275,
+ "grad_norm": 0.020972849801182747,
+ "learning_rate": 0.0006,
+ "loss": 4.783311367034912,
+ "step": 3620
+ },
+ {
+ "epoch": 50.29344978165939,
+ "grad_norm": 0.016956273466348648,
+ "learning_rate": 0.0006,
+ "loss": 4.955767631530762,
+ "step": 3621
+ },
+ {
+ "epoch": 50.30742358078603,
+ "grad_norm": 0.018045252189040184,
+ "learning_rate": 0.0006,
+ "loss": 4.869322776794434,
+ "step": 3622
+ },
+ {
+ "epoch": 50.32139737991266,
+ "grad_norm": 0.018199188634753227,
+ "learning_rate": 0.0006,
+ "loss": 4.95680046081543,
+ "step": 3623
+ },
+ {
+ "epoch": 50.3353711790393,
+ "grad_norm": 0.016688082367181778,
+ "learning_rate": 0.0006,
+ "loss": 4.783535957336426,
+ "step": 3624
+ },
+ {
+ "epoch": 50.34934497816594,
+ "grad_norm": 0.017540352419018745,
+ "learning_rate": 0.0006,
+ "loss": 4.8236083984375,
+ "step": 3625
+ },
+ {
+ "epoch": 50.36331877729258,
+ "grad_norm": 0.018054187297821045,
+ "learning_rate": 0.0006,
+ "loss": 4.764655113220215,
+ "step": 3626
+ },
+ {
+ "epoch": 50.377292576419215,
+ "grad_norm": 0.01741131953895092,
+ "learning_rate": 0.0006,
+ "loss": 4.875613689422607,
+ "step": 3627
+ },
+ {
+ "epoch": 50.391266375545854,
+ "grad_norm": 0.01682269014418125,
+ "learning_rate": 0.0006,
+ "loss": 4.818509101867676,
+ "step": 3628
+ },
+ {
+ "epoch": 50.40524017467249,
+ "grad_norm": 0.016652341932058334,
+ "learning_rate": 0.0006,
+ "loss": 4.974020481109619,
+ "step": 3629
+ },
+ {
+ "epoch": 50.419213973799124,
+ "grad_norm": 0.02035282365977764,
+ "learning_rate": 0.0006,
+ "loss": 4.893037796020508,
+ "step": 3630
+ },
+ {
+ "epoch": 50.43318777292576,
+ "grad_norm": 0.019907061010599136,
+ "learning_rate": 0.0006,
+ "loss": 4.848119735717773,
+ "step": 3631
+ },
+ {
+ "epoch": 50.4471615720524,
+ "grad_norm": 0.023861493915319443,
+ "learning_rate": 0.0006,
+ "loss": 4.819353103637695,
+ "step": 3632
+ },
+ {
+ "epoch": 50.46113537117904,
+ "grad_norm": 0.025254026055336,
+ "learning_rate": 0.0006,
+ "loss": 4.769312858581543,
+ "step": 3633
+ },
+ {
+ "epoch": 50.47510917030568,
+ "grad_norm": 0.026154937222599983,
+ "learning_rate": 0.0006,
+ "loss": 4.7586774826049805,
+ "step": 3634
+ },
+ {
+ "epoch": 50.48908296943232,
+ "grad_norm": 0.026342090219259262,
+ "learning_rate": 0.0006,
+ "loss": 4.861218452453613,
+ "step": 3635
+ },
+ {
+ "epoch": 50.50305676855895,
+ "grad_norm": 0.028169559314846992,
+ "learning_rate": 0.0006,
+ "loss": 4.818380355834961,
+ "step": 3636
+ },
+ {
+ "epoch": 50.51703056768559,
+ "grad_norm": 0.02820264920592308,
+ "learning_rate": 0.0006,
+ "loss": 4.892871856689453,
+ "step": 3637
+ },
+ {
+ "epoch": 50.531004366812226,
+ "grad_norm": 0.03345080465078354,
+ "learning_rate": 0.0006,
+ "loss": 4.811084747314453,
+ "step": 3638
+ },
+ {
+ "epoch": 50.544978165938865,
+ "grad_norm": 0.04005736857652664,
+ "learning_rate": 0.0006,
+ "loss": 4.765410900115967,
+ "step": 3639
+ },
+ {
+ "epoch": 50.5589519650655,
+ "grad_norm": 0.04333870857954025,
+ "learning_rate": 0.0006,
+ "loss": 4.956520080566406,
+ "step": 3640
+ },
+ {
+ "epoch": 50.57292576419214,
+ "grad_norm": 0.04216952994465828,
+ "learning_rate": 0.0006,
+ "loss": 4.793441295623779,
+ "step": 3641
+ },
+ {
+ "epoch": 50.58689956331878,
+ "grad_norm": 0.02936951443552971,
+ "learning_rate": 0.0006,
+ "loss": 4.8447957038879395,
+ "step": 3642
+ },
+ {
+ "epoch": 50.60087336244541,
+ "grad_norm": 0.026937026530504227,
+ "learning_rate": 0.0006,
+ "loss": 4.962215423583984,
+ "step": 3643
+ },
+ {
+ "epoch": 50.61484716157205,
+ "grad_norm": 0.02628890983760357,
+ "learning_rate": 0.0006,
+ "loss": 4.883061408996582,
+ "step": 3644
+ },
+ {
+ "epoch": 50.62882096069869,
+ "grad_norm": 0.025984834879636765,
+ "learning_rate": 0.0006,
+ "loss": 4.768662452697754,
+ "step": 3645
+ },
+ {
+ "epoch": 50.64279475982533,
+ "grad_norm": 0.023998012766242027,
+ "learning_rate": 0.0006,
+ "loss": 4.851325035095215,
+ "step": 3646
+ },
+ {
+ "epoch": 50.65676855895197,
+ "grad_norm": 0.02143184095621109,
+ "learning_rate": 0.0006,
+ "loss": 4.8274688720703125,
+ "step": 3647
+ },
+ {
+ "epoch": 50.670742358078606,
+ "grad_norm": 0.02488900162279606,
+ "learning_rate": 0.0006,
+ "loss": 4.831140518188477,
+ "step": 3648
+ },
+ {
+ "epoch": 50.68471615720524,
+ "grad_norm": 0.02330533228814602,
+ "learning_rate": 0.0006,
+ "loss": 4.850462913513184,
+ "step": 3649
+ },
+ {
+ "epoch": 50.698689956331876,
+ "grad_norm": 0.02121943049132824,
+ "learning_rate": 0.0006,
+ "loss": 4.870253086090088,
+ "step": 3650
+ },
+ {
+ "epoch": 50.712663755458514,
+ "grad_norm": 0.02080935798585415,
+ "learning_rate": 0.0006,
+ "loss": 4.798770904541016,
+ "step": 3651
+ },
+ {
+ "epoch": 50.72663755458515,
+ "grad_norm": 0.019047029316425323,
+ "learning_rate": 0.0006,
+ "loss": 4.8427300453186035,
+ "step": 3652
+ },
+ {
+ "epoch": 50.74061135371179,
+ "grad_norm": 0.01824103854596615,
+ "learning_rate": 0.0006,
+ "loss": 4.820980548858643,
+ "step": 3653
+ },
+ {
+ "epoch": 50.75458515283843,
+ "grad_norm": 0.018129967153072357,
+ "learning_rate": 0.0006,
+ "loss": 4.946732997894287,
+ "step": 3654
+ },
+ {
+ "epoch": 50.76855895196506,
+ "grad_norm": 0.01784815452992916,
+ "learning_rate": 0.0006,
+ "loss": 4.829516410827637,
+ "step": 3655
+ },
+ {
+ "epoch": 50.7825327510917,
+ "grad_norm": 0.020021555945277214,
+ "learning_rate": 0.0006,
+ "loss": 4.780428409576416,
+ "step": 3656
+ },
+ {
+ "epoch": 50.79650655021834,
+ "grad_norm": 0.01863253302872181,
+ "learning_rate": 0.0006,
+ "loss": 4.791567802429199,
+ "step": 3657
+ },
+ {
+ "epoch": 50.81048034934498,
+ "grad_norm": 0.019151683896780014,
+ "learning_rate": 0.0006,
+ "loss": 4.824892997741699,
+ "step": 3658
+ },
+ {
+ "epoch": 50.82445414847162,
+ "grad_norm": 0.01891648955643177,
+ "learning_rate": 0.0006,
+ "loss": 4.789605140686035,
+ "step": 3659
+ },
+ {
+ "epoch": 50.838427947598255,
+ "grad_norm": 0.016019510105252266,
+ "learning_rate": 0.0006,
+ "loss": 4.779555320739746,
+ "step": 3660
+ },
+ {
+ "epoch": 50.852401746724894,
+ "grad_norm": 0.016096491366624832,
+ "learning_rate": 0.0006,
+ "loss": 4.753842353820801,
+ "step": 3661
+ },
+ {
+ "epoch": 50.866375545851525,
+ "grad_norm": 0.01693476364016533,
+ "learning_rate": 0.0006,
+ "loss": 4.768477439880371,
+ "step": 3662
+ },
+ {
+ "epoch": 50.880349344978164,
+ "grad_norm": 0.01749575324356556,
+ "learning_rate": 0.0006,
+ "loss": 4.779143333435059,
+ "step": 3663
+ },
+ {
+ "epoch": 50.8943231441048,
+ "grad_norm": 0.016851043328642845,
+ "learning_rate": 0.0006,
+ "loss": 4.780387878417969,
+ "step": 3664
+ },
+ {
+ "epoch": 50.90829694323144,
+ "grad_norm": 0.019045202061533928,
+ "learning_rate": 0.0006,
+ "loss": 4.769913673400879,
+ "step": 3665
+ },
+ {
+ "epoch": 50.92227074235808,
+ "grad_norm": 0.01954123005270958,
+ "learning_rate": 0.0006,
+ "loss": 4.769009590148926,
+ "step": 3666
+ },
+ {
+ "epoch": 50.93624454148472,
+ "grad_norm": 0.022527096793055534,
+ "learning_rate": 0.0006,
+ "loss": 4.789215087890625,
+ "step": 3667
+ },
+ {
+ "epoch": 50.95021834061135,
+ "grad_norm": 0.02871634066104889,
+ "learning_rate": 0.0006,
+ "loss": 4.749727725982666,
+ "step": 3668
+ },
+ {
+ "epoch": 50.96419213973799,
+ "grad_norm": 0.04045053571462631,
+ "learning_rate": 0.0006,
+ "loss": 4.909038543701172,
+ "step": 3669
+ },
+ {
+ "epoch": 50.97816593886463,
+ "grad_norm": 0.04156879335641861,
+ "learning_rate": 0.0006,
+ "loss": 4.876831531524658,
+ "step": 3670
+ },
+ {
+ "epoch": 50.992139737991266,
+ "grad_norm": 0.02621397376060486,
+ "learning_rate": 0.0006,
+ "loss": 4.797173500061035,
+ "step": 3671
+ },
+ {
+ "epoch": 51.0,
+ "grad_norm": 0.03389597684144974,
+ "learning_rate": 0.0006,
+ "loss": 4.8428850173950195,
+ "step": 3672
+ },
+ {
+ "epoch": 51.0,
+ "eval_loss": 4.989261627197266,
+ "eval_runtime": 56.358,
+ "eval_samples_per_second": 43.33,
+ "eval_steps_per_second": 1.366,
+ "step": 3672
+ },
+ {
+ "epoch": 51.01397379912664,
+ "grad_norm": 0.03216972202062607,
+ "learning_rate": 0.0006,
+ "loss": 4.755102634429932,
+ "step": 3673
+ },
+ {
+ "epoch": 51.02794759825328,
+ "grad_norm": 0.03190374746918678,
+ "learning_rate": 0.0006,
+ "loss": 4.720424652099609,
+ "step": 3674
+ },
+ {
+ "epoch": 51.041921397379916,
+ "grad_norm": 0.031727299094200134,
+ "learning_rate": 0.0006,
+ "loss": 4.803277969360352,
+ "step": 3675
+ },
+ {
+ "epoch": 51.05589519650655,
+ "grad_norm": 0.02473079040646553,
+ "learning_rate": 0.0006,
+ "loss": 4.761087417602539,
+ "step": 3676
+ },
+ {
+ "epoch": 51.069868995633186,
+ "grad_norm": 0.020360209047794342,
+ "learning_rate": 0.0006,
+ "loss": 4.753884792327881,
+ "step": 3677
+ },
+ {
+ "epoch": 51.083842794759825,
+ "grad_norm": 0.02078923024237156,
+ "learning_rate": 0.0006,
+ "loss": 4.647130489349365,
+ "step": 3678
+ },
+ {
+ "epoch": 51.09781659388646,
+ "grad_norm": 0.019387200474739075,
+ "learning_rate": 0.0006,
+ "loss": 4.74135160446167,
+ "step": 3679
+ },
+ {
+ "epoch": 51.1117903930131,
+ "grad_norm": 0.01874949410557747,
+ "learning_rate": 0.0006,
+ "loss": 4.795329570770264,
+ "step": 3680
+ },
+ {
+ "epoch": 51.12576419213974,
+ "grad_norm": 0.019331837072968483,
+ "learning_rate": 0.0006,
+ "loss": 4.897967338562012,
+ "step": 3681
+ },
+ {
+ "epoch": 51.13973799126637,
+ "grad_norm": 0.017066050320863724,
+ "learning_rate": 0.0006,
+ "loss": 4.825351715087891,
+ "step": 3682
+ },
+ {
+ "epoch": 51.15371179039301,
+ "grad_norm": 0.01757810264825821,
+ "learning_rate": 0.0006,
+ "loss": 4.719335079193115,
+ "step": 3683
+ },
+ {
+ "epoch": 51.16768558951965,
+ "grad_norm": 0.018174266442656517,
+ "learning_rate": 0.0006,
+ "loss": 4.732680320739746,
+ "step": 3684
+ },
+ {
+ "epoch": 51.18165938864629,
+ "grad_norm": 0.018497727811336517,
+ "learning_rate": 0.0006,
+ "loss": 4.731050491333008,
+ "step": 3685
+ },
+ {
+ "epoch": 51.19563318777293,
+ "grad_norm": 0.0186338871717453,
+ "learning_rate": 0.0006,
+ "loss": 4.748035430908203,
+ "step": 3686
+ },
+ {
+ "epoch": 51.209606986899566,
+ "grad_norm": 0.018531711772084236,
+ "learning_rate": 0.0006,
+ "loss": 4.9007463455200195,
+ "step": 3687
+ },
+ {
+ "epoch": 51.223580786026204,
+ "grad_norm": 0.018221354112029076,
+ "learning_rate": 0.0006,
+ "loss": 4.84077262878418,
+ "step": 3688
+ },
+ {
+ "epoch": 51.237554585152836,
+ "grad_norm": 0.0203656405210495,
+ "learning_rate": 0.0006,
+ "loss": 4.73009729385376,
+ "step": 3689
+ },
+ {
+ "epoch": 51.251528384279474,
+ "grad_norm": 0.01964893378317356,
+ "learning_rate": 0.0006,
+ "loss": 4.772583961486816,
+ "step": 3690
+ },
+ {
+ "epoch": 51.26550218340611,
+ "grad_norm": 0.019606510177254677,
+ "learning_rate": 0.0006,
+ "loss": 4.599812030792236,
+ "step": 3691
+ },
+ {
+ "epoch": 51.27947598253275,
+ "grad_norm": 0.020501255989074707,
+ "learning_rate": 0.0006,
+ "loss": 4.809898853302002,
+ "step": 3692
+ },
+ {
+ "epoch": 51.29344978165939,
+ "grad_norm": 0.01870838552713394,
+ "learning_rate": 0.0006,
+ "loss": 4.741898059844971,
+ "step": 3693
+ },
+ {
+ "epoch": 51.30742358078603,
+ "grad_norm": 0.01802559196949005,
+ "learning_rate": 0.0006,
+ "loss": 4.746562480926514,
+ "step": 3694
+ },
+ {
+ "epoch": 51.32139737991266,
+ "grad_norm": 0.02053048275411129,
+ "learning_rate": 0.0006,
+ "loss": 4.69932746887207,
+ "step": 3695
+ },
+ {
+ "epoch": 51.3353711790393,
+ "grad_norm": 0.02661825157701969,
+ "learning_rate": 0.0006,
+ "loss": 4.68748664855957,
+ "step": 3696
+ },
+ {
+ "epoch": 51.34934497816594,
+ "grad_norm": 0.028269147500395775,
+ "learning_rate": 0.0006,
+ "loss": 4.731565475463867,
+ "step": 3697
+ },
+ {
+ "epoch": 51.36331877729258,
+ "grad_norm": 0.02650964818894863,
+ "learning_rate": 0.0006,
+ "loss": 4.733857154846191,
+ "step": 3698
+ },
+ {
+ "epoch": 51.377292576419215,
+ "grad_norm": 0.023696066811680794,
+ "learning_rate": 0.0006,
+ "loss": 4.812362194061279,
+ "step": 3699
+ },
+ {
+ "epoch": 51.391266375545854,
+ "grad_norm": 0.02365880273282528,
+ "learning_rate": 0.0006,
+ "loss": 4.733426094055176,
+ "step": 3700
+ },
+ {
+ "epoch": 51.40524017467249,
+ "grad_norm": 0.026249447837471962,
+ "learning_rate": 0.0006,
+ "loss": 4.715622901916504,
+ "step": 3701
+ },
+ {
+ "epoch": 51.419213973799124,
+ "grad_norm": 0.03301899880170822,
+ "learning_rate": 0.0006,
+ "loss": 4.7223663330078125,
+ "step": 3702
+ },
+ {
+ "epoch": 51.43318777292576,
+ "grad_norm": 0.044686660170555115,
+ "learning_rate": 0.0006,
+ "loss": 4.8121161460876465,
+ "step": 3703
+ },
+ {
+ "epoch": 51.4471615720524,
+ "grad_norm": 0.04733522981405258,
+ "learning_rate": 0.0006,
+ "loss": 4.841394424438477,
+ "step": 3704
+ },
+ {
+ "epoch": 51.46113537117904,
+ "grad_norm": 0.0395209975540638,
+ "learning_rate": 0.0006,
+ "loss": 4.773837089538574,
+ "step": 3705
+ },
+ {
+ "epoch": 51.47510917030568,
+ "grad_norm": 0.032790958881378174,
+ "learning_rate": 0.0006,
+ "loss": 4.833066463470459,
+ "step": 3706
+ },
+ {
+ "epoch": 51.48908296943232,
+ "grad_norm": 0.030578427016735077,
+ "learning_rate": 0.0006,
+ "loss": 4.685190200805664,
+ "step": 3707
+ },
+ {
+ "epoch": 51.50305676855895,
+ "grad_norm": 0.031619083136320114,
+ "learning_rate": 0.0006,
+ "loss": 4.740378379821777,
+ "step": 3708
+ },
+ {
+ "epoch": 51.51703056768559,
+ "grad_norm": 0.02385825850069523,
+ "learning_rate": 0.0006,
+ "loss": 4.812911510467529,
+ "step": 3709
+ },
+ {
+ "epoch": 51.531004366812226,
+ "grad_norm": 0.023523425683379173,
+ "learning_rate": 0.0006,
+ "loss": 4.669580459594727,
+ "step": 3710
+ },
+ {
+ "epoch": 51.544978165938865,
+ "grad_norm": 0.021316280588507652,
+ "learning_rate": 0.0006,
+ "loss": 4.76888370513916,
+ "step": 3711
+ },
+ {
+ "epoch": 51.5589519650655,
+ "grad_norm": 0.024663880467414856,
+ "learning_rate": 0.0006,
+ "loss": 4.702592372894287,
+ "step": 3712
+ },
+ {
+ "epoch": 51.57292576419214,
+ "grad_norm": 0.024173280224204063,
+ "learning_rate": 0.0006,
+ "loss": 4.710604667663574,
+ "step": 3713
+ },
+ {
+ "epoch": 51.58689956331878,
+ "grad_norm": 0.02754114381968975,
+ "learning_rate": 0.0006,
+ "loss": 4.746356964111328,
+ "step": 3714
+ },
+ {
+ "epoch": 51.60087336244541,
+ "grad_norm": 0.027903622016310692,
+ "learning_rate": 0.0006,
+ "loss": 4.7546491622924805,
+ "step": 3715
+ },
+ {
+ "epoch": 51.61484716157205,
+ "grad_norm": 0.024871617555618286,
+ "learning_rate": 0.0006,
+ "loss": 4.605746269226074,
+ "step": 3716
+ },
+ {
+ "epoch": 51.62882096069869,
+ "grad_norm": 0.021239031106233597,
+ "learning_rate": 0.0006,
+ "loss": 4.6942458152771,
+ "step": 3717
+ },
+ {
+ "epoch": 51.64279475982533,
+ "grad_norm": 0.01925622671842575,
+ "learning_rate": 0.0006,
+ "loss": 4.593558311462402,
+ "step": 3718
+ },
+ {
+ "epoch": 51.65676855895197,
+ "grad_norm": 0.01956283673644066,
+ "learning_rate": 0.0006,
+ "loss": 4.631172180175781,
+ "step": 3719
+ },
+ {
+ "epoch": 51.670742358078606,
+ "grad_norm": 0.02269871160387993,
+ "learning_rate": 0.0006,
+ "loss": 4.729236602783203,
+ "step": 3720
+ },
+ {
+ "epoch": 51.68471615720524,
+ "grad_norm": 0.02253916673362255,
+ "learning_rate": 0.0006,
+ "loss": 4.7191033363342285,
+ "step": 3721
+ },
+ {
+ "epoch": 51.698689956331876,
+ "grad_norm": 0.026231050491333008,
+ "learning_rate": 0.0006,
+ "loss": 4.811850547790527,
+ "step": 3722
+ },
+ {
+ "epoch": 51.712663755458514,
+ "grad_norm": 0.02558821812272072,
+ "learning_rate": 0.0006,
+ "loss": 4.681466102600098,
+ "step": 3723
+ },
+ {
+ "epoch": 51.72663755458515,
+ "grad_norm": 0.02124209702014923,
+ "learning_rate": 0.0006,
+ "loss": 4.693664073944092,
+ "step": 3724
+ },
+ {
+ "epoch": 51.74061135371179,
+ "grad_norm": 0.016638319939374924,
+ "learning_rate": 0.0006,
+ "loss": 4.771440029144287,
+ "step": 3725
+ },
+ {
+ "epoch": 51.75458515283843,
+ "grad_norm": 0.018048347905278206,
+ "learning_rate": 0.0006,
+ "loss": 4.643252372741699,
+ "step": 3726
+ },
+ {
+ "epoch": 51.76855895196506,
+ "grad_norm": 0.019904717803001404,
+ "learning_rate": 0.0006,
+ "loss": 4.66530704498291,
+ "step": 3727
+ },
+ {
+ "epoch": 51.7825327510917,
+ "grad_norm": 0.0194614976644516,
+ "learning_rate": 0.0006,
+ "loss": 4.711943626403809,
+ "step": 3728
+ },
+ {
+ "epoch": 51.79650655021834,
+ "grad_norm": 0.019175555557012558,
+ "learning_rate": 0.0006,
+ "loss": 4.721931457519531,
+ "step": 3729
+ },
+ {
+ "epoch": 51.81048034934498,
+ "grad_norm": 0.02093447372317314,
+ "learning_rate": 0.0006,
+ "loss": 4.633059024810791,
+ "step": 3730
+ },
+ {
+ "epoch": 51.82445414847162,
+ "grad_norm": 0.01930283196270466,
+ "learning_rate": 0.0006,
+ "loss": 4.659373760223389,
+ "step": 3731
+ },
+ {
+ "epoch": 51.838427947598255,
+ "grad_norm": 0.015350676141679287,
+ "learning_rate": 0.0006,
+ "loss": 4.763708114624023,
+ "step": 3732
+ },
+ {
+ "epoch": 51.852401746724894,
+ "grad_norm": 0.014139448292553425,
+ "learning_rate": 0.0006,
+ "loss": 4.753891468048096,
+ "step": 3733
+ },
+ {
+ "epoch": 51.866375545851525,
+ "grad_norm": 0.016215872019529343,
+ "learning_rate": 0.0006,
+ "loss": 4.658967018127441,
+ "step": 3734
+ },
+ {
+ "epoch": 51.880349344978164,
+ "grad_norm": 0.017328619956970215,
+ "learning_rate": 0.0006,
+ "loss": 4.696579933166504,
+ "step": 3735
+ },
+ {
+ "epoch": 51.8943231441048,
+ "grad_norm": 0.018103396520018578,
+ "learning_rate": 0.0006,
+ "loss": 4.627464771270752,
+ "step": 3736
+ },
+ {
+ "epoch": 51.90829694323144,
+ "grad_norm": 0.01715235412120819,
+ "learning_rate": 0.0006,
+ "loss": 4.7368011474609375,
+ "step": 3737
+ },
+ {
+ "epoch": 51.92227074235808,
+ "grad_norm": 0.014790094457566738,
+ "learning_rate": 0.0006,
+ "loss": 4.629441738128662,
+ "step": 3738
+ },
+ {
+ "epoch": 51.93624454148472,
+ "grad_norm": 0.017114296555519104,
+ "learning_rate": 0.0006,
+ "loss": 4.761256694793701,
+ "step": 3739
+ },
+ {
+ "epoch": 51.95021834061135,
+ "grad_norm": 0.019009845331311226,
+ "learning_rate": 0.0006,
+ "loss": 4.789155006408691,
+ "step": 3740
+ },
+ {
+ "epoch": 51.96419213973799,
+ "grad_norm": 0.019696692004799843,
+ "learning_rate": 0.0006,
+ "loss": 4.721269607543945,
+ "step": 3741
+ },
+ {
+ "epoch": 51.97816593886463,
+ "grad_norm": 0.020542822778224945,
+ "learning_rate": 0.0006,
+ "loss": 4.687038898468018,
+ "step": 3742
+ },
+ {
+ "epoch": 51.992139737991266,
+ "grad_norm": 0.021267160773277283,
+ "learning_rate": 0.0006,
+ "loss": 4.747860908508301,
+ "step": 3743
+ },
+ {
+ "epoch": 52.0,
+ "grad_norm": 0.025049567222595215,
+ "learning_rate": 0.0006,
+ "loss": 4.8473100662231445,
+ "step": 3744
+ },
+ {
+ "epoch": 52.0,
+ "eval_loss": 4.943043231964111,
+ "eval_runtime": 56.4929,
+ "eval_samples_per_second": 43.227,
+ "eval_steps_per_second": 1.363,
+ "step": 3744
+ },
+ {
+ "epoch": 52.01397379912664,
+ "grad_norm": 0.03054986707866192,
+ "learning_rate": 0.0006,
+ "loss": 4.856931686401367,
+ "step": 3745
+ },
+ {
+ "epoch": 52.02794759825328,
+ "grad_norm": 0.03838375583291054,
+ "learning_rate": 0.0006,
+ "loss": 4.699807167053223,
+ "step": 3746
+ },
+ {
+ "epoch": 52.041921397379916,
+ "grad_norm": 0.03383488208055496,
+ "learning_rate": 0.0006,
+ "loss": 4.6626787185668945,
+ "step": 3747
+ },
+ {
+ "epoch": 52.05589519650655,
+ "grad_norm": 0.0201561376452446,
+ "learning_rate": 0.0006,
+ "loss": 4.621831893920898,
+ "step": 3748
+ },
+ {
+ "epoch": 52.069868995633186,
+ "grad_norm": 0.028749438002705574,
+ "learning_rate": 0.0006,
+ "loss": 4.682540416717529,
+ "step": 3749
+ },
+ {
+ "epoch": 52.083842794759825,
+ "grad_norm": 0.02681174874305725,
+ "learning_rate": 0.0006,
+ "loss": 4.629997253417969,
+ "step": 3750
+ },
+ {
+ "epoch": 52.09781659388646,
+ "grad_norm": 0.024063318967819214,
+ "learning_rate": 0.0006,
+ "loss": 4.573021411895752,
+ "step": 3751
+ },
+ {
+ "epoch": 52.1117903930131,
+ "grad_norm": 0.02324330434203148,
+ "learning_rate": 0.0006,
+ "loss": 4.739465236663818,
+ "step": 3752
+ },
+ {
+ "epoch": 52.12576419213974,
+ "grad_norm": 0.0225426834076643,
+ "learning_rate": 0.0006,
+ "loss": 4.7335333824157715,
+ "step": 3753
+ },
+ {
+ "epoch": 52.13973799126637,
+ "grad_norm": 0.019593428820371628,
+ "learning_rate": 0.0006,
+ "loss": 4.704958438873291,
+ "step": 3754
+ },
+ {
+ "epoch": 52.15371179039301,
+ "grad_norm": 0.020777082070708275,
+ "learning_rate": 0.0006,
+ "loss": 4.638696670532227,
+ "step": 3755
+ },
+ {
+ "epoch": 52.16768558951965,
+ "grad_norm": 0.021525604650378227,
+ "learning_rate": 0.0006,
+ "loss": 4.643706798553467,
+ "step": 3756
+ },
+ {
+ "epoch": 52.18165938864629,
+ "grad_norm": 0.018943654373288155,
+ "learning_rate": 0.0006,
+ "loss": 4.5173211097717285,
+ "step": 3757
+ },
+ {
+ "epoch": 52.19563318777293,
+ "grad_norm": 0.01856580562889576,
+ "learning_rate": 0.0006,
+ "loss": 4.678788661956787,
+ "step": 3758
+ },
+ {
+ "epoch": 52.209606986899566,
+ "grad_norm": 0.019417259842157364,
+ "learning_rate": 0.0006,
+ "loss": 4.587244987487793,
+ "step": 3759
+ },
+ {
+ "epoch": 52.223580786026204,
+ "grad_norm": 0.01911911740899086,
+ "learning_rate": 0.0006,
+ "loss": 4.688503265380859,
+ "step": 3760
+ },
+ {
+ "epoch": 52.237554585152836,
+ "grad_norm": 0.021017879247665405,
+ "learning_rate": 0.0006,
+ "loss": 4.65829610824585,
+ "step": 3761
+ },
+ {
+ "epoch": 52.251528384279474,
+ "grad_norm": 0.02048875205218792,
+ "learning_rate": 0.0006,
+ "loss": 4.624055862426758,
+ "step": 3762
+ },
+ {
+ "epoch": 52.26550218340611,
+ "grad_norm": 0.01932772807776928,
+ "learning_rate": 0.0006,
+ "loss": 4.556562900543213,
+ "step": 3763
+ },
+ {
+ "epoch": 52.27947598253275,
+ "grad_norm": 0.01871614344418049,
+ "learning_rate": 0.0006,
+ "loss": 4.581201076507568,
+ "step": 3764
+ },
+ {
+ "epoch": 52.29344978165939,
+ "grad_norm": 0.017253106459975243,
+ "learning_rate": 0.0006,
+ "loss": 4.605219841003418,
+ "step": 3765
+ },
+ {
+ "epoch": 52.30742358078603,
+ "grad_norm": 0.01831238530576229,
+ "learning_rate": 0.0006,
+ "loss": 4.759900093078613,
+ "step": 3766
+ },
+ {
+ "epoch": 52.32139737991266,
+ "grad_norm": 0.017946207895874977,
+ "learning_rate": 0.0006,
+ "loss": 4.5598344802856445,
+ "step": 3767
+ },
+ {
+ "epoch": 52.3353711790393,
+ "grad_norm": 0.01680106669664383,
+ "learning_rate": 0.0006,
+ "loss": 4.605034828186035,
+ "step": 3768
+ },
+ {
+ "epoch": 52.34934497816594,
+ "grad_norm": 0.021482540294528008,
+ "learning_rate": 0.0006,
+ "loss": 4.620411396026611,
+ "step": 3769
+ },
+ {
+ "epoch": 52.36331877729258,
+ "grad_norm": 0.02210252732038498,
+ "learning_rate": 0.0006,
+ "loss": 4.667577743530273,
+ "step": 3770
+ },
+ {
+ "epoch": 52.377292576419215,
+ "grad_norm": 0.02237485907971859,
+ "learning_rate": 0.0006,
+ "loss": 4.628812789916992,
+ "step": 3771
+ },
+ {
+ "epoch": 52.391266375545854,
+ "grad_norm": 0.02196541801095009,
+ "learning_rate": 0.0006,
+ "loss": 4.572988510131836,
+ "step": 3772
+ },
+ {
+ "epoch": 52.40524017467249,
+ "grad_norm": 0.01930229924619198,
+ "learning_rate": 0.0006,
+ "loss": 4.588339805603027,
+ "step": 3773
+ },
+ {
+ "epoch": 52.419213973799124,
+ "grad_norm": 0.02222434990108013,
+ "learning_rate": 0.0006,
+ "loss": 4.766482353210449,
+ "step": 3774
+ },
+ {
+ "epoch": 52.43318777292576,
+ "grad_norm": 0.028806306421756744,
+ "learning_rate": 0.0006,
+ "loss": 4.617892742156982,
+ "step": 3775
+ },
+ {
+ "epoch": 52.4471615720524,
+ "grad_norm": 0.03306042402982712,
+ "learning_rate": 0.0006,
+ "loss": 4.665334701538086,
+ "step": 3776
+ },
+ {
+ "epoch": 52.46113537117904,
+ "grad_norm": 0.0358956903219223,
+ "learning_rate": 0.0006,
+ "loss": 4.566668510437012,
+ "step": 3777
+ },
+ {
+ "epoch": 52.47510917030568,
+ "grad_norm": 0.030406497418880463,
+ "learning_rate": 0.0006,
+ "loss": 4.513612270355225,
+ "step": 3778
+ },
+ {
+ "epoch": 52.48908296943232,
+ "grad_norm": 0.030295712873339653,
+ "learning_rate": 0.0006,
+ "loss": 4.640189170837402,
+ "step": 3779
+ },
+ {
+ "epoch": 52.50305676855895,
+ "grad_norm": 0.024981003254652023,
+ "learning_rate": 0.0006,
+ "loss": 4.57558012008667,
+ "step": 3780
+ },
+ {
+ "epoch": 52.51703056768559,
+ "grad_norm": 0.024990713223814964,
+ "learning_rate": 0.0006,
+ "loss": 4.708589553833008,
+ "step": 3781
+ },
+ {
+ "epoch": 52.531004366812226,
+ "grad_norm": 0.025034775957465172,
+ "learning_rate": 0.0006,
+ "loss": 4.70468282699585,
+ "step": 3782
+ },
+ {
+ "epoch": 52.544978165938865,
+ "grad_norm": 0.028243759647011757,
+ "learning_rate": 0.0006,
+ "loss": 4.712060451507568,
+ "step": 3783
+ },
+ {
+ "epoch": 52.5589519650655,
+ "grad_norm": 0.026984544470906258,
+ "learning_rate": 0.0006,
+ "loss": 4.505004405975342,
+ "step": 3784
+ },
+ {
+ "epoch": 52.57292576419214,
+ "grad_norm": 0.050843529403209686,
+ "learning_rate": 0.0006,
+ "loss": 4.593271255493164,
+ "step": 3785
+ },
+ {
+ "epoch": 52.58689956331878,
+ "grad_norm": 0.25422802567481995,
+ "learning_rate": 0.0006,
+ "loss": 4.7178802490234375,
+ "step": 3786
+ },
+ {
+ "epoch": 52.60087336244541,
+ "grad_norm": 0.6078406572341919,
+ "learning_rate": 0.0006,
+ "loss": 6.144372940063477,
+ "step": 3787
+ },
+ {
+ "epoch": 52.61484716157205,
+ "grad_norm": 0.25933101773262024,
+ "learning_rate": 0.0006,
+ "loss": 5.971083641052246,
+ "step": 3788
+ },
+ {
+ "epoch": 52.62882096069869,
+ "grad_norm": 0.294657826423645,
+ "learning_rate": 0.0006,
+ "loss": 6.204090118408203,
+ "step": 3789
+ },
+ {
+ "epoch": 52.64279475982533,
+ "grad_norm": 0.23641522228717804,
+ "learning_rate": 0.0006,
+ "loss": 6.869574546813965,
+ "step": 3790
+ },
+ {
+ "epoch": 52.65676855895197,
+ "grad_norm": 0.23717311024665833,
+ "learning_rate": 0.0006,
+ "loss": 6.739034175872803,
+ "step": 3791
+ },
+ {
+ "epoch": 52.670742358078606,
+ "grad_norm": 0.10300362855195999,
+ "learning_rate": 0.0006,
+ "loss": 6.49191951751709,
+ "step": 3792
+ },
+ {
+ "epoch": 52.68471615720524,
+ "grad_norm": 0.1484968513250351,
+ "learning_rate": 0.0006,
+ "loss": 6.6021270751953125,
+ "step": 3793
+ },
+ {
+ "epoch": 52.698689956331876,
+ "grad_norm": 0.08052115142345428,
+ "learning_rate": 0.0006,
+ "loss": 6.432796478271484,
+ "step": 3794
+ },
+ {
+ "epoch": 52.712663755458514,
+ "grad_norm": 0.06869253516197205,
+ "learning_rate": 0.0006,
+ "loss": 6.212784767150879,
+ "step": 3795
+ },
+ {
+ "epoch": 52.72663755458515,
+ "grad_norm": 0.05589582398533821,
+ "learning_rate": 0.0006,
+ "loss": 6.168288707733154,
+ "step": 3796
+ },
+ {
+ "epoch": 52.74061135371179,
+ "grad_norm": 0.04840940609574318,
+ "learning_rate": 0.0006,
+ "loss": 6.163174629211426,
+ "step": 3797
+ },
+ {
+ "epoch": 52.75458515283843,
+ "grad_norm": 0.04578601196408272,
+ "learning_rate": 0.0006,
+ "loss": 6.038390636444092,
+ "step": 3798
+ },
+ {
+ "epoch": 52.76855895196506,
+ "grad_norm": 0.04102053865790367,
+ "learning_rate": 0.0006,
+ "loss": 5.929539680480957,
+ "step": 3799
+ },
+ {
+ "epoch": 52.7825327510917,
+ "grad_norm": 0.038398485630750656,
+ "learning_rate": 0.0006,
+ "loss": 6.007378101348877,
+ "step": 3800
+ },
+ {
+ "epoch": 52.79650655021834,
+ "grad_norm": 0.03466642275452614,
+ "learning_rate": 0.0006,
+ "loss": 5.767291069030762,
+ "step": 3801
+ },
+ {
+ "epoch": 52.81048034934498,
+ "grad_norm": 0.03328905999660492,
+ "learning_rate": 0.0006,
+ "loss": 5.785032272338867,
+ "step": 3802
+ },
+ {
+ "epoch": 52.82445414847162,
+ "grad_norm": 0.03186681121587753,
+ "learning_rate": 0.0006,
+ "loss": 5.6942057609558105,
+ "step": 3803
+ },
+ {
+ "epoch": 52.838427947598255,
+ "grad_norm": 0.030553551390767097,
+ "learning_rate": 0.0006,
+ "loss": 5.680774688720703,
+ "step": 3804
+ },
+ {
+ "epoch": 52.852401746724894,
+ "grad_norm": 0.02578839473426342,
+ "learning_rate": 0.0006,
+ "loss": 5.703461647033691,
+ "step": 3805
+ },
+ {
+ "epoch": 52.866375545851525,
+ "grad_norm": 0.026261312887072563,
+ "learning_rate": 0.0006,
+ "loss": 5.642549514770508,
+ "step": 3806
+ },
+ {
+ "epoch": 52.880349344978164,
+ "grad_norm": 0.02782430872321129,
+ "learning_rate": 0.0006,
+ "loss": 5.563597679138184,
+ "step": 3807
+ },
+ {
+ "epoch": 52.8943231441048,
+ "grad_norm": 0.024467799812555313,
+ "learning_rate": 0.0006,
+ "loss": 5.583395957946777,
+ "step": 3808
+ },
+ {
+ "epoch": 52.90829694323144,
+ "grad_norm": 0.022602897137403488,
+ "learning_rate": 0.0006,
+ "loss": 5.50742769241333,
+ "step": 3809
+ },
+ {
+ "epoch": 52.92227074235808,
+ "grad_norm": 0.024292029440402985,
+ "learning_rate": 0.0006,
+ "loss": 5.367804527282715,
+ "step": 3810
+ },
+ {
+ "epoch": 52.93624454148472,
+ "grad_norm": 0.021062295883893967,
+ "learning_rate": 0.0006,
+ "loss": 5.423705101013184,
+ "step": 3811
+ },
+ {
+ "epoch": 52.95021834061135,
+ "grad_norm": 0.022226469591259956,
+ "learning_rate": 0.0006,
+ "loss": 5.344039440155029,
+ "step": 3812
+ },
+ {
+ "epoch": 52.96419213973799,
+ "grad_norm": 0.023691752925515175,
+ "learning_rate": 0.0006,
+ "loss": 5.2581892013549805,
+ "step": 3813
+ },
+ {
+ "epoch": 52.97816593886463,
+ "grad_norm": 0.02181999757885933,
+ "learning_rate": 0.0006,
+ "loss": 5.2710113525390625,
+ "step": 3814
+ },
+ {
+ "epoch": 52.992139737991266,
+ "grad_norm": 0.025295857340097427,
+ "learning_rate": 0.0006,
+ "loss": 5.278355598449707,
+ "step": 3815
+ },
+ {
+ "epoch": 53.0,
+ "grad_norm": 0.02216828241944313,
+ "learning_rate": 0.0006,
+ "loss": 5.127945899963379,
+ "step": 3816
+ },
+ {
+ "epoch": 53.0,
+ "eval_loss": 5.3411736488342285,
+ "eval_runtime": 56.2837,
+ "eval_samples_per_second": 43.387,
+ "eval_steps_per_second": 1.368,
+ "step": 3816
+ },
+ {
+ "epoch": 53.01397379912664,
+ "grad_norm": 0.022496026009321213,
+ "learning_rate": 0.0006,
+ "loss": 5.19655704498291,
+ "step": 3817
+ },
+ {
+ "epoch": 53.02794759825328,
+ "grad_norm": 0.03201168775558472,
+ "learning_rate": 0.0006,
+ "loss": 5.094621181488037,
+ "step": 3818
+ },
+ {
+ "epoch": 53.041921397379916,
+ "grad_norm": 0.02683178335428238,
+ "learning_rate": 0.0006,
+ "loss": 5.19899845123291,
+ "step": 3819
+ },
+ {
+ "epoch": 53.05589519650655,
+ "grad_norm": 0.022659169510006905,
+ "learning_rate": 0.0006,
+ "loss": 5.184103012084961,
+ "step": 3820
+ },
+ {
+ "epoch": 53.069868995633186,
+ "grad_norm": 0.024784443899989128,
+ "learning_rate": 0.0006,
+ "loss": 5.124286651611328,
+ "step": 3821
+ },
+ {
+ "epoch": 53.083842794759825,
+ "grad_norm": 0.028710370883345604,
+ "learning_rate": 0.0006,
+ "loss": 5.130426406860352,
+ "step": 3822
+ },
+ {
+ "epoch": 53.09781659388646,
+ "grad_norm": 0.02366088517010212,
+ "learning_rate": 0.0006,
+ "loss": 5.1453399658203125,
+ "step": 3823
+ },
+ {
+ "epoch": 53.1117903930131,
+ "grad_norm": 0.01813938096165657,
+ "learning_rate": 0.0006,
+ "loss": 5.05271053314209,
+ "step": 3824
+ },
+ {
+ "epoch": 53.12576419213974,
+ "grad_norm": 0.02051435597240925,
+ "learning_rate": 0.0006,
+ "loss": 5.023038864135742,
+ "step": 3825
+ },
+ {
+ "epoch": 53.13973799126637,
+ "grad_norm": 0.02185754105448723,
+ "learning_rate": 0.0006,
+ "loss": 4.8950018882751465,
+ "step": 3826
+ },
+ {
+ "epoch": 53.15371179039301,
+ "grad_norm": 0.017223743721842766,
+ "learning_rate": 0.0006,
+ "loss": 5.078621864318848,
+ "step": 3827
+ },
+ {
+ "epoch": 53.16768558951965,
+ "grad_norm": 0.01832217164337635,
+ "learning_rate": 0.0006,
+ "loss": 4.921341419219971,
+ "step": 3828
+ },
+ {
+ "epoch": 53.18165938864629,
+ "grad_norm": 0.02114005573093891,
+ "learning_rate": 0.0006,
+ "loss": 4.997865676879883,
+ "step": 3829
+ },
+ {
+ "epoch": 53.19563318777293,
+ "grad_norm": 0.021426431834697723,
+ "learning_rate": 0.0006,
+ "loss": 4.877785682678223,
+ "step": 3830
+ },
+ {
+ "epoch": 53.209606986899566,
+ "grad_norm": 0.021327560767531395,
+ "learning_rate": 0.0006,
+ "loss": 4.847512245178223,
+ "step": 3831
+ },
+ {
+ "epoch": 53.223580786026204,
+ "grad_norm": 0.027825530618429184,
+ "learning_rate": 0.0006,
+ "loss": 5.004823207855225,
+ "step": 3832
+ },
+ {
+ "epoch": 53.237554585152836,
+ "grad_norm": 0.03756481409072876,
+ "learning_rate": 0.0006,
+ "loss": 4.817290306091309,
+ "step": 3833
+ },
+ {
+ "epoch": 53.251528384279474,
+ "grad_norm": 0.04933721572160721,
+ "learning_rate": 0.0006,
+ "loss": 4.863317489624023,
+ "step": 3834
+ },
+ {
+ "epoch": 53.26550218340611,
+ "grad_norm": 0.04588694870471954,
+ "learning_rate": 0.0006,
+ "loss": 4.864742755889893,
+ "step": 3835
+ },
+ {
+ "epoch": 53.27947598253275,
+ "grad_norm": 0.02728264033794403,
+ "learning_rate": 0.0006,
+ "loss": 4.9068756103515625,
+ "step": 3836
+ },
+ {
+ "epoch": 53.29344978165939,
+ "grad_norm": 0.036240726709365845,
+ "learning_rate": 0.0006,
+ "loss": 4.902045249938965,
+ "step": 3837
+ },
+ {
+ "epoch": 53.30742358078603,
+ "grad_norm": 0.02703195810317993,
+ "learning_rate": 0.0006,
+ "loss": 4.908980369567871,
+ "step": 3838
+ },
+ {
+ "epoch": 53.32139737991266,
+ "grad_norm": 0.024894049391150475,
+ "learning_rate": 0.0006,
+ "loss": 4.794848442077637,
+ "step": 3839
+ },
+ {
+ "epoch": 53.3353711790393,
+ "grad_norm": 0.02389993891119957,
+ "learning_rate": 0.0006,
+ "loss": 4.8234734535217285,
+ "step": 3840
+ },
+ {
+ "epoch": 53.34934497816594,
+ "grad_norm": 0.02486400306224823,
+ "learning_rate": 0.0006,
+ "loss": 4.851463794708252,
+ "step": 3841
+ },
+ {
+ "epoch": 53.36331877729258,
+ "grad_norm": 0.02613813802599907,
+ "learning_rate": 0.0006,
+ "loss": 4.860444068908691,
+ "step": 3842
+ },
+ {
+ "epoch": 53.377292576419215,
+ "grad_norm": 0.0229744054377079,
+ "learning_rate": 0.0006,
+ "loss": 4.709606170654297,
+ "step": 3843
+ },
+ {
+ "epoch": 53.391266375545854,
+ "grad_norm": 0.02123042568564415,
+ "learning_rate": 0.0006,
+ "loss": 4.654613494873047,
+ "step": 3844
+ },
+ {
+ "epoch": 53.40524017467249,
+ "grad_norm": 0.017352338880300522,
+ "learning_rate": 0.0006,
+ "loss": 4.769761085510254,
+ "step": 3845
+ },
+ {
+ "epoch": 53.419213973799124,
+ "grad_norm": 0.019808724522590637,
+ "learning_rate": 0.0006,
+ "loss": 4.98514461517334,
+ "step": 3846
+ },
+ {
+ "epoch": 53.43318777292576,
+ "grad_norm": 0.018776152282953262,
+ "learning_rate": 0.0006,
+ "loss": 4.749990463256836,
+ "step": 3847
+ },
+ {
+ "epoch": 53.4471615720524,
+ "grad_norm": 0.02017521858215332,
+ "learning_rate": 0.0006,
+ "loss": 4.757381439208984,
+ "step": 3848
+ },
+ {
+ "epoch": 53.46113537117904,
+ "grad_norm": 0.019509416073560715,
+ "learning_rate": 0.0006,
+ "loss": 4.8111572265625,
+ "step": 3849
+ },
+ {
+ "epoch": 53.47510917030568,
+ "grad_norm": 0.015543580986559391,
+ "learning_rate": 0.0006,
+ "loss": 4.7501678466796875,
+ "step": 3850
+ },
+ {
+ "epoch": 53.48908296943232,
+ "grad_norm": 0.017941651865839958,
+ "learning_rate": 0.0006,
+ "loss": 4.701838493347168,
+ "step": 3851
+ },
+ {
+ "epoch": 53.50305676855895,
+ "grad_norm": 0.01678212359547615,
+ "learning_rate": 0.0006,
+ "loss": 4.787798881530762,
+ "step": 3852
+ },
+ {
+ "epoch": 53.51703056768559,
+ "grad_norm": 0.019050654023885727,
+ "learning_rate": 0.0006,
+ "loss": 4.754859924316406,
+ "step": 3853
+ },
+ {
+ "epoch": 53.531004366812226,
+ "grad_norm": 0.017406078055500984,
+ "learning_rate": 0.0006,
+ "loss": 4.752772808074951,
+ "step": 3854
+ },
+ {
+ "epoch": 53.544978165938865,
+ "grad_norm": 0.01773344725370407,
+ "learning_rate": 0.0006,
+ "loss": 4.651812553405762,
+ "step": 3855
+ },
+ {
+ "epoch": 53.5589519650655,
+ "grad_norm": 0.017221346497535706,
+ "learning_rate": 0.0006,
+ "loss": 4.792244911193848,
+ "step": 3856
+ },
+ {
+ "epoch": 53.57292576419214,
+ "grad_norm": 0.01722758449614048,
+ "learning_rate": 0.0006,
+ "loss": 4.7902960777282715,
+ "step": 3857
+ },
+ {
+ "epoch": 53.58689956331878,
+ "grad_norm": 0.020445192232728004,
+ "learning_rate": 0.0006,
+ "loss": 4.67392110824585,
+ "step": 3858
+ },
+ {
+ "epoch": 53.60087336244541,
+ "grad_norm": 0.02267114631831646,
+ "learning_rate": 0.0006,
+ "loss": 4.796089172363281,
+ "step": 3859
+ },
+ {
+ "epoch": 53.61484716157205,
+ "grad_norm": 0.027129942551255226,
+ "learning_rate": 0.0006,
+ "loss": 4.774303913116455,
+ "step": 3860
+ },
+ {
+ "epoch": 53.62882096069869,
+ "grad_norm": 0.025454839691519737,
+ "learning_rate": 0.0006,
+ "loss": 4.708813190460205,
+ "step": 3861
+ },
+ {
+ "epoch": 53.64279475982533,
+ "grad_norm": 0.019248757511377335,
+ "learning_rate": 0.0006,
+ "loss": 4.760467529296875,
+ "step": 3862
+ },
+ {
+ "epoch": 53.65676855895197,
+ "grad_norm": 0.01791176199913025,
+ "learning_rate": 0.0006,
+ "loss": 4.789203643798828,
+ "step": 3863
+ },
+ {
+ "epoch": 53.670742358078606,
+ "grad_norm": 0.016306612640619278,
+ "learning_rate": 0.0006,
+ "loss": 4.732605934143066,
+ "step": 3864
+ },
+ {
+ "epoch": 53.68471615720524,
+ "grad_norm": 0.017836976796388626,
+ "learning_rate": 0.0006,
+ "loss": 4.691871643066406,
+ "step": 3865
+ },
+ {
+ "epoch": 53.698689956331876,
+ "grad_norm": 0.015102861449122429,
+ "learning_rate": 0.0006,
+ "loss": 4.724156856536865,
+ "step": 3866
+ },
+ {
+ "epoch": 53.712663755458514,
+ "grad_norm": 0.017458142712712288,
+ "learning_rate": 0.0006,
+ "loss": 4.73524808883667,
+ "step": 3867
+ },
+ {
+ "epoch": 53.72663755458515,
+ "grad_norm": 0.01591932773590088,
+ "learning_rate": 0.0006,
+ "loss": 4.705888748168945,
+ "step": 3868
+ },
+ {
+ "epoch": 53.74061135371179,
+ "grad_norm": 0.01713593676686287,
+ "learning_rate": 0.0006,
+ "loss": 4.644516944885254,
+ "step": 3869
+ },
+ {
+ "epoch": 53.75458515283843,
+ "grad_norm": 0.01905885338783264,
+ "learning_rate": 0.0006,
+ "loss": 4.698416233062744,
+ "step": 3870
+ },
+ {
+ "epoch": 53.76855895196506,
+ "grad_norm": 0.01766437292098999,
+ "learning_rate": 0.0006,
+ "loss": 4.6796555519104,
+ "step": 3871
+ },
+ {
+ "epoch": 53.7825327510917,
+ "grad_norm": 0.01552243810147047,
+ "learning_rate": 0.0006,
+ "loss": 4.78138542175293,
+ "step": 3872
+ },
+ {
+ "epoch": 53.79650655021834,
+ "grad_norm": 0.014191331341862679,
+ "learning_rate": 0.0006,
+ "loss": 4.703588485717773,
+ "step": 3873
+ },
+ {
+ "epoch": 53.81048034934498,
+ "grad_norm": 0.015872279182076454,
+ "learning_rate": 0.0006,
+ "loss": 4.705611228942871,
+ "step": 3874
+ },
+ {
+ "epoch": 53.82445414847162,
+ "grad_norm": 0.01787560060620308,
+ "learning_rate": 0.0006,
+ "loss": 4.8227949142456055,
+ "step": 3875
+ },
+ {
+ "epoch": 53.838427947598255,
+ "grad_norm": 0.0158209390938282,
+ "learning_rate": 0.0006,
+ "loss": 4.645227432250977,
+ "step": 3876
+ },
+ {
+ "epoch": 53.852401746724894,
+ "grad_norm": 0.014362678863108158,
+ "learning_rate": 0.0006,
+ "loss": 4.7225422859191895,
+ "step": 3877
+ },
+ {
+ "epoch": 53.866375545851525,
+ "grad_norm": 0.014899161644279957,
+ "learning_rate": 0.0006,
+ "loss": 4.685914039611816,
+ "step": 3878
+ },
+ {
+ "epoch": 53.880349344978164,
+ "grad_norm": 0.01628505438566208,
+ "learning_rate": 0.0006,
+ "loss": 4.674065589904785,
+ "step": 3879
+ },
+ {
+ "epoch": 53.8943231441048,
+ "grad_norm": 0.01873331144452095,
+ "learning_rate": 0.0006,
+ "loss": 4.811744213104248,
+ "step": 3880
+ },
+ {
+ "epoch": 53.90829694323144,
+ "grad_norm": 0.02229069173336029,
+ "learning_rate": 0.0006,
+ "loss": 4.721514701843262,
+ "step": 3881
+ },
+ {
+ "epoch": 53.92227074235808,
+ "grad_norm": 0.028012916445732117,
+ "learning_rate": 0.0006,
+ "loss": 4.773603439331055,
+ "step": 3882
+ },
+ {
+ "epoch": 53.93624454148472,
+ "grad_norm": 0.036322783678770065,
+ "learning_rate": 0.0006,
+ "loss": 4.787134170532227,
+ "step": 3883
+ },
+ {
+ "epoch": 53.95021834061135,
+ "grad_norm": 0.048149142414331436,
+ "learning_rate": 0.0006,
+ "loss": 4.736196517944336,
+ "step": 3884
+ },
+ {
+ "epoch": 53.96419213973799,
+ "grad_norm": 0.051918916404247284,
+ "learning_rate": 0.0006,
+ "loss": 4.686728000640869,
+ "step": 3885
+ },
+ {
+ "epoch": 53.97816593886463,
+ "grad_norm": 0.0481342189013958,
+ "learning_rate": 0.0006,
+ "loss": 4.733730792999268,
+ "step": 3886
+ },
+ {
+ "epoch": 53.992139737991266,
+ "grad_norm": 0.045424818992614746,
+ "learning_rate": 0.0006,
+ "loss": 4.748723983764648,
+ "step": 3887
+ },
+ {
+ "epoch": 54.0,
+ "grad_norm": 0.04153316840529442,
+ "learning_rate": 0.0006,
+ "loss": 4.715220928192139,
+ "step": 3888
+ },
+ {
+ "epoch": 54.0,
+ "eval_loss": 4.964215278625488,
+ "eval_runtime": 56.7757,
+ "eval_samples_per_second": 43.011,
+ "eval_steps_per_second": 1.356,
+ "step": 3888
+ },
+ {
+ "epoch": 54.01397379912664,
+ "grad_norm": 0.03459750860929489,
+ "learning_rate": 0.0006,
+ "loss": 4.661557674407959,
+ "step": 3889
+ },
+ {
+ "epoch": 54.02794759825328,
+ "grad_norm": 0.025409208610653877,
+ "learning_rate": 0.0006,
+ "loss": 4.7065582275390625,
+ "step": 3890
+ },
+ {
+ "epoch": 54.041921397379916,
+ "grad_norm": 0.03202550485730171,
+ "learning_rate": 0.0006,
+ "loss": 4.751070976257324,
+ "step": 3891
+ },
+ {
+ "epoch": 54.05589519650655,
+ "grad_norm": 0.02968018501996994,
+ "learning_rate": 0.0006,
+ "loss": 4.635388374328613,
+ "step": 3892
+ },
+ {
+ "epoch": 54.069868995633186,
+ "grad_norm": 0.02757406234741211,
+ "learning_rate": 0.0006,
+ "loss": 4.706214427947998,
+ "step": 3893
+ },
+ {
+ "epoch": 54.083842794759825,
+ "grad_norm": 0.025223620235919952,
+ "learning_rate": 0.0006,
+ "loss": 4.703792572021484,
+ "step": 3894
+ },
+ {
+ "epoch": 54.09781659388646,
+ "grad_norm": 0.02431778982281685,
+ "learning_rate": 0.0006,
+ "loss": 4.658116340637207,
+ "step": 3895
+ },
+ {
+ "epoch": 54.1117903930131,
+ "grad_norm": 0.02792172320187092,
+ "learning_rate": 0.0006,
+ "loss": 4.638895034790039,
+ "step": 3896
+ },
+ {
+ "epoch": 54.12576419213974,
+ "grad_norm": 0.032253511250019073,
+ "learning_rate": 0.0006,
+ "loss": 4.693902492523193,
+ "step": 3897
+ },
+ {
+ "epoch": 54.13973799126637,
+ "grad_norm": 0.027044931426644325,
+ "learning_rate": 0.0006,
+ "loss": 4.575805187225342,
+ "step": 3898
+ },
+ {
+ "epoch": 54.15371179039301,
+ "grad_norm": 0.023270361125469208,
+ "learning_rate": 0.0006,
+ "loss": 4.67645263671875,
+ "step": 3899
+ },
+ {
+ "epoch": 54.16768558951965,
+ "grad_norm": 0.01929079182446003,
+ "learning_rate": 0.0006,
+ "loss": 4.68062686920166,
+ "step": 3900
+ },
+ {
+ "epoch": 54.18165938864629,
+ "grad_norm": 0.020343473181128502,
+ "learning_rate": 0.0006,
+ "loss": 4.708086967468262,
+ "step": 3901
+ },
+ {
+ "epoch": 54.19563318777293,
+ "grad_norm": 0.0211793202906847,
+ "learning_rate": 0.0006,
+ "loss": 4.677323341369629,
+ "step": 3902
+ },
+ {
+ "epoch": 54.209606986899566,
+ "grad_norm": 0.020341217517852783,
+ "learning_rate": 0.0006,
+ "loss": 4.760937690734863,
+ "step": 3903
+ },
+ {
+ "epoch": 54.223580786026204,
+ "grad_norm": 0.019421054050326347,
+ "learning_rate": 0.0006,
+ "loss": 4.508834362030029,
+ "step": 3904
+ },
+ {
+ "epoch": 54.237554585152836,
+ "grad_norm": 0.016701679676771164,
+ "learning_rate": 0.0006,
+ "loss": 4.664399147033691,
+ "step": 3905
+ },
+ {
+ "epoch": 54.251528384279474,
+ "grad_norm": 0.016561470925807953,
+ "learning_rate": 0.0006,
+ "loss": 4.742990493774414,
+ "step": 3906
+ },
+ {
+ "epoch": 54.26550218340611,
+ "grad_norm": 0.017446085810661316,
+ "learning_rate": 0.0006,
+ "loss": 4.63038444519043,
+ "step": 3907
+ },
+ {
+ "epoch": 54.27947598253275,
+ "grad_norm": 0.014340086840093136,
+ "learning_rate": 0.0006,
+ "loss": 4.688791275024414,
+ "step": 3908
+ },
+ {
+ "epoch": 54.29344978165939,
+ "grad_norm": 0.013580698519945145,
+ "learning_rate": 0.0006,
+ "loss": 4.625302314758301,
+ "step": 3909
+ },
+ {
+ "epoch": 54.30742358078603,
+ "grad_norm": 0.014931539073586464,
+ "learning_rate": 0.0006,
+ "loss": 4.6167497634887695,
+ "step": 3910
+ },
+ {
+ "epoch": 54.32139737991266,
+ "grad_norm": 0.013492150232195854,
+ "learning_rate": 0.0006,
+ "loss": 4.657521724700928,
+ "step": 3911
+ },
+ {
+ "epoch": 54.3353711790393,
+ "grad_norm": 0.015151413157582283,
+ "learning_rate": 0.0006,
+ "loss": 4.634521484375,
+ "step": 3912
+ },
+ {
+ "epoch": 54.34934497816594,
+ "grad_norm": 0.01302040833979845,
+ "learning_rate": 0.0006,
+ "loss": 4.718928337097168,
+ "step": 3913
+ },
+ {
+ "epoch": 54.36331877729258,
+ "grad_norm": 0.013433529995381832,
+ "learning_rate": 0.0006,
+ "loss": 4.584803581237793,
+ "step": 3914
+ },
+ {
+ "epoch": 54.377292576419215,
+ "grad_norm": 0.012987039051949978,
+ "learning_rate": 0.0006,
+ "loss": 4.6393585205078125,
+ "step": 3915
+ },
+ {
+ "epoch": 54.391266375545854,
+ "grad_norm": 0.013695075176656246,
+ "learning_rate": 0.0006,
+ "loss": 4.556313514709473,
+ "step": 3916
+ },
+ {
+ "epoch": 54.40524017467249,
+ "grad_norm": 0.013718255795538425,
+ "learning_rate": 0.0006,
+ "loss": 4.576223850250244,
+ "step": 3917
+ },
+ {
+ "epoch": 54.419213973799124,
+ "grad_norm": 0.013699112460017204,
+ "learning_rate": 0.0006,
+ "loss": 4.63889217376709,
+ "step": 3918
+ },
+ {
+ "epoch": 54.43318777292576,
+ "grad_norm": 0.013516264036297798,
+ "learning_rate": 0.0006,
+ "loss": 4.669532299041748,
+ "step": 3919
+ },
+ {
+ "epoch": 54.4471615720524,
+ "grad_norm": 0.013597294688224792,
+ "learning_rate": 0.0006,
+ "loss": 4.6644287109375,
+ "step": 3920
+ },
+ {
+ "epoch": 54.46113537117904,
+ "grad_norm": 0.013185849413275719,
+ "learning_rate": 0.0006,
+ "loss": 4.636900901794434,
+ "step": 3921
+ },
+ {
+ "epoch": 54.47510917030568,
+ "grad_norm": 0.014311990700662136,
+ "learning_rate": 0.0006,
+ "loss": 4.6047587394714355,
+ "step": 3922
+ },
+ {
+ "epoch": 54.48908296943232,
+ "grad_norm": 0.012426027096807957,
+ "learning_rate": 0.0006,
+ "loss": 4.6161394119262695,
+ "step": 3923
+ },
+ {
+ "epoch": 54.50305676855895,
+ "grad_norm": 0.01283493172377348,
+ "learning_rate": 0.0006,
+ "loss": 4.584190368652344,
+ "step": 3924
+ },
+ {
+ "epoch": 54.51703056768559,
+ "grad_norm": 0.01339943427592516,
+ "learning_rate": 0.0006,
+ "loss": 4.624184608459473,
+ "step": 3925
+ },
+ {
+ "epoch": 54.531004366812226,
+ "grad_norm": 0.012859724462032318,
+ "learning_rate": 0.0006,
+ "loss": 4.6174635887146,
+ "step": 3926
+ },
+ {
+ "epoch": 54.544978165938865,
+ "grad_norm": 0.014420023187994957,
+ "learning_rate": 0.0006,
+ "loss": 4.536544322967529,
+ "step": 3927
+ },
+ {
+ "epoch": 54.5589519650655,
+ "grad_norm": 0.014237185008823872,
+ "learning_rate": 0.0006,
+ "loss": 4.681514739990234,
+ "step": 3928
+ },
+ {
+ "epoch": 54.57292576419214,
+ "grad_norm": 0.013340328820049763,
+ "learning_rate": 0.0006,
+ "loss": 4.579689979553223,
+ "step": 3929
+ },
+ {
+ "epoch": 54.58689956331878,
+ "grad_norm": 0.013894530944526196,
+ "learning_rate": 0.0006,
+ "loss": 4.638958930969238,
+ "step": 3930
+ },
+ {
+ "epoch": 54.60087336244541,
+ "grad_norm": 0.016181141138076782,
+ "learning_rate": 0.0006,
+ "loss": 4.573508262634277,
+ "step": 3931
+ },
+ {
+ "epoch": 54.61484716157205,
+ "grad_norm": 0.013221992179751396,
+ "learning_rate": 0.0006,
+ "loss": 4.6229047775268555,
+ "step": 3932
+ },
+ {
+ "epoch": 54.62882096069869,
+ "grad_norm": 0.013241561129689217,
+ "learning_rate": 0.0006,
+ "loss": 4.641152858734131,
+ "step": 3933
+ },
+ {
+ "epoch": 54.64279475982533,
+ "grad_norm": 0.016237538307905197,
+ "learning_rate": 0.0006,
+ "loss": 4.6421217918396,
+ "step": 3934
+ },
+ {
+ "epoch": 54.65676855895197,
+ "grad_norm": 0.020866507664322853,
+ "learning_rate": 0.0006,
+ "loss": 4.646058082580566,
+ "step": 3935
+ },
+ {
+ "epoch": 54.670742358078606,
+ "grad_norm": 0.022182893007993698,
+ "learning_rate": 0.0006,
+ "loss": 4.554501533508301,
+ "step": 3936
+ },
+ {
+ "epoch": 54.68471615720524,
+ "grad_norm": 0.02254868485033512,
+ "learning_rate": 0.0006,
+ "loss": 4.563869953155518,
+ "step": 3937
+ },
+ {
+ "epoch": 54.698689956331876,
+ "grad_norm": 0.019730940461158752,
+ "learning_rate": 0.0006,
+ "loss": 4.592165946960449,
+ "step": 3938
+ },
+ {
+ "epoch": 54.712663755458514,
+ "grad_norm": 0.015146561898291111,
+ "learning_rate": 0.0006,
+ "loss": 4.6228742599487305,
+ "step": 3939
+ },
+ {
+ "epoch": 54.72663755458515,
+ "grad_norm": 0.01824716106057167,
+ "learning_rate": 0.0006,
+ "loss": 4.716236114501953,
+ "step": 3940
+ },
+ {
+ "epoch": 54.74061135371179,
+ "grad_norm": 0.017730550840497017,
+ "learning_rate": 0.0006,
+ "loss": 4.650232315063477,
+ "step": 3941
+ },
+ {
+ "epoch": 54.75458515283843,
+ "grad_norm": 0.014875595457851887,
+ "learning_rate": 0.0006,
+ "loss": 4.6377458572387695,
+ "step": 3942
+ },
+ {
+ "epoch": 54.76855895196506,
+ "grad_norm": 0.01613209955394268,
+ "learning_rate": 0.0006,
+ "loss": 4.5768537521362305,
+ "step": 3943
+ },
+ {
+ "epoch": 54.7825327510917,
+ "grad_norm": 0.017311280593276024,
+ "learning_rate": 0.0006,
+ "loss": 4.625744342803955,
+ "step": 3944
+ },
+ {
+ "epoch": 54.79650655021834,
+ "grad_norm": 0.018770085647702217,
+ "learning_rate": 0.0006,
+ "loss": 4.613787651062012,
+ "step": 3945
+ },
+ {
+ "epoch": 54.81048034934498,
+ "grad_norm": 0.0165257528424263,
+ "learning_rate": 0.0006,
+ "loss": 4.5991315841674805,
+ "step": 3946
+ },
+ {
+ "epoch": 54.82445414847162,
+ "grad_norm": 0.017433857545256615,
+ "learning_rate": 0.0006,
+ "loss": 4.513872146606445,
+ "step": 3947
+ },
+ {
+ "epoch": 54.838427947598255,
+ "grad_norm": 0.020126905292272568,
+ "learning_rate": 0.0006,
+ "loss": 4.616185188293457,
+ "step": 3948
+ },
+ {
+ "epoch": 54.852401746724894,
+ "grad_norm": 0.01819358579814434,
+ "learning_rate": 0.0006,
+ "loss": 4.535691261291504,
+ "step": 3949
+ },
+ {
+ "epoch": 54.866375545851525,
+ "grad_norm": 0.01899927295744419,
+ "learning_rate": 0.0006,
+ "loss": 4.595199108123779,
+ "step": 3950
+ },
+ {
+ "epoch": 54.880349344978164,
+ "grad_norm": 0.02005566656589508,
+ "learning_rate": 0.0006,
+ "loss": 4.633289337158203,
+ "step": 3951
+ },
+ {
+ "epoch": 54.8943231441048,
+ "grad_norm": 0.021570220589637756,
+ "learning_rate": 0.0006,
+ "loss": 4.592503547668457,
+ "step": 3952
+ },
+ {
+ "epoch": 54.90829694323144,
+ "grad_norm": 0.01883506216108799,
+ "learning_rate": 0.0006,
+ "loss": 4.6597795486450195,
+ "step": 3953
+ },
+ {
+ "epoch": 54.92227074235808,
+ "grad_norm": 0.018070021644234657,
+ "learning_rate": 0.0006,
+ "loss": 4.601474285125732,
+ "step": 3954
+ },
+ {
+ "epoch": 54.93624454148472,
+ "grad_norm": 0.018090378493070602,
+ "learning_rate": 0.0006,
+ "loss": 4.569846153259277,
+ "step": 3955
+ },
+ {
+ "epoch": 54.95021834061135,
+ "grad_norm": 0.018751563504338264,
+ "learning_rate": 0.0006,
+ "loss": 4.646064281463623,
+ "step": 3956
+ },
+ {
+ "epoch": 54.96419213973799,
+ "grad_norm": 0.017512254416942596,
+ "learning_rate": 0.0006,
+ "loss": 4.606670379638672,
+ "step": 3957
+ },
+ {
+ "epoch": 54.97816593886463,
+ "grad_norm": 0.018771110102534294,
+ "learning_rate": 0.0006,
+ "loss": 4.604315757751465,
+ "step": 3958
+ },
+ {
+ "epoch": 54.992139737991266,
+ "grad_norm": 0.01611141301691532,
+ "learning_rate": 0.0006,
+ "loss": 4.643245220184326,
+ "step": 3959
+ },
+ {
+ "epoch": 55.0,
+ "grad_norm": 0.0156728383153677,
+ "learning_rate": 0.0006,
+ "loss": 4.6739044189453125,
+ "step": 3960
+ },
+ {
+ "epoch": 55.0,
+ "eval_loss": 4.851032257080078,
+ "eval_runtime": 56.8699,
+ "eval_samples_per_second": 42.94,
+ "eval_steps_per_second": 1.354,
+ "step": 3960
+ },
+ {
+ "epoch": 55.01397379912664,
+ "grad_norm": 0.016214672476053238,
+ "learning_rate": 0.0006,
+ "loss": 4.571749687194824,
+ "step": 3961
+ },
+ {
+ "epoch": 55.02794759825328,
+ "grad_norm": 0.01649155467748642,
+ "learning_rate": 0.0006,
+ "loss": 4.553592205047607,
+ "step": 3962
+ },
+ {
+ "epoch": 55.041921397379916,
+ "grad_norm": 0.016793379560112953,
+ "learning_rate": 0.0006,
+ "loss": 4.650628566741943,
+ "step": 3963
+ },
+ {
+ "epoch": 55.05589519650655,
+ "grad_norm": 0.0191937405616045,
+ "learning_rate": 0.0006,
+ "loss": 4.620532989501953,
+ "step": 3964
+ },
+ {
+ "epoch": 55.069868995633186,
+ "grad_norm": 0.020976556465029716,
+ "learning_rate": 0.0006,
+ "loss": 4.6329545974731445,
+ "step": 3965
+ },
+ {
+ "epoch": 55.083842794759825,
+ "grad_norm": 0.023450477048754692,
+ "learning_rate": 0.0006,
+ "loss": 4.60828971862793,
+ "step": 3966
+ },
+ {
+ "epoch": 55.09781659388646,
+ "grad_norm": 0.026778312399983406,
+ "learning_rate": 0.0006,
+ "loss": 4.617481708526611,
+ "step": 3967
+ },
+ {
+ "epoch": 55.1117903930131,
+ "grad_norm": 0.02333568036556244,
+ "learning_rate": 0.0006,
+ "loss": 4.580722332000732,
+ "step": 3968
+ },
+ {
+ "epoch": 55.12576419213974,
+ "grad_norm": 0.018491310998797417,
+ "learning_rate": 0.0006,
+ "loss": 4.590953350067139,
+ "step": 3969
+ },
+ {
+ "epoch": 55.13973799126637,
+ "grad_norm": 0.020339803770184517,
+ "learning_rate": 0.0006,
+ "loss": 4.518374443054199,
+ "step": 3970
+ },
+ {
+ "epoch": 55.15371179039301,
+ "grad_norm": 0.01759706623852253,
+ "learning_rate": 0.0006,
+ "loss": 4.560522079467773,
+ "step": 3971
+ },
+ {
+ "epoch": 55.16768558951965,
+ "grad_norm": 0.015875980257987976,
+ "learning_rate": 0.0006,
+ "loss": 4.4795708656311035,
+ "step": 3972
+ },
+ {
+ "epoch": 55.18165938864629,
+ "grad_norm": 0.01650134287774563,
+ "learning_rate": 0.0006,
+ "loss": 4.57435941696167,
+ "step": 3973
+ },
+ {
+ "epoch": 55.19563318777293,
+ "grad_norm": 0.01734331250190735,
+ "learning_rate": 0.0006,
+ "loss": 4.661007881164551,
+ "step": 3974
+ },
+ {
+ "epoch": 55.209606986899566,
+ "grad_norm": 0.01651303470134735,
+ "learning_rate": 0.0006,
+ "loss": 4.511148452758789,
+ "step": 3975
+ },
+ {
+ "epoch": 55.223580786026204,
+ "grad_norm": 0.016189761459827423,
+ "learning_rate": 0.0006,
+ "loss": 4.592350006103516,
+ "step": 3976
+ },
+ {
+ "epoch": 55.237554585152836,
+ "grad_norm": 0.01833222061395645,
+ "learning_rate": 0.0006,
+ "loss": 4.506649971008301,
+ "step": 3977
+ },
+ {
+ "epoch": 55.251528384279474,
+ "grad_norm": 0.02433410845696926,
+ "learning_rate": 0.0006,
+ "loss": 4.49500036239624,
+ "step": 3978
+ },
+ {
+ "epoch": 55.26550218340611,
+ "grad_norm": 0.0314289927482605,
+ "learning_rate": 0.0006,
+ "loss": 4.507624626159668,
+ "step": 3979
+ },
+ {
+ "epoch": 55.27947598253275,
+ "grad_norm": 0.03013679012656212,
+ "learning_rate": 0.0006,
+ "loss": 4.533641815185547,
+ "step": 3980
+ },
+ {
+ "epoch": 55.29344978165939,
+ "grad_norm": 0.024524061009287834,
+ "learning_rate": 0.0006,
+ "loss": 4.5721116065979,
+ "step": 3981
+ },
+ {
+ "epoch": 55.30742358078603,
+ "grad_norm": 0.02640336938202381,
+ "learning_rate": 0.0006,
+ "loss": 4.652521133422852,
+ "step": 3982
+ },
+ {
+ "epoch": 55.32139737991266,
+ "grad_norm": 0.021309101954102516,
+ "learning_rate": 0.0006,
+ "loss": 4.600375175476074,
+ "step": 3983
+ },
+ {
+ "epoch": 55.3353711790393,
+ "grad_norm": 0.023008549585938454,
+ "learning_rate": 0.0006,
+ "loss": 4.53914737701416,
+ "step": 3984
+ },
+ {
+ "epoch": 55.34934497816594,
+ "grad_norm": 0.022491468116641045,
+ "learning_rate": 0.0006,
+ "loss": 4.661163330078125,
+ "step": 3985
+ },
+ {
+ "epoch": 55.36331877729258,
+ "grad_norm": 0.01904081553220749,
+ "learning_rate": 0.0006,
+ "loss": 4.5426177978515625,
+ "step": 3986
+ },
+ {
+ "epoch": 55.377292576419215,
+ "grad_norm": 0.019229518249630928,
+ "learning_rate": 0.0006,
+ "loss": 4.606066703796387,
+ "step": 3987
+ },
+ {
+ "epoch": 55.391266375545854,
+ "grad_norm": 0.01781383901834488,
+ "learning_rate": 0.0006,
+ "loss": 4.54642391204834,
+ "step": 3988
+ },
+ {
+ "epoch": 55.40524017467249,
+ "grad_norm": 0.0174600500613451,
+ "learning_rate": 0.0006,
+ "loss": 4.711891174316406,
+ "step": 3989
+ },
+ {
+ "epoch": 55.419213973799124,
+ "grad_norm": 0.018154339864850044,
+ "learning_rate": 0.0006,
+ "loss": 4.527122974395752,
+ "step": 3990
+ },
+ {
+ "epoch": 55.43318777292576,
+ "grad_norm": 0.01914648339152336,
+ "learning_rate": 0.0006,
+ "loss": 4.641975402832031,
+ "step": 3991
+ },
+ {
+ "epoch": 55.4471615720524,
+ "grad_norm": 0.017118625342845917,
+ "learning_rate": 0.0006,
+ "loss": 4.511425971984863,
+ "step": 3992
+ },
+ {
+ "epoch": 55.46113537117904,
+ "grad_norm": 0.01551489345729351,
+ "learning_rate": 0.0006,
+ "loss": 4.551141738891602,
+ "step": 3993
+ },
+ {
+ "epoch": 55.47510917030568,
+ "grad_norm": 0.015072065405547619,
+ "learning_rate": 0.0006,
+ "loss": 4.651115894317627,
+ "step": 3994
+ },
+ {
+ "epoch": 55.48908296943232,
+ "grad_norm": 0.017010986804962158,
+ "learning_rate": 0.0006,
+ "loss": 4.583282470703125,
+ "step": 3995
+ },
+ {
+ "epoch": 55.50305676855895,
+ "grad_norm": 0.01764831133186817,
+ "learning_rate": 0.0006,
+ "loss": 4.593364238739014,
+ "step": 3996
+ },
+ {
+ "epoch": 55.51703056768559,
+ "grad_norm": 0.01762884296476841,
+ "learning_rate": 0.0006,
+ "loss": 4.45671272277832,
+ "step": 3997
+ },
+ {
+ "epoch": 55.531004366812226,
+ "grad_norm": 0.01600457727909088,
+ "learning_rate": 0.0006,
+ "loss": 4.511260032653809,
+ "step": 3998
+ },
+ {
+ "epoch": 55.544978165938865,
+ "grad_norm": 0.015706021338701248,
+ "learning_rate": 0.0006,
+ "loss": 4.47593355178833,
+ "step": 3999
+ },
+ {
+ "epoch": 55.5589519650655,
+ "grad_norm": 0.017718996852636337,
+ "learning_rate": 0.0006,
+ "loss": 4.610543251037598,
+ "step": 4000
+ },
+ {
+ "epoch": 55.57292576419214,
+ "grad_norm": 0.015073757618665695,
+ "learning_rate": 0.0006,
+ "loss": 4.466620445251465,
+ "step": 4001
+ },
+ {
+ "epoch": 55.58689956331878,
+ "grad_norm": 0.01789247989654541,
+ "learning_rate": 0.0006,
+ "loss": 4.503409385681152,
+ "step": 4002
+ },
+ {
+ "epoch": 55.60087336244541,
+ "grad_norm": 0.022329598665237427,
+ "learning_rate": 0.0006,
+ "loss": 4.62099552154541,
+ "step": 4003
+ },
+ {
+ "epoch": 55.61484716157205,
+ "grad_norm": 0.02343829534947872,
+ "learning_rate": 0.0006,
+ "loss": 4.620813369750977,
+ "step": 4004
+ },
+ {
+ "epoch": 55.62882096069869,
+ "grad_norm": 0.023279687389731407,
+ "learning_rate": 0.0006,
+ "loss": 4.545765399932861,
+ "step": 4005
+ },
+ {
+ "epoch": 55.64279475982533,
+ "grad_norm": 0.020817464217543602,
+ "learning_rate": 0.0006,
+ "loss": 4.561470031738281,
+ "step": 4006
+ },
+ {
+ "epoch": 55.65676855895197,
+ "grad_norm": 0.020270058885216713,
+ "learning_rate": 0.0006,
+ "loss": 4.5875020027160645,
+ "step": 4007
+ },
+ {
+ "epoch": 55.670742358078606,
+ "grad_norm": 0.021085843443870544,
+ "learning_rate": 0.0006,
+ "loss": 4.504560470581055,
+ "step": 4008
+ },
+ {
+ "epoch": 55.68471615720524,
+ "grad_norm": 0.018507730215787888,
+ "learning_rate": 0.0006,
+ "loss": 4.520644664764404,
+ "step": 4009
+ },
+ {
+ "epoch": 55.698689956331876,
+ "grad_norm": 0.020571516826748848,
+ "learning_rate": 0.0006,
+ "loss": 4.539643287658691,
+ "step": 4010
+ },
+ {
+ "epoch": 55.712663755458514,
+ "grad_norm": 0.021484510973095894,
+ "learning_rate": 0.0006,
+ "loss": 4.628833293914795,
+ "step": 4011
+ },
+ {
+ "epoch": 55.72663755458515,
+ "grad_norm": 0.017693819478154182,
+ "learning_rate": 0.0006,
+ "loss": 4.6023783683776855,
+ "step": 4012
+ },
+ {
+ "epoch": 55.74061135371179,
+ "grad_norm": 0.020186059176921844,
+ "learning_rate": 0.0006,
+ "loss": 4.634884834289551,
+ "step": 4013
+ },
+ {
+ "epoch": 55.75458515283843,
+ "grad_norm": 0.01835780404508114,
+ "learning_rate": 0.0006,
+ "loss": 4.577791213989258,
+ "step": 4014
+ },
+ {
+ "epoch": 55.76855895196506,
+ "grad_norm": 0.018162988126277924,
+ "learning_rate": 0.0006,
+ "loss": 4.641597270965576,
+ "step": 4015
+ },
+ {
+ "epoch": 55.7825327510917,
+ "grad_norm": 0.018909504637122154,
+ "learning_rate": 0.0006,
+ "loss": 4.487447738647461,
+ "step": 4016
+ },
+ {
+ "epoch": 55.79650655021834,
+ "grad_norm": 0.018972251564264297,
+ "learning_rate": 0.0006,
+ "loss": 4.590317726135254,
+ "step": 4017
+ },
+ {
+ "epoch": 55.81048034934498,
+ "grad_norm": 0.01793607883155346,
+ "learning_rate": 0.0006,
+ "loss": 4.4994611740112305,
+ "step": 4018
+ },
+ {
+ "epoch": 55.82445414847162,
+ "grad_norm": 0.02123514749109745,
+ "learning_rate": 0.0006,
+ "loss": 4.479006767272949,
+ "step": 4019
+ },
+ {
+ "epoch": 55.838427947598255,
+ "grad_norm": 0.021246599033474922,
+ "learning_rate": 0.0006,
+ "loss": 4.4738640785217285,
+ "step": 4020
+ },
+ {
+ "epoch": 55.852401746724894,
+ "grad_norm": 0.020131602883338928,
+ "learning_rate": 0.0006,
+ "loss": 4.600308418273926,
+ "step": 4021
+ },
+ {
+ "epoch": 55.866375545851525,
+ "grad_norm": 0.019159460440278053,
+ "learning_rate": 0.0006,
+ "loss": 4.540908336639404,
+ "step": 4022
+ },
+ {
+ "epoch": 55.880349344978164,
+ "grad_norm": 0.023114705458283424,
+ "learning_rate": 0.0006,
+ "loss": 4.641247749328613,
+ "step": 4023
+ },
+ {
+ "epoch": 55.8943231441048,
+ "grad_norm": 0.02296135015785694,
+ "learning_rate": 0.0006,
+ "loss": 4.444624423980713,
+ "step": 4024
+ },
+ {
+ "epoch": 55.90829694323144,
+ "grad_norm": 0.02204655110836029,
+ "learning_rate": 0.0006,
+ "loss": 4.506349563598633,
+ "step": 4025
+ },
+ {
+ "epoch": 55.92227074235808,
+ "grad_norm": 0.019338225945830345,
+ "learning_rate": 0.0006,
+ "loss": 4.616047382354736,
+ "step": 4026
+ },
+ {
+ "epoch": 55.93624454148472,
+ "grad_norm": 0.020357884466648102,
+ "learning_rate": 0.0006,
+ "loss": 4.48170280456543,
+ "step": 4027
+ },
+ {
+ "epoch": 55.95021834061135,
+ "grad_norm": 0.02426772564649582,
+ "learning_rate": 0.0006,
+ "loss": 4.576877117156982,
+ "step": 4028
+ },
+ {
+ "epoch": 55.96419213973799,
+ "grad_norm": 0.023351816460490227,
+ "learning_rate": 0.0006,
+ "loss": 4.458885192871094,
+ "step": 4029
+ },
+ {
+ "epoch": 55.97816593886463,
+ "grad_norm": 0.019892243668437004,
+ "learning_rate": 0.0006,
+ "loss": 4.520229816436768,
+ "step": 4030
+ },
+ {
+ "epoch": 55.992139737991266,
+ "grad_norm": 0.018331089988350868,
+ "learning_rate": 0.0006,
+ "loss": 4.432201862335205,
+ "step": 4031
+ },
+ {
+ "epoch": 56.0,
+ "grad_norm": 0.021300828084349632,
+ "learning_rate": 0.0006,
+ "loss": 4.617072582244873,
+ "step": 4032
+ },
+ {
+ "epoch": 56.0,
+ "eval_loss": 4.811229705810547,
+ "eval_runtime": 56.4846,
+ "eval_samples_per_second": 43.233,
+ "eval_steps_per_second": 1.363,
+ "step": 4032
+ },
+ {
+ "epoch": 56.01397379912664,
+ "grad_norm": 0.02199401892721653,
+ "learning_rate": 0.0006,
+ "loss": 4.513965606689453,
+ "step": 4033
+ },
+ {
+ "epoch": 56.02794759825328,
+ "grad_norm": 0.022109858691692352,
+ "learning_rate": 0.0006,
+ "loss": 4.557568550109863,
+ "step": 4034
+ },
+ {
+ "epoch": 56.041921397379916,
+ "grad_norm": 0.020103694871068,
+ "learning_rate": 0.0006,
+ "loss": 4.569512844085693,
+ "step": 4035
+ },
+ {
+ "epoch": 56.05589519650655,
+ "grad_norm": 0.01898941583931446,
+ "learning_rate": 0.0006,
+ "loss": 4.623138904571533,
+ "step": 4036
+ },
+ {
+ "epoch": 56.069868995633186,
+ "grad_norm": 0.02227010577917099,
+ "learning_rate": 0.0006,
+ "loss": 4.433316230773926,
+ "step": 4037
+ },
+ {
+ "epoch": 56.083842794759825,
+ "grad_norm": 0.024683045223355293,
+ "learning_rate": 0.0006,
+ "loss": 4.458296775817871,
+ "step": 4038
+ },
+ {
+ "epoch": 56.09781659388646,
+ "grad_norm": 0.02900184690952301,
+ "learning_rate": 0.0006,
+ "loss": 4.531060218811035,
+ "step": 4039
+ },
+ {
+ "epoch": 56.1117903930131,
+ "grad_norm": 0.025557566434144974,
+ "learning_rate": 0.0006,
+ "loss": 4.56685733795166,
+ "step": 4040
+ },
+ {
+ "epoch": 56.12576419213974,
+ "grad_norm": 0.021414656192064285,
+ "learning_rate": 0.0006,
+ "loss": 4.469037055969238,
+ "step": 4041
+ },
+ {
+ "epoch": 56.13973799126637,
+ "grad_norm": 0.023262616246938705,
+ "learning_rate": 0.0006,
+ "loss": 4.493876934051514,
+ "step": 4042
+ },
+ {
+ "epoch": 56.15371179039301,
+ "grad_norm": 0.023011326789855957,
+ "learning_rate": 0.0006,
+ "loss": 4.4572906494140625,
+ "step": 4043
+ },
+ {
+ "epoch": 56.16768558951965,
+ "grad_norm": 0.02177615463733673,
+ "learning_rate": 0.0006,
+ "loss": 4.615389823913574,
+ "step": 4044
+ },
+ {
+ "epoch": 56.18165938864629,
+ "grad_norm": 0.021527713164687157,
+ "learning_rate": 0.0006,
+ "loss": 4.613762855529785,
+ "step": 4045
+ },
+ {
+ "epoch": 56.19563318777293,
+ "grad_norm": 0.01982325315475464,
+ "learning_rate": 0.0006,
+ "loss": 4.53670597076416,
+ "step": 4046
+ },
+ {
+ "epoch": 56.209606986899566,
+ "grad_norm": 0.022484654560685158,
+ "learning_rate": 0.0006,
+ "loss": 4.605565071105957,
+ "step": 4047
+ },
+ {
+ "epoch": 56.223580786026204,
+ "grad_norm": 0.02346140146255493,
+ "learning_rate": 0.0006,
+ "loss": 4.6265950202941895,
+ "step": 4048
+ },
+ {
+ "epoch": 56.237554585152836,
+ "grad_norm": 0.01876714453101158,
+ "learning_rate": 0.0006,
+ "loss": 4.481525421142578,
+ "step": 4049
+ },
+ {
+ "epoch": 56.251528384279474,
+ "grad_norm": 0.01926851086318493,
+ "learning_rate": 0.0006,
+ "loss": 4.630258083343506,
+ "step": 4050
+ },
+ {
+ "epoch": 56.26550218340611,
+ "grad_norm": 0.020858589559793472,
+ "learning_rate": 0.0006,
+ "loss": 4.5265913009643555,
+ "step": 4051
+ },
+ {
+ "epoch": 56.27947598253275,
+ "grad_norm": 0.01889280416071415,
+ "learning_rate": 0.0006,
+ "loss": 4.581275463104248,
+ "step": 4052
+ },
+ {
+ "epoch": 56.29344978165939,
+ "grad_norm": 0.01801002025604248,
+ "learning_rate": 0.0006,
+ "loss": 4.498557090759277,
+ "step": 4053
+ },
+ {
+ "epoch": 56.30742358078603,
+ "grad_norm": 0.02252180315554142,
+ "learning_rate": 0.0006,
+ "loss": 4.565267086029053,
+ "step": 4054
+ },
+ {
+ "epoch": 56.32139737991266,
+ "grad_norm": 0.02306070365011692,
+ "learning_rate": 0.0006,
+ "loss": 4.3591389656066895,
+ "step": 4055
+ },
+ {
+ "epoch": 56.3353711790393,
+ "grad_norm": 0.020292513072490692,
+ "learning_rate": 0.0006,
+ "loss": 4.407517433166504,
+ "step": 4056
+ },
+ {
+ "epoch": 56.34934497816594,
+ "grad_norm": 0.019154123961925507,
+ "learning_rate": 0.0006,
+ "loss": 4.482261657714844,
+ "step": 4057
+ },
+ {
+ "epoch": 56.36331877729258,
+ "grad_norm": 0.017338937148451805,
+ "learning_rate": 0.0006,
+ "loss": 4.4463582038879395,
+ "step": 4058
+ },
+ {
+ "epoch": 56.377292576419215,
+ "grad_norm": 0.01730559952557087,
+ "learning_rate": 0.0006,
+ "loss": 4.542474746704102,
+ "step": 4059
+ },
+ {
+ "epoch": 56.391266375545854,
+ "grad_norm": 0.017029426991939545,
+ "learning_rate": 0.0006,
+ "loss": 4.552488327026367,
+ "step": 4060
+ },
+ {
+ "epoch": 56.40524017467249,
+ "grad_norm": 0.01745498552918434,
+ "learning_rate": 0.0006,
+ "loss": 4.52928352355957,
+ "step": 4061
+ },
+ {
+ "epoch": 56.419213973799124,
+ "grad_norm": 0.0168515145778656,
+ "learning_rate": 0.0006,
+ "loss": 4.4516801834106445,
+ "step": 4062
+ },
+ {
+ "epoch": 56.43318777292576,
+ "grad_norm": 0.018545417115092278,
+ "learning_rate": 0.0006,
+ "loss": 4.569811820983887,
+ "step": 4063
+ },
+ {
+ "epoch": 56.4471615720524,
+ "grad_norm": 0.02227838523685932,
+ "learning_rate": 0.0006,
+ "loss": 4.412224769592285,
+ "step": 4064
+ },
+ {
+ "epoch": 56.46113537117904,
+ "grad_norm": 0.023338504135608673,
+ "learning_rate": 0.0006,
+ "loss": 4.490184307098389,
+ "step": 4065
+ },
+ {
+ "epoch": 56.47510917030568,
+ "grad_norm": 0.02257535606622696,
+ "learning_rate": 0.0006,
+ "loss": 4.479152679443359,
+ "step": 4066
+ },
+ {
+ "epoch": 56.48908296943232,
+ "grad_norm": 0.022737598046660423,
+ "learning_rate": 0.0006,
+ "loss": 4.418235778808594,
+ "step": 4067
+ },
+ {
+ "epoch": 56.50305676855895,
+ "grad_norm": 0.024603573605418205,
+ "learning_rate": 0.0006,
+ "loss": 4.45781135559082,
+ "step": 4068
+ },
+ {
+ "epoch": 56.51703056768559,
+ "grad_norm": 0.024953778833150864,
+ "learning_rate": 0.0006,
+ "loss": 4.593907833099365,
+ "step": 4069
+ },
+ {
+ "epoch": 56.531004366812226,
+ "grad_norm": 0.02195645309984684,
+ "learning_rate": 0.0006,
+ "loss": 4.467643737792969,
+ "step": 4070
+ },
+ {
+ "epoch": 56.544978165938865,
+ "grad_norm": 0.017126578837633133,
+ "learning_rate": 0.0006,
+ "loss": 4.555484771728516,
+ "step": 4071
+ },
+ {
+ "epoch": 56.5589519650655,
+ "grad_norm": 0.019482605159282684,
+ "learning_rate": 0.0006,
+ "loss": 4.430915832519531,
+ "step": 4072
+ },
+ {
+ "epoch": 56.57292576419214,
+ "grad_norm": 0.0219392292201519,
+ "learning_rate": 0.0006,
+ "loss": 4.593365669250488,
+ "step": 4073
+ },
+ {
+ "epoch": 56.58689956331878,
+ "grad_norm": 0.02420145645737648,
+ "learning_rate": 0.0006,
+ "loss": 4.491467475891113,
+ "step": 4074
+ },
+ {
+ "epoch": 56.60087336244541,
+ "grad_norm": 0.021523723378777504,
+ "learning_rate": 0.0006,
+ "loss": 4.463634490966797,
+ "step": 4075
+ },
+ {
+ "epoch": 56.61484716157205,
+ "grad_norm": 0.020242048427462578,
+ "learning_rate": 0.0006,
+ "loss": 4.577968597412109,
+ "step": 4076
+ },
+ {
+ "epoch": 56.62882096069869,
+ "grad_norm": 0.017771173268556595,
+ "learning_rate": 0.0006,
+ "loss": 4.568269729614258,
+ "step": 4077
+ },
+ {
+ "epoch": 56.64279475982533,
+ "grad_norm": 0.01598544418811798,
+ "learning_rate": 0.0006,
+ "loss": 4.476841926574707,
+ "step": 4078
+ },
+ {
+ "epoch": 56.65676855895197,
+ "grad_norm": 0.0168446097522974,
+ "learning_rate": 0.0006,
+ "loss": 4.480965614318848,
+ "step": 4079
+ },
+ {
+ "epoch": 56.670742358078606,
+ "grad_norm": 0.014948660507798195,
+ "learning_rate": 0.0006,
+ "loss": 4.535400390625,
+ "step": 4080
+ },
+ {
+ "epoch": 56.68471615720524,
+ "grad_norm": 0.014899153262376785,
+ "learning_rate": 0.0006,
+ "loss": 4.476048946380615,
+ "step": 4081
+ },
+ {
+ "epoch": 56.698689956331876,
+ "grad_norm": 0.014051870442926884,
+ "learning_rate": 0.0006,
+ "loss": 4.533884048461914,
+ "step": 4082
+ },
+ {
+ "epoch": 56.712663755458514,
+ "grad_norm": 0.014708572067320347,
+ "learning_rate": 0.0006,
+ "loss": 4.586690902709961,
+ "step": 4083
+ },
+ {
+ "epoch": 56.72663755458515,
+ "grad_norm": 0.014797299169003963,
+ "learning_rate": 0.0006,
+ "loss": 4.446291446685791,
+ "step": 4084
+ },
+ {
+ "epoch": 56.74061135371179,
+ "grad_norm": 0.015256541781127453,
+ "learning_rate": 0.0006,
+ "loss": 4.553572654724121,
+ "step": 4085
+ },
+ {
+ "epoch": 56.75458515283843,
+ "grad_norm": 0.016042491421103477,
+ "learning_rate": 0.0006,
+ "loss": 4.507801055908203,
+ "step": 4086
+ },
+ {
+ "epoch": 56.76855895196506,
+ "grad_norm": 0.018835173919796944,
+ "learning_rate": 0.0006,
+ "loss": 4.393911838531494,
+ "step": 4087
+ },
+ {
+ "epoch": 56.7825327510917,
+ "grad_norm": 0.021994033828377724,
+ "learning_rate": 0.0006,
+ "loss": 4.543368339538574,
+ "step": 4088
+ },
+ {
+ "epoch": 56.79650655021834,
+ "grad_norm": 0.0230990182608366,
+ "learning_rate": 0.0006,
+ "loss": 4.507420539855957,
+ "step": 4089
+ },
+ {
+ "epoch": 56.81048034934498,
+ "grad_norm": 0.018818842247128487,
+ "learning_rate": 0.0006,
+ "loss": 4.499941349029541,
+ "step": 4090
+ },
+ {
+ "epoch": 56.82445414847162,
+ "grad_norm": 0.018257278949022293,
+ "learning_rate": 0.0006,
+ "loss": 4.380057334899902,
+ "step": 4091
+ },
+ {
+ "epoch": 56.838427947598255,
+ "grad_norm": 0.020073018968105316,
+ "learning_rate": 0.0006,
+ "loss": 4.572811603546143,
+ "step": 4092
+ },
+ {
+ "epoch": 56.852401746724894,
+ "grad_norm": 0.02078658528625965,
+ "learning_rate": 0.0006,
+ "loss": 4.446835517883301,
+ "step": 4093
+ },
+ {
+ "epoch": 56.866375545851525,
+ "grad_norm": 0.01793118566274643,
+ "learning_rate": 0.0006,
+ "loss": 4.517854690551758,
+ "step": 4094
+ },
+ {
+ "epoch": 56.880349344978164,
+ "grad_norm": 0.01608281210064888,
+ "learning_rate": 0.0006,
+ "loss": 4.506482124328613,
+ "step": 4095
+ },
+ {
+ "epoch": 56.8943231441048,
+ "grad_norm": 0.017405368387699127,
+ "learning_rate": 0.0006,
+ "loss": 4.439718246459961,
+ "step": 4096
+ },
+ {
+ "epoch": 56.90829694323144,
+ "grad_norm": 0.017930863425135612,
+ "learning_rate": 0.0006,
+ "loss": 4.409365653991699,
+ "step": 4097
+ },
+ {
+ "epoch": 56.92227074235808,
+ "grad_norm": 0.015757670626044273,
+ "learning_rate": 0.0006,
+ "loss": 4.505547523498535,
+ "step": 4098
+ },
+ {
+ "epoch": 56.93624454148472,
+ "grad_norm": 0.01580345816910267,
+ "learning_rate": 0.0006,
+ "loss": 4.462774753570557,
+ "step": 4099
+ },
+ {
+ "epoch": 56.95021834061135,
+ "grad_norm": 0.01659180410206318,
+ "learning_rate": 0.0006,
+ "loss": 4.49345588684082,
+ "step": 4100
+ },
+ {
+ "epoch": 56.96419213973799,
+ "grad_norm": 0.01877259835600853,
+ "learning_rate": 0.0006,
+ "loss": 4.574443340301514,
+ "step": 4101
+ },
+ {
+ "epoch": 56.97816593886463,
+ "grad_norm": 0.01977839693427086,
+ "learning_rate": 0.0006,
+ "loss": 4.574034690856934,
+ "step": 4102
+ },
+ {
+ "epoch": 56.992139737991266,
+ "grad_norm": 0.020627597346901894,
+ "learning_rate": 0.0006,
+ "loss": 4.50969123840332,
+ "step": 4103
+ },
+ {
+ "epoch": 57.0,
+ "grad_norm": 0.025128807872533798,
+ "learning_rate": 0.0006,
+ "loss": 4.401662826538086,
+ "step": 4104
+ },
+ {
+ "epoch": 57.0,
+ "eval_loss": 4.779252052307129,
+ "eval_runtime": 56.8104,
+ "eval_samples_per_second": 42.985,
+ "eval_steps_per_second": 1.355,
+ "step": 4104
+ },
+ {
+ "epoch": 57.01397379912664,
+ "grad_norm": 0.027712570503354073,
+ "learning_rate": 0.0006,
+ "loss": 4.472955703735352,
+ "step": 4105
+ },
+ {
+ "epoch": 57.02794759825328,
+ "grad_norm": 0.02164197526872158,
+ "learning_rate": 0.0006,
+ "loss": 4.387596130371094,
+ "step": 4106
+ },
+ {
+ "epoch": 57.041921397379916,
+ "grad_norm": 0.017972752451896667,
+ "learning_rate": 0.0006,
+ "loss": 4.4326276779174805,
+ "step": 4107
+ },
+ {
+ "epoch": 57.05589519650655,
+ "grad_norm": 0.020811019465327263,
+ "learning_rate": 0.0006,
+ "loss": 4.452951431274414,
+ "step": 4108
+ },
+ {
+ "epoch": 57.069868995633186,
+ "grad_norm": 0.019913259893655777,
+ "learning_rate": 0.0006,
+ "loss": 4.448301315307617,
+ "step": 4109
+ },
+ {
+ "epoch": 57.083842794759825,
+ "grad_norm": 0.017409102991223335,
+ "learning_rate": 0.0006,
+ "loss": 4.4108357429504395,
+ "step": 4110
+ },
+ {
+ "epoch": 57.09781659388646,
+ "grad_norm": 0.019374269992113113,
+ "learning_rate": 0.0006,
+ "loss": 4.398412227630615,
+ "step": 4111
+ },
+ {
+ "epoch": 57.1117903930131,
+ "grad_norm": 0.015521776862442493,
+ "learning_rate": 0.0006,
+ "loss": 4.508594989776611,
+ "step": 4112
+ },
+ {
+ "epoch": 57.12576419213974,
+ "grad_norm": 0.017201920971274376,
+ "learning_rate": 0.0006,
+ "loss": 4.449558734893799,
+ "step": 4113
+ },
+ {
+ "epoch": 57.13973799126637,
+ "grad_norm": 0.018062541261315346,
+ "learning_rate": 0.0006,
+ "loss": 4.517874240875244,
+ "step": 4114
+ },
+ {
+ "epoch": 57.15371179039301,
+ "grad_norm": 0.015846073627471924,
+ "learning_rate": 0.0006,
+ "loss": 4.427501678466797,
+ "step": 4115
+ },
+ {
+ "epoch": 57.16768558951965,
+ "grad_norm": 0.01849362626671791,
+ "learning_rate": 0.0006,
+ "loss": 4.433181285858154,
+ "step": 4116
+ },
+ {
+ "epoch": 57.18165938864629,
+ "grad_norm": 0.01948804222047329,
+ "learning_rate": 0.0006,
+ "loss": 4.446122646331787,
+ "step": 4117
+ },
+ {
+ "epoch": 57.19563318777293,
+ "grad_norm": 0.017354389652609825,
+ "learning_rate": 0.0006,
+ "loss": 4.431153774261475,
+ "step": 4118
+ },
+ {
+ "epoch": 57.209606986899566,
+ "grad_norm": 0.015817373991012573,
+ "learning_rate": 0.0006,
+ "loss": 4.448566913604736,
+ "step": 4119
+ },
+ {
+ "epoch": 57.223580786026204,
+ "grad_norm": 0.01635785959661007,
+ "learning_rate": 0.0006,
+ "loss": 4.408031463623047,
+ "step": 4120
+ },
+ {
+ "epoch": 57.237554585152836,
+ "grad_norm": 0.014528338797390461,
+ "learning_rate": 0.0006,
+ "loss": 4.541668891906738,
+ "step": 4121
+ },
+ {
+ "epoch": 57.251528384279474,
+ "grad_norm": 0.01589174196124077,
+ "learning_rate": 0.0006,
+ "loss": 4.513199329376221,
+ "step": 4122
+ },
+ {
+ "epoch": 57.26550218340611,
+ "grad_norm": 0.016093695536255836,
+ "learning_rate": 0.0006,
+ "loss": 4.516178131103516,
+ "step": 4123
+ },
+ {
+ "epoch": 57.27947598253275,
+ "grad_norm": 0.01788361556828022,
+ "learning_rate": 0.0006,
+ "loss": 4.524880409240723,
+ "step": 4124
+ },
+ {
+ "epoch": 57.29344978165939,
+ "grad_norm": 0.01591576263308525,
+ "learning_rate": 0.0006,
+ "loss": 4.44764518737793,
+ "step": 4125
+ },
+ {
+ "epoch": 57.30742358078603,
+ "grad_norm": 0.015452570281922817,
+ "learning_rate": 0.0006,
+ "loss": 4.46844482421875,
+ "step": 4126
+ },
+ {
+ "epoch": 57.32139737991266,
+ "grad_norm": 0.016404492780566216,
+ "learning_rate": 0.0006,
+ "loss": 4.443678855895996,
+ "step": 4127
+ },
+ {
+ "epoch": 57.3353711790393,
+ "grad_norm": 0.019967148080468178,
+ "learning_rate": 0.0006,
+ "loss": 4.430230140686035,
+ "step": 4128
+ },
+ {
+ "epoch": 57.34934497816594,
+ "grad_norm": 0.02594015561044216,
+ "learning_rate": 0.0006,
+ "loss": 4.47061824798584,
+ "step": 4129
+ },
+ {
+ "epoch": 57.36331877729258,
+ "grad_norm": 0.031096890568733215,
+ "learning_rate": 0.0006,
+ "loss": 4.462275981903076,
+ "step": 4130
+ },
+ {
+ "epoch": 57.377292576419215,
+ "grad_norm": 0.02834898792207241,
+ "learning_rate": 0.0006,
+ "loss": 4.41143798828125,
+ "step": 4131
+ },
+ {
+ "epoch": 57.391266375545854,
+ "grad_norm": 0.023621153086423874,
+ "learning_rate": 0.0006,
+ "loss": 4.562548637390137,
+ "step": 4132
+ },
+ {
+ "epoch": 57.40524017467249,
+ "grad_norm": 0.02277296595275402,
+ "learning_rate": 0.0006,
+ "loss": 4.456247329711914,
+ "step": 4133
+ },
+ {
+ "epoch": 57.419213973799124,
+ "grad_norm": 0.019868768751621246,
+ "learning_rate": 0.0006,
+ "loss": 4.524300575256348,
+ "step": 4134
+ },
+ {
+ "epoch": 57.43318777292576,
+ "grad_norm": 0.01887078955769539,
+ "learning_rate": 0.0006,
+ "loss": 4.540097236633301,
+ "step": 4135
+ },
+ {
+ "epoch": 57.4471615720524,
+ "grad_norm": 0.024268802255392075,
+ "learning_rate": 0.0006,
+ "loss": 4.397233963012695,
+ "step": 4136
+ },
+ {
+ "epoch": 57.46113537117904,
+ "grad_norm": 0.025739721953868866,
+ "learning_rate": 0.0006,
+ "loss": 4.354194641113281,
+ "step": 4137
+ },
+ {
+ "epoch": 57.47510917030568,
+ "grad_norm": 0.02442290261387825,
+ "learning_rate": 0.0006,
+ "loss": 4.303822994232178,
+ "step": 4138
+ },
+ {
+ "epoch": 57.48908296943232,
+ "grad_norm": 0.022215209901332855,
+ "learning_rate": 0.0006,
+ "loss": 4.4870452880859375,
+ "step": 4139
+ },
+ {
+ "epoch": 57.50305676855895,
+ "grad_norm": 0.01789158768951893,
+ "learning_rate": 0.0006,
+ "loss": 4.343809127807617,
+ "step": 4140
+ },
+ {
+ "epoch": 57.51703056768559,
+ "grad_norm": 0.020381739363074303,
+ "learning_rate": 0.0006,
+ "loss": 4.53312873840332,
+ "step": 4141
+ },
+ {
+ "epoch": 57.531004366812226,
+ "grad_norm": 0.019585080444812775,
+ "learning_rate": 0.0006,
+ "loss": 4.510162353515625,
+ "step": 4142
+ },
+ {
+ "epoch": 57.544978165938865,
+ "grad_norm": 0.01985127478837967,
+ "learning_rate": 0.0006,
+ "loss": 4.3654022216796875,
+ "step": 4143
+ },
+ {
+ "epoch": 57.5589519650655,
+ "grad_norm": 0.01962905190885067,
+ "learning_rate": 0.0006,
+ "loss": 4.560830116271973,
+ "step": 4144
+ },
+ {
+ "epoch": 57.57292576419214,
+ "grad_norm": 0.02679629437625408,
+ "learning_rate": 0.0006,
+ "loss": 4.300804138183594,
+ "step": 4145
+ },
+ {
+ "epoch": 57.58689956331878,
+ "grad_norm": 0.039830856025218964,
+ "learning_rate": 0.0006,
+ "loss": 4.54150915145874,
+ "step": 4146
+ },
+ {
+ "epoch": 57.60087336244541,
+ "grad_norm": 0.049527477473020554,
+ "learning_rate": 0.0006,
+ "loss": 4.405792236328125,
+ "step": 4147
+ },
+ {
+ "epoch": 57.61484716157205,
+ "grad_norm": 0.042966946959495544,
+ "learning_rate": 0.0006,
+ "loss": 4.500643730163574,
+ "step": 4148
+ },
+ {
+ "epoch": 57.62882096069869,
+ "grad_norm": 0.08999177813529968,
+ "learning_rate": 0.0006,
+ "loss": 4.504790782928467,
+ "step": 4149
+ },
+ {
+ "epoch": 57.64279475982533,
+ "grad_norm": 0.44693124294281006,
+ "learning_rate": 0.0006,
+ "loss": 4.845344066619873,
+ "step": 4150
+ },
+ {
+ "epoch": 57.65676855895197,
+ "grad_norm": 0.7143144011497498,
+ "learning_rate": 0.0006,
+ "loss": 6.907845497131348,
+ "step": 4151
+ },
+ {
+ "epoch": 57.670742358078606,
+ "grad_norm": 0.25239595770835876,
+ "learning_rate": 0.0006,
+ "loss": 7.082579612731934,
+ "step": 4152
+ },
+ {
+ "epoch": 57.68471615720524,
+ "grad_norm": 0.26550790667533875,
+ "learning_rate": 0.0006,
+ "loss": 7.50180196762085,
+ "step": 4153
+ },
+ {
+ "epoch": 57.698689956331876,
+ "grad_norm": 0.14735378324985504,
+ "learning_rate": 0.0006,
+ "loss": 7.214590072631836,
+ "step": 4154
+ },
+ {
+ "epoch": 57.712663755458514,
+ "grad_norm": 0.17685334384441376,
+ "learning_rate": 0.0006,
+ "loss": 6.861320495605469,
+ "step": 4155
+ },
+ {
+ "epoch": 57.72663755458515,
+ "grad_norm": 0.18231339752674103,
+ "learning_rate": 0.0006,
+ "loss": 6.791839599609375,
+ "step": 4156
+ },
+ {
+ "epoch": 57.74061135371179,
+ "grad_norm": 0.09230359643697739,
+ "learning_rate": 0.0006,
+ "loss": 6.760223388671875,
+ "step": 4157
+ },
+ {
+ "epoch": 57.75458515283843,
+ "grad_norm": 0.06918249279260635,
+ "learning_rate": 0.0006,
+ "loss": 6.532999515533447,
+ "step": 4158
+ },
+ {
+ "epoch": 57.76855895196506,
+ "grad_norm": 0.0901007279753685,
+ "learning_rate": 0.0006,
+ "loss": 6.52907133102417,
+ "step": 4159
+ },
+ {
+ "epoch": 57.7825327510917,
+ "grad_norm": 0.05678229779005051,
+ "learning_rate": 0.0006,
+ "loss": 6.295290470123291,
+ "step": 4160
+ },
+ {
+ "epoch": 57.79650655021834,
+ "grad_norm": 0.04450210556387901,
+ "learning_rate": 0.0006,
+ "loss": 6.14900541305542,
+ "step": 4161
+ },
+ {
+ "epoch": 57.81048034934498,
+ "grad_norm": 0.049191512167453766,
+ "learning_rate": 0.0006,
+ "loss": 6.19650411605835,
+ "step": 4162
+ },
+ {
+ "epoch": 57.82445414847162,
+ "grad_norm": 0.03908967226743698,
+ "learning_rate": 0.0006,
+ "loss": 6.054948806762695,
+ "step": 4163
+ },
+ {
+ "epoch": 57.838427947598255,
+ "grad_norm": 0.04023387283086777,
+ "learning_rate": 0.0006,
+ "loss": 6.034873962402344,
+ "step": 4164
+ },
+ {
+ "epoch": 57.852401746724894,
+ "grad_norm": 0.03304268419742584,
+ "learning_rate": 0.0006,
+ "loss": 5.769852161407471,
+ "step": 4165
+ },
+ {
+ "epoch": 57.866375545851525,
+ "grad_norm": 0.03432450816035271,
+ "learning_rate": 0.0006,
+ "loss": 5.788785457611084,
+ "step": 4166
+ },
+ {
+ "epoch": 57.880349344978164,
+ "grad_norm": 0.029965840280056,
+ "learning_rate": 0.0006,
+ "loss": 5.815243721008301,
+ "step": 4167
+ },
+ {
+ "epoch": 57.8943231441048,
+ "grad_norm": 0.03637698292732239,
+ "learning_rate": 0.0006,
+ "loss": 5.7204060554504395,
+ "step": 4168
+ },
+ {
+ "epoch": 57.90829694323144,
+ "grad_norm": 0.03477517142891884,
+ "learning_rate": 0.0006,
+ "loss": 5.605930328369141,
+ "step": 4169
+ },
+ {
+ "epoch": 57.92227074235808,
+ "grad_norm": 0.05061859264969826,
+ "learning_rate": 0.0006,
+ "loss": 5.547054767608643,
+ "step": 4170
+ },
+ {
+ "epoch": 57.93624454148472,
+ "grad_norm": 0.06299655884504318,
+ "learning_rate": 0.0006,
+ "loss": 5.552166938781738,
+ "step": 4171
+ },
+ {
+ "epoch": 57.95021834061135,
+ "grad_norm": 0.04215948283672333,
+ "learning_rate": 0.0006,
+ "loss": 5.375457763671875,
+ "step": 4172
+ },
+ {
+ "epoch": 57.96419213973799,
+ "grad_norm": 0.0348566472530365,
+ "learning_rate": 0.0006,
+ "loss": 5.418606281280518,
+ "step": 4173
+ },
+ {
+ "epoch": 57.97816593886463,
+ "grad_norm": 0.03807530924677849,
+ "learning_rate": 0.0006,
+ "loss": 5.403156280517578,
+ "step": 4174
+ },
+ {
+ "epoch": 57.992139737991266,
+ "grad_norm": 0.02989993803203106,
+ "learning_rate": 0.0006,
+ "loss": 5.355951309204102,
+ "step": 4175
+ },
+ {
+ "epoch": 58.0,
+ "grad_norm": 0.034300558269023895,
+ "learning_rate": 0.0006,
+ "loss": 5.44920539855957,
+ "step": 4176
+ },
+ {
+ "epoch": 58.0,
+ "eval_loss": 5.416731834411621,
+ "eval_runtime": 56.7934,
+ "eval_samples_per_second": 42.998,
+ "eval_steps_per_second": 1.356,
+ "step": 4176
+ },
+ {
+ "epoch": 58.01397379912664,
+ "grad_norm": 0.033961448818445206,
+ "learning_rate": 0.0006,
+ "loss": 5.250707626342773,
+ "step": 4177
+ },
+ {
+ "epoch": 58.02794759825328,
+ "grad_norm": 0.027343502268195152,
+ "learning_rate": 0.0006,
+ "loss": 5.149238109588623,
+ "step": 4178
+ },
+ {
+ "epoch": 58.041921397379916,
+ "grad_norm": 0.03033650480210781,
+ "learning_rate": 0.0006,
+ "loss": 5.072463512420654,
+ "step": 4179
+ },
+ {
+ "epoch": 58.05589519650655,
+ "grad_norm": 0.028262868523597717,
+ "learning_rate": 0.0006,
+ "loss": 5.109417915344238,
+ "step": 4180
+ },
+ {
+ "epoch": 58.069868995633186,
+ "grad_norm": 0.02373034693300724,
+ "learning_rate": 0.0006,
+ "loss": 5.082464218139648,
+ "step": 4181
+ },
+ {
+ "epoch": 58.083842794759825,
+ "grad_norm": 0.025879688560962677,
+ "learning_rate": 0.0006,
+ "loss": 5.056792259216309,
+ "step": 4182
+ },
+ {
+ "epoch": 58.09781659388646,
+ "grad_norm": 0.026252275332808495,
+ "learning_rate": 0.0006,
+ "loss": 5.064239978790283,
+ "step": 4183
+ },
+ {
+ "epoch": 58.1117903930131,
+ "grad_norm": 0.028177792206406593,
+ "learning_rate": 0.0006,
+ "loss": 4.922310829162598,
+ "step": 4184
+ },
+ {
+ "epoch": 58.12576419213974,
+ "grad_norm": 0.02630820870399475,
+ "learning_rate": 0.0006,
+ "loss": 4.992884635925293,
+ "step": 4185
+ },
+ {
+ "epoch": 58.13973799126637,
+ "grad_norm": 0.024890903383493423,
+ "learning_rate": 0.0006,
+ "loss": 4.910036563873291,
+ "step": 4186
+ },
+ {
+ "epoch": 58.15371179039301,
+ "grad_norm": 0.02822992391884327,
+ "learning_rate": 0.0006,
+ "loss": 4.866640090942383,
+ "step": 4187
+ },
+ {
+ "epoch": 58.16768558951965,
+ "grad_norm": 0.03253607079386711,
+ "learning_rate": 0.0006,
+ "loss": 4.9112958908081055,
+ "step": 4188
+ },
+ {
+ "epoch": 58.18165938864629,
+ "grad_norm": 0.03765944764018059,
+ "learning_rate": 0.0006,
+ "loss": 4.774468421936035,
+ "step": 4189
+ },
+ {
+ "epoch": 58.19563318777293,
+ "grad_norm": 0.039874881505966187,
+ "learning_rate": 0.0006,
+ "loss": 4.831678867340088,
+ "step": 4190
+ },
+ {
+ "epoch": 58.209606986899566,
+ "grad_norm": 0.034595925360918045,
+ "learning_rate": 0.0006,
+ "loss": 4.8243794441223145,
+ "step": 4191
+ },
+ {
+ "epoch": 58.223580786026204,
+ "grad_norm": 0.04840189591050148,
+ "learning_rate": 0.0006,
+ "loss": 4.854156494140625,
+ "step": 4192
+ },
+ {
+ "epoch": 58.237554585152836,
+ "grad_norm": 0.06358753889799118,
+ "learning_rate": 0.0006,
+ "loss": 4.836430549621582,
+ "step": 4193
+ },
+ {
+ "epoch": 58.251528384279474,
+ "grad_norm": 0.0507555715739727,
+ "learning_rate": 0.0006,
+ "loss": 4.88357400894165,
+ "step": 4194
+ },
+ {
+ "epoch": 58.26550218340611,
+ "grad_norm": 0.04384690150618553,
+ "learning_rate": 0.0006,
+ "loss": 4.774049282073975,
+ "step": 4195
+ },
+ {
+ "epoch": 58.27947598253275,
+ "grad_norm": 0.03246928006410599,
+ "learning_rate": 0.0006,
+ "loss": 4.686605930328369,
+ "step": 4196
+ },
+ {
+ "epoch": 58.29344978165939,
+ "grad_norm": 0.028878789395093918,
+ "learning_rate": 0.0006,
+ "loss": 4.796070575714111,
+ "step": 4197
+ },
+ {
+ "epoch": 58.30742358078603,
+ "grad_norm": 0.028175361454486847,
+ "learning_rate": 0.0006,
+ "loss": 4.728490829467773,
+ "step": 4198
+ },
+ {
+ "epoch": 58.32139737991266,
+ "grad_norm": 0.029428014531731606,
+ "learning_rate": 0.0006,
+ "loss": 4.7159624099731445,
+ "step": 4199
+ },
+ {
+ "epoch": 58.3353711790393,
+ "grad_norm": 0.031227827072143555,
+ "learning_rate": 0.0006,
+ "loss": 4.702625274658203,
+ "step": 4200
+ },
+ {
+ "epoch": 58.34934497816594,
+ "grad_norm": 0.028843428939580917,
+ "learning_rate": 0.0006,
+ "loss": 4.811799049377441,
+ "step": 4201
+ },
+ {
+ "epoch": 58.36331877729258,
+ "grad_norm": 0.026780391111969948,
+ "learning_rate": 0.0006,
+ "loss": 4.81662654876709,
+ "step": 4202
+ },
+ {
+ "epoch": 58.377292576419215,
+ "grad_norm": 0.019699757918715477,
+ "learning_rate": 0.0006,
+ "loss": 4.713685989379883,
+ "step": 4203
+ },
+ {
+ "epoch": 58.391266375545854,
+ "grad_norm": 0.024389250203967094,
+ "learning_rate": 0.0006,
+ "loss": 4.72354793548584,
+ "step": 4204
+ },
+ {
+ "epoch": 58.40524017467249,
+ "grad_norm": 0.02218274027109146,
+ "learning_rate": 0.0006,
+ "loss": 4.674125671386719,
+ "step": 4205
+ },
+ {
+ "epoch": 58.419213973799124,
+ "grad_norm": 0.02041488140821457,
+ "learning_rate": 0.0006,
+ "loss": 4.7909698486328125,
+ "step": 4206
+ },
+ {
+ "epoch": 58.43318777292576,
+ "grad_norm": 0.018561935052275658,
+ "learning_rate": 0.0006,
+ "loss": 4.808770179748535,
+ "step": 4207
+ },
+ {
+ "epoch": 58.4471615720524,
+ "grad_norm": 0.01708846725523472,
+ "learning_rate": 0.0006,
+ "loss": 4.701626777648926,
+ "step": 4208
+ },
+ {
+ "epoch": 58.46113537117904,
+ "grad_norm": 0.0222734697163105,
+ "learning_rate": 0.0006,
+ "loss": 4.505410194396973,
+ "step": 4209
+ },
+ {
+ "epoch": 58.47510917030568,
+ "grad_norm": 0.03197057917714119,
+ "learning_rate": 0.0006,
+ "loss": 4.610360145568848,
+ "step": 4210
+ },
+ {
+ "epoch": 58.48908296943232,
+ "grad_norm": 0.06854520738124847,
+ "learning_rate": 0.0006,
+ "loss": 4.625649452209473,
+ "step": 4211
+ },
+ {
+ "epoch": 58.50305676855895,
+ "grad_norm": 0.09981521219015121,
+ "learning_rate": 0.0006,
+ "loss": 4.832150459289551,
+ "step": 4212
+ },
+ {
+ "epoch": 58.51703056768559,
+ "grad_norm": 0.046356040984392166,
+ "learning_rate": 0.0006,
+ "loss": 4.5641632080078125,
+ "step": 4213
+ },
+ {
+ "epoch": 58.531004366812226,
+ "grad_norm": 0.03882667422294617,
+ "learning_rate": 0.0006,
+ "loss": 4.666974067687988,
+ "step": 4214
+ },
+ {
+ "epoch": 58.544978165938865,
+ "grad_norm": 0.033203233033418655,
+ "learning_rate": 0.0006,
+ "loss": 4.629413604736328,
+ "step": 4215
+ },
+ {
+ "epoch": 58.5589519650655,
+ "grad_norm": 0.02806735597550869,
+ "learning_rate": 0.0006,
+ "loss": 4.763245582580566,
+ "step": 4216
+ },
+ {
+ "epoch": 58.57292576419214,
+ "grad_norm": 0.026728278025984764,
+ "learning_rate": 0.0006,
+ "loss": 4.629597187042236,
+ "step": 4217
+ },
+ {
+ "epoch": 58.58689956331878,
+ "grad_norm": 0.024380959570407867,
+ "learning_rate": 0.0006,
+ "loss": 4.633614540100098,
+ "step": 4218
+ },
+ {
+ "epoch": 58.60087336244541,
+ "grad_norm": 0.022312544286251068,
+ "learning_rate": 0.0006,
+ "loss": 4.671379566192627,
+ "step": 4219
+ },
+ {
+ "epoch": 58.61484716157205,
+ "grad_norm": 0.022049568593502045,
+ "learning_rate": 0.0006,
+ "loss": 4.661557197570801,
+ "step": 4220
+ },
+ {
+ "epoch": 58.62882096069869,
+ "grad_norm": 0.02119818702340126,
+ "learning_rate": 0.0006,
+ "loss": 4.688562393188477,
+ "step": 4221
+ },
+ {
+ "epoch": 58.64279475982533,
+ "grad_norm": 0.019769301638007164,
+ "learning_rate": 0.0006,
+ "loss": 4.607748031616211,
+ "step": 4222
+ },
+ {
+ "epoch": 58.65676855895197,
+ "grad_norm": 0.02124079130589962,
+ "learning_rate": 0.0006,
+ "loss": 4.570640563964844,
+ "step": 4223
+ },
+ {
+ "epoch": 58.670742358078606,
+ "grad_norm": 0.016344040632247925,
+ "learning_rate": 0.0006,
+ "loss": 4.676113128662109,
+ "step": 4224
+ },
+ {
+ "epoch": 58.68471615720524,
+ "grad_norm": 0.01735256239771843,
+ "learning_rate": 0.0006,
+ "loss": 4.62088680267334,
+ "step": 4225
+ },
+ {
+ "epoch": 58.698689956331876,
+ "grad_norm": 0.019323458895087242,
+ "learning_rate": 0.0006,
+ "loss": 4.579405307769775,
+ "step": 4226
+ },
+ {
+ "epoch": 58.712663755458514,
+ "grad_norm": 0.01760544814169407,
+ "learning_rate": 0.0006,
+ "loss": 4.625407695770264,
+ "step": 4227
+ },
+ {
+ "epoch": 58.72663755458515,
+ "grad_norm": 0.015593250282108784,
+ "learning_rate": 0.0006,
+ "loss": 4.6326117515563965,
+ "step": 4228
+ },
+ {
+ "epoch": 58.74061135371179,
+ "grad_norm": 0.014357523061335087,
+ "learning_rate": 0.0006,
+ "loss": 4.6869401931762695,
+ "step": 4229
+ },
+ {
+ "epoch": 58.75458515283843,
+ "grad_norm": 0.01604357920587063,
+ "learning_rate": 0.0006,
+ "loss": 4.690784931182861,
+ "step": 4230
+ },
+ {
+ "epoch": 58.76855895196506,
+ "grad_norm": 0.01673845760524273,
+ "learning_rate": 0.0006,
+ "loss": 4.513410568237305,
+ "step": 4231
+ },
+ {
+ "epoch": 58.7825327510917,
+ "grad_norm": 0.013657779432833195,
+ "learning_rate": 0.0006,
+ "loss": 4.637057781219482,
+ "step": 4232
+ },
+ {
+ "epoch": 58.79650655021834,
+ "grad_norm": 0.013373114168643951,
+ "learning_rate": 0.0006,
+ "loss": 4.719967842102051,
+ "step": 4233
+ },
+ {
+ "epoch": 58.81048034934498,
+ "grad_norm": 0.016345568001270294,
+ "learning_rate": 0.0006,
+ "loss": 4.604109764099121,
+ "step": 4234
+ },
+ {
+ "epoch": 58.82445414847162,
+ "grad_norm": 0.01573033444583416,
+ "learning_rate": 0.0006,
+ "loss": 4.6471428871154785,
+ "step": 4235
+ },
+ {
+ "epoch": 58.838427947598255,
+ "grad_norm": 0.01517449039965868,
+ "learning_rate": 0.0006,
+ "loss": 4.722684860229492,
+ "step": 4236
+ },
+ {
+ "epoch": 58.852401746724894,
+ "grad_norm": 0.0145102022215724,
+ "learning_rate": 0.0006,
+ "loss": 4.662420272827148,
+ "step": 4237
+ },
+ {
+ "epoch": 58.866375545851525,
+ "grad_norm": 0.01457217987626791,
+ "learning_rate": 0.0006,
+ "loss": 4.585169792175293,
+ "step": 4238
+ },
+ {
+ "epoch": 58.880349344978164,
+ "grad_norm": 0.01406773366034031,
+ "learning_rate": 0.0006,
+ "loss": 4.5413665771484375,
+ "step": 4239
+ },
+ {
+ "epoch": 58.8943231441048,
+ "grad_norm": 0.012653871439397335,
+ "learning_rate": 0.0006,
+ "loss": 4.553636074066162,
+ "step": 4240
+ },
+ {
+ "epoch": 58.90829694323144,
+ "grad_norm": 0.01261200476437807,
+ "learning_rate": 0.0006,
+ "loss": 4.534193992614746,
+ "step": 4241
+ },
+ {
+ "epoch": 58.92227074235808,
+ "grad_norm": 0.012955864891409874,
+ "learning_rate": 0.0006,
+ "loss": 4.697248458862305,
+ "step": 4242
+ },
+ {
+ "epoch": 58.93624454148472,
+ "grad_norm": 0.012134749442338943,
+ "learning_rate": 0.0006,
+ "loss": 4.557095527648926,
+ "step": 4243
+ },
+ {
+ "epoch": 58.95021834061135,
+ "grad_norm": 0.012164677493274212,
+ "learning_rate": 0.0006,
+ "loss": 4.623579025268555,
+ "step": 4244
+ },
+ {
+ "epoch": 58.96419213973799,
+ "grad_norm": 0.013802947476506233,
+ "learning_rate": 0.0006,
+ "loss": 4.562838077545166,
+ "step": 4245
+ },
+ {
+ "epoch": 58.97816593886463,
+ "grad_norm": 0.011420476250350475,
+ "learning_rate": 0.0006,
+ "loss": 4.62265682220459,
+ "step": 4246
+ },
+ {
+ "epoch": 58.992139737991266,
+ "grad_norm": 0.01169576682150364,
+ "learning_rate": 0.0006,
+ "loss": 4.727449417114258,
+ "step": 4247
+ },
+ {
+ "epoch": 59.0,
+ "grad_norm": 0.013577710837125778,
+ "learning_rate": 0.0006,
+ "loss": 4.536325454711914,
+ "step": 4248
+ },
+ {
+ "epoch": 59.0,
+ "eval_loss": 4.86527681350708,
+ "eval_runtime": 60.4463,
+ "eval_samples_per_second": 40.399,
+ "eval_steps_per_second": 1.274,
+ "step": 4248
+ },
+ {
+ "epoch": 59.01397379912664,
+ "grad_norm": 0.015251655131578445,
+ "learning_rate": 0.0006,
+ "loss": 4.634598731994629,
+ "step": 4249
+ },
+ {
+ "epoch": 59.02794759825328,
+ "grad_norm": 0.017414938658475876,
+ "learning_rate": 0.0006,
+ "loss": 4.521329879760742,
+ "step": 4250
+ },
+ {
+ "epoch": 59.041921397379916,
+ "grad_norm": 0.01656588353216648,
+ "learning_rate": 0.0006,
+ "loss": 4.446192264556885,
+ "step": 4251
+ },
+ {
+ "epoch": 59.05589519650655,
+ "grad_norm": 0.012038925662636757,
+ "learning_rate": 0.0006,
+ "loss": 4.523855209350586,
+ "step": 4252
+ },
+ {
+ "epoch": 59.069868995633186,
+ "grad_norm": 0.014403033070266247,
+ "learning_rate": 0.0006,
+ "loss": 4.561079978942871,
+ "step": 4253
+ },
+ {
+ "epoch": 59.083842794759825,
+ "grad_norm": 0.016243597492575645,
+ "learning_rate": 0.0006,
+ "loss": 4.673602104187012,
+ "step": 4254
+ },
+ {
+ "epoch": 59.09781659388646,
+ "grad_norm": 0.016492877155542374,
+ "learning_rate": 0.0006,
+ "loss": 4.564824104309082,
+ "step": 4255
+ },
+ {
+ "epoch": 59.1117903930131,
+ "grad_norm": 0.015980370342731476,
+ "learning_rate": 0.0006,
+ "loss": 4.339249134063721,
+ "step": 4256
+ },
+ {
+ "epoch": 59.12576419213974,
+ "grad_norm": 0.019687356427311897,
+ "learning_rate": 0.0006,
+ "loss": 4.481834888458252,
+ "step": 4257
+ },
+ {
+ "epoch": 59.13973799126637,
+ "grad_norm": 0.02983798086643219,
+ "learning_rate": 0.0006,
+ "loss": 4.4495697021484375,
+ "step": 4258
+ },
+ {
+ "epoch": 59.15371179039301,
+ "grad_norm": 0.04600981995463371,
+ "learning_rate": 0.0006,
+ "loss": 4.567838668823242,
+ "step": 4259
+ },
+ {
+ "epoch": 59.16768558951965,
+ "grad_norm": 0.04852001741528511,
+ "learning_rate": 0.0006,
+ "loss": 4.594701766967773,
+ "step": 4260
+ },
+ {
+ "epoch": 59.18165938864629,
+ "grad_norm": 0.03429504856467247,
+ "learning_rate": 0.0006,
+ "loss": 4.573735237121582,
+ "step": 4261
+ },
+ {
+ "epoch": 59.19563318777293,
+ "grad_norm": 0.029345160350203514,
+ "learning_rate": 0.0006,
+ "loss": 4.488258361816406,
+ "step": 4262
+ },
+ {
+ "epoch": 59.209606986899566,
+ "grad_norm": 0.01961623504757881,
+ "learning_rate": 0.0006,
+ "loss": 4.545222282409668,
+ "step": 4263
+ },
+ {
+ "epoch": 59.223580786026204,
+ "grad_norm": 0.020406130701303482,
+ "learning_rate": 0.0006,
+ "loss": 4.471898078918457,
+ "step": 4264
+ },
+ {
+ "epoch": 59.237554585152836,
+ "grad_norm": 0.020500272512435913,
+ "learning_rate": 0.0006,
+ "loss": 4.538337707519531,
+ "step": 4265
+ },
+ {
+ "epoch": 59.251528384279474,
+ "grad_norm": 0.020727017894387245,
+ "learning_rate": 0.0006,
+ "loss": 4.437284469604492,
+ "step": 4266
+ },
+ {
+ "epoch": 59.26550218340611,
+ "grad_norm": 0.03395998477935791,
+ "learning_rate": 0.0006,
+ "loss": 4.459124565124512,
+ "step": 4267
+ },
+ {
+ "epoch": 59.27947598253275,
+ "grad_norm": 0.04635605961084366,
+ "learning_rate": 0.0006,
+ "loss": 4.520228385925293,
+ "step": 4268
+ },
+ {
+ "epoch": 59.29344978165939,
+ "grad_norm": 0.04253006726503372,
+ "learning_rate": 0.0006,
+ "loss": 4.578673362731934,
+ "step": 4269
+ },
+ {
+ "epoch": 59.30742358078603,
+ "grad_norm": 0.02403509058058262,
+ "learning_rate": 0.0006,
+ "loss": 4.484639644622803,
+ "step": 4270
+ },
+ {
+ "epoch": 59.32139737991266,
+ "grad_norm": 0.02254190482199192,
+ "learning_rate": 0.0006,
+ "loss": 4.511314868927002,
+ "step": 4271
+ },
+ {
+ "epoch": 59.3353711790393,
+ "grad_norm": 0.021585190668702126,
+ "learning_rate": 0.0006,
+ "loss": 4.530239105224609,
+ "step": 4272
+ },
+ {
+ "epoch": 59.34934497816594,
+ "grad_norm": 0.021210316568613052,
+ "learning_rate": 0.0006,
+ "loss": 4.593514442443848,
+ "step": 4273
+ },
+ {
+ "epoch": 59.36331877729258,
+ "grad_norm": 0.019317107275128365,
+ "learning_rate": 0.0006,
+ "loss": 4.5591535568237305,
+ "step": 4274
+ },
+ {
+ "epoch": 59.377292576419215,
+ "grad_norm": 0.018059708178043365,
+ "learning_rate": 0.0006,
+ "loss": 4.4499592781066895,
+ "step": 4275
+ },
+ {
+ "epoch": 59.391266375545854,
+ "grad_norm": 0.017847446724772453,
+ "learning_rate": 0.0006,
+ "loss": 4.473999500274658,
+ "step": 4276
+ },
+ {
+ "epoch": 59.40524017467249,
+ "grad_norm": 0.014944756403565407,
+ "learning_rate": 0.0006,
+ "loss": 4.5934529304504395,
+ "step": 4277
+ },
+ {
+ "epoch": 59.419213973799124,
+ "grad_norm": 0.01663832925260067,
+ "learning_rate": 0.0006,
+ "loss": 4.504262924194336,
+ "step": 4278
+ },
+ {
+ "epoch": 59.43318777292576,
+ "grad_norm": 0.015571796335279942,
+ "learning_rate": 0.0006,
+ "loss": 4.512310981750488,
+ "step": 4279
+ },
+ {
+ "epoch": 59.4471615720524,
+ "grad_norm": 0.015896400436758995,
+ "learning_rate": 0.0006,
+ "loss": 4.452714920043945,
+ "step": 4280
+ },
+ {
+ "epoch": 59.46113537117904,
+ "grad_norm": 0.016010284423828125,
+ "learning_rate": 0.0006,
+ "loss": 4.6188063621521,
+ "step": 4281
+ },
+ {
+ "epoch": 59.47510917030568,
+ "grad_norm": 0.015075593255460262,
+ "learning_rate": 0.0006,
+ "loss": 4.477978229522705,
+ "step": 4282
+ },
+ {
+ "epoch": 59.48908296943232,
+ "grad_norm": 0.014698950573801994,
+ "learning_rate": 0.0006,
+ "loss": 4.544955730438232,
+ "step": 4283
+ },
+ {
+ "epoch": 59.50305676855895,
+ "grad_norm": 0.0136245833709836,
+ "learning_rate": 0.0006,
+ "loss": 4.530460357666016,
+ "step": 4284
+ },
+ {
+ "epoch": 59.51703056768559,
+ "grad_norm": 0.013118310831487179,
+ "learning_rate": 0.0006,
+ "loss": 4.614217758178711,
+ "step": 4285
+ },
+ {
+ "epoch": 59.531004366812226,
+ "grad_norm": 0.01295961532741785,
+ "learning_rate": 0.0006,
+ "loss": 4.43156099319458,
+ "step": 4286
+ },
+ {
+ "epoch": 59.544978165938865,
+ "grad_norm": 0.012429811991751194,
+ "learning_rate": 0.0006,
+ "loss": 4.517333030700684,
+ "step": 4287
+ },
+ {
+ "epoch": 59.5589519650655,
+ "grad_norm": 0.01261830423027277,
+ "learning_rate": 0.0006,
+ "loss": 4.569780349731445,
+ "step": 4288
+ },
+ {
+ "epoch": 59.57292576419214,
+ "grad_norm": 0.012394499965012074,
+ "learning_rate": 0.0006,
+ "loss": 4.5970869064331055,
+ "step": 4289
+ },
+ {
+ "epoch": 59.58689956331878,
+ "grad_norm": 0.01290759164839983,
+ "learning_rate": 0.0006,
+ "loss": 4.437100410461426,
+ "step": 4290
+ },
+ {
+ "epoch": 59.60087336244541,
+ "grad_norm": 0.012042169459164143,
+ "learning_rate": 0.0006,
+ "loss": 4.567404747009277,
+ "step": 4291
+ },
+ {
+ "epoch": 59.61484716157205,
+ "grad_norm": 0.012338520959019661,
+ "learning_rate": 0.0006,
+ "loss": 4.476193428039551,
+ "step": 4292
+ },
+ {
+ "epoch": 59.62882096069869,
+ "grad_norm": 0.01321258582174778,
+ "learning_rate": 0.0006,
+ "loss": 4.566680431365967,
+ "step": 4293
+ },
+ {
+ "epoch": 59.64279475982533,
+ "grad_norm": 0.013478122651576996,
+ "learning_rate": 0.0006,
+ "loss": 4.487398147583008,
+ "step": 4294
+ },
+ {
+ "epoch": 59.65676855895197,
+ "grad_norm": 0.01537603884935379,
+ "learning_rate": 0.0006,
+ "loss": 4.449337005615234,
+ "step": 4295
+ },
+ {
+ "epoch": 59.670742358078606,
+ "grad_norm": 0.017140565440058708,
+ "learning_rate": 0.0006,
+ "loss": 4.5174360275268555,
+ "step": 4296
+ },
+ {
+ "epoch": 59.68471615720524,
+ "grad_norm": 0.017931461334228516,
+ "learning_rate": 0.0006,
+ "loss": 4.522104740142822,
+ "step": 4297
+ },
+ {
+ "epoch": 59.698689956331876,
+ "grad_norm": 0.018037918955087662,
+ "learning_rate": 0.0006,
+ "loss": 4.496928691864014,
+ "step": 4298
+ },
+ {
+ "epoch": 59.712663755458514,
+ "grad_norm": 0.01684504561126232,
+ "learning_rate": 0.0006,
+ "loss": 4.5332417488098145,
+ "step": 4299
+ },
+ {
+ "epoch": 59.72663755458515,
+ "grad_norm": 0.016268383711576462,
+ "learning_rate": 0.0006,
+ "loss": 4.468893051147461,
+ "step": 4300
+ },
+ {
+ "epoch": 59.74061135371179,
+ "grad_norm": 0.014859385788440704,
+ "learning_rate": 0.0006,
+ "loss": 4.551092147827148,
+ "step": 4301
+ },
+ {
+ "epoch": 59.75458515283843,
+ "grad_norm": 0.013589047826826572,
+ "learning_rate": 0.0006,
+ "loss": 4.470579147338867,
+ "step": 4302
+ },
+ {
+ "epoch": 59.76855895196506,
+ "grad_norm": 0.015704551711678505,
+ "learning_rate": 0.0006,
+ "loss": 4.330068111419678,
+ "step": 4303
+ },
+ {
+ "epoch": 59.7825327510917,
+ "grad_norm": 0.017869921401143074,
+ "learning_rate": 0.0006,
+ "loss": 4.428255081176758,
+ "step": 4304
+ },
+ {
+ "epoch": 59.79650655021834,
+ "grad_norm": 0.019828537479043007,
+ "learning_rate": 0.0006,
+ "loss": 4.496023178100586,
+ "step": 4305
+ },
+ {
+ "epoch": 59.81048034934498,
+ "grad_norm": 0.015856236219406128,
+ "learning_rate": 0.0006,
+ "loss": 4.427152156829834,
+ "step": 4306
+ },
+ {
+ "epoch": 59.82445414847162,
+ "grad_norm": 0.013349304907023907,
+ "learning_rate": 0.0006,
+ "loss": 4.417333602905273,
+ "step": 4307
+ },
+ {
+ "epoch": 59.838427947598255,
+ "grad_norm": 0.014111637137830257,
+ "learning_rate": 0.0006,
+ "loss": 4.4507575035095215,
+ "step": 4308
+ },
+ {
+ "epoch": 59.852401746724894,
+ "grad_norm": 0.015186597593128681,
+ "learning_rate": 0.0006,
+ "loss": 4.375116348266602,
+ "step": 4309
+ },
+ {
+ "epoch": 59.866375545851525,
+ "grad_norm": 0.01604730449616909,
+ "learning_rate": 0.0006,
+ "loss": 4.604518890380859,
+ "step": 4310
+ },
+ {
+ "epoch": 59.880349344978164,
+ "grad_norm": 0.015308464877307415,
+ "learning_rate": 0.0006,
+ "loss": 4.507232189178467,
+ "step": 4311
+ },
+ {
+ "epoch": 59.8943231441048,
+ "grad_norm": 0.014017206616699696,
+ "learning_rate": 0.0006,
+ "loss": 4.398487091064453,
+ "step": 4312
+ },
+ {
+ "epoch": 59.90829694323144,
+ "grad_norm": 0.012911394238471985,
+ "learning_rate": 0.0006,
+ "loss": 4.567872047424316,
+ "step": 4313
+ },
+ {
+ "epoch": 59.92227074235808,
+ "grad_norm": 0.012775142677128315,
+ "learning_rate": 0.0006,
+ "loss": 4.562105655670166,
+ "step": 4314
+ },
+ {
+ "epoch": 59.93624454148472,
+ "grad_norm": 0.014750408008694649,
+ "learning_rate": 0.0006,
+ "loss": 4.456263542175293,
+ "step": 4315
+ },
+ {
+ "epoch": 59.95021834061135,
+ "grad_norm": 0.016726011410355568,
+ "learning_rate": 0.0006,
+ "loss": 4.360722064971924,
+ "step": 4316
+ },
+ {
+ "epoch": 59.96419213973799,
+ "grad_norm": 0.01734776981174946,
+ "learning_rate": 0.0006,
+ "loss": 4.572544574737549,
+ "step": 4317
+ },
+ {
+ "epoch": 59.97816593886463,
+ "grad_norm": 0.01822415366768837,
+ "learning_rate": 0.0006,
+ "loss": 4.490501880645752,
+ "step": 4318
+ },
+ {
+ "epoch": 59.992139737991266,
+ "grad_norm": 0.014638577587902546,
+ "learning_rate": 0.0006,
+ "loss": 4.440951824188232,
+ "step": 4319
+ },
+ {
+ "epoch": 60.0,
+ "grad_norm": 0.01255202479660511,
+ "learning_rate": 0.0006,
+ "loss": 4.559123516082764,
+ "step": 4320
+ },
+ {
+ "epoch": 60.0,
+ "eval_loss": 4.741089344024658,
+ "eval_runtime": 56.8997,
+ "eval_samples_per_second": 42.918,
+ "eval_steps_per_second": 1.353,
+ "step": 4320
+ },
+ {
+ "epoch": 60.01397379912664,
+ "grad_norm": 0.016548393294215202,
+ "learning_rate": 0.0006,
+ "loss": 4.442024230957031,
+ "step": 4321
+ },
+ {
+ "epoch": 60.02794759825328,
+ "grad_norm": 0.019759127870202065,
+ "learning_rate": 0.0006,
+ "loss": 4.394334316253662,
+ "step": 4322
+ },
+ {
+ "epoch": 60.041921397379916,
+ "grad_norm": 0.018325170502066612,
+ "learning_rate": 0.0006,
+ "loss": 4.318270683288574,
+ "step": 4323
+ },
+ {
+ "epoch": 60.05589519650655,
+ "grad_norm": 0.015289463102817535,
+ "learning_rate": 0.0006,
+ "loss": 4.318228721618652,
+ "step": 4324
+ },
+ {
+ "epoch": 60.069868995633186,
+ "grad_norm": 0.015424872748553753,
+ "learning_rate": 0.0006,
+ "loss": 4.3605499267578125,
+ "step": 4325
+ },
+ {
+ "epoch": 60.083842794759825,
+ "grad_norm": 0.014085669070482254,
+ "learning_rate": 0.0006,
+ "loss": 4.359585762023926,
+ "step": 4326
+ },
+ {
+ "epoch": 60.09781659388646,
+ "grad_norm": 0.014004210010170937,
+ "learning_rate": 0.0006,
+ "loss": 4.447783470153809,
+ "step": 4327
+ },
+ {
+ "epoch": 60.1117903930131,
+ "grad_norm": 0.013514582999050617,
+ "learning_rate": 0.0006,
+ "loss": 4.4850239753723145,
+ "step": 4328
+ },
+ {
+ "epoch": 60.12576419213974,
+ "grad_norm": 0.013049394823610783,
+ "learning_rate": 0.0006,
+ "loss": 4.528675556182861,
+ "step": 4329
+ },
+ {
+ "epoch": 60.13973799126637,
+ "grad_norm": 0.01378058921545744,
+ "learning_rate": 0.0006,
+ "loss": 4.444823741912842,
+ "step": 4330
+ },
+ {
+ "epoch": 60.15371179039301,
+ "grad_norm": 0.013752233237028122,
+ "learning_rate": 0.0006,
+ "loss": 4.5205206871032715,
+ "step": 4331
+ },
+ {
+ "epoch": 60.16768558951965,
+ "grad_norm": 0.013431803323328495,
+ "learning_rate": 0.0006,
+ "loss": 4.547320365905762,
+ "step": 4332
+ },
+ {
+ "epoch": 60.18165938864629,
+ "grad_norm": 0.014544196426868439,
+ "learning_rate": 0.0006,
+ "loss": 4.517143249511719,
+ "step": 4333
+ },
+ {
+ "epoch": 60.19563318777293,
+ "grad_norm": 0.01612176187336445,
+ "learning_rate": 0.0006,
+ "loss": 4.517232894897461,
+ "step": 4334
+ },
+ {
+ "epoch": 60.209606986899566,
+ "grad_norm": 0.01803239807486534,
+ "learning_rate": 0.0006,
+ "loss": 4.377574920654297,
+ "step": 4335
+ },
+ {
+ "epoch": 60.223580786026204,
+ "grad_norm": 0.014871489256620407,
+ "learning_rate": 0.0006,
+ "loss": 4.538285255432129,
+ "step": 4336
+ },
+ {
+ "epoch": 60.237554585152836,
+ "grad_norm": 0.013927377760410309,
+ "learning_rate": 0.0006,
+ "loss": 4.386175632476807,
+ "step": 4337
+ },
+ {
+ "epoch": 60.251528384279474,
+ "grad_norm": 0.014757219702005386,
+ "learning_rate": 0.0006,
+ "loss": 4.528553009033203,
+ "step": 4338
+ },
+ {
+ "epoch": 60.26550218340611,
+ "grad_norm": 0.01433873176574707,
+ "learning_rate": 0.0006,
+ "loss": 4.476929187774658,
+ "step": 4339
+ },
+ {
+ "epoch": 60.27947598253275,
+ "grad_norm": 0.015768803656101227,
+ "learning_rate": 0.0006,
+ "loss": 4.50970458984375,
+ "step": 4340
+ },
+ {
+ "epoch": 60.29344978165939,
+ "grad_norm": 0.015651429072022438,
+ "learning_rate": 0.0006,
+ "loss": 4.427133083343506,
+ "step": 4341
+ },
+ {
+ "epoch": 60.30742358078603,
+ "grad_norm": 0.01673000678420067,
+ "learning_rate": 0.0006,
+ "loss": 4.447723865509033,
+ "step": 4342
+ },
+ {
+ "epoch": 60.32139737991266,
+ "grad_norm": 0.018481194972991943,
+ "learning_rate": 0.0006,
+ "loss": 4.559832572937012,
+ "step": 4343
+ },
+ {
+ "epoch": 60.3353711790393,
+ "grad_norm": 0.016731027513742447,
+ "learning_rate": 0.0006,
+ "loss": 4.470893859863281,
+ "step": 4344
+ },
+ {
+ "epoch": 60.34934497816594,
+ "grad_norm": 0.01504011545330286,
+ "learning_rate": 0.0006,
+ "loss": 4.483014106750488,
+ "step": 4345
+ },
+ {
+ "epoch": 60.36331877729258,
+ "grad_norm": 0.01602456159889698,
+ "learning_rate": 0.0006,
+ "loss": 4.548428058624268,
+ "step": 4346
+ },
+ {
+ "epoch": 60.377292576419215,
+ "grad_norm": 0.016027364879846573,
+ "learning_rate": 0.0006,
+ "loss": 4.426779747009277,
+ "step": 4347
+ },
+ {
+ "epoch": 60.391266375545854,
+ "grad_norm": 0.01873827911913395,
+ "learning_rate": 0.0006,
+ "loss": 4.353686332702637,
+ "step": 4348
+ },
+ {
+ "epoch": 60.40524017467249,
+ "grad_norm": 0.01907406374812126,
+ "learning_rate": 0.0006,
+ "loss": 4.502945423126221,
+ "step": 4349
+ },
+ {
+ "epoch": 60.419213973799124,
+ "grad_norm": 0.016083354130387306,
+ "learning_rate": 0.0006,
+ "loss": 4.431731700897217,
+ "step": 4350
+ },
+ {
+ "epoch": 60.43318777292576,
+ "grad_norm": 0.015293709933757782,
+ "learning_rate": 0.0006,
+ "loss": 4.480112552642822,
+ "step": 4351
+ },
+ {
+ "epoch": 60.4471615720524,
+ "grad_norm": 0.01688101328909397,
+ "learning_rate": 0.0006,
+ "loss": 4.50853157043457,
+ "step": 4352
+ },
+ {
+ "epoch": 60.46113537117904,
+ "grad_norm": 0.02094118855893612,
+ "learning_rate": 0.0006,
+ "loss": 4.359930038452148,
+ "step": 4353
+ },
+ {
+ "epoch": 60.47510917030568,
+ "grad_norm": 0.018188107758760452,
+ "learning_rate": 0.0006,
+ "loss": 4.549592018127441,
+ "step": 4354
+ },
+ {
+ "epoch": 60.48908296943232,
+ "grad_norm": 0.01462319865822792,
+ "learning_rate": 0.0006,
+ "loss": 4.49467658996582,
+ "step": 4355
+ },
+ {
+ "epoch": 60.50305676855895,
+ "grad_norm": 0.017508579418063164,
+ "learning_rate": 0.0006,
+ "loss": 4.398318767547607,
+ "step": 4356
+ },
+ {
+ "epoch": 60.51703056768559,
+ "grad_norm": 0.018200315535068512,
+ "learning_rate": 0.0006,
+ "loss": 4.44169282913208,
+ "step": 4357
+ },
+ {
+ "epoch": 60.531004366812226,
+ "grad_norm": 0.01717468351125717,
+ "learning_rate": 0.0006,
+ "loss": 4.481400489807129,
+ "step": 4358
+ },
+ {
+ "epoch": 60.544978165938865,
+ "grad_norm": 0.019329151138663292,
+ "learning_rate": 0.0006,
+ "loss": 4.578839302062988,
+ "step": 4359
+ },
+ {
+ "epoch": 60.5589519650655,
+ "grad_norm": 0.016109555959701538,
+ "learning_rate": 0.0006,
+ "loss": 4.461043357849121,
+ "step": 4360
+ },
+ {
+ "epoch": 60.57292576419214,
+ "grad_norm": 0.01494457945227623,
+ "learning_rate": 0.0006,
+ "loss": 4.505096912384033,
+ "step": 4361
+ },
+ {
+ "epoch": 60.58689956331878,
+ "grad_norm": 0.01729399710893631,
+ "learning_rate": 0.0006,
+ "loss": 4.366747856140137,
+ "step": 4362
+ },
+ {
+ "epoch": 60.60087336244541,
+ "grad_norm": 0.01705821603536606,
+ "learning_rate": 0.0006,
+ "loss": 4.42534065246582,
+ "step": 4363
+ },
+ {
+ "epoch": 60.61484716157205,
+ "grad_norm": 0.018391354009509087,
+ "learning_rate": 0.0006,
+ "loss": 4.350294589996338,
+ "step": 4364
+ },
+ {
+ "epoch": 60.62882096069869,
+ "grad_norm": 0.016989829018712044,
+ "learning_rate": 0.0006,
+ "loss": 4.45905876159668,
+ "step": 4365
+ },
+ {
+ "epoch": 60.64279475982533,
+ "grad_norm": 0.017192212864756584,
+ "learning_rate": 0.0006,
+ "loss": 4.526487350463867,
+ "step": 4366
+ },
+ {
+ "epoch": 60.65676855895197,
+ "grad_norm": 0.018091315403580666,
+ "learning_rate": 0.0006,
+ "loss": 4.309333324432373,
+ "step": 4367
+ },
+ {
+ "epoch": 60.670742358078606,
+ "grad_norm": 0.014051459729671478,
+ "learning_rate": 0.0006,
+ "loss": 4.338171005249023,
+ "step": 4368
+ },
+ {
+ "epoch": 60.68471615720524,
+ "grad_norm": 0.013166418299078941,
+ "learning_rate": 0.0006,
+ "loss": 4.409758567810059,
+ "step": 4369
+ },
+ {
+ "epoch": 60.698689956331876,
+ "grad_norm": 0.014260428957641125,
+ "learning_rate": 0.0006,
+ "loss": 4.524338245391846,
+ "step": 4370
+ },
+ {
+ "epoch": 60.712663755458514,
+ "grad_norm": 0.017019255086779594,
+ "learning_rate": 0.0006,
+ "loss": 4.387436866760254,
+ "step": 4371
+ },
+ {
+ "epoch": 60.72663755458515,
+ "grad_norm": 0.017070675268769264,
+ "learning_rate": 0.0006,
+ "loss": 4.437321662902832,
+ "step": 4372
+ },
+ {
+ "epoch": 60.74061135371179,
+ "grad_norm": 0.015197164379060268,
+ "learning_rate": 0.0006,
+ "loss": 4.371205806732178,
+ "step": 4373
+ },
+ {
+ "epoch": 60.75458515283843,
+ "grad_norm": 0.014791185967624187,
+ "learning_rate": 0.0006,
+ "loss": 4.44439697265625,
+ "step": 4374
+ },
+ {
+ "epoch": 60.76855895196506,
+ "grad_norm": 0.016458261758089066,
+ "learning_rate": 0.0006,
+ "loss": 4.445584297180176,
+ "step": 4375
+ },
+ {
+ "epoch": 60.7825327510917,
+ "grad_norm": 0.017706342041492462,
+ "learning_rate": 0.0006,
+ "loss": 4.348708152770996,
+ "step": 4376
+ },
+ {
+ "epoch": 60.79650655021834,
+ "grad_norm": 0.017561476677656174,
+ "learning_rate": 0.0006,
+ "loss": 4.521929740905762,
+ "step": 4377
+ },
+ {
+ "epoch": 60.81048034934498,
+ "grad_norm": 0.016557445749640465,
+ "learning_rate": 0.0006,
+ "loss": 4.545806884765625,
+ "step": 4378
+ },
+ {
+ "epoch": 60.82445414847162,
+ "grad_norm": 0.016953222453594208,
+ "learning_rate": 0.0006,
+ "loss": 4.4230875968933105,
+ "step": 4379
+ },
+ {
+ "epoch": 60.838427947598255,
+ "grad_norm": 0.017094967886805534,
+ "learning_rate": 0.0006,
+ "loss": 4.423985481262207,
+ "step": 4380
+ },
+ {
+ "epoch": 60.852401746724894,
+ "grad_norm": 0.014787515625357628,
+ "learning_rate": 0.0006,
+ "loss": 4.39473819732666,
+ "step": 4381
+ },
+ {
+ "epoch": 60.866375545851525,
+ "grad_norm": 0.016208263114094734,
+ "learning_rate": 0.0006,
+ "loss": 4.3810038566589355,
+ "step": 4382
+ },
+ {
+ "epoch": 60.880349344978164,
+ "grad_norm": 0.017844321206212044,
+ "learning_rate": 0.0006,
+ "loss": 4.35086727142334,
+ "step": 4383
+ },
+ {
+ "epoch": 60.8943231441048,
+ "grad_norm": 0.017539294436573982,
+ "learning_rate": 0.0006,
+ "loss": 4.460260391235352,
+ "step": 4384
+ },
+ {
+ "epoch": 60.90829694323144,
+ "grad_norm": 0.01612042263150215,
+ "learning_rate": 0.0006,
+ "loss": 4.370500564575195,
+ "step": 4385
+ },
+ {
+ "epoch": 60.92227074235808,
+ "grad_norm": 0.015381601639091969,
+ "learning_rate": 0.0006,
+ "loss": 4.435868263244629,
+ "step": 4386
+ },
+ {
+ "epoch": 60.93624454148472,
+ "grad_norm": 0.01603585295379162,
+ "learning_rate": 0.0006,
+ "loss": 4.434070587158203,
+ "step": 4387
+ },
+ {
+ "epoch": 60.95021834061135,
+ "grad_norm": 0.016268130391836166,
+ "learning_rate": 0.0006,
+ "loss": 4.404343128204346,
+ "step": 4388
+ },
+ {
+ "epoch": 60.96419213973799,
+ "grad_norm": 0.014980032108724117,
+ "learning_rate": 0.0006,
+ "loss": 4.399213790893555,
+ "step": 4389
+ },
+ {
+ "epoch": 60.97816593886463,
+ "grad_norm": 0.016899267211556435,
+ "learning_rate": 0.0006,
+ "loss": 4.508069038391113,
+ "step": 4390
+ },
+ {
+ "epoch": 60.992139737991266,
+ "grad_norm": 0.013756908476352692,
+ "learning_rate": 0.0006,
+ "loss": 4.541858673095703,
+ "step": 4391
+ },
+ {
+ "epoch": 61.0,
+ "grad_norm": 0.01578413136303425,
+ "learning_rate": 0.0006,
+ "loss": 4.543513298034668,
+ "step": 4392
+ },
+ {
+ "epoch": 61.0,
+ "eval_loss": 4.756770610809326,
+ "eval_runtime": 56.4572,
+ "eval_samples_per_second": 43.254,
+ "eval_steps_per_second": 1.364,
+ "step": 4392
+ },
+ {
+ "epoch": 61.01397379912664,
+ "grad_norm": 0.013387506827712059,
+ "learning_rate": 0.0006,
+ "loss": 4.475942611694336,
+ "step": 4393
+ },
+ {
+ "epoch": 61.02794759825328,
+ "grad_norm": 0.014278904534876347,
+ "learning_rate": 0.0006,
+ "loss": 4.442403793334961,
+ "step": 4394
+ },
+ {
+ "epoch": 61.041921397379916,
+ "grad_norm": 0.014359553344547749,
+ "learning_rate": 0.0006,
+ "loss": 4.509463310241699,
+ "step": 4395
+ },
+ {
+ "epoch": 61.05589519650655,
+ "grad_norm": 0.01718050241470337,
+ "learning_rate": 0.0006,
+ "loss": 4.354458808898926,
+ "step": 4396
+ },
+ {
+ "epoch": 61.069868995633186,
+ "grad_norm": 0.015988312661647797,
+ "learning_rate": 0.0006,
+ "loss": 4.484508514404297,
+ "step": 4397
+ },
+ {
+ "epoch": 61.083842794759825,
+ "grad_norm": 0.015466735698282719,
+ "learning_rate": 0.0006,
+ "loss": 4.326908111572266,
+ "step": 4398
+ },
+ {
+ "epoch": 61.09781659388646,
+ "grad_norm": 0.014541332609951496,
+ "learning_rate": 0.0006,
+ "loss": 4.405823230743408,
+ "step": 4399
+ },
+ {
+ "epoch": 61.1117903930131,
+ "grad_norm": 0.014803987927734852,
+ "learning_rate": 0.0006,
+ "loss": 4.375386714935303,
+ "step": 4400
+ },
+ {
+ "epoch": 61.12576419213974,
+ "grad_norm": 0.013962093740701675,
+ "learning_rate": 0.0006,
+ "loss": 4.300838470458984,
+ "step": 4401
+ },
+ {
+ "epoch": 61.13973799126637,
+ "grad_norm": 0.015213954262435436,
+ "learning_rate": 0.0006,
+ "loss": 4.388306617736816,
+ "step": 4402
+ },
+ {
+ "epoch": 61.15371179039301,
+ "grad_norm": 0.018965506926178932,
+ "learning_rate": 0.0006,
+ "loss": 4.388340950012207,
+ "step": 4403
+ },
+ {
+ "epoch": 61.16768558951965,
+ "grad_norm": 0.01958410255610943,
+ "learning_rate": 0.0006,
+ "loss": 4.35012149810791,
+ "step": 4404
+ },
+ {
+ "epoch": 61.18165938864629,
+ "grad_norm": 0.01938101276755333,
+ "learning_rate": 0.0006,
+ "loss": 4.368372917175293,
+ "step": 4405
+ },
+ {
+ "epoch": 61.19563318777293,
+ "grad_norm": 0.016556516289711,
+ "learning_rate": 0.0006,
+ "loss": 4.486198425292969,
+ "step": 4406
+ },
+ {
+ "epoch": 61.209606986899566,
+ "grad_norm": 0.017031289637088776,
+ "learning_rate": 0.0006,
+ "loss": 4.421676158905029,
+ "step": 4407
+ },
+ {
+ "epoch": 61.223580786026204,
+ "grad_norm": 0.018023479729890823,
+ "learning_rate": 0.0006,
+ "loss": 4.38751220703125,
+ "step": 4408
+ },
+ {
+ "epoch": 61.237554585152836,
+ "grad_norm": 0.0178135447204113,
+ "learning_rate": 0.0006,
+ "loss": 4.336528301239014,
+ "step": 4409
+ },
+ {
+ "epoch": 61.251528384279474,
+ "grad_norm": 0.016782190650701523,
+ "learning_rate": 0.0006,
+ "loss": 4.386693954467773,
+ "step": 4410
+ },
+ {
+ "epoch": 61.26550218340611,
+ "grad_norm": 0.015494848601520061,
+ "learning_rate": 0.0006,
+ "loss": 4.308948040008545,
+ "step": 4411
+ },
+ {
+ "epoch": 61.27947598253275,
+ "grad_norm": 0.015002378262579441,
+ "learning_rate": 0.0006,
+ "loss": 4.314206123352051,
+ "step": 4412
+ },
+ {
+ "epoch": 61.29344978165939,
+ "grad_norm": 0.015798920765519142,
+ "learning_rate": 0.0006,
+ "loss": 4.502900123596191,
+ "step": 4413
+ },
+ {
+ "epoch": 61.30742358078603,
+ "grad_norm": 0.015444336459040642,
+ "learning_rate": 0.0006,
+ "loss": 4.473160743713379,
+ "step": 4414
+ },
+ {
+ "epoch": 61.32139737991266,
+ "grad_norm": 0.017025096341967583,
+ "learning_rate": 0.0006,
+ "loss": 4.460621356964111,
+ "step": 4415
+ },
+ {
+ "epoch": 61.3353711790393,
+ "grad_norm": 0.019502537325024605,
+ "learning_rate": 0.0006,
+ "loss": 4.385551452636719,
+ "step": 4416
+ },
+ {
+ "epoch": 61.34934497816594,
+ "grad_norm": 0.018529541790485382,
+ "learning_rate": 0.0006,
+ "loss": 4.425390720367432,
+ "step": 4417
+ },
+ {
+ "epoch": 61.36331877729258,
+ "grad_norm": 0.013155706226825714,
+ "learning_rate": 0.0006,
+ "loss": 4.467020034790039,
+ "step": 4418
+ },
+ {
+ "epoch": 61.377292576419215,
+ "grad_norm": 0.01518090907484293,
+ "learning_rate": 0.0006,
+ "loss": 4.369664192199707,
+ "step": 4419
+ },
+ {
+ "epoch": 61.391266375545854,
+ "grad_norm": 0.01626511849462986,
+ "learning_rate": 0.0006,
+ "loss": 4.463231086730957,
+ "step": 4420
+ },
+ {
+ "epoch": 61.40524017467249,
+ "grad_norm": 0.01701558195054531,
+ "learning_rate": 0.0006,
+ "loss": 4.43708610534668,
+ "step": 4421
+ },
+ {
+ "epoch": 61.419213973799124,
+ "grad_norm": 0.01761862449347973,
+ "learning_rate": 0.0006,
+ "loss": 4.3314208984375,
+ "step": 4422
+ },
+ {
+ "epoch": 61.43318777292576,
+ "grad_norm": 0.017084648832678795,
+ "learning_rate": 0.0006,
+ "loss": 4.39676570892334,
+ "step": 4423
+ },
+ {
+ "epoch": 61.4471615720524,
+ "grad_norm": 0.01667078025639057,
+ "learning_rate": 0.0006,
+ "loss": 4.3789873123168945,
+ "step": 4424
+ },
+ {
+ "epoch": 61.46113537117904,
+ "grad_norm": 0.01877436414361,
+ "learning_rate": 0.0006,
+ "loss": 4.415275573730469,
+ "step": 4425
+ },
+ {
+ "epoch": 61.47510917030568,
+ "grad_norm": 0.020250486209988594,
+ "learning_rate": 0.0006,
+ "loss": 4.478635787963867,
+ "step": 4426
+ },
+ {
+ "epoch": 61.48908296943232,
+ "grad_norm": 0.020244868472218513,
+ "learning_rate": 0.0006,
+ "loss": 4.296854019165039,
+ "step": 4427
+ },
+ {
+ "epoch": 61.50305676855895,
+ "grad_norm": 0.018723875284194946,
+ "learning_rate": 0.0006,
+ "loss": 4.3760881423950195,
+ "step": 4428
+ },
+ {
+ "epoch": 61.51703056768559,
+ "grad_norm": 0.015138164162635803,
+ "learning_rate": 0.0006,
+ "loss": 4.39974308013916,
+ "step": 4429
+ },
+ {
+ "epoch": 61.531004366812226,
+ "grad_norm": 0.015111138112843037,
+ "learning_rate": 0.0006,
+ "loss": 4.34659481048584,
+ "step": 4430
+ },
+ {
+ "epoch": 61.544978165938865,
+ "grad_norm": 0.0164666585624218,
+ "learning_rate": 0.0006,
+ "loss": 4.415737152099609,
+ "step": 4431
+ },
+ {
+ "epoch": 61.5589519650655,
+ "grad_norm": 0.020891312509775162,
+ "learning_rate": 0.0006,
+ "loss": 4.300319671630859,
+ "step": 4432
+ },
+ {
+ "epoch": 61.57292576419214,
+ "grad_norm": 0.02089156024158001,
+ "learning_rate": 0.0006,
+ "loss": 4.373029708862305,
+ "step": 4433
+ },
+ {
+ "epoch": 61.58689956331878,
+ "grad_norm": 0.020114559680223465,
+ "learning_rate": 0.0006,
+ "loss": 4.434499740600586,
+ "step": 4434
+ },
+ {
+ "epoch": 61.60087336244541,
+ "grad_norm": 0.01791365258395672,
+ "learning_rate": 0.0006,
+ "loss": 4.466872215270996,
+ "step": 4435
+ },
+ {
+ "epoch": 61.61484716157205,
+ "grad_norm": 0.01879458874464035,
+ "learning_rate": 0.0006,
+ "loss": 4.5176520347595215,
+ "step": 4436
+ },
+ {
+ "epoch": 61.62882096069869,
+ "grad_norm": 0.016371937468647957,
+ "learning_rate": 0.0006,
+ "loss": 4.3507585525512695,
+ "step": 4437
+ },
+ {
+ "epoch": 61.64279475982533,
+ "grad_norm": 0.01648143120110035,
+ "learning_rate": 0.0006,
+ "loss": 4.398778438568115,
+ "step": 4438
+ },
+ {
+ "epoch": 61.65676855895197,
+ "grad_norm": 0.017240576446056366,
+ "learning_rate": 0.0006,
+ "loss": 4.513933181762695,
+ "step": 4439
+ },
+ {
+ "epoch": 61.670742358078606,
+ "grad_norm": 0.01734284684062004,
+ "learning_rate": 0.0006,
+ "loss": 4.315343856811523,
+ "step": 4440
+ },
+ {
+ "epoch": 61.68471615720524,
+ "grad_norm": 0.01639772206544876,
+ "learning_rate": 0.0006,
+ "loss": 4.4641218185424805,
+ "step": 4441
+ },
+ {
+ "epoch": 61.698689956331876,
+ "grad_norm": 0.0164708960801363,
+ "learning_rate": 0.0006,
+ "loss": 4.439968585968018,
+ "step": 4442
+ },
+ {
+ "epoch": 61.712663755458514,
+ "grad_norm": 0.01757933758199215,
+ "learning_rate": 0.0006,
+ "loss": 4.408327102661133,
+ "step": 4443
+ },
+ {
+ "epoch": 61.72663755458515,
+ "grad_norm": 0.01789589412510395,
+ "learning_rate": 0.0006,
+ "loss": 4.478257179260254,
+ "step": 4444
+ },
+ {
+ "epoch": 61.74061135371179,
+ "grad_norm": 0.01822184957563877,
+ "learning_rate": 0.0006,
+ "loss": 4.412046909332275,
+ "step": 4445
+ },
+ {
+ "epoch": 61.75458515283843,
+ "grad_norm": 0.018624618649482727,
+ "learning_rate": 0.0006,
+ "loss": 4.427122116088867,
+ "step": 4446
+ },
+ {
+ "epoch": 61.76855895196506,
+ "grad_norm": 0.018674887716770172,
+ "learning_rate": 0.0006,
+ "loss": 4.429147720336914,
+ "step": 4447
+ },
+ {
+ "epoch": 61.7825327510917,
+ "grad_norm": 0.015317119657993317,
+ "learning_rate": 0.0006,
+ "loss": 4.2808451652526855,
+ "step": 4448
+ },
+ {
+ "epoch": 61.79650655021834,
+ "grad_norm": 0.015400845557451248,
+ "learning_rate": 0.0006,
+ "loss": 4.420193672180176,
+ "step": 4449
+ },
+ {
+ "epoch": 61.81048034934498,
+ "grad_norm": 0.016743650659918785,
+ "learning_rate": 0.0006,
+ "loss": 4.498254776000977,
+ "step": 4450
+ },
+ {
+ "epoch": 61.82445414847162,
+ "grad_norm": 0.014330781064927578,
+ "learning_rate": 0.0006,
+ "loss": 4.445384979248047,
+ "step": 4451
+ },
+ {
+ "epoch": 61.838427947598255,
+ "grad_norm": 0.015046479180455208,
+ "learning_rate": 0.0006,
+ "loss": 4.368041038513184,
+ "step": 4452
+ },
+ {
+ "epoch": 61.852401746724894,
+ "grad_norm": 0.01630168780684471,
+ "learning_rate": 0.0006,
+ "loss": 4.460142135620117,
+ "step": 4453
+ },
+ {
+ "epoch": 61.866375545851525,
+ "grad_norm": 0.01637885719537735,
+ "learning_rate": 0.0006,
+ "loss": 4.416606426239014,
+ "step": 4454
+ },
+ {
+ "epoch": 61.880349344978164,
+ "grad_norm": 0.016785888001322746,
+ "learning_rate": 0.0006,
+ "loss": 4.452545642852783,
+ "step": 4455
+ },
+ {
+ "epoch": 61.8943231441048,
+ "grad_norm": 0.015044555068016052,
+ "learning_rate": 0.0006,
+ "loss": 4.427334785461426,
+ "step": 4456
+ },
+ {
+ "epoch": 61.90829694323144,
+ "grad_norm": 0.013699211180210114,
+ "learning_rate": 0.0006,
+ "loss": 4.408761024475098,
+ "step": 4457
+ },
+ {
+ "epoch": 61.92227074235808,
+ "grad_norm": 0.014021803624927998,
+ "learning_rate": 0.0006,
+ "loss": 4.543607711791992,
+ "step": 4458
+ },
+ {
+ "epoch": 61.93624454148472,
+ "grad_norm": 0.015438460744917393,
+ "learning_rate": 0.0006,
+ "loss": 4.313363075256348,
+ "step": 4459
+ },
+ {
+ "epoch": 61.95021834061135,
+ "grad_norm": 0.017515428364276886,
+ "learning_rate": 0.0006,
+ "loss": 4.310412406921387,
+ "step": 4460
+ },
+ {
+ "epoch": 61.96419213973799,
+ "grad_norm": 0.016683880239725113,
+ "learning_rate": 0.0006,
+ "loss": 4.3378448486328125,
+ "step": 4461
+ },
+ {
+ "epoch": 61.97816593886463,
+ "grad_norm": 0.019384048879146576,
+ "learning_rate": 0.0006,
+ "loss": 4.450568675994873,
+ "step": 4462
+ },
+ {
+ "epoch": 61.992139737991266,
+ "grad_norm": 0.020651493221521378,
+ "learning_rate": 0.0006,
+ "loss": 4.453197479248047,
+ "step": 4463
+ },
+ {
+ "epoch": 62.0,
+ "grad_norm": 0.020991764962673187,
+ "learning_rate": 0.0006,
+ "loss": 4.441527366638184,
+ "step": 4464
+ },
+ {
+ "epoch": 62.0,
+ "eval_loss": 4.700193881988525,
+ "eval_runtime": 56.6098,
+ "eval_samples_per_second": 43.137,
+ "eval_steps_per_second": 1.36,
+ "step": 4464
+ },
+ {
+ "epoch": 62.01397379912664,
+ "grad_norm": 0.01811373233795166,
+ "learning_rate": 0.0006,
+ "loss": 4.37346076965332,
+ "step": 4465
+ },
+ {
+ "epoch": 62.02794759825328,
+ "grad_norm": 0.01667127199470997,
+ "learning_rate": 0.0006,
+ "loss": 4.237151145935059,
+ "step": 4466
+ },
+ {
+ "epoch": 62.041921397379916,
+ "grad_norm": 0.014971534721553326,
+ "learning_rate": 0.0006,
+ "loss": 4.367053031921387,
+ "step": 4467
+ },
+ {
+ "epoch": 62.05589519650655,
+ "grad_norm": 0.015655893832445145,
+ "learning_rate": 0.0006,
+ "loss": 4.437426567077637,
+ "step": 4468
+ },
+ {
+ "epoch": 62.069868995633186,
+ "grad_norm": 0.014935447834432125,
+ "learning_rate": 0.0006,
+ "loss": 4.3635125160217285,
+ "step": 4469
+ },
+ {
+ "epoch": 62.083842794759825,
+ "grad_norm": 0.015363802202045918,
+ "learning_rate": 0.0006,
+ "loss": 4.40492057800293,
+ "step": 4470
+ },
+ {
+ "epoch": 62.09781659388646,
+ "grad_norm": 0.0159201230853796,
+ "learning_rate": 0.0006,
+ "loss": 4.392058849334717,
+ "step": 4471
+ },
+ {
+ "epoch": 62.1117903930131,
+ "grad_norm": 0.014598241075873375,
+ "learning_rate": 0.0006,
+ "loss": 4.369944095611572,
+ "step": 4472
+ },
+ {
+ "epoch": 62.12576419213974,
+ "grad_norm": 0.015394841320812702,
+ "learning_rate": 0.0006,
+ "loss": 4.425097942352295,
+ "step": 4473
+ },
+ {
+ "epoch": 62.13973799126637,
+ "grad_norm": 0.015913628041744232,
+ "learning_rate": 0.0006,
+ "loss": 4.420596122741699,
+ "step": 4474
+ },
+ {
+ "epoch": 62.15371179039301,
+ "grad_norm": 0.01575258933007717,
+ "learning_rate": 0.0006,
+ "loss": 4.32290506362915,
+ "step": 4475
+ },
+ {
+ "epoch": 62.16768558951965,
+ "grad_norm": 0.0155344782397151,
+ "learning_rate": 0.0006,
+ "loss": 4.273068428039551,
+ "step": 4476
+ },
+ {
+ "epoch": 62.18165938864629,
+ "grad_norm": 0.014947040006518364,
+ "learning_rate": 0.0006,
+ "loss": 4.384868621826172,
+ "step": 4477
+ },
+ {
+ "epoch": 62.19563318777293,
+ "grad_norm": 0.01557591650635004,
+ "learning_rate": 0.0006,
+ "loss": 4.458410263061523,
+ "step": 4478
+ },
+ {
+ "epoch": 62.209606986899566,
+ "grad_norm": 0.015182198956608772,
+ "learning_rate": 0.0006,
+ "loss": 4.300738334655762,
+ "step": 4479
+ },
+ {
+ "epoch": 62.223580786026204,
+ "grad_norm": 0.01634717360138893,
+ "learning_rate": 0.0006,
+ "loss": 4.436875820159912,
+ "step": 4480
+ },
+ {
+ "epoch": 62.237554585152836,
+ "grad_norm": 0.018813718110322952,
+ "learning_rate": 0.0006,
+ "loss": 4.454665184020996,
+ "step": 4481
+ },
+ {
+ "epoch": 62.251528384279474,
+ "grad_norm": 0.020965630188584328,
+ "learning_rate": 0.0006,
+ "loss": 4.357341766357422,
+ "step": 4482
+ },
+ {
+ "epoch": 62.26550218340611,
+ "grad_norm": 0.020738353952765465,
+ "learning_rate": 0.0006,
+ "loss": 4.456644058227539,
+ "step": 4483
+ },
+ {
+ "epoch": 62.27947598253275,
+ "grad_norm": 0.018813546746969223,
+ "learning_rate": 0.0006,
+ "loss": 4.381941795349121,
+ "step": 4484
+ },
+ {
+ "epoch": 62.29344978165939,
+ "grad_norm": 0.015431761741638184,
+ "learning_rate": 0.0006,
+ "loss": 4.41126823425293,
+ "step": 4485
+ },
+ {
+ "epoch": 62.30742358078603,
+ "grad_norm": 0.016335977241396904,
+ "learning_rate": 0.0006,
+ "loss": 4.419488906860352,
+ "step": 4486
+ },
+ {
+ "epoch": 62.32139737991266,
+ "grad_norm": 0.01832406409084797,
+ "learning_rate": 0.0006,
+ "loss": 4.579026222229004,
+ "step": 4487
+ },
+ {
+ "epoch": 62.3353711790393,
+ "grad_norm": 0.01790742017328739,
+ "learning_rate": 0.0006,
+ "loss": 4.289964199066162,
+ "step": 4488
+ },
+ {
+ "epoch": 62.34934497816594,
+ "grad_norm": 0.017266202718019485,
+ "learning_rate": 0.0006,
+ "loss": 4.415959358215332,
+ "step": 4489
+ },
+ {
+ "epoch": 62.36331877729258,
+ "grad_norm": 0.015976132825016975,
+ "learning_rate": 0.0006,
+ "loss": 4.358587265014648,
+ "step": 4490
+ },
+ {
+ "epoch": 62.377292576419215,
+ "grad_norm": 0.016873881220817566,
+ "learning_rate": 0.0006,
+ "loss": 4.354607105255127,
+ "step": 4491
+ },
+ {
+ "epoch": 62.391266375545854,
+ "grad_norm": 0.01476745493710041,
+ "learning_rate": 0.0006,
+ "loss": 4.384904384613037,
+ "step": 4492
+ },
+ {
+ "epoch": 62.40524017467249,
+ "grad_norm": 0.013810782693326473,
+ "learning_rate": 0.0006,
+ "loss": 4.459230899810791,
+ "step": 4493
+ },
+ {
+ "epoch": 62.419213973799124,
+ "grad_norm": 0.015150421299040318,
+ "learning_rate": 0.0006,
+ "loss": 4.369471549987793,
+ "step": 4494
+ },
+ {
+ "epoch": 62.43318777292576,
+ "grad_norm": 0.014539610594511032,
+ "learning_rate": 0.0006,
+ "loss": 4.3564982414245605,
+ "step": 4495
+ },
+ {
+ "epoch": 62.4471615720524,
+ "grad_norm": 0.015554017387330532,
+ "learning_rate": 0.0006,
+ "loss": 4.414379119873047,
+ "step": 4496
+ },
+ {
+ "epoch": 62.46113537117904,
+ "grad_norm": 0.016608357429504395,
+ "learning_rate": 0.0006,
+ "loss": 4.311585426330566,
+ "step": 4497
+ },
+ {
+ "epoch": 62.47510917030568,
+ "grad_norm": 0.01885881833732128,
+ "learning_rate": 0.0006,
+ "loss": 4.306629180908203,
+ "step": 4498
+ },
+ {
+ "epoch": 62.48908296943232,
+ "grad_norm": 0.022128663957118988,
+ "learning_rate": 0.0006,
+ "loss": 4.345483779907227,
+ "step": 4499
+ },
+ {
+ "epoch": 62.50305676855895,
+ "grad_norm": 0.023401018232107162,
+ "learning_rate": 0.0006,
+ "loss": 4.398204326629639,
+ "step": 4500
+ },
+ {
+ "epoch": 62.51703056768559,
+ "grad_norm": 0.022080009803175926,
+ "learning_rate": 0.0006,
+ "loss": 4.375027179718018,
+ "step": 4501
+ },
+ {
+ "epoch": 62.531004366812226,
+ "grad_norm": 0.020004864782094955,
+ "learning_rate": 0.0006,
+ "loss": 4.33494758605957,
+ "step": 4502
+ },
+ {
+ "epoch": 62.544978165938865,
+ "grad_norm": 0.01786700263619423,
+ "learning_rate": 0.0006,
+ "loss": 4.4428300857543945,
+ "step": 4503
+ },
+ {
+ "epoch": 62.5589519650655,
+ "grad_norm": 0.01885388419032097,
+ "learning_rate": 0.0006,
+ "loss": 4.4138946533203125,
+ "step": 4504
+ },
+ {
+ "epoch": 62.57292576419214,
+ "grad_norm": 0.019890939816832542,
+ "learning_rate": 0.0006,
+ "loss": 4.278632164001465,
+ "step": 4505
+ },
+ {
+ "epoch": 62.58689956331878,
+ "grad_norm": 0.02109031192958355,
+ "learning_rate": 0.0006,
+ "loss": 4.3756327629089355,
+ "step": 4506
+ },
+ {
+ "epoch": 62.60087336244541,
+ "grad_norm": 0.020813899114727974,
+ "learning_rate": 0.0006,
+ "loss": 4.410213470458984,
+ "step": 4507
+ },
+ {
+ "epoch": 62.61484716157205,
+ "grad_norm": 0.021655689924955368,
+ "learning_rate": 0.0006,
+ "loss": 4.30633020401001,
+ "step": 4508
+ },
+ {
+ "epoch": 62.62882096069869,
+ "grad_norm": 0.018155649304389954,
+ "learning_rate": 0.0006,
+ "loss": 4.407284736633301,
+ "step": 4509
+ },
+ {
+ "epoch": 62.64279475982533,
+ "grad_norm": 0.017765069380402565,
+ "learning_rate": 0.0006,
+ "loss": 4.321715354919434,
+ "step": 4510
+ },
+ {
+ "epoch": 62.65676855895197,
+ "grad_norm": 0.017724554985761642,
+ "learning_rate": 0.0006,
+ "loss": 4.493371963500977,
+ "step": 4511
+ },
+ {
+ "epoch": 62.670742358078606,
+ "grad_norm": 0.01492092851549387,
+ "learning_rate": 0.0006,
+ "loss": 4.443025588989258,
+ "step": 4512
+ },
+ {
+ "epoch": 62.68471615720524,
+ "grad_norm": 0.01615324057638645,
+ "learning_rate": 0.0006,
+ "loss": 4.3289594650268555,
+ "step": 4513
+ },
+ {
+ "epoch": 62.698689956331876,
+ "grad_norm": 0.016569366678595543,
+ "learning_rate": 0.0006,
+ "loss": 4.369720458984375,
+ "step": 4514
+ },
+ {
+ "epoch": 62.712663755458514,
+ "grad_norm": 0.016509901732206345,
+ "learning_rate": 0.0006,
+ "loss": 4.292316436767578,
+ "step": 4515
+ },
+ {
+ "epoch": 62.72663755458515,
+ "grad_norm": 0.014270003885030746,
+ "learning_rate": 0.0006,
+ "loss": 4.349854469299316,
+ "step": 4516
+ },
+ {
+ "epoch": 62.74061135371179,
+ "grad_norm": 0.014235340990126133,
+ "learning_rate": 0.0006,
+ "loss": 4.473727226257324,
+ "step": 4517
+ },
+ {
+ "epoch": 62.75458515283843,
+ "grad_norm": 0.015163225121796131,
+ "learning_rate": 0.0006,
+ "loss": 4.452878475189209,
+ "step": 4518
+ },
+ {
+ "epoch": 62.76855895196506,
+ "grad_norm": 0.015748055651783943,
+ "learning_rate": 0.0006,
+ "loss": 4.521087169647217,
+ "step": 4519
+ },
+ {
+ "epoch": 62.7825327510917,
+ "grad_norm": 0.01661108434200287,
+ "learning_rate": 0.0006,
+ "loss": 4.443885326385498,
+ "step": 4520
+ },
+ {
+ "epoch": 62.79650655021834,
+ "grad_norm": 0.01610301062464714,
+ "learning_rate": 0.0006,
+ "loss": 4.312990188598633,
+ "step": 4521
+ },
+ {
+ "epoch": 62.81048034934498,
+ "grad_norm": 0.017007583752274513,
+ "learning_rate": 0.0006,
+ "loss": 4.408937454223633,
+ "step": 4522
+ },
+ {
+ "epoch": 62.82445414847162,
+ "grad_norm": 0.01749461516737938,
+ "learning_rate": 0.0006,
+ "loss": 4.346066951751709,
+ "step": 4523
+ },
+ {
+ "epoch": 62.838427947598255,
+ "grad_norm": 0.018243493512272835,
+ "learning_rate": 0.0006,
+ "loss": 4.357658386230469,
+ "step": 4524
+ },
+ {
+ "epoch": 62.852401746724894,
+ "grad_norm": 0.017499985173344612,
+ "learning_rate": 0.0006,
+ "loss": 4.374061584472656,
+ "step": 4525
+ },
+ {
+ "epoch": 62.866375545851525,
+ "grad_norm": 0.014536325819790363,
+ "learning_rate": 0.0006,
+ "loss": 4.189321994781494,
+ "step": 4526
+ },
+ {
+ "epoch": 62.880349344978164,
+ "grad_norm": 0.014728891663253307,
+ "learning_rate": 0.0006,
+ "loss": 4.425910949707031,
+ "step": 4527
+ },
+ {
+ "epoch": 62.8943231441048,
+ "grad_norm": 0.015013772062957287,
+ "learning_rate": 0.0006,
+ "loss": 4.337319850921631,
+ "step": 4528
+ },
+ {
+ "epoch": 62.90829694323144,
+ "grad_norm": 0.01626027375459671,
+ "learning_rate": 0.0006,
+ "loss": 4.330339431762695,
+ "step": 4529
+ },
+ {
+ "epoch": 62.92227074235808,
+ "grad_norm": 0.014773444272577763,
+ "learning_rate": 0.0006,
+ "loss": 4.4188737869262695,
+ "step": 4530
+ },
+ {
+ "epoch": 62.93624454148472,
+ "grad_norm": 0.01422932744026184,
+ "learning_rate": 0.0006,
+ "loss": 4.355986595153809,
+ "step": 4531
+ },
+ {
+ "epoch": 62.95021834061135,
+ "grad_norm": 0.012170110829174519,
+ "learning_rate": 0.0006,
+ "loss": 4.405722618103027,
+ "step": 4532
+ },
+ {
+ "epoch": 62.96419213973799,
+ "grad_norm": 0.013233253732323647,
+ "learning_rate": 0.0006,
+ "loss": 4.377286911010742,
+ "step": 4533
+ },
+ {
+ "epoch": 62.97816593886463,
+ "grad_norm": 0.014497299678623676,
+ "learning_rate": 0.0006,
+ "loss": 4.508796691894531,
+ "step": 4534
+ },
+ {
+ "epoch": 62.992139737991266,
+ "grad_norm": 0.014718741178512573,
+ "learning_rate": 0.0006,
+ "loss": 4.370919704437256,
+ "step": 4535
+ },
+ {
+ "epoch": 63.0,
+ "grad_norm": 0.015833502635359764,
+ "learning_rate": 0.0006,
+ "loss": 4.258753776550293,
+ "step": 4536
+ },
+ {
+ "epoch": 63.0,
+ "eval_loss": 4.714163303375244,
+ "eval_runtime": 57.0269,
+ "eval_samples_per_second": 42.822,
+ "eval_steps_per_second": 1.35,
+ "step": 4536
+ },
+ {
+ "epoch": 63.01397379912664,
+ "grad_norm": 0.014277510344982147,
+ "learning_rate": 0.0006,
+ "loss": 4.3112640380859375,
+ "step": 4537
+ },
+ {
+ "epoch": 63.02794759825328,
+ "grad_norm": 0.015377935022115707,
+ "learning_rate": 0.0006,
+ "loss": 4.342782974243164,
+ "step": 4538
+ },
+ {
+ "epoch": 63.041921397379916,
+ "grad_norm": 0.015406244434416294,
+ "learning_rate": 0.0006,
+ "loss": 4.474056720733643,
+ "step": 4539
+ },
+ {
+ "epoch": 63.05589519650655,
+ "grad_norm": 0.018121499568223953,
+ "learning_rate": 0.0006,
+ "loss": 4.397963523864746,
+ "step": 4540
+ },
+ {
+ "epoch": 63.069868995633186,
+ "grad_norm": 0.022634636610746384,
+ "learning_rate": 0.0006,
+ "loss": 4.306631088256836,
+ "step": 4541
+ },
+ {
+ "epoch": 63.083842794759825,
+ "grad_norm": 0.026448970660567284,
+ "learning_rate": 0.0006,
+ "loss": 4.283175468444824,
+ "step": 4542
+ },
+ {
+ "epoch": 63.09781659388646,
+ "grad_norm": 0.024104537442326546,
+ "learning_rate": 0.0006,
+ "loss": 4.315826416015625,
+ "step": 4543
+ },
+ {
+ "epoch": 63.1117903930131,
+ "grad_norm": 0.019040480256080627,
+ "learning_rate": 0.0006,
+ "loss": 4.44057035446167,
+ "step": 4544
+ },
+ {
+ "epoch": 63.12576419213974,
+ "grad_norm": 0.02006000280380249,
+ "learning_rate": 0.0006,
+ "loss": 4.405603408813477,
+ "step": 4545
+ },
+ {
+ "epoch": 63.13973799126637,
+ "grad_norm": 0.019980071112513542,
+ "learning_rate": 0.0006,
+ "loss": 4.393726348876953,
+ "step": 4546
+ },
+ {
+ "epoch": 63.15371179039301,
+ "grad_norm": 0.01641865260899067,
+ "learning_rate": 0.0006,
+ "loss": 4.272546768188477,
+ "step": 4547
+ },
+ {
+ "epoch": 63.16768558951965,
+ "grad_norm": 0.01890353485941887,
+ "learning_rate": 0.0006,
+ "loss": 4.382566928863525,
+ "step": 4548
+ },
+ {
+ "epoch": 63.18165938864629,
+ "grad_norm": 0.018370937556028366,
+ "learning_rate": 0.0006,
+ "loss": 4.423953533172607,
+ "step": 4549
+ },
+ {
+ "epoch": 63.19563318777293,
+ "grad_norm": 0.014985294081270695,
+ "learning_rate": 0.0006,
+ "loss": 4.308627128601074,
+ "step": 4550
+ },
+ {
+ "epoch": 63.209606986899566,
+ "grad_norm": 0.017513327300548553,
+ "learning_rate": 0.0006,
+ "loss": 4.248819351196289,
+ "step": 4551
+ },
+ {
+ "epoch": 63.223580786026204,
+ "grad_norm": 0.019464161247015,
+ "learning_rate": 0.0006,
+ "loss": 4.359729766845703,
+ "step": 4552
+ },
+ {
+ "epoch": 63.237554585152836,
+ "grad_norm": 0.018678035587072372,
+ "learning_rate": 0.0006,
+ "loss": 4.3299665451049805,
+ "step": 4553
+ },
+ {
+ "epoch": 63.251528384279474,
+ "grad_norm": 0.017909523099660873,
+ "learning_rate": 0.0006,
+ "loss": 4.246147155761719,
+ "step": 4554
+ },
+ {
+ "epoch": 63.26550218340611,
+ "grad_norm": 0.01747293211519718,
+ "learning_rate": 0.0006,
+ "loss": 4.41165018081665,
+ "step": 4555
+ },
+ {
+ "epoch": 63.27947598253275,
+ "grad_norm": 0.018670853227376938,
+ "learning_rate": 0.0006,
+ "loss": 4.409883499145508,
+ "step": 4556
+ },
+ {
+ "epoch": 63.29344978165939,
+ "grad_norm": 0.0189529862254858,
+ "learning_rate": 0.0006,
+ "loss": 4.418654918670654,
+ "step": 4557
+ },
+ {
+ "epoch": 63.30742358078603,
+ "grad_norm": 0.018394414335489273,
+ "learning_rate": 0.0006,
+ "loss": 4.38355827331543,
+ "step": 4558
+ },
+ {
+ "epoch": 63.32139737991266,
+ "grad_norm": 0.018317820504307747,
+ "learning_rate": 0.0006,
+ "loss": 4.456843376159668,
+ "step": 4559
+ },
+ {
+ "epoch": 63.3353711790393,
+ "grad_norm": 0.019225597381591797,
+ "learning_rate": 0.0006,
+ "loss": 4.232264518737793,
+ "step": 4560
+ },
+ {
+ "epoch": 63.34934497816594,
+ "grad_norm": 0.016663681715726852,
+ "learning_rate": 0.0006,
+ "loss": 4.393517017364502,
+ "step": 4561
+ },
+ {
+ "epoch": 63.36331877729258,
+ "grad_norm": 0.014975365251302719,
+ "learning_rate": 0.0006,
+ "loss": 4.375183582305908,
+ "step": 4562
+ },
+ {
+ "epoch": 63.377292576419215,
+ "grad_norm": 0.01771732047200203,
+ "learning_rate": 0.0006,
+ "loss": 4.3073410987854,
+ "step": 4563
+ },
+ {
+ "epoch": 63.391266375545854,
+ "grad_norm": 0.01663549244403839,
+ "learning_rate": 0.0006,
+ "loss": 4.311003684997559,
+ "step": 4564
+ },
+ {
+ "epoch": 63.40524017467249,
+ "grad_norm": 0.016301177442073822,
+ "learning_rate": 0.0006,
+ "loss": 4.326807022094727,
+ "step": 4565
+ },
+ {
+ "epoch": 63.419213973799124,
+ "grad_norm": 0.016781188547611237,
+ "learning_rate": 0.0006,
+ "loss": 4.306562423706055,
+ "step": 4566
+ },
+ {
+ "epoch": 63.43318777292576,
+ "grad_norm": 0.01767745055258274,
+ "learning_rate": 0.0006,
+ "loss": 4.367082595825195,
+ "step": 4567
+ },
+ {
+ "epoch": 63.4471615720524,
+ "grad_norm": 0.016728565096855164,
+ "learning_rate": 0.0006,
+ "loss": 4.327400207519531,
+ "step": 4568
+ },
+ {
+ "epoch": 63.46113537117904,
+ "grad_norm": 0.016289403662085533,
+ "learning_rate": 0.0006,
+ "loss": 4.392712116241455,
+ "step": 4569
+ },
+ {
+ "epoch": 63.47510917030568,
+ "grad_norm": 0.016233716160058975,
+ "learning_rate": 0.0006,
+ "loss": 4.354372024536133,
+ "step": 4570
+ },
+ {
+ "epoch": 63.48908296943232,
+ "grad_norm": 0.016547974199056625,
+ "learning_rate": 0.0006,
+ "loss": 4.309050559997559,
+ "step": 4571
+ },
+ {
+ "epoch": 63.50305676855895,
+ "grad_norm": 0.018466509878635406,
+ "learning_rate": 0.0006,
+ "loss": 4.318345069885254,
+ "step": 4572
+ },
+ {
+ "epoch": 63.51703056768559,
+ "grad_norm": 0.017685720697045326,
+ "learning_rate": 0.0006,
+ "loss": 4.399726867675781,
+ "step": 4573
+ },
+ {
+ "epoch": 63.531004366812226,
+ "grad_norm": 0.015801234170794487,
+ "learning_rate": 0.0006,
+ "loss": 4.328495502471924,
+ "step": 4574
+ },
+ {
+ "epoch": 63.544978165938865,
+ "grad_norm": 0.0188873540610075,
+ "learning_rate": 0.0006,
+ "loss": 4.253244400024414,
+ "step": 4575
+ },
+ {
+ "epoch": 63.5589519650655,
+ "grad_norm": 0.01868271268904209,
+ "learning_rate": 0.0006,
+ "loss": 4.463746070861816,
+ "step": 4576
+ },
+ {
+ "epoch": 63.57292576419214,
+ "grad_norm": 0.01844949834048748,
+ "learning_rate": 0.0006,
+ "loss": 4.400300025939941,
+ "step": 4577
+ },
+ {
+ "epoch": 63.58689956331878,
+ "grad_norm": 0.015256480313837528,
+ "learning_rate": 0.0006,
+ "loss": 4.445844650268555,
+ "step": 4578
+ },
+ {
+ "epoch": 63.60087336244541,
+ "grad_norm": 0.015305760316550732,
+ "learning_rate": 0.0006,
+ "loss": 4.317327499389648,
+ "step": 4579
+ },
+ {
+ "epoch": 63.61484716157205,
+ "grad_norm": 0.015200769528746605,
+ "learning_rate": 0.0006,
+ "loss": 4.25112247467041,
+ "step": 4580
+ },
+ {
+ "epoch": 63.62882096069869,
+ "grad_norm": 0.015919538214802742,
+ "learning_rate": 0.0006,
+ "loss": 4.277937889099121,
+ "step": 4581
+ },
+ {
+ "epoch": 63.64279475982533,
+ "grad_norm": 0.01572665013372898,
+ "learning_rate": 0.0006,
+ "loss": 4.302237510681152,
+ "step": 4582
+ },
+ {
+ "epoch": 63.65676855895197,
+ "grad_norm": 0.016428731381893158,
+ "learning_rate": 0.0006,
+ "loss": 4.36762809753418,
+ "step": 4583
+ },
+ {
+ "epoch": 63.670742358078606,
+ "grad_norm": 0.01706082746386528,
+ "learning_rate": 0.0006,
+ "loss": 4.236526966094971,
+ "step": 4584
+ },
+ {
+ "epoch": 63.68471615720524,
+ "grad_norm": 0.017850136384367943,
+ "learning_rate": 0.0006,
+ "loss": 4.264129638671875,
+ "step": 4585
+ },
+ {
+ "epoch": 63.698689956331876,
+ "grad_norm": 0.016079438850283623,
+ "learning_rate": 0.0006,
+ "loss": 4.449799060821533,
+ "step": 4586
+ },
+ {
+ "epoch": 63.712663755458514,
+ "grad_norm": 0.014996383339166641,
+ "learning_rate": 0.0006,
+ "loss": 4.404574394226074,
+ "step": 4587
+ },
+ {
+ "epoch": 63.72663755458515,
+ "grad_norm": 0.015910431742668152,
+ "learning_rate": 0.0006,
+ "loss": 4.348101615905762,
+ "step": 4588
+ },
+ {
+ "epoch": 63.74061135371179,
+ "grad_norm": 0.01630021072924137,
+ "learning_rate": 0.0006,
+ "loss": 4.3193817138671875,
+ "step": 4589
+ },
+ {
+ "epoch": 63.75458515283843,
+ "grad_norm": 0.014571773819625378,
+ "learning_rate": 0.0006,
+ "loss": 4.374402046203613,
+ "step": 4590
+ },
+ {
+ "epoch": 63.76855895196506,
+ "grad_norm": 0.01662345975637436,
+ "learning_rate": 0.0006,
+ "loss": 4.329075813293457,
+ "step": 4591
+ },
+ {
+ "epoch": 63.7825327510917,
+ "grad_norm": 0.016604198142886162,
+ "learning_rate": 0.0006,
+ "loss": 4.436126232147217,
+ "step": 4592
+ },
+ {
+ "epoch": 63.79650655021834,
+ "grad_norm": 0.016578884795308113,
+ "learning_rate": 0.0006,
+ "loss": 4.361468315124512,
+ "step": 4593
+ },
+ {
+ "epoch": 63.81048034934498,
+ "grad_norm": 0.016339825466275215,
+ "learning_rate": 0.0006,
+ "loss": 4.449892044067383,
+ "step": 4594
+ },
+ {
+ "epoch": 63.82445414847162,
+ "grad_norm": 0.016805455088615417,
+ "learning_rate": 0.0006,
+ "loss": 4.415866851806641,
+ "step": 4595
+ },
+ {
+ "epoch": 63.838427947598255,
+ "grad_norm": 0.017366213724017143,
+ "learning_rate": 0.0006,
+ "loss": 4.361730575561523,
+ "step": 4596
+ },
+ {
+ "epoch": 63.852401746724894,
+ "grad_norm": 0.015599003992974758,
+ "learning_rate": 0.0006,
+ "loss": 4.363637924194336,
+ "step": 4597
+ },
+ {
+ "epoch": 63.866375545851525,
+ "grad_norm": 0.015674632042646408,
+ "learning_rate": 0.0006,
+ "loss": 4.332045078277588,
+ "step": 4598
+ },
+ {
+ "epoch": 63.880349344978164,
+ "grad_norm": 0.015350443311035633,
+ "learning_rate": 0.0006,
+ "loss": 4.320063591003418,
+ "step": 4599
+ },
+ {
+ "epoch": 63.8943231441048,
+ "grad_norm": 0.015388146974146366,
+ "learning_rate": 0.0006,
+ "loss": 4.398199081420898,
+ "step": 4600
+ },
+ {
+ "epoch": 63.90829694323144,
+ "grad_norm": 0.01450140681117773,
+ "learning_rate": 0.0006,
+ "loss": 4.244076728820801,
+ "step": 4601
+ },
+ {
+ "epoch": 63.92227074235808,
+ "grad_norm": 0.01637491025030613,
+ "learning_rate": 0.0006,
+ "loss": 4.213549613952637,
+ "step": 4602
+ },
+ {
+ "epoch": 63.93624454148472,
+ "grad_norm": 0.0159814041107893,
+ "learning_rate": 0.0006,
+ "loss": 4.421775817871094,
+ "step": 4603
+ },
+ {
+ "epoch": 63.95021834061135,
+ "grad_norm": 0.017386827617883682,
+ "learning_rate": 0.0006,
+ "loss": 4.340023994445801,
+ "step": 4604
+ },
+ {
+ "epoch": 63.96419213973799,
+ "grad_norm": 0.019679656252264977,
+ "learning_rate": 0.0006,
+ "loss": 4.352534294128418,
+ "step": 4605
+ },
+ {
+ "epoch": 63.97816593886463,
+ "grad_norm": 0.022391516715288162,
+ "learning_rate": 0.0006,
+ "loss": 4.3538312911987305,
+ "step": 4606
+ },
+ {
+ "epoch": 63.992139737991266,
+ "grad_norm": 0.023281006142497063,
+ "learning_rate": 0.0006,
+ "loss": 4.432089328765869,
+ "step": 4607
+ },
+ {
+ "epoch": 64.0,
+ "grad_norm": 0.019916288554668427,
+ "learning_rate": 0.0006,
+ "loss": 4.466937065124512,
+ "step": 4608
+ },
+ {
+ "epoch": 64.0,
+ "eval_loss": 4.659008502960205,
+ "eval_runtime": 56.0019,
+ "eval_samples_per_second": 43.606,
+ "eval_steps_per_second": 1.375,
+ "step": 4608
+ },
+ {
+ "epoch": 64.01397379912663,
+ "grad_norm": 0.017351187765598297,
+ "learning_rate": 0.0006,
+ "loss": 4.294129371643066,
+ "step": 4609
+ },
+ {
+ "epoch": 64.02794759825328,
+ "grad_norm": 0.02405865117907524,
+ "learning_rate": 0.0006,
+ "loss": 4.266717910766602,
+ "step": 4610
+ },
+ {
+ "epoch": 64.04192139737991,
+ "grad_norm": 0.022158386185765266,
+ "learning_rate": 0.0006,
+ "loss": 4.395932197570801,
+ "step": 4611
+ },
+ {
+ "epoch": 64.05589519650655,
+ "grad_norm": 0.019238505512475967,
+ "learning_rate": 0.0006,
+ "loss": 4.266704082489014,
+ "step": 4612
+ },
+ {
+ "epoch": 64.06986899563319,
+ "grad_norm": 0.021670425310730934,
+ "learning_rate": 0.0006,
+ "loss": 4.3160552978515625,
+ "step": 4613
+ },
+ {
+ "epoch": 64.08384279475983,
+ "grad_norm": 0.02076641097664833,
+ "learning_rate": 0.0006,
+ "loss": 4.415626525878906,
+ "step": 4614
+ },
+ {
+ "epoch": 64.09781659388646,
+ "grad_norm": 0.018223078921437263,
+ "learning_rate": 0.0006,
+ "loss": 4.292387008666992,
+ "step": 4615
+ },
+ {
+ "epoch": 64.1117903930131,
+ "grad_norm": 0.017401937395334244,
+ "learning_rate": 0.0006,
+ "loss": 4.424524307250977,
+ "step": 4616
+ },
+ {
+ "epoch": 64.12576419213974,
+ "grad_norm": 0.01663743332028389,
+ "learning_rate": 0.0006,
+ "loss": 4.350032806396484,
+ "step": 4617
+ },
+ {
+ "epoch": 64.13973799126637,
+ "grad_norm": 0.018716629594564438,
+ "learning_rate": 0.0006,
+ "loss": 4.309157371520996,
+ "step": 4618
+ },
+ {
+ "epoch": 64.15371179039302,
+ "grad_norm": 0.019569825381040573,
+ "learning_rate": 0.0006,
+ "loss": 4.358292579650879,
+ "step": 4619
+ },
+ {
+ "epoch": 64.16768558951965,
+ "grad_norm": 0.018994422629475594,
+ "learning_rate": 0.0006,
+ "loss": 4.3338823318481445,
+ "step": 4620
+ },
+ {
+ "epoch": 64.18165938864628,
+ "grad_norm": 0.017690688371658325,
+ "learning_rate": 0.0006,
+ "loss": 4.414468765258789,
+ "step": 4621
+ },
+ {
+ "epoch": 64.19563318777293,
+ "grad_norm": 0.0170609001070261,
+ "learning_rate": 0.0006,
+ "loss": 4.4337568283081055,
+ "step": 4622
+ },
+ {
+ "epoch": 64.20960698689956,
+ "grad_norm": 0.016984131187200546,
+ "learning_rate": 0.0006,
+ "loss": 4.304839134216309,
+ "step": 4623
+ },
+ {
+ "epoch": 64.2235807860262,
+ "grad_norm": 0.01619867794215679,
+ "learning_rate": 0.0006,
+ "loss": 4.389623165130615,
+ "step": 4624
+ },
+ {
+ "epoch": 64.23755458515284,
+ "grad_norm": 0.015428110957145691,
+ "learning_rate": 0.0006,
+ "loss": 4.277515888214111,
+ "step": 4625
+ },
+ {
+ "epoch": 64.25152838427948,
+ "grad_norm": 0.015909532085061073,
+ "learning_rate": 0.0006,
+ "loss": 4.266029357910156,
+ "step": 4626
+ },
+ {
+ "epoch": 64.26550218340611,
+ "grad_norm": 0.014296560548245907,
+ "learning_rate": 0.0006,
+ "loss": 4.286067485809326,
+ "step": 4627
+ },
+ {
+ "epoch": 64.27947598253274,
+ "grad_norm": 0.014065971598029137,
+ "learning_rate": 0.0006,
+ "loss": 4.353528022766113,
+ "step": 4628
+ },
+ {
+ "epoch": 64.29344978165939,
+ "grad_norm": 0.014386293478310108,
+ "learning_rate": 0.0006,
+ "loss": 4.318715572357178,
+ "step": 4629
+ },
+ {
+ "epoch": 64.30742358078602,
+ "grad_norm": 0.014001123607158661,
+ "learning_rate": 0.0006,
+ "loss": 4.362497329711914,
+ "step": 4630
+ },
+ {
+ "epoch": 64.32139737991267,
+ "grad_norm": 0.013408979400992393,
+ "learning_rate": 0.0006,
+ "loss": 4.417397499084473,
+ "step": 4631
+ },
+ {
+ "epoch": 64.3353711790393,
+ "grad_norm": 0.014960093423724174,
+ "learning_rate": 0.0006,
+ "loss": 4.359710693359375,
+ "step": 4632
+ },
+ {
+ "epoch": 64.34934497816595,
+ "grad_norm": 0.014658777043223381,
+ "learning_rate": 0.0006,
+ "loss": 4.318822383880615,
+ "step": 4633
+ },
+ {
+ "epoch": 64.36331877729258,
+ "grad_norm": 0.013766744174063206,
+ "learning_rate": 0.0006,
+ "loss": 4.265679836273193,
+ "step": 4634
+ },
+ {
+ "epoch": 64.37729257641921,
+ "grad_norm": 0.013821255415678024,
+ "learning_rate": 0.0006,
+ "loss": 4.3452911376953125,
+ "step": 4635
+ },
+ {
+ "epoch": 64.39126637554585,
+ "grad_norm": 0.013541066087782383,
+ "learning_rate": 0.0006,
+ "loss": 4.429254531860352,
+ "step": 4636
+ },
+ {
+ "epoch": 64.40524017467249,
+ "grad_norm": 0.013985645025968552,
+ "learning_rate": 0.0006,
+ "loss": 4.304452896118164,
+ "step": 4637
+ },
+ {
+ "epoch": 64.41921397379913,
+ "grad_norm": 0.015079808421432972,
+ "learning_rate": 0.0006,
+ "loss": 4.296828746795654,
+ "step": 4638
+ },
+ {
+ "epoch": 64.43318777292576,
+ "grad_norm": 0.014896944165229797,
+ "learning_rate": 0.0006,
+ "loss": 4.299156665802002,
+ "step": 4639
+ },
+ {
+ "epoch": 64.44716157205241,
+ "grad_norm": 0.013770041987299919,
+ "learning_rate": 0.0006,
+ "loss": 4.351439952850342,
+ "step": 4640
+ },
+ {
+ "epoch": 64.46113537117904,
+ "grad_norm": 0.01531178504228592,
+ "learning_rate": 0.0006,
+ "loss": 4.373715877532959,
+ "step": 4641
+ },
+ {
+ "epoch": 64.47510917030567,
+ "grad_norm": 0.014024450443685055,
+ "learning_rate": 0.0006,
+ "loss": 4.368053436279297,
+ "step": 4642
+ },
+ {
+ "epoch": 64.48908296943232,
+ "grad_norm": 0.013864024542272091,
+ "learning_rate": 0.0006,
+ "loss": 4.378564834594727,
+ "step": 4643
+ },
+ {
+ "epoch": 64.50305676855895,
+ "grad_norm": 0.013925445266067982,
+ "learning_rate": 0.0006,
+ "loss": 4.261663436889648,
+ "step": 4644
+ },
+ {
+ "epoch": 64.5170305676856,
+ "grad_norm": 0.015011338517069817,
+ "learning_rate": 0.0006,
+ "loss": 4.337285995483398,
+ "step": 4645
+ },
+ {
+ "epoch": 64.53100436681223,
+ "grad_norm": 0.016941986978054047,
+ "learning_rate": 0.0006,
+ "loss": 4.336877822875977,
+ "step": 4646
+ },
+ {
+ "epoch": 64.54497816593886,
+ "grad_norm": 0.01719909906387329,
+ "learning_rate": 0.0006,
+ "loss": 4.149813175201416,
+ "step": 4647
+ },
+ {
+ "epoch": 64.5589519650655,
+ "grad_norm": 0.01608402095735073,
+ "learning_rate": 0.0006,
+ "loss": 4.191280364990234,
+ "step": 4648
+ },
+ {
+ "epoch": 64.57292576419214,
+ "grad_norm": 0.015284820459783077,
+ "learning_rate": 0.0006,
+ "loss": 4.35953426361084,
+ "step": 4649
+ },
+ {
+ "epoch": 64.58689956331878,
+ "grad_norm": 0.014855682849884033,
+ "learning_rate": 0.0006,
+ "loss": 4.308401107788086,
+ "step": 4650
+ },
+ {
+ "epoch": 64.60087336244541,
+ "grad_norm": 0.017553111538290977,
+ "learning_rate": 0.0006,
+ "loss": 4.279720306396484,
+ "step": 4651
+ },
+ {
+ "epoch": 64.61484716157206,
+ "grad_norm": 0.019506998360157013,
+ "learning_rate": 0.0006,
+ "loss": 4.319524765014648,
+ "step": 4652
+ },
+ {
+ "epoch": 64.62882096069869,
+ "grad_norm": 0.020293425768613815,
+ "learning_rate": 0.0006,
+ "loss": 4.340523719787598,
+ "step": 4653
+ },
+ {
+ "epoch": 64.64279475982532,
+ "grad_norm": 0.01664556749165058,
+ "learning_rate": 0.0006,
+ "loss": 4.296785831451416,
+ "step": 4654
+ },
+ {
+ "epoch": 64.65676855895197,
+ "grad_norm": 0.012592585757374763,
+ "learning_rate": 0.0006,
+ "loss": 4.464595317840576,
+ "step": 4655
+ },
+ {
+ "epoch": 64.6707423580786,
+ "grad_norm": 0.01549589168280363,
+ "learning_rate": 0.0006,
+ "loss": 4.275053024291992,
+ "step": 4656
+ },
+ {
+ "epoch": 64.68471615720524,
+ "grad_norm": 0.01705172471702099,
+ "learning_rate": 0.0006,
+ "loss": 4.388431549072266,
+ "step": 4657
+ },
+ {
+ "epoch": 64.69868995633188,
+ "grad_norm": 0.015438582748174667,
+ "learning_rate": 0.0006,
+ "loss": 4.312070846557617,
+ "step": 4658
+ },
+ {
+ "epoch": 64.71266375545852,
+ "grad_norm": 0.014448083005845547,
+ "learning_rate": 0.0006,
+ "loss": 4.237599849700928,
+ "step": 4659
+ },
+ {
+ "epoch": 64.72663755458515,
+ "grad_norm": 0.015267071314156055,
+ "learning_rate": 0.0006,
+ "loss": 4.29893684387207,
+ "step": 4660
+ },
+ {
+ "epoch": 64.74061135371178,
+ "grad_norm": 0.016615847125649452,
+ "learning_rate": 0.0006,
+ "loss": 4.432273864746094,
+ "step": 4661
+ },
+ {
+ "epoch": 64.75458515283843,
+ "grad_norm": 0.016584362834692,
+ "learning_rate": 0.0006,
+ "loss": 4.202266216278076,
+ "step": 4662
+ },
+ {
+ "epoch": 64.76855895196506,
+ "grad_norm": 0.018313994631171227,
+ "learning_rate": 0.0006,
+ "loss": 4.367966651916504,
+ "step": 4663
+ },
+ {
+ "epoch": 64.78253275109171,
+ "grad_norm": 0.017946481704711914,
+ "learning_rate": 0.0006,
+ "loss": 4.205172538757324,
+ "step": 4664
+ },
+ {
+ "epoch": 64.79650655021834,
+ "grad_norm": 0.017744354903697968,
+ "learning_rate": 0.0006,
+ "loss": 4.295661926269531,
+ "step": 4665
+ },
+ {
+ "epoch": 64.81048034934499,
+ "grad_norm": 0.016103824600577354,
+ "learning_rate": 0.0006,
+ "loss": 4.276543617248535,
+ "step": 4666
+ },
+ {
+ "epoch": 64.82445414847162,
+ "grad_norm": 0.014590012840926647,
+ "learning_rate": 0.0006,
+ "loss": 4.414247035980225,
+ "step": 4667
+ },
+ {
+ "epoch": 64.83842794759825,
+ "grad_norm": 0.01782386749982834,
+ "learning_rate": 0.0006,
+ "loss": 4.203220367431641,
+ "step": 4668
+ },
+ {
+ "epoch": 64.8524017467249,
+ "grad_norm": 0.022122733294963837,
+ "learning_rate": 0.0006,
+ "loss": 4.38078498840332,
+ "step": 4669
+ },
+ {
+ "epoch": 64.86637554585153,
+ "grad_norm": 0.0250293780118227,
+ "learning_rate": 0.0006,
+ "loss": 4.3310136795043945,
+ "step": 4670
+ },
+ {
+ "epoch": 64.88034934497817,
+ "grad_norm": 0.022438062354922295,
+ "learning_rate": 0.0006,
+ "loss": 4.428631782531738,
+ "step": 4671
+ },
+ {
+ "epoch": 64.8943231441048,
+ "grad_norm": 0.018518192693591118,
+ "learning_rate": 0.0006,
+ "loss": 4.428250312805176,
+ "step": 4672
+ },
+ {
+ "epoch": 64.90829694323143,
+ "grad_norm": 0.018105225637555122,
+ "learning_rate": 0.0006,
+ "loss": 4.336209297180176,
+ "step": 4673
+ },
+ {
+ "epoch": 64.92227074235808,
+ "grad_norm": 0.017564399167895317,
+ "learning_rate": 0.0006,
+ "loss": 4.416011810302734,
+ "step": 4674
+ },
+ {
+ "epoch": 64.93624454148471,
+ "grad_norm": 0.018747031688690186,
+ "learning_rate": 0.0006,
+ "loss": 4.419938087463379,
+ "step": 4675
+ },
+ {
+ "epoch": 64.95021834061136,
+ "grad_norm": 0.018686484545469284,
+ "learning_rate": 0.0006,
+ "loss": 4.347185134887695,
+ "step": 4676
+ },
+ {
+ "epoch": 64.96419213973799,
+ "grad_norm": 0.019329898059368134,
+ "learning_rate": 0.0006,
+ "loss": 4.343749046325684,
+ "step": 4677
+ },
+ {
+ "epoch": 64.97816593886463,
+ "grad_norm": 0.018517782911658287,
+ "learning_rate": 0.0006,
+ "loss": 4.264604568481445,
+ "step": 4678
+ },
+ {
+ "epoch": 64.99213973799127,
+ "grad_norm": 0.016579125076532364,
+ "learning_rate": 0.0006,
+ "loss": 4.303423881530762,
+ "step": 4679
+ },
+ {
+ "epoch": 65.0,
+ "grad_norm": 0.01617567613720894,
+ "learning_rate": 0.0006,
+ "loss": 4.326556205749512,
+ "step": 4680
+ },
+ {
+ "epoch": 65.0,
+ "eval_loss": 4.655026912689209,
+ "eval_runtime": 57.1299,
+ "eval_samples_per_second": 42.745,
+ "eval_steps_per_second": 1.348,
+ "step": 4680
+ },
+ {
+ "epoch": 65.01397379912663,
+ "grad_norm": 0.01607249490916729,
+ "learning_rate": 0.0006,
+ "loss": 4.276607036590576,
+ "step": 4681
+ },
+ {
+ "epoch": 65.02794759825328,
+ "grad_norm": 0.016779497265815735,
+ "learning_rate": 0.0006,
+ "loss": 4.412464618682861,
+ "step": 4682
+ },
+ {
+ "epoch": 65.04192139737991,
+ "grad_norm": 0.018161555752158165,
+ "learning_rate": 0.0006,
+ "loss": 4.199329376220703,
+ "step": 4683
+ },
+ {
+ "epoch": 65.05589519650655,
+ "grad_norm": 0.017508309334516525,
+ "learning_rate": 0.0006,
+ "loss": 4.217567443847656,
+ "step": 4684
+ },
+ {
+ "epoch": 65.06986899563319,
+ "grad_norm": 0.015716979280114174,
+ "learning_rate": 0.0006,
+ "loss": 4.322023391723633,
+ "step": 4685
+ },
+ {
+ "epoch": 65.08384279475983,
+ "grad_norm": 0.015113635919988155,
+ "learning_rate": 0.0006,
+ "loss": 4.333515167236328,
+ "step": 4686
+ },
+ {
+ "epoch": 65.09781659388646,
+ "grad_norm": 0.01437560748308897,
+ "learning_rate": 0.0006,
+ "loss": 4.295609474182129,
+ "step": 4687
+ },
+ {
+ "epoch": 65.1117903930131,
+ "grad_norm": 0.013074862770736217,
+ "learning_rate": 0.0006,
+ "loss": 4.314120769500732,
+ "step": 4688
+ },
+ {
+ "epoch": 65.12576419213974,
+ "grad_norm": 0.012908284552395344,
+ "learning_rate": 0.0006,
+ "loss": 4.216890811920166,
+ "step": 4689
+ },
+ {
+ "epoch": 65.13973799126637,
+ "grad_norm": 0.013001754879951477,
+ "learning_rate": 0.0006,
+ "loss": 4.287717342376709,
+ "step": 4690
+ },
+ {
+ "epoch": 65.15371179039302,
+ "grad_norm": 0.013694263994693756,
+ "learning_rate": 0.0006,
+ "loss": 4.333837985992432,
+ "step": 4691
+ },
+ {
+ "epoch": 65.16768558951965,
+ "grad_norm": 0.014461833983659744,
+ "learning_rate": 0.0006,
+ "loss": 4.323545455932617,
+ "step": 4692
+ },
+ {
+ "epoch": 65.18165938864628,
+ "grad_norm": 0.013670675456523895,
+ "learning_rate": 0.0006,
+ "loss": 4.3313493728637695,
+ "step": 4693
+ },
+ {
+ "epoch": 65.19563318777293,
+ "grad_norm": 0.013094350695610046,
+ "learning_rate": 0.0006,
+ "loss": 4.23858118057251,
+ "step": 4694
+ },
+ {
+ "epoch": 65.20960698689956,
+ "grad_norm": 0.013996944762766361,
+ "learning_rate": 0.0006,
+ "loss": 4.428705215454102,
+ "step": 4695
+ },
+ {
+ "epoch": 65.2235807860262,
+ "grad_norm": 0.013295048847794533,
+ "learning_rate": 0.0006,
+ "loss": 4.238950252532959,
+ "step": 4696
+ },
+ {
+ "epoch": 65.23755458515284,
+ "grad_norm": 0.013380616903305054,
+ "learning_rate": 0.0006,
+ "loss": 4.273185729980469,
+ "step": 4697
+ },
+ {
+ "epoch": 65.25152838427948,
+ "grad_norm": 0.015062890015542507,
+ "learning_rate": 0.0006,
+ "loss": 4.29472017288208,
+ "step": 4698
+ },
+ {
+ "epoch": 65.26550218340611,
+ "grad_norm": 0.017317263409495354,
+ "learning_rate": 0.0006,
+ "loss": 4.337954998016357,
+ "step": 4699
+ },
+ {
+ "epoch": 65.27947598253274,
+ "grad_norm": 0.017698992043733597,
+ "learning_rate": 0.0006,
+ "loss": 4.336437225341797,
+ "step": 4700
+ },
+ {
+ "epoch": 65.29344978165939,
+ "grad_norm": 0.01628575474023819,
+ "learning_rate": 0.0006,
+ "loss": 4.302372455596924,
+ "step": 4701
+ },
+ {
+ "epoch": 65.30742358078602,
+ "grad_norm": 0.014531854540109634,
+ "learning_rate": 0.0006,
+ "loss": 4.256450653076172,
+ "step": 4702
+ },
+ {
+ "epoch": 65.32139737991267,
+ "grad_norm": 0.014549105428159237,
+ "learning_rate": 0.0006,
+ "loss": 4.302240371704102,
+ "step": 4703
+ },
+ {
+ "epoch": 65.3353711790393,
+ "grad_norm": 0.014881663955748081,
+ "learning_rate": 0.0006,
+ "loss": 4.3316121101379395,
+ "step": 4704
+ },
+ {
+ "epoch": 65.34934497816595,
+ "grad_norm": 0.014513437636196613,
+ "learning_rate": 0.0006,
+ "loss": 4.441212177276611,
+ "step": 4705
+ },
+ {
+ "epoch": 65.36331877729258,
+ "grad_norm": 0.014122087508440018,
+ "learning_rate": 0.0006,
+ "loss": 4.343296527862549,
+ "step": 4706
+ },
+ {
+ "epoch": 65.37729257641921,
+ "grad_norm": 0.014817348681390285,
+ "learning_rate": 0.0006,
+ "loss": 4.195674896240234,
+ "step": 4707
+ },
+ {
+ "epoch": 65.39126637554585,
+ "grad_norm": 0.01787056028842926,
+ "learning_rate": 0.0006,
+ "loss": 4.3507914543151855,
+ "step": 4708
+ },
+ {
+ "epoch": 65.40524017467249,
+ "grad_norm": 0.015295136719942093,
+ "learning_rate": 0.0006,
+ "loss": 4.234372138977051,
+ "step": 4709
+ },
+ {
+ "epoch": 65.41921397379913,
+ "grad_norm": 0.014324234798550606,
+ "learning_rate": 0.0006,
+ "loss": 4.2907867431640625,
+ "step": 4710
+ },
+ {
+ "epoch": 65.43318777292576,
+ "grad_norm": 0.016090670600533485,
+ "learning_rate": 0.0006,
+ "loss": 4.177244186401367,
+ "step": 4711
+ },
+ {
+ "epoch": 65.44716157205241,
+ "grad_norm": 0.0174604132771492,
+ "learning_rate": 0.0006,
+ "loss": 4.335577964782715,
+ "step": 4712
+ },
+ {
+ "epoch": 65.46113537117904,
+ "grad_norm": 0.016731899231672287,
+ "learning_rate": 0.0006,
+ "loss": 4.30197811126709,
+ "step": 4713
+ },
+ {
+ "epoch": 65.47510917030567,
+ "grad_norm": 0.017423735931515694,
+ "learning_rate": 0.0006,
+ "loss": 4.3528852462768555,
+ "step": 4714
+ },
+ {
+ "epoch": 65.48908296943232,
+ "grad_norm": 0.01780976541340351,
+ "learning_rate": 0.0006,
+ "loss": 4.185036659240723,
+ "step": 4715
+ },
+ {
+ "epoch": 65.50305676855895,
+ "grad_norm": 0.019977200776338577,
+ "learning_rate": 0.0006,
+ "loss": 4.328766345977783,
+ "step": 4716
+ },
+ {
+ "epoch": 65.5170305676856,
+ "grad_norm": 0.020284287631511688,
+ "learning_rate": 0.0006,
+ "loss": 4.285494804382324,
+ "step": 4717
+ },
+ {
+ "epoch": 65.53100436681223,
+ "grad_norm": 0.019168002530932426,
+ "learning_rate": 0.0006,
+ "loss": 4.328560829162598,
+ "step": 4718
+ },
+ {
+ "epoch": 65.54497816593886,
+ "grad_norm": 0.019217776134610176,
+ "learning_rate": 0.0006,
+ "loss": 4.189998626708984,
+ "step": 4719
+ },
+ {
+ "epoch": 65.5589519650655,
+ "grad_norm": 0.018960488960146904,
+ "learning_rate": 0.0006,
+ "loss": 4.4995012283325195,
+ "step": 4720
+ },
+ {
+ "epoch": 65.57292576419214,
+ "grad_norm": 0.01941288262605667,
+ "learning_rate": 0.0006,
+ "loss": 4.26570987701416,
+ "step": 4721
+ },
+ {
+ "epoch": 65.58689956331878,
+ "grad_norm": 0.020263448357582092,
+ "learning_rate": 0.0006,
+ "loss": 4.411645889282227,
+ "step": 4722
+ },
+ {
+ "epoch": 65.60087336244541,
+ "grad_norm": 0.02144799567759037,
+ "learning_rate": 0.0006,
+ "loss": 4.163675308227539,
+ "step": 4723
+ },
+ {
+ "epoch": 65.61484716157206,
+ "grad_norm": 0.019790446385741234,
+ "learning_rate": 0.0006,
+ "loss": 4.2361063957214355,
+ "step": 4724
+ },
+ {
+ "epoch": 65.62882096069869,
+ "grad_norm": 0.01829805225133896,
+ "learning_rate": 0.0006,
+ "loss": 4.449810028076172,
+ "step": 4725
+ },
+ {
+ "epoch": 65.64279475982532,
+ "grad_norm": 0.01571105420589447,
+ "learning_rate": 0.0006,
+ "loss": 4.281406879425049,
+ "step": 4726
+ },
+ {
+ "epoch": 65.65676855895197,
+ "grad_norm": 0.016477147117257118,
+ "learning_rate": 0.0006,
+ "loss": 4.293408393859863,
+ "step": 4727
+ },
+ {
+ "epoch": 65.6707423580786,
+ "grad_norm": 0.016766032204031944,
+ "learning_rate": 0.0006,
+ "loss": 4.378252983093262,
+ "step": 4728
+ },
+ {
+ "epoch": 65.68471615720524,
+ "grad_norm": 0.0159933939576149,
+ "learning_rate": 0.0006,
+ "loss": 4.3065667152404785,
+ "step": 4729
+ },
+ {
+ "epoch": 65.69868995633188,
+ "grad_norm": 0.01658209227025509,
+ "learning_rate": 0.0006,
+ "loss": 4.174026966094971,
+ "step": 4730
+ },
+ {
+ "epoch": 65.71266375545852,
+ "grad_norm": 0.015081380493938923,
+ "learning_rate": 0.0006,
+ "loss": 4.360222339630127,
+ "step": 4731
+ },
+ {
+ "epoch": 65.72663755458515,
+ "grad_norm": 0.015110890381038189,
+ "learning_rate": 0.0006,
+ "loss": 4.218810558319092,
+ "step": 4732
+ },
+ {
+ "epoch": 65.74061135371178,
+ "grad_norm": 0.015300953760743141,
+ "learning_rate": 0.0006,
+ "loss": 4.331979274749756,
+ "step": 4733
+ },
+ {
+ "epoch": 65.75458515283843,
+ "grad_norm": 0.01715504191815853,
+ "learning_rate": 0.0006,
+ "loss": 4.349112033843994,
+ "step": 4734
+ },
+ {
+ "epoch": 65.76855895196506,
+ "grad_norm": 0.015870366245508194,
+ "learning_rate": 0.0006,
+ "loss": 4.329850196838379,
+ "step": 4735
+ },
+ {
+ "epoch": 65.78253275109171,
+ "grad_norm": 0.01541937980800867,
+ "learning_rate": 0.0006,
+ "loss": 4.3400163650512695,
+ "step": 4736
+ },
+ {
+ "epoch": 65.79650655021834,
+ "grad_norm": 0.01738942787051201,
+ "learning_rate": 0.0006,
+ "loss": 4.220798492431641,
+ "step": 4737
+ },
+ {
+ "epoch": 65.81048034934499,
+ "grad_norm": 0.01685990020632744,
+ "learning_rate": 0.0006,
+ "loss": 4.177238464355469,
+ "step": 4738
+ },
+ {
+ "epoch": 65.82445414847162,
+ "grad_norm": 0.016183465719223022,
+ "learning_rate": 0.0006,
+ "loss": 4.327216625213623,
+ "step": 4739
+ },
+ {
+ "epoch": 65.83842794759825,
+ "grad_norm": 0.01677694357931614,
+ "learning_rate": 0.0006,
+ "loss": 4.27766227722168,
+ "step": 4740
+ },
+ {
+ "epoch": 65.8524017467249,
+ "grad_norm": 0.01645066775381565,
+ "learning_rate": 0.0006,
+ "loss": 4.288347244262695,
+ "step": 4741
+ },
+ {
+ "epoch": 65.86637554585153,
+ "grad_norm": 0.018034635111689568,
+ "learning_rate": 0.0006,
+ "loss": 4.292458534240723,
+ "step": 4742
+ },
+ {
+ "epoch": 65.88034934497817,
+ "grad_norm": 0.017973346635699272,
+ "learning_rate": 0.0006,
+ "loss": 4.3117265701293945,
+ "step": 4743
+ },
+ {
+ "epoch": 65.8943231441048,
+ "grad_norm": 0.016605669632554054,
+ "learning_rate": 0.0006,
+ "loss": 4.280615329742432,
+ "step": 4744
+ },
+ {
+ "epoch": 65.90829694323143,
+ "grad_norm": 0.016298074275255203,
+ "learning_rate": 0.0006,
+ "loss": 4.301933288574219,
+ "step": 4745
+ },
+ {
+ "epoch": 65.92227074235808,
+ "grad_norm": 0.015383216552436352,
+ "learning_rate": 0.0006,
+ "loss": 4.352383136749268,
+ "step": 4746
+ },
+ {
+ "epoch": 65.93624454148471,
+ "grad_norm": 0.016222182661294937,
+ "learning_rate": 0.0006,
+ "loss": 4.305725574493408,
+ "step": 4747
+ },
+ {
+ "epoch": 65.95021834061136,
+ "grad_norm": 0.01691329851746559,
+ "learning_rate": 0.0006,
+ "loss": 4.29502010345459,
+ "step": 4748
+ },
+ {
+ "epoch": 65.96419213973799,
+ "grad_norm": 0.017637677490711212,
+ "learning_rate": 0.0006,
+ "loss": 4.3133039474487305,
+ "step": 4749
+ },
+ {
+ "epoch": 65.97816593886463,
+ "grad_norm": 0.021362047642469406,
+ "learning_rate": 0.0006,
+ "loss": 4.458834648132324,
+ "step": 4750
+ },
+ {
+ "epoch": 65.99213973799127,
+ "grad_norm": 0.023553011938929558,
+ "learning_rate": 0.0006,
+ "loss": 4.298727989196777,
+ "step": 4751
+ },
+ {
+ "epoch": 66.0,
+ "grad_norm": 0.02154950052499771,
+ "learning_rate": 0.0006,
+ "loss": 4.403392791748047,
+ "step": 4752
+ },
+ {
+ "epoch": 66.0,
+ "eval_loss": 4.724340438842773,
+ "eval_runtime": 56.9555,
+ "eval_samples_per_second": 42.876,
+ "eval_steps_per_second": 1.352,
+ "step": 4752
+ },
+ {
+ "epoch": 66.01397379912663,
+ "grad_norm": 0.01916784606873989,
+ "learning_rate": 0.0006,
+ "loss": 4.22931432723999,
+ "step": 4753
+ },
+ {
+ "epoch": 66.02794759825328,
+ "grad_norm": 0.017492054030299187,
+ "learning_rate": 0.0006,
+ "loss": 4.271522521972656,
+ "step": 4754
+ },
+ {
+ "epoch": 66.04192139737991,
+ "grad_norm": 0.018408095464110374,
+ "learning_rate": 0.0006,
+ "loss": 4.337705135345459,
+ "step": 4755
+ },
+ {
+ "epoch": 66.05589519650655,
+ "grad_norm": 0.019220108166337013,
+ "learning_rate": 0.0006,
+ "loss": 4.280340194702148,
+ "step": 4756
+ },
+ {
+ "epoch": 66.06986899563319,
+ "grad_norm": 0.01773735322058201,
+ "learning_rate": 0.0006,
+ "loss": 4.390201568603516,
+ "step": 4757
+ },
+ {
+ "epoch": 66.08384279475983,
+ "grad_norm": 0.01679125428199768,
+ "learning_rate": 0.0006,
+ "loss": 4.299282073974609,
+ "step": 4758
+ },
+ {
+ "epoch": 66.09781659388646,
+ "grad_norm": 0.01479059923440218,
+ "learning_rate": 0.0006,
+ "loss": 4.472630500793457,
+ "step": 4759
+ },
+ {
+ "epoch": 66.1117903930131,
+ "grad_norm": 0.015130104497075081,
+ "learning_rate": 0.0006,
+ "loss": 4.301194190979004,
+ "step": 4760
+ },
+ {
+ "epoch": 66.12576419213974,
+ "grad_norm": 0.0158979594707489,
+ "learning_rate": 0.0006,
+ "loss": 4.375751972198486,
+ "step": 4761
+ },
+ {
+ "epoch": 66.13973799126637,
+ "grad_norm": 0.014432408846914768,
+ "learning_rate": 0.0006,
+ "loss": 4.279788017272949,
+ "step": 4762
+ },
+ {
+ "epoch": 66.15371179039302,
+ "grad_norm": 0.013026784174144268,
+ "learning_rate": 0.0006,
+ "loss": 4.255035400390625,
+ "step": 4763
+ },
+ {
+ "epoch": 66.16768558951965,
+ "grad_norm": 0.014422653242945671,
+ "learning_rate": 0.0006,
+ "loss": 4.261832237243652,
+ "step": 4764
+ },
+ {
+ "epoch": 66.18165938864628,
+ "grad_norm": 0.014613019302487373,
+ "learning_rate": 0.0006,
+ "loss": 4.382184028625488,
+ "step": 4765
+ },
+ {
+ "epoch": 66.19563318777293,
+ "grad_norm": 0.014989894814789295,
+ "learning_rate": 0.0006,
+ "loss": 4.40528678894043,
+ "step": 4766
+ },
+ {
+ "epoch": 66.20960698689956,
+ "grad_norm": 0.018190350383520126,
+ "learning_rate": 0.0006,
+ "loss": 4.302414894104004,
+ "step": 4767
+ },
+ {
+ "epoch": 66.2235807860262,
+ "grad_norm": 0.020887890830636024,
+ "learning_rate": 0.0006,
+ "loss": 4.289920806884766,
+ "step": 4768
+ },
+ {
+ "epoch": 66.23755458515284,
+ "grad_norm": 0.02376876398921013,
+ "learning_rate": 0.0006,
+ "loss": 4.297855377197266,
+ "step": 4769
+ },
+ {
+ "epoch": 66.25152838427948,
+ "grad_norm": 0.02152269519865513,
+ "learning_rate": 0.0006,
+ "loss": 4.1972832679748535,
+ "step": 4770
+ },
+ {
+ "epoch": 66.26550218340611,
+ "grad_norm": 0.015572802163660526,
+ "learning_rate": 0.0006,
+ "loss": 4.307487487792969,
+ "step": 4771
+ },
+ {
+ "epoch": 66.27947598253274,
+ "grad_norm": 0.015228749252855778,
+ "learning_rate": 0.0006,
+ "loss": 4.103250503540039,
+ "step": 4772
+ },
+ {
+ "epoch": 66.29344978165939,
+ "grad_norm": 0.016925504431128502,
+ "learning_rate": 0.0006,
+ "loss": 4.247461318969727,
+ "step": 4773
+ },
+ {
+ "epoch": 66.30742358078602,
+ "grad_norm": 0.01613711006939411,
+ "learning_rate": 0.0006,
+ "loss": 4.4699273109436035,
+ "step": 4774
+ },
+ {
+ "epoch": 66.32139737991267,
+ "grad_norm": 0.018332140520215034,
+ "learning_rate": 0.0006,
+ "loss": 4.22969388961792,
+ "step": 4775
+ },
+ {
+ "epoch": 66.3353711790393,
+ "grad_norm": 0.019575074315071106,
+ "learning_rate": 0.0006,
+ "loss": 4.29397439956665,
+ "step": 4776
+ },
+ {
+ "epoch": 66.34934497816595,
+ "grad_norm": 0.018177419900894165,
+ "learning_rate": 0.0006,
+ "loss": 4.221899509429932,
+ "step": 4777
+ },
+ {
+ "epoch": 66.36331877729258,
+ "grad_norm": 0.014292889274656773,
+ "learning_rate": 0.0006,
+ "loss": 4.325457572937012,
+ "step": 4778
+ },
+ {
+ "epoch": 66.37729257641921,
+ "grad_norm": 0.015098796226084232,
+ "learning_rate": 0.0006,
+ "loss": 4.364837169647217,
+ "step": 4779
+ },
+ {
+ "epoch": 66.39126637554585,
+ "grad_norm": 0.013992635533213615,
+ "learning_rate": 0.0006,
+ "loss": 4.299082279205322,
+ "step": 4780
+ },
+ {
+ "epoch": 66.40524017467249,
+ "grad_norm": 0.01504511572420597,
+ "learning_rate": 0.0006,
+ "loss": 4.230877876281738,
+ "step": 4781
+ },
+ {
+ "epoch": 66.41921397379913,
+ "grad_norm": 0.014875639230012894,
+ "learning_rate": 0.0006,
+ "loss": 4.368405818939209,
+ "step": 4782
+ },
+ {
+ "epoch": 66.43318777292576,
+ "grad_norm": 0.014306887984275818,
+ "learning_rate": 0.0006,
+ "loss": 4.389184474945068,
+ "step": 4783
+ },
+ {
+ "epoch": 66.44716157205241,
+ "grad_norm": 0.012959563173353672,
+ "learning_rate": 0.0006,
+ "loss": 4.37934684753418,
+ "step": 4784
+ },
+ {
+ "epoch": 66.46113537117904,
+ "grad_norm": 0.013206703588366508,
+ "learning_rate": 0.0006,
+ "loss": 4.442799091339111,
+ "step": 4785
+ },
+ {
+ "epoch": 66.47510917030567,
+ "grad_norm": 0.014848909340798855,
+ "learning_rate": 0.0006,
+ "loss": 4.203481674194336,
+ "step": 4786
+ },
+ {
+ "epoch": 66.48908296943232,
+ "grad_norm": 0.015637682750821114,
+ "learning_rate": 0.0006,
+ "loss": 4.27552604675293,
+ "step": 4787
+ },
+ {
+ "epoch": 66.50305676855895,
+ "grad_norm": 0.016017816960811615,
+ "learning_rate": 0.0006,
+ "loss": 4.296402931213379,
+ "step": 4788
+ },
+ {
+ "epoch": 66.5170305676856,
+ "grad_norm": 0.01577814109623432,
+ "learning_rate": 0.0006,
+ "loss": 4.301733016967773,
+ "step": 4789
+ },
+ {
+ "epoch": 66.53100436681223,
+ "grad_norm": 0.015293175354599953,
+ "learning_rate": 0.0006,
+ "loss": 4.369877815246582,
+ "step": 4790
+ },
+ {
+ "epoch": 66.54497816593886,
+ "grad_norm": 0.014262278564274311,
+ "learning_rate": 0.0006,
+ "loss": 4.3314738273620605,
+ "step": 4791
+ },
+ {
+ "epoch": 66.5589519650655,
+ "grad_norm": 0.015123482793569565,
+ "learning_rate": 0.0006,
+ "loss": 4.313591957092285,
+ "step": 4792
+ },
+ {
+ "epoch": 66.57292576419214,
+ "grad_norm": 0.015458766371011734,
+ "learning_rate": 0.0006,
+ "loss": 4.3045854568481445,
+ "step": 4793
+ },
+ {
+ "epoch": 66.58689956331878,
+ "grad_norm": 0.017217233777046204,
+ "learning_rate": 0.0006,
+ "loss": 4.269944190979004,
+ "step": 4794
+ },
+ {
+ "epoch": 66.60087336244541,
+ "grad_norm": 0.01962282694876194,
+ "learning_rate": 0.0006,
+ "loss": 4.271735191345215,
+ "step": 4795
+ },
+ {
+ "epoch": 66.61484716157206,
+ "grad_norm": 0.01917066052556038,
+ "learning_rate": 0.0006,
+ "loss": 4.288168430328369,
+ "step": 4796
+ },
+ {
+ "epoch": 66.62882096069869,
+ "grad_norm": 0.018211524933576584,
+ "learning_rate": 0.0006,
+ "loss": 4.280660629272461,
+ "step": 4797
+ },
+ {
+ "epoch": 66.64279475982532,
+ "grad_norm": 0.01936524175107479,
+ "learning_rate": 0.0006,
+ "loss": 4.3710551261901855,
+ "step": 4798
+ },
+ {
+ "epoch": 66.65676855895197,
+ "grad_norm": 0.01882351003587246,
+ "learning_rate": 0.0006,
+ "loss": 4.342750549316406,
+ "step": 4799
+ },
+ {
+ "epoch": 66.6707423580786,
+ "grad_norm": 0.017149122431874275,
+ "learning_rate": 0.0006,
+ "loss": 4.32134485244751,
+ "step": 4800
+ },
+ {
+ "epoch": 66.68471615720524,
+ "grad_norm": 0.01699327677488327,
+ "learning_rate": 0.0006,
+ "loss": 4.248893737792969,
+ "step": 4801
+ },
+ {
+ "epoch": 66.69868995633188,
+ "grad_norm": 0.016873255372047424,
+ "learning_rate": 0.0006,
+ "loss": 4.277350425720215,
+ "step": 4802
+ },
+ {
+ "epoch": 66.71266375545852,
+ "grad_norm": 0.016592837870121002,
+ "learning_rate": 0.0006,
+ "loss": 4.221665859222412,
+ "step": 4803
+ },
+ {
+ "epoch": 66.72663755458515,
+ "grad_norm": 0.015887131914496422,
+ "learning_rate": 0.0006,
+ "loss": 4.328040599822998,
+ "step": 4804
+ },
+ {
+ "epoch": 66.74061135371178,
+ "grad_norm": 0.014057286083698273,
+ "learning_rate": 0.0006,
+ "loss": 4.3294548988342285,
+ "step": 4805
+ },
+ {
+ "epoch": 66.75458515283843,
+ "grad_norm": 0.013135905377566814,
+ "learning_rate": 0.0006,
+ "loss": 4.254383563995361,
+ "step": 4806
+ },
+ {
+ "epoch": 66.76855895196506,
+ "grad_norm": 0.013632557354867458,
+ "learning_rate": 0.0006,
+ "loss": 4.342815399169922,
+ "step": 4807
+ },
+ {
+ "epoch": 66.78253275109171,
+ "grad_norm": 0.015512688085436821,
+ "learning_rate": 0.0006,
+ "loss": 4.297835350036621,
+ "step": 4808
+ },
+ {
+ "epoch": 66.79650655021834,
+ "grad_norm": 0.01621832884848118,
+ "learning_rate": 0.0006,
+ "loss": 4.314773082733154,
+ "step": 4809
+ },
+ {
+ "epoch": 66.81048034934499,
+ "grad_norm": 0.016196032986044884,
+ "learning_rate": 0.0006,
+ "loss": 4.29209566116333,
+ "step": 4810
+ },
+ {
+ "epoch": 66.82445414847162,
+ "grad_norm": 0.015463353134691715,
+ "learning_rate": 0.0006,
+ "loss": 4.262795448303223,
+ "step": 4811
+ },
+ {
+ "epoch": 66.83842794759825,
+ "grad_norm": 0.016143787652254105,
+ "learning_rate": 0.0006,
+ "loss": 4.274284362792969,
+ "step": 4812
+ },
+ {
+ "epoch": 66.8524017467249,
+ "grad_norm": 0.01610061153769493,
+ "learning_rate": 0.0006,
+ "loss": 4.337160587310791,
+ "step": 4813
+ },
+ {
+ "epoch": 66.86637554585153,
+ "grad_norm": 0.01582838036119938,
+ "learning_rate": 0.0006,
+ "loss": 4.192723274230957,
+ "step": 4814
+ },
+ {
+ "epoch": 66.88034934497817,
+ "grad_norm": 0.015081540681421757,
+ "learning_rate": 0.0006,
+ "loss": 4.370708465576172,
+ "step": 4815
+ },
+ {
+ "epoch": 66.8943231441048,
+ "grad_norm": 0.015929723158478737,
+ "learning_rate": 0.0006,
+ "loss": 4.340244293212891,
+ "step": 4816
+ },
+ {
+ "epoch": 66.90829694323143,
+ "grad_norm": 0.01784711331129074,
+ "learning_rate": 0.0006,
+ "loss": 4.260406970977783,
+ "step": 4817
+ },
+ {
+ "epoch": 66.92227074235808,
+ "grad_norm": 0.017434703186154366,
+ "learning_rate": 0.0006,
+ "loss": 4.286900043487549,
+ "step": 4818
+ },
+ {
+ "epoch": 66.93624454148471,
+ "grad_norm": 0.016262540593743324,
+ "learning_rate": 0.0006,
+ "loss": 4.292142391204834,
+ "step": 4819
+ },
+ {
+ "epoch": 66.95021834061136,
+ "grad_norm": 0.016122829169034958,
+ "learning_rate": 0.0006,
+ "loss": 4.291485786437988,
+ "step": 4820
+ },
+ {
+ "epoch": 66.96419213973799,
+ "grad_norm": 0.014361178502440453,
+ "learning_rate": 0.0006,
+ "loss": 4.275425910949707,
+ "step": 4821
+ },
+ {
+ "epoch": 66.97816593886463,
+ "grad_norm": 0.01373987551778555,
+ "learning_rate": 0.0006,
+ "loss": 4.223120212554932,
+ "step": 4822
+ },
+ {
+ "epoch": 66.99213973799127,
+ "grad_norm": 0.015687376260757446,
+ "learning_rate": 0.0006,
+ "loss": 4.355381011962891,
+ "step": 4823
+ },
+ {
+ "epoch": 67.0,
+ "grad_norm": 0.01998465321958065,
+ "learning_rate": 0.0006,
+ "loss": 4.264053821563721,
+ "step": 4824
+ },
+ {
+ "epoch": 67.0,
+ "eval_loss": 4.662398338317871,
+ "eval_runtime": 56.7928,
+ "eval_samples_per_second": 42.998,
+ "eval_steps_per_second": 1.356,
+ "step": 4824
+ },
+ {
+ "epoch": 67.01397379912663,
+ "grad_norm": 0.021736836060881615,
+ "learning_rate": 0.0006,
+ "loss": 4.243354797363281,
+ "step": 4825
+ },
+ {
+ "epoch": 67.02794759825328,
+ "grad_norm": 0.021476037800312042,
+ "learning_rate": 0.0006,
+ "loss": 4.218207836151123,
+ "step": 4826
+ },
+ {
+ "epoch": 67.04192139737991,
+ "grad_norm": 0.017188388854265213,
+ "learning_rate": 0.0006,
+ "loss": 4.18048095703125,
+ "step": 4827
+ },
+ {
+ "epoch": 67.05589519650655,
+ "grad_norm": 0.017293719574809074,
+ "learning_rate": 0.0006,
+ "loss": 4.239145278930664,
+ "step": 4828
+ },
+ {
+ "epoch": 67.06986899563319,
+ "grad_norm": 0.017383819445967674,
+ "learning_rate": 0.0006,
+ "loss": 4.114765167236328,
+ "step": 4829
+ },
+ {
+ "epoch": 67.08384279475983,
+ "grad_norm": 0.01651529036462307,
+ "learning_rate": 0.0006,
+ "loss": 4.351023197174072,
+ "step": 4830
+ },
+ {
+ "epoch": 67.09781659388646,
+ "grad_norm": 0.015893027186393738,
+ "learning_rate": 0.0006,
+ "loss": 4.261432647705078,
+ "step": 4831
+ },
+ {
+ "epoch": 67.1117903930131,
+ "grad_norm": 0.013381517492234707,
+ "learning_rate": 0.0006,
+ "loss": 4.172064781188965,
+ "step": 4832
+ },
+ {
+ "epoch": 67.12576419213974,
+ "grad_norm": 0.014642654918134212,
+ "learning_rate": 0.0006,
+ "loss": 4.356171131134033,
+ "step": 4833
+ },
+ {
+ "epoch": 67.13973799126637,
+ "grad_norm": 0.01425282284617424,
+ "learning_rate": 0.0006,
+ "loss": 4.303495407104492,
+ "step": 4834
+ },
+ {
+ "epoch": 67.15371179039302,
+ "grad_norm": 0.014574599452316761,
+ "learning_rate": 0.0006,
+ "loss": 4.2850847244262695,
+ "step": 4835
+ },
+ {
+ "epoch": 67.16768558951965,
+ "grad_norm": 0.012676282785832882,
+ "learning_rate": 0.0006,
+ "loss": 4.297845840454102,
+ "step": 4836
+ },
+ {
+ "epoch": 67.18165938864628,
+ "grad_norm": 0.013948136940598488,
+ "learning_rate": 0.0006,
+ "loss": 4.324530601501465,
+ "step": 4837
+ },
+ {
+ "epoch": 67.19563318777293,
+ "grad_norm": 0.014195773750543594,
+ "learning_rate": 0.0006,
+ "loss": 4.31236457824707,
+ "step": 4838
+ },
+ {
+ "epoch": 67.20960698689956,
+ "grad_norm": 0.013631382025778294,
+ "learning_rate": 0.0006,
+ "loss": 4.086362838745117,
+ "step": 4839
+ },
+ {
+ "epoch": 67.2235807860262,
+ "grad_norm": 0.012798587791621685,
+ "learning_rate": 0.0006,
+ "loss": 4.328471660614014,
+ "step": 4840
+ },
+ {
+ "epoch": 67.23755458515284,
+ "grad_norm": 0.014443321153521538,
+ "learning_rate": 0.0006,
+ "loss": 4.278019905090332,
+ "step": 4841
+ },
+ {
+ "epoch": 67.25152838427948,
+ "grad_norm": 0.015355420298874378,
+ "learning_rate": 0.0006,
+ "loss": 4.333611011505127,
+ "step": 4842
+ },
+ {
+ "epoch": 67.26550218340611,
+ "grad_norm": 0.016487296670675278,
+ "learning_rate": 0.0006,
+ "loss": 4.351336479187012,
+ "step": 4843
+ },
+ {
+ "epoch": 67.27947598253274,
+ "grad_norm": 0.017422541975975037,
+ "learning_rate": 0.0006,
+ "loss": 4.386096477508545,
+ "step": 4844
+ },
+ {
+ "epoch": 67.29344978165939,
+ "grad_norm": 0.015790918841958046,
+ "learning_rate": 0.0006,
+ "loss": 4.301377773284912,
+ "step": 4845
+ },
+ {
+ "epoch": 67.30742358078602,
+ "grad_norm": 0.016427364200353622,
+ "learning_rate": 0.0006,
+ "loss": 4.2838945388793945,
+ "step": 4846
+ },
+ {
+ "epoch": 67.32139737991267,
+ "grad_norm": 0.017050761729478836,
+ "learning_rate": 0.0006,
+ "loss": 4.282417297363281,
+ "step": 4847
+ },
+ {
+ "epoch": 67.3353711790393,
+ "grad_norm": 0.015006253495812416,
+ "learning_rate": 0.0006,
+ "loss": 4.352236270904541,
+ "step": 4848
+ },
+ {
+ "epoch": 67.34934497816595,
+ "grad_norm": 0.012996627949178219,
+ "learning_rate": 0.0006,
+ "loss": 4.309242248535156,
+ "step": 4849
+ },
+ {
+ "epoch": 67.36331877729258,
+ "grad_norm": 0.01367615070194006,
+ "learning_rate": 0.0006,
+ "loss": 4.211801528930664,
+ "step": 4850
+ },
+ {
+ "epoch": 67.37729257641921,
+ "grad_norm": 0.014775963500142097,
+ "learning_rate": 0.0006,
+ "loss": 4.360859394073486,
+ "step": 4851
+ },
+ {
+ "epoch": 67.39126637554585,
+ "grad_norm": 0.015854304656386375,
+ "learning_rate": 0.0006,
+ "loss": 4.211830139160156,
+ "step": 4852
+ },
+ {
+ "epoch": 67.40524017467249,
+ "grad_norm": 0.016888894140720367,
+ "learning_rate": 0.0006,
+ "loss": 4.216549873352051,
+ "step": 4853
+ },
+ {
+ "epoch": 67.41921397379913,
+ "grad_norm": 0.0170894768089056,
+ "learning_rate": 0.0006,
+ "loss": 4.263486385345459,
+ "step": 4854
+ },
+ {
+ "epoch": 67.43318777292576,
+ "grad_norm": 0.015829868614673615,
+ "learning_rate": 0.0006,
+ "loss": 4.360196113586426,
+ "step": 4855
+ },
+ {
+ "epoch": 67.44716157205241,
+ "grad_norm": 0.015220202505588531,
+ "learning_rate": 0.0006,
+ "loss": 4.243794918060303,
+ "step": 4856
+ },
+ {
+ "epoch": 67.46113537117904,
+ "grad_norm": 0.01413388829678297,
+ "learning_rate": 0.0006,
+ "loss": 4.286624908447266,
+ "step": 4857
+ },
+ {
+ "epoch": 67.47510917030567,
+ "grad_norm": 0.01426626369357109,
+ "learning_rate": 0.0006,
+ "loss": 4.257316589355469,
+ "step": 4858
+ },
+ {
+ "epoch": 67.48908296943232,
+ "grad_norm": 0.015957845374941826,
+ "learning_rate": 0.0006,
+ "loss": 4.300443649291992,
+ "step": 4859
+ },
+ {
+ "epoch": 67.50305676855895,
+ "grad_norm": 0.015499051660299301,
+ "learning_rate": 0.0006,
+ "loss": 4.278162002563477,
+ "step": 4860
+ },
+ {
+ "epoch": 67.5170305676856,
+ "grad_norm": 0.015145624987781048,
+ "learning_rate": 0.0006,
+ "loss": 4.267950534820557,
+ "step": 4861
+ },
+ {
+ "epoch": 67.53100436681223,
+ "grad_norm": 0.014610803686082363,
+ "learning_rate": 0.0006,
+ "loss": 4.247499942779541,
+ "step": 4862
+ },
+ {
+ "epoch": 67.54497816593886,
+ "grad_norm": 0.014860909432172775,
+ "learning_rate": 0.0006,
+ "loss": 4.263200759887695,
+ "step": 4863
+ },
+ {
+ "epoch": 67.5589519650655,
+ "grad_norm": 0.01718984916806221,
+ "learning_rate": 0.0006,
+ "loss": 4.290019989013672,
+ "step": 4864
+ },
+ {
+ "epoch": 67.57292576419214,
+ "grad_norm": 0.01558963768184185,
+ "learning_rate": 0.0006,
+ "loss": 4.223816871643066,
+ "step": 4865
+ },
+ {
+ "epoch": 67.58689956331878,
+ "grad_norm": 0.014927188865840435,
+ "learning_rate": 0.0006,
+ "loss": 4.363586902618408,
+ "step": 4866
+ },
+ {
+ "epoch": 67.60087336244541,
+ "grad_norm": 0.01573595404624939,
+ "learning_rate": 0.0006,
+ "loss": 4.254765510559082,
+ "step": 4867
+ },
+ {
+ "epoch": 67.61484716157206,
+ "grad_norm": 0.015618929639458656,
+ "learning_rate": 0.0006,
+ "loss": 4.25072717666626,
+ "step": 4868
+ },
+ {
+ "epoch": 67.62882096069869,
+ "grad_norm": 0.01605590060353279,
+ "learning_rate": 0.0006,
+ "loss": 4.25208854675293,
+ "step": 4869
+ },
+ {
+ "epoch": 67.64279475982532,
+ "grad_norm": 0.01804601214826107,
+ "learning_rate": 0.0006,
+ "loss": 4.351354122161865,
+ "step": 4870
+ },
+ {
+ "epoch": 67.65676855895197,
+ "grad_norm": 0.020595047622919083,
+ "learning_rate": 0.0006,
+ "loss": 4.193732738494873,
+ "step": 4871
+ },
+ {
+ "epoch": 67.6707423580786,
+ "grad_norm": 0.023037966340780258,
+ "learning_rate": 0.0006,
+ "loss": 4.27754020690918,
+ "step": 4872
+ },
+ {
+ "epoch": 67.68471615720524,
+ "grad_norm": 0.020680895075201988,
+ "learning_rate": 0.0006,
+ "loss": 4.314730167388916,
+ "step": 4873
+ },
+ {
+ "epoch": 67.69868995633188,
+ "grad_norm": 0.01887434534728527,
+ "learning_rate": 0.0006,
+ "loss": 4.252305030822754,
+ "step": 4874
+ },
+ {
+ "epoch": 67.71266375545852,
+ "grad_norm": 0.016333596780896187,
+ "learning_rate": 0.0006,
+ "loss": 4.146123886108398,
+ "step": 4875
+ },
+ {
+ "epoch": 67.72663755458515,
+ "grad_norm": 0.015688810497522354,
+ "learning_rate": 0.0006,
+ "loss": 4.2814154624938965,
+ "step": 4876
+ },
+ {
+ "epoch": 67.74061135371178,
+ "grad_norm": 0.016270726919174194,
+ "learning_rate": 0.0006,
+ "loss": 4.323200225830078,
+ "step": 4877
+ },
+ {
+ "epoch": 67.75458515283843,
+ "grad_norm": 0.016824988648295403,
+ "learning_rate": 0.0006,
+ "loss": 4.269501686096191,
+ "step": 4878
+ },
+ {
+ "epoch": 67.76855895196506,
+ "grad_norm": 0.01730562001466751,
+ "learning_rate": 0.0006,
+ "loss": 4.187993049621582,
+ "step": 4879
+ },
+ {
+ "epoch": 67.78253275109171,
+ "grad_norm": 0.01546655036509037,
+ "learning_rate": 0.0006,
+ "loss": 4.340415000915527,
+ "step": 4880
+ },
+ {
+ "epoch": 67.79650655021834,
+ "grad_norm": 0.014173958450555801,
+ "learning_rate": 0.0006,
+ "loss": 4.217419624328613,
+ "step": 4881
+ },
+ {
+ "epoch": 67.81048034934499,
+ "grad_norm": 0.016810474917292595,
+ "learning_rate": 0.0006,
+ "loss": 4.352302074432373,
+ "step": 4882
+ },
+ {
+ "epoch": 67.82445414847162,
+ "grad_norm": 0.0175373163074255,
+ "learning_rate": 0.0006,
+ "loss": 4.20377254486084,
+ "step": 4883
+ },
+ {
+ "epoch": 67.83842794759825,
+ "grad_norm": 0.015734676271677017,
+ "learning_rate": 0.0006,
+ "loss": 4.147495746612549,
+ "step": 4884
+ },
+ {
+ "epoch": 67.8524017467249,
+ "grad_norm": 0.015091817826032639,
+ "learning_rate": 0.0006,
+ "loss": 4.272252082824707,
+ "step": 4885
+ },
+ {
+ "epoch": 67.86637554585153,
+ "grad_norm": 0.015000219456851482,
+ "learning_rate": 0.0006,
+ "loss": 4.302773952484131,
+ "step": 4886
+ },
+ {
+ "epoch": 67.88034934497817,
+ "grad_norm": 0.013827184215188026,
+ "learning_rate": 0.0006,
+ "loss": 4.337242126464844,
+ "step": 4887
+ },
+ {
+ "epoch": 67.8943231441048,
+ "grad_norm": 0.015308346599340439,
+ "learning_rate": 0.0006,
+ "loss": 4.140895366668701,
+ "step": 4888
+ },
+ {
+ "epoch": 67.90829694323143,
+ "grad_norm": 0.015356178395450115,
+ "learning_rate": 0.0006,
+ "loss": 4.312541484832764,
+ "step": 4889
+ },
+ {
+ "epoch": 67.92227074235808,
+ "grad_norm": 0.0137353939935565,
+ "learning_rate": 0.0006,
+ "loss": 4.173184394836426,
+ "step": 4890
+ },
+ {
+ "epoch": 67.93624454148471,
+ "grad_norm": 0.01332368515431881,
+ "learning_rate": 0.0006,
+ "loss": 4.191199779510498,
+ "step": 4891
+ },
+ {
+ "epoch": 67.95021834061136,
+ "grad_norm": 0.01488130260258913,
+ "learning_rate": 0.0006,
+ "loss": 4.331752300262451,
+ "step": 4892
+ },
+ {
+ "epoch": 67.96419213973799,
+ "grad_norm": 0.015449733473360538,
+ "learning_rate": 0.0006,
+ "loss": 4.2385454177856445,
+ "step": 4893
+ },
+ {
+ "epoch": 67.97816593886463,
+ "grad_norm": 0.01556103490293026,
+ "learning_rate": 0.0006,
+ "loss": 4.262353897094727,
+ "step": 4894
+ },
+ {
+ "epoch": 67.99213973799127,
+ "grad_norm": 0.014613849110901356,
+ "learning_rate": 0.0006,
+ "loss": 4.223945617675781,
+ "step": 4895
+ },
+ {
+ "epoch": 68.0,
+ "grad_norm": 0.015323462896049023,
+ "learning_rate": 0.0006,
+ "loss": 4.317168712615967,
+ "step": 4896
+ },
+ {
+ "epoch": 68.0,
+ "eval_loss": 4.671603679656982,
+ "eval_runtime": 56.7317,
+ "eval_samples_per_second": 43.045,
+ "eval_steps_per_second": 1.357,
+ "step": 4896
+ },
+ {
+ "epoch": 68.01397379912663,
+ "grad_norm": 0.01910698413848877,
+ "learning_rate": 0.0006,
+ "loss": 4.206167221069336,
+ "step": 4897
+ },
+ {
+ "epoch": 68.02794759825328,
+ "grad_norm": 0.025112485513091087,
+ "learning_rate": 0.0006,
+ "loss": 4.384706974029541,
+ "step": 4898
+ },
+ {
+ "epoch": 68.04192139737991,
+ "grad_norm": 0.02618720941245556,
+ "learning_rate": 0.0006,
+ "loss": 4.321046829223633,
+ "step": 4899
+ },
+ {
+ "epoch": 68.05589519650655,
+ "grad_norm": 0.020853538066148758,
+ "learning_rate": 0.0006,
+ "loss": 4.217638969421387,
+ "step": 4900
+ },
+ {
+ "epoch": 68.06986899563319,
+ "grad_norm": 0.020025717094540596,
+ "learning_rate": 0.0006,
+ "loss": 4.306824684143066,
+ "step": 4901
+ },
+ {
+ "epoch": 68.08384279475983,
+ "grad_norm": 0.023868650197982788,
+ "learning_rate": 0.0006,
+ "loss": 4.327455997467041,
+ "step": 4902
+ },
+ {
+ "epoch": 68.09781659388646,
+ "grad_norm": 0.022279081866145134,
+ "learning_rate": 0.0006,
+ "loss": 4.400699615478516,
+ "step": 4903
+ },
+ {
+ "epoch": 68.1117903930131,
+ "grad_norm": 0.018965687602758408,
+ "learning_rate": 0.0006,
+ "loss": 4.19097900390625,
+ "step": 4904
+ },
+ {
+ "epoch": 68.12576419213974,
+ "grad_norm": 0.01936378888785839,
+ "learning_rate": 0.0006,
+ "loss": 4.206711769104004,
+ "step": 4905
+ },
+ {
+ "epoch": 68.13973799126637,
+ "grad_norm": 0.01811872608959675,
+ "learning_rate": 0.0006,
+ "loss": 4.204622268676758,
+ "step": 4906
+ },
+ {
+ "epoch": 68.15371179039302,
+ "grad_norm": 0.018530402332544327,
+ "learning_rate": 0.0006,
+ "loss": 4.289602279663086,
+ "step": 4907
+ },
+ {
+ "epoch": 68.16768558951965,
+ "grad_norm": 0.01755962148308754,
+ "learning_rate": 0.0006,
+ "loss": 4.317145347595215,
+ "step": 4908
+ },
+ {
+ "epoch": 68.18165938864628,
+ "grad_norm": 0.020413702353835106,
+ "learning_rate": 0.0006,
+ "loss": 4.2882843017578125,
+ "step": 4909
+ },
+ {
+ "epoch": 68.19563318777293,
+ "grad_norm": 0.020064299926161766,
+ "learning_rate": 0.0006,
+ "loss": 4.248260021209717,
+ "step": 4910
+ },
+ {
+ "epoch": 68.20960698689956,
+ "grad_norm": 0.017415888607501984,
+ "learning_rate": 0.0006,
+ "loss": 4.370295524597168,
+ "step": 4911
+ },
+ {
+ "epoch": 68.2235807860262,
+ "grad_norm": 0.015005858615040779,
+ "learning_rate": 0.0006,
+ "loss": 4.319567680358887,
+ "step": 4912
+ },
+ {
+ "epoch": 68.23755458515284,
+ "grad_norm": 0.016349777579307556,
+ "learning_rate": 0.0006,
+ "loss": 4.301776885986328,
+ "step": 4913
+ },
+ {
+ "epoch": 68.25152838427948,
+ "grad_norm": 0.015779560431838036,
+ "learning_rate": 0.0006,
+ "loss": 4.271171569824219,
+ "step": 4914
+ },
+ {
+ "epoch": 68.26550218340611,
+ "grad_norm": 0.013237644918262959,
+ "learning_rate": 0.0006,
+ "loss": 4.324908256530762,
+ "step": 4915
+ },
+ {
+ "epoch": 68.27947598253274,
+ "grad_norm": 0.014076225459575653,
+ "learning_rate": 0.0006,
+ "loss": 4.177175998687744,
+ "step": 4916
+ },
+ {
+ "epoch": 68.29344978165939,
+ "grad_norm": 0.013906744308769703,
+ "learning_rate": 0.0006,
+ "loss": 4.11588191986084,
+ "step": 4917
+ },
+ {
+ "epoch": 68.30742358078602,
+ "grad_norm": 0.013939064927399158,
+ "learning_rate": 0.0006,
+ "loss": 4.183963775634766,
+ "step": 4918
+ },
+ {
+ "epoch": 68.32139737991267,
+ "grad_norm": 0.014572170563042164,
+ "learning_rate": 0.0006,
+ "loss": 4.19586181640625,
+ "step": 4919
+ },
+ {
+ "epoch": 68.3353711790393,
+ "grad_norm": 0.014085350558161736,
+ "learning_rate": 0.0006,
+ "loss": 4.170892238616943,
+ "step": 4920
+ },
+ {
+ "epoch": 68.34934497816595,
+ "grad_norm": 0.014357863925397396,
+ "learning_rate": 0.0006,
+ "loss": 4.321155548095703,
+ "step": 4921
+ },
+ {
+ "epoch": 68.36331877729258,
+ "grad_norm": 0.01461675763130188,
+ "learning_rate": 0.0006,
+ "loss": 4.293334007263184,
+ "step": 4922
+ },
+ {
+ "epoch": 68.37729257641921,
+ "grad_norm": 0.014828486368060112,
+ "learning_rate": 0.0006,
+ "loss": 4.1439714431762695,
+ "step": 4923
+ },
+ {
+ "epoch": 68.39126637554585,
+ "grad_norm": 0.013730363920331001,
+ "learning_rate": 0.0006,
+ "loss": 4.287973880767822,
+ "step": 4924
+ },
+ {
+ "epoch": 68.40524017467249,
+ "grad_norm": 0.013852203264832497,
+ "learning_rate": 0.0006,
+ "loss": 4.248824119567871,
+ "step": 4925
+ },
+ {
+ "epoch": 68.41921397379913,
+ "grad_norm": 0.014005862176418304,
+ "learning_rate": 0.0006,
+ "loss": 4.293768882751465,
+ "step": 4926
+ },
+ {
+ "epoch": 68.43318777292576,
+ "grad_norm": 0.014435851015150547,
+ "learning_rate": 0.0006,
+ "loss": 4.204813003540039,
+ "step": 4927
+ },
+ {
+ "epoch": 68.44716157205241,
+ "grad_norm": 0.015305934473872185,
+ "learning_rate": 0.0006,
+ "loss": 4.329364776611328,
+ "step": 4928
+ },
+ {
+ "epoch": 68.46113537117904,
+ "grad_norm": 0.014356814324855804,
+ "learning_rate": 0.0006,
+ "loss": 4.264206409454346,
+ "step": 4929
+ },
+ {
+ "epoch": 68.47510917030567,
+ "grad_norm": 0.01432347297668457,
+ "learning_rate": 0.0006,
+ "loss": 4.29440450668335,
+ "step": 4930
+ },
+ {
+ "epoch": 68.48908296943232,
+ "grad_norm": 0.012672432698309422,
+ "learning_rate": 0.0006,
+ "loss": 4.258024215698242,
+ "step": 4931
+ },
+ {
+ "epoch": 68.50305676855895,
+ "grad_norm": 0.014209977351129055,
+ "learning_rate": 0.0006,
+ "loss": 4.216668128967285,
+ "step": 4932
+ },
+ {
+ "epoch": 68.5170305676856,
+ "grad_norm": 0.014606663957238197,
+ "learning_rate": 0.0006,
+ "loss": 4.294184684753418,
+ "step": 4933
+ },
+ {
+ "epoch": 68.53100436681223,
+ "grad_norm": 0.014986990951001644,
+ "learning_rate": 0.0006,
+ "loss": 4.382591247558594,
+ "step": 4934
+ },
+ {
+ "epoch": 68.54497816593886,
+ "grad_norm": 0.013998783193528652,
+ "learning_rate": 0.0006,
+ "loss": 4.195143222808838,
+ "step": 4935
+ },
+ {
+ "epoch": 68.5589519650655,
+ "grad_norm": 0.015011102892458439,
+ "learning_rate": 0.0006,
+ "loss": 4.313785076141357,
+ "step": 4936
+ },
+ {
+ "epoch": 68.57292576419214,
+ "grad_norm": 0.01641209051012993,
+ "learning_rate": 0.0006,
+ "loss": 4.322803020477295,
+ "step": 4937
+ },
+ {
+ "epoch": 68.58689956331878,
+ "grad_norm": 0.016207464039325714,
+ "learning_rate": 0.0006,
+ "loss": 4.311431884765625,
+ "step": 4938
+ },
+ {
+ "epoch": 68.60087336244541,
+ "grad_norm": 0.01747811958193779,
+ "learning_rate": 0.0006,
+ "loss": 4.310845375061035,
+ "step": 4939
+ },
+ {
+ "epoch": 68.61484716157206,
+ "grad_norm": 0.019475717097520828,
+ "learning_rate": 0.0006,
+ "loss": 4.338624954223633,
+ "step": 4940
+ },
+ {
+ "epoch": 68.62882096069869,
+ "grad_norm": 0.018641719594597816,
+ "learning_rate": 0.0006,
+ "loss": 4.34913969039917,
+ "step": 4941
+ },
+ {
+ "epoch": 68.64279475982532,
+ "grad_norm": 0.018376432359218597,
+ "learning_rate": 0.0006,
+ "loss": 4.28464412689209,
+ "step": 4942
+ },
+ {
+ "epoch": 68.65676855895197,
+ "grad_norm": 0.017095010727643967,
+ "learning_rate": 0.0006,
+ "loss": 4.217957496643066,
+ "step": 4943
+ },
+ {
+ "epoch": 68.6707423580786,
+ "grad_norm": 0.016321983188390732,
+ "learning_rate": 0.0006,
+ "loss": 4.220042705535889,
+ "step": 4944
+ },
+ {
+ "epoch": 68.68471615720524,
+ "grad_norm": 0.017128000035881996,
+ "learning_rate": 0.0006,
+ "loss": 4.293500900268555,
+ "step": 4945
+ },
+ {
+ "epoch": 68.69868995633188,
+ "grad_norm": 0.015592411160469055,
+ "learning_rate": 0.0006,
+ "loss": 4.195798873901367,
+ "step": 4946
+ },
+ {
+ "epoch": 68.71266375545852,
+ "grad_norm": 0.016607079654932022,
+ "learning_rate": 0.0006,
+ "loss": 4.2654337882995605,
+ "step": 4947
+ },
+ {
+ "epoch": 68.72663755458515,
+ "grad_norm": 0.01545762736350298,
+ "learning_rate": 0.0006,
+ "loss": 4.303828239440918,
+ "step": 4948
+ },
+ {
+ "epoch": 68.74061135371178,
+ "grad_norm": 0.014328422956168652,
+ "learning_rate": 0.0006,
+ "loss": 4.29379940032959,
+ "step": 4949
+ },
+ {
+ "epoch": 68.75458515283843,
+ "grad_norm": 0.016516804695129395,
+ "learning_rate": 0.0006,
+ "loss": 4.2317214012146,
+ "step": 4950
+ },
+ {
+ "epoch": 68.76855895196506,
+ "grad_norm": 0.016324171796441078,
+ "learning_rate": 0.0006,
+ "loss": 4.278043746948242,
+ "step": 4951
+ },
+ {
+ "epoch": 68.78253275109171,
+ "grad_norm": 0.015511159785091877,
+ "learning_rate": 0.0006,
+ "loss": 4.188244342803955,
+ "step": 4952
+ },
+ {
+ "epoch": 68.79650655021834,
+ "grad_norm": 0.01676103286445141,
+ "learning_rate": 0.0006,
+ "loss": 4.2169976234436035,
+ "step": 4953
+ },
+ {
+ "epoch": 68.81048034934499,
+ "grad_norm": 0.015428312122821808,
+ "learning_rate": 0.0006,
+ "loss": 4.4004716873168945,
+ "step": 4954
+ },
+ {
+ "epoch": 68.82445414847162,
+ "grad_norm": 0.0162410456687212,
+ "learning_rate": 0.0006,
+ "loss": 4.282752513885498,
+ "step": 4955
+ },
+ {
+ "epoch": 68.83842794759825,
+ "grad_norm": 0.017578862607479095,
+ "learning_rate": 0.0006,
+ "loss": 4.098004341125488,
+ "step": 4956
+ },
+ {
+ "epoch": 68.8524017467249,
+ "grad_norm": 0.01805749163031578,
+ "learning_rate": 0.0006,
+ "loss": 4.277988433837891,
+ "step": 4957
+ },
+ {
+ "epoch": 68.86637554585153,
+ "grad_norm": 0.016728045418858528,
+ "learning_rate": 0.0006,
+ "loss": 4.29766321182251,
+ "step": 4958
+ },
+ {
+ "epoch": 68.88034934497817,
+ "grad_norm": 0.014899657107889652,
+ "learning_rate": 0.0006,
+ "loss": 4.329736709594727,
+ "step": 4959
+ },
+ {
+ "epoch": 68.8943231441048,
+ "grad_norm": 0.013441305607557297,
+ "learning_rate": 0.0006,
+ "loss": 4.2873215675354,
+ "step": 4960
+ },
+ {
+ "epoch": 68.90829694323143,
+ "grad_norm": 0.01554128248244524,
+ "learning_rate": 0.0006,
+ "loss": 4.278327941894531,
+ "step": 4961
+ },
+ {
+ "epoch": 68.92227074235808,
+ "grad_norm": 0.01806696318089962,
+ "learning_rate": 0.0006,
+ "loss": 4.38275146484375,
+ "step": 4962
+ },
+ {
+ "epoch": 68.93624454148471,
+ "grad_norm": 0.017025409266352654,
+ "learning_rate": 0.0006,
+ "loss": 4.210447311401367,
+ "step": 4963
+ },
+ {
+ "epoch": 68.95021834061136,
+ "grad_norm": 0.01597420684993267,
+ "learning_rate": 0.0006,
+ "loss": 4.309865951538086,
+ "step": 4964
+ },
+ {
+ "epoch": 68.96419213973799,
+ "grad_norm": 0.015835624188184738,
+ "learning_rate": 0.0006,
+ "loss": 4.225904941558838,
+ "step": 4965
+ },
+ {
+ "epoch": 68.97816593886463,
+ "grad_norm": 0.018170492723584175,
+ "learning_rate": 0.0006,
+ "loss": 4.248478889465332,
+ "step": 4966
+ },
+ {
+ "epoch": 68.99213973799127,
+ "grad_norm": 0.017372991889715195,
+ "learning_rate": 0.0006,
+ "loss": 4.283486843109131,
+ "step": 4967
+ },
+ {
+ "epoch": 69.0,
+ "grad_norm": 0.016305390745401382,
+ "learning_rate": 0.0006,
+ "loss": 4.273636817932129,
+ "step": 4968
+ },
+ {
+ "epoch": 69.0,
+ "eval_loss": 4.66873836517334,
+ "eval_runtime": 57.0791,
+ "eval_samples_per_second": 42.783,
+ "eval_steps_per_second": 1.349,
+ "step": 4968
+ },
+ {
+ "epoch": 69.01397379912663,
+ "grad_norm": 0.016567355021834373,
+ "learning_rate": 0.0006,
+ "loss": 4.269988059997559,
+ "step": 4969
+ },
+ {
+ "epoch": 69.02794759825328,
+ "grad_norm": 0.017476852983236313,
+ "learning_rate": 0.0006,
+ "loss": 4.2518720626831055,
+ "step": 4970
+ },
+ {
+ "epoch": 69.04192139737991,
+ "grad_norm": 0.017070338129997253,
+ "learning_rate": 0.0006,
+ "loss": 4.238024711608887,
+ "step": 4971
+ },
+ {
+ "epoch": 69.05589519650655,
+ "grad_norm": 0.01639373227953911,
+ "learning_rate": 0.0006,
+ "loss": 4.188558101654053,
+ "step": 4972
+ },
+ {
+ "epoch": 69.06986899563319,
+ "grad_norm": 0.01856113411486149,
+ "learning_rate": 0.0006,
+ "loss": 4.339556694030762,
+ "step": 4973
+ },
+ {
+ "epoch": 69.08384279475983,
+ "grad_norm": 0.01762578636407852,
+ "learning_rate": 0.0006,
+ "loss": 4.141518592834473,
+ "step": 4974
+ },
+ {
+ "epoch": 69.09781659388646,
+ "grad_norm": 0.015437949448823929,
+ "learning_rate": 0.0006,
+ "loss": 4.225461959838867,
+ "step": 4975
+ },
+ {
+ "epoch": 69.1117903930131,
+ "grad_norm": 0.015293503180146217,
+ "learning_rate": 0.0006,
+ "loss": 4.290884971618652,
+ "step": 4976
+ },
+ {
+ "epoch": 69.12576419213974,
+ "grad_norm": 0.016259880736470222,
+ "learning_rate": 0.0006,
+ "loss": 4.121919631958008,
+ "step": 4977
+ },
+ {
+ "epoch": 69.13973799126637,
+ "grad_norm": 0.017225584015250206,
+ "learning_rate": 0.0006,
+ "loss": 4.136716842651367,
+ "step": 4978
+ },
+ {
+ "epoch": 69.15371179039302,
+ "grad_norm": 0.01699289306998253,
+ "learning_rate": 0.0006,
+ "loss": 4.193601608276367,
+ "step": 4979
+ },
+ {
+ "epoch": 69.16768558951965,
+ "grad_norm": 0.015613161027431488,
+ "learning_rate": 0.0006,
+ "loss": 4.288212299346924,
+ "step": 4980
+ },
+ {
+ "epoch": 69.18165938864628,
+ "grad_norm": 0.01643933728337288,
+ "learning_rate": 0.0006,
+ "loss": 4.324707984924316,
+ "step": 4981
+ },
+ {
+ "epoch": 69.19563318777293,
+ "grad_norm": 0.01648905873298645,
+ "learning_rate": 0.0006,
+ "loss": 4.287806510925293,
+ "step": 4982
+ },
+ {
+ "epoch": 69.20960698689956,
+ "grad_norm": 0.015444950200617313,
+ "learning_rate": 0.0006,
+ "loss": 4.200622081756592,
+ "step": 4983
+ },
+ {
+ "epoch": 69.2235807860262,
+ "grad_norm": 0.016034258529543877,
+ "learning_rate": 0.0006,
+ "loss": 4.2193827629089355,
+ "step": 4984
+ },
+ {
+ "epoch": 69.23755458515284,
+ "grad_norm": 0.017984388396143913,
+ "learning_rate": 0.0006,
+ "loss": 4.18237829208374,
+ "step": 4985
+ },
+ {
+ "epoch": 69.25152838427948,
+ "grad_norm": 0.0195505041629076,
+ "learning_rate": 0.0006,
+ "loss": 4.216727256774902,
+ "step": 4986
+ },
+ {
+ "epoch": 69.26550218340611,
+ "grad_norm": 0.017779147252440453,
+ "learning_rate": 0.0006,
+ "loss": 4.278628826141357,
+ "step": 4987
+ },
+ {
+ "epoch": 69.27947598253274,
+ "grad_norm": 0.017119476571679115,
+ "learning_rate": 0.0006,
+ "loss": 4.2409138679504395,
+ "step": 4988
+ },
+ {
+ "epoch": 69.29344978165939,
+ "grad_norm": 0.016840364784002304,
+ "learning_rate": 0.0006,
+ "loss": 4.207575798034668,
+ "step": 4989
+ },
+ {
+ "epoch": 69.30742358078602,
+ "grad_norm": 0.014593573287129402,
+ "learning_rate": 0.0006,
+ "loss": 4.164905548095703,
+ "step": 4990
+ },
+ {
+ "epoch": 69.32139737991267,
+ "grad_norm": 0.013922630809247494,
+ "learning_rate": 0.0006,
+ "loss": 4.2696990966796875,
+ "step": 4991
+ },
+ {
+ "epoch": 69.3353711790393,
+ "grad_norm": 0.01577564887702465,
+ "learning_rate": 0.0006,
+ "loss": 4.201589107513428,
+ "step": 4992
+ },
+ {
+ "epoch": 69.34934497816595,
+ "grad_norm": 0.016392096877098083,
+ "learning_rate": 0.0006,
+ "loss": 4.248831748962402,
+ "step": 4993
+ },
+ {
+ "epoch": 69.36331877729258,
+ "grad_norm": 0.015961311757564545,
+ "learning_rate": 0.0006,
+ "loss": 4.181260108947754,
+ "step": 4994
+ },
+ {
+ "epoch": 69.37729257641921,
+ "grad_norm": 0.015350551344454288,
+ "learning_rate": 0.0006,
+ "loss": 4.230352878570557,
+ "step": 4995
+ },
+ {
+ "epoch": 69.39126637554585,
+ "grad_norm": 0.015870466828346252,
+ "learning_rate": 0.0006,
+ "loss": 4.298305511474609,
+ "step": 4996
+ },
+ {
+ "epoch": 69.40524017467249,
+ "grad_norm": 0.013716642744839191,
+ "learning_rate": 0.0006,
+ "loss": 4.220139503479004,
+ "step": 4997
+ },
+ {
+ "epoch": 69.41921397379913,
+ "grad_norm": 0.014431777410209179,
+ "learning_rate": 0.0006,
+ "loss": 4.288130760192871,
+ "step": 4998
+ },
+ {
+ "epoch": 69.43318777292576,
+ "grad_norm": 0.01426647324115038,
+ "learning_rate": 0.0006,
+ "loss": 4.295767784118652,
+ "step": 4999
+ },
+ {
+ "epoch": 69.44716157205241,
+ "grad_norm": 0.014286615885794163,
+ "learning_rate": 0.0006,
+ "loss": 4.38106632232666,
+ "step": 5000
+ },
+ {
+ "epoch": 69.46113537117904,
+ "grad_norm": 0.017990585416555405,
+ "learning_rate": 0.0006,
+ "loss": 4.1593017578125,
+ "step": 5001
+ },
+ {
+ "epoch": 69.47510917030567,
+ "grad_norm": 0.018846049904823303,
+ "learning_rate": 0.0006,
+ "loss": 4.216091632843018,
+ "step": 5002
+ },
+ {
+ "epoch": 69.48908296943232,
+ "grad_norm": 0.018179181963205338,
+ "learning_rate": 0.0006,
+ "loss": 4.240332126617432,
+ "step": 5003
+ },
+ {
+ "epoch": 69.50305676855895,
+ "grad_norm": 0.016693448647856712,
+ "learning_rate": 0.0006,
+ "loss": 4.256624221801758,
+ "step": 5004
+ },
+ {
+ "epoch": 69.5170305676856,
+ "grad_norm": 0.01718989387154579,
+ "learning_rate": 0.0006,
+ "loss": 4.300940990447998,
+ "step": 5005
+ },
+ {
+ "epoch": 69.53100436681223,
+ "grad_norm": 0.020236829295754433,
+ "learning_rate": 0.0006,
+ "loss": 4.289146423339844,
+ "step": 5006
+ },
+ {
+ "epoch": 69.54497816593886,
+ "grad_norm": 0.024860285222530365,
+ "learning_rate": 0.0006,
+ "loss": 4.216216087341309,
+ "step": 5007
+ },
+ {
+ "epoch": 69.5589519650655,
+ "grad_norm": 0.02559969387948513,
+ "learning_rate": 0.0006,
+ "loss": 4.262890815734863,
+ "step": 5008
+ },
+ {
+ "epoch": 69.57292576419214,
+ "grad_norm": 0.019180957227945328,
+ "learning_rate": 0.0006,
+ "loss": 4.230309009552002,
+ "step": 5009
+ },
+ {
+ "epoch": 69.58689956331878,
+ "grad_norm": 0.020442591980099678,
+ "learning_rate": 0.0006,
+ "loss": 4.291759490966797,
+ "step": 5010
+ },
+ {
+ "epoch": 69.60087336244541,
+ "grad_norm": 0.0212619137018919,
+ "learning_rate": 0.0006,
+ "loss": 4.3529839515686035,
+ "step": 5011
+ },
+ {
+ "epoch": 69.61484716157206,
+ "grad_norm": 0.02067175693809986,
+ "learning_rate": 0.0006,
+ "loss": 4.33197021484375,
+ "step": 5012
+ },
+ {
+ "epoch": 69.62882096069869,
+ "grad_norm": 0.01802872307598591,
+ "learning_rate": 0.0006,
+ "loss": 4.2101969718933105,
+ "step": 5013
+ },
+ {
+ "epoch": 69.64279475982532,
+ "grad_norm": 0.01787702925503254,
+ "learning_rate": 0.0006,
+ "loss": 4.270336627960205,
+ "step": 5014
+ },
+ {
+ "epoch": 69.65676855895197,
+ "grad_norm": 0.01586683839559555,
+ "learning_rate": 0.0006,
+ "loss": 4.222309589385986,
+ "step": 5015
+ },
+ {
+ "epoch": 69.6707423580786,
+ "grad_norm": 0.01717999391257763,
+ "learning_rate": 0.0006,
+ "loss": 4.253704071044922,
+ "step": 5016
+ },
+ {
+ "epoch": 69.68471615720524,
+ "grad_norm": 0.01594279147684574,
+ "learning_rate": 0.0006,
+ "loss": 4.344658851623535,
+ "step": 5017
+ },
+ {
+ "epoch": 69.69868995633188,
+ "grad_norm": 0.014419942162930965,
+ "learning_rate": 0.0006,
+ "loss": 4.2144880294799805,
+ "step": 5018
+ },
+ {
+ "epoch": 69.71266375545852,
+ "grad_norm": 0.015824224799871445,
+ "learning_rate": 0.0006,
+ "loss": 4.226252555847168,
+ "step": 5019
+ },
+ {
+ "epoch": 69.72663755458515,
+ "grad_norm": 0.013329979032278061,
+ "learning_rate": 0.0006,
+ "loss": 4.238587379455566,
+ "step": 5020
+ },
+ {
+ "epoch": 69.74061135371178,
+ "grad_norm": 0.013745979405939579,
+ "learning_rate": 0.0006,
+ "loss": 4.317209243774414,
+ "step": 5021
+ },
+ {
+ "epoch": 69.75458515283843,
+ "grad_norm": 0.014064167626202106,
+ "learning_rate": 0.0006,
+ "loss": 4.316982746124268,
+ "step": 5022
+ },
+ {
+ "epoch": 69.76855895196506,
+ "grad_norm": 0.014611301012337208,
+ "learning_rate": 0.0006,
+ "loss": 4.212728500366211,
+ "step": 5023
+ },
+ {
+ "epoch": 69.78253275109171,
+ "grad_norm": 0.014947416260838509,
+ "learning_rate": 0.0006,
+ "loss": 4.319550514221191,
+ "step": 5024
+ },
+ {
+ "epoch": 69.79650655021834,
+ "grad_norm": 0.014780526980757713,
+ "learning_rate": 0.0006,
+ "loss": 4.1888933181762695,
+ "step": 5025
+ },
+ {
+ "epoch": 69.81048034934499,
+ "grad_norm": 0.01611459068953991,
+ "learning_rate": 0.0006,
+ "loss": 4.239692211151123,
+ "step": 5026
+ },
+ {
+ "epoch": 69.82445414847162,
+ "grad_norm": 0.015395848080515862,
+ "learning_rate": 0.0006,
+ "loss": 4.308794021606445,
+ "step": 5027
+ },
+ {
+ "epoch": 69.83842794759825,
+ "grad_norm": 0.016113396733999252,
+ "learning_rate": 0.0006,
+ "loss": 4.328035354614258,
+ "step": 5028
+ },
+ {
+ "epoch": 69.8524017467249,
+ "grad_norm": 0.016386888921260834,
+ "learning_rate": 0.0006,
+ "loss": 4.272579193115234,
+ "step": 5029
+ },
+ {
+ "epoch": 69.86637554585153,
+ "grad_norm": 0.013675205409526825,
+ "learning_rate": 0.0006,
+ "loss": 4.226552963256836,
+ "step": 5030
+ },
+ {
+ "epoch": 69.88034934497817,
+ "grad_norm": 0.01472668256610632,
+ "learning_rate": 0.0006,
+ "loss": 4.224167346954346,
+ "step": 5031
+ },
+ {
+ "epoch": 69.8943231441048,
+ "grad_norm": 0.014958228915929794,
+ "learning_rate": 0.0006,
+ "loss": 4.232619285583496,
+ "step": 5032
+ },
+ {
+ "epoch": 69.90829694323143,
+ "grad_norm": 0.012926832772791386,
+ "learning_rate": 0.0006,
+ "loss": 4.27107048034668,
+ "step": 5033
+ },
+ {
+ "epoch": 69.92227074235808,
+ "grad_norm": 0.013138801790773869,
+ "learning_rate": 0.0006,
+ "loss": 4.288477420806885,
+ "step": 5034
+ },
+ {
+ "epoch": 69.93624454148471,
+ "grad_norm": 0.013415461406111717,
+ "learning_rate": 0.0006,
+ "loss": 4.2033257484436035,
+ "step": 5035
+ },
+ {
+ "epoch": 69.95021834061136,
+ "grad_norm": 0.013807233422994614,
+ "learning_rate": 0.0006,
+ "loss": 4.274764537811279,
+ "step": 5036
+ },
+ {
+ "epoch": 69.96419213973799,
+ "grad_norm": 0.013929193839430809,
+ "learning_rate": 0.0006,
+ "loss": 4.1854352951049805,
+ "step": 5037
+ },
+ {
+ "epoch": 69.97816593886463,
+ "grad_norm": 0.015497001819312572,
+ "learning_rate": 0.0006,
+ "loss": 4.226498126983643,
+ "step": 5038
+ },
+ {
+ "epoch": 69.99213973799127,
+ "grad_norm": 0.01874048449099064,
+ "learning_rate": 0.0006,
+ "loss": 4.304137706756592,
+ "step": 5039
+ },
+ {
+ "epoch": 70.0,
+ "grad_norm": 0.021745791658759117,
+ "learning_rate": 0.0006,
+ "loss": 4.206809997558594,
+ "step": 5040
+ },
+ {
+ "epoch": 70.0,
+ "eval_loss": 4.675539970397949,
+ "eval_runtime": 56.9034,
+ "eval_samples_per_second": 42.915,
+ "eval_steps_per_second": 1.353,
+ "step": 5040
+ },
+ {
+ "epoch": 70.01397379912663,
+ "grad_norm": 0.02124839462339878,
+ "learning_rate": 0.0006,
+ "loss": 4.198199272155762,
+ "step": 5041
+ },
+ {
+ "epoch": 70.02794759825328,
+ "grad_norm": 0.019432688131928444,
+ "learning_rate": 0.0006,
+ "loss": 4.187261581420898,
+ "step": 5042
+ },
+ {
+ "epoch": 70.04192139737991,
+ "grad_norm": 0.01769290305674076,
+ "learning_rate": 0.0006,
+ "loss": 4.155284881591797,
+ "step": 5043
+ },
+ {
+ "epoch": 70.05589519650655,
+ "grad_norm": 0.01751520484685898,
+ "learning_rate": 0.0006,
+ "loss": 4.386857986450195,
+ "step": 5044
+ },
+ {
+ "epoch": 70.06986899563319,
+ "grad_norm": 0.018138699233531952,
+ "learning_rate": 0.0006,
+ "loss": 4.117342472076416,
+ "step": 5045
+ },
+ {
+ "epoch": 70.08384279475983,
+ "grad_norm": 0.016331054270267487,
+ "learning_rate": 0.0006,
+ "loss": 4.225295066833496,
+ "step": 5046
+ },
+ {
+ "epoch": 70.09781659388646,
+ "grad_norm": 0.01658729277551174,
+ "learning_rate": 0.0006,
+ "loss": 4.143835067749023,
+ "step": 5047
+ },
+ {
+ "epoch": 70.1117903930131,
+ "grad_norm": 0.016367698088288307,
+ "learning_rate": 0.0006,
+ "loss": 4.1469268798828125,
+ "step": 5048
+ },
+ {
+ "epoch": 70.12576419213974,
+ "grad_norm": 0.015151629224419594,
+ "learning_rate": 0.0006,
+ "loss": 4.117629051208496,
+ "step": 5049
+ },
+ {
+ "epoch": 70.13973799126637,
+ "grad_norm": 0.016434503719210625,
+ "learning_rate": 0.0006,
+ "loss": 4.107848644256592,
+ "step": 5050
+ },
+ {
+ "epoch": 70.15371179039302,
+ "grad_norm": 0.017837485298514366,
+ "learning_rate": 0.0006,
+ "loss": 4.10837459564209,
+ "step": 5051
+ },
+ {
+ "epoch": 70.16768558951965,
+ "grad_norm": 0.0165961142629385,
+ "learning_rate": 0.0006,
+ "loss": 4.169414043426514,
+ "step": 5052
+ },
+ {
+ "epoch": 70.18165938864628,
+ "grad_norm": 0.0157014112919569,
+ "learning_rate": 0.0006,
+ "loss": 4.0629706382751465,
+ "step": 5053
+ },
+ {
+ "epoch": 70.19563318777293,
+ "grad_norm": 0.01581883803009987,
+ "learning_rate": 0.0006,
+ "loss": 4.220012664794922,
+ "step": 5054
+ },
+ {
+ "epoch": 70.20960698689956,
+ "grad_norm": 0.015285917557775974,
+ "learning_rate": 0.0006,
+ "loss": 4.275337219238281,
+ "step": 5055
+ },
+ {
+ "epoch": 70.2235807860262,
+ "grad_norm": 0.014984549954533577,
+ "learning_rate": 0.0006,
+ "loss": 4.284764289855957,
+ "step": 5056
+ },
+ {
+ "epoch": 70.23755458515284,
+ "grad_norm": 0.015172012150287628,
+ "learning_rate": 0.0006,
+ "loss": 4.234498023986816,
+ "step": 5057
+ },
+ {
+ "epoch": 70.25152838427948,
+ "grad_norm": 0.016846632584929466,
+ "learning_rate": 0.0006,
+ "loss": 4.181689262390137,
+ "step": 5058
+ },
+ {
+ "epoch": 70.26550218340611,
+ "grad_norm": 0.01920711249113083,
+ "learning_rate": 0.0006,
+ "loss": 4.235280513763428,
+ "step": 5059
+ },
+ {
+ "epoch": 70.27947598253274,
+ "grad_norm": 0.017765291035175323,
+ "learning_rate": 0.0006,
+ "loss": 4.238955497741699,
+ "step": 5060
+ },
+ {
+ "epoch": 70.29344978165939,
+ "grad_norm": 0.015436639077961445,
+ "learning_rate": 0.0006,
+ "loss": 4.232516288757324,
+ "step": 5061
+ },
+ {
+ "epoch": 70.30742358078602,
+ "grad_norm": 0.015660086646676064,
+ "learning_rate": 0.0006,
+ "loss": 4.260349273681641,
+ "step": 5062
+ },
+ {
+ "epoch": 70.32139737991267,
+ "grad_norm": 0.016095709055662155,
+ "learning_rate": 0.0006,
+ "loss": 4.168505668640137,
+ "step": 5063
+ },
+ {
+ "epoch": 70.3353711790393,
+ "grad_norm": 0.015001940540969372,
+ "learning_rate": 0.0006,
+ "loss": 4.161739349365234,
+ "step": 5064
+ },
+ {
+ "epoch": 70.34934497816595,
+ "grad_norm": 0.014668774791061878,
+ "learning_rate": 0.0006,
+ "loss": 4.181029796600342,
+ "step": 5065
+ },
+ {
+ "epoch": 70.36331877729258,
+ "grad_norm": 0.013266284950077534,
+ "learning_rate": 0.0006,
+ "loss": 4.3052873611450195,
+ "step": 5066
+ },
+ {
+ "epoch": 70.37729257641921,
+ "grad_norm": 0.015396138653159142,
+ "learning_rate": 0.0006,
+ "loss": 4.34413480758667,
+ "step": 5067
+ },
+ {
+ "epoch": 70.39126637554585,
+ "grad_norm": 0.015511523932218552,
+ "learning_rate": 0.0006,
+ "loss": 4.235037803649902,
+ "step": 5068
+ },
+ {
+ "epoch": 70.40524017467249,
+ "grad_norm": 0.014825738966464996,
+ "learning_rate": 0.0006,
+ "loss": 4.262448310852051,
+ "step": 5069
+ },
+ {
+ "epoch": 70.41921397379913,
+ "grad_norm": 0.015309952199459076,
+ "learning_rate": 0.0006,
+ "loss": 4.258711814880371,
+ "step": 5070
+ },
+ {
+ "epoch": 70.43318777292576,
+ "grad_norm": 0.017614055424928665,
+ "learning_rate": 0.0006,
+ "loss": 4.286028861999512,
+ "step": 5071
+ },
+ {
+ "epoch": 70.44716157205241,
+ "grad_norm": 0.0166039876639843,
+ "learning_rate": 0.0006,
+ "loss": 4.149197578430176,
+ "step": 5072
+ },
+ {
+ "epoch": 70.46113537117904,
+ "grad_norm": 0.016911419108510017,
+ "learning_rate": 0.0006,
+ "loss": 4.271981239318848,
+ "step": 5073
+ },
+ {
+ "epoch": 70.47510917030567,
+ "grad_norm": 0.016806820407509804,
+ "learning_rate": 0.0006,
+ "loss": 4.213315010070801,
+ "step": 5074
+ },
+ {
+ "epoch": 70.48908296943232,
+ "grad_norm": 0.015221742913126945,
+ "learning_rate": 0.0006,
+ "loss": 4.171091079711914,
+ "step": 5075
+ },
+ {
+ "epoch": 70.50305676855895,
+ "grad_norm": 0.014175347983837128,
+ "learning_rate": 0.0006,
+ "loss": 4.185309410095215,
+ "step": 5076
+ },
+ {
+ "epoch": 70.5170305676856,
+ "grad_norm": 0.013897798955440521,
+ "learning_rate": 0.0006,
+ "loss": 4.250119209289551,
+ "step": 5077
+ },
+ {
+ "epoch": 70.53100436681223,
+ "grad_norm": 0.014791003428399563,
+ "learning_rate": 0.0006,
+ "loss": 4.2504682540893555,
+ "step": 5078
+ },
+ {
+ "epoch": 70.54497816593886,
+ "grad_norm": 0.016700128093361855,
+ "learning_rate": 0.0006,
+ "loss": 4.235134124755859,
+ "step": 5079
+ },
+ {
+ "epoch": 70.5589519650655,
+ "grad_norm": 0.016679363325238228,
+ "learning_rate": 0.0006,
+ "loss": 4.32028865814209,
+ "step": 5080
+ },
+ {
+ "epoch": 70.57292576419214,
+ "grad_norm": 0.015117401257157326,
+ "learning_rate": 0.0006,
+ "loss": 4.210995674133301,
+ "step": 5081
+ },
+ {
+ "epoch": 70.58689956331878,
+ "grad_norm": 0.014441263861954212,
+ "learning_rate": 0.0006,
+ "loss": 4.27849006652832,
+ "step": 5082
+ },
+ {
+ "epoch": 70.60087336244541,
+ "grad_norm": 0.01523780357092619,
+ "learning_rate": 0.0006,
+ "loss": 4.197360992431641,
+ "step": 5083
+ },
+ {
+ "epoch": 70.61484716157206,
+ "grad_norm": 0.016201738268136978,
+ "learning_rate": 0.0006,
+ "loss": 4.28239107131958,
+ "step": 5084
+ },
+ {
+ "epoch": 70.62882096069869,
+ "grad_norm": 0.016003292053937912,
+ "learning_rate": 0.0006,
+ "loss": 4.209715843200684,
+ "step": 5085
+ },
+ {
+ "epoch": 70.64279475982532,
+ "grad_norm": 0.01472893450409174,
+ "learning_rate": 0.0006,
+ "loss": 4.169078826904297,
+ "step": 5086
+ },
+ {
+ "epoch": 70.65676855895197,
+ "grad_norm": 0.01604326069355011,
+ "learning_rate": 0.0006,
+ "loss": 4.090967178344727,
+ "step": 5087
+ },
+ {
+ "epoch": 70.6707423580786,
+ "grad_norm": 0.01598356105387211,
+ "learning_rate": 0.0006,
+ "loss": 4.16630220413208,
+ "step": 5088
+ },
+ {
+ "epoch": 70.68471615720524,
+ "grad_norm": 0.016988780349493027,
+ "learning_rate": 0.0006,
+ "loss": 4.117822170257568,
+ "step": 5089
+ },
+ {
+ "epoch": 70.69868995633188,
+ "grad_norm": 0.018582893535494804,
+ "learning_rate": 0.0006,
+ "loss": 4.192902088165283,
+ "step": 5090
+ },
+ {
+ "epoch": 70.71266375545852,
+ "grad_norm": 0.01793155074119568,
+ "learning_rate": 0.0006,
+ "loss": 4.314806938171387,
+ "step": 5091
+ },
+ {
+ "epoch": 70.72663755458515,
+ "grad_norm": 0.016286427155137062,
+ "learning_rate": 0.0006,
+ "loss": 4.2137556076049805,
+ "step": 5092
+ },
+ {
+ "epoch": 70.74061135371178,
+ "grad_norm": 0.014718376100063324,
+ "learning_rate": 0.0006,
+ "loss": 4.3073015213012695,
+ "step": 5093
+ },
+ {
+ "epoch": 70.75458515283843,
+ "grad_norm": 0.014675077050924301,
+ "learning_rate": 0.0006,
+ "loss": 4.160961151123047,
+ "step": 5094
+ },
+ {
+ "epoch": 70.76855895196506,
+ "grad_norm": 0.015550913289189339,
+ "learning_rate": 0.0006,
+ "loss": 4.322887420654297,
+ "step": 5095
+ },
+ {
+ "epoch": 70.78253275109171,
+ "grad_norm": 0.01453235000371933,
+ "learning_rate": 0.0006,
+ "loss": 4.2915873527526855,
+ "step": 5096
+ },
+ {
+ "epoch": 70.79650655021834,
+ "grad_norm": 0.01459164172410965,
+ "learning_rate": 0.0006,
+ "loss": 4.247288227081299,
+ "step": 5097
+ },
+ {
+ "epoch": 70.81048034934499,
+ "grad_norm": 0.015395080670714378,
+ "learning_rate": 0.0006,
+ "loss": 4.360955715179443,
+ "step": 5098
+ },
+ {
+ "epoch": 70.82445414847162,
+ "grad_norm": 0.01317683607339859,
+ "learning_rate": 0.0006,
+ "loss": 4.181969165802002,
+ "step": 5099
+ },
+ {
+ "epoch": 70.83842794759825,
+ "grad_norm": 0.013917932286858559,
+ "learning_rate": 0.0006,
+ "loss": 4.264626979827881,
+ "step": 5100
+ },
+ {
+ "epoch": 70.8524017467249,
+ "grad_norm": 0.015268695540726185,
+ "learning_rate": 0.0006,
+ "loss": 4.287857532501221,
+ "step": 5101
+ },
+ {
+ "epoch": 70.86637554585153,
+ "grad_norm": 0.015531329438090324,
+ "learning_rate": 0.0006,
+ "loss": 4.294188499450684,
+ "step": 5102
+ },
+ {
+ "epoch": 70.88034934497817,
+ "grad_norm": 0.015607129782438278,
+ "learning_rate": 0.0006,
+ "loss": 4.331040859222412,
+ "step": 5103
+ },
+ {
+ "epoch": 70.8943231441048,
+ "grad_norm": 0.015297149308025837,
+ "learning_rate": 0.0006,
+ "loss": 4.243650436401367,
+ "step": 5104
+ },
+ {
+ "epoch": 70.90829694323143,
+ "grad_norm": 0.014790025539696217,
+ "learning_rate": 0.0006,
+ "loss": 4.159859657287598,
+ "step": 5105
+ },
+ {
+ "epoch": 70.92227074235808,
+ "grad_norm": 0.013466274365782738,
+ "learning_rate": 0.0006,
+ "loss": 4.221992492675781,
+ "step": 5106
+ },
+ {
+ "epoch": 70.93624454148471,
+ "grad_norm": 0.01361898984760046,
+ "learning_rate": 0.0006,
+ "loss": 4.278751373291016,
+ "step": 5107
+ },
+ {
+ "epoch": 70.95021834061136,
+ "grad_norm": 0.01325751468539238,
+ "learning_rate": 0.0006,
+ "loss": 4.166001319885254,
+ "step": 5108
+ },
+ {
+ "epoch": 70.96419213973799,
+ "grad_norm": 0.012727087363600731,
+ "learning_rate": 0.0006,
+ "loss": 4.219997406005859,
+ "step": 5109
+ },
+ {
+ "epoch": 70.97816593886463,
+ "grad_norm": 0.013966867700219154,
+ "learning_rate": 0.0006,
+ "loss": 4.207491397857666,
+ "step": 5110
+ },
+ {
+ "epoch": 70.99213973799127,
+ "grad_norm": 0.015053587965667248,
+ "learning_rate": 0.0006,
+ "loss": 4.343245983123779,
+ "step": 5111
+ },
+ {
+ "epoch": 71.0,
+ "grad_norm": 0.017723368480801582,
+ "learning_rate": 0.0006,
+ "loss": 4.356122016906738,
+ "step": 5112
+ },
+ {
+ "epoch": 71.0,
+ "eval_loss": 4.598303318023682,
+ "eval_runtime": 56.7521,
+ "eval_samples_per_second": 43.029,
+ "eval_steps_per_second": 1.357,
+ "step": 5112
+ },
+ {
+ "epoch": 71.01397379912663,
+ "grad_norm": 0.018996279686689377,
+ "learning_rate": 0.0006,
+ "loss": 4.234772205352783,
+ "step": 5113
+ },
+ {
+ "epoch": 71.02794759825328,
+ "grad_norm": 0.02041555754840374,
+ "learning_rate": 0.0006,
+ "loss": 4.275816917419434,
+ "step": 5114
+ },
+ {
+ "epoch": 71.04192139737991,
+ "grad_norm": 0.021955285221338272,
+ "learning_rate": 0.0006,
+ "loss": 4.166038513183594,
+ "step": 5115
+ },
+ {
+ "epoch": 71.05589519650655,
+ "grad_norm": 0.021308740600943565,
+ "learning_rate": 0.0006,
+ "loss": 4.13063907623291,
+ "step": 5116
+ },
+ {
+ "epoch": 71.06986899563319,
+ "grad_norm": 0.01747054047882557,
+ "learning_rate": 0.0006,
+ "loss": 4.200522422790527,
+ "step": 5117
+ },
+ {
+ "epoch": 71.08384279475983,
+ "grad_norm": 0.018544532358646393,
+ "learning_rate": 0.0006,
+ "loss": 4.173915386199951,
+ "step": 5118
+ },
+ {
+ "epoch": 71.09781659388646,
+ "grad_norm": 0.0182361900806427,
+ "learning_rate": 0.0006,
+ "loss": 4.27392578125,
+ "step": 5119
+ },
+ {
+ "epoch": 71.1117903930131,
+ "grad_norm": 0.017062073573470116,
+ "learning_rate": 0.0006,
+ "loss": 4.219576358795166,
+ "step": 5120
+ },
+ {
+ "epoch": 71.12576419213974,
+ "grad_norm": 0.016289666295051575,
+ "learning_rate": 0.0006,
+ "loss": 4.133376598358154,
+ "step": 5121
+ },
+ {
+ "epoch": 71.13973799126637,
+ "grad_norm": 0.014864642173051834,
+ "learning_rate": 0.0006,
+ "loss": 4.168276786804199,
+ "step": 5122
+ },
+ {
+ "epoch": 71.15371179039302,
+ "grad_norm": 0.015309540554881096,
+ "learning_rate": 0.0006,
+ "loss": 4.175103187561035,
+ "step": 5123
+ },
+ {
+ "epoch": 71.16768558951965,
+ "grad_norm": 0.01556059904396534,
+ "learning_rate": 0.0006,
+ "loss": 4.289899826049805,
+ "step": 5124
+ },
+ {
+ "epoch": 71.18165938864628,
+ "grad_norm": 0.01479957066476345,
+ "learning_rate": 0.0006,
+ "loss": 4.341495037078857,
+ "step": 5125
+ },
+ {
+ "epoch": 71.19563318777293,
+ "grad_norm": 0.014762179926037788,
+ "learning_rate": 0.0006,
+ "loss": 4.252333641052246,
+ "step": 5126
+ },
+ {
+ "epoch": 71.20960698689956,
+ "grad_norm": 0.014897473156452179,
+ "learning_rate": 0.0006,
+ "loss": 4.2244181632995605,
+ "step": 5127
+ },
+ {
+ "epoch": 71.2235807860262,
+ "grad_norm": 0.012990163639187813,
+ "learning_rate": 0.0006,
+ "loss": 4.237288475036621,
+ "step": 5128
+ },
+ {
+ "epoch": 71.23755458515284,
+ "grad_norm": 0.012858416885137558,
+ "learning_rate": 0.0006,
+ "loss": 4.265738487243652,
+ "step": 5129
+ },
+ {
+ "epoch": 71.25152838427948,
+ "grad_norm": 0.013454788364470005,
+ "learning_rate": 0.0006,
+ "loss": 4.289131164550781,
+ "step": 5130
+ },
+ {
+ "epoch": 71.26550218340611,
+ "grad_norm": 0.013717238791286945,
+ "learning_rate": 0.0006,
+ "loss": 4.2844014167785645,
+ "step": 5131
+ },
+ {
+ "epoch": 71.27947598253274,
+ "grad_norm": 0.0135411461815238,
+ "learning_rate": 0.0006,
+ "loss": 4.139082908630371,
+ "step": 5132
+ },
+ {
+ "epoch": 71.29344978165939,
+ "grad_norm": 0.01406062301248312,
+ "learning_rate": 0.0006,
+ "loss": 4.226719856262207,
+ "step": 5133
+ },
+ {
+ "epoch": 71.30742358078602,
+ "grad_norm": 0.014027190394699574,
+ "learning_rate": 0.0006,
+ "loss": 4.191699981689453,
+ "step": 5134
+ },
+ {
+ "epoch": 71.32139737991267,
+ "grad_norm": 0.013809907250106335,
+ "learning_rate": 0.0006,
+ "loss": 4.242542266845703,
+ "step": 5135
+ },
+ {
+ "epoch": 71.3353711790393,
+ "grad_norm": 0.013507804833352566,
+ "learning_rate": 0.0006,
+ "loss": 4.068206310272217,
+ "step": 5136
+ },
+ {
+ "epoch": 71.34934497816595,
+ "grad_norm": 0.01432533748447895,
+ "learning_rate": 0.0006,
+ "loss": 4.378222942352295,
+ "step": 5137
+ },
+ {
+ "epoch": 71.36331877729258,
+ "grad_norm": 0.01753002591431141,
+ "learning_rate": 0.0006,
+ "loss": 4.174706935882568,
+ "step": 5138
+ },
+ {
+ "epoch": 71.37729257641921,
+ "grad_norm": 0.020048344507813454,
+ "learning_rate": 0.0006,
+ "loss": 4.19920015335083,
+ "step": 5139
+ },
+ {
+ "epoch": 71.39126637554585,
+ "grad_norm": 0.018551401793956757,
+ "learning_rate": 0.0006,
+ "loss": 4.175182342529297,
+ "step": 5140
+ },
+ {
+ "epoch": 71.40524017467249,
+ "grad_norm": 0.01598959229886532,
+ "learning_rate": 0.0006,
+ "loss": 4.2305169105529785,
+ "step": 5141
+ },
+ {
+ "epoch": 71.41921397379913,
+ "grad_norm": 0.016029560938477516,
+ "learning_rate": 0.0006,
+ "loss": 4.230090618133545,
+ "step": 5142
+ },
+ {
+ "epoch": 71.43318777292576,
+ "grad_norm": 0.01595657505095005,
+ "learning_rate": 0.0006,
+ "loss": 4.138402938842773,
+ "step": 5143
+ },
+ {
+ "epoch": 71.44716157205241,
+ "grad_norm": 0.01623230241239071,
+ "learning_rate": 0.0006,
+ "loss": 4.196106433868408,
+ "step": 5144
+ },
+ {
+ "epoch": 71.46113537117904,
+ "grad_norm": 0.014057597145438194,
+ "learning_rate": 0.0006,
+ "loss": 4.2599897384643555,
+ "step": 5145
+ },
+ {
+ "epoch": 71.47510917030567,
+ "grad_norm": 0.017161158844828606,
+ "learning_rate": 0.0006,
+ "loss": 4.183909893035889,
+ "step": 5146
+ },
+ {
+ "epoch": 71.48908296943232,
+ "grad_norm": 0.017038581892848015,
+ "learning_rate": 0.0006,
+ "loss": 4.373933792114258,
+ "step": 5147
+ },
+ {
+ "epoch": 71.50305676855895,
+ "grad_norm": 0.015851367264986038,
+ "learning_rate": 0.0006,
+ "loss": 4.2620086669921875,
+ "step": 5148
+ },
+ {
+ "epoch": 71.5170305676856,
+ "grad_norm": 0.014598303474485874,
+ "learning_rate": 0.0006,
+ "loss": 4.197151184082031,
+ "step": 5149
+ },
+ {
+ "epoch": 71.53100436681223,
+ "grad_norm": 0.01339802611619234,
+ "learning_rate": 0.0006,
+ "loss": 4.110706329345703,
+ "step": 5150
+ },
+ {
+ "epoch": 71.54497816593886,
+ "grad_norm": 0.015567775815725327,
+ "learning_rate": 0.0006,
+ "loss": 4.1210246086120605,
+ "step": 5151
+ },
+ {
+ "epoch": 71.5589519650655,
+ "grad_norm": 0.015616007149219513,
+ "learning_rate": 0.0006,
+ "loss": 4.129796028137207,
+ "step": 5152
+ },
+ {
+ "epoch": 71.57292576419214,
+ "grad_norm": 0.014198964461684227,
+ "learning_rate": 0.0006,
+ "loss": 4.193232536315918,
+ "step": 5153
+ },
+ {
+ "epoch": 71.58689956331878,
+ "grad_norm": 0.015226383693516254,
+ "learning_rate": 0.0006,
+ "loss": 4.223779678344727,
+ "step": 5154
+ },
+ {
+ "epoch": 71.60087336244541,
+ "grad_norm": 0.015610109083354473,
+ "learning_rate": 0.0006,
+ "loss": 4.223179817199707,
+ "step": 5155
+ },
+ {
+ "epoch": 71.61484716157206,
+ "grad_norm": 0.016034303233027458,
+ "learning_rate": 0.0006,
+ "loss": 4.196593284606934,
+ "step": 5156
+ },
+ {
+ "epoch": 71.62882096069869,
+ "grad_norm": 0.015648581087589264,
+ "learning_rate": 0.0006,
+ "loss": 4.341497421264648,
+ "step": 5157
+ },
+ {
+ "epoch": 71.64279475982532,
+ "grad_norm": 0.014117540791630745,
+ "learning_rate": 0.0006,
+ "loss": 4.194307327270508,
+ "step": 5158
+ },
+ {
+ "epoch": 71.65676855895197,
+ "grad_norm": 0.015612686984241009,
+ "learning_rate": 0.0006,
+ "loss": 4.290152549743652,
+ "step": 5159
+ },
+ {
+ "epoch": 71.6707423580786,
+ "grad_norm": 0.01722259446978569,
+ "learning_rate": 0.0006,
+ "loss": 4.258018493652344,
+ "step": 5160
+ },
+ {
+ "epoch": 71.68471615720524,
+ "grad_norm": 0.015923351049423218,
+ "learning_rate": 0.0006,
+ "loss": 4.21684455871582,
+ "step": 5161
+ },
+ {
+ "epoch": 71.69868995633188,
+ "grad_norm": 0.013663450255990028,
+ "learning_rate": 0.0006,
+ "loss": 4.175601959228516,
+ "step": 5162
+ },
+ {
+ "epoch": 71.71266375545852,
+ "grad_norm": 0.017075596377253532,
+ "learning_rate": 0.0006,
+ "loss": 4.205138683319092,
+ "step": 5163
+ },
+ {
+ "epoch": 71.72663755458515,
+ "grad_norm": 0.01723487675189972,
+ "learning_rate": 0.0006,
+ "loss": 4.08839225769043,
+ "step": 5164
+ },
+ {
+ "epoch": 71.74061135371178,
+ "grad_norm": 0.017355505377054214,
+ "learning_rate": 0.0006,
+ "loss": 4.165130615234375,
+ "step": 5165
+ },
+ {
+ "epoch": 71.75458515283843,
+ "grad_norm": 0.01868261583149433,
+ "learning_rate": 0.0006,
+ "loss": 4.16977071762085,
+ "step": 5166
+ },
+ {
+ "epoch": 71.76855895196506,
+ "grad_norm": 0.01897321827709675,
+ "learning_rate": 0.0006,
+ "loss": 4.173497200012207,
+ "step": 5167
+ },
+ {
+ "epoch": 71.78253275109171,
+ "grad_norm": 0.02045825682580471,
+ "learning_rate": 0.0006,
+ "loss": 4.176192283630371,
+ "step": 5168
+ },
+ {
+ "epoch": 71.79650655021834,
+ "grad_norm": 0.022895624861121178,
+ "learning_rate": 0.0006,
+ "loss": 4.163009166717529,
+ "step": 5169
+ },
+ {
+ "epoch": 71.81048034934499,
+ "grad_norm": 0.02062244527041912,
+ "learning_rate": 0.0006,
+ "loss": 4.156868934631348,
+ "step": 5170
+ },
+ {
+ "epoch": 71.82445414847162,
+ "grad_norm": 0.0176058579236269,
+ "learning_rate": 0.0006,
+ "loss": 4.154374122619629,
+ "step": 5171
+ },
+ {
+ "epoch": 71.83842794759825,
+ "grad_norm": 0.018049508333206177,
+ "learning_rate": 0.0006,
+ "loss": 4.296828269958496,
+ "step": 5172
+ },
+ {
+ "epoch": 71.8524017467249,
+ "grad_norm": 0.019466130062937737,
+ "learning_rate": 0.0006,
+ "loss": 4.19450569152832,
+ "step": 5173
+ },
+ {
+ "epoch": 71.86637554585153,
+ "grad_norm": 0.01838165521621704,
+ "learning_rate": 0.0006,
+ "loss": 4.11812686920166,
+ "step": 5174
+ },
+ {
+ "epoch": 71.88034934497817,
+ "grad_norm": 0.01608051173388958,
+ "learning_rate": 0.0006,
+ "loss": 4.240464210510254,
+ "step": 5175
+ },
+ {
+ "epoch": 71.8943231441048,
+ "grad_norm": 0.015807967633008957,
+ "learning_rate": 0.0006,
+ "loss": 4.275581359863281,
+ "step": 5176
+ },
+ {
+ "epoch": 71.90829694323143,
+ "grad_norm": 0.017366575077176094,
+ "learning_rate": 0.0006,
+ "loss": 4.064750671386719,
+ "step": 5177
+ },
+ {
+ "epoch": 71.92227074235808,
+ "grad_norm": 0.016241202130913734,
+ "learning_rate": 0.0006,
+ "loss": 4.203697681427002,
+ "step": 5178
+ },
+ {
+ "epoch": 71.93624454148471,
+ "grad_norm": 0.019105732440948486,
+ "learning_rate": 0.0006,
+ "loss": 4.140021800994873,
+ "step": 5179
+ },
+ {
+ "epoch": 71.95021834061136,
+ "grad_norm": 0.018492674455046654,
+ "learning_rate": 0.0006,
+ "loss": 4.2534074783325195,
+ "step": 5180
+ },
+ {
+ "epoch": 71.96419213973799,
+ "grad_norm": 0.019301170483231544,
+ "learning_rate": 0.0006,
+ "loss": 4.381392478942871,
+ "step": 5181
+ },
+ {
+ "epoch": 71.97816593886463,
+ "grad_norm": 0.019571911543607712,
+ "learning_rate": 0.0006,
+ "loss": 4.131056785583496,
+ "step": 5182
+ },
+ {
+ "epoch": 71.99213973799127,
+ "grad_norm": 0.01859748177230358,
+ "learning_rate": 0.0006,
+ "loss": 4.278930187225342,
+ "step": 5183
+ },
+ {
+ "epoch": 72.0,
+ "grad_norm": 0.02017994411289692,
+ "learning_rate": 0.0006,
+ "loss": 4.116464614868164,
+ "step": 5184
+ },
+ {
+ "epoch": 72.0,
+ "eval_loss": 4.60784912109375,
+ "eval_runtime": 56.3242,
+ "eval_samples_per_second": 43.356,
+ "eval_steps_per_second": 1.367,
+ "step": 5184
+ },
+ {
+ "epoch": 72.01397379912663,
+ "grad_norm": 0.030090296640992165,
+ "learning_rate": 0.0006,
+ "loss": 4.218588352203369,
+ "step": 5185
+ },
+ {
+ "epoch": 72.02794759825328,
+ "grad_norm": 1.330443024635315,
+ "learning_rate": 0.0006,
+ "loss": 4.417870044708252,
+ "step": 5186
+ },
+ {
+ "epoch": 72.04192139737991,
+ "grad_norm": 0.6991518139839172,
+ "learning_rate": 0.0006,
+ "loss": 8.414234161376953,
+ "step": 5187
+ },
+ {
+ "epoch": 72.05589519650655,
+ "grad_norm": 0.2631438374519348,
+ "learning_rate": 0.0006,
+ "loss": 7.41729736328125,
+ "step": 5188
+ },
+ {
+ "epoch": 72.06986899563319,
+ "grad_norm": 0.12391754984855652,
+ "learning_rate": 0.0006,
+ "loss": 7.209164619445801,
+ "step": 5189
+ },
+ {
+ "epoch": 72.08384279475983,
+ "grad_norm": 0.13519170880317688,
+ "learning_rate": 0.0006,
+ "loss": 6.943539619445801,
+ "step": 5190
+ },
+ {
+ "epoch": 72.09781659388646,
+ "grad_norm": 1.258439302444458,
+ "learning_rate": 0.0006,
+ "loss": 10.267714500427246,
+ "step": 5191
+ },
+ {
+ "epoch": 72.1117903930131,
+ "grad_norm": 1.0296908617019653,
+ "learning_rate": 0.0006,
+ "loss": 9.51768970489502,
+ "step": 5192
+ },
+ {
+ "epoch": 72.12576419213974,
+ "grad_norm": 0.12273016571998596,
+ "learning_rate": 0.0006,
+ "loss": 7.634477615356445,
+ "step": 5193
+ },
+ {
+ "epoch": 72.13973799126637,
+ "grad_norm": 0.06170198693871498,
+ "learning_rate": 0.0006,
+ "loss": 7.592115879058838,
+ "step": 5194
+ },
+ {
+ "epoch": 72.15371179039302,
+ "grad_norm": 0.06295108795166016,
+ "learning_rate": 0.0006,
+ "loss": 7.576314449310303,
+ "step": 5195
+ },
+ {
+ "epoch": 72.16768558951965,
+ "grad_norm": 0.0666993111371994,
+ "learning_rate": 0.0006,
+ "loss": 7.53862190246582,
+ "step": 5196
+ },
+ {
+ "epoch": 72.18165938864628,
+ "grad_norm": 0.052215222269296646,
+ "learning_rate": 0.0006,
+ "loss": 7.502930641174316,
+ "step": 5197
+ },
+ {
+ "epoch": 72.19563318777293,
+ "grad_norm": 0.04978961497545242,
+ "learning_rate": 0.0006,
+ "loss": 7.471630096435547,
+ "step": 5198
+ },
+ {
+ "epoch": 72.20960698689956,
+ "grad_norm": 0.07275046408176422,
+ "learning_rate": 0.0006,
+ "loss": 7.496924877166748,
+ "step": 5199
+ },
+ {
+ "epoch": 72.2235807860262,
+ "grad_norm": 0.04491819068789482,
+ "learning_rate": 0.0006,
+ "loss": 7.424494743347168,
+ "step": 5200
+ },
+ {
+ "epoch": 72.23755458515284,
+ "grad_norm": 0.043718695640563965,
+ "learning_rate": 0.0006,
+ "loss": 7.441090106964111,
+ "step": 5201
+ },
+ {
+ "epoch": 72.25152838427948,
+ "grad_norm": 0.04483497887849808,
+ "learning_rate": 0.0006,
+ "loss": 7.3753862380981445,
+ "step": 5202
+ },
+ {
+ "epoch": 72.26550218340611,
+ "grad_norm": 0.03463944420218468,
+ "learning_rate": 0.0006,
+ "loss": 7.375782012939453,
+ "step": 5203
+ },
+ {
+ "epoch": 72.27947598253274,
+ "grad_norm": 0.0397045724093914,
+ "learning_rate": 0.0006,
+ "loss": 7.358898639678955,
+ "step": 5204
+ },
+ {
+ "epoch": 72.29344978165939,
+ "grad_norm": 0.029339438304305077,
+ "learning_rate": 0.0006,
+ "loss": 7.3774871826171875,
+ "step": 5205
+ },
+ {
+ "epoch": 72.30742358078602,
+ "grad_norm": 0.025469699874520302,
+ "learning_rate": 0.0006,
+ "loss": 7.324203014373779,
+ "step": 5206
+ },
+ {
+ "epoch": 72.32139737991267,
+ "grad_norm": 0.030497785657644272,
+ "learning_rate": 0.0006,
+ "loss": 7.329351902008057,
+ "step": 5207
+ },
+ {
+ "epoch": 72.3353711790393,
+ "grad_norm": 0.03382078558206558,
+ "learning_rate": 0.0006,
+ "loss": 7.289940357208252,
+ "step": 5208
+ },
+ {
+ "epoch": 72.34934497816595,
+ "grad_norm": 0.019869405776262283,
+ "learning_rate": 0.0006,
+ "loss": 7.272582054138184,
+ "step": 5209
+ },
+ {
+ "epoch": 72.36331877729258,
+ "grad_norm": 0.02385753206908703,
+ "learning_rate": 0.0006,
+ "loss": 7.2654266357421875,
+ "step": 5210
+ },
+ {
+ "epoch": 72.37729257641921,
+ "grad_norm": 0.02794312685728073,
+ "learning_rate": 0.0006,
+ "loss": 7.250753402709961,
+ "step": 5211
+ },
+ {
+ "epoch": 72.39126637554585,
+ "grad_norm": 0.02445952221751213,
+ "learning_rate": 0.0006,
+ "loss": 7.293970108032227,
+ "step": 5212
+ },
+ {
+ "epoch": 72.40524017467249,
+ "grad_norm": 0.019660986959934235,
+ "learning_rate": 0.0006,
+ "loss": 7.1948771476745605,
+ "step": 5213
+ },
+ {
+ "epoch": 72.41921397379913,
+ "grad_norm": 0.024110861122608185,
+ "learning_rate": 0.0006,
+ "loss": 7.224499225616455,
+ "step": 5214
+ },
+ {
+ "epoch": 72.43318777292576,
+ "grad_norm": 0.021630609408020973,
+ "learning_rate": 0.0006,
+ "loss": 7.1736884117126465,
+ "step": 5215
+ },
+ {
+ "epoch": 72.44716157205241,
+ "grad_norm": 0.02185880020260811,
+ "learning_rate": 0.0006,
+ "loss": 7.167229652404785,
+ "step": 5216
+ },
+ {
+ "epoch": 72.46113537117904,
+ "grad_norm": 0.018913568928837776,
+ "learning_rate": 0.0006,
+ "loss": 7.196551322937012,
+ "step": 5217
+ },
+ {
+ "epoch": 72.47510917030567,
+ "grad_norm": 0.01802486553788185,
+ "learning_rate": 0.0006,
+ "loss": 7.144286632537842,
+ "step": 5218
+ },
+ {
+ "epoch": 72.48908296943232,
+ "grad_norm": 0.020386485382914543,
+ "learning_rate": 0.0006,
+ "loss": 7.196260929107666,
+ "step": 5219
+ },
+ {
+ "epoch": 72.50305676855895,
+ "grad_norm": 0.020077696070075035,
+ "learning_rate": 0.0006,
+ "loss": 7.138718605041504,
+ "step": 5220
+ },
+ {
+ "epoch": 72.5170305676856,
+ "grad_norm": 0.017224183306097984,
+ "learning_rate": 0.0006,
+ "loss": 7.119478702545166,
+ "step": 5221
+ },
+ {
+ "epoch": 72.53100436681223,
+ "grad_norm": 0.020847434177994728,
+ "learning_rate": 0.0006,
+ "loss": 7.068922996520996,
+ "step": 5222
+ },
+ {
+ "epoch": 72.54497816593886,
+ "grad_norm": 0.021764913573861122,
+ "learning_rate": 0.0006,
+ "loss": 7.071352958679199,
+ "step": 5223
+ },
+ {
+ "epoch": 72.5589519650655,
+ "grad_norm": 0.02697215974330902,
+ "learning_rate": 0.0006,
+ "loss": 7.033344268798828,
+ "step": 5224
+ },
+ {
+ "epoch": 72.57292576419214,
+ "grad_norm": 0.042392998933792114,
+ "learning_rate": 0.0006,
+ "loss": 7.0589919090271,
+ "step": 5225
+ },
+ {
+ "epoch": 72.58689956331878,
+ "grad_norm": 0.06616316735744476,
+ "learning_rate": 0.0006,
+ "loss": 7.060244083404541,
+ "step": 5226
+ },
+ {
+ "epoch": 72.60087336244541,
+ "grad_norm": 0.0828910544514656,
+ "learning_rate": 0.0006,
+ "loss": 7.057173728942871,
+ "step": 5227
+ },
+ {
+ "epoch": 72.61484716157206,
+ "grad_norm": 0.046767015010118484,
+ "learning_rate": 0.0006,
+ "loss": 7.010563850402832,
+ "step": 5228
+ },
+ {
+ "epoch": 72.62882096069869,
+ "grad_norm": 0.03831760957837105,
+ "learning_rate": 0.0006,
+ "loss": 6.9028825759887695,
+ "step": 5229
+ },
+ {
+ "epoch": 72.64279475982532,
+ "grad_norm": 0.05109567195177078,
+ "learning_rate": 0.0006,
+ "loss": 6.923257827758789,
+ "step": 5230
+ },
+ {
+ "epoch": 72.65676855895197,
+ "grad_norm": 0.02897934429347515,
+ "learning_rate": 0.0006,
+ "loss": 6.823014259338379,
+ "step": 5231
+ },
+ {
+ "epoch": 72.6707423580786,
+ "grad_norm": 0.03769192099571228,
+ "learning_rate": 0.0006,
+ "loss": 6.87416410446167,
+ "step": 5232
+ },
+ {
+ "epoch": 72.68471615720524,
+ "grad_norm": 0.03677637130022049,
+ "learning_rate": 0.0006,
+ "loss": 6.7649030685424805,
+ "step": 5233
+ },
+ {
+ "epoch": 72.69868995633188,
+ "grad_norm": 0.0371948778629303,
+ "learning_rate": 0.0006,
+ "loss": 6.749821662902832,
+ "step": 5234
+ },
+ {
+ "epoch": 72.71266375545852,
+ "grad_norm": 0.03057795763015747,
+ "learning_rate": 0.0006,
+ "loss": 6.748945236206055,
+ "step": 5235
+ },
+ {
+ "epoch": 72.72663755458515,
+ "grad_norm": 0.028109321370720863,
+ "learning_rate": 0.0006,
+ "loss": 6.663548946380615,
+ "step": 5236
+ },
+ {
+ "epoch": 72.74061135371178,
+ "grad_norm": 0.02442420832812786,
+ "learning_rate": 0.0006,
+ "loss": 6.636251449584961,
+ "step": 5237
+ },
+ {
+ "epoch": 72.75458515283843,
+ "grad_norm": 0.031841423362493515,
+ "learning_rate": 0.0006,
+ "loss": 6.606883525848389,
+ "step": 5238
+ },
+ {
+ "epoch": 72.76855895196506,
+ "grad_norm": 0.028100458905100822,
+ "learning_rate": 0.0006,
+ "loss": 6.603745937347412,
+ "step": 5239
+ },
+ {
+ "epoch": 72.78253275109171,
+ "grad_norm": 0.03222901374101639,
+ "learning_rate": 0.0006,
+ "loss": 6.463825225830078,
+ "step": 5240
+ },
+ {
+ "epoch": 72.79650655021834,
+ "grad_norm": 0.04766521975398064,
+ "learning_rate": 0.0006,
+ "loss": 6.43734073638916,
+ "step": 5241
+ },
+ {
+ "epoch": 72.81048034934499,
+ "grad_norm": 0.09058783948421478,
+ "learning_rate": 0.0006,
+ "loss": 6.448520183563232,
+ "step": 5242
+ },
+ {
+ "epoch": 72.82445414847162,
+ "grad_norm": 0.08995158225297928,
+ "learning_rate": 0.0006,
+ "loss": 6.522093296051025,
+ "step": 5243
+ },
+ {
+ "epoch": 72.83842794759825,
+ "grad_norm": 0.05491040274500847,
+ "learning_rate": 0.0006,
+ "loss": 6.37354040145874,
+ "step": 5244
+ },
+ {
+ "epoch": 72.8524017467249,
+ "grad_norm": 0.05198514088988304,
+ "learning_rate": 0.0006,
+ "loss": 6.362032413482666,
+ "step": 5245
+ },
+ {
+ "epoch": 72.86637554585153,
+ "grad_norm": 0.05464765429496765,
+ "learning_rate": 0.0006,
+ "loss": 6.326773643493652,
+ "step": 5246
+ },
+ {
+ "epoch": 72.88034934497817,
+ "grad_norm": 0.07337361574172974,
+ "learning_rate": 0.0006,
+ "loss": 6.297532081604004,
+ "step": 5247
+ },
+ {
+ "epoch": 72.8943231441048,
+ "grad_norm": 0.10061230510473251,
+ "learning_rate": 0.0006,
+ "loss": 6.359557628631592,
+ "step": 5248
+ },
+ {
+ "epoch": 72.90829694323143,
+ "grad_norm": 0.10005959123373032,
+ "learning_rate": 0.0006,
+ "loss": 6.422270774841309,
+ "step": 5249
+ },
+ {
+ "epoch": 72.92227074235808,
+ "grad_norm": 0.08989301323890686,
+ "learning_rate": 0.0006,
+ "loss": 6.289371490478516,
+ "step": 5250
+ },
+ {
+ "epoch": 72.93624454148471,
+ "grad_norm": 0.06371799111366272,
+ "learning_rate": 0.0006,
+ "loss": 6.199618339538574,
+ "step": 5251
+ },
+ {
+ "epoch": 72.95021834061136,
+ "grad_norm": 0.05932944267988205,
+ "learning_rate": 0.0006,
+ "loss": 6.111152648925781,
+ "step": 5252
+ },
+ {
+ "epoch": 72.96419213973799,
+ "grad_norm": 0.04735502973198891,
+ "learning_rate": 0.0006,
+ "loss": 6.026026248931885,
+ "step": 5253
+ },
+ {
+ "epoch": 72.97816593886463,
+ "grad_norm": 0.04270688816905022,
+ "learning_rate": 0.0006,
+ "loss": 6.01606559753418,
+ "step": 5254
+ },
+ {
+ "epoch": 72.99213973799127,
+ "grad_norm": 0.03774143010377884,
+ "learning_rate": 0.0006,
+ "loss": 5.8935441970825195,
+ "step": 5255
+ },
+ {
+ "epoch": 73.0,
+ "grad_norm": 0.04313631355762482,
+ "learning_rate": 0.0006,
+ "loss": 5.842061996459961,
+ "step": 5256
+ },
+ {
+ "epoch": 73.0,
+ "eval_loss": 5.9985785484313965,
+ "eval_runtime": 56.3891,
+ "eval_samples_per_second": 43.306,
+ "eval_steps_per_second": 1.366,
+ "step": 5256
+ },
+ {
+ "epoch": 73.01397379912663,
+ "grad_norm": 0.045038655400276184,
+ "learning_rate": 0.0006,
+ "loss": 5.8209662437438965,
+ "step": 5257
+ },
+ {
+ "epoch": 73.02794759825328,
+ "grad_norm": 0.03706910088658333,
+ "learning_rate": 0.0006,
+ "loss": 5.806189060211182,
+ "step": 5258
+ },
+ {
+ "epoch": 73.04192139737991,
+ "grad_norm": 0.04051685333251953,
+ "learning_rate": 0.0006,
+ "loss": 5.706869125366211,
+ "step": 5259
+ },
+ {
+ "epoch": 73.05589519650655,
+ "grad_norm": 0.04849153012037277,
+ "learning_rate": 0.0006,
+ "loss": 5.696385383605957,
+ "step": 5260
+ },
+ {
+ "epoch": 73.06986899563319,
+ "grad_norm": 0.06124403700232506,
+ "learning_rate": 0.0006,
+ "loss": 5.618223667144775,
+ "step": 5261
+ },
+ {
+ "epoch": 73.08384279475983,
+ "grad_norm": 0.07375937700271606,
+ "learning_rate": 0.0006,
+ "loss": 5.631045341491699,
+ "step": 5262
+ },
+ {
+ "epoch": 73.09781659388646,
+ "grad_norm": 0.07413844019174576,
+ "learning_rate": 0.0006,
+ "loss": 5.6532182693481445,
+ "step": 5263
+ },
+ {
+ "epoch": 73.1117903930131,
+ "grad_norm": 0.0633770152926445,
+ "learning_rate": 0.0006,
+ "loss": 5.57209587097168,
+ "step": 5264
+ },
+ {
+ "epoch": 73.12576419213974,
+ "grad_norm": 0.05436766892671585,
+ "learning_rate": 0.0006,
+ "loss": 5.53944730758667,
+ "step": 5265
+ },
+ {
+ "epoch": 73.13973799126637,
+ "grad_norm": 0.04639481008052826,
+ "learning_rate": 0.0006,
+ "loss": 5.502948760986328,
+ "step": 5266
+ },
+ {
+ "epoch": 73.15371179039302,
+ "grad_norm": 0.07556558400392532,
+ "learning_rate": 0.0006,
+ "loss": 5.429069995880127,
+ "step": 5267
+ },
+ {
+ "epoch": 73.16768558951965,
+ "grad_norm": 0.1127929836511612,
+ "learning_rate": 0.0006,
+ "loss": 5.545061111450195,
+ "step": 5268
+ },
+ {
+ "epoch": 73.18165938864628,
+ "grad_norm": 0.07620745897293091,
+ "learning_rate": 0.0006,
+ "loss": 5.507102966308594,
+ "step": 5269
+ },
+ {
+ "epoch": 73.19563318777293,
+ "grad_norm": 0.05076538026332855,
+ "learning_rate": 0.0006,
+ "loss": 5.329682350158691,
+ "step": 5270
+ },
+ {
+ "epoch": 73.20960698689956,
+ "grad_norm": 0.05407547950744629,
+ "learning_rate": 0.0006,
+ "loss": 5.3048553466796875,
+ "step": 5271
+ },
+ {
+ "epoch": 73.2235807860262,
+ "grad_norm": 0.059095583856105804,
+ "learning_rate": 0.0006,
+ "loss": 5.292568683624268,
+ "step": 5272
+ },
+ {
+ "epoch": 73.23755458515284,
+ "grad_norm": 0.06193305924534798,
+ "learning_rate": 0.0006,
+ "loss": 5.1870927810668945,
+ "step": 5273
+ },
+ {
+ "epoch": 73.25152838427948,
+ "grad_norm": 0.062261562794446945,
+ "learning_rate": 0.0006,
+ "loss": 5.256819248199463,
+ "step": 5274
+ },
+ {
+ "epoch": 73.26550218340611,
+ "grad_norm": 0.0732516348361969,
+ "learning_rate": 0.0006,
+ "loss": 5.256464004516602,
+ "step": 5275
+ },
+ {
+ "epoch": 73.27947598253274,
+ "grad_norm": 0.12177979946136475,
+ "learning_rate": 0.0006,
+ "loss": 5.284239768981934,
+ "step": 5276
+ },
+ {
+ "epoch": 73.29344978165939,
+ "grad_norm": 0.10459653288125992,
+ "learning_rate": 0.0006,
+ "loss": 5.269713401794434,
+ "step": 5277
+ },
+ {
+ "epoch": 73.30742358078602,
+ "grad_norm": 0.07854694873094559,
+ "learning_rate": 0.0006,
+ "loss": 5.3245391845703125,
+ "step": 5278
+ },
+ {
+ "epoch": 73.32139737991267,
+ "grad_norm": 0.06030094623565674,
+ "learning_rate": 0.0006,
+ "loss": 5.033702850341797,
+ "step": 5279
+ },
+ {
+ "epoch": 73.3353711790393,
+ "grad_norm": 0.05168461427092552,
+ "learning_rate": 0.0006,
+ "loss": 5.140530109405518,
+ "step": 5280
+ },
+ {
+ "epoch": 73.34934497816595,
+ "grad_norm": 0.041064050048589706,
+ "learning_rate": 0.0006,
+ "loss": 5.1085896492004395,
+ "step": 5281
+ },
+ {
+ "epoch": 73.36331877729258,
+ "grad_norm": 0.03788219019770622,
+ "learning_rate": 0.0006,
+ "loss": 5.045009613037109,
+ "step": 5282
+ },
+ {
+ "epoch": 73.37729257641921,
+ "grad_norm": 0.03851071000099182,
+ "learning_rate": 0.0006,
+ "loss": 5.055008888244629,
+ "step": 5283
+ },
+ {
+ "epoch": 73.39126637554585,
+ "grad_norm": 0.03448539599776268,
+ "learning_rate": 0.0006,
+ "loss": 5.10657262802124,
+ "step": 5284
+ },
+ {
+ "epoch": 73.40524017467249,
+ "grad_norm": 0.026618091389536858,
+ "learning_rate": 0.0006,
+ "loss": 5.0342912673950195,
+ "step": 5285
+ },
+ {
+ "epoch": 73.41921397379913,
+ "grad_norm": 0.03060489147901535,
+ "learning_rate": 0.0006,
+ "loss": 5.022700786590576,
+ "step": 5286
+ },
+ {
+ "epoch": 73.43318777292576,
+ "grad_norm": 0.03362000361084938,
+ "learning_rate": 0.0006,
+ "loss": 4.891584396362305,
+ "step": 5287
+ },
+ {
+ "epoch": 73.44716157205241,
+ "grad_norm": 0.03887505829334259,
+ "learning_rate": 0.0006,
+ "loss": 4.772258758544922,
+ "step": 5288
+ },
+ {
+ "epoch": 73.46113537117904,
+ "grad_norm": 0.07157225161790848,
+ "learning_rate": 0.0006,
+ "loss": 4.997142791748047,
+ "step": 5289
+ },
+ {
+ "epoch": 73.47510917030567,
+ "grad_norm": 0.10180624574422836,
+ "learning_rate": 0.0006,
+ "loss": 4.908722877502441,
+ "step": 5290
+ },
+ {
+ "epoch": 73.48908296943232,
+ "grad_norm": 0.06529122591018677,
+ "learning_rate": 0.0006,
+ "loss": 4.98001766204834,
+ "step": 5291
+ },
+ {
+ "epoch": 73.50305676855895,
+ "grad_norm": 0.055171865969896317,
+ "learning_rate": 0.0006,
+ "loss": 4.877438068389893,
+ "step": 5292
+ },
+ {
+ "epoch": 73.5170305676856,
+ "grad_norm": 0.0318770632147789,
+ "learning_rate": 0.0006,
+ "loss": 4.836934566497803,
+ "step": 5293
+ },
+ {
+ "epoch": 73.53100436681223,
+ "grad_norm": 0.042072564363479614,
+ "learning_rate": 0.0006,
+ "loss": 4.919489860534668,
+ "step": 5294
+ },
+ {
+ "epoch": 73.54497816593886,
+ "grad_norm": 0.029197754338383675,
+ "learning_rate": 0.0006,
+ "loss": 4.805579662322998,
+ "step": 5295
+ },
+ {
+ "epoch": 73.5589519650655,
+ "grad_norm": 0.03249497339129448,
+ "learning_rate": 0.0006,
+ "loss": 4.704806327819824,
+ "step": 5296
+ },
+ {
+ "epoch": 73.57292576419214,
+ "grad_norm": 0.025050945580005646,
+ "learning_rate": 0.0006,
+ "loss": 4.936512470245361,
+ "step": 5297
+ },
+ {
+ "epoch": 73.58689956331878,
+ "grad_norm": 0.029073381796479225,
+ "learning_rate": 0.0006,
+ "loss": 4.757592678070068,
+ "step": 5298
+ },
+ {
+ "epoch": 73.60087336244541,
+ "grad_norm": 0.03465264290571213,
+ "learning_rate": 0.0006,
+ "loss": 4.729528427124023,
+ "step": 5299
+ },
+ {
+ "epoch": 73.61484716157206,
+ "grad_norm": 0.054795749485492706,
+ "learning_rate": 0.0006,
+ "loss": 4.77252197265625,
+ "step": 5300
+ },
+ {
+ "epoch": 73.62882096069869,
+ "grad_norm": 0.18129196763038635,
+ "learning_rate": 0.0006,
+ "loss": 4.901700019836426,
+ "step": 5301
+ },
+ {
+ "epoch": 73.64279475982532,
+ "grad_norm": 0.23274236917495728,
+ "learning_rate": 0.0006,
+ "loss": 5.207432270050049,
+ "step": 5302
+ },
+ {
+ "epoch": 73.65676855895197,
+ "grad_norm": 0.18162772059440613,
+ "learning_rate": 0.0006,
+ "loss": 5.326632976531982,
+ "step": 5303
+ },
+ {
+ "epoch": 73.6707423580786,
+ "grad_norm": 0.11858413368463516,
+ "learning_rate": 0.0006,
+ "loss": 5.3730058670043945,
+ "step": 5304
+ },
+ {
+ "epoch": 73.68471615720524,
+ "grad_norm": 0.06031268090009689,
+ "learning_rate": 0.0006,
+ "loss": 5.330936431884766,
+ "step": 5305
+ },
+ {
+ "epoch": 73.69868995633188,
+ "grad_norm": 0.059304628521203995,
+ "learning_rate": 0.0006,
+ "loss": 5.280514717102051,
+ "step": 5306
+ },
+ {
+ "epoch": 73.71266375545852,
+ "grad_norm": 0.04702131822705269,
+ "learning_rate": 0.0006,
+ "loss": 5.288274765014648,
+ "step": 5307
+ },
+ {
+ "epoch": 73.72663755458515,
+ "grad_norm": 0.037792228162288666,
+ "learning_rate": 0.0006,
+ "loss": 5.173289775848389,
+ "step": 5308
+ },
+ {
+ "epoch": 73.74061135371178,
+ "grad_norm": 0.037763576954603195,
+ "learning_rate": 0.0006,
+ "loss": 5.084990978240967,
+ "step": 5309
+ },
+ {
+ "epoch": 73.75458515283843,
+ "grad_norm": 0.03337346389889717,
+ "learning_rate": 0.0006,
+ "loss": 5.121417045593262,
+ "step": 5310
+ },
+ {
+ "epoch": 73.76855895196506,
+ "grad_norm": 0.03184802085161209,
+ "learning_rate": 0.0006,
+ "loss": 5.208918571472168,
+ "step": 5311
+ },
+ {
+ "epoch": 73.78253275109171,
+ "grad_norm": 0.02832048572599888,
+ "learning_rate": 0.0006,
+ "loss": 4.923543453216553,
+ "step": 5312
+ },
+ {
+ "epoch": 73.79650655021834,
+ "grad_norm": 0.028297102078795433,
+ "learning_rate": 0.0006,
+ "loss": 4.944367408752441,
+ "step": 5313
+ },
+ {
+ "epoch": 73.81048034934499,
+ "grad_norm": 0.027701659128069878,
+ "learning_rate": 0.0006,
+ "loss": 4.858168601989746,
+ "step": 5314
+ },
+ {
+ "epoch": 73.82445414847162,
+ "grad_norm": 0.02214639261364937,
+ "learning_rate": 0.0006,
+ "loss": 4.887052536010742,
+ "step": 5315
+ },
+ {
+ "epoch": 73.83842794759825,
+ "grad_norm": 0.01995093561708927,
+ "learning_rate": 0.0006,
+ "loss": 4.885797500610352,
+ "step": 5316
+ },
+ {
+ "epoch": 73.8524017467249,
+ "grad_norm": 0.020473985001444817,
+ "learning_rate": 0.0006,
+ "loss": 4.917325019836426,
+ "step": 5317
+ },
+ {
+ "epoch": 73.86637554585153,
+ "grad_norm": 0.0192868635058403,
+ "learning_rate": 0.0006,
+ "loss": 4.801746845245361,
+ "step": 5318
+ },
+ {
+ "epoch": 73.88034934497817,
+ "grad_norm": 0.017108267173171043,
+ "learning_rate": 0.0006,
+ "loss": 4.893620491027832,
+ "step": 5319
+ },
+ {
+ "epoch": 73.8943231441048,
+ "grad_norm": 0.01718071475625038,
+ "learning_rate": 0.0006,
+ "loss": 4.940695285797119,
+ "step": 5320
+ },
+ {
+ "epoch": 73.90829694323143,
+ "grad_norm": 0.017821557819843292,
+ "learning_rate": 0.0006,
+ "loss": 4.767613410949707,
+ "step": 5321
+ },
+ {
+ "epoch": 73.92227074235808,
+ "grad_norm": 0.015262950211763382,
+ "learning_rate": 0.0006,
+ "loss": 4.910315036773682,
+ "step": 5322
+ },
+ {
+ "epoch": 73.93624454148471,
+ "grad_norm": 0.01769006811082363,
+ "learning_rate": 0.0006,
+ "loss": 4.7743449211120605,
+ "step": 5323
+ },
+ {
+ "epoch": 73.95021834061136,
+ "grad_norm": 0.014460788108408451,
+ "learning_rate": 0.0006,
+ "loss": 4.743494987487793,
+ "step": 5324
+ },
+ {
+ "epoch": 73.96419213973799,
+ "grad_norm": 0.01403993833810091,
+ "learning_rate": 0.0006,
+ "loss": 4.666640281677246,
+ "step": 5325
+ },
+ {
+ "epoch": 73.97816593886463,
+ "grad_norm": 0.013997809961438179,
+ "learning_rate": 0.0006,
+ "loss": 4.698747158050537,
+ "step": 5326
+ },
+ {
+ "epoch": 73.99213973799127,
+ "grad_norm": 0.014244182966649532,
+ "learning_rate": 0.0006,
+ "loss": 4.745250701904297,
+ "step": 5327
+ },
+ {
+ "epoch": 74.0,
+ "grad_norm": 0.015144186094403267,
+ "learning_rate": 0.0006,
+ "loss": 4.667946815490723,
+ "step": 5328
+ },
+ {
+ "epoch": 74.0,
+ "eval_loss": 5.005440711975098,
+ "eval_runtime": 56.7836,
+ "eval_samples_per_second": 43.005,
+ "eval_steps_per_second": 1.356,
+ "step": 5328
+ },
+ {
+ "epoch": 74.01397379912663,
+ "grad_norm": 0.020384902134537697,
+ "learning_rate": 0.0006,
+ "loss": 4.626955509185791,
+ "step": 5329
+ },
+ {
+ "epoch": 74.02794759825328,
+ "grad_norm": 0.03624720871448517,
+ "learning_rate": 0.0006,
+ "loss": 4.713998317718506,
+ "step": 5330
+ },
+ {
+ "epoch": 74.04192139737991,
+ "grad_norm": 0.05701754242181778,
+ "learning_rate": 0.0006,
+ "loss": 4.717874526977539,
+ "step": 5331
+ },
+ {
+ "epoch": 74.05589519650655,
+ "grad_norm": 0.03942621126770973,
+ "learning_rate": 0.0006,
+ "loss": 4.7026190757751465,
+ "step": 5332
+ },
+ {
+ "epoch": 74.06986899563319,
+ "grad_norm": 0.017803272232413292,
+ "learning_rate": 0.0006,
+ "loss": 4.702548027038574,
+ "step": 5333
+ },
+ {
+ "epoch": 74.08384279475983,
+ "grad_norm": 0.017024677246809006,
+ "learning_rate": 0.0006,
+ "loss": 4.69752311706543,
+ "step": 5334
+ },
+ {
+ "epoch": 74.09781659388646,
+ "grad_norm": 0.016467854380607605,
+ "learning_rate": 0.0006,
+ "loss": 4.624876976013184,
+ "step": 5335
+ },
+ {
+ "epoch": 74.1117903930131,
+ "grad_norm": 0.016569215804338455,
+ "learning_rate": 0.0006,
+ "loss": 4.645040512084961,
+ "step": 5336
+ },
+ {
+ "epoch": 74.12576419213974,
+ "grad_norm": 0.012872067280113697,
+ "learning_rate": 0.0006,
+ "loss": 4.64582633972168,
+ "step": 5337
+ },
+ {
+ "epoch": 74.13973799126637,
+ "grad_norm": 0.015689387917518616,
+ "learning_rate": 0.0006,
+ "loss": 4.596040725708008,
+ "step": 5338
+ },
+ {
+ "epoch": 74.15371179039302,
+ "grad_norm": 0.013082082383334637,
+ "learning_rate": 0.0006,
+ "loss": 4.648033142089844,
+ "step": 5339
+ },
+ {
+ "epoch": 74.16768558951965,
+ "grad_norm": 0.013728172518312931,
+ "learning_rate": 0.0006,
+ "loss": 4.657873153686523,
+ "step": 5340
+ },
+ {
+ "epoch": 74.18165938864628,
+ "grad_norm": 0.013120594434440136,
+ "learning_rate": 0.0006,
+ "loss": 4.696074485778809,
+ "step": 5341
+ },
+ {
+ "epoch": 74.19563318777293,
+ "grad_norm": 0.01431245356798172,
+ "learning_rate": 0.0006,
+ "loss": 4.5679121017456055,
+ "step": 5342
+ },
+ {
+ "epoch": 74.20960698689956,
+ "grad_norm": 0.013994511216878891,
+ "learning_rate": 0.0006,
+ "loss": 4.5236406326293945,
+ "step": 5343
+ },
+ {
+ "epoch": 74.2235807860262,
+ "grad_norm": 0.011971707455813885,
+ "learning_rate": 0.0006,
+ "loss": 4.626547813415527,
+ "step": 5344
+ },
+ {
+ "epoch": 74.23755458515284,
+ "grad_norm": 0.012805287726223469,
+ "learning_rate": 0.0006,
+ "loss": 4.556386947631836,
+ "step": 5345
+ },
+ {
+ "epoch": 74.25152838427948,
+ "grad_norm": 0.011694584973156452,
+ "learning_rate": 0.0006,
+ "loss": 4.656987190246582,
+ "step": 5346
+ },
+ {
+ "epoch": 74.26550218340611,
+ "grad_norm": 0.013396821916103363,
+ "learning_rate": 0.0006,
+ "loss": 4.526467323303223,
+ "step": 5347
+ },
+ {
+ "epoch": 74.27947598253274,
+ "grad_norm": 0.014398031868040562,
+ "learning_rate": 0.0006,
+ "loss": 4.595100402832031,
+ "step": 5348
+ },
+ {
+ "epoch": 74.29344978165939,
+ "grad_norm": 0.02165800705552101,
+ "learning_rate": 0.0006,
+ "loss": 4.460549354553223,
+ "step": 5349
+ },
+ {
+ "epoch": 74.30742358078602,
+ "grad_norm": 0.03276512026786804,
+ "learning_rate": 0.0006,
+ "loss": 4.498189449310303,
+ "step": 5350
+ },
+ {
+ "epoch": 74.32139737991267,
+ "grad_norm": 0.04616224020719528,
+ "learning_rate": 0.0006,
+ "loss": 4.5268659591674805,
+ "step": 5351
+ },
+ {
+ "epoch": 74.3353711790393,
+ "grad_norm": 0.03863786533474922,
+ "learning_rate": 0.0006,
+ "loss": 4.554747104644775,
+ "step": 5352
+ },
+ {
+ "epoch": 74.34934497816595,
+ "grad_norm": 0.01749865710735321,
+ "learning_rate": 0.0006,
+ "loss": 4.6656036376953125,
+ "step": 5353
+ },
+ {
+ "epoch": 74.36331877729258,
+ "grad_norm": 0.021814396604895592,
+ "learning_rate": 0.0006,
+ "loss": 4.607456207275391,
+ "step": 5354
+ },
+ {
+ "epoch": 74.37729257641921,
+ "grad_norm": 0.017018496990203857,
+ "learning_rate": 0.0006,
+ "loss": 4.54205322265625,
+ "step": 5355
+ },
+ {
+ "epoch": 74.39126637554585,
+ "grad_norm": 0.018297340720891953,
+ "learning_rate": 0.0006,
+ "loss": 4.608870506286621,
+ "step": 5356
+ },
+ {
+ "epoch": 74.40524017467249,
+ "grad_norm": 0.01473482046276331,
+ "learning_rate": 0.0006,
+ "loss": 4.384891510009766,
+ "step": 5357
+ },
+ {
+ "epoch": 74.41921397379913,
+ "grad_norm": 0.01579415239393711,
+ "learning_rate": 0.0006,
+ "loss": 4.588859558105469,
+ "step": 5358
+ },
+ {
+ "epoch": 74.43318777292576,
+ "grad_norm": 0.013906026259064674,
+ "learning_rate": 0.0006,
+ "loss": 4.525175094604492,
+ "step": 5359
+ },
+ {
+ "epoch": 74.44716157205241,
+ "grad_norm": 0.014163361862301826,
+ "learning_rate": 0.0006,
+ "loss": 4.495802879333496,
+ "step": 5360
+ },
+ {
+ "epoch": 74.46113537117904,
+ "grad_norm": 0.014009335078299046,
+ "learning_rate": 0.0006,
+ "loss": 4.5960869789123535,
+ "step": 5361
+ },
+ {
+ "epoch": 74.47510917030567,
+ "grad_norm": 0.012765258550643921,
+ "learning_rate": 0.0006,
+ "loss": 4.515678405761719,
+ "step": 5362
+ },
+ {
+ "epoch": 74.48908296943232,
+ "grad_norm": 0.014071918092668056,
+ "learning_rate": 0.0006,
+ "loss": 4.394025802612305,
+ "step": 5363
+ },
+ {
+ "epoch": 74.50305676855895,
+ "grad_norm": 0.01244515273720026,
+ "learning_rate": 0.0006,
+ "loss": 4.576985836029053,
+ "step": 5364
+ },
+ {
+ "epoch": 74.5170305676856,
+ "grad_norm": 0.012975298799574375,
+ "learning_rate": 0.0006,
+ "loss": 4.540657997131348,
+ "step": 5365
+ },
+ {
+ "epoch": 74.53100436681223,
+ "grad_norm": 0.012380947358906269,
+ "learning_rate": 0.0006,
+ "loss": 4.55731725692749,
+ "step": 5366
+ },
+ {
+ "epoch": 74.54497816593886,
+ "grad_norm": 0.012577617540955544,
+ "learning_rate": 0.0006,
+ "loss": 4.500924110412598,
+ "step": 5367
+ },
+ {
+ "epoch": 74.5589519650655,
+ "grad_norm": 0.012089254334568977,
+ "learning_rate": 0.0006,
+ "loss": 4.561795234680176,
+ "step": 5368
+ },
+ {
+ "epoch": 74.57292576419214,
+ "grad_norm": 0.012024011462926865,
+ "learning_rate": 0.0006,
+ "loss": 4.430635452270508,
+ "step": 5369
+ },
+ {
+ "epoch": 74.58689956331878,
+ "grad_norm": 0.012361960485577583,
+ "learning_rate": 0.0006,
+ "loss": 4.4664106369018555,
+ "step": 5370
+ },
+ {
+ "epoch": 74.60087336244541,
+ "grad_norm": 0.012149893678724766,
+ "learning_rate": 0.0006,
+ "loss": 4.522271156311035,
+ "step": 5371
+ },
+ {
+ "epoch": 74.61484716157206,
+ "grad_norm": 0.01166840922087431,
+ "learning_rate": 0.0006,
+ "loss": 4.3870463371276855,
+ "step": 5372
+ },
+ {
+ "epoch": 74.62882096069869,
+ "grad_norm": 0.01160407718271017,
+ "learning_rate": 0.0006,
+ "loss": 4.504255771636963,
+ "step": 5373
+ },
+ {
+ "epoch": 74.64279475982532,
+ "grad_norm": 0.011259177699685097,
+ "learning_rate": 0.0006,
+ "loss": 4.507267951965332,
+ "step": 5374
+ },
+ {
+ "epoch": 74.65676855895197,
+ "grad_norm": 0.012285715900361538,
+ "learning_rate": 0.0006,
+ "loss": 4.472966194152832,
+ "step": 5375
+ },
+ {
+ "epoch": 74.6707423580786,
+ "grad_norm": 0.011214440688490868,
+ "learning_rate": 0.0006,
+ "loss": 4.5840253829956055,
+ "step": 5376
+ },
+ {
+ "epoch": 74.68471615720524,
+ "grad_norm": 0.012394879944622517,
+ "learning_rate": 0.0006,
+ "loss": 4.417830944061279,
+ "step": 5377
+ },
+ {
+ "epoch": 74.69868995633188,
+ "grad_norm": 0.012360049411654472,
+ "learning_rate": 0.0006,
+ "loss": 4.535454750061035,
+ "step": 5378
+ },
+ {
+ "epoch": 74.71266375545852,
+ "grad_norm": 0.011961652897298336,
+ "learning_rate": 0.0006,
+ "loss": 4.403409957885742,
+ "step": 5379
+ },
+ {
+ "epoch": 74.72663755458515,
+ "grad_norm": 0.012078274972736835,
+ "learning_rate": 0.0006,
+ "loss": 4.448366641998291,
+ "step": 5380
+ },
+ {
+ "epoch": 74.74061135371178,
+ "grad_norm": 0.013239333406090736,
+ "learning_rate": 0.0006,
+ "loss": 4.543450355529785,
+ "step": 5381
+ },
+ {
+ "epoch": 74.75458515283843,
+ "grad_norm": 0.01476441603153944,
+ "learning_rate": 0.0006,
+ "loss": 4.488677978515625,
+ "step": 5382
+ },
+ {
+ "epoch": 74.76855895196506,
+ "grad_norm": 0.013968009501695633,
+ "learning_rate": 0.0006,
+ "loss": 4.603639602661133,
+ "step": 5383
+ },
+ {
+ "epoch": 74.78253275109171,
+ "grad_norm": 0.012322458438575268,
+ "learning_rate": 0.0006,
+ "loss": 4.637874603271484,
+ "step": 5384
+ },
+ {
+ "epoch": 74.79650655021834,
+ "grad_norm": 0.013352593407034874,
+ "learning_rate": 0.0006,
+ "loss": 4.474833965301514,
+ "step": 5385
+ },
+ {
+ "epoch": 74.81048034934499,
+ "grad_norm": 0.014671728946268559,
+ "learning_rate": 0.0006,
+ "loss": 4.516847610473633,
+ "step": 5386
+ },
+ {
+ "epoch": 74.82445414847162,
+ "grad_norm": 0.017163321375846863,
+ "learning_rate": 0.0006,
+ "loss": 4.462325096130371,
+ "step": 5387
+ },
+ {
+ "epoch": 74.83842794759825,
+ "grad_norm": 0.02012511156499386,
+ "learning_rate": 0.0006,
+ "loss": 4.355982780456543,
+ "step": 5388
+ },
+ {
+ "epoch": 74.8524017467249,
+ "grad_norm": 0.01881697215139866,
+ "learning_rate": 0.0006,
+ "loss": 4.451120376586914,
+ "step": 5389
+ },
+ {
+ "epoch": 74.86637554585153,
+ "grad_norm": 0.017279407009482384,
+ "learning_rate": 0.0006,
+ "loss": 4.461202621459961,
+ "step": 5390
+ },
+ {
+ "epoch": 74.88034934497817,
+ "grad_norm": 0.015470407903194427,
+ "learning_rate": 0.0006,
+ "loss": 4.4968671798706055,
+ "step": 5391
+ },
+ {
+ "epoch": 74.8943231441048,
+ "grad_norm": 0.01411783043295145,
+ "learning_rate": 0.0006,
+ "loss": 4.558697700500488,
+ "step": 5392
+ },
+ {
+ "epoch": 74.90829694323143,
+ "grad_norm": 0.017178812995553017,
+ "learning_rate": 0.0006,
+ "loss": 4.254153251647949,
+ "step": 5393
+ },
+ {
+ "epoch": 74.92227074235808,
+ "grad_norm": 0.019193004816770554,
+ "learning_rate": 0.0006,
+ "loss": 4.394747257232666,
+ "step": 5394
+ },
+ {
+ "epoch": 74.93624454148471,
+ "grad_norm": 0.018942881375551224,
+ "learning_rate": 0.0006,
+ "loss": 4.459539413452148,
+ "step": 5395
+ },
+ {
+ "epoch": 74.95021834061136,
+ "grad_norm": 0.01641223020851612,
+ "learning_rate": 0.0006,
+ "loss": 4.485820770263672,
+ "step": 5396
+ },
+ {
+ "epoch": 74.96419213973799,
+ "grad_norm": 0.014485753141343594,
+ "learning_rate": 0.0006,
+ "loss": 4.5068583488464355,
+ "step": 5397
+ },
+ {
+ "epoch": 74.97816593886463,
+ "grad_norm": 0.018242299556732178,
+ "learning_rate": 0.0006,
+ "loss": 4.508835792541504,
+ "step": 5398
+ },
+ {
+ "epoch": 74.99213973799127,
+ "grad_norm": 0.016491873189806938,
+ "learning_rate": 0.0006,
+ "loss": 4.4760870933532715,
+ "step": 5399
+ },
+ {
+ "epoch": 75.0,
+ "grad_norm": 0.014543353579938412,
+ "learning_rate": 0.0006,
+ "loss": 4.540217399597168,
+ "step": 5400
+ },
+ {
+ "epoch": 75.0,
+ "eval_loss": 4.7617573738098145,
+ "eval_runtime": 57.321,
+ "eval_samples_per_second": 42.602,
+ "eval_steps_per_second": 1.343,
+ "step": 5400
+ },
+ {
+ "epoch": 75.01397379912663,
+ "grad_norm": 0.014417114667594433,
+ "learning_rate": 0.0006,
+ "loss": 4.332956790924072,
+ "step": 5401
+ },
+ {
+ "epoch": 75.02794759825328,
+ "grad_norm": 0.01607639715075493,
+ "learning_rate": 0.0006,
+ "loss": 4.352453708648682,
+ "step": 5402
+ },
+ {
+ "epoch": 75.04192139737991,
+ "grad_norm": 0.015438318252563477,
+ "learning_rate": 0.0006,
+ "loss": 4.387139320373535,
+ "step": 5403
+ },
+ {
+ "epoch": 75.05589519650655,
+ "grad_norm": 0.015539007261395454,
+ "learning_rate": 0.0006,
+ "loss": 4.453171730041504,
+ "step": 5404
+ },
+ {
+ "epoch": 75.06986899563319,
+ "grad_norm": 0.017361903563141823,
+ "learning_rate": 0.0006,
+ "loss": 4.512143135070801,
+ "step": 5405
+ },
+ {
+ "epoch": 75.08384279475983,
+ "grad_norm": 0.01815599389374256,
+ "learning_rate": 0.0006,
+ "loss": 4.508997440338135,
+ "step": 5406
+ },
+ {
+ "epoch": 75.09781659388646,
+ "grad_norm": 0.01620607078075409,
+ "learning_rate": 0.0006,
+ "loss": 4.322444915771484,
+ "step": 5407
+ },
+ {
+ "epoch": 75.1117903930131,
+ "grad_norm": 0.014773265458643436,
+ "learning_rate": 0.0006,
+ "loss": 4.359251976013184,
+ "step": 5408
+ },
+ {
+ "epoch": 75.12576419213974,
+ "grad_norm": 0.014099023304879665,
+ "learning_rate": 0.0006,
+ "loss": 4.313380241394043,
+ "step": 5409
+ },
+ {
+ "epoch": 75.13973799126637,
+ "grad_norm": 0.014387735165655613,
+ "learning_rate": 0.0006,
+ "loss": 4.435023307800293,
+ "step": 5410
+ },
+ {
+ "epoch": 75.15371179039302,
+ "grad_norm": 0.014886162243783474,
+ "learning_rate": 0.0006,
+ "loss": 4.335939407348633,
+ "step": 5411
+ },
+ {
+ "epoch": 75.16768558951965,
+ "grad_norm": 0.015766102820634842,
+ "learning_rate": 0.0006,
+ "loss": 4.408674240112305,
+ "step": 5412
+ },
+ {
+ "epoch": 75.18165938864628,
+ "grad_norm": 0.016383284702897072,
+ "learning_rate": 0.0006,
+ "loss": 4.385453224182129,
+ "step": 5413
+ },
+ {
+ "epoch": 75.19563318777293,
+ "grad_norm": 0.013535384088754654,
+ "learning_rate": 0.0006,
+ "loss": 4.292785167694092,
+ "step": 5414
+ },
+ {
+ "epoch": 75.20960698689956,
+ "grad_norm": 0.01354036945849657,
+ "learning_rate": 0.0006,
+ "loss": 4.306252479553223,
+ "step": 5415
+ },
+ {
+ "epoch": 75.2235807860262,
+ "grad_norm": 0.015818078070878983,
+ "learning_rate": 0.0006,
+ "loss": 4.425487995147705,
+ "step": 5416
+ },
+ {
+ "epoch": 75.23755458515284,
+ "grad_norm": 0.014363568276166916,
+ "learning_rate": 0.0006,
+ "loss": 4.395087242126465,
+ "step": 5417
+ },
+ {
+ "epoch": 75.25152838427948,
+ "grad_norm": 0.013895470649003983,
+ "learning_rate": 0.0006,
+ "loss": 4.364090442657471,
+ "step": 5418
+ },
+ {
+ "epoch": 75.26550218340611,
+ "grad_norm": 0.014178668148815632,
+ "learning_rate": 0.0006,
+ "loss": 4.44619083404541,
+ "step": 5419
+ },
+ {
+ "epoch": 75.27947598253274,
+ "grad_norm": 0.0152514623478055,
+ "learning_rate": 0.0006,
+ "loss": 4.348445892333984,
+ "step": 5420
+ },
+ {
+ "epoch": 75.29344978165939,
+ "grad_norm": 0.015397852286696434,
+ "learning_rate": 0.0006,
+ "loss": 4.476832389831543,
+ "step": 5421
+ },
+ {
+ "epoch": 75.30742358078602,
+ "grad_norm": 0.018009863793849945,
+ "learning_rate": 0.0006,
+ "loss": 4.431162357330322,
+ "step": 5422
+ },
+ {
+ "epoch": 75.32139737991267,
+ "grad_norm": 0.017038367688655853,
+ "learning_rate": 0.0006,
+ "loss": 4.474787712097168,
+ "step": 5423
+ },
+ {
+ "epoch": 75.3353711790393,
+ "grad_norm": 0.013763288035988808,
+ "learning_rate": 0.0006,
+ "loss": 4.470419883728027,
+ "step": 5424
+ },
+ {
+ "epoch": 75.34934497816595,
+ "grad_norm": 0.013972110114991665,
+ "learning_rate": 0.0006,
+ "loss": 4.373147964477539,
+ "step": 5425
+ },
+ {
+ "epoch": 75.36331877729258,
+ "grad_norm": 0.017197875306010246,
+ "learning_rate": 0.0006,
+ "loss": 4.494963645935059,
+ "step": 5426
+ },
+ {
+ "epoch": 75.37729257641921,
+ "grad_norm": 0.01687047816812992,
+ "learning_rate": 0.0006,
+ "loss": 4.435934066772461,
+ "step": 5427
+ },
+ {
+ "epoch": 75.39126637554585,
+ "grad_norm": 0.01578272506594658,
+ "learning_rate": 0.0006,
+ "loss": 4.37918758392334,
+ "step": 5428
+ },
+ {
+ "epoch": 75.40524017467249,
+ "grad_norm": 0.015956247225403786,
+ "learning_rate": 0.0006,
+ "loss": 4.455668926239014,
+ "step": 5429
+ },
+ {
+ "epoch": 75.41921397379913,
+ "grad_norm": 0.013495155610144138,
+ "learning_rate": 0.0006,
+ "loss": 4.4271039962768555,
+ "step": 5430
+ },
+ {
+ "epoch": 75.43318777292576,
+ "grad_norm": 0.012851621024310589,
+ "learning_rate": 0.0006,
+ "loss": 4.213107109069824,
+ "step": 5431
+ },
+ {
+ "epoch": 75.44716157205241,
+ "grad_norm": 0.013351158238947392,
+ "learning_rate": 0.0006,
+ "loss": 4.3070783615112305,
+ "step": 5432
+ },
+ {
+ "epoch": 75.46113537117904,
+ "grad_norm": 0.014765217900276184,
+ "learning_rate": 0.0006,
+ "loss": 4.385021209716797,
+ "step": 5433
+ },
+ {
+ "epoch": 75.47510917030567,
+ "grad_norm": 0.016091614961624146,
+ "learning_rate": 0.0006,
+ "loss": 4.312644004821777,
+ "step": 5434
+ },
+ {
+ "epoch": 75.48908296943232,
+ "grad_norm": 0.01612013578414917,
+ "learning_rate": 0.0006,
+ "loss": 4.4482011795043945,
+ "step": 5435
+ },
+ {
+ "epoch": 75.50305676855895,
+ "grad_norm": 0.013157603330910206,
+ "learning_rate": 0.0006,
+ "loss": 4.427262306213379,
+ "step": 5436
+ },
+ {
+ "epoch": 75.5170305676856,
+ "grad_norm": 0.016039809212088585,
+ "learning_rate": 0.0006,
+ "loss": 4.396635055541992,
+ "step": 5437
+ },
+ {
+ "epoch": 75.53100436681223,
+ "grad_norm": 0.01794283092021942,
+ "learning_rate": 0.0006,
+ "loss": 4.476870059967041,
+ "step": 5438
+ },
+ {
+ "epoch": 75.54497816593886,
+ "grad_norm": 0.017030486837029457,
+ "learning_rate": 0.0006,
+ "loss": 4.361156463623047,
+ "step": 5439
+ },
+ {
+ "epoch": 75.5589519650655,
+ "grad_norm": 0.017915161326527596,
+ "learning_rate": 0.0006,
+ "loss": 4.286147117614746,
+ "step": 5440
+ },
+ {
+ "epoch": 75.57292576419214,
+ "grad_norm": 0.017514433711767197,
+ "learning_rate": 0.0006,
+ "loss": 4.394091606140137,
+ "step": 5441
+ },
+ {
+ "epoch": 75.58689956331878,
+ "grad_norm": 0.015128589235246181,
+ "learning_rate": 0.0006,
+ "loss": 4.383126735687256,
+ "step": 5442
+ },
+ {
+ "epoch": 75.60087336244541,
+ "grad_norm": 0.0202616099268198,
+ "learning_rate": 0.0006,
+ "loss": 4.336188316345215,
+ "step": 5443
+ },
+ {
+ "epoch": 75.61484716157206,
+ "grad_norm": 0.019978471100330353,
+ "learning_rate": 0.0006,
+ "loss": 4.461453437805176,
+ "step": 5444
+ },
+ {
+ "epoch": 75.62882096069869,
+ "grad_norm": 0.01738741435110569,
+ "learning_rate": 0.0006,
+ "loss": 4.371197700500488,
+ "step": 5445
+ },
+ {
+ "epoch": 75.64279475982532,
+ "grad_norm": 0.016157809644937515,
+ "learning_rate": 0.0006,
+ "loss": 4.3312835693359375,
+ "step": 5446
+ },
+ {
+ "epoch": 75.65676855895197,
+ "grad_norm": 0.016394613310694695,
+ "learning_rate": 0.0006,
+ "loss": 4.380190849304199,
+ "step": 5447
+ },
+ {
+ "epoch": 75.6707423580786,
+ "grad_norm": 0.014727243222296238,
+ "learning_rate": 0.0006,
+ "loss": 4.308028221130371,
+ "step": 5448
+ },
+ {
+ "epoch": 75.68471615720524,
+ "grad_norm": 0.01695958711206913,
+ "learning_rate": 0.0006,
+ "loss": 4.399503707885742,
+ "step": 5449
+ },
+ {
+ "epoch": 75.69868995633188,
+ "grad_norm": 0.015720795840024948,
+ "learning_rate": 0.0006,
+ "loss": 4.342249870300293,
+ "step": 5450
+ },
+ {
+ "epoch": 75.71266375545852,
+ "grad_norm": 0.01574772410094738,
+ "learning_rate": 0.0006,
+ "loss": 4.331503868103027,
+ "step": 5451
+ },
+ {
+ "epoch": 75.72663755458515,
+ "grad_norm": 0.014812648296356201,
+ "learning_rate": 0.0006,
+ "loss": 4.332413673400879,
+ "step": 5452
+ },
+ {
+ "epoch": 75.74061135371178,
+ "grad_norm": 0.012850426137447357,
+ "learning_rate": 0.0006,
+ "loss": 4.2884063720703125,
+ "step": 5453
+ },
+ {
+ "epoch": 75.75458515283843,
+ "grad_norm": 0.012876970693469048,
+ "learning_rate": 0.0006,
+ "loss": 4.399855613708496,
+ "step": 5454
+ },
+ {
+ "epoch": 75.76855895196506,
+ "grad_norm": 0.012958863750100136,
+ "learning_rate": 0.0006,
+ "loss": 4.411185264587402,
+ "step": 5455
+ },
+ {
+ "epoch": 75.78253275109171,
+ "grad_norm": 0.012197135016322136,
+ "learning_rate": 0.0006,
+ "loss": 4.272603988647461,
+ "step": 5456
+ },
+ {
+ "epoch": 75.79650655021834,
+ "grad_norm": 0.013483566232025623,
+ "learning_rate": 0.0006,
+ "loss": 4.3564348220825195,
+ "step": 5457
+ },
+ {
+ "epoch": 75.81048034934499,
+ "grad_norm": 0.012365429662168026,
+ "learning_rate": 0.0006,
+ "loss": 4.354825973510742,
+ "step": 5458
+ },
+ {
+ "epoch": 75.82445414847162,
+ "grad_norm": 0.012238141149282455,
+ "learning_rate": 0.0006,
+ "loss": 4.440438747406006,
+ "step": 5459
+ },
+ {
+ "epoch": 75.83842794759825,
+ "grad_norm": 0.013380305841565132,
+ "learning_rate": 0.0006,
+ "loss": 4.2813215255737305,
+ "step": 5460
+ },
+ {
+ "epoch": 75.8524017467249,
+ "grad_norm": 0.01443779468536377,
+ "learning_rate": 0.0006,
+ "loss": 4.289078235626221,
+ "step": 5461
+ },
+ {
+ "epoch": 75.86637554585153,
+ "grad_norm": 0.012646574527025223,
+ "learning_rate": 0.0006,
+ "loss": 4.374177932739258,
+ "step": 5462
+ },
+ {
+ "epoch": 75.88034934497817,
+ "grad_norm": 0.012403388507664204,
+ "learning_rate": 0.0006,
+ "loss": 4.343016147613525,
+ "step": 5463
+ },
+ {
+ "epoch": 75.8943231441048,
+ "grad_norm": 0.01319252047687769,
+ "learning_rate": 0.0006,
+ "loss": 4.309957027435303,
+ "step": 5464
+ },
+ {
+ "epoch": 75.90829694323143,
+ "grad_norm": 0.012597428634762764,
+ "learning_rate": 0.0006,
+ "loss": 4.403654098510742,
+ "step": 5465
+ },
+ {
+ "epoch": 75.92227074235808,
+ "grad_norm": 0.012724226340651512,
+ "learning_rate": 0.0006,
+ "loss": 4.326869010925293,
+ "step": 5466
+ },
+ {
+ "epoch": 75.93624454148471,
+ "grad_norm": 0.012698772363364697,
+ "learning_rate": 0.0006,
+ "loss": 4.268430233001709,
+ "step": 5467
+ },
+ {
+ "epoch": 75.95021834061136,
+ "grad_norm": 0.01376490481197834,
+ "learning_rate": 0.0006,
+ "loss": 4.42446231842041,
+ "step": 5468
+ },
+ {
+ "epoch": 75.96419213973799,
+ "grad_norm": 0.013175144791603088,
+ "learning_rate": 0.0006,
+ "loss": 4.374675750732422,
+ "step": 5469
+ },
+ {
+ "epoch": 75.97816593886463,
+ "grad_norm": 0.012731322087347507,
+ "learning_rate": 0.0006,
+ "loss": 4.372966766357422,
+ "step": 5470
+ },
+ {
+ "epoch": 75.99213973799127,
+ "grad_norm": 0.01360505260527134,
+ "learning_rate": 0.0006,
+ "loss": 4.3095903396606445,
+ "step": 5471
+ },
+ {
+ "epoch": 76.0,
+ "grad_norm": 0.014011901803314686,
+ "learning_rate": 0.0006,
+ "loss": 4.26954460144043,
+ "step": 5472
+ },
+ {
+ "epoch": 76.0,
+ "eval_loss": 4.689640998840332,
+ "eval_runtime": 56.9095,
+ "eval_samples_per_second": 42.91,
+ "eval_steps_per_second": 1.353,
+ "step": 5472
+ },
+ {
+ "epoch": 76.01397379912663,
+ "grad_norm": 0.015351055189967155,
+ "learning_rate": 0.0006,
+ "loss": 4.348268985748291,
+ "step": 5473
+ },
+ {
+ "epoch": 76.02794759825328,
+ "grad_norm": 0.017164217308163643,
+ "learning_rate": 0.0006,
+ "loss": 4.244204998016357,
+ "step": 5474
+ },
+ {
+ "epoch": 76.04192139737991,
+ "grad_norm": 0.021624932065606117,
+ "learning_rate": 0.0006,
+ "loss": 4.311168670654297,
+ "step": 5475
+ },
+ {
+ "epoch": 76.05589519650655,
+ "grad_norm": 0.028712468221783638,
+ "learning_rate": 0.0006,
+ "loss": 4.322287082672119,
+ "step": 5476
+ },
+ {
+ "epoch": 76.06986899563319,
+ "grad_norm": 0.027915487065911293,
+ "learning_rate": 0.0006,
+ "loss": 4.381207466125488,
+ "step": 5477
+ },
+ {
+ "epoch": 76.08384279475983,
+ "grad_norm": 0.017955303192138672,
+ "learning_rate": 0.0006,
+ "loss": 4.2635016441345215,
+ "step": 5478
+ },
+ {
+ "epoch": 76.09781659388646,
+ "grad_norm": 0.019113514572381973,
+ "learning_rate": 0.0006,
+ "loss": 4.399471282958984,
+ "step": 5479
+ },
+ {
+ "epoch": 76.1117903930131,
+ "grad_norm": 0.017502542585134506,
+ "learning_rate": 0.0006,
+ "loss": 4.321204662322998,
+ "step": 5480
+ },
+ {
+ "epoch": 76.12576419213974,
+ "grad_norm": 0.01852525770664215,
+ "learning_rate": 0.0006,
+ "loss": 4.217884063720703,
+ "step": 5481
+ },
+ {
+ "epoch": 76.13973799126637,
+ "grad_norm": 0.019836828112602234,
+ "learning_rate": 0.0006,
+ "loss": 4.278532028198242,
+ "step": 5482
+ },
+ {
+ "epoch": 76.15371179039302,
+ "grad_norm": 0.02101975679397583,
+ "learning_rate": 0.0006,
+ "loss": 4.246720314025879,
+ "step": 5483
+ },
+ {
+ "epoch": 76.16768558951965,
+ "grad_norm": 0.021202990785241127,
+ "learning_rate": 0.0006,
+ "loss": 4.517517566680908,
+ "step": 5484
+ },
+ {
+ "epoch": 76.18165938864628,
+ "grad_norm": 0.021428769454360008,
+ "learning_rate": 0.0006,
+ "loss": 4.3373517990112305,
+ "step": 5485
+ },
+ {
+ "epoch": 76.19563318777293,
+ "grad_norm": 0.018263190984725952,
+ "learning_rate": 0.0006,
+ "loss": 4.402856826782227,
+ "step": 5486
+ },
+ {
+ "epoch": 76.20960698689956,
+ "grad_norm": 0.019025536254048347,
+ "learning_rate": 0.0006,
+ "loss": 4.314227104187012,
+ "step": 5487
+ },
+ {
+ "epoch": 76.2235807860262,
+ "grad_norm": 0.020918630063533783,
+ "learning_rate": 0.0006,
+ "loss": 4.454369068145752,
+ "step": 5488
+ },
+ {
+ "epoch": 76.23755458515284,
+ "grad_norm": 0.019307617098093033,
+ "learning_rate": 0.0006,
+ "loss": 4.247509002685547,
+ "step": 5489
+ },
+ {
+ "epoch": 76.25152838427948,
+ "grad_norm": 0.014956331811845303,
+ "learning_rate": 0.0006,
+ "loss": 4.365175247192383,
+ "step": 5490
+ },
+ {
+ "epoch": 76.26550218340611,
+ "grad_norm": 0.017213059589266777,
+ "learning_rate": 0.0006,
+ "loss": 4.384005069732666,
+ "step": 5491
+ },
+ {
+ "epoch": 76.27947598253274,
+ "grad_norm": 0.017575597390532494,
+ "learning_rate": 0.0006,
+ "loss": 4.41934871673584,
+ "step": 5492
+ },
+ {
+ "epoch": 76.29344978165939,
+ "grad_norm": 0.015879256650805473,
+ "learning_rate": 0.0006,
+ "loss": 4.26675271987915,
+ "step": 5493
+ },
+ {
+ "epoch": 76.30742358078602,
+ "grad_norm": 0.01564522460103035,
+ "learning_rate": 0.0006,
+ "loss": 4.277339935302734,
+ "step": 5494
+ },
+ {
+ "epoch": 76.32139737991267,
+ "grad_norm": 0.016414586454629898,
+ "learning_rate": 0.0006,
+ "loss": 4.4138593673706055,
+ "step": 5495
+ },
+ {
+ "epoch": 76.3353711790393,
+ "grad_norm": 0.014732337556779385,
+ "learning_rate": 0.0006,
+ "loss": 4.281006336212158,
+ "step": 5496
+ },
+ {
+ "epoch": 76.34934497816595,
+ "grad_norm": 0.015461455099284649,
+ "learning_rate": 0.0006,
+ "loss": 4.263422966003418,
+ "step": 5497
+ },
+ {
+ "epoch": 76.36331877729258,
+ "grad_norm": 0.014529814012348652,
+ "learning_rate": 0.0006,
+ "loss": 4.36165189743042,
+ "step": 5498
+ },
+ {
+ "epoch": 76.37729257641921,
+ "grad_norm": 0.01469483319669962,
+ "learning_rate": 0.0006,
+ "loss": 4.309642791748047,
+ "step": 5499
+ },
+ {
+ "epoch": 76.39126637554585,
+ "grad_norm": 0.016490574926137924,
+ "learning_rate": 0.0006,
+ "loss": 4.288150787353516,
+ "step": 5500
+ },
+ {
+ "epoch": 76.40524017467249,
+ "grad_norm": 0.014593565836548805,
+ "learning_rate": 0.0006,
+ "loss": 4.271614074707031,
+ "step": 5501
+ },
+ {
+ "epoch": 76.41921397379913,
+ "grad_norm": 0.012796535156667233,
+ "learning_rate": 0.0006,
+ "loss": 4.420042037963867,
+ "step": 5502
+ },
+ {
+ "epoch": 76.43318777292576,
+ "grad_norm": 0.016086025163531303,
+ "learning_rate": 0.0006,
+ "loss": 4.289943695068359,
+ "step": 5503
+ },
+ {
+ "epoch": 76.44716157205241,
+ "grad_norm": 0.015439470298588276,
+ "learning_rate": 0.0006,
+ "loss": 4.3023271560668945,
+ "step": 5504
+ },
+ {
+ "epoch": 76.46113537117904,
+ "grad_norm": 0.015066367574036121,
+ "learning_rate": 0.0006,
+ "loss": 4.197081565856934,
+ "step": 5505
+ },
+ {
+ "epoch": 76.47510917030567,
+ "grad_norm": 0.016653263941407204,
+ "learning_rate": 0.0006,
+ "loss": 4.333785057067871,
+ "step": 5506
+ },
+ {
+ "epoch": 76.48908296943232,
+ "grad_norm": 0.014791973866522312,
+ "learning_rate": 0.0006,
+ "loss": 4.323811054229736,
+ "step": 5507
+ },
+ {
+ "epoch": 76.50305676855895,
+ "grad_norm": 0.014694827608764172,
+ "learning_rate": 0.0006,
+ "loss": 4.195852279663086,
+ "step": 5508
+ },
+ {
+ "epoch": 76.5170305676856,
+ "grad_norm": 0.015120396390557289,
+ "learning_rate": 0.0006,
+ "loss": 4.42379903793335,
+ "step": 5509
+ },
+ {
+ "epoch": 76.53100436681223,
+ "grad_norm": 0.014584025368094444,
+ "learning_rate": 0.0006,
+ "loss": 4.2702531814575195,
+ "step": 5510
+ },
+ {
+ "epoch": 76.54497816593886,
+ "grad_norm": 0.013981613330543041,
+ "learning_rate": 0.0006,
+ "loss": 4.383029937744141,
+ "step": 5511
+ },
+ {
+ "epoch": 76.5589519650655,
+ "grad_norm": 0.012930642813444138,
+ "learning_rate": 0.0006,
+ "loss": 4.296136856079102,
+ "step": 5512
+ },
+ {
+ "epoch": 76.57292576419214,
+ "grad_norm": 0.013060989789664745,
+ "learning_rate": 0.0006,
+ "loss": 4.346380233764648,
+ "step": 5513
+ },
+ {
+ "epoch": 76.58689956331878,
+ "grad_norm": 0.013209798373281956,
+ "learning_rate": 0.0006,
+ "loss": 4.316067218780518,
+ "step": 5514
+ },
+ {
+ "epoch": 76.60087336244541,
+ "grad_norm": 0.012423216365277767,
+ "learning_rate": 0.0006,
+ "loss": 4.370274543762207,
+ "step": 5515
+ },
+ {
+ "epoch": 76.61484716157206,
+ "grad_norm": 0.013205411843955517,
+ "learning_rate": 0.0006,
+ "loss": 4.429969310760498,
+ "step": 5516
+ },
+ {
+ "epoch": 76.62882096069869,
+ "grad_norm": 0.013049827888607979,
+ "learning_rate": 0.0006,
+ "loss": 4.378549098968506,
+ "step": 5517
+ },
+ {
+ "epoch": 76.64279475982532,
+ "grad_norm": 0.012075880542397499,
+ "learning_rate": 0.0006,
+ "loss": 4.348293304443359,
+ "step": 5518
+ },
+ {
+ "epoch": 76.65676855895197,
+ "grad_norm": 0.014932847581803799,
+ "learning_rate": 0.0006,
+ "loss": 4.282538890838623,
+ "step": 5519
+ },
+ {
+ "epoch": 76.6707423580786,
+ "grad_norm": 0.015615998767316341,
+ "learning_rate": 0.0006,
+ "loss": 4.274641036987305,
+ "step": 5520
+ },
+ {
+ "epoch": 76.68471615720524,
+ "grad_norm": 0.01424513477832079,
+ "learning_rate": 0.0006,
+ "loss": 4.243166923522949,
+ "step": 5521
+ },
+ {
+ "epoch": 76.69868995633188,
+ "grad_norm": 0.012775546871125698,
+ "learning_rate": 0.0006,
+ "loss": 4.335611820220947,
+ "step": 5522
+ },
+ {
+ "epoch": 76.71266375545852,
+ "grad_norm": 0.014461303129792213,
+ "learning_rate": 0.0006,
+ "loss": 4.246976375579834,
+ "step": 5523
+ },
+ {
+ "epoch": 76.72663755458515,
+ "grad_norm": 0.0145436255261302,
+ "learning_rate": 0.0006,
+ "loss": 4.275895118713379,
+ "step": 5524
+ },
+ {
+ "epoch": 76.74061135371178,
+ "grad_norm": 0.014055085368454456,
+ "learning_rate": 0.0006,
+ "loss": 4.331108093261719,
+ "step": 5525
+ },
+ {
+ "epoch": 76.75458515283843,
+ "grad_norm": 0.011933894827961922,
+ "learning_rate": 0.0006,
+ "loss": 4.293721675872803,
+ "step": 5526
+ },
+ {
+ "epoch": 76.76855895196506,
+ "grad_norm": 0.01277392078191042,
+ "learning_rate": 0.0006,
+ "loss": 4.312374114990234,
+ "step": 5527
+ },
+ {
+ "epoch": 76.78253275109171,
+ "grad_norm": 0.014227936044335365,
+ "learning_rate": 0.0006,
+ "loss": 4.238729000091553,
+ "step": 5528
+ },
+ {
+ "epoch": 76.79650655021834,
+ "grad_norm": 0.01488024927675724,
+ "learning_rate": 0.0006,
+ "loss": 4.24587345123291,
+ "step": 5529
+ },
+ {
+ "epoch": 76.81048034934499,
+ "grad_norm": 0.015341250225901604,
+ "learning_rate": 0.0006,
+ "loss": 4.257378578186035,
+ "step": 5530
+ },
+ {
+ "epoch": 76.82445414847162,
+ "grad_norm": 0.015179437585175037,
+ "learning_rate": 0.0006,
+ "loss": 4.2812910079956055,
+ "step": 5531
+ },
+ {
+ "epoch": 76.83842794759825,
+ "grad_norm": 0.012814724817872047,
+ "learning_rate": 0.0006,
+ "loss": 4.2309465408325195,
+ "step": 5532
+ },
+ {
+ "epoch": 76.8524017467249,
+ "grad_norm": 0.013810716569423676,
+ "learning_rate": 0.0006,
+ "loss": 4.325292110443115,
+ "step": 5533
+ },
+ {
+ "epoch": 76.86637554585153,
+ "grad_norm": 0.017468232661485672,
+ "learning_rate": 0.0006,
+ "loss": 4.243788719177246,
+ "step": 5534
+ },
+ {
+ "epoch": 76.88034934497817,
+ "grad_norm": 0.01739857904613018,
+ "learning_rate": 0.0006,
+ "loss": 4.345390796661377,
+ "step": 5535
+ },
+ {
+ "epoch": 76.8943231441048,
+ "grad_norm": 0.015329061076045036,
+ "learning_rate": 0.0006,
+ "loss": 4.415142059326172,
+ "step": 5536
+ },
+ {
+ "epoch": 76.90829694323143,
+ "grad_norm": 0.01598426327109337,
+ "learning_rate": 0.0006,
+ "loss": 4.299558639526367,
+ "step": 5537
+ },
+ {
+ "epoch": 76.92227074235808,
+ "grad_norm": 0.01621454581618309,
+ "learning_rate": 0.0006,
+ "loss": 4.284354209899902,
+ "step": 5538
+ },
+ {
+ "epoch": 76.93624454148471,
+ "grad_norm": 0.015956968069076538,
+ "learning_rate": 0.0006,
+ "loss": 4.293837547302246,
+ "step": 5539
+ },
+ {
+ "epoch": 76.95021834061136,
+ "grad_norm": 0.01776493526995182,
+ "learning_rate": 0.0006,
+ "loss": 4.229729175567627,
+ "step": 5540
+ },
+ {
+ "epoch": 76.96419213973799,
+ "grad_norm": 0.015317477285861969,
+ "learning_rate": 0.0006,
+ "loss": 4.3328447341918945,
+ "step": 5541
+ },
+ {
+ "epoch": 76.97816593886463,
+ "grad_norm": 0.015140251256525517,
+ "learning_rate": 0.0006,
+ "loss": 4.310840606689453,
+ "step": 5542
+ },
+ {
+ "epoch": 76.99213973799127,
+ "grad_norm": 0.015326812863349915,
+ "learning_rate": 0.0006,
+ "loss": 4.230832099914551,
+ "step": 5543
+ },
+ {
+ "epoch": 77.0,
+ "grad_norm": 0.01607763208448887,
+ "learning_rate": 0.0006,
+ "loss": 4.2236328125,
+ "step": 5544
+ },
+ {
+ "epoch": 77.0,
+ "eval_loss": 4.662221431732178,
+ "eval_runtime": 56.5194,
+ "eval_samples_per_second": 43.206,
+ "eval_steps_per_second": 1.362,
+ "step": 5544
+ },
+ {
+ "epoch": 77.01397379912663,
+ "grad_norm": 0.014839510433375835,
+ "learning_rate": 0.0006,
+ "loss": 4.254111289978027,
+ "step": 5545
+ },
+ {
+ "epoch": 77.02794759825328,
+ "grad_norm": 0.01519741676747799,
+ "learning_rate": 0.0006,
+ "loss": 4.1607184410095215,
+ "step": 5546
+ },
+ {
+ "epoch": 77.04192139737991,
+ "grad_norm": 0.015472603030502796,
+ "learning_rate": 0.0006,
+ "loss": 4.323320388793945,
+ "step": 5547
+ },
+ {
+ "epoch": 77.05589519650655,
+ "grad_norm": 0.017937948927283287,
+ "learning_rate": 0.0006,
+ "loss": 4.308490753173828,
+ "step": 5548
+ },
+ {
+ "epoch": 77.06986899563319,
+ "grad_norm": 0.020890841260552406,
+ "learning_rate": 0.0006,
+ "loss": 4.259101867675781,
+ "step": 5549
+ },
+ {
+ "epoch": 77.08384279475983,
+ "grad_norm": 0.01892857998609543,
+ "learning_rate": 0.0006,
+ "loss": 4.386970520019531,
+ "step": 5550
+ },
+ {
+ "epoch": 77.09781659388646,
+ "grad_norm": 0.016583630815148354,
+ "learning_rate": 0.0006,
+ "loss": 4.233604907989502,
+ "step": 5551
+ },
+ {
+ "epoch": 77.1117903930131,
+ "grad_norm": 0.016063468530774117,
+ "learning_rate": 0.0006,
+ "loss": 4.339544296264648,
+ "step": 5552
+ },
+ {
+ "epoch": 77.12576419213974,
+ "grad_norm": 0.018326779827475548,
+ "learning_rate": 0.0006,
+ "loss": 4.367204666137695,
+ "step": 5553
+ },
+ {
+ "epoch": 77.13973799126637,
+ "grad_norm": 0.021590933203697205,
+ "learning_rate": 0.0006,
+ "loss": 4.352685928344727,
+ "step": 5554
+ },
+ {
+ "epoch": 77.15371179039302,
+ "grad_norm": 0.019643642008304596,
+ "learning_rate": 0.0006,
+ "loss": 4.258535385131836,
+ "step": 5555
+ },
+ {
+ "epoch": 77.16768558951965,
+ "grad_norm": 0.018476149067282677,
+ "learning_rate": 0.0006,
+ "loss": 4.232568740844727,
+ "step": 5556
+ },
+ {
+ "epoch": 77.18165938864628,
+ "grad_norm": 0.01584405079483986,
+ "learning_rate": 0.0006,
+ "loss": 4.258371353149414,
+ "step": 5557
+ },
+ {
+ "epoch": 77.19563318777293,
+ "grad_norm": 0.01575201191008091,
+ "learning_rate": 0.0006,
+ "loss": 4.183218002319336,
+ "step": 5558
+ },
+ {
+ "epoch": 77.20960698689956,
+ "grad_norm": 0.016871824860572815,
+ "learning_rate": 0.0006,
+ "loss": 4.194563865661621,
+ "step": 5559
+ },
+ {
+ "epoch": 77.2235807860262,
+ "grad_norm": 0.02059333398938179,
+ "learning_rate": 0.0006,
+ "loss": 4.195281982421875,
+ "step": 5560
+ },
+ {
+ "epoch": 77.23755458515284,
+ "grad_norm": 0.022640453651547432,
+ "learning_rate": 0.0006,
+ "loss": 4.279615879058838,
+ "step": 5561
+ },
+ {
+ "epoch": 77.25152838427948,
+ "grad_norm": 0.020335959270596504,
+ "learning_rate": 0.0006,
+ "loss": 4.25040864944458,
+ "step": 5562
+ },
+ {
+ "epoch": 77.26550218340611,
+ "grad_norm": 0.020010793581604958,
+ "learning_rate": 0.0006,
+ "loss": 4.291116714477539,
+ "step": 5563
+ },
+ {
+ "epoch": 77.27947598253274,
+ "grad_norm": 0.02109045907855034,
+ "learning_rate": 0.0006,
+ "loss": 4.364088535308838,
+ "step": 5564
+ },
+ {
+ "epoch": 77.29344978165939,
+ "grad_norm": 0.019519446417689323,
+ "learning_rate": 0.0006,
+ "loss": 4.329166412353516,
+ "step": 5565
+ },
+ {
+ "epoch": 77.30742358078602,
+ "grad_norm": 0.01689804159104824,
+ "learning_rate": 0.0006,
+ "loss": 4.244176864624023,
+ "step": 5566
+ },
+ {
+ "epoch": 77.32139737991267,
+ "grad_norm": 0.01893681474030018,
+ "learning_rate": 0.0006,
+ "loss": 4.239068984985352,
+ "step": 5567
+ },
+ {
+ "epoch": 77.3353711790393,
+ "grad_norm": 0.015998193994164467,
+ "learning_rate": 0.0006,
+ "loss": 4.173398017883301,
+ "step": 5568
+ },
+ {
+ "epoch": 77.34934497816595,
+ "grad_norm": 0.013892280869185925,
+ "learning_rate": 0.0006,
+ "loss": 4.287654876708984,
+ "step": 5569
+ },
+ {
+ "epoch": 77.36331877729258,
+ "grad_norm": 0.017060643061995506,
+ "learning_rate": 0.0006,
+ "loss": 4.398447036743164,
+ "step": 5570
+ },
+ {
+ "epoch": 77.37729257641921,
+ "grad_norm": 0.015168849378824234,
+ "learning_rate": 0.0006,
+ "loss": 4.15665864944458,
+ "step": 5571
+ },
+ {
+ "epoch": 77.39126637554585,
+ "grad_norm": 0.014886523596942425,
+ "learning_rate": 0.0006,
+ "loss": 4.196673393249512,
+ "step": 5572
+ },
+ {
+ "epoch": 77.40524017467249,
+ "grad_norm": 0.016941796988248825,
+ "learning_rate": 0.0006,
+ "loss": 4.314562797546387,
+ "step": 5573
+ },
+ {
+ "epoch": 77.41921397379913,
+ "grad_norm": 0.015353812836110592,
+ "learning_rate": 0.0006,
+ "loss": 4.409579277038574,
+ "step": 5574
+ },
+ {
+ "epoch": 77.43318777292576,
+ "grad_norm": 0.015192852355539799,
+ "learning_rate": 0.0006,
+ "loss": 4.392045974731445,
+ "step": 5575
+ },
+ {
+ "epoch": 77.44716157205241,
+ "grad_norm": 0.014232569374144077,
+ "learning_rate": 0.0006,
+ "loss": 4.21639347076416,
+ "step": 5576
+ },
+ {
+ "epoch": 77.46113537117904,
+ "grad_norm": 0.014453226700425148,
+ "learning_rate": 0.0006,
+ "loss": 4.2802276611328125,
+ "step": 5577
+ },
+ {
+ "epoch": 77.47510917030567,
+ "grad_norm": 0.012881236150860786,
+ "learning_rate": 0.0006,
+ "loss": 4.375980854034424,
+ "step": 5578
+ },
+ {
+ "epoch": 77.48908296943232,
+ "grad_norm": 0.01583976298570633,
+ "learning_rate": 0.0006,
+ "loss": 4.264520645141602,
+ "step": 5579
+ },
+ {
+ "epoch": 77.50305676855895,
+ "grad_norm": 0.014171584509313107,
+ "learning_rate": 0.0006,
+ "loss": 4.225279808044434,
+ "step": 5580
+ },
+ {
+ "epoch": 77.5170305676856,
+ "grad_norm": 0.014715159311890602,
+ "learning_rate": 0.0006,
+ "loss": 4.369834899902344,
+ "step": 5581
+ },
+ {
+ "epoch": 77.53100436681223,
+ "grad_norm": 0.016554079949855804,
+ "learning_rate": 0.0006,
+ "loss": 4.2976179122924805,
+ "step": 5582
+ },
+ {
+ "epoch": 77.54497816593886,
+ "grad_norm": 0.015095747075974941,
+ "learning_rate": 0.0006,
+ "loss": 4.294407844543457,
+ "step": 5583
+ },
+ {
+ "epoch": 77.5589519650655,
+ "grad_norm": 0.012726361863315105,
+ "learning_rate": 0.0006,
+ "loss": 4.221401214599609,
+ "step": 5584
+ },
+ {
+ "epoch": 77.57292576419214,
+ "grad_norm": 0.014999309554696083,
+ "learning_rate": 0.0006,
+ "loss": 4.243929386138916,
+ "step": 5585
+ },
+ {
+ "epoch": 77.58689956331878,
+ "grad_norm": 0.016526037827134132,
+ "learning_rate": 0.0006,
+ "loss": 4.374139785766602,
+ "step": 5586
+ },
+ {
+ "epoch": 77.60087336244541,
+ "grad_norm": 0.01648339442908764,
+ "learning_rate": 0.0006,
+ "loss": 4.392425060272217,
+ "step": 5587
+ },
+ {
+ "epoch": 77.61484716157206,
+ "grad_norm": 0.015269213356077671,
+ "learning_rate": 0.0006,
+ "loss": 4.33986759185791,
+ "step": 5588
+ },
+ {
+ "epoch": 77.62882096069869,
+ "grad_norm": 0.01690507121384144,
+ "learning_rate": 0.0006,
+ "loss": 4.307476043701172,
+ "step": 5589
+ },
+ {
+ "epoch": 77.64279475982532,
+ "grad_norm": 0.016438685357570648,
+ "learning_rate": 0.0006,
+ "loss": 4.2406206130981445,
+ "step": 5590
+ },
+ {
+ "epoch": 77.65676855895197,
+ "grad_norm": 0.0157600287348032,
+ "learning_rate": 0.0006,
+ "loss": 4.314070224761963,
+ "step": 5591
+ },
+ {
+ "epoch": 77.6707423580786,
+ "grad_norm": 0.015127947553992271,
+ "learning_rate": 0.0006,
+ "loss": 4.22475528717041,
+ "step": 5592
+ },
+ {
+ "epoch": 77.68471615720524,
+ "grad_norm": 0.016632677987217903,
+ "learning_rate": 0.0006,
+ "loss": 4.26954984664917,
+ "step": 5593
+ },
+ {
+ "epoch": 77.69868995633188,
+ "grad_norm": 0.013150460086762905,
+ "learning_rate": 0.0006,
+ "loss": 4.369131088256836,
+ "step": 5594
+ },
+ {
+ "epoch": 77.71266375545852,
+ "grad_norm": 0.014513005502521992,
+ "learning_rate": 0.0006,
+ "loss": 4.3248796463012695,
+ "step": 5595
+ },
+ {
+ "epoch": 77.72663755458515,
+ "grad_norm": 0.014090991578996181,
+ "learning_rate": 0.0006,
+ "loss": 4.268381118774414,
+ "step": 5596
+ },
+ {
+ "epoch": 77.74061135371178,
+ "grad_norm": 0.01633097417652607,
+ "learning_rate": 0.0006,
+ "loss": 4.238439083099365,
+ "step": 5597
+ },
+ {
+ "epoch": 77.75458515283843,
+ "grad_norm": 0.01710933819413185,
+ "learning_rate": 0.0006,
+ "loss": 4.399697780609131,
+ "step": 5598
+ },
+ {
+ "epoch": 77.76855895196506,
+ "grad_norm": 0.019251767545938492,
+ "learning_rate": 0.0006,
+ "loss": 4.290450096130371,
+ "step": 5599
+ },
+ {
+ "epoch": 77.78253275109171,
+ "grad_norm": 0.01849171705543995,
+ "learning_rate": 0.0006,
+ "loss": 4.267734527587891,
+ "step": 5600
+ },
+ {
+ "epoch": 77.79650655021834,
+ "grad_norm": 0.014173115603625774,
+ "learning_rate": 0.0006,
+ "loss": 4.288799285888672,
+ "step": 5601
+ },
+ {
+ "epoch": 77.81048034934499,
+ "grad_norm": 0.014572718180716038,
+ "learning_rate": 0.0006,
+ "loss": 4.3790435791015625,
+ "step": 5602
+ },
+ {
+ "epoch": 77.82445414847162,
+ "grad_norm": 0.016375336796045303,
+ "learning_rate": 0.0006,
+ "loss": 4.231725692749023,
+ "step": 5603
+ },
+ {
+ "epoch": 77.83842794759825,
+ "grad_norm": 0.017326701432466507,
+ "learning_rate": 0.0006,
+ "loss": 4.211495399475098,
+ "step": 5604
+ },
+ {
+ "epoch": 77.8524017467249,
+ "grad_norm": 0.016399390995502472,
+ "learning_rate": 0.0006,
+ "loss": 4.386699199676514,
+ "step": 5605
+ },
+ {
+ "epoch": 77.86637554585153,
+ "grad_norm": 0.018269989639520645,
+ "learning_rate": 0.0006,
+ "loss": 4.254856109619141,
+ "step": 5606
+ },
+ {
+ "epoch": 77.88034934497817,
+ "grad_norm": 0.018443435430526733,
+ "learning_rate": 0.0006,
+ "loss": 4.391961097717285,
+ "step": 5607
+ },
+ {
+ "epoch": 77.8943231441048,
+ "grad_norm": 0.01784636080265045,
+ "learning_rate": 0.0006,
+ "loss": 4.273881912231445,
+ "step": 5608
+ },
+ {
+ "epoch": 77.90829694323143,
+ "grad_norm": 0.01782994158565998,
+ "learning_rate": 0.0006,
+ "loss": 4.26833963394165,
+ "step": 5609
+ },
+ {
+ "epoch": 77.92227074235808,
+ "grad_norm": 0.016260741278529167,
+ "learning_rate": 0.0006,
+ "loss": 4.306807994842529,
+ "step": 5610
+ },
+ {
+ "epoch": 77.93624454148471,
+ "grad_norm": 0.015482903458178043,
+ "learning_rate": 0.0006,
+ "loss": 4.187826633453369,
+ "step": 5611
+ },
+ {
+ "epoch": 77.95021834061136,
+ "grad_norm": 0.015193257480859756,
+ "learning_rate": 0.0006,
+ "loss": 4.286060333251953,
+ "step": 5612
+ },
+ {
+ "epoch": 77.96419213973799,
+ "grad_norm": 0.015110115520656109,
+ "learning_rate": 0.0006,
+ "loss": 4.2828192710876465,
+ "step": 5613
+ },
+ {
+ "epoch": 77.97816593886463,
+ "grad_norm": 0.015973486006259918,
+ "learning_rate": 0.0006,
+ "loss": 4.317939758300781,
+ "step": 5614
+ },
+ {
+ "epoch": 77.99213973799127,
+ "grad_norm": 0.017363594844937325,
+ "learning_rate": 0.0006,
+ "loss": 4.287721633911133,
+ "step": 5615
+ },
+ {
+ "epoch": 78.0,
+ "grad_norm": 0.015277642756700516,
+ "learning_rate": 0.0006,
+ "loss": 4.356618881225586,
+ "step": 5616
+ },
+ {
+ "epoch": 78.0,
+ "eval_loss": 4.642678737640381,
+ "eval_runtime": 56.7335,
+ "eval_samples_per_second": 43.043,
+ "eval_steps_per_second": 1.357,
+ "step": 5616
+ },
+ {
+ "epoch": 78.01397379912663,
+ "grad_norm": 0.016794051975011826,
+ "learning_rate": 0.0006,
+ "loss": 4.279930591583252,
+ "step": 5617
+ },
+ {
+ "epoch": 78.02794759825328,
+ "grad_norm": 0.01632673107087612,
+ "learning_rate": 0.0006,
+ "loss": 4.235943794250488,
+ "step": 5618
+ },
+ {
+ "epoch": 78.04192139737991,
+ "grad_norm": 0.01437876746058464,
+ "learning_rate": 0.0006,
+ "loss": 4.2542243003845215,
+ "step": 5619
+ },
+ {
+ "epoch": 78.05589519650655,
+ "grad_norm": 0.014993119053542614,
+ "learning_rate": 0.0006,
+ "loss": 4.262046813964844,
+ "step": 5620
+ },
+ {
+ "epoch": 78.06986899563319,
+ "grad_norm": 0.015721119940280914,
+ "learning_rate": 0.0006,
+ "loss": 4.173436164855957,
+ "step": 5621
+ },
+ {
+ "epoch": 78.08384279475983,
+ "grad_norm": 0.01334085687994957,
+ "learning_rate": 0.0006,
+ "loss": 4.322175979614258,
+ "step": 5622
+ },
+ {
+ "epoch": 78.09781659388646,
+ "grad_norm": 0.015561388805508614,
+ "learning_rate": 0.0006,
+ "loss": 4.242125988006592,
+ "step": 5623
+ },
+ {
+ "epoch": 78.1117903930131,
+ "grad_norm": 0.014418602921068668,
+ "learning_rate": 0.0006,
+ "loss": 4.266186714172363,
+ "step": 5624
+ },
+ {
+ "epoch": 78.12576419213974,
+ "grad_norm": 0.013739909045398235,
+ "learning_rate": 0.0006,
+ "loss": 4.231633186340332,
+ "step": 5625
+ },
+ {
+ "epoch": 78.13973799126637,
+ "grad_norm": 0.01402067206799984,
+ "learning_rate": 0.0006,
+ "loss": 4.188406944274902,
+ "step": 5626
+ },
+ {
+ "epoch": 78.15371179039302,
+ "grad_norm": 0.013639481738209724,
+ "learning_rate": 0.0006,
+ "loss": 4.1586594581604,
+ "step": 5627
+ },
+ {
+ "epoch": 78.16768558951965,
+ "grad_norm": 0.012858159840106964,
+ "learning_rate": 0.0006,
+ "loss": 4.360448360443115,
+ "step": 5628
+ },
+ {
+ "epoch": 78.18165938864628,
+ "grad_norm": 0.013089739717543125,
+ "learning_rate": 0.0006,
+ "loss": 4.344371795654297,
+ "step": 5629
+ },
+ {
+ "epoch": 78.19563318777293,
+ "grad_norm": 0.01382314320653677,
+ "learning_rate": 0.0006,
+ "loss": 4.20378303527832,
+ "step": 5630
+ },
+ {
+ "epoch": 78.20960698689956,
+ "grad_norm": 0.014206619001924992,
+ "learning_rate": 0.0006,
+ "loss": 4.239891529083252,
+ "step": 5631
+ },
+ {
+ "epoch": 78.2235807860262,
+ "grad_norm": 0.01472168043255806,
+ "learning_rate": 0.0006,
+ "loss": 4.282382488250732,
+ "step": 5632
+ },
+ {
+ "epoch": 78.23755458515284,
+ "grad_norm": 0.014237025752663612,
+ "learning_rate": 0.0006,
+ "loss": 4.241545677185059,
+ "step": 5633
+ },
+ {
+ "epoch": 78.25152838427948,
+ "grad_norm": 0.013231472112238407,
+ "learning_rate": 0.0006,
+ "loss": 4.184896469116211,
+ "step": 5634
+ },
+ {
+ "epoch": 78.26550218340611,
+ "grad_norm": 0.013538471423089504,
+ "learning_rate": 0.0006,
+ "loss": 4.325181007385254,
+ "step": 5635
+ },
+ {
+ "epoch": 78.27947598253274,
+ "grad_norm": 0.013541961088776588,
+ "learning_rate": 0.0006,
+ "loss": 4.368722915649414,
+ "step": 5636
+ },
+ {
+ "epoch": 78.29344978165939,
+ "grad_norm": 0.013446723110973835,
+ "learning_rate": 0.0006,
+ "loss": 4.301422595977783,
+ "step": 5637
+ },
+ {
+ "epoch": 78.30742358078602,
+ "grad_norm": 0.014753809198737144,
+ "learning_rate": 0.0006,
+ "loss": 4.259127616882324,
+ "step": 5638
+ },
+ {
+ "epoch": 78.32139737991267,
+ "grad_norm": 0.013256008736789227,
+ "learning_rate": 0.0006,
+ "loss": 4.284032821655273,
+ "step": 5639
+ },
+ {
+ "epoch": 78.3353711790393,
+ "grad_norm": 0.014272629283368587,
+ "learning_rate": 0.0006,
+ "loss": 4.234989166259766,
+ "step": 5640
+ },
+ {
+ "epoch": 78.34934497816595,
+ "grad_norm": 0.014756826683878899,
+ "learning_rate": 0.0006,
+ "loss": 4.143499374389648,
+ "step": 5641
+ },
+ {
+ "epoch": 78.36331877729258,
+ "grad_norm": 0.016549205407500267,
+ "learning_rate": 0.0006,
+ "loss": 4.311034202575684,
+ "step": 5642
+ },
+ {
+ "epoch": 78.37729257641921,
+ "grad_norm": 0.017735283821821213,
+ "learning_rate": 0.0006,
+ "loss": 4.246767520904541,
+ "step": 5643
+ },
+ {
+ "epoch": 78.39126637554585,
+ "grad_norm": 0.017370613291859627,
+ "learning_rate": 0.0006,
+ "loss": 4.2460832595825195,
+ "step": 5644
+ },
+ {
+ "epoch": 78.40524017467249,
+ "grad_norm": 0.014642846770584583,
+ "learning_rate": 0.0006,
+ "loss": 4.303811550140381,
+ "step": 5645
+ },
+ {
+ "epoch": 78.41921397379913,
+ "grad_norm": 0.013445726595818996,
+ "learning_rate": 0.0006,
+ "loss": 4.25921630859375,
+ "step": 5646
+ },
+ {
+ "epoch": 78.43318777292576,
+ "grad_norm": 0.014254065230488777,
+ "learning_rate": 0.0006,
+ "loss": 4.129446506500244,
+ "step": 5647
+ },
+ {
+ "epoch": 78.44716157205241,
+ "grad_norm": 0.01883019506931305,
+ "learning_rate": 0.0006,
+ "loss": 4.191716194152832,
+ "step": 5648
+ },
+ {
+ "epoch": 78.46113537117904,
+ "grad_norm": 0.0204437468200922,
+ "learning_rate": 0.0006,
+ "loss": 4.333085536956787,
+ "step": 5649
+ },
+ {
+ "epoch": 78.47510917030567,
+ "grad_norm": 0.01652318798005581,
+ "learning_rate": 0.0006,
+ "loss": 4.135462760925293,
+ "step": 5650
+ },
+ {
+ "epoch": 78.48908296943232,
+ "grad_norm": 0.012753551825881004,
+ "learning_rate": 0.0006,
+ "loss": 4.195051193237305,
+ "step": 5651
+ },
+ {
+ "epoch": 78.50305676855895,
+ "grad_norm": 0.01499118935316801,
+ "learning_rate": 0.0006,
+ "loss": 4.157092094421387,
+ "step": 5652
+ },
+ {
+ "epoch": 78.5170305676856,
+ "grad_norm": 0.015383945778012276,
+ "learning_rate": 0.0006,
+ "loss": 4.253818035125732,
+ "step": 5653
+ },
+ {
+ "epoch": 78.53100436681223,
+ "grad_norm": 0.015721332281827927,
+ "learning_rate": 0.0006,
+ "loss": 4.228775978088379,
+ "step": 5654
+ },
+ {
+ "epoch": 78.54497816593886,
+ "grad_norm": 0.014595463871955872,
+ "learning_rate": 0.0006,
+ "loss": 4.135197639465332,
+ "step": 5655
+ },
+ {
+ "epoch": 78.5589519650655,
+ "grad_norm": 0.015059644356369972,
+ "learning_rate": 0.0006,
+ "loss": 4.255614280700684,
+ "step": 5656
+ },
+ {
+ "epoch": 78.57292576419214,
+ "grad_norm": 0.013833344914019108,
+ "learning_rate": 0.0006,
+ "loss": 4.301250457763672,
+ "step": 5657
+ },
+ {
+ "epoch": 78.58689956331878,
+ "grad_norm": 0.013665376231074333,
+ "learning_rate": 0.0006,
+ "loss": 4.1734232902526855,
+ "step": 5658
+ },
+ {
+ "epoch": 78.60087336244541,
+ "grad_norm": 0.01539086177945137,
+ "learning_rate": 0.0006,
+ "loss": 4.221459865570068,
+ "step": 5659
+ },
+ {
+ "epoch": 78.61484716157206,
+ "grad_norm": 0.013313917443156242,
+ "learning_rate": 0.0006,
+ "loss": 4.255619049072266,
+ "step": 5660
+ },
+ {
+ "epoch": 78.62882096069869,
+ "grad_norm": 0.01447741687297821,
+ "learning_rate": 0.0006,
+ "loss": 4.298219680786133,
+ "step": 5661
+ },
+ {
+ "epoch": 78.64279475982532,
+ "grad_norm": 0.012796915136277676,
+ "learning_rate": 0.0006,
+ "loss": 4.2542405128479,
+ "step": 5662
+ },
+ {
+ "epoch": 78.65676855895197,
+ "grad_norm": 0.014058141969144344,
+ "learning_rate": 0.0006,
+ "loss": 4.284477233886719,
+ "step": 5663
+ },
+ {
+ "epoch": 78.6707423580786,
+ "grad_norm": 0.012705069966614246,
+ "learning_rate": 0.0006,
+ "loss": 4.2826666831970215,
+ "step": 5664
+ },
+ {
+ "epoch": 78.68471615720524,
+ "grad_norm": 0.013245890848338604,
+ "learning_rate": 0.0006,
+ "loss": 4.371612548828125,
+ "step": 5665
+ },
+ {
+ "epoch": 78.69868995633188,
+ "grad_norm": 0.014192677102982998,
+ "learning_rate": 0.0006,
+ "loss": 4.267344951629639,
+ "step": 5666
+ },
+ {
+ "epoch": 78.71266375545852,
+ "grad_norm": 0.015346777625381947,
+ "learning_rate": 0.0006,
+ "loss": 4.222797393798828,
+ "step": 5667
+ },
+ {
+ "epoch": 78.72663755458515,
+ "grad_norm": 0.014068621210753918,
+ "learning_rate": 0.0006,
+ "loss": 4.172964572906494,
+ "step": 5668
+ },
+ {
+ "epoch": 78.74061135371178,
+ "grad_norm": 0.014364945702254772,
+ "learning_rate": 0.0006,
+ "loss": 4.259319305419922,
+ "step": 5669
+ },
+ {
+ "epoch": 78.75458515283843,
+ "grad_norm": 0.017337387427687645,
+ "learning_rate": 0.0006,
+ "loss": 4.29533576965332,
+ "step": 5670
+ },
+ {
+ "epoch": 78.76855895196506,
+ "grad_norm": 0.017416715621948242,
+ "learning_rate": 0.0006,
+ "loss": 4.340364456176758,
+ "step": 5671
+ },
+ {
+ "epoch": 78.78253275109171,
+ "grad_norm": 0.015925846993923187,
+ "learning_rate": 0.0006,
+ "loss": 4.177454948425293,
+ "step": 5672
+ },
+ {
+ "epoch": 78.79650655021834,
+ "grad_norm": 0.013130443170666695,
+ "learning_rate": 0.0006,
+ "loss": 4.277454853057861,
+ "step": 5673
+ },
+ {
+ "epoch": 78.81048034934499,
+ "grad_norm": 0.014628687873482704,
+ "learning_rate": 0.0006,
+ "loss": 4.273757457733154,
+ "step": 5674
+ },
+ {
+ "epoch": 78.82445414847162,
+ "grad_norm": 0.01601659320294857,
+ "learning_rate": 0.0006,
+ "loss": 4.281774520874023,
+ "step": 5675
+ },
+ {
+ "epoch": 78.83842794759825,
+ "grad_norm": 0.015443485230207443,
+ "learning_rate": 0.0006,
+ "loss": 4.199937343597412,
+ "step": 5676
+ },
+ {
+ "epoch": 78.8524017467249,
+ "grad_norm": 0.013934246264398098,
+ "learning_rate": 0.0006,
+ "loss": 4.331369876861572,
+ "step": 5677
+ },
+ {
+ "epoch": 78.86637554585153,
+ "grad_norm": 0.01572083681821823,
+ "learning_rate": 0.0006,
+ "loss": 4.2828474044799805,
+ "step": 5678
+ },
+ {
+ "epoch": 78.88034934497817,
+ "grad_norm": 0.013444120064377785,
+ "learning_rate": 0.0006,
+ "loss": 4.183859825134277,
+ "step": 5679
+ },
+ {
+ "epoch": 78.8943231441048,
+ "grad_norm": 0.01591254211962223,
+ "learning_rate": 0.0006,
+ "loss": 4.295256614685059,
+ "step": 5680
+ },
+ {
+ "epoch": 78.90829694323143,
+ "grad_norm": 0.017556684091687202,
+ "learning_rate": 0.0006,
+ "loss": 4.132272720336914,
+ "step": 5681
+ },
+ {
+ "epoch": 78.92227074235808,
+ "grad_norm": 0.018184322863817215,
+ "learning_rate": 0.0006,
+ "loss": 4.251728057861328,
+ "step": 5682
+ },
+ {
+ "epoch": 78.93624454148471,
+ "grad_norm": 0.016207285225391388,
+ "learning_rate": 0.0006,
+ "loss": 4.205544471740723,
+ "step": 5683
+ },
+ {
+ "epoch": 78.95021834061136,
+ "grad_norm": 0.014478376135230064,
+ "learning_rate": 0.0006,
+ "loss": 4.182462215423584,
+ "step": 5684
+ },
+ {
+ "epoch": 78.96419213973799,
+ "grad_norm": 0.014255235902965069,
+ "learning_rate": 0.0006,
+ "loss": 4.253366947174072,
+ "step": 5685
+ },
+ {
+ "epoch": 78.97816593886463,
+ "grad_norm": 0.01524396613240242,
+ "learning_rate": 0.0006,
+ "loss": 4.221786022186279,
+ "step": 5686
+ },
+ {
+ "epoch": 78.99213973799127,
+ "grad_norm": 0.015057693235576153,
+ "learning_rate": 0.0006,
+ "loss": 4.208454608917236,
+ "step": 5687
+ },
+ {
+ "epoch": 79.0,
+ "grad_norm": 0.015478022396564484,
+ "learning_rate": 0.0006,
+ "loss": 4.336383819580078,
+ "step": 5688
+ },
+ {
+ "epoch": 79.0,
+ "eval_loss": 4.641749382019043,
+ "eval_runtime": 56.8167,
+ "eval_samples_per_second": 42.98,
+ "eval_steps_per_second": 1.355,
+ "step": 5688
+ },
+ {
+ "epoch": 79.01397379912663,
+ "grad_norm": 0.014143316075205803,
+ "learning_rate": 0.0006,
+ "loss": 4.260974884033203,
+ "step": 5689
+ },
+ {
+ "epoch": 79.02794759825328,
+ "grad_norm": 0.013120883144438267,
+ "learning_rate": 0.0006,
+ "loss": 4.139886856079102,
+ "step": 5690
+ },
+ {
+ "epoch": 79.04192139737991,
+ "grad_norm": 0.015019661746919155,
+ "learning_rate": 0.0006,
+ "loss": 4.267004013061523,
+ "step": 5691
+ },
+ {
+ "epoch": 79.05589519650655,
+ "grad_norm": 0.014891284517943859,
+ "learning_rate": 0.0006,
+ "loss": 4.205019950866699,
+ "step": 5692
+ },
+ {
+ "epoch": 79.06986899563319,
+ "grad_norm": 0.015527728945016861,
+ "learning_rate": 0.0006,
+ "loss": 4.1513566970825195,
+ "step": 5693
+ },
+ {
+ "epoch": 79.08384279475983,
+ "grad_norm": 0.017527181655168533,
+ "learning_rate": 0.0006,
+ "loss": 4.148828983306885,
+ "step": 5694
+ },
+ {
+ "epoch": 79.09781659388646,
+ "grad_norm": 0.01632206328213215,
+ "learning_rate": 0.0006,
+ "loss": 4.195789813995361,
+ "step": 5695
+ },
+ {
+ "epoch": 79.1117903930131,
+ "grad_norm": 0.01590503193438053,
+ "learning_rate": 0.0006,
+ "loss": 4.199155330657959,
+ "step": 5696
+ },
+ {
+ "epoch": 79.12576419213974,
+ "grad_norm": 0.016489997506141663,
+ "learning_rate": 0.0006,
+ "loss": 4.155154705047607,
+ "step": 5697
+ },
+ {
+ "epoch": 79.13973799126637,
+ "grad_norm": 0.0161368977278471,
+ "learning_rate": 0.0006,
+ "loss": 4.164384841918945,
+ "step": 5698
+ },
+ {
+ "epoch": 79.15371179039302,
+ "grad_norm": 0.018574809655547142,
+ "learning_rate": 0.0006,
+ "loss": 4.1898193359375,
+ "step": 5699
+ },
+ {
+ "epoch": 79.16768558951965,
+ "grad_norm": 0.018987983465194702,
+ "learning_rate": 0.0006,
+ "loss": 4.219719409942627,
+ "step": 5700
+ },
+ {
+ "epoch": 79.18165938864628,
+ "grad_norm": 0.01668606884777546,
+ "learning_rate": 0.0006,
+ "loss": 4.152734756469727,
+ "step": 5701
+ },
+ {
+ "epoch": 79.19563318777293,
+ "grad_norm": 0.014046299271285534,
+ "learning_rate": 0.0006,
+ "loss": 4.152964115142822,
+ "step": 5702
+ },
+ {
+ "epoch": 79.20960698689956,
+ "grad_norm": 0.01628032699227333,
+ "learning_rate": 0.0006,
+ "loss": 4.299479961395264,
+ "step": 5703
+ },
+ {
+ "epoch": 79.2235807860262,
+ "grad_norm": 0.017998795956373215,
+ "learning_rate": 0.0006,
+ "loss": 4.231203079223633,
+ "step": 5704
+ },
+ {
+ "epoch": 79.23755458515284,
+ "grad_norm": 0.014799647964537144,
+ "learning_rate": 0.0006,
+ "loss": 4.3530192375183105,
+ "step": 5705
+ },
+ {
+ "epoch": 79.25152838427948,
+ "grad_norm": 0.014640678651630878,
+ "learning_rate": 0.0006,
+ "loss": 4.287322998046875,
+ "step": 5706
+ },
+ {
+ "epoch": 79.26550218340611,
+ "grad_norm": 0.0172616895288229,
+ "learning_rate": 0.0006,
+ "loss": 4.148009777069092,
+ "step": 5707
+ },
+ {
+ "epoch": 79.27947598253274,
+ "grad_norm": 0.01811790093779564,
+ "learning_rate": 0.0006,
+ "loss": 4.269433975219727,
+ "step": 5708
+ },
+ {
+ "epoch": 79.29344978165939,
+ "grad_norm": 0.01764000579714775,
+ "learning_rate": 0.0006,
+ "loss": 4.292638778686523,
+ "step": 5709
+ },
+ {
+ "epoch": 79.30742358078602,
+ "grad_norm": 0.018030935898423195,
+ "learning_rate": 0.0006,
+ "loss": 4.267889976501465,
+ "step": 5710
+ },
+ {
+ "epoch": 79.32139737991267,
+ "grad_norm": 0.017860891297459602,
+ "learning_rate": 0.0006,
+ "loss": 4.211110591888428,
+ "step": 5711
+ },
+ {
+ "epoch": 79.3353711790393,
+ "grad_norm": 0.017295073717832565,
+ "learning_rate": 0.0006,
+ "loss": 4.224645614624023,
+ "step": 5712
+ },
+ {
+ "epoch": 79.34934497816595,
+ "grad_norm": 0.015891101211309433,
+ "learning_rate": 0.0006,
+ "loss": 4.224797248840332,
+ "step": 5713
+ },
+ {
+ "epoch": 79.36331877729258,
+ "grad_norm": 0.014680417254567146,
+ "learning_rate": 0.0006,
+ "loss": 4.182242393493652,
+ "step": 5714
+ },
+ {
+ "epoch": 79.37729257641921,
+ "grad_norm": 0.016621140763163567,
+ "learning_rate": 0.0006,
+ "loss": 4.186334609985352,
+ "step": 5715
+ },
+ {
+ "epoch": 79.39126637554585,
+ "grad_norm": 0.016207249835133553,
+ "learning_rate": 0.0006,
+ "loss": 4.277074813842773,
+ "step": 5716
+ },
+ {
+ "epoch": 79.40524017467249,
+ "grad_norm": 0.015375478193163872,
+ "learning_rate": 0.0006,
+ "loss": 4.154317855834961,
+ "step": 5717
+ },
+ {
+ "epoch": 79.41921397379913,
+ "grad_norm": 0.013453228399157524,
+ "learning_rate": 0.0006,
+ "loss": 4.265270233154297,
+ "step": 5718
+ },
+ {
+ "epoch": 79.43318777292576,
+ "grad_norm": 0.014312694780528545,
+ "learning_rate": 0.0006,
+ "loss": 4.232220649719238,
+ "step": 5719
+ },
+ {
+ "epoch": 79.44716157205241,
+ "grad_norm": 0.014759554527699947,
+ "learning_rate": 0.0006,
+ "loss": 4.272275924682617,
+ "step": 5720
+ },
+ {
+ "epoch": 79.46113537117904,
+ "grad_norm": 0.015885505825281143,
+ "learning_rate": 0.0006,
+ "loss": 4.2229719161987305,
+ "step": 5721
+ },
+ {
+ "epoch": 79.47510917030567,
+ "grad_norm": 0.016516581177711487,
+ "learning_rate": 0.0006,
+ "loss": 4.200624465942383,
+ "step": 5722
+ },
+ {
+ "epoch": 79.48908296943232,
+ "grad_norm": 0.018254555761814117,
+ "learning_rate": 0.0006,
+ "loss": 4.19648551940918,
+ "step": 5723
+ },
+ {
+ "epoch": 79.50305676855895,
+ "grad_norm": 0.01467090006917715,
+ "learning_rate": 0.0006,
+ "loss": 4.135190486907959,
+ "step": 5724
+ },
+ {
+ "epoch": 79.5170305676856,
+ "grad_norm": 0.013474797829985619,
+ "learning_rate": 0.0006,
+ "loss": 4.311361312866211,
+ "step": 5725
+ },
+ {
+ "epoch": 79.53100436681223,
+ "grad_norm": 0.01742764748632908,
+ "learning_rate": 0.0006,
+ "loss": 4.1845903396606445,
+ "step": 5726
+ },
+ {
+ "epoch": 79.54497816593886,
+ "grad_norm": 0.018786882981657982,
+ "learning_rate": 0.0006,
+ "loss": 4.163942337036133,
+ "step": 5727
+ },
+ {
+ "epoch": 79.5589519650655,
+ "grad_norm": 0.017831219360232353,
+ "learning_rate": 0.0006,
+ "loss": 4.236685276031494,
+ "step": 5728
+ },
+ {
+ "epoch": 79.57292576419214,
+ "grad_norm": 0.016334986314177513,
+ "learning_rate": 0.0006,
+ "loss": 4.255853652954102,
+ "step": 5729
+ },
+ {
+ "epoch": 79.58689956331878,
+ "grad_norm": 0.0138487434014678,
+ "learning_rate": 0.0006,
+ "loss": 4.227874279022217,
+ "step": 5730
+ },
+ {
+ "epoch": 79.60087336244541,
+ "grad_norm": 0.01531760673969984,
+ "learning_rate": 0.0006,
+ "loss": 4.221432685852051,
+ "step": 5731
+ },
+ {
+ "epoch": 79.61484716157206,
+ "grad_norm": 0.01603361964225769,
+ "learning_rate": 0.0006,
+ "loss": 4.278326511383057,
+ "step": 5732
+ },
+ {
+ "epoch": 79.62882096069869,
+ "grad_norm": 0.013149895705282688,
+ "learning_rate": 0.0006,
+ "loss": 4.2474517822265625,
+ "step": 5733
+ },
+ {
+ "epoch": 79.64279475982532,
+ "grad_norm": 0.013428786769509315,
+ "learning_rate": 0.0006,
+ "loss": 4.1701178550720215,
+ "step": 5734
+ },
+ {
+ "epoch": 79.65676855895197,
+ "grad_norm": 0.014305585995316505,
+ "learning_rate": 0.0006,
+ "loss": 4.250778675079346,
+ "step": 5735
+ },
+ {
+ "epoch": 79.6707423580786,
+ "grad_norm": 0.014544566161930561,
+ "learning_rate": 0.0006,
+ "loss": 4.243236541748047,
+ "step": 5736
+ },
+ {
+ "epoch": 79.68471615720524,
+ "grad_norm": 0.013524414040148258,
+ "learning_rate": 0.0006,
+ "loss": 4.255457878112793,
+ "step": 5737
+ },
+ {
+ "epoch": 79.69868995633188,
+ "grad_norm": 0.014750641770660877,
+ "learning_rate": 0.0006,
+ "loss": 4.202816486358643,
+ "step": 5738
+ },
+ {
+ "epoch": 79.71266375545852,
+ "grad_norm": 0.015090773813426495,
+ "learning_rate": 0.0006,
+ "loss": 4.3053178787231445,
+ "step": 5739
+ },
+ {
+ "epoch": 79.72663755458515,
+ "grad_norm": 0.01606169529259205,
+ "learning_rate": 0.0006,
+ "loss": 4.224017143249512,
+ "step": 5740
+ },
+ {
+ "epoch": 79.74061135371178,
+ "grad_norm": 0.016303520649671555,
+ "learning_rate": 0.0006,
+ "loss": 4.190760612487793,
+ "step": 5741
+ },
+ {
+ "epoch": 79.75458515283843,
+ "grad_norm": 0.01465672068297863,
+ "learning_rate": 0.0006,
+ "loss": 4.24506950378418,
+ "step": 5742
+ },
+ {
+ "epoch": 79.76855895196506,
+ "grad_norm": 0.012616301886737347,
+ "learning_rate": 0.0006,
+ "loss": 4.174039840698242,
+ "step": 5743
+ },
+ {
+ "epoch": 79.78253275109171,
+ "grad_norm": 0.01358284242451191,
+ "learning_rate": 0.0006,
+ "loss": 4.236380577087402,
+ "step": 5744
+ },
+ {
+ "epoch": 79.79650655021834,
+ "grad_norm": 0.01557177770882845,
+ "learning_rate": 0.0006,
+ "loss": 4.102962493896484,
+ "step": 5745
+ },
+ {
+ "epoch": 79.81048034934499,
+ "grad_norm": 0.015584317035973072,
+ "learning_rate": 0.0006,
+ "loss": 4.242603302001953,
+ "step": 5746
+ },
+ {
+ "epoch": 79.82445414847162,
+ "grad_norm": 0.012597334571182728,
+ "learning_rate": 0.0006,
+ "loss": 4.202791690826416,
+ "step": 5747
+ },
+ {
+ "epoch": 79.83842794759825,
+ "grad_norm": 0.013722898438572884,
+ "learning_rate": 0.0006,
+ "loss": 4.204328536987305,
+ "step": 5748
+ },
+ {
+ "epoch": 79.8524017467249,
+ "grad_norm": 0.014669225551187992,
+ "learning_rate": 0.0006,
+ "loss": 4.164331912994385,
+ "step": 5749
+ },
+ {
+ "epoch": 79.86637554585153,
+ "grad_norm": 0.015012740157544613,
+ "learning_rate": 0.0006,
+ "loss": 4.260565280914307,
+ "step": 5750
+ },
+ {
+ "epoch": 79.88034934497817,
+ "grad_norm": 0.016026340425014496,
+ "learning_rate": 0.0006,
+ "loss": 4.299723148345947,
+ "step": 5751
+ },
+ {
+ "epoch": 79.8943231441048,
+ "grad_norm": 0.01626049168407917,
+ "learning_rate": 0.0006,
+ "loss": 4.207252502441406,
+ "step": 5752
+ },
+ {
+ "epoch": 79.90829694323143,
+ "grad_norm": 0.015415354631841183,
+ "learning_rate": 0.0006,
+ "loss": 4.151934623718262,
+ "step": 5753
+ },
+ {
+ "epoch": 79.92227074235808,
+ "grad_norm": 0.014158242382109165,
+ "learning_rate": 0.0006,
+ "loss": 4.3461995124816895,
+ "step": 5754
+ },
+ {
+ "epoch": 79.93624454148471,
+ "grad_norm": 0.01585628278553486,
+ "learning_rate": 0.0006,
+ "loss": 4.334761619567871,
+ "step": 5755
+ },
+ {
+ "epoch": 79.95021834061136,
+ "grad_norm": 0.013710759580135345,
+ "learning_rate": 0.0006,
+ "loss": 4.180401802062988,
+ "step": 5756
+ },
+ {
+ "epoch": 79.96419213973799,
+ "grad_norm": 0.014159204438328743,
+ "learning_rate": 0.0006,
+ "loss": 4.322331428527832,
+ "step": 5757
+ },
+ {
+ "epoch": 79.97816593886463,
+ "grad_norm": 0.015153195708990097,
+ "learning_rate": 0.0006,
+ "loss": 4.287478923797607,
+ "step": 5758
+ },
+ {
+ "epoch": 79.99213973799127,
+ "grad_norm": 0.014100227504968643,
+ "learning_rate": 0.0006,
+ "loss": 4.231807708740234,
+ "step": 5759
+ },
+ {
+ "epoch": 80.0,
+ "grad_norm": 0.015248682349920273,
+ "learning_rate": 0.0006,
+ "loss": 4.261106014251709,
+ "step": 5760
+ },
+ {
+ "epoch": 80.0,
+ "eval_loss": 4.610264301300049,
+ "eval_runtime": 56.514,
+ "eval_samples_per_second": 43.211,
+ "eval_steps_per_second": 1.362,
+ "step": 5760
+ },
+ {
+ "epoch": 80.01397379912663,
+ "grad_norm": 0.014730271883308887,
+ "learning_rate": 0.0006,
+ "loss": 4.108556747436523,
+ "step": 5761
+ },
+ {
+ "epoch": 80.02794759825328,
+ "grad_norm": 0.014575716108083725,
+ "learning_rate": 0.0006,
+ "loss": 4.169658660888672,
+ "step": 5762
+ },
+ {
+ "epoch": 80.04192139737991,
+ "grad_norm": 0.014706958085298538,
+ "learning_rate": 0.0006,
+ "loss": 4.252662658691406,
+ "step": 5763
+ },
+ {
+ "epoch": 80.05589519650655,
+ "grad_norm": 0.01561224739998579,
+ "learning_rate": 0.0006,
+ "loss": 4.193892955780029,
+ "step": 5764
+ },
+ {
+ "epoch": 80.06986899563319,
+ "grad_norm": 0.014702324755489826,
+ "learning_rate": 0.0006,
+ "loss": 4.1889238357543945,
+ "step": 5765
+ },
+ {
+ "epoch": 80.08384279475983,
+ "grad_norm": 0.015565741807222366,
+ "learning_rate": 0.0006,
+ "loss": 4.243549823760986,
+ "step": 5766
+ },
+ {
+ "epoch": 80.09781659388646,
+ "grad_norm": 0.01647181063890457,
+ "learning_rate": 0.0006,
+ "loss": 4.275446891784668,
+ "step": 5767
+ },
+ {
+ "epoch": 80.1117903930131,
+ "grad_norm": 0.018177716061472893,
+ "learning_rate": 0.0006,
+ "loss": 4.12040376663208,
+ "step": 5768
+ },
+ {
+ "epoch": 80.12576419213974,
+ "grad_norm": 0.01745191030204296,
+ "learning_rate": 0.0006,
+ "loss": 4.140371322631836,
+ "step": 5769
+ },
+ {
+ "epoch": 80.13973799126637,
+ "grad_norm": 0.01568993367254734,
+ "learning_rate": 0.0006,
+ "loss": 4.258280277252197,
+ "step": 5770
+ },
+ {
+ "epoch": 80.15371179039302,
+ "grad_norm": 0.01636597141623497,
+ "learning_rate": 0.0006,
+ "loss": 4.179052352905273,
+ "step": 5771
+ },
+ {
+ "epoch": 80.16768558951965,
+ "grad_norm": 0.016839100047945976,
+ "learning_rate": 0.0006,
+ "loss": 4.127636432647705,
+ "step": 5772
+ },
+ {
+ "epoch": 80.18165938864628,
+ "grad_norm": 0.015471016988158226,
+ "learning_rate": 0.0006,
+ "loss": 4.181325912475586,
+ "step": 5773
+ },
+ {
+ "epoch": 80.19563318777293,
+ "grad_norm": 0.015836089849472046,
+ "learning_rate": 0.0006,
+ "loss": 4.142168998718262,
+ "step": 5774
+ },
+ {
+ "epoch": 80.20960698689956,
+ "grad_norm": 0.014782954007387161,
+ "learning_rate": 0.0006,
+ "loss": 4.308799743652344,
+ "step": 5775
+ },
+ {
+ "epoch": 80.2235807860262,
+ "grad_norm": 0.014040893875062466,
+ "learning_rate": 0.0006,
+ "loss": 4.216936111450195,
+ "step": 5776
+ },
+ {
+ "epoch": 80.23755458515284,
+ "grad_norm": 0.01420280709862709,
+ "learning_rate": 0.0006,
+ "loss": 4.235230445861816,
+ "step": 5777
+ },
+ {
+ "epoch": 80.25152838427948,
+ "grad_norm": 0.013341007754206657,
+ "learning_rate": 0.0006,
+ "loss": 4.369109153747559,
+ "step": 5778
+ },
+ {
+ "epoch": 80.26550218340611,
+ "grad_norm": 0.014208327978849411,
+ "learning_rate": 0.0006,
+ "loss": 4.219112396240234,
+ "step": 5779
+ },
+ {
+ "epoch": 80.27947598253274,
+ "grad_norm": 0.013825052417814732,
+ "learning_rate": 0.0006,
+ "loss": 4.1983489990234375,
+ "step": 5780
+ },
+ {
+ "epoch": 80.29344978165939,
+ "grad_norm": 0.012916515581309795,
+ "learning_rate": 0.0006,
+ "loss": 4.204582691192627,
+ "step": 5781
+ },
+ {
+ "epoch": 80.30742358078602,
+ "grad_norm": 0.013071243651211262,
+ "learning_rate": 0.0006,
+ "loss": 4.17643928527832,
+ "step": 5782
+ },
+ {
+ "epoch": 80.32139737991267,
+ "grad_norm": 0.015507174655795097,
+ "learning_rate": 0.0006,
+ "loss": 4.172976016998291,
+ "step": 5783
+ },
+ {
+ "epoch": 80.3353711790393,
+ "grad_norm": 0.014967243187129498,
+ "learning_rate": 0.0006,
+ "loss": 4.141168594360352,
+ "step": 5784
+ },
+ {
+ "epoch": 80.34934497816595,
+ "grad_norm": 0.015629924833774567,
+ "learning_rate": 0.0006,
+ "loss": 4.253526210784912,
+ "step": 5785
+ },
+ {
+ "epoch": 80.36331877729258,
+ "grad_norm": 0.01607954502105713,
+ "learning_rate": 0.0006,
+ "loss": 4.249395370483398,
+ "step": 5786
+ },
+ {
+ "epoch": 80.37729257641921,
+ "grad_norm": 0.01500706560909748,
+ "learning_rate": 0.0006,
+ "loss": 4.279012680053711,
+ "step": 5787
+ },
+ {
+ "epoch": 80.39126637554585,
+ "grad_norm": 0.014343900606036186,
+ "learning_rate": 0.0006,
+ "loss": 4.268509864807129,
+ "step": 5788
+ },
+ {
+ "epoch": 80.40524017467249,
+ "grad_norm": 0.014682197012007236,
+ "learning_rate": 0.0006,
+ "loss": 4.206334114074707,
+ "step": 5789
+ },
+ {
+ "epoch": 80.41921397379913,
+ "grad_norm": 0.014989163726568222,
+ "learning_rate": 0.0006,
+ "loss": 4.143601417541504,
+ "step": 5790
+ },
+ {
+ "epoch": 80.43318777292576,
+ "grad_norm": 0.015001608058810234,
+ "learning_rate": 0.0006,
+ "loss": 4.164111137390137,
+ "step": 5791
+ },
+ {
+ "epoch": 80.44716157205241,
+ "grad_norm": 0.015222722664475441,
+ "learning_rate": 0.0006,
+ "loss": 4.2369914054870605,
+ "step": 5792
+ },
+ {
+ "epoch": 80.46113537117904,
+ "grad_norm": 0.014126183465123177,
+ "learning_rate": 0.0006,
+ "loss": 4.2209930419921875,
+ "step": 5793
+ },
+ {
+ "epoch": 80.47510917030567,
+ "grad_norm": 0.013065959326922894,
+ "learning_rate": 0.0006,
+ "loss": 4.231184959411621,
+ "step": 5794
+ },
+ {
+ "epoch": 80.48908296943232,
+ "grad_norm": 0.013833962380886078,
+ "learning_rate": 0.0006,
+ "loss": 4.226434707641602,
+ "step": 5795
+ },
+ {
+ "epoch": 80.50305676855895,
+ "grad_norm": 0.01387125626206398,
+ "learning_rate": 0.0006,
+ "loss": 4.241328716278076,
+ "step": 5796
+ },
+ {
+ "epoch": 80.5170305676856,
+ "grad_norm": 0.014145063236355782,
+ "learning_rate": 0.0006,
+ "loss": 4.2268805503845215,
+ "step": 5797
+ },
+ {
+ "epoch": 80.53100436681223,
+ "grad_norm": 0.012990240007638931,
+ "learning_rate": 0.0006,
+ "loss": 4.1718339920043945,
+ "step": 5798
+ },
+ {
+ "epoch": 80.54497816593886,
+ "grad_norm": 0.014185411855578423,
+ "learning_rate": 0.0006,
+ "loss": 4.232397556304932,
+ "step": 5799
+ },
+ {
+ "epoch": 80.5589519650655,
+ "grad_norm": 0.014260523952543736,
+ "learning_rate": 0.0006,
+ "loss": 4.282734394073486,
+ "step": 5800
+ },
+ {
+ "epoch": 80.57292576419214,
+ "grad_norm": 0.012542120181024075,
+ "learning_rate": 0.0006,
+ "loss": 4.134551525115967,
+ "step": 5801
+ },
+ {
+ "epoch": 80.58689956331878,
+ "grad_norm": 0.013078657910227776,
+ "learning_rate": 0.0006,
+ "loss": 4.229815483093262,
+ "step": 5802
+ },
+ {
+ "epoch": 80.60087336244541,
+ "grad_norm": 0.012821056880056858,
+ "learning_rate": 0.0006,
+ "loss": 4.208278179168701,
+ "step": 5803
+ },
+ {
+ "epoch": 80.61484716157206,
+ "grad_norm": 0.012964180670678616,
+ "learning_rate": 0.0006,
+ "loss": 4.188477039337158,
+ "step": 5804
+ },
+ {
+ "epoch": 80.62882096069869,
+ "grad_norm": 0.01631746254861355,
+ "learning_rate": 0.0006,
+ "loss": 4.3214497566223145,
+ "step": 5805
+ },
+ {
+ "epoch": 80.64279475982532,
+ "grad_norm": 0.01726640947163105,
+ "learning_rate": 0.0006,
+ "loss": 4.279359817504883,
+ "step": 5806
+ },
+ {
+ "epoch": 80.65676855895197,
+ "grad_norm": 0.018778258934617043,
+ "learning_rate": 0.0006,
+ "loss": 4.294047832489014,
+ "step": 5807
+ },
+ {
+ "epoch": 80.6707423580786,
+ "grad_norm": 0.01796831749379635,
+ "learning_rate": 0.0006,
+ "loss": 4.314162254333496,
+ "step": 5808
+ },
+ {
+ "epoch": 80.68471615720524,
+ "grad_norm": 0.016495538875460625,
+ "learning_rate": 0.0006,
+ "loss": 4.273871421813965,
+ "step": 5809
+ },
+ {
+ "epoch": 80.69868995633188,
+ "grad_norm": 0.013590306043624878,
+ "learning_rate": 0.0006,
+ "loss": 4.14961051940918,
+ "step": 5810
+ },
+ {
+ "epoch": 80.71266375545852,
+ "grad_norm": 0.014670426957309246,
+ "learning_rate": 0.0006,
+ "loss": 4.211853981018066,
+ "step": 5811
+ },
+ {
+ "epoch": 80.72663755458515,
+ "grad_norm": 0.01641341857612133,
+ "learning_rate": 0.0006,
+ "loss": 4.211948871612549,
+ "step": 5812
+ },
+ {
+ "epoch": 80.74061135371178,
+ "grad_norm": 0.016365470364689827,
+ "learning_rate": 0.0006,
+ "loss": 4.200774192810059,
+ "step": 5813
+ },
+ {
+ "epoch": 80.75458515283843,
+ "grad_norm": 0.01613970287144184,
+ "learning_rate": 0.0006,
+ "loss": 4.288322448730469,
+ "step": 5814
+ },
+ {
+ "epoch": 80.76855895196506,
+ "grad_norm": 0.016742777079343796,
+ "learning_rate": 0.0006,
+ "loss": 4.23801326751709,
+ "step": 5815
+ },
+ {
+ "epoch": 80.78253275109171,
+ "grad_norm": 0.015351069159805775,
+ "learning_rate": 0.0006,
+ "loss": 4.263832092285156,
+ "step": 5816
+ },
+ {
+ "epoch": 80.79650655021834,
+ "grad_norm": 0.013435336761176586,
+ "learning_rate": 0.0006,
+ "loss": 4.160857200622559,
+ "step": 5817
+ },
+ {
+ "epoch": 80.81048034934499,
+ "grad_norm": 0.01474886853247881,
+ "learning_rate": 0.0006,
+ "loss": 4.216025352478027,
+ "step": 5818
+ },
+ {
+ "epoch": 80.82445414847162,
+ "grad_norm": 0.01586935855448246,
+ "learning_rate": 0.0006,
+ "loss": 4.256402969360352,
+ "step": 5819
+ },
+ {
+ "epoch": 80.83842794759825,
+ "grad_norm": 0.014596572145819664,
+ "learning_rate": 0.0006,
+ "loss": 4.142584800720215,
+ "step": 5820
+ },
+ {
+ "epoch": 80.8524017467249,
+ "grad_norm": 0.01501593366265297,
+ "learning_rate": 0.0006,
+ "loss": 4.250462055206299,
+ "step": 5821
+ },
+ {
+ "epoch": 80.86637554585153,
+ "grad_norm": 0.014773511327803135,
+ "learning_rate": 0.0006,
+ "loss": 4.253305912017822,
+ "step": 5822
+ },
+ {
+ "epoch": 80.88034934497817,
+ "grad_norm": 0.013246326707303524,
+ "learning_rate": 0.0006,
+ "loss": 4.170881271362305,
+ "step": 5823
+ },
+ {
+ "epoch": 80.8943231441048,
+ "grad_norm": 0.013013756833970547,
+ "learning_rate": 0.0006,
+ "loss": 4.264721870422363,
+ "step": 5824
+ },
+ {
+ "epoch": 80.90829694323143,
+ "grad_norm": 0.013155903667211533,
+ "learning_rate": 0.0006,
+ "loss": 4.274667739868164,
+ "step": 5825
+ },
+ {
+ "epoch": 80.92227074235808,
+ "grad_norm": 0.01448952779173851,
+ "learning_rate": 0.0006,
+ "loss": 4.24777889251709,
+ "step": 5826
+ },
+ {
+ "epoch": 80.93624454148471,
+ "grad_norm": 0.01415231078863144,
+ "learning_rate": 0.0006,
+ "loss": 4.22014856338501,
+ "step": 5827
+ },
+ {
+ "epoch": 80.95021834061136,
+ "grad_norm": 0.014796273782849312,
+ "learning_rate": 0.0006,
+ "loss": 4.22036600112915,
+ "step": 5828
+ },
+ {
+ "epoch": 80.96419213973799,
+ "grad_norm": 0.013005326502025127,
+ "learning_rate": 0.0006,
+ "loss": 4.276123046875,
+ "step": 5829
+ },
+ {
+ "epoch": 80.97816593886463,
+ "grad_norm": 0.012851177714765072,
+ "learning_rate": 0.0006,
+ "loss": 4.248575687408447,
+ "step": 5830
+ },
+ {
+ "epoch": 80.99213973799127,
+ "grad_norm": 0.013740002177655697,
+ "learning_rate": 0.0006,
+ "loss": 4.051140308380127,
+ "step": 5831
+ },
+ {
+ "epoch": 81.0,
+ "grad_norm": 0.015296104364097118,
+ "learning_rate": 0.0006,
+ "loss": 4.251155853271484,
+ "step": 5832
+ },
+ {
+ "epoch": 81.0,
+ "eval_loss": 4.594127655029297,
+ "eval_runtime": 56.0959,
+ "eval_samples_per_second": 43.533,
+ "eval_steps_per_second": 1.373,
+ "step": 5832
+ },
+ {
+ "epoch": 81.01397379912663,
+ "grad_norm": 0.015804558992385864,
+ "learning_rate": 0.0006,
+ "loss": 4.152894020080566,
+ "step": 5833
+ },
+ {
+ "epoch": 81.02794759825328,
+ "grad_norm": 0.017000854015350342,
+ "learning_rate": 0.0006,
+ "loss": 4.266067981719971,
+ "step": 5834
+ },
+ {
+ "epoch": 81.04192139737991,
+ "grad_norm": 0.016351008787751198,
+ "learning_rate": 0.0006,
+ "loss": 4.192703723907471,
+ "step": 5835
+ },
+ {
+ "epoch": 81.05589519650655,
+ "grad_norm": 0.014703895896673203,
+ "learning_rate": 0.0006,
+ "loss": 4.23554801940918,
+ "step": 5836
+ },
+ {
+ "epoch": 81.06986899563319,
+ "grad_norm": 0.01619596965610981,
+ "learning_rate": 0.0006,
+ "loss": 4.233000755310059,
+ "step": 5837
+ },
+ {
+ "epoch": 81.08384279475983,
+ "grad_norm": 0.01815493032336235,
+ "learning_rate": 0.0006,
+ "loss": 4.200490951538086,
+ "step": 5838
+ },
+ {
+ "epoch": 81.09781659388646,
+ "grad_norm": 0.018305106088519096,
+ "learning_rate": 0.0006,
+ "loss": 4.154242515563965,
+ "step": 5839
+ },
+ {
+ "epoch": 81.1117903930131,
+ "grad_norm": 0.0200856514275074,
+ "learning_rate": 0.0006,
+ "loss": 4.25820779800415,
+ "step": 5840
+ },
+ {
+ "epoch": 81.12576419213974,
+ "grad_norm": 0.02101440541446209,
+ "learning_rate": 0.0006,
+ "loss": 4.168209075927734,
+ "step": 5841
+ },
+ {
+ "epoch": 81.13973799126637,
+ "grad_norm": 0.022829681634902954,
+ "learning_rate": 0.0006,
+ "loss": 4.25560998916626,
+ "step": 5842
+ },
+ {
+ "epoch": 81.15371179039302,
+ "grad_norm": 0.0195082426071167,
+ "learning_rate": 0.0006,
+ "loss": 4.2035627365112305,
+ "step": 5843
+ },
+ {
+ "epoch": 81.16768558951965,
+ "grad_norm": 0.01716248318552971,
+ "learning_rate": 0.0006,
+ "loss": 4.11322021484375,
+ "step": 5844
+ },
+ {
+ "epoch": 81.18165938864628,
+ "grad_norm": 0.02034664712846279,
+ "learning_rate": 0.0006,
+ "loss": 4.179389953613281,
+ "step": 5845
+ },
+ {
+ "epoch": 81.19563318777293,
+ "grad_norm": 0.023578710854053497,
+ "learning_rate": 0.0006,
+ "loss": 4.265695571899414,
+ "step": 5846
+ },
+ {
+ "epoch": 81.20960698689956,
+ "grad_norm": 0.019244182854890823,
+ "learning_rate": 0.0006,
+ "loss": 4.292043209075928,
+ "step": 5847
+ },
+ {
+ "epoch": 81.2235807860262,
+ "grad_norm": 0.01577291637659073,
+ "learning_rate": 0.0006,
+ "loss": 4.250776290893555,
+ "step": 5848
+ },
+ {
+ "epoch": 81.23755458515284,
+ "grad_norm": 0.01617090217769146,
+ "learning_rate": 0.0006,
+ "loss": 4.193098068237305,
+ "step": 5849
+ },
+ {
+ "epoch": 81.25152838427948,
+ "grad_norm": 0.016090184450149536,
+ "learning_rate": 0.0006,
+ "loss": 4.112357139587402,
+ "step": 5850
+ },
+ {
+ "epoch": 81.26550218340611,
+ "grad_norm": 0.01441365759819746,
+ "learning_rate": 0.0006,
+ "loss": 4.114147186279297,
+ "step": 5851
+ },
+ {
+ "epoch": 81.27947598253274,
+ "grad_norm": 0.014122162945568562,
+ "learning_rate": 0.0006,
+ "loss": 4.276175022125244,
+ "step": 5852
+ },
+ {
+ "epoch": 81.29344978165939,
+ "grad_norm": 0.014859487302601337,
+ "learning_rate": 0.0006,
+ "loss": 4.094699859619141,
+ "step": 5853
+ },
+ {
+ "epoch": 81.30742358078602,
+ "grad_norm": 0.013624267652630806,
+ "learning_rate": 0.0006,
+ "loss": 4.2648773193359375,
+ "step": 5854
+ },
+ {
+ "epoch": 81.32139737991267,
+ "grad_norm": 0.01417913381010294,
+ "learning_rate": 0.0006,
+ "loss": 4.228567123413086,
+ "step": 5855
+ },
+ {
+ "epoch": 81.3353711790393,
+ "grad_norm": 0.014560805633664131,
+ "learning_rate": 0.0006,
+ "loss": 4.148930549621582,
+ "step": 5856
+ },
+ {
+ "epoch": 81.34934497816595,
+ "grad_norm": 0.013662154786288738,
+ "learning_rate": 0.0006,
+ "loss": 4.192107677459717,
+ "step": 5857
+ },
+ {
+ "epoch": 81.36331877729258,
+ "grad_norm": 0.013434980995953083,
+ "learning_rate": 0.0006,
+ "loss": 4.162052154541016,
+ "step": 5858
+ },
+ {
+ "epoch": 81.37729257641921,
+ "grad_norm": 0.01300742756575346,
+ "learning_rate": 0.0006,
+ "loss": 4.262385368347168,
+ "step": 5859
+ },
+ {
+ "epoch": 81.39126637554585,
+ "grad_norm": 0.012544351629912853,
+ "learning_rate": 0.0006,
+ "loss": 4.215371131896973,
+ "step": 5860
+ },
+ {
+ "epoch": 81.40524017467249,
+ "grad_norm": 0.013663645833730698,
+ "learning_rate": 0.0006,
+ "loss": 4.247312545776367,
+ "step": 5861
+ },
+ {
+ "epoch": 81.41921397379913,
+ "grad_norm": 0.013127296231687069,
+ "learning_rate": 0.0006,
+ "loss": 4.087685585021973,
+ "step": 5862
+ },
+ {
+ "epoch": 81.43318777292576,
+ "grad_norm": 0.014081938192248344,
+ "learning_rate": 0.0006,
+ "loss": 4.272648811340332,
+ "step": 5863
+ },
+ {
+ "epoch": 81.44716157205241,
+ "grad_norm": 0.015470641665160656,
+ "learning_rate": 0.0006,
+ "loss": 4.278578758239746,
+ "step": 5864
+ },
+ {
+ "epoch": 81.46113537117904,
+ "grad_norm": 0.016524532809853554,
+ "learning_rate": 0.0006,
+ "loss": 4.118268013000488,
+ "step": 5865
+ },
+ {
+ "epoch": 81.47510917030567,
+ "grad_norm": 0.016242722049355507,
+ "learning_rate": 0.0006,
+ "loss": 4.182405471801758,
+ "step": 5866
+ },
+ {
+ "epoch": 81.48908296943232,
+ "grad_norm": 0.01557017583400011,
+ "learning_rate": 0.0006,
+ "loss": 4.229862213134766,
+ "step": 5867
+ },
+ {
+ "epoch": 81.50305676855895,
+ "grad_norm": 0.01562870666384697,
+ "learning_rate": 0.0006,
+ "loss": 4.27754020690918,
+ "step": 5868
+ },
+ {
+ "epoch": 81.5170305676856,
+ "grad_norm": 0.018606074154376984,
+ "learning_rate": 0.0006,
+ "loss": 4.148681640625,
+ "step": 5869
+ },
+ {
+ "epoch": 81.53100436681223,
+ "grad_norm": 0.017740730196237564,
+ "learning_rate": 0.0006,
+ "loss": 4.321932792663574,
+ "step": 5870
+ },
+ {
+ "epoch": 81.54497816593886,
+ "grad_norm": 0.014517677947878838,
+ "learning_rate": 0.0006,
+ "loss": 4.22862434387207,
+ "step": 5871
+ },
+ {
+ "epoch": 81.5589519650655,
+ "grad_norm": 0.01389714702963829,
+ "learning_rate": 0.0006,
+ "loss": 4.198000907897949,
+ "step": 5872
+ },
+ {
+ "epoch": 81.57292576419214,
+ "grad_norm": 0.015348682180047035,
+ "learning_rate": 0.0006,
+ "loss": 4.293485164642334,
+ "step": 5873
+ },
+ {
+ "epoch": 81.58689956331878,
+ "grad_norm": 0.015965791419148445,
+ "learning_rate": 0.0006,
+ "loss": 4.23481559753418,
+ "step": 5874
+ },
+ {
+ "epoch": 81.60087336244541,
+ "grad_norm": 0.015662726014852524,
+ "learning_rate": 0.0006,
+ "loss": 4.158148765563965,
+ "step": 5875
+ },
+ {
+ "epoch": 81.61484716157206,
+ "grad_norm": 0.014402080327272415,
+ "learning_rate": 0.0006,
+ "loss": 4.132723808288574,
+ "step": 5876
+ },
+ {
+ "epoch": 81.62882096069869,
+ "grad_norm": 0.016846125945448875,
+ "learning_rate": 0.0006,
+ "loss": 4.1654462814331055,
+ "step": 5877
+ },
+ {
+ "epoch": 81.64279475982532,
+ "grad_norm": 0.0196097232401371,
+ "learning_rate": 0.0006,
+ "loss": 4.228450298309326,
+ "step": 5878
+ },
+ {
+ "epoch": 81.65676855895197,
+ "grad_norm": 0.016005294397473335,
+ "learning_rate": 0.0006,
+ "loss": 4.230222225189209,
+ "step": 5879
+ },
+ {
+ "epoch": 81.6707423580786,
+ "grad_norm": 0.014685059897601604,
+ "learning_rate": 0.0006,
+ "loss": 4.1367387771606445,
+ "step": 5880
+ },
+ {
+ "epoch": 81.68471615720524,
+ "grad_norm": 0.01647144928574562,
+ "learning_rate": 0.0006,
+ "loss": 4.096675395965576,
+ "step": 5881
+ },
+ {
+ "epoch": 81.69868995633188,
+ "grad_norm": 0.015978610143065453,
+ "learning_rate": 0.0006,
+ "loss": 4.232292175292969,
+ "step": 5882
+ },
+ {
+ "epoch": 81.71266375545852,
+ "grad_norm": 0.01510288193821907,
+ "learning_rate": 0.0006,
+ "loss": 4.274591445922852,
+ "step": 5883
+ },
+ {
+ "epoch": 81.72663755458515,
+ "grad_norm": 0.014942476525902748,
+ "learning_rate": 0.0006,
+ "loss": 4.161482810974121,
+ "step": 5884
+ },
+ {
+ "epoch": 81.74061135371178,
+ "grad_norm": 0.014595299027860165,
+ "learning_rate": 0.0006,
+ "loss": 4.204291343688965,
+ "step": 5885
+ },
+ {
+ "epoch": 81.75458515283843,
+ "grad_norm": 0.014952429570257664,
+ "learning_rate": 0.0006,
+ "loss": 4.242499351501465,
+ "step": 5886
+ },
+ {
+ "epoch": 81.76855895196506,
+ "grad_norm": 0.014761554077267647,
+ "learning_rate": 0.0006,
+ "loss": 4.20393180847168,
+ "step": 5887
+ },
+ {
+ "epoch": 81.78253275109171,
+ "grad_norm": 0.014749647118151188,
+ "learning_rate": 0.0006,
+ "loss": 4.227565765380859,
+ "step": 5888
+ },
+ {
+ "epoch": 81.79650655021834,
+ "grad_norm": 0.014384493231773376,
+ "learning_rate": 0.0006,
+ "loss": 4.282228946685791,
+ "step": 5889
+ },
+ {
+ "epoch": 81.81048034934499,
+ "grad_norm": 0.01260330155491829,
+ "learning_rate": 0.0006,
+ "loss": 4.2882866859436035,
+ "step": 5890
+ },
+ {
+ "epoch": 81.82445414847162,
+ "grad_norm": 0.014055311679840088,
+ "learning_rate": 0.0006,
+ "loss": 4.22351598739624,
+ "step": 5891
+ },
+ {
+ "epoch": 81.83842794759825,
+ "grad_norm": 0.013674065470695496,
+ "learning_rate": 0.0006,
+ "loss": 4.240716934204102,
+ "step": 5892
+ },
+ {
+ "epoch": 81.8524017467249,
+ "grad_norm": 0.01225196197628975,
+ "learning_rate": 0.0006,
+ "loss": 4.2958598136901855,
+ "step": 5893
+ },
+ {
+ "epoch": 81.86637554585153,
+ "grad_norm": 0.013538680970668793,
+ "learning_rate": 0.0006,
+ "loss": 4.24342679977417,
+ "step": 5894
+ },
+ {
+ "epoch": 81.88034934497817,
+ "grad_norm": 0.013151394203305244,
+ "learning_rate": 0.0006,
+ "loss": 4.081999778747559,
+ "step": 5895
+ },
+ {
+ "epoch": 81.8943231441048,
+ "grad_norm": 0.014260678552091122,
+ "learning_rate": 0.0006,
+ "loss": 4.156423568725586,
+ "step": 5896
+ },
+ {
+ "epoch": 81.90829694323143,
+ "grad_norm": 0.01329776830971241,
+ "learning_rate": 0.0006,
+ "loss": 4.1713995933532715,
+ "step": 5897
+ },
+ {
+ "epoch": 81.92227074235808,
+ "grad_norm": 0.014623276889324188,
+ "learning_rate": 0.0006,
+ "loss": 4.244891166687012,
+ "step": 5898
+ },
+ {
+ "epoch": 81.93624454148471,
+ "grad_norm": 0.015760324895381927,
+ "learning_rate": 0.0006,
+ "loss": 4.076528549194336,
+ "step": 5899
+ },
+ {
+ "epoch": 81.95021834061136,
+ "grad_norm": 0.014331766404211521,
+ "learning_rate": 0.0006,
+ "loss": 4.283037185668945,
+ "step": 5900
+ },
+ {
+ "epoch": 81.96419213973799,
+ "grad_norm": 0.012383842840790749,
+ "learning_rate": 0.0006,
+ "loss": 4.117308616638184,
+ "step": 5901
+ },
+ {
+ "epoch": 81.97816593886463,
+ "grad_norm": 0.01250660140067339,
+ "learning_rate": 0.0006,
+ "loss": 4.241250991821289,
+ "step": 5902
+ },
+ {
+ "epoch": 81.99213973799127,
+ "grad_norm": 0.012276900932192802,
+ "learning_rate": 0.0006,
+ "loss": 4.141091823577881,
+ "step": 5903
+ },
+ {
+ "epoch": 82.0,
+ "grad_norm": 0.01588737778365612,
+ "learning_rate": 0.0006,
+ "loss": 4.065532684326172,
+ "step": 5904
+ },
+ {
+ "epoch": 82.0,
+ "eval_loss": 4.612304210662842,
+ "eval_runtime": 57.5411,
+ "eval_samples_per_second": 42.439,
+ "eval_steps_per_second": 1.338,
+ "step": 5904
+ },
+ {
+ "epoch": 82.01397379912663,
+ "grad_norm": 0.017433544620871544,
+ "learning_rate": 0.0006,
+ "loss": 4.202497482299805,
+ "step": 5905
+ },
+ {
+ "epoch": 82.02794759825328,
+ "grad_norm": 0.017260415479540825,
+ "learning_rate": 0.0006,
+ "loss": 4.099949359893799,
+ "step": 5906
+ },
+ {
+ "epoch": 82.04192139737991,
+ "grad_norm": 0.01577908545732498,
+ "learning_rate": 0.0006,
+ "loss": 4.17806339263916,
+ "step": 5907
+ },
+ {
+ "epoch": 82.05589519650655,
+ "grad_norm": 0.016093818470835686,
+ "learning_rate": 0.0006,
+ "loss": 4.219491004943848,
+ "step": 5908
+ },
+ {
+ "epoch": 82.06986899563319,
+ "grad_norm": 0.015426302328705788,
+ "learning_rate": 0.0006,
+ "loss": 4.207982063293457,
+ "step": 5909
+ },
+ {
+ "epoch": 82.08384279475983,
+ "grad_norm": 0.01605442352592945,
+ "learning_rate": 0.0006,
+ "loss": 4.265942573547363,
+ "step": 5910
+ },
+ {
+ "epoch": 82.09781659388646,
+ "grad_norm": 0.01787225715816021,
+ "learning_rate": 0.0006,
+ "loss": 4.151535511016846,
+ "step": 5911
+ },
+ {
+ "epoch": 82.1117903930131,
+ "grad_norm": 0.01750164106488228,
+ "learning_rate": 0.0006,
+ "loss": 4.133655071258545,
+ "step": 5912
+ },
+ {
+ "epoch": 82.12576419213974,
+ "grad_norm": 0.01602957211434841,
+ "learning_rate": 0.0006,
+ "loss": 4.176972389221191,
+ "step": 5913
+ },
+ {
+ "epoch": 82.13973799126637,
+ "grad_norm": 0.01553812064230442,
+ "learning_rate": 0.0006,
+ "loss": 4.125692367553711,
+ "step": 5914
+ },
+ {
+ "epoch": 82.15371179039302,
+ "grad_norm": 0.013978508301079273,
+ "learning_rate": 0.0006,
+ "loss": 4.207271575927734,
+ "step": 5915
+ },
+ {
+ "epoch": 82.16768558951965,
+ "grad_norm": 0.013561118394136429,
+ "learning_rate": 0.0006,
+ "loss": 4.238930702209473,
+ "step": 5916
+ },
+ {
+ "epoch": 82.18165938864628,
+ "grad_norm": 0.013285507448017597,
+ "learning_rate": 0.0006,
+ "loss": 4.129464626312256,
+ "step": 5917
+ },
+ {
+ "epoch": 82.19563318777293,
+ "grad_norm": 0.013903839513659477,
+ "learning_rate": 0.0006,
+ "loss": 4.214109420776367,
+ "step": 5918
+ },
+ {
+ "epoch": 82.20960698689956,
+ "grad_norm": 0.014426837675273418,
+ "learning_rate": 0.0006,
+ "loss": 4.143196105957031,
+ "step": 5919
+ },
+ {
+ "epoch": 82.2235807860262,
+ "grad_norm": 0.015347165986895561,
+ "learning_rate": 0.0006,
+ "loss": 4.235439300537109,
+ "step": 5920
+ },
+ {
+ "epoch": 82.23755458515284,
+ "grad_norm": 0.014266039244830608,
+ "learning_rate": 0.0006,
+ "loss": 3.991617202758789,
+ "step": 5921
+ },
+ {
+ "epoch": 82.25152838427948,
+ "grad_norm": 0.013101593591272831,
+ "learning_rate": 0.0006,
+ "loss": 4.266017913818359,
+ "step": 5922
+ },
+ {
+ "epoch": 82.26550218340611,
+ "grad_norm": 0.013741881586611271,
+ "learning_rate": 0.0006,
+ "loss": 4.224879741668701,
+ "step": 5923
+ },
+ {
+ "epoch": 82.27947598253274,
+ "grad_norm": 0.014206945896148682,
+ "learning_rate": 0.0006,
+ "loss": 4.102272987365723,
+ "step": 5924
+ },
+ {
+ "epoch": 82.29344978165939,
+ "grad_norm": 0.013798723928630352,
+ "learning_rate": 0.0006,
+ "loss": 4.30418586730957,
+ "step": 5925
+ },
+ {
+ "epoch": 82.30742358078602,
+ "grad_norm": 0.015470122918486595,
+ "learning_rate": 0.0006,
+ "loss": 4.2513322830200195,
+ "step": 5926
+ },
+ {
+ "epoch": 82.32139737991267,
+ "grad_norm": 0.01655442826449871,
+ "learning_rate": 0.0006,
+ "loss": 4.223002910614014,
+ "step": 5927
+ },
+ {
+ "epoch": 82.3353711790393,
+ "grad_norm": 0.017170004546642303,
+ "learning_rate": 0.0006,
+ "loss": 4.253904819488525,
+ "step": 5928
+ },
+ {
+ "epoch": 82.34934497816595,
+ "grad_norm": 0.017595812678337097,
+ "learning_rate": 0.0006,
+ "loss": 4.184927940368652,
+ "step": 5929
+ },
+ {
+ "epoch": 82.36331877729258,
+ "grad_norm": 0.015073256567120552,
+ "learning_rate": 0.0006,
+ "loss": 4.202239036560059,
+ "step": 5930
+ },
+ {
+ "epoch": 82.37729257641921,
+ "grad_norm": 0.014310941100120544,
+ "learning_rate": 0.0006,
+ "loss": 4.124934673309326,
+ "step": 5931
+ },
+ {
+ "epoch": 82.39126637554585,
+ "grad_norm": 0.013671821914613247,
+ "learning_rate": 0.0006,
+ "loss": 4.138919353485107,
+ "step": 5932
+ },
+ {
+ "epoch": 82.40524017467249,
+ "grad_norm": 0.014103314839303493,
+ "learning_rate": 0.0006,
+ "loss": 4.211289405822754,
+ "step": 5933
+ },
+ {
+ "epoch": 82.41921397379913,
+ "grad_norm": 0.014272668398916721,
+ "learning_rate": 0.0006,
+ "loss": 4.06290340423584,
+ "step": 5934
+ },
+ {
+ "epoch": 82.43318777292576,
+ "grad_norm": 0.01387016475200653,
+ "learning_rate": 0.0006,
+ "loss": 4.101334571838379,
+ "step": 5935
+ },
+ {
+ "epoch": 82.44716157205241,
+ "grad_norm": 0.013318683952093124,
+ "learning_rate": 0.0006,
+ "loss": 4.206409931182861,
+ "step": 5936
+ },
+ {
+ "epoch": 82.46113537117904,
+ "grad_norm": 0.013781268149614334,
+ "learning_rate": 0.0006,
+ "loss": 4.189474105834961,
+ "step": 5937
+ },
+ {
+ "epoch": 82.47510917030567,
+ "grad_norm": 0.013195851817727089,
+ "learning_rate": 0.0006,
+ "loss": 4.276607036590576,
+ "step": 5938
+ },
+ {
+ "epoch": 82.48908296943232,
+ "grad_norm": 0.014358163811266422,
+ "learning_rate": 0.0006,
+ "loss": 4.203262805938721,
+ "step": 5939
+ },
+ {
+ "epoch": 82.50305676855895,
+ "grad_norm": 0.01403636671602726,
+ "learning_rate": 0.0006,
+ "loss": 4.079258441925049,
+ "step": 5940
+ },
+ {
+ "epoch": 82.5170305676856,
+ "grad_norm": 0.013272535055875778,
+ "learning_rate": 0.0006,
+ "loss": 4.19197940826416,
+ "step": 5941
+ },
+ {
+ "epoch": 82.53100436681223,
+ "grad_norm": 0.013787264935672283,
+ "learning_rate": 0.0006,
+ "loss": 4.12995719909668,
+ "step": 5942
+ },
+ {
+ "epoch": 82.54497816593886,
+ "grad_norm": 0.015020057559013367,
+ "learning_rate": 0.0006,
+ "loss": 4.148890495300293,
+ "step": 5943
+ },
+ {
+ "epoch": 82.5589519650655,
+ "grad_norm": 0.014199157245457172,
+ "learning_rate": 0.0006,
+ "loss": 4.201516151428223,
+ "step": 5944
+ },
+ {
+ "epoch": 82.57292576419214,
+ "grad_norm": 0.013827123679220676,
+ "learning_rate": 0.0006,
+ "loss": 4.240894317626953,
+ "step": 5945
+ },
+ {
+ "epoch": 82.58689956331878,
+ "grad_norm": 0.014767560176551342,
+ "learning_rate": 0.0006,
+ "loss": 4.173320770263672,
+ "step": 5946
+ },
+ {
+ "epoch": 82.60087336244541,
+ "grad_norm": 0.01612078957259655,
+ "learning_rate": 0.0006,
+ "loss": 4.266448020935059,
+ "step": 5947
+ },
+ {
+ "epoch": 82.61484716157206,
+ "grad_norm": 0.016706543043255806,
+ "learning_rate": 0.0006,
+ "loss": 4.229398250579834,
+ "step": 5948
+ },
+ {
+ "epoch": 82.62882096069869,
+ "grad_norm": 0.015630796551704407,
+ "learning_rate": 0.0006,
+ "loss": 4.181282997131348,
+ "step": 5949
+ },
+ {
+ "epoch": 82.64279475982532,
+ "grad_norm": 0.013103844597935677,
+ "learning_rate": 0.0006,
+ "loss": 4.309617042541504,
+ "step": 5950
+ },
+ {
+ "epoch": 82.65676855895197,
+ "grad_norm": 0.016136541962623596,
+ "learning_rate": 0.0006,
+ "loss": 4.224159240722656,
+ "step": 5951
+ },
+ {
+ "epoch": 82.6707423580786,
+ "grad_norm": 0.016160814091563225,
+ "learning_rate": 0.0006,
+ "loss": 4.267970085144043,
+ "step": 5952
+ },
+ {
+ "epoch": 82.68471615720524,
+ "grad_norm": 0.01672680303454399,
+ "learning_rate": 0.0006,
+ "loss": 4.230506896972656,
+ "step": 5953
+ },
+ {
+ "epoch": 82.69868995633188,
+ "grad_norm": 0.0159586351364851,
+ "learning_rate": 0.0006,
+ "loss": 4.159542083740234,
+ "step": 5954
+ },
+ {
+ "epoch": 82.71266375545852,
+ "grad_norm": 0.01461219135671854,
+ "learning_rate": 0.0006,
+ "loss": 4.27863073348999,
+ "step": 5955
+ },
+ {
+ "epoch": 82.72663755458515,
+ "grad_norm": 0.015088613145053387,
+ "learning_rate": 0.0006,
+ "loss": 4.272418975830078,
+ "step": 5956
+ },
+ {
+ "epoch": 82.74061135371178,
+ "grad_norm": 0.01594499684870243,
+ "learning_rate": 0.0006,
+ "loss": 4.271326065063477,
+ "step": 5957
+ },
+ {
+ "epoch": 82.75458515283843,
+ "grad_norm": 0.016648447141051292,
+ "learning_rate": 0.0006,
+ "loss": 4.141478538513184,
+ "step": 5958
+ },
+ {
+ "epoch": 82.76855895196506,
+ "grad_norm": 0.01540316641330719,
+ "learning_rate": 0.0006,
+ "loss": 4.111814498901367,
+ "step": 5959
+ },
+ {
+ "epoch": 82.78253275109171,
+ "grad_norm": 0.01634364202618599,
+ "learning_rate": 0.0006,
+ "loss": 4.3113112449646,
+ "step": 5960
+ },
+ {
+ "epoch": 82.79650655021834,
+ "grad_norm": 0.014768977649509907,
+ "learning_rate": 0.0006,
+ "loss": 4.184217929840088,
+ "step": 5961
+ },
+ {
+ "epoch": 82.81048034934499,
+ "grad_norm": 0.013453755527734756,
+ "learning_rate": 0.0006,
+ "loss": 4.247729301452637,
+ "step": 5962
+ },
+ {
+ "epoch": 82.82445414847162,
+ "grad_norm": 0.014731907285749912,
+ "learning_rate": 0.0006,
+ "loss": 4.352839469909668,
+ "step": 5963
+ },
+ {
+ "epoch": 82.83842794759825,
+ "grad_norm": 0.016047345474362373,
+ "learning_rate": 0.0006,
+ "loss": 4.257283687591553,
+ "step": 5964
+ },
+ {
+ "epoch": 82.8524017467249,
+ "grad_norm": 0.017141170799732208,
+ "learning_rate": 0.0006,
+ "loss": 4.22850227355957,
+ "step": 5965
+ },
+ {
+ "epoch": 82.86637554585153,
+ "grad_norm": 0.01370396837592125,
+ "learning_rate": 0.0006,
+ "loss": 4.1169257164001465,
+ "step": 5966
+ },
+ {
+ "epoch": 82.88034934497817,
+ "grad_norm": 0.012060081586241722,
+ "learning_rate": 0.0006,
+ "loss": 4.195242404937744,
+ "step": 5967
+ },
+ {
+ "epoch": 82.8943231441048,
+ "grad_norm": 0.015528378076851368,
+ "learning_rate": 0.0006,
+ "loss": 4.294939994812012,
+ "step": 5968
+ },
+ {
+ "epoch": 82.90829694323143,
+ "grad_norm": 0.018151041120290756,
+ "learning_rate": 0.0006,
+ "loss": 4.137192726135254,
+ "step": 5969
+ },
+ {
+ "epoch": 82.92227074235808,
+ "grad_norm": 0.017895735800266266,
+ "learning_rate": 0.0006,
+ "loss": 4.174036026000977,
+ "step": 5970
+ },
+ {
+ "epoch": 82.93624454148471,
+ "grad_norm": 0.015610386617481709,
+ "learning_rate": 0.0006,
+ "loss": 4.230136871337891,
+ "step": 5971
+ },
+ {
+ "epoch": 82.95021834061136,
+ "grad_norm": 0.01339928898960352,
+ "learning_rate": 0.0006,
+ "loss": 4.199074745178223,
+ "step": 5972
+ },
+ {
+ "epoch": 82.96419213973799,
+ "grad_norm": 0.015454214997589588,
+ "learning_rate": 0.0006,
+ "loss": 4.216109752655029,
+ "step": 5973
+ },
+ {
+ "epoch": 82.97816593886463,
+ "grad_norm": 0.014978132210671902,
+ "learning_rate": 0.0006,
+ "loss": 4.131511688232422,
+ "step": 5974
+ },
+ {
+ "epoch": 82.99213973799127,
+ "grad_norm": 0.015611122362315655,
+ "learning_rate": 0.0006,
+ "loss": 4.246518135070801,
+ "step": 5975
+ },
+ {
+ "epoch": 83.0,
+ "grad_norm": 0.016682764515280724,
+ "learning_rate": 0.0006,
+ "loss": 4.110680103302002,
+ "step": 5976
+ },
+ {
+ "epoch": 83.0,
+ "eval_loss": 4.641707897186279,
+ "eval_runtime": 57.2799,
+ "eval_samples_per_second": 42.633,
+ "eval_steps_per_second": 1.344,
+ "step": 5976
+ },
+ {
+ "epoch": 83.01397379912663,
+ "grad_norm": 0.015007982030510902,
+ "learning_rate": 0.0006,
+ "loss": 4.100912094116211,
+ "step": 5977
+ },
+ {
+ "epoch": 83.02794759825328,
+ "grad_norm": 0.017609436064958572,
+ "learning_rate": 0.0006,
+ "loss": 4.233399868011475,
+ "step": 5978
+ },
+ {
+ "epoch": 83.04192139737991,
+ "grad_norm": 0.019817272201180458,
+ "learning_rate": 0.0006,
+ "loss": 4.191173076629639,
+ "step": 5979
+ },
+ {
+ "epoch": 83.05589519650655,
+ "grad_norm": 0.017752140760421753,
+ "learning_rate": 0.0006,
+ "loss": 4.044437885284424,
+ "step": 5980
+ },
+ {
+ "epoch": 83.06986899563319,
+ "grad_norm": 0.015482760034501553,
+ "learning_rate": 0.0006,
+ "loss": 4.319931983947754,
+ "step": 5981
+ },
+ {
+ "epoch": 83.08384279475983,
+ "grad_norm": 0.015153988264501095,
+ "learning_rate": 0.0006,
+ "loss": 4.095277309417725,
+ "step": 5982
+ },
+ {
+ "epoch": 83.09781659388646,
+ "grad_norm": 0.017025282606482506,
+ "learning_rate": 0.0006,
+ "loss": 4.235110759735107,
+ "step": 5983
+ },
+ {
+ "epoch": 83.1117903930131,
+ "grad_norm": 0.01566586270928383,
+ "learning_rate": 0.0006,
+ "loss": 4.1934685707092285,
+ "step": 5984
+ },
+ {
+ "epoch": 83.12576419213974,
+ "grad_norm": 0.016419263556599617,
+ "learning_rate": 0.0006,
+ "loss": 4.234419822692871,
+ "step": 5985
+ },
+ {
+ "epoch": 83.13973799126637,
+ "grad_norm": 0.017132675275206566,
+ "learning_rate": 0.0006,
+ "loss": 4.221447944641113,
+ "step": 5986
+ },
+ {
+ "epoch": 83.15371179039302,
+ "grad_norm": 0.016138629987835884,
+ "learning_rate": 0.0006,
+ "loss": 4.199830055236816,
+ "step": 5987
+ },
+ {
+ "epoch": 83.16768558951965,
+ "grad_norm": 0.016210174188017845,
+ "learning_rate": 0.0006,
+ "loss": 4.081717491149902,
+ "step": 5988
+ },
+ {
+ "epoch": 83.18165938864628,
+ "grad_norm": 0.016055308282375336,
+ "learning_rate": 0.0006,
+ "loss": 4.247655391693115,
+ "step": 5989
+ },
+ {
+ "epoch": 83.19563318777293,
+ "grad_norm": 0.016132952645421028,
+ "learning_rate": 0.0006,
+ "loss": 4.262535572052002,
+ "step": 5990
+ },
+ {
+ "epoch": 83.20960698689956,
+ "grad_norm": 0.01779935695230961,
+ "learning_rate": 0.0006,
+ "loss": 4.153120040893555,
+ "step": 5991
+ },
+ {
+ "epoch": 83.2235807860262,
+ "grad_norm": 0.018360646441578865,
+ "learning_rate": 0.0006,
+ "loss": 4.232644081115723,
+ "step": 5992
+ },
+ {
+ "epoch": 83.23755458515284,
+ "grad_norm": 0.019629675894975662,
+ "learning_rate": 0.0006,
+ "loss": 3.9758734703063965,
+ "step": 5993
+ },
+ {
+ "epoch": 83.25152838427948,
+ "grad_norm": 0.017819996923208237,
+ "learning_rate": 0.0006,
+ "loss": 4.196932315826416,
+ "step": 5994
+ },
+ {
+ "epoch": 83.26550218340611,
+ "grad_norm": 0.01557681243866682,
+ "learning_rate": 0.0006,
+ "loss": 4.122654914855957,
+ "step": 5995
+ },
+ {
+ "epoch": 83.27947598253274,
+ "grad_norm": 0.017858853563666344,
+ "learning_rate": 0.0006,
+ "loss": 4.185330867767334,
+ "step": 5996
+ },
+ {
+ "epoch": 83.29344978165939,
+ "grad_norm": 0.018065953627228737,
+ "learning_rate": 0.0006,
+ "loss": 4.218471050262451,
+ "step": 5997
+ },
+ {
+ "epoch": 83.30742358078602,
+ "grad_norm": 0.017247939482331276,
+ "learning_rate": 0.0006,
+ "loss": 4.127076148986816,
+ "step": 5998
+ },
+ {
+ "epoch": 83.32139737991267,
+ "grad_norm": 0.017047952860593796,
+ "learning_rate": 0.0006,
+ "loss": 4.131356716156006,
+ "step": 5999
+ },
+ {
+ "epoch": 83.3353711790393,
+ "grad_norm": 0.015905629843473434,
+ "learning_rate": 0.0006,
+ "loss": 4.2239789962768555,
+ "step": 6000
+ },
+ {
+ "epoch": 83.34934497816595,
+ "grad_norm": 0.016522347927093506,
+ "learning_rate": 0.0006,
+ "loss": 4.054897785186768,
+ "step": 6001
+ },
+ {
+ "epoch": 83.36331877729258,
+ "grad_norm": 0.01622571051120758,
+ "learning_rate": 0.0006,
+ "loss": 4.284210681915283,
+ "step": 6002
+ },
+ {
+ "epoch": 83.37729257641921,
+ "grad_norm": 0.01640467159450054,
+ "learning_rate": 0.0006,
+ "loss": 4.268840789794922,
+ "step": 6003
+ },
+ {
+ "epoch": 83.39126637554585,
+ "grad_norm": 0.015698859468102455,
+ "learning_rate": 0.0006,
+ "loss": 4.179793357849121,
+ "step": 6004
+ },
+ {
+ "epoch": 83.40524017467249,
+ "grad_norm": 0.016694528982043266,
+ "learning_rate": 0.0006,
+ "loss": 4.132297515869141,
+ "step": 6005
+ },
+ {
+ "epoch": 83.41921397379913,
+ "grad_norm": 0.015345900319516659,
+ "learning_rate": 0.0006,
+ "loss": 4.090163230895996,
+ "step": 6006
+ },
+ {
+ "epoch": 83.43318777292576,
+ "grad_norm": 0.014722758904099464,
+ "learning_rate": 0.0006,
+ "loss": 4.179166793823242,
+ "step": 6007
+ },
+ {
+ "epoch": 83.44716157205241,
+ "grad_norm": 0.014884887263178825,
+ "learning_rate": 0.0006,
+ "loss": 4.137659072875977,
+ "step": 6008
+ },
+ {
+ "epoch": 83.46113537117904,
+ "grad_norm": 0.014041902497410774,
+ "learning_rate": 0.0006,
+ "loss": 4.1202073097229,
+ "step": 6009
+ },
+ {
+ "epoch": 83.47510917030567,
+ "grad_norm": 0.012799333781003952,
+ "learning_rate": 0.0006,
+ "loss": 4.092607498168945,
+ "step": 6010
+ },
+ {
+ "epoch": 83.48908296943232,
+ "grad_norm": 0.013272707350552082,
+ "learning_rate": 0.0006,
+ "loss": 4.132715225219727,
+ "step": 6011
+ },
+ {
+ "epoch": 83.50305676855895,
+ "grad_norm": 0.012934371829032898,
+ "learning_rate": 0.0006,
+ "loss": 4.230278968811035,
+ "step": 6012
+ },
+ {
+ "epoch": 83.5170305676856,
+ "grad_norm": 0.0120006762444973,
+ "learning_rate": 0.0006,
+ "loss": 4.140068531036377,
+ "step": 6013
+ },
+ {
+ "epoch": 83.53100436681223,
+ "grad_norm": 0.013616700656712055,
+ "learning_rate": 0.0006,
+ "loss": 4.20107364654541,
+ "step": 6014
+ },
+ {
+ "epoch": 83.54497816593886,
+ "grad_norm": 0.013957851566374302,
+ "learning_rate": 0.0006,
+ "loss": 4.124945163726807,
+ "step": 6015
+ },
+ {
+ "epoch": 83.5589519650655,
+ "grad_norm": 0.012762055732309818,
+ "learning_rate": 0.0006,
+ "loss": 4.022052764892578,
+ "step": 6016
+ },
+ {
+ "epoch": 83.57292576419214,
+ "grad_norm": 0.013080576434731483,
+ "learning_rate": 0.0006,
+ "loss": 4.103252410888672,
+ "step": 6017
+ },
+ {
+ "epoch": 83.58689956331878,
+ "grad_norm": 0.013957500457763672,
+ "learning_rate": 0.0006,
+ "loss": 4.235089302062988,
+ "step": 6018
+ },
+ {
+ "epoch": 83.60087336244541,
+ "grad_norm": 0.014707164838910103,
+ "learning_rate": 0.0006,
+ "loss": 4.191570281982422,
+ "step": 6019
+ },
+ {
+ "epoch": 83.61484716157206,
+ "grad_norm": 0.012807869352400303,
+ "learning_rate": 0.0006,
+ "loss": 4.294543266296387,
+ "step": 6020
+ },
+ {
+ "epoch": 83.62882096069869,
+ "grad_norm": 0.012027994729578495,
+ "learning_rate": 0.0006,
+ "loss": 4.13849401473999,
+ "step": 6021
+ },
+ {
+ "epoch": 83.64279475982532,
+ "grad_norm": 0.012147724628448486,
+ "learning_rate": 0.0006,
+ "loss": 4.127481937408447,
+ "step": 6022
+ },
+ {
+ "epoch": 83.65676855895197,
+ "grad_norm": 0.013241392560303211,
+ "learning_rate": 0.0006,
+ "loss": 4.218972682952881,
+ "step": 6023
+ },
+ {
+ "epoch": 83.6707423580786,
+ "grad_norm": 0.014071648940443993,
+ "learning_rate": 0.0006,
+ "loss": 4.096724033355713,
+ "step": 6024
+ },
+ {
+ "epoch": 83.68471615720524,
+ "grad_norm": 0.015212797559797764,
+ "learning_rate": 0.0006,
+ "loss": 4.27097225189209,
+ "step": 6025
+ },
+ {
+ "epoch": 83.69868995633188,
+ "grad_norm": 0.017647385597229004,
+ "learning_rate": 0.0006,
+ "loss": 4.209835529327393,
+ "step": 6026
+ },
+ {
+ "epoch": 83.71266375545852,
+ "grad_norm": 0.01709647662937641,
+ "learning_rate": 0.0006,
+ "loss": 4.042106628417969,
+ "step": 6027
+ },
+ {
+ "epoch": 83.72663755458515,
+ "grad_norm": 0.014530927874147892,
+ "learning_rate": 0.0006,
+ "loss": 4.181731700897217,
+ "step": 6028
+ },
+ {
+ "epoch": 83.74061135371178,
+ "grad_norm": 0.015356711111962795,
+ "learning_rate": 0.0006,
+ "loss": 4.314183235168457,
+ "step": 6029
+ },
+ {
+ "epoch": 83.75458515283843,
+ "grad_norm": 0.015500029549002647,
+ "learning_rate": 0.0006,
+ "loss": 4.06776762008667,
+ "step": 6030
+ },
+ {
+ "epoch": 83.76855895196506,
+ "grad_norm": 0.015351332724094391,
+ "learning_rate": 0.0006,
+ "loss": 4.185141563415527,
+ "step": 6031
+ },
+ {
+ "epoch": 83.78253275109171,
+ "grad_norm": 0.016678154468536377,
+ "learning_rate": 0.0006,
+ "loss": 4.179529666900635,
+ "step": 6032
+ },
+ {
+ "epoch": 83.79650655021834,
+ "grad_norm": 0.0156265702098608,
+ "learning_rate": 0.0006,
+ "loss": 4.238584041595459,
+ "step": 6033
+ },
+ {
+ "epoch": 83.81048034934499,
+ "grad_norm": 0.016418596729636192,
+ "learning_rate": 0.0006,
+ "loss": 4.1593523025512695,
+ "step": 6034
+ },
+ {
+ "epoch": 83.82445414847162,
+ "grad_norm": 0.016113586723804474,
+ "learning_rate": 0.0006,
+ "loss": 4.157470226287842,
+ "step": 6035
+ },
+ {
+ "epoch": 83.83842794759825,
+ "grad_norm": 0.014434042386710644,
+ "learning_rate": 0.0006,
+ "loss": 4.196226119995117,
+ "step": 6036
+ },
+ {
+ "epoch": 83.8524017467249,
+ "grad_norm": 0.012647042982280254,
+ "learning_rate": 0.0006,
+ "loss": 4.130589485168457,
+ "step": 6037
+ },
+ {
+ "epoch": 83.86637554585153,
+ "grad_norm": 0.014412877149879932,
+ "learning_rate": 0.0006,
+ "loss": 4.154845237731934,
+ "step": 6038
+ },
+ {
+ "epoch": 83.88034934497817,
+ "grad_norm": 0.014222027733922005,
+ "learning_rate": 0.0006,
+ "loss": 4.189724922180176,
+ "step": 6039
+ },
+ {
+ "epoch": 83.8943231441048,
+ "grad_norm": 0.013084255158901215,
+ "learning_rate": 0.0006,
+ "loss": 4.303338527679443,
+ "step": 6040
+ },
+ {
+ "epoch": 83.90829694323143,
+ "grad_norm": 0.012659232132136822,
+ "learning_rate": 0.0006,
+ "loss": 4.08192253112793,
+ "step": 6041
+ },
+ {
+ "epoch": 83.92227074235808,
+ "grad_norm": 0.01246276218444109,
+ "learning_rate": 0.0006,
+ "loss": 4.283324718475342,
+ "step": 6042
+ },
+ {
+ "epoch": 83.93624454148471,
+ "grad_norm": 0.01340118795633316,
+ "learning_rate": 0.0006,
+ "loss": 4.117908477783203,
+ "step": 6043
+ },
+ {
+ "epoch": 83.95021834061136,
+ "grad_norm": 0.014496722258627415,
+ "learning_rate": 0.0006,
+ "loss": 4.110318660736084,
+ "step": 6044
+ },
+ {
+ "epoch": 83.96419213973799,
+ "grad_norm": 0.013348298147320747,
+ "learning_rate": 0.0006,
+ "loss": 4.185139179229736,
+ "step": 6045
+ },
+ {
+ "epoch": 83.97816593886463,
+ "grad_norm": 0.013294277712702751,
+ "learning_rate": 0.0006,
+ "loss": 4.271645545959473,
+ "step": 6046
+ },
+ {
+ "epoch": 83.99213973799127,
+ "grad_norm": 0.014431594870984554,
+ "learning_rate": 0.0006,
+ "loss": 4.161383152008057,
+ "step": 6047
+ },
+ {
+ "epoch": 84.0,
+ "grad_norm": 0.016268927603960037,
+ "learning_rate": 0.0006,
+ "loss": 4.178324222564697,
+ "step": 6048
+ },
+ {
+ "epoch": 84.0,
+ "eval_loss": 4.6307549476623535,
+ "eval_runtime": 56.8521,
+ "eval_samples_per_second": 42.954,
+ "eval_steps_per_second": 1.354,
+ "step": 6048
+ },
+ {
+ "epoch": 84.01397379912663,
+ "grad_norm": 0.014864908531308174,
+ "learning_rate": 0.0006,
+ "loss": 4.121300220489502,
+ "step": 6049
+ },
+ {
+ "epoch": 84.02794759825328,
+ "grad_norm": 0.014863367192447186,
+ "learning_rate": 0.0006,
+ "loss": 4.1650543212890625,
+ "step": 6050
+ },
+ {
+ "epoch": 84.04192139737991,
+ "grad_norm": 0.014379646629095078,
+ "learning_rate": 0.0006,
+ "loss": 4.0201239585876465,
+ "step": 6051
+ },
+ {
+ "epoch": 84.05589519650655,
+ "grad_norm": 0.014871201477944851,
+ "learning_rate": 0.0006,
+ "loss": 4.0820698738098145,
+ "step": 6052
+ },
+ {
+ "epoch": 84.06986899563319,
+ "grad_norm": 0.013085179962217808,
+ "learning_rate": 0.0006,
+ "loss": 4.223984718322754,
+ "step": 6053
+ },
+ {
+ "epoch": 84.08384279475983,
+ "grad_norm": 0.012485949322581291,
+ "learning_rate": 0.0006,
+ "loss": 4.182206153869629,
+ "step": 6054
+ },
+ {
+ "epoch": 84.09781659388646,
+ "grad_norm": 0.013653291389346123,
+ "learning_rate": 0.0006,
+ "loss": 4.142557144165039,
+ "step": 6055
+ },
+ {
+ "epoch": 84.1117903930131,
+ "grad_norm": 0.013930532149970531,
+ "learning_rate": 0.0006,
+ "loss": 4.13972282409668,
+ "step": 6056
+ },
+ {
+ "epoch": 84.12576419213974,
+ "grad_norm": 0.01297379657626152,
+ "learning_rate": 0.0006,
+ "loss": 4.144664764404297,
+ "step": 6057
+ },
+ {
+ "epoch": 84.13973799126637,
+ "grad_norm": 0.013210492208600044,
+ "learning_rate": 0.0006,
+ "loss": 4.253764629364014,
+ "step": 6058
+ },
+ {
+ "epoch": 84.15371179039302,
+ "grad_norm": 0.014650176279246807,
+ "learning_rate": 0.0006,
+ "loss": 4.246866226196289,
+ "step": 6059
+ },
+ {
+ "epoch": 84.16768558951965,
+ "grad_norm": 0.01482292078435421,
+ "learning_rate": 0.0006,
+ "loss": 4.140187740325928,
+ "step": 6060
+ },
+ {
+ "epoch": 84.18165938864628,
+ "grad_norm": 0.01592552848160267,
+ "learning_rate": 0.0006,
+ "loss": 4.170705318450928,
+ "step": 6061
+ },
+ {
+ "epoch": 84.19563318777293,
+ "grad_norm": 0.017152156680822372,
+ "learning_rate": 0.0006,
+ "loss": 4.126092433929443,
+ "step": 6062
+ },
+ {
+ "epoch": 84.20960698689956,
+ "grad_norm": 0.015880905091762543,
+ "learning_rate": 0.0006,
+ "loss": 4.161938190460205,
+ "step": 6063
+ },
+ {
+ "epoch": 84.2235807860262,
+ "grad_norm": 0.015451471321284771,
+ "learning_rate": 0.0006,
+ "loss": 4.263461112976074,
+ "step": 6064
+ },
+ {
+ "epoch": 84.23755458515284,
+ "grad_norm": 0.014365996234118938,
+ "learning_rate": 0.0006,
+ "loss": 4.14488410949707,
+ "step": 6065
+ },
+ {
+ "epoch": 84.25152838427948,
+ "grad_norm": 0.016364749521017075,
+ "learning_rate": 0.0006,
+ "loss": 4.112489700317383,
+ "step": 6066
+ },
+ {
+ "epoch": 84.26550218340611,
+ "grad_norm": 0.01791754551231861,
+ "learning_rate": 0.0006,
+ "loss": 4.133513450622559,
+ "step": 6067
+ },
+ {
+ "epoch": 84.27947598253274,
+ "grad_norm": 0.016609208658337593,
+ "learning_rate": 0.0006,
+ "loss": 4.201829433441162,
+ "step": 6068
+ },
+ {
+ "epoch": 84.29344978165939,
+ "grad_norm": 0.017587820068001747,
+ "learning_rate": 0.0006,
+ "loss": 4.203151702880859,
+ "step": 6069
+ },
+ {
+ "epoch": 84.30742358078602,
+ "grad_norm": 0.017047742381691933,
+ "learning_rate": 0.0006,
+ "loss": 4.077008247375488,
+ "step": 6070
+ },
+ {
+ "epoch": 84.32139737991267,
+ "grad_norm": 0.01524646207690239,
+ "learning_rate": 0.0006,
+ "loss": 4.198038101196289,
+ "step": 6071
+ },
+ {
+ "epoch": 84.3353711790393,
+ "grad_norm": 0.014196421951055527,
+ "learning_rate": 0.0006,
+ "loss": 4.164245128631592,
+ "step": 6072
+ },
+ {
+ "epoch": 84.34934497816595,
+ "grad_norm": 0.016291776672005653,
+ "learning_rate": 0.0006,
+ "loss": 4.150152206420898,
+ "step": 6073
+ },
+ {
+ "epoch": 84.36331877729258,
+ "grad_norm": 0.015627920627593994,
+ "learning_rate": 0.0006,
+ "loss": 4.110006332397461,
+ "step": 6074
+ },
+ {
+ "epoch": 84.37729257641921,
+ "grad_norm": 0.013251487165689468,
+ "learning_rate": 0.0006,
+ "loss": 4.101463317871094,
+ "step": 6075
+ },
+ {
+ "epoch": 84.39126637554585,
+ "grad_norm": 0.013601605780422688,
+ "learning_rate": 0.0006,
+ "loss": 4.191105842590332,
+ "step": 6076
+ },
+ {
+ "epoch": 84.40524017467249,
+ "grad_norm": 0.014878741465508938,
+ "learning_rate": 0.0006,
+ "loss": 4.153046131134033,
+ "step": 6077
+ },
+ {
+ "epoch": 84.41921397379913,
+ "grad_norm": 0.014783897437155247,
+ "learning_rate": 0.0006,
+ "loss": 4.362000465393066,
+ "step": 6078
+ },
+ {
+ "epoch": 84.43318777292576,
+ "grad_norm": 0.014221332035958767,
+ "learning_rate": 0.0006,
+ "loss": 4.144364833831787,
+ "step": 6079
+ },
+ {
+ "epoch": 84.44716157205241,
+ "grad_norm": 0.014911034144461155,
+ "learning_rate": 0.0006,
+ "loss": 4.088857650756836,
+ "step": 6080
+ },
+ {
+ "epoch": 84.46113537117904,
+ "grad_norm": 0.014924770221114159,
+ "learning_rate": 0.0006,
+ "loss": 4.1677446365356445,
+ "step": 6081
+ },
+ {
+ "epoch": 84.47510917030567,
+ "grad_norm": 0.014394409023225307,
+ "learning_rate": 0.0006,
+ "loss": 4.087103843688965,
+ "step": 6082
+ },
+ {
+ "epoch": 84.48908296943232,
+ "grad_norm": 0.013926676474511623,
+ "learning_rate": 0.0006,
+ "loss": 3.9972851276397705,
+ "step": 6083
+ },
+ {
+ "epoch": 84.50305676855895,
+ "grad_norm": 0.013069954700767994,
+ "learning_rate": 0.0006,
+ "loss": 3.998237133026123,
+ "step": 6084
+ },
+ {
+ "epoch": 84.5170305676856,
+ "grad_norm": 0.012748011387884617,
+ "learning_rate": 0.0006,
+ "loss": 4.195858955383301,
+ "step": 6085
+ },
+ {
+ "epoch": 84.53100436681223,
+ "grad_norm": 0.012278077192604542,
+ "learning_rate": 0.0006,
+ "loss": 4.102165222167969,
+ "step": 6086
+ },
+ {
+ "epoch": 84.54497816593886,
+ "grad_norm": 0.014259163290262222,
+ "learning_rate": 0.0006,
+ "loss": 4.092750549316406,
+ "step": 6087
+ },
+ {
+ "epoch": 84.5589519650655,
+ "grad_norm": 0.014553683809936047,
+ "learning_rate": 0.0006,
+ "loss": 4.1412763595581055,
+ "step": 6088
+ },
+ {
+ "epoch": 84.57292576419214,
+ "grad_norm": 0.013605243526399136,
+ "learning_rate": 0.0006,
+ "loss": 4.192173480987549,
+ "step": 6089
+ },
+ {
+ "epoch": 84.58689956331878,
+ "grad_norm": 0.014259358868002892,
+ "learning_rate": 0.0006,
+ "loss": 4.356374740600586,
+ "step": 6090
+ },
+ {
+ "epoch": 84.60087336244541,
+ "grad_norm": 0.014166025444865227,
+ "learning_rate": 0.0006,
+ "loss": 4.165982246398926,
+ "step": 6091
+ },
+ {
+ "epoch": 84.61484716157206,
+ "grad_norm": 0.012198803946375847,
+ "learning_rate": 0.0006,
+ "loss": 4.287566184997559,
+ "step": 6092
+ },
+ {
+ "epoch": 84.62882096069869,
+ "grad_norm": 0.013794238679111004,
+ "learning_rate": 0.0006,
+ "loss": 4.199882984161377,
+ "step": 6093
+ },
+ {
+ "epoch": 84.64279475982532,
+ "grad_norm": 0.014757642522454262,
+ "learning_rate": 0.0006,
+ "loss": 4.116573333740234,
+ "step": 6094
+ },
+ {
+ "epoch": 84.65676855895197,
+ "grad_norm": 0.014591689221560955,
+ "learning_rate": 0.0006,
+ "loss": 4.240896224975586,
+ "step": 6095
+ },
+ {
+ "epoch": 84.6707423580786,
+ "grad_norm": 0.015217579901218414,
+ "learning_rate": 0.0006,
+ "loss": 4.257142066955566,
+ "step": 6096
+ },
+ {
+ "epoch": 84.68471615720524,
+ "grad_norm": 0.015371371991932392,
+ "learning_rate": 0.0006,
+ "loss": 4.181037425994873,
+ "step": 6097
+ },
+ {
+ "epoch": 84.69868995633188,
+ "grad_norm": 0.014412983320653439,
+ "learning_rate": 0.0006,
+ "loss": 4.2207794189453125,
+ "step": 6098
+ },
+ {
+ "epoch": 84.71266375545852,
+ "grad_norm": 0.013773671351373196,
+ "learning_rate": 0.0006,
+ "loss": 4.211123466491699,
+ "step": 6099
+ },
+ {
+ "epoch": 84.72663755458515,
+ "grad_norm": 0.012838046066462994,
+ "learning_rate": 0.0006,
+ "loss": 4.175459384918213,
+ "step": 6100
+ },
+ {
+ "epoch": 84.74061135371178,
+ "grad_norm": 0.014094929210841656,
+ "learning_rate": 0.0006,
+ "loss": 4.1724958419799805,
+ "step": 6101
+ },
+ {
+ "epoch": 84.75458515283843,
+ "grad_norm": 0.014573262073099613,
+ "learning_rate": 0.0006,
+ "loss": 4.15885066986084,
+ "step": 6102
+ },
+ {
+ "epoch": 84.76855895196506,
+ "grad_norm": 0.011964819394052029,
+ "learning_rate": 0.0006,
+ "loss": 3.9804892539978027,
+ "step": 6103
+ },
+ {
+ "epoch": 84.78253275109171,
+ "grad_norm": 0.014140539802610874,
+ "learning_rate": 0.0006,
+ "loss": 4.203708648681641,
+ "step": 6104
+ },
+ {
+ "epoch": 84.79650655021834,
+ "grad_norm": 0.016618940979242325,
+ "learning_rate": 0.0006,
+ "loss": 4.147684097290039,
+ "step": 6105
+ },
+ {
+ "epoch": 84.81048034934499,
+ "grad_norm": 0.015094385482370853,
+ "learning_rate": 0.0006,
+ "loss": 4.392223834991455,
+ "step": 6106
+ },
+ {
+ "epoch": 84.82445414847162,
+ "grad_norm": 0.014386717230081558,
+ "learning_rate": 0.0006,
+ "loss": 4.253332138061523,
+ "step": 6107
+ },
+ {
+ "epoch": 84.83842794759825,
+ "grad_norm": 0.013877755962312222,
+ "learning_rate": 0.0006,
+ "loss": 4.075056076049805,
+ "step": 6108
+ },
+ {
+ "epoch": 84.8524017467249,
+ "grad_norm": 0.014062155969440937,
+ "learning_rate": 0.0006,
+ "loss": 4.072048664093018,
+ "step": 6109
+ },
+ {
+ "epoch": 84.86637554585153,
+ "grad_norm": 0.014445619657635689,
+ "learning_rate": 0.0006,
+ "loss": 4.151832580566406,
+ "step": 6110
+ },
+ {
+ "epoch": 84.88034934497817,
+ "grad_norm": 0.015116363763809204,
+ "learning_rate": 0.0006,
+ "loss": 4.088900566101074,
+ "step": 6111
+ },
+ {
+ "epoch": 84.8943231441048,
+ "grad_norm": 0.014425474219024181,
+ "learning_rate": 0.0006,
+ "loss": 4.200483322143555,
+ "step": 6112
+ },
+ {
+ "epoch": 84.90829694323143,
+ "grad_norm": 0.01449573878198862,
+ "learning_rate": 0.0006,
+ "loss": 4.239742279052734,
+ "step": 6113
+ },
+ {
+ "epoch": 84.92227074235808,
+ "grad_norm": 0.015698261559009552,
+ "learning_rate": 0.0006,
+ "loss": 4.197391510009766,
+ "step": 6114
+ },
+ {
+ "epoch": 84.93624454148471,
+ "grad_norm": 0.014545733109116554,
+ "learning_rate": 0.0006,
+ "loss": 4.035508155822754,
+ "step": 6115
+ },
+ {
+ "epoch": 84.95021834061136,
+ "grad_norm": 0.012796483002603054,
+ "learning_rate": 0.0006,
+ "loss": 4.195248126983643,
+ "step": 6116
+ },
+ {
+ "epoch": 84.96419213973799,
+ "grad_norm": 0.01264485064893961,
+ "learning_rate": 0.0006,
+ "loss": 4.158118724822998,
+ "step": 6117
+ },
+ {
+ "epoch": 84.97816593886463,
+ "grad_norm": 0.013695276342332363,
+ "learning_rate": 0.0006,
+ "loss": 4.255288124084473,
+ "step": 6118
+ },
+ {
+ "epoch": 84.99213973799127,
+ "grad_norm": 0.013459564186632633,
+ "learning_rate": 0.0006,
+ "loss": 4.204586505889893,
+ "step": 6119
+ },
+ {
+ "epoch": 85.0,
+ "grad_norm": 0.014312652871012688,
+ "learning_rate": 0.0006,
+ "loss": 4.167879104614258,
+ "step": 6120
+ },
+ {
+ "epoch": 85.0,
+ "eval_loss": 4.608152389526367,
+ "eval_runtime": 57.1903,
+ "eval_samples_per_second": 42.7,
+ "eval_steps_per_second": 1.346,
+ "step": 6120
+ },
+ {
+ "epoch": 85.01397379912663,
+ "grad_norm": 0.013918368145823479,
+ "learning_rate": 0.0006,
+ "loss": 4.129991054534912,
+ "step": 6121
+ },
+ {
+ "epoch": 85.02794759825328,
+ "grad_norm": 0.01604384370148182,
+ "learning_rate": 0.0006,
+ "loss": 4.233926773071289,
+ "step": 6122
+ },
+ {
+ "epoch": 85.04192139737991,
+ "grad_norm": 0.016565652564167976,
+ "learning_rate": 0.0006,
+ "loss": 4.129604339599609,
+ "step": 6123
+ },
+ {
+ "epoch": 85.05589519650655,
+ "grad_norm": 0.0149751678109169,
+ "learning_rate": 0.0006,
+ "loss": 4.165839195251465,
+ "step": 6124
+ },
+ {
+ "epoch": 85.06986899563319,
+ "grad_norm": 0.013812461867928505,
+ "learning_rate": 0.0006,
+ "loss": 4.153792858123779,
+ "step": 6125
+ },
+ {
+ "epoch": 85.08384279475983,
+ "grad_norm": 0.014432745985686779,
+ "learning_rate": 0.0006,
+ "loss": 4.170895576477051,
+ "step": 6126
+ },
+ {
+ "epoch": 85.09781659388646,
+ "grad_norm": 0.015208868309855461,
+ "learning_rate": 0.0006,
+ "loss": 4.10639762878418,
+ "step": 6127
+ },
+ {
+ "epoch": 85.1117903930131,
+ "grad_norm": 0.013209272176027298,
+ "learning_rate": 0.0006,
+ "loss": 4.200364112854004,
+ "step": 6128
+ },
+ {
+ "epoch": 85.12576419213974,
+ "grad_norm": 0.014807002618908882,
+ "learning_rate": 0.0006,
+ "loss": 4.172666549682617,
+ "step": 6129
+ },
+ {
+ "epoch": 85.13973799126637,
+ "grad_norm": 0.015662945806980133,
+ "learning_rate": 0.0006,
+ "loss": 4.145219326019287,
+ "step": 6130
+ },
+ {
+ "epoch": 85.15371179039302,
+ "grad_norm": 0.014598497189581394,
+ "learning_rate": 0.0006,
+ "loss": 4.275618553161621,
+ "step": 6131
+ },
+ {
+ "epoch": 85.16768558951965,
+ "grad_norm": 0.013778576627373695,
+ "learning_rate": 0.0006,
+ "loss": 4.204716682434082,
+ "step": 6132
+ },
+ {
+ "epoch": 85.18165938864628,
+ "grad_norm": 0.015762265771627426,
+ "learning_rate": 0.0006,
+ "loss": 4.145524024963379,
+ "step": 6133
+ },
+ {
+ "epoch": 85.19563318777293,
+ "grad_norm": 0.015579195693135262,
+ "learning_rate": 0.0006,
+ "loss": 4.150625228881836,
+ "step": 6134
+ },
+ {
+ "epoch": 85.20960698689956,
+ "grad_norm": 0.016641216352581978,
+ "learning_rate": 0.0006,
+ "loss": 4.141819953918457,
+ "step": 6135
+ },
+ {
+ "epoch": 85.2235807860262,
+ "grad_norm": 0.01632271707057953,
+ "learning_rate": 0.0006,
+ "loss": 4.030620574951172,
+ "step": 6136
+ },
+ {
+ "epoch": 85.23755458515284,
+ "grad_norm": 0.015564918518066406,
+ "learning_rate": 0.0006,
+ "loss": 4.147753715515137,
+ "step": 6137
+ },
+ {
+ "epoch": 85.25152838427948,
+ "grad_norm": 0.0161129217594862,
+ "learning_rate": 0.0006,
+ "loss": 4.128524303436279,
+ "step": 6138
+ },
+ {
+ "epoch": 85.26550218340611,
+ "grad_norm": 0.017766550183296204,
+ "learning_rate": 0.0006,
+ "loss": 4.134856224060059,
+ "step": 6139
+ },
+ {
+ "epoch": 85.27947598253274,
+ "grad_norm": 0.01869054324924946,
+ "learning_rate": 0.0006,
+ "loss": 4.231565475463867,
+ "step": 6140
+ },
+ {
+ "epoch": 85.29344978165939,
+ "grad_norm": 0.01917010359466076,
+ "learning_rate": 0.0006,
+ "loss": 4.21175479888916,
+ "step": 6141
+ },
+ {
+ "epoch": 85.30742358078602,
+ "grad_norm": 0.018116485327482224,
+ "learning_rate": 0.0006,
+ "loss": 4.122644424438477,
+ "step": 6142
+ },
+ {
+ "epoch": 85.32139737991267,
+ "grad_norm": 0.015949107706546783,
+ "learning_rate": 0.0006,
+ "loss": 4.135458946228027,
+ "step": 6143
+ },
+ {
+ "epoch": 85.3353711790393,
+ "grad_norm": 0.014783683232963085,
+ "learning_rate": 0.0006,
+ "loss": 4.118994235992432,
+ "step": 6144
+ },
+ {
+ "epoch": 85.34934497816595,
+ "grad_norm": 0.015406692400574684,
+ "learning_rate": 0.0006,
+ "loss": 4.157268524169922,
+ "step": 6145
+ },
+ {
+ "epoch": 85.36331877729258,
+ "grad_norm": 0.013893205672502518,
+ "learning_rate": 0.0006,
+ "loss": 4.056166648864746,
+ "step": 6146
+ },
+ {
+ "epoch": 85.37729257641921,
+ "grad_norm": 0.014380923472344875,
+ "learning_rate": 0.0006,
+ "loss": 4.206905364990234,
+ "step": 6147
+ },
+ {
+ "epoch": 85.39126637554585,
+ "grad_norm": 0.015669258311390877,
+ "learning_rate": 0.0006,
+ "loss": 4.213881492614746,
+ "step": 6148
+ },
+ {
+ "epoch": 85.40524017467249,
+ "grad_norm": 0.01553428452461958,
+ "learning_rate": 0.0006,
+ "loss": 4.1814374923706055,
+ "step": 6149
+ },
+ {
+ "epoch": 85.41921397379913,
+ "grad_norm": 0.01697700284421444,
+ "learning_rate": 0.0006,
+ "loss": 4.083939552307129,
+ "step": 6150
+ },
+ {
+ "epoch": 85.43318777292576,
+ "grad_norm": 0.016639070585370064,
+ "learning_rate": 0.0006,
+ "loss": 4.279311180114746,
+ "step": 6151
+ },
+ {
+ "epoch": 85.44716157205241,
+ "grad_norm": 0.016285281628370285,
+ "learning_rate": 0.0006,
+ "loss": 4.244845390319824,
+ "step": 6152
+ },
+ {
+ "epoch": 85.46113537117904,
+ "grad_norm": 0.016071317717432976,
+ "learning_rate": 0.0006,
+ "loss": 4.089814186096191,
+ "step": 6153
+ },
+ {
+ "epoch": 85.47510917030567,
+ "grad_norm": 0.01640418916940689,
+ "learning_rate": 0.0006,
+ "loss": 4.168088912963867,
+ "step": 6154
+ },
+ {
+ "epoch": 85.48908296943232,
+ "grad_norm": 0.014781414531171322,
+ "learning_rate": 0.0006,
+ "loss": 4.196092128753662,
+ "step": 6155
+ },
+ {
+ "epoch": 85.50305676855895,
+ "grad_norm": 0.013393213972449303,
+ "learning_rate": 0.0006,
+ "loss": 4.193429946899414,
+ "step": 6156
+ },
+ {
+ "epoch": 85.5170305676856,
+ "grad_norm": 0.01467406190931797,
+ "learning_rate": 0.0006,
+ "loss": 4.216714859008789,
+ "step": 6157
+ },
+ {
+ "epoch": 85.53100436681223,
+ "grad_norm": 0.014950765296816826,
+ "learning_rate": 0.0006,
+ "loss": 4.093854904174805,
+ "step": 6158
+ },
+ {
+ "epoch": 85.54497816593886,
+ "grad_norm": 0.014655301347374916,
+ "learning_rate": 0.0006,
+ "loss": 4.09483528137207,
+ "step": 6159
+ },
+ {
+ "epoch": 85.5589519650655,
+ "grad_norm": 0.01425520796328783,
+ "learning_rate": 0.0006,
+ "loss": 4.090396881103516,
+ "step": 6160
+ },
+ {
+ "epoch": 85.57292576419214,
+ "grad_norm": 0.01403278298676014,
+ "learning_rate": 0.0006,
+ "loss": 4.088724136352539,
+ "step": 6161
+ },
+ {
+ "epoch": 85.58689956331878,
+ "grad_norm": 0.01364967692643404,
+ "learning_rate": 0.0006,
+ "loss": 4.095558166503906,
+ "step": 6162
+ },
+ {
+ "epoch": 85.60087336244541,
+ "grad_norm": 0.013085267506539822,
+ "learning_rate": 0.0006,
+ "loss": 4.163962364196777,
+ "step": 6163
+ },
+ {
+ "epoch": 85.61484716157206,
+ "grad_norm": 0.0136906448751688,
+ "learning_rate": 0.0006,
+ "loss": 4.197134017944336,
+ "step": 6164
+ },
+ {
+ "epoch": 85.62882096069869,
+ "grad_norm": 0.013472451828420162,
+ "learning_rate": 0.0006,
+ "loss": 4.1578369140625,
+ "step": 6165
+ },
+ {
+ "epoch": 85.64279475982532,
+ "grad_norm": 0.012329773046076298,
+ "learning_rate": 0.0006,
+ "loss": 4.231907844543457,
+ "step": 6166
+ },
+ {
+ "epoch": 85.65676855895197,
+ "grad_norm": 0.011977200396358967,
+ "learning_rate": 0.0006,
+ "loss": 4.1552958488464355,
+ "step": 6167
+ },
+ {
+ "epoch": 85.6707423580786,
+ "grad_norm": 0.013049956411123276,
+ "learning_rate": 0.0006,
+ "loss": 4.17852783203125,
+ "step": 6168
+ },
+ {
+ "epoch": 85.68471615720524,
+ "grad_norm": 0.011440741829574108,
+ "learning_rate": 0.0006,
+ "loss": 4.144646644592285,
+ "step": 6169
+ },
+ {
+ "epoch": 85.69868995633188,
+ "grad_norm": 0.012472936883568764,
+ "learning_rate": 0.0006,
+ "loss": 4.2306413650512695,
+ "step": 6170
+ },
+ {
+ "epoch": 85.71266375545852,
+ "grad_norm": 0.013015356846153736,
+ "learning_rate": 0.0006,
+ "loss": 4.193439960479736,
+ "step": 6171
+ },
+ {
+ "epoch": 85.72663755458515,
+ "grad_norm": 0.013832319527864456,
+ "learning_rate": 0.0006,
+ "loss": 4.18073034286499,
+ "step": 6172
+ },
+ {
+ "epoch": 85.74061135371178,
+ "grad_norm": 0.014613982290029526,
+ "learning_rate": 0.0006,
+ "loss": 4.223531246185303,
+ "step": 6173
+ },
+ {
+ "epoch": 85.75458515283843,
+ "grad_norm": 0.014209083281457424,
+ "learning_rate": 0.0006,
+ "loss": 4.142666816711426,
+ "step": 6174
+ },
+ {
+ "epoch": 85.76855895196506,
+ "grad_norm": 0.013396669179201126,
+ "learning_rate": 0.0006,
+ "loss": 4.185786247253418,
+ "step": 6175
+ },
+ {
+ "epoch": 85.78253275109171,
+ "grad_norm": 0.01364929135888815,
+ "learning_rate": 0.0006,
+ "loss": 4.125226974487305,
+ "step": 6176
+ },
+ {
+ "epoch": 85.79650655021834,
+ "grad_norm": 0.012561003677546978,
+ "learning_rate": 0.0006,
+ "loss": 4.084519386291504,
+ "step": 6177
+ },
+ {
+ "epoch": 85.81048034934499,
+ "grad_norm": 0.013508481904864311,
+ "learning_rate": 0.0006,
+ "loss": 4.193173885345459,
+ "step": 6178
+ },
+ {
+ "epoch": 85.82445414847162,
+ "grad_norm": 0.014154613018035889,
+ "learning_rate": 0.0006,
+ "loss": 4.257272720336914,
+ "step": 6179
+ },
+ {
+ "epoch": 85.83842794759825,
+ "grad_norm": 0.012478531338274479,
+ "learning_rate": 0.0006,
+ "loss": 4.161380767822266,
+ "step": 6180
+ },
+ {
+ "epoch": 85.8524017467249,
+ "grad_norm": 0.013468451797962189,
+ "learning_rate": 0.0006,
+ "loss": 4.121395111083984,
+ "step": 6181
+ },
+ {
+ "epoch": 85.86637554585153,
+ "grad_norm": 0.015155194327235222,
+ "learning_rate": 0.0006,
+ "loss": 4.242815971374512,
+ "step": 6182
+ },
+ {
+ "epoch": 85.88034934497817,
+ "grad_norm": 0.016223272308707237,
+ "learning_rate": 0.0006,
+ "loss": 4.176027297973633,
+ "step": 6183
+ },
+ {
+ "epoch": 85.8943231441048,
+ "grad_norm": 0.013741295784711838,
+ "learning_rate": 0.0006,
+ "loss": 4.131257057189941,
+ "step": 6184
+ },
+ {
+ "epoch": 85.90829694323143,
+ "grad_norm": 0.014982067979872227,
+ "learning_rate": 0.0006,
+ "loss": 4.237896919250488,
+ "step": 6185
+ },
+ {
+ "epoch": 85.92227074235808,
+ "grad_norm": 0.018197322264313698,
+ "learning_rate": 0.0006,
+ "loss": 4.154962539672852,
+ "step": 6186
+ },
+ {
+ "epoch": 85.93624454148471,
+ "grad_norm": 0.01971200481057167,
+ "learning_rate": 0.0006,
+ "loss": 4.16819429397583,
+ "step": 6187
+ },
+ {
+ "epoch": 85.95021834061136,
+ "grad_norm": 0.018008720129728317,
+ "learning_rate": 0.0006,
+ "loss": 4.146164417266846,
+ "step": 6188
+ },
+ {
+ "epoch": 85.96419213973799,
+ "grad_norm": 0.015859264880418777,
+ "learning_rate": 0.0006,
+ "loss": 4.208099365234375,
+ "step": 6189
+ },
+ {
+ "epoch": 85.97816593886463,
+ "grad_norm": 0.013846252113580704,
+ "learning_rate": 0.0006,
+ "loss": 4.1881184577941895,
+ "step": 6190
+ },
+ {
+ "epoch": 85.99213973799127,
+ "grad_norm": 0.018641939386725426,
+ "learning_rate": 0.0006,
+ "loss": 4.127632141113281,
+ "step": 6191
+ },
+ {
+ "epoch": 86.0,
+ "grad_norm": 0.02385227382183075,
+ "learning_rate": 0.0006,
+ "loss": 4.281900405883789,
+ "step": 6192
+ },
+ {
+ "epoch": 86.0,
+ "eval_loss": 4.625458717346191,
+ "eval_runtime": 57.0634,
+ "eval_samples_per_second": 42.795,
+ "eval_steps_per_second": 1.349,
+ "step": 6192
+ },
+ {
+ "epoch": 86.01397379912663,
+ "grad_norm": 0.019032124429941177,
+ "learning_rate": 0.0006,
+ "loss": 4.2163190841674805,
+ "step": 6193
+ },
+ {
+ "epoch": 86.02794759825328,
+ "grad_norm": 0.018049925565719604,
+ "learning_rate": 0.0006,
+ "loss": 4.1534037590026855,
+ "step": 6194
+ },
+ {
+ "epoch": 86.04192139737991,
+ "grad_norm": 0.019116360694169998,
+ "learning_rate": 0.0006,
+ "loss": 4.212411403656006,
+ "step": 6195
+ },
+ {
+ "epoch": 86.05589519650655,
+ "grad_norm": 0.018967336043715477,
+ "learning_rate": 0.0006,
+ "loss": 4.190918922424316,
+ "step": 6196
+ },
+ {
+ "epoch": 86.06986899563319,
+ "grad_norm": 0.01945611648261547,
+ "learning_rate": 0.0006,
+ "loss": 4.18459415435791,
+ "step": 6197
+ },
+ {
+ "epoch": 86.08384279475983,
+ "grad_norm": 0.018104076385498047,
+ "learning_rate": 0.0006,
+ "loss": 4.0942816734313965,
+ "step": 6198
+ },
+ {
+ "epoch": 86.09781659388646,
+ "grad_norm": 0.01815992221236229,
+ "learning_rate": 0.0006,
+ "loss": 4.011547565460205,
+ "step": 6199
+ },
+ {
+ "epoch": 86.1117903930131,
+ "grad_norm": 0.017574431374669075,
+ "learning_rate": 0.0006,
+ "loss": 4.051593780517578,
+ "step": 6200
+ },
+ {
+ "epoch": 86.12576419213974,
+ "grad_norm": 0.01516508124768734,
+ "learning_rate": 0.0006,
+ "loss": 4.239673614501953,
+ "step": 6201
+ },
+ {
+ "epoch": 86.13973799126637,
+ "grad_norm": 0.015689702704548836,
+ "learning_rate": 0.0006,
+ "loss": 4.1085734367370605,
+ "step": 6202
+ },
+ {
+ "epoch": 86.15371179039302,
+ "grad_norm": 0.01596900261938572,
+ "learning_rate": 0.0006,
+ "loss": 4.1509599685668945,
+ "step": 6203
+ },
+ {
+ "epoch": 86.16768558951965,
+ "grad_norm": 0.01550096832215786,
+ "learning_rate": 0.0006,
+ "loss": 4.160238265991211,
+ "step": 6204
+ },
+ {
+ "epoch": 86.18165938864628,
+ "grad_norm": 0.014601529575884342,
+ "learning_rate": 0.0006,
+ "loss": 4.029219150543213,
+ "step": 6205
+ },
+ {
+ "epoch": 86.19563318777293,
+ "grad_norm": 0.014006740413606167,
+ "learning_rate": 0.0006,
+ "loss": 4.117501258850098,
+ "step": 6206
+ },
+ {
+ "epoch": 86.20960698689956,
+ "grad_norm": 0.013191957026720047,
+ "learning_rate": 0.0006,
+ "loss": 4.252002716064453,
+ "step": 6207
+ },
+ {
+ "epoch": 86.2235807860262,
+ "grad_norm": 0.013031814247369766,
+ "learning_rate": 0.0006,
+ "loss": 4.209517478942871,
+ "step": 6208
+ },
+ {
+ "epoch": 86.23755458515284,
+ "grad_norm": 0.013605714775621891,
+ "learning_rate": 0.0006,
+ "loss": 4.176309585571289,
+ "step": 6209
+ },
+ {
+ "epoch": 86.25152838427948,
+ "grad_norm": 0.012753087095916271,
+ "learning_rate": 0.0006,
+ "loss": 4.097990036010742,
+ "step": 6210
+ },
+ {
+ "epoch": 86.26550218340611,
+ "grad_norm": 0.012940628454089165,
+ "learning_rate": 0.0006,
+ "loss": 4.085589408874512,
+ "step": 6211
+ },
+ {
+ "epoch": 86.27947598253274,
+ "grad_norm": 0.014286619611084461,
+ "learning_rate": 0.0006,
+ "loss": 4.199261665344238,
+ "step": 6212
+ },
+ {
+ "epoch": 86.29344978165939,
+ "grad_norm": 0.014643525704741478,
+ "learning_rate": 0.0006,
+ "loss": 4.23718786239624,
+ "step": 6213
+ },
+ {
+ "epoch": 86.30742358078602,
+ "grad_norm": 0.015182147733867168,
+ "learning_rate": 0.0006,
+ "loss": 4.219855308532715,
+ "step": 6214
+ },
+ {
+ "epoch": 86.32139737991267,
+ "grad_norm": 0.01530102826654911,
+ "learning_rate": 0.0006,
+ "loss": 4.190094947814941,
+ "step": 6215
+ },
+ {
+ "epoch": 86.3353711790393,
+ "grad_norm": 0.015492603182792664,
+ "learning_rate": 0.0006,
+ "loss": 4.153780460357666,
+ "step": 6216
+ },
+ {
+ "epoch": 86.34934497816595,
+ "grad_norm": 0.015339907258749008,
+ "learning_rate": 0.0006,
+ "loss": 4.182865619659424,
+ "step": 6217
+ },
+ {
+ "epoch": 86.36331877729258,
+ "grad_norm": 0.015576880425214767,
+ "learning_rate": 0.0006,
+ "loss": 4.196417331695557,
+ "step": 6218
+ },
+ {
+ "epoch": 86.37729257641921,
+ "grad_norm": 0.014805737882852554,
+ "learning_rate": 0.0006,
+ "loss": 4.265018463134766,
+ "step": 6219
+ },
+ {
+ "epoch": 86.39126637554585,
+ "grad_norm": 0.01597665622830391,
+ "learning_rate": 0.0006,
+ "loss": 4.2864837646484375,
+ "step": 6220
+ },
+ {
+ "epoch": 86.40524017467249,
+ "grad_norm": 0.01693643443286419,
+ "learning_rate": 0.0006,
+ "loss": 4.118192672729492,
+ "step": 6221
+ },
+ {
+ "epoch": 86.41921397379913,
+ "grad_norm": 0.016648951917886734,
+ "learning_rate": 0.0006,
+ "loss": 4.23712682723999,
+ "step": 6222
+ },
+ {
+ "epoch": 86.43318777292576,
+ "grad_norm": 0.014809766784310341,
+ "learning_rate": 0.0006,
+ "loss": 4.251391410827637,
+ "step": 6223
+ },
+ {
+ "epoch": 86.44716157205241,
+ "grad_norm": 0.014333373866975307,
+ "learning_rate": 0.0006,
+ "loss": 4.17243766784668,
+ "step": 6224
+ },
+ {
+ "epoch": 86.46113537117904,
+ "grad_norm": 0.013912711292505264,
+ "learning_rate": 0.0006,
+ "loss": 4.250836372375488,
+ "step": 6225
+ },
+ {
+ "epoch": 86.47510917030567,
+ "grad_norm": 0.014821341261267662,
+ "learning_rate": 0.0006,
+ "loss": 4.073995590209961,
+ "step": 6226
+ },
+ {
+ "epoch": 86.48908296943232,
+ "grad_norm": 0.016634922474622726,
+ "learning_rate": 0.0006,
+ "loss": 4.155867576599121,
+ "step": 6227
+ },
+ {
+ "epoch": 86.50305676855895,
+ "grad_norm": 0.01748695969581604,
+ "learning_rate": 0.0006,
+ "loss": 4.139344692230225,
+ "step": 6228
+ },
+ {
+ "epoch": 86.5170305676856,
+ "grad_norm": 0.01569363661110401,
+ "learning_rate": 0.0006,
+ "loss": 4.23923397064209,
+ "step": 6229
+ },
+ {
+ "epoch": 86.53100436681223,
+ "grad_norm": 0.014037121087312698,
+ "learning_rate": 0.0006,
+ "loss": 4.200334072113037,
+ "step": 6230
+ },
+ {
+ "epoch": 86.54497816593886,
+ "grad_norm": 0.01401279866695404,
+ "learning_rate": 0.0006,
+ "loss": 4.035246849060059,
+ "step": 6231
+ },
+ {
+ "epoch": 86.5589519650655,
+ "grad_norm": 0.014603731222450733,
+ "learning_rate": 0.0006,
+ "loss": 4.141385555267334,
+ "step": 6232
+ },
+ {
+ "epoch": 86.57292576419214,
+ "grad_norm": 0.015113704837858677,
+ "learning_rate": 0.0006,
+ "loss": 4.130824089050293,
+ "step": 6233
+ },
+ {
+ "epoch": 86.58689956331878,
+ "grad_norm": 0.01588696800172329,
+ "learning_rate": 0.0006,
+ "loss": 4.05495023727417,
+ "step": 6234
+ },
+ {
+ "epoch": 86.60087336244541,
+ "grad_norm": 0.01593640074133873,
+ "learning_rate": 0.0006,
+ "loss": 4.22935152053833,
+ "step": 6235
+ },
+ {
+ "epoch": 86.61484716157206,
+ "grad_norm": 0.014292445965111256,
+ "learning_rate": 0.0006,
+ "loss": 4.309839248657227,
+ "step": 6236
+ },
+ {
+ "epoch": 86.62882096069869,
+ "grad_norm": 0.015471610240638256,
+ "learning_rate": 0.0006,
+ "loss": 4.268278121948242,
+ "step": 6237
+ },
+ {
+ "epoch": 86.64279475982532,
+ "grad_norm": 0.016643544659018517,
+ "learning_rate": 0.0006,
+ "loss": 4.135109901428223,
+ "step": 6238
+ },
+ {
+ "epoch": 86.65676855895197,
+ "grad_norm": 0.014733029529452324,
+ "learning_rate": 0.0006,
+ "loss": 4.060197830200195,
+ "step": 6239
+ },
+ {
+ "epoch": 86.6707423580786,
+ "grad_norm": 0.014869216829538345,
+ "learning_rate": 0.0006,
+ "loss": 4.126917362213135,
+ "step": 6240
+ },
+ {
+ "epoch": 86.68471615720524,
+ "grad_norm": 0.01608203910291195,
+ "learning_rate": 0.0006,
+ "loss": 4.183716773986816,
+ "step": 6241
+ },
+ {
+ "epoch": 86.69868995633188,
+ "grad_norm": 0.016178175806999207,
+ "learning_rate": 0.0006,
+ "loss": 4.118875503540039,
+ "step": 6242
+ },
+ {
+ "epoch": 86.71266375545852,
+ "grad_norm": 0.014576597139239311,
+ "learning_rate": 0.0006,
+ "loss": 4.261541366577148,
+ "step": 6243
+ },
+ {
+ "epoch": 86.72663755458515,
+ "grad_norm": 0.014466420747339725,
+ "learning_rate": 0.0006,
+ "loss": 4.307284355163574,
+ "step": 6244
+ },
+ {
+ "epoch": 86.74061135371178,
+ "grad_norm": 0.014042644761502743,
+ "learning_rate": 0.0006,
+ "loss": 4.2729997634887695,
+ "step": 6245
+ },
+ {
+ "epoch": 86.75458515283843,
+ "grad_norm": 0.014589524827897549,
+ "learning_rate": 0.0006,
+ "loss": 4.072981834411621,
+ "step": 6246
+ },
+ {
+ "epoch": 86.76855895196506,
+ "grad_norm": 0.018765483051538467,
+ "learning_rate": 0.0006,
+ "loss": 4.264769554138184,
+ "step": 6247
+ },
+ {
+ "epoch": 86.78253275109171,
+ "grad_norm": 0.019397731870412827,
+ "learning_rate": 0.0006,
+ "loss": 4.183379173278809,
+ "step": 6248
+ },
+ {
+ "epoch": 86.79650655021834,
+ "grad_norm": 0.017525402829051018,
+ "learning_rate": 0.0006,
+ "loss": 4.1494245529174805,
+ "step": 6249
+ },
+ {
+ "epoch": 86.81048034934499,
+ "grad_norm": 0.016310200095176697,
+ "learning_rate": 0.0006,
+ "loss": 4.149104118347168,
+ "step": 6250
+ },
+ {
+ "epoch": 86.82445414847162,
+ "grad_norm": 0.01506423857063055,
+ "learning_rate": 0.0006,
+ "loss": 4.22091817855835,
+ "step": 6251
+ },
+ {
+ "epoch": 86.83842794759825,
+ "grad_norm": 0.014200284145772457,
+ "learning_rate": 0.0006,
+ "loss": 4.177569389343262,
+ "step": 6252
+ },
+ {
+ "epoch": 86.8524017467249,
+ "grad_norm": 0.015016552060842514,
+ "learning_rate": 0.0006,
+ "loss": 4.089859962463379,
+ "step": 6253
+ },
+ {
+ "epoch": 86.86637554585153,
+ "grad_norm": 0.016054971143603325,
+ "learning_rate": 0.0006,
+ "loss": 4.157166004180908,
+ "step": 6254
+ },
+ {
+ "epoch": 86.88034934497817,
+ "grad_norm": 0.014388521201908588,
+ "learning_rate": 0.0006,
+ "loss": 4.1437578201293945,
+ "step": 6255
+ },
+ {
+ "epoch": 86.8943231441048,
+ "grad_norm": 0.015705369412899017,
+ "learning_rate": 0.0006,
+ "loss": 4.045689582824707,
+ "step": 6256
+ },
+ {
+ "epoch": 86.90829694323143,
+ "grad_norm": 0.014192395843565464,
+ "learning_rate": 0.0006,
+ "loss": 4.098793029785156,
+ "step": 6257
+ },
+ {
+ "epoch": 86.92227074235808,
+ "grad_norm": 0.013751034624874592,
+ "learning_rate": 0.0006,
+ "loss": 3.989920139312744,
+ "step": 6258
+ },
+ {
+ "epoch": 86.93624454148471,
+ "grad_norm": 0.01582193188369274,
+ "learning_rate": 0.0006,
+ "loss": 4.283449172973633,
+ "step": 6259
+ },
+ {
+ "epoch": 86.95021834061136,
+ "grad_norm": 0.01476013008505106,
+ "learning_rate": 0.0006,
+ "loss": 4.221401691436768,
+ "step": 6260
+ },
+ {
+ "epoch": 86.96419213973799,
+ "grad_norm": 0.01346337515860796,
+ "learning_rate": 0.0006,
+ "loss": 4.05896520614624,
+ "step": 6261
+ },
+ {
+ "epoch": 86.97816593886463,
+ "grad_norm": 0.012899307534098625,
+ "learning_rate": 0.0006,
+ "loss": 4.159560203552246,
+ "step": 6262
+ },
+ {
+ "epoch": 86.99213973799127,
+ "grad_norm": 0.013150574639439583,
+ "learning_rate": 0.0006,
+ "loss": 4.217330455780029,
+ "step": 6263
+ },
+ {
+ "epoch": 87.0,
+ "grad_norm": 0.014940370805561543,
+ "learning_rate": 0.0006,
+ "loss": 4.1611833572387695,
+ "step": 6264
+ },
+ {
+ "epoch": 87.0,
+ "eval_loss": 4.666043758392334,
+ "eval_runtime": 56.5978,
+ "eval_samples_per_second": 43.147,
+ "eval_steps_per_second": 1.36,
+ "step": 6264
+ },
+ {
+ "epoch": 87.01397379912663,
+ "grad_norm": 0.013721180148422718,
+ "learning_rate": 0.0006,
+ "loss": 4.142454147338867,
+ "step": 6265
+ },
+ {
+ "epoch": 87.02794759825328,
+ "grad_norm": 0.014805255457758904,
+ "learning_rate": 0.0006,
+ "loss": 4.121677875518799,
+ "step": 6266
+ },
+ {
+ "epoch": 87.04192139737991,
+ "grad_norm": 0.015576514415442944,
+ "learning_rate": 0.0006,
+ "loss": 4.1188435554504395,
+ "step": 6267
+ },
+ {
+ "epoch": 87.05589519650655,
+ "grad_norm": 0.01605502888560295,
+ "learning_rate": 0.0006,
+ "loss": 4.255428314208984,
+ "step": 6268
+ },
+ {
+ "epoch": 87.06986899563319,
+ "grad_norm": 0.014933343045413494,
+ "learning_rate": 0.0006,
+ "loss": 4.122596740722656,
+ "step": 6269
+ },
+ {
+ "epoch": 87.08384279475983,
+ "grad_norm": 0.014630894176661968,
+ "learning_rate": 0.0006,
+ "loss": 4.148405075073242,
+ "step": 6270
+ },
+ {
+ "epoch": 87.09781659388646,
+ "grad_norm": 0.014211163856089115,
+ "learning_rate": 0.0006,
+ "loss": 4.058030128479004,
+ "step": 6271
+ },
+ {
+ "epoch": 87.1117903930131,
+ "grad_norm": 0.014060168527066708,
+ "learning_rate": 0.0006,
+ "loss": 4.146586894989014,
+ "step": 6272
+ },
+ {
+ "epoch": 87.12576419213974,
+ "grad_norm": 0.014740100130438805,
+ "learning_rate": 0.0006,
+ "loss": 4.077739715576172,
+ "step": 6273
+ },
+ {
+ "epoch": 87.13973799126637,
+ "grad_norm": 0.013970664702355862,
+ "learning_rate": 0.0006,
+ "loss": 4.038252830505371,
+ "step": 6274
+ },
+ {
+ "epoch": 87.15371179039302,
+ "grad_norm": 0.014399951323866844,
+ "learning_rate": 0.0006,
+ "loss": 4.143700122833252,
+ "step": 6275
+ },
+ {
+ "epoch": 87.16768558951965,
+ "grad_norm": 0.01496410183608532,
+ "learning_rate": 0.0006,
+ "loss": 4.105609893798828,
+ "step": 6276
+ },
+ {
+ "epoch": 87.18165938864628,
+ "grad_norm": 0.014287355355918407,
+ "learning_rate": 0.0006,
+ "loss": 4.173645973205566,
+ "step": 6277
+ },
+ {
+ "epoch": 87.19563318777293,
+ "grad_norm": 0.0141292130574584,
+ "learning_rate": 0.0006,
+ "loss": 4.068511486053467,
+ "step": 6278
+ },
+ {
+ "epoch": 87.20960698689956,
+ "grad_norm": 0.01393523346632719,
+ "learning_rate": 0.0006,
+ "loss": 4.060687065124512,
+ "step": 6279
+ },
+ {
+ "epoch": 87.2235807860262,
+ "grad_norm": 0.01431021187454462,
+ "learning_rate": 0.0006,
+ "loss": 4.219209671020508,
+ "step": 6280
+ },
+ {
+ "epoch": 87.23755458515284,
+ "grad_norm": 0.014856027439236641,
+ "learning_rate": 0.0006,
+ "loss": 4.146749973297119,
+ "step": 6281
+ },
+ {
+ "epoch": 87.25152838427948,
+ "grad_norm": 0.013677539303898811,
+ "learning_rate": 0.0006,
+ "loss": 3.9859797954559326,
+ "step": 6282
+ },
+ {
+ "epoch": 87.26550218340611,
+ "grad_norm": 0.014078967273235321,
+ "learning_rate": 0.0006,
+ "loss": 4.129668235778809,
+ "step": 6283
+ },
+ {
+ "epoch": 87.27947598253274,
+ "grad_norm": 0.015564394183456898,
+ "learning_rate": 0.0006,
+ "loss": 4.0007829666137695,
+ "step": 6284
+ },
+ {
+ "epoch": 87.29344978165939,
+ "grad_norm": 0.015077749267220497,
+ "learning_rate": 0.0006,
+ "loss": 4.112264156341553,
+ "step": 6285
+ },
+ {
+ "epoch": 87.30742358078602,
+ "grad_norm": 0.015341218560934067,
+ "learning_rate": 0.0006,
+ "loss": 4.112249374389648,
+ "step": 6286
+ },
+ {
+ "epoch": 87.32139737991267,
+ "grad_norm": 0.016399366781115532,
+ "learning_rate": 0.0006,
+ "loss": 4.176196098327637,
+ "step": 6287
+ },
+ {
+ "epoch": 87.3353711790393,
+ "grad_norm": 0.01599763333797455,
+ "learning_rate": 0.0006,
+ "loss": 4.163441181182861,
+ "step": 6288
+ },
+ {
+ "epoch": 87.34934497816595,
+ "grad_norm": 0.015439776703715324,
+ "learning_rate": 0.0006,
+ "loss": 4.11902379989624,
+ "step": 6289
+ },
+ {
+ "epoch": 87.36331877729258,
+ "grad_norm": 0.015909943729639053,
+ "learning_rate": 0.0006,
+ "loss": 4.16621732711792,
+ "step": 6290
+ },
+ {
+ "epoch": 87.37729257641921,
+ "grad_norm": 0.016710078343749046,
+ "learning_rate": 0.0006,
+ "loss": 4.044960975646973,
+ "step": 6291
+ },
+ {
+ "epoch": 87.39126637554585,
+ "grad_norm": 0.015883933752775192,
+ "learning_rate": 0.0006,
+ "loss": 4.065184593200684,
+ "step": 6292
+ },
+ {
+ "epoch": 87.40524017467249,
+ "grad_norm": 0.013912747614085674,
+ "learning_rate": 0.0006,
+ "loss": 4.102450370788574,
+ "step": 6293
+ },
+ {
+ "epoch": 87.41921397379913,
+ "grad_norm": 0.013170420192182064,
+ "learning_rate": 0.0006,
+ "loss": 4.234364032745361,
+ "step": 6294
+ },
+ {
+ "epoch": 87.43318777292576,
+ "grad_norm": 0.014043302275240421,
+ "learning_rate": 0.0006,
+ "loss": 4.129565238952637,
+ "step": 6295
+ },
+ {
+ "epoch": 87.44716157205241,
+ "grad_norm": 0.015010139904916286,
+ "learning_rate": 0.0006,
+ "loss": 4.166678428649902,
+ "step": 6296
+ },
+ {
+ "epoch": 87.46113537117904,
+ "grad_norm": 0.014721784740686417,
+ "learning_rate": 0.0006,
+ "loss": 4.182618618011475,
+ "step": 6297
+ },
+ {
+ "epoch": 87.47510917030567,
+ "grad_norm": 0.015713468194007874,
+ "learning_rate": 0.0006,
+ "loss": 4.103855133056641,
+ "step": 6298
+ },
+ {
+ "epoch": 87.48908296943232,
+ "grad_norm": 0.017204247415065765,
+ "learning_rate": 0.0006,
+ "loss": 4.140937328338623,
+ "step": 6299
+ },
+ {
+ "epoch": 87.50305676855895,
+ "grad_norm": 0.016856668516993523,
+ "learning_rate": 0.0006,
+ "loss": 4.147859573364258,
+ "step": 6300
+ },
+ {
+ "epoch": 87.5170305676856,
+ "grad_norm": 0.017205018550157547,
+ "learning_rate": 0.0006,
+ "loss": 4.1667327880859375,
+ "step": 6301
+ },
+ {
+ "epoch": 87.53100436681223,
+ "grad_norm": 0.014821539632976055,
+ "learning_rate": 0.0006,
+ "loss": 4.089935302734375,
+ "step": 6302
+ },
+ {
+ "epoch": 87.54497816593886,
+ "grad_norm": 0.01446484588086605,
+ "learning_rate": 0.0006,
+ "loss": 4.243657112121582,
+ "step": 6303
+ },
+ {
+ "epoch": 87.5589519650655,
+ "grad_norm": 0.014374484308063984,
+ "learning_rate": 0.0006,
+ "loss": 4.14356803894043,
+ "step": 6304
+ },
+ {
+ "epoch": 87.57292576419214,
+ "grad_norm": 0.01420109998434782,
+ "learning_rate": 0.0006,
+ "loss": 4.1273651123046875,
+ "step": 6305
+ },
+ {
+ "epoch": 87.58689956331878,
+ "grad_norm": 0.014810388907790184,
+ "learning_rate": 0.0006,
+ "loss": 4.222290515899658,
+ "step": 6306
+ },
+ {
+ "epoch": 87.60087336244541,
+ "grad_norm": 0.015826983377337456,
+ "learning_rate": 0.0006,
+ "loss": 4.137996196746826,
+ "step": 6307
+ },
+ {
+ "epoch": 87.61484716157206,
+ "grad_norm": 0.014457454904913902,
+ "learning_rate": 0.0006,
+ "loss": 4.0780744552612305,
+ "step": 6308
+ },
+ {
+ "epoch": 87.62882096069869,
+ "grad_norm": 0.014458604156970978,
+ "learning_rate": 0.0006,
+ "loss": 4.1287922859191895,
+ "step": 6309
+ },
+ {
+ "epoch": 87.64279475982532,
+ "grad_norm": 0.015282071195542812,
+ "learning_rate": 0.0006,
+ "loss": 4.189508438110352,
+ "step": 6310
+ },
+ {
+ "epoch": 87.65676855895197,
+ "grad_norm": 0.017536642029881477,
+ "learning_rate": 0.0006,
+ "loss": 4.070173263549805,
+ "step": 6311
+ },
+ {
+ "epoch": 87.6707423580786,
+ "grad_norm": 0.017303580418229103,
+ "learning_rate": 0.0006,
+ "loss": 4.119133949279785,
+ "step": 6312
+ },
+ {
+ "epoch": 87.68471615720524,
+ "grad_norm": 0.016148580238223076,
+ "learning_rate": 0.0006,
+ "loss": 3.994354486465454,
+ "step": 6313
+ },
+ {
+ "epoch": 87.69868995633188,
+ "grad_norm": 0.01706942543387413,
+ "learning_rate": 0.0006,
+ "loss": 4.183113098144531,
+ "step": 6314
+ },
+ {
+ "epoch": 87.71266375545852,
+ "grad_norm": 0.016971010714769363,
+ "learning_rate": 0.0006,
+ "loss": 4.104086875915527,
+ "step": 6315
+ },
+ {
+ "epoch": 87.72663755458515,
+ "grad_norm": 0.015516847372055054,
+ "learning_rate": 0.0006,
+ "loss": 4.2409772872924805,
+ "step": 6316
+ },
+ {
+ "epoch": 87.74061135371178,
+ "grad_norm": 0.015352269634604454,
+ "learning_rate": 0.0006,
+ "loss": 4.026427268981934,
+ "step": 6317
+ },
+ {
+ "epoch": 87.75458515283843,
+ "grad_norm": 0.017702415585517883,
+ "learning_rate": 0.0006,
+ "loss": 4.177840232849121,
+ "step": 6318
+ },
+ {
+ "epoch": 87.76855895196506,
+ "grad_norm": 0.01629541628062725,
+ "learning_rate": 0.0006,
+ "loss": 4.229039192199707,
+ "step": 6319
+ },
+ {
+ "epoch": 87.78253275109171,
+ "grad_norm": 0.01503711100667715,
+ "learning_rate": 0.0006,
+ "loss": 4.214350700378418,
+ "step": 6320
+ },
+ {
+ "epoch": 87.79650655021834,
+ "grad_norm": 0.014913883060216904,
+ "learning_rate": 0.0006,
+ "loss": 4.168880939483643,
+ "step": 6321
+ },
+ {
+ "epoch": 87.81048034934499,
+ "grad_norm": 0.014513909816741943,
+ "learning_rate": 0.0006,
+ "loss": 4.162267684936523,
+ "step": 6322
+ },
+ {
+ "epoch": 87.82445414847162,
+ "grad_norm": 0.014635873027145863,
+ "learning_rate": 0.0006,
+ "loss": 4.220067977905273,
+ "step": 6323
+ },
+ {
+ "epoch": 87.83842794759825,
+ "grad_norm": 0.017102796584367752,
+ "learning_rate": 0.0006,
+ "loss": 4.1493377685546875,
+ "step": 6324
+ },
+ {
+ "epoch": 87.8524017467249,
+ "grad_norm": 0.015533884055912495,
+ "learning_rate": 0.0006,
+ "loss": 4.0646209716796875,
+ "step": 6325
+ },
+ {
+ "epoch": 87.86637554585153,
+ "grad_norm": 0.014617550186812878,
+ "learning_rate": 0.0006,
+ "loss": 4.1390533447265625,
+ "step": 6326
+ },
+ {
+ "epoch": 87.88034934497817,
+ "grad_norm": 0.014142967760562897,
+ "learning_rate": 0.0006,
+ "loss": 4.133934020996094,
+ "step": 6327
+ },
+ {
+ "epoch": 87.8943231441048,
+ "grad_norm": 0.013182433322072029,
+ "learning_rate": 0.0006,
+ "loss": 4.125327110290527,
+ "step": 6328
+ },
+ {
+ "epoch": 87.90829694323143,
+ "grad_norm": 0.013928229920566082,
+ "learning_rate": 0.0006,
+ "loss": 4.142777442932129,
+ "step": 6329
+ },
+ {
+ "epoch": 87.92227074235808,
+ "grad_norm": 0.015337013639509678,
+ "learning_rate": 0.0006,
+ "loss": 4.140210151672363,
+ "step": 6330
+ },
+ {
+ "epoch": 87.93624454148471,
+ "grad_norm": 0.014873536303639412,
+ "learning_rate": 0.0006,
+ "loss": 4.102636337280273,
+ "step": 6331
+ },
+ {
+ "epoch": 87.95021834061136,
+ "grad_norm": 0.01387431938201189,
+ "learning_rate": 0.0006,
+ "loss": 4.149059295654297,
+ "step": 6332
+ },
+ {
+ "epoch": 87.96419213973799,
+ "grad_norm": 0.013234943151473999,
+ "learning_rate": 0.0006,
+ "loss": 4.1564459800720215,
+ "step": 6333
+ },
+ {
+ "epoch": 87.97816593886463,
+ "grad_norm": 0.013781017623841763,
+ "learning_rate": 0.0006,
+ "loss": 4.189839839935303,
+ "step": 6334
+ },
+ {
+ "epoch": 87.99213973799127,
+ "grad_norm": 0.013733026571571827,
+ "learning_rate": 0.0006,
+ "loss": 4.29277229309082,
+ "step": 6335
+ },
+ {
+ "epoch": 88.0,
+ "grad_norm": 0.014752327464520931,
+ "learning_rate": 0.0006,
+ "loss": 4.163867950439453,
+ "step": 6336
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 28800,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 400,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": true,
+ "should_log": false,
+ "should_save": false,
+ "should_training_stop": false
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 2.701259757972357e+18,
+ "train_batch_size": 8,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6336/training_args.bin b/runs/l2r50-i2-fulle-lm/checkpoint-6336/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..cad232831c222cc7f8b938b8bbe42f61e7d89c3a
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-6336/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:f7c40e3bfa2f5ac1450ca1243e3f2a895d44a06ed5995e6aefd97f78d229d8c8
+size 4792
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6624/chat_template.jinja b/runs/l2r50-i2-fulle-lm/checkpoint-6624/chat_template.jinja
new file mode 100644
index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-6624/chat_template.jinja
@@ -0,0 +1,85 @@
+{%- if tools %}
+ {{- '<|im_start|>system\n' }}
+ {%- if messages[0].role == 'system' %}
+ {{- messages[0].content + '\n\n' }}
+ {%- endif %}
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }}
+ {%- for tool in tools %}
+ {{- "\n" }}
+ {{- tool | tojson }}
+ {%- endfor %}
+ {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }}
+{%- else %}
+ {%- if messages[0].role == 'system' %}
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
+ {%- endif %}
+{%- endif %}
+{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
+{%- for message in messages[::-1] %}
+ {%- set index = (messages|length - 1) - loop.index0 %}
+ {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %}
+ {%- set ns.multi_step_tool = false %}
+ {%- set ns.last_query_index = index %}
+ {%- endif %}
+{%- endfor %}
+{%- for message in messages %}
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
+ {%- elif message.role == "assistant" %}
+ {%- set content = message.content %}
+ {%- set reasoning_content = '' %}
+ {%- if message.reasoning_content is defined and message.reasoning_content is not none %}
+ {%- set reasoning_content = message.reasoning_content %}
+ {%- else %}
+ {%- if '' in message.content %}
+ {%- set content = message.content.split('')[-1].lstrip('\n') %}
+ {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %}
+ {%- endif %}
+ {%- endif %}
+ {%- if loop.index0 > ns.last_query_index %}
+ {%- if loop.last or (not loop.last and reasoning_content) %}
+ {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- if message.tool_calls %}
+ {%- for tool_call in message.tool_calls %}
+ {%- if (loop.first and content) or (not loop.first) %}
+ {{- '\n' }}
+ {%- endif %}
+ {%- if tool_call.function %}
+ {%- set tool_call = tool_call.function %}
+ {%- endif %}
+ {{- '\n{"name": "' }}
+ {{- tool_call.name }}
+ {{- '", "arguments": ' }}
+ {%- if tool_call.arguments is string %}
+ {{- tool_call.arguments }}
+ {%- else %}
+ {{- tool_call.arguments | tojson }}
+ {%- endif %}
+ {{- '}\n' }}
+ {%- endfor %}
+ {%- endif %}
+ {{- '<|im_end|>\n' }}
+ {%- elif message.role == "tool" %}
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
+ {{- '<|im_start|>user' }}
+ {%- endif %}
+ {{- '\n\n' }}
+ {{- message.content }}
+ {{- '\n' }}
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
+ {{- '<|im_end|>\n' }}
+ {%- endif %}
+ {%- endif %}
+{%- endfor %}
+{%- if add_generation_prompt %}
+ {{- '<|im_start|>assistant\n' }}
+ {%- if enable_thinking is defined and enable_thinking is false %}
+ {{- '\n\n\n\n' }}
+ {%- endif %}
+{%- endif %}
\ No newline at end of file
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6624/config.json b/runs/l2r50-i2-fulle-lm/checkpoint-6624/config.json
new file mode 100644
index 0000000000000000000000000000000000000000..22f250953c6ce8fc91ee4411856dd73041a415e2
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-6624/config.json
@@ -0,0 +1,32 @@
+{
+ "architectures": [
+ "LlamaForCausalLM"
+ ],
+ "attention_bias": false,
+ "attention_dropout": 0.0,
+ "bos_token_id": null,
+ "dtype": "float32",
+ "eos_token_id": 151645,
+ "head_dim": 128,
+ "hidden_act": "silu",
+ "hidden_size": 512,
+ "initializer_range": 0.02,
+ "intermediate_size": 1536,
+ "max_position_embeddings": 2048,
+ "mlp_bias": false,
+ "model_type": "llama",
+ "num_attention_heads": 4,
+ "num_hidden_layers": 20,
+ "num_key_value_heads": 4,
+ "pad_token_id": 151645,
+ "pretraining_tp": 1,
+ "rms_norm_eps": 1e-06,
+ "rope_parameters": {
+ "rope_theta": 10000.0,
+ "rope_type": "default"
+ },
+ "tie_word_embeddings": true,
+ "transformers_version": "5.5.0",
+ "use_cache": false,
+ "vocab_size": 151673
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6624/generation_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-6624/generation_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-6624/generation_config.json
@@ -0,0 +1,11 @@
+{
+ "_from_model_config": true,
+ "eos_token_id": [
+ 151645
+ ],
+ "output_attentions": false,
+ "output_hidden_states": false,
+ "pad_token_id": 151645,
+ "transformers_version": "5.5.0",
+ "use_cache": true
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6624/model.safetensors b/runs/l2r50-i2-fulle-lm/checkpoint-6624/model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..f487010d7fbcb1b812acb3764df63a019459ab80
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-6624/model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:49a736af76c89dc83131d9cfcbb9f97ed70927d70d2fe82dbda55116bc87ddaf
+size 583360328
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6624/optimizer.pt b/runs/l2r50-i2-fulle-lm/checkpoint-6624/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..0d2cba0c5c5c417ed400c491752c71740d1963da
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-6624/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:e175ac6cb4d3df111ed04fca76ecb917ea42b452481a30c602abbc84283e744e
+size 1166833530
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6624/rng_state_0.pth b/runs/l2r50-i2-fulle-lm/checkpoint-6624/rng_state_0.pth
new file mode 100644
index 0000000000000000000000000000000000000000..f7893f89b98ec10f03213b6e8beac3fa8d5e4926
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-6624/rng_state_0.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:c6d5bec1a5bb83a83aa97513fe583f8fc14f4840be26666e8cd7efa07536abe1
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6624/rng_state_1.pth b/runs/l2r50-i2-fulle-lm/checkpoint-6624/rng_state_1.pth
new file mode 100644
index 0000000000000000000000000000000000000000..33aa04c3b93e992e98a21fd16e67f87011843073
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-6624/rng_state_1.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:bb8e96dd1435044d27f9f8cbcf84f8dd6c8f0f0749cf034035894986b2419766
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6624/rng_state_2.pth b/runs/l2r50-i2-fulle-lm/checkpoint-6624/rng_state_2.pth
new file mode 100644
index 0000000000000000000000000000000000000000..8fff922a7792a11b3253e8dd288ca2c009b12152
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-6624/rng_state_2.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:ff27d8e23cd1e341ec2c1a3a22e887cd6b3e72b202f7c20aac714e64b8ad1825
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6624/rng_state_3.pth b/runs/l2r50-i2-fulle-lm/checkpoint-6624/rng_state_3.pth
new file mode 100644
index 0000000000000000000000000000000000000000..9434f1eaa62b162a1f26d844c9cefc8483a39897
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-6624/rng_state_3.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:4af1228878c5edb039eb391166ff4f3d6973c92e66fb7cd96e3fb2e64d0a2b3e
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6624/scheduler.pt b/runs/l2r50-i2-fulle-lm/checkpoint-6624/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..1fb4fbb9cb8eafb61c9d89947d329fb67f4dd4e6
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-6624/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:aeab2b26cd983e17d83b05e83f37ef1706928049bbb8ff0e989945b479e25df2
+size 1064
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6624/tokenizer.json b/runs/l2r50-i2-fulle-lm/checkpoint-6624/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..53786016f070e11184813f3c83101dc5e93e392e
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-6624/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:c3c63fda1646a555448e8f03662542e04b11d103721319e4d9b267d7da03ceb8
+size 11424463
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6624/tokenizer_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-6624/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..98e2fb9b2cc54785dc9d871b8e503f8b8f2defe7
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-6624/tokenizer_config.json
@@ -0,0 +1,20 @@
+{
+ "add_prefix_space": false,
+ "backend": "tokenizers",
+ "bos_token": null,
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "<|im_end|>",
+ "errors": "replace",
+ "extra_special_tokens": [
+ "<|l2r_pred|>",
+ "<|r2l_pred|>",
+ "<|next_1_pred|>",
+ "<|next_2_pred|>"
+ ],
+ "is_local": false,
+ "model_max_length": 131072,
+ "pad_token": "<|im_end|>",
+ "split_special_tokens": false,
+ "tokenizer_class": "Qwen2Tokenizer",
+ "unk_token": null
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6624/trainer_state.json b/runs/l2r50-i2-fulle-lm/checkpoint-6624/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..30d8cf60cffd4ada8c2c7a0e02e68accd0abe0d1
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-6624/trainer_state.json
@@ -0,0 +1,47130 @@
+{
+ "best_global_step": null,
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 92.0,
+ "eval_steps": 500,
+ "global_step": 6624,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.013973799126637555,
+ "grad_norm": 0.1739506870508194,
+ "learning_rate": 0.0,
+ "loss": 11.99299430847168,
+ "step": 1
+ },
+ {
+ "epoch": 0.02794759825327511,
+ "grad_norm": 0.17216597497463226,
+ "learning_rate": 5.999999999999999e-06,
+ "loss": 11.988704681396484,
+ "step": 2
+ },
+ {
+ "epoch": 0.04192139737991266,
+ "grad_norm": 0.17459715902805328,
+ "learning_rate": 1.1999999999999999e-05,
+ "loss": 11.945581436157227,
+ "step": 3
+ },
+ {
+ "epoch": 0.05589519650655022,
+ "grad_norm": 0.17577414214611053,
+ "learning_rate": 1.7999999999999997e-05,
+ "loss": 11.859772682189941,
+ "step": 4
+ },
+ {
+ "epoch": 0.06986899563318777,
+ "grad_norm": 0.16163651645183563,
+ "learning_rate": 2.3999999999999997e-05,
+ "loss": 11.744415283203125,
+ "step": 5
+ },
+ {
+ "epoch": 0.08384279475982533,
+ "grad_norm": 0.14409932494163513,
+ "learning_rate": 2.9999999999999997e-05,
+ "loss": 11.618099212646484,
+ "step": 6
+ },
+ {
+ "epoch": 0.09781659388646288,
+ "grad_norm": 0.13079848885536194,
+ "learning_rate": 3.5999999999999994e-05,
+ "loss": 11.505935668945312,
+ "step": 7
+ },
+ {
+ "epoch": 0.11179039301310044,
+ "grad_norm": 0.11964613944292068,
+ "learning_rate": 4.2e-05,
+ "loss": 11.416532516479492,
+ "step": 8
+ },
+ {
+ "epoch": 0.125764192139738,
+ "grad_norm": 0.11345142871141434,
+ "learning_rate": 4.7999999999999994e-05,
+ "loss": 11.33206558227539,
+ "step": 9
+ },
+ {
+ "epoch": 0.13973799126637554,
+ "grad_norm": 0.11033952981233597,
+ "learning_rate": 5.399999999999999e-05,
+ "loss": 11.259052276611328,
+ "step": 10
+ },
+ {
+ "epoch": 0.1537117903930131,
+ "grad_norm": 0.10776489973068237,
+ "learning_rate": 5.9999999999999995e-05,
+ "loss": 11.203384399414062,
+ "step": 11
+ },
+ {
+ "epoch": 0.16768558951965065,
+ "grad_norm": 0.10721946507692337,
+ "learning_rate": 6.599999999999999e-05,
+ "loss": 11.153295516967773,
+ "step": 12
+ },
+ {
+ "epoch": 0.18165938864628822,
+ "grad_norm": 0.10673430562019348,
+ "learning_rate": 7.199999999999999e-05,
+ "loss": 11.110200881958008,
+ "step": 13
+ },
+ {
+ "epoch": 0.19563318777292577,
+ "grad_norm": 0.10630254447460175,
+ "learning_rate": 7.8e-05,
+ "loss": 11.067477226257324,
+ "step": 14
+ },
+ {
+ "epoch": 0.2096069868995633,
+ "grad_norm": 0.10631615668535233,
+ "learning_rate": 8.4e-05,
+ "loss": 11.022941589355469,
+ "step": 15
+ },
+ {
+ "epoch": 0.22358078602620088,
+ "grad_norm": 0.10609443485736847,
+ "learning_rate": 8.999999999999999e-05,
+ "loss": 10.970746040344238,
+ "step": 16
+ },
+ {
+ "epoch": 0.23755458515283842,
+ "grad_norm": 0.10648232698440552,
+ "learning_rate": 9.599999999999999e-05,
+ "loss": 10.91267204284668,
+ "step": 17
+ },
+ {
+ "epoch": 0.251528384279476,
+ "grad_norm": 0.10584589838981628,
+ "learning_rate": 0.000102,
+ "loss": 10.854547500610352,
+ "step": 18
+ },
+ {
+ "epoch": 0.26550218340611353,
+ "grad_norm": 0.10591720044612885,
+ "learning_rate": 0.00010799999999999998,
+ "loss": 10.788917541503906,
+ "step": 19
+ },
+ {
+ "epoch": 0.2794759825327511,
+ "grad_norm": 0.10611672699451447,
+ "learning_rate": 0.00011399999999999999,
+ "loss": 10.717556953430176,
+ "step": 20
+ },
+ {
+ "epoch": 0.2934497816593886,
+ "grad_norm": 0.10455359518527985,
+ "learning_rate": 0.00011999999999999999,
+ "loss": 10.650186538696289,
+ "step": 21
+ },
+ {
+ "epoch": 0.3074235807860262,
+ "grad_norm": 0.1047525629401207,
+ "learning_rate": 0.00012599999999999997,
+ "loss": 10.56808090209961,
+ "step": 22
+ },
+ {
+ "epoch": 0.32139737991266376,
+ "grad_norm": 0.10511605441570282,
+ "learning_rate": 0.00013199999999999998,
+ "loss": 10.484028816223145,
+ "step": 23
+ },
+ {
+ "epoch": 0.3353711790393013,
+ "grad_norm": 0.10467177629470825,
+ "learning_rate": 0.000138,
+ "loss": 10.399947166442871,
+ "step": 24
+ },
+ {
+ "epoch": 0.34934497816593885,
+ "grad_norm": 0.10432136058807373,
+ "learning_rate": 0.00014399999999999998,
+ "loss": 10.312714576721191,
+ "step": 25
+ },
+ {
+ "epoch": 0.36331877729257644,
+ "grad_norm": 0.10527262091636658,
+ "learning_rate": 0.00015,
+ "loss": 10.212261199951172,
+ "step": 26
+ },
+ {
+ "epoch": 0.377292576419214,
+ "grad_norm": 0.103610560297966,
+ "learning_rate": 0.000156,
+ "loss": 10.132034301757812,
+ "step": 27
+ },
+ {
+ "epoch": 0.39126637554585153,
+ "grad_norm": 0.10553059726953506,
+ "learning_rate": 0.000162,
+ "loss": 10.020434379577637,
+ "step": 28
+ },
+ {
+ "epoch": 0.4052401746724891,
+ "grad_norm": 0.1037299782037735,
+ "learning_rate": 0.000168,
+ "loss": 9.934466361999512,
+ "step": 29
+ },
+ {
+ "epoch": 0.4192139737991266,
+ "grad_norm": 0.1042739674448967,
+ "learning_rate": 0.00017399999999999997,
+ "loss": 9.822539329528809,
+ "step": 30
+ },
+ {
+ "epoch": 0.4331877729257642,
+ "grad_norm": 0.10300865024328232,
+ "learning_rate": 0.00017999999999999998,
+ "loss": 9.734994888305664,
+ "step": 31
+ },
+ {
+ "epoch": 0.44716157205240176,
+ "grad_norm": 0.10291654616594315,
+ "learning_rate": 0.000186,
+ "loss": 9.624505996704102,
+ "step": 32
+ },
+ {
+ "epoch": 0.4611353711790393,
+ "grad_norm": 0.10210851579904556,
+ "learning_rate": 0.00019199999999999998,
+ "loss": 9.518205642700195,
+ "step": 33
+ },
+ {
+ "epoch": 0.47510917030567684,
+ "grad_norm": 0.10222529619932175,
+ "learning_rate": 0.000198,
+ "loss": 9.412222862243652,
+ "step": 34
+ },
+ {
+ "epoch": 0.4890829694323144,
+ "grad_norm": 0.09992019087076187,
+ "learning_rate": 0.000204,
+ "loss": 9.326566696166992,
+ "step": 35
+ },
+ {
+ "epoch": 0.503056768558952,
+ "grad_norm": 0.09931084513664246,
+ "learning_rate": 0.00020999999999999998,
+ "loss": 9.210111618041992,
+ "step": 36
+ },
+ {
+ "epoch": 0.5170305676855895,
+ "grad_norm": 0.09928256273269653,
+ "learning_rate": 0.00021599999999999996,
+ "loss": 9.100493431091309,
+ "step": 37
+ },
+ {
+ "epoch": 0.5310043668122271,
+ "grad_norm": 0.09843816608190536,
+ "learning_rate": 0.00022199999999999998,
+ "loss": 8.996543884277344,
+ "step": 38
+ },
+ {
+ "epoch": 0.5449781659388646,
+ "grad_norm": 0.09825374186038971,
+ "learning_rate": 0.00022799999999999999,
+ "loss": 8.888991355895996,
+ "step": 39
+ },
+ {
+ "epoch": 0.5589519650655022,
+ "grad_norm": 0.09602385759353638,
+ "learning_rate": 0.000234,
+ "loss": 8.802907943725586,
+ "step": 40
+ },
+ {
+ "epoch": 0.5729257641921397,
+ "grad_norm": 0.09414064139127731,
+ "learning_rate": 0.00023999999999999998,
+ "loss": 8.715423583984375,
+ "step": 41
+ },
+ {
+ "epoch": 0.5868995633187772,
+ "grad_norm": 0.09273893386125565,
+ "learning_rate": 0.00024599999999999996,
+ "loss": 8.61374282836914,
+ "step": 42
+ },
+ {
+ "epoch": 0.6008733624454149,
+ "grad_norm": 0.09079599380493164,
+ "learning_rate": 0.00025199999999999995,
+ "loss": 8.502904891967773,
+ "step": 43
+ },
+ {
+ "epoch": 0.6148471615720524,
+ "grad_norm": 0.08605825901031494,
+ "learning_rate": 0.000258,
+ "loss": 8.453008651733398,
+ "step": 44
+ },
+ {
+ "epoch": 0.62882096069869,
+ "grad_norm": 0.08338996022939682,
+ "learning_rate": 0.00026399999999999997,
+ "loss": 8.365163803100586,
+ "step": 45
+ },
+ {
+ "epoch": 0.6427947598253275,
+ "grad_norm": 0.08150867372751236,
+ "learning_rate": 0.00027,
+ "loss": 8.285372734069824,
+ "step": 46
+ },
+ {
+ "epoch": 0.6567685589519651,
+ "grad_norm": 0.07794807851314545,
+ "learning_rate": 0.000276,
+ "loss": 8.185811996459961,
+ "step": 47
+ },
+ {
+ "epoch": 0.6707423580786026,
+ "grad_norm": 0.07592720538377762,
+ "learning_rate": 0.00028199999999999997,
+ "loss": 8.098482131958008,
+ "step": 48
+ },
+ {
+ "epoch": 0.6847161572052402,
+ "grad_norm": 0.0709897130727768,
+ "learning_rate": 0.00028799999999999995,
+ "loss": 8.042091369628906,
+ "step": 49
+ },
+ {
+ "epoch": 0.6986899563318777,
+ "grad_norm": 0.06582305580377579,
+ "learning_rate": 0.000294,
+ "loss": 7.972844123840332,
+ "step": 50
+ },
+ {
+ "epoch": 0.7126637554585152,
+ "grad_norm": 0.059452351182699203,
+ "learning_rate": 0.0003,
+ "loss": 7.924579620361328,
+ "step": 51
+ },
+ {
+ "epoch": 0.7266375545851529,
+ "grad_norm": 0.052426718175411224,
+ "learning_rate": 0.00030599999999999996,
+ "loss": 7.876336097717285,
+ "step": 52
+ },
+ {
+ "epoch": 0.7406113537117904,
+ "grad_norm": 0.04768150672316551,
+ "learning_rate": 0.000312,
+ "loss": 7.8102707862854,
+ "step": 53
+ },
+ {
+ "epoch": 0.754585152838428,
+ "grad_norm": 0.03970075398683548,
+ "learning_rate": 0.000318,
+ "loss": 7.791913986206055,
+ "step": 54
+ },
+ {
+ "epoch": 0.7685589519650655,
+ "grad_norm": 0.03559752553701401,
+ "learning_rate": 0.000324,
+ "loss": 7.756960868835449,
+ "step": 55
+ },
+ {
+ "epoch": 0.7825327510917031,
+ "grad_norm": 0.029321353882551193,
+ "learning_rate": 0.00033,
+ "loss": 7.724878787994385,
+ "step": 56
+ },
+ {
+ "epoch": 0.7965065502183406,
+ "grad_norm": 0.023485427722334862,
+ "learning_rate": 0.000336,
+ "loss": 7.72553014755249,
+ "step": 57
+ },
+ {
+ "epoch": 0.8104803493449781,
+ "grad_norm": 0.019319474697113037,
+ "learning_rate": 0.00034199999999999996,
+ "loss": 7.709493160247803,
+ "step": 58
+ },
+ {
+ "epoch": 0.8244541484716157,
+ "grad_norm": 0.016705691814422607,
+ "learning_rate": 0.00034799999999999995,
+ "loss": 7.6873297691345215,
+ "step": 59
+ },
+ {
+ "epoch": 0.8384279475982532,
+ "grad_norm": 0.017667734995484352,
+ "learning_rate": 0.00035399999999999993,
+ "loss": 7.677059173583984,
+ "step": 60
+ },
+ {
+ "epoch": 0.8524017467248908,
+ "grad_norm": 0.01671992614865303,
+ "learning_rate": 0.00035999999999999997,
+ "loss": 7.643512725830078,
+ "step": 61
+ },
+ {
+ "epoch": 0.8663755458515284,
+ "grad_norm": 0.012288172729313374,
+ "learning_rate": 0.00036599999999999995,
+ "loss": 7.658343315124512,
+ "step": 62
+ },
+ {
+ "epoch": 0.880349344978166,
+ "grad_norm": 0.012004776857793331,
+ "learning_rate": 0.000372,
+ "loss": 7.649759292602539,
+ "step": 63
+ },
+ {
+ "epoch": 0.8943231441048035,
+ "grad_norm": 0.012487445957958698,
+ "learning_rate": 0.00037799999999999997,
+ "loss": 7.631503105163574,
+ "step": 64
+ },
+ {
+ "epoch": 0.9082969432314411,
+ "grad_norm": 0.01489241048693657,
+ "learning_rate": 0.00038399999999999996,
+ "loss": 7.647110462188721,
+ "step": 65
+ },
+ {
+ "epoch": 0.9222707423580786,
+ "grad_norm": 0.01428980939090252,
+ "learning_rate": 0.00039,
+ "loss": 7.636983871459961,
+ "step": 66
+ },
+ {
+ "epoch": 0.9362445414847161,
+ "grad_norm": 0.010711144655942917,
+ "learning_rate": 0.000396,
+ "loss": 7.621130466461182,
+ "step": 67
+ },
+ {
+ "epoch": 0.9502183406113537,
+ "grad_norm": 0.010178886353969574,
+ "learning_rate": 0.000402,
+ "loss": 7.593924522399902,
+ "step": 68
+ },
+ {
+ "epoch": 0.9641921397379912,
+ "grad_norm": 0.010604874230921268,
+ "learning_rate": 0.000408,
+ "loss": 7.6015496253967285,
+ "step": 69
+ },
+ {
+ "epoch": 0.9781659388646288,
+ "grad_norm": 0.009751426056027412,
+ "learning_rate": 0.0004139999999999999,
+ "loss": 7.571964263916016,
+ "step": 70
+ },
+ {
+ "epoch": 0.9921397379912664,
+ "grad_norm": 0.009816481731832027,
+ "learning_rate": 0.00041999999999999996,
+ "loss": 7.57344913482666,
+ "step": 71
+ },
+ {
+ "epoch": 1.0,
+ "grad_norm": 0.008665637113153934,
+ "learning_rate": 0.00042599999999999995,
+ "loss": 7.579050064086914,
+ "step": 72
+ },
+ {
+ "epoch": 1.0,
+ "eval_loss": 7.581204891204834,
+ "eval_runtime": 59.0091,
+ "eval_samples_per_second": 41.383,
+ "eval_steps_per_second": 1.305,
+ "step": 72
+ },
+ {
+ "epoch": 1.0139737991266375,
+ "grad_norm": 0.009684201329946518,
+ "learning_rate": 0.00043199999999999993,
+ "loss": 7.568594932556152,
+ "step": 73
+ },
+ {
+ "epoch": 1.027947598253275,
+ "grad_norm": 0.025783710181713104,
+ "learning_rate": 0.00043799999999999997,
+ "loss": 7.578948497772217,
+ "step": 74
+ },
+ {
+ "epoch": 1.0419213973799126,
+ "grad_norm": 0.0327819399535656,
+ "learning_rate": 0.00044399999999999995,
+ "loss": 7.548173904418945,
+ "step": 75
+ },
+ {
+ "epoch": 1.0558951965065502,
+ "grad_norm": 0.021604498848319054,
+ "learning_rate": 0.00045,
+ "loss": 7.552953720092773,
+ "step": 76
+ },
+ {
+ "epoch": 1.0698689956331877,
+ "grad_norm": 0.010230105370283127,
+ "learning_rate": 0.00045599999999999997,
+ "loss": 7.5213212966918945,
+ "step": 77
+ },
+ {
+ "epoch": 1.0838427947598253,
+ "grad_norm": 0.023315589874982834,
+ "learning_rate": 0.00046199999999999995,
+ "loss": 7.501640319824219,
+ "step": 78
+ },
+ {
+ "epoch": 1.0978165938864628,
+ "grad_norm": 0.01109944935888052,
+ "learning_rate": 0.000468,
+ "loss": 7.539519786834717,
+ "step": 79
+ },
+ {
+ "epoch": 1.1117903930131003,
+ "grad_norm": 0.026093758642673492,
+ "learning_rate": 0.000474,
+ "loss": 7.525928974151611,
+ "step": 80
+ },
+ {
+ "epoch": 1.125764192139738,
+ "grad_norm": 0.03274895250797272,
+ "learning_rate": 0.00047999999999999996,
+ "loss": 7.535602569580078,
+ "step": 81
+ },
+ {
+ "epoch": 1.1397379912663754,
+ "grad_norm": 0.01543602254241705,
+ "learning_rate": 0.000486,
+ "loss": 7.513683795928955,
+ "step": 82
+ },
+ {
+ "epoch": 1.1537117903930132,
+ "grad_norm": 0.028377249836921692,
+ "learning_rate": 0.0004919999999999999,
+ "loss": 7.49384880065918,
+ "step": 83
+ },
+ {
+ "epoch": 1.1676855895196507,
+ "grad_norm": 0.04121481627225876,
+ "learning_rate": 0.000498,
+ "loss": 7.488447189331055,
+ "step": 84
+ },
+ {
+ "epoch": 1.1816593886462883,
+ "grad_norm": 0.013282302767038345,
+ "learning_rate": 0.0005039999999999999,
+ "loss": 7.463010787963867,
+ "step": 85
+ },
+ {
+ "epoch": 1.1956331877729258,
+ "grad_norm": 0.03200364485383034,
+ "learning_rate": 0.0005099999999999999,
+ "loss": 7.473515033721924,
+ "step": 86
+ },
+ {
+ "epoch": 1.2096069868995634,
+ "grad_norm": 0.020768029615283012,
+ "learning_rate": 0.000516,
+ "loss": 7.464688777923584,
+ "step": 87
+ },
+ {
+ "epoch": 1.223580786026201,
+ "grad_norm": 0.021715788170695305,
+ "learning_rate": 0.000522,
+ "loss": 7.4835028648376465,
+ "step": 88
+ },
+ {
+ "epoch": 1.2375545851528384,
+ "grad_norm": 0.026255754753947258,
+ "learning_rate": 0.0005279999999999999,
+ "loss": 7.443392753601074,
+ "step": 89
+ },
+ {
+ "epoch": 1.251528384279476,
+ "grad_norm": 0.022325951606035233,
+ "learning_rate": 0.000534,
+ "loss": 7.457373142242432,
+ "step": 90
+ },
+ {
+ "epoch": 1.2655021834061135,
+ "grad_norm": 0.009270387701690197,
+ "learning_rate": 0.00054,
+ "loss": 7.455986976623535,
+ "step": 91
+ },
+ {
+ "epoch": 1.279475982532751,
+ "grad_norm": 0.02369544468820095,
+ "learning_rate": 0.0005459999999999999,
+ "loss": 7.442216873168945,
+ "step": 92
+ },
+ {
+ "epoch": 1.2934497816593886,
+ "grad_norm": 0.008474506437778473,
+ "learning_rate": 0.000552,
+ "loss": 7.4294915199279785,
+ "step": 93
+ },
+ {
+ "epoch": 1.3074235807860262,
+ "grad_norm": 0.017043249681591988,
+ "learning_rate": 0.000558,
+ "loss": 7.406932830810547,
+ "step": 94
+ },
+ {
+ "epoch": 1.3213973799126637,
+ "grad_norm": 0.011813412420451641,
+ "learning_rate": 0.0005639999999999999,
+ "loss": 7.415900230407715,
+ "step": 95
+ },
+ {
+ "epoch": 1.3353711790393012,
+ "grad_norm": 0.01086883433163166,
+ "learning_rate": 0.00057,
+ "loss": 7.452998638153076,
+ "step": 96
+ },
+ {
+ "epoch": 1.3493449781659388,
+ "grad_norm": 0.014700951054692268,
+ "learning_rate": 0.0005759999999999999,
+ "loss": 7.410249710083008,
+ "step": 97
+ },
+ {
+ "epoch": 1.3633187772925766,
+ "grad_norm": 0.014512976631522179,
+ "learning_rate": 0.0005819999999999999,
+ "loss": 7.429132461547852,
+ "step": 98
+ },
+ {
+ "epoch": 1.3772925764192139,
+ "grad_norm": 0.012519920244812965,
+ "learning_rate": 0.000588,
+ "loss": 7.430817604064941,
+ "step": 99
+ },
+ {
+ "epoch": 1.3912663755458516,
+ "grad_norm": 0.007132664322853088,
+ "learning_rate": 0.0005939999999999999,
+ "loss": 7.405234336853027,
+ "step": 100
+ },
+ {
+ "epoch": 1.405240174672489,
+ "grad_norm": 0.019489388912916183,
+ "learning_rate": 0.0006,
+ "loss": 7.427791595458984,
+ "step": 101
+ },
+ {
+ "epoch": 1.4192139737991267,
+ "grad_norm": 0.027252664789557457,
+ "learning_rate": 0.0006,
+ "loss": 7.41761589050293,
+ "step": 102
+ },
+ {
+ "epoch": 1.4331877729257643,
+ "grad_norm": 0.02501637488603592,
+ "learning_rate": 0.0006,
+ "loss": 7.416735649108887,
+ "step": 103
+ },
+ {
+ "epoch": 1.4471615720524018,
+ "grad_norm": 0.00989772193133831,
+ "learning_rate": 0.0006,
+ "loss": 7.3756632804870605,
+ "step": 104
+ },
+ {
+ "epoch": 1.4611353711790394,
+ "grad_norm": 0.01622949168086052,
+ "learning_rate": 0.0006,
+ "loss": 7.400811672210693,
+ "step": 105
+ },
+ {
+ "epoch": 1.475109170305677,
+ "grad_norm": 0.012407763861119747,
+ "learning_rate": 0.0006,
+ "loss": 7.37675666809082,
+ "step": 106
+ },
+ {
+ "epoch": 1.4890829694323144,
+ "grad_norm": 0.010353565216064453,
+ "learning_rate": 0.0006,
+ "loss": 7.4013142585754395,
+ "step": 107
+ },
+ {
+ "epoch": 1.503056768558952,
+ "grad_norm": 0.0120553532615304,
+ "learning_rate": 0.0006,
+ "loss": 7.379873275756836,
+ "step": 108
+ },
+ {
+ "epoch": 1.5170305676855895,
+ "grad_norm": 0.01610225811600685,
+ "learning_rate": 0.0006,
+ "loss": 7.371631145477295,
+ "step": 109
+ },
+ {
+ "epoch": 1.531004366812227,
+ "grad_norm": 0.01159879844635725,
+ "learning_rate": 0.0006,
+ "loss": 7.389007568359375,
+ "step": 110
+ },
+ {
+ "epoch": 1.5449781659388646,
+ "grad_norm": 0.008013816550374031,
+ "learning_rate": 0.0006,
+ "loss": 7.395920753479004,
+ "step": 111
+ },
+ {
+ "epoch": 1.5589519650655022,
+ "grad_norm": 0.016334809362888336,
+ "learning_rate": 0.0006,
+ "loss": 7.393121719360352,
+ "step": 112
+ },
+ {
+ "epoch": 1.5729257641921397,
+ "grad_norm": 0.018821721896529198,
+ "learning_rate": 0.0006,
+ "loss": 7.367543697357178,
+ "step": 113
+ },
+ {
+ "epoch": 1.5868995633187772,
+ "grad_norm": 0.014088994823396206,
+ "learning_rate": 0.0006,
+ "loss": 7.385537624359131,
+ "step": 114
+ },
+ {
+ "epoch": 1.600873362445415,
+ "grad_norm": 0.005864050704985857,
+ "learning_rate": 0.0006,
+ "loss": 7.365346908569336,
+ "step": 115
+ },
+ {
+ "epoch": 1.6148471615720523,
+ "grad_norm": 0.011745430529117584,
+ "learning_rate": 0.0006,
+ "loss": 7.390227317810059,
+ "step": 116
+ },
+ {
+ "epoch": 1.62882096069869,
+ "grad_norm": 0.017205573618412018,
+ "learning_rate": 0.0006,
+ "loss": 7.3558549880981445,
+ "step": 117
+ },
+ {
+ "epoch": 1.6427947598253274,
+ "grad_norm": 0.021256346255540848,
+ "learning_rate": 0.0006,
+ "loss": 7.377346992492676,
+ "step": 118
+ },
+ {
+ "epoch": 1.6567685589519652,
+ "grad_norm": 0.019335204735398293,
+ "learning_rate": 0.0006,
+ "loss": 7.346194267272949,
+ "step": 119
+ },
+ {
+ "epoch": 1.6707423580786025,
+ "grad_norm": 0.008134003728628159,
+ "learning_rate": 0.0006,
+ "loss": 7.3063578605651855,
+ "step": 120
+ },
+ {
+ "epoch": 1.6847161572052403,
+ "grad_norm": 0.017613831907510757,
+ "learning_rate": 0.0006,
+ "loss": 7.338171005249023,
+ "step": 121
+ },
+ {
+ "epoch": 1.6986899563318776,
+ "grad_norm": 0.00890734139829874,
+ "learning_rate": 0.0006,
+ "loss": 7.355384826660156,
+ "step": 122
+ },
+ {
+ "epoch": 1.7126637554585153,
+ "grad_norm": 0.013448155485093594,
+ "learning_rate": 0.0006,
+ "loss": 7.344296455383301,
+ "step": 123
+ },
+ {
+ "epoch": 1.726637554585153,
+ "grad_norm": 0.009241633117198944,
+ "learning_rate": 0.0006,
+ "loss": 7.3454060554504395,
+ "step": 124
+ },
+ {
+ "epoch": 1.7406113537117904,
+ "grad_norm": 0.018561311066150665,
+ "learning_rate": 0.0006,
+ "loss": 7.32348108291626,
+ "step": 125
+ },
+ {
+ "epoch": 1.754585152838428,
+ "grad_norm": 0.025667704641819,
+ "learning_rate": 0.0006,
+ "loss": 7.356142520904541,
+ "step": 126
+ },
+ {
+ "epoch": 1.7685589519650655,
+ "grad_norm": 0.036232151091098785,
+ "learning_rate": 0.0006,
+ "loss": 7.363539695739746,
+ "step": 127
+ },
+ {
+ "epoch": 1.782532751091703,
+ "grad_norm": 0.059483714401721954,
+ "learning_rate": 0.0006,
+ "loss": 7.402331352233887,
+ "step": 128
+ },
+ {
+ "epoch": 1.7965065502183406,
+ "grad_norm": 0.05365009233355522,
+ "learning_rate": 0.0006,
+ "loss": 7.40457820892334,
+ "step": 129
+ },
+ {
+ "epoch": 1.8104803493449781,
+ "grad_norm": 0.02738415263593197,
+ "learning_rate": 0.0006,
+ "loss": 7.3756208419799805,
+ "step": 130
+ },
+ {
+ "epoch": 1.8244541484716157,
+ "grad_norm": 0.035017091780900955,
+ "learning_rate": 0.0006,
+ "loss": 7.379263877868652,
+ "step": 131
+ },
+ {
+ "epoch": 1.8384279475982532,
+ "grad_norm": 0.02819863148033619,
+ "learning_rate": 0.0006,
+ "loss": 7.327673435211182,
+ "step": 132
+ },
+ {
+ "epoch": 1.8524017467248908,
+ "grad_norm": 0.017754238098859787,
+ "learning_rate": 0.0006,
+ "loss": 7.333420276641846,
+ "step": 133
+ },
+ {
+ "epoch": 1.8663755458515285,
+ "grad_norm": 0.024685733020305634,
+ "learning_rate": 0.0006,
+ "loss": 7.34493350982666,
+ "step": 134
+ },
+ {
+ "epoch": 1.8803493449781659,
+ "grad_norm": 0.019426532089710236,
+ "learning_rate": 0.0006,
+ "loss": 7.328448295593262,
+ "step": 135
+ },
+ {
+ "epoch": 1.8943231441048036,
+ "grad_norm": 0.02106391079723835,
+ "learning_rate": 0.0006,
+ "loss": 7.327917098999023,
+ "step": 136
+ },
+ {
+ "epoch": 1.908296943231441,
+ "grad_norm": 0.02056879736483097,
+ "learning_rate": 0.0006,
+ "loss": 7.3311991691589355,
+ "step": 137
+ },
+ {
+ "epoch": 1.9222707423580787,
+ "grad_norm": 0.016865141689777374,
+ "learning_rate": 0.0006,
+ "loss": 7.312766075134277,
+ "step": 138
+ },
+ {
+ "epoch": 1.936244541484716,
+ "grad_norm": 0.017549855634570122,
+ "learning_rate": 0.0006,
+ "loss": 7.31191349029541,
+ "step": 139
+ },
+ {
+ "epoch": 1.9502183406113538,
+ "grad_norm": 0.015930157154798508,
+ "learning_rate": 0.0006,
+ "loss": 7.330901622772217,
+ "step": 140
+ },
+ {
+ "epoch": 1.9641921397379911,
+ "grad_norm": 0.013283558189868927,
+ "learning_rate": 0.0006,
+ "loss": 7.298318386077881,
+ "step": 141
+ },
+ {
+ "epoch": 1.9781659388646289,
+ "grad_norm": 0.012023469433188438,
+ "learning_rate": 0.0006,
+ "loss": 7.282337188720703,
+ "step": 142
+ },
+ {
+ "epoch": 1.9921397379912664,
+ "grad_norm": 0.01459511648863554,
+ "learning_rate": 0.0006,
+ "loss": 7.282924175262451,
+ "step": 143
+ },
+ {
+ "epoch": 2.0,
+ "grad_norm": 0.017474235966801643,
+ "learning_rate": 0.0006,
+ "loss": 7.254333972930908,
+ "step": 144
+ },
+ {
+ "epoch": 2.0,
+ "eval_loss": 7.318708896636963,
+ "eval_runtime": 59.6555,
+ "eval_samples_per_second": 40.935,
+ "eval_steps_per_second": 1.291,
+ "step": 144
+ },
+ {
+ "epoch": 2.0139737991266378,
+ "grad_norm": 0.019757507368922234,
+ "learning_rate": 0.0006,
+ "loss": 7.30288553237915,
+ "step": 145
+ },
+ {
+ "epoch": 2.027947598253275,
+ "grad_norm": 0.014571248553693295,
+ "learning_rate": 0.0006,
+ "loss": 7.274681568145752,
+ "step": 146
+ },
+ {
+ "epoch": 2.041921397379913,
+ "grad_norm": 0.015744198113679886,
+ "learning_rate": 0.0006,
+ "loss": 7.278632640838623,
+ "step": 147
+ },
+ {
+ "epoch": 2.05589519650655,
+ "grad_norm": 0.028048858046531677,
+ "learning_rate": 0.0006,
+ "loss": 7.26558780670166,
+ "step": 148
+ },
+ {
+ "epoch": 2.069868995633188,
+ "grad_norm": 0.06547307223081589,
+ "learning_rate": 0.0006,
+ "loss": 7.32567834854126,
+ "step": 149
+ },
+ {
+ "epoch": 2.0838427947598253,
+ "grad_norm": 0.05969763547182083,
+ "learning_rate": 0.0006,
+ "loss": 7.395836353302002,
+ "step": 150
+ },
+ {
+ "epoch": 2.097816593886463,
+ "grad_norm": 0.032574281096458435,
+ "learning_rate": 0.0006,
+ "loss": 7.318150997161865,
+ "step": 151
+ },
+ {
+ "epoch": 2.1117903930131003,
+ "grad_norm": 0.029452962800860405,
+ "learning_rate": 0.0006,
+ "loss": 7.308066368103027,
+ "step": 152
+ },
+ {
+ "epoch": 2.125764192139738,
+ "grad_norm": 0.03161991387605667,
+ "learning_rate": 0.0006,
+ "loss": 7.306290626525879,
+ "step": 153
+ },
+ {
+ "epoch": 2.1397379912663754,
+ "grad_norm": 0.027096295729279518,
+ "learning_rate": 0.0006,
+ "loss": 7.297796249389648,
+ "step": 154
+ },
+ {
+ "epoch": 2.153711790393013,
+ "grad_norm": 0.028657056391239166,
+ "learning_rate": 0.0006,
+ "loss": 7.317424774169922,
+ "step": 155
+ },
+ {
+ "epoch": 2.1676855895196505,
+ "grad_norm": 0.02046995982527733,
+ "learning_rate": 0.0006,
+ "loss": 7.297986030578613,
+ "step": 156
+ },
+ {
+ "epoch": 2.1816593886462883,
+ "grad_norm": 0.02220369316637516,
+ "learning_rate": 0.0006,
+ "loss": 7.268237590789795,
+ "step": 157
+ },
+ {
+ "epoch": 2.1956331877729256,
+ "grad_norm": 0.018381357192993164,
+ "learning_rate": 0.0006,
+ "loss": 7.259415626525879,
+ "step": 158
+ },
+ {
+ "epoch": 2.2096069868995634,
+ "grad_norm": 0.0204413291066885,
+ "learning_rate": 0.0006,
+ "loss": 7.258173942565918,
+ "step": 159
+ },
+ {
+ "epoch": 2.2235807860262007,
+ "grad_norm": 0.015874288976192474,
+ "learning_rate": 0.0006,
+ "loss": 7.281834602355957,
+ "step": 160
+ },
+ {
+ "epoch": 2.2375545851528384,
+ "grad_norm": 0.01634068600833416,
+ "learning_rate": 0.0006,
+ "loss": 7.2469868659973145,
+ "step": 161
+ },
+ {
+ "epoch": 2.251528384279476,
+ "grad_norm": 0.01728747971355915,
+ "learning_rate": 0.0006,
+ "loss": 7.280995845794678,
+ "step": 162
+ },
+ {
+ "epoch": 2.2655021834061135,
+ "grad_norm": 0.016341188922524452,
+ "learning_rate": 0.0006,
+ "loss": 7.273357391357422,
+ "step": 163
+ },
+ {
+ "epoch": 2.279475982532751,
+ "grad_norm": 0.016125807538628578,
+ "learning_rate": 0.0006,
+ "loss": 7.273510932922363,
+ "step": 164
+ },
+ {
+ "epoch": 2.2934497816593886,
+ "grad_norm": 0.01629287749528885,
+ "learning_rate": 0.0006,
+ "loss": 7.267029762268066,
+ "step": 165
+ },
+ {
+ "epoch": 2.3074235807860264,
+ "grad_norm": 0.015296131372451782,
+ "learning_rate": 0.0006,
+ "loss": 7.211989402770996,
+ "step": 166
+ },
+ {
+ "epoch": 2.3213973799126637,
+ "grad_norm": 0.01709570363163948,
+ "learning_rate": 0.0006,
+ "loss": 7.242851734161377,
+ "step": 167
+ },
+ {
+ "epoch": 2.3353711790393015,
+ "grad_norm": 0.012552388943731785,
+ "learning_rate": 0.0006,
+ "loss": 7.241696834564209,
+ "step": 168
+ },
+ {
+ "epoch": 2.349344978165939,
+ "grad_norm": 0.014571291394531727,
+ "learning_rate": 0.0006,
+ "loss": 7.270681381225586,
+ "step": 169
+ },
+ {
+ "epoch": 2.3633187772925766,
+ "grad_norm": 0.013389437459409237,
+ "learning_rate": 0.0006,
+ "loss": 7.191600799560547,
+ "step": 170
+ },
+ {
+ "epoch": 2.377292576419214,
+ "grad_norm": 0.011895314790308475,
+ "learning_rate": 0.0006,
+ "loss": 7.213610649108887,
+ "step": 171
+ },
+ {
+ "epoch": 2.3912663755458516,
+ "grad_norm": 0.01345642190426588,
+ "learning_rate": 0.0006,
+ "loss": 7.193787574768066,
+ "step": 172
+ },
+ {
+ "epoch": 2.405240174672489,
+ "grad_norm": 0.0091794328764081,
+ "learning_rate": 0.0006,
+ "loss": 7.233364105224609,
+ "step": 173
+ },
+ {
+ "epoch": 2.4192139737991267,
+ "grad_norm": 0.013767086900770664,
+ "learning_rate": 0.0006,
+ "loss": 7.217272758483887,
+ "step": 174
+ },
+ {
+ "epoch": 2.433187772925764,
+ "grad_norm": 0.015134960412979126,
+ "learning_rate": 0.0006,
+ "loss": 7.192791938781738,
+ "step": 175
+ },
+ {
+ "epoch": 2.447161572052402,
+ "grad_norm": 0.01790648326277733,
+ "learning_rate": 0.0006,
+ "loss": 7.229467391967773,
+ "step": 176
+ },
+ {
+ "epoch": 2.461135371179039,
+ "grad_norm": 0.015362377278506756,
+ "learning_rate": 0.0006,
+ "loss": 7.180346488952637,
+ "step": 177
+ },
+ {
+ "epoch": 2.475109170305677,
+ "grad_norm": 0.010505498386919498,
+ "learning_rate": 0.0006,
+ "loss": 7.188833236694336,
+ "step": 178
+ },
+ {
+ "epoch": 2.489082969432314,
+ "grad_norm": 0.009129747748374939,
+ "learning_rate": 0.0006,
+ "loss": 7.203153133392334,
+ "step": 179
+ },
+ {
+ "epoch": 2.503056768558952,
+ "grad_norm": 0.01147883478552103,
+ "learning_rate": 0.0006,
+ "loss": 7.163824558258057,
+ "step": 180
+ },
+ {
+ "epoch": 2.5170305676855893,
+ "grad_norm": 0.011338942684233189,
+ "learning_rate": 0.0006,
+ "loss": 7.16287899017334,
+ "step": 181
+ },
+ {
+ "epoch": 2.531004366812227,
+ "grad_norm": 0.013277675025165081,
+ "learning_rate": 0.0006,
+ "loss": 7.155905723571777,
+ "step": 182
+ },
+ {
+ "epoch": 2.544978165938865,
+ "grad_norm": 0.015269347466528416,
+ "learning_rate": 0.0006,
+ "loss": 7.189371109008789,
+ "step": 183
+ },
+ {
+ "epoch": 2.558951965065502,
+ "grad_norm": 0.017905596643686295,
+ "learning_rate": 0.0006,
+ "loss": 7.177547931671143,
+ "step": 184
+ },
+ {
+ "epoch": 2.5729257641921395,
+ "grad_norm": 0.024431610479950905,
+ "learning_rate": 0.0006,
+ "loss": 7.192169189453125,
+ "step": 185
+ },
+ {
+ "epoch": 2.5868995633187772,
+ "grad_norm": 0.03656961768865585,
+ "learning_rate": 0.0006,
+ "loss": 7.124387741088867,
+ "step": 186
+ },
+ {
+ "epoch": 2.600873362445415,
+ "grad_norm": 0.04337048903107643,
+ "learning_rate": 0.0006,
+ "loss": 7.140231609344482,
+ "step": 187
+ },
+ {
+ "epoch": 2.6148471615720523,
+ "grad_norm": 0.04047228768467903,
+ "learning_rate": 0.0006,
+ "loss": 7.171927452087402,
+ "step": 188
+ },
+ {
+ "epoch": 2.62882096069869,
+ "grad_norm": 0.07825770974159241,
+ "learning_rate": 0.0006,
+ "loss": 7.207631587982178,
+ "step": 189
+ },
+ {
+ "epoch": 2.6427947598253274,
+ "grad_norm": 0.07118063420057297,
+ "learning_rate": 0.0006,
+ "loss": 7.209488391876221,
+ "step": 190
+ },
+ {
+ "epoch": 2.656768558951965,
+ "grad_norm": 0.05153293535113335,
+ "learning_rate": 0.0006,
+ "loss": 7.14793062210083,
+ "step": 191
+ },
+ {
+ "epoch": 2.6707423580786025,
+ "grad_norm": 0.03318839520215988,
+ "learning_rate": 0.0006,
+ "loss": 7.163010597229004,
+ "step": 192
+ },
+ {
+ "epoch": 2.6847161572052403,
+ "grad_norm": 0.029333539307117462,
+ "learning_rate": 0.0006,
+ "loss": 7.159955978393555,
+ "step": 193
+ },
+ {
+ "epoch": 2.6986899563318776,
+ "grad_norm": 0.029405880719423294,
+ "learning_rate": 0.0006,
+ "loss": 7.136178493499756,
+ "step": 194
+ },
+ {
+ "epoch": 2.7126637554585153,
+ "grad_norm": 0.037815988063812256,
+ "learning_rate": 0.0006,
+ "loss": 7.183186054229736,
+ "step": 195
+ },
+ {
+ "epoch": 2.726637554585153,
+ "grad_norm": 0.020307250320911407,
+ "learning_rate": 0.0006,
+ "loss": 7.133164882659912,
+ "step": 196
+ },
+ {
+ "epoch": 2.7406113537117904,
+ "grad_norm": 0.026048069819808006,
+ "learning_rate": 0.0006,
+ "loss": 7.124977111816406,
+ "step": 197
+ },
+ {
+ "epoch": 2.7545851528384278,
+ "grad_norm": 0.027316724881529808,
+ "learning_rate": 0.0006,
+ "loss": 7.07265567779541,
+ "step": 198
+ },
+ {
+ "epoch": 2.7685589519650655,
+ "grad_norm": 0.021629009395837784,
+ "learning_rate": 0.0006,
+ "loss": 7.150020599365234,
+ "step": 199
+ },
+ {
+ "epoch": 2.7825327510917033,
+ "grad_norm": 0.01605929434299469,
+ "learning_rate": 0.0006,
+ "loss": 7.10919189453125,
+ "step": 200
+ },
+ {
+ "epoch": 2.7965065502183406,
+ "grad_norm": 0.020800089463591576,
+ "learning_rate": 0.0006,
+ "loss": 7.132045269012451,
+ "step": 201
+ },
+ {
+ "epoch": 2.810480349344978,
+ "grad_norm": 0.019225774332880974,
+ "learning_rate": 0.0006,
+ "loss": 7.128364562988281,
+ "step": 202
+ },
+ {
+ "epoch": 2.8244541484716157,
+ "grad_norm": 0.01561372634023428,
+ "learning_rate": 0.0006,
+ "loss": 7.083669662475586,
+ "step": 203
+ },
+ {
+ "epoch": 2.8384279475982535,
+ "grad_norm": 0.021969519555568695,
+ "learning_rate": 0.0006,
+ "loss": 7.050841331481934,
+ "step": 204
+ },
+ {
+ "epoch": 2.8524017467248908,
+ "grad_norm": 0.021453695371747017,
+ "learning_rate": 0.0006,
+ "loss": 7.097085475921631,
+ "step": 205
+ },
+ {
+ "epoch": 2.8663755458515285,
+ "grad_norm": 0.022836022078990936,
+ "learning_rate": 0.0006,
+ "loss": 7.10405969619751,
+ "step": 206
+ },
+ {
+ "epoch": 2.880349344978166,
+ "grad_norm": 0.02430851384997368,
+ "learning_rate": 0.0006,
+ "loss": 7.045181751251221,
+ "step": 207
+ },
+ {
+ "epoch": 2.8943231441048036,
+ "grad_norm": 0.020166993141174316,
+ "learning_rate": 0.0006,
+ "loss": 7.068183898925781,
+ "step": 208
+ },
+ {
+ "epoch": 2.908296943231441,
+ "grad_norm": 0.0118505684658885,
+ "learning_rate": 0.0006,
+ "loss": 7.07132625579834,
+ "step": 209
+ },
+ {
+ "epoch": 2.9222707423580787,
+ "grad_norm": 0.023772427812218666,
+ "learning_rate": 0.0006,
+ "loss": 7.055238723754883,
+ "step": 210
+ },
+ {
+ "epoch": 2.936244541484716,
+ "grad_norm": 0.02350712940096855,
+ "learning_rate": 0.0006,
+ "loss": 7.055898666381836,
+ "step": 211
+ },
+ {
+ "epoch": 2.950218340611354,
+ "grad_norm": 0.017244907096028328,
+ "learning_rate": 0.0006,
+ "loss": 7.0099711418151855,
+ "step": 212
+ },
+ {
+ "epoch": 2.964192139737991,
+ "grad_norm": 0.021565575152635574,
+ "learning_rate": 0.0006,
+ "loss": 6.991751670837402,
+ "step": 213
+ },
+ {
+ "epoch": 2.978165938864629,
+ "grad_norm": 0.03241860866546631,
+ "learning_rate": 0.0006,
+ "loss": 7.005980491638184,
+ "step": 214
+ },
+ {
+ "epoch": 2.992139737991266,
+ "grad_norm": 0.04613726586103439,
+ "learning_rate": 0.0006,
+ "loss": 7.027859687805176,
+ "step": 215
+ },
+ {
+ "epoch": 3.0,
+ "grad_norm": 0.06404894590377808,
+ "learning_rate": 0.0006,
+ "loss": 7.123725891113281,
+ "step": 216
+ },
+ {
+ "epoch": 3.0,
+ "eval_loss": 7.070647716522217,
+ "eval_runtime": 58.5801,
+ "eval_samples_per_second": 41.687,
+ "eval_steps_per_second": 1.314,
+ "step": 216
+ },
+ {
+ "epoch": 3.0139737991266378,
+ "grad_norm": 0.05542432889342308,
+ "learning_rate": 0.0006,
+ "loss": 7.078299522399902,
+ "step": 217
+ },
+ {
+ "epoch": 3.027947598253275,
+ "grad_norm": 0.026484526693820953,
+ "learning_rate": 0.0006,
+ "loss": 6.961159706115723,
+ "step": 218
+ },
+ {
+ "epoch": 3.041921397379913,
+ "grad_norm": 0.028276391327381134,
+ "learning_rate": 0.0006,
+ "loss": 7.019961833953857,
+ "step": 219
+ },
+ {
+ "epoch": 3.05589519650655,
+ "grad_norm": 0.024486595764756203,
+ "learning_rate": 0.0006,
+ "loss": 7.039677619934082,
+ "step": 220
+ },
+ {
+ "epoch": 3.069868995633188,
+ "grad_norm": 0.017620893195271492,
+ "learning_rate": 0.0006,
+ "loss": 6.9270782470703125,
+ "step": 221
+ },
+ {
+ "epoch": 3.0838427947598253,
+ "grad_norm": 0.02547302283346653,
+ "learning_rate": 0.0006,
+ "loss": 6.924233436584473,
+ "step": 222
+ },
+ {
+ "epoch": 3.097816593886463,
+ "grad_norm": 0.02539009042084217,
+ "learning_rate": 0.0006,
+ "loss": 6.973979949951172,
+ "step": 223
+ },
+ {
+ "epoch": 3.1117903930131003,
+ "grad_norm": 0.03256227448582649,
+ "learning_rate": 0.0006,
+ "loss": 6.933725357055664,
+ "step": 224
+ },
+ {
+ "epoch": 3.125764192139738,
+ "grad_norm": 0.04651800915598869,
+ "learning_rate": 0.0006,
+ "loss": 6.951044082641602,
+ "step": 225
+ },
+ {
+ "epoch": 3.1397379912663754,
+ "grad_norm": 0.04822975769639015,
+ "learning_rate": 0.0006,
+ "loss": 6.9614410400390625,
+ "step": 226
+ },
+ {
+ "epoch": 3.153711790393013,
+ "grad_norm": 0.031165020540356636,
+ "learning_rate": 0.0006,
+ "loss": 6.91193151473999,
+ "step": 227
+ },
+ {
+ "epoch": 3.1676855895196505,
+ "grad_norm": 0.029754292219877243,
+ "learning_rate": 0.0006,
+ "loss": 6.91802978515625,
+ "step": 228
+ },
+ {
+ "epoch": 3.1816593886462883,
+ "grad_norm": 0.02193152718245983,
+ "learning_rate": 0.0006,
+ "loss": 6.928090572357178,
+ "step": 229
+ },
+ {
+ "epoch": 3.1956331877729256,
+ "grad_norm": 0.02428470179438591,
+ "learning_rate": 0.0006,
+ "loss": 6.90396785736084,
+ "step": 230
+ },
+ {
+ "epoch": 3.2096069868995634,
+ "grad_norm": 0.023374345153570175,
+ "learning_rate": 0.0006,
+ "loss": 6.885858535766602,
+ "step": 231
+ },
+ {
+ "epoch": 3.2235807860262007,
+ "grad_norm": 0.023405000567436218,
+ "learning_rate": 0.0006,
+ "loss": 6.8855767250061035,
+ "step": 232
+ },
+ {
+ "epoch": 3.2375545851528384,
+ "grad_norm": 0.017277134582400322,
+ "learning_rate": 0.0006,
+ "loss": 6.901449203491211,
+ "step": 233
+ },
+ {
+ "epoch": 3.251528384279476,
+ "grad_norm": 0.019496750086545944,
+ "learning_rate": 0.0006,
+ "loss": 6.885687828063965,
+ "step": 234
+ },
+ {
+ "epoch": 3.2655021834061135,
+ "grad_norm": 0.016893276944756508,
+ "learning_rate": 0.0006,
+ "loss": 6.872369766235352,
+ "step": 235
+ },
+ {
+ "epoch": 3.279475982532751,
+ "grad_norm": 0.02058715932071209,
+ "learning_rate": 0.0006,
+ "loss": 6.823280334472656,
+ "step": 236
+ },
+ {
+ "epoch": 3.2934497816593886,
+ "grad_norm": 0.015530755743384361,
+ "learning_rate": 0.0006,
+ "loss": 6.860499858856201,
+ "step": 237
+ },
+ {
+ "epoch": 3.3074235807860264,
+ "grad_norm": 0.01934926211833954,
+ "learning_rate": 0.0006,
+ "loss": 6.817448616027832,
+ "step": 238
+ },
+ {
+ "epoch": 3.3213973799126637,
+ "grad_norm": 0.017413552850484848,
+ "learning_rate": 0.0006,
+ "loss": 6.881955146789551,
+ "step": 239
+ },
+ {
+ "epoch": 3.3353711790393015,
+ "grad_norm": 0.026478875428438187,
+ "learning_rate": 0.0006,
+ "loss": 6.827404499053955,
+ "step": 240
+ },
+ {
+ "epoch": 3.349344978165939,
+ "grad_norm": 0.032611701637506485,
+ "learning_rate": 0.0006,
+ "loss": 6.78285551071167,
+ "step": 241
+ },
+ {
+ "epoch": 3.3633187772925766,
+ "grad_norm": 0.041900306940078735,
+ "learning_rate": 0.0006,
+ "loss": 6.783053874969482,
+ "step": 242
+ },
+ {
+ "epoch": 3.377292576419214,
+ "grad_norm": 0.04149497672915459,
+ "learning_rate": 0.0006,
+ "loss": 6.745220184326172,
+ "step": 243
+ },
+ {
+ "epoch": 3.3912663755458516,
+ "grad_norm": 0.023949826136231422,
+ "learning_rate": 0.0006,
+ "loss": 6.841533184051514,
+ "step": 244
+ },
+ {
+ "epoch": 3.405240174672489,
+ "grad_norm": 0.028601842001080513,
+ "learning_rate": 0.0006,
+ "loss": 6.852852821350098,
+ "step": 245
+ },
+ {
+ "epoch": 3.4192139737991267,
+ "grad_norm": 0.027365155518054962,
+ "learning_rate": 0.0006,
+ "loss": 6.74376106262207,
+ "step": 246
+ },
+ {
+ "epoch": 3.433187772925764,
+ "grad_norm": 0.02856568619608879,
+ "learning_rate": 0.0006,
+ "loss": 6.74906063079834,
+ "step": 247
+ },
+ {
+ "epoch": 3.447161572052402,
+ "grad_norm": 0.019172416999936104,
+ "learning_rate": 0.0006,
+ "loss": 6.809206008911133,
+ "step": 248
+ },
+ {
+ "epoch": 3.461135371179039,
+ "grad_norm": 0.02161746844649315,
+ "learning_rate": 0.0006,
+ "loss": 6.789888381958008,
+ "step": 249
+ },
+ {
+ "epoch": 3.475109170305677,
+ "grad_norm": 0.020273666828870773,
+ "learning_rate": 0.0006,
+ "loss": 6.73512077331543,
+ "step": 250
+ },
+ {
+ "epoch": 3.489082969432314,
+ "grad_norm": 0.022329283878207207,
+ "learning_rate": 0.0006,
+ "loss": 6.826148509979248,
+ "step": 251
+ },
+ {
+ "epoch": 3.503056768558952,
+ "grad_norm": 0.02284284122288227,
+ "learning_rate": 0.0006,
+ "loss": 6.748521327972412,
+ "step": 252
+ },
+ {
+ "epoch": 3.5170305676855893,
+ "grad_norm": 0.026368247345089912,
+ "learning_rate": 0.0006,
+ "loss": 6.645879745483398,
+ "step": 253
+ },
+ {
+ "epoch": 3.531004366812227,
+ "grad_norm": 0.03627244383096695,
+ "learning_rate": 0.0006,
+ "loss": 6.666064739227295,
+ "step": 254
+ },
+ {
+ "epoch": 3.544978165938865,
+ "grad_norm": 0.04140935093164444,
+ "learning_rate": 0.0006,
+ "loss": 6.712275981903076,
+ "step": 255
+ },
+ {
+ "epoch": 3.558951965065502,
+ "grad_norm": 0.0390462800860405,
+ "learning_rate": 0.0006,
+ "loss": 6.665900230407715,
+ "step": 256
+ },
+ {
+ "epoch": 3.5729257641921395,
+ "grad_norm": 0.05178583785891533,
+ "learning_rate": 0.0006,
+ "loss": 6.753298759460449,
+ "step": 257
+ },
+ {
+ "epoch": 3.5868995633187772,
+ "grad_norm": 0.054328061640262604,
+ "learning_rate": 0.0006,
+ "loss": 6.732633113861084,
+ "step": 258
+ },
+ {
+ "epoch": 3.600873362445415,
+ "grad_norm": 0.04590460658073425,
+ "learning_rate": 0.0006,
+ "loss": 6.65250301361084,
+ "step": 259
+ },
+ {
+ "epoch": 3.6148471615720523,
+ "grad_norm": 0.044331666082143784,
+ "learning_rate": 0.0006,
+ "loss": 6.68654203414917,
+ "step": 260
+ },
+ {
+ "epoch": 3.62882096069869,
+ "grad_norm": 0.05308730527758598,
+ "learning_rate": 0.0006,
+ "loss": 6.642482757568359,
+ "step": 261
+ },
+ {
+ "epoch": 3.6427947598253274,
+ "grad_norm": 0.03999016433954239,
+ "learning_rate": 0.0006,
+ "loss": 6.625227451324463,
+ "step": 262
+ },
+ {
+ "epoch": 3.656768558951965,
+ "grad_norm": 0.04088086634874344,
+ "learning_rate": 0.0006,
+ "loss": 6.67623233795166,
+ "step": 263
+ },
+ {
+ "epoch": 3.6707423580786025,
+ "grad_norm": 0.03615636005997658,
+ "learning_rate": 0.0006,
+ "loss": 6.6869001388549805,
+ "step": 264
+ },
+ {
+ "epoch": 3.6847161572052403,
+ "grad_norm": 0.02817360684275627,
+ "learning_rate": 0.0006,
+ "loss": 6.674601078033447,
+ "step": 265
+ },
+ {
+ "epoch": 3.6986899563318776,
+ "grad_norm": 0.032586123794317245,
+ "learning_rate": 0.0006,
+ "loss": 6.612994194030762,
+ "step": 266
+ },
+ {
+ "epoch": 3.7126637554585153,
+ "grad_norm": 0.031117867678403854,
+ "learning_rate": 0.0006,
+ "loss": 6.6244354248046875,
+ "step": 267
+ },
+ {
+ "epoch": 3.726637554585153,
+ "grad_norm": 0.028330687433481216,
+ "learning_rate": 0.0006,
+ "loss": 6.669557094573975,
+ "step": 268
+ },
+ {
+ "epoch": 3.7406113537117904,
+ "grad_norm": 0.021489202976226807,
+ "learning_rate": 0.0006,
+ "loss": 6.632035732269287,
+ "step": 269
+ },
+ {
+ "epoch": 3.7545851528384278,
+ "grad_norm": 0.024269424378871918,
+ "learning_rate": 0.0006,
+ "loss": 6.660867691040039,
+ "step": 270
+ },
+ {
+ "epoch": 3.7685589519650655,
+ "grad_norm": 0.02237873338162899,
+ "learning_rate": 0.0006,
+ "loss": 6.602587699890137,
+ "step": 271
+ },
+ {
+ "epoch": 3.7825327510917033,
+ "grad_norm": 0.026503929868340492,
+ "learning_rate": 0.0006,
+ "loss": 6.599200248718262,
+ "step": 272
+ },
+ {
+ "epoch": 3.7965065502183406,
+ "grad_norm": 0.03687124326825142,
+ "learning_rate": 0.0006,
+ "loss": 6.567296504974365,
+ "step": 273
+ },
+ {
+ "epoch": 3.810480349344978,
+ "grad_norm": 0.04708952456712723,
+ "learning_rate": 0.0006,
+ "loss": 6.621265411376953,
+ "step": 274
+ },
+ {
+ "epoch": 3.8244541484716157,
+ "grad_norm": 0.041846614331007004,
+ "learning_rate": 0.0006,
+ "loss": 6.550536155700684,
+ "step": 275
+ },
+ {
+ "epoch": 3.8384279475982535,
+ "grad_norm": 0.031249675899744034,
+ "learning_rate": 0.0006,
+ "loss": 6.54278564453125,
+ "step": 276
+ },
+ {
+ "epoch": 3.8524017467248908,
+ "grad_norm": 0.023082256317138672,
+ "learning_rate": 0.0006,
+ "loss": 6.544787406921387,
+ "step": 277
+ },
+ {
+ "epoch": 3.8663755458515285,
+ "grad_norm": 0.020995570346713066,
+ "learning_rate": 0.0006,
+ "loss": 6.518143653869629,
+ "step": 278
+ },
+ {
+ "epoch": 3.880349344978166,
+ "grad_norm": 0.020839324221014977,
+ "learning_rate": 0.0006,
+ "loss": 6.546667098999023,
+ "step": 279
+ },
+ {
+ "epoch": 3.8943231441048036,
+ "grad_norm": 0.018377432599663734,
+ "learning_rate": 0.0006,
+ "loss": 6.478307723999023,
+ "step": 280
+ },
+ {
+ "epoch": 3.908296943231441,
+ "grad_norm": 0.021713724359869957,
+ "learning_rate": 0.0006,
+ "loss": 6.485574245452881,
+ "step": 281
+ },
+ {
+ "epoch": 3.9222707423580787,
+ "grad_norm": 0.01359301246702671,
+ "learning_rate": 0.0006,
+ "loss": 6.570268630981445,
+ "step": 282
+ },
+ {
+ "epoch": 3.936244541484716,
+ "grad_norm": 0.016233332455158234,
+ "learning_rate": 0.0006,
+ "loss": 6.57150936126709,
+ "step": 283
+ },
+ {
+ "epoch": 3.950218340611354,
+ "grad_norm": 0.020687608048319817,
+ "learning_rate": 0.0006,
+ "loss": 6.527956008911133,
+ "step": 284
+ },
+ {
+ "epoch": 3.964192139737991,
+ "grad_norm": 0.02032964862883091,
+ "learning_rate": 0.0006,
+ "loss": 6.468169689178467,
+ "step": 285
+ },
+ {
+ "epoch": 3.978165938864629,
+ "grad_norm": 0.016676288098096848,
+ "learning_rate": 0.0006,
+ "loss": 6.448418617248535,
+ "step": 286
+ },
+ {
+ "epoch": 3.992139737991266,
+ "grad_norm": 0.0168539360165596,
+ "learning_rate": 0.0006,
+ "loss": 6.461180686950684,
+ "step": 287
+ },
+ {
+ "epoch": 4.0,
+ "grad_norm": 0.02666233666241169,
+ "learning_rate": 0.0006,
+ "loss": 6.344893932342529,
+ "step": 288
+ },
+ {
+ "epoch": 4.0,
+ "eval_loss": 6.5384979248046875,
+ "eval_runtime": 58.814,
+ "eval_samples_per_second": 41.521,
+ "eval_steps_per_second": 1.309,
+ "step": 288
+ },
+ {
+ "epoch": 4.013973799126638,
+ "grad_norm": 0.04154708981513977,
+ "learning_rate": 0.0006,
+ "loss": 6.41855525970459,
+ "step": 289
+ },
+ {
+ "epoch": 4.0279475982532755,
+ "grad_norm": 0.04895668104290962,
+ "learning_rate": 0.0006,
+ "loss": 6.499302864074707,
+ "step": 290
+ },
+ {
+ "epoch": 4.041921397379912,
+ "grad_norm": 0.034464482218027115,
+ "learning_rate": 0.0006,
+ "loss": 6.438611030578613,
+ "step": 291
+ },
+ {
+ "epoch": 4.05589519650655,
+ "grad_norm": 0.06845773756504059,
+ "learning_rate": 0.0006,
+ "loss": 6.538361072540283,
+ "step": 292
+ },
+ {
+ "epoch": 4.069868995633188,
+ "grad_norm": 0.053768858313560486,
+ "learning_rate": 0.0006,
+ "loss": 6.527524948120117,
+ "step": 293
+ },
+ {
+ "epoch": 4.083842794759826,
+ "grad_norm": 0.039037931710481644,
+ "learning_rate": 0.0006,
+ "loss": 6.503847122192383,
+ "step": 294
+ },
+ {
+ "epoch": 4.097816593886463,
+ "grad_norm": 0.03962196782231331,
+ "learning_rate": 0.0006,
+ "loss": 6.475178241729736,
+ "step": 295
+ },
+ {
+ "epoch": 4.1117903930131,
+ "grad_norm": 0.049101535230875015,
+ "learning_rate": 0.0006,
+ "loss": 6.474587917327881,
+ "step": 296
+ },
+ {
+ "epoch": 4.125764192139738,
+ "grad_norm": 0.02830282226204872,
+ "learning_rate": 0.0006,
+ "loss": 6.45065975189209,
+ "step": 297
+ },
+ {
+ "epoch": 4.139737991266376,
+ "grad_norm": 0.023153482005000114,
+ "learning_rate": 0.0006,
+ "loss": 6.5218706130981445,
+ "step": 298
+ },
+ {
+ "epoch": 4.153711790393013,
+ "grad_norm": 0.025064002722501755,
+ "learning_rate": 0.0006,
+ "loss": 6.456475257873535,
+ "step": 299
+ },
+ {
+ "epoch": 4.1676855895196505,
+ "grad_norm": 0.02357092685997486,
+ "learning_rate": 0.0006,
+ "loss": 6.485952377319336,
+ "step": 300
+ },
+ {
+ "epoch": 4.181659388646288,
+ "grad_norm": 0.021832305938005447,
+ "learning_rate": 0.0006,
+ "loss": 6.425506591796875,
+ "step": 301
+ },
+ {
+ "epoch": 4.195633187772926,
+ "grad_norm": 0.0227900929749012,
+ "learning_rate": 0.0006,
+ "loss": 6.439155578613281,
+ "step": 302
+ },
+ {
+ "epoch": 4.209606986899563,
+ "grad_norm": 0.015961607918143272,
+ "learning_rate": 0.0006,
+ "loss": 6.460862636566162,
+ "step": 303
+ },
+ {
+ "epoch": 4.223580786026201,
+ "grad_norm": 0.017876390367746353,
+ "learning_rate": 0.0006,
+ "loss": 6.44564151763916,
+ "step": 304
+ },
+ {
+ "epoch": 4.2375545851528384,
+ "grad_norm": 0.013703204691410065,
+ "learning_rate": 0.0006,
+ "loss": 6.418414115905762,
+ "step": 305
+ },
+ {
+ "epoch": 4.251528384279476,
+ "grad_norm": 0.015240306034684181,
+ "learning_rate": 0.0006,
+ "loss": 6.411238670349121,
+ "step": 306
+ },
+ {
+ "epoch": 4.265502183406113,
+ "grad_norm": 0.01587662845849991,
+ "learning_rate": 0.0006,
+ "loss": 6.395642280578613,
+ "step": 307
+ },
+ {
+ "epoch": 4.279475982532751,
+ "grad_norm": 0.014650700613856316,
+ "learning_rate": 0.0006,
+ "loss": 6.302793025970459,
+ "step": 308
+ },
+ {
+ "epoch": 4.293449781659389,
+ "grad_norm": 0.016533225774765015,
+ "learning_rate": 0.0006,
+ "loss": 6.380588054656982,
+ "step": 309
+ },
+ {
+ "epoch": 4.307423580786026,
+ "grad_norm": 0.019963741302490234,
+ "learning_rate": 0.0006,
+ "loss": 6.345336437225342,
+ "step": 310
+ },
+ {
+ "epoch": 4.321397379912664,
+ "grad_norm": 0.02237936295568943,
+ "learning_rate": 0.0006,
+ "loss": 6.387770652770996,
+ "step": 311
+ },
+ {
+ "epoch": 4.335371179039301,
+ "grad_norm": 0.026658328250050545,
+ "learning_rate": 0.0006,
+ "loss": 6.323662757873535,
+ "step": 312
+ },
+ {
+ "epoch": 4.349344978165939,
+ "grad_norm": 0.02852284163236618,
+ "learning_rate": 0.0006,
+ "loss": 6.382204055786133,
+ "step": 313
+ },
+ {
+ "epoch": 4.3633187772925766,
+ "grad_norm": 0.028815090656280518,
+ "learning_rate": 0.0006,
+ "loss": 6.402390480041504,
+ "step": 314
+ },
+ {
+ "epoch": 4.377292576419214,
+ "grad_norm": 0.029393529519438744,
+ "learning_rate": 0.0006,
+ "loss": 6.280278205871582,
+ "step": 315
+ },
+ {
+ "epoch": 4.391266375545851,
+ "grad_norm": 0.02186041697859764,
+ "learning_rate": 0.0006,
+ "loss": 6.302707672119141,
+ "step": 316
+ },
+ {
+ "epoch": 4.405240174672489,
+ "grad_norm": 0.02071218006312847,
+ "learning_rate": 0.0006,
+ "loss": 6.3365020751953125,
+ "step": 317
+ },
+ {
+ "epoch": 4.419213973799127,
+ "grad_norm": 0.02779117226600647,
+ "learning_rate": 0.0006,
+ "loss": 6.252157211303711,
+ "step": 318
+ },
+ {
+ "epoch": 4.4331877729257645,
+ "grad_norm": 0.03470654785633087,
+ "learning_rate": 0.0006,
+ "loss": 6.319070339202881,
+ "step": 319
+ },
+ {
+ "epoch": 4.447161572052401,
+ "grad_norm": 0.037067804485559464,
+ "learning_rate": 0.0006,
+ "loss": 6.314116477966309,
+ "step": 320
+ },
+ {
+ "epoch": 4.461135371179039,
+ "grad_norm": 0.032182756811380386,
+ "learning_rate": 0.0006,
+ "loss": 6.267298698425293,
+ "step": 321
+ },
+ {
+ "epoch": 4.475109170305677,
+ "grad_norm": 0.026305217295885086,
+ "learning_rate": 0.0006,
+ "loss": 6.331688404083252,
+ "step": 322
+ },
+ {
+ "epoch": 4.489082969432315,
+ "grad_norm": 0.027228357270359993,
+ "learning_rate": 0.0006,
+ "loss": 6.31139612197876,
+ "step": 323
+ },
+ {
+ "epoch": 4.503056768558952,
+ "grad_norm": 0.025781169533729553,
+ "learning_rate": 0.0006,
+ "loss": 6.194684982299805,
+ "step": 324
+ },
+ {
+ "epoch": 4.517030567685589,
+ "grad_norm": 0.029186977073550224,
+ "learning_rate": 0.0006,
+ "loss": 6.31306266784668,
+ "step": 325
+ },
+ {
+ "epoch": 4.531004366812227,
+ "grad_norm": 0.023608777672052383,
+ "learning_rate": 0.0006,
+ "loss": 6.285243034362793,
+ "step": 326
+ },
+ {
+ "epoch": 4.544978165938865,
+ "grad_norm": 0.02267594076693058,
+ "learning_rate": 0.0006,
+ "loss": 6.30342960357666,
+ "step": 327
+ },
+ {
+ "epoch": 4.558951965065502,
+ "grad_norm": 0.01926310732960701,
+ "learning_rate": 0.0006,
+ "loss": 6.280606269836426,
+ "step": 328
+ },
+ {
+ "epoch": 4.5729257641921395,
+ "grad_norm": 0.026484837755560875,
+ "learning_rate": 0.0006,
+ "loss": 6.237980365753174,
+ "step": 329
+ },
+ {
+ "epoch": 4.586899563318777,
+ "grad_norm": 0.026228854432702065,
+ "learning_rate": 0.0006,
+ "loss": 6.3018293380737305,
+ "step": 330
+ },
+ {
+ "epoch": 4.600873362445415,
+ "grad_norm": 0.022096967324614525,
+ "learning_rate": 0.0006,
+ "loss": 6.253863334655762,
+ "step": 331
+ },
+ {
+ "epoch": 4.614847161572053,
+ "grad_norm": 0.027223890647292137,
+ "learning_rate": 0.0006,
+ "loss": 6.166882038116455,
+ "step": 332
+ },
+ {
+ "epoch": 4.62882096069869,
+ "grad_norm": 0.03160124272108078,
+ "learning_rate": 0.0006,
+ "loss": 6.230309963226318,
+ "step": 333
+ },
+ {
+ "epoch": 4.642794759825327,
+ "grad_norm": 0.04110539332032204,
+ "learning_rate": 0.0006,
+ "loss": 6.2334818840026855,
+ "step": 334
+ },
+ {
+ "epoch": 4.656768558951965,
+ "grad_norm": 0.04355933889746666,
+ "learning_rate": 0.0006,
+ "loss": 6.160956382751465,
+ "step": 335
+ },
+ {
+ "epoch": 4.670742358078603,
+ "grad_norm": 0.03611086308956146,
+ "learning_rate": 0.0006,
+ "loss": 6.23399543762207,
+ "step": 336
+ },
+ {
+ "epoch": 4.68471615720524,
+ "grad_norm": 0.051811400800943375,
+ "learning_rate": 0.0006,
+ "loss": 6.241855144500732,
+ "step": 337
+ },
+ {
+ "epoch": 4.698689956331878,
+ "grad_norm": 0.057434193789958954,
+ "learning_rate": 0.0006,
+ "loss": 6.220993995666504,
+ "step": 338
+ },
+ {
+ "epoch": 4.712663755458515,
+ "grad_norm": 0.04420953616499901,
+ "learning_rate": 0.0006,
+ "loss": 6.300461292266846,
+ "step": 339
+ },
+ {
+ "epoch": 4.726637554585153,
+ "grad_norm": 0.03515457361936569,
+ "learning_rate": 0.0006,
+ "loss": 6.269349098205566,
+ "step": 340
+ },
+ {
+ "epoch": 4.74061135371179,
+ "grad_norm": 0.03222600743174553,
+ "learning_rate": 0.0006,
+ "loss": 6.248613357543945,
+ "step": 341
+ },
+ {
+ "epoch": 4.754585152838428,
+ "grad_norm": 0.0384046845138073,
+ "learning_rate": 0.0006,
+ "loss": 6.266979217529297,
+ "step": 342
+ },
+ {
+ "epoch": 4.7685589519650655,
+ "grad_norm": 0.0514867827296257,
+ "learning_rate": 0.0006,
+ "loss": 6.234103202819824,
+ "step": 343
+ },
+ {
+ "epoch": 4.782532751091703,
+ "grad_norm": 0.050747793167829514,
+ "learning_rate": 0.0006,
+ "loss": 6.322582244873047,
+ "step": 344
+ },
+ {
+ "epoch": 4.796506550218341,
+ "grad_norm": 0.040878988802433014,
+ "learning_rate": 0.0006,
+ "loss": 6.212509632110596,
+ "step": 345
+ },
+ {
+ "epoch": 4.810480349344978,
+ "grad_norm": 0.034655820578336716,
+ "learning_rate": 0.0006,
+ "loss": 6.265879154205322,
+ "step": 346
+ },
+ {
+ "epoch": 4.824454148471616,
+ "grad_norm": 0.04202224686741829,
+ "learning_rate": 0.0006,
+ "loss": 6.2896409034729,
+ "step": 347
+ },
+ {
+ "epoch": 4.8384279475982535,
+ "grad_norm": 0.03632277995347977,
+ "learning_rate": 0.0006,
+ "loss": 6.253917217254639,
+ "step": 348
+ },
+ {
+ "epoch": 4.85240174672489,
+ "grad_norm": 0.028645765036344528,
+ "learning_rate": 0.0006,
+ "loss": 6.207965850830078,
+ "step": 349
+ },
+ {
+ "epoch": 4.866375545851528,
+ "grad_norm": 0.025305170565843582,
+ "learning_rate": 0.0006,
+ "loss": 6.168068885803223,
+ "step": 350
+ },
+ {
+ "epoch": 4.880349344978166,
+ "grad_norm": 0.022137803956866264,
+ "learning_rate": 0.0006,
+ "loss": 6.235283374786377,
+ "step": 351
+ },
+ {
+ "epoch": 4.894323144104804,
+ "grad_norm": 0.01814538985490799,
+ "learning_rate": 0.0006,
+ "loss": 6.259641647338867,
+ "step": 352
+ },
+ {
+ "epoch": 4.908296943231441,
+ "grad_norm": 0.018465086817741394,
+ "learning_rate": 0.0006,
+ "loss": 6.111738681793213,
+ "step": 353
+ },
+ {
+ "epoch": 4.922270742358078,
+ "grad_norm": 0.018638933077454567,
+ "learning_rate": 0.0006,
+ "loss": 6.225642204284668,
+ "step": 354
+ },
+ {
+ "epoch": 4.936244541484716,
+ "grad_norm": 0.016461260616779327,
+ "learning_rate": 0.0006,
+ "loss": 6.140647888183594,
+ "step": 355
+ },
+ {
+ "epoch": 4.950218340611354,
+ "grad_norm": 0.015458385460078716,
+ "learning_rate": 0.0006,
+ "loss": 6.155224800109863,
+ "step": 356
+ },
+ {
+ "epoch": 4.964192139737992,
+ "grad_norm": 0.014934048056602478,
+ "learning_rate": 0.0006,
+ "loss": 6.101555824279785,
+ "step": 357
+ },
+ {
+ "epoch": 4.978165938864628,
+ "grad_norm": 0.013979545794427395,
+ "learning_rate": 0.0006,
+ "loss": 6.161718845367432,
+ "step": 358
+ },
+ {
+ "epoch": 4.992139737991266,
+ "grad_norm": 0.01640394888818264,
+ "learning_rate": 0.0006,
+ "loss": 6.055768966674805,
+ "step": 359
+ },
+ {
+ "epoch": 5.0,
+ "grad_norm": 0.015135680325329304,
+ "learning_rate": 0.0006,
+ "loss": 6.169695854187012,
+ "step": 360
+ },
+ {
+ "epoch": 5.0,
+ "eval_loss": 6.16408109664917,
+ "eval_runtime": 59.3649,
+ "eval_samples_per_second": 41.135,
+ "eval_steps_per_second": 1.297,
+ "step": 360
+ },
+ {
+ "epoch": 5.013973799126638,
+ "grad_norm": 0.012849048711359501,
+ "learning_rate": 0.0006,
+ "loss": 6.055830001831055,
+ "step": 361
+ },
+ {
+ "epoch": 5.0279475982532755,
+ "grad_norm": 0.018169576302170753,
+ "learning_rate": 0.0006,
+ "loss": 6.082655429840088,
+ "step": 362
+ },
+ {
+ "epoch": 5.041921397379912,
+ "grad_norm": 0.024293430149555206,
+ "learning_rate": 0.0006,
+ "loss": 6.109032154083252,
+ "step": 363
+ },
+ {
+ "epoch": 5.05589519650655,
+ "grad_norm": 0.03266787901520729,
+ "learning_rate": 0.0006,
+ "loss": 6.030613899230957,
+ "step": 364
+ },
+ {
+ "epoch": 5.069868995633188,
+ "grad_norm": 0.03572266176342964,
+ "learning_rate": 0.0006,
+ "loss": 6.050390720367432,
+ "step": 365
+ },
+ {
+ "epoch": 5.083842794759826,
+ "grad_norm": 0.032625213265419006,
+ "learning_rate": 0.0006,
+ "loss": 6.169313430786133,
+ "step": 366
+ },
+ {
+ "epoch": 5.097816593886463,
+ "grad_norm": 0.04051872715353966,
+ "learning_rate": 0.0006,
+ "loss": 6.10398006439209,
+ "step": 367
+ },
+ {
+ "epoch": 5.1117903930131,
+ "grad_norm": 0.0351213738322258,
+ "learning_rate": 0.0006,
+ "loss": 6.121973514556885,
+ "step": 368
+ },
+ {
+ "epoch": 5.125764192139738,
+ "grad_norm": 0.03523759916424751,
+ "learning_rate": 0.0006,
+ "loss": 6.0626654624938965,
+ "step": 369
+ },
+ {
+ "epoch": 5.139737991266376,
+ "grad_norm": 0.02706182189285755,
+ "learning_rate": 0.0006,
+ "loss": 6.049354076385498,
+ "step": 370
+ },
+ {
+ "epoch": 5.153711790393013,
+ "grad_norm": 0.024212589487433434,
+ "learning_rate": 0.0006,
+ "loss": 6.034826278686523,
+ "step": 371
+ },
+ {
+ "epoch": 5.1676855895196505,
+ "grad_norm": 0.024798082187771797,
+ "learning_rate": 0.0006,
+ "loss": 6.004058837890625,
+ "step": 372
+ },
+ {
+ "epoch": 5.181659388646288,
+ "grad_norm": 0.02325567416846752,
+ "learning_rate": 0.0006,
+ "loss": 5.986461639404297,
+ "step": 373
+ },
+ {
+ "epoch": 5.195633187772926,
+ "grad_norm": 0.019060570746660233,
+ "learning_rate": 0.0006,
+ "loss": 6.059736251831055,
+ "step": 374
+ },
+ {
+ "epoch": 5.209606986899563,
+ "grad_norm": 0.016822976991534233,
+ "learning_rate": 0.0006,
+ "loss": 6.072957515716553,
+ "step": 375
+ },
+ {
+ "epoch": 5.223580786026201,
+ "grad_norm": 0.017218658700585365,
+ "learning_rate": 0.0006,
+ "loss": 6.031739234924316,
+ "step": 376
+ },
+ {
+ "epoch": 5.2375545851528384,
+ "grad_norm": 0.014889383688569069,
+ "learning_rate": 0.0006,
+ "loss": 6.025674343109131,
+ "step": 377
+ },
+ {
+ "epoch": 5.251528384279476,
+ "grad_norm": 0.01708107627928257,
+ "learning_rate": 0.0006,
+ "loss": 6.090451240539551,
+ "step": 378
+ },
+ {
+ "epoch": 5.265502183406113,
+ "grad_norm": 0.020260317251086235,
+ "learning_rate": 0.0006,
+ "loss": 6.094257354736328,
+ "step": 379
+ },
+ {
+ "epoch": 5.279475982532751,
+ "grad_norm": 0.020110400393605232,
+ "learning_rate": 0.0006,
+ "loss": 6.021188735961914,
+ "step": 380
+ },
+ {
+ "epoch": 5.293449781659389,
+ "grad_norm": 0.02027830481529236,
+ "learning_rate": 0.0006,
+ "loss": 6.022156238555908,
+ "step": 381
+ },
+ {
+ "epoch": 5.307423580786026,
+ "grad_norm": 0.022747062146663666,
+ "learning_rate": 0.0006,
+ "loss": 6.011836051940918,
+ "step": 382
+ },
+ {
+ "epoch": 5.321397379912664,
+ "grad_norm": 0.01990230567753315,
+ "learning_rate": 0.0006,
+ "loss": 6.003054618835449,
+ "step": 383
+ },
+ {
+ "epoch": 5.335371179039301,
+ "grad_norm": 0.01596238650381565,
+ "learning_rate": 0.0006,
+ "loss": 5.991410732269287,
+ "step": 384
+ },
+ {
+ "epoch": 5.349344978165939,
+ "grad_norm": 0.015847105532884598,
+ "learning_rate": 0.0006,
+ "loss": 6.100622177124023,
+ "step": 385
+ },
+ {
+ "epoch": 5.3633187772925766,
+ "grad_norm": 0.016179397702217102,
+ "learning_rate": 0.0006,
+ "loss": 6.032659530639648,
+ "step": 386
+ },
+ {
+ "epoch": 5.377292576419214,
+ "grad_norm": 0.018256602808833122,
+ "learning_rate": 0.0006,
+ "loss": 5.961983680725098,
+ "step": 387
+ },
+ {
+ "epoch": 5.391266375545851,
+ "grad_norm": 0.02805912122130394,
+ "learning_rate": 0.0006,
+ "loss": 5.983541965484619,
+ "step": 388
+ },
+ {
+ "epoch": 5.405240174672489,
+ "grad_norm": 0.039082255214452744,
+ "learning_rate": 0.0006,
+ "loss": 6.021474361419678,
+ "step": 389
+ },
+ {
+ "epoch": 5.419213973799127,
+ "grad_norm": 0.036757659167051315,
+ "learning_rate": 0.0006,
+ "loss": 6.005046367645264,
+ "step": 390
+ },
+ {
+ "epoch": 5.4331877729257645,
+ "grad_norm": 0.035277366638183594,
+ "learning_rate": 0.0006,
+ "loss": 6.036296844482422,
+ "step": 391
+ },
+ {
+ "epoch": 5.447161572052401,
+ "grad_norm": 0.034404635429382324,
+ "learning_rate": 0.0006,
+ "loss": 5.994539260864258,
+ "step": 392
+ },
+ {
+ "epoch": 5.461135371179039,
+ "grad_norm": 0.0377206988632679,
+ "learning_rate": 0.0006,
+ "loss": 5.941534519195557,
+ "step": 393
+ },
+ {
+ "epoch": 5.475109170305677,
+ "grad_norm": 0.0389922633767128,
+ "learning_rate": 0.0006,
+ "loss": 5.983644485473633,
+ "step": 394
+ },
+ {
+ "epoch": 5.489082969432315,
+ "grad_norm": 0.04176778718829155,
+ "learning_rate": 0.0006,
+ "loss": 6.030922889709473,
+ "step": 395
+ },
+ {
+ "epoch": 5.503056768558952,
+ "grad_norm": 0.0337153784930706,
+ "learning_rate": 0.0006,
+ "loss": 5.961367130279541,
+ "step": 396
+ },
+ {
+ "epoch": 5.517030567685589,
+ "grad_norm": 0.036116816103458405,
+ "learning_rate": 0.0006,
+ "loss": 5.819280624389648,
+ "step": 397
+ },
+ {
+ "epoch": 5.531004366812227,
+ "grad_norm": 0.030724041163921356,
+ "learning_rate": 0.0006,
+ "loss": 5.906380653381348,
+ "step": 398
+ },
+ {
+ "epoch": 5.544978165938865,
+ "grad_norm": 0.030264202505350113,
+ "learning_rate": 0.0006,
+ "loss": 5.975833415985107,
+ "step": 399
+ },
+ {
+ "epoch": 5.558951965065502,
+ "grad_norm": 0.032096318900585175,
+ "learning_rate": 0.0006,
+ "loss": 6.001348972320557,
+ "step": 400
+ },
+ {
+ "epoch": 5.5729257641921395,
+ "grad_norm": 0.030579356476664543,
+ "learning_rate": 0.0006,
+ "loss": 5.928768634796143,
+ "step": 401
+ },
+ {
+ "epoch": 5.586899563318777,
+ "grad_norm": 0.028241973370313644,
+ "learning_rate": 0.0006,
+ "loss": 5.92582893371582,
+ "step": 402
+ },
+ {
+ "epoch": 5.600873362445415,
+ "grad_norm": 0.023086953908205032,
+ "learning_rate": 0.0006,
+ "loss": 5.892926216125488,
+ "step": 403
+ },
+ {
+ "epoch": 5.614847161572053,
+ "grad_norm": 0.02547992393374443,
+ "learning_rate": 0.0006,
+ "loss": 5.903195381164551,
+ "step": 404
+ },
+ {
+ "epoch": 5.62882096069869,
+ "grad_norm": 0.023089831694960594,
+ "learning_rate": 0.0006,
+ "loss": 5.912033557891846,
+ "step": 405
+ },
+ {
+ "epoch": 5.642794759825327,
+ "grad_norm": 0.025446368381381035,
+ "learning_rate": 0.0006,
+ "loss": 5.926138401031494,
+ "step": 406
+ },
+ {
+ "epoch": 5.656768558951965,
+ "grad_norm": 0.03748747706413269,
+ "learning_rate": 0.0006,
+ "loss": 5.913451194763184,
+ "step": 407
+ },
+ {
+ "epoch": 5.670742358078603,
+ "grad_norm": 0.03493810072541237,
+ "learning_rate": 0.0006,
+ "loss": 5.97633695602417,
+ "step": 408
+ },
+ {
+ "epoch": 5.68471615720524,
+ "grad_norm": 0.023131275549530983,
+ "learning_rate": 0.0006,
+ "loss": 5.899571418762207,
+ "step": 409
+ },
+ {
+ "epoch": 5.698689956331878,
+ "grad_norm": 0.02986014075577259,
+ "learning_rate": 0.0006,
+ "loss": 5.894903182983398,
+ "step": 410
+ },
+ {
+ "epoch": 5.712663755458515,
+ "grad_norm": 0.030171144753694534,
+ "learning_rate": 0.0006,
+ "loss": 5.876595497131348,
+ "step": 411
+ },
+ {
+ "epoch": 5.726637554585153,
+ "grad_norm": 0.029377546161413193,
+ "learning_rate": 0.0006,
+ "loss": 5.911637306213379,
+ "step": 412
+ },
+ {
+ "epoch": 5.74061135371179,
+ "grad_norm": 0.029829490929841995,
+ "learning_rate": 0.0006,
+ "loss": 5.819629669189453,
+ "step": 413
+ },
+ {
+ "epoch": 5.754585152838428,
+ "grad_norm": 0.024139299988746643,
+ "learning_rate": 0.0006,
+ "loss": 5.81728458404541,
+ "step": 414
+ },
+ {
+ "epoch": 5.7685589519650655,
+ "grad_norm": 0.021453432738780975,
+ "learning_rate": 0.0006,
+ "loss": 5.987326145172119,
+ "step": 415
+ },
+ {
+ "epoch": 5.782532751091703,
+ "grad_norm": 0.017942246049642563,
+ "learning_rate": 0.0006,
+ "loss": 5.893008232116699,
+ "step": 416
+ },
+ {
+ "epoch": 5.796506550218341,
+ "grad_norm": 0.019724637269973755,
+ "learning_rate": 0.0006,
+ "loss": 5.912441253662109,
+ "step": 417
+ },
+ {
+ "epoch": 5.810480349344978,
+ "grad_norm": 0.018597912043333054,
+ "learning_rate": 0.0006,
+ "loss": 5.860394477844238,
+ "step": 418
+ },
+ {
+ "epoch": 5.824454148471616,
+ "grad_norm": 0.016232412308454514,
+ "learning_rate": 0.0006,
+ "loss": 5.94991397857666,
+ "step": 419
+ },
+ {
+ "epoch": 5.8384279475982535,
+ "grad_norm": 0.01667204685509205,
+ "learning_rate": 0.0006,
+ "loss": 5.8506364822387695,
+ "step": 420
+ },
+ {
+ "epoch": 5.85240174672489,
+ "grad_norm": 0.014232151210308075,
+ "learning_rate": 0.0006,
+ "loss": 5.928424835205078,
+ "step": 421
+ },
+ {
+ "epoch": 5.866375545851528,
+ "grad_norm": 0.01570476032793522,
+ "learning_rate": 0.0006,
+ "loss": 5.953692436218262,
+ "step": 422
+ },
+ {
+ "epoch": 5.880349344978166,
+ "grad_norm": 0.01586179807782173,
+ "learning_rate": 0.0006,
+ "loss": 5.820014476776123,
+ "step": 423
+ },
+ {
+ "epoch": 5.894323144104804,
+ "grad_norm": 0.014729145914316177,
+ "learning_rate": 0.0006,
+ "loss": 5.860154628753662,
+ "step": 424
+ },
+ {
+ "epoch": 5.908296943231441,
+ "grad_norm": 0.013111459091305733,
+ "learning_rate": 0.0006,
+ "loss": 5.85988187789917,
+ "step": 425
+ },
+ {
+ "epoch": 5.922270742358078,
+ "grad_norm": 0.012937922962009907,
+ "learning_rate": 0.0006,
+ "loss": 5.756265640258789,
+ "step": 426
+ },
+ {
+ "epoch": 5.936244541484716,
+ "grad_norm": 0.01377453189343214,
+ "learning_rate": 0.0006,
+ "loss": 5.8333048820495605,
+ "step": 427
+ },
+ {
+ "epoch": 5.950218340611354,
+ "grad_norm": 0.014054795727133751,
+ "learning_rate": 0.0006,
+ "loss": 5.803333759307861,
+ "step": 428
+ },
+ {
+ "epoch": 5.964192139737992,
+ "grad_norm": 0.01674959622323513,
+ "learning_rate": 0.0006,
+ "loss": 5.8718485832214355,
+ "step": 429
+ },
+ {
+ "epoch": 5.978165938864628,
+ "grad_norm": 0.018733017146587372,
+ "learning_rate": 0.0006,
+ "loss": 5.875979900360107,
+ "step": 430
+ },
+ {
+ "epoch": 5.992139737991266,
+ "grad_norm": 0.02088983915746212,
+ "learning_rate": 0.0006,
+ "loss": 5.763634204864502,
+ "step": 431
+ },
+ {
+ "epoch": 6.0,
+ "grad_norm": 0.026705985888838768,
+ "learning_rate": 0.0006,
+ "loss": 5.6922478675842285,
+ "step": 432
+ },
+ {
+ "epoch": 6.0,
+ "eval_loss": 5.847958087921143,
+ "eval_runtime": 58.5451,
+ "eval_samples_per_second": 41.711,
+ "eval_steps_per_second": 1.315,
+ "step": 432
+ },
+ {
+ "epoch": 6.013973799126638,
+ "grad_norm": 0.03181544691324234,
+ "learning_rate": 0.0006,
+ "loss": 5.786368370056152,
+ "step": 433
+ },
+ {
+ "epoch": 6.0279475982532755,
+ "grad_norm": 0.03238112851977348,
+ "learning_rate": 0.0006,
+ "loss": 5.810310363769531,
+ "step": 434
+ },
+ {
+ "epoch": 6.041921397379912,
+ "grad_norm": 0.03916756808757782,
+ "learning_rate": 0.0006,
+ "loss": 5.7761993408203125,
+ "step": 435
+ },
+ {
+ "epoch": 6.05589519650655,
+ "grad_norm": 0.044009730219841,
+ "learning_rate": 0.0006,
+ "loss": 5.898112773895264,
+ "step": 436
+ },
+ {
+ "epoch": 6.069868995633188,
+ "grad_norm": 0.045235246419906616,
+ "learning_rate": 0.0006,
+ "loss": 5.876371383666992,
+ "step": 437
+ },
+ {
+ "epoch": 6.083842794759826,
+ "grad_norm": 0.050218384712934494,
+ "learning_rate": 0.0006,
+ "loss": 5.805103302001953,
+ "step": 438
+ },
+ {
+ "epoch": 6.097816593886463,
+ "grad_norm": 0.05444490164518356,
+ "learning_rate": 0.0006,
+ "loss": 5.826424598693848,
+ "step": 439
+ },
+ {
+ "epoch": 6.1117903930131,
+ "grad_norm": 0.04890982061624527,
+ "learning_rate": 0.0006,
+ "loss": 5.7735395431518555,
+ "step": 440
+ },
+ {
+ "epoch": 6.125764192139738,
+ "grad_norm": 0.05478639155626297,
+ "learning_rate": 0.0006,
+ "loss": 5.823677062988281,
+ "step": 441
+ },
+ {
+ "epoch": 6.139737991266376,
+ "grad_norm": 0.057562313973903656,
+ "learning_rate": 0.0006,
+ "loss": 5.889334678649902,
+ "step": 442
+ },
+ {
+ "epoch": 6.153711790393013,
+ "grad_norm": 0.06447703391313553,
+ "learning_rate": 0.0006,
+ "loss": 5.84621524810791,
+ "step": 443
+ },
+ {
+ "epoch": 6.1676855895196505,
+ "grad_norm": 0.048861872404813766,
+ "learning_rate": 0.0006,
+ "loss": 5.931595802307129,
+ "step": 444
+ },
+ {
+ "epoch": 6.181659388646288,
+ "grad_norm": 0.05521339178085327,
+ "learning_rate": 0.0006,
+ "loss": 5.8229780197143555,
+ "step": 445
+ },
+ {
+ "epoch": 6.195633187772926,
+ "grad_norm": 0.053666990250349045,
+ "learning_rate": 0.0006,
+ "loss": 5.879191875457764,
+ "step": 446
+ },
+ {
+ "epoch": 6.209606986899563,
+ "grad_norm": 0.0451025515794754,
+ "learning_rate": 0.0006,
+ "loss": 5.801628112792969,
+ "step": 447
+ },
+ {
+ "epoch": 6.223580786026201,
+ "grad_norm": 0.04291655868291855,
+ "learning_rate": 0.0006,
+ "loss": 5.914680480957031,
+ "step": 448
+ },
+ {
+ "epoch": 6.2375545851528384,
+ "grad_norm": 0.035102490335702896,
+ "learning_rate": 0.0006,
+ "loss": 5.843682765960693,
+ "step": 449
+ },
+ {
+ "epoch": 6.251528384279476,
+ "grad_norm": 0.03403995931148529,
+ "learning_rate": 0.0006,
+ "loss": 5.827154159545898,
+ "step": 450
+ },
+ {
+ "epoch": 6.265502183406113,
+ "grad_norm": 0.03444375470280647,
+ "learning_rate": 0.0006,
+ "loss": 5.800136566162109,
+ "step": 451
+ },
+ {
+ "epoch": 6.279475982532751,
+ "grad_norm": 0.03165159001946449,
+ "learning_rate": 0.0006,
+ "loss": 5.906252384185791,
+ "step": 452
+ },
+ {
+ "epoch": 6.293449781659389,
+ "grad_norm": 0.026153093203902245,
+ "learning_rate": 0.0006,
+ "loss": 5.7423200607299805,
+ "step": 453
+ },
+ {
+ "epoch": 6.307423580786026,
+ "grad_norm": 0.024357657879590988,
+ "learning_rate": 0.0006,
+ "loss": 5.701364517211914,
+ "step": 454
+ },
+ {
+ "epoch": 6.321397379912664,
+ "grad_norm": 0.021508049219846725,
+ "learning_rate": 0.0006,
+ "loss": 5.8328094482421875,
+ "step": 455
+ },
+ {
+ "epoch": 6.335371179039301,
+ "grad_norm": 0.017313921824097633,
+ "learning_rate": 0.0006,
+ "loss": 5.792145252227783,
+ "step": 456
+ },
+ {
+ "epoch": 6.349344978165939,
+ "grad_norm": 0.018563397228717804,
+ "learning_rate": 0.0006,
+ "loss": 5.732672214508057,
+ "step": 457
+ },
+ {
+ "epoch": 6.3633187772925766,
+ "grad_norm": 0.016568679362535477,
+ "learning_rate": 0.0006,
+ "loss": 5.81948184967041,
+ "step": 458
+ },
+ {
+ "epoch": 6.377292576419214,
+ "grad_norm": 0.014133770950138569,
+ "learning_rate": 0.0006,
+ "loss": 5.734982490539551,
+ "step": 459
+ },
+ {
+ "epoch": 6.391266375545851,
+ "grad_norm": 0.014114447869360447,
+ "learning_rate": 0.0006,
+ "loss": 5.70073127746582,
+ "step": 460
+ },
+ {
+ "epoch": 6.405240174672489,
+ "grad_norm": 0.012907393276691437,
+ "learning_rate": 0.0006,
+ "loss": 5.758626937866211,
+ "step": 461
+ },
+ {
+ "epoch": 6.419213973799127,
+ "grad_norm": 0.012673444114625454,
+ "learning_rate": 0.0006,
+ "loss": 5.775138854980469,
+ "step": 462
+ },
+ {
+ "epoch": 6.4331877729257645,
+ "grad_norm": 0.01311410591006279,
+ "learning_rate": 0.0006,
+ "loss": 5.70920991897583,
+ "step": 463
+ },
+ {
+ "epoch": 6.447161572052401,
+ "grad_norm": 0.012935544364154339,
+ "learning_rate": 0.0006,
+ "loss": 5.781479835510254,
+ "step": 464
+ },
+ {
+ "epoch": 6.461135371179039,
+ "grad_norm": 0.01439738366752863,
+ "learning_rate": 0.0006,
+ "loss": 5.691315650939941,
+ "step": 465
+ },
+ {
+ "epoch": 6.475109170305677,
+ "grad_norm": 0.01205186452716589,
+ "learning_rate": 0.0006,
+ "loss": 5.662499904632568,
+ "step": 466
+ },
+ {
+ "epoch": 6.489082969432315,
+ "grad_norm": 0.011575652286410332,
+ "learning_rate": 0.0006,
+ "loss": 5.6271772384643555,
+ "step": 467
+ },
+ {
+ "epoch": 6.503056768558952,
+ "grad_norm": 0.011672502383589745,
+ "learning_rate": 0.0006,
+ "loss": 5.761662006378174,
+ "step": 468
+ },
+ {
+ "epoch": 6.517030567685589,
+ "grad_norm": 0.011541751213371754,
+ "learning_rate": 0.0006,
+ "loss": 5.762078285217285,
+ "step": 469
+ },
+ {
+ "epoch": 6.531004366812227,
+ "grad_norm": 0.011396365240216255,
+ "learning_rate": 0.0006,
+ "loss": 5.67873477935791,
+ "step": 470
+ },
+ {
+ "epoch": 6.544978165938865,
+ "grad_norm": 0.010688499547541142,
+ "learning_rate": 0.0006,
+ "loss": 5.717051982879639,
+ "step": 471
+ },
+ {
+ "epoch": 6.558951965065502,
+ "grad_norm": 0.012224317528307438,
+ "learning_rate": 0.0006,
+ "loss": 5.707948684692383,
+ "step": 472
+ },
+ {
+ "epoch": 6.5729257641921395,
+ "grad_norm": 0.011856825090944767,
+ "learning_rate": 0.0006,
+ "loss": 5.70878791809082,
+ "step": 473
+ },
+ {
+ "epoch": 6.586899563318777,
+ "grad_norm": 0.01199064590036869,
+ "learning_rate": 0.0006,
+ "loss": 5.708697319030762,
+ "step": 474
+ },
+ {
+ "epoch": 6.600873362445415,
+ "grad_norm": 0.01219446212053299,
+ "learning_rate": 0.0006,
+ "loss": 5.6061577796936035,
+ "step": 475
+ },
+ {
+ "epoch": 6.614847161572053,
+ "grad_norm": 0.013397484086453915,
+ "learning_rate": 0.0006,
+ "loss": 5.624789714813232,
+ "step": 476
+ },
+ {
+ "epoch": 6.62882096069869,
+ "grad_norm": 0.01483136136084795,
+ "learning_rate": 0.0006,
+ "loss": 5.721141338348389,
+ "step": 477
+ },
+ {
+ "epoch": 6.642794759825327,
+ "grad_norm": 0.0194488987326622,
+ "learning_rate": 0.0006,
+ "loss": 5.604279518127441,
+ "step": 478
+ },
+ {
+ "epoch": 6.656768558951965,
+ "grad_norm": 0.02079896256327629,
+ "learning_rate": 0.0006,
+ "loss": 5.586322784423828,
+ "step": 479
+ },
+ {
+ "epoch": 6.670742358078603,
+ "grad_norm": 0.018996229395270348,
+ "learning_rate": 0.0006,
+ "loss": 5.5987772941589355,
+ "step": 480
+ },
+ {
+ "epoch": 6.68471615720524,
+ "grad_norm": 0.015779603272676468,
+ "learning_rate": 0.0006,
+ "loss": 5.588602066040039,
+ "step": 481
+ },
+ {
+ "epoch": 6.698689956331878,
+ "grad_norm": 0.015322973020374775,
+ "learning_rate": 0.0006,
+ "loss": 5.686467170715332,
+ "step": 482
+ },
+ {
+ "epoch": 6.712663755458515,
+ "grad_norm": 0.015282213687896729,
+ "learning_rate": 0.0006,
+ "loss": 5.664676666259766,
+ "step": 483
+ },
+ {
+ "epoch": 6.726637554585153,
+ "grad_norm": 0.016589272767305374,
+ "learning_rate": 0.0006,
+ "loss": 5.666739463806152,
+ "step": 484
+ },
+ {
+ "epoch": 6.74061135371179,
+ "grad_norm": 0.01811421848833561,
+ "learning_rate": 0.0006,
+ "loss": 5.647593021392822,
+ "step": 485
+ },
+ {
+ "epoch": 6.754585152838428,
+ "grad_norm": 0.019652029499411583,
+ "learning_rate": 0.0006,
+ "loss": 5.628726959228516,
+ "step": 486
+ },
+ {
+ "epoch": 6.7685589519650655,
+ "grad_norm": 0.02118665538728237,
+ "learning_rate": 0.0006,
+ "loss": 5.6580705642700195,
+ "step": 487
+ },
+ {
+ "epoch": 6.782532751091703,
+ "grad_norm": 0.02237832546234131,
+ "learning_rate": 0.0006,
+ "loss": 5.635931968688965,
+ "step": 488
+ },
+ {
+ "epoch": 6.796506550218341,
+ "grad_norm": 0.023897631093859673,
+ "learning_rate": 0.0006,
+ "loss": 5.517949104309082,
+ "step": 489
+ },
+ {
+ "epoch": 6.810480349344978,
+ "grad_norm": 0.02442227490246296,
+ "learning_rate": 0.0006,
+ "loss": 5.600021839141846,
+ "step": 490
+ },
+ {
+ "epoch": 6.824454148471616,
+ "grad_norm": 0.024675287306308746,
+ "learning_rate": 0.0006,
+ "loss": 5.6639909744262695,
+ "step": 491
+ },
+ {
+ "epoch": 6.8384279475982535,
+ "grad_norm": 0.030372392386198044,
+ "learning_rate": 0.0006,
+ "loss": 5.582888603210449,
+ "step": 492
+ },
+ {
+ "epoch": 6.85240174672489,
+ "grad_norm": 0.03237690031528473,
+ "learning_rate": 0.0006,
+ "loss": 5.573896884918213,
+ "step": 493
+ },
+ {
+ "epoch": 6.866375545851528,
+ "grad_norm": 0.03511589393019676,
+ "learning_rate": 0.0006,
+ "loss": 5.565878868103027,
+ "step": 494
+ },
+ {
+ "epoch": 6.880349344978166,
+ "grad_norm": 0.03883938118815422,
+ "learning_rate": 0.0006,
+ "loss": 5.58540678024292,
+ "step": 495
+ },
+ {
+ "epoch": 6.894323144104804,
+ "grad_norm": 0.04175502806901932,
+ "learning_rate": 0.0006,
+ "loss": 5.6095194816589355,
+ "step": 496
+ },
+ {
+ "epoch": 6.908296943231441,
+ "grad_norm": 0.04013441503047943,
+ "learning_rate": 0.0006,
+ "loss": 5.654401779174805,
+ "step": 497
+ },
+ {
+ "epoch": 6.922270742358078,
+ "grad_norm": 0.04618770629167557,
+ "learning_rate": 0.0006,
+ "loss": 5.6483473777771,
+ "step": 498
+ },
+ {
+ "epoch": 6.936244541484716,
+ "grad_norm": 0.04507308453321457,
+ "learning_rate": 0.0006,
+ "loss": 5.5856218338012695,
+ "step": 499
+ },
+ {
+ "epoch": 6.950218340611354,
+ "grad_norm": 0.03151383996009827,
+ "learning_rate": 0.0006,
+ "loss": 5.594086647033691,
+ "step": 500
+ },
+ {
+ "epoch": 6.964192139737992,
+ "grad_norm": 0.026883797720074654,
+ "learning_rate": 0.0006,
+ "loss": 5.607676029205322,
+ "step": 501
+ },
+ {
+ "epoch": 6.978165938864628,
+ "grad_norm": 0.02981836162507534,
+ "learning_rate": 0.0006,
+ "loss": 5.563666820526123,
+ "step": 502
+ },
+ {
+ "epoch": 6.992139737991266,
+ "grad_norm": 0.03233477845788002,
+ "learning_rate": 0.0006,
+ "loss": 5.667543411254883,
+ "step": 503
+ },
+ {
+ "epoch": 7.0,
+ "grad_norm": 0.029927456751465797,
+ "learning_rate": 0.0006,
+ "loss": 5.733482837677002,
+ "step": 504
+ },
+ {
+ "epoch": 7.0,
+ "eval_loss": 5.679076671600342,
+ "eval_runtime": 59.5965,
+ "eval_samples_per_second": 40.976,
+ "eval_steps_per_second": 1.292,
+ "step": 504
+ },
+ {
+ "epoch": 7.013973799126638,
+ "grad_norm": 0.027081597596406937,
+ "learning_rate": 0.0006,
+ "loss": 5.634487152099609,
+ "step": 505
+ },
+ {
+ "epoch": 7.0279475982532755,
+ "grad_norm": 0.031959421932697296,
+ "learning_rate": 0.0006,
+ "loss": 5.643294811248779,
+ "step": 506
+ },
+ {
+ "epoch": 7.041921397379912,
+ "grad_norm": 0.02802063524723053,
+ "learning_rate": 0.0006,
+ "loss": 5.573239326477051,
+ "step": 507
+ },
+ {
+ "epoch": 7.05589519650655,
+ "grad_norm": 0.032579705119132996,
+ "learning_rate": 0.0006,
+ "loss": 5.627150058746338,
+ "step": 508
+ },
+ {
+ "epoch": 7.069868995633188,
+ "grad_norm": 0.02503928542137146,
+ "learning_rate": 0.0006,
+ "loss": 5.6045308113098145,
+ "step": 509
+ },
+ {
+ "epoch": 7.083842794759826,
+ "grad_norm": 0.02421317622065544,
+ "learning_rate": 0.0006,
+ "loss": 5.522153377532959,
+ "step": 510
+ },
+ {
+ "epoch": 7.097816593886463,
+ "grad_norm": 0.021983426064252853,
+ "learning_rate": 0.0006,
+ "loss": 5.632939338684082,
+ "step": 511
+ },
+ {
+ "epoch": 7.1117903930131,
+ "grad_norm": 0.021933183073997498,
+ "learning_rate": 0.0006,
+ "loss": 5.592185974121094,
+ "step": 512
+ },
+ {
+ "epoch": 7.125764192139738,
+ "grad_norm": 0.02287031151354313,
+ "learning_rate": 0.0006,
+ "loss": 5.613700866699219,
+ "step": 513
+ },
+ {
+ "epoch": 7.139737991266376,
+ "grad_norm": 0.02281602844595909,
+ "learning_rate": 0.0006,
+ "loss": 5.551383018493652,
+ "step": 514
+ },
+ {
+ "epoch": 7.153711790393013,
+ "grad_norm": 0.021050360053777695,
+ "learning_rate": 0.0006,
+ "loss": 5.55275821685791,
+ "step": 515
+ },
+ {
+ "epoch": 7.1676855895196505,
+ "grad_norm": 0.018299926072359085,
+ "learning_rate": 0.0006,
+ "loss": 5.528225421905518,
+ "step": 516
+ },
+ {
+ "epoch": 7.181659388646288,
+ "grad_norm": 0.02024853602051735,
+ "learning_rate": 0.0006,
+ "loss": 5.522453784942627,
+ "step": 517
+ },
+ {
+ "epoch": 7.195633187772926,
+ "grad_norm": 0.02135239914059639,
+ "learning_rate": 0.0006,
+ "loss": 5.527522563934326,
+ "step": 518
+ },
+ {
+ "epoch": 7.209606986899563,
+ "grad_norm": 0.023394184187054634,
+ "learning_rate": 0.0006,
+ "loss": 5.4535017013549805,
+ "step": 519
+ },
+ {
+ "epoch": 7.223580786026201,
+ "grad_norm": 0.025952080264687538,
+ "learning_rate": 0.0006,
+ "loss": 5.5374674797058105,
+ "step": 520
+ },
+ {
+ "epoch": 7.2375545851528384,
+ "grad_norm": 0.022287718951702118,
+ "learning_rate": 0.0006,
+ "loss": 5.493753433227539,
+ "step": 521
+ },
+ {
+ "epoch": 7.251528384279476,
+ "grad_norm": 0.02018447406589985,
+ "learning_rate": 0.0006,
+ "loss": 5.510575294494629,
+ "step": 522
+ },
+ {
+ "epoch": 7.265502183406113,
+ "grad_norm": 0.02050507813692093,
+ "learning_rate": 0.0006,
+ "loss": 5.466026782989502,
+ "step": 523
+ },
+ {
+ "epoch": 7.279475982532751,
+ "grad_norm": 0.023688802495598793,
+ "learning_rate": 0.0006,
+ "loss": 5.485815525054932,
+ "step": 524
+ },
+ {
+ "epoch": 7.293449781659389,
+ "grad_norm": 0.021078843623399734,
+ "learning_rate": 0.0006,
+ "loss": 5.415581703186035,
+ "step": 525
+ },
+ {
+ "epoch": 7.307423580786026,
+ "grad_norm": 0.018266011029481888,
+ "learning_rate": 0.0006,
+ "loss": 5.402815818786621,
+ "step": 526
+ },
+ {
+ "epoch": 7.321397379912664,
+ "grad_norm": 0.019693493843078613,
+ "learning_rate": 0.0006,
+ "loss": 5.504674911499023,
+ "step": 527
+ },
+ {
+ "epoch": 7.335371179039301,
+ "grad_norm": 0.02031373605132103,
+ "learning_rate": 0.0006,
+ "loss": 5.503837585449219,
+ "step": 528
+ },
+ {
+ "epoch": 7.349344978165939,
+ "grad_norm": 0.017636626958847046,
+ "learning_rate": 0.0006,
+ "loss": 5.530580997467041,
+ "step": 529
+ },
+ {
+ "epoch": 7.3633187772925766,
+ "grad_norm": 0.01787244901061058,
+ "learning_rate": 0.0006,
+ "loss": 5.450218677520752,
+ "step": 530
+ },
+ {
+ "epoch": 7.377292576419214,
+ "grad_norm": 0.0170669574290514,
+ "learning_rate": 0.0006,
+ "loss": 5.481570243835449,
+ "step": 531
+ },
+ {
+ "epoch": 7.391266375545851,
+ "grad_norm": 0.01802165061235428,
+ "learning_rate": 0.0006,
+ "loss": 5.386394500732422,
+ "step": 532
+ },
+ {
+ "epoch": 7.405240174672489,
+ "grad_norm": 0.022949472069740295,
+ "learning_rate": 0.0006,
+ "loss": 5.424929618835449,
+ "step": 533
+ },
+ {
+ "epoch": 7.419213973799127,
+ "grad_norm": 0.034213390201330185,
+ "learning_rate": 0.0006,
+ "loss": 5.346663475036621,
+ "step": 534
+ },
+ {
+ "epoch": 7.4331877729257645,
+ "grad_norm": 0.03847593814134598,
+ "learning_rate": 0.0006,
+ "loss": 5.418482780456543,
+ "step": 535
+ },
+ {
+ "epoch": 7.447161572052401,
+ "grad_norm": 0.029393676668405533,
+ "learning_rate": 0.0006,
+ "loss": 5.456090927124023,
+ "step": 536
+ },
+ {
+ "epoch": 7.461135371179039,
+ "grad_norm": 0.03388667106628418,
+ "learning_rate": 0.0006,
+ "loss": 5.572983741760254,
+ "step": 537
+ },
+ {
+ "epoch": 7.475109170305677,
+ "grad_norm": 0.039690613746643066,
+ "learning_rate": 0.0006,
+ "loss": 5.414067268371582,
+ "step": 538
+ },
+ {
+ "epoch": 7.489082969432315,
+ "grad_norm": 0.03630776330828667,
+ "learning_rate": 0.0006,
+ "loss": 5.523739814758301,
+ "step": 539
+ },
+ {
+ "epoch": 7.503056768558952,
+ "grad_norm": 0.03356870263814926,
+ "learning_rate": 0.0006,
+ "loss": 5.444767951965332,
+ "step": 540
+ },
+ {
+ "epoch": 7.517030567685589,
+ "grad_norm": 0.030812304466962814,
+ "learning_rate": 0.0006,
+ "loss": 5.461244106292725,
+ "step": 541
+ },
+ {
+ "epoch": 7.531004366812227,
+ "grad_norm": 0.03319217637181282,
+ "learning_rate": 0.0006,
+ "loss": 5.502161026000977,
+ "step": 542
+ },
+ {
+ "epoch": 7.544978165938865,
+ "grad_norm": 0.032997481524944305,
+ "learning_rate": 0.0006,
+ "loss": 5.4162278175354,
+ "step": 543
+ },
+ {
+ "epoch": 7.558951965065502,
+ "grad_norm": 0.03364962339401245,
+ "learning_rate": 0.0006,
+ "loss": 5.415736198425293,
+ "step": 544
+ },
+ {
+ "epoch": 7.5729257641921395,
+ "grad_norm": 0.0344221405684948,
+ "learning_rate": 0.0006,
+ "loss": 5.541967868804932,
+ "step": 545
+ },
+ {
+ "epoch": 7.586899563318777,
+ "grad_norm": 0.029609503224492073,
+ "learning_rate": 0.0006,
+ "loss": 5.440614223480225,
+ "step": 546
+ },
+ {
+ "epoch": 7.600873362445415,
+ "grad_norm": 0.029055926948785782,
+ "learning_rate": 0.0006,
+ "loss": 5.462865352630615,
+ "step": 547
+ },
+ {
+ "epoch": 7.614847161572053,
+ "grad_norm": 0.02658848837018013,
+ "learning_rate": 0.0006,
+ "loss": 5.444467544555664,
+ "step": 548
+ },
+ {
+ "epoch": 7.62882096069869,
+ "grad_norm": 0.026276404038071632,
+ "learning_rate": 0.0006,
+ "loss": 5.462278842926025,
+ "step": 549
+ },
+ {
+ "epoch": 7.642794759825327,
+ "grad_norm": 0.0286889486014843,
+ "learning_rate": 0.0006,
+ "loss": 5.4849534034729,
+ "step": 550
+ },
+ {
+ "epoch": 7.656768558951965,
+ "grad_norm": 0.026849817484617233,
+ "learning_rate": 0.0006,
+ "loss": 5.431473731994629,
+ "step": 551
+ },
+ {
+ "epoch": 7.670742358078603,
+ "grad_norm": 0.02312396466732025,
+ "learning_rate": 0.0006,
+ "loss": 5.45462703704834,
+ "step": 552
+ },
+ {
+ "epoch": 7.68471615720524,
+ "grad_norm": 0.026197485625743866,
+ "learning_rate": 0.0006,
+ "loss": 5.452552318572998,
+ "step": 553
+ },
+ {
+ "epoch": 7.698689956331878,
+ "grad_norm": 0.02747255191206932,
+ "learning_rate": 0.0006,
+ "loss": 5.447169780731201,
+ "step": 554
+ },
+ {
+ "epoch": 7.712663755458515,
+ "grad_norm": 0.028123432770371437,
+ "learning_rate": 0.0006,
+ "loss": 5.394400596618652,
+ "step": 555
+ },
+ {
+ "epoch": 7.726637554585153,
+ "grad_norm": 0.02599870041012764,
+ "learning_rate": 0.0006,
+ "loss": 5.398341178894043,
+ "step": 556
+ },
+ {
+ "epoch": 7.74061135371179,
+ "grad_norm": 0.022171657532453537,
+ "learning_rate": 0.0006,
+ "loss": 5.368820667266846,
+ "step": 557
+ },
+ {
+ "epoch": 7.754585152838428,
+ "grad_norm": 0.022309480234980583,
+ "learning_rate": 0.0006,
+ "loss": 5.416772365570068,
+ "step": 558
+ },
+ {
+ "epoch": 7.7685589519650655,
+ "grad_norm": 0.024074165150523186,
+ "learning_rate": 0.0006,
+ "loss": 5.359002113342285,
+ "step": 559
+ },
+ {
+ "epoch": 7.782532751091703,
+ "grad_norm": 0.02636653557419777,
+ "learning_rate": 0.0006,
+ "loss": 5.484368324279785,
+ "step": 560
+ },
+ {
+ "epoch": 7.796506550218341,
+ "grad_norm": 0.02203752100467682,
+ "learning_rate": 0.0006,
+ "loss": 5.375498294830322,
+ "step": 561
+ },
+ {
+ "epoch": 7.810480349344978,
+ "grad_norm": 0.019924819469451904,
+ "learning_rate": 0.0006,
+ "loss": 5.334723472595215,
+ "step": 562
+ },
+ {
+ "epoch": 7.824454148471616,
+ "grad_norm": 0.018755966797471046,
+ "learning_rate": 0.0006,
+ "loss": 5.348781108856201,
+ "step": 563
+ },
+ {
+ "epoch": 7.8384279475982535,
+ "grad_norm": 0.016747845336794853,
+ "learning_rate": 0.0006,
+ "loss": 5.433079719543457,
+ "step": 564
+ },
+ {
+ "epoch": 7.85240174672489,
+ "grad_norm": 0.0170395877212286,
+ "learning_rate": 0.0006,
+ "loss": 5.346794128417969,
+ "step": 565
+ },
+ {
+ "epoch": 7.866375545851528,
+ "grad_norm": 0.016074176877737045,
+ "learning_rate": 0.0006,
+ "loss": 5.3277740478515625,
+ "step": 566
+ },
+ {
+ "epoch": 7.880349344978166,
+ "grad_norm": 0.014119806699454784,
+ "learning_rate": 0.0006,
+ "loss": 5.483551979064941,
+ "step": 567
+ },
+ {
+ "epoch": 7.894323144104804,
+ "grad_norm": 0.015270394273102283,
+ "learning_rate": 0.0006,
+ "loss": 5.378519535064697,
+ "step": 568
+ },
+ {
+ "epoch": 7.908296943231441,
+ "grad_norm": 0.015677539631724358,
+ "learning_rate": 0.0006,
+ "loss": 5.290210247039795,
+ "step": 569
+ },
+ {
+ "epoch": 7.922270742358078,
+ "grad_norm": 0.015930423513054848,
+ "learning_rate": 0.0006,
+ "loss": 5.427360534667969,
+ "step": 570
+ },
+ {
+ "epoch": 7.936244541484716,
+ "grad_norm": 0.016641564667224884,
+ "learning_rate": 0.0006,
+ "loss": 5.301599979400635,
+ "step": 571
+ },
+ {
+ "epoch": 7.950218340611354,
+ "grad_norm": 0.018293552100658417,
+ "learning_rate": 0.0006,
+ "loss": 5.277889251708984,
+ "step": 572
+ },
+ {
+ "epoch": 7.964192139737992,
+ "grad_norm": 0.017618799582123756,
+ "learning_rate": 0.0006,
+ "loss": 5.297194480895996,
+ "step": 573
+ },
+ {
+ "epoch": 7.978165938864628,
+ "grad_norm": 0.015034242533147335,
+ "learning_rate": 0.0006,
+ "loss": 5.42428731918335,
+ "step": 574
+ },
+ {
+ "epoch": 7.992139737991266,
+ "grad_norm": 0.01617511734366417,
+ "learning_rate": 0.0006,
+ "loss": 5.420044898986816,
+ "step": 575
+ },
+ {
+ "epoch": 8.0,
+ "grad_norm": 0.01837257295846939,
+ "learning_rate": 0.0006,
+ "loss": 5.240387916564941,
+ "step": 576
+ },
+ {
+ "epoch": 8.0,
+ "eval_loss": 5.423932075500488,
+ "eval_runtime": 58.0841,
+ "eval_samples_per_second": 42.042,
+ "eval_steps_per_second": 1.326,
+ "step": 576
+ },
+ {
+ "epoch": 8.013973799126637,
+ "grad_norm": 0.020953809842467308,
+ "learning_rate": 0.0006,
+ "loss": 5.4320783615112305,
+ "step": 577
+ },
+ {
+ "epoch": 8.027947598253276,
+ "grad_norm": 0.0227745920419693,
+ "learning_rate": 0.0006,
+ "loss": 5.310197830200195,
+ "step": 578
+ },
+ {
+ "epoch": 8.041921397379912,
+ "grad_norm": 0.021372603252530098,
+ "learning_rate": 0.0006,
+ "loss": 5.339468002319336,
+ "step": 579
+ },
+ {
+ "epoch": 8.055895196506551,
+ "grad_norm": 0.022401731461286545,
+ "learning_rate": 0.0006,
+ "loss": 5.279829978942871,
+ "step": 580
+ },
+ {
+ "epoch": 8.069868995633188,
+ "grad_norm": 0.023942379280924797,
+ "learning_rate": 0.0006,
+ "loss": 5.372323989868164,
+ "step": 581
+ },
+ {
+ "epoch": 8.083842794759825,
+ "grad_norm": 0.024380534887313843,
+ "learning_rate": 0.0006,
+ "loss": 5.253486633300781,
+ "step": 582
+ },
+ {
+ "epoch": 8.097816593886463,
+ "grad_norm": 0.03246625140309334,
+ "learning_rate": 0.0006,
+ "loss": 5.305774211883545,
+ "step": 583
+ },
+ {
+ "epoch": 8.1117903930131,
+ "grad_norm": 0.030399736016988754,
+ "learning_rate": 0.0006,
+ "loss": 5.370976448059082,
+ "step": 584
+ },
+ {
+ "epoch": 8.125764192139737,
+ "grad_norm": 0.027148913592100143,
+ "learning_rate": 0.0006,
+ "loss": 5.368010520935059,
+ "step": 585
+ },
+ {
+ "epoch": 8.139737991266376,
+ "grad_norm": 0.02925540879368782,
+ "learning_rate": 0.0006,
+ "loss": 5.303013801574707,
+ "step": 586
+ },
+ {
+ "epoch": 8.153711790393013,
+ "grad_norm": 0.02915453538298607,
+ "learning_rate": 0.0006,
+ "loss": 5.329667091369629,
+ "step": 587
+ },
+ {
+ "epoch": 8.167685589519651,
+ "grad_norm": 0.031732626259326935,
+ "learning_rate": 0.0006,
+ "loss": 5.3634934425354,
+ "step": 588
+ },
+ {
+ "epoch": 8.181659388646288,
+ "grad_norm": 0.036166731268167496,
+ "learning_rate": 0.0006,
+ "loss": 5.223980903625488,
+ "step": 589
+ },
+ {
+ "epoch": 8.195633187772925,
+ "grad_norm": 0.035654108971357346,
+ "learning_rate": 0.0006,
+ "loss": 5.361059188842773,
+ "step": 590
+ },
+ {
+ "epoch": 8.209606986899564,
+ "grad_norm": 0.0365324430167675,
+ "learning_rate": 0.0006,
+ "loss": 5.336151599884033,
+ "step": 591
+ },
+ {
+ "epoch": 8.2235807860262,
+ "grad_norm": 0.03578880429267883,
+ "learning_rate": 0.0006,
+ "loss": 5.434076309204102,
+ "step": 592
+ },
+ {
+ "epoch": 8.237554585152838,
+ "grad_norm": 0.035734012722969055,
+ "learning_rate": 0.0006,
+ "loss": 5.330893516540527,
+ "step": 593
+ },
+ {
+ "epoch": 8.251528384279476,
+ "grad_norm": 0.03196857124567032,
+ "learning_rate": 0.0006,
+ "loss": 5.306009769439697,
+ "step": 594
+ },
+ {
+ "epoch": 8.265502183406113,
+ "grad_norm": 0.032042086124420166,
+ "learning_rate": 0.0006,
+ "loss": 5.343267917633057,
+ "step": 595
+ },
+ {
+ "epoch": 8.279475982532752,
+ "grad_norm": 0.03160746395587921,
+ "learning_rate": 0.0006,
+ "loss": 5.2353715896606445,
+ "step": 596
+ },
+ {
+ "epoch": 8.293449781659389,
+ "grad_norm": 0.026689818128943443,
+ "learning_rate": 0.0006,
+ "loss": 5.254042625427246,
+ "step": 597
+ },
+ {
+ "epoch": 8.307423580786025,
+ "grad_norm": 0.02880188450217247,
+ "learning_rate": 0.0006,
+ "loss": 5.326833248138428,
+ "step": 598
+ },
+ {
+ "epoch": 8.321397379912664,
+ "grad_norm": 0.027516381815075874,
+ "learning_rate": 0.0006,
+ "loss": 5.342336177825928,
+ "step": 599
+ },
+ {
+ "epoch": 8.335371179039301,
+ "grad_norm": 0.027875030413269997,
+ "learning_rate": 0.0006,
+ "loss": 5.200719833374023,
+ "step": 600
+ },
+ {
+ "epoch": 8.34934497816594,
+ "grad_norm": 0.0268265288323164,
+ "learning_rate": 0.0006,
+ "loss": 5.249449729919434,
+ "step": 601
+ },
+ {
+ "epoch": 8.363318777292577,
+ "grad_norm": 0.024824608117341995,
+ "learning_rate": 0.0006,
+ "loss": 5.318952560424805,
+ "step": 602
+ },
+ {
+ "epoch": 8.377292576419213,
+ "grad_norm": 0.022990627214312553,
+ "learning_rate": 0.0006,
+ "loss": 5.244993209838867,
+ "step": 603
+ },
+ {
+ "epoch": 8.391266375545852,
+ "grad_norm": 0.022804604843258858,
+ "learning_rate": 0.0006,
+ "loss": 5.3236846923828125,
+ "step": 604
+ },
+ {
+ "epoch": 8.405240174672489,
+ "grad_norm": 0.02419872209429741,
+ "learning_rate": 0.0006,
+ "loss": 5.255486488342285,
+ "step": 605
+ },
+ {
+ "epoch": 8.419213973799126,
+ "grad_norm": 0.021952059119939804,
+ "learning_rate": 0.0006,
+ "loss": 5.3511552810668945,
+ "step": 606
+ },
+ {
+ "epoch": 8.433187772925764,
+ "grad_norm": 0.022375497967004776,
+ "learning_rate": 0.0006,
+ "loss": 5.294790267944336,
+ "step": 607
+ },
+ {
+ "epoch": 8.447161572052401,
+ "grad_norm": 0.019242815673351288,
+ "learning_rate": 0.0006,
+ "loss": 5.252618789672852,
+ "step": 608
+ },
+ {
+ "epoch": 8.46113537117904,
+ "grad_norm": 0.0173486340790987,
+ "learning_rate": 0.0006,
+ "loss": 5.30147647857666,
+ "step": 609
+ },
+ {
+ "epoch": 8.475109170305677,
+ "grad_norm": 0.01612604409456253,
+ "learning_rate": 0.0006,
+ "loss": 5.23960542678833,
+ "step": 610
+ },
+ {
+ "epoch": 8.489082969432314,
+ "grad_norm": 0.014293976128101349,
+ "learning_rate": 0.0006,
+ "loss": 5.119932651519775,
+ "step": 611
+ },
+ {
+ "epoch": 8.503056768558952,
+ "grad_norm": 0.014453536830842495,
+ "learning_rate": 0.0006,
+ "loss": 5.166794776916504,
+ "step": 612
+ },
+ {
+ "epoch": 8.51703056768559,
+ "grad_norm": 0.014685769565403461,
+ "learning_rate": 0.0006,
+ "loss": 5.264822959899902,
+ "step": 613
+ },
+ {
+ "epoch": 8.531004366812226,
+ "grad_norm": 0.01427740603685379,
+ "learning_rate": 0.0006,
+ "loss": 5.278433799743652,
+ "step": 614
+ },
+ {
+ "epoch": 8.544978165938865,
+ "grad_norm": 0.014864951372146606,
+ "learning_rate": 0.0006,
+ "loss": 5.214512348175049,
+ "step": 615
+ },
+ {
+ "epoch": 8.558951965065502,
+ "grad_norm": 0.01565164513885975,
+ "learning_rate": 0.0006,
+ "loss": 5.183066368103027,
+ "step": 616
+ },
+ {
+ "epoch": 8.57292576419214,
+ "grad_norm": 0.01779816672205925,
+ "learning_rate": 0.0006,
+ "loss": 5.163185119628906,
+ "step": 617
+ },
+ {
+ "epoch": 8.586899563318777,
+ "grad_norm": 0.017254827544093132,
+ "learning_rate": 0.0006,
+ "loss": 5.243555068969727,
+ "step": 618
+ },
+ {
+ "epoch": 8.600873362445414,
+ "grad_norm": 0.01650584116578102,
+ "learning_rate": 0.0006,
+ "loss": 5.234884738922119,
+ "step": 619
+ },
+ {
+ "epoch": 8.614847161572053,
+ "grad_norm": 0.017021115869283676,
+ "learning_rate": 0.0006,
+ "loss": 5.12271785736084,
+ "step": 620
+ },
+ {
+ "epoch": 8.62882096069869,
+ "grad_norm": 0.01773759163916111,
+ "learning_rate": 0.0006,
+ "loss": 5.242153167724609,
+ "step": 621
+ },
+ {
+ "epoch": 8.642794759825328,
+ "grad_norm": 0.015679042786359787,
+ "learning_rate": 0.0006,
+ "loss": 5.170779228210449,
+ "step": 622
+ },
+ {
+ "epoch": 8.656768558951965,
+ "grad_norm": 0.013760549947619438,
+ "learning_rate": 0.0006,
+ "loss": 5.238644599914551,
+ "step": 623
+ },
+ {
+ "epoch": 8.670742358078602,
+ "grad_norm": 0.014571522362530231,
+ "learning_rate": 0.0006,
+ "loss": 5.177056312561035,
+ "step": 624
+ },
+ {
+ "epoch": 8.68471615720524,
+ "grad_norm": 0.016209015622735023,
+ "learning_rate": 0.0006,
+ "loss": 5.163750648498535,
+ "step": 625
+ },
+ {
+ "epoch": 8.698689956331878,
+ "grad_norm": 0.016813941299915314,
+ "learning_rate": 0.0006,
+ "loss": 5.183428764343262,
+ "step": 626
+ },
+ {
+ "epoch": 8.712663755458514,
+ "grad_norm": 0.020985357463359833,
+ "learning_rate": 0.0006,
+ "loss": 5.229465007781982,
+ "step": 627
+ },
+ {
+ "epoch": 8.726637554585153,
+ "grad_norm": 0.02679632417857647,
+ "learning_rate": 0.0006,
+ "loss": 5.257368087768555,
+ "step": 628
+ },
+ {
+ "epoch": 8.74061135371179,
+ "grad_norm": 0.025756070390343666,
+ "learning_rate": 0.0006,
+ "loss": 5.253736972808838,
+ "step": 629
+ },
+ {
+ "epoch": 8.754585152838429,
+ "grad_norm": 0.02643490768969059,
+ "learning_rate": 0.0006,
+ "loss": 5.228633403778076,
+ "step": 630
+ },
+ {
+ "epoch": 8.768558951965066,
+ "grad_norm": 0.03303210437297821,
+ "learning_rate": 0.0006,
+ "loss": 5.163540363311768,
+ "step": 631
+ },
+ {
+ "epoch": 8.782532751091702,
+ "grad_norm": 0.030432431027293205,
+ "learning_rate": 0.0006,
+ "loss": 5.243169784545898,
+ "step": 632
+ },
+ {
+ "epoch": 8.796506550218341,
+ "grad_norm": 0.028631288558244705,
+ "learning_rate": 0.0006,
+ "loss": 5.166398048400879,
+ "step": 633
+ },
+ {
+ "epoch": 8.810480349344978,
+ "grad_norm": 0.026188310235738754,
+ "learning_rate": 0.0006,
+ "loss": 5.128122329711914,
+ "step": 634
+ },
+ {
+ "epoch": 8.824454148471617,
+ "grad_norm": 0.025308528915047646,
+ "learning_rate": 0.0006,
+ "loss": 5.211284637451172,
+ "step": 635
+ },
+ {
+ "epoch": 8.838427947598253,
+ "grad_norm": 0.020657729357481003,
+ "learning_rate": 0.0006,
+ "loss": 5.161575794219971,
+ "step": 636
+ },
+ {
+ "epoch": 8.85240174672489,
+ "grad_norm": 0.021901234984397888,
+ "learning_rate": 0.0006,
+ "loss": 5.20050048828125,
+ "step": 637
+ },
+ {
+ "epoch": 8.866375545851529,
+ "grad_norm": 0.021219318732619286,
+ "learning_rate": 0.0006,
+ "loss": 5.211699485778809,
+ "step": 638
+ },
+ {
+ "epoch": 8.880349344978166,
+ "grad_norm": 0.0200974028557539,
+ "learning_rate": 0.0006,
+ "loss": 5.092503547668457,
+ "step": 639
+ },
+ {
+ "epoch": 8.894323144104803,
+ "grad_norm": 0.023804882541298866,
+ "learning_rate": 0.0006,
+ "loss": 5.216068267822266,
+ "step": 640
+ },
+ {
+ "epoch": 8.908296943231441,
+ "grad_norm": 0.026088010519742966,
+ "learning_rate": 0.0006,
+ "loss": 5.155592918395996,
+ "step": 641
+ },
+ {
+ "epoch": 8.922270742358078,
+ "grad_norm": 0.02601276896893978,
+ "learning_rate": 0.0006,
+ "loss": 5.197238922119141,
+ "step": 642
+ },
+ {
+ "epoch": 8.936244541484717,
+ "grad_norm": 0.020387211814522743,
+ "learning_rate": 0.0006,
+ "loss": 5.1955437660217285,
+ "step": 643
+ },
+ {
+ "epoch": 8.950218340611354,
+ "grad_norm": 0.019214622676372528,
+ "learning_rate": 0.0006,
+ "loss": 5.206346035003662,
+ "step": 644
+ },
+ {
+ "epoch": 8.96419213973799,
+ "grad_norm": 0.019674314185976982,
+ "learning_rate": 0.0006,
+ "loss": 5.219857215881348,
+ "step": 645
+ },
+ {
+ "epoch": 8.97816593886463,
+ "grad_norm": 0.020318234339356422,
+ "learning_rate": 0.0006,
+ "loss": 5.191132545471191,
+ "step": 646
+ },
+ {
+ "epoch": 8.992139737991266,
+ "grad_norm": 0.021428676322102547,
+ "learning_rate": 0.0006,
+ "loss": 5.214695930480957,
+ "step": 647
+ },
+ {
+ "epoch": 9.0,
+ "grad_norm": 0.023235054686665535,
+ "learning_rate": 0.0006,
+ "loss": 5.229091644287109,
+ "step": 648
+ },
+ {
+ "epoch": 9.0,
+ "eval_loss": 5.289593696594238,
+ "eval_runtime": 56.9256,
+ "eval_samples_per_second": 42.898,
+ "eval_steps_per_second": 1.353,
+ "step": 648
+ },
+ {
+ "epoch": 9.013973799126637,
+ "grad_norm": 0.02320289984345436,
+ "learning_rate": 0.0006,
+ "loss": 4.983669281005859,
+ "step": 649
+ },
+ {
+ "epoch": 9.027947598253276,
+ "grad_norm": 0.018521282821893692,
+ "learning_rate": 0.0006,
+ "loss": 5.200575828552246,
+ "step": 650
+ },
+ {
+ "epoch": 9.041921397379912,
+ "grad_norm": 0.0180169939994812,
+ "learning_rate": 0.0006,
+ "loss": 5.148033618927002,
+ "step": 651
+ },
+ {
+ "epoch": 9.055895196506551,
+ "grad_norm": 0.019969860091805458,
+ "learning_rate": 0.0006,
+ "loss": 5.099125385284424,
+ "step": 652
+ },
+ {
+ "epoch": 9.069868995633188,
+ "grad_norm": 0.01643305830657482,
+ "learning_rate": 0.0006,
+ "loss": 5.092347145080566,
+ "step": 653
+ },
+ {
+ "epoch": 9.083842794759825,
+ "grad_norm": 0.016022363677620888,
+ "learning_rate": 0.0006,
+ "loss": 5.129279613494873,
+ "step": 654
+ },
+ {
+ "epoch": 9.097816593886463,
+ "grad_norm": 0.01657041162252426,
+ "learning_rate": 0.0006,
+ "loss": 5.159562587738037,
+ "step": 655
+ },
+ {
+ "epoch": 9.1117903930131,
+ "grad_norm": 0.017924228683114052,
+ "learning_rate": 0.0006,
+ "loss": 5.165497779846191,
+ "step": 656
+ },
+ {
+ "epoch": 9.125764192139737,
+ "grad_norm": 0.021592361852526665,
+ "learning_rate": 0.0006,
+ "loss": 5.204775810241699,
+ "step": 657
+ },
+ {
+ "epoch": 9.139737991266376,
+ "grad_norm": 0.026924120262265205,
+ "learning_rate": 0.0006,
+ "loss": 5.128296852111816,
+ "step": 658
+ },
+ {
+ "epoch": 9.153711790393013,
+ "grad_norm": 0.023156899958848953,
+ "learning_rate": 0.0006,
+ "loss": 5.1255340576171875,
+ "step": 659
+ },
+ {
+ "epoch": 9.167685589519651,
+ "grad_norm": 0.020670367404818535,
+ "learning_rate": 0.0006,
+ "loss": 5.128479957580566,
+ "step": 660
+ },
+ {
+ "epoch": 9.181659388646288,
+ "grad_norm": 0.025447173044085503,
+ "learning_rate": 0.0006,
+ "loss": 5.154821872711182,
+ "step": 661
+ },
+ {
+ "epoch": 9.195633187772925,
+ "grad_norm": 0.027940068393945694,
+ "learning_rate": 0.0006,
+ "loss": 5.143466472625732,
+ "step": 662
+ },
+ {
+ "epoch": 9.209606986899564,
+ "grad_norm": 0.026847844943404198,
+ "learning_rate": 0.0006,
+ "loss": 5.045665740966797,
+ "step": 663
+ },
+ {
+ "epoch": 9.2235807860262,
+ "grad_norm": 0.02340601570904255,
+ "learning_rate": 0.0006,
+ "loss": 5.21933126449585,
+ "step": 664
+ },
+ {
+ "epoch": 9.237554585152838,
+ "grad_norm": 0.02340371161699295,
+ "learning_rate": 0.0006,
+ "loss": 4.942949295043945,
+ "step": 665
+ },
+ {
+ "epoch": 9.251528384279476,
+ "grad_norm": 0.02221992425620556,
+ "learning_rate": 0.0006,
+ "loss": 5.197099208831787,
+ "step": 666
+ },
+ {
+ "epoch": 9.265502183406113,
+ "grad_norm": 0.023508677259087563,
+ "learning_rate": 0.0006,
+ "loss": 5.022897720336914,
+ "step": 667
+ },
+ {
+ "epoch": 9.279475982532752,
+ "grad_norm": 0.026120394468307495,
+ "learning_rate": 0.0006,
+ "loss": 5.087725639343262,
+ "step": 668
+ },
+ {
+ "epoch": 9.293449781659389,
+ "grad_norm": 0.027273228392004967,
+ "learning_rate": 0.0006,
+ "loss": 5.090963840484619,
+ "step": 669
+ },
+ {
+ "epoch": 9.307423580786025,
+ "grad_norm": 0.03241586685180664,
+ "learning_rate": 0.0006,
+ "loss": 5.047842979431152,
+ "step": 670
+ },
+ {
+ "epoch": 9.321397379912664,
+ "grad_norm": 0.030245667323470116,
+ "learning_rate": 0.0006,
+ "loss": 5.104803085327148,
+ "step": 671
+ },
+ {
+ "epoch": 9.335371179039301,
+ "grad_norm": 0.027698364108800888,
+ "learning_rate": 0.0006,
+ "loss": 5.0644097328186035,
+ "step": 672
+ },
+ {
+ "epoch": 9.34934497816594,
+ "grad_norm": 0.029692554846405983,
+ "learning_rate": 0.0006,
+ "loss": 5.111942291259766,
+ "step": 673
+ },
+ {
+ "epoch": 9.363318777292577,
+ "grad_norm": 0.03328656405210495,
+ "learning_rate": 0.0006,
+ "loss": 5.1574554443359375,
+ "step": 674
+ },
+ {
+ "epoch": 9.377292576419213,
+ "grad_norm": 0.031938180327415466,
+ "learning_rate": 0.0006,
+ "loss": 5.190371990203857,
+ "step": 675
+ },
+ {
+ "epoch": 9.391266375545852,
+ "grad_norm": 0.024858945980668068,
+ "learning_rate": 0.0006,
+ "loss": 5.121420860290527,
+ "step": 676
+ },
+ {
+ "epoch": 9.405240174672489,
+ "grad_norm": 0.023033203557133675,
+ "learning_rate": 0.0006,
+ "loss": 5.0974345207214355,
+ "step": 677
+ },
+ {
+ "epoch": 9.419213973799126,
+ "grad_norm": 0.022083545103669167,
+ "learning_rate": 0.0006,
+ "loss": 5.046319961547852,
+ "step": 678
+ },
+ {
+ "epoch": 9.433187772925764,
+ "grad_norm": 0.0228437427431345,
+ "learning_rate": 0.0006,
+ "loss": 5.119963645935059,
+ "step": 679
+ },
+ {
+ "epoch": 9.447161572052401,
+ "grad_norm": 0.0242351982742548,
+ "learning_rate": 0.0006,
+ "loss": 5.0444560050964355,
+ "step": 680
+ },
+ {
+ "epoch": 9.46113537117904,
+ "grad_norm": 0.02401185780763626,
+ "learning_rate": 0.0006,
+ "loss": 5.063238620758057,
+ "step": 681
+ },
+ {
+ "epoch": 9.475109170305677,
+ "grad_norm": 0.02560959942638874,
+ "learning_rate": 0.0006,
+ "loss": 5.035224914550781,
+ "step": 682
+ },
+ {
+ "epoch": 9.489082969432314,
+ "grad_norm": 0.024985626339912415,
+ "learning_rate": 0.0006,
+ "loss": 5.049284934997559,
+ "step": 683
+ },
+ {
+ "epoch": 9.503056768558952,
+ "grad_norm": 0.023178115487098694,
+ "learning_rate": 0.0006,
+ "loss": 5.2266621589660645,
+ "step": 684
+ },
+ {
+ "epoch": 9.51703056768559,
+ "grad_norm": 0.0215512253344059,
+ "learning_rate": 0.0006,
+ "loss": 5.1875715255737305,
+ "step": 685
+ },
+ {
+ "epoch": 9.531004366812226,
+ "grad_norm": 0.01662837713956833,
+ "learning_rate": 0.0006,
+ "loss": 5.137012481689453,
+ "step": 686
+ },
+ {
+ "epoch": 9.544978165938865,
+ "grad_norm": 0.01952764391899109,
+ "learning_rate": 0.0006,
+ "loss": 5.072500228881836,
+ "step": 687
+ },
+ {
+ "epoch": 9.558951965065502,
+ "grad_norm": 0.017521077767014503,
+ "learning_rate": 0.0006,
+ "loss": 5.045711994171143,
+ "step": 688
+ },
+ {
+ "epoch": 9.57292576419214,
+ "grad_norm": 0.017355265095829964,
+ "learning_rate": 0.0006,
+ "loss": 5.02881383895874,
+ "step": 689
+ },
+ {
+ "epoch": 9.586899563318777,
+ "grad_norm": 0.015587719157338142,
+ "learning_rate": 0.0006,
+ "loss": 5.132787704467773,
+ "step": 690
+ },
+ {
+ "epoch": 9.600873362445414,
+ "grad_norm": 0.017469845712184906,
+ "learning_rate": 0.0006,
+ "loss": 5.105447292327881,
+ "step": 691
+ },
+ {
+ "epoch": 9.614847161572053,
+ "grad_norm": 0.019630100578069687,
+ "learning_rate": 0.0006,
+ "loss": 5.032464981079102,
+ "step": 692
+ },
+ {
+ "epoch": 9.62882096069869,
+ "grad_norm": 0.01784994639456272,
+ "learning_rate": 0.0006,
+ "loss": 5.081345558166504,
+ "step": 693
+ },
+ {
+ "epoch": 9.642794759825328,
+ "grad_norm": 0.01872754842042923,
+ "learning_rate": 0.0006,
+ "loss": 4.9772114753723145,
+ "step": 694
+ },
+ {
+ "epoch": 9.656768558951965,
+ "grad_norm": 0.01999032311141491,
+ "learning_rate": 0.0006,
+ "loss": 5.089740753173828,
+ "step": 695
+ },
+ {
+ "epoch": 9.670742358078602,
+ "grad_norm": 0.01710767112672329,
+ "learning_rate": 0.0006,
+ "loss": 4.9869890213012695,
+ "step": 696
+ },
+ {
+ "epoch": 9.68471615720524,
+ "grad_norm": 0.015144161880016327,
+ "learning_rate": 0.0006,
+ "loss": 5.022564888000488,
+ "step": 697
+ },
+ {
+ "epoch": 9.698689956331878,
+ "grad_norm": 0.01618543453514576,
+ "learning_rate": 0.0006,
+ "loss": 5.035921573638916,
+ "step": 698
+ },
+ {
+ "epoch": 9.712663755458514,
+ "grad_norm": 0.014653387479484081,
+ "learning_rate": 0.0006,
+ "loss": 4.914456367492676,
+ "step": 699
+ },
+ {
+ "epoch": 9.726637554585153,
+ "grad_norm": 0.013434977270662785,
+ "learning_rate": 0.0006,
+ "loss": 5.063764572143555,
+ "step": 700
+ },
+ {
+ "epoch": 9.74061135371179,
+ "grad_norm": 0.015128864906728268,
+ "learning_rate": 0.0006,
+ "loss": 5.052367687225342,
+ "step": 701
+ },
+ {
+ "epoch": 9.754585152838429,
+ "grad_norm": 0.014210768043994904,
+ "learning_rate": 0.0006,
+ "loss": 5.016489028930664,
+ "step": 702
+ },
+ {
+ "epoch": 9.768558951965066,
+ "grad_norm": 0.015191680751740932,
+ "learning_rate": 0.0006,
+ "loss": 4.990176200866699,
+ "step": 703
+ },
+ {
+ "epoch": 9.782532751091702,
+ "grad_norm": 0.018285127356648445,
+ "learning_rate": 0.0006,
+ "loss": 5.015164375305176,
+ "step": 704
+ },
+ {
+ "epoch": 9.796506550218341,
+ "grad_norm": 0.0204079058021307,
+ "learning_rate": 0.0006,
+ "loss": 5.091142654418945,
+ "step": 705
+ },
+ {
+ "epoch": 9.810480349344978,
+ "grad_norm": 0.01980586163699627,
+ "learning_rate": 0.0006,
+ "loss": 5.03350830078125,
+ "step": 706
+ },
+ {
+ "epoch": 9.824454148471617,
+ "grad_norm": 0.02048729732632637,
+ "learning_rate": 0.0006,
+ "loss": 5.03611421585083,
+ "step": 707
+ },
+ {
+ "epoch": 9.838427947598253,
+ "grad_norm": 0.02058163844048977,
+ "learning_rate": 0.0006,
+ "loss": 4.9613447189331055,
+ "step": 708
+ },
+ {
+ "epoch": 9.85240174672489,
+ "grad_norm": 0.018004924058914185,
+ "learning_rate": 0.0006,
+ "loss": 4.966541767120361,
+ "step": 709
+ },
+ {
+ "epoch": 9.866375545851529,
+ "grad_norm": 0.0198152307420969,
+ "learning_rate": 0.0006,
+ "loss": 5.01750373840332,
+ "step": 710
+ },
+ {
+ "epoch": 9.880349344978166,
+ "grad_norm": 0.018532969057559967,
+ "learning_rate": 0.0006,
+ "loss": 5.097424507141113,
+ "step": 711
+ },
+ {
+ "epoch": 9.894323144104803,
+ "grad_norm": 0.018939971923828125,
+ "learning_rate": 0.0006,
+ "loss": 5.177223205566406,
+ "step": 712
+ },
+ {
+ "epoch": 9.908296943231441,
+ "grad_norm": 0.017969170585274696,
+ "learning_rate": 0.0006,
+ "loss": 5.118013858795166,
+ "step": 713
+ },
+ {
+ "epoch": 9.922270742358078,
+ "grad_norm": 0.018021270632743835,
+ "learning_rate": 0.0006,
+ "loss": 5.034787178039551,
+ "step": 714
+ },
+ {
+ "epoch": 9.936244541484717,
+ "grad_norm": 0.017009710893034935,
+ "learning_rate": 0.0006,
+ "loss": 5.003920555114746,
+ "step": 715
+ },
+ {
+ "epoch": 9.950218340611354,
+ "grad_norm": 0.017722049728035927,
+ "learning_rate": 0.0006,
+ "loss": 5.031939506530762,
+ "step": 716
+ },
+ {
+ "epoch": 9.96419213973799,
+ "grad_norm": 0.017134087160229683,
+ "learning_rate": 0.0006,
+ "loss": 5.093474388122559,
+ "step": 717
+ },
+ {
+ "epoch": 9.97816593886463,
+ "grad_norm": 0.018511613830924034,
+ "learning_rate": 0.0006,
+ "loss": 5.078604698181152,
+ "step": 718
+ },
+ {
+ "epoch": 9.992139737991266,
+ "grad_norm": 0.021206334233283997,
+ "learning_rate": 0.0006,
+ "loss": 5.052572727203369,
+ "step": 719
+ },
+ {
+ "epoch": 10.0,
+ "grad_norm": 0.02239062264561653,
+ "learning_rate": 0.0006,
+ "loss": 5.063547134399414,
+ "step": 720
+ },
+ {
+ "epoch": 10.0,
+ "eval_loss": 5.173044681549072,
+ "eval_runtime": 56.7287,
+ "eval_samples_per_second": 43.047,
+ "eval_steps_per_second": 1.357,
+ "step": 720
+ },
+ {
+ "epoch": 10.013973799126637,
+ "grad_norm": 0.020757969468832016,
+ "learning_rate": 0.0006,
+ "loss": 5.045648574829102,
+ "step": 721
+ },
+ {
+ "epoch": 10.027947598253276,
+ "grad_norm": 0.022697702050209045,
+ "learning_rate": 0.0006,
+ "loss": 4.993833065032959,
+ "step": 722
+ },
+ {
+ "epoch": 10.041921397379912,
+ "grad_norm": 0.02434331737458706,
+ "learning_rate": 0.0006,
+ "loss": 5.073219299316406,
+ "step": 723
+ },
+ {
+ "epoch": 10.055895196506551,
+ "grad_norm": 0.025499660521745682,
+ "learning_rate": 0.0006,
+ "loss": 5.019072532653809,
+ "step": 724
+ },
+ {
+ "epoch": 10.069868995633188,
+ "grad_norm": 0.029281053692102432,
+ "learning_rate": 0.0006,
+ "loss": 4.959851264953613,
+ "step": 725
+ },
+ {
+ "epoch": 10.083842794759825,
+ "grad_norm": 0.033878110349178314,
+ "learning_rate": 0.0006,
+ "loss": 5.132717609405518,
+ "step": 726
+ },
+ {
+ "epoch": 10.097816593886463,
+ "grad_norm": 0.03385764732956886,
+ "learning_rate": 0.0006,
+ "loss": 5.0482892990112305,
+ "step": 727
+ },
+ {
+ "epoch": 10.1117903930131,
+ "grad_norm": 0.03466648980975151,
+ "learning_rate": 0.0006,
+ "loss": 4.946855545043945,
+ "step": 728
+ },
+ {
+ "epoch": 10.125764192139737,
+ "grad_norm": 0.028546305373311043,
+ "learning_rate": 0.0006,
+ "loss": 5.027774333953857,
+ "step": 729
+ },
+ {
+ "epoch": 10.139737991266376,
+ "grad_norm": 0.030575869604945183,
+ "learning_rate": 0.0006,
+ "loss": 4.9768476486206055,
+ "step": 730
+ },
+ {
+ "epoch": 10.153711790393013,
+ "grad_norm": 0.03614303469657898,
+ "learning_rate": 0.0006,
+ "loss": 5.078113555908203,
+ "step": 731
+ },
+ {
+ "epoch": 10.167685589519651,
+ "grad_norm": 0.040991660207509995,
+ "learning_rate": 0.0006,
+ "loss": 5.033185958862305,
+ "step": 732
+ },
+ {
+ "epoch": 10.181659388646288,
+ "grad_norm": 0.04551699757575989,
+ "learning_rate": 0.0006,
+ "loss": 5.0820746421813965,
+ "step": 733
+ },
+ {
+ "epoch": 10.195633187772925,
+ "grad_norm": 0.040193989872932434,
+ "learning_rate": 0.0006,
+ "loss": 5.059700965881348,
+ "step": 734
+ },
+ {
+ "epoch": 10.209606986899564,
+ "grad_norm": 0.035851605236530304,
+ "learning_rate": 0.0006,
+ "loss": 5.142467498779297,
+ "step": 735
+ },
+ {
+ "epoch": 10.2235807860262,
+ "grad_norm": 0.034200169146060944,
+ "learning_rate": 0.0006,
+ "loss": 5.072312355041504,
+ "step": 736
+ },
+ {
+ "epoch": 10.237554585152838,
+ "grad_norm": 0.033621110022068024,
+ "learning_rate": 0.0006,
+ "loss": 5.08037805557251,
+ "step": 737
+ },
+ {
+ "epoch": 10.251528384279476,
+ "grad_norm": 0.037372369319200516,
+ "learning_rate": 0.0006,
+ "loss": 4.998425483703613,
+ "step": 738
+ },
+ {
+ "epoch": 10.265502183406113,
+ "grad_norm": 0.037044707685709,
+ "learning_rate": 0.0006,
+ "loss": 4.923993110656738,
+ "step": 739
+ },
+ {
+ "epoch": 10.279475982532752,
+ "grad_norm": 0.030454808846116066,
+ "learning_rate": 0.0006,
+ "loss": 5.0889997482299805,
+ "step": 740
+ },
+ {
+ "epoch": 10.293449781659389,
+ "grad_norm": 0.030969638377428055,
+ "learning_rate": 0.0006,
+ "loss": 4.984112739562988,
+ "step": 741
+ },
+ {
+ "epoch": 10.307423580786025,
+ "grad_norm": 0.0276983380317688,
+ "learning_rate": 0.0006,
+ "loss": 4.955412864685059,
+ "step": 742
+ },
+ {
+ "epoch": 10.321397379912664,
+ "grad_norm": 0.02716052532196045,
+ "learning_rate": 0.0006,
+ "loss": 5.0098876953125,
+ "step": 743
+ },
+ {
+ "epoch": 10.335371179039301,
+ "grad_norm": 0.02403552085161209,
+ "learning_rate": 0.0006,
+ "loss": 5.165160655975342,
+ "step": 744
+ },
+ {
+ "epoch": 10.34934497816594,
+ "grad_norm": 0.0257862601429224,
+ "learning_rate": 0.0006,
+ "loss": 5.001547813415527,
+ "step": 745
+ },
+ {
+ "epoch": 10.363318777292577,
+ "grad_norm": 0.02300378680229187,
+ "learning_rate": 0.0006,
+ "loss": 5.038888931274414,
+ "step": 746
+ },
+ {
+ "epoch": 10.377292576419213,
+ "grad_norm": 0.019766854122281075,
+ "learning_rate": 0.0006,
+ "loss": 4.924291610717773,
+ "step": 747
+ },
+ {
+ "epoch": 10.391266375545852,
+ "grad_norm": 0.01888395845890045,
+ "learning_rate": 0.0006,
+ "loss": 5.059971332550049,
+ "step": 748
+ },
+ {
+ "epoch": 10.405240174672489,
+ "grad_norm": 0.015577499754726887,
+ "learning_rate": 0.0006,
+ "loss": 4.987686634063721,
+ "step": 749
+ },
+ {
+ "epoch": 10.419213973799126,
+ "grad_norm": 0.016012346372008324,
+ "learning_rate": 0.0006,
+ "loss": 5.015157699584961,
+ "step": 750
+ },
+ {
+ "epoch": 10.433187772925764,
+ "grad_norm": 0.016277998685836792,
+ "learning_rate": 0.0006,
+ "loss": 5.004931449890137,
+ "step": 751
+ },
+ {
+ "epoch": 10.447161572052401,
+ "grad_norm": 0.017504561692476273,
+ "learning_rate": 0.0006,
+ "loss": 4.902827739715576,
+ "step": 752
+ },
+ {
+ "epoch": 10.46113537117904,
+ "grad_norm": 0.01569022797048092,
+ "learning_rate": 0.0006,
+ "loss": 4.954863548278809,
+ "step": 753
+ },
+ {
+ "epoch": 10.475109170305677,
+ "grad_norm": 0.014114422723650932,
+ "learning_rate": 0.0006,
+ "loss": 5.00656795501709,
+ "step": 754
+ },
+ {
+ "epoch": 10.489082969432314,
+ "grad_norm": 0.01573832333087921,
+ "learning_rate": 0.0006,
+ "loss": 4.978353023529053,
+ "step": 755
+ },
+ {
+ "epoch": 10.503056768558952,
+ "grad_norm": 0.016480036079883575,
+ "learning_rate": 0.0006,
+ "loss": 4.953327655792236,
+ "step": 756
+ },
+ {
+ "epoch": 10.51703056768559,
+ "grad_norm": 0.014008025638759136,
+ "learning_rate": 0.0006,
+ "loss": 5.050840377807617,
+ "step": 757
+ },
+ {
+ "epoch": 10.531004366812226,
+ "grad_norm": 0.013147998601198196,
+ "learning_rate": 0.0006,
+ "loss": 4.974964141845703,
+ "step": 758
+ },
+ {
+ "epoch": 10.544978165938865,
+ "grad_norm": 0.013807603158056736,
+ "learning_rate": 0.0006,
+ "loss": 4.927907943725586,
+ "step": 759
+ },
+ {
+ "epoch": 10.558951965065502,
+ "grad_norm": 0.013555224984884262,
+ "learning_rate": 0.0006,
+ "loss": 4.980690956115723,
+ "step": 760
+ },
+ {
+ "epoch": 10.57292576419214,
+ "grad_norm": 0.013045408762991428,
+ "learning_rate": 0.0006,
+ "loss": 4.9436116218566895,
+ "step": 761
+ },
+ {
+ "epoch": 10.586899563318777,
+ "grad_norm": 0.012858862057328224,
+ "learning_rate": 0.0006,
+ "loss": 4.9595770835876465,
+ "step": 762
+ },
+ {
+ "epoch": 10.600873362445414,
+ "grad_norm": 0.01468253880739212,
+ "learning_rate": 0.0006,
+ "loss": 4.90725040435791,
+ "step": 763
+ },
+ {
+ "epoch": 10.614847161572053,
+ "grad_norm": 0.013224679045379162,
+ "learning_rate": 0.0006,
+ "loss": 4.877615928649902,
+ "step": 764
+ },
+ {
+ "epoch": 10.62882096069869,
+ "grad_norm": 0.013555348850786686,
+ "learning_rate": 0.0006,
+ "loss": 4.9456610679626465,
+ "step": 765
+ },
+ {
+ "epoch": 10.642794759825328,
+ "grad_norm": 0.014387160539627075,
+ "learning_rate": 0.0006,
+ "loss": 4.9233717918396,
+ "step": 766
+ },
+ {
+ "epoch": 10.656768558951965,
+ "grad_norm": 0.014354088343679905,
+ "learning_rate": 0.0006,
+ "loss": 4.983511924743652,
+ "step": 767
+ },
+ {
+ "epoch": 10.670742358078602,
+ "grad_norm": 0.01332700252532959,
+ "learning_rate": 0.0006,
+ "loss": 4.934416770935059,
+ "step": 768
+ },
+ {
+ "epoch": 10.68471615720524,
+ "grad_norm": 0.011543313041329384,
+ "learning_rate": 0.0006,
+ "loss": 4.96388053894043,
+ "step": 769
+ },
+ {
+ "epoch": 10.698689956331878,
+ "grad_norm": 0.012330878525972366,
+ "learning_rate": 0.0006,
+ "loss": 4.946396350860596,
+ "step": 770
+ },
+ {
+ "epoch": 10.712663755458514,
+ "grad_norm": 0.01379779726266861,
+ "learning_rate": 0.0006,
+ "loss": 4.981019020080566,
+ "step": 771
+ },
+ {
+ "epoch": 10.726637554585153,
+ "grad_norm": 0.014612431637942791,
+ "learning_rate": 0.0006,
+ "loss": 4.894567966461182,
+ "step": 772
+ },
+ {
+ "epoch": 10.74061135371179,
+ "grad_norm": 0.013062899932265282,
+ "learning_rate": 0.0006,
+ "loss": 5.033038139343262,
+ "step": 773
+ },
+ {
+ "epoch": 10.754585152838429,
+ "grad_norm": 0.013898049481213093,
+ "learning_rate": 0.0006,
+ "loss": 4.823636054992676,
+ "step": 774
+ },
+ {
+ "epoch": 10.768558951965066,
+ "grad_norm": 0.01502019353210926,
+ "learning_rate": 0.0006,
+ "loss": 5.020156383514404,
+ "step": 775
+ },
+ {
+ "epoch": 10.782532751091702,
+ "grad_norm": 0.013810204342007637,
+ "learning_rate": 0.0006,
+ "loss": 4.935647010803223,
+ "step": 776
+ },
+ {
+ "epoch": 10.796506550218341,
+ "grad_norm": 0.013616513460874557,
+ "learning_rate": 0.0006,
+ "loss": 4.901486396789551,
+ "step": 777
+ },
+ {
+ "epoch": 10.810480349344978,
+ "grad_norm": 0.013601024635136127,
+ "learning_rate": 0.0006,
+ "loss": 4.931873321533203,
+ "step": 778
+ },
+ {
+ "epoch": 10.824454148471617,
+ "grad_norm": 0.013668350875377655,
+ "learning_rate": 0.0006,
+ "loss": 4.937256813049316,
+ "step": 779
+ },
+ {
+ "epoch": 10.838427947598253,
+ "grad_norm": 0.01581823080778122,
+ "learning_rate": 0.0006,
+ "loss": 4.9611406326293945,
+ "step": 780
+ },
+ {
+ "epoch": 10.85240174672489,
+ "grad_norm": 0.019478484988212585,
+ "learning_rate": 0.0006,
+ "loss": 4.9212236404418945,
+ "step": 781
+ },
+ {
+ "epoch": 10.866375545851529,
+ "grad_norm": 0.026921333745121956,
+ "learning_rate": 0.0006,
+ "loss": 4.88501501083374,
+ "step": 782
+ },
+ {
+ "epoch": 10.880349344978166,
+ "grad_norm": 0.03155914694070816,
+ "learning_rate": 0.0006,
+ "loss": 4.989967346191406,
+ "step": 783
+ },
+ {
+ "epoch": 10.894323144104803,
+ "grad_norm": 0.029689429327845573,
+ "learning_rate": 0.0006,
+ "loss": 4.936000823974609,
+ "step": 784
+ },
+ {
+ "epoch": 10.908296943231441,
+ "grad_norm": 0.026626063510775566,
+ "learning_rate": 0.0006,
+ "loss": 4.984785079956055,
+ "step": 785
+ },
+ {
+ "epoch": 10.922270742358078,
+ "grad_norm": 0.026168212294578552,
+ "learning_rate": 0.0006,
+ "loss": 4.908282279968262,
+ "step": 786
+ },
+ {
+ "epoch": 10.936244541484717,
+ "grad_norm": 0.02828158251941204,
+ "learning_rate": 0.0006,
+ "loss": 4.926390647888184,
+ "step": 787
+ },
+ {
+ "epoch": 10.950218340611354,
+ "grad_norm": 0.02649078331887722,
+ "learning_rate": 0.0006,
+ "loss": 5.00400972366333,
+ "step": 788
+ },
+ {
+ "epoch": 10.96419213973799,
+ "grad_norm": 0.0256856270134449,
+ "learning_rate": 0.0006,
+ "loss": 5.0317487716674805,
+ "step": 789
+ },
+ {
+ "epoch": 10.97816593886463,
+ "grad_norm": 0.024728883057832718,
+ "learning_rate": 0.0006,
+ "loss": 4.9299635887146,
+ "step": 790
+ },
+ {
+ "epoch": 10.992139737991266,
+ "grad_norm": 0.026241116225719452,
+ "learning_rate": 0.0006,
+ "loss": 4.947319984436035,
+ "step": 791
+ },
+ {
+ "epoch": 11.0,
+ "grad_norm": 0.023526743054389954,
+ "learning_rate": 0.0006,
+ "loss": 4.94429349899292,
+ "step": 792
+ },
+ {
+ "epoch": 11.0,
+ "eval_loss": 5.104345798492432,
+ "eval_runtime": 57.0955,
+ "eval_samples_per_second": 42.77,
+ "eval_steps_per_second": 1.349,
+ "step": 792
+ },
+ {
+ "epoch": 11.013973799126637,
+ "grad_norm": 0.0216788612306118,
+ "learning_rate": 0.0006,
+ "loss": 4.881950855255127,
+ "step": 793
+ },
+ {
+ "epoch": 11.027947598253276,
+ "grad_norm": 0.01830463856458664,
+ "learning_rate": 0.0006,
+ "loss": 4.971016883850098,
+ "step": 794
+ },
+ {
+ "epoch": 11.041921397379912,
+ "grad_norm": 0.02204521745443344,
+ "learning_rate": 0.0006,
+ "loss": 4.894623756408691,
+ "step": 795
+ },
+ {
+ "epoch": 11.055895196506551,
+ "grad_norm": 0.0210255216807127,
+ "learning_rate": 0.0006,
+ "loss": 4.834009170532227,
+ "step": 796
+ },
+ {
+ "epoch": 11.069868995633188,
+ "grad_norm": 0.019797448068857193,
+ "learning_rate": 0.0006,
+ "loss": 5.086714744567871,
+ "step": 797
+ },
+ {
+ "epoch": 11.083842794759825,
+ "grad_norm": 0.02036641724407673,
+ "learning_rate": 0.0006,
+ "loss": 4.920056343078613,
+ "step": 798
+ },
+ {
+ "epoch": 11.097816593886463,
+ "grad_norm": 0.01966066285967827,
+ "learning_rate": 0.0006,
+ "loss": 4.879859447479248,
+ "step": 799
+ },
+ {
+ "epoch": 11.1117903930131,
+ "grad_norm": 0.0202149897813797,
+ "learning_rate": 0.0006,
+ "loss": 4.933640480041504,
+ "step": 800
+ },
+ {
+ "epoch": 11.125764192139737,
+ "grad_norm": 0.01893039606511593,
+ "learning_rate": 0.0006,
+ "loss": 4.901970863342285,
+ "step": 801
+ },
+ {
+ "epoch": 11.139737991266376,
+ "grad_norm": 0.016040155664086342,
+ "learning_rate": 0.0006,
+ "loss": 4.905491828918457,
+ "step": 802
+ },
+ {
+ "epoch": 11.153711790393013,
+ "grad_norm": 0.01720350608229637,
+ "learning_rate": 0.0006,
+ "loss": 4.798816680908203,
+ "step": 803
+ },
+ {
+ "epoch": 11.167685589519651,
+ "grad_norm": 0.017002852633595467,
+ "learning_rate": 0.0006,
+ "loss": 4.851859092712402,
+ "step": 804
+ },
+ {
+ "epoch": 11.181659388646288,
+ "grad_norm": 0.015594680793583393,
+ "learning_rate": 0.0006,
+ "loss": 4.90328311920166,
+ "step": 805
+ },
+ {
+ "epoch": 11.195633187772925,
+ "grad_norm": 0.014904462732374668,
+ "learning_rate": 0.0006,
+ "loss": 4.904215335845947,
+ "step": 806
+ },
+ {
+ "epoch": 11.209606986899564,
+ "grad_norm": 0.01366228237748146,
+ "learning_rate": 0.0006,
+ "loss": 4.9813055992126465,
+ "step": 807
+ },
+ {
+ "epoch": 11.2235807860262,
+ "grad_norm": 0.014719638973474503,
+ "learning_rate": 0.0006,
+ "loss": 4.927306175231934,
+ "step": 808
+ },
+ {
+ "epoch": 11.237554585152838,
+ "grad_norm": 0.015016036108136177,
+ "learning_rate": 0.0006,
+ "loss": 4.874215126037598,
+ "step": 809
+ },
+ {
+ "epoch": 11.251528384279476,
+ "grad_norm": 0.014925424009561539,
+ "learning_rate": 0.0006,
+ "loss": 4.838257789611816,
+ "step": 810
+ },
+ {
+ "epoch": 11.265502183406113,
+ "grad_norm": 0.014335823245346546,
+ "learning_rate": 0.0006,
+ "loss": 4.940608024597168,
+ "step": 811
+ },
+ {
+ "epoch": 11.279475982532752,
+ "grad_norm": 0.014987512491643429,
+ "learning_rate": 0.0006,
+ "loss": 4.831531524658203,
+ "step": 812
+ },
+ {
+ "epoch": 11.293449781659389,
+ "grad_norm": 0.016087457537651062,
+ "learning_rate": 0.0006,
+ "loss": 4.863227367401123,
+ "step": 813
+ },
+ {
+ "epoch": 11.307423580786025,
+ "grad_norm": 0.01671300269663334,
+ "learning_rate": 0.0006,
+ "loss": 4.8535661697387695,
+ "step": 814
+ },
+ {
+ "epoch": 11.321397379912664,
+ "grad_norm": 0.015229142270982265,
+ "learning_rate": 0.0006,
+ "loss": 4.948635101318359,
+ "step": 815
+ },
+ {
+ "epoch": 11.335371179039301,
+ "grad_norm": 0.01532587967813015,
+ "learning_rate": 0.0006,
+ "loss": 4.9035797119140625,
+ "step": 816
+ },
+ {
+ "epoch": 11.34934497816594,
+ "grad_norm": 0.015715476125478745,
+ "learning_rate": 0.0006,
+ "loss": 4.894709587097168,
+ "step": 817
+ },
+ {
+ "epoch": 11.363318777292577,
+ "grad_norm": 0.014156977646052837,
+ "learning_rate": 0.0006,
+ "loss": 4.960206985473633,
+ "step": 818
+ },
+ {
+ "epoch": 11.377292576419213,
+ "grad_norm": 0.017076566815376282,
+ "learning_rate": 0.0006,
+ "loss": 5.005850791931152,
+ "step": 819
+ },
+ {
+ "epoch": 11.391266375545852,
+ "grad_norm": 0.0172084029763937,
+ "learning_rate": 0.0006,
+ "loss": 4.903355598449707,
+ "step": 820
+ },
+ {
+ "epoch": 11.405240174672489,
+ "grad_norm": 0.0180258359760046,
+ "learning_rate": 0.0006,
+ "loss": 4.842667579650879,
+ "step": 821
+ },
+ {
+ "epoch": 11.419213973799126,
+ "grad_norm": 0.022130120545625687,
+ "learning_rate": 0.0006,
+ "loss": 4.799970626831055,
+ "step": 822
+ },
+ {
+ "epoch": 11.433187772925764,
+ "grad_norm": 0.02384897693991661,
+ "learning_rate": 0.0006,
+ "loss": 4.860746383666992,
+ "step": 823
+ },
+ {
+ "epoch": 11.447161572052401,
+ "grad_norm": 0.021881572902202606,
+ "learning_rate": 0.0006,
+ "loss": 4.918107032775879,
+ "step": 824
+ },
+ {
+ "epoch": 11.46113537117904,
+ "grad_norm": 0.023085080087184906,
+ "learning_rate": 0.0006,
+ "loss": 4.878946304321289,
+ "step": 825
+ },
+ {
+ "epoch": 11.475109170305677,
+ "grad_norm": 0.02291012555360794,
+ "learning_rate": 0.0006,
+ "loss": 4.958887100219727,
+ "step": 826
+ },
+ {
+ "epoch": 11.489082969432314,
+ "grad_norm": 0.0199968870729208,
+ "learning_rate": 0.0006,
+ "loss": 4.80955696105957,
+ "step": 827
+ },
+ {
+ "epoch": 11.503056768558952,
+ "grad_norm": 0.018702229484915733,
+ "learning_rate": 0.0006,
+ "loss": 5.0276408195495605,
+ "step": 828
+ },
+ {
+ "epoch": 11.51703056768559,
+ "grad_norm": 0.017787311226129532,
+ "learning_rate": 0.0006,
+ "loss": 4.96018123626709,
+ "step": 829
+ },
+ {
+ "epoch": 11.531004366812226,
+ "grad_norm": 0.01704501546919346,
+ "learning_rate": 0.0006,
+ "loss": 4.914553165435791,
+ "step": 830
+ },
+ {
+ "epoch": 11.544978165938865,
+ "grad_norm": 0.017519677057862282,
+ "learning_rate": 0.0006,
+ "loss": 4.798098087310791,
+ "step": 831
+ },
+ {
+ "epoch": 11.558951965065502,
+ "grad_norm": 0.018511991947889328,
+ "learning_rate": 0.0006,
+ "loss": 4.817785263061523,
+ "step": 832
+ },
+ {
+ "epoch": 11.57292576419214,
+ "grad_norm": 0.01876644790172577,
+ "learning_rate": 0.0006,
+ "loss": 4.818995475769043,
+ "step": 833
+ },
+ {
+ "epoch": 11.586899563318777,
+ "grad_norm": 0.01753568835556507,
+ "learning_rate": 0.0006,
+ "loss": 4.811938285827637,
+ "step": 834
+ },
+ {
+ "epoch": 11.600873362445414,
+ "grad_norm": 0.017046531662344933,
+ "learning_rate": 0.0006,
+ "loss": 4.914674282073975,
+ "step": 835
+ },
+ {
+ "epoch": 11.614847161572053,
+ "grad_norm": 0.01802094094455242,
+ "learning_rate": 0.0006,
+ "loss": 4.993284225463867,
+ "step": 836
+ },
+ {
+ "epoch": 11.62882096069869,
+ "grad_norm": 0.019493764266371727,
+ "learning_rate": 0.0006,
+ "loss": 4.895628452301025,
+ "step": 837
+ },
+ {
+ "epoch": 11.642794759825328,
+ "grad_norm": 0.019461210817098618,
+ "learning_rate": 0.0006,
+ "loss": 4.869269371032715,
+ "step": 838
+ },
+ {
+ "epoch": 11.656768558951965,
+ "grad_norm": 0.01981567218899727,
+ "learning_rate": 0.0006,
+ "loss": 4.863052845001221,
+ "step": 839
+ },
+ {
+ "epoch": 11.670742358078602,
+ "grad_norm": 0.019894491881132126,
+ "learning_rate": 0.0006,
+ "loss": 4.8775787353515625,
+ "step": 840
+ },
+ {
+ "epoch": 11.68471615720524,
+ "grad_norm": 0.01875786855816841,
+ "learning_rate": 0.0006,
+ "loss": 4.817763328552246,
+ "step": 841
+ },
+ {
+ "epoch": 11.698689956331878,
+ "grad_norm": 0.020926695317029953,
+ "learning_rate": 0.0006,
+ "loss": 4.7470526695251465,
+ "step": 842
+ },
+ {
+ "epoch": 11.712663755458514,
+ "grad_norm": 0.021477103233337402,
+ "learning_rate": 0.0006,
+ "loss": 4.762682914733887,
+ "step": 843
+ },
+ {
+ "epoch": 11.726637554585153,
+ "grad_norm": 0.020628711208701134,
+ "learning_rate": 0.0006,
+ "loss": 4.807985782623291,
+ "step": 844
+ },
+ {
+ "epoch": 11.74061135371179,
+ "grad_norm": 0.021489612758159637,
+ "learning_rate": 0.0006,
+ "loss": 4.843091011047363,
+ "step": 845
+ },
+ {
+ "epoch": 11.754585152838429,
+ "grad_norm": 0.021194491535425186,
+ "learning_rate": 0.0006,
+ "loss": 4.8300251960754395,
+ "step": 846
+ },
+ {
+ "epoch": 11.768558951965066,
+ "grad_norm": 0.01833350397646427,
+ "learning_rate": 0.0006,
+ "loss": 5.000202178955078,
+ "step": 847
+ },
+ {
+ "epoch": 11.782532751091702,
+ "grad_norm": 0.018096931278705597,
+ "learning_rate": 0.0006,
+ "loss": 4.880648136138916,
+ "step": 848
+ },
+ {
+ "epoch": 11.796506550218341,
+ "grad_norm": 0.0197161752730608,
+ "learning_rate": 0.0006,
+ "loss": 4.788316249847412,
+ "step": 849
+ },
+ {
+ "epoch": 11.810480349344978,
+ "grad_norm": 0.0203181654214859,
+ "learning_rate": 0.0006,
+ "loss": 4.870944976806641,
+ "step": 850
+ },
+ {
+ "epoch": 11.824454148471617,
+ "grad_norm": 0.01908455230295658,
+ "learning_rate": 0.0006,
+ "loss": 4.890144348144531,
+ "step": 851
+ },
+ {
+ "epoch": 11.838427947598253,
+ "grad_norm": 0.021281801164150238,
+ "learning_rate": 0.0006,
+ "loss": 4.89578914642334,
+ "step": 852
+ },
+ {
+ "epoch": 11.85240174672489,
+ "grad_norm": 0.02124273031949997,
+ "learning_rate": 0.0006,
+ "loss": 4.883749008178711,
+ "step": 853
+ },
+ {
+ "epoch": 11.866375545851529,
+ "grad_norm": 0.020969398319721222,
+ "learning_rate": 0.0006,
+ "loss": 4.809142589569092,
+ "step": 854
+ },
+ {
+ "epoch": 11.880349344978166,
+ "grad_norm": 0.01967657171189785,
+ "learning_rate": 0.0006,
+ "loss": 4.821864604949951,
+ "step": 855
+ },
+ {
+ "epoch": 11.894323144104803,
+ "grad_norm": 0.020887333899736404,
+ "learning_rate": 0.0006,
+ "loss": 4.917985916137695,
+ "step": 856
+ },
+ {
+ "epoch": 11.908296943231441,
+ "grad_norm": 0.019203083589673042,
+ "learning_rate": 0.0006,
+ "loss": 4.824960708618164,
+ "step": 857
+ },
+ {
+ "epoch": 11.922270742358078,
+ "grad_norm": 0.020162401720881462,
+ "learning_rate": 0.0006,
+ "loss": 4.772617816925049,
+ "step": 858
+ },
+ {
+ "epoch": 11.936244541484717,
+ "grad_norm": 0.02251487225294113,
+ "learning_rate": 0.0006,
+ "loss": 4.735418319702148,
+ "step": 859
+ },
+ {
+ "epoch": 11.950218340611354,
+ "grad_norm": 0.020313527435064316,
+ "learning_rate": 0.0006,
+ "loss": 4.939233779907227,
+ "step": 860
+ },
+ {
+ "epoch": 11.96419213973799,
+ "grad_norm": 0.021635599434375763,
+ "learning_rate": 0.0006,
+ "loss": 4.861807346343994,
+ "step": 861
+ },
+ {
+ "epoch": 11.97816593886463,
+ "grad_norm": 0.02285311557352543,
+ "learning_rate": 0.0006,
+ "loss": 4.932732582092285,
+ "step": 862
+ },
+ {
+ "epoch": 11.992139737991266,
+ "grad_norm": 0.02072535641491413,
+ "learning_rate": 0.0006,
+ "loss": 4.796767711639404,
+ "step": 863
+ },
+ {
+ "epoch": 12.0,
+ "grad_norm": 0.020588304847478867,
+ "learning_rate": 0.0006,
+ "loss": 4.997537612915039,
+ "step": 864
+ },
+ {
+ "epoch": 12.0,
+ "eval_loss": 5.044651985168457,
+ "eval_runtime": 56.9145,
+ "eval_samples_per_second": 42.906,
+ "eval_steps_per_second": 1.353,
+ "step": 864
+ },
+ {
+ "epoch": 12.013973799126637,
+ "grad_norm": 0.017597166821360588,
+ "learning_rate": 0.0006,
+ "loss": 4.830524444580078,
+ "step": 865
+ },
+ {
+ "epoch": 12.027947598253276,
+ "grad_norm": 0.01716788113117218,
+ "learning_rate": 0.0006,
+ "loss": 4.894325256347656,
+ "step": 866
+ },
+ {
+ "epoch": 12.041921397379912,
+ "grad_norm": 0.017712457105517387,
+ "learning_rate": 0.0006,
+ "loss": 4.868297576904297,
+ "step": 867
+ },
+ {
+ "epoch": 12.055895196506551,
+ "grad_norm": 0.021486390382051468,
+ "learning_rate": 0.0006,
+ "loss": 4.922049522399902,
+ "step": 868
+ },
+ {
+ "epoch": 12.069868995633188,
+ "grad_norm": 0.024560654535889626,
+ "learning_rate": 0.0006,
+ "loss": 4.734241008758545,
+ "step": 869
+ },
+ {
+ "epoch": 12.083842794759825,
+ "grad_norm": 0.025870252400636673,
+ "learning_rate": 0.0006,
+ "loss": 4.868566989898682,
+ "step": 870
+ },
+ {
+ "epoch": 12.097816593886463,
+ "grad_norm": 0.0277389045804739,
+ "learning_rate": 0.0006,
+ "loss": 4.755006790161133,
+ "step": 871
+ },
+ {
+ "epoch": 12.1117903930131,
+ "grad_norm": 0.029100844636559486,
+ "learning_rate": 0.0006,
+ "loss": 4.909667015075684,
+ "step": 872
+ },
+ {
+ "epoch": 12.125764192139737,
+ "grad_norm": 0.031833067536354065,
+ "learning_rate": 0.0006,
+ "loss": 4.858674049377441,
+ "step": 873
+ },
+ {
+ "epoch": 12.139737991266376,
+ "grad_norm": 0.032709237188100815,
+ "learning_rate": 0.0006,
+ "loss": 4.823696136474609,
+ "step": 874
+ },
+ {
+ "epoch": 12.153711790393013,
+ "grad_norm": 0.02995380386710167,
+ "learning_rate": 0.0006,
+ "loss": 4.850235462188721,
+ "step": 875
+ },
+ {
+ "epoch": 12.167685589519651,
+ "grad_norm": 0.034098610281944275,
+ "learning_rate": 0.0006,
+ "loss": 4.8169989585876465,
+ "step": 876
+ },
+ {
+ "epoch": 12.181659388646288,
+ "grad_norm": 0.03377080336213112,
+ "learning_rate": 0.0006,
+ "loss": 4.90738582611084,
+ "step": 877
+ },
+ {
+ "epoch": 12.195633187772925,
+ "grad_norm": 0.029780825600028038,
+ "learning_rate": 0.0006,
+ "loss": 4.811163902282715,
+ "step": 878
+ },
+ {
+ "epoch": 12.209606986899564,
+ "grad_norm": 0.03118024580180645,
+ "learning_rate": 0.0006,
+ "loss": 4.861286640167236,
+ "step": 879
+ },
+ {
+ "epoch": 12.2235807860262,
+ "grad_norm": 0.033052004873752594,
+ "learning_rate": 0.0006,
+ "loss": 4.865021228790283,
+ "step": 880
+ },
+ {
+ "epoch": 12.237554585152838,
+ "grad_norm": 0.029070032760500908,
+ "learning_rate": 0.0006,
+ "loss": 4.947303771972656,
+ "step": 881
+ },
+ {
+ "epoch": 12.251528384279476,
+ "grad_norm": 0.028355715796351433,
+ "learning_rate": 0.0006,
+ "loss": 4.804648399353027,
+ "step": 882
+ },
+ {
+ "epoch": 12.265502183406113,
+ "grad_norm": 0.031257227063179016,
+ "learning_rate": 0.0006,
+ "loss": 4.757718086242676,
+ "step": 883
+ },
+ {
+ "epoch": 12.279475982532752,
+ "grad_norm": 0.031438373029232025,
+ "learning_rate": 0.0006,
+ "loss": 4.851268768310547,
+ "step": 884
+ },
+ {
+ "epoch": 12.293449781659389,
+ "grad_norm": 0.030492138117551804,
+ "learning_rate": 0.0006,
+ "loss": 4.7891035079956055,
+ "step": 885
+ },
+ {
+ "epoch": 12.307423580786025,
+ "grad_norm": 0.026115991175174713,
+ "learning_rate": 0.0006,
+ "loss": 4.86474609375,
+ "step": 886
+ },
+ {
+ "epoch": 12.321397379912664,
+ "grad_norm": 0.024184465408325195,
+ "learning_rate": 0.0006,
+ "loss": 4.860456943511963,
+ "step": 887
+ },
+ {
+ "epoch": 12.335371179039301,
+ "grad_norm": 0.025437982752919197,
+ "learning_rate": 0.0006,
+ "loss": 4.806251049041748,
+ "step": 888
+ },
+ {
+ "epoch": 12.34934497816594,
+ "grad_norm": 0.022390231490135193,
+ "learning_rate": 0.0006,
+ "loss": 4.807086944580078,
+ "step": 889
+ },
+ {
+ "epoch": 12.363318777292577,
+ "grad_norm": 0.020174294710159302,
+ "learning_rate": 0.0006,
+ "loss": 4.78292179107666,
+ "step": 890
+ },
+ {
+ "epoch": 12.377292576419213,
+ "grad_norm": 0.019734324887394905,
+ "learning_rate": 0.0006,
+ "loss": 4.784509658813477,
+ "step": 891
+ },
+ {
+ "epoch": 12.391266375545852,
+ "grad_norm": 0.016733523458242416,
+ "learning_rate": 0.0006,
+ "loss": 4.811755180358887,
+ "step": 892
+ },
+ {
+ "epoch": 12.405240174672489,
+ "grad_norm": 0.01570146717131138,
+ "learning_rate": 0.0006,
+ "loss": 4.720088005065918,
+ "step": 893
+ },
+ {
+ "epoch": 12.419213973799126,
+ "grad_norm": 0.01601393334567547,
+ "learning_rate": 0.0006,
+ "loss": 4.783543109893799,
+ "step": 894
+ },
+ {
+ "epoch": 12.433187772925764,
+ "grad_norm": 0.014035487547516823,
+ "learning_rate": 0.0006,
+ "loss": 4.848340034484863,
+ "step": 895
+ },
+ {
+ "epoch": 12.447161572052401,
+ "grad_norm": 0.015248487703502178,
+ "learning_rate": 0.0006,
+ "loss": 4.773287773132324,
+ "step": 896
+ },
+ {
+ "epoch": 12.46113537117904,
+ "grad_norm": 0.01532268151640892,
+ "learning_rate": 0.0006,
+ "loss": 4.799646854400635,
+ "step": 897
+ },
+ {
+ "epoch": 12.475109170305677,
+ "grad_norm": 0.013982534408569336,
+ "learning_rate": 0.0006,
+ "loss": 4.8090500831604,
+ "step": 898
+ },
+ {
+ "epoch": 12.489082969432314,
+ "grad_norm": 0.014081164263188839,
+ "learning_rate": 0.0006,
+ "loss": 4.751888751983643,
+ "step": 899
+ },
+ {
+ "epoch": 12.503056768558952,
+ "grad_norm": 0.01380133256316185,
+ "learning_rate": 0.0006,
+ "loss": 4.78144645690918,
+ "step": 900
+ },
+ {
+ "epoch": 12.51703056768559,
+ "grad_norm": 0.013555269688367844,
+ "learning_rate": 0.0006,
+ "loss": 4.779257774353027,
+ "step": 901
+ },
+ {
+ "epoch": 12.531004366812226,
+ "grad_norm": 0.013075701892375946,
+ "learning_rate": 0.0006,
+ "loss": 4.820705413818359,
+ "step": 902
+ },
+ {
+ "epoch": 12.544978165938865,
+ "grad_norm": 0.01276316400617361,
+ "learning_rate": 0.0006,
+ "loss": 4.836733818054199,
+ "step": 903
+ },
+ {
+ "epoch": 12.558951965065502,
+ "grad_norm": 0.013033892959356308,
+ "learning_rate": 0.0006,
+ "loss": 4.764178276062012,
+ "step": 904
+ },
+ {
+ "epoch": 12.57292576419214,
+ "grad_norm": 0.013996980153024197,
+ "learning_rate": 0.0006,
+ "loss": 4.767983436584473,
+ "step": 905
+ },
+ {
+ "epoch": 12.586899563318777,
+ "grad_norm": 0.015222184360027313,
+ "learning_rate": 0.0006,
+ "loss": 4.759721755981445,
+ "step": 906
+ },
+ {
+ "epoch": 12.600873362445414,
+ "grad_norm": 0.01623941957950592,
+ "learning_rate": 0.0006,
+ "loss": 4.849811553955078,
+ "step": 907
+ },
+ {
+ "epoch": 12.614847161572053,
+ "grad_norm": 0.015711182728409767,
+ "learning_rate": 0.0006,
+ "loss": 4.7829084396362305,
+ "step": 908
+ },
+ {
+ "epoch": 12.62882096069869,
+ "grad_norm": 0.014137467369437218,
+ "learning_rate": 0.0006,
+ "loss": 4.73321533203125,
+ "step": 909
+ },
+ {
+ "epoch": 12.642794759825328,
+ "grad_norm": 0.012738462537527084,
+ "learning_rate": 0.0006,
+ "loss": 4.870052337646484,
+ "step": 910
+ },
+ {
+ "epoch": 12.656768558951965,
+ "grad_norm": 0.013683130033314228,
+ "learning_rate": 0.0006,
+ "loss": 4.82658052444458,
+ "step": 911
+ },
+ {
+ "epoch": 12.670742358078602,
+ "grad_norm": 0.014409830793738365,
+ "learning_rate": 0.0006,
+ "loss": 4.707494735717773,
+ "step": 912
+ },
+ {
+ "epoch": 12.68471615720524,
+ "grad_norm": 0.013806809671223164,
+ "learning_rate": 0.0006,
+ "loss": 4.778928279876709,
+ "step": 913
+ },
+ {
+ "epoch": 12.698689956331878,
+ "grad_norm": 0.012857912108302116,
+ "learning_rate": 0.0006,
+ "loss": 4.890717506408691,
+ "step": 914
+ },
+ {
+ "epoch": 12.712663755458514,
+ "grad_norm": 0.013828758150339127,
+ "learning_rate": 0.0006,
+ "loss": 4.8357439041137695,
+ "step": 915
+ },
+ {
+ "epoch": 12.726637554585153,
+ "grad_norm": 0.01478351280093193,
+ "learning_rate": 0.0006,
+ "loss": 4.869986534118652,
+ "step": 916
+ },
+ {
+ "epoch": 12.74061135371179,
+ "grad_norm": 0.01326421182602644,
+ "learning_rate": 0.0006,
+ "loss": 4.830646514892578,
+ "step": 917
+ },
+ {
+ "epoch": 12.754585152838429,
+ "grad_norm": 0.013335433788597584,
+ "learning_rate": 0.0006,
+ "loss": 4.7786149978637695,
+ "step": 918
+ },
+ {
+ "epoch": 12.768558951965066,
+ "grad_norm": 0.014261147007346153,
+ "learning_rate": 0.0006,
+ "loss": 4.788693428039551,
+ "step": 919
+ },
+ {
+ "epoch": 12.782532751091702,
+ "grad_norm": 0.014274783432483673,
+ "learning_rate": 0.0006,
+ "loss": 4.77295446395874,
+ "step": 920
+ },
+ {
+ "epoch": 12.796506550218341,
+ "grad_norm": 0.015074674971401691,
+ "learning_rate": 0.0006,
+ "loss": 4.830754280090332,
+ "step": 921
+ },
+ {
+ "epoch": 12.810480349344978,
+ "grad_norm": 0.015599515289068222,
+ "learning_rate": 0.0006,
+ "loss": 4.837759017944336,
+ "step": 922
+ },
+ {
+ "epoch": 12.824454148471617,
+ "grad_norm": 0.014595242217183113,
+ "learning_rate": 0.0006,
+ "loss": 4.882965564727783,
+ "step": 923
+ },
+ {
+ "epoch": 12.838427947598253,
+ "grad_norm": 0.014951467514038086,
+ "learning_rate": 0.0006,
+ "loss": 4.709748268127441,
+ "step": 924
+ },
+ {
+ "epoch": 12.85240174672489,
+ "grad_norm": 0.01539295818656683,
+ "learning_rate": 0.0006,
+ "loss": 4.791160583496094,
+ "step": 925
+ },
+ {
+ "epoch": 12.866375545851529,
+ "grad_norm": 0.015987906605005264,
+ "learning_rate": 0.0006,
+ "loss": 4.661876678466797,
+ "step": 926
+ },
+ {
+ "epoch": 12.880349344978166,
+ "grad_norm": 0.016527341678738594,
+ "learning_rate": 0.0006,
+ "loss": 4.721952438354492,
+ "step": 927
+ },
+ {
+ "epoch": 12.894323144104803,
+ "grad_norm": 0.01637161523103714,
+ "learning_rate": 0.0006,
+ "loss": 4.704446792602539,
+ "step": 928
+ },
+ {
+ "epoch": 12.908296943231441,
+ "grad_norm": 0.015221050009131432,
+ "learning_rate": 0.0006,
+ "loss": 4.771907806396484,
+ "step": 929
+ },
+ {
+ "epoch": 12.922270742358078,
+ "grad_norm": 0.016242166981101036,
+ "learning_rate": 0.0006,
+ "loss": 4.806267738342285,
+ "step": 930
+ },
+ {
+ "epoch": 12.936244541484717,
+ "grad_norm": 0.01628357358276844,
+ "learning_rate": 0.0006,
+ "loss": 4.7252326011657715,
+ "step": 931
+ },
+ {
+ "epoch": 12.950218340611354,
+ "grad_norm": 0.017428133636713028,
+ "learning_rate": 0.0006,
+ "loss": 4.765689849853516,
+ "step": 932
+ },
+ {
+ "epoch": 12.96419213973799,
+ "grad_norm": 0.01870061084628105,
+ "learning_rate": 0.0006,
+ "loss": 4.814488887786865,
+ "step": 933
+ },
+ {
+ "epoch": 12.97816593886463,
+ "grad_norm": 0.019796060398221016,
+ "learning_rate": 0.0006,
+ "loss": 4.739185810089111,
+ "step": 934
+ },
+ {
+ "epoch": 12.992139737991266,
+ "grad_norm": 0.01985820196568966,
+ "learning_rate": 0.0006,
+ "loss": 4.810608863830566,
+ "step": 935
+ },
+ {
+ "epoch": 13.0,
+ "grad_norm": 0.020834477618336678,
+ "learning_rate": 0.0006,
+ "loss": 4.727858066558838,
+ "step": 936
+ },
+ {
+ "epoch": 13.0,
+ "eval_loss": 4.979116916656494,
+ "eval_runtime": 57.2877,
+ "eval_samples_per_second": 42.627,
+ "eval_steps_per_second": 1.344,
+ "step": 936
+ },
+ {
+ "epoch": 13.013973799126637,
+ "grad_norm": 0.02025519125163555,
+ "learning_rate": 0.0006,
+ "loss": 4.628875732421875,
+ "step": 937
+ },
+ {
+ "epoch": 13.027947598253276,
+ "grad_norm": 0.019596368074417114,
+ "learning_rate": 0.0006,
+ "loss": 4.694699287414551,
+ "step": 938
+ },
+ {
+ "epoch": 13.041921397379912,
+ "grad_norm": 0.020494934171438217,
+ "learning_rate": 0.0006,
+ "loss": 4.691821098327637,
+ "step": 939
+ },
+ {
+ "epoch": 13.055895196506551,
+ "grad_norm": 0.020121760666370392,
+ "learning_rate": 0.0006,
+ "loss": 4.71478271484375,
+ "step": 940
+ },
+ {
+ "epoch": 13.069868995633188,
+ "grad_norm": 0.021091405302286148,
+ "learning_rate": 0.0006,
+ "loss": 4.695915222167969,
+ "step": 941
+ },
+ {
+ "epoch": 13.083842794759825,
+ "grad_norm": 0.019787423312664032,
+ "learning_rate": 0.0006,
+ "loss": 4.8652448654174805,
+ "step": 942
+ },
+ {
+ "epoch": 13.097816593886463,
+ "grad_norm": 0.0185833927243948,
+ "learning_rate": 0.0006,
+ "loss": 4.7065887451171875,
+ "step": 943
+ },
+ {
+ "epoch": 13.1117903930131,
+ "grad_norm": 0.021220475435256958,
+ "learning_rate": 0.0006,
+ "loss": 4.741232395172119,
+ "step": 944
+ },
+ {
+ "epoch": 13.125764192139737,
+ "grad_norm": 0.02136622555553913,
+ "learning_rate": 0.0006,
+ "loss": 4.705967903137207,
+ "step": 945
+ },
+ {
+ "epoch": 13.139737991266376,
+ "grad_norm": 0.0191587433218956,
+ "learning_rate": 0.0006,
+ "loss": 4.758667945861816,
+ "step": 946
+ },
+ {
+ "epoch": 13.153711790393013,
+ "grad_norm": 0.018290022388100624,
+ "learning_rate": 0.0006,
+ "loss": 4.760666847229004,
+ "step": 947
+ },
+ {
+ "epoch": 13.167685589519651,
+ "grad_norm": 0.021704409271478653,
+ "learning_rate": 0.0006,
+ "loss": 4.750555992126465,
+ "step": 948
+ },
+ {
+ "epoch": 13.181659388646288,
+ "grad_norm": 0.021668918430805206,
+ "learning_rate": 0.0006,
+ "loss": 4.840903282165527,
+ "step": 949
+ },
+ {
+ "epoch": 13.195633187772925,
+ "grad_norm": 0.019497141242027283,
+ "learning_rate": 0.0006,
+ "loss": 4.738969326019287,
+ "step": 950
+ },
+ {
+ "epoch": 13.209606986899564,
+ "grad_norm": 0.018500596284866333,
+ "learning_rate": 0.0006,
+ "loss": 4.75154972076416,
+ "step": 951
+ },
+ {
+ "epoch": 13.2235807860262,
+ "grad_norm": 0.01872239261865616,
+ "learning_rate": 0.0006,
+ "loss": 4.811738014221191,
+ "step": 952
+ },
+ {
+ "epoch": 13.237554585152838,
+ "grad_norm": 0.020321547985076904,
+ "learning_rate": 0.0006,
+ "loss": 4.815384387969971,
+ "step": 953
+ },
+ {
+ "epoch": 13.251528384279476,
+ "grad_norm": 0.02283575013279915,
+ "learning_rate": 0.0006,
+ "loss": 4.74582052230835,
+ "step": 954
+ },
+ {
+ "epoch": 13.265502183406113,
+ "grad_norm": 0.02474398910999298,
+ "learning_rate": 0.0006,
+ "loss": 4.738375186920166,
+ "step": 955
+ },
+ {
+ "epoch": 13.279475982532752,
+ "grad_norm": 0.02065764181315899,
+ "learning_rate": 0.0006,
+ "loss": 4.746875286102295,
+ "step": 956
+ },
+ {
+ "epoch": 13.293449781659389,
+ "grad_norm": 0.020534520968794823,
+ "learning_rate": 0.0006,
+ "loss": 4.656914710998535,
+ "step": 957
+ },
+ {
+ "epoch": 13.307423580786025,
+ "grad_norm": 0.0206378772854805,
+ "learning_rate": 0.0006,
+ "loss": 4.842615604400635,
+ "step": 958
+ },
+ {
+ "epoch": 13.321397379912664,
+ "grad_norm": 0.017935654148459435,
+ "learning_rate": 0.0006,
+ "loss": 4.717827796936035,
+ "step": 959
+ },
+ {
+ "epoch": 13.335371179039301,
+ "grad_norm": 0.016958363354206085,
+ "learning_rate": 0.0006,
+ "loss": 4.681070327758789,
+ "step": 960
+ },
+ {
+ "epoch": 13.34934497816594,
+ "grad_norm": 0.015735412016510963,
+ "learning_rate": 0.0006,
+ "loss": 4.7957048416137695,
+ "step": 961
+ },
+ {
+ "epoch": 13.363318777292577,
+ "grad_norm": 0.01563699170947075,
+ "learning_rate": 0.0006,
+ "loss": 4.701192378997803,
+ "step": 962
+ },
+ {
+ "epoch": 13.377292576419213,
+ "grad_norm": 0.015207760035991669,
+ "learning_rate": 0.0006,
+ "loss": 4.73256254196167,
+ "step": 963
+ },
+ {
+ "epoch": 13.391266375545852,
+ "grad_norm": 0.015396883711218834,
+ "learning_rate": 0.0006,
+ "loss": 4.76412296295166,
+ "step": 964
+ },
+ {
+ "epoch": 13.405240174672489,
+ "grad_norm": 0.015690583735704422,
+ "learning_rate": 0.0006,
+ "loss": 4.791157245635986,
+ "step": 965
+ },
+ {
+ "epoch": 13.419213973799126,
+ "grad_norm": 0.01755724661052227,
+ "learning_rate": 0.0006,
+ "loss": 4.845666885375977,
+ "step": 966
+ },
+ {
+ "epoch": 13.433187772925764,
+ "grad_norm": 0.019443422555923462,
+ "learning_rate": 0.0006,
+ "loss": 4.5902299880981445,
+ "step": 967
+ },
+ {
+ "epoch": 13.447161572052401,
+ "grad_norm": 0.02118338644504547,
+ "learning_rate": 0.0006,
+ "loss": 4.740627288818359,
+ "step": 968
+ },
+ {
+ "epoch": 13.46113537117904,
+ "grad_norm": 0.02174839936196804,
+ "learning_rate": 0.0006,
+ "loss": 4.814560890197754,
+ "step": 969
+ },
+ {
+ "epoch": 13.475109170305677,
+ "grad_norm": 0.022860374301671982,
+ "learning_rate": 0.0006,
+ "loss": 4.66440486907959,
+ "step": 970
+ },
+ {
+ "epoch": 13.489082969432314,
+ "grad_norm": 0.024742677807807922,
+ "learning_rate": 0.0006,
+ "loss": 4.7451653480529785,
+ "step": 971
+ },
+ {
+ "epoch": 13.503056768558952,
+ "grad_norm": 0.027068907395005226,
+ "learning_rate": 0.0006,
+ "loss": 4.802584171295166,
+ "step": 972
+ },
+ {
+ "epoch": 13.51703056768559,
+ "grad_norm": 0.026316625997424126,
+ "learning_rate": 0.0006,
+ "loss": 4.842397689819336,
+ "step": 973
+ },
+ {
+ "epoch": 13.531004366812226,
+ "grad_norm": 0.024610962718725204,
+ "learning_rate": 0.0006,
+ "loss": 4.650721073150635,
+ "step": 974
+ },
+ {
+ "epoch": 13.544978165938865,
+ "grad_norm": 0.024122603237628937,
+ "learning_rate": 0.0006,
+ "loss": 4.6984357833862305,
+ "step": 975
+ },
+ {
+ "epoch": 13.558951965065502,
+ "grad_norm": 0.022699708119034767,
+ "learning_rate": 0.0006,
+ "loss": 4.793007850646973,
+ "step": 976
+ },
+ {
+ "epoch": 13.57292576419214,
+ "grad_norm": 0.021706534549593925,
+ "learning_rate": 0.0006,
+ "loss": 4.691362380981445,
+ "step": 977
+ },
+ {
+ "epoch": 13.586899563318777,
+ "grad_norm": 0.021363291889429092,
+ "learning_rate": 0.0006,
+ "loss": 4.76762580871582,
+ "step": 978
+ },
+ {
+ "epoch": 13.600873362445414,
+ "grad_norm": 0.023476677015423775,
+ "learning_rate": 0.0006,
+ "loss": 4.777202606201172,
+ "step": 979
+ },
+ {
+ "epoch": 13.614847161572053,
+ "grad_norm": 0.02457100711762905,
+ "learning_rate": 0.0006,
+ "loss": 4.590703010559082,
+ "step": 980
+ },
+ {
+ "epoch": 13.62882096069869,
+ "grad_norm": 0.02469099499285221,
+ "learning_rate": 0.0006,
+ "loss": 4.640185356140137,
+ "step": 981
+ },
+ {
+ "epoch": 13.642794759825328,
+ "grad_norm": 0.024914603680372238,
+ "learning_rate": 0.0006,
+ "loss": 4.772201061248779,
+ "step": 982
+ },
+ {
+ "epoch": 13.656768558951965,
+ "grad_norm": 0.025133173912763596,
+ "learning_rate": 0.0006,
+ "loss": 4.743337154388428,
+ "step": 983
+ },
+ {
+ "epoch": 13.670742358078602,
+ "grad_norm": 0.02034137211740017,
+ "learning_rate": 0.0006,
+ "loss": 4.752127647399902,
+ "step": 984
+ },
+ {
+ "epoch": 13.68471615720524,
+ "grad_norm": 0.019645661115646362,
+ "learning_rate": 0.0006,
+ "loss": 4.838453769683838,
+ "step": 985
+ },
+ {
+ "epoch": 13.698689956331878,
+ "grad_norm": 0.01900297962129116,
+ "learning_rate": 0.0006,
+ "loss": 4.804868221282959,
+ "step": 986
+ },
+ {
+ "epoch": 13.712663755458514,
+ "grad_norm": 0.020292341709136963,
+ "learning_rate": 0.0006,
+ "loss": 4.609743118286133,
+ "step": 987
+ },
+ {
+ "epoch": 13.726637554585153,
+ "grad_norm": 0.017014775425195694,
+ "learning_rate": 0.0006,
+ "loss": 4.740612983703613,
+ "step": 988
+ },
+ {
+ "epoch": 13.74061135371179,
+ "grad_norm": 0.015199844725430012,
+ "learning_rate": 0.0006,
+ "loss": 4.719701766967773,
+ "step": 989
+ },
+ {
+ "epoch": 13.754585152838429,
+ "grad_norm": 0.015125414356589317,
+ "learning_rate": 0.0006,
+ "loss": 4.686114311218262,
+ "step": 990
+ },
+ {
+ "epoch": 13.768558951965066,
+ "grad_norm": 0.014779540710151196,
+ "learning_rate": 0.0006,
+ "loss": 4.674347877502441,
+ "step": 991
+ },
+ {
+ "epoch": 13.782532751091702,
+ "grad_norm": 0.015255163423717022,
+ "learning_rate": 0.0006,
+ "loss": 4.6309309005737305,
+ "step": 992
+ },
+ {
+ "epoch": 13.796506550218341,
+ "grad_norm": 0.013692197389900684,
+ "learning_rate": 0.0006,
+ "loss": 4.698497295379639,
+ "step": 993
+ },
+ {
+ "epoch": 13.810480349344978,
+ "grad_norm": 0.014529623091220856,
+ "learning_rate": 0.0006,
+ "loss": 4.628443717956543,
+ "step": 994
+ },
+ {
+ "epoch": 13.824454148471617,
+ "grad_norm": 0.01576659083366394,
+ "learning_rate": 0.0006,
+ "loss": 4.788015842437744,
+ "step": 995
+ },
+ {
+ "epoch": 13.838427947598253,
+ "grad_norm": 0.015834277495741844,
+ "learning_rate": 0.0006,
+ "loss": 4.672703742980957,
+ "step": 996
+ },
+ {
+ "epoch": 13.85240174672489,
+ "grad_norm": 0.015539503656327724,
+ "learning_rate": 0.0006,
+ "loss": 4.731503486633301,
+ "step": 997
+ },
+ {
+ "epoch": 13.866375545851529,
+ "grad_norm": 0.013463746756315231,
+ "learning_rate": 0.0006,
+ "loss": 4.7006330490112305,
+ "step": 998
+ },
+ {
+ "epoch": 13.880349344978166,
+ "grad_norm": 0.016377726569771767,
+ "learning_rate": 0.0006,
+ "loss": 4.636074066162109,
+ "step": 999
+ },
+ {
+ "epoch": 13.894323144104803,
+ "grad_norm": 0.016660034656524658,
+ "learning_rate": 0.0006,
+ "loss": 4.742693901062012,
+ "step": 1000
+ },
+ {
+ "epoch": 13.908296943231441,
+ "grad_norm": 0.014902938157320023,
+ "learning_rate": 0.0006,
+ "loss": 4.769811630249023,
+ "step": 1001
+ },
+ {
+ "epoch": 13.922270742358078,
+ "grad_norm": 0.014883476309478283,
+ "learning_rate": 0.0006,
+ "loss": 4.731159210205078,
+ "step": 1002
+ },
+ {
+ "epoch": 13.936244541484717,
+ "grad_norm": 0.014494217932224274,
+ "learning_rate": 0.0006,
+ "loss": 4.659485816955566,
+ "step": 1003
+ },
+ {
+ "epoch": 13.950218340611354,
+ "grad_norm": 0.015426991507411003,
+ "learning_rate": 0.0006,
+ "loss": 4.755229949951172,
+ "step": 1004
+ },
+ {
+ "epoch": 13.96419213973799,
+ "grad_norm": 0.016648586839437485,
+ "learning_rate": 0.0006,
+ "loss": 4.74111270904541,
+ "step": 1005
+ },
+ {
+ "epoch": 13.97816593886463,
+ "grad_norm": 0.016980541869997978,
+ "learning_rate": 0.0006,
+ "loss": 4.74072265625,
+ "step": 1006
+ },
+ {
+ "epoch": 13.992139737991266,
+ "grad_norm": 0.018070152029395103,
+ "learning_rate": 0.0006,
+ "loss": 4.732975006103516,
+ "step": 1007
+ },
+ {
+ "epoch": 14.0,
+ "grad_norm": 0.019910665228962898,
+ "learning_rate": 0.0006,
+ "loss": 4.634697914123535,
+ "step": 1008
+ },
+ {
+ "epoch": 14.0,
+ "eval_loss": 4.850640296936035,
+ "eval_runtime": 57.3039,
+ "eval_samples_per_second": 42.615,
+ "eval_steps_per_second": 1.344,
+ "step": 1008
+ },
+ {
+ "epoch": 14.013973799126637,
+ "grad_norm": 0.01980053074657917,
+ "learning_rate": 0.0006,
+ "loss": 4.628744602203369,
+ "step": 1009
+ },
+ {
+ "epoch": 14.027947598253276,
+ "grad_norm": 0.02497623674571514,
+ "learning_rate": 0.0006,
+ "loss": 4.75930118560791,
+ "step": 1010
+ },
+ {
+ "epoch": 14.041921397379912,
+ "grad_norm": 0.02794097363948822,
+ "learning_rate": 0.0006,
+ "loss": 4.691340446472168,
+ "step": 1011
+ },
+ {
+ "epoch": 14.055895196506551,
+ "grad_norm": 0.023725535720586777,
+ "learning_rate": 0.0006,
+ "loss": 4.733664512634277,
+ "step": 1012
+ },
+ {
+ "epoch": 14.069868995633188,
+ "grad_norm": 0.022966474294662476,
+ "learning_rate": 0.0006,
+ "loss": 4.664808750152588,
+ "step": 1013
+ },
+ {
+ "epoch": 14.083842794759825,
+ "grad_norm": 0.02530563250184059,
+ "learning_rate": 0.0006,
+ "loss": 4.8047285079956055,
+ "step": 1014
+ },
+ {
+ "epoch": 14.097816593886463,
+ "grad_norm": 0.02491082064807415,
+ "learning_rate": 0.0006,
+ "loss": 4.667956352233887,
+ "step": 1015
+ },
+ {
+ "epoch": 14.1117903930131,
+ "grad_norm": 0.022110294550657272,
+ "learning_rate": 0.0006,
+ "loss": 4.666140556335449,
+ "step": 1016
+ },
+ {
+ "epoch": 14.125764192139737,
+ "grad_norm": 0.02065849117934704,
+ "learning_rate": 0.0006,
+ "loss": 4.581960678100586,
+ "step": 1017
+ },
+ {
+ "epoch": 14.139737991266376,
+ "grad_norm": 0.01941298507153988,
+ "learning_rate": 0.0006,
+ "loss": 4.614226341247559,
+ "step": 1018
+ },
+ {
+ "epoch": 14.153711790393013,
+ "grad_norm": 0.022441856563091278,
+ "learning_rate": 0.0006,
+ "loss": 4.638352394104004,
+ "step": 1019
+ },
+ {
+ "epoch": 14.167685589519651,
+ "grad_norm": 0.021519694477319717,
+ "learning_rate": 0.0006,
+ "loss": 4.700160980224609,
+ "step": 1020
+ },
+ {
+ "epoch": 14.181659388646288,
+ "grad_norm": 0.018956486135721207,
+ "learning_rate": 0.0006,
+ "loss": 4.649593353271484,
+ "step": 1021
+ },
+ {
+ "epoch": 14.195633187772925,
+ "grad_norm": 0.016681140288710594,
+ "learning_rate": 0.0006,
+ "loss": 4.603952407836914,
+ "step": 1022
+ },
+ {
+ "epoch": 14.209606986899564,
+ "grad_norm": 0.01788809522986412,
+ "learning_rate": 0.0006,
+ "loss": 4.669160842895508,
+ "step": 1023
+ },
+ {
+ "epoch": 14.2235807860262,
+ "grad_norm": 0.016047228127717972,
+ "learning_rate": 0.0006,
+ "loss": 4.6367950439453125,
+ "step": 1024
+ },
+ {
+ "epoch": 14.237554585152838,
+ "grad_norm": 0.01614239625632763,
+ "learning_rate": 0.0006,
+ "loss": 4.664529323577881,
+ "step": 1025
+ },
+ {
+ "epoch": 14.251528384279476,
+ "grad_norm": 0.016362549737095833,
+ "learning_rate": 0.0006,
+ "loss": 4.702373504638672,
+ "step": 1026
+ },
+ {
+ "epoch": 14.265502183406113,
+ "grad_norm": 0.015800409018993378,
+ "learning_rate": 0.0006,
+ "loss": 4.7594428062438965,
+ "step": 1027
+ },
+ {
+ "epoch": 14.279475982532752,
+ "grad_norm": 0.016730139032006264,
+ "learning_rate": 0.0006,
+ "loss": 4.677244186401367,
+ "step": 1028
+ },
+ {
+ "epoch": 14.293449781659389,
+ "grad_norm": 0.01661759801208973,
+ "learning_rate": 0.0006,
+ "loss": 4.633679389953613,
+ "step": 1029
+ },
+ {
+ "epoch": 14.307423580786025,
+ "grad_norm": 0.017673347145318985,
+ "learning_rate": 0.0006,
+ "loss": 4.649107456207275,
+ "step": 1030
+ },
+ {
+ "epoch": 14.321397379912664,
+ "grad_norm": 0.019243352115154266,
+ "learning_rate": 0.0006,
+ "loss": 4.611257553100586,
+ "step": 1031
+ },
+ {
+ "epoch": 14.335371179039301,
+ "grad_norm": 0.019817164167761803,
+ "learning_rate": 0.0006,
+ "loss": 4.588086128234863,
+ "step": 1032
+ },
+ {
+ "epoch": 14.34934497816594,
+ "grad_norm": 0.022292211651802063,
+ "learning_rate": 0.0006,
+ "loss": 4.783066749572754,
+ "step": 1033
+ },
+ {
+ "epoch": 14.363318777292577,
+ "grad_norm": 0.02140096202492714,
+ "learning_rate": 0.0006,
+ "loss": 4.604718208312988,
+ "step": 1034
+ },
+ {
+ "epoch": 14.377292576419213,
+ "grad_norm": 0.0187817495316267,
+ "learning_rate": 0.0006,
+ "loss": 4.721773147583008,
+ "step": 1035
+ },
+ {
+ "epoch": 14.391266375545852,
+ "grad_norm": 0.019879210740327835,
+ "learning_rate": 0.0006,
+ "loss": 4.789286136627197,
+ "step": 1036
+ },
+ {
+ "epoch": 14.405240174672489,
+ "grad_norm": 0.01885785162448883,
+ "learning_rate": 0.0006,
+ "loss": 4.669743537902832,
+ "step": 1037
+ },
+ {
+ "epoch": 14.419213973799126,
+ "grad_norm": 0.018324270844459534,
+ "learning_rate": 0.0006,
+ "loss": 4.65901517868042,
+ "step": 1038
+ },
+ {
+ "epoch": 14.433187772925764,
+ "grad_norm": 0.017664821818470955,
+ "learning_rate": 0.0006,
+ "loss": 4.667839527130127,
+ "step": 1039
+ },
+ {
+ "epoch": 14.447161572052401,
+ "grad_norm": 0.015702905133366585,
+ "learning_rate": 0.0006,
+ "loss": 4.64700984954834,
+ "step": 1040
+ },
+ {
+ "epoch": 14.46113537117904,
+ "grad_norm": 0.016674194484949112,
+ "learning_rate": 0.0006,
+ "loss": 4.694093227386475,
+ "step": 1041
+ },
+ {
+ "epoch": 14.475109170305677,
+ "grad_norm": 0.016869477927684784,
+ "learning_rate": 0.0006,
+ "loss": 4.619544982910156,
+ "step": 1042
+ },
+ {
+ "epoch": 14.489082969432314,
+ "grad_norm": 0.016557540744543076,
+ "learning_rate": 0.0006,
+ "loss": 4.568667411804199,
+ "step": 1043
+ },
+ {
+ "epoch": 14.503056768558952,
+ "grad_norm": 0.014500945806503296,
+ "learning_rate": 0.0006,
+ "loss": 4.561741828918457,
+ "step": 1044
+ },
+ {
+ "epoch": 14.51703056768559,
+ "grad_norm": 0.014228739775717258,
+ "learning_rate": 0.0006,
+ "loss": 4.654277324676514,
+ "step": 1045
+ },
+ {
+ "epoch": 14.531004366812226,
+ "grad_norm": 0.013894238509237766,
+ "learning_rate": 0.0006,
+ "loss": 4.5867719650268555,
+ "step": 1046
+ },
+ {
+ "epoch": 14.544978165938865,
+ "grad_norm": 0.01550068985670805,
+ "learning_rate": 0.0006,
+ "loss": 4.61168098449707,
+ "step": 1047
+ },
+ {
+ "epoch": 14.558951965065502,
+ "grad_norm": 0.015940772369503975,
+ "learning_rate": 0.0006,
+ "loss": 4.691247940063477,
+ "step": 1048
+ },
+ {
+ "epoch": 14.57292576419214,
+ "grad_norm": 0.01714535616338253,
+ "learning_rate": 0.0006,
+ "loss": 4.61183500289917,
+ "step": 1049
+ },
+ {
+ "epoch": 14.586899563318777,
+ "grad_norm": 0.017925038933753967,
+ "learning_rate": 0.0006,
+ "loss": 4.719022750854492,
+ "step": 1050
+ },
+ {
+ "epoch": 14.600873362445414,
+ "grad_norm": 0.018141552805900574,
+ "learning_rate": 0.0006,
+ "loss": 4.760281562805176,
+ "step": 1051
+ },
+ {
+ "epoch": 14.614847161572053,
+ "grad_norm": 0.017131328582763672,
+ "learning_rate": 0.0006,
+ "loss": 4.709680080413818,
+ "step": 1052
+ },
+ {
+ "epoch": 14.62882096069869,
+ "grad_norm": 0.018730048090219498,
+ "learning_rate": 0.0006,
+ "loss": 4.5994720458984375,
+ "step": 1053
+ },
+ {
+ "epoch": 14.642794759825328,
+ "grad_norm": 0.018894298002123833,
+ "learning_rate": 0.0006,
+ "loss": 4.669070720672607,
+ "step": 1054
+ },
+ {
+ "epoch": 14.656768558951965,
+ "grad_norm": 0.017643198370933533,
+ "learning_rate": 0.0006,
+ "loss": 4.641790390014648,
+ "step": 1055
+ },
+ {
+ "epoch": 14.670742358078602,
+ "grad_norm": 0.017647244036197662,
+ "learning_rate": 0.0006,
+ "loss": 4.653899192810059,
+ "step": 1056
+ },
+ {
+ "epoch": 14.68471615720524,
+ "grad_norm": 0.017304198816418648,
+ "learning_rate": 0.0006,
+ "loss": 4.7487473487854,
+ "step": 1057
+ },
+ {
+ "epoch": 14.698689956331878,
+ "grad_norm": 0.01638447679579258,
+ "learning_rate": 0.0006,
+ "loss": 4.71945858001709,
+ "step": 1058
+ },
+ {
+ "epoch": 14.712663755458514,
+ "grad_norm": 0.015892071649432182,
+ "learning_rate": 0.0006,
+ "loss": 4.744903564453125,
+ "step": 1059
+ },
+ {
+ "epoch": 14.726637554585153,
+ "grad_norm": 0.017420614138245583,
+ "learning_rate": 0.0006,
+ "loss": 4.781563758850098,
+ "step": 1060
+ },
+ {
+ "epoch": 14.74061135371179,
+ "grad_norm": 0.016498520970344543,
+ "learning_rate": 0.0006,
+ "loss": 4.622779846191406,
+ "step": 1061
+ },
+ {
+ "epoch": 14.754585152838429,
+ "grad_norm": 0.017482127994298935,
+ "learning_rate": 0.0006,
+ "loss": 4.62825870513916,
+ "step": 1062
+ },
+ {
+ "epoch": 14.768558951965066,
+ "grad_norm": 0.019018379971385002,
+ "learning_rate": 0.0006,
+ "loss": 4.693194389343262,
+ "step": 1063
+ },
+ {
+ "epoch": 14.782532751091702,
+ "grad_norm": 0.01993914321064949,
+ "learning_rate": 0.0006,
+ "loss": 4.742903709411621,
+ "step": 1064
+ },
+ {
+ "epoch": 14.796506550218341,
+ "grad_norm": 0.02301335707306862,
+ "learning_rate": 0.0006,
+ "loss": 4.613251686096191,
+ "step": 1065
+ },
+ {
+ "epoch": 14.810480349344978,
+ "grad_norm": 0.024088222533464432,
+ "learning_rate": 0.0006,
+ "loss": 4.594305992126465,
+ "step": 1066
+ },
+ {
+ "epoch": 14.824454148471617,
+ "grad_norm": 0.02493320405483246,
+ "learning_rate": 0.0006,
+ "loss": 4.700588226318359,
+ "step": 1067
+ },
+ {
+ "epoch": 14.838427947598253,
+ "grad_norm": 0.02501937560737133,
+ "learning_rate": 0.0006,
+ "loss": 4.601742744445801,
+ "step": 1068
+ },
+ {
+ "epoch": 14.85240174672489,
+ "grad_norm": 0.023490751162171364,
+ "learning_rate": 0.0006,
+ "loss": 4.589211463928223,
+ "step": 1069
+ },
+ {
+ "epoch": 14.866375545851529,
+ "grad_norm": 0.019695665687322617,
+ "learning_rate": 0.0006,
+ "loss": 4.635081768035889,
+ "step": 1070
+ },
+ {
+ "epoch": 14.880349344978166,
+ "grad_norm": 0.019276980310678482,
+ "learning_rate": 0.0006,
+ "loss": 4.737002372741699,
+ "step": 1071
+ },
+ {
+ "epoch": 14.894323144104803,
+ "grad_norm": 0.022430511191487312,
+ "learning_rate": 0.0006,
+ "loss": 4.808295726776123,
+ "step": 1072
+ },
+ {
+ "epoch": 14.908296943231441,
+ "grad_norm": 0.021543916314840317,
+ "learning_rate": 0.0006,
+ "loss": 4.559475421905518,
+ "step": 1073
+ },
+ {
+ "epoch": 14.922270742358078,
+ "grad_norm": 0.019885241985321045,
+ "learning_rate": 0.0006,
+ "loss": 4.647948741912842,
+ "step": 1074
+ },
+ {
+ "epoch": 14.936244541484717,
+ "grad_norm": 0.02115175500512123,
+ "learning_rate": 0.0006,
+ "loss": 4.680505752563477,
+ "step": 1075
+ },
+ {
+ "epoch": 14.950218340611354,
+ "grad_norm": 0.02156151458621025,
+ "learning_rate": 0.0006,
+ "loss": 4.673477649688721,
+ "step": 1076
+ },
+ {
+ "epoch": 14.96419213973799,
+ "grad_norm": 0.02184601128101349,
+ "learning_rate": 0.0006,
+ "loss": 4.644402027130127,
+ "step": 1077
+ },
+ {
+ "epoch": 14.97816593886463,
+ "grad_norm": 0.0211983360350132,
+ "learning_rate": 0.0006,
+ "loss": 4.821558475494385,
+ "step": 1078
+ },
+ {
+ "epoch": 14.992139737991266,
+ "grad_norm": 0.019636567682027817,
+ "learning_rate": 0.0006,
+ "loss": 4.578868865966797,
+ "step": 1079
+ },
+ {
+ "epoch": 15.0,
+ "grad_norm": 0.019997350871562958,
+ "learning_rate": 0.0006,
+ "loss": 4.691307544708252,
+ "step": 1080
+ },
+ {
+ "epoch": 15.0,
+ "eval_loss": 4.872478008270264,
+ "eval_runtime": 56.8513,
+ "eval_samples_per_second": 42.954,
+ "eval_steps_per_second": 1.354,
+ "step": 1080
+ },
+ {
+ "epoch": 15.013973799126637,
+ "grad_norm": 0.018830662593245506,
+ "learning_rate": 0.0006,
+ "loss": 4.761779308319092,
+ "step": 1081
+ },
+ {
+ "epoch": 15.027947598253276,
+ "grad_norm": 0.016876481473445892,
+ "learning_rate": 0.0006,
+ "loss": 4.61148738861084,
+ "step": 1082
+ },
+ {
+ "epoch": 15.041921397379912,
+ "grad_norm": 0.015550863929092884,
+ "learning_rate": 0.0006,
+ "loss": 4.676124095916748,
+ "step": 1083
+ },
+ {
+ "epoch": 15.055895196506551,
+ "grad_norm": 0.017178896814584732,
+ "learning_rate": 0.0006,
+ "loss": 4.588935852050781,
+ "step": 1084
+ },
+ {
+ "epoch": 15.069868995633188,
+ "grad_norm": 0.015465447679162025,
+ "learning_rate": 0.0006,
+ "loss": 4.605499267578125,
+ "step": 1085
+ },
+ {
+ "epoch": 15.083842794759825,
+ "grad_norm": 0.01606195978820324,
+ "learning_rate": 0.0006,
+ "loss": 4.723373889923096,
+ "step": 1086
+ },
+ {
+ "epoch": 15.097816593886463,
+ "grad_norm": 0.016521936282515526,
+ "learning_rate": 0.0006,
+ "loss": 4.646608352661133,
+ "step": 1087
+ },
+ {
+ "epoch": 15.1117903930131,
+ "grad_norm": 0.01589205674827099,
+ "learning_rate": 0.0006,
+ "loss": 4.498138427734375,
+ "step": 1088
+ },
+ {
+ "epoch": 15.125764192139737,
+ "grad_norm": 0.01524856686592102,
+ "learning_rate": 0.0006,
+ "loss": 4.594562530517578,
+ "step": 1089
+ },
+ {
+ "epoch": 15.139737991266376,
+ "grad_norm": 0.014714641496539116,
+ "learning_rate": 0.0006,
+ "loss": 4.630744934082031,
+ "step": 1090
+ },
+ {
+ "epoch": 15.153711790393013,
+ "grad_norm": 0.015333331190049648,
+ "learning_rate": 0.0006,
+ "loss": 4.682388782501221,
+ "step": 1091
+ },
+ {
+ "epoch": 15.167685589519651,
+ "grad_norm": 0.01550073828548193,
+ "learning_rate": 0.0006,
+ "loss": 4.692652702331543,
+ "step": 1092
+ },
+ {
+ "epoch": 15.181659388646288,
+ "grad_norm": 0.014783318154513836,
+ "learning_rate": 0.0006,
+ "loss": 4.6223835945129395,
+ "step": 1093
+ },
+ {
+ "epoch": 15.195633187772925,
+ "grad_norm": 0.014372751116752625,
+ "learning_rate": 0.0006,
+ "loss": 4.52687406539917,
+ "step": 1094
+ },
+ {
+ "epoch": 15.209606986899564,
+ "grad_norm": 0.014779130928218365,
+ "learning_rate": 0.0006,
+ "loss": 4.637009620666504,
+ "step": 1095
+ },
+ {
+ "epoch": 15.2235807860262,
+ "grad_norm": 0.015697073191404343,
+ "learning_rate": 0.0006,
+ "loss": 4.682981014251709,
+ "step": 1096
+ },
+ {
+ "epoch": 15.237554585152838,
+ "grad_norm": 0.016523664817214012,
+ "learning_rate": 0.0006,
+ "loss": 4.705722808837891,
+ "step": 1097
+ },
+ {
+ "epoch": 15.251528384279476,
+ "grad_norm": 0.017633996903896332,
+ "learning_rate": 0.0006,
+ "loss": 4.574686050415039,
+ "step": 1098
+ },
+ {
+ "epoch": 15.265502183406113,
+ "grad_norm": 0.019955076277256012,
+ "learning_rate": 0.0006,
+ "loss": 4.634450912475586,
+ "step": 1099
+ },
+ {
+ "epoch": 15.279475982532752,
+ "grad_norm": 0.02132618986070156,
+ "learning_rate": 0.0006,
+ "loss": 4.56110143661499,
+ "step": 1100
+ },
+ {
+ "epoch": 15.293449781659389,
+ "grad_norm": 0.0207593385130167,
+ "learning_rate": 0.0006,
+ "loss": 4.629396438598633,
+ "step": 1101
+ },
+ {
+ "epoch": 15.307423580786025,
+ "grad_norm": 0.017882540822029114,
+ "learning_rate": 0.0006,
+ "loss": 4.730963230133057,
+ "step": 1102
+ },
+ {
+ "epoch": 15.321397379912664,
+ "grad_norm": 0.01917876861989498,
+ "learning_rate": 0.0006,
+ "loss": 4.64585542678833,
+ "step": 1103
+ },
+ {
+ "epoch": 15.335371179039301,
+ "grad_norm": 0.01943974196910858,
+ "learning_rate": 0.0006,
+ "loss": 4.656181335449219,
+ "step": 1104
+ },
+ {
+ "epoch": 15.34934497816594,
+ "grad_norm": 0.019904715940356255,
+ "learning_rate": 0.0006,
+ "loss": 4.619329452514648,
+ "step": 1105
+ },
+ {
+ "epoch": 15.363318777292577,
+ "grad_norm": 0.019174346700310707,
+ "learning_rate": 0.0006,
+ "loss": 4.647892475128174,
+ "step": 1106
+ },
+ {
+ "epoch": 15.377292576419213,
+ "grad_norm": 0.018244758248329163,
+ "learning_rate": 0.0006,
+ "loss": 4.56166934967041,
+ "step": 1107
+ },
+ {
+ "epoch": 15.391266375545852,
+ "grad_norm": 0.018914205953478813,
+ "learning_rate": 0.0006,
+ "loss": 4.648063659667969,
+ "step": 1108
+ },
+ {
+ "epoch": 15.405240174672489,
+ "grad_norm": 0.01961071416735649,
+ "learning_rate": 0.0006,
+ "loss": 4.665525436401367,
+ "step": 1109
+ },
+ {
+ "epoch": 15.419213973799126,
+ "grad_norm": 0.01841311901807785,
+ "learning_rate": 0.0006,
+ "loss": 4.611342430114746,
+ "step": 1110
+ },
+ {
+ "epoch": 15.433187772925764,
+ "grad_norm": 0.01863427273929119,
+ "learning_rate": 0.0006,
+ "loss": 4.664256572723389,
+ "step": 1111
+ },
+ {
+ "epoch": 15.447161572052401,
+ "grad_norm": 0.020506877452135086,
+ "learning_rate": 0.0006,
+ "loss": 4.631237506866455,
+ "step": 1112
+ },
+ {
+ "epoch": 15.46113537117904,
+ "grad_norm": 0.02223833277821541,
+ "learning_rate": 0.0006,
+ "loss": 4.668654441833496,
+ "step": 1113
+ },
+ {
+ "epoch": 15.475109170305677,
+ "grad_norm": 0.023336702957749367,
+ "learning_rate": 0.0006,
+ "loss": 4.562833309173584,
+ "step": 1114
+ },
+ {
+ "epoch": 15.489082969432314,
+ "grad_norm": 0.023052413016557693,
+ "learning_rate": 0.0006,
+ "loss": 4.621532917022705,
+ "step": 1115
+ },
+ {
+ "epoch": 15.503056768558952,
+ "grad_norm": 0.023420924320816994,
+ "learning_rate": 0.0006,
+ "loss": 4.665465354919434,
+ "step": 1116
+ },
+ {
+ "epoch": 15.51703056768559,
+ "grad_norm": 0.024071281775832176,
+ "learning_rate": 0.0006,
+ "loss": 4.595980644226074,
+ "step": 1117
+ },
+ {
+ "epoch": 15.531004366812226,
+ "grad_norm": 0.023406682536005974,
+ "learning_rate": 0.0006,
+ "loss": 4.642635345458984,
+ "step": 1118
+ },
+ {
+ "epoch": 15.544978165938865,
+ "grad_norm": 0.02138647809624672,
+ "learning_rate": 0.0006,
+ "loss": 4.704513072967529,
+ "step": 1119
+ },
+ {
+ "epoch": 15.558951965065502,
+ "grad_norm": 0.0221384447067976,
+ "learning_rate": 0.0006,
+ "loss": 4.567503452301025,
+ "step": 1120
+ },
+ {
+ "epoch": 15.57292576419214,
+ "grad_norm": 0.020879851654171944,
+ "learning_rate": 0.0006,
+ "loss": 4.664057731628418,
+ "step": 1121
+ },
+ {
+ "epoch": 15.586899563318777,
+ "grad_norm": 0.021478967741131783,
+ "learning_rate": 0.0006,
+ "loss": 4.539627552032471,
+ "step": 1122
+ },
+ {
+ "epoch": 15.600873362445414,
+ "grad_norm": 0.022748656570911407,
+ "learning_rate": 0.0006,
+ "loss": 4.5654497146606445,
+ "step": 1123
+ },
+ {
+ "epoch": 15.614847161572053,
+ "grad_norm": 0.021870248019695282,
+ "learning_rate": 0.0006,
+ "loss": 4.56899356842041,
+ "step": 1124
+ },
+ {
+ "epoch": 15.62882096069869,
+ "grad_norm": 0.021711355075240135,
+ "learning_rate": 0.0006,
+ "loss": 4.657808303833008,
+ "step": 1125
+ },
+ {
+ "epoch": 15.642794759825328,
+ "grad_norm": 0.022810909897089005,
+ "learning_rate": 0.0006,
+ "loss": 4.6526689529418945,
+ "step": 1126
+ },
+ {
+ "epoch": 15.656768558951965,
+ "grad_norm": 0.022561343386769295,
+ "learning_rate": 0.0006,
+ "loss": 4.534887790679932,
+ "step": 1127
+ },
+ {
+ "epoch": 15.670742358078602,
+ "grad_norm": 0.019909802824258804,
+ "learning_rate": 0.0006,
+ "loss": 4.600790023803711,
+ "step": 1128
+ },
+ {
+ "epoch": 15.68471615720524,
+ "grad_norm": 0.020092325285077095,
+ "learning_rate": 0.0006,
+ "loss": 4.773830413818359,
+ "step": 1129
+ },
+ {
+ "epoch": 15.698689956331878,
+ "grad_norm": 0.018834445625543594,
+ "learning_rate": 0.0006,
+ "loss": 4.628868103027344,
+ "step": 1130
+ },
+ {
+ "epoch": 15.712663755458514,
+ "grad_norm": 0.017552798613905907,
+ "learning_rate": 0.0006,
+ "loss": 4.647500038146973,
+ "step": 1131
+ },
+ {
+ "epoch": 15.726637554585153,
+ "grad_norm": 0.01751251146197319,
+ "learning_rate": 0.0006,
+ "loss": 4.590545654296875,
+ "step": 1132
+ },
+ {
+ "epoch": 15.74061135371179,
+ "grad_norm": 0.01699099875986576,
+ "learning_rate": 0.0006,
+ "loss": 4.509699821472168,
+ "step": 1133
+ },
+ {
+ "epoch": 15.754585152838429,
+ "grad_norm": 0.015808649361133575,
+ "learning_rate": 0.0006,
+ "loss": 4.606992244720459,
+ "step": 1134
+ },
+ {
+ "epoch": 15.768558951965066,
+ "grad_norm": 0.01655443385243416,
+ "learning_rate": 0.0006,
+ "loss": 4.585219860076904,
+ "step": 1135
+ },
+ {
+ "epoch": 15.782532751091702,
+ "grad_norm": 0.01708163693547249,
+ "learning_rate": 0.0006,
+ "loss": 4.668147087097168,
+ "step": 1136
+ },
+ {
+ "epoch": 15.796506550218341,
+ "grad_norm": 0.017161913216114044,
+ "learning_rate": 0.0006,
+ "loss": 4.7170634269714355,
+ "step": 1137
+ },
+ {
+ "epoch": 15.810480349344978,
+ "grad_norm": 0.017262674868106842,
+ "learning_rate": 0.0006,
+ "loss": 4.572132110595703,
+ "step": 1138
+ },
+ {
+ "epoch": 15.824454148471617,
+ "grad_norm": 0.017713576555252075,
+ "learning_rate": 0.0006,
+ "loss": 4.702538967132568,
+ "step": 1139
+ },
+ {
+ "epoch": 15.838427947598253,
+ "grad_norm": 0.018730733543634415,
+ "learning_rate": 0.0006,
+ "loss": 4.629721641540527,
+ "step": 1140
+ },
+ {
+ "epoch": 15.85240174672489,
+ "grad_norm": 0.017960187047719955,
+ "learning_rate": 0.0006,
+ "loss": 4.661039352416992,
+ "step": 1141
+ },
+ {
+ "epoch": 15.866375545851529,
+ "grad_norm": 0.01750079356133938,
+ "learning_rate": 0.0006,
+ "loss": 4.503710746765137,
+ "step": 1142
+ },
+ {
+ "epoch": 15.880349344978166,
+ "grad_norm": 0.016630422323942184,
+ "learning_rate": 0.0006,
+ "loss": 4.6582136154174805,
+ "step": 1143
+ },
+ {
+ "epoch": 15.894323144104803,
+ "grad_norm": 0.015896832570433617,
+ "learning_rate": 0.0006,
+ "loss": 4.596693992614746,
+ "step": 1144
+ },
+ {
+ "epoch": 15.908296943231441,
+ "grad_norm": 0.01566407084465027,
+ "learning_rate": 0.0006,
+ "loss": 4.6337199211120605,
+ "step": 1145
+ },
+ {
+ "epoch": 15.922270742358078,
+ "grad_norm": 0.016823330894112587,
+ "learning_rate": 0.0006,
+ "loss": 4.686994552612305,
+ "step": 1146
+ },
+ {
+ "epoch": 15.936244541484717,
+ "grad_norm": 0.016208255663514137,
+ "learning_rate": 0.0006,
+ "loss": 4.6205549240112305,
+ "step": 1147
+ },
+ {
+ "epoch": 15.950218340611354,
+ "grad_norm": 0.0171707421541214,
+ "learning_rate": 0.0006,
+ "loss": 4.582968711853027,
+ "step": 1148
+ },
+ {
+ "epoch": 15.96419213973799,
+ "grad_norm": 0.017009446397423744,
+ "learning_rate": 0.0006,
+ "loss": 4.510274887084961,
+ "step": 1149
+ },
+ {
+ "epoch": 15.97816593886463,
+ "grad_norm": 0.01565811224281788,
+ "learning_rate": 0.0006,
+ "loss": 4.55834436416626,
+ "step": 1150
+ },
+ {
+ "epoch": 15.992139737991266,
+ "grad_norm": 0.015276037156581879,
+ "learning_rate": 0.0006,
+ "loss": 4.537870407104492,
+ "step": 1151
+ },
+ {
+ "epoch": 16.0,
+ "grad_norm": 0.017305882647633553,
+ "learning_rate": 0.0006,
+ "loss": 4.611305236816406,
+ "step": 1152
+ },
+ {
+ "epoch": 16.0,
+ "eval_loss": 4.775513172149658,
+ "eval_runtime": 56.7341,
+ "eval_samples_per_second": 43.043,
+ "eval_steps_per_second": 1.357,
+ "step": 1152
+ },
+ {
+ "epoch": 16.01397379912664,
+ "grad_norm": 0.016752436757087708,
+ "learning_rate": 0.0006,
+ "loss": 4.541312217712402,
+ "step": 1153
+ },
+ {
+ "epoch": 16.027947598253274,
+ "grad_norm": 0.01643134094774723,
+ "learning_rate": 0.0006,
+ "loss": 4.585402965545654,
+ "step": 1154
+ },
+ {
+ "epoch": 16.041921397379912,
+ "grad_norm": 0.01888720504939556,
+ "learning_rate": 0.0006,
+ "loss": 4.597825050354004,
+ "step": 1155
+ },
+ {
+ "epoch": 16.05589519650655,
+ "grad_norm": 0.022044632583856583,
+ "learning_rate": 0.0006,
+ "loss": 4.532772064208984,
+ "step": 1156
+ },
+ {
+ "epoch": 16.069868995633186,
+ "grad_norm": 0.022986987605690956,
+ "learning_rate": 0.0006,
+ "loss": 4.551351547241211,
+ "step": 1157
+ },
+ {
+ "epoch": 16.083842794759825,
+ "grad_norm": 0.020116791129112244,
+ "learning_rate": 0.0006,
+ "loss": 4.484149932861328,
+ "step": 1158
+ },
+ {
+ "epoch": 16.097816593886463,
+ "grad_norm": 0.021442247554659843,
+ "learning_rate": 0.0006,
+ "loss": 4.515316486358643,
+ "step": 1159
+ },
+ {
+ "epoch": 16.111790393013102,
+ "grad_norm": 0.020974772050976753,
+ "learning_rate": 0.0006,
+ "loss": 4.51606559753418,
+ "step": 1160
+ },
+ {
+ "epoch": 16.125764192139737,
+ "grad_norm": 0.021274864673614502,
+ "learning_rate": 0.0006,
+ "loss": 4.587329864501953,
+ "step": 1161
+ },
+ {
+ "epoch": 16.139737991266376,
+ "grad_norm": 0.022799784317612648,
+ "learning_rate": 0.0006,
+ "loss": 4.568149089813232,
+ "step": 1162
+ },
+ {
+ "epoch": 16.153711790393015,
+ "grad_norm": 0.021784933283925056,
+ "learning_rate": 0.0006,
+ "loss": 4.515336990356445,
+ "step": 1163
+ },
+ {
+ "epoch": 16.16768558951965,
+ "grad_norm": 0.024261271581053734,
+ "learning_rate": 0.0006,
+ "loss": 4.5127081871032715,
+ "step": 1164
+ },
+ {
+ "epoch": 16.18165938864629,
+ "grad_norm": 0.02248375490307808,
+ "learning_rate": 0.0006,
+ "loss": 4.754251003265381,
+ "step": 1165
+ },
+ {
+ "epoch": 16.195633187772927,
+ "grad_norm": 0.02136492356657982,
+ "learning_rate": 0.0006,
+ "loss": 4.58193826675415,
+ "step": 1166
+ },
+ {
+ "epoch": 16.209606986899562,
+ "grad_norm": 0.020727066323161125,
+ "learning_rate": 0.0006,
+ "loss": 4.564209938049316,
+ "step": 1167
+ },
+ {
+ "epoch": 16.2235807860262,
+ "grad_norm": 0.017789119854569435,
+ "learning_rate": 0.0006,
+ "loss": 4.698680877685547,
+ "step": 1168
+ },
+ {
+ "epoch": 16.23755458515284,
+ "grad_norm": 0.017603034153580666,
+ "learning_rate": 0.0006,
+ "loss": 4.501171588897705,
+ "step": 1169
+ },
+ {
+ "epoch": 16.251528384279474,
+ "grad_norm": 0.017345862463116646,
+ "learning_rate": 0.0006,
+ "loss": 4.593590259552002,
+ "step": 1170
+ },
+ {
+ "epoch": 16.265502183406113,
+ "grad_norm": 0.018047425895929337,
+ "learning_rate": 0.0006,
+ "loss": 4.597467422485352,
+ "step": 1171
+ },
+ {
+ "epoch": 16.27947598253275,
+ "grad_norm": 0.01947222836315632,
+ "learning_rate": 0.0006,
+ "loss": 4.5072221755981445,
+ "step": 1172
+ },
+ {
+ "epoch": 16.29344978165939,
+ "grad_norm": 0.019598500803112984,
+ "learning_rate": 0.0006,
+ "loss": 4.601351737976074,
+ "step": 1173
+ },
+ {
+ "epoch": 16.307423580786025,
+ "grad_norm": 0.020002514123916626,
+ "learning_rate": 0.0006,
+ "loss": 4.580431938171387,
+ "step": 1174
+ },
+ {
+ "epoch": 16.321397379912664,
+ "grad_norm": 0.0216530729085207,
+ "learning_rate": 0.0006,
+ "loss": 4.424681663513184,
+ "step": 1175
+ },
+ {
+ "epoch": 16.335371179039303,
+ "grad_norm": 0.01949235238134861,
+ "learning_rate": 0.0006,
+ "loss": 4.751883029937744,
+ "step": 1176
+ },
+ {
+ "epoch": 16.349344978165938,
+ "grad_norm": 0.017779016867280006,
+ "learning_rate": 0.0006,
+ "loss": 4.541490077972412,
+ "step": 1177
+ },
+ {
+ "epoch": 16.363318777292577,
+ "grad_norm": 0.01909121684730053,
+ "learning_rate": 0.0006,
+ "loss": 4.594141006469727,
+ "step": 1178
+ },
+ {
+ "epoch": 16.377292576419215,
+ "grad_norm": 0.019894849509000778,
+ "learning_rate": 0.0006,
+ "loss": 4.555070400238037,
+ "step": 1179
+ },
+ {
+ "epoch": 16.39126637554585,
+ "grad_norm": 0.01993032731115818,
+ "learning_rate": 0.0006,
+ "loss": 4.521401405334473,
+ "step": 1180
+ },
+ {
+ "epoch": 16.40524017467249,
+ "grad_norm": 0.0210788045078516,
+ "learning_rate": 0.0006,
+ "loss": 4.672842979431152,
+ "step": 1181
+ },
+ {
+ "epoch": 16.419213973799128,
+ "grad_norm": 0.02274598926305771,
+ "learning_rate": 0.0006,
+ "loss": 4.629504203796387,
+ "step": 1182
+ },
+ {
+ "epoch": 16.433187772925763,
+ "grad_norm": 0.020448120310902596,
+ "learning_rate": 0.0006,
+ "loss": 4.548209190368652,
+ "step": 1183
+ },
+ {
+ "epoch": 16.4471615720524,
+ "grad_norm": 0.017565643414855003,
+ "learning_rate": 0.0006,
+ "loss": 4.529054641723633,
+ "step": 1184
+ },
+ {
+ "epoch": 16.46113537117904,
+ "grad_norm": 0.019504237920045853,
+ "learning_rate": 0.0006,
+ "loss": 4.48182487487793,
+ "step": 1185
+ },
+ {
+ "epoch": 16.475109170305675,
+ "grad_norm": 0.016823742538690567,
+ "learning_rate": 0.0006,
+ "loss": 4.608696937561035,
+ "step": 1186
+ },
+ {
+ "epoch": 16.489082969432314,
+ "grad_norm": 0.01602894812822342,
+ "learning_rate": 0.0006,
+ "loss": 4.620062351226807,
+ "step": 1187
+ },
+ {
+ "epoch": 16.503056768558952,
+ "grad_norm": 0.016615642234683037,
+ "learning_rate": 0.0006,
+ "loss": 4.69378662109375,
+ "step": 1188
+ },
+ {
+ "epoch": 16.51703056768559,
+ "grad_norm": 0.018576730042696,
+ "learning_rate": 0.0006,
+ "loss": 4.537041664123535,
+ "step": 1189
+ },
+ {
+ "epoch": 16.531004366812226,
+ "grad_norm": 0.018611151725053787,
+ "learning_rate": 0.0006,
+ "loss": 4.496169090270996,
+ "step": 1190
+ },
+ {
+ "epoch": 16.544978165938865,
+ "grad_norm": 0.018622413277626038,
+ "learning_rate": 0.0006,
+ "loss": 4.557358264923096,
+ "step": 1191
+ },
+ {
+ "epoch": 16.558951965065503,
+ "grad_norm": 0.01823544315993786,
+ "learning_rate": 0.0006,
+ "loss": 4.730792999267578,
+ "step": 1192
+ },
+ {
+ "epoch": 16.57292576419214,
+ "grad_norm": 0.018796846270561218,
+ "learning_rate": 0.0006,
+ "loss": 4.781791687011719,
+ "step": 1193
+ },
+ {
+ "epoch": 16.586899563318777,
+ "grad_norm": 0.01728292554616928,
+ "learning_rate": 0.0006,
+ "loss": 4.515257835388184,
+ "step": 1194
+ },
+ {
+ "epoch": 16.600873362445416,
+ "grad_norm": 0.01699589006602764,
+ "learning_rate": 0.0006,
+ "loss": 4.5996904373168945,
+ "step": 1195
+ },
+ {
+ "epoch": 16.61484716157205,
+ "grad_norm": 0.015905970707535744,
+ "learning_rate": 0.0006,
+ "loss": 4.601301193237305,
+ "step": 1196
+ },
+ {
+ "epoch": 16.62882096069869,
+ "grad_norm": 0.01338571310043335,
+ "learning_rate": 0.0006,
+ "loss": 4.557708740234375,
+ "step": 1197
+ },
+ {
+ "epoch": 16.64279475982533,
+ "grad_norm": 0.015298855490982533,
+ "learning_rate": 0.0006,
+ "loss": 4.542287349700928,
+ "step": 1198
+ },
+ {
+ "epoch": 16.656768558951963,
+ "grad_norm": 0.014753367751836777,
+ "learning_rate": 0.0006,
+ "loss": 4.595787048339844,
+ "step": 1199
+ },
+ {
+ "epoch": 16.670742358078602,
+ "grad_norm": 0.015527483075857162,
+ "learning_rate": 0.0006,
+ "loss": 4.607529163360596,
+ "step": 1200
+ },
+ {
+ "epoch": 16.68471615720524,
+ "grad_norm": 0.015042162500321865,
+ "learning_rate": 0.0006,
+ "loss": 4.686626434326172,
+ "step": 1201
+ },
+ {
+ "epoch": 16.69868995633188,
+ "grad_norm": 0.0159393772482872,
+ "learning_rate": 0.0006,
+ "loss": 4.47656774520874,
+ "step": 1202
+ },
+ {
+ "epoch": 16.712663755458514,
+ "grad_norm": 0.01532787922769785,
+ "learning_rate": 0.0006,
+ "loss": 4.55678653717041,
+ "step": 1203
+ },
+ {
+ "epoch": 16.726637554585153,
+ "grad_norm": 0.016332164406776428,
+ "learning_rate": 0.0006,
+ "loss": 4.51441764831543,
+ "step": 1204
+ },
+ {
+ "epoch": 16.74061135371179,
+ "grad_norm": 0.018901726230978966,
+ "learning_rate": 0.0006,
+ "loss": 4.57227897644043,
+ "step": 1205
+ },
+ {
+ "epoch": 16.754585152838427,
+ "grad_norm": 0.018589433282613754,
+ "learning_rate": 0.0006,
+ "loss": 4.550070762634277,
+ "step": 1206
+ },
+ {
+ "epoch": 16.768558951965066,
+ "grad_norm": 0.01634034886956215,
+ "learning_rate": 0.0006,
+ "loss": 4.586910724639893,
+ "step": 1207
+ },
+ {
+ "epoch": 16.782532751091704,
+ "grad_norm": 0.017599433660507202,
+ "learning_rate": 0.0006,
+ "loss": 4.5891923904418945,
+ "step": 1208
+ },
+ {
+ "epoch": 16.79650655021834,
+ "grad_norm": 0.01878446340560913,
+ "learning_rate": 0.0006,
+ "loss": 4.57227087020874,
+ "step": 1209
+ },
+ {
+ "epoch": 16.810480349344978,
+ "grad_norm": 0.01803469844162464,
+ "learning_rate": 0.0006,
+ "loss": 4.583845138549805,
+ "step": 1210
+ },
+ {
+ "epoch": 16.824454148471617,
+ "grad_norm": 0.018505144864320755,
+ "learning_rate": 0.0006,
+ "loss": 4.561151504516602,
+ "step": 1211
+ },
+ {
+ "epoch": 16.83842794759825,
+ "grad_norm": 0.01920081488788128,
+ "learning_rate": 0.0006,
+ "loss": 4.506834030151367,
+ "step": 1212
+ },
+ {
+ "epoch": 16.85240174672489,
+ "grad_norm": 0.0177712831646204,
+ "learning_rate": 0.0006,
+ "loss": 4.557195663452148,
+ "step": 1213
+ },
+ {
+ "epoch": 16.86637554585153,
+ "grad_norm": 0.019113918766379356,
+ "learning_rate": 0.0006,
+ "loss": 4.651227951049805,
+ "step": 1214
+ },
+ {
+ "epoch": 16.880349344978168,
+ "grad_norm": 0.02012539468705654,
+ "learning_rate": 0.0006,
+ "loss": 4.4952239990234375,
+ "step": 1215
+ },
+ {
+ "epoch": 16.894323144104803,
+ "grad_norm": 0.019993646070361137,
+ "learning_rate": 0.0006,
+ "loss": 4.517433166503906,
+ "step": 1216
+ },
+ {
+ "epoch": 16.90829694323144,
+ "grad_norm": 0.020041441544890404,
+ "learning_rate": 0.0006,
+ "loss": 4.489407539367676,
+ "step": 1217
+ },
+ {
+ "epoch": 16.92227074235808,
+ "grad_norm": 0.0183611661195755,
+ "learning_rate": 0.0006,
+ "loss": 4.5435638427734375,
+ "step": 1218
+ },
+ {
+ "epoch": 16.936244541484715,
+ "grad_norm": 0.019138429313898087,
+ "learning_rate": 0.0006,
+ "loss": 4.545949459075928,
+ "step": 1219
+ },
+ {
+ "epoch": 16.950218340611354,
+ "grad_norm": 0.018554266542196274,
+ "learning_rate": 0.0006,
+ "loss": 4.579620361328125,
+ "step": 1220
+ },
+ {
+ "epoch": 16.964192139737992,
+ "grad_norm": 0.0193545650690794,
+ "learning_rate": 0.0006,
+ "loss": 4.588255405426025,
+ "step": 1221
+ },
+ {
+ "epoch": 16.978165938864628,
+ "grad_norm": 0.019389165565371513,
+ "learning_rate": 0.0006,
+ "loss": 4.520698547363281,
+ "step": 1222
+ },
+ {
+ "epoch": 16.992139737991266,
+ "grad_norm": 0.016601379960775375,
+ "learning_rate": 0.0006,
+ "loss": 4.4809064865112305,
+ "step": 1223
+ },
+ {
+ "epoch": 17.0,
+ "grad_norm": 0.01771303080022335,
+ "learning_rate": 0.0006,
+ "loss": 4.721122741699219,
+ "step": 1224
+ },
+ {
+ "epoch": 17.0,
+ "eval_loss": 4.7903666496276855,
+ "eval_runtime": 56.9826,
+ "eval_samples_per_second": 42.855,
+ "eval_steps_per_second": 1.351,
+ "step": 1224
+ },
+ {
+ "epoch": 17.01397379912664,
+ "grad_norm": 0.017720289528369904,
+ "learning_rate": 0.0006,
+ "loss": 4.523834705352783,
+ "step": 1225
+ },
+ {
+ "epoch": 17.027947598253274,
+ "grad_norm": 0.020137161016464233,
+ "learning_rate": 0.0006,
+ "loss": 4.574061870574951,
+ "step": 1226
+ },
+ {
+ "epoch": 17.041921397379912,
+ "grad_norm": 0.02002314105629921,
+ "learning_rate": 0.0006,
+ "loss": 4.512407302856445,
+ "step": 1227
+ },
+ {
+ "epoch": 17.05589519650655,
+ "grad_norm": 0.019628183916211128,
+ "learning_rate": 0.0006,
+ "loss": 4.502533435821533,
+ "step": 1228
+ },
+ {
+ "epoch": 17.069868995633186,
+ "grad_norm": 0.019808458164334297,
+ "learning_rate": 0.0006,
+ "loss": 4.463156700134277,
+ "step": 1229
+ },
+ {
+ "epoch": 17.083842794759825,
+ "grad_norm": 0.021250786259770393,
+ "learning_rate": 0.0006,
+ "loss": 4.622682094573975,
+ "step": 1230
+ },
+ {
+ "epoch": 17.097816593886463,
+ "grad_norm": 0.022031694650650024,
+ "learning_rate": 0.0006,
+ "loss": 4.451882839202881,
+ "step": 1231
+ },
+ {
+ "epoch": 17.111790393013102,
+ "grad_norm": 0.01898987777531147,
+ "learning_rate": 0.0006,
+ "loss": 4.529331684112549,
+ "step": 1232
+ },
+ {
+ "epoch": 17.125764192139737,
+ "grad_norm": 0.018749834969639778,
+ "learning_rate": 0.0006,
+ "loss": 4.435815334320068,
+ "step": 1233
+ },
+ {
+ "epoch": 17.139737991266376,
+ "grad_norm": 0.01896674931049347,
+ "learning_rate": 0.0006,
+ "loss": 4.484029293060303,
+ "step": 1234
+ },
+ {
+ "epoch": 17.153711790393015,
+ "grad_norm": 0.018286997452378273,
+ "learning_rate": 0.0006,
+ "loss": 4.523510932922363,
+ "step": 1235
+ },
+ {
+ "epoch": 17.16768558951965,
+ "grad_norm": 0.017273059114813805,
+ "learning_rate": 0.0006,
+ "loss": 4.451767444610596,
+ "step": 1236
+ },
+ {
+ "epoch": 17.18165938864629,
+ "grad_norm": 0.016499150544404984,
+ "learning_rate": 0.0006,
+ "loss": 4.5568647384643555,
+ "step": 1237
+ },
+ {
+ "epoch": 17.195633187772927,
+ "grad_norm": 0.015934469178318977,
+ "learning_rate": 0.0006,
+ "loss": 4.4639997482299805,
+ "step": 1238
+ },
+ {
+ "epoch": 17.209606986899562,
+ "grad_norm": 0.01714209094643593,
+ "learning_rate": 0.0006,
+ "loss": 4.547614097595215,
+ "step": 1239
+ },
+ {
+ "epoch": 17.2235807860262,
+ "grad_norm": 0.019167495891451836,
+ "learning_rate": 0.0006,
+ "loss": 4.571097373962402,
+ "step": 1240
+ },
+ {
+ "epoch": 17.23755458515284,
+ "grad_norm": 0.018276996910572052,
+ "learning_rate": 0.0006,
+ "loss": 4.489764213562012,
+ "step": 1241
+ },
+ {
+ "epoch": 17.251528384279474,
+ "grad_norm": 0.021296866238117218,
+ "learning_rate": 0.0006,
+ "loss": 4.526913642883301,
+ "step": 1242
+ },
+ {
+ "epoch": 17.265502183406113,
+ "grad_norm": 0.022606629878282547,
+ "learning_rate": 0.0006,
+ "loss": 4.521624565124512,
+ "step": 1243
+ },
+ {
+ "epoch": 17.27947598253275,
+ "grad_norm": 0.021204711869359016,
+ "learning_rate": 0.0006,
+ "loss": 4.500864028930664,
+ "step": 1244
+ },
+ {
+ "epoch": 17.29344978165939,
+ "grad_norm": 0.01988849975168705,
+ "learning_rate": 0.0006,
+ "loss": 4.45162296295166,
+ "step": 1245
+ },
+ {
+ "epoch": 17.307423580786025,
+ "grad_norm": 0.020268384367227554,
+ "learning_rate": 0.0006,
+ "loss": 4.536198616027832,
+ "step": 1246
+ },
+ {
+ "epoch": 17.321397379912664,
+ "grad_norm": 0.020185677334666252,
+ "learning_rate": 0.0006,
+ "loss": 4.440262794494629,
+ "step": 1247
+ },
+ {
+ "epoch": 17.335371179039303,
+ "grad_norm": 0.020106853917241096,
+ "learning_rate": 0.0006,
+ "loss": 4.475907325744629,
+ "step": 1248
+ },
+ {
+ "epoch": 17.349344978165938,
+ "grad_norm": 0.018883297219872475,
+ "learning_rate": 0.0006,
+ "loss": 4.640661716461182,
+ "step": 1249
+ },
+ {
+ "epoch": 17.363318777292577,
+ "grad_norm": 0.01852620206773281,
+ "learning_rate": 0.0006,
+ "loss": 4.578243255615234,
+ "step": 1250
+ },
+ {
+ "epoch": 17.377292576419215,
+ "grad_norm": 0.01886451430618763,
+ "learning_rate": 0.0006,
+ "loss": 4.546026229858398,
+ "step": 1251
+ },
+ {
+ "epoch": 17.39126637554585,
+ "grad_norm": 0.01777452416718006,
+ "learning_rate": 0.0006,
+ "loss": 4.521542549133301,
+ "step": 1252
+ },
+ {
+ "epoch": 17.40524017467249,
+ "grad_norm": 0.016626447439193726,
+ "learning_rate": 0.0006,
+ "loss": 4.52629280090332,
+ "step": 1253
+ },
+ {
+ "epoch": 17.419213973799128,
+ "grad_norm": 0.01599477231502533,
+ "learning_rate": 0.0006,
+ "loss": 4.4796037673950195,
+ "step": 1254
+ },
+ {
+ "epoch": 17.433187772925763,
+ "grad_norm": 0.017021458595991135,
+ "learning_rate": 0.0006,
+ "loss": 4.468813419342041,
+ "step": 1255
+ },
+ {
+ "epoch": 17.4471615720524,
+ "grad_norm": 0.016925333067774773,
+ "learning_rate": 0.0006,
+ "loss": 4.584898948669434,
+ "step": 1256
+ },
+ {
+ "epoch": 17.46113537117904,
+ "grad_norm": 0.01714998669922352,
+ "learning_rate": 0.0006,
+ "loss": 4.523544788360596,
+ "step": 1257
+ },
+ {
+ "epoch": 17.475109170305675,
+ "grad_norm": 0.01802288182079792,
+ "learning_rate": 0.0006,
+ "loss": 4.595581531524658,
+ "step": 1258
+ },
+ {
+ "epoch": 17.489082969432314,
+ "grad_norm": 0.0180289875715971,
+ "learning_rate": 0.0006,
+ "loss": 4.394440650939941,
+ "step": 1259
+ },
+ {
+ "epoch": 17.503056768558952,
+ "grad_norm": 0.018504245206713676,
+ "learning_rate": 0.0006,
+ "loss": 4.4794416427612305,
+ "step": 1260
+ },
+ {
+ "epoch": 17.51703056768559,
+ "grad_norm": 0.019973233342170715,
+ "learning_rate": 0.0006,
+ "loss": 4.4241437911987305,
+ "step": 1261
+ },
+ {
+ "epoch": 17.531004366812226,
+ "grad_norm": 0.020063083618879318,
+ "learning_rate": 0.0006,
+ "loss": 4.48068380355835,
+ "step": 1262
+ },
+ {
+ "epoch": 17.544978165938865,
+ "grad_norm": 0.02066640742123127,
+ "learning_rate": 0.0006,
+ "loss": 4.628993034362793,
+ "step": 1263
+ },
+ {
+ "epoch": 17.558951965065503,
+ "grad_norm": 0.02206416428089142,
+ "learning_rate": 0.0006,
+ "loss": 4.564790725708008,
+ "step": 1264
+ },
+ {
+ "epoch": 17.57292576419214,
+ "grad_norm": 0.020640335977077484,
+ "learning_rate": 0.0006,
+ "loss": 4.578705787658691,
+ "step": 1265
+ },
+ {
+ "epoch": 17.586899563318777,
+ "grad_norm": 0.019093405455350876,
+ "learning_rate": 0.0006,
+ "loss": 4.5023603439331055,
+ "step": 1266
+ },
+ {
+ "epoch": 17.600873362445416,
+ "grad_norm": 0.019459193572402,
+ "learning_rate": 0.0006,
+ "loss": 4.594531059265137,
+ "step": 1267
+ },
+ {
+ "epoch": 17.61484716157205,
+ "grad_norm": 0.02021171525120735,
+ "learning_rate": 0.0006,
+ "loss": 4.770299911499023,
+ "step": 1268
+ },
+ {
+ "epoch": 17.62882096069869,
+ "grad_norm": 0.021161191165447235,
+ "learning_rate": 0.0006,
+ "loss": 4.5380401611328125,
+ "step": 1269
+ },
+ {
+ "epoch": 17.64279475982533,
+ "grad_norm": 0.021013403311371803,
+ "learning_rate": 0.0006,
+ "loss": 4.455524921417236,
+ "step": 1270
+ },
+ {
+ "epoch": 17.656768558951963,
+ "grad_norm": 0.020336592569947243,
+ "learning_rate": 0.0006,
+ "loss": 4.533672332763672,
+ "step": 1271
+ },
+ {
+ "epoch": 17.670742358078602,
+ "grad_norm": 0.020159434527158737,
+ "learning_rate": 0.0006,
+ "loss": 4.596151351928711,
+ "step": 1272
+ },
+ {
+ "epoch": 17.68471615720524,
+ "grad_norm": 0.020432960242033005,
+ "learning_rate": 0.0006,
+ "loss": 4.55531644821167,
+ "step": 1273
+ },
+ {
+ "epoch": 17.69868995633188,
+ "grad_norm": 0.025466229766607285,
+ "learning_rate": 0.0006,
+ "loss": 4.566912651062012,
+ "step": 1274
+ },
+ {
+ "epoch": 17.712663755458514,
+ "grad_norm": 0.026537558063864708,
+ "learning_rate": 0.0006,
+ "loss": 4.526003837585449,
+ "step": 1275
+ },
+ {
+ "epoch": 17.726637554585153,
+ "grad_norm": 0.024397607892751694,
+ "learning_rate": 0.0006,
+ "loss": 4.541755676269531,
+ "step": 1276
+ },
+ {
+ "epoch": 17.74061135371179,
+ "grad_norm": 0.022062312811613083,
+ "learning_rate": 0.0006,
+ "loss": 4.6532793045043945,
+ "step": 1277
+ },
+ {
+ "epoch": 17.754585152838427,
+ "grad_norm": 0.02204282023012638,
+ "learning_rate": 0.0006,
+ "loss": 4.582825660705566,
+ "step": 1278
+ },
+ {
+ "epoch": 17.768558951965066,
+ "grad_norm": 0.020844416692852974,
+ "learning_rate": 0.0006,
+ "loss": 4.522336483001709,
+ "step": 1279
+ },
+ {
+ "epoch": 17.782532751091704,
+ "grad_norm": 0.019815703853964806,
+ "learning_rate": 0.0006,
+ "loss": 4.607398509979248,
+ "step": 1280
+ },
+ {
+ "epoch": 17.79650655021834,
+ "grad_norm": 0.016676954925060272,
+ "learning_rate": 0.0006,
+ "loss": 4.446206092834473,
+ "step": 1281
+ },
+ {
+ "epoch": 17.810480349344978,
+ "grad_norm": 0.0166452769190073,
+ "learning_rate": 0.0006,
+ "loss": 4.505931854248047,
+ "step": 1282
+ },
+ {
+ "epoch": 17.824454148471617,
+ "grad_norm": 0.015928907319903374,
+ "learning_rate": 0.0006,
+ "loss": 4.455267429351807,
+ "step": 1283
+ },
+ {
+ "epoch": 17.83842794759825,
+ "grad_norm": 0.015219560824334621,
+ "learning_rate": 0.0006,
+ "loss": 4.5623674392700195,
+ "step": 1284
+ },
+ {
+ "epoch": 17.85240174672489,
+ "grad_norm": 0.015125355683267117,
+ "learning_rate": 0.0006,
+ "loss": 4.554104804992676,
+ "step": 1285
+ },
+ {
+ "epoch": 17.86637554585153,
+ "grad_norm": 0.015612718649208546,
+ "learning_rate": 0.0006,
+ "loss": 4.484847068786621,
+ "step": 1286
+ },
+ {
+ "epoch": 17.880349344978168,
+ "grad_norm": 0.015215662308037281,
+ "learning_rate": 0.0006,
+ "loss": 4.4802093505859375,
+ "step": 1287
+ },
+ {
+ "epoch": 17.894323144104803,
+ "grad_norm": 0.01511684525758028,
+ "learning_rate": 0.0006,
+ "loss": 4.537298679351807,
+ "step": 1288
+ },
+ {
+ "epoch": 17.90829694323144,
+ "grad_norm": 0.013841322623193264,
+ "learning_rate": 0.0006,
+ "loss": 4.427488327026367,
+ "step": 1289
+ },
+ {
+ "epoch": 17.92227074235808,
+ "grad_norm": 0.015115964226424694,
+ "learning_rate": 0.0006,
+ "loss": 4.4547119140625,
+ "step": 1290
+ },
+ {
+ "epoch": 17.936244541484715,
+ "grad_norm": 0.016411006450653076,
+ "learning_rate": 0.0006,
+ "loss": 4.52994966506958,
+ "step": 1291
+ },
+ {
+ "epoch": 17.950218340611354,
+ "grad_norm": 0.017313152551651,
+ "learning_rate": 0.0006,
+ "loss": 4.5337653160095215,
+ "step": 1292
+ },
+ {
+ "epoch": 17.964192139737992,
+ "grad_norm": 0.016559796407818794,
+ "learning_rate": 0.0006,
+ "loss": 4.5277910232543945,
+ "step": 1293
+ },
+ {
+ "epoch": 17.978165938864628,
+ "grad_norm": 0.015854761004447937,
+ "learning_rate": 0.0006,
+ "loss": 4.6334309577941895,
+ "step": 1294
+ },
+ {
+ "epoch": 17.992139737991266,
+ "grad_norm": 0.015658238902688026,
+ "learning_rate": 0.0006,
+ "loss": 4.613372802734375,
+ "step": 1295
+ },
+ {
+ "epoch": 18.0,
+ "grad_norm": 0.0169155802577734,
+ "learning_rate": 0.0006,
+ "loss": 4.527356147766113,
+ "step": 1296
+ },
+ {
+ "epoch": 18.0,
+ "eval_loss": 4.764543056488037,
+ "eval_runtime": 56.5998,
+ "eval_samples_per_second": 43.145,
+ "eval_steps_per_second": 1.36,
+ "step": 1296
+ },
+ {
+ "epoch": 18.01397379912664,
+ "grad_norm": 0.015348963439464569,
+ "learning_rate": 0.0006,
+ "loss": 4.466423988342285,
+ "step": 1297
+ },
+ {
+ "epoch": 18.027947598253274,
+ "grad_norm": 0.01768588088452816,
+ "learning_rate": 0.0006,
+ "loss": 4.545933246612549,
+ "step": 1298
+ },
+ {
+ "epoch": 18.041921397379912,
+ "grad_norm": 0.017381226643919945,
+ "learning_rate": 0.0006,
+ "loss": 4.494060039520264,
+ "step": 1299
+ },
+ {
+ "epoch": 18.05589519650655,
+ "grad_norm": 0.019002335146069527,
+ "learning_rate": 0.0006,
+ "loss": 4.395200729370117,
+ "step": 1300
+ },
+ {
+ "epoch": 18.069868995633186,
+ "grad_norm": 0.01827932707965374,
+ "learning_rate": 0.0006,
+ "loss": 4.424848556518555,
+ "step": 1301
+ },
+ {
+ "epoch": 18.083842794759825,
+ "grad_norm": 0.016383031383156776,
+ "learning_rate": 0.0006,
+ "loss": 4.494722366333008,
+ "step": 1302
+ },
+ {
+ "epoch": 18.097816593886463,
+ "grad_norm": 0.016265157610177994,
+ "learning_rate": 0.0006,
+ "loss": 4.416522979736328,
+ "step": 1303
+ },
+ {
+ "epoch": 18.111790393013102,
+ "grad_norm": 0.01702386513352394,
+ "learning_rate": 0.0006,
+ "loss": 4.600103378295898,
+ "step": 1304
+ },
+ {
+ "epoch": 18.125764192139737,
+ "grad_norm": 0.018495498225092888,
+ "learning_rate": 0.0006,
+ "loss": 4.588401794433594,
+ "step": 1305
+ },
+ {
+ "epoch": 18.139737991266376,
+ "grad_norm": 0.017921684309840202,
+ "learning_rate": 0.0006,
+ "loss": 4.478075981140137,
+ "step": 1306
+ },
+ {
+ "epoch": 18.153711790393015,
+ "grad_norm": 0.01801278069615364,
+ "learning_rate": 0.0006,
+ "loss": 4.413186550140381,
+ "step": 1307
+ },
+ {
+ "epoch": 18.16768558951965,
+ "grad_norm": 0.017211275175213814,
+ "learning_rate": 0.0006,
+ "loss": 4.461341857910156,
+ "step": 1308
+ },
+ {
+ "epoch": 18.18165938864629,
+ "grad_norm": 0.017589328810572624,
+ "learning_rate": 0.0006,
+ "loss": 4.546969413757324,
+ "step": 1309
+ },
+ {
+ "epoch": 18.195633187772927,
+ "grad_norm": 0.01797698810696602,
+ "learning_rate": 0.0006,
+ "loss": 4.530797004699707,
+ "step": 1310
+ },
+ {
+ "epoch": 18.209606986899562,
+ "grad_norm": 0.020408619195222855,
+ "learning_rate": 0.0006,
+ "loss": 4.62999153137207,
+ "step": 1311
+ },
+ {
+ "epoch": 18.2235807860262,
+ "grad_norm": 0.01994038000702858,
+ "learning_rate": 0.0006,
+ "loss": 4.586811065673828,
+ "step": 1312
+ },
+ {
+ "epoch": 18.23755458515284,
+ "grad_norm": 0.019381340593099594,
+ "learning_rate": 0.0006,
+ "loss": 4.4769368171691895,
+ "step": 1313
+ },
+ {
+ "epoch": 18.251528384279474,
+ "grad_norm": 0.020484555512666702,
+ "learning_rate": 0.0006,
+ "loss": 4.524302959442139,
+ "step": 1314
+ },
+ {
+ "epoch": 18.265502183406113,
+ "grad_norm": 0.02426416426897049,
+ "learning_rate": 0.0006,
+ "loss": 4.561705589294434,
+ "step": 1315
+ },
+ {
+ "epoch": 18.27947598253275,
+ "grad_norm": 0.024748101830482483,
+ "learning_rate": 0.0006,
+ "loss": 4.4654316902160645,
+ "step": 1316
+ },
+ {
+ "epoch": 18.29344978165939,
+ "grad_norm": 0.020703328773379326,
+ "learning_rate": 0.0006,
+ "loss": 4.53033447265625,
+ "step": 1317
+ },
+ {
+ "epoch": 18.307423580786025,
+ "grad_norm": 0.01765458658337593,
+ "learning_rate": 0.0006,
+ "loss": 4.5154266357421875,
+ "step": 1318
+ },
+ {
+ "epoch": 18.321397379912664,
+ "grad_norm": 0.018646420910954475,
+ "learning_rate": 0.0006,
+ "loss": 4.5224714279174805,
+ "step": 1319
+ },
+ {
+ "epoch": 18.335371179039303,
+ "grad_norm": 0.01907474733889103,
+ "learning_rate": 0.0006,
+ "loss": 4.5055437088012695,
+ "step": 1320
+ },
+ {
+ "epoch": 18.349344978165938,
+ "grad_norm": 0.018179846927523613,
+ "learning_rate": 0.0006,
+ "loss": 4.462161064147949,
+ "step": 1321
+ },
+ {
+ "epoch": 18.363318777292577,
+ "grad_norm": 0.016697878018021584,
+ "learning_rate": 0.0006,
+ "loss": 4.473718643188477,
+ "step": 1322
+ },
+ {
+ "epoch": 18.377292576419215,
+ "grad_norm": 0.017504019662737846,
+ "learning_rate": 0.0006,
+ "loss": 4.474348068237305,
+ "step": 1323
+ },
+ {
+ "epoch": 18.39126637554585,
+ "grad_norm": 0.018889913335442543,
+ "learning_rate": 0.0006,
+ "loss": 4.4531636238098145,
+ "step": 1324
+ },
+ {
+ "epoch": 18.40524017467249,
+ "grad_norm": 0.019167251884937286,
+ "learning_rate": 0.0006,
+ "loss": 4.4278364181518555,
+ "step": 1325
+ },
+ {
+ "epoch": 18.419213973799128,
+ "grad_norm": 0.015895741060376167,
+ "learning_rate": 0.0006,
+ "loss": 4.474592208862305,
+ "step": 1326
+ },
+ {
+ "epoch": 18.433187772925763,
+ "grad_norm": 0.01531882956624031,
+ "learning_rate": 0.0006,
+ "loss": 4.403757572174072,
+ "step": 1327
+ },
+ {
+ "epoch": 18.4471615720524,
+ "grad_norm": 0.01565849967300892,
+ "learning_rate": 0.0006,
+ "loss": 4.507664680480957,
+ "step": 1328
+ },
+ {
+ "epoch": 18.46113537117904,
+ "grad_norm": 0.01602315716445446,
+ "learning_rate": 0.0006,
+ "loss": 4.5648908615112305,
+ "step": 1329
+ },
+ {
+ "epoch": 18.475109170305675,
+ "grad_norm": 0.016467366367578506,
+ "learning_rate": 0.0006,
+ "loss": 4.428762912750244,
+ "step": 1330
+ },
+ {
+ "epoch": 18.489082969432314,
+ "grad_norm": 0.01604127697646618,
+ "learning_rate": 0.0006,
+ "loss": 4.527771472930908,
+ "step": 1331
+ },
+ {
+ "epoch": 18.503056768558952,
+ "grad_norm": 0.018825586885213852,
+ "learning_rate": 0.0006,
+ "loss": 4.485386848449707,
+ "step": 1332
+ },
+ {
+ "epoch": 18.51703056768559,
+ "grad_norm": 0.019258007407188416,
+ "learning_rate": 0.0006,
+ "loss": 4.414972305297852,
+ "step": 1333
+ },
+ {
+ "epoch": 18.531004366812226,
+ "grad_norm": 0.018192021176218987,
+ "learning_rate": 0.0006,
+ "loss": 4.412166595458984,
+ "step": 1334
+ },
+ {
+ "epoch": 18.544978165938865,
+ "grad_norm": 0.016120698302984238,
+ "learning_rate": 0.0006,
+ "loss": 4.484317779541016,
+ "step": 1335
+ },
+ {
+ "epoch": 18.558951965065503,
+ "grad_norm": 0.016496622934937477,
+ "learning_rate": 0.0006,
+ "loss": 4.514986515045166,
+ "step": 1336
+ },
+ {
+ "epoch": 18.57292576419214,
+ "grad_norm": 0.01795116998255253,
+ "learning_rate": 0.0006,
+ "loss": 4.41473388671875,
+ "step": 1337
+ },
+ {
+ "epoch": 18.586899563318777,
+ "grad_norm": 0.019062276929616928,
+ "learning_rate": 0.0006,
+ "loss": 4.601996898651123,
+ "step": 1338
+ },
+ {
+ "epoch": 18.600873362445416,
+ "grad_norm": 0.017174631357192993,
+ "learning_rate": 0.0006,
+ "loss": 4.571218490600586,
+ "step": 1339
+ },
+ {
+ "epoch": 18.61484716157205,
+ "grad_norm": 0.014015606604516506,
+ "learning_rate": 0.0006,
+ "loss": 4.526606559753418,
+ "step": 1340
+ },
+ {
+ "epoch": 18.62882096069869,
+ "grad_norm": 0.015358510427176952,
+ "learning_rate": 0.0006,
+ "loss": 4.604286193847656,
+ "step": 1341
+ },
+ {
+ "epoch": 18.64279475982533,
+ "grad_norm": 0.01619391329586506,
+ "learning_rate": 0.0006,
+ "loss": 4.560883522033691,
+ "step": 1342
+ },
+ {
+ "epoch": 18.656768558951963,
+ "grad_norm": 0.015972964465618134,
+ "learning_rate": 0.0006,
+ "loss": 4.479755401611328,
+ "step": 1343
+ },
+ {
+ "epoch": 18.670742358078602,
+ "grad_norm": 0.016861025243997574,
+ "learning_rate": 0.0006,
+ "loss": 4.474471092224121,
+ "step": 1344
+ },
+ {
+ "epoch": 18.68471615720524,
+ "grad_norm": 0.01822078227996826,
+ "learning_rate": 0.0006,
+ "loss": 4.509489059448242,
+ "step": 1345
+ },
+ {
+ "epoch": 18.69868995633188,
+ "grad_norm": 0.019415007904171944,
+ "learning_rate": 0.0006,
+ "loss": 4.508941173553467,
+ "step": 1346
+ },
+ {
+ "epoch": 18.712663755458514,
+ "grad_norm": 0.019856126978993416,
+ "learning_rate": 0.0006,
+ "loss": 4.446670055389404,
+ "step": 1347
+ },
+ {
+ "epoch": 18.726637554585153,
+ "grad_norm": 0.01842520199716091,
+ "learning_rate": 0.0006,
+ "loss": 4.524422645568848,
+ "step": 1348
+ },
+ {
+ "epoch": 18.74061135371179,
+ "grad_norm": 0.017196480184793472,
+ "learning_rate": 0.0006,
+ "loss": 4.498466968536377,
+ "step": 1349
+ },
+ {
+ "epoch": 18.754585152838427,
+ "grad_norm": 0.01823010854423046,
+ "learning_rate": 0.0006,
+ "loss": 4.534684181213379,
+ "step": 1350
+ },
+ {
+ "epoch": 18.768558951965066,
+ "grad_norm": 0.020398065447807312,
+ "learning_rate": 0.0006,
+ "loss": 4.456494331359863,
+ "step": 1351
+ },
+ {
+ "epoch": 18.782532751091704,
+ "grad_norm": 0.022992262616753578,
+ "learning_rate": 0.0006,
+ "loss": 4.476374626159668,
+ "step": 1352
+ },
+ {
+ "epoch": 18.79650655021834,
+ "grad_norm": 0.02235420234501362,
+ "learning_rate": 0.0006,
+ "loss": 4.587776184082031,
+ "step": 1353
+ },
+ {
+ "epoch": 18.810480349344978,
+ "grad_norm": 0.02093626745045185,
+ "learning_rate": 0.0006,
+ "loss": 4.424172401428223,
+ "step": 1354
+ },
+ {
+ "epoch": 18.824454148471617,
+ "grad_norm": 0.01997302658855915,
+ "learning_rate": 0.0006,
+ "loss": 4.448066711425781,
+ "step": 1355
+ },
+ {
+ "epoch": 18.83842794759825,
+ "grad_norm": 0.018265459686517715,
+ "learning_rate": 0.0006,
+ "loss": 4.436610698699951,
+ "step": 1356
+ },
+ {
+ "epoch": 18.85240174672489,
+ "grad_norm": 0.01673833839595318,
+ "learning_rate": 0.0006,
+ "loss": 4.5502214431762695,
+ "step": 1357
+ },
+ {
+ "epoch": 18.86637554585153,
+ "grad_norm": 0.017908833920955658,
+ "learning_rate": 0.0006,
+ "loss": 4.550530433654785,
+ "step": 1358
+ },
+ {
+ "epoch": 18.880349344978168,
+ "grad_norm": 0.01796654611825943,
+ "learning_rate": 0.0006,
+ "loss": 4.531463146209717,
+ "step": 1359
+ },
+ {
+ "epoch": 18.894323144104803,
+ "grad_norm": 0.01646183803677559,
+ "learning_rate": 0.0006,
+ "loss": 4.498582363128662,
+ "step": 1360
+ },
+ {
+ "epoch": 18.90829694323144,
+ "grad_norm": 0.015430327504873276,
+ "learning_rate": 0.0006,
+ "loss": 4.52571439743042,
+ "step": 1361
+ },
+ {
+ "epoch": 18.92227074235808,
+ "grad_norm": 0.016527952626347542,
+ "learning_rate": 0.0006,
+ "loss": 4.547013759613037,
+ "step": 1362
+ },
+ {
+ "epoch": 18.936244541484715,
+ "grad_norm": 0.016263660043478012,
+ "learning_rate": 0.0006,
+ "loss": 4.468754291534424,
+ "step": 1363
+ },
+ {
+ "epoch": 18.950218340611354,
+ "grad_norm": 0.016347210854291916,
+ "learning_rate": 0.0006,
+ "loss": 4.450741767883301,
+ "step": 1364
+ },
+ {
+ "epoch": 18.964192139737992,
+ "grad_norm": 0.01701398566365242,
+ "learning_rate": 0.0006,
+ "loss": 4.534899711608887,
+ "step": 1365
+ },
+ {
+ "epoch": 18.978165938864628,
+ "grad_norm": 0.017693081870675087,
+ "learning_rate": 0.0006,
+ "loss": 4.460553169250488,
+ "step": 1366
+ },
+ {
+ "epoch": 18.992139737991266,
+ "grad_norm": 0.018746308982372284,
+ "learning_rate": 0.0006,
+ "loss": 4.538792610168457,
+ "step": 1367
+ },
+ {
+ "epoch": 19.0,
+ "grad_norm": 0.020099416375160217,
+ "learning_rate": 0.0006,
+ "loss": 4.449617385864258,
+ "step": 1368
+ },
+ {
+ "epoch": 19.0,
+ "eval_loss": 4.7756571769714355,
+ "eval_runtime": 57.1804,
+ "eval_samples_per_second": 42.707,
+ "eval_steps_per_second": 1.347,
+ "step": 1368
+ },
+ {
+ "epoch": 19.01397379912664,
+ "grad_norm": 0.018726933747529984,
+ "learning_rate": 0.0006,
+ "loss": 4.446059703826904,
+ "step": 1369
+ },
+ {
+ "epoch": 19.027947598253274,
+ "grad_norm": 0.01956087350845337,
+ "learning_rate": 0.0006,
+ "loss": 4.4520039558410645,
+ "step": 1370
+ },
+ {
+ "epoch": 19.041921397379912,
+ "grad_norm": 0.020092811435461044,
+ "learning_rate": 0.0006,
+ "loss": 4.543343544006348,
+ "step": 1371
+ },
+ {
+ "epoch": 19.05589519650655,
+ "grad_norm": 0.01843622885644436,
+ "learning_rate": 0.0006,
+ "loss": 4.535579204559326,
+ "step": 1372
+ },
+ {
+ "epoch": 19.069868995633186,
+ "grad_norm": 0.016989488154649734,
+ "learning_rate": 0.0006,
+ "loss": 4.52316951751709,
+ "step": 1373
+ },
+ {
+ "epoch": 19.083842794759825,
+ "grad_norm": 0.01805422455072403,
+ "learning_rate": 0.0006,
+ "loss": 4.5393266677856445,
+ "step": 1374
+ },
+ {
+ "epoch": 19.097816593886463,
+ "grad_norm": 0.01874629408121109,
+ "learning_rate": 0.0006,
+ "loss": 4.422059059143066,
+ "step": 1375
+ },
+ {
+ "epoch": 19.111790393013102,
+ "grad_norm": 0.01804221048951149,
+ "learning_rate": 0.0006,
+ "loss": 4.50508451461792,
+ "step": 1376
+ },
+ {
+ "epoch": 19.125764192139737,
+ "grad_norm": 0.01785474829375744,
+ "learning_rate": 0.0006,
+ "loss": 4.412599563598633,
+ "step": 1377
+ },
+ {
+ "epoch": 19.139737991266376,
+ "grad_norm": 0.01971791312098503,
+ "learning_rate": 0.0006,
+ "loss": 4.518318176269531,
+ "step": 1378
+ },
+ {
+ "epoch": 19.153711790393015,
+ "grad_norm": 0.02061633951961994,
+ "learning_rate": 0.0006,
+ "loss": 4.332393169403076,
+ "step": 1379
+ },
+ {
+ "epoch": 19.16768558951965,
+ "grad_norm": 0.020563429221510887,
+ "learning_rate": 0.0006,
+ "loss": 4.445863723754883,
+ "step": 1380
+ },
+ {
+ "epoch": 19.18165938864629,
+ "grad_norm": 0.019995713606476784,
+ "learning_rate": 0.0006,
+ "loss": 4.497312068939209,
+ "step": 1381
+ },
+ {
+ "epoch": 19.195633187772927,
+ "grad_norm": 0.019823070615530014,
+ "learning_rate": 0.0006,
+ "loss": 4.475085258483887,
+ "step": 1382
+ },
+ {
+ "epoch": 19.209606986899562,
+ "grad_norm": 0.019039355218410492,
+ "learning_rate": 0.0006,
+ "loss": 4.50432014465332,
+ "step": 1383
+ },
+ {
+ "epoch": 19.2235807860262,
+ "grad_norm": 0.017876973375678062,
+ "learning_rate": 0.0006,
+ "loss": 4.384788513183594,
+ "step": 1384
+ },
+ {
+ "epoch": 19.23755458515284,
+ "grad_norm": 0.018697699531912804,
+ "learning_rate": 0.0006,
+ "loss": 4.410778045654297,
+ "step": 1385
+ },
+ {
+ "epoch": 19.251528384279474,
+ "grad_norm": 0.018826806917786598,
+ "learning_rate": 0.0006,
+ "loss": 4.524034023284912,
+ "step": 1386
+ },
+ {
+ "epoch": 19.265502183406113,
+ "grad_norm": 0.01824226602911949,
+ "learning_rate": 0.0006,
+ "loss": 4.526393890380859,
+ "step": 1387
+ },
+ {
+ "epoch": 19.27947598253275,
+ "grad_norm": 0.01753208599984646,
+ "learning_rate": 0.0006,
+ "loss": 4.493703365325928,
+ "step": 1388
+ },
+ {
+ "epoch": 19.29344978165939,
+ "grad_norm": 0.01742432825267315,
+ "learning_rate": 0.0006,
+ "loss": 4.392831802368164,
+ "step": 1389
+ },
+ {
+ "epoch": 19.307423580786025,
+ "grad_norm": 0.017099356278777122,
+ "learning_rate": 0.0006,
+ "loss": 4.578668594360352,
+ "step": 1390
+ },
+ {
+ "epoch": 19.321397379912664,
+ "grad_norm": 0.017946023494005203,
+ "learning_rate": 0.0006,
+ "loss": 4.456718444824219,
+ "step": 1391
+ },
+ {
+ "epoch": 19.335371179039303,
+ "grad_norm": 0.017705943435430527,
+ "learning_rate": 0.0006,
+ "loss": 4.553682327270508,
+ "step": 1392
+ },
+ {
+ "epoch": 19.349344978165938,
+ "grad_norm": 0.016647523269057274,
+ "learning_rate": 0.0006,
+ "loss": 4.525016784667969,
+ "step": 1393
+ },
+ {
+ "epoch": 19.363318777292577,
+ "grad_norm": 0.015407416969537735,
+ "learning_rate": 0.0006,
+ "loss": 4.500231742858887,
+ "step": 1394
+ },
+ {
+ "epoch": 19.377292576419215,
+ "grad_norm": 0.017264019697904587,
+ "learning_rate": 0.0006,
+ "loss": 4.4615349769592285,
+ "step": 1395
+ },
+ {
+ "epoch": 19.39126637554585,
+ "grad_norm": 0.01871555484831333,
+ "learning_rate": 0.0006,
+ "loss": 4.281026840209961,
+ "step": 1396
+ },
+ {
+ "epoch": 19.40524017467249,
+ "grad_norm": 0.017361650243401527,
+ "learning_rate": 0.0006,
+ "loss": 4.479382514953613,
+ "step": 1397
+ },
+ {
+ "epoch": 19.419213973799128,
+ "grad_norm": 0.019781362265348434,
+ "learning_rate": 0.0006,
+ "loss": 4.402320861816406,
+ "step": 1398
+ },
+ {
+ "epoch": 19.433187772925763,
+ "grad_norm": 0.019043298438191414,
+ "learning_rate": 0.0006,
+ "loss": 4.476983070373535,
+ "step": 1399
+ },
+ {
+ "epoch": 19.4471615720524,
+ "grad_norm": 0.017137985676527023,
+ "learning_rate": 0.0006,
+ "loss": 4.418689250946045,
+ "step": 1400
+ },
+ {
+ "epoch": 19.46113537117904,
+ "grad_norm": 0.016481833532452583,
+ "learning_rate": 0.0006,
+ "loss": 4.579145431518555,
+ "step": 1401
+ },
+ {
+ "epoch": 19.475109170305675,
+ "grad_norm": 0.016507714986801147,
+ "learning_rate": 0.0006,
+ "loss": 4.519158363342285,
+ "step": 1402
+ },
+ {
+ "epoch": 19.489082969432314,
+ "grad_norm": 0.018092872574925423,
+ "learning_rate": 0.0006,
+ "loss": 4.513748645782471,
+ "step": 1403
+ },
+ {
+ "epoch": 19.503056768558952,
+ "grad_norm": 0.018153440207242966,
+ "learning_rate": 0.0006,
+ "loss": 4.645384311676025,
+ "step": 1404
+ },
+ {
+ "epoch": 19.51703056768559,
+ "grad_norm": 0.01657119207084179,
+ "learning_rate": 0.0006,
+ "loss": 4.399757385253906,
+ "step": 1405
+ },
+ {
+ "epoch": 19.531004366812226,
+ "grad_norm": 0.0156280305236578,
+ "learning_rate": 0.0006,
+ "loss": 4.467836380004883,
+ "step": 1406
+ },
+ {
+ "epoch": 19.544978165938865,
+ "grad_norm": 0.016838403418660164,
+ "learning_rate": 0.0006,
+ "loss": 4.529176235198975,
+ "step": 1407
+ },
+ {
+ "epoch": 19.558951965065503,
+ "grad_norm": 0.01743236929178238,
+ "learning_rate": 0.0006,
+ "loss": 4.440032482147217,
+ "step": 1408
+ },
+ {
+ "epoch": 19.57292576419214,
+ "grad_norm": 0.01553610060364008,
+ "learning_rate": 0.0006,
+ "loss": 4.425588607788086,
+ "step": 1409
+ },
+ {
+ "epoch": 19.586899563318777,
+ "grad_norm": 0.017259271815419197,
+ "learning_rate": 0.0006,
+ "loss": 4.405452728271484,
+ "step": 1410
+ },
+ {
+ "epoch": 19.600873362445416,
+ "grad_norm": 0.015568481758236885,
+ "learning_rate": 0.0006,
+ "loss": 4.444279193878174,
+ "step": 1411
+ },
+ {
+ "epoch": 19.61484716157205,
+ "grad_norm": 0.015331248752772808,
+ "learning_rate": 0.0006,
+ "loss": 4.451846122741699,
+ "step": 1412
+ },
+ {
+ "epoch": 19.62882096069869,
+ "grad_norm": 0.015667378902435303,
+ "learning_rate": 0.0006,
+ "loss": 4.533931255340576,
+ "step": 1413
+ },
+ {
+ "epoch": 19.64279475982533,
+ "grad_norm": 0.016177602112293243,
+ "learning_rate": 0.0006,
+ "loss": 4.467313289642334,
+ "step": 1414
+ },
+ {
+ "epoch": 19.656768558951963,
+ "grad_norm": 0.01733129471540451,
+ "learning_rate": 0.0006,
+ "loss": 4.496893882751465,
+ "step": 1415
+ },
+ {
+ "epoch": 19.670742358078602,
+ "grad_norm": 0.016539594158530235,
+ "learning_rate": 0.0006,
+ "loss": 4.401961326599121,
+ "step": 1416
+ },
+ {
+ "epoch": 19.68471615720524,
+ "grad_norm": 0.016058551147580147,
+ "learning_rate": 0.0006,
+ "loss": 4.503375053405762,
+ "step": 1417
+ },
+ {
+ "epoch": 19.69868995633188,
+ "grad_norm": 0.01715777814388275,
+ "learning_rate": 0.0006,
+ "loss": 4.3808369636535645,
+ "step": 1418
+ },
+ {
+ "epoch": 19.712663755458514,
+ "grad_norm": 0.01681119203567505,
+ "learning_rate": 0.0006,
+ "loss": 4.5469560623168945,
+ "step": 1419
+ },
+ {
+ "epoch": 19.726637554585153,
+ "grad_norm": 0.01634550653398037,
+ "learning_rate": 0.0006,
+ "loss": 4.328197956085205,
+ "step": 1420
+ },
+ {
+ "epoch": 19.74061135371179,
+ "grad_norm": 0.01549347210675478,
+ "learning_rate": 0.0006,
+ "loss": 4.534621238708496,
+ "step": 1421
+ },
+ {
+ "epoch": 19.754585152838427,
+ "grad_norm": 0.01654665172100067,
+ "learning_rate": 0.0006,
+ "loss": 4.4724225997924805,
+ "step": 1422
+ },
+ {
+ "epoch": 19.768558951965066,
+ "grad_norm": 0.017119025811553,
+ "learning_rate": 0.0006,
+ "loss": 4.3897247314453125,
+ "step": 1423
+ },
+ {
+ "epoch": 19.782532751091704,
+ "grad_norm": 0.01603223942220211,
+ "learning_rate": 0.0006,
+ "loss": 4.412731170654297,
+ "step": 1424
+ },
+ {
+ "epoch": 19.79650655021834,
+ "grad_norm": 0.016774067655205727,
+ "learning_rate": 0.0006,
+ "loss": 4.54707145690918,
+ "step": 1425
+ },
+ {
+ "epoch": 19.810480349344978,
+ "grad_norm": 0.016453025862574577,
+ "learning_rate": 0.0006,
+ "loss": 4.462501049041748,
+ "step": 1426
+ },
+ {
+ "epoch": 19.824454148471617,
+ "grad_norm": 0.01597629487514496,
+ "learning_rate": 0.0006,
+ "loss": 4.229758262634277,
+ "step": 1427
+ },
+ {
+ "epoch": 19.83842794759825,
+ "grad_norm": 0.017999034374952316,
+ "learning_rate": 0.0006,
+ "loss": 4.389263153076172,
+ "step": 1428
+ },
+ {
+ "epoch": 19.85240174672489,
+ "grad_norm": 0.02165898121893406,
+ "learning_rate": 0.0006,
+ "loss": 4.342130661010742,
+ "step": 1429
+ },
+ {
+ "epoch": 19.86637554585153,
+ "grad_norm": 0.02222822792828083,
+ "learning_rate": 0.0006,
+ "loss": 4.492661476135254,
+ "step": 1430
+ },
+ {
+ "epoch": 19.880349344978168,
+ "grad_norm": 0.019378066062927246,
+ "learning_rate": 0.0006,
+ "loss": 4.4614105224609375,
+ "step": 1431
+ },
+ {
+ "epoch": 19.894323144104803,
+ "grad_norm": 0.019908083602786064,
+ "learning_rate": 0.0006,
+ "loss": 4.370929718017578,
+ "step": 1432
+ },
+ {
+ "epoch": 19.90829694323144,
+ "grad_norm": 0.022296471521258354,
+ "learning_rate": 0.0006,
+ "loss": 4.438320159912109,
+ "step": 1433
+ },
+ {
+ "epoch": 19.92227074235808,
+ "grad_norm": 0.021810343489050865,
+ "learning_rate": 0.0006,
+ "loss": 4.462067127227783,
+ "step": 1434
+ },
+ {
+ "epoch": 19.936244541484715,
+ "grad_norm": 0.02129209227859974,
+ "learning_rate": 0.0006,
+ "loss": 4.541261196136475,
+ "step": 1435
+ },
+ {
+ "epoch": 19.950218340611354,
+ "grad_norm": 0.02079629711806774,
+ "learning_rate": 0.0006,
+ "loss": 4.36712646484375,
+ "step": 1436
+ },
+ {
+ "epoch": 19.964192139737992,
+ "grad_norm": 0.022501740604639053,
+ "learning_rate": 0.0006,
+ "loss": 4.5279083251953125,
+ "step": 1437
+ },
+ {
+ "epoch": 19.978165938864628,
+ "grad_norm": 0.019121676683425903,
+ "learning_rate": 0.0006,
+ "loss": 4.477789878845215,
+ "step": 1438
+ },
+ {
+ "epoch": 19.992139737991266,
+ "grad_norm": 0.017429588362574577,
+ "learning_rate": 0.0006,
+ "loss": 4.4023847579956055,
+ "step": 1439
+ },
+ {
+ "epoch": 20.0,
+ "grad_norm": 0.018857665359973907,
+ "learning_rate": 0.0006,
+ "loss": 4.494743347167969,
+ "step": 1440
+ },
+ {
+ "epoch": 20.0,
+ "eval_loss": 4.694555759429932,
+ "eval_runtime": 56.4146,
+ "eval_samples_per_second": 43.287,
+ "eval_steps_per_second": 1.365,
+ "step": 1440
+ },
+ {
+ "epoch": 20.01397379912664,
+ "grad_norm": 0.016505861654877663,
+ "learning_rate": 0.0006,
+ "loss": 4.506229877471924,
+ "step": 1441
+ },
+ {
+ "epoch": 20.027947598253274,
+ "grad_norm": 0.017571842297911644,
+ "learning_rate": 0.0006,
+ "loss": 4.384073257446289,
+ "step": 1442
+ },
+ {
+ "epoch": 20.041921397379912,
+ "grad_norm": 0.019398365169763565,
+ "learning_rate": 0.0006,
+ "loss": 4.482177734375,
+ "step": 1443
+ },
+ {
+ "epoch": 20.05589519650655,
+ "grad_norm": 0.02011062018573284,
+ "learning_rate": 0.0006,
+ "loss": 4.440773963928223,
+ "step": 1444
+ },
+ {
+ "epoch": 20.069868995633186,
+ "grad_norm": 0.019026953727006912,
+ "learning_rate": 0.0006,
+ "loss": 4.507168769836426,
+ "step": 1445
+ },
+ {
+ "epoch": 20.083842794759825,
+ "grad_norm": 0.020506031811237335,
+ "learning_rate": 0.0006,
+ "loss": 4.469079971313477,
+ "step": 1446
+ },
+ {
+ "epoch": 20.097816593886463,
+ "grad_norm": 0.019608162343502045,
+ "learning_rate": 0.0006,
+ "loss": 4.395927429199219,
+ "step": 1447
+ },
+ {
+ "epoch": 20.111790393013102,
+ "grad_norm": 0.019419124349951744,
+ "learning_rate": 0.0006,
+ "loss": 4.484235763549805,
+ "step": 1448
+ },
+ {
+ "epoch": 20.125764192139737,
+ "grad_norm": 0.021775128319859505,
+ "learning_rate": 0.0006,
+ "loss": 4.333173751831055,
+ "step": 1449
+ },
+ {
+ "epoch": 20.139737991266376,
+ "grad_norm": 0.025512272492051125,
+ "learning_rate": 0.0006,
+ "loss": 4.460362434387207,
+ "step": 1450
+ },
+ {
+ "epoch": 20.153711790393015,
+ "grad_norm": 0.025157401338219643,
+ "learning_rate": 0.0006,
+ "loss": 4.4361395835876465,
+ "step": 1451
+ },
+ {
+ "epoch": 20.16768558951965,
+ "grad_norm": 0.019862616434693336,
+ "learning_rate": 0.0006,
+ "loss": 4.426807403564453,
+ "step": 1452
+ },
+ {
+ "epoch": 20.18165938864629,
+ "grad_norm": 0.01932152360677719,
+ "learning_rate": 0.0006,
+ "loss": 4.375677108764648,
+ "step": 1453
+ },
+ {
+ "epoch": 20.195633187772927,
+ "grad_norm": 0.01760999858379364,
+ "learning_rate": 0.0006,
+ "loss": 4.42259407043457,
+ "step": 1454
+ },
+ {
+ "epoch": 20.209606986899562,
+ "grad_norm": 0.016998112201690674,
+ "learning_rate": 0.0006,
+ "loss": 4.481280326843262,
+ "step": 1455
+ },
+ {
+ "epoch": 20.2235807860262,
+ "grad_norm": 0.017199190333485603,
+ "learning_rate": 0.0006,
+ "loss": 4.496712684631348,
+ "step": 1456
+ },
+ {
+ "epoch": 20.23755458515284,
+ "grad_norm": 0.016833679750561714,
+ "learning_rate": 0.0006,
+ "loss": 4.40736722946167,
+ "step": 1457
+ },
+ {
+ "epoch": 20.251528384279474,
+ "grad_norm": 0.017546730116009712,
+ "learning_rate": 0.0006,
+ "loss": 4.4915266036987305,
+ "step": 1458
+ },
+ {
+ "epoch": 20.265502183406113,
+ "grad_norm": 0.016595078632235527,
+ "learning_rate": 0.0006,
+ "loss": 4.460815906524658,
+ "step": 1459
+ },
+ {
+ "epoch": 20.27947598253275,
+ "grad_norm": 0.017728568986058235,
+ "learning_rate": 0.0006,
+ "loss": 4.346048831939697,
+ "step": 1460
+ },
+ {
+ "epoch": 20.29344978165939,
+ "grad_norm": 0.01849197782576084,
+ "learning_rate": 0.0006,
+ "loss": 4.399681091308594,
+ "step": 1461
+ },
+ {
+ "epoch": 20.307423580786025,
+ "grad_norm": 0.018336588516831398,
+ "learning_rate": 0.0006,
+ "loss": 4.414183616638184,
+ "step": 1462
+ },
+ {
+ "epoch": 20.321397379912664,
+ "grad_norm": 0.018086519092321396,
+ "learning_rate": 0.0006,
+ "loss": 4.346586227416992,
+ "step": 1463
+ },
+ {
+ "epoch": 20.335371179039303,
+ "grad_norm": 0.019697299227118492,
+ "learning_rate": 0.0006,
+ "loss": 4.4420952796936035,
+ "step": 1464
+ },
+ {
+ "epoch": 20.349344978165938,
+ "grad_norm": 0.02198074571788311,
+ "learning_rate": 0.0006,
+ "loss": 4.588729381561279,
+ "step": 1465
+ },
+ {
+ "epoch": 20.363318777292577,
+ "grad_norm": 0.022813329473137856,
+ "learning_rate": 0.0006,
+ "loss": 4.49729585647583,
+ "step": 1466
+ },
+ {
+ "epoch": 20.377292576419215,
+ "grad_norm": 0.022006649523973465,
+ "learning_rate": 0.0006,
+ "loss": 4.475641250610352,
+ "step": 1467
+ },
+ {
+ "epoch": 20.39126637554585,
+ "grad_norm": 0.023188438266515732,
+ "learning_rate": 0.0006,
+ "loss": 4.404322624206543,
+ "step": 1468
+ },
+ {
+ "epoch": 20.40524017467249,
+ "grad_norm": 0.022226877510547638,
+ "learning_rate": 0.0006,
+ "loss": 4.50904655456543,
+ "step": 1469
+ },
+ {
+ "epoch": 20.419213973799128,
+ "grad_norm": 0.021082637831568718,
+ "learning_rate": 0.0006,
+ "loss": 4.339011192321777,
+ "step": 1470
+ },
+ {
+ "epoch": 20.433187772925763,
+ "grad_norm": 0.02013542130589485,
+ "learning_rate": 0.0006,
+ "loss": 4.516942024230957,
+ "step": 1471
+ },
+ {
+ "epoch": 20.4471615720524,
+ "grad_norm": 0.01898968778550625,
+ "learning_rate": 0.0006,
+ "loss": 4.315804481506348,
+ "step": 1472
+ },
+ {
+ "epoch": 20.46113537117904,
+ "grad_norm": 0.019450880587100983,
+ "learning_rate": 0.0006,
+ "loss": 4.4794921875,
+ "step": 1473
+ },
+ {
+ "epoch": 20.475109170305675,
+ "grad_norm": 0.018577802926301956,
+ "learning_rate": 0.0006,
+ "loss": 4.422323703765869,
+ "step": 1474
+ },
+ {
+ "epoch": 20.489082969432314,
+ "grad_norm": 0.01778412237763405,
+ "learning_rate": 0.0006,
+ "loss": 4.429789066314697,
+ "step": 1475
+ },
+ {
+ "epoch": 20.503056768558952,
+ "grad_norm": 0.017134476453065872,
+ "learning_rate": 0.0006,
+ "loss": 4.416459083557129,
+ "step": 1476
+ },
+ {
+ "epoch": 20.51703056768559,
+ "grad_norm": 0.017153557389974594,
+ "learning_rate": 0.0006,
+ "loss": 4.446170806884766,
+ "step": 1477
+ },
+ {
+ "epoch": 20.531004366812226,
+ "grad_norm": 0.014998532831668854,
+ "learning_rate": 0.0006,
+ "loss": 4.419447898864746,
+ "step": 1478
+ },
+ {
+ "epoch": 20.544978165938865,
+ "grad_norm": 0.015877850353717804,
+ "learning_rate": 0.0006,
+ "loss": 4.312338829040527,
+ "step": 1479
+ },
+ {
+ "epoch": 20.558951965065503,
+ "grad_norm": 0.01569489948451519,
+ "learning_rate": 0.0006,
+ "loss": 4.572905540466309,
+ "step": 1480
+ },
+ {
+ "epoch": 20.57292576419214,
+ "grad_norm": 0.01591474376618862,
+ "learning_rate": 0.0006,
+ "loss": 4.414907455444336,
+ "step": 1481
+ },
+ {
+ "epoch": 20.586899563318777,
+ "grad_norm": 0.015395903028547764,
+ "learning_rate": 0.0006,
+ "loss": 4.549131393432617,
+ "step": 1482
+ },
+ {
+ "epoch": 20.600873362445416,
+ "grad_norm": 0.015715358778834343,
+ "learning_rate": 0.0006,
+ "loss": 4.530194282531738,
+ "step": 1483
+ },
+ {
+ "epoch": 20.61484716157205,
+ "grad_norm": 0.016042975708842278,
+ "learning_rate": 0.0006,
+ "loss": 4.322149276733398,
+ "step": 1484
+ },
+ {
+ "epoch": 20.62882096069869,
+ "grad_norm": 0.01664186641573906,
+ "learning_rate": 0.0006,
+ "loss": 4.424866676330566,
+ "step": 1485
+ },
+ {
+ "epoch": 20.64279475982533,
+ "grad_norm": 0.01631534844636917,
+ "learning_rate": 0.0006,
+ "loss": 4.441752910614014,
+ "step": 1486
+ },
+ {
+ "epoch": 20.656768558951963,
+ "grad_norm": 0.014057058840990067,
+ "learning_rate": 0.0006,
+ "loss": 4.399982452392578,
+ "step": 1487
+ },
+ {
+ "epoch": 20.670742358078602,
+ "grad_norm": 0.015076273120939732,
+ "learning_rate": 0.0006,
+ "loss": 4.441324234008789,
+ "step": 1488
+ },
+ {
+ "epoch": 20.68471615720524,
+ "grad_norm": 0.014138277620077133,
+ "learning_rate": 0.0006,
+ "loss": 4.4312543869018555,
+ "step": 1489
+ },
+ {
+ "epoch": 20.69868995633188,
+ "grad_norm": 0.013762440532445908,
+ "learning_rate": 0.0006,
+ "loss": 4.387807846069336,
+ "step": 1490
+ },
+ {
+ "epoch": 20.712663755458514,
+ "grad_norm": 0.015058411285281181,
+ "learning_rate": 0.0006,
+ "loss": 4.594747543334961,
+ "step": 1491
+ },
+ {
+ "epoch": 20.726637554585153,
+ "grad_norm": 0.013827123679220676,
+ "learning_rate": 0.0006,
+ "loss": 4.434604167938232,
+ "step": 1492
+ },
+ {
+ "epoch": 20.74061135371179,
+ "grad_norm": 0.015315295197069645,
+ "learning_rate": 0.0006,
+ "loss": 4.588066101074219,
+ "step": 1493
+ },
+ {
+ "epoch": 20.754585152838427,
+ "grad_norm": 0.018678732216358185,
+ "learning_rate": 0.0006,
+ "loss": 4.3514204025268555,
+ "step": 1494
+ },
+ {
+ "epoch": 20.768558951965066,
+ "grad_norm": 0.019521350041031837,
+ "learning_rate": 0.0006,
+ "loss": 4.437658786773682,
+ "step": 1495
+ },
+ {
+ "epoch": 20.782532751091704,
+ "grad_norm": 0.018363403156399727,
+ "learning_rate": 0.0006,
+ "loss": 4.495002269744873,
+ "step": 1496
+ },
+ {
+ "epoch": 20.79650655021834,
+ "grad_norm": 0.016830725595355034,
+ "learning_rate": 0.0006,
+ "loss": 4.402471542358398,
+ "step": 1497
+ },
+ {
+ "epoch": 20.810480349344978,
+ "grad_norm": 0.01722072623670101,
+ "learning_rate": 0.0006,
+ "loss": 4.437653541564941,
+ "step": 1498
+ },
+ {
+ "epoch": 20.824454148471617,
+ "grad_norm": 0.017459599301218987,
+ "learning_rate": 0.0006,
+ "loss": 4.366603851318359,
+ "step": 1499
+ },
+ {
+ "epoch": 20.83842794759825,
+ "grad_norm": 0.018902219831943512,
+ "learning_rate": 0.0006,
+ "loss": 4.501561164855957,
+ "step": 1500
+ },
+ {
+ "epoch": 20.85240174672489,
+ "grad_norm": 0.020814096555113792,
+ "learning_rate": 0.0006,
+ "loss": 4.535295486450195,
+ "step": 1501
+ },
+ {
+ "epoch": 20.86637554585153,
+ "grad_norm": 0.018990976735949516,
+ "learning_rate": 0.0006,
+ "loss": 4.333580017089844,
+ "step": 1502
+ },
+ {
+ "epoch": 20.880349344978168,
+ "grad_norm": 0.01837880350649357,
+ "learning_rate": 0.0006,
+ "loss": 4.313388824462891,
+ "step": 1503
+ },
+ {
+ "epoch": 20.894323144104803,
+ "grad_norm": 0.019820360466837883,
+ "learning_rate": 0.0006,
+ "loss": 4.388157844543457,
+ "step": 1504
+ },
+ {
+ "epoch": 20.90829694323144,
+ "grad_norm": 0.020181143656373024,
+ "learning_rate": 0.0006,
+ "loss": 4.572224140167236,
+ "step": 1505
+ },
+ {
+ "epoch": 20.92227074235808,
+ "grad_norm": 0.021206077188253403,
+ "learning_rate": 0.0006,
+ "loss": 4.3127970695495605,
+ "step": 1506
+ },
+ {
+ "epoch": 20.936244541484715,
+ "grad_norm": 0.017962060868740082,
+ "learning_rate": 0.0006,
+ "loss": 4.384734630584717,
+ "step": 1507
+ },
+ {
+ "epoch": 20.950218340611354,
+ "grad_norm": 0.01737220585346222,
+ "learning_rate": 0.0006,
+ "loss": 4.4075751304626465,
+ "step": 1508
+ },
+ {
+ "epoch": 20.964192139737992,
+ "grad_norm": 0.01839268207550049,
+ "learning_rate": 0.0006,
+ "loss": 4.553779125213623,
+ "step": 1509
+ },
+ {
+ "epoch": 20.978165938864628,
+ "grad_norm": 0.020102640613913536,
+ "learning_rate": 0.0006,
+ "loss": 4.399398326873779,
+ "step": 1510
+ },
+ {
+ "epoch": 20.992139737991266,
+ "grad_norm": 0.018831463530659676,
+ "learning_rate": 0.0006,
+ "loss": 4.358499526977539,
+ "step": 1511
+ },
+ {
+ "epoch": 21.0,
+ "grad_norm": 0.018254734575748444,
+ "learning_rate": 0.0006,
+ "loss": 4.384119510650635,
+ "step": 1512
+ },
+ {
+ "epoch": 21.0,
+ "eval_loss": 4.7524237632751465,
+ "eval_runtime": 57.8812,
+ "eval_samples_per_second": 42.19,
+ "eval_steps_per_second": 1.33,
+ "step": 1512
+ },
+ {
+ "epoch": 21.01397379912664,
+ "grad_norm": 0.019728587940335274,
+ "learning_rate": 0.0006,
+ "loss": 4.32363224029541,
+ "step": 1513
+ },
+ {
+ "epoch": 21.027947598253274,
+ "grad_norm": 0.02216625027358532,
+ "learning_rate": 0.0006,
+ "loss": 4.436988353729248,
+ "step": 1514
+ },
+ {
+ "epoch": 21.041921397379912,
+ "grad_norm": 0.02227606810629368,
+ "learning_rate": 0.0006,
+ "loss": 4.406063556671143,
+ "step": 1515
+ },
+ {
+ "epoch": 21.05589519650655,
+ "grad_norm": 0.020187009125947952,
+ "learning_rate": 0.0006,
+ "loss": 4.324402332305908,
+ "step": 1516
+ },
+ {
+ "epoch": 21.069868995633186,
+ "grad_norm": 0.02095157466828823,
+ "learning_rate": 0.0006,
+ "loss": 4.456544876098633,
+ "step": 1517
+ },
+ {
+ "epoch": 21.083842794759825,
+ "grad_norm": 0.02146013453602791,
+ "learning_rate": 0.0006,
+ "loss": 4.368387222290039,
+ "step": 1518
+ },
+ {
+ "epoch": 21.097816593886463,
+ "grad_norm": 0.02162035182118416,
+ "learning_rate": 0.0006,
+ "loss": 4.48838996887207,
+ "step": 1519
+ },
+ {
+ "epoch": 21.111790393013102,
+ "grad_norm": 0.021745676174759865,
+ "learning_rate": 0.0006,
+ "loss": 4.30778694152832,
+ "step": 1520
+ },
+ {
+ "epoch": 21.125764192139737,
+ "grad_norm": 0.02173588052392006,
+ "learning_rate": 0.0006,
+ "loss": 4.433149814605713,
+ "step": 1521
+ },
+ {
+ "epoch": 21.139737991266376,
+ "grad_norm": 0.020950648933649063,
+ "learning_rate": 0.0006,
+ "loss": 4.520089149475098,
+ "step": 1522
+ },
+ {
+ "epoch": 21.153711790393015,
+ "grad_norm": 0.020684752613306046,
+ "learning_rate": 0.0006,
+ "loss": 4.372437477111816,
+ "step": 1523
+ },
+ {
+ "epoch": 21.16768558951965,
+ "grad_norm": 0.019472869113087654,
+ "learning_rate": 0.0006,
+ "loss": 4.437015533447266,
+ "step": 1524
+ },
+ {
+ "epoch": 21.18165938864629,
+ "grad_norm": 0.01846962980926037,
+ "learning_rate": 0.0006,
+ "loss": 4.449531555175781,
+ "step": 1525
+ },
+ {
+ "epoch": 21.195633187772927,
+ "grad_norm": 0.017377803102135658,
+ "learning_rate": 0.0006,
+ "loss": 4.458279132843018,
+ "step": 1526
+ },
+ {
+ "epoch": 21.209606986899562,
+ "grad_norm": 0.016002200543880463,
+ "learning_rate": 0.0006,
+ "loss": 4.266297817230225,
+ "step": 1527
+ },
+ {
+ "epoch": 21.2235807860262,
+ "grad_norm": 0.015575903467833996,
+ "learning_rate": 0.0006,
+ "loss": 4.342632293701172,
+ "step": 1528
+ },
+ {
+ "epoch": 21.23755458515284,
+ "grad_norm": 0.015634698793292046,
+ "learning_rate": 0.0006,
+ "loss": 4.338623523712158,
+ "step": 1529
+ },
+ {
+ "epoch": 21.251528384279474,
+ "grad_norm": 0.01449581328779459,
+ "learning_rate": 0.0006,
+ "loss": 4.354330062866211,
+ "step": 1530
+ },
+ {
+ "epoch": 21.265502183406113,
+ "grad_norm": 0.01487616915255785,
+ "learning_rate": 0.0006,
+ "loss": 4.486835479736328,
+ "step": 1531
+ },
+ {
+ "epoch": 21.27947598253275,
+ "grad_norm": 0.016346555203199387,
+ "learning_rate": 0.0006,
+ "loss": 4.407098770141602,
+ "step": 1532
+ },
+ {
+ "epoch": 21.29344978165939,
+ "grad_norm": 0.01682182401418686,
+ "learning_rate": 0.0006,
+ "loss": 4.436809539794922,
+ "step": 1533
+ },
+ {
+ "epoch": 21.307423580786025,
+ "grad_norm": 0.017704004421830177,
+ "learning_rate": 0.0006,
+ "loss": 4.3854780197143555,
+ "step": 1534
+ },
+ {
+ "epoch": 21.321397379912664,
+ "grad_norm": 0.018437420949339867,
+ "learning_rate": 0.0006,
+ "loss": 4.413283824920654,
+ "step": 1535
+ },
+ {
+ "epoch": 21.335371179039303,
+ "grad_norm": 0.016911081969738007,
+ "learning_rate": 0.0006,
+ "loss": 4.357306003570557,
+ "step": 1536
+ },
+ {
+ "epoch": 21.349344978165938,
+ "grad_norm": 0.01642223633825779,
+ "learning_rate": 0.0006,
+ "loss": 4.422322750091553,
+ "step": 1537
+ },
+ {
+ "epoch": 21.363318777292577,
+ "grad_norm": 0.01780877821147442,
+ "learning_rate": 0.0006,
+ "loss": 4.387620449066162,
+ "step": 1538
+ },
+ {
+ "epoch": 21.377292576419215,
+ "grad_norm": 0.01653975248336792,
+ "learning_rate": 0.0006,
+ "loss": 4.327267646789551,
+ "step": 1539
+ },
+ {
+ "epoch": 21.39126637554585,
+ "grad_norm": 0.01572156324982643,
+ "learning_rate": 0.0006,
+ "loss": 4.475979804992676,
+ "step": 1540
+ },
+ {
+ "epoch": 21.40524017467249,
+ "grad_norm": 0.017156066372990608,
+ "learning_rate": 0.0006,
+ "loss": 4.4011125564575195,
+ "step": 1541
+ },
+ {
+ "epoch": 21.419213973799128,
+ "grad_norm": 0.01788829080760479,
+ "learning_rate": 0.0006,
+ "loss": 4.4174089431762695,
+ "step": 1542
+ },
+ {
+ "epoch": 21.433187772925763,
+ "grad_norm": 0.01728985086083412,
+ "learning_rate": 0.0006,
+ "loss": 4.429244041442871,
+ "step": 1543
+ },
+ {
+ "epoch": 21.4471615720524,
+ "grad_norm": 0.016534466296434402,
+ "learning_rate": 0.0006,
+ "loss": 4.326068878173828,
+ "step": 1544
+ },
+ {
+ "epoch": 21.46113537117904,
+ "grad_norm": 0.016582584008574486,
+ "learning_rate": 0.0006,
+ "loss": 4.2975969314575195,
+ "step": 1545
+ },
+ {
+ "epoch": 21.475109170305675,
+ "grad_norm": 0.018034178763628006,
+ "learning_rate": 0.0006,
+ "loss": 4.501319408416748,
+ "step": 1546
+ },
+ {
+ "epoch": 21.489082969432314,
+ "grad_norm": 0.017827702686190605,
+ "learning_rate": 0.0006,
+ "loss": 4.502284526824951,
+ "step": 1547
+ },
+ {
+ "epoch": 21.503056768558952,
+ "grad_norm": 0.017182299867272377,
+ "learning_rate": 0.0006,
+ "loss": 4.380828857421875,
+ "step": 1548
+ },
+ {
+ "epoch": 21.51703056768559,
+ "grad_norm": 0.019101588055491447,
+ "learning_rate": 0.0006,
+ "loss": 4.31564998626709,
+ "step": 1549
+ },
+ {
+ "epoch": 21.531004366812226,
+ "grad_norm": 0.0204677302390337,
+ "learning_rate": 0.0006,
+ "loss": 4.3840131759643555,
+ "step": 1550
+ },
+ {
+ "epoch": 21.544978165938865,
+ "grad_norm": 0.0188993439078331,
+ "learning_rate": 0.0006,
+ "loss": 4.4257731437683105,
+ "step": 1551
+ },
+ {
+ "epoch": 21.558951965065503,
+ "grad_norm": 0.01730230636894703,
+ "learning_rate": 0.0006,
+ "loss": 4.445132255554199,
+ "step": 1552
+ },
+ {
+ "epoch": 21.57292576419214,
+ "grad_norm": 0.01801574043929577,
+ "learning_rate": 0.0006,
+ "loss": 4.427064895629883,
+ "step": 1553
+ },
+ {
+ "epoch": 21.586899563318777,
+ "grad_norm": 0.018417565152049065,
+ "learning_rate": 0.0006,
+ "loss": 4.319595813751221,
+ "step": 1554
+ },
+ {
+ "epoch": 21.600873362445416,
+ "grad_norm": 0.01683160290122032,
+ "learning_rate": 0.0006,
+ "loss": 4.349218368530273,
+ "step": 1555
+ },
+ {
+ "epoch": 21.61484716157205,
+ "grad_norm": 0.016973106190562248,
+ "learning_rate": 0.0006,
+ "loss": 4.375990390777588,
+ "step": 1556
+ },
+ {
+ "epoch": 21.62882096069869,
+ "grad_norm": 0.015688583254814148,
+ "learning_rate": 0.0006,
+ "loss": 4.408170223236084,
+ "step": 1557
+ },
+ {
+ "epoch": 21.64279475982533,
+ "grad_norm": 0.01573043316602707,
+ "learning_rate": 0.0006,
+ "loss": 4.416255950927734,
+ "step": 1558
+ },
+ {
+ "epoch": 21.656768558951963,
+ "grad_norm": 0.01744293048977852,
+ "learning_rate": 0.0006,
+ "loss": 4.477141380310059,
+ "step": 1559
+ },
+ {
+ "epoch": 21.670742358078602,
+ "grad_norm": 0.017503513023257256,
+ "learning_rate": 0.0006,
+ "loss": 4.3367486000061035,
+ "step": 1560
+ },
+ {
+ "epoch": 21.68471615720524,
+ "grad_norm": 0.016592150554060936,
+ "learning_rate": 0.0006,
+ "loss": 4.314785480499268,
+ "step": 1561
+ },
+ {
+ "epoch": 21.69868995633188,
+ "grad_norm": 0.015774084255099297,
+ "learning_rate": 0.0006,
+ "loss": 4.448209285736084,
+ "step": 1562
+ },
+ {
+ "epoch": 21.712663755458514,
+ "grad_norm": 0.015779150649905205,
+ "learning_rate": 0.0006,
+ "loss": 4.392690658569336,
+ "step": 1563
+ },
+ {
+ "epoch": 21.726637554585153,
+ "grad_norm": 0.015392648056149483,
+ "learning_rate": 0.0006,
+ "loss": 4.437741756439209,
+ "step": 1564
+ },
+ {
+ "epoch": 21.74061135371179,
+ "grad_norm": 0.01530836895108223,
+ "learning_rate": 0.0006,
+ "loss": 4.424595832824707,
+ "step": 1565
+ },
+ {
+ "epoch": 21.754585152838427,
+ "grad_norm": 0.015600779093801975,
+ "learning_rate": 0.0006,
+ "loss": 4.335417747497559,
+ "step": 1566
+ },
+ {
+ "epoch": 21.768558951965066,
+ "grad_norm": 0.014646518044173717,
+ "learning_rate": 0.0006,
+ "loss": 4.279081344604492,
+ "step": 1567
+ },
+ {
+ "epoch": 21.782532751091704,
+ "grad_norm": 0.014132864773273468,
+ "learning_rate": 0.0006,
+ "loss": 4.345950126647949,
+ "step": 1568
+ },
+ {
+ "epoch": 21.79650655021834,
+ "grad_norm": 0.01425724383443594,
+ "learning_rate": 0.0006,
+ "loss": 4.277588844299316,
+ "step": 1569
+ },
+ {
+ "epoch": 21.810480349344978,
+ "grad_norm": 0.015396572649478912,
+ "learning_rate": 0.0006,
+ "loss": 4.326196670532227,
+ "step": 1570
+ },
+ {
+ "epoch": 21.824454148471617,
+ "grad_norm": 0.01683669537305832,
+ "learning_rate": 0.0006,
+ "loss": 4.3294172286987305,
+ "step": 1571
+ },
+ {
+ "epoch": 21.83842794759825,
+ "grad_norm": 0.016365647315979004,
+ "learning_rate": 0.0006,
+ "loss": 4.523321628570557,
+ "step": 1572
+ },
+ {
+ "epoch": 21.85240174672489,
+ "grad_norm": 0.017142774537205696,
+ "learning_rate": 0.0006,
+ "loss": 4.475842475891113,
+ "step": 1573
+ },
+ {
+ "epoch": 21.86637554585153,
+ "grad_norm": 0.015950839966535568,
+ "learning_rate": 0.0006,
+ "loss": 4.397187232971191,
+ "step": 1574
+ },
+ {
+ "epoch": 21.880349344978168,
+ "grad_norm": 0.016180362552404404,
+ "learning_rate": 0.0006,
+ "loss": 4.414778709411621,
+ "step": 1575
+ },
+ {
+ "epoch": 21.894323144104803,
+ "grad_norm": 0.017615556716918945,
+ "learning_rate": 0.0006,
+ "loss": 4.348567962646484,
+ "step": 1576
+ },
+ {
+ "epoch": 21.90829694323144,
+ "grad_norm": 0.01896277628839016,
+ "learning_rate": 0.0006,
+ "loss": 4.431762218475342,
+ "step": 1577
+ },
+ {
+ "epoch": 21.92227074235808,
+ "grad_norm": 0.01990971714258194,
+ "learning_rate": 0.0006,
+ "loss": 4.431285858154297,
+ "step": 1578
+ },
+ {
+ "epoch": 21.936244541484715,
+ "grad_norm": 0.0201258547604084,
+ "learning_rate": 0.0006,
+ "loss": 4.309872627258301,
+ "step": 1579
+ },
+ {
+ "epoch": 21.950218340611354,
+ "grad_norm": 0.019620198756456375,
+ "learning_rate": 0.0006,
+ "loss": 4.484044551849365,
+ "step": 1580
+ },
+ {
+ "epoch": 21.964192139737992,
+ "grad_norm": 0.01771123707294464,
+ "learning_rate": 0.0006,
+ "loss": 4.368093013763428,
+ "step": 1581
+ },
+ {
+ "epoch": 21.978165938864628,
+ "grad_norm": 0.017501654103398323,
+ "learning_rate": 0.0006,
+ "loss": 4.372651100158691,
+ "step": 1582
+ },
+ {
+ "epoch": 21.992139737991266,
+ "grad_norm": 0.01693608984351158,
+ "learning_rate": 0.0006,
+ "loss": 4.379167556762695,
+ "step": 1583
+ },
+ {
+ "epoch": 22.0,
+ "grad_norm": 0.02002849616110325,
+ "learning_rate": 0.0006,
+ "loss": 4.329627990722656,
+ "step": 1584
+ },
+ {
+ "epoch": 22.0,
+ "eval_loss": 4.723696231842041,
+ "eval_runtime": 56.1758,
+ "eval_samples_per_second": 43.471,
+ "eval_steps_per_second": 1.371,
+ "step": 1584
+ },
+ {
+ "epoch": 22.01397379912664,
+ "grad_norm": 0.02030305378139019,
+ "learning_rate": 0.0006,
+ "loss": 4.355566024780273,
+ "step": 1585
+ },
+ {
+ "epoch": 22.027947598253274,
+ "grad_norm": 0.018638934940099716,
+ "learning_rate": 0.0006,
+ "loss": 4.314009189605713,
+ "step": 1586
+ },
+ {
+ "epoch": 22.041921397379912,
+ "grad_norm": 0.019917072728276253,
+ "learning_rate": 0.0006,
+ "loss": 4.367525577545166,
+ "step": 1587
+ },
+ {
+ "epoch": 22.05589519650655,
+ "grad_norm": 0.020785167813301086,
+ "learning_rate": 0.0006,
+ "loss": 4.375561714172363,
+ "step": 1588
+ },
+ {
+ "epoch": 22.069868995633186,
+ "grad_norm": 0.021214453503489494,
+ "learning_rate": 0.0006,
+ "loss": 4.418305397033691,
+ "step": 1589
+ },
+ {
+ "epoch": 22.083842794759825,
+ "grad_norm": 0.019051678478717804,
+ "learning_rate": 0.0006,
+ "loss": 4.215126991271973,
+ "step": 1590
+ },
+ {
+ "epoch": 22.097816593886463,
+ "grad_norm": 0.02323366142809391,
+ "learning_rate": 0.0006,
+ "loss": 4.306807994842529,
+ "step": 1591
+ },
+ {
+ "epoch": 22.111790393013102,
+ "grad_norm": 0.026186738163232803,
+ "learning_rate": 0.0006,
+ "loss": 4.408508777618408,
+ "step": 1592
+ },
+ {
+ "epoch": 22.125764192139737,
+ "grad_norm": 0.027361227199435234,
+ "learning_rate": 0.0006,
+ "loss": 4.329322814941406,
+ "step": 1593
+ },
+ {
+ "epoch": 22.139737991266376,
+ "grad_norm": 0.026564646512269974,
+ "learning_rate": 0.0006,
+ "loss": 4.346642017364502,
+ "step": 1594
+ },
+ {
+ "epoch": 22.153711790393015,
+ "grad_norm": 0.025073667988181114,
+ "learning_rate": 0.0006,
+ "loss": 4.3984198570251465,
+ "step": 1595
+ },
+ {
+ "epoch": 22.16768558951965,
+ "grad_norm": 0.02347794733941555,
+ "learning_rate": 0.0006,
+ "loss": 4.352538108825684,
+ "step": 1596
+ },
+ {
+ "epoch": 22.18165938864629,
+ "grad_norm": 0.02203468047082424,
+ "learning_rate": 0.0006,
+ "loss": 4.298148155212402,
+ "step": 1597
+ },
+ {
+ "epoch": 22.195633187772927,
+ "grad_norm": 0.020621957257390022,
+ "learning_rate": 0.0006,
+ "loss": 4.480309963226318,
+ "step": 1598
+ },
+ {
+ "epoch": 22.209606986899562,
+ "grad_norm": 0.023710055276751518,
+ "learning_rate": 0.0006,
+ "loss": 4.339317321777344,
+ "step": 1599
+ },
+ {
+ "epoch": 22.2235807860262,
+ "grad_norm": 0.0233482476323843,
+ "learning_rate": 0.0006,
+ "loss": 4.436896324157715,
+ "step": 1600
+ },
+ {
+ "epoch": 22.23755458515284,
+ "grad_norm": 0.020122403278946877,
+ "learning_rate": 0.0006,
+ "loss": 4.236002445220947,
+ "step": 1601
+ },
+ {
+ "epoch": 22.251528384279474,
+ "grad_norm": 0.01957266591489315,
+ "learning_rate": 0.0006,
+ "loss": 4.497328281402588,
+ "step": 1602
+ },
+ {
+ "epoch": 22.265502183406113,
+ "grad_norm": 0.018964895978569984,
+ "learning_rate": 0.0006,
+ "loss": 4.36360502243042,
+ "step": 1603
+ },
+ {
+ "epoch": 22.27947598253275,
+ "grad_norm": 0.018902769312262535,
+ "learning_rate": 0.0006,
+ "loss": 4.427093505859375,
+ "step": 1604
+ },
+ {
+ "epoch": 22.29344978165939,
+ "grad_norm": 0.018918119370937347,
+ "learning_rate": 0.0006,
+ "loss": 4.294950485229492,
+ "step": 1605
+ },
+ {
+ "epoch": 22.307423580786025,
+ "grad_norm": 0.019315775483846664,
+ "learning_rate": 0.0006,
+ "loss": 4.427952289581299,
+ "step": 1606
+ },
+ {
+ "epoch": 22.321397379912664,
+ "grad_norm": 0.018350522965192795,
+ "learning_rate": 0.0006,
+ "loss": 4.272945404052734,
+ "step": 1607
+ },
+ {
+ "epoch": 22.335371179039303,
+ "grad_norm": 0.01759534701704979,
+ "learning_rate": 0.0006,
+ "loss": 4.348023891448975,
+ "step": 1608
+ },
+ {
+ "epoch": 22.349344978165938,
+ "grad_norm": 0.01646057888865471,
+ "learning_rate": 0.0006,
+ "loss": 4.278518199920654,
+ "step": 1609
+ },
+ {
+ "epoch": 22.363318777292577,
+ "grad_norm": 0.015875060111284256,
+ "learning_rate": 0.0006,
+ "loss": 4.423527240753174,
+ "step": 1610
+ },
+ {
+ "epoch": 22.377292576419215,
+ "grad_norm": 0.01638529635965824,
+ "learning_rate": 0.0006,
+ "loss": 4.321723937988281,
+ "step": 1611
+ },
+ {
+ "epoch": 22.39126637554585,
+ "grad_norm": 0.016942061483860016,
+ "learning_rate": 0.0006,
+ "loss": 4.346164703369141,
+ "step": 1612
+ },
+ {
+ "epoch": 22.40524017467249,
+ "grad_norm": 0.015209621749818325,
+ "learning_rate": 0.0006,
+ "loss": 4.434340000152588,
+ "step": 1613
+ },
+ {
+ "epoch": 22.419213973799128,
+ "grad_norm": 0.014423009008169174,
+ "learning_rate": 0.0006,
+ "loss": 4.342560768127441,
+ "step": 1614
+ },
+ {
+ "epoch": 22.433187772925763,
+ "grad_norm": 0.014369525946676731,
+ "learning_rate": 0.0006,
+ "loss": 4.296586990356445,
+ "step": 1615
+ },
+ {
+ "epoch": 22.4471615720524,
+ "grad_norm": 0.01392540242522955,
+ "learning_rate": 0.0006,
+ "loss": 4.433176040649414,
+ "step": 1616
+ },
+ {
+ "epoch": 22.46113537117904,
+ "grad_norm": 0.013206899166107178,
+ "learning_rate": 0.0006,
+ "loss": 4.32962703704834,
+ "step": 1617
+ },
+ {
+ "epoch": 22.475109170305675,
+ "grad_norm": 0.013427077792584896,
+ "learning_rate": 0.0006,
+ "loss": 4.420437335968018,
+ "step": 1618
+ },
+ {
+ "epoch": 22.489082969432314,
+ "grad_norm": 0.014194196090102196,
+ "learning_rate": 0.0006,
+ "loss": 4.410478591918945,
+ "step": 1619
+ },
+ {
+ "epoch": 22.503056768558952,
+ "grad_norm": 0.01362364087253809,
+ "learning_rate": 0.0006,
+ "loss": 4.338932037353516,
+ "step": 1620
+ },
+ {
+ "epoch": 22.51703056768559,
+ "grad_norm": 0.014299589209258556,
+ "learning_rate": 0.0006,
+ "loss": 4.34089994430542,
+ "step": 1621
+ },
+ {
+ "epoch": 22.531004366812226,
+ "grad_norm": 0.014061295427381992,
+ "learning_rate": 0.0006,
+ "loss": 4.29201602935791,
+ "step": 1622
+ },
+ {
+ "epoch": 22.544978165938865,
+ "grad_norm": 0.013945229351520538,
+ "learning_rate": 0.0006,
+ "loss": 4.326052188873291,
+ "step": 1623
+ },
+ {
+ "epoch": 22.558951965065503,
+ "grad_norm": 0.01387068536132574,
+ "learning_rate": 0.0006,
+ "loss": 4.393521308898926,
+ "step": 1624
+ },
+ {
+ "epoch": 22.57292576419214,
+ "grad_norm": 0.013931898400187492,
+ "learning_rate": 0.0006,
+ "loss": 4.3739752769470215,
+ "step": 1625
+ },
+ {
+ "epoch": 22.586899563318777,
+ "grad_norm": 0.01497740764170885,
+ "learning_rate": 0.0006,
+ "loss": 4.3991618156433105,
+ "step": 1626
+ },
+ {
+ "epoch": 22.600873362445416,
+ "grad_norm": 0.015842905268073082,
+ "learning_rate": 0.0006,
+ "loss": 4.405340671539307,
+ "step": 1627
+ },
+ {
+ "epoch": 22.61484716157205,
+ "grad_norm": 0.01580633595585823,
+ "learning_rate": 0.0006,
+ "loss": 4.280102729797363,
+ "step": 1628
+ },
+ {
+ "epoch": 22.62882096069869,
+ "grad_norm": 0.016567382961511612,
+ "learning_rate": 0.0006,
+ "loss": 4.407657146453857,
+ "step": 1629
+ },
+ {
+ "epoch": 22.64279475982533,
+ "grad_norm": 0.01648622378706932,
+ "learning_rate": 0.0006,
+ "loss": 4.392590522766113,
+ "step": 1630
+ },
+ {
+ "epoch": 22.656768558951963,
+ "grad_norm": 0.017066307365894318,
+ "learning_rate": 0.0006,
+ "loss": 4.324862003326416,
+ "step": 1631
+ },
+ {
+ "epoch": 22.670742358078602,
+ "grad_norm": 0.01637062057852745,
+ "learning_rate": 0.0006,
+ "loss": 4.34598445892334,
+ "step": 1632
+ },
+ {
+ "epoch": 22.68471615720524,
+ "grad_norm": 0.018038960173726082,
+ "learning_rate": 0.0006,
+ "loss": 4.320711612701416,
+ "step": 1633
+ },
+ {
+ "epoch": 22.69868995633188,
+ "grad_norm": 0.018096428364515305,
+ "learning_rate": 0.0006,
+ "loss": 4.536527156829834,
+ "step": 1634
+ },
+ {
+ "epoch": 22.712663755458514,
+ "grad_norm": 0.017889291048049927,
+ "learning_rate": 0.0006,
+ "loss": 4.388131141662598,
+ "step": 1635
+ },
+ {
+ "epoch": 22.726637554585153,
+ "grad_norm": 0.018771521747112274,
+ "learning_rate": 0.0006,
+ "loss": 4.357668876647949,
+ "step": 1636
+ },
+ {
+ "epoch": 22.74061135371179,
+ "grad_norm": 0.02100873365998268,
+ "learning_rate": 0.0006,
+ "loss": 4.293972015380859,
+ "step": 1637
+ },
+ {
+ "epoch": 22.754585152838427,
+ "grad_norm": 0.017715785652399063,
+ "learning_rate": 0.0006,
+ "loss": 4.349849224090576,
+ "step": 1638
+ },
+ {
+ "epoch": 22.768558951965066,
+ "grad_norm": 0.015546333976089954,
+ "learning_rate": 0.0006,
+ "loss": 4.311728000640869,
+ "step": 1639
+ },
+ {
+ "epoch": 22.782532751091704,
+ "grad_norm": 0.01605990156531334,
+ "learning_rate": 0.0006,
+ "loss": 4.499123573303223,
+ "step": 1640
+ },
+ {
+ "epoch": 22.79650655021834,
+ "grad_norm": 0.016642142087221146,
+ "learning_rate": 0.0006,
+ "loss": 4.4319257736206055,
+ "step": 1641
+ },
+ {
+ "epoch": 22.810480349344978,
+ "grad_norm": 0.016120387241244316,
+ "learning_rate": 0.0006,
+ "loss": 4.4373650550842285,
+ "step": 1642
+ },
+ {
+ "epoch": 22.824454148471617,
+ "grad_norm": 0.015174890868365765,
+ "learning_rate": 0.0006,
+ "loss": 4.314082145690918,
+ "step": 1643
+ },
+ {
+ "epoch": 22.83842794759825,
+ "grad_norm": 0.01551714539527893,
+ "learning_rate": 0.0006,
+ "loss": 4.360363960266113,
+ "step": 1644
+ },
+ {
+ "epoch": 22.85240174672489,
+ "grad_norm": 0.014966742135584354,
+ "learning_rate": 0.0006,
+ "loss": 4.346253395080566,
+ "step": 1645
+ },
+ {
+ "epoch": 22.86637554585153,
+ "grad_norm": 0.014288028702139854,
+ "learning_rate": 0.0006,
+ "loss": 4.464366436004639,
+ "step": 1646
+ },
+ {
+ "epoch": 22.880349344978168,
+ "grad_norm": 0.015291682444512844,
+ "learning_rate": 0.0006,
+ "loss": 4.504101753234863,
+ "step": 1647
+ },
+ {
+ "epoch": 22.894323144104803,
+ "grad_norm": 0.014534426853060722,
+ "learning_rate": 0.0006,
+ "loss": 4.462172031402588,
+ "step": 1648
+ },
+ {
+ "epoch": 22.90829694323144,
+ "grad_norm": 0.014129819348454475,
+ "learning_rate": 0.0006,
+ "loss": 4.350126266479492,
+ "step": 1649
+ },
+ {
+ "epoch": 22.92227074235808,
+ "grad_norm": 0.014932668767869473,
+ "learning_rate": 0.0006,
+ "loss": 4.374392032623291,
+ "step": 1650
+ },
+ {
+ "epoch": 22.936244541484715,
+ "grad_norm": 0.014647853560745716,
+ "learning_rate": 0.0006,
+ "loss": 4.252264499664307,
+ "step": 1651
+ },
+ {
+ "epoch": 22.950218340611354,
+ "grad_norm": 0.013553360477089882,
+ "learning_rate": 0.0006,
+ "loss": 4.379729270935059,
+ "step": 1652
+ },
+ {
+ "epoch": 22.964192139737992,
+ "grad_norm": 0.014606194570660591,
+ "learning_rate": 0.0006,
+ "loss": 4.363372802734375,
+ "step": 1653
+ },
+ {
+ "epoch": 22.978165938864628,
+ "grad_norm": 0.015000631101429462,
+ "learning_rate": 0.0006,
+ "loss": 4.408048629760742,
+ "step": 1654
+ },
+ {
+ "epoch": 22.992139737991266,
+ "grad_norm": 0.015370816923677921,
+ "learning_rate": 0.0006,
+ "loss": 4.297451019287109,
+ "step": 1655
+ },
+ {
+ "epoch": 23.0,
+ "grad_norm": 0.01709144562482834,
+ "learning_rate": 0.0006,
+ "loss": 4.294776916503906,
+ "step": 1656
+ },
+ {
+ "epoch": 23.0,
+ "eval_loss": 4.678618907928467,
+ "eval_runtime": 56.1668,
+ "eval_samples_per_second": 43.478,
+ "eval_steps_per_second": 1.371,
+ "step": 1656
+ },
+ {
+ "epoch": 23.01397379912664,
+ "grad_norm": 0.01727290451526642,
+ "learning_rate": 0.0006,
+ "loss": 4.311701774597168,
+ "step": 1657
+ },
+ {
+ "epoch": 23.027947598253274,
+ "grad_norm": 0.020645759999752045,
+ "learning_rate": 0.0006,
+ "loss": 4.429897308349609,
+ "step": 1658
+ },
+ {
+ "epoch": 23.041921397379912,
+ "grad_norm": 0.02204226516187191,
+ "learning_rate": 0.0006,
+ "loss": 4.365151882171631,
+ "step": 1659
+ },
+ {
+ "epoch": 23.05589519650655,
+ "grad_norm": 0.023069269955158234,
+ "learning_rate": 0.0006,
+ "loss": 4.3710832595825195,
+ "step": 1660
+ },
+ {
+ "epoch": 23.069868995633186,
+ "grad_norm": 0.024151179939508438,
+ "learning_rate": 0.0006,
+ "loss": 4.403108596801758,
+ "step": 1661
+ },
+ {
+ "epoch": 23.083842794759825,
+ "grad_norm": 0.02511233650147915,
+ "learning_rate": 0.0006,
+ "loss": 4.414128303527832,
+ "step": 1662
+ },
+ {
+ "epoch": 23.097816593886463,
+ "grad_norm": 0.025802113115787506,
+ "learning_rate": 0.0006,
+ "loss": 4.388265609741211,
+ "step": 1663
+ },
+ {
+ "epoch": 23.111790393013102,
+ "grad_norm": 0.02377348765730858,
+ "learning_rate": 0.0006,
+ "loss": 4.357505798339844,
+ "step": 1664
+ },
+ {
+ "epoch": 23.125764192139737,
+ "grad_norm": 0.020876651629805565,
+ "learning_rate": 0.0006,
+ "loss": 4.23043966293335,
+ "step": 1665
+ },
+ {
+ "epoch": 23.139737991266376,
+ "grad_norm": 0.01959727331995964,
+ "learning_rate": 0.0006,
+ "loss": 4.280970096588135,
+ "step": 1666
+ },
+ {
+ "epoch": 23.153711790393015,
+ "grad_norm": 0.018678616732358932,
+ "learning_rate": 0.0006,
+ "loss": 4.346716403961182,
+ "step": 1667
+ },
+ {
+ "epoch": 23.16768558951965,
+ "grad_norm": 0.02108944021165371,
+ "learning_rate": 0.0006,
+ "loss": 4.277904510498047,
+ "step": 1668
+ },
+ {
+ "epoch": 23.18165938864629,
+ "grad_norm": 0.020221566781401634,
+ "learning_rate": 0.0006,
+ "loss": 4.367693901062012,
+ "step": 1669
+ },
+ {
+ "epoch": 23.195633187772927,
+ "grad_norm": 0.019599711522459984,
+ "learning_rate": 0.0006,
+ "loss": 4.257092475891113,
+ "step": 1670
+ },
+ {
+ "epoch": 23.209606986899562,
+ "grad_norm": 0.019833361729979515,
+ "learning_rate": 0.0006,
+ "loss": 4.540737628936768,
+ "step": 1671
+ },
+ {
+ "epoch": 23.2235807860262,
+ "grad_norm": 0.020352095365524292,
+ "learning_rate": 0.0006,
+ "loss": 4.360923767089844,
+ "step": 1672
+ },
+ {
+ "epoch": 23.23755458515284,
+ "grad_norm": 0.01939748041331768,
+ "learning_rate": 0.0006,
+ "loss": 4.393110275268555,
+ "step": 1673
+ },
+ {
+ "epoch": 23.251528384279474,
+ "grad_norm": 0.01884354278445244,
+ "learning_rate": 0.0006,
+ "loss": 4.383938789367676,
+ "step": 1674
+ },
+ {
+ "epoch": 23.265502183406113,
+ "grad_norm": 0.02008470520377159,
+ "learning_rate": 0.0006,
+ "loss": 4.346344947814941,
+ "step": 1675
+ },
+ {
+ "epoch": 23.27947598253275,
+ "grad_norm": 0.019065195694565773,
+ "learning_rate": 0.0006,
+ "loss": 4.36611795425415,
+ "step": 1676
+ },
+ {
+ "epoch": 23.29344978165939,
+ "grad_norm": 0.01938541606068611,
+ "learning_rate": 0.0006,
+ "loss": 4.4009599685668945,
+ "step": 1677
+ },
+ {
+ "epoch": 23.307423580786025,
+ "grad_norm": 0.019044723361730576,
+ "learning_rate": 0.0006,
+ "loss": 4.340169906616211,
+ "step": 1678
+ },
+ {
+ "epoch": 23.321397379912664,
+ "grad_norm": 0.018850378692150116,
+ "learning_rate": 0.0006,
+ "loss": 4.37674617767334,
+ "step": 1679
+ },
+ {
+ "epoch": 23.335371179039303,
+ "grad_norm": 0.019076095893979073,
+ "learning_rate": 0.0006,
+ "loss": 4.226874351501465,
+ "step": 1680
+ },
+ {
+ "epoch": 23.349344978165938,
+ "grad_norm": 0.017543038353323936,
+ "learning_rate": 0.0006,
+ "loss": 4.337751388549805,
+ "step": 1681
+ },
+ {
+ "epoch": 23.363318777292577,
+ "grad_norm": 0.017170218750834465,
+ "learning_rate": 0.0006,
+ "loss": 4.282512664794922,
+ "step": 1682
+ },
+ {
+ "epoch": 23.377292576419215,
+ "grad_norm": 0.01574796624481678,
+ "learning_rate": 0.0006,
+ "loss": 4.30638313293457,
+ "step": 1683
+ },
+ {
+ "epoch": 23.39126637554585,
+ "grad_norm": 0.015160846523940563,
+ "learning_rate": 0.0006,
+ "loss": 4.336716175079346,
+ "step": 1684
+ },
+ {
+ "epoch": 23.40524017467249,
+ "grad_norm": 0.01534526702016592,
+ "learning_rate": 0.0006,
+ "loss": 4.2174787521362305,
+ "step": 1685
+ },
+ {
+ "epoch": 23.419213973799128,
+ "grad_norm": 0.015289999544620514,
+ "learning_rate": 0.0006,
+ "loss": 4.272546291351318,
+ "step": 1686
+ },
+ {
+ "epoch": 23.433187772925763,
+ "grad_norm": 0.014894185587763786,
+ "learning_rate": 0.0006,
+ "loss": 4.296045303344727,
+ "step": 1687
+ },
+ {
+ "epoch": 23.4471615720524,
+ "grad_norm": 0.014869502745568752,
+ "learning_rate": 0.0006,
+ "loss": 4.425461769104004,
+ "step": 1688
+ },
+ {
+ "epoch": 23.46113537117904,
+ "grad_norm": 0.01474483497440815,
+ "learning_rate": 0.0006,
+ "loss": 4.361867427825928,
+ "step": 1689
+ },
+ {
+ "epoch": 23.475109170305675,
+ "grad_norm": 0.015299245715141296,
+ "learning_rate": 0.0006,
+ "loss": 4.364523410797119,
+ "step": 1690
+ },
+ {
+ "epoch": 23.489082969432314,
+ "grad_norm": 0.015201340429484844,
+ "learning_rate": 0.0006,
+ "loss": 4.285182952880859,
+ "step": 1691
+ },
+ {
+ "epoch": 23.503056768558952,
+ "grad_norm": 0.016057293862104416,
+ "learning_rate": 0.0006,
+ "loss": 4.284873962402344,
+ "step": 1692
+ },
+ {
+ "epoch": 23.51703056768559,
+ "grad_norm": 0.015871062874794006,
+ "learning_rate": 0.0006,
+ "loss": 4.326470851898193,
+ "step": 1693
+ },
+ {
+ "epoch": 23.531004366812226,
+ "grad_norm": 0.01586288772523403,
+ "learning_rate": 0.0006,
+ "loss": 4.324883460998535,
+ "step": 1694
+ },
+ {
+ "epoch": 23.544978165938865,
+ "grad_norm": 0.016092827543616295,
+ "learning_rate": 0.0006,
+ "loss": 4.3258867263793945,
+ "step": 1695
+ },
+ {
+ "epoch": 23.558951965065503,
+ "grad_norm": 0.01582016795873642,
+ "learning_rate": 0.0006,
+ "loss": 4.358892440795898,
+ "step": 1696
+ },
+ {
+ "epoch": 23.57292576419214,
+ "grad_norm": 0.01688452623784542,
+ "learning_rate": 0.0006,
+ "loss": 4.280842304229736,
+ "step": 1697
+ },
+ {
+ "epoch": 23.586899563318777,
+ "grad_norm": 0.015865886583924294,
+ "learning_rate": 0.0006,
+ "loss": 4.245567321777344,
+ "step": 1698
+ },
+ {
+ "epoch": 23.600873362445416,
+ "grad_norm": 0.015303662046790123,
+ "learning_rate": 0.0006,
+ "loss": 4.383112907409668,
+ "step": 1699
+ },
+ {
+ "epoch": 23.61484716157205,
+ "grad_norm": 0.015649516135454178,
+ "learning_rate": 0.0006,
+ "loss": 4.519082069396973,
+ "step": 1700
+ },
+ {
+ "epoch": 23.62882096069869,
+ "grad_norm": 0.0167120099067688,
+ "learning_rate": 0.0006,
+ "loss": 4.447847366333008,
+ "step": 1701
+ },
+ {
+ "epoch": 23.64279475982533,
+ "grad_norm": 0.016368716955184937,
+ "learning_rate": 0.0006,
+ "loss": 4.434186935424805,
+ "step": 1702
+ },
+ {
+ "epoch": 23.656768558951963,
+ "grad_norm": 0.017379503697156906,
+ "learning_rate": 0.0006,
+ "loss": 4.317594528198242,
+ "step": 1703
+ },
+ {
+ "epoch": 23.670742358078602,
+ "grad_norm": 0.018573811277747154,
+ "learning_rate": 0.0006,
+ "loss": 4.205793857574463,
+ "step": 1704
+ },
+ {
+ "epoch": 23.68471615720524,
+ "grad_norm": 0.01749090477824211,
+ "learning_rate": 0.0006,
+ "loss": 4.3985700607299805,
+ "step": 1705
+ },
+ {
+ "epoch": 23.69868995633188,
+ "grad_norm": 0.0165668074041605,
+ "learning_rate": 0.0006,
+ "loss": 4.329226970672607,
+ "step": 1706
+ },
+ {
+ "epoch": 23.712663755458514,
+ "grad_norm": 0.016913846135139465,
+ "learning_rate": 0.0006,
+ "loss": 4.388485908508301,
+ "step": 1707
+ },
+ {
+ "epoch": 23.726637554585153,
+ "grad_norm": 0.016813859343528748,
+ "learning_rate": 0.0006,
+ "loss": 4.3395915031433105,
+ "step": 1708
+ },
+ {
+ "epoch": 23.74061135371179,
+ "grad_norm": 0.01676975190639496,
+ "learning_rate": 0.0006,
+ "loss": 4.254745960235596,
+ "step": 1709
+ },
+ {
+ "epoch": 23.754585152838427,
+ "grad_norm": 0.016370350494980812,
+ "learning_rate": 0.0006,
+ "loss": 4.2612223625183105,
+ "step": 1710
+ },
+ {
+ "epoch": 23.768558951965066,
+ "grad_norm": 0.01696198247373104,
+ "learning_rate": 0.0006,
+ "loss": 4.327524662017822,
+ "step": 1711
+ },
+ {
+ "epoch": 23.782532751091704,
+ "grad_norm": 0.0189108457416296,
+ "learning_rate": 0.0006,
+ "loss": 4.325839519500732,
+ "step": 1712
+ },
+ {
+ "epoch": 23.79650655021834,
+ "grad_norm": 0.01862351782619953,
+ "learning_rate": 0.0006,
+ "loss": 4.333662509918213,
+ "step": 1713
+ },
+ {
+ "epoch": 23.810480349344978,
+ "grad_norm": 0.01655014418065548,
+ "learning_rate": 0.0006,
+ "loss": 4.39305305480957,
+ "step": 1714
+ },
+ {
+ "epoch": 23.824454148471617,
+ "grad_norm": 0.015899376943707466,
+ "learning_rate": 0.0006,
+ "loss": 4.390053749084473,
+ "step": 1715
+ },
+ {
+ "epoch": 23.83842794759825,
+ "grad_norm": 0.016519056633114815,
+ "learning_rate": 0.0006,
+ "loss": 4.255903244018555,
+ "step": 1716
+ },
+ {
+ "epoch": 23.85240174672489,
+ "grad_norm": 0.0167539119720459,
+ "learning_rate": 0.0006,
+ "loss": 4.255727767944336,
+ "step": 1717
+ },
+ {
+ "epoch": 23.86637554585153,
+ "grad_norm": 0.0171353816986084,
+ "learning_rate": 0.0006,
+ "loss": 4.376740455627441,
+ "step": 1718
+ },
+ {
+ "epoch": 23.880349344978168,
+ "grad_norm": 0.016719117760658264,
+ "learning_rate": 0.0006,
+ "loss": 4.3376688957214355,
+ "step": 1719
+ },
+ {
+ "epoch": 23.894323144104803,
+ "grad_norm": 0.01561494916677475,
+ "learning_rate": 0.0006,
+ "loss": 4.3345746994018555,
+ "step": 1720
+ },
+ {
+ "epoch": 23.90829694323144,
+ "grad_norm": 0.016303174197673798,
+ "learning_rate": 0.0006,
+ "loss": 4.332371234893799,
+ "step": 1721
+ },
+ {
+ "epoch": 23.92227074235808,
+ "grad_norm": 0.016722865402698517,
+ "learning_rate": 0.0006,
+ "loss": 4.267125129699707,
+ "step": 1722
+ },
+ {
+ "epoch": 23.936244541484715,
+ "grad_norm": 0.017072124406695366,
+ "learning_rate": 0.0006,
+ "loss": 4.322751998901367,
+ "step": 1723
+ },
+ {
+ "epoch": 23.950218340611354,
+ "grad_norm": 0.01699932850897312,
+ "learning_rate": 0.0006,
+ "loss": 4.378422737121582,
+ "step": 1724
+ },
+ {
+ "epoch": 23.964192139737992,
+ "grad_norm": 0.01742720976471901,
+ "learning_rate": 0.0006,
+ "loss": 4.345083236694336,
+ "step": 1725
+ },
+ {
+ "epoch": 23.978165938864628,
+ "grad_norm": 0.018797501921653748,
+ "learning_rate": 0.0006,
+ "loss": 4.290443420410156,
+ "step": 1726
+ },
+ {
+ "epoch": 23.992139737991266,
+ "grad_norm": 0.018090683966875076,
+ "learning_rate": 0.0006,
+ "loss": 4.377580642700195,
+ "step": 1727
+ },
+ {
+ "epoch": 24.0,
+ "grad_norm": 0.017305459827184677,
+ "learning_rate": 0.0006,
+ "loss": 4.364426612854004,
+ "step": 1728
+ },
+ {
+ "epoch": 24.0,
+ "eval_loss": 4.647762298583984,
+ "eval_runtime": 56.5402,
+ "eval_samples_per_second": 43.191,
+ "eval_steps_per_second": 1.362,
+ "step": 1728
+ },
+ {
+ "epoch": 24.01397379912664,
+ "grad_norm": 0.016493607312440872,
+ "learning_rate": 0.0006,
+ "loss": 4.252255916595459,
+ "step": 1729
+ },
+ {
+ "epoch": 24.027947598253274,
+ "grad_norm": 0.016529450193047523,
+ "learning_rate": 0.0006,
+ "loss": 4.352746963500977,
+ "step": 1730
+ },
+ {
+ "epoch": 24.041921397379912,
+ "grad_norm": 0.016940191388130188,
+ "learning_rate": 0.0006,
+ "loss": 4.301870822906494,
+ "step": 1731
+ },
+ {
+ "epoch": 24.05589519650655,
+ "grad_norm": 0.01752558909356594,
+ "learning_rate": 0.0006,
+ "loss": 4.2762451171875,
+ "step": 1732
+ },
+ {
+ "epoch": 24.069868995633186,
+ "grad_norm": 0.017112884670495987,
+ "learning_rate": 0.0006,
+ "loss": 4.2659149169921875,
+ "step": 1733
+ },
+ {
+ "epoch": 24.083842794759825,
+ "grad_norm": 0.01900586113333702,
+ "learning_rate": 0.0006,
+ "loss": 4.306058883666992,
+ "step": 1734
+ },
+ {
+ "epoch": 24.097816593886463,
+ "grad_norm": 0.019447680562734604,
+ "learning_rate": 0.0006,
+ "loss": 4.450128555297852,
+ "step": 1735
+ },
+ {
+ "epoch": 24.111790393013102,
+ "grad_norm": 0.01948186755180359,
+ "learning_rate": 0.0006,
+ "loss": 4.202869415283203,
+ "step": 1736
+ },
+ {
+ "epoch": 24.125764192139737,
+ "grad_norm": 0.01716383546590805,
+ "learning_rate": 0.0006,
+ "loss": 4.318876266479492,
+ "step": 1737
+ },
+ {
+ "epoch": 24.139737991266376,
+ "grad_norm": 0.017049988731741905,
+ "learning_rate": 0.0006,
+ "loss": 4.237238883972168,
+ "step": 1738
+ },
+ {
+ "epoch": 24.153711790393015,
+ "grad_norm": 0.015169923193752766,
+ "learning_rate": 0.0006,
+ "loss": 4.289426803588867,
+ "step": 1739
+ },
+ {
+ "epoch": 24.16768558951965,
+ "grad_norm": 0.014760667458176613,
+ "learning_rate": 0.0006,
+ "loss": 4.356924057006836,
+ "step": 1740
+ },
+ {
+ "epoch": 24.18165938864629,
+ "grad_norm": 0.015127750113606453,
+ "learning_rate": 0.0006,
+ "loss": 4.142916679382324,
+ "step": 1741
+ },
+ {
+ "epoch": 24.195633187772927,
+ "grad_norm": 0.01549555640667677,
+ "learning_rate": 0.0006,
+ "loss": 4.345563888549805,
+ "step": 1742
+ },
+ {
+ "epoch": 24.209606986899562,
+ "grad_norm": 0.016393091529607773,
+ "learning_rate": 0.0006,
+ "loss": 4.417495250701904,
+ "step": 1743
+ },
+ {
+ "epoch": 24.2235807860262,
+ "grad_norm": 0.01785266026854515,
+ "learning_rate": 0.0006,
+ "loss": 4.3553547859191895,
+ "step": 1744
+ },
+ {
+ "epoch": 24.23755458515284,
+ "grad_norm": 0.017574617639183998,
+ "learning_rate": 0.0006,
+ "loss": 4.241024494171143,
+ "step": 1745
+ },
+ {
+ "epoch": 24.251528384279474,
+ "grad_norm": 0.019060306251049042,
+ "learning_rate": 0.0006,
+ "loss": 4.3685150146484375,
+ "step": 1746
+ },
+ {
+ "epoch": 24.265502183406113,
+ "grad_norm": 0.0214633010327816,
+ "learning_rate": 0.0006,
+ "loss": 4.323866844177246,
+ "step": 1747
+ },
+ {
+ "epoch": 24.27947598253275,
+ "grad_norm": 0.021055003628134727,
+ "learning_rate": 0.0006,
+ "loss": 4.278757095336914,
+ "step": 1748
+ },
+ {
+ "epoch": 24.29344978165939,
+ "grad_norm": 0.018618572503328323,
+ "learning_rate": 0.0006,
+ "loss": 4.28849983215332,
+ "step": 1749
+ },
+ {
+ "epoch": 24.307423580786025,
+ "grad_norm": 0.018666435033082962,
+ "learning_rate": 0.0006,
+ "loss": 4.368590354919434,
+ "step": 1750
+ },
+ {
+ "epoch": 24.321397379912664,
+ "grad_norm": 0.020626146346330643,
+ "learning_rate": 0.0006,
+ "loss": 4.2951836585998535,
+ "step": 1751
+ },
+ {
+ "epoch": 24.335371179039303,
+ "grad_norm": 0.01929759792983532,
+ "learning_rate": 0.0006,
+ "loss": 4.3664703369140625,
+ "step": 1752
+ },
+ {
+ "epoch": 24.349344978165938,
+ "grad_norm": 0.020956002175807953,
+ "learning_rate": 0.0006,
+ "loss": 4.345308303833008,
+ "step": 1753
+ },
+ {
+ "epoch": 24.363318777292577,
+ "grad_norm": 0.019227957352995872,
+ "learning_rate": 0.0006,
+ "loss": 4.349849224090576,
+ "step": 1754
+ },
+ {
+ "epoch": 24.377292576419215,
+ "grad_norm": 0.015970097854733467,
+ "learning_rate": 0.0006,
+ "loss": 4.223012447357178,
+ "step": 1755
+ },
+ {
+ "epoch": 24.39126637554585,
+ "grad_norm": 0.015951646491885185,
+ "learning_rate": 0.0006,
+ "loss": 4.236420631408691,
+ "step": 1756
+ },
+ {
+ "epoch": 24.40524017467249,
+ "grad_norm": 0.016032515093684196,
+ "learning_rate": 0.0006,
+ "loss": 4.226899147033691,
+ "step": 1757
+ },
+ {
+ "epoch": 24.419213973799128,
+ "grad_norm": 0.01573132909834385,
+ "learning_rate": 0.0006,
+ "loss": 4.315916061401367,
+ "step": 1758
+ },
+ {
+ "epoch": 24.433187772925763,
+ "grad_norm": 0.015367005951702595,
+ "learning_rate": 0.0006,
+ "loss": 4.358939170837402,
+ "step": 1759
+ },
+ {
+ "epoch": 24.4471615720524,
+ "grad_norm": 0.01669907011091709,
+ "learning_rate": 0.0006,
+ "loss": 4.232147216796875,
+ "step": 1760
+ },
+ {
+ "epoch": 24.46113537117904,
+ "grad_norm": 0.018708152696490288,
+ "learning_rate": 0.0006,
+ "loss": 4.329852104187012,
+ "step": 1761
+ },
+ {
+ "epoch": 24.475109170305675,
+ "grad_norm": 0.019035987555980682,
+ "learning_rate": 0.0006,
+ "loss": 4.260340213775635,
+ "step": 1762
+ },
+ {
+ "epoch": 24.489082969432314,
+ "grad_norm": 0.01832144521176815,
+ "learning_rate": 0.0006,
+ "loss": 4.359543800354004,
+ "step": 1763
+ },
+ {
+ "epoch": 24.503056768558952,
+ "grad_norm": 0.015313294716179371,
+ "learning_rate": 0.0006,
+ "loss": 4.279088020324707,
+ "step": 1764
+ },
+ {
+ "epoch": 24.51703056768559,
+ "grad_norm": 0.015803923830389977,
+ "learning_rate": 0.0006,
+ "loss": 4.230682849884033,
+ "step": 1765
+ },
+ {
+ "epoch": 24.531004366812226,
+ "grad_norm": 0.016748374328017235,
+ "learning_rate": 0.0006,
+ "loss": 4.349661827087402,
+ "step": 1766
+ },
+ {
+ "epoch": 24.544978165938865,
+ "grad_norm": 0.015732428058981895,
+ "learning_rate": 0.0006,
+ "loss": 4.372441291809082,
+ "step": 1767
+ },
+ {
+ "epoch": 24.558951965065503,
+ "grad_norm": 0.016564829275012016,
+ "learning_rate": 0.0006,
+ "loss": 4.323090553283691,
+ "step": 1768
+ },
+ {
+ "epoch": 24.57292576419214,
+ "grad_norm": 0.01603010483086109,
+ "learning_rate": 0.0006,
+ "loss": 4.347898006439209,
+ "step": 1769
+ },
+ {
+ "epoch": 24.586899563318777,
+ "grad_norm": 0.01674809865653515,
+ "learning_rate": 0.0006,
+ "loss": 4.269165515899658,
+ "step": 1770
+ },
+ {
+ "epoch": 24.600873362445416,
+ "grad_norm": 0.017333930358290672,
+ "learning_rate": 0.0006,
+ "loss": 4.353298187255859,
+ "step": 1771
+ },
+ {
+ "epoch": 24.61484716157205,
+ "grad_norm": 0.017519580200314522,
+ "learning_rate": 0.0006,
+ "loss": 4.265199661254883,
+ "step": 1772
+ },
+ {
+ "epoch": 24.62882096069869,
+ "grad_norm": 0.017675306648015976,
+ "learning_rate": 0.0006,
+ "loss": 4.322596549987793,
+ "step": 1773
+ },
+ {
+ "epoch": 24.64279475982533,
+ "grad_norm": 0.017166534438729286,
+ "learning_rate": 0.0006,
+ "loss": 4.413900375366211,
+ "step": 1774
+ },
+ {
+ "epoch": 24.656768558951963,
+ "grad_norm": 0.016362346708774567,
+ "learning_rate": 0.0006,
+ "loss": 4.307236671447754,
+ "step": 1775
+ },
+ {
+ "epoch": 24.670742358078602,
+ "grad_norm": 0.016086872667074203,
+ "learning_rate": 0.0006,
+ "loss": 4.235316276550293,
+ "step": 1776
+ },
+ {
+ "epoch": 24.68471615720524,
+ "grad_norm": 0.01689537614583969,
+ "learning_rate": 0.0006,
+ "loss": 4.390591621398926,
+ "step": 1777
+ },
+ {
+ "epoch": 24.69868995633188,
+ "grad_norm": 0.016836725175380707,
+ "learning_rate": 0.0006,
+ "loss": 4.2990031242370605,
+ "step": 1778
+ },
+ {
+ "epoch": 24.712663755458514,
+ "grad_norm": 0.015627335757017136,
+ "learning_rate": 0.0006,
+ "loss": 4.166745185852051,
+ "step": 1779
+ },
+ {
+ "epoch": 24.726637554585153,
+ "grad_norm": 0.015007016249001026,
+ "learning_rate": 0.0006,
+ "loss": 4.295273303985596,
+ "step": 1780
+ },
+ {
+ "epoch": 24.74061135371179,
+ "grad_norm": 0.016084209084510803,
+ "learning_rate": 0.0006,
+ "loss": 4.342424392700195,
+ "step": 1781
+ },
+ {
+ "epoch": 24.754585152838427,
+ "grad_norm": 0.01595073565840721,
+ "learning_rate": 0.0006,
+ "loss": 4.3060102462768555,
+ "step": 1782
+ },
+ {
+ "epoch": 24.768558951965066,
+ "grad_norm": 0.016113461926579475,
+ "learning_rate": 0.0006,
+ "loss": 4.267908096313477,
+ "step": 1783
+ },
+ {
+ "epoch": 24.782532751091704,
+ "grad_norm": 0.01564556173980236,
+ "learning_rate": 0.0006,
+ "loss": 4.324642658233643,
+ "step": 1784
+ },
+ {
+ "epoch": 24.79650655021834,
+ "grad_norm": 0.01575268618762493,
+ "learning_rate": 0.0006,
+ "loss": 4.328181743621826,
+ "step": 1785
+ },
+ {
+ "epoch": 24.810480349344978,
+ "grad_norm": 0.016830824315547943,
+ "learning_rate": 0.0006,
+ "loss": 4.247906684875488,
+ "step": 1786
+ },
+ {
+ "epoch": 24.824454148471617,
+ "grad_norm": 0.017989547923207283,
+ "learning_rate": 0.0006,
+ "loss": 4.228281021118164,
+ "step": 1787
+ },
+ {
+ "epoch": 24.83842794759825,
+ "grad_norm": 0.016783086583018303,
+ "learning_rate": 0.0006,
+ "loss": 4.2880964279174805,
+ "step": 1788
+ },
+ {
+ "epoch": 24.85240174672489,
+ "grad_norm": 0.016467561945319176,
+ "learning_rate": 0.0006,
+ "loss": 4.36118745803833,
+ "step": 1789
+ },
+ {
+ "epoch": 24.86637554585153,
+ "grad_norm": 0.01780523732304573,
+ "learning_rate": 0.0006,
+ "loss": 4.172321319580078,
+ "step": 1790
+ },
+ {
+ "epoch": 24.880349344978168,
+ "grad_norm": 0.017102031037211418,
+ "learning_rate": 0.0006,
+ "loss": 4.352826118469238,
+ "step": 1791
+ },
+ {
+ "epoch": 24.894323144104803,
+ "grad_norm": 0.016409218311309814,
+ "learning_rate": 0.0006,
+ "loss": 4.212162017822266,
+ "step": 1792
+ },
+ {
+ "epoch": 24.90829694323144,
+ "grad_norm": 0.018660584464669228,
+ "learning_rate": 0.0006,
+ "loss": 4.346684455871582,
+ "step": 1793
+ },
+ {
+ "epoch": 24.92227074235808,
+ "grad_norm": 0.017389440909028053,
+ "learning_rate": 0.0006,
+ "loss": 4.197913646697998,
+ "step": 1794
+ },
+ {
+ "epoch": 24.936244541484715,
+ "grad_norm": 0.01596485823392868,
+ "learning_rate": 0.0006,
+ "loss": 4.370617866516113,
+ "step": 1795
+ },
+ {
+ "epoch": 24.950218340611354,
+ "grad_norm": 0.01709776371717453,
+ "learning_rate": 0.0006,
+ "loss": 4.313847064971924,
+ "step": 1796
+ },
+ {
+ "epoch": 24.964192139737992,
+ "grad_norm": 0.016571134328842163,
+ "learning_rate": 0.0006,
+ "loss": 4.1895856857299805,
+ "step": 1797
+ },
+ {
+ "epoch": 24.978165938864628,
+ "grad_norm": 0.016851048916578293,
+ "learning_rate": 0.0006,
+ "loss": 4.350772857666016,
+ "step": 1798
+ },
+ {
+ "epoch": 24.992139737991266,
+ "grad_norm": 0.017007440328598022,
+ "learning_rate": 0.0006,
+ "loss": 4.399008750915527,
+ "step": 1799
+ },
+ {
+ "epoch": 25.0,
+ "grad_norm": 0.01791212148964405,
+ "learning_rate": 0.0006,
+ "loss": 4.306121826171875,
+ "step": 1800
+ },
+ {
+ "epoch": 25.0,
+ "eval_loss": 4.636417865753174,
+ "eval_runtime": 56.8296,
+ "eval_samples_per_second": 42.971,
+ "eval_steps_per_second": 1.355,
+ "step": 1800
+ },
+ {
+ "epoch": 25.01397379912664,
+ "grad_norm": 0.01729617640376091,
+ "learning_rate": 0.0006,
+ "loss": 4.257018089294434,
+ "step": 1801
+ },
+ {
+ "epoch": 25.027947598253274,
+ "grad_norm": 0.017466643825173378,
+ "learning_rate": 0.0006,
+ "loss": 4.307605266571045,
+ "step": 1802
+ },
+ {
+ "epoch": 25.041921397379912,
+ "grad_norm": 0.018614279106259346,
+ "learning_rate": 0.0006,
+ "loss": 4.360141277313232,
+ "step": 1803
+ },
+ {
+ "epoch": 25.05589519650655,
+ "grad_norm": 0.02010771445930004,
+ "learning_rate": 0.0006,
+ "loss": 4.196428298950195,
+ "step": 1804
+ },
+ {
+ "epoch": 25.069868995633186,
+ "grad_norm": 0.02090228535234928,
+ "learning_rate": 0.0006,
+ "loss": 4.288947105407715,
+ "step": 1805
+ },
+ {
+ "epoch": 25.083842794759825,
+ "grad_norm": 0.02050255797803402,
+ "learning_rate": 0.0006,
+ "loss": 4.259582042694092,
+ "step": 1806
+ },
+ {
+ "epoch": 25.097816593886463,
+ "grad_norm": 0.021844089031219482,
+ "learning_rate": 0.0006,
+ "loss": 4.405138969421387,
+ "step": 1807
+ },
+ {
+ "epoch": 25.111790393013102,
+ "grad_norm": 0.022287018597126007,
+ "learning_rate": 0.0006,
+ "loss": 4.333793640136719,
+ "step": 1808
+ },
+ {
+ "epoch": 25.125764192139737,
+ "grad_norm": 0.02288329415023327,
+ "learning_rate": 0.0006,
+ "loss": 4.3910980224609375,
+ "step": 1809
+ },
+ {
+ "epoch": 25.139737991266376,
+ "grad_norm": 0.02373199723660946,
+ "learning_rate": 0.0006,
+ "loss": 4.32921028137207,
+ "step": 1810
+ },
+ {
+ "epoch": 25.153711790393015,
+ "grad_norm": 0.024960234761238098,
+ "learning_rate": 0.0006,
+ "loss": 4.295609474182129,
+ "step": 1811
+ },
+ {
+ "epoch": 25.16768558951965,
+ "grad_norm": 0.026743892580270767,
+ "learning_rate": 0.0006,
+ "loss": 4.315308094024658,
+ "step": 1812
+ },
+ {
+ "epoch": 25.18165938864629,
+ "grad_norm": 0.022138185799121857,
+ "learning_rate": 0.0006,
+ "loss": 4.360208511352539,
+ "step": 1813
+ },
+ {
+ "epoch": 25.195633187772927,
+ "grad_norm": 0.022120574489235878,
+ "learning_rate": 0.0006,
+ "loss": 4.1973772048950195,
+ "step": 1814
+ },
+ {
+ "epoch": 25.209606986899562,
+ "grad_norm": 0.02230502851307392,
+ "learning_rate": 0.0006,
+ "loss": 4.285558700561523,
+ "step": 1815
+ },
+ {
+ "epoch": 25.2235807860262,
+ "grad_norm": 0.022301986813545227,
+ "learning_rate": 0.0006,
+ "loss": 4.238343238830566,
+ "step": 1816
+ },
+ {
+ "epoch": 25.23755458515284,
+ "grad_norm": 0.020050447434186935,
+ "learning_rate": 0.0006,
+ "loss": 4.298235893249512,
+ "step": 1817
+ },
+ {
+ "epoch": 25.251528384279474,
+ "grad_norm": 0.01987724006175995,
+ "learning_rate": 0.0006,
+ "loss": 4.301384449005127,
+ "step": 1818
+ },
+ {
+ "epoch": 25.265502183406113,
+ "grad_norm": 0.020440855994820595,
+ "learning_rate": 0.0006,
+ "loss": 4.269218444824219,
+ "step": 1819
+ },
+ {
+ "epoch": 25.27947598253275,
+ "grad_norm": 0.01891893707215786,
+ "learning_rate": 0.0006,
+ "loss": 4.2401957511901855,
+ "step": 1820
+ },
+ {
+ "epoch": 25.29344978165939,
+ "grad_norm": 0.01903688907623291,
+ "learning_rate": 0.0006,
+ "loss": 4.139670372009277,
+ "step": 1821
+ },
+ {
+ "epoch": 25.307423580786025,
+ "grad_norm": 0.018856938928365707,
+ "learning_rate": 0.0006,
+ "loss": 4.233214378356934,
+ "step": 1822
+ },
+ {
+ "epoch": 25.321397379912664,
+ "grad_norm": 0.017791323363780975,
+ "learning_rate": 0.0006,
+ "loss": 4.2342023849487305,
+ "step": 1823
+ },
+ {
+ "epoch": 25.335371179039303,
+ "grad_norm": 0.017304565757513046,
+ "learning_rate": 0.0006,
+ "loss": 4.355198860168457,
+ "step": 1824
+ },
+ {
+ "epoch": 25.349344978165938,
+ "grad_norm": 0.01576206088066101,
+ "learning_rate": 0.0006,
+ "loss": 4.114147186279297,
+ "step": 1825
+ },
+ {
+ "epoch": 25.363318777292577,
+ "grad_norm": 0.015105154365301132,
+ "learning_rate": 0.0006,
+ "loss": 4.313795566558838,
+ "step": 1826
+ },
+ {
+ "epoch": 25.377292576419215,
+ "grad_norm": 0.01563587784767151,
+ "learning_rate": 0.0006,
+ "loss": 4.308282852172852,
+ "step": 1827
+ },
+ {
+ "epoch": 25.39126637554585,
+ "grad_norm": 0.015980906784534454,
+ "learning_rate": 0.0006,
+ "loss": 4.345149040222168,
+ "step": 1828
+ },
+ {
+ "epoch": 25.40524017467249,
+ "grad_norm": 0.01669284701347351,
+ "learning_rate": 0.0006,
+ "loss": 4.237611293792725,
+ "step": 1829
+ },
+ {
+ "epoch": 25.419213973799128,
+ "grad_norm": 0.016202064231038094,
+ "learning_rate": 0.0006,
+ "loss": 4.3072309494018555,
+ "step": 1830
+ },
+ {
+ "epoch": 25.433187772925763,
+ "grad_norm": 0.01478662807494402,
+ "learning_rate": 0.0006,
+ "loss": 4.276215076446533,
+ "step": 1831
+ },
+ {
+ "epoch": 25.4471615720524,
+ "grad_norm": 0.015168731100857258,
+ "learning_rate": 0.0006,
+ "loss": 4.37509822845459,
+ "step": 1832
+ },
+ {
+ "epoch": 25.46113537117904,
+ "grad_norm": 0.01648980937898159,
+ "learning_rate": 0.0006,
+ "loss": 4.349167346954346,
+ "step": 1833
+ },
+ {
+ "epoch": 25.475109170305675,
+ "grad_norm": 0.01708398200571537,
+ "learning_rate": 0.0006,
+ "loss": 4.346580505371094,
+ "step": 1834
+ },
+ {
+ "epoch": 25.489082969432314,
+ "grad_norm": 0.017370784655213356,
+ "learning_rate": 0.0006,
+ "loss": 4.299466133117676,
+ "step": 1835
+ },
+ {
+ "epoch": 25.503056768558952,
+ "grad_norm": 0.016248228028416634,
+ "learning_rate": 0.0006,
+ "loss": 4.184610366821289,
+ "step": 1836
+ },
+ {
+ "epoch": 25.51703056768559,
+ "grad_norm": 0.017106913030147552,
+ "learning_rate": 0.0006,
+ "loss": 4.366856098175049,
+ "step": 1837
+ },
+ {
+ "epoch": 25.531004366812226,
+ "grad_norm": 0.016071965917944908,
+ "learning_rate": 0.0006,
+ "loss": 4.297691345214844,
+ "step": 1838
+ },
+ {
+ "epoch": 25.544978165938865,
+ "grad_norm": 0.016451986506581306,
+ "learning_rate": 0.0006,
+ "loss": 4.229983329772949,
+ "step": 1839
+ },
+ {
+ "epoch": 25.558951965065503,
+ "grad_norm": 0.016195081174373627,
+ "learning_rate": 0.0006,
+ "loss": 4.259893417358398,
+ "step": 1840
+ },
+ {
+ "epoch": 25.57292576419214,
+ "grad_norm": 0.0165175162255764,
+ "learning_rate": 0.0006,
+ "loss": 4.3348236083984375,
+ "step": 1841
+ },
+ {
+ "epoch": 25.586899563318777,
+ "grad_norm": 0.01651517115533352,
+ "learning_rate": 0.0006,
+ "loss": 4.319541931152344,
+ "step": 1842
+ },
+ {
+ "epoch": 25.600873362445416,
+ "grad_norm": 0.016143690794706345,
+ "learning_rate": 0.0006,
+ "loss": 4.345652103424072,
+ "step": 1843
+ },
+ {
+ "epoch": 25.61484716157205,
+ "grad_norm": 0.015140696428716183,
+ "learning_rate": 0.0006,
+ "loss": 4.212047576904297,
+ "step": 1844
+ },
+ {
+ "epoch": 25.62882096069869,
+ "grad_norm": 0.014637443237006664,
+ "learning_rate": 0.0006,
+ "loss": 4.187453269958496,
+ "step": 1845
+ },
+ {
+ "epoch": 25.64279475982533,
+ "grad_norm": 0.014585614204406738,
+ "learning_rate": 0.0006,
+ "loss": 4.253936767578125,
+ "step": 1846
+ },
+ {
+ "epoch": 25.656768558951963,
+ "grad_norm": 0.015158289112150669,
+ "learning_rate": 0.0006,
+ "loss": 4.34612512588501,
+ "step": 1847
+ },
+ {
+ "epoch": 25.670742358078602,
+ "grad_norm": 0.015202849172055721,
+ "learning_rate": 0.0006,
+ "loss": 4.270634651184082,
+ "step": 1848
+ },
+ {
+ "epoch": 25.68471615720524,
+ "grad_norm": 0.015363575890660286,
+ "learning_rate": 0.0006,
+ "loss": 4.271080493927002,
+ "step": 1849
+ },
+ {
+ "epoch": 25.69868995633188,
+ "grad_norm": 0.014360226690769196,
+ "learning_rate": 0.0006,
+ "loss": 4.375033855438232,
+ "step": 1850
+ },
+ {
+ "epoch": 25.712663755458514,
+ "grad_norm": 0.014263181947171688,
+ "learning_rate": 0.0006,
+ "loss": 4.424138069152832,
+ "step": 1851
+ },
+ {
+ "epoch": 25.726637554585153,
+ "grad_norm": 0.014398688450455666,
+ "learning_rate": 0.0006,
+ "loss": 4.410154819488525,
+ "step": 1852
+ },
+ {
+ "epoch": 25.74061135371179,
+ "grad_norm": 0.016131488606333733,
+ "learning_rate": 0.0006,
+ "loss": 4.324014663696289,
+ "step": 1853
+ },
+ {
+ "epoch": 25.754585152838427,
+ "grad_norm": 0.017568735405802727,
+ "learning_rate": 0.0006,
+ "loss": 4.367861747741699,
+ "step": 1854
+ },
+ {
+ "epoch": 25.768558951965066,
+ "grad_norm": 0.017907511442899704,
+ "learning_rate": 0.0006,
+ "loss": 4.352965354919434,
+ "step": 1855
+ },
+ {
+ "epoch": 25.782532751091704,
+ "grad_norm": 0.016918016597628593,
+ "learning_rate": 0.0006,
+ "loss": 4.252397537231445,
+ "step": 1856
+ },
+ {
+ "epoch": 25.79650655021834,
+ "grad_norm": 0.017274610698223114,
+ "learning_rate": 0.0006,
+ "loss": 4.342309951782227,
+ "step": 1857
+ },
+ {
+ "epoch": 25.810480349344978,
+ "grad_norm": 0.017879825085401535,
+ "learning_rate": 0.0006,
+ "loss": 4.330987453460693,
+ "step": 1858
+ },
+ {
+ "epoch": 25.824454148471617,
+ "grad_norm": 0.01736099272966385,
+ "learning_rate": 0.0006,
+ "loss": 4.307255744934082,
+ "step": 1859
+ },
+ {
+ "epoch": 25.83842794759825,
+ "grad_norm": 0.016575824469327927,
+ "learning_rate": 0.0006,
+ "loss": 4.262016773223877,
+ "step": 1860
+ },
+ {
+ "epoch": 25.85240174672489,
+ "grad_norm": 0.015181140042841434,
+ "learning_rate": 0.0006,
+ "loss": 4.291863441467285,
+ "step": 1861
+ },
+ {
+ "epoch": 25.86637554585153,
+ "grad_norm": 0.015965687111020088,
+ "learning_rate": 0.0006,
+ "loss": 4.424836158752441,
+ "step": 1862
+ },
+ {
+ "epoch": 25.880349344978168,
+ "grad_norm": 0.01687219925224781,
+ "learning_rate": 0.0006,
+ "loss": 4.343531131744385,
+ "step": 1863
+ },
+ {
+ "epoch": 25.894323144104803,
+ "grad_norm": 0.015115504153072834,
+ "learning_rate": 0.0006,
+ "loss": 4.357808589935303,
+ "step": 1864
+ },
+ {
+ "epoch": 25.90829694323144,
+ "grad_norm": 0.014150998555123806,
+ "learning_rate": 0.0006,
+ "loss": 4.324550151824951,
+ "step": 1865
+ },
+ {
+ "epoch": 25.92227074235808,
+ "grad_norm": 0.014989510178565979,
+ "learning_rate": 0.0006,
+ "loss": 4.380428314208984,
+ "step": 1866
+ },
+ {
+ "epoch": 25.936244541484715,
+ "grad_norm": 0.01527960691601038,
+ "learning_rate": 0.0006,
+ "loss": 4.399374961853027,
+ "step": 1867
+ },
+ {
+ "epoch": 25.950218340611354,
+ "grad_norm": 0.015434633940458298,
+ "learning_rate": 0.0006,
+ "loss": 4.260984420776367,
+ "step": 1868
+ },
+ {
+ "epoch": 25.964192139737992,
+ "grad_norm": 0.015370228327810764,
+ "learning_rate": 0.0006,
+ "loss": 4.3340253829956055,
+ "step": 1869
+ },
+ {
+ "epoch": 25.978165938864628,
+ "grad_norm": 0.015348542481660843,
+ "learning_rate": 0.0006,
+ "loss": 4.369132995605469,
+ "step": 1870
+ },
+ {
+ "epoch": 25.992139737991266,
+ "grad_norm": 0.015354936942458153,
+ "learning_rate": 0.0006,
+ "loss": 4.381937026977539,
+ "step": 1871
+ },
+ {
+ "epoch": 26.0,
+ "grad_norm": 0.01796272210776806,
+ "learning_rate": 0.0006,
+ "loss": 4.286768436431885,
+ "step": 1872
+ },
+ {
+ "epoch": 26.0,
+ "eval_loss": 4.576984405517578,
+ "eval_runtime": 57.2091,
+ "eval_samples_per_second": 42.685,
+ "eval_steps_per_second": 1.346,
+ "step": 1872
+ },
+ {
+ "epoch": 26.01397379912664,
+ "grad_norm": 0.01748845726251602,
+ "learning_rate": 0.0006,
+ "loss": 4.337504863739014,
+ "step": 1873
+ },
+ {
+ "epoch": 26.027947598253274,
+ "grad_norm": 0.019244296476244926,
+ "learning_rate": 0.0006,
+ "loss": 4.264280796051025,
+ "step": 1874
+ },
+ {
+ "epoch": 26.041921397379912,
+ "grad_norm": 0.0215692650526762,
+ "learning_rate": 0.0006,
+ "loss": 4.350830078125,
+ "step": 1875
+ },
+ {
+ "epoch": 26.05589519650655,
+ "grad_norm": 0.022189252078533173,
+ "learning_rate": 0.0006,
+ "loss": 4.299098968505859,
+ "step": 1876
+ },
+ {
+ "epoch": 26.069868995633186,
+ "grad_norm": 0.022384043782949448,
+ "learning_rate": 0.0006,
+ "loss": 4.1185173988342285,
+ "step": 1877
+ },
+ {
+ "epoch": 26.083842794759825,
+ "grad_norm": 0.02076014317572117,
+ "learning_rate": 0.0006,
+ "loss": 4.297612190246582,
+ "step": 1878
+ },
+ {
+ "epoch": 26.097816593886463,
+ "grad_norm": 0.020843051373958588,
+ "learning_rate": 0.0006,
+ "loss": 4.294098854064941,
+ "step": 1879
+ },
+ {
+ "epoch": 26.111790393013102,
+ "grad_norm": 0.020654456689953804,
+ "learning_rate": 0.0006,
+ "loss": 4.267856597900391,
+ "step": 1880
+ },
+ {
+ "epoch": 26.125764192139737,
+ "grad_norm": 0.01860472559928894,
+ "learning_rate": 0.0006,
+ "loss": 4.284232139587402,
+ "step": 1881
+ },
+ {
+ "epoch": 26.139737991266376,
+ "grad_norm": 0.018719421699643135,
+ "learning_rate": 0.0006,
+ "loss": 4.277214527130127,
+ "step": 1882
+ },
+ {
+ "epoch": 26.153711790393015,
+ "grad_norm": 0.020903829485177994,
+ "learning_rate": 0.0006,
+ "loss": 4.254947662353516,
+ "step": 1883
+ },
+ {
+ "epoch": 26.16768558951965,
+ "grad_norm": 0.02020149491727352,
+ "learning_rate": 0.0006,
+ "loss": 4.2526936531066895,
+ "step": 1884
+ },
+ {
+ "epoch": 26.18165938864629,
+ "grad_norm": 0.0180392786860466,
+ "learning_rate": 0.0006,
+ "loss": 4.309922695159912,
+ "step": 1885
+ },
+ {
+ "epoch": 26.195633187772927,
+ "grad_norm": 0.018163125962018967,
+ "learning_rate": 0.0006,
+ "loss": 4.13860559463501,
+ "step": 1886
+ },
+ {
+ "epoch": 26.209606986899562,
+ "grad_norm": 0.01853291131556034,
+ "learning_rate": 0.0006,
+ "loss": 4.190926551818848,
+ "step": 1887
+ },
+ {
+ "epoch": 26.2235807860262,
+ "grad_norm": 0.01757677271962166,
+ "learning_rate": 0.0006,
+ "loss": 4.163999557495117,
+ "step": 1888
+ },
+ {
+ "epoch": 26.23755458515284,
+ "grad_norm": 0.01767992228269577,
+ "learning_rate": 0.0006,
+ "loss": 4.184430122375488,
+ "step": 1889
+ },
+ {
+ "epoch": 26.251528384279474,
+ "grad_norm": 0.018139973282814026,
+ "learning_rate": 0.0006,
+ "loss": 4.300766944885254,
+ "step": 1890
+ },
+ {
+ "epoch": 26.265502183406113,
+ "grad_norm": 0.020184461027383804,
+ "learning_rate": 0.0006,
+ "loss": 4.232537269592285,
+ "step": 1891
+ },
+ {
+ "epoch": 26.27947598253275,
+ "grad_norm": 0.02028856985270977,
+ "learning_rate": 0.0006,
+ "loss": 4.324883460998535,
+ "step": 1892
+ },
+ {
+ "epoch": 26.29344978165939,
+ "grad_norm": 0.017120616510510445,
+ "learning_rate": 0.0006,
+ "loss": 4.23493766784668,
+ "step": 1893
+ },
+ {
+ "epoch": 26.307423580786025,
+ "grad_norm": 0.016202978789806366,
+ "learning_rate": 0.0006,
+ "loss": 4.197602272033691,
+ "step": 1894
+ },
+ {
+ "epoch": 26.321397379912664,
+ "grad_norm": 0.017231421545147896,
+ "learning_rate": 0.0006,
+ "loss": 4.302001953125,
+ "step": 1895
+ },
+ {
+ "epoch": 26.335371179039303,
+ "grad_norm": 0.01725655607879162,
+ "learning_rate": 0.0006,
+ "loss": 4.356800079345703,
+ "step": 1896
+ },
+ {
+ "epoch": 26.349344978165938,
+ "grad_norm": 0.016657443717122078,
+ "learning_rate": 0.0006,
+ "loss": 4.266307830810547,
+ "step": 1897
+ },
+ {
+ "epoch": 26.363318777292577,
+ "grad_norm": 0.01616556942462921,
+ "learning_rate": 0.0006,
+ "loss": 4.307868003845215,
+ "step": 1898
+ },
+ {
+ "epoch": 26.377292576419215,
+ "grad_norm": 0.016621306538581848,
+ "learning_rate": 0.0006,
+ "loss": 4.172905921936035,
+ "step": 1899
+ },
+ {
+ "epoch": 26.39126637554585,
+ "grad_norm": 0.016243206337094307,
+ "learning_rate": 0.0006,
+ "loss": 4.162896156311035,
+ "step": 1900
+ },
+ {
+ "epoch": 26.40524017467249,
+ "grad_norm": 0.016663808375597,
+ "learning_rate": 0.0006,
+ "loss": 4.258232116699219,
+ "step": 1901
+ },
+ {
+ "epoch": 26.419213973799128,
+ "grad_norm": 0.01677478663623333,
+ "learning_rate": 0.0006,
+ "loss": 4.336090087890625,
+ "step": 1902
+ },
+ {
+ "epoch": 26.433187772925763,
+ "grad_norm": 0.01641303487122059,
+ "learning_rate": 0.0006,
+ "loss": 4.191476821899414,
+ "step": 1903
+ },
+ {
+ "epoch": 26.4471615720524,
+ "grad_norm": 0.016700323671102524,
+ "learning_rate": 0.0006,
+ "loss": 4.299427509307861,
+ "step": 1904
+ },
+ {
+ "epoch": 26.46113537117904,
+ "grad_norm": 0.017595035955309868,
+ "learning_rate": 0.0006,
+ "loss": 4.312828063964844,
+ "step": 1905
+ },
+ {
+ "epoch": 26.475109170305675,
+ "grad_norm": 0.020300284028053284,
+ "learning_rate": 0.0006,
+ "loss": 4.28331184387207,
+ "step": 1906
+ },
+ {
+ "epoch": 26.489082969432314,
+ "grad_norm": 0.02198321744799614,
+ "learning_rate": 0.0006,
+ "loss": 4.3280839920043945,
+ "step": 1907
+ },
+ {
+ "epoch": 26.503056768558952,
+ "grad_norm": 0.02003200724720955,
+ "learning_rate": 0.0006,
+ "loss": 4.325623989105225,
+ "step": 1908
+ },
+ {
+ "epoch": 26.51703056768559,
+ "grad_norm": 0.01870287023484707,
+ "learning_rate": 0.0006,
+ "loss": 4.251203536987305,
+ "step": 1909
+ },
+ {
+ "epoch": 26.531004366812226,
+ "grad_norm": 0.01835448481142521,
+ "learning_rate": 0.0006,
+ "loss": 4.335144996643066,
+ "step": 1910
+ },
+ {
+ "epoch": 26.544978165938865,
+ "grad_norm": 0.02046327292919159,
+ "learning_rate": 0.0006,
+ "loss": 4.238023281097412,
+ "step": 1911
+ },
+ {
+ "epoch": 26.558951965065503,
+ "grad_norm": 0.019771311432123184,
+ "learning_rate": 0.0006,
+ "loss": 4.353594779968262,
+ "step": 1912
+ },
+ {
+ "epoch": 26.57292576419214,
+ "grad_norm": 0.018090544268488884,
+ "learning_rate": 0.0006,
+ "loss": 4.304343223571777,
+ "step": 1913
+ },
+ {
+ "epoch": 26.586899563318777,
+ "grad_norm": 0.017489777877926826,
+ "learning_rate": 0.0006,
+ "loss": 4.204375267028809,
+ "step": 1914
+ },
+ {
+ "epoch": 26.600873362445416,
+ "grad_norm": 0.016785891726613045,
+ "learning_rate": 0.0006,
+ "loss": 4.286669731140137,
+ "step": 1915
+ },
+ {
+ "epoch": 26.61484716157205,
+ "grad_norm": 0.016382789239287376,
+ "learning_rate": 0.0006,
+ "loss": 4.347662448883057,
+ "step": 1916
+ },
+ {
+ "epoch": 26.62882096069869,
+ "grad_norm": 0.016581876203417778,
+ "learning_rate": 0.0006,
+ "loss": 4.2716193199157715,
+ "step": 1917
+ },
+ {
+ "epoch": 26.64279475982533,
+ "grad_norm": 0.015627184882760048,
+ "learning_rate": 0.0006,
+ "loss": 4.311489105224609,
+ "step": 1918
+ },
+ {
+ "epoch": 26.656768558951963,
+ "grad_norm": 0.015908483415842056,
+ "learning_rate": 0.0006,
+ "loss": 4.262099742889404,
+ "step": 1919
+ },
+ {
+ "epoch": 26.670742358078602,
+ "grad_norm": 0.015576236881315708,
+ "learning_rate": 0.0006,
+ "loss": 4.275871276855469,
+ "step": 1920
+ },
+ {
+ "epoch": 26.68471615720524,
+ "grad_norm": 0.014621064998209476,
+ "learning_rate": 0.0006,
+ "loss": 4.184260845184326,
+ "step": 1921
+ },
+ {
+ "epoch": 26.69868995633188,
+ "grad_norm": 0.014797811396420002,
+ "learning_rate": 0.0006,
+ "loss": 4.3412628173828125,
+ "step": 1922
+ },
+ {
+ "epoch": 26.712663755458514,
+ "grad_norm": 0.015610570088028908,
+ "learning_rate": 0.0006,
+ "loss": 4.181094646453857,
+ "step": 1923
+ },
+ {
+ "epoch": 26.726637554585153,
+ "grad_norm": 0.016048265621066093,
+ "learning_rate": 0.0006,
+ "loss": 4.365290641784668,
+ "step": 1924
+ },
+ {
+ "epoch": 26.74061135371179,
+ "grad_norm": 0.016604071483016014,
+ "learning_rate": 0.0006,
+ "loss": 4.2628912925720215,
+ "step": 1925
+ },
+ {
+ "epoch": 26.754585152838427,
+ "grad_norm": 0.017044221982359886,
+ "learning_rate": 0.0006,
+ "loss": 4.378512382507324,
+ "step": 1926
+ },
+ {
+ "epoch": 26.768558951965066,
+ "grad_norm": 0.01746384985744953,
+ "learning_rate": 0.0006,
+ "loss": 4.3888773918151855,
+ "step": 1927
+ },
+ {
+ "epoch": 26.782532751091704,
+ "grad_norm": 0.01572711206972599,
+ "learning_rate": 0.0006,
+ "loss": 4.189935684204102,
+ "step": 1928
+ },
+ {
+ "epoch": 26.79650655021834,
+ "grad_norm": 0.014976629987359047,
+ "learning_rate": 0.0006,
+ "loss": 4.280525207519531,
+ "step": 1929
+ },
+ {
+ "epoch": 26.810480349344978,
+ "grad_norm": 0.014804664999246597,
+ "learning_rate": 0.0006,
+ "loss": 4.270630836486816,
+ "step": 1930
+ },
+ {
+ "epoch": 26.824454148471617,
+ "grad_norm": 0.014641920104622841,
+ "learning_rate": 0.0006,
+ "loss": 4.319505214691162,
+ "step": 1931
+ },
+ {
+ "epoch": 26.83842794759825,
+ "grad_norm": 0.014928505755960941,
+ "learning_rate": 0.0006,
+ "loss": 4.180147647857666,
+ "step": 1932
+ },
+ {
+ "epoch": 26.85240174672489,
+ "grad_norm": 0.016130365431308746,
+ "learning_rate": 0.0006,
+ "loss": 4.260660648345947,
+ "step": 1933
+ },
+ {
+ "epoch": 26.86637554585153,
+ "grad_norm": 0.015642493963241577,
+ "learning_rate": 0.0006,
+ "loss": 4.304064750671387,
+ "step": 1934
+ },
+ {
+ "epoch": 26.880349344978168,
+ "grad_norm": 0.015988217666745186,
+ "learning_rate": 0.0006,
+ "loss": 4.2548017501831055,
+ "step": 1935
+ },
+ {
+ "epoch": 26.894323144104803,
+ "grad_norm": 0.016501398757100105,
+ "learning_rate": 0.0006,
+ "loss": 4.260058879852295,
+ "step": 1936
+ },
+ {
+ "epoch": 26.90829694323144,
+ "grad_norm": 0.01645040698349476,
+ "learning_rate": 0.0006,
+ "loss": 4.483081340789795,
+ "step": 1937
+ },
+ {
+ "epoch": 26.92227074235808,
+ "grad_norm": 0.016140863299369812,
+ "learning_rate": 0.0006,
+ "loss": 4.1776838302612305,
+ "step": 1938
+ },
+ {
+ "epoch": 26.936244541484715,
+ "grad_norm": 0.016681019216775894,
+ "learning_rate": 0.0006,
+ "loss": 4.224747657775879,
+ "step": 1939
+ },
+ {
+ "epoch": 26.950218340611354,
+ "grad_norm": 0.01650378853082657,
+ "learning_rate": 0.0006,
+ "loss": 4.305703163146973,
+ "step": 1940
+ },
+ {
+ "epoch": 26.964192139737992,
+ "grad_norm": 0.017051683738827705,
+ "learning_rate": 0.0006,
+ "loss": 4.285835266113281,
+ "step": 1941
+ },
+ {
+ "epoch": 26.978165938864628,
+ "grad_norm": 0.016894327476620674,
+ "learning_rate": 0.0006,
+ "loss": 4.318901538848877,
+ "step": 1942
+ },
+ {
+ "epoch": 26.992139737991266,
+ "grad_norm": 0.017313100397586823,
+ "learning_rate": 0.0006,
+ "loss": 4.281434535980225,
+ "step": 1943
+ },
+ {
+ "epoch": 27.0,
+ "grad_norm": 0.018450884148478508,
+ "learning_rate": 0.0006,
+ "loss": 4.335104942321777,
+ "step": 1944
+ },
+ {
+ "epoch": 27.0,
+ "eval_loss": 4.640687465667725,
+ "eval_runtime": 56.9624,
+ "eval_samples_per_second": 42.87,
+ "eval_steps_per_second": 1.352,
+ "step": 1944
+ },
+ {
+ "epoch": 27.01397379912664,
+ "grad_norm": 0.017755698412656784,
+ "learning_rate": 0.0006,
+ "loss": 4.278824329376221,
+ "step": 1945
+ },
+ {
+ "epoch": 27.027947598253274,
+ "grad_norm": 0.01724310778081417,
+ "learning_rate": 0.0006,
+ "loss": 4.311049461364746,
+ "step": 1946
+ },
+ {
+ "epoch": 27.041921397379912,
+ "grad_norm": 0.01690123789012432,
+ "learning_rate": 0.0006,
+ "loss": 4.261994361877441,
+ "step": 1947
+ },
+ {
+ "epoch": 27.05589519650655,
+ "grad_norm": 0.018406696617603302,
+ "learning_rate": 0.0006,
+ "loss": 4.1925153732299805,
+ "step": 1948
+ },
+ {
+ "epoch": 27.069868995633186,
+ "grad_norm": 0.019945867359638214,
+ "learning_rate": 0.0006,
+ "loss": 4.108537197113037,
+ "step": 1949
+ },
+ {
+ "epoch": 27.083842794759825,
+ "grad_norm": 0.020316628739237785,
+ "learning_rate": 0.0006,
+ "loss": 4.209043025970459,
+ "step": 1950
+ },
+ {
+ "epoch": 27.097816593886463,
+ "grad_norm": 0.019690370187163353,
+ "learning_rate": 0.0006,
+ "loss": 4.195789337158203,
+ "step": 1951
+ },
+ {
+ "epoch": 27.111790393013102,
+ "grad_norm": 0.01785232499241829,
+ "learning_rate": 0.0006,
+ "loss": 4.21466064453125,
+ "step": 1952
+ },
+ {
+ "epoch": 27.125764192139737,
+ "grad_norm": 0.0164767038077116,
+ "learning_rate": 0.0006,
+ "loss": 4.270247459411621,
+ "step": 1953
+ },
+ {
+ "epoch": 27.139737991266376,
+ "grad_norm": 0.018008455634117126,
+ "learning_rate": 0.0006,
+ "loss": 4.201042175292969,
+ "step": 1954
+ },
+ {
+ "epoch": 27.153711790393015,
+ "grad_norm": 0.01807340793311596,
+ "learning_rate": 0.0006,
+ "loss": 4.25289249420166,
+ "step": 1955
+ },
+ {
+ "epoch": 27.16768558951965,
+ "grad_norm": 0.016246909275650978,
+ "learning_rate": 0.0006,
+ "loss": 4.116283416748047,
+ "step": 1956
+ },
+ {
+ "epoch": 27.18165938864629,
+ "grad_norm": 0.017545776441693306,
+ "learning_rate": 0.0006,
+ "loss": 4.316999435424805,
+ "step": 1957
+ },
+ {
+ "epoch": 27.195633187772927,
+ "grad_norm": 0.017915375530719757,
+ "learning_rate": 0.0006,
+ "loss": 4.253859519958496,
+ "step": 1958
+ },
+ {
+ "epoch": 27.209606986899562,
+ "grad_norm": 0.018112223595380783,
+ "learning_rate": 0.0006,
+ "loss": 4.2541022300720215,
+ "step": 1959
+ },
+ {
+ "epoch": 27.2235807860262,
+ "grad_norm": 0.01846139505505562,
+ "learning_rate": 0.0006,
+ "loss": 4.20438289642334,
+ "step": 1960
+ },
+ {
+ "epoch": 27.23755458515284,
+ "grad_norm": 0.01943897269666195,
+ "learning_rate": 0.0006,
+ "loss": 4.2799577713012695,
+ "step": 1961
+ },
+ {
+ "epoch": 27.251528384279474,
+ "grad_norm": 0.018253348767757416,
+ "learning_rate": 0.0006,
+ "loss": 4.1720499992370605,
+ "step": 1962
+ },
+ {
+ "epoch": 27.265502183406113,
+ "grad_norm": 0.018292246386408806,
+ "learning_rate": 0.0006,
+ "loss": 4.28231143951416,
+ "step": 1963
+ },
+ {
+ "epoch": 27.27947598253275,
+ "grad_norm": 0.01943155936896801,
+ "learning_rate": 0.0006,
+ "loss": 4.314976692199707,
+ "step": 1964
+ },
+ {
+ "epoch": 27.29344978165939,
+ "grad_norm": 0.019064994528889656,
+ "learning_rate": 0.0006,
+ "loss": 4.25328254699707,
+ "step": 1965
+ },
+ {
+ "epoch": 27.307423580786025,
+ "grad_norm": 0.017423273995518684,
+ "learning_rate": 0.0006,
+ "loss": 4.195992469787598,
+ "step": 1966
+ },
+ {
+ "epoch": 27.321397379912664,
+ "grad_norm": 0.01677924580872059,
+ "learning_rate": 0.0006,
+ "loss": 4.260043144226074,
+ "step": 1967
+ },
+ {
+ "epoch": 27.335371179039303,
+ "grad_norm": 0.01720673404633999,
+ "learning_rate": 0.0006,
+ "loss": 4.222780704498291,
+ "step": 1968
+ },
+ {
+ "epoch": 27.349344978165938,
+ "grad_norm": 0.017300743609666824,
+ "learning_rate": 0.0006,
+ "loss": 4.109524250030518,
+ "step": 1969
+ },
+ {
+ "epoch": 27.363318777292577,
+ "grad_norm": 0.016065871343016624,
+ "learning_rate": 0.0006,
+ "loss": 4.36108922958374,
+ "step": 1970
+ },
+ {
+ "epoch": 27.377292576419215,
+ "grad_norm": 0.016626616939902306,
+ "learning_rate": 0.0006,
+ "loss": 4.220630645751953,
+ "step": 1971
+ },
+ {
+ "epoch": 27.39126637554585,
+ "grad_norm": 0.016482602804899216,
+ "learning_rate": 0.0006,
+ "loss": 4.344797134399414,
+ "step": 1972
+ },
+ {
+ "epoch": 27.40524017467249,
+ "grad_norm": 0.01727953553199768,
+ "learning_rate": 0.0006,
+ "loss": 4.296645164489746,
+ "step": 1973
+ },
+ {
+ "epoch": 27.419213973799128,
+ "grad_norm": 0.01522731315344572,
+ "learning_rate": 0.0006,
+ "loss": 4.223751068115234,
+ "step": 1974
+ },
+ {
+ "epoch": 27.433187772925763,
+ "grad_norm": 0.01581776700913906,
+ "learning_rate": 0.0006,
+ "loss": 4.351379871368408,
+ "step": 1975
+ },
+ {
+ "epoch": 27.4471615720524,
+ "grad_norm": 0.016394605860114098,
+ "learning_rate": 0.0006,
+ "loss": 4.2412919998168945,
+ "step": 1976
+ },
+ {
+ "epoch": 27.46113537117904,
+ "grad_norm": 0.01751169003546238,
+ "learning_rate": 0.0006,
+ "loss": 4.257606506347656,
+ "step": 1977
+ },
+ {
+ "epoch": 27.475109170305675,
+ "grad_norm": 0.017452310770750046,
+ "learning_rate": 0.0006,
+ "loss": 4.380134582519531,
+ "step": 1978
+ },
+ {
+ "epoch": 27.489082969432314,
+ "grad_norm": 0.017741898074746132,
+ "learning_rate": 0.0006,
+ "loss": 4.279632568359375,
+ "step": 1979
+ },
+ {
+ "epoch": 27.503056768558952,
+ "grad_norm": 0.017680590972304344,
+ "learning_rate": 0.0006,
+ "loss": 4.273406028747559,
+ "step": 1980
+ },
+ {
+ "epoch": 27.51703056768559,
+ "grad_norm": 0.01864718459546566,
+ "learning_rate": 0.0006,
+ "loss": 4.278583526611328,
+ "step": 1981
+ },
+ {
+ "epoch": 27.531004366812226,
+ "grad_norm": 0.019558541476726532,
+ "learning_rate": 0.0006,
+ "loss": 4.211919784545898,
+ "step": 1982
+ },
+ {
+ "epoch": 27.544978165938865,
+ "grad_norm": 0.019467400386929512,
+ "learning_rate": 0.0006,
+ "loss": 4.276822566986084,
+ "step": 1983
+ },
+ {
+ "epoch": 27.558951965065503,
+ "grad_norm": 0.01756172813475132,
+ "learning_rate": 0.0006,
+ "loss": 4.165379047393799,
+ "step": 1984
+ },
+ {
+ "epoch": 27.57292576419214,
+ "grad_norm": 0.017458263784646988,
+ "learning_rate": 0.0006,
+ "loss": 4.2992401123046875,
+ "step": 1985
+ },
+ {
+ "epoch": 27.586899563318777,
+ "grad_norm": 0.018089665099978447,
+ "learning_rate": 0.0006,
+ "loss": 4.243378639221191,
+ "step": 1986
+ },
+ {
+ "epoch": 27.600873362445416,
+ "grad_norm": 0.019535524770617485,
+ "learning_rate": 0.0006,
+ "loss": 4.321477890014648,
+ "step": 1987
+ },
+ {
+ "epoch": 27.61484716157205,
+ "grad_norm": 0.01997970975935459,
+ "learning_rate": 0.0006,
+ "loss": 4.123082637786865,
+ "step": 1988
+ },
+ {
+ "epoch": 27.62882096069869,
+ "grad_norm": 0.01913067139685154,
+ "learning_rate": 0.0006,
+ "loss": 4.289896488189697,
+ "step": 1989
+ },
+ {
+ "epoch": 27.64279475982533,
+ "grad_norm": 0.0169806070625782,
+ "learning_rate": 0.0006,
+ "loss": 4.247004985809326,
+ "step": 1990
+ },
+ {
+ "epoch": 27.656768558951963,
+ "grad_norm": 0.017315007746219635,
+ "learning_rate": 0.0006,
+ "loss": 4.2095441818237305,
+ "step": 1991
+ },
+ {
+ "epoch": 27.670742358078602,
+ "grad_norm": 0.015098759904503822,
+ "learning_rate": 0.0006,
+ "loss": 4.145560264587402,
+ "step": 1992
+ },
+ {
+ "epoch": 27.68471615720524,
+ "grad_norm": 0.016240514814853668,
+ "learning_rate": 0.0006,
+ "loss": 4.254927635192871,
+ "step": 1993
+ },
+ {
+ "epoch": 27.69868995633188,
+ "grad_norm": 0.01621108129620552,
+ "learning_rate": 0.0006,
+ "loss": 4.219725608825684,
+ "step": 1994
+ },
+ {
+ "epoch": 27.712663755458514,
+ "grad_norm": 0.015191009268164635,
+ "learning_rate": 0.0006,
+ "loss": 4.194345474243164,
+ "step": 1995
+ },
+ {
+ "epoch": 27.726637554585153,
+ "grad_norm": 0.014387983828783035,
+ "learning_rate": 0.0006,
+ "loss": 4.211467742919922,
+ "step": 1996
+ },
+ {
+ "epoch": 27.74061135371179,
+ "grad_norm": 0.015460561029613018,
+ "learning_rate": 0.0006,
+ "loss": 4.358700752258301,
+ "step": 1997
+ },
+ {
+ "epoch": 27.754585152838427,
+ "grad_norm": 0.015100076794624329,
+ "learning_rate": 0.0006,
+ "loss": 4.171223163604736,
+ "step": 1998
+ },
+ {
+ "epoch": 27.768558951965066,
+ "grad_norm": 0.015607891604304314,
+ "learning_rate": 0.0006,
+ "loss": 4.217874526977539,
+ "step": 1999
+ },
+ {
+ "epoch": 27.782532751091704,
+ "grad_norm": 0.015650689601898193,
+ "learning_rate": 0.0006,
+ "loss": 4.280780792236328,
+ "step": 2000
+ },
+ {
+ "epoch": 27.79650655021834,
+ "grad_norm": 0.01580994389951229,
+ "learning_rate": 0.0006,
+ "loss": 4.159212112426758,
+ "step": 2001
+ },
+ {
+ "epoch": 27.810480349344978,
+ "grad_norm": 0.016607413068413734,
+ "learning_rate": 0.0006,
+ "loss": 4.238853454589844,
+ "step": 2002
+ },
+ {
+ "epoch": 27.824454148471617,
+ "grad_norm": 0.017383860424160957,
+ "learning_rate": 0.0006,
+ "loss": 4.233944892883301,
+ "step": 2003
+ },
+ {
+ "epoch": 27.83842794759825,
+ "grad_norm": 0.014966659247875214,
+ "learning_rate": 0.0006,
+ "loss": 4.167392730712891,
+ "step": 2004
+ },
+ {
+ "epoch": 27.85240174672489,
+ "grad_norm": 0.014589878730475903,
+ "learning_rate": 0.0006,
+ "loss": 4.2390899658203125,
+ "step": 2005
+ },
+ {
+ "epoch": 27.86637554585153,
+ "grad_norm": 0.017179450020194054,
+ "learning_rate": 0.0006,
+ "loss": 4.365785598754883,
+ "step": 2006
+ },
+ {
+ "epoch": 27.880349344978168,
+ "grad_norm": 0.01893901452422142,
+ "learning_rate": 0.0006,
+ "loss": 4.289430141448975,
+ "step": 2007
+ },
+ {
+ "epoch": 27.894323144104803,
+ "grad_norm": 0.018619712442159653,
+ "learning_rate": 0.0006,
+ "loss": 4.310665130615234,
+ "step": 2008
+ },
+ {
+ "epoch": 27.90829694323144,
+ "grad_norm": 0.019180385395884514,
+ "learning_rate": 0.0006,
+ "loss": 4.27454137802124,
+ "step": 2009
+ },
+ {
+ "epoch": 27.92227074235808,
+ "grad_norm": 0.017811523750424385,
+ "learning_rate": 0.0006,
+ "loss": 4.312562465667725,
+ "step": 2010
+ },
+ {
+ "epoch": 27.936244541484715,
+ "grad_norm": 0.016583573073148727,
+ "learning_rate": 0.0006,
+ "loss": 4.202253341674805,
+ "step": 2011
+ },
+ {
+ "epoch": 27.950218340611354,
+ "grad_norm": 0.01658615842461586,
+ "learning_rate": 0.0006,
+ "loss": 4.183433532714844,
+ "step": 2012
+ },
+ {
+ "epoch": 27.964192139737992,
+ "grad_norm": 0.015327401459217072,
+ "learning_rate": 0.0006,
+ "loss": 4.2125139236450195,
+ "step": 2013
+ },
+ {
+ "epoch": 27.978165938864628,
+ "grad_norm": 0.015004601329565048,
+ "learning_rate": 0.0006,
+ "loss": 4.257392883300781,
+ "step": 2014
+ },
+ {
+ "epoch": 27.992139737991266,
+ "grad_norm": 0.015259161591529846,
+ "learning_rate": 0.0006,
+ "loss": 4.241818428039551,
+ "step": 2015
+ },
+ {
+ "epoch": 28.0,
+ "grad_norm": 0.016808776184916496,
+ "learning_rate": 0.0006,
+ "loss": 4.146241188049316,
+ "step": 2016
+ },
+ {
+ "epoch": 28.0,
+ "eval_loss": 4.663090705871582,
+ "eval_runtime": 56.4559,
+ "eval_samples_per_second": 43.255,
+ "eval_steps_per_second": 1.364,
+ "step": 2016
+ },
+ {
+ "epoch": 28.01397379912664,
+ "grad_norm": 0.015784382820129395,
+ "learning_rate": 0.0006,
+ "loss": 4.044191360473633,
+ "step": 2017
+ },
+ {
+ "epoch": 28.027947598253274,
+ "grad_norm": 0.015636710450053215,
+ "learning_rate": 0.0006,
+ "loss": 4.128754615783691,
+ "step": 2018
+ },
+ {
+ "epoch": 28.041921397379912,
+ "grad_norm": 0.015364975668489933,
+ "learning_rate": 0.0006,
+ "loss": 4.191835403442383,
+ "step": 2019
+ },
+ {
+ "epoch": 28.05589519650655,
+ "grad_norm": 0.015015700832009315,
+ "learning_rate": 0.0006,
+ "loss": 4.120022773742676,
+ "step": 2020
+ },
+ {
+ "epoch": 28.069868995633186,
+ "grad_norm": 0.016666799783706665,
+ "learning_rate": 0.0006,
+ "loss": 4.33906364440918,
+ "step": 2021
+ },
+ {
+ "epoch": 28.083842794759825,
+ "grad_norm": 0.01780104823410511,
+ "learning_rate": 0.0006,
+ "loss": 4.279458999633789,
+ "step": 2022
+ },
+ {
+ "epoch": 28.097816593886463,
+ "grad_norm": 0.01870882883667946,
+ "learning_rate": 0.0006,
+ "loss": 4.34728479385376,
+ "step": 2023
+ },
+ {
+ "epoch": 28.111790393013102,
+ "grad_norm": 0.018406381830573082,
+ "learning_rate": 0.0006,
+ "loss": 4.211311340332031,
+ "step": 2024
+ },
+ {
+ "epoch": 28.125764192139737,
+ "grad_norm": 0.017833448946475983,
+ "learning_rate": 0.0006,
+ "loss": 4.223254203796387,
+ "step": 2025
+ },
+ {
+ "epoch": 28.139737991266376,
+ "grad_norm": 0.017552688717842102,
+ "learning_rate": 0.0006,
+ "loss": 4.208823204040527,
+ "step": 2026
+ },
+ {
+ "epoch": 28.153711790393015,
+ "grad_norm": 0.018500229343771935,
+ "learning_rate": 0.0006,
+ "loss": 4.198025226593018,
+ "step": 2027
+ },
+ {
+ "epoch": 28.16768558951965,
+ "grad_norm": 0.01782156340777874,
+ "learning_rate": 0.0006,
+ "loss": 4.179978370666504,
+ "step": 2028
+ },
+ {
+ "epoch": 28.18165938864629,
+ "grad_norm": 0.019580967724323273,
+ "learning_rate": 0.0006,
+ "loss": 4.291114807128906,
+ "step": 2029
+ },
+ {
+ "epoch": 28.195633187772927,
+ "grad_norm": 0.020804036408662796,
+ "learning_rate": 0.0006,
+ "loss": 4.220607280731201,
+ "step": 2030
+ },
+ {
+ "epoch": 28.209606986899562,
+ "grad_norm": 0.019595693796873093,
+ "learning_rate": 0.0006,
+ "loss": 4.2905731201171875,
+ "step": 2031
+ },
+ {
+ "epoch": 28.2235807860262,
+ "grad_norm": 0.019247086718678474,
+ "learning_rate": 0.0006,
+ "loss": 4.2467360496521,
+ "step": 2032
+ },
+ {
+ "epoch": 28.23755458515284,
+ "grad_norm": 0.01958036608994007,
+ "learning_rate": 0.0006,
+ "loss": 4.286137580871582,
+ "step": 2033
+ },
+ {
+ "epoch": 28.251528384279474,
+ "grad_norm": 0.021354753524065018,
+ "learning_rate": 0.0006,
+ "loss": 4.234755039215088,
+ "step": 2034
+ },
+ {
+ "epoch": 28.265502183406113,
+ "grad_norm": 0.02206375077366829,
+ "learning_rate": 0.0006,
+ "loss": 4.149641036987305,
+ "step": 2035
+ },
+ {
+ "epoch": 28.27947598253275,
+ "grad_norm": 0.02201310358941555,
+ "learning_rate": 0.0006,
+ "loss": 4.1928629875183105,
+ "step": 2036
+ },
+ {
+ "epoch": 28.29344978165939,
+ "grad_norm": 0.01961950585246086,
+ "learning_rate": 0.0006,
+ "loss": 4.211404323577881,
+ "step": 2037
+ },
+ {
+ "epoch": 28.307423580786025,
+ "grad_norm": 0.019211795181035995,
+ "learning_rate": 0.0006,
+ "loss": 4.200976848602295,
+ "step": 2038
+ },
+ {
+ "epoch": 28.321397379912664,
+ "grad_norm": 0.01779370754957199,
+ "learning_rate": 0.0006,
+ "loss": 4.249148845672607,
+ "step": 2039
+ },
+ {
+ "epoch": 28.335371179039303,
+ "grad_norm": 0.018635908141732216,
+ "learning_rate": 0.0006,
+ "loss": 4.11806058883667,
+ "step": 2040
+ },
+ {
+ "epoch": 28.349344978165938,
+ "grad_norm": 0.019792694598436356,
+ "learning_rate": 0.0006,
+ "loss": 4.335224151611328,
+ "step": 2041
+ },
+ {
+ "epoch": 28.363318777292577,
+ "grad_norm": 0.021731717512011528,
+ "learning_rate": 0.0006,
+ "loss": 4.302778244018555,
+ "step": 2042
+ },
+ {
+ "epoch": 28.377292576419215,
+ "grad_norm": 0.020823447033762932,
+ "learning_rate": 0.0006,
+ "loss": 4.260610580444336,
+ "step": 2043
+ },
+ {
+ "epoch": 28.39126637554585,
+ "grad_norm": 0.019385412335395813,
+ "learning_rate": 0.0006,
+ "loss": 4.2247724533081055,
+ "step": 2044
+ },
+ {
+ "epoch": 28.40524017467249,
+ "grad_norm": 0.019539108499884605,
+ "learning_rate": 0.0006,
+ "loss": 4.236655235290527,
+ "step": 2045
+ },
+ {
+ "epoch": 28.419213973799128,
+ "grad_norm": 0.01942690648138523,
+ "learning_rate": 0.0006,
+ "loss": 4.251850605010986,
+ "step": 2046
+ },
+ {
+ "epoch": 28.433187772925763,
+ "grad_norm": 0.019667886197566986,
+ "learning_rate": 0.0006,
+ "loss": 4.222312927246094,
+ "step": 2047
+ },
+ {
+ "epoch": 28.4471615720524,
+ "grad_norm": 0.02029012329876423,
+ "learning_rate": 0.0006,
+ "loss": 4.246252059936523,
+ "step": 2048
+ },
+ {
+ "epoch": 28.46113537117904,
+ "grad_norm": 0.01784469373524189,
+ "learning_rate": 0.0006,
+ "loss": 4.192128658294678,
+ "step": 2049
+ },
+ {
+ "epoch": 28.475109170305675,
+ "grad_norm": 0.0160287544131279,
+ "learning_rate": 0.0006,
+ "loss": 4.208607196807861,
+ "step": 2050
+ },
+ {
+ "epoch": 28.489082969432314,
+ "grad_norm": 0.015348346903920174,
+ "learning_rate": 0.0006,
+ "loss": 4.236598014831543,
+ "step": 2051
+ },
+ {
+ "epoch": 28.503056768558952,
+ "grad_norm": 0.015198206529021263,
+ "learning_rate": 0.0006,
+ "loss": 4.235552787780762,
+ "step": 2052
+ },
+ {
+ "epoch": 28.51703056768559,
+ "grad_norm": 0.01523754745721817,
+ "learning_rate": 0.0006,
+ "loss": 4.1941728591918945,
+ "step": 2053
+ },
+ {
+ "epoch": 28.531004366812226,
+ "grad_norm": 0.015559614636003971,
+ "learning_rate": 0.0006,
+ "loss": 4.294757843017578,
+ "step": 2054
+ },
+ {
+ "epoch": 28.544978165938865,
+ "grad_norm": 0.01568358950316906,
+ "learning_rate": 0.0006,
+ "loss": 4.213065147399902,
+ "step": 2055
+ },
+ {
+ "epoch": 28.558951965065503,
+ "grad_norm": 0.015336516313254833,
+ "learning_rate": 0.0006,
+ "loss": 4.240293979644775,
+ "step": 2056
+ },
+ {
+ "epoch": 28.57292576419214,
+ "grad_norm": 0.015041496604681015,
+ "learning_rate": 0.0006,
+ "loss": 4.273397445678711,
+ "step": 2057
+ },
+ {
+ "epoch": 28.586899563318777,
+ "grad_norm": 0.01610128954052925,
+ "learning_rate": 0.0006,
+ "loss": 4.125369071960449,
+ "step": 2058
+ },
+ {
+ "epoch": 28.600873362445416,
+ "grad_norm": 0.015974976122379303,
+ "learning_rate": 0.0006,
+ "loss": 4.133459091186523,
+ "step": 2059
+ },
+ {
+ "epoch": 28.61484716157205,
+ "grad_norm": 0.015936799347400665,
+ "learning_rate": 0.0006,
+ "loss": 4.306667327880859,
+ "step": 2060
+ },
+ {
+ "epoch": 28.62882096069869,
+ "grad_norm": 0.01795247197151184,
+ "learning_rate": 0.0006,
+ "loss": 4.205156326293945,
+ "step": 2061
+ },
+ {
+ "epoch": 28.64279475982533,
+ "grad_norm": 0.019639814272522926,
+ "learning_rate": 0.0006,
+ "loss": 4.277614593505859,
+ "step": 2062
+ },
+ {
+ "epoch": 28.656768558951963,
+ "grad_norm": 0.01722300611436367,
+ "learning_rate": 0.0006,
+ "loss": 4.191032886505127,
+ "step": 2063
+ },
+ {
+ "epoch": 28.670742358078602,
+ "grad_norm": 0.01612633652985096,
+ "learning_rate": 0.0006,
+ "loss": 4.20717191696167,
+ "step": 2064
+ },
+ {
+ "epoch": 28.68471615720524,
+ "grad_norm": 0.01603279449045658,
+ "learning_rate": 0.0006,
+ "loss": 4.228884220123291,
+ "step": 2065
+ },
+ {
+ "epoch": 28.69868995633188,
+ "grad_norm": 0.016405927017331123,
+ "learning_rate": 0.0006,
+ "loss": 4.201292991638184,
+ "step": 2066
+ },
+ {
+ "epoch": 28.712663755458514,
+ "grad_norm": 0.016644183546304703,
+ "learning_rate": 0.0006,
+ "loss": 4.239282131195068,
+ "step": 2067
+ },
+ {
+ "epoch": 28.726637554585153,
+ "grad_norm": 0.0175587497651577,
+ "learning_rate": 0.0006,
+ "loss": 4.289038181304932,
+ "step": 2068
+ },
+ {
+ "epoch": 28.74061135371179,
+ "grad_norm": 0.018223397433757782,
+ "learning_rate": 0.0006,
+ "loss": 4.2266998291015625,
+ "step": 2069
+ },
+ {
+ "epoch": 28.754585152838427,
+ "grad_norm": 0.018680868670344353,
+ "learning_rate": 0.0006,
+ "loss": 4.348849296569824,
+ "step": 2070
+ },
+ {
+ "epoch": 28.768558951965066,
+ "grad_norm": 0.016726728528738022,
+ "learning_rate": 0.0006,
+ "loss": 4.292080879211426,
+ "step": 2071
+ },
+ {
+ "epoch": 28.782532751091704,
+ "grad_norm": 0.017386795952916145,
+ "learning_rate": 0.0006,
+ "loss": 4.208870887756348,
+ "step": 2072
+ },
+ {
+ "epoch": 28.79650655021834,
+ "grad_norm": 0.01865178905427456,
+ "learning_rate": 0.0006,
+ "loss": 4.054150581359863,
+ "step": 2073
+ },
+ {
+ "epoch": 28.810480349344978,
+ "grad_norm": 0.01776815392076969,
+ "learning_rate": 0.0006,
+ "loss": 4.202404499053955,
+ "step": 2074
+ },
+ {
+ "epoch": 28.824454148471617,
+ "grad_norm": 0.01643497310578823,
+ "learning_rate": 0.0006,
+ "loss": 4.131855010986328,
+ "step": 2075
+ },
+ {
+ "epoch": 28.83842794759825,
+ "grad_norm": 0.01574273779988289,
+ "learning_rate": 0.0006,
+ "loss": 4.296788692474365,
+ "step": 2076
+ },
+ {
+ "epoch": 28.85240174672489,
+ "grad_norm": 0.015613908879458904,
+ "learning_rate": 0.0006,
+ "loss": 4.210949897766113,
+ "step": 2077
+ },
+ {
+ "epoch": 28.86637554585153,
+ "grad_norm": 0.015543212182819843,
+ "learning_rate": 0.0006,
+ "loss": 4.218136310577393,
+ "step": 2078
+ },
+ {
+ "epoch": 28.880349344978168,
+ "grad_norm": 0.014681251719594002,
+ "learning_rate": 0.0006,
+ "loss": 4.19401741027832,
+ "step": 2079
+ },
+ {
+ "epoch": 28.894323144104803,
+ "grad_norm": 0.014393828809261322,
+ "learning_rate": 0.0006,
+ "loss": 4.198468208312988,
+ "step": 2080
+ },
+ {
+ "epoch": 28.90829694323144,
+ "grad_norm": 0.01580170728266239,
+ "learning_rate": 0.0006,
+ "loss": 4.224969863891602,
+ "step": 2081
+ },
+ {
+ "epoch": 28.92227074235808,
+ "grad_norm": 0.01484165620058775,
+ "learning_rate": 0.0006,
+ "loss": 4.12236213684082,
+ "step": 2082
+ },
+ {
+ "epoch": 28.936244541484715,
+ "grad_norm": 0.01497623696923256,
+ "learning_rate": 0.0006,
+ "loss": 4.24141788482666,
+ "step": 2083
+ },
+ {
+ "epoch": 28.950218340611354,
+ "grad_norm": 0.015080046840012074,
+ "learning_rate": 0.0006,
+ "loss": 4.284151077270508,
+ "step": 2084
+ },
+ {
+ "epoch": 28.964192139737992,
+ "grad_norm": 0.016207212582230568,
+ "learning_rate": 0.0006,
+ "loss": 4.298693656921387,
+ "step": 2085
+ },
+ {
+ "epoch": 28.978165938864628,
+ "grad_norm": 0.015474417246878147,
+ "learning_rate": 0.0006,
+ "loss": 4.2412261962890625,
+ "step": 2086
+ },
+ {
+ "epoch": 28.992139737991266,
+ "grad_norm": 0.01515351701527834,
+ "learning_rate": 0.0006,
+ "loss": 4.3054423332214355,
+ "step": 2087
+ },
+ {
+ "epoch": 29.0,
+ "grad_norm": 0.016411345452070236,
+ "learning_rate": 0.0006,
+ "loss": 4.104830741882324,
+ "step": 2088
+ },
+ {
+ "epoch": 29.0,
+ "eval_loss": 4.661567687988281,
+ "eval_runtime": 57.1244,
+ "eval_samples_per_second": 42.749,
+ "eval_steps_per_second": 1.348,
+ "step": 2088
+ },
+ {
+ "epoch": 29.01397379912664,
+ "grad_norm": 0.01665342226624489,
+ "learning_rate": 0.0006,
+ "loss": 4.311391830444336,
+ "step": 2089
+ },
+ {
+ "epoch": 29.027947598253274,
+ "grad_norm": 0.018053214997053146,
+ "learning_rate": 0.0006,
+ "loss": 4.135597229003906,
+ "step": 2090
+ },
+ {
+ "epoch": 29.041921397379912,
+ "grad_norm": 0.018213748931884766,
+ "learning_rate": 0.0006,
+ "loss": 4.303299903869629,
+ "step": 2091
+ },
+ {
+ "epoch": 29.05589519650655,
+ "grad_norm": 0.018460242077708244,
+ "learning_rate": 0.0006,
+ "loss": 4.20850944519043,
+ "step": 2092
+ },
+ {
+ "epoch": 29.069868995633186,
+ "grad_norm": 0.018365809693932533,
+ "learning_rate": 0.0006,
+ "loss": 4.275847434997559,
+ "step": 2093
+ },
+ {
+ "epoch": 29.083842794759825,
+ "grad_norm": 0.018346186727285385,
+ "learning_rate": 0.0006,
+ "loss": 4.202091217041016,
+ "step": 2094
+ },
+ {
+ "epoch": 29.097816593886463,
+ "grad_norm": 0.017920587211847305,
+ "learning_rate": 0.0006,
+ "loss": 4.182092666625977,
+ "step": 2095
+ },
+ {
+ "epoch": 29.111790393013102,
+ "grad_norm": 0.01812003180384636,
+ "learning_rate": 0.0006,
+ "loss": 4.259721755981445,
+ "step": 2096
+ },
+ {
+ "epoch": 29.125764192139737,
+ "grad_norm": 0.01700986735522747,
+ "learning_rate": 0.0006,
+ "loss": 4.067573070526123,
+ "step": 2097
+ },
+ {
+ "epoch": 29.139737991266376,
+ "grad_norm": 0.018069760873913765,
+ "learning_rate": 0.0006,
+ "loss": 4.234400749206543,
+ "step": 2098
+ },
+ {
+ "epoch": 29.153711790393015,
+ "grad_norm": 0.01905428245663643,
+ "learning_rate": 0.0006,
+ "loss": 4.16609001159668,
+ "step": 2099
+ },
+ {
+ "epoch": 29.16768558951965,
+ "grad_norm": 0.018711242824792862,
+ "learning_rate": 0.0006,
+ "loss": 4.249088764190674,
+ "step": 2100
+ },
+ {
+ "epoch": 29.18165938864629,
+ "grad_norm": 0.018721306696534157,
+ "learning_rate": 0.0006,
+ "loss": 4.14540958404541,
+ "step": 2101
+ },
+ {
+ "epoch": 29.195633187772927,
+ "grad_norm": 0.020139874890446663,
+ "learning_rate": 0.0006,
+ "loss": 4.049248695373535,
+ "step": 2102
+ },
+ {
+ "epoch": 29.209606986899562,
+ "grad_norm": 0.020849574357271194,
+ "learning_rate": 0.0006,
+ "loss": 4.235224723815918,
+ "step": 2103
+ },
+ {
+ "epoch": 29.2235807860262,
+ "grad_norm": 0.018693549558520317,
+ "learning_rate": 0.0006,
+ "loss": 4.232961177825928,
+ "step": 2104
+ },
+ {
+ "epoch": 29.23755458515284,
+ "grad_norm": 0.017889857292175293,
+ "learning_rate": 0.0006,
+ "loss": 4.171422004699707,
+ "step": 2105
+ },
+ {
+ "epoch": 29.251528384279474,
+ "grad_norm": 0.016690224409103394,
+ "learning_rate": 0.0006,
+ "loss": 4.244717597961426,
+ "step": 2106
+ },
+ {
+ "epoch": 29.265502183406113,
+ "grad_norm": 0.01708616502583027,
+ "learning_rate": 0.0006,
+ "loss": 4.125068664550781,
+ "step": 2107
+ },
+ {
+ "epoch": 29.27947598253275,
+ "grad_norm": 0.017928237095475197,
+ "learning_rate": 0.0006,
+ "loss": 4.208292007446289,
+ "step": 2108
+ },
+ {
+ "epoch": 29.29344978165939,
+ "grad_norm": 0.018250394612550735,
+ "learning_rate": 0.0006,
+ "loss": 4.118630886077881,
+ "step": 2109
+ },
+ {
+ "epoch": 29.307423580786025,
+ "grad_norm": 0.020160246640443802,
+ "learning_rate": 0.0006,
+ "loss": 4.206025123596191,
+ "step": 2110
+ },
+ {
+ "epoch": 29.321397379912664,
+ "grad_norm": 0.020045453682541847,
+ "learning_rate": 0.0006,
+ "loss": 4.2758684158325195,
+ "step": 2111
+ },
+ {
+ "epoch": 29.335371179039303,
+ "grad_norm": 0.019556907936930656,
+ "learning_rate": 0.0006,
+ "loss": 4.188272953033447,
+ "step": 2112
+ },
+ {
+ "epoch": 29.349344978165938,
+ "grad_norm": 0.019185200333595276,
+ "learning_rate": 0.0006,
+ "loss": 4.17451286315918,
+ "step": 2113
+ },
+ {
+ "epoch": 29.363318777292577,
+ "grad_norm": 0.01816629432141781,
+ "learning_rate": 0.0006,
+ "loss": 4.165246963500977,
+ "step": 2114
+ },
+ {
+ "epoch": 29.377292576419215,
+ "grad_norm": 0.01865146867930889,
+ "learning_rate": 0.0006,
+ "loss": 4.148184776306152,
+ "step": 2115
+ },
+ {
+ "epoch": 29.39126637554585,
+ "grad_norm": 0.018338464200496674,
+ "learning_rate": 0.0006,
+ "loss": 4.288936614990234,
+ "step": 2116
+ },
+ {
+ "epoch": 29.40524017467249,
+ "grad_norm": 0.018986187875270844,
+ "learning_rate": 0.0006,
+ "loss": 4.183379650115967,
+ "step": 2117
+ },
+ {
+ "epoch": 29.419213973799128,
+ "grad_norm": 0.018409637734293938,
+ "learning_rate": 0.0006,
+ "loss": 4.057574272155762,
+ "step": 2118
+ },
+ {
+ "epoch": 29.433187772925763,
+ "grad_norm": 0.016664542257785797,
+ "learning_rate": 0.0006,
+ "loss": 4.18517541885376,
+ "step": 2119
+ },
+ {
+ "epoch": 29.4471615720524,
+ "grad_norm": 0.016486089676618576,
+ "learning_rate": 0.0006,
+ "loss": 4.179367542266846,
+ "step": 2120
+ },
+ {
+ "epoch": 29.46113537117904,
+ "grad_norm": 0.01787523552775383,
+ "learning_rate": 0.0006,
+ "loss": 4.184126377105713,
+ "step": 2121
+ },
+ {
+ "epoch": 29.475109170305675,
+ "grad_norm": 0.016574613749980927,
+ "learning_rate": 0.0006,
+ "loss": 4.290170192718506,
+ "step": 2122
+ },
+ {
+ "epoch": 29.489082969432314,
+ "grad_norm": 0.01643376611173153,
+ "learning_rate": 0.0006,
+ "loss": 4.052913665771484,
+ "step": 2123
+ },
+ {
+ "epoch": 29.503056768558952,
+ "grad_norm": 0.016503969207406044,
+ "learning_rate": 0.0006,
+ "loss": 4.335184097290039,
+ "step": 2124
+ },
+ {
+ "epoch": 29.51703056768559,
+ "grad_norm": 0.017596479505300522,
+ "learning_rate": 0.0006,
+ "loss": 4.165539741516113,
+ "step": 2125
+ },
+ {
+ "epoch": 29.531004366812226,
+ "grad_norm": 0.017468582838773727,
+ "learning_rate": 0.0006,
+ "loss": 4.315964698791504,
+ "step": 2126
+ },
+ {
+ "epoch": 29.544978165938865,
+ "grad_norm": 0.01724618673324585,
+ "learning_rate": 0.0006,
+ "loss": 4.243681907653809,
+ "step": 2127
+ },
+ {
+ "epoch": 29.558951965065503,
+ "grad_norm": 0.018384616822004318,
+ "learning_rate": 0.0006,
+ "loss": 4.112447261810303,
+ "step": 2128
+ },
+ {
+ "epoch": 29.57292576419214,
+ "grad_norm": 0.01824074238538742,
+ "learning_rate": 0.0006,
+ "loss": 4.236065864562988,
+ "step": 2129
+ },
+ {
+ "epoch": 29.586899563318777,
+ "grad_norm": 0.01625971868634224,
+ "learning_rate": 0.0006,
+ "loss": 4.18143892288208,
+ "step": 2130
+ },
+ {
+ "epoch": 29.600873362445416,
+ "grad_norm": 0.015537494793534279,
+ "learning_rate": 0.0006,
+ "loss": 4.225147247314453,
+ "step": 2131
+ },
+ {
+ "epoch": 29.61484716157205,
+ "grad_norm": 0.015818974003195763,
+ "learning_rate": 0.0006,
+ "loss": 4.3100175857543945,
+ "step": 2132
+ },
+ {
+ "epoch": 29.62882096069869,
+ "grad_norm": 0.015909671783447266,
+ "learning_rate": 0.0006,
+ "loss": 4.169775009155273,
+ "step": 2133
+ },
+ {
+ "epoch": 29.64279475982533,
+ "grad_norm": 0.01769658550620079,
+ "learning_rate": 0.0006,
+ "loss": 4.278904914855957,
+ "step": 2134
+ },
+ {
+ "epoch": 29.656768558951963,
+ "grad_norm": 0.015641704201698303,
+ "learning_rate": 0.0006,
+ "loss": 4.2426228523254395,
+ "step": 2135
+ },
+ {
+ "epoch": 29.670742358078602,
+ "grad_norm": 0.01565658301115036,
+ "learning_rate": 0.0006,
+ "loss": 4.232687950134277,
+ "step": 2136
+ },
+ {
+ "epoch": 29.68471615720524,
+ "grad_norm": 0.016485940665006638,
+ "learning_rate": 0.0006,
+ "loss": 4.209839820861816,
+ "step": 2137
+ },
+ {
+ "epoch": 29.69868995633188,
+ "grad_norm": 0.017622729763388634,
+ "learning_rate": 0.0006,
+ "loss": 4.196017265319824,
+ "step": 2138
+ },
+ {
+ "epoch": 29.712663755458514,
+ "grad_norm": 0.018320564180612564,
+ "learning_rate": 0.0006,
+ "loss": 4.225683689117432,
+ "step": 2139
+ },
+ {
+ "epoch": 29.726637554585153,
+ "grad_norm": 0.018140794709324837,
+ "learning_rate": 0.0006,
+ "loss": 4.311944961547852,
+ "step": 2140
+ },
+ {
+ "epoch": 29.74061135371179,
+ "grad_norm": 0.016227100044488907,
+ "learning_rate": 0.0006,
+ "loss": 4.230257511138916,
+ "step": 2141
+ },
+ {
+ "epoch": 29.754585152838427,
+ "grad_norm": 0.015759488567709923,
+ "learning_rate": 0.0006,
+ "loss": 4.119174003601074,
+ "step": 2142
+ },
+ {
+ "epoch": 29.768558951965066,
+ "grad_norm": 0.01757766678929329,
+ "learning_rate": 0.0006,
+ "loss": 4.165826797485352,
+ "step": 2143
+ },
+ {
+ "epoch": 29.782532751091704,
+ "grad_norm": 0.019661923870444298,
+ "learning_rate": 0.0006,
+ "loss": 4.392203330993652,
+ "step": 2144
+ },
+ {
+ "epoch": 29.79650655021834,
+ "grad_norm": 0.019478755071759224,
+ "learning_rate": 0.0006,
+ "loss": 4.348371982574463,
+ "step": 2145
+ },
+ {
+ "epoch": 29.810480349344978,
+ "grad_norm": 0.019149569794535637,
+ "learning_rate": 0.0006,
+ "loss": 4.34104061126709,
+ "step": 2146
+ },
+ {
+ "epoch": 29.824454148471617,
+ "grad_norm": 0.017707521095871925,
+ "learning_rate": 0.0006,
+ "loss": 4.228226661682129,
+ "step": 2147
+ },
+ {
+ "epoch": 29.83842794759825,
+ "grad_norm": 0.015615378506481647,
+ "learning_rate": 0.0006,
+ "loss": 4.175541877746582,
+ "step": 2148
+ },
+ {
+ "epoch": 29.85240174672489,
+ "grad_norm": 0.016111129894852638,
+ "learning_rate": 0.0006,
+ "loss": 4.267936706542969,
+ "step": 2149
+ },
+ {
+ "epoch": 29.86637554585153,
+ "grad_norm": 0.01528779324144125,
+ "learning_rate": 0.0006,
+ "loss": 4.186014175415039,
+ "step": 2150
+ },
+ {
+ "epoch": 29.880349344978168,
+ "grad_norm": 0.014925646595656872,
+ "learning_rate": 0.0006,
+ "loss": 4.174860000610352,
+ "step": 2151
+ },
+ {
+ "epoch": 29.894323144104803,
+ "grad_norm": 0.016917673870921135,
+ "learning_rate": 0.0006,
+ "loss": 4.33908748626709,
+ "step": 2152
+ },
+ {
+ "epoch": 29.90829694323144,
+ "grad_norm": 0.016007075086236,
+ "learning_rate": 0.0006,
+ "loss": 4.266141414642334,
+ "step": 2153
+ },
+ {
+ "epoch": 29.92227074235808,
+ "grad_norm": 0.016178956255316734,
+ "learning_rate": 0.0006,
+ "loss": 4.271215915679932,
+ "step": 2154
+ },
+ {
+ "epoch": 29.936244541484715,
+ "grad_norm": 0.016180871054530144,
+ "learning_rate": 0.0006,
+ "loss": 4.278801441192627,
+ "step": 2155
+ },
+ {
+ "epoch": 29.950218340611354,
+ "grad_norm": 0.015335503034293652,
+ "learning_rate": 0.0006,
+ "loss": 4.194978713989258,
+ "step": 2156
+ },
+ {
+ "epoch": 29.964192139737992,
+ "grad_norm": 0.016580764204263687,
+ "learning_rate": 0.0006,
+ "loss": 4.290809631347656,
+ "step": 2157
+ },
+ {
+ "epoch": 29.978165938864628,
+ "grad_norm": 0.01623128168284893,
+ "learning_rate": 0.0006,
+ "loss": 4.205451488494873,
+ "step": 2158
+ },
+ {
+ "epoch": 29.992139737991266,
+ "grad_norm": 0.014904248528182507,
+ "learning_rate": 0.0006,
+ "loss": 4.24924373626709,
+ "step": 2159
+ },
+ {
+ "epoch": 30.0,
+ "grad_norm": 0.017089299857616425,
+ "learning_rate": 0.0006,
+ "loss": 4.222072601318359,
+ "step": 2160
+ },
+ {
+ "epoch": 30.0,
+ "eval_loss": 4.585411071777344,
+ "eval_runtime": 57.1164,
+ "eval_samples_per_second": 42.755,
+ "eval_steps_per_second": 1.348,
+ "step": 2160
+ },
+ {
+ "epoch": 30.01397379912664,
+ "grad_norm": 0.017226383090019226,
+ "learning_rate": 0.0006,
+ "loss": 4.19467830657959,
+ "step": 2161
+ },
+ {
+ "epoch": 30.027947598253274,
+ "grad_norm": 0.01635066606104374,
+ "learning_rate": 0.0006,
+ "loss": 4.176875591278076,
+ "step": 2162
+ },
+ {
+ "epoch": 30.041921397379912,
+ "grad_norm": 0.01735362410545349,
+ "learning_rate": 0.0006,
+ "loss": 4.13667106628418,
+ "step": 2163
+ },
+ {
+ "epoch": 30.05589519650655,
+ "grad_norm": 0.017047109082341194,
+ "learning_rate": 0.0006,
+ "loss": 4.136329650878906,
+ "step": 2164
+ },
+ {
+ "epoch": 30.069868995633186,
+ "grad_norm": 0.018116053193807602,
+ "learning_rate": 0.0006,
+ "loss": 4.228706359863281,
+ "step": 2165
+ },
+ {
+ "epoch": 30.083842794759825,
+ "grad_norm": 0.01791389286518097,
+ "learning_rate": 0.0006,
+ "loss": 4.1597747802734375,
+ "step": 2166
+ },
+ {
+ "epoch": 30.097816593886463,
+ "grad_norm": 0.017793502658605576,
+ "learning_rate": 0.0006,
+ "loss": 4.0807976722717285,
+ "step": 2167
+ },
+ {
+ "epoch": 30.111790393013102,
+ "grad_norm": 0.01827503927052021,
+ "learning_rate": 0.0006,
+ "loss": 4.170646667480469,
+ "step": 2168
+ },
+ {
+ "epoch": 30.125764192139737,
+ "grad_norm": 0.017228564247488976,
+ "learning_rate": 0.0006,
+ "loss": 4.146934509277344,
+ "step": 2169
+ },
+ {
+ "epoch": 30.139737991266376,
+ "grad_norm": 0.01757437363266945,
+ "learning_rate": 0.0006,
+ "loss": 4.1859130859375,
+ "step": 2170
+ },
+ {
+ "epoch": 30.153711790393015,
+ "grad_norm": 0.01701093092560768,
+ "learning_rate": 0.0006,
+ "loss": 4.126981258392334,
+ "step": 2171
+ },
+ {
+ "epoch": 30.16768558951965,
+ "grad_norm": 0.015220489352941513,
+ "learning_rate": 0.0006,
+ "loss": 4.164018630981445,
+ "step": 2172
+ },
+ {
+ "epoch": 30.18165938864629,
+ "grad_norm": 0.016254238784313202,
+ "learning_rate": 0.0006,
+ "loss": 4.258275508880615,
+ "step": 2173
+ },
+ {
+ "epoch": 30.195633187772927,
+ "grad_norm": 0.016609614714980125,
+ "learning_rate": 0.0006,
+ "loss": 4.222233772277832,
+ "step": 2174
+ },
+ {
+ "epoch": 30.209606986899562,
+ "grad_norm": 0.016531318426132202,
+ "learning_rate": 0.0006,
+ "loss": 4.144489288330078,
+ "step": 2175
+ },
+ {
+ "epoch": 30.2235807860262,
+ "grad_norm": 0.01567668654024601,
+ "learning_rate": 0.0006,
+ "loss": 4.181916236877441,
+ "step": 2176
+ },
+ {
+ "epoch": 30.23755458515284,
+ "grad_norm": 0.017195681110024452,
+ "learning_rate": 0.0006,
+ "loss": 4.189693927764893,
+ "step": 2177
+ },
+ {
+ "epoch": 30.251528384279474,
+ "grad_norm": 0.016882948577404022,
+ "learning_rate": 0.0006,
+ "loss": 4.21604061126709,
+ "step": 2178
+ },
+ {
+ "epoch": 30.265502183406113,
+ "grad_norm": 0.016261622309684753,
+ "learning_rate": 0.0006,
+ "loss": 4.227639198303223,
+ "step": 2179
+ },
+ {
+ "epoch": 30.27947598253275,
+ "grad_norm": 0.016411812976002693,
+ "learning_rate": 0.0006,
+ "loss": 4.130362510681152,
+ "step": 2180
+ },
+ {
+ "epoch": 30.29344978165939,
+ "grad_norm": 0.01732019893825054,
+ "learning_rate": 0.0006,
+ "loss": 4.230443000793457,
+ "step": 2181
+ },
+ {
+ "epoch": 30.307423580786025,
+ "grad_norm": 0.01825905591249466,
+ "learning_rate": 0.0006,
+ "loss": 4.175817489624023,
+ "step": 2182
+ },
+ {
+ "epoch": 30.321397379912664,
+ "grad_norm": 0.01887713558971882,
+ "learning_rate": 0.0006,
+ "loss": 4.169020652770996,
+ "step": 2183
+ },
+ {
+ "epoch": 30.335371179039303,
+ "grad_norm": 0.019391661509871483,
+ "learning_rate": 0.0006,
+ "loss": 4.176318645477295,
+ "step": 2184
+ },
+ {
+ "epoch": 30.349344978165938,
+ "grad_norm": 0.020840002223849297,
+ "learning_rate": 0.0006,
+ "loss": 4.174228668212891,
+ "step": 2185
+ },
+ {
+ "epoch": 30.363318777292577,
+ "grad_norm": 0.020193178206682205,
+ "learning_rate": 0.0006,
+ "loss": 4.127919673919678,
+ "step": 2186
+ },
+ {
+ "epoch": 30.377292576419215,
+ "grad_norm": 0.018449561670422554,
+ "learning_rate": 0.0006,
+ "loss": 4.229487895965576,
+ "step": 2187
+ },
+ {
+ "epoch": 30.39126637554585,
+ "grad_norm": 0.01711983233690262,
+ "learning_rate": 0.0006,
+ "loss": 4.201651573181152,
+ "step": 2188
+ },
+ {
+ "epoch": 30.40524017467249,
+ "grad_norm": 0.017293395474553108,
+ "learning_rate": 0.0006,
+ "loss": 4.168704032897949,
+ "step": 2189
+ },
+ {
+ "epoch": 30.419213973799128,
+ "grad_norm": 0.018526358529925346,
+ "learning_rate": 0.0006,
+ "loss": 4.171864032745361,
+ "step": 2190
+ },
+ {
+ "epoch": 30.433187772925763,
+ "grad_norm": 0.019287291914224625,
+ "learning_rate": 0.0006,
+ "loss": 4.039534568786621,
+ "step": 2191
+ },
+ {
+ "epoch": 30.4471615720524,
+ "grad_norm": 0.016902444884181023,
+ "learning_rate": 0.0006,
+ "loss": 4.282787799835205,
+ "step": 2192
+ },
+ {
+ "epoch": 30.46113537117904,
+ "grad_norm": 0.01656750962138176,
+ "learning_rate": 0.0006,
+ "loss": 4.213098526000977,
+ "step": 2193
+ },
+ {
+ "epoch": 30.475109170305675,
+ "grad_norm": 0.016805065795779228,
+ "learning_rate": 0.0006,
+ "loss": 4.1583476066589355,
+ "step": 2194
+ },
+ {
+ "epoch": 30.489082969432314,
+ "grad_norm": 0.017166638746857643,
+ "learning_rate": 0.0006,
+ "loss": 4.188634395599365,
+ "step": 2195
+ },
+ {
+ "epoch": 30.503056768558952,
+ "grad_norm": 0.01592106744647026,
+ "learning_rate": 0.0006,
+ "loss": 4.234043121337891,
+ "step": 2196
+ },
+ {
+ "epoch": 30.51703056768559,
+ "grad_norm": 0.015270989388227463,
+ "learning_rate": 0.0006,
+ "loss": 4.301420211791992,
+ "step": 2197
+ },
+ {
+ "epoch": 30.531004366812226,
+ "grad_norm": 0.015194724313914776,
+ "learning_rate": 0.0006,
+ "loss": 4.18703031539917,
+ "step": 2198
+ },
+ {
+ "epoch": 30.544978165938865,
+ "grad_norm": 0.015587746165692806,
+ "learning_rate": 0.0006,
+ "loss": 4.138928413391113,
+ "step": 2199
+ },
+ {
+ "epoch": 30.558951965065503,
+ "grad_norm": 0.016029570251703262,
+ "learning_rate": 0.0006,
+ "loss": 4.290863990783691,
+ "step": 2200
+ },
+ {
+ "epoch": 30.57292576419214,
+ "grad_norm": 0.015651309862732887,
+ "learning_rate": 0.0006,
+ "loss": 4.272062301635742,
+ "step": 2201
+ },
+ {
+ "epoch": 30.586899563318777,
+ "grad_norm": 0.015992306172847748,
+ "learning_rate": 0.0006,
+ "loss": 4.2487406730651855,
+ "step": 2202
+ },
+ {
+ "epoch": 30.600873362445416,
+ "grad_norm": 0.014969157986342907,
+ "learning_rate": 0.0006,
+ "loss": 4.115347862243652,
+ "step": 2203
+ },
+ {
+ "epoch": 30.61484716157205,
+ "grad_norm": 0.014639027416706085,
+ "learning_rate": 0.0006,
+ "loss": 4.2207489013671875,
+ "step": 2204
+ },
+ {
+ "epoch": 30.62882096069869,
+ "grad_norm": 0.015059413388371468,
+ "learning_rate": 0.0006,
+ "loss": 4.274044990539551,
+ "step": 2205
+ },
+ {
+ "epoch": 30.64279475982533,
+ "grad_norm": 0.0166641678661108,
+ "learning_rate": 0.0006,
+ "loss": 4.191373348236084,
+ "step": 2206
+ },
+ {
+ "epoch": 30.656768558951963,
+ "grad_norm": 0.017806345596909523,
+ "learning_rate": 0.0006,
+ "loss": 4.139800548553467,
+ "step": 2207
+ },
+ {
+ "epoch": 30.670742358078602,
+ "grad_norm": 0.018712421879172325,
+ "learning_rate": 0.0006,
+ "loss": 4.035965442657471,
+ "step": 2208
+ },
+ {
+ "epoch": 30.68471615720524,
+ "grad_norm": 0.018457243219017982,
+ "learning_rate": 0.0006,
+ "loss": 4.2579240798950195,
+ "step": 2209
+ },
+ {
+ "epoch": 30.69868995633188,
+ "grad_norm": 0.01920999586582184,
+ "learning_rate": 0.0006,
+ "loss": 4.131112098693848,
+ "step": 2210
+ },
+ {
+ "epoch": 30.712663755458514,
+ "grad_norm": 0.01729944534599781,
+ "learning_rate": 0.0006,
+ "loss": 4.254059314727783,
+ "step": 2211
+ },
+ {
+ "epoch": 30.726637554585153,
+ "grad_norm": 0.016914231702685356,
+ "learning_rate": 0.0006,
+ "loss": 4.233428955078125,
+ "step": 2212
+ },
+ {
+ "epoch": 30.74061135371179,
+ "grad_norm": 0.01913098618388176,
+ "learning_rate": 0.0006,
+ "loss": 4.23525333404541,
+ "step": 2213
+ },
+ {
+ "epoch": 30.754585152838427,
+ "grad_norm": 0.019150281324982643,
+ "learning_rate": 0.0006,
+ "loss": 4.213695049285889,
+ "step": 2214
+ },
+ {
+ "epoch": 30.768558951965066,
+ "grad_norm": 0.018947452306747437,
+ "learning_rate": 0.0006,
+ "loss": 4.151697158813477,
+ "step": 2215
+ },
+ {
+ "epoch": 30.782532751091704,
+ "grad_norm": 0.016058508306741714,
+ "learning_rate": 0.0006,
+ "loss": 4.220280170440674,
+ "step": 2216
+ },
+ {
+ "epoch": 30.79650655021834,
+ "grad_norm": 0.017488619312644005,
+ "learning_rate": 0.0006,
+ "loss": 4.154428482055664,
+ "step": 2217
+ },
+ {
+ "epoch": 30.810480349344978,
+ "grad_norm": 0.017355090007185936,
+ "learning_rate": 0.0006,
+ "loss": 4.256086349487305,
+ "step": 2218
+ },
+ {
+ "epoch": 30.824454148471617,
+ "grad_norm": 0.016636032611131668,
+ "learning_rate": 0.0006,
+ "loss": 4.227564334869385,
+ "step": 2219
+ },
+ {
+ "epoch": 30.83842794759825,
+ "grad_norm": 0.016273891553282738,
+ "learning_rate": 0.0006,
+ "loss": 4.200466156005859,
+ "step": 2220
+ },
+ {
+ "epoch": 30.85240174672489,
+ "grad_norm": 0.016939911991357803,
+ "learning_rate": 0.0006,
+ "loss": 4.1571221351623535,
+ "step": 2221
+ },
+ {
+ "epoch": 30.86637554585153,
+ "grad_norm": 0.018332522362470627,
+ "learning_rate": 0.0006,
+ "loss": 4.228398323059082,
+ "step": 2222
+ },
+ {
+ "epoch": 30.880349344978168,
+ "grad_norm": 0.019508114084601402,
+ "learning_rate": 0.0006,
+ "loss": 4.222960472106934,
+ "step": 2223
+ },
+ {
+ "epoch": 30.894323144104803,
+ "grad_norm": 0.018440278246998787,
+ "learning_rate": 0.0006,
+ "loss": 4.261051177978516,
+ "step": 2224
+ },
+ {
+ "epoch": 30.90829694323144,
+ "grad_norm": 0.018113229423761368,
+ "learning_rate": 0.0006,
+ "loss": 4.109235763549805,
+ "step": 2225
+ },
+ {
+ "epoch": 30.92227074235808,
+ "grad_norm": 0.015392083674669266,
+ "learning_rate": 0.0006,
+ "loss": 4.21950101852417,
+ "step": 2226
+ },
+ {
+ "epoch": 30.936244541484715,
+ "grad_norm": 0.01633290946483612,
+ "learning_rate": 0.0006,
+ "loss": 4.176053047180176,
+ "step": 2227
+ },
+ {
+ "epoch": 30.950218340611354,
+ "grad_norm": 0.016236383467912674,
+ "learning_rate": 0.0006,
+ "loss": 4.314309120178223,
+ "step": 2228
+ },
+ {
+ "epoch": 30.964192139737992,
+ "grad_norm": 0.015284373424947262,
+ "learning_rate": 0.0006,
+ "loss": 4.303133487701416,
+ "step": 2229
+ },
+ {
+ "epoch": 30.978165938864628,
+ "grad_norm": 0.015982897952198982,
+ "learning_rate": 0.0006,
+ "loss": 4.25739860534668,
+ "step": 2230
+ },
+ {
+ "epoch": 30.992139737991266,
+ "grad_norm": 0.016682198271155357,
+ "learning_rate": 0.0006,
+ "loss": 4.292210578918457,
+ "step": 2231
+ },
+ {
+ "epoch": 31.0,
+ "grad_norm": 0.017521467059850693,
+ "learning_rate": 0.0006,
+ "loss": 4.170637130737305,
+ "step": 2232
+ },
+ {
+ "epoch": 31.0,
+ "eval_loss": 4.642786979675293,
+ "eval_runtime": 57.029,
+ "eval_samples_per_second": 42.82,
+ "eval_steps_per_second": 1.35,
+ "step": 2232
+ },
+ {
+ "epoch": 31.01397379912664,
+ "grad_norm": 0.016607852652668953,
+ "learning_rate": 0.0006,
+ "loss": 4.106224536895752,
+ "step": 2233
+ },
+ {
+ "epoch": 31.027947598253274,
+ "grad_norm": 0.018448904156684875,
+ "learning_rate": 0.0006,
+ "loss": 4.224459648132324,
+ "step": 2234
+ },
+ {
+ "epoch": 31.041921397379912,
+ "grad_norm": 0.020481228828430176,
+ "learning_rate": 0.0006,
+ "loss": 4.305548667907715,
+ "step": 2235
+ },
+ {
+ "epoch": 31.05589519650655,
+ "grad_norm": 0.01999707892537117,
+ "learning_rate": 0.0006,
+ "loss": 4.253663539886475,
+ "step": 2236
+ },
+ {
+ "epoch": 31.069868995633186,
+ "grad_norm": 0.020428303629159927,
+ "learning_rate": 0.0006,
+ "loss": 4.166954040527344,
+ "step": 2237
+ },
+ {
+ "epoch": 31.083842794759825,
+ "grad_norm": 0.022608213126659393,
+ "learning_rate": 0.0006,
+ "loss": 4.172819137573242,
+ "step": 2238
+ },
+ {
+ "epoch": 31.097816593886463,
+ "grad_norm": 0.024669520556926727,
+ "learning_rate": 0.0006,
+ "loss": 4.106586456298828,
+ "step": 2239
+ },
+ {
+ "epoch": 31.111790393013102,
+ "grad_norm": 0.02247351035475731,
+ "learning_rate": 0.0006,
+ "loss": 4.160271644592285,
+ "step": 2240
+ },
+ {
+ "epoch": 31.125764192139737,
+ "grad_norm": 0.0207810141146183,
+ "learning_rate": 0.0006,
+ "loss": 4.272491931915283,
+ "step": 2241
+ },
+ {
+ "epoch": 31.139737991266376,
+ "grad_norm": 0.021576499566435814,
+ "learning_rate": 0.0006,
+ "loss": 4.171474456787109,
+ "step": 2242
+ },
+ {
+ "epoch": 31.153711790393015,
+ "grad_norm": 0.02337106689810753,
+ "learning_rate": 0.0006,
+ "loss": 4.094712734222412,
+ "step": 2243
+ },
+ {
+ "epoch": 31.16768558951965,
+ "grad_norm": 0.02219875156879425,
+ "learning_rate": 0.0006,
+ "loss": 4.1800994873046875,
+ "step": 2244
+ },
+ {
+ "epoch": 31.18165938864629,
+ "grad_norm": 0.022680504247546196,
+ "learning_rate": 0.0006,
+ "loss": 4.2054362297058105,
+ "step": 2245
+ },
+ {
+ "epoch": 31.195633187772927,
+ "grad_norm": 0.0208734143525362,
+ "learning_rate": 0.0006,
+ "loss": 4.06068229675293,
+ "step": 2246
+ },
+ {
+ "epoch": 31.209606986899562,
+ "grad_norm": 0.01811066083610058,
+ "learning_rate": 0.0006,
+ "loss": 4.2465128898620605,
+ "step": 2247
+ },
+ {
+ "epoch": 31.2235807860262,
+ "grad_norm": 0.018571637570858,
+ "learning_rate": 0.0006,
+ "loss": 4.191955089569092,
+ "step": 2248
+ },
+ {
+ "epoch": 31.23755458515284,
+ "grad_norm": 0.018598807975649834,
+ "learning_rate": 0.0006,
+ "loss": 4.200872421264648,
+ "step": 2249
+ },
+ {
+ "epoch": 31.251528384279474,
+ "grad_norm": 0.017577625811100006,
+ "learning_rate": 0.0006,
+ "loss": 4.148700714111328,
+ "step": 2250
+ },
+ {
+ "epoch": 31.265502183406113,
+ "grad_norm": 0.01817980408668518,
+ "learning_rate": 0.0006,
+ "loss": 4.256978988647461,
+ "step": 2251
+ },
+ {
+ "epoch": 31.27947598253275,
+ "grad_norm": 0.017983486875891685,
+ "learning_rate": 0.0006,
+ "loss": 4.175005912780762,
+ "step": 2252
+ },
+ {
+ "epoch": 31.29344978165939,
+ "grad_norm": 0.016751961782574654,
+ "learning_rate": 0.0006,
+ "loss": 4.237518310546875,
+ "step": 2253
+ },
+ {
+ "epoch": 31.307423580786025,
+ "grad_norm": 0.016407202929258347,
+ "learning_rate": 0.0006,
+ "loss": 4.197269439697266,
+ "step": 2254
+ },
+ {
+ "epoch": 31.321397379912664,
+ "grad_norm": 0.01578645594418049,
+ "learning_rate": 0.0006,
+ "loss": 4.179751396179199,
+ "step": 2255
+ },
+ {
+ "epoch": 31.335371179039303,
+ "grad_norm": 0.016001226380467415,
+ "learning_rate": 0.0006,
+ "loss": 4.08821964263916,
+ "step": 2256
+ },
+ {
+ "epoch": 31.349344978165938,
+ "grad_norm": 0.015122811309993267,
+ "learning_rate": 0.0006,
+ "loss": 4.047828674316406,
+ "step": 2257
+ },
+ {
+ "epoch": 31.363318777292577,
+ "grad_norm": 0.015399965457618237,
+ "learning_rate": 0.0006,
+ "loss": 4.1506524085998535,
+ "step": 2258
+ },
+ {
+ "epoch": 31.377292576419215,
+ "grad_norm": 0.01534605398774147,
+ "learning_rate": 0.0006,
+ "loss": 4.27952766418457,
+ "step": 2259
+ },
+ {
+ "epoch": 31.39126637554585,
+ "grad_norm": 0.015461381524801254,
+ "learning_rate": 0.0006,
+ "loss": 4.27102518081665,
+ "step": 2260
+ },
+ {
+ "epoch": 31.40524017467249,
+ "grad_norm": 0.016769153997302055,
+ "learning_rate": 0.0006,
+ "loss": 4.191592216491699,
+ "step": 2261
+ },
+ {
+ "epoch": 31.419213973799128,
+ "grad_norm": 0.016557959839701653,
+ "learning_rate": 0.0006,
+ "loss": 4.159705638885498,
+ "step": 2262
+ },
+ {
+ "epoch": 31.433187772925763,
+ "grad_norm": 0.01659569889307022,
+ "learning_rate": 0.0006,
+ "loss": 4.2136664390563965,
+ "step": 2263
+ },
+ {
+ "epoch": 31.4471615720524,
+ "grad_norm": 0.017101937904953957,
+ "learning_rate": 0.0006,
+ "loss": 4.2716498374938965,
+ "step": 2264
+ },
+ {
+ "epoch": 31.46113537117904,
+ "grad_norm": 0.01722985878586769,
+ "learning_rate": 0.0006,
+ "loss": 4.300534248352051,
+ "step": 2265
+ },
+ {
+ "epoch": 31.475109170305675,
+ "grad_norm": 0.016800185665488243,
+ "learning_rate": 0.0006,
+ "loss": 4.113240718841553,
+ "step": 2266
+ },
+ {
+ "epoch": 31.489082969432314,
+ "grad_norm": 0.01603875495493412,
+ "learning_rate": 0.0006,
+ "loss": 4.157105445861816,
+ "step": 2267
+ },
+ {
+ "epoch": 31.503056768558952,
+ "grad_norm": 0.015903932973742485,
+ "learning_rate": 0.0006,
+ "loss": 4.233669757843018,
+ "step": 2268
+ },
+ {
+ "epoch": 31.51703056768559,
+ "grad_norm": 0.0160983856767416,
+ "learning_rate": 0.0006,
+ "loss": 4.147704124450684,
+ "step": 2269
+ },
+ {
+ "epoch": 31.531004366812226,
+ "grad_norm": 0.01604730263352394,
+ "learning_rate": 0.0006,
+ "loss": 4.1465911865234375,
+ "step": 2270
+ },
+ {
+ "epoch": 31.544978165938865,
+ "grad_norm": 0.016205785796046257,
+ "learning_rate": 0.0006,
+ "loss": 4.106704235076904,
+ "step": 2271
+ },
+ {
+ "epoch": 31.558951965065503,
+ "grad_norm": 0.017008036375045776,
+ "learning_rate": 0.0006,
+ "loss": 4.191000938415527,
+ "step": 2272
+ },
+ {
+ "epoch": 31.57292576419214,
+ "grad_norm": 0.01699635200202465,
+ "learning_rate": 0.0006,
+ "loss": 4.187042236328125,
+ "step": 2273
+ },
+ {
+ "epoch": 31.586899563318777,
+ "grad_norm": 0.01692967116832733,
+ "learning_rate": 0.0006,
+ "loss": 4.208805561065674,
+ "step": 2274
+ },
+ {
+ "epoch": 31.600873362445416,
+ "grad_norm": 0.017608756199479103,
+ "learning_rate": 0.0006,
+ "loss": 4.200512886047363,
+ "step": 2275
+ },
+ {
+ "epoch": 31.61484716157205,
+ "grad_norm": 0.016590990126132965,
+ "learning_rate": 0.0006,
+ "loss": 4.194705009460449,
+ "step": 2276
+ },
+ {
+ "epoch": 31.62882096069869,
+ "grad_norm": 0.0170235987752676,
+ "learning_rate": 0.0006,
+ "loss": 4.210326194763184,
+ "step": 2277
+ },
+ {
+ "epoch": 31.64279475982533,
+ "grad_norm": 0.01604801043868065,
+ "learning_rate": 0.0006,
+ "loss": 4.196024417877197,
+ "step": 2278
+ },
+ {
+ "epoch": 31.656768558951963,
+ "grad_norm": 0.015694202855229378,
+ "learning_rate": 0.0006,
+ "loss": 4.178236484527588,
+ "step": 2279
+ },
+ {
+ "epoch": 31.670742358078602,
+ "grad_norm": 0.015473959036171436,
+ "learning_rate": 0.0006,
+ "loss": 4.259339809417725,
+ "step": 2280
+ },
+ {
+ "epoch": 31.68471615720524,
+ "grad_norm": 0.014689362607896328,
+ "learning_rate": 0.0006,
+ "loss": 3.9702324867248535,
+ "step": 2281
+ },
+ {
+ "epoch": 31.69868995633188,
+ "grad_norm": 0.015245123766362667,
+ "learning_rate": 0.0006,
+ "loss": 4.197025775909424,
+ "step": 2282
+ },
+ {
+ "epoch": 31.712663755458514,
+ "grad_norm": 0.016148347407579422,
+ "learning_rate": 0.0006,
+ "loss": 4.231856346130371,
+ "step": 2283
+ },
+ {
+ "epoch": 31.726637554585153,
+ "grad_norm": 0.015169818885624409,
+ "learning_rate": 0.0006,
+ "loss": 4.230169296264648,
+ "step": 2284
+ },
+ {
+ "epoch": 31.74061135371179,
+ "grad_norm": 0.014843013137578964,
+ "learning_rate": 0.0006,
+ "loss": 4.195633411407471,
+ "step": 2285
+ },
+ {
+ "epoch": 31.754585152838427,
+ "grad_norm": 0.015229340642690659,
+ "learning_rate": 0.0006,
+ "loss": 4.231280326843262,
+ "step": 2286
+ },
+ {
+ "epoch": 31.768558951965066,
+ "grad_norm": 0.01551708485931158,
+ "learning_rate": 0.0006,
+ "loss": 4.17064094543457,
+ "step": 2287
+ },
+ {
+ "epoch": 31.782532751091704,
+ "grad_norm": 0.015863575041294098,
+ "learning_rate": 0.0006,
+ "loss": 4.1579484939575195,
+ "step": 2288
+ },
+ {
+ "epoch": 31.79650655021834,
+ "grad_norm": 0.015118198469281197,
+ "learning_rate": 0.0006,
+ "loss": 4.1626787185668945,
+ "step": 2289
+ },
+ {
+ "epoch": 31.810480349344978,
+ "grad_norm": 0.015220776200294495,
+ "learning_rate": 0.0006,
+ "loss": 4.135746955871582,
+ "step": 2290
+ },
+ {
+ "epoch": 31.824454148471617,
+ "grad_norm": 0.014529784210026264,
+ "learning_rate": 0.0006,
+ "loss": 4.164931297302246,
+ "step": 2291
+ },
+ {
+ "epoch": 31.83842794759825,
+ "grad_norm": 0.015533296391367912,
+ "learning_rate": 0.0006,
+ "loss": 4.112464904785156,
+ "step": 2292
+ },
+ {
+ "epoch": 31.85240174672489,
+ "grad_norm": 0.016562335193157196,
+ "learning_rate": 0.0006,
+ "loss": 4.164898872375488,
+ "step": 2293
+ },
+ {
+ "epoch": 31.86637554585153,
+ "grad_norm": 0.016258113086223602,
+ "learning_rate": 0.0006,
+ "loss": 4.1979265213012695,
+ "step": 2294
+ },
+ {
+ "epoch": 31.880349344978168,
+ "grad_norm": 0.01558110024780035,
+ "learning_rate": 0.0006,
+ "loss": 4.174675941467285,
+ "step": 2295
+ },
+ {
+ "epoch": 31.894323144104803,
+ "grad_norm": 0.016412515193223953,
+ "learning_rate": 0.0006,
+ "loss": 4.217909336090088,
+ "step": 2296
+ },
+ {
+ "epoch": 31.90829694323144,
+ "grad_norm": 0.015118442475795746,
+ "learning_rate": 0.0006,
+ "loss": 4.137143611907959,
+ "step": 2297
+ },
+ {
+ "epoch": 31.92227074235808,
+ "grad_norm": 0.01506161317229271,
+ "learning_rate": 0.0006,
+ "loss": 4.196122169494629,
+ "step": 2298
+ },
+ {
+ "epoch": 31.936244541484715,
+ "grad_norm": 0.017165271565318108,
+ "learning_rate": 0.0006,
+ "loss": 4.14834451675415,
+ "step": 2299
+ },
+ {
+ "epoch": 31.950218340611354,
+ "grad_norm": 0.01696440577507019,
+ "learning_rate": 0.0006,
+ "loss": 4.218868732452393,
+ "step": 2300
+ },
+ {
+ "epoch": 31.964192139737992,
+ "grad_norm": 0.017349394038319588,
+ "learning_rate": 0.0006,
+ "loss": 4.2754011154174805,
+ "step": 2301
+ },
+ {
+ "epoch": 31.978165938864628,
+ "grad_norm": 0.018387921154499054,
+ "learning_rate": 0.0006,
+ "loss": 4.282623291015625,
+ "step": 2302
+ },
+ {
+ "epoch": 31.992139737991266,
+ "grad_norm": 0.018837926909327507,
+ "learning_rate": 0.0006,
+ "loss": 4.185425281524658,
+ "step": 2303
+ },
+ {
+ "epoch": 32.0,
+ "grad_norm": 0.018998464569449425,
+ "learning_rate": 0.0006,
+ "loss": 4.151291847229004,
+ "step": 2304
+ },
+ {
+ "epoch": 32.0,
+ "eval_loss": 4.522614002227783,
+ "eval_runtime": 57.0106,
+ "eval_samples_per_second": 42.834,
+ "eval_steps_per_second": 1.351,
+ "step": 2304
+ },
+ {
+ "epoch": 32.01397379912664,
+ "grad_norm": 0.019036833196878433,
+ "learning_rate": 0.0006,
+ "loss": 4.183202266693115,
+ "step": 2305
+ },
+ {
+ "epoch": 32.02794759825328,
+ "grad_norm": 0.02207397110760212,
+ "learning_rate": 0.0006,
+ "loss": 4.17875862121582,
+ "step": 2306
+ },
+ {
+ "epoch": 32.041921397379916,
+ "grad_norm": 0.022008279338479042,
+ "learning_rate": 0.0006,
+ "loss": 4.08042049407959,
+ "step": 2307
+ },
+ {
+ "epoch": 32.05589519650655,
+ "grad_norm": 0.022407017648220062,
+ "learning_rate": 0.0006,
+ "loss": 4.112189769744873,
+ "step": 2308
+ },
+ {
+ "epoch": 32.069868995633186,
+ "grad_norm": 0.021599747240543365,
+ "learning_rate": 0.0006,
+ "loss": 4.126455307006836,
+ "step": 2309
+ },
+ {
+ "epoch": 32.083842794759825,
+ "grad_norm": 0.020340966060757637,
+ "learning_rate": 0.0006,
+ "loss": 4.080538272857666,
+ "step": 2310
+ },
+ {
+ "epoch": 32.09781659388646,
+ "grad_norm": 0.020956892520189285,
+ "learning_rate": 0.0006,
+ "loss": 4.171943664550781,
+ "step": 2311
+ },
+ {
+ "epoch": 32.1117903930131,
+ "grad_norm": 0.019966667518019676,
+ "learning_rate": 0.0006,
+ "loss": 4.144007682800293,
+ "step": 2312
+ },
+ {
+ "epoch": 32.12576419213974,
+ "grad_norm": 0.018256952986121178,
+ "learning_rate": 0.0006,
+ "loss": 4.208107948303223,
+ "step": 2313
+ },
+ {
+ "epoch": 32.13973799126637,
+ "grad_norm": 0.017289170995354652,
+ "learning_rate": 0.0006,
+ "loss": 4.182024955749512,
+ "step": 2314
+ },
+ {
+ "epoch": 32.15371179039301,
+ "grad_norm": 0.017656700685620308,
+ "learning_rate": 0.0006,
+ "loss": 4.068373680114746,
+ "step": 2315
+ },
+ {
+ "epoch": 32.16768558951965,
+ "grad_norm": 0.017903361469507217,
+ "learning_rate": 0.0006,
+ "loss": 4.2345428466796875,
+ "step": 2316
+ },
+ {
+ "epoch": 32.18165938864629,
+ "grad_norm": 0.01785109005868435,
+ "learning_rate": 0.0006,
+ "loss": 4.219776153564453,
+ "step": 2317
+ },
+ {
+ "epoch": 32.19563318777293,
+ "grad_norm": 0.01887257769703865,
+ "learning_rate": 0.0006,
+ "loss": 4.139206409454346,
+ "step": 2318
+ },
+ {
+ "epoch": 32.209606986899566,
+ "grad_norm": 0.01924685761332512,
+ "learning_rate": 0.0006,
+ "loss": 4.121052265167236,
+ "step": 2319
+ },
+ {
+ "epoch": 32.223580786026204,
+ "grad_norm": 0.020574072375893593,
+ "learning_rate": 0.0006,
+ "loss": 4.13262414932251,
+ "step": 2320
+ },
+ {
+ "epoch": 32.237554585152836,
+ "grad_norm": 0.023849064484238625,
+ "learning_rate": 0.0006,
+ "loss": 4.239119529724121,
+ "step": 2321
+ },
+ {
+ "epoch": 32.251528384279474,
+ "grad_norm": 0.026239361613988876,
+ "learning_rate": 0.0006,
+ "loss": 4.248993873596191,
+ "step": 2322
+ },
+ {
+ "epoch": 32.26550218340611,
+ "grad_norm": 0.023980243131518364,
+ "learning_rate": 0.0006,
+ "loss": 4.159852504730225,
+ "step": 2323
+ },
+ {
+ "epoch": 32.27947598253275,
+ "grad_norm": 0.020442405715584755,
+ "learning_rate": 0.0006,
+ "loss": 4.133546829223633,
+ "step": 2324
+ },
+ {
+ "epoch": 32.29344978165939,
+ "grad_norm": 0.020661186426877975,
+ "learning_rate": 0.0006,
+ "loss": 4.078243732452393,
+ "step": 2325
+ },
+ {
+ "epoch": 32.30742358078603,
+ "grad_norm": 0.02046387456357479,
+ "learning_rate": 0.0006,
+ "loss": 4.172012805938721,
+ "step": 2326
+ },
+ {
+ "epoch": 32.32139737991266,
+ "grad_norm": 0.02089597098529339,
+ "learning_rate": 0.0006,
+ "loss": 4.0604705810546875,
+ "step": 2327
+ },
+ {
+ "epoch": 32.3353711790393,
+ "grad_norm": 0.01875591278076172,
+ "learning_rate": 0.0006,
+ "loss": 4.125911235809326,
+ "step": 2328
+ },
+ {
+ "epoch": 32.34934497816594,
+ "grad_norm": 0.018606796860694885,
+ "learning_rate": 0.0006,
+ "loss": 4.097362518310547,
+ "step": 2329
+ },
+ {
+ "epoch": 32.36331877729258,
+ "grad_norm": 0.01766626164317131,
+ "learning_rate": 0.0006,
+ "loss": 4.118616580963135,
+ "step": 2330
+ },
+ {
+ "epoch": 32.377292576419215,
+ "grad_norm": 0.017692960798740387,
+ "learning_rate": 0.0006,
+ "loss": 4.1457672119140625,
+ "step": 2331
+ },
+ {
+ "epoch": 32.391266375545854,
+ "grad_norm": 0.017619704827666283,
+ "learning_rate": 0.0006,
+ "loss": 4.175961494445801,
+ "step": 2332
+ },
+ {
+ "epoch": 32.40524017467249,
+ "grad_norm": 0.01784413494169712,
+ "learning_rate": 0.0006,
+ "loss": 4.051291465759277,
+ "step": 2333
+ },
+ {
+ "epoch": 32.419213973799124,
+ "grad_norm": 0.018508171662688255,
+ "learning_rate": 0.0006,
+ "loss": 4.0638556480407715,
+ "step": 2334
+ },
+ {
+ "epoch": 32.43318777292576,
+ "grad_norm": 0.01731436885893345,
+ "learning_rate": 0.0006,
+ "loss": 4.1805419921875,
+ "step": 2335
+ },
+ {
+ "epoch": 32.4471615720524,
+ "grad_norm": 0.015886498615145683,
+ "learning_rate": 0.0006,
+ "loss": 4.151026725769043,
+ "step": 2336
+ },
+ {
+ "epoch": 32.46113537117904,
+ "grad_norm": 0.016583820804953575,
+ "learning_rate": 0.0006,
+ "loss": 4.203604698181152,
+ "step": 2337
+ },
+ {
+ "epoch": 32.47510917030568,
+ "grad_norm": 0.0176097322255373,
+ "learning_rate": 0.0006,
+ "loss": 4.135159015655518,
+ "step": 2338
+ },
+ {
+ "epoch": 32.48908296943232,
+ "grad_norm": 0.016657903790473938,
+ "learning_rate": 0.0006,
+ "loss": 4.126681804656982,
+ "step": 2339
+ },
+ {
+ "epoch": 32.50305676855895,
+ "grad_norm": 0.016866130754351616,
+ "learning_rate": 0.0006,
+ "loss": 4.162341117858887,
+ "step": 2340
+ },
+ {
+ "epoch": 32.51703056768559,
+ "grad_norm": 0.016286680474877357,
+ "learning_rate": 0.0006,
+ "loss": 4.261801242828369,
+ "step": 2341
+ },
+ {
+ "epoch": 32.531004366812226,
+ "grad_norm": 0.01481365505605936,
+ "learning_rate": 0.0006,
+ "loss": 4.110016822814941,
+ "step": 2342
+ },
+ {
+ "epoch": 32.544978165938865,
+ "grad_norm": 0.016605013981461525,
+ "learning_rate": 0.0006,
+ "loss": 4.230024814605713,
+ "step": 2343
+ },
+ {
+ "epoch": 32.5589519650655,
+ "grad_norm": 0.01645175740122795,
+ "learning_rate": 0.0006,
+ "loss": 4.174612045288086,
+ "step": 2344
+ },
+ {
+ "epoch": 32.57292576419214,
+ "grad_norm": 0.014725026674568653,
+ "learning_rate": 0.0006,
+ "loss": 4.066371440887451,
+ "step": 2345
+ },
+ {
+ "epoch": 32.58689956331878,
+ "grad_norm": 0.015003837645053864,
+ "learning_rate": 0.0006,
+ "loss": 4.193343639373779,
+ "step": 2346
+ },
+ {
+ "epoch": 32.60087336244541,
+ "grad_norm": 0.015346302650868893,
+ "learning_rate": 0.0006,
+ "loss": 4.210659980773926,
+ "step": 2347
+ },
+ {
+ "epoch": 32.61484716157205,
+ "grad_norm": 0.015135962516069412,
+ "learning_rate": 0.0006,
+ "loss": 4.204618453979492,
+ "step": 2348
+ },
+ {
+ "epoch": 32.62882096069869,
+ "grad_norm": 0.01601956970989704,
+ "learning_rate": 0.0006,
+ "loss": 4.145564556121826,
+ "step": 2349
+ },
+ {
+ "epoch": 32.64279475982533,
+ "grad_norm": 0.01517851185053587,
+ "learning_rate": 0.0006,
+ "loss": 4.051673889160156,
+ "step": 2350
+ },
+ {
+ "epoch": 32.65676855895197,
+ "grad_norm": 0.014894292689859867,
+ "learning_rate": 0.0006,
+ "loss": 4.184818267822266,
+ "step": 2351
+ },
+ {
+ "epoch": 32.670742358078606,
+ "grad_norm": 0.015720834955573082,
+ "learning_rate": 0.0006,
+ "loss": 4.220036506652832,
+ "step": 2352
+ },
+ {
+ "epoch": 32.68471615720524,
+ "grad_norm": 0.01573433168232441,
+ "learning_rate": 0.0006,
+ "loss": 4.076773166656494,
+ "step": 2353
+ },
+ {
+ "epoch": 32.698689956331876,
+ "grad_norm": 0.015828052535653114,
+ "learning_rate": 0.0006,
+ "loss": 4.116329669952393,
+ "step": 2354
+ },
+ {
+ "epoch": 32.712663755458514,
+ "grad_norm": 0.015462259761989117,
+ "learning_rate": 0.0006,
+ "loss": 4.1155242919921875,
+ "step": 2355
+ },
+ {
+ "epoch": 32.72663755458515,
+ "grad_norm": 0.015763849020004272,
+ "learning_rate": 0.0006,
+ "loss": 4.169687747955322,
+ "step": 2356
+ },
+ {
+ "epoch": 32.74061135371179,
+ "grad_norm": 0.014935145154595375,
+ "learning_rate": 0.0006,
+ "loss": 4.2090911865234375,
+ "step": 2357
+ },
+ {
+ "epoch": 32.75458515283843,
+ "grad_norm": 0.015119184739887714,
+ "learning_rate": 0.0006,
+ "loss": 4.131838798522949,
+ "step": 2358
+ },
+ {
+ "epoch": 32.76855895196506,
+ "grad_norm": 0.014833525754511356,
+ "learning_rate": 0.0006,
+ "loss": 4.17618465423584,
+ "step": 2359
+ },
+ {
+ "epoch": 32.7825327510917,
+ "grad_norm": 0.01588359661400318,
+ "learning_rate": 0.0006,
+ "loss": 4.163723945617676,
+ "step": 2360
+ },
+ {
+ "epoch": 32.79650655021834,
+ "grad_norm": 0.0168188214302063,
+ "learning_rate": 0.0006,
+ "loss": 4.178189277648926,
+ "step": 2361
+ },
+ {
+ "epoch": 32.81048034934498,
+ "grad_norm": 0.01726095750927925,
+ "learning_rate": 0.0006,
+ "loss": 4.144223690032959,
+ "step": 2362
+ },
+ {
+ "epoch": 32.82445414847162,
+ "grad_norm": 0.01701410301029682,
+ "learning_rate": 0.0006,
+ "loss": 4.209896087646484,
+ "step": 2363
+ },
+ {
+ "epoch": 32.838427947598255,
+ "grad_norm": 0.01682264916598797,
+ "learning_rate": 0.0006,
+ "loss": 4.26223087310791,
+ "step": 2364
+ },
+ {
+ "epoch": 32.852401746724894,
+ "grad_norm": 0.015597456134855747,
+ "learning_rate": 0.0006,
+ "loss": 4.292664051055908,
+ "step": 2365
+ },
+ {
+ "epoch": 32.866375545851525,
+ "grad_norm": 0.015258579514920712,
+ "learning_rate": 0.0006,
+ "loss": 4.165006637573242,
+ "step": 2366
+ },
+ {
+ "epoch": 32.880349344978164,
+ "grad_norm": 0.01662200503051281,
+ "learning_rate": 0.0006,
+ "loss": 4.309786319732666,
+ "step": 2367
+ },
+ {
+ "epoch": 32.8943231441048,
+ "grad_norm": 0.015691161155700684,
+ "learning_rate": 0.0006,
+ "loss": 4.167202949523926,
+ "step": 2368
+ },
+ {
+ "epoch": 32.90829694323144,
+ "grad_norm": 0.014870637096464634,
+ "learning_rate": 0.0006,
+ "loss": 4.142415523529053,
+ "step": 2369
+ },
+ {
+ "epoch": 32.92227074235808,
+ "grad_norm": 0.01567252166569233,
+ "learning_rate": 0.0006,
+ "loss": 4.2312092781066895,
+ "step": 2370
+ },
+ {
+ "epoch": 32.93624454148472,
+ "grad_norm": 0.014948616735637188,
+ "learning_rate": 0.0006,
+ "loss": 4.1514129638671875,
+ "step": 2371
+ },
+ {
+ "epoch": 32.95021834061135,
+ "grad_norm": 0.014654500409960747,
+ "learning_rate": 0.0006,
+ "loss": 4.205010890960693,
+ "step": 2372
+ },
+ {
+ "epoch": 32.96419213973799,
+ "grad_norm": 0.015059148892760277,
+ "learning_rate": 0.0006,
+ "loss": 4.148405075073242,
+ "step": 2373
+ },
+ {
+ "epoch": 32.97816593886463,
+ "grad_norm": 0.015026576817035675,
+ "learning_rate": 0.0006,
+ "loss": 4.182535171508789,
+ "step": 2374
+ },
+ {
+ "epoch": 32.992139737991266,
+ "grad_norm": 0.016394320875406265,
+ "learning_rate": 0.0006,
+ "loss": 4.129051208496094,
+ "step": 2375
+ },
+ {
+ "epoch": 33.0,
+ "grad_norm": 0.01867043413221836,
+ "learning_rate": 0.0006,
+ "loss": 4.185556411743164,
+ "step": 2376
+ },
+ {
+ "epoch": 33.0,
+ "eval_loss": 4.640405654907227,
+ "eval_runtime": 56.5861,
+ "eval_samples_per_second": 43.155,
+ "eval_steps_per_second": 1.361,
+ "step": 2376
+ },
+ {
+ "epoch": 33.01397379912664,
+ "grad_norm": 0.018626727163791656,
+ "learning_rate": 0.0006,
+ "loss": 4.143031597137451,
+ "step": 2377
+ },
+ {
+ "epoch": 33.02794759825328,
+ "grad_norm": 0.019283363595604897,
+ "learning_rate": 0.0006,
+ "loss": 4.117845058441162,
+ "step": 2378
+ },
+ {
+ "epoch": 33.041921397379916,
+ "grad_norm": 0.018185822293162346,
+ "learning_rate": 0.0006,
+ "loss": 4.076862812042236,
+ "step": 2379
+ },
+ {
+ "epoch": 33.05589519650655,
+ "grad_norm": 0.019411196932196617,
+ "learning_rate": 0.0006,
+ "loss": 4.124156475067139,
+ "step": 2380
+ },
+ {
+ "epoch": 33.069868995633186,
+ "grad_norm": 0.019248466938734055,
+ "learning_rate": 0.0006,
+ "loss": 4.201735019683838,
+ "step": 2381
+ },
+ {
+ "epoch": 33.083842794759825,
+ "grad_norm": 0.018856720998883247,
+ "learning_rate": 0.0006,
+ "loss": 4.00296688079834,
+ "step": 2382
+ },
+ {
+ "epoch": 33.09781659388646,
+ "grad_norm": 0.018554294481873512,
+ "learning_rate": 0.0006,
+ "loss": 4.270577430725098,
+ "step": 2383
+ },
+ {
+ "epoch": 33.1117903930131,
+ "grad_norm": 0.019946224987506866,
+ "learning_rate": 0.0006,
+ "loss": 4.119589328765869,
+ "step": 2384
+ },
+ {
+ "epoch": 33.12576419213974,
+ "grad_norm": 0.021395670250058174,
+ "learning_rate": 0.0006,
+ "loss": 4.128847122192383,
+ "step": 2385
+ },
+ {
+ "epoch": 33.13973799126637,
+ "grad_norm": 0.02009674906730652,
+ "learning_rate": 0.0006,
+ "loss": 4.183228969573975,
+ "step": 2386
+ },
+ {
+ "epoch": 33.15371179039301,
+ "grad_norm": 0.01820223033428192,
+ "learning_rate": 0.0006,
+ "loss": 4.131601810455322,
+ "step": 2387
+ },
+ {
+ "epoch": 33.16768558951965,
+ "grad_norm": 0.018653156235814095,
+ "learning_rate": 0.0006,
+ "loss": 4.116430282592773,
+ "step": 2388
+ },
+ {
+ "epoch": 33.18165938864629,
+ "grad_norm": 0.016828063875436783,
+ "learning_rate": 0.0006,
+ "loss": 4.054663181304932,
+ "step": 2389
+ },
+ {
+ "epoch": 33.19563318777293,
+ "grad_norm": 0.017025409266352654,
+ "learning_rate": 0.0006,
+ "loss": 4.110044956207275,
+ "step": 2390
+ },
+ {
+ "epoch": 33.209606986899566,
+ "grad_norm": 0.016608452424407005,
+ "learning_rate": 0.0006,
+ "loss": 4.102668285369873,
+ "step": 2391
+ },
+ {
+ "epoch": 33.223580786026204,
+ "grad_norm": 0.01743577979505062,
+ "learning_rate": 0.0006,
+ "loss": 4.166791915893555,
+ "step": 2392
+ },
+ {
+ "epoch": 33.237554585152836,
+ "grad_norm": 0.017253734171390533,
+ "learning_rate": 0.0006,
+ "loss": 4.217832088470459,
+ "step": 2393
+ },
+ {
+ "epoch": 33.251528384279474,
+ "grad_norm": 0.017937595024704933,
+ "learning_rate": 0.0006,
+ "loss": 4.2173967361450195,
+ "step": 2394
+ },
+ {
+ "epoch": 33.26550218340611,
+ "grad_norm": 0.01985993981361389,
+ "learning_rate": 0.0006,
+ "loss": 4.17849588394165,
+ "step": 2395
+ },
+ {
+ "epoch": 33.27947598253275,
+ "grad_norm": 0.019243355840444565,
+ "learning_rate": 0.0006,
+ "loss": 4.086523056030273,
+ "step": 2396
+ },
+ {
+ "epoch": 33.29344978165939,
+ "grad_norm": 0.01924416422843933,
+ "learning_rate": 0.0006,
+ "loss": 4.128574371337891,
+ "step": 2397
+ },
+ {
+ "epoch": 33.30742358078603,
+ "grad_norm": 0.01983758620917797,
+ "learning_rate": 0.0006,
+ "loss": 4.061581611633301,
+ "step": 2398
+ },
+ {
+ "epoch": 33.32139737991266,
+ "grad_norm": 0.01930462196469307,
+ "learning_rate": 0.0006,
+ "loss": 4.091731071472168,
+ "step": 2399
+ },
+ {
+ "epoch": 33.3353711790393,
+ "grad_norm": 0.017988424748182297,
+ "learning_rate": 0.0006,
+ "loss": 4.155599117279053,
+ "step": 2400
+ },
+ {
+ "epoch": 33.34934497816594,
+ "grad_norm": 0.015194767154753208,
+ "learning_rate": 0.0006,
+ "loss": 4.144134044647217,
+ "step": 2401
+ },
+ {
+ "epoch": 33.36331877729258,
+ "grad_norm": 0.01536081824451685,
+ "learning_rate": 0.0006,
+ "loss": 4.127272129058838,
+ "step": 2402
+ },
+ {
+ "epoch": 33.377292576419215,
+ "grad_norm": 0.015431297942996025,
+ "learning_rate": 0.0006,
+ "loss": 4.3213019371032715,
+ "step": 2403
+ },
+ {
+ "epoch": 33.391266375545854,
+ "grad_norm": 0.015740280970931053,
+ "learning_rate": 0.0006,
+ "loss": 4.121467590332031,
+ "step": 2404
+ },
+ {
+ "epoch": 33.40524017467249,
+ "grad_norm": 0.015379384160041809,
+ "learning_rate": 0.0006,
+ "loss": 4.095216751098633,
+ "step": 2405
+ },
+ {
+ "epoch": 33.419213973799124,
+ "grad_norm": 0.01528073474764824,
+ "learning_rate": 0.0006,
+ "loss": 4.13047981262207,
+ "step": 2406
+ },
+ {
+ "epoch": 33.43318777292576,
+ "grad_norm": 0.015739979222416878,
+ "learning_rate": 0.0006,
+ "loss": 4.2533135414123535,
+ "step": 2407
+ },
+ {
+ "epoch": 33.4471615720524,
+ "grad_norm": 0.01519764307886362,
+ "learning_rate": 0.0006,
+ "loss": 4.131782054901123,
+ "step": 2408
+ },
+ {
+ "epoch": 33.46113537117904,
+ "grad_norm": 0.015425996854901314,
+ "learning_rate": 0.0006,
+ "loss": 4.02811336517334,
+ "step": 2409
+ },
+ {
+ "epoch": 33.47510917030568,
+ "grad_norm": 0.015378830023109913,
+ "learning_rate": 0.0006,
+ "loss": 4.08652925491333,
+ "step": 2410
+ },
+ {
+ "epoch": 33.48908296943232,
+ "grad_norm": 0.01473075058311224,
+ "learning_rate": 0.0006,
+ "loss": 4.214364528656006,
+ "step": 2411
+ },
+ {
+ "epoch": 33.50305676855895,
+ "grad_norm": 0.015141637064516544,
+ "learning_rate": 0.0006,
+ "loss": 4.155712127685547,
+ "step": 2412
+ },
+ {
+ "epoch": 33.51703056768559,
+ "grad_norm": 0.01641593500971794,
+ "learning_rate": 0.0006,
+ "loss": 4.097907066345215,
+ "step": 2413
+ },
+ {
+ "epoch": 33.531004366812226,
+ "grad_norm": 0.015182029455900192,
+ "learning_rate": 0.0006,
+ "loss": 4.238499641418457,
+ "step": 2414
+ },
+ {
+ "epoch": 33.544978165938865,
+ "grad_norm": 0.01643732376396656,
+ "learning_rate": 0.0006,
+ "loss": 4.0661492347717285,
+ "step": 2415
+ },
+ {
+ "epoch": 33.5589519650655,
+ "grad_norm": 0.015614228323101997,
+ "learning_rate": 0.0006,
+ "loss": 4.120795249938965,
+ "step": 2416
+ },
+ {
+ "epoch": 33.57292576419214,
+ "grad_norm": 0.016066187992691994,
+ "learning_rate": 0.0006,
+ "loss": 4.096465110778809,
+ "step": 2417
+ },
+ {
+ "epoch": 33.58689956331878,
+ "grad_norm": 0.017214342951774597,
+ "learning_rate": 0.0006,
+ "loss": 4.242628574371338,
+ "step": 2418
+ },
+ {
+ "epoch": 33.60087336244541,
+ "grad_norm": 0.01787586510181427,
+ "learning_rate": 0.0006,
+ "loss": 4.149589538574219,
+ "step": 2419
+ },
+ {
+ "epoch": 33.61484716157205,
+ "grad_norm": 0.019455241039395332,
+ "learning_rate": 0.0006,
+ "loss": 4.267796039581299,
+ "step": 2420
+ },
+ {
+ "epoch": 33.62882096069869,
+ "grad_norm": 0.01984674669802189,
+ "learning_rate": 0.0006,
+ "loss": 4.175220966339111,
+ "step": 2421
+ },
+ {
+ "epoch": 33.64279475982533,
+ "grad_norm": 0.021933717653155327,
+ "learning_rate": 0.0006,
+ "loss": 4.106647491455078,
+ "step": 2422
+ },
+ {
+ "epoch": 33.65676855895197,
+ "grad_norm": 0.02123589813709259,
+ "learning_rate": 0.0006,
+ "loss": 4.204996109008789,
+ "step": 2423
+ },
+ {
+ "epoch": 33.670742358078606,
+ "grad_norm": 0.019814589992165565,
+ "learning_rate": 0.0006,
+ "loss": 4.186171531677246,
+ "step": 2424
+ },
+ {
+ "epoch": 33.68471615720524,
+ "grad_norm": 0.02059728465974331,
+ "learning_rate": 0.0006,
+ "loss": 4.170509338378906,
+ "step": 2425
+ },
+ {
+ "epoch": 33.698689956331876,
+ "grad_norm": 0.018483608961105347,
+ "learning_rate": 0.0006,
+ "loss": 4.085585594177246,
+ "step": 2426
+ },
+ {
+ "epoch": 33.712663755458514,
+ "grad_norm": 0.018403852358460426,
+ "learning_rate": 0.0006,
+ "loss": 4.2551774978637695,
+ "step": 2427
+ },
+ {
+ "epoch": 33.72663755458515,
+ "grad_norm": 0.018653830513358116,
+ "learning_rate": 0.0006,
+ "loss": 4.091766357421875,
+ "step": 2428
+ },
+ {
+ "epoch": 33.74061135371179,
+ "grad_norm": 0.017692334949970245,
+ "learning_rate": 0.0006,
+ "loss": 4.100838661193848,
+ "step": 2429
+ },
+ {
+ "epoch": 33.75458515283843,
+ "grad_norm": 0.016029126942157745,
+ "learning_rate": 0.0006,
+ "loss": 4.1530680656433105,
+ "step": 2430
+ },
+ {
+ "epoch": 33.76855895196506,
+ "grad_norm": 0.016095684841275215,
+ "learning_rate": 0.0006,
+ "loss": 4.20544958114624,
+ "step": 2431
+ },
+ {
+ "epoch": 33.7825327510917,
+ "grad_norm": 0.016375798732042313,
+ "learning_rate": 0.0006,
+ "loss": 4.199177265167236,
+ "step": 2432
+ },
+ {
+ "epoch": 33.79650655021834,
+ "grad_norm": 0.01590786688029766,
+ "learning_rate": 0.0006,
+ "loss": 4.200799465179443,
+ "step": 2433
+ },
+ {
+ "epoch": 33.81048034934498,
+ "grad_norm": 0.014896097593009472,
+ "learning_rate": 0.0006,
+ "loss": 4.185850143432617,
+ "step": 2434
+ },
+ {
+ "epoch": 33.82445414847162,
+ "grad_norm": 0.01438513956964016,
+ "learning_rate": 0.0006,
+ "loss": 4.214962005615234,
+ "step": 2435
+ },
+ {
+ "epoch": 33.838427947598255,
+ "grad_norm": 0.014459339901804924,
+ "learning_rate": 0.0006,
+ "loss": 4.196972370147705,
+ "step": 2436
+ },
+ {
+ "epoch": 33.852401746724894,
+ "grad_norm": 0.014782169833779335,
+ "learning_rate": 0.0006,
+ "loss": 4.0885910987854,
+ "step": 2437
+ },
+ {
+ "epoch": 33.866375545851525,
+ "grad_norm": 0.01520370040088892,
+ "learning_rate": 0.0006,
+ "loss": 4.0323262214660645,
+ "step": 2438
+ },
+ {
+ "epoch": 33.880349344978164,
+ "grad_norm": 0.014056282117962837,
+ "learning_rate": 0.0006,
+ "loss": 4.1386518478393555,
+ "step": 2439
+ },
+ {
+ "epoch": 33.8943231441048,
+ "grad_norm": 0.013567990623414516,
+ "learning_rate": 0.0006,
+ "loss": 4.095754623413086,
+ "step": 2440
+ },
+ {
+ "epoch": 33.90829694323144,
+ "grad_norm": 0.013306716457009315,
+ "learning_rate": 0.0006,
+ "loss": 4.1623077392578125,
+ "step": 2441
+ },
+ {
+ "epoch": 33.92227074235808,
+ "grad_norm": 0.014047150500118732,
+ "learning_rate": 0.0006,
+ "loss": 4.142541885375977,
+ "step": 2442
+ },
+ {
+ "epoch": 33.93624454148472,
+ "grad_norm": 0.01377617847174406,
+ "learning_rate": 0.0006,
+ "loss": 4.268551826477051,
+ "step": 2443
+ },
+ {
+ "epoch": 33.95021834061135,
+ "grad_norm": 0.014671621844172478,
+ "learning_rate": 0.0006,
+ "loss": 4.198196887969971,
+ "step": 2444
+ },
+ {
+ "epoch": 33.96419213973799,
+ "grad_norm": 0.014806744642555714,
+ "learning_rate": 0.0006,
+ "loss": 4.057899475097656,
+ "step": 2445
+ },
+ {
+ "epoch": 33.97816593886463,
+ "grad_norm": 0.015734272077679634,
+ "learning_rate": 0.0006,
+ "loss": 4.163505554199219,
+ "step": 2446
+ },
+ {
+ "epoch": 33.992139737991266,
+ "grad_norm": 0.015250146389007568,
+ "learning_rate": 0.0006,
+ "loss": 4.076294898986816,
+ "step": 2447
+ },
+ {
+ "epoch": 34.0,
+ "grad_norm": 0.016733702272176743,
+ "learning_rate": 0.0006,
+ "loss": 4.138420104980469,
+ "step": 2448
+ },
+ {
+ "epoch": 34.0,
+ "eval_loss": 4.542503356933594,
+ "eval_runtime": 56.983,
+ "eval_samples_per_second": 42.855,
+ "eval_steps_per_second": 1.351,
+ "step": 2448
+ },
+ {
+ "epoch": 34.01397379912664,
+ "grad_norm": 0.015866613015532494,
+ "learning_rate": 0.0006,
+ "loss": 4.032081127166748,
+ "step": 2449
+ },
+ {
+ "epoch": 34.02794759825328,
+ "grad_norm": 0.01946570910513401,
+ "learning_rate": 0.0006,
+ "loss": 4.122320175170898,
+ "step": 2450
+ },
+ {
+ "epoch": 34.041921397379916,
+ "grad_norm": 0.020265938714146614,
+ "learning_rate": 0.0006,
+ "loss": 4.06460428237915,
+ "step": 2451
+ },
+ {
+ "epoch": 34.05589519650655,
+ "grad_norm": 0.018887171521782875,
+ "learning_rate": 0.0006,
+ "loss": 4.096906661987305,
+ "step": 2452
+ },
+ {
+ "epoch": 34.069868995633186,
+ "grad_norm": 0.017304692417383194,
+ "learning_rate": 0.0006,
+ "loss": 4.137261390686035,
+ "step": 2453
+ },
+ {
+ "epoch": 34.083842794759825,
+ "grad_norm": 0.017029082402586937,
+ "learning_rate": 0.0006,
+ "loss": 4.081112861633301,
+ "step": 2454
+ },
+ {
+ "epoch": 34.09781659388646,
+ "grad_norm": 0.016414813697338104,
+ "learning_rate": 0.0006,
+ "loss": 4.140783309936523,
+ "step": 2455
+ },
+ {
+ "epoch": 34.1117903930131,
+ "grad_norm": 0.018238281831145287,
+ "learning_rate": 0.0006,
+ "loss": 4.161116600036621,
+ "step": 2456
+ },
+ {
+ "epoch": 34.12576419213974,
+ "grad_norm": 0.019929232075810432,
+ "learning_rate": 0.0006,
+ "loss": 4.243722438812256,
+ "step": 2457
+ },
+ {
+ "epoch": 34.13973799126637,
+ "grad_norm": 0.019375786185264587,
+ "learning_rate": 0.0006,
+ "loss": 4.134873390197754,
+ "step": 2458
+ },
+ {
+ "epoch": 34.15371179039301,
+ "grad_norm": 0.018151137977838516,
+ "learning_rate": 0.0006,
+ "loss": 4.098745346069336,
+ "step": 2459
+ },
+ {
+ "epoch": 34.16768558951965,
+ "grad_norm": 0.017054934054613113,
+ "learning_rate": 0.0006,
+ "loss": 4.194662570953369,
+ "step": 2460
+ },
+ {
+ "epoch": 34.18165938864629,
+ "grad_norm": 0.018220555037260056,
+ "learning_rate": 0.0006,
+ "loss": 4.128866195678711,
+ "step": 2461
+ },
+ {
+ "epoch": 34.19563318777293,
+ "grad_norm": 0.01692879945039749,
+ "learning_rate": 0.0006,
+ "loss": 4.062477111816406,
+ "step": 2462
+ },
+ {
+ "epoch": 34.209606986899566,
+ "grad_norm": 0.01674199476838112,
+ "learning_rate": 0.0006,
+ "loss": 4.186149597167969,
+ "step": 2463
+ },
+ {
+ "epoch": 34.223580786026204,
+ "grad_norm": 0.016527455300092697,
+ "learning_rate": 0.0006,
+ "loss": 4.0734028816223145,
+ "step": 2464
+ },
+ {
+ "epoch": 34.237554585152836,
+ "grad_norm": 0.016448386013507843,
+ "learning_rate": 0.0006,
+ "loss": 4.098896026611328,
+ "step": 2465
+ },
+ {
+ "epoch": 34.251528384279474,
+ "grad_norm": 0.014644953422248363,
+ "learning_rate": 0.0006,
+ "loss": 4.1031904220581055,
+ "step": 2466
+ },
+ {
+ "epoch": 34.26550218340611,
+ "grad_norm": 0.016649341210722923,
+ "learning_rate": 0.0006,
+ "loss": 4.148791313171387,
+ "step": 2467
+ },
+ {
+ "epoch": 34.27947598253275,
+ "grad_norm": 0.017442714422941208,
+ "learning_rate": 0.0006,
+ "loss": 4.070799827575684,
+ "step": 2468
+ },
+ {
+ "epoch": 34.29344978165939,
+ "grad_norm": 0.017852013930678368,
+ "learning_rate": 0.0006,
+ "loss": 4.031924247741699,
+ "step": 2469
+ },
+ {
+ "epoch": 34.30742358078603,
+ "grad_norm": 0.016371270641684532,
+ "learning_rate": 0.0006,
+ "loss": 4.149420738220215,
+ "step": 2470
+ },
+ {
+ "epoch": 34.32139737991266,
+ "grad_norm": 0.015846814960241318,
+ "learning_rate": 0.0006,
+ "loss": 4.15427303314209,
+ "step": 2471
+ },
+ {
+ "epoch": 34.3353711790393,
+ "grad_norm": 0.017411163076758385,
+ "learning_rate": 0.0006,
+ "loss": 4.1304192543029785,
+ "step": 2472
+ },
+ {
+ "epoch": 34.34934497816594,
+ "grad_norm": 0.01780046708881855,
+ "learning_rate": 0.0006,
+ "loss": 4.191732406616211,
+ "step": 2473
+ },
+ {
+ "epoch": 34.36331877729258,
+ "grad_norm": 0.01958383060991764,
+ "learning_rate": 0.0006,
+ "loss": 4.068355083465576,
+ "step": 2474
+ },
+ {
+ "epoch": 34.377292576419215,
+ "grad_norm": 0.022173317149281502,
+ "learning_rate": 0.0006,
+ "loss": 4.182929039001465,
+ "step": 2475
+ },
+ {
+ "epoch": 34.391266375545854,
+ "grad_norm": 0.022389156743884087,
+ "learning_rate": 0.0006,
+ "loss": 4.110218048095703,
+ "step": 2476
+ },
+ {
+ "epoch": 34.40524017467249,
+ "grad_norm": 0.022478511556982994,
+ "learning_rate": 0.0006,
+ "loss": 4.095479488372803,
+ "step": 2477
+ },
+ {
+ "epoch": 34.419213973799124,
+ "grad_norm": 0.019504647701978683,
+ "learning_rate": 0.0006,
+ "loss": 4.101964950561523,
+ "step": 2478
+ },
+ {
+ "epoch": 34.43318777292576,
+ "grad_norm": 0.019378576427698135,
+ "learning_rate": 0.0006,
+ "loss": 4.14620304107666,
+ "step": 2479
+ },
+ {
+ "epoch": 34.4471615720524,
+ "grad_norm": 0.017993204295635223,
+ "learning_rate": 0.0006,
+ "loss": 4.183492660522461,
+ "step": 2480
+ },
+ {
+ "epoch": 34.46113537117904,
+ "grad_norm": 0.01814805157482624,
+ "learning_rate": 0.0006,
+ "loss": 4.151247978210449,
+ "step": 2481
+ },
+ {
+ "epoch": 34.47510917030568,
+ "grad_norm": 0.018622448667883873,
+ "learning_rate": 0.0006,
+ "loss": 4.020938396453857,
+ "step": 2482
+ },
+ {
+ "epoch": 34.48908296943232,
+ "grad_norm": 0.018360132351517677,
+ "learning_rate": 0.0006,
+ "loss": 4.064734935760498,
+ "step": 2483
+ },
+ {
+ "epoch": 34.50305676855895,
+ "grad_norm": 0.017715711146593094,
+ "learning_rate": 0.0006,
+ "loss": 4.185687065124512,
+ "step": 2484
+ },
+ {
+ "epoch": 34.51703056768559,
+ "grad_norm": 0.01714295893907547,
+ "learning_rate": 0.0006,
+ "loss": 4.1044721603393555,
+ "step": 2485
+ },
+ {
+ "epoch": 34.531004366812226,
+ "grad_norm": 0.016943803057074547,
+ "learning_rate": 0.0006,
+ "loss": 4.106837272644043,
+ "step": 2486
+ },
+ {
+ "epoch": 34.544978165938865,
+ "grad_norm": 0.0182423684746027,
+ "learning_rate": 0.0006,
+ "loss": 4.1041340827941895,
+ "step": 2487
+ },
+ {
+ "epoch": 34.5589519650655,
+ "grad_norm": 0.01636587455868721,
+ "learning_rate": 0.0006,
+ "loss": 3.9735546112060547,
+ "step": 2488
+ },
+ {
+ "epoch": 34.57292576419214,
+ "grad_norm": 0.014593546278774738,
+ "learning_rate": 0.0006,
+ "loss": 4.041980266571045,
+ "step": 2489
+ },
+ {
+ "epoch": 34.58689956331878,
+ "grad_norm": 0.015436965972185135,
+ "learning_rate": 0.0006,
+ "loss": 4.301148414611816,
+ "step": 2490
+ },
+ {
+ "epoch": 34.60087336244541,
+ "grad_norm": 0.015075269155204296,
+ "learning_rate": 0.0006,
+ "loss": 4.118527412414551,
+ "step": 2491
+ },
+ {
+ "epoch": 34.61484716157205,
+ "grad_norm": 0.01701514609158039,
+ "learning_rate": 0.0006,
+ "loss": 4.245968818664551,
+ "step": 2492
+ },
+ {
+ "epoch": 34.62882096069869,
+ "grad_norm": 0.018047887831926346,
+ "learning_rate": 0.0006,
+ "loss": 4.227261066436768,
+ "step": 2493
+ },
+ {
+ "epoch": 34.64279475982533,
+ "grad_norm": 0.018499085679650307,
+ "learning_rate": 0.0006,
+ "loss": 4.121360778808594,
+ "step": 2494
+ },
+ {
+ "epoch": 34.65676855895197,
+ "grad_norm": 0.018987711519002914,
+ "learning_rate": 0.0006,
+ "loss": 4.131564617156982,
+ "step": 2495
+ },
+ {
+ "epoch": 34.670742358078606,
+ "grad_norm": 0.01834476925432682,
+ "learning_rate": 0.0006,
+ "loss": 4.224941253662109,
+ "step": 2496
+ },
+ {
+ "epoch": 34.68471615720524,
+ "grad_norm": 0.017507942393422127,
+ "learning_rate": 0.0006,
+ "loss": 4.1394548416137695,
+ "step": 2497
+ },
+ {
+ "epoch": 34.698689956331876,
+ "grad_norm": 0.01680580899119377,
+ "learning_rate": 0.0006,
+ "loss": 4.1073431968688965,
+ "step": 2498
+ },
+ {
+ "epoch": 34.712663755458514,
+ "grad_norm": 0.015233514830470085,
+ "learning_rate": 0.0006,
+ "loss": 4.136994361877441,
+ "step": 2499
+ },
+ {
+ "epoch": 34.72663755458515,
+ "grad_norm": 0.015482917428016663,
+ "learning_rate": 0.0006,
+ "loss": 4.110719680786133,
+ "step": 2500
+ },
+ {
+ "epoch": 34.74061135371179,
+ "grad_norm": 0.015940461307764053,
+ "learning_rate": 0.0006,
+ "loss": 4.207851886749268,
+ "step": 2501
+ },
+ {
+ "epoch": 34.75458515283843,
+ "grad_norm": 0.015355597250163555,
+ "learning_rate": 0.0006,
+ "loss": 4.169469833374023,
+ "step": 2502
+ },
+ {
+ "epoch": 34.76855895196506,
+ "grad_norm": 0.015974000096321106,
+ "learning_rate": 0.0006,
+ "loss": 4.110061168670654,
+ "step": 2503
+ },
+ {
+ "epoch": 34.7825327510917,
+ "grad_norm": 0.016088927164673805,
+ "learning_rate": 0.0006,
+ "loss": 4.155452728271484,
+ "step": 2504
+ },
+ {
+ "epoch": 34.79650655021834,
+ "grad_norm": 0.016156120225787163,
+ "learning_rate": 0.0006,
+ "loss": 4.178092956542969,
+ "step": 2505
+ },
+ {
+ "epoch": 34.81048034934498,
+ "grad_norm": 0.014928647316992283,
+ "learning_rate": 0.0006,
+ "loss": 4.05457878112793,
+ "step": 2506
+ },
+ {
+ "epoch": 34.82445414847162,
+ "grad_norm": 0.015289328061044216,
+ "learning_rate": 0.0006,
+ "loss": 4.18869686126709,
+ "step": 2507
+ },
+ {
+ "epoch": 34.838427947598255,
+ "grad_norm": 0.015799039974808693,
+ "learning_rate": 0.0006,
+ "loss": 4.106073379516602,
+ "step": 2508
+ },
+ {
+ "epoch": 34.852401746724894,
+ "grad_norm": 0.015824418514966965,
+ "learning_rate": 0.0006,
+ "loss": 4.0634589195251465,
+ "step": 2509
+ },
+ {
+ "epoch": 34.866375545851525,
+ "grad_norm": 0.015321656130254269,
+ "learning_rate": 0.0006,
+ "loss": 4.093090057373047,
+ "step": 2510
+ },
+ {
+ "epoch": 34.880349344978164,
+ "grad_norm": 0.014380166307091713,
+ "learning_rate": 0.0006,
+ "loss": 4.1750640869140625,
+ "step": 2511
+ },
+ {
+ "epoch": 34.8943231441048,
+ "grad_norm": 0.014962567947804928,
+ "learning_rate": 0.0006,
+ "loss": 4.185277938842773,
+ "step": 2512
+ },
+ {
+ "epoch": 34.90829694323144,
+ "grad_norm": 0.014467543922364712,
+ "learning_rate": 0.0006,
+ "loss": 4.1600799560546875,
+ "step": 2513
+ },
+ {
+ "epoch": 34.92227074235808,
+ "grad_norm": 0.01545181218534708,
+ "learning_rate": 0.0006,
+ "loss": 4.096299648284912,
+ "step": 2514
+ },
+ {
+ "epoch": 34.93624454148472,
+ "grad_norm": 0.01605340465903282,
+ "learning_rate": 0.0006,
+ "loss": 4.203323841094971,
+ "step": 2515
+ },
+ {
+ "epoch": 34.95021834061135,
+ "grad_norm": 0.015203815884888172,
+ "learning_rate": 0.0006,
+ "loss": 4.201385021209717,
+ "step": 2516
+ },
+ {
+ "epoch": 34.96419213973799,
+ "grad_norm": 0.015149409882724285,
+ "learning_rate": 0.0006,
+ "loss": 4.088164329528809,
+ "step": 2517
+ },
+ {
+ "epoch": 34.97816593886463,
+ "grad_norm": 0.014885558746755123,
+ "learning_rate": 0.0006,
+ "loss": 4.073911666870117,
+ "step": 2518
+ },
+ {
+ "epoch": 34.992139737991266,
+ "grad_norm": 0.015003419481217861,
+ "learning_rate": 0.0006,
+ "loss": 4.157113552093506,
+ "step": 2519
+ },
+ {
+ "epoch": 35.0,
+ "grad_norm": 0.01641916297376156,
+ "learning_rate": 0.0006,
+ "loss": 3.970052719116211,
+ "step": 2520
+ },
+ {
+ "epoch": 35.0,
+ "eval_loss": 4.641692161560059,
+ "eval_runtime": 56.8836,
+ "eval_samples_per_second": 42.93,
+ "eval_steps_per_second": 1.354,
+ "step": 2520
+ },
+ {
+ "epoch": 35.01397379912664,
+ "grad_norm": 0.015894673764705658,
+ "learning_rate": 0.0006,
+ "loss": 4.104437351226807,
+ "step": 2521
+ },
+ {
+ "epoch": 35.02794759825328,
+ "grad_norm": 0.0173867866396904,
+ "learning_rate": 0.0006,
+ "loss": 4.091726303100586,
+ "step": 2522
+ },
+ {
+ "epoch": 35.041921397379916,
+ "grad_norm": 0.019272832199931145,
+ "learning_rate": 0.0006,
+ "loss": 4.217905044555664,
+ "step": 2523
+ },
+ {
+ "epoch": 35.05589519650655,
+ "grad_norm": 0.021820371970534325,
+ "learning_rate": 0.0006,
+ "loss": 4.111428260803223,
+ "step": 2524
+ },
+ {
+ "epoch": 35.069868995633186,
+ "grad_norm": 0.02340741828083992,
+ "learning_rate": 0.0006,
+ "loss": 4.115771293640137,
+ "step": 2525
+ },
+ {
+ "epoch": 35.083842794759825,
+ "grad_norm": 0.023760691285133362,
+ "learning_rate": 0.0006,
+ "loss": 4.132383346557617,
+ "step": 2526
+ },
+ {
+ "epoch": 35.09781659388646,
+ "grad_norm": 0.02073468267917633,
+ "learning_rate": 0.0006,
+ "loss": 4.272691249847412,
+ "step": 2527
+ },
+ {
+ "epoch": 35.1117903930131,
+ "grad_norm": 0.02197718247771263,
+ "learning_rate": 0.0006,
+ "loss": 4.105586528778076,
+ "step": 2528
+ },
+ {
+ "epoch": 35.12576419213974,
+ "grad_norm": 0.0200254675000906,
+ "learning_rate": 0.0006,
+ "loss": 3.968820095062256,
+ "step": 2529
+ },
+ {
+ "epoch": 35.13973799126637,
+ "grad_norm": 0.01929587684571743,
+ "learning_rate": 0.0006,
+ "loss": 4.045468807220459,
+ "step": 2530
+ },
+ {
+ "epoch": 35.15371179039301,
+ "grad_norm": 0.01886162906885147,
+ "learning_rate": 0.0006,
+ "loss": 4.185571670532227,
+ "step": 2531
+ },
+ {
+ "epoch": 35.16768558951965,
+ "grad_norm": 0.019789839163422585,
+ "learning_rate": 0.0006,
+ "loss": 3.9734551906585693,
+ "step": 2532
+ },
+ {
+ "epoch": 35.18165938864629,
+ "grad_norm": 0.0198142658919096,
+ "learning_rate": 0.0006,
+ "loss": 4.2662506103515625,
+ "step": 2533
+ },
+ {
+ "epoch": 35.19563318777293,
+ "grad_norm": 0.02087012305855751,
+ "learning_rate": 0.0006,
+ "loss": 4.175931930541992,
+ "step": 2534
+ },
+ {
+ "epoch": 35.209606986899566,
+ "grad_norm": 0.020294535905122757,
+ "learning_rate": 0.0006,
+ "loss": 4.060324668884277,
+ "step": 2535
+ },
+ {
+ "epoch": 35.223580786026204,
+ "grad_norm": 0.01908983662724495,
+ "learning_rate": 0.0006,
+ "loss": 4.217939853668213,
+ "step": 2536
+ },
+ {
+ "epoch": 35.237554585152836,
+ "grad_norm": 0.0178892333060503,
+ "learning_rate": 0.0006,
+ "loss": 4.190216541290283,
+ "step": 2537
+ },
+ {
+ "epoch": 35.251528384279474,
+ "grad_norm": 0.017892315983772278,
+ "learning_rate": 0.0006,
+ "loss": 4.067419052124023,
+ "step": 2538
+ },
+ {
+ "epoch": 35.26550218340611,
+ "grad_norm": 0.018556250259280205,
+ "learning_rate": 0.0006,
+ "loss": 4.105958461761475,
+ "step": 2539
+ },
+ {
+ "epoch": 35.27947598253275,
+ "grad_norm": 0.01944422349333763,
+ "learning_rate": 0.0006,
+ "loss": 4.155775547027588,
+ "step": 2540
+ },
+ {
+ "epoch": 35.29344978165939,
+ "grad_norm": 0.017333146184682846,
+ "learning_rate": 0.0006,
+ "loss": 4.244513511657715,
+ "step": 2541
+ },
+ {
+ "epoch": 35.30742358078603,
+ "grad_norm": 0.01639244332909584,
+ "learning_rate": 0.0006,
+ "loss": 4.167815685272217,
+ "step": 2542
+ },
+ {
+ "epoch": 35.32139737991266,
+ "grad_norm": 0.01610143668949604,
+ "learning_rate": 0.0006,
+ "loss": 4.121486663818359,
+ "step": 2543
+ },
+ {
+ "epoch": 35.3353711790393,
+ "grad_norm": 0.01656411588191986,
+ "learning_rate": 0.0006,
+ "loss": 4.1755499839782715,
+ "step": 2544
+ },
+ {
+ "epoch": 35.34934497816594,
+ "grad_norm": 0.015075297094881535,
+ "learning_rate": 0.0006,
+ "loss": 4.102992534637451,
+ "step": 2545
+ },
+ {
+ "epoch": 35.36331877729258,
+ "grad_norm": 0.01538170501589775,
+ "learning_rate": 0.0006,
+ "loss": 4.161603927612305,
+ "step": 2546
+ },
+ {
+ "epoch": 35.377292576419215,
+ "grad_norm": 0.01483312901109457,
+ "learning_rate": 0.0006,
+ "loss": 4.085355758666992,
+ "step": 2547
+ },
+ {
+ "epoch": 35.391266375545854,
+ "grad_norm": 0.014894738793373108,
+ "learning_rate": 0.0006,
+ "loss": 4.083585262298584,
+ "step": 2548
+ },
+ {
+ "epoch": 35.40524017467249,
+ "grad_norm": 0.01475780550390482,
+ "learning_rate": 0.0006,
+ "loss": 4.012645721435547,
+ "step": 2549
+ },
+ {
+ "epoch": 35.419213973799124,
+ "grad_norm": 0.014531653374433517,
+ "learning_rate": 0.0006,
+ "loss": 4.197113990783691,
+ "step": 2550
+ },
+ {
+ "epoch": 35.43318777292576,
+ "grad_norm": 0.015316493809223175,
+ "learning_rate": 0.0006,
+ "loss": 4.227970123291016,
+ "step": 2551
+ },
+ {
+ "epoch": 35.4471615720524,
+ "grad_norm": 0.014685571193695068,
+ "learning_rate": 0.0006,
+ "loss": 4.089415550231934,
+ "step": 2552
+ },
+ {
+ "epoch": 35.46113537117904,
+ "grad_norm": 0.015337035991251469,
+ "learning_rate": 0.0006,
+ "loss": 4.080583572387695,
+ "step": 2553
+ },
+ {
+ "epoch": 35.47510917030568,
+ "grad_norm": 0.016754208132624626,
+ "learning_rate": 0.0006,
+ "loss": 4.163935661315918,
+ "step": 2554
+ },
+ {
+ "epoch": 35.48908296943232,
+ "grad_norm": 0.017535798251628876,
+ "learning_rate": 0.0006,
+ "loss": 4.093027591705322,
+ "step": 2555
+ },
+ {
+ "epoch": 35.50305676855895,
+ "grad_norm": 0.017336975783109665,
+ "learning_rate": 0.0006,
+ "loss": 4.127961158752441,
+ "step": 2556
+ },
+ {
+ "epoch": 35.51703056768559,
+ "grad_norm": 0.016242152079939842,
+ "learning_rate": 0.0006,
+ "loss": 4.155593395233154,
+ "step": 2557
+ },
+ {
+ "epoch": 35.531004366812226,
+ "grad_norm": 0.015121322125196457,
+ "learning_rate": 0.0006,
+ "loss": 3.984004020690918,
+ "step": 2558
+ },
+ {
+ "epoch": 35.544978165938865,
+ "grad_norm": 0.015178884379565716,
+ "learning_rate": 0.0006,
+ "loss": 4.111564636230469,
+ "step": 2559
+ },
+ {
+ "epoch": 35.5589519650655,
+ "grad_norm": 0.014777131378650665,
+ "learning_rate": 0.0006,
+ "loss": 4.041402816772461,
+ "step": 2560
+ },
+ {
+ "epoch": 35.57292576419214,
+ "grad_norm": 0.015197676606476307,
+ "learning_rate": 0.0006,
+ "loss": 4.133706092834473,
+ "step": 2561
+ },
+ {
+ "epoch": 35.58689956331878,
+ "grad_norm": 0.015905972570180893,
+ "learning_rate": 0.0006,
+ "loss": 4.1069135665893555,
+ "step": 2562
+ },
+ {
+ "epoch": 35.60087336244541,
+ "grad_norm": 0.015942877158522606,
+ "learning_rate": 0.0006,
+ "loss": 4.0720624923706055,
+ "step": 2563
+ },
+ {
+ "epoch": 35.61484716157205,
+ "grad_norm": 0.016697583720088005,
+ "learning_rate": 0.0006,
+ "loss": 4.167448043823242,
+ "step": 2564
+ },
+ {
+ "epoch": 35.62882096069869,
+ "grad_norm": 0.015603577718138695,
+ "learning_rate": 0.0006,
+ "loss": 4.127327919006348,
+ "step": 2565
+ },
+ {
+ "epoch": 35.64279475982533,
+ "grad_norm": 0.017153460532426834,
+ "learning_rate": 0.0006,
+ "loss": 4.056543350219727,
+ "step": 2566
+ },
+ {
+ "epoch": 35.65676855895197,
+ "grad_norm": 0.016731958836317062,
+ "learning_rate": 0.0006,
+ "loss": 4.0681843757629395,
+ "step": 2567
+ },
+ {
+ "epoch": 35.670742358078606,
+ "grad_norm": 0.016167359426617622,
+ "learning_rate": 0.0006,
+ "loss": 4.174992084503174,
+ "step": 2568
+ },
+ {
+ "epoch": 35.68471615720524,
+ "grad_norm": 0.01615108922123909,
+ "learning_rate": 0.0006,
+ "loss": 4.1383161544799805,
+ "step": 2569
+ },
+ {
+ "epoch": 35.698689956331876,
+ "grad_norm": 0.015750816091895103,
+ "learning_rate": 0.0006,
+ "loss": 4.031595706939697,
+ "step": 2570
+ },
+ {
+ "epoch": 35.712663755458514,
+ "grad_norm": 0.015416189096868038,
+ "learning_rate": 0.0006,
+ "loss": 4.067498207092285,
+ "step": 2571
+ },
+ {
+ "epoch": 35.72663755458515,
+ "grad_norm": 0.01570792682468891,
+ "learning_rate": 0.0006,
+ "loss": 4.1338300704956055,
+ "step": 2572
+ },
+ {
+ "epoch": 35.74061135371179,
+ "grad_norm": 0.016367752104997635,
+ "learning_rate": 0.0006,
+ "loss": 4.152523040771484,
+ "step": 2573
+ },
+ {
+ "epoch": 35.75458515283843,
+ "grad_norm": 0.016349803656339645,
+ "learning_rate": 0.0006,
+ "loss": 4.101985454559326,
+ "step": 2574
+ },
+ {
+ "epoch": 35.76855895196506,
+ "grad_norm": 0.01663660630583763,
+ "learning_rate": 0.0006,
+ "loss": 4.190130710601807,
+ "step": 2575
+ },
+ {
+ "epoch": 35.7825327510917,
+ "grad_norm": 0.015599170699715614,
+ "learning_rate": 0.0006,
+ "loss": 4.0792436599731445,
+ "step": 2576
+ },
+ {
+ "epoch": 35.79650655021834,
+ "grad_norm": 0.015472427010536194,
+ "learning_rate": 0.0006,
+ "loss": 4.186446666717529,
+ "step": 2577
+ },
+ {
+ "epoch": 35.81048034934498,
+ "grad_norm": 0.015520439483225346,
+ "learning_rate": 0.0006,
+ "loss": 4.246635437011719,
+ "step": 2578
+ },
+ {
+ "epoch": 35.82445414847162,
+ "grad_norm": 0.016998281702399254,
+ "learning_rate": 0.0006,
+ "loss": 4.13435173034668,
+ "step": 2579
+ },
+ {
+ "epoch": 35.838427947598255,
+ "grad_norm": 0.016290361061692238,
+ "learning_rate": 0.0006,
+ "loss": 4.107428073883057,
+ "step": 2580
+ },
+ {
+ "epoch": 35.852401746724894,
+ "grad_norm": 0.0165516659617424,
+ "learning_rate": 0.0006,
+ "loss": 4.1612138748168945,
+ "step": 2581
+ },
+ {
+ "epoch": 35.866375545851525,
+ "grad_norm": 0.01736091636121273,
+ "learning_rate": 0.0006,
+ "loss": 4.1072869300842285,
+ "step": 2582
+ },
+ {
+ "epoch": 35.880349344978164,
+ "grad_norm": 0.01787031814455986,
+ "learning_rate": 0.0006,
+ "loss": 4.1076812744140625,
+ "step": 2583
+ },
+ {
+ "epoch": 35.8943231441048,
+ "grad_norm": 0.015437985770404339,
+ "learning_rate": 0.0006,
+ "loss": 4.097488880157471,
+ "step": 2584
+ },
+ {
+ "epoch": 35.90829694323144,
+ "grad_norm": 0.016306741163134575,
+ "learning_rate": 0.0006,
+ "loss": 4.105017185211182,
+ "step": 2585
+ },
+ {
+ "epoch": 35.92227074235808,
+ "grad_norm": 0.015723245218396187,
+ "learning_rate": 0.0006,
+ "loss": 4.205283164978027,
+ "step": 2586
+ },
+ {
+ "epoch": 35.93624454148472,
+ "grad_norm": 0.014626456424593925,
+ "learning_rate": 0.0006,
+ "loss": 4.086552619934082,
+ "step": 2587
+ },
+ {
+ "epoch": 35.95021834061135,
+ "grad_norm": 0.014348454773426056,
+ "learning_rate": 0.0006,
+ "loss": 4.171329975128174,
+ "step": 2588
+ },
+ {
+ "epoch": 35.96419213973799,
+ "grad_norm": 0.015598480589687824,
+ "learning_rate": 0.0006,
+ "loss": 4.140815258026123,
+ "step": 2589
+ },
+ {
+ "epoch": 35.97816593886463,
+ "grad_norm": 0.017126062884926796,
+ "learning_rate": 0.0006,
+ "loss": 4.157176971435547,
+ "step": 2590
+ },
+ {
+ "epoch": 35.992139737991266,
+ "grad_norm": 0.01860199309885502,
+ "learning_rate": 0.0006,
+ "loss": 4.252880096435547,
+ "step": 2591
+ },
+ {
+ "epoch": 36.0,
+ "grad_norm": 0.021780220791697502,
+ "learning_rate": 0.0006,
+ "loss": 4.249861717224121,
+ "step": 2592
+ },
+ {
+ "epoch": 36.0,
+ "eval_loss": 4.606466293334961,
+ "eval_runtime": 56.6217,
+ "eval_samples_per_second": 43.128,
+ "eval_steps_per_second": 1.36,
+ "step": 2592
+ },
+ {
+ "epoch": 36.01397379912664,
+ "grad_norm": 0.020892828702926636,
+ "learning_rate": 0.0006,
+ "loss": 4.102055549621582,
+ "step": 2593
+ },
+ {
+ "epoch": 36.02794759825328,
+ "grad_norm": 0.020306255668401718,
+ "learning_rate": 0.0006,
+ "loss": 4.184308052062988,
+ "step": 2594
+ },
+ {
+ "epoch": 36.041921397379916,
+ "grad_norm": 0.02025097794830799,
+ "learning_rate": 0.0006,
+ "loss": 3.9691858291625977,
+ "step": 2595
+ },
+ {
+ "epoch": 36.05589519650655,
+ "grad_norm": 0.020254382863640785,
+ "learning_rate": 0.0006,
+ "loss": 4.02957820892334,
+ "step": 2596
+ },
+ {
+ "epoch": 36.069868995633186,
+ "grad_norm": 0.019307196140289307,
+ "learning_rate": 0.0006,
+ "loss": 4.033181667327881,
+ "step": 2597
+ },
+ {
+ "epoch": 36.083842794759825,
+ "grad_norm": 0.0198476854711771,
+ "learning_rate": 0.0006,
+ "loss": 4.0816874504089355,
+ "step": 2598
+ },
+ {
+ "epoch": 36.09781659388646,
+ "grad_norm": 0.019916515797376633,
+ "learning_rate": 0.0006,
+ "loss": 4.164889812469482,
+ "step": 2599
+ },
+ {
+ "epoch": 36.1117903930131,
+ "grad_norm": 0.0189144778996706,
+ "learning_rate": 0.0006,
+ "loss": 4.038427352905273,
+ "step": 2600
+ },
+ {
+ "epoch": 36.12576419213974,
+ "grad_norm": 0.01754768192768097,
+ "learning_rate": 0.0006,
+ "loss": 4.010427474975586,
+ "step": 2601
+ },
+ {
+ "epoch": 36.13973799126637,
+ "grad_norm": 0.016862990334630013,
+ "learning_rate": 0.0006,
+ "loss": 4.035970211029053,
+ "step": 2602
+ },
+ {
+ "epoch": 36.15371179039301,
+ "grad_norm": 0.019616752862930298,
+ "learning_rate": 0.0006,
+ "loss": 4.079898834228516,
+ "step": 2603
+ },
+ {
+ "epoch": 36.16768558951965,
+ "grad_norm": 0.020131465047597885,
+ "learning_rate": 0.0006,
+ "loss": 4.144684791564941,
+ "step": 2604
+ },
+ {
+ "epoch": 36.18165938864629,
+ "grad_norm": 0.02072521299123764,
+ "learning_rate": 0.0006,
+ "loss": 4.013660430908203,
+ "step": 2605
+ },
+ {
+ "epoch": 36.19563318777293,
+ "grad_norm": 0.01924615167081356,
+ "learning_rate": 0.0006,
+ "loss": 4.071398735046387,
+ "step": 2606
+ },
+ {
+ "epoch": 36.209606986899566,
+ "grad_norm": 0.02013954520225525,
+ "learning_rate": 0.0006,
+ "loss": 4.127094268798828,
+ "step": 2607
+ },
+ {
+ "epoch": 36.223580786026204,
+ "grad_norm": 0.02042008936405182,
+ "learning_rate": 0.0006,
+ "loss": 4.154261589050293,
+ "step": 2608
+ },
+ {
+ "epoch": 36.237554585152836,
+ "grad_norm": 0.021998876705765724,
+ "learning_rate": 0.0006,
+ "loss": 4.204150199890137,
+ "step": 2609
+ },
+ {
+ "epoch": 36.251528384279474,
+ "grad_norm": 0.022946683689951897,
+ "learning_rate": 0.0006,
+ "loss": 4.013632774353027,
+ "step": 2610
+ },
+ {
+ "epoch": 36.26550218340611,
+ "grad_norm": 0.02061830274760723,
+ "learning_rate": 0.0006,
+ "loss": 4.171084403991699,
+ "step": 2611
+ },
+ {
+ "epoch": 36.27947598253275,
+ "grad_norm": 0.01883152686059475,
+ "learning_rate": 0.0006,
+ "loss": 4.149848461151123,
+ "step": 2612
+ },
+ {
+ "epoch": 36.29344978165939,
+ "grad_norm": 0.019945282489061356,
+ "learning_rate": 0.0006,
+ "loss": 4.174656867980957,
+ "step": 2613
+ },
+ {
+ "epoch": 36.30742358078603,
+ "grad_norm": 0.020751051604747772,
+ "learning_rate": 0.0006,
+ "loss": 4.080747127532959,
+ "step": 2614
+ },
+ {
+ "epoch": 36.32139737991266,
+ "grad_norm": 0.01879691146314144,
+ "learning_rate": 0.0006,
+ "loss": 4.057356357574463,
+ "step": 2615
+ },
+ {
+ "epoch": 36.3353711790393,
+ "grad_norm": 0.019037071615457535,
+ "learning_rate": 0.0006,
+ "loss": 4.204607009887695,
+ "step": 2616
+ },
+ {
+ "epoch": 36.34934497816594,
+ "grad_norm": 0.01872078701853752,
+ "learning_rate": 0.0006,
+ "loss": 4.050586700439453,
+ "step": 2617
+ },
+ {
+ "epoch": 36.36331877729258,
+ "grad_norm": 0.018092811107635498,
+ "learning_rate": 0.0006,
+ "loss": 4.102832794189453,
+ "step": 2618
+ },
+ {
+ "epoch": 36.377292576419215,
+ "grad_norm": 0.017010487616062164,
+ "learning_rate": 0.0006,
+ "loss": 4.191788673400879,
+ "step": 2619
+ },
+ {
+ "epoch": 36.391266375545854,
+ "grad_norm": 0.01691143773496151,
+ "learning_rate": 0.0006,
+ "loss": 4.160456657409668,
+ "step": 2620
+ },
+ {
+ "epoch": 36.40524017467249,
+ "grad_norm": 0.017046816647052765,
+ "learning_rate": 0.0006,
+ "loss": 4.073828220367432,
+ "step": 2621
+ },
+ {
+ "epoch": 36.419213973799124,
+ "grad_norm": 0.01707981713116169,
+ "learning_rate": 0.0006,
+ "loss": 4.0980095863342285,
+ "step": 2622
+ },
+ {
+ "epoch": 36.43318777292576,
+ "grad_norm": 0.01630452275276184,
+ "learning_rate": 0.0006,
+ "loss": 4.101321220397949,
+ "step": 2623
+ },
+ {
+ "epoch": 36.4471615720524,
+ "grad_norm": 0.01576915942132473,
+ "learning_rate": 0.0006,
+ "loss": 4.158249855041504,
+ "step": 2624
+ },
+ {
+ "epoch": 36.46113537117904,
+ "grad_norm": 0.016179712489247322,
+ "learning_rate": 0.0006,
+ "loss": 4.077852249145508,
+ "step": 2625
+ },
+ {
+ "epoch": 36.47510917030568,
+ "grad_norm": 0.01609169878065586,
+ "learning_rate": 0.0006,
+ "loss": 4.187317848205566,
+ "step": 2626
+ },
+ {
+ "epoch": 36.48908296943232,
+ "grad_norm": 0.017148254439234734,
+ "learning_rate": 0.0006,
+ "loss": 4.109071731567383,
+ "step": 2627
+ },
+ {
+ "epoch": 36.50305676855895,
+ "grad_norm": 0.015842624008655548,
+ "learning_rate": 0.0006,
+ "loss": 4.145096302032471,
+ "step": 2628
+ },
+ {
+ "epoch": 36.51703056768559,
+ "grad_norm": 0.014906194061040878,
+ "learning_rate": 0.0006,
+ "loss": 4.028440475463867,
+ "step": 2629
+ },
+ {
+ "epoch": 36.531004366812226,
+ "grad_norm": 0.01485038734972477,
+ "learning_rate": 0.0006,
+ "loss": 4.175201416015625,
+ "step": 2630
+ },
+ {
+ "epoch": 36.544978165938865,
+ "grad_norm": 0.015834150835871696,
+ "learning_rate": 0.0006,
+ "loss": 4.076434135437012,
+ "step": 2631
+ },
+ {
+ "epoch": 36.5589519650655,
+ "grad_norm": 0.015411358326673508,
+ "learning_rate": 0.0006,
+ "loss": 4.03912878036499,
+ "step": 2632
+ },
+ {
+ "epoch": 36.57292576419214,
+ "grad_norm": 0.014762846753001213,
+ "learning_rate": 0.0006,
+ "loss": 4.019827365875244,
+ "step": 2633
+ },
+ {
+ "epoch": 36.58689956331878,
+ "grad_norm": 0.015269497409462929,
+ "learning_rate": 0.0006,
+ "loss": 4.116957187652588,
+ "step": 2634
+ },
+ {
+ "epoch": 36.60087336244541,
+ "grad_norm": 0.015171617269515991,
+ "learning_rate": 0.0006,
+ "loss": 4.2069926261901855,
+ "step": 2635
+ },
+ {
+ "epoch": 36.61484716157205,
+ "grad_norm": 0.014770902693271637,
+ "learning_rate": 0.0006,
+ "loss": 4.092021942138672,
+ "step": 2636
+ },
+ {
+ "epoch": 36.62882096069869,
+ "grad_norm": 0.014957531355321407,
+ "learning_rate": 0.0006,
+ "loss": 4.16819953918457,
+ "step": 2637
+ },
+ {
+ "epoch": 36.64279475982533,
+ "grad_norm": 0.014774451963603497,
+ "learning_rate": 0.0006,
+ "loss": 4.046073913574219,
+ "step": 2638
+ },
+ {
+ "epoch": 36.65676855895197,
+ "grad_norm": 0.014320445246994495,
+ "learning_rate": 0.0006,
+ "loss": 3.977792501449585,
+ "step": 2639
+ },
+ {
+ "epoch": 36.670742358078606,
+ "grad_norm": 0.01413130946457386,
+ "learning_rate": 0.0006,
+ "loss": 4.130649566650391,
+ "step": 2640
+ },
+ {
+ "epoch": 36.68471615720524,
+ "grad_norm": 0.014336307533085346,
+ "learning_rate": 0.0006,
+ "loss": 4.161357402801514,
+ "step": 2641
+ },
+ {
+ "epoch": 36.698689956331876,
+ "grad_norm": 0.014516931027173996,
+ "learning_rate": 0.0006,
+ "loss": 4.14248514175415,
+ "step": 2642
+ },
+ {
+ "epoch": 36.712663755458514,
+ "grad_norm": 0.014377126470208168,
+ "learning_rate": 0.0006,
+ "loss": 4.062711715698242,
+ "step": 2643
+ },
+ {
+ "epoch": 36.72663755458515,
+ "grad_norm": 0.014332668855786324,
+ "learning_rate": 0.0006,
+ "loss": 4.06423807144165,
+ "step": 2644
+ },
+ {
+ "epoch": 36.74061135371179,
+ "grad_norm": 0.014030571095645428,
+ "learning_rate": 0.0006,
+ "loss": 4.178613662719727,
+ "step": 2645
+ },
+ {
+ "epoch": 36.75458515283843,
+ "grad_norm": 0.014376776292920113,
+ "learning_rate": 0.0006,
+ "loss": 4.073130130767822,
+ "step": 2646
+ },
+ {
+ "epoch": 36.76855895196506,
+ "grad_norm": 0.013827716931700706,
+ "learning_rate": 0.0006,
+ "loss": 4.043869972229004,
+ "step": 2647
+ },
+ {
+ "epoch": 36.7825327510917,
+ "grad_norm": 0.015548762865364552,
+ "learning_rate": 0.0006,
+ "loss": 4.134049415588379,
+ "step": 2648
+ },
+ {
+ "epoch": 36.79650655021834,
+ "grad_norm": 0.017041970044374466,
+ "learning_rate": 0.0006,
+ "loss": 4.205024242401123,
+ "step": 2649
+ },
+ {
+ "epoch": 36.81048034934498,
+ "grad_norm": 0.017638275399804115,
+ "learning_rate": 0.0006,
+ "loss": 4.199924468994141,
+ "step": 2650
+ },
+ {
+ "epoch": 36.82445414847162,
+ "grad_norm": 0.017771627753973007,
+ "learning_rate": 0.0006,
+ "loss": 4.15706729888916,
+ "step": 2651
+ },
+ {
+ "epoch": 36.838427947598255,
+ "grad_norm": 0.015933562070131302,
+ "learning_rate": 0.0006,
+ "loss": 4.220742225646973,
+ "step": 2652
+ },
+ {
+ "epoch": 36.852401746724894,
+ "grad_norm": 0.01668144389986992,
+ "learning_rate": 0.0006,
+ "loss": 4.161426544189453,
+ "step": 2653
+ },
+ {
+ "epoch": 36.866375545851525,
+ "grad_norm": 0.018014680594205856,
+ "learning_rate": 0.0006,
+ "loss": 4.183046817779541,
+ "step": 2654
+ },
+ {
+ "epoch": 36.880349344978164,
+ "grad_norm": 0.018808670341968536,
+ "learning_rate": 0.0006,
+ "loss": 4.105293273925781,
+ "step": 2655
+ },
+ {
+ "epoch": 36.8943231441048,
+ "grad_norm": 0.018003476783633232,
+ "learning_rate": 0.0006,
+ "loss": 4.103171348571777,
+ "step": 2656
+ },
+ {
+ "epoch": 36.90829694323144,
+ "grad_norm": 0.016865914687514305,
+ "learning_rate": 0.0006,
+ "loss": 4.125634670257568,
+ "step": 2657
+ },
+ {
+ "epoch": 36.92227074235808,
+ "grad_norm": 0.018371153622865677,
+ "learning_rate": 0.0006,
+ "loss": 4.192461013793945,
+ "step": 2658
+ },
+ {
+ "epoch": 36.93624454148472,
+ "grad_norm": 0.019045377150177956,
+ "learning_rate": 0.0006,
+ "loss": 4.198105812072754,
+ "step": 2659
+ },
+ {
+ "epoch": 36.95021834061135,
+ "grad_norm": 0.01594168320298195,
+ "learning_rate": 0.0006,
+ "loss": 4.169788360595703,
+ "step": 2660
+ },
+ {
+ "epoch": 36.96419213973799,
+ "grad_norm": 0.015645820647478104,
+ "learning_rate": 0.0006,
+ "loss": 4.047277450561523,
+ "step": 2661
+ },
+ {
+ "epoch": 36.97816593886463,
+ "grad_norm": 0.016313156113028526,
+ "learning_rate": 0.0006,
+ "loss": 4.087392807006836,
+ "step": 2662
+ },
+ {
+ "epoch": 36.992139737991266,
+ "grad_norm": 0.017050625756382942,
+ "learning_rate": 0.0006,
+ "loss": 4.0908002853393555,
+ "step": 2663
+ },
+ {
+ "epoch": 37.0,
+ "grad_norm": 0.017730968073010445,
+ "learning_rate": 0.0006,
+ "loss": 4.110957622528076,
+ "step": 2664
+ },
+ {
+ "epoch": 37.0,
+ "eval_loss": 4.563793182373047,
+ "eval_runtime": 56.5922,
+ "eval_samples_per_second": 43.151,
+ "eval_steps_per_second": 1.361,
+ "step": 2664
+ },
+ {
+ "epoch": 37.01397379912664,
+ "grad_norm": 0.015687910839915276,
+ "learning_rate": 0.0006,
+ "loss": 4.143589019775391,
+ "step": 2665
+ },
+ {
+ "epoch": 37.02794759825328,
+ "grad_norm": 0.017451366409659386,
+ "learning_rate": 0.0006,
+ "loss": 4.0326666831970215,
+ "step": 2666
+ },
+ {
+ "epoch": 37.041921397379916,
+ "grad_norm": 0.01909538544714451,
+ "learning_rate": 0.0006,
+ "loss": 3.9706249237060547,
+ "step": 2667
+ },
+ {
+ "epoch": 37.05589519650655,
+ "grad_norm": 0.01988379657268524,
+ "learning_rate": 0.0006,
+ "loss": 3.9531009197235107,
+ "step": 2668
+ },
+ {
+ "epoch": 37.069868995633186,
+ "grad_norm": 0.02035115472972393,
+ "learning_rate": 0.0006,
+ "loss": 3.962214469909668,
+ "step": 2669
+ },
+ {
+ "epoch": 37.083842794759825,
+ "grad_norm": 0.01952001079916954,
+ "learning_rate": 0.0006,
+ "loss": 4.182262897491455,
+ "step": 2670
+ },
+ {
+ "epoch": 37.09781659388646,
+ "grad_norm": 0.01854325644671917,
+ "learning_rate": 0.0006,
+ "loss": 4.127663612365723,
+ "step": 2671
+ },
+ {
+ "epoch": 37.1117903930131,
+ "grad_norm": 0.017757616937160492,
+ "learning_rate": 0.0006,
+ "loss": 4.100804805755615,
+ "step": 2672
+ },
+ {
+ "epoch": 37.12576419213974,
+ "grad_norm": 0.016712479293346405,
+ "learning_rate": 0.0006,
+ "loss": 4.026284217834473,
+ "step": 2673
+ },
+ {
+ "epoch": 37.13973799126637,
+ "grad_norm": 0.0168139711022377,
+ "learning_rate": 0.0006,
+ "loss": 4.183882713317871,
+ "step": 2674
+ },
+ {
+ "epoch": 37.15371179039301,
+ "grad_norm": 0.017060423269867897,
+ "learning_rate": 0.0006,
+ "loss": 4.131832599639893,
+ "step": 2675
+ },
+ {
+ "epoch": 37.16768558951965,
+ "grad_norm": 0.01726631447672844,
+ "learning_rate": 0.0006,
+ "loss": 3.9156057834625244,
+ "step": 2676
+ },
+ {
+ "epoch": 37.18165938864629,
+ "grad_norm": 0.016348622739315033,
+ "learning_rate": 0.0006,
+ "loss": 4.1731367111206055,
+ "step": 2677
+ },
+ {
+ "epoch": 37.19563318777293,
+ "grad_norm": 0.016640059649944305,
+ "learning_rate": 0.0006,
+ "loss": 4.10997200012207,
+ "step": 2678
+ },
+ {
+ "epoch": 37.209606986899566,
+ "grad_norm": 0.01601342298090458,
+ "learning_rate": 0.0006,
+ "loss": 4.081145286560059,
+ "step": 2679
+ },
+ {
+ "epoch": 37.223580786026204,
+ "grad_norm": 0.01735631190240383,
+ "learning_rate": 0.0006,
+ "loss": 4.202108860015869,
+ "step": 2680
+ },
+ {
+ "epoch": 37.237554585152836,
+ "grad_norm": 0.019164685159921646,
+ "learning_rate": 0.0006,
+ "loss": 4.104109764099121,
+ "step": 2681
+ },
+ {
+ "epoch": 37.251528384279474,
+ "grad_norm": 0.01881759613752365,
+ "learning_rate": 0.0006,
+ "loss": 4.145532131195068,
+ "step": 2682
+ },
+ {
+ "epoch": 37.26550218340611,
+ "grad_norm": 0.017656555399298668,
+ "learning_rate": 0.0006,
+ "loss": 4.012782096862793,
+ "step": 2683
+ },
+ {
+ "epoch": 37.27947598253275,
+ "grad_norm": 0.01624823734164238,
+ "learning_rate": 0.0006,
+ "loss": 4.001727104187012,
+ "step": 2684
+ },
+ {
+ "epoch": 37.29344978165939,
+ "grad_norm": 0.016973139718174934,
+ "learning_rate": 0.0006,
+ "loss": 4.095399856567383,
+ "step": 2685
+ },
+ {
+ "epoch": 37.30742358078603,
+ "grad_norm": 0.015503671951591969,
+ "learning_rate": 0.0006,
+ "loss": 4.034431457519531,
+ "step": 2686
+ },
+ {
+ "epoch": 37.32139737991266,
+ "grad_norm": 0.015370250679552555,
+ "learning_rate": 0.0006,
+ "loss": 4.075836181640625,
+ "step": 2687
+ },
+ {
+ "epoch": 37.3353711790393,
+ "grad_norm": 0.01564021222293377,
+ "learning_rate": 0.0006,
+ "loss": 4.125202178955078,
+ "step": 2688
+ },
+ {
+ "epoch": 37.34934497816594,
+ "grad_norm": 0.016717975959181786,
+ "learning_rate": 0.0006,
+ "loss": 4.153017044067383,
+ "step": 2689
+ },
+ {
+ "epoch": 37.36331877729258,
+ "grad_norm": 0.015263617038726807,
+ "learning_rate": 0.0006,
+ "loss": 4.021814823150635,
+ "step": 2690
+ },
+ {
+ "epoch": 37.377292576419215,
+ "grad_norm": 0.016497142612934113,
+ "learning_rate": 0.0006,
+ "loss": 4.069095611572266,
+ "step": 2691
+ },
+ {
+ "epoch": 37.391266375545854,
+ "grad_norm": 0.01669355481863022,
+ "learning_rate": 0.0006,
+ "loss": 4.174587726593018,
+ "step": 2692
+ },
+ {
+ "epoch": 37.40524017467249,
+ "grad_norm": 0.019695773720741272,
+ "learning_rate": 0.0006,
+ "loss": 4.0073041915893555,
+ "step": 2693
+ },
+ {
+ "epoch": 37.419213973799124,
+ "grad_norm": 0.023720644414424896,
+ "learning_rate": 0.0006,
+ "loss": 4.111948013305664,
+ "step": 2694
+ },
+ {
+ "epoch": 37.43318777292576,
+ "grad_norm": 0.02399350143969059,
+ "learning_rate": 0.0006,
+ "loss": 4.081164360046387,
+ "step": 2695
+ },
+ {
+ "epoch": 37.4471615720524,
+ "grad_norm": 0.019630271941423416,
+ "learning_rate": 0.0006,
+ "loss": 4.065731048583984,
+ "step": 2696
+ },
+ {
+ "epoch": 37.46113537117904,
+ "grad_norm": 0.019679008051753044,
+ "learning_rate": 0.0006,
+ "loss": 3.993852376937866,
+ "step": 2697
+ },
+ {
+ "epoch": 37.47510917030568,
+ "grad_norm": 0.020597299560904503,
+ "learning_rate": 0.0006,
+ "loss": 4.275555610656738,
+ "step": 2698
+ },
+ {
+ "epoch": 37.48908296943232,
+ "grad_norm": 0.01983889564871788,
+ "learning_rate": 0.0006,
+ "loss": 4.254708290100098,
+ "step": 2699
+ },
+ {
+ "epoch": 37.50305676855895,
+ "grad_norm": 0.019150668755173683,
+ "learning_rate": 0.0006,
+ "loss": 4.097885608673096,
+ "step": 2700
+ },
+ {
+ "epoch": 37.51703056768559,
+ "grad_norm": 0.018847715109586716,
+ "learning_rate": 0.0006,
+ "loss": 4.054081439971924,
+ "step": 2701
+ },
+ {
+ "epoch": 37.531004366812226,
+ "grad_norm": 0.019607288762927055,
+ "learning_rate": 0.0006,
+ "loss": 3.9891440868377686,
+ "step": 2702
+ },
+ {
+ "epoch": 37.544978165938865,
+ "grad_norm": 0.018863150849938393,
+ "learning_rate": 0.0006,
+ "loss": 4.084877014160156,
+ "step": 2703
+ },
+ {
+ "epoch": 37.5589519650655,
+ "grad_norm": 0.017340239137411118,
+ "learning_rate": 0.0006,
+ "loss": 4.113290786743164,
+ "step": 2704
+ },
+ {
+ "epoch": 37.57292576419214,
+ "grad_norm": 0.015628693625330925,
+ "learning_rate": 0.0006,
+ "loss": 4.1483001708984375,
+ "step": 2705
+ },
+ {
+ "epoch": 37.58689956331878,
+ "grad_norm": 0.016493283212184906,
+ "learning_rate": 0.0006,
+ "loss": 4.1040143966674805,
+ "step": 2706
+ },
+ {
+ "epoch": 37.60087336244541,
+ "grad_norm": 0.016732443124055862,
+ "learning_rate": 0.0006,
+ "loss": 4.07205057144165,
+ "step": 2707
+ },
+ {
+ "epoch": 37.61484716157205,
+ "grad_norm": 0.017176156863570213,
+ "learning_rate": 0.0006,
+ "loss": 4.1279754638671875,
+ "step": 2708
+ },
+ {
+ "epoch": 37.62882096069869,
+ "grad_norm": 0.015730608254671097,
+ "learning_rate": 0.0006,
+ "loss": 4.086658000946045,
+ "step": 2709
+ },
+ {
+ "epoch": 37.64279475982533,
+ "grad_norm": 0.016111090779304504,
+ "learning_rate": 0.0006,
+ "loss": 4.064335823059082,
+ "step": 2710
+ },
+ {
+ "epoch": 37.65676855895197,
+ "grad_norm": 0.01769379712641239,
+ "learning_rate": 0.0006,
+ "loss": 4.116047382354736,
+ "step": 2711
+ },
+ {
+ "epoch": 37.670742358078606,
+ "grad_norm": 0.01683102548122406,
+ "learning_rate": 0.0006,
+ "loss": 4.095986843109131,
+ "step": 2712
+ },
+ {
+ "epoch": 37.68471615720524,
+ "grad_norm": 0.015726450830698013,
+ "learning_rate": 0.0006,
+ "loss": 4.10775089263916,
+ "step": 2713
+ },
+ {
+ "epoch": 37.698689956331876,
+ "grad_norm": 0.014760235324501991,
+ "learning_rate": 0.0006,
+ "loss": 4.1234893798828125,
+ "step": 2714
+ },
+ {
+ "epoch": 37.712663755458514,
+ "grad_norm": 0.014933438040316105,
+ "learning_rate": 0.0006,
+ "loss": 4.079200744628906,
+ "step": 2715
+ },
+ {
+ "epoch": 37.72663755458515,
+ "grad_norm": 0.015723997727036476,
+ "learning_rate": 0.0006,
+ "loss": 4.07865047454834,
+ "step": 2716
+ },
+ {
+ "epoch": 37.74061135371179,
+ "grad_norm": 0.014539001509547234,
+ "learning_rate": 0.0006,
+ "loss": 4.036242485046387,
+ "step": 2717
+ },
+ {
+ "epoch": 37.75458515283843,
+ "grad_norm": 0.014844979159533978,
+ "learning_rate": 0.0006,
+ "loss": 4.1051483154296875,
+ "step": 2718
+ },
+ {
+ "epoch": 37.76855895196506,
+ "grad_norm": 0.015939844772219658,
+ "learning_rate": 0.0006,
+ "loss": 4.017253875732422,
+ "step": 2719
+ },
+ {
+ "epoch": 37.7825327510917,
+ "grad_norm": 0.014771555550396442,
+ "learning_rate": 0.0006,
+ "loss": 4.0694661140441895,
+ "step": 2720
+ },
+ {
+ "epoch": 37.79650655021834,
+ "grad_norm": 0.015777921304106712,
+ "learning_rate": 0.0006,
+ "loss": 3.947587490081787,
+ "step": 2721
+ },
+ {
+ "epoch": 37.81048034934498,
+ "grad_norm": 0.017451291903853416,
+ "learning_rate": 0.0006,
+ "loss": 4.093625068664551,
+ "step": 2722
+ },
+ {
+ "epoch": 37.82445414847162,
+ "grad_norm": 0.016829246655106544,
+ "learning_rate": 0.0006,
+ "loss": 4.09030818939209,
+ "step": 2723
+ },
+ {
+ "epoch": 37.838427947598255,
+ "grad_norm": 0.016984397545456886,
+ "learning_rate": 0.0006,
+ "loss": 4.140626907348633,
+ "step": 2724
+ },
+ {
+ "epoch": 37.852401746724894,
+ "grad_norm": 0.01732603646814823,
+ "learning_rate": 0.0006,
+ "loss": 4.128276824951172,
+ "step": 2725
+ },
+ {
+ "epoch": 37.866375545851525,
+ "grad_norm": 0.01744619384407997,
+ "learning_rate": 0.0006,
+ "loss": 4.103130340576172,
+ "step": 2726
+ },
+ {
+ "epoch": 37.880349344978164,
+ "grad_norm": 0.016331320628523827,
+ "learning_rate": 0.0006,
+ "loss": 4.037951469421387,
+ "step": 2727
+ },
+ {
+ "epoch": 37.8943231441048,
+ "grad_norm": 0.015919430181384087,
+ "learning_rate": 0.0006,
+ "loss": 4.121513366699219,
+ "step": 2728
+ },
+ {
+ "epoch": 37.90829694323144,
+ "grad_norm": 0.016153695061802864,
+ "learning_rate": 0.0006,
+ "loss": 4.099384307861328,
+ "step": 2729
+ },
+ {
+ "epoch": 37.92227074235808,
+ "grad_norm": 0.01628800667822361,
+ "learning_rate": 0.0006,
+ "loss": 4.092325687408447,
+ "step": 2730
+ },
+ {
+ "epoch": 37.93624454148472,
+ "grad_norm": 0.017285684123635292,
+ "learning_rate": 0.0006,
+ "loss": 3.9828906059265137,
+ "step": 2731
+ },
+ {
+ "epoch": 37.95021834061135,
+ "grad_norm": 0.016047554090619087,
+ "learning_rate": 0.0006,
+ "loss": 4.100419044494629,
+ "step": 2732
+ },
+ {
+ "epoch": 37.96419213973799,
+ "grad_norm": 0.01562986709177494,
+ "learning_rate": 0.0006,
+ "loss": 4.065832614898682,
+ "step": 2733
+ },
+ {
+ "epoch": 37.97816593886463,
+ "grad_norm": 0.01677163504064083,
+ "learning_rate": 0.0006,
+ "loss": 4.077518463134766,
+ "step": 2734
+ },
+ {
+ "epoch": 37.992139737991266,
+ "grad_norm": 0.01755492016673088,
+ "learning_rate": 0.0006,
+ "loss": 4.1076154708862305,
+ "step": 2735
+ },
+ {
+ "epoch": 38.0,
+ "grad_norm": 0.01929762214422226,
+ "learning_rate": 0.0006,
+ "loss": 4.09603214263916,
+ "step": 2736
+ },
+ {
+ "epoch": 38.0,
+ "eval_loss": 4.582442283630371,
+ "eval_runtime": 62.8649,
+ "eval_samples_per_second": 38.845,
+ "eval_steps_per_second": 1.225,
+ "step": 2736
+ },
+ {
+ "epoch": 38.01397379912664,
+ "grad_norm": 0.019129136577248573,
+ "learning_rate": 0.0006,
+ "loss": 4.095005989074707,
+ "step": 2737
+ },
+ {
+ "epoch": 38.02794759825328,
+ "grad_norm": 0.019167408347129822,
+ "learning_rate": 0.0006,
+ "loss": 4.054426670074463,
+ "step": 2738
+ },
+ {
+ "epoch": 38.041921397379916,
+ "grad_norm": 0.01789460889995098,
+ "learning_rate": 0.0006,
+ "loss": 4.0434417724609375,
+ "step": 2739
+ },
+ {
+ "epoch": 38.05589519650655,
+ "grad_norm": 0.015947027131915092,
+ "learning_rate": 0.0006,
+ "loss": 4.074516296386719,
+ "step": 2740
+ },
+ {
+ "epoch": 38.069868995633186,
+ "grad_norm": 0.016667252406477928,
+ "learning_rate": 0.0006,
+ "loss": 4.108545303344727,
+ "step": 2741
+ },
+ {
+ "epoch": 38.083842794759825,
+ "grad_norm": 0.016014399006962776,
+ "learning_rate": 0.0006,
+ "loss": 4.026925086975098,
+ "step": 2742
+ },
+ {
+ "epoch": 38.09781659388646,
+ "grad_norm": 0.015397695824503899,
+ "learning_rate": 0.0006,
+ "loss": 4.163970470428467,
+ "step": 2743
+ },
+ {
+ "epoch": 38.1117903930131,
+ "grad_norm": 0.01563289947807789,
+ "learning_rate": 0.0006,
+ "loss": 4.116578578948975,
+ "step": 2744
+ },
+ {
+ "epoch": 38.12576419213974,
+ "grad_norm": 0.015512318350374699,
+ "learning_rate": 0.0006,
+ "loss": 4.096311569213867,
+ "step": 2745
+ },
+ {
+ "epoch": 38.13973799126637,
+ "grad_norm": 0.015426642261445522,
+ "learning_rate": 0.0006,
+ "loss": 4.017065048217773,
+ "step": 2746
+ },
+ {
+ "epoch": 38.15371179039301,
+ "grad_norm": 0.015115435235202312,
+ "learning_rate": 0.0006,
+ "loss": 4.005434989929199,
+ "step": 2747
+ },
+ {
+ "epoch": 38.16768558951965,
+ "grad_norm": 0.015235894359648228,
+ "learning_rate": 0.0006,
+ "loss": 3.982743263244629,
+ "step": 2748
+ },
+ {
+ "epoch": 38.18165938864629,
+ "grad_norm": 0.016095101833343506,
+ "learning_rate": 0.0006,
+ "loss": 4.0849385261535645,
+ "step": 2749
+ },
+ {
+ "epoch": 38.19563318777293,
+ "grad_norm": 0.015201197937130928,
+ "learning_rate": 0.0006,
+ "loss": 4.103863716125488,
+ "step": 2750
+ },
+ {
+ "epoch": 38.209606986899566,
+ "grad_norm": 0.015687478706240654,
+ "learning_rate": 0.0006,
+ "loss": 4.12074089050293,
+ "step": 2751
+ },
+ {
+ "epoch": 38.223580786026204,
+ "grad_norm": 0.015201589092612267,
+ "learning_rate": 0.0006,
+ "loss": 4.056872367858887,
+ "step": 2752
+ },
+ {
+ "epoch": 38.237554585152836,
+ "grad_norm": 0.015154894441366196,
+ "learning_rate": 0.0006,
+ "loss": 4.089670181274414,
+ "step": 2753
+ },
+ {
+ "epoch": 38.251528384279474,
+ "grad_norm": 0.01586608588695526,
+ "learning_rate": 0.0006,
+ "loss": 4.155317306518555,
+ "step": 2754
+ },
+ {
+ "epoch": 38.26550218340611,
+ "grad_norm": 0.01609654724597931,
+ "learning_rate": 0.0006,
+ "loss": 3.995800495147705,
+ "step": 2755
+ },
+ {
+ "epoch": 38.27947598253275,
+ "grad_norm": 0.015495185740292072,
+ "learning_rate": 0.0006,
+ "loss": 3.995769739151001,
+ "step": 2756
+ },
+ {
+ "epoch": 38.29344978165939,
+ "grad_norm": 0.017042381688952446,
+ "learning_rate": 0.0006,
+ "loss": 4.132832050323486,
+ "step": 2757
+ },
+ {
+ "epoch": 38.30742358078603,
+ "grad_norm": 0.018706614151597023,
+ "learning_rate": 0.0006,
+ "loss": 4.038820743560791,
+ "step": 2758
+ },
+ {
+ "epoch": 38.32139737991266,
+ "grad_norm": 0.018316026777029037,
+ "learning_rate": 0.0006,
+ "loss": 3.972052574157715,
+ "step": 2759
+ },
+ {
+ "epoch": 38.3353711790393,
+ "grad_norm": 0.018375389277935028,
+ "learning_rate": 0.0006,
+ "loss": 4.152097702026367,
+ "step": 2760
+ },
+ {
+ "epoch": 38.34934497816594,
+ "grad_norm": 0.016879886388778687,
+ "learning_rate": 0.0006,
+ "loss": 4.020869731903076,
+ "step": 2761
+ },
+ {
+ "epoch": 38.36331877729258,
+ "grad_norm": 0.015728183090686798,
+ "learning_rate": 0.0006,
+ "loss": 4.008016586303711,
+ "step": 2762
+ },
+ {
+ "epoch": 38.377292576419215,
+ "grad_norm": 0.016147758811712265,
+ "learning_rate": 0.0006,
+ "loss": 4.035854339599609,
+ "step": 2763
+ },
+ {
+ "epoch": 38.391266375545854,
+ "grad_norm": 0.017774797976017,
+ "learning_rate": 0.0006,
+ "loss": 4.122506141662598,
+ "step": 2764
+ },
+ {
+ "epoch": 38.40524017467249,
+ "grad_norm": 0.016590069979429245,
+ "learning_rate": 0.0006,
+ "loss": 4.082847595214844,
+ "step": 2765
+ },
+ {
+ "epoch": 38.419213973799124,
+ "grad_norm": 0.016544528305530548,
+ "learning_rate": 0.0006,
+ "loss": 4.184270858764648,
+ "step": 2766
+ },
+ {
+ "epoch": 38.43318777292576,
+ "grad_norm": 0.017972925677895546,
+ "learning_rate": 0.0006,
+ "loss": 3.987558603286743,
+ "step": 2767
+ },
+ {
+ "epoch": 38.4471615720524,
+ "grad_norm": 0.018562227487564087,
+ "learning_rate": 0.0006,
+ "loss": 4.1193461418151855,
+ "step": 2768
+ },
+ {
+ "epoch": 38.46113537117904,
+ "grad_norm": 0.018896685913205147,
+ "learning_rate": 0.0006,
+ "loss": 4.057808876037598,
+ "step": 2769
+ },
+ {
+ "epoch": 38.47510917030568,
+ "grad_norm": 0.016793809831142426,
+ "learning_rate": 0.0006,
+ "loss": 4.069523334503174,
+ "step": 2770
+ },
+ {
+ "epoch": 38.48908296943232,
+ "grad_norm": 0.01698676310479641,
+ "learning_rate": 0.0006,
+ "loss": 3.997624158859253,
+ "step": 2771
+ },
+ {
+ "epoch": 38.50305676855895,
+ "grad_norm": 0.016784202307462692,
+ "learning_rate": 0.0006,
+ "loss": 4.174795627593994,
+ "step": 2772
+ },
+ {
+ "epoch": 38.51703056768559,
+ "grad_norm": 0.016785014420747757,
+ "learning_rate": 0.0006,
+ "loss": 3.999321222305298,
+ "step": 2773
+ },
+ {
+ "epoch": 38.531004366812226,
+ "grad_norm": 0.018615184351801872,
+ "learning_rate": 0.0006,
+ "loss": 4.191237926483154,
+ "step": 2774
+ },
+ {
+ "epoch": 38.544978165938865,
+ "grad_norm": 0.019090790301561356,
+ "learning_rate": 0.0006,
+ "loss": 4.106966018676758,
+ "step": 2775
+ },
+ {
+ "epoch": 38.5589519650655,
+ "grad_norm": 0.018298320472240448,
+ "learning_rate": 0.0006,
+ "loss": 4.063363552093506,
+ "step": 2776
+ },
+ {
+ "epoch": 38.57292576419214,
+ "grad_norm": 0.018541090190410614,
+ "learning_rate": 0.0006,
+ "loss": 4.101193428039551,
+ "step": 2777
+ },
+ {
+ "epoch": 38.58689956331878,
+ "grad_norm": 0.01849970407783985,
+ "learning_rate": 0.0006,
+ "loss": 4.071335792541504,
+ "step": 2778
+ },
+ {
+ "epoch": 38.60087336244541,
+ "grad_norm": 0.018309077247977257,
+ "learning_rate": 0.0006,
+ "loss": 4.245305061340332,
+ "step": 2779
+ },
+ {
+ "epoch": 38.61484716157205,
+ "grad_norm": 0.017681840807199478,
+ "learning_rate": 0.0006,
+ "loss": 4.081173419952393,
+ "step": 2780
+ },
+ {
+ "epoch": 38.62882096069869,
+ "grad_norm": 0.016286224126815796,
+ "learning_rate": 0.0006,
+ "loss": 4.024351596832275,
+ "step": 2781
+ },
+ {
+ "epoch": 38.64279475982533,
+ "grad_norm": 0.015837235376238823,
+ "learning_rate": 0.0006,
+ "loss": 4.028205394744873,
+ "step": 2782
+ },
+ {
+ "epoch": 38.65676855895197,
+ "grad_norm": 0.016062596812844276,
+ "learning_rate": 0.0006,
+ "loss": 4.131122589111328,
+ "step": 2783
+ },
+ {
+ "epoch": 38.670742358078606,
+ "grad_norm": 0.01711239665746689,
+ "learning_rate": 0.0006,
+ "loss": 4.124416351318359,
+ "step": 2784
+ },
+ {
+ "epoch": 38.68471615720524,
+ "grad_norm": 0.015304913744330406,
+ "learning_rate": 0.0006,
+ "loss": 4.056122779846191,
+ "step": 2785
+ },
+ {
+ "epoch": 38.698689956331876,
+ "grad_norm": 0.015971887856721878,
+ "learning_rate": 0.0006,
+ "loss": 4.132555961608887,
+ "step": 2786
+ },
+ {
+ "epoch": 38.712663755458514,
+ "grad_norm": 0.015787221491336823,
+ "learning_rate": 0.0006,
+ "loss": 4.051596164703369,
+ "step": 2787
+ },
+ {
+ "epoch": 38.72663755458515,
+ "grad_norm": 0.01605444774031639,
+ "learning_rate": 0.0006,
+ "loss": 4.191808700561523,
+ "step": 2788
+ },
+ {
+ "epoch": 38.74061135371179,
+ "grad_norm": 0.016383009031414986,
+ "learning_rate": 0.0006,
+ "loss": 4.0648417472839355,
+ "step": 2789
+ },
+ {
+ "epoch": 38.75458515283843,
+ "grad_norm": 0.016135554760694504,
+ "learning_rate": 0.0006,
+ "loss": 3.988563060760498,
+ "step": 2790
+ },
+ {
+ "epoch": 38.76855895196506,
+ "grad_norm": 0.016634158790111542,
+ "learning_rate": 0.0006,
+ "loss": 4.133493423461914,
+ "step": 2791
+ },
+ {
+ "epoch": 38.7825327510917,
+ "grad_norm": 0.01688210666179657,
+ "learning_rate": 0.0006,
+ "loss": 4.091277122497559,
+ "step": 2792
+ },
+ {
+ "epoch": 38.79650655021834,
+ "grad_norm": 0.016102122142910957,
+ "learning_rate": 0.0006,
+ "loss": 4.051219940185547,
+ "step": 2793
+ },
+ {
+ "epoch": 38.81048034934498,
+ "grad_norm": 0.016585038974881172,
+ "learning_rate": 0.0006,
+ "loss": 4.058587074279785,
+ "step": 2794
+ },
+ {
+ "epoch": 38.82445414847162,
+ "grad_norm": 0.017700951546430588,
+ "learning_rate": 0.0006,
+ "loss": 4.016030311584473,
+ "step": 2795
+ },
+ {
+ "epoch": 38.838427947598255,
+ "grad_norm": 0.017922203987836838,
+ "learning_rate": 0.0006,
+ "loss": 4.079977989196777,
+ "step": 2796
+ },
+ {
+ "epoch": 38.852401746724894,
+ "grad_norm": 0.017299890518188477,
+ "learning_rate": 0.0006,
+ "loss": 4.062984943389893,
+ "step": 2797
+ },
+ {
+ "epoch": 38.866375545851525,
+ "grad_norm": 0.018648581579327583,
+ "learning_rate": 0.0006,
+ "loss": 4.065717697143555,
+ "step": 2798
+ },
+ {
+ "epoch": 38.880349344978164,
+ "grad_norm": 0.019057920202612877,
+ "learning_rate": 0.0006,
+ "loss": 4.227892875671387,
+ "step": 2799
+ },
+ {
+ "epoch": 38.8943231441048,
+ "grad_norm": 0.018824422731995583,
+ "learning_rate": 0.0006,
+ "loss": 4.116058349609375,
+ "step": 2800
+ },
+ {
+ "epoch": 38.90829694323144,
+ "grad_norm": 0.019084978848695755,
+ "learning_rate": 0.0006,
+ "loss": 4.061677932739258,
+ "step": 2801
+ },
+ {
+ "epoch": 38.92227074235808,
+ "grad_norm": 0.017842981964349747,
+ "learning_rate": 0.0006,
+ "loss": 4.083813667297363,
+ "step": 2802
+ },
+ {
+ "epoch": 38.93624454148472,
+ "grad_norm": 0.01880858838558197,
+ "learning_rate": 0.0006,
+ "loss": 4.152561187744141,
+ "step": 2803
+ },
+ {
+ "epoch": 38.95021834061135,
+ "grad_norm": 0.017400087788701057,
+ "learning_rate": 0.0006,
+ "loss": 4.154635429382324,
+ "step": 2804
+ },
+ {
+ "epoch": 38.96419213973799,
+ "grad_norm": 0.017682507634162903,
+ "learning_rate": 0.0006,
+ "loss": 4.151128768920898,
+ "step": 2805
+ },
+ {
+ "epoch": 38.97816593886463,
+ "grad_norm": 0.015803052112460136,
+ "learning_rate": 0.0006,
+ "loss": 4.10498571395874,
+ "step": 2806
+ },
+ {
+ "epoch": 38.992139737991266,
+ "grad_norm": 0.01582511141896248,
+ "learning_rate": 0.0006,
+ "loss": 4.140159606933594,
+ "step": 2807
+ },
+ {
+ "epoch": 39.0,
+ "grad_norm": 0.01925666816532612,
+ "learning_rate": 0.0006,
+ "loss": 4.126394271850586,
+ "step": 2808
+ },
+ {
+ "epoch": 39.0,
+ "eval_loss": 4.615617275238037,
+ "eval_runtime": 56.8895,
+ "eval_samples_per_second": 42.925,
+ "eval_steps_per_second": 1.354,
+ "step": 2808
+ },
+ {
+ "epoch": 39.01397379912664,
+ "grad_norm": 0.01677853614091873,
+ "learning_rate": 0.0006,
+ "loss": 4.102395057678223,
+ "step": 2809
+ },
+ {
+ "epoch": 39.02794759825328,
+ "grad_norm": 0.017901137471199036,
+ "learning_rate": 0.0006,
+ "loss": 3.976609230041504,
+ "step": 2810
+ },
+ {
+ "epoch": 39.041921397379916,
+ "grad_norm": 0.017978668212890625,
+ "learning_rate": 0.0006,
+ "loss": 3.944416046142578,
+ "step": 2811
+ },
+ {
+ "epoch": 39.05589519650655,
+ "grad_norm": 0.019967492669820786,
+ "learning_rate": 0.0006,
+ "loss": 3.992234468460083,
+ "step": 2812
+ },
+ {
+ "epoch": 39.069868995633186,
+ "grad_norm": 0.018937544897198677,
+ "learning_rate": 0.0006,
+ "loss": 4.028573036193848,
+ "step": 2813
+ },
+ {
+ "epoch": 39.083842794759825,
+ "grad_norm": 0.016852952539920807,
+ "learning_rate": 0.0006,
+ "loss": 4.062442779541016,
+ "step": 2814
+ },
+ {
+ "epoch": 39.09781659388646,
+ "grad_norm": 0.016643647104501724,
+ "learning_rate": 0.0006,
+ "loss": 4.08871603012085,
+ "step": 2815
+ },
+ {
+ "epoch": 39.1117903930131,
+ "grad_norm": 0.016872866079211235,
+ "learning_rate": 0.0006,
+ "loss": 3.9514245986938477,
+ "step": 2816
+ },
+ {
+ "epoch": 39.12576419213974,
+ "grad_norm": 0.018045669421553612,
+ "learning_rate": 0.0006,
+ "loss": 4.038842678070068,
+ "step": 2817
+ },
+ {
+ "epoch": 39.13973799126637,
+ "grad_norm": 0.017853403463959694,
+ "learning_rate": 0.0006,
+ "loss": 3.93709135055542,
+ "step": 2818
+ },
+ {
+ "epoch": 39.15371179039301,
+ "grad_norm": 0.017507443204522133,
+ "learning_rate": 0.0006,
+ "loss": 3.98665714263916,
+ "step": 2819
+ },
+ {
+ "epoch": 39.16768558951965,
+ "grad_norm": 0.01766853965818882,
+ "learning_rate": 0.0006,
+ "loss": 3.937126636505127,
+ "step": 2820
+ },
+ {
+ "epoch": 39.18165938864629,
+ "grad_norm": 0.016660021618008614,
+ "learning_rate": 0.0006,
+ "loss": 3.8397014141082764,
+ "step": 2821
+ },
+ {
+ "epoch": 39.19563318777293,
+ "grad_norm": 0.017719309777021408,
+ "learning_rate": 0.0006,
+ "loss": 4.003579616546631,
+ "step": 2822
+ },
+ {
+ "epoch": 39.209606986899566,
+ "grad_norm": 0.018234414979815483,
+ "learning_rate": 0.0006,
+ "loss": 4.053859710693359,
+ "step": 2823
+ },
+ {
+ "epoch": 39.223580786026204,
+ "grad_norm": 0.018777024000883102,
+ "learning_rate": 0.0006,
+ "loss": 4.056182861328125,
+ "step": 2824
+ },
+ {
+ "epoch": 39.237554585152836,
+ "grad_norm": 0.017443502321839333,
+ "learning_rate": 0.0006,
+ "loss": 4.025068283081055,
+ "step": 2825
+ },
+ {
+ "epoch": 39.251528384279474,
+ "grad_norm": 0.016500331461429596,
+ "learning_rate": 0.0006,
+ "loss": 4.101631164550781,
+ "step": 2826
+ },
+ {
+ "epoch": 39.26550218340611,
+ "grad_norm": 0.016101544722914696,
+ "learning_rate": 0.0006,
+ "loss": 4.027082443237305,
+ "step": 2827
+ },
+ {
+ "epoch": 39.27947598253275,
+ "grad_norm": 0.015471171587705612,
+ "learning_rate": 0.0006,
+ "loss": 4.022909641265869,
+ "step": 2828
+ },
+ {
+ "epoch": 39.29344978165939,
+ "grad_norm": 0.015779748558998108,
+ "learning_rate": 0.0006,
+ "loss": 4.18796968460083,
+ "step": 2829
+ },
+ {
+ "epoch": 39.30742358078603,
+ "grad_norm": 0.015522679314017296,
+ "learning_rate": 0.0006,
+ "loss": 4.005777359008789,
+ "step": 2830
+ },
+ {
+ "epoch": 39.32139737991266,
+ "grad_norm": 0.015226504765450954,
+ "learning_rate": 0.0006,
+ "loss": 4.108416557312012,
+ "step": 2831
+ },
+ {
+ "epoch": 39.3353711790393,
+ "grad_norm": 0.015095778740942478,
+ "learning_rate": 0.0006,
+ "loss": 4.08029842376709,
+ "step": 2832
+ },
+ {
+ "epoch": 39.34934497816594,
+ "grad_norm": 0.015601031482219696,
+ "learning_rate": 0.0006,
+ "loss": 3.989492416381836,
+ "step": 2833
+ },
+ {
+ "epoch": 39.36331877729258,
+ "grad_norm": 0.014825052581727505,
+ "learning_rate": 0.0006,
+ "loss": 3.998520612716675,
+ "step": 2834
+ },
+ {
+ "epoch": 39.377292576419215,
+ "grad_norm": 0.014821100980043411,
+ "learning_rate": 0.0006,
+ "loss": 4.001777172088623,
+ "step": 2835
+ },
+ {
+ "epoch": 39.391266375545854,
+ "grad_norm": 0.015333754941821098,
+ "learning_rate": 0.0006,
+ "loss": 4.11985969543457,
+ "step": 2836
+ },
+ {
+ "epoch": 39.40524017467249,
+ "grad_norm": 0.015940699726343155,
+ "learning_rate": 0.0006,
+ "loss": 4.025373935699463,
+ "step": 2837
+ },
+ {
+ "epoch": 39.419213973799124,
+ "grad_norm": 0.015096287243068218,
+ "learning_rate": 0.0006,
+ "loss": 4.068551063537598,
+ "step": 2838
+ },
+ {
+ "epoch": 39.43318777292576,
+ "grad_norm": 0.01533548254519701,
+ "learning_rate": 0.0006,
+ "loss": 3.9800801277160645,
+ "step": 2839
+ },
+ {
+ "epoch": 39.4471615720524,
+ "grad_norm": 0.014648732729256153,
+ "learning_rate": 0.0006,
+ "loss": 3.9670190811157227,
+ "step": 2840
+ },
+ {
+ "epoch": 39.46113537117904,
+ "grad_norm": 0.015759671106934547,
+ "learning_rate": 0.0006,
+ "loss": 4.050443649291992,
+ "step": 2841
+ },
+ {
+ "epoch": 39.47510917030568,
+ "grad_norm": 0.016054073348641396,
+ "learning_rate": 0.0006,
+ "loss": 4.22734260559082,
+ "step": 2842
+ },
+ {
+ "epoch": 39.48908296943232,
+ "grad_norm": 0.0160264540463686,
+ "learning_rate": 0.0006,
+ "loss": 4.116174221038818,
+ "step": 2843
+ },
+ {
+ "epoch": 39.50305676855895,
+ "grad_norm": 0.015426610596477985,
+ "learning_rate": 0.0006,
+ "loss": 4.000178337097168,
+ "step": 2844
+ },
+ {
+ "epoch": 39.51703056768559,
+ "grad_norm": 0.01520642638206482,
+ "learning_rate": 0.0006,
+ "loss": 4.218410968780518,
+ "step": 2845
+ },
+ {
+ "epoch": 39.531004366812226,
+ "grad_norm": 0.015305949375033379,
+ "learning_rate": 0.0006,
+ "loss": 4.025891304016113,
+ "step": 2846
+ },
+ {
+ "epoch": 39.544978165938865,
+ "grad_norm": 0.015461564064025879,
+ "learning_rate": 0.0006,
+ "loss": 4.100950717926025,
+ "step": 2847
+ },
+ {
+ "epoch": 39.5589519650655,
+ "grad_norm": 0.01712625101208687,
+ "learning_rate": 0.0006,
+ "loss": 4.0620856285095215,
+ "step": 2848
+ },
+ {
+ "epoch": 39.57292576419214,
+ "grad_norm": 0.017133301123976707,
+ "learning_rate": 0.0006,
+ "loss": 4.097138404846191,
+ "step": 2849
+ },
+ {
+ "epoch": 39.58689956331878,
+ "grad_norm": 0.01683625392615795,
+ "learning_rate": 0.0006,
+ "loss": 4.084364891052246,
+ "step": 2850
+ },
+ {
+ "epoch": 39.60087336244541,
+ "grad_norm": 0.0165710486471653,
+ "learning_rate": 0.0006,
+ "loss": 4.042361259460449,
+ "step": 2851
+ },
+ {
+ "epoch": 39.61484716157205,
+ "grad_norm": 0.01660393364727497,
+ "learning_rate": 0.0006,
+ "loss": 4.1395487785339355,
+ "step": 2852
+ },
+ {
+ "epoch": 39.62882096069869,
+ "grad_norm": 0.017025845125317574,
+ "learning_rate": 0.0006,
+ "loss": 4.1207075119018555,
+ "step": 2853
+ },
+ {
+ "epoch": 39.64279475982533,
+ "grad_norm": 0.01660497486591339,
+ "learning_rate": 0.0006,
+ "loss": 4.103846073150635,
+ "step": 2854
+ },
+ {
+ "epoch": 39.65676855895197,
+ "grad_norm": 0.01607946865260601,
+ "learning_rate": 0.0006,
+ "loss": 4.189016342163086,
+ "step": 2855
+ },
+ {
+ "epoch": 39.670742358078606,
+ "grad_norm": 0.01614505983889103,
+ "learning_rate": 0.0006,
+ "loss": 4.079860687255859,
+ "step": 2856
+ },
+ {
+ "epoch": 39.68471615720524,
+ "grad_norm": 0.017085086554288864,
+ "learning_rate": 0.0006,
+ "loss": 3.9774293899536133,
+ "step": 2857
+ },
+ {
+ "epoch": 39.698689956331876,
+ "grad_norm": 0.01816520094871521,
+ "learning_rate": 0.0006,
+ "loss": 4.113666534423828,
+ "step": 2858
+ },
+ {
+ "epoch": 39.712663755458514,
+ "grad_norm": 0.016597526147961617,
+ "learning_rate": 0.0006,
+ "loss": 4.068242073059082,
+ "step": 2859
+ },
+ {
+ "epoch": 39.72663755458515,
+ "grad_norm": 0.01672235131263733,
+ "learning_rate": 0.0006,
+ "loss": 4.024444580078125,
+ "step": 2860
+ },
+ {
+ "epoch": 39.74061135371179,
+ "grad_norm": 0.020093556493520737,
+ "learning_rate": 0.0006,
+ "loss": 4.064124584197998,
+ "step": 2861
+ },
+ {
+ "epoch": 39.75458515283843,
+ "grad_norm": 0.023141304031014442,
+ "learning_rate": 0.0006,
+ "loss": 4.130439758300781,
+ "step": 2862
+ },
+ {
+ "epoch": 39.76855895196506,
+ "grad_norm": 0.02289372682571411,
+ "learning_rate": 0.0006,
+ "loss": 4.14794921875,
+ "step": 2863
+ },
+ {
+ "epoch": 39.7825327510917,
+ "grad_norm": 0.01919020712375641,
+ "learning_rate": 0.0006,
+ "loss": 4.123961925506592,
+ "step": 2864
+ },
+ {
+ "epoch": 39.79650655021834,
+ "grad_norm": 0.016932329162955284,
+ "learning_rate": 0.0006,
+ "loss": 4.0944132804870605,
+ "step": 2865
+ },
+ {
+ "epoch": 39.81048034934498,
+ "grad_norm": 0.016644535586237907,
+ "learning_rate": 0.0006,
+ "loss": 4.124269962310791,
+ "step": 2866
+ },
+ {
+ "epoch": 39.82445414847162,
+ "grad_norm": 0.016302216798067093,
+ "learning_rate": 0.0006,
+ "loss": 4.0267205238342285,
+ "step": 2867
+ },
+ {
+ "epoch": 39.838427947598255,
+ "grad_norm": 0.01673300378024578,
+ "learning_rate": 0.0006,
+ "loss": 4.10650634765625,
+ "step": 2868
+ },
+ {
+ "epoch": 39.852401746724894,
+ "grad_norm": 0.016581356525421143,
+ "learning_rate": 0.0006,
+ "loss": 4.1547956466674805,
+ "step": 2869
+ },
+ {
+ "epoch": 39.866375545851525,
+ "grad_norm": 0.01691797934472561,
+ "learning_rate": 0.0006,
+ "loss": 4.056936740875244,
+ "step": 2870
+ },
+ {
+ "epoch": 39.880349344978164,
+ "grad_norm": 0.01596726104617119,
+ "learning_rate": 0.0006,
+ "loss": 3.9891414642333984,
+ "step": 2871
+ },
+ {
+ "epoch": 39.8943231441048,
+ "grad_norm": 0.015615091659128666,
+ "learning_rate": 0.0006,
+ "loss": 4.137948036193848,
+ "step": 2872
+ },
+ {
+ "epoch": 39.90829694323144,
+ "grad_norm": 0.015703413635492325,
+ "learning_rate": 0.0006,
+ "loss": 3.9844930171966553,
+ "step": 2873
+ },
+ {
+ "epoch": 39.92227074235808,
+ "grad_norm": 0.015453094616532326,
+ "learning_rate": 0.0006,
+ "loss": 4.098443508148193,
+ "step": 2874
+ },
+ {
+ "epoch": 39.93624454148472,
+ "grad_norm": 0.01449158601462841,
+ "learning_rate": 0.0006,
+ "loss": 4.090852737426758,
+ "step": 2875
+ },
+ {
+ "epoch": 39.95021834061135,
+ "grad_norm": 0.01537515688687563,
+ "learning_rate": 0.0006,
+ "loss": 4.239612579345703,
+ "step": 2876
+ },
+ {
+ "epoch": 39.96419213973799,
+ "grad_norm": 0.015989921987056732,
+ "learning_rate": 0.0006,
+ "loss": 4.142073154449463,
+ "step": 2877
+ },
+ {
+ "epoch": 39.97816593886463,
+ "grad_norm": 0.015985611826181412,
+ "learning_rate": 0.0006,
+ "loss": 4.100076675415039,
+ "step": 2878
+ },
+ {
+ "epoch": 39.992139737991266,
+ "grad_norm": 0.015260833315551281,
+ "learning_rate": 0.0006,
+ "loss": 4.068576335906982,
+ "step": 2879
+ },
+ {
+ "epoch": 40.0,
+ "grad_norm": 0.018047941848635674,
+ "learning_rate": 0.0006,
+ "loss": 4.005878925323486,
+ "step": 2880
+ },
+ {
+ "epoch": 40.0,
+ "eval_loss": 4.622848987579346,
+ "eval_runtime": 56.9618,
+ "eval_samples_per_second": 42.871,
+ "eval_steps_per_second": 1.352,
+ "step": 2880
+ },
+ {
+ "epoch": 40.01397379912664,
+ "grad_norm": 0.015976838767528534,
+ "learning_rate": 0.0006,
+ "loss": 4.035689353942871,
+ "step": 2881
+ },
+ {
+ "epoch": 40.02794759825328,
+ "grad_norm": 0.015176220797002316,
+ "learning_rate": 0.0006,
+ "loss": 3.932197332382202,
+ "step": 2882
+ },
+ {
+ "epoch": 40.041921397379916,
+ "grad_norm": 0.016676228493452072,
+ "learning_rate": 0.0006,
+ "loss": 4.014171123504639,
+ "step": 2883
+ },
+ {
+ "epoch": 40.05589519650655,
+ "grad_norm": 0.01698281429708004,
+ "learning_rate": 0.0006,
+ "loss": 4.117893218994141,
+ "step": 2884
+ },
+ {
+ "epoch": 40.069868995633186,
+ "grad_norm": 0.017872486263513565,
+ "learning_rate": 0.0006,
+ "loss": 4.132119178771973,
+ "step": 2885
+ },
+ {
+ "epoch": 40.083842794759825,
+ "grad_norm": 0.017615554854273796,
+ "learning_rate": 0.0006,
+ "loss": 4.072000980377197,
+ "step": 2886
+ },
+ {
+ "epoch": 40.09781659388646,
+ "grad_norm": 0.017469394952058792,
+ "learning_rate": 0.0006,
+ "loss": 4.017726421356201,
+ "step": 2887
+ },
+ {
+ "epoch": 40.1117903930131,
+ "grad_norm": 0.01954762637615204,
+ "learning_rate": 0.0006,
+ "loss": 4.014443874359131,
+ "step": 2888
+ },
+ {
+ "epoch": 40.12576419213974,
+ "grad_norm": 0.018935853615403175,
+ "learning_rate": 0.0006,
+ "loss": 4.0042948722839355,
+ "step": 2889
+ },
+ {
+ "epoch": 40.13973799126637,
+ "grad_norm": 0.019382692873477936,
+ "learning_rate": 0.0006,
+ "loss": 4.06160831451416,
+ "step": 2890
+ },
+ {
+ "epoch": 40.15371179039301,
+ "grad_norm": 0.019677115604281425,
+ "learning_rate": 0.0006,
+ "loss": 4.13466215133667,
+ "step": 2891
+ },
+ {
+ "epoch": 40.16768558951965,
+ "grad_norm": 0.020365122705698013,
+ "learning_rate": 0.0006,
+ "loss": 4.095000267028809,
+ "step": 2892
+ },
+ {
+ "epoch": 40.18165938864629,
+ "grad_norm": 0.021842138841748238,
+ "learning_rate": 0.0006,
+ "loss": 3.9877853393554688,
+ "step": 2893
+ },
+ {
+ "epoch": 40.19563318777293,
+ "grad_norm": 0.02354881353676319,
+ "learning_rate": 0.0006,
+ "loss": 3.9430112838745117,
+ "step": 2894
+ },
+ {
+ "epoch": 40.209606986899566,
+ "grad_norm": 0.026602735742926598,
+ "learning_rate": 0.0006,
+ "loss": 4.012076377868652,
+ "step": 2895
+ },
+ {
+ "epoch": 40.223580786026204,
+ "grad_norm": 0.023976484313607216,
+ "learning_rate": 0.0006,
+ "loss": 3.9768595695495605,
+ "step": 2896
+ },
+ {
+ "epoch": 40.237554585152836,
+ "grad_norm": 0.02336616814136505,
+ "learning_rate": 0.0006,
+ "loss": 4.1096038818359375,
+ "step": 2897
+ },
+ {
+ "epoch": 40.251528384279474,
+ "grad_norm": 0.02426023595035076,
+ "learning_rate": 0.0006,
+ "loss": 4.157189846038818,
+ "step": 2898
+ },
+ {
+ "epoch": 40.26550218340611,
+ "grad_norm": 0.021572696045041084,
+ "learning_rate": 0.0006,
+ "loss": 3.979175090789795,
+ "step": 2899
+ },
+ {
+ "epoch": 40.27947598253275,
+ "grad_norm": 0.08300432562828064,
+ "learning_rate": 0.0006,
+ "loss": 4.092336654663086,
+ "step": 2900
+ },
+ {
+ "epoch": 40.29344978165939,
+ "grad_norm": 0.7323967814445496,
+ "learning_rate": 0.0006,
+ "loss": 5.337990760803223,
+ "step": 2901
+ },
+ {
+ "epoch": 40.30742358078603,
+ "grad_norm": 0.3363337814807892,
+ "learning_rate": 0.0006,
+ "loss": 6.40043830871582,
+ "step": 2902
+ },
+ {
+ "epoch": 40.32139737991266,
+ "grad_norm": 0.598686933517456,
+ "learning_rate": 0.0006,
+ "loss": 7.004765510559082,
+ "step": 2903
+ },
+ {
+ "epoch": 40.3353711790393,
+ "grad_norm": 0.5998364686965942,
+ "learning_rate": 0.0006,
+ "loss": 8.184747695922852,
+ "step": 2904
+ },
+ {
+ "epoch": 40.34934497816594,
+ "grad_norm": 1.408208966255188,
+ "learning_rate": 0.0006,
+ "loss": 8.88978385925293,
+ "step": 2905
+ },
+ {
+ "epoch": 40.36331877729258,
+ "grad_norm": 0.7829785943031311,
+ "learning_rate": 0.0006,
+ "loss": 11.13601303100586,
+ "step": 2906
+ },
+ {
+ "epoch": 40.377292576419215,
+ "grad_norm": 0.7769215106964111,
+ "learning_rate": 0.0006,
+ "loss": 9.90416145324707,
+ "step": 2907
+ },
+ {
+ "epoch": 40.391266375545854,
+ "grad_norm": 0.46508079767227173,
+ "learning_rate": 0.0006,
+ "loss": 9.647209167480469,
+ "step": 2908
+ },
+ {
+ "epoch": 40.40524017467249,
+ "grad_norm": 1.0228968858718872,
+ "learning_rate": 0.0006,
+ "loss": 11.685434341430664,
+ "step": 2909
+ },
+ {
+ "epoch": 40.419213973799124,
+ "grad_norm": 0.20553778111934662,
+ "learning_rate": 0.0006,
+ "loss": 8.621708869934082,
+ "step": 2910
+ },
+ {
+ "epoch": 40.43318777292576,
+ "grad_norm": 0.18466424942016602,
+ "learning_rate": 0.0006,
+ "loss": 8.590787887573242,
+ "step": 2911
+ },
+ {
+ "epoch": 40.4471615720524,
+ "grad_norm": 0.10458594560623169,
+ "learning_rate": 0.0006,
+ "loss": 8.324458122253418,
+ "step": 2912
+ },
+ {
+ "epoch": 40.46113537117904,
+ "grad_norm": 0.1096842959523201,
+ "learning_rate": 0.0006,
+ "loss": 8.300851821899414,
+ "step": 2913
+ },
+ {
+ "epoch": 40.47510917030568,
+ "grad_norm": 0.10501226782798767,
+ "learning_rate": 0.0006,
+ "loss": 8.21009635925293,
+ "step": 2914
+ },
+ {
+ "epoch": 40.48908296943232,
+ "grad_norm": 0.13704165816307068,
+ "learning_rate": 0.0006,
+ "loss": 8.141355514526367,
+ "step": 2915
+ },
+ {
+ "epoch": 40.50305676855895,
+ "grad_norm": 0.12129108607769012,
+ "learning_rate": 0.0006,
+ "loss": 8.074612617492676,
+ "step": 2916
+ },
+ {
+ "epoch": 40.51703056768559,
+ "grad_norm": 0.10028897225856781,
+ "learning_rate": 0.0006,
+ "loss": 8.00384521484375,
+ "step": 2917
+ },
+ {
+ "epoch": 40.531004366812226,
+ "grad_norm": 0.11764246970415115,
+ "learning_rate": 0.0006,
+ "loss": 8.078598022460938,
+ "step": 2918
+ },
+ {
+ "epoch": 40.544978165938865,
+ "grad_norm": 0.10081959515810013,
+ "learning_rate": 0.0006,
+ "loss": 8.008938789367676,
+ "step": 2919
+ },
+ {
+ "epoch": 40.5589519650655,
+ "grad_norm": 0.1003170907497406,
+ "learning_rate": 0.0006,
+ "loss": 7.962940216064453,
+ "step": 2920
+ },
+ {
+ "epoch": 40.57292576419214,
+ "grad_norm": 0.08763492107391357,
+ "learning_rate": 0.0006,
+ "loss": 7.909419536590576,
+ "step": 2921
+ },
+ {
+ "epoch": 40.58689956331878,
+ "grad_norm": 0.0636807233095169,
+ "learning_rate": 0.0006,
+ "loss": 7.878905773162842,
+ "step": 2922
+ },
+ {
+ "epoch": 40.60087336244541,
+ "grad_norm": 0.06005936488509178,
+ "learning_rate": 0.0006,
+ "loss": 7.850980758666992,
+ "step": 2923
+ },
+ {
+ "epoch": 40.61484716157205,
+ "grad_norm": 0.06415296345949173,
+ "learning_rate": 0.0006,
+ "loss": 7.830353260040283,
+ "step": 2924
+ },
+ {
+ "epoch": 40.62882096069869,
+ "grad_norm": 0.0732007846236229,
+ "learning_rate": 0.0006,
+ "loss": 7.830746650695801,
+ "step": 2925
+ },
+ {
+ "epoch": 40.64279475982533,
+ "grad_norm": 0.06628110259771347,
+ "learning_rate": 0.0006,
+ "loss": 7.82661771774292,
+ "step": 2926
+ },
+ {
+ "epoch": 40.65676855895197,
+ "grad_norm": 0.05544218048453331,
+ "learning_rate": 0.0006,
+ "loss": 7.814305782318115,
+ "step": 2927
+ },
+ {
+ "epoch": 40.670742358078606,
+ "grad_norm": 0.050327103585004807,
+ "learning_rate": 0.0006,
+ "loss": 7.764055252075195,
+ "step": 2928
+ },
+ {
+ "epoch": 40.68471615720524,
+ "grad_norm": 0.04918425530195236,
+ "learning_rate": 0.0006,
+ "loss": 7.744625091552734,
+ "step": 2929
+ },
+ {
+ "epoch": 40.698689956331876,
+ "grad_norm": 0.042862266302108765,
+ "learning_rate": 0.0006,
+ "loss": 7.764426231384277,
+ "step": 2930
+ },
+ {
+ "epoch": 40.712663755458514,
+ "grad_norm": 0.04338089004158974,
+ "learning_rate": 0.0006,
+ "loss": 7.712530136108398,
+ "step": 2931
+ },
+ {
+ "epoch": 40.72663755458515,
+ "grad_norm": 0.04127737879753113,
+ "learning_rate": 0.0006,
+ "loss": 7.742180347442627,
+ "step": 2932
+ },
+ {
+ "epoch": 40.74061135371179,
+ "grad_norm": 0.03358631581068039,
+ "learning_rate": 0.0006,
+ "loss": 7.714834213256836,
+ "step": 2933
+ },
+ {
+ "epoch": 40.75458515283843,
+ "grad_norm": 0.02845790982246399,
+ "learning_rate": 0.0006,
+ "loss": 7.7021026611328125,
+ "step": 2934
+ },
+ {
+ "epoch": 40.76855895196506,
+ "grad_norm": 0.03012419492006302,
+ "learning_rate": 0.0006,
+ "loss": 7.712377548217773,
+ "step": 2935
+ },
+ {
+ "epoch": 40.7825327510917,
+ "grad_norm": 0.032981179654598236,
+ "learning_rate": 0.0006,
+ "loss": 7.666385650634766,
+ "step": 2936
+ },
+ {
+ "epoch": 40.79650655021834,
+ "grad_norm": 0.03357074037194252,
+ "learning_rate": 0.0006,
+ "loss": 7.649097919464111,
+ "step": 2937
+ },
+ {
+ "epoch": 40.81048034934498,
+ "grad_norm": 0.030914708971977234,
+ "learning_rate": 0.0006,
+ "loss": 7.649196624755859,
+ "step": 2938
+ },
+ {
+ "epoch": 40.82445414847162,
+ "grad_norm": 0.026296131312847137,
+ "learning_rate": 0.0006,
+ "loss": 7.629964828491211,
+ "step": 2939
+ },
+ {
+ "epoch": 40.838427947598255,
+ "grad_norm": 0.026677245274186134,
+ "learning_rate": 0.0006,
+ "loss": 7.599973678588867,
+ "step": 2940
+ },
+ {
+ "epoch": 40.852401746724894,
+ "grad_norm": 0.023208219558000565,
+ "learning_rate": 0.0006,
+ "loss": 7.601624488830566,
+ "step": 2941
+ },
+ {
+ "epoch": 40.866375545851525,
+ "grad_norm": 0.020741475746035576,
+ "learning_rate": 0.0006,
+ "loss": 7.603902339935303,
+ "step": 2942
+ },
+ {
+ "epoch": 40.880349344978164,
+ "grad_norm": 0.019221751019358635,
+ "learning_rate": 0.0006,
+ "loss": 7.5535888671875,
+ "step": 2943
+ },
+ {
+ "epoch": 40.8943231441048,
+ "grad_norm": 0.019491951912641525,
+ "learning_rate": 0.0006,
+ "loss": 7.54597282409668,
+ "step": 2944
+ },
+ {
+ "epoch": 40.90829694323144,
+ "grad_norm": 0.019702572375535965,
+ "learning_rate": 0.0006,
+ "loss": 7.537567138671875,
+ "step": 2945
+ },
+ {
+ "epoch": 40.92227074235808,
+ "grad_norm": 0.01768800988793373,
+ "learning_rate": 0.0006,
+ "loss": 7.512800216674805,
+ "step": 2946
+ },
+ {
+ "epoch": 40.93624454148472,
+ "grad_norm": 0.01926138624548912,
+ "learning_rate": 0.0006,
+ "loss": 7.509932518005371,
+ "step": 2947
+ },
+ {
+ "epoch": 40.95021834061135,
+ "grad_norm": 0.019030006602406502,
+ "learning_rate": 0.0006,
+ "loss": 7.497983932495117,
+ "step": 2948
+ },
+ {
+ "epoch": 40.96419213973799,
+ "grad_norm": 0.017313841730356216,
+ "learning_rate": 0.0006,
+ "loss": 7.463518142700195,
+ "step": 2949
+ },
+ {
+ "epoch": 40.97816593886463,
+ "grad_norm": 0.01310762483626604,
+ "learning_rate": 0.0006,
+ "loss": 7.474668502807617,
+ "step": 2950
+ },
+ {
+ "epoch": 40.992139737991266,
+ "grad_norm": 0.013008282519876957,
+ "learning_rate": 0.0006,
+ "loss": 7.45982551574707,
+ "step": 2951
+ },
+ {
+ "epoch": 41.0,
+ "grad_norm": 0.01441363524645567,
+ "learning_rate": 0.0006,
+ "loss": 7.446529388427734,
+ "step": 2952
+ },
+ {
+ "epoch": 41.0,
+ "eval_loss": 7.486120223999023,
+ "eval_runtime": 70.4986,
+ "eval_samples_per_second": 34.639,
+ "eval_steps_per_second": 1.092,
+ "step": 2952
+ },
+ {
+ "epoch": 41.01397379912664,
+ "grad_norm": 0.014061855152249336,
+ "learning_rate": 0.0006,
+ "loss": 7.4692182540893555,
+ "step": 2953
+ },
+ {
+ "epoch": 41.02794759825328,
+ "grad_norm": 0.01587500050663948,
+ "learning_rate": 0.0006,
+ "loss": 7.4182024002075195,
+ "step": 2954
+ },
+ {
+ "epoch": 41.041921397379916,
+ "grad_norm": 0.014783984050154686,
+ "learning_rate": 0.0006,
+ "loss": 7.466778755187988,
+ "step": 2955
+ },
+ {
+ "epoch": 41.05589519650655,
+ "grad_norm": 0.011560854502022266,
+ "learning_rate": 0.0006,
+ "loss": 7.438002586364746,
+ "step": 2956
+ },
+ {
+ "epoch": 41.069868995633186,
+ "grad_norm": 0.00970957800745964,
+ "learning_rate": 0.0006,
+ "loss": 7.442694187164307,
+ "step": 2957
+ },
+ {
+ "epoch": 41.083842794759825,
+ "grad_norm": 0.009633459150791168,
+ "learning_rate": 0.0006,
+ "loss": 7.458884239196777,
+ "step": 2958
+ },
+ {
+ "epoch": 41.09781659388646,
+ "grad_norm": 0.010664350353181362,
+ "learning_rate": 0.0006,
+ "loss": 7.48378324508667,
+ "step": 2959
+ },
+ {
+ "epoch": 41.1117903930131,
+ "grad_norm": 0.011440591886639595,
+ "learning_rate": 0.0006,
+ "loss": 7.474484920501709,
+ "step": 2960
+ },
+ {
+ "epoch": 41.12576419213974,
+ "grad_norm": 0.01150908600538969,
+ "learning_rate": 0.0006,
+ "loss": 7.417902946472168,
+ "step": 2961
+ },
+ {
+ "epoch": 41.13973799126637,
+ "grad_norm": 0.010375697165727615,
+ "learning_rate": 0.0006,
+ "loss": 7.403279781341553,
+ "step": 2962
+ },
+ {
+ "epoch": 41.15371179039301,
+ "grad_norm": 0.009075703099370003,
+ "learning_rate": 0.0006,
+ "loss": 7.447347164154053,
+ "step": 2963
+ },
+ {
+ "epoch": 41.16768558951965,
+ "grad_norm": 0.008892177604138851,
+ "learning_rate": 0.0006,
+ "loss": 7.402681350708008,
+ "step": 2964
+ },
+ {
+ "epoch": 41.18165938864629,
+ "grad_norm": 0.009492084383964539,
+ "learning_rate": 0.0006,
+ "loss": 7.3871355056762695,
+ "step": 2965
+ },
+ {
+ "epoch": 41.19563318777293,
+ "grad_norm": 0.009139460511505604,
+ "learning_rate": 0.0006,
+ "loss": 7.419617176055908,
+ "step": 2966
+ },
+ {
+ "epoch": 41.209606986899566,
+ "grad_norm": 0.007959138602018356,
+ "learning_rate": 0.0006,
+ "loss": 7.374949932098389,
+ "step": 2967
+ },
+ {
+ "epoch": 41.223580786026204,
+ "grad_norm": 0.007571548223495483,
+ "learning_rate": 0.0006,
+ "loss": 7.4205241203308105,
+ "step": 2968
+ },
+ {
+ "epoch": 41.237554585152836,
+ "grad_norm": 0.007885400205850601,
+ "learning_rate": 0.0006,
+ "loss": 7.416543006896973,
+ "step": 2969
+ },
+ {
+ "epoch": 41.251528384279474,
+ "grad_norm": 0.008650158531963825,
+ "learning_rate": 0.0006,
+ "loss": 7.357326984405518,
+ "step": 2970
+ },
+ {
+ "epoch": 41.26550218340611,
+ "grad_norm": 0.007483262103050947,
+ "learning_rate": 0.0006,
+ "loss": 7.410891056060791,
+ "step": 2971
+ },
+ {
+ "epoch": 41.27947598253275,
+ "grad_norm": 0.00869277399033308,
+ "learning_rate": 0.0006,
+ "loss": 7.392127990722656,
+ "step": 2972
+ },
+ {
+ "epoch": 41.29344978165939,
+ "grad_norm": 0.00885078776627779,
+ "learning_rate": 0.0006,
+ "loss": 7.405449867248535,
+ "step": 2973
+ },
+ {
+ "epoch": 41.30742358078603,
+ "grad_norm": 0.00958226341754198,
+ "learning_rate": 0.0006,
+ "loss": 7.399868488311768,
+ "step": 2974
+ },
+ {
+ "epoch": 41.32139737991266,
+ "grad_norm": 0.006991599686443806,
+ "learning_rate": 0.0006,
+ "loss": 7.35369873046875,
+ "step": 2975
+ },
+ {
+ "epoch": 41.3353711790393,
+ "grad_norm": 0.007415142375975847,
+ "learning_rate": 0.0006,
+ "loss": 7.358353137969971,
+ "step": 2976
+ },
+ {
+ "epoch": 41.34934497816594,
+ "grad_norm": 0.007024371065199375,
+ "learning_rate": 0.0006,
+ "loss": 7.4070892333984375,
+ "step": 2977
+ },
+ {
+ "epoch": 41.36331877729258,
+ "grad_norm": 0.007234332151710987,
+ "learning_rate": 0.0006,
+ "loss": 7.379120349884033,
+ "step": 2978
+ },
+ {
+ "epoch": 41.377292576419215,
+ "grad_norm": 0.00798892229795456,
+ "learning_rate": 0.0006,
+ "loss": 7.365455627441406,
+ "step": 2979
+ },
+ {
+ "epoch": 41.391266375545854,
+ "grad_norm": 0.007303398102521896,
+ "learning_rate": 0.0006,
+ "loss": 7.341798782348633,
+ "step": 2980
+ },
+ {
+ "epoch": 41.40524017467249,
+ "grad_norm": 0.007119669578969479,
+ "learning_rate": 0.0006,
+ "loss": 7.363280296325684,
+ "step": 2981
+ },
+ {
+ "epoch": 41.419213973799124,
+ "grad_norm": 0.007457619532942772,
+ "learning_rate": 0.0006,
+ "loss": 7.39388370513916,
+ "step": 2982
+ },
+ {
+ "epoch": 41.43318777292576,
+ "grad_norm": 0.007905688136816025,
+ "learning_rate": 0.0006,
+ "loss": 7.380652904510498,
+ "step": 2983
+ },
+ {
+ "epoch": 41.4471615720524,
+ "grad_norm": 0.006212855689227581,
+ "learning_rate": 0.0006,
+ "loss": 7.398769378662109,
+ "step": 2984
+ },
+ {
+ "epoch": 41.46113537117904,
+ "grad_norm": 0.0070641543716192245,
+ "learning_rate": 0.0006,
+ "loss": 7.3503546714782715,
+ "step": 2985
+ },
+ {
+ "epoch": 41.47510917030568,
+ "grad_norm": 0.00787805300205946,
+ "learning_rate": 0.0006,
+ "loss": 7.347678184509277,
+ "step": 2986
+ },
+ {
+ "epoch": 41.48908296943232,
+ "grad_norm": 0.006760426796972752,
+ "learning_rate": 0.0006,
+ "loss": 7.380006790161133,
+ "step": 2987
+ },
+ {
+ "epoch": 41.50305676855895,
+ "grad_norm": 0.00772385997697711,
+ "learning_rate": 0.0006,
+ "loss": 7.363680362701416,
+ "step": 2988
+ },
+ {
+ "epoch": 41.51703056768559,
+ "grad_norm": 0.007373564876616001,
+ "learning_rate": 0.0006,
+ "loss": 7.333014488220215,
+ "step": 2989
+ },
+ {
+ "epoch": 41.531004366812226,
+ "grad_norm": 0.006717653013765812,
+ "learning_rate": 0.0006,
+ "loss": 7.360076427459717,
+ "step": 2990
+ },
+ {
+ "epoch": 41.544978165938865,
+ "grad_norm": 0.007191413082182407,
+ "learning_rate": 0.0006,
+ "loss": 7.397030830383301,
+ "step": 2991
+ },
+ {
+ "epoch": 41.5589519650655,
+ "grad_norm": 0.007831960916519165,
+ "learning_rate": 0.0006,
+ "loss": 7.369959831237793,
+ "step": 2992
+ },
+ {
+ "epoch": 41.57292576419214,
+ "grad_norm": 0.006947481073439121,
+ "learning_rate": 0.0006,
+ "loss": 7.333254337310791,
+ "step": 2993
+ },
+ {
+ "epoch": 41.58689956331878,
+ "grad_norm": 0.006243606563657522,
+ "learning_rate": 0.0006,
+ "loss": 7.327378273010254,
+ "step": 2994
+ },
+ {
+ "epoch": 41.60087336244541,
+ "grad_norm": 0.006298339460045099,
+ "learning_rate": 0.0006,
+ "loss": 7.323482036590576,
+ "step": 2995
+ },
+ {
+ "epoch": 41.61484716157205,
+ "grad_norm": 0.007587412837892771,
+ "learning_rate": 0.0006,
+ "loss": 7.313789367675781,
+ "step": 2996
+ },
+ {
+ "epoch": 41.62882096069869,
+ "grad_norm": 0.007140746805816889,
+ "learning_rate": 0.0006,
+ "loss": 7.356067657470703,
+ "step": 2997
+ },
+ {
+ "epoch": 41.64279475982533,
+ "grad_norm": 0.008315625600516796,
+ "learning_rate": 0.0006,
+ "loss": 7.3339643478393555,
+ "step": 2998
+ },
+ {
+ "epoch": 41.65676855895197,
+ "grad_norm": 0.006547102238982916,
+ "learning_rate": 0.0006,
+ "loss": 7.342191696166992,
+ "step": 2999
+ },
+ {
+ "epoch": 41.670742358078606,
+ "grad_norm": 0.006068122107535601,
+ "learning_rate": 0.0006,
+ "loss": 7.321015357971191,
+ "step": 3000
+ },
+ {
+ "epoch": 41.68471615720524,
+ "grad_norm": 0.007316991221159697,
+ "learning_rate": 0.0006,
+ "loss": 7.346490859985352,
+ "step": 3001
+ },
+ {
+ "epoch": 41.698689956331876,
+ "grad_norm": 0.0067388564348220825,
+ "learning_rate": 0.0006,
+ "loss": 7.340723991394043,
+ "step": 3002
+ },
+ {
+ "epoch": 41.712663755458514,
+ "grad_norm": 0.006839963141828775,
+ "learning_rate": 0.0006,
+ "loss": 7.329365253448486,
+ "step": 3003
+ },
+ {
+ "epoch": 41.72663755458515,
+ "grad_norm": 0.00805041566491127,
+ "learning_rate": 0.0006,
+ "loss": 7.298934459686279,
+ "step": 3004
+ },
+ {
+ "epoch": 41.74061135371179,
+ "grad_norm": 0.007197503466159105,
+ "learning_rate": 0.0006,
+ "loss": 7.331496238708496,
+ "step": 3005
+ },
+ {
+ "epoch": 41.75458515283843,
+ "grad_norm": 0.007220805156975985,
+ "learning_rate": 0.0006,
+ "loss": 7.317695140838623,
+ "step": 3006
+ },
+ {
+ "epoch": 41.76855895196506,
+ "grad_norm": 0.00874276738613844,
+ "learning_rate": 0.0006,
+ "loss": 7.272614479064941,
+ "step": 3007
+ },
+ {
+ "epoch": 41.7825327510917,
+ "grad_norm": 0.008386565372347832,
+ "learning_rate": 0.0006,
+ "loss": 7.295244216918945,
+ "step": 3008
+ },
+ {
+ "epoch": 41.79650655021834,
+ "grad_norm": 0.0077730282209813595,
+ "learning_rate": 0.0006,
+ "loss": 7.3395490646362305,
+ "step": 3009
+ },
+ {
+ "epoch": 41.81048034934498,
+ "grad_norm": 0.008441480807960033,
+ "learning_rate": 0.0006,
+ "loss": 7.319921493530273,
+ "step": 3010
+ },
+ {
+ "epoch": 41.82445414847162,
+ "grad_norm": 0.007398269604891539,
+ "learning_rate": 0.0006,
+ "loss": 7.3029465675354,
+ "step": 3011
+ },
+ {
+ "epoch": 41.838427947598255,
+ "grad_norm": 0.006878677289932966,
+ "learning_rate": 0.0006,
+ "loss": 7.318467140197754,
+ "step": 3012
+ },
+ {
+ "epoch": 41.852401746724894,
+ "grad_norm": 0.007616598624736071,
+ "learning_rate": 0.0006,
+ "loss": 7.3112077713012695,
+ "step": 3013
+ },
+ {
+ "epoch": 41.866375545851525,
+ "grad_norm": 0.011629699729382992,
+ "learning_rate": 0.0006,
+ "loss": 7.32523250579834,
+ "step": 3014
+ },
+ {
+ "epoch": 41.880349344978164,
+ "grad_norm": 0.011011249385774136,
+ "learning_rate": 0.0006,
+ "loss": 7.287247657775879,
+ "step": 3015
+ },
+ {
+ "epoch": 41.8943231441048,
+ "grad_norm": 0.010208295658230782,
+ "learning_rate": 0.0006,
+ "loss": 7.292117118835449,
+ "step": 3016
+ },
+ {
+ "epoch": 41.90829694323144,
+ "grad_norm": 0.008651095442473888,
+ "learning_rate": 0.0006,
+ "loss": 7.302201271057129,
+ "step": 3017
+ },
+ {
+ "epoch": 41.92227074235808,
+ "grad_norm": 0.006692920345813036,
+ "learning_rate": 0.0006,
+ "loss": 7.319568634033203,
+ "step": 3018
+ },
+ {
+ "epoch": 41.93624454148472,
+ "grad_norm": 0.010033348575234413,
+ "learning_rate": 0.0006,
+ "loss": 7.2738566398620605,
+ "step": 3019
+ },
+ {
+ "epoch": 41.95021834061135,
+ "grad_norm": 0.011790832504630089,
+ "learning_rate": 0.0006,
+ "loss": 7.298540115356445,
+ "step": 3020
+ },
+ {
+ "epoch": 41.96419213973799,
+ "grad_norm": 0.018622662872076035,
+ "learning_rate": 0.0006,
+ "loss": 7.267182350158691,
+ "step": 3021
+ },
+ {
+ "epoch": 41.97816593886463,
+ "grad_norm": 0.02520257607102394,
+ "learning_rate": 0.0006,
+ "loss": 7.261566162109375,
+ "step": 3022
+ },
+ {
+ "epoch": 41.992139737991266,
+ "grad_norm": 0.036332644522190094,
+ "learning_rate": 0.0006,
+ "loss": 7.270719051361084,
+ "step": 3023
+ },
+ {
+ "epoch": 42.0,
+ "grad_norm": 0.04717305675148964,
+ "learning_rate": 0.0006,
+ "loss": 7.293332099914551,
+ "step": 3024
+ },
+ {
+ "epoch": 42.0,
+ "eval_loss": 7.354280471801758,
+ "eval_runtime": 56.4353,
+ "eval_samples_per_second": 43.271,
+ "eval_steps_per_second": 1.364,
+ "step": 3024
+ },
+ {
+ "epoch": 42.01397379912664,
+ "grad_norm": 0.06661626696586609,
+ "learning_rate": 0.0006,
+ "loss": 7.338189601898193,
+ "step": 3025
+ },
+ {
+ "epoch": 42.02794759825328,
+ "grad_norm": 0.05452054366469383,
+ "learning_rate": 0.0006,
+ "loss": 7.316764831542969,
+ "step": 3026
+ },
+ {
+ "epoch": 42.041921397379916,
+ "grad_norm": 0.025638775900006294,
+ "learning_rate": 0.0006,
+ "loss": 7.34204626083374,
+ "step": 3027
+ },
+ {
+ "epoch": 42.05589519650655,
+ "grad_norm": 0.022398799657821655,
+ "learning_rate": 0.0006,
+ "loss": 7.262149333953857,
+ "step": 3028
+ },
+ {
+ "epoch": 42.069868995633186,
+ "grad_norm": 0.03713499382138252,
+ "learning_rate": 0.0006,
+ "loss": 7.27695369720459,
+ "step": 3029
+ },
+ {
+ "epoch": 42.083842794759825,
+ "grad_norm": 0.014949247241020203,
+ "learning_rate": 0.0006,
+ "loss": 7.296508312225342,
+ "step": 3030
+ },
+ {
+ "epoch": 42.09781659388646,
+ "grad_norm": 0.02703995071351528,
+ "learning_rate": 0.0006,
+ "loss": 7.258018493652344,
+ "step": 3031
+ },
+ {
+ "epoch": 42.1117903930131,
+ "grad_norm": 0.020352816209197044,
+ "learning_rate": 0.0006,
+ "loss": 7.241450309753418,
+ "step": 3032
+ },
+ {
+ "epoch": 42.12576419213974,
+ "grad_norm": 0.014578479342162609,
+ "learning_rate": 0.0006,
+ "loss": 7.257323265075684,
+ "step": 3033
+ },
+ {
+ "epoch": 42.13973799126637,
+ "grad_norm": 0.026053044945001602,
+ "learning_rate": 0.0006,
+ "loss": 7.254680633544922,
+ "step": 3034
+ },
+ {
+ "epoch": 42.15371179039301,
+ "grad_norm": 0.015120275318622589,
+ "learning_rate": 0.0006,
+ "loss": 7.261074066162109,
+ "step": 3035
+ },
+ {
+ "epoch": 42.16768558951965,
+ "grad_norm": 0.016822177916765213,
+ "learning_rate": 0.0006,
+ "loss": 7.279064178466797,
+ "step": 3036
+ },
+ {
+ "epoch": 42.18165938864629,
+ "grad_norm": 0.016367588192224503,
+ "learning_rate": 0.0006,
+ "loss": 7.251429557800293,
+ "step": 3037
+ },
+ {
+ "epoch": 42.19563318777293,
+ "grad_norm": 0.014350418001413345,
+ "learning_rate": 0.0006,
+ "loss": 7.259264945983887,
+ "step": 3038
+ },
+ {
+ "epoch": 42.209606986899566,
+ "grad_norm": 0.017078228294849396,
+ "learning_rate": 0.0006,
+ "loss": 7.227105140686035,
+ "step": 3039
+ },
+ {
+ "epoch": 42.223580786026204,
+ "grad_norm": 0.011759317480027676,
+ "learning_rate": 0.0006,
+ "loss": 7.275888442993164,
+ "step": 3040
+ },
+ {
+ "epoch": 42.237554585152836,
+ "grad_norm": 0.013019340112805367,
+ "learning_rate": 0.0006,
+ "loss": 7.2517008781433105,
+ "step": 3041
+ },
+ {
+ "epoch": 42.251528384279474,
+ "grad_norm": 0.01178441196680069,
+ "learning_rate": 0.0006,
+ "loss": 7.23576545715332,
+ "step": 3042
+ },
+ {
+ "epoch": 42.26550218340611,
+ "grad_norm": 0.014518891461193562,
+ "learning_rate": 0.0006,
+ "loss": 7.253895282745361,
+ "step": 3043
+ },
+ {
+ "epoch": 42.27947598253275,
+ "grad_norm": 0.018159950152039528,
+ "learning_rate": 0.0006,
+ "loss": 7.278058052062988,
+ "step": 3044
+ },
+ {
+ "epoch": 42.29344978165939,
+ "grad_norm": 0.007587796077132225,
+ "learning_rate": 0.0006,
+ "loss": 7.278578758239746,
+ "step": 3045
+ },
+ {
+ "epoch": 42.30742358078603,
+ "grad_norm": 0.015994979068636894,
+ "learning_rate": 0.0006,
+ "loss": 7.239841461181641,
+ "step": 3046
+ },
+ {
+ "epoch": 42.32139737991266,
+ "grad_norm": 0.017830168828368187,
+ "learning_rate": 0.0006,
+ "loss": 7.2443976402282715,
+ "step": 3047
+ },
+ {
+ "epoch": 42.3353711790393,
+ "grad_norm": 0.008738759905099869,
+ "learning_rate": 0.0006,
+ "loss": 7.232264518737793,
+ "step": 3048
+ },
+ {
+ "epoch": 42.34934497816594,
+ "grad_norm": 0.0175449438393116,
+ "learning_rate": 0.0006,
+ "loss": 7.228000640869141,
+ "step": 3049
+ },
+ {
+ "epoch": 42.36331877729258,
+ "grad_norm": 0.019611820578575134,
+ "learning_rate": 0.0006,
+ "loss": 7.234999179840088,
+ "step": 3050
+ },
+ {
+ "epoch": 42.377292576419215,
+ "grad_norm": 0.015302863903343678,
+ "learning_rate": 0.0006,
+ "loss": 7.242547988891602,
+ "step": 3051
+ },
+ {
+ "epoch": 42.391266375545854,
+ "grad_norm": 0.025919731706380844,
+ "learning_rate": 0.0006,
+ "loss": 7.216652870178223,
+ "step": 3052
+ },
+ {
+ "epoch": 42.40524017467249,
+ "grad_norm": 0.034587468951940536,
+ "learning_rate": 0.0006,
+ "loss": 7.245336532592773,
+ "step": 3053
+ },
+ {
+ "epoch": 42.419213973799124,
+ "grad_norm": 0.019198840484023094,
+ "learning_rate": 0.0006,
+ "loss": 7.256758689880371,
+ "step": 3054
+ },
+ {
+ "epoch": 42.43318777292576,
+ "grad_norm": 0.012316389009356499,
+ "learning_rate": 0.0006,
+ "loss": 7.209102630615234,
+ "step": 3055
+ },
+ {
+ "epoch": 42.4471615720524,
+ "grad_norm": 0.023512162268161774,
+ "learning_rate": 0.0006,
+ "loss": 7.250742435455322,
+ "step": 3056
+ },
+ {
+ "epoch": 42.46113537117904,
+ "grad_norm": 0.025457851588726044,
+ "learning_rate": 0.0006,
+ "loss": 7.263757705688477,
+ "step": 3057
+ },
+ {
+ "epoch": 42.47510917030568,
+ "grad_norm": 0.02192617766559124,
+ "learning_rate": 0.0006,
+ "loss": 7.246977806091309,
+ "step": 3058
+ },
+ {
+ "epoch": 42.48908296943232,
+ "grad_norm": 0.02541281282901764,
+ "learning_rate": 0.0006,
+ "loss": 7.232361316680908,
+ "step": 3059
+ },
+ {
+ "epoch": 42.50305676855895,
+ "grad_norm": 0.036353763192892075,
+ "learning_rate": 0.0006,
+ "loss": 7.242556571960449,
+ "step": 3060
+ },
+ {
+ "epoch": 42.51703056768559,
+ "grad_norm": 0.04555872827768326,
+ "learning_rate": 0.0006,
+ "loss": 7.228940963745117,
+ "step": 3061
+ },
+ {
+ "epoch": 42.531004366812226,
+ "grad_norm": 0.04563378915190697,
+ "learning_rate": 0.0006,
+ "loss": 7.235276222229004,
+ "step": 3062
+ },
+ {
+ "epoch": 42.544978165938865,
+ "grad_norm": 0.04275760427117348,
+ "learning_rate": 0.0006,
+ "loss": 7.238895416259766,
+ "step": 3063
+ },
+ {
+ "epoch": 42.5589519650655,
+ "grad_norm": 0.04535885155200958,
+ "learning_rate": 0.0006,
+ "loss": 7.205336570739746,
+ "step": 3064
+ },
+ {
+ "epoch": 42.57292576419214,
+ "grad_norm": 0.015534617006778717,
+ "learning_rate": 0.0006,
+ "loss": 7.238393783569336,
+ "step": 3065
+ },
+ {
+ "epoch": 42.58689956331878,
+ "grad_norm": 0.03286299481987953,
+ "learning_rate": 0.0006,
+ "loss": 7.216453552246094,
+ "step": 3066
+ },
+ {
+ "epoch": 42.60087336244541,
+ "grad_norm": 0.025453168898820877,
+ "learning_rate": 0.0006,
+ "loss": 7.203750133514404,
+ "step": 3067
+ },
+ {
+ "epoch": 42.61484716157205,
+ "grad_norm": 0.021501649171113968,
+ "learning_rate": 0.0006,
+ "loss": 7.241621971130371,
+ "step": 3068
+ },
+ {
+ "epoch": 42.62882096069869,
+ "grad_norm": 0.02743634767830372,
+ "learning_rate": 0.0006,
+ "loss": 7.22076940536499,
+ "step": 3069
+ },
+ {
+ "epoch": 42.64279475982533,
+ "grad_norm": 0.020890483632683754,
+ "learning_rate": 0.0006,
+ "loss": 7.218412399291992,
+ "step": 3070
+ },
+ {
+ "epoch": 42.65676855895197,
+ "grad_norm": 0.022499850019812584,
+ "learning_rate": 0.0006,
+ "loss": 7.217696189880371,
+ "step": 3071
+ },
+ {
+ "epoch": 42.670742358078606,
+ "grad_norm": 0.01653696969151497,
+ "learning_rate": 0.0006,
+ "loss": 7.156417369842529,
+ "step": 3072
+ },
+ {
+ "epoch": 42.68471615720524,
+ "grad_norm": 0.01834225095808506,
+ "learning_rate": 0.0006,
+ "loss": 7.211247444152832,
+ "step": 3073
+ },
+ {
+ "epoch": 42.698689956331876,
+ "grad_norm": 0.020137229934334755,
+ "learning_rate": 0.0006,
+ "loss": 7.232499599456787,
+ "step": 3074
+ },
+ {
+ "epoch": 42.712663755458514,
+ "grad_norm": 0.012671459466218948,
+ "learning_rate": 0.0006,
+ "loss": 7.172441482543945,
+ "step": 3075
+ },
+ {
+ "epoch": 42.72663755458515,
+ "grad_norm": 0.016803130507469177,
+ "learning_rate": 0.0006,
+ "loss": 7.152103424072266,
+ "step": 3076
+ },
+ {
+ "epoch": 42.74061135371179,
+ "grad_norm": 0.01623712107539177,
+ "learning_rate": 0.0006,
+ "loss": 7.187032699584961,
+ "step": 3077
+ },
+ {
+ "epoch": 42.75458515283843,
+ "grad_norm": 0.015938356518745422,
+ "learning_rate": 0.0006,
+ "loss": 7.171140670776367,
+ "step": 3078
+ },
+ {
+ "epoch": 42.76855895196506,
+ "grad_norm": 0.02580736204981804,
+ "learning_rate": 0.0006,
+ "loss": 7.204472541809082,
+ "step": 3079
+ },
+ {
+ "epoch": 42.7825327510917,
+ "grad_norm": 0.023372117429971695,
+ "learning_rate": 0.0006,
+ "loss": 7.165554523468018,
+ "step": 3080
+ },
+ {
+ "epoch": 42.79650655021834,
+ "grad_norm": 0.04440586268901825,
+ "learning_rate": 0.0006,
+ "loss": 7.169431209564209,
+ "step": 3081
+ },
+ {
+ "epoch": 42.81048034934498,
+ "grad_norm": 0.06797406077384949,
+ "learning_rate": 0.0006,
+ "loss": 7.168193340301514,
+ "step": 3082
+ },
+ {
+ "epoch": 42.82445414847162,
+ "grad_norm": 0.0914321169257164,
+ "learning_rate": 0.0006,
+ "loss": 7.204671859741211,
+ "step": 3083
+ },
+ {
+ "epoch": 42.838427947598255,
+ "grad_norm": 0.0506940595805645,
+ "learning_rate": 0.0006,
+ "loss": 7.15949821472168,
+ "step": 3084
+ },
+ {
+ "epoch": 42.852401746724894,
+ "grad_norm": 0.04819890484213829,
+ "learning_rate": 0.0006,
+ "loss": 7.192837238311768,
+ "step": 3085
+ },
+ {
+ "epoch": 42.866375545851525,
+ "grad_norm": 0.040044233202934265,
+ "learning_rate": 0.0006,
+ "loss": 7.1600165367126465,
+ "step": 3086
+ },
+ {
+ "epoch": 42.880349344978164,
+ "grad_norm": 0.02573299966752529,
+ "learning_rate": 0.0006,
+ "loss": 7.153427600860596,
+ "step": 3087
+ },
+ {
+ "epoch": 42.8943231441048,
+ "grad_norm": 0.046942535787820816,
+ "learning_rate": 0.0006,
+ "loss": 7.16767692565918,
+ "step": 3088
+ },
+ {
+ "epoch": 42.90829694323144,
+ "grad_norm": 0.02683161571621895,
+ "learning_rate": 0.0006,
+ "loss": 7.17144250869751,
+ "step": 3089
+ },
+ {
+ "epoch": 42.92227074235808,
+ "grad_norm": 0.05363674834370613,
+ "learning_rate": 0.0006,
+ "loss": 7.120128631591797,
+ "step": 3090
+ },
+ {
+ "epoch": 42.93624454148472,
+ "grad_norm": 0.023756247013807297,
+ "learning_rate": 0.0006,
+ "loss": 7.169703006744385,
+ "step": 3091
+ },
+ {
+ "epoch": 42.95021834061135,
+ "grad_norm": 0.04216783866286278,
+ "learning_rate": 0.0006,
+ "loss": 7.1206865310668945,
+ "step": 3092
+ },
+ {
+ "epoch": 42.96419213973799,
+ "grad_norm": 0.02219565585255623,
+ "learning_rate": 0.0006,
+ "loss": 7.157215118408203,
+ "step": 3093
+ },
+ {
+ "epoch": 42.97816593886463,
+ "grad_norm": 0.03527127206325531,
+ "learning_rate": 0.0006,
+ "loss": 7.166672706604004,
+ "step": 3094
+ },
+ {
+ "epoch": 42.992139737991266,
+ "grad_norm": 0.017748817801475525,
+ "learning_rate": 0.0006,
+ "loss": 7.154747009277344,
+ "step": 3095
+ },
+ {
+ "epoch": 43.0,
+ "grad_norm": 0.033165838569402695,
+ "learning_rate": 0.0006,
+ "loss": 7.150866508483887,
+ "step": 3096
+ },
+ {
+ "epoch": 43.0,
+ "eval_loss": 7.144403457641602,
+ "eval_runtime": 56.8853,
+ "eval_samples_per_second": 42.928,
+ "eval_steps_per_second": 1.354,
+ "step": 3096
+ },
+ {
+ "epoch": 43.01397379912664,
+ "grad_norm": 0.015199829824268818,
+ "learning_rate": 0.0006,
+ "loss": 7.122735023498535,
+ "step": 3097
+ },
+ {
+ "epoch": 43.02794759825328,
+ "grad_norm": 0.03220256045460701,
+ "learning_rate": 0.0006,
+ "loss": 7.143553733825684,
+ "step": 3098
+ },
+ {
+ "epoch": 43.041921397379916,
+ "grad_norm": 0.014687920920550823,
+ "learning_rate": 0.0006,
+ "loss": 7.080960750579834,
+ "step": 3099
+ },
+ {
+ "epoch": 43.05589519650655,
+ "grad_norm": 0.023522932082414627,
+ "learning_rate": 0.0006,
+ "loss": 7.137356758117676,
+ "step": 3100
+ },
+ {
+ "epoch": 43.069868995633186,
+ "grad_norm": 0.014808216132223606,
+ "learning_rate": 0.0006,
+ "loss": 7.09143590927124,
+ "step": 3101
+ },
+ {
+ "epoch": 43.083842794759825,
+ "grad_norm": 0.029339246451854706,
+ "learning_rate": 0.0006,
+ "loss": 7.122198104858398,
+ "step": 3102
+ },
+ {
+ "epoch": 43.09781659388646,
+ "grad_norm": 0.022953078150749207,
+ "learning_rate": 0.0006,
+ "loss": 7.092505931854248,
+ "step": 3103
+ },
+ {
+ "epoch": 43.1117903930131,
+ "grad_norm": 0.04590052738785744,
+ "learning_rate": 0.0006,
+ "loss": 7.113100051879883,
+ "step": 3104
+ },
+ {
+ "epoch": 43.12576419213974,
+ "grad_norm": 0.0524442233145237,
+ "learning_rate": 0.0006,
+ "loss": 7.147022247314453,
+ "step": 3105
+ },
+ {
+ "epoch": 43.13973799126637,
+ "grad_norm": 0.0315794013440609,
+ "learning_rate": 0.0006,
+ "loss": 7.094045162200928,
+ "step": 3106
+ },
+ {
+ "epoch": 43.15371179039301,
+ "grad_norm": 0.05112722888588905,
+ "learning_rate": 0.0006,
+ "loss": 7.054388523101807,
+ "step": 3107
+ },
+ {
+ "epoch": 43.16768558951965,
+ "grad_norm": 0.03146776929497719,
+ "learning_rate": 0.0006,
+ "loss": 7.116674423217773,
+ "step": 3108
+ },
+ {
+ "epoch": 43.18165938864629,
+ "grad_norm": 0.03391246497631073,
+ "learning_rate": 0.0006,
+ "loss": 7.150345802307129,
+ "step": 3109
+ },
+ {
+ "epoch": 43.19563318777293,
+ "grad_norm": 0.026957903057336807,
+ "learning_rate": 0.0006,
+ "loss": 7.058531284332275,
+ "step": 3110
+ },
+ {
+ "epoch": 43.209606986899566,
+ "grad_norm": 0.041532568633556366,
+ "learning_rate": 0.0006,
+ "loss": 7.085498809814453,
+ "step": 3111
+ },
+ {
+ "epoch": 43.223580786026204,
+ "grad_norm": 0.020001588389277458,
+ "learning_rate": 0.0006,
+ "loss": 7.131487846374512,
+ "step": 3112
+ },
+ {
+ "epoch": 43.237554585152836,
+ "grad_norm": 0.02268712781369686,
+ "learning_rate": 0.0006,
+ "loss": 7.0790863037109375,
+ "step": 3113
+ },
+ {
+ "epoch": 43.251528384279474,
+ "grad_norm": 0.023363925516605377,
+ "learning_rate": 0.0006,
+ "loss": 7.067816734313965,
+ "step": 3114
+ },
+ {
+ "epoch": 43.26550218340611,
+ "grad_norm": 0.019367508590221405,
+ "learning_rate": 0.0006,
+ "loss": 7.042041778564453,
+ "step": 3115
+ },
+ {
+ "epoch": 43.27947598253275,
+ "grad_norm": 0.02745090238749981,
+ "learning_rate": 0.0006,
+ "loss": 7.064733028411865,
+ "step": 3116
+ },
+ {
+ "epoch": 43.29344978165939,
+ "grad_norm": 0.02543802745640278,
+ "learning_rate": 0.0006,
+ "loss": 7.058948040008545,
+ "step": 3117
+ },
+ {
+ "epoch": 43.30742358078603,
+ "grad_norm": 0.018874092027544975,
+ "learning_rate": 0.0006,
+ "loss": 7.009946346282959,
+ "step": 3118
+ },
+ {
+ "epoch": 43.32139737991266,
+ "grad_norm": 0.024662822484970093,
+ "learning_rate": 0.0006,
+ "loss": 7.015913963317871,
+ "step": 3119
+ },
+ {
+ "epoch": 43.3353711790393,
+ "grad_norm": 0.02441416308283806,
+ "learning_rate": 0.0006,
+ "loss": 7.061622142791748,
+ "step": 3120
+ },
+ {
+ "epoch": 43.34934497816594,
+ "grad_norm": 0.031887322664260864,
+ "learning_rate": 0.0006,
+ "loss": 7.046883583068848,
+ "step": 3121
+ },
+ {
+ "epoch": 43.36331877729258,
+ "grad_norm": 0.021930739283561707,
+ "learning_rate": 0.0006,
+ "loss": 7.0309624671936035,
+ "step": 3122
+ },
+ {
+ "epoch": 43.377292576419215,
+ "grad_norm": 0.02722783200442791,
+ "learning_rate": 0.0006,
+ "loss": 7.072924613952637,
+ "step": 3123
+ },
+ {
+ "epoch": 43.391266375545854,
+ "grad_norm": 0.01483179721981287,
+ "learning_rate": 0.0006,
+ "loss": 7.028624534606934,
+ "step": 3124
+ },
+ {
+ "epoch": 43.40524017467249,
+ "grad_norm": 0.024793945252895355,
+ "learning_rate": 0.0006,
+ "loss": 7.001623153686523,
+ "step": 3125
+ },
+ {
+ "epoch": 43.419213973799124,
+ "grad_norm": 0.020148713141679764,
+ "learning_rate": 0.0006,
+ "loss": 6.9938812255859375,
+ "step": 3126
+ },
+ {
+ "epoch": 43.43318777292576,
+ "grad_norm": 0.02413850650191307,
+ "learning_rate": 0.0006,
+ "loss": 7.001243591308594,
+ "step": 3127
+ },
+ {
+ "epoch": 43.4471615720524,
+ "grad_norm": 0.028549203649163246,
+ "learning_rate": 0.0006,
+ "loss": 6.987671375274658,
+ "step": 3128
+ },
+ {
+ "epoch": 43.46113537117904,
+ "grad_norm": 0.025988785549998283,
+ "learning_rate": 0.0006,
+ "loss": 6.9551167488098145,
+ "step": 3129
+ },
+ {
+ "epoch": 43.47510917030568,
+ "grad_norm": 0.027140729129314423,
+ "learning_rate": 0.0006,
+ "loss": 6.9547810554504395,
+ "step": 3130
+ },
+ {
+ "epoch": 43.48908296943232,
+ "grad_norm": 0.030130082741379738,
+ "learning_rate": 0.0006,
+ "loss": 6.993186950683594,
+ "step": 3131
+ },
+ {
+ "epoch": 43.50305676855895,
+ "grad_norm": 0.025996994227170944,
+ "learning_rate": 0.0006,
+ "loss": 6.984624862670898,
+ "step": 3132
+ },
+ {
+ "epoch": 43.51703056768559,
+ "grad_norm": 0.02423008903861046,
+ "learning_rate": 0.0006,
+ "loss": 6.98813009262085,
+ "step": 3133
+ },
+ {
+ "epoch": 43.531004366812226,
+ "grad_norm": 0.029665837064385414,
+ "learning_rate": 0.0006,
+ "loss": 6.986610412597656,
+ "step": 3134
+ },
+ {
+ "epoch": 43.544978165938865,
+ "grad_norm": 0.028768595308065414,
+ "learning_rate": 0.0006,
+ "loss": 6.898797035217285,
+ "step": 3135
+ },
+ {
+ "epoch": 43.5589519650655,
+ "grad_norm": 0.03111208975315094,
+ "learning_rate": 0.0006,
+ "loss": 6.923741817474365,
+ "step": 3136
+ },
+ {
+ "epoch": 43.57292576419214,
+ "grad_norm": 0.0368187390267849,
+ "learning_rate": 0.0006,
+ "loss": 6.994343280792236,
+ "step": 3137
+ },
+ {
+ "epoch": 43.58689956331878,
+ "grad_norm": 0.04076763987541199,
+ "learning_rate": 0.0006,
+ "loss": 6.909458160400391,
+ "step": 3138
+ },
+ {
+ "epoch": 43.60087336244541,
+ "grad_norm": 0.03281061723828316,
+ "learning_rate": 0.0006,
+ "loss": 6.924851417541504,
+ "step": 3139
+ },
+ {
+ "epoch": 43.61484716157205,
+ "grad_norm": 0.025486988946795464,
+ "learning_rate": 0.0006,
+ "loss": 6.962926864624023,
+ "step": 3140
+ },
+ {
+ "epoch": 43.62882096069869,
+ "grad_norm": 0.04583485797047615,
+ "learning_rate": 0.0006,
+ "loss": 6.876555442810059,
+ "step": 3141
+ },
+ {
+ "epoch": 43.64279475982533,
+ "grad_norm": 0.05460785701870918,
+ "learning_rate": 0.0006,
+ "loss": 6.952932357788086,
+ "step": 3142
+ },
+ {
+ "epoch": 43.65676855895197,
+ "grad_norm": 0.040917906910181046,
+ "learning_rate": 0.0006,
+ "loss": 6.9624924659729,
+ "step": 3143
+ },
+ {
+ "epoch": 43.670742358078606,
+ "grad_norm": 0.04009668901562691,
+ "learning_rate": 0.0006,
+ "loss": 6.94033145904541,
+ "step": 3144
+ },
+ {
+ "epoch": 43.68471615720524,
+ "grad_norm": 0.04500625655055046,
+ "learning_rate": 0.0006,
+ "loss": 6.961839199066162,
+ "step": 3145
+ },
+ {
+ "epoch": 43.698689956331876,
+ "grad_norm": 0.03182258829474449,
+ "learning_rate": 0.0006,
+ "loss": 6.930580139160156,
+ "step": 3146
+ },
+ {
+ "epoch": 43.712663755458514,
+ "grad_norm": 0.053621504455804825,
+ "learning_rate": 0.0006,
+ "loss": 6.950235366821289,
+ "step": 3147
+ },
+ {
+ "epoch": 43.72663755458515,
+ "grad_norm": 0.031919289380311966,
+ "learning_rate": 0.0006,
+ "loss": 6.895111560821533,
+ "step": 3148
+ },
+ {
+ "epoch": 43.74061135371179,
+ "grad_norm": 0.042713891714811325,
+ "learning_rate": 0.0006,
+ "loss": 6.940885543823242,
+ "step": 3149
+ },
+ {
+ "epoch": 43.75458515283843,
+ "grad_norm": 0.02385653182864189,
+ "learning_rate": 0.0006,
+ "loss": 6.948801517486572,
+ "step": 3150
+ },
+ {
+ "epoch": 43.76855895196506,
+ "grad_norm": 0.04100663587450981,
+ "learning_rate": 0.0006,
+ "loss": 6.970411777496338,
+ "step": 3151
+ },
+ {
+ "epoch": 43.7825327510917,
+ "grad_norm": 0.02729583904147148,
+ "learning_rate": 0.0006,
+ "loss": 6.899379253387451,
+ "step": 3152
+ },
+ {
+ "epoch": 43.79650655021834,
+ "grad_norm": 0.03128620237112045,
+ "learning_rate": 0.0006,
+ "loss": 6.931319236755371,
+ "step": 3153
+ },
+ {
+ "epoch": 43.81048034934498,
+ "grad_norm": 0.025488659739494324,
+ "learning_rate": 0.0006,
+ "loss": 6.8486409187316895,
+ "step": 3154
+ },
+ {
+ "epoch": 43.82445414847162,
+ "grad_norm": 0.03679376840591431,
+ "learning_rate": 0.0006,
+ "loss": 6.840811252593994,
+ "step": 3155
+ },
+ {
+ "epoch": 43.838427947598255,
+ "grad_norm": 0.020578034222126007,
+ "learning_rate": 0.0006,
+ "loss": 6.9088616371154785,
+ "step": 3156
+ },
+ {
+ "epoch": 43.852401746724894,
+ "grad_norm": 0.02935841865837574,
+ "learning_rate": 0.0006,
+ "loss": 6.806676864624023,
+ "step": 3157
+ },
+ {
+ "epoch": 43.866375545851525,
+ "grad_norm": 0.023319244384765625,
+ "learning_rate": 0.0006,
+ "loss": 6.878457546234131,
+ "step": 3158
+ },
+ {
+ "epoch": 43.880349344978164,
+ "grad_norm": 0.02779589593410492,
+ "learning_rate": 0.0006,
+ "loss": 6.822447776794434,
+ "step": 3159
+ },
+ {
+ "epoch": 43.8943231441048,
+ "grad_norm": 0.01879825070500374,
+ "learning_rate": 0.0006,
+ "loss": 6.854121208190918,
+ "step": 3160
+ },
+ {
+ "epoch": 43.90829694323144,
+ "grad_norm": 0.023286139592528343,
+ "learning_rate": 0.0006,
+ "loss": 6.836827278137207,
+ "step": 3161
+ },
+ {
+ "epoch": 43.92227074235808,
+ "grad_norm": 0.030358297750353813,
+ "learning_rate": 0.0006,
+ "loss": 6.860037803649902,
+ "step": 3162
+ },
+ {
+ "epoch": 43.93624454148472,
+ "grad_norm": 0.029621105641126633,
+ "learning_rate": 0.0006,
+ "loss": 6.884025573730469,
+ "step": 3163
+ },
+ {
+ "epoch": 43.95021834061135,
+ "grad_norm": 0.029829824343323708,
+ "learning_rate": 0.0006,
+ "loss": 6.8702392578125,
+ "step": 3164
+ },
+ {
+ "epoch": 43.96419213973799,
+ "grad_norm": 0.036892473697662354,
+ "learning_rate": 0.0006,
+ "loss": 6.806652069091797,
+ "step": 3165
+ },
+ {
+ "epoch": 43.97816593886463,
+ "grad_norm": 0.03206709027290344,
+ "learning_rate": 0.0006,
+ "loss": 6.828495979309082,
+ "step": 3166
+ },
+ {
+ "epoch": 43.992139737991266,
+ "grad_norm": 0.02744000405073166,
+ "learning_rate": 0.0006,
+ "loss": 6.832002639770508,
+ "step": 3167
+ },
+ {
+ "epoch": 44.0,
+ "grad_norm": 0.02205660194158554,
+ "learning_rate": 0.0006,
+ "loss": 6.898828506469727,
+ "step": 3168
+ },
+ {
+ "epoch": 44.0,
+ "eval_loss": 6.8150553703308105,
+ "eval_runtime": 56.784,
+ "eval_samples_per_second": 43.005,
+ "eval_steps_per_second": 1.356,
+ "step": 3168
+ },
+ {
+ "epoch": 44.01397379912664,
+ "grad_norm": 0.030421260744333267,
+ "learning_rate": 0.0006,
+ "loss": 6.775057792663574,
+ "step": 3169
+ },
+ {
+ "epoch": 44.02794759825328,
+ "grad_norm": 0.0465611033141613,
+ "learning_rate": 0.0006,
+ "loss": 6.849390506744385,
+ "step": 3170
+ },
+ {
+ "epoch": 44.041921397379916,
+ "grad_norm": 0.05977972596883774,
+ "learning_rate": 0.0006,
+ "loss": 6.725576400756836,
+ "step": 3171
+ },
+ {
+ "epoch": 44.05589519650655,
+ "grad_norm": 0.05432697385549545,
+ "learning_rate": 0.0006,
+ "loss": 6.768024444580078,
+ "step": 3172
+ },
+ {
+ "epoch": 44.069868995633186,
+ "grad_norm": 0.04850967228412628,
+ "learning_rate": 0.0006,
+ "loss": 6.799457550048828,
+ "step": 3173
+ },
+ {
+ "epoch": 44.083842794759825,
+ "grad_norm": 0.06201786920428276,
+ "learning_rate": 0.0006,
+ "loss": 6.897234916687012,
+ "step": 3174
+ },
+ {
+ "epoch": 44.09781659388646,
+ "grad_norm": 0.04695591703057289,
+ "learning_rate": 0.0006,
+ "loss": 6.826470375061035,
+ "step": 3175
+ },
+ {
+ "epoch": 44.1117903930131,
+ "grad_norm": 0.03850164636969566,
+ "learning_rate": 0.0006,
+ "loss": 6.835901260375977,
+ "step": 3176
+ },
+ {
+ "epoch": 44.12576419213974,
+ "grad_norm": 0.04250155761837959,
+ "learning_rate": 0.0006,
+ "loss": 6.782310485839844,
+ "step": 3177
+ },
+ {
+ "epoch": 44.13973799126637,
+ "grad_norm": 0.041465479880571365,
+ "learning_rate": 0.0006,
+ "loss": 6.785732746124268,
+ "step": 3178
+ },
+ {
+ "epoch": 44.15371179039301,
+ "grad_norm": 0.02527807280421257,
+ "learning_rate": 0.0006,
+ "loss": 6.78061580657959,
+ "step": 3179
+ },
+ {
+ "epoch": 44.16768558951965,
+ "grad_norm": 0.03660374879837036,
+ "learning_rate": 0.0006,
+ "loss": 6.783992767333984,
+ "step": 3180
+ },
+ {
+ "epoch": 44.18165938864629,
+ "grad_norm": 0.027567964047193527,
+ "learning_rate": 0.0006,
+ "loss": 6.765413284301758,
+ "step": 3181
+ },
+ {
+ "epoch": 44.19563318777293,
+ "grad_norm": 0.024301424622535706,
+ "learning_rate": 0.0006,
+ "loss": 6.760368347167969,
+ "step": 3182
+ },
+ {
+ "epoch": 44.209606986899566,
+ "grad_norm": 0.0266821701079607,
+ "learning_rate": 0.0006,
+ "loss": 6.71417236328125,
+ "step": 3183
+ },
+ {
+ "epoch": 44.223580786026204,
+ "grad_norm": 0.028125090524554253,
+ "learning_rate": 0.0006,
+ "loss": 6.646126747131348,
+ "step": 3184
+ },
+ {
+ "epoch": 44.237554585152836,
+ "grad_norm": 0.0185737032443285,
+ "learning_rate": 0.0006,
+ "loss": 6.712944030761719,
+ "step": 3185
+ },
+ {
+ "epoch": 44.251528384279474,
+ "grad_norm": 0.025836674496531487,
+ "learning_rate": 0.0006,
+ "loss": 6.707030296325684,
+ "step": 3186
+ },
+ {
+ "epoch": 44.26550218340611,
+ "grad_norm": 0.029744870960712433,
+ "learning_rate": 0.0006,
+ "loss": 6.6271071434021,
+ "step": 3187
+ },
+ {
+ "epoch": 44.27947598253275,
+ "grad_norm": 0.03021426685154438,
+ "learning_rate": 0.0006,
+ "loss": 6.677826404571533,
+ "step": 3188
+ },
+ {
+ "epoch": 44.29344978165939,
+ "grad_norm": 0.032727014273405075,
+ "learning_rate": 0.0006,
+ "loss": 6.657764434814453,
+ "step": 3189
+ },
+ {
+ "epoch": 44.30742358078603,
+ "grad_norm": 0.03696778789162636,
+ "learning_rate": 0.0006,
+ "loss": 6.632883548736572,
+ "step": 3190
+ },
+ {
+ "epoch": 44.32139737991266,
+ "grad_norm": 0.034630827605724335,
+ "learning_rate": 0.0006,
+ "loss": 6.759583473205566,
+ "step": 3191
+ },
+ {
+ "epoch": 44.3353711790393,
+ "grad_norm": 0.030517758801579475,
+ "learning_rate": 0.0006,
+ "loss": 6.702268600463867,
+ "step": 3192
+ },
+ {
+ "epoch": 44.34934497816594,
+ "grad_norm": 0.04105190932750702,
+ "learning_rate": 0.0006,
+ "loss": 6.705792427062988,
+ "step": 3193
+ },
+ {
+ "epoch": 44.36331877729258,
+ "grad_norm": 0.06659174710512161,
+ "learning_rate": 0.0006,
+ "loss": 6.73862361907959,
+ "step": 3194
+ },
+ {
+ "epoch": 44.377292576419215,
+ "grad_norm": 0.08013619482517242,
+ "learning_rate": 0.0006,
+ "loss": 6.710282802581787,
+ "step": 3195
+ },
+ {
+ "epoch": 44.391266375545854,
+ "grad_norm": 0.049850963056087494,
+ "learning_rate": 0.0006,
+ "loss": 6.684974670410156,
+ "step": 3196
+ },
+ {
+ "epoch": 44.40524017467249,
+ "grad_norm": 0.061247408390045166,
+ "learning_rate": 0.0006,
+ "loss": 6.716034412384033,
+ "step": 3197
+ },
+ {
+ "epoch": 44.419213973799124,
+ "grad_norm": 0.05868857726454735,
+ "learning_rate": 0.0006,
+ "loss": 6.589107513427734,
+ "step": 3198
+ },
+ {
+ "epoch": 44.43318777292576,
+ "grad_norm": 0.048878561705350876,
+ "learning_rate": 0.0006,
+ "loss": 6.694855690002441,
+ "step": 3199
+ },
+ {
+ "epoch": 44.4471615720524,
+ "grad_norm": 0.028748324140906334,
+ "learning_rate": 0.0006,
+ "loss": 6.629969596862793,
+ "step": 3200
+ },
+ {
+ "epoch": 44.46113537117904,
+ "grad_norm": 0.04060306400060654,
+ "learning_rate": 0.0006,
+ "loss": 6.721431255340576,
+ "step": 3201
+ },
+ {
+ "epoch": 44.47510917030568,
+ "grad_norm": 0.03833824768662453,
+ "learning_rate": 0.0006,
+ "loss": 6.625029563903809,
+ "step": 3202
+ },
+ {
+ "epoch": 44.48908296943232,
+ "grad_norm": 0.03791997581720352,
+ "learning_rate": 0.0006,
+ "loss": 6.669187545776367,
+ "step": 3203
+ },
+ {
+ "epoch": 44.50305676855895,
+ "grad_norm": 0.03259684890508652,
+ "learning_rate": 0.0006,
+ "loss": 6.665643692016602,
+ "step": 3204
+ },
+ {
+ "epoch": 44.51703056768559,
+ "grad_norm": 0.040658626705408096,
+ "learning_rate": 0.0006,
+ "loss": 6.639796257019043,
+ "step": 3205
+ },
+ {
+ "epoch": 44.531004366812226,
+ "grad_norm": 0.03691193461418152,
+ "learning_rate": 0.0006,
+ "loss": 6.692244529724121,
+ "step": 3206
+ },
+ {
+ "epoch": 44.544978165938865,
+ "grad_norm": 0.02814812771975994,
+ "learning_rate": 0.0006,
+ "loss": 6.61978006362915,
+ "step": 3207
+ },
+ {
+ "epoch": 44.5589519650655,
+ "grad_norm": 0.022155074402689934,
+ "learning_rate": 0.0006,
+ "loss": 6.625905990600586,
+ "step": 3208
+ },
+ {
+ "epoch": 44.57292576419214,
+ "grad_norm": 0.030794885009527206,
+ "learning_rate": 0.0006,
+ "loss": 6.678500652313232,
+ "step": 3209
+ },
+ {
+ "epoch": 44.58689956331878,
+ "grad_norm": 0.03201768547296524,
+ "learning_rate": 0.0006,
+ "loss": 6.6015729904174805,
+ "step": 3210
+ },
+ {
+ "epoch": 44.60087336244541,
+ "grad_norm": 0.025138361379504204,
+ "learning_rate": 0.0006,
+ "loss": 6.568234443664551,
+ "step": 3211
+ },
+ {
+ "epoch": 44.61484716157205,
+ "grad_norm": 0.02969353087246418,
+ "learning_rate": 0.0006,
+ "loss": 6.588366508483887,
+ "step": 3212
+ },
+ {
+ "epoch": 44.62882096069869,
+ "grad_norm": 0.03136962279677391,
+ "learning_rate": 0.0006,
+ "loss": 6.479961395263672,
+ "step": 3213
+ },
+ {
+ "epoch": 44.64279475982533,
+ "grad_norm": 0.030178170651197433,
+ "learning_rate": 0.0006,
+ "loss": 6.622910976409912,
+ "step": 3214
+ },
+ {
+ "epoch": 44.65676855895197,
+ "grad_norm": 0.02682245895266533,
+ "learning_rate": 0.0006,
+ "loss": 6.597707748413086,
+ "step": 3215
+ },
+ {
+ "epoch": 44.670742358078606,
+ "grad_norm": 0.025154557079076767,
+ "learning_rate": 0.0006,
+ "loss": 6.630541801452637,
+ "step": 3216
+ },
+ {
+ "epoch": 44.68471615720524,
+ "grad_norm": 0.03564463183283806,
+ "learning_rate": 0.0006,
+ "loss": 6.542823791503906,
+ "step": 3217
+ },
+ {
+ "epoch": 44.698689956331876,
+ "grad_norm": 0.049477651715278625,
+ "learning_rate": 0.0006,
+ "loss": 6.646030426025391,
+ "step": 3218
+ },
+ {
+ "epoch": 44.712663755458514,
+ "grad_norm": 0.06536535918712616,
+ "learning_rate": 0.0006,
+ "loss": 6.595463275909424,
+ "step": 3219
+ },
+ {
+ "epoch": 44.72663755458515,
+ "grad_norm": 0.07710316777229309,
+ "learning_rate": 0.0006,
+ "loss": 6.55524206161499,
+ "step": 3220
+ },
+ {
+ "epoch": 44.74061135371179,
+ "grad_norm": 0.09571163356304169,
+ "learning_rate": 0.0006,
+ "loss": 6.659764289855957,
+ "step": 3221
+ },
+ {
+ "epoch": 44.75458515283843,
+ "grad_norm": 0.055076830089092255,
+ "learning_rate": 0.0006,
+ "loss": 6.658882141113281,
+ "step": 3222
+ },
+ {
+ "epoch": 44.76855895196506,
+ "grad_norm": 0.07671099156141281,
+ "learning_rate": 0.0006,
+ "loss": 6.605809211730957,
+ "step": 3223
+ },
+ {
+ "epoch": 44.7825327510917,
+ "grad_norm": 0.05295584350824356,
+ "learning_rate": 0.0006,
+ "loss": 6.641455173492432,
+ "step": 3224
+ },
+ {
+ "epoch": 44.79650655021834,
+ "grad_norm": 0.09128042310476303,
+ "learning_rate": 0.0006,
+ "loss": 6.691066741943359,
+ "step": 3225
+ },
+ {
+ "epoch": 44.81048034934498,
+ "grad_norm": 0.05362958461046219,
+ "learning_rate": 0.0006,
+ "loss": 6.573406219482422,
+ "step": 3226
+ },
+ {
+ "epoch": 44.82445414847162,
+ "grad_norm": 0.046910062432289124,
+ "learning_rate": 0.0006,
+ "loss": 6.550876617431641,
+ "step": 3227
+ },
+ {
+ "epoch": 44.838427947598255,
+ "grad_norm": 0.03494197130203247,
+ "learning_rate": 0.0006,
+ "loss": 6.540179252624512,
+ "step": 3228
+ },
+ {
+ "epoch": 44.852401746724894,
+ "grad_norm": 0.04716387391090393,
+ "learning_rate": 0.0006,
+ "loss": 6.490459442138672,
+ "step": 3229
+ },
+ {
+ "epoch": 44.866375545851525,
+ "grad_norm": 0.04190506786108017,
+ "learning_rate": 0.0006,
+ "loss": 6.509925365447998,
+ "step": 3230
+ },
+ {
+ "epoch": 44.880349344978164,
+ "grad_norm": 0.04749064892530441,
+ "learning_rate": 0.0006,
+ "loss": 6.478779315948486,
+ "step": 3231
+ },
+ {
+ "epoch": 44.8943231441048,
+ "grad_norm": 0.028002185747027397,
+ "learning_rate": 0.0006,
+ "loss": 6.528904914855957,
+ "step": 3232
+ },
+ {
+ "epoch": 44.90829694323144,
+ "grad_norm": 0.036473166197538376,
+ "learning_rate": 0.0006,
+ "loss": 6.549042701721191,
+ "step": 3233
+ },
+ {
+ "epoch": 44.92227074235808,
+ "grad_norm": 0.02543315477669239,
+ "learning_rate": 0.0006,
+ "loss": 6.498875141143799,
+ "step": 3234
+ },
+ {
+ "epoch": 44.93624454148472,
+ "grad_norm": 0.029906807467341423,
+ "learning_rate": 0.0006,
+ "loss": 6.453019142150879,
+ "step": 3235
+ },
+ {
+ "epoch": 44.95021834061135,
+ "grad_norm": 0.027231857180595398,
+ "learning_rate": 0.0006,
+ "loss": 6.428829193115234,
+ "step": 3236
+ },
+ {
+ "epoch": 44.96419213973799,
+ "grad_norm": 0.02506154775619507,
+ "learning_rate": 0.0006,
+ "loss": 6.420858383178711,
+ "step": 3237
+ },
+ {
+ "epoch": 44.97816593886463,
+ "grad_norm": 0.0269472673535347,
+ "learning_rate": 0.0006,
+ "loss": 6.484989643096924,
+ "step": 3238
+ },
+ {
+ "epoch": 44.992139737991266,
+ "grad_norm": 0.023410487920045853,
+ "learning_rate": 0.0006,
+ "loss": 6.454830169677734,
+ "step": 3239
+ },
+ {
+ "epoch": 45.0,
+ "grad_norm": 0.034175023436546326,
+ "learning_rate": 0.0006,
+ "loss": 6.430975437164307,
+ "step": 3240
+ },
+ {
+ "epoch": 45.0,
+ "eval_loss": 6.458896160125732,
+ "eval_runtime": 56.6934,
+ "eval_samples_per_second": 43.074,
+ "eval_steps_per_second": 1.358,
+ "step": 3240
+ },
+ {
+ "epoch": 45.01397379912664,
+ "grad_norm": 0.045031629502773285,
+ "learning_rate": 0.0006,
+ "loss": 6.4243011474609375,
+ "step": 3241
+ },
+ {
+ "epoch": 45.02794759825328,
+ "grad_norm": 0.04968734085559845,
+ "learning_rate": 0.0006,
+ "loss": 6.380091667175293,
+ "step": 3242
+ },
+ {
+ "epoch": 45.041921397379916,
+ "grad_norm": 0.047145187854766846,
+ "learning_rate": 0.0006,
+ "loss": 6.477178573608398,
+ "step": 3243
+ },
+ {
+ "epoch": 45.05589519650655,
+ "grad_norm": 0.04304948449134827,
+ "learning_rate": 0.0006,
+ "loss": 6.399726867675781,
+ "step": 3244
+ },
+ {
+ "epoch": 45.069868995633186,
+ "grad_norm": 0.04466010630130768,
+ "learning_rate": 0.0006,
+ "loss": 6.3415679931640625,
+ "step": 3245
+ },
+ {
+ "epoch": 45.083842794759825,
+ "grad_norm": 0.04305306077003479,
+ "learning_rate": 0.0006,
+ "loss": 6.3413310050964355,
+ "step": 3246
+ },
+ {
+ "epoch": 45.09781659388646,
+ "grad_norm": 0.043064575642347336,
+ "learning_rate": 0.0006,
+ "loss": 6.4838714599609375,
+ "step": 3247
+ },
+ {
+ "epoch": 45.1117903930131,
+ "grad_norm": 0.049503277987241745,
+ "learning_rate": 0.0006,
+ "loss": 6.38473653793335,
+ "step": 3248
+ },
+ {
+ "epoch": 45.12576419213974,
+ "grad_norm": 0.04613238573074341,
+ "learning_rate": 0.0006,
+ "loss": 6.370769500732422,
+ "step": 3249
+ },
+ {
+ "epoch": 45.13973799126637,
+ "grad_norm": 0.052119478583335876,
+ "learning_rate": 0.0006,
+ "loss": 6.362321853637695,
+ "step": 3250
+ },
+ {
+ "epoch": 45.15371179039301,
+ "grad_norm": 0.045110076665878296,
+ "learning_rate": 0.0006,
+ "loss": 6.369844913482666,
+ "step": 3251
+ },
+ {
+ "epoch": 45.16768558951965,
+ "grad_norm": 0.03798583149909973,
+ "learning_rate": 0.0006,
+ "loss": 6.324633598327637,
+ "step": 3252
+ },
+ {
+ "epoch": 45.18165938864629,
+ "grad_norm": 0.03744659945368767,
+ "learning_rate": 0.0006,
+ "loss": 6.349421501159668,
+ "step": 3253
+ },
+ {
+ "epoch": 45.19563318777293,
+ "grad_norm": 0.03767292574048042,
+ "learning_rate": 0.0006,
+ "loss": 6.334238052368164,
+ "step": 3254
+ },
+ {
+ "epoch": 45.209606986899566,
+ "grad_norm": 0.03577509894967079,
+ "learning_rate": 0.0006,
+ "loss": 6.3453826904296875,
+ "step": 3255
+ },
+ {
+ "epoch": 45.223580786026204,
+ "grad_norm": 0.03403240442276001,
+ "learning_rate": 0.0006,
+ "loss": 6.2836174964904785,
+ "step": 3256
+ },
+ {
+ "epoch": 45.237554585152836,
+ "grad_norm": 0.03517331928014755,
+ "learning_rate": 0.0006,
+ "loss": 6.281718730926514,
+ "step": 3257
+ },
+ {
+ "epoch": 45.251528384279474,
+ "grad_norm": 0.04409139230847359,
+ "learning_rate": 0.0006,
+ "loss": 6.320815086364746,
+ "step": 3258
+ },
+ {
+ "epoch": 45.26550218340611,
+ "grad_norm": 0.054750364273786545,
+ "learning_rate": 0.0006,
+ "loss": 6.343618392944336,
+ "step": 3259
+ },
+ {
+ "epoch": 45.27947598253275,
+ "grad_norm": 0.07612067461013794,
+ "learning_rate": 0.0006,
+ "loss": 6.310614585876465,
+ "step": 3260
+ },
+ {
+ "epoch": 45.29344978165939,
+ "grad_norm": 0.08438985049724579,
+ "learning_rate": 0.0006,
+ "loss": 6.2369184494018555,
+ "step": 3261
+ },
+ {
+ "epoch": 45.30742358078603,
+ "grad_norm": 0.0782201811671257,
+ "learning_rate": 0.0006,
+ "loss": 6.254452705383301,
+ "step": 3262
+ },
+ {
+ "epoch": 45.32139737991266,
+ "grad_norm": 0.09928355365991592,
+ "learning_rate": 0.0006,
+ "loss": 6.3112077713012695,
+ "step": 3263
+ },
+ {
+ "epoch": 45.3353711790393,
+ "grad_norm": 0.07233133912086487,
+ "learning_rate": 0.0006,
+ "loss": 6.408087253570557,
+ "step": 3264
+ },
+ {
+ "epoch": 45.34934497816594,
+ "grad_norm": 0.07478684186935425,
+ "learning_rate": 0.0006,
+ "loss": 6.3770036697387695,
+ "step": 3265
+ },
+ {
+ "epoch": 45.36331877729258,
+ "grad_norm": 0.06036846712231636,
+ "learning_rate": 0.0006,
+ "loss": 6.33865213394165,
+ "step": 3266
+ },
+ {
+ "epoch": 45.377292576419215,
+ "grad_norm": 0.05783458054065704,
+ "learning_rate": 0.0006,
+ "loss": 6.363203048706055,
+ "step": 3267
+ },
+ {
+ "epoch": 45.391266375545854,
+ "grad_norm": 0.05443909019231796,
+ "learning_rate": 0.0006,
+ "loss": 6.287069797515869,
+ "step": 3268
+ },
+ {
+ "epoch": 45.40524017467249,
+ "grad_norm": 0.0557909794151783,
+ "learning_rate": 0.0006,
+ "loss": 6.297784805297852,
+ "step": 3269
+ },
+ {
+ "epoch": 45.419213973799124,
+ "grad_norm": 0.03483372926712036,
+ "learning_rate": 0.0006,
+ "loss": 6.280734062194824,
+ "step": 3270
+ },
+ {
+ "epoch": 45.43318777292576,
+ "grad_norm": 0.0425543375313282,
+ "learning_rate": 0.0006,
+ "loss": 6.33565616607666,
+ "step": 3271
+ },
+ {
+ "epoch": 45.4471615720524,
+ "grad_norm": 0.03230159357190132,
+ "learning_rate": 0.0006,
+ "loss": 6.197698593139648,
+ "step": 3272
+ },
+ {
+ "epoch": 45.46113537117904,
+ "grad_norm": 0.036897290498018265,
+ "learning_rate": 0.0006,
+ "loss": 6.241239547729492,
+ "step": 3273
+ },
+ {
+ "epoch": 45.47510917030568,
+ "grad_norm": 0.031103838235139847,
+ "learning_rate": 0.0006,
+ "loss": 6.159435749053955,
+ "step": 3274
+ },
+ {
+ "epoch": 45.48908296943232,
+ "grad_norm": 0.027116449549794197,
+ "learning_rate": 0.0006,
+ "loss": 6.212225914001465,
+ "step": 3275
+ },
+ {
+ "epoch": 45.50305676855895,
+ "grad_norm": 0.029923981055617332,
+ "learning_rate": 0.0006,
+ "loss": 6.157269477844238,
+ "step": 3276
+ },
+ {
+ "epoch": 45.51703056768559,
+ "grad_norm": 0.02521538734436035,
+ "learning_rate": 0.0006,
+ "loss": 6.146918773651123,
+ "step": 3277
+ },
+ {
+ "epoch": 45.531004366812226,
+ "grad_norm": 0.02758920192718506,
+ "learning_rate": 0.0006,
+ "loss": 6.0446977615356445,
+ "step": 3278
+ },
+ {
+ "epoch": 45.544978165938865,
+ "grad_norm": 0.024328220635652542,
+ "learning_rate": 0.0006,
+ "loss": 6.099508285522461,
+ "step": 3279
+ },
+ {
+ "epoch": 45.5589519650655,
+ "grad_norm": 0.027425771579146385,
+ "learning_rate": 0.0006,
+ "loss": 6.084299087524414,
+ "step": 3280
+ },
+ {
+ "epoch": 45.57292576419214,
+ "grad_norm": 0.026927651837468147,
+ "learning_rate": 0.0006,
+ "loss": 6.146334171295166,
+ "step": 3281
+ },
+ {
+ "epoch": 45.58689956331878,
+ "grad_norm": 0.04054379090666771,
+ "learning_rate": 0.0006,
+ "loss": 6.1301493644714355,
+ "step": 3282
+ },
+ {
+ "epoch": 45.60087336244541,
+ "grad_norm": 0.04456290975213051,
+ "learning_rate": 0.0006,
+ "loss": 6.152511119842529,
+ "step": 3283
+ },
+ {
+ "epoch": 45.61484716157205,
+ "grad_norm": 0.052821505814790726,
+ "learning_rate": 0.0006,
+ "loss": 6.113544940948486,
+ "step": 3284
+ },
+ {
+ "epoch": 45.62882096069869,
+ "grad_norm": 0.07251739501953125,
+ "learning_rate": 0.0006,
+ "loss": 6.118298530578613,
+ "step": 3285
+ },
+ {
+ "epoch": 45.64279475982533,
+ "grad_norm": 0.09433556348085403,
+ "learning_rate": 0.0006,
+ "loss": 6.170897483825684,
+ "step": 3286
+ },
+ {
+ "epoch": 45.65676855895197,
+ "grad_norm": 0.057683493942022324,
+ "learning_rate": 0.0006,
+ "loss": 6.1747822761535645,
+ "step": 3287
+ },
+ {
+ "epoch": 45.670742358078606,
+ "grad_norm": 0.06363270431756973,
+ "learning_rate": 0.0006,
+ "loss": 6.1617889404296875,
+ "step": 3288
+ },
+ {
+ "epoch": 45.68471615720524,
+ "grad_norm": 0.07528169453144073,
+ "learning_rate": 0.0006,
+ "loss": 6.230060577392578,
+ "step": 3289
+ },
+ {
+ "epoch": 45.698689956331876,
+ "grad_norm": 0.07816170156002045,
+ "learning_rate": 0.0006,
+ "loss": 6.171257019042969,
+ "step": 3290
+ },
+ {
+ "epoch": 45.712663755458514,
+ "grad_norm": 0.05860164389014244,
+ "learning_rate": 0.0006,
+ "loss": 6.095484256744385,
+ "step": 3291
+ },
+ {
+ "epoch": 45.72663755458515,
+ "grad_norm": 0.04747429117560387,
+ "learning_rate": 0.0006,
+ "loss": 6.060067176818848,
+ "step": 3292
+ },
+ {
+ "epoch": 45.74061135371179,
+ "grad_norm": 0.04514474794268608,
+ "learning_rate": 0.0006,
+ "loss": 6.100218772888184,
+ "step": 3293
+ },
+ {
+ "epoch": 45.75458515283843,
+ "grad_norm": 0.05168681964278221,
+ "learning_rate": 0.0006,
+ "loss": 6.0621867179870605,
+ "step": 3294
+ },
+ {
+ "epoch": 45.76855895196506,
+ "grad_norm": 0.04665876924991608,
+ "learning_rate": 0.0006,
+ "loss": 6.118292808532715,
+ "step": 3295
+ },
+ {
+ "epoch": 45.7825327510917,
+ "grad_norm": 0.049374762922525406,
+ "learning_rate": 0.0006,
+ "loss": 6.098394393920898,
+ "step": 3296
+ },
+ {
+ "epoch": 45.79650655021834,
+ "grad_norm": 0.03793549910187721,
+ "learning_rate": 0.0006,
+ "loss": 6.069541931152344,
+ "step": 3297
+ },
+ {
+ "epoch": 45.81048034934498,
+ "grad_norm": 0.036362383514642715,
+ "learning_rate": 0.0006,
+ "loss": 6.1188435554504395,
+ "step": 3298
+ },
+ {
+ "epoch": 45.82445414847162,
+ "grad_norm": 0.03397097811102867,
+ "learning_rate": 0.0006,
+ "loss": 6.117589950561523,
+ "step": 3299
+ },
+ {
+ "epoch": 45.838427947598255,
+ "grad_norm": 0.030977753922343254,
+ "learning_rate": 0.0006,
+ "loss": 6.0236711502075195,
+ "step": 3300
+ },
+ {
+ "epoch": 45.852401746724894,
+ "grad_norm": 0.02800217643380165,
+ "learning_rate": 0.0006,
+ "loss": 6.034825325012207,
+ "step": 3301
+ },
+ {
+ "epoch": 45.866375545851525,
+ "grad_norm": 0.03129778802394867,
+ "learning_rate": 0.0006,
+ "loss": 5.980271339416504,
+ "step": 3302
+ },
+ {
+ "epoch": 45.880349344978164,
+ "grad_norm": 0.025886094197630882,
+ "learning_rate": 0.0006,
+ "loss": 5.966419219970703,
+ "step": 3303
+ },
+ {
+ "epoch": 45.8943231441048,
+ "grad_norm": 0.025712328031659126,
+ "learning_rate": 0.0006,
+ "loss": 5.938652515411377,
+ "step": 3304
+ },
+ {
+ "epoch": 45.90829694323144,
+ "grad_norm": 0.022508298978209496,
+ "learning_rate": 0.0006,
+ "loss": 5.982312202453613,
+ "step": 3305
+ },
+ {
+ "epoch": 45.92227074235808,
+ "grad_norm": 0.024880951270461082,
+ "learning_rate": 0.0006,
+ "loss": 5.931150436401367,
+ "step": 3306
+ },
+ {
+ "epoch": 45.93624454148472,
+ "grad_norm": 0.030007576569914818,
+ "learning_rate": 0.0006,
+ "loss": 5.936639308929443,
+ "step": 3307
+ },
+ {
+ "epoch": 45.95021834061135,
+ "grad_norm": 0.033314041793346405,
+ "learning_rate": 0.0006,
+ "loss": 5.944248676300049,
+ "step": 3308
+ },
+ {
+ "epoch": 45.96419213973799,
+ "grad_norm": 0.043078381568193436,
+ "learning_rate": 0.0006,
+ "loss": 5.971286773681641,
+ "step": 3309
+ },
+ {
+ "epoch": 45.97816593886463,
+ "grad_norm": 0.059296440333127975,
+ "learning_rate": 0.0006,
+ "loss": 5.910497665405273,
+ "step": 3310
+ },
+ {
+ "epoch": 45.992139737991266,
+ "grad_norm": 0.06232666224241257,
+ "learning_rate": 0.0006,
+ "loss": 5.929817199707031,
+ "step": 3311
+ },
+ {
+ "epoch": 46.0,
+ "grad_norm": 0.05303940549492836,
+ "learning_rate": 0.0006,
+ "loss": 5.910822868347168,
+ "step": 3312
+ },
+ {
+ "epoch": 46.0,
+ "eval_loss": 5.977851390838623,
+ "eval_runtime": 56.2435,
+ "eval_samples_per_second": 43.418,
+ "eval_steps_per_second": 1.369,
+ "step": 3312
+ },
+ {
+ "epoch": 46.01397379912664,
+ "grad_norm": 0.06943798810243607,
+ "learning_rate": 0.0006,
+ "loss": 5.854922294616699,
+ "step": 3313
+ },
+ {
+ "epoch": 46.02794759825328,
+ "grad_norm": 0.05236278846859932,
+ "learning_rate": 0.0006,
+ "loss": 5.876611232757568,
+ "step": 3314
+ },
+ {
+ "epoch": 46.041921397379916,
+ "grad_norm": 0.06240001320838928,
+ "learning_rate": 0.0006,
+ "loss": 5.878912448883057,
+ "step": 3315
+ },
+ {
+ "epoch": 46.05589519650655,
+ "grad_norm": 0.07401192933320999,
+ "learning_rate": 0.0006,
+ "loss": 5.985136032104492,
+ "step": 3316
+ },
+ {
+ "epoch": 46.069868995633186,
+ "grad_norm": 0.0826529711484909,
+ "learning_rate": 0.0006,
+ "loss": 5.939439296722412,
+ "step": 3317
+ },
+ {
+ "epoch": 46.083842794759825,
+ "grad_norm": 0.07999005168676376,
+ "learning_rate": 0.0006,
+ "loss": 5.982336044311523,
+ "step": 3318
+ },
+ {
+ "epoch": 46.09781659388646,
+ "grad_norm": 0.07014869153499603,
+ "learning_rate": 0.0006,
+ "loss": 5.896534442901611,
+ "step": 3319
+ },
+ {
+ "epoch": 46.1117903930131,
+ "grad_norm": 0.05795128643512726,
+ "learning_rate": 0.0006,
+ "loss": 5.843111991882324,
+ "step": 3320
+ },
+ {
+ "epoch": 46.12576419213974,
+ "grad_norm": 0.06374258548021317,
+ "learning_rate": 0.0006,
+ "loss": 5.78752326965332,
+ "step": 3321
+ },
+ {
+ "epoch": 46.13973799126637,
+ "grad_norm": 0.045524805784225464,
+ "learning_rate": 0.0006,
+ "loss": 5.86755895614624,
+ "step": 3322
+ },
+ {
+ "epoch": 46.15371179039301,
+ "grad_norm": 0.03773504123091698,
+ "learning_rate": 0.0006,
+ "loss": 5.83627462387085,
+ "step": 3323
+ },
+ {
+ "epoch": 46.16768558951965,
+ "grad_norm": 0.035815201699733734,
+ "learning_rate": 0.0006,
+ "loss": 5.798238277435303,
+ "step": 3324
+ },
+ {
+ "epoch": 46.18165938864629,
+ "grad_norm": 0.03756783902645111,
+ "learning_rate": 0.0006,
+ "loss": 5.845197677612305,
+ "step": 3325
+ },
+ {
+ "epoch": 46.19563318777293,
+ "grad_norm": 0.03569113463163376,
+ "learning_rate": 0.0006,
+ "loss": 5.888932704925537,
+ "step": 3326
+ },
+ {
+ "epoch": 46.209606986899566,
+ "grad_norm": 0.029467813670635223,
+ "learning_rate": 0.0006,
+ "loss": 5.792047500610352,
+ "step": 3327
+ },
+ {
+ "epoch": 46.223580786026204,
+ "grad_norm": 0.025698406621813774,
+ "learning_rate": 0.0006,
+ "loss": 5.783734321594238,
+ "step": 3328
+ },
+ {
+ "epoch": 46.237554585152836,
+ "grad_norm": 0.028426671400666237,
+ "learning_rate": 0.0006,
+ "loss": 5.762084007263184,
+ "step": 3329
+ },
+ {
+ "epoch": 46.251528384279474,
+ "grad_norm": 0.03099622018635273,
+ "learning_rate": 0.0006,
+ "loss": 5.719109058380127,
+ "step": 3330
+ },
+ {
+ "epoch": 46.26550218340611,
+ "grad_norm": 0.0310219656676054,
+ "learning_rate": 0.0006,
+ "loss": 5.737057209014893,
+ "step": 3331
+ },
+ {
+ "epoch": 46.27947598253275,
+ "grad_norm": 0.02713446132838726,
+ "learning_rate": 0.0006,
+ "loss": 5.758657455444336,
+ "step": 3332
+ },
+ {
+ "epoch": 46.29344978165939,
+ "grad_norm": 0.03099844604730606,
+ "learning_rate": 0.0006,
+ "loss": 5.742086410522461,
+ "step": 3333
+ },
+ {
+ "epoch": 46.30742358078603,
+ "grad_norm": 0.030258730053901672,
+ "learning_rate": 0.0006,
+ "loss": 5.767426490783691,
+ "step": 3334
+ },
+ {
+ "epoch": 46.32139737991266,
+ "grad_norm": 0.029034852981567383,
+ "learning_rate": 0.0006,
+ "loss": 5.651792049407959,
+ "step": 3335
+ },
+ {
+ "epoch": 46.3353711790393,
+ "grad_norm": 0.03138665854930878,
+ "learning_rate": 0.0006,
+ "loss": 5.664980888366699,
+ "step": 3336
+ },
+ {
+ "epoch": 46.34934497816594,
+ "grad_norm": 0.040546901524066925,
+ "learning_rate": 0.0006,
+ "loss": 5.665165901184082,
+ "step": 3337
+ },
+ {
+ "epoch": 46.36331877729258,
+ "grad_norm": 0.06395485252141953,
+ "learning_rate": 0.0006,
+ "loss": 5.7776713371276855,
+ "step": 3338
+ },
+ {
+ "epoch": 46.377292576419215,
+ "grad_norm": 0.10088056325912476,
+ "learning_rate": 0.0006,
+ "loss": 5.755645751953125,
+ "step": 3339
+ },
+ {
+ "epoch": 46.391266375545854,
+ "grad_norm": 0.09527938067913055,
+ "learning_rate": 0.0006,
+ "loss": 5.838809490203857,
+ "step": 3340
+ },
+ {
+ "epoch": 46.40524017467249,
+ "grad_norm": 0.0849648267030716,
+ "learning_rate": 0.0006,
+ "loss": 5.781932830810547,
+ "step": 3341
+ },
+ {
+ "epoch": 46.419213973799124,
+ "grad_norm": 0.10142666101455688,
+ "learning_rate": 0.0006,
+ "loss": 5.838970184326172,
+ "step": 3342
+ },
+ {
+ "epoch": 46.43318777292576,
+ "grad_norm": 0.42731454968452454,
+ "learning_rate": 0.0006,
+ "loss": 6.071902275085449,
+ "step": 3343
+ },
+ {
+ "epoch": 46.4471615720524,
+ "grad_norm": 0.13102488219738007,
+ "learning_rate": 0.0006,
+ "loss": 6.8635993003845215,
+ "step": 3344
+ },
+ {
+ "epoch": 46.46113537117904,
+ "grad_norm": 0.23691938817501068,
+ "learning_rate": 0.0006,
+ "loss": 7.313158988952637,
+ "step": 3345
+ },
+ {
+ "epoch": 46.47510917030568,
+ "grad_norm": 0.11670669168233871,
+ "learning_rate": 0.0006,
+ "loss": 7.240363121032715,
+ "step": 3346
+ },
+ {
+ "epoch": 46.48908296943232,
+ "grad_norm": 0.09985820949077606,
+ "learning_rate": 0.0006,
+ "loss": 7.113332271575928,
+ "step": 3347
+ },
+ {
+ "epoch": 46.50305676855895,
+ "grad_norm": 0.11406408250331879,
+ "learning_rate": 0.0006,
+ "loss": 7.051285743713379,
+ "step": 3348
+ },
+ {
+ "epoch": 46.51703056768559,
+ "grad_norm": 0.07900336384773254,
+ "learning_rate": 0.0006,
+ "loss": 7.014365196228027,
+ "step": 3349
+ },
+ {
+ "epoch": 46.531004366812226,
+ "grad_norm": 0.05380333960056305,
+ "learning_rate": 0.0006,
+ "loss": 6.8868818283081055,
+ "step": 3350
+ },
+ {
+ "epoch": 46.544978165938865,
+ "grad_norm": 0.05868449807167053,
+ "learning_rate": 0.0006,
+ "loss": 6.783599853515625,
+ "step": 3351
+ },
+ {
+ "epoch": 46.5589519650655,
+ "grad_norm": 0.04428785294294357,
+ "learning_rate": 0.0006,
+ "loss": 6.798511505126953,
+ "step": 3352
+ },
+ {
+ "epoch": 46.57292576419214,
+ "grad_norm": 0.055527277290821075,
+ "learning_rate": 0.0006,
+ "loss": 6.694507598876953,
+ "step": 3353
+ },
+ {
+ "epoch": 46.58689956331878,
+ "grad_norm": 0.03872432932257652,
+ "learning_rate": 0.0006,
+ "loss": 6.736631393432617,
+ "step": 3354
+ },
+ {
+ "epoch": 46.60087336244541,
+ "grad_norm": 0.04330779239535332,
+ "learning_rate": 0.0006,
+ "loss": 6.576806545257568,
+ "step": 3355
+ },
+ {
+ "epoch": 46.61484716157205,
+ "grad_norm": 0.036767736077308655,
+ "learning_rate": 0.0006,
+ "loss": 6.559871196746826,
+ "step": 3356
+ },
+ {
+ "epoch": 46.62882096069869,
+ "grad_norm": 0.039857108145952225,
+ "learning_rate": 0.0006,
+ "loss": 6.5749831199646,
+ "step": 3357
+ },
+ {
+ "epoch": 46.64279475982533,
+ "grad_norm": 0.048719972372055054,
+ "learning_rate": 0.0006,
+ "loss": 6.4915618896484375,
+ "step": 3358
+ },
+ {
+ "epoch": 46.65676855895197,
+ "grad_norm": 0.042849015444517136,
+ "learning_rate": 0.0006,
+ "loss": 6.413516044616699,
+ "step": 3359
+ },
+ {
+ "epoch": 46.670742358078606,
+ "grad_norm": 0.02871009334921837,
+ "learning_rate": 0.0006,
+ "loss": 6.320952415466309,
+ "step": 3360
+ },
+ {
+ "epoch": 46.68471615720524,
+ "grad_norm": 0.03193620592355728,
+ "learning_rate": 0.0006,
+ "loss": 6.4176435470581055,
+ "step": 3361
+ },
+ {
+ "epoch": 46.698689956331876,
+ "grad_norm": 0.027417024597525597,
+ "learning_rate": 0.0006,
+ "loss": 6.342319488525391,
+ "step": 3362
+ },
+ {
+ "epoch": 46.712663755458514,
+ "grad_norm": 0.03305937349796295,
+ "learning_rate": 0.0006,
+ "loss": 6.246112823486328,
+ "step": 3363
+ },
+ {
+ "epoch": 46.72663755458515,
+ "grad_norm": 0.031403183937072754,
+ "learning_rate": 0.0006,
+ "loss": 6.149503231048584,
+ "step": 3364
+ },
+ {
+ "epoch": 46.74061135371179,
+ "grad_norm": 0.031563788652420044,
+ "learning_rate": 0.0006,
+ "loss": 6.278808116912842,
+ "step": 3365
+ },
+ {
+ "epoch": 46.75458515283843,
+ "grad_norm": 0.028366589918732643,
+ "learning_rate": 0.0006,
+ "loss": 6.057036876678467,
+ "step": 3366
+ },
+ {
+ "epoch": 46.76855895196506,
+ "grad_norm": 0.023715803399682045,
+ "learning_rate": 0.0006,
+ "loss": 6.111324310302734,
+ "step": 3367
+ },
+ {
+ "epoch": 46.7825327510917,
+ "grad_norm": 0.026546811684966087,
+ "learning_rate": 0.0006,
+ "loss": 6.105669975280762,
+ "step": 3368
+ },
+ {
+ "epoch": 46.79650655021834,
+ "grad_norm": 0.023531019687652588,
+ "learning_rate": 0.0006,
+ "loss": 6.068072319030762,
+ "step": 3369
+ },
+ {
+ "epoch": 46.81048034934498,
+ "grad_norm": 0.024480128660798073,
+ "learning_rate": 0.0006,
+ "loss": 5.947896957397461,
+ "step": 3370
+ },
+ {
+ "epoch": 46.82445414847162,
+ "grad_norm": 0.02517310529947281,
+ "learning_rate": 0.0006,
+ "loss": 6.030037879943848,
+ "step": 3371
+ },
+ {
+ "epoch": 46.838427947598255,
+ "grad_norm": 0.026723403483629227,
+ "learning_rate": 0.0006,
+ "loss": 5.963776111602783,
+ "step": 3372
+ },
+ {
+ "epoch": 46.852401746724894,
+ "grad_norm": 0.03163504600524902,
+ "learning_rate": 0.0006,
+ "loss": 5.893466949462891,
+ "step": 3373
+ },
+ {
+ "epoch": 46.866375545851525,
+ "grad_norm": 0.04058683291077614,
+ "learning_rate": 0.0006,
+ "loss": 5.9202799797058105,
+ "step": 3374
+ },
+ {
+ "epoch": 46.880349344978164,
+ "grad_norm": 0.04964117333292961,
+ "learning_rate": 0.0006,
+ "loss": 5.9366559982299805,
+ "step": 3375
+ },
+ {
+ "epoch": 46.8943231441048,
+ "grad_norm": 0.055406033992767334,
+ "learning_rate": 0.0006,
+ "loss": 5.912230968475342,
+ "step": 3376
+ },
+ {
+ "epoch": 46.90829694323144,
+ "grad_norm": 0.04643625393509865,
+ "learning_rate": 0.0006,
+ "loss": 5.87260103225708,
+ "step": 3377
+ },
+ {
+ "epoch": 46.92227074235808,
+ "grad_norm": 0.029325876384973526,
+ "learning_rate": 0.0006,
+ "loss": 5.854763984680176,
+ "step": 3378
+ },
+ {
+ "epoch": 46.93624454148472,
+ "grad_norm": 0.03159382566809654,
+ "learning_rate": 0.0006,
+ "loss": 5.880882263183594,
+ "step": 3379
+ },
+ {
+ "epoch": 46.95021834061135,
+ "grad_norm": 0.030224012210965157,
+ "learning_rate": 0.0006,
+ "loss": 5.787360191345215,
+ "step": 3380
+ },
+ {
+ "epoch": 46.96419213973799,
+ "grad_norm": 0.028368130326271057,
+ "learning_rate": 0.0006,
+ "loss": 5.817195892333984,
+ "step": 3381
+ },
+ {
+ "epoch": 46.97816593886463,
+ "grad_norm": 0.029411399737000465,
+ "learning_rate": 0.0006,
+ "loss": 5.783403396606445,
+ "step": 3382
+ },
+ {
+ "epoch": 46.992139737991266,
+ "grad_norm": 0.029927462339401245,
+ "learning_rate": 0.0006,
+ "loss": 5.662444591522217,
+ "step": 3383
+ },
+ {
+ "epoch": 47.0,
+ "grad_norm": 0.02872961387038231,
+ "learning_rate": 0.0006,
+ "loss": 5.742781639099121,
+ "step": 3384
+ },
+ {
+ "epoch": 47.0,
+ "eval_loss": 5.870260238647461,
+ "eval_runtime": 57.3734,
+ "eval_samples_per_second": 42.563,
+ "eval_steps_per_second": 1.342,
+ "step": 3384
+ },
+ {
+ "epoch": 47.01397379912664,
+ "grad_norm": 0.028542131185531616,
+ "learning_rate": 0.0006,
+ "loss": 5.753256797790527,
+ "step": 3385
+ },
+ {
+ "epoch": 47.02794759825328,
+ "grad_norm": 0.03227119892835617,
+ "learning_rate": 0.0006,
+ "loss": 5.722246170043945,
+ "step": 3386
+ },
+ {
+ "epoch": 47.041921397379916,
+ "grad_norm": 0.031461723148822784,
+ "learning_rate": 0.0006,
+ "loss": 5.740520000457764,
+ "step": 3387
+ },
+ {
+ "epoch": 47.05589519650655,
+ "grad_norm": 0.03184220567345619,
+ "learning_rate": 0.0006,
+ "loss": 5.720890045166016,
+ "step": 3388
+ },
+ {
+ "epoch": 47.069868995633186,
+ "grad_norm": 0.030015580356121063,
+ "learning_rate": 0.0006,
+ "loss": 5.597814559936523,
+ "step": 3389
+ },
+ {
+ "epoch": 47.083842794759825,
+ "grad_norm": 0.029931779950857162,
+ "learning_rate": 0.0006,
+ "loss": 5.5412468910217285,
+ "step": 3390
+ },
+ {
+ "epoch": 47.09781659388646,
+ "grad_norm": 0.03206505626440048,
+ "learning_rate": 0.0006,
+ "loss": 5.6967597007751465,
+ "step": 3391
+ },
+ {
+ "epoch": 47.1117903930131,
+ "grad_norm": 0.03978161886334419,
+ "learning_rate": 0.0006,
+ "loss": 5.59661340713501,
+ "step": 3392
+ },
+ {
+ "epoch": 47.12576419213974,
+ "grad_norm": 0.04619080200791359,
+ "learning_rate": 0.0006,
+ "loss": 5.795788764953613,
+ "step": 3393
+ },
+ {
+ "epoch": 47.13973799126637,
+ "grad_norm": 0.049085382372140884,
+ "learning_rate": 0.0006,
+ "loss": 5.667335510253906,
+ "step": 3394
+ },
+ {
+ "epoch": 47.15371179039301,
+ "grad_norm": 0.04830858111381531,
+ "learning_rate": 0.0006,
+ "loss": 5.668806552886963,
+ "step": 3395
+ },
+ {
+ "epoch": 47.16768558951965,
+ "grad_norm": 0.0583178736269474,
+ "learning_rate": 0.0006,
+ "loss": 5.692285537719727,
+ "step": 3396
+ },
+ {
+ "epoch": 47.18165938864629,
+ "grad_norm": 0.05182253196835518,
+ "learning_rate": 0.0006,
+ "loss": 5.678645133972168,
+ "step": 3397
+ },
+ {
+ "epoch": 47.19563318777293,
+ "grad_norm": 0.04011905938386917,
+ "learning_rate": 0.0006,
+ "loss": 5.661550998687744,
+ "step": 3398
+ },
+ {
+ "epoch": 47.209606986899566,
+ "grad_norm": 0.03888486698269844,
+ "learning_rate": 0.0006,
+ "loss": 5.643001556396484,
+ "step": 3399
+ },
+ {
+ "epoch": 47.223580786026204,
+ "grad_norm": 0.030658265575766563,
+ "learning_rate": 0.0006,
+ "loss": 5.659090995788574,
+ "step": 3400
+ },
+ {
+ "epoch": 47.237554585152836,
+ "grad_norm": 0.03609256446361542,
+ "learning_rate": 0.0006,
+ "loss": 5.544501781463623,
+ "step": 3401
+ },
+ {
+ "epoch": 47.251528384279474,
+ "grad_norm": 0.033589769154787064,
+ "learning_rate": 0.0006,
+ "loss": 5.463878631591797,
+ "step": 3402
+ },
+ {
+ "epoch": 47.26550218340611,
+ "grad_norm": 0.035128287971019745,
+ "learning_rate": 0.0006,
+ "loss": 5.5680060386657715,
+ "step": 3403
+ },
+ {
+ "epoch": 47.27947598253275,
+ "grad_norm": 0.038684550672769547,
+ "learning_rate": 0.0006,
+ "loss": 5.587898254394531,
+ "step": 3404
+ },
+ {
+ "epoch": 47.29344978165939,
+ "grad_norm": 0.043566226959228516,
+ "learning_rate": 0.0006,
+ "loss": 5.580896854400635,
+ "step": 3405
+ },
+ {
+ "epoch": 47.30742358078603,
+ "grad_norm": 0.04428374394774437,
+ "learning_rate": 0.0006,
+ "loss": 5.603460311889648,
+ "step": 3406
+ },
+ {
+ "epoch": 47.32139737991266,
+ "grad_norm": 0.037368934601545334,
+ "learning_rate": 0.0006,
+ "loss": 5.564740180969238,
+ "step": 3407
+ },
+ {
+ "epoch": 47.3353711790393,
+ "grad_norm": 0.04092350974678993,
+ "learning_rate": 0.0006,
+ "loss": 5.584153652191162,
+ "step": 3408
+ },
+ {
+ "epoch": 47.34934497816594,
+ "grad_norm": 0.045031506568193436,
+ "learning_rate": 0.0006,
+ "loss": 5.601186275482178,
+ "step": 3409
+ },
+ {
+ "epoch": 47.36331877729258,
+ "grad_norm": 0.041830409318208694,
+ "learning_rate": 0.0006,
+ "loss": 5.512788772583008,
+ "step": 3410
+ },
+ {
+ "epoch": 47.377292576419215,
+ "grad_norm": 0.037633974105119705,
+ "learning_rate": 0.0006,
+ "loss": 5.503828525543213,
+ "step": 3411
+ },
+ {
+ "epoch": 47.391266375545854,
+ "grad_norm": 0.03473588079214096,
+ "learning_rate": 0.0006,
+ "loss": 5.570557594299316,
+ "step": 3412
+ },
+ {
+ "epoch": 47.40524017467249,
+ "grad_norm": 0.039325181394815445,
+ "learning_rate": 0.0006,
+ "loss": 5.427783966064453,
+ "step": 3413
+ },
+ {
+ "epoch": 47.419213973799124,
+ "grad_norm": 0.04034624621272087,
+ "learning_rate": 0.0006,
+ "loss": 5.572591304779053,
+ "step": 3414
+ },
+ {
+ "epoch": 47.43318777292576,
+ "grad_norm": 0.05245184153318405,
+ "learning_rate": 0.0006,
+ "loss": 5.505208969116211,
+ "step": 3415
+ },
+ {
+ "epoch": 47.4471615720524,
+ "grad_norm": 0.0568675734102726,
+ "learning_rate": 0.0006,
+ "loss": 5.589373588562012,
+ "step": 3416
+ },
+ {
+ "epoch": 47.46113537117904,
+ "grad_norm": 0.048336587846279144,
+ "learning_rate": 0.0006,
+ "loss": 5.522005081176758,
+ "step": 3417
+ },
+ {
+ "epoch": 47.47510917030568,
+ "grad_norm": 0.04194975271821022,
+ "learning_rate": 0.0006,
+ "loss": 5.387576103210449,
+ "step": 3418
+ },
+ {
+ "epoch": 47.48908296943232,
+ "grad_norm": 0.04019433259963989,
+ "learning_rate": 0.0006,
+ "loss": 5.522615432739258,
+ "step": 3419
+ },
+ {
+ "epoch": 47.50305676855895,
+ "grad_norm": 0.05709933117032051,
+ "learning_rate": 0.0006,
+ "loss": 5.420286178588867,
+ "step": 3420
+ },
+ {
+ "epoch": 47.51703056768559,
+ "grad_norm": 0.09130338579416275,
+ "learning_rate": 0.0006,
+ "loss": 5.484622478485107,
+ "step": 3421
+ },
+ {
+ "epoch": 47.531004366812226,
+ "grad_norm": 0.2757541835308075,
+ "learning_rate": 0.0006,
+ "loss": 5.821832656860352,
+ "step": 3422
+ },
+ {
+ "epoch": 47.544978165938865,
+ "grad_norm": 0.28292080760002136,
+ "learning_rate": 0.0006,
+ "loss": 6.579934120178223,
+ "step": 3423
+ },
+ {
+ "epoch": 47.5589519650655,
+ "grad_norm": 0.15293604135513306,
+ "learning_rate": 0.0006,
+ "loss": 6.478213310241699,
+ "step": 3424
+ },
+ {
+ "epoch": 47.57292576419214,
+ "grad_norm": 0.11284393817186356,
+ "learning_rate": 0.0006,
+ "loss": 6.4066386222839355,
+ "step": 3425
+ },
+ {
+ "epoch": 47.58689956331878,
+ "grad_norm": 0.10933073610067368,
+ "learning_rate": 0.0006,
+ "loss": 6.3306169509887695,
+ "step": 3426
+ },
+ {
+ "epoch": 47.60087336244541,
+ "grad_norm": 0.05837641656398773,
+ "learning_rate": 0.0006,
+ "loss": 6.250000953674316,
+ "step": 3427
+ },
+ {
+ "epoch": 47.61484716157205,
+ "grad_norm": 0.05987229570746422,
+ "learning_rate": 0.0006,
+ "loss": 6.152463436126709,
+ "step": 3428
+ },
+ {
+ "epoch": 47.62882096069869,
+ "grad_norm": 0.04064938426017761,
+ "learning_rate": 0.0006,
+ "loss": 6.117897033691406,
+ "step": 3429
+ },
+ {
+ "epoch": 47.64279475982533,
+ "grad_norm": 0.04565057530999184,
+ "learning_rate": 0.0006,
+ "loss": 6.042291641235352,
+ "step": 3430
+ },
+ {
+ "epoch": 47.65676855895197,
+ "grad_norm": 0.03475063294172287,
+ "learning_rate": 0.0006,
+ "loss": 6.0071868896484375,
+ "step": 3431
+ },
+ {
+ "epoch": 47.670742358078606,
+ "grad_norm": 0.03809220343828201,
+ "learning_rate": 0.0006,
+ "loss": 6.0128021240234375,
+ "step": 3432
+ },
+ {
+ "epoch": 47.68471615720524,
+ "grad_norm": 0.03307219594717026,
+ "learning_rate": 0.0006,
+ "loss": 5.894664287567139,
+ "step": 3433
+ },
+ {
+ "epoch": 47.698689956331876,
+ "grad_norm": 0.028824346140027046,
+ "learning_rate": 0.0006,
+ "loss": 5.914865493774414,
+ "step": 3434
+ },
+ {
+ "epoch": 47.712663755458514,
+ "grad_norm": 0.027923690155148506,
+ "learning_rate": 0.0006,
+ "loss": 5.851759910583496,
+ "step": 3435
+ },
+ {
+ "epoch": 47.72663755458515,
+ "grad_norm": 0.025492684915661812,
+ "learning_rate": 0.0006,
+ "loss": 5.749186038970947,
+ "step": 3436
+ },
+ {
+ "epoch": 47.74061135371179,
+ "grad_norm": 0.025105174630880356,
+ "learning_rate": 0.0006,
+ "loss": 5.803566932678223,
+ "step": 3437
+ },
+ {
+ "epoch": 47.75458515283843,
+ "grad_norm": 0.023298880085349083,
+ "learning_rate": 0.0006,
+ "loss": 5.76276159286499,
+ "step": 3438
+ },
+ {
+ "epoch": 47.76855895196506,
+ "grad_norm": 0.02468576282262802,
+ "learning_rate": 0.0006,
+ "loss": 5.833852767944336,
+ "step": 3439
+ },
+ {
+ "epoch": 47.7825327510917,
+ "grad_norm": 0.022730689495801926,
+ "learning_rate": 0.0006,
+ "loss": 5.736724376678467,
+ "step": 3440
+ },
+ {
+ "epoch": 47.79650655021834,
+ "grad_norm": 0.021335354074835777,
+ "learning_rate": 0.0006,
+ "loss": 5.644248008728027,
+ "step": 3441
+ },
+ {
+ "epoch": 47.81048034934498,
+ "grad_norm": 0.02288101427257061,
+ "learning_rate": 0.0006,
+ "loss": 5.7011942863464355,
+ "step": 3442
+ },
+ {
+ "epoch": 47.82445414847162,
+ "grad_norm": 0.01961585320532322,
+ "learning_rate": 0.0006,
+ "loss": 5.63215446472168,
+ "step": 3443
+ },
+ {
+ "epoch": 47.838427947598255,
+ "grad_norm": 0.017659205943346024,
+ "learning_rate": 0.0006,
+ "loss": 5.744845390319824,
+ "step": 3444
+ },
+ {
+ "epoch": 47.852401746724894,
+ "grad_norm": 0.02281784638762474,
+ "learning_rate": 0.0006,
+ "loss": 5.696993350982666,
+ "step": 3445
+ },
+ {
+ "epoch": 47.866375545851525,
+ "grad_norm": 0.016394242644309998,
+ "learning_rate": 0.0006,
+ "loss": 5.659870147705078,
+ "step": 3446
+ },
+ {
+ "epoch": 47.880349344978164,
+ "grad_norm": 0.01967368647456169,
+ "learning_rate": 0.0006,
+ "loss": 5.638072490692139,
+ "step": 3447
+ },
+ {
+ "epoch": 47.8943231441048,
+ "grad_norm": 0.017326267436146736,
+ "learning_rate": 0.0006,
+ "loss": 5.638836860656738,
+ "step": 3448
+ },
+ {
+ "epoch": 47.90829694323144,
+ "grad_norm": 0.017778463661670685,
+ "learning_rate": 0.0006,
+ "loss": 5.634466648101807,
+ "step": 3449
+ },
+ {
+ "epoch": 47.92227074235808,
+ "grad_norm": 0.017269758507609367,
+ "learning_rate": 0.0006,
+ "loss": 5.506277561187744,
+ "step": 3450
+ },
+ {
+ "epoch": 47.93624454148472,
+ "grad_norm": 0.01733289286494255,
+ "learning_rate": 0.0006,
+ "loss": 5.544455528259277,
+ "step": 3451
+ },
+ {
+ "epoch": 47.95021834061135,
+ "grad_norm": 0.017787795513868332,
+ "learning_rate": 0.0006,
+ "loss": 5.498943328857422,
+ "step": 3452
+ },
+ {
+ "epoch": 47.96419213973799,
+ "grad_norm": 0.015770932659506798,
+ "learning_rate": 0.0006,
+ "loss": 5.537227630615234,
+ "step": 3453
+ },
+ {
+ "epoch": 47.97816593886463,
+ "grad_norm": 0.016132809221744537,
+ "learning_rate": 0.0006,
+ "loss": 5.489081382751465,
+ "step": 3454
+ },
+ {
+ "epoch": 47.992139737991266,
+ "grad_norm": 0.01749664731323719,
+ "learning_rate": 0.0006,
+ "loss": 5.454263687133789,
+ "step": 3455
+ },
+ {
+ "epoch": 48.0,
+ "grad_norm": 0.016711296513676643,
+ "learning_rate": 0.0006,
+ "loss": 5.530783653259277,
+ "step": 3456
+ },
+ {
+ "epoch": 48.0,
+ "eval_loss": 5.596925258636475,
+ "eval_runtime": 56.9612,
+ "eval_samples_per_second": 42.871,
+ "eval_steps_per_second": 1.352,
+ "step": 3456
+ },
+ {
+ "epoch": 48.01397379912664,
+ "grad_norm": 0.01867641694843769,
+ "learning_rate": 0.0006,
+ "loss": 5.506853103637695,
+ "step": 3457
+ },
+ {
+ "epoch": 48.02794759825328,
+ "grad_norm": 0.023250792175531387,
+ "learning_rate": 0.0006,
+ "loss": 5.4074177742004395,
+ "step": 3458
+ },
+ {
+ "epoch": 48.041921397379916,
+ "grad_norm": 0.01970936730504036,
+ "learning_rate": 0.0006,
+ "loss": 5.508777618408203,
+ "step": 3459
+ },
+ {
+ "epoch": 48.05589519650655,
+ "grad_norm": 0.01879260502755642,
+ "learning_rate": 0.0006,
+ "loss": 5.412424087524414,
+ "step": 3460
+ },
+ {
+ "epoch": 48.069868995633186,
+ "grad_norm": 0.017420092597603798,
+ "learning_rate": 0.0006,
+ "loss": 5.43361759185791,
+ "step": 3461
+ },
+ {
+ "epoch": 48.083842794759825,
+ "grad_norm": 0.018707862123847008,
+ "learning_rate": 0.0006,
+ "loss": 5.476800918579102,
+ "step": 3462
+ },
+ {
+ "epoch": 48.09781659388646,
+ "grad_norm": 0.023185882717370987,
+ "learning_rate": 0.0006,
+ "loss": 5.465272426605225,
+ "step": 3463
+ },
+ {
+ "epoch": 48.1117903930131,
+ "grad_norm": 0.023570368066430092,
+ "learning_rate": 0.0006,
+ "loss": 5.391471862792969,
+ "step": 3464
+ },
+ {
+ "epoch": 48.12576419213974,
+ "grad_norm": 0.02651144750416279,
+ "learning_rate": 0.0006,
+ "loss": 5.422799110412598,
+ "step": 3465
+ },
+ {
+ "epoch": 48.13973799126637,
+ "grad_norm": 0.03316329047083855,
+ "learning_rate": 0.0006,
+ "loss": 5.3952202796936035,
+ "step": 3466
+ },
+ {
+ "epoch": 48.15371179039301,
+ "grad_norm": 0.04012278467416763,
+ "learning_rate": 0.0006,
+ "loss": 5.540853023529053,
+ "step": 3467
+ },
+ {
+ "epoch": 48.16768558951965,
+ "grad_norm": 0.04763203114271164,
+ "learning_rate": 0.0006,
+ "loss": 5.387890338897705,
+ "step": 3468
+ },
+ {
+ "epoch": 48.18165938864629,
+ "grad_norm": 0.05129165202379227,
+ "learning_rate": 0.0006,
+ "loss": 5.414722442626953,
+ "step": 3469
+ },
+ {
+ "epoch": 48.19563318777293,
+ "grad_norm": 0.04438071325421333,
+ "learning_rate": 0.0006,
+ "loss": 5.386244773864746,
+ "step": 3470
+ },
+ {
+ "epoch": 48.209606986899566,
+ "grad_norm": 0.040234919637441635,
+ "learning_rate": 0.0006,
+ "loss": 5.431617736816406,
+ "step": 3471
+ },
+ {
+ "epoch": 48.223580786026204,
+ "grad_norm": 0.03217008337378502,
+ "learning_rate": 0.0006,
+ "loss": 5.30168342590332,
+ "step": 3472
+ },
+ {
+ "epoch": 48.237554585152836,
+ "grad_norm": 0.03087974339723587,
+ "learning_rate": 0.0006,
+ "loss": 5.316389083862305,
+ "step": 3473
+ },
+ {
+ "epoch": 48.251528384279474,
+ "grad_norm": 0.02789347991347313,
+ "learning_rate": 0.0006,
+ "loss": 5.385970592498779,
+ "step": 3474
+ },
+ {
+ "epoch": 48.26550218340611,
+ "grad_norm": 0.024792097508907318,
+ "learning_rate": 0.0006,
+ "loss": 5.362329959869385,
+ "step": 3475
+ },
+ {
+ "epoch": 48.27947598253275,
+ "grad_norm": 0.025869522243738174,
+ "learning_rate": 0.0006,
+ "loss": 5.28400182723999,
+ "step": 3476
+ },
+ {
+ "epoch": 48.29344978165939,
+ "grad_norm": 0.021339882165193558,
+ "learning_rate": 0.0006,
+ "loss": 5.36669921875,
+ "step": 3477
+ },
+ {
+ "epoch": 48.30742358078603,
+ "grad_norm": 0.021462520584464073,
+ "learning_rate": 0.0006,
+ "loss": 5.311222076416016,
+ "step": 3478
+ },
+ {
+ "epoch": 48.32139737991266,
+ "grad_norm": 0.02193879708647728,
+ "learning_rate": 0.0006,
+ "loss": 5.291441917419434,
+ "step": 3479
+ },
+ {
+ "epoch": 48.3353711790393,
+ "grad_norm": 0.01963093690574169,
+ "learning_rate": 0.0006,
+ "loss": 5.342780113220215,
+ "step": 3480
+ },
+ {
+ "epoch": 48.34934497816594,
+ "grad_norm": 0.020634736865758896,
+ "learning_rate": 0.0006,
+ "loss": 5.144141674041748,
+ "step": 3481
+ },
+ {
+ "epoch": 48.36331877729258,
+ "grad_norm": 0.02126968279480934,
+ "learning_rate": 0.0006,
+ "loss": 5.232624530792236,
+ "step": 3482
+ },
+ {
+ "epoch": 48.377292576419215,
+ "grad_norm": 0.021203994750976562,
+ "learning_rate": 0.0006,
+ "loss": 5.290219783782959,
+ "step": 3483
+ },
+ {
+ "epoch": 48.391266375545854,
+ "grad_norm": 0.02193225920200348,
+ "learning_rate": 0.0006,
+ "loss": 5.29197359085083,
+ "step": 3484
+ },
+ {
+ "epoch": 48.40524017467249,
+ "grad_norm": 0.023174211382865906,
+ "learning_rate": 0.0006,
+ "loss": 5.204349517822266,
+ "step": 3485
+ },
+ {
+ "epoch": 48.419213973799124,
+ "grad_norm": 0.024161914363503456,
+ "learning_rate": 0.0006,
+ "loss": 5.339351654052734,
+ "step": 3486
+ },
+ {
+ "epoch": 48.43318777292576,
+ "grad_norm": 0.02221524342894554,
+ "learning_rate": 0.0006,
+ "loss": 5.263001441955566,
+ "step": 3487
+ },
+ {
+ "epoch": 48.4471615720524,
+ "grad_norm": 0.026632830500602722,
+ "learning_rate": 0.0006,
+ "loss": 5.221281051635742,
+ "step": 3488
+ },
+ {
+ "epoch": 48.46113537117904,
+ "grad_norm": 0.034923870116472244,
+ "learning_rate": 0.0006,
+ "loss": 5.296039581298828,
+ "step": 3489
+ },
+ {
+ "epoch": 48.47510917030568,
+ "grad_norm": 0.046184979379177094,
+ "learning_rate": 0.0006,
+ "loss": 5.212769985198975,
+ "step": 3490
+ },
+ {
+ "epoch": 48.48908296943232,
+ "grad_norm": 0.05154098570346832,
+ "learning_rate": 0.0006,
+ "loss": 5.337864875793457,
+ "step": 3491
+ },
+ {
+ "epoch": 48.50305676855895,
+ "grad_norm": 0.042967259883880615,
+ "learning_rate": 0.0006,
+ "loss": 5.267958641052246,
+ "step": 3492
+ },
+ {
+ "epoch": 48.51703056768559,
+ "grad_norm": 0.05356583371758461,
+ "learning_rate": 0.0006,
+ "loss": 5.1404218673706055,
+ "step": 3493
+ },
+ {
+ "epoch": 48.531004366812226,
+ "grad_norm": 0.08219032734632492,
+ "learning_rate": 0.0006,
+ "loss": 5.300687789916992,
+ "step": 3494
+ },
+ {
+ "epoch": 48.544978165938865,
+ "grad_norm": 0.19204089045524597,
+ "learning_rate": 0.0006,
+ "loss": 5.456025123596191,
+ "step": 3495
+ },
+ {
+ "epoch": 48.5589519650655,
+ "grad_norm": 0.25917696952819824,
+ "learning_rate": 0.0006,
+ "loss": 6.156761169433594,
+ "step": 3496
+ },
+ {
+ "epoch": 48.57292576419214,
+ "grad_norm": 0.1566283404827118,
+ "learning_rate": 0.0006,
+ "loss": 5.862346649169922,
+ "step": 3497
+ },
+ {
+ "epoch": 48.58689956331878,
+ "grad_norm": 0.1761486828327179,
+ "learning_rate": 0.0006,
+ "loss": 6.012795448303223,
+ "step": 3498
+ },
+ {
+ "epoch": 48.60087336244541,
+ "grad_norm": 0.09857621043920517,
+ "learning_rate": 0.0006,
+ "loss": 5.9248151779174805,
+ "step": 3499
+ },
+ {
+ "epoch": 48.61484716157205,
+ "grad_norm": 0.05968055501580238,
+ "learning_rate": 0.0006,
+ "loss": 5.780706882476807,
+ "step": 3500
+ },
+ {
+ "epoch": 48.62882096069869,
+ "grad_norm": 0.057433564215898514,
+ "learning_rate": 0.0006,
+ "loss": 5.645388603210449,
+ "step": 3501
+ },
+ {
+ "epoch": 48.64279475982533,
+ "grad_norm": 0.039659615606069565,
+ "learning_rate": 0.0006,
+ "loss": 5.695138931274414,
+ "step": 3502
+ },
+ {
+ "epoch": 48.65676855895197,
+ "grad_norm": 0.04019695147871971,
+ "learning_rate": 0.0006,
+ "loss": 5.672435283660889,
+ "step": 3503
+ },
+ {
+ "epoch": 48.670742358078606,
+ "grad_norm": 0.03033018670976162,
+ "learning_rate": 0.0006,
+ "loss": 5.682260513305664,
+ "step": 3504
+ },
+ {
+ "epoch": 48.68471615720524,
+ "grad_norm": 0.03215787187218666,
+ "learning_rate": 0.0006,
+ "loss": 5.604182243347168,
+ "step": 3505
+ },
+ {
+ "epoch": 48.698689956331876,
+ "grad_norm": 0.026865126565098763,
+ "learning_rate": 0.0006,
+ "loss": 5.64040470123291,
+ "step": 3506
+ },
+ {
+ "epoch": 48.712663755458514,
+ "grad_norm": 0.02831609919667244,
+ "learning_rate": 0.0006,
+ "loss": 5.552720069885254,
+ "step": 3507
+ },
+ {
+ "epoch": 48.72663755458515,
+ "grad_norm": 0.024226613342761993,
+ "learning_rate": 0.0006,
+ "loss": 5.556946277618408,
+ "step": 3508
+ },
+ {
+ "epoch": 48.74061135371179,
+ "grad_norm": 0.025393934920430183,
+ "learning_rate": 0.0006,
+ "loss": 5.476178169250488,
+ "step": 3509
+ },
+ {
+ "epoch": 48.75458515283843,
+ "grad_norm": 0.01984412968158722,
+ "learning_rate": 0.0006,
+ "loss": 5.471088409423828,
+ "step": 3510
+ },
+ {
+ "epoch": 48.76855895196506,
+ "grad_norm": 0.02035129815340042,
+ "learning_rate": 0.0006,
+ "loss": 5.438812732696533,
+ "step": 3511
+ },
+ {
+ "epoch": 48.7825327510917,
+ "grad_norm": 0.0208350270986557,
+ "learning_rate": 0.0006,
+ "loss": 5.381746292114258,
+ "step": 3512
+ },
+ {
+ "epoch": 48.79650655021834,
+ "grad_norm": 0.017783882096409798,
+ "learning_rate": 0.0006,
+ "loss": 5.351675510406494,
+ "step": 3513
+ },
+ {
+ "epoch": 48.81048034934498,
+ "grad_norm": 0.01671770215034485,
+ "learning_rate": 0.0006,
+ "loss": 5.329428672790527,
+ "step": 3514
+ },
+ {
+ "epoch": 48.82445414847162,
+ "grad_norm": 0.017891215160489082,
+ "learning_rate": 0.0006,
+ "loss": 5.212074279785156,
+ "step": 3515
+ },
+ {
+ "epoch": 48.838427947598255,
+ "grad_norm": 0.016847791150212288,
+ "learning_rate": 0.0006,
+ "loss": 5.299534797668457,
+ "step": 3516
+ },
+ {
+ "epoch": 48.852401746724894,
+ "grad_norm": 0.01593637838959694,
+ "learning_rate": 0.0006,
+ "loss": 5.30466890335083,
+ "step": 3517
+ },
+ {
+ "epoch": 48.866375545851525,
+ "grad_norm": 0.014713102020323277,
+ "learning_rate": 0.0006,
+ "loss": 5.294345378875732,
+ "step": 3518
+ },
+ {
+ "epoch": 48.880349344978164,
+ "grad_norm": 0.01433857623487711,
+ "learning_rate": 0.0006,
+ "loss": 5.2148542404174805,
+ "step": 3519
+ },
+ {
+ "epoch": 48.8943231441048,
+ "grad_norm": 0.014979379251599312,
+ "learning_rate": 0.0006,
+ "loss": 5.298582077026367,
+ "step": 3520
+ },
+ {
+ "epoch": 48.90829694323144,
+ "grad_norm": 0.013494192622601986,
+ "learning_rate": 0.0006,
+ "loss": 5.309783935546875,
+ "step": 3521
+ },
+ {
+ "epoch": 48.92227074235808,
+ "grad_norm": 0.014020012691617012,
+ "learning_rate": 0.0006,
+ "loss": 5.284879684448242,
+ "step": 3522
+ },
+ {
+ "epoch": 48.93624454148472,
+ "grad_norm": 0.01348058134317398,
+ "learning_rate": 0.0006,
+ "loss": 5.249116897583008,
+ "step": 3523
+ },
+ {
+ "epoch": 48.95021834061135,
+ "grad_norm": 0.013280218467116356,
+ "learning_rate": 0.0006,
+ "loss": 5.273489952087402,
+ "step": 3524
+ },
+ {
+ "epoch": 48.96419213973799,
+ "grad_norm": 0.012647110037505627,
+ "learning_rate": 0.0006,
+ "loss": 5.306652069091797,
+ "step": 3525
+ },
+ {
+ "epoch": 48.97816593886463,
+ "grad_norm": 0.012847724370658398,
+ "learning_rate": 0.0006,
+ "loss": 5.235745429992676,
+ "step": 3526
+ },
+ {
+ "epoch": 48.992139737991266,
+ "grad_norm": 0.011820166371762753,
+ "learning_rate": 0.0006,
+ "loss": 5.238091468811035,
+ "step": 3527
+ },
+ {
+ "epoch": 49.0,
+ "grad_norm": 0.01255585066974163,
+ "learning_rate": 0.0006,
+ "loss": 5.231295585632324,
+ "step": 3528
+ },
+ {
+ "epoch": 49.0,
+ "eval_loss": 5.394587993621826,
+ "eval_runtime": 56.3734,
+ "eval_samples_per_second": 43.318,
+ "eval_steps_per_second": 1.366,
+ "step": 3528
+ },
+ {
+ "epoch": 49.01397379912664,
+ "grad_norm": 0.011031972244381905,
+ "learning_rate": 0.0006,
+ "loss": 5.1354570388793945,
+ "step": 3529
+ },
+ {
+ "epoch": 49.02794759825328,
+ "grad_norm": 0.010631263256072998,
+ "learning_rate": 0.0006,
+ "loss": 5.177639961242676,
+ "step": 3530
+ },
+ {
+ "epoch": 49.041921397379916,
+ "grad_norm": 0.01133775431662798,
+ "learning_rate": 0.0006,
+ "loss": 5.283428192138672,
+ "step": 3531
+ },
+ {
+ "epoch": 49.05589519650655,
+ "grad_norm": 0.01154172234237194,
+ "learning_rate": 0.0006,
+ "loss": 5.206364154815674,
+ "step": 3532
+ },
+ {
+ "epoch": 49.069868995633186,
+ "grad_norm": 0.010192356072366238,
+ "learning_rate": 0.0006,
+ "loss": 5.2385735511779785,
+ "step": 3533
+ },
+ {
+ "epoch": 49.083842794759825,
+ "grad_norm": 0.012789200060069561,
+ "learning_rate": 0.0006,
+ "loss": 5.071800231933594,
+ "step": 3534
+ },
+ {
+ "epoch": 49.09781659388646,
+ "grad_norm": 0.011282820254564285,
+ "learning_rate": 0.0006,
+ "loss": 5.218020439147949,
+ "step": 3535
+ },
+ {
+ "epoch": 49.1117903930131,
+ "grad_norm": 0.012153448536992073,
+ "learning_rate": 0.0006,
+ "loss": 5.134698867797852,
+ "step": 3536
+ },
+ {
+ "epoch": 49.12576419213974,
+ "grad_norm": 0.011920543387532234,
+ "learning_rate": 0.0006,
+ "loss": 5.141620635986328,
+ "step": 3537
+ },
+ {
+ "epoch": 49.13973799126637,
+ "grad_norm": 0.010949329473078251,
+ "learning_rate": 0.0006,
+ "loss": 5.116259574890137,
+ "step": 3538
+ },
+ {
+ "epoch": 49.15371179039301,
+ "grad_norm": 0.012362448498606682,
+ "learning_rate": 0.0006,
+ "loss": 5.090177059173584,
+ "step": 3539
+ },
+ {
+ "epoch": 49.16768558951965,
+ "grad_norm": 0.011575071141123772,
+ "learning_rate": 0.0006,
+ "loss": 5.093441009521484,
+ "step": 3540
+ },
+ {
+ "epoch": 49.18165938864629,
+ "grad_norm": 0.011126236990094185,
+ "learning_rate": 0.0006,
+ "loss": 5.060235023498535,
+ "step": 3541
+ },
+ {
+ "epoch": 49.19563318777293,
+ "grad_norm": 0.0107401292771101,
+ "learning_rate": 0.0006,
+ "loss": 5.140704154968262,
+ "step": 3542
+ },
+ {
+ "epoch": 49.209606986899566,
+ "grad_norm": 0.011202906258404255,
+ "learning_rate": 0.0006,
+ "loss": 5.10855770111084,
+ "step": 3543
+ },
+ {
+ "epoch": 49.223580786026204,
+ "grad_norm": 0.011083454824984074,
+ "learning_rate": 0.0006,
+ "loss": 5.092751502990723,
+ "step": 3544
+ },
+ {
+ "epoch": 49.237554585152836,
+ "grad_norm": 0.009523800574243069,
+ "learning_rate": 0.0006,
+ "loss": 5.13733434677124,
+ "step": 3545
+ },
+ {
+ "epoch": 49.251528384279474,
+ "grad_norm": 0.011298353783786297,
+ "learning_rate": 0.0006,
+ "loss": 5.036048889160156,
+ "step": 3546
+ },
+ {
+ "epoch": 49.26550218340611,
+ "grad_norm": 0.010676483623683453,
+ "learning_rate": 0.0006,
+ "loss": 5.051901340484619,
+ "step": 3547
+ },
+ {
+ "epoch": 49.27947598253275,
+ "grad_norm": 0.011463399045169353,
+ "learning_rate": 0.0006,
+ "loss": 4.983170986175537,
+ "step": 3548
+ },
+ {
+ "epoch": 49.29344978165939,
+ "grad_norm": 0.010603402741253376,
+ "learning_rate": 0.0006,
+ "loss": 5.112311363220215,
+ "step": 3549
+ },
+ {
+ "epoch": 49.30742358078603,
+ "grad_norm": 0.010535670444369316,
+ "learning_rate": 0.0006,
+ "loss": 5.085227966308594,
+ "step": 3550
+ },
+ {
+ "epoch": 49.32139737991266,
+ "grad_norm": 0.011843928135931492,
+ "learning_rate": 0.0006,
+ "loss": 5.07575798034668,
+ "step": 3551
+ },
+ {
+ "epoch": 49.3353711790393,
+ "grad_norm": 0.010587952099740505,
+ "learning_rate": 0.0006,
+ "loss": 5.019237041473389,
+ "step": 3552
+ },
+ {
+ "epoch": 49.34934497816594,
+ "grad_norm": 0.012285659089684486,
+ "learning_rate": 0.0006,
+ "loss": 5.144603729248047,
+ "step": 3553
+ },
+ {
+ "epoch": 49.36331877729258,
+ "grad_norm": 0.013119431212544441,
+ "learning_rate": 0.0006,
+ "loss": 5.035256862640381,
+ "step": 3554
+ },
+ {
+ "epoch": 49.377292576419215,
+ "grad_norm": 0.012138537131249905,
+ "learning_rate": 0.0006,
+ "loss": 4.996667861938477,
+ "step": 3555
+ },
+ {
+ "epoch": 49.391266375545854,
+ "grad_norm": 0.012142804451286793,
+ "learning_rate": 0.0006,
+ "loss": 4.95058536529541,
+ "step": 3556
+ },
+ {
+ "epoch": 49.40524017467249,
+ "grad_norm": 0.012205676175653934,
+ "learning_rate": 0.0006,
+ "loss": 5.096931457519531,
+ "step": 3557
+ },
+ {
+ "epoch": 49.419213973799124,
+ "grad_norm": 0.012086744420230389,
+ "learning_rate": 0.0006,
+ "loss": 4.99384880065918,
+ "step": 3558
+ },
+ {
+ "epoch": 49.43318777292576,
+ "grad_norm": 0.012918103486299515,
+ "learning_rate": 0.0006,
+ "loss": 5.031294822692871,
+ "step": 3559
+ },
+ {
+ "epoch": 49.4471615720524,
+ "grad_norm": 0.012006713077425957,
+ "learning_rate": 0.0006,
+ "loss": 5.0547943115234375,
+ "step": 3560
+ },
+ {
+ "epoch": 49.46113537117904,
+ "grad_norm": 0.0120399734005332,
+ "learning_rate": 0.0006,
+ "loss": 5.014238357543945,
+ "step": 3561
+ },
+ {
+ "epoch": 49.47510917030568,
+ "grad_norm": 0.013306990265846252,
+ "learning_rate": 0.0006,
+ "loss": 4.972925186157227,
+ "step": 3562
+ },
+ {
+ "epoch": 49.48908296943232,
+ "grad_norm": 0.013878636993467808,
+ "learning_rate": 0.0006,
+ "loss": 5.079033851623535,
+ "step": 3563
+ },
+ {
+ "epoch": 49.50305676855895,
+ "grad_norm": 0.012454566545784473,
+ "learning_rate": 0.0006,
+ "loss": 4.93708610534668,
+ "step": 3564
+ },
+ {
+ "epoch": 49.51703056768559,
+ "grad_norm": 0.014075061306357384,
+ "learning_rate": 0.0006,
+ "loss": 4.965582370758057,
+ "step": 3565
+ },
+ {
+ "epoch": 49.531004366812226,
+ "grad_norm": 0.012525617145001888,
+ "learning_rate": 0.0006,
+ "loss": 4.968148708343506,
+ "step": 3566
+ },
+ {
+ "epoch": 49.544978165938865,
+ "grad_norm": 0.012418225407600403,
+ "learning_rate": 0.0006,
+ "loss": 5.06901741027832,
+ "step": 3567
+ },
+ {
+ "epoch": 49.5589519650655,
+ "grad_norm": 0.015034239739179611,
+ "learning_rate": 0.0006,
+ "loss": 4.984277725219727,
+ "step": 3568
+ },
+ {
+ "epoch": 49.57292576419214,
+ "grad_norm": 0.014329628087580204,
+ "learning_rate": 0.0006,
+ "loss": 4.951558589935303,
+ "step": 3569
+ },
+ {
+ "epoch": 49.58689956331878,
+ "grad_norm": 0.017769070342183113,
+ "learning_rate": 0.0006,
+ "loss": 4.966532230377197,
+ "step": 3570
+ },
+ {
+ "epoch": 49.60087336244541,
+ "grad_norm": 0.01909453794360161,
+ "learning_rate": 0.0006,
+ "loss": 4.919890880584717,
+ "step": 3571
+ },
+ {
+ "epoch": 49.61484716157205,
+ "grad_norm": 0.017519688233733177,
+ "learning_rate": 0.0006,
+ "loss": 5.031286716461182,
+ "step": 3572
+ },
+ {
+ "epoch": 49.62882096069869,
+ "grad_norm": 0.017143897712230682,
+ "learning_rate": 0.0006,
+ "loss": 4.969764709472656,
+ "step": 3573
+ },
+ {
+ "epoch": 49.64279475982533,
+ "grad_norm": 0.01688285917043686,
+ "learning_rate": 0.0006,
+ "loss": 5.0733442306518555,
+ "step": 3574
+ },
+ {
+ "epoch": 49.65676855895197,
+ "grad_norm": 0.018547311425209045,
+ "learning_rate": 0.0006,
+ "loss": 4.980863571166992,
+ "step": 3575
+ },
+ {
+ "epoch": 49.670742358078606,
+ "grad_norm": 0.018877137452363968,
+ "learning_rate": 0.0006,
+ "loss": 4.954646110534668,
+ "step": 3576
+ },
+ {
+ "epoch": 49.68471615720524,
+ "grad_norm": 0.018681600689888,
+ "learning_rate": 0.0006,
+ "loss": 4.959702014923096,
+ "step": 3577
+ },
+ {
+ "epoch": 49.698689956331876,
+ "grad_norm": 0.018947336822748184,
+ "learning_rate": 0.0006,
+ "loss": 5.070502758026123,
+ "step": 3578
+ },
+ {
+ "epoch": 49.712663755458514,
+ "grad_norm": 0.018817484378814697,
+ "learning_rate": 0.0006,
+ "loss": 5.056508541107178,
+ "step": 3579
+ },
+ {
+ "epoch": 49.72663755458515,
+ "grad_norm": 0.01931724324822426,
+ "learning_rate": 0.0006,
+ "loss": 4.922451972961426,
+ "step": 3580
+ },
+ {
+ "epoch": 49.74061135371179,
+ "grad_norm": 0.022047819569706917,
+ "learning_rate": 0.0006,
+ "loss": 5.053170204162598,
+ "step": 3581
+ },
+ {
+ "epoch": 49.75458515283843,
+ "grad_norm": 0.02279801294207573,
+ "learning_rate": 0.0006,
+ "loss": 4.8768439292907715,
+ "step": 3582
+ },
+ {
+ "epoch": 49.76855895196506,
+ "grad_norm": 0.020661601796746254,
+ "learning_rate": 0.0006,
+ "loss": 4.907928466796875,
+ "step": 3583
+ },
+ {
+ "epoch": 49.7825327510917,
+ "grad_norm": 0.020479509606957436,
+ "learning_rate": 0.0006,
+ "loss": 5.040694236755371,
+ "step": 3584
+ },
+ {
+ "epoch": 49.79650655021834,
+ "grad_norm": 0.021654162555933,
+ "learning_rate": 0.0006,
+ "loss": 5.018434524536133,
+ "step": 3585
+ },
+ {
+ "epoch": 49.81048034934498,
+ "grad_norm": 0.022486824542284012,
+ "learning_rate": 0.0006,
+ "loss": 5.086538314819336,
+ "step": 3586
+ },
+ {
+ "epoch": 49.82445414847162,
+ "grad_norm": 0.02141369879245758,
+ "learning_rate": 0.0006,
+ "loss": 5.035694122314453,
+ "step": 3587
+ },
+ {
+ "epoch": 49.838427947598255,
+ "grad_norm": 0.015926677733659744,
+ "learning_rate": 0.0006,
+ "loss": 4.905038833618164,
+ "step": 3588
+ },
+ {
+ "epoch": 49.852401746724894,
+ "grad_norm": 0.015775298699736595,
+ "learning_rate": 0.0006,
+ "loss": 5.003902435302734,
+ "step": 3589
+ },
+ {
+ "epoch": 49.866375545851525,
+ "grad_norm": 0.01872118003666401,
+ "learning_rate": 0.0006,
+ "loss": 4.891060829162598,
+ "step": 3590
+ },
+ {
+ "epoch": 49.880349344978164,
+ "grad_norm": 0.02066192403435707,
+ "learning_rate": 0.0006,
+ "loss": 5.002897262573242,
+ "step": 3591
+ },
+ {
+ "epoch": 49.8943231441048,
+ "grad_norm": 0.02348714880645275,
+ "learning_rate": 0.0006,
+ "loss": 4.9189958572387695,
+ "step": 3592
+ },
+ {
+ "epoch": 49.90829694323144,
+ "grad_norm": 0.020402103662490845,
+ "learning_rate": 0.0006,
+ "loss": 4.992661476135254,
+ "step": 3593
+ },
+ {
+ "epoch": 49.92227074235808,
+ "grad_norm": 0.020466072484850883,
+ "learning_rate": 0.0006,
+ "loss": 4.962968826293945,
+ "step": 3594
+ },
+ {
+ "epoch": 49.93624454148472,
+ "grad_norm": 0.01922314241528511,
+ "learning_rate": 0.0006,
+ "loss": 4.897730827331543,
+ "step": 3595
+ },
+ {
+ "epoch": 49.95021834061135,
+ "grad_norm": 0.017228754237294197,
+ "learning_rate": 0.0006,
+ "loss": 4.964842796325684,
+ "step": 3596
+ },
+ {
+ "epoch": 49.96419213973799,
+ "grad_norm": 0.016021743416786194,
+ "learning_rate": 0.0006,
+ "loss": 4.910758972167969,
+ "step": 3597
+ },
+ {
+ "epoch": 49.97816593886463,
+ "grad_norm": 0.017985103651881218,
+ "learning_rate": 0.0006,
+ "loss": 4.916072368621826,
+ "step": 3598
+ },
+ {
+ "epoch": 49.992139737991266,
+ "grad_norm": 0.018780240789055824,
+ "learning_rate": 0.0006,
+ "loss": 4.94658088684082,
+ "step": 3599
+ },
+ {
+ "epoch": 50.0,
+ "grad_norm": 0.01776314526796341,
+ "learning_rate": 0.0006,
+ "loss": 4.969357967376709,
+ "step": 3600
+ },
+ {
+ "epoch": 50.0,
+ "eval_loss": 5.141158580780029,
+ "eval_runtime": 56.2332,
+ "eval_samples_per_second": 43.426,
+ "eval_steps_per_second": 1.369,
+ "step": 3600
+ },
+ {
+ "epoch": 50.01397379912664,
+ "grad_norm": 0.016308696940541267,
+ "learning_rate": 0.0006,
+ "loss": 4.924664497375488,
+ "step": 3601
+ },
+ {
+ "epoch": 50.02794759825328,
+ "grad_norm": 0.018810199573636055,
+ "learning_rate": 0.0006,
+ "loss": 4.82293701171875,
+ "step": 3602
+ },
+ {
+ "epoch": 50.041921397379916,
+ "grad_norm": 0.02227160707116127,
+ "learning_rate": 0.0006,
+ "loss": 4.9612298011779785,
+ "step": 3603
+ },
+ {
+ "epoch": 50.05589519650655,
+ "grad_norm": 0.025266462936997414,
+ "learning_rate": 0.0006,
+ "loss": 4.963835716247559,
+ "step": 3604
+ },
+ {
+ "epoch": 50.069868995633186,
+ "grad_norm": 0.028588181361556053,
+ "learning_rate": 0.0006,
+ "loss": 4.992118835449219,
+ "step": 3605
+ },
+ {
+ "epoch": 50.083842794759825,
+ "grad_norm": 0.02739487960934639,
+ "learning_rate": 0.0006,
+ "loss": 4.915547847747803,
+ "step": 3606
+ },
+ {
+ "epoch": 50.09781659388646,
+ "grad_norm": 0.027000300586223602,
+ "learning_rate": 0.0006,
+ "loss": 4.883801460266113,
+ "step": 3607
+ },
+ {
+ "epoch": 50.1117903930131,
+ "grad_norm": 0.023468483239412308,
+ "learning_rate": 0.0006,
+ "loss": 5.006168365478516,
+ "step": 3608
+ },
+ {
+ "epoch": 50.12576419213974,
+ "grad_norm": 0.026753634214401245,
+ "learning_rate": 0.0006,
+ "loss": 4.92526912689209,
+ "step": 3609
+ },
+ {
+ "epoch": 50.13973799126637,
+ "grad_norm": 0.026816733181476593,
+ "learning_rate": 0.0006,
+ "loss": 4.847546100616455,
+ "step": 3610
+ },
+ {
+ "epoch": 50.15371179039301,
+ "grad_norm": 0.027362391352653503,
+ "learning_rate": 0.0006,
+ "loss": 4.862548351287842,
+ "step": 3611
+ },
+ {
+ "epoch": 50.16768558951965,
+ "grad_norm": 0.03450845181941986,
+ "learning_rate": 0.0006,
+ "loss": 4.963202953338623,
+ "step": 3612
+ },
+ {
+ "epoch": 50.18165938864629,
+ "grad_norm": 0.03801408037543297,
+ "learning_rate": 0.0006,
+ "loss": 4.770240306854248,
+ "step": 3613
+ },
+ {
+ "epoch": 50.19563318777293,
+ "grad_norm": 0.03579140082001686,
+ "learning_rate": 0.0006,
+ "loss": 4.797981262207031,
+ "step": 3614
+ },
+ {
+ "epoch": 50.209606986899566,
+ "grad_norm": 0.029393760487437248,
+ "learning_rate": 0.0006,
+ "loss": 4.798581600189209,
+ "step": 3615
+ },
+ {
+ "epoch": 50.223580786026204,
+ "grad_norm": 0.026111142709851265,
+ "learning_rate": 0.0006,
+ "loss": 4.990712642669678,
+ "step": 3616
+ },
+ {
+ "epoch": 50.237554585152836,
+ "grad_norm": 0.02213387005031109,
+ "learning_rate": 0.0006,
+ "loss": 4.865842342376709,
+ "step": 3617
+ },
+ {
+ "epoch": 50.251528384279474,
+ "grad_norm": 0.025450684130191803,
+ "learning_rate": 0.0006,
+ "loss": 4.903746128082275,
+ "step": 3618
+ },
+ {
+ "epoch": 50.26550218340611,
+ "grad_norm": 0.020915117114782333,
+ "learning_rate": 0.0006,
+ "loss": 4.951366424560547,
+ "step": 3619
+ },
+ {
+ "epoch": 50.27947598253275,
+ "grad_norm": 0.020972849801182747,
+ "learning_rate": 0.0006,
+ "loss": 4.783311367034912,
+ "step": 3620
+ },
+ {
+ "epoch": 50.29344978165939,
+ "grad_norm": 0.016956273466348648,
+ "learning_rate": 0.0006,
+ "loss": 4.955767631530762,
+ "step": 3621
+ },
+ {
+ "epoch": 50.30742358078603,
+ "grad_norm": 0.018045252189040184,
+ "learning_rate": 0.0006,
+ "loss": 4.869322776794434,
+ "step": 3622
+ },
+ {
+ "epoch": 50.32139737991266,
+ "grad_norm": 0.018199188634753227,
+ "learning_rate": 0.0006,
+ "loss": 4.95680046081543,
+ "step": 3623
+ },
+ {
+ "epoch": 50.3353711790393,
+ "grad_norm": 0.016688082367181778,
+ "learning_rate": 0.0006,
+ "loss": 4.783535957336426,
+ "step": 3624
+ },
+ {
+ "epoch": 50.34934497816594,
+ "grad_norm": 0.017540352419018745,
+ "learning_rate": 0.0006,
+ "loss": 4.8236083984375,
+ "step": 3625
+ },
+ {
+ "epoch": 50.36331877729258,
+ "grad_norm": 0.018054187297821045,
+ "learning_rate": 0.0006,
+ "loss": 4.764655113220215,
+ "step": 3626
+ },
+ {
+ "epoch": 50.377292576419215,
+ "grad_norm": 0.01741131953895092,
+ "learning_rate": 0.0006,
+ "loss": 4.875613689422607,
+ "step": 3627
+ },
+ {
+ "epoch": 50.391266375545854,
+ "grad_norm": 0.01682269014418125,
+ "learning_rate": 0.0006,
+ "loss": 4.818509101867676,
+ "step": 3628
+ },
+ {
+ "epoch": 50.40524017467249,
+ "grad_norm": 0.016652341932058334,
+ "learning_rate": 0.0006,
+ "loss": 4.974020481109619,
+ "step": 3629
+ },
+ {
+ "epoch": 50.419213973799124,
+ "grad_norm": 0.02035282365977764,
+ "learning_rate": 0.0006,
+ "loss": 4.893037796020508,
+ "step": 3630
+ },
+ {
+ "epoch": 50.43318777292576,
+ "grad_norm": 0.019907061010599136,
+ "learning_rate": 0.0006,
+ "loss": 4.848119735717773,
+ "step": 3631
+ },
+ {
+ "epoch": 50.4471615720524,
+ "grad_norm": 0.023861493915319443,
+ "learning_rate": 0.0006,
+ "loss": 4.819353103637695,
+ "step": 3632
+ },
+ {
+ "epoch": 50.46113537117904,
+ "grad_norm": 0.025254026055336,
+ "learning_rate": 0.0006,
+ "loss": 4.769312858581543,
+ "step": 3633
+ },
+ {
+ "epoch": 50.47510917030568,
+ "grad_norm": 0.026154937222599983,
+ "learning_rate": 0.0006,
+ "loss": 4.7586774826049805,
+ "step": 3634
+ },
+ {
+ "epoch": 50.48908296943232,
+ "grad_norm": 0.026342090219259262,
+ "learning_rate": 0.0006,
+ "loss": 4.861218452453613,
+ "step": 3635
+ },
+ {
+ "epoch": 50.50305676855895,
+ "grad_norm": 0.028169559314846992,
+ "learning_rate": 0.0006,
+ "loss": 4.818380355834961,
+ "step": 3636
+ },
+ {
+ "epoch": 50.51703056768559,
+ "grad_norm": 0.02820264920592308,
+ "learning_rate": 0.0006,
+ "loss": 4.892871856689453,
+ "step": 3637
+ },
+ {
+ "epoch": 50.531004366812226,
+ "grad_norm": 0.03345080465078354,
+ "learning_rate": 0.0006,
+ "loss": 4.811084747314453,
+ "step": 3638
+ },
+ {
+ "epoch": 50.544978165938865,
+ "grad_norm": 0.04005736857652664,
+ "learning_rate": 0.0006,
+ "loss": 4.765410900115967,
+ "step": 3639
+ },
+ {
+ "epoch": 50.5589519650655,
+ "grad_norm": 0.04333870857954025,
+ "learning_rate": 0.0006,
+ "loss": 4.956520080566406,
+ "step": 3640
+ },
+ {
+ "epoch": 50.57292576419214,
+ "grad_norm": 0.04216952994465828,
+ "learning_rate": 0.0006,
+ "loss": 4.793441295623779,
+ "step": 3641
+ },
+ {
+ "epoch": 50.58689956331878,
+ "grad_norm": 0.02936951443552971,
+ "learning_rate": 0.0006,
+ "loss": 4.8447957038879395,
+ "step": 3642
+ },
+ {
+ "epoch": 50.60087336244541,
+ "grad_norm": 0.026937026530504227,
+ "learning_rate": 0.0006,
+ "loss": 4.962215423583984,
+ "step": 3643
+ },
+ {
+ "epoch": 50.61484716157205,
+ "grad_norm": 0.02628890983760357,
+ "learning_rate": 0.0006,
+ "loss": 4.883061408996582,
+ "step": 3644
+ },
+ {
+ "epoch": 50.62882096069869,
+ "grad_norm": 0.025984834879636765,
+ "learning_rate": 0.0006,
+ "loss": 4.768662452697754,
+ "step": 3645
+ },
+ {
+ "epoch": 50.64279475982533,
+ "grad_norm": 0.023998012766242027,
+ "learning_rate": 0.0006,
+ "loss": 4.851325035095215,
+ "step": 3646
+ },
+ {
+ "epoch": 50.65676855895197,
+ "grad_norm": 0.02143184095621109,
+ "learning_rate": 0.0006,
+ "loss": 4.8274688720703125,
+ "step": 3647
+ },
+ {
+ "epoch": 50.670742358078606,
+ "grad_norm": 0.02488900162279606,
+ "learning_rate": 0.0006,
+ "loss": 4.831140518188477,
+ "step": 3648
+ },
+ {
+ "epoch": 50.68471615720524,
+ "grad_norm": 0.02330533228814602,
+ "learning_rate": 0.0006,
+ "loss": 4.850462913513184,
+ "step": 3649
+ },
+ {
+ "epoch": 50.698689956331876,
+ "grad_norm": 0.02121943049132824,
+ "learning_rate": 0.0006,
+ "loss": 4.870253086090088,
+ "step": 3650
+ },
+ {
+ "epoch": 50.712663755458514,
+ "grad_norm": 0.02080935798585415,
+ "learning_rate": 0.0006,
+ "loss": 4.798770904541016,
+ "step": 3651
+ },
+ {
+ "epoch": 50.72663755458515,
+ "grad_norm": 0.019047029316425323,
+ "learning_rate": 0.0006,
+ "loss": 4.8427300453186035,
+ "step": 3652
+ },
+ {
+ "epoch": 50.74061135371179,
+ "grad_norm": 0.01824103854596615,
+ "learning_rate": 0.0006,
+ "loss": 4.820980548858643,
+ "step": 3653
+ },
+ {
+ "epoch": 50.75458515283843,
+ "grad_norm": 0.018129967153072357,
+ "learning_rate": 0.0006,
+ "loss": 4.946732997894287,
+ "step": 3654
+ },
+ {
+ "epoch": 50.76855895196506,
+ "grad_norm": 0.01784815452992916,
+ "learning_rate": 0.0006,
+ "loss": 4.829516410827637,
+ "step": 3655
+ },
+ {
+ "epoch": 50.7825327510917,
+ "grad_norm": 0.020021555945277214,
+ "learning_rate": 0.0006,
+ "loss": 4.780428409576416,
+ "step": 3656
+ },
+ {
+ "epoch": 50.79650655021834,
+ "grad_norm": 0.01863253302872181,
+ "learning_rate": 0.0006,
+ "loss": 4.791567802429199,
+ "step": 3657
+ },
+ {
+ "epoch": 50.81048034934498,
+ "grad_norm": 0.019151683896780014,
+ "learning_rate": 0.0006,
+ "loss": 4.824892997741699,
+ "step": 3658
+ },
+ {
+ "epoch": 50.82445414847162,
+ "grad_norm": 0.01891648955643177,
+ "learning_rate": 0.0006,
+ "loss": 4.789605140686035,
+ "step": 3659
+ },
+ {
+ "epoch": 50.838427947598255,
+ "grad_norm": 0.016019510105252266,
+ "learning_rate": 0.0006,
+ "loss": 4.779555320739746,
+ "step": 3660
+ },
+ {
+ "epoch": 50.852401746724894,
+ "grad_norm": 0.016096491366624832,
+ "learning_rate": 0.0006,
+ "loss": 4.753842353820801,
+ "step": 3661
+ },
+ {
+ "epoch": 50.866375545851525,
+ "grad_norm": 0.01693476364016533,
+ "learning_rate": 0.0006,
+ "loss": 4.768477439880371,
+ "step": 3662
+ },
+ {
+ "epoch": 50.880349344978164,
+ "grad_norm": 0.01749575324356556,
+ "learning_rate": 0.0006,
+ "loss": 4.779143333435059,
+ "step": 3663
+ },
+ {
+ "epoch": 50.8943231441048,
+ "grad_norm": 0.016851043328642845,
+ "learning_rate": 0.0006,
+ "loss": 4.780387878417969,
+ "step": 3664
+ },
+ {
+ "epoch": 50.90829694323144,
+ "grad_norm": 0.019045202061533928,
+ "learning_rate": 0.0006,
+ "loss": 4.769913673400879,
+ "step": 3665
+ },
+ {
+ "epoch": 50.92227074235808,
+ "grad_norm": 0.01954123005270958,
+ "learning_rate": 0.0006,
+ "loss": 4.769009590148926,
+ "step": 3666
+ },
+ {
+ "epoch": 50.93624454148472,
+ "grad_norm": 0.022527096793055534,
+ "learning_rate": 0.0006,
+ "loss": 4.789215087890625,
+ "step": 3667
+ },
+ {
+ "epoch": 50.95021834061135,
+ "grad_norm": 0.02871634066104889,
+ "learning_rate": 0.0006,
+ "loss": 4.749727725982666,
+ "step": 3668
+ },
+ {
+ "epoch": 50.96419213973799,
+ "grad_norm": 0.04045053571462631,
+ "learning_rate": 0.0006,
+ "loss": 4.909038543701172,
+ "step": 3669
+ },
+ {
+ "epoch": 50.97816593886463,
+ "grad_norm": 0.04156879335641861,
+ "learning_rate": 0.0006,
+ "loss": 4.876831531524658,
+ "step": 3670
+ },
+ {
+ "epoch": 50.992139737991266,
+ "grad_norm": 0.02621397376060486,
+ "learning_rate": 0.0006,
+ "loss": 4.797173500061035,
+ "step": 3671
+ },
+ {
+ "epoch": 51.0,
+ "grad_norm": 0.03389597684144974,
+ "learning_rate": 0.0006,
+ "loss": 4.8428850173950195,
+ "step": 3672
+ },
+ {
+ "epoch": 51.0,
+ "eval_loss": 4.989261627197266,
+ "eval_runtime": 56.358,
+ "eval_samples_per_second": 43.33,
+ "eval_steps_per_second": 1.366,
+ "step": 3672
+ },
+ {
+ "epoch": 51.01397379912664,
+ "grad_norm": 0.03216972202062607,
+ "learning_rate": 0.0006,
+ "loss": 4.755102634429932,
+ "step": 3673
+ },
+ {
+ "epoch": 51.02794759825328,
+ "grad_norm": 0.03190374746918678,
+ "learning_rate": 0.0006,
+ "loss": 4.720424652099609,
+ "step": 3674
+ },
+ {
+ "epoch": 51.041921397379916,
+ "grad_norm": 0.031727299094200134,
+ "learning_rate": 0.0006,
+ "loss": 4.803277969360352,
+ "step": 3675
+ },
+ {
+ "epoch": 51.05589519650655,
+ "grad_norm": 0.02473079040646553,
+ "learning_rate": 0.0006,
+ "loss": 4.761087417602539,
+ "step": 3676
+ },
+ {
+ "epoch": 51.069868995633186,
+ "grad_norm": 0.020360209047794342,
+ "learning_rate": 0.0006,
+ "loss": 4.753884792327881,
+ "step": 3677
+ },
+ {
+ "epoch": 51.083842794759825,
+ "grad_norm": 0.02078923024237156,
+ "learning_rate": 0.0006,
+ "loss": 4.647130489349365,
+ "step": 3678
+ },
+ {
+ "epoch": 51.09781659388646,
+ "grad_norm": 0.019387200474739075,
+ "learning_rate": 0.0006,
+ "loss": 4.74135160446167,
+ "step": 3679
+ },
+ {
+ "epoch": 51.1117903930131,
+ "grad_norm": 0.01874949410557747,
+ "learning_rate": 0.0006,
+ "loss": 4.795329570770264,
+ "step": 3680
+ },
+ {
+ "epoch": 51.12576419213974,
+ "grad_norm": 0.019331837072968483,
+ "learning_rate": 0.0006,
+ "loss": 4.897967338562012,
+ "step": 3681
+ },
+ {
+ "epoch": 51.13973799126637,
+ "grad_norm": 0.017066050320863724,
+ "learning_rate": 0.0006,
+ "loss": 4.825351715087891,
+ "step": 3682
+ },
+ {
+ "epoch": 51.15371179039301,
+ "grad_norm": 0.01757810264825821,
+ "learning_rate": 0.0006,
+ "loss": 4.719335079193115,
+ "step": 3683
+ },
+ {
+ "epoch": 51.16768558951965,
+ "grad_norm": 0.018174266442656517,
+ "learning_rate": 0.0006,
+ "loss": 4.732680320739746,
+ "step": 3684
+ },
+ {
+ "epoch": 51.18165938864629,
+ "grad_norm": 0.018497727811336517,
+ "learning_rate": 0.0006,
+ "loss": 4.731050491333008,
+ "step": 3685
+ },
+ {
+ "epoch": 51.19563318777293,
+ "grad_norm": 0.0186338871717453,
+ "learning_rate": 0.0006,
+ "loss": 4.748035430908203,
+ "step": 3686
+ },
+ {
+ "epoch": 51.209606986899566,
+ "grad_norm": 0.018531711772084236,
+ "learning_rate": 0.0006,
+ "loss": 4.9007463455200195,
+ "step": 3687
+ },
+ {
+ "epoch": 51.223580786026204,
+ "grad_norm": 0.018221354112029076,
+ "learning_rate": 0.0006,
+ "loss": 4.84077262878418,
+ "step": 3688
+ },
+ {
+ "epoch": 51.237554585152836,
+ "grad_norm": 0.0203656405210495,
+ "learning_rate": 0.0006,
+ "loss": 4.73009729385376,
+ "step": 3689
+ },
+ {
+ "epoch": 51.251528384279474,
+ "grad_norm": 0.01964893378317356,
+ "learning_rate": 0.0006,
+ "loss": 4.772583961486816,
+ "step": 3690
+ },
+ {
+ "epoch": 51.26550218340611,
+ "grad_norm": 0.019606510177254677,
+ "learning_rate": 0.0006,
+ "loss": 4.599812030792236,
+ "step": 3691
+ },
+ {
+ "epoch": 51.27947598253275,
+ "grad_norm": 0.020501255989074707,
+ "learning_rate": 0.0006,
+ "loss": 4.809898853302002,
+ "step": 3692
+ },
+ {
+ "epoch": 51.29344978165939,
+ "grad_norm": 0.01870838552713394,
+ "learning_rate": 0.0006,
+ "loss": 4.741898059844971,
+ "step": 3693
+ },
+ {
+ "epoch": 51.30742358078603,
+ "grad_norm": 0.01802559196949005,
+ "learning_rate": 0.0006,
+ "loss": 4.746562480926514,
+ "step": 3694
+ },
+ {
+ "epoch": 51.32139737991266,
+ "grad_norm": 0.02053048275411129,
+ "learning_rate": 0.0006,
+ "loss": 4.69932746887207,
+ "step": 3695
+ },
+ {
+ "epoch": 51.3353711790393,
+ "grad_norm": 0.02661825157701969,
+ "learning_rate": 0.0006,
+ "loss": 4.68748664855957,
+ "step": 3696
+ },
+ {
+ "epoch": 51.34934497816594,
+ "grad_norm": 0.028269147500395775,
+ "learning_rate": 0.0006,
+ "loss": 4.731565475463867,
+ "step": 3697
+ },
+ {
+ "epoch": 51.36331877729258,
+ "grad_norm": 0.02650964818894863,
+ "learning_rate": 0.0006,
+ "loss": 4.733857154846191,
+ "step": 3698
+ },
+ {
+ "epoch": 51.377292576419215,
+ "grad_norm": 0.023696066811680794,
+ "learning_rate": 0.0006,
+ "loss": 4.812362194061279,
+ "step": 3699
+ },
+ {
+ "epoch": 51.391266375545854,
+ "grad_norm": 0.02365880273282528,
+ "learning_rate": 0.0006,
+ "loss": 4.733426094055176,
+ "step": 3700
+ },
+ {
+ "epoch": 51.40524017467249,
+ "grad_norm": 0.026249447837471962,
+ "learning_rate": 0.0006,
+ "loss": 4.715622901916504,
+ "step": 3701
+ },
+ {
+ "epoch": 51.419213973799124,
+ "grad_norm": 0.03301899880170822,
+ "learning_rate": 0.0006,
+ "loss": 4.7223663330078125,
+ "step": 3702
+ },
+ {
+ "epoch": 51.43318777292576,
+ "grad_norm": 0.044686660170555115,
+ "learning_rate": 0.0006,
+ "loss": 4.8121161460876465,
+ "step": 3703
+ },
+ {
+ "epoch": 51.4471615720524,
+ "grad_norm": 0.04733522981405258,
+ "learning_rate": 0.0006,
+ "loss": 4.841394424438477,
+ "step": 3704
+ },
+ {
+ "epoch": 51.46113537117904,
+ "grad_norm": 0.0395209975540638,
+ "learning_rate": 0.0006,
+ "loss": 4.773837089538574,
+ "step": 3705
+ },
+ {
+ "epoch": 51.47510917030568,
+ "grad_norm": 0.032790958881378174,
+ "learning_rate": 0.0006,
+ "loss": 4.833066463470459,
+ "step": 3706
+ },
+ {
+ "epoch": 51.48908296943232,
+ "grad_norm": 0.030578427016735077,
+ "learning_rate": 0.0006,
+ "loss": 4.685190200805664,
+ "step": 3707
+ },
+ {
+ "epoch": 51.50305676855895,
+ "grad_norm": 0.031619083136320114,
+ "learning_rate": 0.0006,
+ "loss": 4.740378379821777,
+ "step": 3708
+ },
+ {
+ "epoch": 51.51703056768559,
+ "grad_norm": 0.02385825850069523,
+ "learning_rate": 0.0006,
+ "loss": 4.812911510467529,
+ "step": 3709
+ },
+ {
+ "epoch": 51.531004366812226,
+ "grad_norm": 0.023523425683379173,
+ "learning_rate": 0.0006,
+ "loss": 4.669580459594727,
+ "step": 3710
+ },
+ {
+ "epoch": 51.544978165938865,
+ "grad_norm": 0.021316280588507652,
+ "learning_rate": 0.0006,
+ "loss": 4.76888370513916,
+ "step": 3711
+ },
+ {
+ "epoch": 51.5589519650655,
+ "grad_norm": 0.024663880467414856,
+ "learning_rate": 0.0006,
+ "loss": 4.702592372894287,
+ "step": 3712
+ },
+ {
+ "epoch": 51.57292576419214,
+ "grad_norm": 0.024173280224204063,
+ "learning_rate": 0.0006,
+ "loss": 4.710604667663574,
+ "step": 3713
+ },
+ {
+ "epoch": 51.58689956331878,
+ "grad_norm": 0.02754114381968975,
+ "learning_rate": 0.0006,
+ "loss": 4.746356964111328,
+ "step": 3714
+ },
+ {
+ "epoch": 51.60087336244541,
+ "grad_norm": 0.027903622016310692,
+ "learning_rate": 0.0006,
+ "loss": 4.7546491622924805,
+ "step": 3715
+ },
+ {
+ "epoch": 51.61484716157205,
+ "grad_norm": 0.024871617555618286,
+ "learning_rate": 0.0006,
+ "loss": 4.605746269226074,
+ "step": 3716
+ },
+ {
+ "epoch": 51.62882096069869,
+ "grad_norm": 0.021239031106233597,
+ "learning_rate": 0.0006,
+ "loss": 4.6942458152771,
+ "step": 3717
+ },
+ {
+ "epoch": 51.64279475982533,
+ "grad_norm": 0.01925622671842575,
+ "learning_rate": 0.0006,
+ "loss": 4.593558311462402,
+ "step": 3718
+ },
+ {
+ "epoch": 51.65676855895197,
+ "grad_norm": 0.01956283673644066,
+ "learning_rate": 0.0006,
+ "loss": 4.631172180175781,
+ "step": 3719
+ },
+ {
+ "epoch": 51.670742358078606,
+ "grad_norm": 0.02269871160387993,
+ "learning_rate": 0.0006,
+ "loss": 4.729236602783203,
+ "step": 3720
+ },
+ {
+ "epoch": 51.68471615720524,
+ "grad_norm": 0.02253916673362255,
+ "learning_rate": 0.0006,
+ "loss": 4.7191033363342285,
+ "step": 3721
+ },
+ {
+ "epoch": 51.698689956331876,
+ "grad_norm": 0.026231050491333008,
+ "learning_rate": 0.0006,
+ "loss": 4.811850547790527,
+ "step": 3722
+ },
+ {
+ "epoch": 51.712663755458514,
+ "grad_norm": 0.02558821812272072,
+ "learning_rate": 0.0006,
+ "loss": 4.681466102600098,
+ "step": 3723
+ },
+ {
+ "epoch": 51.72663755458515,
+ "grad_norm": 0.02124209702014923,
+ "learning_rate": 0.0006,
+ "loss": 4.693664073944092,
+ "step": 3724
+ },
+ {
+ "epoch": 51.74061135371179,
+ "grad_norm": 0.016638319939374924,
+ "learning_rate": 0.0006,
+ "loss": 4.771440029144287,
+ "step": 3725
+ },
+ {
+ "epoch": 51.75458515283843,
+ "grad_norm": 0.018048347905278206,
+ "learning_rate": 0.0006,
+ "loss": 4.643252372741699,
+ "step": 3726
+ },
+ {
+ "epoch": 51.76855895196506,
+ "grad_norm": 0.019904717803001404,
+ "learning_rate": 0.0006,
+ "loss": 4.66530704498291,
+ "step": 3727
+ },
+ {
+ "epoch": 51.7825327510917,
+ "grad_norm": 0.0194614976644516,
+ "learning_rate": 0.0006,
+ "loss": 4.711943626403809,
+ "step": 3728
+ },
+ {
+ "epoch": 51.79650655021834,
+ "grad_norm": 0.019175555557012558,
+ "learning_rate": 0.0006,
+ "loss": 4.721931457519531,
+ "step": 3729
+ },
+ {
+ "epoch": 51.81048034934498,
+ "grad_norm": 0.02093447372317314,
+ "learning_rate": 0.0006,
+ "loss": 4.633059024810791,
+ "step": 3730
+ },
+ {
+ "epoch": 51.82445414847162,
+ "grad_norm": 0.01930283196270466,
+ "learning_rate": 0.0006,
+ "loss": 4.659373760223389,
+ "step": 3731
+ },
+ {
+ "epoch": 51.838427947598255,
+ "grad_norm": 0.015350676141679287,
+ "learning_rate": 0.0006,
+ "loss": 4.763708114624023,
+ "step": 3732
+ },
+ {
+ "epoch": 51.852401746724894,
+ "grad_norm": 0.014139448292553425,
+ "learning_rate": 0.0006,
+ "loss": 4.753891468048096,
+ "step": 3733
+ },
+ {
+ "epoch": 51.866375545851525,
+ "grad_norm": 0.016215872019529343,
+ "learning_rate": 0.0006,
+ "loss": 4.658967018127441,
+ "step": 3734
+ },
+ {
+ "epoch": 51.880349344978164,
+ "grad_norm": 0.017328619956970215,
+ "learning_rate": 0.0006,
+ "loss": 4.696579933166504,
+ "step": 3735
+ },
+ {
+ "epoch": 51.8943231441048,
+ "grad_norm": 0.018103396520018578,
+ "learning_rate": 0.0006,
+ "loss": 4.627464771270752,
+ "step": 3736
+ },
+ {
+ "epoch": 51.90829694323144,
+ "grad_norm": 0.01715235412120819,
+ "learning_rate": 0.0006,
+ "loss": 4.7368011474609375,
+ "step": 3737
+ },
+ {
+ "epoch": 51.92227074235808,
+ "grad_norm": 0.014790094457566738,
+ "learning_rate": 0.0006,
+ "loss": 4.629441738128662,
+ "step": 3738
+ },
+ {
+ "epoch": 51.93624454148472,
+ "grad_norm": 0.017114296555519104,
+ "learning_rate": 0.0006,
+ "loss": 4.761256694793701,
+ "step": 3739
+ },
+ {
+ "epoch": 51.95021834061135,
+ "grad_norm": 0.019009845331311226,
+ "learning_rate": 0.0006,
+ "loss": 4.789155006408691,
+ "step": 3740
+ },
+ {
+ "epoch": 51.96419213973799,
+ "grad_norm": 0.019696692004799843,
+ "learning_rate": 0.0006,
+ "loss": 4.721269607543945,
+ "step": 3741
+ },
+ {
+ "epoch": 51.97816593886463,
+ "grad_norm": 0.020542822778224945,
+ "learning_rate": 0.0006,
+ "loss": 4.687038898468018,
+ "step": 3742
+ },
+ {
+ "epoch": 51.992139737991266,
+ "grad_norm": 0.021267160773277283,
+ "learning_rate": 0.0006,
+ "loss": 4.747860908508301,
+ "step": 3743
+ },
+ {
+ "epoch": 52.0,
+ "grad_norm": 0.025049567222595215,
+ "learning_rate": 0.0006,
+ "loss": 4.8473100662231445,
+ "step": 3744
+ },
+ {
+ "epoch": 52.0,
+ "eval_loss": 4.943043231964111,
+ "eval_runtime": 56.4929,
+ "eval_samples_per_second": 43.227,
+ "eval_steps_per_second": 1.363,
+ "step": 3744
+ },
+ {
+ "epoch": 52.01397379912664,
+ "grad_norm": 0.03054986707866192,
+ "learning_rate": 0.0006,
+ "loss": 4.856931686401367,
+ "step": 3745
+ },
+ {
+ "epoch": 52.02794759825328,
+ "grad_norm": 0.03838375583291054,
+ "learning_rate": 0.0006,
+ "loss": 4.699807167053223,
+ "step": 3746
+ },
+ {
+ "epoch": 52.041921397379916,
+ "grad_norm": 0.03383488208055496,
+ "learning_rate": 0.0006,
+ "loss": 4.6626787185668945,
+ "step": 3747
+ },
+ {
+ "epoch": 52.05589519650655,
+ "grad_norm": 0.0201561376452446,
+ "learning_rate": 0.0006,
+ "loss": 4.621831893920898,
+ "step": 3748
+ },
+ {
+ "epoch": 52.069868995633186,
+ "grad_norm": 0.028749438002705574,
+ "learning_rate": 0.0006,
+ "loss": 4.682540416717529,
+ "step": 3749
+ },
+ {
+ "epoch": 52.083842794759825,
+ "grad_norm": 0.02681174874305725,
+ "learning_rate": 0.0006,
+ "loss": 4.629997253417969,
+ "step": 3750
+ },
+ {
+ "epoch": 52.09781659388646,
+ "grad_norm": 0.024063318967819214,
+ "learning_rate": 0.0006,
+ "loss": 4.573021411895752,
+ "step": 3751
+ },
+ {
+ "epoch": 52.1117903930131,
+ "grad_norm": 0.02324330434203148,
+ "learning_rate": 0.0006,
+ "loss": 4.739465236663818,
+ "step": 3752
+ },
+ {
+ "epoch": 52.12576419213974,
+ "grad_norm": 0.0225426834076643,
+ "learning_rate": 0.0006,
+ "loss": 4.7335333824157715,
+ "step": 3753
+ },
+ {
+ "epoch": 52.13973799126637,
+ "grad_norm": 0.019593428820371628,
+ "learning_rate": 0.0006,
+ "loss": 4.704958438873291,
+ "step": 3754
+ },
+ {
+ "epoch": 52.15371179039301,
+ "grad_norm": 0.020777082070708275,
+ "learning_rate": 0.0006,
+ "loss": 4.638696670532227,
+ "step": 3755
+ },
+ {
+ "epoch": 52.16768558951965,
+ "grad_norm": 0.021525604650378227,
+ "learning_rate": 0.0006,
+ "loss": 4.643706798553467,
+ "step": 3756
+ },
+ {
+ "epoch": 52.18165938864629,
+ "grad_norm": 0.018943654373288155,
+ "learning_rate": 0.0006,
+ "loss": 4.5173211097717285,
+ "step": 3757
+ },
+ {
+ "epoch": 52.19563318777293,
+ "grad_norm": 0.01856580562889576,
+ "learning_rate": 0.0006,
+ "loss": 4.678788661956787,
+ "step": 3758
+ },
+ {
+ "epoch": 52.209606986899566,
+ "grad_norm": 0.019417259842157364,
+ "learning_rate": 0.0006,
+ "loss": 4.587244987487793,
+ "step": 3759
+ },
+ {
+ "epoch": 52.223580786026204,
+ "grad_norm": 0.01911911740899086,
+ "learning_rate": 0.0006,
+ "loss": 4.688503265380859,
+ "step": 3760
+ },
+ {
+ "epoch": 52.237554585152836,
+ "grad_norm": 0.021017879247665405,
+ "learning_rate": 0.0006,
+ "loss": 4.65829610824585,
+ "step": 3761
+ },
+ {
+ "epoch": 52.251528384279474,
+ "grad_norm": 0.02048875205218792,
+ "learning_rate": 0.0006,
+ "loss": 4.624055862426758,
+ "step": 3762
+ },
+ {
+ "epoch": 52.26550218340611,
+ "grad_norm": 0.01932772807776928,
+ "learning_rate": 0.0006,
+ "loss": 4.556562900543213,
+ "step": 3763
+ },
+ {
+ "epoch": 52.27947598253275,
+ "grad_norm": 0.01871614344418049,
+ "learning_rate": 0.0006,
+ "loss": 4.581201076507568,
+ "step": 3764
+ },
+ {
+ "epoch": 52.29344978165939,
+ "grad_norm": 0.017253106459975243,
+ "learning_rate": 0.0006,
+ "loss": 4.605219841003418,
+ "step": 3765
+ },
+ {
+ "epoch": 52.30742358078603,
+ "grad_norm": 0.01831238530576229,
+ "learning_rate": 0.0006,
+ "loss": 4.759900093078613,
+ "step": 3766
+ },
+ {
+ "epoch": 52.32139737991266,
+ "grad_norm": 0.017946207895874977,
+ "learning_rate": 0.0006,
+ "loss": 4.5598344802856445,
+ "step": 3767
+ },
+ {
+ "epoch": 52.3353711790393,
+ "grad_norm": 0.01680106669664383,
+ "learning_rate": 0.0006,
+ "loss": 4.605034828186035,
+ "step": 3768
+ },
+ {
+ "epoch": 52.34934497816594,
+ "grad_norm": 0.021482540294528008,
+ "learning_rate": 0.0006,
+ "loss": 4.620411396026611,
+ "step": 3769
+ },
+ {
+ "epoch": 52.36331877729258,
+ "grad_norm": 0.02210252732038498,
+ "learning_rate": 0.0006,
+ "loss": 4.667577743530273,
+ "step": 3770
+ },
+ {
+ "epoch": 52.377292576419215,
+ "grad_norm": 0.02237485907971859,
+ "learning_rate": 0.0006,
+ "loss": 4.628812789916992,
+ "step": 3771
+ },
+ {
+ "epoch": 52.391266375545854,
+ "grad_norm": 0.02196541801095009,
+ "learning_rate": 0.0006,
+ "loss": 4.572988510131836,
+ "step": 3772
+ },
+ {
+ "epoch": 52.40524017467249,
+ "grad_norm": 0.01930229924619198,
+ "learning_rate": 0.0006,
+ "loss": 4.588339805603027,
+ "step": 3773
+ },
+ {
+ "epoch": 52.419213973799124,
+ "grad_norm": 0.02222434990108013,
+ "learning_rate": 0.0006,
+ "loss": 4.766482353210449,
+ "step": 3774
+ },
+ {
+ "epoch": 52.43318777292576,
+ "grad_norm": 0.028806306421756744,
+ "learning_rate": 0.0006,
+ "loss": 4.617892742156982,
+ "step": 3775
+ },
+ {
+ "epoch": 52.4471615720524,
+ "grad_norm": 0.03306042402982712,
+ "learning_rate": 0.0006,
+ "loss": 4.665334701538086,
+ "step": 3776
+ },
+ {
+ "epoch": 52.46113537117904,
+ "grad_norm": 0.0358956903219223,
+ "learning_rate": 0.0006,
+ "loss": 4.566668510437012,
+ "step": 3777
+ },
+ {
+ "epoch": 52.47510917030568,
+ "grad_norm": 0.030406497418880463,
+ "learning_rate": 0.0006,
+ "loss": 4.513612270355225,
+ "step": 3778
+ },
+ {
+ "epoch": 52.48908296943232,
+ "grad_norm": 0.030295712873339653,
+ "learning_rate": 0.0006,
+ "loss": 4.640189170837402,
+ "step": 3779
+ },
+ {
+ "epoch": 52.50305676855895,
+ "grad_norm": 0.024981003254652023,
+ "learning_rate": 0.0006,
+ "loss": 4.57558012008667,
+ "step": 3780
+ },
+ {
+ "epoch": 52.51703056768559,
+ "grad_norm": 0.024990713223814964,
+ "learning_rate": 0.0006,
+ "loss": 4.708589553833008,
+ "step": 3781
+ },
+ {
+ "epoch": 52.531004366812226,
+ "grad_norm": 0.025034775957465172,
+ "learning_rate": 0.0006,
+ "loss": 4.70468282699585,
+ "step": 3782
+ },
+ {
+ "epoch": 52.544978165938865,
+ "grad_norm": 0.028243759647011757,
+ "learning_rate": 0.0006,
+ "loss": 4.712060451507568,
+ "step": 3783
+ },
+ {
+ "epoch": 52.5589519650655,
+ "grad_norm": 0.026984544470906258,
+ "learning_rate": 0.0006,
+ "loss": 4.505004405975342,
+ "step": 3784
+ },
+ {
+ "epoch": 52.57292576419214,
+ "grad_norm": 0.050843529403209686,
+ "learning_rate": 0.0006,
+ "loss": 4.593271255493164,
+ "step": 3785
+ },
+ {
+ "epoch": 52.58689956331878,
+ "grad_norm": 0.25422802567481995,
+ "learning_rate": 0.0006,
+ "loss": 4.7178802490234375,
+ "step": 3786
+ },
+ {
+ "epoch": 52.60087336244541,
+ "grad_norm": 0.6078406572341919,
+ "learning_rate": 0.0006,
+ "loss": 6.144372940063477,
+ "step": 3787
+ },
+ {
+ "epoch": 52.61484716157205,
+ "grad_norm": 0.25933101773262024,
+ "learning_rate": 0.0006,
+ "loss": 5.971083641052246,
+ "step": 3788
+ },
+ {
+ "epoch": 52.62882096069869,
+ "grad_norm": 0.294657826423645,
+ "learning_rate": 0.0006,
+ "loss": 6.204090118408203,
+ "step": 3789
+ },
+ {
+ "epoch": 52.64279475982533,
+ "grad_norm": 0.23641522228717804,
+ "learning_rate": 0.0006,
+ "loss": 6.869574546813965,
+ "step": 3790
+ },
+ {
+ "epoch": 52.65676855895197,
+ "grad_norm": 0.23717311024665833,
+ "learning_rate": 0.0006,
+ "loss": 6.739034175872803,
+ "step": 3791
+ },
+ {
+ "epoch": 52.670742358078606,
+ "grad_norm": 0.10300362855195999,
+ "learning_rate": 0.0006,
+ "loss": 6.49191951751709,
+ "step": 3792
+ },
+ {
+ "epoch": 52.68471615720524,
+ "grad_norm": 0.1484968513250351,
+ "learning_rate": 0.0006,
+ "loss": 6.6021270751953125,
+ "step": 3793
+ },
+ {
+ "epoch": 52.698689956331876,
+ "grad_norm": 0.08052115142345428,
+ "learning_rate": 0.0006,
+ "loss": 6.432796478271484,
+ "step": 3794
+ },
+ {
+ "epoch": 52.712663755458514,
+ "grad_norm": 0.06869253516197205,
+ "learning_rate": 0.0006,
+ "loss": 6.212784767150879,
+ "step": 3795
+ },
+ {
+ "epoch": 52.72663755458515,
+ "grad_norm": 0.05589582398533821,
+ "learning_rate": 0.0006,
+ "loss": 6.168288707733154,
+ "step": 3796
+ },
+ {
+ "epoch": 52.74061135371179,
+ "grad_norm": 0.04840940609574318,
+ "learning_rate": 0.0006,
+ "loss": 6.163174629211426,
+ "step": 3797
+ },
+ {
+ "epoch": 52.75458515283843,
+ "grad_norm": 0.04578601196408272,
+ "learning_rate": 0.0006,
+ "loss": 6.038390636444092,
+ "step": 3798
+ },
+ {
+ "epoch": 52.76855895196506,
+ "grad_norm": 0.04102053865790367,
+ "learning_rate": 0.0006,
+ "loss": 5.929539680480957,
+ "step": 3799
+ },
+ {
+ "epoch": 52.7825327510917,
+ "grad_norm": 0.038398485630750656,
+ "learning_rate": 0.0006,
+ "loss": 6.007378101348877,
+ "step": 3800
+ },
+ {
+ "epoch": 52.79650655021834,
+ "grad_norm": 0.03466642275452614,
+ "learning_rate": 0.0006,
+ "loss": 5.767291069030762,
+ "step": 3801
+ },
+ {
+ "epoch": 52.81048034934498,
+ "grad_norm": 0.03328905999660492,
+ "learning_rate": 0.0006,
+ "loss": 5.785032272338867,
+ "step": 3802
+ },
+ {
+ "epoch": 52.82445414847162,
+ "grad_norm": 0.03186681121587753,
+ "learning_rate": 0.0006,
+ "loss": 5.6942057609558105,
+ "step": 3803
+ },
+ {
+ "epoch": 52.838427947598255,
+ "grad_norm": 0.030553551390767097,
+ "learning_rate": 0.0006,
+ "loss": 5.680774688720703,
+ "step": 3804
+ },
+ {
+ "epoch": 52.852401746724894,
+ "grad_norm": 0.02578839473426342,
+ "learning_rate": 0.0006,
+ "loss": 5.703461647033691,
+ "step": 3805
+ },
+ {
+ "epoch": 52.866375545851525,
+ "grad_norm": 0.026261312887072563,
+ "learning_rate": 0.0006,
+ "loss": 5.642549514770508,
+ "step": 3806
+ },
+ {
+ "epoch": 52.880349344978164,
+ "grad_norm": 0.02782430872321129,
+ "learning_rate": 0.0006,
+ "loss": 5.563597679138184,
+ "step": 3807
+ },
+ {
+ "epoch": 52.8943231441048,
+ "grad_norm": 0.024467799812555313,
+ "learning_rate": 0.0006,
+ "loss": 5.583395957946777,
+ "step": 3808
+ },
+ {
+ "epoch": 52.90829694323144,
+ "grad_norm": 0.022602897137403488,
+ "learning_rate": 0.0006,
+ "loss": 5.50742769241333,
+ "step": 3809
+ },
+ {
+ "epoch": 52.92227074235808,
+ "grad_norm": 0.024292029440402985,
+ "learning_rate": 0.0006,
+ "loss": 5.367804527282715,
+ "step": 3810
+ },
+ {
+ "epoch": 52.93624454148472,
+ "grad_norm": 0.021062295883893967,
+ "learning_rate": 0.0006,
+ "loss": 5.423705101013184,
+ "step": 3811
+ },
+ {
+ "epoch": 52.95021834061135,
+ "grad_norm": 0.022226469591259956,
+ "learning_rate": 0.0006,
+ "loss": 5.344039440155029,
+ "step": 3812
+ },
+ {
+ "epoch": 52.96419213973799,
+ "grad_norm": 0.023691752925515175,
+ "learning_rate": 0.0006,
+ "loss": 5.2581892013549805,
+ "step": 3813
+ },
+ {
+ "epoch": 52.97816593886463,
+ "grad_norm": 0.02181999757885933,
+ "learning_rate": 0.0006,
+ "loss": 5.2710113525390625,
+ "step": 3814
+ },
+ {
+ "epoch": 52.992139737991266,
+ "grad_norm": 0.025295857340097427,
+ "learning_rate": 0.0006,
+ "loss": 5.278355598449707,
+ "step": 3815
+ },
+ {
+ "epoch": 53.0,
+ "grad_norm": 0.02216828241944313,
+ "learning_rate": 0.0006,
+ "loss": 5.127945899963379,
+ "step": 3816
+ },
+ {
+ "epoch": 53.0,
+ "eval_loss": 5.3411736488342285,
+ "eval_runtime": 56.2837,
+ "eval_samples_per_second": 43.387,
+ "eval_steps_per_second": 1.368,
+ "step": 3816
+ },
+ {
+ "epoch": 53.01397379912664,
+ "grad_norm": 0.022496026009321213,
+ "learning_rate": 0.0006,
+ "loss": 5.19655704498291,
+ "step": 3817
+ },
+ {
+ "epoch": 53.02794759825328,
+ "grad_norm": 0.03201168775558472,
+ "learning_rate": 0.0006,
+ "loss": 5.094621181488037,
+ "step": 3818
+ },
+ {
+ "epoch": 53.041921397379916,
+ "grad_norm": 0.02683178335428238,
+ "learning_rate": 0.0006,
+ "loss": 5.19899845123291,
+ "step": 3819
+ },
+ {
+ "epoch": 53.05589519650655,
+ "grad_norm": 0.022659169510006905,
+ "learning_rate": 0.0006,
+ "loss": 5.184103012084961,
+ "step": 3820
+ },
+ {
+ "epoch": 53.069868995633186,
+ "grad_norm": 0.024784443899989128,
+ "learning_rate": 0.0006,
+ "loss": 5.124286651611328,
+ "step": 3821
+ },
+ {
+ "epoch": 53.083842794759825,
+ "grad_norm": 0.028710370883345604,
+ "learning_rate": 0.0006,
+ "loss": 5.130426406860352,
+ "step": 3822
+ },
+ {
+ "epoch": 53.09781659388646,
+ "grad_norm": 0.02366088517010212,
+ "learning_rate": 0.0006,
+ "loss": 5.1453399658203125,
+ "step": 3823
+ },
+ {
+ "epoch": 53.1117903930131,
+ "grad_norm": 0.01813938096165657,
+ "learning_rate": 0.0006,
+ "loss": 5.05271053314209,
+ "step": 3824
+ },
+ {
+ "epoch": 53.12576419213974,
+ "grad_norm": 0.02051435597240925,
+ "learning_rate": 0.0006,
+ "loss": 5.023038864135742,
+ "step": 3825
+ },
+ {
+ "epoch": 53.13973799126637,
+ "grad_norm": 0.02185754105448723,
+ "learning_rate": 0.0006,
+ "loss": 4.8950018882751465,
+ "step": 3826
+ },
+ {
+ "epoch": 53.15371179039301,
+ "grad_norm": 0.017223743721842766,
+ "learning_rate": 0.0006,
+ "loss": 5.078621864318848,
+ "step": 3827
+ },
+ {
+ "epoch": 53.16768558951965,
+ "grad_norm": 0.01832217164337635,
+ "learning_rate": 0.0006,
+ "loss": 4.921341419219971,
+ "step": 3828
+ },
+ {
+ "epoch": 53.18165938864629,
+ "grad_norm": 0.02114005573093891,
+ "learning_rate": 0.0006,
+ "loss": 4.997865676879883,
+ "step": 3829
+ },
+ {
+ "epoch": 53.19563318777293,
+ "grad_norm": 0.021426431834697723,
+ "learning_rate": 0.0006,
+ "loss": 4.877785682678223,
+ "step": 3830
+ },
+ {
+ "epoch": 53.209606986899566,
+ "grad_norm": 0.021327560767531395,
+ "learning_rate": 0.0006,
+ "loss": 4.847512245178223,
+ "step": 3831
+ },
+ {
+ "epoch": 53.223580786026204,
+ "grad_norm": 0.027825530618429184,
+ "learning_rate": 0.0006,
+ "loss": 5.004823207855225,
+ "step": 3832
+ },
+ {
+ "epoch": 53.237554585152836,
+ "grad_norm": 0.03756481409072876,
+ "learning_rate": 0.0006,
+ "loss": 4.817290306091309,
+ "step": 3833
+ },
+ {
+ "epoch": 53.251528384279474,
+ "grad_norm": 0.04933721572160721,
+ "learning_rate": 0.0006,
+ "loss": 4.863317489624023,
+ "step": 3834
+ },
+ {
+ "epoch": 53.26550218340611,
+ "grad_norm": 0.04588694870471954,
+ "learning_rate": 0.0006,
+ "loss": 4.864742755889893,
+ "step": 3835
+ },
+ {
+ "epoch": 53.27947598253275,
+ "grad_norm": 0.02728264033794403,
+ "learning_rate": 0.0006,
+ "loss": 4.9068756103515625,
+ "step": 3836
+ },
+ {
+ "epoch": 53.29344978165939,
+ "grad_norm": 0.036240726709365845,
+ "learning_rate": 0.0006,
+ "loss": 4.902045249938965,
+ "step": 3837
+ },
+ {
+ "epoch": 53.30742358078603,
+ "grad_norm": 0.02703195810317993,
+ "learning_rate": 0.0006,
+ "loss": 4.908980369567871,
+ "step": 3838
+ },
+ {
+ "epoch": 53.32139737991266,
+ "grad_norm": 0.024894049391150475,
+ "learning_rate": 0.0006,
+ "loss": 4.794848442077637,
+ "step": 3839
+ },
+ {
+ "epoch": 53.3353711790393,
+ "grad_norm": 0.02389993891119957,
+ "learning_rate": 0.0006,
+ "loss": 4.8234734535217285,
+ "step": 3840
+ },
+ {
+ "epoch": 53.34934497816594,
+ "grad_norm": 0.02486400306224823,
+ "learning_rate": 0.0006,
+ "loss": 4.851463794708252,
+ "step": 3841
+ },
+ {
+ "epoch": 53.36331877729258,
+ "grad_norm": 0.02613813802599907,
+ "learning_rate": 0.0006,
+ "loss": 4.860444068908691,
+ "step": 3842
+ },
+ {
+ "epoch": 53.377292576419215,
+ "grad_norm": 0.0229744054377079,
+ "learning_rate": 0.0006,
+ "loss": 4.709606170654297,
+ "step": 3843
+ },
+ {
+ "epoch": 53.391266375545854,
+ "grad_norm": 0.02123042568564415,
+ "learning_rate": 0.0006,
+ "loss": 4.654613494873047,
+ "step": 3844
+ },
+ {
+ "epoch": 53.40524017467249,
+ "grad_norm": 0.017352338880300522,
+ "learning_rate": 0.0006,
+ "loss": 4.769761085510254,
+ "step": 3845
+ },
+ {
+ "epoch": 53.419213973799124,
+ "grad_norm": 0.019808724522590637,
+ "learning_rate": 0.0006,
+ "loss": 4.98514461517334,
+ "step": 3846
+ },
+ {
+ "epoch": 53.43318777292576,
+ "grad_norm": 0.018776152282953262,
+ "learning_rate": 0.0006,
+ "loss": 4.749990463256836,
+ "step": 3847
+ },
+ {
+ "epoch": 53.4471615720524,
+ "grad_norm": 0.02017521858215332,
+ "learning_rate": 0.0006,
+ "loss": 4.757381439208984,
+ "step": 3848
+ },
+ {
+ "epoch": 53.46113537117904,
+ "grad_norm": 0.019509416073560715,
+ "learning_rate": 0.0006,
+ "loss": 4.8111572265625,
+ "step": 3849
+ },
+ {
+ "epoch": 53.47510917030568,
+ "grad_norm": 0.015543580986559391,
+ "learning_rate": 0.0006,
+ "loss": 4.7501678466796875,
+ "step": 3850
+ },
+ {
+ "epoch": 53.48908296943232,
+ "grad_norm": 0.017941651865839958,
+ "learning_rate": 0.0006,
+ "loss": 4.701838493347168,
+ "step": 3851
+ },
+ {
+ "epoch": 53.50305676855895,
+ "grad_norm": 0.01678212359547615,
+ "learning_rate": 0.0006,
+ "loss": 4.787798881530762,
+ "step": 3852
+ },
+ {
+ "epoch": 53.51703056768559,
+ "grad_norm": 0.019050654023885727,
+ "learning_rate": 0.0006,
+ "loss": 4.754859924316406,
+ "step": 3853
+ },
+ {
+ "epoch": 53.531004366812226,
+ "grad_norm": 0.017406078055500984,
+ "learning_rate": 0.0006,
+ "loss": 4.752772808074951,
+ "step": 3854
+ },
+ {
+ "epoch": 53.544978165938865,
+ "grad_norm": 0.01773344725370407,
+ "learning_rate": 0.0006,
+ "loss": 4.651812553405762,
+ "step": 3855
+ },
+ {
+ "epoch": 53.5589519650655,
+ "grad_norm": 0.017221346497535706,
+ "learning_rate": 0.0006,
+ "loss": 4.792244911193848,
+ "step": 3856
+ },
+ {
+ "epoch": 53.57292576419214,
+ "grad_norm": 0.01722758449614048,
+ "learning_rate": 0.0006,
+ "loss": 4.7902960777282715,
+ "step": 3857
+ },
+ {
+ "epoch": 53.58689956331878,
+ "grad_norm": 0.020445192232728004,
+ "learning_rate": 0.0006,
+ "loss": 4.67392110824585,
+ "step": 3858
+ },
+ {
+ "epoch": 53.60087336244541,
+ "grad_norm": 0.02267114631831646,
+ "learning_rate": 0.0006,
+ "loss": 4.796089172363281,
+ "step": 3859
+ },
+ {
+ "epoch": 53.61484716157205,
+ "grad_norm": 0.027129942551255226,
+ "learning_rate": 0.0006,
+ "loss": 4.774303913116455,
+ "step": 3860
+ },
+ {
+ "epoch": 53.62882096069869,
+ "grad_norm": 0.025454839691519737,
+ "learning_rate": 0.0006,
+ "loss": 4.708813190460205,
+ "step": 3861
+ },
+ {
+ "epoch": 53.64279475982533,
+ "grad_norm": 0.019248757511377335,
+ "learning_rate": 0.0006,
+ "loss": 4.760467529296875,
+ "step": 3862
+ },
+ {
+ "epoch": 53.65676855895197,
+ "grad_norm": 0.01791176199913025,
+ "learning_rate": 0.0006,
+ "loss": 4.789203643798828,
+ "step": 3863
+ },
+ {
+ "epoch": 53.670742358078606,
+ "grad_norm": 0.016306612640619278,
+ "learning_rate": 0.0006,
+ "loss": 4.732605934143066,
+ "step": 3864
+ },
+ {
+ "epoch": 53.68471615720524,
+ "grad_norm": 0.017836976796388626,
+ "learning_rate": 0.0006,
+ "loss": 4.691871643066406,
+ "step": 3865
+ },
+ {
+ "epoch": 53.698689956331876,
+ "grad_norm": 0.015102861449122429,
+ "learning_rate": 0.0006,
+ "loss": 4.724156856536865,
+ "step": 3866
+ },
+ {
+ "epoch": 53.712663755458514,
+ "grad_norm": 0.017458142712712288,
+ "learning_rate": 0.0006,
+ "loss": 4.73524808883667,
+ "step": 3867
+ },
+ {
+ "epoch": 53.72663755458515,
+ "grad_norm": 0.01591932773590088,
+ "learning_rate": 0.0006,
+ "loss": 4.705888748168945,
+ "step": 3868
+ },
+ {
+ "epoch": 53.74061135371179,
+ "grad_norm": 0.01713593676686287,
+ "learning_rate": 0.0006,
+ "loss": 4.644516944885254,
+ "step": 3869
+ },
+ {
+ "epoch": 53.75458515283843,
+ "grad_norm": 0.01905885338783264,
+ "learning_rate": 0.0006,
+ "loss": 4.698416233062744,
+ "step": 3870
+ },
+ {
+ "epoch": 53.76855895196506,
+ "grad_norm": 0.01766437292098999,
+ "learning_rate": 0.0006,
+ "loss": 4.6796555519104,
+ "step": 3871
+ },
+ {
+ "epoch": 53.7825327510917,
+ "grad_norm": 0.01552243810147047,
+ "learning_rate": 0.0006,
+ "loss": 4.78138542175293,
+ "step": 3872
+ },
+ {
+ "epoch": 53.79650655021834,
+ "grad_norm": 0.014191331341862679,
+ "learning_rate": 0.0006,
+ "loss": 4.703588485717773,
+ "step": 3873
+ },
+ {
+ "epoch": 53.81048034934498,
+ "grad_norm": 0.015872279182076454,
+ "learning_rate": 0.0006,
+ "loss": 4.705611228942871,
+ "step": 3874
+ },
+ {
+ "epoch": 53.82445414847162,
+ "grad_norm": 0.01787560060620308,
+ "learning_rate": 0.0006,
+ "loss": 4.8227949142456055,
+ "step": 3875
+ },
+ {
+ "epoch": 53.838427947598255,
+ "grad_norm": 0.0158209390938282,
+ "learning_rate": 0.0006,
+ "loss": 4.645227432250977,
+ "step": 3876
+ },
+ {
+ "epoch": 53.852401746724894,
+ "grad_norm": 0.014362678863108158,
+ "learning_rate": 0.0006,
+ "loss": 4.7225422859191895,
+ "step": 3877
+ },
+ {
+ "epoch": 53.866375545851525,
+ "grad_norm": 0.014899161644279957,
+ "learning_rate": 0.0006,
+ "loss": 4.685914039611816,
+ "step": 3878
+ },
+ {
+ "epoch": 53.880349344978164,
+ "grad_norm": 0.01628505438566208,
+ "learning_rate": 0.0006,
+ "loss": 4.674065589904785,
+ "step": 3879
+ },
+ {
+ "epoch": 53.8943231441048,
+ "grad_norm": 0.01873331144452095,
+ "learning_rate": 0.0006,
+ "loss": 4.811744213104248,
+ "step": 3880
+ },
+ {
+ "epoch": 53.90829694323144,
+ "grad_norm": 0.02229069173336029,
+ "learning_rate": 0.0006,
+ "loss": 4.721514701843262,
+ "step": 3881
+ },
+ {
+ "epoch": 53.92227074235808,
+ "grad_norm": 0.028012916445732117,
+ "learning_rate": 0.0006,
+ "loss": 4.773603439331055,
+ "step": 3882
+ },
+ {
+ "epoch": 53.93624454148472,
+ "grad_norm": 0.036322783678770065,
+ "learning_rate": 0.0006,
+ "loss": 4.787134170532227,
+ "step": 3883
+ },
+ {
+ "epoch": 53.95021834061135,
+ "grad_norm": 0.048149142414331436,
+ "learning_rate": 0.0006,
+ "loss": 4.736196517944336,
+ "step": 3884
+ },
+ {
+ "epoch": 53.96419213973799,
+ "grad_norm": 0.051918916404247284,
+ "learning_rate": 0.0006,
+ "loss": 4.686728000640869,
+ "step": 3885
+ },
+ {
+ "epoch": 53.97816593886463,
+ "grad_norm": 0.0481342189013958,
+ "learning_rate": 0.0006,
+ "loss": 4.733730792999268,
+ "step": 3886
+ },
+ {
+ "epoch": 53.992139737991266,
+ "grad_norm": 0.045424818992614746,
+ "learning_rate": 0.0006,
+ "loss": 4.748723983764648,
+ "step": 3887
+ },
+ {
+ "epoch": 54.0,
+ "grad_norm": 0.04153316840529442,
+ "learning_rate": 0.0006,
+ "loss": 4.715220928192139,
+ "step": 3888
+ },
+ {
+ "epoch": 54.0,
+ "eval_loss": 4.964215278625488,
+ "eval_runtime": 56.7757,
+ "eval_samples_per_second": 43.011,
+ "eval_steps_per_second": 1.356,
+ "step": 3888
+ },
+ {
+ "epoch": 54.01397379912664,
+ "grad_norm": 0.03459750860929489,
+ "learning_rate": 0.0006,
+ "loss": 4.661557674407959,
+ "step": 3889
+ },
+ {
+ "epoch": 54.02794759825328,
+ "grad_norm": 0.025409208610653877,
+ "learning_rate": 0.0006,
+ "loss": 4.7065582275390625,
+ "step": 3890
+ },
+ {
+ "epoch": 54.041921397379916,
+ "grad_norm": 0.03202550485730171,
+ "learning_rate": 0.0006,
+ "loss": 4.751070976257324,
+ "step": 3891
+ },
+ {
+ "epoch": 54.05589519650655,
+ "grad_norm": 0.02968018501996994,
+ "learning_rate": 0.0006,
+ "loss": 4.635388374328613,
+ "step": 3892
+ },
+ {
+ "epoch": 54.069868995633186,
+ "grad_norm": 0.02757406234741211,
+ "learning_rate": 0.0006,
+ "loss": 4.706214427947998,
+ "step": 3893
+ },
+ {
+ "epoch": 54.083842794759825,
+ "grad_norm": 0.025223620235919952,
+ "learning_rate": 0.0006,
+ "loss": 4.703792572021484,
+ "step": 3894
+ },
+ {
+ "epoch": 54.09781659388646,
+ "grad_norm": 0.02431778982281685,
+ "learning_rate": 0.0006,
+ "loss": 4.658116340637207,
+ "step": 3895
+ },
+ {
+ "epoch": 54.1117903930131,
+ "grad_norm": 0.02792172320187092,
+ "learning_rate": 0.0006,
+ "loss": 4.638895034790039,
+ "step": 3896
+ },
+ {
+ "epoch": 54.12576419213974,
+ "grad_norm": 0.032253511250019073,
+ "learning_rate": 0.0006,
+ "loss": 4.693902492523193,
+ "step": 3897
+ },
+ {
+ "epoch": 54.13973799126637,
+ "grad_norm": 0.027044931426644325,
+ "learning_rate": 0.0006,
+ "loss": 4.575805187225342,
+ "step": 3898
+ },
+ {
+ "epoch": 54.15371179039301,
+ "grad_norm": 0.023270361125469208,
+ "learning_rate": 0.0006,
+ "loss": 4.67645263671875,
+ "step": 3899
+ },
+ {
+ "epoch": 54.16768558951965,
+ "grad_norm": 0.01929079182446003,
+ "learning_rate": 0.0006,
+ "loss": 4.68062686920166,
+ "step": 3900
+ },
+ {
+ "epoch": 54.18165938864629,
+ "grad_norm": 0.020343473181128502,
+ "learning_rate": 0.0006,
+ "loss": 4.708086967468262,
+ "step": 3901
+ },
+ {
+ "epoch": 54.19563318777293,
+ "grad_norm": 0.0211793202906847,
+ "learning_rate": 0.0006,
+ "loss": 4.677323341369629,
+ "step": 3902
+ },
+ {
+ "epoch": 54.209606986899566,
+ "grad_norm": 0.020341217517852783,
+ "learning_rate": 0.0006,
+ "loss": 4.760937690734863,
+ "step": 3903
+ },
+ {
+ "epoch": 54.223580786026204,
+ "grad_norm": 0.019421054050326347,
+ "learning_rate": 0.0006,
+ "loss": 4.508834362030029,
+ "step": 3904
+ },
+ {
+ "epoch": 54.237554585152836,
+ "grad_norm": 0.016701679676771164,
+ "learning_rate": 0.0006,
+ "loss": 4.664399147033691,
+ "step": 3905
+ },
+ {
+ "epoch": 54.251528384279474,
+ "grad_norm": 0.016561470925807953,
+ "learning_rate": 0.0006,
+ "loss": 4.742990493774414,
+ "step": 3906
+ },
+ {
+ "epoch": 54.26550218340611,
+ "grad_norm": 0.017446085810661316,
+ "learning_rate": 0.0006,
+ "loss": 4.63038444519043,
+ "step": 3907
+ },
+ {
+ "epoch": 54.27947598253275,
+ "grad_norm": 0.014340086840093136,
+ "learning_rate": 0.0006,
+ "loss": 4.688791275024414,
+ "step": 3908
+ },
+ {
+ "epoch": 54.29344978165939,
+ "grad_norm": 0.013580698519945145,
+ "learning_rate": 0.0006,
+ "loss": 4.625302314758301,
+ "step": 3909
+ },
+ {
+ "epoch": 54.30742358078603,
+ "grad_norm": 0.014931539073586464,
+ "learning_rate": 0.0006,
+ "loss": 4.6167497634887695,
+ "step": 3910
+ },
+ {
+ "epoch": 54.32139737991266,
+ "grad_norm": 0.013492150232195854,
+ "learning_rate": 0.0006,
+ "loss": 4.657521724700928,
+ "step": 3911
+ },
+ {
+ "epoch": 54.3353711790393,
+ "grad_norm": 0.015151413157582283,
+ "learning_rate": 0.0006,
+ "loss": 4.634521484375,
+ "step": 3912
+ },
+ {
+ "epoch": 54.34934497816594,
+ "grad_norm": 0.01302040833979845,
+ "learning_rate": 0.0006,
+ "loss": 4.718928337097168,
+ "step": 3913
+ },
+ {
+ "epoch": 54.36331877729258,
+ "grad_norm": 0.013433529995381832,
+ "learning_rate": 0.0006,
+ "loss": 4.584803581237793,
+ "step": 3914
+ },
+ {
+ "epoch": 54.377292576419215,
+ "grad_norm": 0.012987039051949978,
+ "learning_rate": 0.0006,
+ "loss": 4.6393585205078125,
+ "step": 3915
+ },
+ {
+ "epoch": 54.391266375545854,
+ "grad_norm": 0.013695075176656246,
+ "learning_rate": 0.0006,
+ "loss": 4.556313514709473,
+ "step": 3916
+ },
+ {
+ "epoch": 54.40524017467249,
+ "grad_norm": 0.013718255795538425,
+ "learning_rate": 0.0006,
+ "loss": 4.576223850250244,
+ "step": 3917
+ },
+ {
+ "epoch": 54.419213973799124,
+ "grad_norm": 0.013699112460017204,
+ "learning_rate": 0.0006,
+ "loss": 4.63889217376709,
+ "step": 3918
+ },
+ {
+ "epoch": 54.43318777292576,
+ "grad_norm": 0.013516264036297798,
+ "learning_rate": 0.0006,
+ "loss": 4.669532299041748,
+ "step": 3919
+ },
+ {
+ "epoch": 54.4471615720524,
+ "grad_norm": 0.013597294688224792,
+ "learning_rate": 0.0006,
+ "loss": 4.6644287109375,
+ "step": 3920
+ },
+ {
+ "epoch": 54.46113537117904,
+ "grad_norm": 0.013185849413275719,
+ "learning_rate": 0.0006,
+ "loss": 4.636900901794434,
+ "step": 3921
+ },
+ {
+ "epoch": 54.47510917030568,
+ "grad_norm": 0.014311990700662136,
+ "learning_rate": 0.0006,
+ "loss": 4.6047587394714355,
+ "step": 3922
+ },
+ {
+ "epoch": 54.48908296943232,
+ "grad_norm": 0.012426027096807957,
+ "learning_rate": 0.0006,
+ "loss": 4.6161394119262695,
+ "step": 3923
+ },
+ {
+ "epoch": 54.50305676855895,
+ "grad_norm": 0.01283493172377348,
+ "learning_rate": 0.0006,
+ "loss": 4.584190368652344,
+ "step": 3924
+ },
+ {
+ "epoch": 54.51703056768559,
+ "grad_norm": 0.01339943427592516,
+ "learning_rate": 0.0006,
+ "loss": 4.624184608459473,
+ "step": 3925
+ },
+ {
+ "epoch": 54.531004366812226,
+ "grad_norm": 0.012859724462032318,
+ "learning_rate": 0.0006,
+ "loss": 4.6174635887146,
+ "step": 3926
+ },
+ {
+ "epoch": 54.544978165938865,
+ "grad_norm": 0.014420023187994957,
+ "learning_rate": 0.0006,
+ "loss": 4.536544322967529,
+ "step": 3927
+ },
+ {
+ "epoch": 54.5589519650655,
+ "grad_norm": 0.014237185008823872,
+ "learning_rate": 0.0006,
+ "loss": 4.681514739990234,
+ "step": 3928
+ },
+ {
+ "epoch": 54.57292576419214,
+ "grad_norm": 0.013340328820049763,
+ "learning_rate": 0.0006,
+ "loss": 4.579689979553223,
+ "step": 3929
+ },
+ {
+ "epoch": 54.58689956331878,
+ "grad_norm": 0.013894530944526196,
+ "learning_rate": 0.0006,
+ "loss": 4.638958930969238,
+ "step": 3930
+ },
+ {
+ "epoch": 54.60087336244541,
+ "grad_norm": 0.016181141138076782,
+ "learning_rate": 0.0006,
+ "loss": 4.573508262634277,
+ "step": 3931
+ },
+ {
+ "epoch": 54.61484716157205,
+ "grad_norm": 0.013221992179751396,
+ "learning_rate": 0.0006,
+ "loss": 4.6229047775268555,
+ "step": 3932
+ },
+ {
+ "epoch": 54.62882096069869,
+ "grad_norm": 0.013241561129689217,
+ "learning_rate": 0.0006,
+ "loss": 4.641152858734131,
+ "step": 3933
+ },
+ {
+ "epoch": 54.64279475982533,
+ "grad_norm": 0.016237538307905197,
+ "learning_rate": 0.0006,
+ "loss": 4.6421217918396,
+ "step": 3934
+ },
+ {
+ "epoch": 54.65676855895197,
+ "grad_norm": 0.020866507664322853,
+ "learning_rate": 0.0006,
+ "loss": 4.646058082580566,
+ "step": 3935
+ },
+ {
+ "epoch": 54.670742358078606,
+ "grad_norm": 0.022182893007993698,
+ "learning_rate": 0.0006,
+ "loss": 4.554501533508301,
+ "step": 3936
+ },
+ {
+ "epoch": 54.68471615720524,
+ "grad_norm": 0.02254868485033512,
+ "learning_rate": 0.0006,
+ "loss": 4.563869953155518,
+ "step": 3937
+ },
+ {
+ "epoch": 54.698689956331876,
+ "grad_norm": 0.019730940461158752,
+ "learning_rate": 0.0006,
+ "loss": 4.592165946960449,
+ "step": 3938
+ },
+ {
+ "epoch": 54.712663755458514,
+ "grad_norm": 0.015146561898291111,
+ "learning_rate": 0.0006,
+ "loss": 4.6228742599487305,
+ "step": 3939
+ },
+ {
+ "epoch": 54.72663755458515,
+ "grad_norm": 0.01824716106057167,
+ "learning_rate": 0.0006,
+ "loss": 4.716236114501953,
+ "step": 3940
+ },
+ {
+ "epoch": 54.74061135371179,
+ "grad_norm": 0.017730550840497017,
+ "learning_rate": 0.0006,
+ "loss": 4.650232315063477,
+ "step": 3941
+ },
+ {
+ "epoch": 54.75458515283843,
+ "grad_norm": 0.014875595457851887,
+ "learning_rate": 0.0006,
+ "loss": 4.6377458572387695,
+ "step": 3942
+ },
+ {
+ "epoch": 54.76855895196506,
+ "grad_norm": 0.01613209955394268,
+ "learning_rate": 0.0006,
+ "loss": 4.5768537521362305,
+ "step": 3943
+ },
+ {
+ "epoch": 54.7825327510917,
+ "grad_norm": 0.017311280593276024,
+ "learning_rate": 0.0006,
+ "loss": 4.625744342803955,
+ "step": 3944
+ },
+ {
+ "epoch": 54.79650655021834,
+ "grad_norm": 0.018770085647702217,
+ "learning_rate": 0.0006,
+ "loss": 4.613787651062012,
+ "step": 3945
+ },
+ {
+ "epoch": 54.81048034934498,
+ "grad_norm": 0.0165257528424263,
+ "learning_rate": 0.0006,
+ "loss": 4.5991315841674805,
+ "step": 3946
+ },
+ {
+ "epoch": 54.82445414847162,
+ "grad_norm": 0.017433857545256615,
+ "learning_rate": 0.0006,
+ "loss": 4.513872146606445,
+ "step": 3947
+ },
+ {
+ "epoch": 54.838427947598255,
+ "grad_norm": 0.020126905292272568,
+ "learning_rate": 0.0006,
+ "loss": 4.616185188293457,
+ "step": 3948
+ },
+ {
+ "epoch": 54.852401746724894,
+ "grad_norm": 0.01819358579814434,
+ "learning_rate": 0.0006,
+ "loss": 4.535691261291504,
+ "step": 3949
+ },
+ {
+ "epoch": 54.866375545851525,
+ "grad_norm": 0.01899927295744419,
+ "learning_rate": 0.0006,
+ "loss": 4.595199108123779,
+ "step": 3950
+ },
+ {
+ "epoch": 54.880349344978164,
+ "grad_norm": 0.02005566656589508,
+ "learning_rate": 0.0006,
+ "loss": 4.633289337158203,
+ "step": 3951
+ },
+ {
+ "epoch": 54.8943231441048,
+ "grad_norm": 0.021570220589637756,
+ "learning_rate": 0.0006,
+ "loss": 4.592503547668457,
+ "step": 3952
+ },
+ {
+ "epoch": 54.90829694323144,
+ "grad_norm": 0.01883506216108799,
+ "learning_rate": 0.0006,
+ "loss": 4.6597795486450195,
+ "step": 3953
+ },
+ {
+ "epoch": 54.92227074235808,
+ "grad_norm": 0.018070021644234657,
+ "learning_rate": 0.0006,
+ "loss": 4.601474285125732,
+ "step": 3954
+ },
+ {
+ "epoch": 54.93624454148472,
+ "grad_norm": 0.018090378493070602,
+ "learning_rate": 0.0006,
+ "loss": 4.569846153259277,
+ "step": 3955
+ },
+ {
+ "epoch": 54.95021834061135,
+ "grad_norm": 0.018751563504338264,
+ "learning_rate": 0.0006,
+ "loss": 4.646064281463623,
+ "step": 3956
+ },
+ {
+ "epoch": 54.96419213973799,
+ "grad_norm": 0.017512254416942596,
+ "learning_rate": 0.0006,
+ "loss": 4.606670379638672,
+ "step": 3957
+ },
+ {
+ "epoch": 54.97816593886463,
+ "grad_norm": 0.018771110102534294,
+ "learning_rate": 0.0006,
+ "loss": 4.604315757751465,
+ "step": 3958
+ },
+ {
+ "epoch": 54.992139737991266,
+ "grad_norm": 0.01611141301691532,
+ "learning_rate": 0.0006,
+ "loss": 4.643245220184326,
+ "step": 3959
+ },
+ {
+ "epoch": 55.0,
+ "grad_norm": 0.0156728383153677,
+ "learning_rate": 0.0006,
+ "loss": 4.6739044189453125,
+ "step": 3960
+ },
+ {
+ "epoch": 55.0,
+ "eval_loss": 4.851032257080078,
+ "eval_runtime": 56.8699,
+ "eval_samples_per_second": 42.94,
+ "eval_steps_per_second": 1.354,
+ "step": 3960
+ },
+ {
+ "epoch": 55.01397379912664,
+ "grad_norm": 0.016214672476053238,
+ "learning_rate": 0.0006,
+ "loss": 4.571749687194824,
+ "step": 3961
+ },
+ {
+ "epoch": 55.02794759825328,
+ "grad_norm": 0.01649155467748642,
+ "learning_rate": 0.0006,
+ "loss": 4.553592205047607,
+ "step": 3962
+ },
+ {
+ "epoch": 55.041921397379916,
+ "grad_norm": 0.016793379560112953,
+ "learning_rate": 0.0006,
+ "loss": 4.650628566741943,
+ "step": 3963
+ },
+ {
+ "epoch": 55.05589519650655,
+ "grad_norm": 0.0191937405616045,
+ "learning_rate": 0.0006,
+ "loss": 4.620532989501953,
+ "step": 3964
+ },
+ {
+ "epoch": 55.069868995633186,
+ "grad_norm": 0.020976556465029716,
+ "learning_rate": 0.0006,
+ "loss": 4.6329545974731445,
+ "step": 3965
+ },
+ {
+ "epoch": 55.083842794759825,
+ "grad_norm": 0.023450477048754692,
+ "learning_rate": 0.0006,
+ "loss": 4.60828971862793,
+ "step": 3966
+ },
+ {
+ "epoch": 55.09781659388646,
+ "grad_norm": 0.026778312399983406,
+ "learning_rate": 0.0006,
+ "loss": 4.617481708526611,
+ "step": 3967
+ },
+ {
+ "epoch": 55.1117903930131,
+ "grad_norm": 0.02333568036556244,
+ "learning_rate": 0.0006,
+ "loss": 4.580722332000732,
+ "step": 3968
+ },
+ {
+ "epoch": 55.12576419213974,
+ "grad_norm": 0.018491310998797417,
+ "learning_rate": 0.0006,
+ "loss": 4.590953350067139,
+ "step": 3969
+ },
+ {
+ "epoch": 55.13973799126637,
+ "grad_norm": 0.020339803770184517,
+ "learning_rate": 0.0006,
+ "loss": 4.518374443054199,
+ "step": 3970
+ },
+ {
+ "epoch": 55.15371179039301,
+ "grad_norm": 0.01759706623852253,
+ "learning_rate": 0.0006,
+ "loss": 4.560522079467773,
+ "step": 3971
+ },
+ {
+ "epoch": 55.16768558951965,
+ "grad_norm": 0.015875980257987976,
+ "learning_rate": 0.0006,
+ "loss": 4.4795708656311035,
+ "step": 3972
+ },
+ {
+ "epoch": 55.18165938864629,
+ "grad_norm": 0.01650134287774563,
+ "learning_rate": 0.0006,
+ "loss": 4.57435941696167,
+ "step": 3973
+ },
+ {
+ "epoch": 55.19563318777293,
+ "grad_norm": 0.01734331250190735,
+ "learning_rate": 0.0006,
+ "loss": 4.661007881164551,
+ "step": 3974
+ },
+ {
+ "epoch": 55.209606986899566,
+ "grad_norm": 0.01651303470134735,
+ "learning_rate": 0.0006,
+ "loss": 4.511148452758789,
+ "step": 3975
+ },
+ {
+ "epoch": 55.223580786026204,
+ "grad_norm": 0.016189761459827423,
+ "learning_rate": 0.0006,
+ "loss": 4.592350006103516,
+ "step": 3976
+ },
+ {
+ "epoch": 55.237554585152836,
+ "grad_norm": 0.01833222061395645,
+ "learning_rate": 0.0006,
+ "loss": 4.506649971008301,
+ "step": 3977
+ },
+ {
+ "epoch": 55.251528384279474,
+ "grad_norm": 0.02433410845696926,
+ "learning_rate": 0.0006,
+ "loss": 4.49500036239624,
+ "step": 3978
+ },
+ {
+ "epoch": 55.26550218340611,
+ "grad_norm": 0.0314289927482605,
+ "learning_rate": 0.0006,
+ "loss": 4.507624626159668,
+ "step": 3979
+ },
+ {
+ "epoch": 55.27947598253275,
+ "grad_norm": 0.03013679012656212,
+ "learning_rate": 0.0006,
+ "loss": 4.533641815185547,
+ "step": 3980
+ },
+ {
+ "epoch": 55.29344978165939,
+ "grad_norm": 0.024524061009287834,
+ "learning_rate": 0.0006,
+ "loss": 4.5721116065979,
+ "step": 3981
+ },
+ {
+ "epoch": 55.30742358078603,
+ "grad_norm": 0.02640336938202381,
+ "learning_rate": 0.0006,
+ "loss": 4.652521133422852,
+ "step": 3982
+ },
+ {
+ "epoch": 55.32139737991266,
+ "grad_norm": 0.021309101954102516,
+ "learning_rate": 0.0006,
+ "loss": 4.600375175476074,
+ "step": 3983
+ },
+ {
+ "epoch": 55.3353711790393,
+ "grad_norm": 0.023008549585938454,
+ "learning_rate": 0.0006,
+ "loss": 4.53914737701416,
+ "step": 3984
+ },
+ {
+ "epoch": 55.34934497816594,
+ "grad_norm": 0.022491468116641045,
+ "learning_rate": 0.0006,
+ "loss": 4.661163330078125,
+ "step": 3985
+ },
+ {
+ "epoch": 55.36331877729258,
+ "grad_norm": 0.01904081553220749,
+ "learning_rate": 0.0006,
+ "loss": 4.5426177978515625,
+ "step": 3986
+ },
+ {
+ "epoch": 55.377292576419215,
+ "grad_norm": 0.019229518249630928,
+ "learning_rate": 0.0006,
+ "loss": 4.606066703796387,
+ "step": 3987
+ },
+ {
+ "epoch": 55.391266375545854,
+ "grad_norm": 0.01781383901834488,
+ "learning_rate": 0.0006,
+ "loss": 4.54642391204834,
+ "step": 3988
+ },
+ {
+ "epoch": 55.40524017467249,
+ "grad_norm": 0.0174600500613451,
+ "learning_rate": 0.0006,
+ "loss": 4.711891174316406,
+ "step": 3989
+ },
+ {
+ "epoch": 55.419213973799124,
+ "grad_norm": 0.018154339864850044,
+ "learning_rate": 0.0006,
+ "loss": 4.527122974395752,
+ "step": 3990
+ },
+ {
+ "epoch": 55.43318777292576,
+ "grad_norm": 0.01914648339152336,
+ "learning_rate": 0.0006,
+ "loss": 4.641975402832031,
+ "step": 3991
+ },
+ {
+ "epoch": 55.4471615720524,
+ "grad_norm": 0.017118625342845917,
+ "learning_rate": 0.0006,
+ "loss": 4.511425971984863,
+ "step": 3992
+ },
+ {
+ "epoch": 55.46113537117904,
+ "grad_norm": 0.01551489345729351,
+ "learning_rate": 0.0006,
+ "loss": 4.551141738891602,
+ "step": 3993
+ },
+ {
+ "epoch": 55.47510917030568,
+ "grad_norm": 0.015072065405547619,
+ "learning_rate": 0.0006,
+ "loss": 4.651115894317627,
+ "step": 3994
+ },
+ {
+ "epoch": 55.48908296943232,
+ "grad_norm": 0.017010986804962158,
+ "learning_rate": 0.0006,
+ "loss": 4.583282470703125,
+ "step": 3995
+ },
+ {
+ "epoch": 55.50305676855895,
+ "grad_norm": 0.01764831133186817,
+ "learning_rate": 0.0006,
+ "loss": 4.593364238739014,
+ "step": 3996
+ },
+ {
+ "epoch": 55.51703056768559,
+ "grad_norm": 0.01762884296476841,
+ "learning_rate": 0.0006,
+ "loss": 4.45671272277832,
+ "step": 3997
+ },
+ {
+ "epoch": 55.531004366812226,
+ "grad_norm": 0.01600457727909088,
+ "learning_rate": 0.0006,
+ "loss": 4.511260032653809,
+ "step": 3998
+ },
+ {
+ "epoch": 55.544978165938865,
+ "grad_norm": 0.015706021338701248,
+ "learning_rate": 0.0006,
+ "loss": 4.47593355178833,
+ "step": 3999
+ },
+ {
+ "epoch": 55.5589519650655,
+ "grad_norm": 0.017718996852636337,
+ "learning_rate": 0.0006,
+ "loss": 4.610543251037598,
+ "step": 4000
+ },
+ {
+ "epoch": 55.57292576419214,
+ "grad_norm": 0.015073757618665695,
+ "learning_rate": 0.0006,
+ "loss": 4.466620445251465,
+ "step": 4001
+ },
+ {
+ "epoch": 55.58689956331878,
+ "grad_norm": 0.01789247989654541,
+ "learning_rate": 0.0006,
+ "loss": 4.503409385681152,
+ "step": 4002
+ },
+ {
+ "epoch": 55.60087336244541,
+ "grad_norm": 0.022329598665237427,
+ "learning_rate": 0.0006,
+ "loss": 4.62099552154541,
+ "step": 4003
+ },
+ {
+ "epoch": 55.61484716157205,
+ "grad_norm": 0.02343829534947872,
+ "learning_rate": 0.0006,
+ "loss": 4.620813369750977,
+ "step": 4004
+ },
+ {
+ "epoch": 55.62882096069869,
+ "grad_norm": 0.023279687389731407,
+ "learning_rate": 0.0006,
+ "loss": 4.545765399932861,
+ "step": 4005
+ },
+ {
+ "epoch": 55.64279475982533,
+ "grad_norm": 0.020817464217543602,
+ "learning_rate": 0.0006,
+ "loss": 4.561470031738281,
+ "step": 4006
+ },
+ {
+ "epoch": 55.65676855895197,
+ "grad_norm": 0.020270058885216713,
+ "learning_rate": 0.0006,
+ "loss": 4.5875020027160645,
+ "step": 4007
+ },
+ {
+ "epoch": 55.670742358078606,
+ "grad_norm": 0.021085843443870544,
+ "learning_rate": 0.0006,
+ "loss": 4.504560470581055,
+ "step": 4008
+ },
+ {
+ "epoch": 55.68471615720524,
+ "grad_norm": 0.018507730215787888,
+ "learning_rate": 0.0006,
+ "loss": 4.520644664764404,
+ "step": 4009
+ },
+ {
+ "epoch": 55.698689956331876,
+ "grad_norm": 0.020571516826748848,
+ "learning_rate": 0.0006,
+ "loss": 4.539643287658691,
+ "step": 4010
+ },
+ {
+ "epoch": 55.712663755458514,
+ "grad_norm": 0.021484510973095894,
+ "learning_rate": 0.0006,
+ "loss": 4.628833293914795,
+ "step": 4011
+ },
+ {
+ "epoch": 55.72663755458515,
+ "grad_norm": 0.017693819478154182,
+ "learning_rate": 0.0006,
+ "loss": 4.6023783683776855,
+ "step": 4012
+ },
+ {
+ "epoch": 55.74061135371179,
+ "grad_norm": 0.020186059176921844,
+ "learning_rate": 0.0006,
+ "loss": 4.634884834289551,
+ "step": 4013
+ },
+ {
+ "epoch": 55.75458515283843,
+ "grad_norm": 0.01835780404508114,
+ "learning_rate": 0.0006,
+ "loss": 4.577791213989258,
+ "step": 4014
+ },
+ {
+ "epoch": 55.76855895196506,
+ "grad_norm": 0.018162988126277924,
+ "learning_rate": 0.0006,
+ "loss": 4.641597270965576,
+ "step": 4015
+ },
+ {
+ "epoch": 55.7825327510917,
+ "grad_norm": 0.018909504637122154,
+ "learning_rate": 0.0006,
+ "loss": 4.487447738647461,
+ "step": 4016
+ },
+ {
+ "epoch": 55.79650655021834,
+ "grad_norm": 0.018972251564264297,
+ "learning_rate": 0.0006,
+ "loss": 4.590317726135254,
+ "step": 4017
+ },
+ {
+ "epoch": 55.81048034934498,
+ "grad_norm": 0.01793607883155346,
+ "learning_rate": 0.0006,
+ "loss": 4.4994611740112305,
+ "step": 4018
+ },
+ {
+ "epoch": 55.82445414847162,
+ "grad_norm": 0.02123514749109745,
+ "learning_rate": 0.0006,
+ "loss": 4.479006767272949,
+ "step": 4019
+ },
+ {
+ "epoch": 55.838427947598255,
+ "grad_norm": 0.021246599033474922,
+ "learning_rate": 0.0006,
+ "loss": 4.4738640785217285,
+ "step": 4020
+ },
+ {
+ "epoch": 55.852401746724894,
+ "grad_norm": 0.020131602883338928,
+ "learning_rate": 0.0006,
+ "loss": 4.600308418273926,
+ "step": 4021
+ },
+ {
+ "epoch": 55.866375545851525,
+ "grad_norm": 0.019159460440278053,
+ "learning_rate": 0.0006,
+ "loss": 4.540908336639404,
+ "step": 4022
+ },
+ {
+ "epoch": 55.880349344978164,
+ "grad_norm": 0.023114705458283424,
+ "learning_rate": 0.0006,
+ "loss": 4.641247749328613,
+ "step": 4023
+ },
+ {
+ "epoch": 55.8943231441048,
+ "grad_norm": 0.02296135015785694,
+ "learning_rate": 0.0006,
+ "loss": 4.444624423980713,
+ "step": 4024
+ },
+ {
+ "epoch": 55.90829694323144,
+ "grad_norm": 0.02204655110836029,
+ "learning_rate": 0.0006,
+ "loss": 4.506349563598633,
+ "step": 4025
+ },
+ {
+ "epoch": 55.92227074235808,
+ "grad_norm": 0.019338225945830345,
+ "learning_rate": 0.0006,
+ "loss": 4.616047382354736,
+ "step": 4026
+ },
+ {
+ "epoch": 55.93624454148472,
+ "grad_norm": 0.020357884466648102,
+ "learning_rate": 0.0006,
+ "loss": 4.48170280456543,
+ "step": 4027
+ },
+ {
+ "epoch": 55.95021834061135,
+ "grad_norm": 0.02426772564649582,
+ "learning_rate": 0.0006,
+ "loss": 4.576877117156982,
+ "step": 4028
+ },
+ {
+ "epoch": 55.96419213973799,
+ "grad_norm": 0.023351816460490227,
+ "learning_rate": 0.0006,
+ "loss": 4.458885192871094,
+ "step": 4029
+ },
+ {
+ "epoch": 55.97816593886463,
+ "grad_norm": 0.019892243668437004,
+ "learning_rate": 0.0006,
+ "loss": 4.520229816436768,
+ "step": 4030
+ },
+ {
+ "epoch": 55.992139737991266,
+ "grad_norm": 0.018331089988350868,
+ "learning_rate": 0.0006,
+ "loss": 4.432201862335205,
+ "step": 4031
+ },
+ {
+ "epoch": 56.0,
+ "grad_norm": 0.021300828084349632,
+ "learning_rate": 0.0006,
+ "loss": 4.617072582244873,
+ "step": 4032
+ },
+ {
+ "epoch": 56.0,
+ "eval_loss": 4.811229705810547,
+ "eval_runtime": 56.4846,
+ "eval_samples_per_second": 43.233,
+ "eval_steps_per_second": 1.363,
+ "step": 4032
+ },
+ {
+ "epoch": 56.01397379912664,
+ "grad_norm": 0.02199401892721653,
+ "learning_rate": 0.0006,
+ "loss": 4.513965606689453,
+ "step": 4033
+ },
+ {
+ "epoch": 56.02794759825328,
+ "grad_norm": 0.022109858691692352,
+ "learning_rate": 0.0006,
+ "loss": 4.557568550109863,
+ "step": 4034
+ },
+ {
+ "epoch": 56.041921397379916,
+ "grad_norm": 0.020103694871068,
+ "learning_rate": 0.0006,
+ "loss": 4.569512844085693,
+ "step": 4035
+ },
+ {
+ "epoch": 56.05589519650655,
+ "grad_norm": 0.01898941583931446,
+ "learning_rate": 0.0006,
+ "loss": 4.623138904571533,
+ "step": 4036
+ },
+ {
+ "epoch": 56.069868995633186,
+ "grad_norm": 0.02227010577917099,
+ "learning_rate": 0.0006,
+ "loss": 4.433316230773926,
+ "step": 4037
+ },
+ {
+ "epoch": 56.083842794759825,
+ "grad_norm": 0.024683045223355293,
+ "learning_rate": 0.0006,
+ "loss": 4.458296775817871,
+ "step": 4038
+ },
+ {
+ "epoch": 56.09781659388646,
+ "grad_norm": 0.02900184690952301,
+ "learning_rate": 0.0006,
+ "loss": 4.531060218811035,
+ "step": 4039
+ },
+ {
+ "epoch": 56.1117903930131,
+ "grad_norm": 0.025557566434144974,
+ "learning_rate": 0.0006,
+ "loss": 4.56685733795166,
+ "step": 4040
+ },
+ {
+ "epoch": 56.12576419213974,
+ "grad_norm": 0.021414656192064285,
+ "learning_rate": 0.0006,
+ "loss": 4.469037055969238,
+ "step": 4041
+ },
+ {
+ "epoch": 56.13973799126637,
+ "grad_norm": 0.023262616246938705,
+ "learning_rate": 0.0006,
+ "loss": 4.493876934051514,
+ "step": 4042
+ },
+ {
+ "epoch": 56.15371179039301,
+ "grad_norm": 0.023011326789855957,
+ "learning_rate": 0.0006,
+ "loss": 4.4572906494140625,
+ "step": 4043
+ },
+ {
+ "epoch": 56.16768558951965,
+ "grad_norm": 0.02177615463733673,
+ "learning_rate": 0.0006,
+ "loss": 4.615389823913574,
+ "step": 4044
+ },
+ {
+ "epoch": 56.18165938864629,
+ "grad_norm": 0.021527713164687157,
+ "learning_rate": 0.0006,
+ "loss": 4.613762855529785,
+ "step": 4045
+ },
+ {
+ "epoch": 56.19563318777293,
+ "grad_norm": 0.01982325315475464,
+ "learning_rate": 0.0006,
+ "loss": 4.53670597076416,
+ "step": 4046
+ },
+ {
+ "epoch": 56.209606986899566,
+ "grad_norm": 0.022484654560685158,
+ "learning_rate": 0.0006,
+ "loss": 4.605565071105957,
+ "step": 4047
+ },
+ {
+ "epoch": 56.223580786026204,
+ "grad_norm": 0.02346140146255493,
+ "learning_rate": 0.0006,
+ "loss": 4.6265950202941895,
+ "step": 4048
+ },
+ {
+ "epoch": 56.237554585152836,
+ "grad_norm": 0.01876714453101158,
+ "learning_rate": 0.0006,
+ "loss": 4.481525421142578,
+ "step": 4049
+ },
+ {
+ "epoch": 56.251528384279474,
+ "grad_norm": 0.01926851086318493,
+ "learning_rate": 0.0006,
+ "loss": 4.630258083343506,
+ "step": 4050
+ },
+ {
+ "epoch": 56.26550218340611,
+ "grad_norm": 0.020858589559793472,
+ "learning_rate": 0.0006,
+ "loss": 4.5265913009643555,
+ "step": 4051
+ },
+ {
+ "epoch": 56.27947598253275,
+ "grad_norm": 0.01889280416071415,
+ "learning_rate": 0.0006,
+ "loss": 4.581275463104248,
+ "step": 4052
+ },
+ {
+ "epoch": 56.29344978165939,
+ "grad_norm": 0.01801002025604248,
+ "learning_rate": 0.0006,
+ "loss": 4.498557090759277,
+ "step": 4053
+ },
+ {
+ "epoch": 56.30742358078603,
+ "grad_norm": 0.02252180315554142,
+ "learning_rate": 0.0006,
+ "loss": 4.565267086029053,
+ "step": 4054
+ },
+ {
+ "epoch": 56.32139737991266,
+ "grad_norm": 0.02306070365011692,
+ "learning_rate": 0.0006,
+ "loss": 4.3591389656066895,
+ "step": 4055
+ },
+ {
+ "epoch": 56.3353711790393,
+ "grad_norm": 0.020292513072490692,
+ "learning_rate": 0.0006,
+ "loss": 4.407517433166504,
+ "step": 4056
+ },
+ {
+ "epoch": 56.34934497816594,
+ "grad_norm": 0.019154123961925507,
+ "learning_rate": 0.0006,
+ "loss": 4.482261657714844,
+ "step": 4057
+ },
+ {
+ "epoch": 56.36331877729258,
+ "grad_norm": 0.017338937148451805,
+ "learning_rate": 0.0006,
+ "loss": 4.4463582038879395,
+ "step": 4058
+ },
+ {
+ "epoch": 56.377292576419215,
+ "grad_norm": 0.01730559952557087,
+ "learning_rate": 0.0006,
+ "loss": 4.542474746704102,
+ "step": 4059
+ },
+ {
+ "epoch": 56.391266375545854,
+ "grad_norm": 0.017029426991939545,
+ "learning_rate": 0.0006,
+ "loss": 4.552488327026367,
+ "step": 4060
+ },
+ {
+ "epoch": 56.40524017467249,
+ "grad_norm": 0.01745498552918434,
+ "learning_rate": 0.0006,
+ "loss": 4.52928352355957,
+ "step": 4061
+ },
+ {
+ "epoch": 56.419213973799124,
+ "grad_norm": 0.0168515145778656,
+ "learning_rate": 0.0006,
+ "loss": 4.4516801834106445,
+ "step": 4062
+ },
+ {
+ "epoch": 56.43318777292576,
+ "grad_norm": 0.018545417115092278,
+ "learning_rate": 0.0006,
+ "loss": 4.569811820983887,
+ "step": 4063
+ },
+ {
+ "epoch": 56.4471615720524,
+ "grad_norm": 0.02227838523685932,
+ "learning_rate": 0.0006,
+ "loss": 4.412224769592285,
+ "step": 4064
+ },
+ {
+ "epoch": 56.46113537117904,
+ "grad_norm": 0.023338504135608673,
+ "learning_rate": 0.0006,
+ "loss": 4.490184307098389,
+ "step": 4065
+ },
+ {
+ "epoch": 56.47510917030568,
+ "grad_norm": 0.02257535606622696,
+ "learning_rate": 0.0006,
+ "loss": 4.479152679443359,
+ "step": 4066
+ },
+ {
+ "epoch": 56.48908296943232,
+ "grad_norm": 0.022737598046660423,
+ "learning_rate": 0.0006,
+ "loss": 4.418235778808594,
+ "step": 4067
+ },
+ {
+ "epoch": 56.50305676855895,
+ "grad_norm": 0.024603573605418205,
+ "learning_rate": 0.0006,
+ "loss": 4.45781135559082,
+ "step": 4068
+ },
+ {
+ "epoch": 56.51703056768559,
+ "grad_norm": 0.024953778833150864,
+ "learning_rate": 0.0006,
+ "loss": 4.593907833099365,
+ "step": 4069
+ },
+ {
+ "epoch": 56.531004366812226,
+ "grad_norm": 0.02195645309984684,
+ "learning_rate": 0.0006,
+ "loss": 4.467643737792969,
+ "step": 4070
+ },
+ {
+ "epoch": 56.544978165938865,
+ "grad_norm": 0.017126578837633133,
+ "learning_rate": 0.0006,
+ "loss": 4.555484771728516,
+ "step": 4071
+ },
+ {
+ "epoch": 56.5589519650655,
+ "grad_norm": 0.019482605159282684,
+ "learning_rate": 0.0006,
+ "loss": 4.430915832519531,
+ "step": 4072
+ },
+ {
+ "epoch": 56.57292576419214,
+ "grad_norm": 0.0219392292201519,
+ "learning_rate": 0.0006,
+ "loss": 4.593365669250488,
+ "step": 4073
+ },
+ {
+ "epoch": 56.58689956331878,
+ "grad_norm": 0.02420145645737648,
+ "learning_rate": 0.0006,
+ "loss": 4.491467475891113,
+ "step": 4074
+ },
+ {
+ "epoch": 56.60087336244541,
+ "grad_norm": 0.021523723378777504,
+ "learning_rate": 0.0006,
+ "loss": 4.463634490966797,
+ "step": 4075
+ },
+ {
+ "epoch": 56.61484716157205,
+ "grad_norm": 0.020242048427462578,
+ "learning_rate": 0.0006,
+ "loss": 4.577968597412109,
+ "step": 4076
+ },
+ {
+ "epoch": 56.62882096069869,
+ "grad_norm": 0.017771173268556595,
+ "learning_rate": 0.0006,
+ "loss": 4.568269729614258,
+ "step": 4077
+ },
+ {
+ "epoch": 56.64279475982533,
+ "grad_norm": 0.01598544418811798,
+ "learning_rate": 0.0006,
+ "loss": 4.476841926574707,
+ "step": 4078
+ },
+ {
+ "epoch": 56.65676855895197,
+ "grad_norm": 0.0168446097522974,
+ "learning_rate": 0.0006,
+ "loss": 4.480965614318848,
+ "step": 4079
+ },
+ {
+ "epoch": 56.670742358078606,
+ "grad_norm": 0.014948660507798195,
+ "learning_rate": 0.0006,
+ "loss": 4.535400390625,
+ "step": 4080
+ },
+ {
+ "epoch": 56.68471615720524,
+ "grad_norm": 0.014899153262376785,
+ "learning_rate": 0.0006,
+ "loss": 4.476048946380615,
+ "step": 4081
+ },
+ {
+ "epoch": 56.698689956331876,
+ "grad_norm": 0.014051870442926884,
+ "learning_rate": 0.0006,
+ "loss": 4.533884048461914,
+ "step": 4082
+ },
+ {
+ "epoch": 56.712663755458514,
+ "grad_norm": 0.014708572067320347,
+ "learning_rate": 0.0006,
+ "loss": 4.586690902709961,
+ "step": 4083
+ },
+ {
+ "epoch": 56.72663755458515,
+ "grad_norm": 0.014797299169003963,
+ "learning_rate": 0.0006,
+ "loss": 4.446291446685791,
+ "step": 4084
+ },
+ {
+ "epoch": 56.74061135371179,
+ "grad_norm": 0.015256541781127453,
+ "learning_rate": 0.0006,
+ "loss": 4.553572654724121,
+ "step": 4085
+ },
+ {
+ "epoch": 56.75458515283843,
+ "grad_norm": 0.016042491421103477,
+ "learning_rate": 0.0006,
+ "loss": 4.507801055908203,
+ "step": 4086
+ },
+ {
+ "epoch": 56.76855895196506,
+ "grad_norm": 0.018835173919796944,
+ "learning_rate": 0.0006,
+ "loss": 4.393911838531494,
+ "step": 4087
+ },
+ {
+ "epoch": 56.7825327510917,
+ "grad_norm": 0.021994033828377724,
+ "learning_rate": 0.0006,
+ "loss": 4.543368339538574,
+ "step": 4088
+ },
+ {
+ "epoch": 56.79650655021834,
+ "grad_norm": 0.0230990182608366,
+ "learning_rate": 0.0006,
+ "loss": 4.507420539855957,
+ "step": 4089
+ },
+ {
+ "epoch": 56.81048034934498,
+ "grad_norm": 0.018818842247128487,
+ "learning_rate": 0.0006,
+ "loss": 4.499941349029541,
+ "step": 4090
+ },
+ {
+ "epoch": 56.82445414847162,
+ "grad_norm": 0.018257278949022293,
+ "learning_rate": 0.0006,
+ "loss": 4.380057334899902,
+ "step": 4091
+ },
+ {
+ "epoch": 56.838427947598255,
+ "grad_norm": 0.020073018968105316,
+ "learning_rate": 0.0006,
+ "loss": 4.572811603546143,
+ "step": 4092
+ },
+ {
+ "epoch": 56.852401746724894,
+ "grad_norm": 0.02078658528625965,
+ "learning_rate": 0.0006,
+ "loss": 4.446835517883301,
+ "step": 4093
+ },
+ {
+ "epoch": 56.866375545851525,
+ "grad_norm": 0.01793118566274643,
+ "learning_rate": 0.0006,
+ "loss": 4.517854690551758,
+ "step": 4094
+ },
+ {
+ "epoch": 56.880349344978164,
+ "grad_norm": 0.01608281210064888,
+ "learning_rate": 0.0006,
+ "loss": 4.506482124328613,
+ "step": 4095
+ },
+ {
+ "epoch": 56.8943231441048,
+ "grad_norm": 0.017405368387699127,
+ "learning_rate": 0.0006,
+ "loss": 4.439718246459961,
+ "step": 4096
+ },
+ {
+ "epoch": 56.90829694323144,
+ "grad_norm": 0.017930863425135612,
+ "learning_rate": 0.0006,
+ "loss": 4.409365653991699,
+ "step": 4097
+ },
+ {
+ "epoch": 56.92227074235808,
+ "grad_norm": 0.015757670626044273,
+ "learning_rate": 0.0006,
+ "loss": 4.505547523498535,
+ "step": 4098
+ },
+ {
+ "epoch": 56.93624454148472,
+ "grad_norm": 0.01580345816910267,
+ "learning_rate": 0.0006,
+ "loss": 4.462774753570557,
+ "step": 4099
+ },
+ {
+ "epoch": 56.95021834061135,
+ "grad_norm": 0.01659180410206318,
+ "learning_rate": 0.0006,
+ "loss": 4.49345588684082,
+ "step": 4100
+ },
+ {
+ "epoch": 56.96419213973799,
+ "grad_norm": 0.01877259835600853,
+ "learning_rate": 0.0006,
+ "loss": 4.574443340301514,
+ "step": 4101
+ },
+ {
+ "epoch": 56.97816593886463,
+ "grad_norm": 0.01977839693427086,
+ "learning_rate": 0.0006,
+ "loss": 4.574034690856934,
+ "step": 4102
+ },
+ {
+ "epoch": 56.992139737991266,
+ "grad_norm": 0.020627597346901894,
+ "learning_rate": 0.0006,
+ "loss": 4.50969123840332,
+ "step": 4103
+ },
+ {
+ "epoch": 57.0,
+ "grad_norm": 0.025128807872533798,
+ "learning_rate": 0.0006,
+ "loss": 4.401662826538086,
+ "step": 4104
+ },
+ {
+ "epoch": 57.0,
+ "eval_loss": 4.779252052307129,
+ "eval_runtime": 56.8104,
+ "eval_samples_per_second": 42.985,
+ "eval_steps_per_second": 1.355,
+ "step": 4104
+ },
+ {
+ "epoch": 57.01397379912664,
+ "grad_norm": 0.027712570503354073,
+ "learning_rate": 0.0006,
+ "loss": 4.472955703735352,
+ "step": 4105
+ },
+ {
+ "epoch": 57.02794759825328,
+ "grad_norm": 0.02164197526872158,
+ "learning_rate": 0.0006,
+ "loss": 4.387596130371094,
+ "step": 4106
+ },
+ {
+ "epoch": 57.041921397379916,
+ "grad_norm": 0.017972752451896667,
+ "learning_rate": 0.0006,
+ "loss": 4.4326276779174805,
+ "step": 4107
+ },
+ {
+ "epoch": 57.05589519650655,
+ "grad_norm": 0.020811019465327263,
+ "learning_rate": 0.0006,
+ "loss": 4.452951431274414,
+ "step": 4108
+ },
+ {
+ "epoch": 57.069868995633186,
+ "grad_norm": 0.019913259893655777,
+ "learning_rate": 0.0006,
+ "loss": 4.448301315307617,
+ "step": 4109
+ },
+ {
+ "epoch": 57.083842794759825,
+ "grad_norm": 0.017409102991223335,
+ "learning_rate": 0.0006,
+ "loss": 4.4108357429504395,
+ "step": 4110
+ },
+ {
+ "epoch": 57.09781659388646,
+ "grad_norm": 0.019374269992113113,
+ "learning_rate": 0.0006,
+ "loss": 4.398412227630615,
+ "step": 4111
+ },
+ {
+ "epoch": 57.1117903930131,
+ "grad_norm": 0.015521776862442493,
+ "learning_rate": 0.0006,
+ "loss": 4.508594989776611,
+ "step": 4112
+ },
+ {
+ "epoch": 57.12576419213974,
+ "grad_norm": 0.017201920971274376,
+ "learning_rate": 0.0006,
+ "loss": 4.449558734893799,
+ "step": 4113
+ },
+ {
+ "epoch": 57.13973799126637,
+ "grad_norm": 0.018062541261315346,
+ "learning_rate": 0.0006,
+ "loss": 4.517874240875244,
+ "step": 4114
+ },
+ {
+ "epoch": 57.15371179039301,
+ "grad_norm": 0.015846073627471924,
+ "learning_rate": 0.0006,
+ "loss": 4.427501678466797,
+ "step": 4115
+ },
+ {
+ "epoch": 57.16768558951965,
+ "grad_norm": 0.01849362626671791,
+ "learning_rate": 0.0006,
+ "loss": 4.433181285858154,
+ "step": 4116
+ },
+ {
+ "epoch": 57.18165938864629,
+ "grad_norm": 0.01948804222047329,
+ "learning_rate": 0.0006,
+ "loss": 4.446122646331787,
+ "step": 4117
+ },
+ {
+ "epoch": 57.19563318777293,
+ "grad_norm": 0.017354389652609825,
+ "learning_rate": 0.0006,
+ "loss": 4.431153774261475,
+ "step": 4118
+ },
+ {
+ "epoch": 57.209606986899566,
+ "grad_norm": 0.015817373991012573,
+ "learning_rate": 0.0006,
+ "loss": 4.448566913604736,
+ "step": 4119
+ },
+ {
+ "epoch": 57.223580786026204,
+ "grad_norm": 0.01635785959661007,
+ "learning_rate": 0.0006,
+ "loss": 4.408031463623047,
+ "step": 4120
+ },
+ {
+ "epoch": 57.237554585152836,
+ "grad_norm": 0.014528338797390461,
+ "learning_rate": 0.0006,
+ "loss": 4.541668891906738,
+ "step": 4121
+ },
+ {
+ "epoch": 57.251528384279474,
+ "grad_norm": 0.01589174196124077,
+ "learning_rate": 0.0006,
+ "loss": 4.513199329376221,
+ "step": 4122
+ },
+ {
+ "epoch": 57.26550218340611,
+ "grad_norm": 0.016093695536255836,
+ "learning_rate": 0.0006,
+ "loss": 4.516178131103516,
+ "step": 4123
+ },
+ {
+ "epoch": 57.27947598253275,
+ "grad_norm": 0.01788361556828022,
+ "learning_rate": 0.0006,
+ "loss": 4.524880409240723,
+ "step": 4124
+ },
+ {
+ "epoch": 57.29344978165939,
+ "grad_norm": 0.01591576263308525,
+ "learning_rate": 0.0006,
+ "loss": 4.44764518737793,
+ "step": 4125
+ },
+ {
+ "epoch": 57.30742358078603,
+ "grad_norm": 0.015452570281922817,
+ "learning_rate": 0.0006,
+ "loss": 4.46844482421875,
+ "step": 4126
+ },
+ {
+ "epoch": 57.32139737991266,
+ "grad_norm": 0.016404492780566216,
+ "learning_rate": 0.0006,
+ "loss": 4.443678855895996,
+ "step": 4127
+ },
+ {
+ "epoch": 57.3353711790393,
+ "grad_norm": 0.019967148080468178,
+ "learning_rate": 0.0006,
+ "loss": 4.430230140686035,
+ "step": 4128
+ },
+ {
+ "epoch": 57.34934497816594,
+ "grad_norm": 0.02594015561044216,
+ "learning_rate": 0.0006,
+ "loss": 4.47061824798584,
+ "step": 4129
+ },
+ {
+ "epoch": 57.36331877729258,
+ "grad_norm": 0.031096890568733215,
+ "learning_rate": 0.0006,
+ "loss": 4.462275981903076,
+ "step": 4130
+ },
+ {
+ "epoch": 57.377292576419215,
+ "grad_norm": 0.02834898792207241,
+ "learning_rate": 0.0006,
+ "loss": 4.41143798828125,
+ "step": 4131
+ },
+ {
+ "epoch": 57.391266375545854,
+ "grad_norm": 0.023621153086423874,
+ "learning_rate": 0.0006,
+ "loss": 4.562548637390137,
+ "step": 4132
+ },
+ {
+ "epoch": 57.40524017467249,
+ "grad_norm": 0.02277296595275402,
+ "learning_rate": 0.0006,
+ "loss": 4.456247329711914,
+ "step": 4133
+ },
+ {
+ "epoch": 57.419213973799124,
+ "grad_norm": 0.019868768751621246,
+ "learning_rate": 0.0006,
+ "loss": 4.524300575256348,
+ "step": 4134
+ },
+ {
+ "epoch": 57.43318777292576,
+ "grad_norm": 0.01887078955769539,
+ "learning_rate": 0.0006,
+ "loss": 4.540097236633301,
+ "step": 4135
+ },
+ {
+ "epoch": 57.4471615720524,
+ "grad_norm": 0.024268802255392075,
+ "learning_rate": 0.0006,
+ "loss": 4.397233963012695,
+ "step": 4136
+ },
+ {
+ "epoch": 57.46113537117904,
+ "grad_norm": 0.025739721953868866,
+ "learning_rate": 0.0006,
+ "loss": 4.354194641113281,
+ "step": 4137
+ },
+ {
+ "epoch": 57.47510917030568,
+ "grad_norm": 0.02442290261387825,
+ "learning_rate": 0.0006,
+ "loss": 4.303822994232178,
+ "step": 4138
+ },
+ {
+ "epoch": 57.48908296943232,
+ "grad_norm": 0.022215209901332855,
+ "learning_rate": 0.0006,
+ "loss": 4.4870452880859375,
+ "step": 4139
+ },
+ {
+ "epoch": 57.50305676855895,
+ "grad_norm": 0.01789158768951893,
+ "learning_rate": 0.0006,
+ "loss": 4.343809127807617,
+ "step": 4140
+ },
+ {
+ "epoch": 57.51703056768559,
+ "grad_norm": 0.020381739363074303,
+ "learning_rate": 0.0006,
+ "loss": 4.53312873840332,
+ "step": 4141
+ },
+ {
+ "epoch": 57.531004366812226,
+ "grad_norm": 0.019585080444812775,
+ "learning_rate": 0.0006,
+ "loss": 4.510162353515625,
+ "step": 4142
+ },
+ {
+ "epoch": 57.544978165938865,
+ "grad_norm": 0.01985127478837967,
+ "learning_rate": 0.0006,
+ "loss": 4.3654022216796875,
+ "step": 4143
+ },
+ {
+ "epoch": 57.5589519650655,
+ "grad_norm": 0.01962905190885067,
+ "learning_rate": 0.0006,
+ "loss": 4.560830116271973,
+ "step": 4144
+ },
+ {
+ "epoch": 57.57292576419214,
+ "grad_norm": 0.02679629437625408,
+ "learning_rate": 0.0006,
+ "loss": 4.300804138183594,
+ "step": 4145
+ },
+ {
+ "epoch": 57.58689956331878,
+ "grad_norm": 0.039830856025218964,
+ "learning_rate": 0.0006,
+ "loss": 4.54150915145874,
+ "step": 4146
+ },
+ {
+ "epoch": 57.60087336244541,
+ "grad_norm": 0.049527477473020554,
+ "learning_rate": 0.0006,
+ "loss": 4.405792236328125,
+ "step": 4147
+ },
+ {
+ "epoch": 57.61484716157205,
+ "grad_norm": 0.042966946959495544,
+ "learning_rate": 0.0006,
+ "loss": 4.500643730163574,
+ "step": 4148
+ },
+ {
+ "epoch": 57.62882096069869,
+ "grad_norm": 0.08999177813529968,
+ "learning_rate": 0.0006,
+ "loss": 4.504790782928467,
+ "step": 4149
+ },
+ {
+ "epoch": 57.64279475982533,
+ "grad_norm": 0.44693124294281006,
+ "learning_rate": 0.0006,
+ "loss": 4.845344066619873,
+ "step": 4150
+ },
+ {
+ "epoch": 57.65676855895197,
+ "grad_norm": 0.7143144011497498,
+ "learning_rate": 0.0006,
+ "loss": 6.907845497131348,
+ "step": 4151
+ },
+ {
+ "epoch": 57.670742358078606,
+ "grad_norm": 0.25239595770835876,
+ "learning_rate": 0.0006,
+ "loss": 7.082579612731934,
+ "step": 4152
+ },
+ {
+ "epoch": 57.68471615720524,
+ "grad_norm": 0.26550790667533875,
+ "learning_rate": 0.0006,
+ "loss": 7.50180196762085,
+ "step": 4153
+ },
+ {
+ "epoch": 57.698689956331876,
+ "grad_norm": 0.14735378324985504,
+ "learning_rate": 0.0006,
+ "loss": 7.214590072631836,
+ "step": 4154
+ },
+ {
+ "epoch": 57.712663755458514,
+ "grad_norm": 0.17685334384441376,
+ "learning_rate": 0.0006,
+ "loss": 6.861320495605469,
+ "step": 4155
+ },
+ {
+ "epoch": 57.72663755458515,
+ "grad_norm": 0.18231339752674103,
+ "learning_rate": 0.0006,
+ "loss": 6.791839599609375,
+ "step": 4156
+ },
+ {
+ "epoch": 57.74061135371179,
+ "grad_norm": 0.09230359643697739,
+ "learning_rate": 0.0006,
+ "loss": 6.760223388671875,
+ "step": 4157
+ },
+ {
+ "epoch": 57.75458515283843,
+ "grad_norm": 0.06918249279260635,
+ "learning_rate": 0.0006,
+ "loss": 6.532999515533447,
+ "step": 4158
+ },
+ {
+ "epoch": 57.76855895196506,
+ "grad_norm": 0.0901007279753685,
+ "learning_rate": 0.0006,
+ "loss": 6.52907133102417,
+ "step": 4159
+ },
+ {
+ "epoch": 57.7825327510917,
+ "grad_norm": 0.05678229779005051,
+ "learning_rate": 0.0006,
+ "loss": 6.295290470123291,
+ "step": 4160
+ },
+ {
+ "epoch": 57.79650655021834,
+ "grad_norm": 0.04450210556387901,
+ "learning_rate": 0.0006,
+ "loss": 6.14900541305542,
+ "step": 4161
+ },
+ {
+ "epoch": 57.81048034934498,
+ "grad_norm": 0.049191512167453766,
+ "learning_rate": 0.0006,
+ "loss": 6.19650411605835,
+ "step": 4162
+ },
+ {
+ "epoch": 57.82445414847162,
+ "grad_norm": 0.03908967226743698,
+ "learning_rate": 0.0006,
+ "loss": 6.054948806762695,
+ "step": 4163
+ },
+ {
+ "epoch": 57.838427947598255,
+ "grad_norm": 0.04023387283086777,
+ "learning_rate": 0.0006,
+ "loss": 6.034873962402344,
+ "step": 4164
+ },
+ {
+ "epoch": 57.852401746724894,
+ "grad_norm": 0.03304268419742584,
+ "learning_rate": 0.0006,
+ "loss": 5.769852161407471,
+ "step": 4165
+ },
+ {
+ "epoch": 57.866375545851525,
+ "grad_norm": 0.03432450816035271,
+ "learning_rate": 0.0006,
+ "loss": 5.788785457611084,
+ "step": 4166
+ },
+ {
+ "epoch": 57.880349344978164,
+ "grad_norm": 0.029965840280056,
+ "learning_rate": 0.0006,
+ "loss": 5.815243721008301,
+ "step": 4167
+ },
+ {
+ "epoch": 57.8943231441048,
+ "grad_norm": 0.03637698292732239,
+ "learning_rate": 0.0006,
+ "loss": 5.7204060554504395,
+ "step": 4168
+ },
+ {
+ "epoch": 57.90829694323144,
+ "grad_norm": 0.03477517142891884,
+ "learning_rate": 0.0006,
+ "loss": 5.605930328369141,
+ "step": 4169
+ },
+ {
+ "epoch": 57.92227074235808,
+ "grad_norm": 0.05061859264969826,
+ "learning_rate": 0.0006,
+ "loss": 5.547054767608643,
+ "step": 4170
+ },
+ {
+ "epoch": 57.93624454148472,
+ "grad_norm": 0.06299655884504318,
+ "learning_rate": 0.0006,
+ "loss": 5.552166938781738,
+ "step": 4171
+ },
+ {
+ "epoch": 57.95021834061135,
+ "grad_norm": 0.04215948283672333,
+ "learning_rate": 0.0006,
+ "loss": 5.375457763671875,
+ "step": 4172
+ },
+ {
+ "epoch": 57.96419213973799,
+ "grad_norm": 0.0348566472530365,
+ "learning_rate": 0.0006,
+ "loss": 5.418606281280518,
+ "step": 4173
+ },
+ {
+ "epoch": 57.97816593886463,
+ "grad_norm": 0.03807530924677849,
+ "learning_rate": 0.0006,
+ "loss": 5.403156280517578,
+ "step": 4174
+ },
+ {
+ "epoch": 57.992139737991266,
+ "grad_norm": 0.02989993803203106,
+ "learning_rate": 0.0006,
+ "loss": 5.355951309204102,
+ "step": 4175
+ },
+ {
+ "epoch": 58.0,
+ "grad_norm": 0.034300558269023895,
+ "learning_rate": 0.0006,
+ "loss": 5.44920539855957,
+ "step": 4176
+ },
+ {
+ "epoch": 58.0,
+ "eval_loss": 5.416731834411621,
+ "eval_runtime": 56.7934,
+ "eval_samples_per_second": 42.998,
+ "eval_steps_per_second": 1.356,
+ "step": 4176
+ },
+ {
+ "epoch": 58.01397379912664,
+ "grad_norm": 0.033961448818445206,
+ "learning_rate": 0.0006,
+ "loss": 5.250707626342773,
+ "step": 4177
+ },
+ {
+ "epoch": 58.02794759825328,
+ "grad_norm": 0.027343502268195152,
+ "learning_rate": 0.0006,
+ "loss": 5.149238109588623,
+ "step": 4178
+ },
+ {
+ "epoch": 58.041921397379916,
+ "grad_norm": 0.03033650480210781,
+ "learning_rate": 0.0006,
+ "loss": 5.072463512420654,
+ "step": 4179
+ },
+ {
+ "epoch": 58.05589519650655,
+ "grad_norm": 0.028262868523597717,
+ "learning_rate": 0.0006,
+ "loss": 5.109417915344238,
+ "step": 4180
+ },
+ {
+ "epoch": 58.069868995633186,
+ "grad_norm": 0.02373034693300724,
+ "learning_rate": 0.0006,
+ "loss": 5.082464218139648,
+ "step": 4181
+ },
+ {
+ "epoch": 58.083842794759825,
+ "grad_norm": 0.025879688560962677,
+ "learning_rate": 0.0006,
+ "loss": 5.056792259216309,
+ "step": 4182
+ },
+ {
+ "epoch": 58.09781659388646,
+ "grad_norm": 0.026252275332808495,
+ "learning_rate": 0.0006,
+ "loss": 5.064239978790283,
+ "step": 4183
+ },
+ {
+ "epoch": 58.1117903930131,
+ "grad_norm": 0.028177792206406593,
+ "learning_rate": 0.0006,
+ "loss": 4.922310829162598,
+ "step": 4184
+ },
+ {
+ "epoch": 58.12576419213974,
+ "grad_norm": 0.02630820870399475,
+ "learning_rate": 0.0006,
+ "loss": 4.992884635925293,
+ "step": 4185
+ },
+ {
+ "epoch": 58.13973799126637,
+ "grad_norm": 0.024890903383493423,
+ "learning_rate": 0.0006,
+ "loss": 4.910036563873291,
+ "step": 4186
+ },
+ {
+ "epoch": 58.15371179039301,
+ "grad_norm": 0.02822992391884327,
+ "learning_rate": 0.0006,
+ "loss": 4.866640090942383,
+ "step": 4187
+ },
+ {
+ "epoch": 58.16768558951965,
+ "grad_norm": 0.03253607079386711,
+ "learning_rate": 0.0006,
+ "loss": 4.9112958908081055,
+ "step": 4188
+ },
+ {
+ "epoch": 58.18165938864629,
+ "grad_norm": 0.03765944764018059,
+ "learning_rate": 0.0006,
+ "loss": 4.774468421936035,
+ "step": 4189
+ },
+ {
+ "epoch": 58.19563318777293,
+ "grad_norm": 0.039874881505966187,
+ "learning_rate": 0.0006,
+ "loss": 4.831678867340088,
+ "step": 4190
+ },
+ {
+ "epoch": 58.209606986899566,
+ "grad_norm": 0.034595925360918045,
+ "learning_rate": 0.0006,
+ "loss": 4.8243794441223145,
+ "step": 4191
+ },
+ {
+ "epoch": 58.223580786026204,
+ "grad_norm": 0.04840189591050148,
+ "learning_rate": 0.0006,
+ "loss": 4.854156494140625,
+ "step": 4192
+ },
+ {
+ "epoch": 58.237554585152836,
+ "grad_norm": 0.06358753889799118,
+ "learning_rate": 0.0006,
+ "loss": 4.836430549621582,
+ "step": 4193
+ },
+ {
+ "epoch": 58.251528384279474,
+ "grad_norm": 0.0507555715739727,
+ "learning_rate": 0.0006,
+ "loss": 4.88357400894165,
+ "step": 4194
+ },
+ {
+ "epoch": 58.26550218340611,
+ "grad_norm": 0.04384690150618553,
+ "learning_rate": 0.0006,
+ "loss": 4.774049282073975,
+ "step": 4195
+ },
+ {
+ "epoch": 58.27947598253275,
+ "grad_norm": 0.03246928006410599,
+ "learning_rate": 0.0006,
+ "loss": 4.686605930328369,
+ "step": 4196
+ },
+ {
+ "epoch": 58.29344978165939,
+ "grad_norm": 0.028878789395093918,
+ "learning_rate": 0.0006,
+ "loss": 4.796070575714111,
+ "step": 4197
+ },
+ {
+ "epoch": 58.30742358078603,
+ "grad_norm": 0.028175361454486847,
+ "learning_rate": 0.0006,
+ "loss": 4.728490829467773,
+ "step": 4198
+ },
+ {
+ "epoch": 58.32139737991266,
+ "grad_norm": 0.029428014531731606,
+ "learning_rate": 0.0006,
+ "loss": 4.7159624099731445,
+ "step": 4199
+ },
+ {
+ "epoch": 58.3353711790393,
+ "grad_norm": 0.031227827072143555,
+ "learning_rate": 0.0006,
+ "loss": 4.702625274658203,
+ "step": 4200
+ },
+ {
+ "epoch": 58.34934497816594,
+ "grad_norm": 0.028843428939580917,
+ "learning_rate": 0.0006,
+ "loss": 4.811799049377441,
+ "step": 4201
+ },
+ {
+ "epoch": 58.36331877729258,
+ "grad_norm": 0.026780391111969948,
+ "learning_rate": 0.0006,
+ "loss": 4.81662654876709,
+ "step": 4202
+ },
+ {
+ "epoch": 58.377292576419215,
+ "grad_norm": 0.019699757918715477,
+ "learning_rate": 0.0006,
+ "loss": 4.713685989379883,
+ "step": 4203
+ },
+ {
+ "epoch": 58.391266375545854,
+ "grad_norm": 0.024389250203967094,
+ "learning_rate": 0.0006,
+ "loss": 4.72354793548584,
+ "step": 4204
+ },
+ {
+ "epoch": 58.40524017467249,
+ "grad_norm": 0.02218274027109146,
+ "learning_rate": 0.0006,
+ "loss": 4.674125671386719,
+ "step": 4205
+ },
+ {
+ "epoch": 58.419213973799124,
+ "grad_norm": 0.02041488140821457,
+ "learning_rate": 0.0006,
+ "loss": 4.7909698486328125,
+ "step": 4206
+ },
+ {
+ "epoch": 58.43318777292576,
+ "grad_norm": 0.018561935052275658,
+ "learning_rate": 0.0006,
+ "loss": 4.808770179748535,
+ "step": 4207
+ },
+ {
+ "epoch": 58.4471615720524,
+ "grad_norm": 0.01708846725523472,
+ "learning_rate": 0.0006,
+ "loss": 4.701626777648926,
+ "step": 4208
+ },
+ {
+ "epoch": 58.46113537117904,
+ "grad_norm": 0.0222734697163105,
+ "learning_rate": 0.0006,
+ "loss": 4.505410194396973,
+ "step": 4209
+ },
+ {
+ "epoch": 58.47510917030568,
+ "grad_norm": 0.03197057917714119,
+ "learning_rate": 0.0006,
+ "loss": 4.610360145568848,
+ "step": 4210
+ },
+ {
+ "epoch": 58.48908296943232,
+ "grad_norm": 0.06854520738124847,
+ "learning_rate": 0.0006,
+ "loss": 4.625649452209473,
+ "step": 4211
+ },
+ {
+ "epoch": 58.50305676855895,
+ "grad_norm": 0.09981521219015121,
+ "learning_rate": 0.0006,
+ "loss": 4.832150459289551,
+ "step": 4212
+ },
+ {
+ "epoch": 58.51703056768559,
+ "grad_norm": 0.046356040984392166,
+ "learning_rate": 0.0006,
+ "loss": 4.5641632080078125,
+ "step": 4213
+ },
+ {
+ "epoch": 58.531004366812226,
+ "grad_norm": 0.03882667422294617,
+ "learning_rate": 0.0006,
+ "loss": 4.666974067687988,
+ "step": 4214
+ },
+ {
+ "epoch": 58.544978165938865,
+ "grad_norm": 0.033203233033418655,
+ "learning_rate": 0.0006,
+ "loss": 4.629413604736328,
+ "step": 4215
+ },
+ {
+ "epoch": 58.5589519650655,
+ "grad_norm": 0.02806735597550869,
+ "learning_rate": 0.0006,
+ "loss": 4.763245582580566,
+ "step": 4216
+ },
+ {
+ "epoch": 58.57292576419214,
+ "grad_norm": 0.026728278025984764,
+ "learning_rate": 0.0006,
+ "loss": 4.629597187042236,
+ "step": 4217
+ },
+ {
+ "epoch": 58.58689956331878,
+ "grad_norm": 0.024380959570407867,
+ "learning_rate": 0.0006,
+ "loss": 4.633614540100098,
+ "step": 4218
+ },
+ {
+ "epoch": 58.60087336244541,
+ "grad_norm": 0.022312544286251068,
+ "learning_rate": 0.0006,
+ "loss": 4.671379566192627,
+ "step": 4219
+ },
+ {
+ "epoch": 58.61484716157205,
+ "grad_norm": 0.022049568593502045,
+ "learning_rate": 0.0006,
+ "loss": 4.661557197570801,
+ "step": 4220
+ },
+ {
+ "epoch": 58.62882096069869,
+ "grad_norm": 0.02119818702340126,
+ "learning_rate": 0.0006,
+ "loss": 4.688562393188477,
+ "step": 4221
+ },
+ {
+ "epoch": 58.64279475982533,
+ "grad_norm": 0.019769301638007164,
+ "learning_rate": 0.0006,
+ "loss": 4.607748031616211,
+ "step": 4222
+ },
+ {
+ "epoch": 58.65676855895197,
+ "grad_norm": 0.02124079130589962,
+ "learning_rate": 0.0006,
+ "loss": 4.570640563964844,
+ "step": 4223
+ },
+ {
+ "epoch": 58.670742358078606,
+ "grad_norm": 0.016344040632247925,
+ "learning_rate": 0.0006,
+ "loss": 4.676113128662109,
+ "step": 4224
+ },
+ {
+ "epoch": 58.68471615720524,
+ "grad_norm": 0.01735256239771843,
+ "learning_rate": 0.0006,
+ "loss": 4.62088680267334,
+ "step": 4225
+ },
+ {
+ "epoch": 58.698689956331876,
+ "grad_norm": 0.019323458895087242,
+ "learning_rate": 0.0006,
+ "loss": 4.579405307769775,
+ "step": 4226
+ },
+ {
+ "epoch": 58.712663755458514,
+ "grad_norm": 0.01760544814169407,
+ "learning_rate": 0.0006,
+ "loss": 4.625407695770264,
+ "step": 4227
+ },
+ {
+ "epoch": 58.72663755458515,
+ "grad_norm": 0.015593250282108784,
+ "learning_rate": 0.0006,
+ "loss": 4.6326117515563965,
+ "step": 4228
+ },
+ {
+ "epoch": 58.74061135371179,
+ "grad_norm": 0.014357523061335087,
+ "learning_rate": 0.0006,
+ "loss": 4.6869401931762695,
+ "step": 4229
+ },
+ {
+ "epoch": 58.75458515283843,
+ "grad_norm": 0.01604357920587063,
+ "learning_rate": 0.0006,
+ "loss": 4.690784931182861,
+ "step": 4230
+ },
+ {
+ "epoch": 58.76855895196506,
+ "grad_norm": 0.01673845760524273,
+ "learning_rate": 0.0006,
+ "loss": 4.513410568237305,
+ "step": 4231
+ },
+ {
+ "epoch": 58.7825327510917,
+ "grad_norm": 0.013657779432833195,
+ "learning_rate": 0.0006,
+ "loss": 4.637057781219482,
+ "step": 4232
+ },
+ {
+ "epoch": 58.79650655021834,
+ "grad_norm": 0.013373114168643951,
+ "learning_rate": 0.0006,
+ "loss": 4.719967842102051,
+ "step": 4233
+ },
+ {
+ "epoch": 58.81048034934498,
+ "grad_norm": 0.016345568001270294,
+ "learning_rate": 0.0006,
+ "loss": 4.604109764099121,
+ "step": 4234
+ },
+ {
+ "epoch": 58.82445414847162,
+ "grad_norm": 0.01573033444583416,
+ "learning_rate": 0.0006,
+ "loss": 4.6471428871154785,
+ "step": 4235
+ },
+ {
+ "epoch": 58.838427947598255,
+ "grad_norm": 0.01517449039965868,
+ "learning_rate": 0.0006,
+ "loss": 4.722684860229492,
+ "step": 4236
+ },
+ {
+ "epoch": 58.852401746724894,
+ "grad_norm": 0.0145102022215724,
+ "learning_rate": 0.0006,
+ "loss": 4.662420272827148,
+ "step": 4237
+ },
+ {
+ "epoch": 58.866375545851525,
+ "grad_norm": 0.01457217987626791,
+ "learning_rate": 0.0006,
+ "loss": 4.585169792175293,
+ "step": 4238
+ },
+ {
+ "epoch": 58.880349344978164,
+ "grad_norm": 0.01406773366034031,
+ "learning_rate": 0.0006,
+ "loss": 4.5413665771484375,
+ "step": 4239
+ },
+ {
+ "epoch": 58.8943231441048,
+ "grad_norm": 0.012653871439397335,
+ "learning_rate": 0.0006,
+ "loss": 4.553636074066162,
+ "step": 4240
+ },
+ {
+ "epoch": 58.90829694323144,
+ "grad_norm": 0.01261200476437807,
+ "learning_rate": 0.0006,
+ "loss": 4.534193992614746,
+ "step": 4241
+ },
+ {
+ "epoch": 58.92227074235808,
+ "grad_norm": 0.012955864891409874,
+ "learning_rate": 0.0006,
+ "loss": 4.697248458862305,
+ "step": 4242
+ },
+ {
+ "epoch": 58.93624454148472,
+ "grad_norm": 0.012134749442338943,
+ "learning_rate": 0.0006,
+ "loss": 4.557095527648926,
+ "step": 4243
+ },
+ {
+ "epoch": 58.95021834061135,
+ "grad_norm": 0.012164677493274212,
+ "learning_rate": 0.0006,
+ "loss": 4.623579025268555,
+ "step": 4244
+ },
+ {
+ "epoch": 58.96419213973799,
+ "grad_norm": 0.013802947476506233,
+ "learning_rate": 0.0006,
+ "loss": 4.562838077545166,
+ "step": 4245
+ },
+ {
+ "epoch": 58.97816593886463,
+ "grad_norm": 0.011420476250350475,
+ "learning_rate": 0.0006,
+ "loss": 4.62265682220459,
+ "step": 4246
+ },
+ {
+ "epoch": 58.992139737991266,
+ "grad_norm": 0.01169576682150364,
+ "learning_rate": 0.0006,
+ "loss": 4.727449417114258,
+ "step": 4247
+ },
+ {
+ "epoch": 59.0,
+ "grad_norm": 0.013577710837125778,
+ "learning_rate": 0.0006,
+ "loss": 4.536325454711914,
+ "step": 4248
+ },
+ {
+ "epoch": 59.0,
+ "eval_loss": 4.86527681350708,
+ "eval_runtime": 60.4463,
+ "eval_samples_per_second": 40.399,
+ "eval_steps_per_second": 1.274,
+ "step": 4248
+ },
+ {
+ "epoch": 59.01397379912664,
+ "grad_norm": 0.015251655131578445,
+ "learning_rate": 0.0006,
+ "loss": 4.634598731994629,
+ "step": 4249
+ },
+ {
+ "epoch": 59.02794759825328,
+ "grad_norm": 0.017414938658475876,
+ "learning_rate": 0.0006,
+ "loss": 4.521329879760742,
+ "step": 4250
+ },
+ {
+ "epoch": 59.041921397379916,
+ "grad_norm": 0.01656588353216648,
+ "learning_rate": 0.0006,
+ "loss": 4.446192264556885,
+ "step": 4251
+ },
+ {
+ "epoch": 59.05589519650655,
+ "grad_norm": 0.012038925662636757,
+ "learning_rate": 0.0006,
+ "loss": 4.523855209350586,
+ "step": 4252
+ },
+ {
+ "epoch": 59.069868995633186,
+ "grad_norm": 0.014403033070266247,
+ "learning_rate": 0.0006,
+ "loss": 4.561079978942871,
+ "step": 4253
+ },
+ {
+ "epoch": 59.083842794759825,
+ "grad_norm": 0.016243597492575645,
+ "learning_rate": 0.0006,
+ "loss": 4.673602104187012,
+ "step": 4254
+ },
+ {
+ "epoch": 59.09781659388646,
+ "grad_norm": 0.016492877155542374,
+ "learning_rate": 0.0006,
+ "loss": 4.564824104309082,
+ "step": 4255
+ },
+ {
+ "epoch": 59.1117903930131,
+ "grad_norm": 0.015980370342731476,
+ "learning_rate": 0.0006,
+ "loss": 4.339249134063721,
+ "step": 4256
+ },
+ {
+ "epoch": 59.12576419213974,
+ "grad_norm": 0.019687356427311897,
+ "learning_rate": 0.0006,
+ "loss": 4.481834888458252,
+ "step": 4257
+ },
+ {
+ "epoch": 59.13973799126637,
+ "grad_norm": 0.02983798086643219,
+ "learning_rate": 0.0006,
+ "loss": 4.4495697021484375,
+ "step": 4258
+ },
+ {
+ "epoch": 59.15371179039301,
+ "grad_norm": 0.04600981995463371,
+ "learning_rate": 0.0006,
+ "loss": 4.567838668823242,
+ "step": 4259
+ },
+ {
+ "epoch": 59.16768558951965,
+ "grad_norm": 0.04852001741528511,
+ "learning_rate": 0.0006,
+ "loss": 4.594701766967773,
+ "step": 4260
+ },
+ {
+ "epoch": 59.18165938864629,
+ "grad_norm": 0.03429504856467247,
+ "learning_rate": 0.0006,
+ "loss": 4.573735237121582,
+ "step": 4261
+ },
+ {
+ "epoch": 59.19563318777293,
+ "grad_norm": 0.029345160350203514,
+ "learning_rate": 0.0006,
+ "loss": 4.488258361816406,
+ "step": 4262
+ },
+ {
+ "epoch": 59.209606986899566,
+ "grad_norm": 0.01961623504757881,
+ "learning_rate": 0.0006,
+ "loss": 4.545222282409668,
+ "step": 4263
+ },
+ {
+ "epoch": 59.223580786026204,
+ "grad_norm": 0.020406130701303482,
+ "learning_rate": 0.0006,
+ "loss": 4.471898078918457,
+ "step": 4264
+ },
+ {
+ "epoch": 59.237554585152836,
+ "grad_norm": 0.020500272512435913,
+ "learning_rate": 0.0006,
+ "loss": 4.538337707519531,
+ "step": 4265
+ },
+ {
+ "epoch": 59.251528384279474,
+ "grad_norm": 0.020727017894387245,
+ "learning_rate": 0.0006,
+ "loss": 4.437284469604492,
+ "step": 4266
+ },
+ {
+ "epoch": 59.26550218340611,
+ "grad_norm": 0.03395998477935791,
+ "learning_rate": 0.0006,
+ "loss": 4.459124565124512,
+ "step": 4267
+ },
+ {
+ "epoch": 59.27947598253275,
+ "grad_norm": 0.04635605961084366,
+ "learning_rate": 0.0006,
+ "loss": 4.520228385925293,
+ "step": 4268
+ },
+ {
+ "epoch": 59.29344978165939,
+ "grad_norm": 0.04253006726503372,
+ "learning_rate": 0.0006,
+ "loss": 4.578673362731934,
+ "step": 4269
+ },
+ {
+ "epoch": 59.30742358078603,
+ "grad_norm": 0.02403509058058262,
+ "learning_rate": 0.0006,
+ "loss": 4.484639644622803,
+ "step": 4270
+ },
+ {
+ "epoch": 59.32139737991266,
+ "grad_norm": 0.02254190482199192,
+ "learning_rate": 0.0006,
+ "loss": 4.511314868927002,
+ "step": 4271
+ },
+ {
+ "epoch": 59.3353711790393,
+ "grad_norm": 0.021585190668702126,
+ "learning_rate": 0.0006,
+ "loss": 4.530239105224609,
+ "step": 4272
+ },
+ {
+ "epoch": 59.34934497816594,
+ "grad_norm": 0.021210316568613052,
+ "learning_rate": 0.0006,
+ "loss": 4.593514442443848,
+ "step": 4273
+ },
+ {
+ "epoch": 59.36331877729258,
+ "grad_norm": 0.019317107275128365,
+ "learning_rate": 0.0006,
+ "loss": 4.5591535568237305,
+ "step": 4274
+ },
+ {
+ "epoch": 59.377292576419215,
+ "grad_norm": 0.018059708178043365,
+ "learning_rate": 0.0006,
+ "loss": 4.4499592781066895,
+ "step": 4275
+ },
+ {
+ "epoch": 59.391266375545854,
+ "grad_norm": 0.017847446724772453,
+ "learning_rate": 0.0006,
+ "loss": 4.473999500274658,
+ "step": 4276
+ },
+ {
+ "epoch": 59.40524017467249,
+ "grad_norm": 0.014944756403565407,
+ "learning_rate": 0.0006,
+ "loss": 4.5934529304504395,
+ "step": 4277
+ },
+ {
+ "epoch": 59.419213973799124,
+ "grad_norm": 0.01663832925260067,
+ "learning_rate": 0.0006,
+ "loss": 4.504262924194336,
+ "step": 4278
+ },
+ {
+ "epoch": 59.43318777292576,
+ "grad_norm": 0.015571796335279942,
+ "learning_rate": 0.0006,
+ "loss": 4.512310981750488,
+ "step": 4279
+ },
+ {
+ "epoch": 59.4471615720524,
+ "grad_norm": 0.015896400436758995,
+ "learning_rate": 0.0006,
+ "loss": 4.452714920043945,
+ "step": 4280
+ },
+ {
+ "epoch": 59.46113537117904,
+ "grad_norm": 0.016010284423828125,
+ "learning_rate": 0.0006,
+ "loss": 4.6188063621521,
+ "step": 4281
+ },
+ {
+ "epoch": 59.47510917030568,
+ "grad_norm": 0.015075593255460262,
+ "learning_rate": 0.0006,
+ "loss": 4.477978229522705,
+ "step": 4282
+ },
+ {
+ "epoch": 59.48908296943232,
+ "grad_norm": 0.014698950573801994,
+ "learning_rate": 0.0006,
+ "loss": 4.544955730438232,
+ "step": 4283
+ },
+ {
+ "epoch": 59.50305676855895,
+ "grad_norm": 0.0136245833709836,
+ "learning_rate": 0.0006,
+ "loss": 4.530460357666016,
+ "step": 4284
+ },
+ {
+ "epoch": 59.51703056768559,
+ "grad_norm": 0.013118310831487179,
+ "learning_rate": 0.0006,
+ "loss": 4.614217758178711,
+ "step": 4285
+ },
+ {
+ "epoch": 59.531004366812226,
+ "grad_norm": 0.01295961532741785,
+ "learning_rate": 0.0006,
+ "loss": 4.43156099319458,
+ "step": 4286
+ },
+ {
+ "epoch": 59.544978165938865,
+ "grad_norm": 0.012429811991751194,
+ "learning_rate": 0.0006,
+ "loss": 4.517333030700684,
+ "step": 4287
+ },
+ {
+ "epoch": 59.5589519650655,
+ "grad_norm": 0.01261830423027277,
+ "learning_rate": 0.0006,
+ "loss": 4.569780349731445,
+ "step": 4288
+ },
+ {
+ "epoch": 59.57292576419214,
+ "grad_norm": 0.012394499965012074,
+ "learning_rate": 0.0006,
+ "loss": 4.5970869064331055,
+ "step": 4289
+ },
+ {
+ "epoch": 59.58689956331878,
+ "grad_norm": 0.01290759164839983,
+ "learning_rate": 0.0006,
+ "loss": 4.437100410461426,
+ "step": 4290
+ },
+ {
+ "epoch": 59.60087336244541,
+ "grad_norm": 0.012042169459164143,
+ "learning_rate": 0.0006,
+ "loss": 4.567404747009277,
+ "step": 4291
+ },
+ {
+ "epoch": 59.61484716157205,
+ "grad_norm": 0.012338520959019661,
+ "learning_rate": 0.0006,
+ "loss": 4.476193428039551,
+ "step": 4292
+ },
+ {
+ "epoch": 59.62882096069869,
+ "grad_norm": 0.01321258582174778,
+ "learning_rate": 0.0006,
+ "loss": 4.566680431365967,
+ "step": 4293
+ },
+ {
+ "epoch": 59.64279475982533,
+ "grad_norm": 0.013478122651576996,
+ "learning_rate": 0.0006,
+ "loss": 4.487398147583008,
+ "step": 4294
+ },
+ {
+ "epoch": 59.65676855895197,
+ "grad_norm": 0.01537603884935379,
+ "learning_rate": 0.0006,
+ "loss": 4.449337005615234,
+ "step": 4295
+ },
+ {
+ "epoch": 59.670742358078606,
+ "grad_norm": 0.017140565440058708,
+ "learning_rate": 0.0006,
+ "loss": 4.5174360275268555,
+ "step": 4296
+ },
+ {
+ "epoch": 59.68471615720524,
+ "grad_norm": 0.017931461334228516,
+ "learning_rate": 0.0006,
+ "loss": 4.522104740142822,
+ "step": 4297
+ },
+ {
+ "epoch": 59.698689956331876,
+ "grad_norm": 0.018037918955087662,
+ "learning_rate": 0.0006,
+ "loss": 4.496928691864014,
+ "step": 4298
+ },
+ {
+ "epoch": 59.712663755458514,
+ "grad_norm": 0.01684504561126232,
+ "learning_rate": 0.0006,
+ "loss": 4.5332417488098145,
+ "step": 4299
+ },
+ {
+ "epoch": 59.72663755458515,
+ "grad_norm": 0.016268383711576462,
+ "learning_rate": 0.0006,
+ "loss": 4.468893051147461,
+ "step": 4300
+ },
+ {
+ "epoch": 59.74061135371179,
+ "grad_norm": 0.014859385788440704,
+ "learning_rate": 0.0006,
+ "loss": 4.551092147827148,
+ "step": 4301
+ },
+ {
+ "epoch": 59.75458515283843,
+ "grad_norm": 0.013589047826826572,
+ "learning_rate": 0.0006,
+ "loss": 4.470579147338867,
+ "step": 4302
+ },
+ {
+ "epoch": 59.76855895196506,
+ "grad_norm": 0.015704551711678505,
+ "learning_rate": 0.0006,
+ "loss": 4.330068111419678,
+ "step": 4303
+ },
+ {
+ "epoch": 59.7825327510917,
+ "grad_norm": 0.017869921401143074,
+ "learning_rate": 0.0006,
+ "loss": 4.428255081176758,
+ "step": 4304
+ },
+ {
+ "epoch": 59.79650655021834,
+ "grad_norm": 0.019828537479043007,
+ "learning_rate": 0.0006,
+ "loss": 4.496023178100586,
+ "step": 4305
+ },
+ {
+ "epoch": 59.81048034934498,
+ "grad_norm": 0.015856236219406128,
+ "learning_rate": 0.0006,
+ "loss": 4.427152156829834,
+ "step": 4306
+ },
+ {
+ "epoch": 59.82445414847162,
+ "grad_norm": 0.013349304907023907,
+ "learning_rate": 0.0006,
+ "loss": 4.417333602905273,
+ "step": 4307
+ },
+ {
+ "epoch": 59.838427947598255,
+ "grad_norm": 0.014111637137830257,
+ "learning_rate": 0.0006,
+ "loss": 4.4507575035095215,
+ "step": 4308
+ },
+ {
+ "epoch": 59.852401746724894,
+ "grad_norm": 0.015186597593128681,
+ "learning_rate": 0.0006,
+ "loss": 4.375116348266602,
+ "step": 4309
+ },
+ {
+ "epoch": 59.866375545851525,
+ "grad_norm": 0.01604730449616909,
+ "learning_rate": 0.0006,
+ "loss": 4.604518890380859,
+ "step": 4310
+ },
+ {
+ "epoch": 59.880349344978164,
+ "grad_norm": 0.015308464877307415,
+ "learning_rate": 0.0006,
+ "loss": 4.507232189178467,
+ "step": 4311
+ },
+ {
+ "epoch": 59.8943231441048,
+ "grad_norm": 0.014017206616699696,
+ "learning_rate": 0.0006,
+ "loss": 4.398487091064453,
+ "step": 4312
+ },
+ {
+ "epoch": 59.90829694323144,
+ "grad_norm": 0.012911394238471985,
+ "learning_rate": 0.0006,
+ "loss": 4.567872047424316,
+ "step": 4313
+ },
+ {
+ "epoch": 59.92227074235808,
+ "grad_norm": 0.012775142677128315,
+ "learning_rate": 0.0006,
+ "loss": 4.562105655670166,
+ "step": 4314
+ },
+ {
+ "epoch": 59.93624454148472,
+ "grad_norm": 0.014750408008694649,
+ "learning_rate": 0.0006,
+ "loss": 4.456263542175293,
+ "step": 4315
+ },
+ {
+ "epoch": 59.95021834061135,
+ "grad_norm": 0.016726011410355568,
+ "learning_rate": 0.0006,
+ "loss": 4.360722064971924,
+ "step": 4316
+ },
+ {
+ "epoch": 59.96419213973799,
+ "grad_norm": 0.01734776981174946,
+ "learning_rate": 0.0006,
+ "loss": 4.572544574737549,
+ "step": 4317
+ },
+ {
+ "epoch": 59.97816593886463,
+ "grad_norm": 0.01822415366768837,
+ "learning_rate": 0.0006,
+ "loss": 4.490501880645752,
+ "step": 4318
+ },
+ {
+ "epoch": 59.992139737991266,
+ "grad_norm": 0.014638577587902546,
+ "learning_rate": 0.0006,
+ "loss": 4.440951824188232,
+ "step": 4319
+ },
+ {
+ "epoch": 60.0,
+ "grad_norm": 0.01255202479660511,
+ "learning_rate": 0.0006,
+ "loss": 4.559123516082764,
+ "step": 4320
+ },
+ {
+ "epoch": 60.0,
+ "eval_loss": 4.741089344024658,
+ "eval_runtime": 56.8997,
+ "eval_samples_per_second": 42.918,
+ "eval_steps_per_second": 1.353,
+ "step": 4320
+ },
+ {
+ "epoch": 60.01397379912664,
+ "grad_norm": 0.016548393294215202,
+ "learning_rate": 0.0006,
+ "loss": 4.442024230957031,
+ "step": 4321
+ },
+ {
+ "epoch": 60.02794759825328,
+ "grad_norm": 0.019759127870202065,
+ "learning_rate": 0.0006,
+ "loss": 4.394334316253662,
+ "step": 4322
+ },
+ {
+ "epoch": 60.041921397379916,
+ "grad_norm": 0.018325170502066612,
+ "learning_rate": 0.0006,
+ "loss": 4.318270683288574,
+ "step": 4323
+ },
+ {
+ "epoch": 60.05589519650655,
+ "grad_norm": 0.015289463102817535,
+ "learning_rate": 0.0006,
+ "loss": 4.318228721618652,
+ "step": 4324
+ },
+ {
+ "epoch": 60.069868995633186,
+ "grad_norm": 0.015424872748553753,
+ "learning_rate": 0.0006,
+ "loss": 4.3605499267578125,
+ "step": 4325
+ },
+ {
+ "epoch": 60.083842794759825,
+ "grad_norm": 0.014085669070482254,
+ "learning_rate": 0.0006,
+ "loss": 4.359585762023926,
+ "step": 4326
+ },
+ {
+ "epoch": 60.09781659388646,
+ "grad_norm": 0.014004210010170937,
+ "learning_rate": 0.0006,
+ "loss": 4.447783470153809,
+ "step": 4327
+ },
+ {
+ "epoch": 60.1117903930131,
+ "grad_norm": 0.013514582999050617,
+ "learning_rate": 0.0006,
+ "loss": 4.4850239753723145,
+ "step": 4328
+ },
+ {
+ "epoch": 60.12576419213974,
+ "grad_norm": 0.013049394823610783,
+ "learning_rate": 0.0006,
+ "loss": 4.528675556182861,
+ "step": 4329
+ },
+ {
+ "epoch": 60.13973799126637,
+ "grad_norm": 0.01378058921545744,
+ "learning_rate": 0.0006,
+ "loss": 4.444823741912842,
+ "step": 4330
+ },
+ {
+ "epoch": 60.15371179039301,
+ "grad_norm": 0.013752233237028122,
+ "learning_rate": 0.0006,
+ "loss": 4.5205206871032715,
+ "step": 4331
+ },
+ {
+ "epoch": 60.16768558951965,
+ "grad_norm": 0.013431803323328495,
+ "learning_rate": 0.0006,
+ "loss": 4.547320365905762,
+ "step": 4332
+ },
+ {
+ "epoch": 60.18165938864629,
+ "grad_norm": 0.014544196426868439,
+ "learning_rate": 0.0006,
+ "loss": 4.517143249511719,
+ "step": 4333
+ },
+ {
+ "epoch": 60.19563318777293,
+ "grad_norm": 0.01612176187336445,
+ "learning_rate": 0.0006,
+ "loss": 4.517232894897461,
+ "step": 4334
+ },
+ {
+ "epoch": 60.209606986899566,
+ "grad_norm": 0.01803239807486534,
+ "learning_rate": 0.0006,
+ "loss": 4.377574920654297,
+ "step": 4335
+ },
+ {
+ "epoch": 60.223580786026204,
+ "grad_norm": 0.014871489256620407,
+ "learning_rate": 0.0006,
+ "loss": 4.538285255432129,
+ "step": 4336
+ },
+ {
+ "epoch": 60.237554585152836,
+ "grad_norm": 0.013927377760410309,
+ "learning_rate": 0.0006,
+ "loss": 4.386175632476807,
+ "step": 4337
+ },
+ {
+ "epoch": 60.251528384279474,
+ "grad_norm": 0.014757219702005386,
+ "learning_rate": 0.0006,
+ "loss": 4.528553009033203,
+ "step": 4338
+ },
+ {
+ "epoch": 60.26550218340611,
+ "grad_norm": 0.01433873176574707,
+ "learning_rate": 0.0006,
+ "loss": 4.476929187774658,
+ "step": 4339
+ },
+ {
+ "epoch": 60.27947598253275,
+ "grad_norm": 0.015768803656101227,
+ "learning_rate": 0.0006,
+ "loss": 4.50970458984375,
+ "step": 4340
+ },
+ {
+ "epoch": 60.29344978165939,
+ "grad_norm": 0.015651429072022438,
+ "learning_rate": 0.0006,
+ "loss": 4.427133083343506,
+ "step": 4341
+ },
+ {
+ "epoch": 60.30742358078603,
+ "grad_norm": 0.01673000678420067,
+ "learning_rate": 0.0006,
+ "loss": 4.447723865509033,
+ "step": 4342
+ },
+ {
+ "epoch": 60.32139737991266,
+ "grad_norm": 0.018481194972991943,
+ "learning_rate": 0.0006,
+ "loss": 4.559832572937012,
+ "step": 4343
+ },
+ {
+ "epoch": 60.3353711790393,
+ "grad_norm": 0.016731027513742447,
+ "learning_rate": 0.0006,
+ "loss": 4.470893859863281,
+ "step": 4344
+ },
+ {
+ "epoch": 60.34934497816594,
+ "grad_norm": 0.01504011545330286,
+ "learning_rate": 0.0006,
+ "loss": 4.483014106750488,
+ "step": 4345
+ },
+ {
+ "epoch": 60.36331877729258,
+ "grad_norm": 0.01602456159889698,
+ "learning_rate": 0.0006,
+ "loss": 4.548428058624268,
+ "step": 4346
+ },
+ {
+ "epoch": 60.377292576419215,
+ "grad_norm": 0.016027364879846573,
+ "learning_rate": 0.0006,
+ "loss": 4.426779747009277,
+ "step": 4347
+ },
+ {
+ "epoch": 60.391266375545854,
+ "grad_norm": 0.01873827911913395,
+ "learning_rate": 0.0006,
+ "loss": 4.353686332702637,
+ "step": 4348
+ },
+ {
+ "epoch": 60.40524017467249,
+ "grad_norm": 0.01907406374812126,
+ "learning_rate": 0.0006,
+ "loss": 4.502945423126221,
+ "step": 4349
+ },
+ {
+ "epoch": 60.419213973799124,
+ "grad_norm": 0.016083354130387306,
+ "learning_rate": 0.0006,
+ "loss": 4.431731700897217,
+ "step": 4350
+ },
+ {
+ "epoch": 60.43318777292576,
+ "grad_norm": 0.015293709933757782,
+ "learning_rate": 0.0006,
+ "loss": 4.480112552642822,
+ "step": 4351
+ },
+ {
+ "epoch": 60.4471615720524,
+ "grad_norm": 0.01688101328909397,
+ "learning_rate": 0.0006,
+ "loss": 4.50853157043457,
+ "step": 4352
+ },
+ {
+ "epoch": 60.46113537117904,
+ "grad_norm": 0.02094118855893612,
+ "learning_rate": 0.0006,
+ "loss": 4.359930038452148,
+ "step": 4353
+ },
+ {
+ "epoch": 60.47510917030568,
+ "grad_norm": 0.018188107758760452,
+ "learning_rate": 0.0006,
+ "loss": 4.549592018127441,
+ "step": 4354
+ },
+ {
+ "epoch": 60.48908296943232,
+ "grad_norm": 0.01462319865822792,
+ "learning_rate": 0.0006,
+ "loss": 4.49467658996582,
+ "step": 4355
+ },
+ {
+ "epoch": 60.50305676855895,
+ "grad_norm": 0.017508579418063164,
+ "learning_rate": 0.0006,
+ "loss": 4.398318767547607,
+ "step": 4356
+ },
+ {
+ "epoch": 60.51703056768559,
+ "grad_norm": 0.018200315535068512,
+ "learning_rate": 0.0006,
+ "loss": 4.44169282913208,
+ "step": 4357
+ },
+ {
+ "epoch": 60.531004366812226,
+ "grad_norm": 0.01717468351125717,
+ "learning_rate": 0.0006,
+ "loss": 4.481400489807129,
+ "step": 4358
+ },
+ {
+ "epoch": 60.544978165938865,
+ "grad_norm": 0.019329151138663292,
+ "learning_rate": 0.0006,
+ "loss": 4.578839302062988,
+ "step": 4359
+ },
+ {
+ "epoch": 60.5589519650655,
+ "grad_norm": 0.016109555959701538,
+ "learning_rate": 0.0006,
+ "loss": 4.461043357849121,
+ "step": 4360
+ },
+ {
+ "epoch": 60.57292576419214,
+ "grad_norm": 0.01494457945227623,
+ "learning_rate": 0.0006,
+ "loss": 4.505096912384033,
+ "step": 4361
+ },
+ {
+ "epoch": 60.58689956331878,
+ "grad_norm": 0.01729399710893631,
+ "learning_rate": 0.0006,
+ "loss": 4.366747856140137,
+ "step": 4362
+ },
+ {
+ "epoch": 60.60087336244541,
+ "grad_norm": 0.01705821603536606,
+ "learning_rate": 0.0006,
+ "loss": 4.42534065246582,
+ "step": 4363
+ },
+ {
+ "epoch": 60.61484716157205,
+ "grad_norm": 0.018391354009509087,
+ "learning_rate": 0.0006,
+ "loss": 4.350294589996338,
+ "step": 4364
+ },
+ {
+ "epoch": 60.62882096069869,
+ "grad_norm": 0.016989829018712044,
+ "learning_rate": 0.0006,
+ "loss": 4.45905876159668,
+ "step": 4365
+ },
+ {
+ "epoch": 60.64279475982533,
+ "grad_norm": 0.017192212864756584,
+ "learning_rate": 0.0006,
+ "loss": 4.526487350463867,
+ "step": 4366
+ },
+ {
+ "epoch": 60.65676855895197,
+ "grad_norm": 0.018091315403580666,
+ "learning_rate": 0.0006,
+ "loss": 4.309333324432373,
+ "step": 4367
+ },
+ {
+ "epoch": 60.670742358078606,
+ "grad_norm": 0.014051459729671478,
+ "learning_rate": 0.0006,
+ "loss": 4.338171005249023,
+ "step": 4368
+ },
+ {
+ "epoch": 60.68471615720524,
+ "grad_norm": 0.013166418299078941,
+ "learning_rate": 0.0006,
+ "loss": 4.409758567810059,
+ "step": 4369
+ },
+ {
+ "epoch": 60.698689956331876,
+ "grad_norm": 0.014260428957641125,
+ "learning_rate": 0.0006,
+ "loss": 4.524338245391846,
+ "step": 4370
+ },
+ {
+ "epoch": 60.712663755458514,
+ "grad_norm": 0.017019255086779594,
+ "learning_rate": 0.0006,
+ "loss": 4.387436866760254,
+ "step": 4371
+ },
+ {
+ "epoch": 60.72663755458515,
+ "grad_norm": 0.017070675268769264,
+ "learning_rate": 0.0006,
+ "loss": 4.437321662902832,
+ "step": 4372
+ },
+ {
+ "epoch": 60.74061135371179,
+ "grad_norm": 0.015197164379060268,
+ "learning_rate": 0.0006,
+ "loss": 4.371205806732178,
+ "step": 4373
+ },
+ {
+ "epoch": 60.75458515283843,
+ "grad_norm": 0.014791185967624187,
+ "learning_rate": 0.0006,
+ "loss": 4.44439697265625,
+ "step": 4374
+ },
+ {
+ "epoch": 60.76855895196506,
+ "grad_norm": 0.016458261758089066,
+ "learning_rate": 0.0006,
+ "loss": 4.445584297180176,
+ "step": 4375
+ },
+ {
+ "epoch": 60.7825327510917,
+ "grad_norm": 0.017706342041492462,
+ "learning_rate": 0.0006,
+ "loss": 4.348708152770996,
+ "step": 4376
+ },
+ {
+ "epoch": 60.79650655021834,
+ "grad_norm": 0.017561476677656174,
+ "learning_rate": 0.0006,
+ "loss": 4.521929740905762,
+ "step": 4377
+ },
+ {
+ "epoch": 60.81048034934498,
+ "grad_norm": 0.016557445749640465,
+ "learning_rate": 0.0006,
+ "loss": 4.545806884765625,
+ "step": 4378
+ },
+ {
+ "epoch": 60.82445414847162,
+ "grad_norm": 0.016953222453594208,
+ "learning_rate": 0.0006,
+ "loss": 4.4230875968933105,
+ "step": 4379
+ },
+ {
+ "epoch": 60.838427947598255,
+ "grad_norm": 0.017094967886805534,
+ "learning_rate": 0.0006,
+ "loss": 4.423985481262207,
+ "step": 4380
+ },
+ {
+ "epoch": 60.852401746724894,
+ "grad_norm": 0.014787515625357628,
+ "learning_rate": 0.0006,
+ "loss": 4.39473819732666,
+ "step": 4381
+ },
+ {
+ "epoch": 60.866375545851525,
+ "grad_norm": 0.016208263114094734,
+ "learning_rate": 0.0006,
+ "loss": 4.3810038566589355,
+ "step": 4382
+ },
+ {
+ "epoch": 60.880349344978164,
+ "grad_norm": 0.017844321206212044,
+ "learning_rate": 0.0006,
+ "loss": 4.35086727142334,
+ "step": 4383
+ },
+ {
+ "epoch": 60.8943231441048,
+ "grad_norm": 0.017539294436573982,
+ "learning_rate": 0.0006,
+ "loss": 4.460260391235352,
+ "step": 4384
+ },
+ {
+ "epoch": 60.90829694323144,
+ "grad_norm": 0.01612042263150215,
+ "learning_rate": 0.0006,
+ "loss": 4.370500564575195,
+ "step": 4385
+ },
+ {
+ "epoch": 60.92227074235808,
+ "grad_norm": 0.015381601639091969,
+ "learning_rate": 0.0006,
+ "loss": 4.435868263244629,
+ "step": 4386
+ },
+ {
+ "epoch": 60.93624454148472,
+ "grad_norm": 0.01603585295379162,
+ "learning_rate": 0.0006,
+ "loss": 4.434070587158203,
+ "step": 4387
+ },
+ {
+ "epoch": 60.95021834061135,
+ "grad_norm": 0.016268130391836166,
+ "learning_rate": 0.0006,
+ "loss": 4.404343128204346,
+ "step": 4388
+ },
+ {
+ "epoch": 60.96419213973799,
+ "grad_norm": 0.014980032108724117,
+ "learning_rate": 0.0006,
+ "loss": 4.399213790893555,
+ "step": 4389
+ },
+ {
+ "epoch": 60.97816593886463,
+ "grad_norm": 0.016899267211556435,
+ "learning_rate": 0.0006,
+ "loss": 4.508069038391113,
+ "step": 4390
+ },
+ {
+ "epoch": 60.992139737991266,
+ "grad_norm": 0.013756908476352692,
+ "learning_rate": 0.0006,
+ "loss": 4.541858673095703,
+ "step": 4391
+ },
+ {
+ "epoch": 61.0,
+ "grad_norm": 0.01578413136303425,
+ "learning_rate": 0.0006,
+ "loss": 4.543513298034668,
+ "step": 4392
+ },
+ {
+ "epoch": 61.0,
+ "eval_loss": 4.756770610809326,
+ "eval_runtime": 56.4572,
+ "eval_samples_per_second": 43.254,
+ "eval_steps_per_second": 1.364,
+ "step": 4392
+ },
+ {
+ "epoch": 61.01397379912664,
+ "grad_norm": 0.013387506827712059,
+ "learning_rate": 0.0006,
+ "loss": 4.475942611694336,
+ "step": 4393
+ },
+ {
+ "epoch": 61.02794759825328,
+ "grad_norm": 0.014278904534876347,
+ "learning_rate": 0.0006,
+ "loss": 4.442403793334961,
+ "step": 4394
+ },
+ {
+ "epoch": 61.041921397379916,
+ "grad_norm": 0.014359553344547749,
+ "learning_rate": 0.0006,
+ "loss": 4.509463310241699,
+ "step": 4395
+ },
+ {
+ "epoch": 61.05589519650655,
+ "grad_norm": 0.01718050241470337,
+ "learning_rate": 0.0006,
+ "loss": 4.354458808898926,
+ "step": 4396
+ },
+ {
+ "epoch": 61.069868995633186,
+ "grad_norm": 0.015988312661647797,
+ "learning_rate": 0.0006,
+ "loss": 4.484508514404297,
+ "step": 4397
+ },
+ {
+ "epoch": 61.083842794759825,
+ "grad_norm": 0.015466735698282719,
+ "learning_rate": 0.0006,
+ "loss": 4.326908111572266,
+ "step": 4398
+ },
+ {
+ "epoch": 61.09781659388646,
+ "grad_norm": 0.014541332609951496,
+ "learning_rate": 0.0006,
+ "loss": 4.405823230743408,
+ "step": 4399
+ },
+ {
+ "epoch": 61.1117903930131,
+ "grad_norm": 0.014803987927734852,
+ "learning_rate": 0.0006,
+ "loss": 4.375386714935303,
+ "step": 4400
+ },
+ {
+ "epoch": 61.12576419213974,
+ "grad_norm": 0.013962093740701675,
+ "learning_rate": 0.0006,
+ "loss": 4.300838470458984,
+ "step": 4401
+ },
+ {
+ "epoch": 61.13973799126637,
+ "grad_norm": 0.015213954262435436,
+ "learning_rate": 0.0006,
+ "loss": 4.388306617736816,
+ "step": 4402
+ },
+ {
+ "epoch": 61.15371179039301,
+ "grad_norm": 0.018965506926178932,
+ "learning_rate": 0.0006,
+ "loss": 4.388340950012207,
+ "step": 4403
+ },
+ {
+ "epoch": 61.16768558951965,
+ "grad_norm": 0.01958410255610943,
+ "learning_rate": 0.0006,
+ "loss": 4.35012149810791,
+ "step": 4404
+ },
+ {
+ "epoch": 61.18165938864629,
+ "grad_norm": 0.01938101276755333,
+ "learning_rate": 0.0006,
+ "loss": 4.368372917175293,
+ "step": 4405
+ },
+ {
+ "epoch": 61.19563318777293,
+ "grad_norm": 0.016556516289711,
+ "learning_rate": 0.0006,
+ "loss": 4.486198425292969,
+ "step": 4406
+ },
+ {
+ "epoch": 61.209606986899566,
+ "grad_norm": 0.017031289637088776,
+ "learning_rate": 0.0006,
+ "loss": 4.421676158905029,
+ "step": 4407
+ },
+ {
+ "epoch": 61.223580786026204,
+ "grad_norm": 0.018023479729890823,
+ "learning_rate": 0.0006,
+ "loss": 4.38751220703125,
+ "step": 4408
+ },
+ {
+ "epoch": 61.237554585152836,
+ "grad_norm": 0.0178135447204113,
+ "learning_rate": 0.0006,
+ "loss": 4.336528301239014,
+ "step": 4409
+ },
+ {
+ "epoch": 61.251528384279474,
+ "grad_norm": 0.016782190650701523,
+ "learning_rate": 0.0006,
+ "loss": 4.386693954467773,
+ "step": 4410
+ },
+ {
+ "epoch": 61.26550218340611,
+ "grad_norm": 0.015494848601520061,
+ "learning_rate": 0.0006,
+ "loss": 4.308948040008545,
+ "step": 4411
+ },
+ {
+ "epoch": 61.27947598253275,
+ "grad_norm": 0.015002378262579441,
+ "learning_rate": 0.0006,
+ "loss": 4.314206123352051,
+ "step": 4412
+ },
+ {
+ "epoch": 61.29344978165939,
+ "grad_norm": 0.015798920765519142,
+ "learning_rate": 0.0006,
+ "loss": 4.502900123596191,
+ "step": 4413
+ },
+ {
+ "epoch": 61.30742358078603,
+ "grad_norm": 0.015444336459040642,
+ "learning_rate": 0.0006,
+ "loss": 4.473160743713379,
+ "step": 4414
+ },
+ {
+ "epoch": 61.32139737991266,
+ "grad_norm": 0.017025096341967583,
+ "learning_rate": 0.0006,
+ "loss": 4.460621356964111,
+ "step": 4415
+ },
+ {
+ "epoch": 61.3353711790393,
+ "grad_norm": 0.019502537325024605,
+ "learning_rate": 0.0006,
+ "loss": 4.385551452636719,
+ "step": 4416
+ },
+ {
+ "epoch": 61.34934497816594,
+ "grad_norm": 0.018529541790485382,
+ "learning_rate": 0.0006,
+ "loss": 4.425390720367432,
+ "step": 4417
+ },
+ {
+ "epoch": 61.36331877729258,
+ "grad_norm": 0.013155706226825714,
+ "learning_rate": 0.0006,
+ "loss": 4.467020034790039,
+ "step": 4418
+ },
+ {
+ "epoch": 61.377292576419215,
+ "grad_norm": 0.01518090907484293,
+ "learning_rate": 0.0006,
+ "loss": 4.369664192199707,
+ "step": 4419
+ },
+ {
+ "epoch": 61.391266375545854,
+ "grad_norm": 0.01626511849462986,
+ "learning_rate": 0.0006,
+ "loss": 4.463231086730957,
+ "step": 4420
+ },
+ {
+ "epoch": 61.40524017467249,
+ "grad_norm": 0.01701558195054531,
+ "learning_rate": 0.0006,
+ "loss": 4.43708610534668,
+ "step": 4421
+ },
+ {
+ "epoch": 61.419213973799124,
+ "grad_norm": 0.01761862449347973,
+ "learning_rate": 0.0006,
+ "loss": 4.3314208984375,
+ "step": 4422
+ },
+ {
+ "epoch": 61.43318777292576,
+ "grad_norm": 0.017084648832678795,
+ "learning_rate": 0.0006,
+ "loss": 4.39676570892334,
+ "step": 4423
+ },
+ {
+ "epoch": 61.4471615720524,
+ "grad_norm": 0.01667078025639057,
+ "learning_rate": 0.0006,
+ "loss": 4.3789873123168945,
+ "step": 4424
+ },
+ {
+ "epoch": 61.46113537117904,
+ "grad_norm": 0.01877436414361,
+ "learning_rate": 0.0006,
+ "loss": 4.415275573730469,
+ "step": 4425
+ },
+ {
+ "epoch": 61.47510917030568,
+ "grad_norm": 0.020250486209988594,
+ "learning_rate": 0.0006,
+ "loss": 4.478635787963867,
+ "step": 4426
+ },
+ {
+ "epoch": 61.48908296943232,
+ "grad_norm": 0.020244868472218513,
+ "learning_rate": 0.0006,
+ "loss": 4.296854019165039,
+ "step": 4427
+ },
+ {
+ "epoch": 61.50305676855895,
+ "grad_norm": 0.018723875284194946,
+ "learning_rate": 0.0006,
+ "loss": 4.3760881423950195,
+ "step": 4428
+ },
+ {
+ "epoch": 61.51703056768559,
+ "grad_norm": 0.015138164162635803,
+ "learning_rate": 0.0006,
+ "loss": 4.39974308013916,
+ "step": 4429
+ },
+ {
+ "epoch": 61.531004366812226,
+ "grad_norm": 0.015111138112843037,
+ "learning_rate": 0.0006,
+ "loss": 4.34659481048584,
+ "step": 4430
+ },
+ {
+ "epoch": 61.544978165938865,
+ "grad_norm": 0.0164666585624218,
+ "learning_rate": 0.0006,
+ "loss": 4.415737152099609,
+ "step": 4431
+ },
+ {
+ "epoch": 61.5589519650655,
+ "grad_norm": 0.020891312509775162,
+ "learning_rate": 0.0006,
+ "loss": 4.300319671630859,
+ "step": 4432
+ },
+ {
+ "epoch": 61.57292576419214,
+ "grad_norm": 0.02089156024158001,
+ "learning_rate": 0.0006,
+ "loss": 4.373029708862305,
+ "step": 4433
+ },
+ {
+ "epoch": 61.58689956331878,
+ "grad_norm": 0.020114559680223465,
+ "learning_rate": 0.0006,
+ "loss": 4.434499740600586,
+ "step": 4434
+ },
+ {
+ "epoch": 61.60087336244541,
+ "grad_norm": 0.01791365258395672,
+ "learning_rate": 0.0006,
+ "loss": 4.466872215270996,
+ "step": 4435
+ },
+ {
+ "epoch": 61.61484716157205,
+ "grad_norm": 0.01879458874464035,
+ "learning_rate": 0.0006,
+ "loss": 4.5176520347595215,
+ "step": 4436
+ },
+ {
+ "epoch": 61.62882096069869,
+ "grad_norm": 0.016371937468647957,
+ "learning_rate": 0.0006,
+ "loss": 4.3507585525512695,
+ "step": 4437
+ },
+ {
+ "epoch": 61.64279475982533,
+ "grad_norm": 0.01648143120110035,
+ "learning_rate": 0.0006,
+ "loss": 4.398778438568115,
+ "step": 4438
+ },
+ {
+ "epoch": 61.65676855895197,
+ "grad_norm": 0.017240576446056366,
+ "learning_rate": 0.0006,
+ "loss": 4.513933181762695,
+ "step": 4439
+ },
+ {
+ "epoch": 61.670742358078606,
+ "grad_norm": 0.01734284684062004,
+ "learning_rate": 0.0006,
+ "loss": 4.315343856811523,
+ "step": 4440
+ },
+ {
+ "epoch": 61.68471615720524,
+ "grad_norm": 0.01639772206544876,
+ "learning_rate": 0.0006,
+ "loss": 4.4641218185424805,
+ "step": 4441
+ },
+ {
+ "epoch": 61.698689956331876,
+ "grad_norm": 0.0164708960801363,
+ "learning_rate": 0.0006,
+ "loss": 4.439968585968018,
+ "step": 4442
+ },
+ {
+ "epoch": 61.712663755458514,
+ "grad_norm": 0.01757933758199215,
+ "learning_rate": 0.0006,
+ "loss": 4.408327102661133,
+ "step": 4443
+ },
+ {
+ "epoch": 61.72663755458515,
+ "grad_norm": 0.01789589412510395,
+ "learning_rate": 0.0006,
+ "loss": 4.478257179260254,
+ "step": 4444
+ },
+ {
+ "epoch": 61.74061135371179,
+ "grad_norm": 0.01822184957563877,
+ "learning_rate": 0.0006,
+ "loss": 4.412046909332275,
+ "step": 4445
+ },
+ {
+ "epoch": 61.75458515283843,
+ "grad_norm": 0.018624618649482727,
+ "learning_rate": 0.0006,
+ "loss": 4.427122116088867,
+ "step": 4446
+ },
+ {
+ "epoch": 61.76855895196506,
+ "grad_norm": 0.018674887716770172,
+ "learning_rate": 0.0006,
+ "loss": 4.429147720336914,
+ "step": 4447
+ },
+ {
+ "epoch": 61.7825327510917,
+ "grad_norm": 0.015317119657993317,
+ "learning_rate": 0.0006,
+ "loss": 4.2808451652526855,
+ "step": 4448
+ },
+ {
+ "epoch": 61.79650655021834,
+ "grad_norm": 0.015400845557451248,
+ "learning_rate": 0.0006,
+ "loss": 4.420193672180176,
+ "step": 4449
+ },
+ {
+ "epoch": 61.81048034934498,
+ "grad_norm": 0.016743650659918785,
+ "learning_rate": 0.0006,
+ "loss": 4.498254776000977,
+ "step": 4450
+ },
+ {
+ "epoch": 61.82445414847162,
+ "grad_norm": 0.014330781064927578,
+ "learning_rate": 0.0006,
+ "loss": 4.445384979248047,
+ "step": 4451
+ },
+ {
+ "epoch": 61.838427947598255,
+ "grad_norm": 0.015046479180455208,
+ "learning_rate": 0.0006,
+ "loss": 4.368041038513184,
+ "step": 4452
+ },
+ {
+ "epoch": 61.852401746724894,
+ "grad_norm": 0.01630168780684471,
+ "learning_rate": 0.0006,
+ "loss": 4.460142135620117,
+ "step": 4453
+ },
+ {
+ "epoch": 61.866375545851525,
+ "grad_norm": 0.01637885719537735,
+ "learning_rate": 0.0006,
+ "loss": 4.416606426239014,
+ "step": 4454
+ },
+ {
+ "epoch": 61.880349344978164,
+ "grad_norm": 0.016785888001322746,
+ "learning_rate": 0.0006,
+ "loss": 4.452545642852783,
+ "step": 4455
+ },
+ {
+ "epoch": 61.8943231441048,
+ "grad_norm": 0.015044555068016052,
+ "learning_rate": 0.0006,
+ "loss": 4.427334785461426,
+ "step": 4456
+ },
+ {
+ "epoch": 61.90829694323144,
+ "grad_norm": 0.013699211180210114,
+ "learning_rate": 0.0006,
+ "loss": 4.408761024475098,
+ "step": 4457
+ },
+ {
+ "epoch": 61.92227074235808,
+ "grad_norm": 0.014021803624927998,
+ "learning_rate": 0.0006,
+ "loss": 4.543607711791992,
+ "step": 4458
+ },
+ {
+ "epoch": 61.93624454148472,
+ "grad_norm": 0.015438460744917393,
+ "learning_rate": 0.0006,
+ "loss": 4.313363075256348,
+ "step": 4459
+ },
+ {
+ "epoch": 61.95021834061135,
+ "grad_norm": 0.017515428364276886,
+ "learning_rate": 0.0006,
+ "loss": 4.310412406921387,
+ "step": 4460
+ },
+ {
+ "epoch": 61.96419213973799,
+ "grad_norm": 0.016683880239725113,
+ "learning_rate": 0.0006,
+ "loss": 4.3378448486328125,
+ "step": 4461
+ },
+ {
+ "epoch": 61.97816593886463,
+ "grad_norm": 0.019384048879146576,
+ "learning_rate": 0.0006,
+ "loss": 4.450568675994873,
+ "step": 4462
+ },
+ {
+ "epoch": 61.992139737991266,
+ "grad_norm": 0.020651493221521378,
+ "learning_rate": 0.0006,
+ "loss": 4.453197479248047,
+ "step": 4463
+ },
+ {
+ "epoch": 62.0,
+ "grad_norm": 0.020991764962673187,
+ "learning_rate": 0.0006,
+ "loss": 4.441527366638184,
+ "step": 4464
+ },
+ {
+ "epoch": 62.0,
+ "eval_loss": 4.700193881988525,
+ "eval_runtime": 56.6098,
+ "eval_samples_per_second": 43.137,
+ "eval_steps_per_second": 1.36,
+ "step": 4464
+ },
+ {
+ "epoch": 62.01397379912664,
+ "grad_norm": 0.01811373233795166,
+ "learning_rate": 0.0006,
+ "loss": 4.37346076965332,
+ "step": 4465
+ },
+ {
+ "epoch": 62.02794759825328,
+ "grad_norm": 0.01667127199470997,
+ "learning_rate": 0.0006,
+ "loss": 4.237151145935059,
+ "step": 4466
+ },
+ {
+ "epoch": 62.041921397379916,
+ "grad_norm": 0.014971534721553326,
+ "learning_rate": 0.0006,
+ "loss": 4.367053031921387,
+ "step": 4467
+ },
+ {
+ "epoch": 62.05589519650655,
+ "grad_norm": 0.015655893832445145,
+ "learning_rate": 0.0006,
+ "loss": 4.437426567077637,
+ "step": 4468
+ },
+ {
+ "epoch": 62.069868995633186,
+ "grad_norm": 0.014935447834432125,
+ "learning_rate": 0.0006,
+ "loss": 4.3635125160217285,
+ "step": 4469
+ },
+ {
+ "epoch": 62.083842794759825,
+ "grad_norm": 0.015363802202045918,
+ "learning_rate": 0.0006,
+ "loss": 4.40492057800293,
+ "step": 4470
+ },
+ {
+ "epoch": 62.09781659388646,
+ "grad_norm": 0.0159201230853796,
+ "learning_rate": 0.0006,
+ "loss": 4.392058849334717,
+ "step": 4471
+ },
+ {
+ "epoch": 62.1117903930131,
+ "grad_norm": 0.014598241075873375,
+ "learning_rate": 0.0006,
+ "loss": 4.369944095611572,
+ "step": 4472
+ },
+ {
+ "epoch": 62.12576419213974,
+ "grad_norm": 0.015394841320812702,
+ "learning_rate": 0.0006,
+ "loss": 4.425097942352295,
+ "step": 4473
+ },
+ {
+ "epoch": 62.13973799126637,
+ "grad_norm": 0.015913628041744232,
+ "learning_rate": 0.0006,
+ "loss": 4.420596122741699,
+ "step": 4474
+ },
+ {
+ "epoch": 62.15371179039301,
+ "grad_norm": 0.01575258933007717,
+ "learning_rate": 0.0006,
+ "loss": 4.32290506362915,
+ "step": 4475
+ },
+ {
+ "epoch": 62.16768558951965,
+ "grad_norm": 0.0155344782397151,
+ "learning_rate": 0.0006,
+ "loss": 4.273068428039551,
+ "step": 4476
+ },
+ {
+ "epoch": 62.18165938864629,
+ "grad_norm": 0.014947040006518364,
+ "learning_rate": 0.0006,
+ "loss": 4.384868621826172,
+ "step": 4477
+ },
+ {
+ "epoch": 62.19563318777293,
+ "grad_norm": 0.01557591650635004,
+ "learning_rate": 0.0006,
+ "loss": 4.458410263061523,
+ "step": 4478
+ },
+ {
+ "epoch": 62.209606986899566,
+ "grad_norm": 0.015182198956608772,
+ "learning_rate": 0.0006,
+ "loss": 4.300738334655762,
+ "step": 4479
+ },
+ {
+ "epoch": 62.223580786026204,
+ "grad_norm": 0.01634717360138893,
+ "learning_rate": 0.0006,
+ "loss": 4.436875820159912,
+ "step": 4480
+ },
+ {
+ "epoch": 62.237554585152836,
+ "grad_norm": 0.018813718110322952,
+ "learning_rate": 0.0006,
+ "loss": 4.454665184020996,
+ "step": 4481
+ },
+ {
+ "epoch": 62.251528384279474,
+ "grad_norm": 0.020965630188584328,
+ "learning_rate": 0.0006,
+ "loss": 4.357341766357422,
+ "step": 4482
+ },
+ {
+ "epoch": 62.26550218340611,
+ "grad_norm": 0.020738353952765465,
+ "learning_rate": 0.0006,
+ "loss": 4.456644058227539,
+ "step": 4483
+ },
+ {
+ "epoch": 62.27947598253275,
+ "grad_norm": 0.018813546746969223,
+ "learning_rate": 0.0006,
+ "loss": 4.381941795349121,
+ "step": 4484
+ },
+ {
+ "epoch": 62.29344978165939,
+ "grad_norm": 0.015431761741638184,
+ "learning_rate": 0.0006,
+ "loss": 4.41126823425293,
+ "step": 4485
+ },
+ {
+ "epoch": 62.30742358078603,
+ "grad_norm": 0.016335977241396904,
+ "learning_rate": 0.0006,
+ "loss": 4.419488906860352,
+ "step": 4486
+ },
+ {
+ "epoch": 62.32139737991266,
+ "grad_norm": 0.01832406409084797,
+ "learning_rate": 0.0006,
+ "loss": 4.579026222229004,
+ "step": 4487
+ },
+ {
+ "epoch": 62.3353711790393,
+ "grad_norm": 0.01790742017328739,
+ "learning_rate": 0.0006,
+ "loss": 4.289964199066162,
+ "step": 4488
+ },
+ {
+ "epoch": 62.34934497816594,
+ "grad_norm": 0.017266202718019485,
+ "learning_rate": 0.0006,
+ "loss": 4.415959358215332,
+ "step": 4489
+ },
+ {
+ "epoch": 62.36331877729258,
+ "grad_norm": 0.015976132825016975,
+ "learning_rate": 0.0006,
+ "loss": 4.358587265014648,
+ "step": 4490
+ },
+ {
+ "epoch": 62.377292576419215,
+ "grad_norm": 0.016873881220817566,
+ "learning_rate": 0.0006,
+ "loss": 4.354607105255127,
+ "step": 4491
+ },
+ {
+ "epoch": 62.391266375545854,
+ "grad_norm": 0.01476745493710041,
+ "learning_rate": 0.0006,
+ "loss": 4.384904384613037,
+ "step": 4492
+ },
+ {
+ "epoch": 62.40524017467249,
+ "grad_norm": 0.013810782693326473,
+ "learning_rate": 0.0006,
+ "loss": 4.459230899810791,
+ "step": 4493
+ },
+ {
+ "epoch": 62.419213973799124,
+ "grad_norm": 0.015150421299040318,
+ "learning_rate": 0.0006,
+ "loss": 4.369471549987793,
+ "step": 4494
+ },
+ {
+ "epoch": 62.43318777292576,
+ "grad_norm": 0.014539610594511032,
+ "learning_rate": 0.0006,
+ "loss": 4.3564982414245605,
+ "step": 4495
+ },
+ {
+ "epoch": 62.4471615720524,
+ "grad_norm": 0.015554017387330532,
+ "learning_rate": 0.0006,
+ "loss": 4.414379119873047,
+ "step": 4496
+ },
+ {
+ "epoch": 62.46113537117904,
+ "grad_norm": 0.016608357429504395,
+ "learning_rate": 0.0006,
+ "loss": 4.311585426330566,
+ "step": 4497
+ },
+ {
+ "epoch": 62.47510917030568,
+ "grad_norm": 0.01885881833732128,
+ "learning_rate": 0.0006,
+ "loss": 4.306629180908203,
+ "step": 4498
+ },
+ {
+ "epoch": 62.48908296943232,
+ "grad_norm": 0.022128663957118988,
+ "learning_rate": 0.0006,
+ "loss": 4.345483779907227,
+ "step": 4499
+ },
+ {
+ "epoch": 62.50305676855895,
+ "grad_norm": 0.023401018232107162,
+ "learning_rate": 0.0006,
+ "loss": 4.398204326629639,
+ "step": 4500
+ },
+ {
+ "epoch": 62.51703056768559,
+ "grad_norm": 0.022080009803175926,
+ "learning_rate": 0.0006,
+ "loss": 4.375027179718018,
+ "step": 4501
+ },
+ {
+ "epoch": 62.531004366812226,
+ "grad_norm": 0.020004864782094955,
+ "learning_rate": 0.0006,
+ "loss": 4.33494758605957,
+ "step": 4502
+ },
+ {
+ "epoch": 62.544978165938865,
+ "grad_norm": 0.01786700263619423,
+ "learning_rate": 0.0006,
+ "loss": 4.4428300857543945,
+ "step": 4503
+ },
+ {
+ "epoch": 62.5589519650655,
+ "grad_norm": 0.01885388419032097,
+ "learning_rate": 0.0006,
+ "loss": 4.4138946533203125,
+ "step": 4504
+ },
+ {
+ "epoch": 62.57292576419214,
+ "grad_norm": 0.019890939816832542,
+ "learning_rate": 0.0006,
+ "loss": 4.278632164001465,
+ "step": 4505
+ },
+ {
+ "epoch": 62.58689956331878,
+ "grad_norm": 0.02109031192958355,
+ "learning_rate": 0.0006,
+ "loss": 4.3756327629089355,
+ "step": 4506
+ },
+ {
+ "epoch": 62.60087336244541,
+ "grad_norm": 0.020813899114727974,
+ "learning_rate": 0.0006,
+ "loss": 4.410213470458984,
+ "step": 4507
+ },
+ {
+ "epoch": 62.61484716157205,
+ "grad_norm": 0.021655689924955368,
+ "learning_rate": 0.0006,
+ "loss": 4.30633020401001,
+ "step": 4508
+ },
+ {
+ "epoch": 62.62882096069869,
+ "grad_norm": 0.018155649304389954,
+ "learning_rate": 0.0006,
+ "loss": 4.407284736633301,
+ "step": 4509
+ },
+ {
+ "epoch": 62.64279475982533,
+ "grad_norm": 0.017765069380402565,
+ "learning_rate": 0.0006,
+ "loss": 4.321715354919434,
+ "step": 4510
+ },
+ {
+ "epoch": 62.65676855895197,
+ "grad_norm": 0.017724554985761642,
+ "learning_rate": 0.0006,
+ "loss": 4.493371963500977,
+ "step": 4511
+ },
+ {
+ "epoch": 62.670742358078606,
+ "grad_norm": 0.01492092851549387,
+ "learning_rate": 0.0006,
+ "loss": 4.443025588989258,
+ "step": 4512
+ },
+ {
+ "epoch": 62.68471615720524,
+ "grad_norm": 0.01615324057638645,
+ "learning_rate": 0.0006,
+ "loss": 4.3289594650268555,
+ "step": 4513
+ },
+ {
+ "epoch": 62.698689956331876,
+ "grad_norm": 0.016569366678595543,
+ "learning_rate": 0.0006,
+ "loss": 4.369720458984375,
+ "step": 4514
+ },
+ {
+ "epoch": 62.712663755458514,
+ "grad_norm": 0.016509901732206345,
+ "learning_rate": 0.0006,
+ "loss": 4.292316436767578,
+ "step": 4515
+ },
+ {
+ "epoch": 62.72663755458515,
+ "grad_norm": 0.014270003885030746,
+ "learning_rate": 0.0006,
+ "loss": 4.349854469299316,
+ "step": 4516
+ },
+ {
+ "epoch": 62.74061135371179,
+ "grad_norm": 0.014235340990126133,
+ "learning_rate": 0.0006,
+ "loss": 4.473727226257324,
+ "step": 4517
+ },
+ {
+ "epoch": 62.75458515283843,
+ "grad_norm": 0.015163225121796131,
+ "learning_rate": 0.0006,
+ "loss": 4.452878475189209,
+ "step": 4518
+ },
+ {
+ "epoch": 62.76855895196506,
+ "grad_norm": 0.015748055651783943,
+ "learning_rate": 0.0006,
+ "loss": 4.521087169647217,
+ "step": 4519
+ },
+ {
+ "epoch": 62.7825327510917,
+ "grad_norm": 0.01661108434200287,
+ "learning_rate": 0.0006,
+ "loss": 4.443885326385498,
+ "step": 4520
+ },
+ {
+ "epoch": 62.79650655021834,
+ "grad_norm": 0.01610301062464714,
+ "learning_rate": 0.0006,
+ "loss": 4.312990188598633,
+ "step": 4521
+ },
+ {
+ "epoch": 62.81048034934498,
+ "grad_norm": 0.017007583752274513,
+ "learning_rate": 0.0006,
+ "loss": 4.408937454223633,
+ "step": 4522
+ },
+ {
+ "epoch": 62.82445414847162,
+ "grad_norm": 0.01749461516737938,
+ "learning_rate": 0.0006,
+ "loss": 4.346066951751709,
+ "step": 4523
+ },
+ {
+ "epoch": 62.838427947598255,
+ "grad_norm": 0.018243493512272835,
+ "learning_rate": 0.0006,
+ "loss": 4.357658386230469,
+ "step": 4524
+ },
+ {
+ "epoch": 62.852401746724894,
+ "grad_norm": 0.017499985173344612,
+ "learning_rate": 0.0006,
+ "loss": 4.374061584472656,
+ "step": 4525
+ },
+ {
+ "epoch": 62.866375545851525,
+ "grad_norm": 0.014536325819790363,
+ "learning_rate": 0.0006,
+ "loss": 4.189321994781494,
+ "step": 4526
+ },
+ {
+ "epoch": 62.880349344978164,
+ "grad_norm": 0.014728891663253307,
+ "learning_rate": 0.0006,
+ "loss": 4.425910949707031,
+ "step": 4527
+ },
+ {
+ "epoch": 62.8943231441048,
+ "grad_norm": 0.015013772062957287,
+ "learning_rate": 0.0006,
+ "loss": 4.337319850921631,
+ "step": 4528
+ },
+ {
+ "epoch": 62.90829694323144,
+ "grad_norm": 0.01626027375459671,
+ "learning_rate": 0.0006,
+ "loss": 4.330339431762695,
+ "step": 4529
+ },
+ {
+ "epoch": 62.92227074235808,
+ "grad_norm": 0.014773444272577763,
+ "learning_rate": 0.0006,
+ "loss": 4.4188737869262695,
+ "step": 4530
+ },
+ {
+ "epoch": 62.93624454148472,
+ "grad_norm": 0.01422932744026184,
+ "learning_rate": 0.0006,
+ "loss": 4.355986595153809,
+ "step": 4531
+ },
+ {
+ "epoch": 62.95021834061135,
+ "grad_norm": 0.012170110829174519,
+ "learning_rate": 0.0006,
+ "loss": 4.405722618103027,
+ "step": 4532
+ },
+ {
+ "epoch": 62.96419213973799,
+ "grad_norm": 0.013233253732323647,
+ "learning_rate": 0.0006,
+ "loss": 4.377286911010742,
+ "step": 4533
+ },
+ {
+ "epoch": 62.97816593886463,
+ "grad_norm": 0.014497299678623676,
+ "learning_rate": 0.0006,
+ "loss": 4.508796691894531,
+ "step": 4534
+ },
+ {
+ "epoch": 62.992139737991266,
+ "grad_norm": 0.014718741178512573,
+ "learning_rate": 0.0006,
+ "loss": 4.370919704437256,
+ "step": 4535
+ },
+ {
+ "epoch": 63.0,
+ "grad_norm": 0.015833502635359764,
+ "learning_rate": 0.0006,
+ "loss": 4.258753776550293,
+ "step": 4536
+ },
+ {
+ "epoch": 63.0,
+ "eval_loss": 4.714163303375244,
+ "eval_runtime": 57.0269,
+ "eval_samples_per_second": 42.822,
+ "eval_steps_per_second": 1.35,
+ "step": 4536
+ },
+ {
+ "epoch": 63.01397379912664,
+ "grad_norm": 0.014277510344982147,
+ "learning_rate": 0.0006,
+ "loss": 4.3112640380859375,
+ "step": 4537
+ },
+ {
+ "epoch": 63.02794759825328,
+ "grad_norm": 0.015377935022115707,
+ "learning_rate": 0.0006,
+ "loss": 4.342782974243164,
+ "step": 4538
+ },
+ {
+ "epoch": 63.041921397379916,
+ "grad_norm": 0.015406244434416294,
+ "learning_rate": 0.0006,
+ "loss": 4.474056720733643,
+ "step": 4539
+ },
+ {
+ "epoch": 63.05589519650655,
+ "grad_norm": 0.018121499568223953,
+ "learning_rate": 0.0006,
+ "loss": 4.397963523864746,
+ "step": 4540
+ },
+ {
+ "epoch": 63.069868995633186,
+ "grad_norm": 0.022634636610746384,
+ "learning_rate": 0.0006,
+ "loss": 4.306631088256836,
+ "step": 4541
+ },
+ {
+ "epoch": 63.083842794759825,
+ "grad_norm": 0.026448970660567284,
+ "learning_rate": 0.0006,
+ "loss": 4.283175468444824,
+ "step": 4542
+ },
+ {
+ "epoch": 63.09781659388646,
+ "grad_norm": 0.024104537442326546,
+ "learning_rate": 0.0006,
+ "loss": 4.315826416015625,
+ "step": 4543
+ },
+ {
+ "epoch": 63.1117903930131,
+ "grad_norm": 0.019040480256080627,
+ "learning_rate": 0.0006,
+ "loss": 4.44057035446167,
+ "step": 4544
+ },
+ {
+ "epoch": 63.12576419213974,
+ "grad_norm": 0.02006000280380249,
+ "learning_rate": 0.0006,
+ "loss": 4.405603408813477,
+ "step": 4545
+ },
+ {
+ "epoch": 63.13973799126637,
+ "grad_norm": 0.019980071112513542,
+ "learning_rate": 0.0006,
+ "loss": 4.393726348876953,
+ "step": 4546
+ },
+ {
+ "epoch": 63.15371179039301,
+ "grad_norm": 0.01641865260899067,
+ "learning_rate": 0.0006,
+ "loss": 4.272546768188477,
+ "step": 4547
+ },
+ {
+ "epoch": 63.16768558951965,
+ "grad_norm": 0.01890353485941887,
+ "learning_rate": 0.0006,
+ "loss": 4.382566928863525,
+ "step": 4548
+ },
+ {
+ "epoch": 63.18165938864629,
+ "grad_norm": 0.018370937556028366,
+ "learning_rate": 0.0006,
+ "loss": 4.423953533172607,
+ "step": 4549
+ },
+ {
+ "epoch": 63.19563318777293,
+ "grad_norm": 0.014985294081270695,
+ "learning_rate": 0.0006,
+ "loss": 4.308627128601074,
+ "step": 4550
+ },
+ {
+ "epoch": 63.209606986899566,
+ "grad_norm": 0.017513327300548553,
+ "learning_rate": 0.0006,
+ "loss": 4.248819351196289,
+ "step": 4551
+ },
+ {
+ "epoch": 63.223580786026204,
+ "grad_norm": 0.019464161247015,
+ "learning_rate": 0.0006,
+ "loss": 4.359729766845703,
+ "step": 4552
+ },
+ {
+ "epoch": 63.237554585152836,
+ "grad_norm": 0.018678035587072372,
+ "learning_rate": 0.0006,
+ "loss": 4.3299665451049805,
+ "step": 4553
+ },
+ {
+ "epoch": 63.251528384279474,
+ "grad_norm": 0.017909523099660873,
+ "learning_rate": 0.0006,
+ "loss": 4.246147155761719,
+ "step": 4554
+ },
+ {
+ "epoch": 63.26550218340611,
+ "grad_norm": 0.01747293211519718,
+ "learning_rate": 0.0006,
+ "loss": 4.41165018081665,
+ "step": 4555
+ },
+ {
+ "epoch": 63.27947598253275,
+ "grad_norm": 0.018670853227376938,
+ "learning_rate": 0.0006,
+ "loss": 4.409883499145508,
+ "step": 4556
+ },
+ {
+ "epoch": 63.29344978165939,
+ "grad_norm": 0.0189529862254858,
+ "learning_rate": 0.0006,
+ "loss": 4.418654918670654,
+ "step": 4557
+ },
+ {
+ "epoch": 63.30742358078603,
+ "grad_norm": 0.018394414335489273,
+ "learning_rate": 0.0006,
+ "loss": 4.38355827331543,
+ "step": 4558
+ },
+ {
+ "epoch": 63.32139737991266,
+ "grad_norm": 0.018317820504307747,
+ "learning_rate": 0.0006,
+ "loss": 4.456843376159668,
+ "step": 4559
+ },
+ {
+ "epoch": 63.3353711790393,
+ "grad_norm": 0.019225597381591797,
+ "learning_rate": 0.0006,
+ "loss": 4.232264518737793,
+ "step": 4560
+ },
+ {
+ "epoch": 63.34934497816594,
+ "grad_norm": 0.016663681715726852,
+ "learning_rate": 0.0006,
+ "loss": 4.393517017364502,
+ "step": 4561
+ },
+ {
+ "epoch": 63.36331877729258,
+ "grad_norm": 0.014975365251302719,
+ "learning_rate": 0.0006,
+ "loss": 4.375183582305908,
+ "step": 4562
+ },
+ {
+ "epoch": 63.377292576419215,
+ "grad_norm": 0.01771732047200203,
+ "learning_rate": 0.0006,
+ "loss": 4.3073410987854,
+ "step": 4563
+ },
+ {
+ "epoch": 63.391266375545854,
+ "grad_norm": 0.01663549244403839,
+ "learning_rate": 0.0006,
+ "loss": 4.311003684997559,
+ "step": 4564
+ },
+ {
+ "epoch": 63.40524017467249,
+ "grad_norm": 0.016301177442073822,
+ "learning_rate": 0.0006,
+ "loss": 4.326807022094727,
+ "step": 4565
+ },
+ {
+ "epoch": 63.419213973799124,
+ "grad_norm": 0.016781188547611237,
+ "learning_rate": 0.0006,
+ "loss": 4.306562423706055,
+ "step": 4566
+ },
+ {
+ "epoch": 63.43318777292576,
+ "grad_norm": 0.01767745055258274,
+ "learning_rate": 0.0006,
+ "loss": 4.367082595825195,
+ "step": 4567
+ },
+ {
+ "epoch": 63.4471615720524,
+ "grad_norm": 0.016728565096855164,
+ "learning_rate": 0.0006,
+ "loss": 4.327400207519531,
+ "step": 4568
+ },
+ {
+ "epoch": 63.46113537117904,
+ "grad_norm": 0.016289403662085533,
+ "learning_rate": 0.0006,
+ "loss": 4.392712116241455,
+ "step": 4569
+ },
+ {
+ "epoch": 63.47510917030568,
+ "grad_norm": 0.016233716160058975,
+ "learning_rate": 0.0006,
+ "loss": 4.354372024536133,
+ "step": 4570
+ },
+ {
+ "epoch": 63.48908296943232,
+ "grad_norm": 0.016547974199056625,
+ "learning_rate": 0.0006,
+ "loss": 4.309050559997559,
+ "step": 4571
+ },
+ {
+ "epoch": 63.50305676855895,
+ "grad_norm": 0.018466509878635406,
+ "learning_rate": 0.0006,
+ "loss": 4.318345069885254,
+ "step": 4572
+ },
+ {
+ "epoch": 63.51703056768559,
+ "grad_norm": 0.017685720697045326,
+ "learning_rate": 0.0006,
+ "loss": 4.399726867675781,
+ "step": 4573
+ },
+ {
+ "epoch": 63.531004366812226,
+ "grad_norm": 0.015801234170794487,
+ "learning_rate": 0.0006,
+ "loss": 4.328495502471924,
+ "step": 4574
+ },
+ {
+ "epoch": 63.544978165938865,
+ "grad_norm": 0.0188873540610075,
+ "learning_rate": 0.0006,
+ "loss": 4.253244400024414,
+ "step": 4575
+ },
+ {
+ "epoch": 63.5589519650655,
+ "grad_norm": 0.01868271268904209,
+ "learning_rate": 0.0006,
+ "loss": 4.463746070861816,
+ "step": 4576
+ },
+ {
+ "epoch": 63.57292576419214,
+ "grad_norm": 0.01844949834048748,
+ "learning_rate": 0.0006,
+ "loss": 4.400300025939941,
+ "step": 4577
+ },
+ {
+ "epoch": 63.58689956331878,
+ "grad_norm": 0.015256480313837528,
+ "learning_rate": 0.0006,
+ "loss": 4.445844650268555,
+ "step": 4578
+ },
+ {
+ "epoch": 63.60087336244541,
+ "grad_norm": 0.015305760316550732,
+ "learning_rate": 0.0006,
+ "loss": 4.317327499389648,
+ "step": 4579
+ },
+ {
+ "epoch": 63.61484716157205,
+ "grad_norm": 0.015200769528746605,
+ "learning_rate": 0.0006,
+ "loss": 4.25112247467041,
+ "step": 4580
+ },
+ {
+ "epoch": 63.62882096069869,
+ "grad_norm": 0.015919538214802742,
+ "learning_rate": 0.0006,
+ "loss": 4.277937889099121,
+ "step": 4581
+ },
+ {
+ "epoch": 63.64279475982533,
+ "grad_norm": 0.01572665013372898,
+ "learning_rate": 0.0006,
+ "loss": 4.302237510681152,
+ "step": 4582
+ },
+ {
+ "epoch": 63.65676855895197,
+ "grad_norm": 0.016428731381893158,
+ "learning_rate": 0.0006,
+ "loss": 4.36762809753418,
+ "step": 4583
+ },
+ {
+ "epoch": 63.670742358078606,
+ "grad_norm": 0.01706082746386528,
+ "learning_rate": 0.0006,
+ "loss": 4.236526966094971,
+ "step": 4584
+ },
+ {
+ "epoch": 63.68471615720524,
+ "grad_norm": 0.017850136384367943,
+ "learning_rate": 0.0006,
+ "loss": 4.264129638671875,
+ "step": 4585
+ },
+ {
+ "epoch": 63.698689956331876,
+ "grad_norm": 0.016079438850283623,
+ "learning_rate": 0.0006,
+ "loss": 4.449799060821533,
+ "step": 4586
+ },
+ {
+ "epoch": 63.712663755458514,
+ "grad_norm": 0.014996383339166641,
+ "learning_rate": 0.0006,
+ "loss": 4.404574394226074,
+ "step": 4587
+ },
+ {
+ "epoch": 63.72663755458515,
+ "grad_norm": 0.015910431742668152,
+ "learning_rate": 0.0006,
+ "loss": 4.348101615905762,
+ "step": 4588
+ },
+ {
+ "epoch": 63.74061135371179,
+ "grad_norm": 0.01630021072924137,
+ "learning_rate": 0.0006,
+ "loss": 4.3193817138671875,
+ "step": 4589
+ },
+ {
+ "epoch": 63.75458515283843,
+ "grad_norm": 0.014571773819625378,
+ "learning_rate": 0.0006,
+ "loss": 4.374402046203613,
+ "step": 4590
+ },
+ {
+ "epoch": 63.76855895196506,
+ "grad_norm": 0.01662345975637436,
+ "learning_rate": 0.0006,
+ "loss": 4.329075813293457,
+ "step": 4591
+ },
+ {
+ "epoch": 63.7825327510917,
+ "grad_norm": 0.016604198142886162,
+ "learning_rate": 0.0006,
+ "loss": 4.436126232147217,
+ "step": 4592
+ },
+ {
+ "epoch": 63.79650655021834,
+ "grad_norm": 0.016578884795308113,
+ "learning_rate": 0.0006,
+ "loss": 4.361468315124512,
+ "step": 4593
+ },
+ {
+ "epoch": 63.81048034934498,
+ "grad_norm": 0.016339825466275215,
+ "learning_rate": 0.0006,
+ "loss": 4.449892044067383,
+ "step": 4594
+ },
+ {
+ "epoch": 63.82445414847162,
+ "grad_norm": 0.016805455088615417,
+ "learning_rate": 0.0006,
+ "loss": 4.415866851806641,
+ "step": 4595
+ },
+ {
+ "epoch": 63.838427947598255,
+ "grad_norm": 0.017366213724017143,
+ "learning_rate": 0.0006,
+ "loss": 4.361730575561523,
+ "step": 4596
+ },
+ {
+ "epoch": 63.852401746724894,
+ "grad_norm": 0.015599003992974758,
+ "learning_rate": 0.0006,
+ "loss": 4.363637924194336,
+ "step": 4597
+ },
+ {
+ "epoch": 63.866375545851525,
+ "grad_norm": 0.015674632042646408,
+ "learning_rate": 0.0006,
+ "loss": 4.332045078277588,
+ "step": 4598
+ },
+ {
+ "epoch": 63.880349344978164,
+ "grad_norm": 0.015350443311035633,
+ "learning_rate": 0.0006,
+ "loss": 4.320063591003418,
+ "step": 4599
+ },
+ {
+ "epoch": 63.8943231441048,
+ "grad_norm": 0.015388146974146366,
+ "learning_rate": 0.0006,
+ "loss": 4.398199081420898,
+ "step": 4600
+ },
+ {
+ "epoch": 63.90829694323144,
+ "grad_norm": 0.01450140681117773,
+ "learning_rate": 0.0006,
+ "loss": 4.244076728820801,
+ "step": 4601
+ },
+ {
+ "epoch": 63.92227074235808,
+ "grad_norm": 0.01637491025030613,
+ "learning_rate": 0.0006,
+ "loss": 4.213549613952637,
+ "step": 4602
+ },
+ {
+ "epoch": 63.93624454148472,
+ "grad_norm": 0.0159814041107893,
+ "learning_rate": 0.0006,
+ "loss": 4.421775817871094,
+ "step": 4603
+ },
+ {
+ "epoch": 63.95021834061135,
+ "grad_norm": 0.017386827617883682,
+ "learning_rate": 0.0006,
+ "loss": 4.340023994445801,
+ "step": 4604
+ },
+ {
+ "epoch": 63.96419213973799,
+ "grad_norm": 0.019679656252264977,
+ "learning_rate": 0.0006,
+ "loss": 4.352534294128418,
+ "step": 4605
+ },
+ {
+ "epoch": 63.97816593886463,
+ "grad_norm": 0.022391516715288162,
+ "learning_rate": 0.0006,
+ "loss": 4.3538312911987305,
+ "step": 4606
+ },
+ {
+ "epoch": 63.992139737991266,
+ "grad_norm": 0.023281006142497063,
+ "learning_rate": 0.0006,
+ "loss": 4.432089328765869,
+ "step": 4607
+ },
+ {
+ "epoch": 64.0,
+ "grad_norm": 0.019916288554668427,
+ "learning_rate": 0.0006,
+ "loss": 4.466937065124512,
+ "step": 4608
+ },
+ {
+ "epoch": 64.0,
+ "eval_loss": 4.659008502960205,
+ "eval_runtime": 56.0019,
+ "eval_samples_per_second": 43.606,
+ "eval_steps_per_second": 1.375,
+ "step": 4608
+ },
+ {
+ "epoch": 64.01397379912663,
+ "grad_norm": 0.017351187765598297,
+ "learning_rate": 0.0006,
+ "loss": 4.294129371643066,
+ "step": 4609
+ },
+ {
+ "epoch": 64.02794759825328,
+ "grad_norm": 0.02405865117907524,
+ "learning_rate": 0.0006,
+ "loss": 4.266717910766602,
+ "step": 4610
+ },
+ {
+ "epoch": 64.04192139737991,
+ "grad_norm": 0.022158386185765266,
+ "learning_rate": 0.0006,
+ "loss": 4.395932197570801,
+ "step": 4611
+ },
+ {
+ "epoch": 64.05589519650655,
+ "grad_norm": 0.019238505512475967,
+ "learning_rate": 0.0006,
+ "loss": 4.266704082489014,
+ "step": 4612
+ },
+ {
+ "epoch": 64.06986899563319,
+ "grad_norm": 0.021670425310730934,
+ "learning_rate": 0.0006,
+ "loss": 4.3160552978515625,
+ "step": 4613
+ },
+ {
+ "epoch": 64.08384279475983,
+ "grad_norm": 0.02076641097664833,
+ "learning_rate": 0.0006,
+ "loss": 4.415626525878906,
+ "step": 4614
+ },
+ {
+ "epoch": 64.09781659388646,
+ "grad_norm": 0.018223078921437263,
+ "learning_rate": 0.0006,
+ "loss": 4.292387008666992,
+ "step": 4615
+ },
+ {
+ "epoch": 64.1117903930131,
+ "grad_norm": 0.017401937395334244,
+ "learning_rate": 0.0006,
+ "loss": 4.424524307250977,
+ "step": 4616
+ },
+ {
+ "epoch": 64.12576419213974,
+ "grad_norm": 0.01663743332028389,
+ "learning_rate": 0.0006,
+ "loss": 4.350032806396484,
+ "step": 4617
+ },
+ {
+ "epoch": 64.13973799126637,
+ "grad_norm": 0.018716629594564438,
+ "learning_rate": 0.0006,
+ "loss": 4.309157371520996,
+ "step": 4618
+ },
+ {
+ "epoch": 64.15371179039302,
+ "grad_norm": 0.019569825381040573,
+ "learning_rate": 0.0006,
+ "loss": 4.358292579650879,
+ "step": 4619
+ },
+ {
+ "epoch": 64.16768558951965,
+ "grad_norm": 0.018994422629475594,
+ "learning_rate": 0.0006,
+ "loss": 4.3338823318481445,
+ "step": 4620
+ },
+ {
+ "epoch": 64.18165938864628,
+ "grad_norm": 0.017690688371658325,
+ "learning_rate": 0.0006,
+ "loss": 4.414468765258789,
+ "step": 4621
+ },
+ {
+ "epoch": 64.19563318777293,
+ "grad_norm": 0.0170609001070261,
+ "learning_rate": 0.0006,
+ "loss": 4.4337568283081055,
+ "step": 4622
+ },
+ {
+ "epoch": 64.20960698689956,
+ "grad_norm": 0.016984131187200546,
+ "learning_rate": 0.0006,
+ "loss": 4.304839134216309,
+ "step": 4623
+ },
+ {
+ "epoch": 64.2235807860262,
+ "grad_norm": 0.01619867794215679,
+ "learning_rate": 0.0006,
+ "loss": 4.389623165130615,
+ "step": 4624
+ },
+ {
+ "epoch": 64.23755458515284,
+ "grad_norm": 0.015428110957145691,
+ "learning_rate": 0.0006,
+ "loss": 4.277515888214111,
+ "step": 4625
+ },
+ {
+ "epoch": 64.25152838427948,
+ "grad_norm": 0.015909532085061073,
+ "learning_rate": 0.0006,
+ "loss": 4.266029357910156,
+ "step": 4626
+ },
+ {
+ "epoch": 64.26550218340611,
+ "grad_norm": 0.014296560548245907,
+ "learning_rate": 0.0006,
+ "loss": 4.286067485809326,
+ "step": 4627
+ },
+ {
+ "epoch": 64.27947598253274,
+ "grad_norm": 0.014065971598029137,
+ "learning_rate": 0.0006,
+ "loss": 4.353528022766113,
+ "step": 4628
+ },
+ {
+ "epoch": 64.29344978165939,
+ "grad_norm": 0.014386293478310108,
+ "learning_rate": 0.0006,
+ "loss": 4.318715572357178,
+ "step": 4629
+ },
+ {
+ "epoch": 64.30742358078602,
+ "grad_norm": 0.014001123607158661,
+ "learning_rate": 0.0006,
+ "loss": 4.362497329711914,
+ "step": 4630
+ },
+ {
+ "epoch": 64.32139737991267,
+ "grad_norm": 0.013408979400992393,
+ "learning_rate": 0.0006,
+ "loss": 4.417397499084473,
+ "step": 4631
+ },
+ {
+ "epoch": 64.3353711790393,
+ "grad_norm": 0.014960093423724174,
+ "learning_rate": 0.0006,
+ "loss": 4.359710693359375,
+ "step": 4632
+ },
+ {
+ "epoch": 64.34934497816595,
+ "grad_norm": 0.014658777043223381,
+ "learning_rate": 0.0006,
+ "loss": 4.318822383880615,
+ "step": 4633
+ },
+ {
+ "epoch": 64.36331877729258,
+ "grad_norm": 0.013766744174063206,
+ "learning_rate": 0.0006,
+ "loss": 4.265679836273193,
+ "step": 4634
+ },
+ {
+ "epoch": 64.37729257641921,
+ "grad_norm": 0.013821255415678024,
+ "learning_rate": 0.0006,
+ "loss": 4.3452911376953125,
+ "step": 4635
+ },
+ {
+ "epoch": 64.39126637554585,
+ "grad_norm": 0.013541066087782383,
+ "learning_rate": 0.0006,
+ "loss": 4.429254531860352,
+ "step": 4636
+ },
+ {
+ "epoch": 64.40524017467249,
+ "grad_norm": 0.013985645025968552,
+ "learning_rate": 0.0006,
+ "loss": 4.304452896118164,
+ "step": 4637
+ },
+ {
+ "epoch": 64.41921397379913,
+ "grad_norm": 0.015079808421432972,
+ "learning_rate": 0.0006,
+ "loss": 4.296828746795654,
+ "step": 4638
+ },
+ {
+ "epoch": 64.43318777292576,
+ "grad_norm": 0.014896944165229797,
+ "learning_rate": 0.0006,
+ "loss": 4.299156665802002,
+ "step": 4639
+ },
+ {
+ "epoch": 64.44716157205241,
+ "grad_norm": 0.013770041987299919,
+ "learning_rate": 0.0006,
+ "loss": 4.351439952850342,
+ "step": 4640
+ },
+ {
+ "epoch": 64.46113537117904,
+ "grad_norm": 0.01531178504228592,
+ "learning_rate": 0.0006,
+ "loss": 4.373715877532959,
+ "step": 4641
+ },
+ {
+ "epoch": 64.47510917030567,
+ "grad_norm": 0.014024450443685055,
+ "learning_rate": 0.0006,
+ "loss": 4.368053436279297,
+ "step": 4642
+ },
+ {
+ "epoch": 64.48908296943232,
+ "grad_norm": 0.013864024542272091,
+ "learning_rate": 0.0006,
+ "loss": 4.378564834594727,
+ "step": 4643
+ },
+ {
+ "epoch": 64.50305676855895,
+ "grad_norm": 0.013925445266067982,
+ "learning_rate": 0.0006,
+ "loss": 4.261663436889648,
+ "step": 4644
+ },
+ {
+ "epoch": 64.5170305676856,
+ "grad_norm": 0.015011338517069817,
+ "learning_rate": 0.0006,
+ "loss": 4.337285995483398,
+ "step": 4645
+ },
+ {
+ "epoch": 64.53100436681223,
+ "grad_norm": 0.016941986978054047,
+ "learning_rate": 0.0006,
+ "loss": 4.336877822875977,
+ "step": 4646
+ },
+ {
+ "epoch": 64.54497816593886,
+ "grad_norm": 0.01719909906387329,
+ "learning_rate": 0.0006,
+ "loss": 4.149813175201416,
+ "step": 4647
+ },
+ {
+ "epoch": 64.5589519650655,
+ "grad_norm": 0.01608402095735073,
+ "learning_rate": 0.0006,
+ "loss": 4.191280364990234,
+ "step": 4648
+ },
+ {
+ "epoch": 64.57292576419214,
+ "grad_norm": 0.015284820459783077,
+ "learning_rate": 0.0006,
+ "loss": 4.35953426361084,
+ "step": 4649
+ },
+ {
+ "epoch": 64.58689956331878,
+ "grad_norm": 0.014855682849884033,
+ "learning_rate": 0.0006,
+ "loss": 4.308401107788086,
+ "step": 4650
+ },
+ {
+ "epoch": 64.60087336244541,
+ "grad_norm": 0.017553111538290977,
+ "learning_rate": 0.0006,
+ "loss": 4.279720306396484,
+ "step": 4651
+ },
+ {
+ "epoch": 64.61484716157206,
+ "grad_norm": 0.019506998360157013,
+ "learning_rate": 0.0006,
+ "loss": 4.319524765014648,
+ "step": 4652
+ },
+ {
+ "epoch": 64.62882096069869,
+ "grad_norm": 0.020293425768613815,
+ "learning_rate": 0.0006,
+ "loss": 4.340523719787598,
+ "step": 4653
+ },
+ {
+ "epoch": 64.64279475982532,
+ "grad_norm": 0.01664556749165058,
+ "learning_rate": 0.0006,
+ "loss": 4.296785831451416,
+ "step": 4654
+ },
+ {
+ "epoch": 64.65676855895197,
+ "grad_norm": 0.012592585757374763,
+ "learning_rate": 0.0006,
+ "loss": 4.464595317840576,
+ "step": 4655
+ },
+ {
+ "epoch": 64.6707423580786,
+ "grad_norm": 0.01549589168280363,
+ "learning_rate": 0.0006,
+ "loss": 4.275053024291992,
+ "step": 4656
+ },
+ {
+ "epoch": 64.68471615720524,
+ "grad_norm": 0.01705172471702099,
+ "learning_rate": 0.0006,
+ "loss": 4.388431549072266,
+ "step": 4657
+ },
+ {
+ "epoch": 64.69868995633188,
+ "grad_norm": 0.015438582748174667,
+ "learning_rate": 0.0006,
+ "loss": 4.312070846557617,
+ "step": 4658
+ },
+ {
+ "epoch": 64.71266375545852,
+ "grad_norm": 0.014448083005845547,
+ "learning_rate": 0.0006,
+ "loss": 4.237599849700928,
+ "step": 4659
+ },
+ {
+ "epoch": 64.72663755458515,
+ "grad_norm": 0.015267071314156055,
+ "learning_rate": 0.0006,
+ "loss": 4.29893684387207,
+ "step": 4660
+ },
+ {
+ "epoch": 64.74061135371178,
+ "grad_norm": 0.016615847125649452,
+ "learning_rate": 0.0006,
+ "loss": 4.432273864746094,
+ "step": 4661
+ },
+ {
+ "epoch": 64.75458515283843,
+ "grad_norm": 0.016584362834692,
+ "learning_rate": 0.0006,
+ "loss": 4.202266216278076,
+ "step": 4662
+ },
+ {
+ "epoch": 64.76855895196506,
+ "grad_norm": 0.018313994631171227,
+ "learning_rate": 0.0006,
+ "loss": 4.367966651916504,
+ "step": 4663
+ },
+ {
+ "epoch": 64.78253275109171,
+ "grad_norm": 0.017946481704711914,
+ "learning_rate": 0.0006,
+ "loss": 4.205172538757324,
+ "step": 4664
+ },
+ {
+ "epoch": 64.79650655021834,
+ "grad_norm": 0.017744354903697968,
+ "learning_rate": 0.0006,
+ "loss": 4.295661926269531,
+ "step": 4665
+ },
+ {
+ "epoch": 64.81048034934499,
+ "grad_norm": 0.016103824600577354,
+ "learning_rate": 0.0006,
+ "loss": 4.276543617248535,
+ "step": 4666
+ },
+ {
+ "epoch": 64.82445414847162,
+ "grad_norm": 0.014590012840926647,
+ "learning_rate": 0.0006,
+ "loss": 4.414247035980225,
+ "step": 4667
+ },
+ {
+ "epoch": 64.83842794759825,
+ "grad_norm": 0.01782386749982834,
+ "learning_rate": 0.0006,
+ "loss": 4.203220367431641,
+ "step": 4668
+ },
+ {
+ "epoch": 64.8524017467249,
+ "grad_norm": 0.022122733294963837,
+ "learning_rate": 0.0006,
+ "loss": 4.38078498840332,
+ "step": 4669
+ },
+ {
+ "epoch": 64.86637554585153,
+ "grad_norm": 0.0250293780118227,
+ "learning_rate": 0.0006,
+ "loss": 4.3310136795043945,
+ "step": 4670
+ },
+ {
+ "epoch": 64.88034934497817,
+ "grad_norm": 0.022438062354922295,
+ "learning_rate": 0.0006,
+ "loss": 4.428631782531738,
+ "step": 4671
+ },
+ {
+ "epoch": 64.8943231441048,
+ "grad_norm": 0.018518192693591118,
+ "learning_rate": 0.0006,
+ "loss": 4.428250312805176,
+ "step": 4672
+ },
+ {
+ "epoch": 64.90829694323143,
+ "grad_norm": 0.018105225637555122,
+ "learning_rate": 0.0006,
+ "loss": 4.336209297180176,
+ "step": 4673
+ },
+ {
+ "epoch": 64.92227074235808,
+ "grad_norm": 0.017564399167895317,
+ "learning_rate": 0.0006,
+ "loss": 4.416011810302734,
+ "step": 4674
+ },
+ {
+ "epoch": 64.93624454148471,
+ "grad_norm": 0.018747031688690186,
+ "learning_rate": 0.0006,
+ "loss": 4.419938087463379,
+ "step": 4675
+ },
+ {
+ "epoch": 64.95021834061136,
+ "grad_norm": 0.018686484545469284,
+ "learning_rate": 0.0006,
+ "loss": 4.347185134887695,
+ "step": 4676
+ },
+ {
+ "epoch": 64.96419213973799,
+ "grad_norm": 0.019329898059368134,
+ "learning_rate": 0.0006,
+ "loss": 4.343749046325684,
+ "step": 4677
+ },
+ {
+ "epoch": 64.97816593886463,
+ "grad_norm": 0.018517782911658287,
+ "learning_rate": 0.0006,
+ "loss": 4.264604568481445,
+ "step": 4678
+ },
+ {
+ "epoch": 64.99213973799127,
+ "grad_norm": 0.016579125076532364,
+ "learning_rate": 0.0006,
+ "loss": 4.303423881530762,
+ "step": 4679
+ },
+ {
+ "epoch": 65.0,
+ "grad_norm": 0.01617567613720894,
+ "learning_rate": 0.0006,
+ "loss": 4.326556205749512,
+ "step": 4680
+ },
+ {
+ "epoch": 65.0,
+ "eval_loss": 4.655026912689209,
+ "eval_runtime": 57.1299,
+ "eval_samples_per_second": 42.745,
+ "eval_steps_per_second": 1.348,
+ "step": 4680
+ },
+ {
+ "epoch": 65.01397379912663,
+ "grad_norm": 0.01607249490916729,
+ "learning_rate": 0.0006,
+ "loss": 4.276607036590576,
+ "step": 4681
+ },
+ {
+ "epoch": 65.02794759825328,
+ "grad_norm": 0.016779497265815735,
+ "learning_rate": 0.0006,
+ "loss": 4.412464618682861,
+ "step": 4682
+ },
+ {
+ "epoch": 65.04192139737991,
+ "grad_norm": 0.018161555752158165,
+ "learning_rate": 0.0006,
+ "loss": 4.199329376220703,
+ "step": 4683
+ },
+ {
+ "epoch": 65.05589519650655,
+ "grad_norm": 0.017508309334516525,
+ "learning_rate": 0.0006,
+ "loss": 4.217567443847656,
+ "step": 4684
+ },
+ {
+ "epoch": 65.06986899563319,
+ "grad_norm": 0.015716979280114174,
+ "learning_rate": 0.0006,
+ "loss": 4.322023391723633,
+ "step": 4685
+ },
+ {
+ "epoch": 65.08384279475983,
+ "grad_norm": 0.015113635919988155,
+ "learning_rate": 0.0006,
+ "loss": 4.333515167236328,
+ "step": 4686
+ },
+ {
+ "epoch": 65.09781659388646,
+ "grad_norm": 0.01437560748308897,
+ "learning_rate": 0.0006,
+ "loss": 4.295609474182129,
+ "step": 4687
+ },
+ {
+ "epoch": 65.1117903930131,
+ "grad_norm": 0.013074862770736217,
+ "learning_rate": 0.0006,
+ "loss": 4.314120769500732,
+ "step": 4688
+ },
+ {
+ "epoch": 65.12576419213974,
+ "grad_norm": 0.012908284552395344,
+ "learning_rate": 0.0006,
+ "loss": 4.216890811920166,
+ "step": 4689
+ },
+ {
+ "epoch": 65.13973799126637,
+ "grad_norm": 0.013001754879951477,
+ "learning_rate": 0.0006,
+ "loss": 4.287717342376709,
+ "step": 4690
+ },
+ {
+ "epoch": 65.15371179039302,
+ "grad_norm": 0.013694263994693756,
+ "learning_rate": 0.0006,
+ "loss": 4.333837985992432,
+ "step": 4691
+ },
+ {
+ "epoch": 65.16768558951965,
+ "grad_norm": 0.014461833983659744,
+ "learning_rate": 0.0006,
+ "loss": 4.323545455932617,
+ "step": 4692
+ },
+ {
+ "epoch": 65.18165938864628,
+ "grad_norm": 0.013670675456523895,
+ "learning_rate": 0.0006,
+ "loss": 4.3313493728637695,
+ "step": 4693
+ },
+ {
+ "epoch": 65.19563318777293,
+ "grad_norm": 0.013094350695610046,
+ "learning_rate": 0.0006,
+ "loss": 4.23858118057251,
+ "step": 4694
+ },
+ {
+ "epoch": 65.20960698689956,
+ "grad_norm": 0.013996944762766361,
+ "learning_rate": 0.0006,
+ "loss": 4.428705215454102,
+ "step": 4695
+ },
+ {
+ "epoch": 65.2235807860262,
+ "grad_norm": 0.013295048847794533,
+ "learning_rate": 0.0006,
+ "loss": 4.238950252532959,
+ "step": 4696
+ },
+ {
+ "epoch": 65.23755458515284,
+ "grad_norm": 0.013380616903305054,
+ "learning_rate": 0.0006,
+ "loss": 4.273185729980469,
+ "step": 4697
+ },
+ {
+ "epoch": 65.25152838427948,
+ "grad_norm": 0.015062890015542507,
+ "learning_rate": 0.0006,
+ "loss": 4.29472017288208,
+ "step": 4698
+ },
+ {
+ "epoch": 65.26550218340611,
+ "grad_norm": 0.017317263409495354,
+ "learning_rate": 0.0006,
+ "loss": 4.337954998016357,
+ "step": 4699
+ },
+ {
+ "epoch": 65.27947598253274,
+ "grad_norm": 0.017698992043733597,
+ "learning_rate": 0.0006,
+ "loss": 4.336437225341797,
+ "step": 4700
+ },
+ {
+ "epoch": 65.29344978165939,
+ "grad_norm": 0.01628575474023819,
+ "learning_rate": 0.0006,
+ "loss": 4.302372455596924,
+ "step": 4701
+ },
+ {
+ "epoch": 65.30742358078602,
+ "grad_norm": 0.014531854540109634,
+ "learning_rate": 0.0006,
+ "loss": 4.256450653076172,
+ "step": 4702
+ },
+ {
+ "epoch": 65.32139737991267,
+ "grad_norm": 0.014549105428159237,
+ "learning_rate": 0.0006,
+ "loss": 4.302240371704102,
+ "step": 4703
+ },
+ {
+ "epoch": 65.3353711790393,
+ "grad_norm": 0.014881663955748081,
+ "learning_rate": 0.0006,
+ "loss": 4.3316121101379395,
+ "step": 4704
+ },
+ {
+ "epoch": 65.34934497816595,
+ "grad_norm": 0.014513437636196613,
+ "learning_rate": 0.0006,
+ "loss": 4.441212177276611,
+ "step": 4705
+ },
+ {
+ "epoch": 65.36331877729258,
+ "grad_norm": 0.014122087508440018,
+ "learning_rate": 0.0006,
+ "loss": 4.343296527862549,
+ "step": 4706
+ },
+ {
+ "epoch": 65.37729257641921,
+ "grad_norm": 0.014817348681390285,
+ "learning_rate": 0.0006,
+ "loss": 4.195674896240234,
+ "step": 4707
+ },
+ {
+ "epoch": 65.39126637554585,
+ "grad_norm": 0.01787056028842926,
+ "learning_rate": 0.0006,
+ "loss": 4.3507914543151855,
+ "step": 4708
+ },
+ {
+ "epoch": 65.40524017467249,
+ "grad_norm": 0.015295136719942093,
+ "learning_rate": 0.0006,
+ "loss": 4.234372138977051,
+ "step": 4709
+ },
+ {
+ "epoch": 65.41921397379913,
+ "grad_norm": 0.014324234798550606,
+ "learning_rate": 0.0006,
+ "loss": 4.2907867431640625,
+ "step": 4710
+ },
+ {
+ "epoch": 65.43318777292576,
+ "grad_norm": 0.016090670600533485,
+ "learning_rate": 0.0006,
+ "loss": 4.177244186401367,
+ "step": 4711
+ },
+ {
+ "epoch": 65.44716157205241,
+ "grad_norm": 0.0174604132771492,
+ "learning_rate": 0.0006,
+ "loss": 4.335577964782715,
+ "step": 4712
+ },
+ {
+ "epoch": 65.46113537117904,
+ "grad_norm": 0.016731899231672287,
+ "learning_rate": 0.0006,
+ "loss": 4.30197811126709,
+ "step": 4713
+ },
+ {
+ "epoch": 65.47510917030567,
+ "grad_norm": 0.017423735931515694,
+ "learning_rate": 0.0006,
+ "loss": 4.3528852462768555,
+ "step": 4714
+ },
+ {
+ "epoch": 65.48908296943232,
+ "grad_norm": 0.01780976541340351,
+ "learning_rate": 0.0006,
+ "loss": 4.185036659240723,
+ "step": 4715
+ },
+ {
+ "epoch": 65.50305676855895,
+ "grad_norm": 0.019977200776338577,
+ "learning_rate": 0.0006,
+ "loss": 4.328766345977783,
+ "step": 4716
+ },
+ {
+ "epoch": 65.5170305676856,
+ "grad_norm": 0.020284287631511688,
+ "learning_rate": 0.0006,
+ "loss": 4.285494804382324,
+ "step": 4717
+ },
+ {
+ "epoch": 65.53100436681223,
+ "grad_norm": 0.019168002530932426,
+ "learning_rate": 0.0006,
+ "loss": 4.328560829162598,
+ "step": 4718
+ },
+ {
+ "epoch": 65.54497816593886,
+ "grad_norm": 0.019217776134610176,
+ "learning_rate": 0.0006,
+ "loss": 4.189998626708984,
+ "step": 4719
+ },
+ {
+ "epoch": 65.5589519650655,
+ "grad_norm": 0.018960488960146904,
+ "learning_rate": 0.0006,
+ "loss": 4.4995012283325195,
+ "step": 4720
+ },
+ {
+ "epoch": 65.57292576419214,
+ "grad_norm": 0.01941288262605667,
+ "learning_rate": 0.0006,
+ "loss": 4.26570987701416,
+ "step": 4721
+ },
+ {
+ "epoch": 65.58689956331878,
+ "grad_norm": 0.020263448357582092,
+ "learning_rate": 0.0006,
+ "loss": 4.411645889282227,
+ "step": 4722
+ },
+ {
+ "epoch": 65.60087336244541,
+ "grad_norm": 0.02144799567759037,
+ "learning_rate": 0.0006,
+ "loss": 4.163675308227539,
+ "step": 4723
+ },
+ {
+ "epoch": 65.61484716157206,
+ "grad_norm": 0.019790446385741234,
+ "learning_rate": 0.0006,
+ "loss": 4.2361063957214355,
+ "step": 4724
+ },
+ {
+ "epoch": 65.62882096069869,
+ "grad_norm": 0.01829805225133896,
+ "learning_rate": 0.0006,
+ "loss": 4.449810028076172,
+ "step": 4725
+ },
+ {
+ "epoch": 65.64279475982532,
+ "grad_norm": 0.01571105420589447,
+ "learning_rate": 0.0006,
+ "loss": 4.281406879425049,
+ "step": 4726
+ },
+ {
+ "epoch": 65.65676855895197,
+ "grad_norm": 0.016477147117257118,
+ "learning_rate": 0.0006,
+ "loss": 4.293408393859863,
+ "step": 4727
+ },
+ {
+ "epoch": 65.6707423580786,
+ "grad_norm": 0.016766032204031944,
+ "learning_rate": 0.0006,
+ "loss": 4.378252983093262,
+ "step": 4728
+ },
+ {
+ "epoch": 65.68471615720524,
+ "grad_norm": 0.0159933939576149,
+ "learning_rate": 0.0006,
+ "loss": 4.3065667152404785,
+ "step": 4729
+ },
+ {
+ "epoch": 65.69868995633188,
+ "grad_norm": 0.01658209227025509,
+ "learning_rate": 0.0006,
+ "loss": 4.174026966094971,
+ "step": 4730
+ },
+ {
+ "epoch": 65.71266375545852,
+ "grad_norm": 0.015081380493938923,
+ "learning_rate": 0.0006,
+ "loss": 4.360222339630127,
+ "step": 4731
+ },
+ {
+ "epoch": 65.72663755458515,
+ "grad_norm": 0.015110890381038189,
+ "learning_rate": 0.0006,
+ "loss": 4.218810558319092,
+ "step": 4732
+ },
+ {
+ "epoch": 65.74061135371178,
+ "grad_norm": 0.015300953760743141,
+ "learning_rate": 0.0006,
+ "loss": 4.331979274749756,
+ "step": 4733
+ },
+ {
+ "epoch": 65.75458515283843,
+ "grad_norm": 0.01715504191815853,
+ "learning_rate": 0.0006,
+ "loss": 4.349112033843994,
+ "step": 4734
+ },
+ {
+ "epoch": 65.76855895196506,
+ "grad_norm": 0.015870366245508194,
+ "learning_rate": 0.0006,
+ "loss": 4.329850196838379,
+ "step": 4735
+ },
+ {
+ "epoch": 65.78253275109171,
+ "grad_norm": 0.01541937980800867,
+ "learning_rate": 0.0006,
+ "loss": 4.3400163650512695,
+ "step": 4736
+ },
+ {
+ "epoch": 65.79650655021834,
+ "grad_norm": 0.01738942787051201,
+ "learning_rate": 0.0006,
+ "loss": 4.220798492431641,
+ "step": 4737
+ },
+ {
+ "epoch": 65.81048034934499,
+ "grad_norm": 0.01685990020632744,
+ "learning_rate": 0.0006,
+ "loss": 4.177238464355469,
+ "step": 4738
+ },
+ {
+ "epoch": 65.82445414847162,
+ "grad_norm": 0.016183465719223022,
+ "learning_rate": 0.0006,
+ "loss": 4.327216625213623,
+ "step": 4739
+ },
+ {
+ "epoch": 65.83842794759825,
+ "grad_norm": 0.01677694357931614,
+ "learning_rate": 0.0006,
+ "loss": 4.27766227722168,
+ "step": 4740
+ },
+ {
+ "epoch": 65.8524017467249,
+ "grad_norm": 0.01645066775381565,
+ "learning_rate": 0.0006,
+ "loss": 4.288347244262695,
+ "step": 4741
+ },
+ {
+ "epoch": 65.86637554585153,
+ "grad_norm": 0.018034635111689568,
+ "learning_rate": 0.0006,
+ "loss": 4.292458534240723,
+ "step": 4742
+ },
+ {
+ "epoch": 65.88034934497817,
+ "grad_norm": 0.017973346635699272,
+ "learning_rate": 0.0006,
+ "loss": 4.3117265701293945,
+ "step": 4743
+ },
+ {
+ "epoch": 65.8943231441048,
+ "grad_norm": 0.016605669632554054,
+ "learning_rate": 0.0006,
+ "loss": 4.280615329742432,
+ "step": 4744
+ },
+ {
+ "epoch": 65.90829694323143,
+ "grad_norm": 0.016298074275255203,
+ "learning_rate": 0.0006,
+ "loss": 4.301933288574219,
+ "step": 4745
+ },
+ {
+ "epoch": 65.92227074235808,
+ "grad_norm": 0.015383216552436352,
+ "learning_rate": 0.0006,
+ "loss": 4.352383136749268,
+ "step": 4746
+ },
+ {
+ "epoch": 65.93624454148471,
+ "grad_norm": 0.016222182661294937,
+ "learning_rate": 0.0006,
+ "loss": 4.305725574493408,
+ "step": 4747
+ },
+ {
+ "epoch": 65.95021834061136,
+ "grad_norm": 0.01691329851746559,
+ "learning_rate": 0.0006,
+ "loss": 4.29502010345459,
+ "step": 4748
+ },
+ {
+ "epoch": 65.96419213973799,
+ "grad_norm": 0.017637677490711212,
+ "learning_rate": 0.0006,
+ "loss": 4.3133039474487305,
+ "step": 4749
+ },
+ {
+ "epoch": 65.97816593886463,
+ "grad_norm": 0.021362047642469406,
+ "learning_rate": 0.0006,
+ "loss": 4.458834648132324,
+ "step": 4750
+ },
+ {
+ "epoch": 65.99213973799127,
+ "grad_norm": 0.023553011938929558,
+ "learning_rate": 0.0006,
+ "loss": 4.298727989196777,
+ "step": 4751
+ },
+ {
+ "epoch": 66.0,
+ "grad_norm": 0.02154950052499771,
+ "learning_rate": 0.0006,
+ "loss": 4.403392791748047,
+ "step": 4752
+ },
+ {
+ "epoch": 66.0,
+ "eval_loss": 4.724340438842773,
+ "eval_runtime": 56.9555,
+ "eval_samples_per_second": 42.876,
+ "eval_steps_per_second": 1.352,
+ "step": 4752
+ },
+ {
+ "epoch": 66.01397379912663,
+ "grad_norm": 0.01916784606873989,
+ "learning_rate": 0.0006,
+ "loss": 4.22931432723999,
+ "step": 4753
+ },
+ {
+ "epoch": 66.02794759825328,
+ "grad_norm": 0.017492054030299187,
+ "learning_rate": 0.0006,
+ "loss": 4.271522521972656,
+ "step": 4754
+ },
+ {
+ "epoch": 66.04192139737991,
+ "grad_norm": 0.018408095464110374,
+ "learning_rate": 0.0006,
+ "loss": 4.337705135345459,
+ "step": 4755
+ },
+ {
+ "epoch": 66.05589519650655,
+ "grad_norm": 0.019220108166337013,
+ "learning_rate": 0.0006,
+ "loss": 4.280340194702148,
+ "step": 4756
+ },
+ {
+ "epoch": 66.06986899563319,
+ "grad_norm": 0.01773735322058201,
+ "learning_rate": 0.0006,
+ "loss": 4.390201568603516,
+ "step": 4757
+ },
+ {
+ "epoch": 66.08384279475983,
+ "grad_norm": 0.01679125428199768,
+ "learning_rate": 0.0006,
+ "loss": 4.299282073974609,
+ "step": 4758
+ },
+ {
+ "epoch": 66.09781659388646,
+ "grad_norm": 0.01479059923440218,
+ "learning_rate": 0.0006,
+ "loss": 4.472630500793457,
+ "step": 4759
+ },
+ {
+ "epoch": 66.1117903930131,
+ "grad_norm": 0.015130104497075081,
+ "learning_rate": 0.0006,
+ "loss": 4.301194190979004,
+ "step": 4760
+ },
+ {
+ "epoch": 66.12576419213974,
+ "grad_norm": 0.0158979594707489,
+ "learning_rate": 0.0006,
+ "loss": 4.375751972198486,
+ "step": 4761
+ },
+ {
+ "epoch": 66.13973799126637,
+ "grad_norm": 0.014432408846914768,
+ "learning_rate": 0.0006,
+ "loss": 4.279788017272949,
+ "step": 4762
+ },
+ {
+ "epoch": 66.15371179039302,
+ "grad_norm": 0.013026784174144268,
+ "learning_rate": 0.0006,
+ "loss": 4.255035400390625,
+ "step": 4763
+ },
+ {
+ "epoch": 66.16768558951965,
+ "grad_norm": 0.014422653242945671,
+ "learning_rate": 0.0006,
+ "loss": 4.261832237243652,
+ "step": 4764
+ },
+ {
+ "epoch": 66.18165938864628,
+ "grad_norm": 0.014613019302487373,
+ "learning_rate": 0.0006,
+ "loss": 4.382184028625488,
+ "step": 4765
+ },
+ {
+ "epoch": 66.19563318777293,
+ "grad_norm": 0.014989894814789295,
+ "learning_rate": 0.0006,
+ "loss": 4.40528678894043,
+ "step": 4766
+ },
+ {
+ "epoch": 66.20960698689956,
+ "grad_norm": 0.018190350383520126,
+ "learning_rate": 0.0006,
+ "loss": 4.302414894104004,
+ "step": 4767
+ },
+ {
+ "epoch": 66.2235807860262,
+ "grad_norm": 0.020887890830636024,
+ "learning_rate": 0.0006,
+ "loss": 4.289920806884766,
+ "step": 4768
+ },
+ {
+ "epoch": 66.23755458515284,
+ "grad_norm": 0.02376876398921013,
+ "learning_rate": 0.0006,
+ "loss": 4.297855377197266,
+ "step": 4769
+ },
+ {
+ "epoch": 66.25152838427948,
+ "grad_norm": 0.02152269519865513,
+ "learning_rate": 0.0006,
+ "loss": 4.1972832679748535,
+ "step": 4770
+ },
+ {
+ "epoch": 66.26550218340611,
+ "grad_norm": 0.015572802163660526,
+ "learning_rate": 0.0006,
+ "loss": 4.307487487792969,
+ "step": 4771
+ },
+ {
+ "epoch": 66.27947598253274,
+ "grad_norm": 0.015228749252855778,
+ "learning_rate": 0.0006,
+ "loss": 4.103250503540039,
+ "step": 4772
+ },
+ {
+ "epoch": 66.29344978165939,
+ "grad_norm": 0.016925504431128502,
+ "learning_rate": 0.0006,
+ "loss": 4.247461318969727,
+ "step": 4773
+ },
+ {
+ "epoch": 66.30742358078602,
+ "grad_norm": 0.01613711006939411,
+ "learning_rate": 0.0006,
+ "loss": 4.4699273109436035,
+ "step": 4774
+ },
+ {
+ "epoch": 66.32139737991267,
+ "grad_norm": 0.018332140520215034,
+ "learning_rate": 0.0006,
+ "loss": 4.22969388961792,
+ "step": 4775
+ },
+ {
+ "epoch": 66.3353711790393,
+ "grad_norm": 0.019575074315071106,
+ "learning_rate": 0.0006,
+ "loss": 4.29397439956665,
+ "step": 4776
+ },
+ {
+ "epoch": 66.34934497816595,
+ "grad_norm": 0.018177419900894165,
+ "learning_rate": 0.0006,
+ "loss": 4.221899509429932,
+ "step": 4777
+ },
+ {
+ "epoch": 66.36331877729258,
+ "grad_norm": 0.014292889274656773,
+ "learning_rate": 0.0006,
+ "loss": 4.325457572937012,
+ "step": 4778
+ },
+ {
+ "epoch": 66.37729257641921,
+ "grad_norm": 0.015098796226084232,
+ "learning_rate": 0.0006,
+ "loss": 4.364837169647217,
+ "step": 4779
+ },
+ {
+ "epoch": 66.39126637554585,
+ "grad_norm": 0.013992635533213615,
+ "learning_rate": 0.0006,
+ "loss": 4.299082279205322,
+ "step": 4780
+ },
+ {
+ "epoch": 66.40524017467249,
+ "grad_norm": 0.01504511572420597,
+ "learning_rate": 0.0006,
+ "loss": 4.230877876281738,
+ "step": 4781
+ },
+ {
+ "epoch": 66.41921397379913,
+ "grad_norm": 0.014875639230012894,
+ "learning_rate": 0.0006,
+ "loss": 4.368405818939209,
+ "step": 4782
+ },
+ {
+ "epoch": 66.43318777292576,
+ "grad_norm": 0.014306887984275818,
+ "learning_rate": 0.0006,
+ "loss": 4.389184474945068,
+ "step": 4783
+ },
+ {
+ "epoch": 66.44716157205241,
+ "grad_norm": 0.012959563173353672,
+ "learning_rate": 0.0006,
+ "loss": 4.37934684753418,
+ "step": 4784
+ },
+ {
+ "epoch": 66.46113537117904,
+ "grad_norm": 0.013206703588366508,
+ "learning_rate": 0.0006,
+ "loss": 4.442799091339111,
+ "step": 4785
+ },
+ {
+ "epoch": 66.47510917030567,
+ "grad_norm": 0.014848909340798855,
+ "learning_rate": 0.0006,
+ "loss": 4.203481674194336,
+ "step": 4786
+ },
+ {
+ "epoch": 66.48908296943232,
+ "grad_norm": 0.015637682750821114,
+ "learning_rate": 0.0006,
+ "loss": 4.27552604675293,
+ "step": 4787
+ },
+ {
+ "epoch": 66.50305676855895,
+ "grad_norm": 0.016017816960811615,
+ "learning_rate": 0.0006,
+ "loss": 4.296402931213379,
+ "step": 4788
+ },
+ {
+ "epoch": 66.5170305676856,
+ "grad_norm": 0.01577814109623432,
+ "learning_rate": 0.0006,
+ "loss": 4.301733016967773,
+ "step": 4789
+ },
+ {
+ "epoch": 66.53100436681223,
+ "grad_norm": 0.015293175354599953,
+ "learning_rate": 0.0006,
+ "loss": 4.369877815246582,
+ "step": 4790
+ },
+ {
+ "epoch": 66.54497816593886,
+ "grad_norm": 0.014262278564274311,
+ "learning_rate": 0.0006,
+ "loss": 4.3314738273620605,
+ "step": 4791
+ },
+ {
+ "epoch": 66.5589519650655,
+ "grad_norm": 0.015123482793569565,
+ "learning_rate": 0.0006,
+ "loss": 4.313591957092285,
+ "step": 4792
+ },
+ {
+ "epoch": 66.57292576419214,
+ "grad_norm": 0.015458766371011734,
+ "learning_rate": 0.0006,
+ "loss": 4.3045854568481445,
+ "step": 4793
+ },
+ {
+ "epoch": 66.58689956331878,
+ "grad_norm": 0.017217233777046204,
+ "learning_rate": 0.0006,
+ "loss": 4.269944190979004,
+ "step": 4794
+ },
+ {
+ "epoch": 66.60087336244541,
+ "grad_norm": 0.01962282694876194,
+ "learning_rate": 0.0006,
+ "loss": 4.271735191345215,
+ "step": 4795
+ },
+ {
+ "epoch": 66.61484716157206,
+ "grad_norm": 0.01917066052556038,
+ "learning_rate": 0.0006,
+ "loss": 4.288168430328369,
+ "step": 4796
+ },
+ {
+ "epoch": 66.62882096069869,
+ "grad_norm": 0.018211524933576584,
+ "learning_rate": 0.0006,
+ "loss": 4.280660629272461,
+ "step": 4797
+ },
+ {
+ "epoch": 66.64279475982532,
+ "grad_norm": 0.01936524175107479,
+ "learning_rate": 0.0006,
+ "loss": 4.3710551261901855,
+ "step": 4798
+ },
+ {
+ "epoch": 66.65676855895197,
+ "grad_norm": 0.01882351003587246,
+ "learning_rate": 0.0006,
+ "loss": 4.342750549316406,
+ "step": 4799
+ },
+ {
+ "epoch": 66.6707423580786,
+ "grad_norm": 0.017149122431874275,
+ "learning_rate": 0.0006,
+ "loss": 4.32134485244751,
+ "step": 4800
+ },
+ {
+ "epoch": 66.68471615720524,
+ "grad_norm": 0.01699327677488327,
+ "learning_rate": 0.0006,
+ "loss": 4.248893737792969,
+ "step": 4801
+ },
+ {
+ "epoch": 66.69868995633188,
+ "grad_norm": 0.016873255372047424,
+ "learning_rate": 0.0006,
+ "loss": 4.277350425720215,
+ "step": 4802
+ },
+ {
+ "epoch": 66.71266375545852,
+ "grad_norm": 0.016592837870121002,
+ "learning_rate": 0.0006,
+ "loss": 4.221665859222412,
+ "step": 4803
+ },
+ {
+ "epoch": 66.72663755458515,
+ "grad_norm": 0.015887131914496422,
+ "learning_rate": 0.0006,
+ "loss": 4.328040599822998,
+ "step": 4804
+ },
+ {
+ "epoch": 66.74061135371178,
+ "grad_norm": 0.014057286083698273,
+ "learning_rate": 0.0006,
+ "loss": 4.3294548988342285,
+ "step": 4805
+ },
+ {
+ "epoch": 66.75458515283843,
+ "grad_norm": 0.013135905377566814,
+ "learning_rate": 0.0006,
+ "loss": 4.254383563995361,
+ "step": 4806
+ },
+ {
+ "epoch": 66.76855895196506,
+ "grad_norm": 0.013632557354867458,
+ "learning_rate": 0.0006,
+ "loss": 4.342815399169922,
+ "step": 4807
+ },
+ {
+ "epoch": 66.78253275109171,
+ "grad_norm": 0.015512688085436821,
+ "learning_rate": 0.0006,
+ "loss": 4.297835350036621,
+ "step": 4808
+ },
+ {
+ "epoch": 66.79650655021834,
+ "grad_norm": 0.01621832884848118,
+ "learning_rate": 0.0006,
+ "loss": 4.314773082733154,
+ "step": 4809
+ },
+ {
+ "epoch": 66.81048034934499,
+ "grad_norm": 0.016196032986044884,
+ "learning_rate": 0.0006,
+ "loss": 4.29209566116333,
+ "step": 4810
+ },
+ {
+ "epoch": 66.82445414847162,
+ "grad_norm": 0.015463353134691715,
+ "learning_rate": 0.0006,
+ "loss": 4.262795448303223,
+ "step": 4811
+ },
+ {
+ "epoch": 66.83842794759825,
+ "grad_norm": 0.016143787652254105,
+ "learning_rate": 0.0006,
+ "loss": 4.274284362792969,
+ "step": 4812
+ },
+ {
+ "epoch": 66.8524017467249,
+ "grad_norm": 0.01610061153769493,
+ "learning_rate": 0.0006,
+ "loss": 4.337160587310791,
+ "step": 4813
+ },
+ {
+ "epoch": 66.86637554585153,
+ "grad_norm": 0.01582838036119938,
+ "learning_rate": 0.0006,
+ "loss": 4.192723274230957,
+ "step": 4814
+ },
+ {
+ "epoch": 66.88034934497817,
+ "grad_norm": 0.015081540681421757,
+ "learning_rate": 0.0006,
+ "loss": 4.370708465576172,
+ "step": 4815
+ },
+ {
+ "epoch": 66.8943231441048,
+ "grad_norm": 0.015929723158478737,
+ "learning_rate": 0.0006,
+ "loss": 4.340244293212891,
+ "step": 4816
+ },
+ {
+ "epoch": 66.90829694323143,
+ "grad_norm": 0.01784711331129074,
+ "learning_rate": 0.0006,
+ "loss": 4.260406970977783,
+ "step": 4817
+ },
+ {
+ "epoch": 66.92227074235808,
+ "grad_norm": 0.017434703186154366,
+ "learning_rate": 0.0006,
+ "loss": 4.286900043487549,
+ "step": 4818
+ },
+ {
+ "epoch": 66.93624454148471,
+ "grad_norm": 0.016262540593743324,
+ "learning_rate": 0.0006,
+ "loss": 4.292142391204834,
+ "step": 4819
+ },
+ {
+ "epoch": 66.95021834061136,
+ "grad_norm": 0.016122829169034958,
+ "learning_rate": 0.0006,
+ "loss": 4.291485786437988,
+ "step": 4820
+ },
+ {
+ "epoch": 66.96419213973799,
+ "grad_norm": 0.014361178502440453,
+ "learning_rate": 0.0006,
+ "loss": 4.275425910949707,
+ "step": 4821
+ },
+ {
+ "epoch": 66.97816593886463,
+ "grad_norm": 0.01373987551778555,
+ "learning_rate": 0.0006,
+ "loss": 4.223120212554932,
+ "step": 4822
+ },
+ {
+ "epoch": 66.99213973799127,
+ "grad_norm": 0.015687376260757446,
+ "learning_rate": 0.0006,
+ "loss": 4.355381011962891,
+ "step": 4823
+ },
+ {
+ "epoch": 67.0,
+ "grad_norm": 0.01998465321958065,
+ "learning_rate": 0.0006,
+ "loss": 4.264053821563721,
+ "step": 4824
+ },
+ {
+ "epoch": 67.0,
+ "eval_loss": 4.662398338317871,
+ "eval_runtime": 56.7928,
+ "eval_samples_per_second": 42.998,
+ "eval_steps_per_second": 1.356,
+ "step": 4824
+ },
+ {
+ "epoch": 67.01397379912663,
+ "grad_norm": 0.021736836060881615,
+ "learning_rate": 0.0006,
+ "loss": 4.243354797363281,
+ "step": 4825
+ },
+ {
+ "epoch": 67.02794759825328,
+ "grad_norm": 0.021476037800312042,
+ "learning_rate": 0.0006,
+ "loss": 4.218207836151123,
+ "step": 4826
+ },
+ {
+ "epoch": 67.04192139737991,
+ "grad_norm": 0.017188388854265213,
+ "learning_rate": 0.0006,
+ "loss": 4.18048095703125,
+ "step": 4827
+ },
+ {
+ "epoch": 67.05589519650655,
+ "grad_norm": 0.017293719574809074,
+ "learning_rate": 0.0006,
+ "loss": 4.239145278930664,
+ "step": 4828
+ },
+ {
+ "epoch": 67.06986899563319,
+ "grad_norm": 0.017383819445967674,
+ "learning_rate": 0.0006,
+ "loss": 4.114765167236328,
+ "step": 4829
+ },
+ {
+ "epoch": 67.08384279475983,
+ "grad_norm": 0.01651529036462307,
+ "learning_rate": 0.0006,
+ "loss": 4.351023197174072,
+ "step": 4830
+ },
+ {
+ "epoch": 67.09781659388646,
+ "grad_norm": 0.015893027186393738,
+ "learning_rate": 0.0006,
+ "loss": 4.261432647705078,
+ "step": 4831
+ },
+ {
+ "epoch": 67.1117903930131,
+ "grad_norm": 0.013381517492234707,
+ "learning_rate": 0.0006,
+ "loss": 4.172064781188965,
+ "step": 4832
+ },
+ {
+ "epoch": 67.12576419213974,
+ "grad_norm": 0.014642654918134212,
+ "learning_rate": 0.0006,
+ "loss": 4.356171131134033,
+ "step": 4833
+ },
+ {
+ "epoch": 67.13973799126637,
+ "grad_norm": 0.01425282284617424,
+ "learning_rate": 0.0006,
+ "loss": 4.303495407104492,
+ "step": 4834
+ },
+ {
+ "epoch": 67.15371179039302,
+ "grad_norm": 0.014574599452316761,
+ "learning_rate": 0.0006,
+ "loss": 4.2850847244262695,
+ "step": 4835
+ },
+ {
+ "epoch": 67.16768558951965,
+ "grad_norm": 0.012676282785832882,
+ "learning_rate": 0.0006,
+ "loss": 4.297845840454102,
+ "step": 4836
+ },
+ {
+ "epoch": 67.18165938864628,
+ "grad_norm": 0.013948136940598488,
+ "learning_rate": 0.0006,
+ "loss": 4.324530601501465,
+ "step": 4837
+ },
+ {
+ "epoch": 67.19563318777293,
+ "grad_norm": 0.014195773750543594,
+ "learning_rate": 0.0006,
+ "loss": 4.31236457824707,
+ "step": 4838
+ },
+ {
+ "epoch": 67.20960698689956,
+ "grad_norm": 0.013631382025778294,
+ "learning_rate": 0.0006,
+ "loss": 4.086362838745117,
+ "step": 4839
+ },
+ {
+ "epoch": 67.2235807860262,
+ "grad_norm": 0.012798587791621685,
+ "learning_rate": 0.0006,
+ "loss": 4.328471660614014,
+ "step": 4840
+ },
+ {
+ "epoch": 67.23755458515284,
+ "grad_norm": 0.014443321153521538,
+ "learning_rate": 0.0006,
+ "loss": 4.278019905090332,
+ "step": 4841
+ },
+ {
+ "epoch": 67.25152838427948,
+ "grad_norm": 0.015355420298874378,
+ "learning_rate": 0.0006,
+ "loss": 4.333611011505127,
+ "step": 4842
+ },
+ {
+ "epoch": 67.26550218340611,
+ "grad_norm": 0.016487296670675278,
+ "learning_rate": 0.0006,
+ "loss": 4.351336479187012,
+ "step": 4843
+ },
+ {
+ "epoch": 67.27947598253274,
+ "grad_norm": 0.017422541975975037,
+ "learning_rate": 0.0006,
+ "loss": 4.386096477508545,
+ "step": 4844
+ },
+ {
+ "epoch": 67.29344978165939,
+ "grad_norm": 0.015790918841958046,
+ "learning_rate": 0.0006,
+ "loss": 4.301377773284912,
+ "step": 4845
+ },
+ {
+ "epoch": 67.30742358078602,
+ "grad_norm": 0.016427364200353622,
+ "learning_rate": 0.0006,
+ "loss": 4.2838945388793945,
+ "step": 4846
+ },
+ {
+ "epoch": 67.32139737991267,
+ "grad_norm": 0.017050761729478836,
+ "learning_rate": 0.0006,
+ "loss": 4.282417297363281,
+ "step": 4847
+ },
+ {
+ "epoch": 67.3353711790393,
+ "grad_norm": 0.015006253495812416,
+ "learning_rate": 0.0006,
+ "loss": 4.352236270904541,
+ "step": 4848
+ },
+ {
+ "epoch": 67.34934497816595,
+ "grad_norm": 0.012996627949178219,
+ "learning_rate": 0.0006,
+ "loss": 4.309242248535156,
+ "step": 4849
+ },
+ {
+ "epoch": 67.36331877729258,
+ "grad_norm": 0.01367615070194006,
+ "learning_rate": 0.0006,
+ "loss": 4.211801528930664,
+ "step": 4850
+ },
+ {
+ "epoch": 67.37729257641921,
+ "grad_norm": 0.014775963500142097,
+ "learning_rate": 0.0006,
+ "loss": 4.360859394073486,
+ "step": 4851
+ },
+ {
+ "epoch": 67.39126637554585,
+ "grad_norm": 0.015854304656386375,
+ "learning_rate": 0.0006,
+ "loss": 4.211830139160156,
+ "step": 4852
+ },
+ {
+ "epoch": 67.40524017467249,
+ "grad_norm": 0.016888894140720367,
+ "learning_rate": 0.0006,
+ "loss": 4.216549873352051,
+ "step": 4853
+ },
+ {
+ "epoch": 67.41921397379913,
+ "grad_norm": 0.0170894768089056,
+ "learning_rate": 0.0006,
+ "loss": 4.263486385345459,
+ "step": 4854
+ },
+ {
+ "epoch": 67.43318777292576,
+ "grad_norm": 0.015829868614673615,
+ "learning_rate": 0.0006,
+ "loss": 4.360196113586426,
+ "step": 4855
+ },
+ {
+ "epoch": 67.44716157205241,
+ "grad_norm": 0.015220202505588531,
+ "learning_rate": 0.0006,
+ "loss": 4.243794918060303,
+ "step": 4856
+ },
+ {
+ "epoch": 67.46113537117904,
+ "grad_norm": 0.01413388829678297,
+ "learning_rate": 0.0006,
+ "loss": 4.286624908447266,
+ "step": 4857
+ },
+ {
+ "epoch": 67.47510917030567,
+ "grad_norm": 0.01426626369357109,
+ "learning_rate": 0.0006,
+ "loss": 4.257316589355469,
+ "step": 4858
+ },
+ {
+ "epoch": 67.48908296943232,
+ "grad_norm": 0.015957845374941826,
+ "learning_rate": 0.0006,
+ "loss": 4.300443649291992,
+ "step": 4859
+ },
+ {
+ "epoch": 67.50305676855895,
+ "grad_norm": 0.015499051660299301,
+ "learning_rate": 0.0006,
+ "loss": 4.278162002563477,
+ "step": 4860
+ },
+ {
+ "epoch": 67.5170305676856,
+ "grad_norm": 0.015145624987781048,
+ "learning_rate": 0.0006,
+ "loss": 4.267950534820557,
+ "step": 4861
+ },
+ {
+ "epoch": 67.53100436681223,
+ "grad_norm": 0.014610803686082363,
+ "learning_rate": 0.0006,
+ "loss": 4.247499942779541,
+ "step": 4862
+ },
+ {
+ "epoch": 67.54497816593886,
+ "grad_norm": 0.014860909432172775,
+ "learning_rate": 0.0006,
+ "loss": 4.263200759887695,
+ "step": 4863
+ },
+ {
+ "epoch": 67.5589519650655,
+ "grad_norm": 0.01718984916806221,
+ "learning_rate": 0.0006,
+ "loss": 4.290019989013672,
+ "step": 4864
+ },
+ {
+ "epoch": 67.57292576419214,
+ "grad_norm": 0.01558963768184185,
+ "learning_rate": 0.0006,
+ "loss": 4.223816871643066,
+ "step": 4865
+ },
+ {
+ "epoch": 67.58689956331878,
+ "grad_norm": 0.014927188865840435,
+ "learning_rate": 0.0006,
+ "loss": 4.363586902618408,
+ "step": 4866
+ },
+ {
+ "epoch": 67.60087336244541,
+ "grad_norm": 0.01573595404624939,
+ "learning_rate": 0.0006,
+ "loss": 4.254765510559082,
+ "step": 4867
+ },
+ {
+ "epoch": 67.61484716157206,
+ "grad_norm": 0.015618929639458656,
+ "learning_rate": 0.0006,
+ "loss": 4.25072717666626,
+ "step": 4868
+ },
+ {
+ "epoch": 67.62882096069869,
+ "grad_norm": 0.01605590060353279,
+ "learning_rate": 0.0006,
+ "loss": 4.25208854675293,
+ "step": 4869
+ },
+ {
+ "epoch": 67.64279475982532,
+ "grad_norm": 0.01804601214826107,
+ "learning_rate": 0.0006,
+ "loss": 4.351354122161865,
+ "step": 4870
+ },
+ {
+ "epoch": 67.65676855895197,
+ "grad_norm": 0.020595047622919083,
+ "learning_rate": 0.0006,
+ "loss": 4.193732738494873,
+ "step": 4871
+ },
+ {
+ "epoch": 67.6707423580786,
+ "grad_norm": 0.023037966340780258,
+ "learning_rate": 0.0006,
+ "loss": 4.27754020690918,
+ "step": 4872
+ },
+ {
+ "epoch": 67.68471615720524,
+ "grad_norm": 0.020680895075201988,
+ "learning_rate": 0.0006,
+ "loss": 4.314730167388916,
+ "step": 4873
+ },
+ {
+ "epoch": 67.69868995633188,
+ "grad_norm": 0.01887434534728527,
+ "learning_rate": 0.0006,
+ "loss": 4.252305030822754,
+ "step": 4874
+ },
+ {
+ "epoch": 67.71266375545852,
+ "grad_norm": 0.016333596780896187,
+ "learning_rate": 0.0006,
+ "loss": 4.146123886108398,
+ "step": 4875
+ },
+ {
+ "epoch": 67.72663755458515,
+ "grad_norm": 0.015688810497522354,
+ "learning_rate": 0.0006,
+ "loss": 4.2814154624938965,
+ "step": 4876
+ },
+ {
+ "epoch": 67.74061135371178,
+ "grad_norm": 0.016270726919174194,
+ "learning_rate": 0.0006,
+ "loss": 4.323200225830078,
+ "step": 4877
+ },
+ {
+ "epoch": 67.75458515283843,
+ "grad_norm": 0.016824988648295403,
+ "learning_rate": 0.0006,
+ "loss": 4.269501686096191,
+ "step": 4878
+ },
+ {
+ "epoch": 67.76855895196506,
+ "grad_norm": 0.01730562001466751,
+ "learning_rate": 0.0006,
+ "loss": 4.187993049621582,
+ "step": 4879
+ },
+ {
+ "epoch": 67.78253275109171,
+ "grad_norm": 0.01546655036509037,
+ "learning_rate": 0.0006,
+ "loss": 4.340415000915527,
+ "step": 4880
+ },
+ {
+ "epoch": 67.79650655021834,
+ "grad_norm": 0.014173958450555801,
+ "learning_rate": 0.0006,
+ "loss": 4.217419624328613,
+ "step": 4881
+ },
+ {
+ "epoch": 67.81048034934499,
+ "grad_norm": 0.016810474917292595,
+ "learning_rate": 0.0006,
+ "loss": 4.352302074432373,
+ "step": 4882
+ },
+ {
+ "epoch": 67.82445414847162,
+ "grad_norm": 0.0175373163074255,
+ "learning_rate": 0.0006,
+ "loss": 4.20377254486084,
+ "step": 4883
+ },
+ {
+ "epoch": 67.83842794759825,
+ "grad_norm": 0.015734676271677017,
+ "learning_rate": 0.0006,
+ "loss": 4.147495746612549,
+ "step": 4884
+ },
+ {
+ "epoch": 67.8524017467249,
+ "grad_norm": 0.015091817826032639,
+ "learning_rate": 0.0006,
+ "loss": 4.272252082824707,
+ "step": 4885
+ },
+ {
+ "epoch": 67.86637554585153,
+ "grad_norm": 0.015000219456851482,
+ "learning_rate": 0.0006,
+ "loss": 4.302773952484131,
+ "step": 4886
+ },
+ {
+ "epoch": 67.88034934497817,
+ "grad_norm": 0.013827184215188026,
+ "learning_rate": 0.0006,
+ "loss": 4.337242126464844,
+ "step": 4887
+ },
+ {
+ "epoch": 67.8943231441048,
+ "grad_norm": 0.015308346599340439,
+ "learning_rate": 0.0006,
+ "loss": 4.140895366668701,
+ "step": 4888
+ },
+ {
+ "epoch": 67.90829694323143,
+ "grad_norm": 0.015356178395450115,
+ "learning_rate": 0.0006,
+ "loss": 4.312541484832764,
+ "step": 4889
+ },
+ {
+ "epoch": 67.92227074235808,
+ "grad_norm": 0.0137353939935565,
+ "learning_rate": 0.0006,
+ "loss": 4.173184394836426,
+ "step": 4890
+ },
+ {
+ "epoch": 67.93624454148471,
+ "grad_norm": 0.01332368515431881,
+ "learning_rate": 0.0006,
+ "loss": 4.191199779510498,
+ "step": 4891
+ },
+ {
+ "epoch": 67.95021834061136,
+ "grad_norm": 0.01488130260258913,
+ "learning_rate": 0.0006,
+ "loss": 4.331752300262451,
+ "step": 4892
+ },
+ {
+ "epoch": 67.96419213973799,
+ "grad_norm": 0.015449733473360538,
+ "learning_rate": 0.0006,
+ "loss": 4.2385454177856445,
+ "step": 4893
+ },
+ {
+ "epoch": 67.97816593886463,
+ "grad_norm": 0.01556103490293026,
+ "learning_rate": 0.0006,
+ "loss": 4.262353897094727,
+ "step": 4894
+ },
+ {
+ "epoch": 67.99213973799127,
+ "grad_norm": 0.014613849110901356,
+ "learning_rate": 0.0006,
+ "loss": 4.223945617675781,
+ "step": 4895
+ },
+ {
+ "epoch": 68.0,
+ "grad_norm": 0.015323462896049023,
+ "learning_rate": 0.0006,
+ "loss": 4.317168712615967,
+ "step": 4896
+ },
+ {
+ "epoch": 68.0,
+ "eval_loss": 4.671603679656982,
+ "eval_runtime": 56.7317,
+ "eval_samples_per_second": 43.045,
+ "eval_steps_per_second": 1.357,
+ "step": 4896
+ },
+ {
+ "epoch": 68.01397379912663,
+ "grad_norm": 0.01910698413848877,
+ "learning_rate": 0.0006,
+ "loss": 4.206167221069336,
+ "step": 4897
+ },
+ {
+ "epoch": 68.02794759825328,
+ "grad_norm": 0.025112485513091087,
+ "learning_rate": 0.0006,
+ "loss": 4.384706974029541,
+ "step": 4898
+ },
+ {
+ "epoch": 68.04192139737991,
+ "grad_norm": 0.02618720941245556,
+ "learning_rate": 0.0006,
+ "loss": 4.321046829223633,
+ "step": 4899
+ },
+ {
+ "epoch": 68.05589519650655,
+ "grad_norm": 0.020853538066148758,
+ "learning_rate": 0.0006,
+ "loss": 4.217638969421387,
+ "step": 4900
+ },
+ {
+ "epoch": 68.06986899563319,
+ "grad_norm": 0.020025717094540596,
+ "learning_rate": 0.0006,
+ "loss": 4.306824684143066,
+ "step": 4901
+ },
+ {
+ "epoch": 68.08384279475983,
+ "grad_norm": 0.023868650197982788,
+ "learning_rate": 0.0006,
+ "loss": 4.327455997467041,
+ "step": 4902
+ },
+ {
+ "epoch": 68.09781659388646,
+ "grad_norm": 0.022279081866145134,
+ "learning_rate": 0.0006,
+ "loss": 4.400699615478516,
+ "step": 4903
+ },
+ {
+ "epoch": 68.1117903930131,
+ "grad_norm": 0.018965687602758408,
+ "learning_rate": 0.0006,
+ "loss": 4.19097900390625,
+ "step": 4904
+ },
+ {
+ "epoch": 68.12576419213974,
+ "grad_norm": 0.01936378888785839,
+ "learning_rate": 0.0006,
+ "loss": 4.206711769104004,
+ "step": 4905
+ },
+ {
+ "epoch": 68.13973799126637,
+ "grad_norm": 0.01811872608959675,
+ "learning_rate": 0.0006,
+ "loss": 4.204622268676758,
+ "step": 4906
+ },
+ {
+ "epoch": 68.15371179039302,
+ "grad_norm": 0.018530402332544327,
+ "learning_rate": 0.0006,
+ "loss": 4.289602279663086,
+ "step": 4907
+ },
+ {
+ "epoch": 68.16768558951965,
+ "grad_norm": 0.01755962148308754,
+ "learning_rate": 0.0006,
+ "loss": 4.317145347595215,
+ "step": 4908
+ },
+ {
+ "epoch": 68.18165938864628,
+ "grad_norm": 0.020413702353835106,
+ "learning_rate": 0.0006,
+ "loss": 4.2882843017578125,
+ "step": 4909
+ },
+ {
+ "epoch": 68.19563318777293,
+ "grad_norm": 0.020064299926161766,
+ "learning_rate": 0.0006,
+ "loss": 4.248260021209717,
+ "step": 4910
+ },
+ {
+ "epoch": 68.20960698689956,
+ "grad_norm": 0.017415888607501984,
+ "learning_rate": 0.0006,
+ "loss": 4.370295524597168,
+ "step": 4911
+ },
+ {
+ "epoch": 68.2235807860262,
+ "grad_norm": 0.015005858615040779,
+ "learning_rate": 0.0006,
+ "loss": 4.319567680358887,
+ "step": 4912
+ },
+ {
+ "epoch": 68.23755458515284,
+ "grad_norm": 0.016349777579307556,
+ "learning_rate": 0.0006,
+ "loss": 4.301776885986328,
+ "step": 4913
+ },
+ {
+ "epoch": 68.25152838427948,
+ "grad_norm": 0.015779560431838036,
+ "learning_rate": 0.0006,
+ "loss": 4.271171569824219,
+ "step": 4914
+ },
+ {
+ "epoch": 68.26550218340611,
+ "grad_norm": 0.013237644918262959,
+ "learning_rate": 0.0006,
+ "loss": 4.324908256530762,
+ "step": 4915
+ },
+ {
+ "epoch": 68.27947598253274,
+ "grad_norm": 0.014076225459575653,
+ "learning_rate": 0.0006,
+ "loss": 4.177175998687744,
+ "step": 4916
+ },
+ {
+ "epoch": 68.29344978165939,
+ "grad_norm": 0.013906744308769703,
+ "learning_rate": 0.0006,
+ "loss": 4.11588191986084,
+ "step": 4917
+ },
+ {
+ "epoch": 68.30742358078602,
+ "grad_norm": 0.013939064927399158,
+ "learning_rate": 0.0006,
+ "loss": 4.183963775634766,
+ "step": 4918
+ },
+ {
+ "epoch": 68.32139737991267,
+ "grad_norm": 0.014572170563042164,
+ "learning_rate": 0.0006,
+ "loss": 4.19586181640625,
+ "step": 4919
+ },
+ {
+ "epoch": 68.3353711790393,
+ "grad_norm": 0.014085350558161736,
+ "learning_rate": 0.0006,
+ "loss": 4.170892238616943,
+ "step": 4920
+ },
+ {
+ "epoch": 68.34934497816595,
+ "grad_norm": 0.014357863925397396,
+ "learning_rate": 0.0006,
+ "loss": 4.321155548095703,
+ "step": 4921
+ },
+ {
+ "epoch": 68.36331877729258,
+ "grad_norm": 0.01461675763130188,
+ "learning_rate": 0.0006,
+ "loss": 4.293334007263184,
+ "step": 4922
+ },
+ {
+ "epoch": 68.37729257641921,
+ "grad_norm": 0.014828486368060112,
+ "learning_rate": 0.0006,
+ "loss": 4.1439714431762695,
+ "step": 4923
+ },
+ {
+ "epoch": 68.39126637554585,
+ "grad_norm": 0.013730363920331001,
+ "learning_rate": 0.0006,
+ "loss": 4.287973880767822,
+ "step": 4924
+ },
+ {
+ "epoch": 68.40524017467249,
+ "grad_norm": 0.013852203264832497,
+ "learning_rate": 0.0006,
+ "loss": 4.248824119567871,
+ "step": 4925
+ },
+ {
+ "epoch": 68.41921397379913,
+ "grad_norm": 0.014005862176418304,
+ "learning_rate": 0.0006,
+ "loss": 4.293768882751465,
+ "step": 4926
+ },
+ {
+ "epoch": 68.43318777292576,
+ "grad_norm": 0.014435851015150547,
+ "learning_rate": 0.0006,
+ "loss": 4.204813003540039,
+ "step": 4927
+ },
+ {
+ "epoch": 68.44716157205241,
+ "grad_norm": 0.015305934473872185,
+ "learning_rate": 0.0006,
+ "loss": 4.329364776611328,
+ "step": 4928
+ },
+ {
+ "epoch": 68.46113537117904,
+ "grad_norm": 0.014356814324855804,
+ "learning_rate": 0.0006,
+ "loss": 4.264206409454346,
+ "step": 4929
+ },
+ {
+ "epoch": 68.47510917030567,
+ "grad_norm": 0.01432347297668457,
+ "learning_rate": 0.0006,
+ "loss": 4.29440450668335,
+ "step": 4930
+ },
+ {
+ "epoch": 68.48908296943232,
+ "grad_norm": 0.012672432698309422,
+ "learning_rate": 0.0006,
+ "loss": 4.258024215698242,
+ "step": 4931
+ },
+ {
+ "epoch": 68.50305676855895,
+ "grad_norm": 0.014209977351129055,
+ "learning_rate": 0.0006,
+ "loss": 4.216668128967285,
+ "step": 4932
+ },
+ {
+ "epoch": 68.5170305676856,
+ "grad_norm": 0.014606663957238197,
+ "learning_rate": 0.0006,
+ "loss": 4.294184684753418,
+ "step": 4933
+ },
+ {
+ "epoch": 68.53100436681223,
+ "grad_norm": 0.014986990951001644,
+ "learning_rate": 0.0006,
+ "loss": 4.382591247558594,
+ "step": 4934
+ },
+ {
+ "epoch": 68.54497816593886,
+ "grad_norm": 0.013998783193528652,
+ "learning_rate": 0.0006,
+ "loss": 4.195143222808838,
+ "step": 4935
+ },
+ {
+ "epoch": 68.5589519650655,
+ "grad_norm": 0.015011102892458439,
+ "learning_rate": 0.0006,
+ "loss": 4.313785076141357,
+ "step": 4936
+ },
+ {
+ "epoch": 68.57292576419214,
+ "grad_norm": 0.01641209051012993,
+ "learning_rate": 0.0006,
+ "loss": 4.322803020477295,
+ "step": 4937
+ },
+ {
+ "epoch": 68.58689956331878,
+ "grad_norm": 0.016207464039325714,
+ "learning_rate": 0.0006,
+ "loss": 4.311431884765625,
+ "step": 4938
+ },
+ {
+ "epoch": 68.60087336244541,
+ "grad_norm": 0.01747811958193779,
+ "learning_rate": 0.0006,
+ "loss": 4.310845375061035,
+ "step": 4939
+ },
+ {
+ "epoch": 68.61484716157206,
+ "grad_norm": 0.019475717097520828,
+ "learning_rate": 0.0006,
+ "loss": 4.338624954223633,
+ "step": 4940
+ },
+ {
+ "epoch": 68.62882096069869,
+ "grad_norm": 0.018641719594597816,
+ "learning_rate": 0.0006,
+ "loss": 4.34913969039917,
+ "step": 4941
+ },
+ {
+ "epoch": 68.64279475982532,
+ "grad_norm": 0.018376432359218597,
+ "learning_rate": 0.0006,
+ "loss": 4.28464412689209,
+ "step": 4942
+ },
+ {
+ "epoch": 68.65676855895197,
+ "grad_norm": 0.017095010727643967,
+ "learning_rate": 0.0006,
+ "loss": 4.217957496643066,
+ "step": 4943
+ },
+ {
+ "epoch": 68.6707423580786,
+ "grad_norm": 0.016321983188390732,
+ "learning_rate": 0.0006,
+ "loss": 4.220042705535889,
+ "step": 4944
+ },
+ {
+ "epoch": 68.68471615720524,
+ "grad_norm": 0.017128000035881996,
+ "learning_rate": 0.0006,
+ "loss": 4.293500900268555,
+ "step": 4945
+ },
+ {
+ "epoch": 68.69868995633188,
+ "grad_norm": 0.015592411160469055,
+ "learning_rate": 0.0006,
+ "loss": 4.195798873901367,
+ "step": 4946
+ },
+ {
+ "epoch": 68.71266375545852,
+ "grad_norm": 0.016607079654932022,
+ "learning_rate": 0.0006,
+ "loss": 4.2654337882995605,
+ "step": 4947
+ },
+ {
+ "epoch": 68.72663755458515,
+ "grad_norm": 0.01545762736350298,
+ "learning_rate": 0.0006,
+ "loss": 4.303828239440918,
+ "step": 4948
+ },
+ {
+ "epoch": 68.74061135371178,
+ "grad_norm": 0.014328422956168652,
+ "learning_rate": 0.0006,
+ "loss": 4.29379940032959,
+ "step": 4949
+ },
+ {
+ "epoch": 68.75458515283843,
+ "grad_norm": 0.016516804695129395,
+ "learning_rate": 0.0006,
+ "loss": 4.2317214012146,
+ "step": 4950
+ },
+ {
+ "epoch": 68.76855895196506,
+ "grad_norm": 0.016324171796441078,
+ "learning_rate": 0.0006,
+ "loss": 4.278043746948242,
+ "step": 4951
+ },
+ {
+ "epoch": 68.78253275109171,
+ "grad_norm": 0.015511159785091877,
+ "learning_rate": 0.0006,
+ "loss": 4.188244342803955,
+ "step": 4952
+ },
+ {
+ "epoch": 68.79650655021834,
+ "grad_norm": 0.01676103286445141,
+ "learning_rate": 0.0006,
+ "loss": 4.2169976234436035,
+ "step": 4953
+ },
+ {
+ "epoch": 68.81048034934499,
+ "grad_norm": 0.015428312122821808,
+ "learning_rate": 0.0006,
+ "loss": 4.4004716873168945,
+ "step": 4954
+ },
+ {
+ "epoch": 68.82445414847162,
+ "grad_norm": 0.0162410456687212,
+ "learning_rate": 0.0006,
+ "loss": 4.282752513885498,
+ "step": 4955
+ },
+ {
+ "epoch": 68.83842794759825,
+ "grad_norm": 0.017578862607479095,
+ "learning_rate": 0.0006,
+ "loss": 4.098004341125488,
+ "step": 4956
+ },
+ {
+ "epoch": 68.8524017467249,
+ "grad_norm": 0.01805749163031578,
+ "learning_rate": 0.0006,
+ "loss": 4.277988433837891,
+ "step": 4957
+ },
+ {
+ "epoch": 68.86637554585153,
+ "grad_norm": 0.016728045418858528,
+ "learning_rate": 0.0006,
+ "loss": 4.29766321182251,
+ "step": 4958
+ },
+ {
+ "epoch": 68.88034934497817,
+ "grad_norm": 0.014899657107889652,
+ "learning_rate": 0.0006,
+ "loss": 4.329736709594727,
+ "step": 4959
+ },
+ {
+ "epoch": 68.8943231441048,
+ "grad_norm": 0.013441305607557297,
+ "learning_rate": 0.0006,
+ "loss": 4.2873215675354,
+ "step": 4960
+ },
+ {
+ "epoch": 68.90829694323143,
+ "grad_norm": 0.01554128248244524,
+ "learning_rate": 0.0006,
+ "loss": 4.278327941894531,
+ "step": 4961
+ },
+ {
+ "epoch": 68.92227074235808,
+ "grad_norm": 0.01806696318089962,
+ "learning_rate": 0.0006,
+ "loss": 4.38275146484375,
+ "step": 4962
+ },
+ {
+ "epoch": 68.93624454148471,
+ "grad_norm": 0.017025409266352654,
+ "learning_rate": 0.0006,
+ "loss": 4.210447311401367,
+ "step": 4963
+ },
+ {
+ "epoch": 68.95021834061136,
+ "grad_norm": 0.01597420684993267,
+ "learning_rate": 0.0006,
+ "loss": 4.309865951538086,
+ "step": 4964
+ },
+ {
+ "epoch": 68.96419213973799,
+ "grad_norm": 0.015835624188184738,
+ "learning_rate": 0.0006,
+ "loss": 4.225904941558838,
+ "step": 4965
+ },
+ {
+ "epoch": 68.97816593886463,
+ "grad_norm": 0.018170492723584175,
+ "learning_rate": 0.0006,
+ "loss": 4.248478889465332,
+ "step": 4966
+ },
+ {
+ "epoch": 68.99213973799127,
+ "grad_norm": 0.017372991889715195,
+ "learning_rate": 0.0006,
+ "loss": 4.283486843109131,
+ "step": 4967
+ },
+ {
+ "epoch": 69.0,
+ "grad_norm": 0.016305390745401382,
+ "learning_rate": 0.0006,
+ "loss": 4.273636817932129,
+ "step": 4968
+ },
+ {
+ "epoch": 69.0,
+ "eval_loss": 4.66873836517334,
+ "eval_runtime": 57.0791,
+ "eval_samples_per_second": 42.783,
+ "eval_steps_per_second": 1.349,
+ "step": 4968
+ },
+ {
+ "epoch": 69.01397379912663,
+ "grad_norm": 0.016567355021834373,
+ "learning_rate": 0.0006,
+ "loss": 4.269988059997559,
+ "step": 4969
+ },
+ {
+ "epoch": 69.02794759825328,
+ "grad_norm": 0.017476852983236313,
+ "learning_rate": 0.0006,
+ "loss": 4.2518720626831055,
+ "step": 4970
+ },
+ {
+ "epoch": 69.04192139737991,
+ "grad_norm": 0.017070338129997253,
+ "learning_rate": 0.0006,
+ "loss": 4.238024711608887,
+ "step": 4971
+ },
+ {
+ "epoch": 69.05589519650655,
+ "grad_norm": 0.01639373227953911,
+ "learning_rate": 0.0006,
+ "loss": 4.188558101654053,
+ "step": 4972
+ },
+ {
+ "epoch": 69.06986899563319,
+ "grad_norm": 0.01856113411486149,
+ "learning_rate": 0.0006,
+ "loss": 4.339556694030762,
+ "step": 4973
+ },
+ {
+ "epoch": 69.08384279475983,
+ "grad_norm": 0.01762578636407852,
+ "learning_rate": 0.0006,
+ "loss": 4.141518592834473,
+ "step": 4974
+ },
+ {
+ "epoch": 69.09781659388646,
+ "grad_norm": 0.015437949448823929,
+ "learning_rate": 0.0006,
+ "loss": 4.225461959838867,
+ "step": 4975
+ },
+ {
+ "epoch": 69.1117903930131,
+ "grad_norm": 0.015293503180146217,
+ "learning_rate": 0.0006,
+ "loss": 4.290884971618652,
+ "step": 4976
+ },
+ {
+ "epoch": 69.12576419213974,
+ "grad_norm": 0.016259880736470222,
+ "learning_rate": 0.0006,
+ "loss": 4.121919631958008,
+ "step": 4977
+ },
+ {
+ "epoch": 69.13973799126637,
+ "grad_norm": 0.017225584015250206,
+ "learning_rate": 0.0006,
+ "loss": 4.136716842651367,
+ "step": 4978
+ },
+ {
+ "epoch": 69.15371179039302,
+ "grad_norm": 0.01699289306998253,
+ "learning_rate": 0.0006,
+ "loss": 4.193601608276367,
+ "step": 4979
+ },
+ {
+ "epoch": 69.16768558951965,
+ "grad_norm": 0.015613161027431488,
+ "learning_rate": 0.0006,
+ "loss": 4.288212299346924,
+ "step": 4980
+ },
+ {
+ "epoch": 69.18165938864628,
+ "grad_norm": 0.01643933728337288,
+ "learning_rate": 0.0006,
+ "loss": 4.324707984924316,
+ "step": 4981
+ },
+ {
+ "epoch": 69.19563318777293,
+ "grad_norm": 0.01648905873298645,
+ "learning_rate": 0.0006,
+ "loss": 4.287806510925293,
+ "step": 4982
+ },
+ {
+ "epoch": 69.20960698689956,
+ "grad_norm": 0.015444950200617313,
+ "learning_rate": 0.0006,
+ "loss": 4.200622081756592,
+ "step": 4983
+ },
+ {
+ "epoch": 69.2235807860262,
+ "grad_norm": 0.016034258529543877,
+ "learning_rate": 0.0006,
+ "loss": 4.2193827629089355,
+ "step": 4984
+ },
+ {
+ "epoch": 69.23755458515284,
+ "grad_norm": 0.017984388396143913,
+ "learning_rate": 0.0006,
+ "loss": 4.18237829208374,
+ "step": 4985
+ },
+ {
+ "epoch": 69.25152838427948,
+ "grad_norm": 0.0195505041629076,
+ "learning_rate": 0.0006,
+ "loss": 4.216727256774902,
+ "step": 4986
+ },
+ {
+ "epoch": 69.26550218340611,
+ "grad_norm": 0.017779147252440453,
+ "learning_rate": 0.0006,
+ "loss": 4.278628826141357,
+ "step": 4987
+ },
+ {
+ "epoch": 69.27947598253274,
+ "grad_norm": 0.017119476571679115,
+ "learning_rate": 0.0006,
+ "loss": 4.2409138679504395,
+ "step": 4988
+ },
+ {
+ "epoch": 69.29344978165939,
+ "grad_norm": 0.016840364784002304,
+ "learning_rate": 0.0006,
+ "loss": 4.207575798034668,
+ "step": 4989
+ },
+ {
+ "epoch": 69.30742358078602,
+ "grad_norm": 0.014593573287129402,
+ "learning_rate": 0.0006,
+ "loss": 4.164905548095703,
+ "step": 4990
+ },
+ {
+ "epoch": 69.32139737991267,
+ "grad_norm": 0.013922630809247494,
+ "learning_rate": 0.0006,
+ "loss": 4.2696990966796875,
+ "step": 4991
+ },
+ {
+ "epoch": 69.3353711790393,
+ "grad_norm": 0.01577564887702465,
+ "learning_rate": 0.0006,
+ "loss": 4.201589107513428,
+ "step": 4992
+ },
+ {
+ "epoch": 69.34934497816595,
+ "grad_norm": 0.016392096877098083,
+ "learning_rate": 0.0006,
+ "loss": 4.248831748962402,
+ "step": 4993
+ },
+ {
+ "epoch": 69.36331877729258,
+ "grad_norm": 0.015961311757564545,
+ "learning_rate": 0.0006,
+ "loss": 4.181260108947754,
+ "step": 4994
+ },
+ {
+ "epoch": 69.37729257641921,
+ "grad_norm": 0.015350551344454288,
+ "learning_rate": 0.0006,
+ "loss": 4.230352878570557,
+ "step": 4995
+ },
+ {
+ "epoch": 69.39126637554585,
+ "grad_norm": 0.015870466828346252,
+ "learning_rate": 0.0006,
+ "loss": 4.298305511474609,
+ "step": 4996
+ },
+ {
+ "epoch": 69.40524017467249,
+ "grad_norm": 0.013716642744839191,
+ "learning_rate": 0.0006,
+ "loss": 4.220139503479004,
+ "step": 4997
+ },
+ {
+ "epoch": 69.41921397379913,
+ "grad_norm": 0.014431777410209179,
+ "learning_rate": 0.0006,
+ "loss": 4.288130760192871,
+ "step": 4998
+ },
+ {
+ "epoch": 69.43318777292576,
+ "grad_norm": 0.01426647324115038,
+ "learning_rate": 0.0006,
+ "loss": 4.295767784118652,
+ "step": 4999
+ },
+ {
+ "epoch": 69.44716157205241,
+ "grad_norm": 0.014286615885794163,
+ "learning_rate": 0.0006,
+ "loss": 4.38106632232666,
+ "step": 5000
+ },
+ {
+ "epoch": 69.46113537117904,
+ "grad_norm": 0.017990585416555405,
+ "learning_rate": 0.0006,
+ "loss": 4.1593017578125,
+ "step": 5001
+ },
+ {
+ "epoch": 69.47510917030567,
+ "grad_norm": 0.018846049904823303,
+ "learning_rate": 0.0006,
+ "loss": 4.216091632843018,
+ "step": 5002
+ },
+ {
+ "epoch": 69.48908296943232,
+ "grad_norm": 0.018179181963205338,
+ "learning_rate": 0.0006,
+ "loss": 4.240332126617432,
+ "step": 5003
+ },
+ {
+ "epoch": 69.50305676855895,
+ "grad_norm": 0.016693448647856712,
+ "learning_rate": 0.0006,
+ "loss": 4.256624221801758,
+ "step": 5004
+ },
+ {
+ "epoch": 69.5170305676856,
+ "grad_norm": 0.01718989387154579,
+ "learning_rate": 0.0006,
+ "loss": 4.300940990447998,
+ "step": 5005
+ },
+ {
+ "epoch": 69.53100436681223,
+ "grad_norm": 0.020236829295754433,
+ "learning_rate": 0.0006,
+ "loss": 4.289146423339844,
+ "step": 5006
+ },
+ {
+ "epoch": 69.54497816593886,
+ "grad_norm": 0.024860285222530365,
+ "learning_rate": 0.0006,
+ "loss": 4.216216087341309,
+ "step": 5007
+ },
+ {
+ "epoch": 69.5589519650655,
+ "grad_norm": 0.02559969387948513,
+ "learning_rate": 0.0006,
+ "loss": 4.262890815734863,
+ "step": 5008
+ },
+ {
+ "epoch": 69.57292576419214,
+ "grad_norm": 0.019180957227945328,
+ "learning_rate": 0.0006,
+ "loss": 4.230309009552002,
+ "step": 5009
+ },
+ {
+ "epoch": 69.58689956331878,
+ "grad_norm": 0.020442591980099678,
+ "learning_rate": 0.0006,
+ "loss": 4.291759490966797,
+ "step": 5010
+ },
+ {
+ "epoch": 69.60087336244541,
+ "grad_norm": 0.0212619137018919,
+ "learning_rate": 0.0006,
+ "loss": 4.3529839515686035,
+ "step": 5011
+ },
+ {
+ "epoch": 69.61484716157206,
+ "grad_norm": 0.02067175693809986,
+ "learning_rate": 0.0006,
+ "loss": 4.33197021484375,
+ "step": 5012
+ },
+ {
+ "epoch": 69.62882096069869,
+ "grad_norm": 0.01802872307598591,
+ "learning_rate": 0.0006,
+ "loss": 4.2101969718933105,
+ "step": 5013
+ },
+ {
+ "epoch": 69.64279475982532,
+ "grad_norm": 0.01787702925503254,
+ "learning_rate": 0.0006,
+ "loss": 4.270336627960205,
+ "step": 5014
+ },
+ {
+ "epoch": 69.65676855895197,
+ "grad_norm": 0.01586683839559555,
+ "learning_rate": 0.0006,
+ "loss": 4.222309589385986,
+ "step": 5015
+ },
+ {
+ "epoch": 69.6707423580786,
+ "grad_norm": 0.01717999391257763,
+ "learning_rate": 0.0006,
+ "loss": 4.253704071044922,
+ "step": 5016
+ },
+ {
+ "epoch": 69.68471615720524,
+ "grad_norm": 0.01594279147684574,
+ "learning_rate": 0.0006,
+ "loss": 4.344658851623535,
+ "step": 5017
+ },
+ {
+ "epoch": 69.69868995633188,
+ "grad_norm": 0.014419942162930965,
+ "learning_rate": 0.0006,
+ "loss": 4.2144880294799805,
+ "step": 5018
+ },
+ {
+ "epoch": 69.71266375545852,
+ "grad_norm": 0.015824224799871445,
+ "learning_rate": 0.0006,
+ "loss": 4.226252555847168,
+ "step": 5019
+ },
+ {
+ "epoch": 69.72663755458515,
+ "grad_norm": 0.013329979032278061,
+ "learning_rate": 0.0006,
+ "loss": 4.238587379455566,
+ "step": 5020
+ },
+ {
+ "epoch": 69.74061135371178,
+ "grad_norm": 0.013745979405939579,
+ "learning_rate": 0.0006,
+ "loss": 4.317209243774414,
+ "step": 5021
+ },
+ {
+ "epoch": 69.75458515283843,
+ "grad_norm": 0.014064167626202106,
+ "learning_rate": 0.0006,
+ "loss": 4.316982746124268,
+ "step": 5022
+ },
+ {
+ "epoch": 69.76855895196506,
+ "grad_norm": 0.014611301012337208,
+ "learning_rate": 0.0006,
+ "loss": 4.212728500366211,
+ "step": 5023
+ },
+ {
+ "epoch": 69.78253275109171,
+ "grad_norm": 0.014947416260838509,
+ "learning_rate": 0.0006,
+ "loss": 4.319550514221191,
+ "step": 5024
+ },
+ {
+ "epoch": 69.79650655021834,
+ "grad_norm": 0.014780526980757713,
+ "learning_rate": 0.0006,
+ "loss": 4.1888933181762695,
+ "step": 5025
+ },
+ {
+ "epoch": 69.81048034934499,
+ "grad_norm": 0.01611459068953991,
+ "learning_rate": 0.0006,
+ "loss": 4.239692211151123,
+ "step": 5026
+ },
+ {
+ "epoch": 69.82445414847162,
+ "grad_norm": 0.015395848080515862,
+ "learning_rate": 0.0006,
+ "loss": 4.308794021606445,
+ "step": 5027
+ },
+ {
+ "epoch": 69.83842794759825,
+ "grad_norm": 0.016113396733999252,
+ "learning_rate": 0.0006,
+ "loss": 4.328035354614258,
+ "step": 5028
+ },
+ {
+ "epoch": 69.8524017467249,
+ "grad_norm": 0.016386888921260834,
+ "learning_rate": 0.0006,
+ "loss": 4.272579193115234,
+ "step": 5029
+ },
+ {
+ "epoch": 69.86637554585153,
+ "grad_norm": 0.013675205409526825,
+ "learning_rate": 0.0006,
+ "loss": 4.226552963256836,
+ "step": 5030
+ },
+ {
+ "epoch": 69.88034934497817,
+ "grad_norm": 0.01472668256610632,
+ "learning_rate": 0.0006,
+ "loss": 4.224167346954346,
+ "step": 5031
+ },
+ {
+ "epoch": 69.8943231441048,
+ "grad_norm": 0.014958228915929794,
+ "learning_rate": 0.0006,
+ "loss": 4.232619285583496,
+ "step": 5032
+ },
+ {
+ "epoch": 69.90829694323143,
+ "grad_norm": 0.012926832772791386,
+ "learning_rate": 0.0006,
+ "loss": 4.27107048034668,
+ "step": 5033
+ },
+ {
+ "epoch": 69.92227074235808,
+ "grad_norm": 0.013138801790773869,
+ "learning_rate": 0.0006,
+ "loss": 4.288477420806885,
+ "step": 5034
+ },
+ {
+ "epoch": 69.93624454148471,
+ "grad_norm": 0.013415461406111717,
+ "learning_rate": 0.0006,
+ "loss": 4.2033257484436035,
+ "step": 5035
+ },
+ {
+ "epoch": 69.95021834061136,
+ "grad_norm": 0.013807233422994614,
+ "learning_rate": 0.0006,
+ "loss": 4.274764537811279,
+ "step": 5036
+ },
+ {
+ "epoch": 69.96419213973799,
+ "grad_norm": 0.013929193839430809,
+ "learning_rate": 0.0006,
+ "loss": 4.1854352951049805,
+ "step": 5037
+ },
+ {
+ "epoch": 69.97816593886463,
+ "grad_norm": 0.015497001819312572,
+ "learning_rate": 0.0006,
+ "loss": 4.226498126983643,
+ "step": 5038
+ },
+ {
+ "epoch": 69.99213973799127,
+ "grad_norm": 0.01874048449099064,
+ "learning_rate": 0.0006,
+ "loss": 4.304137706756592,
+ "step": 5039
+ },
+ {
+ "epoch": 70.0,
+ "grad_norm": 0.021745791658759117,
+ "learning_rate": 0.0006,
+ "loss": 4.206809997558594,
+ "step": 5040
+ },
+ {
+ "epoch": 70.0,
+ "eval_loss": 4.675539970397949,
+ "eval_runtime": 56.9034,
+ "eval_samples_per_second": 42.915,
+ "eval_steps_per_second": 1.353,
+ "step": 5040
+ },
+ {
+ "epoch": 70.01397379912663,
+ "grad_norm": 0.02124839462339878,
+ "learning_rate": 0.0006,
+ "loss": 4.198199272155762,
+ "step": 5041
+ },
+ {
+ "epoch": 70.02794759825328,
+ "grad_norm": 0.019432688131928444,
+ "learning_rate": 0.0006,
+ "loss": 4.187261581420898,
+ "step": 5042
+ },
+ {
+ "epoch": 70.04192139737991,
+ "grad_norm": 0.01769290305674076,
+ "learning_rate": 0.0006,
+ "loss": 4.155284881591797,
+ "step": 5043
+ },
+ {
+ "epoch": 70.05589519650655,
+ "grad_norm": 0.01751520484685898,
+ "learning_rate": 0.0006,
+ "loss": 4.386857986450195,
+ "step": 5044
+ },
+ {
+ "epoch": 70.06986899563319,
+ "grad_norm": 0.018138699233531952,
+ "learning_rate": 0.0006,
+ "loss": 4.117342472076416,
+ "step": 5045
+ },
+ {
+ "epoch": 70.08384279475983,
+ "grad_norm": 0.016331054270267487,
+ "learning_rate": 0.0006,
+ "loss": 4.225295066833496,
+ "step": 5046
+ },
+ {
+ "epoch": 70.09781659388646,
+ "grad_norm": 0.01658729277551174,
+ "learning_rate": 0.0006,
+ "loss": 4.143835067749023,
+ "step": 5047
+ },
+ {
+ "epoch": 70.1117903930131,
+ "grad_norm": 0.016367698088288307,
+ "learning_rate": 0.0006,
+ "loss": 4.1469268798828125,
+ "step": 5048
+ },
+ {
+ "epoch": 70.12576419213974,
+ "grad_norm": 0.015151629224419594,
+ "learning_rate": 0.0006,
+ "loss": 4.117629051208496,
+ "step": 5049
+ },
+ {
+ "epoch": 70.13973799126637,
+ "grad_norm": 0.016434503719210625,
+ "learning_rate": 0.0006,
+ "loss": 4.107848644256592,
+ "step": 5050
+ },
+ {
+ "epoch": 70.15371179039302,
+ "grad_norm": 0.017837485298514366,
+ "learning_rate": 0.0006,
+ "loss": 4.10837459564209,
+ "step": 5051
+ },
+ {
+ "epoch": 70.16768558951965,
+ "grad_norm": 0.0165961142629385,
+ "learning_rate": 0.0006,
+ "loss": 4.169414043426514,
+ "step": 5052
+ },
+ {
+ "epoch": 70.18165938864628,
+ "grad_norm": 0.0157014112919569,
+ "learning_rate": 0.0006,
+ "loss": 4.0629706382751465,
+ "step": 5053
+ },
+ {
+ "epoch": 70.19563318777293,
+ "grad_norm": 0.01581883803009987,
+ "learning_rate": 0.0006,
+ "loss": 4.220012664794922,
+ "step": 5054
+ },
+ {
+ "epoch": 70.20960698689956,
+ "grad_norm": 0.015285917557775974,
+ "learning_rate": 0.0006,
+ "loss": 4.275337219238281,
+ "step": 5055
+ },
+ {
+ "epoch": 70.2235807860262,
+ "grad_norm": 0.014984549954533577,
+ "learning_rate": 0.0006,
+ "loss": 4.284764289855957,
+ "step": 5056
+ },
+ {
+ "epoch": 70.23755458515284,
+ "grad_norm": 0.015172012150287628,
+ "learning_rate": 0.0006,
+ "loss": 4.234498023986816,
+ "step": 5057
+ },
+ {
+ "epoch": 70.25152838427948,
+ "grad_norm": 0.016846632584929466,
+ "learning_rate": 0.0006,
+ "loss": 4.181689262390137,
+ "step": 5058
+ },
+ {
+ "epoch": 70.26550218340611,
+ "grad_norm": 0.01920711249113083,
+ "learning_rate": 0.0006,
+ "loss": 4.235280513763428,
+ "step": 5059
+ },
+ {
+ "epoch": 70.27947598253274,
+ "grad_norm": 0.017765291035175323,
+ "learning_rate": 0.0006,
+ "loss": 4.238955497741699,
+ "step": 5060
+ },
+ {
+ "epoch": 70.29344978165939,
+ "grad_norm": 0.015436639077961445,
+ "learning_rate": 0.0006,
+ "loss": 4.232516288757324,
+ "step": 5061
+ },
+ {
+ "epoch": 70.30742358078602,
+ "grad_norm": 0.015660086646676064,
+ "learning_rate": 0.0006,
+ "loss": 4.260349273681641,
+ "step": 5062
+ },
+ {
+ "epoch": 70.32139737991267,
+ "grad_norm": 0.016095709055662155,
+ "learning_rate": 0.0006,
+ "loss": 4.168505668640137,
+ "step": 5063
+ },
+ {
+ "epoch": 70.3353711790393,
+ "grad_norm": 0.015001940540969372,
+ "learning_rate": 0.0006,
+ "loss": 4.161739349365234,
+ "step": 5064
+ },
+ {
+ "epoch": 70.34934497816595,
+ "grad_norm": 0.014668774791061878,
+ "learning_rate": 0.0006,
+ "loss": 4.181029796600342,
+ "step": 5065
+ },
+ {
+ "epoch": 70.36331877729258,
+ "grad_norm": 0.013266284950077534,
+ "learning_rate": 0.0006,
+ "loss": 4.3052873611450195,
+ "step": 5066
+ },
+ {
+ "epoch": 70.37729257641921,
+ "grad_norm": 0.015396138653159142,
+ "learning_rate": 0.0006,
+ "loss": 4.34413480758667,
+ "step": 5067
+ },
+ {
+ "epoch": 70.39126637554585,
+ "grad_norm": 0.015511523932218552,
+ "learning_rate": 0.0006,
+ "loss": 4.235037803649902,
+ "step": 5068
+ },
+ {
+ "epoch": 70.40524017467249,
+ "grad_norm": 0.014825738966464996,
+ "learning_rate": 0.0006,
+ "loss": 4.262448310852051,
+ "step": 5069
+ },
+ {
+ "epoch": 70.41921397379913,
+ "grad_norm": 0.015309952199459076,
+ "learning_rate": 0.0006,
+ "loss": 4.258711814880371,
+ "step": 5070
+ },
+ {
+ "epoch": 70.43318777292576,
+ "grad_norm": 0.017614055424928665,
+ "learning_rate": 0.0006,
+ "loss": 4.286028861999512,
+ "step": 5071
+ },
+ {
+ "epoch": 70.44716157205241,
+ "grad_norm": 0.0166039876639843,
+ "learning_rate": 0.0006,
+ "loss": 4.149197578430176,
+ "step": 5072
+ },
+ {
+ "epoch": 70.46113537117904,
+ "grad_norm": 0.016911419108510017,
+ "learning_rate": 0.0006,
+ "loss": 4.271981239318848,
+ "step": 5073
+ },
+ {
+ "epoch": 70.47510917030567,
+ "grad_norm": 0.016806820407509804,
+ "learning_rate": 0.0006,
+ "loss": 4.213315010070801,
+ "step": 5074
+ },
+ {
+ "epoch": 70.48908296943232,
+ "grad_norm": 0.015221742913126945,
+ "learning_rate": 0.0006,
+ "loss": 4.171091079711914,
+ "step": 5075
+ },
+ {
+ "epoch": 70.50305676855895,
+ "grad_norm": 0.014175347983837128,
+ "learning_rate": 0.0006,
+ "loss": 4.185309410095215,
+ "step": 5076
+ },
+ {
+ "epoch": 70.5170305676856,
+ "grad_norm": 0.013897798955440521,
+ "learning_rate": 0.0006,
+ "loss": 4.250119209289551,
+ "step": 5077
+ },
+ {
+ "epoch": 70.53100436681223,
+ "grad_norm": 0.014791003428399563,
+ "learning_rate": 0.0006,
+ "loss": 4.2504682540893555,
+ "step": 5078
+ },
+ {
+ "epoch": 70.54497816593886,
+ "grad_norm": 0.016700128093361855,
+ "learning_rate": 0.0006,
+ "loss": 4.235134124755859,
+ "step": 5079
+ },
+ {
+ "epoch": 70.5589519650655,
+ "grad_norm": 0.016679363325238228,
+ "learning_rate": 0.0006,
+ "loss": 4.32028865814209,
+ "step": 5080
+ },
+ {
+ "epoch": 70.57292576419214,
+ "grad_norm": 0.015117401257157326,
+ "learning_rate": 0.0006,
+ "loss": 4.210995674133301,
+ "step": 5081
+ },
+ {
+ "epoch": 70.58689956331878,
+ "grad_norm": 0.014441263861954212,
+ "learning_rate": 0.0006,
+ "loss": 4.27849006652832,
+ "step": 5082
+ },
+ {
+ "epoch": 70.60087336244541,
+ "grad_norm": 0.01523780357092619,
+ "learning_rate": 0.0006,
+ "loss": 4.197360992431641,
+ "step": 5083
+ },
+ {
+ "epoch": 70.61484716157206,
+ "grad_norm": 0.016201738268136978,
+ "learning_rate": 0.0006,
+ "loss": 4.28239107131958,
+ "step": 5084
+ },
+ {
+ "epoch": 70.62882096069869,
+ "grad_norm": 0.016003292053937912,
+ "learning_rate": 0.0006,
+ "loss": 4.209715843200684,
+ "step": 5085
+ },
+ {
+ "epoch": 70.64279475982532,
+ "grad_norm": 0.01472893450409174,
+ "learning_rate": 0.0006,
+ "loss": 4.169078826904297,
+ "step": 5086
+ },
+ {
+ "epoch": 70.65676855895197,
+ "grad_norm": 0.01604326069355011,
+ "learning_rate": 0.0006,
+ "loss": 4.090967178344727,
+ "step": 5087
+ },
+ {
+ "epoch": 70.6707423580786,
+ "grad_norm": 0.01598356105387211,
+ "learning_rate": 0.0006,
+ "loss": 4.16630220413208,
+ "step": 5088
+ },
+ {
+ "epoch": 70.68471615720524,
+ "grad_norm": 0.016988780349493027,
+ "learning_rate": 0.0006,
+ "loss": 4.117822170257568,
+ "step": 5089
+ },
+ {
+ "epoch": 70.69868995633188,
+ "grad_norm": 0.018582893535494804,
+ "learning_rate": 0.0006,
+ "loss": 4.192902088165283,
+ "step": 5090
+ },
+ {
+ "epoch": 70.71266375545852,
+ "grad_norm": 0.01793155074119568,
+ "learning_rate": 0.0006,
+ "loss": 4.314806938171387,
+ "step": 5091
+ },
+ {
+ "epoch": 70.72663755458515,
+ "grad_norm": 0.016286427155137062,
+ "learning_rate": 0.0006,
+ "loss": 4.2137556076049805,
+ "step": 5092
+ },
+ {
+ "epoch": 70.74061135371178,
+ "grad_norm": 0.014718376100063324,
+ "learning_rate": 0.0006,
+ "loss": 4.3073015213012695,
+ "step": 5093
+ },
+ {
+ "epoch": 70.75458515283843,
+ "grad_norm": 0.014675077050924301,
+ "learning_rate": 0.0006,
+ "loss": 4.160961151123047,
+ "step": 5094
+ },
+ {
+ "epoch": 70.76855895196506,
+ "grad_norm": 0.015550913289189339,
+ "learning_rate": 0.0006,
+ "loss": 4.322887420654297,
+ "step": 5095
+ },
+ {
+ "epoch": 70.78253275109171,
+ "grad_norm": 0.01453235000371933,
+ "learning_rate": 0.0006,
+ "loss": 4.2915873527526855,
+ "step": 5096
+ },
+ {
+ "epoch": 70.79650655021834,
+ "grad_norm": 0.01459164172410965,
+ "learning_rate": 0.0006,
+ "loss": 4.247288227081299,
+ "step": 5097
+ },
+ {
+ "epoch": 70.81048034934499,
+ "grad_norm": 0.015395080670714378,
+ "learning_rate": 0.0006,
+ "loss": 4.360955715179443,
+ "step": 5098
+ },
+ {
+ "epoch": 70.82445414847162,
+ "grad_norm": 0.01317683607339859,
+ "learning_rate": 0.0006,
+ "loss": 4.181969165802002,
+ "step": 5099
+ },
+ {
+ "epoch": 70.83842794759825,
+ "grad_norm": 0.013917932286858559,
+ "learning_rate": 0.0006,
+ "loss": 4.264626979827881,
+ "step": 5100
+ },
+ {
+ "epoch": 70.8524017467249,
+ "grad_norm": 0.015268695540726185,
+ "learning_rate": 0.0006,
+ "loss": 4.287857532501221,
+ "step": 5101
+ },
+ {
+ "epoch": 70.86637554585153,
+ "grad_norm": 0.015531329438090324,
+ "learning_rate": 0.0006,
+ "loss": 4.294188499450684,
+ "step": 5102
+ },
+ {
+ "epoch": 70.88034934497817,
+ "grad_norm": 0.015607129782438278,
+ "learning_rate": 0.0006,
+ "loss": 4.331040859222412,
+ "step": 5103
+ },
+ {
+ "epoch": 70.8943231441048,
+ "grad_norm": 0.015297149308025837,
+ "learning_rate": 0.0006,
+ "loss": 4.243650436401367,
+ "step": 5104
+ },
+ {
+ "epoch": 70.90829694323143,
+ "grad_norm": 0.014790025539696217,
+ "learning_rate": 0.0006,
+ "loss": 4.159859657287598,
+ "step": 5105
+ },
+ {
+ "epoch": 70.92227074235808,
+ "grad_norm": 0.013466274365782738,
+ "learning_rate": 0.0006,
+ "loss": 4.221992492675781,
+ "step": 5106
+ },
+ {
+ "epoch": 70.93624454148471,
+ "grad_norm": 0.01361898984760046,
+ "learning_rate": 0.0006,
+ "loss": 4.278751373291016,
+ "step": 5107
+ },
+ {
+ "epoch": 70.95021834061136,
+ "grad_norm": 0.01325751468539238,
+ "learning_rate": 0.0006,
+ "loss": 4.166001319885254,
+ "step": 5108
+ },
+ {
+ "epoch": 70.96419213973799,
+ "grad_norm": 0.012727087363600731,
+ "learning_rate": 0.0006,
+ "loss": 4.219997406005859,
+ "step": 5109
+ },
+ {
+ "epoch": 70.97816593886463,
+ "grad_norm": 0.013966867700219154,
+ "learning_rate": 0.0006,
+ "loss": 4.207491397857666,
+ "step": 5110
+ },
+ {
+ "epoch": 70.99213973799127,
+ "grad_norm": 0.015053587965667248,
+ "learning_rate": 0.0006,
+ "loss": 4.343245983123779,
+ "step": 5111
+ },
+ {
+ "epoch": 71.0,
+ "grad_norm": 0.017723368480801582,
+ "learning_rate": 0.0006,
+ "loss": 4.356122016906738,
+ "step": 5112
+ },
+ {
+ "epoch": 71.0,
+ "eval_loss": 4.598303318023682,
+ "eval_runtime": 56.7521,
+ "eval_samples_per_second": 43.029,
+ "eval_steps_per_second": 1.357,
+ "step": 5112
+ },
+ {
+ "epoch": 71.01397379912663,
+ "grad_norm": 0.018996279686689377,
+ "learning_rate": 0.0006,
+ "loss": 4.234772205352783,
+ "step": 5113
+ },
+ {
+ "epoch": 71.02794759825328,
+ "grad_norm": 0.02041555754840374,
+ "learning_rate": 0.0006,
+ "loss": 4.275816917419434,
+ "step": 5114
+ },
+ {
+ "epoch": 71.04192139737991,
+ "grad_norm": 0.021955285221338272,
+ "learning_rate": 0.0006,
+ "loss": 4.166038513183594,
+ "step": 5115
+ },
+ {
+ "epoch": 71.05589519650655,
+ "grad_norm": 0.021308740600943565,
+ "learning_rate": 0.0006,
+ "loss": 4.13063907623291,
+ "step": 5116
+ },
+ {
+ "epoch": 71.06986899563319,
+ "grad_norm": 0.01747054047882557,
+ "learning_rate": 0.0006,
+ "loss": 4.200522422790527,
+ "step": 5117
+ },
+ {
+ "epoch": 71.08384279475983,
+ "grad_norm": 0.018544532358646393,
+ "learning_rate": 0.0006,
+ "loss": 4.173915386199951,
+ "step": 5118
+ },
+ {
+ "epoch": 71.09781659388646,
+ "grad_norm": 0.0182361900806427,
+ "learning_rate": 0.0006,
+ "loss": 4.27392578125,
+ "step": 5119
+ },
+ {
+ "epoch": 71.1117903930131,
+ "grad_norm": 0.017062073573470116,
+ "learning_rate": 0.0006,
+ "loss": 4.219576358795166,
+ "step": 5120
+ },
+ {
+ "epoch": 71.12576419213974,
+ "grad_norm": 0.016289666295051575,
+ "learning_rate": 0.0006,
+ "loss": 4.133376598358154,
+ "step": 5121
+ },
+ {
+ "epoch": 71.13973799126637,
+ "grad_norm": 0.014864642173051834,
+ "learning_rate": 0.0006,
+ "loss": 4.168276786804199,
+ "step": 5122
+ },
+ {
+ "epoch": 71.15371179039302,
+ "grad_norm": 0.015309540554881096,
+ "learning_rate": 0.0006,
+ "loss": 4.175103187561035,
+ "step": 5123
+ },
+ {
+ "epoch": 71.16768558951965,
+ "grad_norm": 0.01556059904396534,
+ "learning_rate": 0.0006,
+ "loss": 4.289899826049805,
+ "step": 5124
+ },
+ {
+ "epoch": 71.18165938864628,
+ "grad_norm": 0.01479957066476345,
+ "learning_rate": 0.0006,
+ "loss": 4.341495037078857,
+ "step": 5125
+ },
+ {
+ "epoch": 71.19563318777293,
+ "grad_norm": 0.014762179926037788,
+ "learning_rate": 0.0006,
+ "loss": 4.252333641052246,
+ "step": 5126
+ },
+ {
+ "epoch": 71.20960698689956,
+ "grad_norm": 0.014897473156452179,
+ "learning_rate": 0.0006,
+ "loss": 4.2244181632995605,
+ "step": 5127
+ },
+ {
+ "epoch": 71.2235807860262,
+ "grad_norm": 0.012990163639187813,
+ "learning_rate": 0.0006,
+ "loss": 4.237288475036621,
+ "step": 5128
+ },
+ {
+ "epoch": 71.23755458515284,
+ "grad_norm": 0.012858416885137558,
+ "learning_rate": 0.0006,
+ "loss": 4.265738487243652,
+ "step": 5129
+ },
+ {
+ "epoch": 71.25152838427948,
+ "grad_norm": 0.013454788364470005,
+ "learning_rate": 0.0006,
+ "loss": 4.289131164550781,
+ "step": 5130
+ },
+ {
+ "epoch": 71.26550218340611,
+ "grad_norm": 0.013717238791286945,
+ "learning_rate": 0.0006,
+ "loss": 4.2844014167785645,
+ "step": 5131
+ },
+ {
+ "epoch": 71.27947598253274,
+ "grad_norm": 0.0135411461815238,
+ "learning_rate": 0.0006,
+ "loss": 4.139082908630371,
+ "step": 5132
+ },
+ {
+ "epoch": 71.29344978165939,
+ "grad_norm": 0.01406062301248312,
+ "learning_rate": 0.0006,
+ "loss": 4.226719856262207,
+ "step": 5133
+ },
+ {
+ "epoch": 71.30742358078602,
+ "grad_norm": 0.014027190394699574,
+ "learning_rate": 0.0006,
+ "loss": 4.191699981689453,
+ "step": 5134
+ },
+ {
+ "epoch": 71.32139737991267,
+ "grad_norm": 0.013809907250106335,
+ "learning_rate": 0.0006,
+ "loss": 4.242542266845703,
+ "step": 5135
+ },
+ {
+ "epoch": 71.3353711790393,
+ "grad_norm": 0.013507804833352566,
+ "learning_rate": 0.0006,
+ "loss": 4.068206310272217,
+ "step": 5136
+ },
+ {
+ "epoch": 71.34934497816595,
+ "grad_norm": 0.01432533748447895,
+ "learning_rate": 0.0006,
+ "loss": 4.378222942352295,
+ "step": 5137
+ },
+ {
+ "epoch": 71.36331877729258,
+ "grad_norm": 0.01753002591431141,
+ "learning_rate": 0.0006,
+ "loss": 4.174706935882568,
+ "step": 5138
+ },
+ {
+ "epoch": 71.37729257641921,
+ "grad_norm": 0.020048344507813454,
+ "learning_rate": 0.0006,
+ "loss": 4.19920015335083,
+ "step": 5139
+ },
+ {
+ "epoch": 71.39126637554585,
+ "grad_norm": 0.018551401793956757,
+ "learning_rate": 0.0006,
+ "loss": 4.175182342529297,
+ "step": 5140
+ },
+ {
+ "epoch": 71.40524017467249,
+ "grad_norm": 0.01598959229886532,
+ "learning_rate": 0.0006,
+ "loss": 4.2305169105529785,
+ "step": 5141
+ },
+ {
+ "epoch": 71.41921397379913,
+ "grad_norm": 0.016029560938477516,
+ "learning_rate": 0.0006,
+ "loss": 4.230090618133545,
+ "step": 5142
+ },
+ {
+ "epoch": 71.43318777292576,
+ "grad_norm": 0.01595657505095005,
+ "learning_rate": 0.0006,
+ "loss": 4.138402938842773,
+ "step": 5143
+ },
+ {
+ "epoch": 71.44716157205241,
+ "grad_norm": 0.01623230241239071,
+ "learning_rate": 0.0006,
+ "loss": 4.196106433868408,
+ "step": 5144
+ },
+ {
+ "epoch": 71.46113537117904,
+ "grad_norm": 0.014057597145438194,
+ "learning_rate": 0.0006,
+ "loss": 4.2599897384643555,
+ "step": 5145
+ },
+ {
+ "epoch": 71.47510917030567,
+ "grad_norm": 0.017161158844828606,
+ "learning_rate": 0.0006,
+ "loss": 4.183909893035889,
+ "step": 5146
+ },
+ {
+ "epoch": 71.48908296943232,
+ "grad_norm": 0.017038581892848015,
+ "learning_rate": 0.0006,
+ "loss": 4.373933792114258,
+ "step": 5147
+ },
+ {
+ "epoch": 71.50305676855895,
+ "grad_norm": 0.015851367264986038,
+ "learning_rate": 0.0006,
+ "loss": 4.2620086669921875,
+ "step": 5148
+ },
+ {
+ "epoch": 71.5170305676856,
+ "grad_norm": 0.014598303474485874,
+ "learning_rate": 0.0006,
+ "loss": 4.197151184082031,
+ "step": 5149
+ },
+ {
+ "epoch": 71.53100436681223,
+ "grad_norm": 0.01339802611619234,
+ "learning_rate": 0.0006,
+ "loss": 4.110706329345703,
+ "step": 5150
+ },
+ {
+ "epoch": 71.54497816593886,
+ "grad_norm": 0.015567775815725327,
+ "learning_rate": 0.0006,
+ "loss": 4.1210246086120605,
+ "step": 5151
+ },
+ {
+ "epoch": 71.5589519650655,
+ "grad_norm": 0.015616007149219513,
+ "learning_rate": 0.0006,
+ "loss": 4.129796028137207,
+ "step": 5152
+ },
+ {
+ "epoch": 71.57292576419214,
+ "grad_norm": 0.014198964461684227,
+ "learning_rate": 0.0006,
+ "loss": 4.193232536315918,
+ "step": 5153
+ },
+ {
+ "epoch": 71.58689956331878,
+ "grad_norm": 0.015226383693516254,
+ "learning_rate": 0.0006,
+ "loss": 4.223779678344727,
+ "step": 5154
+ },
+ {
+ "epoch": 71.60087336244541,
+ "grad_norm": 0.015610109083354473,
+ "learning_rate": 0.0006,
+ "loss": 4.223179817199707,
+ "step": 5155
+ },
+ {
+ "epoch": 71.61484716157206,
+ "grad_norm": 0.016034303233027458,
+ "learning_rate": 0.0006,
+ "loss": 4.196593284606934,
+ "step": 5156
+ },
+ {
+ "epoch": 71.62882096069869,
+ "grad_norm": 0.015648581087589264,
+ "learning_rate": 0.0006,
+ "loss": 4.341497421264648,
+ "step": 5157
+ },
+ {
+ "epoch": 71.64279475982532,
+ "grad_norm": 0.014117540791630745,
+ "learning_rate": 0.0006,
+ "loss": 4.194307327270508,
+ "step": 5158
+ },
+ {
+ "epoch": 71.65676855895197,
+ "grad_norm": 0.015612686984241009,
+ "learning_rate": 0.0006,
+ "loss": 4.290152549743652,
+ "step": 5159
+ },
+ {
+ "epoch": 71.6707423580786,
+ "grad_norm": 0.01722259446978569,
+ "learning_rate": 0.0006,
+ "loss": 4.258018493652344,
+ "step": 5160
+ },
+ {
+ "epoch": 71.68471615720524,
+ "grad_norm": 0.015923351049423218,
+ "learning_rate": 0.0006,
+ "loss": 4.21684455871582,
+ "step": 5161
+ },
+ {
+ "epoch": 71.69868995633188,
+ "grad_norm": 0.013663450255990028,
+ "learning_rate": 0.0006,
+ "loss": 4.175601959228516,
+ "step": 5162
+ },
+ {
+ "epoch": 71.71266375545852,
+ "grad_norm": 0.017075596377253532,
+ "learning_rate": 0.0006,
+ "loss": 4.205138683319092,
+ "step": 5163
+ },
+ {
+ "epoch": 71.72663755458515,
+ "grad_norm": 0.01723487675189972,
+ "learning_rate": 0.0006,
+ "loss": 4.08839225769043,
+ "step": 5164
+ },
+ {
+ "epoch": 71.74061135371178,
+ "grad_norm": 0.017355505377054214,
+ "learning_rate": 0.0006,
+ "loss": 4.165130615234375,
+ "step": 5165
+ },
+ {
+ "epoch": 71.75458515283843,
+ "grad_norm": 0.01868261583149433,
+ "learning_rate": 0.0006,
+ "loss": 4.16977071762085,
+ "step": 5166
+ },
+ {
+ "epoch": 71.76855895196506,
+ "grad_norm": 0.01897321827709675,
+ "learning_rate": 0.0006,
+ "loss": 4.173497200012207,
+ "step": 5167
+ },
+ {
+ "epoch": 71.78253275109171,
+ "grad_norm": 0.02045825682580471,
+ "learning_rate": 0.0006,
+ "loss": 4.176192283630371,
+ "step": 5168
+ },
+ {
+ "epoch": 71.79650655021834,
+ "grad_norm": 0.022895624861121178,
+ "learning_rate": 0.0006,
+ "loss": 4.163009166717529,
+ "step": 5169
+ },
+ {
+ "epoch": 71.81048034934499,
+ "grad_norm": 0.02062244527041912,
+ "learning_rate": 0.0006,
+ "loss": 4.156868934631348,
+ "step": 5170
+ },
+ {
+ "epoch": 71.82445414847162,
+ "grad_norm": 0.0176058579236269,
+ "learning_rate": 0.0006,
+ "loss": 4.154374122619629,
+ "step": 5171
+ },
+ {
+ "epoch": 71.83842794759825,
+ "grad_norm": 0.018049508333206177,
+ "learning_rate": 0.0006,
+ "loss": 4.296828269958496,
+ "step": 5172
+ },
+ {
+ "epoch": 71.8524017467249,
+ "grad_norm": 0.019466130062937737,
+ "learning_rate": 0.0006,
+ "loss": 4.19450569152832,
+ "step": 5173
+ },
+ {
+ "epoch": 71.86637554585153,
+ "grad_norm": 0.01838165521621704,
+ "learning_rate": 0.0006,
+ "loss": 4.11812686920166,
+ "step": 5174
+ },
+ {
+ "epoch": 71.88034934497817,
+ "grad_norm": 0.01608051173388958,
+ "learning_rate": 0.0006,
+ "loss": 4.240464210510254,
+ "step": 5175
+ },
+ {
+ "epoch": 71.8943231441048,
+ "grad_norm": 0.015807967633008957,
+ "learning_rate": 0.0006,
+ "loss": 4.275581359863281,
+ "step": 5176
+ },
+ {
+ "epoch": 71.90829694323143,
+ "grad_norm": 0.017366575077176094,
+ "learning_rate": 0.0006,
+ "loss": 4.064750671386719,
+ "step": 5177
+ },
+ {
+ "epoch": 71.92227074235808,
+ "grad_norm": 0.016241202130913734,
+ "learning_rate": 0.0006,
+ "loss": 4.203697681427002,
+ "step": 5178
+ },
+ {
+ "epoch": 71.93624454148471,
+ "grad_norm": 0.019105732440948486,
+ "learning_rate": 0.0006,
+ "loss": 4.140021800994873,
+ "step": 5179
+ },
+ {
+ "epoch": 71.95021834061136,
+ "grad_norm": 0.018492674455046654,
+ "learning_rate": 0.0006,
+ "loss": 4.2534074783325195,
+ "step": 5180
+ },
+ {
+ "epoch": 71.96419213973799,
+ "grad_norm": 0.019301170483231544,
+ "learning_rate": 0.0006,
+ "loss": 4.381392478942871,
+ "step": 5181
+ },
+ {
+ "epoch": 71.97816593886463,
+ "grad_norm": 0.019571911543607712,
+ "learning_rate": 0.0006,
+ "loss": 4.131056785583496,
+ "step": 5182
+ },
+ {
+ "epoch": 71.99213973799127,
+ "grad_norm": 0.01859748177230358,
+ "learning_rate": 0.0006,
+ "loss": 4.278930187225342,
+ "step": 5183
+ },
+ {
+ "epoch": 72.0,
+ "grad_norm": 0.02017994411289692,
+ "learning_rate": 0.0006,
+ "loss": 4.116464614868164,
+ "step": 5184
+ },
+ {
+ "epoch": 72.0,
+ "eval_loss": 4.60784912109375,
+ "eval_runtime": 56.3242,
+ "eval_samples_per_second": 43.356,
+ "eval_steps_per_second": 1.367,
+ "step": 5184
+ },
+ {
+ "epoch": 72.01397379912663,
+ "grad_norm": 0.030090296640992165,
+ "learning_rate": 0.0006,
+ "loss": 4.218588352203369,
+ "step": 5185
+ },
+ {
+ "epoch": 72.02794759825328,
+ "grad_norm": 1.330443024635315,
+ "learning_rate": 0.0006,
+ "loss": 4.417870044708252,
+ "step": 5186
+ },
+ {
+ "epoch": 72.04192139737991,
+ "grad_norm": 0.6991518139839172,
+ "learning_rate": 0.0006,
+ "loss": 8.414234161376953,
+ "step": 5187
+ },
+ {
+ "epoch": 72.05589519650655,
+ "grad_norm": 0.2631438374519348,
+ "learning_rate": 0.0006,
+ "loss": 7.41729736328125,
+ "step": 5188
+ },
+ {
+ "epoch": 72.06986899563319,
+ "grad_norm": 0.12391754984855652,
+ "learning_rate": 0.0006,
+ "loss": 7.209164619445801,
+ "step": 5189
+ },
+ {
+ "epoch": 72.08384279475983,
+ "grad_norm": 0.13519170880317688,
+ "learning_rate": 0.0006,
+ "loss": 6.943539619445801,
+ "step": 5190
+ },
+ {
+ "epoch": 72.09781659388646,
+ "grad_norm": 1.258439302444458,
+ "learning_rate": 0.0006,
+ "loss": 10.267714500427246,
+ "step": 5191
+ },
+ {
+ "epoch": 72.1117903930131,
+ "grad_norm": 1.0296908617019653,
+ "learning_rate": 0.0006,
+ "loss": 9.51768970489502,
+ "step": 5192
+ },
+ {
+ "epoch": 72.12576419213974,
+ "grad_norm": 0.12273016571998596,
+ "learning_rate": 0.0006,
+ "loss": 7.634477615356445,
+ "step": 5193
+ },
+ {
+ "epoch": 72.13973799126637,
+ "grad_norm": 0.06170198693871498,
+ "learning_rate": 0.0006,
+ "loss": 7.592115879058838,
+ "step": 5194
+ },
+ {
+ "epoch": 72.15371179039302,
+ "grad_norm": 0.06295108795166016,
+ "learning_rate": 0.0006,
+ "loss": 7.576314449310303,
+ "step": 5195
+ },
+ {
+ "epoch": 72.16768558951965,
+ "grad_norm": 0.0666993111371994,
+ "learning_rate": 0.0006,
+ "loss": 7.53862190246582,
+ "step": 5196
+ },
+ {
+ "epoch": 72.18165938864628,
+ "grad_norm": 0.052215222269296646,
+ "learning_rate": 0.0006,
+ "loss": 7.502930641174316,
+ "step": 5197
+ },
+ {
+ "epoch": 72.19563318777293,
+ "grad_norm": 0.04978961497545242,
+ "learning_rate": 0.0006,
+ "loss": 7.471630096435547,
+ "step": 5198
+ },
+ {
+ "epoch": 72.20960698689956,
+ "grad_norm": 0.07275046408176422,
+ "learning_rate": 0.0006,
+ "loss": 7.496924877166748,
+ "step": 5199
+ },
+ {
+ "epoch": 72.2235807860262,
+ "grad_norm": 0.04491819068789482,
+ "learning_rate": 0.0006,
+ "loss": 7.424494743347168,
+ "step": 5200
+ },
+ {
+ "epoch": 72.23755458515284,
+ "grad_norm": 0.043718695640563965,
+ "learning_rate": 0.0006,
+ "loss": 7.441090106964111,
+ "step": 5201
+ },
+ {
+ "epoch": 72.25152838427948,
+ "grad_norm": 0.04483497887849808,
+ "learning_rate": 0.0006,
+ "loss": 7.3753862380981445,
+ "step": 5202
+ },
+ {
+ "epoch": 72.26550218340611,
+ "grad_norm": 0.03463944420218468,
+ "learning_rate": 0.0006,
+ "loss": 7.375782012939453,
+ "step": 5203
+ },
+ {
+ "epoch": 72.27947598253274,
+ "grad_norm": 0.0397045724093914,
+ "learning_rate": 0.0006,
+ "loss": 7.358898639678955,
+ "step": 5204
+ },
+ {
+ "epoch": 72.29344978165939,
+ "grad_norm": 0.029339438304305077,
+ "learning_rate": 0.0006,
+ "loss": 7.3774871826171875,
+ "step": 5205
+ },
+ {
+ "epoch": 72.30742358078602,
+ "grad_norm": 0.025469699874520302,
+ "learning_rate": 0.0006,
+ "loss": 7.324203014373779,
+ "step": 5206
+ },
+ {
+ "epoch": 72.32139737991267,
+ "grad_norm": 0.030497785657644272,
+ "learning_rate": 0.0006,
+ "loss": 7.329351902008057,
+ "step": 5207
+ },
+ {
+ "epoch": 72.3353711790393,
+ "grad_norm": 0.03382078558206558,
+ "learning_rate": 0.0006,
+ "loss": 7.289940357208252,
+ "step": 5208
+ },
+ {
+ "epoch": 72.34934497816595,
+ "grad_norm": 0.019869405776262283,
+ "learning_rate": 0.0006,
+ "loss": 7.272582054138184,
+ "step": 5209
+ },
+ {
+ "epoch": 72.36331877729258,
+ "grad_norm": 0.02385753206908703,
+ "learning_rate": 0.0006,
+ "loss": 7.2654266357421875,
+ "step": 5210
+ },
+ {
+ "epoch": 72.37729257641921,
+ "grad_norm": 0.02794312685728073,
+ "learning_rate": 0.0006,
+ "loss": 7.250753402709961,
+ "step": 5211
+ },
+ {
+ "epoch": 72.39126637554585,
+ "grad_norm": 0.02445952221751213,
+ "learning_rate": 0.0006,
+ "loss": 7.293970108032227,
+ "step": 5212
+ },
+ {
+ "epoch": 72.40524017467249,
+ "grad_norm": 0.019660986959934235,
+ "learning_rate": 0.0006,
+ "loss": 7.1948771476745605,
+ "step": 5213
+ },
+ {
+ "epoch": 72.41921397379913,
+ "grad_norm": 0.024110861122608185,
+ "learning_rate": 0.0006,
+ "loss": 7.224499225616455,
+ "step": 5214
+ },
+ {
+ "epoch": 72.43318777292576,
+ "grad_norm": 0.021630609408020973,
+ "learning_rate": 0.0006,
+ "loss": 7.1736884117126465,
+ "step": 5215
+ },
+ {
+ "epoch": 72.44716157205241,
+ "grad_norm": 0.02185880020260811,
+ "learning_rate": 0.0006,
+ "loss": 7.167229652404785,
+ "step": 5216
+ },
+ {
+ "epoch": 72.46113537117904,
+ "grad_norm": 0.018913568928837776,
+ "learning_rate": 0.0006,
+ "loss": 7.196551322937012,
+ "step": 5217
+ },
+ {
+ "epoch": 72.47510917030567,
+ "grad_norm": 0.01802486553788185,
+ "learning_rate": 0.0006,
+ "loss": 7.144286632537842,
+ "step": 5218
+ },
+ {
+ "epoch": 72.48908296943232,
+ "grad_norm": 0.020386485382914543,
+ "learning_rate": 0.0006,
+ "loss": 7.196260929107666,
+ "step": 5219
+ },
+ {
+ "epoch": 72.50305676855895,
+ "grad_norm": 0.020077696070075035,
+ "learning_rate": 0.0006,
+ "loss": 7.138718605041504,
+ "step": 5220
+ },
+ {
+ "epoch": 72.5170305676856,
+ "grad_norm": 0.017224183306097984,
+ "learning_rate": 0.0006,
+ "loss": 7.119478702545166,
+ "step": 5221
+ },
+ {
+ "epoch": 72.53100436681223,
+ "grad_norm": 0.020847434177994728,
+ "learning_rate": 0.0006,
+ "loss": 7.068922996520996,
+ "step": 5222
+ },
+ {
+ "epoch": 72.54497816593886,
+ "grad_norm": 0.021764913573861122,
+ "learning_rate": 0.0006,
+ "loss": 7.071352958679199,
+ "step": 5223
+ },
+ {
+ "epoch": 72.5589519650655,
+ "grad_norm": 0.02697215974330902,
+ "learning_rate": 0.0006,
+ "loss": 7.033344268798828,
+ "step": 5224
+ },
+ {
+ "epoch": 72.57292576419214,
+ "grad_norm": 0.042392998933792114,
+ "learning_rate": 0.0006,
+ "loss": 7.0589919090271,
+ "step": 5225
+ },
+ {
+ "epoch": 72.58689956331878,
+ "grad_norm": 0.06616316735744476,
+ "learning_rate": 0.0006,
+ "loss": 7.060244083404541,
+ "step": 5226
+ },
+ {
+ "epoch": 72.60087336244541,
+ "grad_norm": 0.0828910544514656,
+ "learning_rate": 0.0006,
+ "loss": 7.057173728942871,
+ "step": 5227
+ },
+ {
+ "epoch": 72.61484716157206,
+ "grad_norm": 0.046767015010118484,
+ "learning_rate": 0.0006,
+ "loss": 7.010563850402832,
+ "step": 5228
+ },
+ {
+ "epoch": 72.62882096069869,
+ "grad_norm": 0.03831760957837105,
+ "learning_rate": 0.0006,
+ "loss": 6.9028825759887695,
+ "step": 5229
+ },
+ {
+ "epoch": 72.64279475982532,
+ "grad_norm": 0.05109567195177078,
+ "learning_rate": 0.0006,
+ "loss": 6.923257827758789,
+ "step": 5230
+ },
+ {
+ "epoch": 72.65676855895197,
+ "grad_norm": 0.02897934429347515,
+ "learning_rate": 0.0006,
+ "loss": 6.823014259338379,
+ "step": 5231
+ },
+ {
+ "epoch": 72.6707423580786,
+ "grad_norm": 0.03769192099571228,
+ "learning_rate": 0.0006,
+ "loss": 6.87416410446167,
+ "step": 5232
+ },
+ {
+ "epoch": 72.68471615720524,
+ "grad_norm": 0.03677637130022049,
+ "learning_rate": 0.0006,
+ "loss": 6.7649030685424805,
+ "step": 5233
+ },
+ {
+ "epoch": 72.69868995633188,
+ "grad_norm": 0.0371948778629303,
+ "learning_rate": 0.0006,
+ "loss": 6.749821662902832,
+ "step": 5234
+ },
+ {
+ "epoch": 72.71266375545852,
+ "grad_norm": 0.03057795763015747,
+ "learning_rate": 0.0006,
+ "loss": 6.748945236206055,
+ "step": 5235
+ },
+ {
+ "epoch": 72.72663755458515,
+ "grad_norm": 0.028109321370720863,
+ "learning_rate": 0.0006,
+ "loss": 6.663548946380615,
+ "step": 5236
+ },
+ {
+ "epoch": 72.74061135371178,
+ "grad_norm": 0.02442420832812786,
+ "learning_rate": 0.0006,
+ "loss": 6.636251449584961,
+ "step": 5237
+ },
+ {
+ "epoch": 72.75458515283843,
+ "grad_norm": 0.031841423362493515,
+ "learning_rate": 0.0006,
+ "loss": 6.606883525848389,
+ "step": 5238
+ },
+ {
+ "epoch": 72.76855895196506,
+ "grad_norm": 0.028100458905100822,
+ "learning_rate": 0.0006,
+ "loss": 6.603745937347412,
+ "step": 5239
+ },
+ {
+ "epoch": 72.78253275109171,
+ "grad_norm": 0.03222901374101639,
+ "learning_rate": 0.0006,
+ "loss": 6.463825225830078,
+ "step": 5240
+ },
+ {
+ "epoch": 72.79650655021834,
+ "grad_norm": 0.04766521975398064,
+ "learning_rate": 0.0006,
+ "loss": 6.43734073638916,
+ "step": 5241
+ },
+ {
+ "epoch": 72.81048034934499,
+ "grad_norm": 0.09058783948421478,
+ "learning_rate": 0.0006,
+ "loss": 6.448520183563232,
+ "step": 5242
+ },
+ {
+ "epoch": 72.82445414847162,
+ "grad_norm": 0.08995158225297928,
+ "learning_rate": 0.0006,
+ "loss": 6.522093296051025,
+ "step": 5243
+ },
+ {
+ "epoch": 72.83842794759825,
+ "grad_norm": 0.05491040274500847,
+ "learning_rate": 0.0006,
+ "loss": 6.37354040145874,
+ "step": 5244
+ },
+ {
+ "epoch": 72.8524017467249,
+ "grad_norm": 0.05198514088988304,
+ "learning_rate": 0.0006,
+ "loss": 6.362032413482666,
+ "step": 5245
+ },
+ {
+ "epoch": 72.86637554585153,
+ "grad_norm": 0.05464765429496765,
+ "learning_rate": 0.0006,
+ "loss": 6.326773643493652,
+ "step": 5246
+ },
+ {
+ "epoch": 72.88034934497817,
+ "grad_norm": 0.07337361574172974,
+ "learning_rate": 0.0006,
+ "loss": 6.297532081604004,
+ "step": 5247
+ },
+ {
+ "epoch": 72.8943231441048,
+ "grad_norm": 0.10061230510473251,
+ "learning_rate": 0.0006,
+ "loss": 6.359557628631592,
+ "step": 5248
+ },
+ {
+ "epoch": 72.90829694323143,
+ "grad_norm": 0.10005959123373032,
+ "learning_rate": 0.0006,
+ "loss": 6.422270774841309,
+ "step": 5249
+ },
+ {
+ "epoch": 72.92227074235808,
+ "grad_norm": 0.08989301323890686,
+ "learning_rate": 0.0006,
+ "loss": 6.289371490478516,
+ "step": 5250
+ },
+ {
+ "epoch": 72.93624454148471,
+ "grad_norm": 0.06371799111366272,
+ "learning_rate": 0.0006,
+ "loss": 6.199618339538574,
+ "step": 5251
+ },
+ {
+ "epoch": 72.95021834061136,
+ "grad_norm": 0.05932944267988205,
+ "learning_rate": 0.0006,
+ "loss": 6.111152648925781,
+ "step": 5252
+ },
+ {
+ "epoch": 72.96419213973799,
+ "grad_norm": 0.04735502973198891,
+ "learning_rate": 0.0006,
+ "loss": 6.026026248931885,
+ "step": 5253
+ },
+ {
+ "epoch": 72.97816593886463,
+ "grad_norm": 0.04270688816905022,
+ "learning_rate": 0.0006,
+ "loss": 6.01606559753418,
+ "step": 5254
+ },
+ {
+ "epoch": 72.99213973799127,
+ "grad_norm": 0.03774143010377884,
+ "learning_rate": 0.0006,
+ "loss": 5.8935441970825195,
+ "step": 5255
+ },
+ {
+ "epoch": 73.0,
+ "grad_norm": 0.04313631355762482,
+ "learning_rate": 0.0006,
+ "loss": 5.842061996459961,
+ "step": 5256
+ },
+ {
+ "epoch": 73.0,
+ "eval_loss": 5.9985785484313965,
+ "eval_runtime": 56.3891,
+ "eval_samples_per_second": 43.306,
+ "eval_steps_per_second": 1.366,
+ "step": 5256
+ },
+ {
+ "epoch": 73.01397379912663,
+ "grad_norm": 0.045038655400276184,
+ "learning_rate": 0.0006,
+ "loss": 5.8209662437438965,
+ "step": 5257
+ },
+ {
+ "epoch": 73.02794759825328,
+ "grad_norm": 0.03706910088658333,
+ "learning_rate": 0.0006,
+ "loss": 5.806189060211182,
+ "step": 5258
+ },
+ {
+ "epoch": 73.04192139737991,
+ "grad_norm": 0.04051685333251953,
+ "learning_rate": 0.0006,
+ "loss": 5.706869125366211,
+ "step": 5259
+ },
+ {
+ "epoch": 73.05589519650655,
+ "grad_norm": 0.04849153012037277,
+ "learning_rate": 0.0006,
+ "loss": 5.696385383605957,
+ "step": 5260
+ },
+ {
+ "epoch": 73.06986899563319,
+ "grad_norm": 0.06124403700232506,
+ "learning_rate": 0.0006,
+ "loss": 5.618223667144775,
+ "step": 5261
+ },
+ {
+ "epoch": 73.08384279475983,
+ "grad_norm": 0.07375937700271606,
+ "learning_rate": 0.0006,
+ "loss": 5.631045341491699,
+ "step": 5262
+ },
+ {
+ "epoch": 73.09781659388646,
+ "grad_norm": 0.07413844019174576,
+ "learning_rate": 0.0006,
+ "loss": 5.6532182693481445,
+ "step": 5263
+ },
+ {
+ "epoch": 73.1117903930131,
+ "grad_norm": 0.0633770152926445,
+ "learning_rate": 0.0006,
+ "loss": 5.57209587097168,
+ "step": 5264
+ },
+ {
+ "epoch": 73.12576419213974,
+ "grad_norm": 0.05436766892671585,
+ "learning_rate": 0.0006,
+ "loss": 5.53944730758667,
+ "step": 5265
+ },
+ {
+ "epoch": 73.13973799126637,
+ "grad_norm": 0.04639481008052826,
+ "learning_rate": 0.0006,
+ "loss": 5.502948760986328,
+ "step": 5266
+ },
+ {
+ "epoch": 73.15371179039302,
+ "grad_norm": 0.07556558400392532,
+ "learning_rate": 0.0006,
+ "loss": 5.429069995880127,
+ "step": 5267
+ },
+ {
+ "epoch": 73.16768558951965,
+ "grad_norm": 0.1127929836511612,
+ "learning_rate": 0.0006,
+ "loss": 5.545061111450195,
+ "step": 5268
+ },
+ {
+ "epoch": 73.18165938864628,
+ "grad_norm": 0.07620745897293091,
+ "learning_rate": 0.0006,
+ "loss": 5.507102966308594,
+ "step": 5269
+ },
+ {
+ "epoch": 73.19563318777293,
+ "grad_norm": 0.05076538026332855,
+ "learning_rate": 0.0006,
+ "loss": 5.329682350158691,
+ "step": 5270
+ },
+ {
+ "epoch": 73.20960698689956,
+ "grad_norm": 0.05407547950744629,
+ "learning_rate": 0.0006,
+ "loss": 5.3048553466796875,
+ "step": 5271
+ },
+ {
+ "epoch": 73.2235807860262,
+ "grad_norm": 0.059095583856105804,
+ "learning_rate": 0.0006,
+ "loss": 5.292568683624268,
+ "step": 5272
+ },
+ {
+ "epoch": 73.23755458515284,
+ "grad_norm": 0.06193305924534798,
+ "learning_rate": 0.0006,
+ "loss": 5.1870927810668945,
+ "step": 5273
+ },
+ {
+ "epoch": 73.25152838427948,
+ "grad_norm": 0.062261562794446945,
+ "learning_rate": 0.0006,
+ "loss": 5.256819248199463,
+ "step": 5274
+ },
+ {
+ "epoch": 73.26550218340611,
+ "grad_norm": 0.0732516348361969,
+ "learning_rate": 0.0006,
+ "loss": 5.256464004516602,
+ "step": 5275
+ },
+ {
+ "epoch": 73.27947598253274,
+ "grad_norm": 0.12177979946136475,
+ "learning_rate": 0.0006,
+ "loss": 5.284239768981934,
+ "step": 5276
+ },
+ {
+ "epoch": 73.29344978165939,
+ "grad_norm": 0.10459653288125992,
+ "learning_rate": 0.0006,
+ "loss": 5.269713401794434,
+ "step": 5277
+ },
+ {
+ "epoch": 73.30742358078602,
+ "grad_norm": 0.07854694873094559,
+ "learning_rate": 0.0006,
+ "loss": 5.3245391845703125,
+ "step": 5278
+ },
+ {
+ "epoch": 73.32139737991267,
+ "grad_norm": 0.06030094623565674,
+ "learning_rate": 0.0006,
+ "loss": 5.033702850341797,
+ "step": 5279
+ },
+ {
+ "epoch": 73.3353711790393,
+ "grad_norm": 0.05168461427092552,
+ "learning_rate": 0.0006,
+ "loss": 5.140530109405518,
+ "step": 5280
+ },
+ {
+ "epoch": 73.34934497816595,
+ "grad_norm": 0.041064050048589706,
+ "learning_rate": 0.0006,
+ "loss": 5.1085896492004395,
+ "step": 5281
+ },
+ {
+ "epoch": 73.36331877729258,
+ "grad_norm": 0.03788219019770622,
+ "learning_rate": 0.0006,
+ "loss": 5.045009613037109,
+ "step": 5282
+ },
+ {
+ "epoch": 73.37729257641921,
+ "grad_norm": 0.03851071000099182,
+ "learning_rate": 0.0006,
+ "loss": 5.055008888244629,
+ "step": 5283
+ },
+ {
+ "epoch": 73.39126637554585,
+ "grad_norm": 0.03448539599776268,
+ "learning_rate": 0.0006,
+ "loss": 5.10657262802124,
+ "step": 5284
+ },
+ {
+ "epoch": 73.40524017467249,
+ "grad_norm": 0.026618091389536858,
+ "learning_rate": 0.0006,
+ "loss": 5.0342912673950195,
+ "step": 5285
+ },
+ {
+ "epoch": 73.41921397379913,
+ "grad_norm": 0.03060489147901535,
+ "learning_rate": 0.0006,
+ "loss": 5.022700786590576,
+ "step": 5286
+ },
+ {
+ "epoch": 73.43318777292576,
+ "grad_norm": 0.03362000361084938,
+ "learning_rate": 0.0006,
+ "loss": 4.891584396362305,
+ "step": 5287
+ },
+ {
+ "epoch": 73.44716157205241,
+ "grad_norm": 0.03887505829334259,
+ "learning_rate": 0.0006,
+ "loss": 4.772258758544922,
+ "step": 5288
+ },
+ {
+ "epoch": 73.46113537117904,
+ "grad_norm": 0.07157225161790848,
+ "learning_rate": 0.0006,
+ "loss": 4.997142791748047,
+ "step": 5289
+ },
+ {
+ "epoch": 73.47510917030567,
+ "grad_norm": 0.10180624574422836,
+ "learning_rate": 0.0006,
+ "loss": 4.908722877502441,
+ "step": 5290
+ },
+ {
+ "epoch": 73.48908296943232,
+ "grad_norm": 0.06529122591018677,
+ "learning_rate": 0.0006,
+ "loss": 4.98001766204834,
+ "step": 5291
+ },
+ {
+ "epoch": 73.50305676855895,
+ "grad_norm": 0.055171865969896317,
+ "learning_rate": 0.0006,
+ "loss": 4.877438068389893,
+ "step": 5292
+ },
+ {
+ "epoch": 73.5170305676856,
+ "grad_norm": 0.0318770632147789,
+ "learning_rate": 0.0006,
+ "loss": 4.836934566497803,
+ "step": 5293
+ },
+ {
+ "epoch": 73.53100436681223,
+ "grad_norm": 0.042072564363479614,
+ "learning_rate": 0.0006,
+ "loss": 4.919489860534668,
+ "step": 5294
+ },
+ {
+ "epoch": 73.54497816593886,
+ "grad_norm": 0.029197754338383675,
+ "learning_rate": 0.0006,
+ "loss": 4.805579662322998,
+ "step": 5295
+ },
+ {
+ "epoch": 73.5589519650655,
+ "grad_norm": 0.03249497339129448,
+ "learning_rate": 0.0006,
+ "loss": 4.704806327819824,
+ "step": 5296
+ },
+ {
+ "epoch": 73.57292576419214,
+ "grad_norm": 0.025050945580005646,
+ "learning_rate": 0.0006,
+ "loss": 4.936512470245361,
+ "step": 5297
+ },
+ {
+ "epoch": 73.58689956331878,
+ "grad_norm": 0.029073381796479225,
+ "learning_rate": 0.0006,
+ "loss": 4.757592678070068,
+ "step": 5298
+ },
+ {
+ "epoch": 73.60087336244541,
+ "grad_norm": 0.03465264290571213,
+ "learning_rate": 0.0006,
+ "loss": 4.729528427124023,
+ "step": 5299
+ },
+ {
+ "epoch": 73.61484716157206,
+ "grad_norm": 0.054795749485492706,
+ "learning_rate": 0.0006,
+ "loss": 4.77252197265625,
+ "step": 5300
+ },
+ {
+ "epoch": 73.62882096069869,
+ "grad_norm": 0.18129196763038635,
+ "learning_rate": 0.0006,
+ "loss": 4.901700019836426,
+ "step": 5301
+ },
+ {
+ "epoch": 73.64279475982532,
+ "grad_norm": 0.23274236917495728,
+ "learning_rate": 0.0006,
+ "loss": 5.207432270050049,
+ "step": 5302
+ },
+ {
+ "epoch": 73.65676855895197,
+ "grad_norm": 0.18162772059440613,
+ "learning_rate": 0.0006,
+ "loss": 5.326632976531982,
+ "step": 5303
+ },
+ {
+ "epoch": 73.6707423580786,
+ "grad_norm": 0.11858413368463516,
+ "learning_rate": 0.0006,
+ "loss": 5.3730058670043945,
+ "step": 5304
+ },
+ {
+ "epoch": 73.68471615720524,
+ "grad_norm": 0.06031268090009689,
+ "learning_rate": 0.0006,
+ "loss": 5.330936431884766,
+ "step": 5305
+ },
+ {
+ "epoch": 73.69868995633188,
+ "grad_norm": 0.059304628521203995,
+ "learning_rate": 0.0006,
+ "loss": 5.280514717102051,
+ "step": 5306
+ },
+ {
+ "epoch": 73.71266375545852,
+ "grad_norm": 0.04702131822705269,
+ "learning_rate": 0.0006,
+ "loss": 5.288274765014648,
+ "step": 5307
+ },
+ {
+ "epoch": 73.72663755458515,
+ "grad_norm": 0.037792228162288666,
+ "learning_rate": 0.0006,
+ "loss": 5.173289775848389,
+ "step": 5308
+ },
+ {
+ "epoch": 73.74061135371178,
+ "grad_norm": 0.037763576954603195,
+ "learning_rate": 0.0006,
+ "loss": 5.084990978240967,
+ "step": 5309
+ },
+ {
+ "epoch": 73.75458515283843,
+ "grad_norm": 0.03337346389889717,
+ "learning_rate": 0.0006,
+ "loss": 5.121417045593262,
+ "step": 5310
+ },
+ {
+ "epoch": 73.76855895196506,
+ "grad_norm": 0.03184802085161209,
+ "learning_rate": 0.0006,
+ "loss": 5.208918571472168,
+ "step": 5311
+ },
+ {
+ "epoch": 73.78253275109171,
+ "grad_norm": 0.02832048572599888,
+ "learning_rate": 0.0006,
+ "loss": 4.923543453216553,
+ "step": 5312
+ },
+ {
+ "epoch": 73.79650655021834,
+ "grad_norm": 0.028297102078795433,
+ "learning_rate": 0.0006,
+ "loss": 4.944367408752441,
+ "step": 5313
+ },
+ {
+ "epoch": 73.81048034934499,
+ "grad_norm": 0.027701659128069878,
+ "learning_rate": 0.0006,
+ "loss": 4.858168601989746,
+ "step": 5314
+ },
+ {
+ "epoch": 73.82445414847162,
+ "grad_norm": 0.02214639261364937,
+ "learning_rate": 0.0006,
+ "loss": 4.887052536010742,
+ "step": 5315
+ },
+ {
+ "epoch": 73.83842794759825,
+ "grad_norm": 0.01995093561708927,
+ "learning_rate": 0.0006,
+ "loss": 4.885797500610352,
+ "step": 5316
+ },
+ {
+ "epoch": 73.8524017467249,
+ "grad_norm": 0.020473985001444817,
+ "learning_rate": 0.0006,
+ "loss": 4.917325019836426,
+ "step": 5317
+ },
+ {
+ "epoch": 73.86637554585153,
+ "grad_norm": 0.0192868635058403,
+ "learning_rate": 0.0006,
+ "loss": 4.801746845245361,
+ "step": 5318
+ },
+ {
+ "epoch": 73.88034934497817,
+ "grad_norm": 0.017108267173171043,
+ "learning_rate": 0.0006,
+ "loss": 4.893620491027832,
+ "step": 5319
+ },
+ {
+ "epoch": 73.8943231441048,
+ "grad_norm": 0.01718071475625038,
+ "learning_rate": 0.0006,
+ "loss": 4.940695285797119,
+ "step": 5320
+ },
+ {
+ "epoch": 73.90829694323143,
+ "grad_norm": 0.017821557819843292,
+ "learning_rate": 0.0006,
+ "loss": 4.767613410949707,
+ "step": 5321
+ },
+ {
+ "epoch": 73.92227074235808,
+ "grad_norm": 0.015262950211763382,
+ "learning_rate": 0.0006,
+ "loss": 4.910315036773682,
+ "step": 5322
+ },
+ {
+ "epoch": 73.93624454148471,
+ "grad_norm": 0.01769006811082363,
+ "learning_rate": 0.0006,
+ "loss": 4.7743449211120605,
+ "step": 5323
+ },
+ {
+ "epoch": 73.95021834061136,
+ "grad_norm": 0.014460788108408451,
+ "learning_rate": 0.0006,
+ "loss": 4.743494987487793,
+ "step": 5324
+ },
+ {
+ "epoch": 73.96419213973799,
+ "grad_norm": 0.01403993833810091,
+ "learning_rate": 0.0006,
+ "loss": 4.666640281677246,
+ "step": 5325
+ },
+ {
+ "epoch": 73.97816593886463,
+ "grad_norm": 0.013997809961438179,
+ "learning_rate": 0.0006,
+ "loss": 4.698747158050537,
+ "step": 5326
+ },
+ {
+ "epoch": 73.99213973799127,
+ "grad_norm": 0.014244182966649532,
+ "learning_rate": 0.0006,
+ "loss": 4.745250701904297,
+ "step": 5327
+ },
+ {
+ "epoch": 74.0,
+ "grad_norm": 0.015144186094403267,
+ "learning_rate": 0.0006,
+ "loss": 4.667946815490723,
+ "step": 5328
+ },
+ {
+ "epoch": 74.0,
+ "eval_loss": 5.005440711975098,
+ "eval_runtime": 56.7836,
+ "eval_samples_per_second": 43.005,
+ "eval_steps_per_second": 1.356,
+ "step": 5328
+ },
+ {
+ "epoch": 74.01397379912663,
+ "grad_norm": 0.020384902134537697,
+ "learning_rate": 0.0006,
+ "loss": 4.626955509185791,
+ "step": 5329
+ },
+ {
+ "epoch": 74.02794759825328,
+ "grad_norm": 0.03624720871448517,
+ "learning_rate": 0.0006,
+ "loss": 4.713998317718506,
+ "step": 5330
+ },
+ {
+ "epoch": 74.04192139737991,
+ "grad_norm": 0.05701754242181778,
+ "learning_rate": 0.0006,
+ "loss": 4.717874526977539,
+ "step": 5331
+ },
+ {
+ "epoch": 74.05589519650655,
+ "grad_norm": 0.03942621126770973,
+ "learning_rate": 0.0006,
+ "loss": 4.7026190757751465,
+ "step": 5332
+ },
+ {
+ "epoch": 74.06986899563319,
+ "grad_norm": 0.017803272232413292,
+ "learning_rate": 0.0006,
+ "loss": 4.702548027038574,
+ "step": 5333
+ },
+ {
+ "epoch": 74.08384279475983,
+ "grad_norm": 0.017024677246809006,
+ "learning_rate": 0.0006,
+ "loss": 4.69752311706543,
+ "step": 5334
+ },
+ {
+ "epoch": 74.09781659388646,
+ "grad_norm": 0.016467854380607605,
+ "learning_rate": 0.0006,
+ "loss": 4.624876976013184,
+ "step": 5335
+ },
+ {
+ "epoch": 74.1117903930131,
+ "grad_norm": 0.016569215804338455,
+ "learning_rate": 0.0006,
+ "loss": 4.645040512084961,
+ "step": 5336
+ },
+ {
+ "epoch": 74.12576419213974,
+ "grad_norm": 0.012872067280113697,
+ "learning_rate": 0.0006,
+ "loss": 4.64582633972168,
+ "step": 5337
+ },
+ {
+ "epoch": 74.13973799126637,
+ "grad_norm": 0.015689387917518616,
+ "learning_rate": 0.0006,
+ "loss": 4.596040725708008,
+ "step": 5338
+ },
+ {
+ "epoch": 74.15371179039302,
+ "grad_norm": 0.013082082383334637,
+ "learning_rate": 0.0006,
+ "loss": 4.648033142089844,
+ "step": 5339
+ },
+ {
+ "epoch": 74.16768558951965,
+ "grad_norm": 0.013728172518312931,
+ "learning_rate": 0.0006,
+ "loss": 4.657873153686523,
+ "step": 5340
+ },
+ {
+ "epoch": 74.18165938864628,
+ "grad_norm": 0.013120594434440136,
+ "learning_rate": 0.0006,
+ "loss": 4.696074485778809,
+ "step": 5341
+ },
+ {
+ "epoch": 74.19563318777293,
+ "grad_norm": 0.01431245356798172,
+ "learning_rate": 0.0006,
+ "loss": 4.5679121017456055,
+ "step": 5342
+ },
+ {
+ "epoch": 74.20960698689956,
+ "grad_norm": 0.013994511216878891,
+ "learning_rate": 0.0006,
+ "loss": 4.5236406326293945,
+ "step": 5343
+ },
+ {
+ "epoch": 74.2235807860262,
+ "grad_norm": 0.011971707455813885,
+ "learning_rate": 0.0006,
+ "loss": 4.626547813415527,
+ "step": 5344
+ },
+ {
+ "epoch": 74.23755458515284,
+ "grad_norm": 0.012805287726223469,
+ "learning_rate": 0.0006,
+ "loss": 4.556386947631836,
+ "step": 5345
+ },
+ {
+ "epoch": 74.25152838427948,
+ "grad_norm": 0.011694584973156452,
+ "learning_rate": 0.0006,
+ "loss": 4.656987190246582,
+ "step": 5346
+ },
+ {
+ "epoch": 74.26550218340611,
+ "grad_norm": 0.013396821916103363,
+ "learning_rate": 0.0006,
+ "loss": 4.526467323303223,
+ "step": 5347
+ },
+ {
+ "epoch": 74.27947598253274,
+ "grad_norm": 0.014398031868040562,
+ "learning_rate": 0.0006,
+ "loss": 4.595100402832031,
+ "step": 5348
+ },
+ {
+ "epoch": 74.29344978165939,
+ "grad_norm": 0.02165800705552101,
+ "learning_rate": 0.0006,
+ "loss": 4.460549354553223,
+ "step": 5349
+ },
+ {
+ "epoch": 74.30742358078602,
+ "grad_norm": 0.03276512026786804,
+ "learning_rate": 0.0006,
+ "loss": 4.498189449310303,
+ "step": 5350
+ },
+ {
+ "epoch": 74.32139737991267,
+ "grad_norm": 0.04616224020719528,
+ "learning_rate": 0.0006,
+ "loss": 4.5268659591674805,
+ "step": 5351
+ },
+ {
+ "epoch": 74.3353711790393,
+ "grad_norm": 0.03863786533474922,
+ "learning_rate": 0.0006,
+ "loss": 4.554747104644775,
+ "step": 5352
+ },
+ {
+ "epoch": 74.34934497816595,
+ "grad_norm": 0.01749865710735321,
+ "learning_rate": 0.0006,
+ "loss": 4.6656036376953125,
+ "step": 5353
+ },
+ {
+ "epoch": 74.36331877729258,
+ "grad_norm": 0.021814396604895592,
+ "learning_rate": 0.0006,
+ "loss": 4.607456207275391,
+ "step": 5354
+ },
+ {
+ "epoch": 74.37729257641921,
+ "grad_norm": 0.017018496990203857,
+ "learning_rate": 0.0006,
+ "loss": 4.54205322265625,
+ "step": 5355
+ },
+ {
+ "epoch": 74.39126637554585,
+ "grad_norm": 0.018297340720891953,
+ "learning_rate": 0.0006,
+ "loss": 4.608870506286621,
+ "step": 5356
+ },
+ {
+ "epoch": 74.40524017467249,
+ "grad_norm": 0.01473482046276331,
+ "learning_rate": 0.0006,
+ "loss": 4.384891510009766,
+ "step": 5357
+ },
+ {
+ "epoch": 74.41921397379913,
+ "grad_norm": 0.01579415239393711,
+ "learning_rate": 0.0006,
+ "loss": 4.588859558105469,
+ "step": 5358
+ },
+ {
+ "epoch": 74.43318777292576,
+ "grad_norm": 0.013906026259064674,
+ "learning_rate": 0.0006,
+ "loss": 4.525175094604492,
+ "step": 5359
+ },
+ {
+ "epoch": 74.44716157205241,
+ "grad_norm": 0.014163361862301826,
+ "learning_rate": 0.0006,
+ "loss": 4.495802879333496,
+ "step": 5360
+ },
+ {
+ "epoch": 74.46113537117904,
+ "grad_norm": 0.014009335078299046,
+ "learning_rate": 0.0006,
+ "loss": 4.5960869789123535,
+ "step": 5361
+ },
+ {
+ "epoch": 74.47510917030567,
+ "grad_norm": 0.012765258550643921,
+ "learning_rate": 0.0006,
+ "loss": 4.515678405761719,
+ "step": 5362
+ },
+ {
+ "epoch": 74.48908296943232,
+ "grad_norm": 0.014071918092668056,
+ "learning_rate": 0.0006,
+ "loss": 4.394025802612305,
+ "step": 5363
+ },
+ {
+ "epoch": 74.50305676855895,
+ "grad_norm": 0.01244515273720026,
+ "learning_rate": 0.0006,
+ "loss": 4.576985836029053,
+ "step": 5364
+ },
+ {
+ "epoch": 74.5170305676856,
+ "grad_norm": 0.012975298799574375,
+ "learning_rate": 0.0006,
+ "loss": 4.540657997131348,
+ "step": 5365
+ },
+ {
+ "epoch": 74.53100436681223,
+ "grad_norm": 0.012380947358906269,
+ "learning_rate": 0.0006,
+ "loss": 4.55731725692749,
+ "step": 5366
+ },
+ {
+ "epoch": 74.54497816593886,
+ "grad_norm": 0.012577617540955544,
+ "learning_rate": 0.0006,
+ "loss": 4.500924110412598,
+ "step": 5367
+ },
+ {
+ "epoch": 74.5589519650655,
+ "grad_norm": 0.012089254334568977,
+ "learning_rate": 0.0006,
+ "loss": 4.561795234680176,
+ "step": 5368
+ },
+ {
+ "epoch": 74.57292576419214,
+ "grad_norm": 0.012024011462926865,
+ "learning_rate": 0.0006,
+ "loss": 4.430635452270508,
+ "step": 5369
+ },
+ {
+ "epoch": 74.58689956331878,
+ "grad_norm": 0.012361960485577583,
+ "learning_rate": 0.0006,
+ "loss": 4.4664106369018555,
+ "step": 5370
+ },
+ {
+ "epoch": 74.60087336244541,
+ "grad_norm": 0.012149893678724766,
+ "learning_rate": 0.0006,
+ "loss": 4.522271156311035,
+ "step": 5371
+ },
+ {
+ "epoch": 74.61484716157206,
+ "grad_norm": 0.01166840922087431,
+ "learning_rate": 0.0006,
+ "loss": 4.3870463371276855,
+ "step": 5372
+ },
+ {
+ "epoch": 74.62882096069869,
+ "grad_norm": 0.01160407718271017,
+ "learning_rate": 0.0006,
+ "loss": 4.504255771636963,
+ "step": 5373
+ },
+ {
+ "epoch": 74.64279475982532,
+ "grad_norm": 0.011259177699685097,
+ "learning_rate": 0.0006,
+ "loss": 4.507267951965332,
+ "step": 5374
+ },
+ {
+ "epoch": 74.65676855895197,
+ "grad_norm": 0.012285715900361538,
+ "learning_rate": 0.0006,
+ "loss": 4.472966194152832,
+ "step": 5375
+ },
+ {
+ "epoch": 74.6707423580786,
+ "grad_norm": 0.011214440688490868,
+ "learning_rate": 0.0006,
+ "loss": 4.5840253829956055,
+ "step": 5376
+ },
+ {
+ "epoch": 74.68471615720524,
+ "grad_norm": 0.012394879944622517,
+ "learning_rate": 0.0006,
+ "loss": 4.417830944061279,
+ "step": 5377
+ },
+ {
+ "epoch": 74.69868995633188,
+ "grad_norm": 0.012360049411654472,
+ "learning_rate": 0.0006,
+ "loss": 4.535454750061035,
+ "step": 5378
+ },
+ {
+ "epoch": 74.71266375545852,
+ "grad_norm": 0.011961652897298336,
+ "learning_rate": 0.0006,
+ "loss": 4.403409957885742,
+ "step": 5379
+ },
+ {
+ "epoch": 74.72663755458515,
+ "grad_norm": 0.012078274972736835,
+ "learning_rate": 0.0006,
+ "loss": 4.448366641998291,
+ "step": 5380
+ },
+ {
+ "epoch": 74.74061135371178,
+ "grad_norm": 0.013239333406090736,
+ "learning_rate": 0.0006,
+ "loss": 4.543450355529785,
+ "step": 5381
+ },
+ {
+ "epoch": 74.75458515283843,
+ "grad_norm": 0.01476441603153944,
+ "learning_rate": 0.0006,
+ "loss": 4.488677978515625,
+ "step": 5382
+ },
+ {
+ "epoch": 74.76855895196506,
+ "grad_norm": 0.013968009501695633,
+ "learning_rate": 0.0006,
+ "loss": 4.603639602661133,
+ "step": 5383
+ },
+ {
+ "epoch": 74.78253275109171,
+ "grad_norm": 0.012322458438575268,
+ "learning_rate": 0.0006,
+ "loss": 4.637874603271484,
+ "step": 5384
+ },
+ {
+ "epoch": 74.79650655021834,
+ "grad_norm": 0.013352593407034874,
+ "learning_rate": 0.0006,
+ "loss": 4.474833965301514,
+ "step": 5385
+ },
+ {
+ "epoch": 74.81048034934499,
+ "grad_norm": 0.014671728946268559,
+ "learning_rate": 0.0006,
+ "loss": 4.516847610473633,
+ "step": 5386
+ },
+ {
+ "epoch": 74.82445414847162,
+ "grad_norm": 0.017163321375846863,
+ "learning_rate": 0.0006,
+ "loss": 4.462325096130371,
+ "step": 5387
+ },
+ {
+ "epoch": 74.83842794759825,
+ "grad_norm": 0.02012511156499386,
+ "learning_rate": 0.0006,
+ "loss": 4.355982780456543,
+ "step": 5388
+ },
+ {
+ "epoch": 74.8524017467249,
+ "grad_norm": 0.01881697215139866,
+ "learning_rate": 0.0006,
+ "loss": 4.451120376586914,
+ "step": 5389
+ },
+ {
+ "epoch": 74.86637554585153,
+ "grad_norm": 0.017279407009482384,
+ "learning_rate": 0.0006,
+ "loss": 4.461202621459961,
+ "step": 5390
+ },
+ {
+ "epoch": 74.88034934497817,
+ "grad_norm": 0.015470407903194427,
+ "learning_rate": 0.0006,
+ "loss": 4.4968671798706055,
+ "step": 5391
+ },
+ {
+ "epoch": 74.8943231441048,
+ "grad_norm": 0.01411783043295145,
+ "learning_rate": 0.0006,
+ "loss": 4.558697700500488,
+ "step": 5392
+ },
+ {
+ "epoch": 74.90829694323143,
+ "grad_norm": 0.017178812995553017,
+ "learning_rate": 0.0006,
+ "loss": 4.254153251647949,
+ "step": 5393
+ },
+ {
+ "epoch": 74.92227074235808,
+ "grad_norm": 0.019193004816770554,
+ "learning_rate": 0.0006,
+ "loss": 4.394747257232666,
+ "step": 5394
+ },
+ {
+ "epoch": 74.93624454148471,
+ "grad_norm": 0.018942881375551224,
+ "learning_rate": 0.0006,
+ "loss": 4.459539413452148,
+ "step": 5395
+ },
+ {
+ "epoch": 74.95021834061136,
+ "grad_norm": 0.01641223020851612,
+ "learning_rate": 0.0006,
+ "loss": 4.485820770263672,
+ "step": 5396
+ },
+ {
+ "epoch": 74.96419213973799,
+ "grad_norm": 0.014485753141343594,
+ "learning_rate": 0.0006,
+ "loss": 4.5068583488464355,
+ "step": 5397
+ },
+ {
+ "epoch": 74.97816593886463,
+ "grad_norm": 0.018242299556732178,
+ "learning_rate": 0.0006,
+ "loss": 4.508835792541504,
+ "step": 5398
+ },
+ {
+ "epoch": 74.99213973799127,
+ "grad_norm": 0.016491873189806938,
+ "learning_rate": 0.0006,
+ "loss": 4.4760870933532715,
+ "step": 5399
+ },
+ {
+ "epoch": 75.0,
+ "grad_norm": 0.014543353579938412,
+ "learning_rate": 0.0006,
+ "loss": 4.540217399597168,
+ "step": 5400
+ },
+ {
+ "epoch": 75.0,
+ "eval_loss": 4.7617573738098145,
+ "eval_runtime": 57.321,
+ "eval_samples_per_second": 42.602,
+ "eval_steps_per_second": 1.343,
+ "step": 5400
+ },
+ {
+ "epoch": 75.01397379912663,
+ "grad_norm": 0.014417114667594433,
+ "learning_rate": 0.0006,
+ "loss": 4.332956790924072,
+ "step": 5401
+ },
+ {
+ "epoch": 75.02794759825328,
+ "grad_norm": 0.01607639715075493,
+ "learning_rate": 0.0006,
+ "loss": 4.352453708648682,
+ "step": 5402
+ },
+ {
+ "epoch": 75.04192139737991,
+ "grad_norm": 0.015438318252563477,
+ "learning_rate": 0.0006,
+ "loss": 4.387139320373535,
+ "step": 5403
+ },
+ {
+ "epoch": 75.05589519650655,
+ "grad_norm": 0.015539007261395454,
+ "learning_rate": 0.0006,
+ "loss": 4.453171730041504,
+ "step": 5404
+ },
+ {
+ "epoch": 75.06986899563319,
+ "grad_norm": 0.017361903563141823,
+ "learning_rate": 0.0006,
+ "loss": 4.512143135070801,
+ "step": 5405
+ },
+ {
+ "epoch": 75.08384279475983,
+ "grad_norm": 0.01815599389374256,
+ "learning_rate": 0.0006,
+ "loss": 4.508997440338135,
+ "step": 5406
+ },
+ {
+ "epoch": 75.09781659388646,
+ "grad_norm": 0.01620607078075409,
+ "learning_rate": 0.0006,
+ "loss": 4.322444915771484,
+ "step": 5407
+ },
+ {
+ "epoch": 75.1117903930131,
+ "grad_norm": 0.014773265458643436,
+ "learning_rate": 0.0006,
+ "loss": 4.359251976013184,
+ "step": 5408
+ },
+ {
+ "epoch": 75.12576419213974,
+ "grad_norm": 0.014099023304879665,
+ "learning_rate": 0.0006,
+ "loss": 4.313380241394043,
+ "step": 5409
+ },
+ {
+ "epoch": 75.13973799126637,
+ "grad_norm": 0.014387735165655613,
+ "learning_rate": 0.0006,
+ "loss": 4.435023307800293,
+ "step": 5410
+ },
+ {
+ "epoch": 75.15371179039302,
+ "grad_norm": 0.014886162243783474,
+ "learning_rate": 0.0006,
+ "loss": 4.335939407348633,
+ "step": 5411
+ },
+ {
+ "epoch": 75.16768558951965,
+ "grad_norm": 0.015766102820634842,
+ "learning_rate": 0.0006,
+ "loss": 4.408674240112305,
+ "step": 5412
+ },
+ {
+ "epoch": 75.18165938864628,
+ "grad_norm": 0.016383284702897072,
+ "learning_rate": 0.0006,
+ "loss": 4.385453224182129,
+ "step": 5413
+ },
+ {
+ "epoch": 75.19563318777293,
+ "grad_norm": 0.013535384088754654,
+ "learning_rate": 0.0006,
+ "loss": 4.292785167694092,
+ "step": 5414
+ },
+ {
+ "epoch": 75.20960698689956,
+ "grad_norm": 0.01354036945849657,
+ "learning_rate": 0.0006,
+ "loss": 4.306252479553223,
+ "step": 5415
+ },
+ {
+ "epoch": 75.2235807860262,
+ "grad_norm": 0.015818078070878983,
+ "learning_rate": 0.0006,
+ "loss": 4.425487995147705,
+ "step": 5416
+ },
+ {
+ "epoch": 75.23755458515284,
+ "grad_norm": 0.014363568276166916,
+ "learning_rate": 0.0006,
+ "loss": 4.395087242126465,
+ "step": 5417
+ },
+ {
+ "epoch": 75.25152838427948,
+ "grad_norm": 0.013895470649003983,
+ "learning_rate": 0.0006,
+ "loss": 4.364090442657471,
+ "step": 5418
+ },
+ {
+ "epoch": 75.26550218340611,
+ "grad_norm": 0.014178668148815632,
+ "learning_rate": 0.0006,
+ "loss": 4.44619083404541,
+ "step": 5419
+ },
+ {
+ "epoch": 75.27947598253274,
+ "grad_norm": 0.0152514623478055,
+ "learning_rate": 0.0006,
+ "loss": 4.348445892333984,
+ "step": 5420
+ },
+ {
+ "epoch": 75.29344978165939,
+ "grad_norm": 0.015397852286696434,
+ "learning_rate": 0.0006,
+ "loss": 4.476832389831543,
+ "step": 5421
+ },
+ {
+ "epoch": 75.30742358078602,
+ "grad_norm": 0.018009863793849945,
+ "learning_rate": 0.0006,
+ "loss": 4.431162357330322,
+ "step": 5422
+ },
+ {
+ "epoch": 75.32139737991267,
+ "grad_norm": 0.017038367688655853,
+ "learning_rate": 0.0006,
+ "loss": 4.474787712097168,
+ "step": 5423
+ },
+ {
+ "epoch": 75.3353711790393,
+ "grad_norm": 0.013763288035988808,
+ "learning_rate": 0.0006,
+ "loss": 4.470419883728027,
+ "step": 5424
+ },
+ {
+ "epoch": 75.34934497816595,
+ "grad_norm": 0.013972110114991665,
+ "learning_rate": 0.0006,
+ "loss": 4.373147964477539,
+ "step": 5425
+ },
+ {
+ "epoch": 75.36331877729258,
+ "grad_norm": 0.017197875306010246,
+ "learning_rate": 0.0006,
+ "loss": 4.494963645935059,
+ "step": 5426
+ },
+ {
+ "epoch": 75.37729257641921,
+ "grad_norm": 0.01687047816812992,
+ "learning_rate": 0.0006,
+ "loss": 4.435934066772461,
+ "step": 5427
+ },
+ {
+ "epoch": 75.39126637554585,
+ "grad_norm": 0.01578272506594658,
+ "learning_rate": 0.0006,
+ "loss": 4.37918758392334,
+ "step": 5428
+ },
+ {
+ "epoch": 75.40524017467249,
+ "grad_norm": 0.015956247225403786,
+ "learning_rate": 0.0006,
+ "loss": 4.455668926239014,
+ "step": 5429
+ },
+ {
+ "epoch": 75.41921397379913,
+ "grad_norm": 0.013495155610144138,
+ "learning_rate": 0.0006,
+ "loss": 4.4271039962768555,
+ "step": 5430
+ },
+ {
+ "epoch": 75.43318777292576,
+ "grad_norm": 0.012851621024310589,
+ "learning_rate": 0.0006,
+ "loss": 4.213107109069824,
+ "step": 5431
+ },
+ {
+ "epoch": 75.44716157205241,
+ "grad_norm": 0.013351158238947392,
+ "learning_rate": 0.0006,
+ "loss": 4.3070783615112305,
+ "step": 5432
+ },
+ {
+ "epoch": 75.46113537117904,
+ "grad_norm": 0.014765217900276184,
+ "learning_rate": 0.0006,
+ "loss": 4.385021209716797,
+ "step": 5433
+ },
+ {
+ "epoch": 75.47510917030567,
+ "grad_norm": 0.016091614961624146,
+ "learning_rate": 0.0006,
+ "loss": 4.312644004821777,
+ "step": 5434
+ },
+ {
+ "epoch": 75.48908296943232,
+ "grad_norm": 0.01612013578414917,
+ "learning_rate": 0.0006,
+ "loss": 4.4482011795043945,
+ "step": 5435
+ },
+ {
+ "epoch": 75.50305676855895,
+ "grad_norm": 0.013157603330910206,
+ "learning_rate": 0.0006,
+ "loss": 4.427262306213379,
+ "step": 5436
+ },
+ {
+ "epoch": 75.5170305676856,
+ "grad_norm": 0.016039809212088585,
+ "learning_rate": 0.0006,
+ "loss": 4.396635055541992,
+ "step": 5437
+ },
+ {
+ "epoch": 75.53100436681223,
+ "grad_norm": 0.01794283092021942,
+ "learning_rate": 0.0006,
+ "loss": 4.476870059967041,
+ "step": 5438
+ },
+ {
+ "epoch": 75.54497816593886,
+ "grad_norm": 0.017030486837029457,
+ "learning_rate": 0.0006,
+ "loss": 4.361156463623047,
+ "step": 5439
+ },
+ {
+ "epoch": 75.5589519650655,
+ "grad_norm": 0.017915161326527596,
+ "learning_rate": 0.0006,
+ "loss": 4.286147117614746,
+ "step": 5440
+ },
+ {
+ "epoch": 75.57292576419214,
+ "grad_norm": 0.017514433711767197,
+ "learning_rate": 0.0006,
+ "loss": 4.394091606140137,
+ "step": 5441
+ },
+ {
+ "epoch": 75.58689956331878,
+ "grad_norm": 0.015128589235246181,
+ "learning_rate": 0.0006,
+ "loss": 4.383126735687256,
+ "step": 5442
+ },
+ {
+ "epoch": 75.60087336244541,
+ "grad_norm": 0.0202616099268198,
+ "learning_rate": 0.0006,
+ "loss": 4.336188316345215,
+ "step": 5443
+ },
+ {
+ "epoch": 75.61484716157206,
+ "grad_norm": 0.019978471100330353,
+ "learning_rate": 0.0006,
+ "loss": 4.461453437805176,
+ "step": 5444
+ },
+ {
+ "epoch": 75.62882096069869,
+ "grad_norm": 0.01738741435110569,
+ "learning_rate": 0.0006,
+ "loss": 4.371197700500488,
+ "step": 5445
+ },
+ {
+ "epoch": 75.64279475982532,
+ "grad_norm": 0.016157809644937515,
+ "learning_rate": 0.0006,
+ "loss": 4.3312835693359375,
+ "step": 5446
+ },
+ {
+ "epoch": 75.65676855895197,
+ "grad_norm": 0.016394613310694695,
+ "learning_rate": 0.0006,
+ "loss": 4.380190849304199,
+ "step": 5447
+ },
+ {
+ "epoch": 75.6707423580786,
+ "grad_norm": 0.014727243222296238,
+ "learning_rate": 0.0006,
+ "loss": 4.308028221130371,
+ "step": 5448
+ },
+ {
+ "epoch": 75.68471615720524,
+ "grad_norm": 0.01695958711206913,
+ "learning_rate": 0.0006,
+ "loss": 4.399503707885742,
+ "step": 5449
+ },
+ {
+ "epoch": 75.69868995633188,
+ "grad_norm": 0.015720795840024948,
+ "learning_rate": 0.0006,
+ "loss": 4.342249870300293,
+ "step": 5450
+ },
+ {
+ "epoch": 75.71266375545852,
+ "grad_norm": 0.01574772410094738,
+ "learning_rate": 0.0006,
+ "loss": 4.331503868103027,
+ "step": 5451
+ },
+ {
+ "epoch": 75.72663755458515,
+ "grad_norm": 0.014812648296356201,
+ "learning_rate": 0.0006,
+ "loss": 4.332413673400879,
+ "step": 5452
+ },
+ {
+ "epoch": 75.74061135371178,
+ "grad_norm": 0.012850426137447357,
+ "learning_rate": 0.0006,
+ "loss": 4.2884063720703125,
+ "step": 5453
+ },
+ {
+ "epoch": 75.75458515283843,
+ "grad_norm": 0.012876970693469048,
+ "learning_rate": 0.0006,
+ "loss": 4.399855613708496,
+ "step": 5454
+ },
+ {
+ "epoch": 75.76855895196506,
+ "grad_norm": 0.012958863750100136,
+ "learning_rate": 0.0006,
+ "loss": 4.411185264587402,
+ "step": 5455
+ },
+ {
+ "epoch": 75.78253275109171,
+ "grad_norm": 0.012197135016322136,
+ "learning_rate": 0.0006,
+ "loss": 4.272603988647461,
+ "step": 5456
+ },
+ {
+ "epoch": 75.79650655021834,
+ "grad_norm": 0.013483566232025623,
+ "learning_rate": 0.0006,
+ "loss": 4.3564348220825195,
+ "step": 5457
+ },
+ {
+ "epoch": 75.81048034934499,
+ "grad_norm": 0.012365429662168026,
+ "learning_rate": 0.0006,
+ "loss": 4.354825973510742,
+ "step": 5458
+ },
+ {
+ "epoch": 75.82445414847162,
+ "grad_norm": 0.012238141149282455,
+ "learning_rate": 0.0006,
+ "loss": 4.440438747406006,
+ "step": 5459
+ },
+ {
+ "epoch": 75.83842794759825,
+ "grad_norm": 0.013380305841565132,
+ "learning_rate": 0.0006,
+ "loss": 4.2813215255737305,
+ "step": 5460
+ },
+ {
+ "epoch": 75.8524017467249,
+ "grad_norm": 0.01443779468536377,
+ "learning_rate": 0.0006,
+ "loss": 4.289078235626221,
+ "step": 5461
+ },
+ {
+ "epoch": 75.86637554585153,
+ "grad_norm": 0.012646574527025223,
+ "learning_rate": 0.0006,
+ "loss": 4.374177932739258,
+ "step": 5462
+ },
+ {
+ "epoch": 75.88034934497817,
+ "grad_norm": 0.012403388507664204,
+ "learning_rate": 0.0006,
+ "loss": 4.343016147613525,
+ "step": 5463
+ },
+ {
+ "epoch": 75.8943231441048,
+ "grad_norm": 0.01319252047687769,
+ "learning_rate": 0.0006,
+ "loss": 4.309957027435303,
+ "step": 5464
+ },
+ {
+ "epoch": 75.90829694323143,
+ "grad_norm": 0.012597428634762764,
+ "learning_rate": 0.0006,
+ "loss": 4.403654098510742,
+ "step": 5465
+ },
+ {
+ "epoch": 75.92227074235808,
+ "grad_norm": 0.012724226340651512,
+ "learning_rate": 0.0006,
+ "loss": 4.326869010925293,
+ "step": 5466
+ },
+ {
+ "epoch": 75.93624454148471,
+ "grad_norm": 0.012698772363364697,
+ "learning_rate": 0.0006,
+ "loss": 4.268430233001709,
+ "step": 5467
+ },
+ {
+ "epoch": 75.95021834061136,
+ "grad_norm": 0.01376490481197834,
+ "learning_rate": 0.0006,
+ "loss": 4.42446231842041,
+ "step": 5468
+ },
+ {
+ "epoch": 75.96419213973799,
+ "grad_norm": 0.013175144791603088,
+ "learning_rate": 0.0006,
+ "loss": 4.374675750732422,
+ "step": 5469
+ },
+ {
+ "epoch": 75.97816593886463,
+ "grad_norm": 0.012731322087347507,
+ "learning_rate": 0.0006,
+ "loss": 4.372966766357422,
+ "step": 5470
+ },
+ {
+ "epoch": 75.99213973799127,
+ "grad_norm": 0.01360505260527134,
+ "learning_rate": 0.0006,
+ "loss": 4.3095903396606445,
+ "step": 5471
+ },
+ {
+ "epoch": 76.0,
+ "grad_norm": 0.014011901803314686,
+ "learning_rate": 0.0006,
+ "loss": 4.26954460144043,
+ "step": 5472
+ },
+ {
+ "epoch": 76.0,
+ "eval_loss": 4.689640998840332,
+ "eval_runtime": 56.9095,
+ "eval_samples_per_second": 42.91,
+ "eval_steps_per_second": 1.353,
+ "step": 5472
+ },
+ {
+ "epoch": 76.01397379912663,
+ "grad_norm": 0.015351055189967155,
+ "learning_rate": 0.0006,
+ "loss": 4.348268985748291,
+ "step": 5473
+ },
+ {
+ "epoch": 76.02794759825328,
+ "grad_norm": 0.017164217308163643,
+ "learning_rate": 0.0006,
+ "loss": 4.244204998016357,
+ "step": 5474
+ },
+ {
+ "epoch": 76.04192139737991,
+ "grad_norm": 0.021624932065606117,
+ "learning_rate": 0.0006,
+ "loss": 4.311168670654297,
+ "step": 5475
+ },
+ {
+ "epoch": 76.05589519650655,
+ "grad_norm": 0.028712468221783638,
+ "learning_rate": 0.0006,
+ "loss": 4.322287082672119,
+ "step": 5476
+ },
+ {
+ "epoch": 76.06986899563319,
+ "grad_norm": 0.027915487065911293,
+ "learning_rate": 0.0006,
+ "loss": 4.381207466125488,
+ "step": 5477
+ },
+ {
+ "epoch": 76.08384279475983,
+ "grad_norm": 0.017955303192138672,
+ "learning_rate": 0.0006,
+ "loss": 4.2635016441345215,
+ "step": 5478
+ },
+ {
+ "epoch": 76.09781659388646,
+ "grad_norm": 0.019113514572381973,
+ "learning_rate": 0.0006,
+ "loss": 4.399471282958984,
+ "step": 5479
+ },
+ {
+ "epoch": 76.1117903930131,
+ "grad_norm": 0.017502542585134506,
+ "learning_rate": 0.0006,
+ "loss": 4.321204662322998,
+ "step": 5480
+ },
+ {
+ "epoch": 76.12576419213974,
+ "grad_norm": 0.01852525770664215,
+ "learning_rate": 0.0006,
+ "loss": 4.217884063720703,
+ "step": 5481
+ },
+ {
+ "epoch": 76.13973799126637,
+ "grad_norm": 0.019836828112602234,
+ "learning_rate": 0.0006,
+ "loss": 4.278532028198242,
+ "step": 5482
+ },
+ {
+ "epoch": 76.15371179039302,
+ "grad_norm": 0.02101975679397583,
+ "learning_rate": 0.0006,
+ "loss": 4.246720314025879,
+ "step": 5483
+ },
+ {
+ "epoch": 76.16768558951965,
+ "grad_norm": 0.021202990785241127,
+ "learning_rate": 0.0006,
+ "loss": 4.517517566680908,
+ "step": 5484
+ },
+ {
+ "epoch": 76.18165938864628,
+ "grad_norm": 0.021428769454360008,
+ "learning_rate": 0.0006,
+ "loss": 4.3373517990112305,
+ "step": 5485
+ },
+ {
+ "epoch": 76.19563318777293,
+ "grad_norm": 0.018263190984725952,
+ "learning_rate": 0.0006,
+ "loss": 4.402856826782227,
+ "step": 5486
+ },
+ {
+ "epoch": 76.20960698689956,
+ "grad_norm": 0.019025536254048347,
+ "learning_rate": 0.0006,
+ "loss": 4.314227104187012,
+ "step": 5487
+ },
+ {
+ "epoch": 76.2235807860262,
+ "grad_norm": 0.020918630063533783,
+ "learning_rate": 0.0006,
+ "loss": 4.454369068145752,
+ "step": 5488
+ },
+ {
+ "epoch": 76.23755458515284,
+ "grad_norm": 0.019307617098093033,
+ "learning_rate": 0.0006,
+ "loss": 4.247509002685547,
+ "step": 5489
+ },
+ {
+ "epoch": 76.25152838427948,
+ "grad_norm": 0.014956331811845303,
+ "learning_rate": 0.0006,
+ "loss": 4.365175247192383,
+ "step": 5490
+ },
+ {
+ "epoch": 76.26550218340611,
+ "grad_norm": 0.017213059589266777,
+ "learning_rate": 0.0006,
+ "loss": 4.384005069732666,
+ "step": 5491
+ },
+ {
+ "epoch": 76.27947598253274,
+ "grad_norm": 0.017575597390532494,
+ "learning_rate": 0.0006,
+ "loss": 4.41934871673584,
+ "step": 5492
+ },
+ {
+ "epoch": 76.29344978165939,
+ "grad_norm": 0.015879256650805473,
+ "learning_rate": 0.0006,
+ "loss": 4.26675271987915,
+ "step": 5493
+ },
+ {
+ "epoch": 76.30742358078602,
+ "grad_norm": 0.01564522460103035,
+ "learning_rate": 0.0006,
+ "loss": 4.277339935302734,
+ "step": 5494
+ },
+ {
+ "epoch": 76.32139737991267,
+ "grad_norm": 0.016414586454629898,
+ "learning_rate": 0.0006,
+ "loss": 4.4138593673706055,
+ "step": 5495
+ },
+ {
+ "epoch": 76.3353711790393,
+ "grad_norm": 0.014732337556779385,
+ "learning_rate": 0.0006,
+ "loss": 4.281006336212158,
+ "step": 5496
+ },
+ {
+ "epoch": 76.34934497816595,
+ "grad_norm": 0.015461455099284649,
+ "learning_rate": 0.0006,
+ "loss": 4.263422966003418,
+ "step": 5497
+ },
+ {
+ "epoch": 76.36331877729258,
+ "grad_norm": 0.014529814012348652,
+ "learning_rate": 0.0006,
+ "loss": 4.36165189743042,
+ "step": 5498
+ },
+ {
+ "epoch": 76.37729257641921,
+ "grad_norm": 0.01469483319669962,
+ "learning_rate": 0.0006,
+ "loss": 4.309642791748047,
+ "step": 5499
+ },
+ {
+ "epoch": 76.39126637554585,
+ "grad_norm": 0.016490574926137924,
+ "learning_rate": 0.0006,
+ "loss": 4.288150787353516,
+ "step": 5500
+ },
+ {
+ "epoch": 76.40524017467249,
+ "grad_norm": 0.014593565836548805,
+ "learning_rate": 0.0006,
+ "loss": 4.271614074707031,
+ "step": 5501
+ },
+ {
+ "epoch": 76.41921397379913,
+ "grad_norm": 0.012796535156667233,
+ "learning_rate": 0.0006,
+ "loss": 4.420042037963867,
+ "step": 5502
+ },
+ {
+ "epoch": 76.43318777292576,
+ "grad_norm": 0.016086025163531303,
+ "learning_rate": 0.0006,
+ "loss": 4.289943695068359,
+ "step": 5503
+ },
+ {
+ "epoch": 76.44716157205241,
+ "grad_norm": 0.015439470298588276,
+ "learning_rate": 0.0006,
+ "loss": 4.3023271560668945,
+ "step": 5504
+ },
+ {
+ "epoch": 76.46113537117904,
+ "grad_norm": 0.015066367574036121,
+ "learning_rate": 0.0006,
+ "loss": 4.197081565856934,
+ "step": 5505
+ },
+ {
+ "epoch": 76.47510917030567,
+ "grad_norm": 0.016653263941407204,
+ "learning_rate": 0.0006,
+ "loss": 4.333785057067871,
+ "step": 5506
+ },
+ {
+ "epoch": 76.48908296943232,
+ "grad_norm": 0.014791973866522312,
+ "learning_rate": 0.0006,
+ "loss": 4.323811054229736,
+ "step": 5507
+ },
+ {
+ "epoch": 76.50305676855895,
+ "grad_norm": 0.014694827608764172,
+ "learning_rate": 0.0006,
+ "loss": 4.195852279663086,
+ "step": 5508
+ },
+ {
+ "epoch": 76.5170305676856,
+ "grad_norm": 0.015120396390557289,
+ "learning_rate": 0.0006,
+ "loss": 4.42379903793335,
+ "step": 5509
+ },
+ {
+ "epoch": 76.53100436681223,
+ "grad_norm": 0.014584025368094444,
+ "learning_rate": 0.0006,
+ "loss": 4.2702531814575195,
+ "step": 5510
+ },
+ {
+ "epoch": 76.54497816593886,
+ "grad_norm": 0.013981613330543041,
+ "learning_rate": 0.0006,
+ "loss": 4.383029937744141,
+ "step": 5511
+ },
+ {
+ "epoch": 76.5589519650655,
+ "grad_norm": 0.012930642813444138,
+ "learning_rate": 0.0006,
+ "loss": 4.296136856079102,
+ "step": 5512
+ },
+ {
+ "epoch": 76.57292576419214,
+ "grad_norm": 0.013060989789664745,
+ "learning_rate": 0.0006,
+ "loss": 4.346380233764648,
+ "step": 5513
+ },
+ {
+ "epoch": 76.58689956331878,
+ "grad_norm": 0.013209798373281956,
+ "learning_rate": 0.0006,
+ "loss": 4.316067218780518,
+ "step": 5514
+ },
+ {
+ "epoch": 76.60087336244541,
+ "grad_norm": 0.012423216365277767,
+ "learning_rate": 0.0006,
+ "loss": 4.370274543762207,
+ "step": 5515
+ },
+ {
+ "epoch": 76.61484716157206,
+ "grad_norm": 0.013205411843955517,
+ "learning_rate": 0.0006,
+ "loss": 4.429969310760498,
+ "step": 5516
+ },
+ {
+ "epoch": 76.62882096069869,
+ "grad_norm": 0.013049827888607979,
+ "learning_rate": 0.0006,
+ "loss": 4.378549098968506,
+ "step": 5517
+ },
+ {
+ "epoch": 76.64279475982532,
+ "grad_norm": 0.012075880542397499,
+ "learning_rate": 0.0006,
+ "loss": 4.348293304443359,
+ "step": 5518
+ },
+ {
+ "epoch": 76.65676855895197,
+ "grad_norm": 0.014932847581803799,
+ "learning_rate": 0.0006,
+ "loss": 4.282538890838623,
+ "step": 5519
+ },
+ {
+ "epoch": 76.6707423580786,
+ "grad_norm": 0.015615998767316341,
+ "learning_rate": 0.0006,
+ "loss": 4.274641036987305,
+ "step": 5520
+ },
+ {
+ "epoch": 76.68471615720524,
+ "grad_norm": 0.01424513477832079,
+ "learning_rate": 0.0006,
+ "loss": 4.243166923522949,
+ "step": 5521
+ },
+ {
+ "epoch": 76.69868995633188,
+ "grad_norm": 0.012775546871125698,
+ "learning_rate": 0.0006,
+ "loss": 4.335611820220947,
+ "step": 5522
+ },
+ {
+ "epoch": 76.71266375545852,
+ "grad_norm": 0.014461303129792213,
+ "learning_rate": 0.0006,
+ "loss": 4.246976375579834,
+ "step": 5523
+ },
+ {
+ "epoch": 76.72663755458515,
+ "grad_norm": 0.0145436255261302,
+ "learning_rate": 0.0006,
+ "loss": 4.275895118713379,
+ "step": 5524
+ },
+ {
+ "epoch": 76.74061135371178,
+ "grad_norm": 0.014055085368454456,
+ "learning_rate": 0.0006,
+ "loss": 4.331108093261719,
+ "step": 5525
+ },
+ {
+ "epoch": 76.75458515283843,
+ "grad_norm": 0.011933894827961922,
+ "learning_rate": 0.0006,
+ "loss": 4.293721675872803,
+ "step": 5526
+ },
+ {
+ "epoch": 76.76855895196506,
+ "grad_norm": 0.01277392078191042,
+ "learning_rate": 0.0006,
+ "loss": 4.312374114990234,
+ "step": 5527
+ },
+ {
+ "epoch": 76.78253275109171,
+ "grad_norm": 0.014227936044335365,
+ "learning_rate": 0.0006,
+ "loss": 4.238729000091553,
+ "step": 5528
+ },
+ {
+ "epoch": 76.79650655021834,
+ "grad_norm": 0.01488024927675724,
+ "learning_rate": 0.0006,
+ "loss": 4.24587345123291,
+ "step": 5529
+ },
+ {
+ "epoch": 76.81048034934499,
+ "grad_norm": 0.015341250225901604,
+ "learning_rate": 0.0006,
+ "loss": 4.257378578186035,
+ "step": 5530
+ },
+ {
+ "epoch": 76.82445414847162,
+ "grad_norm": 0.015179437585175037,
+ "learning_rate": 0.0006,
+ "loss": 4.2812910079956055,
+ "step": 5531
+ },
+ {
+ "epoch": 76.83842794759825,
+ "grad_norm": 0.012814724817872047,
+ "learning_rate": 0.0006,
+ "loss": 4.2309465408325195,
+ "step": 5532
+ },
+ {
+ "epoch": 76.8524017467249,
+ "grad_norm": 0.013810716569423676,
+ "learning_rate": 0.0006,
+ "loss": 4.325292110443115,
+ "step": 5533
+ },
+ {
+ "epoch": 76.86637554585153,
+ "grad_norm": 0.017468232661485672,
+ "learning_rate": 0.0006,
+ "loss": 4.243788719177246,
+ "step": 5534
+ },
+ {
+ "epoch": 76.88034934497817,
+ "grad_norm": 0.01739857904613018,
+ "learning_rate": 0.0006,
+ "loss": 4.345390796661377,
+ "step": 5535
+ },
+ {
+ "epoch": 76.8943231441048,
+ "grad_norm": 0.015329061076045036,
+ "learning_rate": 0.0006,
+ "loss": 4.415142059326172,
+ "step": 5536
+ },
+ {
+ "epoch": 76.90829694323143,
+ "grad_norm": 0.01598426327109337,
+ "learning_rate": 0.0006,
+ "loss": 4.299558639526367,
+ "step": 5537
+ },
+ {
+ "epoch": 76.92227074235808,
+ "grad_norm": 0.01621454581618309,
+ "learning_rate": 0.0006,
+ "loss": 4.284354209899902,
+ "step": 5538
+ },
+ {
+ "epoch": 76.93624454148471,
+ "grad_norm": 0.015956968069076538,
+ "learning_rate": 0.0006,
+ "loss": 4.293837547302246,
+ "step": 5539
+ },
+ {
+ "epoch": 76.95021834061136,
+ "grad_norm": 0.01776493526995182,
+ "learning_rate": 0.0006,
+ "loss": 4.229729175567627,
+ "step": 5540
+ },
+ {
+ "epoch": 76.96419213973799,
+ "grad_norm": 0.015317477285861969,
+ "learning_rate": 0.0006,
+ "loss": 4.3328447341918945,
+ "step": 5541
+ },
+ {
+ "epoch": 76.97816593886463,
+ "grad_norm": 0.015140251256525517,
+ "learning_rate": 0.0006,
+ "loss": 4.310840606689453,
+ "step": 5542
+ },
+ {
+ "epoch": 76.99213973799127,
+ "grad_norm": 0.015326812863349915,
+ "learning_rate": 0.0006,
+ "loss": 4.230832099914551,
+ "step": 5543
+ },
+ {
+ "epoch": 77.0,
+ "grad_norm": 0.01607763208448887,
+ "learning_rate": 0.0006,
+ "loss": 4.2236328125,
+ "step": 5544
+ },
+ {
+ "epoch": 77.0,
+ "eval_loss": 4.662221431732178,
+ "eval_runtime": 56.5194,
+ "eval_samples_per_second": 43.206,
+ "eval_steps_per_second": 1.362,
+ "step": 5544
+ },
+ {
+ "epoch": 77.01397379912663,
+ "grad_norm": 0.014839510433375835,
+ "learning_rate": 0.0006,
+ "loss": 4.254111289978027,
+ "step": 5545
+ },
+ {
+ "epoch": 77.02794759825328,
+ "grad_norm": 0.01519741676747799,
+ "learning_rate": 0.0006,
+ "loss": 4.1607184410095215,
+ "step": 5546
+ },
+ {
+ "epoch": 77.04192139737991,
+ "grad_norm": 0.015472603030502796,
+ "learning_rate": 0.0006,
+ "loss": 4.323320388793945,
+ "step": 5547
+ },
+ {
+ "epoch": 77.05589519650655,
+ "grad_norm": 0.017937948927283287,
+ "learning_rate": 0.0006,
+ "loss": 4.308490753173828,
+ "step": 5548
+ },
+ {
+ "epoch": 77.06986899563319,
+ "grad_norm": 0.020890841260552406,
+ "learning_rate": 0.0006,
+ "loss": 4.259101867675781,
+ "step": 5549
+ },
+ {
+ "epoch": 77.08384279475983,
+ "grad_norm": 0.01892857998609543,
+ "learning_rate": 0.0006,
+ "loss": 4.386970520019531,
+ "step": 5550
+ },
+ {
+ "epoch": 77.09781659388646,
+ "grad_norm": 0.016583630815148354,
+ "learning_rate": 0.0006,
+ "loss": 4.233604907989502,
+ "step": 5551
+ },
+ {
+ "epoch": 77.1117903930131,
+ "grad_norm": 0.016063468530774117,
+ "learning_rate": 0.0006,
+ "loss": 4.339544296264648,
+ "step": 5552
+ },
+ {
+ "epoch": 77.12576419213974,
+ "grad_norm": 0.018326779827475548,
+ "learning_rate": 0.0006,
+ "loss": 4.367204666137695,
+ "step": 5553
+ },
+ {
+ "epoch": 77.13973799126637,
+ "grad_norm": 0.021590933203697205,
+ "learning_rate": 0.0006,
+ "loss": 4.352685928344727,
+ "step": 5554
+ },
+ {
+ "epoch": 77.15371179039302,
+ "grad_norm": 0.019643642008304596,
+ "learning_rate": 0.0006,
+ "loss": 4.258535385131836,
+ "step": 5555
+ },
+ {
+ "epoch": 77.16768558951965,
+ "grad_norm": 0.018476149067282677,
+ "learning_rate": 0.0006,
+ "loss": 4.232568740844727,
+ "step": 5556
+ },
+ {
+ "epoch": 77.18165938864628,
+ "grad_norm": 0.01584405079483986,
+ "learning_rate": 0.0006,
+ "loss": 4.258371353149414,
+ "step": 5557
+ },
+ {
+ "epoch": 77.19563318777293,
+ "grad_norm": 0.01575201191008091,
+ "learning_rate": 0.0006,
+ "loss": 4.183218002319336,
+ "step": 5558
+ },
+ {
+ "epoch": 77.20960698689956,
+ "grad_norm": 0.016871824860572815,
+ "learning_rate": 0.0006,
+ "loss": 4.194563865661621,
+ "step": 5559
+ },
+ {
+ "epoch": 77.2235807860262,
+ "grad_norm": 0.02059333398938179,
+ "learning_rate": 0.0006,
+ "loss": 4.195281982421875,
+ "step": 5560
+ },
+ {
+ "epoch": 77.23755458515284,
+ "grad_norm": 0.022640453651547432,
+ "learning_rate": 0.0006,
+ "loss": 4.279615879058838,
+ "step": 5561
+ },
+ {
+ "epoch": 77.25152838427948,
+ "grad_norm": 0.020335959270596504,
+ "learning_rate": 0.0006,
+ "loss": 4.25040864944458,
+ "step": 5562
+ },
+ {
+ "epoch": 77.26550218340611,
+ "grad_norm": 0.020010793581604958,
+ "learning_rate": 0.0006,
+ "loss": 4.291116714477539,
+ "step": 5563
+ },
+ {
+ "epoch": 77.27947598253274,
+ "grad_norm": 0.02109045907855034,
+ "learning_rate": 0.0006,
+ "loss": 4.364088535308838,
+ "step": 5564
+ },
+ {
+ "epoch": 77.29344978165939,
+ "grad_norm": 0.019519446417689323,
+ "learning_rate": 0.0006,
+ "loss": 4.329166412353516,
+ "step": 5565
+ },
+ {
+ "epoch": 77.30742358078602,
+ "grad_norm": 0.01689804159104824,
+ "learning_rate": 0.0006,
+ "loss": 4.244176864624023,
+ "step": 5566
+ },
+ {
+ "epoch": 77.32139737991267,
+ "grad_norm": 0.01893681474030018,
+ "learning_rate": 0.0006,
+ "loss": 4.239068984985352,
+ "step": 5567
+ },
+ {
+ "epoch": 77.3353711790393,
+ "grad_norm": 0.015998193994164467,
+ "learning_rate": 0.0006,
+ "loss": 4.173398017883301,
+ "step": 5568
+ },
+ {
+ "epoch": 77.34934497816595,
+ "grad_norm": 0.013892280869185925,
+ "learning_rate": 0.0006,
+ "loss": 4.287654876708984,
+ "step": 5569
+ },
+ {
+ "epoch": 77.36331877729258,
+ "grad_norm": 0.017060643061995506,
+ "learning_rate": 0.0006,
+ "loss": 4.398447036743164,
+ "step": 5570
+ },
+ {
+ "epoch": 77.37729257641921,
+ "grad_norm": 0.015168849378824234,
+ "learning_rate": 0.0006,
+ "loss": 4.15665864944458,
+ "step": 5571
+ },
+ {
+ "epoch": 77.39126637554585,
+ "grad_norm": 0.014886523596942425,
+ "learning_rate": 0.0006,
+ "loss": 4.196673393249512,
+ "step": 5572
+ },
+ {
+ "epoch": 77.40524017467249,
+ "grad_norm": 0.016941796988248825,
+ "learning_rate": 0.0006,
+ "loss": 4.314562797546387,
+ "step": 5573
+ },
+ {
+ "epoch": 77.41921397379913,
+ "grad_norm": 0.015353812836110592,
+ "learning_rate": 0.0006,
+ "loss": 4.409579277038574,
+ "step": 5574
+ },
+ {
+ "epoch": 77.43318777292576,
+ "grad_norm": 0.015192852355539799,
+ "learning_rate": 0.0006,
+ "loss": 4.392045974731445,
+ "step": 5575
+ },
+ {
+ "epoch": 77.44716157205241,
+ "grad_norm": 0.014232569374144077,
+ "learning_rate": 0.0006,
+ "loss": 4.21639347076416,
+ "step": 5576
+ },
+ {
+ "epoch": 77.46113537117904,
+ "grad_norm": 0.014453226700425148,
+ "learning_rate": 0.0006,
+ "loss": 4.2802276611328125,
+ "step": 5577
+ },
+ {
+ "epoch": 77.47510917030567,
+ "grad_norm": 0.012881236150860786,
+ "learning_rate": 0.0006,
+ "loss": 4.375980854034424,
+ "step": 5578
+ },
+ {
+ "epoch": 77.48908296943232,
+ "grad_norm": 0.01583976298570633,
+ "learning_rate": 0.0006,
+ "loss": 4.264520645141602,
+ "step": 5579
+ },
+ {
+ "epoch": 77.50305676855895,
+ "grad_norm": 0.014171584509313107,
+ "learning_rate": 0.0006,
+ "loss": 4.225279808044434,
+ "step": 5580
+ },
+ {
+ "epoch": 77.5170305676856,
+ "grad_norm": 0.014715159311890602,
+ "learning_rate": 0.0006,
+ "loss": 4.369834899902344,
+ "step": 5581
+ },
+ {
+ "epoch": 77.53100436681223,
+ "grad_norm": 0.016554079949855804,
+ "learning_rate": 0.0006,
+ "loss": 4.2976179122924805,
+ "step": 5582
+ },
+ {
+ "epoch": 77.54497816593886,
+ "grad_norm": 0.015095747075974941,
+ "learning_rate": 0.0006,
+ "loss": 4.294407844543457,
+ "step": 5583
+ },
+ {
+ "epoch": 77.5589519650655,
+ "grad_norm": 0.012726361863315105,
+ "learning_rate": 0.0006,
+ "loss": 4.221401214599609,
+ "step": 5584
+ },
+ {
+ "epoch": 77.57292576419214,
+ "grad_norm": 0.014999309554696083,
+ "learning_rate": 0.0006,
+ "loss": 4.243929386138916,
+ "step": 5585
+ },
+ {
+ "epoch": 77.58689956331878,
+ "grad_norm": 0.016526037827134132,
+ "learning_rate": 0.0006,
+ "loss": 4.374139785766602,
+ "step": 5586
+ },
+ {
+ "epoch": 77.60087336244541,
+ "grad_norm": 0.01648339442908764,
+ "learning_rate": 0.0006,
+ "loss": 4.392425060272217,
+ "step": 5587
+ },
+ {
+ "epoch": 77.61484716157206,
+ "grad_norm": 0.015269213356077671,
+ "learning_rate": 0.0006,
+ "loss": 4.33986759185791,
+ "step": 5588
+ },
+ {
+ "epoch": 77.62882096069869,
+ "grad_norm": 0.01690507121384144,
+ "learning_rate": 0.0006,
+ "loss": 4.307476043701172,
+ "step": 5589
+ },
+ {
+ "epoch": 77.64279475982532,
+ "grad_norm": 0.016438685357570648,
+ "learning_rate": 0.0006,
+ "loss": 4.2406206130981445,
+ "step": 5590
+ },
+ {
+ "epoch": 77.65676855895197,
+ "grad_norm": 0.0157600287348032,
+ "learning_rate": 0.0006,
+ "loss": 4.314070224761963,
+ "step": 5591
+ },
+ {
+ "epoch": 77.6707423580786,
+ "grad_norm": 0.015127947553992271,
+ "learning_rate": 0.0006,
+ "loss": 4.22475528717041,
+ "step": 5592
+ },
+ {
+ "epoch": 77.68471615720524,
+ "grad_norm": 0.016632677987217903,
+ "learning_rate": 0.0006,
+ "loss": 4.26954984664917,
+ "step": 5593
+ },
+ {
+ "epoch": 77.69868995633188,
+ "grad_norm": 0.013150460086762905,
+ "learning_rate": 0.0006,
+ "loss": 4.369131088256836,
+ "step": 5594
+ },
+ {
+ "epoch": 77.71266375545852,
+ "grad_norm": 0.014513005502521992,
+ "learning_rate": 0.0006,
+ "loss": 4.3248796463012695,
+ "step": 5595
+ },
+ {
+ "epoch": 77.72663755458515,
+ "grad_norm": 0.014090991578996181,
+ "learning_rate": 0.0006,
+ "loss": 4.268381118774414,
+ "step": 5596
+ },
+ {
+ "epoch": 77.74061135371178,
+ "grad_norm": 0.01633097417652607,
+ "learning_rate": 0.0006,
+ "loss": 4.238439083099365,
+ "step": 5597
+ },
+ {
+ "epoch": 77.75458515283843,
+ "grad_norm": 0.01710933819413185,
+ "learning_rate": 0.0006,
+ "loss": 4.399697780609131,
+ "step": 5598
+ },
+ {
+ "epoch": 77.76855895196506,
+ "grad_norm": 0.019251767545938492,
+ "learning_rate": 0.0006,
+ "loss": 4.290450096130371,
+ "step": 5599
+ },
+ {
+ "epoch": 77.78253275109171,
+ "grad_norm": 0.01849171705543995,
+ "learning_rate": 0.0006,
+ "loss": 4.267734527587891,
+ "step": 5600
+ },
+ {
+ "epoch": 77.79650655021834,
+ "grad_norm": 0.014173115603625774,
+ "learning_rate": 0.0006,
+ "loss": 4.288799285888672,
+ "step": 5601
+ },
+ {
+ "epoch": 77.81048034934499,
+ "grad_norm": 0.014572718180716038,
+ "learning_rate": 0.0006,
+ "loss": 4.3790435791015625,
+ "step": 5602
+ },
+ {
+ "epoch": 77.82445414847162,
+ "grad_norm": 0.016375336796045303,
+ "learning_rate": 0.0006,
+ "loss": 4.231725692749023,
+ "step": 5603
+ },
+ {
+ "epoch": 77.83842794759825,
+ "grad_norm": 0.017326701432466507,
+ "learning_rate": 0.0006,
+ "loss": 4.211495399475098,
+ "step": 5604
+ },
+ {
+ "epoch": 77.8524017467249,
+ "grad_norm": 0.016399390995502472,
+ "learning_rate": 0.0006,
+ "loss": 4.386699199676514,
+ "step": 5605
+ },
+ {
+ "epoch": 77.86637554585153,
+ "grad_norm": 0.018269989639520645,
+ "learning_rate": 0.0006,
+ "loss": 4.254856109619141,
+ "step": 5606
+ },
+ {
+ "epoch": 77.88034934497817,
+ "grad_norm": 0.018443435430526733,
+ "learning_rate": 0.0006,
+ "loss": 4.391961097717285,
+ "step": 5607
+ },
+ {
+ "epoch": 77.8943231441048,
+ "grad_norm": 0.01784636080265045,
+ "learning_rate": 0.0006,
+ "loss": 4.273881912231445,
+ "step": 5608
+ },
+ {
+ "epoch": 77.90829694323143,
+ "grad_norm": 0.01782994158565998,
+ "learning_rate": 0.0006,
+ "loss": 4.26833963394165,
+ "step": 5609
+ },
+ {
+ "epoch": 77.92227074235808,
+ "grad_norm": 0.016260741278529167,
+ "learning_rate": 0.0006,
+ "loss": 4.306807994842529,
+ "step": 5610
+ },
+ {
+ "epoch": 77.93624454148471,
+ "grad_norm": 0.015482903458178043,
+ "learning_rate": 0.0006,
+ "loss": 4.187826633453369,
+ "step": 5611
+ },
+ {
+ "epoch": 77.95021834061136,
+ "grad_norm": 0.015193257480859756,
+ "learning_rate": 0.0006,
+ "loss": 4.286060333251953,
+ "step": 5612
+ },
+ {
+ "epoch": 77.96419213973799,
+ "grad_norm": 0.015110115520656109,
+ "learning_rate": 0.0006,
+ "loss": 4.2828192710876465,
+ "step": 5613
+ },
+ {
+ "epoch": 77.97816593886463,
+ "grad_norm": 0.015973486006259918,
+ "learning_rate": 0.0006,
+ "loss": 4.317939758300781,
+ "step": 5614
+ },
+ {
+ "epoch": 77.99213973799127,
+ "grad_norm": 0.017363594844937325,
+ "learning_rate": 0.0006,
+ "loss": 4.287721633911133,
+ "step": 5615
+ },
+ {
+ "epoch": 78.0,
+ "grad_norm": 0.015277642756700516,
+ "learning_rate": 0.0006,
+ "loss": 4.356618881225586,
+ "step": 5616
+ },
+ {
+ "epoch": 78.0,
+ "eval_loss": 4.642678737640381,
+ "eval_runtime": 56.7335,
+ "eval_samples_per_second": 43.043,
+ "eval_steps_per_second": 1.357,
+ "step": 5616
+ },
+ {
+ "epoch": 78.01397379912663,
+ "grad_norm": 0.016794051975011826,
+ "learning_rate": 0.0006,
+ "loss": 4.279930591583252,
+ "step": 5617
+ },
+ {
+ "epoch": 78.02794759825328,
+ "grad_norm": 0.01632673107087612,
+ "learning_rate": 0.0006,
+ "loss": 4.235943794250488,
+ "step": 5618
+ },
+ {
+ "epoch": 78.04192139737991,
+ "grad_norm": 0.01437876746058464,
+ "learning_rate": 0.0006,
+ "loss": 4.2542243003845215,
+ "step": 5619
+ },
+ {
+ "epoch": 78.05589519650655,
+ "grad_norm": 0.014993119053542614,
+ "learning_rate": 0.0006,
+ "loss": 4.262046813964844,
+ "step": 5620
+ },
+ {
+ "epoch": 78.06986899563319,
+ "grad_norm": 0.015721119940280914,
+ "learning_rate": 0.0006,
+ "loss": 4.173436164855957,
+ "step": 5621
+ },
+ {
+ "epoch": 78.08384279475983,
+ "grad_norm": 0.01334085687994957,
+ "learning_rate": 0.0006,
+ "loss": 4.322175979614258,
+ "step": 5622
+ },
+ {
+ "epoch": 78.09781659388646,
+ "grad_norm": 0.015561388805508614,
+ "learning_rate": 0.0006,
+ "loss": 4.242125988006592,
+ "step": 5623
+ },
+ {
+ "epoch": 78.1117903930131,
+ "grad_norm": 0.014418602921068668,
+ "learning_rate": 0.0006,
+ "loss": 4.266186714172363,
+ "step": 5624
+ },
+ {
+ "epoch": 78.12576419213974,
+ "grad_norm": 0.013739909045398235,
+ "learning_rate": 0.0006,
+ "loss": 4.231633186340332,
+ "step": 5625
+ },
+ {
+ "epoch": 78.13973799126637,
+ "grad_norm": 0.01402067206799984,
+ "learning_rate": 0.0006,
+ "loss": 4.188406944274902,
+ "step": 5626
+ },
+ {
+ "epoch": 78.15371179039302,
+ "grad_norm": 0.013639481738209724,
+ "learning_rate": 0.0006,
+ "loss": 4.1586594581604,
+ "step": 5627
+ },
+ {
+ "epoch": 78.16768558951965,
+ "grad_norm": 0.012858159840106964,
+ "learning_rate": 0.0006,
+ "loss": 4.360448360443115,
+ "step": 5628
+ },
+ {
+ "epoch": 78.18165938864628,
+ "grad_norm": 0.013089739717543125,
+ "learning_rate": 0.0006,
+ "loss": 4.344371795654297,
+ "step": 5629
+ },
+ {
+ "epoch": 78.19563318777293,
+ "grad_norm": 0.01382314320653677,
+ "learning_rate": 0.0006,
+ "loss": 4.20378303527832,
+ "step": 5630
+ },
+ {
+ "epoch": 78.20960698689956,
+ "grad_norm": 0.014206619001924992,
+ "learning_rate": 0.0006,
+ "loss": 4.239891529083252,
+ "step": 5631
+ },
+ {
+ "epoch": 78.2235807860262,
+ "grad_norm": 0.01472168043255806,
+ "learning_rate": 0.0006,
+ "loss": 4.282382488250732,
+ "step": 5632
+ },
+ {
+ "epoch": 78.23755458515284,
+ "grad_norm": 0.014237025752663612,
+ "learning_rate": 0.0006,
+ "loss": 4.241545677185059,
+ "step": 5633
+ },
+ {
+ "epoch": 78.25152838427948,
+ "grad_norm": 0.013231472112238407,
+ "learning_rate": 0.0006,
+ "loss": 4.184896469116211,
+ "step": 5634
+ },
+ {
+ "epoch": 78.26550218340611,
+ "grad_norm": 0.013538471423089504,
+ "learning_rate": 0.0006,
+ "loss": 4.325181007385254,
+ "step": 5635
+ },
+ {
+ "epoch": 78.27947598253274,
+ "grad_norm": 0.013541961088776588,
+ "learning_rate": 0.0006,
+ "loss": 4.368722915649414,
+ "step": 5636
+ },
+ {
+ "epoch": 78.29344978165939,
+ "grad_norm": 0.013446723110973835,
+ "learning_rate": 0.0006,
+ "loss": 4.301422595977783,
+ "step": 5637
+ },
+ {
+ "epoch": 78.30742358078602,
+ "grad_norm": 0.014753809198737144,
+ "learning_rate": 0.0006,
+ "loss": 4.259127616882324,
+ "step": 5638
+ },
+ {
+ "epoch": 78.32139737991267,
+ "grad_norm": 0.013256008736789227,
+ "learning_rate": 0.0006,
+ "loss": 4.284032821655273,
+ "step": 5639
+ },
+ {
+ "epoch": 78.3353711790393,
+ "grad_norm": 0.014272629283368587,
+ "learning_rate": 0.0006,
+ "loss": 4.234989166259766,
+ "step": 5640
+ },
+ {
+ "epoch": 78.34934497816595,
+ "grad_norm": 0.014756826683878899,
+ "learning_rate": 0.0006,
+ "loss": 4.143499374389648,
+ "step": 5641
+ },
+ {
+ "epoch": 78.36331877729258,
+ "grad_norm": 0.016549205407500267,
+ "learning_rate": 0.0006,
+ "loss": 4.311034202575684,
+ "step": 5642
+ },
+ {
+ "epoch": 78.37729257641921,
+ "grad_norm": 0.017735283821821213,
+ "learning_rate": 0.0006,
+ "loss": 4.246767520904541,
+ "step": 5643
+ },
+ {
+ "epoch": 78.39126637554585,
+ "grad_norm": 0.017370613291859627,
+ "learning_rate": 0.0006,
+ "loss": 4.2460832595825195,
+ "step": 5644
+ },
+ {
+ "epoch": 78.40524017467249,
+ "grad_norm": 0.014642846770584583,
+ "learning_rate": 0.0006,
+ "loss": 4.303811550140381,
+ "step": 5645
+ },
+ {
+ "epoch": 78.41921397379913,
+ "grad_norm": 0.013445726595818996,
+ "learning_rate": 0.0006,
+ "loss": 4.25921630859375,
+ "step": 5646
+ },
+ {
+ "epoch": 78.43318777292576,
+ "grad_norm": 0.014254065230488777,
+ "learning_rate": 0.0006,
+ "loss": 4.129446506500244,
+ "step": 5647
+ },
+ {
+ "epoch": 78.44716157205241,
+ "grad_norm": 0.01883019506931305,
+ "learning_rate": 0.0006,
+ "loss": 4.191716194152832,
+ "step": 5648
+ },
+ {
+ "epoch": 78.46113537117904,
+ "grad_norm": 0.0204437468200922,
+ "learning_rate": 0.0006,
+ "loss": 4.333085536956787,
+ "step": 5649
+ },
+ {
+ "epoch": 78.47510917030567,
+ "grad_norm": 0.01652318798005581,
+ "learning_rate": 0.0006,
+ "loss": 4.135462760925293,
+ "step": 5650
+ },
+ {
+ "epoch": 78.48908296943232,
+ "grad_norm": 0.012753551825881004,
+ "learning_rate": 0.0006,
+ "loss": 4.195051193237305,
+ "step": 5651
+ },
+ {
+ "epoch": 78.50305676855895,
+ "grad_norm": 0.01499118935316801,
+ "learning_rate": 0.0006,
+ "loss": 4.157092094421387,
+ "step": 5652
+ },
+ {
+ "epoch": 78.5170305676856,
+ "grad_norm": 0.015383945778012276,
+ "learning_rate": 0.0006,
+ "loss": 4.253818035125732,
+ "step": 5653
+ },
+ {
+ "epoch": 78.53100436681223,
+ "grad_norm": 0.015721332281827927,
+ "learning_rate": 0.0006,
+ "loss": 4.228775978088379,
+ "step": 5654
+ },
+ {
+ "epoch": 78.54497816593886,
+ "grad_norm": 0.014595463871955872,
+ "learning_rate": 0.0006,
+ "loss": 4.135197639465332,
+ "step": 5655
+ },
+ {
+ "epoch": 78.5589519650655,
+ "grad_norm": 0.015059644356369972,
+ "learning_rate": 0.0006,
+ "loss": 4.255614280700684,
+ "step": 5656
+ },
+ {
+ "epoch": 78.57292576419214,
+ "grad_norm": 0.013833344914019108,
+ "learning_rate": 0.0006,
+ "loss": 4.301250457763672,
+ "step": 5657
+ },
+ {
+ "epoch": 78.58689956331878,
+ "grad_norm": 0.013665376231074333,
+ "learning_rate": 0.0006,
+ "loss": 4.1734232902526855,
+ "step": 5658
+ },
+ {
+ "epoch": 78.60087336244541,
+ "grad_norm": 0.01539086177945137,
+ "learning_rate": 0.0006,
+ "loss": 4.221459865570068,
+ "step": 5659
+ },
+ {
+ "epoch": 78.61484716157206,
+ "grad_norm": 0.013313917443156242,
+ "learning_rate": 0.0006,
+ "loss": 4.255619049072266,
+ "step": 5660
+ },
+ {
+ "epoch": 78.62882096069869,
+ "grad_norm": 0.01447741687297821,
+ "learning_rate": 0.0006,
+ "loss": 4.298219680786133,
+ "step": 5661
+ },
+ {
+ "epoch": 78.64279475982532,
+ "grad_norm": 0.012796915136277676,
+ "learning_rate": 0.0006,
+ "loss": 4.2542405128479,
+ "step": 5662
+ },
+ {
+ "epoch": 78.65676855895197,
+ "grad_norm": 0.014058141969144344,
+ "learning_rate": 0.0006,
+ "loss": 4.284477233886719,
+ "step": 5663
+ },
+ {
+ "epoch": 78.6707423580786,
+ "grad_norm": 0.012705069966614246,
+ "learning_rate": 0.0006,
+ "loss": 4.2826666831970215,
+ "step": 5664
+ },
+ {
+ "epoch": 78.68471615720524,
+ "grad_norm": 0.013245890848338604,
+ "learning_rate": 0.0006,
+ "loss": 4.371612548828125,
+ "step": 5665
+ },
+ {
+ "epoch": 78.69868995633188,
+ "grad_norm": 0.014192677102982998,
+ "learning_rate": 0.0006,
+ "loss": 4.267344951629639,
+ "step": 5666
+ },
+ {
+ "epoch": 78.71266375545852,
+ "grad_norm": 0.015346777625381947,
+ "learning_rate": 0.0006,
+ "loss": 4.222797393798828,
+ "step": 5667
+ },
+ {
+ "epoch": 78.72663755458515,
+ "grad_norm": 0.014068621210753918,
+ "learning_rate": 0.0006,
+ "loss": 4.172964572906494,
+ "step": 5668
+ },
+ {
+ "epoch": 78.74061135371178,
+ "grad_norm": 0.014364945702254772,
+ "learning_rate": 0.0006,
+ "loss": 4.259319305419922,
+ "step": 5669
+ },
+ {
+ "epoch": 78.75458515283843,
+ "grad_norm": 0.017337387427687645,
+ "learning_rate": 0.0006,
+ "loss": 4.29533576965332,
+ "step": 5670
+ },
+ {
+ "epoch": 78.76855895196506,
+ "grad_norm": 0.017416715621948242,
+ "learning_rate": 0.0006,
+ "loss": 4.340364456176758,
+ "step": 5671
+ },
+ {
+ "epoch": 78.78253275109171,
+ "grad_norm": 0.015925846993923187,
+ "learning_rate": 0.0006,
+ "loss": 4.177454948425293,
+ "step": 5672
+ },
+ {
+ "epoch": 78.79650655021834,
+ "grad_norm": 0.013130443170666695,
+ "learning_rate": 0.0006,
+ "loss": 4.277454853057861,
+ "step": 5673
+ },
+ {
+ "epoch": 78.81048034934499,
+ "grad_norm": 0.014628687873482704,
+ "learning_rate": 0.0006,
+ "loss": 4.273757457733154,
+ "step": 5674
+ },
+ {
+ "epoch": 78.82445414847162,
+ "grad_norm": 0.01601659320294857,
+ "learning_rate": 0.0006,
+ "loss": 4.281774520874023,
+ "step": 5675
+ },
+ {
+ "epoch": 78.83842794759825,
+ "grad_norm": 0.015443485230207443,
+ "learning_rate": 0.0006,
+ "loss": 4.199937343597412,
+ "step": 5676
+ },
+ {
+ "epoch": 78.8524017467249,
+ "grad_norm": 0.013934246264398098,
+ "learning_rate": 0.0006,
+ "loss": 4.331369876861572,
+ "step": 5677
+ },
+ {
+ "epoch": 78.86637554585153,
+ "grad_norm": 0.01572083681821823,
+ "learning_rate": 0.0006,
+ "loss": 4.2828474044799805,
+ "step": 5678
+ },
+ {
+ "epoch": 78.88034934497817,
+ "grad_norm": 0.013444120064377785,
+ "learning_rate": 0.0006,
+ "loss": 4.183859825134277,
+ "step": 5679
+ },
+ {
+ "epoch": 78.8943231441048,
+ "grad_norm": 0.01591254211962223,
+ "learning_rate": 0.0006,
+ "loss": 4.295256614685059,
+ "step": 5680
+ },
+ {
+ "epoch": 78.90829694323143,
+ "grad_norm": 0.017556684091687202,
+ "learning_rate": 0.0006,
+ "loss": 4.132272720336914,
+ "step": 5681
+ },
+ {
+ "epoch": 78.92227074235808,
+ "grad_norm": 0.018184322863817215,
+ "learning_rate": 0.0006,
+ "loss": 4.251728057861328,
+ "step": 5682
+ },
+ {
+ "epoch": 78.93624454148471,
+ "grad_norm": 0.016207285225391388,
+ "learning_rate": 0.0006,
+ "loss": 4.205544471740723,
+ "step": 5683
+ },
+ {
+ "epoch": 78.95021834061136,
+ "grad_norm": 0.014478376135230064,
+ "learning_rate": 0.0006,
+ "loss": 4.182462215423584,
+ "step": 5684
+ },
+ {
+ "epoch": 78.96419213973799,
+ "grad_norm": 0.014255235902965069,
+ "learning_rate": 0.0006,
+ "loss": 4.253366947174072,
+ "step": 5685
+ },
+ {
+ "epoch": 78.97816593886463,
+ "grad_norm": 0.01524396613240242,
+ "learning_rate": 0.0006,
+ "loss": 4.221786022186279,
+ "step": 5686
+ },
+ {
+ "epoch": 78.99213973799127,
+ "grad_norm": 0.015057693235576153,
+ "learning_rate": 0.0006,
+ "loss": 4.208454608917236,
+ "step": 5687
+ },
+ {
+ "epoch": 79.0,
+ "grad_norm": 0.015478022396564484,
+ "learning_rate": 0.0006,
+ "loss": 4.336383819580078,
+ "step": 5688
+ },
+ {
+ "epoch": 79.0,
+ "eval_loss": 4.641749382019043,
+ "eval_runtime": 56.8167,
+ "eval_samples_per_second": 42.98,
+ "eval_steps_per_second": 1.355,
+ "step": 5688
+ },
+ {
+ "epoch": 79.01397379912663,
+ "grad_norm": 0.014143316075205803,
+ "learning_rate": 0.0006,
+ "loss": 4.260974884033203,
+ "step": 5689
+ },
+ {
+ "epoch": 79.02794759825328,
+ "grad_norm": 0.013120883144438267,
+ "learning_rate": 0.0006,
+ "loss": 4.139886856079102,
+ "step": 5690
+ },
+ {
+ "epoch": 79.04192139737991,
+ "grad_norm": 0.015019661746919155,
+ "learning_rate": 0.0006,
+ "loss": 4.267004013061523,
+ "step": 5691
+ },
+ {
+ "epoch": 79.05589519650655,
+ "grad_norm": 0.014891284517943859,
+ "learning_rate": 0.0006,
+ "loss": 4.205019950866699,
+ "step": 5692
+ },
+ {
+ "epoch": 79.06986899563319,
+ "grad_norm": 0.015527728945016861,
+ "learning_rate": 0.0006,
+ "loss": 4.1513566970825195,
+ "step": 5693
+ },
+ {
+ "epoch": 79.08384279475983,
+ "grad_norm": 0.017527181655168533,
+ "learning_rate": 0.0006,
+ "loss": 4.148828983306885,
+ "step": 5694
+ },
+ {
+ "epoch": 79.09781659388646,
+ "grad_norm": 0.01632206328213215,
+ "learning_rate": 0.0006,
+ "loss": 4.195789813995361,
+ "step": 5695
+ },
+ {
+ "epoch": 79.1117903930131,
+ "grad_norm": 0.01590503193438053,
+ "learning_rate": 0.0006,
+ "loss": 4.199155330657959,
+ "step": 5696
+ },
+ {
+ "epoch": 79.12576419213974,
+ "grad_norm": 0.016489997506141663,
+ "learning_rate": 0.0006,
+ "loss": 4.155154705047607,
+ "step": 5697
+ },
+ {
+ "epoch": 79.13973799126637,
+ "grad_norm": 0.0161368977278471,
+ "learning_rate": 0.0006,
+ "loss": 4.164384841918945,
+ "step": 5698
+ },
+ {
+ "epoch": 79.15371179039302,
+ "grad_norm": 0.018574809655547142,
+ "learning_rate": 0.0006,
+ "loss": 4.1898193359375,
+ "step": 5699
+ },
+ {
+ "epoch": 79.16768558951965,
+ "grad_norm": 0.018987983465194702,
+ "learning_rate": 0.0006,
+ "loss": 4.219719409942627,
+ "step": 5700
+ },
+ {
+ "epoch": 79.18165938864628,
+ "grad_norm": 0.01668606884777546,
+ "learning_rate": 0.0006,
+ "loss": 4.152734756469727,
+ "step": 5701
+ },
+ {
+ "epoch": 79.19563318777293,
+ "grad_norm": 0.014046299271285534,
+ "learning_rate": 0.0006,
+ "loss": 4.152964115142822,
+ "step": 5702
+ },
+ {
+ "epoch": 79.20960698689956,
+ "grad_norm": 0.01628032699227333,
+ "learning_rate": 0.0006,
+ "loss": 4.299479961395264,
+ "step": 5703
+ },
+ {
+ "epoch": 79.2235807860262,
+ "grad_norm": 0.017998795956373215,
+ "learning_rate": 0.0006,
+ "loss": 4.231203079223633,
+ "step": 5704
+ },
+ {
+ "epoch": 79.23755458515284,
+ "grad_norm": 0.014799647964537144,
+ "learning_rate": 0.0006,
+ "loss": 4.3530192375183105,
+ "step": 5705
+ },
+ {
+ "epoch": 79.25152838427948,
+ "grad_norm": 0.014640678651630878,
+ "learning_rate": 0.0006,
+ "loss": 4.287322998046875,
+ "step": 5706
+ },
+ {
+ "epoch": 79.26550218340611,
+ "grad_norm": 0.0172616895288229,
+ "learning_rate": 0.0006,
+ "loss": 4.148009777069092,
+ "step": 5707
+ },
+ {
+ "epoch": 79.27947598253274,
+ "grad_norm": 0.01811790093779564,
+ "learning_rate": 0.0006,
+ "loss": 4.269433975219727,
+ "step": 5708
+ },
+ {
+ "epoch": 79.29344978165939,
+ "grad_norm": 0.01764000579714775,
+ "learning_rate": 0.0006,
+ "loss": 4.292638778686523,
+ "step": 5709
+ },
+ {
+ "epoch": 79.30742358078602,
+ "grad_norm": 0.018030935898423195,
+ "learning_rate": 0.0006,
+ "loss": 4.267889976501465,
+ "step": 5710
+ },
+ {
+ "epoch": 79.32139737991267,
+ "grad_norm": 0.017860891297459602,
+ "learning_rate": 0.0006,
+ "loss": 4.211110591888428,
+ "step": 5711
+ },
+ {
+ "epoch": 79.3353711790393,
+ "grad_norm": 0.017295073717832565,
+ "learning_rate": 0.0006,
+ "loss": 4.224645614624023,
+ "step": 5712
+ },
+ {
+ "epoch": 79.34934497816595,
+ "grad_norm": 0.015891101211309433,
+ "learning_rate": 0.0006,
+ "loss": 4.224797248840332,
+ "step": 5713
+ },
+ {
+ "epoch": 79.36331877729258,
+ "grad_norm": 0.014680417254567146,
+ "learning_rate": 0.0006,
+ "loss": 4.182242393493652,
+ "step": 5714
+ },
+ {
+ "epoch": 79.37729257641921,
+ "grad_norm": 0.016621140763163567,
+ "learning_rate": 0.0006,
+ "loss": 4.186334609985352,
+ "step": 5715
+ },
+ {
+ "epoch": 79.39126637554585,
+ "grad_norm": 0.016207249835133553,
+ "learning_rate": 0.0006,
+ "loss": 4.277074813842773,
+ "step": 5716
+ },
+ {
+ "epoch": 79.40524017467249,
+ "grad_norm": 0.015375478193163872,
+ "learning_rate": 0.0006,
+ "loss": 4.154317855834961,
+ "step": 5717
+ },
+ {
+ "epoch": 79.41921397379913,
+ "grad_norm": 0.013453228399157524,
+ "learning_rate": 0.0006,
+ "loss": 4.265270233154297,
+ "step": 5718
+ },
+ {
+ "epoch": 79.43318777292576,
+ "grad_norm": 0.014312694780528545,
+ "learning_rate": 0.0006,
+ "loss": 4.232220649719238,
+ "step": 5719
+ },
+ {
+ "epoch": 79.44716157205241,
+ "grad_norm": 0.014759554527699947,
+ "learning_rate": 0.0006,
+ "loss": 4.272275924682617,
+ "step": 5720
+ },
+ {
+ "epoch": 79.46113537117904,
+ "grad_norm": 0.015885505825281143,
+ "learning_rate": 0.0006,
+ "loss": 4.2229719161987305,
+ "step": 5721
+ },
+ {
+ "epoch": 79.47510917030567,
+ "grad_norm": 0.016516581177711487,
+ "learning_rate": 0.0006,
+ "loss": 4.200624465942383,
+ "step": 5722
+ },
+ {
+ "epoch": 79.48908296943232,
+ "grad_norm": 0.018254555761814117,
+ "learning_rate": 0.0006,
+ "loss": 4.19648551940918,
+ "step": 5723
+ },
+ {
+ "epoch": 79.50305676855895,
+ "grad_norm": 0.01467090006917715,
+ "learning_rate": 0.0006,
+ "loss": 4.135190486907959,
+ "step": 5724
+ },
+ {
+ "epoch": 79.5170305676856,
+ "grad_norm": 0.013474797829985619,
+ "learning_rate": 0.0006,
+ "loss": 4.311361312866211,
+ "step": 5725
+ },
+ {
+ "epoch": 79.53100436681223,
+ "grad_norm": 0.01742764748632908,
+ "learning_rate": 0.0006,
+ "loss": 4.1845903396606445,
+ "step": 5726
+ },
+ {
+ "epoch": 79.54497816593886,
+ "grad_norm": 0.018786882981657982,
+ "learning_rate": 0.0006,
+ "loss": 4.163942337036133,
+ "step": 5727
+ },
+ {
+ "epoch": 79.5589519650655,
+ "grad_norm": 0.017831219360232353,
+ "learning_rate": 0.0006,
+ "loss": 4.236685276031494,
+ "step": 5728
+ },
+ {
+ "epoch": 79.57292576419214,
+ "grad_norm": 0.016334986314177513,
+ "learning_rate": 0.0006,
+ "loss": 4.255853652954102,
+ "step": 5729
+ },
+ {
+ "epoch": 79.58689956331878,
+ "grad_norm": 0.0138487434014678,
+ "learning_rate": 0.0006,
+ "loss": 4.227874279022217,
+ "step": 5730
+ },
+ {
+ "epoch": 79.60087336244541,
+ "grad_norm": 0.01531760673969984,
+ "learning_rate": 0.0006,
+ "loss": 4.221432685852051,
+ "step": 5731
+ },
+ {
+ "epoch": 79.61484716157206,
+ "grad_norm": 0.01603361964225769,
+ "learning_rate": 0.0006,
+ "loss": 4.278326511383057,
+ "step": 5732
+ },
+ {
+ "epoch": 79.62882096069869,
+ "grad_norm": 0.013149895705282688,
+ "learning_rate": 0.0006,
+ "loss": 4.2474517822265625,
+ "step": 5733
+ },
+ {
+ "epoch": 79.64279475982532,
+ "grad_norm": 0.013428786769509315,
+ "learning_rate": 0.0006,
+ "loss": 4.1701178550720215,
+ "step": 5734
+ },
+ {
+ "epoch": 79.65676855895197,
+ "grad_norm": 0.014305585995316505,
+ "learning_rate": 0.0006,
+ "loss": 4.250778675079346,
+ "step": 5735
+ },
+ {
+ "epoch": 79.6707423580786,
+ "grad_norm": 0.014544566161930561,
+ "learning_rate": 0.0006,
+ "loss": 4.243236541748047,
+ "step": 5736
+ },
+ {
+ "epoch": 79.68471615720524,
+ "grad_norm": 0.013524414040148258,
+ "learning_rate": 0.0006,
+ "loss": 4.255457878112793,
+ "step": 5737
+ },
+ {
+ "epoch": 79.69868995633188,
+ "grad_norm": 0.014750641770660877,
+ "learning_rate": 0.0006,
+ "loss": 4.202816486358643,
+ "step": 5738
+ },
+ {
+ "epoch": 79.71266375545852,
+ "grad_norm": 0.015090773813426495,
+ "learning_rate": 0.0006,
+ "loss": 4.3053178787231445,
+ "step": 5739
+ },
+ {
+ "epoch": 79.72663755458515,
+ "grad_norm": 0.01606169529259205,
+ "learning_rate": 0.0006,
+ "loss": 4.224017143249512,
+ "step": 5740
+ },
+ {
+ "epoch": 79.74061135371178,
+ "grad_norm": 0.016303520649671555,
+ "learning_rate": 0.0006,
+ "loss": 4.190760612487793,
+ "step": 5741
+ },
+ {
+ "epoch": 79.75458515283843,
+ "grad_norm": 0.01465672068297863,
+ "learning_rate": 0.0006,
+ "loss": 4.24506950378418,
+ "step": 5742
+ },
+ {
+ "epoch": 79.76855895196506,
+ "grad_norm": 0.012616301886737347,
+ "learning_rate": 0.0006,
+ "loss": 4.174039840698242,
+ "step": 5743
+ },
+ {
+ "epoch": 79.78253275109171,
+ "grad_norm": 0.01358284242451191,
+ "learning_rate": 0.0006,
+ "loss": 4.236380577087402,
+ "step": 5744
+ },
+ {
+ "epoch": 79.79650655021834,
+ "grad_norm": 0.01557177770882845,
+ "learning_rate": 0.0006,
+ "loss": 4.102962493896484,
+ "step": 5745
+ },
+ {
+ "epoch": 79.81048034934499,
+ "grad_norm": 0.015584317035973072,
+ "learning_rate": 0.0006,
+ "loss": 4.242603302001953,
+ "step": 5746
+ },
+ {
+ "epoch": 79.82445414847162,
+ "grad_norm": 0.012597334571182728,
+ "learning_rate": 0.0006,
+ "loss": 4.202791690826416,
+ "step": 5747
+ },
+ {
+ "epoch": 79.83842794759825,
+ "grad_norm": 0.013722898438572884,
+ "learning_rate": 0.0006,
+ "loss": 4.204328536987305,
+ "step": 5748
+ },
+ {
+ "epoch": 79.8524017467249,
+ "grad_norm": 0.014669225551187992,
+ "learning_rate": 0.0006,
+ "loss": 4.164331912994385,
+ "step": 5749
+ },
+ {
+ "epoch": 79.86637554585153,
+ "grad_norm": 0.015012740157544613,
+ "learning_rate": 0.0006,
+ "loss": 4.260565280914307,
+ "step": 5750
+ },
+ {
+ "epoch": 79.88034934497817,
+ "grad_norm": 0.016026340425014496,
+ "learning_rate": 0.0006,
+ "loss": 4.299723148345947,
+ "step": 5751
+ },
+ {
+ "epoch": 79.8943231441048,
+ "grad_norm": 0.01626049168407917,
+ "learning_rate": 0.0006,
+ "loss": 4.207252502441406,
+ "step": 5752
+ },
+ {
+ "epoch": 79.90829694323143,
+ "grad_norm": 0.015415354631841183,
+ "learning_rate": 0.0006,
+ "loss": 4.151934623718262,
+ "step": 5753
+ },
+ {
+ "epoch": 79.92227074235808,
+ "grad_norm": 0.014158242382109165,
+ "learning_rate": 0.0006,
+ "loss": 4.3461995124816895,
+ "step": 5754
+ },
+ {
+ "epoch": 79.93624454148471,
+ "grad_norm": 0.01585628278553486,
+ "learning_rate": 0.0006,
+ "loss": 4.334761619567871,
+ "step": 5755
+ },
+ {
+ "epoch": 79.95021834061136,
+ "grad_norm": 0.013710759580135345,
+ "learning_rate": 0.0006,
+ "loss": 4.180401802062988,
+ "step": 5756
+ },
+ {
+ "epoch": 79.96419213973799,
+ "grad_norm": 0.014159204438328743,
+ "learning_rate": 0.0006,
+ "loss": 4.322331428527832,
+ "step": 5757
+ },
+ {
+ "epoch": 79.97816593886463,
+ "grad_norm": 0.015153195708990097,
+ "learning_rate": 0.0006,
+ "loss": 4.287478923797607,
+ "step": 5758
+ },
+ {
+ "epoch": 79.99213973799127,
+ "grad_norm": 0.014100227504968643,
+ "learning_rate": 0.0006,
+ "loss": 4.231807708740234,
+ "step": 5759
+ },
+ {
+ "epoch": 80.0,
+ "grad_norm": 0.015248682349920273,
+ "learning_rate": 0.0006,
+ "loss": 4.261106014251709,
+ "step": 5760
+ },
+ {
+ "epoch": 80.0,
+ "eval_loss": 4.610264301300049,
+ "eval_runtime": 56.514,
+ "eval_samples_per_second": 43.211,
+ "eval_steps_per_second": 1.362,
+ "step": 5760
+ },
+ {
+ "epoch": 80.01397379912663,
+ "grad_norm": 0.014730271883308887,
+ "learning_rate": 0.0006,
+ "loss": 4.108556747436523,
+ "step": 5761
+ },
+ {
+ "epoch": 80.02794759825328,
+ "grad_norm": 0.014575716108083725,
+ "learning_rate": 0.0006,
+ "loss": 4.169658660888672,
+ "step": 5762
+ },
+ {
+ "epoch": 80.04192139737991,
+ "grad_norm": 0.014706958085298538,
+ "learning_rate": 0.0006,
+ "loss": 4.252662658691406,
+ "step": 5763
+ },
+ {
+ "epoch": 80.05589519650655,
+ "grad_norm": 0.01561224739998579,
+ "learning_rate": 0.0006,
+ "loss": 4.193892955780029,
+ "step": 5764
+ },
+ {
+ "epoch": 80.06986899563319,
+ "grad_norm": 0.014702324755489826,
+ "learning_rate": 0.0006,
+ "loss": 4.1889238357543945,
+ "step": 5765
+ },
+ {
+ "epoch": 80.08384279475983,
+ "grad_norm": 0.015565741807222366,
+ "learning_rate": 0.0006,
+ "loss": 4.243549823760986,
+ "step": 5766
+ },
+ {
+ "epoch": 80.09781659388646,
+ "grad_norm": 0.01647181063890457,
+ "learning_rate": 0.0006,
+ "loss": 4.275446891784668,
+ "step": 5767
+ },
+ {
+ "epoch": 80.1117903930131,
+ "grad_norm": 0.018177716061472893,
+ "learning_rate": 0.0006,
+ "loss": 4.12040376663208,
+ "step": 5768
+ },
+ {
+ "epoch": 80.12576419213974,
+ "grad_norm": 0.01745191030204296,
+ "learning_rate": 0.0006,
+ "loss": 4.140371322631836,
+ "step": 5769
+ },
+ {
+ "epoch": 80.13973799126637,
+ "grad_norm": 0.01568993367254734,
+ "learning_rate": 0.0006,
+ "loss": 4.258280277252197,
+ "step": 5770
+ },
+ {
+ "epoch": 80.15371179039302,
+ "grad_norm": 0.01636597141623497,
+ "learning_rate": 0.0006,
+ "loss": 4.179052352905273,
+ "step": 5771
+ },
+ {
+ "epoch": 80.16768558951965,
+ "grad_norm": 0.016839100047945976,
+ "learning_rate": 0.0006,
+ "loss": 4.127636432647705,
+ "step": 5772
+ },
+ {
+ "epoch": 80.18165938864628,
+ "grad_norm": 0.015471016988158226,
+ "learning_rate": 0.0006,
+ "loss": 4.181325912475586,
+ "step": 5773
+ },
+ {
+ "epoch": 80.19563318777293,
+ "grad_norm": 0.015836089849472046,
+ "learning_rate": 0.0006,
+ "loss": 4.142168998718262,
+ "step": 5774
+ },
+ {
+ "epoch": 80.20960698689956,
+ "grad_norm": 0.014782954007387161,
+ "learning_rate": 0.0006,
+ "loss": 4.308799743652344,
+ "step": 5775
+ },
+ {
+ "epoch": 80.2235807860262,
+ "grad_norm": 0.014040893875062466,
+ "learning_rate": 0.0006,
+ "loss": 4.216936111450195,
+ "step": 5776
+ },
+ {
+ "epoch": 80.23755458515284,
+ "grad_norm": 0.01420280709862709,
+ "learning_rate": 0.0006,
+ "loss": 4.235230445861816,
+ "step": 5777
+ },
+ {
+ "epoch": 80.25152838427948,
+ "grad_norm": 0.013341007754206657,
+ "learning_rate": 0.0006,
+ "loss": 4.369109153747559,
+ "step": 5778
+ },
+ {
+ "epoch": 80.26550218340611,
+ "grad_norm": 0.014208327978849411,
+ "learning_rate": 0.0006,
+ "loss": 4.219112396240234,
+ "step": 5779
+ },
+ {
+ "epoch": 80.27947598253274,
+ "grad_norm": 0.013825052417814732,
+ "learning_rate": 0.0006,
+ "loss": 4.1983489990234375,
+ "step": 5780
+ },
+ {
+ "epoch": 80.29344978165939,
+ "grad_norm": 0.012916515581309795,
+ "learning_rate": 0.0006,
+ "loss": 4.204582691192627,
+ "step": 5781
+ },
+ {
+ "epoch": 80.30742358078602,
+ "grad_norm": 0.013071243651211262,
+ "learning_rate": 0.0006,
+ "loss": 4.17643928527832,
+ "step": 5782
+ },
+ {
+ "epoch": 80.32139737991267,
+ "grad_norm": 0.015507174655795097,
+ "learning_rate": 0.0006,
+ "loss": 4.172976016998291,
+ "step": 5783
+ },
+ {
+ "epoch": 80.3353711790393,
+ "grad_norm": 0.014967243187129498,
+ "learning_rate": 0.0006,
+ "loss": 4.141168594360352,
+ "step": 5784
+ },
+ {
+ "epoch": 80.34934497816595,
+ "grad_norm": 0.015629924833774567,
+ "learning_rate": 0.0006,
+ "loss": 4.253526210784912,
+ "step": 5785
+ },
+ {
+ "epoch": 80.36331877729258,
+ "grad_norm": 0.01607954502105713,
+ "learning_rate": 0.0006,
+ "loss": 4.249395370483398,
+ "step": 5786
+ },
+ {
+ "epoch": 80.37729257641921,
+ "grad_norm": 0.01500706560909748,
+ "learning_rate": 0.0006,
+ "loss": 4.279012680053711,
+ "step": 5787
+ },
+ {
+ "epoch": 80.39126637554585,
+ "grad_norm": 0.014343900606036186,
+ "learning_rate": 0.0006,
+ "loss": 4.268509864807129,
+ "step": 5788
+ },
+ {
+ "epoch": 80.40524017467249,
+ "grad_norm": 0.014682197012007236,
+ "learning_rate": 0.0006,
+ "loss": 4.206334114074707,
+ "step": 5789
+ },
+ {
+ "epoch": 80.41921397379913,
+ "grad_norm": 0.014989163726568222,
+ "learning_rate": 0.0006,
+ "loss": 4.143601417541504,
+ "step": 5790
+ },
+ {
+ "epoch": 80.43318777292576,
+ "grad_norm": 0.015001608058810234,
+ "learning_rate": 0.0006,
+ "loss": 4.164111137390137,
+ "step": 5791
+ },
+ {
+ "epoch": 80.44716157205241,
+ "grad_norm": 0.015222722664475441,
+ "learning_rate": 0.0006,
+ "loss": 4.2369914054870605,
+ "step": 5792
+ },
+ {
+ "epoch": 80.46113537117904,
+ "grad_norm": 0.014126183465123177,
+ "learning_rate": 0.0006,
+ "loss": 4.2209930419921875,
+ "step": 5793
+ },
+ {
+ "epoch": 80.47510917030567,
+ "grad_norm": 0.013065959326922894,
+ "learning_rate": 0.0006,
+ "loss": 4.231184959411621,
+ "step": 5794
+ },
+ {
+ "epoch": 80.48908296943232,
+ "grad_norm": 0.013833962380886078,
+ "learning_rate": 0.0006,
+ "loss": 4.226434707641602,
+ "step": 5795
+ },
+ {
+ "epoch": 80.50305676855895,
+ "grad_norm": 0.01387125626206398,
+ "learning_rate": 0.0006,
+ "loss": 4.241328716278076,
+ "step": 5796
+ },
+ {
+ "epoch": 80.5170305676856,
+ "grad_norm": 0.014145063236355782,
+ "learning_rate": 0.0006,
+ "loss": 4.2268805503845215,
+ "step": 5797
+ },
+ {
+ "epoch": 80.53100436681223,
+ "grad_norm": 0.012990240007638931,
+ "learning_rate": 0.0006,
+ "loss": 4.1718339920043945,
+ "step": 5798
+ },
+ {
+ "epoch": 80.54497816593886,
+ "grad_norm": 0.014185411855578423,
+ "learning_rate": 0.0006,
+ "loss": 4.232397556304932,
+ "step": 5799
+ },
+ {
+ "epoch": 80.5589519650655,
+ "grad_norm": 0.014260523952543736,
+ "learning_rate": 0.0006,
+ "loss": 4.282734394073486,
+ "step": 5800
+ },
+ {
+ "epoch": 80.57292576419214,
+ "grad_norm": 0.012542120181024075,
+ "learning_rate": 0.0006,
+ "loss": 4.134551525115967,
+ "step": 5801
+ },
+ {
+ "epoch": 80.58689956331878,
+ "grad_norm": 0.013078657910227776,
+ "learning_rate": 0.0006,
+ "loss": 4.229815483093262,
+ "step": 5802
+ },
+ {
+ "epoch": 80.60087336244541,
+ "grad_norm": 0.012821056880056858,
+ "learning_rate": 0.0006,
+ "loss": 4.208278179168701,
+ "step": 5803
+ },
+ {
+ "epoch": 80.61484716157206,
+ "grad_norm": 0.012964180670678616,
+ "learning_rate": 0.0006,
+ "loss": 4.188477039337158,
+ "step": 5804
+ },
+ {
+ "epoch": 80.62882096069869,
+ "grad_norm": 0.01631746254861355,
+ "learning_rate": 0.0006,
+ "loss": 4.3214497566223145,
+ "step": 5805
+ },
+ {
+ "epoch": 80.64279475982532,
+ "grad_norm": 0.01726640947163105,
+ "learning_rate": 0.0006,
+ "loss": 4.279359817504883,
+ "step": 5806
+ },
+ {
+ "epoch": 80.65676855895197,
+ "grad_norm": 0.018778258934617043,
+ "learning_rate": 0.0006,
+ "loss": 4.294047832489014,
+ "step": 5807
+ },
+ {
+ "epoch": 80.6707423580786,
+ "grad_norm": 0.01796831749379635,
+ "learning_rate": 0.0006,
+ "loss": 4.314162254333496,
+ "step": 5808
+ },
+ {
+ "epoch": 80.68471615720524,
+ "grad_norm": 0.016495538875460625,
+ "learning_rate": 0.0006,
+ "loss": 4.273871421813965,
+ "step": 5809
+ },
+ {
+ "epoch": 80.69868995633188,
+ "grad_norm": 0.013590306043624878,
+ "learning_rate": 0.0006,
+ "loss": 4.14961051940918,
+ "step": 5810
+ },
+ {
+ "epoch": 80.71266375545852,
+ "grad_norm": 0.014670426957309246,
+ "learning_rate": 0.0006,
+ "loss": 4.211853981018066,
+ "step": 5811
+ },
+ {
+ "epoch": 80.72663755458515,
+ "grad_norm": 0.01641341857612133,
+ "learning_rate": 0.0006,
+ "loss": 4.211948871612549,
+ "step": 5812
+ },
+ {
+ "epoch": 80.74061135371178,
+ "grad_norm": 0.016365470364689827,
+ "learning_rate": 0.0006,
+ "loss": 4.200774192810059,
+ "step": 5813
+ },
+ {
+ "epoch": 80.75458515283843,
+ "grad_norm": 0.01613970287144184,
+ "learning_rate": 0.0006,
+ "loss": 4.288322448730469,
+ "step": 5814
+ },
+ {
+ "epoch": 80.76855895196506,
+ "grad_norm": 0.016742777079343796,
+ "learning_rate": 0.0006,
+ "loss": 4.23801326751709,
+ "step": 5815
+ },
+ {
+ "epoch": 80.78253275109171,
+ "grad_norm": 0.015351069159805775,
+ "learning_rate": 0.0006,
+ "loss": 4.263832092285156,
+ "step": 5816
+ },
+ {
+ "epoch": 80.79650655021834,
+ "grad_norm": 0.013435336761176586,
+ "learning_rate": 0.0006,
+ "loss": 4.160857200622559,
+ "step": 5817
+ },
+ {
+ "epoch": 80.81048034934499,
+ "grad_norm": 0.01474886853247881,
+ "learning_rate": 0.0006,
+ "loss": 4.216025352478027,
+ "step": 5818
+ },
+ {
+ "epoch": 80.82445414847162,
+ "grad_norm": 0.01586935855448246,
+ "learning_rate": 0.0006,
+ "loss": 4.256402969360352,
+ "step": 5819
+ },
+ {
+ "epoch": 80.83842794759825,
+ "grad_norm": 0.014596572145819664,
+ "learning_rate": 0.0006,
+ "loss": 4.142584800720215,
+ "step": 5820
+ },
+ {
+ "epoch": 80.8524017467249,
+ "grad_norm": 0.01501593366265297,
+ "learning_rate": 0.0006,
+ "loss": 4.250462055206299,
+ "step": 5821
+ },
+ {
+ "epoch": 80.86637554585153,
+ "grad_norm": 0.014773511327803135,
+ "learning_rate": 0.0006,
+ "loss": 4.253305912017822,
+ "step": 5822
+ },
+ {
+ "epoch": 80.88034934497817,
+ "grad_norm": 0.013246326707303524,
+ "learning_rate": 0.0006,
+ "loss": 4.170881271362305,
+ "step": 5823
+ },
+ {
+ "epoch": 80.8943231441048,
+ "grad_norm": 0.013013756833970547,
+ "learning_rate": 0.0006,
+ "loss": 4.264721870422363,
+ "step": 5824
+ },
+ {
+ "epoch": 80.90829694323143,
+ "grad_norm": 0.013155903667211533,
+ "learning_rate": 0.0006,
+ "loss": 4.274667739868164,
+ "step": 5825
+ },
+ {
+ "epoch": 80.92227074235808,
+ "grad_norm": 0.01448952779173851,
+ "learning_rate": 0.0006,
+ "loss": 4.24777889251709,
+ "step": 5826
+ },
+ {
+ "epoch": 80.93624454148471,
+ "grad_norm": 0.01415231078863144,
+ "learning_rate": 0.0006,
+ "loss": 4.22014856338501,
+ "step": 5827
+ },
+ {
+ "epoch": 80.95021834061136,
+ "grad_norm": 0.014796273782849312,
+ "learning_rate": 0.0006,
+ "loss": 4.22036600112915,
+ "step": 5828
+ },
+ {
+ "epoch": 80.96419213973799,
+ "grad_norm": 0.013005326502025127,
+ "learning_rate": 0.0006,
+ "loss": 4.276123046875,
+ "step": 5829
+ },
+ {
+ "epoch": 80.97816593886463,
+ "grad_norm": 0.012851177714765072,
+ "learning_rate": 0.0006,
+ "loss": 4.248575687408447,
+ "step": 5830
+ },
+ {
+ "epoch": 80.99213973799127,
+ "grad_norm": 0.013740002177655697,
+ "learning_rate": 0.0006,
+ "loss": 4.051140308380127,
+ "step": 5831
+ },
+ {
+ "epoch": 81.0,
+ "grad_norm": 0.015296104364097118,
+ "learning_rate": 0.0006,
+ "loss": 4.251155853271484,
+ "step": 5832
+ },
+ {
+ "epoch": 81.0,
+ "eval_loss": 4.594127655029297,
+ "eval_runtime": 56.0959,
+ "eval_samples_per_second": 43.533,
+ "eval_steps_per_second": 1.373,
+ "step": 5832
+ },
+ {
+ "epoch": 81.01397379912663,
+ "grad_norm": 0.015804558992385864,
+ "learning_rate": 0.0006,
+ "loss": 4.152894020080566,
+ "step": 5833
+ },
+ {
+ "epoch": 81.02794759825328,
+ "grad_norm": 0.017000854015350342,
+ "learning_rate": 0.0006,
+ "loss": 4.266067981719971,
+ "step": 5834
+ },
+ {
+ "epoch": 81.04192139737991,
+ "grad_norm": 0.016351008787751198,
+ "learning_rate": 0.0006,
+ "loss": 4.192703723907471,
+ "step": 5835
+ },
+ {
+ "epoch": 81.05589519650655,
+ "grad_norm": 0.014703895896673203,
+ "learning_rate": 0.0006,
+ "loss": 4.23554801940918,
+ "step": 5836
+ },
+ {
+ "epoch": 81.06986899563319,
+ "grad_norm": 0.01619596965610981,
+ "learning_rate": 0.0006,
+ "loss": 4.233000755310059,
+ "step": 5837
+ },
+ {
+ "epoch": 81.08384279475983,
+ "grad_norm": 0.01815493032336235,
+ "learning_rate": 0.0006,
+ "loss": 4.200490951538086,
+ "step": 5838
+ },
+ {
+ "epoch": 81.09781659388646,
+ "grad_norm": 0.018305106088519096,
+ "learning_rate": 0.0006,
+ "loss": 4.154242515563965,
+ "step": 5839
+ },
+ {
+ "epoch": 81.1117903930131,
+ "grad_norm": 0.0200856514275074,
+ "learning_rate": 0.0006,
+ "loss": 4.25820779800415,
+ "step": 5840
+ },
+ {
+ "epoch": 81.12576419213974,
+ "grad_norm": 0.02101440541446209,
+ "learning_rate": 0.0006,
+ "loss": 4.168209075927734,
+ "step": 5841
+ },
+ {
+ "epoch": 81.13973799126637,
+ "grad_norm": 0.022829681634902954,
+ "learning_rate": 0.0006,
+ "loss": 4.25560998916626,
+ "step": 5842
+ },
+ {
+ "epoch": 81.15371179039302,
+ "grad_norm": 0.0195082426071167,
+ "learning_rate": 0.0006,
+ "loss": 4.2035627365112305,
+ "step": 5843
+ },
+ {
+ "epoch": 81.16768558951965,
+ "grad_norm": 0.01716248318552971,
+ "learning_rate": 0.0006,
+ "loss": 4.11322021484375,
+ "step": 5844
+ },
+ {
+ "epoch": 81.18165938864628,
+ "grad_norm": 0.02034664712846279,
+ "learning_rate": 0.0006,
+ "loss": 4.179389953613281,
+ "step": 5845
+ },
+ {
+ "epoch": 81.19563318777293,
+ "grad_norm": 0.023578710854053497,
+ "learning_rate": 0.0006,
+ "loss": 4.265695571899414,
+ "step": 5846
+ },
+ {
+ "epoch": 81.20960698689956,
+ "grad_norm": 0.019244182854890823,
+ "learning_rate": 0.0006,
+ "loss": 4.292043209075928,
+ "step": 5847
+ },
+ {
+ "epoch": 81.2235807860262,
+ "grad_norm": 0.01577291637659073,
+ "learning_rate": 0.0006,
+ "loss": 4.250776290893555,
+ "step": 5848
+ },
+ {
+ "epoch": 81.23755458515284,
+ "grad_norm": 0.01617090217769146,
+ "learning_rate": 0.0006,
+ "loss": 4.193098068237305,
+ "step": 5849
+ },
+ {
+ "epoch": 81.25152838427948,
+ "grad_norm": 0.016090184450149536,
+ "learning_rate": 0.0006,
+ "loss": 4.112357139587402,
+ "step": 5850
+ },
+ {
+ "epoch": 81.26550218340611,
+ "grad_norm": 0.01441365759819746,
+ "learning_rate": 0.0006,
+ "loss": 4.114147186279297,
+ "step": 5851
+ },
+ {
+ "epoch": 81.27947598253274,
+ "grad_norm": 0.014122162945568562,
+ "learning_rate": 0.0006,
+ "loss": 4.276175022125244,
+ "step": 5852
+ },
+ {
+ "epoch": 81.29344978165939,
+ "grad_norm": 0.014859487302601337,
+ "learning_rate": 0.0006,
+ "loss": 4.094699859619141,
+ "step": 5853
+ },
+ {
+ "epoch": 81.30742358078602,
+ "grad_norm": 0.013624267652630806,
+ "learning_rate": 0.0006,
+ "loss": 4.2648773193359375,
+ "step": 5854
+ },
+ {
+ "epoch": 81.32139737991267,
+ "grad_norm": 0.01417913381010294,
+ "learning_rate": 0.0006,
+ "loss": 4.228567123413086,
+ "step": 5855
+ },
+ {
+ "epoch": 81.3353711790393,
+ "grad_norm": 0.014560805633664131,
+ "learning_rate": 0.0006,
+ "loss": 4.148930549621582,
+ "step": 5856
+ },
+ {
+ "epoch": 81.34934497816595,
+ "grad_norm": 0.013662154786288738,
+ "learning_rate": 0.0006,
+ "loss": 4.192107677459717,
+ "step": 5857
+ },
+ {
+ "epoch": 81.36331877729258,
+ "grad_norm": 0.013434980995953083,
+ "learning_rate": 0.0006,
+ "loss": 4.162052154541016,
+ "step": 5858
+ },
+ {
+ "epoch": 81.37729257641921,
+ "grad_norm": 0.01300742756575346,
+ "learning_rate": 0.0006,
+ "loss": 4.262385368347168,
+ "step": 5859
+ },
+ {
+ "epoch": 81.39126637554585,
+ "grad_norm": 0.012544351629912853,
+ "learning_rate": 0.0006,
+ "loss": 4.215371131896973,
+ "step": 5860
+ },
+ {
+ "epoch": 81.40524017467249,
+ "grad_norm": 0.013663645833730698,
+ "learning_rate": 0.0006,
+ "loss": 4.247312545776367,
+ "step": 5861
+ },
+ {
+ "epoch": 81.41921397379913,
+ "grad_norm": 0.013127296231687069,
+ "learning_rate": 0.0006,
+ "loss": 4.087685585021973,
+ "step": 5862
+ },
+ {
+ "epoch": 81.43318777292576,
+ "grad_norm": 0.014081938192248344,
+ "learning_rate": 0.0006,
+ "loss": 4.272648811340332,
+ "step": 5863
+ },
+ {
+ "epoch": 81.44716157205241,
+ "grad_norm": 0.015470641665160656,
+ "learning_rate": 0.0006,
+ "loss": 4.278578758239746,
+ "step": 5864
+ },
+ {
+ "epoch": 81.46113537117904,
+ "grad_norm": 0.016524532809853554,
+ "learning_rate": 0.0006,
+ "loss": 4.118268013000488,
+ "step": 5865
+ },
+ {
+ "epoch": 81.47510917030567,
+ "grad_norm": 0.016242722049355507,
+ "learning_rate": 0.0006,
+ "loss": 4.182405471801758,
+ "step": 5866
+ },
+ {
+ "epoch": 81.48908296943232,
+ "grad_norm": 0.01557017583400011,
+ "learning_rate": 0.0006,
+ "loss": 4.229862213134766,
+ "step": 5867
+ },
+ {
+ "epoch": 81.50305676855895,
+ "grad_norm": 0.01562870666384697,
+ "learning_rate": 0.0006,
+ "loss": 4.27754020690918,
+ "step": 5868
+ },
+ {
+ "epoch": 81.5170305676856,
+ "grad_norm": 0.018606074154376984,
+ "learning_rate": 0.0006,
+ "loss": 4.148681640625,
+ "step": 5869
+ },
+ {
+ "epoch": 81.53100436681223,
+ "grad_norm": 0.017740730196237564,
+ "learning_rate": 0.0006,
+ "loss": 4.321932792663574,
+ "step": 5870
+ },
+ {
+ "epoch": 81.54497816593886,
+ "grad_norm": 0.014517677947878838,
+ "learning_rate": 0.0006,
+ "loss": 4.22862434387207,
+ "step": 5871
+ },
+ {
+ "epoch": 81.5589519650655,
+ "grad_norm": 0.01389714702963829,
+ "learning_rate": 0.0006,
+ "loss": 4.198000907897949,
+ "step": 5872
+ },
+ {
+ "epoch": 81.57292576419214,
+ "grad_norm": 0.015348682180047035,
+ "learning_rate": 0.0006,
+ "loss": 4.293485164642334,
+ "step": 5873
+ },
+ {
+ "epoch": 81.58689956331878,
+ "grad_norm": 0.015965791419148445,
+ "learning_rate": 0.0006,
+ "loss": 4.23481559753418,
+ "step": 5874
+ },
+ {
+ "epoch": 81.60087336244541,
+ "grad_norm": 0.015662726014852524,
+ "learning_rate": 0.0006,
+ "loss": 4.158148765563965,
+ "step": 5875
+ },
+ {
+ "epoch": 81.61484716157206,
+ "grad_norm": 0.014402080327272415,
+ "learning_rate": 0.0006,
+ "loss": 4.132723808288574,
+ "step": 5876
+ },
+ {
+ "epoch": 81.62882096069869,
+ "grad_norm": 0.016846125945448875,
+ "learning_rate": 0.0006,
+ "loss": 4.1654462814331055,
+ "step": 5877
+ },
+ {
+ "epoch": 81.64279475982532,
+ "grad_norm": 0.0196097232401371,
+ "learning_rate": 0.0006,
+ "loss": 4.228450298309326,
+ "step": 5878
+ },
+ {
+ "epoch": 81.65676855895197,
+ "grad_norm": 0.016005294397473335,
+ "learning_rate": 0.0006,
+ "loss": 4.230222225189209,
+ "step": 5879
+ },
+ {
+ "epoch": 81.6707423580786,
+ "grad_norm": 0.014685059897601604,
+ "learning_rate": 0.0006,
+ "loss": 4.1367387771606445,
+ "step": 5880
+ },
+ {
+ "epoch": 81.68471615720524,
+ "grad_norm": 0.01647144928574562,
+ "learning_rate": 0.0006,
+ "loss": 4.096675395965576,
+ "step": 5881
+ },
+ {
+ "epoch": 81.69868995633188,
+ "grad_norm": 0.015978610143065453,
+ "learning_rate": 0.0006,
+ "loss": 4.232292175292969,
+ "step": 5882
+ },
+ {
+ "epoch": 81.71266375545852,
+ "grad_norm": 0.01510288193821907,
+ "learning_rate": 0.0006,
+ "loss": 4.274591445922852,
+ "step": 5883
+ },
+ {
+ "epoch": 81.72663755458515,
+ "grad_norm": 0.014942476525902748,
+ "learning_rate": 0.0006,
+ "loss": 4.161482810974121,
+ "step": 5884
+ },
+ {
+ "epoch": 81.74061135371178,
+ "grad_norm": 0.014595299027860165,
+ "learning_rate": 0.0006,
+ "loss": 4.204291343688965,
+ "step": 5885
+ },
+ {
+ "epoch": 81.75458515283843,
+ "grad_norm": 0.014952429570257664,
+ "learning_rate": 0.0006,
+ "loss": 4.242499351501465,
+ "step": 5886
+ },
+ {
+ "epoch": 81.76855895196506,
+ "grad_norm": 0.014761554077267647,
+ "learning_rate": 0.0006,
+ "loss": 4.20393180847168,
+ "step": 5887
+ },
+ {
+ "epoch": 81.78253275109171,
+ "grad_norm": 0.014749647118151188,
+ "learning_rate": 0.0006,
+ "loss": 4.227565765380859,
+ "step": 5888
+ },
+ {
+ "epoch": 81.79650655021834,
+ "grad_norm": 0.014384493231773376,
+ "learning_rate": 0.0006,
+ "loss": 4.282228946685791,
+ "step": 5889
+ },
+ {
+ "epoch": 81.81048034934499,
+ "grad_norm": 0.01260330155491829,
+ "learning_rate": 0.0006,
+ "loss": 4.2882866859436035,
+ "step": 5890
+ },
+ {
+ "epoch": 81.82445414847162,
+ "grad_norm": 0.014055311679840088,
+ "learning_rate": 0.0006,
+ "loss": 4.22351598739624,
+ "step": 5891
+ },
+ {
+ "epoch": 81.83842794759825,
+ "grad_norm": 0.013674065470695496,
+ "learning_rate": 0.0006,
+ "loss": 4.240716934204102,
+ "step": 5892
+ },
+ {
+ "epoch": 81.8524017467249,
+ "grad_norm": 0.01225196197628975,
+ "learning_rate": 0.0006,
+ "loss": 4.2958598136901855,
+ "step": 5893
+ },
+ {
+ "epoch": 81.86637554585153,
+ "grad_norm": 0.013538680970668793,
+ "learning_rate": 0.0006,
+ "loss": 4.24342679977417,
+ "step": 5894
+ },
+ {
+ "epoch": 81.88034934497817,
+ "grad_norm": 0.013151394203305244,
+ "learning_rate": 0.0006,
+ "loss": 4.081999778747559,
+ "step": 5895
+ },
+ {
+ "epoch": 81.8943231441048,
+ "grad_norm": 0.014260678552091122,
+ "learning_rate": 0.0006,
+ "loss": 4.156423568725586,
+ "step": 5896
+ },
+ {
+ "epoch": 81.90829694323143,
+ "grad_norm": 0.01329776830971241,
+ "learning_rate": 0.0006,
+ "loss": 4.1713995933532715,
+ "step": 5897
+ },
+ {
+ "epoch": 81.92227074235808,
+ "grad_norm": 0.014623276889324188,
+ "learning_rate": 0.0006,
+ "loss": 4.244891166687012,
+ "step": 5898
+ },
+ {
+ "epoch": 81.93624454148471,
+ "grad_norm": 0.015760324895381927,
+ "learning_rate": 0.0006,
+ "loss": 4.076528549194336,
+ "step": 5899
+ },
+ {
+ "epoch": 81.95021834061136,
+ "grad_norm": 0.014331766404211521,
+ "learning_rate": 0.0006,
+ "loss": 4.283037185668945,
+ "step": 5900
+ },
+ {
+ "epoch": 81.96419213973799,
+ "grad_norm": 0.012383842840790749,
+ "learning_rate": 0.0006,
+ "loss": 4.117308616638184,
+ "step": 5901
+ },
+ {
+ "epoch": 81.97816593886463,
+ "grad_norm": 0.01250660140067339,
+ "learning_rate": 0.0006,
+ "loss": 4.241250991821289,
+ "step": 5902
+ },
+ {
+ "epoch": 81.99213973799127,
+ "grad_norm": 0.012276900932192802,
+ "learning_rate": 0.0006,
+ "loss": 4.141091823577881,
+ "step": 5903
+ },
+ {
+ "epoch": 82.0,
+ "grad_norm": 0.01588737778365612,
+ "learning_rate": 0.0006,
+ "loss": 4.065532684326172,
+ "step": 5904
+ },
+ {
+ "epoch": 82.0,
+ "eval_loss": 4.612304210662842,
+ "eval_runtime": 57.5411,
+ "eval_samples_per_second": 42.439,
+ "eval_steps_per_second": 1.338,
+ "step": 5904
+ },
+ {
+ "epoch": 82.01397379912663,
+ "grad_norm": 0.017433544620871544,
+ "learning_rate": 0.0006,
+ "loss": 4.202497482299805,
+ "step": 5905
+ },
+ {
+ "epoch": 82.02794759825328,
+ "grad_norm": 0.017260415479540825,
+ "learning_rate": 0.0006,
+ "loss": 4.099949359893799,
+ "step": 5906
+ },
+ {
+ "epoch": 82.04192139737991,
+ "grad_norm": 0.01577908545732498,
+ "learning_rate": 0.0006,
+ "loss": 4.17806339263916,
+ "step": 5907
+ },
+ {
+ "epoch": 82.05589519650655,
+ "grad_norm": 0.016093818470835686,
+ "learning_rate": 0.0006,
+ "loss": 4.219491004943848,
+ "step": 5908
+ },
+ {
+ "epoch": 82.06986899563319,
+ "grad_norm": 0.015426302328705788,
+ "learning_rate": 0.0006,
+ "loss": 4.207982063293457,
+ "step": 5909
+ },
+ {
+ "epoch": 82.08384279475983,
+ "grad_norm": 0.01605442352592945,
+ "learning_rate": 0.0006,
+ "loss": 4.265942573547363,
+ "step": 5910
+ },
+ {
+ "epoch": 82.09781659388646,
+ "grad_norm": 0.01787225715816021,
+ "learning_rate": 0.0006,
+ "loss": 4.151535511016846,
+ "step": 5911
+ },
+ {
+ "epoch": 82.1117903930131,
+ "grad_norm": 0.01750164106488228,
+ "learning_rate": 0.0006,
+ "loss": 4.133655071258545,
+ "step": 5912
+ },
+ {
+ "epoch": 82.12576419213974,
+ "grad_norm": 0.01602957211434841,
+ "learning_rate": 0.0006,
+ "loss": 4.176972389221191,
+ "step": 5913
+ },
+ {
+ "epoch": 82.13973799126637,
+ "grad_norm": 0.01553812064230442,
+ "learning_rate": 0.0006,
+ "loss": 4.125692367553711,
+ "step": 5914
+ },
+ {
+ "epoch": 82.15371179039302,
+ "grad_norm": 0.013978508301079273,
+ "learning_rate": 0.0006,
+ "loss": 4.207271575927734,
+ "step": 5915
+ },
+ {
+ "epoch": 82.16768558951965,
+ "grad_norm": 0.013561118394136429,
+ "learning_rate": 0.0006,
+ "loss": 4.238930702209473,
+ "step": 5916
+ },
+ {
+ "epoch": 82.18165938864628,
+ "grad_norm": 0.013285507448017597,
+ "learning_rate": 0.0006,
+ "loss": 4.129464626312256,
+ "step": 5917
+ },
+ {
+ "epoch": 82.19563318777293,
+ "grad_norm": 0.013903839513659477,
+ "learning_rate": 0.0006,
+ "loss": 4.214109420776367,
+ "step": 5918
+ },
+ {
+ "epoch": 82.20960698689956,
+ "grad_norm": 0.014426837675273418,
+ "learning_rate": 0.0006,
+ "loss": 4.143196105957031,
+ "step": 5919
+ },
+ {
+ "epoch": 82.2235807860262,
+ "grad_norm": 0.015347165986895561,
+ "learning_rate": 0.0006,
+ "loss": 4.235439300537109,
+ "step": 5920
+ },
+ {
+ "epoch": 82.23755458515284,
+ "grad_norm": 0.014266039244830608,
+ "learning_rate": 0.0006,
+ "loss": 3.991617202758789,
+ "step": 5921
+ },
+ {
+ "epoch": 82.25152838427948,
+ "grad_norm": 0.013101593591272831,
+ "learning_rate": 0.0006,
+ "loss": 4.266017913818359,
+ "step": 5922
+ },
+ {
+ "epoch": 82.26550218340611,
+ "grad_norm": 0.013741881586611271,
+ "learning_rate": 0.0006,
+ "loss": 4.224879741668701,
+ "step": 5923
+ },
+ {
+ "epoch": 82.27947598253274,
+ "grad_norm": 0.014206945896148682,
+ "learning_rate": 0.0006,
+ "loss": 4.102272987365723,
+ "step": 5924
+ },
+ {
+ "epoch": 82.29344978165939,
+ "grad_norm": 0.013798723928630352,
+ "learning_rate": 0.0006,
+ "loss": 4.30418586730957,
+ "step": 5925
+ },
+ {
+ "epoch": 82.30742358078602,
+ "grad_norm": 0.015470122918486595,
+ "learning_rate": 0.0006,
+ "loss": 4.2513322830200195,
+ "step": 5926
+ },
+ {
+ "epoch": 82.32139737991267,
+ "grad_norm": 0.01655442826449871,
+ "learning_rate": 0.0006,
+ "loss": 4.223002910614014,
+ "step": 5927
+ },
+ {
+ "epoch": 82.3353711790393,
+ "grad_norm": 0.017170004546642303,
+ "learning_rate": 0.0006,
+ "loss": 4.253904819488525,
+ "step": 5928
+ },
+ {
+ "epoch": 82.34934497816595,
+ "grad_norm": 0.017595812678337097,
+ "learning_rate": 0.0006,
+ "loss": 4.184927940368652,
+ "step": 5929
+ },
+ {
+ "epoch": 82.36331877729258,
+ "grad_norm": 0.015073256567120552,
+ "learning_rate": 0.0006,
+ "loss": 4.202239036560059,
+ "step": 5930
+ },
+ {
+ "epoch": 82.37729257641921,
+ "grad_norm": 0.014310941100120544,
+ "learning_rate": 0.0006,
+ "loss": 4.124934673309326,
+ "step": 5931
+ },
+ {
+ "epoch": 82.39126637554585,
+ "grad_norm": 0.013671821914613247,
+ "learning_rate": 0.0006,
+ "loss": 4.138919353485107,
+ "step": 5932
+ },
+ {
+ "epoch": 82.40524017467249,
+ "grad_norm": 0.014103314839303493,
+ "learning_rate": 0.0006,
+ "loss": 4.211289405822754,
+ "step": 5933
+ },
+ {
+ "epoch": 82.41921397379913,
+ "grad_norm": 0.014272668398916721,
+ "learning_rate": 0.0006,
+ "loss": 4.06290340423584,
+ "step": 5934
+ },
+ {
+ "epoch": 82.43318777292576,
+ "grad_norm": 0.01387016475200653,
+ "learning_rate": 0.0006,
+ "loss": 4.101334571838379,
+ "step": 5935
+ },
+ {
+ "epoch": 82.44716157205241,
+ "grad_norm": 0.013318683952093124,
+ "learning_rate": 0.0006,
+ "loss": 4.206409931182861,
+ "step": 5936
+ },
+ {
+ "epoch": 82.46113537117904,
+ "grad_norm": 0.013781268149614334,
+ "learning_rate": 0.0006,
+ "loss": 4.189474105834961,
+ "step": 5937
+ },
+ {
+ "epoch": 82.47510917030567,
+ "grad_norm": 0.013195851817727089,
+ "learning_rate": 0.0006,
+ "loss": 4.276607036590576,
+ "step": 5938
+ },
+ {
+ "epoch": 82.48908296943232,
+ "grad_norm": 0.014358163811266422,
+ "learning_rate": 0.0006,
+ "loss": 4.203262805938721,
+ "step": 5939
+ },
+ {
+ "epoch": 82.50305676855895,
+ "grad_norm": 0.01403636671602726,
+ "learning_rate": 0.0006,
+ "loss": 4.079258441925049,
+ "step": 5940
+ },
+ {
+ "epoch": 82.5170305676856,
+ "grad_norm": 0.013272535055875778,
+ "learning_rate": 0.0006,
+ "loss": 4.19197940826416,
+ "step": 5941
+ },
+ {
+ "epoch": 82.53100436681223,
+ "grad_norm": 0.013787264935672283,
+ "learning_rate": 0.0006,
+ "loss": 4.12995719909668,
+ "step": 5942
+ },
+ {
+ "epoch": 82.54497816593886,
+ "grad_norm": 0.015020057559013367,
+ "learning_rate": 0.0006,
+ "loss": 4.148890495300293,
+ "step": 5943
+ },
+ {
+ "epoch": 82.5589519650655,
+ "grad_norm": 0.014199157245457172,
+ "learning_rate": 0.0006,
+ "loss": 4.201516151428223,
+ "step": 5944
+ },
+ {
+ "epoch": 82.57292576419214,
+ "grad_norm": 0.013827123679220676,
+ "learning_rate": 0.0006,
+ "loss": 4.240894317626953,
+ "step": 5945
+ },
+ {
+ "epoch": 82.58689956331878,
+ "grad_norm": 0.014767560176551342,
+ "learning_rate": 0.0006,
+ "loss": 4.173320770263672,
+ "step": 5946
+ },
+ {
+ "epoch": 82.60087336244541,
+ "grad_norm": 0.01612078957259655,
+ "learning_rate": 0.0006,
+ "loss": 4.266448020935059,
+ "step": 5947
+ },
+ {
+ "epoch": 82.61484716157206,
+ "grad_norm": 0.016706543043255806,
+ "learning_rate": 0.0006,
+ "loss": 4.229398250579834,
+ "step": 5948
+ },
+ {
+ "epoch": 82.62882096069869,
+ "grad_norm": 0.015630796551704407,
+ "learning_rate": 0.0006,
+ "loss": 4.181282997131348,
+ "step": 5949
+ },
+ {
+ "epoch": 82.64279475982532,
+ "grad_norm": 0.013103844597935677,
+ "learning_rate": 0.0006,
+ "loss": 4.309617042541504,
+ "step": 5950
+ },
+ {
+ "epoch": 82.65676855895197,
+ "grad_norm": 0.016136541962623596,
+ "learning_rate": 0.0006,
+ "loss": 4.224159240722656,
+ "step": 5951
+ },
+ {
+ "epoch": 82.6707423580786,
+ "grad_norm": 0.016160814091563225,
+ "learning_rate": 0.0006,
+ "loss": 4.267970085144043,
+ "step": 5952
+ },
+ {
+ "epoch": 82.68471615720524,
+ "grad_norm": 0.01672680303454399,
+ "learning_rate": 0.0006,
+ "loss": 4.230506896972656,
+ "step": 5953
+ },
+ {
+ "epoch": 82.69868995633188,
+ "grad_norm": 0.0159586351364851,
+ "learning_rate": 0.0006,
+ "loss": 4.159542083740234,
+ "step": 5954
+ },
+ {
+ "epoch": 82.71266375545852,
+ "grad_norm": 0.01461219135671854,
+ "learning_rate": 0.0006,
+ "loss": 4.27863073348999,
+ "step": 5955
+ },
+ {
+ "epoch": 82.72663755458515,
+ "grad_norm": 0.015088613145053387,
+ "learning_rate": 0.0006,
+ "loss": 4.272418975830078,
+ "step": 5956
+ },
+ {
+ "epoch": 82.74061135371178,
+ "grad_norm": 0.01594499684870243,
+ "learning_rate": 0.0006,
+ "loss": 4.271326065063477,
+ "step": 5957
+ },
+ {
+ "epoch": 82.75458515283843,
+ "grad_norm": 0.016648447141051292,
+ "learning_rate": 0.0006,
+ "loss": 4.141478538513184,
+ "step": 5958
+ },
+ {
+ "epoch": 82.76855895196506,
+ "grad_norm": 0.01540316641330719,
+ "learning_rate": 0.0006,
+ "loss": 4.111814498901367,
+ "step": 5959
+ },
+ {
+ "epoch": 82.78253275109171,
+ "grad_norm": 0.01634364202618599,
+ "learning_rate": 0.0006,
+ "loss": 4.3113112449646,
+ "step": 5960
+ },
+ {
+ "epoch": 82.79650655021834,
+ "grad_norm": 0.014768977649509907,
+ "learning_rate": 0.0006,
+ "loss": 4.184217929840088,
+ "step": 5961
+ },
+ {
+ "epoch": 82.81048034934499,
+ "grad_norm": 0.013453755527734756,
+ "learning_rate": 0.0006,
+ "loss": 4.247729301452637,
+ "step": 5962
+ },
+ {
+ "epoch": 82.82445414847162,
+ "grad_norm": 0.014731907285749912,
+ "learning_rate": 0.0006,
+ "loss": 4.352839469909668,
+ "step": 5963
+ },
+ {
+ "epoch": 82.83842794759825,
+ "grad_norm": 0.016047345474362373,
+ "learning_rate": 0.0006,
+ "loss": 4.257283687591553,
+ "step": 5964
+ },
+ {
+ "epoch": 82.8524017467249,
+ "grad_norm": 0.017141170799732208,
+ "learning_rate": 0.0006,
+ "loss": 4.22850227355957,
+ "step": 5965
+ },
+ {
+ "epoch": 82.86637554585153,
+ "grad_norm": 0.01370396837592125,
+ "learning_rate": 0.0006,
+ "loss": 4.1169257164001465,
+ "step": 5966
+ },
+ {
+ "epoch": 82.88034934497817,
+ "grad_norm": 0.012060081586241722,
+ "learning_rate": 0.0006,
+ "loss": 4.195242404937744,
+ "step": 5967
+ },
+ {
+ "epoch": 82.8943231441048,
+ "grad_norm": 0.015528378076851368,
+ "learning_rate": 0.0006,
+ "loss": 4.294939994812012,
+ "step": 5968
+ },
+ {
+ "epoch": 82.90829694323143,
+ "grad_norm": 0.018151041120290756,
+ "learning_rate": 0.0006,
+ "loss": 4.137192726135254,
+ "step": 5969
+ },
+ {
+ "epoch": 82.92227074235808,
+ "grad_norm": 0.017895735800266266,
+ "learning_rate": 0.0006,
+ "loss": 4.174036026000977,
+ "step": 5970
+ },
+ {
+ "epoch": 82.93624454148471,
+ "grad_norm": 0.015610386617481709,
+ "learning_rate": 0.0006,
+ "loss": 4.230136871337891,
+ "step": 5971
+ },
+ {
+ "epoch": 82.95021834061136,
+ "grad_norm": 0.01339928898960352,
+ "learning_rate": 0.0006,
+ "loss": 4.199074745178223,
+ "step": 5972
+ },
+ {
+ "epoch": 82.96419213973799,
+ "grad_norm": 0.015454214997589588,
+ "learning_rate": 0.0006,
+ "loss": 4.216109752655029,
+ "step": 5973
+ },
+ {
+ "epoch": 82.97816593886463,
+ "grad_norm": 0.014978132210671902,
+ "learning_rate": 0.0006,
+ "loss": 4.131511688232422,
+ "step": 5974
+ },
+ {
+ "epoch": 82.99213973799127,
+ "grad_norm": 0.015611122362315655,
+ "learning_rate": 0.0006,
+ "loss": 4.246518135070801,
+ "step": 5975
+ },
+ {
+ "epoch": 83.0,
+ "grad_norm": 0.016682764515280724,
+ "learning_rate": 0.0006,
+ "loss": 4.110680103302002,
+ "step": 5976
+ },
+ {
+ "epoch": 83.0,
+ "eval_loss": 4.641707897186279,
+ "eval_runtime": 57.2799,
+ "eval_samples_per_second": 42.633,
+ "eval_steps_per_second": 1.344,
+ "step": 5976
+ },
+ {
+ "epoch": 83.01397379912663,
+ "grad_norm": 0.015007982030510902,
+ "learning_rate": 0.0006,
+ "loss": 4.100912094116211,
+ "step": 5977
+ },
+ {
+ "epoch": 83.02794759825328,
+ "grad_norm": 0.017609436064958572,
+ "learning_rate": 0.0006,
+ "loss": 4.233399868011475,
+ "step": 5978
+ },
+ {
+ "epoch": 83.04192139737991,
+ "grad_norm": 0.019817272201180458,
+ "learning_rate": 0.0006,
+ "loss": 4.191173076629639,
+ "step": 5979
+ },
+ {
+ "epoch": 83.05589519650655,
+ "grad_norm": 0.017752140760421753,
+ "learning_rate": 0.0006,
+ "loss": 4.044437885284424,
+ "step": 5980
+ },
+ {
+ "epoch": 83.06986899563319,
+ "grad_norm": 0.015482760034501553,
+ "learning_rate": 0.0006,
+ "loss": 4.319931983947754,
+ "step": 5981
+ },
+ {
+ "epoch": 83.08384279475983,
+ "grad_norm": 0.015153988264501095,
+ "learning_rate": 0.0006,
+ "loss": 4.095277309417725,
+ "step": 5982
+ },
+ {
+ "epoch": 83.09781659388646,
+ "grad_norm": 0.017025282606482506,
+ "learning_rate": 0.0006,
+ "loss": 4.235110759735107,
+ "step": 5983
+ },
+ {
+ "epoch": 83.1117903930131,
+ "grad_norm": 0.01566586270928383,
+ "learning_rate": 0.0006,
+ "loss": 4.1934685707092285,
+ "step": 5984
+ },
+ {
+ "epoch": 83.12576419213974,
+ "grad_norm": 0.016419263556599617,
+ "learning_rate": 0.0006,
+ "loss": 4.234419822692871,
+ "step": 5985
+ },
+ {
+ "epoch": 83.13973799126637,
+ "grad_norm": 0.017132675275206566,
+ "learning_rate": 0.0006,
+ "loss": 4.221447944641113,
+ "step": 5986
+ },
+ {
+ "epoch": 83.15371179039302,
+ "grad_norm": 0.016138629987835884,
+ "learning_rate": 0.0006,
+ "loss": 4.199830055236816,
+ "step": 5987
+ },
+ {
+ "epoch": 83.16768558951965,
+ "grad_norm": 0.016210174188017845,
+ "learning_rate": 0.0006,
+ "loss": 4.081717491149902,
+ "step": 5988
+ },
+ {
+ "epoch": 83.18165938864628,
+ "grad_norm": 0.016055308282375336,
+ "learning_rate": 0.0006,
+ "loss": 4.247655391693115,
+ "step": 5989
+ },
+ {
+ "epoch": 83.19563318777293,
+ "grad_norm": 0.016132952645421028,
+ "learning_rate": 0.0006,
+ "loss": 4.262535572052002,
+ "step": 5990
+ },
+ {
+ "epoch": 83.20960698689956,
+ "grad_norm": 0.01779935695230961,
+ "learning_rate": 0.0006,
+ "loss": 4.153120040893555,
+ "step": 5991
+ },
+ {
+ "epoch": 83.2235807860262,
+ "grad_norm": 0.018360646441578865,
+ "learning_rate": 0.0006,
+ "loss": 4.232644081115723,
+ "step": 5992
+ },
+ {
+ "epoch": 83.23755458515284,
+ "grad_norm": 0.019629675894975662,
+ "learning_rate": 0.0006,
+ "loss": 3.9758734703063965,
+ "step": 5993
+ },
+ {
+ "epoch": 83.25152838427948,
+ "grad_norm": 0.017819996923208237,
+ "learning_rate": 0.0006,
+ "loss": 4.196932315826416,
+ "step": 5994
+ },
+ {
+ "epoch": 83.26550218340611,
+ "grad_norm": 0.01557681243866682,
+ "learning_rate": 0.0006,
+ "loss": 4.122654914855957,
+ "step": 5995
+ },
+ {
+ "epoch": 83.27947598253274,
+ "grad_norm": 0.017858853563666344,
+ "learning_rate": 0.0006,
+ "loss": 4.185330867767334,
+ "step": 5996
+ },
+ {
+ "epoch": 83.29344978165939,
+ "grad_norm": 0.018065953627228737,
+ "learning_rate": 0.0006,
+ "loss": 4.218471050262451,
+ "step": 5997
+ },
+ {
+ "epoch": 83.30742358078602,
+ "grad_norm": 0.017247939482331276,
+ "learning_rate": 0.0006,
+ "loss": 4.127076148986816,
+ "step": 5998
+ },
+ {
+ "epoch": 83.32139737991267,
+ "grad_norm": 0.017047952860593796,
+ "learning_rate": 0.0006,
+ "loss": 4.131356716156006,
+ "step": 5999
+ },
+ {
+ "epoch": 83.3353711790393,
+ "grad_norm": 0.015905629843473434,
+ "learning_rate": 0.0006,
+ "loss": 4.2239789962768555,
+ "step": 6000
+ },
+ {
+ "epoch": 83.34934497816595,
+ "grad_norm": 0.016522347927093506,
+ "learning_rate": 0.0006,
+ "loss": 4.054897785186768,
+ "step": 6001
+ },
+ {
+ "epoch": 83.36331877729258,
+ "grad_norm": 0.01622571051120758,
+ "learning_rate": 0.0006,
+ "loss": 4.284210681915283,
+ "step": 6002
+ },
+ {
+ "epoch": 83.37729257641921,
+ "grad_norm": 0.01640467159450054,
+ "learning_rate": 0.0006,
+ "loss": 4.268840789794922,
+ "step": 6003
+ },
+ {
+ "epoch": 83.39126637554585,
+ "grad_norm": 0.015698859468102455,
+ "learning_rate": 0.0006,
+ "loss": 4.179793357849121,
+ "step": 6004
+ },
+ {
+ "epoch": 83.40524017467249,
+ "grad_norm": 0.016694528982043266,
+ "learning_rate": 0.0006,
+ "loss": 4.132297515869141,
+ "step": 6005
+ },
+ {
+ "epoch": 83.41921397379913,
+ "grad_norm": 0.015345900319516659,
+ "learning_rate": 0.0006,
+ "loss": 4.090163230895996,
+ "step": 6006
+ },
+ {
+ "epoch": 83.43318777292576,
+ "grad_norm": 0.014722758904099464,
+ "learning_rate": 0.0006,
+ "loss": 4.179166793823242,
+ "step": 6007
+ },
+ {
+ "epoch": 83.44716157205241,
+ "grad_norm": 0.014884887263178825,
+ "learning_rate": 0.0006,
+ "loss": 4.137659072875977,
+ "step": 6008
+ },
+ {
+ "epoch": 83.46113537117904,
+ "grad_norm": 0.014041902497410774,
+ "learning_rate": 0.0006,
+ "loss": 4.1202073097229,
+ "step": 6009
+ },
+ {
+ "epoch": 83.47510917030567,
+ "grad_norm": 0.012799333781003952,
+ "learning_rate": 0.0006,
+ "loss": 4.092607498168945,
+ "step": 6010
+ },
+ {
+ "epoch": 83.48908296943232,
+ "grad_norm": 0.013272707350552082,
+ "learning_rate": 0.0006,
+ "loss": 4.132715225219727,
+ "step": 6011
+ },
+ {
+ "epoch": 83.50305676855895,
+ "grad_norm": 0.012934371829032898,
+ "learning_rate": 0.0006,
+ "loss": 4.230278968811035,
+ "step": 6012
+ },
+ {
+ "epoch": 83.5170305676856,
+ "grad_norm": 0.0120006762444973,
+ "learning_rate": 0.0006,
+ "loss": 4.140068531036377,
+ "step": 6013
+ },
+ {
+ "epoch": 83.53100436681223,
+ "grad_norm": 0.013616700656712055,
+ "learning_rate": 0.0006,
+ "loss": 4.20107364654541,
+ "step": 6014
+ },
+ {
+ "epoch": 83.54497816593886,
+ "grad_norm": 0.013957851566374302,
+ "learning_rate": 0.0006,
+ "loss": 4.124945163726807,
+ "step": 6015
+ },
+ {
+ "epoch": 83.5589519650655,
+ "grad_norm": 0.012762055732309818,
+ "learning_rate": 0.0006,
+ "loss": 4.022052764892578,
+ "step": 6016
+ },
+ {
+ "epoch": 83.57292576419214,
+ "grad_norm": 0.013080576434731483,
+ "learning_rate": 0.0006,
+ "loss": 4.103252410888672,
+ "step": 6017
+ },
+ {
+ "epoch": 83.58689956331878,
+ "grad_norm": 0.013957500457763672,
+ "learning_rate": 0.0006,
+ "loss": 4.235089302062988,
+ "step": 6018
+ },
+ {
+ "epoch": 83.60087336244541,
+ "grad_norm": 0.014707164838910103,
+ "learning_rate": 0.0006,
+ "loss": 4.191570281982422,
+ "step": 6019
+ },
+ {
+ "epoch": 83.61484716157206,
+ "grad_norm": 0.012807869352400303,
+ "learning_rate": 0.0006,
+ "loss": 4.294543266296387,
+ "step": 6020
+ },
+ {
+ "epoch": 83.62882096069869,
+ "grad_norm": 0.012027994729578495,
+ "learning_rate": 0.0006,
+ "loss": 4.13849401473999,
+ "step": 6021
+ },
+ {
+ "epoch": 83.64279475982532,
+ "grad_norm": 0.012147724628448486,
+ "learning_rate": 0.0006,
+ "loss": 4.127481937408447,
+ "step": 6022
+ },
+ {
+ "epoch": 83.65676855895197,
+ "grad_norm": 0.013241392560303211,
+ "learning_rate": 0.0006,
+ "loss": 4.218972682952881,
+ "step": 6023
+ },
+ {
+ "epoch": 83.6707423580786,
+ "grad_norm": 0.014071648940443993,
+ "learning_rate": 0.0006,
+ "loss": 4.096724033355713,
+ "step": 6024
+ },
+ {
+ "epoch": 83.68471615720524,
+ "grad_norm": 0.015212797559797764,
+ "learning_rate": 0.0006,
+ "loss": 4.27097225189209,
+ "step": 6025
+ },
+ {
+ "epoch": 83.69868995633188,
+ "grad_norm": 0.017647385597229004,
+ "learning_rate": 0.0006,
+ "loss": 4.209835529327393,
+ "step": 6026
+ },
+ {
+ "epoch": 83.71266375545852,
+ "grad_norm": 0.01709647662937641,
+ "learning_rate": 0.0006,
+ "loss": 4.042106628417969,
+ "step": 6027
+ },
+ {
+ "epoch": 83.72663755458515,
+ "grad_norm": 0.014530927874147892,
+ "learning_rate": 0.0006,
+ "loss": 4.181731700897217,
+ "step": 6028
+ },
+ {
+ "epoch": 83.74061135371178,
+ "grad_norm": 0.015356711111962795,
+ "learning_rate": 0.0006,
+ "loss": 4.314183235168457,
+ "step": 6029
+ },
+ {
+ "epoch": 83.75458515283843,
+ "grad_norm": 0.015500029549002647,
+ "learning_rate": 0.0006,
+ "loss": 4.06776762008667,
+ "step": 6030
+ },
+ {
+ "epoch": 83.76855895196506,
+ "grad_norm": 0.015351332724094391,
+ "learning_rate": 0.0006,
+ "loss": 4.185141563415527,
+ "step": 6031
+ },
+ {
+ "epoch": 83.78253275109171,
+ "grad_norm": 0.016678154468536377,
+ "learning_rate": 0.0006,
+ "loss": 4.179529666900635,
+ "step": 6032
+ },
+ {
+ "epoch": 83.79650655021834,
+ "grad_norm": 0.0156265702098608,
+ "learning_rate": 0.0006,
+ "loss": 4.238584041595459,
+ "step": 6033
+ },
+ {
+ "epoch": 83.81048034934499,
+ "grad_norm": 0.016418596729636192,
+ "learning_rate": 0.0006,
+ "loss": 4.1593523025512695,
+ "step": 6034
+ },
+ {
+ "epoch": 83.82445414847162,
+ "grad_norm": 0.016113586723804474,
+ "learning_rate": 0.0006,
+ "loss": 4.157470226287842,
+ "step": 6035
+ },
+ {
+ "epoch": 83.83842794759825,
+ "grad_norm": 0.014434042386710644,
+ "learning_rate": 0.0006,
+ "loss": 4.196226119995117,
+ "step": 6036
+ },
+ {
+ "epoch": 83.8524017467249,
+ "grad_norm": 0.012647042982280254,
+ "learning_rate": 0.0006,
+ "loss": 4.130589485168457,
+ "step": 6037
+ },
+ {
+ "epoch": 83.86637554585153,
+ "grad_norm": 0.014412877149879932,
+ "learning_rate": 0.0006,
+ "loss": 4.154845237731934,
+ "step": 6038
+ },
+ {
+ "epoch": 83.88034934497817,
+ "grad_norm": 0.014222027733922005,
+ "learning_rate": 0.0006,
+ "loss": 4.189724922180176,
+ "step": 6039
+ },
+ {
+ "epoch": 83.8943231441048,
+ "grad_norm": 0.013084255158901215,
+ "learning_rate": 0.0006,
+ "loss": 4.303338527679443,
+ "step": 6040
+ },
+ {
+ "epoch": 83.90829694323143,
+ "grad_norm": 0.012659232132136822,
+ "learning_rate": 0.0006,
+ "loss": 4.08192253112793,
+ "step": 6041
+ },
+ {
+ "epoch": 83.92227074235808,
+ "grad_norm": 0.01246276218444109,
+ "learning_rate": 0.0006,
+ "loss": 4.283324718475342,
+ "step": 6042
+ },
+ {
+ "epoch": 83.93624454148471,
+ "grad_norm": 0.01340118795633316,
+ "learning_rate": 0.0006,
+ "loss": 4.117908477783203,
+ "step": 6043
+ },
+ {
+ "epoch": 83.95021834061136,
+ "grad_norm": 0.014496722258627415,
+ "learning_rate": 0.0006,
+ "loss": 4.110318660736084,
+ "step": 6044
+ },
+ {
+ "epoch": 83.96419213973799,
+ "grad_norm": 0.013348298147320747,
+ "learning_rate": 0.0006,
+ "loss": 4.185139179229736,
+ "step": 6045
+ },
+ {
+ "epoch": 83.97816593886463,
+ "grad_norm": 0.013294277712702751,
+ "learning_rate": 0.0006,
+ "loss": 4.271645545959473,
+ "step": 6046
+ },
+ {
+ "epoch": 83.99213973799127,
+ "grad_norm": 0.014431594870984554,
+ "learning_rate": 0.0006,
+ "loss": 4.161383152008057,
+ "step": 6047
+ },
+ {
+ "epoch": 84.0,
+ "grad_norm": 0.016268927603960037,
+ "learning_rate": 0.0006,
+ "loss": 4.178324222564697,
+ "step": 6048
+ },
+ {
+ "epoch": 84.0,
+ "eval_loss": 4.6307549476623535,
+ "eval_runtime": 56.8521,
+ "eval_samples_per_second": 42.954,
+ "eval_steps_per_second": 1.354,
+ "step": 6048
+ },
+ {
+ "epoch": 84.01397379912663,
+ "grad_norm": 0.014864908531308174,
+ "learning_rate": 0.0006,
+ "loss": 4.121300220489502,
+ "step": 6049
+ },
+ {
+ "epoch": 84.02794759825328,
+ "grad_norm": 0.014863367192447186,
+ "learning_rate": 0.0006,
+ "loss": 4.1650543212890625,
+ "step": 6050
+ },
+ {
+ "epoch": 84.04192139737991,
+ "grad_norm": 0.014379646629095078,
+ "learning_rate": 0.0006,
+ "loss": 4.0201239585876465,
+ "step": 6051
+ },
+ {
+ "epoch": 84.05589519650655,
+ "grad_norm": 0.014871201477944851,
+ "learning_rate": 0.0006,
+ "loss": 4.0820698738098145,
+ "step": 6052
+ },
+ {
+ "epoch": 84.06986899563319,
+ "grad_norm": 0.013085179962217808,
+ "learning_rate": 0.0006,
+ "loss": 4.223984718322754,
+ "step": 6053
+ },
+ {
+ "epoch": 84.08384279475983,
+ "grad_norm": 0.012485949322581291,
+ "learning_rate": 0.0006,
+ "loss": 4.182206153869629,
+ "step": 6054
+ },
+ {
+ "epoch": 84.09781659388646,
+ "grad_norm": 0.013653291389346123,
+ "learning_rate": 0.0006,
+ "loss": 4.142557144165039,
+ "step": 6055
+ },
+ {
+ "epoch": 84.1117903930131,
+ "grad_norm": 0.013930532149970531,
+ "learning_rate": 0.0006,
+ "loss": 4.13972282409668,
+ "step": 6056
+ },
+ {
+ "epoch": 84.12576419213974,
+ "grad_norm": 0.01297379657626152,
+ "learning_rate": 0.0006,
+ "loss": 4.144664764404297,
+ "step": 6057
+ },
+ {
+ "epoch": 84.13973799126637,
+ "grad_norm": 0.013210492208600044,
+ "learning_rate": 0.0006,
+ "loss": 4.253764629364014,
+ "step": 6058
+ },
+ {
+ "epoch": 84.15371179039302,
+ "grad_norm": 0.014650176279246807,
+ "learning_rate": 0.0006,
+ "loss": 4.246866226196289,
+ "step": 6059
+ },
+ {
+ "epoch": 84.16768558951965,
+ "grad_norm": 0.01482292078435421,
+ "learning_rate": 0.0006,
+ "loss": 4.140187740325928,
+ "step": 6060
+ },
+ {
+ "epoch": 84.18165938864628,
+ "grad_norm": 0.01592552848160267,
+ "learning_rate": 0.0006,
+ "loss": 4.170705318450928,
+ "step": 6061
+ },
+ {
+ "epoch": 84.19563318777293,
+ "grad_norm": 0.017152156680822372,
+ "learning_rate": 0.0006,
+ "loss": 4.126092433929443,
+ "step": 6062
+ },
+ {
+ "epoch": 84.20960698689956,
+ "grad_norm": 0.015880905091762543,
+ "learning_rate": 0.0006,
+ "loss": 4.161938190460205,
+ "step": 6063
+ },
+ {
+ "epoch": 84.2235807860262,
+ "grad_norm": 0.015451471321284771,
+ "learning_rate": 0.0006,
+ "loss": 4.263461112976074,
+ "step": 6064
+ },
+ {
+ "epoch": 84.23755458515284,
+ "grad_norm": 0.014365996234118938,
+ "learning_rate": 0.0006,
+ "loss": 4.14488410949707,
+ "step": 6065
+ },
+ {
+ "epoch": 84.25152838427948,
+ "grad_norm": 0.016364749521017075,
+ "learning_rate": 0.0006,
+ "loss": 4.112489700317383,
+ "step": 6066
+ },
+ {
+ "epoch": 84.26550218340611,
+ "grad_norm": 0.01791754551231861,
+ "learning_rate": 0.0006,
+ "loss": 4.133513450622559,
+ "step": 6067
+ },
+ {
+ "epoch": 84.27947598253274,
+ "grad_norm": 0.016609208658337593,
+ "learning_rate": 0.0006,
+ "loss": 4.201829433441162,
+ "step": 6068
+ },
+ {
+ "epoch": 84.29344978165939,
+ "grad_norm": 0.017587820068001747,
+ "learning_rate": 0.0006,
+ "loss": 4.203151702880859,
+ "step": 6069
+ },
+ {
+ "epoch": 84.30742358078602,
+ "grad_norm": 0.017047742381691933,
+ "learning_rate": 0.0006,
+ "loss": 4.077008247375488,
+ "step": 6070
+ },
+ {
+ "epoch": 84.32139737991267,
+ "grad_norm": 0.01524646207690239,
+ "learning_rate": 0.0006,
+ "loss": 4.198038101196289,
+ "step": 6071
+ },
+ {
+ "epoch": 84.3353711790393,
+ "grad_norm": 0.014196421951055527,
+ "learning_rate": 0.0006,
+ "loss": 4.164245128631592,
+ "step": 6072
+ },
+ {
+ "epoch": 84.34934497816595,
+ "grad_norm": 0.016291776672005653,
+ "learning_rate": 0.0006,
+ "loss": 4.150152206420898,
+ "step": 6073
+ },
+ {
+ "epoch": 84.36331877729258,
+ "grad_norm": 0.015627920627593994,
+ "learning_rate": 0.0006,
+ "loss": 4.110006332397461,
+ "step": 6074
+ },
+ {
+ "epoch": 84.37729257641921,
+ "grad_norm": 0.013251487165689468,
+ "learning_rate": 0.0006,
+ "loss": 4.101463317871094,
+ "step": 6075
+ },
+ {
+ "epoch": 84.39126637554585,
+ "grad_norm": 0.013601605780422688,
+ "learning_rate": 0.0006,
+ "loss": 4.191105842590332,
+ "step": 6076
+ },
+ {
+ "epoch": 84.40524017467249,
+ "grad_norm": 0.014878741465508938,
+ "learning_rate": 0.0006,
+ "loss": 4.153046131134033,
+ "step": 6077
+ },
+ {
+ "epoch": 84.41921397379913,
+ "grad_norm": 0.014783897437155247,
+ "learning_rate": 0.0006,
+ "loss": 4.362000465393066,
+ "step": 6078
+ },
+ {
+ "epoch": 84.43318777292576,
+ "grad_norm": 0.014221332035958767,
+ "learning_rate": 0.0006,
+ "loss": 4.144364833831787,
+ "step": 6079
+ },
+ {
+ "epoch": 84.44716157205241,
+ "grad_norm": 0.014911034144461155,
+ "learning_rate": 0.0006,
+ "loss": 4.088857650756836,
+ "step": 6080
+ },
+ {
+ "epoch": 84.46113537117904,
+ "grad_norm": 0.014924770221114159,
+ "learning_rate": 0.0006,
+ "loss": 4.1677446365356445,
+ "step": 6081
+ },
+ {
+ "epoch": 84.47510917030567,
+ "grad_norm": 0.014394409023225307,
+ "learning_rate": 0.0006,
+ "loss": 4.087103843688965,
+ "step": 6082
+ },
+ {
+ "epoch": 84.48908296943232,
+ "grad_norm": 0.013926676474511623,
+ "learning_rate": 0.0006,
+ "loss": 3.9972851276397705,
+ "step": 6083
+ },
+ {
+ "epoch": 84.50305676855895,
+ "grad_norm": 0.013069954700767994,
+ "learning_rate": 0.0006,
+ "loss": 3.998237133026123,
+ "step": 6084
+ },
+ {
+ "epoch": 84.5170305676856,
+ "grad_norm": 0.012748011387884617,
+ "learning_rate": 0.0006,
+ "loss": 4.195858955383301,
+ "step": 6085
+ },
+ {
+ "epoch": 84.53100436681223,
+ "grad_norm": 0.012278077192604542,
+ "learning_rate": 0.0006,
+ "loss": 4.102165222167969,
+ "step": 6086
+ },
+ {
+ "epoch": 84.54497816593886,
+ "grad_norm": 0.014259163290262222,
+ "learning_rate": 0.0006,
+ "loss": 4.092750549316406,
+ "step": 6087
+ },
+ {
+ "epoch": 84.5589519650655,
+ "grad_norm": 0.014553683809936047,
+ "learning_rate": 0.0006,
+ "loss": 4.1412763595581055,
+ "step": 6088
+ },
+ {
+ "epoch": 84.57292576419214,
+ "grad_norm": 0.013605243526399136,
+ "learning_rate": 0.0006,
+ "loss": 4.192173480987549,
+ "step": 6089
+ },
+ {
+ "epoch": 84.58689956331878,
+ "grad_norm": 0.014259358868002892,
+ "learning_rate": 0.0006,
+ "loss": 4.356374740600586,
+ "step": 6090
+ },
+ {
+ "epoch": 84.60087336244541,
+ "grad_norm": 0.014166025444865227,
+ "learning_rate": 0.0006,
+ "loss": 4.165982246398926,
+ "step": 6091
+ },
+ {
+ "epoch": 84.61484716157206,
+ "grad_norm": 0.012198803946375847,
+ "learning_rate": 0.0006,
+ "loss": 4.287566184997559,
+ "step": 6092
+ },
+ {
+ "epoch": 84.62882096069869,
+ "grad_norm": 0.013794238679111004,
+ "learning_rate": 0.0006,
+ "loss": 4.199882984161377,
+ "step": 6093
+ },
+ {
+ "epoch": 84.64279475982532,
+ "grad_norm": 0.014757642522454262,
+ "learning_rate": 0.0006,
+ "loss": 4.116573333740234,
+ "step": 6094
+ },
+ {
+ "epoch": 84.65676855895197,
+ "grad_norm": 0.014591689221560955,
+ "learning_rate": 0.0006,
+ "loss": 4.240896224975586,
+ "step": 6095
+ },
+ {
+ "epoch": 84.6707423580786,
+ "grad_norm": 0.015217579901218414,
+ "learning_rate": 0.0006,
+ "loss": 4.257142066955566,
+ "step": 6096
+ },
+ {
+ "epoch": 84.68471615720524,
+ "grad_norm": 0.015371371991932392,
+ "learning_rate": 0.0006,
+ "loss": 4.181037425994873,
+ "step": 6097
+ },
+ {
+ "epoch": 84.69868995633188,
+ "grad_norm": 0.014412983320653439,
+ "learning_rate": 0.0006,
+ "loss": 4.2207794189453125,
+ "step": 6098
+ },
+ {
+ "epoch": 84.71266375545852,
+ "grad_norm": 0.013773671351373196,
+ "learning_rate": 0.0006,
+ "loss": 4.211123466491699,
+ "step": 6099
+ },
+ {
+ "epoch": 84.72663755458515,
+ "grad_norm": 0.012838046066462994,
+ "learning_rate": 0.0006,
+ "loss": 4.175459384918213,
+ "step": 6100
+ },
+ {
+ "epoch": 84.74061135371178,
+ "grad_norm": 0.014094929210841656,
+ "learning_rate": 0.0006,
+ "loss": 4.1724958419799805,
+ "step": 6101
+ },
+ {
+ "epoch": 84.75458515283843,
+ "grad_norm": 0.014573262073099613,
+ "learning_rate": 0.0006,
+ "loss": 4.15885066986084,
+ "step": 6102
+ },
+ {
+ "epoch": 84.76855895196506,
+ "grad_norm": 0.011964819394052029,
+ "learning_rate": 0.0006,
+ "loss": 3.9804892539978027,
+ "step": 6103
+ },
+ {
+ "epoch": 84.78253275109171,
+ "grad_norm": 0.014140539802610874,
+ "learning_rate": 0.0006,
+ "loss": 4.203708648681641,
+ "step": 6104
+ },
+ {
+ "epoch": 84.79650655021834,
+ "grad_norm": 0.016618940979242325,
+ "learning_rate": 0.0006,
+ "loss": 4.147684097290039,
+ "step": 6105
+ },
+ {
+ "epoch": 84.81048034934499,
+ "grad_norm": 0.015094385482370853,
+ "learning_rate": 0.0006,
+ "loss": 4.392223834991455,
+ "step": 6106
+ },
+ {
+ "epoch": 84.82445414847162,
+ "grad_norm": 0.014386717230081558,
+ "learning_rate": 0.0006,
+ "loss": 4.253332138061523,
+ "step": 6107
+ },
+ {
+ "epoch": 84.83842794759825,
+ "grad_norm": 0.013877755962312222,
+ "learning_rate": 0.0006,
+ "loss": 4.075056076049805,
+ "step": 6108
+ },
+ {
+ "epoch": 84.8524017467249,
+ "grad_norm": 0.014062155969440937,
+ "learning_rate": 0.0006,
+ "loss": 4.072048664093018,
+ "step": 6109
+ },
+ {
+ "epoch": 84.86637554585153,
+ "grad_norm": 0.014445619657635689,
+ "learning_rate": 0.0006,
+ "loss": 4.151832580566406,
+ "step": 6110
+ },
+ {
+ "epoch": 84.88034934497817,
+ "grad_norm": 0.015116363763809204,
+ "learning_rate": 0.0006,
+ "loss": 4.088900566101074,
+ "step": 6111
+ },
+ {
+ "epoch": 84.8943231441048,
+ "grad_norm": 0.014425474219024181,
+ "learning_rate": 0.0006,
+ "loss": 4.200483322143555,
+ "step": 6112
+ },
+ {
+ "epoch": 84.90829694323143,
+ "grad_norm": 0.01449573878198862,
+ "learning_rate": 0.0006,
+ "loss": 4.239742279052734,
+ "step": 6113
+ },
+ {
+ "epoch": 84.92227074235808,
+ "grad_norm": 0.015698261559009552,
+ "learning_rate": 0.0006,
+ "loss": 4.197391510009766,
+ "step": 6114
+ },
+ {
+ "epoch": 84.93624454148471,
+ "grad_norm": 0.014545733109116554,
+ "learning_rate": 0.0006,
+ "loss": 4.035508155822754,
+ "step": 6115
+ },
+ {
+ "epoch": 84.95021834061136,
+ "grad_norm": 0.012796483002603054,
+ "learning_rate": 0.0006,
+ "loss": 4.195248126983643,
+ "step": 6116
+ },
+ {
+ "epoch": 84.96419213973799,
+ "grad_norm": 0.01264485064893961,
+ "learning_rate": 0.0006,
+ "loss": 4.158118724822998,
+ "step": 6117
+ },
+ {
+ "epoch": 84.97816593886463,
+ "grad_norm": 0.013695276342332363,
+ "learning_rate": 0.0006,
+ "loss": 4.255288124084473,
+ "step": 6118
+ },
+ {
+ "epoch": 84.99213973799127,
+ "grad_norm": 0.013459564186632633,
+ "learning_rate": 0.0006,
+ "loss": 4.204586505889893,
+ "step": 6119
+ },
+ {
+ "epoch": 85.0,
+ "grad_norm": 0.014312652871012688,
+ "learning_rate": 0.0006,
+ "loss": 4.167879104614258,
+ "step": 6120
+ },
+ {
+ "epoch": 85.0,
+ "eval_loss": 4.608152389526367,
+ "eval_runtime": 57.1903,
+ "eval_samples_per_second": 42.7,
+ "eval_steps_per_second": 1.346,
+ "step": 6120
+ },
+ {
+ "epoch": 85.01397379912663,
+ "grad_norm": 0.013918368145823479,
+ "learning_rate": 0.0006,
+ "loss": 4.129991054534912,
+ "step": 6121
+ },
+ {
+ "epoch": 85.02794759825328,
+ "grad_norm": 0.01604384370148182,
+ "learning_rate": 0.0006,
+ "loss": 4.233926773071289,
+ "step": 6122
+ },
+ {
+ "epoch": 85.04192139737991,
+ "grad_norm": 0.016565652564167976,
+ "learning_rate": 0.0006,
+ "loss": 4.129604339599609,
+ "step": 6123
+ },
+ {
+ "epoch": 85.05589519650655,
+ "grad_norm": 0.0149751678109169,
+ "learning_rate": 0.0006,
+ "loss": 4.165839195251465,
+ "step": 6124
+ },
+ {
+ "epoch": 85.06986899563319,
+ "grad_norm": 0.013812461867928505,
+ "learning_rate": 0.0006,
+ "loss": 4.153792858123779,
+ "step": 6125
+ },
+ {
+ "epoch": 85.08384279475983,
+ "grad_norm": 0.014432745985686779,
+ "learning_rate": 0.0006,
+ "loss": 4.170895576477051,
+ "step": 6126
+ },
+ {
+ "epoch": 85.09781659388646,
+ "grad_norm": 0.015208868309855461,
+ "learning_rate": 0.0006,
+ "loss": 4.10639762878418,
+ "step": 6127
+ },
+ {
+ "epoch": 85.1117903930131,
+ "grad_norm": 0.013209272176027298,
+ "learning_rate": 0.0006,
+ "loss": 4.200364112854004,
+ "step": 6128
+ },
+ {
+ "epoch": 85.12576419213974,
+ "grad_norm": 0.014807002618908882,
+ "learning_rate": 0.0006,
+ "loss": 4.172666549682617,
+ "step": 6129
+ },
+ {
+ "epoch": 85.13973799126637,
+ "grad_norm": 0.015662945806980133,
+ "learning_rate": 0.0006,
+ "loss": 4.145219326019287,
+ "step": 6130
+ },
+ {
+ "epoch": 85.15371179039302,
+ "grad_norm": 0.014598497189581394,
+ "learning_rate": 0.0006,
+ "loss": 4.275618553161621,
+ "step": 6131
+ },
+ {
+ "epoch": 85.16768558951965,
+ "grad_norm": 0.013778576627373695,
+ "learning_rate": 0.0006,
+ "loss": 4.204716682434082,
+ "step": 6132
+ },
+ {
+ "epoch": 85.18165938864628,
+ "grad_norm": 0.015762265771627426,
+ "learning_rate": 0.0006,
+ "loss": 4.145524024963379,
+ "step": 6133
+ },
+ {
+ "epoch": 85.19563318777293,
+ "grad_norm": 0.015579195693135262,
+ "learning_rate": 0.0006,
+ "loss": 4.150625228881836,
+ "step": 6134
+ },
+ {
+ "epoch": 85.20960698689956,
+ "grad_norm": 0.016641216352581978,
+ "learning_rate": 0.0006,
+ "loss": 4.141819953918457,
+ "step": 6135
+ },
+ {
+ "epoch": 85.2235807860262,
+ "grad_norm": 0.01632271707057953,
+ "learning_rate": 0.0006,
+ "loss": 4.030620574951172,
+ "step": 6136
+ },
+ {
+ "epoch": 85.23755458515284,
+ "grad_norm": 0.015564918518066406,
+ "learning_rate": 0.0006,
+ "loss": 4.147753715515137,
+ "step": 6137
+ },
+ {
+ "epoch": 85.25152838427948,
+ "grad_norm": 0.0161129217594862,
+ "learning_rate": 0.0006,
+ "loss": 4.128524303436279,
+ "step": 6138
+ },
+ {
+ "epoch": 85.26550218340611,
+ "grad_norm": 0.017766550183296204,
+ "learning_rate": 0.0006,
+ "loss": 4.134856224060059,
+ "step": 6139
+ },
+ {
+ "epoch": 85.27947598253274,
+ "grad_norm": 0.01869054324924946,
+ "learning_rate": 0.0006,
+ "loss": 4.231565475463867,
+ "step": 6140
+ },
+ {
+ "epoch": 85.29344978165939,
+ "grad_norm": 0.01917010359466076,
+ "learning_rate": 0.0006,
+ "loss": 4.21175479888916,
+ "step": 6141
+ },
+ {
+ "epoch": 85.30742358078602,
+ "grad_norm": 0.018116485327482224,
+ "learning_rate": 0.0006,
+ "loss": 4.122644424438477,
+ "step": 6142
+ },
+ {
+ "epoch": 85.32139737991267,
+ "grad_norm": 0.015949107706546783,
+ "learning_rate": 0.0006,
+ "loss": 4.135458946228027,
+ "step": 6143
+ },
+ {
+ "epoch": 85.3353711790393,
+ "grad_norm": 0.014783683232963085,
+ "learning_rate": 0.0006,
+ "loss": 4.118994235992432,
+ "step": 6144
+ },
+ {
+ "epoch": 85.34934497816595,
+ "grad_norm": 0.015406692400574684,
+ "learning_rate": 0.0006,
+ "loss": 4.157268524169922,
+ "step": 6145
+ },
+ {
+ "epoch": 85.36331877729258,
+ "grad_norm": 0.013893205672502518,
+ "learning_rate": 0.0006,
+ "loss": 4.056166648864746,
+ "step": 6146
+ },
+ {
+ "epoch": 85.37729257641921,
+ "grad_norm": 0.014380923472344875,
+ "learning_rate": 0.0006,
+ "loss": 4.206905364990234,
+ "step": 6147
+ },
+ {
+ "epoch": 85.39126637554585,
+ "grad_norm": 0.015669258311390877,
+ "learning_rate": 0.0006,
+ "loss": 4.213881492614746,
+ "step": 6148
+ },
+ {
+ "epoch": 85.40524017467249,
+ "grad_norm": 0.01553428452461958,
+ "learning_rate": 0.0006,
+ "loss": 4.1814374923706055,
+ "step": 6149
+ },
+ {
+ "epoch": 85.41921397379913,
+ "grad_norm": 0.01697700284421444,
+ "learning_rate": 0.0006,
+ "loss": 4.083939552307129,
+ "step": 6150
+ },
+ {
+ "epoch": 85.43318777292576,
+ "grad_norm": 0.016639070585370064,
+ "learning_rate": 0.0006,
+ "loss": 4.279311180114746,
+ "step": 6151
+ },
+ {
+ "epoch": 85.44716157205241,
+ "grad_norm": 0.016285281628370285,
+ "learning_rate": 0.0006,
+ "loss": 4.244845390319824,
+ "step": 6152
+ },
+ {
+ "epoch": 85.46113537117904,
+ "grad_norm": 0.016071317717432976,
+ "learning_rate": 0.0006,
+ "loss": 4.089814186096191,
+ "step": 6153
+ },
+ {
+ "epoch": 85.47510917030567,
+ "grad_norm": 0.01640418916940689,
+ "learning_rate": 0.0006,
+ "loss": 4.168088912963867,
+ "step": 6154
+ },
+ {
+ "epoch": 85.48908296943232,
+ "grad_norm": 0.014781414531171322,
+ "learning_rate": 0.0006,
+ "loss": 4.196092128753662,
+ "step": 6155
+ },
+ {
+ "epoch": 85.50305676855895,
+ "grad_norm": 0.013393213972449303,
+ "learning_rate": 0.0006,
+ "loss": 4.193429946899414,
+ "step": 6156
+ },
+ {
+ "epoch": 85.5170305676856,
+ "grad_norm": 0.01467406190931797,
+ "learning_rate": 0.0006,
+ "loss": 4.216714859008789,
+ "step": 6157
+ },
+ {
+ "epoch": 85.53100436681223,
+ "grad_norm": 0.014950765296816826,
+ "learning_rate": 0.0006,
+ "loss": 4.093854904174805,
+ "step": 6158
+ },
+ {
+ "epoch": 85.54497816593886,
+ "grad_norm": 0.014655301347374916,
+ "learning_rate": 0.0006,
+ "loss": 4.09483528137207,
+ "step": 6159
+ },
+ {
+ "epoch": 85.5589519650655,
+ "grad_norm": 0.01425520796328783,
+ "learning_rate": 0.0006,
+ "loss": 4.090396881103516,
+ "step": 6160
+ },
+ {
+ "epoch": 85.57292576419214,
+ "grad_norm": 0.01403278298676014,
+ "learning_rate": 0.0006,
+ "loss": 4.088724136352539,
+ "step": 6161
+ },
+ {
+ "epoch": 85.58689956331878,
+ "grad_norm": 0.01364967692643404,
+ "learning_rate": 0.0006,
+ "loss": 4.095558166503906,
+ "step": 6162
+ },
+ {
+ "epoch": 85.60087336244541,
+ "grad_norm": 0.013085267506539822,
+ "learning_rate": 0.0006,
+ "loss": 4.163962364196777,
+ "step": 6163
+ },
+ {
+ "epoch": 85.61484716157206,
+ "grad_norm": 0.0136906448751688,
+ "learning_rate": 0.0006,
+ "loss": 4.197134017944336,
+ "step": 6164
+ },
+ {
+ "epoch": 85.62882096069869,
+ "grad_norm": 0.013472451828420162,
+ "learning_rate": 0.0006,
+ "loss": 4.1578369140625,
+ "step": 6165
+ },
+ {
+ "epoch": 85.64279475982532,
+ "grad_norm": 0.012329773046076298,
+ "learning_rate": 0.0006,
+ "loss": 4.231907844543457,
+ "step": 6166
+ },
+ {
+ "epoch": 85.65676855895197,
+ "grad_norm": 0.011977200396358967,
+ "learning_rate": 0.0006,
+ "loss": 4.1552958488464355,
+ "step": 6167
+ },
+ {
+ "epoch": 85.6707423580786,
+ "grad_norm": 0.013049956411123276,
+ "learning_rate": 0.0006,
+ "loss": 4.17852783203125,
+ "step": 6168
+ },
+ {
+ "epoch": 85.68471615720524,
+ "grad_norm": 0.011440741829574108,
+ "learning_rate": 0.0006,
+ "loss": 4.144646644592285,
+ "step": 6169
+ },
+ {
+ "epoch": 85.69868995633188,
+ "grad_norm": 0.012472936883568764,
+ "learning_rate": 0.0006,
+ "loss": 4.2306413650512695,
+ "step": 6170
+ },
+ {
+ "epoch": 85.71266375545852,
+ "grad_norm": 0.013015356846153736,
+ "learning_rate": 0.0006,
+ "loss": 4.193439960479736,
+ "step": 6171
+ },
+ {
+ "epoch": 85.72663755458515,
+ "grad_norm": 0.013832319527864456,
+ "learning_rate": 0.0006,
+ "loss": 4.18073034286499,
+ "step": 6172
+ },
+ {
+ "epoch": 85.74061135371178,
+ "grad_norm": 0.014613982290029526,
+ "learning_rate": 0.0006,
+ "loss": 4.223531246185303,
+ "step": 6173
+ },
+ {
+ "epoch": 85.75458515283843,
+ "grad_norm": 0.014209083281457424,
+ "learning_rate": 0.0006,
+ "loss": 4.142666816711426,
+ "step": 6174
+ },
+ {
+ "epoch": 85.76855895196506,
+ "grad_norm": 0.013396669179201126,
+ "learning_rate": 0.0006,
+ "loss": 4.185786247253418,
+ "step": 6175
+ },
+ {
+ "epoch": 85.78253275109171,
+ "grad_norm": 0.01364929135888815,
+ "learning_rate": 0.0006,
+ "loss": 4.125226974487305,
+ "step": 6176
+ },
+ {
+ "epoch": 85.79650655021834,
+ "grad_norm": 0.012561003677546978,
+ "learning_rate": 0.0006,
+ "loss": 4.084519386291504,
+ "step": 6177
+ },
+ {
+ "epoch": 85.81048034934499,
+ "grad_norm": 0.013508481904864311,
+ "learning_rate": 0.0006,
+ "loss": 4.193173885345459,
+ "step": 6178
+ },
+ {
+ "epoch": 85.82445414847162,
+ "grad_norm": 0.014154613018035889,
+ "learning_rate": 0.0006,
+ "loss": 4.257272720336914,
+ "step": 6179
+ },
+ {
+ "epoch": 85.83842794759825,
+ "grad_norm": 0.012478531338274479,
+ "learning_rate": 0.0006,
+ "loss": 4.161380767822266,
+ "step": 6180
+ },
+ {
+ "epoch": 85.8524017467249,
+ "grad_norm": 0.013468451797962189,
+ "learning_rate": 0.0006,
+ "loss": 4.121395111083984,
+ "step": 6181
+ },
+ {
+ "epoch": 85.86637554585153,
+ "grad_norm": 0.015155194327235222,
+ "learning_rate": 0.0006,
+ "loss": 4.242815971374512,
+ "step": 6182
+ },
+ {
+ "epoch": 85.88034934497817,
+ "grad_norm": 0.016223272308707237,
+ "learning_rate": 0.0006,
+ "loss": 4.176027297973633,
+ "step": 6183
+ },
+ {
+ "epoch": 85.8943231441048,
+ "grad_norm": 0.013741295784711838,
+ "learning_rate": 0.0006,
+ "loss": 4.131257057189941,
+ "step": 6184
+ },
+ {
+ "epoch": 85.90829694323143,
+ "grad_norm": 0.014982067979872227,
+ "learning_rate": 0.0006,
+ "loss": 4.237896919250488,
+ "step": 6185
+ },
+ {
+ "epoch": 85.92227074235808,
+ "grad_norm": 0.018197322264313698,
+ "learning_rate": 0.0006,
+ "loss": 4.154962539672852,
+ "step": 6186
+ },
+ {
+ "epoch": 85.93624454148471,
+ "grad_norm": 0.01971200481057167,
+ "learning_rate": 0.0006,
+ "loss": 4.16819429397583,
+ "step": 6187
+ },
+ {
+ "epoch": 85.95021834061136,
+ "grad_norm": 0.018008720129728317,
+ "learning_rate": 0.0006,
+ "loss": 4.146164417266846,
+ "step": 6188
+ },
+ {
+ "epoch": 85.96419213973799,
+ "grad_norm": 0.015859264880418777,
+ "learning_rate": 0.0006,
+ "loss": 4.208099365234375,
+ "step": 6189
+ },
+ {
+ "epoch": 85.97816593886463,
+ "grad_norm": 0.013846252113580704,
+ "learning_rate": 0.0006,
+ "loss": 4.1881184577941895,
+ "step": 6190
+ },
+ {
+ "epoch": 85.99213973799127,
+ "grad_norm": 0.018641939386725426,
+ "learning_rate": 0.0006,
+ "loss": 4.127632141113281,
+ "step": 6191
+ },
+ {
+ "epoch": 86.0,
+ "grad_norm": 0.02385227382183075,
+ "learning_rate": 0.0006,
+ "loss": 4.281900405883789,
+ "step": 6192
+ },
+ {
+ "epoch": 86.0,
+ "eval_loss": 4.625458717346191,
+ "eval_runtime": 57.0634,
+ "eval_samples_per_second": 42.795,
+ "eval_steps_per_second": 1.349,
+ "step": 6192
+ },
+ {
+ "epoch": 86.01397379912663,
+ "grad_norm": 0.019032124429941177,
+ "learning_rate": 0.0006,
+ "loss": 4.2163190841674805,
+ "step": 6193
+ },
+ {
+ "epoch": 86.02794759825328,
+ "grad_norm": 0.018049925565719604,
+ "learning_rate": 0.0006,
+ "loss": 4.1534037590026855,
+ "step": 6194
+ },
+ {
+ "epoch": 86.04192139737991,
+ "grad_norm": 0.019116360694169998,
+ "learning_rate": 0.0006,
+ "loss": 4.212411403656006,
+ "step": 6195
+ },
+ {
+ "epoch": 86.05589519650655,
+ "grad_norm": 0.018967336043715477,
+ "learning_rate": 0.0006,
+ "loss": 4.190918922424316,
+ "step": 6196
+ },
+ {
+ "epoch": 86.06986899563319,
+ "grad_norm": 0.01945611648261547,
+ "learning_rate": 0.0006,
+ "loss": 4.18459415435791,
+ "step": 6197
+ },
+ {
+ "epoch": 86.08384279475983,
+ "grad_norm": 0.018104076385498047,
+ "learning_rate": 0.0006,
+ "loss": 4.0942816734313965,
+ "step": 6198
+ },
+ {
+ "epoch": 86.09781659388646,
+ "grad_norm": 0.01815992221236229,
+ "learning_rate": 0.0006,
+ "loss": 4.011547565460205,
+ "step": 6199
+ },
+ {
+ "epoch": 86.1117903930131,
+ "grad_norm": 0.017574431374669075,
+ "learning_rate": 0.0006,
+ "loss": 4.051593780517578,
+ "step": 6200
+ },
+ {
+ "epoch": 86.12576419213974,
+ "grad_norm": 0.01516508124768734,
+ "learning_rate": 0.0006,
+ "loss": 4.239673614501953,
+ "step": 6201
+ },
+ {
+ "epoch": 86.13973799126637,
+ "grad_norm": 0.015689702704548836,
+ "learning_rate": 0.0006,
+ "loss": 4.1085734367370605,
+ "step": 6202
+ },
+ {
+ "epoch": 86.15371179039302,
+ "grad_norm": 0.01596900261938572,
+ "learning_rate": 0.0006,
+ "loss": 4.1509599685668945,
+ "step": 6203
+ },
+ {
+ "epoch": 86.16768558951965,
+ "grad_norm": 0.01550096832215786,
+ "learning_rate": 0.0006,
+ "loss": 4.160238265991211,
+ "step": 6204
+ },
+ {
+ "epoch": 86.18165938864628,
+ "grad_norm": 0.014601529575884342,
+ "learning_rate": 0.0006,
+ "loss": 4.029219150543213,
+ "step": 6205
+ },
+ {
+ "epoch": 86.19563318777293,
+ "grad_norm": 0.014006740413606167,
+ "learning_rate": 0.0006,
+ "loss": 4.117501258850098,
+ "step": 6206
+ },
+ {
+ "epoch": 86.20960698689956,
+ "grad_norm": 0.013191957026720047,
+ "learning_rate": 0.0006,
+ "loss": 4.252002716064453,
+ "step": 6207
+ },
+ {
+ "epoch": 86.2235807860262,
+ "grad_norm": 0.013031814247369766,
+ "learning_rate": 0.0006,
+ "loss": 4.209517478942871,
+ "step": 6208
+ },
+ {
+ "epoch": 86.23755458515284,
+ "grad_norm": 0.013605714775621891,
+ "learning_rate": 0.0006,
+ "loss": 4.176309585571289,
+ "step": 6209
+ },
+ {
+ "epoch": 86.25152838427948,
+ "grad_norm": 0.012753087095916271,
+ "learning_rate": 0.0006,
+ "loss": 4.097990036010742,
+ "step": 6210
+ },
+ {
+ "epoch": 86.26550218340611,
+ "grad_norm": 0.012940628454089165,
+ "learning_rate": 0.0006,
+ "loss": 4.085589408874512,
+ "step": 6211
+ },
+ {
+ "epoch": 86.27947598253274,
+ "grad_norm": 0.014286619611084461,
+ "learning_rate": 0.0006,
+ "loss": 4.199261665344238,
+ "step": 6212
+ },
+ {
+ "epoch": 86.29344978165939,
+ "grad_norm": 0.014643525704741478,
+ "learning_rate": 0.0006,
+ "loss": 4.23718786239624,
+ "step": 6213
+ },
+ {
+ "epoch": 86.30742358078602,
+ "grad_norm": 0.015182147733867168,
+ "learning_rate": 0.0006,
+ "loss": 4.219855308532715,
+ "step": 6214
+ },
+ {
+ "epoch": 86.32139737991267,
+ "grad_norm": 0.01530102826654911,
+ "learning_rate": 0.0006,
+ "loss": 4.190094947814941,
+ "step": 6215
+ },
+ {
+ "epoch": 86.3353711790393,
+ "grad_norm": 0.015492603182792664,
+ "learning_rate": 0.0006,
+ "loss": 4.153780460357666,
+ "step": 6216
+ },
+ {
+ "epoch": 86.34934497816595,
+ "grad_norm": 0.015339907258749008,
+ "learning_rate": 0.0006,
+ "loss": 4.182865619659424,
+ "step": 6217
+ },
+ {
+ "epoch": 86.36331877729258,
+ "grad_norm": 0.015576880425214767,
+ "learning_rate": 0.0006,
+ "loss": 4.196417331695557,
+ "step": 6218
+ },
+ {
+ "epoch": 86.37729257641921,
+ "grad_norm": 0.014805737882852554,
+ "learning_rate": 0.0006,
+ "loss": 4.265018463134766,
+ "step": 6219
+ },
+ {
+ "epoch": 86.39126637554585,
+ "grad_norm": 0.01597665622830391,
+ "learning_rate": 0.0006,
+ "loss": 4.2864837646484375,
+ "step": 6220
+ },
+ {
+ "epoch": 86.40524017467249,
+ "grad_norm": 0.01693643443286419,
+ "learning_rate": 0.0006,
+ "loss": 4.118192672729492,
+ "step": 6221
+ },
+ {
+ "epoch": 86.41921397379913,
+ "grad_norm": 0.016648951917886734,
+ "learning_rate": 0.0006,
+ "loss": 4.23712682723999,
+ "step": 6222
+ },
+ {
+ "epoch": 86.43318777292576,
+ "grad_norm": 0.014809766784310341,
+ "learning_rate": 0.0006,
+ "loss": 4.251391410827637,
+ "step": 6223
+ },
+ {
+ "epoch": 86.44716157205241,
+ "grad_norm": 0.014333373866975307,
+ "learning_rate": 0.0006,
+ "loss": 4.17243766784668,
+ "step": 6224
+ },
+ {
+ "epoch": 86.46113537117904,
+ "grad_norm": 0.013912711292505264,
+ "learning_rate": 0.0006,
+ "loss": 4.250836372375488,
+ "step": 6225
+ },
+ {
+ "epoch": 86.47510917030567,
+ "grad_norm": 0.014821341261267662,
+ "learning_rate": 0.0006,
+ "loss": 4.073995590209961,
+ "step": 6226
+ },
+ {
+ "epoch": 86.48908296943232,
+ "grad_norm": 0.016634922474622726,
+ "learning_rate": 0.0006,
+ "loss": 4.155867576599121,
+ "step": 6227
+ },
+ {
+ "epoch": 86.50305676855895,
+ "grad_norm": 0.01748695969581604,
+ "learning_rate": 0.0006,
+ "loss": 4.139344692230225,
+ "step": 6228
+ },
+ {
+ "epoch": 86.5170305676856,
+ "grad_norm": 0.01569363661110401,
+ "learning_rate": 0.0006,
+ "loss": 4.23923397064209,
+ "step": 6229
+ },
+ {
+ "epoch": 86.53100436681223,
+ "grad_norm": 0.014037121087312698,
+ "learning_rate": 0.0006,
+ "loss": 4.200334072113037,
+ "step": 6230
+ },
+ {
+ "epoch": 86.54497816593886,
+ "grad_norm": 0.01401279866695404,
+ "learning_rate": 0.0006,
+ "loss": 4.035246849060059,
+ "step": 6231
+ },
+ {
+ "epoch": 86.5589519650655,
+ "grad_norm": 0.014603731222450733,
+ "learning_rate": 0.0006,
+ "loss": 4.141385555267334,
+ "step": 6232
+ },
+ {
+ "epoch": 86.57292576419214,
+ "grad_norm": 0.015113704837858677,
+ "learning_rate": 0.0006,
+ "loss": 4.130824089050293,
+ "step": 6233
+ },
+ {
+ "epoch": 86.58689956331878,
+ "grad_norm": 0.01588696800172329,
+ "learning_rate": 0.0006,
+ "loss": 4.05495023727417,
+ "step": 6234
+ },
+ {
+ "epoch": 86.60087336244541,
+ "grad_norm": 0.01593640074133873,
+ "learning_rate": 0.0006,
+ "loss": 4.22935152053833,
+ "step": 6235
+ },
+ {
+ "epoch": 86.61484716157206,
+ "grad_norm": 0.014292445965111256,
+ "learning_rate": 0.0006,
+ "loss": 4.309839248657227,
+ "step": 6236
+ },
+ {
+ "epoch": 86.62882096069869,
+ "grad_norm": 0.015471610240638256,
+ "learning_rate": 0.0006,
+ "loss": 4.268278121948242,
+ "step": 6237
+ },
+ {
+ "epoch": 86.64279475982532,
+ "grad_norm": 0.016643544659018517,
+ "learning_rate": 0.0006,
+ "loss": 4.135109901428223,
+ "step": 6238
+ },
+ {
+ "epoch": 86.65676855895197,
+ "grad_norm": 0.014733029529452324,
+ "learning_rate": 0.0006,
+ "loss": 4.060197830200195,
+ "step": 6239
+ },
+ {
+ "epoch": 86.6707423580786,
+ "grad_norm": 0.014869216829538345,
+ "learning_rate": 0.0006,
+ "loss": 4.126917362213135,
+ "step": 6240
+ },
+ {
+ "epoch": 86.68471615720524,
+ "grad_norm": 0.01608203910291195,
+ "learning_rate": 0.0006,
+ "loss": 4.183716773986816,
+ "step": 6241
+ },
+ {
+ "epoch": 86.69868995633188,
+ "grad_norm": 0.016178175806999207,
+ "learning_rate": 0.0006,
+ "loss": 4.118875503540039,
+ "step": 6242
+ },
+ {
+ "epoch": 86.71266375545852,
+ "grad_norm": 0.014576597139239311,
+ "learning_rate": 0.0006,
+ "loss": 4.261541366577148,
+ "step": 6243
+ },
+ {
+ "epoch": 86.72663755458515,
+ "grad_norm": 0.014466420747339725,
+ "learning_rate": 0.0006,
+ "loss": 4.307284355163574,
+ "step": 6244
+ },
+ {
+ "epoch": 86.74061135371178,
+ "grad_norm": 0.014042644761502743,
+ "learning_rate": 0.0006,
+ "loss": 4.2729997634887695,
+ "step": 6245
+ },
+ {
+ "epoch": 86.75458515283843,
+ "grad_norm": 0.014589524827897549,
+ "learning_rate": 0.0006,
+ "loss": 4.072981834411621,
+ "step": 6246
+ },
+ {
+ "epoch": 86.76855895196506,
+ "grad_norm": 0.018765483051538467,
+ "learning_rate": 0.0006,
+ "loss": 4.264769554138184,
+ "step": 6247
+ },
+ {
+ "epoch": 86.78253275109171,
+ "grad_norm": 0.019397731870412827,
+ "learning_rate": 0.0006,
+ "loss": 4.183379173278809,
+ "step": 6248
+ },
+ {
+ "epoch": 86.79650655021834,
+ "grad_norm": 0.017525402829051018,
+ "learning_rate": 0.0006,
+ "loss": 4.1494245529174805,
+ "step": 6249
+ },
+ {
+ "epoch": 86.81048034934499,
+ "grad_norm": 0.016310200095176697,
+ "learning_rate": 0.0006,
+ "loss": 4.149104118347168,
+ "step": 6250
+ },
+ {
+ "epoch": 86.82445414847162,
+ "grad_norm": 0.01506423857063055,
+ "learning_rate": 0.0006,
+ "loss": 4.22091817855835,
+ "step": 6251
+ },
+ {
+ "epoch": 86.83842794759825,
+ "grad_norm": 0.014200284145772457,
+ "learning_rate": 0.0006,
+ "loss": 4.177569389343262,
+ "step": 6252
+ },
+ {
+ "epoch": 86.8524017467249,
+ "grad_norm": 0.015016552060842514,
+ "learning_rate": 0.0006,
+ "loss": 4.089859962463379,
+ "step": 6253
+ },
+ {
+ "epoch": 86.86637554585153,
+ "grad_norm": 0.016054971143603325,
+ "learning_rate": 0.0006,
+ "loss": 4.157166004180908,
+ "step": 6254
+ },
+ {
+ "epoch": 86.88034934497817,
+ "grad_norm": 0.014388521201908588,
+ "learning_rate": 0.0006,
+ "loss": 4.1437578201293945,
+ "step": 6255
+ },
+ {
+ "epoch": 86.8943231441048,
+ "grad_norm": 0.015705369412899017,
+ "learning_rate": 0.0006,
+ "loss": 4.045689582824707,
+ "step": 6256
+ },
+ {
+ "epoch": 86.90829694323143,
+ "grad_norm": 0.014192395843565464,
+ "learning_rate": 0.0006,
+ "loss": 4.098793029785156,
+ "step": 6257
+ },
+ {
+ "epoch": 86.92227074235808,
+ "grad_norm": 0.013751034624874592,
+ "learning_rate": 0.0006,
+ "loss": 3.989920139312744,
+ "step": 6258
+ },
+ {
+ "epoch": 86.93624454148471,
+ "grad_norm": 0.01582193188369274,
+ "learning_rate": 0.0006,
+ "loss": 4.283449172973633,
+ "step": 6259
+ },
+ {
+ "epoch": 86.95021834061136,
+ "grad_norm": 0.01476013008505106,
+ "learning_rate": 0.0006,
+ "loss": 4.221401691436768,
+ "step": 6260
+ },
+ {
+ "epoch": 86.96419213973799,
+ "grad_norm": 0.01346337515860796,
+ "learning_rate": 0.0006,
+ "loss": 4.05896520614624,
+ "step": 6261
+ },
+ {
+ "epoch": 86.97816593886463,
+ "grad_norm": 0.012899307534098625,
+ "learning_rate": 0.0006,
+ "loss": 4.159560203552246,
+ "step": 6262
+ },
+ {
+ "epoch": 86.99213973799127,
+ "grad_norm": 0.013150574639439583,
+ "learning_rate": 0.0006,
+ "loss": 4.217330455780029,
+ "step": 6263
+ },
+ {
+ "epoch": 87.0,
+ "grad_norm": 0.014940370805561543,
+ "learning_rate": 0.0006,
+ "loss": 4.1611833572387695,
+ "step": 6264
+ },
+ {
+ "epoch": 87.0,
+ "eval_loss": 4.666043758392334,
+ "eval_runtime": 56.5978,
+ "eval_samples_per_second": 43.147,
+ "eval_steps_per_second": 1.36,
+ "step": 6264
+ },
+ {
+ "epoch": 87.01397379912663,
+ "grad_norm": 0.013721180148422718,
+ "learning_rate": 0.0006,
+ "loss": 4.142454147338867,
+ "step": 6265
+ },
+ {
+ "epoch": 87.02794759825328,
+ "grad_norm": 0.014805255457758904,
+ "learning_rate": 0.0006,
+ "loss": 4.121677875518799,
+ "step": 6266
+ },
+ {
+ "epoch": 87.04192139737991,
+ "grad_norm": 0.015576514415442944,
+ "learning_rate": 0.0006,
+ "loss": 4.1188435554504395,
+ "step": 6267
+ },
+ {
+ "epoch": 87.05589519650655,
+ "grad_norm": 0.01605502888560295,
+ "learning_rate": 0.0006,
+ "loss": 4.255428314208984,
+ "step": 6268
+ },
+ {
+ "epoch": 87.06986899563319,
+ "grad_norm": 0.014933343045413494,
+ "learning_rate": 0.0006,
+ "loss": 4.122596740722656,
+ "step": 6269
+ },
+ {
+ "epoch": 87.08384279475983,
+ "grad_norm": 0.014630894176661968,
+ "learning_rate": 0.0006,
+ "loss": 4.148405075073242,
+ "step": 6270
+ },
+ {
+ "epoch": 87.09781659388646,
+ "grad_norm": 0.014211163856089115,
+ "learning_rate": 0.0006,
+ "loss": 4.058030128479004,
+ "step": 6271
+ },
+ {
+ "epoch": 87.1117903930131,
+ "grad_norm": 0.014060168527066708,
+ "learning_rate": 0.0006,
+ "loss": 4.146586894989014,
+ "step": 6272
+ },
+ {
+ "epoch": 87.12576419213974,
+ "grad_norm": 0.014740100130438805,
+ "learning_rate": 0.0006,
+ "loss": 4.077739715576172,
+ "step": 6273
+ },
+ {
+ "epoch": 87.13973799126637,
+ "grad_norm": 0.013970664702355862,
+ "learning_rate": 0.0006,
+ "loss": 4.038252830505371,
+ "step": 6274
+ },
+ {
+ "epoch": 87.15371179039302,
+ "grad_norm": 0.014399951323866844,
+ "learning_rate": 0.0006,
+ "loss": 4.143700122833252,
+ "step": 6275
+ },
+ {
+ "epoch": 87.16768558951965,
+ "grad_norm": 0.01496410183608532,
+ "learning_rate": 0.0006,
+ "loss": 4.105609893798828,
+ "step": 6276
+ },
+ {
+ "epoch": 87.18165938864628,
+ "grad_norm": 0.014287355355918407,
+ "learning_rate": 0.0006,
+ "loss": 4.173645973205566,
+ "step": 6277
+ },
+ {
+ "epoch": 87.19563318777293,
+ "grad_norm": 0.0141292130574584,
+ "learning_rate": 0.0006,
+ "loss": 4.068511486053467,
+ "step": 6278
+ },
+ {
+ "epoch": 87.20960698689956,
+ "grad_norm": 0.01393523346632719,
+ "learning_rate": 0.0006,
+ "loss": 4.060687065124512,
+ "step": 6279
+ },
+ {
+ "epoch": 87.2235807860262,
+ "grad_norm": 0.01431021187454462,
+ "learning_rate": 0.0006,
+ "loss": 4.219209671020508,
+ "step": 6280
+ },
+ {
+ "epoch": 87.23755458515284,
+ "grad_norm": 0.014856027439236641,
+ "learning_rate": 0.0006,
+ "loss": 4.146749973297119,
+ "step": 6281
+ },
+ {
+ "epoch": 87.25152838427948,
+ "grad_norm": 0.013677539303898811,
+ "learning_rate": 0.0006,
+ "loss": 3.9859797954559326,
+ "step": 6282
+ },
+ {
+ "epoch": 87.26550218340611,
+ "grad_norm": 0.014078967273235321,
+ "learning_rate": 0.0006,
+ "loss": 4.129668235778809,
+ "step": 6283
+ },
+ {
+ "epoch": 87.27947598253274,
+ "grad_norm": 0.015564394183456898,
+ "learning_rate": 0.0006,
+ "loss": 4.0007829666137695,
+ "step": 6284
+ },
+ {
+ "epoch": 87.29344978165939,
+ "grad_norm": 0.015077749267220497,
+ "learning_rate": 0.0006,
+ "loss": 4.112264156341553,
+ "step": 6285
+ },
+ {
+ "epoch": 87.30742358078602,
+ "grad_norm": 0.015341218560934067,
+ "learning_rate": 0.0006,
+ "loss": 4.112249374389648,
+ "step": 6286
+ },
+ {
+ "epoch": 87.32139737991267,
+ "grad_norm": 0.016399366781115532,
+ "learning_rate": 0.0006,
+ "loss": 4.176196098327637,
+ "step": 6287
+ },
+ {
+ "epoch": 87.3353711790393,
+ "grad_norm": 0.01599763333797455,
+ "learning_rate": 0.0006,
+ "loss": 4.163441181182861,
+ "step": 6288
+ },
+ {
+ "epoch": 87.34934497816595,
+ "grad_norm": 0.015439776703715324,
+ "learning_rate": 0.0006,
+ "loss": 4.11902379989624,
+ "step": 6289
+ },
+ {
+ "epoch": 87.36331877729258,
+ "grad_norm": 0.015909943729639053,
+ "learning_rate": 0.0006,
+ "loss": 4.16621732711792,
+ "step": 6290
+ },
+ {
+ "epoch": 87.37729257641921,
+ "grad_norm": 0.016710078343749046,
+ "learning_rate": 0.0006,
+ "loss": 4.044960975646973,
+ "step": 6291
+ },
+ {
+ "epoch": 87.39126637554585,
+ "grad_norm": 0.015883933752775192,
+ "learning_rate": 0.0006,
+ "loss": 4.065184593200684,
+ "step": 6292
+ },
+ {
+ "epoch": 87.40524017467249,
+ "grad_norm": 0.013912747614085674,
+ "learning_rate": 0.0006,
+ "loss": 4.102450370788574,
+ "step": 6293
+ },
+ {
+ "epoch": 87.41921397379913,
+ "grad_norm": 0.013170420192182064,
+ "learning_rate": 0.0006,
+ "loss": 4.234364032745361,
+ "step": 6294
+ },
+ {
+ "epoch": 87.43318777292576,
+ "grad_norm": 0.014043302275240421,
+ "learning_rate": 0.0006,
+ "loss": 4.129565238952637,
+ "step": 6295
+ },
+ {
+ "epoch": 87.44716157205241,
+ "grad_norm": 0.015010139904916286,
+ "learning_rate": 0.0006,
+ "loss": 4.166678428649902,
+ "step": 6296
+ },
+ {
+ "epoch": 87.46113537117904,
+ "grad_norm": 0.014721784740686417,
+ "learning_rate": 0.0006,
+ "loss": 4.182618618011475,
+ "step": 6297
+ },
+ {
+ "epoch": 87.47510917030567,
+ "grad_norm": 0.015713468194007874,
+ "learning_rate": 0.0006,
+ "loss": 4.103855133056641,
+ "step": 6298
+ },
+ {
+ "epoch": 87.48908296943232,
+ "grad_norm": 0.017204247415065765,
+ "learning_rate": 0.0006,
+ "loss": 4.140937328338623,
+ "step": 6299
+ },
+ {
+ "epoch": 87.50305676855895,
+ "grad_norm": 0.016856668516993523,
+ "learning_rate": 0.0006,
+ "loss": 4.147859573364258,
+ "step": 6300
+ },
+ {
+ "epoch": 87.5170305676856,
+ "grad_norm": 0.017205018550157547,
+ "learning_rate": 0.0006,
+ "loss": 4.1667327880859375,
+ "step": 6301
+ },
+ {
+ "epoch": 87.53100436681223,
+ "grad_norm": 0.014821539632976055,
+ "learning_rate": 0.0006,
+ "loss": 4.089935302734375,
+ "step": 6302
+ },
+ {
+ "epoch": 87.54497816593886,
+ "grad_norm": 0.01446484588086605,
+ "learning_rate": 0.0006,
+ "loss": 4.243657112121582,
+ "step": 6303
+ },
+ {
+ "epoch": 87.5589519650655,
+ "grad_norm": 0.014374484308063984,
+ "learning_rate": 0.0006,
+ "loss": 4.14356803894043,
+ "step": 6304
+ },
+ {
+ "epoch": 87.57292576419214,
+ "grad_norm": 0.01420109998434782,
+ "learning_rate": 0.0006,
+ "loss": 4.1273651123046875,
+ "step": 6305
+ },
+ {
+ "epoch": 87.58689956331878,
+ "grad_norm": 0.014810388907790184,
+ "learning_rate": 0.0006,
+ "loss": 4.222290515899658,
+ "step": 6306
+ },
+ {
+ "epoch": 87.60087336244541,
+ "grad_norm": 0.015826983377337456,
+ "learning_rate": 0.0006,
+ "loss": 4.137996196746826,
+ "step": 6307
+ },
+ {
+ "epoch": 87.61484716157206,
+ "grad_norm": 0.014457454904913902,
+ "learning_rate": 0.0006,
+ "loss": 4.0780744552612305,
+ "step": 6308
+ },
+ {
+ "epoch": 87.62882096069869,
+ "grad_norm": 0.014458604156970978,
+ "learning_rate": 0.0006,
+ "loss": 4.1287922859191895,
+ "step": 6309
+ },
+ {
+ "epoch": 87.64279475982532,
+ "grad_norm": 0.015282071195542812,
+ "learning_rate": 0.0006,
+ "loss": 4.189508438110352,
+ "step": 6310
+ },
+ {
+ "epoch": 87.65676855895197,
+ "grad_norm": 0.017536642029881477,
+ "learning_rate": 0.0006,
+ "loss": 4.070173263549805,
+ "step": 6311
+ },
+ {
+ "epoch": 87.6707423580786,
+ "grad_norm": 0.017303580418229103,
+ "learning_rate": 0.0006,
+ "loss": 4.119133949279785,
+ "step": 6312
+ },
+ {
+ "epoch": 87.68471615720524,
+ "grad_norm": 0.016148580238223076,
+ "learning_rate": 0.0006,
+ "loss": 3.994354486465454,
+ "step": 6313
+ },
+ {
+ "epoch": 87.69868995633188,
+ "grad_norm": 0.01706942543387413,
+ "learning_rate": 0.0006,
+ "loss": 4.183113098144531,
+ "step": 6314
+ },
+ {
+ "epoch": 87.71266375545852,
+ "grad_norm": 0.016971010714769363,
+ "learning_rate": 0.0006,
+ "loss": 4.104086875915527,
+ "step": 6315
+ },
+ {
+ "epoch": 87.72663755458515,
+ "grad_norm": 0.015516847372055054,
+ "learning_rate": 0.0006,
+ "loss": 4.2409772872924805,
+ "step": 6316
+ },
+ {
+ "epoch": 87.74061135371178,
+ "grad_norm": 0.015352269634604454,
+ "learning_rate": 0.0006,
+ "loss": 4.026427268981934,
+ "step": 6317
+ },
+ {
+ "epoch": 87.75458515283843,
+ "grad_norm": 0.017702415585517883,
+ "learning_rate": 0.0006,
+ "loss": 4.177840232849121,
+ "step": 6318
+ },
+ {
+ "epoch": 87.76855895196506,
+ "grad_norm": 0.01629541628062725,
+ "learning_rate": 0.0006,
+ "loss": 4.229039192199707,
+ "step": 6319
+ },
+ {
+ "epoch": 87.78253275109171,
+ "grad_norm": 0.01503711100667715,
+ "learning_rate": 0.0006,
+ "loss": 4.214350700378418,
+ "step": 6320
+ },
+ {
+ "epoch": 87.79650655021834,
+ "grad_norm": 0.014913883060216904,
+ "learning_rate": 0.0006,
+ "loss": 4.168880939483643,
+ "step": 6321
+ },
+ {
+ "epoch": 87.81048034934499,
+ "grad_norm": 0.014513909816741943,
+ "learning_rate": 0.0006,
+ "loss": 4.162267684936523,
+ "step": 6322
+ },
+ {
+ "epoch": 87.82445414847162,
+ "grad_norm": 0.014635873027145863,
+ "learning_rate": 0.0006,
+ "loss": 4.220067977905273,
+ "step": 6323
+ },
+ {
+ "epoch": 87.83842794759825,
+ "grad_norm": 0.017102796584367752,
+ "learning_rate": 0.0006,
+ "loss": 4.1493377685546875,
+ "step": 6324
+ },
+ {
+ "epoch": 87.8524017467249,
+ "grad_norm": 0.015533884055912495,
+ "learning_rate": 0.0006,
+ "loss": 4.0646209716796875,
+ "step": 6325
+ },
+ {
+ "epoch": 87.86637554585153,
+ "grad_norm": 0.014617550186812878,
+ "learning_rate": 0.0006,
+ "loss": 4.1390533447265625,
+ "step": 6326
+ },
+ {
+ "epoch": 87.88034934497817,
+ "grad_norm": 0.014142967760562897,
+ "learning_rate": 0.0006,
+ "loss": 4.133934020996094,
+ "step": 6327
+ },
+ {
+ "epoch": 87.8943231441048,
+ "grad_norm": 0.013182433322072029,
+ "learning_rate": 0.0006,
+ "loss": 4.125327110290527,
+ "step": 6328
+ },
+ {
+ "epoch": 87.90829694323143,
+ "grad_norm": 0.013928229920566082,
+ "learning_rate": 0.0006,
+ "loss": 4.142777442932129,
+ "step": 6329
+ },
+ {
+ "epoch": 87.92227074235808,
+ "grad_norm": 0.015337013639509678,
+ "learning_rate": 0.0006,
+ "loss": 4.140210151672363,
+ "step": 6330
+ },
+ {
+ "epoch": 87.93624454148471,
+ "grad_norm": 0.014873536303639412,
+ "learning_rate": 0.0006,
+ "loss": 4.102636337280273,
+ "step": 6331
+ },
+ {
+ "epoch": 87.95021834061136,
+ "grad_norm": 0.01387431938201189,
+ "learning_rate": 0.0006,
+ "loss": 4.149059295654297,
+ "step": 6332
+ },
+ {
+ "epoch": 87.96419213973799,
+ "grad_norm": 0.013234943151473999,
+ "learning_rate": 0.0006,
+ "loss": 4.1564459800720215,
+ "step": 6333
+ },
+ {
+ "epoch": 87.97816593886463,
+ "grad_norm": 0.013781017623841763,
+ "learning_rate": 0.0006,
+ "loss": 4.189839839935303,
+ "step": 6334
+ },
+ {
+ "epoch": 87.99213973799127,
+ "grad_norm": 0.013733026571571827,
+ "learning_rate": 0.0006,
+ "loss": 4.29277229309082,
+ "step": 6335
+ },
+ {
+ "epoch": 88.0,
+ "grad_norm": 0.014752327464520931,
+ "learning_rate": 0.0006,
+ "loss": 4.163867950439453,
+ "step": 6336
+ },
+ {
+ "epoch": 88.0,
+ "eval_loss": 4.535566329956055,
+ "eval_runtime": 56.4677,
+ "eval_samples_per_second": 43.246,
+ "eval_steps_per_second": 1.364,
+ "step": 6336
+ },
+ {
+ "epoch": 88.01397379912663,
+ "grad_norm": 0.013958732597529888,
+ "learning_rate": 0.0006,
+ "loss": 4.200681209564209,
+ "step": 6337
+ },
+ {
+ "epoch": 88.02794759825328,
+ "grad_norm": 0.015283545479178429,
+ "learning_rate": 0.0006,
+ "loss": 4.1736016273498535,
+ "step": 6338
+ },
+ {
+ "epoch": 88.04192139737991,
+ "grad_norm": 0.016192059963941574,
+ "learning_rate": 0.0006,
+ "loss": 4.040589809417725,
+ "step": 6339
+ },
+ {
+ "epoch": 88.05589519650655,
+ "grad_norm": 0.018096154555678368,
+ "learning_rate": 0.0006,
+ "loss": 4.146696090698242,
+ "step": 6340
+ },
+ {
+ "epoch": 88.06986899563319,
+ "grad_norm": 0.01797664538025856,
+ "learning_rate": 0.0006,
+ "loss": 4.007248878479004,
+ "step": 6341
+ },
+ {
+ "epoch": 88.08384279475983,
+ "grad_norm": 0.016275612637400627,
+ "learning_rate": 0.0006,
+ "loss": 4.193206787109375,
+ "step": 6342
+ },
+ {
+ "epoch": 88.09781659388646,
+ "grad_norm": 0.014210258610546589,
+ "learning_rate": 0.0006,
+ "loss": 4.064548492431641,
+ "step": 6343
+ },
+ {
+ "epoch": 88.1117903930131,
+ "grad_norm": 0.016085151582956314,
+ "learning_rate": 0.0006,
+ "loss": 4.003668785095215,
+ "step": 6344
+ },
+ {
+ "epoch": 88.12576419213974,
+ "grad_norm": 0.01930483989417553,
+ "learning_rate": 0.0006,
+ "loss": 4.236578941345215,
+ "step": 6345
+ },
+ {
+ "epoch": 88.13973799126637,
+ "grad_norm": 0.017140906304121017,
+ "learning_rate": 0.0006,
+ "loss": 3.9593005180358887,
+ "step": 6346
+ },
+ {
+ "epoch": 88.15371179039302,
+ "grad_norm": 0.01521560549736023,
+ "learning_rate": 0.0006,
+ "loss": 4.137644290924072,
+ "step": 6347
+ },
+ {
+ "epoch": 88.16768558951965,
+ "grad_norm": 0.015788141638040543,
+ "learning_rate": 0.0006,
+ "loss": 4.124845027923584,
+ "step": 6348
+ },
+ {
+ "epoch": 88.18165938864628,
+ "grad_norm": 0.01784597896039486,
+ "learning_rate": 0.0006,
+ "loss": 4.139750957489014,
+ "step": 6349
+ },
+ {
+ "epoch": 88.19563318777293,
+ "grad_norm": 0.016266558319330215,
+ "learning_rate": 0.0006,
+ "loss": 4.074851989746094,
+ "step": 6350
+ },
+ {
+ "epoch": 88.20960698689956,
+ "grad_norm": 0.014024483971297741,
+ "learning_rate": 0.0006,
+ "loss": 4.164689064025879,
+ "step": 6351
+ },
+ {
+ "epoch": 88.2235807860262,
+ "grad_norm": 0.015694420784711838,
+ "learning_rate": 0.0006,
+ "loss": 4.219979286193848,
+ "step": 6352
+ },
+ {
+ "epoch": 88.23755458515284,
+ "grad_norm": 0.014897859655320644,
+ "learning_rate": 0.0006,
+ "loss": 4.050738334655762,
+ "step": 6353
+ },
+ {
+ "epoch": 88.25152838427948,
+ "grad_norm": 0.014827374368906021,
+ "learning_rate": 0.0006,
+ "loss": 4.137135028839111,
+ "step": 6354
+ },
+ {
+ "epoch": 88.26550218340611,
+ "grad_norm": 0.01454879529774189,
+ "learning_rate": 0.0006,
+ "loss": 3.9980897903442383,
+ "step": 6355
+ },
+ {
+ "epoch": 88.27947598253274,
+ "grad_norm": 0.015623947605490685,
+ "learning_rate": 0.0006,
+ "loss": 4.226551055908203,
+ "step": 6356
+ },
+ {
+ "epoch": 88.29344978165939,
+ "grad_norm": 0.015527035109698772,
+ "learning_rate": 0.0006,
+ "loss": 4.11053466796875,
+ "step": 6357
+ },
+ {
+ "epoch": 88.30742358078602,
+ "grad_norm": 0.015242574736475945,
+ "learning_rate": 0.0006,
+ "loss": 4.105966567993164,
+ "step": 6358
+ },
+ {
+ "epoch": 88.32139737991267,
+ "grad_norm": 0.01375538855791092,
+ "learning_rate": 0.0006,
+ "loss": 4.148453712463379,
+ "step": 6359
+ },
+ {
+ "epoch": 88.3353711790393,
+ "grad_norm": 0.01595226116478443,
+ "learning_rate": 0.0006,
+ "loss": 4.095836162567139,
+ "step": 6360
+ },
+ {
+ "epoch": 88.34934497816595,
+ "grad_norm": 0.015252740122377872,
+ "learning_rate": 0.0006,
+ "loss": 4.136335849761963,
+ "step": 6361
+ },
+ {
+ "epoch": 88.36331877729258,
+ "grad_norm": 0.014405866153538227,
+ "learning_rate": 0.0006,
+ "loss": 4.084691047668457,
+ "step": 6362
+ },
+ {
+ "epoch": 88.37729257641921,
+ "grad_norm": 0.014755490235984325,
+ "learning_rate": 0.0006,
+ "loss": 4.095475673675537,
+ "step": 6363
+ },
+ {
+ "epoch": 88.39126637554585,
+ "grad_norm": 0.015339708887040615,
+ "learning_rate": 0.0006,
+ "loss": 4.163136959075928,
+ "step": 6364
+ },
+ {
+ "epoch": 88.40524017467249,
+ "grad_norm": 0.014667164534330368,
+ "learning_rate": 0.0006,
+ "loss": 4.062265396118164,
+ "step": 6365
+ },
+ {
+ "epoch": 88.41921397379913,
+ "grad_norm": 0.014588423073291779,
+ "learning_rate": 0.0006,
+ "loss": 4.13359260559082,
+ "step": 6366
+ },
+ {
+ "epoch": 88.43318777292576,
+ "grad_norm": 0.014005377888679504,
+ "learning_rate": 0.0006,
+ "loss": 4.219711780548096,
+ "step": 6367
+ },
+ {
+ "epoch": 88.44716157205241,
+ "grad_norm": 0.012769930064678192,
+ "learning_rate": 0.0006,
+ "loss": 4.114658355712891,
+ "step": 6368
+ },
+ {
+ "epoch": 88.46113537117904,
+ "grad_norm": 0.013632693327963352,
+ "learning_rate": 0.0006,
+ "loss": 4.0675225257873535,
+ "step": 6369
+ },
+ {
+ "epoch": 88.47510917030567,
+ "grad_norm": 0.013829804956912994,
+ "learning_rate": 0.0006,
+ "loss": 4.1125969886779785,
+ "step": 6370
+ },
+ {
+ "epoch": 88.48908296943232,
+ "grad_norm": 0.013882109895348549,
+ "learning_rate": 0.0006,
+ "loss": 4.1627092361450195,
+ "step": 6371
+ },
+ {
+ "epoch": 88.50305676855895,
+ "grad_norm": 0.014023828320205212,
+ "learning_rate": 0.0006,
+ "loss": 4.01038932800293,
+ "step": 6372
+ },
+ {
+ "epoch": 88.5170305676856,
+ "grad_norm": 0.014840706251561642,
+ "learning_rate": 0.0006,
+ "loss": 4.117877960205078,
+ "step": 6373
+ },
+ {
+ "epoch": 88.53100436681223,
+ "grad_norm": 0.014165930449962616,
+ "learning_rate": 0.0006,
+ "loss": 4.1864237785339355,
+ "step": 6374
+ },
+ {
+ "epoch": 88.54497816593886,
+ "grad_norm": 0.014390201307833195,
+ "learning_rate": 0.0006,
+ "loss": 4.1044602394104,
+ "step": 6375
+ },
+ {
+ "epoch": 88.5589519650655,
+ "grad_norm": 0.015037395060062408,
+ "learning_rate": 0.0006,
+ "loss": 4.101107120513916,
+ "step": 6376
+ },
+ {
+ "epoch": 88.57292576419214,
+ "grad_norm": 0.01671452447772026,
+ "learning_rate": 0.0006,
+ "loss": 4.170581817626953,
+ "step": 6377
+ },
+ {
+ "epoch": 88.58689956331878,
+ "grad_norm": 0.01993076130747795,
+ "learning_rate": 0.0006,
+ "loss": 4.261412620544434,
+ "step": 6378
+ },
+ {
+ "epoch": 88.60087336244541,
+ "grad_norm": 0.023721016943454742,
+ "learning_rate": 0.0006,
+ "loss": 4.199363708496094,
+ "step": 6379
+ },
+ {
+ "epoch": 88.61484716157206,
+ "grad_norm": 0.023726163432002068,
+ "learning_rate": 0.0006,
+ "loss": 4.061285018920898,
+ "step": 6380
+ },
+ {
+ "epoch": 88.62882096069869,
+ "grad_norm": 0.018352877348661423,
+ "learning_rate": 0.0006,
+ "loss": 4.0721282958984375,
+ "step": 6381
+ },
+ {
+ "epoch": 88.64279475982532,
+ "grad_norm": 0.01679830253124237,
+ "learning_rate": 0.0006,
+ "loss": 4.1717939376831055,
+ "step": 6382
+ },
+ {
+ "epoch": 88.65676855895197,
+ "grad_norm": 0.017193637788295746,
+ "learning_rate": 0.0006,
+ "loss": 4.230783939361572,
+ "step": 6383
+ },
+ {
+ "epoch": 88.6707423580786,
+ "grad_norm": 0.01581464521586895,
+ "learning_rate": 0.0006,
+ "loss": 4.029906272888184,
+ "step": 6384
+ },
+ {
+ "epoch": 88.68471615720524,
+ "grad_norm": 0.014049704186618328,
+ "learning_rate": 0.0006,
+ "loss": 4.192776679992676,
+ "step": 6385
+ },
+ {
+ "epoch": 88.69868995633188,
+ "grad_norm": 0.017214450985193253,
+ "learning_rate": 0.0006,
+ "loss": 4.248916149139404,
+ "step": 6386
+ },
+ {
+ "epoch": 88.71266375545852,
+ "grad_norm": 0.0173103716224432,
+ "learning_rate": 0.0006,
+ "loss": 4.124347686767578,
+ "step": 6387
+ },
+ {
+ "epoch": 88.72663755458515,
+ "grad_norm": 0.015745092183351517,
+ "learning_rate": 0.0006,
+ "loss": 4.192878246307373,
+ "step": 6388
+ },
+ {
+ "epoch": 88.74061135371178,
+ "grad_norm": 0.013750772923231125,
+ "learning_rate": 0.0006,
+ "loss": 4.1564483642578125,
+ "step": 6389
+ },
+ {
+ "epoch": 88.75458515283843,
+ "grad_norm": 0.014969858340919018,
+ "learning_rate": 0.0006,
+ "loss": 4.0980424880981445,
+ "step": 6390
+ },
+ {
+ "epoch": 88.76855895196506,
+ "grad_norm": 0.01287173479795456,
+ "learning_rate": 0.0006,
+ "loss": 4.145712852478027,
+ "step": 6391
+ },
+ {
+ "epoch": 88.78253275109171,
+ "grad_norm": 0.012716300785541534,
+ "learning_rate": 0.0006,
+ "loss": 4.214226722717285,
+ "step": 6392
+ },
+ {
+ "epoch": 88.79650655021834,
+ "grad_norm": 0.013855204917490482,
+ "learning_rate": 0.0006,
+ "loss": 4.1667256355285645,
+ "step": 6393
+ },
+ {
+ "epoch": 88.81048034934499,
+ "grad_norm": 0.013650864362716675,
+ "learning_rate": 0.0006,
+ "loss": 4.161942481994629,
+ "step": 6394
+ },
+ {
+ "epoch": 88.82445414847162,
+ "grad_norm": 0.013780126348137856,
+ "learning_rate": 0.0006,
+ "loss": 4.087825775146484,
+ "step": 6395
+ },
+ {
+ "epoch": 88.83842794759825,
+ "grad_norm": 0.014029948972165585,
+ "learning_rate": 0.0006,
+ "loss": 4.142813205718994,
+ "step": 6396
+ },
+ {
+ "epoch": 88.8524017467249,
+ "grad_norm": 0.01322692260146141,
+ "learning_rate": 0.0006,
+ "loss": 4.106776714324951,
+ "step": 6397
+ },
+ {
+ "epoch": 88.86637554585153,
+ "grad_norm": 0.013522760011255741,
+ "learning_rate": 0.0006,
+ "loss": 4.090996265411377,
+ "step": 6398
+ },
+ {
+ "epoch": 88.88034934497817,
+ "grad_norm": 0.014474684372544289,
+ "learning_rate": 0.0006,
+ "loss": 4.124624252319336,
+ "step": 6399
+ },
+ {
+ "epoch": 88.8943231441048,
+ "grad_norm": 0.014490422792732716,
+ "learning_rate": 0.0006,
+ "loss": 4.065431594848633,
+ "step": 6400
+ },
+ {
+ "epoch": 88.90829694323143,
+ "grad_norm": 0.015379710122942924,
+ "learning_rate": 0.0006,
+ "loss": 4.207939624786377,
+ "step": 6401
+ },
+ {
+ "epoch": 88.92227074235808,
+ "grad_norm": 0.015914618968963623,
+ "learning_rate": 0.0006,
+ "loss": 4.041152000427246,
+ "step": 6402
+ },
+ {
+ "epoch": 88.93624454148471,
+ "grad_norm": 0.01480048056691885,
+ "learning_rate": 0.0006,
+ "loss": 4.138754844665527,
+ "step": 6403
+ },
+ {
+ "epoch": 88.95021834061136,
+ "grad_norm": 0.013734851032495499,
+ "learning_rate": 0.0006,
+ "loss": 4.25323486328125,
+ "step": 6404
+ },
+ {
+ "epoch": 88.96419213973799,
+ "grad_norm": 0.01452311035245657,
+ "learning_rate": 0.0006,
+ "loss": 4.115228176116943,
+ "step": 6405
+ },
+ {
+ "epoch": 88.97816593886463,
+ "grad_norm": 0.016197681427001953,
+ "learning_rate": 0.0006,
+ "loss": 4.099531173706055,
+ "step": 6406
+ },
+ {
+ "epoch": 88.99213973799127,
+ "grad_norm": 0.017169523984193802,
+ "learning_rate": 0.0006,
+ "loss": 4.182981491088867,
+ "step": 6407
+ },
+ {
+ "epoch": 89.0,
+ "grad_norm": 0.017855646088719368,
+ "learning_rate": 0.0006,
+ "loss": 4.1470136642456055,
+ "step": 6408
+ },
+ {
+ "epoch": 89.0,
+ "eval_loss": 4.6035614013671875,
+ "eval_runtime": 57.0625,
+ "eval_samples_per_second": 42.795,
+ "eval_steps_per_second": 1.349,
+ "step": 6408
+ },
+ {
+ "epoch": 89.01397379912663,
+ "grad_norm": 0.014834101311862469,
+ "learning_rate": 0.0006,
+ "loss": 4.159243583679199,
+ "step": 6409
+ },
+ {
+ "epoch": 89.02794759825328,
+ "grad_norm": 0.015316772274672985,
+ "learning_rate": 0.0006,
+ "loss": 4.0709123611450195,
+ "step": 6410
+ },
+ {
+ "epoch": 89.04192139737991,
+ "grad_norm": 0.01576591096818447,
+ "learning_rate": 0.0006,
+ "loss": 4.0921196937561035,
+ "step": 6411
+ },
+ {
+ "epoch": 89.05589519650655,
+ "grad_norm": 0.015948045998811722,
+ "learning_rate": 0.0006,
+ "loss": 4.181877136230469,
+ "step": 6412
+ },
+ {
+ "epoch": 89.06986899563319,
+ "grad_norm": 0.016612380743026733,
+ "learning_rate": 0.0006,
+ "loss": 4.081249237060547,
+ "step": 6413
+ },
+ {
+ "epoch": 89.08384279475983,
+ "grad_norm": 0.017455685883760452,
+ "learning_rate": 0.0006,
+ "loss": 4.175990104675293,
+ "step": 6414
+ },
+ {
+ "epoch": 89.09781659388646,
+ "grad_norm": 0.015040521509945393,
+ "learning_rate": 0.0006,
+ "loss": 4.121378421783447,
+ "step": 6415
+ },
+ {
+ "epoch": 89.1117903930131,
+ "grad_norm": 0.014513209462165833,
+ "learning_rate": 0.0006,
+ "loss": 3.996107816696167,
+ "step": 6416
+ },
+ {
+ "epoch": 89.12576419213974,
+ "grad_norm": 0.015455638989806175,
+ "learning_rate": 0.0006,
+ "loss": 4.073675155639648,
+ "step": 6417
+ },
+ {
+ "epoch": 89.13973799126637,
+ "grad_norm": 0.015468129888176918,
+ "learning_rate": 0.0006,
+ "loss": 4.130242824554443,
+ "step": 6418
+ },
+ {
+ "epoch": 89.15371179039302,
+ "grad_norm": 0.015284297987818718,
+ "learning_rate": 0.0006,
+ "loss": 4.18555212020874,
+ "step": 6419
+ },
+ {
+ "epoch": 89.16768558951965,
+ "grad_norm": 0.01476874016225338,
+ "learning_rate": 0.0006,
+ "loss": 4.134175777435303,
+ "step": 6420
+ },
+ {
+ "epoch": 89.18165938864628,
+ "grad_norm": 0.015836201608181,
+ "learning_rate": 0.0006,
+ "loss": 4.094529151916504,
+ "step": 6421
+ },
+ {
+ "epoch": 89.19563318777293,
+ "grad_norm": 0.0161928478628397,
+ "learning_rate": 0.0006,
+ "loss": 4.12647819519043,
+ "step": 6422
+ },
+ {
+ "epoch": 89.20960698689956,
+ "grad_norm": 0.015680814161896706,
+ "learning_rate": 0.0006,
+ "loss": 3.9620392322540283,
+ "step": 6423
+ },
+ {
+ "epoch": 89.2235807860262,
+ "grad_norm": 0.01584681309759617,
+ "learning_rate": 0.0006,
+ "loss": 4.049304008483887,
+ "step": 6424
+ },
+ {
+ "epoch": 89.23755458515284,
+ "grad_norm": 0.01452907919883728,
+ "learning_rate": 0.0006,
+ "loss": 4.064027786254883,
+ "step": 6425
+ },
+ {
+ "epoch": 89.25152838427948,
+ "grad_norm": 0.014002552255988121,
+ "learning_rate": 0.0006,
+ "loss": 4.151862621307373,
+ "step": 6426
+ },
+ {
+ "epoch": 89.26550218340611,
+ "grad_norm": 0.014658625237643719,
+ "learning_rate": 0.0006,
+ "loss": 4.1672210693359375,
+ "step": 6427
+ },
+ {
+ "epoch": 89.27947598253274,
+ "grad_norm": 0.015021766535937786,
+ "learning_rate": 0.0006,
+ "loss": 3.9742586612701416,
+ "step": 6428
+ },
+ {
+ "epoch": 89.29344978165939,
+ "grad_norm": 0.014681321568787098,
+ "learning_rate": 0.0006,
+ "loss": 4.024159908294678,
+ "step": 6429
+ },
+ {
+ "epoch": 89.30742358078602,
+ "grad_norm": 0.014161549508571625,
+ "learning_rate": 0.0006,
+ "loss": 4.105640888214111,
+ "step": 6430
+ },
+ {
+ "epoch": 89.32139737991267,
+ "grad_norm": 0.014801453799009323,
+ "learning_rate": 0.0006,
+ "loss": 4.123049736022949,
+ "step": 6431
+ },
+ {
+ "epoch": 89.3353711790393,
+ "grad_norm": 0.014307181350886822,
+ "learning_rate": 0.0006,
+ "loss": 4.086954593658447,
+ "step": 6432
+ },
+ {
+ "epoch": 89.34934497816595,
+ "grad_norm": 0.013299369253218174,
+ "learning_rate": 0.0006,
+ "loss": 4.109533786773682,
+ "step": 6433
+ },
+ {
+ "epoch": 89.36331877729258,
+ "grad_norm": 0.014388669282197952,
+ "learning_rate": 0.0006,
+ "loss": 4.222370147705078,
+ "step": 6434
+ },
+ {
+ "epoch": 89.37729257641921,
+ "grad_norm": 0.014501879923045635,
+ "learning_rate": 0.0006,
+ "loss": 4.224178314208984,
+ "step": 6435
+ },
+ {
+ "epoch": 89.39126637554585,
+ "grad_norm": 0.014109021984040737,
+ "learning_rate": 0.0006,
+ "loss": 4.112702369689941,
+ "step": 6436
+ },
+ {
+ "epoch": 89.40524017467249,
+ "grad_norm": 0.013521002605557442,
+ "learning_rate": 0.0006,
+ "loss": 4.010124683380127,
+ "step": 6437
+ },
+ {
+ "epoch": 89.41921397379913,
+ "grad_norm": 0.013288144953548908,
+ "learning_rate": 0.0006,
+ "loss": 4.127592086791992,
+ "step": 6438
+ },
+ {
+ "epoch": 89.43318777292576,
+ "grad_norm": 0.013882486149668694,
+ "learning_rate": 0.0006,
+ "loss": 4.031871318817139,
+ "step": 6439
+ },
+ {
+ "epoch": 89.44716157205241,
+ "grad_norm": 0.01289335172623396,
+ "learning_rate": 0.0006,
+ "loss": 4.0757551193237305,
+ "step": 6440
+ },
+ {
+ "epoch": 89.46113537117904,
+ "grad_norm": 0.01356302760541439,
+ "learning_rate": 0.0006,
+ "loss": 4.075678825378418,
+ "step": 6441
+ },
+ {
+ "epoch": 89.47510917030567,
+ "grad_norm": 0.01447434164583683,
+ "learning_rate": 0.0006,
+ "loss": 4.093453407287598,
+ "step": 6442
+ },
+ {
+ "epoch": 89.48908296943232,
+ "grad_norm": 0.014443481341004372,
+ "learning_rate": 0.0006,
+ "loss": 4.075372219085693,
+ "step": 6443
+ },
+ {
+ "epoch": 89.50305676855895,
+ "grad_norm": 0.01314218994230032,
+ "learning_rate": 0.0006,
+ "loss": 4.171533584594727,
+ "step": 6444
+ },
+ {
+ "epoch": 89.5170305676856,
+ "grad_norm": 0.01348346658051014,
+ "learning_rate": 0.0006,
+ "loss": 4.227236270904541,
+ "step": 6445
+ },
+ {
+ "epoch": 89.53100436681223,
+ "grad_norm": 0.013370679691433907,
+ "learning_rate": 0.0006,
+ "loss": 3.973062515258789,
+ "step": 6446
+ },
+ {
+ "epoch": 89.54497816593886,
+ "grad_norm": 0.01410642359405756,
+ "learning_rate": 0.0006,
+ "loss": 4.231175422668457,
+ "step": 6447
+ },
+ {
+ "epoch": 89.5589519650655,
+ "grad_norm": 0.01419337373226881,
+ "learning_rate": 0.0006,
+ "loss": 4.145868301391602,
+ "step": 6448
+ },
+ {
+ "epoch": 89.57292576419214,
+ "grad_norm": 0.013691053725779057,
+ "learning_rate": 0.0006,
+ "loss": 4.175131797790527,
+ "step": 6449
+ },
+ {
+ "epoch": 89.58689956331878,
+ "grad_norm": 0.013714492321014404,
+ "learning_rate": 0.0006,
+ "loss": 4.251893997192383,
+ "step": 6450
+ },
+ {
+ "epoch": 89.60087336244541,
+ "grad_norm": 0.01393197476863861,
+ "learning_rate": 0.0006,
+ "loss": 4.047079086303711,
+ "step": 6451
+ },
+ {
+ "epoch": 89.61484716157206,
+ "grad_norm": 0.013957309536635876,
+ "learning_rate": 0.0006,
+ "loss": 4.128530502319336,
+ "step": 6452
+ },
+ {
+ "epoch": 89.62882096069869,
+ "grad_norm": 0.013346714898943901,
+ "learning_rate": 0.0006,
+ "loss": 4.2296929359436035,
+ "step": 6453
+ },
+ {
+ "epoch": 89.64279475982532,
+ "grad_norm": 0.014243829064071178,
+ "learning_rate": 0.0006,
+ "loss": 4.128233909606934,
+ "step": 6454
+ },
+ {
+ "epoch": 89.65676855895197,
+ "grad_norm": 0.01606106385588646,
+ "learning_rate": 0.0006,
+ "loss": 4.202202320098877,
+ "step": 6455
+ },
+ {
+ "epoch": 89.6707423580786,
+ "grad_norm": 0.016074348241090775,
+ "learning_rate": 0.0006,
+ "loss": 4.215323448181152,
+ "step": 6456
+ },
+ {
+ "epoch": 89.68471615720524,
+ "grad_norm": 0.01611054316163063,
+ "learning_rate": 0.0006,
+ "loss": 4.096758842468262,
+ "step": 6457
+ },
+ {
+ "epoch": 89.69868995633188,
+ "grad_norm": 0.013788217678666115,
+ "learning_rate": 0.0006,
+ "loss": 4.098560810089111,
+ "step": 6458
+ },
+ {
+ "epoch": 89.71266375545852,
+ "grad_norm": 0.014408969320356846,
+ "learning_rate": 0.0006,
+ "loss": 4.107398509979248,
+ "step": 6459
+ },
+ {
+ "epoch": 89.72663755458515,
+ "grad_norm": 0.014765267260372639,
+ "learning_rate": 0.0006,
+ "loss": 4.057718276977539,
+ "step": 6460
+ },
+ {
+ "epoch": 89.74061135371178,
+ "grad_norm": 0.017395280301570892,
+ "learning_rate": 0.0006,
+ "loss": 4.100795269012451,
+ "step": 6461
+ },
+ {
+ "epoch": 89.75458515283843,
+ "grad_norm": 0.016821149736642838,
+ "learning_rate": 0.0006,
+ "loss": 4.153731346130371,
+ "step": 6462
+ },
+ {
+ "epoch": 89.76855895196506,
+ "grad_norm": 0.015633124858140945,
+ "learning_rate": 0.0006,
+ "loss": 4.061585903167725,
+ "step": 6463
+ },
+ {
+ "epoch": 89.78253275109171,
+ "grad_norm": 0.01451497245579958,
+ "learning_rate": 0.0006,
+ "loss": 4.1624884605407715,
+ "step": 6464
+ },
+ {
+ "epoch": 89.79650655021834,
+ "grad_norm": 0.013868392445147038,
+ "learning_rate": 0.0006,
+ "loss": 4.178536415100098,
+ "step": 6465
+ },
+ {
+ "epoch": 89.81048034934499,
+ "grad_norm": 0.014259875752031803,
+ "learning_rate": 0.0006,
+ "loss": 4.073114395141602,
+ "step": 6466
+ },
+ {
+ "epoch": 89.82445414847162,
+ "grad_norm": 0.013651594519615173,
+ "learning_rate": 0.0006,
+ "loss": 4.050265312194824,
+ "step": 6467
+ },
+ {
+ "epoch": 89.83842794759825,
+ "grad_norm": 0.014063458889722824,
+ "learning_rate": 0.0006,
+ "loss": 4.2014689445495605,
+ "step": 6468
+ },
+ {
+ "epoch": 89.8524017467249,
+ "grad_norm": 0.014179578050971031,
+ "learning_rate": 0.0006,
+ "loss": 4.110550880432129,
+ "step": 6469
+ },
+ {
+ "epoch": 89.86637554585153,
+ "grad_norm": 0.013616854324936867,
+ "learning_rate": 0.0006,
+ "loss": 4.1776580810546875,
+ "step": 6470
+ },
+ {
+ "epoch": 89.88034934497817,
+ "grad_norm": 0.014351330697536469,
+ "learning_rate": 0.0006,
+ "loss": 4.058526992797852,
+ "step": 6471
+ },
+ {
+ "epoch": 89.8943231441048,
+ "grad_norm": 0.0141244325786829,
+ "learning_rate": 0.0006,
+ "loss": 4.068673610687256,
+ "step": 6472
+ },
+ {
+ "epoch": 89.90829694323143,
+ "grad_norm": 0.01552131213247776,
+ "learning_rate": 0.0006,
+ "loss": 4.086787223815918,
+ "step": 6473
+ },
+ {
+ "epoch": 89.92227074235808,
+ "grad_norm": 0.014614894054830074,
+ "learning_rate": 0.0006,
+ "loss": 4.2190656661987305,
+ "step": 6474
+ },
+ {
+ "epoch": 89.93624454148471,
+ "grad_norm": 0.014369758777320385,
+ "learning_rate": 0.0006,
+ "loss": 4.117021560668945,
+ "step": 6475
+ },
+ {
+ "epoch": 89.95021834061136,
+ "grad_norm": 0.015352046117186546,
+ "learning_rate": 0.0006,
+ "loss": 4.167664527893066,
+ "step": 6476
+ },
+ {
+ "epoch": 89.96419213973799,
+ "grad_norm": 0.014314834959805012,
+ "learning_rate": 0.0006,
+ "loss": 4.160182952880859,
+ "step": 6477
+ },
+ {
+ "epoch": 89.97816593886463,
+ "grad_norm": 0.013511341996490955,
+ "learning_rate": 0.0006,
+ "loss": 4.096034049987793,
+ "step": 6478
+ },
+ {
+ "epoch": 89.99213973799127,
+ "grad_norm": 0.01308482326567173,
+ "learning_rate": 0.0006,
+ "loss": 4.062689781188965,
+ "step": 6479
+ },
+ {
+ "epoch": 90.0,
+ "grad_norm": 0.01442806702107191,
+ "learning_rate": 0.0006,
+ "loss": 4.1955885887146,
+ "step": 6480
+ },
+ {
+ "epoch": 90.0,
+ "eval_loss": 4.644347667694092,
+ "eval_runtime": 56.8677,
+ "eval_samples_per_second": 42.942,
+ "eval_steps_per_second": 1.354,
+ "step": 6480
+ },
+ {
+ "epoch": 90.01397379912663,
+ "grad_norm": 0.014839007519185543,
+ "learning_rate": 0.0006,
+ "loss": 4.171868324279785,
+ "step": 6481
+ },
+ {
+ "epoch": 90.02794759825328,
+ "grad_norm": 0.01476900465786457,
+ "learning_rate": 0.0006,
+ "loss": 4.119754791259766,
+ "step": 6482
+ },
+ {
+ "epoch": 90.04192139737991,
+ "grad_norm": 0.015179301612079144,
+ "learning_rate": 0.0006,
+ "loss": 4.06740140914917,
+ "step": 6483
+ },
+ {
+ "epoch": 90.05589519650655,
+ "grad_norm": 0.015552980825304985,
+ "learning_rate": 0.0006,
+ "loss": 4.112268447875977,
+ "step": 6484
+ },
+ {
+ "epoch": 90.06986899563319,
+ "grad_norm": 0.014930234290659428,
+ "learning_rate": 0.0006,
+ "loss": 4.071866989135742,
+ "step": 6485
+ },
+ {
+ "epoch": 90.08384279475983,
+ "grad_norm": 0.014124148525297642,
+ "learning_rate": 0.0006,
+ "loss": 4.013731956481934,
+ "step": 6486
+ },
+ {
+ "epoch": 90.09781659388646,
+ "grad_norm": 0.014204252511262894,
+ "learning_rate": 0.0006,
+ "loss": 4.133620738983154,
+ "step": 6487
+ },
+ {
+ "epoch": 90.1117903930131,
+ "grad_norm": 0.014845543541014194,
+ "learning_rate": 0.0006,
+ "loss": 4.122956275939941,
+ "step": 6488
+ },
+ {
+ "epoch": 90.12576419213974,
+ "grad_norm": 0.0146744754165411,
+ "learning_rate": 0.0006,
+ "loss": 4.087121963500977,
+ "step": 6489
+ },
+ {
+ "epoch": 90.13973799126637,
+ "grad_norm": 0.016007190570235252,
+ "learning_rate": 0.0006,
+ "loss": 3.98740816116333,
+ "step": 6490
+ },
+ {
+ "epoch": 90.15371179039302,
+ "grad_norm": 0.018072661012411118,
+ "learning_rate": 0.0006,
+ "loss": 4.1155009269714355,
+ "step": 6491
+ },
+ {
+ "epoch": 90.16768558951965,
+ "grad_norm": 0.016868166625499725,
+ "learning_rate": 0.0006,
+ "loss": 4.123727321624756,
+ "step": 6492
+ },
+ {
+ "epoch": 90.18165938864628,
+ "grad_norm": 0.017091726884245872,
+ "learning_rate": 0.0006,
+ "loss": 4.1405229568481445,
+ "step": 6493
+ },
+ {
+ "epoch": 90.19563318777293,
+ "grad_norm": 0.018329855054616928,
+ "learning_rate": 0.0006,
+ "loss": 4.093900680541992,
+ "step": 6494
+ },
+ {
+ "epoch": 90.20960698689956,
+ "grad_norm": 0.018906457349658012,
+ "learning_rate": 0.0006,
+ "loss": 4.0552849769592285,
+ "step": 6495
+ },
+ {
+ "epoch": 90.2235807860262,
+ "grad_norm": 0.017275869846343994,
+ "learning_rate": 0.0006,
+ "loss": 4.1310625076293945,
+ "step": 6496
+ },
+ {
+ "epoch": 90.23755458515284,
+ "grad_norm": 0.015029897913336754,
+ "learning_rate": 0.0006,
+ "loss": 4.100166320800781,
+ "step": 6497
+ },
+ {
+ "epoch": 90.25152838427948,
+ "grad_norm": 0.014891199767589569,
+ "learning_rate": 0.0006,
+ "loss": 4.1314802169799805,
+ "step": 6498
+ },
+ {
+ "epoch": 90.26550218340611,
+ "grad_norm": 0.014877690002322197,
+ "learning_rate": 0.0006,
+ "loss": 4.076098442077637,
+ "step": 6499
+ },
+ {
+ "epoch": 90.27947598253274,
+ "grad_norm": 0.014670864678919315,
+ "learning_rate": 0.0006,
+ "loss": 4.066137313842773,
+ "step": 6500
+ },
+ {
+ "epoch": 90.29344978165939,
+ "grad_norm": 0.01444234699010849,
+ "learning_rate": 0.0006,
+ "loss": 4.013591766357422,
+ "step": 6501
+ },
+ {
+ "epoch": 90.30742358078602,
+ "grad_norm": 0.01339263841509819,
+ "learning_rate": 0.0006,
+ "loss": 4.117959499359131,
+ "step": 6502
+ },
+ {
+ "epoch": 90.32139737991267,
+ "grad_norm": 0.014753810130059719,
+ "learning_rate": 0.0006,
+ "loss": 4.042947769165039,
+ "step": 6503
+ },
+ {
+ "epoch": 90.3353711790393,
+ "grad_norm": 0.014947964809834957,
+ "learning_rate": 0.0006,
+ "loss": 4.112810134887695,
+ "step": 6504
+ },
+ {
+ "epoch": 90.34934497816595,
+ "grad_norm": 0.015155353583395481,
+ "learning_rate": 0.0006,
+ "loss": 4.213214874267578,
+ "step": 6505
+ },
+ {
+ "epoch": 90.36331877729258,
+ "grad_norm": 0.015502216294407845,
+ "learning_rate": 0.0006,
+ "loss": 4.127619743347168,
+ "step": 6506
+ },
+ {
+ "epoch": 90.37729257641921,
+ "grad_norm": 0.015377781353890896,
+ "learning_rate": 0.0006,
+ "loss": 4.135862350463867,
+ "step": 6507
+ },
+ {
+ "epoch": 90.39126637554585,
+ "grad_norm": 0.01662006787955761,
+ "learning_rate": 0.0006,
+ "loss": 4.076877593994141,
+ "step": 6508
+ },
+ {
+ "epoch": 90.40524017467249,
+ "grad_norm": 0.016324395313858986,
+ "learning_rate": 0.0006,
+ "loss": 4.214895725250244,
+ "step": 6509
+ },
+ {
+ "epoch": 90.41921397379913,
+ "grad_norm": 0.016099024564027786,
+ "learning_rate": 0.0006,
+ "loss": 4.139717102050781,
+ "step": 6510
+ },
+ {
+ "epoch": 90.43318777292576,
+ "grad_norm": 0.014403503388166428,
+ "learning_rate": 0.0006,
+ "loss": 4.040343284606934,
+ "step": 6511
+ },
+ {
+ "epoch": 90.44716157205241,
+ "grad_norm": 0.01413416862487793,
+ "learning_rate": 0.0006,
+ "loss": 4.087702751159668,
+ "step": 6512
+ },
+ {
+ "epoch": 90.46113537117904,
+ "grad_norm": 0.01603989489376545,
+ "learning_rate": 0.0006,
+ "loss": 4.173558712005615,
+ "step": 6513
+ },
+ {
+ "epoch": 90.47510917030567,
+ "grad_norm": 0.016846206039190292,
+ "learning_rate": 0.0006,
+ "loss": 4.1955461502075195,
+ "step": 6514
+ },
+ {
+ "epoch": 90.48908296943232,
+ "grad_norm": 0.01652812585234642,
+ "learning_rate": 0.0006,
+ "loss": 4.061838150024414,
+ "step": 6515
+ },
+ {
+ "epoch": 90.50305676855895,
+ "grad_norm": 0.015100528486073017,
+ "learning_rate": 0.0006,
+ "loss": 4.041808605194092,
+ "step": 6516
+ },
+ {
+ "epoch": 90.5170305676856,
+ "grad_norm": 0.013867822475731373,
+ "learning_rate": 0.0006,
+ "loss": 4.12662410736084,
+ "step": 6517
+ },
+ {
+ "epoch": 90.53100436681223,
+ "grad_norm": 0.013684107922017574,
+ "learning_rate": 0.0006,
+ "loss": 4.1254048347473145,
+ "step": 6518
+ },
+ {
+ "epoch": 90.54497816593886,
+ "grad_norm": 0.014200013130903244,
+ "learning_rate": 0.0006,
+ "loss": 4.183216094970703,
+ "step": 6519
+ },
+ {
+ "epoch": 90.5589519650655,
+ "grad_norm": 0.013707022182643414,
+ "learning_rate": 0.0006,
+ "loss": 4.07176399230957,
+ "step": 6520
+ },
+ {
+ "epoch": 90.57292576419214,
+ "grad_norm": 0.013996582478284836,
+ "learning_rate": 0.0006,
+ "loss": 4.006575584411621,
+ "step": 6521
+ },
+ {
+ "epoch": 90.58689956331878,
+ "grad_norm": 0.014886210672557354,
+ "learning_rate": 0.0006,
+ "loss": 4.118956565856934,
+ "step": 6522
+ },
+ {
+ "epoch": 90.60087336244541,
+ "grad_norm": 0.014022685587406158,
+ "learning_rate": 0.0006,
+ "loss": 3.96986722946167,
+ "step": 6523
+ },
+ {
+ "epoch": 90.61484716157206,
+ "grad_norm": 0.014547634869813919,
+ "learning_rate": 0.0006,
+ "loss": 4.122886657714844,
+ "step": 6524
+ },
+ {
+ "epoch": 90.62882096069869,
+ "grad_norm": 0.013682851567864418,
+ "learning_rate": 0.0006,
+ "loss": 3.988292694091797,
+ "step": 6525
+ },
+ {
+ "epoch": 90.64279475982532,
+ "grad_norm": 0.014016421511769295,
+ "learning_rate": 0.0006,
+ "loss": 4.110270977020264,
+ "step": 6526
+ },
+ {
+ "epoch": 90.65676855895197,
+ "grad_norm": 0.013392233289778233,
+ "learning_rate": 0.0006,
+ "loss": 4.065059661865234,
+ "step": 6527
+ },
+ {
+ "epoch": 90.6707423580786,
+ "grad_norm": 0.012107795104384422,
+ "learning_rate": 0.0006,
+ "loss": 4.034260272979736,
+ "step": 6528
+ },
+ {
+ "epoch": 90.68471615720524,
+ "grad_norm": 0.013198604807257652,
+ "learning_rate": 0.0006,
+ "loss": 4.162083625793457,
+ "step": 6529
+ },
+ {
+ "epoch": 90.69868995633188,
+ "grad_norm": 0.012380579486489296,
+ "learning_rate": 0.0006,
+ "loss": 4.040456295013428,
+ "step": 6530
+ },
+ {
+ "epoch": 90.71266375545852,
+ "grad_norm": 0.012784135527908802,
+ "learning_rate": 0.0006,
+ "loss": 4.0823798179626465,
+ "step": 6531
+ },
+ {
+ "epoch": 90.72663755458515,
+ "grad_norm": 0.012920220382511616,
+ "learning_rate": 0.0006,
+ "loss": 4.181064128875732,
+ "step": 6532
+ },
+ {
+ "epoch": 90.74061135371178,
+ "grad_norm": 0.01338814664632082,
+ "learning_rate": 0.0006,
+ "loss": 4.111660003662109,
+ "step": 6533
+ },
+ {
+ "epoch": 90.75458515283843,
+ "grad_norm": 0.013628040440380573,
+ "learning_rate": 0.0006,
+ "loss": 4.237852573394775,
+ "step": 6534
+ },
+ {
+ "epoch": 90.76855895196506,
+ "grad_norm": 0.013384854421019554,
+ "learning_rate": 0.0006,
+ "loss": 4.121862888336182,
+ "step": 6535
+ },
+ {
+ "epoch": 90.78253275109171,
+ "grad_norm": 0.014726929366588593,
+ "learning_rate": 0.0006,
+ "loss": 4.1695990562438965,
+ "step": 6536
+ },
+ {
+ "epoch": 90.79650655021834,
+ "grad_norm": 0.0161006897687912,
+ "learning_rate": 0.0006,
+ "loss": 4.17794132232666,
+ "step": 6537
+ },
+ {
+ "epoch": 90.81048034934499,
+ "grad_norm": 0.016761884093284607,
+ "learning_rate": 0.0006,
+ "loss": 4.093764781951904,
+ "step": 6538
+ },
+ {
+ "epoch": 90.82445414847162,
+ "grad_norm": 0.017246825620532036,
+ "learning_rate": 0.0006,
+ "loss": 3.9382331371307373,
+ "step": 6539
+ },
+ {
+ "epoch": 90.83842794759825,
+ "grad_norm": 0.018141290172934532,
+ "learning_rate": 0.0006,
+ "loss": 3.973609209060669,
+ "step": 6540
+ },
+ {
+ "epoch": 90.8524017467249,
+ "grad_norm": 0.018129250034689903,
+ "learning_rate": 0.0006,
+ "loss": 4.227414131164551,
+ "step": 6541
+ },
+ {
+ "epoch": 90.86637554585153,
+ "grad_norm": 0.018234573304653168,
+ "learning_rate": 0.0006,
+ "loss": 4.151870250701904,
+ "step": 6542
+ },
+ {
+ "epoch": 90.88034934497817,
+ "grad_norm": 0.01560172438621521,
+ "learning_rate": 0.0006,
+ "loss": 4.127397537231445,
+ "step": 6543
+ },
+ {
+ "epoch": 90.8943231441048,
+ "grad_norm": 0.014367681927978992,
+ "learning_rate": 0.0006,
+ "loss": 4.17820930480957,
+ "step": 6544
+ },
+ {
+ "epoch": 90.90829694323143,
+ "grad_norm": 0.016071319580078125,
+ "learning_rate": 0.0006,
+ "loss": 4.059854507446289,
+ "step": 6545
+ },
+ {
+ "epoch": 90.92227074235808,
+ "grad_norm": 0.01589169353246689,
+ "learning_rate": 0.0006,
+ "loss": 4.282748222351074,
+ "step": 6546
+ },
+ {
+ "epoch": 90.93624454148471,
+ "grad_norm": 0.013639772310853004,
+ "learning_rate": 0.0006,
+ "loss": 4.147368907928467,
+ "step": 6547
+ },
+ {
+ "epoch": 90.95021834061136,
+ "grad_norm": 0.014248427003622055,
+ "learning_rate": 0.0006,
+ "loss": 4.240248680114746,
+ "step": 6548
+ },
+ {
+ "epoch": 90.96419213973799,
+ "grad_norm": 0.014600320719182491,
+ "learning_rate": 0.0006,
+ "loss": 4.122123718261719,
+ "step": 6549
+ },
+ {
+ "epoch": 90.97816593886463,
+ "grad_norm": 0.014371675439178944,
+ "learning_rate": 0.0006,
+ "loss": 4.013302803039551,
+ "step": 6550
+ },
+ {
+ "epoch": 90.99213973799127,
+ "grad_norm": 0.013133526779711246,
+ "learning_rate": 0.0006,
+ "loss": 4.097136497497559,
+ "step": 6551
+ },
+ {
+ "epoch": 91.0,
+ "grad_norm": 0.015599353238940239,
+ "learning_rate": 0.0006,
+ "loss": 4.198554039001465,
+ "step": 6552
+ },
+ {
+ "epoch": 91.0,
+ "eval_loss": 4.603907585144043,
+ "eval_runtime": 56.9418,
+ "eval_samples_per_second": 42.886,
+ "eval_steps_per_second": 1.352,
+ "step": 6552
+ },
+ {
+ "epoch": 91.01397379912663,
+ "grad_norm": 0.015917908400297165,
+ "learning_rate": 0.0006,
+ "loss": 4.133877754211426,
+ "step": 6553
+ },
+ {
+ "epoch": 91.02794759825328,
+ "grad_norm": 0.015485800802707672,
+ "learning_rate": 0.0006,
+ "loss": 4.111617088317871,
+ "step": 6554
+ },
+ {
+ "epoch": 91.04192139737991,
+ "grad_norm": 0.012969603762030602,
+ "learning_rate": 0.0006,
+ "loss": 4.064828395843506,
+ "step": 6555
+ },
+ {
+ "epoch": 91.05589519650655,
+ "grad_norm": 0.014482097700238228,
+ "learning_rate": 0.0006,
+ "loss": 4.101710319519043,
+ "step": 6556
+ },
+ {
+ "epoch": 91.06986899563319,
+ "grad_norm": 0.017242740839719772,
+ "learning_rate": 0.0006,
+ "loss": 4.015342712402344,
+ "step": 6557
+ },
+ {
+ "epoch": 91.08384279475983,
+ "grad_norm": 0.017100440338253975,
+ "learning_rate": 0.0006,
+ "loss": 3.9801602363586426,
+ "step": 6558
+ },
+ {
+ "epoch": 91.09781659388646,
+ "grad_norm": 0.015969131141901016,
+ "learning_rate": 0.0006,
+ "loss": 4.049958229064941,
+ "step": 6559
+ },
+ {
+ "epoch": 91.1117903930131,
+ "grad_norm": 0.016018927097320557,
+ "learning_rate": 0.0006,
+ "loss": 4.094139099121094,
+ "step": 6560
+ },
+ {
+ "epoch": 91.12576419213974,
+ "grad_norm": 0.018131926655769348,
+ "learning_rate": 0.0006,
+ "loss": 3.9499974250793457,
+ "step": 6561
+ },
+ {
+ "epoch": 91.13973799126637,
+ "grad_norm": 0.019425220787525177,
+ "learning_rate": 0.0006,
+ "loss": 3.9910552501678467,
+ "step": 6562
+ },
+ {
+ "epoch": 91.15371179039302,
+ "grad_norm": 0.01908663846552372,
+ "learning_rate": 0.0006,
+ "loss": 4.032179832458496,
+ "step": 6563
+ },
+ {
+ "epoch": 91.16768558951965,
+ "grad_norm": 0.01871207356452942,
+ "learning_rate": 0.0006,
+ "loss": 4.112380027770996,
+ "step": 6564
+ },
+ {
+ "epoch": 91.18165938864628,
+ "grad_norm": 0.020867953076958656,
+ "learning_rate": 0.0006,
+ "loss": 4.06105899810791,
+ "step": 6565
+ },
+ {
+ "epoch": 91.19563318777293,
+ "grad_norm": 0.019053412601351738,
+ "learning_rate": 0.0006,
+ "loss": 3.992504596710205,
+ "step": 6566
+ },
+ {
+ "epoch": 91.20960698689956,
+ "grad_norm": 0.016084402799606323,
+ "learning_rate": 0.0006,
+ "loss": 4.100559234619141,
+ "step": 6567
+ },
+ {
+ "epoch": 91.2235807860262,
+ "grad_norm": 0.016857590526342392,
+ "learning_rate": 0.0006,
+ "loss": 4.1567277908325195,
+ "step": 6568
+ },
+ {
+ "epoch": 91.23755458515284,
+ "grad_norm": 0.01702957972884178,
+ "learning_rate": 0.0006,
+ "loss": 4.234771728515625,
+ "step": 6569
+ },
+ {
+ "epoch": 91.25152838427948,
+ "grad_norm": 0.016869494691491127,
+ "learning_rate": 0.0006,
+ "loss": 3.9931387901306152,
+ "step": 6570
+ },
+ {
+ "epoch": 91.26550218340611,
+ "grad_norm": 0.01696581020951271,
+ "learning_rate": 0.0006,
+ "loss": 4.184908866882324,
+ "step": 6571
+ },
+ {
+ "epoch": 91.27947598253274,
+ "grad_norm": 0.014982126653194427,
+ "learning_rate": 0.0006,
+ "loss": 4.036668300628662,
+ "step": 6572
+ },
+ {
+ "epoch": 91.29344978165939,
+ "grad_norm": 0.014363301917910576,
+ "learning_rate": 0.0006,
+ "loss": 4.074010848999023,
+ "step": 6573
+ },
+ {
+ "epoch": 91.30742358078602,
+ "grad_norm": 0.015492337755858898,
+ "learning_rate": 0.0006,
+ "loss": 4.127626419067383,
+ "step": 6574
+ },
+ {
+ "epoch": 91.32139737991267,
+ "grad_norm": 0.015375050716102123,
+ "learning_rate": 0.0006,
+ "loss": 4.159229755401611,
+ "step": 6575
+ },
+ {
+ "epoch": 91.3353711790393,
+ "grad_norm": 0.015068565495312214,
+ "learning_rate": 0.0006,
+ "loss": 4.172623634338379,
+ "step": 6576
+ },
+ {
+ "epoch": 91.34934497816595,
+ "grad_norm": 0.0140256742015481,
+ "learning_rate": 0.0006,
+ "loss": 4.207065105438232,
+ "step": 6577
+ },
+ {
+ "epoch": 91.36331877729258,
+ "grad_norm": 0.014220152981579304,
+ "learning_rate": 0.0006,
+ "loss": 4.131444931030273,
+ "step": 6578
+ },
+ {
+ "epoch": 91.37729257641921,
+ "grad_norm": 0.0143149234354496,
+ "learning_rate": 0.0006,
+ "loss": 4.065218448638916,
+ "step": 6579
+ },
+ {
+ "epoch": 91.39126637554585,
+ "grad_norm": 0.013149659149348736,
+ "learning_rate": 0.0006,
+ "loss": 4.106743335723877,
+ "step": 6580
+ },
+ {
+ "epoch": 91.40524017467249,
+ "grad_norm": 0.014550330117344856,
+ "learning_rate": 0.0006,
+ "loss": 4.123722553253174,
+ "step": 6581
+ },
+ {
+ "epoch": 91.41921397379913,
+ "grad_norm": 0.014611710794270039,
+ "learning_rate": 0.0006,
+ "loss": 4.079610824584961,
+ "step": 6582
+ },
+ {
+ "epoch": 91.43318777292576,
+ "grad_norm": 0.013369769789278507,
+ "learning_rate": 0.0006,
+ "loss": 4.133187770843506,
+ "step": 6583
+ },
+ {
+ "epoch": 91.44716157205241,
+ "grad_norm": 0.013243545778095722,
+ "learning_rate": 0.0006,
+ "loss": 4.042607307434082,
+ "step": 6584
+ },
+ {
+ "epoch": 91.46113537117904,
+ "grad_norm": 0.013388266786932945,
+ "learning_rate": 0.0006,
+ "loss": 4.14863395690918,
+ "step": 6585
+ },
+ {
+ "epoch": 91.47510917030567,
+ "grad_norm": 0.014333273284137249,
+ "learning_rate": 0.0006,
+ "loss": 4.109699726104736,
+ "step": 6586
+ },
+ {
+ "epoch": 91.48908296943232,
+ "grad_norm": 0.012742757797241211,
+ "learning_rate": 0.0006,
+ "loss": 4.126582145690918,
+ "step": 6587
+ },
+ {
+ "epoch": 91.50305676855895,
+ "grad_norm": 0.012880304828286171,
+ "learning_rate": 0.0006,
+ "loss": 4.205300331115723,
+ "step": 6588
+ },
+ {
+ "epoch": 91.5170305676856,
+ "grad_norm": 0.013896043412387371,
+ "learning_rate": 0.0006,
+ "loss": 4.132510185241699,
+ "step": 6589
+ },
+ {
+ "epoch": 91.53100436681223,
+ "grad_norm": 0.014361281879246235,
+ "learning_rate": 0.0006,
+ "loss": 4.1487932205200195,
+ "step": 6590
+ },
+ {
+ "epoch": 91.54497816593886,
+ "grad_norm": 0.015137741342186928,
+ "learning_rate": 0.0006,
+ "loss": 4.1564531326293945,
+ "step": 6591
+ },
+ {
+ "epoch": 91.5589519650655,
+ "grad_norm": 0.014660682529211044,
+ "learning_rate": 0.0006,
+ "loss": 4.056064128875732,
+ "step": 6592
+ },
+ {
+ "epoch": 91.57292576419214,
+ "grad_norm": 0.014888476580381393,
+ "learning_rate": 0.0006,
+ "loss": 4.0633344650268555,
+ "step": 6593
+ },
+ {
+ "epoch": 91.58689956331878,
+ "grad_norm": 0.012162433005869389,
+ "learning_rate": 0.0006,
+ "loss": 4.026293754577637,
+ "step": 6594
+ },
+ {
+ "epoch": 91.60087336244541,
+ "grad_norm": 0.012892271392047405,
+ "learning_rate": 0.0006,
+ "loss": 4.267086029052734,
+ "step": 6595
+ },
+ {
+ "epoch": 91.61484716157206,
+ "grad_norm": 0.01450759544968605,
+ "learning_rate": 0.0006,
+ "loss": 4.149359703063965,
+ "step": 6596
+ },
+ {
+ "epoch": 91.62882096069869,
+ "grad_norm": 0.01419843453913927,
+ "learning_rate": 0.0006,
+ "loss": 4.094183921813965,
+ "step": 6597
+ },
+ {
+ "epoch": 91.64279475982532,
+ "grad_norm": 0.014401616528630257,
+ "learning_rate": 0.0006,
+ "loss": 4.125172138214111,
+ "step": 6598
+ },
+ {
+ "epoch": 91.65676855895197,
+ "grad_norm": 0.013945513404905796,
+ "learning_rate": 0.0006,
+ "loss": 4.146623134613037,
+ "step": 6599
+ },
+ {
+ "epoch": 91.6707423580786,
+ "grad_norm": 0.013516047969460487,
+ "learning_rate": 0.0006,
+ "loss": 4.160526275634766,
+ "step": 6600
+ },
+ {
+ "epoch": 91.68471615720524,
+ "grad_norm": 0.013262378051877022,
+ "learning_rate": 0.0006,
+ "loss": 4.074923038482666,
+ "step": 6601
+ },
+ {
+ "epoch": 91.69868995633188,
+ "grad_norm": 0.014365669339895248,
+ "learning_rate": 0.0006,
+ "loss": 4.118094444274902,
+ "step": 6602
+ },
+ {
+ "epoch": 91.71266375545852,
+ "grad_norm": 0.016713734716176987,
+ "learning_rate": 0.0006,
+ "loss": 4.021799087524414,
+ "step": 6603
+ },
+ {
+ "epoch": 91.72663755458515,
+ "grad_norm": 0.01988680474460125,
+ "learning_rate": 0.0006,
+ "loss": 4.238436222076416,
+ "step": 6604
+ },
+ {
+ "epoch": 91.74061135371178,
+ "grad_norm": 0.022749733179807663,
+ "learning_rate": 0.0006,
+ "loss": 4.010470390319824,
+ "step": 6605
+ },
+ {
+ "epoch": 91.75458515283843,
+ "grad_norm": 0.022533461451530457,
+ "learning_rate": 0.0006,
+ "loss": 4.152201175689697,
+ "step": 6606
+ },
+ {
+ "epoch": 91.76855895196506,
+ "grad_norm": 0.02021082676947117,
+ "learning_rate": 0.0006,
+ "loss": 4.120572090148926,
+ "step": 6607
+ },
+ {
+ "epoch": 91.78253275109171,
+ "grad_norm": 0.016298197209835052,
+ "learning_rate": 0.0006,
+ "loss": 4.115815162658691,
+ "step": 6608
+ },
+ {
+ "epoch": 91.79650655021834,
+ "grad_norm": 0.017562786117196083,
+ "learning_rate": 0.0006,
+ "loss": 4.109724044799805,
+ "step": 6609
+ },
+ {
+ "epoch": 91.81048034934499,
+ "grad_norm": 0.018153153359889984,
+ "learning_rate": 0.0006,
+ "loss": 4.143759727478027,
+ "step": 6610
+ },
+ {
+ "epoch": 91.82445414847162,
+ "grad_norm": 0.014762306585907936,
+ "learning_rate": 0.0006,
+ "loss": 4.25596809387207,
+ "step": 6611
+ },
+ {
+ "epoch": 91.83842794759825,
+ "grad_norm": 0.015074343420565128,
+ "learning_rate": 0.0006,
+ "loss": 4.207723140716553,
+ "step": 6612
+ },
+ {
+ "epoch": 91.8524017467249,
+ "grad_norm": 0.015028579160571098,
+ "learning_rate": 0.0006,
+ "loss": 4.01950216293335,
+ "step": 6613
+ },
+ {
+ "epoch": 91.86637554585153,
+ "grad_norm": 0.013996795751154423,
+ "learning_rate": 0.0006,
+ "loss": 4.132997989654541,
+ "step": 6614
+ },
+ {
+ "epoch": 91.88034934497817,
+ "grad_norm": 0.014978236518800259,
+ "learning_rate": 0.0006,
+ "loss": 4.132402420043945,
+ "step": 6615
+ },
+ {
+ "epoch": 91.8943231441048,
+ "grad_norm": 0.013111517764627934,
+ "learning_rate": 0.0006,
+ "loss": 4.1223649978637695,
+ "step": 6616
+ },
+ {
+ "epoch": 91.90829694323143,
+ "grad_norm": 0.013872234150767326,
+ "learning_rate": 0.0006,
+ "loss": 4.096590042114258,
+ "step": 6617
+ },
+ {
+ "epoch": 91.92227074235808,
+ "grad_norm": 0.013642470352351665,
+ "learning_rate": 0.0006,
+ "loss": 4.236227035522461,
+ "step": 6618
+ },
+ {
+ "epoch": 91.93624454148471,
+ "grad_norm": 0.014684873633086681,
+ "learning_rate": 0.0006,
+ "loss": 4.037961959838867,
+ "step": 6619
+ },
+ {
+ "epoch": 91.95021834061136,
+ "grad_norm": 0.014498206786811352,
+ "learning_rate": 0.0006,
+ "loss": 4.043325424194336,
+ "step": 6620
+ },
+ {
+ "epoch": 91.96419213973799,
+ "grad_norm": 0.012768656946718693,
+ "learning_rate": 0.0006,
+ "loss": 4.138803482055664,
+ "step": 6621
+ },
+ {
+ "epoch": 91.97816593886463,
+ "grad_norm": 0.012155905365943909,
+ "learning_rate": 0.0006,
+ "loss": 4.163424968719482,
+ "step": 6622
+ },
+ {
+ "epoch": 91.99213973799127,
+ "grad_norm": 0.012584930285811424,
+ "learning_rate": 0.0006,
+ "loss": 4.086629867553711,
+ "step": 6623
+ },
+ {
+ "epoch": 92.0,
+ "grad_norm": 0.01578151248395443,
+ "learning_rate": 0.0006,
+ "loss": 4.108386993408203,
+ "step": 6624
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 28800,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 400,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": true,
+ "should_log": false,
+ "should_save": false,
+ "should_training_stop": false
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 2.824044292425646e+18,
+ "train_batch_size": 8,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6624/training_args.bin b/runs/l2r50-i2-fulle-lm/checkpoint-6624/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..cad232831c222cc7f8b938b8bbe42f61e7d89c3a
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-6624/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:f7c40e3bfa2f5ac1450ca1243e3f2a895d44a06ed5995e6aefd97f78d229d8c8
+size 4792
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6912/chat_template.jinja b/runs/l2r50-i2-fulle-lm/checkpoint-6912/chat_template.jinja
new file mode 100644
index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-6912/chat_template.jinja
@@ -0,0 +1,85 @@
+{%- if tools %}
+ {{- '<|im_start|>system\n' }}
+ {%- if messages[0].role == 'system' %}
+ {{- messages[0].content + '\n\n' }}
+ {%- endif %}
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }}
+ {%- for tool in tools %}
+ {{- "\n" }}
+ {{- tool | tojson }}
+ {%- endfor %}
+ {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }}
+{%- else %}
+ {%- if messages[0].role == 'system' %}
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
+ {%- endif %}
+{%- endif %}
+{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
+{%- for message in messages[::-1] %}
+ {%- set index = (messages|length - 1) - loop.index0 %}
+ {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %}
+ {%- set ns.multi_step_tool = false %}
+ {%- set ns.last_query_index = index %}
+ {%- endif %}
+{%- endfor %}
+{%- for message in messages %}
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
+ {%- elif message.role == "assistant" %}
+ {%- set content = message.content %}
+ {%- set reasoning_content = '' %}
+ {%- if message.reasoning_content is defined and message.reasoning_content is not none %}
+ {%- set reasoning_content = message.reasoning_content %}
+ {%- else %}
+ {%- if '' in message.content %}
+ {%- set content = message.content.split('')[-1].lstrip('\n') %}
+ {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %}
+ {%- endif %}
+ {%- endif %}
+ {%- if loop.index0 > ns.last_query_index %}
+ {%- if loop.last or (not loop.last and reasoning_content) %}
+ {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- if message.tool_calls %}
+ {%- for tool_call in message.tool_calls %}
+ {%- if (loop.first and content) or (not loop.first) %}
+ {{- '\n' }}
+ {%- endif %}
+ {%- if tool_call.function %}
+ {%- set tool_call = tool_call.function %}
+ {%- endif %}
+ {{- '\n{"name": "' }}
+ {{- tool_call.name }}
+ {{- '", "arguments": ' }}
+ {%- if tool_call.arguments is string %}
+ {{- tool_call.arguments }}
+ {%- else %}
+ {{- tool_call.arguments | tojson }}
+ {%- endif %}
+ {{- '}\n' }}
+ {%- endfor %}
+ {%- endif %}
+ {{- '<|im_end|>\n' }}
+ {%- elif message.role == "tool" %}
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
+ {{- '<|im_start|>user' }}
+ {%- endif %}
+ {{- '\n\n' }}
+ {{- message.content }}
+ {{- '\n' }}
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
+ {{- '<|im_end|>\n' }}
+ {%- endif %}
+ {%- endif %}
+{%- endfor %}
+{%- if add_generation_prompt %}
+ {{- '<|im_start|>assistant\n' }}
+ {%- if enable_thinking is defined and enable_thinking is false %}
+ {{- '\n\n\n\n' }}
+ {%- endif %}
+{%- endif %}
\ No newline at end of file
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6912/config.json b/runs/l2r50-i2-fulle-lm/checkpoint-6912/config.json
new file mode 100644
index 0000000000000000000000000000000000000000..22f250953c6ce8fc91ee4411856dd73041a415e2
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-6912/config.json
@@ -0,0 +1,32 @@
+{
+ "architectures": [
+ "LlamaForCausalLM"
+ ],
+ "attention_bias": false,
+ "attention_dropout": 0.0,
+ "bos_token_id": null,
+ "dtype": "float32",
+ "eos_token_id": 151645,
+ "head_dim": 128,
+ "hidden_act": "silu",
+ "hidden_size": 512,
+ "initializer_range": 0.02,
+ "intermediate_size": 1536,
+ "max_position_embeddings": 2048,
+ "mlp_bias": false,
+ "model_type": "llama",
+ "num_attention_heads": 4,
+ "num_hidden_layers": 20,
+ "num_key_value_heads": 4,
+ "pad_token_id": 151645,
+ "pretraining_tp": 1,
+ "rms_norm_eps": 1e-06,
+ "rope_parameters": {
+ "rope_theta": 10000.0,
+ "rope_type": "default"
+ },
+ "tie_word_embeddings": true,
+ "transformers_version": "5.5.0",
+ "use_cache": false,
+ "vocab_size": 151673
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6912/generation_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-6912/generation_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-6912/generation_config.json
@@ -0,0 +1,11 @@
+{
+ "_from_model_config": true,
+ "eos_token_id": [
+ 151645
+ ],
+ "output_attentions": false,
+ "output_hidden_states": false,
+ "pad_token_id": 151645,
+ "transformers_version": "5.5.0",
+ "use_cache": true
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6912/model.safetensors b/runs/l2r50-i2-fulle-lm/checkpoint-6912/model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..3e4c07da00f264e39f747e36434e38eec1c1c2cb
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-6912/model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:0f71ef97c4ea096ec954b9050ef9e62a495733b2317bbf71f968b5a2413758c9
+size 583360328
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6912/optimizer.pt b/runs/l2r50-i2-fulle-lm/checkpoint-6912/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..d15bceb7ef69347bf8cf9394a5d2bd6e5910741f
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-6912/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:999438d24e47e87ad0493588d6f4cf83f9e0860905cd03e026fcb40d7c1a8a1e
+size 1166833530
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6912/rng_state_0.pth b/runs/l2r50-i2-fulle-lm/checkpoint-6912/rng_state_0.pth
new file mode 100644
index 0000000000000000000000000000000000000000..c3ef1d6a71abfca102fd404f57354f3ca09ee701
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-6912/rng_state_0.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:ac169ba05fadeecfe00246f94826dca416d3fc2b1287c215fafb508d79aec1a2
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6912/rng_state_1.pth b/runs/l2r50-i2-fulle-lm/checkpoint-6912/rng_state_1.pth
new file mode 100644
index 0000000000000000000000000000000000000000..5fbf0d7eae7fc8d4dda5a5846601c548e141e1d0
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-6912/rng_state_1.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:40d6b8065c0a4634082fe5db4e072a73cbf9df22d0f465f7470ee490e85b8f06
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6912/rng_state_2.pth b/runs/l2r50-i2-fulle-lm/checkpoint-6912/rng_state_2.pth
new file mode 100644
index 0000000000000000000000000000000000000000..0c1428cff0ef70846cc7092b335c53bf4fc384b0
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-6912/rng_state_2.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:6b339702460b81003d8a5ee7890a1776405165efac6a55c01b2b259ef1c27513
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6912/rng_state_3.pth b/runs/l2r50-i2-fulle-lm/checkpoint-6912/rng_state_3.pth
new file mode 100644
index 0000000000000000000000000000000000000000..cd8a97b2a3ecba19c6ed2d04fbb06d29237b4ec3
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-6912/rng_state_3.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:ef39912fd9c5e4d2a49b252fb2e37c1d5bb81162d1ad99afecfcd996b765fec9
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6912/scheduler.pt b/runs/l2r50-i2-fulle-lm/checkpoint-6912/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..80771355cefe444a789dc3e83678b2d55d3b8db3
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-6912/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:0d0b661f1b18caa6118d27fef199882abc2f72f0aaf9b16c81f9627cbcf08e00
+size 1064
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6912/tokenizer.json b/runs/l2r50-i2-fulle-lm/checkpoint-6912/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..53786016f070e11184813f3c83101dc5e93e392e
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-6912/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:c3c63fda1646a555448e8f03662542e04b11d103721319e4d9b267d7da03ceb8
+size 11424463
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6912/tokenizer_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-6912/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..98e2fb9b2cc54785dc9d871b8e503f8b8f2defe7
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-6912/tokenizer_config.json
@@ -0,0 +1,20 @@
+{
+ "add_prefix_space": false,
+ "backend": "tokenizers",
+ "bos_token": null,
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "<|im_end|>",
+ "errors": "replace",
+ "extra_special_tokens": [
+ "<|l2r_pred|>",
+ "<|r2l_pred|>",
+ "<|next_1_pred|>",
+ "<|next_2_pred|>"
+ ],
+ "is_local": false,
+ "model_max_length": 131072,
+ "pad_token": "<|im_end|>",
+ "split_special_tokens": false,
+ "tokenizer_class": "Qwen2Tokenizer",
+ "unk_token": null
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6912/trainer_state.json b/runs/l2r50-i2-fulle-lm/checkpoint-6912/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..ca31af02fdcf45c3b7dcdf6824df426dec965568
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-6912/trainer_state.json
@@ -0,0 +1,49178 @@
+{
+ "best_global_step": null,
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 96.0,
+ "eval_steps": 500,
+ "global_step": 6912,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.013973799126637555,
+ "grad_norm": 0.1739506870508194,
+ "learning_rate": 0.0,
+ "loss": 11.99299430847168,
+ "step": 1
+ },
+ {
+ "epoch": 0.02794759825327511,
+ "grad_norm": 0.17216597497463226,
+ "learning_rate": 5.999999999999999e-06,
+ "loss": 11.988704681396484,
+ "step": 2
+ },
+ {
+ "epoch": 0.04192139737991266,
+ "grad_norm": 0.17459715902805328,
+ "learning_rate": 1.1999999999999999e-05,
+ "loss": 11.945581436157227,
+ "step": 3
+ },
+ {
+ "epoch": 0.05589519650655022,
+ "grad_norm": 0.17577414214611053,
+ "learning_rate": 1.7999999999999997e-05,
+ "loss": 11.859772682189941,
+ "step": 4
+ },
+ {
+ "epoch": 0.06986899563318777,
+ "grad_norm": 0.16163651645183563,
+ "learning_rate": 2.3999999999999997e-05,
+ "loss": 11.744415283203125,
+ "step": 5
+ },
+ {
+ "epoch": 0.08384279475982533,
+ "grad_norm": 0.14409932494163513,
+ "learning_rate": 2.9999999999999997e-05,
+ "loss": 11.618099212646484,
+ "step": 6
+ },
+ {
+ "epoch": 0.09781659388646288,
+ "grad_norm": 0.13079848885536194,
+ "learning_rate": 3.5999999999999994e-05,
+ "loss": 11.505935668945312,
+ "step": 7
+ },
+ {
+ "epoch": 0.11179039301310044,
+ "grad_norm": 0.11964613944292068,
+ "learning_rate": 4.2e-05,
+ "loss": 11.416532516479492,
+ "step": 8
+ },
+ {
+ "epoch": 0.125764192139738,
+ "grad_norm": 0.11345142871141434,
+ "learning_rate": 4.7999999999999994e-05,
+ "loss": 11.33206558227539,
+ "step": 9
+ },
+ {
+ "epoch": 0.13973799126637554,
+ "grad_norm": 0.11033952981233597,
+ "learning_rate": 5.399999999999999e-05,
+ "loss": 11.259052276611328,
+ "step": 10
+ },
+ {
+ "epoch": 0.1537117903930131,
+ "grad_norm": 0.10776489973068237,
+ "learning_rate": 5.9999999999999995e-05,
+ "loss": 11.203384399414062,
+ "step": 11
+ },
+ {
+ "epoch": 0.16768558951965065,
+ "grad_norm": 0.10721946507692337,
+ "learning_rate": 6.599999999999999e-05,
+ "loss": 11.153295516967773,
+ "step": 12
+ },
+ {
+ "epoch": 0.18165938864628822,
+ "grad_norm": 0.10673430562019348,
+ "learning_rate": 7.199999999999999e-05,
+ "loss": 11.110200881958008,
+ "step": 13
+ },
+ {
+ "epoch": 0.19563318777292577,
+ "grad_norm": 0.10630254447460175,
+ "learning_rate": 7.8e-05,
+ "loss": 11.067477226257324,
+ "step": 14
+ },
+ {
+ "epoch": 0.2096069868995633,
+ "grad_norm": 0.10631615668535233,
+ "learning_rate": 8.4e-05,
+ "loss": 11.022941589355469,
+ "step": 15
+ },
+ {
+ "epoch": 0.22358078602620088,
+ "grad_norm": 0.10609443485736847,
+ "learning_rate": 8.999999999999999e-05,
+ "loss": 10.970746040344238,
+ "step": 16
+ },
+ {
+ "epoch": 0.23755458515283842,
+ "grad_norm": 0.10648232698440552,
+ "learning_rate": 9.599999999999999e-05,
+ "loss": 10.91267204284668,
+ "step": 17
+ },
+ {
+ "epoch": 0.251528384279476,
+ "grad_norm": 0.10584589838981628,
+ "learning_rate": 0.000102,
+ "loss": 10.854547500610352,
+ "step": 18
+ },
+ {
+ "epoch": 0.26550218340611353,
+ "grad_norm": 0.10591720044612885,
+ "learning_rate": 0.00010799999999999998,
+ "loss": 10.788917541503906,
+ "step": 19
+ },
+ {
+ "epoch": 0.2794759825327511,
+ "grad_norm": 0.10611672699451447,
+ "learning_rate": 0.00011399999999999999,
+ "loss": 10.717556953430176,
+ "step": 20
+ },
+ {
+ "epoch": 0.2934497816593886,
+ "grad_norm": 0.10455359518527985,
+ "learning_rate": 0.00011999999999999999,
+ "loss": 10.650186538696289,
+ "step": 21
+ },
+ {
+ "epoch": 0.3074235807860262,
+ "grad_norm": 0.1047525629401207,
+ "learning_rate": 0.00012599999999999997,
+ "loss": 10.56808090209961,
+ "step": 22
+ },
+ {
+ "epoch": 0.32139737991266376,
+ "grad_norm": 0.10511605441570282,
+ "learning_rate": 0.00013199999999999998,
+ "loss": 10.484028816223145,
+ "step": 23
+ },
+ {
+ "epoch": 0.3353711790393013,
+ "grad_norm": 0.10467177629470825,
+ "learning_rate": 0.000138,
+ "loss": 10.399947166442871,
+ "step": 24
+ },
+ {
+ "epoch": 0.34934497816593885,
+ "grad_norm": 0.10432136058807373,
+ "learning_rate": 0.00014399999999999998,
+ "loss": 10.312714576721191,
+ "step": 25
+ },
+ {
+ "epoch": 0.36331877729257644,
+ "grad_norm": 0.10527262091636658,
+ "learning_rate": 0.00015,
+ "loss": 10.212261199951172,
+ "step": 26
+ },
+ {
+ "epoch": 0.377292576419214,
+ "grad_norm": 0.103610560297966,
+ "learning_rate": 0.000156,
+ "loss": 10.132034301757812,
+ "step": 27
+ },
+ {
+ "epoch": 0.39126637554585153,
+ "grad_norm": 0.10553059726953506,
+ "learning_rate": 0.000162,
+ "loss": 10.020434379577637,
+ "step": 28
+ },
+ {
+ "epoch": 0.4052401746724891,
+ "grad_norm": 0.1037299782037735,
+ "learning_rate": 0.000168,
+ "loss": 9.934466361999512,
+ "step": 29
+ },
+ {
+ "epoch": 0.4192139737991266,
+ "grad_norm": 0.1042739674448967,
+ "learning_rate": 0.00017399999999999997,
+ "loss": 9.822539329528809,
+ "step": 30
+ },
+ {
+ "epoch": 0.4331877729257642,
+ "grad_norm": 0.10300865024328232,
+ "learning_rate": 0.00017999999999999998,
+ "loss": 9.734994888305664,
+ "step": 31
+ },
+ {
+ "epoch": 0.44716157205240176,
+ "grad_norm": 0.10291654616594315,
+ "learning_rate": 0.000186,
+ "loss": 9.624505996704102,
+ "step": 32
+ },
+ {
+ "epoch": 0.4611353711790393,
+ "grad_norm": 0.10210851579904556,
+ "learning_rate": 0.00019199999999999998,
+ "loss": 9.518205642700195,
+ "step": 33
+ },
+ {
+ "epoch": 0.47510917030567684,
+ "grad_norm": 0.10222529619932175,
+ "learning_rate": 0.000198,
+ "loss": 9.412222862243652,
+ "step": 34
+ },
+ {
+ "epoch": 0.4890829694323144,
+ "grad_norm": 0.09992019087076187,
+ "learning_rate": 0.000204,
+ "loss": 9.326566696166992,
+ "step": 35
+ },
+ {
+ "epoch": 0.503056768558952,
+ "grad_norm": 0.09931084513664246,
+ "learning_rate": 0.00020999999999999998,
+ "loss": 9.210111618041992,
+ "step": 36
+ },
+ {
+ "epoch": 0.5170305676855895,
+ "grad_norm": 0.09928256273269653,
+ "learning_rate": 0.00021599999999999996,
+ "loss": 9.100493431091309,
+ "step": 37
+ },
+ {
+ "epoch": 0.5310043668122271,
+ "grad_norm": 0.09843816608190536,
+ "learning_rate": 0.00022199999999999998,
+ "loss": 8.996543884277344,
+ "step": 38
+ },
+ {
+ "epoch": 0.5449781659388646,
+ "grad_norm": 0.09825374186038971,
+ "learning_rate": 0.00022799999999999999,
+ "loss": 8.888991355895996,
+ "step": 39
+ },
+ {
+ "epoch": 0.5589519650655022,
+ "grad_norm": 0.09602385759353638,
+ "learning_rate": 0.000234,
+ "loss": 8.802907943725586,
+ "step": 40
+ },
+ {
+ "epoch": 0.5729257641921397,
+ "grad_norm": 0.09414064139127731,
+ "learning_rate": 0.00023999999999999998,
+ "loss": 8.715423583984375,
+ "step": 41
+ },
+ {
+ "epoch": 0.5868995633187772,
+ "grad_norm": 0.09273893386125565,
+ "learning_rate": 0.00024599999999999996,
+ "loss": 8.61374282836914,
+ "step": 42
+ },
+ {
+ "epoch": 0.6008733624454149,
+ "grad_norm": 0.09079599380493164,
+ "learning_rate": 0.00025199999999999995,
+ "loss": 8.502904891967773,
+ "step": 43
+ },
+ {
+ "epoch": 0.6148471615720524,
+ "grad_norm": 0.08605825901031494,
+ "learning_rate": 0.000258,
+ "loss": 8.453008651733398,
+ "step": 44
+ },
+ {
+ "epoch": 0.62882096069869,
+ "grad_norm": 0.08338996022939682,
+ "learning_rate": 0.00026399999999999997,
+ "loss": 8.365163803100586,
+ "step": 45
+ },
+ {
+ "epoch": 0.6427947598253275,
+ "grad_norm": 0.08150867372751236,
+ "learning_rate": 0.00027,
+ "loss": 8.285372734069824,
+ "step": 46
+ },
+ {
+ "epoch": 0.6567685589519651,
+ "grad_norm": 0.07794807851314545,
+ "learning_rate": 0.000276,
+ "loss": 8.185811996459961,
+ "step": 47
+ },
+ {
+ "epoch": 0.6707423580786026,
+ "grad_norm": 0.07592720538377762,
+ "learning_rate": 0.00028199999999999997,
+ "loss": 8.098482131958008,
+ "step": 48
+ },
+ {
+ "epoch": 0.6847161572052402,
+ "grad_norm": 0.0709897130727768,
+ "learning_rate": 0.00028799999999999995,
+ "loss": 8.042091369628906,
+ "step": 49
+ },
+ {
+ "epoch": 0.6986899563318777,
+ "grad_norm": 0.06582305580377579,
+ "learning_rate": 0.000294,
+ "loss": 7.972844123840332,
+ "step": 50
+ },
+ {
+ "epoch": 0.7126637554585152,
+ "grad_norm": 0.059452351182699203,
+ "learning_rate": 0.0003,
+ "loss": 7.924579620361328,
+ "step": 51
+ },
+ {
+ "epoch": 0.7266375545851529,
+ "grad_norm": 0.052426718175411224,
+ "learning_rate": 0.00030599999999999996,
+ "loss": 7.876336097717285,
+ "step": 52
+ },
+ {
+ "epoch": 0.7406113537117904,
+ "grad_norm": 0.04768150672316551,
+ "learning_rate": 0.000312,
+ "loss": 7.8102707862854,
+ "step": 53
+ },
+ {
+ "epoch": 0.754585152838428,
+ "grad_norm": 0.03970075398683548,
+ "learning_rate": 0.000318,
+ "loss": 7.791913986206055,
+ "step": 54
+ },
+ {
+ "epoch": 0.7685589519650655,
+ "grad_norm": 0.03559752553701401,
+ "learning_rate": 0.000324,
+ "loss": 7.756960868835449,
+ "step": 55
+ },
+ {
+ "epoch": 0.7825327510917031,
+ "grad_norm": 0.029321353882551193,
+ "learning_rate": 0.00033,
+ "loss": 7.724878787994385,
+ "step": 56
+ },
+ {
+ "epoch": 0.7965065502183406,
+ "grad_norm": 0.023485427722334862,
+ "learning_rate": 0.000336,
+ "loss": 7.72553014755249,
+ "step": 57
+ },
+ {
+ "epoch": 0.8104803493449781,
+ "grad_norm": 0.019319474697113037,
+ "learning_rate": 0.00034199999999999996,
+ "loss": 7.709493160247803,
+ "step": 58
+ },
+ {
+ "epoch": 0.8244541484716157,
+ "grad_norm": 0.016705691814422607,
+ "learning_rate": 0.00034799999999999995,
+ "loss": 7.6873297691345215,
+ "step": 59
+ },
+ {
+ "epoch": 0.8384279475982532,
+ "grad_norm": 0.017667734995484352,
+ "learning_rate": 0.00035399999999999993,
+ "loss": 7.677059173583984,
+ "step": 60
+ },
+ {
+ "epoch": 0.8524017467248908,
+ "grad_norm": 0.01671992614865303,
+ "learning_rate": 0.00035999999999999997,
+ "loss": 7.643512725830078,
+ "step": 61
+ },
+ {
+ "epoch": 0.8663755458515284,
+ "grad_norm": 0.012288172729313374,
+ "learning_rate": 0.00036599999999999995,
+ "loss": 7.658343315124512,
+ "step": 62
+ },
+ {
+ "epoch": 0.880349344978166,
+ "grad_norm": 0.012004776857793331,
+ "learning_rate": 0.000372,
+ "loss": 7.649759292602539,
+ "step": 63
+ },
+ {
+ "epoch": 0.8943231441048035,
+ "grad_norm": 0.012487445957958698,
+ "learning_rate": 0.00037799999999999997,
+ "loss": 7.631503105163574,
+ "step": 64
+ },
+ {
+ "epoch": 0.9082969432314411,
+ "grad_norm": 0.01489241048693657,
+ "learning_rate": 0.00038399999999999996,
+ "loss": 7.647110462188721,
+ "step": 65
+ },
+ {
+ "epoch": 0.9222707423580786,
+ "grad_norm": 0.01428980939090252,
+ "learning_rate": 0.00039,
+ "loss": 7.636983871459961,
+ "step": 66
+ },
+ {
+ "epoch": 0.9362445414847161,
+ "grad_norm": 0.010711144655942917,
+ "learning_rate": 0.000396,
+ "loss": 7.621130466461182,
+ "step": 67
+ },
+ {
+ "epoch": 0.9502183406113537,
+ "grad_norm": 0.010178886353969574,
+ "learning_rate": 0.000402,
+ "loss": 7.593924522399902,
+ "step": 68
+ },
+ {
+ "epoch": 0.9641921397379912,
+ "grad_norm": 0.010604874230921268,
+ "learning_rate": 0.000408,
+ "loss": 7.6015496253967285,
+ "step": 69
+ },
+ {
+ "epoch": 0.9781659388646288,
+ "grad_norm": 0.009751426056027412,
+ "learning_rate": 0.0004139999999999999,
+ "loss": 7.571964263916016,
+ "step": 70
+ },
+ {
+ "epoch": 0.9921397379912664,
+ "grad_norm": 0.009816481731832027,
+ "learning_rate": 0.00041999999999999996,
+ "loss": 7.57344913482666,
+ "step": 71
+ },
+ {
+ "epoch": 1.0,
+ "grad_norm": 0.008665637113153934,
+ "learning_rate": 0.00042599999999999995,
+ "loss": 7.579050064086914,
+ "step": 72
+ },
+ {
+ "epoch": 1.0,
+ "eval_loss": 7.581204891204834,
+ "eval_runtime": 59.0091,
+ "eval_samples_per_second": 41.383,
+ "eval_steps_per_second": 1.305,
+ "step": 72
+ },
+ {
+ "epoch": 1.0139737991266375,
+ "grad_norm": 0.009684201329946518,
+ "learning_rate": 0.00043199999999999993,
+ "loss": 7.568594932556152,
+ "step": 73
+ },
+ {
+ "epoch": 1.027947598253275,
+ "grad_norm": 0.025783710181713104,
+ "learning_rate": 0.00043799999999999997,
+ "loss": 7.578948497772217,
+ "step": 74
+ },
+ {
+ "epoch": 1.0419213973799126,
+ "grad_norm": 0.0327819399535656,
+ "learning_rate": 0.00044399999999999995,
+ "loss": 7.548173904418945,
+ "step": 75
+ },
+ {
+ "epoch": 1.0558951965065502,
+ "grad_norm": 0.021604498848319054,
+ "learning_rate": 0.00045,
+ "loss": 7.552953720092773,
+ "step": 76
+ },
+ {
+ "epoch": 1.0698689956331877,
+ "grad_norm": 0.010230105370283127,
+ "learning_rate": 0.00045599999999999997,
+ "loss": 7.5213212966918945,
+ "step": 77
+ },
+ {
+ "epoch": 1.0838427947598253,
+ "grad_norm": 0.023315589874982834,
+ "learning_rate": 0.00046199999999999995,
+ "loss": 7.501640319824219,
+ "step": 78
+ },
+ {
+ "epoch": 1.0978165938864628,
+ "grad_norm": 0.01109944935888052,
+ "learning_rate": 0.000468,
+ "loss": 7.539519786834717,
+ "step": 79
+ },
+ {
+ "epoch": 1.1117903930131003,
+ "grad_norm": 0.026093758642673492,
+ "learning_rate": 0.000474,
+ "loss": 7.525928974151611,
+ "step": 80
+ },
+ {
+ "epoch": 1.125764192139738,
+ "grad_norm": 0.03274895250797272,
+ "learning_rate": 0.00047999999999999996,
+ "loss": 7.535602569580078,
+ "step": 81
+ },
+ {
+ "epoch": 1.1397379912663754,
+ "grad_norm": 0.01543602254241705,
+ "learning_rate": 0.000486,
+ "loss": 7.513683795928955,
+ "step": 82
+ },
+ {
+ "epoch": 1.1537117903930132,
+ "grad_norm": 0.028377249836921692,
+ "learning_rate": 0.0004919999999999999,
+ "loss": 7.49384880065918,
+ "step": 83
+ },
+ {
+ "epoch": 1.1676855895196507,
+ "grad_norm": 0.04121481627225876,
+ "learning_rate": 0.000498,
+ "loss": 7.488447189331055,
+ "step": 84
+ },
+ {
+ "epoch": 1.1816593886462883,
+ "grad_norm": 0.013282302767038345,
+ "learning_rate": 0.0005039999999999999,
+ "loss": 7.463010787963867,
+ "step": 85
+ },
+ {
+ "epoch": 1.1956331877729258,
+ "grad_norm": 0.03200364485383034,
+ "learning_rate": 0.0005099999999999999,
+ "loss": 7.473515033721924,
+ "step": 86
+ },
+ {
+ "epoch": 1.2096069868995634,
+ "grad_norm": 0.020768029615283012,
+ "learning_rate": 0.000516,
+ "loss": 7.464688777923584,
+ "step": 87
+ },
+ {
+ "epoch": 1.223580786026201,
+ "grad_norm": 0.021715788170695305,
+ "learning_rate": 0.000522,
+ "loss": 7.4835028648376465,
+ "step": 88
+ },
+ {
+ "epoch": 1.2375545851528384,
+ "grad_norm": 0.026255754753947258,
+ "learning_rate": 0.0005279999999999999,
+ "loss": 7.443392753601074,
+ "step": 89
+ },
+ {
+ "epoch": 1.251528384279476,
+ "grad_norm": 0.022325951606035233,
+ "learning_rate": 0.000534,
+ "loss": 7.457373142242432,
+ "step": 90
+ },
+ {
+ "epoch": 1.2655021834061135,
+ "grad_norm": 0.009270387701690197,
+ "learning_rate": 0.00054,
+ "loss": 7.455986976623535,
+ "step": 91
+ },
+ {
+ "epoch": 1.279475982532751,
+ "grad_norm": 0.02369544468820095,
+ "learning_rate": 0.0005459999999999999,
+ "loss": 7.442216873168945,
+ "step": 92
+ },
+ {
+ "epoch": 1.2934497816593886,
+ "grad_norm": 0.008474506437778473,
+ "learning_rate": 0.000552,
+ "loss": 7.4294915199279785,
+ "step": 93
+ },
+ {
+ "epoch": 1.3074235807860262,
+ "grad_norm": 0.017043249681591988,
+ "learning_rate": 0.000558,
+ "loss": 7.406932830810547,
+ "step": 94
+ },
+ {
+ "epoch": 1.3213973799126637,
+ "grad_norm": 0.011813412420451641,
+ "learning_rate": 0.0005639999999999999,
+ "loss": 7.415900230407715,
+ "step": 95
+ },
+ {
+ "epoch": 1.3353711790393012,
+ "grad_norm": 0.01086883433163166,
+ "learning_rate": 0.00057,
+ "loss": 7.452998638153076,
+ "step": 96
+ },
+ {
+ "epoch": 1.3493449781659388,
+ "grad_norm": 0.014700951054692268,
+ "learning_rate": 0.0005759999999999999,
+ "loss": 7.410249710083008,
+ "step": 97
+ },
+ {
+ "epoch": 1.3633187772925766,
+ "grad_norm": 0.014512976631522179,
+ "learning_rate": 0.0005819999999999999,
+ "loss": 7.429132461547852,
+ "step": 98
+ },
+ {
+ "epoch": 1.3772925764192139,
+ "grad_norm": 0.012519920244812965,
+ "learning_rate": 0.000588,
+ "loss": 7.430817604064941,
+ "step": 99
+ },
+ {
+ "epoch": 1.3912663755458516,
+ "grad_norm": 0.007132664322853088,
+ "learning_rate": 0.0005939999999999999,
+ "loss": 7.405234336853027,
+ "step": 100
+ },
+ {
+ "epoch": 1.405240174672489,
+ "grad_norm": 0.019489388912916183,
+ "learning_rate": 0.0006,
+ "loss": 7.427791595458984,
+ "step": 101
+ },
+ {
+ "epoch": 1.4192139737991267,
+ "grad_norm": 0.027252664789557457,
+ "learning_rate": 0.0006,
+ "loss": 7.41761589050293,
+ "step": 102
+ },
+ {
+ "epoch": 1.4331877729257643,
+ "grad_norm": 0.02501637488603592,
+ "learning_rate": 0.0006,
+ "loss": 7.416735649108887,
+ "step": 103
+ },
+ {
+ "epoch": 1.4471615720524018,
+ "grad_norm": 0.00989772193133831,
+ "learning_rate": 0.0006,
+ "loss": 7.3756632804870605,
+ "step": 104
+ },
+ {
+ "epoch": 1.4611353711790394,
+ "grad_norm": 0.01622949168086052,
+ "learning_rate": 0.0006,
+ "loss": 7.400811672210693,
+ "step": 105
+ },
+ {
+ "epoch": 1.475109170305677,
+ "grad_norm": 0.012407763861119747,
+ "learning_rate": 0.0006,
+ "loss": 7.37675666809082,
+ "step": 106
+ },
+ {
+ "epoch": 1.4890829694323144,
+ "grad_norm": 0.010353565216064453,
+ "learning_rate": 0.0006,
+ "loss": 7.4013142585754395,
+ "step": 107
+ },
+ {
+ "epoch": 1.503056768558952,
+ "grad_norm": 0.0120553532615304,
+ "learning_rate": 0.0006,
+ "loss": 7.379873275756836,
+ "step": 108
+ },
+ {
+ "epoch": 1.5170305676855895,
+ "grad_norm": 0.01610225811600685,
+ "learning_rate": 0.0006,
+ "loss": 7.371631145477295,
+ "step": 109
+ },
+ {
+ "epoch": 1.531004366812227,
+ "grad_norm": 0.01159879844635725,
+ "learning_rate": 0.0006,
+ "loss": 7.389007568359375,
+ "step": 110
+ },
+ {
+ "epoch": 1.5449781659388646,
+ "grad_norm": 0.008013816550374031,
+ "learning_rate": 0.0006,
+ "loss": 7.395920753479004,
+ "step": 111
+ },
+ {
+ "epoch": 1.5589519650655022,
+ "grad_norm": 0.016334809362888336,
+ "learning_rate": 0.0006,
+ "loss": 7.393121719360352,
+ "step": 112
+ },
+ {
+ "epoch": 1.5729257641921397,
+ "grad_norm": 0.018821721896529198,
+ "learning_rate": 0.0006,
+ "loss": 7.367543697357178,
+ "step": 113
+ },
+ {
+ "epoch": 1.5868995633187772,
+ "grad_norm": 0.014088994823396206,
+ "learning_rate": 0.0006,
+ "loss": 7.385537624359131,
+ "step": 114
+ },
+ {
+ "epoch": 1.600873362445415,
+ "grad_norm": 0.005864050704985857,
+ "learning_rate": 0.0006,
+ "loss": 7.365346908569336,
+ "step": 115
+ },
+ {
+ "epoch": 1.6148471615720523,
+ "grad_norm": 0.011745430529117584,
+ "learning_rate": 0.0006,
+ "loss": 7.390227317810059,
+ "step": 116
+ },
+ {
+ "epoch": 1.62882096069869,
+ "grad_norm": 0.017205573618412018,
+ "learning_rate": 0.0006,
+ "loss": 7.3558549880981445,
+ "step": 117
+ },
+ {
+ "epoch": 1.6427947598253274,
+ "grad_norm": 0.021256346255540848,
+ "learning_rate": 0.0006,
+ "loss": 7.377346992492676,
+ "step": 118
+ },
+ {
+ "epoch": 1.6567685589519652,
+ "grad_norm": 0.019335204735398293,
+ "learning_rate": 0.0006,
+ "loss": 7.346194267272949,
+ "step": 119
+ },
+ {
+ "epoch": 1.6707423580786025,
+ "grad_norm": 0.008134003728628159,
+ "learning_rate": 0.0006,
+ "loss": 7.3063578605651855,
+ "step": 120
+ },
+ {
+ "epoch": 1.6847161572052403,
+ "grad_norm": 0.017613831907510757,
+ "learning_rate": 0.0006,
+ "loss": 7.338171005249023,
+ "step": 121
+ },
+ {
+ "epoch": 1.6986899563318776,
+ "grad_norm": 0.00890734139829874,
+ "learning_rate": 0.0006,
+ "loss": 7.355384826660156,
+ "step": 122
+ },
+ {
+ "epoch": 1.7126637554585153,
+ "grad_norm": 0.013448155485093594,
+ "learning_rate": 0.0006,
+ "loss": 7.344296455383301,
+ "step": 123
+ },
+ {
+ "epoch": 1.726637554585153,
+ "grad_norm": 0.009241633117198944,
+ "learning_rate": 0.0006,
+ "loss": 7.3454060554504395,
+ "step": 124
+ },
+ {
+ "epoch": 1.7406113537117904,
+ "grad_norm": 0.018561311066150665,
+ "learning_rate": 0.0006,
+ "loss": 7.32348108291626,
+ "step": 125
+ },
+ {
+ "epoch": 1.754585152838428,
+ "grad_norm": 0.025667704641819,
+ "learning_rate": 0.0006,
+ "loss": 7.356142520904541,
+ "step": 126
+ },
+ {
+ "epoch": 1.7685589519650655,
+ "grad_norm": 0.036232151091098785,
+ "learning_rate": 0.0006,
+ "loss": 7.363539695739746,
+ "step": 127
+ },
+ {
+ "epoch": 1.782532751091703,
+ "grad_norm": 0.059483714401721954,
+ "learning_rate": 0.0006,
+ "loss": 7.402331352233887,
+ "step": 128
+ },
+ {
+ "epoch": 1.7965065502183406,
+ "grad_norm": 0.05365009233355522,
+ "learning_rate": 0.0006,
+ "loss": 7.40457820892334,
+ "step": 129
+ },
+ {
+ "epoch": 1.8104803493449781,
+ "grad_norm": 0.02738415263593197,
+ "learning_rate": 0.0006,
+ "loss": 7.3756208419799805,
+ "step": 130
+ },
+ {
+ "epoch": 1.8244541484716157,
+ "grad_norm": 0.035017091780900955,
+ "learning_rate": 0.0006,
+ "loss": 7.379263877868652,
+ "step": 131
+ },
+ {
+ "epoch": 1.8384279475982532,
+ "grad_norm": 0.02819863148033619,
+ "learning_rate": 0.0006,
+ "loss": 7.327673435211182,
+ "step": 132
+ },
+ {
+ "epoch": 1.8524017467248908,
+ "grad_norm": 0.017754238098859787,
+ "learning_rate": 0.0006,
+ "loss": 7.333420276641846,
+ "step": 133
+ },
+ {
+ "epoch": 1.8663755458515285,
+ "grad_norm": 0.024685733020305634,
+ "learning_rate": 0.0006,
+ "loss": 7.34493350982666,
+ "step": 134
+ },
+ {
+ "epoch": 1.8803493449781659,
+ "grad_norm": 0.019426532089710236,
+ "learning_rate": 0.0006,
+ "loss": 7.328448295593262,
+ "step": 135
+ },
+ {
+ "epoch": 1.8943231441048036,
+ "grad_norm": 0.02106391079723835,
+ "learning_rate": 0.0006,
+ "loss": 7.327917098999023,
+ "step": 136
+ },
+ {
+ "epoch": 1.908296943231441,
+ "grad_norm": 0.02056879736483097,
+ "learning_rate": 0.0006,
+ "loss": 7.3311991691589355,
+ "step": 137
+ },
+ {
+ "epoch": 1.9222707423580787,
+ "grad_norm": 0.016865141689777374,
+ "learning_rate": 0.0006,
+ "loss": 7.312766075134277,
+ "step": 138
+ },
+ {
+ "epoch": 1.936244541484716,
+ "grad_norm": 0.017549855634570122,
+ "learning_rate": 0.0006,
+ "loss": 7.31191349029541,
+ "step": 139
+ },
+ {
+ "epoch": 1.9502183406113538,
+ "grad_norm": 0.015930157154798508,
+ "learning_rate": 0.0006,
+ "loss": 7.330901622772217,
+ "step": 140
+ },
+ {
+ "epoch": 1.9641921397379911,
+ "grad_norm": 0.013283558189868927,
+ "learning_rate": 0.0006,
+ "loss": 7.298318386077881,
+ "step": 141
+ },
+ {
+ "epoch": 1.9781659388646289,
+ "grad_norm": 0.012023469433188438,
+ "learning_rate": 0.0006,
+ "loss": 7.282337188720703,
+ "step": 142
+ },
+ {
+ "epoch": 1.9921397379912664,
+ "grad_norm": 0.01459511648863554,
+ "learning_rate": 0.0006,
+ "loss": 7.282924175262451,
+ "step": 143
+ },
+ {
+ "epoch": 2.0,
+ "grad_norm": 0.017474235966801643,
+ "learning_rate": 0.0006,
+ "loss": 7.254333972930908,
+ "step": 144
+ },
+ {
+ "epoch": 2.0,
+ "eval_loss": 7.318708896636963,
+ "eval_runtime": 59.6555,
+ "eval_samples_per_second": 40.935,
+ "eval_steps_per_second": 1.291,
+ "step": 144
+ },
+ {
+ "epoch": 2.0139737991266378,
+ "grad_norm": 0.019757507368922234,
+ "learning_rate": 0.0006,
+ "loss": 7.30288553237915,
+ "step": 145
+ },
+ {
+ "epoch": 2.027947598253275,
+ "grad_norm": 0.014571248553693295,
+ "learning_rate": 0.0006,
+ "loss": 7.274681568145752,
+ "step": 146
+ },
+ {
+ "epoch": 2.041921397379913,
+ "grad_norm": 0.015744198113679886,
+ "learning_rate": 0.0006,
+ "loss": 7.278632640838623,
+ "step": 147
+ },
+ {
+ "epoch": 2.05589519650655,
+ "grad_norm": 0.028048858046531677,
+ "learning_rate": 0.0006,
+ "loss": 7.26558780670166,
+ "step": 148
+ },
+ {
+ "epoch": 2.069868995633188,
+ "grad_norm": 0.06547307223081589,
+ "learning_rate": 0.0006,
+ "loss": 7.32567834854126,
+ "step": 149
+ },
+ {
+ "epoch": 2.0838427947598253,
+ "grad_norm": 0.05969763547182083,
+ "learning_rate": 0.0006,
+ "loss": 7.395836353302002,
+ "step": 150
+ },
+ {
+ "epoch": 2.097816593886463,
+ "grad_norm": 0.032574281096458435,
+ "learning_rate": 0.0006,
+ "loss": 7.318150997161865,
+ "step": 151
+ },
+ {
+ "epoch": 2.1117903930131003,
+ "grad_norm": 0.029452962800860405,
+ "learning_rate": 0.0006,
+ "loss": 7.308066368103027,
+ "step": 152
+ },
+ {
+ "epoch": 2.125764192139738,
+ "grad_norm": 0.03161991387605667,
+ "learning_rate": 0.0006,
+ "loss": 7.306290626525879,
+ "step": 153
+ },
+ {
+ "epoch": 2.1397379912663754,
+ "grad_norm": 0.027096295729279518,
+ "learning_rate": 0.0006,
+ "loss": 7.297796249389648,
+ "step": 154
+ },
+ {
+ "epoch": 2.153711790393013,
+ "grad_norm": 0.028657056391239166,
+ "learning_rate": 0.0006,
+ "loss": 7.317424774169922,
+ "step": 155
+ },
+ {
+ "epoch": 2.1676855895196505,
+ "grad_norm": 0.02046995982527733,
+ "learning_rate": 0.0006,
+ "loss": 7.297986030578613,
+ "step": 156
+ },
+ {
+ "epoch": 2.1816593886462883,
+ "grad_norm": 0.02220369316637516,
+ "learning_rate": 0.0006,
+ "loss": 7.268237590789795,
+ "step": 157
+ },
+ {
+ "epoch": 2.1956331877729256,
+ "grad_norm": 0.018381357192993164,
+ "learning_rate": 0.0006,
+ "loss": 7.259415626525879,
+ "step": 158
+ },
+ {
+ "epoch": 2.2096069868995634,
+ "grad_norm": 0.0204413291066885,
+ "learning_rate": 0.0006,
+ "loss": 7.258173942565918,
+ "step": 159
+ },
+ {
+ "epoch": 2.2235807860262007,
+ "grad_norm": 0.015874288976192474,
+ "learning_rate": 0.0006,
+ "loss": 7.281834602355957,
+ "step": 160
+ },
+ {
+ "epoch": 2.2375545851528384,
+ "grad_norm": 0.01634068600833416,
+ "learning_rate": 0.0006,
+ "loss": 7.2469868659973145,
+ "step": 161
+ },
+ {
+ "epoch": 2.251528384279476,
+ "grad_norm": 0.01728747971355915,
+ "learning_rate": 0.0006,
+ "loss": 7.280995845794678,
+ "step": 162
+ },
+ {
+ "epoch": 2.2655021834061135,
+ "grad_norm": 0.016341188922524452,
+ "learning_rate": 0.0006,
+ "loss": 7.273357391357422,
+ "step": 163
+ },
+ {
+ "epoch": 2.279475982532751,
+ "grad_norm": 0.016125807538628578,
+ "learning_rate": 0.0006,
+ "loss": 7.273510932922363,
+ "step": 164
+ },
+ {
+ "epoch": 2.2934497816593886,
+ "grad_norm": 0.01629287749528885,
+ "learning_rate": 0.0006,
+ "loss": 7.267029762268066,
+ "step": 165
+ },
+ {
+ "epoch": 2.3074235807860264,
+ "grad_norm": 0.015296131372451782,
+ "learning_rate": 0.0006,
+ "loss": 7.211989402770996,
+ "step": 166
+ },
+ {
+ "epoch": 2.3213973799126637,
+ "grad_norm": 0.01709570363163948,
+ "learning_rate": 0.0006,
+ "loss": 7.242851734161377,
+ "step": 167
+ },
+ {
+ "epoch": 2.3353711790393015,
+ "grad_norm": 0.012552388943731785,
+ "learning_rate": 0.0006,
+ "loss": 7.241696834564209,
+ "step": 168
+ },
+ {
+ "epoch": 2.349344978165939,
+ "grad_norm": 0.014571291394531727,
+ "learning_rate": 0.0006,
+ "loss": 7.270681381225586,
+ "step": 169
+ },
+ {
+ "epoch": 2.3633187772925766,
+ "grad_norm": 0.013389437459409237,
+ "learning_rate": 0.0006,
+ "loss": 7.191600799560547,
+ "step": 170
+ },
+ {
+ "epoch": 2.377292576419214,
+ "grad_norm": 0.011895314790308475,
+ "learning_rate": 0.0006,
+ "loss": 7.213610649108887,
+ "step": 171
+ },
+ {
+ "epoch": 2.3912663755458516,
+ "grad_norm": 0.01345642190426588,
+ "learning_rate": 0.0006,
+ "loss": 7.193787574768066,
+ "step": 172
+ },
+ {
+ "epoch": 2.405240174672489,
+ "grad_norm": 0.0091794328764081,
+ "learning_rate": 0.0006,
+ "loss": 7.233364105224609,
+ "step": 173
+ },
+ {
+ "epoch": 2.4192139737991267,
+ "grad_norm": 0.013767086900770664,
+ "learning_rate": 0.0006,
+ "loss": 7.217272758483887,
+ "step": 174
+ },
+ {
+ "epoch": 2.433187772925764,
+ "grad_norm": 0.015134960412979126,
+ "learning_rate": 0.0006,
+ "loss": 7.192791938781738,
+ "step": 175
+ },
+ {
+ "epoch": 2.447161572052402,
+ "grad_norm": 0.01790648326277733,
+ "learning_rate": 0.0006,
+ "loss": 7.229467391967773,
+ "step": 176
+ },
+ {
+ "epoch": 2.461135371179039,
+ "grad_norm": 0.015362377278506756,
+ "learning_rate": 0.0006,
+ "loss": 7.180346488952637,
+ "step": 177
+ },
+ {
+ "epoch": 2.475109170305677,
+ "grad_norm": 0.010505498386919498,
+ "learning_rate": 0.0006,
+ "loss": 7.188833236694336,
+ "step": 178
+ },
+ {
+ "epoch": 2.489082969432314,
+ "grad_norm": 0.009129747748374939,
+ "learning_rate": 0.0006,
+ "loss": 7.203153133392334,
+ "step": 179
+ },
+ {
+ "epoch": 2.503056768558952,
+ "grad_norm": 0.01147883478552103,
+ "learning_rate": 0.0006,
+ "loss": 7.163824558258057,
+ "step": 180
+ },
+ {
+ "epoch": 2.5170305676855893,
+ "grad_norm": 0.011338942684233189,
+ "learning_rate": 0.0006,
+ "loss": 7.16287899017334,
+ "step": 181
+ },
+ {
+ "epoch": 2.531004366812227,
+ "grad_norm": 0.013277675025165081,
+ "learning_rate": 0.0006,
+ "loss": 7.155905723571777,
+ "step": 182
+ },
+ {
+ "epoch": 2.544978165938865,
+ "grad_norm": 0.015269347466528416,
+ "learning_rate": 0.0006,
+ "loss": 7.189371109008789,
+ "step": 183
+ },
+ {
+ "epoch": 2.558951965065502,
+ "grad_norm": 0.017905596643686295,
+ "learning_rate": 0.0006,
+ "loss": 7.177547931671143,
+ "step": 184
+ },
+ {
+ "epoch": 2.5729257641921395,
+ "grad_norm": 0.024431610479950905,
+ "learning_rate": 0.0006,
+ "loss": 7.192169189453125,
+ "step": 185
+ },
+ {
+ "epoch": 2.5868995633187772,
+ "grad_norm": 0.03656961768865585,
+ "learning_rate": 0.0006,
+ "loss": 7.124387741088867,
+ "step": 186
+ },
+ {
+ "epoch": 2.600873362445415,
+ "grad_norm": 0.04337048903107643,
+ "learning_rate": 0.0006,
+ "loss": 7.140231609344482,
+ "step": 187
+ },
+ {
+ "epoch": 2.6148471615720523,
+ "grad_norm": 0.04047228768467903,
+ "learning_rate": 0.0006,
+ "loss": 7.171927452087402,
+ "step": 188
+ },
+ {
+ "epoch": 2.62882096069869,
+ "grad_norm": 0.07825770974159241,
+ "learning_rate": 0.0006,
+ "loss": 7.207631587982178,
+ "step": 189
+ },
+ {
+ "epoch": 2.6427947598253274,
+ "grad_norm": 0.07118063420057297,
+ "learning_rate": 0.0006,
+ "loss": 7.209488391876221,
+ "step": 190
+ },
+ {
+ "epoch": 2.656768558951965,
+ "grad_norm": 0.05153293535113335,
+ "learning_rate": 0.0006,
+ "loss": 7.14793062210083,
+ "step": 191
+ },
+ {
+ "epoch": 2.6707423580786025,
+ "grad_norm": 0.03318839520215988,
+ "learning_rate": 0.0006,
+ "loss": 7.163010597229004,
+ "step": 192
+ },
+ {
+ "epoch": 2.6847161572052403,
+ "grad_norm": 0.029333539307117462,
+ "learning_rate": 0.0006,
+ "loss": 7.159955978393555,
+ "step": 193
+ },
+ {
+ "epoch": 2.6986899563318776,
+ "grad_norm": 0.029405880719423294,
+ "learning_rate": 0.0006,
+ "loss": 7.136178493499756,
+ "step": 194
+ },
+ {
+ "epoch": 2.7126637554585153,
+ "grad_norm": 0.037815988063812256,
+ "learning_rate": 0.0006,
+ "loss": 7.183186054229736,
+ "step": 195
+ },
+ {
+ "epoch": 2.726637554585153,
+ "grad_norm": 0.020307250320911407,
+ "learning_rate": 0.0006,
+ "loss": 7.133164882659912,
+ "step": 196
+ },
+ {
+ "epoch": 2.7406113537117904,
+ "grad_norm": 0.026048069819808006,
+ "learning_rate": 0.0006,
+ "loss": 7.124977111816406,
+ "step": 197
+ },
+ {
+ "epoch": 2.7545851528384278,
+ "grad_norm": 0.027316724881529808,
+ "learning_rate": 0.0006,
+ "loss": 7.07265567779541,
+ "step": 198
+ },
+ {
+ "epoch": 2.7685589519650655,
+ "grad_norm": 0.021629009395837784,
+ "learning_rate": 0.0006,
+ "loss": 7.150020599365234,
+ "step": 199
+ },
+ {
+ "epoch": 2.7825327510917033,
+ "grad_norm": 0.01605929434299469,
+ "learning_rate": 0.0006,
+ "loss": 7.10919189453125,
+ "step": 200
+ },
+ {
+ "epoch": 2.7965065502183406,
+ "grad_norm": 0.020800089463591576,
+ "learning_rate": 0.0006,
+ "loss": 7.132045269012451,
+ "step": 201
+ },
+ {
+ "epoch": 2.810480349344978,
+ "grad_norm": 0.019225774332880974,
+ "learning_rate": 0.0006,
+ "loss": 7.128364562988281,
+ "step": 202
+ },
+ {
+ "epoch": 2.8244541484716157,
+ "grad_norm": 0.01561372634023428,
+ "learning_rate": 0.0006,
+ "loss": 7.083669662475586,
+ "step": 203
+ },
+ {
+ "epoch": 2.8384279475982535,
+ "grad_norm": 0.021969519555568695,
+ "learning_rate": 0.0006,
+ "loss": 7.050841331481934,
+ "step": 204
+ },
+ {
+ "epoch": 2.8524017467248908,
+ "grad_norm": 0.021453695371747017,
+ "learning_rate": 0.0006,
+ "loss": 7.097085475921631,
+ "step": 205
+ },
+ {
+ "epoch": 2.8663755458515285,
+ "grad_norm": 0.022836022078990936,
+ "learning_rate": 0.0006,
+ "loss": 7.10405969619751,
+ "step": 206
+ },
+ {
+ "epoch": 2.880349344978166,
+ "grad_norm": 0.02430851384997368,
+ "learning_rate": 0.0006,
+ "loss": 7.045181751251221,
+ "step": 207
+ },
+ {
+ "epoch": 2.8943231441048036,
+ "grad_norm": 0.020166993141174316,
+ "learning_rate": 0.0006,
+ "loss": 7.068183898925781,
+ "step": 208
+ },
+ {
+ "epoch": 2.908296943231441,
+ "grad_norm": 0.0118505684658885,
+ "learning_rate": 0.0006,
+ "loss": 7.07132625579834,
+ "step": 209
+ },
+ {
+ "epoch": 2.9222707423580787,
+ "grad_norm": 0.023772427812218666,
+ "learning_rate": 0.0006,
+ "loss": 7.055238723754883,
+ "step": 210
+ },
+ {
+ "epoch": 2.936244541484716,
+ "grad_norm": 0.02350712940096855,
+ "learning_rate": 0.0006,
+ "loss": 7.055898666381836,
+ "step": 211
+ },
+ {
+ "epoch": 2.950218340611354,
+ "grad_norm": 0.017244907096028328,
+ "learning_rate": 0.0006,
+ "loss": 7.0099711418151855,
+ "step": 212
+ },
+ {
+ "epoch": 2.964192139737991,
+ "grad_norm": 0.021565575152635574,
+ "learning_rate": 0.0006,
+ "loss": 6.991751670837402,
+ "step": 213
+ },
+ {
+ "epoch": 2.978165938864629,
+ "grad_norm": 0.03241860866546631,
+ "learning_rate": 0.0006,
+ "loss": 7.005980491638184,
+ "step": 214
+ },
+ {
+ "epoch": 2.992139737991266,
+ "grad_norm": 0.04613726586103439,
+ "learning_rate": 0.0006,
+ "loss": 7.027859687805176,
+ "step": 215
+ },
+ {
+ "epoch": 3.0,
+ "grad_norm": 0.06404894590377808,
+ "learning_rate": 0.0006,
+ "loss": 7.123725891113281,
+ "step": 216
+ },
+ {
+ "epoch": 3.0,
+ "eval_loss": 7.070647716522217,
+ "eval_runtime": 58.5801,
+ "eval_samples_per_second": 41.687,
+ "eval_steps_per_second": 1.314,
+ "step": 216
+ },
+ {
+ "epoch": 3.0139737991266378,
+ "grad_norm": 0.05542432889342308,
+ "learning_rate": 0.0006,
+ "loss": 7.078299522399902,
+ "step": 217
+ },
+ {
+ "epoch": 3.027947598253275,
+ "grad_norm": 0.026484526693820953,
+ "learning_rate": 0.0006,
+ "loss": 6.961159706115723,
+ "step": 218
+ },
+ {
+ "epoch": 3.041921397379913,
+ "grad_norm": 0.028276391327381134,
+ "learning_rate": 0.0006,
+ "loss": 7.019961833953857,
+ "step": 219
+ },
+ {
+ "epoch": 3.05589519650655,
+ "grad_norm": 0.024486595764756203,
+ "learning_rate": 0.0006,
+ "loss": 7.039677619934082,
+ "step": 220
+ },
+ {
+ "epoch": 3.069868995633188,
+ "grad_norm": 0.017620893195271492,
+ "learning_rate": 0.0006,
+ "loss": 6.9270782470703125,
+ "step": 221
+ },
+ {
+ "epoch": 3.0838427947598253,
+ "grad_norm": 0.02547302283346653,
+ "learning_rate": 0.0006,
+ "loss": 6.924233436584473,
+ "step": 222
+ },
+ {
+ "epoch": 3.097816593886463,
+ "grad_norm": 0.02539009042084217,
+ "learning_rate": 0.0006,
+ "loss": 6.973979949951172,
+ "step": 223
+ },
+ {
+ "epoch": 3.1117903930131003,
+ "grad_norm": 0.03256227448582649,
+ "learning_rate": 0.0006,
+ "loss": 6.933725357055664,
+ "step": 224
+ },
+ {
+ "epoch": 3.125764192139738,
+ "grad_norm": 0.04651800915598869,
+ "learning_rate": 0.0006,
+ "loss": 6.951044082641602,
+ "step": 225
+ },
+ {
+ "epoch": 3.1397379912663754,
+ "grad_norm": 0.04822975769639015,
+ "learning_rate": 0.0006,
+ "loss": 6.9614410400390625,
+ "step": 226
+ },
+ {
+ "epoch": 3.153711790393013,
+ "grad_norm": 0.031165020540356636,
+ "learning_rate": 0.0006,
+ "loss": 6.91193151473999,
+ "step": 227
+ },
+ {
+ "epoch": 3.1676855895196505,
+ "grad_norm": 0.029754292219877243,
+ "learning_rate": 0.0006,
+ "loss": 6.91802978515625,
+ "step": 228
+ },
+ {
+ "epoch": 3.1816593886462883,
+ "grad_norm": 0.02193152718245983,
+ "learning_rate": 0.0006,
+ "loss": 6.928090572357178,
+ "step": 229
+ },
+ {
+ "epoch": 3.1956331877729256,
+ "grad_norm": 0.02428470179438591,
+ "learning_rate": 0.0006,
+ "loss": 6.90396785736084,
+ "step": 230
+ },
+ {
+ "epoch": 3.2096069868995634,
+ "grad_norm": 0.023374345153570175,
+ "learning_rate": 0.0006,
+ "loss": 6.885858535766602,
+ "step": 231
+ },
+ {
+ "epoch": 3.2235807860262007,
+ "grad_norm": 0.023405000567436218,
+ "learning_rate": 0.0006,
+ "loss": 6.8855767250061035,
+ "step": 232
+ },
+ {
+ "epoch": 3.2375545851528384,
+ "grad_norm": 0.017277134582400322,
+ "learning_rate": 0.0006,
+ "loss": 6.901449203491211,
+ "step": 233
+ },
+ {
+ "epoch": 3.251528384279476,
+ "grad_norm": 0.019496750086545944,
+ "learning_rate": 0.0006,
+ "loss": 6.885687828063965,
+ "step": 234
+ },
+ {
+ "epoch": 3.2655021834061135,
+ "grad_norm": 0.016893276944756508,
+ "learning_rate": 0.0006,
+ "loss": 6.872369766235352,
+ "step": 235
+ },
+ {
+ "epoch": 3.279475982532751,
+ "grad_norm": 0.02058715932071209,
+ "learning_rate": 0.0006,
+ "loss": 6.823280334472656,
+ "step": 236
+ },
+ {
+ "epoch": 3.2934497816593886,
+ "grad_norm": 0.015530755743384361,
+ "learning_rate": 0.0006,
+ "loss": 6.860499858856201,
+ "step": 237
+ },
+ {
+ "epoch": 3.3074235807860264,
+ "grad_norm": 0.01934926211833954,
+ "learning_rate": 0.0006,
+ "loss": 6.817448616027832,
+ "step": 238
+ },
+ {
+ "epoch": 3.3213973799126637,
+ "grad_norm": 0.017413552850484848,
+ "learning_rate": 0.0006,
+ "loss": 6.881955146789551,
+ "step": 239
+ },
+ {
+ "epoch": 3.3353711790393015,
+ "grad_norm": 0.026478875428438187,
+ "learning_rate": 0.0006,
+ "loss": 6.827404499053955,
+ "step": 240
+ },
+ {
+ "epoch": 3.349344978165939,
+ "grad_norm": 0.032611701637506485,
+ "learning_rate": 0.0006,
+ "loss": 6.78285551071167,
+ "step": 241
+ },
+ {
+ "epoch": 3.3633187772925766,
+ "grad_norm": 0.041900306940078735,
+ "learning_rate": 0.0006,
+ "loss": 6.783053874969482,
+ "step": 242
+ },
+ {
+ "epoch": 3.377292576419214,
+ "grad_norm": 0.04149497672915459,
+ "learning_rate": 0.0006,
+ "loss": 6.745220184326172,
+ "step": 243
+ },
+ {
+ "epoch": 3.3912663755458516,
+ "grad_norm": 0.023949826136231422,
+ "learning_rate": 0.0006,
+ "loss": 6.841533184051514,
+ "step": 244
+ },
+ {
+ "epoch": 3.405240174672489,
+ "grad_norm": 0.028601842001080513,
+ "learning_rate": 0.0006,
+ "loss": 6.852852821350098,
+ "step": 245
+ },
+ {
+ "epoch": 3.4192139737991267,
+ "grad_norm": 0.027365155518054962,
+ "learning_rate": 0.0006,
+ "loss": 6.74376106262207,
+ "step": 246
+ },
+ {
+ "epoch": 3.433187772925764,
+ "grad_norm": 0.02856568619608879,
+ "learning_rate": 0.0006,
+ "loss": 6.74906063079834,
+ "step": 247
+ },
+ {
+ "epoch": 3.447161572052402,
+ "grad_norm": 0.019172416999936104,
+ "learning_rate": 0.0006,
+ "loss": 6.809206008911133,
+ "step": 248
+ },
+ {
+ "epoch": 3.461135371179039,
+ "grad_norm": 0.02161746844649315,
+ "learning_rate": 0.0006,
+ "loss": 6.789888381958008,
+ "step": 249
+ },
+ {
+ "epoch": 3.475109170305677,
+ "grad_norm": 0.020273666828870773,
+ "learning_rate": 0.0006,
+ "loss": 6.73512077331543,
+ "step": 250
+ },
+ {
+ "epoch": 3.489082969432314,
+ "grad_norm": 0.022329283878207207,
+ "learning_rate": 0.0006,
+ "loss": 6.826148509979248,
+ "step": 251
+ },
+ {
+ "epoch": 3.503056768558952,
+ "grad_norm": 0.02284284122288227,
+ "learning_rate": 0.0006,
+ "loss": 6.748521327972412,
+ "step": 252
+ },
+ {
+ "epoch": 3.5170305676855893,
+ "grad_norm": 0.026368247345089912,
+ "learning_rate": 0.0006,
+ "loss": 6.645879745483398,
+ "step": 253
+ },
+ {
+ "epoch": 3.531004366812227,
+ "grad_norm": 0.03627244383096695,
+ "learning_rate": 0.0006,
+ "loss": 6.666064739227295,
+ "step": 254
+ },
+ {
+ "epoch": 3.544978165938865,
+ "grad_norm": 0.04140935093164444,
+ "learning_rate": 0.0006,
+ "loss": 6.712275981903076,
+ "step": 255
+ },
+ {
+ "epoch": 3.558951965065502,
+ "grad_norm": 0.0390462800860405,
+ "learning_rate": 0.0006,
+ "loss": 6.665900230407715,
+ "step": 256
+ },
+ {
+ "epoch": 3.5729257641921395,
+ "grad_norm": 0.05178583785891533,
+ "learning_rate": 0.0006,
+ "loss": 6.753298759460449,
+ "step": 257
+ },
+ {
+ "epoch": 3.5868995633187772,
+ "grad_norm": 0.054328061640262604,
+ "learning_rate": 0.0006,
+ "loss": 6.732633113861084,
+ "step": 258
+ },
+ {
+ "epoch": 3.600873362445415,
+ "grad_norm": 0.04590460658073425,
+ "learning_rate": 0.0006,
+ "loss": 6.65250301361084,
+ "step": 259
+ },
+ {
+ "epoch": 3.6148471615720523,
+ "grad_norm": 0.044331666082143784,
+ "learning_rate": 0.0006,
+ "loss": 6.68654203414917,
+ "step": 260
+ },
+ {
+ "epoch": 3.62882096069869,
+ "grad_norm": 0.05308730527758598,
+ "learning_rate": 0.0006,
+ "loss": 6.642482757568359,
+ "step": 261
+ },
+ {
+ "epoch": 3.6427947598253274,
+ "grad_norm": 0.03999016433954239,
+ "learning_rate": 0.0006,
+ "loss": 6.625227451324463,
+ "step": 262
+ },
+ {
+ "epoch": 3.656768558951965,
+ "grad_norm": 0.04088086634874344,
+ "learning_rate": 0.0006,
+ "loss": 6.67623233795166,
+ "step": 263
+ },
+ {
+ "epoch": 3.6707423580786025,
+ "grad_norm": 0.03615636005997658,
+ "learning_rate": 0.0006,
+ "loss": 6.6869001388549805,
+ "step": 264
+ },
+ {
+ "epoch": 3.6847161572052403,
+ "grad_norm": 0.02817360684275627,
+ "learning_rate": 0.0006,
+ "loss": 6.674601078033447,
+ "step": 265
+ },
+ {
+ "epoch": 3.6986899563318776,
+ "grad_norm": 0.032586123794317245,
+ "learning_rate": 0.0006,
+ "loss": 6.612994194030762,
+ "step": 266
+ },
+ {
+ "epoch": 3.7126637554585153,
+ "grad_norm": 0.031117867678403854,
+ "learning_rate": 0.0006,
+ "loss": 6.6244354248046875,
+ "step": 267
+ },
+ {
+ "epoch": 3.726637554585153,
+ "grad_norm": 0.028330687433481216,
+ "learning_rate": 0.0006,
+ "loss": 6.669557094573975,
+ "step": 268
+ },
+ {
+ "epoch": 3.7406113537117904,
+ "grad_norm": 0.021489202976226807,
+ "learning_rate": 0.0006,
+ "loss": 6.632035732269287,
+ "step": 269
+ },
+ {
+ "epoch": 3.7545851528384278,
+ "grad_norm": 0.024269424378871918,
+ "learning_rate": 0.0006,
+ "loss": 6.660867691040039,
+ "step": 270
+ },
+ {
+ "epoch": 3.7685589519650655,
+ "grad_norm": 0.02237873338162899,
+ "learning_rate": 0.0006,
+ "loss": 6.602587699890137,
+ "step": 271
+ },
+ {
+ "epoch": 3.7825327510917033,
+ "grad_norm": 0.026503929868340492,
+ "learning_rate": 0.0006,
+ "loss": 6.599200248718262,
+ "step": 272
+ },
+ {
+ "epoch": 3.7965065502183406,
+ "grad_norm": 0.03687124326825142,
+ "learning_rate": 0.0006,
+ "loss": 6.567296504974365,
+ "step": 273
+ },
+ {
+ "epoch": 3.810480349344978,
+ "grad_norm": 0.04708952456712723,
+ "learning_rate": 0.0006,
+ "loss": 6.621265411376953,
+ "step": 274
+ },
+ {
+ "epoch": 3.8244541484716157,
+ "grad_norm": 0.041846614331007004,
+ "learning_rate": 0.0006,
+ "loss": 6.550536155700684,
+ "step": 275
+ },
+ {
+ "epoch": 3.8384279475982535,
+ "grad_norm": 0.031249675899744034,
+ "learning_rate": 0.0006,
+ "loss": 6.54278564453125,
+ "step": 276
+ },
+ {
+ "epoch": 3.8524017467248908,
+ "grad_norm": 0.023082256317138672,
+ "learning_rate": 0.0006,
+ "loss": 6.544787406921387,
+ "step": 277
+ },
+ {
+ "epoch": 3.8663755458515285,
+ "grad_norm": 0.020995570346713066,
+ "learning_rate": 0.0006,
+ "loss": 6.518143653869629,
+ "step": 278
+ },
+ {
+ "epoch": 3.880349344978166,
+ "grad_norm": 0.020839324221014977,
+ "learning_rate": 0.0006,
+ "loss": 6.546667098999023,
+ "step": 279
+ },
+ {
+ "epoch": 3.8943231441048036,
+ "grad_norm": 0.018377432599663734,
+ "learning_rate": 0.0006,
+ "loss": 6.478307723999023,
+ "step": 280
+ },
+ {
+ "epoch": 3.908296943231441,
+ "grad_norm": 0.021713724359869957,
+ "learning_rate": 0.0006,
+ "loss": 6.485574245452881,
+ "step": 281
+ },
+ {
+ "epoch": 3.9222707423580787,
+ "grad_norm": 0.01359301246702671,
+ "learning_rate": 0.0006,
+ "loss": 6.570268630981445,
+ "step": 282
+ },
+ {
+ "epoch": 3.936244541484716,
+ "grad_norm": 0.016233332455158234,
+ "learning_rate": 0.0006,
+ "loss": 6.57150936126709,
+ "step": 283
+ },
+ {
+ "epoch": 3.950218340611354,
+ "grad_norm": 0.020687608048319817,
+ "learning_rate": 0.0006,
+ "loss": 6.527956008911133,
+ "step": 284
+ },
+ {
+ "epoch": 3.964192139737991,
+ "grad_norm": 0.02032964862883091,
+ "learning_rate": 0.0006,
+ "loss": 6.468169689178467,
+ "step": 285
+ },
+ {
+ "epoch": 3.978165938864629,
+ "grad_norm": 0.016676288098096848,
+ "learning_rate": 0.0006,
+ "loss": 6.448418617248535,
+ "step": 286
+ },
+ {
+ "epoch": 3.992139737991266,
+ "grad_norm": 0.0168539360165596,
+ "learning_rate": 0.0006,
+ "loss": 6.461180686950684,
+ "step": 287
+ },
+ {
+ "epoch": 4.0,
+ "grad_norm": 0.02666233666241169,
+ "learning_rate": 0.0006,
+ "loss": 6.344893932342529,
+ "step": 288
+ },
+ {
+ "epoch": 4.0,
+ "eval_loss": 6.5384979248046875,
+ "eval_runtime": 58.814,
+ "eval_samples_per_second": 41.521,
+ "eval_steps_per_second": 1.309,
+ "step": 288
+ },
+ {
+ "epoch": 4.013973799126638,
+ "grad_norm": 0.04154708981513977,
+ "learning_rate": 0.0006,
+ "loss": 6.41855525970459,
+ "step": 289
+ },
+ {
+ "epoch": 4.0279475982532755,
+ "grad_norm": 0.04895668104290962,
+ "learning_rate": 0.0006,
+ "loss": 6.499302864074707,
+ "step": 290
+ },
+ {
+ "epoch": 4.041921397379912,
+ "grad_norm": 0.034464482218027115,
+ "learning_rate": 0.0006,
+ "loss": 6.438611030578613,
+ "step": 291
+ },
+ {
+ "epoch": 4.05589519650655,
+ "grad_norm": 0.06845773756504059,
+ "learning_rate": 0.0006,
+ "loss": 6.538361072540283,
+ "step": 292
+ },
+ {
+ "epoch": 4.069868995633188,
+ "grad_norm": 0.053768858313560486,
+ "learning_rate": 0.0006,
+ "loss": 6.527524948120117,
+ "step": 293
+ },
+ {
+ "epoch": 4.083842794759826,
+ "grad_norm": 0.039037931710481644,
+ "learning_rate": 0.0006,
+ "loss": 6.503847122192383,
+ "step": 294
+ },
+ {
+ "epoch": 4.097816593886463,
+ "grad_norm": 0.03962196782231331,
+ "learning_rate": 0.0006,
+ "loss": 6.475178241729736,
+ "step": 295
+ },
+ {
+ "epoch": 4.1117903930131,
+ "grad_norm": 0.049101535230875015,
+ "learning_rate": 0.0006,
+ "loss": 6.474587917327881,
+ "step": 296
+ },
+ {
+ "epoch": 4.125764192139738,
+ "grad_norm": 0.02830282226204872,
+ "learning_rate": 0.0006,
+ "loss": 6.45065975189209,
+ "step": 297
+ },
+ {
+ "epoch": 4.139737991266376,
+ "grad_norm": 0.023153482005000114,
+ "learning_rate": 0.0006,
+ "loss": 6.5218706130981445,
+ "step": 298
+ },
+ {
+ "epoch": 4.153711790393013,
+ "grad_norm": 0.025064002722501755,
+ "learning_rate": 0.0006,
+ "loss": 6.456475257873535,
+ "step": 299
+ },
+ {
+ "epoch": 4.1676855895196505,
+ "grad_norm": 0.02357092685997486,
+ "learning_rate": 0.0006,
+ "loss": 6.485952377319336,
+ "step": 300
+ },
+ {
+ "epoch": 4.181659388646288,
+ "grad_norm": 0.021832305938005447,
+ "learning_rate": 0.0006,
+ "loss": 6.425506591796875,
+ "step": 301
+ },
+ {
+ "epoch": 4.195633187772926,
+ "grad_norm": 0.0227900929749012,
+ "learning_rate": 0.0006,
+ "loss": 6.439155578613281,
+ "step": 302
+ },
+ {
+ "epoch": 4.209606986899563,
+ "grad_norm": 0.015961607918143272,
+ "learning_rate": 0.0006,
+ "loss": 6.460862636566162,
+ "step": 303
+ },
+ {
+ "epoch": 4.223580786026201,
+ "grad_norm": 0.017876390367746353,
+ "learning_rate": 0.0006,
+ "loss": 6.44564151763916,
+ "step": 304
+ },
+ {
+ "epoch": 4.2375545851528384,
+ "grad_norm": 0.013703204691410065,
+ "learning_rate": 0.0006,
+ "loss": 6.418414115905762,
+ "step": 305
+ },
+ {
+ "epoch": 4.251528384279476,
+ "grad_norm": 0.015240306034684181,
+ "learning_rate": 0.0006,
+ "loss": 6.411238670349121,
+ "step": 306
+ },
+ {
+ "epoch": 4.265502183406113,
+ "grad_norm": 0.01587662845849991,
+ "learning_rate": 0.0006,
+ "loss": 6.395642280578613,
+ "step": 307
+ },
+ {
+ "epoch": 4.279475982532751,
+ "grad_norm": 0.014650700613856316,
+ "learning_rate": 0.0006,
+ "loss": 6.302793025970459,
+ "step": 308
+ },
+ {
+ "epoch": 4.293449781659389,
+ "grad_norm": 0.016533225774765015,
+ "learning_rate": 0.0006,
+ "loss": 6.380588054656982,
+ "step": 309
+ },
+ {
+ "epoch": 4.307423580786026,
+ "grad_norm": 0.019963741302490234,
+ "learning_rate": 0.0006,
+ "loss": 6.345336437225342,
+ "step": 310
+ },
+ {
+ "epoch": 4.321397379912664,
+ "grad_norm": 0.02237936295568943,
+ "learning_rate": 0.0006,
+ "loss": 6.387770652770996,
+ "step": 311
+ },
+ {
+ "epoch": 4.335371179039301,
+ "grad_norm": 0.026658328250050545,
+ "learning_rate": 0.0006,
+ "loss": 6.323662757873535,
+ "step": 312
+ },
+ {
+ "epoch": 4.349344978165939,
+ "grad_norm": 0.02852284163236618,
+ "learning_rate": 0.0006,
+ "loss": 6.382204055786133,
+ "step": 313
+ },
+ {
+ "epoch": 4.3633187772925766,
+ "grad_norm": 0.028815090656280518,
+ "learning_rate": 0.0006,
+ "loss": 6.402390480041504,
+ "step": 314
+ },
+ {
+ "epoch": 4.377292576419214,
+ "grad_norm": 0.029393529519438744,
+ "learning_rate": 0.0006,
+ "loss": 6.280278205871582,
+ "step": 315
+ },
+ {
+ "epoch": 4.391266375545851,
+ "grad_norm": 0.02186041697859764,
+ "learning_rate": 0.0006,
+ "loss": 6.302707672119141,
+ "step": 316
+ },
+ {
+ "epoch": 4.405240174672489,
+ "grad_norm": 0.02071218006312847,
+ "learning_rate": 0.0006,
+ "loss": 6.3365020751953125,
+ "step": 317
+ },
+ {
+ "epoch": 4.419213973799127,
+ "grad_norm": 0.02779117226600647,
+ "learning_rate": 0.0006,
+ "loss": 6.252157211303711,
+ "step": 318
+ },
+ {
+ "epoch": 4.4331877729257645,
+ "grad_norm": 0.03470654785633087,
+ "learning_rate": 0.0006,
+ "loss": 6.319070339202881,
+ "step": 319
+ },
+ {
+ "epoch": 4.447161572052401,
+ "grad_norm": 0.037067804485559464,
+ "learning_rate": 0.0006,
+ "loss": 6.314116477966309,
+ "step": 320
+ },
+ {
+ "epoch": 4.461135371179039,
+ "grad_norm": 0.032182756811380386,
+ "learning_rate": 0.0006,
+ "loss": 6.267298698425293,
+ "step": 321
+ },
+ {
+ "epoch": 4.475109170305677,
+ "grad_norm": 0.026305217295885086,
+ "learning_rate": 0.0006,
+ "loss": 6.331688404083252,
+ "step": 322
+ },
+ {
+ "epoch": 4.489082969432315,
+ "grad_norm": 0.027228357270359993,
+ "learning_rate": 0.0006,
+ "loss": 6.31139612197876,
+ "step": 323
+ },
+ {
+ "epoch": 4.503056768558952,
+ "grad_norm": 0.025781169533729553,
+ "learning_rate": 0.0006,
+ "loss": 6.194684982299805,
+ "step": 324
+ },
+ {
+ "epoch": 4.517030567685589,
+ "grad_norm": 0.029186977073550224,
+ "learning_rate": 0.0006,
+ "loss": 6.31306266784668,
+ "step": 325
+ },
+ {
+ "epoch": 4.531004366812227,
+ "grad_norm": 0.023608777672052383,
+ "learning_rate": 0.0006,
+ "loss": 6.285243034362793,
+ "step": 326
+ },
+ {
+ "epoch": 4.544978165938865,
+ "grad_norm": 0.02267594076693058,
+ "learning_rate": 0.0006,
+ "loss": 6.30342960357666,
+ "step": 327
+ },
+ {
+ "epoch": 4.558951965065502,
+ "grad_norm": 0.01926310732960701,
+ "learning_rate": 0.0006,
+ "loss": 6.280606269836426,
+ "step": 328
+ },
+ {
+ "epoch": 4.5729257641921395,
+ "grad_norm": 0.026484837755560875,
+ "learning_rate": 0.0006,
+ "loss": 6.237980365753174,
+ "step": 329
+ },
+ {
+ "epoch": 4.586899563318777,
+ "grad_norm": 0.026228854432702065,
+ "learning_rate": 0.0006,
+ "loss": 6.3018293380737305,
+ "step": 330
+ },
+ {
+ "epoch": 4.600873362445415,
+ "grad_norm": 0.022096967324614525,
+ "learning_rate": 0.0006,
+ "loss": 6.253863334655762,
+ "step": 331
+ },
+ {
+ "epoch": 4.614847161572053,
+ "grad_norm": 0.027223890647292137,
+ "learning_rate": 0.0006,
+ "loss": 6.166882038116455,
+ "step": 332
+ },
+ {
+ "epoch": 4.62882096069869,
+ "grad_norm": 0.03160124272108078,
+ "learning_rate": 0.0006,
+ "loss": 6.230309963226318,
+ "step": 333
+ },
+ {
+ "epoch": 4.642794759825327,
+ "grad_norm": 0.04110539332032204,
+ "learning_rate": 0.0006,
+ "loss": 6.2334818840026855,
+ "step": 334
+ },
+ {
+ "epoch": 4.656768558951965,
+ "grad_norm": 0.04355933889746666,
+ "learning_rate": 0.0006,
+ "loss": 6.160956382751465,
+ "step": 335
+ },
+ {
+ "epoch": 4.670742358078603,
+ "grad_norm": 0.03611086308956146,
+ "learning_rate": 0.0006,
+ "loss": 6.23399543762207,
+ "step": 336
+ },
+ {
+ "epoch": 4.68471615720524,
+ "grad_norm": 0.051811400800943375,
+ "learning_rate": 0.0006,
+ "loss": 6.241855144500732,
+ "step": 337
+ },
+ {
+ "epoch": 4.698689956331878,
+ "grad_norm": 0.057434193789958954,
+ "learning_rate": 0.0006,
+ "loss": 6.220993995666504,
+ "step": 338
+ },
+ {
+ "epoch": 4.712663755458515,
+ "grad_norm": 0.04420953616499901,
+ "learning_rate": 0.0006,
+ "loss": 6.300461292266846,
+ "step": 339
+ },
+ {
+ "epoch": 4.726637554585153,
+ "grad_norm": 0.03515457361936569,
+ "learning_rate": 0.0006,
+ "loss": 6.269349098205566,
+ "step": 340
+ },
+ {
+ "epoch": 4.74061135371179,
+ "grad_norm": 0.03222600743174553,
+ "learning_rate": 0.0006,
+ "loss": 6.248613357543945,
+ "step": 341
+ },
+ {
+ "epoch": 4.754585152838428,
+ "grad_norm": 0.0384046845138073,
+ "learning_rate": 0.0006,
+ "loss": 6.266979217529297,
+ "step": 342
+ },
+ {
+ "epoch": 4.7685589519650655,
+ "grad_norm": 0.0514867827296257,
+ "learning_rate": 0.0006,
+ "loss": 6.234103202819824,
+ "step": 343
+ },
+ {
+ "epoch": 4.782532751091703,
+ "grad_norm": 0.050747793167829514,
+ "learning_rate": 0.0006,
+ "loss": 6.322582244873047,
+ "step": 344
+ },
+ {
+ "epoch": 4.796506550218341,
+ "grad_norm": 0.040878988802433014,
+ "learning_rate": 0.0006,
+ "loss": 6.212509632110596,
+ "step": 345
+ },
+ {
+ "epoch": 4.810480349344978,
+ "grad_norm": 0.034655820578336716,
+ "learning_rate": 0.0006,
+ "loss": 6.265879154205322,
+ "step": 346
+ },
+ {
+ "epoch": 4.824454148471616,
+ "grad_norm": 0.04202224686741829,
+ "learning_rate": 0.0006,
+ "loss": 6.2896409034729,
+ "step": 347
+ },
+ {
+ "epoch": 4.8384279475982535,
+ "grad_norm": 0.03632277995347977,
+ "learning_rate": 0.0006,
+ "loss": 6.253917217254639,
+ "step": 348
+ },
+ {
+ "epoch": 4.85240174672489,
+ "grad_norm": 0.028645765036344528,
+ "learning_rate": 0.0006,
+ "loss": 6.207965850830078,
+ "step": 349
+ },
+ {
+ "epoch": 4.866375545851528,
+ "grad_norm": 0.025305170565843582,
+ "learning_rate": 0.0006,
+ "loss": 6.168068885803223,
+ "step": 350
+ },
+ {
+ "epoch": 4.880349344978166,
+ "grad_norm": 0.022137803956866264,
+ "learning_rate": 0.0006,
+ "loss": 6.235283374786377,
+ "step": 351
+ },
+ {
+ "epoch": 4.894323144104804,
+ "grad_norm": 0.01814538985490799,
+ "learning_rate": 0.0006,
+ "loss": 6.259641647338867,
+ "step": 352
+ },
+ {
+ "epoch": 4.908296943231441,
+ "grad_norm": 0.018465086817741394,
+ "learning_rate": 0.0006,
+ "loss": 6.111738681793213,
+ "step": 353
+ },
+ {
+ "epoch": 4.922270742358078,
+ "grad_norm": 0.018638933077454567,
+ "learning_rate": 0.0006,
+ "loss": 6.225642204284668,
+ "step": 354
+ },
+ {
+ "epoch": 4.936244541484716,
+ "grad_norm": 0.016461260616779327,
+ "learning_rate": 0.0006,
+ "loss": 6.140647888183594,
+ "step": 355
+ },
+ {
+ "epoch": 4.950218340611354,
+ "grad_norm": 0.015458385460078716,
+ "learning_rate": 0.0006,
+ "loss": 6.155224800109863,
+ "step": 356
+ },
+ {
+ "epoch": 4.964192139737992,
+ "grad_norm": 0.014934048056602478,
+ "learning_rate": 0.0006,
+ "loss": 6.101555824279785,
+ "step": 357
+ },
+ {
+ "epoch": 4.978165938864628,
+ "grad_norm": 0.013979545794427395,
+ "learning_rate": 0.0006,
+ "loss": 6.161718845367432,
+ "step": 358
+ },
+ {
+ "epoch": 4.992139737991266,
+ "grad_norm": 0.01640394888818264,
+ "learning_rate": 0.0006,
+ "loss": 6.055768966674805,
+ "step": 359
+ },
+ {
+ "epoch": 5.0,
+ "grad_norm": 0.015135680325329304,
+ "learning_rate": 0.0006,
+ "loss": 6.169695854187012,
+ "step": 360
+ },
+ {
+ "epoch": 5.0,
+ "eval_loss": 6.16408109664917,
+ "eval_runtime": 59.3649,
+ "eval_samples_per_second": 41.135,
+ "eval_steps_per_second": 1.297,
+ "step": 360
+ },
+ {
+ "epoch": 5.013973799126638,
+ "grad_norm": 0.012849048711359501,
+ "learning_rate": 0.0006,
+ "loss": 6.055830001831055,
+ "step": 361
+ },
+ {
+ "epoch": 5.0279475982532755,
+ "grad_norm": 0.018169576302170753,
+ "learning_rate": 0.0006,
+ "loss": 6.082655429840088,
+ "step": 362
+ },
+ {
+ "epoch": 5.041921397379912,
+ "grad_norm": 0.024293430149555206,
+ "learning_rate": 0.0006,
+ "loss": 6.109032154083252,
+ "step": 363
+ },
+ {
+ "epoch": 5.05589519650655,
+ "grad_norm": 0.03266787901520729,
+ "learning_rate": 0.0006,
+ "loss": 6.030613899230957,
+ "step": 364
+ },
+ {
+ "epoch": 5.069868995633188,
+ "grad_norm": 0.03572266176342964,
+ "learning_rate": 0.0006,
+ "loss": 6.050390720367432,
+ "step": 365
+ },
+ {
+ "epoch": 5.083842794759826,
+ "grad_norm": 0.032625213265419006,
+ "learning_rate": 0.0006,
+ "loss": 6.169313430786133,
+ "step": 366
+ },
+ {
+ "epoch": 5.097816593886463,
+ "grad_norm": 0.04051872715353966,
+ "learning_rate": 0.0006,
+ "loss": 6.10398006439209,
+ "step": 367
+ },
+ {
+ "epoch": 5.1117903930131,
+ "grad_norm": 0.0351213738322258,
+ "learning_rate": 0.0006,
+ "loss": 6.121973514556885,
+ "step": 368
+ },
+ {
+ "epoch": 5.125764192139738,
+ "grad_norm": 0.03523759916424751,
+ "learning_rate": 0.0006,
+ "loss": 6.0626654624938965,
+ "step": 369
+ },
+ {
+ "epoch": 5.139737991266376,
+ "grad_norm": 0.02706182189285755,
+ "learning_rate": 0.0006,
+ "loss": 6.049354076385498,
+ "step": 370
+ },
+ {
+ "epoch": 5.153711790393013,
+ "grad_norm": 0.024212589487433434,
+ "learning_rate": 0.0006,
+ "loss": 6.034826278686523,
+ "step": 371
+ },
+ {
+ "epoch": 5.1676855895196505,
+ "grad_norm": 0.024798082187771797,
+ "learning_rate": 0.0006,
+ "loss": 6.004058837890625,
+ "step": 372
+ },
+ {
+ "epoch": 5.181659388646288,
+ "grad_norm": 0.02325567416846752,
+ "learning_rate": 0.0006,
+ "loss": 5.986461639404297,
+ "step": 373
+ },
+ {
+ "epoch": 5.195633187772926,
+ "grad_norm": 0.019060570746660233,
+ "learning_rate": 0.0006,
+ "loss": 6.059736251831055,
+ "step": 374
+ },
+ {
+ "epoch": 5.209606986899563,
+ "grad_norm": 0.016822976991534233,
+ "learning_rate": 0.0006,
+ "loss": 6.072957515716553,
+ "step": 375
+ },
+ {
+ "epoch": 5.223580786026201,
+ "grad_norm": 0.017218658700585365,
+ "learning_rate": 0.0006,
+ "loss": 6.031739234924316,
+ "step": 376
+ },
+ {
+ "epoch": 5.2375545851528384,
+ "grad_norm": 0.014889383688569069,
+ "learning_rate": 0.0006,
+ "loss": 6.025674343109131,
+ "step": 377
+ },
+ {
+ "epoch": 5.251528384279476,
+ "grad_norm": 0.01708107627928257,
+ "learning_rate": 0.0006,
+ "loss": 6.090451240539551,
+ "step": 378
+ },
+ {
+ "epoch": 5.265502183406113,
+ "grad_norm": 0.020260317251086235,
+ "learning_rate": 0.0006,
+ "loss": 6.094257354736328,
+ "step": 379
+ },
+ {
+ "epoch": 5.279475982532751,
+ "grad_norm": 0.020110400393605232,
+ "learning_rate": 0.0006,
+ "loss": 6.021188735961914,
+ "step": 380
+ },
+ {
+ "epoch": 5.293449781659389,
+ "grad_norm": 0.02027830481529236,
+ "learning_rate": 0.0006,
+ "loss": 6.022156238555908,
+ "step": 381
+ },
+ {
+ "epoch": 5.307423580786026,
+ "grad_norm": 0.022747062146663666,
+ "learning_rate": 0.0006,
+ "loss": 6.011836051940918,
+ "step": 382
+ },
+ {
+ "epoch": 5.321397379912664,
+ "grad_norm": 0.01990230567753315,
+ "learning_rate": 0.0006,
+ "loss": 6.003054618835449,
+ "step": 383
+ },
+ {
+ "epoch": 5.335371179039301,
+ "grad_norm": 0.01596238650381565,
+ "learning_rate": 0.0006,
+ "loss": 5.991410732269287,
+ "step": 384
+ },
+ {
+ "epoch": 5.349344978165939,
+ "grad_norm": 0.015847105532884598,
+ "learning_rate": 0.0006,
+ "loss": 6.100622177124023,
+ "step": 385
+ },
+ {
+ "epoch": 5.3633187772925766,
+ "grad_norm": 0.016179397702217102,
+ "learning_rate": 0.0006,
+ "loss": 6.032659530639648,
+ "step": 386
+ },
+ {
+ "epoch": 5.377292576419214,
+ "grad_norm": 0.018256602808833122,
+ "learning_rate": 0.0006,
+ "loss": 5.961983680725098,
+ "step": 387
+ },
+ {
+ "epoch": 5.391266375545851,
+ "grad_norm": 0.02805912122130394,
+ "learning_rate": 0.0006,
+ "loss": 5.983541965484619,
+ "step": 388
+ },
+ {
+ "epoch": 5.405240174672489,
+ "grad_norm": 0.039082255214452744,
+ "learning_rate": 0.0006,
+ "loss": 6.021474361419678,
+ "step": 389
+ },
+ {
+ "epoch": 5.419213973799127,
+ "grad_norm": 0.036757659167051315,
+ "learning_rate": 0.0006,
+ "loss": 6.005046367645264,
+ "step": 390
+ },
+ {
+ "epoch": 5.4331877729257645,
+ "grad_norm": 0.035277366638183594,
+ "learning_rate": 0.0006,
+ "loss": 6.036296844482422,
+ "step": 391
+ },
+ {
+ "epoch": 5.447161572052401,
+ "grad_norm": 0.034404635429382324,
+ "learning_rate": 0.0006,
+ "loss": 5.994539260864258,
+ "step": 392
+ },
+ {
+ "epoch": 5.461135371179039,
+ "grad_norm": 0.0377206988632679,
+ "learning_rate": 0.0006,
+ "loss": 5.941534519195557,
+ "step": 393
+ },
+ {
+ "epoch": 5.475109170305677,
+ "grad_norm": 0.0389922633767128,
+ "learning_rate": 0.0006,
+ "loss": 5.983644485473633,
+ "step": 394
+ },
+ {
+ "epoch": 5.489082969432315,
+ "grad_norm": 0.04176778718829155,
+ "learning_rate": 0.0006,
+ "loss": 6.030922889709473,
+ "step": 395
+ },
+ {
+ "epoch": 5.503056768558952,
+ "grad_norm": 0.0337153784930706,
+ "learning_rate": 0.0006,
+ "loss": 5.961367130279541,
+ "step": 396
+ },
+ {
+ "epoch": 5.517030567685589,
+ "grad_norm": 0.036116816103458405,
+ "learning_rate": 0.0006,
+ "loss": 5.819280624389648,
+ "step": 397
+ },
+ {
+ "epoch": 5.531004366812227,
+ "grad_norm": 0.030724041163921356,
+ "learning_rate": 0.0006,
+ "loss": 5.906380653381348,
+ "step": 398
+ },
+ {
+ "epoch": 5.544978165938865,
+ "grad_norm": 0.030264202505350113,
+ "learning_rate": 0.0006,
+ "loss": 5.975833415985107,
+ "step": 399
+ },
+ {
+ "epoch": 5.558951965065502,
+ "grad_norm": 0.032096318900585175,
+ "learning_rate": 0.0006,
+ "loss": 6.001348972320557,
+ "step": 400
+ },
+ {
+ "epoch": 5.5729257641921395,
+ "grad_norm": 0.030579356476664543,
+ "learning_rate": 0.0006,
+ "loss": 5.928768634796143,
+ "step": 401
+ },
+ {
+ "epoch": 5.586899563318777,
+ "grad_norm": 0.028241973370313644,
+ "learning_rate": 0.0006,
+ "loss": 5.92582893371582,
+ "step": 402
+ },
+ {
+ "epoch": 5.600873362445415,
+ "grad_norm": 0.023086953908205032,
+ "learning_rate": 0.0006,
+ "loss": 5.892926216125488,
+ "step": 403
+ },
+ {
+ "epoch": 5.614847161572053,
+ "grad_norm": 0.02547992393374443,
+ "learning_rate": 0.0006,
+ "loss": 5.903195381164551,
+ "step": 404
+ },
+ {
+ "epoch": 5.62882096069869,
+ "grad_norm": 0.023089831694960594,
+ "learning_rate": 0.0006,
+ "loss": 5.912033557891846,
+ "step": 405
+ },
+ {
+ "epoch": 5.642794759825327,
+ "grad_norm": 0.025446368381381035,
+ "learning_rate": 0.0006,
+ "loss": 5.926138401031494,
+ "step": 406
+ },
+ {
+ "epoch": 5.656768558951965,
+ "grad_norm": 0.03748747706413269,
+ "learning_rate": 0.0006,
+ "loss": 5.913451194763184,
+ "step": 407
+ },
+ {
+ "epoch": 5.670742358078603,
+ "grad_norm": 0.03493810072541237,
+ "learning_rate": 0.0006,
+ "loss": 5.97633695602417,
+ "step": 408
+ },
+ {
+ "epoch": 5.68471615720524,
+ "grad_norm": 0.023131275549530983,
+ "learning_rate": 0.0006,
+ "loss": 5.899571418762207,
+ "step": 409
+ },
+ {
+ "epoch": 5.698689956331878,
+ "grad_norm": 0.02986014075577259,
+ "learning_rate": 0.0006,
+ "loss": 5.894903182983398,
+ "step": 410
+ },
+ {
+ "epoch": 5.712663755458515,
+ "grad_norm": 0.030171144753694534,
+ "learning_rate": 0.0006,
+ "loss": 5.876595497131348,
+ "step": 411
+ },
+ {
+ "epoch": 5.726637554585153,
+ "grad_norm": 0.029377546161413193,
+ "learning_rate": 0.0006,
+ "loss": 5.911637306213379,
+ "step": 412
+ },
+ {
+ "epoch": 5.74061135371179,
+ "grad_norm": 0.029829490929841995,
+ "learning_rate": 0.0006,
+ "loss": 5.819629669189453,
+ "step": 413
+ },
+ {
+ "epoch": 5.754585152838428,
+ "grad_norm": 0.024139299988746643,
+ "learning_rate": 0.0006,
+ "loss": 5.81728458404541,
+ "step": 414
+ },
+ {
+ "epoch": 5.7685589519650655,
+ "grad_norm": 0.021453432738780975,
+ "learning_rate": 0.0006,
+ "loss": 5.987326145172119,
+ "step": 415
+ },
+ {
+ "epoch": 5.782532751091703,
+ "grad_norm": 0.017942246049642563,
+ "learning_rate": 0.0006,
+ "loss": 5.893008232116699,
+ "step": 416
+ },
+ {
+ "epoch": 5.796506550218341,
+ "grad_norm": 0.019724637269973755,
+ "learning_rate": 0.0006,
+ "loss": 5.912441253662109,
+ "step": 417
+ },
+ {
+ "epoch": 5.810480349344978,
+ "grad_norm": 0.018597912043333054,
+ "learning_rate": 0.0006,
+ "loss": 5.860394477844238,
+ "step": 418
+ },
+ {
+ "epoch": 5.824454148471616,
+ "grad_norm": 0.016232412308454514,
+ "learning_rate": 0.0006,
+ "loss": 5.94991397857666,
+ "step": 419
+ },
+ {
+ "epoch": 5.8384279475982535,
+ "grad_norm": 0.01667204685509205,
+ "learning_rate": 0.0006,
+ "loss": 5.8506364822387695,
+ "step": 420
+ },
+ {
+ "epoch": 5.85240174672489,
+ "grad_norm": 0.014232151210308075,
+ "learning_rate": 0.0006,
+ "loss": 5.928424835205078,
+ "step": 421
+ },
+ {
+ "epoch": 5.866375545851528,
+ "grad_norm": 0.01570476032793522,
+ "learning_rate": 0.0006,
+ "loss": 5.953692436218262,
+ "step": 422
+ },
+ {
+ "epoch": 5.880349344978166,
+ "grad_norm": 0.01586179807782173,
+ "learning_rate": 0.0006,
+ "loss": 5.820014476776123,
+ "step": 423
+ },
+ {
+ "epoch": 5.894323144104804,
+ "grad_norm": 0.014729145914316177,
+ "learning_rate": 0.0006,
+ "loss": 5.860154628753662,
+ "step": 424
+ },
+ {
+ "epoch": 5.908296943231441,
+ "grad_norm": 0.013111459091305733,
+ "learning_rate": 0.0006,
+ "loss": 5.85988187789917,
+ "step": 425
+ },
+ {
+ "epoch": 5.922270742358078,
+ "grad_norm": 0.012937922962009907,
+ "learning_rate": 0.0006,
+ "loss": 5.756265640258789,
+ "step": 426
+ },
+ {
+ "epoch": 5.936244541484716,
+ "grad_norm": 0.01377453189343214,
+ "learning_rate": 0.0006,
+ "loss": 5.8333048820495605,
+ "step": 427
+ },
+ {
+ "epoch": 5.950218340611354,
+ "grad_norm": 0.014054795727133751,
+ "learning_rate": 0.0006,
+ "loss": 5.803333759307861,
+ "step": 428
+ },
+ {
+ "epoch": 5.964192139737992,
+ "grad_norm": 0.01674959622323513,
+ "learning_rate": 0.0006,
+ "loss": 5.8718485832214355,
+ "step": 429
+ },
+ {
+ "epoch": 5.978165938864628,
+ "grad_norm": 0.018733017146587372,
+ "learning_rate": 0.0006,
+ "loss": 5.875979900360107,
+ "step": 430
+ },
+ {
+ "epoch": 5.992139737991266,
+ "grad_norm": 0.02088983915746212,
+ "learning_rate": 0.0006,
+ "loss": 5.763634204864502,
+ "step": 431
+ },
+ {
+ "epoch": 6.0,
+ "grad_norm": 0.026705985888838768,
+ "learning_rate": 0.0006,
+ "loss": 5.6922478675842285,
+ "step": 432
+ },
+ {
+ "epoch": 6.0,
+ "eval_loss": 5.847958087921143,
+ "eval_runtime": 58.5451,
+ "eval_samples_per_second": 41.711,
+ "eval_steps_per_second": 1.315,
+ "step": 432
+ },
+ {
+ "epoch": 6.013973799126638,
+ "grad_norm": 0.03181544691324234,
+ "learning_rate": 0.0006,
+ "loss": 5.786368370056152,
+ "step": 433
+ },
+ {
+ "epoch": 6.0279475982532755,
+ "grad_norm": 0.03238112851977348,
+ "learning_rate": 0.0006,
+ "loss": 5.810310363769531,
+ "step": 434
+ },
+ {
+ "epoch": 6.041921397379912,
+ "grad_norm": 0.03916756808757782,
+ "learning_rate": 0.0006,
+ "loss": 5.7761993408203125,
+ "step": 435
+ },
+ {
+ "epoch": 6.05589519650655,
+ "grad_norm": 0.044009730219841,
+ "learning_rate": 0.0006,
+ "loss": 5.898112773895264,
+ "step": 436
+ },
+ {
+ "epoch": 6.069868995633188,
+ "grad_norm": 0.045235246419906616,
+ "learning_rate": 0.0006,
+ "loss": 5.876371383666992,
+ "step": 437
+ },
+ {
+ "epoch": 6.083842794759826,
+ "grad_norm": 0.050218384712934494,
+ "learning_rate": 0.0006,
+ "loss": 5.805103302001953,
+ "step": 438
+ },
+ {
+ "epoch": 6.097816593886463,
+ "grad_norm": 0.05444490164518356,
+ "learning_rate": 0.0006,
+ "loss": 5.826424598693848,
+ "step": 439
+ },
+ {
+ "epoch": 6.1117903930131,
+ "grad_norm": 0.04890982061624527,
+ "learning_rate": 0.0006,
+ "loss": 5.7735395431518555,
+ "step": 440
+ },
+ {
+ "epoch": 6.125764192139738,
+ "grad_norm": 0.05478639155626297,
+ "learning_rate": 0.0006,
+ "loss": 5.823677062988281,
+ "step": 441
+ },
+ {
+ "epoch": 6.139737991266376,
+ "grad_norm": 0.057562313973903656,
+ "learning_rate": 0.0006,
+ "loss": 5.889334678649902,
+ "step": 442
+ },
+ {
+ "epoch": 6.153711790393013,
+ "grad_norm": 0.06447703391313553,
+ "learning_rate": 0.0006,
+ "loss": 5.84621524810791,
+ "step": 443
+ },
+ {
+ "epoch": 6.1676855895196505,
+ "grad_norm": 0.048861872404813766,
+ "learning_rate": 0.0006,
+ "loss": 5.931595802307129,
+ "step": 444
+ },
+ {
+ "epoch": 6.181659388646288,
+ "grad_norm": 0.05521339178085327,
+ "learning_rate": 0.0006,
+ "loss": 5.8229780197143555,
+ "step": 445
+ },
+ {
+ "epoch": 6.195633187772926,
+ "grad_norm": 0.053666990250349045,
+ "learning_rate": 0.0006,
+ "loss": 5.879191875457764,
+ "step": 446
+ },
+ {
+ "epoch": 6.209606986899563,
+ "grad_norm": 0.0451025515794754,
+ "learning_rate": 0.0006,
+ "loss": 5.801628112792969,
+ "step": 447
+ },
+ {
+ "epoch": 6.223580786026201,
+ "grad_norm": 0.04291655868291855,
+ "learning_rate": 0.0006,
+ "loss": 5.914680480957031,
+ "step": 448
+ },
+ {
+ "epoch": 6.2375545851528384,
+ "grad_norm": 0.035102490335702896,
+ "learning_rate": 0.0006,
+ "loss": 5.843682765960693,
+ "step": 449
+ },
+ {
+ "epoch": 6.251528384279476,
+ "grad_norm": 0.03403995931148529,
+ "learning_rate": 0.0006,
+ "loss": 5.827154159545898,
+ "step": 450
+ },
+ {
+ "epoch": 6.265502183406113,
+ "grad_norm": 0.03444375470280647,
+ "learning_rate": 0.0006,
+ "loss": 5.800136566162109,
+ "step": 451
+ },
+ {
+ "epoch": 6.279475982532751,
+ "grad_norm": 0.03165159001946449,
+ "learning_rate": 0.0006,
+ "loss": 5.906252384185791,
+ "step": 452
+ },
+ {
+ "epoch": 6.293449781659389,
+ "grad_norm": 0.026153093203902245,
+ "learning_rate": 0.0006,
+ "loss": 5.7423200607299805,
+ "step": 453
+ },
+ {
+ "epoch": 6.307423580786026,
+ "grad_norm": 0.024357657879590988,
+ "learning_rate": 0.0006,
+ "loss": 5.701364517211914,
+ "step": 454
+ },
+ {
+ "epoch": 6.321397379912664,
+ "grad_norm": 0.021508049219846725,
+ "learning_rate": 0.0006,
+ "loss": 5.8328094482421875,
+ "step": 455
+ },
+ {
+ "epoch": 6.335371179039301,
+ "grad_norm": 0.017313921824097633,
+ "learning_rate": 0.0006,
+ "loss": 5.792145252227783,
+ "step": 456
+ },
+ {
+ "epoch": 6.349344978165939,
+ "grad_norm": 0.018563397228717804,
+ "learning_rate": 0.0006,
+ "loss": 5.732672214508057,
+ "step": 457
+ },
+ {
+ "epoch": 6.3633187772925766,
+ "grad_norm": 0.016568679362535477,
+ "learning_rate": 0.0006,
+ "loss": 5.81948184967041,
+ "step": 458
+ },
+ {
+ "epoch": 6.377292576419214,
+ "grad_norm": 0.014133770950138569,
+ "learning_rate": 0.0006,
+ "loss": 5.734982490539551,
+ "step": 459
+ },
+ {
+ "epoch": 6.391266375545851,
+ "grad_norm": 0.014114447869360447,
+ "learning_rate": 0.0006,
+ "loss": 5.70073127746582,
+ "step": 460
+ },
+ {
+ "epoch": 6.405240174672489,
+ "grad_norm": 0.012907393276691437,
+ "learning_rate": 0.0006,
+ "loss": 5.758626937866211,
+ "step": 461
+ },
+ {
+ "epoch": 6.419213973799127,
+ "grad_norm": 0.012673444114625454,
+ "learning_rate": 0.0006,
+ "loss": 5.775138854980469,
+ "step": 462
+ },
+ {
+ "epoch": 6.4331877729257645,
+ "grad_norm": 0.01311410591006279,
+ "learning_rate": 0.0006,
+ "loss": 5.70920991897583,
+ "step": 463
+ },
+ {
+ "epoch": 6.447161572052401,
+ "grad_norm": 0.012935544364154339,
+ "learning_rate": 0.0006,
+ "loss": 5.781479835510254,
+ "step": 464
+ },
+ {
+ "epoch": 6.461135371179039,
+ "grad_norm": 0.01439738366752863,
+ "learning_rate": 0.0006,
+ "loss": 5.691315650939941,
+ "step": 465
+ },
+ {
+ "epoch": 6.475109170305677,
+ "grad_norm": 0.01205186452716589,
+ "learning_rate": 0.0006,
+ "loss": 5.662499904632568,
+ "step": 466
+ },
+ {
+ "epoch": 6.489082969432315,
+ "grad_norm": 0.011575652286410332,
+ "learning_rate": 0.0006,
+ "loss": 5.6271772384643555,
+ "step": 467
+ },
+ {
+ "epoch": 6.503056768558952,
+ "grad_norm": 0.011672502383589745,
+ "learning_rate": 0.0006,
+ "loss": 5.761662006378174,
+ "step": 468
+ },
+ {
+ "epoch": 6.517030567685589,
+ "grad_norm": 0.011541751213371754,
+ "learning_rate": 0.0006,
+ "loss": 5.762078285217285,
+ "step": 469
+ },
+ {
+ "epoch": 6.531004366812227,
+ "grad_norm": 0.011396365240216255,
+ "learning_rate": 0.0006,
+ "loss": 5.67873477935791,
+ "step": 470
+ },
+ {
+ "epoch": 6.544978165938865,
+ "grad_norm": 0.010688499547541142,
+ "learning_rate": 0.0006,
+ "loss": 5.717051982879639,
+ "step": 471
+ },
+ {
+ "epoch": 6.558951965065502,
+ "grad_norm": 0.012224317528307438,
+ "learning_rate": 0.0006,
+ "loss": 5.707948684692383,
+ "step": 472
+ },
+ {
+ "epoch": 6.5729257641921395,
+ "grad_norm": 0.011856825090944767,
+ "learning_rate": 0.0006,
+ "loss": 5.70878791809082,
+ "step": 473
+ },
+ {
+ "epoch": 6.586899563318777,
+ "grad_norm": 0.01199064590036869,
+ "learning_rate": 0.0006,
+ "loss": 5.708697319030762,
+ "step": 474
+ },
+ {
+ "epoch": 6.600873362445415,
+ "grad_norm": 0.01219446212053299,
+ "learning_rate": 0.0006,
+ "loss": 5.6061577796936035,
+ "step": 475
+ },
+ {
+ "epoch": 6.614847161572053,
+ "grad_norm": 0.013397484086453915,
+ "learning_rate": 0.0006,
+ "loss": 5.624789714813232,
+ "step": 476
+ },
+ {
+ "epoch": 6.62882096069869,
+ "grad_norm": 0.01483136136084795,
+ "learning_rate": 0.0006,
+ "loss": 5.721141338348389,
+ "step": 477
+ },
+ {
+ "epoch": 6.642794759825327,
+ "grad_norm": 0.0194488987326622,
+ "learning_rate": 0.0006,
+ "loss": 5.604279518127441,
+ "step": 478
+ },
+ {
+ "epoch": 6.656768558951965,
+ "grad_norm": 0.02079896256327629,
+ "learning_rate": 0.0006,
+ "loss": 5.586322784423828,
+ "step": 479
+ },
+ {
+ "epoch": 6.670742358078603,
+ "grad_norm": 0.018996229395270348,
+ "learning_rate": 0.0006,
+ "loss": 5.5987772941589355,
+ "step": 480
+ },
+ {
+ "epoch": 6.68471615720524,
+ "grad_norm": 0.015779603272676468,
+ "learning_rate": 0.0006,
+ "loss": 5.588602066040039,
+ "step": 481
+ },
+ {
+ "epoch": 6.698689956331878,
+ "grad_norm": 0.015322973020374775,
+ "learning_rate": 0.0006,
+ "loss": 5.686467170715332,
+ "step": 482
+ },
+ {
+ "epoch": 6.712663755458515,
+ "grad_norm": 0.015282213687896729,
+ "learning_rate": 0.0006,
+ "loss": 5.664676666259766,
+ "step": 483
+ },
+ {
+ "epoch": 6.726637554585153,
+ "grad_norm": 0.016589272767305374,
+ "learning_rate": 0.0006,
+ "loss": 5.666739463806152,
+ "step": 484
+ },
+ {
+ "epoch": 6.74061135371179,
+ "grad_norm": 0.01811421848833561,
+ "learning_rate": 0.0006,
+ "loss": 5.647593021392822,
+ "step": 485
+ },
+ {
+ "epoch": 6.754585152838428,
+ "grad_norm": 0.019652029499411583,
+ "learning_rate": 0.0006,
+ "loss": 5.628726959228516,
+ "step": 486
+ },
+ {
+ "epoch": 6.7685589519650655,
+ "grad_norm": 0.02118665538728237,
+ "learning_rate": 0.0006,
+ "loss": 5.6580705642700195,
+ "step": 487
+ },
+ {
+ "epoch": 6.782532751091703,
+ "grad_norm": 0.02237832546234131,
+ "learning_rate": 0.0006,
+ "loss": 5.635931968688965,
+ "step": 488
+ },
+ {
+ "epoch": 6.796506550218341,
+ "grad_norm": 0.023897631093859673,
+ "learning_rate": 0.0006,
+ "loss": 5.517949104309082,
+ "step": 489
+ },
+ {
+ "epoch": 6.810480349344978,
+ "grad_norm": 0.02442227490246296,
+ "learning_rate": 0.0006,
+ "loss": 5.600021839141846,
+ "step": 490
+ },
+ {
+ "epoch": 6.824454148471616,
+ "grad_norm": 0.024675287306308746,
+ "learning_rate": 0.0006,
+ "loss": 5.6639909744262695,
+ "step": 491
+ },
+ {
+ "epoch": 6.8384279475982535,
+ "grad_norm": 0.030372392386198044,
+ "learning_rate": 0.0006,
+ "loss": 5.582888603210449,
+ "step": 492
+ },
+ {
+ "epoch": 6.85240174672489,
+ "grad_norm": 0.03237690031528473,
+ "learning_rate": 0.0006,
+ "loss": 5.573896884918213,
+ "step": 493
+ },
+ {
+ "epoch": 6.866375545851528,
+ "grad_norm": 0.03511589393019676,
+ "learning_rate": 0.0006,
+ "loss": 5.565878868103027,
+ "step": 494
+ },
+ {
+ "epoch": 6.880349344978166,
+ "grad_norm": 0.03883938118815422,
+ "learning_rate": 0.0006,
+ "loss": 5.58540678024292,
+ "step": 495
+ },
+ {
+ "epoch": 6.894323144104804,
+ "grad_norm": 0.04175502806901932,
+ "learning_rate": 0.0006,
+ "loss": 5.6095194816589355,
+ "step": 496
+ },
+ {
+ "epoch": 6.908296943231441,
+ "grad_norm": 0.04013441503047943,
+ "learning_rate": 0.0006,
+ "loss": 5.654401779174805,
+ "step": 497
+ },
+ {
+ "epoch": 6.922270742358078,
+ "grad_norm": 0.04618770629167557,
+ "learning_rate": 0.0006,
+ "loss": 5.6483473777771,
+ "step": 498
+ },
+ {
+ "epoch": 6.936244541484716,
+ "grad_norm": 0.04507308453321457,
+ "learning_rate": 0.0006,
+ "loss": 5.5856218338012695,
+ "step": 499
+ },
+ {
+ "epoch": 6.950218340611354,
+ "grad_norm": 0.03151383996009827,
+ "learning_rate": 0.0006,
+ "loss": 5.594086647033691,
+ "step": 500
+ },
+ {
+ "epoch": 6.964192139737992,
+ "grad_norm": 0.026883797720074654,
+ "learning_rate": 0.0006,
+ "loss": 5.607676029205322,
+ "step": 501
+ },
+ {
+ "epoch": 6.978165938864628,
+ "grad_norm": 0.02981836162507534,
+ "learning_rate": 0.0006,
+ "loss": 5.563666820526123,
+ "step": 502
+ },
+ {
+ "epoch": 6.992139737991266,
+ "grad_norm": 0.03233477845788002,
+ "learning_rate": 0.0006,
+ "loss": 5.667543411254883,
+ "step": 503
+ },
+ {
+ "epoch": 7.0,
+ "grad_norm": 0.029927456751465797,
+ "learning_rate": 0.0006,
+ "loss": 5.733482837677002,
+ "step": 504
+ },
+ {
+ "epoch": 7.0,
+ "eval_loss": 5.679076671600342,
+ "eval_runtime": 59.5965,
+ "eval_samples_per_second": 40.976,
+ "eval_steps_per_second": 1.292,
+ "step": 504
+ },
+ {
+ "epoch": 7.013973799126638,
+ "grad_norm": 0.027081597596406937,
+ "learning_rate": 0.0006,
+ "loss": 5.634487152099609,
+ "step": 505
+ },
+ {
+ "epoch": 7.0279475982532755,
+ "grad_norm": 0.031959421932697296,
+ "learning_rate": 0.0006,
+ "loss": 5.643294811248779,
+ "step": 506
+ },
+ {
+ "epoch": 7.041921397379912,
+ "grad_norm": 0.02802063524723053,
+ "learning_rate": 0.0006,
+ "loss": 5.573239326477051,
+ "step": 507
+ },
+ {
+ "epoch": 7.05589519650655,
+ "grad_norm": 0.032579705119132996,
+ "learning_rate": 0.0006,
+ "loss": 5.627150058746338,
+ "step": 508
+ },
+ {
+ "epoch": 7.069868995633188,
+ "grad_norm": 0.02503928542137146,
+ "learning_rate": 0.0006,
+ "loss": 5.6045308113098145,
+ "step": 509
+ },
+ {
+ "epoch": 7.083842794759826,
+ "grad_norm": 0.02421317622065544,
+ "learning_rate": 0.0006,
+ "loss": 5.522153377532959,
+ "step": 510
+ },
+ {
+ "epoch": 7.097816593886463,
+ "grad_norm": 0.021983426064252853,
+ "learning_rate": 0.0006,
+ "loss": 5.632939338684082,
+ "step": 511
+ },
+ {
+ "epoch": 7.1117903930131,
+ "grad_norm": 0.021933183073997498,
+ "learning_rate": 0.0006,
+ "loss": 5.592185974121094,
+ "step": 512
+ },
+ {
+ "epoch": 7.125764192139738,
+ "grad_norm": 0.02287031151354313,
+ "learning_rate": 0.0006,
+ "loss": 5.613700866699219,
+ "step": 513
+ },
+ {
+ "epoch": 7.139737991266376,
+ "grad_norm": 0.02281602844595909,
+ "learning_rate": 0.0006,
+ "loss": 5.551383018493652,
+ "step": 514
+ },
+ {
+ "epoch": 7.153711790393013,
+ "grad_norm": 0.021050360053777695,
+ "learning_rate": 0.0006,
+ "loss": 5.55275821685791,
+ "step": 515
+ },
+ {
+ "epoch": 7.1676855895196505,
+ "grad_norm": 0.018299926072359085,
+ "learning_rate": 0.0006,
+ "loss": 5.528225421905518,
+ "step": 516
+ },
+ {
+ "epoch": 7.181659388646288,
+ "grad_norm": 0.02024853602051735,
+ "learning_rate": 0.0006,
+ "loss": 5.522453784942627,
+ "step": 517
+ },
+ {
+ "epoch": 7.195633187772926,
+ "grad_norm": 0.02135239914059639,
+ "learning_rate": 0.0006,
+ "loss": 5.527522563934326,
+ "step": 518
+ },
+ {
+ "epoch": 7.209606986899563,
+ "grad_norm": 0.023394184187054634,
+ "learning_rate": 0.0006,
+ "loss": 5.4535017013549805,
+ "step": 519
+ },
+ {
+ "epoch": 7.223580786026201,
+ "grad_norm": 0.025952080264687538,
+ "learning_rate": 0.0006,
+ "loss": 5.5374674797058105,
+ "step": 520
+ },
+ {
+ "epoch": 7.2375545851528384,
+ "grad_norm": 0.022287718951702118,
+ "learning_rate": 0.0006,
+ "loss": 5.493753433227539,
+ "step": 521
+ },
+ {
+ "epoch": 7.251528384279476,
+ "grad_norm": 0.02018447406589985,
+ "learning_rate": 0.0006,
+ "loss": 5.510575294494629,
+ "step": 522
+ },
+ {
+ "epoch": 7.265502183406113,
+ "grad_norm": 0.02050507813692093,
+ "learning_rate": 0.0006,
+ "loss": 5.466026782989502,
+ "step": 523
+ },
+ {
+ "epoch": 7.279475982532751,
+ "grad_norm": 0.023688802495598793,
+ "learning_rate": 0.0006,
+ "loss": 5.485815525054932,
+ "step": 524
+ },
+ {
+ "epoch": 7.293449781659389,
+ "grad_norm": 0.021078843623399734,
+ "learning_rate": 0.0006,
+ "loss": 5.415581703186035,
+ "step": 525
+ },
+ {
+ "epoch": 7.307423580786026,
+ "grad_norm": 0.018266011029481888,
+ "learning_rate": 0.0006,
+ "loss": 5.402815818786621,
+ "step": 526
+ },
+ {
+ "epoch": 7.321397379912664,
+ "grad_norm": 0.019693493843078613,
+ "learning_rate": 0.0006,
+ "loss": 5.504674911499023,
+ "step": 527
+ },
+ {
+ "epoch": 7.335371179039301,
+ "grad_norm": 0.02031373605132103,
+ "learning_rate": 0.0006,
+ "loss": 5.503837585449219,
+ "step": 528
+ },
+ {
+ "epoch": 7.349344978165939,
+ "grad_norm": 0.017636626958847046,
+ "learning_rate": 0.0006,
+ "loss": 5.530580997467041,
+ "step": 529
+ },
+ {
+ "epoch": 7.3633187772925766,
+ "grad_norm": 0.01787244901061058,
+ "learning_rate": 0.0006,
+ "loss": 5.450218677520752,
+ "step": 530
+ },
+ {
+ "epoch": 7.377292576419214,
+ "grad_norm": 0.0170669574290514,
+ "learning_rate": 0.0006,
+ "loss": 5.481570243835449,
+ "step": 531
+ },
+ {
+ "epoch": 7.391266375545851,
+ "grad_norm": 0.01802165061235428,
+ "learning_rate": 0.0006,
+ "loss": 5.386394500732422,
+ "step": 532
+ },
+ {
+ "epoch": 7.405240174672489,
+ "grad_norm": 0.022949472069740295,
+ "learning_rate": 0.0006,
+ "loss": 5.424929618835449,
+ "step": 533
+ },
+ {
+ "epoch": 7.419213973799127,
+ "grad_norm": 0.034213390201330185,
+ "learning_rate": 0.0006,
+ "loss": 5.346663475036621,
+ "step": 534
+ },
+ {
+ "epoch": 7.4331877729257645,
+ "grad_norm": 0.03847593814134598,
+ "learning_rate": 0.0006,
+ "loss": 5.418482780456543,
+ "step": 535
+ },
+ {
+ "epoch": 7.447161572052401,
+ "grad_norm": 0.029393676668405533,
+ "learning_rate": 0.0006,
+ "loss": 5.456090927124023,
+ "step": 536
+ },
+ {
+ "epoch": 7.461135371179039,
+ "grad_norm": 0.03388667106628418,
+ "learning_rate": 0.0006,
+ "loss": 5.572983741760254,
+ "step": 537
+ },
+ {
+ "epoch": 7.475109170305677,
+ "grad_norm": 0.039690613746643066,
+ "learning_rate": 0.0006,
+ "loss": 5.414067268371582,
+ "step": 538
+ },
+ {
+ "epoch": 7.489082969432315,
+ "grad_norm": 0.03630776330828667,
+ "learning_rate": 0.0006,
+ "loss": 5.523739814758301,
+ "step": 539
+ },
+ {
+ "epoch": 7.503056768558952,
+ "grad_norm": 0.03356870263814926,
+ "learning_rate": 0.0006,
+ "loss": 5.444767951965332,
+ "step": 540
+ },
+ {
+ "epoch": 7.517030567685589,
+ "grad_norm": 0.030812304466962814,
+ "learning_rate": 0.0006,
+ "loss": 5.461244106292725,
+ "step": 541
+ },
+ {
+ "epoch": 7.531004366812227,
+ "grad_norm": 0.03319217637181282,
+ "learning_rate": 0.0006,
+ "loss": 5.502161026000977,
+ "step": 542
+ },
+ {
+ "epoch": 7.544978165938865,
+ "grad_norm": 0.032997481524944305,
+ "learning_rate": 0.0006,
+ "loss": 5.4162278175354,
+ "step": 543
+ },
+ {
+ "epoch": 7.558951965065502,
+ "grad_norm": 0.03364962339401245,
+ "learning_rate": 0.0006,
+ "loss": 5.415736198425293,
+ "step": 544
+ },
+ {
+ "epoch": 7.5729257641921395,
+ "grad_norm": 0.0344221405684948,
+ "learning_rate": 0.0006,
+ "loss": 5.541967868804932,
+ "step": 545
+ },
+ {
+ "epoch": 7.586899563318777,
+ "grad_norm": 0.029609503224492073,
+ "learning_rate": 0.0006,
+ "loss": 5.440614223480225,
+ "step": 546
+ },
+ {
+ "epoch": 7.600873362445415,
+ "grad_norm": 0.029055926948785782,
+ "learning_rate": 0.0006,
+ "loss": 5.462865352630615,
+ "step": 547
+ },
+ {
+ "epoch": 7.614847161572053,
+ "grad_norm": 0.02658848837018013,
+ "learning_rate": 0.0006,
+ "loss": 5.444467544555664,
+ "step": 548
+ },
+ {
+ "epoch": 7.62882096069869,
+ "grad_norm": 0.026276404038071632,
+ "learning_rate": 0.0006,
+ "loss": 5.462278842926025,
+ "step": 549
+ },
+ {
+ "epoch": 7.642794759825327,
+ "grad_norm": 0.0286889486014843,
+ "learning_rate": 0.0006,
+ "loss": 5.4849534034729,
+ "step": 550
+ },
+ {
+ "epoch": 7.656768558951965,
+ "grad_norm": 0.026849817484617233,
+ "learning_rate": 0.0006,
+ "loss": 5.431473731994629,
+ "step": 551
+ },
+ {
+ "epoch": 7.670742358078603,
+ "grad_norm": 0.02312396466732025,
+ "learning_rate": 0.0006,
+ "loss": 5.45462703704834,
+ "step": 552
+ },
+ {
+ "epoch": 7.68471615720524,
+ "grad_norm": 0.026197485625743866,
+ "learning_rate": 0.0006,
+ "loss": 5.452552318572998,
+ "step": 553
+ },
+ {
+ "epoch": 7.698689956331878,
+ "grad_norm": 0.02747255191206932,
+ "learning_rate": 0.0006,
+ "loss": 5.447169780731201,
+ "step": 554
+ },
+ {
+ "epoch": 7.712663755458515,
+ "grad_norm": 0.028123432770371437,
+ "learning_rate": 0.0006,
+ "loss": 5.394400596618652,
+ "step": 555
+ },
+ {
+ "epoch": 7.726637554585153,
+ "grad_norm": 0.02599870041012764,
+ "learning_rate": 0.0006,
+ "loss": 5.398341178894043,
+ "step": 556
+ },
+ {
+ "epoch": 7.74061135371179,
+ "grad_norm": 0.022171657532453537,
+ "learning_rate": 0.0006,
+ "loss": 5.368820667266846,
+ "step": 557
+ },
+ {
+ "epoch": 7.754585152838428,
+ "grad_norm": 0.022309480234980583,
+ "learning_rate": 0.0006,
+ "loss": 5.416772365570068,
+ "step": 558
+ },
+ {
+ "epoch": 7.7685589519650655,
+ "grad_norm": 0.024074165150523186,
+ "learning_rate": 0.0006,
+ "loss": 5.359002113342285,
+ "step": 559
+ },
+ {
+ "epoch": 7.782532751091703,
+ "grad_norm": 0.02636653557419777,
+ "learning_rate": 0.0006,
+ "loss": 5.484368324279785,
+ "step": 560
+ },
+ {
+ "epoch": 7.796506550218341,
+ "grad_norm": 0.02203752100467682,
+ "learning_rate": 0.0006,
+ "loss": 5.375498294830322,
+ "step": 561
+ },
+ {
+ "epoch": 7.810480349344978,
+ "grad_norm": 0.019924819469451904,
+ "learning_rate": 0.0006,
+ "loss": 5.334723472595215,
+ "step": 562
+ },
+ {
+ "epoch": 7.824454148471616,
+ "grad_norm": 0.018755966797471046,
+ "learning_rate": 0.0006,
+ "loss": 5.348781108856201,
+ "step": 563
+ },
+ {
+ "epoch": 7.8384279475982535,
+ "grad_norm": 0.016747845336794853,
+ "learning_rate": 0.0006,
+ "loss": 5.433079719543457,
+ "step": 564
+ },
+ {
+ "epoch": 7.85240174672489,
+ "grad_norm": 0.0170395877212286,
+ "learning_rate": 0.0006,
+ "loss": 5.346794128417969,
+ "step": 565
+ },
+ {
+ "epoch": 7.866375545851528,
+ "grad_norm": 0.016074176877737045,
+ "learning_rate": 0.0006,
+ "loss": 5.3277740478515625,
+ "step": 566
+ },
+ {
+ "epoch": 7.880349344978166,
+ "grad_norm": 0.014119806699454784,
+ "learning_rate": 0.0006,
+ "loss": 5.483551979064941,
+ "step": 567
+ },
+ {
+ "epoch": 7.894323144104804,
+ "grad_norm": 0.015270394273102283,
+ "learning_rate": 0.0006,
+ "loss": 5.378519535064697,
+ "step": 568
+ },
+ {
+ "epoch": 7.908296943231441,
+ "grad_norm": 0.015677539631724358,
+ "learning_rate": 0.0006,
+ "loss": 5.290210247039795,
+ "step": 569
+ },
+ {
+ "epoch": 7.922270742358078,
+ "grad_norm": 0.015930423513054848,
+ "learning_rate": 0.0006,
+ "loss": 5.427360534667969,
+ "step": 570
+ },
+ {
+ "epoch": 7.936244541484716,
+ "grad_norm": 0.016641564667224884,
+ "learning_rate": 0.0006,
+ "loss": 5.301599979400635,
+ "step": 571
+ },
+ {
+ "epoch": 7.950218340611354,
+ "grad_norm": 0.018293552100658417,
+ "learning_rate": 0.0006,
+ "loss": 5.277889251708984,
+ "step": 572
+ },
+ {
+ "epoch": 7.964192139737992,
+ "grad_norm": 0.017618799582123756,
+ "learning_rate": 0.0006,
+ "loss": 5.297194480895996,
+ "step": 573
+ },
+ {
+ "epoch": 7.978165938864628,
+ "grad_norm": 0.015034242533147335,
+ "learning_rate": 0.0006,
+ "loss": 5.42428731918335,
+ "step": 574
+ },
+ {
+ "epoch": 7.992139737991266,
+ "grad_norm": 0.01617511734366417,
+ "learning_rate": 0.0006,
+ "loss": 5.420044898986816,
+ "step": 575
+ },
+ {
+ "epoch": 8.0,
+ "grad_norm": 0.01837257295846939,
+ "learning_rate": 0.0006,
+ "loss": 5.240387916564941,
+ "step": 576
+ },
+ {
+ "epoch": 8.0,
+ "eval_loss": 5.423932075500488,
+ "eval_runtime": 58.0841,
+ "eval_samples_per_second": 42.042,
+ "eval_steps_per_second": 1.326,
+ "step": 576
+ },
+ {
+ "epoch": 8.013973799126637,
+ "grad_norm": 0.020953809842467308,
+ "learning_rate": 0.0006,
+ "loss": 5.4320783615112305,
+ "step": 577
+ },
+ {
+ "epoch": 8.027947598253276,
+ "grad_norm": 0.0227745920419693,
+ "learning_rate": 0.0006,
+ "loss": 5.310197830200195,
+ "step": 578
+ },
+ {
+ "epoch": 8.041921397379912,
+ "grad_norm": 0.021372603252530098,
+ "learning_rate": 0.0006,
+ "loss": 5.339468002319336,
+ "step": 579
+ },
+ {
+ "epoch": 8.055895196506551,
+ "grad_norm": 0.022401731461286545,
+ "learning_rate": 0.0006,
+ "loss": 5.279829978942871,
+ "step": 580
+ },
+ {
+ "epoch": 8.069868995633188,
+ "grad_norm": 0.023942379280924797,
+ "learning_rate": 0.0006,
+ "loss": 5.372323989868164,
+ "step": 581
+ },
+ {
+ "epoch": 8.083842794759825,
+ "grad_norm": 0.024380534887313843,
+ "learning_rate": 0.0006,
+ "loss": 5.253486633300781,
+ "step": 582
+ },
+ {
+ "epoch": 8.097816593886463,
+ "grad_norm": 0.03246625140309334,
+ "learning_rate": 0.0006,
+ "loss": 5.305774211883545,
+ "step": 583
+ },
+ {
+ "epoch": 8.1117903930131,
+ "grad_norm": 0.030399736016988754,
+ "learning_rate": 0.0006,
+ "loss": 5.370976448059082,
+ "step": 584
+ },
+ {
+ "epoch": 8.125764192139737,
+ "grad_norm": 0.027148913592100143,
+ "learning_rate": 0.0006,
+ "loss": 5.368010520935059,
+ "step": 585
+ },
+ {
+ "epoch": 8.139737991266376,
+ "grad_norm": 0.02925540879368782,
+ "learning_rate": 0.0006,
+ "loss": 5.303013801574707,
+ "step": 586
+ },
+ {
+ "epoch": 8.153711790393013,
+ "grad_norm": 0.02915453538298607,
+ "learning_rate": 0.0006,
+ "loss": 5.329667091369629,
+ "step": 587
+ },
+ {
+ "epoch": 8.167685589519651,
+ "grad_norm": 0.031732626259326935,
+ "learning_rate": 0.0006,
+ "loss": 5.3634934425354,
+ "step": 588
+ },
+ {
+ "epoch": 8.181659388646288,
+ "grad_norm": 0.036166731268167496,
+ "learning_rate": 0.0006,
+ "loss": 5.223980903625488,
+ "step": 589
+ },
+ {
+ "epoch": 8.195633187772925,
+ "grad_norm": 0.035654108971357346,
+ "learning_rate": 0.0006,
+ "loss": 5.361059188842773,
+ "step": 590
+ },
+ {
+ "epoch": 8.209606986899564,
+ "grad_norm": 0.0365324430167675,
+ "learning_rate": 0.0006,
+ "loss": 5.336151599884033,
+ "step": 591
+ },
+ {
+ "epoch": 8.2235807860262,
+ "grad_norm": 0.03578880429267883,
+ "learning_rate": 0.0006,
+ "loss": 5.434076309204102,
+ "step": 592
+ },
+ {
+ "epoch": 8.237554585152838,
+ "grad_norm": 0.035734012722969055,
+ "learning_rate": 0.0006,
+ "loss": 5.330893516540527,
+ "step": 593
+ },
+ {
+ "epoch": 8.251528384279476,
+ "grad_norm": 0.03196857124567032,
+ "learning_rate": 0.0006,
+ "loss": 5.306009769439697,
+ "step": 594
+ },
+ {
+ "epoch": 8.265502183406113,
+ "grad_norm": 0.032042086124420166,
+ "learning_rate": 0.0006,
+ "loss": 5.343267917633057,
+ "step": 595
+ },
+ {
+ "epoch": 8.279475982532752,
+ "grad_norm": 0.03160746395587921,
+ "learning_rate": 0.0006,
+ "loss": 5.2353715896606445,
+ "step": 596
+ },
+ {
+ "epoch": 8.293449781659389,
+ "grad_norm": 0.026689818128943443,
+ "learning_rate": 0.0006,
+ "loss": 5.254042625427246,
+ "step": 597
+ },
+ {
+ "epoch": 8.307423580786025,
+ "grad_norm": 0.02880188450217247,
+ "learning_rate": 0.0006,
+ "loss": 5.326833248138428,
+ "step": 598
+ },
+ {
+ "epoch": 8.321397379912664,
+ "grad_norm": 0.027516381815075874,
+ "learning_rate": 0.0006,
+ "loss": 5.342336177825928,
+ "step": 599
+ },
+ {
+ "epoch": 8.335371179039301,
+ "grad_norm": 0.027875030413269997,
+ "learning_rate": 0.0006,
+ "loss": 5.200719833374023,
+ "step": 600
+ },
+ {
+ "epoch": 8.34934497816594,
+ "grad_norm": 0.0268265288323164,
+ "learning_rate": 0.0006,
+ "loss": 5.249449729919434,
+ "step": 601
+ },
+ {
+ "epoch": 8.363318777292577,
+ "grad_norm": 0.024824608117341995,
+ "learning_rate": 0.0006,
+ "loss": 5.318952560424805,
+ "step": 602
+ },
+ {
+ "epoch": 8.377292576419213,
+ "grad_norm": 0.022990627214312553,
+ "learning_rate": 0.0006,
+ "loss": 5.244993209838867,
+ "step": 603
+ },
+ {
+ "epoch": 8.391266375545852,
+ "grad_norm": 0.022804604843258858,
+ "learning_rate": 0.0006,
+ "loss": 5.3236846923828125,
+ "step": 604
+ },
+ {
+ "epoch": 8.405240174672489,
+ "grad_norm": 0.02419872209429741,
+ "learning_rate": 0.0006,
+ "loss": 5.255486488342285,
+ "step": 605
+ },
+ {
+ "epoch": 8.419213973799126,
+ "grad_norm": 0.021952059119939804,
+ "learning_rate": 0.0006,
+ "loss": 5.3511552810668945,
+ "step": 606
+ },
+ {
+ "epoch": 8.433187772925764,
+ "grad_norm": 0.022375497967004776,
+ "learning_rate": 0.0006,
+ "loss": 5.294790267944336,
+ "step": 607
+ },
+ {
+ "epoch": 8.447161572052401,
+ "grad_norm": 0.019242815673351288,
+ "learning_rate": 0.0006,
+ "loss": 5.252618789672852,
+ "step": 608
+ },
+ {
+ "epoch": 8.46113537117904,
+ "grad_norm": 0.0173486340790987,
+ "learning_rate": 0.0006,
+ "loss": 5.30147647857666,
+ "step": 609
+ },
+ {
+ "epoch": 8.475109170305677,
+ "grad_norm": 0.01612604409456253,
+ "learning_rate": 0.0006,
+ "loss": 5.23960542678833,
+ "step": 610
+ },
+ {
+ "epoch": 8.489082969432314,
+ "grad_norm": 0.014293976128101349,
+ "learning_rate": 0.0006,
+ "loss": 5.119932651519775,
+ "step": 611
+ },
+ {
+ "epoch": 8.503056768558952,
+ "grad_norm": 0.014453536830842495,
+ "learning_rate": 0.0006,
+ "loss": 5.166794776916504,
+ "step": 612
+ },
+ {
+ "epoch": 8.51703056768559,
+ "grad_norm": 0.014685769565403461,
+ "learning_rate": 0.0006,
+ "loss": 5.264822959899902,
+ "step": 613
+ },
+ {
+ "epoch": 8.531004366812226,
+ "grad_norm": 0.01427740603685379,
+ "learning_rate": 0.0006,
+ "loss": 5.278433799743652,
+ "step": 614
+ },
+ {
+ "epoch": 8.544978165938865,
+ "grad_norm": 0.014864951372146606,
+ "learning_rate": 0.0006,
+ "loss": 5.214512348175049,
+ "step": 615
+ },
+ {
+ "epoch": 8.558951965065502,
+ "grad_norm": 0.01565164513885975,
+ "learning_rate": 0.0006,
+ "loss": 5.183066368103027,
+ "step": 616
+ },
+ {
+ "epoch": 8.57292576419214,
+ "grad_norm": 0.01779816672205925,
+ "learning_rate": 0.0006,
+ "loss": 5.163185119628906,
+ "step": 617
+ },
+ {
+ "epoch": 8.586899563318777,
+ "grad_norm": 0.017254827544093132,
+ "learning_rate": 0.0006,
+ "loss": 5.243555068969727,
+ "step": 618
+ },
+ {
+ "epoch": 8.600873362445414,
+ "grad_norm": 0.01650584116578102,
+ "learning_rate": 0.0006,
+ "loss": 5.234884738922119,
+ "step": 619
+ },
+ {
+ "epoch": 8.614847161572053,
+ "grad_norm": 0.017021115869283676,
+ "learning_rate": 0.0006,
+ "loss": 5.12271785736084,
+ "step": 620
+ },
+ {
+ "epoch": 8.62882096069869,
+ "grad_norm": 0.01773759163916111,
+ "learning_rate": 0.0006,
+ "loss": 5.242153167724609,
+ "step": 621
+ },
+ {
+ "epoch": 8.642794759825328,
+ "grad_norm": 0.015679042786359787,
+ "learning_rate": 0.0006,
+ "loss": 5.170779228210449,
+ "step": 622
+ },
+ {
+ "epoch": 8.656768558951965,
+ "grad_norm": 0.013760549947619438,
+ "learning_rate": 0.0006,
+ "loss": 5.238644599914551,
+ "step": 623
+ },
+ {
+ "epoch": 8.670742358078602,
+ "grad_norm": 0.014571522362530231,
+ "learning_rate": 0.0006,
+ "loss": 5.177056312561035,
+ "step": 624
+ },
+ {
+ "epoch": 8.68471615720524,
+ "grad_norm": 0.016209015622735023,
+ "learning_rate": 0.0006,
+ "loss": 5.163750648498535,
+ "step": 625
+ },
+ {
+ "epoch": 8.698689956331878,
+ "grad_norm": 0.016813941299915314,
+ "learning_rate": 0.0006,
+ "loss": 5.183428764343262,
+ "step": 626
+ },
+ {
+ "epoch": 8.712663755458514,
+ "grad_norm": 0.020985357463359833,
+ "learning_rate": 0.0006,
+ "loss": 5.229465007781982,
+ "step": 627
+ },
+ {
+ "epoch": 8.726637554585153,
+ "grad_norm": 0.02679632417857647,
+ "learning_rate": 0.0006,
+ "loss": 5.257368087768555,
+ "step": 628
+ },
+ {
+ "epoch": 8.74061135371179,
+ "grad_norm": 0.025756070390343666,
+ "learning_rate": 0.0006,
+ "loss": 5.253736972808838,
+ "step": 629
+ },
+ {
+ "epoch": 8.754585152838429,
+ "grad_norm": 0.02643490768969059,
+ "learning_rate": 0.0006,
+ "loss": 5.228633403778076,
+ "step": 630
+ },
+ {
+ "epoch": 8.768558951965066,
+ "grad_norm": 0.03303210437297821,
+ "learning_rate": 0.0006,
+ "loss": 5.163540363311768,
+ "step": 631
+ },
+ {
+ "epoch": 8.782532751091702,
+ "grad_norm": 0.030432431027293205,
+ "learning_rate": 0.0006,
+ "loss": 5.243169784545898,
+ "step": 632
+ },
+ {
+ "epoch": 8.796506550218341,
+ "grad_norm": 0.028631288558244705,
+ "learning_rate": 0.0006,
+ "loss": 5.166398048400879,
+ "step": 633
+ },
+ {
+ "epoch": 8.810480349344978,
+ "grad_norm": 0.026188310235738754,
+ "learning_rate": 0.0006,
+ "loss": 5.128122329711914,
+ "step": 634
+ },
+ {
+ "epoch": 8.824454148471617,
+ "grad_norm": 0.025308528915047646,
+ "learning_rate": 0.0006,
+ "loss": 5.211284637451172,
+ "step": 635
+ },
+ {
+ "epoch": 8.838427947598253,
+ "grad_norm": 0.020657729357481003,
+ "learning_rate": 0.0006,
+ "loss": 5.161575794219971,
+ "step": 636
+ },
+ {
+ "epoch": 8.85240174672489,
+ "grad_norm": 0.021901234984397888,
+ "learning_rate": 0.0006,
+ "loss": 5.20050048828125,
+ "step": 637
+ },
+ {
+ "epoch": 8.866375545851529,
+ "grad_norm": 0.021219318732619286,
+ "learning_rate": 0.0006,
+ "loss": 5.211699485778809,
+ "step": 638
+ },
+ {
+ "epoch": 8.880349344978166,
+ "grad_norm": 0.0200974028557539,
+ "learning_rate": 0.0006,
+ "loss": 5.092503547668457,
+ "step": 639
+ },
+ {
+ "epoch": 8.894323144104803,
+ "grad_norm": 0.023804882541298866,
+ "learning_rate": 0.0006,
+ "loss": 5.216068267822266,
+ "step": 640
+ },
+ {
+ "epoch": 8.908296943231441,
+ "grad_norm": 0.026088010519742966,
+ "learning_rate": 0.0006,
+ "loss": 5.155592918395996,
+ "step": 641
+ },
+ {
+ "epoch": 8.922270742358078,
+ "grad_norm": 0.02601276896893978,
+ "learning_rate": 0.0006,
+ "loss": 5.197238922119141,
+ "step": 642
+ },
+ {
+ "epoch": 8.936244541484717,
+ "grad_norm": 0.020387211814522743,
+ "learning_rate": 0.0006,
+ "loss": 5.1955437660217285,
+ "step": 643
+ },
+ {
+ "epoch": 8.950218340611354,
+ "grad_norm": 0.019214622676372528,
+ "learning_rate": 0.0006,
+ "loss": 5.206346035003662,
+ "step": 644
+ },
+ {
+ "epoch": 8.96419213973799,
+ "grad_norm": 0.019674314185976982,
+ "learning_rate": 0.0006,
+ "loss": 5.219857215881348,
+ "step": 645
+ },
+ {
+ "epoch": 8.97816593886463,
+ "grad_norm": 0.020318234339356422,
+ "learning_rate": 0.0006,
+ "loss": 5.191132545471191,
+ "step": 646
+ },
+ {
+ "epoch": 8.992139737991266,
+ "grad_norm": 0.021428676322102547,
+ "learning_rate": 0.0006,
+ "loss": 5.214695930480957,
+ "step": 647
+ },
+ {
+ "epoch": 9.0,
+ "grad_norm": 0.023235054686665535,
+ "learning_rate": 0.0006,
+ "loss": 5.229091644287109,
+ "step": 648
+ },
+ {
+ "epoch": 9.0,
+ "eval_loss": 5.289593696594238,
+ "eval_runtime": 56.9256,
+ "eval_samples_per_second": 42.898,
+ "eval_steps_per_second": 1.353,
+ "step": 648
+ },
+ {
+ "epoch": 9.013973799126637,
+ "grad_norm": 0.02320289984345436,
+ "learning_rate": 0.0006,
+ "loss": 4.983669281005859,
+ "step": 649
+ },
+ {
+ "epoch": 9.027947598253276,
+ "grad_norm": 0.018521282821893692,
+ "learning_rate": 0.0006,
+ "loss": 5.200575828552246,
+ "step": 650
+ },
+ {
+ "epoch": 9.041921397379912,
+ "grad_norm": 0.0180169939994812,
+ "learning_rate": 0.0006,
+ "loss": 5.148033618927002,
+ "step": 651
+ },
+ {
+ "epoch": 9.055895196506551,
+ "grad_norm": 0.019969860091805458,
+ "learning_rate": 0.0006,
+ "loss": 5.099125385284424,
+ "step": 652
+ },
+ {
+ "epoch": 9.069868995633188,
+ "grad_norm": 0.01643305830657482,
+ "learning_rate": 0.0006,
+ "loss": 5.092347145080566,
+ "step": 653
+ },
+ {
+ "epoch": 9.083842794759825,
+ "grad_norm": 0.016022363677620888,
+ "learning_rate": 0.0006,
+ "loss": 5.129279613494873,
+ "step": 654
+ },
+ {
+ "epoch": 9.097816593886463,
+ "grad_norm": 0.01657041162252426,
+ "learning_rate": 0.0006,
+ "loss": 5.159562587738037,
+ "step": 655
+ },
+ {
+ "epoch": 9.1117903930131,
+ "grad_norm": 0.017924228683114052,
+ "learning_rate": 0.0006,
+ "loss": 5.165497779846191,
+ "step": 656
+ },
+ {
+ "epoch": 9.125764192139737,
+ "grad_norm": 0.021592361852526665,
+ "learning_rate": 0.0006,
+ "loss": 5.204775810241699,
+ "step": 657
+ },
+ {
+ "epoch": 9.139737991266376,
+ "grad_norm": 0.026924120262265205,
+ "learning_rate": 0.0006,
+ "loss": 5.128296852111816,
+ "step": 658
+ },
+ {
+ "epoch": 9.153711790393013,
+ "grad_norm": 0.023156899958848953,
+ "learning_rate": 0.0006,
+ "loss": 5.1255340576171875,
+ "step": 659
+ },
+ {
+ "epoch": 9.167685589519651,
+ "grad_norm": 0.020670367404818535,
+ "learning_rate": 0.0006,
+ "loss": 5.128479957580566,
+ "step": 660
+ },
+ {
+ "epoch": 9.181659388646288,
+ "grad_norm": 0.025447173044085503,
+ "learning_rate": 0.0006,
+ "loss": 5.154821872711182,
+ "step": 661
+ },
+ {
+ "epoch": 9.195633187772925,
+ "grad_norm": 0.027940068393945694,
+ "learning_rate": 0.0006,
+ "loss": 5.143466472625732,
+ "step": 662
+ },
+ {
+ "epoch": 9.209606986899564,
+ "grad_norm": 0.026847844943404198,
+ "learning_rate": 0.0006,
+ "loss": 5.045665740966797,
+ "step": 663
+ },
+ {
+ "epoch": 9.2235807860262,
+ "grad_norm": 0.02340601570904255,
+ "learning_rate": 0.0006,
+ "loss": 5.21933126449585,
+ "step": 664
+ },
+ {
+ "epoch": 9.237554585152838,
+ "grad_norm": 0.02340371161699295,
+ "learning_rate": 0.0006,
+ "loss": 4.942949295043945,
+ "step": 665
+ },
+ {
+ "epoch": 9.251528384279476,
+ "grad_norm": 0.02221992425620556,
+ "learning_rate": 0.0006,
+ "loss": 5.197099208831787,
+ "step": 666
+ },
+ {
+ "epoch": 9.265502183406113,
+ "grad_norm": 0.023508677259087563,
+ "learning_rate": 0.0006,
+ "loss": 5.022897720336914,
+ "step": 667
+ },
+ {
+ "epoch": 9.279475982532752,
+ "grad_norm": 0.026120394468307495,
+ "learning_rate": 0.0006,
+ "loss": 5.087725639343262,
+ "step": 668
+ },
+ {
+ "epoch": 9.293449781659389,
+ "grad_norm": 0.027273228392004967,
+ "learning_rate": 0.0006,
+ "loss": 5.090963840484619,
+ "step": 669
+ },
+ {
+ "epoch": 9.307423580786025,
+ "grad_norm": 0.03241586685180664,
+ "learning_rate": 0.0006,
+ "loss": 5.047842979431152,
+ "step": 670
+ },
+ {
+ "epoch": 9.321397379912664,
+ "grad_norm": 0.030245667323470116,
+ "learning_rate": 0.0006,
+ "loss": 5.104803085327148,
+ "step": 671
+ },
+ {
+ "epoch": 9.335371179039301,
+ "grad_norm": 0.027698364108800888,
+ "learning_rate": 0.0006,
+ "loss": 5.0644097328186035,
+ "step": 672
+ },
+ {
+ "epoch": 9.34934497816594,
+ "grad_norm": 0.029692554846405983,
+ "learning_rate": 0.0006,
+ "loss": 5.111942291259766,
+ "step": 673
+ },
+ {
+ "epoch": 9.363318777292577,
+ "grad_norm": 0.03328656405210495,
+ "learning_rate": 0.0006,
+ "loss": 5.1574554443359375,
+ "step": 674
+ },
+ {
+ "epoch": 9.377292576419213,
+ "grad_norm": 0.031938180327415466,
+ "learning_rate": 0.0006,
+ "loss": 5.190371990203857,
+ "step": 675
+ },
+ {
+ "epoch": 9.391266375545852,
+ "grad_norm": 0.024858945980668068,
+ "learning_rate": 0.0006,
+ "loss": 5.121420860290527,
+ "step": 676
+ },
+ {
+ "epoch": 9.405240174672489,
+ "grad_norm": 0.023033203557133675,
+ "learning_rate": 0.0006,
+ "loss": 5.0974345207214355,
+ "step": 677
+ },
+ {
+ "epoch": 9.419213973799126,
+ "grad_norm": 0.022083545103669167,
+ "learning_rate": 0.0006,
+ "loss": 5.046319961547852,
+ "step": 678
+ },
+ {
+ "epoch": 9.433187772925764,
+ "grad_norm": 0.0228437427431345,
+ "learning_rate": 0.0006,
+ "loss": 5.119963645935059,
+ "step": 679
+ },
+ {
+ "epoch": 9.447161572052401,
+ "grad_norm": 0.0242351982742548,
+ "learning_rate": 0.0006,
+ "loss": 5.0444560050964355,
+ "step": 680
+ },
+ {
+ "epoch": 9.46113537117904,
+ "grad_norm": 0.02401185780763626,
+ "learning_rate": 0.0006,
+ "loss": 5.063238620758057,
+ "step": 681
+ },
+ {
+ "epoch": 9.475109170305677,
+ "grad_norm": 0.02560959942638874,
+ "learning_rate": 0.0006,
+ "loss": 5.035224914550781,
+ "step": 682
+ },
+ {
+ "epoch": 9.489082969432314,
+ "grad_norm": 0.024985626339912415,
+ "learning_rate": 0.0006,
+ "loss": 5.049284934997559,
+ "step": 683
+ },
+ {
+ "epoch": 9.503056768558952,
+ "grad_norm": 0.023178115487098694,
+ "learning_rate": 0.0006,
+ "loss": 5.2266621589660645,
+ "step": 684
+ },
+ {
+ "epoch": 9.51703056768559,
+ "grad_norm": 0.0215512253344059,
+ "learning_rate": 0.0006,
+ "loss": 5.1875715255737305,
+ "step": 685
+ },
+ {
+ "epoch": 9.531004366812226,
+ "grad_norm": 0.01662837713956833,
+ "learning_rate": 0.0006,
+ "loss": 5.137012481689453,
+ "step": 686
+ },
+ {
+ "epoch": 9.544978165938865,
+ "grad_norm": 0.01952764391899109,
+ "learning_rate": 0.0006,
+ "loss": 5.072500228881836,
+ "step": 687
+ },
+ {
+ "epoch": 9.558951965065502,
+ "grad_norm": 0.017521077767014503,
+ "learning_rate": 0.0006,
+ "loss": 5.045711994171143,
+ "step": 688
+ },
+ {
+ "epoch": 9.57292576419214,
+ "grad_norm": 0.017355265095829964,
+ "learning_rate": 0.0006,
+ "loss": 5.02881383895874,
+ "step": 689
+ },
+ {
+ "epoch": 9.586899563318777,
+ "grad_norm": 0.015587719157338142,
+ "learning_rate": 0.0006,
+ "loss": 5.132787704467773,
+ "step": 690
+ },
+ {
+ "epoch": 9.600873362445414,
+ "grad_norm": 0.017469845712184906,
+ "learning_rate": 0.0006,
+ "loss": 5.105447292327881,
+ "step": 691
+ },
+ {
+ "epoch": 9.614847161572053,
+ "grad_norm": 0.019630100578069687,
+ "learning_rate": 0.0006,
+ "loss": 5.032464981079102,
+ "step": 692
+ },
+ {
+ "epoch": 9.62882096069869,
+ "grad_norm": 0.01784994639456272,
+ "learning_rate": 0.0006,
+ "loss": 5.081345558166504,
+ "step": 693
+ },
+ {
+ "epoch": 9.642794759825328,
+ "grad_norm": 0.01872754842042923,
+ "learning_rate": 0.0006,
+ "loss": 4.9772114753723145,
+ "step": 694
+ },
+ {
+ "epoch": 9.656768558951965,
+ "grad_norm": 0.01999032311141491,
+ "learning_rate": 0.0006,
+ "loss": 5.089740753173828,
+ "step": 695
+ },
+ {
+ "epoch": 9.670742358078602,
+ "grad_norm": 0.01710767112672329,
+ "learning_rate": 0.0006,
+ "loss": 4.9869890213012695,
+ "step": 696
+ },
+ {
+ "epoch": 9.68471615720524,
+ "grad_norm": 0.015144161880016327,
+ "learning_rate": 0.0006,
+ "loss": 5.022564888000488,
+ "step": 697
+ },
+ {
+ "epoch": 9.698689956331878,
+ "grad_norm": 0.01618543453514576,
+ "learning_rate": 0.0006,
+ "loss": 5.035921573638916,
+ "step": 698
+ },
+ {
+ "epoch": 9.712663755458514,
+ "grad_norm": 0.014653387479484081,
+ "learning_rate": 0.0006,
+ "loss": 4.914456367492676,
+ "step": 699
+ },
+ {
+ "epoch": 9.726637554585153,
+ "grad_norm": 0.013434977270662785,
+ "learning_rate": 0.0006,
+ "loss": 5.063764572143555,
+ "step": 700
+ },
+ {
+ "epoch": 9.74061135371179,
+ "grad_norm": 0.015128864906728268,
+ "learning_rate": 0.0006,
+ "loss": 5.052367687225342,
+ "step": 701
+ },
+ {
+ "epoch": 9.754585152838429,
+ "grad_norm": 0.014210768043994904,
+ "learning_rate": 0.0006,
+ "loss": 5.016489028930664,
+ "step": 702
+ },
+ {
+ "epoch": 9.768558951965066,
+ "grad_norm": 0.015191680751740932,
+ "learning_rate": 0.0006,
+ "loss": 4.990176200866699,
+ "step": 703
+ },
+ {
+ "epoch": 9.782532751091702,
+ "grad_norm": 0.018285127356648445,
+ "learning_rate": 0.0006,
+ "loss": 5.015164375305176,
+ "step": 704
+ },
+ {
+ "epoch": 9.796506550218341,
+ "grad_norm": 0.0204079058021307,
+ "learning_rate": 0.0006,
+ "loss": 5.091142654418945,
+ "step": 705
+ },
+ {
+ "epoch": 9.810480349344978,
+ "grad_norm": 0.01980586163699627,
+ "learning_rate": 0.0006,
+ "loss": 5.03350830078125,
+ "step": 706
+ },
+ {
+ "epoch": 9.824454148471617,
+ "grad_norm": 0.02048729732632637,
+ "learning_rate": 0.0006,
+ "loss": 5.03611421585083,
+ "step": 707
+ },
+ {
+ "epoch": 9.838427947598253,
+ "grad_norm": 0.02058163844048977,
+ "learning_rate": 0.0006,
+ "loss": 4.9613447189331055,
+ "step": 708
+ },
+ {
+ "epoch": 9.85240174672489,
+ "grad_norm": 0.018004924058914185,
+ "learning_rate": 0.0006,
+ "loss": 4.966541767120361,
+ "step": 709
+ },
+ {
+ "epoch": 9.866375545851529,
+ "grad_norm": 0.0198152307420969,
+ "learning_rate": 0.0006,
+ "loss": 5.01750373840332,
+ "step": 710
+ },
+ {
+ "epoch": 9.880349344978166,
+ "grad_norm": 0.018532969057559967,
+ "learning_rate": 0.0006,
+ "loss": 5.097424507141113,
+ "step": 711
+ },
+ {
+ "epoch": 9.894323144104803,
+ "grad_norm": 0.018939971923828125,
+ "learning_rate": 0.0006,
+ "loss": 5.177223205566406,
+ "step": 712
+ },
+ {
+ "epoch": 9.908296943231441,
+ "grad_norm": 0.017969170585274696,
+ "learning_rate": 0.0006,
+ "loss": 5.118013858795166,
+ "step": 713
+ },
+ {
+ "epoch": 9.922270742358078,
+ "grad_norm": 0.018021270632743835,
+ "learning_rate": 0.0006,
+ "loss": 5.034787178039551,
+ "step": 714
+ },
+ {
+ "epoch": 9.936244541484717,
+ "grad_norm": 0.017009710893034935,
+ "learning_rate": 0.0006,
+ "loss": 5.003920555114746,
+ "step": 715
+ },
+ {
+ "epoch": 9.950218340611354,
+ "grad_norm": 0.017722049728035927,
+ "learning_rate": 0.0006,
+ "loss": 5.031939506530762,
+ "step": 716
+ },
+ {
+ "epoch": 9.96419213973799,
+ "grad_norm": 0.017134087160229683,
+ "learning_rate": 0.0006,
+ "loss": 5.093474388122559,
+ "step": 717
+ },
+ {
+ "epoch": 9.97816593886463,
+ "grad_norm": 0.018511613830924034,
+ "learning_rate": 0.0006,
+ "loss": 5.078604698181152,
+ "step": 718
+ },
+ {
+ "epoch": 9.992139737991266,
+ "grad_norm": 0.021206334233283997,
+ "learning_rate": 0.0006,
+ "loss": 5.052572727203369,
+ "step": 719
+ },
+ {
+ "epoch": 10.0,
+ "grad_norm": 0.02239062264561653,
+ "learning_rate": 0.0006,
+ "loss": 5.063547134399414,
+ "step": 720
+ },
+ {
+ "epoch": 10.0,
+ "eval_loss": 5.173044681549072,
+ "eval_runtime": 56.7287,
+ "eval_samples_per_second": 43.047,
+ "eval_steps_per_second": 1.357,
+ "step": 720
+ },
+ {
+ "epoch": 10.013973799126637,
+ "grad_norm": 0.020757969468832016,
+ "learning_rate": 0.0006,
+ "loss": 5.045648574829102,
+ "step": 721
+ },
+ {
+ "epoch": 10.027947598253276,
+ "grad_norm": 0.022697702050209045,
+ "learning_rate": 0.0006,
+ "loss": 4.993833065032959,
+ "step": 722
+ },
+ {
+ "epoch": 10.041921397379912,
+ "grad_norm": 0.02434331737458706,
+ "learning_rate": 0.0006,
+ "loss": 5.073219299316406,
+ "step": 723
+ },
+ {
+ "epoch": 10.055895196506551,
+ "grad_norm": 0.025499660521745682,
+ "learning_rate": 0.0006,
+ "loss": 5.019072532653809,
+ "step": 724
+ },
+ {
+ "epoch": 10.069868995633188,
+ "grad_norm": 0.029281053692102432,
+ "learning_rate": 0.0006,
+ "loss": 4.959851264953613,
+ "step": 725
+ },
+ {
+ "epoch": 10.083842794759825,
+ "grad_norm": 0.033878110349178314,
+ "learning_rate": 0.0006,
+ "loss": 5.132717609405518,
+ "step": 726
+ },
+ {
+ "epoch": 10.097816593886463,
+ "grad_norm": 0.03385764732956886,
+ "learning_rate": 0.0006,
+ "loss": 5.0482892990112305,
+ "step": 727
+ },
+ {
+ "epoch": 10.1117903930131,
+ "grad_norm": 0.03466648980975151,
+ "learning_rate": 0.0006,
+ "loss": 4.946855545043945,
+ "step": 728
+ },
+ {
+ "epoch": 10.125764192139737,
+ "grad_norm": 0.028546305373311043,
+ "learning_rate": 0.0006,
+ "loss": 5.027774333953857,
+ "step": 729
+ },
+ {
+ "epoch": 10.139737991266376,
+ "grad_norm": 0.030575869604945183,
+ "learning_rate": 0.0006,
+ "loss": 4.9768476486206055,
+ "step": 730
+ },
+ {
+ "epoch": 10.153711790393013,
+ "grad_norm": 0.03614303469657898,
+ "learning_rate": 0.0006,
+ "loss": 5.078113555908203,
+ "step": 731
+ },
+ {
+ "epoch": 10.167685589519651,
+ "grad_norm": 0.040991660207509995,
+ "learning_rate": 0.0006,
+ "loss": 5.033185958862305,
+ "step": 732
+ },
+ {
+ "epoch": 10.181659388646288,
+ "grad_norm": 0.04551699757575989,
+ "learning_rate": 0.0006,
+ "loss": 5.0820746421813965,
+ "step": 733
+ },
+ {
+ "epoch": 10.195633187772925,
+ "grad_norm": 0.040193989872932434,
+ "learning_rate": 0.0006,
+ "loss": 5.059700965881348,
+ "step": 734
+ },
+ {
+ "epoch": 10.209606986899564,
+ "grad_norm": 0.035851605236530304,
+ "learning_rate": 0.0006,
+ "loss": 5.142467498779297,
+ "step": 735
+ },
+ {
+ "epoch": 10.2235807860262,
+ "grad_norm": 0.034200169146060944,
+ "learning_rate": 0.0006,
+ "loss": 5.072312355041504,
+ "step": 736
+ },
+ {
+ "epoch": 10.237554585152838,
+ "grad_norm": 0.033621110022068024,
+ "learning_rate": 0.0006,
+ "loss": 5.08037805557251,
+ "step": 737
+ },
+ {
+ "epoch": 10.251528384279476,
+ "grad_norm": 0.037372369319200516,
+ "learning_rate": 0.0006,
+ "loss": 4.998425483703613,
+ "step": 738
+ },
+ {
+ "epoch": 10.265502183406113,
+ "grad_norm": 0.037044707685709,
+ "learning_rate": 0.0006,
+ "loss": 4.923993110656738,
+ "step": 739
+ },
+ {
+ "epoch": 10.279475982532752,
+ "grad_norm": 0.030454808846116066,
+ "learning_rate": 0.0006,
+ "loss": 5.0889997482299805,
+ "step": 740
+ },
+ {
+ "epoch": 10.293449781659389,
+ "grad_norm": 0.030969638377428055,
+ "learning_rate": 0.0006,
+ "loss": 4.984112739562988,
+ "step": 741
+ },
+ {
+ "epoch": 10.307423580786025,
+ "grad_norm": 0.0276983380317688,
+ "learning_rate": 0.0006,
+ "loss": 4.955412864685059,
+ "step": 742
+ },
+ {
+ "epoch": 10.321397379912664,
+ "grad_norm": 0.02716052532196045,
+ "learning_rate": 0.0006,
+ "loss": 5.0098876953125,
+ "step": 743
+ },
+ {
+ "epoch": 10.335371179039301,
+ "grad_norm": 0.02403552085161209,
+ "learning_rate": 0.0006,
+ "loss": 5.165160655975342,
+ "step": 744
+ },
+ {
+ "epoch": 10.34934497816594,
+ "grad_norm": 0.0257862601429224,
+ "learning_rate": 0.0006,
+ "loss": 5.001547813415527,
+ "step": 745
+ },
+ {
+ "epoch": 10.363318777292577,
+ "grad_norm": 0.02300378680229187,
+ "learning_rate": 0.0006,
+ "loss": 5.038888931274414,
+ "step": 746
+ },
+ {
+ "epoch": 10.377292576419213,
+ "grad_norm": 0.019766854122281075,
+ "learning_rate": 0.0006,
+ "loss": 4.924291610717773,
+ "step": 747
+ },
+ {
+ "epoch": 10.391266375545852,
+ "grad_norm": 0.01888395845890045,
+ "learning_rate": 0.0006,
+ "loss": 5.059971332550049,
+ "step": 748
+ },
+ {
+ "epoch": 10.405240174672489,
+ "grad_norm": 0.015577499754726887,
+ "learning_rate": 0.0006,
+ "loss": 4.987686634063721,
+ "step": 749
+ },
+ {
+ "epoch": 10.419213973799126,
+ "grad_norm": 0.016012346372008324,
+ "learning_rate": 0.0006,
+ "loss": 5.015157699584961,
+ "step": 750
+ },
+ {
+ "epoch": 10.433187772925764,
+ "grad_norm": 0.016277998685836792,
+ "learning_rate": 0.0006,
+ "loss": 5.004931449890137,
+ "step": 751
+ },
+ {
+ "epoch": 10.447161572052401,
+ "grad_norm": 0.017504561692476273,
+ "learning_rate": 0.0006,
+ "loss": 4.902827739715576,
+ "step": 752
+ },
+ {
+ "epoch": 10.46113537117904,
+ "grad_norm": 0.01569022797048092,
+ "learning_rate": 0.0006,
+ "loss": 4.954863548278809,
+ "step": 753
+ },
+ {
+ "epoch": 10.475109170305677,
+ "grad_norm": 0.014114422723650932,
+ "learning_rate": 0.0006,
+ "loss": 5.00656795501709,
+ "step": 754
+ },
+ {
+ "epoch": 10.489082969432314,
+ "grad_norm": 0.01573832333087921,
+ "learning_rate": 0.0006,
+ "loss": 4.978353023529053,
+ "step": 755
+ },
+ {
+ "epoch": 10.503056768558952,
+ "grad_norm": 0.016480036079883575,
+ "learning_rate": 0.0006,
+ "loss": 4.953327655792236,
+ "step": 756
+ },
+ {
+ "epoch": 10.51703056768559,
+ "grad_norm": 0.014008025638759136,
+ "learning_rate": 0.0006,
+ "loss": 5.050840377807617,
+ "step": 757
+ },
+ {
+ "epoch": 10.531004366812226,
+ "grad_norm": 0.013147998601198196,
+ "learning_rate": 0.0006,
+ "loss": 4.974964141845703,
+ "step": 758
+ },
+ {
+ "epoch": 10.544978165938865,
+ "grad_norm": 0.013807603158056736,
+ "learning_rate": 0.0006,
+ "loss": 4.927907943725586,
+ "step": 759
+ },
+ {
+ "epoch": 10.558951965065502,
+ "grad_norm": 0.013555224984884262,
+ "learning_rate": 0.0006,
+ "loss": 4.980690956115723,
+ "step": 760
+ },
+ {
+ "epoch": 10.57292576419214,
+ "grad_norm": 0.013045408762991428,
+ "learning_rate": 0.0006,
+ "loss": 4.9436116218566895,
+ "step": 761
+ },
+ {
+ "epoch": 10.586899563318777,
+ "grad_norm": 0.012858862057328224,
+ "learning_rate": 0.0006,
+ "loss": 4.9595770835876465,
+ "step": 762
+ },
+ {
+ "epoch": 10.600873362445414,
+ "grad_norm": 0.01468253880739212,
+ "learning_rate": 0.0006,
+ "loss": 4.90725040435791,
+ "step": 763
+ },
+ {
+ "epoch": 10.614847161572053,
+ "grad_norm": 0.013224679045379162,
+ "learning_rate": 0.0006,
+ "loss": 4.877615928649902,
+ "step": 764
+ },
+ {
+ "epoch": 10.62882096069869,
+ "grad_norm": 0.013555348850786686,
+ "learning_rate": 0.0006,
+ "loss": 4.9456610679626465,
+ "step": 765
+ },
+ {
+ "epoch": 10.642794759825328,
+ "grad_norm": 0.014387160539627075,
+ "learning_rate": 0.0006,
+ "loss": 4.9233717918396,
+ "step": 766
+ },
+ {
+ "epoch": 10.656768558951965,
+ "grad_norm": 0.014354088343679905,
+ "learning_rate": 0.0006,
+ "loss": 4.983511924743652,
+ "step": 767
+ },
+ {
+ "epoch": 10.670742358078602,
+ "grad_norm": 0.01332700252532959,
+ "learning_rate": 0.0006,
+ "loss": 4.934416770935059,
+ "step": 768
+ },
+ {
+ "epoch": 10.68471615720524,
+ "grad_norm": 0.011543313041329384,
+ "learning_rate": 0.0006,
+ "loss": 4.96388053894043,
+ "step": 769
+ },
+ {
+ "epoch": 10.698689956331878,
+ "grad_norm": 0.012330878525972366,
+ "learning_rate": 0.0006,
+ "loss": 4.946396350860596,
+ "step": 770
+ },
+ {
+ "epoch": 10.712663755458514,
+ "grad_norm": 0.01379779726266861,
+ "learning_rate": 0.0006,
+ "loss": 4.981019020080566,
+ "step": 771
+ },
+ {
+ "epoch": 10.726637554585153,
+ "grad_norm": 0.014612431637942791,
+ "learning_rate": 0.0006,
+ "loss": 4.894567966461182,
+ "step": 772
+ },
+ {
+ "epoch": 10.74061135371179,
+ "grad_norm": 0.013062899932265282,
+ "learning_rate": 0.0006,
+ "loss": 5.033038139343262,
+ "step": 773
+ },
+ {
+ "epoch": 10.754585152838429,
+ "grad_norm": 0.013898049481213093,
+ "learning_rate": 0.0006,
+ "loss": 4.823636054992676,
+ "step": 774
+ },
+ {
+ "epoch": 10.768558951965066,
+ "grad_norm": 0.01502019353210926,
+ "learning_rate": 0.0006,
+ "loss": 5.020156383514404,
+ "step": 775
+ },
+ {
+ "epoch": 10.782532751091702,
+ "grad_norm": 0.013810204342007637,
+ "learning_rate": 0.0006,
+ "loss": 4.935647010803223,
+ "step": 776
+ },
+ {
+ "epoch": 10.796506550218341,
+ "grad_norm": 0.013616513460874557,
+ "learning_rate": 0.0006,
+ "loss": 4.901486396789551,
+ "step": 777
+ },
+ {
+ "epoch": 10.810480349344978,
+ "grad_norm": 0.013601024635136127,
+ "learning_rate": 0.0006,
+ "loss": 4.931873321533203,
+ "step": 778
+ },
+ {
+ "epoch": 10.824454148471617,
+ "grad_norm": 0.013668350875377655,
+ "learning_rate": 0.0006,
+ "loss": 4.937256813049316,
+ "step": 779
+ },
+ {
+ "epoch": 10.838427947598253,
+ "grad_norm": 0.01581823080778122,
+ "learning_rate": 0.0006,
+ "loss": 4.9611406326293945,
+ "step": 780
+ },
+ {
+ "epoch": 10.85240174672489,
+ "grad_norm": 0.019478484988212585,
+ "learning_rate": 0.0006,
+ "loss": 4.9212236404418945,
+ "step": 781
+ },
+ {
+ "epoch": 10.866375545851529,
+ "grad_norm": 0.026921333745121956,
+ "learning_rate": 0.0006,
+ "loss": 4.88501501083374,
+ "step": 782
+ },
+ {
+ "epoch": 10.880349344978166,
+ "grad_norm": 0.03155914694070816,
+ "learning_rate": 0.0006,
+ "loss": 4.989967346191406,
+ "step": 783
+ },
+ {
+ "epoch": 10.894323144104803,
+ "grad_norm": 0.029689429327845573,
+ "learning_rate": 0.0006,
+ "loss": 4.936000823974609,
+ "step": 784
+ },
+ {
+ "epoch": 10.908296943231441,
+ "grad_norm": 0.026626063510775566,
+ "learning_rate": 0.0006,
+ "loss": 4.984785079956055,
+ "step": 785
+ },
+ {
+ "epoch": 10.922270742358078,
+ "grad_norm": 0.026168212294578552,
+ "learning_rate": 0.0006,
+ "loss": 4.908282279968262,
+ "step": 786
+ },
+ {
+ "epoch": 10.936244541484717,
+ "grad_norm": 0.02828158251941204,
+ "learning_rate": 0.0006,
+ "loss": 4.926390647888184,
+ "step": 787
+ },
+ {
+ "epoch": 10.950218340611354,
+ "grad_norm": 0.02649078331887722,
+ "learning_rate": 0.0006,
+ "loss": 5.00400972366333,
+ "step": 788
+ },
+ {
+ "epoch": 10.96419213973799,
+ "grad_norm": 0.0256856270134449,
+ "learning_rate": 0.0006,
+ "loss": 5.0317487716674805,
+ "step": 789
+ },
+ {
+ "epoch": 10.97816593886463,
+ "grad_norm": 0.024728883057832718,
+ "learning_rate": 0.0006,
+ "loss": 4.9299635887146,
+ "step": 790
+ },
+ {
+ "epoch": 10.992139737991266,
+ "grad_norm": 0.026241116225719452,
+ "learning_rate": 0.0006,
+ "loss": 4.947319984436035,
+ "step": 791
+ },
+ {
+ "epoch": 11.0,
+ "grad_norm": 0.023526743054389954,
+ "learning_rate": 0.0006,
+ "loss": 4.94429349899292,
+ "step": 792
+ },
+ {
+ "epoch": 11.0,
+ "eval_loss": 5.104345798492432,
+ "eval_runtime": 57.0955,
+ "eval_samples_per_second": 42.77,
+ "eval_steps_per_second": 1.349,
+ "step": 792
+ },
+ {
+ "epoch": 11.013973799126637,
+ "grad_norm": 0.0216788612306118,
+ "learning_rate": 0.0006,
+ "loss": 4.881950855255127,
+ "step": 793
+ },
+ {
+ "epoch": 11.027947598253276,
+ "grad_norm": 0.01830463856458664,
+ "learning_rate": 0.0006,
+ "loss": 4.971016883850098,
+ "step": 794
+ },
+ {
+ "epoch": 11.041921397379912,
+ "grad_norm": 0.02204521745443344,
+ "learning_rate": 0.0006,
+ "loss": 4.894623756408691,
+ "step": 795
+ },
+ {
+ "epoch": 11.055895196506551,
+ "grad_norm": 0.0210255216807127,
+ "learning_rate": 0.0006,
+ "loss": 4.834009170532227,
+ "step": 796
+ },
+ {
+ "epoch": 11.069868995633188,
+ "grad_norm": 0.019797448068857193,
+ "learning_rate": 0.0006,
+ "loss": 5.086714744567871,
+ "step": 797
+ },
+ {
+ "epoch": 11.083842794759825,
+ "grad_norm": 0.02036641724407673,
+ "learning_rate": 0.0006,
+ "loss": 4.920056343078613,
+ "step": 798
+ },
+ {
+ "epoch": 11.097816593886463,
+ "grad_norm": 0.01966066285967827,
+ "learning_rate": 0.0006,
+ "loss": 4.879859447479248,
+ "step": 799
+ },
+ {
+ "epoch": 11.1117903930131,
+ "grad_norm": 0.0202149897813797,
+ "learning_rate": 0.0006,
+ "loss": 4.933640480041504,
+ "step": 800
+ },
+ {
+ "epoch": 11.125764192139737,
+ "grad_norm": 0.01893039606511593,
+ "learning_rate": 0.0006,
+ "loss": 4.901970863342285,
+ "step": 801
+ },
+ {
+ "epoch": 11.139737991266376,
+ "grad_norm": 0.016040155664086342,
+ "learning_rate": 0.0006,
+ "loss": 4.905491828918457,
+ "step": 802
+ },
+ {
+ "epoch": 11.153711790393013,
+ "grad_norm": 0.01720350608229637,
+ "learning_rate": 0.0006,
+ "loss": 4.798816680908203,
+ "step": 803
+ },
+ {
+ "epoch": 11.167685589519651,
+ "grad_norm": 0.017002852633595467,
+ "learning_rate": 0.0006,
+ "loss": 4.851859092712402,
+ "step": 804
+ },
+ {
+ "epoch": 11.181659388646288,
+ "grad_norm": 0.015594680793583393,
+ "learning_rate": 0.0006,
+ "loss": 4.90328311920166,
+ "step": 805
+ },
+ {
+ "epoch": 11.195633187772925,
+ "grad_norm": 0.014904462732374668,
+ "learning_rate": 0.0006,
+ "loss": 4.904215335845947,
+ "step": 806
+ },
+ {
+ "epoch": 11.209606986899564,
+ "grad_norm": 0.01366228237748146,
+ "learning_rate": 0.0006,
+ "loss": 4.9813055992126465,
+ "step": 807
+ },
+ {
+ "epoch": 11.2235807860262,
+ "grad_norm": 0.014719638973474503,
+ "learning_rate": 0.0006,
+ "loss": 4.927306175231934,
+ "step": 808
+ },
+ {
+ "epoch": 11.237554585152838,
+ "grad_norm": 0.015016036108136177,
+ "learning_rate": 0.0006,
+ "loss": 4.874215126037598,
+ "step": 809
+ },
+ {
+ "epoch": 11.251528384279476,
+ "grad_norm": 0.014925424009561539,
+ "learning_rate": 0.0006,
+ "loss": 4.838257789611816,
+ "step": 810
+ },
+ {
+ "epoch": 11.265502183406113,
+ "grad_norm": 0.014335823245346546,
+ "learning_rate": 0.0006,
+ "loss": 4.940608024597168,
+ "step": 811
+ },
+ {
+ "epoch": 11.279475982532752,
+ "grad_norm": 0.014987512491643429,
+ "learning_rate": 0.0006,
+ "loss": 4.831531524658203,
+ "step": 812
+ },
+ {
+ "epoch": 11.293449781659389,
+ "grad_norm": 0.016087457537651062,
+ "learning_rate": 0.0006,
+ "loss": 4.863227367401123,
+ "step": 813
+ },
+ {
+ "epoch": 11.307423580786025,
+ "grad_norm": 0.01671300269663334,
+ "learning_rate": 0.0006,
+ "loss": 4.8535661697387695,
+ "step": 814
+ },
+ {
+ "epoch": 11.321397379912664,
+ "grad_norm": 0.015229142270982265,
+ "learning_rate": 0.0006,
+ "loss": 4.948635101318359,
+ "step": 815
+ },
+ {
+ "epoch": 11.335371179039301,
+ "grad_norm": 0.01532587967813015,
+ "learning_rate": 0.0006,
+ "loss": 4.9035797119140625,
+ "step": 816
+ },
+ {
+ "epoch": 11.34934497816594,
+ "grad_norm": 0.015715476125478745,
+ "learning_rate": 0.0006,
+ "loss": 4.894709587097168,
+ "step": 817
+ },
+ {
+ "epoch": 11.363318777292577,
+ "grad_norm": 0.014156977646052837,
+ "learning_rate": 0.0006,
+ "loss": 4.960206985473633,
+ "step": 818
+ },
+ {
+ "epoch": 11.377292576419213,
+ "grad_norm": 0.017076566815376282,
+ "learning_rate": 0.0006,
+ "loss": 5.005850791931152,
+ "step": 819
+ },
+ {
+ "epoch": 11.391266375545852,
+ "grad_norm": 0.0172084029763937,
+ "learning_rate": 0.0006,
+ "loss": 4.903355598449707,
+ "step": 820
+ },
+ {
+ "epoch": 11.405240174672489,
+ "grad_norm": 0.0180258359760046,
+ "learning_rate": 0.0006,
+ "loss": 4.842667579650879,
+ "step": 821
+ },
+ {
+ "epoch": 11.419213973799126,
+ "grad_norm": 0.022130120545625687,
+ "learning_rate": 0.0006,
+ "loss": 4.799970626831055,
+ "step": 822
+ },
+ {
+ "epoch": 11.433187772925764,
+ "grad_norm": 0.02384897693991661,
+ "learning_rate": 0.0006,
+ "loss": 4.860746383666992,
+ "step": 823
+ },
+ {
+ "epoch": 11.447161572052401,
+ "grad_norm": 0.021881572902202606,
+ "learning_rate": 0.0006,
+ "loss": 4.918107032775879,
+ "step": 824
+ },
+ {
+ "epoch": 11.46113537117904,
+ "grad_norm": 0.023085080087184906,
+ "learning_rate": 0.0006,
+ "loss": 4.878946304321289,
+ "step": 825
+ },
+ {
+ "epoch": 11.475109170305677,
+ "grad_norm": 0.02291012555360794,
+ "learning_rate": 0.0006,
+ "loss": 4.958887100219727,
+ "step": 826
+ },
+ {
+ "epoch": 11.489082969432314,
+ "grad_norm": 0.0199968870729208,
+ "learning_rate": 0.0006,
+ "loss": 4.80955696105957,
+ "step": 827
+ },
+ {
+ "epoch": 11.503056768558952,
+ "grad_norm": 0.018702229484915733,
+ "learning_rate": 0.0006,
+ "loss": 5.0276408195495605,
+ "step": 828
+ },
+ {
+ "epoch": 11.51703056768559,
+ "grad_norm": 0.017787311226129532,
+ "learning_rate": 0.0006,
+ "loss": 4.96018123626709,
+ "step": 829
+ },
+ {
+ "epoch": 11.531004366812226,
+ "grad_norm": 0.01704501546919346,
+ "learning_rate": 0.0006,
+ "loss": 4.914553165435791,
+ "step": 830
+ },
+ {
+ "epoch": 11.544978165938865,
+ "grad_norm": 0.017519677057862282,
+ "learning_rate": 0.0006,
+ "loss": 4.798098087310791,
+ "step": 831
+ },
+ {
+ "epoch": 11.558951965065502,
+ "grad_norm": 0.018511991947889328,
+ "learning_rate": 0.0006,
+ "loss": 4.817785263061523,
+ "step": 832
+ },
+ {
+ "epoch": 11.57292576419214,
+ "grad_norm": 0.01876644790172577,
+ "learning_rate": 0.0006,
+ "loss": 4.818995475769043,
+ "step": 833
+ },
+ {
+ "epoch": 11.586899563318777,
+ "grad_norm": 0.01753568835556507,
+ "learning_rate": 0.0006,
+ "loss": 4.811938285827637,
+ "step": 834
+ },
+ {
+ "epoch": 11.600873362445414,
+ "grad_norm": 0.017046531662344933,
+ "learning_rate": 0.0006,
+ "loss": 4.914674282073975,
+ "step": 835
+ },
+ {
+ "epoch": 11.614847161572053,
+ "grad_norm": 0.01802094094455242,
+ "learning_rate": 0.0006,
+ "loss": 4.993284225463867,
+ "step": 836
+ },
+ {
+ "epoch": 11.62882096069869,
+ "grad_norm": 0.019493764266371727,
+ "learning_rate": 0.0006,
+ "loss": 4.895628452301025,
+ "step": 837
+ },
+ {
+ "epoch": 11.642794759825328,
+ "grad_norm": 0.019461210817098618,
+ "learning_rate": 0.0006,
+ "loss": 4.869269371032715,
+ "step": 838
+ },
+ {
+ "epoch": 11.656768558951965,
+ "grad_norm": 0.01981567218899727,
+ "learning_rate": 0.0006,
+ "loss": 4.863052845001221,
+ "step": 839
+ },
+ {
+ "epoch": 11.670742358078602,
+ "grad_norm": 0.019894491881132126,
+ "learning_rate": 0.0006,
+ "loss": 4.8775787353515625,
+ "step": 840
+ },
+ {
+ "epoch": 11.68471615720524,
+ "grad_norm": 0.01875786855816841,
+ "learning_rate": 0.0006,
+ "loss": 4.817763328552246,
+ "step": 841
+ },
+ {
+ "epoch": 11.698689956331878,
+ "grad_norm": 0.020926695317029953,
+ "learning_rate": 0.0006,
+ "loss": 4.7470526695251465,
+ "step": 842
+ },
+ {
+ "epoch": 11.712663755458514,
+ "grad_norm": 0.021477103233337402,
+ "learning_rate": 0.0006,
+ "loss": 4.762682914733887,
+ "step": 843
+ },
+ {
+ "epoch": 11.726637554585153,
+ "grad_norm": 0.020628711208701134,
+ "learning_rate": 0.0006,
+ "loss": 4.807985782623291,
+ "step": 844
+ },
+ {
+ "epoch": 11.74061135371179,
+ "grad_norm": 0.021489612758159637,
+ "learning_rate": 0.0006,
+ "loss": 4.843091011047363,
+ "step": 845
+ },
+ {
+ "epoch": 11.754585152838429,
+ "grad_norm": 0.021194491535425186,
+ "learning_rate": 0.0006,
+ "loss": 4.8300251960754395,
+ "step": 846
+ },
+ {
+ "epoch": 11.768558951965066,
+ "grad_norm": 0.01833350397646427,
+ "learning_rate": 0.0006,
+ "loss": 5.000202178955078,
+ "step": 847
+ },
+ {
+ "epoch": 11.782532751091702,
+ "grad_norm": 0.018096931278705597,
+ "learning_rate": 0.0006,
+ "loss": 4.880648136138916,
+ "step": 848
+ },
+ {
+ "epoch": 11.796506550218341,
+ "grad_norm": 0.0197161752730608,
+ "learning_rate": 0.0006,
+ "loss": 4.788316249847412,
+ "step": 849
+ },
+ {
+ "epoch": 11.810480349344978,
+ "grad_norm": 0.0203181654214859,
+ "learning_rate": 0.0006,
+ "loss": 4.870944976806641,
+ "step": 850
+ },
+ {
+ "epoch": 11.824454148471617,
+ "grad_norm": 0.01908455230295658,
+ "learning_rate": 0.0006,
+ "loss": 4.890144348144531,
+ "step": 851
+ },
+ {
+ "epoch": 11.838427947598253,
+ "grad_norm": 0.021281801164150238,
+ "learning_rate": 0.0006,
+ "loss": 4.89578914642334,
+ "step": 852
+ },
+ {
+ "epoch": 11.85240174672489,
+ "grad_norm": 0.02124273031949997,
+ "learning_rate": 0.0006,
+ "loss": 4.883749008178711,
+ "step": 853
+ },
+ {
+ "epoch": 11.866375545851529,
+ "grad_norm": 0.020969398319721222,
+ "learning_rate": 0.0006,
+ "loss": 4.809142589569092,
+ "step": 854
+ },
+ {
+ "epoch": 11.880349344978166,
+ "grad_norm": 0.01967657171189785,
+ "learning_rate": 0.0006,
+ "loss": 4.821864604949951,
+ "step": 855
+ },
+ {
+ "epoch": 11.894323144104803,
+ "grad_norm": 0.020887333899736404,
+ "learning_rate": 0.0006,
+ "loss": 4.917985916137695,
+ "step": 856
+ },
+ {
+ "epoch": 11.908296943231441,
+ "grad_norm": 0.019203083589673042,
+ "learning_rate": 0.0006,
+ "loss": 4.824960708618164,
+ "step": 857
+ },
+ {
+ "epoch": 11.922270742358078,
+ "grad_norm": 0.020162401720881462,
+ "learning_rate": 0.0006,
+ "loss": 4.772617816925049,
+ "step": 858
+ },
+ {
+ "epoch": 11.936244541484717,
+ "grad_norm": 0.02251487225294113,
+ "learning_rate": 0.0006,
+ "loss": 4.735418319702148,
+ "step": 859
+ },
+ {
+ "epoch": 11.950218340611354,
+ "grad_norm": 0.020313527435064316,
+ "learning_rate": 0.0006,
+ "loss": 4.939233779907227,
+ "step": 860
+ },
+ {
+ "epoch": 11.96419213973799,
+ "grad_norm": 0.021635599434375763,
+ "learning_rate": 0.0006,
+ "loss": 4.861807346343994,
+ "step": 861
+ },
+ {
+ "epoch": 11.97816593886463,
+ "grad_norm": 0.02285311557352543,
+ "learning_rate": 0.0006,
+ "loss": 4.932732582092285,
+ "step": 862
+ },
+ {
+ "epoch": 11.992139737991266,
+ "grad_norm": 0.02072535641491413,
+ "learning_rate": 0.0006,
+ "loss": 4.796767711639404,
+ "step": 863
+ },
+ {
+ "epoch": 12.0,
+ "grad_norm": 0.020588304847478867,
+ "learning_rate": 0.0006,
+ "loss": 4.997537612915039,
+ "step": 864
+ },
+ {
+ "epoch": 12.0,
+ "eval_loss": 5.044651985168457,
+ "eval_runtime": 56.9145,
+ "eval_samples_per_second": 42.906,
+ "eval_steps_per_second": 1.353,
+ "step": 864
+ },
+ {
+ "epoch": 12.013973799126637,
+ "grad_norm": 0.017597166821360588,
+ "learning_rate": 0.0006,
+ "loss": 4.830524444580078,
+ "step": 865
+ },
+ {
+ "epoch": 12.027947598253276,
+ "grad_norm": 0.01716788113117218,
+ "learning_rate": 0.0006,
+ "loss": 4.894325256347656,
+ "step": 866
+ },
+ {
+ "epoch": 12.041921397379912,
+ "grad_norm": 0.017712457105517387,
+ "learning_rate": 0.0006,
+ "loss": 4.868297576904297,
+ "step": 867
+ },
+ {
+ "epoch": 12.055895196506551,
+ "grad_norm": 0.021486390382051468,
+ "learning_rate": 0.0006,
+ "loss": 4.922049522399902,
+ "step": 868
+ },
+ {
+ "epoch": 12.069868995633188,
+ "grad_norm": 0.024560654535889626,
+ "learning_rate": 0.0006,
+ "loss": 4.734241008758545,
+ "step": 869
+ },
+ {
+ "epoch": 12.083842794759825,
+ "grad_norm": 0.025870252400636673,
+ "learning_rate": 0.0006,
+ "loss": 4.868566989898682,
+ "step": 870
+ },
+ {
+ "epoch": 12.097816593886463,
+ "grad_norm": 0.0277389045804739,
+ "learning_rate": 0.0006,
+ "loss": 4.755006790161133,
+ "step": 871
+ },
+ {
+ "epoch": 12.1117903930131,
+ "grad_norm": 0.029100844636559486,
+ "learning_rate": 0.0006,
+ "loss": 4.909667015075684,
+ "step": 872
+ },
+ {
+ "epoch": 12.125764192139737,
+ "grad_norm": 0.031833067536354065,
+ "learning_rate": 0.0006,
+ "loss": 4.858674049377441,
+ "step": 873
+ },
+ {
+ "epoch": 12.139737991266376,
+ "grad_norm": 0.032709237188100815,
+ "learning_rate": 0.0006,
+ "loss": 4.823696136474609,
+ "step": 874
+ },
+ {
+ "epoch": 12.153711790393013,
+ "grad_norm": 0.02995380386710167,
+ "learning_rate": 0.0006,
+ "loss": 4.850235462188721,
+ "step": 875
+ },
+ {
+ "epoch": 12.167685589519651,
+ "grad_norm": 0.034098610281944275,
+ "learning_rate": 0.0006,
+ "loss": 4.8169989585876465,
+ "step": 876
+ },
+ {
+ "epoch": 12.181659388646288,
+ "grad_norm": 0.03377080336213112,
+ "learning_rate": 0.0006,
+ "loss": 4.90738582611084,
+ "step": 877
+ },
+ {
+ "epoch": 12.195633187772925,
+ "grad_norm": 0.029780825600028038,
+ "learning_rate": 0.0006,
+ "loss": 4.811163902282715,
+ "step": 878
+ },
+ {
+ "epoch": 12.209606986899564,
+ "grad_norm": 0.03118024580180645,
+ "learning_rate": 0.0006,
+ "loss": 4.861286640167236,
+ "step": 879
+ },
+ {
+ "epoch": 12.2235807860262,
+ "grad_norm": 0.033052004873752594,
+ "learning_rate": 0.0006,
+ "loss": 4.865021228790283,
+ "step": 880
+ },
+ {
+ "epoch": 12.237554585152838,
+ "grad_norm": 0.029070032760500908,
+ "learning_rate": 0.0006,
+ "loss": 4.947303771972656,
+ "step": 881
+ },
+ {
+ "epoch": 12.251528384279476,
+ "grad_norm": 0.028355715796351433,
+ "learning_rate": 0.0006,
+ "loss": 4.804648399353027,
+ "step": 882
+ },
+ {
+ "epoch": 12.265502183406113,
+ "grad_norm": 0.031257227063179016,
+ "learning_rate": 0.0006,
+ "loss": 4.757718086242676,
+ "step": 883
+ },
+ {
+ "epoch": 12.279475982532752,
+ "grad_norm": 0.031438373029232025,
+ "learning_rate": 0.0006,
+ "loss": 4.851268768310547,
+ "step": 884
+ },
+ {
+ "epoch": 12.293449781659389,
+ "grad_norm": 0.030492138117551804,
+ "learning_rate": 0.0006,
+ "loss": 4.7891035079956055,
+ "step": 885
+ },
+ {
+ "epoch": 12.307423580786025,
+ "grad_norm": 0.026115991175174713,
+ "learning_rate": 0.0006,
+ "loss": 4.86474609375,
+ "step": 886
+ },
+ {
+ "epoch": 12.321397379912664,
+ "grad_norm": 0.024184465408325195,
+ "learning_rate": 0.0006,
+ "loss": 4.860456943511963,
+ "step": 887
+ },
+ {
+ "epoch": 12.335371179039301,
+ "grad_norm": 0.025437982752919197,
+ "learning_rate": 0.0006,
+ "loss": 4.806251049041748,
+ "step": 888
+ },
+ {
+ "epoch": 12.34934497816594,
+ "grad_norm": 0.022390231490135193,
+ "learning_rate": 0.0006,
+ "loss": 4.807086944580078,
+ "step": 889
+ },
+ {
+ "epoch": 12.363318777292577,
+ "grad_norm": 0.020174294710159302,
+ "learning_rate": 0.0006,
+ "loss": 4.78292179107666,
+ "step": 890
+ },
+ {
+ "epoch": 12.377292576419213,
+ "grad_norm": 0.019734324887394905,
+ "learning_rate": 0.0006,
+ "loss": 4.784509658813477,
+ "step": 891
+ },
+ {
+ "epoch": 12.391266375545852,
+ "grad_norm": 0.016733523458242416,
+ "learning_rate": 0.0006,
+ "loss": 4.811755180358887,
+ "step": 892
+ },
+ {
+ "epoch": 12.405240174672489,
+ "grad_norm": 0.01570146717131138,
+ "learning_rate": 0.0006,
+ "loss": 4.720088005065918,
+ "step": 893
+ },
+ {
+ "epoch": 12.419213973799126,
+ "grad_norm": 0.01601393334567547,
+ "learning_rate": 0.0006,
+ "loss": 4.783543109893799,
+ "step": 894
+ },
+ {
+ "epoch": 12.433187772925764,
+ "grad_norm": 0.014035487547516823,
+ "learning_rate": 0.0006,
+ "loss": 4.848340034484863,
+ "step": 895
+ },
+ {
+ "epoch": 12.447161572052401,
+ "grad_norm": 0.015248487703502178,
+ "learning_rate": 0.0006,
+ "loss": 4.773287773132324,
+ "step": 896
+ },
+ {
+ "epoch": 12.46113537117904,
+ "grad_norm": 0.01532268151640892,
+ "learning_rate": 0.0006,
+ "loss": 4.799646854400635,
+ "step": 897
+ },
+ {
+ "epoch": 12.475109170305677,
+ "grad_norm": 0.013982534408569336,
+ "learning_rate": 0.0006,
+ "loss": 4.8090500831604,
+ "step": 898
+ },
+ {
+ "epoch": 12.489082969432314,
+ "grad_norm": 0.014081164263188839,
+ "learning_rate": 0.0006,
+ "loss": 4.751888751983643,
+ "step": 899
+ },
+ {
+ "epoch": 12.503056768558952,
+ "grad_norm": 0.01380133256316185,
+ "learning_rate": 0.0006,
+ "loss": 4.78144645690918,
+ "step": 900
+ },
+ {
+ "epoch": 12.51703056768559,
+ "grad_norm": 0.013555269688367844,
+ "learning_rate": 0.0006,
+ "loss": 4.779257774353027,
+ "step": 901
+ },
+ {
+ "epoch": 12.531004366812226,
+ "grad_norm": 0.013075701892375946,
+ "learning_rate": 0.0006,
+ "loss": 4.820705413818359,
+ "step": 902
+ },
+ {
+ "epoch": 12.544978165938865,
+ "grad_norm": 0.01276316400617361,
+ "learning_rate": 0.0006,
+ "loss": 4.836733818054199,
+ "step": 903
+ },
+ {
+ "epoch": 12.558951965065502,
+ "grad_norm": 0.013033892959356308,
+ "learning_rate": 0.0006,
+ "loss": 4.764178276062012,
+ "step": 904
+ },
+ {
+ "epoch": 12.57292576419214,
+ "grad_norm": 0.013996980153024197,
+ "learning_rate": 0.0006,
+ "loss": 4.767983436584473,
+ "step": 905
+ },
+ {
+ "epoch": 12.586899563318777,
+ "grad_norm": 0.015222184360027313,
+ "learning_rate": 0.0006,
+ "loss": 4.759721755981445,
+ "step": 906
+ },
+ {
+ "epoch": 12.600873362445414,
+ "grad_norm": 0.01623941957950592,
+ "learning_rate": 0.0006,
+ "loss": 4.849811553955078,
+ "step": 907
+ },
+ {
+ "epoch": 12.614847161572053,
+ "grad_norm": 0.015711182728409767,
+ "learning_rate": 0.0006,
+ "loss": 4.7829084396362305,
+ "step": 908
+ },
+ {
+ "epoch": 12.62882096069869,
+ "grad_norm": 0.014137467369437218,
+ "learning_rate": 0.0006,
+ "loss": 4.73321533203125,
+ "step": 909
+ },
+ {
+ "epoch": 12.642794759825328,
+ "grad_norm": 0.012738462537527084,
+ "learning_rate": 0.0006,
+ "loss": 4.870052337646484,
+ "step": 910
+ },
+ {
+ "epoch": 12.656768558951965,
+ "grad_norm": 0.013683130033314228,
+ "learning_rate": 0.0006,
+ "loss": 4.82658052444458,
+ "step": 911
+ },
+ {
+ "epoch": 12.670742358078602,
+ "grad_norm": 0.014409830793738365,
+ "learning_rate": 0.0006,
+ "loss": 4.707494735717773,
+ "step": 912
+ },
+ {
+ "epoch": 12.68471615720524,
+ "grad_norm": 0.013806809671223164,
+ "learning_rate": 0.0006,
+ "loss": 4.778928279876709,
+ "step": 913
+ },
+ {
+ "epoch": 12.698689956331878,
+ "grad_norm": 0.012857912108302116,
+ "learning_rate": 0.0006,
+ "loss": 4.890717506408691,
+ "step": 914
+ },
+ {
+ "epoch": 12.712663755458514,
+ "grad_norm": 0.013828758150339127,
+ "learning_rate": 0.0006,
+ "loss": 4.8357439041137695,
+ "step": 915
+ },
+ {
+ "epoch": 12.726637554585153,
+ "grad_norm": 0.01478351280093193,
+ "learning_rate": 0.0006,
+ "loss": 4.869986534118652,
+ "step": 916
+ },
+ {
+ "epoch": 12.74061135371179,
+ "grad_norm": 0.01326421182602644,
+ "learning_rate": 0.0006,
+ "loss": 4.830646514892578,
+ "step": 917
+ },
+ {
+ "epoch": 12.754585152838429,
+ "grad_norm": 0.013335433788597584,
+ "learning_rate": 0.0006,
+ "loss": 4.7786149978637695,
+ "step": 918
+ },
+ {
+ "epoch": 12.768558951965066,
+ "grad_norm": 0.014261147007346153,
+ "learning_rate": 0.0006,
+ "loss": 4.788693428039551,
+ "step": 919
+ },
+ {
+ "epoch": 12.782532751091702,
+ "grad_norm": 0.014274783432483673,
+ "learning_rate": 0.0006,
+ "loss": 4.77295446395874,
+ "step": 920
+ },
+ {
+ "epoch": 12.796506550218341,
+ "grad_norm": 0.015074674971401691,
+ "learning_rate": 0.0006,
+ "loss": 4.830754280090332,
+ "step": 921
+ },
+ {
+ "epoch": 12.810480349344978,
+ "grad_norm": 0.015599515289068222,
+ "learning_rate": 0.0006,
+ "loss": 4.837759017944336,
+ "step": 922
+ },
+ {
+ "epoch": 12.824454148471617,
+ "grad_norm": 0.014595242217183113,
+ "learning_rate": 0.0006,
+ "loss": 4.882965564727783,
+ "step": 923
+ },
+ {
+ "epoch": 12.838427947598253,
+ "grad_norm": 0.014951467514038086,
+ "learning_rate": 0.0006,
+ "loss": 4.709748268127441,
+ "step": 924
+ },
+ {
+ "epoch": 12.85240174672489,
+ "grad_norm": 0.01539295818656683,
+ "learning_rate": 0.0006,
+ "loss": 4.791160583496094,
+ "step": 925
+ },
+ {
+ "epoch": 12.866375545851529,
+ "grad_norm": 0.015987906605005264,
+ "learning_rate": 0.0006,
+ "loss": 4.661876678466797,
+ "step": 926
+ },
+ {
+ "epoch": 12.880349344978166,
+ "grad_norm": 0.016527341678738594,
+ "learning_rate": 0.0006,
+ "loss": 4.721952438354492,
+ "step": 927
+ },
+ {
+ "epoch": 12.894323144104803,
+ "grad_norm": 0.01637161523103714,
+ "learning_rate": 0.0006,
+ "loss": 4.704446792602539,
+ "step": 928
+ },
+ {
+ "epoch": 12.908296943231441,
+ "grad_norm": 0.015221050009131432,
+ "learning_rate": 0.0006,
+ "loss": 4.771907806396484,
+ "step": 929
+ },
+ {
+ "epoch": 12.922270742358078,
+ "grad_norm": 0.016242166981101036,
+ "learning_rate": 0.0006,
+ "loss": 4.806267738342285,
+ "step": 930
+ },
+ {
+ "epoch": 12.936244541484717,
+ "grad_norm": 0.01628357358276844,
+ "learning_rate": 0.0006,
+ "loss": 4.7252326011657715,
+ "step": 931
+ },
+ {
+ "epoch": 12.950218340611354,
+ "grad_norm": 0.017428133636713028,
+ "learning_rate": 0.0006,
+ "loss": 4.765689849853516,
+ "step": 932
+ },
+ {
+ "epoch": 12.96419213973799,
+ "grad_norm": 0.01870061084628105,
+ "learning_rate": 0.0006,
+ "loss": 4.814488887786865,
+ "step": 933
+ },
+ {
+ "epoch": 12.97816593886463,
+ "grad_norm": 0.019796060398221016,
+ "learning_rate": 0.0006,
+ "loss": 4.739185810089111,
+ "step": 934
+ },
+ {
+ "epoch": 12.992139737991266,
+ "grad_norm": 0.01985820196568966,
+ "learning_rate": 0.0006,
+ "loss": 4.810608863830566,
+ "step": 935
+ },
+ {
+ "epoch": 13.0,
+ "grad_norm": 0.020834477618336678,
+ "learning_rate": 0.0006,
+ "loss": 4.727858066558838,
+ "step": 936
+ },
+ {
+ "epoch": 13.0,
+ "eval_loss": 4.979116916656494,
+ "eval_runtime": 57.2877,
+ "eval_samples_per_second": 42.627,
+ "eval_steps_per_second": 1.344,
+ "step": 936
+ },
+ {
+ "epoch": 13.013973799126637,
+ "grad_norm": 0.02025519125163555,
+ "learning_rate": 0.0006,
+ "loss": 4.628875732421875,
+ "step": 937
+ },
+ {
+ "epoch": 13.027947598253276,
+ "grad_norm": 0.019596368074417114,
+ "learning_rate": 0.0006,
+ "loss": 4.694699287414551,
+ "step": 938
+ },
+ {
+ "epoch": 13.041921397379912,
+ "grad_norm": 0.020494934171438217,
+ "learning_rate": 0.0006,
+ "loss": 4.691821098327637,
+ "step": 939
+ },
+ {
+ "epoch": 13.055895196506551,
+ "grad_norm": 0.020121760666370392,
+ "learning_rate": 0.0006,
+ "loss": 4.71478271484375,
+ "step": 940
+ },
+ {
+ "epoch": 13.069868995633188,
+ "grad_norm": 0.021091405302286148,
+ "learning_rate": 0.0006,
+ "loss": 4.695915222167969,
+ "step": 941
+ },
+ {
+ "epoch": 13.083842794759825,
+ "grad_norm": 0.019787423312664032,
+ "learning_rate": 0.0006,
+ "loss": 4.8652448654174805,
+ "step": 942
+ },
+ {
+ "epoch": 13.097816593886463,
+ "grad_norm": 0.0185833927243948,
+ "learning_rate": 0.0006,
+ "loss": 4.7065887451171875,
+ "step": 943
+ },
+ {
+ "epoch": 13.1117903930131,
+ "grad_norm": 0.021220475435256958,
+ "learning_rate": 0.0006,
+ "loss": 4.741232395172119,
+ "step": 944
+ },
+ {
+ "epoch": 13.125764192139737,
+ "grad_norm": 0.02136622555553913,
+ "learning_rate": 0.0006,
+ "loss": 4.705967903137207,
+ "step": 945
+ },
+ {
+ "epoch": 13.139737991266376,
+ "grad_norm": 0.0191587433218956,
+ "learning_rate": 0.0006,
+ "loss": 4.758667945861816,
+ "step": 946
+ },
+ {
+ "epoch": 13.153711790393013,
+ "grad_norm": 0.018290022388100624,
+ "learning_rate": 0.0006,
+ "loss": 4.760666847229004,
+ "step": 947
+ },
+ {
+ "epoch": 13.167685589519651,
+ "grad_norm": 0.021704409271478653,
+ "learning_rate": 0.0006,
+ "loss": 4.750555992126465,
+ "step": 948
+ },
+ {
+ "epoch": 13.181659388646288,
+ "grad_norm": 0.021668918430805206,
+ "learning_rate": 0.0006,
+ "loss": 4.840903282165527,
+ "step": 949
+ },
+ {
+ "epoch": 13.195633187772925,
+ "grad_norm": 0.019497141242027283,
+ "learning_rate": 0.0006,
+ "loss": 4.738969326019287,
+ "step": 950
+ },
+ {
+ "epoch": 13.209606986899564,
+ "grad_norm": 0.018500596284866333,
+ "learning_rate": 0.0006,
+ "loss": 4.75154972076416,
+ "step": 951
+ },
+ {
+ "epoch": 13.2235807860262,
+ "grad_norm": 0.01872239261865616,
+ "learning_rate": 0.0006,
+ "loss": 4.811738014221191,
+ "step": 952
+ },
+ {
+ "epoch": 13.237554585152838,
+ "grad_norm": 0.020321547985076904,
+ "learning_rate": 0.0006,
+ "loss": 4.815384387969971,
+ "step": 953
+ },
+ {
+ "epoch": 13.251528384279476,
+ "grad_norm": 0.02283575013279915,
+ "learning_rate": 0.0006,
+ "loss": 4.74582052230835,
+ "step": 954
+ },
+ {
+ "epoch": 13.265502183406113,
+ "grad_norm": 0.02474398910999298,
+ "learning_rate": 0.0006,
+ "loss": 4.738375186920166,
+ "step": 955
+ },
+ {
+ "epoch": 13.279475982532752,
+ "grad_norm": 0.02065764181315899,
+ "learning_rate": 0.0006,
+ "loss": 4.746875286102295,
+ "step": 956
+ },
+ {
+ "epoch": 13.293449781659389,
+ "grad_norm": 0.020534520968794823,
+ "learning_rate": 0.0006,
+ "loss": 4.656914710998535,
+ "step": 957
+ },
+ {
+ "epoch": 13.307423580786025,
+ "grad_norm": 0.0206378772854805,
+ "learning_rate": 0.0006,
+ "loss": 4.842615604400635,
+ "step": 958
+ },
+ {
+ "epoch": 13.321397379912664,
+ "grad_norm": 0.017935654148459435,
+ "learning_rate": 0.0006,
+ "loss": 4.717827796936035,
+ "step": 959
+ },
+ {
+ "epoch": 13.335371179039301,
+ "grad_norm": 0.016958363354206085,
+ "learning_rate": 0.0006,
+ "loss": 4.681070327758789,
+ "step": 960
+ },
+ {
+ "epoch": 13.34934497816594,
+ "grad_norm": 0.015735412016510963,
+ "learning_rate": 0.0006,
+ "loss": 4.7957048416137695,
+ "step": 961
+ },
+ {
+ "epoch": 13.363318777292577,
+ "grad_norm": 0.01563699170947075,
+ "learning_rate": 0.0006,
+ "loss": 4.701192378997803,
+ "step": 962
+ },
+ {
+ "epoch": 13.377292576419213,
+ "grad_norm": 0.015207760035991669,
+ "learning_rate": 0.0006,
+ "loss": 4.73256254196167,
+ "step": 963
+ },
+ {
+ "epoch": 13.391266375545852,
+ "grad_norm": 0.015396883711218834,
+ "learning_rate": 0.0006,
+ "loss": 4.76412296295166,
+ "step": 964
+ },
+ {
+ "epoch": 13.405240174672489,
+ "grad_norm": 0.015690583735704422,
+ "learning_rate": 0.0006,
+ "loss": 4.791157245635986,
+ "step": 965
+ },
+ {
+ "epoch": 13.419213973799126,
+ "grad_norm": 0.01755724661052227,
+ "learning_rate": 0.0006,
+ "loss": 4.845666885375977,
+ "step": 966
+ },
+ {
+ "epoch": 13.433187772925764,
+ "grad_norm": 0.019443422555923462,
+ "learning_rate": 0.0006,
+ "loss": 4.5902299880981445,
+ "step": 967
+ },
+ {
+ "epoch": 13.447161572052401,
+ "grad_norm": 0.02118338644504547,
+ "learning_rate": 0.0006,
+ "loss": 4.740627288818359,
+ "step": 968
+ },
+ {
+ "epoch": 13.46113537117904,
+ "grad_norm": 0.02174839936196804,
+ "learning_rate": 0.0006,
+ "loss": 4.814560890197754,
+ "step": 969
+ },
+ {
+ "epoch": 13.475109170305677,
+ "grad_norm": 0.022860374301671982,
+ "learning_rate": 0.0006,
+ "loss": 4.66440486907959,
+ "step": 970
+ },
+ {
+ "epoch": 13.489082969432314,
+ "grad_norm": 0.024742677807807922,
+ "learning_rate": 0.0006,
+ "loss": 4.7451653480529785,
+ "step": 971
+ },
+ {
+ "epoch": 13.503056768558952,
+ "grad_norm": 0.027068907395005226,
+ "learning_rate": 0.0006,
+ "loss": 4.802584171295166,
+ "step": 972
+ },
+ {
+ "epoch": 13.51703056768559,
+ "grad_norm": 0.026316625997424126,
+ "learning_rate": 0.0006,
+ "loss": 4.842397689819336,
+ "step": 973
+ },
+ {
+ "epoch": 13.531004366812226,
+ "grad_norm": 0.024610962718725204,
+ "learning_rate": 0.0006,
+ "loss": 4.650721073150635,
+ "step": 974
+ },
+ {
+ "epoch": 13.544978165938865,
+ "grad_norm": 0.024122603237628937,
+ "learning_rate": 0.0006,
+ "loss": 4.6984357833862305,
+ "step": 975
+ },
+ {
+ "epoch": 13.558951965065502,
+ "grad_norm": 0.022699708119034767,
+ "learning_rate": 0.0006,
+ "loss": 4.793007850646973,
+ "step": 976
+ },
+ {
+ "epoch": 13.57292576419214,
+ "grad_norm": 0.021706534549593925,
+ "learning_rate": 0.0006,
+ "loss": 4.691362380981445,
+ "step": 977
+ },
+ {
+ "epoch": 13.586899563318777,
+ "grad_norm": 0.021363291889429092,
+ "learning_rate": 0.0006,
+ "loss": 4.76762580871582,
+ "step": 978
+ },
+ {
+ "epoch": 13.600873362445414,
+ "grad_norm": 0.023476677015423775,
+ "learning_rate": 0.0006,
+ "loss": 4.777202606201172,
+ "step": 979
+ },
+ {
+ "epoch": 13.614847161572053,
+ "grad_norm": 0.02457100711762905,
+ "learning_rate": 0.0006,
+ "loss": 4.590703010559082,
+ "step": 980
+ },
+ {
+ "epoch": 13.62882096069869,
+ "grad_norm": 0.02469099499285221,
+ "learning_rate": 0.0006,
+ "loss": 4.640185356140137,
+ "step": 981
+ },
+ {
+ "epoch": 13.642794759825328,
+ "grad_norm": 0.024914603680372238,
+ "learning_rate": 0.0006,
+ "loss": 4.772201061248779,
+ "step": 982
+ },
+ {
+ "epoch": 13.656768558951965,
+ "grad_norm": 0.025133173912763596,
+ "learning_rate": 0.0006,
+ "loss": 4.743337154388428,
+ "step": 983
+ },
+ {
+ "epoch": 13.670742358078602,
+ "grad_norm": 0.02034137211740017,
+ "learning_rate": 0.0006,
+ "loss": 4.752127647399902,
+ "step": 984
+ },
+ {
+ "epoch": 13.68471615720524,
+ "grad_norm": 0.019645661115646362,
+ "learning_rate": 0.0006,
+ "loss": 4.838453769683838,
+ "step": 985
+ },
+ {
+ "epoch": 13.698689956331878,
+ "grad_norm": 0.01900297962129116,
+ "learning_rate": 0.0006,
+ "loss": 4.804868221282959,
+ "step": 986
+ },
+ {
+ "epoch": 13.712663755458514,
+ "grad_norm": 0.020292341709136963,
+ "learning_rate": 0.0006,
+ "loss": 4.609743118286133,
+ "step": 987
+ },
+ {
+ "epoch": 13.726637554585153,
+ "grad_norm": 0.017014775425195694,
+ "learning_rate": 0.0006,
+ "loss": 4.740612983703613,
+ "step": 988
+ },
+ {
+ "epoch": 13.74061135371179,
+ "grad_norm": 0.015199844725430012,
+ "learning_rate": 0.0006,
+ "loss": 4.719701766967773,
+ "step": 989
+ },
+ {
+ "epoch": 13.754585152838429,
+ "grad_norm": 0.015125414356589317,
+ "learning_rate": 0.0006,
+ "loss": 4.686114311218262,
+ "step": 990
+ },
+ {
+ "epoch": 13.768558951965066,
+ "grad_norm": 0.014779540710151196,
+ "learning_rate": 0.0006,
+ "loss": 4.674347877502441,
+ "step": 991
+ },
+ {
+ "epoch": 13.782532751091702,
+ "grad_norm": 0.015255163423717022,
+ "learning_rate": 0.0006,
+ "loss": 4.6309309005737305,
+ "step": 992
+ },
+ {
+ "epoch": 13.796506550218341,
+ "grad_norm": 0.013692197389900684,
+ "learning_rate": 0.0006,
+ "loss": 4.698497295379639,
+ "step": 993
+ },
+ {
+ "epoch": 13.810480349344978,
+ "grad_norm": 0.014529623091220856,
+ "learning_rate": 0.0006,
+ "loss": 4.628443717956543,
+ "step": 994
+ },
+ {
+ "epoch": 13.824454148471617,
+ "grad_norm": 0.01576659083366394,
+ "learning_rate": 0.0006,
+ "loss": 4.788015842437744,
+ "step": 995
+ },
+ {
+ "epoch": 13.838427947598253,
+ "grad_norm": 0.015834277495741844,
+ "learning_rate": 0.0006,
+ "loss": 4.672703742980957,
+ "step": 996
+ },
+ {
+ "epoch": 13.85240174672489,
+ "grad_norm": 0.015539503656327724,
+ "learning_rate": 0.0006,
+ "loss": 4.731503486633301,
+ "step": 997
+ },
+ {
+ "epoch": 13.866375545851529,
+ "grad_norm": 0.013463746756315231,
+ "learning_rate": 0.0006,
+ "loss": 4.7006330490112305,
+ "step": 998
+ },
+ {
+ "epoch": 13.880349344978166,
+ "grad_norm": 0.016377726569771767,
+ "learning_rate": 0.0006,
+ "loss": 4.636074066162109,
+ "step": 999
+ },
+ {
+ "epoch": 13.894323144104803,
+ "grad_norm": 0.016660034656524658,
+ "learning_rate": 0.0006,
+ "loss": 4.742693901062012,
+ "step": 1000
+ },
+ {
+ "epoch": 13.908296943231441,
+ "grad_norm": 0.014902938157320023,
+ "learning_rate": 0.0006,
+ "loss": 4.769811630249023,
+ "step": 1001
+ },
+ {
+ "epoch": 13.922270742358078,
+ "grad_norm": 0.014883476309478283,
+ "learning_rate": 0.0006,
+ "loss": 4.731159210205078,
+ "step": 1002
+ },
+ {
+ "epoch": 13.936244541484717,
+ "grad_norm": 0.014494217932224274,
+ "learning_rate": 0.0006,
+ "loss": 4.659485816955566,
+ "step": 1003
+ },
+ {
+ "epoch": 13.950218340611354,
+ "grad_norm": 0.015426991507411003,
+ "learning_rate": 0.0006,
+ "loss": 4.755229949951172,
+ "step": 1004
+ },
+ {
+ "epoch": 13.96419213973799,
+ "grad_norm": 0.016648586839437485,
+ "learning_rate": 0.0006,
+ "loss": 4.74111270904541,
+ "step": 1005
+ },
+ {
+ "epoch": 13.97816593886463,
+ "grad_norm": 0.016980541869997978,
+ "learning_rate": 0.0006,
+ "loss": 4.74072265625,
+ "step": 1006
+ },
+ {
+ "epoch": 13.992139737991266,
+ "grad_norm": 0.018070152029395103,
+ "learning_rate": 0.0006,
+ "loss": 4.732975006103516,
+ "step": 1007
+ },
+ {
+ "epoch": 14.0,
+ "grad_norm": 0.019910665228962898,
+ "learning_rate": 0.0006,
+ "loss": 4.634697914123535,
+ "step": 1008
+ },
+ {
+ "epoch": 14.0,
+ "eval_loss": 4.850640296936035,
+ "eval_runtime": 57.3039,
+ "eval_samples_per_second": 42.615,
+ "eval_steps_per_second": 1.344,
+ "step": 1008
+ },
+ {
+ "epoch": 14.013973799126637,
+ "grad_norm": 0.01980053074657917,
+ "learning_rate": 0.0006,
+ "loss": 4.628744602203369,
+ "step": 1009
+ },
+ {
+ "epoch": 14.027947598253276,
+ "grad_norm": 0.02497623674571514,
+ "learning_rate": 0.0006,
+ "loss": 4.75930118560791,
+ "step": 1010
+ },
+ {
+ "epoch": 14.041921397379912,
+ "grad_norm": 0.02794097363948822,
+ "learning_rate": 0.0006,
+ "loss": 4.691340446472168,
+ "step": 1011
+ },
+ {
+ "epoch": 14.055895196506551,
+ "grad_norm": 0.023725535720586777,
+ "learning_rate": 0.0006,
+ "loss": 4.733664512634277,
+ "step": 1012
+ },
+ {
+ "epoch": 14.069868995633188,
+ "grad_norm": 0.022966474294662476,
+ "learning_rate": 0.0006,
+ "loss": 4.664808750152588,
+ "step": 1013
+ },
+ {
+ "epoch": 14.083842794759825,
+ "grad_norm": 0.02530563250184059,
+ "learning_rate": 0.0006,
+ "loss": 4.8047285079956055,
+ "step": 1014
+ },
+ {
+ "epoch": 14.097816593886463,
+ "grad_norm": 0.02491082064807415,
+ "learning_rate": 0.0006,
+ "loss": 4.667956352233887,
+ "step": 1015
+ },
+ {
+ "epoch": 14.1117903930131,
+ "grad_norm": 0.022110294550657272,
+ "learning_rate": 0.0006,
+ "loss": 4.666140556335449,
+ "step": 1016
+ },
+ {
+ "epoch": 14.125764192139737,
+ "grad_norm": 0.02065849117934704,
+ "learning_rate": 0.0006,
+ "loss": 4.581960678100586,
+ "step": 1017
+ },
+ {
+ "epoch": 14.139737991266376,
+ "grad_norm": 0.01941298507153988,
+ "learning_rate": 0.0006,
+ "loss": 4.614226341247559,
+ "step": 1018
+ },
+ {
+ "epoch": 14.153711790393013,
+ "grad_norm": 0.022441856563091278,
+ "learning_rate": 0.0006,
+ "loss": 4.638352394104004,
+ "step": 1019
+ },
+ {
+ "epoch": 14.167685589519651,
+ "grad_norm": 0.021519694477319717,
+ "learning_rate": 0.0006,
+ "loss": 4.700160980224609,
+ "step": 1020
+ },
+ {
+ "epoch": 14.181659388646288,
+ "grad_norm": 0.018956486135721207,
+ "learning_rate": 0.0006,
+ "loss": 4.649593353271484,
+ "step": 1021
+ },
+ {
+ "epoch": 14.195633187772925,
+ "grad_norm": 0.016681140288710594,
+ "learning_rate": 0.0006,
+ "loss": 4.603952407836914,
+ "step": 1022
+ },
+ {
+ "epoch": 14.209606986899564,
+ "grad_norm": 0.01788809522986412,
+ "learning_rate": 0.0006,
+ "loss": 4.669160842895508,
+ "step": 1023
+ },
+ {
+ "epoch": 14.2235807860262,
+ "grad_norm": 0.016047228127717972,
+ "learning_rate": 0.0006,
+ "loss": 4.6367950439453125,
+ "step": 1024
+ },
+ {
+ "epoch": 14.237554585152838,
+ "grad_norm": 0.01614239625632763,
+ "learning_rate": 0.0006,
+ "loss": 4.664529323577881,
+ "step": 1025
+ },
+ {
+ "epoch": 14.251528384279476,
+ "grad_norm": 0.016362549737095833,
+ "learning_rate": 0.0006,
+ "loss": 4.702373504638672,
+ "step": 1026
+ },
+ {
+ "epoch": 14.265502183406113,
+ "grad_norm": 0.015800409018993378,
+ "learning_rate": 0.0006,
+ "loss": 4.7594428062438965,
+ "step": 1027
+ },
+ {
+ "epoch": 14.279475982532752,
+ "grad_norm": 0.016730139032006264,
+ "learning_rate": 0.0006,
+ "loss": 4.677244186401367,
+ "step": 1028
+ },
+ {
+ "epoch": 14.293449781659389,
+ "grad_norm": 0.01661759801208973,
+ "learning_rate": 0.0006,
+ "loss": 4.633679389953613,
+ "step": 1029
+ },
+ {
+ "epoch": 14.307423580786025,
+ "grad_norm": 0.017673347145318985,
+ "learning_rate": 0.0006,
+ "loss": 4.649107456207275,
+ "step": 1030
+ },
+ {
+ "epoch": 14.321397379912664,
+ "grad_norm": 0.019243352115154266,
+ "learning_rate": 0.0006,
+ "loss": 4.611257553100586,
+ "step": 1031
+ },
+ {
+ "epoch": 14.335371179039301,
+ "grad_norm": 0.019817164167761803,
+ "learning_rate": 0.0006,
+ "loss": 4.588086128234863,
+ "step": 1032
+ },
+ {
+ "epoch": 14.34934497816594,
+ "grad_norm": 0.022292211651802063,
+ "learning_rate": 0.0006,
+ "loss": 4.783066749572754,
+ "step": 1033
+ },
+ {
+ "epoch": 14.363318777292577,
+ "grad_norm": 0.02140096202492714,
+ "learning_rate": 0.0006,
+ "loss": 4.604718208312988,
+ "step": 1034
+ },
+ {
+ "epoch": 14.377292576419213,
+ "grad_norm": 0.0187817495316267,
+ "learning_rate": 0.0006,
+ "loss": 4.721773147583008,
+ "step": 1035
+ },
+ {
+ "epoch": 14.391266375545852,
+ "grad_norm": 0.019879210740327835,
+ "learning_rate": 0.0006,
+ "loss": 4.789286136627197,
+ "step": 1036
+ },
+ {
+ "epoch": 14.405240174672489,
+ "grad_norm": 0.01885785162448883,
+ "learning_rate": 0.0006,
+ "loss": 4.669743537902832,
+ "step": 1037
+ },
+ {
+ "epoch": 14.419213973799126,
+ "grad_norm": 0.018324270844459534,
+ "learning_rate": 0.0006,
+ "loss": 4.65901517868042,
+ "step": 1038
+ },
+ {
+ "epoch": 14.433187772925764,
+ "grad_norm": 0.017664821818470955,
+ "learning_rate": 0.0006,
+ "loss": 4.667839527130127,
+ "step": 1039
+ },
+ {
+ "epoch": 14.447161572052401,
+ "grad_norm": 0.015702905133366585,
+ "learning_rate": 0.0006,
+ "loss": 4.64700984954834,
+ "step": 1040
+ },
+ {
+ "epoch": 14.46113537117904,
+ "grad_norm": 0.016674194484949112,
+ "learning_rate": 0.0006,
+ "loss": 4.694093227386475,
+ "step": 1041
+ },
+ {
+ "epoch": 14.475109170305677,
+ "grad_norm": 0.016869477927684784,
+ "learning_rate": 0.0006,
+ "loss": 4.619544982910156,
+ "step": 1042
+ },
+ {
+ "epoch": 14.489082969432314,
+ "grad_norm": 0.016557540744543076,
+ "learning_rate": 0.0006,
+ "loss": 4.568667411804199,
+ "step": 1043
+ },
+ {
+ "epoch": 14.503056768558952,
+ "grad_norm": 0.014500945806503296,
+ "learning_rate": 0.0006,
+ "loss": 4.561741828918457,
+ "step": 1044
+ },
+ {
+ "epoch": 14.51703056768559,
+ "grad_norm": 0.014228739775717258,
+ "learning_rate": 0.0006,
+ "loss": 4.654277324676514,
+ "step": 1045
+ },
+ {
+ "epoch": 14.531004366812226,
+ "grad_norm": 0.013894238509237766,
+ "learning_rate": 0.0006,
+ "loss": 4.5867719650268555,
+ "step": 1046
+ },
+ {
+ "epoch": 14.544978165938865,
+ "grad_norm": 0.01550068985670805,
+ "learning_rate": 0.0006,
+ "loss": 4.61168098449707,
+ "step": 1047
+ },
+ {
+ "epoch": 14.558951965065502,
+ "grad_norm": 0.015940772369503975,
+ "learning_rate": 0.0006,
+ "loss": 4.691247940063477,
+ "step": 1048
+ },
+ {
+ "epoch": 14.57292576419214,
+ "grad_norm": 0.01714535616338253,
+ "learning_rate": 0.0006,
+ "loss": 4.61183500289917,
+ "step": 1049
+ },
+ {
+ "epoch": 14.586899563318777,
+ "grad_norm": 0.017925038933753967,
+ "learning_rate": 0.0006,
+ "loss": 4.719022750854492,
+ "step": 1050
+ },
+ {
+ "epoch": 14.600873362445414,
+ "grad_norm": 0.018141552805900574,
+ "learning_rate": 0.0006,
+ "loss": 4.760281562805176,
+ "step": 1051
+ },
+ {
+ "epoch": 14.614847161572053,
+ "grad_norm": 0.017131328582763672,
+ "learning_rate": 0.0006,
+ "loss": 4.709680080413818,
+ "step": 1052
+ },
+ {
+ "epoch": 14.62882096069869,
+ "grad_norm": 0.018730048090219498,
+ "learning_rate": 0.0006,
+ "loss": 4.5994720458984375,
+ "step": 1053
+ },
+ {
+ "epoch": 14.642794759825328,
+ "grad_norm": 0.018894298002123833,
+ "learning_rate": 0.0006,
+ "loss": 4.669070720672607,
+ "step": 1054
+ },
+ {
+ "epoch": 14.656768558951965,
+ "grad_norm": 0.017643198370933533,
+ "learning_rate": 0.0006,
+ "loss": 4.641790390014648,
+ "step": 1055
+ },
+ {
+ "epoch": 14.670742358078602,
+ "grad_norm": 0.017647244036197662,
+ "learning_rate": 0.0006,
+ "loss": 4.653899192810059,
+ "step": 1056
+ },
+ {
+ "epoch": 14.68471615720524,
+ "grad_norm": 0.017304198816418648,
+ "learning_rate": 0.0006,
+ "loss": 4.7487473487854,
+ "step": 1057
+ },
+ {
+ "epoch": 14.698689956331878,
+ "grad_norm": 0.01638447679579258,
+ "learning_rate": 0.0006,
+ "loss": 4.71945858001709,
+ "step": 1058
+ },
+ {
+ "epoch": 14.712663755458514,
+ "grad_norm": 0.015892071649432182,
+ "learning_rate": 0.0006,
+ "loss": 4.744903564453125,
+ "step": 1059
+ },
+ {
+ "epoch": 14.726637554585153,
+ "grad_norm": 0.017420614138245583,
+ "learning_rate": 0.0006,
+ "loss": 4.781563758850098,
+ "step": 1060
+ },
+ {
+ "epoch": 14.74061135371179,
+ "grad_norm": 0.016498520970344543,
+ "learning_rate": 0.0006,
+ "loss": 4.622779846191406,
+ "step": 1061
+ },
+ {
+ "epoch": 14.754585152838429,
+ "grad_norm": 0.017482127994298935,
+ "learning_rate": 0.0006,
+ "loss": 4.62825870513916,
+ "step": 1062
+ },
+ {
+ "epoch": 14.768558951965066,
+ "grad_norm": 0.019018379971385002,
+ "learning_rate": 0.0006,
+ "loss": 4.693194389343262,
+ "step": 1063
+ },
+ {
+ "epoch": 14.782532751091702,
+ "grad_norm": 0.01993914321064949,
+ "learning_rate": 0.0006,
+ "loss": 4.742903709411621,
+ "step": 1064
+ },
+ {
+ "epoch": 14.796506550218341,
+ "grad_norm": 0.02301335707306862,
+ "learning_rate": 0.0006,
+ "loss": 4.613251686096191,
+ "step": 1065
+ },
+ {
+ "epoch": 14.810480349344978,
+ "grad_norm": 0.024088222533464432,
+ "learning_rate": 0.0006,
+ "loss": 4.594305992126465,
+ "step": 1066
+ },
+ {
+ "epoch": 14.824454148471617,
+ "grad_norm": 0.02493320405483246,
+ "learning_rate": 0.0006,
+ "loss": 4.700588226318359,
+ "step": 1067
+ },
+ {
+ "epoch": 14.838427947598253,
+ "grad_norm": 0.02501937560737133,
+ "learning_rate": 0.0006,
+ "loss": 4.601742744445801,
+ "step": 1068
+ },
+ {
+ "epoch": 14.85240174672489,
+ "grad_norm": 0.023490751162171364,
+ "learning_rate": 0.0006,
+ "loss": 4.589211463928223,
+ "step": 1069
+ },
+ {
+ "epoch": 14.866375545851529,
+ "grad_norm": 0.019695665687322617,
+ "learning_rate": 0.0006,
+ "loss": 4.635081768035889,
+ "step": 1070
+ },
+ {
+ "epoch": 14.880349344978166,
+ "grad_norm": 0.019276980310678482,
+ "learning_rate": 0.0006,
+ "loss": 4.737002372741699,
+ "step": 1071
+ },
+ {
+ "epoch": 14.894323144104803,
+ "grad_norm": 0.022430511191487312,
+ "learning_rate": 0.0006,
+ "loss": 4.808295726776123,
+ "step": 1072
+ },
+ {
+ "epoch": 14.908296943231441,
+ "grad_norm": 0.021543916314840317,
+ "learning_rate": 0.0006,
+ "loss": 4.559475421905518,
+ "step": 1073
+ },
+ {
+ "epoch": 14.922270742358078,
+ "grad_norm": 0.019885241985321045,
+ "learning_rate": 0.0006,
+ "loss": 4.647948741912842,
+ "step": 1074
+ },
+ {
+ "epoch": 14.936244541484717,
+ "grad_norm": 0.02115175500512123,
+ "learning_rate": 0.0006,
+ "loss": 4.680505752563477,
+ "step": 1075
+ },
+ {
+ "epoch": 14.950218340611354,
+ "grad_norm": 0.02156151458621025,
+ "learning_rate": 0.0006,
+ "loss": 4.673477649688721,
+ "step": 1076
+ },
+ {
+ "epoch": 14.96419213973799,
+ "grad_norm": 0.02184601128101349,
+ "learning_rate": 0.0006,
+ "loss": 4.644402027130127,
+ "step": 1077
+ },
+ {
+ "epoch": 14.97816593886463,
+ "grad_norm": 0.0211983360350132,
+ "learning_rate": 0.0006,
+ "loss": 4.821558475494385,
+ "step": 1078
+ },
+ {
+ "epoch": 14.992139737991266,
+ "grad_norm": 0.019636567682027817,
+ "learning_rate": 0.0006,
+ "loss": 4.578868865966797,
+ "step": 1079
+ },
+ {
+ "epoch": 15.0,
+ "grad_norm": 0.019997350871562958,
+ "learning_rate": 0.0006,
+ "loss": 4.691307544708252,
+ "step": 1080
+ },
+ {
+ "epoch": 15.0,
+ "eval_loss": 4.872478008270264,
+ "eval_runtime": 56.8513,
+ "eval_samples_per_second": 42.954,
+ "eval_steps_per_second": 1.354,
+ "step": 1080
+ },
+ {
+ "epoch": 15.013973799126637,
+ "grad_norm": 0.018830662593245506,
+ "learning_rate": 0.0006,
+ "loss": 4.761779308319092,
+ "step": 1081
+ },
+ {
+ "epoch": 15.027947598253276,
+ "grad_norm": 0.016876481473445892,
+ "learning_rate": 0.0006,
+ "loss": 4.61148738861084,
+ "step": 1082
+ },
+ {
+ "epoch": 15.041921397379912,
+ "grad_norm": 0.015550863929092884,
+ "learning_rate": 0.0006,
+ "loss": 4.676124095916748,
+ "step": 1083
+ },
+ {
+ "epoch": 15.055895196506551,
+ "grad_norm": 0.017178896814584732,
+ "learning_rate": 0.0006,
+ "loss": 4.588935852050781,
+ "step": 1084
+ },
+ {
+ "epoch": 15.069868995633188,
+ "grad_norm": 0.015465447679162025,
+ "learning_rate": 0.0006,
+ "loss": 4.605499267578125,
+ "step": 1085
+ },
+ {
+ "epoch": 15.083842794759825,
+ "grad_norm": 0.01606195978820324,
+ "learning_rate": 0.0006,
+ "loss": 4.723373889923096,
+ "step": 1086
+ },
+ {
+ "epoch": 15.097816593886463,
+ "grad_norm": 0.016521936282515526,
+ "learning_rate": 0.0006,
+ "loss": 4.646608352661133,
+ "step": 1087
+ },
+ {
+ "epoch": 15.1117903930131,
+ "grad_norm": 0.01589205674827099,
+ "learning_rate": 0.0006,
+ "loss": 4.498138427734375,
+ "step": 1088
+ },
+ {
+ "epoch": 15.125764192139737,
+ "grad_norm": 0.01524856686592102,
+ "learning_rate": 0.0006,
+ "loss": 4.594562530517578,
+ "step": 1089
+ },
+ {
+ "epoch": 15.139737991266376,
+ "grad_norm": 0.014714641496539116,
+ "learning_rate": 0.0006,
+ "loss": 4.630744934082031,
+ "step": 1090
+ },
+ {
+ "epoch": 15.153711790393013,
+ "grad_norm": 0.015333331190049648,
+ "learning_rate": 0.0006,
+ "loss": 4.682388782501221,
+ "step": 1091
+ },
+ {
+ "epoch": 15.167685589519651,
+ "grad_norm": 0.01550073828548193,
+ "learning_rate": 0.0006,
+ "loss": 4.692652702331543,
+ "step": 1092
+ },
+ {
+ "epoch": 15.181659388646288,
+ "grad_norm": 0.014783318154513836,
+ "learning_rate": 0.0006,
+ "loss": 4.6223835945129395,
+ "step": 1093
+ },
+ {
+ "epoch": 15.195633187772925,
+ "grad_norm": 0.014372751116752625,
+ "learning_rate": 0.0006,
+ "loss": 4.52687406539917,
+ "step": 1094
+ },
+ {
+ "epoch": 15.209606986899564,
+ "grad_norm": 0.014779130928218365,
+ "learning_rate": 0.0006,
+ "loss": 4.637009620666504,
+ "step": 1095
+ },
+ {
+ "epoch": 15.2235807860262,
+ "grad_norm": 0.015697073191404343,
+ "learning_rate": 0.0006,
+ "loss": 4.682981014251709,
+ "step": 1096
+ },
+ {
+ "epoch": 15.237554585152838,
+ "grad_norm": 0.016523664817214012,
+ "learning_rate": 0.0006,
+ "loss": 4.705722808837891,
+ "step": 1097
+ },
+ {
+ "epoch": 15.251528384279476,
+ "grad_norm": 0.017633996903896332,
+ "learning_rate": 0.0006,
+ "loss": 4.574686050415039,
+ "step": 1098
+ },
+ {
+ "epoch": 15.265502183406113,
+ "grad_norm": 0.019955076277256012,
+ "learning_rate": 0.0006,
+ "loss": 4.634450912475586,
+ "step": 1099
+ },
+ {
+ "epoch": 15.279475982532752,
+ "grad_norm": 0.02132618986070156,
+ "learning_rate": 0.0006,
+ "loss": 4.56110143661499,
+ "step": 1100
+ },
+ {
+ "epoch": 15.293449781659389,
+ "grad_norm": 0.0207593385130167,
+ "learning_rate": 0.0006,
+ "loss": 4.629396438598633,
+ "step": 1101
+ },
+ {
+ "epoch": 15.307423580786025,
+ "grad_norm": 0.017882540822029114,
+ "learning_rate": 0.0006,
+ "loss": 4.730963230133057,
+ "step": 1102
+ },
+ {
+ "epoch": 15.321397379912664,
+ "grad_norm": 0.01917876861989498,
+ "learning_rate": 0.0006,
+ "loss": 4.64585542678833,
+ "step": 1103
+ },
+ {
+ "epoch": 15.335371179039301,
+ "grad_norm": 0.01943974196910858,
+ "learning_rate": 0.0006,
+ "loss": 4.656181335449219,
+ "step": 1104
+ },
+ {
+ "epoch": 15.34934497816594,
+ "grad_norm": 0.019904715940356255,
+ "learning_rate": 0.0006,
+ "loss": 4.619329452514648,
+ "step": 1105
+ },
+ {
+ "epoch": 15.363318777292577,
+ "grad_norm": 0.019174346700310707,
+ "learning_rate": 0.0006,
+ "loss": 4.647892475128174,
+ "step": 1106
+ },
+ {
+ "epoch": 15.377292576419213,
+ "grad_norm": 0.018244758248329163,
+ "learning_rate": 0.0006,
+ "loss": 4.56166934967041,
+ "step": 1107
+ },
+ {
+ "epoch": 15.391266375545852,
+ "grad_norm": 0.018914205953478813,
+ "learning_rate": 0.0006,
+ "loss": 4.648063659667969,
+ "step": 1108
+ },
+ {
+ "epoch": 15.405240174672489,
+ "grad_norm": 0.01961071416735649,
+ "learning_rate": 0.0006,
+ "loss": 4.665525436401367,
+ "step": 1109
+ },
+ {
+ "epoch": 15.419213973799126,
+ "grad_norm": 0.01841311901807785,
+ "learning_rate": 0.0006,
+ "loss": 4.611342430114746,
+ "step": 1110
+ },
+ {
+ "epoch": 15.433187772925764,
+ "grad_norm": 0.01863427273929119,
+ "learning_rate": 0.0006,
+ "loss": 4.664256572723389,
+ "step": 1111
+ },
+ {
+ "epoch": 15.447161572052401,
+ "grad_norm": 0.020506877452135086,
+ "learning_rate": 0.0006,
+ "loss": 4.631237506866455,
+ "step": 1112
+ },
+ {
+ "epoch": 15.46113537117904,
+ "grad_norm": 0.02223833277821541,
+ "learning_rate": 0.0006,
+ "loss": 4.668654441833496,
+ "step": 1113
+ },
+ {
+ "epoch": 15.475109170305677,
+ "grad_norm": 0.023336702957749367,
+ "learning_rate": 0.0006,
+ "loss": 4.562833309173584,
+ "step": 1114
+ },
+ {
+ "epoch": 15.489082969432314,
+ "grad_norm": 0.023052413016557693,
+ "learning_rate": 0.0006,
+ "loss": 4.621532917022705,
+ "step": 1115
+ },
+ {
+ "epoch": 15.503056768558952,
+ "grad_norm": 0.023420924320816994,
+ "learning_rate": 0.0006,
+ "loss": 4.665465354919434,
+ "step": 1116
+ },
+ {
+ "epoch": 15.51703056768559,
+ "grad_norm": 0.024071281775832176,
+ "learning_rate": 0.0006,
+ "loss": 4.595980644226074,
+ "step": 1117
+ },
+ {
+ "epoch": 15.531004366812226,
+ "grad_norm": 0.023406682536005974,
+ "learning_rate": 0.0006,
+ "loss": 4.642635345458984,
+ "step": 1118
+ },
+ {
+ "epoch": 15.544978165938865,
+ "grad_norm": 0.02138647809624672,
+ "learning_rate": 0.0006,
+ "loss": 4.704513072967529,
+ "step": 1119
+ },
+ {
+ "epoch": 15.558951965065502,
+ "grad_norm": 0.0221384447067976,
+ "learning_rate": 0.0006,
+ "loss": 4.567503452301025,
+ "step": 1120
+ },
+ {
+ "epoch": 15.57292576419214,
+ "grad_norm": 0.020879851654171944,
+ "learning_rate": 0.0006,
+ "loss": 4.664057731628418,
+ "step": 1121
+ },
+ {
+ "epoch": 15.586899563318777,
+ "grad_norm": 0.021478967741131783,
+ "learning_rate": 0.0006,
+ "loss": 4.539627552032471,
+ "step": 1122
+ },
+ {
+ "epoch": 15.600873362445414,
+ "grad_norm": 0.022748656570911407,
+ "learning_rate": 0.0006,
+ "loss": 4.5654497146606445,
+ "step": 1123
+ },
+ {
+ "epoch": 15.614847161572053,
+ "grad_norm": 0.021870248019695282,
+ "learning_rate": 0.0006,
+ "loss": 4.56899356842041,
+ "step": 1124
+ },
+ {
+ "epoch": 15.62882096069869,
+ "grad_norm": 0.021711355075240135,
+ "learning_rate": 0.0006,
+ "loss": 4.657808303833008,
+ "step": 1125
+ },
+ {
+ "epoch": 15.642794759825328,
+ "grad_norm": 0.022810909897089005,
+ "learning_rate": 0.0006,
+ "loss": 4.6526689529418945,
+ "step": 1126
+ },
+ {
+ "epoch": 15.656768558951965,
+ "grad_norm": 0.022561343386769295,
+ "learning_rate": 0.0006,
+ "loss": 4.534887790679932,
+ "step": 1127
+ },
+ {
+ "epoch": 15.670742358078602,
+ "grad_norm": 0.019909802824258804,
+ "learning_rate": 0.0006,
+ "loss": 4.600790023803711,
+ "step": 1128
+ },
+ {
+ "epoch": 15.68471615720524,
+ "grad_norm": 0.020092325285077095,
+ "learning_rate": 0.0006,
+ "loss": 4.773830413818359,
+ "step": 1129
+ },
+ {
+ "epoch": 15.698689956331878,
+ "grad_norm": 0.018834445625543594,
+ "learning_rate": 0.0006,
+ "loss": 4.628868103027344,
+ "step": 1130
+ },
+ {
+ "epoch": 15.712663755458514,
+ "grad_norm": 0.017552798613905907,
+ "learning_rate": 0.0006,
+ "loss": 4.647500038146973,
+ "step": 1131
+ },
+ {
+ "epoch": 15.726637554585153,
+ "grad_norm": 0.01751251146197319,
+ "learning_rate": 0.0006,
+ "loss": 4.590545654296875,
+ "step": 1132
+ },
+ {
+ "epoch": 15.74061135371179,
+ "grad_norm": 0.01699099875986576,
+ "learning_rate": 0.0006,
+ "loss": 4.509699821472168,
+ "step": 1133
+ },
+ {
+ "epoch": 15.754585152838429,
+ "grad_norm": 0.015808649361133575,
+ "learning_rate": 0.0006,
+ "loss": 4.606992244720459,
+ "step": 1134
+ },
+ {
+ "epoch": 15.768558951965066,
+ "grad_norm": 0.01655443385243416,
+ "learning_rate": 0.0006,
+ "loss": 4.585219860076904,
+ "step": 1135
+ },
+ {
+ "epoch": 15.782532751091702,
+ "grad_norm": 0.01708163693547249,
+ "learning_rate": 0.0006,
+ "loss": 4.668147087097168,
+ "step": 1136
+ },
+ {
+ "epoch": 15.796506550218341,
+ "grad_norm": 0.017161913216114044,
+ "learning_rate": 0.0006,
+ "loss": 4.7170634269714355,
+ "step": 1137
+ },
+ {
+ "epoch": 15.810480349344978,
+ "grad_norm": 0.017262674868106842,
+ "learning_rate": 0.0006,
+ "loss": 4.572132110595703,
+ "step": 1138
+ },
+ {
+ "epoch": 15.824454148471617,
+ "grad_norm": 0.017713576555252075,
+ "learning_rate": 0.0006,
+ "loss": 4.702538967132568,
+ "step": 1139
+ },
+ {
+ "epoch": 15.838427947598253,
+ "grad_norm": 0.018730733543634415,
+ "learning_rate": 0.0006,
+ "loss": 4.629721641540527,
+ "step": 1140
+ },
+ {
+ "epoch": 15.85240174672489,
+ "grad_norm": 0.017960187047719955,
+ "learning_rate": 0.0006,
+ "loss": 4.661039352416992,
+ "step": 1141
+ },
+ {
+ "epoch": 15.866375545851529,
+ "grad_norm": 0.01750079356133938,
+ "learning_rate": 0.0006,
+ "loss": 4.503710746765137,
+ "step": 1142
+ },
+ {
+ "epoch": 15.880349344978166,
+ "grad_norm": 0.016630422323942184,
+ "learning_rate": 0.0006,
+ "loss": 4.6582136154174805,
+ "step": 1143
+ },
+ {
+ "epoch": 15.894323144104803,
+ "grad_norm": 0.015896832570433617,
+ "learning_rate": 0.0006,
+ "loss": 4.596693992614746,
+ "step": 1144
+ },
+ {
+ "epoch": 15.908296943231441,
+ "grad_norm": 0.01566407084465027,
+ "learning_rate": 0.0006,
+ "loss": 4.6337199211120605,
+ "step": 1145
+ },
+ {
+ "epoch": 15.922270742358078,
+ "grad_norm": 0.016823330894112587,
+ "learning_rate": 0.0006,
+ "loss": 4.686994552612305,
+ "step": 1146
+ },
+ {
+ "epoch": 15.936244541484717,
+ "grad_norm": 0.016208255663514137,
+ "learning_rate": 0.0006,
+ "loss": 4.6205549240112305,
+ "step": 1147
+ },
+ {
+ "epoch": 15.950218340611354,
+ "grad_norm": 0.0171707421541214,
+ "learning_rate": 0.0006,
+ "loss": 4.582968711853027,
+ "step": 1148
+ },
+ {
+ "epoch": 15.96419213973799,
+ "grad_norm": 0.017009446397423744,
+ "learning_rate": 0.0006,
+ "loss": 4.510274887084961,
+ "step": 1149
+ },
+ {
+ "epoch": 15.97816593886463,
+ "grad_norm": 0.01565811224281788,
+ "learning_rate": 0.0006,
+ "loss": 4.55834436416626,
+ "step": 1150
+ },
+ {
+ "epoch": 15.992139737991266,
+ "grad_norm": 0.015276037156581879,
+ "learning_rate": 0.0006,
+ "loss": 4.537870407104492,
+ "step": 1151
+ },
+ {
+ "epoch": 16.0,
+ "grad_norm": 0.017305882647633553,
+ "learning_rate": 0.0006,
+ "loss": 4.611305236816406,
+ "step": 1152
+ },
+ {
+ "epoch": 16.0,
+ "eval_loss": 4.775513172149658,
+ "eval_runtime": 56.7341,
+ "eval_samples_per_second": 43.043,
+ "eval_steps_per_second": 1.357,
+ "step": 1152
+ },
+ {
+ "epoch": 16.01397379912664,
+ "grad_norm": 0.016752436757087708,
+ "learning_rate": 0.0006,
+ "loss": 4.541312217712402,
+ "step": 1153
+ },
+ {
+ "epoch": 16.027947598253274,
+ "grad_norm": 0.01643134094774723,
+ "learning_rate": 0.0006,
+ "loss": 4.585402965545654,
+ "step": 1154
+ },
+ {
+ "epoch": 16.041921397379912,
+ "grad_norm": 0.01888720504939556,
+ "learning_rate": 0.0006,
+ "loss": 4.597825050354004,
+ "step": 1155
+ },
+ {
+ "epoch": 16.05589519650655,
+ "grad_norm": 0.022044632583856583,
+ "learning_rate": 0.0006,
+ "loss": 4.532772064208984,
+ "step": 1156
+ },
+ {
+ "epoch": 16.069868995633186,
+ "grad_norm": 0.022986987605690956,
+ "learning_rate": 0.0006,
+ "loss": 4.551351547241211,
+ "step": 1157
+ },
+ {
+ "epoch": 16.083842794759825,
+ "grad_norm": 0.020116791129112244,
+ "learning_rate": 0.0006,
+ "loss": 4.484149932861328,
+ "step": 1158
+ },
+ {
+ "epoch": 16.097816593886463,
+ "grad_norm": 0.021442247554659843,
+ "learning_rate": 0.0006,
+ "loss": 4.515316486358643,
+ "step": 1159
+ },
+ {
+ "epoch": 16.111790393013102,
+ "grad_norm": 0.020974772050976753,
+ "learning_rate": 0.0006,
+ "loss": 4.51606559753418,
+ "step": 1160
+ },
+ {
+ "epoch": 16.125764192139737,
+ "grad_norm": 0.021274864673614502,
+ "learning_rate": 0.0006,
+ "loss": 4.587329864501953,
+ "step": 1161
+ },
+ {
+ "epoch": 16.139737991266376,
+ "grad_norm": 0.022799784317612648,
+ "learning_rate": 0.0006,
+ "loss": 4.568149089813232,
+ "step": 1162
+ },
+ {
+ "epoch": 16.153711790393015,
+ "grad_norm": 0.021784933283925056,
+ "learning_rate": 0.0006,
+ "loss": 4.515336990356445,
+ "step": 1163
+ },
+ {
+ "epoch": 16.16768558951965,
+ "grad_norm": 0.024261271581053734,
+ "learning_rate": 0.0006,
+ "loss": 4.5127081871032715,
+ "step": 1164
+ },
+ {
+ "epoch": 16.18165938864629,
+ "grad_norm": 0.02248375490307808,
+ "learning_rate": 0.0006,
+ "loss": 4.754251003265381,
+ "step": 1165
+ },
+ {
+ "epoch": 16.195633187772927,
+ "grad_norm": 0.02136492356657982,
+ "learning_rate": 0.0006,
+ "loss": 4.58193826675415,
+ "step": 1166
+ },
+ {
+ "epoch": 16.209606986899562,
+ "grad_norm": 0.020727066323161125,
+ "learning_rate": 0.0006,
+ "loss": 4.564209938049316,
+ "step": 1167
+ },
+ {
+ "epoch": 16.2235807860262,
+ "grad_norm": 0.017789119854569435,
+ "learning_rate": 0.0006,
+ "loss": 4.698680877685547,
+ "step": 1168
+ },
+ {
+ "epoch": 16.23755458515284,
+ "grad_norm": 0.017603034153580666,
+ "learning_rate": 0.0006,
+ "loss": 4.501171588897705,
+ "step": 1169
+ },
+ {
+ "epoch": 16.251528384279474,
+ "grad_norm": 0.017345862463116646,
+ "learning_rate": 0.0006,
+ "loss": 4.593590259552002,
+ "step": 1170
+ },
+ {
+ "epoch": 16.265502183406113,
+ "grad_norm": 0.018047425895929337,
+ "learning_rate": 0.0006,
+ "loss": 4.597467422485352,
+ "step": 1171
+ },
+ {
+ "epoch": 16.27947598253275,
+ "grad_norm": 0.01947222836315632,
+ "learning_rate": 0.0006,
+ "loss": 4.5072221755981445,
+ "step": 1172
+ },
+ {
+ "epoch": 16.29344978165939,
+ "grad_norm": 0.019598500803112984,
+ "learning_rate": 0.0006,
+ "loss": 4.601351737976074,
+ "step": 1173
+ },
+ {
+ "epoch": 16.307423580786025,
+ "grad_norm": 0.020002514123916626,
+ "learning_rate": 0.0006,
+ "loss": 4.580431938171387,
+ "step": 1174
+ },
+ {
+ "epoch": 16.321397379912664,
+ "grad_norm": 0.0216530729085207,
+ "learning_rate": 0.0006,
+ "loss": 4.424681663513184,
+ "step": 1175
+ },
+ {
+ "epoch": 16.335371179039303,
+ "grad_norm": 0.01949235238134861,
+ "learning_rate": 0.0006,
+ "loss": 4.751883029937744,
+ "step": 1176
+ },
+ {
+ "epoch": 16.349344978165938,
+ "grad_norm": 0.017779016867280006,
+ "learning_rate": 0.0006,
+ "loss": 4.541490077972412,
+ "step": 1177
+ },
+ {
+ "epoch": 16.363318777292577,
+ "grad_norm": 0.01909121684730053,
+ "learning_rate": 0.0006,
+ "loss": 4.594141006469727,
+ "step": 1178
+ },
+ {
+ "epoch": 16.377292576419215,
+ "grad_norm": 0.019894849509000778,
+ "learning_rate": 0.0006,
+ "loss": 4.555070400238037,
+ "step": 1179
+ },
+ {
+ "epoch": 16.39126637554585,
+ "grad_norm": 0.01993032731115818,
+ "learning_rate": 0.0006,
+ "loss": 4.521401405334473,
+ "step": 1180
+ },
+ {
+ "epoch": 16.40524017467249,
+ "grad_norm": 0.0210788045078516,
+ "learning_rate": 0.0006,
+ "loss": 4.672842979431152,
+ "step": 1181
+ },
+ {
+ "epoch": 16.419213973799128,
+ "grad_norm": 0.02274598926305771,
+ "learning_rate": 0.0006,
+ "loss": 4.629504203796387,
+ "step": 1182
+ },
+ {
+ "epoch": 16.433187772925763,
+ "grad_norm": 0.020448120310902596,
+ "learning_rate": 0.0006,
+ "loss": 4.548209190368652,
+ "step": 1183
+ },
+ {
+ "epoch": 16.4471615720524,
+ "grad_norm": 0.017565643414855003,
+ "learning_rate": 0.0006,
+ "loss": 4.529054641723633,
+ "step": 1184
+ },
+ {
+ "epoch": 16.46113537117904,
+ "grad_norm": 0.019504237920045853,
+ "learning_rate": 0.0006,
+ "loss": 4.48182487487793,
+ "step": 1185
+ },
+ {
+ "epoch": 16.475109170305675,
+ "grad_norm": 0.016823742538690567,
+ "learning_rate": 0.0006,
+ "loss": 4.608696937561035,
+ "step": 1186
+ },
+ {
+ "epoch": 16.489082969432314,
+ "grad_norm": 0.01602894812822342,
+ "learning_rate": 0.0006,
+ "loss": 4.620062351226807,
+ "step": 1187
+ },
+ {
+ "epoch": 16.503056768558952,
+ "grad_norm": 0.016615642234683037,
+ "learning_rate": 0.0006,
+ "loss": 4.69378662109375,
+ "step": 1188
+ },
+ {
+ "epoch": 16.51703056768559,
+ "grad_norm": 0.018576730042696,
+ "learning_rate": 0.0006,
+ "loss": 4.537041664123535,
+ "step": 1189
+ },
+ {
+ "epoch": 16.531004366812226,
+ "grad_norm": 0.018611151725053787,
+ "learning_rate": 0.0006,
+ "loss": 4.496169090270996,
+ "step": 1190
+ },
+ {
+ "epoch": 16.544978165938865,
+ "grad_norm": 0.018622413277626038,
+ "learning_rate": 0.0006,
+ "loss": 4.557358264923096,
+ "step": 1191
+ },
+ {
+ "epoch": 16.558951965065503,
+ "grad_norm": 0.01823544315993786,
+ "learning_rate": 0.0006,
+ "loss": 4.730792999267578,
+ "step": 1192
+ },
+ {
+ "epoch": 16.57292576419214,
+ "grad_norm": 0.018796846270561218,
+ "learning_rate": 0.0006,
+ "loss": 4.781791687011719,
+ "step": 1193
+ },
+ {
+ "epoch": 16.586899563318777,
+ "grad_norm": 0.01728292554616928,
+ "learning_rate": 0.0006,
+ "loss": 4.515257835388184,
+ "step": 1194
+ },
+ {
+ "epoch": 16.600873362445416,
+ "grad_norm": 0.01699589006602764,
+ "learning_rate": 0.0006,
+ "loss": 4.5996904373168945,
+ "step": 1195
+ },
+ {
+ "epoch": 16.61484716157205,
+ "grad_norm": 0.015905970707535744,
+ "learning_rate": 0.0006,
+ "loss": 4.601301193237305,
+ "step": 1196
+ },
+ {
+ "epoch": 16.62882096069869,
+ "grad_norm": 0.01338571310043335,
+ "learning_rate": 0.0006,
+ "loss": 4.557708740234375,
+ "step": 1197
+ },
+ {
+ "epoch": 16.64279475982533,
+ "grad_norm": 0.015298855490982533,
+ "learning_rate": 0.0006,
+ "loss": 4.542287349700928,
+ "step": 1198
+ },
+ {
+ "epoch": 16.656768558951963,
+ "grad_norm": 0.014753367751836777,
+ "learning_rate": 0.0006,
+ "loss": 4.595787048339844,
+ "step": 1199
+ },
+ {
+ "epoch": 16.670742358078602,
+ "grad_norm": 0.015527483075857162,
+ "learning_rate": 0.0006,
+ "loss": 4.607529163360596,
+ "step": 1200
+ },
+ {
+ "epoch": 16.68471615720524,
+ "grad_norm": 0.015042162500321865,
+ "learning_rate": 0.0006,
+ "loss": 4.686626434326172,
+ "step": 1201
+ },
+ {
+ "epoch": 16.69868995633188,
+ "grad_norm": 0.0159393772482872,
+ "learning_rate": 0.0006,
+ "loss": 4.47656774520874,
+ "step": 1202
+ },
+ {
+ "epoch": 16.712663755458514,
+ "grad_norm": 0.01532787922769785,
+ "learning_rate": 0.0006,
+ "loss": 4.55678653717041,
+ "step": 1203
+ },
+ {
+ "epoch": 16.726637554585153,
+ "grad_norm": 0.016332164406776428,
+ "learning_rate": 0.0006,
+ "loss": 4.51441764831543,
+ "step": 1204
+ },
+ {
+ "epoch": 16.74061135371179,
+ "grad_norm": 0.018901726230978966,
+ "learning_rate": 0.0006,
+ "loss": 4.57227897644043,
+ "step": 1205
+ },
+ {
+ "epoch": 16.754585152838427,
+ "grad_norm": 0.018589433282613754,
+ "learning_rate": 0.0006,
+ "loss": 4.550070762634277,
+ "step": 1206
+ },
+ {
+ "epoch": 16.768558951965066,
+ "grad_norm": 0.01634034886956215,
+ "learning_rate": 0.0006,
+ "loss": 4.586910724639893,
+ "step": 1207
+ },
+ {
+ "epoch": 16.782532751091704,
+ "grad_norm": 0.017599433660507202,
+ "learning_rate": 0.0006,
+ "loss": 4.5891923904418945,
+ "step": 1208
+ },
+ {
+ "epoch": 16.79650655021834,
+ "grad_norm": 0.01878446340560913,
+ "learning_rate": 0.0006,
+ "loss": 4.57227087020874,
+ "step": 1209
+ },
+ {
+ "epoch": 16.810480349344978,
+ "grad_norm": 0.01803469844162464,
+ "learning_rate": 0.0006,
+ "loss": 4.583845138549805,
+ "step": 1210
+ },
+ {
+ "epoch": 16.824454148471617,
+ "grad_norm": 0.018505144864320755,
+ "learning_rate": 0.0006,
+ "loss": 4.561151504516602,
+ "step": 1211
+ },
+ {
+ "epoch": 16.83842794759825,
+ "grad_norm": 0.01920081488788128,
+ "learning_rate": 0.0006,
+ "loss": 4.506834030151367,
+ "step": 1212
+ },
+ {
+ "epoch": 16.85240174672489,
+ "grad_norm": 0.0177712831646204,
+ "learning_rate": 0.0006,
+ "loss": 4.557195663452148,
+ "step": 1213
+ },
+ {
+ "epoch": 16.86637554585153,
+ "grad_norm": 0.019113918766379356,
+ "learning_rate": 0.0006,
+ "loss": 4.651227951049805,
+ "step": 1214
+ },
+ {
+ "epoch": 16.880349344978168,
+ "grad_norm": 0.02012539468705654,
+ "learning_rate": 0.0006,
+ "loss": 4.4952239990234375,
+ "step": 1215
+ },
+ {
+ "epoch": 16.894323144104803,
+ "grad_norm": 0.019993646070361137,
+ "learning_rate": 0.0006,
+ "loss": 4.517433166503906,
+ "step": 1216
+ },
+ {
+ "epoch": 16.90829694323144,
+ "grad_norm": 0.020041441544890404,
+ "learning_rate": 0.0006,
+ "loss": 4.489407539367676,
+ "step": 1217
+ },
+ {
+ "epoch": 16.92227074235808,
+ "grad_norm": 0.0183611661195755,
+ "learning_rate": 0.0006,
+ "loss": 4.5435638427734375,
+ "step": 1218
+ },
+ {
+ "epoch": 16.936244541484715,
+ "grad_norm": 0.019138429313898087,
+ "learning_rate": 0.0006,
+ "loss": 4.545949459075928,
+ "step": 1219
+ },
+ {
+ "epoch": 16.950218340611354,
+ "grad_norm": 0.018554266542196274,
+ "learning_rate": 0.0006,
+ "loss": 4.579620361328125,
+ "step": 1220
+ },
+ {
+ "epoch": 16.964192139737992,
+ "grad_norm": 0.0193545650690794,
+ "learning_rate": 0.0006,
+ "loss": 4.588255405426025,
+ "step": 1221
+ },
+ {
+ "epoch": 16.978165938864628,
+ "grad_norm": 0.019389165565371513,
+ "learning_rate": 0.0006,
+ "loss": 4.520698547363281,
+ "step": 1222
+ },
+ {
+ "epoch": 16.992139737991266,
+ "grad_norm": 0.016601379960775375,
+ "learning_rate": 0.0006,
+ "loss": 4.4809064865112305,
+ "step": 1223
+ },
+ {
+ "epoch": 17.0,
+ "grad_norm": 0.01771303080022335,
+ "learning_rate": 0.0006,
+ "loss": 4.721122741699219,
+ "step": 1224
+ },
+ {
+ "epoch": 17.0,
+ "eval_loss": 4.7903666496276855,
+ "eval_runtime": 56.9826,
+ "eval_samples_per_second": 42.855,
+ "eval_steps_per_second": 1.351,
+ "step": 1224
+ },
+ {
+ "epoch": 17.01397379912664,
+ "grad_norm": 0.017720289528369904,
+ "learning_rate": 0.0006,
+ "loss": 4.523834705352783,
+ "step": 1225
+ },
+ {
+ "epoch": 17.027947598253274,
+ "grad_norm": 0.020137161016464233,
+ "learning_rate": 0.0006,
+ "loss": 4.574061870574951,
+ "step": 1226
+ },
+ {
+ "epoch": 17.041921397379912,
+ "grad_norm": 0.02002314105629921,
+ "learning_rate": 0.0006,
+ "loss": 4.512407302856445,
+ "step": 1227
+ },
+ {
+ "epoch": 17.05589519650655,
+ "grad_norm": 0.019628183916211128,
+ "learning_rate": 0.0006,
+ "loss": 4.502533435821533,
+ "step": 1228
+ },
+ {
+ "epoch": 17.069868995633186,
+ "grad_norm": 0.019808458164334297,
+ "learning_rate": 0.0006,
+ "loss": 4.463156700134277,
+ "step": 1229
+ },
+ {
+ "epoch": 17.083842794759825,
+ "grad_norm": 0.021250786259770393,
+ "learning_rate": 0.0006,
+ "loss": 4.622682094573975,
+ "step": 1230
+ },
+ {
+ "epoch": 17.097816593886463,
+ "grad_norm": 0.022031694650650024,
+ "learning_rate": 0.0006,
+ "loss": 4.451882839202881,
+ "step": 1231
+ },
+ {
+ "epoch": 17.111790393013102,
+ "grad_norm": 0.01898987777531147,
+ "learning_rate": 0.0006,
+ "loss": 4.529331684112549,
+ "step": 1232
+ },
+ {
+ "epoch": 17.125764192139737,
+ "grad_norm": 0.018749834969639778,
+ "learning_rate": 0.0006,
+ "loss": 4.435815334320068,
+ "step": 1233
+ },
+ {
+ "epoch": 17.139737991266376,
+ "grad_norm": 0.01896674931049347,
+ "learning_rate": 0.0006,
+ "loss": 4.484029293060303,
+ "step": 1234
+ },
+ {
+ "epoch": 17.153711790393015,
+ "grad_norm": 0.018286997452378273,
+ "learning_rate": 0.0006,
+ "loss": 4.523510932922363,
+ "step": 1235
+ },
+ {
+ "epoch": 17.16768558951965,
+ "grad_norm": 0.017273059114813805,
+ "learning_rate": 0.0006,
+ "loss": 4.451767444610596,
+ "step": 1236
+ },
+ {
+ "epoch": 17.18165938864629,
+ "grad_norm": 0.016499150544404984,
+ "learning_rate": 0.0006,
+ "loss": 4.5568647384643555,
+ "step": 1237
+ },
+ {
+ "epoch": 17.195633187772927,
+ "grad_norm": 0.015934469178318977,
+ "learning_rate": 0.0006,
+ "loss": 4.4639997482299805,
+ "step": 1238
+ },
+ {
+ "epoch": 17.209606986899562,
+ "grad_norm": 0.01714209094643593,
+ "learning_rate": 0.0006,
+ "loss": 4.547614097595215,
+ "step": 1239
+ },
+ {
+ "epoch": 17.2235807860262,
+ "grad_norm": 0.019167495891451836,
+ "learning_rate": 0.0006,
+ "loss": 4.571097373962402,
+ "step": 1240
+ },
+ {
+ "epoch": 17.23755458515284,
+ "grad_norm": 0.018276996910572052,
+ "learning_rate": 0.0006,
+ "loss": 4.489764213562012,
+ "step": 1241
+ },
+ {
+ "epoch": 17.251528384279474,
+ "grad_norm": 0.021296866238117218,
+ "learning_rate": 0.0006,
+ "loss": 4.526913642883301,
+ "step": 1242
+ },
+ {
+ "epoch": 17.265502183406113,
+ "grad_norm": 0.022606629878282547,
+ "learning_rate": 0.0006,
+ "loss": 4.521624565124512,
+ "step": 1243
+ },
+ {
+ "epoch": 17.27947598253275,
+ "grad_norm": 0.021204711869359016,
+ "learning_rate": 0.0006,
+ "loss": 4.500864028930664,
+ "step": 1244
+ },
+ {
+ "epoch": 17.29344978165939,
+ "grad_norm": 0.01988849975168705,
+ "learning_rate": 0.0006,
+ "loss": 4.45162296295166,
+ "step": 1245
+ },
+ {
+ "epoch": 17.307423580786025,
+ "grad_norm": 0.020268384367227554,
+ "learning_rate": 0.0006,
+ "loss": 4.536198616027832,
+ "step": 1246
+ },
+ {
+ "epoch": 17.321397379912664,
+ "grad_norm": 0.020185677334666252,
+ "learning_rate": 0.0006,
+ "loss": 4.440262794494629,
+ "step": 1247
+ },
+ {
+ "epoch": 17.335371179039303,
+ "grad_norm": 0.020106853917241096,
+ "learning_rate": 0.0006,
+ "loss": 4.475907325744629,
+ "step": 1248
+ },
+ {
+ "epoch": 17.349344978165938,
+ "grad_norm": 0.018883297219872475,
+ "learning_rate": 0.0006,
+ "loss": 4.640661716461182,
+ "step": 1249
+ },
+ {
+ "epoch": 17.363318777292577,
+ "grad_norm": 0.01852620206773281,
+ "learning_rate": 0.0006,
+ "loss": 4.578243255615234,
+ "step": 1250
+ },
+ {
+ "epoch": 17.377292576419215,
+ "grad_norm": 0.01886451430618763,
+ "learning_rate": 0.0006,
+ "loss": 4.546026229858398,
+ "step": 1251
+ },
+ {
+ "epoch": 17.39126637554585,
+ "grad_norm": 0.01777452416718006,
+ "learning_rate": 0.0006,
+ "loss": 4.521542549133301,
+ "step": 1252
+ },
+ {
+ "epoch": 17.40524017467249,
+ "grad_norm": 0.016626447439193726,
+ "learning_rate": 0.0006,
+ "loss": 4.52629280090332,
+ "step": 1253
+ },
+ {
+ "epoch": 17.419213973799128,
+ "grad_norm": 0.01599477231502533,
+ "learning_rate": 0.0006,
+ "loss": 4.4796037673950195,
+ "step": 1254
+ },
+ {
+ "epoch": 17.433187772925763,
+ "grad_norm": 0.017021458595991135,
+ "learning_rate": 0.0006,
+ "loss": 4.468813419342041,
+ "step": 1255
+ },
+ {
+ "epoch": 17.4471615720524,
+ "grad_norm": 0.016925333067774773,
+ "learning_rate": 0.0006,
+ "loss": 4.584898948669434,
+ "step": 1256
+ },
+ {
+ "epoch": 17.46113537117904,
+ "grad_norm": 0.01714998669922352,
+ "learning_rate": 0.0006,
+ "loss": 4.523544788360596,
+ "step": 1257
+ },
+ {
+ "epoch": 17.475109170305675,
+ "grad_norm": 0.01802288182079792,
+ "learning_rate": 0.0006,
+ "loss": 4.595581531524658,
+ "step": 1258
+ },
+ {
+ "epoch": 17.489082969432314,
+ "grad_norm": 0.0180289875715971,
+ "learning_rate": 0.0006,
+ "loss": 4.394440650939941,
+ "step": 1259
+ },
+ {
+ "epoch": 17.503056768558952,
+ "grad_norm": 0.018504245206713676,
+ "learning_rate": 0.0006,
+ "loss": 4.4794416427612305,
+ "step": 1260
+ },
+ {
+ "epoch": 17.51703056768559,
+ "grad_norm": 0.019973233342170715,
+ "learning_rate": 0.0006,
+ "loss": 4.4241437911987305,
+ "step": 1261
+ },
+ {
+ "epoch": 17.531004366812226,
+ "grad_norm": 0.020063083618879318,
+ "learning_rate": 0.0006,
+ "loss": 4.48068380355835,
+ "step": 1262
+ },
+ {
+ "epoch": 17.544978165938865,
+ "grad_norm": 0.02066640742123127,
+ "learning_rate": 0.0006,
+ "loss": 4.628993034362793,
+ "step": 1263
+ },
+ {
+ "epoch": 17.558951965065503,
+ "grad_norm": 0.02206416428089142,
+ "learning_rate": 0.0006,
+ "loss": 4.564790725708008,
+ "step": 1264
+ },
+ {
+ "epoch": 17.57292576419214,
+ "grad_norm": 0.020640335977077484,
+ "learning_rate": 0.0006,
+ "loss": 4.578705787658691,
+ "step": 1265
+ },
+ {
+ "epoch": 17.586899563318777,
+ "grad_norm": 0.019093405455350876,
+ "learning_rate": 0.0006,
+ "loss": 4.5023603439331055,
+ "step": 1266
+ },
+ {
+ "epoch": 17.600873362445416,
+ "grad_norm": 0.019459193572402,
+ "learning_rate": 0.0006,
+ "loss": 4.594531059265137,
+ "step": 1267
+ },
+ {
+ "epoch": 17.61484716157205,
+ "grad_norm": 0.02021171525120735,
+ "learning_rate": 0.0006,
+ "loss": 4.770299911499023,
+ "step": 1268
+ },
+ {
+ "epoch": 17.62882096069869,
+ "grad_norm": 0.021161191165447235,
+ "learning_rate": 0.0006,
+ "loss": 4.5380401611328125,
+ "step": 1269
+ },
+ {
+ "epoch": 17.64279475982533,
+ "grad_norm": 0.021013403311371803,
+ "learning_rate": 0.0006,
+ "loss": 4.455524921417236,
+ "step": 1270
+ },
+ {
+ "epoch": 17.656768558951963,
+ "grad_norm": 0.020336592569947243,
+ "learning_rate": 0.0006,
+ "loss": 4.533672332763672,
+ "step": 1271
+ },
+ {
+ "epoch": 17.670742358078602,
+ "grad_norm": 0.020159434527158737,
+ "learning_rate": 0.0006,
+ "loss": 4.596151351928711,
+ "step": 1272
+ },
+ {
+ "epoch": 17.68471615720524,
+ "grad_norm": 0.020432960242033005,
+ "learning_rate": 0.0006,
+ "loss": 4.55531644821167,
+ "step": 1273
+ },
+ {
+ "epoch": 17.69868995633188,
+ "grad_norm": 0.025466229766607285,
+ "learning_rate": 0.0006,
+ "loss": 4.566912651062012,
+ "step": 1274
+ },
+ {
+ "epoch": 17.712663755458514,
+ "grad_norm": 0.026537558063864708,
+ "learning_rate": 0.0006,
+ "loss": 4.526003837585449,
+ "step": 1275
+ },
+ {
+ "epoch": 17.726637554585153,
+ "grad_norm": 0.024397607892751694,
+ "learning_rate": 0.0006,
+ "loss": 4.541755676269531,
+ "step": 1276
+ },
+ {
+ "epoch": 17.74061135371179,
+ "grad_norm": 0.022062312811613083,
+ "learning_rate": 0.0006,
+ "loss": 4.6532793045043945,
+ "step": 1277
+ },
+ {
+ "epoch": 17.754585152838427,
+ "grad_norm": 0.02204282023012638,
+ "learning_rate": 0.0006,
+ "loss": 4.582825660705566,
+ "step": 1278
+ },
+ {
+ "epoch": 17.768558951965066,
+ "grad_norm": 0.020844416692852974,
+ "learning_rate": 0.0006,
+ "loss": 4.522336483001709,
+ "step": 1279
+ },
+ {
+ "epoch": 17.782532751091704,
+ "grad_norm": 0.019815703853964806,
+ "learning_rate": 0.0006,
+ "loss": 4.607398509979248,
+ "step": 1280
+ },
+ {
+ "epoch": 17.79650655021834,
+ "grad_norm": 0.016676954925060272,
+ "learning_rate": 0.0006,
+ "loss": 4.446206092834473,
+ "step": 1281
+ },
+ {
+ "epoch": 17.810480349344978,
+ "grad_norm": 0.0166452769190073,
+ "learning_rate": 0.0006,
+ "loss": 4.505931854248047,
+ "step": 1282
+ },
+ {
+ "epoch": 17.824454148471617,
+ "grad_norm": 0.015928907319903374,
+ "learning_rate": 0.0006,
+ "loss": 4.455267429351807,
+ "step": 1283
+ },
+ {
+ "epoch": 17.83842794759825,
+ "grad_norm": 0.015219560824334621,
+ "learning_rate": 0.0006,
+ "loss": 4.5623674392700195,
+ "step": 1284
+ },
+ {
+ "epoch": 17.85240174672489,
+ "grad_norm": 0.015125355683267117,
+ "learning_rate": 0.0006,
+ "loss": 4.554104804992676,
+ "step": 1285
+ },
+ {
+ "epoch": 17.86637554585153,
+ "grad_norm": 0.015612718649208546,
+ "learning_rate": 0.0006,
+ "loss": 4.484847068786621,
+ "step": 1286
+ },
+ {
+ "epoch": 17.880349344978168,
+ "grad_norm": 0.015215662308037281,
+ "learning_rate": 0.0006,
+ "loss": 4.4802093505859375,
+ "step": 1287
+ },
+ {
+ "epoch": 17.894323144104803,
+ "grad_norm": 0.01511684525758028,
+ "learning_rate": 0.0006,
+ "loss": 4.537298679351807,
+ "step": 1288
+ },
+ {
+ "epoch": 17.90829694323144,
+ "grad_norm": 0.013841322623193264,
+ "learning_rate": 0.0006,
+ "loss": 4.427488327026367,
+ "step": 1289
+ },
+ {
+ "epoch": 17.92227074235808,
+ "grad_norm": 0.015115964226424694,
+ "learning_rate": 0.0006,
+ "loss": 4.4547119140625,
+ "step": 1290
+ },
+ {
+ "epoch": 17.936244541484715,
+ "grad_norm": 0.016411006450653076,
+ "learning_rate": 0.0006,
+ "loss": 4.52994966506958,
+ "step": 1291
+ },
+ {
+ "epoch": 17.950218340611354,
+ "grad_norm": 0.017313152551651,
+ "learning_rate": 0.0006,
+ "loss": 4.5337653160095215,
+ "step": 1292
+ },
+ {
+ "epoch": 17.964192139737992,
+ "grad_norm": 0.016559796407818794,
+ "learning_rate": 0.0006,
+ "loss": 4.5277910232543945,
+ "step": 1293
+ },
+ {
+ "epoch": 17.978165938864628,
+ "grad_norm": 0.015854761004447937,
+ "learning_rate": 0.0006,
+ "loss": 4.6334309577941895,
+ "step": 1294
+ },
+ {
+ "epoch": 17.992139737991266,
+ "grad_norm": 0.015658238902688026,
+ "learning_rate": 0.0006,
+ "loss": 4.613372802734375,
+ "step": 1295
+ },
+ {
+ "epoch": 18.0,
+ "grad_norm": 0.0169155802577734,
+ "learning_rate": 0.0006,
+ "loss": 4.527356147766113,
+ "step": 1296
+ },
+ {
+ "epoch": 18.0,
+ "eval_loss": 4.764543056488037,
+ "eval_runtime": 56.5998,
+ "eval_samples_per_second": 43.145,
+ "eval_steps_per_second": 1.36,
+ "step": 1296
+ },
+ {
+ "epoch": 18.01397379912664,
+ "grad_norm": 0.015348963439464569,
+ "learning_rate": 0.0006,
+ "loss": 4.466423988342285,
+ "step": 1297
+ },
+ {
+ "epoch": 18.027947598253274,
+ "grad_norm": 0.01768588088452816,
+ "learning_rate": 0.0006,
+ "loss": 4.545933246612549,
+ "step": 1298
+ },
+ {
+ "epoch": 18.041921397379912,
+ "grad_norm": 0.017381226643919945,
+ "learning_rate": 0.0006,
+ "loss": 4.494060039520264,
+ "step": 1299
+ },
+ {
+ "epoch": 18.05589519650655,
+ "grad_norm": 0.019002335146069527,
+ "learning_rate": 0.0006,
+ "loss": 4.395200729370117,
+ "step": 1300
+ },
+ {
+ "epoch": 18.069868995633186,
+ "grad_norm": 0.01827932707965374,
+ "learning_rate": 0.0006,
+ "loss": 4.424848556518555,
+ "step": 1301
+ },
+ {
+ "epoch": 18.083842794759825,
+ "grad_norm": 0.016383031383156776,
+ "learning_rate": 0.0006,
+ "loss": 4.494722366333008,
+ "step": 1302
+ },
+ {
+ "epoch": 18.097816593886463,
+ "grad_norm": 0.016265157610177994,
+ "learning_rate": 0.0006,
+ "loss": 4.416522979736328,
+ "step": 1303
+ },
+ {
+ "epoch": 18.111790393013102,
+ "grad_norm": 0.01702386513352394,
+ "learning_rate": 0.0006,
+ "loss": 4.600103378295898,
+ "step": 1304
+ },
+ {
+ "epoch": 18.125764192139737,
+ "grad_norm": 0.018495498225092888,
+ "learning_rate": 0.0006,
+ "loss": 4.588401794433594,
+ "step": 1305
+ },
+ {
+ "epoch": 18.139737991266376,
+ "grad_norm": 0.017921684309840202,
+ "learning_rate": 0.0006,
+ "loss": 4.478075981140137,
+ "step": 1306
+ },
+ {
+ "epoch": 18.153711790393015,
+ "grad_norm": 0.01801278069615364,
+ "learning_rate": 0.0006,
+ "loss": 4.413186550140381,
+ "step": 1307
+ },
+ {
+ "epoch": 18.16768558951965,
+ "grad_norm": 0.017211275175213814,
+ "learning_rate": 0.0006,
+ "loss": 4.461341857910156,
+ "step": 1308
+ },
+ {
+ "epoch": 18.18165938864629,
+ "grad_norm": 0.017589328810572624,
+ "learning_rate": 0.0006,
+ "loss": 4.546969413757324,
+ "step": 1309
+ },
+ {
+ "epoch": 18.195633187772927,
+ "grad_norm": 0.01797698810696602,
+ "learning_rate": 0.0006,
+ "loss": 4.530797004699707,
+ "step": 1310
+ },
+ {
+ "epoch": 18.209606986899562,
+ "grad_norm": 0.020408619195222855,
+ "learning_rate": 0.0006,
+ "loss": 4.62999153137207,
+ "step": 1311
+ },
+ {
+ "epoch": 18.2235807860262,
+ "grad_norm": 0.01994038000702858,
+ "learning_rate": 0.0006,
+ "loss": 4.586811065673828,
+ "step": 1312
+ },
+ {
+ "epoch": 18.23755458515284,
+ "grad_norm": 0.019381340593099594,
+ "learning_rate": 0.0006,
+ "loss": 4.4769368171691895,
+ "step": 1313
+ },
+ {
+ "epoch": 18.251528384279474,
+ "grad_norm": 0.020484555512666702,
+ "learning_rate": 0.0006,
+ "loss": 4.524302959442139,
+ "step": 1314
+ },
+ {
+ "epoch": 18.265502183406113,
+ "grad_norm": 0.02426416426897049,
+ "learning_rate": 0.0006,
+ "loss": 4.561705589294434,
+ "step": 1315
+ },
+ {
+ "epoch": 18.27947598253275,
+ "grad_norm": 0.024748101830482483,
+ "learning_rate": 0.0006,
+ "loss": 4.4654316902160645,
+ "step": 1316
+ },
+ {
+ "epoch": 18.29344978165939,
+ "grad_norm": 0.020703328773379326,
+ "learning_rate": 0.0006,
+ "loss": 4.53033447265625,
+ "step": 1317
+ },
+ {
+ "epoch": 18.307423580786025,
+ "grad_norm": 0.01765458658337593,
+ "learning_rate": 0.0006,
+ "loss": 4.5154266357421875,
+ "step": 1318
+ },
+ {
+ "epoch": 18.321397379912664,
+ "grad_norm": 0.018646420910954475,
+ "learning_rate": 0.0006,
+ "loss": 4.5224714279174805,
+ "step": 1319
+ },
+ {
+ "epoch": 18.335371179039303,
+ "grad_norm": 0.01907474733889103,
+ "learning_rate": 0.0006,
+ "loss": 4.5055437088012695,
+ "step": 1320
+ },
+ {
+ "epoch": 18.349344978165938,
+ "grad_norm": 0.018179846927523613,
+ "learning_rate": 0.0006,
+ "loss": 4.462161064147949,
+ "step": 1321
+ },
+ {
+ "epoch": 18.363318777292577,
+ "grad_norm": 0.016697878018021584,
+ "learning_rate": 0.0006,
+ "loss": 4.473718643188477,
+ "step": 1322
+ },
+ {
+ "epoch": 18.377292576419215,
+ "grad_norm": 0.017504019662737846,
+ "learning_rate": 0.0006,
+ "loss": 4.474348068237305,
+ "step": 1323
+ },
+ {
+ "epoch": 18.39126637554585,
+ "grad_norm": 0.018889913335442543,
+ "learning_rate": 0.0006,
+ "loss": 4.4531636238098145,
+ "step": 1324
+ },
+ {
+ "epoch": 18.40524017467249,
+ "grad_norm": 0.019167251884937286,
+ "learning_rate": 0.0006,
+ "loss": 4.4278364181518555,
+ "step": 1325
+ },
+ {
+ "epoch": 18.419213973799128,
+ "grad_norm": 0.015895741060376167,
+ "learning_rate": 0.0006,
+ "loss": 4.474592208862305,
+ "step": 1326
+ },
+ {
+ "epoch": 18.433187772925763,
+ "grad_norm": 0.01531882956624031,
+ "learning_rate": 0.0006,
+ "loss": 4.403757572174072,
+ "step": 1327
+ },
+ {
+ "epoch": 18.4471615720524,
+ "grad_norm": 0.01565849967300892,
+ "learning_rate": 0.0006,
+ "loss": 4.507664680480957,
+ "step": 1328
+ },
+ {
+ "epoch": 18.46113537117904,
+ "grad_norm": 0.01602315716445446,
+ "learning_rate": 0.0006,
+ "loss": 4.5648908615112305,
+ "step": 1329
+ },
+ {
+ "epoch": 18.475109170305675,
+ "grad_norm": 0.016467366367578506,
+ "learning_rate": 0.0006,
+ "loss": 4.428762912750244,
+ "step": 1330
+ },
+ {
+ "epoch": 18.489082969432314,
+ "grad_norm": 0.01604127697646618,
+ "learning_rate": 0.0006,
+ "loss": 4.527771472930908,
+ "step": 1331
+ },
+ {
+ "epoch": 18.503056768558952,
+ "grad_norm": 0.018825586885213852,
+ "learning_rate": 0.0006,
+ "loss": 4.485386848449707,
+ "step": 1332
+ },
+ {
+ "epoch": 18.51703056768559,
+ "grad_norm": 0.019258007407188416,
+ "learning_rate": 0.0006,
+ "loss": 4.414972305297852,
+ "step": 1333
+ },
+ {
+ "epoch": 18.531004366812226,
+ "grad_norm": 0.018192021176218987,
+ "learning_rate": 0.0006,
+ "loss": 4.412166595458984,
+ "step": 1334
+ },
+ {
+ "epoch": 18.544978165938865,
+ "grad_norm": 0.016120698302984238,
+ "learning_rate": 0.0006,
+ "loss": 4.484317779541016,
+ "step": 1335
+ },
+ {
+ "epoch": 18.558951965065503,
+ "grad_norm": 0.016496622934937477,
+ "learning_rate": 0.0006,
+ "loss": 4.514986515045166,
+ "step": 1336
+ },
+ {
+ "epoch": 18.57292576419214,
+ "grad_norm": 0.01795116998255253,
+ "learning_rate": 0.0006,
+ "loss": 4.41473388671875,
+ "step": 1337
+ },
+ {
+ "epoch": 18.586899563318777,
+ "grad_norm": 0.019062276929616928,
+ "learning_rate": 0.0006,
+ "loss": 4.601996898651123,
+ "step": 1338
+ },
+ {
+ "epoch": 18.600873362445416,
+ "grad_norm": 0.017174631357192993,
+ "learning_rate": 0.0006,
+ "loss": 4.571218490600586,
+ "step": 1339
+ },
+ {
+ "epoch": 18.61484716157205,
+ "grad_norm": 0.014015606604516506,
+ "learning_rate": 0.0006,
+ "loss": 4.526606559753418,
+ "step": 1340
+ },
+ {
+ "epoch": 18.62882096069869,
+ "grad_norm": 0.015358510427176952,
+ "learning_rate": 0.0006,
+ "loss": 4.604286193847656,
+ "step": 1341
+ },
+ {
+ "epoch": 18.64279475982533,
+ "grad_norm": 0.01619391329586506,
+ "learning_rate": 0.0006,
+ "loss": 4.560883522033691,
+ "step": 1342
+ },
+ {
+ "epoch": 18.656768558951963,
+ "grad_norm": 0.015972964465618134,
+ "learning_rate": 0.0006,
+ "loss": 4.479755401611328,
+ "step": 1343
+ },
+ {
+ "epoch": 18.670742358078602,
+ "grad_norm": 0.016861025243997574,
+ "learning_rate": 0.0006,
+ "loss": 4.474471092224121,
+ "step": 1344
+ },
+ {
+ "epoch": 18.68471615720524,
+ "grad_norm": 0.01822078227996826,
+ "learning_rate": 0.0006,
+ "loss": 4.509489059448242,
+ "step": 1345
+ },
+ {
+ "epoch": 18.69868995633188,
+ "grad_norm": 0.019415007904171944,
+ "learning_rate": 0.0006,
+ "loss": 4.508941173553467,
+ "step": 1346
+ },
+ {
+ "epoch": 18.712663755458514,
+ "grad_norm": 0.019856126978993416,
+ "learning_rate": 0.0006,
+ "loss": 4.446670055389404,
+ "step": 1347
+ },
+ {
+ "epoch": 18.726637554585153,
+ "grad_norm": 0.01842520199716091,
+ "learning_rate": 0.0006,
+ "loss": 4.524422645568848,
+ "step": 1348
+ },
+ {
+ "epoch": 18.74061135371179,
+ "grad_norm": 0.017196480184793472,
+ "learning_rate": 0.0006,
+ "loss": 4.498466968536377,
+ "step": 1349
+ },
+ {
+ "epoch": 18.754585152838427,
+ "grad_norm": 0.01823010854423046,
+ "learning_rate": 0.0006,
+ "loss": 4.534684181213379,
+ "step": 1350
+ },
+ {
+ "epoch": 18.768558951965066,
+ "grad_norm": 0.020398065447807312,
+ "learning_rate": 0.0006,
+ "loss": 4.456494331359863,
+ "step": 1351
+ },
+ {
+ "epoch": 18.782532751091704,
+ "grad_norm": 0.022992262616753578,
+ "learning_rate": 0.0006,
+ "loss": 4.476374626159668,
+ "step": 1352
+ },
+ {
+ "epoch": 18.79650655021834,
+ "grad_norm": 0.02235420234501362,
+ "learning_rate": 0.0006,
+ "loss": 4.587776184082031,
+ "step": 1353
+ },
+ {
+ "epoch": 18.810480349344978,
+ "grad_norm": 0.02093626745045185,
+ "learning_rate": 0.0006,
+ "loss": 4.424172401428223,
+ "step": 1354
+ },
+ {
+ "epoch": 18.824454148471617,
+ "grad_norm": 0.01997302658855915,
+ "learning_rate": 0.0006,
+ "loss": 4.448066711425781,
+ "step": 1355
+ },
+ {
+ "epoch": 18.83842794759825,
+ "grad_norm": 0.018265459686517715,
+ "learning_rate": 0.0006,
+ "loss": 4.436610698699951,
+ "step": 1356
+ },
+ {
+ "epoch": 18.85240174672489,
+ "grad_norm": 0.01673833839595318,
+ "learning_rate": 0.0006,
+ "loss": 4.5502214431762695,
+ "step": 1357
+ },
+ {
+ "epoch": 18.86637554585153,
+ "grad_norm": 0.017908833920955658,
+ "learning_rate": 0.0006,
+ "loss": 4.550530433654785,
+ "step": 1358
+ },
+ {
+ "epoch": 18.880349344978168,
+ "grad_norm": 0.01796654611825943,
+ "learning_rate": 0.0006,
+ "loss": 4.531463146209717,
+ "step": 1359
+ },
+ {
+ "epoch": 18.894323144104803,
+ "grad_norm": 0.01646183803677559,
+ "learning_rate": 0.0006,
+ "loss": 4.498582363128662,
+ "step": 1360
+ },
+ {
+ "epoch": 18.90829694323144,
+ "grad_norm": 0.015430327504873276,
+ "learning_rate": 0.0006,
+ "loss": 4.52571439743042,
+ "step": 1361
+ },
+ {
+ "epoch": 18.92227074235808,
+ "grad_norm": 0.016527952626347542,
+ "learning_rate": 0.0006,
+ "loss": 4.547013759613037,
+ "step": 1362
+ },
+ {
+ "epoch": 18.936244541484715,
+ "grad_norm": 0.016263660043478012,
+ "learning_rate": 0.0006,
+ "loss": 4.468754291534424,
+ "step": 1363
+ },
+ {
+ "epoch": 18.950218340611354,
+ "grad_norm": 0.016347210854291916,
+ "learning_rate": 0.0006,
+ "loss": 4.450741767883301,
+ "step": 1364
+ },
+ {
+ "epoch": 18.964192139737992,
+ "grad_norm": 0.01701398566365242,
+ "learning_rate": 0.0006,
+ "loss": 4.534899711608887,
+ "step": 1365
+ },
+ {
+ "epoch": 18.978165938864628,
+ "grad_norm": 0.017693081870675087,
+ "learning_rate": 0.0006,
+ "loss": 4.460553169250488,
+ "step": 1366
+ },
+ {
+ "epoch": 18.992139737991266,
+ "grad_norm": 0.018746308982372284,
+ "learning_rate": 0.0006,
+ "loss": 4.538792610168457,
+ "step": 1367
+ },
+ {
+ "epoch": 19.0,
+ "grad_norm": 0.020099416375160217,
+ "learning_rate": 0.0006,
+ "loss": 4.449617385864258,
+ "step": 1368
+ },
+ {
+ "epoch": 19.0,
+ "eval_loss": 4.7756571769714355,
+ "eval_runtime": 57.1804,
+ "eval_samples_per_second": 42.707,
+ "eval_steps_per_second": 1.347,
+ "step": 1368
+ },
+ {
+ "epoch": 19.01397379912664,
+ "grad_norm": 0.018726933747529984,
+ "learning_rate": 0.0006,
+ "loss": 4.446059703826904,
+ "step": 1369
+ },
+ {
+ "epoch": 19.027947598253274,
+ "grad_norm": 0.01956087350845337,
+ "learning_rate": 0.0006,
+ "loss": 4.4520039558410645,
+ "step": 1370
+ },
+ {
+ "epoch": 19.041921397379912,
+ "grad_norm": 0.020092811435461044,
+ "learning_rate": 0.0006,
+ "loss": 4.543343544006348,
+ "step": 1371
+ },
+ {
+ "epoch": 19.05589519650655,
+ "grad_norm": 0.01843622885644436,
+ "learning_rate": 0.0006,
+ "loss": 4.535579204559326,
+ "step": 1372
+ },
+ {
+ "epoch": 19.069868995633186,
+ "grad_norm": 0.016989488154649734,
+ "learning_rate": 0.0006,
+ "loss": 4.52316951751709,
+ "step": 1373
+ },
+ {
+ "epoch": 19.083842794759825,
+ "grad_norm": 0.01805422455072403,
+ "learning_rate": 0.0006,
+ "loss": 4.5393266677856445,
+ "step": 1374
+ },
+ {
+ "epoch": 19.097816593886463,
+ "grad_norm": 0.01874629408121109,
+ "learning_rate": 0.0006,
+ "loss": 4.422059059143066,
+ "step": 1375
+ },
+ {
+ "epoch": 19.111790393013102,
+ "grad_norm": 0.01804221048951149,
+ "learning_rate": 0.0006,
+ "loss": 4.50508451461792,
+ "step": 1376
+ },
+ {
+ "epoch": 19.125764192139737,
+ "grad_norm": 0.01785474829375744,
+ "learning_rate": 0.0006,
+ "loss": 4.412599563598633,
+ "step": 1377
+ },
+ {
+ "epoch": 19.139737991266376,
+ "grad_norm": 0.01971791312098503,
+ "learning_rate": 0.0006,
+ "loss": 4.518318176269531,
+ "step": 1378
+ },
+ {
+ "epoch": 19.153711790393015,
+ "grad_norm": 0.02061633951961994,
+ "learning_rate": 0.0006,
+ "loss": 4.332393169403076,
+ "step": 1379
+ },
+ {
+ "epoch": 19.16768558951965,
+ "grad_norm": 0.020563429221510887,
+ "learning_rate": 0.0006,
+ "loss": 4.445863723754883,
+ "step": 1380
+ },
+ {
+ "epoch": 19.18165938864629,
+ "grad_norm": 0.019995713606476784,
+ "learning_rate": 0.0006,
+ "loss": 4.497312068939209,
+ "step": 1381
+ },
+ {
+ "epoch": 19.195633187772927,
+ "grad_norm": 0.019823070615530014,
+ "learning_rate": 0.0006,
+ "loss": 4.475085258483887,
+ "step": 1382
+ },
+ {
+ "epoch": 19.209606986899562,
+ "grad_norm": 0.019039355218410492,
+ "learning_rate": 0.0006,
+ "loss": 4.50432014465332,
+ "step": 1383
+ },
+ {
+ "epoch": 19.2235807860262,
+ "grad_norm": 0.017876973375678062,
+ "learning_rate": 0.0006,
+ "loss": 4.384788513183594,
+ "step": 1384
+ },
+ {
+ "epoch": 19.23755458515284,
+ "grad_norm": 0.018697699531912804,
+ "learning_rate": 0.0006,
+ "loss": 4.410778045654297,
+ "step": 1385
+ },
+ {
+ "epoch": 19.251528384279474,
+ "grad_norm": 0.018826806917786598,
+ "learning_rate": 0.0006,
+ "loss": 4.524034023284912,
+ "step": 1386
+ },
+ {
+ "epoch": 19.265502183406113,
+ "grad_norm": 0.01824226602911949,
+ "learning_rate": 0.0006,
+ "loss": 4.526393890380859,
+ "step": 1387
+ },
+ {
+ "epoch": 19.27947598253275,
+ "grad_norm": 0.01753208599984646,
+ "learning_rate": 0.0006,
+ "loss": 4.493703365325928,
+ "step": 1388
+ },
+ {
+ "epoch": 19.29344978165939,
+ "grad_norm": 0.01742432825267315,
+ "learning_rate": 0.0006,
+ "loss": 4.392831802368164,
+ "step": 1389
+ },
+ {
+ "epoch": 19.307423580786025,
+ "grad_norm": 0.017099356278777122,
+ "learning_rate": 0.0006,
+ "loss": 4.578668594360352,
+ "step": 1390
+ },
+ {
+ "epoch": 19.321397379912664,
+ "grad_norm": 0.017946023494005203,
+ "learning_rate": 0.0006,
+ "loss": 4.456718444824219,
+ "step": 1391
+ },
+ {
+ "epoch": 19.335371179039303,
+ "grad_norm": 0.017705943435430527,
+ "learning_rate": 0.0006,
+ "loss": 4.553682327270508,
+ "step": 1392
+ },
+ {
+ "epoch": 19.349344978165938,
+ "grad_norm": 0.016647523269057274,
+ "learning_rate": 0.0006,
+ "loss": 4.525016784667969,
+ "step": 1393
+ },
+ {
+ "epoch": 19.363318777292577,
+ "grad_norm": 0.015407416969537735,
+ "learning_rate": 0.0006,
+ "loss": 4.500231742858887,
+ "step": 1394
+ },
+ {
+ "epoch": 19.377292576419215,
+ "grad_norm": 0.017264019697904587,
+ "learning_rate": 0.0006,
+ "loss": 4.4615349769592285,
+ "step": 1395
+ },
+ {
+ "epoch": 19.39126637554585,
+ "grad_norm": 0.01871555484831333,
+ "learning_rate": 0.0006,
+ "loss": 4.281026840209961,
+ "step": 1396
+ },
+ {
+ "epoch": 19.40524017467249,
+ "grad_norm": 0.017361650243401527,
+ "learning_rate": 0.0006,
+ "loss": 4.479382514953613,
+ "step": 1397
+ },
+ {
+ "epoch": 19.419213973799128,
+ "grad_norm": 0.019781362265348434,
+ "learning_rate": 0.0006,
+ "loss": 4.402320861816406,
+ "step": 1398
+ },
+ {
+ "epoch": 19.433187772925763,
+ "grad_norm": 0.019043298438191414,
+ "learning_rate": 0.0006,
+ "loss": 4.476983070373535,
+ "step": 1399
+ },
+ {
+ "epoch": 19.4471615720524,
+ "grad_norm": 0.017137985676527023,
+ "learning_rate": 0.0006,
+ "loss": 4.418689250946045,
+ "step": 1400
+ },
+ {
+ "epoch": 19.46113537117904,
+ "grad_norm": 0.016481833532452583,
+ "learning_rate": 0.0006,
+ "loss": 4.579145431518555,
+ "step": 1401
+ },
+ {
+ "epoch": 19.475109170305675,
+ "grad_norm": 0.016507714986801147,
+ "learning_rate": 0.0006,
+ "loss": 4.519158363342285,
+ "step": 1402
+ },
+ {
+ "epoch": 19.489082969432314,
+ "grad_norm": 0.018092872574925423,
+ "learning_rate": 0.0006,
+ "loss": 4.513748645782471,
+ "step": 1403
+ },
+ {
+ "epoch": 19.503056768558952,
+ "grad_norm": 0.018153440207242966,
+ "learning_rate": 0.0006,
+ "loss": 4.645384311676025,
+ "step": 1404
+ },
+ {
+ "epoch": 19.51703056768559,
+ "grad_norm": 0.01657119207084179,
+ "learning_rate": 0.0006,
+ "loss": 4.399757385253906,
+ "step": 1405
+ },
+ {
+ "epoch": 19.531004366812226,
+ "grad_norm": 0.0156280305236578,
+ "learning_rate": 0.0006,
+ "loss": 4.467836380004883,
+ "step": 1406
+ },
+ {
+ "epoch": 19.544978165938865,
+ "grad_norm": 0.016838403418660164,
+ "learning_rate": 0.0006,
+ "loss": 4.529176235198975,
+ "step": 1407
+ },
+ {
+ "epoch": 19.558951965065503,
+ "grad_norm": 0.01743236929178238,
+ "learning_rate": 0.0006,
+ "loss": 4.440032482147217,
+ "step": 1408
+ },
+ {
+ "epoch": 19.57292576419214,
+ "grad_norm": 0.01553610060364008,
+ "learning_rate": 0.0006,
+ "loss": 4.425588607788086,
+ "step": 1409
+ },
+ {
+ "epoch": 19.586899563318777,
+ "grad_norm": 0.017259271815419197,
+ "learning_rate": 0.0006,
+ "loss": 4.405452728271484,
+ "step": 1410
+ },
+ {
+ "epoch": 19.600873362445416,
+ "grad_norm": 0.015568481758236885,
+ "learning_rate": 0.0006,
+ "loss": 4.444279193878174,
+ "step": 1411
+ },
+ {
+ "epoch": 19.61484716157205,
+ "grad_norm": 0.015331248752772808,
+ "learning_rate": 0.0006,
+ "loss": 4.451846122741699,
+ "step": 1412
+ },
+ {
+ "epoch": 19.62882096069869,
+ "grad_norm": 0.015667378902435303,
+ "learning_rate": 0.0006,
+ "loss": 4.533931255340576,
+ "step": 1413
+ },
+ {
+ "epoch": 19.64279475982533,
+ "grad_norm": 0.016177602112293243,
+ "learning_rate": 0.0006,
+ "loss": 4.467313289642334,
+ "step": 1414
+ },
+ {
+ "epoch": 19.656768558951963,
+ "grad_norm": 0.01733129471540451,
+ "learning_rate": 0.0006,
+ "loss": 4.496893882751465,
+ "step": 1415
+ },
+ {
+ "epoch": 19.670742358078602,
+ "grad_norm": 0.016539594158530235,
+ "learning_rate": 0.0006,
+ "loss": 4.401961326599121,
+ "step": 1416
+ },
+ {
+ "epoch": 19.68471615720524,
+ "grad_norm": 0.016058551147580147,
+ "learning_rate": 0.0006,
+ "loss": 4.503375053405762,
+ "step": 1417
+ },
+ {
+ "epoch": 19.69868995633188,
+ "grad_norm": 0.01715777814388275,
+ "learning_rate": 0.0006,
+ "loss": 4.3808369636535645,
+ "step": 1418
+ },
+ {
+ "epoch": 19.712663755458514,
+ "grad_norm": 0.01681119203567505,
+ "learning_rate": 0.0006,
+ "loss": 4.5469560623168945,
+ "step": 1419
+ },
+ {
+ "epoch": 19.726637554585153,
+ "grad_norm": 0.01634550653398037,
+ "learning_rate": 0.0006,
+ "loss": 4.328197956085205,
+ "step": 1420
+ },
+ {
+ "epoch": 19.74061135371179,
+ "grad_norm": 0.01549347210675478,
+ "learning_rate": 0.0006,
+ "loss": 4.534621238708496,
+ "step": 1421
+ },
+ {
+ "epoch": 19.754585152838427,
+ "grad_norm": 0.01654665172100067,
+ "learning_rate": 0.0006,
+ "loss": 4.4724225997924805,
+ "step": 1422
+ },
+ {
+ "epoch": 19.768558951965066,
+ "grad_norm": 0.017119025811553,
+ "learning_rate": 0.0006,
+ "loss": 4.3897247314453125,
+ "step": 1423
+ },
+ {
+ "epoch": 19.782532751091704,
+ "grad_norm": 0.01603223942220211,
+ "learning_rate": 0.0006,
+ "loss": 4.412731170654297,
+ "step": 1424
+ },
+ {
+ "epoch": 19.79650655021834,
+ "grad_norm": 0.016774067655205727,
+ "learning_rate": 0.0006,
+ "loss": 4.54707145690918,
+ "step": 1425
+ },
+ {
+ "epoch": 19.810480349344978,
+ "grad_norm": 0.016453025862574577,
+ "learning_rate": 0.0006,
+ "loss": 4.462501049041748,
+ "step": 1426
+ },
+ {
+ "epoch": 19.824454148471617,
+ "grad_norm": 0.01597629487514496,
+ "learning_rate": 0.0006,
+ "loss": 4.229758262634277,
+ "step": 1427
+ },
+ {
+ "epoch": 19.83842794759825,
+ "grad_norm": 0.017999034374952316,
+ "learning_rate": 0.0006,
+ "loss": 4.389263153076172,
+ "step": 1428
+ },
+ {
+ "epoch": 19.85240174672489,
+ "grad_norm": 0.02165898121893406,
+ "learning_rate": 0.0006,
+ "loss": 4.342130661010742,
+ "step": 1429
+ },
+ {
+ "epoch": 19.86637554585153,
+ "grad_norm": 0.02222822792828083,
+ "learning_rate": 0.0006,
+ "loss": 4.492661476135254,
+ "step": 1430
+ },
+ {
+ "epoch": 19.880349344978168,
+ "grad_norm": 0.019378066062927246,
+ "learning_rate": 0.0006,
+ "loss": 4.4614105224609375,
+ "step": 1431
+ },
+ {
+ "epoch": 19.894323144104803,
+ "grad_norm": 0.019908083602786064,
+ "learning_rate": 0.0006,
+ "loss": 4.370929718017578,
+ "step": 1432
+ },
+ {
+ "epoch": 19.90829694323144,
+ "grad_norm": 0.022296471521258354,
+ "learning_rate": 0.0006,
+ "loss": 4.438320159912109,
+ "step": 1433
+ },
+ {
+ "epoch": 19.92227074235808,
+ "grad_norm": 0.021810343489050865,
+ "learning_rate": 0.0006,
+ "loss": 4.462067127227783,
+ "step": 1434
+ },
+ {
+ "epoch": 19.936244541484715,
+ "grad_norm": 0.02129209227859974,
+ "learning_rate": 0.0006,
+ "loss": 4.541261196136475,
+ "step": 1435
+ },
+ {
+ "epoch": 19.950218340611354,
+ "grad_norm": 0.02079629711806774,
+ "learning_rate": 0.0006,
+ "loss": 4.36712646484375,
+ "step": 1436
+ },
+ {
+ "epoch": 19.964192139737992,
+ "grad_norm": 0.022501740604639053,
+ "learning_rate": 0.0006,
+ "loss": 4.5279083251953125,
+ "step": 1437
+ },
+ {
+ "epoch": 19.978165938864628,
+ "grad_norm": 0.019121676683425903,
+ "learning_rate": 0.0006,
+ "loss": 4.477789878845215,
+ "step": 1438
+ },
+ {
+ "epoch": 19.992139737991266,
+ "grad_norm": 0.017429588362574577,
+ "learning_rate": 0.0006,
+ "loss": 4.4023847579956055,
+ "step": 1439
+ },
+ {
+ "epoch": 20.0,
+ "grad_norm": 0.018857665359973907,
+ "learning_rate": 0.0006,
+ "loss": 4.494743347167969,
+ "step": 1440
+ },
+ {
+ "epoch": 20.0,
+ "eval_loss": 4.694555759429932,
+ "eval_runtime": 56.4146,
+ "eval_samples_per_second": 43.287,
+ "eval_steps_per_second": 1.365,
+ "step": 1440
+ },
+ {
+ "epoch": 20.01397379912664,
+ "grad_norm": 0.016505861654877663,
+ "learning_rate": 0.0006,
+ "loss": 4.506229877471924,
+ "step": 1441
+ },
+ {
+ "epoch": 20.027947598253274,
+ "grad_norm": 0.017571842297911644,
+ "learning_rate": 0.0006,
+ "loss": 4.384073257446289,
+ "step": 1442
+ },
+ {
+ "epoch": 20.041921397379912,
+ "grad_norm": 0.019398365169763565,
+ "learning_rate": 0.0006,
+ "loss": 4.482177734375,
+ "step": 1443
+ },
+ {
+ "epoch": 20.05589519650655,
+ "grad_norm": 0.02011062018573284,
+ "learning_rate": 0.0006,
+ "loss": 4.440773963928223,
+ "step": 1444
+ },
+ {
+ "epoch": 20.069868995633186,
+ "grad_norm": 0.019026953727006912,
+ "learning_rate": 0.0006,
+ "loss": 4.507168769836426,
+ "step": 1445
+ },
+ {
+ "epoch": 20.083842794759825,
+ "grad_norm": 0.020506031811237335,
+ "learning_rate": 0.0006,
+ "loss": 4.469079971313477,
+ "step": 1446
+ },
+ {
+ "epoch": 20.097816593886463,
+ "grad_norm": 0.019608162343502045,
+ "learning_rate": 0.0006,
+ "loss": 4.395927429199219,
+ "step": 1447
+ },
+ {
+ "epoch": 20.111790393013102,
+ "grad_norm": 0.019419124349951744,
+ "learning_rate": 0.0006,
+ "loss": 4.484235763549805,
+ "step": 1448
+ },
+ {
+ "epoch": 20.125764192139737,
+ "grad_norm": 0.021775128319859505,
+ "learning_rate": 0.0006,
+ "loss": 4.333173751831055,
+ "step": 1449
+ },
+ {
+ "epoch": 20.139737991266376,
+ "grad_norm": 0.025512272492051125,
+ "learning_rate": 0.0006,
+ "loss": 4.460362434387207,
+ "step": 1450
+ },
+ {
+ "epoch": 20.153711790393015,
+ "grad_norm": 0.025157401338219643,
+ "learning_rate": 0.0006,
+ "loss": 4.4361395835876465,
+ "step": 1451
+ },
+ {
+ "epoch": 20.16768558951965,
+ "grad_norm": 0.019862616434693336,
+ "learning_rate": 0.0006,
+ "loss": 4.426807403564453,
+ "step": 1452
+ },
+ {
+ "epoch": 20.18165938864629,
+ "grad_norm": 0.01932152360677719,
+ "learning_rate": 0.0006,
+ "loss": 4.375677108764648,
+ "step": 1453
+ },
+ {
+ "epoch": 20.195633187772927,
+ "grad_norm": 0.01760999858379364,
+ "learning_rate": 0.0006,
+ "loss": 4.42259407043457,
+ "step": 1454
+ },
+ {
+ "epoch": 20.209606986899562,
+ "grad_norm": 0.016998112201690674,
+ "learning_rate": 0.0006,
+ "loss": 4.481280326843262,
+ "step": 1455
+ },
+ {
+ "epoch": 20.2235807860262,
+ "grad_norm": 0.017199190333485603,
+ "learning_rate": 0.0006,
+ "loss": 4.496712684631348,
+ "step": 1456
+ },
+ {
+ "epoch": 20.23755458515284,
+ "grad_norm": 0.016833679750561714,
+ "learning_rate": 0.0006,
+ "loss": 4.40736722946167,
+ "step": 1457
+ },
+ {
+ "epoch": 20.251528384279474,
+ "grad_norm": 0.017546730116009712,
+ "learning_rate": 0.0006,
+ "loss": 4.4915266036987305,
+ "step": 1458
+ },
+ {
+ "epoch": 20.265502183406113,
+ "grad_norm": 0.016595078632235527,
+ "learning_rate": 0.0006,
+ "loss": 4.460815906524658,
+ "step": 1459
+ },
+ {
+ "epoch": 20.27947598253275,
+ "grad_norm": 0.017728568986058235,
+ "learning_rate": 0.0006,
+ "loss": 4.346048831939697,
+ "step": 1460
+ },
+ {
+ "epoch": 20.29344978165939,
+ "grad_norm": 0.01849197782576084,
+ "learning_rate": 0.0006,
+ "loss": 4.399681091308594,
+ "step": 1461
+ },
+ {
+ "epoch": 20.307423580786025,
+ "grad_norm": 0.018336588516831398,
+ "learning_rate": 0.0006,
+ "loss": 4.414183616638184,
+ "step": 1462
+ },
+ {
+ "epoch": 20.321397379912664,
+ "grad_norm": 0.018086519092321396,
+ "learning_rate": 0.0006,
+ "loss": 4.346586227416992,
+ "step": 1463
+ },
+ {
+ "epoch": 20.335371179039303,
+ "grad_norm": 0.019697299227118492,
+ "learning_rate": 0.0006,
+ "loss": 4.4420952796936035,
+ "step": 1464
+ },
+ {
+ "epoch": 20.349344978165938,
+ "grad_norm": 0.02198074571788311,
+ "learning_rate": 0.0006,
+ "loss": 4.588729381561279,
+ "step": 1465
+ },
+ {
+ "epoch": 20.363318777292577,
+ "grad_norm": 0.022813329473137856,
+ "learning_rate": 0.0006,
+ "loss": 4.49729585647583,
+ "step": 1466
+ },
+ {
+ "epoch": 20.377292576419215,
+ "grad_norm": 0.022006649523973465,
+ "learning_rate": 0.0006,
+ "loss": 4.475641250610352,
+ "step": 1467
+ },
+ {
+ "epoch": 20.39126637554585,
+ "grad_norm": 0.023188438266515732,
+ "learning_rate": 0.0006,
+ "loss": 4.404322624206543,
+ "step": 1468
+ },
+ {
+ "epoch": 20.40524017467249,
+ "grad_norm": 0.022226877510547638,
+ "learning_rate": 0.0006,
+ "loss": 4.50904655456543,
+ "step": 1469
+ },
+ {
+ "epoch": 20.419213973799128,
+ "grad_norm": 0.021082637831568718,
+ "learning_rate": 0.0006,
+ "loss": 4.339011192321777,
+ "step": 1470
+ },
+ {
+ "epoch": 20.433187772925763,
+ "grad_norm": 0.02013542130589485,
+ "learning_rate": 0.0006,
+ "loss": 4.516942024230957,
+ "step": 1471
+ },
+ {
+ "epoch": 20.4471615720524,
+ "grad_norm": 0.01898968778550625,
+ "learning_rate": 0.0006,
+ "loss": 4.315804481506348,
+ "step": 1472
+ },
+ {
+ "epoch": 20.46113537117904,
+ "grad_norm": 0.019450880587100983,
+ "learning_rate": 0.0006,
+ "loss": 4.4794921875,
+ "step": 1473
+ },
+ {
+ "epoch": 20.475109170305675,
+ "grad_norm": 0.018577802926301956,
+ "learning_rate": 0.0006,
+ "loss": 4.422323703765869,
+ "step": 1474
+ },
+ {
+ "epoch": 20.489082969432314,
+ "grad_norm": 0.01778412237763405,
+ "learning_rate": 0.0006,
+ "loss": 4.429789066314697,
+ "step": 1475
+ },
+ {
+ "epoch": 20.503056768558952,
+ "grad_norm": 0.017134476453065872,
+ "learning_rate": 0.0006,
+ "loss": 4.416459083557129,
+ "step": 1476
+ },
+ {
+ "epoch": 20.51703056768559,
+ "grad_norm": 0.017153557389974594,
+ "learning_rate": 0.0006,
+ "loss": 4.446170806884766,
+ "step": 1477
+ },
+ {
+ "epoch": 20.531004366812226,
+ "grad_norm": 0.014998532831668854,
+ "learning_rate": 0.0006,
+ "loss": 4.419447898864746,
+ "step": 1478
+ },
+ {
+ "epoch": 20.544978165938865,
+ "grad_norm": 0.015877850353717804,
+ "learning_rate": 0.0006,
+ "loss": 4.312338829040527,
+ "step": 1479
+ },
+ {
+ "epoch": 20.558951965065503,
+ "grad_norm": 0.01569489948451519,
+ "learning_rate": 0.0006,
+ "loss": 4.572905540466309,
+ "step": 1480
+ },
+ {
+ "epoch": 20.57292576419214,
+ "grad_norm": 0.01591474376618862,
+ "learning_rate": 0.0006,
+ "loss": 4.414907455444336,
+ "step": 1481
+ },
+ {
+ "epoch": 20.586899563318777,
+ "grad_norm": 0.015395903028547764,
+ "learning_rate": 0.0006,
+ "loss": 4.549131393432617,
+ "step": 1482
+ },
+ {
+ "epoch": 20.600873362445416,
+ "grad_norm": 0.015715358778834343,
+ "learning_rate": 0.0006,
+ "loss": 4.530194282531738,
+ "step": 1483
+ },
+ {
+ "epoch": 20.61484716157205,
+ "grad_norm": 0.016042975708842278,
+ "learning_rate": 0.0006,
+ "loss": 4.322149276733398,
+ "step": 1484
+ },
+ {
+ "epoch": 20.62882096069869,
+ "grad_norm": 0.01664186641573906,
+ "learning_rate": 0.0006,
+ "loss": 4.424866676330566,
+ "step": 1485
+ },
+ {
+ "epoch": 20.64279475982533,
+ "grad_norm": 0.01631534844636917,
+ "learning_rate": 0.0006,
+ "loss": 4.441752910614014,
+ "step": 1486
+ },
+ {
+ "epoch": 20.656768558951963,
+ "grad_norm": 0.014057058840990067,
+ "learning_rate": 0.0006,
+ "loss": 4.399982452392578,
+ "step": 1487
+ },
+ {
+ "epoch": 20.670742358078602,
+ "grad_norm": 0.015076273120939732,
+ "learning_rate": 0.0006,
+ "loss": 4.441324234008789,
+ "step": 1488
+ },
+ {
+ "epoch": 20.68471615720524,
+ "grad_norm": 0.014138277620077133,
+ "learning_rate": 0.0006,
+ "loss": 4.4312543869018555,
+ "step": 1489
+ },
+ {
+ "epoch": 20.69868995633188,
+ "grad_norm": 0.013762440532445908,
+ "learning_rate": 0.0006,
+ "loss": 4.387807846069336,
+ "step": 1490
+ },
+ {
+ "epoch": 20.712663755458514,
+ "grad_norm": 0.015058411285281181,
+ "learning_rate": 0.0006,
+ "loss": 4.594747543334961,
+ "step": 1491
+ },
+ {
+ "epoch": 20.726637554585153,
+ "grad_norm": 0.013827123679220676,
+ "learning_rate": 0.0006,
+ "loss": 4.434604167938232,
+ "step": 1492
+ },
+ {
+ "epoch": 20.74061135371179,
+ "grad_norm": 0.015315295197069645,
+ "learning_rate": 0.0006,
+ "loss": 4.588066101074219,
+ "step": 1493
+ },
+ {
+ "epoch": 20.754585152838427,
+ "grad_norm": 0.018678732216358185,
+ "learning_rate": 0.0006,
+ "loss": 4.3514204025268555,
+ "step": 1494
+ },
+ {
+ "epoch": 20.768558951965066,
+ "grad_norm": 0.019521350041031837,
+ "learning_rate": 0.0006,
+ "loss": 4.437658786773682,
+ "step": 1495
+ },
+ {
+ "epoch": 20.782532751091704,
+ "grad_norm": 0.018363403156399727,
+ "learning_rate": 0.0006,
+ "loss": 4.495002269744873,
+ "step": 1496
+ },
+ {
+ "epoch": 20.79650655021834,
+ "grad_norm": 0.016830725595355034,
+ "learning_rate": 0.0006,
+ "loss": 4.402471542358398,
+ "step": 1497
+ },
+ {
+ "epoch": 20.810480349344978,
+ "grad_norm": 0.01722072623670101,
+ "learning_rate": 0.0006,
+ "loss": 4.437653541564941,
+ "step": 1498
+ },
+ {
+ "epoch": 20.824454148471617,
+ "grad_norm": 0.017459599301218987,
+ "learning_rate": 0.0006,
+ "loss": 4.366603851318359,
+ "step": 1499
+ },
+ {
+ "epoch": 20.83842794759825,
+ "grad_norm": 0.018902219831943512,
+ "learning_rate": 0.0006,
+ "loss": 4.501561164855957,
+ "step": 1500
+ },
+ {
+ "epoch": 20.85240174672489,
+ "grad_norm": 0.020814096555113792,
+ "learning_rate": 0.0006,
+ "loss": 4.535295486450195,
+ "step": 1501
+ },
+ {
+ "epoch": 20.86637554585153,
+ "grad_norm": 0.018990976735949516,
+ "learning_rate": 0.0006,
+ "loss": 4.333580017089844,
+ "step": 1502
+ },
+ {
+ "epoch": 20.880349344978168,
+ "grad_norm": 0.01837880350649357,
+ "learning_rate": 0.0006,
+ "loss": 4.313388824462891,
+ "step": 1503
+ },
+ {
+ "epoch": 20.894323144104803,
+ "grad_norm": 0.019820360466837883,
+ "learning_rate": 0.0006,
+ "loss": 4.388157844543457,
+ "step": 1504
+ },
+ {
+ "epoch": 20.90829694323144,
+ "grad_norm": 0.020181143656373024,
+ "learning_rate": 0.0006,
+ "loss": 4.572224140167236,
+ "step": 1505
+ },
+ {
+ "epoch": 20.92227074235808,
+ "grad_norm": 0.021206077188253403,
+ "learning_rate": 0.0006,
+ "loss": 4.3127970695495605,
+ "step": 1506
+ },
+ {
+ "epoch": 20.936244541484715,
+ "grad_norm": 0.017962060868740082,
+ "learning_rate": 0.0006,
+ "loss": 4.384734630584717,
+ "step": 1507
+ },
+ {
+ "epoch": 20.950218340611354,
+ "grad_norm": 0.01737220585346222,
+ "learning_rate": 0.0006,
+ "loss": 4.4075751304626465,
+ "step": 1508
+ },
+ {
+ "epoch": 20.964192139737992,
+ "grad_norm": 0.01839268207550049,
+ "learning_rate": 0.0006,
+ "loss": 4.553779125213623,
+ "step": 1509
+ },
+ {
+ "epoch": 20.978165938864628,
+ "grad_norm": 0.020102640613913536,
+ "learning_rate": 0.0006,
+ "loss": 4.399398326873779,
+ "step": 1510
+ },
+ {
+ "epoch": 20.992139737991266,
+ "grad_norm": 0.018831463530659676,
+ "learning_rate": 0.0006,
+ "loss": 4.358499526977539,
+ "step": 1511
+ },
+ {
+ "epoch": 21.0,
+ "grad_norm": 0.018254734575748444,
+ "learning_rate": 0.0006,
+ "loss": 4.384119510650635,
+ "step": 1512
+ },
+ {
+ "epoch": 21.0,
+ "eval_loss": 4.7524237632751465,
+ "eval_runtime": 57.8812,
+ "eval_samples_per_second": 42.19,
+ "eval_steps_per_second": 1.33,
+ "step": 1512
+ },
+ {
+ "epoch": 21.01397379912664,
+ "grad_norm": 0.019728587940335274,
+ "learning_rate": 0.0006,
+ "loss": 4.32363224029541,
+ "step": 1513
+ },
+ {
+ "epoch": 21.027947598253274,
+ "grad_norm": 0.02216625027358532,
+ "learning_rate": 0.0006,
+ "loss": 4.436988353729248,
+ "step": 1514
+ },
+ {
+ "epoch": 21.041921397379912,
+ "grad_norm": 0.02227606810629368,
+ "learning_rate": 0.0006,
+ "loss": 4.406063556671143,
+ "step": 1515
+ },
+ {
+ "epoch": 21.05589519650655,
+ "grad_norm": 0.020187009125947952,
+ "learning_rate": 0.0006,
+ "loss": 4.324402332305908,
+ "step": 1516
+ },
+ {
+ "epoch": 21.069868995633186,
+ "grad_norm": 0.02095157466828823,
+ "learning_rate": 0.0006,
+ "loss": 4.456544876098633,
+ "step": 1517
+ },
+ {
+ "epoch": 21.083842794759825,
+ "grad_norm": 0.02146013453602791,
+ "learning_rate": 0.0006,
+ "loss": 4.368387222290039,
+ "step": 1518
+ },
+ {
+ "epoch": 21.097816593886463,
+ "grad_norm": 0.02162035182118416,
+ "learning_rate": 0.0006,
+ "loss": 4.48838996887207,
+ "step": 1519
+ },
+ {
+ "epoch": 21.111790393013102,
+ "grad_norm": 0.021745676174759865,
+ "learning_rate": 0.0006,
+ "loss": 4.30778694152832,
+ "step": 1520
+ },
+ {
+ "epoch": 21.125764192139737,
+ "grad_norm": 0.02173588052392006,
+ "learning_rate": 0.0006,
+ "loss": 4.433149814605713,
+ "step": 1521
+ },
+ {
+ "epoch": 21.139737991266376,
+ "grad_norm": 0.020950648933649063,
+ "learning_rate": 0.0006,
+ "loss": 4.520089149475098,
+ "step": 1522
+ },
+ {
+ "epoch": 21.153711790393015,
+ "grad_norm": 0.020684752613306046,
+ "learning_rate": 0.0006,
+ "loss": 4.372437477111816,
+ "step": 1523
+ },
+ {
+ "epoch": 21.16768558951965,
+ "grad_norm": 0.019472869113087654,
+ "learning_rate": 0.0006,
+ "loss": 4.437015533447266,
+ "step": 1524
+ },
+ {
+ "epoch": 21.18165938864629,
+ "grad_norm": 0.01846962980926037,
+ "learning_rate": 0.0006,
+ "loss": 4.449531555175781,
+ "step": 1525
+ },
+ {
+ "epoch": 21.195633187772927,
+ "grad_norm": 0.017377803102135658,
+ "learning_rate": 0.0006,
+ "loss": 4.458279132843018,
+ "step": 1526
+ },
+ {
+ "epoch": 21.209606986899562,
+ "grad_norm": 0.016002200543880463,
+ "learning_rate": 0.0006,
+ "loss": 4.266297817230225,
+ "step": 1527
+ },
+ {
+ "epoch": 21.2235807860262,
+ "grad_norm": 0.015575903467833996,
+ "learning_rate": 0.0006,
+ "loss": 4.342632293701172,
+ "step": 1528
+ },
+ {
+ "epoch": 21.23755458515284,
+ "grad_norm": 0.015634698793292046,
+ "learning_rate": 0.0006,
+ "loss": 4.338623523712158,
+ "step": 1529
+ },
+ {
+ "epoch": 21.251528384279474,
+ "grad_norm": 0.01449581328779459,
+ "learning_rate": 0.0006,
+ "loss": 4.354330062866211,
+ "step": 1530
+ },
+ {
+ "epoch": 21.265502183406113,
+ "grad_norm": 0.01487616915255785,
+ "learning_rate": 0.0006,
+ "loss": 4.486835479736328,
+ "step": 1531
+ },
+ {
+ "epoch": 21.27947598253275,
+ "grad_norm": 0.016346555203199387,
+ "learning_rate": 0.0006,
+ "loss": 4.407098770141602,
+ "step": 1532
+ },
+ {
+ "epoch": 21.29344978165939,
+ "grad_norm": 0.01682182401418686,
+ "learning_rate": 0.0006,
+ "loss": 4.436809539794922,
+ "step": 1533
+ },
+ {
+ "epoch": 21.307423580786025,
+ "grad_norm": 0.017704004421830177,
+ "learning_rate": 0.0006,
+ "loss": 4.3854780197143555,
+ "step": 1534
+ },
+ {
+ "epoch": 21.321397379912664,
+ "grad_norm": 0.018437420949339867,
+ "learning_rate": 0.0006,
+ "loss": 4.413283824920654,
+ "step": 1535
+ },
+ {
+ "epoch": 21.335371179039303,
+ "grad_norm": 0.016911081969738007,
+ "learning_rate": 0.0006,
+ "loss": 4.357306003570557,
+ "step": 1536
+ },
+ {
+ "epoch": 21.349344978165938,
+ "grad_norm": 0.01642223633825779,
+ "learning_rate": 0.0006,
+ "loss": 4.422322750091553,
+ "step": 1537
+ },
+ {
+ "epoch": 21.363318777292577,
+ "grad_norm": 0.01780877821147442,
+ "learning_rate": 0.0006,
+ "loss": 4.387620449066162,
+ "step": 1538
+ },
+ {
+ "epoch": 21.377292576419215,
+ "grad_norm": 0.01653975248336792,
+ "learning_rate": 0.0006,
+ "loss": 4.327267646789551,
+ "step": 1539
+ },
+ {
+ "epoch": 21.39126637554585,
+ "grad_norm": 0.01572156324982643,
+ "learning_rate": 0.0006,
+ "loss": 4.475979804992676,
+ "step": 1540
+ },
+ {
+ "epoch": 21.40524017467249,
+ "grad_norm": 0.017156066372990608,
+ "learning_rate": 0.0006,
+ "loss": 4.4011125564575195,
+ "step": 1541
+ },
+ {
+ "epoch": 21.419213973799128,
+ "grad_norm": 0.01788829080760479,
+ "learning_rate": 0.0006,
+ "loss": 4.4174089431762695,
+ "step": 1542
+ },
+ {
+ "epoch": 21.433187772925763,
+ "grad_norm": 0.01728985086083412,
+ "learning_rate": 0.0006,
+ "loss": 4.429244041442871,
+ "step": 1543
+ },
+ {
+ "epoch": 21.4471615720524,
+ "grad_norm": 0.016534466296434402,
+ "learning_rate": 0.0006,
+ "loss": 4.326068878173828,
+ "step": 1544
+ },
+ {
+ "epoch": 21.46113537117904,
+ "grad_norm": 0.016582584008574486,
+ "learning_rate": 0.0006,
+ "loss": 4.2975969314575195,
+ "step": 1545
+ },
+ {
+ "epoch": 21.475109170305675,
+ "grad_norm": 0.018034178763628006,
+ "learning_rate": 0.0006,
+ "loss": 4.501319408416748,
+ "step": 1546
+ },
+ {
+ "epoch": 21.489082969432314,
+ "grad_norm": 0.017827702686190605,
+ "learning_rate": 0.0006,
+ "loss": 4.502284526824951,
+ "step": 1547
+ },
+ {
+ "epoch": 21.503056768558952,
+ "grad_norm": 0.017182299867272377,
+ "learning_rate": 0.0006,
+ "loss": 4.380828857421875,
+ "step": 1548
+ },
+ {
+ "epoch": 21.51703056768559,
+ "grad_norm": 0.019101588055491447,
+ "learning_rate": 0.0006,
+ "loss": 4.31564998626709,
+ "step": 1549
+ },
+ {
+ "epoch": 21.531004366812226,
+ "grad_norm": 0.0204677302390337,
+ "learning_rate": 0.0006,
+ "loss": 4.3840131759643555,
+ "step": 1550
+ },
+ {
+ "epoch": 21.544978165938865,
+ "grad_norm": 0.0188993439078331,
+ "learning_rate": 0.0006,
+ "loss": 4.4257731437683105,
+ "step": 1551
+ },
+ {
+ "epoch": 21.558951965065503,
+ "grad_norm": 0.01730230636894703,
+ "learning_rate": 0.0006,
+ "loss": 4.445132255554199,
+ "step": 1552
+ },
+ {
+ "epoch": 21.57292576419214,
+ "grad_norm": 0.01801574043929577,
+ "learning_rate": 0.0006,
+ "loss": 4.427064895629883,
+ "step": 1553
+ },
+ {
+ "epoch": 21.586899563318777,
+ "grad_norm": 0.018417565152049065,
+ "learning_rate": 0.0006,
+ "loss": 4.319595813751221,
+ "step": 1554
+ },
+ {
+ "epoch": 21.600873362445416,
+ "grad_norm": 0.01683160290122032,
+ "learning_rate": 0.0006,
+ "loss": 4.349218368530273,
+ "step": 1555
+ },
+ {
+ "epoch": 21.61484716157205,
+ "grad_norm": 0.016973106190562248,
+ "learning_rate": 0.0006,
+ "loss": 4.375990390777588,
+ "step": 1556
+ },
+ {
+ "epoch": 21.62882096069869,
+ "grad_norm": 0.015688583254814148,
+ "learning_rate": 0.0006,
+ "loss": 4.408170223236084,
+ "step": 1557
+ },
+ {
+ "epoch": 21.64279475982533,
+ "grad_norm": 0.01573043316602707,
+ "learning_rate": 0.0006,
+ "loss": 4.416255950927734,
+ "step": 1558
+ },
+ {
+ "epoch": 21.656768558951963,
+ "grad_norm": 0.01744293048977852,
+ "learning_rate": 0.0006,
+ "loss": 4.477141380310059,
+ "step": 1559
+ },
+ {
+ "epoch": 21.670742358078602,
+ "grad_norm": 0.017503513023257256,
+ "learning_rate": 0.0006,
+ "loss": 4.3367486000061035,
+ "step": 1560
+ },
+ {
+ "epoch": 21.68471615720524,
+ "grad_norm": 0.016592150554060936,
+ "learning_rate": 0.0006,
+ "loss": 4.314785480499268,
+ "step": 1561
+ },
+ {
+ "epoch": 21.69868995633188,
+ "grad_norm": 0.015774084255099297,
+ "learning_rate": 0.0006,
+ "loss": 4.448209285736084,
+ "step": 1562
+ },
+ {
+ "epoch": 21.712663755458514,
+ "grad_norm": 0.015779150649905205,
+ "learning_rate": 0.0006,
+ "loss": 4.392690658569336,
+ "step": 1563
+ },
+ {
+ "epoch": 21.726637554585153,
+ "grad_norm": 0.015392648056149483,
+ "learning_rate": 0.0006,
+ "loss": 4.437741756439209,
+ "step": 1564
+ },
+ {
+ "epoch": 21.74061135371179,
+ "grad_norm": 0.01530836895108223,
+ "learning_rate": 0.0006,
+ "loss": 4.424595832824707,
+ "step": 1565
+ },
+ {
+ "epoch": 21.754585152838427,
+ "grad_norm": 0.015600779093801975,
+ "learning_rate": 0.0006,
+ "loss": 4.335417747497559,
+ "step": 1566
+ },
+ {
+ "epoch": 21.768558951965066,
+ "grad_norm": 0.014646518044173717,
+ "learning_rate": 0.0006,
+ "loss": 4.279081344604492,
+ "step": 1567
+ },
+ {
+ "epoch": 21.782532751091704,
+ "grad_norm": 0.014132864773273468,
+ "learning_rate": 0.0006,
+ "loss": 4.345950126647949,
+ "step": 1568
+ },
+ {
+ "epoch": 21.79650655021834,
+ "grad_norm": 0.01425724383443594,
+ "learning_rate": 0.0006,
+ "loss": 4.277588844299316,
+ "step": 1569
+ },
+ {
+ "epoch": 21.810480349344978,
+ "grad_norm": 0.015396572649478912,
+ "learning_rate": 0.0006,
+ "loss": 4.326196670532227,
+ "step": 1570
+ },
+ {
+ "epoch": 21.824454148471617,
+ "grad_norm": 0.01683669537305832,
+ "learning_rate": 0.0006,
+ "loss": 4.3294172286987305,
+ "step": 1571
+ },
+ {
+ "epoch": 21.83842794759825,
+ "grad_norm": 0.016365647315979004,
+ "learning_rate": 0.0006,
+ "loss": 4.523321628570557,
+ "step": 1572
+ },
+ {
+ "epoch": 21.85240174672489,
+ "grad_norm": 0.017142774537205696,
+ "learning_rate": 0.0006,
+ "loss": 4.475842475891113,
+ "step": 1573
+ },
+ {
+ "epoch": 21.86637554585153,
+ "grad_norm": 0.015950839966535568,
+ "learning_rate": 0.0006,
+ "loss": 4.397187232971191,
+ "step": 1574
+ },
+ {
+ "epoch": 21.880349344978168,
+ "grad_norm": 0.016180362552404404,
+ "learning_rate": 0.0006,
+ "loss": 4.414778709411621,
+ "step": 1575
+ },
+ {
+ "epoch": 21.894323144104803,
+ "grad_norm": 0.017615556716918945,
+ "learning_rate": 0.0006,
+ "loss": 4.348567962646484,
+ "step": 1576
+ },
+ {
+ "epoch": 21.90829694323144,
+ "grad_norm": 0.01896277628839016,
+ "learning_rate": 0.0006,
+ "loss": 4.431762218475342,
+ "step": 1577
+ },
+ {
+ "epoch": 21.92227074235808,
+ "grad_norm": 0.01990971714258194,
+ "learning_rate": 0.0006,
+ "loss": 4.431285858154297,
+ "step": 1578
+ },
+ {
+ "epoch": 21.936244541484715,
+ "grad_norm": 0.0201258547604084,
+ "learning_rate": 0.0006,
+ "loss": 4.309872627258301,
+ "step": 1579
+ },
+ {
+ "epoch": 21.950218340611354,
+ "grad_norm": 0.019620198756456375,
+ "learning_rate": 0.0006,
+ "loss": 4.484044551849365,
+ "step": 1580
+ },
+ {
+ "epoch": 21.964192139737992,
+ "grad_norm": 0.01771123707294464,
+ "learning_rate": 0.0006,
+ "loss": 4.368093013763428,
+ "step": 1581
+ },
+ {
+ "epoch": 21.978165938864628,
+ "grad_norm": 0.017501654103398323,
+ "learning_rate": 0.0006,
+ "loss": 4.372651100158691,
+ "step": 1582
+ },
+ {
+ "epoch": 21.992139737991266,
+ "grad_norm": 0.01693608984351158,
+ "learning_rate": 0.0006,
+ "loss": 4.379167556762695,
+ "step": 1583
+ },
+ {
+ "epoch": 22.0,
+ "grad_norm": 0.02002849616110325,
+ "learning_rate": 0.0006,
+ "loss": 4.329627990722656,
+ "step": 1584
+ },
+ {
+ "epoch": 22.0,
+ "eval_loss": 4.723696231842041,
+ "eval_runtime": 56.1758,
+ "eval_samples_per_second": 43.471,
+ "eval_steps_per_second": 1.371,
+ "step": 1584
+ },
+ {
+ "epoch": 22.01397379912664,
+ "grad_norm": 0.02030305378139019,
+ "learning_rate": 0.0006,
+ "loss": 4.355566024780273,
+ "step": 1585
+ },
+ {
+ "epoch": 22.027947598253274,
+ "grad_norm": 0.018638934940099716,
+ "learning_rate": 0.0006,
+ "loss": 4.314009189605713,
+ "step": 1586
+ },
+ {
+ "epoch": 22.041921397379912,
+ "grad_norm": 0.019917072728276253,
+ "learning_rate": 0.0006,
+ "loss": 4.367525577545166,
+ "step": 1587
+ },
+ {
+ "epoch": 22.05589519650655,
+ "grad_norm": 0.020785167813301086,
+ "learning_rate": 0.0006,
+ "loss": 4.375561714172363,
+ "step": 1588
+ },
+ {
+ "epoch": 22.069868995633186,
+ "grad_norm": 0.021214453503489494,
+ "learning_rate": 0.0006,
+ "loss": 4.418305397033691,
+ "step": 1589
+ },
+ {
+ "epoch": 22.083842794759825,
+ "grad_norm": 0.019051678478717804,
+ "learning_rate": 0.0006,
+ "loss": 4.215126991271973,
+ "step": 1590
+ },
+ {
+ "epoch": 22.097816593886463,
+ "grad_norm": 0.02323366142809391,
+ "learning_rate": 0.0006,
+ "loss": 4.306807994842529,
+ "step": 1591
+ },
+ {
+ "epoch": 22.111790393013102,
+ "grad_norm": 0.026186738163232803,
+ "learning_rate": 0.0006,
+ "loss": 4.408508777618408,
+ "step": 1592
+ },
+ {
+ "epoch": 22.125764192139737,
+ "grad_norm": 0.027361227199435234,
+ "learning_rate": 0.0006,
+ "loss": 4.329322814941406,
+ "step": 1593
+ },
+ {
+ "epoch": 22.139737991266376,
+ "grad_norm": 0.026564646512269974,
+ "learning_rate": 0.0006,
+ "loss": 4.346642017364502,
+ "step": 1594
+ },
+ {
+ "epoch": 22.153711790393015,
+ "grad_norm": 0.025073667988181114,
+ "learning_rate": 0.0006,
+ "loss": 4.3984198570251465,
+ "step": 1595
+ },
+ {
+ "epoch": 22.16768558951965,
+ "grad_norm": 0.02347794733941555,
+ "learning_rate": 0.0006,
+ "loss": 4.352538108825684,
+ "step": 1596
+ },
+ {
+ "epoch": 22.18165938864629,
+ "grad_norm": 0.02203468047082424,
+ "learning_rate": 0.0006,
+ "loss": 4.298148155212402,
+ "step": 1597
+ },
+ {
+ "epoch": 22.195633187772927,
+ "grad_norm": 0.020621957257390022,
+ "learning_rate": 0.0006,
+ "loss": 4.480309963226318,
+ "step": 1598
+ },
+ {
+ "epoch": 22.209606986899562,
+ "grad_norm": 0.023710055276751518,
+ "learning_rate": 0.0006,
+ "loss": 4.339317321777344,
+ "step": 1599
+ },
+ {
+ "epoch": 22.2235807860262,
+ "grad_norm": 0.0233482476323843,
+ "learning_rate": 0.0006,
+ "loss": 4.436896324157715,
+ "step": 1600
+ },
+ {
+ "epoch": 22.23755458515284,
+ "grad_norm": 0.020122403278946877,
+ "learning_rate": 0.0006,
+ "loss": 4.236002445220947,
+ "step": 1601
+ },
+ {
+ "epoch": 22.251528384279474,
+ "grad_norm": 0.01957266591489315,
+ "learning_rate": 0.0006,
+ "loss": 4.497328281402588,
+ "step": 1602
+ },
+ {
+ "epoch": 22.265502183406113,
+ "grad_norm": 0.018964895978569984,
+ "learning_rate": 0.0006,
+ "loss": 4.36360502243042,
+ "step": 1603
+ },
+ {
+ "epoch": 22.27947598253275,
+ "grad_norm": 0.018902769312262535,
+ "learning_rate": 0.0006,
+ "loss": 4.427093505859375,
+ "step": 1604
+ },
+ {
+ "epoch": 22.29344978165939,
+ "grad_norm": 0.018918119370937347,
+ "learning_rate": 0.0006,
+ "loss": 4.294950485229492,
+ "step": 1605
+ },
+ {
+ "epoch": 22.307423580786025,
+ "grad_norm": 0.019315775483846664,
+ "learning_rate": 0.0006,
+ "loss": 4.427952289581299,
+ "step": 1606
+ },
+ {
+ "epoch": 22.321397379912664,
+ "grad_norm": 0.018350522965192795,
+ "learning_rate": 0.0006,
+ "loss": 4.272945404052734,
+ "step": 1607
+ },
+ {
+ "epoch": 22.335371179039303,
+ "grad_norm": 0.01759534701704979,
+ "learning_rate": 0.0006,
+ "loss": 4.348023891448975,
+ "step": 1608
+ },
+ {
+ "epoch": 22.349344978165938,
+ "grad_norm": 0.01646057888865471,
+ "learning_rate": 0.0006,
+ "loss": 4.278518199920654,
+ "step": 1609
+ },
+ {
+ "epoch": 22.363318777292577,
+ "grad_norm": 0.015875060111284256,
+ "learning_rate": 0.0006,
+ "loss": 4.423527240753174,
+ "step": 1610
+ },
+ {
+ "epoch": 22.377292576419215,
+ "grad_norm": 0.01638529635965824,
+ "learning_rate": 0.0006,
+ "loss": 4.321723937988281,
+ "step": 1611
+ },
+ {
+ "epoch": 22.39126637554585,
+ "grad_norm": 0.016942061483860016,
+ "learning_rate": 0.0006,
+ "loss": 4.346164703369141,
+ "step": 1612
+ },
+ {
+ "epoch": 22.40524017467249,
+ "grad_norm": 0.015209621749818325,
+ "learning_rate": 0.0006,
+ "loss": 4.434340000152588,
+ "step": 1613
+ },
+ {
+ "epoch": 22.419213973799128,
+ "grad_norm": 0.014423009008169174,
+ "learning_rate": 0.0006,
+ "loss": 4.342560768127441,
+ "step": 1614
+ },
+ {
+ "epoch": 22.433187772925763,
+ "grad_norm": 0.014369525946676731,
+ "learning_rate": 0.0006,
+ "loss": 4.296586990356445,
+ "step": 1615
+ },
+ {
+ "epoch": 22.4471615720524,
+ "grad_norm": 0.01392540242522955,
+ "learning_rate": 0.0006,
+ "loss": 4.433176040649414,
+ "step": 1616
+ },
+ {
+ "epoch": 22.46113537117904,
+ "grad_norm": 0.013206899166107178,
+ "learning_rate": 0.0006,
+ "loss": 4.32962703704834,
+ "step": 1617
+ },
+ {
+ "epoch": 22.475109170305675,
+ "grad_norm": 0.013427077792584896,
+ "learning_rate": 0.0006,
+ "loss": 4.420437335968018,
+ "step": 1618
+ },
+ {
+ "epoch": 22.489082969432314,
+ "grad_norm": 0.014194196090102196,
+ "learning_rate": 0.0006,
+ "loss": 4.410478591918945,
+ "step": 1619
+ },
+ {
+ "epoch": 22.503056768558952,
+ "grad_norm": 0.01362364087253809,
+ "learning_rate": 0.0006,
+ "loss": 4.338932037353516,
+ "step": 1620
+ },
+ {
+ "epoch": 22.51703056768559,
+ "grad_norm": 0.014299589209258556,
+ "learning_rate": 0.0006,
+ "loss": 4.34089994430542,
+ "step": 1621
+ },
+ {
+ "epoch": 22.531004366812226,
+ "grad_norm": 0.014061295427381992,
+ "learning_rate": 0.0006,
+ "loss": 4.29201602935791,
+ "step": 1622
+ },
+ {
+ "epoch": 22.544978165938865,
+ "grad_norm": 0.013945229351520538,
+ "learning_rate": 0.0006,
+ "loss": 4.326052188873291,
+ "step": 1623
+ },
+ {
+ "epoch": 22.558951965065503,
+ "grad_norm": 0.01387068536132574,
+ "learning_rate": 0.0006,
+ "loss": 4.393521308898926,
+ "step": 1624
+ },
+ {
+ "epoch": 22.57292576419214,
+ "grad_norm": 0.013931898400187492,
+ "learning_rate": 0.0006,
+ "loss": 4.3739752769470215,
+ "step": 1625
+ },
+ {
+ "epoch": 22.586899563318777,
+ "grad_norm": 0.01497740764170885,
+ "learning_rate": 0.0006,
+ "loss": 4.3991618156433105,
+ "step": 1626
+ },
+ {
+ "epoch": 22.600873362445416,
+ "grad_norm": 0.015842905268073082,
+ "learning_rate": 0.0006,
+ "loss": 4.405340671539307,
+ "step": 1627
+ },
+ {
+ "epoch": 22.61484716157205,
+ "grad_norm": 0.01580633595585823,
+ "learning_rate": 0.0006,
+ "loss": 4.280102729797363,
+ "step": 1628
+ },
+ {
+ "epoch": 22.62882096069869,
+ "grad_norm": 0.016567382961511612,
+ "learning_rate": 0.0006,
+ "loss": 4.407657146453857,
+ "step": 1629
+ },
+ {
+ "epoch": 22.64279475982533,
+ "grad_norm": 0.01648622378706932,
+ "learning_rate": 0.0006,
+ "loss": 4.392590522766113,
+ "step": 1630
+ },
+ {
+ "epoch": 22.656768558951963,
+ "grad_norm": 0.017066307365894318,
+ "learning_rate": 0.0006,
+ "loss": 4.324862003326416,
+ "step": 1631
+ },
+ {
+ "epoch": 22.670742358078602,
+ "grad_norm": 0.01637062057852745,
+ "learning_rate": 0.0006,
+ "loss": 4.34598445892334,
+ "step": 1632
+ },
+ {
+ "epoch": 22.68471615720524,
+ "grad_norm": 0.018038960173726082,
+ "learning_rate": 0.0006,
+ "loss": 4.320711612701416,
+ "step": 1633
+ },
+ {
+ "epoch": 22.69868995633188,
+ "grad_norm": 0.018096428364515305,
+ "learning_rate": 0.0006,
+ "loss": 4.536527156829834,
+ "step": 1634
+ },
+ {
+ "epoch": 22.712663755458514,
+ "grad_norm": 0.017889291048049927,
+ "learning_rate": 0.0006,
+ "loss": 4.388131141662598,
+ "step": 1635
+ },
+ {
+ "epoch": 22.726637554585153,
+ "grad_norm": 0.018771521747112274,
+ "learning_rate": 0.0006,
+ "loss": 4.357668876647949,
+ "step": 1636
+ },
+ {
+ "epoch": 22.74061135371179,
+ "grad_norm": 0.02100873365998268,
+ "learning_rate": 0.0006,
+ "loss": 4.293972015380859,
+ "step": 1637
+ },
+ {
+ "epoch": 22.754585152838427,
+ "grad_norm": 0.017715785652399063,
+ "learning_rate": 0.0006,
+ "loss": 4.349849224090576,
+ "step": 1638
+ },
+ {
+ "epoch": 22.768558951965066,
+ "grad_norm": 0.015546333976089954,
+ "learning_rate": 0.0006,
+ "loss": 4.311728000640869,
+ "step": 1639
+ },
+ {
+ "epoch": 22.782532751091704,
+ "grad_norm": 0.01605990156531334,
+ "learning_rate": 0.0006,
+ "loss": 4.499123573303223,
+ "step": 1640
+ },
+ {
+ "epoch": 22.79650655021834,
+ "grad_norm": 0.016642142087221146,
+ "learning_rate": 0.0006,
+ "loss": 4.4319257736206055,
+ "step": 1641
+ },
+ {
+ "epoch": 22.810480349344978,
+ "grad_norm": 0.016120387241244316,
+ "learning_rate": 0.0006,
+ "loss": 4.4373650550842285,
+ "step": 1642
+ },
+ {
+ "epoch": 22.824454148471617,
+ "grad_norm": 0.015174890868365765,
+ "learning_rate": 0.0006,
+ "loss": 4.314082145690918,
+ "step": 1643
+ },
+ {
+ "epoch": 22.83842794759825,
+ "grad_norm": 0.01551714539527893,
+ "learning_rate": 0.0006,
+ "loss": 4.360363960266113,
+ "step": 1644
+ },
+ {
+ "epoch": 22.85240174672489,
+ "grad_norm": 0.014966742135584354,
+ "learning_rate": 0.0006,
+ "loss": 4.346253395080566,
+ "step": 1645
+ },
+ {
+ "epoch": 22.86637554585153,
+ "grad_norm": 0.014288028702139854,
+ "learning_rate": 0.0006,
+ "loss": 4.464366436004639,
+ "step": 1646
+ },
+ {
+ "epoch": 22.880349344978168,
+ "grad_norm": 0.015291682444512844,
+ "learning_rate": 0.0006,
+ "loss": 4.504101753234863,
+ "step": 1647
+ },
+ {
+ "epoch": 22.894323144104803,
+ "grad_norm": 0.014534426853060722,
+ "learning_rate": 0.0006,
+ "loss": 4.462172031402588,
+ "step": 1648
+ },
+ {
+ "epoch": 22.90829694323144,
+ "grad_norm": 0.014129819348454475,
+ "learning_rate": 0.0006,
+ "loss": 4.350126266479492,
+ "step": 1649
+ },
+ {
+ "epoch": 22.92227074235808,
+ "grad_norm": 0.014932668767869473,
+ "learning_rate": 0.0006,
+ "loss": 4.374392032623291,
+ "step": 1650
+ },
+ {
+ "epoch": 22.936244541484715,
+ "grad_norm": 0.014647853560745716,
+ "learning_rate": 0.0006,
+ "loss": 4.252264499664307,
+ "step": 1651
+ },
+ {
+ "epoch": 22.950218340611354,
+ "grad_norm": 0.013553360477089882,
+ "learning_rate": 0.0006,
+ "loss": 4.379729270935059,
+ "step": 1652
+ },
+ {
+ "epoch": 22.964192139737992,
+ "grad_norm": 0.014606194570660591,
+ "learning_rate": 0.0006,
+ "loss": 4.363372802734375,
+ "step": 1653
+ },
+ {
+ "epoch": 22.978165938864628,
+ "grad_norm": 0.015000631101429462,
+ "learning_rate": 0.0006,
+ "loss": 4.408048629760742,
+ "step": 1654
+ },
+ {
+ "epoch": 22.992139737991266,
+ "grad_norm": 0.015370816923677921,
+ "learning_rate": 0.0006,
+ "loss": 4.297451019287109,
+ "step": 1655
+ },
+ {
+ "epoch": 23.0,
+ "grad_norm": 0.01709144562482834,
+ "learning_rate": 0.0006,
+ "loss": 4.294776916503906,
+ "step": 1656
+ },
+ {
+ "epoch": 23.0,
+ "eval_loss": 4.678618907928467,
+ "eval_runtime": 56.1668,
+ "eval_samples_per_second": 43.478,
+ "eval_steps_per_second": 1.371,
+ "step": 1656
+ },
+ {
+ "epoch": 23.01397379912664,
+ "grad_norm": 0.01727290451526642,
+ "learning_rate": 0.0006,
+ "loss": 4.311701774597168,
+ "step": 1657
+ },
+ {
+ "epoch": 23.027947598253274,
+ "grad_norm": 0.020645759999752045,
+ "learning_rate": 0.0006,
+ "loss": 4.429897308349609,
+ "step": 1658
+ },
+ {
+ "epoch": 23.041921397379912,
+ "grad_norm": 0.02204226516187191,
+ "learning_rate": 0.0006,
+ "loss": 4.365151882171631,
+ "step": 1659
+ },
+ {
+ "epoch": 23.05589519650655,
+ "grad_norm": 0.023069269955158234,
+ "learning_rate": 0.0006,
+ "loss": 4.3710832595825195,
+ "step": 1660
+ },
+ {
+ "epoch": 23.069868995633186,
+ "grad_norm": 0.024151179939508438,
+ "learning_rate": 0.0006,
+ "loss": 4.403108596801758,
+ "step": 1661
+ },
+ {
+ "epoch": 23.083842794759825,
+ "grad_norm": 0.02511233650147915,
+ "learning_rate": 0.0006,
+ "loss": 4.414128303527832,
+ "step": 1662
+ },
+ {
+ "epoch": 23.097816593886463,
+ "grad_norm": 0.025802113115787506,
+ "learning_rate": 0.0006,
+ "loss": 4.388265609741211,
+ "step": 1663
+ },
+ {
+ "epoch": 23.111790393013102,
+ "grad_norm": 0.02377348765730858,
+ "learning_rate": 0.0006,
+ "loss": 4.357505798339844,
+ "step": 1664
+ },
+ {
+ "epoch": 23.125764192139737,
+ "grad_norm": 0.020876651629805565,
+ "learning_rate": 0.0006,
+ "loss": 4.23043966293335,
+ "step": 1665
+ },
+ {
+ "epoch": 23.139737991266376,
+ "grad_norm": 0.01959727331995964,
+ "learning_rate": 0.0006,
+ "loss": 4.280970096588135,
+ "step": 1666
+ },
+ {
+ "epoch": 23.153711790393015,
+ "grad_norm": 0.018678616732358932,
+ "learning_rate": 0.0006,
+ "loss": 4.346716403961182,
+ "step": 1667
+ },
+ {
+ "epoch": 23.16768558951965,
+ "grad_norm": 0.02108944021165371,
+ "learning_rate": 0.0006,
+ "loss": 4.277904510498047,
+ "step": 1668
+ },
+ {
+ "epoch": 23.18165938864629,
+ "grad_norm": 0.020221566781401634,
+ "learning_rate": 0.0006,
+ "loss": 4.367693901062012,
+ "step": 1669
+ },
+ {
+ "epoch": 23.195633187772927,
+ "grad_norm": 0.019599711522459984,
+ "learning_rate": 0.0006,
+ "loss": 4.257092475891113,
+ "step": 1670
+ },
+ {
+ "epoch": 23.209606986899562,
+ "grad_norm": 0.019833361729979515,
+ "learning_rate": 0.0006,
+ "loss": 4.540737628936768,
+ "step": 1671
+ },
+ {
+ "epoch": 23.2235807860262,
+ "grad_norm": 0.020352095365524292,
+ "learning_rate": 0.0006,
+ "loss": 4.360923767089844,
+ "step": 1672
+ },
+ {
+ "epoch": 23.23755458515284,
+ "grad_norm": 0.01939748041331768,
+ "learning_rate": 0.0006,
+ "loss": 4.393110275268555,
+ "step": 1673
+ },
+ {
+ "epoch": 23.251528384279474,
+ "grad_norm": 0.01884354278445244,
+ "learning_rate": 0.0006,
+ "loss": 4.383938789367676,
+ "step": 1674
+ },
+ {
+ "epoch": 23.265502183406113,
+ "grad_norm": 0.02008470520377159,
+ "learning_rate": 0.0006,
+ "loss": 4.346344947814941,
+ "step": 1675
+ },
+ {
+ "epoch": 23.27947598253275,
+ "grad_norm": 0.019065195694565773,
+ "learning_rate": 0.0006,
+ "loss": 4.36611795425415,
+ "step": 1676
+ },
+ {
+ "epoch": 23.29344978165939,
+ "grad_norm": 0.01938541606068611,
+ "learning_rate": 0.0006,
+ "loss": 4.4009599685668945,
+ "step": 1677
+ },
+ {
+ "epoch": 23.307423580786025,
+ "grad_norm": 0.019044723361730576,
+ "learning_rate": 0.0006,
+ "loss": 4.340169906616211,
+ "step": 1678
+ },
+ {
+ "epoch": 23.321397379912664,
+ "grad_norm": 0.018850378692150116,
+ "learning_rate": 0.0006,
+ "loss": 4.37674617767334,
+ "step": 1679
+ },
+ {
+ "epoch": 23.335371179039303,
+ "grad_norm": 0.019076095893979073,
+ "learning_rate": 0.0006,
+ "loss": 4.226874351501465,
+ "step": 1680
+ },
+ {
+ "epoch": 23.349344978165938,
+ "grad_norm": 0.017543038353323936,
+ "learning_rate": 0.0006,
+ "loss": 4.337751388549805,
+ "step": 1681
+ },
+ {
+ "epoch": 23.363318777292577,
+ "grad_norm": 0.017170218750834465,
+ "learning_rate": 0.0006,
+ "loss": 4.282512664794922,
+ "step": 1682
+ },
+ {
+ "epoch": 23.377292576419215,
+ "grad_norm": 0.01574796624481678,
+ "learning_rate": 0.0006,
+ "loss": 4.30638313293457,
+ "step": 1683
+ },
+ {
+ "epoch": 23.39126637554585,
+ "grad_norm": 0.015160846523940563,
+ "learning_rate": 0.0006,
+ "loss": 4.336716175079346,
+ "step": 1684
+ },
+ {
+ "epoch": 23.40524017467249,
+ "grad_norm": 0.01534526702016592,
+ "learning_rate": 0.0006,
+ "loss": 4.2174787521362305,
+ "step": 1685
+ },
+ {
+ "epoch": 23.419213973799128,
+ "grad_norm": 0.015289999544620514,
+ "learning_rate": 0.0006,
+ "loss": 4.272546291351318,
+ "step": 1686
+ },
+ {
+ "epoch": 23.433187772925763,
+ "grad_norm": 0.014894185587763786,
+ "learning_rate": 0.0006,
+ "loss": 4.296045303344727,
+ "step": 1687
+ },
+ {
+ "epoch": 23.4471615720524,
+ "grad_norm": 0.014869502745568752,
+ "learning_rate": 0.0006,
+ "loss": 4.425461769104004,
+ "step": 1688
+ },
+ {
+ "epoch": 23.46113537117904,
+ "grad_norm": 0.01474483497440815,
+ "learning_rate": 0.0006,
+ "loss": 4.361867427825928,
+ "step": 1689
+ },
+ {
+ "epoch": 23.475109170305675,
+ "grad_norm": 0.015299245715141296,
+ "learning_rate": 0.0006,
+ "loss": 4.364523410797119,
+ "step": 1690
+ },
+ {
+ "epoch": 23.489082969432314,
+ "grad_norm": 0.015201340429484844,
+ "learning_rate": 0.0006,
+ "loss": 4.285182952880859,
+ "step": 1691
+ },
+ {
+ "epoch": 23.503056768558952,
+ "grad_norm": 0.016057293862104416,
+ "learning_rate": 0.0006,
+ "loss": 4.284873962402344,
+ "step": 1692
+ },
+ {
+ "epoch": 23.51703056768559,
+ "grad_norm": 0.015871062874794006,
+ "learning_rate": 0.0006,
+ "loss": 4.326470851898193,
+ "step": 1693
+ },
+ {
+ "epoch": 23.531004366812226,
+ "grad_norm": 0.01586288772523403,
+ "learning_rate": 0.0006,
+ "loss": 4.324883460998535,
+ "step": 1694
+ },
+ {
+ "epoch": 23.544978165938865,
+ "grad_norm": 0.016092827543616295,
+ "learning_rate": 0.0006,
+ "loss": 4.3258867263793945,
+ "step": 1695
+ },
+ {
+ "epoch": 23.558951965065503,
+ "grad_norm": 0.01582016795873642,
+ "learning_rate": 0.0006,
+ "loss": 4.358892440795898,
+ "step": 1696
+ },
+ {
+ "epoch": 23.57292576419214,
+ "grad_norm": 0.01688452623784542,
+ "learning_rate": 0.0006,
+ "loss": 4.280842304229736,
+ "step": 1697
+ },
+ {
+ "epoch": 23.586899563318777,
+ "grad_norm": 0.015865886583924294,
+ "learning_rate": 0.0006,
+ "loss": 4.245567321777344,
+ "step": 1698
+ },
+ {
+ "epoch": 23.600873362445416,
+ "grad_norm": 0.015303662046790123,
+ "learning_rate": 0.0006,
+ "loss": 4.383112907409668,
+ "step": 1699
+ },
+ {
+ "epoch": 23.61484716157205,
+ "grad_norm": 0.015649516135454178,
+ "learning_rate": 0.0006,
+ "loss": 4.519082069396973,
+ "step": 1700
+ },
+ {
+ "epoch": 23.62882096069869,
+ "grad_norm": 0.0167120099067688,
+ "learning_rate": 0.0006,
+ "loss": 4.447847366333008,
+ "step": 1701
+ },
+ {
+ "epoch": 23.64279475982533,
+ "grad_norm": 0.016368716955184937,
+ "learning_rate": 0.0006,
+ "loss": 4.434186935424805,
+ "step": 1702
+ },
+ {
+ "epoch": 23.656768558951963,
+ "grad_norm": 0.017379503697156906,
+ "learning_rate": 0.0006,
+ "loss": 4.317594528198242,
+ "step": 1703
+ },
+ {
+ "epoch": 23.670742358078602,
+ "grad_norm": 0.018573811277747154,
+ "learning_rate": 0.0006,
+ "loss": 4.205793857574463,
+ "step": 1704
+ },
+ {
+ "epoch": 23.68471615720524,
+ "grad_norm": 0.01749090477824211,
+ "learning_rate": 0.0006,
+ "loss": 4.3985700607299805,
+ "step": 1705
+ },
+ {
+ "epoch": 23.69868995633188,
+ "grad_norm": 0.0165668074041605,
+ "learning_rate": 0.0006,
+ "loss": 4.329226970672607,
+ "step": 1706
+ },
+ {
+ "epoch": 23.712663755458514,
+ "grad_norm": 0.016913846135139465,
+ "learning_rate": 0.0006,
+ "loss": 4.388485908508301,
+ "step": 1707
+ },
+ {
+ "epoch": 23.726637554585153,
+ "grad_norm": 0.016813859343528748,
+ "learning_rate": 0.0006,
+ "loss": 4.3395915031433105,
+ "step": 1708
+ },
+ {
+ "epoch": 23.74061135371179,
+ "grad_norm": 0.01676975190639496,
+ "learning_rate": 0.0006,
+ "loss": 4.254745960235596,
+ "step": 1709
+ },
+ {
+ "epoch": 23.754585152838427,
+ "grad_norm": 0.016370350494980812,
+ "learning_rate": 0.0006,
+ "loss": 4.2612223625183105,
+ "step": 1710
+ },
+ {
+ "epoch": 23.768558951965066,
+ "grad_norm": 0.01696198247373104,
+ "learning_rate": 0.0006,
+ "loss": 4.327524662017822,
+ "step": 1711
+ },
+ {
+ "epoch": 23.782532751091704,
+ "grad_norm": 0.0189108457416296,
+ "learning_rate": 0.0006,
+ "loss": 4.325839519500732,
+ "step": 1712
+ },
+ {
+ "epoch": 23.79650655021834,
+ "grad_norm": 0.01862351782619953,
+ "learning_rate": 0.0006,
+ "loss": 4.333662509918213,
+ "step": 1713
+ },
+ {
+ "epoch": 23.810480349344978,
+ "grad_norm": 0.01655014418065548,
+ "learning_rate": 0.0006,
+ "loss": 4.39305305480957,
+ "step": 1714
+ },
+ {
+ "epoch": 23.824454148471617,
+ "grad_norm": 0.015899376943707466,
+ "learning_rate": 0.0006,
+ "loss": 4.390053749084473,
+ "step": 1715
+ },
+ {
+ "epoch": 23.83842794759825,
+ "grad_norm": 0.016519056633114815,
+ "learning_rate": 0.0006,
+ "loss": 4.255903244018555,
+ "step": 1716
+ },
+ {
+ "epoch": 23.85240174672489,
+ "grad_norm": 0.0167539119720459,
+ "learning_rate": 0.0006,
+ "loss": 4.255727767944336,
+ "step": 1717
+ },
+ {
+ "epoch": 23.86637554585153,
+ "grad_norm": 0.0171353816986084,
+ "learning_rate": 0.0006,
+ "loss": 4.376740455627441,
+ "step": 1718
+ },
+ {
+ "epoch": 23.880349344978168,
+ "grad_norm": 0.016719117760658264,
+ "learning_rate": 0.0006,
+ "loss": 4.3376688957214355,
+ "step": 1719
+ },
+ {
+ "epoch": 23.894323144104803,
+ "grad_norm": 0.01561494916677475,
+ "learning_rate": 0.0006,
+ "loss": 4.3345746994018555,
+ "step": 1720
+ },
+ {
+ "epoch": 23.90829694323144,
+ "grad_norm": 0.016303174197673798,
+ "learning_rate": 0.0006,
+ "loss": 4.332371234893799,
+ "step": 1721
+ },
+ {
+ "epoch": 23.92227074235808,
+ "grad_norm": 0.016722865402698517,
+ "learning_rate": 0.0006,
+ "loss": 4.267125129699707,
+ "step": 1722
+ },
+ {
+ "epoch": 23.936244541484715,
+ "grad_norm": 0.017072124406695366,
+ "learning_rate": 0.0006,
+ "loss": 4.322751998901367,
+ "step": 1723
+ },
+ {
+ "epoch": 23.950218340611354,
+ "grad_norm": 0.01699932850897312,
+ "learning_rate": 0.0006,
+ "loss": 4.378422737121582,
+ "step": 1724
+ },
+ {
+ "epoch": 23.964192139737992,
+ "grad_norm": 0.01742720976471901,
+ "learning_rate": 0.0006,
+ "loss": 4.345083236694336,
+ "step": 1725
+ },
+ {
+ "epoch": 23.978165938864628,
+ "grad_norm": 0.018797501921653748,
+ "learning_rate": 0.0006,
+ "loss": 4.290443420410156,
+ "step": 1726
+ },
+ {
+ "epoch": 23.992139737991266,
+ "grad_norm": 0.018090683966875076,
+ "learning_rate": 0.0006,
+ "loss": 4.377580642700195,
+ "step": 1727
+ },
+ {
+ "epoch": 24.0,
+ "grad_norm": 0.017305459827184677,
+ "learning_rate": 0.0006,
+ "loss": 4.364426612854004,
+ "step": 1728
+ },
+ {
+ "epoch": 24.0,
+ "eval_loss": 4.647762298583984,
+ "eval_runtime": 56.5402,
+ "eval_samples_per_second": 43.191,
+ "eval_steps_per_second": 1.362,
+ "step": 1728
+ },
+ {
+ "epoch": 24.01397379912664,
+ "grad_norm": 0.016493607312440872,
+ "learning_rate": 0.0006,
+ "loss": 4.252255916595459,
+ "step": 1729
+ },
+ {
+ "epoch": 24.027947598253274,
+ "grad_norm": 0.016529450193047523,
+ "learning_rate": 0.0006,
+ "loss": 4.352746963500977,
+ "step": 1730
+ },
+ {
+ "epoch": 24.041921397379912,
+ "grad_norm": 0.016940191388130188,
+ "learning_rate": 0.0006,
+ "loss": 4.301870822906494,
+ "step": 1731
+ },
+ {
+ "epoch": 24.05589519650655,
+ "grad_norm": 0.01752558909356594,
+ "learning_rate": 0.0006,
+ "loss": 4.2762451171875,
+ "step": 1732
+ },
+ {
+ "epoch": 24.069868995633186,
+ "grad_norm": 0.017112884670495987,
+ "learning_rate": 0.0006,
+ "loss": 4.2659149169921875,
+ "step": 1733
+ },
+ {
+ "epoch": 24.083842794759825,
+ "grad_norm": 0.01900586113333702,
+ "learning_rate": 0.0006,
+ "loss": 4.306058883666992,
+ "step": 1734
+ },
+ {
+ "epoch": 24.097816593886463,
+ "grad_norm": 0.019447680562734604,
+ "learning_rate": 0.0006,
+ "loss": 4.450128555297852,
+ "step": 1735
+ },
+ {
+ "epoch": 24.111790393013102,
+ "grad_norm": 0.01948186755180359,
+ "learning_rate": 0.0006,
+ "loss": 4.202869415283203,
+ "step": 1736
+ },
+ {
+ "epoch": 24.125764192139737,
+ "grad_norm": 0.01716383546590805,
+ "learning_rate": 0.0006,
+ "loss": 4.318876266479492,
+ "step": 1737
+ },
+ {
+ "epoch": 24.139737991266376,
+ "grad_norm": 0.017049988731741905,
+ "learning_rate": 0.0006,
+ "loss": 4.237238883972168,
+ "step": 1738
+ },
+ {
+ "epoch": 24.153711790393015,
+ "grad_norm": 0.015169923193752766,
+ "learning_rate": 0.0006,
+ "loss": 4.289426803588867,
+ "step": 1739
+ },
+ {
+ "epoch": 24.16768558951965,
+ "grad_norm": 0.014760667458176613,
+ "learning_rate": 0.0006,
+ "loss": 4.356924057006836,
+ "step": 1740
+ },
+ {
+ "epoch": 24.18165938864629,
+ "grad_norm": 0.015127750113606453,
+ "learning_rate": 0.0006,
+ "loss": 4.142916679382324,
+ "step": 1741
+ },
+ {
+ "epoch": 24.195633187772927,
+ "grad_norm": 0.01549555640667677,
+ "learning_rate": 0.0006,
+ "loss": 4.345563888549805,
+ "step": 1742
+ },
+ {
+ "epoch": 24.209606986899562,
+ "grad_norm": 0.016393091529607773,
+ "learning_rate": 0.0006,
+ "loss": 4.417495250701904,
+ "step": 1743
+ },
+ {
+ "epoch": 24.2235807860262,
+ "grad_norm": 0.01785266026854515,
+ "learning_rate": 0.0006,
+ "loss": 4.3553547859191895,
+ "step": 1744
+ },
+ {
+ "epoch": 24.23755458515284,
+ "grad_norm": 0.017574617639183998,
+ "learning_rate": 0.0006,
+ "loss": 4.241024494171143,
+ "step": 1745
+ },
+ {
+ "epoch": 24.251528384279474,
+ "grad_norm": 0.019060306251049042,
+ "learning_rate": 0.0006,
+ "loss": 4.3685150146484375,
+ "step": 1746
+ },
+ {
+ "epoch": 24.265502183406113,
+ "grad_norm": 0.0214633010327816,
+ "learning_rate": 0.0006,
+ "loss": 4.323866844177246,
+ "step": 1747
+ },
+ {
+ "epoch": 24.27947598253275,
+ "grad_norm": 0.021055003628134727,
+ "learning_rate": 0.0006,
+ "loss": 4.278757095336914,
+ "step": 1748
+ },
+ {
+ "epoch": 24.29344978165939,
+ "grad_norm": 0.018618572503328323,
+ "learning_rate": 0.0006,
+ "loss": 4.28849983215332,
+ "step": 1749
+ },
+ {
+ "epoch": 24.307423580786025,
+ "grad_norm": 0.018666435033082962,
+ "learning_rate": 0.0006,
+ "loss": 4.368590354919434,
+ "step": 1750
+ },
+ {
+ "epoch": 24.321397379912664,
+ "grad_norm": 0.020626146346330643,
+ "learning_rate": 0.0006,
+ "loss": 4.2951836585998535,
+ "step": 1751
+ },
+ {
+ "epoch": 24.335371179039303,
+ "grad_norm": 0.01929759792983532,
+ "learning_rate": 0.0006,
+ "loss": 4.3664703369140625,
+ "step": 1752
+ },
+ {
+ "epoch": 24.349344978165938,
+ "grad_norm": 0.020956002175807953,
+ "learning_rate": 0.0006,
+ "loss": 4.345308303833008,
+ "step": 1753
+ },
+ {
+ "epoch": 24.363318777292577,
+ "grad_norm": 0.019227957352995872,
+ "learning_rate": 0.0006,
+ "loss": 4.349849224090576,
+ "step": 1754
+ },
+ {
+ "epoch": 24.377292576419215,
+ "grad_norm": 0.015970097854733467,
+ "learning_rate": 0.0006,
+ "loss": 4.223012447357178,
+ "step": 1755
+ },
+ {
+ "epoch": 24.39126637554585,
+ "grad_norm": 0.015951646491885185,
+ "learning_rate": 0.0006,
+ "loss": 4.236420631408691,
+ "step": 1756
+ },
+ {
+ "epoch": 24.40524017467249,
+ "grad_norm": 0.016032515093684196,
+ "learning_rate": 0.0006,
+ "loss": 4.226899147033691,
+ "step": 1757
+ },
+ {
+ "epoch": 24.419213973799128,
+ "grad_norm": 0.01573132909834385,
+ "learning_rate": 0.0006,
+ "loss": 4.315916061401367,
+ "step": 1758
+ },
+ {
+ "epoch": 24.433187772925763,
+ "grad_norm": 0.015367005951702595,
+ "learning_rate": 0.0006,
+ "loss": 4.358939170837402,
+ "step": 1759
+ },
+ {
+ "epoch": 24.4471615720524,
+ "grad_norm": 0.01669907011091709,
+ "learning_rate": 0.0006,
+ "loss": 4.232147216796875,
+ "step": 1760
+ },
+ {
+ "epoch": 24.46113537117904,
+ "grad_norm": 0.018708152696490288,
+ "learning_rate": 0.0006,
+ "loss": 4.329852104187012,
+ "step": 1761
+ },
+ {
+ "epoch": 24.475109170305675,
+ "grad_norm": 0.019035987555980682,
+ "learning_rate": 0.0006,
+ "loss": 4.260340213775635,
+ "step": 1762
+ },
+ {
+ "epoch": 24.489082969432314,
+ "grad_norm": 0.01832144521176815,
+ "learning_rate": 0.0006,
+ "loss": 4.359543800354004,
+ "step": 1763
+ },
+ {
+ "epoch": 24.503056768558952,
+ "grad_norm": 0.015313294716179371,
+ "learning_rate": 0.0006,
+ "loss": 4.279088020324707,
+ "step": 1764
+ },
+ {
+ "epoch": 24.51703056768559,
+ "grad_norm": 0.015803923830389977,
+ "learning_rate": 0.0006,
+ "loss": 4.230682849884033,
+ "step": 1765
+ },
+ {
+ "epoch": 24.531004366812226,
+ "grad_norm": 0.016748374328017235,
+ "learning_rate": 0.0006,
+ "loss": 4.349661827087402,
+ "step": 1766
+ },
+ {
+ "epoch": 24.544978165938865,
+ "grad_norm": 0.015732428058981895,
+ "learning_rate": 0.0006,
+ "loss": 4.372441291809082,
+ "step": 1767
+ },
+ {
+ "epoch": 24.558951965065503,
+ "grad_norm": 0.016564829275012016,
+ "learning_rate": 0.0006,
+ "loss": 4.323090553283691,
+ "step": 1768
+ },
+ {
+ "epoch": 24.57292576419214,
+ "grad_norm": 0.01603010483086109,
+ "learning_rate": 0.0006,
+ "loss": 4.347898006439209,
+ "step": 1769
+ },
+ {
+ "epoch": 24.586899563318777,
+ "grad_norm": 0.01674809865653515,
+ "learning_rate": 0.0006,
+ "loss": 4.269165515899658,
+ "step": 1770
+ },
+ {
+ "epoch": 24.600873362445416,
+ "grad_norm": 0.017333930358290672,
+ "learning_rate": 0.0006,
+ "loss": 4.353298187255859,
+ "step": 1771
+ },
+ {
+ "epoch": 24.61484716157205,
+ "grad_norm": 0.017519580200314522,
+ "learning_rate": 0.0006,
+ "loss": 4.265199661254883,
+ "step": 1772
+ },
+ {
+ "epoch": 24.62882096069869,
+ "grad_norm": 0.017675306648015976,
+ "learning_rate": 0.0006,
+ "loss": 4.322596549987793,
+ "step": 1773
+ },
+ {
+ "epoch": 24.64279475982533,
+ "grad_norm": 0.017166534438729286,
+ "learning_rate": 0.0006,
+ "loss": 4.413900375366211,
+ "step": 1774
+ },
+ {
+ "epoch": 24.656768558951963,
+ "grad_norm": 0.016362346708774567,
+ "learning_rate": 0.0006,
+ "loss": 4.307236671447754,
+ "step": 1775
+ },
+ {
+ "epoch": 24.670742358078602,
+ "grad_norm": 0.016086872667074203,
+ "learning_rate": 0.0006,
+ "loss": 4.235316276550293,
+ "step": 1776
+ },
+ {
+ "epoch": 24.68471615720524,
+ "grad_norm": 0.01689537614583969,
+ "learning_rate": 0.0006,
+ "loss": 4.390591621398926,
+ "step": 1777
+ },
+ {
+ "epoch": 24.69868995633188,
+ "grad_norm": 0.016836725175380707,
+ "learning_rate": 0.0006,
+ "loss": 4.2990031242370605,
+ "step": 1778
+ },
+ {
+ "epoch": 24.712663755458514,
+ "grad_norm": 0.015627335757017136,
+ "learning_rate": 0.0006,
+ "loss": 4.166745185852051,
+ "step": 1779
+ },
+ {
+ "epoch": 24.726637554585153,
+ "grad_norm": 0.015007016249001026,
+ "learning_rate": 0.0006,
+ "loss": 4.295273303985596,
+ "step": 1780
+ },
+ {
+ "epoch": 24.74061135371179,
+ "grad_norm": 0.016084209084510803,
+ "learning_rate": 0.0006,
+ "loss": 4.342424392700195,
+ "step": 1781
+ },
+ {
+ "epoch": 24.754585152838427,
+ "grad_norm": 0.01595073565840721,
+ "learning_rate": 0.0006,
+ "loss": 4.3060102462768555,
+ "step": 1782
+ },
+ {
+ "epoch": 24.768558951965066,
+ "grad_norm": 0.016113461926579475,
+ "learning_rate": 0.0006,
+ "loss": 4.267908096313477,
+ "step": 1783
+ },
+ {
+ "epoch": 24.782532751091704,
+ "grad_norm": 0.01564556173980236,
+ "learning_rate": 0.0006,
+ "loss": 4.324642658233643,
+ "step": 1784
+ },
+ {
+ "epoch": 24.79650655021834,
+ "grad_norm": 0.01575268618762493,
+ "learning_rate": 0.0006,
+ "loss": 4.328181743621826,
+ "step": 1785
+ },
+ {
+ "epoch": 24.810480349344978,
+ "grad_norm": 0.016830824315547943,
+ "learning_rate": 0.0006,
+ "loss": 4.247906684875488,
+ "step": 1786
+ },
+ {
+ "epoch": 24.824454148471617,
+ "grad_norm": 0.017989547923207283,
+ "learning_rate": 0.0006,
+ "loss": 4.228281021118164,
+ "step": 1787
+ },
+ {
+ "epoch": 24.83842794759825,
+ "grad_norm": 0.016783086583018303,
+ "learning_rate": 0.0006,
+ "loss": 4.2880964279174805,
+ "step": 1788
+ },
+ {
+ "epoch": 24.85240174672489,
+ "grad_norm": 0.016467561945319176,
+ "learning_rate": 0.0006,
+ "loss": 4.36118745803833,
+ "step": 1789
+ },
+ {
+ "epoch": 24.86637554585153,
+ "grad_norm": 0.01780523732304573,
+ "learning_rate": 0.0006,
+ "loss": 4.172321319580078,
+ "step": 1790
+ },
+ {
+ "epoch": 24.880349344978168,
+ "grad_norm": 0.017102031037211418,
+ "learning_rate": 0.0006,
+ "loss": 4.352826118469238,
+ "step": 1791
+ },
+ {
+ "epoch": 24.894323144104803,
+ "grad_norm": 0.016409218311309814,
+ "learning_rate": 0.0006,
+ "loss": 4.212162017822266,
+ "step": 1792
+ },
+ {
+ "epoch": 24.90829694323144,
+ "grad_norm": 0.018660584464669228,
+ "learning_rate": 0.0006,
+ "loss": 4.346684455871582,
+ "step": 1793
+ },
+ {
+ "epoch": 24.92227074235808,
+ "grad_norm": 0.017389440909028053,
+ "learning_rate": 0.0006,
+ "loss": 4.197913646697998,
+ "step": 1794
+ },
+ {
+ "epoch": 24.936244541484715,
+ "grad_norm": 0.01596485823392868,
+ "learning_rate": 0.0006,
+ "loss": 4.370617866516113,
+ "step": 1795
+ },
+ {
+ "epoch": 24.950218340611354,
+ "grad_norm": 0.01709776371717453,
+ "learning_rate": 0.0006,
+ "loss": 4.313847064971924,
+ "step": 1796
+ },
+ {
+ "epoch": 24.964192139737992,
+ "grad_norm": 0.016571134328842163,
+ "learning_rate": 0.0006,
+ "loss": 4.1895856857299805,
+ "step": 1797
+ },
+ {
+ "epoch": 24.978165938864628,
+ "grad_norm": 0.016851048916578293,
+ "learning_rate": 0.0006,
+ "loss": 4.350772857666016,
+ "step": 1798
+ },
+ {
+ "epoch": 24.992139737991266,
+ "grad_norm": 0.017007440328598022,
+ "learning_rate": 0.0006,
+ "loss": 4.399008750915527,
+ "step": 1799
+ },
+ {
+ "epoch": 25.0,
+ "grad_norm": 0.01791212148964405,
+ "learning_rate": 0.0006,
+ "loss": 4.306121826171875,
+ "step": 1800
+ },
+ {
+ "epoch": 25.0,
+ "eval_loss": 4.636417865753174,
+ "eval_runtime": 56.8296,
+ "eval_samples_per_second": 42.971,
+ "eval_steps_per_second": 1.355,
+ "step": 1800
+ },
+ {
+ "epoch": 25.01397379912664,
+ "grad_norm": 0.01729617640376091,
+ "learning_rate": 0.0006,
+ "loss": 4.257018089294434,
+ "step": 1801
+ },
+ {
+ "epoch": 25.027947598253274,
+ "grad_norm": 0.017466643825173378,
+ "learning_rate": 0.0006,
+ "loss": 4.307605266571045,
+ "step": 1802
+ },
+ {
+ "epoch": 25.041921397379912,
+ "grad_norm": 0.018614279106259346,
+ "learning_rate": 0.0006,
+ "loss": 4.360141277313232,
+ "step": 1803
+ },
+ {
+ "epoch": 25.05589519650655,
+ "grad_norm": 0.02010771445930004,
+ "learning_rate": 0.0006,
+ "loss": 4.196428298950195,
+ "step": 1804
+ },
+ {
+ "epoch": 25.069868995633186,
+ "grad_norm": 0.02090228535234928,
+ "learning_rate": 0.0006,
+ "loss": 4.288947105407715,
+ "step": 1805
+ },
+ {
+ "epoch": 25.083842794759825,
+ "grad_norm": 0.02050255797803402,
+ "learning_rate": 0.0006,
+ "loss": 4.259582042694092,
+ "step": 1806
+ },
+ {
+ "epoch": 25.097816593886463,
+ "grad_norm": 0.021844089031219482,
+ "learning_rate": 0.0006,
+ "loss": 4.405138969421387,
+ "step": 1807
+ },
+ {
+ "epoch": 25.111790393013102,
+ "grad_norm": 0.022287018597126007,
+ "learning_rate": 0.0006,
+ "loss": 4.333793640136719,
+ "step": 1808
+ },
+ {
+ "epoch": 25.125764192139737,
+ "grad_norm": 0.02288329415023327,
+ "learning_rate": 0.0006,
+ "loss": 4.3910980224609375,
+ "step": 1809
+ },
+ {
+ "epoch": 25.139737991266376,
+ "grad_norm": 0.02373199723660946,
+ "learning_rate": 0.0006,
+ "loss": 4.32921028137207,
+ "step": 1810
+ },
+ {
+ "epoch": 25.153711790393015,
+ "grad_norm": 0.024960234761238098,
+ "learning_rate": 0.0006,
+ "loss": 4.295609474182129,
+ "step": 1811
+ },
+ {
+ "epoch": 25.16768558951965,
+ "grad_norm": 0.026743892580270767,
+ "learning_rate": 0.0006,
+ "loss": 4.315308094024658,
+ "step": 1812
+ },
+ {
+ "epoch": 25.18165938864629,
+ "grad_norm": 0.022138185799121857,
+ "learning_rate": 0.0006,
+ "loss": 4.360208511352539,
+ "step": 1813
+ },
+ {
+ "epoch": 25.195633187772927,
+ "grad_norm": 0.022120574489235878,
+ "learning_rate": 0.0006,
+ "loss": 4.1973772048950195,
+ "step": 1814
+ },
+ {
+ "epoch": 25.209606986899562,
+ "grad_norm": 0.02230502851307392,
+ "learning_rate": 0.0006,
+ "loss": 4.285558700561523,
+ "step": 1815
+ },
+ {
+ "epoch": 25.2235807860262,
+ "grad_norm": 0.022301986813545227,
+ "learning_rate": 0.0006,
+ "loss": 4.238343238830566,
+ "step": 1816
+ },
+ {
+ "epoch": 25.23755458515284,
+ "grad_norm": 0.020050447434186935,
+ "learning_rate": 0.0006,
+ "loss": 4.298235893249512,
+ "step": 1817
+ },
+ {
+ "epoch": 25.251528384279474,
+ "grad_norm": 0.01987724006175995,
+ "learning_rate": 0.0006,
+ "loss": 4.301384449005127,
+ "step": 1818
+ },
+ {
+ "epoch": 25.265502183406113,
+ "grad_norm": 0.020440855994820595,
+ "learning_rate": 0.0006,
+ "loss": 4.269218444824219,
+ "step": 1819
+ },
+ {
+ "epoch": 25.27947598253275,
+ "grad_norm": 0.01891893707215786,
+ "learning_rate": 0.0006,
+ "loss": 4.2401957511901855,
+ "step": 1820
+ },
+ {
+ "epoch": 25.29344978165939,
+ "grad_norm": 0.01903688907623291,
+ "learning_rate": 0.0006,
+ "loss": 4.139670372009277,
+ "step": 1821
+ },
+ {
+ "epoch": 25.307423580786025,
+ "grad_norm": 0.018856938928365707,
+ "learning_rate": 0.0006,
+ "loss": 4.233214378356934,
+ "step": 1822
+ },
+ {
+ "epoch": 25.321397379912664,
+ "grad_norm": 0.017791323363780975,
+ "learning_rate": 0.0006,
+ "loss": 4.2342023849487305,
+ "step": 1823
+ },
+ {
+ "epoch": 25.335371179039303,
+ "grad_norm": 0.017304565757513046,
+ "learning_rate": 0.0006,
+ "loss": 4.355198860168457,
+ "step": 1824
+ },
+ {
+ "epoch": 25.349344978165938,
+ "grad_norm": 0.01576206088066101,
+ "learning_rate": 0.0006,
+ "loss": 4.114147186279297,
+ "step": 1825
+ },
+ {
+ "epoch": 25.363318777292577,
+ "grad_norm": 0.015105154365301132,
+ "learning_rate": 0.0006,
+ "loss": 4.313795566558838,
+ "step": 1826
+ },
+ {
+ "epoch": 25.377292576419215,
+ "grad_norm": 0.01563587784767151,
+ "learning_rate": 0.0006,
+ "loss": 4.308282852172852,
+ "step": 1827
+ },
+ {
+ "epoch": 25.39126637554585,
+ "grad_norm": 0.015980906784534454,
+ "learning_rate": 0.0006,
+ "loss": 4.345149040222168,
+ "step": 1828
+ },
+ {
+ "epoch": 25.40524017467249,
+ "grad_norm": 0.01669284701347351,
+ "learning_rate": 0.0006,
+ "loss": 4.237611293792725,
+ "step": 1829
+ },
+ {
+ "epoch": 25.419213973799128,
+ "grad_norm": 0.016202064231038094,
+ "learning_rate": 0.0006,
+ "loss": 4.3072309494018555,
+ "step": 1830
+ },
+ {
+ "epoch": 25.433187772925763,
+ "grad_norm": 0.01478662807494402,
+ "learning_rate": 0.0006,
+ "loss": 4.276215076446533,
+ "step": 1831
+ },
+ {
+ "epoch": 25.4471615720524,
+ "grad_norm": 0.015168731100857258,
+ "learning_rate": 0.0006,
+ "loss": 4.37509822845459,
+ "step": 1832
+ },
+ {
+ "epoch": 25.46113537117904,
+ "grad_norm": 0.01648980937898159,
+ "learning_rate": 0.0006,
+ "loss": 4.349167346954346,
+ "step": 1833
+ },
+ {
+ "epoch": 25.475109170305675,
+ "grad_norm": 0.01708398200571537,
+ "learning_rate": 0.0006,
+ "loss": 4.346580505371094,
+ "step": 1834
+ },
+ {
+ "epoch": 25.489082969432314,
+ "grad_norm": 0.017370784655213356,
+ "learning_rate": 0.0006,
+ "loss": 4.299466133117676,
+ "step": 1835
+ },
+ {
+ "epoch": 25.503056768558952,
+ "grad_norm": 0.016248228028416634,
+ "learning_rate": 0.0006,
+ "loss": 4.184610366821289,
+ "step": 1836
+ },
+ {
+ "epoch": 25.51703056768559,
+ "grad_norm": 0.017106913030147552,
+ "learning_rate": 0.0006,
+ "loss": 4.366856098175049,
+ "step": 1837
+ },
+ {
+ "epoch": 25.531004366812226,
+ "grad_norm": 0.016071965917944908,
+ "learning_rate": 0.0006,
+ "loss": 4.297691345214844,
+ "step": 1838
+ },
+ {
+ "epoch": 25.544978165938865,
+ "grad_norm": 0.016451986506581306,
+ "learning_rate": 0.0006,
+ "loss": 4.229983329772949,
+ "step": 1839
+ },
+ {
+ "epoch": 25.558951965065503,
+ "grad_norm": 0.016195081174373627,
+ "learning_rate": 0.0006,
+ "loss": 4.259893417358398,
+ "step": 1840
+ },
+ {
+ "epoch": 25.57292576419214,
+ "grad_norm": 0.0165175162255764,
+ "learning_rate": 0.0006,
+ "loss": 4.3348236083984375,
+ "step": 1841
+ },
+ {
+ "epoch": 25.586899563318777,
+ "grad_norm": 0.01651517115533352,
+ "learning_rate": 0.0006,
+ "loss": 4.319541931152344,
+ "step": 1842
+ },
+ {
+ "epoch": 25.600873362445416,
+ "grad_norm": 0.016143690794706345,
+ "learning_rate": 0.0006,
+ "loss": 4.345652103424072,
+ "step": 1843
+ },
+ {
+ "epoch": 25.61484716157205,
+ "grad_norm": 0.015140696428716183,
+ "learning_rate": 0.0006,
+ "loss": 4.212047576904297,
+ "step": 1844
+ },
+ {
+ "epoch": 25.62882096069869,
+ "grad_norm": 0.014637443237006664,
+ "learning_rate": 0.0006,
+ "loss": 4.187453269958496,
+ "step": 1845
+ },
+ {
+ "epoch": 25.64279475982533,
+ "grad_norm": 0.014585614204406738,
+ "learning_rate": 0.0006,
+ "loss": 4.253936767578125,
+ "step": 1846
+ },
+ {
+ "epoch": 25.656768558951963,
+ "grad_norm": 0.015158289112150669,
+ "learning_rate": 0.0006,
+ "loss": 4.34612512588501,
+ "step": 1847
+ },
+ {
+ "epoch": 25.670742358078602,
+ "grad_norm": 0.015202849172055721,
+ "learning_rate": 0.0006,
+ "loss": 4.270634651184082,
+ "step": 1848
+ },
+ {
+ "epoch": 25.68471615720524,
+ "grad_norm": 0.015363575890660286,
+ "learning_rate": 0.0006,
+ "loss": 4.271080493927002,
+ "step": 1849
+ },
+ {
+ "epoch": 25.69868995633188,
+ "grad_norm": 0.014360226690769196,
+ "learning_rate": 0.0006,
+ "loss": 4.375033855438232,
+ "step": 1850
+ },
+ {
+ "epoch": 25.712663755458514,
+ "grad_norm": 0.014263181947171688,
+ "learning_rate": 0.0006,
+ "loss": 4.424138069152832,
+ "step": 1851
+ },
+ {
+ "epoch": 25.726637554585153,
+ "grad_norm": 0.014398688450455666,
+ "learning_rate": 0.0006,
+ "loss": 4.410154819488525,
+ "step": 1852
+ },
+ {
+ "epoch": 25.74061135371179,
+ "grad_norm": 0.016131488606333733,
+ "learning_rate": 0.0006,
+ "loss": 4.324014663696289,
+ "step": 1853
+ },
+ {
+ "epoch": 25.754585152838427,
+ "grad_norm": 0.017568735405802727,
+ "learning_rate": 0.0006,
+ "loss": 4.367861747741699,
+ "step": 1854
+ },
+ {
+ "epoch": 25.768558951965066,
+ "grad_norm": 0.017907511442899704,
+ "learning_rate": 0.0006,
+ "loss": 4.352965354919434,
+ "step": 1855
+ },
+ {
+ "epoch": 25.782532751091704,
+ "grad_norm": 0.016918016597628593,
+ "learning_rate": 0.0006,
+ "loss": 4.252397537231445,
+ "step": 1856
+ },
+ {
+ "epoch": 25.79650655021834,
+ "grad_norm": 0.017274610698223114,
+ "learning_rate": 0.0006,
+ "loss": 4.342309951782227,
+ "step": 1857
+ },
+ {
+ "epoch": 25.810480349344978,
+ "grad_norm": 0.017879825085401535,
+ "learning_rate": 0.0006,
+ "loss": 4.330987453460693,
+ "step": 1858
+ },
+ {
+ "epoch": 25.824454148471617,
+ "grad_norm": 0.01736099272966385,
+ "learning_rate": 0.0006,
+ "loss": 4.307255744934082,
+ "step": 1859
+ },
+ {
+ "epoch": 25.83842794759825,
+ "grad_norm": 0.016575824469327927,
+ "learning_rate": 0.0006,
+ "loss": 4.262016773223877,
+ "step": 1860
+ },
+ {
+ "epoch": 25.85240174672489,
+ "grad_norm": 0.015181140042841434,
+ "learning_rate": 0.0006,
+ "loss": 4.291863441467285,
+ "step": 1861
+ },
+ {
+ "epoch": 25.86637554585153,
+ "grad_norm": 0.015965687111020088,
+ "learning_rate": 0.0006,
+ "loss": 4.424836158752441,
+ "step": 1862
+ },
+ {
+ "epoch": 25.880349344978168,
+ "grad_norm": 0.01687219925224781,
+ "learning_rate": 0.0006,
+ "loss": 4.343531131744385,
+ "step": 1863
+ },
+ {
+ "epoch": 25.894323144104803,
+ "grad_norm": 0.015115504153072834,
+ "learning_rate": 0.0006,
+ "loss": 4.357808589935303,
+ "step": 1864
+ },
+ {
+ "epoch": 25.90829694323144,
+ "grad_norm": 0.014150998555123806,
+ "learning_rate": 0.0006,
+ "loss": 4.324550151824951,
+ "step": 1865
+ },
+ {
+ "epoch": 25.92227074235808,
+ "grad_norm": 0.014989510178565979,
+ "learning_rate": 0.0006,
+ "loss": 4.380428314208984,
+ "step": 1866
+ },
+ {
+ "epoch": 25.936244541484715,
+ "grad_norm": 0.01527960691601038,
+ "learning_rate": 0.0006,
+ "loss": 4.399374961853027,
+ "step": 1867
+ },
+ {
+ "epoch": 25.950218340611354,
+ "grad_norm": 0.015434633940458298,
+ "learning_rate": 0.0006,
+ "loss": 4.260984420776367,
+ "step": 1868
+ },
+ {
+ "epoch": 25.964192139737992,
+ "grad_norm": 0.015370228327810764,
+ "learning_rate": 0.0006,
+ "loss": 4.3340253829956055,
+ "step": 1869
+ },
+ {
+ "epoch": 25.978165938864628,
+ "grad_norm": 0.015348542481660843,
+ "learning_rate": 0.0006,
+ "loss": 4.369132995605469,
+ "step": 1870
+ },
+ {
+ "epoch": 25.992139737991266,
+ "grad_norm": 0.015354936942458153,
+ "learning_rate": 0.0006,
+ "loss": 4.381937026977539,
+ "step": 1871
+ },
+ {
+ "epoch": 26.0,
+ "grad_norm": 0.01796272210776806,
+ "learning_rate": 0.0006,
+ "loss": 4.286768436431885,
+ "step": 1872
+ },
+ {
+ "epoch": 26.0,
+ "eval_loss": 4.576984405517578,
+ "eval_runtime": 57.2091,
+ "eval_samples_per_second": 42.685,
+ "eval_steps_per_second": 1.346,
+ "step": 1872
+ },
+ {
+ "epoch": 26.01397379912664,
+ "grad_norm": 0.01748845726251602,
+ "learning_rate": 0.0006,
+ "loss": 4.337504863739014,
+ "step": 1873
+ },
+ {
+ "epoch": 26.027947598253274,
+ "grad_norm": 0.019244296476244926,
+ "learning_rate": 0.0006,
+ "loss": 4.264280796051025,
+ "step": 1874
+ },
+ {
+ "epoch": 26.041921397379912,
+ "grad_norm": 0.0215692650526762,
+ "learning_rate": 0.0006,
+ "loss": 4.350830078125,
+ "step": 1875
+ },
+ {
+ "epoch": 26.05589519650655,
+ "grad_norm": 0.022189252078533173,
+ "learning_rate": 0.0006,
+ "loss": 4.299098968505859,
+ "step": 1876
+ },
+ {
+ "epoch": 26.069868995633186,
+ "grad_norm": 0.022384043782949448,
+ "learning_rate": 0.0006,
+ "loss": 4.1185173988342285,
+ "step": 1877
+ },
+ {
+ "epoch": 26.083842794759825,
+ "grad_norm": 0.02076014317572117,
+ "learning_rate": 0.0006,
+ "loss": 4.297612190246582,
+ "step": 1878
+ },
+ {
+ "epoch": 26.097816593886463,
+ "grad_norm": 0.020843051373958588,
+ "learning_rate": 0.0006,
+ "loss": 4.294098854064941,
+ "step": 1879
+ },
+ {
+ "epoch": 26.111790393013102,
+ "grad_norm": 0.020654456689953804,
+ "learning_rate": 0.0006,
+ "loss": 4.267856597900391,
+ "step": 1880
+ },
+ {
+ "epoch": 26.125764192139737,
+ "grad_norm": 0.01860472559928894,
+ "learning_rate": 0.0006,
+ "loss": 4.284232139587402,
+ "step": 1881
+ },
+ {
+ "epoch": 26.139737991266376,
+ "grad_norm": 0.018719421699643135,
+ "learning_rate": 0.0006,
+ "loss": 4.277214527130127,
+ "step": 1882
+ },
+ {
+ "epoch": 26.153711790393015,
+ "grad_norm": 0.020903829485177994,
+ "learning_rate": 0.0006,
+ "loss": 4.254947662353516,
+ "step": 1883
+ },
+ {
+ "epoch": 26.16768558951965,
+ "grad_norm": 0.02020149491727352,
+ "learning_rate": 0.0006,
+ "loss": 4.2526936531066895,
+ "step": 1884
+ },
+ {
+ "epoch": 26.18165938864629,
+ "grad_norm": 0.0180392786860466,
+ "learning_rate": 0.0006,
+ "loss": 4.309922695159912,
+ "step": 1885
+ },
+ {
+ "epoch": 26.195633187772927,
+ "grad_norm": 0.018163125962018967,
+ "learning_rate": 0.0006,
+ "loss": 4.13860559463501,
+ "step": 1886
+ },
+ {
+ "epoch": 26.209606986899562,
+ "grad_norm": 0.01853291131556034,
+ "learning_rate": 0.0006,
+ "loss": 4.190926551818848,
+ "step": 1887
+ },
+ {
+ "epoch": 26.2235807860262,
+ "grad_norm": 0.01757677271962166,
+ "learning_rate": 0.0006,
+ "loss": 4.163999557495117,
+ "step": 1888
+ },
+ {
+ "epoch": 26.23755458515284,
+ "grad_norm": 0.01767992228269577,
+ "learning_rate": 0.0006,
+ "loss": 4.184430122375488,
+ "step": 1889
+ },
+ {
+ "epoch": 26.251528384279474,
+ "grad_norm": 0.018139973282814026,
+ "learning_rate": 0.0006,
+ "loss": 4.300766944885254,
+ "step": 1890
+ },
+ {
+ "epoch": 26.265502183406113,
+ "grad_norm": 0.020184461027383804,
+ "learning_rate": 0.0006,
+ "loss": 4.232537269592285,
+ "step": 1891
+ },
+ {
+ "epoch": 26.27947598253275,
+ "grad_norm": 0.02028856985270977,
+ "learning_rate": 0.0006,
+ "loss": 4.324883460998535,
+ "step": 1892
+ },
+ {
+ "epoch": 26.29344978165939,
+ "grad_norm": 0.017120616510510445,
+ "learning_rate": 0.0006,
+ "loss": 4.23493766784668,
+ "step": 1893
+ },
+ {
+ "epoch": 26.307423580786025,
+ "grad_norm": 0.016202978789806366,
+ "learning_rate": 0.0006,
+ "loss": 4.197602272033691,
+ "step": 1894
+ },
+ {
+ "epoch": 26.321397379912664,
+ "grad_norm": 0.017231421545147896,
+ "learning_rate": 0.0006,
+ "loss": 4.302001953125,
+ "step": 1895
+ },
+ {
+ "epoch": 26.335371179039303,
+ "grad_norm": 0.01725655607879162,
+ "learning_rate": 0.0006,
+ "loss": 4.356800079345703,
+ "step": 1896
+ },
+ {
+ "epoch": 26.349344978165938,
+ "grad_norm": 0.016657443717122078,
+ "learning_rate": 0.0006,
+ "loss": 4.266307830810547,
+ "step": 1897
+ },
+ {
+ "epoch": 26.363318777292577,
+ "grad_norm": 0.01616556942462921,
+ "learning_rate": 0.0006,
+ "loss": 4.307868003845215,
+ "step": 1898
+ },
+ {
+ "epoch": 26.377292576419215,
+ "grad_norm": 0.016621306538581848,
+ "learning_rate": 0.0006,
+ "loss": 4.172905921936035,
+ "step": 1899
+ },
+ {
+ "epoch": 26.39126637554585,
+ "grad_norm": 0.016243206337094307,
+ "learning_rate": 0.0006,
+ "loss": 4.162896156311035,
+ "step": 1900
+ },
+ {
+ "epoch": 26.40524017467249,
+ "grad_norm": 0.016663808375597,
+ "learning_rate": 0.0006,
+ "loss": 4.258232116699219,
+ "step": 1901
+ },
+ {
+ "epoch": 26.419213973799128,
+ "grad_norm": 0.01677478663623333,
+ "learning_rate": 0.0006,
+ "loss": 4.336090087890625,
+ "step": 1902
+ },
+ {
+ "epoch": 26.433187772925763,
+ "grad_norm": 0.01641303487122059,
+ "learning_rate": 0.0006,
+ "loss": 4.191476821899414,
+ "step": 1903
+ },
+ {
+ "epoch": 26.4471615720524,
+ "grad_norm": 0.016700323671102524,
+ "learning_rate": 0.0006,
+ "loss": 4.299427509307861,
+ "step": 1904
+ },
+ {
+ "epoch": 26.46113537117904,
+ "grad_norm": 0.017595035955309868,
+ "learning_rate": 0.0006,
+ "loss": 4.312828063964844,
+ "step": 1905
+ },
+ {
+ "epoch": 26.475109170305675,
+ "grad_norm": 0.020300284028053284,
+ "learning_rate": 0.0006,
+ "loss": 4.28331184387207,
+ "step": 1906
+ },
+ {
+ "epoch": 26.489082969432314,
+ "grad_norm": 0.02198321744799614,
+ "learning_rate": 0.0006,
+ "loss": 4.3280839920043945,
+ "step": 1907
+ },
+ {
+ "epoch": 26.503056768558952,
+ "grad_norm": 0.02003200724720955,
+ "learning_rate": 0.0006,
+ "loss": 4.325623989105225,
+ "step": 1908
+ },
+ {
+ "epoch": 26.51703056768559,
+ "grad_norm": 0.01870287023484707,
+ "learning_rate": 0.0006,
+ "loss": 4.251203536987305,
+ "step": 1909
+ },
+ {
+ "epoch": 26.531004366812226,
+ "grad_norm": 0.01835448481142521,
+ "learning_rate": 0.0006,
+ "loss": 4.335144996643066,
+ "step": 1910
+ },
+ {
+ "epoch": 26.544978165938865,
+ "grad_norm": 0.02046327292919159,
+ "learning_rate": 0.0006,
+ "loss": 4.238023281097412,
+ "step": 1911
+ },
+ {
+ "epoch": 26.558951965065503,
+ "grad_norm": 0.019771311432123184,
+ "learning_rate": 0.0006,
+ "loss": 4.353594779968262,
+ "step": 1912
+ },
+ {
+ "epoch": 26.57292576419214,
+ "grad_norm": 0.018090544268488884,
+ "learning_rate": 0.0006,
+ "loss": 4.304343223571777,
+ "step": 1913
+ },
+ {
+ "epoch": 26.586899563318777,
+ "grad_norm": 0.017489777877926826,
+ "learning_rate": 0.0006,
+ "loss": 4.204375267028809,
+ "step": 1914
+ },
+ {
+ "epoch": 26.600873362445416,
+ "grad_norm": 0.016785891726613045,
+ "learning_rate": 0.0006,
+ "loss": 4.286669731140137,
+ "step": 1915
+ },
+ {
+ "epoch": 26.61484716157205,
+ "grad_norm": 0.016382789239287376,
+ "learning_rate": 0.0006,
+ "loss": 4.347662448883057,
+ "step": 1916
+ },
+ {
+ "epoch": 26.62882096069869,
+ "grad_norm": 0.016581876203417778,
+ "learning_rate": 0.0006,
+ "loss": 4.2716193199157715,
+ "step": 1917
+ },
+ {
+ "epoch": 26.64279475982533,
+ "grad_norm": 0.015627184882760048,
+ "learning_rate": 0.0006,
+ "loss": 4.311489105224609,
+ "step": 1918
+ },
+ {
+ "epoch": 26.656768558951963,
+ "grad_norm": 0.015908483415842056,
+ "learning_rate": 0.0006,
+ "loss": 4.262099742889404,
+ "step": 1919
+ },
+ {
+ "epoch": 26.670742358078602,
+ "grad_norm": 0.015576236881315708,
+ "learning_rate": 0.0006,
+ "loss": 4.275871276855469,
+ "step": 1920
+ },
+ {
+ "epoch": 26.68471615720524,
+ "grad_norm": 0.014621064998209476,
+ "learning_rate": 0.0006,
+ "loss": 4.184260845184326,
+ "step": 1921
+ },
+ {
+ "epoch": 26.69868995633188,
+ "grad_norm": 0.014797811396420002,
+ "learning_rate": 0.0006,
+ "loss": 4.3412628173828125,
+ "step": 1922
+ },
+ {
+ "epoch": 26.712663755458514,
+ "grad_norm": 0.015610570088028908,
+ "learning_rate": 0.0006,
+ "loss": 4.181094646453857,
+ "step": 1923
+ },
+ {
+ "epoch": 26.726637554585153,
+ "grad_norm": 0.016048265621066093,
+ "learning_rate": 0.0006,
+ "loss": 4.365290641784668,
+ "step": 1924
+ },
+ {
+ "epoch": 26.74061135371179,
+ "grad_norm": 0.016604071483016014,
+ "learning_rate": 0.0006,
+ "loss": 4.2628912925720215,
+ "step": 1925
+ },
+ {
+ "epoch": 26.754585152838427,
+ "grad_norm": 0.017044221982359886,
+ "learning_rate": 0.0006,
+ "loss": 4.378512382507324,
+ "step": 1926
+ },
+ {
+ "epoch": 26.768558951965066,
+ "grad_norm": 0.01746384985744953,
+ "learning_rate": 0.0006,
+ "loss": 4.3888773918151855,
+ "step": 1927
+ },
+ {
+ "epoch": 26.782532751091704,
+ "grad_norm": 0.01572711206972599,
+ "learning_rate": 0.0006,
+ "loss": 4.189935684204102,
+ "step": 1928
+ },
+ {
+ "epoch": 26.79650655021834,
+ "grad_norm": 0.014976629987359047,
+ "learning_rate": 0.0006,
+ "loss": 4.280525207519531,
+ "step": 1929
+ },
+ {
+ "epoch": 26.810480349344978,
+ "grad_norm": 0.014804664999246597,
+ "learning_rate": 0.0006,
+ "loss": 4.270630836486816,
+ "step": 1930
+ },
+ {
+ "epoch": 26.824454148471617,
+ "grad_norm": 0.014641920104622841,
+ "learning_rate": 0.0006,
+ "loss": 4.319505214691162,
+ "step": 1931
+ },
+ {
+ "epoch": 26.83842794759825,
+ "grad_norm": 0.014928505755960941,
+ "learning_rate": 0.0006,
+ "loss": 4.180147647857666,
+ "step": 1932
+ },
+ {
+ "epoch": 26.85240174672489,
+ "grad_norm": 0.016130365431308746,
+ "learning_rate": 0.0006,
+ "loss": 4.260660648345947,
+ "step": 1933
+ },
+ {
+ "epoch": 26.86637554585153,
+ "grad_norm": 0.015642493963241577,
+ "learning_rate": 0.0006,
+ "loss": 4.304064750671387,
+ "step": 1934
+ },
+ {
+ "epoch": 26.880349344978168,
+ "grad_norm": 0.015988217666745186,
+ "learning_rate": 0.0006,
+ "loss": 4.2548017501831055,
+ "step": 1935
+ },
+ {
+ "epoch": 26.894323144104803,
+ "grad_norm": 0.016501398757100105,
+ "learning_rate": 0.0006,
+ "loss": 4.260058879852295,
+ "step": 1936
+ },
+ {
+ "epoch": 26.90829694323144,
+ "grad_norm": 0.01645040698349476,
+ "learning_rate": 0.0006,
+ "loss": 4.483081340789795,
+ "step": 1937
+ },
+ {
+ "epoch": 26.92227074235808,
+ "grad_norm": 0.016140863299369812,
+ "learning_rate": 0.0006,
+ "loss": 4.1776838302612305,
+ "step": 1938
+ },
+ {
+ "epoch": 26.936244541484715,
+ "grad_norm": 0.016681019216775894,
+ "learning_rate": 0.0006,
+ "loss": 4.224747657775879,
+ "step": 1939
+ },
+ {
+ "epoch": 26.950218340611354,
+ "grad_norm": 0.01650378853082657,
+ "learning_rate": 0.0006,
+ "loss": 4.305703163146973,
+ "step": 1940
+ },
+ {
+ "epoch": 26.964192139737992,
+ "grad_norm": 0.017051683738827705,
+ "learning_rate": 0.0006,
+ "loss": 4.285835266113281,
+ "step": 1941
+ },
+ {
+ "epoch": 26.978165938864628,
+ "grad_norm": 0.016894327476620674,
+ "learning_rate": 0.0006,
+ "loss": 4.318901538848877,
+ "step": 1942
+ },
+ {
+ "epoch": 26.992139737991266,
+ "grad_norm": 0.017313100397586823,
+ "learning_rate": 0.0006,
+ "loss": 4.281434535980225,
+ "step": 1943
+ },
+ {
+ "epoch": 27.0,
+ "grad_norm": 0.018450884148478508,
+ "learning_rate": 0.0006,
+ "loss": 4.335104942321777,
+ "step": 1944
+ },
+ {
+ "epoch": 27.0,
+ "eval_loss": 4.640687465667725,
+ "eval_runtime": 56.9624,
+ "eval_samples_per_second": 42.87,
+ "eval_steps_per_second": 1.352,
+ "step": 1944
+ },
+ {
+ "epoch": 27.01397379912664,
+ "grad_norm": 0.017755698412656784,
+ "learning_rate": 0.0006,
+ "loss": 4.278824329376221,
+ "step": 1945
+ },
+ {
+ "epoch": 27.027947598253274,
+ "grad_norm": 0.01724310778081417,
+ "learning_rate": 0.0006,
+ "loss": 4.311049461364746,
+ "step": 1946
+ },
+ {
+ "epoch": 27.041921397379912,
+ "grad_norm": 0.01690123789012432,
+ "learning_rate": 0.0006,
+ "loss": 4.261994361877441,
+ "step": 1947
+ },
+ {
+ "epoch": 27.05589519650655,
+ "grad_norm": 0.018406696617603302,
+ "learning_rate": 0.0006,
+ "loss": 4.1925153732299805,
+ "step": 1948
+ },
+ {
+ "epoch": 27.069868995633186,
+ "grad_norm": 0.019945867359638214,
+ "learning_rate": 0.0006,
+ "loss": 4.108537197113037,
+ "step": 1949
+ },
+ {
+ "epoch": 27.083842794759825,
+ "grad_norm": 0.020316628739237785,
+ "learning_rate": 0.0006,
+ "loss": 4.209043025970459,
+ "step": 1950
+ },
+ {
+ "epoch": 27.097816593886463,
+ "grad_norm": 0.019690370187163353,
+ "learning_rate": 0.0006,
+ "loss": 4.195789337158203,
+ "step": 1951
+ },
+ {
+ "epoch": 27.111790393013102,
+ "grad_norm": 0.01785232499241829,
+ "learning_rate": 0.0006,
+ "loss": 4.21466064453125,
+ "step": 1952
+ },
+ {
+ "epoch": 27.125764192139737,
+ "grad_norm": 0.0164767038077116,
+ "learning_rate": 0.0006,
+ "loss": 4.270247459411621,
+ "step": 1953
+ },
+ {
+ "epoch": 27.139737991266376,
+ "grad_norm": 0.018008455634117126,
+ "learning_rate": 0.0006,
+ "loss": 4.201042175292969,
+ "step": 1954
+ },
+ {
+ "epoch": 27.153711790393015,
+ "grad_norm": 0.01807340793311596,
+ "learning_rate": 0.0006,
+ "loss": 4.25289249420166,
+ "step": 1955
+ },
+ {
+ "epoch": 27.16768558951965,
+ "grad_norm": 0.016246909275650978,
+ "learning_rate": 0.0006,
+ "loss": 4.116283416748047,
+ "step": 1956
+ },
+ {
+ "epoch": 27.18165938864629,
+ "grad_norm": 0.017545776441693306,
+ "learning_rate": 0.0006,
+ "loss": 4.316999435424805,
+ "step": 1957
+ },
+ {
+ "epoch": 27.195633187772927,
+ "grad_norm": 0.017915375530719757,
+ "learning_rate": 0.0006,
+ "loss": 4.253859519958496,
+ "step": 1958
+ },
+ {
+ "epoch": 27.209606986899562,
+ "grad_norm": 0.018112223595380783,
+ "learning_rate": 0.0006,
+ "loss": 4.2541022300720215,
+ "step": 1959
+ },
+ {
+ "epoch": 27.2235807860262,
+ "grad_norm": 0.01846139505505562,
+ "learning_rate": 0.0006,
+ "loss": 4.20438289642334,
+ "step": 1960
+ },
+ {
+ "epoch": 27.23755458515284,
+ "grad_norm": 0.01943897269666195,
+ "learning_rate": 0.0006,
+ "loss": 4.2799577713012695,
+ "step": 1961
+ },
+ {
+ "epoch": 27.251528384279474,
+ "grad_norm": 0.018253348767757416,
+ "learning_rate": 0.0006,
+ "loss": 4.1720499992370605,
+ "step": 1962
+ },
+ {
+ "epoch": 27.265502183406113,
+ "grad_norm": 0.018292246386408806,
+ "learning_rate": 0.0006,
+ "loss": 4.28231143951416,
+ "step": 1963
+ },
+ {
+ "epoch": 27.27947598253275,
+ "grad_norm": 0.01943155936896801,
+ "learning_rate": 0.0006,
+ "loss": 4.314976692199707,
+ "step": 1964
+ },
+ {
+ "epoch": 27.29344978165939,
+ "grad_norm": 0.019064994528889656,
+ "learning_rate": 0.0006,
+ "loss": 4.25328254699707,
+ "step": 1965
+ },
+ {
+ "epoch": 27.307423580786025,
+ "grad_norm": 0.017423273995518684,
+ "learning_rate": 0.0006,
+ "loss": 4.195992469787598,
+ "step": 1966
+ },
+ {
+ "epoch": 27.321397379912664,
+ "grad_norm": 0.01677924580872059,
+ "learning_rate": 0.0006,
+ "loss": 4.260043144226074,
+ "step": 1967
+ },
+ {
+ "epoch": 27.335371179039303,
+ "grad_norm": 0.01720673404633999,
+ "learning_rate": 0.0006,
+ "loss": 4.222780704498291,
+ "step": 1968
+ },
+ {
+ "epoch": 27.349344978165938,
+ "grad_norm": 0.017300743609666824,
+ "learning_rate": 0.0006,
+ "loss": 4.109524250030518,
+ "step": 1969
+ },
+ {
+ "epoch": 27.363318777292577,
+ "grad_norm": 0.016065871343016624,
+ "learning_rate": 0.0006,
+ "loss": 4.36108922958374,
+ "step": 1970
+ },
+ {
+ "epoch": 27.377292576419215,
+ "grad_norm": 0.016626616939902306,
+ "learning_rate": 0.0006,
+ "loss": 4.220630645751953,
+ "step": 1971
+ },
+ {
+ "epoch": 27.39126637554585,
+ "grad_norm": 0.016482602804899216,
+ "learning_rate": 0.0006,
+ "loss": 4.344797134399414,
+ "step": 1972
+ },
+ {
+ "epoch": 27.40524017467249,
+ "grad_norm": 0.01727953553199768,
+ "learning_rate": 0.0006,
+ "loss": 4.296645164489746,
+ "step": 1973
+ },
+ {
+ "epoch": 27.419213973799128,
+ "grad_norm": 0.01522731315344572,
+ "learning_rate": 0.0006,
+ "loss": 4.223751068115234,
+ "step": 1974
+ },
+ {
+ "epoch": 27.433187772925763,
+ "grad_norm": 0.01581776700913906,
+ "learning_rate": 0.0006,
+ "loss": 4.351379871368408,
+ "step": 1975
+ },
+ {
+ "epoch": 27.4471615720524,
+ "grad_norm": 0.016394605860114098,
+ "learning_rate": 0.0006,
+ "loss": 4.2412919998168945,
+ "step": 1976
+ },
+ {
+ "epoch": 27.46113537117904,
+ "grad_norm": 0.01751169003546238,
+ "learning_rate": 0.0006,
+ "loss": 4.257606506347656,
+ "step": 1977
+ },
+ {
+ "epoch": 27.475109170305675,
+ "grad_norm": 0.017452310770750046,
+ "learning_rate": 0.0006,
+ "loss": 4.380134582519531,
+ "step": 1978
+ },
+ {
+ "epoch": 27.489082969432314,
+ "grad_norm": 0.017741898074746132,
+ "learning_rate": 0.0006,
+ "loss": 4.279632568359375,
+ "step": 1979
+ },
+ {
+ "epoch": 27.503056768558952,
+ "grad_norm": 0.017680590972304344,
+ "learning_rate": 0.0006,
+ "loss": 4.273406028747559,
+ "step": 1980
+ },
+ {
+ "epoch": 27.51703056768559,
+ "grad_norm": 0.01864718459546566,
+ "learning_rate": 0.0006,
+ "loss": 4.278583526611328,
+ "step": 1981
+ },
+ {
+ "epoch": 27.531004366812226,
+ "grad_norm": 0.019558541476726532,
+ "learning_rate": 0.0006,
+ "loss": 4.211919784545898,
+ "step": 1982
+ },
+ {
+ "epoch": 27.544978165938865,
+ "grad_norm": 0.019467400386929512,
+ "learning_rate": 0.0006,
+ "loss": 4.276822566986084,
+ "step": 1983
+ },
+ {
+ "epoch": 27.558951965065503,
+ "grad_norm": 0.01756172813475132,
+ "learning_rate": 0.0006,
+ "loss": 4.165379047393799,
+ "step": 1984
+ },
+ {
+ "epoch": 27.57292576419214,
+ "grad_norm": 0.017458263784646988,
+ "learning_rate": 0.0006,
+ "loss": 4.2992401123046875,
+ "step": 1985
+ },
+ {
+ "epoch": 27.586899563318777,
+ "grad_norm": 0.018089665099978447,
+ "learning_rate": 0.0006,
+ "loss": 4.243378639221191,
+ "step": 1986
+ },
+ {
+ "epoch": 27.600873362445416,
+ "grad_norm": 0.019535524770617485,
+ "learning_rate": 0.0006,
+ "loss": 4.321477890014648,
+ "step": 1987
+ },
+ {
+ "epoch": 27.61484716157205,
+ "grad_norm": 0.01997970975935459,
+ "learning_rate": 0.0006,
+ "loss": 4.123082637786865,
+ "step": 1988
+ },
+ {
+ "epoch": 27.62882096069869,
+ "grad_norm": 0.01913067139685154,
+ "learning_rate": 0.0006,
+ "loss": 4.289896488189697,
+ "step": 1989
+ },
+ {
+ "epoch": 27.64279475982533,
+ "grad_norm": 0.0169806070625782,
+ "learning_rate": 0.0006,
+ "loss": 4.247004985809326,
+ "step": 1990
+ },
+ {
+ "epoch": 27.656768558951963,
+ "grad_norm": 0.017315007746219635,
+ "learning_rate": 0.0006,
+ "loss": 4.2095441818237305,
+ "step": 1991
+ },
+ {
+ "epoch": 27.670742358078602,
+ "grad_norm": 0.015098759904503822,
+ "learning_rate": 0.0006,
+ "loss": 4.145560264587402,
+ "step": 1992
+ },
+ {
+ "epoch": 27.68471615720524,
+ "grad_norm": 0.016240514814853668,
+ "learning_rate": 0.0006,
+ "loss": 4.254927635192871,
+ "step": 1993
+ },
+ {
+ "epoch": 27.69868995633188,
+ "grad_norm": 0.01621108129620552,
+ "learning_rate": 0.0006,
+ "loss": 4.219725608825684,
+ "step": 1994
+ },
+ {
+ "epoch": 27.712663755458514,
+ "grad_norm": 0.015191009268164635,
+ "learning_rate": 0.0006,
+ "loss": 4.194345474243164,
+ "step": 1995
+ },
+ {
+ "epoch": 27.726637554585153,
+ "grad_norm": 0.014387983828783035,
+ "learning_rate": 0.0006,
+ "loss": 4.211467742919922,
+ "step": 1996
+ },
+ {
+ "epoch": 27.74061135371179,
+ "grad_norm": 0.015460561029613018,
+ "learning_rate": 0.0006,
+ "loss": 4.358700752258301,
+ "step": 1997
+ },
+ {
+ "epoch": 27.754585152838427,
+ "grad_norm": 0.015100076794624329,
+ "learning_rate": 0.0006,
+ "loss": 4.171223163604736,
+ "step": 1998
+ },
+ {
+ "epoch": 27.768558951965066,
+ "grad_norm": 0.015607891604304314,
+ "learning_rate": 0.0006,
+ "loss": 4.217874526977539,
+ "step": 1999
+ },
+ {
+ "epoch": 27.782532751091704,
+ "grad_norm": 0.015650689601898193,
+ "learning_rate": 0.0006,
+ "loss": 4.280780792236328,
+ "step": 2000
+ },
+ {
+ "epoch": 27.79650655021834,
+ "grad_norm": 0.01580994389951229,
+ "learning_rate": 0.0006,
+ "loss": 4.159212112426758,
+ "step": 2001
+ },
+ {
+ "epoch": 27.810480349344978,
+ "grad_norm": 0.016607413068413734,
+ "learning_rate": 0.0006,
+ "loss": 4.238853454589844,
+ "step": 2002
+ },
+ {
+ "epoch": 27.824454148471617,
+ "grad_norm": 0.017383860424160957,
+ "learning_rate": 0.0006,
+ "loss": 4.233944892883301,
+ "step": 2003
+ },
+ {
+ "epoch": 27.83842794759825,
+ "grad_norm": 0.014966659247875214,
+ "learning_rate": 0.0006,
+ "loss": 4.167392730712891,
+ "step": 2004
+ },
+ {
+ "epoch": 27.85240174672489,
+ "grad_norm": 0.014589878730475903,
+ "learning_rate": 0.0006,
+ "loss": 4.2390899658203125,
+ "step": 2005
+ },
+ {
+ "epoch": 27.86637554585153,
+ "grad_norm": 0.017179450020194054,
+ "learning_rate": 0.0006,
+ "loss": 4.365785598754883,
+ "step": 2006
+ },
+ {
+ "epoch": 27.880349344978168,
+ "grad_norm": 0.01893901452422142,
+ "learning_rate": 0.0006,
+ "loss": 4.289430141448975,
+ "step": 2007
+ },
+ {
+ "epoch": 27.894323144104803,
+ "grad_norm": 0.018619712442159653,
+ "learning_rate": 0.0006,
+ "loss": 4.310665130615234,
+ "step": 2008
+ },
+ {
+ "epoch": 27.90829694323144,
+ "grad_norm": 0.019180385395884514,
+ "learning_rate": 0.0006,
+ "loss": 4.27454137802124,
+ "step": 2009
+ },
+ {
+ "epoch": 27.92227074235808,
+ "grad_norm": 0.017811523750424385,
+ "learning_rate": 0.0006,
+ "loss": 4.312562465667725,
+ "step": 2010
+ },
+ {
+ "epoch": 27.936244541484715,
+ "grad_norm": 0.016583573073148727,
+ "learning_rate": 0.0006,
+ "loss": 4.202253341674805,
+ "step": 2011
+ },
+ {
+ "epoch": 27.950218340611354,
+ "grad_norm": 0.01658615842461586,
+ "learning_rate": 0.0006,
+ "loss": 4.183433532714844,
+ "step": 2012
+ },
+ {
+ "epoch": 27.964192139737992,
+ "grad_norm": 0.015327401459217072,
+ "learning_rate": 0.0006,
+ "loss": 4.2125139236450195,
+ "step": 2013
+ },
+ {
+ "epoch": 27.978165938864628,
+ "grad_norm": 0.015004601329565048,
+ "learning_rate": 0.0006,
+ "loss": 4.257392883300781,
+ "step": 2014
+ },
+ {
+ "epoch": 27.992139737991266,
+ "grad_norm": 0.015259161591529846,
+ "learning_rate": 0.0006,
+ "loss": 4.241818428039551,
+ "step": 2015
+ },
+ {
+ "epoch": 28.0,
+ "grad_norm": 0.016808776184916496,
+ "learning_rate": 0.0006,
+ "loss": 4.146241188049316,
+ "step": 2016
+ },
+ {
+ "epoch": 28.0,
+ "eval_loss": 4.663090705871582,
+ "eval_runtime": 56.4559,
+ "eval_samples_per_second": 43.255,
+ "eval_steps_per_second": 1.364,
+ "step": 2016
+ },
+ {
+ "epoch": 28.01397379912664,
+ "grad_norm": 0.015784382820129395,
+ "learning_rate": 0.0006,
+ "loss": 4.044191360473633,
+ "step": 2017
+ },
+ {
+ "epoch": 28.027947598253274,
+ "grad_norm": 0.015636710450053215,
+ "learning_rate": 0.0006,
+ "loss": 4.128754615783691,
+ "step": 2018
+ },
+ {
+ "epoch": 28.041921397379912,
+ "grad_norm": 0.015364975668489933,
+ "learning_rate": 0.0006,
+ "loss": 4.191835403442383,
+ "step": 2019
+ },
+ {
+ "epoch": 28.05589519650655,
+ "grad_norm": 0.015015700832009315,
+ "learning_rate": 0.0006,
+ "loss": 4.120022773742676,
+ "step": 2020
+ },
+ {
+ "epoch": 28.069868995633186,
+ "grad_norm": 0.016666799783706665,
+ "learning_rate": 0.0006,
+ "loss": 4.33906364440918,
+ "step": 2021
+ },
+ {
+ "epoch": 28.083842794759825,
+ "grad_norm": 0.01780104823410511,
+ "learning_rate": 0.0006,
+ "loss": 4.279458999633789,
+ "step": 2022
+ },
+ {
+ "epoch": 28.097816593886463,
+ "grad_norm": 0.01870882883667946,
+ "learning_rate": 0.0006,
+ "loss": 4.34728479385376,
+ "step": 2023
+ },
+ {
+ "epoch": 28.111790393013102,
+ "grad_norm": 0.018406381830573082,
+ "learning_rate": 0.0006,
+ "loss": 4.211311340332031,
+ "step": 2024
+ },
+ {
+ "epoch": 28.125764192139737,
+ "grad_norm": 0.017833448946475983,
+ "learning_rate": 0.0006,
+ "loss": 4.223254203796387,
+ "step": 2025
+ },
+ {
+ "epoch": 28.139737991266376,
+ "grad_norm": 0.017552688717842102,
+ "learning_rate": 0.0006,
+ "loss": 4.208823204040527,
+ "step": 2026
+ },
+ {
+ "epoch": 28.153711790393015,
+ "grad_norm": 0.018500229343771935,
+ "learning_rate": 0.0006,
+ "loss": 4.198025226593018,
+ "step": 2027
+ },
+ {
+ "epoch": 28.16768558951965,
+ "grad_norm": 0.01782156340777874,
+ "learning_rate": 0.0006,
+ "loss": 4.179978370666504,
+ "step": 2028
+ },
+ {
+ "epoch": 28.18165938864629,
+ "grad_norm": 0.019580967724323273,
+ "learning_rate": 0.0006,
+ "loss": 4.291114807128906,
+ "step": 2029
+ },
+ {
+ "epoch": 28.195633187772927,
+ "grad_norm": 0.020804036408662796,
+ "learning_rate": 0.0006,
+ "loss": 4.220607280731201,
+ "step": 2030
+ },
+ {
+ "epoch": 28.209606986899562,
+ "grad_norm": 0.019595693796873093,
+ "learning_rate": 0.0006,
+ "loss": 4.2905731201171875,
+ "step": 2031
+ },
+ {
+ "epoch": 28.2235807860262,
+ "grad_norm": 0.019247086718678474,
+ "learning_rate": 0.0006,
+ "loss": 4.2467360496521,
+ "step": 2032
+ },
+ {
+ "epoch": 28.23755458515284,
+ "grad_norm": 0.01958036608994007,
+ "learning_rate": 0.0006,
+ "loss": 4.286137580871582,
+ "step": 2033
+ },
+ {
+ "epoch": 28.251528384279474,
+ "grad_norm": 0.021354753524065018,
+ "learning_rate": 0.0006,
+ "loss": 4.234755039215088,
+ "step": 2034
+ },
+ {
+ "epoch": 28.265502183406113,
+ "grad_norm": 0.02206375077366829,
+ "learning_rate": 0.0006,
+ "loss": 4.149641036987305,
+ "step": 2035
+ },
+ {
+ "epoch": 28.27947598253275,
+ "grad_norm": 0.02201310358941555,
+ "learning_rate": 0.0006,
+ "loss": 4.1928629875183105,
+ "step": 2036
+ },
+ {
+ "epoch": 28.29344978165939,
+ "grad_norm": 0.01961950585246086,
+ "learning_rate": 0.0006,
+ "loss": 4.211404323577881,
+ "step": 2037
+ },
+ {
+ "epoch": 28.307423580786025,
+ "grad_norm": 0.019211795181035995,
+ "learning_rate": 0.0006,
+ "loss": 4.200976848602295,
+ "step": 2038
+ },
+ {
+ "epoch": 28.321397379912664,
+ "grad_norm": 0.01779370754957199,
+ "learning_rate": 0.0006,
+ "loss": 4.249148845672607,
+ "step": 2039
+ },
+ {
+ "epoch": 28.335371179039303,
+ "grad_norm": 0.018635908141732216,
+ "learning_rate": 0.0006,
+ "loss": 4.11806058883667,
+ "step": 2040
+ },
+ {
+ "epoch": 28.349344978165938,
+ "grad_norm": 0.019792694598436356,
+ "learning_rate": 0.0006,
+ "loss": 4.335224151611328,
+ "step": 2041
+ },
+ {
+ "epoch": 28.363318777292577,
+ "grad_norm": 0.021731717512011528,
+ "learning_rate": 0.0006,
+ "loss": 4.302778244018555,
+ "step": 2042
+ },
+ {
+ "epoch": 28.377292576419215,
+ "grad_norm": 0.020823447033762932,
+ "learning_rate": 0.0006,
+ "loss": 4.260610580444336,
+ "step": 2043
+ },
+ {
+ "epoch": 28.39126637554585,
+ "grad_norm": 0.019385412335395813,
+ "learning_rate": 0.0006,
+ "loss": 4.2247724533081055,
+ "step": 2044
+ },
+ {
+ "epoch": 28.40524017467249,
+ "grad_norm": 0.019539108499884605,
+ "learning_rate": 0.0006,
+ "loss": 4.236655235290527,
+ "step": 2045
+ },
+ {
+ "epoch": 28.419213973799128,
+ "grad_norm": 0.01942690648138523,
+ "learning_rate": 0.0006,
+ "loss": 4.251850605010986,
+ "step": 2046
+ },
+ {
+ "epoch": 28.433187772925763,
+ "grad_norm": 0.019667886197566986,
+ "learning_rate": 0.0006,
+ "loss": 4.222312927246094,
+ "step": 2047
+ },
+ {
+ "epoch": 28.4471615720524,
+ "grad_norm": 0.02029012329876423,
+ "learning_rate": 0.0006,
+ "loss": 4.246252059936523,
+ "step": 2048
+ },
+ {
+ "epoch": 28.46113537117904,
+ "grad_norm": 0.01784469373524189,
+ "learning_rate": 0.0006,
+ "loss": 4.192128658294678,
+ "step": 2049
+ },
+ {
+ "epoch": 28.475109170305675,
+ "grad_norm": 0.0160287544131279,
+ "learning_rate": 0.0006,
+ "loss": 4.208607196807861,
+ "step": 2050
+ },
+ {
+ "epoch": 28.489082969432314,
+ "grad_norm": 0.015348346903920174,
+ "learning_rate": 0.0006,
+ "loss": 4.236598014831543,
+ "step": 2051
+ },
+ {
+ "epoch": 28.503056768558952,
+ "grad_norm": 0.015198206529021263,
+ "learning_rate": 0.0006,
+ "loss": 4.235552787780762,
+ "step": 2052
+ },
+ {
+ "epoch": 28.51703056768559,
+ "grad_norm": 0.01523754745721817,
+ "learning_rate": 0.0006,
+ "loss": 4.1941728591918945,
+ "step": 2053
+ },
+ {
+ "epoch": 28.531004366812226,
+ "grad_norm": 0.015559614636003971,
+ "learning_rate": 0.0006,
+ "loss": 4.294757843017578,
+ "step": 2054
+ },
+ {
+ "epoch": 28.544978165938865,
+ "grad_norm": 0.01568358950316906,
+ "learning_rate": 0.0006,
+ "loss": 4.213065147399902,
+ "step": 2055
+ },
+ {
+ "epoch": 28.558951965065503,
+ "grad_norm": 0.015336516313254833,
+ "learning_rate": 0.0006,
+ "loss": 4.240293979644775,
+ "step": 2056
+ },
+ {
+ "epoch": 28.57292576419214,
+ "grad_norm": 0.015041496604681015,
+ "learning_rate": 0.0006,
+ "loss": 4.273397445678711,
+ "step": 2057
+ },
+ {
+ "epoch": 28.586899563318777,
+ "grad_norm": 0.01610128954052925,
+ "learning_rate": 0.0006,
+ "loss": 4.125369071960449,
+ "step": 2058
+ },
+ {
+ "epoch": 28.600873362445416,
+ "grad_norm": 0.015974976122379303,
+ "learning_rate": 0.0006,
+ "loss": 4.133459091186523,
+ "step": 2059
+ },
+ {
+ "epoch": 28.61484716157205,
+ "grad_norm": 0.015936799347400665,
+ "learning_rate": 0.0006,
+ "loss": 4.306667327880859,
+ "step": 2060
+ },
+ {
+ "epoch": 28.62882096069869,
+ "grad_norm": 0.01795247197151184,
+ "learning_rate": 0.0006,
+ "loss": 4.205156326293945,
+ "step": 2061
+ },
+ {
+ "epoch": 28.64279475982533,
+ "grad_norm": 0.019639814272522926,
+ "learning_rate": 0.0006,
+ "loss": 4.277614593505859,
+ "step": 2062
+ },
+ {
+ "epoch": 28.656768558951963,
+ "grad_norm": 0.01722300611436367,
+ "learning_rate": 0.0006,
+ "loss": 4.191032886505127,
+ "step": 2063
+ },
+ {
+ "epoch": 28.670742358078602,
+ "grad_norm": 0.01612633652985096,
+ "learning_rate": 0.0006,
+ "loss": 4.20717191696167,
+ "step": 2064
+ },
+ {
+ "epoch": 28.68471615720524,
+ "grad_norm": 0.01603279449045658,
+ "learning_rate": 0.0006,
+ "loss": 4.228884220123291,
+ "step": 2065
+ },
+ {
+ "epoch": 28.69868995633188,
+ "grad_norm": 0.016405927017331123,
+ "learning_rate": 0.0006,
+ "loss": 4.201292991638184,
+ "step": 2066
+ },
+ {
+ "epoch": 28.712663755458514,
+ "grad_norm": 0.016644183546304703,
+ "learning_rate": 0.0006,
+ "loss": 4.239282131195068,
+ "step": 2067
+ },
+ {
+ "epoch": 28.726637554585153,
+ "grad_norm": 0.0175587497651577,
+ "learning_rate": 0.0006,
+ "loss": 4.289038181304932,
+ "step": 2068
+ },
+ {
+ "epoch": 28.74061135371179,
+ "grad_norm": 0.018223397433757782,
+ "learning_rate": 0.0006,
+ "loss": 4.2266998291015625,
+ "step": 2069
+ },
+ {
+ "epoch": 28.754585152838427,
+ "grad_norm": 0.018680868670344353,
+ "learning_rate": 0.0006,
+ "loss": 4.348849296569824,
+ "step": 2070
+ },
+ {
+ "epoch": 28.768558951965066,
+ "grad_norm": 0.016726728528738022,
+ "learning_rate": 0.0006,
+ "loss": 4.292080879211426,
+ "step": 2071
+ },
+ {
+ "epoch": 28.782532751091704,
+ "grad_norm": 0.017386795952916145,
+ "learning_rate": 0.0006,
+ "loss": 4.208870887756348,
+ "step": 2072
+ },
+ {
+ "epoch": 28.79650655021834,
+ "grad_norm": 0.01865178905427456,
+ "learning_rate": 0.0006,
+ "loss": 4.054150581359863,
+ "step": 2073
+ },
+ {
+ "epoch": 28.810480349344978,
+ "grad_norm": 0.01776815392076969,
+ "learning_rate": 0.0006,
+ "loss": 4.202404499053955,
+ "step": 2074
+ },
+ {
+ "epoch": 28.824454148471617,
+ "grad_norm": 0.01643497310578823,
+ "learning_rate": 0.0006,
+ "loss": 4.131855010986328,
+ "step": 2075
+ },
+ {
+ "epoch": 28.83842794759825,
+ "grad_norm": 0.01574273779988289,
+ "learning_rate": 0.0006,
+ "loss": 4.296788692474365,
+ "step": 2076
+ },
+ {
+ "epoch": 28.85240174672489,
+ "grad_norm": 0.015613908879458904,
+ "learning_rate": 0.0006,
+ "loss": 4.210949897766113,
+ "step": 2077
+ },
+ {
+ "epoch": 28.86637554585153,
+ "grad_norm": 0.015543212182819843,
+ "learning_rate": 0.0006,
+ "loss": 4.218136310577393,
+ "step": 2078
+ },
+ {
+ "epoch": 28.880349344978168,
+ "grad_norm": 0.014681251719594002,
+ "learning_rate": 0.0006,
+ "loss": 4.19401741027832,
+ "step": 2079
+ },
+ {
+ "epoch": 28.894323144104803,
+ "grad_norm": 0.014393828809261322,
+ "learning_rate": 0.0006,
+ "loss": 4.198468208312988,
+ "step": 2080
+ },
+ {
+ "epoch": 28.90829694323144,
+ "grad_norm": 0.01580170728266239,
+ "learning_rate": 0.0006,
+ "loss": 4.224969863891602,
+ "step": 2081
+ },
+ {
+ "epoch": 28.92227074235808,
+ "grad_norm": 0.01484165620058775,
+ "learning_rate": 0.0006,
+ "loss": 4.12236213684082,
+ "step": 2082
+ },
+ {
+ "epoch": 28.936244541484715,
+ "grad_norm": 0.01497623696923256,
+ "learning_rate": 0.0006,
+ "loss": 4.24141788482666,
+ "step": 2083
+ },
+ {
+ "epoch": 28.950218340611354,
+ "grad_norm": 0.015080046840012074,
+ "learning_rate": 0.0006,
+ "loss": 4.284151077270508,
+ "step": 2084
+ },
+ {
+ "epoch": 28.964192139737992,
+ "grad_norm": 0.016207212582230568,
+ "learning_rate": 0.0006,
+ "loss": 4.298693656921387,
+ "step": 2085
+ },
+ {
+ "epoch": 28.978165938864628,
+ "grad_norm": 0.015474417246878147,
+ "learning_rate": 0.0006,
+ "loss": 4.2412261962890625,
+ "step": 2086
+ },
+ {
+ "epoch": 28.992139737991266,
+ "grad_norm": 0.01515351701527834,
+ "learning_rate": 0.0006,
+ "loss": 4.3054423332214355,
+ "step": 2087
+ },
+ {
+ "epoch": 29.0,
+ "grad_norm": 0.016411345452070236,
+ "learning_rate": 0.0006,
+ "loss": 4.104830741882324,
+ "step": 2088
+ },
+ {
+ "epoch": 29.0,
+ "eval_loss": 4.661567687988281,
+ "eval_runtime": 57.1244,
+ "eval_samples_per_second": 42.749,
+ "eval_steps_per_second": 1.348,
+ "step": 2088
+ },
+ {
+ "epoch": 29.01397379912664,
+ "grad_norm": 0.01665342226624489,
+ "learning_rate": 0.0006,
+ "loss": 4.311391830444336,
+ "step": 2089
+ },
+ {
+ "epoch": 29.027947598253274,
+ "grad_norm": 0.018053214997053146,
+ "learning_rate": 0.0006,
+ "loss": 4.135597229003906,
+ "step": 2090
+ },
+ {
+ "epoch": 29.041921397379912,
+ "grad_norm": 0.018213748931884766,
+ "learning_rate": 0.0006,
+ "loss": 4.303299903869629,
+ "step": 2091
+ },
+ {
+ "epoch": 29.05589519650655,
+ "grad_norm": 0.018460242077708244,
+ "learning_rate": 0.0006,
+ "loss": 4.20850944519043,
+ "step": 2092
+ },
+ {
+ "epoch": 29.069868995633186,
+ "grad_norm": 0.018365809693932533,
+ "learning_rate": 0.0006,
+ "loss": 4.275847434997559,
+ "step": 2093
+ },
+ {
+ "epoch": 29.083842794759825,
+ "grad_norm": 0.018346186727285385,
+ "learning_rate": 0.0006,
+ "loss": 4.202091217041016,
+ "step": 2094
+ },
+ {
+ "epoch": 29.097816593886463,
+ "grad_norm": 0.017920587211847305,
+ "learning_rate": 0.0006,
+ "loss": 4.182092666625977,
+ "step": 2095
+ },
+ {
+ "epoch": 29.111790393013102,
+ "grad_norm": 0.01812003180384636,
+ "learning_rate": 0.0006,
+ "loss": 4.259721755981445,
+ "step": 2096
+ },
+ {
+ "epoch": 29.125764192139737,
+ "grad_norm": 0.01700986735522747,
+ "learning_rate": 0.0006,
+ "loss": 4.067573070526123,
+ "step": 2097
+ },
+ {
+ "epoch": 29.139737991266376,
+ "grad_norm": 0.018069760873913765,
+ "learning_rate": 0.0006,
+ "loss": 4.234400749206543,
+ "step": 2098
+ },
+ {
+ "epoch": 29.153711790393015,
+ "grad_norm": 0.01905428245663643,
+ "learning_rate": 0.0006,
+ "loss": 4.16609001159668,
+ "step": 2099
+ },
+ {
+ "epoch": 29.16768558951965,
+ "grad_norm": 0.018711242824792862,
+ "learning_rate": 0.0006,
+ "loss": 4.249088764190674,
+ "step": 2100
+ },
+ {
+ "epoch": 29.18165938864629,
+ "grad_norm": 0.018721306696534157,
+ "learning_rate": 0.0006,
+ "loss": 4.14540958404541,
+ "step": 2101
+ },
+ {
+ "epoch": 29.195633187772927,
+ "grad_norm": 0.020139874890446663,
+ "learning_rate": 0.0006,
+ "loss": 4.049248695373535,
+ "step": 2102
+ },
+ {
+ "epoch": 29.209606986899562,
+ "grad_norm": 0.020849574357271194,
+ "learning_rate": 0.0006,
+ "loss": 4.235224723815918,
+ "step": 2103
+ },
+ {
+ "epoch": 29.2235807860262,
+ "grad_norm": 0.018693549558520317,
+ "learning_rate": 0.0006,
+ "loss": 4.232961177825928,
+ "step": 2104
+ },
+ {
+ "epoch": 29.23755458515284,
+ "grad_norm": 0.017889857292175293,
+ "learning_rate": 0.0006,
+ "loss": 4.171422004699707,
+ "step": 2105
+ },
+ {
+ "epoch": 29.251528384279474,
+ "grad_norm": 0.016690224409103394,
+ "learning_rate": 0.0006,
+ "loss": 4.244717597961426,
+ "step": 2106
+ },
+ {
+ "epoch": 29.265502183406113,
+ "grad_norm": 0.01708616502583027,
+ "learning_rate": 0.0006,
+ "loss": 4.125068664550781,
+ "step": 2107
+ },
+ {
+ "epoch": 29.27947598253275,
+ "grad_norm": 0.017928237095475197,
+ "learning_rate": 0.0006,
+ "loss": 4.208292007446289,
+ "step": 2108
+ },
+ {
+ "epoch": 29.29344978165939,
+ "grad_norm": 0.018250394612550735,
+ "learning_rate": 0.0006,
+ "loss": 4.118630886077881,
+ "step": 2109
+ },
+ {
+ "epoch": 29.307423580786025,
+ "grad_norm": 0.020160246640443802,
+ "learning_rate": 0.0006,
+ "loss": 4.206025123596191,
+ "step": 2110
+ },
+ {
+ "epoch": 29.321397379912664,
+ "grad_norm": 0.020045453682541847,
+ "learning_rate": 0.0006,
+ "loss": 4.2758684158325195,
+ "step": 2111
+ },
+ {
+ "epoch": 29.335371179039303,
+ "grad_norm": 0.019556907936930656,
+ "learning_rate": 0.0006,
+ "loss": 4.188272953033447,
+ "step": 2112
+ },
+ {
+ "epoch": 29.349344978165938,
+ "grad_norm": 0.019185200333595276,
+ "learning_rate": 0.0006,
+ "loss": 4.17451286315918,
+ "step": 2113
+ },
+ {
+ "epoch": 29.363318777292577,
+ "grad_norm": 0.01816629432141781,
+ "learning_rate": 0.0006,
+ "loss": 4.165246963500977,
+ "step": 2114
+ },
+ {
+ "epoch": 29.377292576419215,
+ "grad_norm": 0.01865146867930889,
+ "learning_rate": 0.0006,
+ "loss": 4.148184776306152,
+ "step": 2115
+ },
+ {
+ "epoch": 29.39126637554585,
+ "grad_norm": 0.018338464200496674,
+ "learning_rate": 0.0006,
+ "loss": 4.288936614990234,
+ "step": 2116
+ },
+ {
+ "epoch": 29.40524017467249,
+ "grad_norm": 0.018986187875270844,
+ "learning_rate": 0.0006,
+ "loss": 4.183379650115967,
+ "step": 2117
+ },
+ {
+ "epoch": 29.419213973799128,
+ "grad_norm": 0.018409637734293938,
+ "learning_rate": 0.0006,
+ "loss": 4.057574272155762,
+ "step": 2118
+ },
+ {
+ "epoch": 29.433187772925763,
+ "grad_norm": 0.016664542257785797,
+ "learning_rate": 0.0006,
+ "loss": 4.18517541885376,
+ "step": 2119
+ },
+ {
+ "epoch": 29.4471615720524,
+ "grad_norm": 0.016486089676618576,
+ "learning_rate": 0.0006,
+ "loss": 4.179367542266846,
+ "step": 2120
+ },
+ {
+ "epoch": 29.46113537117904,
+ "grad_norm": 0.01787523552775383,
+ "learning_rate": 0.0006,
+ "loss": 4.184126377105713,
+ "step": 2121
+ },
+ {
+ "epoch": 29.475109170305675,
+ "grad_norm": 0.016574613749980927,
+ "learning_rate": 0.0006,
+ "loss": 4.290170192718506,
+ "step": 2122
+ },
+ {
+ "epoch": 29.489082969432314,
+ "grad_norm": 0.01643376611173153,
+ "learning_rate": 0.0006,
+ "loss": 4.052913665771484,
+ "step": 2123
+ },
+ {
+ "epoch": 29.503056768558952,
+ "grad_norm": 0.016503969207406044,
+ "learning_rate": 0.0006,
+ "loss": 4.335184097290039,
+ "step": 2124
+ },
+ {
+ "epoch": 29.51703056768559,
+ "grad_norm": 0.017596479505300522,
+ "learning_rate": 0.0006,
+ "loss": 4.165539741516113,
+ "step": 2125
+ },
+ {
+ "epoch": 29.531004366812226,
+ "grad_norm": 0.017468582838773727,
+ "learning_rate": 0.0006,
+ "loss": 4.315964698791504,
+ "step": 2126
+ },
+ {
+ "epoch": 29.544978165938865,
+ "grad_norm": 0.01724618673324585,
+ "learning_rate": 0.0006,
+ "loss": 4.243681907653809,
+ "step": 2127
+ },
+ {
+ "epoch": 29.558951965065503,
+ "grad_norm": 0.018384616822004318,
+ "learning_rate": 0.0006,
+ "loss": 4.112447261810303,
+ "step": 2128
+ },
+ {
+ "epoch": 29.57292576419214,
+ "grad_norm": 0.01824074238538742,
+ "learning_rate": 0.0006,
+ "loss": 4.236065864562988,
+ "step": 2129
+ },
+ {
+ "epoch": 29.586899563318777,
+ "grad_norm": 0.01625971868634224,
+ "learning_rate": 0.0006,
+ "loss": 4.18143892288208,
+ "step": 2130
+ },
+ {
+ "epoch": 29.600873362445416,
+ "grad_norm": 0.015537494793534279,
+ "learning_rate": 0.0006,
+ "loss": 4.225147247314453,
+ "step": 2131
+ },
+ {
+ "epoch": 29.61484716157205,
+ "grad_norm": 0.015818974003195763,
+ "learning_rate": 0.0006,
+ "loss": 4.3100175857543945,
+ "step": 2132
+ },
+ {
+ "epoch": 29.62882096069869,
+ "grad_norm": 0.015909671783447266,
+ "learning_rate": 0.0006,
+ "loss": 4.169775009155273,
+ "step": 2133
+ },
+ {
+ "epoch": 29.64279475982533,
+ "grad_norm": 0.01769658550620079,
+ "learning_rate": 0.0006,
+ "loss": 4.278904914855957,
+ "step": 2134
+ },
+ {
+ "epoch": 29.656768558951963,
+ "grad_norm": 0.015641704201698303,
+ "learning_rate": 0.0006,
+ "loss": 4.2426228523254395,
+ "step": 2135
+ },
+ {
+ "epoch": 29.670742358078602,
+ "grad_norm": 0.01565658301115036,
+ "learning_rate": 0.0006,
+ "loss": 4.232687950134277,
+ "step": 2136
+ },
+ {
+ "epoch": 29.68471615720524,
+ "grad_norm": 0.016485940665006638,
+ "learning_rate": 0.0006,
+ "loss": 4.209839820861816,
+ "step": 2137
+ },
+ {
+ "epoch": 29.69868995633188,
+ "grad_norm": 0.017622729763388634,
+ "learning_rate": 0.0006,
+ "loss": 4.196017265319824,
+ "step": 2138
+ },
+ {
+ "epoch": 29.712663755458514,
+ "grad_norm": 0.018320564180612564,
+ "learning_rate": 0.0006,
+ "loss": 4.225683689117432,
+ "step": 2139
+ },
+ {
+ "epoch": 29.726637554585153,
+ "grad_norm": 0.018140794709324837,
+ "learning_rate": 0.0006,
+ "loss": 4.311944961547852,
+ "step": 2140
+ },
+ {
+ "epoch": 29.74061135371179,
+ "grad_norm": 0.016227100044488907,
+ "learning_rate": 0.0006,
+ "loss": 4.230257511138916,
+ "step": 2141
+ },
+ {
+ "epoch": 29.754585152838427,
+ "grad_norm": 0.015759488567709923,
+ "learning_rate": 0.0006,
+ "loss": 4.119174003601074,
+ "step": 2142
+ },
+ {
+ "epoch": 29.768558951965066,
+ "grad_norm": 0.01757766678929329,
+ "learning_rate": 0.0006,
+ "loss": 4.165826797485352,
+ "step": 2143
+ },
+ {
+ "epoch": 29.782532751091704,
+ "grad_norm": 0.019661923870444298,
+ "learning_rate": 0.0006,
+ "loss": 4.392203330993652,
+ "step": 2144
+ },
+ {
+ "epoch": 29.79650655021834,
+ "grad_norm": 0.019478755071759224,
+ "learning_rate": 0.0006,
+ "loss": 4.348371982574463,
+ "step": 2145
+ },
+ {
+ "epoch": 29.810480349344978,
+ "grad_norm": 0.019149569794535637,
+ "learning_rate": 0.0006,
+ "loss": 4.34104061126709,
+ "step": 2146
+ },
+ {
+ "epoch": 29.824454148471617,
+ "grad_norm": 0.017707521095871925,
+ "learning_rate": 0.0006,
+ "loss": 4.228226661682129,
+ "step": 2147
+ },
+ {
+ "epoch": 29.83842794759825,
+ "grad_norm": 0.015615378506481647,
+ "learning_rate": 0.0006,
+ "loss": 4.175541877746582,
+ "step": 2148
+ },
+ {
+ "epoch": 29.85240174672489,
+ "grad_norm": 0.016111129894852638,
+ "learning_rate": 0.0006,
+ "loss": 4.267936706542969,
+ "step": 2149
+ },
+ {
+ "epoch": 29.86637554585153,
+ "grad_norm": 0.01528779324144125,
+ "learning_rate": 0.0006,
+ "loss": 4.186014175415039,
+ "step": 2150
+ },
+ {
+ "epoch": 29.880349344978168,
+ "grad_norm": 0.014925646595656872,
+ "learning_rate": 0.0006,
+ "loss": 4.174860000610352,
+ "step": 2151
+ },
+ {
+ "epoch": 29.894323144104803,
+ "grad_norm": 0.016917673870921135,
+ "learning_rate": 0.0006,
+ "loss": 4.33908748626709,
+ "step": 2152
+ },
+ {
+ "epoch": 29.90829694323144,
+ "grad_norm": 0.016007075086236,
+ "learning_rate": 0.0006,
+ "loss": 4.266141414642334,
+ "step": 2153
+ },
+ {
+ "epoch": 29.92227074235808,
+ "grad_norm": 0.016178956255316734,
+ "learning_rate": 0.0006,
+ "loss": 4.271215915679932,
+ "step": 2154
+ },
+ {
+ "epoch": 29.936244541484715,
+ "grad_norm": 0.016180871054530144,
+ "learning_rate": 0.0006,
+ "loss": 4.278801441192627,
+ "step": 2155
+ },
+ {
+ "epoch": 29.950218340611354,
+ "grad_norm": 0.015335503034293652,
+ "learning_rate": 0.0006,
+ "loss": 4.194978713989258,
+ "step": 2156
+ },
+ {
+ "epoch": 29.964192139737992,
+ "grad_norm": 0.016580764204263687,
+ "learning_rate": 0.0006,
+ "loss": 4.290809631347656,
+ "step": 2157
+ },
+ {
+ "epoch": 29.978165938864628,
+ "grad_norm": 0.01623128168284893,
+ "learning_rate": 0.0006,
+ "loss": 4.205451488494873,
+ "step": 2158
+ },
+ {
+ "epoch": 29.992139737991266,
+ "grad_norm": 0.014904248528182507,
+ "learning_rate": 0.0006,
+ "loss": 4.24924373626709,
+ "step": 2159
+ },
+ {
+ "epoch": 30.0,
+ "grad_norm": 0.017089299857616425,
+ "learning_rate": 0.0006,
+ "loss": 4.222072601318359,
+ "step": 2160
+ },
+ {
+ "epoch": 30.0,
+ "eval_loss": 4.585411071777344,
+ "eval_runtime": 57.1164,
+ "eval_samples_per_second": 42.755,
+ "eval_steps_per_second": 1.348,
+ "step": 2160
+ },
+ {
+ "epoch": 30.01397379912664,
+ "grad_norm": 0.017226383090019226,
+ "learning_rate": 0.0006,
+ "loss": 4.19467830657959,
+ "step": 2161
+ },
+ {
+ "epoch": 30.027947598253274,
+ "grad_norm": 0.01635066606104374,
+ "learning_rate": 0.0006,
+ "loss": 4.176875591278076,
+ "step": 2162
+ },
+ {
+ "epoch": 30.041921397379912,
+ "grad_norm": 0.01735362410545349,
+ "learning_rate": 0.0006,
+ "loss": 4.13667106628418,
+ "step": 2163
+ },
+ {
+ "epoch": 30.05589519650655,
+ "grad_norm": 0.017047109082341194,
+ "learning_rate": 0.0006,
+ "loss": 4.136329650878906,
+ "step": 2164
+ },
+ {
+ "epoch": 30.069868995633186,
+ "grad_norm": 0.018116053193807602,
+ "learning_rate": 0.0006,
+ "loss": 4.228706359863281,
+ "step": 2165
+ },
+ {
+ "epoch": 30.083842794759825,
+ "grad_norm": 0.01791389286518097,
+ "learning_rate": 0.0006,
+ "loss": 4.1597747802734375,
+ "step": 2166
+ },
+ {
+ "epoch": 30.097816593886463,
+ "grad_norm": 0.017793502658605576,
+ "learning_rate": 0.0006,
+ "loss": 4.0807976722717285,
+ "step": 2167
+ },
+ {
+ "epoch": 30.111790393013102,
+ "grad_norm": 0.01827503927052021,
+ "learning_rate": 0.0006,
+ "loss": 4.170646667480469,
+ "step": 2168
+ },
+ {
+ "epoch": 30.125764192139737,
+ "grad_norm": 0.017228564247488976,
+ "learning_rate": 0.0006,
+ "loss": 4.146934509277344,
+ "step": 2169
+ },
+ {
+ "epoch": 30.139737991266376,
+ "grad_norm": 0.01757437363266945,
+ "learning_rate": 0.0006,
+ "loss": 4.1859130859375,
+ "step": 2170
+ },
+ {
+ "epoch": 30.153711790393015,
+ "grad_norm": 0.01701093092560768,
+ "learning_rate": 0.0006,
+ "loss": 4.126981258392334,
+ "step": 2171
+ },
+ {
+ "epoch": 30.16768558951965,
+ "grad_norm": 0.015220489352941513,
+ "learning_rate": 0.0006,
+ "loss": 4.164018630981445,
+ "step": 2172
+ },
+ {
+ "epoch": 30.18165938864629,
+ "grad_norm": 0.016254238784313202,
+ "learning_rate": 0.0006,
+ "loss": 4.258275508880615,
+ "step": 2173
+ },
+ {
+ "epoch": 30.195633187772927,
+ "grad_norm": 0.016609614714980125,
+ "learning_rate": 0.0006,
+ "loss": 4.222233772277832,
+ "step": 2174
+ },
+ {
+ "epoch": 30.209606986899562,
+ "grad_norm": 0.016531318426132202,
+ "learning_rate": 0.0006,
+ "loss": 4.144489288330078,
+ "step": 2175
+ },
+ {
+ "epoch": 30.2235807860262,
+ "grad_norm": 0.01567668654024601,
+ "learning_rate": 0.0006,
+ "loss": 4.181916236877441,
+ "step": 2176
+ },
+ {
+ "epoch": 30.23755458515284,
+ "grad_norm": 0.017195681110024452,
+ "learning_rate": 0.0006,
+ "loss": 4.189693927764893,
+ "step": 2177
+ },
+ {
+ "epoch": 30.251528384279474,
+ "grad_norm": 0.016882948577404022,
+ "learning_rate": 0.0006,
+ "loss": 4.21604061126709,
+ "step": 2178
+ },
+ {
+ "epoch": 30.265502183406113,
+ "grad_norm": 0.016261622309684753,
+ "learning_rate": 0.0006,
+ "loss": 4.227639198303223,
+ "step": 2179
+ },
+ {
+ "epoch": 30.27947598253275,
+ "grad_norm": 0.016411812976002693,
+ "learning_rate": 0.0006,
+ "loss": 4.130362510681152,
+ "step": 2180
+ },
+ {
+ "epoch": 30.29344978165939,
+ "grad_norm": 0.01732019893825054,
+ "learning_rate": 0.0006,
+ "loss": 4.230443000793457,
+ "step": 2181
+ },
+ {
+ "epoch": 30.307423580786025,
+ "grad_norm": 0.01825905591249466,
+ "learning_rate": 0.0006,
+ "loss": 4.175817489624023,
+ "step": 2182
+ },
+ {
+ "epoch": 30.321397379912664,
+ "grad_norm": 0.01887713558971882,
+ "learning_rate": 0.0006,
+ "loss": 4.169020652770996,
+ "step": 2183
+ },
+ {
+ "epoch": 30.335371179039303,
+ "grad_norm": 0.019391661509871483,
+ "learning_rate": 0.0006,
+ "loss": 4.176318645477295,
+ "step": 2184
+ },
+ {
+ "epoch": 30.349344978165938,
+ "grad_norm": 0.020840002223849297,
+ "learning_rate": 0.0006,
+ "loss": 4.174228668212891,
+ "step": 2185
+ },
+ {
+ "epoch": 30.363318777292577,
+ "grad_norm": 0.020193178206682205,
+ "learning_rate": 0.0006,
+ "loss": 4.127919673919678,
+ "step": 2186
+ },
+ {
+ "epoch": 30.377292576419215,
+ "grad_norm": 0.018449561670422554,
+ "learning_rate": 0.0006,
+ "loss": 4.229487895965576,
+ "step": 2187
+ },
+ {
+ "epoch": 30.39126637554585,
+ "grad_norm": 0.01711983233690262,
+ "learning_rate": 0.0006,
+ "loss": 4.201651573181152,
+ "step": 2188
+ },
+ {
+ "epoch": 30.40524017467249,
+ "grad_norm": 0.017293395474553108,
+ "learning_rate": 0.0006,
+ "loss": 4.168704032897949,
+ "step": 2189
+ },
+ {
+ "epoch": 30.419213973799128,
+ "grad_norm": 0.018526358529925346,
+ "learning_rate": 0.0006,
+ "loss": 4.171864032745361,
+ "step": 2190
+ },
+ {
+ "epoch": 30.433187772925763,
+ "grad_norm": 0.019287291914224625,
+ "learning_rate": 0.0006,
+ "loss": 4.039534568786621,
+ "step": 2191
+ },
+ {
+ "epoch": 30.4471615720524,
+ "grad_norm": 0.016902444884181023,
+ "learning_rate": 0.0006,
+ "loss": 4.282787799835205,
+ "step": 2192
+ },
+ {
+ "epoch": 30.46113537117904,
+ "grad_norm": 0.01656750962138176,
+ "learning_rate": 0.0006,
+ "loss": 4.213098526000977,
+ "step": 2193
+ },
+ {
+ "epoch": 30.475109170305675,
+ "grad_norm": 0.016805065795779228,
+ "learning_rate": 0.0006,
+ "loss": 4.1583476066589355,
+ "step": 2194
+ },
+ {
+ "epoch": 30.489082969432314,
+ "grad_norm": 0.017166638746857643,
+ "learning_rate": 0.0006,
+ "loss": 4.188634395599365,
+ "step": 2195
+ },
+ {
+ "epoch": 30.503056768558952,
+ "grad_norm": 0.01592106744647026,
+ "learning_rate": 0.0006,
+ "loss": 4.234043121337891,
+ "step": 2196
+ },
+ {
+ "epoch": 30.51703056768559,
+ "grad_norm": 0.015270989388227463,
+ "learning_rate": 0.0006,
+ "loss": 4.301420211791992,
+ "step": 2197
+ },
+ {
+ "epoch": 30.531004366812226,
+ "grad_norm": 0.015194724313914776,
+ "learning_rate": 0.0006,
+ "loss": 4.18703031539917,
+ "step": 2198
+ },
+ {
+ "epoch": 30.544978165938865,
+ "grad_norm": 0.015587746165692806,
+ "learning_rate": 0.0006,
+ "loss": 4.138928413391113,
+ "step": 2199
+ },
+ {
+ "epoch": 30.558951965065503,
+ "grad_norm": 0.016029570251703262,
+ "learning_rate": 0.0006,
+ "loss": 4.290863990783691,
+ "step": 2200
+ },
+ {
+ "epoch": 30.57292576419214,
+ "grad_norm": 0.015651309862732887,
+ "learning_rate": 0.0006,
+ "loss": 4.272062301635742,
+ "step": 2201
+ },
+ {
+ "epoch": 30.586899563318777,
+ "grad_norm": 0.015992306172847748,
+ "learning_rate": 0.0006,
+ "loss": 4.2487406730651855,
+ "step": 2202
+ },
+ {
+ "epoch": 30.600873362445416,
+ "grad_norm": 0.014969157986342907,
+ "learning_rate": 0.0006,
+ "loss": 4.115347862243652,
+ "step": 2203
+ },
+ {
+ "epoch": 30.61484716157205,
+ "grad_norm": 0.014639027416706085,
+ "learning_rate": 0.0006,
+ "loss": 4.2207489013671875,
+ "step": 2204
+ },
+ {
+ "epoch": 30.62882096069869,
+ "grad_norm": 0.015059413388371468,
+ "learning_rate": 0.0006,
+ "loss": 4.274044990539551,
+ "step": 2205
+ },
+ {
+ "epoch": 30.64279475982533,
+ "grad_norm": 0.0166641678661108,
+ "learning_rate": 0.0006,
+ "loss": 4.191373348236084,
+ "step": 2206
+ },
+ {
+ "epoch": 30.656768558951963,
+ "grad_norm": 0.017806345596909523,
+ "learning_rate": 0.0006,
+ "loss": 4.139800548553467,
+ "step": 2207
+ },
+ {
+ "epoch": 30.670742358078602,
+ "grad_norm": 0.018712421879172325,
+ "learning_rate": 0.0006,
+ "loss": 4.035965442657471,
+ "step": 2208
+ },
+ {
+ "epoch": 30.68471615720524,
+ "grad_norm": 0.018457243219017982,
+ "learning_rate": 0.0006,
+ "loss": 4.2579240798950195,
+ "step": 2209
+ },
+ {
+ "epoch": 30.69868995633188,
+ "grad_norm": 0.01920999586582184,
+ "learning_rate": 0.0006,
+ "loss": 4.131112098693848,
+ "step": 2210
+ },
+ {
+ "epoch": 30.712663755458514,
+ "grad_norm": 0.01729944534599781,
+ "learning_rate": 0.0006,
+ "loss": 4.254059314727783,
+ "step": 2211
+ },
+ {
+ "epoch": 30.726637554585153,
+ "grad_norm": 0.016914231702685356,
+ "learning_rate": 0.0006,
+ "loss": 4.233428955078125,
+ "step": 2212
+ },
+ {
+ "epoch": 30.74061135371179,
+ "grad_norm": 0.01913098618388176,
+ "learning_rate": 0.0006,
+ "loss": 4.23525333404541,
+ "step": 2213
+ },
+ {
+ "epoch": 30.754585152838427,
+ "grad_norm": 0.019150281324982643,
+ "learning_rate": 0.0006,
+ "loss": 4.213695049285889,
+ "step": 2214
+ },
+ {
+ "epoch": 30.768558951965066,
+ "grad_norm": 0.018947452306747437,
+ "learning_rate": 0.0006,
+ "loss": 4.151697158813477,
+ "step": 2215
+ },
+ {
+ "epoch": 30.782532751091704,
+ "grad_norm": 0.016058508306741714,
+ "learning_rate": 0.0006,
+ "loss": 4.220280170440674,
+ "step": 2216
+ },
+ {
+ "epoch": 30.79650655021834,
+ "grad_norm": 0.017488619312644005,
+ "learning_rate": 0.0006,
+ "loss": 4.154428482055664,
+ "step": 2217
+ },
+ {
+ "epoch": 30.810480349344978,
+ "grad_norm": 0.017355090007185936,
+ "learning_rate": 0.0006,
+ "loss": 4.256086349487305,
+ "step": 2218
+ },
+ {
+ "epoch": 30.824454148471617,
+ "grad_norm": 0.016636032611131668,
+ "learning_rate": 0.0006,
+ "loss": 4.227564334869385,
+ "step": 2219
+ },
+ {
+ "epoch": 30.83842794759825,
+ "grad_norm": 0.016273891553282738,
+ "learning_rate": 0.0006,
+ "loss": 4.200466156005859,
+ "step": 2220
+ },
+ {
+ "epoch": 30.85240174672489,
+ "grad_norm": 0.016939911991357803,
+ "learning_rate": 0.0006,
+ "loss": 4.1571221351623535,
+ "step": 2221
+ },
+ {
+ "epoch": 30.86637554585153,
+ "grad_norm": 0.018332522362470627,
+ "learning_rate": 0.0006,
+ "loss": 4.228398323059082,
+ "step": 2222
+ },
+ {
+ "epoch": 30.880349344978168,
+ "grad_norm": 0.019508114084601402,
+ "learning_rate": 0.0006,
+ "loss": 4.222960472106934,
+ "step": 2223
+ },
+ {
+ "epoch": 30.894323144104803,
+ "grad_norm": 0.018440278246998787,
+ "learning_rate": 0.0006,
+ "loss": 4.261051177978516,
+ "step": 2224
+ },
+ {
+ "epoch": 30.90829694323144,
+ "grad_norm": 0.018113229423761368,
+ "learning_rate": 0.0006,
+ "loss": 4.109235763549805,
+ "step": 2225
+ },
+ {
+ "epoch": 30.92227074235808,
+ "grad_norm": 0.015392083674669266,
+ "learning_rate": 0.0006,
+ "loss": 4.21950101852417,
+ "step": 2226
+ },
+ {
+ "epoch": 30.936244541484715,
+ "grad_norm": 0.01633290946483612,
+ "learning_rate": 0.0006,
+ "loss": 4.176053047180176,
+ "step": 2227
+ },
+ {
+ "epoch": 30.950218340611354,
+ "grad_norm": 0.016236383467912674,
+ "learning_rate": 0.0006,
+ "loss": 4.314309120178223,
+ "step": 2228
+ },
+ {
+ "epoch": 30.964192139737992,
+ "grad_norm": 0.015284373424947262,
+ "learning_rate": 0.0006,
+ "loss": 4.303133487701416,
+ "step": 2229
+ },
+ {
+ "epoch": 30.978165938864628,
+ "grad_norm": 0.015982897952198982,
+ "learning_rate": 0.0006,
+ "loss": 4.25739860534668,
+ "step": 2230
+ },
+ {
+ "epoch": 30.992139737991266,
+ "grad_norm": 0.016682198271155357,
+ "learning_rate": 0.0006,
+ "loss": 4.292210578918457,
+ "step": 2231
+ },
+ {
+ "epoch": 31.0,
+ "grad_norm": 0.017521467059850693,
+ "learning_rate": 0.0006,
+ "loss": 4.170637130737305,
+ "step": 2232
+ },
+ {
+ "epoch": 31.0,
+ "eval_loss": 4.642786979675293,
+ "eval_runtime": 57.029,
+ "eval_samples_per_second": 42.82,
+ "eval_steps_per_second": 1.35,
+ "step": 2232
+ },
+ {
+ "epoch": 31.01397379912664,
+ "grad_norm": 0.016607852652668953,
+ "learning_rate": 0.0006,
+ "loss": 4.106224536895752,
+ "step": 2233
+ },
+ {
+ "epoch": 31.027947598253274,
+ "grad_norm": 0.018448904156684875,
+ "learning_rate": 0.0006,
+ "loss": 4.224459648132324,
+ "step": 2234
+ },
+ {
+ "epoch": 31.041921397379912,
+ "grad_norm": 0.020481228828430176,
+ "learning_rate": 0.0006,
+ "loss": 4.305548667907715,
+ "step": 2235
+ },
+ {
+ "epoch": 31.05589519650655,
+ "grad_norm": 0.01999707892537117,
+ "learning_rate": 0.0006,
+ "loss": 4.253663539886475,
+ "step": 2236
+ },
+ {
+ "epoch": 31.069868995633186,
+ "grad_norm": 0.020428303629159927,
+ "learning_rate": 0.0006,
+ "loss": 4.166954040527344,
+ "step": 2237
+ },
+ {
+ "epoch": 31.083842794759825,
+ "grad_norm": 0.022608213126659393,
+ "learning_rate": 0.0006,
+ "loss": 4.172819137573242,
+ "step": 2238
+ },
+ {
+ "epoch": 31.097816593886463,
+ "grad_norm": 0.024669520556926727,
+ "learning_rate": 0.0006,
+ "loss": 4.106586456298828,
+ "step": 2239
+ },
+ {
+ "epoch": 31.111790393013102,
+ "grad_norm": 0.02247351035475731,
+ "learning_rate": 0.0006,
+ "loss": 4.160271644592285,
+ "step": 2240
+ },
+ {
+ "epoch": 31.125764192139737,
+ "grad_norm": 0.0207810141146183,
+ "learning_rate": 0.0006,
+ "loss": 4.272491931915283,
+ "step": 2241
+ },
+ {
+ "epoch": 31.139737991266376,
+ "grad_norm": 0.021576499566435814,
+ "learning_rate": 0.0006,
+ "loss": 4.171474456787109,
+ "step": 2242
+ },
+ {
+ "epoch": 31.153711790393015,
+ "grad_norm": 0.02337106689810753,
+ "learning_rate": 0.0006,
+ "loss": 4.094712734222412,
+ "step": 2243
+ },
+ {
+ "epoch": 31.16768558951965,
+ "grad_norm": 0.02219875156879425,
+ "learning_rate": 0.0006,
+ "loss": 4.1800994873046875,
+ "step": 2244
+ },
+ {
+ "epoch": 31.18165938864629,
+ "grad_norm": 0.022680504247546196,
+ "learning_rate": 0.0006,
+ "loss": 4.2054362297058105,
+ "step": 2245
+ },
+ {
+ "epoch": 31.195633187772927,
+ "grad_norm": 0.0208734143525362,
+ "learning_rate": 0.0006,
+ "loss": 4.06068229675293,
+ "step": 2246
+ },
+ {
+ "epoch": 31.209606986899562,
+ "grad_norm": 0.01811066083610058,
+ "learning_rate": 0.0006,
+ "loss": 4.2465128898620605,
+ "step": 2247
+ },
+ {
+ "epoch": 31.2235807860262,
+ "grad_norm": 0.018571637570858,
+ "learning_rate": 0.0006,
+ "loss": 4.191955089569092,
+ "step": 2248
+ },
+ {
+ "epoch": 31.23755458515284,
+ "grad_norm": 0.018598807975649834,
+ "learning_rate": 0.0006,
+ "loss": 4.200872421264648,
+ "step": 2249
+ },
+ {
+ "epoch": 31.251528384279474,
+ "grad_norm": 0.017577625811100006,
+ "learning_rate": 0.0006,
+ "loss": 4.148700714111328,
+ "step": 2250
+ },
+ {
+ "epoch": 31.265502183406113,
+ "grad_norm": 0.01817980408668518,
+ "learning_rate": 0.0006,
+ "loss": 4.256978988647461,
+ "step": 2251
+ },
+ {
+ "epoch": 31.27947598253275,
+ "grad_norm": 0.017983486875891685,
+ "learning_rate": 0.0006,
+ "loss": 4.175005912780762,
+ "step": 2252
+ },
+ {
+ "epoch": 31.29344978165939,
+ "grad_norm": 0.016751961782574654,
+ "learning_rate": 0.0006,
+ "loss": 4.237518310546875,
+ "step": 2253
+ },
+ {
+ "epoch": 31.307423580786025,
+ "grad_norm": 0.016407202929258347,
+ "learning_rate": 0.0006,
+ "loss": 4.197269439697266,
+ "step": 2254
+ },
+ {
+ "epoch": 31.321397379912664,
+ "grad_norm": 0.01578645594418049,
+ "learning_rate": 0.0006,
+ "loss": 4.179751396179199,
+ "step": 2255
+ },
+ {
+ "epoch": 31.335371179039303,
+ "grad_norm": 0.016001226380467415,
+ "learning_rate": 0.0006,
+ "loss": 4.08821964263916,
+ "step": 2256
+ },
+ {
+ "epoch": 31.349344978165938,
+ "grad_norm": 0.015122811309993267,
+ "learning_rate": 0.0006,
+ "loss": 4.047828674316406,
+ "step": 2257
+ },
+ {
+ "epoch": 31.363318777292577,
+ "grad_norm": 0.015399965457618237,
+ "learning_rate": 0.0006,
+ "loss": 4.1506524085998535,
+ "step": 2258
+ },
+ {
+ "epoch": 31.377292576419215,
+ "grad_norm": 0.01534605398774147,
+ "learning_rate": 0.0006,
+ "loss": 4.27952766418457,
+ "step": 2259
+ },
+ {
+ "epoch": 31.39126637554585,
+ "grad_norm": 0.015461381524801254,
+ "learning_rate": 0.0006,
+ "loss": 4.27102518081665,
+ "step": 2260
+ },
+ {
+ "epoch": 31.40524017467249,
+ "grad_norm": 0.016769153997302055,
+ "learning_rate": 0.0006,
+ "loss": 4.191592216491699,
+ "step": 2261
+ },
+ {
+ "epoch": 31.419213973799128,
+ "grad_norm": 0.016557959839701653,
+ "learning_rate": 0.0006,
+ "loss": 4.159705638885498,
+ "step": 2262
+ },
+ {
+ "epoch": 31.433187772925763,
+ "grad_norm": 0.01659569889307022,
+ "learning_rate": 0.0006,
+ "loss": 4.2136664390563965,
+ "step": 2263
+ },
+ {
+ "epoch": 31.4471615720524,
+ "grad_norm": 0.017101937904953957,
+ "learning_rate": 0.0006,
+ "loss": 4.2716498374938965,
+ "step": 2264
+ },
+ {
+ "epoch": 31.46113537117904,
+ "grad_norm": 0.01722985878586769,
+ "learning_rate": 0.0006,
+ "loss": 4.300534248352051,
+ "step": 2265
+ },
+ {
+ "epoch": 31.475109170305675,
+ "grad_norm": 0.016800185665488243,
+ "learning_rate": 0.0006,
+ "loss": 4.113240718841553,
+ "step": 2266
+ },
+ {
+ "epoch": 31.489082969432314,
+ "grad_norm": 0.01603875495493412,
+ "learning_rate": 0.0006,
+ "loss": 4.157105445861816,
+ "step": 2267
+ },
+ {
+ "epoch": 31.503056768558952,
+ "grad_norm": 0.015903932973742485,
+ "learning_rate": 0.0006,
+ "loss": 4.233669757843018,
+ "step": 2268
+ },
+ {
+ "epoch": 31.51703056768559,
+ "grad_norm": 0.0160983856767416,
+ "learning_rate": 0.0006,
+ "loss": 4.147704124450684,
+ "step": 2269
+ },
+ {
+ "epoch": 31.531004366812226,
+ "grad_norm": 0.01604730263352394,
+ "learning_rate": 0.0006,
+ "loss": 4.1465911865234375,
+ "step": 2270
+ },
+ {
+ "epoch": 31.544978165938865,
+ "grad_norm": 0.016205785796046257,
+ "learning_rate": 0.0006,
+ "loss": 4.106704235076904,
+ "step": 2271
+ },
+ {
+ "epoch": 31.558951965065503,
+ "grad_norm": 0.017008036375045776,
+ "learning_rate": 0.0006,
+ "loss": 4.191000938415527,
+ "step": 2272
+ },
+ {
+ "epoch": 31.57292576419214,
+ "grad_norm": 0.01699635200202465,
+ "learning_rate": 0.0006,
+ "loss": 4.187042236328125,
+ "step": 2273
+ },
+ {
+ "epoch": 31.586899563318777,
+ "grad_norm": 0.01692967116832733,
+ "learning_rate": 0.0006,
+ "loss": 4.208805561065674,
+ "step": 2274
+ },
+ {
+ "epoch": 31.600873362445416,
+ "grad_norm": 0.017608756199479103,
+ "learning_rate": 0.0006,
+ "loss": 4.200512886047363,
+ "step": 2275
+ },
+ {
+ "epoch": 31.61484716157205,
+ "grad_norm": 0.016590990126132965,
+ "learning_rate": 0.0006,
+ "loss": 4.194705009460449,
+ "step": 2276
+ },
+ {
+ "epoch": 31.62882096069869,
+ "grad_norm": 0.0170235987752676,
+ "learning_rate": 0.0006,
+ "loss": 4.210326194763184,
+ "step": 2277
+ },
+ {
+ "epoch": 31.64279475982533,
+ "grad_norm": 0.01604801043868065,
+ "learning_rate": 0.0006,
+ "loss": 4.196024417877197,
+ "step": 2278
+ },
+ {
+ "epoch": 31.656768558951963,
+ "grad_norm": 0.015694202855229378,
+ "learning_rate": 0.0006,
+ "loss": 4.178236484527588,
+ "step": 2279
+ },
+ {
+ "epoch": 31.670742358078602,
+ "grad_norm": 0.015473959036171436,
+ "learning_rate": 0.0006,
+ "loss": 4.259339809417725,
+ "step": 2280
+ },
+ {
+ "epoch": 31.68471615720524,
+ "grad_norm": 0.014689362607896328,
+ "learning_rate": 0.0006,
+ "loss": 3.9702324867248535,
+ "step": 2281
+ },
+ {
+ "epoch": 31.69868995633188,
+ "grad_norm": 0.015245123766362667,
+ "learning_rate": 0.0006,
+ "loss": 4.197025775909424,
+ "step": 2282
+ },
+ {
+ "epoch": 31.712663755458514,
+ "grad_norm": 0.016148347407579422,
+ "learning_rate": 0.0006,
+ "loss": 4.231856346130371,
+ "step": 2283
+ },
+ {
+ "epoch": 31.726637554585153,
+ "grad_norm": 0.015169818885624409,
+ "learning_rate": 0.0006,
+ "loss": 4.230169296264648,
+ "step": 2284
+ },
+ {
+ "epoch": 31.74061135371179,
+ "grad_norm": 0.014843013137578964,
+ "learning_rate": 0.0006,
+ "loss": 4.195633411407471,
+ "step": 2285
+ },
+ {
+ "epoch": 31.754585152838427,
+ "grad_norm": 0.015229340642690659,
+ "learning_rate": 0.0006,
+ "loss": 4.231280326843262,
+ "step": 2286
+ },
+ {
+ "epoch": 31.768558951965066,
+ "grad_norm": 0.01551708485931158,
+ "learning_rate": 0.0006,
+ "loss": 4.17064094543457,
+ "step": 2287
+ },
+ {
+ "epoch": 31.782532751091704,
+ "grad_norm": 0.015863575041294098,
+ "learning_rate": 0.0006,
+ "loss": 4.1579484939575195,
+ "step": 2288
+ },
+ {
+ "epoch": 31.79650655021834,
+ "grad_norm": 0.015118198469281197,
+ "learning_rate": 0.0006,
+ "loss": 4.1626787185668945,
+ "step": 2289
+ },
+ {
+ "epoch": 31.810480349344978,
+ "grad_norm": 0.015220776200294495,
+ "learning_rate": 0.0006,
+ "loss": 4.135746955871582,
+ "step": 2290
+ },
+ {
+ "epoch": 31.824454148471617,
+ "grad_norm": 0.014529784210026264,
+ "learning_rate": 0.0006,
+ "loss": 4.164931297302246,
+ "step": 2291
+ },
+ {
+ "epoch": 31.83842794759825,
+ "grad_norm": 0.015533296391367912,
+ "learning_rate": 0.0006,
+ "loss": 4.112464904785156,
+ "step": 2292
+ },
+ {
+ "epoch": 31.85240174672489,
+ "grad_norm": 0.016562335193157196,
+ "learning_rate": 0.0006,
+ "loss": 4.164898872375488,
+ "step": 2293
+ },
+ {
+ "epoch": 31.86637554585153,
+ "grad_norm": 0.016258113086223602,
+ "learning_rate": 0.0006,
+ "loss": 4.1979265213012695,
+ "step": 2294
+ },
+ {
+ "epoch": 31.880349344978168,
+ "grad_norm": 0.01558110024780035,
+ "learning_rate": 0.0006,
+ "loss": 4.174675941467285,
+ "step": 2295
+ },
+ {
+ "epoch": 31.894323144104803,
+ "grad_norm": 0.016412515193223953,
+ "learning_rate": 0.0006,
+ "loss": 4.217909336090088,
+ "step": 2296
+ },
+ {
+ "epoch": 31.90829694323144,
+ "grad_norm": 0.015118442475795746,
+ "learning_rate": 0.0006,
+ "loss": 4.137143611907959,
+ "step": 2297
+ },
+ {
+ "epoch": 31.92227074235808,
+ "grad_norm": 0.01506161317229271,
+ "learning_rate": 0.0006,
+ "loss": 4.196122169494629,
+ "step": 2298
+ },
+ {
+ "epoch": 31.936244541484715,
+ "grad_norm": 0.017165271565318108,
+ "learning_rate": 0.0006,
+ "loss": 4.14834451675415,
+ "step": 2299
+ },
+ {
+ "epoch": 31.950218340611354,
+ "grad_norm": 0.01696440577507019,
+ "learning_rate": 0.0006,
+ "loss": 4.218868732452393,
+ "step": 2300
+ },
+ {
+ "epoch": 31.964192139737992,
+ "grad_norm": 0.017349394038319588,
+ "learning_rate": 0.0006,
+ "loss": 4.2754011154174805,
+ "step": 2301
+ },
+ {
+ "epoch": 31.978165938864628,
+ "grad_norm": 0.018387921154499054,
+ "learning_rate": 0.0006,
+ "loss": 4.282623291015625,
+ "step": 2302
+ },
+ {
+ "epoch": 31.992139737991266,
+ "grad_norm": 0.018837926909327507,
+ "learning_rate": 0.0006,
+ "loss": 4.185425281524658,
+ "step": 2303
+ },
+ {
+ "epoch": 32.0,
+ "grad_norm": 0.018998464569449425,
+ "learning_rate": 0.0006,
+ "loss": 4.151291847229004,
+ "step": 2304
+ },
+ {
+ "epoch": 32.0,
+ "eval_loss": 4.522614002227783,
+ "eval_runtime": 57.0106,
+ "eval_samples_per_second": 42.834,
+ "eval_steps_per_second": 1.351,
+ "step": 2304
+ },
+ {
+ "epoch": 32.01397379912664,
+ "grad_norm": 0.019036833196878433,
+ "learning_rate": 0.0006,
+ "loss": 4.183202266693115,
+ "step": 2305
+ },
+ {
+ "epoch": 32.02794759825328,
+ "grad_norm": 0.02207397110760212,
+ "learning_rate": 0.0006,
+ "loss": 4.17875862121582,
+ "step": 2306
+ },
+ {
+ "epoch": 32.041921397379916,
+ "grad_norm": 0.022008279338479042,
+ "learning_rate": 0.0006,
+ "loss": 4.08042049407959,
+ "step": 2307
+ },
+ {
+ "epoch": 32.05589519650655,
+ "grad_norm": 0.022407017648220062,
+ "learning_rate": 0.0006,
+ "loss": 4.112189769744873,
+ "step": 2308
+ },
+ {
+ "epoch": 32.069868995633186,
+ "grad_norm": 0.021599747240543365,
+ "learning_rate": 0.0006,
+ "loss": 4.126455307006836,
+ "step": 2309
+ },
+ {
+ "epoch": 32.083842794759825,
+ "grad_norm": 0.020340966060757637,
+ "learning_rate": 0.0006,
+ "loss": 4.080538272857666,
+ "step": 2310
+ },
+ {
+ "epoch": 32.09781659388646,
+ "grad_norm": 0.020956892520189285,
+ "learning_rate": 0.0006,
+ "loss": 4.171943664550781,
+ "step": 2311
+ },
+ {
+ "epoch": 32.1117903930131,
+ "grad_norm": 0.019966667518019676,
+ "learning_rate": 0.0006,
+ "loss": 4.144007682800293,
+ "step": 2312
+ },
+ {
+ "epoch": 32.12576419213974,
+ "grad_norm": 0.018256952986121178,
+ "learning_rate": 0.0006,
+ "loss": 4.208107948303223,
+ "step": 2313
+ },
+ {
+ "epoch": 32.13973799126637,
+ "grad_norm": 0.017289170995354652,
+ "learning_rate": 0.0006,
+ "loss": 4.182024955749512,
+ "step": 2314
+ },
+ {
+ "epoch": 32.15371179039301,
+ "grad_norm": 0.017656700685620308,
+ "learning_rate": 0.0006,
+ "loss": 4.068373680114746,
+ "step": 2315
+ },
+ {
+ "epoch": 32.16768558951965,
+ "grad_norm": 0.017903361469507217,
+ "learning_rate": 0.0006,
+ "loss": 4.2345428466796875,
+ "step": 2316
+ },
+ {
+ "epoch": 32.18165938864629,
+ "grad_norm": 0.01785109005868435,
+ "learning_rate": 0.0006,
+ "loss": 4.219776153564453,
+ "step": 2317
+ },
+ {
+ "epoch": 32.19563318777293,
+ "grad_norm": 0.01887257769703865,
+ "learning_rate": 0.0006,
+ "loss": 4.139206409454346,
+ "step": 2318
+ },
+ {
+ "epoch": 32.209606986899566,
+ "grad_norm": 0.01924685761332512,
+ "learning_rate": 0.0006,
+ "loss": 4.121052265167236,
+ "step": 2319
+ },
+ {
+ "epoch": 32.223580786026204,
+ "grad_norm": 0.020574072375893593,
+ "learning_rate": 0.0006,
+ "loss": 4.13262414932251,
+ "step": 2320
+ },
+ {
+ "epoch": 32.237554585152836,
+ "grad_norm": 0.023849064484238625,
+ "learning_rate": 0.0006,
+ "loss": 4.239119529724121,
+ "step": 2321
+ },
+ {
+ "epoch": 32.251528384279474,
+ "grad_norm": 0.026239361613988876,
+ "learning_rate": 0.0006,
+ "loss": 4.248993873596191,
+ "step": 2322
+ },
+ {
+ "epoch": 32.26550218340611,
+ "grad_norm": 0.023980243131518364,
+ "learning_rate": 0.0006,
+ "loss": 4.159852504730225,
+ "step": 2323
+ },
+ {
+ "epoch": 32.27947598253275,
+ "grad_norm": 0.020442405715584755,
+ "learning_rate": 0.0006,
+ "loss": 4.133546829223633,
+ "step": 2324
+ },
+ {
+ "epoch": 32.29344978165939,
+ "grad_norm": 0.020661186426877975,
+ "learning_rate": 0.0006,
+ "loss": 4.078243732452393,
+ "step": 2325
+ },
+ {
+ "epoch": 32.30742358078603,
+ "grad_norm": 0.02046387456357479,
+ "learning_rate": 0.0006,
+ "loss": 4.172012805938721,
+ "step": 2326
+ },
+ {
+ "epoch": 32.32139737991266,
+ "grad_norm": 0.02089597098529339,
+ "learning_rate": 0.0006,
+ "loss": 4.0604705810546875,
+ "step": 2327
+ },
+ {
+ "epoch": 32.3353711790393,
+ "grad_norm": 0.01875591278076172,
+ "learning_rate": 0.0006,
+ "loss": 4.125911235809326,
+ "step": 2328
+ },
+ {
+ "epoch": 32.34934497816594,
+ "grad_norm": 0.018606796860694885,
+ "learning_rate": 0.0006,
+ "loss": 4.097362518310547,
+ "step": 2329
+ },
+ {
+ "epoch": 32.36331877729258,
+ "grad_norm": 0.01766626164317131,
+ "learning_rate": 0.0006,
+ "loss": 4.118616580963135,
+ "step": 2330
+ },
+ {
+ "epoch": 32.377292576419215,
+ "grad_norm": 0.017692960798740387,
+ "learning_rate": 0.0006,
+ "loss": 4.1457672119140625,
+ "step": 2331
+ },
+ {
+ "epoch": 32.391266375545854,
+ "grad_norm": 0.017619704827666283,
+ "learning_rate": 0.0006,
+ "loss": 4.175961494445801,
+ "step": 2332
+ },
+ {
+ "epoch": 32.40524017467249,
+ "grad_norm": 0.01784413494169712,
+ "learning_rate": 0.0006,
+ "loss": 4.051291465759277,
+ "step": 2333
+ },
+ {
+ "epoch": 32.419213973799124,
+ "grad_norm": 0.018508171662688255,
+ "learning_rate": 0.0006,
+ "loss": 4.0638556480407715,
+ "step": 2334
+ },
+ {
+ "epoch": 32.43318777292576,
+ "grad_norm": 0.01731436885893345,
+ "learning_rate": 0.0006,
+ "loss": 4.1805419921875,
+ "step": 2335
+ },
+ {
+ "epoch": 32.4471615720524,
+ "grad_norm": 0.015886498615145683,
+ "learning_rate": 0.0006,
+ "loss": 4.151026725769043,
+ "step": 2336
+ },
+ {
+ "epoch": 32.46113537117904,
+ "grad_norm": 0.016583820804953575,
+ "learning_rate": 0.0006,
+ "loss": 4.203604698181152,
+ "step": 2337
+ },
+ {
+ "epoch": 32.47510917030568,
+ "grad_norm": 0.0176097322255373,
+ "learning_rate": 0.0006,
+ "loss": 4.135159015655518,
+ "step": 2338
+ },
+ {
+ "epoch": 32.48908296943232,
+ "grad_norm": 0.016657903790473938,
+ "learning_rate": 0.0006,
+ "loss": 4.126681804656982,
+ "step": 2339
+ },
+ {
+ "epoch": 32.50305676855895,
+ "grad_norm": 0.016866130754351616,
+ "learning_rate": 0.0006,
+ "loss": 4.162341117858887,
+ "step": 2340
+ },
+ {
+ "epoch": 32.51703056768559,
+ "grad_norm": 0.016286680474877357,
+ "learning_rate": 0.0006,
+ "loss": 4.261801242828369,
+ "step": 2341
+ },
+ {
+ "epoch": 32.531004366812226,
+ "grad_norm": 0.01481365505605936,
+ "learning_rate": 0.0006,
+ "loss": 4.110016822814941,
+ "step": 2342
+ },
+ {
+ "epoch": 32.544978165938865,
+ "grad_norm": 0.016605013981461525,
+ "learning_rate": 0.0006,
+ "loss": 4.230024814605713,
+ "step": 2343
+ },
+ {
+ "epoch": 32.5589519650655,
+ "grad_norm": 0.01645175740122795,
+ "learning_rate": 0.0006,
+ "loss": 4.174612045288086,
+ "step": 2344
+ },
+ {
+ "epoch": 32.57292576419214,
+ "grad_norm": 0.014725026674568653,
+ "learning_rate": 0.0006,
+ "loss": 4.066371440887451,
+ "step": 2345
+ },
+ {
+ "epoch": 32.58689956331878,
+ "grad_norm": 0.015003837645053864,
+ "learning_rate": 0.0006,
+ "loss": 4.193343639373779,
+ "step": 2346
+ },
+ {
+ "epoch": 32.60087336244541,
+ "grad_norm": 0.015346302650868893,
+ "learning_rate": 0.0006,
+ "loss": 4.210659980773926,
+ "step": 2347
+ },
+ {
+ "epoch": 32.61484716157205,
+ "grad_norm": 0.015135962516069412,
+ "learning_rate": 0.0006,
+ "loss": 4.204618453979492,
+ "step": 2348
+ },
+ {
+ "epoch": 32.62882096069869,
+ "grad_norm": 0.01601956970989704,
+ "learning_rate": 0.0006,
+ "loss": 4.145564556121826,
+ "step": 2349
+ },
+ {
+ "epoch": 32.64279475982533,
+ "grad_norm": 0.01517851185053587,
+ "learning_rate": 0.0006,
+ "loss": 4.051673889160156,
+ "step": 2350
+ },
+ {
+ "epoch": 32.65676855895197,
+ "grad_norm": 0.014894292689859867,
+ "learning_rate": 0.0006,
+ "loss": 4.184818267822266,
+ "step": 2351
+ },
+ {
+ "epoch": 32.670742358078606,
+ "grad_norm": 0.015720834955573082,
+ "learning_rate": 0.0006,
+ "loss": 4.220036506652832,
+ "step": 2352
+ },
+ {
+ "epoch": 32.68471615720524,
+ "grad_norm": 0.01573433168232441,
+ "learning_rate": 0.0006,
+ "loss": 4.076773166656494,
+ "step": 2353
+ },
+ {
+ "epoch": 32.698689956331876,
+ "grad_norm": 0.015828052535653114,
+ "learning_rate": 0.0006,
+ "loss": 4.116329669952393,
+ "step": 2354
+ },
+ {
+ "epoch": 32.712663755458514,
+ "grad_norm": 0.015462259761989117,
+ "learning_rate": 0.0006,
+ "loss": 4.1155242919921875,
+ "step": 2355
+ },
+ {
+ "epoch": 32.72663755458515,
+ "grad_norm": 0.015763849020004272,
+ "learning_rate": 0.0006,
+ "loss": 4.169687747955322,
+ "step": 2356
+ },
+ {
+ "epoch": 32.74061135371179,
+ "grad_norm": 0.014935145154595375,
+ "learning_rate": 0.0006,
+ "loss": 4.2090911865234375,
+ "step": 2357
+ },
+ {
+ "epoch": 32.75458515283843,
+ "grad_norm": 0.015119184739887714,
+ "learning_rate": 0.0006,
+ "loss": 4.131838798522949,
+ "step": 2358
+ },
+ {
+ "epoch": 32.76855895196506,
+ "grad_norm": 0.014833525754511356,
+ "learning_rate": 0.0006,
+ "loss": 4.17618465423584,
+ "step": 2359
+ },
+ {
+ "epoch": 32.7825327510917,
+ "grad_norm": 0.01588359661400318,
+ "learning_rate": 0.0006,
+ "loss": 4.163723945617676,
+ "step": 2360
+ },
+ {
+ "epoch": 32.79650655021834,
+ "grad_norm": 0.0168188214302063,
+ "learning_rate": 0.0006,
+ "loss": 4.178189277648926,
+ "step": 2361
+ },
+ {
+ "epoch": 32.81048034934498,
+ "grad_norm": 0.01726095750927925,
+ "learning_rate": 0.0006,
+ "loss": 4.144223690032959,
+ "step": 2362
+ },
+ {
+ "epoch": 32.82445414847162,
+ "grad_norm": 0.01701410301029682,
+ "learning_rate": 0.0006,
+ "loss": 4.209896087646484,
+ "step": 2363
+ },
+ {
+ "epoch": 32.838427947598255,
+ "grad_norm": 0.01682264916598797,
+ "learning_rate": 0.0006,
+ "loss": 4.26223087310791,
+ "step": 2364
+ },
+ {
+ "epoch": 32.852401746724894,
+ "grad_norm": 0.015597456134855747,
+ "learning_rate": 0.0006,
+ "loss": 4.292664051055908,
+ "step": 2365
+ },
+ {
+ "epoch": 32.866375545851525,
+ "grad_norm": 0.015258579514920712,
+ "learning_rate": 0.0006,
+ "loss": 4.165006637573242,
+ "step": 2366
+ },
+ {
+ "epoch": 32.880349344978164,
+ "grad_norm": 0.01662200503051281,
+ "learning_rate": 0.0006,
+ "loss": 4.309786319732666,
+ "step": 2367
+ },
+ {
+ "epoch": 32.8943231441048,
+ "grad_norm": 0.015691161155700684,
+ "learning_rate": 0.0006,
+ "loss": 4.167202949523926,
+ "step": 2368
+ },
+ {
+ "epoch": 32.90829694323144,
+ "grad_norm": 0.014870637096464634,
+ "learning_rate": 0.0006,
+ "loss": 4.142415523529053,
+ "step": 2369
+ },
+ {
+ "epoch": 32.92227074235808,
+ "grad_norm": 0.01567252166569233,
+ "learning_rate": 0.0006,
+ "loss": 4.2312092781066895,
+ "step": 2370
+ },
+ {
+ "epoch": 32.93624454148472,
+ "grad_norm": 0.014948616735637188,
+ "learning_rate": 0.0006,
+ "loss": 4.1514129638671875,
+ "step": 2371
+ },
+ {
+ "epoch": 32.95021834061135,
+ "grad_norm": 0.014654500409960747,
+ "learning_rate": 0.0006,
+ "loss": 4.205010890960693,
+ "step": 2372
+ },
+ {
+ "epoch": 32.96419213973799,
+ "grad_norm": 0.015059148892760277,
+ "learning_rate": 0.0006,
+ "loss": 4.148405075073242,
+ "step": 2373
+ },
+ {
+ "epoch": 32.97816593886463,
+ "grad_norm": 0.015026576817035675,
+ "learning_rate": 0.0006,
+ "loss": 4.182535171508789,
+ "step": 2374
+ },
+ {
+ "epoch": 32.992139737991266,
+ "grad_norm": 0.016394320875406265,
+ "learning_rate": 0.0006,
+ "loss": 4.129051208496094,
+ "step": 2375
+ },
+ {
+ "epoch": 33.0,
+ "grad_norm": 0.01867043413221836,
+ "learning_rate": 0.0006,
+ "loss": 4.185556411743164,
+ "step": 2376
+ },
+ {
+ "epoch": 33.0,
+ "eval_loss": 4.640405654907227,
+ "eval_runtime": 56.5861,
+ "eval_samples_per_second": 43.155,
+ "eval_steps_per_second": 1.361,
+ "step": 2376
+ },
+ {
+ "epoch": 33.01397379912664,
+ "grad_norm": 0.018626727163791656,
+ "learning_rate": 0.0006,
+ "loss": 4.143031597137451,
+ "step": 2377
+ },
+ {
+ "epoch": 33.02794759825328,
+ "grad_norm": 0.019283363595604897,
+ "learning_rate": 0.0006,
+ "loss": 4.117845058441162,
+ "step": 2378
+ },
+ {
+ "epoch": 33.041921397379916,
+ "grad_norm": 0.018185822293162346,
+ "learning_rate": 0.0006,
+ "loss": 4.076862812042236,
+ "step": 2379
+ },
+ {
+ "epoch": 33.05589519650655,
+ "grad_norm": 0.019411196932196617,
+ "learning_rate": 0.0006,
+ "loss": 4.124156475067139,
+ "step": 2380
+ },
+ {
+ "epoch": 33.069868995633186,
+ "grad_norm": 0.019248466938734055,
+ "learning_rate": 0.0006,
+ "loss": 4.201735019683838,
+ "step": 2381
+ },
+ {
+ "epoch": 33.083842794759825,
+ "grad_norm": 0.018856720998883247,
+ "learning_rate": 0.0006,
+ "loss": 4.00296688079834,
+ "step": 2382
+ },
+ {
+ "epoch": 33.09781659388646,
+ "grad_norm": 0.018554294481873512,
+ "learning_rate": 0.0006,
+ "loss": 4.270577430725098,
+ "step": 2383
+ },
+ {
+ "epoch": 33.1117903930131,
+ "grad_norm": 0.019946224987506866,
+ "learning_rate": 0.0006,
+ "loss": 4.119589328765869,
+ "step": 2384
+ },
+ {
+ "epoch": 33.12576419213974,
+ "grad_norm": 0.021395670250058174,
+ "learning_rate": 0.0006,
+ "loss": 4.128847122192383,
+ "step": 2385
+ },
+ {
+ "epoch": 33.13973799126637,
+ "grad_norm": 0.02009674906730652,
+ "learning_rate": 0.0006,
+ "loss": 4.183228969573975,
+ "step": 2386
+ },
+ {
+ "epoch": 33.15371179039301,
+ "grad_norm": 0.01820223033428192,
+ "learning_rate": 0.0006,
+ "loss": 4.131601810455322,
+ "step": 2387
+ },
+ {
+ "epoch": 33.16768558951965,
+ "grad_norm": 0.018653156235814095,
+ "learning_rate": 0.0006,
+ "loss": 4.116430282592773,
+ "step": 2388
+ },
+ {
+ "epoch": 33.18165938864629,
+ "grad_norm": 0.016828063875436783,
+ "learning_rate": 0.0006,
+ "loss": 4.054663181304932,
+ "step": 2389
+ },
+ {
+ "epoch": 33.19563318777293,
+ "grad_norm": 0.017025409266352654,
+ "learning_rate": 0.0006,
+ "loss": 4.110044956207275,
+ "step": 2390
+ },
+ {
+ "epoch": 33.209606986899566,
+ "grad_norm": 0.016608452424407005,
+ "learning_rate": 0.0006,
+ "loss": 4.102668285369873,
+ "step": 2391
+ },
+ {
+ "epoch": 33.223580786026204,
+ "grad_norm": 0.01743577979505062,
+ "learning_rate": 0.0006,
+ "loss": 4.166791915893555,
+ "step": 2392
+ },
+ {
+ "epoch": 33.237554585152836,
+ "grad_norm": 0.017253734171390533,
+ "learning_rate": 0.0006,
+ "loss": 4.217832088470459,
+ "step": 2393
+ },
+ {
+ "epoch": 33.251528384279474,
+ "grad_norm": 0.017937595024704933,
+ "learning_rate": 0.0006,
+ "loss": 4.2173967361450195,
+ "step": 2394
+ },
+ {
+ "epoch": 33.26550218340611,
+ "grad_norm": 0.01985993981361389,
+ "learning_rate": 0.0006,
+ "loss": 4.17849588394165,
+ "step": 2395
+ },
+ {
+ "epoch": 33.27947598253275,
+ "grad_norm": 0.019243355840444565,
+ "learning_rate": 0.0006,
+ "loss": 4.086523056030273,
+ "step": 2396
+ },
+ {
+ "epoch": 33.29344978165939,
+ "grad_norm": 0.01924416422843933,
+ "learning_rate": 0.0006,
+ "loss": 4.128574371337891,
+ "step": 2397
+ },
+ {
+ "epoch": 33.30742358078603,
+ "grad_norm": 0.01983758620917797,
+ "learning_rate": 0.0006,
+ "loss": 4.061581611633301,
+ "step": 2398
+ },
+ {
+ "epoch": 33.32139737991266,
+ "grad_norm": 0.01930462196469307,
+ "learning_rate": 0.0006,
+ "loss": 4.091731071472168,
+ "step": 2399
+ },
+ {
+ "epoch": 33.3353711790393,
+ "grad_norm": 0.017988424748182297,
+ "learning_rate": 0.0006,
+ "loss": 4.155599117279053,
+ "step": 2400
+ },
+ {
+ "epoch": 33.34934497816594,
+ "grad_norm": 0.015194767154753208,
+ "learning_rate": 0.0006,
+ "loss": 4.144134044647217,
+ "step": 2401
+ },
+ {
+ "epoch": 33.36331877729258,
+ "grad_norm": 0.01536081824451685,
+ "learning_rate": 0.0006,
+ "loss": 4.127272129058838,
+ "step": 2402
+ },
+ {
+ "epoch": 33.377292576419215,
+ "grad_norm": 0.015431297942996025,
+ "learning_rate": 0.0006,
+ "loss": 4.3213019371032715,
+ "step": 2403
+ },
+ {
+ "epoch": 33.391266375545854,
+ "grad_norm": 0.015740280970931053,
+ "learning_rate": 0.0006,
+ "loss": 4.121467590332031,
+ "step": 2404
+ },
+ {
+ "epoch": 33.40524017467249,
+ "grad_norm": 0.015379384160041809,
+ "learning_rate": 0.0006,
+ "loss": 4.095216751098633,
+ "step": 2405
+ },
+ {
+ "epoch": 33.419213973799124,
+ "grad_norm": 0.01528073474764824,
+ "learning_rate": 0.0006,
+ "loss": 4.13047981262207,
+ "step": 2406
+ },
+ {
+ "epoch": 33.43318777292576,
+ "grad_norm": 0.015739979222416878,
+ "learning_rate": 0.0006,
+ "loss": 4.2533135414123535,
+ "step": 2407
+ },
+ {
+ "epoch": 33.4471615720524,
+ "grad_norm": 0.01519764307886362,
+ "learning_rate": 0.0006,
+ "loss": 4.131782054901123,
+ "step": 2408
+ },
+ {
+ "epoch": 33.46113537117904,
+ "grad_norm": 0.015425996854901314,
+ "learning_rate": 0.0006,
+ "loss": 4.02811336517334,
+ "step": 2409
+ },
+ {
+ "epoch": 33.47510917030568,
+ "grad_norm": 0.015378830023109913,
+ "learning_rate": 0.0006,
+ "loss": 4.08652925491333,
+ "step": 2410
+ },
+ {
+ "epoch": 33.48908296943232,
+ "grad_norm": 0.01473075058311224,
+ "learning_rate": 0.0006,
+ "loss": 4.214364528656006,
+ "step": 2411
+ },
+ {
+ "epoch": 33.50305676855895,
+ "grad_norm": 0.015141637064516544,
+ "learning_rate": 0.0006,
+ "loss": 4.155712127685547,
+ "step": 2412
+ },
+ {
+ "epoch": 33.51703056768559,
+ "grad_norm": 0.01641593500971794,
+ "learning_rate": 0.0006,
+ "loss": 4.097907066345215,
+ "step": 2413
+ },
+ {
+ "epoch": 33.531004366812226,
+ "grad_norm": 0.015182029455900192,
+ "learning_rate": 0.0006,
+ "loss": 4.238499641418457,
+ "step": 2414
+ },
+ {
+ "epoch": 33.544978165938865,
+ "grad_norm": 0.01643732376396656,
+ "learning_rate": 0.0006,
+ "loss": 4.0661492347717285,
+ "step": 2415
+ },
+ {
+ "epoch": 33.5589519650655,
+ "grad_norm": 0.015614228323101997,
+ "learning_rate": 0.0006,
+ "loss": 4.120795249938965,
+ "step": 2416
+ },
+ {
+ "epoch": 33.57292576419214,
+ "grad_norm": 0.016066187992691994,
+ "learning_rate": 0.0006,
+ "loss": 4.096465110778809,
+ "step": 2417
+ },
+ {
+ "epoch": 33.58689956331878,
+ "grad_norm": 0.017214342951774597,
+ "learning_rate": 0.0006,
+ "loss": 4.242628574371338,
+ "step": 2418
+ },
+ {
+ "epoch": 33.60087336244541,
+ "grad_norm": 0.01787586510181427,
+ "learning_rate": 0.0006,
+ "loss": 4.149589538574219,
+ "step": 2419
+ },
+ {
+ "epoch": 33.61484716157205,
+ "grad_norm": 0.019455241039395332,
+ "learning_rate": 0.0006,
+ "loss": 4.267796039581299,
+ "step": 2420
+ },
+ {
+ "epoch": 33.62882096069869,
+ "grad_norm": 0.01984674669802189,
+ "learning_rate": 0.0006,
+ "loss": 4.175220966339111,
+ "step": 2421
+ },
+ {
+ "epoch": 33.64279475982533,
+ "grad_norm": 0.021933717653155327,
+ "learning_rate": 0.0006,
+ "loss": 4.106647491455078,
+ "step": 2422
+ },
+ {
+ "epoch": 33.65676855895197,
+ "grad_norm": 0.02123589813709259,
+ "learning_rate": 0.0006,
+ "loss": 4.204996109008789,
+ "step": 2423
+ },
+ {
+ "epoch": 33.670742358078606,
+ "grad_norm": 0.019814589992165565,
+ "learning_rate": 0.0006,
+ "loss": 4.186171531677246,
+ "step": 2424
+ },
+ {
+ "epoch": 33.68471615720524,
+ "grad_norm": 0.02059728465974331,
+ "learning_rate": 0.0006,
+ "loss": 4.170509338378906,
+ "step": 2425
+ },
+ {
+ "epoch": 33.698689956331876,
+ "grad_norm": 0.018483608961105347,
+ "learning_rate": 0.0006,
+ "loss": 4.085585594177246,
+ "step": 2426
+ },
+ {
+ "epoch": 33.712663755458514,
+ "grad_norm": 0.018403852358460426,
+ "learning_rate": 0.0006,
+ "loss": 4.2551774978637695,
+ "step": 2427
+ },
+ {
+ "epoch": 33.72663755458515,
+ "grad_norm": 0.018653830513358116,
+ "learning_rate": 0.0006,
+ "loss": 4.091766357421875,
+ "step": 2428
+ },
+ {
+ "epoch": 33.74061135371179,
+ "grad_norm": 0.017692334949970245,
+ "learning_rate": 0.0006,
+ "loss": 4.100838661193848,
+ "step": 2429
+ },
+ {
+ "epoch": 33.75458515283843,
+ "grad_norm": 0.016029126942157745,
+ "learning_rate": 0.0006,
+ "loss": 4.1530680656433105,
+ "step": 2430
+ },
+ {
+ "epoch": 33.76855895196506,
+ "grad_norm": 0.016095684841275215,
+ "learning_rate": 0.0006,
+ "loss": 4.20544958114624,
+ "step": 2431
+ },
+ {
+ "epoch": 33.7825327510917,
+ "grad_norm": 0.016375798732042313,
+ "learning_rate": 0.0006,
+ "loss": 4.199177265167236,
+ "step": 2432
+ },
+ {
+ "epoch": 33.79650655021834,
+ "grad_norm": 0.01590786688029766,
+ "learning_rate": 0.0006,
+ "loss": 4.200799465179443,
+ "step": 2433
+ },
+ {
+ "epoch": 33.81048034934498,
+ "grad_norm": 0.014896097593009472,
+ "learning_rate": 0.0006,
+ "loss": 4.185850143432617,
+ "step": 2434
+ },
+ {
+ "epoch": 33.82445414847162,
+ "grad_norm": 0.01438513956964016,
+ "learning_rate": 0.0006,
+ "loss": 4.214962005615234,
+ "step": 2435
+ },
+ {
+ "epoch": 33.838427947598255,
+ "grad_norm": 0.014459339901804924,
+ "learning_rate": 0.0006,
+ "loss": 4.196972370147705,
+ "step": 2436
+ },
+ {
+ "epoch": 33.852401746724894,
+ "grad_norm": 0.014782169833779335,
+ "learning_rate": 0.0006,
+ "loss": 4.0885910987854,
+ "step": 2437
+ },
+ {
+ "epoch": 33.866375545851525,
+ "grad_norm": 0.01520370040088892,
+ "learning_rate": 0.0006,
+ "loss": 4.0323262214660645,
+ "step": 2438
+ },
+ {
+ "epoch": 33.880349344978164,
+ "grad_norm": 0.014056282117962837,
+ "learning_rate": 0.0006,
+ "loss": 4.1386518478393555,
+ "step": 2439
+ },
+ {
+ "epoch": 33.8943231441048,
+ "grad_norm": 0.013567990623414516,
+ "learning_rate": 0.0006,
+ "loss": 4.095754623413086,
+ "step": 2440
+ },
+ {
+ "epoch": 33.90829694323144,
+ "grad_norm": 0.013306716457009315,
+ "learning_rate": 0.0006,
+ "loss": 4.1623077392578125,
+ "step": 2441
+ },
+ {
+ "epoch": 33.92227074235808,
+ "grad_norm": 0.014047150500118732,
+ "learning_rate": 0.0006,
+ "loss": 4.142541885375977,
+ "step": 2442
+ },
+ {
+ "epoch": 33.93624454148472,
+ "grad_norm": 0.01377617847174406,
+ "learning_rate": 0.0006,
+ "loss": 4.268551826477051,
+ "step": 2443
+ },
+ {
+ "epoch": 33.95021834061135,
+ "grad_norm": 0.014671621844172478,
+ "learning_rate": 0.0006,
+ "loss": 4.198196887969971,
+ "step": 2444
+ },
+ {
+ "epoch": 33.96419213973799,
+ "grad_norm": 0.014806744642555714,
+ "learning_rate": 0.0006,
+ "loss": 4.057899475097656,
+ "step": 2445
+ },
+ {
+ "epoch": 33.97816593886463,
+ "grad_norm": 0.015734272077679634,
+ "learning_rate": 0.0006,
+ "loss": 4.163505554199219,
+ "step": 2446
+ },
+ {
+ "epoch": 33.992139737991266,
+ "grad_norm": 0.015250146389007568,
+ "learning_rate": 0.0006,
+ "loss": 4.076294898986816,
+ "step": 2447
+ },
+ {
+ "epoch": 34.0,
+ "grad_norm": 0.016733702272176743,
+ "learning_rate": 0.0006,
+ "loss": 4.138420104980469,
+ "step": 2448
+ },
+ {
+ "epoch": 34.0,
+ "eval_loss": 4.542503356933594,
+ "eval_runtime": 56.983,
+ "eval_samples_per_second": 42.855,
+ "eval_steps_per_second": 1.351,
+ "step": 2448
+ },
+ {
+ "epoch": 34.01397379912664,
+ "grad_norm": 0.015866613015532494,
+ "learning_rate": 0.0006,
+ "loss": 4.032081127166748,
+ "step": 2449
+ },
+ {
+ "epoch": 34.02794759825328,
+ "grad_norm": 0.01946570910513401,
+ "learning_rate": 0.0006,
+ "loss": 4.122320175170898,
+ "step": 2450
+ },
+ {
+ "epoch": 34.041921397379916,
+ "grad_norm": 0.020265938714146614,
+ "learning_rate": 0.0006,
+ "loss": 4.06460428237915,
+ "step": 2451
+ },
+ {
+ "epoch": 34.05589519650655,
+ "grad_norm": 0.018887171521782875,
+ "learning_rate": 0.0006,
+ "loss": 4.096906661987305,
+ "step": 2452
+ },
+ {
+ "epoch": 34.069868995633186,
+ "grad_norm": 0.017304692417383194,
+ "learning_rate": 0.0006,
+ "loss": 4.137261390686035,
+ "step": 2453
+ },
+ {
+ "epoch": 34.083842794759825,
+ "grad_norm": 0.017029082402586937,
+ "learning_rate": 0.0006,
+ "loss": 4.081112861633301,
+ "step": 2454
+ },
+ {
+ "epoch": 34.09781659388646,
+ "grad_norm": 0.016414813697338104,
+ "learning_rate": 0.0006,
+ "loss": 4.140783309936523,
+ "step": 2455
+ },
+ {
+ "epoch": 34.1117903930131,
+ "grad_norm": 0.018238281831145287,
+ "learning_rate": 0.0006,
+ "loss": 4.161116600036621,
+ "step": 2456
+ },
+ {
+ "epoch": 34.12576419213974,
+ "grad_norm": 0.019929232075810432,
+ "learning_rate": 0.0006,
+ "loss": 4.243722438812256,
+ "step": 2457
+ },
+ {
+ "epoch": 34.13973799126637,
+ "grad_norm": 0.019375786185264587,
+ "learning_rate": 0.0006,
+ "loss": 4.134873390197754,
+ "step": 2458
+ },
+ {
+ "epoch": 34.15371179039301,
+ "grad_norm": 0.018151137977838516,
+ "learning_rate": 0.0006,
+ "loss": 4.098745346069336,
+ "step": 2459
+ },
+ {
+ "epoch": 34.16768558951965,
+ "grad_norm": 0.017054934054613113,
+ "learning_rate": 0.0006,
+ "loss": 4.194662570953369,
+ "step": 2460
+ },
+ {
+ "epoch": 34.18165938864629,
+ "grad_norm": 0.018220555037260056,
+ "learning_rate": 0.0006,
+ "loss": 4.128866195678711,
+ "step": 2461
+ },
+ {
+ "epoch": 34.19563318777293,
+ "grad_norm": 0.01692879945039749,
+ "learning_rate": 0.0006,
+ "loss": 4.062477111816406,
+ "step": 2462
+ },
+ {
+ "epoch": 34.209606986899566,
+ "grad_norm": 0.01674199476838112,
+ "learning_rate": 0.0006,
+ "loss": 4.186149597167969,
+ "step": 2463
+ },
+ {
+ "epoch": 34.223580786026204,
+ "grad_norm": 0.016527455300092697,
+ "learning_rate": 0.0006,
+ "loss": 4.0734028816223145,
+ "step": 2464
+ },
+ {
+ "epoch": 34.237554585152836,
+ "grad_norm": 0.016448386013507843,
+ "learning_rate": 0.0006,
+ "loss": 4.098896026611328,
+ "step": 2465
+ },
+ {
+ "epoch": 34.251528384279474,
+ "grad_norm": 0.014644953422248363,
+ "learning_rate": 0.0006,
+ "loss": 4.1031904220581055,
+ "step": 2466
+ },
+ {
+ "epoch": 34.26550218340611,
+ "grad_norm": 0.016649341210722923,
+ "learning_rate": 0.0006,
+ "loss": 4.148791313171387,
+ "step": 2467
+ },
+ {
+ "epoch": 34.27947598253275,
+ "grad_norm": 0.017442714422941208,
+ "learning_rate": 0.0006,
+ "loss": 4.070799827575684,
+ "step": 2468
+ },
+ {
+ "epoch": 34.29344978165939,
+ "grad_norm": 0.017852013930678368,
+ "learning_rate": 0.0006,
+ "loss": 4.031924247741699,
+ "step": 2469
+ },
+ {
+ "epoch": 34.30742358078603,
+ "grad_norm": 0.016371270641684532,
+ "learning_rate": 0.0006,
+ "loss": 4.149420738220215,
+ "step": 2470
+ },
+ {
+ "epoch": 34.32139737991266,
+ "grad_norm": 0.015846814960241318,
+ "learning_rate": 0.0006,
+ "loss": 4.15427303314209,
+ "step": 2471
+ },
+ {
+ "epoch": 34.3353711790393,
+ "grad_norm": 0.017411163076758385,
+ "learning_rate": 0.0006,
+ "loss": 4.1304192543029785,
+ "step": 2472
+ },
+ {
+ "epoch": 34.34934497816594,
+ "grad_norm": 0.01780046708881855,
+ "learning_rate": 0.0006,
+ "loss": 4.191732406616211,
+ "step": 2473
+ },
+ {
+ "epoch": 34.36331877729258,
+ "grad_norm": 0.01958383060991764,
+ "learning_rate": 0.0006,
+ "loss": 4.068355083465576,
+ "step": 2474
+ },
+ {
+ "epoch": 34.377292576419215,
+ "grad_norm": 0.022173317149281502,
+ "learning_rate": 0.0006,
+ "loss": 4.182929039001465,
+ "step": 2475
+ },
+ {
+ "epoch": 34.391266375545854,
+ "grad_norm": 0.022389156743884087,
+ "learning_rate": 0.0006,
+ "loss": 4.110218048095703,
+ "step": 2476
+ },
+ {
+ "epoch": 34.40524017467249,
+ "grad_norm": 0.022478511556982994,
+ "learning_rate": 0.0006,
+ "loss": 4.095479488372803,
+ "step": 2477
+ },
+ {
+ "epoch": 34.419213973799124,
+ "grad_norm": 0.019504647701978683,
+ "learning_rate": 0.0006,
+ "loss": 4.101964950561523,
+ "step": 2478
+ },
+ {
+ "epoch": 34.43318777292576,
+ "grad_norm": 0.019378576427698135,
+ "learning_rate": 0.0006,
+ "loss": 4.14620304107666,
+ "step": 2479
+ },
+ {
+ "epoch": 34.4471615720524,
+ "grad_norm": 0.017993204295635223,
+ "learning_rate": 0.0006,
+ "loss": 4.183492660522461,
+ "step": 2480
+ },
+ {
+ "epoch": 34.46113537117904,
+ "grad_norm": 0.01814805157482624,
+ "learning_rate": 0.0006,
+ "loss": 4.151247978210449,
+ "step": 2481
+ },
+ {
+ "epoch": 34.47510917030568,
+ "grad_norm": 0.018622448667883873,
+ "learning_rate": 0.0006,
+ "loss": 4.020938396453857,
+ "step": 2482
+ },
+ {
+ "epoch": 34.48908296943232,
+ "grad_norm": 0.018360132351517677,
+ "learning_rate": 0.0006,
+ "loss": 4.064734935760498,
+ "step": 2483
+ },
+ {
+ "epoch": 34.50305676855895,
+ "grad_norm": 0.017715711146593094,
+ "learning_rate": 0.0006,
+ "loss": 4.185687065124512,
+ "step": 2484
+ },
+ {
+ "epoch": 34.51703056768559,
+ "grad_norm": 0.01714295893907547,
+ "learning_rate": 0.0006,
+ "loss": 4.1044721603393555,
+ "step": 2485
+ },
+ {
+ "epoch": 34.531004366812226,
+ "grad_norm": 0.016943803057074547,
+ "learning_rate": 0.0006,
+ "loss": 4.106837272644043,
+ "step": 2486
+ },
+ {
+ "epoch": 34.544978165938865,
+ "grad_norm": 0.0182423684746027,
+ "learning_rate": 0.0006,
+ "loss": 4.1041340827941895,
+ "step": 2487
+ },
+ {
+ "epoch": 34.5589519650655,
+ "grad_norm": 0.01636587455868721,
+ "learning_rate": 0.0006,
+ "loss": 3.9735546112060547,
+ "step": 2488
+ },
+ {
+ "epoch": 34.57292576419214,
+ "grad_norm": 0.014593546278774738,
+ "learning_rate": 0.0006,
+ "loss": 4.041980266571045,
+ "step": 2489
+ },
+ {
+ "epoch": 34.58689956331878,
+ "grad_norm": 0.015436965972185135,
+ "learning_rate": 0.0006,
+ "loss": 4.301148414611816,
+ "step": 2490
+ },
+ {
+ "epoch": 34.60087336244541,
+ "grad_norm": 0.015075269155204296,
+ "learning_rate": 0.0006,
+ "loss": 4.118527412414551,
+ "step": 2491
+ },
+ {
+ "epoch": 34.61484716157205,
+ "grad_norm": 0.01701514609158039,
+ "learning_rate": 0.0006,
+ "loss": 4.245968818664551,
+ "step": 2492
+ },
+ {
+ "epoch": 34.62882096069869,
+ "grad_norm": 0.018047887831926346,
+ "learning_rate": 0.0006,
+ "loss": 4.227261066436768,
+ "step": 2493
+ },
+ {
+ "epoch": 34.64279475982533,
+ "grad_norm": 0.018499085679650307,
+ "learning_rate": 0.0006,
+ "loss": 4.121360778808594,
+ "step": 2494
+ },
+ {
+ "epoch": 34.65676855895197,
+ "grad_norm": 0.018987711519002914,
+ "learning_rate": 0.0006,
+ "loss": 4.131564617156982,
+ "step": 2495
+ },
+ {
+ "epoch": 34.670742358078606,
+ "grad_norm": 0.01834476925432682,
+ "learning_rate": 0.0006,
+ "loss": 4.224941253662109,
+ "step": 2496
+ },
+ {
+ "epoch": 34.68471615720524,
+ "grad_norm": 0.017507942393422127,
+ "learning_rate": 0.0006,
+ "loss": 4.1394548416137695,
+ "step": 2497
+ },
+ {
+ "epoch": 34.698689956331876,
+ "grad_norm": 0.01680580899119377,
+ "learning_rate": 0.0006,
+ "loss": 4.1073431968688965,
+ "step": 2498
+ },
+ {
+ "epoch": 34.712663755458514,
+ "grad_norm": 0.015233514830470085,
+ "learning_rate": 0.0006,
+ "loss": 4.136994361877441,
+ "step": 2499
+ },
+ {
+ "epoch": 34.72663755458515,
+ "grad_norm": 0.015482917428016663,
+ "learning_rate": 0.0006,
+ "loss": 4.110719680786133,
+ "step": 2500
+ },
+ {
+ "epoch": 34.74061135371179,
+ "grad_norm": 0.015940461307764053,
+ "learning_rate": 0.0006,
+ "loss": 4.207851886749268,
+ "step": 2501
+ },
+ {
+ "epoch": 34.75458515283843,
+ "grad_norm": 0.015355597250163555,
+ "learning_rate": 0.0006,
+ "loss": 4.169469833374023,
+ "step": 2502
+ },
+ {
+ "epoch": 34.76855895196506,
+ "grad_norm": 0.015974000096321106,
+ "learning_rate": 0.0006,
+ "loss": 4.110061168670654,
+ "step": 2503
+ },
+ {
+ "epoch": 34.7825327510917,
+ "grad_norm": 0.016088927164673805,
+ "learning_rate": 0.0006,
+ "loss": 4.155452728271484,
+ "step": 2504
+ },
+ {
+ "epoch": 34.79650655021834,
+ "grad_norm": 0.016156120225787163,
+ "learning_rate": 0.0006,
+ "loss": 4.178092956542969,
+ "step": 2505
+ },
+ {
+ "epoch": 34.81048034934498,
+ "grad_norm": 0.014928647316992283,
+ "learning_rate": 0.0006,
+ "loss": 4.05457878112793,
+ "step": 2506
+ },
+ {
+ "epoch": 34.82445414847162,
+ "grad_norm": 0.015289328061044216,
+ "learning_rate": 0.0006,
+ "loss": 4.18869686126709,
+ "step": 2507
+ },
+ {
+ "epoch": 34.838427947598255,
+ "grad_norm": 0.015799039974808693,
+ "learning_rate": 0.0006,
+ "loss": 4.106073379516602,
+ "step": 2508
+ },
+ {
+ "epoch": 34.852401746724894,
+ "grad_norm": 0.015824418514966965,
+ "learning_rate": 0.0006,
+ "loss": 4.0634589195251465,
+ "step": 2509
+ },
+ {
+ "epoch": 34.866375545851525,
+ "grad_norm": 0.015321656130254269,
+ "learning_rate": 0.0006,
+ "loss": 4.093090057373047,
+ "step": 2510
+ },
+ {
+ "epoch": 34.880349344978164,
+ "grad_norm": 0.014380166307091713,
+ "learning_rate": 0.0006,
+ "loss": 4.1750640869140625,
+ "step": 2511
+ },
+ {
+ "epoch": 34.8943231441048,
+ "grad_norm": 0.014962567947804928,
+ "learning_rate": 0.0006,
+ "loss": 4.185277938842773,
+ "step": 2512
+ },
+ {
+ "epoch": 34.90829694323144,
+ "grad_norm": 0.014467543922364712,
+ "learning_rate": 0.0006,
+ "loss": 4.1600799560546875,
+ "step": 2513
+ },
+ {
+ "epoch": 34.92227074235808,
+ "grad_norm": 0.01545181218534708,
+ "learning_rate": 0.0006,
+ "loss": 4.096299648284912,
+ "step": 2514
+ },
+ {
+ "epoch": 34.93624454148472,
+ "grad_norm": 0.01605340465903282,
+ "learning_rate": 0.0006,
+ "loss": 4.203323841094971,
+ "step": 2515
+ },
+ {
+ "epoch": 34.95021834061135,
+ "grad_norm": 0.015203815884888172,
+ "learning_rate": 0.0006,
+ "loss": 4.201385021209717,
+ "step": 2516
+ },
+ {
+ "epoch": 34.96419213973799,
+ "grad_norm": 0.015149409882724285,
+ "learning_rate": 0.0006,
+ "loss": 4.088164329528809,
+ "step": 2517
+ },
+ {
+ "epoch": 34.97816593886463,
+ "grad_norm": 0.014885558746755123,
+ "learning_rate": 0.0006,
+ "loss": 4.073911666870117,
+ "step": 2518
+ },
+ {
+ "epoch": 34.992139737991266,
+ "grad_norm": 0.015003419481217861,
+ "learning_rate": 0.0006,
+ "loss": 4.157113552093506,
+ "step": 2519
+ },
+ {
+ "epoch": 35.0,
+ "grad_norm": 0.01641916297376156,
+ "learning_rate": 0.0006,
+ "loss": 3.970052719116211,
+ "step": 2520
+ },
+ {
+ "epoch": 35.0,
+ "eval_loss": 4.641692161560059,
+ "eval_runtime": 56.8836,
+ "eval_samples_per_second": 42.93,
+ "eval_steps_per_second": 1.354,
+ "step": 2520
+ },
+ {
+ "epoch": 35.01397379912664,
+ "grad_norm": 0.015894673764705658,
+ "learning_rate": 0.0006,
+ "loss": 4.104437351226807,
+ "step": 2521
+ },
+ {
+ "epoch": 35.02794759825328,
+ "grad_norm": 0.0173867866396904,
+ "learning_rate": 0.0006,
+ "loss": 4.091726303100586,
+ "step": 2522
+ },
+ {
+ "epoch": 35.041921397379916,
+ "grad_norm": 0.019272832199931145,
+ "learning_rate": 0.0006,
+ "loss": 4.217905044555664,
+ "step": 2523
+ },
+ {
+ "epoch": 35.05589519650655,
+ "grad_norm": 0.021820371970534325,
+ "learning_rate": 0.0006,
+ "loss": 4.111428260803223,
+ "step": 2524
+ },
+ {
+ "epoch": 35.069868995633186,
+ "grad_norm": 0.02340741828083992,
+ "learning_rate": 0.0006,
+ "loss": 4.115771293640137,
+ "step": 2525
+ },
+ {
+ "epoch": 35.083842794759825,
+ "grad_norm": 0.023760691285133362,
+ "learning_rate": 0.0006,
+ "loss": 4.132383346557617,
+ "step": 2526
+ },
+ {
+ "epoch": 35.09781659388646,
+ "grad_norm": 0.02073468267917633,
+ "learning_rate": 0.0006,
+ "loss": 4.272691249847412,
+ "step": 2527
+ },
+ {
+ "epoch": 35.1117903930131,
+ "grad_norm": 0.02197718247771263,
+ "learning_rate": 0.0006,
+ "loss": 4.105586528778076,
+ "step": 2528
+ },
+ {
+ "epoch": 35.12576419213974,
+ "grad_norm": 0.0200254675000906,
+ "learning_rate": 0.0006,
+ "loss": 3.968820095062256,
+ "step": 2529
+ },
+ {
+ "epoch": 35.13973799126637,
+ "grad_norm": 0.01929587684571743,
+ "learning_rate": 0.0006,
+ "loss": 4.045468807220459,
+ "step": 2530
+ },
+ {
+ "epoch": 35.15371179039301,
+ "grad_norm": 0.01886162906885147,
+ "learning_rate": 0.0006,
+ "loss": 4.185571670532227,
+ "step": 2531
+ },
+ {
+ "epoch": 35.16768558951965,
+ "grad_norm": 0.019789839163422585,
+ "learning_rate": 0.0006,
+ "loss": 3.9734551906585693,
+ "step": 2532
+ },
+ {
+ "epoch": 35.18165938864629,
+ "grad_norm": 0.0198142658919096,
+ "learning_rate": 0.0006,
+ "loss": 4.2662506103515625,
+ "step": 2533
+ },
+ {
+ "epoch": 35.19563318777293,
+ "grad_norm": 0.02087012305855751,
+ "learning_rate": 0.0006,
+ "loss": 4.175931930541992,
+ "step": 2534
+ },
+ {
+ "epoch": 35.209606986899566,
+ "grad_norm": 0.020294535905122757,
+ "learning_rate": 0.0006,
+ "loss": 4.060324668884277,
+ "step": 2535
+ },
+ {
+ "epoch": 35.223580786026204,
+ "grad_norm": 0.01908983662724495,
+ "learning_rate": 0.0006,
+ "loss": 4.217939853668213,
+ "step": 2536
+ },
+ {
+ "epoch": 35.237554585152836,
+ "grad_norm": 0.0178892333060503,
+ "learning_rate": 0.0006,
+ "loss": 4.190216541290283,
+ "step": 2537
+ },
+ {
+ "epoch": 35.251528384279474,
+ "grad_norm": 0.017892315983772278,
+ "learning_rate": 0.0006,
+ "loss": 4.067419052124023,
+ "step": 2538
+ },
+ {
+ "epoch": 35.26550218340611,
+ "grad_norm": 0.018556250259280205,
+ "learning_rate": 0.0006,
+ "loss": 4.105958461761475,
+ "step": 2539
+ },
+ {
+ "epoch": 35.27947598253275,
+ "grad_norm": 0.01944422349333763,
+ "learning_rate": 0.0006,
+ "loss": 4.155775547027588,
+ "step": 2540
+ },
+ {
+ "epoch": 35.29344978165939,
+ "grad_norm": 0.017333146184682846,
+ "learning_rate": 0.0006,
+ "loss": 4.244513511657715,
+ "step": 2541
+ },
+ {
+ "epoch": 35.30742358078603,
+ "grad_norm": 0.01639244332909584,
+ "learning_rate": 0.0006,
+ "loss": 4.167815685272217,
+ "step": 2542
+ },
+ {
+ "epoch": 35.32139737991266,
+ "grad_norm": 0.01610143668949604,
+ "learning_rate": 0.0006,
+ "loss": 4.121486663818359,
+ "step": 2543
+ },
+ {
+ "epoch": 35.3353711790393,
+ "grad_norm": 0.01656411588191986,
+ "learning_rate": 0.0006,
+ "loss": 4.1755499839782715,
+ "step": 2544
+ },
+ {
+ "epoch": 35.34934497816594,
+ "grad_norm": 0.015075297094881535,
+ "learning_rate": 0.0006,
+ "loss": 4.102992534637451,
+ "step": 2545
+ },
+ {
+ "epoch": 35.36331877729258,
+ "grad_norm": 0.01538170501589775,
+ "learning_rate": 0.0006,
+ "loss": 4.161603927612305,
+ "step": 2546
+ },
+ {
+ "epoch": 35.377292576419215,
+ "grad_norm": 0.01483312901109457,
+ "learning_rate": 0.0006,
+ "loss": 4.085355758666992,
+ "step": 2547
+ },
+ {
+ "epoch": 35.391266375545854,
+ "grad_norm": 0.014894738793373108,
+ "learning_rate": 0.0006,
+ "loss": 4.083585262298584,
+ "step": 2548
+ },
+ {
+ "epoch": 35.40524017467249,
+ "grad_norm": 0.01475780550390482,
+ "learning_rate": 0.0006,
+ "loss": 4.012645721435547,
+ "step": 2549
+ },
+ {
+ "epoch": 35.419213973799124,
+ "grad_norm": 0.014531653374433517,
+ "learning_rate": 0.0006,
+ "loss": 4.197113990783691,
+ "step": 2550
+ },
+ {
+ "epoch": 35.43318777292576,
+ "grad_norm": 0.015316493809223175,
+ "learning_rate": 0.0006,
+ "loss": 4.227970123291016,
+ "step": 2551
+ },
+ {
+ "epoch": 35.4471615720524,
+ "grad_norm": 0.014685571193695068,
+ "learning_rate": 0.0006,
+ "loss": 4.089415550231934,
+ "step": 2552
+ },
+ {
+ "epoch": 35.46113537117904,
+ "grad_norm": 0.015337035991251469,
+ "learning_rate": 0.0006,
+ "loss": 4.080583572387695,
+ "step": 2553
+ },
+ {
+ "epoch": 35.47510917030568,
+ "grad_norm": 0.016754208132624626,
+ "learning_rate": 0.0006,
+ "loss": 4.163935661315918,
+ "step": 2554
+ },
+ {
+ "epoch": 35.48908296943232,
+ "grad_norm": 0.017535798251628876,
+ "learning_rate": 0.0006,
+ "loss": 4.093027591705322,
+ "step": 2555
+ },
+ {
+ "epoch": 35.50305676855895,
+ "grad_norm": 0.017336975783109665,
+ "learning_rate": 0.0006,
+ "loss": 4.127961158752441,
+ "step": 2556
+ },
+ {
+ "epoch": 35.51703056768559,
+ "grad_norm": 0.016242152079939842,
+ "learning_rate": 0.0006,
+ "loss": 4.155593395233154,
+ "step": 2557
+ },
+ {
+ "epoch": 35.531004366812226,
+ "grad_norm": 0.015121322125196457,
+ "learning_rate": 0.0006,
+ "loss": 3.984004020690918,
+ "step": 2558
+ },
+ {
+ "epoch": 35.544978165938865,
+ "grad_norm": 0.015178884379565716,
+ "learning_rate": 0.0006,
+ "loss": 4.111564636230469,
+ "step": 2559
+ },
+ {
+ "epoch": 35.5589519650655,
+ "grad_norm": 0.014777131378650665,
+ "learning_rate": 0.0006,
+ "loss": 4.041402816772461,
+ "step": 2560
+ },
+ {
+ "epoch": 35.57292576419214,
+ "grad_norm": 0.015197676606476307,
+ "learning_rate": 0.0006,
+ "loss": 4.133706092834473,
+ "step": 2561
+ },
+ {
+ "epoch": 35.58689956331878,
+ "grad_norm": 0.015905972570180893,
+ "learning_rate": 0.0006,
+ "loss": 4.1069135665893555,
+ "step": 2562
+ },
+ {
+ "epoch": 35.60087336244541,
+ "grad_norm": 0.015942877158522606,
+ "learning_rate": 0.0006,
+ "loss": 4.0720624923706055,
+ "step": 2563
+ },
+ {
+ "epoch": 35.61484716157205,
+ "grad_norm": 0.016697583720088005,
+ "learning_rate": 0.0006,
+ "loss": 4.167448043823242,
+ "step": 2564
+ },
+ {
+ "epoch": 35.62882096069869,
+ "grad_norm": 0.015603577718138695,
+ "learning_rate": 0.0006,
+ "loss": 4.127327919006348,
+ "step": 2565
+ },
+ {
+ "epoch": 35.64279475982533,
+ "grad_norm": 0.017153460532426834,
+ "learning_rate": 0.0006,
+ "loss": 4.056543350219727,
+ "step": 2566
+ },
+ {
+ "epoch": 35.65676855895197,
+ "grad_norm": 0.016731958836317062,
+ "learning_rate": 0.0006,
+ "loss": 4.0681843757629395,
+ "step": 2567
+ },
+ {
+ "epoch": 35.670742358078606,
+ "grad_norm": 0.016167359426617622,
+ "learning_rate": 0.0006,
+ "loss": 4.174992084503174,
+ "step": 2568
+ },
+ {
+ "epoch": 35.68471615720524,
+ "grad_norm": 0.01615108922123909,
+ "learning_rate": 0.0006,
+ "loss": 4.1383161544799805,
+ "step": 2569
+ },
+ {
+ "epoch": 35.698689956331876,
+ "grad_norm": 0.015750816091895103,
+ "learning_rate": 0.0006,
+ "loss": 4.031595706939697,
+ "step": 2570
+ },
+ {
+ "epoch": 35.712663755458514,
+ "grad_norm": 0.015416189096868038,
+ "learning_rate": 0.0006,
+ "loss": 4.067498207092285,
+ "step": 2571
+ },
+ {
+ "epoch": 35.72663755458515,
+ "grad_norm": 0.01570792682468891,
+ "learning_rate": 0.0006,
+ "loss": 4.1338300704956055,
+ "step": 2572
+ },
+ {
+ "epoch": 35.74061135371179,
+ "grad_norm": 0.016367752104997635,
+ "learning_rate": 0.0006,
+ "loss": 4.152523040771484,
+ "step": 2573
+ },
+ {
+ "epoch": 35.75458515283843,
+ "grad_norm": 0.016349803656339645,
+ "learning_rate": 0.0006,
+ "loss": 4.101985454559326,
+ "step": 2574
+ },
+ {
+ "epoch": 35.76855895196506,
+ "grad_norm": 0.01663660630583763,
+ "learning_rate": 0.0006,
+ "loss": 4.190130710601807,
+ "step": 2575
+ },
+ {
+ "epoch": 35.7825327510917,
+ "grad_norm": 0.015599170699715614,
+ "learning_rate": 0.0006,
+ "loss": 4.0792436599731445,
+ "step": 2576
+ },
+ {
+ "epoch": 35.79650655021834,
+ "grad_norm": 0.015472427010536194,
+ "learning_rate": 0.0006,
+ "loss": 4.186446666717529,
+ "step": 2577
+ },
+ {
+ "epoch": 35.81048034934498,
+ "grad_norm": 0.015520439483225346,
+ "learning_rate": 0.0006,
+ "loss": 4.246635437011719,
+ "step": 2578
+ },
+ {
+ "epoch": 35.82445414847162,
+ "grad_norm": 0.016998281702399254,
+ "learning_rate": 0.0006,
+ "loss": 4.13435173034668,
+ "step": 2579
+ },
+ {
+ "epoch": 35.838427947598255,
+ "grad_norm": 0.016290361061692238,
+ "learning_rate": 0.0006,
+ "loss": 4.107428073883057,
+ "step": 2580
+ },
+ {
+ "epoch": 35.852401746724894,
+ "grad_norm": 0.0165516659617424,
+ "learning_rate": 0.0006,
+ "loss": 4.1612138748168945,
+ "step": 2581
+ },
+ {
+ "epoch": 35.866375545851525,
+ "grad_norm": 0.01736091636121273,
+ "learning_rate": 0.0006,
+ "loss": 4.1072869300842285,
+ "step": 2582
+ },
+ {
+ "epoch": 35.880349344978164,
+ "grad_norm": 0.01787031814455986,
+ "learning_rate": 0.0006,
+ "loss": 4.1076812744140625,
+ "step": 2583
+ },
+ {
+ "epoch": 35.8943231441048,
+ "grad_norm": 0.015437985770404339,
+ "learning_rate": 0.0006,
+ "loss": 4.097488880157471,
+ "step": 2584
+ },
+ {
+ "epoch": 35.90829694323144,
+ "grad_norm": 0.016306741163134575,
+ "learning_rate": 0.0006,
+ "loss": 4.105017185211182,
+ "step": 2585
+ },
+ {
+ "epoch": 35.92227074235808,
+ "grad_norm": 0.015723245218396187,
+ "learning_rate": 0.0006,
+ "loss": 4.205283164978027,
+ "step": 2586
+ },
+ {
+ "epoch": 35.93624454148472,
+ "grad_norm": 0.014626456424593925,
+ "learning_rate": 0.0006,
+ "loss": 4.086552619934082,
+ "step": 2587
+ },
+ {
+ "epoch": 35.95021834061135,
+ "grad_norm": 0.014348454773426056,
+ "learning_rate": 0.0006,
+ "loss": 4.171329975128174,
+ "step": 2588
+ },
+ {
+ "epoch": 35.96419213973799,
+ "grad_norm": 0.015598480589687824,
+ "learning_rate": 0.0006,
+ "loss": 4.140815258026123,
+ "step": 2589
+ },
+ {
+ "epoch": 35.97816593886463,
+ "grad_norm": 0.017126062884926796,
+ "learning_rate": 0.0006,
+ "loss": 4.157176971435547,
+ "step": 2590
+ },
+ {
+ "epoch": 35.992139737991266,
+ "grad_norm": 0.01860199309885502,
+ "learning_rate": 0.0006,
+ "loss": 4.252880096435547,
+ "step": 2591
+ },
+ {
+ "epoch": 36.0,
+ "grad_norm": 0.021780220791697502,
+ "learning_rate": 0.0006,
+ "loss": 4.249861717224121,
+ "step": 2592
+ },
+ {
+ "epoch": 36.0,
+ "eval_loss": 4.606466293334961,
+ "eval_runtime": 56.6217,
+ "eval_samples_per_second": 43.128,
+ "eval_steps_per_second": 1.36,
+ "step": 2592
+ },
+ {
+ "epoch": 36.01397379912664,
+ "grad_norm": 0.020892828702926636,
+ "learning_rate": 0.0006,
+ "loss": 4.102055549621582,
+ "step": 2593
+ },
+ {
+ "epoch": 36.02794759825328,
+ "grad_norm": 0.020306255668401718,
+ "learning_rate": 0.0006,
+ "loss": 4.184308052062988,
+ "step": 2594
+ },
+ {
+ "epoch": 36.041921397379916,
+ "grad_norm": 0.02025097794830799,
+ "learning_rate": 0.0006,
+ "loss": 3.9691858291625977,
+ "step": 2595
+ },
+ {
+ "epoch": 36.05589519650655,
+ "grad_norm": 0.020254382863640785,
+ "learning_rate": 0.0006,
+ "loss": 4.02957820892334,
+ "step": 2596
+ },
+ {
+ "epoch": 36.069868995633186,
+ "grad_norm": 0.019307196140289307,
+ "learning_rate": 0.0006,
+ "loss": 4.033181667327881,
+ "step": 2597
+ },
+ {
+ "epoch": 36.083842794759825,
+ "grad_norm": 0.0198476854711771,
+ "learning_rate": 0.0006,
+ "loss": 4.0816874504089355,
+ "step": 2598
+ },
+ {
+ "epoch": 36.09781659388646,
+ "grad_norm": 0.019916515797376633,
+ "learning_rate": 0.0006,
+ "loss": 4.164889812469482,
+ "step": 2599
+ },
+ {
+ "epoch": 36.1117903930131,
+ "grad_norm": 0.0189144778996706,
+ "learning_rate": 0.0006,
+ "loss": 4.038427352905273,
+ "step": 2600
+ },
+ {
+ "epoch": 36.12576419213974,
+ "grad_norm": 0.01754768192768097,
+ "learning_rate": 0.0006,
+ "loss": 4.010427474975586,
+ "step": 2601
+ },
+ {
+ "epoch": 36.13973799126637,
+ "grad_norm": 0.016862990334630013,
+ "learning_rate": 0.0006,
+ "loss": 4.035970211029053,
+ "step": 2602
+ },
+ {
+ "epoch": 36.15371179039301,
+ "grad_norm": 0.019616752862930298,
+ "learning_rate": 0.0006,
+ "loss": 4.079898834228516,
+ "step": 2603
+ },
+ {
+ "epoch": 36.16768558951965,
+ "grad_norm": 0.020131465047597885,
+ "learning_rate": 0.0006,
+ "loss": 4.144684791564941,
+ "step": 2604
+ },
+ {
+ "epoch": 36.18165938864629,
+ "grad_norm": 0.02072521299123764,
+ "learning_rate": 0.0006,
+ "loss": 4.013660430908203,
+ "step": 2605
+ },
+ {
+ "epoch": 36.19563318777293,
+ "grad_norm": 0.01924615167081356,
+ "learning_rate": 0.0006,
+ "loss": 4.071398735046387,
+ "step": 2606
+ },
+ {
+ "epoch": 36.209606986899566,
+ "grad_norm": 0.02013954520225525,
+ "learning_rate": 0.0006,
+ "loss": 4.127094268798828,
+ "step": 2607
+ },
+ {
+ "epoch": 36.223580786026204,
+ "grad_norm": 0.02042008936405182,
+ "learning_rate": 0.0006,
+ "loss": 4.154261589050293,
+ "step": 2608
+ },
+ {
+ "epoch": 36.237554585152836,
+ "grad_norm": 0.021998876705765724,
+ "learning_rate": 0.0006,
+ "loss": 4.204150199890137,
+ "step": 2609
+ },
+ {
+ "epoch": 36.251528384279474,
+ "grad_norm": 0.022946683689951897,
+ "learning_rate": 0.0006,
+ "loss": 4.013632774353027,
+ "step": 2610
+ },
+ {
+ "epoch": 36.26550218340611,
+ "grad_norm": 0.02061830274760723,
+ "learning_rate": 0.0006,
+ "loss": 4.171084403991699,
+ "step": 2611
+ },
+ {
+ "epoch": 36.27947598253275,
+ "grad_norm": 0.01883152686059475,
+ "learning_rate": 0.0006,
+ "loss": 4.149848461151123,
+ "step": 2612
+ },
+ {
+ "epoch": 36.29344978165939,
+ "grad_norm": 0.019945282489061356,
+ "learning_rate": 0.0006,
+ "loss": 4.174656867980957,
+ "step": 2613
+ },
+ {
+ "epoch": 36.30742358078603,
+ "grad_norm": 0.020751051604747772,
+ "learning_rate": 0.0006,
+ "loss": 4.080747127532959,
+ "step": 2614
+ },
+ {
+ "epoch": 36.32139737991266,
+ "grad_norm": 0.01879691146314144,
+ "learning_rate": 0.0006,
+ "loss": 4.057356357574463,
+ "step": 2615
+ },
+ {
+ "epoch": 36.3353711790393,
+ "grad_norm": 0.019037071615457535,
+ "learning_rate": 0.0006,
+ "loss": 4.204607009887695,
+ "step": 2616
+ },
+ {
+ "epoch": 36.34934497816594,
+ "grad_norm": 0.01872078701853752,
+ "learning_rate": 0.0006,
+ "loss": 4.050586700439453,
+ "step": 2617
+ },
+ {
+ "epoch": 36.36331877729258,
+ "grad_norm": 0.018092811107635498,
+ "learning_rate": 0.0006,
+ "loss": 4.102832794189453,
+ "step": 2618
+ },
+ {
+ "epoch": 36.377292576419215,
+ "grad_norm": 0.017010487616062164,
+ "learning_rate": 0.0006,
+ "loss": 4.191788673400879,
+ "step": 2619
+ },
+ {
+ "epoch": 36.391266375545854,
+ "grad_norm": 0.01691143773496151,
+ "learning_rate": 0.0006,
+ "loss": 4.160456657409668,
+ "step": 2620
+ },
+ {
+ "epoch": 36.40524017467249,
+ "grad_norm": 0.017046816647052765,
+ "learning_rate": 0.0006,
+ "loss": 4.073828220367432,
+ "step": 2621
+ },
+ {
+ "epoch": 36.419213973799124,
+ "grad_norm": 0.01707981713116169,
+ "learning_rate": 0.0006,
+ "loss": 4.0980095863342285,
+ "step": 2622
+ },
+ {
+ "epoch": 36.43318777292576,
+ "grad_norm": 0.01630452275276184,
+ "learning_rate": 0.0006,
+ "loss": 4.101321220397949,
+ "step": 2623
+ },
+ {
+ "epoch": 36.4471615720524,
+ "grad_norm": 0.01576915942132473,
+ "learning_rate": 0.0006,
+ "loss": 4.158249855041504,
+ "step": 2624
+ },
+ {
+ "epoch": 36.46113537117904,
+ "grad_norm": 0.016179712489247322,
+ "learning_rate": 0.0006,
+ "loss": 4.077852249145508,
+ "step": 2625
+ },
+ {
+ "epoch": 36.47510917030568,
+ "grad_norm": 0.01609169878065586,
+ "learning_rate": 0.0006,
+ "loss": 4.187317848205566,
+ "step": 2626
+ },
+ {
+ "epoch": 36.48908296943232,
+ "grad_norm": 0.017148254439234734,
+ "learning_rate": 0.0006,
+ "loss": 4.109071731567383,
+ "step": 2627
+ },
+ {
+ "epoch": 36.50305676855895,
+ "grad_norm": 0.015842624008655548,
+ "learning_rate": 0.0006,
+ "loss": 4.145096302032471,
+ "step": 2628
+ },
+ {
+ "epoch": 36.51703056768559,
+ "grad_norm": 0.014906194061040878,
+ "learning_rate": 0.0006,
+ "loss": 4.028440475463867,
+ "step": 2629
+ },
+ {
+ "epoch": 36.531004366812226,
+ "grad_norm": 0.01485038734972477,
+ "learning_rate": 0.0006,
+ "loss": 4.175201416015625,
+ "step": 2630
+ },
+ {
+ "epoch": 36.544978165938865,
+ "grad_norm": 0.015834150835871696,
+ "learning_rate": 0.0006,
+ "loss": 4.076434135437012,
+ "step": 2631
+ },
+ {
+ "epoch": 36.5589519650655,
+ "grad_norm": 0.015411358326673508,
+ "learning_rate": 0.0006,
+ "loss": 4.03912878036499,
+ "step": 2632
+ },
+ {
+ "epoch": 36.57292576419214,
+ "grad_norm": 0.014762846753001213,
+ "learning_rate": 0.0006,
+ "loss": 4.019827365875244,
+ "step": 2633
+ },
+ {
+ "epoch": 36.58689956331878,
+ "grad_norm": 0.015269497409462929,
+ "learning_rate": 0.0006,
+ "loss": 4.116957187652588,
+ "step": 2634
+ },
+ {
+ "epoch": 36.60087336244541,
+ "grad_norm": 0.015171617269515991,
+ "learning_rate": 0.0006,
+ "loss": 4.2069926261901855,
+ "step": 2635
+ },
+ {
+ "epoch": 36.61484716157205,
+ "grad_norm": 0.014770902693271637,
+ "learning_rate": 0.0006,
+ "loss": 4.092021942138672,
+ "step": 2636
+ },
+ {
+ "epoch": 36.62882096069869,
+ "grad_norm": 0.014957531355321407,
+ "learning_rate": 0.0006,
+ "loss": 4.16819953918457,
+ "step": 2637
+ },
+ {
+ "epoch": 36.64279475982533,
+ "grad_norm": 0.014774451963603497,
+ "learning_rate": 0.0006,
+ "loss": 4.046073913574219,
+ "step": 2638
+ },
+ {
+ "epoch": 36.65676855895197,
+ "grad_norm": 0.014320445246994495,
+ "learning_rate": 0.0006,
+ "loss": 3.977792501449585,
+ "step": 2639
+ },
+ {
+ "epoch": 36.670742358078606,
+ "grad_norm": 0.01413130946457386,
+ "learning_rate": 0.0006,
+ "loss": 4.130649566650391,
+ "step": 2640
+ },
+ {
+ "epoch": 36.68471615720524,
+ "grad_norm": 0.014336307533085346,
+ "learning_rate": 0.0006,
+ "loss": 4.161357402801514,
+ "step": 2641
+ },
+ {
+ "epoch": 36.698689956331876,
+ "grad_norm": 0.014516931027173996,
+ "learning_rate": 0.0006,
+ "loss": 4.14248514175415,
+ "step": 2642
+ },
+ {
+ "epoch": 36.712663755458514,
+ "grad_norm": 0.014377126470208168,
+ "learning_rate": 0.0006,
+ "loss": 4.062711715698242,
+ "step": 2643
+ },
+ {
+ "epoch": 36.72663755458515,
+ "grad_norm": 0.014332668855786324,
+ "learning_rate": 0.0006,
+ "loss": 4.06423807144165,
+ "step": 2644
+ },
+ {
+ "epoch": 36.74061135371179,
+ "grad_norm": 0.014030571095645428,
+ "learning_rate": 0.0006,
+ "loss": 4.178613662719727,
+ "step": 2645
+ },
+ {
+ "epoch": 36.75458515283843,
+ "grad_norm": 0.014376776292920113,
+ "learning_rate": 0.0006,
+ "loss": 4.073130130767822,
+ "step": 2646
+ },
+ {
+ "epoch": 36.76855895196506,
+ "grad_norm": 0.013827716931700706,
+ "learning_rate": 0.0006,
+ "loss": 4.043869972229004,
+ "step": 2647
+ },
+ {
+ "epoch": 36.7825327510917,
+ "grad_norm": 0.015548762865364552,
+ "learning_rate": 0.0006,
+ "loss": 4.134049415588379,
+ "step": 2648
+ },
+ {
+ "epoch": 36.79650655021834,
+ "grad_norm": 0.017041970044374466,
+ "learning_rate": 0.0006,
+ "loss": 4.205024242401123,
+ "step": 2649
+ },
+ {
+ "epoch": 36.81048034934498,
+ "grad_norm": 0.017638275399804115,
+ "learning_rate": 0.0006,
+ "loss": 4.199924468994141,
+ "step": 2650
+ },
+ {
+ "epoch": 36.82445414847162,
+ "grad_norm": 0.017771627753973007,
+ "learning_rate": 0.0006,
+ "loss": 4.15706729888916,
+ "step": 2651
+ },
+ {
+ "epoch": 36.838427947598255,
+ "grad_norm": 0.015933562070131302,
+ "learning_rate": 0.0006,
+ "loss": 4.220742225646973,
+ "step": 2652
+ },
+ {
+ "epoch": 36.852401746724894,
+ "grad_norm": 0.01668144389986992,
+ "learning_rate": 0.0006,
+ "loss": 4.161426544189453,
+ "step": 2653
+ },
+ {
+ "epoch": 36.866375545851525,
+ "grad_norm": 0.018014680594205856,
+ "learning_rate": 0.0006,
+ "loss": 4.183046817779541,
+ "step": 2654
+ },
+ {
+ "epoch": 36.880349344978164,
+ "grad_norm": 0.018808670341968536,
+ "learning_rate": 0.0006,
+ "loss": 4.105293273925781,
+ "step": 2655
+ },
+ {
+ "epoch": 36.8943231441048,
+ "grad_norm": 0.018003476783633232,
+ "learning_rate": 0.0006,
+ "loss": 4.103171348571777,
+ "step": 2656
+ },
+ {
+ "epoch": 36.90829694323144,
+ "grad_norm": 0.016865914687514305,
+ "learning_rate": 0.0006,
+ "loss": 4.125634670257568,
+ "step": 2657
+ },
+ {
+ "epoch": 36.92227074235808,
+ "grad_norm": 0.018371153622865677,
+ "learning_rate": 0.0006,
+ "loss": 4.192461013793945,
+ "step": 2658
+ },
+ {
+ "epoch": 36.93624454148472,
+ "grad_norm": 0.019045377150177956,
+ "learning_rate": 0.0006,
+ "loss": 4.198105812072754,
+ "step": 2659
+ },
+ {
+ "epoch": 36.95021834061135,
+ "grad_norm": 0.01594168320298195,
+ "learning_rate": 0.0006,
+ "loss": 4.169788360595703,
+ "step": 2660
+ },
+ {
+ "epoch": 36.96419213973799,
+ "grad_norm": 0.015645820647478104,
+ "learning_rate": 0.0006,
+ "loss": 4.047277450561523,
+ "step": 2661
+ },
+ {
+ "epoch": 36.97816593886463,
+ "grad_norm": 0.016313156113028526,
+ "learning_rate": 0.0006,
+ "loss": 4.087392807006836,
+ "step": 2662
+ },
+ {
+ "epoch": 36.992139737991266,
+ "grad_norm": 0.017050625756382942,
+ "learning_rate": 0.0006,
+ "loss": 4.0908002853393555,
+ "step": 2663
+ },
+ {
+ "epoch": 37.0,
+ "grad_norm": 0.017730968073010445,
+ "learning_rate": 0.0006,
+ "loss": 4.110957622528076,
+ "step": 2664
+ },
+ {
+ "epoch": 37.0,
+ "eval_loss": 4.563793182373047,
+ "eval_runtime": 56.5922,
+ "eval_samples_per_second": 43.151,
+ "eval_steps_per_second": 1.361,
+ "step": 2664
+ },
+ {
+ "epoch": 37.01397379912664,
+ "grad_norm": 0.015687910839915276,
+ "learning_rate": 0.0006,
+ "loss": 4.143589019775391,
+ "step": 2665
+ },
+ {
+ "epoch": 37.02794759825328,
+ "grad_norm": 0.017451366409659386,
+ "learning_rate": 0.0006,
+ "loss": 4.0326666831970215,
+ "step": 2666
+ },
+ {
+ "epoch": 37.041921397379916,
+ "grad_norm": 0.01909538544714451,
+ "learning_rate": 0.0006,
+ "loss": 3.9706249237060547,
+ "step": 2667
+ },
+ {
+ "epoch": 37.05589519650655,
+ "grad_norm": 0.01988379657268524,
+ "learning_rate": 0.0006,
+ "loss": 3.9531009197235107,
+ "step": 2668
+ },
+ {
+ "epoch": 37.069868995633186,
+ "grad_norm": 0.02035115472972393,
+ "learning_rate": 0.0006,
+ "loss": 3.962214469909668,
+ "step": 2669
+ },
+ {
+ "epoch": 37.083842794759825,
+ "grad_norm": 0.01952001079916954,
+ "learning_rate": 0.0006,
+ "loss": 4.182262897491455,
+ "step": 2670
+ },
+ {
+ "epoch": 37.09781659388646,
+ "grad_norm": 0.01854325644671917,
+ "learning_rate": 0.0006,
+ "loss": 4.127663612365723,
+ "step": 2671
+ },
+ {
+ "epoch": 37.1117903930131,
+ "grad_norm": 0.017757616937160492,
+ "learning_rate": 0.0006,
+ "loss": 4.100804805755615,
+ "step": 2672
+ },
+ {
+ "epoch": 37.12576419213974,
+ "grad_norm": 0.016712479293346405,
+ "learning_rate": 0.0006,
+ "loss": 4.026284217834473,
+ "step": 2673
+ },
+ {
+ "epoch": 37.13973799126637,
+ "grad_norm": 0.0168139711022377,
+ "learning_rate": 0.0006,
+ "loss": 4.183882713317871,
+ "step": 2674
+ },
+ {
+ "epoch": 37.15371179039301,
+ "grad_norm": 0.017060423269867897,
+ "learning_rate": 0.0006,
+ "loss": 4.131832599639893,
+ "step": 2675
+ },
+ {
+ "epoch": 37.16768558951965,
+ "grad_norm": 0.01726631447672844,
+ "learning_rate": 0.0006,
+ "loss": 3.9156057834625244,
+ "step": 2676
+ },
+ {
+ "epoch": 37.18165938864629,
+ "grad_norm": 0.016348622739315033,
+ "learning_rate": 0.0006,
+ "loss": 4.1731367111206055,
+ "step": 2677
+ },
+ {
+ "epoch": 37.19563318777293,
+ "grad_norm": 0.016640059649944305,
+ "learning_rate": 0.0006,
+ "loss": 4.10997200012207,
+ "step": 2678
+ },
+ {
+ "epoch": 37.209606986899566,
+ "grad_norm": 0.01601342298090458,
+ "learning_rate": 0.0006,
+ "loss": 4.081145286560059,
+ "step": 2679
+ },
+ {
+ "epoch": 37.223580786026204,
+ "grad_norm": 0.01735631190240383,
+ "learning_rate": 0.0006,
+ "loss": 4.202108860015869,
+ "step": 2680
+ },
+ {
+ "epoch": 37.237554585152836,
+ "grad_norm": 0.019164685159921646,
+ "learning_rate": 0.0006,
+ "loss": 4.104109764099121,
+ "step": 2681
+ },
+ {
+ "epoch": 37.251528384279474,
+ "grad_norm": 0.01881759613752365,
+ "learning_rate": 0.0006,
+ "loss": 4.145532131195068,
+ "step": 2682
+ },
+ {
+ "epoch": 37.26550218340611,
+ "grad_norm": 0.017656555399298668,
+ "learning_rate": 0.0006,
+ "loss": 4.012782096862793,
+ "step": 2683
+ },
+ {
+ "epoch": 37.27947598253275,
+ "grad_norm": 0.01624823734164238,
+ "learning_rate": 0.0006,
+ "loss": 4.001727104187012,
+ "step": 2684
+ },
+ {
+ "epoch": 37.29344978165939,
+ "grad_norm": 0.016973139718174934,
+ "learning_rate": 0.0006,
+ "loss": 4.095399856567383,
+ "step": 2685
+ },
+ {
+ "epoch": 37.30742358078603,
+ "grad_norm": 0.015503671951591969,
+ "learning_rate": 0.0006,
+ "loss": 4.034431457519531,
+ "step": 2686
+ },
+ {
+ "epoch": 37.32139737991266,
+ "grad_norm": 0.015370250679552555,
+ "learning_rate": 0.0006,
+ "loss": 4.075836181640625,
+ "step": 2687
+ },
+ {
+ "epoch": 37.3353711790393,
+ "grad_norm": 0.01564021222293377,
+ "learning_rate": 0.0006,
+ "loss": 4.125202178955078,
+ "step": 2688
+ },
+ {
+ "epoch": 37.34934497816594,
+ "grad_norm": 0.016717975959181786,
+ "learning_rate": 0.0006,
+ "loss": 4.153017044067383,
+ "step": 2689
+ },
+ {
+ "epoch": 37.36331877729258,
+ "grad_norm": 0.015263617038726807,
+ "learning_rate": 0.0006,
+ "loss": 4.021814823150635,
+ "step": 2690
+ },
+ {
+ "epoch": 37.377292576419215,
+ "grad_norm": 0.016497142612934113,
+ "learning_rate": 0.0006,
+ "loss": 4.069095611572266,
+ "step": 2691
+ },
+ {
+ "epoch": 37.391266375545854,
+ "grad_norm": 0.01669355481863022,
+ "learning_rate": 0.0006,
+ "loss": 4.174587726593018,
+ "step": 2692
+ },
+ {
+ "epoch": 37.40524017467249,
+ "grad_norm": 0.019695773720741272,
+ "learning_rate": 0.0006,
+ "loss": 4.0073041915893555,
+ "step": 2693
+ },
+ {
+ "epoch": 37.419213973799124,
+ "grad_norm": 0.023720644414424896,
+ "learning_rate": 0.0006,
+ "loss": 4.111948013305664,
+ "step": 2694
+ },
+ {
+ "epoch": 37.43318777292576,
+ "grad_norm": 0.02399350143969059,
+ "learning_rate": 0.0006,
+ "loss": 4.081164360046387,
+ "step": 2695
+ },
+ {
+ "epoch": 37.4471615720524,
+ "grad_norm": 0.019630271941423416,
+ "learning_rate": 0.0006,
+ "loss": 4.065731048583984,
+ "step": 2696
+ },
+ {
+ "epoch": 37.46113537117904,
+ "grad_norm": 0.019679008051753044,
+ "learning_rate": 0.0006,
+ "loss": 3.993852376937866,
+ "step": 2697
+ },
+ {
+ "epoch": 37.47510917030568,
+ "grad_norm": 0.020597299560904503,
+ "learning_rate": 0.0006,
+ "loss": 4.275555610656738,
+ "step": 2698
+ },
+ {
+ "epoch": 37.48908296943232,
+ "grad_norm": 0.01983889564871788,
+ "learning_rate": 0.0006,
+ "loss": 4.254708290100098,
+ "step": 2699
+ },
+ {
+ "epoch": 37.50305676855895,
+ "grad_norm": 0.019150668755173683,
+ "learning_rate": 0.0006,
+ "loss": 4.097885608673096,
+ "step": 2700
+ },
+ {
+ "epoch": 37.51703056768559,
+ "grad_norm": 0.018847715109586716,
+ "learning_rate": 0.0006,
+ "loss": 4.054081439971924,
+ "step": 2701
+ },
+ {
+ "epoch": 37.531004366812226,
+ "grad_norm": 0.019607288762927055,
+ "learning_rate": 0.0006,
+ "loss": 3.9891440868377686,
+ "step": 2702
+ },
+ {
+ "epoch": 37.544978165938865,
+ "grad_norm": 0.018863150849938393,
+ "learning_rate": 0.0006,
+ "loss": 4.084877014160156,
+ "step": 2703
+ },
+ {
+ "epoch": 37.5589519650655,
+ "grad_norm": 0.017340239137411118,
+ "learning_rate": 0.0006,
+ "loss": 4.113290786743164,
+ "step": 2704
+ },
+ {
+ "epoch": 37.57292576419214,
+ "grad_norm": 0.015628693625330925,
+ "learning_rate": 0.0006,
+ "loss": 4.1483001708984375,
+ "step": 2705
+ },
+ {
+ "epoch": 37.58689956331878,
+ "grad_norm": 0.016493283212184906,
+ "learning_rate": 0.0006,
+ "loss": 4.1040143966674805,
+ "step": 2706
+ },
+ {
+ "epoch": 37.60087336244541,
+ "grad_norm": 0.016732443124055862,
+ "learning_rate": 0.0006,
+ "loss": 4.07205057144165,
+ "step": 2707
+ },
+ {
+ "epoch": 37.61484716157205,
+ "grad_norm": 0.017176156863570213,
+ "learning_rate": 0.0006,
+ "loss": 4.1279754638671875,
+ "step": 2708
+ },
+ {
+ "epoch": 37.62882096069869,
+ "grad_norm": 0.015730608254671097,
+ "learning_rate": 0.0006,
+ "loss": 4.086658000946045,
+ "step": 2709
+ },
+ {
+ "epoch": 37.64279475982533,
+ "grad_norm": 0.016111090779304504,
+ "learning_rate": 0.0006,
+ "loss": 4.064335823059082,
+ "step": 2710
+ },
+ {
+ "epoch": 37.65676855895197,
+ "grad_norm": 0.01769379712641239,
+ "learning_rate": 0.0006,
+ "loss": 4.116047382354736,
+ "step": 2711
+ },
+ {
+ "epoch": 37.670742358078606,
+ "grad_norm": 0.01683102548122406,
+ "learning_rate": 0.0006,
+ "loss": 4.095986843109131,
+ "step": 2712
+ },
+ {
+ "epoch": 37.68471615720524,
+ "grad_norm": 0.015726450830698013,
+ "learning_rate": 0.0006,
+ "loss": 4.10775089263916,
+ "step": 2713
+ },
+ {
+ "epoch": 37.698689956331876,
+ "grad_norm": 0.014760235324501991,
+ "learning_rate": 0.0006,
+ "loss": 4.1234893798828125,
+ "step": 2714
+ },
+ {
+ "epoch": 37.712663755458514,
+ "grad_norm": 0.014933438040316105,
+ "learning_rate": 0.0006,
+ "loss": 4.079200744628906,
+ "step": 2715
+ },
+ {
+ "epoch": 37.72663755458515,
+ "grad_norm": 0.015723997727036476,
+ "learning_rate": 0.0006,
+ "loss": 4.07865047454834,
+ "step": 2716
+ },
+ {
+ "epoch": 37.74061135371179,
+ "grad_norm": 0.014539001509547234,
+ "learning_rate": 0.0006,
+ "loss": 4.036242485046387,
+ "step": 2717
+ },
+ {
+ "epoch": 37.75458515283843,
+ "grad_norm": 0.014844979159533978,
+ "learning_rate": 0.0006,
+ "loss": 4.1051483154296875,
+ "step": 2718
+ },
+ {
+ "epoch": 37.76855895196506,
+ "grad_norm": 0.015939844772219658,
+ "learning_rate": 0.0006,
+ "loss": 4.017253875732422,
+ "step": 2719
+ },
+ {
+ "epoch": 37.7825327510917,
+ "grad_norm": 0.014771555550396442,
+ "learning_rate": 0.0006,
+ "loss": 4.0694661140441895,
+ "step": 2720
+ },
+ {
+ "epoch": 37.79650655021834,
+ "grad_norm": 0.015777921304106712,
+ "learning_rate": 0.0006,
+ "loss": 3.947587490081787,
+ "step": 2721
+ },
+ {
+ "epoch": 37.81048034934498,
+ "grad_norm": 0.017451291903853416,
+ "learning_rate": 0.0006,
+ "loss": 4.093625068664551,
+ "step": 2722
+ },
+ {
+ "epoch": 37.82445414847162,
+ "grad_norm": 0.016829246655106544,
+ "learning_rate": 0.0006,
+ "loss": 4.09030818939209,
+ "step": 2723
+ },
+ {
+ "epoch": 37.838427947598255,
+ "grad_norm": 0.016984397545456886,
+ "learning_rate": 0.0006,
+ "loss": 4.140626907348633,
+ "step": 2724
+ },
+ {
+ "epoch": 37.852401746724894,
+ "grad_norm": 0.01732603646814823,
+ "learning_rate": 0.0006,
+ "loss": 4.128276824951172,
+ "step": 2725
+ },
+ {
+ "epoch": 37.866375545851525,
+ "grad_norm": 0.01744619384407997,
+ "learning_rate": 0.0006,
+ "loss": 4.103130340576172,
+ "step": 2726
+ },
+ {
+ "epoch": 37.880349344978164,
+ "grad_norm": 0.016331320628523827,
+ "learning_rate": 0.0006,
+ "loss": 4.037951469421387,
+ "step": 2727
+ },
+ {
+ "epoch": 37.8943231441048,
+ "grad_norm": 0.015919430181384087,
+ "learning_rate": 0.0006,
+ "loss": 4.121513366699219,
+ "step": 2728
+ },
+ {
+ "epoch": 37.90829694323144,
+ "grad_norm": 0.016153695061802864,
+ "learning_rate": 0.0006,
+ "loss": 4.099384307861328,
+ "step": 2729
+ },
+ {
+ "epoch": 37.92227074235808,
+ "grad_norm": 0.01628800667822361,
+ "learning_rate": 0.0006,
+ "loss": 4.092325687408447,
+ "step": 2730
+ },
+ {
+ "epoch": 37.93624454148472,
+ "grad_norm": 0.017285684123635292,
+ "learning_rate": 0.0006,
+ "loss": 3.9828906059265137,
+ "step": 2731
+ },
+ {
+ "epoch": 37.95021834061135,
+ "grad_norm": 0.016047554090619087,
+ "learning_rate": 0.0006,
+ "loss": 4.100419044494629,
+ "step": 2732
+ },
+ {
+ "epoch": 37.96419213973799,
+ "grad_norm": 0.01562986709177494,
+ "learning_rate": 0.0006,
+ "loss": 4.065832614898682,
+ "step": 2733
+ },
+ {
+ "epoch": 37.97816593886463,
+ "grad_norm": 0.01677163504064083,
+ "learning_rate": 0.0006,
+ "loss": 4.077518463134766,
+ "step": 2734
+ },
+ {
+ "epoch": 37.992139737991266,
+ "grad_norm": 0.01755492016673088,
+ "learning_rate": 0.0006,
+ "loss": 4.1076154708862305,
+ "step": 2735
+ },
+ {
+ "epoch": 38.0,
+ "grad_norm": 0.01929762214422226,
+ "learning_rate": 0.0006,
+ "loss": 4.09603214263916,
+ "step": 2736
+ },
+ {
+ "epoch": 38.0,
+ "eval_loss": 4.582442283630371,
+ "eval_runtime": 62.8649,
+ "eval_samples_per_second": 38.845,
+ "eval_steps_per_second": 1.225,
+ "step": 2736
+ },
+ {
+ "epoch": 38.01397379912664,
+ "grad_norm": 0.019129136577248573,
+ "learning_rate": 0.0006,
+ "loss": 4.095005989074707,
+ "step": 2737
+ },
+ {
+ "epoch": 38.02794759825328,
+ "grad_norm": 0.019167408347129822,
+ "learning_rate": 0.0006,
+ "loss": 4.054426670074463,
+ "step": 2738
+ },
+ {
+ "epoch": 38.041921397379916,
+ "grad_norm": 0.01789460889995098,
+ "learning_rate": 0.0006,
+ "loss": 4.0434417724609375,
+ "step": 2739
+ },
+ {
+ "epoch": 38.05589519650655,
+ "grad_norm": 0.015947027131915092,
+ "learning_rate": 0.0006,
+ "loss": 4.074516296386719,
+ "step": 2740
+ },
+ {
+ "epoch": 38.069868995633186,
+ "grad_norm": 0.016667252406477928,
+ "learning_rate": 0.0006,
+ "loss": 4.108545303344727,
+ "step": 2741
+ },
+ {
+ "epoch": 38.083842794759825,
+ "grad_norm": 0.016014399006962776,
+ "learning_rate": 0.0006,
+ "loss": 4.026925086975098,
+ "step": 2742
+ },
+ {
+ "epoch": 38.09781659388646,
+ "grad_norm": 0.015397695824503899,
+ "learning_rate": 0.0006,
+ "loss": 4.163970470428467,
+ "step": 2743
+ },
+ {
+ "epoch": 38.1117903930131,
+ "grad_norm": 0.01563289947807789,
+ "learning_rate": 0.0006,
+ "loss": 4.116578578948975,
+ "step": 2744
+ },
+ {
+ "epoch": 38.12576419213974,
+ "grad_norm": 0.015512318350374699,
+ "learning_rate": 0.0006,
+ "loss": 4.096311569213867,
+ "step": 2745
+ },
+ {
+ "epoch": 38.13973799126637,
+ "grad_norm": 0.015426642261445522,
+ "learning_rate": 0.0006,
+ "loss": 4.017065048217773,
+ "step": 2746
+ },
+ {
+ "epoch": 38.15371179039301,
+ "grad_norm": 0.015115435235202312,
+ "learning_rate": 0.0006,
+ "loss": 4.005434989929199,
+ "step": 2747
+ },
+ {
+ "epoch": 38.16768558951965,
+ "grad_norm": 0.015235894359648228,
+ "learning_rate": 0.0006,
+ "loss": 3.982743263244629,
+ "step": 2748
+ },
+ {
+ "epoch": 38.18165938864629,
+ "grad_norm": 0.016095101833343506,
+ "learning_rate": 0.0006,
+ "loss": 4.0849385261535645,
+ "step": 2749
+ },
+ {
+ "epoch": 38.19563318777293,
+ "grad_norm": 0.015201197937130928,
+ "learning_rate": 0.0006,
+ "loss": 4.103863716125488,
+ "step": 2750
+ },
+ {
+ "epoch": 38.209606986899566,
+ "grad_norm": 0.015687478706240654,
+ "learning_rate": 0.0006,
+ "loss": 4.12074089050293,
+ "step": 2751
+ },
+ {
+ "epoch": 38.223580786026204,
+ "grad_norm": 0.015201589092612267,
+ "learning_rate": 0.0006,
+ "loss": 4.056872367858887,
+ "step": 2752
+ },
+ {
+ "epoch": 38.237554585152836,
+ "grad_norm": 0.015154894441366196,
+ "learning_rate": 0.0006,
+ "loss": 4.089670181274414,
+ "step": 2753
+ },
+ {
+ "epoch": 38.251528384279474,
+ "grad_norm": 0.01586608588695526,
+ "learning_rate": 0.0006,
+ "loss": 4.155317306518555,
+ "step": 2754
+ },
+ {
+ "epoch": 38.26550218340611,
+ "grad_norm": 0.01609654724597931,
+ "learning_rate": 0.0006,
+ "loss": 3.995800495147705,
+ "step": 2755
+ },
+ {
+ "epoch": 38.27947598253275,
+ "grad_norm": 0.015495185740292072,
+ "learning_rate": 0.0006,
+ "loss": 3.995769739151001,
+ "step": 2756
+ },
+ {
+ "epoch": 38.29344978165939,
+ "grad_norm": 0.017042381688952446,
+ "learning_rate": 0.0006,
+ "loss": 4.132832050323486,
+ "step": 2757
+ },
+ {
+ "epoch": 38.30742358078603,
+ "grad_norm": 0.018706614151597023,
+ "learning_rate": 0.0006,
+ "loss": 4.038820743560791,
+ "step": 2758
+ },
+ {
+ "epoch": 38.32139737991266,
+ "grad_norm": 0.018316026777029037,
+ "learning_rate": 0.0006,
+ "loss": 3.972052574157715,
+ "step": 2759
+ },
+ {
+ "epoch": 38.3353711790393,
+ "grad_norm": 0.018375389277935028,
+ "learning_rate": 0.0006,
+ "loss": 4.152097702026367,
+ "step": 2760
+ },
+ {
+ "epoch": 38.34934497816594,
+ "grad_norm": 0.016879886388778687,
+ "learning_rate": 0.0006,
+ "loss": 4.020869731903076,
+ "step": 2761
+ },
+ {
+ "epoch": 38.36331877729258,
+ "grad_norm": 0.015728183090686798,
+ "learning_rate": 0.0006,
+ "loss": 4.008016586303711,
+ "step": 2762
+ },
+ {
+ "epoch": 38.377292576419215,
+ "grad_norm": 0.016147758811712265,
+ "learning_rate": 0.0006,
+ "loss": 4.035854339599609,
+ "step": 2763
+ },
+ {
+ "epoch": 38.391266375545854,
+ "grad_norm": 0.017774797976017,
+ "learning_rate": 0.0006,
+ "loss": 4.122506141662598,
+ "step": 2764
+ },
+ {
+ "epoch": 38.40524017467249,
+ "grad_norm": 0.016590069979429245,
+ "learning_rate": 0.0006,
+ "loss": 4.082847595214844,
+ "step": 2765
+ },
+ {
+ "epoch": 38.419213973799124,
+ "grad_norm": 0.016544528305530548,
+ "learning_rate": 0.0006,
+ "loss": 4.184270858764648,
+ "step": 2766
+ },
+ {
+ "epoch": 38.43318777292576,
+ "grad_norm": 0.017972925677895546,
+ "learning_rate": 0.0006,
+ "loss": 3.987558603286743,
+ "step": 2767
+ },
+ {
+ "epoch": 38.4471615720524,
+ "grad_norm": 0.018562227487564087,
+ "learning_rate": 0.0006,
+ "loss": 4.1193461418151855,
+ "step": 2768
+ },
+ {
+ "epoch": 38.46113537117904,
+ "grad_norm": 0.018896685913205147,
+ "learning_rate": 0.0006,
+ "loss": 4.057808876037598,
+ "step": 2769
+ },
+ {
+ "epoch": 38.47510917030568,
+ "grad_norm": 0.016793809831142426,
+ "learning_rate": 0.0006,
+ "loss": 4.069523334503174,
+ "step": 2770
+ },
+ {
+ "epoch": 38.48908296943232,
+ "grad_norm": 0.01698676310479641,
+ "learning_rate": 0.0006,
+ "loss": 3.997624158859253,
+ "step": 2771
+ },
+ {
+ "epoch": 38.50305676855895,
+ "grad_norm": 0.016784202307462692,
+ "learning_rate": 0.0006,
+ "loss": 4.174795627593994,
+ "step": 2772
+ },
+ {
+ "epoch": 38.51703056768559,
+ "grad_norm": 0.016785014420747757,
+ "learning_rate": 0.0006,
+ "loss": 3.999321222305298,
+ "step": 2773
+ },
+ {
+ "epoch": 38.531004366812226,
+ "grad_norm": 0.018615184351801872,
+ "learning_rate": 0.0006,
+ "loss": 4.191237926483154,
+ "step": 2774
+ },
+ {
+ "epoch": 38.544978165938865,
+ "grad_norm": 0.019090790301561356,
+ "learning_rate": 0.0006,
+ "loss": 4.106966018676758,
+ "step": 2775
+ },
+ {
+ "epoch": 38.5589519650655,
+ "grad_norm": 0.018298320472240448,
+ "learning_rate": 0.0006,
+ "loss": 4.063363552093506,
+ "step": 2776
+ },
+ {
+ "epoch": 38.57292576419214,
+ "grad_norm": 0.018541090190410614,
+ "learning_rate": 0.0006,
+ "loss": 4.101193428039551,
+ "step": 2777
+ },
+ {
+ "epoch": 38.58689956331878,
+ "grad_norm": 0.01849970407783985,
+ "learning_rate": 0.0006,
+ "loss": 4.071335792541504,
+ "step": 2778
+ },
+ {
+ "epoch": 38.60087336244541,
+ "grad_norm": 0.018309077247977257,
+ "learning_rate": 0.0006,
+ "loss": 4.245305061340332,
+ "step": 2779
+ },
+ {
+ "epoch": 38.61484716157205,
+ "grad_norm": 0.017681840807199478,
+ "learning_rate": 0.0006,
+ "loss": 4.081173419952393,
+ "step": 2780
+ },
+ {
+ "epoch": 38.62882096069869,
+ "grad_norm": 0.016286224126815796,
+ "learning_rate": 0.0006,
+ "loss": 4.024351596832275,
+ "step": 2781
+ },
+ {
+ "epoch": 38.64279475982533,
+ "grad_norm": 0.015837235376238823,
+ "learning_rate": 0.0006,
+ "loss": 4.028205394744873,
+ "step": 2782
+ },
+ {
+ "epoch": 38.65676855895197,
+ "grad_norm": 0.016062596812844276,
+ "learning_rate": 0.0006,
+ "loss": 4.131122589111328,
+ "step": 2783
+ },
+ {
+ "epoch": 38.670742358078606,
+ "grad_norm": 0.01711239665746689,
+ "learning_rate": 0.0006,
+ "loss": 4.124416351318359,
+ "step": 2784
+ },
+ {
+ "epoch": 38.68471615720524,
+ "grad_norm": 0.015304913744330406,
+ "learning_rate": 0.0006,
+ "loss": 4.056122779846191,
+ "step": 2785
+ },
+ {
+ "epoch": 38.698689956331876,
+ "grad_norm": 0.015971887856721878,
+ "learning_rate": 0.0006,
+ "loss": 4.132555961608887,
+ "step": 2786
+ },
+ {
+ "epoch": 38.712663755458514,
+ "grad_norm": 0.015787221491336823,
+ "learning_rate": 0.0006,
+ "loss": 4.051596164703369,
+ "step": 2787
+ },
+ {
+ "epoch": 38.72663755458515,
+ "grad_norm": 0.01605444774031639,
+ "learning_rate": 0.0006,
+ "loss": 4.191808700561523,
+ "step": 2788
+ },
+ {
+ "epoch": 38.74061135371179,
+ "grad_norm": 0.016383009031414986,
+ "learning_rate": 0.0006,
+ "loss": 4.0648417472839355,
+ "step": 2789
+ },
+ {
+ "epoch": 38.75458515283843,
+ "grad_norm": 0.016135554760694504,
+ "learning_rate": 0.0006,
+ "loss": 3.988563060760498,
+ "step": 2790
+ },
+ {
+ "epoch": 38.76855895196506,
+ "grad_norm": 0.016634158790111542,
+ "learning_rate": 0.0006,
+ "loss": 4.133493423461914,
+ "step": 2791
+ },
+ {
+ "epoch": 38.7825327510917,
+ "grad_norm": 0.01688210666179657,
+ "learning_rate": 0.0006,
+ "loss": 4.091277122497559,
+ "step": 2792
+ },
+ {
+ "epoch": 38.79650655021834,
+ "grad_norm": 0.016102122142910957,
+ "learning_rate": 0.0006,
+ "loss": 4.051219940185547,
+ "step": 2793
+ },
+ {
+ "epoch": 38.81048034934498,
+ "grad_norm": 0.016585038974881172,
+ "learning_rate": 0.0006,
+ "loss": 4.058587074279785,
+ "step": 2794
+ },
+ {
+ "epoch": 38.82445414847162,
+ "grad_norm": 0.017700951546430588,
+ "learning_rate": 0.0006,
+ "loss": 4.016030311584473,
+ "step": 2795
+ },
+ {
+ "epoch": 38.838427947598255,
+ "grad_norm": 0.017922203987836838,
+ "learning_rate": 0.0006,
+ "loss": 4.079977989196777,
+ "step": 2796
+ },
+ {
+ "epoch": 38.852401746724894,
+ "grad_norm": 0.017299890518188477,
+ "learning_rate": 0.0006,
+ "loss": 4.062984943389893,
+ "step": 2797
+ },
+ {
+ "epoch": 38.866375545851525,
+ "grad_norm": 0.018648581579327583,
+ "learning_rate": 0.0006,
+ "loss": 4.065717697143555,
+ "step": 2798
+ },
+ {
+ "epoch": 38.880349344978164,
+ "grad_norm": 0.019057920202612877,
+ "learning_rate": 0.0006,
+ "loss": 4.227892875671387,
+ "step": 2799
+ },
+ {
+ "epoch": 38.8943231441048,
+ "grad_norm": 0.018824422731995583,
+ "learning_rate": 0.0006,
+ "loss": 4.116058349609375,
+ "step": 2800
+ },
+ {
+ "epoch": 38.90829694323144,
+ "grad_norm": 0.019084978848695755,
+ "learning_rate": 0.0006,
+ "loss": 4.061677932739258,
+ "step": 2801
+ },
+ {
+ "epoch": 38.92227074235808,
+ "grad_norm": 0.017842981964349747,
+ "learning_rate": 0.0006,
+ "loss": 4.083813667297363,
+ "step": 2802
+ },
+ {
+ "epoch": 38.93624454148472,
+ "grad_norm": 0.01880858838558197,
+ "learning_rate": 0.0006,
+ "loss": 4.152561187744141,
+ "step": 2803
+ },
+ {
+ "epoch": 38.95021834061135,
+ "grad_norm": 0.017400087788701057,
+ "learning_rate": 0.0006,
+ "loss": 4.154635429382324,
+ "step": 2804
+ },
+ {
+ "epoch": 38.96419213973799,
+ "grad_norm": 0.017682507634162903,
+ "learning_rate": 0.0006,
+ "loss": 4.151128768920898,
+ "step": 2805
+ },
+ {
+ "epoch": 38.97816593886463,
+ "grad_norm": 0.015803052112460136,
+ "learning_rate": 0.0006,
+ "loss": 4.10498571395874,
+ "step": 2806
+ },
+ {
+ "epoch": 38.992139737991266,
+ "grad_norm": 0.01582511141896248,
+ "learning_rate": 0.0006,
+ "loss": 4.140159606933594,
+ "step": 2807
+ },
+ {
+ "epoch": 39.0,
+ "grad_norm": 0.01925666816532612,
+ "learning_rate": 0.0006,
+ "loss": 4.126394271850586,
+ "step": 2808
+ },
+ {
+ "epoch": 39.0,
+ "eval_loss": 4.615617275238037,
+ "eval_runtime": 56.8895,
+ "eval_samples_per_second": 42.925,
+ "eval_steps_per_second": 1.354,
+ "step": 2808
+ },
+ {
+ "epoch": 39.01397379912664,
+ "grad_norm": 0.01677853614091873,
+ "learning_rate": 0.0006,
+ "loss": 4.102395057678223,
+ "step": 2809
+ },
+ {
+ "epoch": 39.02794759825328,
+ "grad_norm": 0.017901137471199036,
+ "learning_rate": 0.0006,
+ "loss": 3.976609230041504,
+ "step": 2810
+ },
+ {
+ "epoch": 39.041921397379916,
+ "grad_norm": 0.017978668212890625,
+ "learning_rate": 0.0006,
+ "loss": 3.944416046142578,
+ "step": 2811
+ },
+ {
+ "epoch": 39.05589519650655,
+ "grad_norm": 0.019967492669820786,
+ "learning_rate": 0.0006,
+ "loss": 3.992234468460083,
+ "step": 2812
+ },
+ {
+ "epoch": 39.069868995633186,
+ "grad_norm": 0.018937544897198677,
+ "learning_rate": 0.0006,
+ "loss": 4.028573036193848,
+ "step": 2813
+ },
+ {
+ "epoch": 39.083842794759825,
+ "grad_norm": 0.016852952539920807,
+ "learning_rate": 0.0006,
+ "loss": 4.062442779541016,
+ "step": 2814
+ },
+ {
+ "epoch": 39.09781659388646,
+ "grad_norm": 0.016643647104501724,
+ "learning_rate": 0.0006,
+ "loss": 4.08871603012085,
+ "step": 2815
+ },
+ {
+ "epoch": 39.1117903930131,
+ "grad_norm": 0.016872866079211235,
+ "learning_rate": 0.0006,
+ "loss": 3.9514245986938477,
+ "step": 2816
+ },
+ {
+ "epoch": 39.12576419213974,
+ "grad_norm": 0.018045669421553612,
+ "learning_rate": 0.0006,
+ "loss": 4.038842678070068,
+ "step": 2817
+ },
+ {
+ "epoch": 39.13973799126637,
+ "grad_norm": 0.017853403463959694,
+ "learning_rate": 0.0006,
+ "loss": 3.93709135055542,
+ "step": 2818
+ },
+ {
+ "epoch": 39.15371179039301,
+ "grad_norm": 0.017507443204522133,
+ "learning_rate": 0.0006,
+ "loss": 3.98665714263916,
+ "step": 2819
+ },
+ {
+ "epoch": 39.16768558951965,
+ "grad_norm": 0.01766853965818882,
+ "learning_rate": 0.0006,
+ "loss": 3.937126636505127,
+ "step": 2820
+ },
+ {
+ "epoch": 39.18165938864629,
+ "grad_norm": 0.016660021618008614,
+ "learning_rate": 0.0006,
+ "loss": 3.8397014141082764,
+ "step": 2821
+ },
+ {
+ "epoch": 39.19563318777293,
+ "grad_norm": 0.017719309777021408,
+ "learning_rate": 0.0006,
+ "loss": 4.003579616546631,
+ "step": 2822
+ },
+ {
+ "epoch": 39.209606986899566,
+ "grad_norm": 0.018234414979815483,
+ "learning_rate": 0.0006,
+ "loss": 4.053859710693359,
+ "step": 2823
+ },
+ {
+ "epoch": 39.223580786026204,
+ "grad_norm": 0.018777024000883102,
+ "learning_rate": 0.0006,
+ "loss": 4.056182861328125,
+ "step": 2824
+ },
+ {
+ "epoch": 39.237554585152836,
+ "grad_norm": 0.017443502321839333,
+ "learning_rate": 0.0006,
+ "loss": 4.025068283081055,
+ "step": 2825
+ },
+ {
+ "epoch": 39.251528384279474,
+ "grad_norm": 0.016500331461429596,
+ "learning_rate": 0.0006,
+ "loss": 4.101631164550781,
+ "step": 2826
+ },
+ {
+ "epoch": 39.26550218340611,
+ "grad_norm": 0.016101544722914696,
+ "learning_rate": 0.0006,
+ "loss": 4.027082443237305,
+ "step": 2827
+ },
+ {
+ "epoch": 39.27947598253275,
+ "grad_norm": 0.015471171587705612,
+ "learning_rate": 0.0006,
+ "loss": 4.022909641265869,
+ "step": 2828
+ },
+ {
+ "epoch": 39.29344978165939,
+ "grad_norm": 0.015779748558998108,
+ "learning_rate": 0.0006,
+ "loss": 4.18796968460083,
+ "step": 2829
+ },
+ {
+ "epoch": 39.30742358078603,
+ "grad_norm": 0.015522679314017296,
+ "learning_rate": 0.0006,
+ "loss": 4.005777359008789,
+ "step": 2830
+ },
+ {
+ "epoch": 39.32139737991266,
+ "grad_norm": 0.015226504765450954,
+ "learning_rate": 0.0006,
+ "loss": 4.108416557312012,
+ "step": 2831
+ },
+ {
+ "epoch": 39.3353711790393,
+ "grad_norm": 0.015095778740942478,
+ "learning_rate": 0.0006,
+ "loss": 4.08029842376709,
+ "step": 2832
+ },
+ {
+ "epoch": 39.34934497816594,
+ "grad_norm": 0.015601031482219696,
+ "learning_rate": 0.0006,
+ "loss": 3.989492416381836,
+ "step": 2833
+ },
+ {
+ "epoch": 39.36331877729258,
+ "grad_norm": 0.014825052581727505,
+ "learning_rate": 0.0006,
+ "loss": 3.998520612716675,
+ "step": 2834
+ },
+ {
+ "epoch": 39.377292576419215,
+ "grad_norm": 0.014821100980043411,
+ "learning_rate": 0.0006,
+ "loss": 4.001777172088623,
+ "step": 2835
+ },
+ {
+ "epoch": 39.391266375545854,
+ "grad_norm": 0.015333754941821098,
+ "learning_rate": 0.0006,
+ "loss": 4.11985969543457,
+ "step": 2836
+ },
+ {
+ "epoch": 39.40524017467249,
+ "grad_norm": 0.015940699726343155,
+ "learning_rate": 0.0006,
+ "loss": 4.025373935699463,
+ "step": 2837
+ },
+ {
+ "epoch": 39.419213973799124,
+ "grad_norm": 0.015096287243068218,
+ "learning_rate": 0.0006,
+ "loss": 4.068551063537598,
+ "step": 2838
+ },
+ {
+ "epoch": 39.43318777292576,
+ "grad_norm": 0.01533548254519701,
+ "learning_rate": 0.0006,
+ "loss": 3.9800801277160645,
+ "step": 2839
+ },
+ {
+ "epoch": 39.4471615720524,
+ "grad_norm": 0.014648732729256153,
+ "learning_rate": 0.0006,
+ "loss": 3.9670190811157227,
+ "step": 2840
+ },
+ {
+ "epoch": 39.46113537117904,
+ "grad_norm": 0.015759671106934547,
+ "learning_rate": 0.0006,
+ "loss": 4.050443649291992,
+ "step": 2841
+ },
+ {
+ "epoch": 39.47510917030568,
+ "grad_norm": 0.016054073348641396,
+ "learning_rate": 0.0006,
+ "loss": 4.22734260559082,
+ "step": 2842
+ },
+ {
+ "epoch": 39.48908296943232,
+ "grad_norm": 0.0160264540463686,
+ "learning_rate": 0.0006,
+ "loss": 4.116174221038818,
+ "step": 2843
+ },
+ {
+ "epoch": 39.50305676855895,
+ "grad_norm": 0.015426610596477985,
+ "learning_rate": 0.0006,
+ "loss": 4.000178337097168,
+ "step": 2844
+ },
+ {
+ "epoch": 39.51703056768559,
+ "grad_norm": 0.01520642638206482,
+ "learning_rate": 0.0006,
+ "loss": 4.218410968780518,
+ "step": 2845
+ },
+ {
+ "epoch": 39.531004366812226,
+ "grad_norm": 0.015305949375033379,
+ "learning_rate": 0.0006,
+ "loss": 4.025891304016113,
+ "step": 2846
+ },
+ {
+ "epoch": 39.544978165938865,
+ "grad_norm": 0.015461564064025879,
+ "learning_rate": 0.0006,
+ "loss": 4.100950717926025,
+ "step": 2847
+ },
+ {
+ "epoch": 39.5589519650655,
+ "grad_norm": 0.01712625101208687,
+ "learning_rate": 0.0006,
+ "loss": 4.0620856285095215,
+ "step": 2848
+ },
+ {
+ "epoch": 39.57292576419214,
+ "grad_norm": 0.017133301123976707,
+ "learning_rate": 0.0006,
+ "loss": 4.097138404846191,
+ "step": 2849
+ },
+ {
+ "epoch": 39.58689956331878,
+ "grad_norm": 0.01683625392615795,
+ "learning_rate": 0.0006,
+ "loss": 4.084364891052246,
+ "step": 2850
+ },
+ {
+ "epoch": 39.60087336244541,
+ "grad_norm": 0.0165710486471653,
+ "learning_rate": 0.0006,
+ "loss": 4.042361259460449,
+ "step": 2851
+ },
+ {
+ "epoch": 39.61484716157205,
+ "grad_norm": 0.01660393364727497,
+ "learning_rate": 0.0006,
+ "loss": 4.1395487785339355,
+ "step": 2852
+ },
+ {
+ "epoch": 39.62882096069869,
+ "grad_norm": 0.017025845125317574,
+ "learning_rate": 0.0006,
+ "loss": 4.1207075119018555,
+ "step": 2853
+ },
+ {
+ "epoch": 39.64279475982533,
+ "grad_norm": 0.01660497486591339,
+ "learning_rate": 0.0006,
+ "loss": 4.103846073150635,
+ "step": 2854
+ },
+ {
+ "epoch": 39.65676855895197,
+ "grad_norm": 0.01607946865260601,
+ "learning_rate": 0.0006,
+ "loss": 4.189016342163086,
+ "step": 2855
+ },
+ {
+ "epoch": 39.670742358078606,
+ "grad_norm": 0.01614505983889103,
+ "learning_rate": 0.0006,
+ "loss": 4.079860687255859,
+ "step": 2856
+ },
+ {
+ "epoch": 39.68471615720524,
+ "grad_norm": 0.017085086554288864,
+ "learning_rate": 0.0006,
+ "loss": 3.9774293899536133,
+ "step": 2857
+ },
+ {
+ "epoch": 39.698689956331876,
+ "grad_norm": 0.01816520094871521,
+ "learning_rate": 0.0006,
+ "loss": 4.113666534423828,
+ "step": 2858
+ },
+ {
+ "epoch": 39.712663755458514,
+ "grad_norm": 0.016597526147961617,
+ "learning_rate": 0.0006,
+ "loss": 4.068242073059082,
+ "step": 2859
+ },
+ {
+ "epoch": 39.72663755458515,
+ "grad_norm": 0.01672235131263733,
+ "learning_rate": 0.0006,
+ "loss": 4.024444580078125,
+ "step": 2860
+ },
+ {
+ "epoch": 39.74061135371179,
+ "grad_norm": 0.020093556493520737,
+ "learning_rate": 0.0006,
+ "loss": 4.064124584197998,
+ "step": 2861
+ },
+ {
+ "epoch": 39.75458515283843,
+ "grad_norm": 0.023141304031014442,
+ "learning_rate": 0.0006,
+ "loss": 4.130439758300781,
+ "step": 2862
+ },
+ {
+ "epoch": 39.76855895196506,
+ "grad_norm": 0.02289372682571411,
+ "learning_rate": 0.0006,
+ "loss": 4.14794921875,
+ "step": 2863
+ },
+ {
+ "epoch": 39.7825327510917,
+ "grad_norm": 0.01919020712375641,
+ "learning_rate": 0.0006,
+ "loss": 4.123961925506592,
+ "step": 2864
+ },
+ {
+ "epoch": 39.79650655021834,
+ "grad_norm": 0.016932329162955284,
+ "learning_rate": 0.0006,
+ "loss": 4.0944132804870605,
+ "step": 2865
+ },
+ {
+ "epoch": 39.81048034934498,
+ "grad_norm": 0.016644535586237907,
+ "learning_rate": 0.0006,
+ "loss": 4.124269962310791,
+ "step": 2866
+ },
+ {
+ "epoch": 39.82445414847162,
+ "grad_norm": 0.016302216798067093,
+ "learning_rate": 0.0006,
+ "loss": 4.0267205238342285,
+ "step": 2867
+ },
+ {
+ "epoch": 39.838427947598255,
+ "grad_norm": 0.01673300378024578,
+ "learning_rate": 0.0006,
+ "loss": 4.10650634765625,
+ "step": 2868
+ },
+ {
+ "epoch": 39.852401746724894,
+ "grad_norm": 0.016581356525421143,
+ "learning_rate": 0.0006,
+ "loss": 4.1547956466674805,
+ "step": 2869
+ },
+ {
+ "epoch": 39.866375545851525,
+ "grad_norm": 0.01691797934472561,
+ "learning_rate": 0.0006,
+ "loss": 4.056936740875244,
+ "step": 2870
+ },
+ {
+ "epoch": 39.880349344978164,
+ "grad_norm": 0.01596726104617119,
+ "learning_rate": 0.0006,
+ "loss": 3.9891414642333984,
+ "step": 2871
+ },
+ {
+ "epoch": 39.8943231441048,
+ "grad_norm": 0.015615091659128666,
+ "learning_rate": 0.0006,
+ "loss": 4.137948036193848,
+ "step": 2872
+ },
+ {
+ "epoch": 39.90829694323144,
+ "grad_norm": 0.015703413635492325,
+ "learning_rate": 0.0006,
+ "loss": 3.9844930171966553,
+ "step": 2873
+ },
+ {
+ "epoch": 39.92227074235808,
+ "grad_norm": 0.015453094616532326,
+ "learning_rate": 0.0006,
+ "loss": 4.098443508148193,
+ "step": 2874
+ },
+ {
+ "epoch": 39.93624454148472,
+ "grad_norm": 0.01449158601462841,
+ "learning_rate": 0.0006,
+ "loss": 4.090852737426758,
+ "step": 2875
+ },
+ {
+ "epoch": 39.95021834061135,
+ "grad_norm": 0.01537515688687563,
+ "learning_rate": 0.0006,
+ "loss": 4.239612579345703,
+ "step": 2876
+ },
+ {
+ "epoch": 39.96419213973799,
+ "grad_norm": 0.015989921987056732,
+ "learning_rate": 0.0006,
+ "loss": 4.142073154449463,
+ "step": 2877
+ },
+ {
+ "epoch": 39.97816593886463,
+ "grad_norm": 0.015985611826181412,
+ "learning_rate": 0.0006,
+ "loss": 4.100076675415039,
+ "step": 2878
+ },
+ {
+ "epoch": 39.992139737991266,
+ "grad_norm": 0.015260833315551281,
+ "learning_rate": 0.0006,
+ "loss": 4.068576335906982,
+ "step": 2879
+ },
+ {
+ "epoch": 40.0,
+ "grad_norm": 0.018047941848635674,
+ "learning_rate": 0.0006,
+ "loss": 4.005878925323486,
+ "step": 2880
+ },
+ {
+ "epoch": 40.0,
+ "eval_loss": 4.622848987579346,
+ "eval_runtime": 56.9618,
+ "eval_samples_per_second": 42.871,
+ "eval_steps_per_second": 1.352,
+ "step": 2880
+ },
+ {
+ "epoch": 40.01397379912664,
+ "grad_norm": 0.015976838767528534,
+ "learning_rate": 0.0006,
+ "loss": 4.035689353942871,
+ "step": 2881
+ },
+ {
+ "epoch": 40.02794759825328,
+ "grad_norm": 0.015176220797002316,
+ "learning_rate": 0.0006,
+ "loss": 3.932197332382202,
+ "step": 2882
+ },
+ {
+ "epoch": 40.041921397379916,
+ "grad_norm": 0.016676228493452072,
+ "learning_rate": 0.0006,
+ "loss": 4.014171123504639,
+ "step": 2883
+ },
+ {
+ "epoch": 40.05589519650655,
+ "grad_norm": 0.01698281429708004,
+ "learning_rate": 0.0006,
+ "loss": 4.117893218994141,
+ "step": 2884
+ },
+ {
+ "epoch": 40.069868995633186,
+ "grad_norm": 0.017872486263513565,
+ "learning_rate": 0.0006,
+ "loss": 4.132119178771973,
+ "step": 2885
+ },
+ {
+ "epoch": 40.083842794759825,
+ "grad_norm": 0.017615554854273796,
+ "learning_rate": 0.0006,
+ "loss": 4.072000980377197,
+ "step": 2886
+ },
+ {
+ "epoch": 40.09781659388646,
+ "grad_norm": 0.017469394952058792,
+ "learning_rate": 0.0006,
+ "loss": 4.017726421356201,
+ "step": 2887
+ },
+ {
+ "epoch": 40.1117903930131,
+ "grad_norm": 0.01954762637615204,
+ "learning_rate": 0.0006,
+ "loss": 4.014443874359131,
+ "step": 2888
+ },
+ {
+ "epoch": 40.12576419213974,
+ "grad_norm": 0.018935853615403175,
+ "learning_rate": 0.0006,
+ "loss": 4.0042948722839355,
+ "step": 2889
+ },
+ {
+ "epoch": 40.13973799126637,
+ "grad_norm": 0.019382692873477936,
+ "learning_rate": 0.0006,
+ "loss": 4.06160831451416,
+ "step": 2890
+ },
+ {
+ "epoch": 40.15371179039301,
+ "grad_norm": 0.019677115604281425,
+ "learning_rate": 0.0006,
+ "loss": 4.13466215133667,
+ "step": 2891
+ },
+ {
+ "epoch": 40.16768558951965,
+ "grad_norm": 0.020365122705698013,
+ "learning_rate": 0.0006,
+ "loss": 4.095000267028809,
+ "step": 2892
+ },
+ {
+ "epoch": 40.18165938864629,
+ "grad_norm": 0.021842138841748238,
+ "learning_rate": 0.0006,
+ "loss": 3.9877853393554688,
+ "step": 2893
+ },
+ {
+ "epoch": 40.19563318777293,
+ "grad_norm": 0.02354881353676319,
+ "learning_rate": 0.0006,
+ "loss": 3.9430112838745117,
+ "step": 2894
+ },
+ {
+ "epoch": 40.209606986899566,
+ "grad_norm": 0.026602735742926598,
+ "learning_rate": 0.0006,
+ "loss": 4.012076377868652,
+ "step": 2895
+ },
+ {
+ "epoch": 40.223580786026204,
+ "grad_norm": 0.023976484313607216,
+ "learning_rate": 0.0006,
+ "loss": 3.9768595695495605,
+ "step": 2896
+ },
+ {
+ "epoch": 40.237554585152836,
+ "grad_norm": 0.02336616814136505,
+ "learning_rate": 0.0006,
+ "loss": 4.1096038818359375,
+ "step": 2897
+ },
+ {
+ "epoch": 40.251528384279474,
+ "grad_norm": 0.02426023595035076,
+ "learning_rate": 0.0006,
+ "loss": 4.157189846038818,
+ "step": 2898
+ },
+ {
+ "epoch": 40.26550218340611,
+ "grad_norm": 0.021572696045041084,
+ "learning_rate": 0.0006,
+ "loss": 3.979175090789795,
+ "step": 2899
+ },
+ {
+ "epoch": 40.27947598253275,
+ "grad_norm": 0.08300432562828064,
+ "learning_rate": 0.0006,
+ "loss": 4.092336654663086,
+ "step": 2900
+ },
+ {
+ "epoch": 40.29344978165939,
+ "grad_norm": 0.7323967814445496,
+ "learning_rate": 0.0006,
+ "loss": 5.337990760803223,
+ "step": 2901
+ },
+ {
+ "epoch": 40.30742358078603,
+ "grad_norm": 0.3363337814807892,
+ "learning_rate": 0.0006,
+ "loss": 6.40043830871582,
+ "step": 2902
+ },
+ {
+ "epoch": 40.32139737991266,
+ "grad_norm": 0.598686933517456,
+ "learning_rate": 0.0006,
+ "loss": 7.004765510559082,
+ "step": 2903
+ },
+ {
+ "epoch": 40.3353711790393,
+ "grad_norm": 0.5998364686965942,
+ "learning_rate": 0.0006,
+ "loss": 8.184747695922852,
+ "step": 2904
+ },
+ {
+ "epoch": 40.34934497816594,
+ "grad_norm": 1.408208966255188,
+ "learning_rate": 0.0006,
+ "loss": 8.88978385925293,
+ "step": 2905
+ },
+ {
+ "epoch": 40.36331877729258,
+ "grad_norm": 0.7829785943031311,
+ "learning_rate": 0.0006,
+ "loss": 11.13601303100586,
+ "step": 2906
+ },
+ {
+ "epoch": 40.377292576419215,
+ "grad_norm": 0.7769215106964111,
+ "learning_rate": 0.0006,
+ "loss": 9.90416145324707,
+ "step": 2907
+ },
+ {
+ "epoch": 40.391266375545854,
+ "grad_norm": 0.46508079767227173,
+ "learning_rate": 0.0006,
+ "loss": 9.647209167480469,
+ "step": 2908
+ },
+ {
+ "epoch": 40.40524017467249,
+ "grad_norm": 1.0228968858718872,
+ "learning_rate": 0.0006,
+ "loss": 11.685434341430664,
+ "step": 2909
+ },
+ {
+ "epoch": 40.419213973799124,
+ "grad_norm": 0.20553778111934662,
+ "learning_rate": 0.0006,
+ "loss": 8.621708869934082,
+ "step": 2910
+ },
+ {
+ "epoch": 40.43318777292576,
+ "grad_norm": 0.18466424942016602,
+ "learning_rate": 0.0006,
+ "loss": 8.590787887573242,
+ "step": 2911
+ },
+ {
+ "epoch": 40.4471615720524,
+ "grad_norm": 0.10458594560623169,
+ "learning_rate": 0.0006,
+ "loss": 8.324458122253418,
+ "step": 2912
+ },
+ {
+ "epoch": 40.46113537117904,
+ "grad_norm": 0.1096842959523201,
+ "learning_rate": 0.0006,
+ "loss": 8.300851821899414,
+ "step": 2913
+ },
+ {
+ "epoch": 40.47510917030568,
+ "grad_norm": 0.10501226782798767,
+ "learning_rate": 0.0006,
+ "loss": 8.21009635925293,
+ "step": 2914
+ },
+ {
+ "epoch": 40.48908296943232,
+ "grad_norm": 0.13704165816307068,
+ "learning_rate": 0.0006,
+ "loss": 8.141355514526367,
+ "step": 2915
+ },
+ {
+ "epoch": 40.50305676855895,
+ "grad_norm": 0.12129108607769012,
+ "learning_rate": 0.0006,
+ "loss": 8.074612617492676,
+ "step": 2916
+ },
+ {
+ "epoch": 40.51703056768559,
+ "grad_norm": 0.10028897225856781,
+ "learning_rate": 0.0006,
+ "loss": 8.00384521484375,
+ "step": 2917
+ },
+ {
+ "epoch": 40.531004366812226,
+ "grad_norm": 0.11764246970415115,
+ "learning_rate": 0.0006,
+ "loss": 8.078598022460938,
+ "step": 2918
+ },
+ {
+ "epoch": 40.544978165938865,
+ "grad_norm": 0.10081959515810013,
+ "learning_rate": 0.0006,
+ "loss": 8.008938789367676,
+ "step": 2919
+ },
+ {
+ "epoch": 40.5589519650655,
+ "grad_norm": 0.1003170907497406,
+ "learning_rate": 0.0006,
+ "loss": 7.962940216064453,
+ "step": 2920
+ },
+ {
+ "epoch": 40.57292576419214,
+ "grad_norm": 0.08763492107391357,
+ "learning_rate": 0.0006,
+ "loss": 7.909419536590576,
+ "step": 2921
+ },
+ {
+ "epoch": 40.58689956331878,
+ "grad_norm": 0.0636807233095169,
+ "learning_rate": 0.0006,
+ "loss": 7.878905773162842,
+ "step": 2922
+ },
+ {
+ "epoch": 40.60087336244541,
+ "grad_norm": 0.06005936488509178,
+ "learning_rate": 0.0006,
+ "loss": 7.850980758666992,
+ "step": 2923
+ },
+ {
+ "epoch": 40.61484716157205,
+ "grad_norm": 0.06415296345949173,
+ "learning_rate": 0.0006,
+ "loss": 7.830353260040283,
+ "step": 2924
+ },
+ {
+ "epoch": 40.62882096069869,
+ "grad_norm": 0.0732007846236229,
+ "learning_rate": 0.0006,
+ "loss": 7.830746650695801,
+ "step": 2925
+ },
+ {
+ "epoch": 40.64279475982533,
+ "grad_norm": 0.06628110259771347,
+ "learning_rate": 0.0006,
+ "loss": 7.82661771774292,
+ "step": 2926
+ },
+ {
+ "epoch": 40.65676855895197,
+ "grad_norm": 0.05544218048453331,
+ "learning_rate": 0.0006,
+ "loss": 7.814305782318115,
+ "step": 2927
+ },
+ {
+ "epoch": 40.670742358078606,
+ "grad_norm": 0.050327103585004807,
+ "learning_rate": 0.0006,
+ "loss": 7.764055252075195,
+ "step": 2928
+ },
+ {
+ "epoch": 40.68471615720524,
+ "grad_norm": 0.04918425530195236,
+ "learning_rate": 0.0006,
+ "loss": 7.744625091552734,
+ "step": 2929
+ },
+ {
+ "epoch": 40.698689956331876,
+ "grad_norm": 0.042862266302108765,
+ "learning_rate": 0.0006,
+ "loss": 7.764426231384277,
+ "step": 2930
+ },
+ {
+ "epoch": 40.712663755458514,
+ "grad_norm": 0.04338089004158974,
+ "learning_rate": 0.0006,
+ "loss": 7.712530136108398,
+ "step": 2931
+ },
+ {
+ "epoch": 40.72663755458515,
+ "grad_norm": 0.04127737879753113,
+ "learning_rate": 0.0006,
+ "loss": 7.742180347442627,
+ "step": 2932
+ },
+ {
+ "epoch": 40.74061135371179,
+ "grad_norm": 0.03358631581068039,
+ "learning_rate": 0.0006,
+ "loss": 7.714834213256836,
+ "step": 2933
+ },
+ {
+ "epoch": 40.75458515283843,
+ "grad_norm": 0.02845790982246399,
+ "learning_rate": 0.0006,
+ "loss": 7.7021026611328125,
+ "step": 2934
+ },
+ {
+ "epoch": 40.76855895196506,
+ "grad_norm": 0.03012419492006302,
+ "learning_rate": 0.0006,
+ "loss": 7.712377548217773,
+ "step": 2935
+ },
+ {
+ "epoch": 40.7825327510917,
+ "grad_norm": 0.032981179654598236,
+ "learning_rate": 0.0006,
+ "loss": 7.666385650634766,
+ "step": 2936
+ },
+ {
+ "epoch": 40.79650655021834,
+ "grad_norm": 0.03357074037194252,
+ "learning_rate": 0.0006,
+ "loss": 7.649097919464111,
+ "step": 2937
+ },
+ {
+ "epoch": 40.81048034934498,
+ "grad_norm": 0.030914708971977234,
+ "learning_rate": 0.0006,
+ "loss": 7.649196624755859,
+ "step": 2938
+ },
+ {
+ "epoch": 40.82445414847162,
+ "grad_norm": 0.026296131312847137,
+ "learning_rate": 0.0006,
+ "loss": 7.629964828491211,
+ "step": 2939
+ },
+ {
+ "epoch": 40.838427947598255,
+ "grad_norm": 0.026677245274186134,
+ "learning_rate": 0.0006,
+ "loss": 7.599973678588867,
+ "step": 2940
+ },
+ {
+ "epoch": 40.852401746724894,
+ "grad_norm": 0.023208219558000565,
+ "learning_rate": 0.0006,
+ "loss": 7.601624488830566,
+ "step": 2941
+ },
+ {
+ "epoch": 40.866375545851525,
+ "grad_norm": 0.020741475746035576,
+ "learning_rate": 0.0006,
+ "loss": 7.603902339935303,
+ "step": 2942
+ },
+ {
+ "epoch": 40.880349344978164,
+ "grad_norm": 0.019221751019358635,
+ "learning_rate": 0.0006,
+ "loss": 7.5535888671875,
+ "step": 2943
+ },
+ {
+ "epoch": 40.8943231441048,
+ "grad_norm": 0.019491951912641525,
+ "learning_rate": 0.0006,
+ "loss": 7.54597282409668,
+ "step": 2944
+ },
+ {
+ "epoch": 40.90829694323144,
+ "grad_norm": 0.019702572375535965,
+ "learning_rate": 0.0006,
+ "loss": 7.537567138671875,
+ "step": 2945
+ },
+ {
+ "epoch": 40.92227074235808,
+ "grad_norm": 0.01768800988793373,
+ "learning_rate": 0.0006,
+ "loss": 7.512800216674805,
+ "step": 2946
+ },
+ {
+ "epoch": 40.93624454148472,
+ "grad_norm": 0.01926138624548912,
+ "learning_rate": 0.0006,
+ "loss": 7.509932518005371,
+ "step": 2947
+ },
+ {
+ "epoch": 40.95021834061135,
+ "grad_norm": 0.019030006602406502,
+ "learning_rate": 0.0006,
+ "loss": 7.497983932495117,
+ "step": 2948
+ },
+ {
+ "epoch": 40.96419213973799,
+ "grad_norm": 0.017313841730356216,
+ "learning_rate": 0.0006,
+ "loss": 7.463518142700195,
+ "step": 2949
+ },
+ {
+ "epoch": 40.97816593886463,
+ "grad_norm": 0.01310762483626604,
+ "learning_rate": 0.0006,
+ "loss": 7.474668502807617,
+ "step": 2950
+ },
+ {
+ "epoch": 40.992139737991266,
+ "grad_norm": 0.013008282519876957,
+ "learning_rate": 0.0006,
+ "loss": 7.45982551574707,
+ "step": 2951
+ },
+ {
+ "epoch": 41.0,
+ "grad_norm": 0.01441363524645567,
+ "learning_rate": 0.0006,
+ "loss": 7.446529388427734,
+ "step": 2952
+ },
+ {
+ "epoch": 41.0,
+ "eval_loss": 7.486120223999023,
+ "eval_runtime": 70.4986,
+ "eval_samples_per_second": 34.639,
+ "eval_steps_per_second": 1.092,
+ "step": 2952
+ },
+ {
+ "epoch": 41.01397379912664,
+ "grad_norm": 0.014061855152249336,
+ "learning_rate": 0.0006,
+ "loss": 7.4692182540893555,
+ "step": 2953
+ },
+ {
+ "epoch": 41.02794759825328,
+ "grad_norm": 0.01587500050663948,
+ "learning_rate": 0.0006,
+ "loss": 7.4182024002075195,
+ "step": 2954
+ },
+ {
+ "epoch": 41.041921397379916,
+ "grad_norm": 0.014783984050154686,
+ "learning_rate": 0.0006,
+ "loss": 7.466778755187988,
+ "step": 2955
+ },
+ {
+ "epoch": 41.05589519650655,
+ "grad_norm": 0.011560854502022266,
+ "learning_rate": 0.0006,
+ "loss": 7.438002586364746,
+ "step": 2956
+ },
+ {
+ "epoch": 41.069868995633186,
+ "grad_norm": 0.00970957800745964,
+ "learning_rate": 0.0006,
+ "loss": 7.442694187164307,
+ "step": 2957
+ },
+ {
+ "epoch": 41.083842794759825,
+ "grad_norm": 0.009633459150791168,
+ "learning_rate": 0.0006,
+ "loss": 7.458884239196777,
+ "step": 2958
+ },
+ {
+ "epoch": 41.09781659388646,
+ "grad_norm": 0.010664350353181362,
+ "learning_rate": 0.0006,
+ "loss": 7.48378324508667,
+ "step": 2959
+ },
+ {
+ "epoch": 41.1117903930131,
+ "grad_norm": 0.011440591886639595,
+ "learning_rate": 0.0006,
+ "loss": 7.474484920501709,
+ "step": 2960
+ },
+ {
+ "epoch": 41.12576419213974,
+ "grad_norm": 0.01150908600538969,
+ "learning_rate": 0.0006,
+ "loss": 7.417902946472168,
+ "step": 2961
+ },
+ {
+ "epoch": 41.13973799126637,
+ "grad_norm": 0.010375697165727615,
+ "learning_rate": 0.0006,
+ "loss": 7.403279781341553,
+ "step": 2962
+ },
+ {
+ "epoch": 41.15371179039301,
+ "grad_norm": 0.009075703099370003,
+ "learning_rate": 0.0006,
+ "loss": 7.447347164154053,
+ "step": 2963
+ },
+ {
+ "epoch": 41.16768558951965,
+ "grad_norm": 0.008892177604138851,
+ "learning_rate": 0.0006,
+ "loss": 7.402681350708008,
+ "step": 2964
+ },
+ {
+ "epoch": 41.18165938864629,
+ "grad_norm": 0.009492084383964539,
+ "learning_rate": 0.0006,
+ "loss": 7.3871355056762695,
+ "step": 2965
+ },
+ {
+ "epoch": 41.19563318777293,
+ "grad_norm": 0.009139460511505604,
+ "learning_rate": 0.0006,
+ "loss": 7.419617176055908,
+ "step": 2966
+ },
+ {
+ "epoch": 41.209606986899566,
+ "grad_norm": 0.007959138602018356,
+ "learning_rate": 0.0006,
+ "loss": 7.374949932098389,
+ "step": 2967
+ },
+ {
+ "epoch": 41.223580786026204,
+ "grad_norm": 0.007571548223495483,
+ "learning_rate": 0.0006,
+ "loss": 7.4205241203308105,
+ "step": 2968
+ },
+ {
+ "epoch": 41.237554585152836,
+ "grad_norm": 0.007885400205850601,
+ "learning_rate": 0.0006,
+ "loss": 7.416543006896973,
+ "step": 2969
+ },
+ {
+ "epoch": 41.251528384279474,
+ "grad_norm": 0.008650158531963825,
+ "learning_rate": 0.0006,
+ "loss": 7.357326984405518,
+ "step": 2970
+ },
+ {
+ "epoch": 41.26550218340611,
+ "grad_norm": 0.007483262103050947,
+ "learning_rate": 0.0006,
+ "loss": 7.410891056060791,
+ "step": 2971
+ },
+ {
+ "epoch": 41.27947598253275,
+ "grad_norm": 0.00869277399033308,
+ "learning_rate": 0.0006,
+ "loss": 7.392127990722656,
+ "step": 2972
+ },
+ {
+ "epoch": 41.29344978165939,
+ "grad_norm": 0.00885078776627779,
+ "learning_rate": 0.0006,
+ "loss": 7.405449867248535,
+ "step": 2973
+ },
+ {
+ "epoch": 41.30742358078603,
+ "grad_norm": 0.00958226341754198,
+ "learning_rate": 0.0006,
+ "loss": 7.399868488311768,
+ "step": 2974
+ },
+ {
+ "epoch": 41.32139737991266,
+ "grad_norm": 0.006991599686443806,
+ "learning_rate": 0.0006,
+ "loss": 7.35369873046875,
+ "step": 2975
+ },
+ {
+ "epoch": 41.3353711790393,
+ "grad_norm": 0.007415142375975847,
+ "learning_rate": 0.0006,
+ "loss": 7.358353137969971,
+ "step": 2976
+ },
+ {
+ "epoch": 41.34934497816594,
+ "grad_norm": 0.007024371065199375,
+ "learning_rate": 0.0006,
+ "loss": 7.4070892333984375,
+ "step": 2977
+ },
+ {
+ "epoch": 41.36331877729258,
+ "grad_norm": 0.007234332151710987,
+ "learning_rate": 0.0006,
+ "loss": 7.379120349884033,
+ "step": 2978
+ },
+ {
+ "epoch": 41.377292576419215,
+ "grad_norm": 0.00798892229795456,
+ "learning_rate": 0.0006,
+ "loss": 7.365455627441406,
+ "step": 2979
+ },
+ {
+ "epoch": 41.391266375545854,
+ "grad_norm": 0.007303398102521896,
+ "learning_rate": 0.0006,
+ "loss": 7.341798782348633,
+ "step": 2980
+ },
+ {
+ "epoch": 41.40524017467249,
+ "grad_norm": 0.007119669578969479,
+ "learning_rate": 0.0006,
+ "loss": 7.363280296325684,
+ "step": 2981
+ },
+ {
+ "epoch": 41.419213973799124,
+ "grad_norm": 0.007457619532942772,
+ "learning_rate": 0.0006,
+ "loss": 7.39388370513916,
+ "step": 2982
+ },
+ {
+ "epoch": 41.43318777292576,
+ "grad_norm": 0.007905688136816025,
+ "learning_rate": 0.0006,
+ "loss": 7.380652904510498,
+ "step": 2983
+ },
+ {
+ "epoch": 41.4471615720524,
+ "grad_norm": 0.006212855689227581,
+ "learning_rate": 0.0006,
+ "loss": 7.398769378662109,
+ "step": 2984
+ },
+ {
+ "epoch": 41.46113537117904,
+ "grad_norm": 0.0070641543716192245,
+ "learning_rate": 0.0006,
+ "loss": 7.3503546714782715,
+ "step": 2985
+ },
+ {
+ "epoch": 41.47510917030568,
+ "grad_norm": 0.00787805300205946,
+ "learning_rate": 0.0006,
+ "loss": 7.347678184509277,
+ "step": 2986
+ },
+ {
+ "epoch": 41.48908296943232,
+ "grad_norm": 0.006760426796972752,
+ "learning_rate": 0.0006,
+ "loss": 7.380006790161133,
+ "step": 2987
+ },
+ {
+ "epoch": 41.50305676855895,
+ "grad_norm": 0.00772385997697711,
+ "learning_rate": 0.0006,
+ "loss": 7.363680362701416,
+ "step": 2988
+ },
+ {
+ "epoch": 41.51703056768559,
+ "grad_norm": 0.007373564876616001,
+ "learning_rate": 0.0006,
+ "loss": 7.333014488220215,
+ "step": 2989
+ },
+ {
+ "epoch": 41.531004366812226,
+ "grad_norm": 0.006717653013765812,
+ "learning_rate": 0.0006,
+ "loss": 7.360076427459717,
+ "step": 2990
+ },
+ {
+ "epoch": 41.544978165938865,
+ "grad_norm": 0.007191413082182407,
+ "learning_rate": 0.0006,
+ "loss": 7.397030830383301,
+ "step": 2991
+ },
+ {
+ "epoch": 41.5589519650655,
+ "grad_norm": 0.007831960916519165,
+ "learning_rate": 0.0006,
+ "loss": 7.369959831237793,
+ "step": 2992
+ },
+ {
+ "epoch": 41.57292576419214,
+ "grad_norm": 0.006947481073439121,
+ "learning_rate": 0.0006,
+ "loss": 7.333254337310791,
+ "step": 2993
+ },
+ {
+ "epoch": 41.58689956331878,
+ "grad_norm": 0.006243606563657522,
+ "learning_rate": 0.0006,
+ "loss": 7.327378273010254,
+ "step": 2994
+ },
+ {
+ "epoch": 41.60087336244541,
+ "grad_norm": 0.006298339460045099,
+ "learning_rate": 0.0006,
+ "loss": 7.323482036590576,
+ "step": 2995
+ },
+ {
+ "epoch": 41.61484716157205,
+ "grad_norm": 0.007587412837892771,
+ "learning_rate": 0.0006,
+ "loss": 7.313789367675781,
+ "step": 2996
+ },
+ {
+ "epoch": 41.62882096069869,
+ "grad_norm": 0.007140746805816889,
+ "learning_rate": 0.0006,
+ "loss": 7.356067657470703,
+ "step": 2997
+ },
+ {
+ "epoch": 41.64279475982533,
+ "grad_norm": 0.008315625600516796,
+ "learning_rate": 0.0006,
+ "loss": 7.3339643478393555,
+ "step": 2998
+ },
+ {
+ "epoch": 41.65676855895197,
+ "grad_norm": 0.006547102238982916,
+ "learning_rate": 0.0006,
+ "loss": 7.342191696166992,
+ "step": 2999
+ },
+ {
+ "epoch": 41.670742358078606,
+ "grad_norm": 0.006068122107535601,
+ "learning_rate": 0.0006,
+ "loss": 7.321015357971191,
+ "step": 3000
+ },
+ {
+ "epoch": 41.68471615720524,
+ "grad_norm": 0.007316991221159697,
+ "learning_rate": 0.0006,
+ "loss": 7.346490859985352,
+ "step": 3001
+ },
+ {
+ "epoch": 41.698689956331876,
+ "grad_norm": 0.0067388564348220825,
+ "learning_rate": 0.0006,
+ "loss": 7.340723991394043,
+ "step": 3002
+ },
+ {
+ "epoch": 41.712663755458514,
+ "grad_norm": 0.006839963141828775,
+ "learning_rate": 0.0006,
+ "loss": 7.329365253448486,
+ "step": 3003
+ },
+ {
+ "epoch": 41.72663755458515,
+ "grad_norm": 0.00805041566491127,
+ "learning_rate": 0.0006,
+ "loss": 7.298934459686279,
+ "step": 3004
+ },
+ {
+ "epoch": 41.74061135371179,
+ "grad_norm": 0.007197503466159105,
+ "learning_rate": 0.0006,
+ "loss": 7.331496238708496,
+ "step": 3005
+ },
+ {
+ "epoch": 41.75458515283843,
+ "grad_norm": 0.007220805156975985,
+ "learning_rate": 0.0006,
+ "loss": 7.317695140838623,
+ "step": 3006
+ },
+ {
+ "epoch": 41.76855895196506,
+ "grad_norm": 0.00874276738613844,
+ "learning_rate": 0.0006,
+ "loss": 7.272614479064941,
+ "step": 3007
+ },
+ {
+ "epoch": 41.7825327510917,
+ "grad_norm": 0.008386565372347832,
+ "learning_rate": 0.0006,
+ "loss": 7.295244216918945,
+ "step": 3008
+ },
+ {
+ "epoch": 41.79650655021834,
+ "grad_norm": 0.0077730282209813595,
+ "learning_rate": 0.0006,
+ "loss": 7.3395490646362305,
+ "step": 3009
+ },
+ {
+ "epoch": 41.81048034934498,
+ "grad_norm": 0.008441480807960033,
+ "learning_rate": 0.0006,
+ "loss": 7.319921493530273,
+ "step": 3010
+ },
+ {
+ "epoch": 41.82445414847162,
+ "grad_norm": 0.007398269604891539,
+ "learning_rate": 0.0006,
+ "loss": 7.3029465675354,
+ "step": 3011
+ },
+ {
+ "epoch": 41.838427947598255,
+ "grad_norm": 0.006878677289932966,
+ "learning_rate": 0.0006,
+ "loss": 7.318467140197754,
+ "step": 3012
+ },
+ {
+ "epoch": 41.852401746724894,
+ "grad_norm": 0.007616598624736071,
+ "learning_rate": 0.0006,
+ "loss": 7.3112077713012695,
+ "step": 3013
+ },
+ {
+ "epoch": 41.866375545851525,
+ "grad_norm": 0.011629699729382992,
+ "learning_rate": 0.0006,
+ "loss": 7.32523250579834,
+ "step": 3014
+ },
+ {
+ "epoch": 41.880349344978164,
+ "grad_norm": 0.011011249385774136,
+ "learning_rate": 0.0006,
+ "loss": 7.287247657775879,
+ "step": 3015
+ },
+ {
+ "epoch": 41.8943231441048,
+ "grad_norm": 0.010208295658230782,
+ "learning_rate": 0.0006,
+ "loss": 7.292117118835449,
+ "step": 3016
+ },
+ {
+ "epoch": 41.90829694323144,
+ "grad_norm": 0.008651095442473888,
+ "learning_rate": 0.0006,
+ "loss": 7.302201271057129,
+ "step": 3017
+ },
+ {
+ "epoch": 41.92227074235808,
+ "grad_norm": 0.006692920345813036,
+ "learning_rate": 0.0006,
+ "loss": 7.319568634033203,
+ "step": 3018
+ },
+ {
+ "epoch": 41.93624454148472,
+ "grad_norm": 0.010033348575234413,
+ "learning_rate": 0.0006,
+ "loss": 7.2738566398620605,
+ "step": 3019
+ },
+ {
+ "epoch": 41.95021834061135,
+ "grad_norm": 0.011790832504630089,
+ "learning_rate": 0.0006,
+ "loss": 7.298540115356445,
+ "step": 3020
+ },
+ {
+ "epoch": 41.96419213973799,
+ "grad_norm": 0.018622662872076035,
+ "learning_rate": 0.0006,
+ "loss": 7.267182350158691,
+ "step": 3021
+ },
+ {
+ "epoch": 41.97816593886463,
+ "grad_norm": 0.02520257607102394,
+ "learning_rate": 0.0006,
+ "loss": 7.261566162109375,
+ "step": 3022
+ },
+ {
+ "epoch": 41.992139737991266,
+ "grad_norm": 0.036332644522190094,
+ "learning_rate": 0.0006,
+ "loss": 7.270719051361084,
+ "step": 3023
+ },
+ {
+ "epoch": 42.0,
+ "grad_norm": 0.04717305675148964,
+ "learning_rate": 0.0006,
+ "loss": 7.293332099914551,
+ "step": 3024
+ },
+ {
+ "epoch": 42.0,
+ "eval_loss": 7.354280471801758,
+ "eval_runtime": 56.4353,
+ "eval_samples_per_second": 43.271,
+ "eval_steps_per_second": 1.364,
+ "step": 3024
+ },
+ {
+ "epoch": 42.01397379912664,
+ "grad_norm": 0.06661626696586609,
+ "learning_rate": 0.0006,
+ "loss": 7.338189601898193,
+ "step": 3025
+ },
+ {
+ "epoch": 42.02794759825328,
+ "grad_norm": 0.05452054366469383,
+ "learning_rate": 0.0006,
+ "loss": 7.316764831542969,
+ "step": 3026
+ },
+ {
+ "epoch": 42.041921397379916,
+ "grad_norm": 0.025638775900006294,
+ "learning_rate": 0.0006,
+ "loss": 7.34204626083374,
+ "step": 3027
+ },
+ {
+ "epoch": 42.05589519650655,
+ "grad_norm": 0.022398799657821655,
+ "learning_rate": 0.0006,
+ "loss": 7.262149333953857,
+ "step": 3028
+ },
+ {
+ "epoch": 42.069868995633186,
+ "grad_norm": 0.03713499382138252,
+ "learning_rate": 0.0006,
+ "loss": 7.27695369720459,
+ "step": 3029
+ },
+ {
+ "epoch": 42.083842794759825,
+ "grad_norm": 0.014949247241020203,
+ "learning_rate": 0.0006,
+ "loss": 7.296508312225342,
+ "step": 3030
+ },
+ {
+ "epoch": 42.09781659388646,
+ "grad_norm": 0.02703995071351528,
+ "learning_rate": 0.0006,
+ "loss": 7.258018493652344,
+ "step": 3031
+ },
+ {
+ "epoch": 42.1117903930131,
+ "grad_norm": 0.020352816209197044,
+ "learning_rate": 0.0006,
+ "loss": 7.241450309753418,
+ "step": 3032
+ },
+ {
+ "epoch": 42.12576419213974,
+ "grad_norm": 0.014578479342162609,
+ "learning_rate": 0.0006,
+ "loss": 7.257323265075684,
+ "step": 3033
+ },
+ {
+ "epoch": 42.13973799126637,
+ "grad_norm": 0.026053044945001602,
+ "learning_rate": 0.0006,
+ "loss": 7.254680633544922,
+ "step": 3034
+ },
+ {
+ "epoch": 42.15371179039301,
+ "grad_norm": 0.015120275318622589,
+ "learning_rate": 0.0006,
+ "loss": 7.261074066162109,
+ "step": 3035
+ },
+ {
+ "epoch": 42.16768558951965,
+ "grad_norm": 0.016822177916765213,
+ "learning_rate": 0.0006,
+ "loss": 7.279064178466797,
+ "step": 3036
+ },
+ {
+ "epoch": 42.18165938864629,
+ "grad_norm": 0.016367588192224503,
+ "learning_rate": 0.0006,
+ "loss": 7.251429557800293,
+ "step": 3037
+ },
+ {
+ "epoch": 42.19563318777293,
+ "grad_norm": 0.014350418001413345,
+ "learning_rate": 0.0006,
+ "loss": 7.259264945983887,
+ "step": 3038
+ },
+ {
+ "epoch": 42.209606986899566,
+ "grad_norm": 0.017078228294849396,
+ "learning_rate": 0.0006,
+ "loss": 7.227105140686035,
+ "step": 3039
+ },
+ {
+ "epoch": 42.223580786026204,
+ "grad_norm": 0.011759317480027676,
+ "learning_rate": 0.0006,
+ "loss": 7.275888442993164,
+ "step": 3040
+ },
+ {
+ "epoch": 42.237554585152836,
+ "grad_norm": 0.013019340112805367,
+ "learning_rate": 0.0006,
+ "loss": 7.2517008781433105,
+ "step": 3041
+ },
+ {
+ "epoch": 42.251528384279474,
+ "grad_norm": 0.01178441196680069,
+ "learning_rate": 0.0006,
+ "loss": 7.23576545715332,
+ "step": 3042
+ },
+ {
+ "epoch": 42.26550218340611,
+ "grad_norm": 0.014518891461193562,
+ "learning_rate": 0.0006,
+ "loss": 7.253895282745361,
+ "step": 3043
+ },
+ {
+ "epoch": 42.27947598253275,
+ "grad_norm": 0.018159950152039528,
+ "learning_rate": 0.0006,
+ "loss": 7.278058052062988,
+ "step": 3044
+ },
+ {
+ "epoch": 42.29344978165939,
+ "grad_norm": 0.007587796077132225,
+ "learning_rate": 0.0006,
+ "loss": 7.278578758239746,
+ "step": 3045
+ },
+ {
+ "epoch": 42.30742358078603,
+ "grad_norm": 0.015994979068636894,
+ "learning_rate": 0.0006,
+ "loss": 7.239841461181641,
+ "step": 3046
+ },
+ {
+ "epoch": 42.32139737991266,
+ "grad_norm": 0.017830168828368187,
+ "learning_rate": 0.0006,
+ "loss": 7.2443976402282715,
+ "step": 3047
+ },
+ {
+ "epoch": 42.3353711790393,
+ "grad_norm": 0.008738759905099869,
+ "learning_rate": 0.0006,
+ "loss": 7.232264518737793,
+ "step": 3048
+ },
+ {
+ "epoch": 42.34934497816594,
+ "grad_norm": 0.0175449438393116,
+ "learning_rate": 0.0006,
+ "loss": 7.228000640869141,
+ "step": 3049
+ },
+ {
+ "epoch": 42.36331877729258,
+ "grad_norm": 0.019611820578575134,
+ "learning_rate": 0.0006,
+ "loss": 7.234999179840088,
+ "step": 3050
+ },
+ {
+ "epoch": 42.377292576419215,
+ "grad_norm": 0.015302863903343678,
+ "learning_rate": 0.0006,
+ "loss": 7.242547988891602,
+ "step": 3051
+ },
+ {
+ "epoch": 42.391266375545854,
+ "grad_norm": 0.025919731706380844,
+ "learning_rate": 0.0006,
+ "loss": 7.216652870178223,
+ "step": 3052
+ },
+ {
+ "epoch": 42.40524017467249,
+ "grad_norm": 0.034587468951940536,
+ "learning_rate": 0.0006,
+ "loss": 7.245336532592773,
+ "step": 3053
+ },
+ {
+ "epoch": 42.419213973799124,
+ "grad_norm": 0.019198840484023094,
+ "learning_rate": 0.0006,
+ "loss": 7.256758689880371,
+ "step": 3054
+ },
+ {
+ "epoch": 42.43318777292576,
+ "grad_norm": 0.012316389009356499,
+ "learning_rate": 0.0006,
+ "loss": 7.209102630615234,
+ "step": 3055
+ },
+ {
+ "epoch": 42.4471615720524,
+ "grad_norm": 0.023512162268161774,
+ "learning_rate": 0.0006,
+ "loss": 7.250742435455322,
+ "step": 3056
+ },
+ {
+ "epoch": 42.46113537117904,
+ "grad_norm": 0.025457851588726044,
+ "learning_rate": 0.0006,
+ "loss": 7.263757705688477,
+ "step": 3057
+ },
+ {
+ "epoch": 42.47510917030568,
+ "grad_norm": 0.02192617766559124,
+ "learning_rate": 0.0006,
+ "loss": 7.246977806091309,
+ "step": 3058
+ },
+ {
+ "epoch": 42.48908296943232,
+ "grad_norm": 0.02541281282901764,
+ "learning_rate": 0.0006,
+ "loss": 7.232361316680908,
+ "step": 3059
+ },
+ {
+ "epoch": 42.50305676855895,
+ "grad_norm": 0.036353763192892075,
+ "learning_rate": 0.0006,
+ "loss": 7.242556571960449,
+ "step": 3060
+ },
+ {
+ "epoch": 42.51703056768559,
+ "grad_norm": 0.04555872827768326,
+ "learning_rate": 0.0006,
+ "loss": 7.228940963745117,
+ "step": 3061
+ },
+ {
+ "epoch": 42.531004366812226,
+ "grad_norm": 0.04563378915190697,
+ "learning_rate": 0.0006,
+ "loss": 7.235276222229004,
+ "step": 3062
+ },
+ {
+ "epoch": 42.544978165938865,
+ "grad_norm": 0.04275760427117348,
+ "learning_rate": 0.0006,
+ "loss": 7.238895416259766,
+ "step": 3063
+ },
+ {
+ "epoch": 42.5589519650655,
+ "grad_norm": 0.04535885155200958,
+ "learning_rate": 0.0006,
+ "loss": 7.205336570739746,
+ "step": 3064
+ },
+ {
+ "epoch": 42.57292576419214,
+ "grad_norm": 0.015534617006778717,
+ "learning_rate": 0.0006,
+ "loss": 7.238393783569336,
+ "step": 3065
+ },
+ {
+ "epoch": 42.58689956331878,
+ "grad_norm": 0.03286299481987953,
+ "learning_rate": 0.0006,
+ "loss": 7.216453552246094,
+ "step": 3066
+ },
+ {
+ "epoch": 42.60087336244541,
+ "grad_norm": 0.025453168898820877,
+ "learning_rate": 0.0006,
+ "loss": 7.203750133514404,
+ "step": 3067
+ },
+ {
+ "epoch": 42.61484716157205,
+ "grad_norm": 0.021501649171113968,
+ "learning_rate": 0.0006,
+ "loss": 7.241621971130371,
+ "step": 3068
+ },
+ {
+ "epoch": 42.62882096069869,
+ "grad_norm": 0.02743634767830372,
+ "learning_rate": 0.0006,
+ "loss": 7.22076940536499,
+ "step": 3069
+ },
+ {
+ "epoch": 42.64279475982533,
+ "grad_norm": 0.020890483632683754,
+ "learning_rate": 0.0006,
+ "loss": 7.218412399291992,
+ "step": 3070
+ },
+ {
+ "epoch": 42.65676855895197,
+ "grad_norm": 0.022499850019812584,
+ "learning_rate": 0.0006,
+ "loss": 7.217696189880371,
+ "step": 3071
+ },
+ {
+ "epoch": 42.670742358078606,
+ "grad_norm": 0.01653696969151497,
+ "learning_rate": 0.0006,
+ "loss": 7.156417369842529,
+ "step": 3072
+ },
+ {
+ "epoch": 42.68471615720524,
+ "grad_norm": 0.01834225095808506,
+ "learning_rate": 0.0006,
+ "loss": 7.211247444152832,
+ "step": 3073
+ },
+ {
+ "epoch": 42.698689956331876,
+ "grad_norm": 0.020137229934334755,
+ "learning_rate": 0.0006,
+ "loss": 7.232499599456787,
+ "step": 3074
+ },
+ {
+ "epoch": 42.712663755458514,
+ "grad_norm": 0.012671459466218948,
+ "learning_rate": 0.0006,
+ "loss": 7.172441482543945,
+ "step": 3075
+ },
+ {
+ "epoch": 42.72663755458515,
+ "grad_norm": 0.016803130507469177,
+ "learning_rate": 0.0006,
+ "loss": 7.152103424072266,
+ "step": 3076
+ },
+ {
+ "epoch": 42.74061135371179,
+ "grad_norm": 0.01623712107539177,
+ "learning_rate": 0.0006,
+ "loss": 7.187032699584961,
+ "step": 3077
+ },
+ {
+ "epoch": 42.75458515283843,
+ "grad_norm": 0.015938356518745422,
+ "learning_rate": 0.0006,
+ "loss": 7.171140670776367,
+ "step": 3078
+ },
+ {
+ "epoch": 42.76855895196506,
+ "grad_norm": 0.02580736204981804,
+ "learning_rate": 0.0006,
+ "loss": 7.204472541809082,
+ "step": 3079
+ },
+ {
+ "epoch": 42.7825327510917,
+ "grad_norm": 0.023372117429971695,
+ "learning_rate": 0.0006,
+ "loss": 7.165554523468018,
+ "step": 3080
+ },
+ {
+ "epoch": 42.79650655021834,
+ "grad_norm": 0.04440586268901825,
+ "learning_rate": 0.0006,
+ "loss": 7.169431209564209,
+ "step": 3081
+ },
+ {
+ "epoch": 42.81048034934498,
+ "grad_norm": 0.06797406077384949,
+ "learning_rate": 0.0006,
+ "loss": 7.168193340301514,
+ "step": 3082
+ },
+ {
+ "epoch": 42.82445414847162,
+ "grad_norm": 0.0914321169257164,
+ "learning_rate": 0.0006,
+ "loss": 7.204671859741211,
+ "step": 3083
+ },
+ {
+ "epoch": 42.838427947598255,
+ "grad_norm": 0.0506940595805645,
+ "learning_rate": 0.0006,
+ "loss": 7.15949821472168,
+ "step": 3084
+ },
+ {
+ "epoch": 42.852401746724894,
+ "grad_norm": 0.04819890484213829,
+ "learning_rate": 0.0006,
+ "loss": 7.192837238311768,
+ "step": 3085
+ },
+ {
+ "epoch": 42.866375545851525,
+ "grad_norm": 0.040044233202934265,
+ "learning_rate": 0.0006,
+ "loss": 7.1600165367126465,
+ "step": 3086
+ },
+ {
+ "epoch": 42.880349344978164,
+ "grad_norm": 0.02573299966752529,
+ "learning_rate": 0.0006,
+ "loss": 7.153427600860596,
+ "step": 3087
+ },
+ {
+ "epoch": 42.8943231441048,
+ "grad_norm": 0.046942535787820816,
+ "learning_rate": 0.0006,
+ "loss": 7.16767692565918,
+ "step": 3088
+ },
+ {
+ "epoch": 42.90829694323144,
+ "grad_norm": 0.02683161571621895,
+ "learning_rate": 0.0006,
+ "loss": 7.17144250869751,
+ "step": 3089
+ },
+ {
+ "epoch": 42.92227074235808,
+ "grad_norm": 0.05363674834370613,
+ "learning_rate": 0.0006,
+ "loss": 7.120128631591797,
+ "step": 3090
+ },
+ {
+ "epoch": 42.93624454148472,
+ "grad_norm": 0.023756247013807297,
+ "learning_rate": 0.0006,
+ "loss": 7.169703006744385,
+ "step": 3091
+ },
+ {
+ "epoch": 42.95021834061135,
+ "grad_norm": 0.04216783866286278,
+ "learning_rate": 0.0006,
+ "loss": 7.1206865310668945,
+ "step": 3092
+ },
+ {
+ "epoch": 42.96419213973799,
+ "grad_norm": 0.02219565585255623,
+ "learning_rate": 0.0006,
+ "loss": 7.157215118408203,
+ "step": 3093
+ },
+ {
+ "epoch": 42.97816593886463,
+ "grad_norm": 0.03527127206325531,
+ "learning_rate": 0.0006,
+ "loss": 7.166672706604004,
+ "step": 3094
+ },
+ {
+ "epoch": 42.992139737991266,
+ "grad_norm": 0.017748817801475525,
+ "learning_rate": 0.0006,
+ "loss": 7.154747009277344,
+ "step": 3095
+ },
+ {
+ "epoch": 43.0,
+ "grad_norm": 0.033165838569402695,
+ "learning_rate": 0.0006,
+ "loss": 7.150866508483887,
+ "step": 3096
+ },
+ {
+ "epoch": 43.0,
+ "eval_loss": 7.144403457641602,
+ "eval_runtime": 56.8853,
+ "eval_samples_per_second": 42.928,
+ "eval_steps_per_second": 1.354,
+ "step": 3096
+ },
+ {
+ "epoch": 43.01397379912664,
+ "grad_norm": 0.015199829824268818,
+ "learning_rate": 0.0006,
+ "loss": 7.122735023498535,
+ "step": 3097
+ },
+ {
+ "epoch": 43.02794759825328,
+ "grad_norm": 0.03220256045460701,
+ "learning_rate": 0.0006,
+ "loss": 7.143553733825684,
+ "step": 3098
+ },
+ {
+ "epoch": 43.041921397379916,
+ "grad_norm": 0.014687920920550823,
+ "learning_rate": 0.0006,
+ "loss": 7.080960750579834,
+ "step": 3099
+ },
+ {
+ "epoch": 43.05589519650655,
+ "grad_norm": 0.023522932082414627,
+ "learning_rate": 0.0006,
+ "loss": 7.137356758117676,
+ "step": 3100
+ },
+ {
+ "epoch": 43.069868995633186,
+ "grad_norm": 0.014808216132223606,
+ "learning_rate": 0.0006,
+ "loss": 7.09143590927124,
+ "step": 3101
+ },
+ {
+ "epoch": 43.083842794759825,
+ "grad_norm": 0.029339246451854706,
+ "learning_rate": 0.0006,
+ "loss": 7.122198104858398,
+ "step": 3102
+ },
+ {
+ "epoch": 43.09781659388646,
+ "grad_norm": 0.022953078150749207,
+ "learning_rate": 0.0006,
+ "loss": 7.092505931854248,
+ "step": 3103
+ },
+ {
+ "epoch": 43.1117903930131,
+ "grad_norm": 0.04590052738785744,
+ "learning_rate": 0.0006,
+ "loss": 7.113100051879883,
+ "step": 3104
+ },
+ {
+ "epoch": 43.12576419213974,
+ "grad_norm": 0.0524442233145237,
+ "learning_rate": 0.0006,
+ "loss": 7.147022247314453,
+ "step": 3105
+ },
+ {
+ "epoch": 43.13973799126637,
+ "grad_norm": 0.0315794013440609,
+ "learning_rate": 0.0006,
+ "loss": 7.094045162200928,
+ "step": 3106
+ },
+ {
+ "epoch": 43.15371179039301,
+ "grad_norm": 0.05112722888588905,
+ "learning_rate": 0.0006,
+ "loss": 7.054388523101807,
+ "step": 3107
+ },
+ {
+ "epoch": 43.16768558951965,
+ "grad_norm": 0.03146776929497719,
+ "learning_rate": 0.0006,
+ "loss": 7.116674423217773,
+ "step": 3108
+ },
+ {
+ "epoch": 43.18165938864629,
+ "grad_norm": 0.03391246497631073,
+ "learning_rate": 0.0006,
+ "loss": 7.150345802307129,
+ "step": 3109
+ },
+ {
+ "epoch": 43.19563318777293,
+ "grad_norm": 0.026957903057336807,
+ "learning_rate": 0.0006,
+ "loss": 7.058531284332275,
+ "step": 3110
+ },
+ {
+ "epoch": 43.209606986899566,
+ "grad_norm": 0.041532568633556366,
+ "learning_rate": 0.0006,
+ "loss": 7.085498809814453,
+ "step": 3111
+ },
+ {
+ "epoch": 43.223580786026204,
+ "grad_norm": 0.020001588389277458,
+ "learning_rate": 0.0006,
+ "loss": 7.131487846374512,
+ "step": 3112
+ },
+ {
+ "epoch": 43.237554585152836,
+ "grad_norm": 0.02268712781369686,
+ "learning_rate": 0.0006,
+ "loss": 7.0790863037109375,
+ "step": 3113
+ },
+ {
+ "epoch": 43.251528384279474,
+ "grad_norm": 0.023363925516605377,
+ "learning_rate": 0.0006,
+ "loss": 7.067816734313965,
+ "step": 3114
+ },
+ {
+ "epoch": 43.26550218340611,
+ "grad_norm": 0.019367508590221405,
+ "learning_rate": 0.0006,
+ "loss": 7.042041778564453,
+ "step": 3115
+ },
+ {
+ "epoch": 43.27947598253275,
+ "grad_norm": 0.02745090238749981,
+ "learning_rate": 0.0006,
+ "loss": 7.064733028411865,
+ "step": 3116
+ },
+ {
+ "epoch": 43.29344978165939,
+ "grad_norm": 0.02543802745640278,
+ "learning_rate": 0.0006,
+ "loss": 7.058948040008545,
+ "step": 3117
+ },
+ {
+ "epoch": 43.30742358078603,
+ "grad_norm": 0.018874092027544975,
+ "learning_rate": 0.0006,
+ "loss": 7.009946346282959,
+ "step": 3118
+ },
+ {
+ "epoch": 43.32139737991266,
+ "grad_norm": 0.024662822484970093,
+ "learning_rate": 0.0006,
+ "loss": 7.015913963317871,
+ "step": 3119
+ },
+ {
+ "epoch": 43.3353711790393,
+ "grad_norm": 0.02441416308283806,
+ "learning_rate": 0.0006,
+ "loss": 7.061622142791748,
+ "step": 3120
+ },
+ {
+ "epoch": 43.34934497816594,
+ "grad_norm": 0.031887322664260864,
+ "learning_rate": 0.0006,
+ "loss": 7.046883583068848,
+ "step": 3121
+ },
+ {
+ "epoch": 43.36331877729258,
+ "grad_norm": 0.021930739283561707,
+ "learning_rate": 0.0006,
+ "loss": 7.0309624671936035,
+ "step": 3122
+ },
+ {
+ "epoch": 43.377292576419215,
+ "grad_norm": 0.02722783200442791,
+ "learning_rate": 0.0006,
+ "loss": 7.072924613952637,
+ "step": 3123
+ },
+ {
+ "epoch": 43.391266375545854,
+ "grad_norm": 0.01483179721981287,
+ "learning_rate": 0.0006,
+ "loss": 7.028624534606934,
+ "step": 3124
+ },
+ {
+ "epoch": 43.40524017467249,
+ "grad_norm": 0.024793945252895355,
+ "learning_rate": 0.0006,
+ "loss": 7.001623153686523,
+ "step": 3125
+ },
+ {
+ "epoch": 43.419213973799124,
+ "grad_norm": 0.020148713141679764,
+ "learning_rate": 0.0006,
+ "loss": 6.9938812255859375,
+ "step": 3126
+ },
+ {
+ "epoch": 43.43318777292576,
+ "grad_norm": 0.02413850650191307,
+ "learning_rate": 0.0006,
+ "loss": 7.001243591308594,
+ "step": 3127
+ },
+ {
+ "epoch": 43.4471615720524,
+ "grad_norm": 0.028549203649163246,
+ "learning_rate": 0.0006,
+ "loss": 6.987671375274658,
+ "step": 3128
+ },
+ {
+ "epoch": 43.46113537117904,
+ "grad_norm": 0.025988785549998283,
+ "learning_rate": 0.0006,
+ "loss": 6.9551167488098145,
+ "step": 3129
+ },
+ {
+ "epoch": 43.47510917030568,
+ "grad_norm": 0.027140729129314423,
+ "learning_rate": 0.0006,
+ "loss": 6.9547810554504395,
+ "step": 3130
+ },
+ {
+ "epoch": 43.48908296943232,
+ "grad_norm": 0.030130082741379738,
+ "learning_rate": 0.0006,
+ "loss": 6.993186950683594,
+ "step": 3131
+ },
+ {
+ "epoch": 43.50305676855895,
+ "grad_norm": 0.025996994227170944,
+ "learning_rate": 0.0006,
+ "loss": 6.984624862670898,
+ "step": 3132
+ },
+ {
+ "epoch": 43.51703056768559,
+ "grad_norm": 0.02423008903861046,
+ "learning_rate": 0.0006,
+ "loss": 6.98813009262085,
+ "step": 3133
+ },
+ {
+ "epoch": 43.531004366812226,
+ "grad_norm": 0.029665837064385414,
+ "learning_rate": 0.0006,
+ "loss": 6.986610412597656,
+ "step": 3134
+ },
+ {
+ "epoch": 43.544978165938865,
+ "grad_norm": 0.028768595308065414,
+ "learning_rate": 0.0006,
+ "loss": 6.898797035217285,
+ "step": 3135
+ },
+ {
+ "epoch": 43.5589519650655,
+ "grad_norm": 0.03111208975315094,
+ "learning_rate": 0.0006,
+ "loss": 6.923741817474365,
+ "step": 3136
+ },
+ {
+ "epoch": 43.57292576419214,
+ "grad_norm": 0.0368187390267849,
+ "learning_rate": 0.0006,
+ "loss": 6.994343280792236,
+ "step": 3137
+ },
+ {
+ "epoch": 43.58689956331878,
+ "grad_norm": 0.04076763987541199,
+ "learning_rate": 0.0006,
+ "loss": 6.909458160400391,
+ "step": 3138
+ },
+ {
+ "epoch": 43.60087336244541,
+ "grad_norm": 0.03281061723828316,
+ "learning_rate": 0.0006,
+ "loss": 6.924851417541504,
+ "step": 3139
+ },
+ {
+ "epoch": 43.61484716157205,
+ "grad_norm": 0.025486988946795464,
+ "learning_rate": 0.0006,
+ "loss": 6.962926864624023,
+ "step": 3140
+ },
+ {
+ "epoch": 43.62882096069869,
+ "grad_norm": 0.04583485797047615,
+ "learning_rate": 0.0006,
+ "loss": 6.876555442810059,
+ "step": 3141
+ },
+ {
+ "epoch": 43.64279475982533,
+ "grad_norm": 0.05460785701870918,
+ "learning_rate": 0.0006,
+ "loss": 6.952932357788086,
+ "step": 3142
+ },
+ {
+ "epoch": 43.65676855895197,
+ "grad_norm": 0.040917906910181046,
+ "learning_rate": 0.0006,
+ "loss": 6.9624924659729,
+ "step": 3143
+ },
+ {
+ "epoch": 43.670742358078606,
+ "grad_norm": 0.04009668901562691,
+ "learning_rate": 0.0006,
+ "loss": 6.94033145904541,
+ "step": 3144
+ },
+ {
+ "epoch": 43.68471615720524,
+ "grad_norm": 0.04500625655055046,
+ "learning_rate": 0.0006,
+ "loss": 6.961839199066162,
+ "step": 3145
+ },
+ {
+ "epoch": 43.698689956331876,
+ "grad_norm": 0.03182258829474449,
+ "learning_rate": 0.0006,
+ "loss": 6.930580139160156,
+ "step": 3146
+ },
+ {
+ "epoch": 43.712663755458514,
+ "grad_norm": 0.053621504455804825,
+ "learning_rate": 0.0006,
+ "loss": 6.950235366821289,
+ "step": 3147
+ },
+ {
+ "epoch": 43.72663755458515,
+ "grad_norm": 0.031919289380311966,
+ "learning_rate": 0.0006,
+ "loss": 6.895111560821533,
+ "step": 3148
+ },
+ {
+ "epoch": 43.74061135371179,
+ "grad_norm": 0.042713891714811325,
+ "learning_rate": 0.0006,
+ "loss": 6.940885543823242,
+ "step": 3149
+ },
+ {
+ "epoch": 43.75458515283843,
+ "grad_norm": 0.02385653182864189,
+ "learning_rate": 0.0006,
+ "loss": 6.948801517486572,
+ "step": 3150
+ },
+ {
+ "epoch": 43.76855895196506,
+ "grad_norm": 0.04100663587450981,
+ "learning_rate": 0.0006,
+ "loss": 6.970411777496338,
+ "step": 3151
+ },
+ {
+ "epoch": 43.7825327510917,
+ "grad_norm": 0.02729583904147148,
+ "learning_rate": 0.0006,
+ "loss": 6.899379253387451,
+ "step": 3152
+ },
+ {
+ "epoch": 43.79650655021834,
+ "grad_norm": 0.03128620237112045,
+ "learning_rate": 0.0006,
+ "loss": 6.931319236755371,
+ "step": 3153
+ },
+ {
+ "epoch": 43.81048034934498,
+ "grad_norm": 0.025488659739494324,
+ "learning_rate": 0.0006,
+ "loss": 6.8486409187316895,
+ "step": 3154
+ },
+ {
+ "epoch": 43.82445414847162,
+ "grad_norm": 0.03679376840591431,
+ "learning_rate": 0.0006,
+ "loss": 6.840811252593994,
+ "step": 3155
+ },
+ {
+ "epoch": 43.838427947598255,
+ "grad_norm": 0.020578034222126007,
+ "learning_rate": 0.0006,
+ "loss": 6.9088616371154785,
+ "step": 3156
+ },
+ {
+ "epoch": 43.852401746724894,
+ "grad_norm": 0.02935841865837574,
+ "learning_rate": 0.0006,
+ "loss": 6.806676864624023,
+ "step": 3157
+ },
+ {
+ "epoch": 43.866375545851525,
+ "grad_norm": 0.023319244384765625,
+ "learning_rate": 0.0006,
+ "loss": 6.878457546234131,
+ "step": 3158
+ },
+ {
+ "epoch": 43.880349344978164,
+ "grad_norm": 0.02779589593410492,
+ "learning_rate": 0.0006,
+ "loss": 6.822447776794434,
+ "step": 3159
+ },
+ {
+ "epoch": 43.8943231441048,
+ "grad_norm": 0.01879825070500374,
+ "learning_rate": 0.0006,
+ "loss": 6.854121208190918,
+ "step": 3160
+ },
+ {
+ "epoch": 43.90829694323144,
+ "grad_norm": 0.023286139592528343,
+ "learning_rate": 0.0006,
+ "loss": 6.836827278137207,
+ "step": 3161
+ },
+ {
+ "epoch": 43.92227074235808,
+ "grad_norm": 0.030358297750353813,
+ "learning_rate": 0.0006,
+ "loss": 6.860037803649902,
+ "step": 3162
+ },
+ {
+ "epoch": 43.93624454148472,
+ "grad_norm": 0.029621105641126633,
+ "learning_rate": 0.0006,
+ "loss": 6.884025573730469,
+ "step": 3163
+ },
+ {
+ "epoch": 43.95021834061135,
+ "grad_norm": 0.029829824343323708,
+ "learning_rate": 0.0006,
+ "loss": 6.8702392578125,
+ "step": 3164
+ },
+ {
+ "epoch": 43.96419213973799,
+ "grad_norm": 0.036892473697662354,
+ "learning_rate": 0.0006,
+ "loss": 6.806652069091797,
+ "step": 3165
+ },
+ {
+ "epoch": 43.97816593886463,
+ "grad_norm": 0.03206709027290344,
+ "learning_rate": 0.0006,
+ "loss": 6.828495979309082,
+ "step": 3166
+ },
+ {
+ "epoch": 43.992139737991266,
+ "grad_norm": 0.02744000405073166,
+ "learning_rate": 0.0006,
+ "loss": 6.832002639770508,
+ "step": 3167
+ },
+ {
+ "epoch": 44.0,
+ "grad_norm": 0.02205660194158554,
+ "learning_rate": 0.0006,
+ "loss": 6.898828506469727,
+ "step": 3168
+ },
+ {
+ "epoch": 44.0,
+ "eval_loss": 6.8150553703308105,
+ "eval_runtime": 56.784,
+ "eval_samples_per_second": 43.005,
+ "eval_steps_per_second": 1.356,
+ "step": 3168
+ },
+ {
+ "epoch": 44.01397379912664,
+ "grad_norm": 0.030421260744333267,
+ "learning_rate": 0.0006,
+ "loss": 6.775057792663574,
+ "step": 3169
+ },
+ {
+ "epoch": 44.02794759825328,
+ "grad_norm": 0.0465611033141613,
+ "learning_rate": 0.0006,
+ "loss": 6.849390506744385,
+ "step": 3170
+ },
+ {
+ "epoch": 44.041921397379916,
+ "grad_norm": 0.05977972596883774,
+ "learning_rate": 0.0006,
+ "loss": 6.725576400756836,
+ "step": 3171
+ },
+ {
+ "epoch": 44.05589519650655,
+ "grad_norm": 0.05432697385549545,
+ "learning_rate": 0.0006,
+ "loss": 6.768024444580078,
+ "step": 3172
+ },
+ {
+ "epoch": 44.069868995633186,
+ "grad_norm": 0.04850967228412628,
+ "learning_rate": 0.0006,
+ "loss": 6.799457550048828,
+ "step": 3173
+ },
+ {
+ "epoch": 44.083842794759825,
+ "grad_norm": 0.06201786920428276,
+ "learning_rate": 0.0006,
+ "loss": 6.897234916687012,
+ "step": 3174
+ },
+ {
+ "epoch": 44.09781659388646,
+ "grad_norm": 0.04695591703057289,
+ "learning_rate": 0.0006,
+ "loss": 6.826470375061035,
+ "step": 3175
+ },
+ {
+ "epoch": 44.1117903930131,
+ "grad_norm": 0.03850164636969566,
+ "learning_rate": 0.0006,
+ "loss": 6.835901260375977,
+ "step": 3176
+ },
+ {
+ "epoch": 44.12576419213974,
+ "grad_norm": 0.04250155761837959,
+ "learning_rate": 0.0006,
+ "loss": 6.782310485839844,
+ "step": 3177
+ },
+ {
+ "epoch": 44.13973799126637,
+ "grad_norm": 0.041465479880571365,
+ "learning_rate": 0.0006,
+ "loss": 6.785732746124268,
+ "step": 3178
+ },
+ {
+ "epoch": 44.15371179039301,
+ "grad_norm": 0.02527807280421257,
+ "learning_rate": 0.0006,
+ "loss": 6.78061580657959,
+ "step": 3179
+ },
+ {
+ "epoch": 44.16768558951965,
+ "grad_norm": 0.03660374879837036,
+ "learning_rate": 0.0006,
+ "loss": 6.783992767333984,
+ "step": 3180
+ },
+ {
+ "epoch": 44.18165938864629,
+ "grad_norm": 0.027567964047193527,
+ "learning_rate": 0.0006,
+ "loss": 6.765413284301758,
+ "step": 3181
+ },
+ {
+ "epoch": 44.19563318777293,
+ "grad_norm": 0.024301424622535706,
+ "learning_rate": 0.0006,
+ "loss": 6.760368347167969,
+ "step": 3182
+ },
+ {
+ "epoch": 44.209606986899566,
+ "grad_norm": 0.0266821701079607,
+ "learning_rate": 0.0006,
+ "loss": 6.71417236328125,
+ "step": 3183
+ },
+ {
+ "epoch": 44.223580786026204,
+ "grad_norm": 0.028125090524554253,
+ "learning_rate": 0.0006,
+ "loss": 6.646126747131348,
+ "step": 3184
+ },
+ {
+ "epoch": 44.237554585152836,
+ "grad_norm": 0.0185737032443285,
+ "learning_rate": 0.0006,
+ "loss": 6.712944030761719,
+ "step": 3185
+ },
+ {
+ "epoch": 44.251528384279474,
+ "grad_norm": 0.025836674496531487,
+ "learning_rate": 0.0006,
+ "loss": 6.707030296325684,
+ "step": 3186
+ },
+ {
+ "epoch": 44.26550218340611,
+ "grad_norm": 0.029744870960712433,
+ "learning_rate": 0.0006,
+ "loss": 6.6271071434021,
+ "step": 3187
+ },
+ {
+ "epoch": 44.27947598253275,
+ "grad_norm": 0.03021426685154438,
+ "learning_rate": 0.0006,
+ "loss": 6.677826404571533,
+ "step": 3188
+ },
+ {
+ "epoch": 44.29344978165939,
+ "grad_norm": 0.032727014273405075,
+ "learning_rate": 0.0006,
+ "loss": 6.657764434814453,
+ "step": 3189
+ },
+ {
+ "epoch": 44.30742358078603,
+ "grad_norm": 0.03696778789162636,
+ "learning_rate": 0.0006,
+ "loss": 6.632883548736572,
+ "step": 3190
+ },
+ {
+ "epoch": 44.32139737991266,
+ "grad_norm": 0.034630827605724335,
+ "learning_rate": 0.0006,
+ "loss": 6.759583473205566,
+ "step": 3191
+ },
+ {
+ "epoch": 44.3353711790393,
+ "grad_norm": 0.030517758801579475,
+ "learning_rate": 0.0006,
+ "loss": 6.702268600463867,
+ "step": 3192
+ },
+ {
+ "epoch": 44.34934497816594,
+ "grad_norm": 0.04105190932750702,
+ "learning_rate": 0.0006,
+ "loss": 6.705792427062988,
+ "step": 3193
+ },
+ {
+ "epoch": 44.36331877729258,
+ "grad_norm": 0.06659174710512161,
+ "learning_rate": 0.0006,
+ "loss": 6.73862361907959,
+ "step": 3194
+ },
+ {
+ "epoch": 44.377292576419215,
+ "grad_norm": 0.08013619482517242,
+ "learning_rate": 0.0006,
+ "loss": 6.710282802581787,
+ "step": 3195
+ },
+ {
+ "epoch": 44.391266375545854,
+ "grad_norm": 0.049850963056087494,
+ "learning_rate": 0.0006,
+ "loss": 6.684974670410156,
+ "step": 3196
+ },
+ {
+ "epoch": 44.40524017467249,
+ "grad_norm": 0.061247408390045166,
+ "learning_rate": 0.0006,
+ "loss": 6.716034412384033,
+ "step": 3197
+ },
+ {
+ "epoch": 44.419213973799124,
+ "grad_norm": 0.05868857726454735,
+ "learning_rate": 0.0006,
+ "loss": 6.589107513427734,
+ "step": 3198
+ },
+ {
+ "epoch": 44.43318777292576,
+ "grad_norm": 0.048878561705350876,
+ "learning_rate": 0.0006,
+ "loss": 6.694855690002441,
+ "step": 3199
+ },
+ {
+ "epoch": 44.4471615720524,
+ "grad_norm": 0.028748324140906334,
+ "learning_rate": 0.0006,
+ "loss": 6.629969596862793,
+ "step": 3200
+ },
+ {
+ "epoch": 44.46113537117904,
+ "grad_norm": 0.04060306400060654,
+ "learning_rate": 0.0006,
+ "loss": 6.721431255340576,
+ "step": 3201
+ },
+ {
+ "epoch": 44.47510917030568,
+ "grad_norm": 0.03833824768662453,
+ "learning_rate": 0.0006,
+ "loss": 6.625029563903809,
+ "step": 3202
+ },
+ {
+ "epoch": 44.48908296943232,
+ "grad_norm": 0.03791997581720352,
+ "learning_rate": 0.0006,
+ "loss": 6.669187545776367,
+ "step": 3203
+ },
+ {
+ "epoch": 44.50305676855895,
+ "grad_norm": 0.03259684890508652,
+ "learning_rate": 0.0006,
+ "loss": 6.665643692016602,
+ "step": 3204
+ },
+ {
+ "epoch": 44.51703056768559,
+ "grad_norm": 0.040658626705408096,
+ "learning_rate": 0.0006,
+ "loss": 6.639796257019043,
+ "step": 3205
+ },
+ {
+ "epoch": 44.531004366812226,
+ "grad_norm": 0.03691193461418152,
+ "learning_rate": 0.0006,
+ "loss": 6.692244529724121,
+ "step": 3206
+ },
+ {
+ "epoch": 44.544978165938865,
+ "grad_norm": 0.02814812771975994,
+ "learning_rate": 0.0006,
+ "loss": 6.61978006362915,
+ "step": 3207
+ },
+ {
+ "epoch": 44.5589519650655,
+ "grad_norm": 0.022155074402689934,
+ "learning_rate": 0.0006,
+ "loss": 6.625905990600586,
+ "step": 3208
+ },
+ {
+ "epoch": 44.57292576419214,
+ "grad_norm": 0.030794885009527206,
+ "learning_rate": 0.0006,
+ "loss": 6.678500652313232,
+ "step": 3209
+ },
+ {
+ "epoch": 44.58689956331878,
+ "grad_norm": 0.03201768547296524,
+ "learning_rate": 0.0006,
+ "loss": 6.6015729904174805,
+ "step": 3210
+ },
+ {
+ "epoch": 44.60087336244541,
+ "grad_norm": 0.025138361379504204,
+ "learning_rate": 0.0006,
+ "loss": 6.568234443664551,
+ "step": 3211
+ },
+ {
+ "epoch": 44.61484716157205,
+ "grad_norm": 0.02969353087246418,
+ "learning_rate": 0.0006,
+ "loss": 6.588366508483887,
+ "step": 3212
+ },
+ {
+ "epoch": 44.62882096069869,
+ "grad_norm": 0.03136962279677391,
+ "learning_rate": 0.0006,
+ "loss": 6.479961395263672,
+ "step": 3213
+ },
+ {
+ "epoch": 44.64279475982533,
+ "grad_norm": 0.030178170651197433,
+ "learning_rate": 0.0006,
+ "loss": 6.622910976409912,
+ "step": 3214
+ },
+ {
+ "epoch": 44.65676855895197,
+ "grad_norm": 0.02682245895266533,
+ "learning_rate": 0.0006,
+ "loss": 6.597707748413086,
+ "step": 3215
+ },
+ {
+ "epoch": 44.670742358078606,
+ "grad_norm": 0.025154557079076767,
+ "learning_rate": 0.0006,
+ "loss": 6.630541801452637,
+ "step": 3216
+ },
+ {
+ "epoch": 44.68471615720524,
+ "grad_norm": 0.03564463183283806,
+ "learning_rate": 0.0006,
+ "loss": 6.542823791503906,
+ "step": 3217
+ },
+ {
+ "epoch": 44.698689956331876,
+ "grad_norm": 0.049477651715278625,
+ "learning_rate": 0.0006,
+ "loss": 6.646030426025391,
+ "step": 3218
+ },
+ {
+ "epoch": 44.712663755458514,
+ "grad_norm": 0.06536535918712616,
+ "learning_rate": 0.0006,
+ "loss": 6.595463275909424,
+ "step": 3219
+ },
+ {
+ "epoch": 44.72663755458515,
+ "grad_norm": 0.07710316777229309,
+ "learning_rate": 0.0006,
+ "loss": 6.55524206161499,
+ "step": 3220
+ },
+ {
+ "epoch": 44.74061135371179,
+ "grad_norm": 0.09571163356304169,
+ "learning_rate": 0.0006,
+ "loss": 6.659764289855957,
+ "step": 3221
+ },
+ {
+ "epoch": 44.75458515283843,
+ "grad_norm": 0.055076830089092255,
+ "learning_rate": 0.0006,
+ "loss": 6.658882141113281,
+ "step": 3222
+ },
+ {
+ "epoch": 44.76855895196506,
+ "grad_norm": 0.07671099156141281,
+ "learning_rate": 0.0006,
+ "loss": 6.605809211730957,
+ "step": 3223
+ },
+ {
+ "epoch": 44.7825327510917,
+ "grad_norm": 0.05295584350824356,
+ "learning_rate": 0.0006,
+ "loss": 6.641455173492432,
+ "step": 3224
+ },
+ {
+ "epoch": 44.79650655021834,
+ "grad_norm": 0.09128042310476303,
+ "learning_rate": 0.0006,
+ "loss": 6.691066741943359,
+ "step": 3225
+ },
+ {
+ "epoch": 44.81048034934498,
+ "grad_norm": 0.05362958461046219,
+ "learning_rate": 0.0006,
+ "loss": 6.573406219482422,
+ "step": 3226
+ },
+ {
+ "epoch": 44.82445414847162,
+ "grad_norm": 0.046910062432289124,
+ "learning_rate": 0.0006,
+ "loss": 6.550876617431641,
+ "step": 3227
+ },
+ {
+ "epoch": 44.838427947598255,
+ "grad_norm": 0.03494197130203247,
+ "learning_rate": 0.0006,
+ "loss": 6.540179252624512,
+ "step": 3228
+ },
+ {
+ "epoch": 44.852401746724894,
+ "grad_norm": 0.04716387391090393,
+ "learning_rate": 0.0006,
+ "loss": 6.490459442138672,
+ "step": 3229
+ },
+ {
+ "epoch": 44.866375545851525,
+ "grad_norm": 0.04190506786108017,
+ "learning_rate": 0.0006,
+ "loss": 6.509925365447998,
+ "step": 3230
+ },
+ {
+ "epoch": 44.880349344978164,
+ "grad_norm": 0.04749064892530441,
+ "learning_rate": 0.0006,
+ "loss": 6.478779315948486,
+ "step": 3231
+ },
+ {
+ "epoch": 44.8943231441048,
+ "grad_norm": 0.028002185747027397,
+ "learning_rate": 0.0006,
+ "loss": 6.528904914855957,
+ "step": 3232
+ },
+ {
+ "epoch": 44.90829694323144,
+ "grad_norm": 0.036473166197538376,
+ "learning_rate": 0.0006,
+ "loss": 6.549042701721191,
+ "step": 3233
+ },
+ {
+ "epoch": 44.92227074235808,
+ "grad_norm": 0.02543315477669239,
+ "learning_rate": 0.0006,
+ "loss": 6.498875141143799,
+ "step": 3234
+ },
+ {
+ "epoch": 44.93624454148472,
+ "grad_norm": 0.029906807467341423,
+ "learning_rate": 0.0006,
+ "loss": 6.453019142150879,
+ "step": 3235
+ },
+ {
+ "epoch": 44.95021834061135,
+ "grad_norm": 0.027231857180595398,
+ "learning_rate": 0.0006,
+ "loss": 6.428829193115234,
+ "step": 3236
+ },
+ {
+ "epoch": 44.96419213973799,
+ "grad_norm": 0.02506154775619507,
+ "learning_rate": 0.0006,
+ "loss": 6.420858383178711,
+ "step": 3237
+ },
+ {
+ "epoch": 44.97816593886463,
+ "grad_norm": 0.0269472673535347,
+ "learning_rate": 0.0006,
+ "loss": 6.484989643096924,
+ "step": 3238
+ },
+ {
+ "epoch": 44.992139737991266,
+ "grad_norm": 0.023410487920045853,
+ "learning_rate": 0.0006,
+ "loss": 6.454830169677734,
+ "step": 3239
+ },
+ {
+ "epoch": 45.0,
+ "grad_norm": 0.034175023436546326,
+ "learning_rate": 0.0006,
+ "loss": 6.430975437164307,
+ "step": 3240
+ },
+ {
+ "epoch": 45.0,
+ "eval_loss": 6.458896160125732,
+ "eval_runtime": 56.6934,
+ "eval_samples_per_second": 43.074,
+ "eval_steps_per_second": 1.358,
+ "step": 3240
+ },
+ {
+ "epoch": 45.01397379912664,
+ "grad_norm": 0.045031629502773285,
+ "learning_rate": 0.0006,
+ "loss": 6.4243011474609375,
+ "step": 3241
+ },
+ {
+ "epoch": 45.02794759825328,
+ "grad_norm": 0.04968734085559845,
+ "learning_rate": 0.0006,
+ "loss": 6.380091667175293,
+ "step": 3242
+ },
+ {
+ "epoch": 45.041921397379916,
+ "grad_norm": 0.047145187854766846,
+ "learning_rate": 0.0006,
+ "loss": 6.477178573608398,
+ "step": 3243
+ },
+ {
+ "epoch": 45.05589519650655,
+ "grad_norm": 0.04304948449134827,
+ "learning_rate": 0.0006,
+ "loss": 6.399726867675781,
+ "step": 3244
+ },
+ {
+ "epoch": 45.069868995633186,
+ "grad_norm": 0.04466010630130768,
+ "learning_rate": 0.0006,
+ "loss": 6.3415679931640625,
+ "step": 3245
+ },
+ {
+ "epoch": 45.083842794759825,
+ "grad_norm": 0.04305306077003479,
+ "learning_rate": 0.0006,
+ "loss": 6.3413310050964355,
+ "step": 3246
+ },
+ {
+ "epoch": 45.09781659388646,
+ "grad_norm": 0.043064575642347336,
+ "learning_rate": 0.0006,
+ "loss": 6.4838714599609375,
+ "step": 3247
+ },
+ {
+ "epoch": 45.1117903930131,
+ "grad_norm": 0.049503277987241745,
+ "learning_rate": 0.0006,
+ "loss": 6.38473653793335,
+ "step": 3248
+ },
+ {
+ "epoch": 45.12576419213974,
+ "grad_norm": 0.04613238573074341,
+ "learning_rate": 0.0006,
+ "loss": 6.370769500732422,
+ "step": 3249
+ },
+ {
+ "epoch": 45.13973799126637,
+ "grad_norm": 0.052119478583335876,
+ "learning_rate": 0.0006,
+ "loss": 6.362321853637695,
+ "step": 3250
+ },
+ {
+ "epoch": 45.15371179039301,
+ "grad_norm": 0.045110076665878296,
+ "learning_rate": 0.0006,
+ "loss": 6.369844913482666,
+ "step": 3251
+ },
+ {
+ "epoch": 45.16768558951965,
+ "grad_norm": 0.03798583149909973,
+ "learning_rate": 0.0006,
+ "loss": 6.324633598327637,
+ "step": 3252
+ },
+ {
+ "epoch": 45.18165938864629,
+ "grad_norm": 0.03744659945368767,
+ "learning_rate": 0.0006,
+ "loss": 6.349421501159668,
+ "step": 3253
+ },
+ {
+ "epoch": 45.19563318777293,
+ "grad_norm": 0.03767292574048042,
+ "learning_rate": 0.0006,
+ "loss": 6.334238052368164,
+ "step": 3254
+ },
+ {
+ "epoch": 45.209606986899566,
+ "grad_norm": 0.03577509894967079,
+ "learning_rate": 0.0006,
+ "loss": 6.3453826904296875,
+ "step": 3255
+ },
+ {
+ "epoch": 45.223580786026204,
+ "grad_norm": 0.03403240442276001,
+ "learning_rate": 0.0006,
+ "loss": 6.2836174964904785,
+ "step": 3256
+ },
+ {
+ "epoch": 45.237554585152836,
+ "grad_norm": 0.03517331928014755,
+ "learning_rate": 0.0006,
+ "loss": 6.281718730926514,
+ "step": 3257
+ },
+ {
+ "epoch": 45.251528384279474,
+ "grad_norm": 0.04409139230847359,
+ "learning_rate": 0.0006,
+ "loss": 6.320815086364746,
+ "step": 3258
+ },
+ {
+ "epoch": 45.26550218340611,
+ "grad_norm": 0.054750364273786545,
+ "learning_rate": 0.0006,
+ "loss": 6.343618392944336,
+ "step": 3259
+ },
+ {
+ "epoch": 45.27947598253275,
+ "grad_norm": 0.07612067461013794,
+ "learning_rate": 0.0006,
+ "loss": 6.310614585876465,
+ "step": 3260
+ },
+ {
+ "epoch": 45.29344978165939,
+ "grad_norm": 0.08438985049724579,
+ "learning_rate": 0.0006,
+ "loss": 6.2369184494018555,
+ "step": 3261
+ },
+ {
+ "epoch": 45.30742358078603,
+ "grad_norm": 0.0782201811671257,
+ "learning_rate": 0.0006,
+ "loss": 6.254452705383301,
+ "step": 3262
+ },
+ {
+ "epoch": 45.32139737991266,
+ "grad_norm": 0.09928355365991592,
+ "learning_rate": 0.0006,
+ "loss": 6.3112077713012695,
+ "step": 3263
+ },
+ {
+ "epoch": 45.3353711790393,
+ "grad_norm": 0.07233133912086487,
+ "learning_rate": 0.0006,
+ "loss": 6.408087253570557,
+ "step": 3264
+ },
+ {
+ "epoch": 45.34934497816594,
+ "grad_norm": 0.07478684186935425,
+ "learning_rate": 0.0006,
+ "loss": 6.3770036697387695,
+ "step": 3265
+ },
+ {
+ "epoch": 45.36331877729258,
+ "grad_norm": 0.06036846712231636,
+ "learning_rate": 0.0006,
+ "loss": 6.33865213394165,
+ "step": 3266
+ },
+ {
+ "epoch": 45.377292576419215,
+ "grad_norm": 0.05783458054065704,
+ "learning_rate": 0.0006,
+ "loss": 6.363203048706055,
+ "step": 3267
+ },
+ {
+ "epoch": 45.391266375545854,
+ "grad_norm": 0.05443909019231796,
+ "learning_rate": 0.0006,
+ "loss": 6.287069797515869,
+ "step": 3268
+ },
+ {
+ "epoch": 45.40524017467249,
+ "grad_norm": 0.0557909794151783,
+ "learning_rate": 0.0006,
+ "loss": 6.297784805297852,
+ "step": 3269
+ },
+ {
+ "epoch": 45.419213973799124,
+ "grad_norm": 0.03483372926712036,
+ "learning_rate": 0.0006,
+ "loss": 6.280734062194824,
+ "step": 3270
+ },
+ {
+ "epoch": 45.43318777292576,
+ "grad_norm": 0.0425543375313282,
+ "learning_rate": 0.0006,
+ "loss": 6.33565616607666,
+ "step": 3271
+ },
+ {
+ "epoch": 45.4471615720524,
+ "grad_norm": 0.03230159357190132,
+ "learning_rate": 0.0006,
+ "loss": 6.197698593139648,
+ "step": 3272
+ },
+ {
+ "epoch": 45.46113537117904,
+ "grad_norm": 0.036897290498018265,
+ "learning_rate": 0.0006,
+ "loss": 6.241239547729492,
+ "step": 3273
+ },
+ {
+ "epoch": 45.47510917030568,
+ "grad_norm": 0.031103838235139847,
+ "learning_rate": 0.0006,
+ "loss": 6.159435749053955,
+ "step": 3274
+ },
+ {
+ "epoch": 45.48908296943232,
+ "grad_norm": 0.027116449549794197,
+ "learning_rate": 0.0006,
+ "loss": 6.212225914001465,
+ "step": 3275
+ },
+ {
+ "epoch": 45.50305676855895,
+ "grad_norm": 0.029923981055617332,
+ "learning_rate": 0.0006,
+ "loss": 6.157269477844238,
+ "step": 3276
+ },
+ {
+ "epoch": 45.51703056768559,
+ "grad_norm": 0.02521538734436035,
+ "learning_rate": 0.0006,
+ "loss": 6.146918773651123,
+ "step": 3277
+ },
+ {
+ "epoch": 45.531004366812226,
+ "grad_norm": 0.02758920192718506,
+ "learning_rate": 0.0006,
+ "loss": 6.0446977615356445,
+ "step": 3278
+ },
+ {
+ "epoch": 45.544978165938865,
+ "grad_norm": 0.024328220635652542,
+ "learning_rate": 0.0006,
+ "loss": 6.099508285522461,
+ "step": 3279
+ },
+ {
+ "epoch": 45.5589519650655,
+ "grad_norm": 0.027425771579146385,
+ "learning_rate": 0.0006,
+ "loss": 6.084299087524414,
+ "step": 3280
+ },
+ {
+ "epoch": 45.57292576419214,
+ "grad_norm": 0.026927651837468147,
+ "learning_rate": 0.0006,
+ "loss": 6.146334171295166,
+ "step": 3281
+ },
+ {
+ "epoch": 45.58689956331878,
+ "grad_norm": 0.04054379090666771,
+ "learning_rate": 0.0006,
+ "loss": 6.1301493644714355,
+ "step": 3282
+ },
+ {
+ "epoch": 45.60087336244541,
+ "grad_norm": 0.04456290975213051,
+ "learning_rate": 0.0006,
+ "loss": 6.152511119842529,
+ "step": 3283
+ },
+ {
+ "epoch": 45.61484716157205,
+ "grad_norm": 0.052821505814790726,
+ "learning_rate": 0.0006,
+ "loss": 6.113544940948486,
+ "step": 3284
+ },
+ {
+ "epoch": 45.62882096069869,
+ "grad_norm": 0.07251739501953125,
+ "learning_rate": 0.0006,
+ "loss": 6.118298530578613,
+ "step": 3285
+ },
+ {
+ "epoch": 45.64279475982533,
+ "grad_norm": 0.09433556348085403,
+ "learning_rate": 0.0006,
+ "loss": 6.170897483825684,
+ "step": 3286
+ },
+ {
+ "epoch": 45.65676855895197,
+ "grad_norm": 0.057683493942022324,
+ "learning_rate": 0.0006,
+ "loss": 6.1747822761535645,
+ "step": 3287
+ },
+ {
+ "epoch": 45.670742358078606,
+ "grad_norm": 0.06363270431756973,
+ "learning_rate": 0.0006,
+ "loss": 6.1617889404296875,
+ "step": 3288
+ },
+ {
+ "epoch": 45.68471615720524,
+ "grad_norm": 0.07528169453144073,
+ "learning_rate": 0.0006,
+ "loss": 6.230060577392578,
+ "step": 3289
+ },
+ {
+ "epoch": 45.698689956331876,
+ "grad_norm": 0.07816170156002045,
+ "learning_rate": 0.0006,
+ "loss": 6.171257019042969,
+ "step": 3290
+ },
+ {
+ "epoch": 45.712663755458514,
+ "grad_norm": 0.05860164389014244,
+ "learning_rate": 0.0006,
+ "loss": 6.095484256744385,
+ "step": 3291
+ },
+ {
+ "epoch": 45.72663755458515,
+ "grad_norm": 0.04747429117560387,
+ "learning_rate": 0.0006,
+ "loss": 6.060067176818848,
+ "step": 3292
+ },
+ {
+ "epoch": 45.74061135371179,
+ "grad_norm": 0.04514474794268608,
+ "learning_rate": 0.0006,
+ "loss": 6.100218772888184,
+ "step": 3293
+ },
+ {
+ "epoch": 45.75458515283843,
+ "grad_norm": 0.05168681964278221,
+ "learning_rate": 0.0006,
+ "loss": 6.0621867179870605,
+ "step": 3294
+ },
+ {
+ "epoch": 45.76855895196506,
+ "grad_norm": 0.04665876924991608,
+ "learning_rate": 0.0006,
+ "loss": 6.118292808532715,
+ "step": 3295
+ },
+ {
+ "epoch": 45.7825327510917,
+ "grad_norm": 0.049374762922525406,
+ "learning_rate": 0.0006,
+ "loss": 6.098394393920898,
+ "step": 3296
+ },
+ {
+ "epoch": 45.79650655021834,
+ "grad_norm": 0.03793549910187721,
+ "learning_rate": 0.0006,
+ "loss": 6.069541931152344,
+ "step": 3297
+ },
+ {
+ "epoch": 45.81048034934498,
+ "grad_norm": 0.036362383514642715,
+ "learning_rate": 0.0006,
+ "loss": 6.1188435554504395,
+ "step": 3298
+ },
+ {
+ "epoch": 45.82445414847162,
+ "grad_norm": 0.03397097811102867,
+ "learning_rate": 0.0006,
+ "loss": 6.117589950561523,
+ "step": 3299
+ },
+ {
+ "epoch": 45.838427947598255,
+ "grad_norm": 0.030977753922343254,
+ "learning_rate": 0.0006,
+ "loss": 6.0236711502075195,
+ "step": 3300
+ },
+ {
+ "epoch": 45.852401746724894,
+ "grad_norm": 0.02800217643380165,
+ "learning_rate": 0.0006,
+ "loss": 6.034825325012207,
+ "step": 3301
+ },
+ {
+ "epoch": 45.866375545851525,
+ "grad_norm": 0.03129778802394867,
+ "learning_rate": 0.0006,
+ "loss": 5.980271339416504,
+ "step": 3302
+ },
+ {
+ "epoch": 45.880349344978164,
+ "grad_norm": 0.025886094197630882,
+ "learning_rate": 0.0006,
+ "loss": 5.966419219970703,
+ "step": 3303
+ },
+ {
+ "epoch": 45.8943231441048,
+ "grad_norm": 0.025712328031659126,
+ "learning_rate": 0.0006,
+ "loss": 5.938652515411377,
+ "step": 3304
+ },
+ {
+ "epoch": 45.90829694323144,
+ "grad_norm": 0.022508298978209496,
+ "learning_rate": 0.0006,
+ "loss": 5.982312202453613,
+ "step": 3305
+ },
+ {
+ "epoch": 45.92227074235808,
+ "grad_norm": 0.024880951270461082,
+ "learning_rate": 0.0006,
+ "loss": 5.931150436401367,
+ "step": 3306
+ },
+ {
+ "epoch": 45.93624454148472,
+ "grad_norm": 0.030007576569914818,
+ "learning_rate": 0.0006,
+ "loss": 5.936639308929443,
+ "step": 3307
+ },
+ {
+ "epoch": 45.95021834061135,
+ "grad_norm": 0.033314041793346405,
+ "learning_rate": 0.0006,
+ "loss": 5.944248676300049,
+ "step": 3308
+ },
+ {
+ "epoch": 45.96419213973799,
+ "grad_norm": 0.043078381568193436,
+ "learning_rate": 0.0006,
+ "loss": 5.971286773681641,
+ "step": 3309
+ },
+ {
+ "epoch": 45.97816593886463,
+ "grad_norm": 0.059296440333127975,
+ "learning_rate": 0.0006,
+ "loss": 5.910497665405273,
+ "step": 3310
+ },
+ {
+ "epoch": 45.992139737991266,
+ "grad_norm": 0.06232666224241257,
+ "learning_rate": 0.0006,
+ "loss": 5.929817199707031,
+ "step": 3311
+ },
+ {
+ "epoch": 46.0,
+ "grad_norm": 0.05303940549492836,
+ "learning_rate": 0.0006,
+ "loss": 5.910822868347168,
+ "step": 3312
+ },
+ {
+ "epoch": 46.0,
+ "eval_loss": 5.977851390838623,
+ "eval_runtime": 56.2435,
+ "eval_samples_per_second": 43.418,
+ "eval_steps_per_second": 1.369,
+ "step": 3312
+ },
+ {
+ "epoch": 46.01397379912664,
+ "grad_norm": 0.06943798810243607,
+ "learning_rate": 0.0006,
+ "loss": 5.854922294616699,
+ "step": 3313
+ },
+ {
+ "epoch": 46.02794759825328,
+ "grad_norm": 0.05236278846859932,
+ "learning_rate": 0.0006,
+ "loss": 5.876611232757568,
+ "step": 3314
+ },
+ {
+ "epoch": 46.041921397379916,
+ "grad_norm": 0.06240001320838928,
+ "learning_rate": 0.0006,
+ "loss": 5.878912448883057,
+ "step": 3315
+ },
+ {
+ "epoch": 46.05589519650655,
+ "grad_norm": 0.07401192933320999,
+ "learning_rate": 0.0006,
+ "loss": 5.985136032104492,
+ "step": 3316
+ },
+ {
+ "epoch": 46.069868995633186,
+ "grad_norm": 0.0826529711484909,
+ "learning_rate": 0.0006,
+ "loss": 5.939439296722412,
+ "step": 3317
+ },
+ {
+ "epoch": 46.083842794759825,
+ "grad_norm": 0.07999005168676376,
+ "learning_rate": 0.0006,
+ "loss": 5.982336044311523,
+ "step": 3318
+ },
+ {
+ "epoch": 46.09781659388646,
+ "grad_norm": 0.07014869153499603,
+ "learning_rate": 0.0006,
+ "loss": 5.896534442901611,
+ "step": 3319
+ },
+ {
+ "epoch": 46.1117903930131,
+ "grad_norm": 0.05795128643512726,
+ "learning_rate": 0.0006,
+ "loss": 5.843111991882324,
+ "step": 3320
+ },
+ {
+ "epoch": 46.12576419213974,
+ "grad_norm": 0.06374258548021317,
+ "learning_rate": 0.0006,
+ "loss": 5.78752326965332,
+ "step": 3321
+ },
+ {
+ "epoch": 46.13973799126637,
+ "grad_norm": 0.045524805784225464,
+ "learning_rate": 0.0006,
+ "loss": 5.86755895614624,
+ "step": 3322
+ },
+ {
+ "epoch": 46.15371179039301,
+ "grad_norm": 0.03773504123091698,
+ "learning_rate": 0.0006,
+ "loss": 5.83627462387085,
+ "step": 3323
+ },
+ {
+ "epoch": 46.16768558951965,
+ "grad_norm": 0.035815201699733734,
+ "learning_rate": 0.0006,
+ "loss": 5.798238277435303,
+ "step": 3324
+ },
+ {
+ "epoch": 46.18165938864629,
+ "grad_norm": 0.03756783902645111,
+ "learning_rate": 0.0006,
+ "loss": 5.845197677612305,
+ "step": 3325
+ },
+ {
+ "epoch": 46.19563318777293,
+ "grad_norm": 0.03569113463163376,
+ "learning_rate": 0.0006,
+ "loss": 5.888932704925537,
+ "step": 3326
+ },
+ {
+ "epoch": 46.209606986899566,
+ "grad_norm": 0.029467813670635223,
+ "learning_rate": 0.0006,
+ "loss": 5.792047500610352,
+ "step": 3327
+ },
+ {
+ "epoch": 46.223580786026204,
+ "grad_norm": 0.025698406621813774,
+ "learning_rate": 0.0006,
+ "loss": 5.783734321594238,
+ "step": 3328
+ },
+ {
+ "epoch": 46.237554585152836,
+ "grad_norm": 0.028426671400666237,
+ "learning_rate": 0.0006,
+ "loss": 5.762084007263184,
+ "step": 3329
+ },
+ {
+ "epoch": 46.251528384279474,
+ "grad_norm": 0.03099622018635273,
+ "learning_rate": 0.0006,
+ "loss": 5.719109058380127,
+ "step": 3330
+ },
+ {
+ "epoch": 46.26550218340611,
+ "grad_norm": 0.0310219656676054,
+ "learning_rate": 0.0006,
+ "loss": 5.737057209014893,
+ "step": 3331
+ },
+ {
+ "epoch": 46.27947598253275,
+ "grad_norm": 0.02713446132838726,
+ "learning_rate": 0.0006,
+ "loss": 5.758657455444336,
+ "step": 3332
+ },
+ {
+ "epoch": 46.29344978165939,
+ "grad_norm": 0.03099844604730606,
+ "learning_rate": 0.0006,
+ "loss": 5.742086410522461,
+ "step": 3333
+ },
+ {
+ "epoch": 46.30742358078603,
+ "grad_norm": 0.030258730053901672,
+ "learning_rate": 0.0006,
+ "loss": 5.767426490783691,
+ "step": 3334
+ },
+ {
+ "epoch": 46.32139737991266,
+ "grad_norm": 0.029034852981567383,
+ "learning_rate": 0.0006,
+ "loss": 5.651792049407959,
+ "step": 3335
+ },
+ {
+ "epoch": 46.3353711790393,
+ "grad_norm": 0.03138665854930878,
+ "learning_rate": 0.0006,
+ "loss": 5.664980888366699,
+ "step": 3336
+ },
+ {
+ "epoch": 46.34934497816594,
+ "grad_norm": 0.040546901524066925,
+ "learning_rate": 0.0006,
+ "loss": 5.665165901184082,
+ "step": 3337
+ },
+ {
+ "epoch": 46.36331877729258,
+ "grad_norm": 0.06395485252141953,
+ "learning_rate": 0.0006,
+ "loss": 5.7776713371276855,
+ "step": 3338
+ },
+ {
+ "epoch": 46.377292576419215,
+ "grad_norm": 0.10088056325912476,
+ "learning_rate": 0.0006,
+ "loss": 5.755645751953125,
+ "step": 3339
+ },
+ {
+ "epoch": 46.391266375545854,
+ "grad_norm": 0.09527938067913055,
+ "learning_rate": 0.0006,
+ "loss": 5.838809490203857,
+ "step": 3340
+ },
+ {
+ "epoch": 46.40524017467249,
+ "grad_norm": 0.0849648267030716,
+ "learning_rate": 0.0006,
+ "loss": 5.781932830810547,
+ "step": 3341
+ },
+ {
+ "epoch": 46.419213973799124,
+ "grad_norm": 0.10142666101455688,
+ "learning_rate": 0.0006,
+ "loss": 5.838970184326172,
+ "step": 3342
+ },
+ {
+ "epoch": 46.43318777292576,
+ "grad_norm": 0.42731454968452454,
+ "learning_rate": 0.0006,
+ "loss": 6.071902275085449,
+ "step": 3343
+ },
+ {
+ "epoch": 46.4471615720524,
+ "grad_norm": 0.13102488219738007,
+ "learning_rate": 0.0006,
+ "loss": 6.8635993003845215,
+ "step": 3344
+ },
+ {
+ "epoch": 46.46113537117904,
+ "grad_norm": 0.23691938817501068,
+ "learning_rate": 0.0006,
+ "loss": 7.313158988952637,
+ "step": 3345
+ },
+ {
+ "epoch": 46.47510917030568,
+ "grad_norm": 0.11670669168233871,
+ "learning_rate": 0.0006,
+ "loss": 7.240363121032715,
+ "step": 3346
+ },
+ {
+ "epoch": 46.48908296943232,
+ "grad_norm": 0.09985820949077606,
+ "learning_rate": 0.0006,
+ "loss": 7.113332271575928,
+ "step": 3347
+ },
+ {
+ "epoch": 46.50305676855895,
+ "grad_norm": 0.11406408250331879,
+ "learning_rate": 0.0006,
+ "loss": 7.051285743713379,
+ "step": 3348
+ },
+ {
+ "epoch": 46.51703056768559,
+ "grad_norm": 0.07900336384773254,
+ "learning_rate": 0.0006,
+ "loss": 7.014365196228027,
+ "step": 3349
+ },
+ {
+ "epoch": 46.531004366812226,
+ "grad_norm": 0.05380333960056305,
+ "learning_rate": 0.0006,
+ "loss": 6.8868818283081055,
+ "step": 3350
+ },
+ {
+ "epoch": 46.544978165938865,
+ "grad_norm": 0.05868449807167053,
+ "learning_rate": 0.0006,
+ "loss": 6.783599853515625,
+ "step": 3351
+ },
+ {
+ "epoch": 46.5589519650655,
+ "grad_norm": 0.04428785294294357,
+ "learning_rate": 0.0006,
+ "loss": 6.798511505126953,
+ "step": 3352
+ },
+ {
+ "epoch": 46.57292576419214,
+ "grad_norm": 0.055527277290821075,
+ "learning_rate": 0.0006,
+ "loss": 6.694507598876953,
+ "step": 3353
+ },
+ {
+ "epoch": 46.58689956331878,
+ "grad_norm": 0.03872432932257652,
+ "learning_rate": 0.0006,
+ "loss": 6.736631393432617,
+ "step": 3354
+ },
+ {
+ "epoch": 46.60087336244541,
+ "grad_norm": 0.04330779239535332,
+ "learning_rate": 0.0006,
+ "loss": 6.576806545257568,
+ "step": 3355
+ },
+ {
+ "epoch": 46.61484716157205,
+ "grad_norm": 0.036767736077308655,
+ "learning_rate": 0.0006,
+ "loss": 6.559871196746826,
+ "step": 3356
+ },
+ {
+ "epoch": 46.62882096069869,
+ "grad_norm": 0.039857108145952225,
+ "learning_rate": 0.0006,
+ "loss": 6.5749831199646,
+ "step": 3357
+ },
+ {
+ "epoch": 46.64279475982533,
+ "grad_norm": 0.048719972372055054,
+ "learning_rate": 0.0006,
+ "loss": 6.4915618896484375,
+ "step": 3358
+ },
+ {
+ "epoch": 46.65676855895197,
+ "grad_norm": 0.042849015444517136,
+ "learning_rate": 0.0006,
+ "loss": 6.413516044616699,
+ "step": 3359
+ },
+ {
+ "epoch": 46.670742358078606,
+ "grad_norm": 0.02871009334921837,
+ "learning_rate": 0.0006,
+ "loss": 6.320952415466309,
+ "step": 3360
+ },
+ {
+ "epoch": 46.68471615720524,
+ "grad_norm": 0.03193620592355728,
+ "learning_rate": 0.0006,
+ "loss": 6.4176435470581055,
+ "step": 3361
+ },
+ {
+ "epoch": 46.698689956331876,
+ "grad_norm": 0.027417024597525597,
+ "learning_rate": 0.0006,
+ "loss": 6.342319488525391,
+ "step": 3362
+ },
+ {
+ "epoch": 46.712663755458514,
+ "grad_norm": 0.03305937349796295,
+ "learning_rate": 0.0006,
+ "loss": 6.246112823486328,
+ "step": 3363
+ },
+ {
+ "epoch": 46.72663755458515,
+ "grad_norm": 0.031403183937072754,
+ "learning_rate": 0.0006,
+ "loss": 6.149503231048584,
+ "step": 3364
+ },
+ {
+ "epoch": 46.74061135371179,
+ "grad_norm": 0.031563788652420044,
+ "learning_rate": 0.0006,
+ "loss": 6.278808116912842,
+ "step": 3365
+ },
+ {
+ "epoch": 46.75458515283843,
+ "grad_norm": 0.028366589918732643,
+ "learning_rate": 0.0006,
+ "loss": 6.057036876678467,
+ "step": 3366
+ },
+ {
+ "epoch": 46.76855895196506,
+ "grad_norm": 0.023715803399682045,
+ "learning_rate": 0.0006,
+ "loss": 6.111324310302734,
+ "step": 3367
+ },
+ {
+ "epoch": 46.7825327510917,
+ "grad_norm": 0.026546811684966087,
+ "learning_rate": 0.0006,
+ "loss": 6.105669975280762,
+ "step": 3368
+ },
+ {
+ "epoch": 46.79650655021834,
+ "grad_norm": 0.023531019687652588,
+ "learning_rate": 0.0006,
+ "loss": 6.068072319030762,
+ "step": 3369
+ },
+ {
+ "epoch": 46.81048034934498,
+ "grad_norm": 0.024480128660798073,
+ "learning_rate": 0.0006,
+ "loss": 5.947896957397461,
+ "step": 3370
+ },
+ {
+ "epoch": 46.82445414847162,
+ "grad_norm": 0.02517310529947281,
+ "learning_rate": 0.0006,
+ "loss": 6.030037879943848,
+ "step": 3371
+ },
+ {
+ "epoch": 46.838427947598255,
+ "grad_norm": 0.026723403483629227,
+ "learning_rate": 0.0006,
+ "loss": 5.963776111602783,
+ "step": 3372
+ },
+ {
+ "epoch": 46.852401746724894,
+ "grad_norm": 0.03163504600524902,
+ "learning_rate": 0.0006,
+ "loss": 5.893466949462891,
+ "step": 3373
+ },
+ {
+ "epoch": 46.866375545851525,
+ "grad_norm": 0.04058683291077614,
+ "learning_rate": 0.0006,
+ "loss": 5.9202799797058105,
+ "step": 3374
+ },
+ {
+ "epoch": 46.880349344978164,
+ "grad_norm": 0.04964117333292961,
+ "learning_rate": 0.0006,
+ "loss": 5.9366559982299805,
+ "step": 3375
+ },
+ {
+ "epoch": 46.8943231441048,
+ "grad_norm": 0.055406033992767334,
+ "learning_rate": 0.0006,
+ "loss": 5.912230968475342,
+ "step": 3376
+ },
+ {
+ "epoch": 46.90829694323144,
+ "grad_norm": 0.04643625393509865,
+ "learning_rate": 0.0006,
+ "loss": 5.87260103225708,
+ "step": 3377
+ },
+ {
+ "epoch": 46.92227074235808,
+ "grad_norm": 0.029325876384973526,
+ "learning_rate": 0.0006,
+ "loss": 5.854763984680176,
+ "step": 3378
+ },
+ {
+ "epoch": 46.93624454148472,
+ "grad_norm": 0.03159382566809654,
+ "learning_rate": 0.0006,
+ "loss": 5.880882263183594,
+ "step": 3379
+ },
+ {
+ "epoch": 46.95021834061135,
+ "grad_norm": 0.030224012210965157,
+ "learning_rate": 0.0006,
+ "loss": 5.787360191345215,
+ "step": 3380
+ },
+ {
+ "epoch": 46.96419213973799,
+ "grad_norm": 0.028368130326271057,
+ "learning_rate": 0.0006,
+ "loss": 5.817195892333984,
+ "step": 3381
+ },
+ {
+ "epoch": 46.97816593886463,
+ "grad_norm": 0.029411399737000465,
+ "learning_rate": 0.0006,
+ "loss": 5.783403396606445,
+ "step": 3382
+ },
+ {
+ "epoch": 46.992139737991266,
+ "grad_norm": 0.029927462339401245,
+ "learning_rate": 0.0006,
+ "loss": 5.662444591522217,
+ "step": 3383
+ },
+ {
+ "epoch": 47.0,
+ "grad_norm": 0.02872961387038231,
+ "learning_rate": 0.0006,
+ "loss": 5.742781639099121,
+ "step": 3384
+ },
+ {
+ "epoch": 47.0,
+ "eval_loss": 5.870260238647461,
+ "eval_runtime": 57.3734,
+ "eval_samples_per_second": 42.563,
+ "eval_steps_per_second": 1.342,
+ "step": 3384
+ },
+ {
+ "epoch": 47.01397379912664,
+ "grad_norm": 0.028542131185531616,
+ "learning_rate": 0.0006,
+ "loss": 5.753256797790527,
+ "step": 3385
+ },
+ {
+ "epoch": 47.02794759825328,
+ "grad_norm": 0.03227119892835617,
+ "learning_rate": 0.0006,
+ "loss": 5.722246170043945,
+ "step": 3386
+ },
+ {
+ "epoch": 47.041921397379916,
+ "grad_norm": 0.031461723148822784,
+ "learning_rate": 0.0006,
+ "loss": 5.740520000457764,
+ "step": 3387
+ },
+ {
+ "epoch": 47.05589519650655,
+ "grad_norm": 0.03184220567345619,
+ "learning_rate": 0.0006,
+ "loss": 5.720890045166016,
+ "step": 3388
+ },
+ {
+ "epoch": 47.069868995633186,
+ "grad_norm": 0.030015580356121063,
+ "learning_rate": 0.0006,
+ "loss": 5.597814559936523,
+ "step": 3389
+ },
+ {
+ "epoch": 47.083842794759825,
+ "grad_norm": 0.029931779950857162,
+ "learning_rate": 0.0006,
+ "loss": 5.5412468910217285,
+ "step": 3390
+ },
+ {
+ "epoch": 47.09781659388646,
+ "grad_norm": 0.03206505626440048,
+ "learning_rate": 0.0006,
+ "loss": 5.6967597007751465,
+ "step": 3391
+ },
+ {
+ "epoch": 47.1117903930131,
+ "grad_norm": 0.03978161886334419,
+ "learning_rate": 0.0006,
+ "loss": 5.59661340713501,
+ "step": 3392
+ },
+ {
+ "epoch": 47.12576419213974,
+ "grad_norm": 0.04619080200791359,
+ "learning_rate": 0.0006,
+ "loss": 5.795788764953613,
+ "step": 3393
+ },
+ {
+ "epoch": 47.13973799126637,
+ "grad_norm": 0.049085382372140884,
+ "learning_rate": 0.0006,
+ "loss": 5.667335510253906,
+ "step": 3394
+ },
+ {
+ "epoch": 47.15371179039301,
+ "grad_norm": 0.04830858111381531,
+ "learning_rate": 0.0006,
+ "loss": 5.668806552886963,
+ "step": 3395
+ },
+ {
+ "epoch": 47.16768558951965,
+ "grad_norm": 0.0583178736269474,
+ "learning_rate": 0.0006,
+ "loss": 5.692285537719727,
+ "step": 3396
+ },
+ {
+ "epoch": 47.18165938864629,
+ "grad_norm": 0.05182253196835518,
+ "learning_rate": 0.0006,
+ "loss": 5.678645133972168,
+ "step": 3397
+ },
+ {
+ "epoch": 47.19563318777293,
+ "grad_norm": 0.04011905938386917,
+ "learning_rate": 0.0006,
+ "loss": 5.661550998687744,
+ "step": 3398
+ },
+ {
+ "epoch": 47.209606986899566,
+ "grad_norm": 0.03888486698269844,
+ "learning_rate": 0.0006,
+ "loss": 5.643001556396484,
+ "step": 3399
+ },
+ {
+ "epoch": 47.223580786026204,
+ "grad_norm": 0.030658265575766563,
+ "learning_rate": 0.0006,
+ "loss": 5.659090995788574,
+ "step": 3400
+ },
+ {
+ "epoch": 47.237554585152836,
+ "grad_norm": 0.03609256446361542,
+ "learning_rate": 0.0006,
+ "loss": 5.544501781463623,
+ "step": 3401
+ },
+ {
+ "epoch": 47.251528384279474,
+ "grad_norm": 0.033589769154787064,
+ "learning_rate": 0.0006,
+ "loss": 5.463878631591797,
+ "step": 3402
+ },
+ {
+ "epoch": 47.26550218340611,
+ "grad_norm": 0.035128287971019745,
+ "learning_rate": 0.0006,
+ "loss": 5.5680060386657715,
+ "step": 3403
+ },
+ {
+ "epoch": 47.27947598253275,
+ "grad_norm": 0.038684550672769547,
+ "learning_rate": 0.0006,
+ "loss": 5.587898254394531,
+ "step": 3404
+ },
+ {
+ "epoch": 47.29344978165939,
+ "grad_norm": 0.043566226959228516,
+ "learning_rate": 0.0006,
+ "loss": 5.580896854400635,
+ "step": 3405
+ },
+ {
+ "epoch": 47.30742358078603,
+ "grad_norm": 0.04428374394774437,
+ "learning_rate": 0.0006,
+ "loss": 5.603460311889648,
+ "step": 3406
+ },
+ {
+ "epoch": 47.32139737991266,
+ "grad_norm": 0.037368934601545334,
+ "learning_rate": 0.0006,
+ "loss": 5.564740180969238,
+ "step": 3407
+ },
+ {
+ "epoch": 47.3353711790393,
+ "grad_norm": 0.04092350974678993,
+ "learning_rate": 0.0006,
+ "loss": 5.584153652191162,
+ "step": 3408
+ },
+ {
+ "epoch": 47.34934497816594,
+ "grad_norm": 0.045031506568193436,
+ "learning_rate": 0.0006,
+ "loss": 5.601186275482178,
+ "step": 3409
+ },
+ {
+ "epoch": 47.36331877729258,
+ "grad_norm": 0.041830409318208694,
+ "learning_rate": 0.0006,
+ "loss": 5.512788772583008,
+ "step": 3410
+ },
+ {
+ "epoch": 47.377292576419215,
+ "grad_norm": 0.037633974105119705,
+ "learning_rate": 0.0006,
+ "loss": 5.503828525543213,
+ "step": 3411
+ },
+ {
+ "epoch": 47.391266375545854,
+ "grad_norm": 0.03473588079214096,
+ "learning_rate": 0.0006,
+ "loss": 5.570557594299316,
+ "step": 3412
+ },
+ {
+ "epoch": 47.40524017467249,
+ "grad_norm": 0.039325181394815445,
+ "learning_rate": 0.0006,
+ "loss": 5.427783966064453,
+ "step": 3413
+ },
+ {
+ "epoch": 47.419213973799124,
+ "grad_norm": 0.04034624621272087,
+ "learning_rate": 0.0006,
+ "loss": 5.572591304779053,
+ "step": 3414
+ },
+ {
+ "epoch": 47.43318777292576,
+ "grad_norm": 0.05245184153318405,
+ "learning_rate": 0.0006,
+ "loss": 5.505208969116211,
+ "step": 3415
+ },
+ {
+ "epoch": 47.4471615720524,
+ "grad_norm": 0.0568675734102726,
+ "learning_rate": 0.0006,
+ "loss": 5.589373588562012,
+ "step": 3416
+ },
+ {
+ "epoch": 47.46113537117904,
+ "grad_norm": 0.048336587846279144,
+ "learning_rate": 0.0006,
+ "loss": 5.522005081176758,
+ "step": 3417
+ },
+ {
+ "epoch": 47.47510917030568,
+ "grad_norm": 0.04194975271821022,
+ "learning_rate": 0.0006,
+ "loss": 5.387576103210449,
+ "step": 3418
+ },
+ {
+ "epoch": 47.48908296943232,
+ "grad_norm": 0.04019433259963989,
+ "learning_rate": 0.0006,
+ "loss": 5.522615432739258,
+ "step": 3419
+ },
+ {
+ "epoch": 47.50305676855895,
+ "grad_norm": 0.05709933117032051,
+ "learning_rate": 0.0006,
+ "loss": 5.420286178588867,
+ "step": 3420
+ },
+ {
+ "epoch": 47.51703056768559,
+ "grad_norm": 0.09130338579416275,
+ "learning_rate": 0.0006,
+ "loss": 5.484622478485107,
+ "step": 3421
+ },
+ {
+ "epoch": 47.531004366812226,
+ "grad_norm": 0.2757541835308075,
+ "learning_rate": 0.0006,
+ "loss": 5.821832656860352,
+ "step": 3422
+ },
+ {
+ "epoch": 47.544978165938865,
+ "grad_norm": 0.28292080760002136,
+ "learning_rate": 0.0006,
+ "loss": 6.579934120178223,
+ "step": 3423
+ },
+ {
+ "epoch": 47.5589519650655,
+ "grad_norm": 0.15293604135513306,
+ "learning_rate": 0.0006,
+ "loss": 6.478213310241699,
+ "step": 3424
+ },
+ {
+ "epoch": 47.57292576419214,
+ "grad_norm": 0.11284393817186356,
+ "learning_rate": 0.0006,
+ "loss": 6.4066386222839355,
+ "step": 3425
+ },
+ {
+ "epoch": 47.58689956331878,
+ "grad_norm": 0.10933073610067368,
+ "learning_rate": 0.0006,
+ "loss": 6.3306169509887695,
+ "step": 3426
+ },
+ {
+ "epoch": 47.60087336244541,
+ "grad_norm": 0.05837641656398773,
+ "learning_rate": 0.0006,
+ "loss": 6.250000953674316,
+ "step": 3427
+ },
+ {
+ "epoch": 47.61484716157205,
+ "grad_norm": 0.05987229570746422,
+ "learning_rate": 0.0006,
+ "loss": 6.152463436126709,
+ "step": 3428
+ },
+ {
+ "epoch": 47.62882096069869,
+ "grad_norm": 0.04064938426017761,
+ "learning_rate": 0.0006,
+ "loss": 6.117897033691406,
+ "step": 3429
+ },
+ {
+ "epoch": 47.64279475982533,
+ "grad_norm": 0.04565057530999184,
+ "learning_rate": 0.0006,
+ "loss": 6.042291641235352,
+ "step": 3430
+ },
+ {
+ "epoch": 47.65676855895197,
+ "grad_norm": 0.03475063294172287,
+ "learning_rate": 0.0006,
+ "loss": 6.0071868896484375,
+ "step": 3431
+ },
+ {
+ "epoch": 47.670742358078606,
+ "grad_norm": 0.03809220343828201,
+ "learning_rate": 0.0006,
+ "loss": 6.0128021240234375,
+ "step": 3432
+ },
+ {
+ "epoch": 47.68471615720524,
+ "grad_norm": 0.03307219594717026,
+ "learning_rate": 0.0006,
+ "loss": 5.894664287567139,
+ "step": 3433
+ },
+ {
+ "epoch": 47.698689956331876,
+ "grad_norm": 0.028824346140027046,
+ "learning_rate": 0.0006,
+ "loss": 5.914865493774414,
+ "step": 3434
+ },
+ {
+ "epoch": 47.712663755458514,
+ "grad_norm": 0.027923690155148506,
+ "learning_rate": 0.0006,
+ "loss": 5.851759910583496,
+ "step": 3435
+ },
+ {
+ "epoch": 47.72663755458515,
+ "grad_norm": 0.025492684915661812,
+ "learning_rate": 0.0006,
+ "loss": 5.749186038970947,
+ "step": 3436
+ },
+ {
+ "epoch": 47.74061135371179,
+ "grad_norm": 0.025105174630880356,
+ "learning_rate": 0.0006,
+ "loss": 5.803566932678223,
+ "step": 3437
+ },
+ {
+ "epoch": 47.75458515283843,
+ "grad_norm": 0.023298880085349083,
+ "learning_rate": 0.0006,
+ "loss": 5.76276159286499,
+ "step": 3438
+ },
+ {
+ "epoch": 47.76855895196506,
+ "grad_norm": 0.02468576282262802,
+ "learning_rate": 0.0006,
+ "loss": 5.833852767944336,
+ "step": 3439
+ },
+ {
+ "epoch": 47.7825327510917,
+ "grad_norm": 0.022730689495801926,
+ "learning_rate": 0.0006,
+ "loss": 5.736724376678467,
+ "step": 3440
+ },
+ {
+ "epoch": 47.79650655021834,
+ "grad_norm": 0.021335354074835777,
+ "learning_rate": 0.0006,
+ "loss": 5.644248008728027,
+ "step": 3441
+ },
+ {
+ "epoch": 47.81048034934498,
+ "grad_norm": 0.02288101427257061,
+ "learning_rate": 0.0006,
+ "loss": 5.7011942863464355,
+ "step": 3442
+ },
+ {
+ "epoch": 47.82445414847162,
+ "grad_norm": 0.01961585320532322,
+ "learning_rate": 0.0006,
+ "loss": 5.63215446472168,
+ "step": 3443
+ },
+ {
+ "epoch": 47.838427947598255,
+ "grad_norm": 0.017659205943346024,
+ "learning_rate": 0.0006,
+ "loss": 5.744845390319824,
+ "step": 3444
+ },
+ {
+ "epoch": 47.852401746724894,
+ "grad_norm": 0.02281784638762474,
+ "learning_rate": 0.0006,
+ "loss": 5.696993350982666,
+ "step": 3445
+ },
+ {
+ "epoch": 47.866375545851525,
+ "grad_norm": 0.016394242644309998,
+ "learning_rate": 0.0006,
+ "loss": 5.659870147705078,
+ "step": 3446
+ },
+ {
+ "epoch": 47.880349344978164,
+ "grad_norm": 0.01967368647456169,
+ "learning_rate": 0.0006,
+ "loss": 5.638072490692139,
+ "step": 3447
+ },
+ {
+ "epoch": 47.8943231441048,
+ "grad_norm": 0.017326267436146736,
+ "learning_rate": 0.0006,
+ "loss": 5.638836860656738,
+ "step": 3448
+ },
+ {
+ "epoch": 47.90829694323144,
+ "grad_norm": 0.017778463661670685,
+ "learning_rate": 0.0006,
+ "loss": 5.634466648101807,
+ "step": 3449
+ },
+ {
+ "epoch": 47.92227074235808,
+ "grad_norm": 0.017269758507609367,
+ "learning_rate": 0.0006,
+ "loss": 5.506277561187744,
+ "step": 3450
+ },
+ {
+ "epoch": 47.93624454148472,
+ "grad_norm": 0.01733289286494255,
+ "learning_rate": 0.0006,
+ "loss": 5.544455528259277,
+ "step": 3451
+ },
+ {
+ "epoch": 47.95021834061135,
+ "grad_norm": 0.017787795513868332,
+ "learning_rate": 0.0006,
+ "loss": 5.498943328857422,
+ "step": 3452
+ },
+ {
+ "epoch": 47.96419213973799,
+ "grad_norm": 0.015770932659506798,
+ "learning_rate": 0.0006,
+ "loss": 5.537227630615234,
+ "step": 3453
+ },
+ {
+ "epoch": 47.97816593886463,
+ "grad_norm": 0.016132809221744537,
+ "learning_rate": 0.0006,
+ "loss": 5.489081382751465,
+ "step": 3454
+ },
+ {
+ "epoch": 47.992139737991266,
+ "grad_norm": 0.01749664731323719,
+ "learning_rate": 0.0006,
+ "loss": 5.454263687133789,
+ "step": 3455
+ },
+ {
+ "epoch": 48.0,
+ "grad_norm": 0.016711296513676643,
+ "learning_rate": 0.0006,
+ "loss": 5.530783653259277,
+ "step": 3456
+ },
+ {
+ "epoch": 48.0,
+ "eval_loss": 5.596925258636475,
+ "eval_runtime": 56.9612,
+ "eval_samples_per_second": 42.871,
+ "eval_steps_per_second": 1.352,
+ "step": 3456
+ },
+ {
+ "epoch": 48.01397379912664,
+ "grad_norm": 0.01867641694843769,
+ "learning_rate": 0.0006,
+ "loss": 5.506853103637695,
+ "step": 3457
+ },
+ {
+ "epoch": 48.02794759825328,
+ "grad_norm": 0.023250792175531387,
+ "learning_rate": 0.0006,
+ "loss": 5.4074177742004395,
+ "step": 3458
+ },
+ {
+ "epoch": 48.041921397379916,
+ "grad_norm": 0.01970936730504036,
+ "learning_rate": 0.0006,
+ "loss": 5.508777618408203,
+ "step": 3459
+ },
+ {
+ "epoch": 48.05589519650655,
+ "grad_norm": 0.01879260502755642,
+ "learning_rate": 0.0006,
+ "loss": 5.412424087524414,
+ "step": 3460
+ },
+ {
+ "epoch": 48.069868995633186,
+ "grad_norm": 0.017420092597603798,
+ "learning_rate": 0.0006,
+ "loss": 5.43361759185791,
+ "step": 3461
+ },
+ {
+ "epoch": 48.083842794759825,
+ "grad_norm": 0.018707862123847008,
+ "learning_rate": 0.0006,
+ "loss": 5.476800918579102,
+ "step": 3462
+ },
+ {
+ "epoch": 48.09781659388646,
+ "grad_norm": 0.023185882717370987,
+ "learning_rate": 0.0006,
+ "loss": 5.465272426605225,
+ "step": 3463
+ },
+ {
+ "epoch": 48.1117903930131,
+ "grad_norm": 0.023570368066430092,
+ "learning_rate": 0.0006,
+ "loss": 5.391471862792969,
+ "step": 3464
+ },
+ {
+ "epoch": 48.12576419213974,
+ "grad_norm": 0.02651144750416279,
+ "learning_rate": 0.0006,
+ "loss": 5.422799110412598,
+ "step": 3465
+ },
+ {
+ "epoch": 48.13973799126637,
+ "grad_norm": 0.03316329047083855,
+ "learning_rate": 0.0006,
+ "loss": 5.3952202796936035,
+ "step": 3466
+ },
+ {
+ "epoch": 48.15371179039301,
+ "grad_norm": 0.04012278467416763,
+ "learning_rate": 0.0006,
+ "loss": 5.540853023529053,
+ "step": 3467
+ },
+ {
+ "epoch": 48.16768558951965,
+ "grad_norm": 0.04763203114271164,
+ "learning_rate": 0.0006,
+ "loss": 5.387890338897705,
+ "step": 3468
+ },
+ {
+ "epoch": 48.18165938864629,
+ "grad_norm": 0.05129165202379227,
+ "learning_rate": 0.0006,
+ "loss": 5.414722442626953,
+ "step": 3469
+ },
+ {
+ "epoch": 48.19563318777293,
+ "grad_norm": 0.04438071325421333,
+ "learning_rate": 0.0006,
+ "loss": 5.386244773864746,
+ "step": 3470
+ },
+ {
+ "epoch": 48.209606986899566,
+ "grad_norm": 0.040234919637441635,
+ "learning_rate": 0.0006,
+ "loss": 5.431617736816406,
+ "step": 3471
+ },
+ {
+ "epoch": 48.223580786026204,
+ "grad_norm": 0.03217008337378502,
+ "learning_rate": 0.0006,
+ "loss": 5.30168342590332,
+ "step": 3472
+ },
+ {
+ "epoch": 48.237554585152836,
+ "grad_norm": 0.03087974339723587,
+ "learning_rate": 0.0006,
+ "loss": 5.316389083862305,
+ "step": 3473
+ },
+ {
+ "epoch": 48.251528384279474,
+ "grad_norm": 0.02789347991347313,
+ "learning_rate": 0.0006,
+ "loss": 5.385970592498779,
+ "step": 3474
+ },
+ {
+ "epoch": 48.26550218340611,
+ "grad_norm": 0.024792097508907318,
+ "learning_rate": 0.0006,
+ "loss": 5.362329959869385,
+ "step": 3475
+ },
+ {
+ "epoch": 48.27947598253275,
+ "grad_norm": 0.025869522243738174,
+ "learning_rate": 0.0006,
+ "loss": 5.28400182723999,
+ "step": 3476
+ },
+ {
+ "epoch": 48.29344978165939,
+ "grad_norm": 0.021339882165193558,
+ "learning_rate": 0.0006,
+ "loss": 5.36669921875,
+ "step": 3477
+ },
+ {
+ "epoch": 48.30742358078603,
+ "grad_norm": 0.021462520584464073,
+ "learning_rate": 0.0006,
+ "loss": 5.311222076416016,
+ "step": 3478
+ },
+ {
+ "epoch": 48.32139737991266,
+ "grad_norm": 0.02193879708647728,
+ "learning_rate": 0.0006,
+ "loss": 5.291441917419434,
+ "step": 3479
+ },
+ {
+ "epoch": 48.3353711790393,
+ "grad_norm": 0.01963093690574169,
+ "learning_rate": 0.0006,
+ "loss": 5.342780113220215,
+ "step": 3480
+ },
+ {
+ "epoch": 48.34934497816594,
+ "grad_norm": 0.020634736865758896,
+ "learning_rate": 0.0006,
+ "loss": 5.144141674041748,
+ "step": 3481
+ },
+ {
+ "epoch": 48.36331877729258,
+ "grad_norm": 0.02126968279480934,
+ "learning_rate": 0.0006,
+ "loss": 5.232624530792236,
+ "step": 3482
+ },
+ {
+ "epoch": 48.377292576419215,
+ "grad_norm": 0.021203994750976562,
+ "learning_rate": 0.0006,
+ "loss": 5.290219783782959,
+ "step": 3483
+ },
+ {
+ "epoch": 48.391266375545854,
+ "grad_norm": 0.02193225920200348,
+ "learning_rate": 0.0006,
+ "loss": 5.29197359085083,
+ "step": 3484
+ },
+ {
+ "epoch": 48.40524017467249,
+ "grad_norm": 0.023174211382865906,
+ "learning_rate": 0.0006,
+ "loss": 5.204349517822266,
+ "step": 3485
+ },
+ {
+ "epoch": 48.419213973799124,
+ "grad_norm": 0.024161914363503456,
+ "learning_rate": 0.0006,
+ "loss": 5.339351654052734,
+ "step": 3486
+ },
+ {
+ "epoch": 48.43318777292576,
+ "grad_norm": 0.02221524342894554,
+ "learning_rate": 0.0006,
+ "loss": 5.263001441955566,
+ "step": 3487
+ },
+ {
+ "epoch": 48.4471615720524,
+ "grad_norm": 0.026632830500602722,
+ "learning_rate": 0.0006,
+ "loss": 5.221281051635742,
+ "step": 3488
+ },
+ {
+ "epoch": 48.46113537117904,
+ "grad_norm": 0.034923870116472244,
+ "learning_rate": 0.0006,
+ "loss": 5.296039581298828,
+ "step": 3489
+ },
+ {
+ "epoch": 48.47510917030568,
+ "grad_norm": 0.046184979379177094,
+ "learning_rate": 0.0006,
+ "loss": 5.212769985198975,
+ "step": 3490
+ },
+ {
+ "epoch": 48.48908296943232,
+ "grad_norm": 0.05154098570346832,
+ "learning_rate": 0.0006,
+ "loss": 5.337864875793457,
+ "step": 3491
+ },
+ {
+ "epoch": 48.50305676855895,
+ "grad_norm": 0.042967259883880615,
+ "learning_rate": 0.0006,
+ "loss": 5.267958641052246,
+ "step": 3492
+ },
+ {
+ "epoch": 48.51703056768559,
+ "grad_norm": 0.05356583371758461,
+ "learning_rate": 0.0006,
+ "loss": 5.1404218673706055,
+ "step": 3493
+ },
+ {
+ "epoch": 48.531004366812226,
+ "grad_norm": 0.08219032734632492,
+ "learning_rate": 0.0006,
+ "loss": 5.300687789916992,
+ "step": 3494
+ },
+ {
+ "epoch": 48.544978165938865,
+ "grad_norm": 0.19204089045524597,
+ "learning_rate": 0.0006,
+ "loss": 5.456025123596191,
+ "step": 3495
+ },
+ {
+ "epoch": 48.5589519650655,
+ "grad_norm": 0.25917696952819824,
+ "learning_rate": 0.0006,
+ "loss": 6.156761169433594,
+ "step": 3496
+ },
+ {
+ "epoch": 48.57292576419214,
+ "grad_norm": 0.1566283404827118,
+ "learning_rate": 0.0006,
+ "loss": 5.862346649169922,
+ "step": 3497
+ },
+ {
+ "epoch": 48.58689956331878,
+ "grad_norm": 0.1761486828327179,
+ "learning_rate": 0.0006,
+ "loss": 6.012795448303223,
+ "step": 3498
+ },
+ {
+ "epoch": 48.60087336244541,
+ "grad_norm": 0.09857621043920517,
+ "learning_rate": 0.0006,
+ "loss": 5.9248151779174805,
+ "step": 3499
+ },
+ {
+ "epoch": 48.61484716157205,
+ "grad_norm": 0.05968055501580238,
+ "learning_rate": 0.0006,
+ "loss": 5.780706882476807,
+ "step": 3500
+ },
+ {
+ "epoch": 48.62882096069869,
+ "grad_norm": 0.057433564215898514,
+ "learning_rate": 0.0006,
+ "loss": 5.645388603210449,
+ "step": 3501
+ },
+ {
+ "epoch": 48.64279475982533,
+ "grad_norm": 0.039659615606069565,
+ "learning_rate": 0.0006,
+ "loss": 5.695138931274414,
+ "step": 3502
+ },
+ {
+ "epoch": 48.65676855895197,
+ "grad_norm": 0.04019695147871971,
+ "learning_rate": 0.0006,
+ "loss": 5.672435283660889,
+ "step": 3503
+ },
+ {
+ "epoch": 48.670742358078606,
+ "grad_norm": 0.03033018670976162,
+ "learning_rate": 0.0006,
+ "loss": 5.682260513305664,
+ "step": 3504
+ },
+ {
+ "epoch": 48.68471615720524,
+ "grad_norm": 0.03215787187218666,
+ "learning_rate": 0.0006,
+ "loss": 5.604182243347168,
+ "step": 3505
+ },
+ {
+ "epoch": 48.698689956331876,
+ "grad_norm": 0.026865126565098763,
+ "learning_rate": 0.0006,
+ "loss": 5.64040470123291,
+ "step": 3506
+ },
+ {
+ "epoch": 48.712663755458514,
+ "grad_norm": 0.02831609919667244,
+ "learning_rate": 0.0006,
+ "loss": 5.552720069885254,
+ "step": 3507
+ },
+ {
+ "epoch": 48.72663755458515,
+ "grad_norm": 0.024226613342761993,
+ "learning_rate": 0.0006,
+ "loss": 5.556946277618408,
+ "step": 3508
+ },
+ {
+ "epoch": 48.74061135371179,
+ "grad_norm": 0.025393934920430183,
+ "learning_rate": 0.0006,
+ "loss": 5.476178169250488,
+ "step": 3509
+ },
+ {
+ "epoch": 48.75458515283843,
+ "grad_norm": 0.01984412968158722,
+ "learning_rate": 0.0006,
+ "loss": 5.471088409423828,
+ "step": 3510
+ },
+ {
+ "epoch": 48.76855895196506,
+ "grad_norm": 0.02035129815340042,
+ "learning_rate": 0.0006,
+ "loss": 5.438812732696533,
+ "step": 3511
+ },
+ {
+ "epoch": 48.7825327510917,
+ "grad_norm": 0.0208350270986557,
+ "learning_rate": 0.0006,
+ "loss": 5.381746292114258,
+ "step": 3512
+ },
+ {
+ "epoch": 48.79650655021834,
+ "grad_norm": 0.017783882096409798,
+ "learning_rate": 0.0006,
+ "loss": 5.351675510406494,
+ "step": 3513
+ },
+ {
+ "epoch": 48.81048034934498,
+ "grad_norm": 0.01671770215034485,
+ "learning_rate": 0.0006,
+ "loss": 5.329428672790527,
+ "step": 3514
+ },
+ {
+ "epoch": 48.82445414847162,
+ "grad_norm": 0.017891215160489082,
+ "learning_rate": 0.0006,
+ "loss": 5.212074279785156,
+ "step": 3515
+ },
+ {
+ "epoch": 48.838427947598255,
+ "grad_norm": 0.016847791150212288,
+ "learning_rate": 0.0006,
+ "loss": 5.299534797668457,
+ "step": 3516
+ },
+ {
+ "epoch": 48.852401746724894,
+ "grad_norm": 0.01593637838959694,
+ "learning_rate": 0.0006,
+ "loss": 5.30466890335083,
+ "step": 3517
+ },
+ {
+ "epoch": 48.866375545851525,
+ "grad_norm": 0.014713102020323277,
+ "learning_rate": 0.0006,
+ "loss": 5.294345378875732,
+ "step": 3518
+ },
+ {
+ "epoch": 48.880349344978164,
+ "grad_norm": 0.01433857623487711,
+ "learning_rate": 0.0006,
+ "loss": 5.2148542404174805,
+ "step": 3519
+ },
+ {
+ "epoch": 48.8943231441048,
+ "grad_norm": 0.014979379251599312,
+ "learning_rate": 0.0006,
+ "loss": 5.298582077026367,
+ "step": 3520
+ },
+ {
+ "epoch": 48.90829694323144,
+ "grad_norm": 0.013494192622601986,
+ "learning_rate": 0.0006,
+ "loss": 5.309783935546875,
+ "step": 3521
+ },
+ {
+ "epoch": 48.92227074235808,
+ "grad_norm": 0.014020012691617012,
+ "learning_rate": 0.0006,
+ "loss": 5.284879684448242,
+ "step": 3522
+ },
+ {
+ "epoch": 48.93624454148472,
+ "grad_norm": 0.01348058134317398,
+ "learning_rate": 0.0006,
+ "loss": 5.249116897583008,
+ "step": 3523
+ },
+ {
+ "epoch": 48.95021834061135,
+ "grad_norm": 0.013280218467116356,
+ "learning_rate": 0.0006,
+ "loss": 5.273489952087402,
+ "step": 3524
+ },
+ {
+ "epoch": 48.96419213973799,
+ "grad_norm": 0.012647110037505627,
+ "learning_rate": 0.0006,
+ "loss": 5.306652069091797,
+ "step": 3525
+ },
+ {
+ "epoch": 48.97816593886463,
+ "grad_norm": 0.012847724370658398,
+ "learning_rate": 0.0006,
+ "loss": 5.235745429992676,
+ "step": 3526
+ },
+ {
+ "epoch": 48.992139737991266,
+ "grad_norm": 0.011820166371762753,
+ "learning_rate": 0.0006,
+ "loss": 5.238091468811035,
+ "step": 3527
+ },
+ {
+ "epoch": 49.0,
+ "grad_norm": 0.01255585066974163,
+ "learning_rate": 0.0006,
+ "loss": 5.231295585632324,
+ "step": 3528
+ },
+ {
+ "epoch": 49.0,
+ "eval_loss": 5.394587993621826,
+ "eval_runtime": 56.3734,
+ "eval_samples_per_second": 43.318,
+ "eval_steps_per_second": 1.366,
+ "step": 3528
+ },
+ {
+ "epoch": 49.01397379912664,
+ "grad_norm": 0.011031972244381905,
+ "learning_rate": 0.0006,
+ "loss": 5.1354570388793945,
+ "step": 3529
+ },
+ {
+ "epoch": 49.02794759825328,
+ "grad_norm": 0.010631263256072998,
+ "learning_rate": 0.0006,
+ "loss": 5.177639961242676,
+ "step": 3530
+ },
+ {
+ "epoch": 49.041921397379916,
+ "grad_norm": 0.01133775431662798,
+ "learning_rate": 0.0006,
+ "loss": 5.283428192138672,
+ "step": 3531
+ },
+ {
+ "epoch": 49.05589519650655,
+ "grad_norm": 0.01154172234237194,
+ "learning_rate": 0.0006,
+ "loss": 5.206364154815674,
+ "step": 3532
+ },
+ {
+ "epoch": 49.069868995633186,
+ "grad_norm": 0.010192356072366238,
+ "learning_rate": 0.0006,
+ "loss": 5.2385735511779785,
+ "step": 3533
+ },
+ {
+ "epoch": 49.083842794759825,
+ "grad_norm": 0.012789200060069561,
+ "learning_rate": 0.0006,
+ "loss": 5.071800231933594,
+ "step": 3534
+ },
+ {
+ "epoch": 49.09781659388646,
+ "grad_norm": 0.011282820254564285,
+ "learning_rate": 0.0006,
+ "loss": 5.218020439147949,
+ "step": 3535
+ },
+ {
+ "epoch": 49.1117903930131,
+ "grad_norm": 0.012153448536992073,
+ "learning_rate": 0.0006,
+ "loss": 5.134698867797852,
+ "step": 3536
+ },
+ {
+ "epoch": 49.12576419213974,
+ "grad_norm": 0.011920543387532234,
+ "learning_rate": 0.0006,
+ "loss": 5.141620635986328,
+ "step": 3537
+ },
+ {
+ "epoch": 49.13973799126637,
+ "grad_norm": 0.010949329473078251,
+ "learning_rate": 0.0006,
+ "loss": 5.116259574890137,
+ "step": 3538
+ },
+ {
+ "epoch": 49.15371179039301,
+ "grad_norm": 0.012362448498606682,
+ "learning_rate": 0.0006,
+ "loss": 5.090177059173584,
+ "step": 3539
+ },
+ {
+ "epoch": 49.16768558951965,
+ "grad_norm": 0.011575071141123772,
+ "learning_rate": 0.0006,
+ "loss": 5.093441009521484,
+ "step": 3540
+ },
+ {
+ "epoch": 49.18165938864629,
+ "grad_norm": 0.011126236990094185,
+ "learning_rate": 0.0006,
+ "loss": 5.060235023498535,
+ "step": 3541
+ },
+ {
+ "epoch": 49.19563318777293,
+ "grad_norm": 0.0107401292771101,
+ "learning_rate": 0.0006,
+ "loss": 5.140704154968262,
+ "step": 3542
+ },
+ {
+ "epoch": 49.209606986899566,
+ "grad_norm": 0.011202906258404255,
+ "learning_rate": 0.0006,
+ "loss": 5.10855770111084,
+ "step": 3543
+ },
+ {
+ "epoch": 49.223580786026204,
+ "grad_norm": 0.011083454824984074,
+ "learning_rate": 0.0006,
+ "loss": 5.092751502990723,
+ "step": 3544
+ },
+ {
+ "epoch": 49.237554585152836,
+ "grad_norm": 0.009523800574243069,
+ "learning_rate": 0.0006,
+ "loss": 5.13733434677124,
+ "step": 3545
+ },
+ {
+ "epoch": 49.251528384279474,
+ "grad_norm": 0.011298353783786297,
+ "learning_rate": 0.0006,
+ "loss": 5.036048889160156,
+ "step": 3546
+ },
+ {
+ "epoch": 49.26550218340611,
+ "grad_norm": 0.010676483623683453,
+ "learning_rate": 0.0006,
+ "loss": 5.051901340484619,
+ "step": 3547
+ },
+ {
+ "epoch": 49.27947598253275,
+ "grad_norm": 0.011463399045169353,
+ "learning_rate": 0.0006,
+ "loss": 4.983170986175537,
+ "step": 3548
+ },
+ {
+ "epoch": 49.29344978165939,
+ "grad_norm": 0.010603402741253376,
+ "learning_rate": 0.0006,
+ "loss": 5.112311363220215,
+ "step": 3549
+ },
+ {
+ "epoch": 49.30742358078603,
+ "grad_norm": 0.010535670444369316,
+ "learning_rate": 0.0006,
+ "loss": 5.085227966308594,
+ "step": 3550
+ },
+ {
+ "epoch": 49.32139737991266,
+ "grad_norm": 0.011843928135931492,
+ "learning_rate": 0.0006,
+ "loss": 5.07575798034668,
+ "step": 3551
+ },
+ {
+ "epoch": 49.3353711790393,
+ "grad_norm": 0.010587952099740505,
+ "learning_rate": 0.0006,
+ "loss": 5.019237041473389,
+ "step": 3552
+ },
+ {
+ "epoch": 49.34934497816594,
+ "grad_norm": 0.012285659089684486,
+ "learning_rate": 0.0006,
+ "loss": 5.144603729248047,
+ "step": 3553
+ },
+ {
+ "epoch": 49.36331877729258,
+ "grad_norm": 0.013119431212544441,
+ "learning_rate": 0.0006,
+ "loss": 5.035256862640381,
+ "step": 3554
+ },
+ {
+ "epoch": 49.377292576419215,
+ "grad_norm": 0.012138537131249905,
+ "learning_rate": 0.0006,
+ "loss": 4.996667861938477,
+ "step": 3555
+ },
+ {
+ "epoch": 49.391266375545854,
+ "grad_norm": 0.012142804451286793,
+ "learning_rate": 0.0006,
+ "loss": 4.95058536529541,
+ "step": 3556
+ },
+ {
+ "epoch": 49.40524017467249,
+ "grad_norm": 0.012205676175653934,
+ "learning_rate": 0.0006,
+ "loss": 5.096931457519531,
+ "step": 3557
+ },
+ {
+ "epoch": 49.419213973799124,
+ "grad_norm": 0.012086744420230389,
+ "learning_rate": 0.0006,
+ "loss": 4.99384880065918,
+ "step": 3558
+ },
+ {
+ "epoch": 49.43318777292576,
+ "grad_norm": 0.012918103486299515,
+ "learning_rate": 0.0006,
+ "loss": 5.031294822692871,
+ "step": 3559
+ },
+ {
+ "epoch": 49.4471615720524,
+ "grad_norm": 0.012006713077425957,
+ "learning_rate": 0.0006,
+ "loss": 5.0547943115234375,
+ "step": 3560
+ },
+ {
+ "epoch": 49.46113537117904,
+ "grad_norm": 0.0120399734005332,
+ "learning_rate": 0.0006,
+ "loss": 5.014238357543945,
+ "step": 3561
+ },
+ {
+ "epoch": 49.47510917030568,
+ "grad_norm": 0.013306990265846252,
+ "learning_rate": 0.0006,
+ "loss": 4.972925186157227,
+ "step": 3562
+ },
+ {
+ "epoch": 49.48908296943232,
+ "grad_norm": 0.013878636993467808,
+ "learning_rate": 0.0006,
+ "loss": 5.079033851623535,
+ "step": 3563
+ },
+ {
+ "epoch": 49.50305676855895,
+ "grad_norm": 0.012454566545784473,
+ "learning_rate": 0.0006,
+ "loss": 4.93708610534668,
+ "step": 3564
+ },
+ {
+ "epoch": 49.51703056768559,
+ "grad_norm": 0.014075061306357384,
+ "learning_rate": 0.0006,
+ "loss": 4.965582370758057,
+ "step": 3565
+ },
+ {
+ "epoch": 49.531004366812226,
+ "grad_norm": 0.012525617145001888,
+ "learning_rate": 0.0006,
+ "loss": 4.968148708343506,
+ "step": 3566
+ },
+ {
+ "epoch": 49.544978165938865,
+ "grad_norm": 0.012418225407600403,
+ "learning_rate": 0.0006,
+ "loss": 5.06901741027832,
+ "step": 3567
+ },
+ {
+ "epoch": 49.5589519650655,
+ "grad_norm": 0.015034239739179611,
+ "learning_rate": 0.0006,
+ "loss": 4.984277725219727,
+ "step": 3568
+ },
+ {
+ "epoch": 49.57292576419214,
+ "grad_norm": 0.014329628087580204,
+ "learning_rate": 0.0006,
+ "loss": 4.951558589935303,
+ "step": 3569
+ },
+ {
+ "epoch": 49.58689956331878,
+ "grad_norm": 0.017769070342183113,
+ "learning_rate": 0.0006,
+ "loss": 4.966532230377197,
+ "step": 3570
+ },
+ {
+ "epoch": 49.60087336244541,
+ "grad_norm": 0.01909453794360161,
+ "learning_rate": 0.0006,
+ "loss": 4.919890880584717,
+ "step": 3571
+ },
+ {
+ "epoch": 49.61484716157205,
+ "grad_norm": 0.017519688233733177,
+ "learning_rate": 0.0006,
+ "loss": 5.031286716461182,
+ "step": 3572
+ },
+ {
+ "epoch": 49.62882096069869,
+ "grad_norm": 0.017143897712230682,
+ "learning_rate": 0.0006,
+ "loss": 4.969764709472656,
+ "step": 3573
+ },
+ {
+ "epoch": 49.64279475982533,
+ "grad_norm": 0.01688285917043686,
+ "learning_rate": 0.0006,
+ "loss": 5.0733442306518555,
+ "step": 3574
+ },
+ {
+ "epoch": 49.65676855895197,
+ "grad_norm": 0.018547311425209045,
+ "learning_rate": 0.0006,
+ "loss": 4.980863571166992,
+ "step": 3575
+ },
+ {
+ "epoch": 49.670742358078606,
+ "grad_norm": 0.018877137452363968,
+ "learning_rate": 0.0006,
+ "loss": 4.954646110534668,
+ "step": 3576
+ },
+ {
+ "epoch": 49.68471615720524,
+ "grad_norm": 0.018681600689888,
+ "learning_rate": 0.0006,
+ "loss": 4.959702014923096,
+ "step": 3577
+ },
+ {
+ "epoch": 49.698689956331876,
+ "grad_norm": 0.018947336822748184,
+ "learning_rate": 0.0006,
+ "loss": 5.070502758026123,
+ "step": 3578
+ },
+ {
+ "epoch": 49.712663755458514,
+ "grad_norm": 0.018817484378814697,
+ "learning_rate": 0.0006,
+ "loss": 5.056508541107178,
+ "step": 3579
+ },
+ {
+ "epoch": 49.72663755458515,
+ "grad_norm": 0.01931724324822426,
+ "learning_rate": 0.0006,
+ "loss": 4.922451972961426,
+ "step": 3580
+ },
+ {
+ "epoch": 49.74061135371179,
+ "grad_norm": 0.022047819569706917,
+ "learning_rate": 0.0006,
+ "loss": 5.053170204162598,
+ "step": 3581
+ },
+ {
+ "epoch": 49.75458515283843,
+ "grad_norm": 0.02279801294207573,
+ "learning_rate": 0.0006,
+ "loss": 4.8768439292907715,
+ "step": 3582
+ },
+ {
+ "epoch": 49.76855895196506,
+ "grad_norm": 0.020661601796746254,
+ "learning_rate": 0.0006,
+ "loss": 4.907928466796875,
+ "step": 3583
+ },
+ {
+ "epoch": 49.7825327510917,
+ "grad_norm": 0.020479509606957436,
+ "learning_rate": 0.0006,
+ "loss": 5.040694236755371,
+ "step": 3584
+ },
+ {
+ "epoch": 49.79650655021834,
+ "grad_norm": 0.021654162555933,
+ "learning_rate": 0.0006,
+ "loss": 5.018434524536133,
+ "step": 3585
+ },
+ {
+ "epoch": 49.81048034934498,
+ "grad_norm": 0.022486824542284012,
+ "learning_rate": 0.0006,
+ "loss": 5.086538314819336,
+ "step": 3586
+ },
+ {
+ "epoch": 49.82445414847162,
+ "grad_norm": 0.02141369879245758,
+ "learning_rate": 0.0006,
+ "loss": 5.035694122314453,
+ "step": 3587
+ },
+ {
+ "epoch": 49.838427947598255,
+ "grad_norm": 0.015926677733659744,
+ "learning_rate": 0.0006,
+ "loss": 4.905038833618164,
+ "step": 3588
+ },
+ {
+ "epoch": 49.852401746724894,
+ "grad_norm": 0.015775298699736595,
+ "learning_rate": 0.0006,
+ "loss": 5.003902435302734,
+ "step": 3589
+ },
+ {
+ "epoch": 49.866375545851525,
+ "grad_norm": 0.01872118003666401,
+ "learning_rate": 0.0006,
+ "loss": 4.891060829162598,
+ "step": 3590
+ },
+ {
+ "epoch": 49.880349344978164,
+ "grad_norm": 0.02066192403435707,
+ "learning_rate": 0.0006,
+ "loss": 5.002897262573242,
+ "step": 3591
+ },
+ {
+ "epoch": 49.8943231441048,
+ "grad_norm": 0.02348714880645275,
+ "learning_rate": 0.0006,
+ "loss": 4.9189958572387695,
+ "step": 3592
+ },
+ {
+ "epoch": 49.90829694323144,
+ "grad_norm": 0.020402103662490845,
+ "learning_rate": 0.0006,
+ "loss": 4.992661476135254,
+ "step": 3593
+ },
+ {
+ "epoch": 49.92227074235808,
+ "grad_norm": 0.020466072484850883,
+ "learning_rate": 0.0006,
+ "loss": 4.962968826293945,
+ "step": 3594
+ },
+ {
+ "epoch": 49.93624454148472,
+ "grad_norm": 0.01922314241528511,
+ "learning_rate": 0.0006,
+ "loss": 4.897730827331543,
+ "step": 3595
+ },
+ {
+ "epoch": 49.95021834061135,
+ "grad_norm": 0.017228754237294197,
+ "learning_rate": 0.0006,
+ "loss": 4.964842796325684,
+ "step": 3596
+ },
+ {
+ "epoch": 49.96419213973799,
+ "grad_norm": 0.016021743416786194,
+ "learning_rate": 0.0006,
+ "loss": 4.910758972167969,
+ "step": 3597
+ },
+ {
+ "epoch": 49.97816593886463,
+ "grad_norm": 0.017985103651881218,
+ "learning_rate": 0.0006,
+ "loss": 4.916072368621826,
+ "step": 3598
+ },
+ {
+ "epoch": 49.992139737991266,
+ "grad_norm": 0.018780240789055824,
+ "learning_rate": 0.0006,
+ "loss": 4.94658088684082,
+ "step": 3599
+ },
+ {
+ "epoch": 50.0,
+ "grad_norm": 0.01776314526796341,
+ "learning_rate": 0.0006,
+ "loss": 4.969357967376709,
+ "step": 3600
+ },
+ {
+ "epoch": 50.0,
+ "eval_loss": 5.141158580780029,
+ "eval_runtime": 56.2332,
+ "eval_samples_per_second": 43.426,
+ "eval_steps_per_second": 1.369,
+ "step": 3600
+ },
+ {
+ "epoch": 50.01397379912664,
+ "grad_norm": 0.016308696940541267,
+ "learning_rate": 0.0006,
+ "loss": 4.924664497375488,
+ "step": 3601
+ },
+ {
+ "epoch": 50.02794759825328,
+ "grad_norm": 0.018810199573636055,
+ "learning_rate": 0.0006,
+ "loss": 4.82293701171875,
+ "step": 3602
+ },
+ {
+ "epoch": 50.041921397379916,
+ "grad_norm": 0.02227160707116127,
+ "learning_rate": 0.0006,
+ "loss": 4.9612298011779785,
+ "step": 3603
+ },
+ {
+ "epoch": 50.05589519650655,
+ "grad_norm": 0.025266462936997414,
+ "learning_rate": 0.0006,
+ "loss": 4.963835716247559,
+ "step": 3604
+ },
+ {
+ "epoch": 50.069868995633186,
+ "grad_norm": 0.028588181361556053,
+ "learning_rate": 0.0006,
+ "loss": 4.992118835449219,
+ "step": 3605
+ },
+ {
+ "epoch": 50.083842794759825,
+ "grad_norm": 0.02739487960934639,
+ "learning_rate": 0.0006,
+ "loss": 4.915547847747803,
+ "step": 3606
+ },
+ {
+ "epoch": 50.09781659388646,
+ "grad_norm": 0.027000300586223602,
+ "learning_rate": 0.0006,
+ "loss": 4.883801460266113,
+ "step": 3607
+ },
+ {
+ "epoch": 50.1117903930131,
+ "grad_norm": 0.023468483239412308,
+ "learning_rate": 0.0006,
+ "loss": 5.006168365478516,
+ "step": 3608
+ },
+ {
+ "epoch": 50.12576419213974,
+ "grad_norm": 0.026753634214401245,
+ "learning_rate": 0.0006,
+ "loss": 4.92526912689209,
+ "step": 3609
+ },
+ {
+ "epoch": 50.13973799126637,
+ "grad_norm": 0.026816733181476593,
+ "learning_rate": 0.0006,
+ "loss": 4.847546100616455,
+ "step": 3610
+ },
+ {
+ "epoch": 50.15371179039301,
+ "grad_norm": 0.027362391352653503,
+ "learning_rate": 0.0006,
+ "loss": 4.862548351287842,
+ "step": 3611
+ },
+ {
+ "epoch": 50.16768558951965,
+ "grad_norm": 0.03450845181941986,
+ "learning_rate": 0.0006,
+ "loss": 4.963202953338623,
+ "step": 3612
+ },
+ {
+ "epoch": 50.18165938864629,
+ "grad_norm": 0.03801408037543297,
+ "learning_rate": 0.0006,
+ "loss": 4.770240306854248,
+ "step": 3613
+ },
+ {
+ "epoch": 50.19563318777293,
+ "grad_norm": 0.03579140082001686,
+ "learning_rate": 0.0006,
+ "loss": 4.797981262207031,
+ "step": 3614
+ },
+ {
+ "epoch": 50.209606986899566,
+ "grad_norm": 0.029393760487437248,
+ "learning_rate": 0.0006,
+ "loss": 4.798581600189209,
+ "step": 3615
+ },
+ {
+ "epoch": 50.223580786026204,
+ "grad_norm": 0.026111142709851265,
+ "learning_rate": 0.0006,
+ "loss": 4.990712642669678,
+ "step": 3616
+ },
+ {
+ "epoch": 50.237554585152836,
+ "grad_norm": 0.02213387005031109,
+ "learning_rate": 0.0006,
+ "loss": 4.865842342376709,
+ "step": 3617
+ },
+ {
+ "epoch": 50.251528384279474,
+ "grad_norm": 0.025450684130191803,
+ "learning_rate": 0.0006,
+ "loss": 4.903746128082275,
+ "step": 3618
+ },
+ {
+ "epoch": 50.26550218340611,
+ "grad_norm": 0.020915117114782333,
+ "learning_rate": 0.0006,
+ "loss": 4.951366424560547,
+ "step": 3619
+ },
+ {
+ "epoch": 50.27947598253275,
+ "grad_norm": 0.020972849801182747,
+ "learning_rate": 0.0006,
+ "loss": 4.783311367034912,
+ "step": 3620
+ },
+ {
+ "epoch": 50.29344978165939,
+ "grad_norm": 0.016956273466348648,
+ "learning_rate": 0.0006,
+ "loss": 4.955767631530762,
+ "step": 3621
+ },
+ {
+ "epoch": 50.30742358078603,
+ "grad_norm": 0.018045252189040184,
+ "learning_rate": 0.0006,
+ "loss": 4.869322776794434,
+ "step": 3622
+ },
+ {
+ "epoch": 50.32139737991266,
+ "grad_norm": 0.018199188634753227,
+ "learning_rate": 0.0006,
+ "loss": 4.95680046081543,
+ "step": 3623
+ },
+ {
+ "epoch": 50.3353711790393,
+ "grad_norm": 0.016688082367181778,
+ "learning_rate": 0.0006,
+ "loss": 4.783535957336426,
+ "step": 3624
+ },
+ {
+ "epoch": 50.34934497816594,
+ "grad_norm": 0.017540352419018745,
+ "learning_rate": 0.0006,
+ "loss": 4.8236083984375,
+ "step": 3625
+ },
+ {
+ "epoch": 50.36331877729258,
+ "grad_norm": 0.018054187297821045,
+ "learning_rate": 0.0006,
+ "loss": 4.764655113220215,
+ "step": 3626
+ },
+ {
+ "epoch": 50.377292576419215,
+ "grad_norm": 0.01741131953895092,
+ "learning_rate": 0.0006,
+ "loss": 4.875613689422607,
+ "step": 3627
+ },
+ {
+ "epoch": 50.391266375545854,
+ "grad_norm": 0.01682269014418125,
+ "learning_rate": 0.0006,
+ "loss": 4.818509101867676,
+ "step": 3628
+ },
+ {
+ "epoch": 50.40524017467249,
+ "grad_norm": 0.016652341932058334,
+ "learning_rate": 0.0006,
+ "loss": 4.974020481109619,
+ "step": 3629
+ },
+ {
+ "epoch": 50.419213973799124,
+ "grad_norm": 0.02035282365977764,
+ "learning_rate": 0.0006,
+ "loss": 4.893037796020508,
+ "step": 3630
+ },
+ {
+ "epoch": 50.43318777292576,
+ "grad_norm": 0.019907061010599136,
+ "learning_rate": 0.0006,
+ "loss": 4.848119735717773,
+ "step": 3631
+ },
+ {
+ "epoch": 50.4471615720524,
+ "grad_norm": 0.023861493915319443,
+ "learning_rate": 0.0006,
+ "loss": 4.819353103637695,
+ "step": 3632
+ },
+ {
+ "epoch": 50.46113537117904,
+ "grad_norm": 0.025254026055336,
+ "learning_rate": 0.0006,
+ "loss": 4.769312858581543,
+ "step": 3633
+ },
+ {
+ "epoch": 50.47510917030568,
+ "grad_norm": 0.026154937222599983,
+ "learning_rate": 0.0006,
+ "loss": 4.7586774826049805,
+ "step": 3634
+ },
+ {
+ "epoch": 50.48908296943232,
+ "grad_norm": 0.026342090219259262,
+ "learning_rate": 0.0006,
+ "loss": 4.861218452453613,
+ "step": 3635
+ },
+ {
+ "epoch": 50.50305676855895,
+ "grad_norm": 0.028169559314846992,
+ "learning_rate": 0.0006,
+ "loss": 4.818380355834961,
+ "step": 3636
+ },
+ {
+ "epoch": 50.51703056768559,
+ "grad_norm": 0.02820264920592308,
+ "learning_rate": 0.0006,
+ "loss": 4.892871856689453,
+ "step": 3637
+ },
+ {
+ "epoch": 50.531004366812226,
+ "grad_norm": 0.03345080465078354,
+ "learning_rate": 0.0006,
+ "loss": 4.811084747314453,
+ "step": 3638
+ },
+ {
+ "epoch": 50.544978165938865,
+ "grad_norm": 0.04005736857652664,
+ "learning_rate": 0.0006,
+ "loss": 4.765410900115967,
+ "step": 3639
+ },
+ {
+ "epoch": 50.5589519650655,
+ "grad_norm": 0.04333870857954025,
+ "learning_rate": 0.0006,
+ "loss": 4.956520080566406,
+ "step": 3640
+ },
+ {
+ "epoch": 50.57292576419214,
+ "grad_norm": 0.04216952994465828,
+ "learning_rate": 0.0006,
+ "loss": 4.793441295623779,
+ "step": 3641
+ },
+ {
+ "epoch": 50.58689956331878,
+ "grad_norm": 0.02936951443552971,
+ "learning_rate": 0.0006,
+ "loss": 4.8447957038879395,
+ "step": 3642
+ },
+ {
+ "epoch": 50.60087336244541,
+ "grad_norm": 0.026937026530504227,
+ "learning_rate": 0.0006,
+ "loss": 4.962215423583984,
+ "step": 3643
+ },
+ {
+ "epoch": 50.61484716157205,
+ "grad_norm": 0.02628890983760357,
+ "learning_rate": 0.0006,
+ "loss": 4.883061408996582,
+ "step": 3644
+ },
+ {
+ "epoch": 50.62882096069869,
+ "grad_norm": 0.025984834879636765,
+ "learning_rate": 0.0006,
+ "loss": 4.768662452697754,
+ "step": 3645
+ },
+ {
+ "epoch": 50.64279475982533,
+ "grad_norm": 0.023998012766242027,
+ "learning_rate": 0.0006,
+ "loss": 4.851325035095215,
+ "step": 3646
+ },
+ {
+ "epoch": 50.65676855895197,
+ "grad_norm": 0.02143184095621109,
+ "learning_rate": 0.0006,
+ "loss": 4.8274688720703125,
+ "step": 3647
+ },
+ {
+ "epoch": 50.670742358078606,
+ "grad_norm": 0.02488900162279606,
+ "learning_rate": 0.0006,
+ "loss": 4.831140518188477,
+ "step": 3648
+ },
+ {
+ "epoch": 50.68471615720524,
+ "grad_norm": 0.02330533228814602,
+ "learning_rate": 0.0006,
+ "loss": 4.850462913513184,
+ "step": 3649
+ },
+ {
+ "epoch": 50.698689956331876,
+ "grad_norm": 0.02121943049132824,
+ "learning_rate": 0.0006,
+ "loss": 4.870253086090088,
+ "step": 3650
+ },
+ {
+ "epoch": 50.712663755458514,
+ "grad_norm": 0.02080935798585415,
+ "learning_rate": 0.0006,
+ "loss": 4.798770904541016,
+ "step": 3651
+ },
+ {
+ "epoch": 50.72663755458515,
+ "grad_norm": 0.019047029316425323,
+ "learning_rate": 0.0006,
+ "loss": 4.8427300453186035,
+ "step": 3652
+ },
+ {
+ "epoch": 50.74061135371179,
+ "grad_norm": 0.01824103854596615,
+ "learning_rate": 0.0006,
+ "loss": 4.820980548858643,
+ "step": 3653
+ },
+ {
+ "epoch": 50.75458515283843,
+ "grad_norm": 0.018129967153072357,
+ "learning_rate": 0.0006,
+ "loss": 4.946732997894287,
+ "step": 3654
+ },
+ {
+ "epoch": 50.76855895196506,
+ "grad_norm": 0.01784815452992916,
+ "learning_rate": 0.0006,
+ "loss": 4.829516410827637,
+ "step": 3655
+ },
+ {
+ "epoch": 50.7825327510917,
+ "grad_norm": 0.020021555945277214,
+ "learning_rate": 0.0006,
+ "loss": 4.780428409576416,
+ "step": 3656
+ },
+ {
+ "epoch": 50.79650655021834,
+ "grad_norm": 0.01863253302872181,
+ "learning_rate": 0.0006,
+ "loss": 4.791567802429199,
+ "step": 3657
+ },
+ {
+ "epoch": 50.81048034934498,
+ "grad_norm": 0.019151683896780014,
+ "learning_rate": 0.0006,
+ "loss": 4.824892997741699,
+ "step": 3658
+ },
+ {
+ "epoch": 50.82445414847162,
+ "grad_norm": 0.01891648955643177,
+ "learning_rate": 0.0006,
+ "loss": 4.789605140686035,
+ "step": 3659
+ },
+ {
+ "epoch": 50.838427947598255,
+ "grad_norm": 0.016019510105252266,
+ "learning_rate": 0.0006,
+ "loss": 4.779555320739746,
+ "step": 3660
+ },
+ {
+ "epoch": 50.852401746724894,
+ "grad_norm": 0.016096491366624832,
+ "learning_rate": 0.0006,
+ "loss": 4.753842353820801,
+ "step": 3661
+ },
+ {
+ "epoch": 50.866375545851525,
+ "grad_norm": 0.01693476364016533,
+ "learning_rate": 0.0006,
+ "loss": 4.768477439880371,
+ "step": 3662
+ },
+ {
+ "epoch": 50.880349344978164,
+ "grad_norm": 0.01749575324356556,
+ "learning_rate": 0.0006,
+ "loss": 4.779143333435059,
+ "step": 3663
+ },
+ {
+ "epoch": 50.8943231441048,
+ "grad_norm": 0.016851043328642845,
+ "learning_rate": 0.0006,
+ "loss": 4.780387878417969,
+ "step": 3664
+ },
+ {
+ "epoch": 50.90829694323144,
+ "grad_norm": 0.019045202061533928,
+ "learning_rate": 0.0006,
+ "loss": 4.769913673400879,
+ "step": 3665
+ },
+ {
+ "epoch": 50.92227074235808,
+ "grad_norm": 0.01954123005270958,
+ "learning_rate": 0.0006,
+ "loss": 4.769009590148926,
+ "step": 3666
+ },
+ {
+ "epoch": 50.93624454148472,
+ "grad_norm": 0.022527096793055534,
+ "learning_rate": 0.0006,
+ "loss": 4.789215087890625,
+ "step": 3667
+ },
+ {
+ "epoch": 50.95021834061135,
+ "grad_norm": 0.02871634066104889,
+ "learning_rate": 0.0006,
+ "loss": 4.749727725982666,
+ "step": 3668
+ },
+ {
+ "epoch": 50.96419213973799,
+ "grad_norm": 0.04045053571462631,
+ "learning_rate": 0.0006,
+ "loss": 4.909038543701172,
+ "step": 3669
+ },
+ {
+ "epoch": 50.97816593886463,
+ "grad_norm": 0.04156879335641861,
+ "learning_rate": 0.0006,
+ "loss": 4.876831531524658,
+ "step": 3670
+ },
+ {
+ "epoch": 50.992139737991266,
+ "grad_norm": 0.02621397376060486,
+ "learning_rate": 0.0006,
+ "loss": 4.797173500061035,
+ "step": 3671
+ },
+ {
+ "epoch": 51.0,
+ "grad_norm": 0.03389597684144974,
+ "learning_rate": 0.0006,
+ "loss": 4.8428850173950195,
+ "step": 3672
+ },
+ {
+ "epoch": 51.0,
+ "eval_loss": 4.989261627197266,
+ "eval_runtime": 56.358,
+ "eval_samples_per_second": 43.33,
+ "eval_steps_per_second": 1.366,
+ "step": 3672
+ },
+ {
+ "epoch": 51.01397379912664,
+ "grad_norm": 0.03216972202062607,
+ "learning_rate": 0.0006,
+ "loss": 4.755102634429932,
+ "step": 3673
+ },
+ {
+ "epoch": 51.02794759825328,
+ "grad_norm": 0.03190374746918678,
+ "learning_rate": 0.0006,
+ "loss": 4.720424652099609,
+ "step": 3674
+ },
+ {
+ "epoch": 51.041921397379916,
+ "grad_norm": 0.031727299094200134,
+ "learning_rate": 0.0006,
+ "loss": 4.803277969360352,
+ "step": 3675
+ },
+ {
+ "epoch": 51.05589519650655,
+ "grad_norm": 0.02473079040646553,
+ "learning_rate": 0.0006,
+ "loss": 4.761087417602539,
+ "step": 3676
+ },
+ {
+ "epoch": 51.069868995633186,
+ "grad_norm": 0.020360209047794342,
+ "learning_rate": 0.0006,
+ "loss": 4.753884792327881,
+ "step": 3677
+ },
+ {
+ "epoch": 51.083842794759825,
+ "grad_norm": 0.02078923024237156,
+ "learning_rate": 0.0006,
+ "loss": 4.647130489349365,
+ "step": 3678
+ },
+ {
+ "epoch": 51.09781659388646,
+ "grad_norm": 0.019387200474739075,
+ "learning_rate": 0.0006,
+ "loss": 4.74135160446167,
+ "step": 3679
+ },
+ {
+ "epoch": 51.1117903930131,
+ "grad_norm": 0.01874949410557747,
+ "learning_rate": 0.0006,
+ "loss": 4.795329570770264,
+ "step": 3680
+ },
+ {
+ "epoch": 51.12576419213974,
+ "grad_norm": 0.019331837072968483,
+ "learning_rate": 0.0006,
+ "loss": 4.897967338562012,
+ "step": 3681
+ },
+ {
+ "epoch": 51.13973799126637,
+ "grad_norm": 0.017066050320863724,
+ "learning_rate": 0.0006,
+ "loss": 4.825351715087891,
+ "step": 3682
+ },
+ {
+ "epoch": 51.15371179039301,
+ "grad_norm": 0.01757810264825821,
+ "learning_rate": 0.0006,
+ "loss": 4.719335079193115,
+ "step": 3683
+ },
+ {
+ "epoch": 51.16768558951965,
+ "grad_norm": 0.018174266442656517,
+ "learning_rate": 0.0006,
+ "loss": 4.732680320739746,
+ "step": 3684
+ },
+ {
+ "epoch": 51.18165938864629,
+ "grad_norm": 0.018497727811336517,
+ "learning_rate": 0.0006,
+ "loss": 4.731050491333008,
+ "step": 3685
+ },
+ {
+ "epoch": 51.19563318777293,
+ "grad_norm": 0.0186338871717453,
+ "learning_rate": 0.0006,
+ "loss": 4.748035430908203,
+ "step": 3686
+ },
+ {
+ "epoch": 51.209606986899566,
+ "grad_norm": 0.018531711772084236,
+ "learning_rate": 0.0006,
+ "loss": 4.9007463455200195,
+ "step": 3687
+ },
+ {
+ "epoch": 51.223580786026204,
+ "grad_norm": 0.018221354112029076,
+ "learning_rate": 0.0006,
+ "loss": 4.84077262878418,
+ "step": 3688
+ },
+ {
+ "epoch": 51.237554585152836,
+ "grad_norm": 0.0203656405210495,
+ "learning_rate": 0.0006,
+ "loss": 4.73009729385376,
+ "step": 3689
+ },
+ {
+ "epoch": 51.251528384279474,
+ "grad_norm": 0.01964893378317356,
+ "learning_rate": 0.0006,
+ "loss": 4.772583961486816,
+ "step": 3690
+ },
+ {
+ "epoch": 51.26550218340611,
+ "grad_norm": 0.019606510177254677,
+ "learning_rate": 0.0006,
+ "loss": 4.599812030792236,
+ "step": 3691
+ },
+ {
+ "epoch": 51.27947598253275,
+ "grad_norm": 0.020501255989074707,
+ "learning_rate": 0.0006,
+ "loss": 4.809898853302002,
+ "step": 3692
+ },
+ {
+ "epoch": 51.29344978165939,
+ "grad_norm": 0.01870838552713394,
+ "learning_rate": 0.0006,
+ "loss": 4.741898059844971,
+ "step": 3693
+ },
+ {
+ "epoch": 51.30742358078603,
+ "grad_norm": 0.01802559196949005,
+ "learning_rate": 0.0006,
+ "loss": 4.746562480926514,
+ "step": 3694
+ },
+ {
+ "epoch": 51.32139737991266,
+ "grad_norm": 0.02053048275411129,
+ "learning_rate": 0.0006,
+ "loss": 4.69932746887207,
+ "step": 3695
+ },
+ {
+ "epoch": 51.3353711790393,
+ "grad_norm": 0.02661825157701969,
+ "learning_rate": 0.0006,
+ "loss": 4.68748664855957,
+ "step": 3696
+ },
+ {
+ "epoch": 51.34934497816594,
+ "grad_norm": 0.028269147500395775,
+ "learning_rate": 0.0006,
+ "loss": 4.731565475463867,
+ "step": 3697
+ },
+ {
+ "epoch": 51.36331877729258,
+ "grad_norm": 0.02650964818894863,
+ "learning_rate": 0.0006,
+ "loss": 4.733857154846191,
+ "step": 3698
+ },
+ {
+ "epoch": 51.377292576419215,
+ "grad_norm": 0.023696066811680794,
+ "learning_rate": 0.0006,
+ "loss": 4.812362194061279,
+ "step": 3699
+ },
+ {
+ "epoch": 51.391266375545854,
+ "grad_norm": 0.02365880273282528,
+ "learning_rate": 0.0006,
+ "loss": 4.733426094055176,
+ "step": 3700
+ },
+ {
+ "epoch": 51.40524017467249,
+ "grad_norm": 0.026249447837471962,
+ "learning_rate": 0.0006,
+ "loss": 4.715622901916504,
+ "step": 3701
+ },
+ {
+ "epoch": 51.419213973799124,
+ "grad_norm": 0.03301899880170822,
+ "learning_rate": 0.0006,
+ "loss": 4.7223663330078125,
+ "step": 3702
+ },
+ {
+ "epoch": 51.43318777292576,
+ "grad_norm": 0.044686660170555115,
+ "learning_rate": 0.0006,
+ "loss": 4.8121161460876465,
+ "step": 3703
+ },
+ {
+ "epoch": 51.4471615720524,
+ "grad_norm": 0.04733522981405258,
+ "learning_rate": 0.0006,
+ "loss": 4.841394424438477,
+ "step": 3704
+ },
+ {
+ "epoch": 51.46113537117904,
+ "grad_norm": 0.0395209975540638,
+ "learning_rate": 0.0006,
+ "loss": 4.773837089538574,
+ "step": 3705
+ },
+ {
+ "epoch": 51.47510917030568,
+ "grad_norm": 0.032790958881378174,
+ "learning_rate": 0.0006,
+ "loss": 4.833066463470459,
+ "step": 3706
+ },
+ {
+ "epoch": 51.48908296943232,
+ "grad_norm": 0.030578427016735077,
+ "learning_rate": 0.0006,
+ "loss": 4.685190200805664,
+ "step": 3707
+ },
+ {
+ "epoch": 51.50305676855895,
+ "grad_norm": 0.031619083136320114,
+ "learning_rate": 0.0006,
+ "loss": 4.740378379821777,
+ "step": 3708
+ },
+ {
+ "epoch": 51.51703056768559,
+ "grad_norm": 0.02385825850069523,
+ "learning_rate": 0.0006,
+ "loss": 4.812911510467529,
+ "step": 3709
+ },
+ {
+ "epoch": 51.531004366812226,
+ "grad_norm": 0.023523425683379173,
+ "learning_rate": 0.0006,
+ "loss": 4.669580459594727,
+ "step": 3710
+ },
+ {
+ "epoch": 51.544978165938865,
+ "grad_norm": 0.021316280588507652,
+ "learning_rate": 0.0006,
+ "loss": 4.76888370513916,
+ "step": 3711
+ },
+ {
+ "epoch": 51.5589519650655,
+ "grad_norm": 0.024663880467414856,
+ "learning_rate": 0.0006,
+ "loss": 4.702592372894287,
+ "step": 3712
+ },
+ {
+ "epoch": 51.57292576419214,
+ "grad_norm": 0.024173280224204063,
+ "learning_rate": 0.0006,
+ "loss": 4.710604667663574,
+ "step": 3713
+ },
+ {
+ "epoch": 51.58689956331878,
+ "grad_norm": 0.02754114381968975,
+ "learning_rate": 0.0006,
+ "loss": 4.746356964111328,
+ "step": 3714
+ },
+ {
+ "epoch": 51.60087336244541,
+ "grad_norm": 0.027903622016310692,
+ "learning_rate": 0.0006,
+ "loss": 4.7546491622924805,
+ "step": 3715
+ },
+ {
+ "epoch": 51.61484716157205,
+ "grad_norm": 0.024871617555618286,
+ "learning_rate": 0.0006,
+ "loss": 4.605746269226074,
+ "step": 3716
+ },
+ {
+ "epoch": 51.62882096069869,
+ "grad_norm": 0.021239031106233597,
+ "learning_rate": 0.0006,
+ "loss": 4.6942458152771,
+ "step": 3717
+ },
+ {
+ "epoch": 51.64279475982533,
+ "grad_norm": 0.01925622671842575,
+ "learning_rate": 0.0006,
+ "loss": 4.593558311462402,
+ "step": 3718
+ },
+ {
+ "epoch": 51.65676855895197,
+ "grad_norm": 0.01956283673644066,
+ "learning_rate": 0.0006,
+ "loss": 4.631172180175781,
+ "step": 3719
+ },
+ {
+ "epoch": 51.670742358078606,
+ "grad_norm": 0.02269871160387993,
+ "learning_rate": 0.0006,
+ "loss": 4.729236602783203,
+ "step": 3720
+ },
+ {
+ "epoch": 51.68471615720524,
+ "grad_norm": 0.02253916673362255,
+ "learning_rate": 0.0006,
+ "loss": 4.7191033363342285,
+ "step": 3721
+ },
+ {
+ "epoch": 51.698689956331876,
+ "grad_norm": 0.026231050491333008,
+ "learning_rate": 0.0006,
+ "loss": 4.811850547790527,
+ "step": 3722
+ },
+ {
+ "epoch": 51.712663755458514,
+ "grad_norm": 0.02558821812272072,
+ "learning_rate": 0.0006,
+ "loss": 4.681466102600098,
+ "step": 3723
+ },
+ {
+ "epoch": 51.72663755458515,
+ "grad_norm": 0.02124209702014923,
+ "learning_rate": 0.0006,
+ "loss": 4.693664073944092,
+ "step": 3724
+ },
+ {
+ "epoch": 51.74061135371179,
+ "grad_norm": 0.016638319939374924,
+ "learning_rate": 0.0006,
+ "loss": 4.771440029144287,
+ "step": 3725
+ },
+ {
+ "epoch": 51.75458515283843,
+ "grad_norm": 0.018048347905278206,
+ "learning_rate": 0.0006,
+ "loss": 4.643252372741699,
+ "step": 3726
+ },
+ {
+ "epoch": 51.76855895196506,
+ "grad_norm": 0.019904717803001404,
+ "learning_rate": 0.0006,
+ "loss": 4.66530704498291,
+ "step": 3727
+ },
+ {
+ "epoch": 51.7825327510917,
+ "grad_norm": 0.0194614976644516,
+ "learning_rate": 0.0006,
+ "loss": 4.711943626403809,
+ "step": 3728
+ },
+ {
+ "epoch": 51.79650655021834,
+ "grad_norm": 0.019175555557012558,
+ "learning_rate": 0.0006,
+ "loss": 4.721931457519531,
+ "step": 3729
+ },
+ {
+ "epoch": 51.81048034934498,
+ "grad_norm": 0.02093447372317314,
+ "learning_rate": 0.0006,
+ "loss": 4.633059024810791,
+ "step": 3730
+ },
+ {
+ "epoch": 51.82445414847162,
+ "grad_norm": 0.01930283196270466,
+ "learning_rate": 0.0006,
+ "loss": 4.659373760223389,
+ "step": 3731
+ },
+ {
+ "epoch": 51.838427947598255,
+ "grad_norm": 0.015350676141679287,
+ "learning_rate": 0.0006,
+ "loss": 4.763708114624023,
+ "step": 3732
+ },
+ {
+ "epoch": 51.852401746724894,
+ "grad_norm": 0.014139448292553425,
+ "learning_rate": 0.0006,
+ "loss": 4.753891468048096,
+ "step": 3733
+ },
+ {
+ "epoch": 51.866375545851525,
+ "grad_norm": 0.016215872019529343,
+ "learning_rate": 0.0006,
+ "loss": 4.658967018127441,
+ "step": 3734
+ },
+ {
+ "epoch": 51.880349344978164,
+ "grad_norm": 0.017328619956970215,
+ "learning_rate": 0.0006,
+ "loss": 4.696579933166504,
+ "step": 3735
+ },
+ {
+ "epoch": 51.8943231441048,
+ "grad_norm": 0.018103396520018578,
+ "learning_rate": 0.0006,
+ "loss": 4.627464771270752,
+ "step": 3736
+ },
+ {
+ "epoch": 51.90829694323144,
+ "grad_norm": 0.01715235412120819,
+ "learning_rate": 0.0006,
+ "loss": 4.7368011474609375,
+ "step": 3737
+ },
+ {
+ "epoch": 51.92227074235808,
+ "grad_norm": 0.014790094457566738,
+ "learning_rate": 0.0006,
+ "loss": 4.629441738128662,
+ "step": 3738
+ },
+ {
+ "epoch": 51.93624454148472,
+ "grad_norm": 0.017114296555519104,
+ "learning_rate": 0.0006,
+ "loss": 4.761256694793701,
+ "step": 3739
+ },
+ {
+ "epoch": 51.95021834061135,
+ "grad_norm": 0.019009845331311226,
+ "learning_rate": 0.0006,
+ "loss": 4.789155006408691,
+ "step": 3740
+ },
+ {
+ "epoch": 51.96419213973799,
+ "grad_norm": 0.019696692004799843,
+ "learning_rate": 0.0006,
+ "loss": 4.721269607543945,
+ "step": 3741
+ },
+ {
+ "epoch": 51.97816593886463,
+ "grad_norm": 0.020542822778224945,
+ "learning_rate": 0.0006,
+ "loss": 4.687038898468018,
+ "step": 3742
+ },
+ {
+ "epoch": 51.992139737991266,
+ "grad_norm": 0.021267160773277283,
+ "learning_rate": 0.0006,
+ "loss": 4.747860908508301,
+ "step": 3743
+ },
+ {
+ "epoch": 52.0,
+ "grad_norm": 0.025049567222595215,
+ "learning_rate": 0.0006,
+ "loss": 4.8473100662231445,
+ "step": 3744
+ },
+ {
+ "epoch": 52.0,
+ "eval_loss": 4.943043231964111,
+ "eval_runtime": 56.4929,
+ "eval_samples_per_second": 43.227,
+ "eval_steps_per_second": 1.363,
+ "step": 3744
+ },
+ {
+ "epoch": 52.01397379912664,
+ "grad_norm": 0.03054986707866192,
+ "learning_rate": 0.0006,
+ "loss": 4.856931686401367,
+ "step": 3745
+ },
+ {
+ "epoch": 52.02794759825328,
+ "grad_norm": 0.03838375583291054,
+ "learning_rate": 0.0006,
+ "loss": 4.699807167053223,
+ "step": 3746
+ },
+ {
+ "epoch": 52.041921397379916,
+ "grad_norm": 0.03383488208055496,
+ "learning_rate": 0.0006,
+ "loss": 4.6626787185668945,
+ "step": 3747
+ },
+ {
+ "epoch": 52.05589519650655,
+ "grad_norm": 0.0201561376452446,
+ "learning_rate": 0.0006,
+ "loss": 4.621831893920898,
+ "step": 3748
+ },
+ {
+ "epoch": 52.069868995633186,
+ "grad_norm": 0.028749438002705574,
+ "learning_rate": 0.0006,
+ "loss": 4.682540416717529,
+ "step": 3749
+ },
+ {
+ "epoch": 52.083842794759825,
+ "grad_norm": 0.02681174874305725,
+ "learning_rate": 0.0006,
+ "loss": 4.629997253417969,
+ "step": 3750
+ },
+ {
+ "epoch": 52.09781659388646,
+ "grad_norm": 0.024063318967819214,
+ "learning_rate": 0.0006,
+ "loss": 4.573021411895752,
+ "step": 3751
+ },
+ {
+ "epoch": 52.1117903930131,
+ "grad_norm": 0.02324330434203148,
+ "learning_rate": 0.0006,
+ "loss": 4.739465236663818,
+ "step": 3752
+ },
+ {
+ "epoch": 52.12576419213974,
+ "grad_norm": 0.0225426834076643,
+ "learning_rate": 0.0006,
+ "loss": 4.7335333824157715,
+ "step": 3753
+ },
+ {
+ "epoch": 52.13973799126637,
+ "grad_norm": 0.019593428820371628,
+ "learning_rate": 0.0006,
+ "loss": 4.704958438873291,
+ "step": 3754
+ },
+ {
+ "epoch": 52.15371179039301,
+ "grad_norm": 0.020777082070708275,
+ "learning_rate": 0.0006,
+ "loss": 4.638696670532227,
+ "step": 3755
+ },
+ {
+ "epoch": 52.16768558951965,
+ "grad_norm": 0.021525604650378227,
+ "learning_rate": 0.0006,
+ "loss": 4.643706798553467,
+ "step": 3756
+ },
+ {
+ "epoch": 52.18165938864629,
+ "grad_norm": 0.018943654373288155,
+ "learning_rate": 0.0006,
+ "loss": 4.5173211097717285,
+ "step": 3757
+ },
+ {
+ "epoch": 52.19563318777293,
+ "grad_norm": 0.01856580562889576,
+ "learning_rate": 0.0006,
+ "loss": 4.678788661956787,
+ "step": 3758
+ },
+ {
+ "epoch": 52.209606986899566,
+ "grad_norm": 0.019417259842157364,
+ "learning_rate": 0.0006,
+ "loss": 4.587244987487793,
+ "step": 3759
+ },
+ {
+ "epoch": 52.223580786026204,
+ "grad_norm": 0.01911911740899086,
+ "learning_rate": 0.0006,
+ "loss": 4.688503265380859,
+ "step": 3760
+ },
+ {
+ "epoch": 52.237554585152836,
+ "grad_norm": 0.021017879247665405,
+ "learning_rate": 0.0006,
+ "loss": 4.65829610824585,
+ "step": 3761
+ },
+ {
+ "epoch": 52.251528384279474,
+ "grad_norm": 0.02048875205218792,
+ "learning_rate": 0.0006,
+ "loss": 4.624055862426758,
+ "step": 3762
+ },
+ {
+ "epoch": 52.26550218340611,
+ "grad_norm": 0.01932772807776928,
+ "learning_rate": 0.0006,
+ "loss": 4.556562900543213,
+ "step": 3763
+ },
+ {
+ "epoch": 52.27947598253275,
+ "grad_norm": 0.01871614344418049,
+ "learning_rate": 0.0006,
+ "loss": 4.581201076507568,
+ "step": 3764
+ },
+ {
+ "epoch": 52.29344978165939,
+ "grad_norm": 0.017253106459975243,
+ "learning_rate": 0.0006,
+ "loss": 4.605219841003418,
+ "step": 3765
+ },
+ {
+ "epoch": 52.30742358078603,
+ "grad_norm": 0.01831238530576229,
+ "learning_rate": 0.0006,
+ "loss": 4.759900093078613,
+ "step": 3766
+ },
+ {
+ "epoch": 52.32139737991266,
+ "grad_norm": 0.017946207895874977,
+ "learning_rate": 0.0006,
+ "loss": 4.5598344802856445,
+ "step": 3767
+ },
+ {
+ "epoch": 52.3353711790393,
+ "grad_norm": 0.01680106669664383,
+ "learning_rate": 0.0006,
+ "loss": 4.605034828186035,
+ "step": 3768
+ },
+ {
+ "epoch": 52.34934497816594,
+ "grad_norm": 0.021482540294528008,
+ "learning_rate": 0.0006,
+ "loss": 4.620411396026611,
+ "step": 3769
+ },
+ {
+ "epoch": 52.36331877729258,
+ "grad_norm": 0.02210252732038498,
+ "learning_rate": 0.0006,
+ "loss": 4.667577743530273,
+ "step": 3770
+ },
+ {
+ "epoch": 52.377292576419215,
+ "grad_norm": 0.02237485907971859,
+ "learning_rate": 0.0006,
+ "loss": 4.628812789916992,
+ "step": 3771
+ },
+ {
+ "epoch": 52.391266375545854,
+ "grad_norm": 0.02196541801095009,
+ "learning_rate": 0.0006,
+ "loss": 4.572988510131836,
+ "step": 3772
+ },
+ {
+ "epoch": 52.40524017467249,
+ "grad_norm": 0.01930229924619198,
+ "learning_rate": 0.0006,
+ "loss": 4.588339805603027,
+ "step": 3773
+ },
+ {
+ "epoch": 52.419213973799124,
+ "grad_norm": 0.02222434990108013,
+ "learning_rate": 0.0006,
+ "loss": 4.766482353210449,
+ "step": 3774
+ },
+ {
+ "epoch": 52.43318777292576,
+ "grad_norm": 0.028806306421756744,
+ "learning_rate": 0.0006,
+ "loss": 4.617892742156982,
+ "step": 3775
+ },
+ {
+ "epoch": 52.4471615720524,
+ "grad_norm": 0.03306042402982712,
+ "learning_rate": 0.0006,
+ "loss": 4.665334701538086,
+ "step": 3776
+ },
+ {
+ "epoch": 52.46113537117904,
+ "grad_norm": 0.0358956903219223,
+ "learning_rate": 0.0006,
+ "loss": 4.566668510437012,
+ "step": 3777
+ },
+ {
+ "epoch": 52.47510917030568,
+ "grad_norm": 0.030406497418880463,
+ "learning_rate": 0.0006,
+ "loss": 4.513612270355225,
+ "step": 3778
+ },
+ {
+ "epoch": 52.48908296943232,
+ "grad_norm": 0.030295712873339653,
+ "learning_rate": 0.0006,
+ "loss": 4.640189170837402,
+ "step": 3779
+ },
+ {
+ "epoch": 52.50305676855895,
+ "grad_norm": 0.024981003254652023,
+ "learning_rate": 0.0006,
+ "loss": 4.57558012008667,
+ "step": 3780
+ },
+ {
+ "epoch": 52.51703056768559,
+ "grad_norm": 0.024990713223814964,
+ "learning_rate": 0.0006,
+ "loss": 4.708589553833008,
+ "step": 3781
+ },
+ {
+ "epoch": 52.531004366812226,
+ "grad_norm": 0.025034775957465172,
+ "learning_rate": 0.0006,
+ "loss": 4.70468282699585,
+ "step": 3782
+ },
+ {
+ "epoch": 52.544978165938865,
+ "grad_norm": 0.028243759647011757,
+ "learning_rate": 0.0006,
+ "loss": 4.712060451507568,
+ "step": 3783
+ },
+ {
+ "epoch": 52.5589519650655,
+ "grad_norm": 0.026984544470906258,
+ "learning_rate": 0.0006,
+ "loss": 4.505004405975342,
+ "step": 3784
+ },
+ {
+ "epoch": 52.57292576419214,
+ "grad_norm": 0.050843529403209686,
+ "learning_rate": 0.0006,
+ "loss": 4.593271255493164,
+ "step": 3785
+ },
+ {
+ "epoch": 52.58689956331878,
+ "grad_norm": 0.25422802567481995,
+ "learning_rate": 0.0006,
+ "loss": 4.7178802490234375,
+ "step": 3786
+ },
+ {
+ "epoch": 52.60087336244541,
+ "grad_norm": 0.6078406572341919,
+ "learning_rate": 0.0006,
+ "loss": 6.144372940063477,
+ "step": 3787
+ },
+ {
+ "epoch": 52.61484716157205,
+ "grad_norm": 0.25933101773262024,
+ "learning_rate": 0.0006,
+ "loss": 5.971083641052246,
+ "step": 3788
+ },
+ {
+ "epoch": 52.62882096069869,
+ "grad_norm": 0.294657826423645,
+ "learning_rate": 0.0006,
+ "loss": 6.204090118408203,
+ "step": 3789
+ },
+ {
+ "epoch": 52.64279475982533,
+ "grad_norm": 0.23641522228717804,
+ "learning_rate": 0.0006,
+ "loss": 6.869574546813965,
+ "step": 3790
+ },
+ {
+ "epoch": 52.65676855895197,
+ "grad_norm": 0.23717311024665833,
+ "learning_rate": 0.0006,
+ "loss": 6.739034175872803,
+ "step": 3791
+ },
+ {
+ "epoch": 52.670742358078606,
+ "grad_norm": 0.10300362855195999,
+ "learning_rate": 0.0006,
+ "loss": 6.49191951751709,
+ "step": 3792
+ },
+ {
+ "epoch": 52.68471615720524,
+ "grad_norm": 0.1484968513250351,
+ "learning_rate": 0.0006,
+ "loss": 6.6021270751953125,
+ "step": 3793
+ },
+ {
+ "epoch": 52.698689956331876,
+ "grad_norm": 0.08052115142345428,
+ "learning_rate": 0.0006,
+ "loss": 6.432796478271484,
+ "step": 3794
+ },
+ {
+ "epoch": 52.712663755458514,
+ "grad_norm": 0.06869253516197205,
+ "learning_rate": 0.0006,
+ "loss": 6.212784767150879,
+ "step": 3795
+ },
+ {
+ "epoch": 52.72663755458515,
+ "grad_norm": 0.05589582398533821,
+ "learning_rate": 0.0006,
+ "loss": 6.168288707733154,
+ "step": 3796
+ },
+ {
+ "epoch": 52.74061135371179,
+ "grad_norm": 0.04840940609574318,
+ "learning_rate": 0.0006,
+ "loss": 6.163174629211426,
+ "step": 3797
+ },
+ {
+ "epoch": 52.75458515283843,
+ "grad_norm": 0.04578601196408272,
+ "learning_rate": 0.0006,
+ "loss": 6.038390636444092,
+ "step": 3798
+ },
+ {
+ "epoch": 52.76855895196506,
+ "grad_norm": 0.04102053865790367,
+ "learning_rate": 0.0006,
+ "loss": 5.929539680480957,
+ "step": 3799
+ },
+ {
+ "epoch": 52.7825327510917,
+ "grad_norm": 0.038398485630750656,
+ "learning_rate": 0.0006,
+ "loss": 6.007378101348877,
+ "step": 3800
+ },
+ {
+ "epoch": 52.79650655021834,
+ "grad_norm": 0.03466642275452614,
+ "learning_rate": 0.0006,
+ "loss": 5.767291069030762,
+ "step": 3801
+ },
+ {
+ "epoch": 52.81048034934498,
+ "grad_norm": 0.03328905999660492,
+ "learning_rate": 0.0006,
+ "loss": 5.785032272338867,
+ "step": 3802
+ },
+ {
+ "epoch": 52.82445414847162,
+ "grad_norm": 0.03186681121587753,
+ "learning_rate": 0.0006,
+ "loss": 5.6942057609558105,
+ "step": 3803
+ },
+ {
+ "epoch": 52.838427947598255,
+ "grad_norm": 0.030553551390767097,
+ "learning_rate": 0.0006,
+ "loss": 5.680774688720703,
+ "step": 3804
+ },
+ {
+ "epoch": 52.852401746724894,
+ "grad_norm": 0.02578839473426342,
+ "learning_rate": 0.0006,
+ "loss": 5.703461647033691,
+ "step": 3805
+ },
+ {
+ "epoch": 52.866375545851525,
+ "grad_norm": 0.026261312887072563,
+ "learning_rate": 0.0006,
+ "loss": 5.642549514770508,
+ "step": 3806
+ },
+ {
+ "epoch": 52.880349344978164,
+ "grad_norm": 0.02782430872321129,
+ "learning_rate": 0.0006,
+ "loss": 5.563597679138184,
+ "step": 3807
+ },
+ {
+ "epoch": 52.8943231441048,
+ "grad_norm": 0.024467799812555313,
+ "learning_rate": 0.0006,
+ "loss": 5.583395957946777,
+ "step": 3808
+ },
+ {
+ "epoch": 52.90829694323144,
+ "grad_norm": 0.022602897137403488,
+ "learning_rate": 0.0006,
+ "loss": 5.50742769241333,
+ "step": 3809
+ },
+ {
+ "epoch": 52.92227074235808,
+ "grad_norm": 0.024292029440402985,
+ "learning_rate": 0.0006,
+ "loss": 5.367804527282715,
+ "step": 3810
+ },
+ {
+ "epoch": 52.93624454148472,
+ "grad_norm": 0.021062295883893967,
+ "learning_rate": 0.0006,
+ "loss": 5.423705101013184,
+ "step": 3811
+ },
+ {
+ "epoch": 52.95021834061135,
+ "grad_norm": 0.022226469591259956,
+ "learning_rate": 0.0006,
+ "loss": 5.344039440155029,
+ "step": 3812
+ },
+ {
+ "epoch": 52.96419213973799,
+ "grad_norm": 0.023691752925515175,
+ "learning_rate": 0.0006,
+ "loss": 5.2581892013549805,
+ "step": 3813
+ },
+ {
+ "epoch": 52.97816593886463,
+ "grad_norm": 0.02181999757885933,
+ "learning_rate": 0.0006,
+ "loss": 5.2710113525390625,
+ "step": 3814
+ },
+ {
+ "epoch": 52.992139737991266,
+ "grad_norm": 0.025295857340097427,
+ "learning_rate": 0.0006,
+ "loss": 5.278355598449707,
+ "step": 3815
+ },
+ {
+ "epoch": 53.0,
+ "grad_norm": 0.02216828241944313,
+ "learning_rate": 0.0006,
+ "loss": 5.127945899963379,
+ "step": 3816
+ },
+ {
+ "epoch": 53.0,
+ "eval_loss": 5.3411736488342285,
+ "eval_runtime": 56.2837,
+ "eval_samples_per_second": 43.387,
+ "eval_steps_per_second": 1.368,
+ "step": 3816
+ },
+ {
+ "epoch": 53.01397379912664,
+ "grad_norm": 0.022496026009321213,
+ "learning_rate": 0.0006,
+ "loss": 5.19655704498291,
+ "step": 3817
+ },
+ {
+ "epoch": 53.02794759825328,
+ "grad_norm": 0.03201168775558472,
+ "learning_rate": 0.0006,
+ "loss": 5.094621181488037,
+ "step": 3818
+ },
+ {
+ "epoch": 53.041921397379916,
+ "grad_norm": 0.02683178335428238,
+ "learning_rate": 0.0006,
+ "loss": 5.19899845123291,
+ "step": 3819
+ },
+ {
+ "epoch": 53.05589519650655,
+ "grad_norm": 0.022659169510006905,
+ "learning_rate": 0.0006,
+ "loss": 5.184103012084961,
+ "step": 3820
+ },
+ {
+ "epoch": 53.069868995633186,
+ "grad_norm": 0.024784443899989128,
+ "learning_rate": 0.0006,
+ "loss": 5.124286651611328,
+ "step": 3821
+ },
+ {
+ "epoch": 53.083842794759825,
+ "grad_norm": 0.028710370883345604,
+ "learning_rate": 0.0006,
+ "loss": 5.130426406860352,
+ "step": 3822
+ },
+ {
+ "epoch": 53.09781659388646,
+ "grad_norm": 0.02366088517010212,
+ "learning_rate": 0.0006,
+ "loss": 5.1453399658203125,
+ "step": 3823
+ },
+ {
+ "epoch": 53.1117903930131,
+ "grad_norm": 0.01813938096165657,
+ "learning_rate": 0.0006,
+ "loss": 5.05271053314209,
+ "step": 3824
+ },
+ {
+ "epoch": 53.12576419213974,
+ "grad_norm": 0.02051435597240925,
+ "learning_rate": 0.0006,
+ "loss": 5.023038864135742,
+ "step": 3825
+ },
+ {
+ "epoch": 53.13973799126637,
+ "grad_norm": 0.02185754105448723,
+ "learning_rate": 0.0006,
+ "loss": 4.8950018882751465,
+ "step": 3826
+ },
+ {
+ "epoch": 53.15371179039301,
+ "grad_norm": 0.017223743721842766,
+ "learning_rate": 0.0006,
+ "loss": 5.078621864318848,
+ "step": 3827
+ },
+ {
+ "epoch": 53.16768558951965,
+ "grad_norm": 0.01832217164337635,
+ "learning_rate": 0.0006,
+ "loss": 4.921341419219971,
+ "step": 3828
+ },
+ {
+ "epoch": 53.18165938864629,
+ "grad_norm": 0.02114005573093891,
+ "learning_rate": 0.0006,
+ "loss": 4.997865676879883,
+ "step": 3829
+ },
+ {
+ "epoch": 53.19563318777293,
+ "grad_norm": 0.021426431834697723,
+ "learning_rate": 0.0006,
+ "loss": 4.877785682678223,
+ "step": 3830
+ },
+ {
+ "epoch": 53.209606986899566,
+ "grad_norm": 0.021327560767531395,
+ "learning_rate": 0.0006,
+ "loss": 4.847512245178223,
+ "step": 3831
+ },
+ {
+ "epoch": 53.223580786026204,
+ "grad_norm": 0.027825530618429184,
+ "learning_rate": 0.0006,
+ "loss": 5.004823207855225,
+ "step": 3832
+ },
+ {
+ "epoch": 53.237554585152836,
+ "grad_norm": 0.03756481409072876,
+ "learning_rate": 0.0006,
+ "loss": 4.817290306091309,
+ "step": 3833
+ },
+ {
+ "epoch": 53.251528384279474,
+ "grad_norm": 0.04933721572160721,
+ "learning_rate": 0.0006,
+ "loss": 4.863317489624023,
+ "step": 3834
+ },
+ {
+ "epoch": 53.26550218340611,
+ "grad_norm": 0.04588694870471954,
+ "learning_rate": 0.0006,
+ "loss": 4.864742755889893,
+ "step": 3835
+ },
+ {
+ "epoch": 53.27947598253275,
+ "grad_norm": 0.02728264033794403,
+ "learning_rate": 0.0006,
+ "loss": 4.9068756103515625,
+ "step": 3836
+ },
+ {
+ "epoch": 53.29344978165939,
+ "grad_norm": 0.036240726709365845,
+ "learning_rate": 0.0006,
+ "loss": 4.902045249938965,
+ "step": 3837
+ },
+ {
+ "epoch": 53.30742358078603,
+ "grad_norm": 0.02703195810317993,
+ "learning_rate": 0.0006,
+ "loss": 4.908980369567871,
+ "step": 3838
+ },
+ {
+ "epoch": 53.32139737991266,
+ "grad_norm": 0.024894049391150475,
+ "learning_rate": 0.0006,
+ "loss": 4.794848442077637,
+ "step": 3839
+ },
+ {
+ "epoch": 53.3353711790393,
+ "grad_norm": 0.02389993891119957,
+ "learning_rate": 0.0006,
+ "loss": 4.8234734535217285,
+ "step": 3840
+ },
+ {
+ "epoch": 53.34934497816594,
+ "grad_norm": 0.02486400306224823,
+ "learning_rate": 0.0006,
+ "loss": 4.851463794708252,
+ "step": 3841
+ },
+ {
+ "epoch": 53.36331877729258,
+ "grad_norm": 0.02613813802599907,
+ "learning_rate": 0.0006,
+ "loss": 4.860444068908691,
+ "step": 3842
+ },
+ {
+ "epoch": 53.377292576419215,
+ "grad_norm": 0.0229744054377079,
+ "learning_rate": 0.0006,
+ "loss": 4.709606170654297,
+ "step": 3843
+ },
+ {
+ "epoch": 53.391266375545854,
+ "grad_norm": 0.02123042568564415,
+ "learning_rate": 0.0006,
+ "loss": 4.654613494873047,
+ "step": 3844
+ },
+ {
+ "epoch": 53.40524017467249,
+ "grad_norm": 0.017352338880300522,
+ "learning_rate": 0.0006,
+ "loss": 4.769761085510254,
+ "step": 3845
+ },
+ {
+ "epoch": 53.419213973799124,
+ "grad_norm": 0.019808724522590637,
+ "learning_rate": 0.0006,
+ "loss": 4.98514461517334,
+ "step": 3846
+ },
+ {
+ "epoch": 53.43318777292576,
+ "grad_norm": 0.018776152282953262,
+ "learning_rate": 0.0006,
+ "loss": 4.749990463256836,
+ "step": 3847
+ },
+ {
+ "epoch": 53.4471615720524,
+ "grad_norm": 0.02017521858215332,
+ "learning_rate": 0.0006,
+ "loss": 4.757381439208984,
+ "step": 3848
+ },
+ {
+ "epoch": 53.46113537117904,
+ "grad_norm": 0.019509416073560715,
+ "learning_rate": 0.0006,
+ "loss": 4.8111572265625,
+ "step": 3849
+ },
+ {
+ "epoch": 53.47510917030568,
+ "grad_norm": 0.015543580986559391,
+ "learning_rate": 0.0006,
+ "loss": 4.7501678466796875,
+ "step": 3850
+ },
+ {
+ "epoch": 53.48908296943232,
+ "grad_norm": 0.017941651865839958,
+ "learning_rate": 0.0006,
+ "loss": 4.701838493347168,
+ "step": 3851
+ },
+ {
+ "epoch": 53.50305676855895,
+ "grad_norm": 0.01678212359547615,
+ "learning_rate": 0.0006,
+ "loss": 4.787798881530762,
+ "step": 3852
+ },
+ {
+ "epoch": 53.51703056768559,
+ "grad_norm": 0.019050654023885727,
+ "learning_rate": 0.0006,
+ "loss": 4.754859924316406,
+ "step": 3853
+ },
+ {
+ "epoch": 53.531004366812226,
+ "grad_norm": 0.017406078055500984,
+ "learning_rate": 0.0006,
+ "loss": 4.752772808074951,
+ "step": 3854
+ },
+ {
+ "epoch": 53.544978165938865,
+ "grad_norm": 0.01773344725370407,
+ "learning_rate": 0.0006,
+ "loss": 4.651812553405762,
+ "step": 3855
+ },
+ {
+ "epoch": 53.5589519650655,
+ "grad_norm": 0.017221346497535706,
+ "learning_rate": 0.0006,
+ "loss": 4.792244911193848,
+ "step": 3856
+ },
+ {
+ "epoch": 53.57292576419214,
+ "grad_norm": 0.01722758449614048,
+ "learning_rate": 0.0006,
+ "loss": 4.7902960777282715,
+ "step": 3857
+ },
+ {
+ "epoch": 53.58689956331878,
+ "grad_norm": 0.020445192232728004,
+ "learning_rate": 0.0006,
+ "loss": 4.67392110824585,
+ "step": 3858
+ },
+ {
+ "epoch": 53.60087336244541,
+ "grad_norm": 0.02267114631831646,
+ "learning_rate": 0.0006,
+ "loss": 4.796089172363281,
+ "step": 3859
+ },
+ {
+ "epoch": 53.61484716157205,
+ "grad_norm": 0.027129942551255226,
+ "learning_rate": 0.0006,
+ "loss": 4.774303913116455,
+ "step": 3860
+ },
+ {
+ "epoch": 53.62882096069869,
+ "grad_norm": 0.025454839691519737,
+ "learning_rate": 0.0006,
+ "loss": 4.708813190460205,
+ "step": 3861
+ },
+ {
+ "epoch": 53.64279475982533,
+ "grad_norm": 0.019248757511377335,
+ "learning_rate": 0.0006,
+ "loss": 4.760467529296875,
+ "step": 3862
+ },
+ {
+ "epoch": 53.65676855895197,
+ "grad_norm": 0.01791176199913025,
+ "learning_rate": 0.0006,
+ "loss": 4.789203643798828,
+ "step": 3863
+ },
+ {
+ "epoch": 53.670742358078606,
+ "grad_norm": 0.016306612640619278,
+ "learning_rate": 0.0006,
+ "loss": 4.732605934143066,
+ "step": 3864
+ },
+ {
+ "epoch": 53.68471615720524,
+ "grad_norm": 0.017836976796388626,
+ "learning_rate": 0.0006,
+ "loss": 4.691871643066406,
+ "step": 3865
+ },
+ {
+ "epoch": 53.698689956331876,
+ "grad_norm": 0.015102861449122429,
+ "learning_rate": 0.0006,
+ "loss": 4.724156856536865,
+ "step": 3866
+ },
+ {
+ "epoch": 53.712663755458514,
+ "grad_norm": 0.017458142712712288,
+ "learning_rate": 0.0006,
+ "loss": 4.73524808883667,
+ "step": 3867
+ },
+ {
+ "epoch": 53.72663755458515,
+ "grad_norm": 0.01591932773590088,
+ "learning_rate": 0.0006,
+ "loss": 4.705888748168945,
+ "step": 3868
+ },
+ {
+ "epoch": 53.74061135371179,
+ "grad_norm": 0.01713593676686287,
+ "learning_rate": 0.0006,
+ "loss": 4.644516944885254,
+ "step": 3869
+ },
+ {
+ "epoch": 53.75458515283843,
+ "grad_norm": 0.01905885338783264,
+ "learning_rate": 0.0006,
+ "loss": 4.698416233062744,
+ "step": 3870
+ },
+ {
+ "epoch": 53.76855895196506,
+ "grad_norm": 0.01766437292098999,
+ "learning_rate": 0.0006,
+ "loss": 4.6796555519104,
+ "step": 3871
+ },
+ {
+ "epoch": 53.7825327510917,
+ "grad_norm": 0.01552243810147047,
+ "learning_rate": 0.0006,
+ "loss": 4.78138542175293,
+ "step": 3872
+ },
+ {
+ "epoch": 53.79650655021834,
+ "grad_norm": 0.014191331341862679,
+ "learning_rate": 0.0006,
+ "loss": 4.703588485717773,
+ "step": 3873
+ },
+ {
+ "epoch": 53.81048034934498,
+ "grad_norm": 0.015872279182076454,
+ "learning_rate": 0.0006,
+ "loss": 4.705611228942871,
+ "step": 3874
+ },
+ {
+ "epoch": 53.82445414847162,
+ "grad_norm": 0.01787560060620308,
+ "learning_rate": 0.0006,
+ "loss": 4.8227949142456055,
+ "step": 3875
+ },
+ {
+ "epoch": 53.838427947598255,
+ "grad_norm": 0.0158209390938282,
+ "learning_rate": 0.0006,
+ "loss": 4.645227432250977,
+ "step": 3876
+ },
+ {
+ "epoch": 53.852401746724894,
+ "grad_norm": 0.014362678863108158,
+ "learning_rate": 0.0006,
+ "loss": 4.7225422859191895,
+ "step": 3877
+ },
+ {
+ "epoch": 53.866375545851525,
+ "grad_norm": 0.014899161644279957,
+ "learning_rate": 0.0006,
+ "loss": 4.685914039611816,
+ "step": 3878
+ },
+ {
+ "epoch": 53.880349344978164,
+ "grad_norm": 0.01628505438566208,
+ "learning_rate": 0.0006,
+ "loss": 4.674065589904785,
+ "step": 3879
+ },
+ {
+ "epoch": 53.8943231441048,
+ "grad_norm": 0.01873331144452095,
+ "learning_rate": 0.0006,
+ "loss": 4.811744213104248,
+ "step": 3880
+ },
+ {
+ "epoch": 53.90829694323144,
+ "grad_norm": 0.02229069173336029,
+ "learning_rate": 0.0006,
+ "loss": 4.721514701843262,
+ "step": 3881
+ },
+ {
+ "epoch": 53.92227074235808,
+ "grad_norm": 0.028012916445732117,
+ "learning_rate": 0.0006,
+ "loss": 4.773603439331055,
+ "step": 3882
+ },
+ {
+ "epoch": 53.93624454148472,
+ "grad_norm": 0.036322783678770065,
+ "learning_rate": 0.0006,
+ "loss": 4.787134170532227,
+ "step": 3883
+ },
+ {
+ "epoch": 53.95021834061135,
+ "grad_norm": 0.048149142414331436,
+ "learning_rate": 0.0006,
+ "loss": 4.736196517944336,
+ "step": 3884
+ },
+ {
+ "epoch": 53.96419213973799,
+ "grad_norm": 0.051918916404247284,
+ "learning_rate": 0.0006,
+ "loss": 4.686728000640869,
+ "step": 3885
+ },
+ {
+ "epoch": 53.97816593886463,
+ "grad_norm": 0.0481342189013958,
+ "learning_rate": 0.0006,
+ "loss": 4.733730792999268,
+ "step": 3886
+ },
+ {
+ "epoch": 53.992139737991266,
+ "grad_norm": 0.045424818992614746,
+ "learning_rate": 0.0006,
+ "loss": 4.748723983764648,
+ "step": 3887
+ },
+ {
+ "epoch": 54.0,
+ "grad_norm": 0.04153316840529442,
+ "learning_rate": 0.0006,
+ "loss": 4.715220928192139,
+ "step": 3888
+ },
+ {
+ "epoch": 54.0,
+ "eval_loss": 4.964215278625488,
+ "eval_runtime": 56.7757,
+ "eval_samples_per_second": 43.011,
+ "eval_steps_per_second": 1.356,
+ "step": 3888
+ },
+ {
+ "epoch": 54.01397379912664,
+ "grad_norm": 0.03459750860929489,
+ "learning_rate": 0.0006,
+ "loss": 4.661557674407959,
+ "step": 3889
+ },
+ {
+ "epoch": 54.02794759825328,
+ "grad_norm": 0.025409208610653877,
+ "learning_rate": 0.0006,
+ "loss": 4.7065582275390625,
+ "step": 3890
+ },
+ {
+ "epoch": 54.041921397379916,
+ "grad_norm": 0.03202550485730171,
+ "learning_rate": 0.0006,
+ "loss": 4.751070976257324,
+ "step": 3891
+ },
+ {
+ "epoch": 54.05589519650655,
+ "grad_norm": 0.02968018501996994,
+ "learning_rate": 0.0006,
+ "loss": 4.635388374328613,
+ "step": 3892
+ },
+ {
+ "epoch": 54.069868995633186,
+ "grad_norm": 0.02757406234741211,
+ "learning_rate": 0.0006,
+ "loss": 4.706214427947998,
+ "step": 3893
+ },
+ {
+ "epoch": 54.083842794759825,
+ "grad_norm": 0.025223620235919952,
+ "learning_rate": 0.0006,
+ "loss": 4.703792572021484,
+ "step": 3894
+ },
+ {
+ "epoch": 54.09781659388646,
+ "grad_norm": 0.02431778982281685,
+ "learning_rate": 0.0006,
+ "loss": 4.658116340637207,
+ "step": 3895
+ },
+ {
+ "epoch": 54.1117903930131,
+ "grad_norm": 0.02792172320187092,
+ "learning_rate": 0.0006,
+ "loss": 4.638895034790039,
+ "step": 3896
+ },
+ {
+ "epoch": 54.12576419213974,
+ "grad_norm": 0.032253511250019073,
+ "learning_rate": 0.0006,
+ "loss": 4.693902492523193,
+ "step": 3897
+ },
+ {
+ "epoch": 54.13973799126637,
+ "grad_norm": 0.027044931426644325,
+ "learning_rate": 0.0006,
+ "loss": 4.575805187225342,
+ "step": 3898
+ },
+ {
+ "epoch": 54.15371179039301,
+ "grad_norm": 0.023270361125469208,
+ "learning_rate": 0.0006,
+ "loss": 4.67645263671875,
+ "step": 3899
+ },
+ {
+ "epoch": 54.16768558951965,
+ "grad_norm": 0.01929079182446003,
+ "learning_rate": 0.0006,
+ "loss": 4.68062686920166,
+ "step": 3900
+ },
+ {
+ "epoch": 54.18165938864629,
+ "grad_norm": 0.020343473181128502,
+ "learning_rate": 0.0006,
+ "loss": 4.708086967468262,
+ "step": 3901
+ },
+ {
+ "epoch": 54.19563318777293,
+ "grad_norm": 0.0211793202906847,
+ "learning_rate": 0.0006,
+ "loss": 4.677323341369629,
+ "step": 3902
+ },
+ {
+ "epoch": 54.209606986899566,
+ "grad_norm": 0.020341217517852783,
+ "learning_rate": 0.0006,
+ "loss": 4.760937690734863,
+ "step": 3903
+ },
+ {
+ "epoch": 54.223580786026204,
+ "grad_norm": 0.019421054050326347,
+ "learning_rate": 0.0006,
+ "loss": 4.508834362030029,
+ "step": 3904
+ },
+ {
+ "epoch": 54.237554585152836,
+ "grad_norm": 0.016701679676771164,
+ "learning_rate": 0.0006,
+ "loss": 4.664399147033691,
+ "step": 3905
+ },
+ {
+ "epoch": 54.251528384279474,
+ "grad_norm": 0.016561470925807953,
+ "learning_rate": 0.0006,
+ "loss": 4.742990493774414,
+ "step": 3906
+ },
+ {
+ "epoch": 54.26550218340611,
+ "grad_norm": 0.017446085810661316,
+ "learning_rate": 0.0006,
+ "loss": 4.63038444519043,
+ "step": 3907
+ },
+ {
+ "epoch": 54.27947598253275,
+ "grad_norm": 0.014340086840093136,
+ "learning_rate": 0.0006,
+ "loss": 4.688791275024414,
+ "step": 3908
+ },
+ {
+ "epoch": 54.29344978165939,
+ "grad_norm": 0.013580698519945145,
+ "learning_rate": 0.0006,
+ "loss": 4.625302314758301,
+ "step": 3909
+ },
+ {
+ "epoch": 54.30742358078603,
+ "grad_norm": 0.014931539073586464,
+ "learning_rate": 0.0006,
+ "loss": 4.6167497634887695,
+ "step": 3910
+ },
+ {
+ "epoch": 54.32139737991266,
+ "grad_norm": 0.013492150232195854,
+ "learning_rate": 0.0006,
+ "loss": 4.657521724700928,
+ "step": 3911
+ },
+ {
+ "epoch": 54.3353711790393,
+ "grad_norm": 0.015151413157582283,
+ "learning_rate": 0.0006,
+ "loss": 4.634521484375,
+ "step": 3912
+ },
+ {
+ "epoch": 54.34934497816594,
+ "grad_norm": 0.01302040833979845,
+ "learning_rate": 0.0006,
+ "loss": 4.718928337097168,
+ "step": 3913
+ },
+ {
+ "epoch": 54.36331877729258,
+ "grad_norm": 0.013433529995381832,
+ "learning_rate": 0.0006,
+ "loss": 4.584803581237793,
+ "step": 3914
+ },
+ {
+ "epoch": 54.377292576419215,
+ "grad_norm": 0.012987039051949978,
+ "learning_rate": 0.0006,
+ "loss": 4.6393585205078125,
+ "step": 3915
+ },
+ {
+ "epoch": 54.391266375545854,
+ "grad_norm": 0.013695075176656246,
+ "learning_rate": 0.0006,
+ "loss": 4.556313514709473,
+ "step": 3916
+ },
+ {
+ "epoch": 54.40524017467249,
+ "grad_norm": 0.013718255795538425,
+ "learning_rate": 0.0006,
+ "loss": 4.576223850250244,
+ "step": 3917
+ },
+ {
+ "epoch": 54.419213973799124,
+ "grad_norm": 0.013699112460017204,
+ "learning_rate": 0.0006,
+ "loss": 4.63889217376709,
+ "step": 3918
+ },
+ {
+ "epoch": 54.43318777292576,
+ "grad_norm": 0.013516264036297798,
+ "learning_rate": 0.0006,
+ "loss": 4.669532299041748,
+ "step": 3919
+ },
+ {
+ "epoch": 54.4471615720524,
+ "grad_norm": 0.013597294688224792,
+ "learning_rate": 0.0006,
+ "loss": 4.6644287109375,
+ "step": 3920
+ },
+ {
+ "epoch": 54.46113537117904,
+ "grad_norm": 0.013185849413275719,
+ "learning_rate": 0.0006,
+ "loss": 4.636900901794434,
+ "step": 3921
+ },
+ {
+ "epoch": 54.47510917030568,
+ "grad_norm": 0.014311990700662136,
+ "learning_rate": 0.0006,
+ "loss": 4.6047587394714355,
+ "step": 3922
+ },
+ {
+ "epoch": 54.48908296943232,
+ "grad_norm": 0.012426027096807957,
+ "learning_rate": 0.0006,
+ "loss": 4.6161394119262695,
+ "step": 3923
+ },
+ {
+ "epoch": 54.50305676855895,
+ "grad_norm": 0.01283493172377348,
+ "learning_rate": 0.0006,
+ "loss": 4.584190368652344,
+ "step": 3924
+ },
+ {
+ "epoch": 54.51703056768559,
+ "grad_norm": 0.01339943427592516,
+ "learning_rate": 0.0006,
+ "loss": 4.624184608459473,
+ "step": 3925
+ },
+ {
+ "epoch": 54.531004366812226,
+ "grad_norm": 0.012859724462032318,
+ "learning_rate": 0.0006,
+ "loss": 4.6174635887146,
+ "step": 3926
+ },
+ {
+ "epoch": 54.544978165938865,
+ "grad_norm": 0.014420023187994957,
+ "learning_rate": 0.0006,
+ "loss": 4.536544322967529,
+ "step": 3927
+ },
+ {
+ "epoch": 54.5589519650655,
+ "grad_norm": 0.014237185008823872,
+ "learning_rate": 0.0006,
+ "loss": 4.681514739990234,
+ "step": 3928
+ },
+ {
+ "epoch": 54.57292576419214,
+ "grad_norm": 0.013340328820049763,
+ "learning_rate": 0.0006,
+ "loss": 4.579689979553223,
+ "step": 3929
+ },
+ {
+ "epoch": 54.58689956331878,
+ "grad_norm": 0.013894530944526196,
+ "learning_rate": 0.0006,
+ "loss": 4.638958930969238,
+ "step": 3930
+ },
+ {
+ "epoch": 54.60087336244541,
+ "grad_norm": 0.016181141138076782,
+ "learning_rate": 0.0006,
+ "loss": 4.573508262634277,
+ "step": 3931
+ },
+ {
+ "epoch": 54.61484716157205,
+ "grad_norm": 0.013221992179751396,
+ "learning_rate": 0.0006,
+ "loss": 4.6229047775268555,
+ "step": 3932
+ },
+ {
+ "epoch": 54.62882096069869,
+ "grad_norm": 0.013241561129689217,
+ "learning_rate": 0.0006,
+ "loss": 4.641152858734131,
+ "step": 3933
+ },
+ {
+ "epoch": 54.64279475982533,
+ "grad_norm": 0.016237538307905197,
+ "learning_rate": 0.0006,
+ "loss": 4.6421217918396,
+ "step": 3934
+ },
+ {
+ "epoch": 54.65676855895197,
+ "grad_norm": 0.020866507664322853,
+ "learning_rate": 0.0006,
+ "loss": 4.646058082580566,
+ "step": 3935
+ },
+ {
+ "epoch": 54.670742358078606,
+ "grad_norm": 0.022182893007993698,
+ "learning_rate": 0.0006,
+ "loss": 4.554501533508301,
+ "step": 3936
+ },
+ {
+ "epoch": 54.68471615720524,
+ "grad_norm": 0.02254868485033512,
+ "learning_rate": 0.0006,
+ "loss": 4.563869953155518,
+ "step": 3937
+ },
+ {
+ "epoch": 54.698689956331876,
+ "grad_norm": 0.019730940461158752,
+ "learning_rate": 0.0006,
+ "loss": 4.592165946960449,
+ "step": 3938
+ },
+ {
+ "epoch": 54.712663755458514,
+ "grad_norm": 0.015146561898291111,
+ "learning_rate": 0.0006,
+ "loss": 4.6228742599487305,
+ "step": 3939
+ },
+ {
+ "epoch": 54.72663755458515,
+ "grad_norm": 0.01824716106057167,
+ "learning_rate": 0.0006,
+ "loss": 4.716236114501953,
+ "step": 3940
+ },
+ {
+ "epoch": 54.74061135371179,
+ "grad_norm": 0.017730550840497017,
+ "learning_rate": 0.0006,
+ "loss": 4.650232315063477,
+ "step": 3941
+ },
+ {
+ "epoch": 54.75458515283843,
+ "grad_norm": 0.014875595457851887,
+ "learning_rate": 0.0006,
+ "loss": 4.6377458572387695,
+ "step": 3942
+ },
+ {
+ "epoch": 54.76855895196506,
+ "grad_norm": 0.01613209955394268,
+ "learning_rate": 0.0006,
+ "loss": 4.5768537521362305,
+ "step": 3943
+ },
+ {
+ "epoch": 54.7825327510917,
+ "grad_norm": 0.017311280593276024,
+ "learning_rate": 0.0006,
+ "loss": 4.625744342803955,
+ "step": 3944
+ },
+ {
+ "epoch": 54.79650655021834,
+ "grad_norm": 0.018770085647702217,
+ "learning_rate": 0.0006,
+ "loss": 4.613787651062012,
+ "step": 3945
+ },
+ {
+ "epoch": 54.81048034934498,
+ "grad_norm": 0.0165257528424263,
+ "learning_rate": 0.0006,
+ "loss": 4.5991315841674805,
+ "step": 3946
+ },
+ {
+ "epoch": 54.82445414847162,
+ "grad_norm": 0.017433857545256615,
+ "learning_rate": 0.0006,
+ "loss": 4.513872146606445,
+ "step": 3947
+ },
+ {
+ "epoch": 54.838427947598255,
+ "grad_norm": 0.020126905292272568,
+ "learning_rate": 0.0006,
+ "loss": 4.616185188293457,
+ "step": 3948
+ },
+ {
+ "epoch": 54.852401746724894,
+ "grad_norm": 0.01819358579814434,
+ "learning_rate": 0.0006,
+ "loss": 4.535691261291504,
+ "step": 3949
+ },
+ {
+ "epoch": 54.866375545851525,
+ "grad_norm": 0.01899927295744419,
+ "learning_rate": 0.0006,
+ "loss": 4.595199108123779,
+ "step": 3950
+ },
+ {
+ "epoch": 54.880349344978164,
+ "grad_norm": 0.02005566656589508,
+ "learning_rate": 0.0006,
+ "loss": 4.633289337158203,
+ "step": 3951
+ },
+ {
+ "epoch": 54.8943231441048,
+ "grad_norm": 0.021570220589637756,
+ "learning_rate": 0.0006,
+ "loss": 4.592503547668457,
+ "step": 3952
+ },
+ {
+ "epoch": 54.90829694323144,
+ "grad_norm": 0.01883506216108799,
+ "learning_rate": 0.0006,
+ "loss": 4.6597795486450195,
+ "step": 3953
+ },
+ {
+ "epoch": 54.92227074235808,
+ "grad_norm": 0.018070021644234657,
+ "learning_rate": 0.0006,
+ "loss": 4.601474285125732,
+ "step": 3954
+ },
+ {
+ "epoch": 54.93624454148472,
+ "grad_norm": 0.018090378493070602,
+ "learning_rate": 0.0006,
+ "loss": 4.569846153259277,
+ "step": 3955
+ },
+ {
+ "epoch": 54.95021834061135,
+ "grad_norm": 0.018751563504338264,
+ "learning_rate": 0.0006,
+ "loss": 4.646064281463623,
+ "step": 3956
+ },
+ {
+ "epoch": 54.96419213973799,
+ "grad_norm": 0.017512254416942596,
+ "learning_rate": 0.0006,
+ "loss": 4.606670379638672,
+ "step": 3957
+ },
+ {
+ "epoch": 54.97816593886463,
+ "grad_norm": 0.018771110102534294,
+ "learning_rate": 0.0006,
+ "loss": 4.604315757751465,
+ "step": 3958
+ },
+ {
+ "epoch": 54.992139737991266,
+ "grad_norm": 0.01611141301691532,
+ "learning_rate": 0.0006,
+ "loss": 4.643245220184326,
+ "step": 3959
+ },
+ {
+ "epoch": 55.0,
+ "grad_norm": 0.0156728383153677,
+ "learning_rate": 0.0006,
+ "loss": 4.6739044189453125,
+ "step": 3960
+ },
+ {
+ "epoch": 55.0,
+ "eval_loss": 4.851032257080078,
+ "eval_runtime": 56.8699,
+ "eval_samples_per_second": 42.94,
+ "eval_steps_per_second": 1.354,
+ "step": 3960
+ },
+ {
+ "epoch": 55.01397379912664,
+ "grad_norm": 0.016214672476053238,
+ "learning_rate": 0.0006,
+ "loss": 4.571749687194824,
+ "step": 3961
+ },
+ {
+ "epoch": 55.02794759825328,
+ "grad_norm": 0.01649155467748642,
+ "learning_rate": 0.0006,
+ "loss": 4.553592205047607,
+ "step": 3962
+ },
+ {
+ "epoch": 55.041921397379916,
+ "grad_norm": 0.016793379560112953,
+ "learning_rate": 0.0006,
+ "loss": 4.650628566741943,
+ "step": 3963
+ },
+ {
+ "epoch": 55.05589519650655,
+ "grad_norm": 0.0191937405616045,
+ "learning_rate": 0.0006,
+ "loss": 4.620532989501953,
+ "step": 3964
+ },
+ {
+ "epoch": 55.069868995633186,
+ "grad_norm": 0.020976556465029716,
+ "learning_rate": 0.0006,
+ "loss": 4.6329545974731445,
+ "step": 3965
+ },
+ {
+ "epoch": 55.083842794759825,
+ "grad_norm": 0.023450477048754692,
+ "learning_rate": 0.0006,
+ "loss": 4.60828971862793,
+ "step": 3966
+ },
+ {
+ "epoch": 55.09781659388646,
+ "grad_norm": 0.026778312399983406,
+ "learning_rate": 0.0006,
+ "loss": 4.617481708526611,
+ "step": 3967
+ },
+ {
+ "epoch": 55.1117903930131,
+ "grad_norm": 0.02333568036556244,
+ "learning_rate": 0.0006,
+ "loss": 4.580722332000732,
+ "step": 3968
+ },
+ {
+ "epoch": 55.12576419213974,
+ "grad_norm": 0.018491310998797417,
+ "learning_rate": 0.0006,
+ "loss": 4.590953350067139,
+ "step": 3969
+ },
+ {
+ "epoch": 55.13973799126637,
+ "grad_norm": 0.020339803770184517,
+ "learning_rate": 0.0006,
+ "loss": 4.518374443054199,
+ "step": 3970
+ },
+ {
+ "epoch": 55.15371179039301,
+ "grad_norm": 0.01759706623852253,
+ "learning_rate": 0.0006,
+ "loss": 4.560522079467773,
+ "step": 3971
+ },
+ {
+ "epoch": 55.16768558951965,
+ "grad_norm": 0.015875980257987976,
+ "learning_rate": 0.0006,
+ "loss": 4.4795708656311035,
+ "step": 3972
+ },
+ {
+ "epoch": 55.18165938864629,
+ "grad_norm": 0.01650134287774563,
+ "learning_rate": 0.0006,
+ "loss": 4.57435941696167,
+ "step": 3973
+ },
+ {
+ "epoch": 55.19563318777293,
+ "grad_norm": 0.01734331250190735,
+ "learning_rate": 0.0006,
+ "loss": 4.661007881164551,
+ "step": 3974
+ },
+ {
+ "epoch": 55.209606986899566,
+ "grad_norm": 0.01651303470134735,
+ "learning_rate": 0.0006,
+ "loss": 4.511148452758789,
+ "step": 3975
+ },
+ {
+ "epoch": 55.223580786026204,
+ "grad_norm": 0.016189761459827423,
+ "learning_rate": 0.0006,
+ "loss": 4.592350006103516,
+ "step": 3976
+ },
+ {
+ "epoch": 55.237554585152836,
+ "grad_norm": 0.01833222061395645,
+ "learning_rate": 0.0006,
+ "loss": 4.506649971008301,
+ "step": 3977
+ },
+ {
+ "epoch": 55.251528384279474,
+ "grad_norm": 0.02433410845696926,
+ "learning_rate": 0.0006,
+ "loss": 4.49500036239624,
+ "step": 3978
+ },
+ {
+ "epoch": 55.26550218340611,
+ "grad_norm": 0.0314289927482605,
+ "learning_rate": 0.0006,
+ "loss": 4.507624626159668,
+ "step": 3979
+ },
+ {
+ "epoch": 55.27947598253275,
+ "grad_norm": 0.03013679012656212,
+ "learning_rate": 0.0006,
+ "loss": 4.533641815185547,
+ "step": 3980
+ },
+ {
+ "epoch": 55.29344978165939,
+ "grad_norm": 0.024524061009287834,
+ "learning_rate": 0.0006,
+ "loss": 4.5721116065979,
+ "step": 3981
+ },
+ {
+ "epoch": 55.30742358078603,
+ "grad_norm": 0.02640336938202381,
+ "learning_rate": 0.0006,
+ "loss": 4.652521133422852,
+ "step": 3982
+ },
+ {
+ "epoch": 55.32139737991266,
+ "grad_norm": 0.021309101954102516,
+ "learning_rate": 0.0006,
+ "loss": 4.600375175476074,
+ "step": 3983
+ },
+ {
+ "epoch": 55.3353711790393,
+ "grad_norm": 0.023008549585938454,
+ "learning_rate": 0.0006,
+ "loss": 4.53914737701416,
+ "step": 3984
+ },
+ {
+ "epoch": 55.34934497816594,
+ "grad_norm": 0.022491468116641045,
+ "learning_rate": 0.0006,
+ "loss": 4.661163330078125,
+ "step": 3985
+ },
+ {
+ "epoch": 55.36331877729258,
+ "grad_norm": 0.01904081553220749,
+ "learning_rate": 0.0006,
+ "loss": 4.5426177978515625,
+ "step": 3986
+ },
+ {
+ "epoch": 55.377292576419215,
+ "grad_norm": 0.019229518249630928,
+ "learning_rate": 0.0006,
+ "loss": 4.606066703796387,
+ "step": 3987
+ },
+ {
+ "epoch": 55.391266375545854,
+ "grad_norm": 0.01781383901834488,
+ "learning_rate": 0.0006,
+ "loss": 4.54642391204834,
+ "step": 3988
+ },
+ {
+ "epoch": 55.40524017467249,
+ "grad_norm": 0.0174600500613451,
+ "learning_rate": 0.0006,
+ "loss": 4.711891174316406,
+ "step": 3989
+ },
+ {
+ "epoch": 55.419213973799124,
+ "grad_norm": 0.018154339864850044,
+ "learning_rate": 0.0006,
+ "loss": 4.527122974395752,
+ "step": 3990
+ },
+ {
+ "epoch": 55.43318777292576,
+ "grad_norm": 0.01914648339152336,
+ "learning_rate": 0.0006,
+ "loss": 4.641975402832031,
+ "step": 3991
+ },
+ {
+ "epoch": 55.4471615720524,
+ "grad_norm": 0.017118625342845917,
+ "learning_rate": 0.0006,
+ "loss": 4.511425971984863,
+ "step": 3992
+ },
+ {
+ "epoch": 55.46113537117904,
+ "grad_norm": 0.01551489345729351,
+ "learning_rate": 0.0006,
+ "loss": 4.551141738891602,
+ "step": 3993
+ },
+ {
+ "epoch": 55.47510917030568,
+ "grad_norm": 0.015072065405547619,
+ "learning_rate": 0.0006,
+ "loss": 4.651115894317627,
+ "step": 3994
+ },
+ {
+ "epoch": 55.48908296943232,
+ "grad_norm": 0.017010986804962158,
+ "learning_rate": 0.0006,
+ "loss": 4.583282470703125,
+ "step": 3995
+ },
+ {
+ "epoch": 55.50305676855895,
+ "grad_norm": 0.01764831133186817,
+ "learning_rate": 0.0006,
+ "loss": 4.593364238739014,
+ "step": 3996
+ },
+ {
+ "epoch": 55.51703056768559,
+ "grad_norm": 0.01762884296476841,
+ "learning_rate": 0.0006,
+ "loss": 4.45671272277832,
+ "step": 3997
+ },
+ {
+ "epoch": 55.531004366812226,
+ "grad_norm": 0.01600457727909088,
+ "learning_rate": 0.0006,
+ "loss": 4.511260032653809,
+ "step": 3998
+ },
+ {
+ "epoch": 55.544978165938865,
+ "grad_norm": 0.015706021338701248,
+ "learning_rate": 0.0006,
+ "loss": 4.47593355178833,
+ "step": 3999
+ },
+ {
+ "epoch": 55.5589519650655,
+ "grad_norm": 0.017718996852636337,
+ "learning_rate": 0.0006,
+ "loss": 4.610543251037598,
+ "step": 4000
+ },
+ {
+ "epoch": 55.57292576419214,
+ "grad_norm": 0.015073757618665695,
+ "learning_rate": 0.0006,
+ "loss": 4.466620445251465,
+ "step": 4001
+ },
+ {
+ "epoch": 55.58689956331878,
+ "grad_norm": 0.01789247989654541,
+ "learning_rate": 0.0006,
+ "loss": 4.503409385681152,
+ "step": 4002
+ },
+ {
+ "epoch": 55.60087336244541,
+ "grad_norm": 0.022329598665237427,
+ "learning_rate": 0.0006,
+ "loss": 4.62099552154541,
+ "step": 4003
+ },
+ {
+ "epoch": 55.61484716157205,
+ "grad_norm": 0.02343829534947872,
+ "learning_rate": 0.0006,
+ "loss": 4.620813369750977,
+ "step": 4004
+ },
+ {
+ "epoch": 55.62882096069869,
+ "grad_norm": 0.023279687389731407,
+ "learning_rate": 0.0006,
+ "loss": 4.545765399932861,
+ "step": 4005
+ },
+ {
+ "epoch": 55.64279475982533,
+ "grad_norm": 0.020817464217543602,
+ "learning_rate": 0.0006,
+ "loss": 4.561470031738281,
+ "step": 4006
+ },
+ {
+ "epoch": 55.65676855895197,
+ "grad_norm": 0.020270058885216713,
+ "learning_rate": 0.0006,
+ "loss": 4.5875020027160645,
+ "step": 4007
+ },
+ {
+ "epoch": 55.670742358078606,
+ "grad_norm": 0.021085843443870544,
+ "learning_rate": 0.0006,
+ "loss": 4.504560470581055,
+ "step": 4008
+ },
+ {
+ "epoch": 55.68471615720524,
+ "grad_norm": 0.018507730215787888,
+ "learning_rate": 0.0006,
+ "loss": 4.520644664764404,
+ "step": 4009
+ },
+ {
+ "epoch": 55.698689956331876,
+ "grad_norm": 0.020571516826748848,
+ "learning_rate": 0.0006,
+ "loss": 4.539643287658691,
+ "step": 4010
+ },
+ {
+ "epoch": 55.712663755458514,
+ "grad_norm": 0.021484510973095894,
+ "learning_rate": 0.0006,
+ "loss": 4.628833293914795,
+ "step": 4011
+ },
+ {
+ "epoch": 55.72663755458515,
+ "grad_norm": 0.017693819478154182,
+ "learning_rate": 0.0006,
+ "loss": 4.6023783683776855,
+ "step": 4012
+ },
+ {
+ "epoch": 55.74061135371179,
+ "grad_norm": 0.020186059176921844,
+ "learning_rate": 0.0006,
+ "loss": 4.634884834289551,
+ "step": 4013
+ },
+ {
+ "epoch": 55.75458515283843,
+ "grad_norm": 0.01835780404508114,
+ "learning_rate": 0.0006,
+ "loss": 4.577791213989258,
+ "step": 4014
+ },
+ {
+ "epoch": 55.76855895196506,
+ "grad_norm": 0.018162988126277924,
+ "learning_rate": 0.0006,
+ "loss": 4.641597270965576,
+ "step": 4015
+ },
+ {
+ "epoch": 55.7825327510917,
+ "grad_norm": 0.018909504637122154,
+ "learning_rate": 0.0006,
+ "loss": 4.487447738647461,
+ "step": 4016
+ },
+ {
+ "epoch": 55.79650655021834,
+ "grad_norm": 0.018972251564264297,
+ "learning_rate": 0.0006,
+ "loss": 4.590317726135254,
+ "step": 4017
+ },
+ {
+ "epoch": 55.81048034934498,
+ "grad_norm": 0.01793607883155346,
+ "learning_rate": 0.0006,
+ "loss": 4.4994611740112305,
+ "step": 4018
+ },
+ {
+ "epoch": 55.82445414847162,
+ "grad_norm": 0.02123514749109745,
+ "learning_rate": 0.0006,
+ "loss": 4.479006767272949,
+ "step": 4019
+ },
+ {
+ "epoch": 55.838427947598255,
+ "grad_norm": 0.021246599033474922,
+ "learning_rate": 0.0006,
+ "loss": 4.4738640785217285,
+ "step": 4020
+ },
+ {
+ "epoch": 55.852401746724894,
+ "grad_norm": 0.020131602883338928,
+ "learning_rate": 0.0006,
+ "loss": 4.600308418273926,
+ "step": 4021
+ },
+ {
+ "epoch": 55.866375545851525,
+ "grad_norm": 0.019159460440278053,
+ "learning_rate": 0.0006,
+ "loss": 4.540908336639404,
+ "step": 4022
+ },
+ {
+ "epoch": 55.880349344978164,
+ "grad_norm": 0.023114705458283424,
+ "learning_rate": 0.0006,
+ "loss": 4.641247749328613,
+ "step": 4023
+ },
+ {
+ "epoch": 55.8943231441048,
+ "grad_norm": 0.02296135015785694,
+ "learning_rate": 0.0006,
+ "loss": 4.444624423980713,
+ "step": 4024
+ },
+ {
+ "epoch": 55.90829694323144,
+ "grad_norm": 0.02204655110836029,
+ "learning_rate": 0.0006,
+ "loss": 4.506349563598633,
+ "step": 4025
+ },
+ {
+ "epoch": 55.92227074235808,
+ "grad_norm": 0.019338225945830345,
+ "learning_rate": 0.0006,
+ "loss": 4.616047382354736,
+ "step": 4026
+ },
+ {
+ "epoch": 55.93624454148472,
+ "grad_norm": 0.020357884466648102,
+ "learning_rate": 0.0006,
+ "loss": 4.48170280456543,
+ "step": 4027
+ },
+ {
+ "epoch": 55.95021834061135,
+ "grad_norm": 0.02426772564649582,
+ "learning_rate": 0.0006,
+ "loss": 4.576877117156982,
+ "step": 4028
+ },
+ {
+ "epoch": 55.96419213973799,
+ "grad_norm": 0.023351816460490227,
+ "learning_rate": 0.0006,
+ "loss": 4.458885192871094,
+ "step": 4029
+ },
+ {
+ "epoch": 55.97816593886463,
+ "grad_norm": 0.019892243668437004,
+ "learning_rate": 0.0006,
+ "loss": 4.520229816436768,
+ "step": 4030
+ },
+ {
+ "epoch": 55.992139737991266,
+ "grad_norm": 0.018331089988350868,
+ "learning_rate": 0.0006,
+ "loss": 4.432201862335205,
+ "step": 4031
+ },
+ {
+ "epoch": 56.0,
+ "grad_norm": 0.021300828084349632,
+ "learning_rate": 0.0006,
+ "loss": 4.617072582244873,
+ "step": 4032
+ },
+ {
+ "epoch": 56.0,
+ "eval_loss": 4.811229705810547,
+ "eval_runtime": 56.4846,
+ "eval_samples_per_second": 43.233,
+ "eval_steps_per_second": 1.363,
+ "step": 4032
+ },
+ {
+ "epoch": 56.01397379912664,
+ "grad_norm": 0.02199401892721653,
+ "learning_rate": 0.0006,
+ "loss": 4.513965606689453,
+ "step": 4033
+ },
+ {
+ "epoch": 56.02794759825328,
+ "grad_norm": 0.022109858691692352,
+ "learning_rate": 0.0006,
+ "loss": 4.557568550109863,
+ "step": 4034
+ },
+ {
+ "epoch": 56.041921397379916,
+ "grad_norm": 0.020103694871068,
+ "learning_rate": 0.0006,
+ "loss": 4.569512844085693,
+ "step": 4035
+ },
+ {
+ "epoch": 56.05589519650655,
+ "grad_norm": 0.01898941583931446,
+ "learning_rate": 0.0006,
+ "loss": 4.623138904571533,
+ "step": 4036
+ },
+ {
+ "epoch": 56.069868995633186,
+ "grad_norm": 0.02227010577917099,
+ "learning_rate": 0.0006,
+ "loss": 4.433316230773926,
+ "step": 4037
+ },
+ {
+ "epoch": 56.083842794759825,
+ "grad_norm": 0.024683045223355293,
+ "learning_rate": 0.0006,
+ "loss": 4.458296775817871,
+ "step": 4038
+ },
+ {
+ "epoch": 56.09781659388646,
+ "grad_norm": 0.02900184690952301,
+ "learning_rate": 0.0006,
+ "loss": 4.531060218811035,
+ "step": 4039
+ },
+ {
+ "epoch": 56.1117903930131,
+ "grad_norm": 0.025557566434144974,
+ "learning_rate": 0.0006,
+ "loss": 4.56685733795166,
+ "step": 4040
+ },
+ {
+ "epoch": 56.12576419213974,
+ "grad_norm": 0.021414656192064285,
+ "learning_rate": 0.0006,
+ "loss": 4.469037055969238,
+ "step": 4041
+ },
+ {
+ "epoch": 56.13973799126637,
+ "grad_norm": 0.023262616246938705,
+ "learning_rate": 0.0006,
+ "loss": 4.493876934051514,
+ "step": 4042
+ },
+ {
+ "epoch": 56.15371179039301,
+ "grad_norm": 0.023011326789855957,
+ "learning_rate": 0.0006,
+ "loss": 4.4572906494140625,
+ "step": 4043
+ },
+ {
+ "epoch": 56.16768558951965,
+ "grad_norm": 0.02177615463733673,
+ "learning_rate": 0.0006,
+ "loss": 4.615389823913574,
+ "step": 4044
+ },
+ {
+ "epoch": 56.18165938864629,
+ "grad_norm": 0.021527713164687157,
+ "learning_rate": 0.0006,
+ "loss": 4.613762855529785,
+ "step": 4045
+ },
+ {
+ "epoch": 56.19563318777293,
+ "grad_norm": 0.01982325315475464,
+ "learning_rate": 0.0006,
+ "loss": 4.53670597076416,
+ "step": 4046
+ },
+ {
+ "epoch": 56.209606986899566,
+ "grad_norm": 0.022484654560685158,
+ "learning_rate": 0.0006,
+ "loss": 4.605565071105957,
+ "step": 4047
+ },
+ {
+ "epoch": 56.223580786026204,
+ "grad_norm": 0.02346140146255493,
+ "learning_rate": 0.0006,
+ "loss": 4.6265950202941895,
+ "step": 4048
+ },
+ {
+ "epoch": 56.237554585152836,
+ "grad_norm": 0.01876714453101158,
+ "learning_rate": 0.0006,
+ "loss": 4.481525421142578,
+ "step": 4049
+ },
+ {
+ "epoch": 56.251528384279474,
+ "grad_norm": 0.01926851086318493,
+ "learning_rate": 0.0006,
+ "loss": 4.630258083343506,
+ "step": 4050
+ },
+ {
+ "epoch": 56.26550218340611,
+ "grad_norm": 0.020858589559793472,
+ "learning_rate": 0.0006,
+ "loss": 4.5265913009643555,
+ "step": 4051
+ },
+ {
+ "epoch": 56.27947598253275,
+ "grad_norm": 0.01889280416071415,
+ "learning_rate": 0.0006,
+ "loss": 4.581275463104248,
+ "step": 4052
+ },
+ {
+ "epoch": 56.29344978165939,
+ "grad_norm": 0.01801002025604248,
+ "learning_rate": 0.0006,
+ "loss": 4.498557090759277,
+ "step": 4053
+ },
+ {
+ "epoch": 56.30742358078603,
+ "grad_norm": 0.02252180315554142,
+ "learning_rate": 0.0006,
+ "loss": 4.565267086029053,
+ "step": 4054
+ },
+ {
+ "epoch": 56.32139737991266,
+ "grad_norm": 0.02306070365011692,
+ "learning_rate": 0.0006,
+ "loss": 4.3591389656066895,
+ "step": 4055
+ },
+ {
+ "epoch": 56.3353711790393,
+ "grad_norm": 0.020292513072490692,
+ "learning_rate": 0.0006,
+ "loss": 4.407517433166504,
+ "step": 4056
+ },
+ {
+ "epoch": 56.34934497816594,
+ "grad_norm": 0.019154123961925507,
+ "learning_rate": 0.0006,
+ "loss": 4.482261657714844,
+ "step": 4057
+ },
+ {
+ "epoch": 56.36331877729258,
+ "grad_norm": 0.017338937148451805,
+ "learning_rate": 0.0006,
+ "loss": 4.4463582038879395,
+ "step": 4058
+ },
+ {
+ "epoch": 56.377292576419215,
+ "grad_norm": 0.01730559952557087,
+ "learning_rate": 0.0006,
+ "loss": 4.542474746704102,
+ "step": 4059
+ },
+ {
+ "epoch": 56.391266375545854,
+ "grad_norm": 0.017029426991939545,
+ "learning_rate": 0.0006,
+ "loss": 4.552488327026367,
+ "step": 4060
+ },
+ {
+ "epoch": 56.40524017467249,
+ "grad_norm": 0.01745498552918434,
+ "learning_rate": 0.0006,
+ "loss": 4.52928352355957,
+ "step": 4061
+ },
+ {
+ "epoch": 56.419213973799124,
+ "grad_norm": 0.0168515145778656,
+ "learning_rate": 0.0006,
+ "loss": 4.4516801834106445,
+ "step": 4062
+ },
+ {
+ "epoch": 56.43318777292576,
+ "grad_norm": 0.018545417115092278,
+ "learning_rate": 0.0006,
+ "loss": 4.569811820983887,
+ "step": 4063
+ },
+ {
+ "epoch": 56.4471615720524,
+ "grad_norm": 0.02227838523685932,
+ "learning_rate": 0.0006,
+ "loss": 4.412224769592285,
+ "step": 4064
+ },
+ {
+ "epoch": 56.46113537117904,
+ "grad_norm": 0.023338504135608673,
+ "learning_rate": 0.0006,
+ "loss": 4.490184307098389,
+ "step": 4065
+ },
+ {
+ "epoch": 56.47510917030568,
+ "grad_norm": 0.02257535606622696,
+ "learning_rate": 0.0006,
+ "loss": 4.479152679443359,
+ "step": 4066
+ },
+ {
+ "epoch": 56.48908296943232,
+ "grad_norm": 0.022737598046660423,
+ "learning_rate": 0.0006,
+ "loss": 4.418235778808594,
+ "step": 4067
+ },
+ {
+ "epoch": 56.50305676855895,
+ "grad_norm": 0.024603573605418205,
+ "learning_rate": 0.0006,
+ "loss": 4.45781135559082,
+ "step": 4068
+ },
+ {
+ "epoch": 56.51703056768559,
+ "grad_norm": 0.024953778833150864,
+ "learning_rate": 0.0006,
+ "loss": 4.593907833099365,
+ "step": 4069
+ },
+ {
+ "epoch": 56.531004366812226,
+ "grad_norm": 0.02195645309984684,
+ "learning_rate": 0.0006,
+ "loss": 4.467643737792969,
+ "step": 4070
+ },
+ {
+ "epoch": 56.544978165938865,
+ "grad_norm": 0.017126578837633133,
+ "learning_rate": 0.0006,
+ "loss": 4.555484771728516,
+ "step": 4071
+ },
+ {
+ "epoch": 56.5589519650655,
+ "grad_norm": 0.019482605159282684,
+ "learning_rate": 0.0006,
+ "loss": 4.430915832519531,
+ "step": 4072
+ },
+ {
+ "epoch": 56.57292576419214,
+ "grad_norm": 0.0219392292201519,
+ "learning_rate": 0.0006,
+ "loss": 4.593365669250488,
+ "step": 4073
+ },
+ {
+ "epoch": 56.58689956331878,
+ "grad_norm": 0.02420145645737648,
+ "learning_rate": 0.0006,
+ "loss": 4.491467475891113,
+ "step": 4074
+ },
+ {
+ "epoch": 56.60087336244541,
+ "grad_norm": 0.021523723378777504,
+ "learning_rate": 0.0006,
+ "loss": 4.463634490966797,
+ "step": 4075
+ },
+ {
+ "epoch": 56.61484716157205,
+ "grad_norm": 0.020242048427462578,
+ "learning_rate": 0.0006,
+ "loss": 4.577968597412109,
+ "step": 4076
+ },
+ {
+ "epoch": 56.62882096069869,
+ "grad_norm": 0.017771173268556595,
+ "learning_rate": 0.0006,
+ "loss": 4.568269729614258,
+ "step": 4077
+ },
+ {
+ "epoch": 56.64279475982533,
+ "grad_norm": 0.01598544418811798,
+ "learning_rate": 0.0006,
+ "loss": 4.476841926574707,
+ "step": 4078
+ },
+ {
+ "epoch": 56.65676855895197,
+ "grad_norm": 0.0168446097522974,
+ "learning_rate": 0.0006,
+ "loss": 4.480965614318848,
+ "step": 4079
+ },
+ {
+ "epoch": 56.670742358078606,
+ "grad_norm": 0.014948660507798195,
+ "learning_rate": 0.0006,
+ "loss": 4.535400390625,
+ "step": 4080
+ },
+ {
+ "epoch": 56.68471615720524,
+ "grad_norm": 0.014899153262376785,
+ "learning_rate": 0.0006,
+ "loss": 4.476048946380615,
+ "step": 4081
+ },
+ {
+ "epoch": 56.698689956331876,
+ "grad_norm": 0.014051870442926884,
+ "learning_rate": 0.0006,
+ "loss": 4.533884048461914,
+ "step": 4082
+ },
+ {
+ "epoch": 56.712663755458514,
+ "grad_norm": 0.014708572067320347,
+ "learning_rate": 0.0006,
+ "loss": 4.586690902709961,
+ "step": 4083
+ },
+ {
+ "epoch": 56.72663755458515,
+ "grad_norm": 0.014797299169003963,
+ "learning_rate": 0.0006,
+ "loss": 4.446291446685791,
+ "step": 4084
+ },
+ {
+ "epoch": 56.74061135371179,
+ "grad_norm": 0.015256541781127453,
+ "learning_rate": 0.0006,
+ "loss": 4.553572654724121,
+ "step": 4085
+ },
+ {
+ "epoch": 56.75458515283843,
+ "grad_norm": 0.016042491421103477,
+ "learning_rate": 0.0006,
+ "loss": 4.507801055908203,
+ "step": 4086
+ },
+ {
+ "epoch": 56.76855895196506,
+ "grad_norm": 0.018835173919796944,
+ "learning_rate": 0.0006,
+ "loss": 4.393911838531494,
+ "step": 4087
+ },
+ {
+ "epoch": 56.7825327510917,
+ "grad_norm": 0.021994033828377724,
+ "learning_rate": 0.0006,
+ "loss": 4.543368339538574,
+ "step": 4088
+ },
+ {
+ "epoch": 56.79650655021834,
+ "grad_norm": 0.0230990182608366,
+ "learning_rate": 0.0006,
+ "loss": 4.507420539855957,
+ "step": 4089
+ },
+ {
+ "epoch": 56.81048034934498,
+ "grad_norm": 0.018818842247128487,
+ "learning_rate": 0.0006,
+ "loss": 4.499941349029541,
+ "step": 4090
+ },
+ {
+ "epoch": 56.82445414847162,
+ "grad_norm": 0.018257278949022293,
+ "learning_rate": 0.0006,
+ "loss": 4.380057334899902,
+ "step": 4091
+ },
+ {
+ "epoch": 56.838427947598255,
+ "grad_norm": 0.020073018968105316,
+ "learning_rate": 0.0006,
+ "loss": 4.572811603546143,
+ "step": 4092
+ },
+ {
+ "epoch": 56.852401746724894,
+ "grad_norm": 0.02078658528625965,
+ "learning_rate": 0.0006,
+ "loss": 4.446835517883301,
+ "step": 4093
+ },
+ {
+ "epoch": 56.866375545851525,
+ "grad_norm": 0.01793118566274643,
+ "learning_rate": 0.0006,
+ "loss": 4.517854690551758,
+ "step": 4094
+ },
+ {
+ "epoch": 56.880349344978164,
+ "grad_norm": 0.01608281210064888,
+ "learning_rate": 0.0006,
+ "loss": 4.506482124328613,
+ "step": 4095
+ },
+ {
+ "epoch": 56.8943231441048,
+ "grad_norm": 0.017405368387699127,
+ "learning_rate": 0.0006,
+ "loss": 4.439718246459961,
+ "step": 4096
+ },
+ {
+ "epoch": 56.90829694323144,
+ "grad_norm": 0.017930863425135612,
+ "learning_rate": 0.0006,
+ "loss": 4.409365653991699,
+ "step": 4097
+ },
+ {
+ "epoch": 56.92227074235808,
+ "grad_norm": 0.015757670626044273,
+ "learning_rate": 0.0006,
+ "loss": 4.505547523498535,
+ "step": 4098
+ },
+ {
+ "epoch": 56.93624454148472,
+ "grad_norm": 0.01580345816910267,
+ "learning_rate": 0.0006,
+ "loss": 4.462774753570557,
+ "step": 4099
+ },
+ {
+ "epoch": 56.95021834061135,
+ "grad_norm": 0.01659180410206318,
+ "learning_rate": 0.0006,
+ "loss": 4.49345588684082,
+ "step": 4100
+ },
+ {
+ "epoch": 56.96419213973799,
+ "grad_norm": 0.01877259835600853,
+ "learning_rate": 0.0006,
+ "loss": 4.574443340301514,
+ "step": 4101
+ },
+ {
+ "epoch": 56.97816593886463,
+ "grad_norm": 0.01977839693427086,
+ "learning_rate": 0.0006,
+ "loss": 4.574034690856934,
+ "step": 4102
+ },
+ {
+ "epoch": 56.992139737991266,
+ "grad_norm": 0.020627597346901894,
+ "learning_rate": 0.0006,
+ "loss": 4.50969123840332,
+ "step": 4103
+ },
+ {
+ "epoch": 57.0,
+ "grad_norm": 0.025128807872533798,
+ "learning_rate": 0.0006,
+ "loss": 4.401662826538086,
+ "step": 4104
+ },
+ {
+ "epoch": 57.0,
+ "eval_loss": 4.779252052307129,
+ "eval_runtime": 56.8104,
+ "eval_samples_per_second": 42.985,
+ "eval_steps_per_second": 1.355,
+ "step": 4104
+ },
+ {
+ "epoch": 57.01397379912664,
+ "grad_norm": 0.027712570503354073,
+ "learning_rate": 0.0006,
+ "loss": 4.472955703735352,
+ "step": 4105
+ },
+ {
+ "epoch": 57.02794759825328,
+ "grad_norm": 0.02164197526872158,
+ "learning_rate": 0.0006,
+ "loss": 4.387596130371094,
+ "step": 4106
+ },
+ {
+ "epoch": 57.041921397379916,
+ "grad_norm": 0.017972752451896667,
+ "learning_rate": 0.0006,
+ "loss": 4.4326276779174805,
+ "step": 4107
+ },
+ {
+ "epoch": 57.05589519650655,
+ "grad_norm": 0.020811019465327263,
+ "learning_rate": 0.0006,
+ "loss": 4.452951431274414,
+ "step": 4108
+ },
+ {
+ "epoch": 57.069868995633186,
+ "grad_norm": 0.019913259893655777,
+ "learning_rate": 0.0006,
+ "loss": 4.448301315307617,
+ "step": 4109
+ },
+ {
+ "epoch": 57.083842794759825,
+ "grad_norm": 0.017409102991223335,
+ "learning_rate": 0.0006,
+ "loss": 4.4108357429504395,
+ "step": 4110
+ },
+ {
+ "epoch": 57.09781659388646,
+ "grad_norm": 0.019374269992113113,
+ "learning_rate": 0.0006,
+ "loss": 4.398412227630615,
+ "step": 4111
+ },
+ {
+ "epoch": 57.1117903930131,
+ "grad_norm": 0.015521776862442493,
+ "learning_rate": 0.0006,
+ "loss": 4.508594989776611,
+ "step": 4112
+ },
+ {
+ "epoch": 57.12576419213974,
+ "grad_norm": 0.017201920971274376,
+ "learning_rate": 0.0006,
+ "loss": 4.449558734893799,
+ "step": 4113
+ },
+ {
+ "epoch": 57.13973799126637,
+ "grad_norm": 0.018062541261315346,
+ "learning_rate": 0.0006,
+ "loss": 4.517874240875244,
+ "step": 4114
+ },
+ {
+ "epoch": 57.15371179039301,
+ "grad_norm": 0.015846073627471924,
+ "learning_rate": 0.0006,
+ "loss": 4.427501678466797,
+ "step": 4115
+ },
+ {
+ "epoch": 57.16768558951965,
+ "grad_norm": 0.01849362626671791,
+ "learning_rate": 0.0006,
+ "loss": 4.433181285858154,
+ "step": 4116
+ },
+ {
+ "epoch": 57.18165938864629,
+ "grad_norm": 0.01948804222047329,
+ "learning_rate": 0.0006,
+ "loss": 4.446122646331787,
+ "step": 4117
+ },
+ {
+ "epoch": 57.19563318777293,
+ "grad_norm": 0.017354389652609825,
+ "learning_rate": 0.0006,
+ "loss": 4.431153774261475,
+ "step": 4118
+ },
+ {
+ "epoch": 57.209606986899566,
+ "grad_norm": 0.015817373991012573,
+ "learning_rate": 0.0006,
+ "loss": 4.448566913604736,
+ "step": 4119
+ },
+ {
+ "epoch": 57.223580786026204,
+ "grad_norm": 0.01635785959661007,
+ "learning_rate": 0.0006,
+ "loss": 4.408031463623047,
+ "step": 4120
+ },
+ {
+ "epoch": 57.237554585152836,
+ "grad_norm": 0.014528338797390461,
+ "learning_rate": 0.0006,
+ "loss": 4.541668891906738,
+ "step": 4121
+ },
+ {
+ "epoch": 57.251528384279474,
+ "grad_norm": 0.01589174196124077,
+ "learning_rate": 0.0006,
+ "loss": 4.513199329376221,
+ "step": 4122
+ },
+ {
+ "epoch": 57.26550218340611,
+ "grad_norm": 0.016093695536255836,
+ "learning_rate": 0.0006,
+ "loss": 4.516178131103516,
+ "step": 4123
+ },
+ {
+ "epoch": 57.27947598253275,
+ "grad_norm": 0.01788361556828022,
+ "learning_rate": 0.0006,
+ "loss": 4.524880409240723,
+ "step": 4124
+ },
+ {
+ "epoch": 57.29344978165939,
+ "grad_norm": 0.01591576263308525,
+ "learning_rate": 0.0006,
+ "loss": 4.44764518737793,
+ "step": 4125
+ },
+ {
+ "epoch": 57.30742358078603,
+ "grad_norm": 0.015452570281922817,
+ "learning_rate": 0.0006,
+ "loss": 4.46844482421875,
+ "step": 4126
+ },
+ {
+ "epoch": 57.32139737991266,
+ "grad_norm": 0.016404492780566216,
+ "learning_rate": 0.0006,
+ "loss": 4.443678855895996,
+ "step": 4127
+ },
+ {
+ "epoch": 57.3353711790393,
+ "grad_norm": 0.019967148080468178,
+ "learning_rate": 0.0006,
+ "loss": 4.430230140686035,
+ "step": 4128
+ },
+ {
+ "epoch": 57.34934497816594,
+ "grad_norm": 0.02594015561044216,
+ "learning_rate": 0.0006,
+ "loss": 4.47061824798584,
+ "step": 4129
+ },
+ {
+ "epoch": 57.36331877729258,
+ "grad_norm": 0.031096890568733215,
+ "learning_rate": 0.0006,
+ "loss": 4.462275981903076,
+ "step": 4130
+ },
+ {
+ "epoch": 57.377292576419215,
+ "grad_norm": 0.02834898792207241,
+ "learning_rate": 0.0006,
+ "loss": 4.41143798828125,
+ "step": 4131
+ },
+ {
+ "epoch": 57.391266375545854,
+ "grad_norm": 0.023621153086423874,
+ "learning_rate": 0.0006,
+ "loss": 4.562548637390137,
+ "step": 4132
+ },
+ {
+ "epoch": 57.40524017467249,
+ "grad_norm": 0.02277296595275402,
+ "learning_rate": 0.0006,
+ "loss": 4.456247329711914,
+ "step": 4133
+ },
+ {
+ "epoch": 57.419213973799124,
+ "grad_norm": 0.019868768751621246,
+ "learning_rate": 0.0006,
+ "loss": 4.524300575256348,
+ "step": 4134
+ },
+ {
+ "epoch": 57.43318777292576,
+ "grad_norm": 0.01887078955769539,
+ "learning_rate": 0.0006,
+ "loss": 4.540097236633301,
+ "step": 4135
+ },
+ {
+ "epoch": 57.4471615720524,
+ "grad_norm": 0.024268802255392075,
+ "learning_rate": 0.0006,
+ "loss": 4.397233963012695,
+ "step": 4136
+ },
+ {
+ "epoch": 57.46113537117904,
+ "grad_norm": 0.025739721953868866,
+ "learning_rate": 0.0006,
+ "loss": 4.354194641113281,
+ "step": 4137
+ },
+ {
+ "epoch": 57.47510917030568,
+ "grad_norm": 0.02442290261387825,
+ "learning_rate": 0.0006,
+ "loss": 4.303822994232178,
+ "step": 4138
+ },
+ {
+ "epoch": 57.48908296943232,
+ "grad_norm": 0.022215209901332855,
+ "learning_rate": 0.0006,
+ "loss": 4.4870452880859375,
+ "step": 4139
+ },
+ {
+ "epoch": 57.50305676855895,
+ "grad_norm": 0.01789158768951893,
+ "learning_rate": 0.0006,
+ "loss": 4.343809127807617,
+ "step": 4140
+ },
+ {
+ "epoch": 57.51703056768559,
+ "grad_norm": 0.020381739363074303,
+ "learning_rate": 0.0006,
+ "loss": 4.53312873840332,
+ "step": 4141
+ },
+ {
+ "epoch": 57.531004366812226,
+ "grad_norm": 0.019585080444812775,
+ "learning_rate": 0.0006,
+ "loss": 4.510162353515625,
+ "step": 4142
+ },
+ {
+ "epoch": 57.544978165938865,
+ "grad_norm": 0.01985127478837967,
+ "learning_rate": 0.0006,
+ "loss": 4.3654022216796875,
+ "step": 4143
+ },
+ {
+ "epoch": 57.5589519650655,
+ "grad_norm": 0.01962905190885067,
+ "learning_rate": 0.0006,
+ "loss": 4.560830116271973,
+ "step": 4144
+ },
+ {
+ "epoch": 57.57292576419214,
+ "grad_norm": 0.02679629437625408,
+ "learning_rate": 0.0006,
+ "loss": 4.300804138183594,
+ "step": 4145
+ },
+ {
+ "epoch": 57.58689956331878,
+ "grad_norm": 0.039830856025218964,
+ "learning_rate": 0.0006,
+ "loss": 4.54150915145874,
+ "step": 4146
+ },
+ {
+ "epoch": 57.60087336244541,
+ "grad_norm": 0.049527477473020554,
+ "learning_rate": 0.0006,
+ "loss": 4.405792236328125,
+ "step": 4147
+ },
+ {
+ "epoch": 57.61484716157205,
+ "grad_norm": 0.042966946959495544,
+ "learning_rate": 0.0006,
+ "loss": 4.500643730163574,
+ "step": 4148
+ },
+ {
+ "epoch": 57.62882096069869,
+ "grad_norm": 0.08999177813529968,
+ "learning_rate": 0.0006,
+ "loss": 4.504790782928467,
+ "step": 4149
+ },
+ {
+ "epoch": 57.64279475982533,
+ "grad_norm": 0.44693124294281006,
+ "learning_rate": 0.0006,
+ "loss": 4.845344066619873,
+ "step": 4150
+ },
+ {
+ "epoch": 57.65676855895197,
+ "grad_norm": 0.7143144011497498,
+ "learning_rate": 0.0006,
+ "loss": 6.907845497131348,
+ "step": 4151
+ },
+ {
+ "epoch": 57.670742358078606,
+ "grad_norm": 0.25239595770835876,
+ "learning_rate": 0.0006,
+ "loss": 7.082579612731934,
+ "step": 4152
+ },
+ {
+ "epoch": 57.68471615720524,
+ "grad_norm": 0.26550790667533875,
+ "learning_rate": 0.0006,
+ "loss": 7.50180196762085,
+ "step": 4153
+ },
+ {
+ "epoch": 57.698689956331876,
+ "grad_norm": 0.14735378324985504,
+ "learning_rate": 0.0006,
+ "loss": 7.214590072631836,
+ "step": 4154
+ },
+ {
+ "epoch": 57.712663755458514,
+ "grad_norm": 0.17685334384441376,
+ "learning_rate": 0.0006,
+ "loss": 6.861320495605469,
+ "step": 4155
+ },
+ {
+ "epoch": 57.72663755458515,
+ "grad_norm": 0.18231339752674103,
+ "learning_rate": 0.0006,
+ "loss": 6.791839599609375,
+ "step": 4156
+ },
+ {
+ "epoch": 57.74061135371179,
+ "grad_norm": 0.09230359643697739,
+ "learning_rate": 0.0006,
+ "loss": 6.760223388671875,
+ "step": 4157
+ },
+ {
+ "epoch": 57.75458515283843,
+ "grad_norm": 0.06918249279260635,
+ "learning_rate": 0.0006,
+ "loss": 6.532999515533447,
+ "step": 4158
+ },
+ {
+ "epoch": 57.76855895196506,
+ "grad_norm": 0.0901007279753685,
+ "learning_rate": 0.0006,
+ "loss": 6.52907133102417,
+ "step": 4159
+ },
+ {
+ "epoch": 57.7825327510917,
+ "grad_norm": 0.05678229779005051,
+ "learning_rate": 0.0006,
+ "loss": 6.295290470123291,
+ "step": 4160
+ },
+ {
+ "epoch": 57.79650655021834,
+ "grad_norm": 0.04450210556387901,
+ "learning_rate": 0.0006,
+ "loss": 6.14900541305542,
+ "step": 4161
+ },
+ {
+ "epoch": 57.81048034934498,
+ "grad_norm": 0.049191512167453766,
+ "learning_rate": 0.0006,
+ "loss": 6.19650411605835,
+ "step": 4162
+ },
+ {
+ "epoch": 57.82445414847162,
+ "grad_norm": 0.03908967226743698,
+ "learning_rate": 0.0006,
+ "loss": 6.054948806762695,
+ "step": 4163
+ },
+ {
+ "epoch": 57.838427947598255,
+ "grad_norm": 0.04023387283086777,
+ "learning_rate": 0.0006,
+ "loss": 6.034873962402344,
+ "step": 4164
+ },
+ {
+ "epoch": 57.852401746724894,
+ "grad_norm": 0.03304268419742584,
+ "learning_rate": 0.0006,
+ "loss": 5.769852161407471,
+ "step": 4165
+ },
+ {
+ "epoch": 57.866375545851525,
+ "grad_norm": 0.03432450816035271,
+ "learning_rate": 0.0006,
+ "loss": 5.788785457611084,
+ "step": 4166
+ },
+ {
+ "epoch": 57.880349344978164,
+ "grad_norm": 0.029965840280056,
+ "learning_rate": 0.0006,
+ "loss": 5.815243721008301,
+ "step": 4167
+ },
+ {
+ "epoch": 57.8943231441048,
+ "grad_norm": 0.03637698292732239,
+ "learning_rate": 0.0006,
+ "loss": 5.7204060554504395,
+ "step": 4168
+ },
+ {
+ "epoch": 57.90829694323144,
+ "grad_norm": 0.03477517142891884,
+ "learning_rate": 0.0006,
+ "loss": 5.605930328369141,
+ "step": 4169
+ },
+ {
+ "epoch": 57.92227074235808,
+ "grad_norm": 0.05061859264969826,
+ "learning_rate": 0.0006,
+ "loss": 5.547054767608643,
+ "step": 4170
+ },
+ {
+ "epoch": 57.93624454148472,
+ "grad_norm": 0.06299655884504318,
+ "learning_rate": 0.0006,
+ "loss": 5.552166938781738,
+ "step": 4171
+ },
+ {
+ "epoch": 57.95021834061135,
+ "grad_norm": 0.04215948283672333,
+ "learning_rate": 0.0006,
+ "loss": 5.375457763671875,
+ "step": 4172
+ },
+ {
+ "epoch": 57.96419213973799,
+ "grad_norm": 0.0348566472530365,
+ "learning_rate": 0.0006,
+ "loss": 5.418606281280518,
+ "step": 4173
+ },
+ {
+ "epoch": 57.97816593886463,
+ "grad_norm": 0.03807530924677849,
+ "learning_rate": 0.0006,
+ "loss": 5.403156280517578,
+ "step": 4174
+ },
+ {
+ "epoch": 57.992139737991266,
+ "grad_norm": 0.02989993803203106,
+ "learning_rate": 0.0006,
+ "loss": 5.355951309204102,
+ "step": 4175
+ },
+ {
+ "epoch": 58.0,
+ "grad_norm": 0.034300558269023895,
+ "learning_rate": 0.0006,
+ "loss": 5.44920539855957,
+ "step": 4176
+ },
+ {
+ "epoch": 58.0,
+ "eval_loss": 5.416731834411621,
+ "eval_runtime": 56.7934,
+ "eval_samples_per_second": 42.998,
+ "eval_steps_per_second": 1.356,
+ "step": 4176
+ },
+ {
+ "epoch": 58.01397379912664,
+ "grad_norm": 0.033961448818445206,
+ "learning_rate": 0.0006,
+ "loss": 5.250707626342773,
+ "step": 4177
+ },
+ {
+ "epoch": 58.02794759825328,
+ "grad_norm": 0.027343502268195152,
+ "learning_rate": 0.0006,
+ "loss": 5.149238109588623,
+ "step": 4178
+ },
+ {
+ "epoch": 58.041921397379916,
+ "grad_norm": 0.03033650480210781,
+ "learning_rate": 0.0006,
+ "loss": 5.072463512420654,
+ "step": 4179
+ },
+ {
+ "epoch": 58.05589519650655,
+ "grad_norm": 0.028262868523597717,
+ "learning_rate": 0.0006,
+ "loss": 5.109417915344238,
+ "step": 4180
+ },
+ {
+ "epoch": 58.069868995633186,
+ "grad_norm": 0.02373034693300724,
+ "learning_rate": 0.0006,
+ "loss": 5.082464218139648,
+ "step": 4181
+ },
+ {
+ "epoch": 58.083842794759825,
+ "grad_norm": 0.025879688560962677,
+ "learning_rate": 0.0006,
+ "loss": 5.056792259216309,
+ "step": 4182
+ },
+ {
+ "epoch": 58.09781659388646,
+ "grad_norm": 0.026252275332808495,
+ "learning_rate": 0.0006,
+ "loss": 5.064239978790283,
+ "step": 4183
+ },
+ {
+ "epoch": 58.1117903930131,
+ "grad_norm": 0.028177792206406593,
+ "learning_rate": 0.0006,
+ "loss": 4.922310829162598,
+ "step": 4184
+ },
+ {
+ "epoch": 58.12576419213974,
+ "grad_norm": 0.02630820870399475,
+ "learning_rate": 0.0006,
+ "loss": 4.992884635925293,
+ "step": 4185
+ },
+ {
+ "epoch": 58.13973799126637,
+ "grad_norm": 0.024890903383493423,
+ "learning_rate": 0.0006,
+ "loss": 4.910036563873291,
+ "step": 4186
+ },
+ {
+ "epoch": 58.15371179039301,
+ "grad_norm": 0.02822992391884327,
+ "learning_rate": 0.0006,
+ "loss": 4.866640090942383,
+ "step": 4187
+ },
+ {
+ "epoch": 58.16768558951965,
+ "grad_norm": 0.03253607079386711,
+ "learning_rate": 0.0006,
+ "loss": 4.9112958908081055,
+ "step": 4188
+ },
+ {
+ "epoch": 58.18165938864629,
+ "grad_norm": 0.03765944764018059,
+ "learning_rate": 0.0006,
+ "loss": 4.774468421936035,
+ "step": 4189
+ },
+ {
+ "epoch": 58.19563318777293,
+ "grad_norm": 0.039874881505966187,
+ "learning_rate": 0.0006,
+ "loss": 4.831678867340088,
+ "step": 4190
+ },
+ {
+ "epoch": 58.209606986899566,
+ "grad_norm": 0.034595925360918045,
+ "learning_rate": 0.0006,
+ "loss": 4.8243794441223145,
+ "step": 4191
+ },
+ {
+ "epoch": 58.223580786026204,
+ "grad_norm": 0.04840189591050148,
+ "learning_rate": 0.0006,
+ "loss": 4.854156494140625,
+ "step": 4192
+ },
+ {
+ "epoch": 58.237554585152836,
+ "grad_norm": 0.06358753889799118,
+ "learning_rate": 0.0006,
+ "loss": 4.836430549621582,
+ "step": 4193
+ },
+ {
+ "epoch": 58.251528384279474,
+ "grad_norm": 0.0507555715739727,
+ "learning_rate": 0.0006,
+ "loss": 4.88357400894165,
+ "step": 4194
+ },
+ {
+ "epoch": 58.26550218340611,
+ "grad_norm": 0.04384690150618553,
+ "learning_rate": 0.0006,
+ "loss": 4.774049282073975,
+ "step": 4195
+ },
+ {
+ "epoch": 58.27947598253275,
+ "grad_norm": 0.03246928006410599,
+ "learning_rate": 0.0006,
+ "loss": 4.686605930328369,
+ "step": 4196
+ },
+ {
+ "epoch": 58.29344978165939,
+ "grad_norm": 0.028878789395093918,
+ "learning_rate": 0.0006,
+ "loss": 4.796070575714111,
+ "step": 4197
+ },
+ {
+ "epoch": 58.30742358078603,
+ "grad_norm": 0.028175361454486847,
+ "learning_rate": 0.0006,
+ "loss": 4.728490829467773,
+ "step": 4198
+ },
+ {
+ "epoch": 58.32139737991266,
+ "grad_norm": 0.029428014531731606,
+ "learning_rate": 0.0006,
+ "loss": 4.7159624099731445,
+ "step": 4199
+ },
+ {
+ "epoch": 58.3353711790393,
+ "grad_norm": 0.031227827072143555,
+ "learning_rate": 0.0006,
+ "loss": 4.702625274658203,
+ "step": 4200
+ },
+ {
+ "epoch": 58.34934497816594,
+ "grad_norm": 0.028843428939580917,
+ "learning_rate": 0.0006,
+ "loss": 4.811799049377441,
+ "step": 4201
+ },
+ {
+ "epoch": 58.36331877729258,
+ "grad_norm": 0.026780391111969948,
+ "learning_rate": 0.0006,
+ "loss": 4.81662654876709,
+ "step": 4202
+ },
+ {
+ "epoch": 58.377292576419215,
+ "grad_norm": 0.019699757918715477,
+ "learning_rate": 0.0006,
+ "loss": 4.713685989379883,
+ "step": 4203
+ },
+ {
+ "epoch": 58.391266375545854,
+ "grad_norm": 0.024389250203967094,
+ "learning_rate": 0.0006,
+ "loss": 4.72354793548584,
+ "step": 4204
+ },
+ {
+ "epoch": 58.40524017467249,
+ "grad_norm": 0.02218274027109146,
+ "learning_rate": 0.0006,
+ "loss": 4.674125671386719,
+ "step": 4205
+ },
+ {
+ "epoch": 58.419213973799124,
+ "grad_norm": 0.02041488140821457,
+ "learning_rate": 0.0006,
+ "loss": 4.7909698486328125,
+ "step": 4206
+ },
+ {
+ "epoch": 58.43318777292576,
+ "grad_norm": 0.018561935052275658,
+ "learning_rate": 0.0006,
+ "loss": 4.808770179748535,
+ "step": 4207
+ },
+ {
+ "epoch": 58.4471615720524,
+ "grad_norm": 0.01708846725523472,
+ "learning_rate": 0.0006,
+ "loss": 4.701626777648926,
+ "step": 4208
+ },
+ {
+ "epoch": 58.46113537117904,
+ "grad_norm": 0.0222734697163105,
+ "learning_rate": 0.0006,
+ "loss": 4.505410194396973,
+ "step": 4209
+ },
+ {
+ "epoch": 58.47510917030568,
+ "grad_norm": 0.03197057917714119,
+ "learning_rate": 0.0006,
+ "loss": 4.610360145568848,
+ "step": 4210
+ },
+ {
+ "epoch": 58.48908296943232,
+ "grad_norm": 0.06854520738124847,
+ "learning_rate": 0.0006,
+ "loss": 4.625649452209473,
+ "step": 4211
+ },
+ {
+ "epoch": 58.50305676855895,
+ "grad_norm": 0.09981521219015121,
+ "learning_rate": 0.0006,
+ "loss": 4.832150459289551,
+ "step": 4212
+ },
+ {
+ "epoch": 58.51703056768559,
+ "grad_norm": 0.046356040984392166,
+ "learning_rate": 0.0006,
+ "loss": 4.5641632080078125,
+ "step": 4213
+ },
+ {
+ "epoch": 58.531004366812226,
+ "grad_norm": 0.03882667422294617,
+ "learning_rate": 0.0006,
+ "loss": 4.666974067687988,
+ "step": 4214
+ },
+ {
+ "epoch": 58.544978165938865,
+ "grad_norm": 0.033203233033418655,
+ "learning_rate": 0.0006,
+ "loss": 4.629413604736328,
+ "step": 4215
+ },
+ {
+ "epoch": 58.5589519650655,
+ "grad_norm": 0.02806735597550869,
+ "learning_rate": 0.0006,
+ "loss": 4.763245582580566,
+ "step": 4216
+ },
+ {
+ "epoch": 58.57292576419214,
+ "grad_norm": 0.026728278025984764,
+ "learning_rate": 0.0006,
+ "loss": 4.629597187042236,
+ "step": 4217
+ },
+ {
+ "epoch": 58.58689956331878,
+ "grad_norm": 0.024380959570407867,
+ "learning_rate": 0.0006,
+ "loss": 4.633614540100098,
+ "step": 4218
+ },
+ {
+ "epoch": 58.60087336244541,
+ "grad_norm": 0.022312544286251068,
+ "learning_rate": 0.0006,
+ "loss": 4.671379566192627,
+ "step": 4219
+ },
+ {
+ "epoch": 58.61484716157205,
+ "grad_norm": 0.022049568593502045,
+ "learning_rate": 0.0006,
+ "loss": 4.661557197570801,
+ "step": 4220
+ },
+ {
+ "epoch": 58.62882096069869,
+ "grad_norm": 0.02119818702340126,
+ "learning_rate": 0.0006,
+ "loss": 4.688562393188477,
+ "step": 4221
+ },
+ {
+ "epoch": 58.64279475982533,
+ "grad_norm": 0.019769301638007164,
+ "learning_rate": 0.0006,
+ "loss": 4.607748031616211,
+ "step": 4222
+ },
+ {
+ "epoch": 58.65676855895197,
+ "grad_norm": 0.02124079130589962,
+ "learning_rate": 0.0006,
+ "loss": 4.570640563964844,
+ "step": 4223
+ },
+ {
+ "epoch": 58.670742358078606,
+ "grad_norm": 0.016344040632247925,
+ "learning_rate": 0.0006,
+ "loss": 4.676113128662109,
+ "step": 4224
+ },
+ {
+ "epoch": 58.68471615720524,
+ "grad_norm": 0.01735256239771843,
+ "learning_rate": 0.0006,
+ "loss": 4.62088680267334,
+ "step": 4225
+ },
+ {
+ "epoch": 58.698689956331876,
+ "grad_norm": 0.019323458895087242,
+ "learning_rate": 0.0006,
+ "loss": 4.579405307769775,
+ "step": 4226
+ },
+ {
+ "epoch": 58.712663755458514,
+ "grad_norm": 0.01760544814169407,
+ "learning_rate": 0.0006,
+ "loss": 4.625407695770264,
+ "step": 4227
+ },
+ {
+ "epoch": 58.72663755458515,
+ "grad_norm": 0.015593250282108784,
+ "learning_rate": 0.0006,
+ "loss": 4.6326117515563965,
+ "step": 4228
+ },
+ {
+ "epoch": 58.74061135371179,
+ "grad_norm": 0.014357523061335087,
+ "learning_rate": 0.0006,
+ "loss": 4.6869401931762695,
+ "step": 4229
+ },
+ {
+ "epoch": 58.75458515283843,
+ "grad_norm": 0.01604357920587063,
+ "learning_rate": 0.0006,
+ "loss": 4.690784931182861,
+ "step": 4230
+ },
+ {
+ "epoch": 58.76855895196506,
+ "grad_norm": 0.01673845760524273,
+ "learning_rate": 0.0006,
+ "loss": 4.513410568237305,
+ "step": 4231
+ },
+ {
+ "epoch": 58.7825327510917,
+ "grad_norm": 0.013657779432833195,
+ "learning_rate": 0.0006,
+ "loss": 4.637057781219482,
+ "step": 4232
+ },
+ {
+ "epoch": 58.79650655021834,
+ "grad_norm": 0.013373114168643951,
+ "learning_rate": 0.0006,
+ "loss": 4.719967842102051,
+ "step": 4233
+ },
+ {
+ "epoch": 58.81048034934498,
+ "grad_norm": 0.016345568001270294,
+ "learning_rate": 0.0006,
+ "loss": 4.604109764099121,
+ "step": 4234
+ },
+ {
+ "epoch": 58.82445414847162,
+ "grad_norm": 0.01573033444583416,
+ "learning_rate": 0.0006,
+ "loss": 4.6471428871154785,
+ "step": 4235
+ },
+ {
+ "epoch": 58.838427947598255,
+ "grad_norm": 0.01517449039965868,
+ "learning_rate": 0.0006,
+ "loss": 4.722684860229492,
+ "step": 4236
+ },
+ {
+ "epoch": 58.852401746724894,
+ "grad_norm": 0.0145102022215724,
+ "learning_rate": 0.0006,
+ "loss": 4.662420272827148,
+ "step": 4237
+ },
+ {
+ "epoch": 58.866375545851525,
+ "grad_norm": 0.01457217987626791,
+ "learning_rate": 0.0006,
+ "loss": 4.585169792175293,
+ "step": 4238
+ },
+ {
+ "epoch": 58.880349344978164,
+ "grad_norm": 0.01406773366034031,
+ "learning_rate": 0.0006,
+ "loss": 4.5413665771484375,
+ "step": 4239
+ },
+ {
+ "epoch": 58.8943231441048,
+ "grad_norm": 0.012653871439397335,
+ "learning_rate": 0.0006,
+ "loss": 4.553636074066162,
+ "step": 4240
+ },
+ {
+ "epoch": 58.90829694323144,
+ "grad_norm": 0.01261200476437807,
+ "learning_rate": 0.0006,
+ "loss": 4.534193992614746,
+ "step": 4241
+ },
+ {
+ "epoch": 58.92227074235808,
+ "grad_norm": 0.012955864891409874,
+ "learning_rate": 0.0006,
+ "loss": 4.697248458862305,
+ "step": 4242
+ },
+ {
+ "epoch": 58.93624454148472,
+ "grad_norm": 0.012134749442338943,
+ "learning_rate": 0.0006,
+ "loss": 4.557095527648926,
+ "step": 4243
+ },
+ {
+ "epoch": 58.95021834061135,
+ "grad_norm": 0.012164677493274212,
+ "learning_rate": 0.0006,
+ "loss": 4.623579025268555,
+ "step": 4244
+ },
+ {
+ "epoch": 58.96419213973799,
+ "grad_norm": 0.013802947476506233,
+ "learning_rate": 0.0006,
+ "loss": 4.562838077545166,
+ "step": 4245
+ },
+ {
+ "epoch": 58.97816593886463,
+ "grad_norm": 0.011420476250350475,
+ "learning_rate": 0.0006,
+ "loss": 4.62265682220459,
+ "step": 4246
+ },
+ {
+ "epoch": 58.992139737991266,
+ "grad_norm": 0.01169576682150364,
+ "learning_rate": 0.0006,
+ "loss": 4.727449417114258,
+ "step": 4247
+ },
+ {
+ "epoch": 59.0,
+ "grad_norm": 0.013577710837125778,
+ "learning_rate": 0.0006,
+ "loss": 4.536325454711914,
+ "step": 4248
+ },
+ {
+ "epoch": 59.0,
+ "eval_loss": 4.86527681350708,
+ "eval_runtime": 60.4463,
+ "eval_samples_per_second": 40.399,
+ "eval_steps_per_second": 1.274,
+ "step": 4248
+ },
+ {
+ "epoch": 59.01397379912664,
+ "grad_norm": 0.015251655131578445,
+ "learning_rate": 0.0006,
+ "loss": 4.634598731994629,
+ "step": 4249
+ },
+ {
+ "epoch": 59.02794759825328,
+ "grad_norm": 0.017414938658475876,
+ "learning_rate": 0.0006,
+ "loss": 4.521329879760742,
+ "step": 4250
+ },
+ {
+ "epoch": 59.041921397379916,
+ "grad_norm": 0.01656588353216648,
+ "learning_rate": 0.0006,
+ "loss": 4.446192264556885,
+ "step": 4251
+ },
+ {
+ "epoch": 59.05589519650655,
+ "grad_norm": 0.012038925662636757,
+ "learning_rate": 0.0006,
+ "loss": 4.523855209350586,
+ "step": 4252
+ },
+ {
+ "epoch": 59.069868995633186,
+ "grad_norm": 0.014403033070266247,
+ "learning_rate": 0.0006,
+ "loss": 4.561079978942871,
+ "step": 4253
+ },
+ {
+ "epoch": 59.083842794759825,
+ "grad_norm": 0.016243597492575645,
+ "learning_rate": 0.0006,
+ "loss": 4.673602104187012,
+ "step": 4254
+ },
+ {
+ "epoch": 59.09781659388646,
+ "grad_norm": 0.016492877155542374,
+ "learning_rate": 0.0006,
+ "loss": 4.564824104309082,
+ "step": 4255
+ },
+ {
+ "epoch": 59.1117903930131,
+ "grad_norm": 0.015980370342731476,
+ "learning_rate": 0.0006,
+ "loss": 4.339249134063721,
+ "step": 4256
+ },
+ {
+ "epoch": 59.12576419213974,
+ "grad_norm": 0.019687356427311897,
+ "learning_rate": 0.0006,
+ "loss": 4.481834888458252,
+ "step": 4257
+ },
+ {
+ "epoch": 59.13973799126637,
+ "grad_norm": 0.02983798086643219,
+ "learning_rate": 0.0006,
+ "loss": 4.4495697021484375,
+ "step": 4258
+ },
+ {
+ "epoch": 59.15371179039301,
+ "grad_norm": 0.04600981995463371,
+ "learning_rate": 0.0006,
+ "loss": 4.567838668823242,
+ "step": 4259
+ },
+ {
+ "epoch": 59.16768558951965,
+ "grad_norm": 0.04852001741528511,
+ "learning_rate": 0.0006,
+ "loss": 4.594701766967773,
+ "step": 4260
+ },
+ {
+ "epoch": 59.18165938864629,
+ "grad_norm": 0.03429504856467247,
+ "learning_rate": 0.0006,
+ "loss": 4.573735237121582,
+ "step": 4261
+ },
+ {
+ "epoch": 59.19563318777293,
+ "grad_norm": 0.029345160350203514,
+ "learning_rate": 0.0006,
+ "loss": 4.488258361816406,
+ "step": 4262
+ },
+ {
+ "epoch": 59.209606986899566,
+ "grad_norm": 0.01961623504757881,
+ "learning_rate": 0.0006,
+ "loss": 4.545222282409668,
+ "step": 4263
+ },
+ {
+ "epoch": 59.223580786026204,
+ "grad_norm": 0.020406130701303482,
+ "learning_rate": 0.0006,
+ "loss": 4.471898078918457,
+ "step": 4264
+ },
+ {
+ "epoch": 59.237554585152836,
+ "grad_norm": 0.020500272512435913,
+ "learning_rate": 0.0006,
+ "loss": 4.538337707519531,
+ "step": 4265
+ },
+ {
+ "epoch": 59.251528384279474,
+ "grad_norm": 0.020727017894387245,
+ "learning_rate": 0.0006,
+ "loss": 4.437284469604492,
+ "step": 4266
+ },
+ {
+ "epoch": 59.26550218340611,
+ "grad_norm": 0.03395998477935791,
+ "learning_rate": 0.0006,
+ "loss": 4.459124565124512,
+ "step": 4267
+ },
+ {
+ "epoch": 59.27947598253275,
+ "grad_norm": 0.04635605961084366,
+ "learning_rate": 0.0006,
+ "loss": 4.520228385925293,
+ "step": 4268
+ },
+ {
+ "epoch": 59.29344978165939,
+ "grad_norm": 0.04253006726503372,
+ "learning_rate": 0.0006,
+ "loss": 4.578673362731934,
+ "step": 4269
+ },
+ {
+ "epoch": 59.30742358078603,
+ "grad_norm": 0.02403509058058262,
+ "learning_rate": 0.0006,
+ "loss": 4.484639644622803,
+ "step": 4270
+ },
+ {
+ "epoch": 59.32139737991266,
+ "grad_norm": 0.02254190482199192,
+ "learning_rate": 0.0006,
+ "loss": 4.511314868927002,
+ "step": 4271
+ },
+ {
+ "epoch": 59.3353711790393,
+ "grad_norm": 0.021585190668702126,
+ "learning_rate": 0.0006,
+ "loss": 4.530239105224609,
+ "step": 4272
+ },
+ {
+ "epoch": 59.34934497816594,
+ "grad_norm": 0.021210316568613052,
+ "learning_rate": 0.0006,
+ "loss": 4.593514442443848,
+ "step": 4273
+ },
+ {
+ "epoch": 59.36331877729258,
+ "grad_norm": 0.019317107275128365,
+ "learning_rate": 0.0006,
+ "loss": 4.5591535568237305,
+ "step": 4274
+ },
+ {
+ "epoch": 59.377292576419215,
+ "grad_norm": 0.018059708178043365,
+ "learning_rate": 0.0006,
+ "loss": 4.4499592781066895,
+ "step": 4275
+ },
+ {
+ "epoch": 59.391266375545854,
+ "grad_norm": 0.017847446724772453,
+ "learning_rate": 0.0006,
+ "loss": 4.473999500274658,
+ "step": 4276
+ },
+ {
+ "epoch": 59.40524017467249,
+ "grad_norm": 0.014944756403565407,
+ "learning_rate": 0.0006,
+ "loss": 4.5934529304504395,
+ "step": 4277
+ },
+ {
+ "epoch": 59.419213973799124,
+ "grad_norm": 0.01663832925260067,
+ "learning_rate": 0.0006,
+ "loss": 4.504262924194336,
+ "step": 4278
+ },
+ {
+ "epoch": 59.43318777292576,
+ "grad_norm": 0.015571796335279942,
+ "learning_rate": 0.0006,
+ "loss": 4.512310981750488,
+ "step": 4279
+ },
+ {
+ "epoch": 59.4471615720524,
+ "grad_norm": 0.015896400436758995,
+ "learning_rate": 0.0006,
+ "loss": 4.452714920043945,
+ "step": 4280
+ },
+ {
+ "epoch": 59.46113537117904,
+ "grad_norm": 0.016010284423828125,
+ "learning_rate": 0.0006,
+ "loss": 4.6188063621521,
+ "step": 4281
+ },
+ {
+ "epoch": 59.47510917030568,
+ "grad_norm": 0.015075593255460262,
+ "learning_rate": 0.0006,
+ "loss": 4.477978229522705,
+ "step": 4282
+ },
+ {
+ "epoch": 59.48908296943232,
+ "grad_norm": 0.014698950573801994,
+ "learning_rate": 0.0006,
+ "loss": 4.544955730438232,
+ "step": 4283
+ },
+ {
+ "epoch": 59.50305676855895,
+ "grad_norm": 0.0136245833709836,
+ "learning_rate": 0.0006,
+ "loss": 4.530460357666016,
+ "step": 4284
+ },
+ {
+ "epoch": 59.51703056768559,
+ "grad_norm": 0.013118310831487179,
+ "learning_rate": 0.0006,
+ "loss": 4.614217758178711,
+ "step": 4285
+ },
+ {
+ "epoch": 59.531004366812226,
+ "grad_norm": 0.01295961532741785,
+ "learning_rate": 0.0006,
+ "loss": 4.43156099319458,
+ "step": 4286
+ },
+ {
+ "epoch": 59.544978165938865,
+ "grad_norm": 0.012429811991751194,
+ "learning_rate": 0.0006,
+ "loss": 4.517333030700684,
+ "step": 4287
+ },
+ {
+ "epoch": 59.5589519650655,
+ "grad_norm": 0.01261830423027277,
+ "learning_rate": 0.0006,
+ "loss": 4.569780349731445,
+ "step": 4288
+ },
+ {
+ "epoch": 59.57292576419214,
+ "grad_norm": 0.012394499965012074,
+ "learning_rate": 0.0006,
+ "loss": 4.5970869064331055,
+ "step": 4289
+ },
+ {
+ "epoch": 59.58689956331878,
+ "grad_norm": 0.01290759164839983,
+ "learning_rate": 0.0006,
+ "loss": 4.437100410461426,
+ "step": 4290
+ },
+ {
+ "epoch": 59.60087336244541,
+ "grad_norm": 0.012042169459164143,
+ "learning_rate": 0.0006,
+ "loss": 4.567404747009277,
+ "step": 4291
+ },
+ {
+ "epoch": 59.61484716157205,
+ "grad_norm": 0.012338520959019661,
+ "learning_rate": 0.0006,
+ "loss": 4.476193428039551,
+ "step": 4292
+ },
+ {
+ "epoch": 59.62882096069869,
+ "grad_norm": 0.01321258582174778,
+ "learning_rate": 0.0006,
+ "loss": 4.566680431365967,
+ "step": 4293
+ },
+ {
+ "epoch": 59.64279475982533,
+ "grad_norm": 0.013478122651576996,
+ "learning_rate": 0.0006,
+ "loss": 4.487398147583008,
+ "step": 4294
+ },
+ {
+ "epoch": 59.65676855895197,
+ "grad_norm": 0.01537603884935379,
+ "learning_rate": 0.0006,
+ "loss": 4.449337005615234,
+ "step": 4295
+ },
+ {
+ "epoch": 59.670742358078606,
+ "grad_norm": 0.017140565440058708,
+ "learning_rate": 0.0006,
+ "loss": 4.5174360275268555,
+ "step": 4296
+ },
+ {
+ "epoch": 59.68471615720524,
+ "grad_norm": 0.017931461334228516,
+ "learning_rate": 0.0006,
+ "loss": 4.522104740142822,
+ "step": 4297
+ },
+ {
+ "epoch": 59.698689956331876,
+ "grad_norm": 0.018037918955087662,
+ "learning_rate": 0.0006,
+ "loss": 4.496928691864014,
+ "step": 4298
+ },
+ {
+ "epoch": 59.712663755458514,
+ "grad_norm": 0.01684504561126232,
+ "learning_rate": 0.0006,
+ "loss": 4.5332417488098145,
+ "step": 4299
+ },
+ {
+ "epoch": 59.72663755458515,
+ "grad_norm": 0.016268383711576462,
+ "learning_rate": 0.0006,
+ "loss": 4.468893051147461,
+ "step": 4300
+ },
+ {
+ "epoch": 59.74061135371179,
+ "grad_norm": 0.014859385788440704,
+ "learning_rate": 0.0006,
+ "loss": 4.551092147827148,
+ "step": 4301
+ },
+ {
+ "epoch": 59.75458515283843,
+ "grad_norm": 0.013589047826826572,
+ "learning_rate": 0.0006,
+ "loss": 4.470579147338867,
+ "step": 4302
+ },
+ {
+ "epoch": 59.76855895196506,
+ "grad_norm": 0.015704551711678505,
+ "learning_rate": 0.0006,
+ "loss": 4.330068111419678,
+ "step": 4303
+ },
+ {
+ "epoch": 59.7825327510917,
+ "grad_norm": 0.017869921401143074,
+ "learning_rate": 0.0006,
+ "loss": 4.428255081176758,
+ "step": 4304
+ },
+ {
+ "epoch": 59.79650655021834,
+ "grad_norm": 0.019828537479043007,
+ "learning_rate": 0.0006,
+ "loss": 4.496023178100586,
+ "step": 4305
+ },
+ {
+ "epoch": 59.81048034934498,
+ "grad_norm": 0.015856236219406128,
+ "learning_rate": 0.0006,
+ "loss": 4.427152156829834,
+ "step": 4306
+ },
+ {
+ "epoch": 59.82445414847162,
+ "grad_norm": 0.013349304907023907,
+ "learning_rate": 0.0006,
+ "loss": 4.417333602905273,
+ "step": 4307
+ },
+ {
+ "epoch": 59.838427947598255,
+ "grad_norm": 0.014111637137830257,
+ "learning_rate": 0.0006,
+ "loss": 4.4507575035095215,
+ "step": 4308
+ },
+ {
+ "epoch": 59.852401746724894,
+ "grad_norm": 0.015186597593128681,
+ "learning_rate": 0.0006,
+ "loss": 4.375116348266602,
+ "step": 4309
+ },
+ {
+ "epoch": 59.866375545851525,
+ "grad_norm": 0.01604730449616909,
+ "learning_rate": 0.0006,
+ "loss": 4.604518890380859,
+ "step": 4310
+ },
+ {
+ "epoch": 59.880349344978164,
+ "grad_norm": 0.015308464877307415,
+ "learning_rate": 0.0006,
+ "loss": 4.507232189178467,
+ "step": 4311
+ },
+ {
+ "epoch": 59.8943231441048,
+ "grad_norm": 0.014017206616699696,
+ "learning_rate": 0.0006,
+ "loss": 4.398487091064453,
+ "step": 4312
+ },
+ {
+ "epoch": 59.90829694323144,
+ "grad_norm": 0.012911394238471985,
+ "learning_rate": 0.0006,
+ "loss": 4.567872047424316,
+ "step": 4313
+ },
+ {
+ "epoch": 59.92227074235808,
+ "grad_norm": 0.012775142677128315,
+ "learning_rate": 0.0006,
+ "loss": 4.562105655670166,
+ "step": 4314
+ },
+ {
+ "epoch": 59.93624454148472,
+ "grad_norm": 0.014750408008694649,
+ "learning_rate": 0.0006,
+ "loss": 4.456263542175293,
+ "step": 4315
+ },
+ {
+ "epoch": 59.95021834061135,
+ "grad_norm": 0.016726011410355568,
+ "learning_rate": 0.0006,
+ "loss": 4.360722064971924,
+ "step": 4316
+ },
+ {
+ "epoch": 59.96419213973799,
+ "grad_norm": 0.01734776981174946,
+ "learning_rate": 0.0006,
+ "loss": 4.572544574737549,
+ "step": 4317
+ },
+ {
+ "epoch": 59.97816593886463,
+ "grad_norm": 0.01822415366768837,
+ "learning_rate": 0.0006,
+ "loss": 4.490501880645752,
+ "step": 4318
+ },
+ {
+ "epoch": 59.992139737991266,
+ "grad_norm": 0.014638577587902546,
+ "learning_rate": 0.0006,
+ "loss": 4.440951824188232,
+ "step": 4319
+ },
+ {
+ "epoch": 60.0,
+ "grad_norm": 0.01255202479660511,
+ "learning_rate": 0.0006,
+ "loss": 4.559123516082764,
+ "step": 4320
+ },
+ {
+ "epoch": 60.0,
+ "eval_loss": 4.741089344024658,
+ "eval_runtime": 56.8997,
+ "eval_samples_per_second": 42.918,
+ "eval_steps_per_second": 1.353,
+ "step": 4320
+ },
+ {
+ "epoch": 60.01397379912664,
+ "grad_norm": 0.016548393294215202,
+ "learning_rate": 0.0006,
+ "loss": 4.442024230957031,
+ "step": 4321
+ },
+ {
+ "epoch": 60.02794759825328,
+ "grad_norm": 0.019759127870202065,
+ "learning_rate": 0.0006,
+ "loss": 4.394334316253662,
+ "step": 4322
+ },
+ {
+ "epoch": 60.041921397379916,
+ "grad_norm": 0.018325170502066612,
+ "learning_rate": 0.0006,
+ "loss": 4.318270683288574,
+ "step": 4323
+ },
+ {
+ "epoch": 60.05589519650655,
+ "grad_norm": 0.015289463102817535,
+ "learning_rate": 0.0006,
+ "loss": 4.318228721618652,
+ "step": 4324
+ },
+ {
+ "epoch": 60.069868995633186,
+ "grad_norm": 0.015424872748553753,
+ "learning_rate": 0.0006,
+ "loss": 4.3605499267578125,
+ "step": 4325
+ },
+ {
+ "epoch": 60.083842794759825,
+ "grad_norm": 0.014085669070482254,
+ "learning_rate": 0.0006,
+ "loss": 4.359585762023926,
+ "step": 4326
+ },
+ {
+ "epoch": 60.09781659388646,
+ "grad_norm": 0.014004210010170937,
+ "learning_rate": 0.0006,
+ "loss": 4.447783470153809,
+ "step": 4327
+ },
+ {
+ "epoch": 60.1117903930131,
+ "grad_norm": 0.013514582999050617,
+ "learning_rate": 0.0006,
+ "loss": 4.4850239753723145,
+ "step": 4328
+ },
+ {
+ "epoch": 60.12576419213974,
+ "grad_norm": 0.013049394823610783,
+ "learning_rate": 0.0006,
+ "loss": 4.528675556182861,
+ "step": 4329
+ },
+ {
+ "epoch": 60.13973799126637,
+ "grad_norm": 0.01378058921545744,
+ "learning_rate": 0.0006,
+ "loss": 4.444823741912842,
+ "step": 4330
+ },
+ {
+ "epoch": 60.15371179039301,
+ "grad_norm": 0.013752233237028122,
+ "learning_rate": 0.0006,
+ "loss": 4.5205206871032715,
+ "step": 4331
+ },
+ {
+ "epoch": 60.16768558951965,
+ "grad_norm": 0.013431803323328495,
+ "learning_rate": 0.0006,
+ "loss": 4.547320365905762,
+ "step": 4332
+ },
+ {
+ "epoch": 60.18165938864629,
+ "grad_norm": 0.014544196426868439,
+ "learning_rate": 0.0006,
+ "loss": 4.517143249511719,
+ "step": 4333
+ },
+ {
+ "epoch": 60.19563318777293,
+ "grad_norm": 0.01612176187336445,
+ "learning_rate": 0.0006,
+ "loss": 4.517232894897461,
+ "step": 4334
+ },
+ {
+ "epoch": 60.209606986899566,
+ "grad_norm": 0.01803239807486534,
+ "learning_rate": 0.0006,
+ "loss": 4.377574920654297,
+ "step": 4335
+ },
+ {
+ "epoch": 60.223580786026204,
+ "grad_norm": 0.014871489256620407,
+ "learning_rate": 0.0006,
+ "loss": 4.538285255432129,
+ "step": 4336
+ },
+ {
+ "epoch": 60.237554585152836,
+ "grad_norm": 0.013927377760410309,
+ "learning_rate": 0.0006,
+ "loss": 4.386175632476807,
+ "step": 4337
+ },
+ {
+ "epoch": 60.251528384279474,
+ "grad_norm": 0.014757219702005386,
+ "learning_rate": 0.0006,
+ "loss": 4.528553009033203,
+ "step": 4338
+ },
+ {
+ "epoch": 60.26550218340611,
+ "grad_norm": 0.01433873176574707,
+ "learning_rate": 0.0006,
+ "loss": 4.476929187774658,
+ "step": 4339
+ },
+ {
+ "epoch": 60.27947598253275,
+ "grad_norm": 0.015768803656101227,
+ "learning_rate": 0.0006,
+ "loss": 4.50970458984375,
+ "step": 4340
+ },
+ {
+ "epoch": 60.29344978165939,
+ "grad_norm": 0.015651429072022438,
+ "learning_rate": 0.0006,
+ "loss": 4.427133083343506,
+ "step": 4341
+ },
+ {
+ "epoch": 60.30742358078603,
+ "grad_norm": 0.01673000678420067,
+ "learning_rate": 0.0006,
+ "loss": 4.447723865509033,
+ "step": 4342
+ },
+ {
+ "epoch": 60.32139737991266,
+ "grad_norm": 0.018481194972991943,
+ "learning_rate": 0.0006,
+ "loss": 4.559832572937012,
+ "step": 4343
+ },
+ {
+ "epoch": 60.3353711790393,
+ "grad_norm": 0.016731027513742447,
+ "learning_rate": 0.0006,
+ "loss": 4.470893859863281,
+ "step": 4344
+ },
+ {
+ "epoch": 60.34934497816594,
+ "grad_norm": 0.01504011545330286,
+ "learning_rate": 0.0006,
+ "loss": 4.483014106750488,
+ "step": 4345
+ },
+ {
+ "epoch": 60.36331877729258,
+ "grad_norm": 0.01602456159889698,
+ "learning_rate": 0.0006,
+ "loss": 4.548428058624268,
+ "step": 4346
+ },
+ {
+ "epoch": 60.377292576419215,
+ "grad_norm": 0.016027364879846573,
+ "learning_rate": 0.0006,
+ "loss": 4.426779747009277,
+ "step": 4347
+ },
+ {
+ "epoch": 60.391266375545854,
+ "grad_norm": 0.01873827911913395,
+ "learning_rate": 0.0006,
+ "loss": 4.353686332702637,
+ "step": 4348
+ },
+ {
+ "epoch": 60.40524017467249,
+ "grad_norm": 0.01907406374812126,
+ "learning_rate": 0.0006,
+ "loss": 4.502945423126221,
+ "step": 4349
+ },
+ {
+ "epoch": 60.419213973799124,
+ "grad_norm": 0.016083354130387306,
+ "learning_rate": 0.0006,
+ "loss": 4.431731700897217,
+ "step": 4350
+ },
+ {
+ "epoch": 60.43318777292576,
+ "grad_norm": 0.015293709933757782,
+ "learning_rate": 0.0006,
+ "loss": 4.480112552642822,
+ "step": 4351
+ },
+ {
+ "epoch": 60.4471615720524,
+ "grad_norm": 0.01688101328909397,
+ "learning_rate": 0.0006,
+ "loss": 4.50853157043457,
+ "step": 4352
+ },
+ {
+ "epoch": 60.46113537117904,
+ "grad_norm": 0.02094118855893612,
+ "learning_rate": 0.0006,
+ "loss": 4.359930038452148,
+ "step": 4353
+ },
+ {
+ "epoch": 60.47510917030568,
+ "grad_norm": 0.018188107758760452,
+ "learning_rate": 0.0006,
+ "loss": 4.549592018127441,
+ "step": 4354
+ },
+ {
+ "epoch": 60.48908296943232,
+ "grad_norm": 0.01462319865822792,
+ "learning_rate": 0.0006,
+ "loss": 4.49467658996582,
+ "step": 4355
+ },
+ {
+ "epoch": 60.50305676855895,
+ "grad_norm": 0.017508579418063164,
+ "learning_rate": 0.0006,
+ "loss": 4.398318767547607,
+ "step": 4356
+ },
+ {
+ "epoch": 60.51703056768559,
+ "grad_norm": 0.018200315535068512,
+ "learning_rate": 0.0006,
+ "loss": 4.44169282913208,
+ "step": 4357
+ },
+ {
+ "epoch": 60.531004366812226,
+ "grad_norm": 0.01717468351125717,
+ "learning_rate": 0.0006,
+ "loss": 4.481400489807129,
+ "step": 4358
+ },
+ {
+ "epoch": 60.544978165938865,
+ "grad_norm": 0.019329151138663292,
+ "learning_rate": 0.0006,
+ "loss": 4.578839302062988,
+ "step": 4359
+ },
+ {
+ "epoch": 60.5589519650655,
+ "grad_norm": 0.016109555959701538,
+ "learning_rate": 0.0006,
+ "loss": 4.461043357849121,
+ "step": 4360
+ },
+ {
+ "epoch": 60.57292576419214,
+ "grad_norm": 0.01494457945227623,
+ "learning_rate": 0.0006,
+ "loss": 4.505096912384033,
+ "step": 4361
+ },
+ {
+ "epoch": 60.58689956331878,
+ "grad_norm": 0.01729399710893631,
+ "learning_rate": 0.0006,
+ "loss": 4.366747856140137,
+ "step": 4362
+ },
+ {
+ "epoch": 60.60087336244541,
+ "grad_norm": 0.01705821603536606,
+ "learning_rate": 0.0006,
+ "loss": 4.42534065246582,
+ "step": 4363
+ },
+ {
+ "epoch": 60.61484716157205,
+ "grad_norm": 0.018391354009509087,
+ "learning_rate": 0.0006,
+ "loss": 4.350294589996338,
+ "step": 4364
+ },
+ {
+ "epoch": 60.62882096069869,
+ "grad_norm": 0.016989829018712044,
+ "learning_rate": 0.0006,
+ "loss": 4.45905876159668,
+ "step": 4365
+ },
+ {
+ "epoch": 60.64279475982533,
+ "grad_norm": 0.017192212864756584,
+ "learning_rate": 0.0006,
+ "loss": 4.526487350463867,
+ "step": 4366
+ },
+ {
+ "epoch": 60.65676855895197,
+ "grad_norm": 0.018091315403580666,
+ "learning_rate": 0.0006,
+ "loss": 4.309333324432373,
+ "step": 4367
+ },
+ {
+ "epoch": 60.670742358078606,
+ "grad_norm": 0.014051459729671478,
+ "learning_rate": 0.0006,
+ "loss": 4.338171005249023,
+ "step": 4368
+ },
+ {
+ "epoch": 60.68471615720524,
+ "grad_norm": 0.013166418299078941,
+ "learning_rate": 0.0006,
+ "loss": 4.409758567810059,
+ "step": 4369
+ },
+ {
+ "epoch": 60.698689956331876,
+ "grad_norm": 0.014260428957641125,
+ "learning_rate": 0.0006,
+ "loss": 4.524338245391846,
+ "step": 4370
+ },
+ {
+ "epoch": 60.712663755458514,
+ "grad_norm": 0.017019255086779594,
+ "learning_rate": 0.0006,
+ "loss": 4.387436866760254,
+ "step": 4371
+ },
+ {
+ "epoch": 60.72663755458515,
+ "grad_norm": 0.017070675268769264,
+ "learning_rate": 0.0006,
+ "loss": 4.437321662902832,
+ "step": 4372
+ },
+ {
+ "epoch": 60.74061135371179,
+ "grad_norm": 0.015197164379060268,
+ "learning_rate": 0.0006,
+ "loss": 4.371205806732178,
+ "step": 4373
+ },
+ {
+ "epoch": 60.75458515283843,
+ "grad_norm": 0.014791185967624187,
+ "learning_rate": 0.0006,
+ "loss": 4.44439697265625,
+ "step": 4374
+ },
+ {
+ "epoch": 60.76855895196506,
+ "grad_norm": 0.016458261758089066,
+ "learning_rate": 0.0006,
+ "loss": 4.445584297180176,
+ "step": 4375
+ },
+ {
+ "epoch": 60.7825327510917,
+ "grad_norm": 0.017706342041492462,
+ "learning_rate": 0.0006,
+ "loss": 4.348708152770996,
+ "step": 4376
+ },
+ {
+ "epoch": 60.79650655021834,
+ "grad_norm": 0.017561476677656174,
+ "learning_rate": 0.0006,
+ "loss": 4.521929740905762,
+ "step": 4377
+ },
+ {
+ "epoch": 60.81048034934498,
+ "grad_norm": 0.016557445749640465,
+ "learning_rate": 0.0006,
+ "loss": 4.545806884765625,
+ "step": 4378
+ },
+ {
+ "epoch": 60.82445414847162,
+ "grad_norm": 0.016953222453594208,
+ "learning_rate": 0.0006,
+ "loss": 4.4230875968933105,
+ "step": 4379
+ },
+ {
+ "epoch": 60.838427947598255,
+ "grad_norm": 0.017094967886805534,
+ "learning_rate": 0.0006,
+ "loss": 4.423985481262207,
+ "step": 4380
+ },
+ {
+ "epoch": 60.852401746724894,
+ "grad_norm": 0.014787515625357628,
+ "learning_rate": 0.0006,
+ "loss": 4.39473819732666,
+ "step": 4381
+ },
+ {
+ "epoch": 60.866375545851525,
+ "grad_norm": 0.016208263114094734,
+ "learning_rate": 0.0006,
+ "loss": 4.3810038566589355,
+ "step": 4382
+ },
+ {
+ "epoch": 60.880349344978164,
+ "grad_norm": 0.017844321206212044,
+ "learning_rate": 0.0006,
+ "loss": 4.35086727142334,
+ "step": 4383
+ },
+ {
+ "epoch": 60.8943231441048,
+ "grad_norm": 0.017539294436573982,
+ "learning_rate": 0.0006,
+ "loss": 4.460260391235352,
+ "step": 4384
+ },
+ {
+ "epoch": 60.90829694323144,
+ "grad_norm": 0.01612042263150215,
+ "learning_rate": 0.0006,
+ "loss": 4.370500564575195,
+ "step": 4385
+ },
+ {
+ "epoch": 60.92227074235808,
+ "grad_norm": 0.015381601639091969,
+ "learning_rate": 0.0006,
+ "loss": 4.435868263244629,
+ "step": 4386
+ },
+ {
+ "epoch": 60.93624454148472,
+ "grad_norm": 0.01603585295379162,
+ "learning_rate": 0.0006,
+ "loss": 4.434070587158203,
+ "step": 4387
+ },
+ {
+ "epoch": 60.95021834061135,
+ "grad_norm": 0.016268130391836166,
+ "learning_rate": 0.0006,
+ "loss": 4.404343128204346,
+ "step": 4388
+ },
+ {
+ "epoch": 60.96419213973799,
+ "grad_norm": 0.014980032108724117,
+ "learning_rate": 0.0006,
+ "loss": 4.399213790893555,
+ "step": 4389
+ },
+ {
+ "epoch": 60.97816593886463,
+ "grad_norm": 0.016899267211556435,
+ "learning_rate": 0.0006,
+ "loss": 4.508069038391113,
+ "step": 4390
+ },
+ {
+ "epoch": 60.992139737991266,
+ "grad_norm": 0.013756908476352692,
+ "learning_rate": 0.0006,
+ "loss": 4.541858673095703,
+ "step": 4391
+ },
+ {
+ "epoch": 61.0,
+ "grad_norm": 0.01578413136303425,
+ "learning_rate": 0.0006,
+ "loss": 4.543513298034668,
+ "step": 4392
+ },
+ {
+ "epoch": 61.0,
+ "eval_loss": 4.756770610809326,
+ "eval_runtime": 56.4572,
+ "eval_samples_per_second": 43.254,
+ "eval_steps_per_second": 1.364,
+ "step": 4392
+ },
+ {
+ "epoch": 61.01397379912664,
+ "grad_norm": 0.013387506827712059,
+ "learning_rate": 0.0006,
+ "loss": 4.475942611694336,
+ "step": 4393
+ },
+ {
+ "epoch": 61.02794759825328,
+ "grad_norm": 0.014278904534876347,
+ "learning_rate": 0.0006,
+ "loss": 4.442403793334961,
+ "step": 4394
+ },
+ {
+ "epoch": 61.041921397379916,
+ "grad_norm": 0.014359553344547749,
+ "learning_rate": 0.0006,
+ "loss": 4.509463310241699,
+ "step": 4395
+ },
+ {
+ "epoch": 61.05589519650655,
+ "grad_norm": 0.01718050241470337,
+ "learning_rate": 0.0006,
+ "loss": 4.354458808898926,
+ "step": 4396
+ },
+ {
+ "epoch": 61.069868995633186,
+ "grad_norm": 0.015988312661647797,
+ "learning_rate": 0.0006,
+ "loss": 4.484508514404297,
+ "step": 4397
+ },
+ {
+ "epoch": 61.083842794759825,
+ "grad_norm": 0.015466735698282719,
+ "learning_rate": 0.0006,
+ "loss": 4.326908111572266,
+ "step": 4398
+ },
+ {
+ "epoch": 61.09781659388646,
+ "grad_norm": 0.014541332609951496,
+ "learning_rate": 0.0006,
+ "loss": 4.405823230743408,
+ "step": 4399
+ },
+ {
+ "epoch": 61.1117903930131,
+ "grad_norm": 0.014803987927734852,
+ "learning_rate": 0.0006,
+ "loss": 4.375386714935303,
+ "step": 4400
+ },
+ {
+ "epoch": 61.12576419213974,
+ "grad_norm": 0.013962093740701675,
+ "learning_rate": 0.0006,
+ "loss": 4.300838470458984,
+ "step": 4401
+ },
+ {
+ "epoch": 61.13973799126637,
+ "grad_norm": 0.015213954262435436,
+ "learning_rate": 0.0006,
+ "loss": 4.388306617736816,
+ "step": 4402
+ },
+ {
+ "epoch": 61.15371179039301,
+ "grad_norm": 0.018965506926178932,
+ "learning_rate": 0.0006,
+ "loss": 4.388340950012207,
+ "step": 4403
+ },
+ {
+ "epoch": 61.16768558951965,
+ "grad_norm": 0.01958410255610943,
+ "learning_rate": 0.0006,
+ "loss": 4.35012149810791,
+ "step": 4404
+ },
+ {
+ "epoch": 61.18165938864629,
+ "grad_norm": 0.01938101276755333,
+ "learning_rate": 0.0006,
+ "loss": 4.368372917175293,
+ "step": 4405
+ },
+ {
+ "epoch": 61.19563318777293,
+ "grad_norm": 0.016556516289711,
+ "learning_rate": 0.0006,
+ "loss": 4.486198425292969,
+ "step": 4406
+ },
+ {
+ "epoch": 61.209606986899566,
+ "grad_norm": 0.017031289637088776,
+ "learning_rate": 0.0006,
+ "loss": 4.421676158905029,
+ "step": 4407
+ },
+ {
+ "epoch": 61.223580786026204,
+ "grad_norm": 0.018023479729890823,
+ "learning_rate": 0.0006,
+ "loss": 4.38751220703125,
+ "step": 4408
+ },
+ {
+ "epoch": 61.237554585152836,
+ "grad_norm": 0.0178135447204113,
+ "learning_rate": 0.0006,
+ "loss": 4.336528301239014,
+ "step": 4409
+ },
+ {
+ "epoch": 61.251528384279474,
+ "grad_norm": 0.016782190650701523,
+ "learning_rate": 0.0006,
+ "loss": 4.386693954467773,
+ "step": 4410
+ },
+ {
+ "epoch": 61.26550218340611,
+ "grad_norm": 0.015494848601520061,
+ "learning_rate": 0.0006,
+ "loss": 4.308948040008545,
+ "step": 4411
+ },
+ {
+ "epoch": 61.27947598253275,
+ "grad_norm": 0.015002378262579441,
+ "learning_rate": 0.0006,
+ "loss": 4.314206123352051,
+ "step": 4412
+ },
+ {
+ "epoch": 61.29344978165939,
+ "grad_norm": 0.015798920765519142,
+ "learning_rate": 0.0006,
+ "loss": 4.502900123596191,
+ "step": 4413
+ },
+ {
+ "epoch": 61.30742358078603,
+ "grad_norm": 0.015444336459040642,
+ "learning_rate": 0.0006,
+ "loss": 4.473160743713379,
+ "step": 4414
+ },
+ {
+ "epoch": 61.32139737991266,
+ "grad_norm": 0.017025096341967583,
+ "learning_rate": 0.0006,
+ "loss": 4.460621356964111,
+ "step": 4415
+ },
+ {
+ "epoch": 61.3353711790393,
+ "grad_norm": 0.019502537325024605,
+ "learning_rate": 0.0006,
+ "loss": 4.385551452636719,
+ "step": 4416
+ },
+ {
+ "epoch": 61.34934497816594,
+ "grad_norm": 0.018529541790485382,
+ "learning_rate": 0.0006,
+ "loss": 4.425390720367432,
+ "step": 4417
+ },
+ {
+ "epoch": 61.36331877729258,
+ "grad_norm": 0.013155706226825714,
+ "learning_rate": 0.0006,
+ "loss": 4.467020034790039,
+ "step": 4418
+ },
+ {
+ "epoch": 61.377292576419215,
+ "grad_norm": 0.01518090907484293,
+ "learning_rate": 0.0006,
+ "loss": 4.369664192199707,
+ "step": 4419
+ },
+ {
+ "epoch": 61.391266375545854,
+ "grad_norm": 0.01626511849462986,
+ "learning_rate": 0.0006,
+ "loss": 4.463231086730957,
+ "step": 4420
+ },
+ {
+ "epoch": 61.40524017467249,
+ "grad_norm": 0.01701558195054531,
+ "learning_rate": 0.0006,
+ "loss": 4.43708610534668,
+ "step": 4421
+ },
+ {
+ "epoch": 61.419213973799124,
+ "grad_norm": 0.01761862449347973,
+ "learning_rate": 0.0006,
+ "loss": 4.3314208984375,
+ "step": 4422
+ },
+ {
+ "epoch": 61.43318777292576,
+ "grad_norm": 0.017084648832678795,
+ "learning_rate": 0.0006,
+ "loss": 4.39676570892334,
+ "step": 4423
+ },
+ {
+ "epoch": 61.4471615720524,
+ "grad_norm": 0.01667078025639057,
+ "learning_rate": 0.0006,
+ "loss": 4.3789873123168945,
+ "step": 4424
+ },
+ {
+ "epoch": 61.46113537117904,
+ "grad_norm": 0.01877436414361,
+ "learning_rate": 0.0006,
+ "loss": 4.415275573730469,
+ "step": 4425
+ },
+ {
+ "epoch": 61.47510917030568,
+ "grad_norm": 0.020250486209988594,
+ "learning_rate": 0.0006,
+ "loss": 4.478635787963867,
+ "step": 4426
+ },
+ {
+ "epoch": 61.48908296943232,
+ "grad_norm": 0.020244868472218513,
+ "learning_rate": 0.0006,
+ "loss": 4.296854019165039,
+ "step": 4427
+ },
+ {
+ "epoch": 61.50305676855895,
+ "grad_norm": 0.018723875284194946,
+ "learning_rate": 0.0006,
+ "loss": 4.3760881423950195,
+ "step": 4428
+ },
+ {
+ "epoch": 61.51703056768559,
+ "grad_norm": 0.015138164162635803,
+ "learning_rate": 0.0006,
+ "loss": 4.39974308013916,
+ "step": 4429
+ },
+ {
+ "epoch": 61.531004366812226,
+ "grad_norm": 0.015111138112843037,
+ "learning_rate": 0.0006,
+ "loss": 4.34659481048584,
+ "step": 4430
+ },
+ {
+ "epoch": 61.544978165938865,
+ "grad_norm": 0.0164666585624218,
+ "learning_rate": 0.0006,
+ "loss": 4.415737152099609,
+ "step": 4431
+ },
+ {
+ "epoch": 61.5589519650655,
+ "grad_norm": 0.020891312509775162,
+ "learning_rate": 0.0006,
+ "loss": 4.300319671630859,
+ "step": 4432
+ },
+ {
+ "epoch": 61.57292576419214,
+ "grad_norm": 0.02089156024158001,
+ "learning_rate": 0.0006,
+ "loss": 4.373029708862305,
+ "step": 4433
+ },
+ {
+ "epoch": 61.58689956331878,
+ "grad_norm": 0.020114559680223465,
+ "learning_rate": 0.0006,
+ "loss": 4.434499740600586,
+ "step": 4434
+ },
+ {
+ "epoch": 61.60087336244541,
+ "grad_norm": 0.01791365258395672,
+ "learning_rate": 0.0006,
+ "loss": 4.466872215270996,
+ "step": 4435
+ },
+ {
+ "epoch": 61.61484716157205,
+ "grad_norm": 0.01879458874464035,
+ "learning_rate": 0.0006,
+ "loss": 4.5176520347595215,
+ "step": 4436
+ },
+ {
+ "epoch": 61.62882096069869,
+ "grad_norm": 0.016371937468647957,
+ "learning_rate": 0.0006,
+ "loss": 4.3507585525512695,
+ "step": 4437
+ },
+ {
+ "epoch": 61.64279475982533,
+ "grad_norm": 0.01648143120110035,
+ "learning_rate": 0.0006,
+ "loss": 4.398778438568115,
+ "step": 4438
+ },
+ {
+ "epoch": 61.65676855895197,
+ "grad_norm": 0.017240576446056366,
+ "learning_rate": 0.0006,
+ "loss": 4.513933181762695,
+ "step": 4439
+ },
+ {
+ "epoch": 61.670742358078606,
+ "grad_norm": 0.01734284684062004,
+ "learning_rate": 0.0006,
+ "loss": 4.315343856811523,
+ "step": 4440
+ },
+ {
+ "epoch": 61.68471615720524,
+ "grad_norm": 0.01639772206544876,
+ "learning_rate": 0.0006,
+ "loss": 4.4641218185424805,
+ "step": 4441
+ },
+ {
+ "epoch": 61.698689956331876,
+ "grad_norm": 0.0164708960801363,
+ "learning_rate": 0.0006,
+ "loss": 4.439968585968018,
+ "step": 4442
+ },
+ {
+ "epoch": 61.712663755458514,
+ "grad_norm": 0.01757933758199215,
+ "learning_rate": 0.0006,
+ "loss": 4.408327102661133,
+ "step": 4443
+ },
+ {
+ "epoch": 61.72663755458515,
+ "grad_norm": 0.01789589412510395,
+ "learning_rate": 0.0006,
+ "loss": 4.478257179260254,
+ "step": 4444
+ },
+ {
+ "epoch": 61.74061135371179,
+ "grad_norm": 0.01822184957563877,
+ "learning_rate": 0.0006,
+ "loss": 4.412046909332275,
+ "step": 4445
+ },
+ {
+ "epoch": 61.75458515283843,
+ "grad_norm": 0.018624618649482727,
+ "learning_rate": 0.0006,
+ "loss": 4.427122116088867,
+ "step": 4446
+ },
+ {
+ "epoch": 61.76855895196506,
+ "grad_norm": 0.018674887716770172,
+ "learning_rate": 0.0006,
+ "loss": 4.429147720336914,
+ "step": 4447
+ },
+ {
+ "epoch": 61.7825327510917,
+ "grad_norm": 0.015317119657993317,
+ "learning_rate": 0.0006,
+ "loss": 4.2808451652526855,
+ "step": 4448
+ },
+ {
+ "epoch": 61.79650655021834,
+ "grad_norm": 0.015400845557451248,
+ "learning_rate": 0.0006,
+ "loss": 4.420193672180176,
+ "step": 4449
+ },
+ {
+ "epoch": 61.81048034934498,
+ "grad_norm": 0.016743650659918785,
+ "learning_rate": 0.0006,
+ "loss": 4.498254776000977,
+ "step": 4450
+ },
+ {
+ "epoch": 61.82445414847162,
+ "grad_norm": 0.014330781064927578,
+ "learning_rate": 0.0006,
+ "loss": 4.445384979248047,
+ "step": 4451
+ },
+ {
+ "epoch": 61.838427947598255,
+ "grad_norm": 0.015046479180455208,
+ "learning_rate": 0.0006,
+ "loss": 4.368041038513184,
+ "step": 4452
+ },
+ {
+ "epoch": 61.852401746724894,
+ "grad_norm": 0.01630168780684471,
+ "learning_rate": 0.0006,
+ "loss": 4.460142135620117,
+ "step": 4453
+ },
+ {
+ "epoch": 61.866375545851525,
+ "grad_norm": 0.01637885719537735,
+ "learning_rate": 0.0006,
+ "loss": 4.416606426239014,
+ "step": 4454
+ },
+ {
+ "epoch": 61.880349344978164,
+ "grad_norm": 0.016785888001322746,
+ "learning_rate": 0.0006,
+ "loss": 4.452545642852783,
+ "step": 4455
+ },
+ {
+ "epoch": 61.8943231441048,
+ "grad_norm": 0.015044555068016052,
+ "learning_rate": 0.0006,
+ "loss": 4.427334785461426,
+ "step": 4456
+ },
+ {
+ "epoch": 61.90829694323144,
+ "grad_norm": 0.013699211180210114,
+ "learning_rate": 0.0006,
+ "loss": 4.408761024475098,
+ "step": 4457
+ },
+ {
+ "epoch": 61.92227074235808,
+ "grad_norm": 0.014021803624927998,
+ "learning_rate": 0.0006,
+ "loss": 4.543607711791992,
+ "step": 4458
+ },
+ {
+ "epoch": 61.93624454148472,
+ "grad_norm": 0.015438460744917393,
+ "learning_rate": 0.0006,
+ "loss": 4.313363075256348,
+ "step": 4459
+ },
+ {
+ "epoch": 61.95021834061135,
+ "grad_norm": 0.017515428364276886,
+ "learning_rate": 0.0006,
+ "loss": 4.310412406921387,
+ "step": 4460
+ },
+ {
+ "epoch": 61.96419213973799,
+ "grad_norm": 0.016683880239725113,
+ "learning_rate": 0.0006,
+ "loss": 4.3378448486328125,
+ "step": 4461
+ },
+ {
+ "epoch": 61.97816593886463,
+ "grad_norm": 0.019384048879146576,
+ "learning_rate": 0.0006,
+ "loss": 4.450568675994873,
+ "step": 4462
+ },
+ {
+ "epoch": 61.992139737991266,
+ "grad_norm": 0.020651493221521378,
+ "learning_rate": 0.0006,
+ "loss": 4.453197479248047,
+ "step": 4463
+ },
+ {
+ "epoch": 62.0,
+ "grad_norm": 0.020991764962673187,
+ "learning_rate": 0.0006,
+ "loss": 4.441527366638184,
+ "step": 4464
+ },
+ {
+ "epoch": 62.0,
+ "eval_loss": 4.700193881988525,
+ "eval_runtime": 56.6098,
+ "eval_samples_per_second": 43.137,
+ "eval_steps_per_second": 1.36,
+ "step": 4464
+ },
+ {
+ "epoch": 62.01397379912664,
+ "grad_norm": 0.01811373233795166,
+ "learning_rate": 0.0006,
+ "loss": 4.37346076965332,
+ "step": 4465
+ },
+ {
+ "epoch": 62.02794759825328,
+ "grad_norm": 0.01667127199470997,
+ "learning_rate": 0.0006,
+ "loss": 4.237151145935059,
+ "step": 4466
+ },
+ {
+ "epoch": 62.041921397379916,
+ "grad_norm": 0.014971534721553326,
+ "learning_rate": 0.0006,
+ "loss": 4.367053031921387,
+ "step": 4467
+ },
+ {
+ "epoch": 62.05589519650655,
+ "grad_norm": 0.015655893832445145,
+ "learning_rate": 0.0006,
+ "loss": 4.437426567077637,
+ "step": 4468
+ },
+ {
+ "epoch": 62.069868995633186,
+ "grad_norm": 0.014935447834432125,
+ "learning_rate": 0.0006,
+ "loss": 4.3635125160217285,
+ "step": 4469
+ },
+ {
+ "epoch": 62.083842794759825,
+ "grad_norm": 0.015363802202045918,
+ "learning_rate": 0.0006,
+ "loss": 4.40492057800293,
+ "step": 4470
+ },
+ {
+ "epoch": 62.09781659388646,
+ "grad_norm": 0.0159201230853796,
+ "learning_rate": 0.0006,
+ "loss": 4.392058849334717,
+ "step": 4471
+ },
+ {
+ "epoch": 62.1117903930131,
+ "grad_norm": 0.014598241075873375,
+ "learning_rate": 0.0006,
+ "loss": 4.369944095611572,
+ "step": 4472
+ },
+ {
+ "epoch": 62.12576419213974,
+ "grad_norm": 0.015394841320812702,
+ "learning_rate": 0.0006,
+ "loss": 4.425097942352295,
+ "step": 4473
+ },
+ {
+ "epoch": 62.13973799126637,
+ "grad_norm": 0.015913628041744232,
+ "learning_rate": 0.0006,
+ "loss": 4.420596122741699,
+ "step": 4474
+ },
+ {
+ "epoch": 62.15371179039301,
+ "grad_norm": 0.01575258933007717,
+ "learning_rate": 0.0006,
+ "loss": 4.32290506362915,
+ "step": 4475
+ },
+ {
+ "epoch": 62.16768558951965,
+ "grad_norm": 0.0155344782397151,
+ "learning_rate": 0.0006,
+ "loss": 4.273068428039551,
+ "step": 4476
+ },
+ {
+ "epoch": 62.18165938864629,
+ "grad_norm": 0.014947040006518364,
+ "learning_rate": 0.0006,
+ "loss": 4.384868621826172,
+ "step": 4477
+ },
+ {
+ "epoch": 62.19563318777293,
+ "grad_norm": 0.01557591650635004,
+ "learning_rate": 0.0006,
+ "loss": 4.458410263061523,
+ "step": 4478
+ },
+ {
+ "epoch": 62.209606986899566,
+ "grad_norm": 0.015182198956608772,
+ "learning_rate": 0.0006,
+ "loss": 4.300738334655762,
+ "step": 4479
+ },
+ {
+ "epoch": 62.223580786026204,
+ "grad_norm": 0.01634717360138893,
+ "learning_rate": 0.0006,
+ "loss": 4.436875820159912,
+ "step": 4480
+ },
+ {
+ "epoch": 62.237554585152836,
+ "grad_norm": 0.018813718110322952,
+ "learning_rate": 0.0006,
+ "loss": 4.454665184020996,
+ "step": 4481
+ },
+ {
+ "epoch": 62.251528384279474,
+ "grad_norm": 0.020965630188584328,
+ "learning_rate": 0.0006,
+ "loss": 4.357341766357422,
+ "step": 4482
+ },
+ {
+ "epoch": 62.26550218340611,
+ "grad_norm": 0.020738353952765465,
+ "learning_rate": 0.0006,
+ "loss": 4.456644058227539,
+ "step": 4483
+ },
+ {
+ "epoch": 62.27947598253275,
+ "grad_norm": 0.018813546746969223,
+ "learning_rate": 0.0006,
+ "loss": 4.381941795349121,
+ "step": 4484
+ },
+ {
+ "epoch": 62.29344978165939,
+ "grad_norm": 0.015431761741638184,
+ "learning_rate": 0.0006,
+ "loss": 4.41126823425293,
+ "step": 4485
+ },
+ {
+ "epoch": 62.30742358078603,
+ "grad_norm": 0.016335977241396904,
+ "learning_rate": 0.0006,
+ "loss": 4.419488906860352,
+ "step": 4486
+ },
+ {
+ "epoch": 62.32139737991266,
+ "grad_norm": 0.01832406409084797,
+ "learning_rate": 0.0006,
+ "loss": 4.579026222229004,
+ "step": 4487
+ },
+ {
+ "epoch": 62.3353711790393,
+ "grad_norm": 0.01790742017328739,
+ "learning_rate": 0.0006,
+ "loss": 4.289964199066162,
+ "step": 4488
+ },
+ {
+ "epoch": 62.34934497816594,
+ "grad_norm": 0.017266202718019485,
+ "learning_rate": 0.0006,
+ "loss": 4.415959358215332,
+ "step": 4489
+ },
+ {
+ "epoch": 62.36331877729258,
+ "grad_norm": 0.015976132825016975,
+ "learning_rate": 0.0006,
+ "loss": 4.358587265014648,
+ "step": 4490
+ },
+ {
+ "epoch": 62.377292576419215,
+ "grad_norm": 0.016873881220817566,
+ "learning_rate": 0.0006,
+ "loss": 4.354607105255127,
+ "step": 4491
+ },
+ {
+ "epoch": 62.391266375545854,
+ "grad_norm": 0.01476745493710041,
+ "learning_rate": 0.0006,
+ "loss": 4.384904384613037,
+ "step": 4492
+ },
+ {
+ "epoch": 62.40524017467249,
+ "grad_norm": 0.013810782693326473,
+ "learning_rate": 0.0006,
+ "loss": 4.459230899810791,
+ "step": 4493
+ },
+ {
+ "epoch": 62.419213973799124,
+ "grad_norm": 0.015150421299040318,
+ "learning_rate": 0.0006,
+ "loss": 4.369471549987793,
+ "step": 4494
+ },
+ {
+ "epoch": 62.43318777292576,
+ "grad_norm": 0.014539610594511032,
+ "learning_rate": 0.0006,
+ "loss": 4.3564982414245605,
+ "step": 4495
+ },
+ {
+ "epoch": 62.4471615720524,
+ "grad_norm": 0.015554017387330532,
+ "learning_rate": 0.0006,
+ "loss": 4.414379119873047,
+ "step": 4496
+ },
+ {
+ "epoch": 62.46113537117904,
+ "grad_norm": 0.016608357429504395,
+ "learning_rate": 0.0006,
+ "loss": 4.311585426330566,
+ "step": 4497
+ },
+ {
+ "epoch": 62.47510917030568,
+ "grad_norm": 0.01885881833732128,
+ "learning_rate": 0.0006,
+ "loss": 4.306629180908203,
+ "step": 4498
+ },
+ {
+ "epoch": 62.48908296943232,
+ "grad_norm": 0.022128663957118988,
+ "learning_rate": 0.0006,
+ "loss": 4.345483779907227,
+ "step": 4499
+ },
+ {
+ "epoch": 62.50305676855895,
+ "grad_norm": 0.023401018232107162,
+ "learning_rate": 0.0006,
+ "loss": 4.398204326629639,
+ "step": 4500
+ },
+ {
+ "epoch": 62.51703056768559,
+ "grad_norm": 0.022080009803175926,
+ "learning_rate": 0.0006,
+ "loss": 4.375027179718018,
+ "step": 4501
+ },
+ {
+ "epoch": 62.531004366812226,
+ "grad_norm": 0.020004864782094955,
+ "learning_rate": 0.0006,
+ "loss": 4.33494758605957,
+ "step": 4502
+ },
+ {
+ "epoch": 62.544978165938865,
+ "grad_norm": 0.01786700263619423,
+ "learning_rate": 0.0006,
+ "loss": 4.4428300857543945,
+ "step": 4503
+ },
+ {
+ "epoch": 62.5589519650655,
+ "grad_norm": 0.01885388419032097,
+ "learning_rate": 0.0006,
+ "loss": 4.4138946533203125,
+ "step": 4504
+ },
+ {
+ "epoch": 62.57292576419214,
+ "grad_norm": 0.019890939816832542,
+ "learning_rate": 0.0006,
+ "loss": 4.278632164001465,
+ "step": 4505
+ },
+ {
+ "epoch": 62.58689956331878,
+ "grad_norm": 0.02109031192958355,
+ "learning_rate": 0.0006,
+ "loss": 4.3756327629089355,
+ "step": 4506
+ },
+ {
+ "epoch": 62.60087336244541,
+ "grad_norm": 0.020813899114727974,
+ "learning_rate": 0.0006,
+ "loss": 4.410213470458984,
+ "step": 4507
+ },
+ {
+ "epoch": 62.61484716157205,
+ "grad_norm": 0.021655689924955368,
+ "learning_rate": 0.0006,
+ "loss": 4.30633020401001,
+ "step": 4508
+ },
+ {
+ "epoch": 62.62882096069869,
+ "grad_norm": 0.018155649304389954,
+ "learning_rate": 0.0006,
+ "loss": 4.407284736633301,
+ "step": 4509
+ },
+ {
+ "epoch": 62.64279475982533,
+ "grad_norm": 0.017765069380402565,
+ "learning_rate": 0.0006,
+ "loss": 4.321715354919434,
+ "step": 4510
+ },
+ {
+ "epoch": 62.65676855895197,
+ "grad_norm": 0.017724554985761642,
+ "learning_rate": 0.0006,
+ "loss": 4.493371963500977,
+ "step": 4511
+ },
+ {
+ "epoch": 62.670742358078606,
+ "grad_norm": 0.01492092851549387,
+ "learning_rate": 0.0006,
+ "loss": 4.443025588989258,
+ "step": 4512
+ },
+ {
+ "epoch": 62.68471615720524,
+ "grad_norm": 0.01615324057638645,
+ "learning_rate": 0.0006,
+ "loss": 4.3289594650268555,
+ "step": 4513
+ },
+ {
+ "epoch": 62.698689956331876,
+ "grad_norm": 0.016569366678595543,
+ "learning_rate": 0.0006,
+ "loss": 4.369720458984375,
+ "step": 4514
+ },
+ {
+ "epoch": 62.712663755458514,
+ "grad_norm": 0.016509901732206345,
+ "learning_rate": 0.0006,
+ "loss": 4.292316436767578,
+ "step": 4515
+ },
+ {
+ "epoch": 62.72663755458515,
+ "grad_norm": 0.014270003885030746,
+ "learning_rate": 0.0006,
+ "loss": 4.349854469299316,
+ "step": 4516
+ },
+ {
+ "epoch": 62.74061135371179,
+ "grad_norm": 0.014235340990126133,
+ "learning_rate": 0.0006,
+ "loss": 4.473727226257324,
+ "step": 4517
+ },
+ {
+ "epoch": 62.75458515283843,
+ "grad_norm": 0.015163225121796131,
+ "learning_rate": 0.0006,
+ "loss": 4.452878475189209,
+ "step": 4518
+ },
+ {
+ "epoch": 62.76855895196506,
+ "grad_norm": 0.015748055651783943,
+ "learning_rate": 0.0006,
+ "loss": 4.521087169647217,
+ "step": 4519
+ },
+ {
+ "epoch": 62.7825327510917,
+ "grad_norm": 0.01661108434200287,
+ "learning_rate": 0.0006,
+ "loss": 4.443885326385498,
+ "step": 4520
+ },
+ {
+ "epoch": 62.79650655021834,
+ "grad_norm": 0.01610301062464714,
+ "learning_rate": 0.0006,
+ "loss": 4.312990188598633,
+ "step": 4521
+ },
+ {
+ "epoch": 62.81048034934498,
+ "grad_norm": 0.017007583752274513,
+ "learning_rate": 0.0006,
+ "loss": 4.408937454223633,
+ "step": 4522
+ },
+ {
+ "epoch": 62.82445414847162,
+ "grad_norm": 0.01749461516737938,
+ "learning_rate": 0.0006,
+ "loss": 4.346066951751709,
+ "step": 4523
+ },
+ {
+ "epoch": 62.838427947598255,
+ "grad_norm": 0.018243493512272835,
+ "learning_rate": 0.0006,
+ "loss": 4.357658386230469,
+ "step": 4524
+ },
+ {
+ "epoch": 62.852401746724894,
+ "grad_norm": 0.017499985173344612,
+ "learning_rate": 0.0006,
+ "loss": 4.374061584472656,
+ "step": 4525
+ },
+ {
+ "epoch": 62.866375545851525,
+ "grad_norm": 0.014536325819790363,
+ "learning_rate": 0.0006,
+ "loss": 4.189321994781494,
+ "step": 4526
+ },
+ {
+ "epoch": 62.880349344978164,
+ "grad_norm": 0.014728891663253307,
+ "learning_rate": 0.0006,
+ "loss": 4.425910949707031,
+ "step": 4527
+ },
+ {
+ "epoch": 62.8943231441048,
+ "grad_norm": 0.015013772062957287,
+ "learning_rate": 0.0006,
+ "loss": 4.337319850921631,
+ "step": 4528
+ },
+ {
+ "epoch": 62.90829694323144,
+ "grad_norm": 0.01626027375459671,
+ "learning_rate": 0.0006,
+ "loss": 4.330339431762695,
+ "step": 4529
+ },
+ {
+ "epoch": 62.92227074235808,
+ "grad_norm": 0.014773444272577763,
+ "learning_rate": 0.0006,
+ "loss": 4.4188737869262695,
+ "step": 4530
+ },
+ {
+ "epoch": 62.93624454148472,
+ "grad_norm": 0.01422932744026184,
+ "learning_rate": 0.0006,
+ "loss": 4.355986595153809,
+ "step": 4531
+ },
+ {
+ "epoch": 62.95021834061135,
+ "grad_norm": 0.012170110829174519,
+ "learning_rate": 0.0006,
+ "loss": 4.405722618103027,
+ "step": 4532
+ },
+ {
+ "epoch": 62.96419213973799,
+ "grad_norm": 0.013233253732323647,
+ "learning_rate": 0.0006,
+ "loss": 4.377286911010742,
+ "step": 4533
+ },
+ {
+ "epoch": 62.97816593886463,
+ "grad_norm": 0.014497299678623676,
+ "learning_rate": 0.0006,
+ "loss": 4.508796691894531,
+ "step": 4534
+ },
+ {
+ "epoch": 62.992139737991266,
+ "grad_norm": 0.014718741178512573,
+ "learning_rate": 0.0006,
+ "loss": 4.370919704437256,
+ "step": 4535
+ },
+ {
+ "epoch": 63.0,
+ "grad_norm": 0.015833502635359764,
+ "learning_rate": 0.0006,
+ "loss": 4.258753776550293,
+ "step": 4536
+ },
+ {
+ "epoch": 63.0,
+ "eval_loss": 4.714163303375244,
+ "eval_runtime": 57.0269,
+ "eval_samples_per_second": 42.822,
+ "eval_steps_per_second": 1.35,
+ "step": 4536
+ },
+ {
+ "epoch": 63.01397379912664,
+ "grad_norm": 0.014277510344982147,
+ "learning_rate": 0.0006,
+ "loss": 4.3112640380859375,
+ "step": 4537
+ },
+ {
+ "epoch": 63.02794759825328,
+ "grad_norm": 0.015377935022115707,
+ "learning_rate": 0.0006,
+ "loss": 4.342782974243164,
+ "step": 4538
+ },
+ {
+ "epoch": 63.041921397379916,
+ "grad_norm": 0.015406244434416294,
+ "learning_rate": 0.0006,
+ "loss": 4.474056720733643,
+ "step": 4539
+ },
+ {
+ "epoch": 63.05589519650655,
+ "grad_norm": 0.018121499568223953,
+ "learning_rate": 0.0006,
+ "loss": 4.397963523864746,
+ "step": 4540
+ },
+ {
+ "epoch": 63.069868995633186,
+ "grad_norm": 0.022634636610746384,
+ "learning_rate": 0.0006,
+ "loss": 4.306631088256836,
+ "step": 4541
+ },
+ {
+ "epoch": 63.083842794759825,
+ "grad_norm": 0.026448970660567284,
+ "learning_rate": 0.0006,
+ "loss": 4.283175468444824,
+ "step": 4542
+ },
+ {
+ "epoch": 63.09781659388646,
+ "grad_norm": 0.024104537442326546,
+ "learning_rate": 0.0006,
+ "loss": 4.315826416015625,
+ "step": 4543
+ },
+ {
+ "epoch": 63.1117903930131,
+ "grad_norm": 0.019040480256080627,
+ "learning_rate": 0.0006,
+ "loss": 4.44057035446167,
+ "step": 4544
+ },
+ {
+ "epoch": 63.12576419213974,
+ "grad_norm": 0.02006000280380249,
+ "learning_rate": 0.0006,
+ "loss": 4.405603408813477,
+ "step": 4545
+ },
+ {
+ "epoch": 63.13973799126637,
+ "grad_norm": 0.019980071112513542,
+ "learning_rate": 0.0006,
+ "loss": 4.393726348876953,
+ "step": 4546
+ },
+ {
+ "epoch": 63.15371179039301,
+ "grad_norm": 0.01641865260899067,
+ "learning_rate": 0.0006,
+ "loss": 4.272546768188477,
+ "step": 4547
+ },
+ {
+ "epoch": 63.16768558951965,
+ "grad_norm": 0.01890353485941887,
+ "learning_rate": 0.0006,
+ "loss": 4.382566928863525,
+ "step": 4548
+ },
+ {
+ "epoch": 63.18165938864629,
+ "grad_norm": 0.018370937556028366,
+ "learning_rate": 0.0006,
+ "loss": 4.423953533172607,
+ "step": 4549
+ },
+ {
+ "epoch": 63.19563318777293,
+ "grad_norm": 0.014985294081270695,
+ "learning_rate": 0.0006,
+ "loss": 4.308627128601074,
+ "step": 4550
+ },
+ {
+ "epoch": 63.209606986899566,
+ "grad_norm": 0.017513327300548553,
+ "learning_rate": 0.0006,
+ "loss": 4.248819351196289,
+ "step": 4551
+ },
+ {
+ "epoch": 63.223580786026204,
+ "grad_norm": 0.019464161247015,
+ "learning_rate": 0.0006,
+ "loss": 4.359729766845703,
+ "step": 4552
+ },
+ {
+ "epoch": 63.237554585152836,
+ "grad_norm": 0.018678035587072372,
+ "learning_rate": 0.0006,
+ "loss": 4.3299665451049805,
+ "step": 4553
+ },
+ {
+ "epoch": 63.251528384279474,
+ "grad_norm": 0.017909523099660873,
+ "learning_rate": 0.0006,
+ "loss": 4.246147155761719,
+ "step": 4554
+ },
+ {
+ "epoch": 63.26550218340611,
+ "grad_norm": 0.01747293211519718,
+ "learning_rate": 0.0006,
+ "loss": 4.41165018081665,
+ "step": 4555
+ },
+ {
+ "epoch": 63.27947598253275,
+ "grad_norm": 0.018670853227376938,
+ "learning_rate": 0.0006,
+ "loss": 4.409883499145508,
+ "step": 4556
+ },
+ {
+ "epoch": 63.29344978165939,
+ "grad_norm": 0.0189529862254858,
+ "learning_rate": 0.0006,
+ "loss": 4.418654918670654,
+ "step": 4557
+ },
+ {
+ "epoch": 63.30742358078603,
+ "grad_norm": 0.018394414335489273,
+ "learning_rate": 0.0006,
+ "loss": 4.38355827331543,
+ "step": 4558
+ },
+ {
+ "epoch": 63.32139737991266,
+ "grad_norm": 0.018317820504307747,
+ "learning_rate": 0.0006,
+ "loss": 4.456843376159668,
+ "step": 4559
+ },
+ {
+ "epoch": 63.3353711790393,
+ "grad_norm": 0.019225597381591797,
+ "learning_rate": 0.0006,
+ "loss": 4.232264518737793,
+ "step": 4560
+ },
+ {
+ "epoch": 63.34934497816594,
+ "grad_norm": 0.016663681715726852,
+ "learning_rate": 0.0006,
+ "loss": 4.393517017364502,
+ "step": 4561
+ },
+ {
+ "epoch": 63.36331877729258,
+ "grad_norm": 0.014975365251302719,
+ "learning_rate": 0.0006,
+ "loss": 4.375183582305908,
+ "step": 4562
+ },
+ {
+ "epoch": 63.377292576419215,
+ "grad_norm": 0.01771732047200203,
+ "learning_rate": 0.0006,
+ "loss": 4.3073410987854,
+ "step": 4563
+ },
+ {
+ "epoch": 63.391266375545854,
+ "grad_norm": 0.01663549244403839,
+ "learning_rate": 0.0006,
+ "loss": 4.311003684997559,
+ "step": 4564
+ },
+ {
+ "epoch": 63.40524017467249,
+ "grad_norm": 0.016301177442073822,
+ "learning_rate": 0.0006,
+ "loss": 4.326807022094727,
+ "step": 4565
+ },
+ {
+ "epoch": 63.419213973799124,
+ "grad_norm": 0.016781188547611237,
+ "learning_rate": 0.0006,
+ "loss": 4.306562423706055,
+ "step": 4566
+ },
+ {
+ "epoch": 63.43318777292576,
+ "grad_norm": 0.01767745055258274,
+ "learning_rate": 0.0006,
+ "loss": 4.367082595825195,
+ "step": 4567
+ },
+ {
+ "epoch": 63.4471615720524,
+ "grad_norm": 0.016728565096855164,
+ "learning_rate": 0.0006,
+ "loss": 4.327400207519531,
+ "step": 4568
+ },
+ {
+ "epoch": 63.46113537117904,
+ "grad_norm": 0.016289403662085533,
+ "learning_rate": 0.0006,
+ "loss": 4.392712116241455,
+ "step": 4569
+ },
+ {
+ "epoch": 63.47510917030568,
+ "grad_norm": 0.016233716160058975,
+ "learning_rate": 0.0006,
+ "loss": 4.354372024536133,
+ "step": 4570
+ },
+ {
+ "epoch": 63.48908296943232,
+ "grad_norm": 0.016547974199056625,
+ "learning_rate": 0.0006,
+ "loss": 4.309050559997559,
+ "step": 4571
+ },
+ {
+ "epoch": 63.50305676855895,
+ "grad_norm": 0.018466509878635406,
+ "learning_rate": 0.0006,
+ "loss": 4.318345069885254,
+ "step": 4572
+ },
+ {
+ "epoch": 63.51703056768559,
+ "grad_norm": 0.017685720697045326,
+ "learning_rate": 0.0006,
+ "loss": 4.399726867675781,
+ "step": 4573
+ },
+ {
+ "epoch": 63.531004366812226,
+ "grad_norm": 0.015801234170794487,
+ "learning_rate": 0.0006,
+ "loss": 4.328495502471924,
+ "step": 4574
+ },
+ {
+ "epoch": 63.544978165938865,
+ "grad_norm": 0.0188873540610075,
+ "learning_rate": 0.0006,
+ "loss": 4.253244400024414,
+ "step": 4575
+ },
+ {
+ "epoch": 63.5589519650655,
+ "grad_norm": 0.01868271268904209,
+ "learning_rate": 0.0006,
+ "loss": 4.463746070861816,
+ "step": 4576
+ },
+ {
+ "epoch": 63.57292576419214,
+ "grad_norm": 0.01844949834048748,
+ "learning_rate": 0.0006,
+ "loss": 4.400300025939941,
+ "step": 4577
+ },
+ {
+ "epoch": 63.58689956331878,
+ "grad_norm": 0.015256480313837528,
+ "learning_rate": 0.0006,
+ "loss": 4.445844650268555,
+ "step": 4578
+ },
+ {
+ "epoch": 63.60087336244541,
+ "grad_norm": 0.015305760316550732,
+ "learning_rate": 0.0006,
+ "loss": 4.317327499389648,
+ "step": 4579
+ },
+ {
+ "epoch": 63.61484716157205,
+ "grad_norm": 0.015200769528746605,
+ "learning_rate": 0.0006,
+ "loss": 4.25112247467041,
+ "step": 4580
+ },
+ {
+ "epoch": 63.62882096069869,
+ "grad_norm": 0.015919538214802742,
+ "learning_rate": 0.0006,
+ "loss": 4.277937889099121,
+ "step": 4581
+ },
+ {
+ "epoch": 63.64279475982533,
+ "grad_norm": 0.01572665013372898,
+ "learning_rate": 0.0006,
+ "loss": 4.302237510681152,
+ "step": 4582
+ },
+ {
+ "epoch": 63.65676855895197,
+ "grad_norm": 0.016428731381893158,
+ "learning_rate": 0.0006,
+ "loss": 4.36762809753418,
+ "step": 4583
+ },
+ {
+ "epoch": 63.670742358078606,
+ "grad_norm": 0.01706082746386528,
+ "learning_rate": 0.0006,
+ "loss": 4.236526966094971,
+ "step": 4584
+ },
+ {
+ "epoch": 63.68471615720524,
+ "grad_norm": 0.017850136384367943,
+ "learning_rate": 0.0006,
+ "loss": 4.264129638671875,
+ "step": 4585
+ },
+ {
+ "epoch": 63.698689956331876,
+ "grad_norm": 0.016079438850283623,
+ "learning_rate": 0.0006,
+ "loss": 4.449799060821533,
+ "step": 4586
+ },
+ {
+ "epoch": 63.712663755458514,
+ "grad_norm": 0.014996383339166641,
+ "learning_rate": 0.0006,
+ "loss": 4.404574394226074,
+ "step": 4587
+ },
+ {
+ "epoch": 63.72663755458515,
+ "grad_norm": 0.015910431742668152,
+ "learning_rate": 0.0006,
+ "loss": 4.348101615905762,
+ "step": 4588
+ },
+ {
+ "epoch": 63.74061135371179,
+ "grad_norm": 0.01630021072924137,
+ "learning_rate": 0.0006,
+ "loss": 4.3193817138671875,
+ "step": 4589
+ },
+ {
+ "epoch": 63.75458515283843,
+ "grad_norm": 0.014571773819625378,
+ "learning_rate": 0.0006,
+ "loss": 4.374402046203613,
+ "step": 4590
+ },
+ {
+ "epoch": 63.76855895196506,
+ "grad_norm": 0.01662345975637436,
+ "learning_rate": 0.0006,
+ "loss": 4.329075813293457,
+ "step": 4591
+ },
+ {
+ "epoch": 63.7825327510917,
+ "grad_norm": 0.016604198142886162,
+ "learning_rate": 0.0006,
+ "loss": 4.436126232147217,
+ "step": 4592
+ },
+ {
+ "epoch": 63.79650655021834,
+ "grad_norm": 0.016578884795308113,
+ "learning_rate": 0.0006,
+ "loss": 4.361468315124512,
+ "step": 4593
+ },
+ {
+ "epoch": 63.81048034934498,
+ "grad_norm": 0.016339825466275215,
+ "learning_rate": 0.0006,
+ "loss": 4.449892044067383,
+ "step": 4594
+ },
+ {
+ "epoch": 63.82445414847162,
+ "grad_norm": 0.016805455088615417,
+ "learning_rate": 0.0006,
+ "loss": 4.415866851806641,
+ "step": 4595
+ },
+ {
+ "epoch": 63.838427947598255,
+ "grad_norm": 0.017366213724017143,
+ "learning_rate": 0.0006,
+ "loss": 4.361730575561523,
+ "step": 4596
+ },
+ {
+ "epoch": 63.852401746724894,
+ "grad_norm": 0.015599003992974758,
+ "learning_rate": 0.0006,
+ "loss": 4.363637924194336,
+ "step": 4597
+ },
+ {
+ "epoch": 63.866375545851525,
+ "grad_norm": 0.015674632042646408,
+ "learning_rate": 0.0006,
+ "loss": 4.332045078277588,
+ "step": 4598
+ },
+ {
+ "epoch": 63.880349344978164,
+ "grad_norm": 0.015350443311035633,
+ "learning_rate": 0.0006,
+ "loss": 4.320063591003418,
+ "step": 4599
+ },
+ {
+ "epoch": 63.8943231441048,
+ "grad_norm": 0.015388146974146366,
+ "learning_rate": 0.0006,
+ "loss": 4.398199081420898,
+ "step": 4600
+ },
+ {
+ "epoch": 63.90829694323144,
+ "grad_norm": 0.01450140681117773,
+ "learning_rate": 0.0006,
+ "loss": 4.244076728820801,
+ "step": 4601
+ },
+ {
+ "epoch": 63.92227074235808,
+ "grad_norm": 0.01637491025030613,
+ "learning_rate": 0.0006,
+ "loss": 4.213549613952637,
+ "step": 4602
+ },
+ {
+ "epoch": 63.93624454148472,
+ "grad_norm": 0.0159814041107893,
+ "learning_rate": 0.0006,
+ "loss": 4.421775817871094,
+ "step": 4603
+ },
+ {
+ "epoch": 63.95021834061135,
+ "grad_norm": 0.017386827617883682,
+ "learning_rate": 0.0006,
+ "loss": 4.340023994445801,
+ "step": 4604
+ },
+ {
+ "epoch": 63.96419213973799,
+ "grad_norm": 0.019679656252264977,
+ "learning_rate": 0.0006,
+ "loss": 4.352534294128418,
+ "step": 4605
+ },
+ {
+ "epoch": 63.97816593886463,
+ "grad_norm": 0.022391516715288162,
+ "learning_rate": 0.0006,
+ "loss": 4.3538312911987305,
+ "step": 4606
+ },
+ {
+ "epoch": 63.992139737991266,
+ "grad_norm": 0.023281006142497063,
+ "learning_rate": 0.0006,
+ "loss": 4.432089328765869,
+ "step": 4607
+ },
+ {
+ "epoch": 64.0,
+ "grad_norm": 0.019916288554668427,
+ "learning_rate": 0.0006,
+ "loss": 4.466937065124512,
+ "step": 4608
+ },
+ {
+ "epoch": 64.0,
+ "eval_loss": 4.659008502960205,
+ "eval_runtime": 56.0019,
+ "eval_samples_per_second": 43.606,
+ "eval_steps_per_second": 1.375,
+ "step": 4608
+ },
+ {
+ "epoch": 64.01397379912663,
+ "grad_norm": 0.017351187765598297,
+ "learning_rate": 0.0006,
+ "loss": 4.294129371643066,
+ "step": 4609
+ },
+ {
+ "epoch": 64.02794759825328,
+ "grad_norm": 0.02405865117907524,
+ "learning_rate": 0.0006,
+ "loss": 4.266717910766602,
+ "step": 4610
+ },
+ {
+ "epoch": 64.04192139737991,
+ "grad_norm": 0.022158386185765266,
+ "learning_rate": 0.0006,
+ "loss": 4.395932197570801,
+ "step": 4611
+ },
+ {
+ "epoch": 64.05589519650655,
+ "grad_norm": 0.019238505512475967,
+ "learning_rate": 0.0006,
+ "loss": 4.266704082489014,
+ "step": 4612
+ },
+ {
+ "epoch": 64.06986899563319,
+ "grad_norm": 0.021670425310730934,
+ "learning_rate": 0.0006,
+ "loss": 4.3160552978515625,
+ "step": 4613
+ },
+ {
+ "epoch": 64.08384279475983,
+ "grad_norm": 0.02076641097664833,
+ "learning_rate": 0.0006,
+ "loss": 4.415626525878906,
+ "step": 4614
+ },
+ {
+ "epoch": 64.09781659388646,
+ "grad_norm": 0.018223078921437263,
+ "learning_rate": 0.0006,
+ "loss": 4.292387008666992,
+ "step": 4615
+ },
+ {
+ "epoch": 64.1117903930131,
+ "grad_norm": 0.017401937395334244,
+ "learning_rate": 0.0006,
+ "loss": 4.424524307250977,
+ "step": 4616
+ },
+ {
+ "epoch": 64.12576419213974,
+ "grad_norm": 0.01663743332028389,
+ "learning_rate": 0.0006,
+ "loss": 4.350032806396484,
+ "step": 4617
+ },
+ {
+ "epoch": 64.13973799126637,
+ "grad_norm": 0.018716629594564438,
+ "learning_rate": 0.0006,
+ "loss": 4.309157371520996,
+ "step": 4618
+ },
+ {
+ "epoch": 64.15371179039302,
+ "grad_norm": 0.019569825381040573,
+ "learning_rate": 0.0006,
+ "loss": 4.358292579650879,
+ "step": 4619
+ },
+ {
+ "epoch": 64.16768558951965,
+ "grad_norm": 0.018994422629475594,
+ "learning_rate": 0.0006,
+ "loss": 4.3338823318481445,
+ "step": 4620
+ },
+ {
+ "epoch": 64.18165938864628,
+ "grad_norm": 0.017690688371658325,
+ "learning_rate": 0.0006,
+ "loss": 4.414468765258789,
+ "step": 4621
+ },
+ {
+ "epoch": 64.19563318777293,
+ "grad_norm": 0.0170609001070261,
+ "learning_rate": 0.0006,
+ "loss": 4.4337568283081055,
+ "step": 4622
+ },
+ {
+ "epoch": 64.20960698689956,
+ "grad_norm": 0.016984131187200546,
+ "learning_rate": 0.0006,
+ "loss": 4.304839134216309,
+ "step": 4623
+ },
+ {
+ "epoch": 64.2235807860262,
+ "grad_norm": 0.01619867794215679,
+ "learning_rate": 0.0006,
+ "loss": 4.389623165130615,
+ "step": 4624
+ },
+ {
+ "epoch": 64.23755458515284,
+ "grad_norm": 0.015428110957145691,
+ "learning_rate": 0.0006,
+ "loss": 4.277515888214111,
+ "step": 4625
+ },
+ {
+ "epoch": 64.25152838427948,
+ "grad_norm": 0.015909532085061073,
+ "learning_rate": 0.0006,
+ "loss": 4.266029357910156,
+ "step": 4626
+ },
+ {
+ "epoch": 64.26550218340611,
+ "grad_norm": 0.014296560548245907,
+ "learning_rate": 0.0006,
+ "loss": 4.286067485809326,
+ "step": 4627
+ },
+ {
+ "epoch": 64.27947598253274,
+ "grad_norm": 0.014065971598029137,
+ "learning_rate": 0.0006,
+ "loss": 4.353528022766113,
+ "step": 4628
+ },
+ {
+ "epoch": 64.29344978165939,
+ "grad_norm": 0.014386293478310108,
+ "learning_rate": 0.0006,
+ "loss": 4.318715572357178,
+ "step": 4629
+ },
+ {
+ "epoch": 64.30742358078602,
+ "grad_norm": 0.014001123607158661,
+ "learning_rate": 0.0006,
+ "loss": 4.362497329711914,
+ "step": 4630
+ },
+ {
+ "epoch": 64.32139737991267,
+ "grad_norm": 0.013408979400992393,
+ "learning_rate": 0.0006,
+ "loss": 4.417397499084473,
+ "step": 4631
+ },
+ {
+ "epoch": 64.3353711790393,
+ "grad_norm": 0.014960093423724174,
+ "learning_rate": 0.0006,
+ "loss": 4.359710693359375,
+ "step": 4632
+ },
+ {
+ "epoch": 64.34934497816595,
+ "grad_norm": 0.014658777043223381,
+ "learning_rate": 0.0006,
+ "loss": 4.318822383880615,
+ "step": 4633
+ },
+ {
+ "epoch": 64.36331877729258,
+ "grad_norm": 0.013766744174063206,
+ "learning_rate": 0.0006,
+ "loss": 4.265679836273193,
+ "step": 4634
+ },
+ {
+ "epoch": 64.37729257641921,
+ "grad_norm": 0.013821255415678024,
+ "learning_rate": 0.0006,
+ "loss": 4.3452911376953125,
+ "step": 4635
+ },
+ {
+ "epoch": 64.39126637554585,
+ "grad_norm": 0.013541066087782383,
+ "learning_rate": 0.0006,
+ "loss": 4.429254531860352,
+ "step": 4636
+ },
+ {
+ "epoch": 64.40524017467249,
+ "grad_norm": 0.013985645025968552,
+ "learning_rate": 0.0006,
+ "loss": 4.304452896118164,
+ "step": 4637
+ },
+ {
+ "epoch": 64.41921397379913,
+ "grad_norm": 0.015079808421432972,
+ "learning_rate": 0.0006,
+ "loss": 4.296828746795654,
+ "step": 4638
+ },
+ {
+ "epoch": 64.43318777292576,
+ "grad_norm": 0.014896944165229797,
+ "learning_rate": 0.0006,
+ "loss": 4.299156665802002,
+ "step": 4639
+ },
+ {
+ "epoch": 64.44716157205241,
+ "grad_norm": 0.013770041987299919,
+ "learning_rate": 0.0006,
+ "loss": 4.351439952850342,
+ "step": 4640
+ },
+ {
+ "epoch": 64.46113537117904,
+ "grad_norm": 0.01531178504228592,
+ "learning_rate": 0.0006,
+ "loss": 4.373715877532959,
+ "step": 4641
+ },
+ {
+ "epoch": 64.47510917030567,
+ "grad_norm": 0.014024450443685055,
+ "learning_rate": 0.0006,
+ "loss": 4.368053436279297,
+ "step": 4642
+ },
+ {
+ "epoch": 64.48908296943232,
+ "grad_norm": 0.013864024542272091,
+ "learning_rate": 0.0006,
+ "loss": 4.378564834594727,
+ "step": 4643
+ },
+ {
+ "epoch": 64.50305676855895,
+ "grad_norm": 0.013925445266067982,
+ "learning_rate": 0.0006,
+ "loss": 4.261663436889648,
+ "step": 4644
+ },
+ {
+ "epoch": 64.5170305676856,
+ "grad_norm": 0.015011338517069817,
+ "learning_rate": 0.0006,
+ "loss": 4.337285995483398,
+ "step": 4645
+ },
+ {
+ "epoch": 64.53100436681223,
+ "grad_norm": 0.016941986978054047,
+ "learning_rate": 0.0006,
+ "loss": 4.336877822875977,
+ "step": 4646
+ },
+ {
+ "epoch": 64.54497816593886,
+ "grad_norm": 0.01719909906387329,
+ "learning_rate": 0.0006,
+ "loss": 4.149813175201416,
+ "step": 4647
+ },
+ {
+ "epoch": 64.5589519650655,
+ "grad_norm": 0.01608402095735073,
+ "learning_rate": 0.0006,
+ "loss": 4.191280364990234,
+ "step": 4648
+ },
+ {
+ "epoch": 64.57292576419214,
+ "grad_norm": 0.015284820459783077,
+ "learning_rate": 0.0006,
+ "loss": 4.35953426361084,
+ "step": 4649
+ },
+ {
+ "epoch": 64.58689956331878,
+ "grad_norm": 0.014855682849884033,
+ "learning_rate": 0.0006,
+ "loss": 4.308401107788086,
+ "step": 4650
+ },
+ {
+ "epoch": 64.60087336244541,
+ "grad_norm": 0.017553111538290977,
+ "learning_rate": 0.0006,
+ "loss": 4.279720306396484,
+ "step": 4651
+ },
+ {
+ "epoch": 64.61484716157206,
+ "grad_norm": 0.019506998360157013,
+ "learning_rate": 0.0006,
+ "loss": 4.319524765014648,
+ "step": 4652
+ },
+ {
+ "epoch": 64.62882096069869,
+ "grad_norm": 0.020293425768613815,
+ "learning_rate": 0.0006,
+ "loss": 4.340523719787598,
+ "step": 4653
+ },
+ {
+ "epoch": 64.64279475982532,
+ "grad_norm": 0.01664556749165058,
+ "learning_rate": 0.0006,
+ "loss": 4.296785831451416,
+ "step": 4654
+ },
+ {
+ "epoch": 64.65676855895197,
+ "grad_norm": 0.012592585757374763,
+ "learning_rate": 0.0006,
+ "loss": 4.464595317840576,
+ "step": 4655
+ },
+ {
+ "epoch": 64.6707423580786,
+ "grad_norm": 0.01549589168280363,
+ "learning_rate": 0.0006,
+ "loss": 4.275053024291992,
+ "step": 4656
+ },
+ {
+ "epoch": 64.68471615720524,
+ "grad_norm": 0.01705172471702099,
+ "learning_rate": 0.0006,
+ "loss": 4.388431549072266,
+ "step": 4657
+ },
+ {
+ "epoch": 64.69868995633188,
+ "grad_norm": 0.015438582748174667,
+ "learning_rate": 0.0006,
+ "loss": 4.312070846557617,
+ "step": 4658
+ },
+ {
+ "epoch": 64.71266375545852,
+ "grad_norm": 0.014448083005845547,
+ "learning_rate": 0.0006,
+ "loss": 4.237599849700928,
+ "step": 4659
+ },
+ {
+ "epoch": 64.72663755458515,
+ "grad_norm": 0.015267071314156055,
+ "learning_rate": 0.0006,
+ "loss": 4.29893684387207,
+ "step": 4660
+ },
+ {
+ "epoch": 64.74061135371178,
+ "grad_norm": 0.016615847125649452,
+ "learning_rate": 0.0006,
+ "loss": 4.432273864746094,
+ "step": 4661
+ },
+ {
+ "epoch": 64.75458515283843,
+ "grad_norm": 0.016584362834692,
+ "learning_rate": 0.0006,
+ "loss": 4.202266216278076,
+ "step": 4662
+ },
+ {
+ "epoch": 64.76855895196506,
+ "grad_norm": 0.018313994631171227,
+ "learning_rate": 0.0006,
+ "loss": 4.367966651916504,
+ "step": 4663
+ },
+ {
+ "epoch": 64.78253275109171,
+ "grad_norm": 0.017946481704711914,
+ "learning_rate": 0.0006,
+ "loss": 4.205172538757324,
+ "step": 4664
+ },
+ {
+ "epoch": 64.79650655021834,
+ "grad_norm": 0.017744354903697968,
+ "learning_rate": 0.0006,
+ "loss": 4.295661926269531,
+ "step": 4665
+ },
+ {
+ "epoch": 64.81048034934499,
+ "grad_norm": 0.016103824600577354,
+ "learning_rate": 0.0006,
+ "loss": 4.276543617248535,
+ "step": 4666
+ },
+ {
+ "epoch": 64.82445414847162,
+ "grad_norm": 0.014590012840926647,
+ "learning_rate": 0.0006,
+ "loss": 4.414247035980225,
+ "step": 4667
+ },
+ {
+ "epoch": 64.83842794759825,
+ "grad_norm": 0.01782386749982834,
+ "learning_rate": 0.0006,
+ "loss": 4.203220367431641,
+ "step": 4668
+ },
+ {
+ "epoch": 64.8524017467249,
+ "grad_norm": 0.022122733294963837,
+ "learning_rate": 0.0006,
+ "loss": 4.38078498840332,
+ "step": 4669
+ },
+ {
+ "epoch": 64.86637554585153,
+ "grad_norm": 0.0250293780118227,
+ "learning_rate": 0.0006,
+ "loss": 4.3310136795043945,
+ "step": 4670
+ },
+ {
+ "epoch": 64.88034934497817,
+ "grad_norm": 0.022438062354922295,
+ "learning_rate": 0.0006,
+ "loss": 4.428631782531738,
+ "step": 4671
+ },
+ {
+ "epoch": 64.8943231441048,
+ "grad_norm": 0.018518192693591118,
+ "learning_rate": 0.0006,
+ "loss": 4.428250312805176,
+ "step": 4672
+ },
+ {
+ "epoch": 64.90829694323143,
+ "grad_norm": 0.018105225637555122,
+ "learning_rate": 0.0006,
+ "loss": 4.336209297180176,
+ "step": 4673
+ },
+ {
+ "epoch": 64.92227074235808,
+ "grad_norm": 0.017564399167895317,
+ "learning_rate": 0.0006,
+ "loss": 4.416011810302734,
+ "step": 4674
+ },
+ {
+ "epoch": 64.93624454148471,
+ "grad_norm": 0.018747031688690186,
+ "learning_rate": 0.0006,
+ "loss": 4.419938087463379,
+ "step": 4675
+ },
+ {
+ "epoch": 64.95021834061136,
+ "grad_norm": 0.018686484545469284,
+ "learning_rate": 0.0006,
+ "loss": 4.347185134887695,
+ "step": 4676
+ },
+ {
+ "epoch": 64.96419213973799,
+ "grad_norm": 0.019329898059368134,
+ "learning_rate": 0.0006,
+ "loss": 4.343749046325684,
+ "step": 4677
+ },
+ {
+ "epoch": 64.97816593886463,
+ "grad_norm": 0.018517782911658287,
+ "learning_rate": 0.0006,
+ "loss": 4.264604568481445,
+ "step": 4678
+ },
+ {
+ "epoch": 64.99213973799127,
+ "grad_norm": 0.016579125076532364,
+ "learning_rate": 0.0006,
+ "loss": 4.303423881530762,
+ "step": 4679
+ },
+ {
+ "epoch": 65.0,
+ "grad_norm": 0.01617567613720894,
+ "learning_rate": 0.0006,
+ "loss": 4.326556205749512,
+ "step": 4680
+ },
+ {
+ "epoch": 65.0,
+ "eval_loss": 4.655026912689209,
+ "eval_runtime": 57.1299,
+ "eval_samples_per_second": 42.745,
+ "eval_steps_per_second": 1.348,
+ "step": 4680
+ },
+ {
+ "epoch": 65.01397379912663,
+ "grad_norm": 0.01607249490916729,
+ "learning_rate": 0.0006,
+ "loss": 4.276607036590576,
+ "step": 4681
+ },
+ {
+ "epoch": 65.02794759825328,
+ "grad_norm": 0.016779497265815735,
+ "learning_rate": 0.0006,
+ "loss": 4.412464618682861,
+ "step": 4682
+ },
+ {
+ "epoch": 65.04192139737991,
+ "grad_norm": 0.018161555752158165,
+ "learning_rate": 0.0006,
+ "loss": 4.199329376220703,
+ "step": 4683
+ },
+ {
+ "epoch": 65.05589519650655,
+ "grad_norm": 0.017508309334516525,
+ "learning_rate": 0.0006,
+ "loss": 4.217567443847656,
+ "step": 4684
+ },
+ {
+ "epoch": 65.06986899563319,
+ "grad_norm": 0.015716979280114174,
+ "learning_rate": 0.0006,
+ "loss": 4.322023391723633,
+ "step": 4685
+ },
+ {
+ "epoch": 65.08384279475983,
+ "grad_norm": 0.015113635919988155,
+ "learning_rate": 0.0006,
+ "loss": 4.333515167236328,
+ "step": 4686
+ },
+ {
+ "epoch": 65.09781659388646,
+ "grad_norm": 0.01437560748308897,
+ "learning_rate": 0.0006,
+ "loss": 4.295609474182129,
+ "step": 4687
+ },
+ {
+ "epoch": 65.1117903930131,
+ "grad_norm": 0.013074862770736217,
+ "learning_rate": 0.0006,
+ "loss": 4.314120769500732,
+ "step": 4688
+ },
+ {
+ "epoch": 65.12576419213974,
+ "grad_norm": 0.012908284552395344,
+ "learning_rate": 0.0006,
+ "loss": 4.216890811920166,
+ "step": 4689
+ },
+ {
+ "epoch": 65.13973799126637,
+ "grad_norm": 0.013001754879951477,
+ "learning_rate": 0.0006,
+ "loss": 4.287717342376709,
+ "step": 4690
+ },
+ {
+ "epoch": 65.15371179039302,
+ "grad_norm": 0.013694263994693756,
+ "learning_rate": 0.0006,
+ "loss": 4.333837985992432,
+ "step": 4691
+ },
+ {
+ "epoch": 65.16768558951965,
+ "grad_norm": 0.014461833983659744,
+ "learning_rate": 0.0006,
+ "loss": 4.323545455932617,
+ "step": 4692
+ },
+ {
+ "epoch": 65.18165938864628,
+ "grad_norm": 0.013670675456523895,
+ "learning_rate": 0.0006,
+ "loss": 4.3313493728637695,
+ "step": 4693
+ },
+ {
+ "epoch": 65.19563318777293,
+ "grad_norm": 0.013094350695610046,
+ "learning_rate": 0.0006,
+ "loss": 4.23858118057251,
+ "step": 4694
+ },
+ {
+ "epoch": 65.20960698689956,
+ "grad_norm": 0.013996944762766361,
+ "learning_rate": 0.0006,
+ "loss": 4.428705215454102,
+ "step": 4695
+ },
+ {
+ "epoch": 65.2235807860262,
+ "grad_norm": 0.013295048847794533,
+ "learning_rate": 0.0006,
+ "loss": 4.238950252532959,
+ "step": 4696
+ },
+ {
+ "epoch": 65.23755458515284,
+ "grad_norm": 0.013380616903305054,
+ "learning_rate": 0.0006,
+ "loss": 4.273185729980469,
+ "step": 4697
+ },
+ {
+ "epoch": 65.25152838427948,
+ "grad_norm": 0.015062890015542507,
+ "learning_rate": 0.0006,
+ "loss": 4.29472017288208,
+ "step": 4698
+ },
+ {
+ "epoch": 65.26550218340611,
+ "grad_norm": 0.017317263409495354,
+ "learning_rate": 0.0006,
+ "loss": 4.337954998016357,
+ "step": 4699
+ },
+ {
+ "epoch": 65.27947598253274,
+ "grad_norm": 0.017698992043733597,
+ "learning_rate": 0.0006,
+ "loss": 4.336437225341797,
+ "step": 4700
+ },
+ {
+ "epoch": 65.29344978165939,
+ "grad_norm": 0.01628575474023819,
+ "learning_rate": 0.0006,
+ "loss": 4.302372455596924,
+ "step": 4701
+ },
+ {
+ "epoch": 65.30742358078602,
+ "grad_norm": 0.014531854540109634,
+ "learning_rate": 0.0006,
+ "loss": 4.256450653076172,
+ "step": 4702
+ },
+ {
+ "epoch": 65.32139737991267,
+ "grad_norm": 0.014549105428159237,
+ "learning_rate": 0.0006,
+ "loss": 4.302240371704102,
+ "step": 4703
+ },
+ {
+ "epoch": 65.3353711790393,
+ "grad_norm": 0.014881663955748081,
+ "learning_rate": 0.0006,
+ "loss": 4.3316121101379395,
+ "step": 4704
+ },
+ {
+ "epoch": 65.34934497816595,
+ "grad_norm": 0.014513437636196613,
+ "learning_rate": 0.0006,
+ "loss": 4.441212177276611,
+ "step": 4705
+ },
+ {
+ "epoch": 65.36331877729258,
+ "grad_norm": 0.014122087508440018,
+ "learning_rate": 0.0006,
+ "loss": 4.343296527862549,
+ "step": 4706
+ },
+ {
+ "epoch": 65.37729257641921,
+ "grad_norm": 0.014817348681390285,
+ "learning_rate": 0.0006,
+ "loss": 4.195674896240234,
+ "step": 4707
+ },
+ {
+ "epoch": 65.39126637554585,
+ "grad_norm": 0.01787056028842926,
+ "learning_rate": 0.0006,
+ "loss": 4.3507914543151855,
+ "step": 4708
+ },
+ {
+ "epoch": 65.40524017467249,
+ "grad_norm": 0.015295136719942093,
+ "learning_rate": 0.0006,
+ "loss": 4.234372138977051,
+ "step": 4709
+ },
+ {
+ "epoch": 65.41921397379913,
+ "grad_norm": 0.014324234798550606,
+ "learning_rate": 0.0006,
+ "loss": 4.2907867431640625,
+ "step": 4710
+ },
+ {
+ "epoch": 65.43318777292576,
+ "grad_norm": 0.016090670600533485,
+ "learning_rate": 0.0006,
+ "loss": 4.177244186401367,
+ "step": 4711
+ },
+ {
+ "epoch": 65.44716157205241,
+ "grad_norm": 0.0174604132771492,
+ "learning_rate": 0.0006,
+ "loss": 4.335577964782715,
+ "step": 4712
+ },
+ {
+ "epoch": 65.46113537117904,
+ "grad_norm": 0.016731899231672287,
+ "learning_rate": 0.0006,
+ "loss": 4.30197811126709,
+ "step": 4713
+ },
+ {
+ "epoch": 65.47510917030567,
+ "grad_norm": 0.017423735931515694,
+ "learning_rate": 0.0006,
+ "loss": 4.3528852462768555,
+ "step": 4714
+ },
+ {
+ "epoch": 65.48908296943232,
+ "grad_norm": 0.01780976541340351,
+ "learning_rate": 0.0006,
+ "loss": 4.185036659240723,
+ "step": 4715
+ },
+ {
+ "epoch": 65.50305676855895,
+ "grad_norm": 0.019977200776338577,
+ "learning_rate": 0.0006,
+ "loss": 4.328766345977783,
+ "step": 4716
+ },
+ {
+ "epoch": 65.5170305676856,
+ "grad_norm": 0.020284287631511688,
+ "learning_rate": 0.0006,
+ "loss": 4.285494804382324,
+ "step": 4717
+ },
+ {
+ "epoch": 65.53100436681223,
+ "grad_norm": 0.019168002530932426,
+ "learning_rate": 0.0006,
+ "loss": 4.328560829162598,
+ "step": 4718
+ },
+ {
+ "epoch": 65.54497816593886,
+ "grad_norm": 0.019217776134610176,
+ "learning_rate": 0.0006,
+ "loss": 4.189998626708984,
+ "step": 4719
+ },
+ {
+ "epoch": 65.5589519650655,
+ "grad_norm": 0.018960488960146904,
+ "learning_rate": 0.0006,
+ "loss": 4.4995012283325195,
+ "step": 4720
+ },
+ {
+ "epoch": 65.57292576419214,
+ "grad_norm": 0.01941288262605667,
+ "learning_rate": 0.0006,
+ "loss": 4.26570987701416,
+ "step": 4721
+ },
+ {
+ "epoch": 65.58689956331878,
+ "grad_norm": 0.020263448357582092,
+ "learning_rate": 0.0006,
+ "loss": 4.411645889282227,
+ "step": 4722
+ },
+ {
+ "epoch": 65.60087336244541,
+ "grad_norm": 0.02144799567759037,
+ "learning_rate": 0.0006,
+ "loss": 4.163675308227539,
+ "step": 4723
+ },
+ {
+ "epoch": 65.61484716157206,
+ "grad_norm": 0.019790446385741234,
+ "learning_rate": 0.0006,
+ "loss": 4.2361063957214355,
+ "step": 4724
+ },
+ {
+ "epoch": 65.62882096069869,
+ "grad_norm": 0.01829805225133896,
+ "learning_rate": 0.0006,
+ "loss": 4.449810028076172,
+ "step": 4725
+ },
+ {
+ "epoch": 65.64279475982532,
+ "grad_norm": 0.01571105420589447,
+ "learning_rate": 0.0006,
+ "loss": 4.281406879425049,
+ "step": 4726
+ },
+ {
+ "epoch": 65.65676855895197,
+ "grad_norm": 0.016477147117257118,
+ "learning_rate": 0.0006,
+ "loss": 4.293408393859863,
+ "step": 4727
+ },
+ {
+ "epoch": 65.6707423580786,
+ "grad_norm": 0.016766032204031944,
+ "learning_rate": 0.0006,
+ "loss": 4.378252983093262,
+ "step": 4728
+ },
+ {
+ "epoch": 65.68471615720524,
+ "grad_norm": 0.0159933939576149,
+ "learning_rate": 0.0006,
+ "loss": 4.3065667152404785,
+ "step": 4729
+ },
+ {
+ "epoch": 65.69868995633188,
+ "grad_norm": 0.01658209227025509,
+ "learning_rate": 0.0006,
+ "loss": 4.174026966094971,
+ "step": 4730
+ },
+ {
+ "epoch": 65.71266375545852,
+ "grad_norm": 0.015081380493938923,
+ "learning_rate": 0.0006,
+ "loss": 4.360222339630127,
+ "step": 4731
+ },
+ {
+ "epoch": 65.72663755458515,
+ "grad_norm": 0.015110890381038189,
+ "learning_rate": 0.0006,
+ "loss": 4.218810558319092,
+ "step": 4732
+ },
+ {
+ "epoch": 65.74061135371178,
+ "grad_norm": 0.015300953760743141,
+ "learning_rate": 0.0006,
+ "loss": 4.331979274749756,
+ "step": 4733
+ },
+ {
+ "epoch": 65.75458515283843,
+ "grad_norm": 0.01715504191815853,
+ "learning_rate": 0.0006,
+ "loss": 4.349112033843994,
+ "step": 4734
+ },
+ {
+ "epoch": 65.76855895196506,
+ "grad_norm": 0.015870366245508194,
+ "learning_rate": 0.0006,
+ "loss": 4.329850196838379,
+ "step": 4735
+ },
+ {
+ "epoch": 65.78253275109171,
+ "grad_norm": 0.01541937980800867,
+ "learning_rate": 0.0006,
+ "loss": 4.3400163650512695,
+ "step": 4736
+ },
+ {
+ "epoch": 65.79650655021834,
+ "grad_norm": 0.01738942787051201,
+ "learning_rate": 0.0006,
+ "loss": 4.220798492431641,
+ "step": 4737
+ },
+ {
+ "epoch": 65.81048034934499,
+ "grad_norm": 0.01685990020632744,
+ "learning_rate": 0.0006,
+ "loss": 4.177238464355469,
+ "step": 4738
+ },
+ {
+ "epoch": 65.82445414847162,
+ "grad_norm": 0.016183465719223022,
+ "learning_rate": 0.0006,
+ "loss": 4.327216625213623,
+ "step": 4739
+ },
+ {
+ "epoch": 65.83842794759825,
+ "grad_norm": 0.01677694357931614,
+ "learning_rate": 0.0006,
+ "loss": 4.27766227722168,
+ "step": 4740
+ },
+ {
+ "epoch": 65.8524017467249,
+ "grad_norm": 0.01645066775381565,
+ "learning_rate": 0.0006,
+ "loss": 4.288347244262695,
+ "step": 4741
+ },
+ {
+ "epoch": 65.86637554585153,
+ "grad_norm": 0.018034635111689568,
+ "learning_rate": 0.0006,
+ "loss": 4.292458534240723,
+ "step": 4742
+ },
+ {
+ "epoch": 65.88034934497817,
+ "grad_norm": 0.017973346635699272,
+ "learning_rate": 0.0006,
+ "loss": 4.3117265701293945,
+ "step": 4743
+ },
+ {
+ "epoch": 65.8943231441048,
+ "grad_norm": 0.016605669632554054,
+ "learning_rate": 0.0006,
+ "loss": 4.280615329742432,
+ "step": 4744
+ },
+ {
+ "epoch": 65.90829694323143,
+ "grad_norm": 0.016298074275255203,
+ "learning_rate": 0.0006,
+ "loss": 4.301933288574219,
+ "step": 4745
+ },
+ {
+ "epoch": 65.92227074235808,
+ "grad_norm": 0.015383216552436352,
+ "learning_rate": 0.0006,
+ "loss": 4.352383136749268,
+ "step": 4746
+ },
+ {
+ "epoch": 65.93624454148471,
+ "grad_norm": 0.016222182661294937,
+ "learning_rate": 0.0006,
+ "loss": 4.305725574493408,
+ "step": 4747
+ },
+ {
+ "epoch": 65.95021834061136,
+ "grad_norm": 0.01691329851746559,
+ "learning_rate": 0.0006,
+ "loss": 4.29502010345459,
+ "step": 4748
+ },
+ {
+ "epoch": 65.96419213973799,
+ "grad_norm": 0.017637677490711212,
+ "learning_rate": 0.0006,
+ "loss": 4.3133039474487305,
+ "step": 4749
+ },
+ {
+ "epoch": 65.97816593886463,
+ "grad_norm": 0.021362047642469406,
+ "learning_rate": 0.0006,
+ "loss": 4.458834648132324,
+ "step": 4750
+ },
+ {
+ "epoch": 65.99213973799127,
+ "grad_norm": 0.023553011938929558,
+ "learning_rate": 0.0006,
+ "loss": 4.298727989196777,
+ "step": 4751
+ },
+ {
+ "epoch": 66.0,
+ "grad_norm": 0.02154950052499771,
+ "learning_rate": 0.0006,
+ "loss": 4.403392791748047,
+ "step": 4752
+ },
+ {
+ "epoch": 66.0,
+ "eval_loss": 4.724340438842773,
+ "eval_runtime": 56.9555,
+ "eval_samples_per_second": 42.876,
+ "eval_steps_per_second": 1.352,
+ "step": 4752
+ },
+ {
+ "epoch": 66.01397379912663,
+ "grad_norm": 0.01916784606873989,
+ "learning_rate": 0.0006,
+ "loss": 4.22931432723999,
+ "step": 4753
+ },
+ {
+ "epoch": 66.02794759825328,
+ "grad_norm": 0.017492054030299187,
+ "learning_rate": 0.0006,
+ "loss": 4.271522521972656,
+ "step": 4754
+ },
+ {
+ "epoch": 66.04192139737991,
+ "grad_norm": 0.018408095464110374,
+ "learning_rate": 0.0006,
+ "loss": 4.337705135345459,
+ "step": 4755
+ },
+ {
+ "epoch": 66.05589519650655,
+ "grad_norm": 0.019220108166337013,
+ "learning_rate": 0.0006,
+ "loss": 4.280340194702148,
+ "step": 4756
+ },
+ {
+ "epoch": 66.06986899563319,
+ "grad_norm": 0.01773735322058201,
+ "learning_rate": 0.0006,
+ "loss": 4.390201568603516,
+ "step": 4757
+ },
+ {
+ "epoch": 66.08384279475983,
+ "grad_norm": 0.01679125428199768,
+ "learning_rate": 0.0006,
+ "loss": 4.299282073974609,
+ "step": 4758
+ },
+ {
+ "epoch": 66.09781659388646,
+ "grad_norm": 0.01479059923440218,
+ "learning_rate": 0.0006,
+ "loss": 4.472630500793457,
+ "step": 4759
+ },
+ {
+ "epoch": 66.1117903930131,
+ "grad_norm": 0.015130104497075081,
+ "learning_rate": 0.0006,
+ "loss": 4.301194190979004,
+ "step": 4760
+ },
+ {
+ "epoch": 66.12576419213974,
+ "grad_norm": 0.0158979594707489,
+ "learning_rate": 0.0006,
+ "loss": 4.375751972198486,
+ "step": 4761
+ },
+ {
+ "epoch": 66.13973799126637,
+ "grad_norm": 0.014432408846914768,
+ "learning_rate": 0.0006,
+ "loss": 4.279788017272949,
+ "step": 4762
+ },
+ {
+ "epoch": 66.15371179039302,
+ "grad_norm": 0.013026784174144268,
+ "learning_rate": 0.0006,
+ "loss": 4.255035400390625,
+ "step": 4763
+ },
+ {
+ "epoch": 66.16768558951965,
+ "grad_norm": 0.014422653242945671,
+ "learning_rate": 0.0006,
+ "loss": 4.261832237243652,
+ "step": 4764
+ },
+ {
+ "epoch": 66.18165938864628,
+ "grad_norm": 0.014613019302487373,
+ "learning_rate": 0.0006,
+ "loss": 4.382184028625488,
+ "step": 4765
+ },
+ {
+ "epoch": 66.19563318777293,
+ "grad_norm": 0.014989894814789295,
+ "learning_rate": 0.0006,
+ "loss": 4.40528678894043,
+ "step": 4766
+ },
+ {
+ "epoch": 66.20960698689956,
+ "grad_norm": 0.018190350383520126,
+ "learning_rate": 0.0006,
+ "loss": 4.302414894104004,
+ "step": 4767
+ },
+ {
+ "epoch": 66.2235807860262,
+ "grad_norm": 0.020887890830636024,
+ "learning_rate": 0.0006,
+ "loss": 4.289920806884766,
+ "step": 4768
+ },
+ {
+ "epoch": 66.23755458515284,
+ "grad_norm": 0.02376876398921013,
+ "learning_rate": 0.0006,
+ "loss": 4.297855377197266,
+ "step": 4769
+ },
+ {
+ "epoch": 66.25152838427948,
+ "grad_norm": 0.02152269519865513,
+ "learning_rate": 0.0006,
+ "loss": 4.1972832679748535,
+ "step": 4770
+ },
+ {
+ "epoch": 66.26550218340611,
+ "grad_norm": 0.015572802163660526,
+ "learning_rate": 0.0006,
+ "loss": 4.307487487792969,
+ "step": 4771
+ },
+ {
+ "epoch": 66.27947598253274,
+ "grad_norm": 0.015228749252855778,
+ "learning_rate": 0.0006,
+ "loss": 4.103250503540039,
+ "step": 4772
+ },
+ {
+ "epoch": 66.29344978165939,
+ "grad_norm": 0.016925504431128502,
+ "learning_rate": 0.0006,
+ "loss": 4.247461318969727,
+ "step": 4773
+ },
+ {
+ "epoch": 66.30742358078602,
+ "grad_norm": 0.01613711006939411,
+ "learning_rate": 0.0006,
+ "loss": 4.4699273109436035,
+ "step": 4774
+ },
+ {
+ "epoch": 66.32139737991267,
+ "grad_norm": 0.018332140520215034,
+ "learning_rate": 0.0006,
+ "loss": 4.22969388961792,
+ "step": 4775
+ },
+ {
+ "epoch": 66.3353711790393,
+ "grad_norm": 0.019575074315071106,
+ "learning_rate": 0.0006,
+ "loss": 4.29397439956665,
+ "step": 4776
+ },
+ {
+ "epoch": 66.34934497816595,
+ "grad_norm": 0.018177419900894165,
+ "learning_rate": 0.0006,
+ "loss": 4.221899509429932,
+ "step": 4777
+ },
+ {
+ "epoch": 66.36331877729258,
+ "grad_norm": 0.014292889274656773,
+ "learning_rate": 0.0006,
+ "loss": 4.325457572937012,
+ "step": 4778
+ },
+ {
+ "epoch": 66.37729257641921,
+ "grad_norm": 0.015098796226084232,
+ "learning_rate": 0.0006,
+ "loss": 4.364837169647217,
+ "step": 4779
+ },
+ {
+ "epoch": 66.39126637554585,
+ "grad_norm": 0.013992635533213615,
+ "learning_rate": 0.0006,
+ "loss": 4.299082279205322,
+ "step": 4780
+ },
+ {
+ "epoch": 66.40524017467249,
+ "grad_norm": 0.01504511572420597,
+ "learning_rate": 0.0006,
+ "loss": 4.230877876281738,
+ "step": 4781
+ },
+ {
+ "epoch": 66.41921397379913,
+ "grad_norm": 0.014875639230012894,
+ "learning_rate": 0.0006,
+ "loss": 4.368405818939209,
+ "step": 4782
+ },
+ {
+ "epoch": 66.43318777292576,
+ "grad_norm": 0.014306887984275818,
+ "learning_rate": 0.0006,
+ "loss": 4.389184474945068,
+ "step": 4783
+ },
+ {
+ "epoch": 66.44716157205241,
+ "grad_norm": 0.012959563173353672,
+ "learning_rate": 0.0006,
+ "loss": 4.37934684753418,
+ "step": 4784
+ },
+ {
+ "epoch": 66.46113537117904,
+ "grad_norm": 0.013206703588366508,
+ "learning_rate": 0.0006,
+ "loss": 4.442799091339111,
+ "step": 4785
+ },
+ {
+ "epoch": 66.47510917030567,
+ "grad_norm": 0.014848909340798855,
+ "learning_rate": 0.0006,
+ "loss": 4.203481674194336,
+ "step": 4786
+ },
+ {
+ "epoch": 66.48908296943232,
+ "grad_norm": 0.015637682750821114,
+ "learning_rate": 0.0006,
+ "loss": 4.27552604675293,
+ "step": 4787
+ },
+ {
+ "epoch": 66.50305676855895,
+ "grad_norm": 0.016017816960811615,
+ "learning_rate": 0.0006,
+ "loss": 4.296402931213379,
+ "step": 4788
+ },
+ {
+ "epoch": 66.5170305676856,
+ "grad_norm": 0.01577814109623432,
+ "learning_rate": 0.0006,
+ "loss": 4.301733016967773,
+ "step": 4789
+ },
+ {
+ "epoch": 66.53100436681223,
+ "grad_norm": 0.015293175354599953,
+ "learning_rate": 0.0006,
+ "loss": 4.369877815246582,
+ "step": 4790
+ },
+ {
+ "epoch": 66.54497816593886,
+ "grad_norm": 0.014262278564274311,
+ "learning_rate": 0.0006,
+ "loss": 4.3314738273620605,
+ "step": 4791
+ },
+ {
+ "epoch": 66.5589519650655,
+ "grad_norm": 0.015123482793569565,
+ "learning_rate": 0.0006,
+ "loss": 4.313591957092285,
+ "step": 4792
+ },
+ {
+ "epoch": 66.57292576419214,
+ "grad_norm": 0.015458766371011734,
+ "learning_rate": 0.0006,
+ "loss": 4.3045854568481445,
+ "step": 4793
+ },
+ {
+ "epoch": 66.58689956331878,
+ "grad_norm": 0.017217233777046204,
+ "learning_rate": 0.0006,
+ "loss": 4.269944190979004,
+ "step": 4794
+ },
+ {
+ "epoch": 66.60087336244541,
+ "grad_norm": 0.01962282694876194,
+ "learning_rate": 0.0006,
+ "loss": 4.271735191345215,
+ "step": 4795
+ },
+ {
+ "epoch": 66.61484716157206,
+ "grad_norm": 0.01917066052556038,
+ "learning_rate": 0.0006,
+ "loss": 4.288168430328369,
+ "step": 4796
+ },
+ {
+ "epoch": 66.62882096069869,
+ "grad_norm": 0.018211524933576584,
+ "learning_rate": 0.0006,
+ "loss": 4.280660629272461,
+ "step": 4797
+ },
+ {
+ "epoch": 66.64279475982532,
+ "grad_norm": 0.01936524175107479,
+ "learning_rate": 0.0006,
+ "loss": 4.3710551261901855,
+ "step": 4798
+ },
+ {
+ "epoch": 66.65676855895197,
+ "grad_norm": 0.01882351003587246,
+ "learning_rate": 0.0006,
+ "loss": 4.342750549316406,
+ "step": 4799
+ },
+ {
+ "epoch": 66.6707423580786,
+ "grad_norm": 0.017149122431874275,
+ "learning_rate": 0.0006,
+ "loss": 4.32134485244751,
+ "step": 4800
+ },
+ {
+ "epoch": 66.68471615720524,
+ "grad_norm": 0.01699327677488327,
+ "learning_rate": 0.0006,
+ "loss": 4.248893737792969,
+ "step": 4801
+ },
+ {
+ "epoch": 66.69868995633188,
+ "grad_norm": 0.016873255372047424,
+ "learning_rate": 0.0006,
+ "loss": 4.277350425720215,
+ "step": 4802
+ },
+ {
+ "epoch": 66.71266375545852,
+ "grad_norm": 0.016592837870121002,
+ "learning_rate": 0.0006,
+ "loss": 4.221665859222412,
+ "step": 4803
+ },
+ {
+ "epoch": 66.72663755458515,
+ "grad_norm": 0.015887131914496422,
+ "learning_rate": 0.0006,
+ "loss": 4.328040599822998,
+ "step": 4804
+ },
+ {
+ "epoch": 66.74061135371178,
+ "grad_norm": 0.014057286083698273,
+ "learning_rate": 0.0006,
+ "loss": 4.3294548988342285,
+ "step": 4805
+ },
+ {
+ "epoch": 66.75458515283843,
+ "grad_norm": 0.013135905377566814,
+ "learning_rate": 0.0006,
+ "loss": 4.254383563995361,
+ "step": 4806
+ },
+ {
+ "epoch": 66.76855895196506,
+ "grad_norm": 0.013632557354867458,
+ "learning_rate": 0.0006,
+ "loss": 4.342815399169922,
+ "step": 4807
+ },
+ {
+ "epoch": 66.78253275109171,
+ "grad_norm": 0.015512688085436821,
+ "learning_rate": 0.0006,
+ "loss": 4.297835350036621,
+ "step": 4808
+ },
+ {
+ "epoch": 66.79650655021834,
+ "grad_norm": 0.01621832884848118,
+ "learning_rate": 0.0006,
+ "loss": 4.314773082733154,
+ "step": 4809
+ },
+ {
+ "epoch": 66.81048034934499,
+ "grad_norm": 0.016196032986044884,
+ "learning_rate": 0.0006,
+ "loss": 4.29209566116333,
+ "step": 4810
+ },
+ {
+ "epoch": 66.82445414847162,
+ "grad_norm": 0.015463353134691715,
+ "learning_rate": 0.0006,
+ "loss": 4.262795448303223,
+ "step": 4811
+ },
+ {
+ "epoch": 66.83842794759825,
+ "grad_norm": 0.016143787652254105,
+ "learning_rate": 0.0006,
+ "loss": 4.274284362792969,
+ "step": 4812
+ },
+ {
+ "epoch": 66.8524017467249,
+ "grad_norm": 0.01610061153769493,
+ "learning_rate": 0.0006,
+ "loss": 4.337160587310791,
+ "step": 4813
+ },
+ {
+ "epoch": 66.86637554585153,
+ "grad_norm": 0.01582838036119938,
+ "learning_rate": 0.0006,
+ "loss": 4.192723274230957,
+ "step": 4814
+ },
+ {
+ "epoch": 66.88034934497817,
+ "grad_norm": 0.015081540681421757,
+ "learning_rate": 0.0006,
+ "loss": 4.370708465576172,
+ "step": 4815
+ },
+ {
+ "epoch": 66.8943231441048,
+ "grad_norm": 0.015929723158478737,
+ "learning_rate": 0.0006,
+ "loss": 4.340244293212891,
+ "step": 4816
+ },
+ {
+ "epoch": 66.90829694323143,
+ "grad_norm": 0.01784711331129074,
+ "learning_rate": 0.0006,
+ "loss": 4.260406970977783,
+ "step": 4817
+ },
+ {
+ "epoch": 66.92227074235808,
+ "grad_norm": 0.017434703186154366,
+ "learning_rate": 0.0006,
+ "loss": 4.286900043487549,
+ "step": 4818
+ },
+ {
+ "epoch": 66.93624454148471,
+ "grad_norm": 0.016262540593743324,
+ "learning_rate": 0.0006,
+ "loss": 4.292142391204834,
+ "step": 4819
+ },
+ {
+ "epoch": 66.95021834061136,
+ "grad_norm": 0.016122829169034958,
+ "learning_rate": 0.0006,
+ "loss": 4.291485786437988,
+ "step": 4820
+ },
+ {
+ "epoch": 66.96419213973799,
+ "grad_norm": 0.014361178502440453,
+ "learning_rate": 0.0006,
+ "loss": 4.275425910949707,
+ "step": 4821
+ },
+ {
+ "epoch": 66.97816593886463,
+ "grad_norm": 0.01373987551778555,
+ "learning_rate": 0.0006,
+ "loss": 4.223120212554932,
+ "step": 4822
+ },
+ {
+ "epoch": 66.99213973799127,
+ "grad_norm": 0.015687376260757446,
+ "learning_rate": 0.0006,
+ "loss": 4.355381011962891,
+ "step": 4823
+ },
+ {
+ "epoch": 67.0,
+ "grad_norm": 0.01998465321958065,
+ "learning_rate": 0.0006,
+ "loss": 4.264053821563721,
+ "step": 4824
+ },
+ {
+ "epoch": 67.0,
+ "eval_loss": 4.662398338317871,
+ "eval_runtime": 56.7928,
+ "eval_samples_per_second": 42.998,
+ "eval_steps_per_second": 1.356,
+ "step": 4824
+ },
+ {
+ "epoch": 67.01397379912663,
+ "grad_norm": 0.021736836060881615,
+ "learning_rate": 0.0006,
+ "loss": 4.243354797363281,
+ "step": 4825
+ },
+ {
+ "epoch": 67.02794759825328,
+ "grad_norm": 0.021476037800312042,
+ "learning_rate": 0.0006,
+ "loss": 4.218207836151123,
+ "step": 4826
+ },
+ {
+ "epoch": 67.04192139737991,
+ "grad_norm": 0.017188388854265213,
+ "learning_rate": 0.0006,
+ "loss": 4.18048095703125,
+ "step": 4827
+ },
+ {
+ "epoch": 67.05589519650655,
+ "grad_norm": 0.017293719574809074,
+ "learning_rate": 0.0006,
+ "loss": 4.239145278930664,
+ "step": 4828
+ },
+ {
+ "epoch": 67.06986899563319,
+ "grad_norm": 0.017383819445967674,
+ "learning_rate": 0.0006,
+ "loss": 4.114765167236328,
+ "step": 4829
+ },
+ {
+ "epoch": 67.08384279475983,
+ "grad_norm": 0.01651529036462307,
+ "learning_rate": 0.0006,
+ "loss": 4.351023197174072,
+ "step": 4830
+ },
+ {
+ "epoch": 67.09781659388646,
+ "grad_norm": 0.015893027186393738,
+ "learning_rate": 0.0006,
+ "loss": 4.261432647705078,
+ "step": 4831
+ },
+ {
+ "epoch": 67.1117903930131,
+ "grad_norm": 0.013381517492234707,
+ "learning_rate": 0.0006,
+ "loss": 4.172064781188965,
+ "step": 4832
+ },
+ {
+ "epoch": 67.12576419213974,
+ "grad_norm": 0.014642654918134212,
+ "learning_rate": 0.0006,
+ "loss": 4.356171131134033,
+ "step": 4833
+ },
+ {
+ "epoch": 67.13973799126637,
+ "grad_norm": 0.01425282284617424,
+ "learning_rate": 0.0006,
+ "loss": 4.303495407104492,
+ "step": 4834
+ },
+ {
+ "epoch": 67.15371179039302,
+ "grad_norm": 0.014574599452316761,
+ "learning_rate": 0.0006,
+ "loss": 4.2850847244262695,
+ "step": 4835
+ },
+ {
+ "epoch": 67.16768558951965,
+ "grad_norm": 0.012676282785832882,
+ "learning_rate": 0.0006,
+ "loss": 4.297845840454102,
+ "step": 4836
+ },
+ {
+ "epoch": 67.18165938864628,
+ "grad_norm": 0.013948136940598488,
+ "learning_rate": 0.0006,
+ "loss": 4.324530601501465,
+ "step": 4837
+ },
+ {
+ "epoch": 67.19563318777293,
+ "grad_norm": 0.014195773750543594,
+ "learning_rate": 0.0006,
+ "loss": 4.31236457824707,
+ "step": 4838
+ },
+ {
+ "epoch": 67.20960698689956,
+ "grad_norm": 0.013631382025778294,
+ "learning_rate": 0.0006,
+ "loss": 4.086362838745117,
+ "step": 4839
+ },
+ {
+ "epoch": 67.2235807860262,
+ "grad_norm": 0.012798587791621685,
+ "learning_rate": 0.0006,
+ "loss": 4.328471660614014,
+ "step": 4840
+ },
+ {
+ "epoch": 67.23755458515284,
+ "grad_norm": 0.014443321153521538,
+ "learning_rate": 0.0006,
+ "loss": 4.278019905090332,
+ "step": 4841
+ },
+ {
+ "epoch": 67.25152838427948,
+ "grad_norm": 0.015355420298874378,
+ "learning_rate": 0.0006,
+ "loss": 4.333611011505127,
+ "step": 4842
+ },
+ {
+ "epoch": 67.26550218340611,
+ "grad_norm": 0.016487296670675278,
+ "learning_rate": 0.0006,
+ "loss": 4.351336479187012,
+ "step": 4843
+ },
+ {
+ "epoch": 67.27947598253274,
+ "grad_norm": 0.017422541975975037,
+ "learning_rate": 0.0006,
+ "loss": 4.386096477508545,
+ "step": 4844
+ },
+ {
+ "epoch": 67.29344978165939,
+ "grad_norm": 0.015790918841958046,
+ "learning_rate": 0.0006,
+ "loss": 4.301377773284912,
+ "step": 4845
+ },
+ {
+ "epoch": 67.30742358078602,
+ "grad_norm": 0.016427364200353622,
+ "learning_rate": 0.0006,
+ "loss": 4.2838945388793945,
+ "step": 4846
+ },
+ {
+ "epoch": 67.32139737991267,
+ "grad_norm": 0.017050761729478836,
+ "learning_rate": 0.0006,
+ "loss": 4.282417297363281,
+ "step": 4847
+ },
+ {
+ "epoch": 67.3353711790393,
+ "grad_norm": 0.015006253495812416,
+ "learning_rate": 0.0006,
+ "loss": 4.352236270904541,
+ "step": 4848
+ },
+ {
+ "epoch": 67.34934497816595,
+ "grad_norm": 0.012996627949178219,
+ "learning_rate": 0.0006,
+ "loss": 4.309242248535156,
+ "step": 4849
+ },
+ {
+ "epoch": 67.36331877729258,
+ "grad_norm": 0.01367615070194006,
+ "learning_rate": 0.0006,
+ "loss": 4.211801528930664,
+ "step": 4850
+ },
+ {
+ "epoch": 67.37729257641921,
+ "grad_norm": 0.014775963500142097,
+ "learning_rate": 0.0006,
+ "loss": 4.360859394073486,
+ "step": 4851
+ },
+ {
+ "epoch": 67.39126637554585,
+ "grad_norm": 0.015854304656386375,
+ "learning_rate": 0.0006,
+ "loss": 4.211830139160156,
+ "step": 4852
+ },
+ {
+ "epoch": 67.40524017467249,
+ "grad_norm": 0.016888894140720367,
+ "learning_rate": 0.0006,
+ "loss": 4.216549873352051,
+ "step": 4853
+ },
+ {
+ "epoch": 67.41921397379913,
+ "grad_norm": 0.0170894768089056,
+ "learning_rate": 0.0006,
+ "loss": 4.263486385345459,
+ "step": 4854
+ },
+ {
+ "epoch": 67.43318777292576,
+ "grad_norm": 0.015829868614673615,
+ "learning_rate": 0.0006,
+ "loss": 4.360196113586426,
+ "step": 4855
+ },
+ {
+ "epoch": 67.44716157205241,
+ "grad_norm": 0.015220202505588531,
+ "learning_rate": 0.0006,
+ "loss": 4.243794918060303,
+ "step": 4856
+ },
+ {
+ "epoch": 67.46113537117904,
+ "grad_norm": 0.01413388829678297,
+ "learning_rate": 0.0006,
+ "loss": 4.286624908447266,
+ "step": 4857
+ },
+ {
+ "epoch": 67.47510917030567,
+ "grad_norm": 0.01426626369357109,
+ "learning_rate": 0.0006,
+ "loss": 4.257316589355469,
+ "step": 4858
+ },
+ {
+ "epoch": 67.48908296943232,
+ "grad_norm": 0.015957845374941826,
+ "learning_rate": 0.0006,
+ "loss": 4.300443649291992,
+ "step": 4859
+ },
+ {
+ "epoch": 67.50305676855895,
+ "grad_norm": 0.015499051660299301,
+ "learning_rate": 0.0006,
+ "loss": 4.278162002563477,
+ "step": 4860
+ },
+ {
+ "epoch": 67.5170305676856,
+ "grad_norm": 0.015145624987781048,
+ "learning_rate": 0.0006,
+ "loss": 4.267950534820557,
+ "step": 4861
+ },
+ {
+ "epoch": 67.53100436681223,
+ "grad_norm": 0.014610803686082363,
+ "learning_rate": 0.0006,
+ "loss": 4.247499942779541,
+ "step": 4862
+ },
+ {
+ "epoch": 67.54497816593886,
+ "grad_norm": 0.014860909432172775,
+ "learning_rate": 0.0006,
+ "loss": 4.263200759887695,
+ "step": 4863
+ },
+ {
+ "epoch": 67.5589519650655,
+ "grad_norm": 0.01718984916806221,
+ "learning_rate": 0.0006,
+ "loss": 4.290019989013672,
+ "step": 4864
+ },
+ {
+ "epoch": 67.57292576419214,
+ "grad_norm": 0.01558963768184185,
+ "learning_rate": 0.0006,
+ "loss": 4.223816871643066,
+ "step": 4865
+ },
+ {
+ "epoch": 67.58689956331878,
+ "grad_norm": 0.014927188865840435,
+ "learning_rate": 0.0006,
+ "loss": 4.363586902618408,
+ "step": 4866
+ },
+ {
+ "epoch": 67.60087336244541,
+ "grad_norm": 0.01573595404624939,
+ "learning_rate": 0.0006,
+ "loss": 4.254765510559082,
+ "step": 4867
+ },
+ {
+ "epoch": 67.61484716157206,
+ "grad_norm": 0.015618929639458656,
+ "learning_rate": 0.0006,
+ "loss": 4.25072717666626,
+ "step": 4868
+ },
+ {
+ "epoch": 67.62882096069869,
+ "grad_norm": 0.01605590060353279,
+ "learning_rate": 0.0006,
+ "loss": 4.25208854675293,
+ "step": 4869
+ },
+ {
+ "epoch": 67.64279475982532,
+ "grad_norm": 0.01804601214826107,
+ "learning_rate": 0.0006,
+ "loss": 4.351354122161865,
+ "step": 4870
+ },
+ {
+ "epoch": 67.65676855895197,
+ "grad_norm": 0.020595047622919083,
+ "learning_rate": 0.0006,
+ "loss": 4.193732738494873,
+ "step": 4871
+ },
+ {
+ "epoch": 67.6707423580786,
+ "grad_norm": 0.023037966340780258,
+ "learning_rate": 0.0006,
+ "loss": 4.27754020690918,
+ "step": 4872
+ },
+ {
+ "epoch": 67.68471615720524,
+ "grad_norm": 0.020680895075201988,
+ "learning_rate": 0.0006,
+ "loss": 4.314730167388916,
+ "step": 4873
+ },
+ {
+ "epoch": 67.69868995633188,
+ "grad_norm": 0.01887434534728527,
+ "learning_rate": 0.0006,
+ "loss": 4.252305030822754,
+ "step": 4874
+ },
+ {
+ "epoch": 67.71266375545852,
+ "grad_norm": 0.016333596780896187,
+ "learning_rate": 0.0006,
+ "loss": 4.146123886108398,
+ "step": 4875
+ },
+ {
+ "epoch": 67.72663755458515,
+ "grad_norm": 0.015688810497522354,
+ "learning_rate": 0.0006,
+ "loss": 4.2814154624938965,
+ "step": 4876
+ },
+ {
+ "epoch": 67.74061135371178,
+ "grad_norm": 0.016270726919174194,
+ "learning_rate": 0.0006,
+ "loss": 4.323200225830078,
+ "step": 4877
+ },
+ {
+ "epoch": 67.75458515283843,
+ "grad_norm": 0.016824988648295403,
+ "learning_rate": 0.0006,
+ "loss": 4.269501686096191,
+ "step": 4878
+ },
+ {
+ "epoch": 67.76855895196506,
+ "grad_norm": 0.01730562001466751,
+ "learning_rate": 0.0006,
+ "loss": 4.187993049621582,
+ "step": 4879
+ },
+ {
+ "epoch": 67.78253275109171,
+ "grad_norm": 0.01546655036509037,
+ "learning_rate": 0.0006,
+ "loss": 4.340415000915527,
+ "step": 4880
+ },
+ {
+ "epoch": 67.79650655021834,
+ "grad_norm": 0.014173958450555801,
+ "learning_rate": 0.0006,
+ "loss": 4.217419624328613,
+ "step": 4881
+ },
+ {
+ "epoch": 67.81048034934499,
+ "grad_norm": 0.016810474917292595,
+ "learning_rate": 0.0006,
+ "loss": 4.352302074432373,
+ "step": 4882
+ },
+ {
+ "epoch": 67.82445414847162,
+ "grad_norm": 0.0175373163074255,
+ "learning_rate": 0.0006,
+ "loss": 4.20377254486084,
+ "step": 4883
+ },
+ {
+ "epoch": 67.83842794759825,
+ "grad_norm": 0.015734676271677017,
+ "learning_rate": 0.0006,
+ "loss": 4.147495746612549,
+ "step": 4884
+ },
+ {
+ "epoch": 67.8524017467249,
+ "grad_norm": 0.015091817826032639,
+ "learning_rate": 0.0006,
+ "loss": 4.272252082824707,
+ "step": 4885
+ },
+ {
+ "epoch": 67.86637554585153,
+ "grad_norm": 0.015000219456851482,
+ "learning_rate": 0.0006,
+ "loss": 4.302773952484131,
+ "step": 4886
+ },
+ {
+ "epoch": 67.88034934497817,
+ "grad_norm": 0.013827184215188026,
+ "learning_rate": 0.0006,
+ "loss": 4.337242126464844,
+ "step": 4887
+ },
+ {
+ "epoch": 67.8943231441048,
+ "grad_norm": 0.015308346599340439,
+ "learning_rate": 0.0006,
+ "loss": 4.140895366668701,
+ "step": 4888
+ },
+ {
+ "epoch": 67.90829694323143,
+ "grad_norm": 0.015356178395450115,
+ "learning_rate": 0.0006,
+ "loss": 4.312541484832764,
+ "step": 4889
+ },
+ {
+ "epoch": 67.92227074235808,
+ "grad_norm": 0.0137353939935565,
+ "learning_rate": 0.0006,
+ "loss": 4.173184394836426,
+ "step": 4890
+ },
+ {
+ "epoch": 67.93624454148471,
+ "grad_norm": 0.01332368515431881,
+ "learning_rate": 0.0006,
+ "loss": 4.191199779510498,
+ "step": 4891
+ },
+ {
+ "epoch": 67.95021834061136,
+ "grad_norm": 0.01488130260258913,
+ "learning_rate": 0.0006,
+ "loss": 4.331752300262451,
+ "step": 4892
+ },
+ {
+ "epoch": 67.96419213973799,
+ "grad_norm": 0.015449733473360538,
+ "learning_rate": 0.0006,
+ "loss": 4.2385454177856445,
+ "step": 4893
+ },
+ {
+ "epoch": 67.97816593886463,
+ "grad_norm": 0.01556103490293026,
+ "learning_rate": 0.0006,
+ "loss": 4.262353897094727,
+ "step": 4894
+ },
+ {
+ "epoch": 67.99213973799127,
+ "grad_norm": 0.014613849110901356,
+ "learning_rate": 0.0006,
+ "loss": 4.223945617675781,
+ "step": 4895
+ },
+ {
+ "epoch": 68.0,
+ "grad_norm": 0.015323462896049023,
+ "learning_rate": 0.0006,
+ "loss": 4.317168712615967,
+ "step": 4896
+ },
+ {
+ "epoch": 68.0,
+ "eval_loss": 4.671603679656982,
+ "eval_runtime": 56.7317,
+ "eval_samples_per_second": 43.045,
+ "eval_steps_per_second": 1.357,
+ "step": 4896
+ },
+ {
+ "epoch": 68.01397379912663,
+ "grad_norm": 0.01910698413848877,
+ "learning_rate": 0.0006,
+ "loss": 4.206167221069336,
+ "step": 4897
+ },
+ {
+ "epoch": 68.02794759825328,
+ "grad_norm": 0.025112485513091087,
+ "learning_rate": 0.0006,
+ "loss": 4.384706974029541,
+ "step": 4898
+ },
+ {
+ "epoch": 68.04192139737991,
+ "grad_norm": 0.02618720941245556,
+ "learning_rate": 0.0006,
+ "loss": 4.321046829223633,
+ "step": 4899
+ },
+ {
+ "epoch": 68.05589519650655,
+ "grad_norm": 0.020853538066148758,
+ "learning_rate": 0.0006,
+ "loss": 4.217638969421387,
+ "step": 4900
+ },
+ {
+ "epoch": 68.06986899563319,
+ "grad_norm": 0.020025717094540596,
+ "learning_rate": 0.0006,
+ "loss": 4.306824684143066,
+ "step": 4901
+ },
+ {
+ "epoch": 68.08384279475983,
+ "grad_norm": 0.023868650197982788,
+ "learning_rate": 0.0006,
+ "loss": 4.327455997467041,
+ "step": 4902
+ },
+ {
+ "epoch": 68.09781659388646,
+ "grad_norm": 0.022279081866145134,
+ "learning_rate": 0.0006,
+ "loss": 4.400699615478516,
+ "step": 4903
+ },
+ {
+ "epoch": 68.1117903930131,
+ "grad_norm": 0.018965687602758408,
+ "learning_rate": 0.0006,
+ "loss": 4.19097900390625,
+ "step": 4904
+ },
+ {
+ "epoch": 68.12576419213974,
+ "grad_norm": 0.01936378888785839,
+ "learning_rate": 0.0006,
+ "loss": 4.206711769104004,
+ "step": 4905
+ },
+ {
+ "epoch": 68.13973799126637,
+ "grad_norm": 0.01811872608959675,
+ "learning_rate": 0.0006,
+ "loss": 4.204622268676758,
+ "step": 4906
+ },
+ {
+ "epoch": 68.15371179039302,
+ "grad_norm": 0.018530402332544327,
+ "learning_rate": 0.0006,
+ "loss": 4.289602279663086,
+ "step": 4907
+ },
+ {
+ "epoch": 68.16768558951965,
+ "grad_norm": 0.01755962148308754,
+ "learning_rate": 0.0006,
+ "loss": 4.317145347595215,
+ "step": 4908
+ },
+ {
+ "epoch": 68.18165938864628,
+ "grad_norm": 0.020413702353835106,
+ "learning_rate": 0.0006,
+ "loss": 4.2882843017578125,
+ "step": 4909
+ },
+ {
+ "epoch": 68.19563318777293,
+ "grad_norm": 0.020064299926161766,
+ "learning_rate": 0.0006,
+ "loss": 4.248260021209717,
+ "step": 4910
+ },
+ {
+ "epoch": 68.20960698689956,
+ "grad_norm": 0.017415888607501984,
+ "learning_rate": 0.0006,
+ "loss": 4.370295524597168,
+ "step": 4911
+ },
+ {
+ "epoch": 68.2235807860262,
+ "grad_norm": 0.015005858615040779,
+ "learning_rate": 0.0006,
+ "loss": 4.319567680358887,
+ "step": 4912
+ },
+ {
+ "epoch": 68.23755458515284,
+ "grad_norm": 0.016349777579307556,
+ "learning_rate": 0.0006,
+ "loss": 4.301776885986328,
+ "step": 4913
+ },
+ {
+ "epoch": 68.25152838427948,
+ "grad_norm": 0.015779560431838036,
+ "learning_rate": 0.0006,
+ "loss": 4.271171569824219,
+ "step": 4914
+ },
+ {
+ "epoch": 68.26550218340611,
+ "grad_norm": 0.013237644918262959,
+ "learning_rate": 0.0006,
+ "loss": 4.324908256530762,
+ "step": 4915
+ },
+ {
+ "epoch": 68.27947598253274,
+ "grad_norm": 0.014076225459575653,
+ "learning_rate": 0.0006,
+ "loss": 4.177175998687744,
+ "step": 4916
+ },
+ {
+ "epoch": 68.29344978165939,
+ "grad_norm": 0.013906744308769703,
+ "learning_rate": 0.0006,
+ "loss": 4.11588191986084,
+ "step": 4917
+ },
+ {
+ "epoch": 68.30742358078602,
+ "grad_norm": 0.013939064927399158,
+ "learning_rate": 0.0006,
+ "loss": 4.183963775634766,
+ "step": 4918
+ },
+ {
+ "epoch": 68.32139737991267,
+ "grad_norm": 0.014572170563042164,
+ "learning_rate": 0.0006,
+ "loss": 4.19586181640625,
+ "step": 4919
+ },
+ {
+ "epoch": 68.3353711790393,
+ "grad_norm": 0.014085350558161736,
+ "learning_rate": 0.0006,
+ "loss": 4.170892238616943,
+ "step": 4920
+ },
+ {
+ "epoch": 68.34934497816595,
+ "grad_norm": 0.014357863925397396,
+ "learning_rate": 0.0006,
+ "loss": 4.321155548095703,
+ "step": 4921
+ },
+ {
+ "epoch": 68.36331877729258,
+ "grad_norm": 0.01461675763130188,
+ "learning_rate": 0.0006,
+ "loss": 4.293334007263184,
+ "step": 4922
+ },
+ {
+ "epoch": 68.37729257641921,
+ "grad_norm": 0.014828486368060112,
+ "learning_rate": 0.0006,
+ "loss": 4.1439714431762695,
+ "step": 4923
+ },
+ {
+ "epoch": 68.39126637554585,
+ "grad_norm": 0.013730363920331001,
+ "learning_rate": 0.0006,
+ "loss": 4.287973880767822,
+ "step": 4924
+ },
+ {
+ "epoch": 68.40524017467249,
+ "grad_norm": 0.013852203264832497,
+ "learning_rate": 0.0006,
+ "loss": 4.248824119567871,
+ "step": 4925
+ },
+ {
+ "epoch": 68.41921397379913,
+ "grad_norm": 0.014005862176418304,
+ "learning_rate": 0.0006,
+ "loss": 4.293768882751465,
+ "step": 4926
+ },
+ {
+ "epoch": 68.43318777292576,
+ "grad_norm": 0.014435851015150547,
+ "learning_rate": 0.0006,
+ "loss": 4.204813003540039,
+ "step": 4927
+ },
+ {
+ "epoch": 68.44716157205241,
+ "grad_norm": 0.015305934473872185,
+ "learning_rate": 0.0006,
+ "loss": 4.329364776611328,
+ "step": 4928
+ },
+ {
+ "epoch": 68.46113537117904,
+ "grad_norm": 0.014356814324855804,
+ "learning_rate": 0.0006,
+ "loss": 4.264206409454346,
+ "step": 4929
+ },
+ {
+ "epoch": 68.47510917030567,
+ "grad_norm": 0.01432347297668457,
+ "learning_rate": 0.0006,
+ "loss": 4.29440450668335,
+ "step": 4930
+ },
+ {
+ "epoch": 68.48908296943232,
+ "grad_norm": 0.012672432698309422,
+ "learning_rate": 0.0006,
+ "loss": 4.258024215698242,
+ "step": 4931
+ },
+ {
+ "epoch": 68.50305676855895,
+ "grad_norm": 0.014209977351129055,
+ "learning_rate": 0.0006,
+ "loss": 4.216668128967285,
+ "step": 4932
+ },
+ {
+ "epoch": 68.5170305676856,
+ "grad_norm": 0.014606663957238197,
+ "learning_rate": 0.0006,
+ "loss": 4.294184684753418,
+ "step": 4933
+ },
+ {
+ "epoch": 68.53100436681223,
+ "grad_norm": 0.014986990951001644,
+ "learning_rate": 0.0006,
+ "loss": 4.382591247558594,
+ "step": 4934
+ },
+ {
+ "epoch": 68.54497816593886,
+ "grad_norm": 0.013998783193528652,
+ "learning_rate": 0.0006,
+ "loss": 4.195143222808838,
+ "step": 4935
+ },
+ {
+ "epoch": 68.5589519650655,
+ "grad_norm": 0.015011102892458439,
+ "learning_rate": 0.0006,
+ "loss": 4.313785076141357,
+ "step": 4936
+ },
+ {
+ "epoch": 68.57292576419214,
+ "grad_norm": 0.01641209051012993,
+ "learning_rate": 0.0006,
+ "loss": 4.322803020477295,
+ "step": 4937
+ },
+ {
+ "epoch": 68.58689956331878,
+ "grad_norm": 0.016207464039325714,
+ "learning_rate": 0.0006,
+ "loss": 4.311431884765625,
+ "step": 4938
+ },
+ {
+ "epoch": 68.60087336244541,
+ "grad_norm": 0.01747811958193779,
+ "learning_rate": 0.0006,
+ "loss": 4.310845375061035,
+ "step": 4939
+ },
+ {
+ "epoch": 68.61484716157206,
+ "grad_norm": 0.019475717097520828,
+ "learning_rate": 0.0006,
+ "loss": 4.338624954223633,
+ "step": 4940
+ },
+ {
+ "epoch": 68.62882096069869,
+ "grad_norm": 0.018641719594597816,
+ "learning_rate": 0.0006,
+ "loss": 4.34913969039917,
+ "step": 4941
+ },
+ {
+ "epoch": 68.64279475982532,
+ "grad_norm": 0.018376432359218597,
+ "learning_rate": 0.0006,
+ "loss": 4.28464412689209,
+ "step": 4942
+ },
+ {
+ "epoch": 68.65676855895197,
+ "grad_norm": 0.017095010727643967,
+ "learning_rate": 0.0006,
+ "loss": 4.217957496643066,
+ "step": 4943
+ },
+ {
+ "epoch": 68.6707423580786,
+ "grad_norm": 0.016321983188390732,
+ "learning_rate": 0.0006,
+ "loss": 4.220042705535889,
+ "step": 4944
+ },
+ {
+ "epoch": 68.68471615720524,
+ "grad_norm": 0.017128000035881996,
+ "learning_rate": 0.0006,
+ "loss": 4.293500900268555,
+ "step": 4945
+ },
+ {
+ "epoch": 68.69868995633188,
+ "grad_norm": 0.015592411160469055,
+ "learning_rate": 0.0006,
+ "loss": 4.195798873901367,
+ "step": 4946
+ },
+ {
+ "epoch": 68.71266375545852,
+ "grad_norm": 0.016607079654932022,
+ "learning_rate": 0.0006,
+ "loss": 4.2654337882995605,
+ "step": 4947
+ },
+ {
+ "epoch": 68.72663755458515,
+ "grad_norm": 0.01545762736350298,
+ "learning_rate": 0.0006,
+ "loss": 4.303828239440918,
+ "step": 4948
+ },
+ {
+ "epoch": 68.74061135371178,
+ "grad_norm": 0.014328422956168652,
+ "learning_rate": 0.0006,
+ "loss": 4.29379940032959,
+ "step": 4949
+ },
+ {
+ "epoch": 68.75458515283843,
+ "grad_norm": 0.016516804695129395,
+ "learning_rate": 0.0006,
+ "loss": 4.2317214012146,
+ "step": 4950
+ },
+ {
+ "epoch": 68.76855895196506,
+ "grad_norm": 0.016324171796441078,
+ "learning_rate": 0.0006,
+ "loss": 4.278043746948242,
+ "step": 4951
+ },
+ {
+ "epoch": 68.78253275109171,
+ "grad_norm": 0.015511159785091877,
+ "learning_rate": 0.0006,
+ "loss": 4.188244342803955,
+ "step": 4952
+ },
+ {
+ "epoch": 68.79650655021834,
+ "grad_norm": 0.01676103286445141,
+ "learning_rate": 0.0006,
+ "loss": 4.2169976234436035,
+ "step": 4953
+ },
+ {
+ "epoch": 68.81048034934499,
+ "grad_norm": 0.015428312122821808,
+ "learning_rate": 0.0006,
+ "loss": 4.4004716873168945,
+ "step": 4954
+ },
+ {
+ "epoch": 68.82445414847162,
+ "grad_norm": 0.0162410456687212,
+ "learning_rate": 0.0006,
+ "loss": 4.282752513885498,
+ "step": 4955
+ },
+ {
+ "epoch": 68.83842794759825,
+ "grad_norm": 0.017578862607479095,
+ "learning_rate": 0.0006,
+ "loss": 4.098004341125488,
+ "step": 4956
+ },
+ {
+ "epoch": 68.8524017467249,
+ "grad_norm": 0.01805749163031578,
+ "learning_rate": 0.0006,
+ "loss": 4.277988433837891,
+ "step": 4957
+ },
+ {
+ "epoch": 68.86637554585153,
+ "grad_norm": 0.016728045418858528,
+ "learning_rate": 0.0006,
+ "loss": 4.29766321182251,
+ "step": 4958
+ },
+ {
+ "epoch": 68.88034934497817,
+ "grad_norm": 0.014899657107889652,
+ "learning_rate": 0.0006,
+ "loss": 4.329736709594727,
+ "step": 4959
+ },
+ {
+ "epoch": 68.8943231441048,
+ "grad_norm": 0.013441305607557297,
+ "learning_rate": 0.0006,
+ "loss": 4.2873215675354,
+ "step": 4960
+ },
+ {
+ "epoch": 68.90829694323143,
+ "grad_norm": 0.01554128248244524,
+ "learning_rate": 0.0006,
+ "loss": 4.278327941894531,
+ "step": 4961
+ },
+ {
+ "epoch": 68.92227074235808,
+ "grad_norm": 0.01806696318089962,
+ "learning_rate": 0.0006,
+ "loss": 4.38275146484375,
+ "step": 4962
+ },
+ {
+ "epoch": 68.93624454148471,
+ "grad_norm": 0.017025409266352654,
+ "learning_rate": 0.0006,
+ "loss": 4.210447311401367,
+ "step": 4963
+ },
+ {
+ "epoch": 68.95021834061136,
+ "grad_norm": 0.01597420684993267,
+ "learning_rate": 0.0006,
+ "loss": 4.309865951538086,
+ "step": 4964
+ },
+ {
+ "epoch": 68.96419213973799,
+ "grad_norm": 0.015835624188184738,
+ "learning_rate": 0.0006,
+ "loss": 4.225904941558838,
+ "step": 4965
+ },
+ {
+ "epoch": 68.97816593886463,
+ "grad_norm": 0.018170492723584175,
+ "learning_rate": 0.0006,
+ "loss": 4.248478889465332,
+ "step": 4966
+ },
+ {
+ "epoch": 68.99213973799127,
+ "grad_norm": 0.017372991889715195,
+ "learning_rate": 0.0006,
+ "loss": 4.283486843109131,
+ "step": 4967
+ },
+ {
+ "epoch": 69.0,
+ "grad_norm": 0.016305390745401382,
+ "learning_rate": 0.0006,
+ "loss": 4.273636817932129,
+ "step": 4968
+ },
+ {
+ "epoch": 69.0,
+ "eval_loss": 4.66873836517334,
+ "eval_runtime": 57.0791,
+ "eval_samples_per_second": 42.783,
+ "eval_steps_per_second": 1.349,
+ "step": 4968
+ },
+ {
+ "epoch": 69.01397379912663,
+ "grad_norm": 0.016567355021834373,
+ "learning_rate": 0.0006,
+ "loss": 4.269988059997559,
+ "step": 4969
+ },
+ {
+ "epoch": 69.02794759825328,
+ "grad_norm": 0.017476852983236313,
+ "learning_rate": 0.0006,
+ "loss": 4.2518720626831055,
+ "step": 4970
+ },
+ {
+ "epoch": 69.04192139737991,
+ "grad_norm": 0.017070338129997253,
+ "learning_rate": 0.0006,
+ "loss": 4.238024711608887,
+ "step": 4971
+ },
+ {
+ "epoch": 69.05589519650655,
+ "grad_norm": 0.01639373227953911,
+ "learning_rate": 0.0006,
+ "loss": 4.188558101654053,
+ "step": 4972
+ },
+ {
+ "epoch": 69.06986899563319,
+ "grad_norm": 0.01856113411486149,
+ "learning_rate": 0.0006,
+ "loss": 4.339556694030762,
+ "step": 4973
+ },
+ {
+ "epoch": 69.08384279475983,
+ "grad_norm": 0.01762578636407852,
+ "learning_rate": 0.0006,
+ "loss": 4.141518592834473,
+ "step": 4974
+ },
+ {
+ "epoch": 69.09781659388646,
+ "grad_norm": 0.015437949448823929,
+ "learning_rate": 0.0006,
+ "loss": 4.225461959838867,
+ "step": 4975
+ },
+ {
+ "epoch": 69.1117903930131,
+ "grad_norm": 0.015293503180146217,
+ "learning_rate": 0.0006,
+ "loss": 4.290884971618652,
+ "step": 4976
+ },
+ {
+ "epoch": 69.12576419213974,
+ "grad_norm": 0.016259880736470222,
+ "learning_rate": 0.0006,
+ "loss": 4.121919631958008,
+ "step": 4977
+ },
+ {
+ "epoch": 69.13973799126637,
+ "grad_norm": 0.017225584015250206,
+ "learning_rate": 0.0006,
+ "loss": 4.136716842651367,
+ "step": 4978
+ },
+ {
+ "epoch": 69.15371179039302,
+ "grad_norm": 0.01699289306998253,
+ "learning_rate": 0.0006,
+ "loss": 4.193601608276367,
+ "step": 4979
+ },
+ {
+ "epoch": 69.16768558951965,
+ "grad_norm": 0.015613161027431488,
+ "learning_rate": 0.0006,
+ "loss": 4.288212299346924,
+ "step": 4980
+ },
+ {
+ "epoch": 69.18165938864628,
+ "grad_norm": 0.01643933728337288,
+ "learning_rate": 0.0006,
+ "loss": 4.324707984924316,
+ "step": 4981
+ },
+ {
+ "epoch": 69.19563318777293,
+ "grad_norm": 0.01648905873298645,
+ "learning_rate": 0.0006,
+ "loss": 4.287806510925293,
+ "step": 4982
+ },
+ {
+ "epoch": 69.20960698689956,
+ "grad_norm": 0.015444950200617313,
+ "learning_rate": 0.0006,
+ "loss": 4.200622081756592,
+ "step": 4983
+ },
+ {
+ "epoch": 69.2235807860262,
+ "grad_norm": 0.016034258529543877,
+ "learning_rate": 0.0006,
+ "loss": 4.2193827629089355,
+ "step": 4984
+ },
+ {
+ "epoch": 69.23755458515284,
+ "grad_norm": 0.017984388396143913,
+ "learning_rate": 0.0006,
+ "loss": 4.18237829208374,
+ "step": 4985
+ },
+ {
+ "epoch": 69.25152838427948,
+ "grad_norm": 0.0195505041629076,
+ "learning_rate": 0.0006,
+ "loss": 4.216727256774902,
+ "step": 4986
+ },
+ {
+ "epoch": 69.26550218340611,
+ "grad_norm": 0.017779147252440453,
+ "learning_rate": 0.0006,
+ "loss": 4.278628826141357,
+ "step": 4987
+ },
+ {
+ "epoch": 69.27947598253274,
+ "grad_norm": 0.017119476571679115,
+ "learning_rate": 0.0006,
+ "loss": 4.2409138679504395,
+ "step": 4988
+ },
+ {
+ "epoch": 69.29344978165939,
+ "grad_norm": 0.016840364784002304,
+ "learning_rate": 0.0006,
+ "loss": 4.207575798034668,
+ "step": 4989
+ },
+ {
+ "epoch": 69.30742358078602,
+ "grad_norm": 0.014593573287129402,
+ "learning_rate": 0.0006,
+ "loss": 4.164905548095703,
+ "step": 4990
+ },
+ {
+ "epoch": 69.32139737991267,
+ "grad_norm": 0.013922630809247494,
+ "learning_rate": 0.0006,
+ "loss": 4.2696990966796875,
+ "step": 4991
+ },
+ {
+ "epoch": 69.3353711790393,
+ "grad_norm": 0.01577564887702465,
+ "learning_rate": 0.0006,
+ "loss": 4.201589107513428,
+ "step": 4992
+ },
+ {
+ "epoch": 69.34934497816595,
+ "grad_norm": 0.016392096877098083,
+ "learning_rate": 0.0006,
+ "loss": 4.248831748962402,
+ "step": 4993
+ },
+ {
+ "epoch": 69.36331877729258,
+ "grad_norm": 0.015961311757564545,
+ "learning_rate": 0.0006,
+ "loss": 4.181260108947754,
+ "step": 4994
+ },
+ {
+ "epoch": 69.37729257641921,
+ "grad_norm": 0.015350551344454288,
+ "learning_rate": 0.0006,
+ "loss": 4.230352878570557,
+ "step": 4995
+ },
+ {
+ "epoch": 69.39126637554585,
+ "grad_norm": 0.015870466828346252,
+ "learning_rate": 0.0006,
+ "loss": 4.298305511474609,
+ "step": 4996
+ },
+ {
+ "epoch": 69.40524017467249,
+ "grad_norm": 0.013716642744839191,
+ "learning_rate": 0.0006,
+ "loss": 4.220139503479004,
+ "step": 4997
+ },
+ {
+ "epoch": 69.41921397379913,
+ "grad_norm": 0.014431777410209179,
+ "learning_rate": 0.0006,
+ "loss": 4.288130760192871,
+ "step": 4998
+ },
+ {
+ "epoch": 69.43318777292576,
+ "grad_norm": 0.01426647324115038,
+ "learning_rate": 0.0006,
+ "loss": 4.295767784118652,
+ "step": 4999
+ },
+ {
+ "epoch": 69.44716157205241,
+ "grad_norm": 0.014286615885794163,
+ "learning_rate": 0.0006,
+ "loss": 4.38106632232666,
+ "step": 5000
+ },
+ {
+ "epoch": 69.46113537117904,
+ "grad_norm": 0.017990585416555405,
+ "learning_rate": 0.0006,
+ "loss": 4.1593017578125,
+ "step": 5001
+ },
+ {
+ "epoch": 69.47510917030567,
+ "grad_norm": 0.018846049904823303,
+ "learning_rate": 0.0006,
+ "loss": 4.216091632843018,
+ "step": 5002
+ },
+ {
+ "epoch": 69.48908296943232,
+ "grad_norm": 0.018179181963205338,
+ "learning_rate": 0.0006,
+ "loss": 4.240332126617432,
+ "step": 5003
+ },
+ {
+ "epoch": 69.50305676855895,
+ "grad_norm": 0.016693448647856712,
+ "learning_rate": 0.0006,
+ "loss": 4.256624221801758,
+ "step": 5004
+ },
+ {
+ "epoch": 69.5170305676856,
+ "grad_norm": 0.01718989387154579,
+ "learning_rate": 0.0006,
+ "loss": 4.300940990447998,
+ "step": 5005
+ },
+ {
+ "epoch": 69.53100436681223,
+ "grad_norm": 0.020236829295754433,
+ "learning_rate": 0.0006,
+ "loss": 4.289146423339844,
+ "step": 5006
+ },
+ {
+ "epoch": 69.54497816593886,
+ "grad_norm": 0.024860285222530365,
+ "learning_rate": 0.0006,
+ "loss": 4.216216087341309,
+ "step": 5007
+ },
+ {
+ "epoch": 69.5589519650655,
+ "grad_norm": 0.02559969387948513,
+ "learning_rate": 0.0006,
+ "loss": 4.262890815734863,
+ "step": 5008
+ },
+ {
+ "epoch": 69.57292576419214,
+ "grad_norm": 0.019180957227945328,
+ "learning_rate": 0.0006,
+ "loss": 4.230309009552002,
+ "step": 5009
+ },
+ {
+ "epoch": 69.58689956331878,
+ "grad_norm": 0.020442591980099678,
+ "learning_rate": 0.0006,
+ "loss": 4.291759490966797,
+ "step": 5010
+ },
+ {
+ "epoch": 69.60087336244541,
+ "grad_norm": 0.0212619137018919,
+ "learning_rate": 0.0006,
+ "loss": 4.3529839515686035,
+ "step": 5011
+ },
+ {
+ "epoch": 69.61484716157206,
+ "grad_norm": 0.02067175693809986,
+ "learning_rate": 0.0006,
+ "loss": 4.33197021484375,
+ "step": 5012
+ },
+ {
+ "epoch": 69.62882096069869,
+ "grad_norm": 0.01802872307598591,
+ "learning_rate": 0.0006,
+ "loss": 4.2101969718933105,
+ "step": 5013
+ },
+ {
+ "epoch": 69.64279475982532,
+ "grad_norm": 0.01787702925503254,
+ "learning_rate": 0.0006,
+ "loss": 4.270336627960205,
+ "step": 5014
+ },
+ {
+ "epoch": 69.65676855895197,
+ "grad_norm": 0.01586683839559555,
+ "learning_rate": 0.0006,
+ "loss": 4.222309589385986,
+ "step": 5015
+ },
+ {
+ "epoch": 69.6707423580786,
+ "grad_norm": 0.01717999391257763,
+ "learning_rate": 0.0006,
+ "loss": 4.253704071044922,
+ "step": 5016
+ },
+ {
+ "epoch": 69.68471615720524,
+ "grad_norm": 0.01594279147684574,
+ "learning_rate": 0.0006,
+ "loss": 4.344658851623535,
+ "step": 5017
+ },
+ {
+ "epoch": 69.69868995633188,
+ "grad_norm": 0.014419942162930965,
+ "learning_rate": 0.0006,
+ "loss": 4.2144880294799805,
+ "step": 5018
+ },
+ {
+ "epoch": 69.71266375545852,
+ "grad_norm": 0.015824224799871445,
+ "learning_rate": 0.0006,
+ "loss": 4.226252555847168,
+ "step": 5019
+ },
+ {
+ "epoch": 69.72663755458515,
+ "grad_norm": 0.013329979032278061,
+ "learning_rate": 0.0006,
+ "loss": 4.238587379455566,
+ "step": 5020
+ },
+ {
+ "epoch": 69.74061135371178,
+ "grad_norm": 0.013745979405939579,
+ "learning_rate": 0.0006,
+ "loss": 4.317209243774414,
+ "step": 5021
+ },
+ {
+ "epoch": 69.75458515283843,
+ "grad_norm": 0.014064167626202106,
+ "learning_rate": 0.0006,
+ "loss": 4.316982746124268,
+ "step": 5022
+ },
+ {
+ "epoch": 69.76855895196506,
+ "grad_norm": 0.014611301012337208,
+ "learning_rate": 0.0006,
+ "loss": 4.212728500366211,
+ "step": 5023
+ },
+ {
+ "epoch": 69.78253275109171,
+ "grad_norm": 0.014947416260838509,
+ "learning_rate": 0.0006,
+ "loss": 4.319550514221191,
+ "step": 5024
+ },
+ {
+ "epoch": 69.79650655021834,
+ "grad_norm": 0.014780526980757713,
+ "learning_rate": 0.0006,
+ "loss": 4.1888933181762695,
+ "step": 5025
+ },
+ {
+ "epoch": 69.81048034934499,
+ "grad_norm": 0.01611459068953991,
+ "learning_rate": 0.0006,
+ "loss": 4.239692211151123,
+ "step": 5026
+ },
+ {
+ "epoch": 69.82445414847162,
+ "grad_norm": 0.015395848080515862,
+ "learning_rate": 0.0006,
+ "loss": 4.308794021606445,
+ "step": 5027
+ },
+ {
+ "epoch": 69.83842794759825,
+ "grad_norm": 0.016113396733999252,
+ "learning_rate": 0.0006,
+ "loss": 4.328035354614258,
+ "step": 5028
+ },
+ {
+ "epoch": 69.8524017467249,
+ "grad_norm": 0.016386888921260834,
+ "learning_rate": 0.0006,
+ "loss": 4.272579193115234,
+ "step": 5029
+ },
+ {
+ "epoch": 69.86637554585153,
+ "grad_norm": 0.013675205409526825,
+ "learning_rate": 0.0006,
+ "loss": 4.226552963256836,
+ "step": 5030
+ },
+ {
+ "epoch": 69.88034934497817,
+ "grad_norm": 0.01472668256610632,
+ "learning_rate": 0.0006,
+ "loss": 4.224167346954346,
+ "step": 5031
+ },
+ {
+ "epoch": 69.8943231441048,
+ "grad_norm": 0.014958228915929794,
+ "learning_rate": 0.0006,
+ "loss": 4.232619285583496,
+ "step": 5032
+ },
+ {
+ "epoch": 69.90829694323143,
+ "grad_norm": 0.012926832772791386,
+ "learning_rate": 0.0006,
+ "loss": 4.27107048034668,
+ "step": 5033
+ },
+ {
+ "epoch": 69.92227074235808,
+ "grad_norm": 0.013138801790773869,
+ "learning_rate": 0.0006,
+ "loss": 4.288477420806885,
+ "step": 5034
+ },
+ {
+ "epoch": 69.93624454148471,
+ "grad_norm": 0.013415461406111717,
+ "learning_rate": 0.0006,
+ "loss": 4.2033257484436035,
+ "step": 5035
+ },
+ {
+ "epoch": 69.95021834061136,
+ "grad_norm": 0.013807233422994614,
+ "learning_rate": 0.0006,
+ "loss": 4.274764537811279,
+ "step": 5036
+ },
+ {
+ "epoch": 69.96419213973799,
+ "grad_norm": 0.013929193839430809,
+ "learning_rate": 0.0006,
+ "loss": 4.1854352951049805,
+ "step": 5037
+ },
+ {
+ "epoch": 69.97816593886463,
+ "grad_norm": 0.015497001819312572,
+ "learning_rate": 0.0006,
+ "loss": 4.226498126983643,
+ "step": 5038
+ },
+ {
+ "epoch": 69.99213973799127,
+ "grad_norm": 0.01874048449099064,
+ "learning_rate": 0.0006,
+ "loss": 4.304137706756592,
+ "step": 5039
+ },
+ {
+ "epoch": 70.0,
+ "grad_norm": 0.021745791658759117,
+ "learning_rate": 0.0006,
+ "loss": 4.206809997558594,
+ "step": 5040
+ },
+ {
+ "epoch": 70.0,
+ "eval_loss": 4.675539970397949,
+ "eval_runtime": 56.9034,
+ "eval_samples_per_second": 42.915,
+ "eval_steps_per_second": 1.353,
+ "step": 5040
+ },
+ {
+ "epoch": 70.01397379912663,
+ "grad_norm": 0.02124839462339878,
+ "learning_rate": 0.0006,
+ "loss": 4.198199272155762,
+ "step": 5041
+ },
+ {
+ "epoch": 70.02794759825328,
+ "grad_norm": 0.019432688131928444,
+ "learning_rate": 0.0006,
+ "loss": 4.187261581420898,
+ "step": 5042
+ },
+ {
+ "epoch": 70.04192139737991,
+ "grad_norm": 0.01769290305674076,
+ "learning_rate": 0.0006,
+ "loss": 4.155284881591797,
+ "step": 5043
+ },
+ {
+ "epoch": 70.05589519650655,
+ "grad_norm": 0.01751520484685898,
+ "learning_rate": 0.0006,
+ "loss": 4.386857986450195,
+ "step": 5044
+ },
+ {
+ "epoch": 70.06986899563319,
+ "grad_norm": 0.018138699233531952,
+ "learning_rate": 0.0006,
+ "loss": 4.117342472076416,
+ "step": 5045
+ },
+ {
+ "epoch": 70.08384279475983,
+ "grad_norm": 0.016331054270267487,
+ "learning_rate": 0.0006,
+ "loss": 4.225295066833496,
+ "step": 5046
+ },
+ {
+ "epoch": 70.09781659388646,
+ "grad_norm": 0.01658729277551174,
+ "learning_rate": 0.0006,
+ "loss": 4.143835067749023,
+ "step": 5047
+ },
+ {
+ "epoch": 70.1117903930131,
+ "grad_norm": 0.016367698088288307,
+ "learning_rate": 0.0006,
+ "loss": 4.1469268798828125,
+ "step": 5048
+ },
+ {
+ "epoch": 70.12576419213974,
+ "grad_norm": 0.015151629224419594,
+ "learning_rate": 0.0006,
+ "loss": 4.117629051208496,
+ "step": 5049
+ },
+ {
+ "epoch": 70.13973799126637,
+ "grad_norm": 0.016434503719210625,
+ "learning_rate": 0.0006,
+ "loss": 4.107848644256592,
+ "step": 5050
+ },
+ {
+ "epoch": 70.15371179039302,
+ "grad_norm": 0.017837485298514366,
+ "learning_rate": 0.0006,
+ "loss": 4.10837459564209,
+ "step": 5051
+ },
+ {
+ "epoch": 70.16768558951965,
+ "grad_norm": 0.0165961142629385,
+ "learning_rate": 0.0006,
+ "loss": 4.169414043426514,
+ "step": 5052
+ },
+ {
+ "epoch": 70.18165938864628,
+ "grad_norm": 0.0157014112919569,
+ "learning_rate": 0.0006,
+ "loss": 4.0629706382751465,
+ "step": 5053
+ },
+ {
+ "epoch": 70.19563318777293,
+ "grad_norm": 0.01581883803009987,
+ "learning_rate": 0.0006,
+ "loss": 4.220012664794922,
+ "step": 5054
+ },
+ {
+ "epoch": 70.20960698689956,
+ "grad_norm": 0.015285917557775974,
+ "learning_rate": 0.0006,
+ "loss": 4.275337219238281,
+ "step": 5055
+ },
+ {
+ "epoch": 70.2235807860262,
+ "grad_norm": 0.014984549954533577,
+ "learning_rate": 0.0006,
+ "loss": 4.284764289855957,
+ "step": 5056
+ },
+ {
+ "epoch": 70.23755458515284,
+ "grad_norm": 0.015172012150287628,
+ "learning_rate": 0.0006,
+ "loss": 4.234498023986816,
+ "step": 5057
+ },
+ {
+ "epoch": 70.25152838427948,
+ "grad_norm": 0.016846632584929466,
+ "learning_rate": 0.0006,
+ "loss": 4.181689262390137,
+ "step": 5058
+ },
+ {
+ "epoch": 70.26550218340611,
+ "grad_norm": 0.01920711249113083,
+ "learning_rate": 0.0006,
+ "loss": 4.235280513763428,
+ "step": 5059
+ },
+ {
+ "epoch": 70.27947598253274,
+ "grad_norm": 0.017765291035175323,
+ "learning_rate": 0.0006,
+ "loss": 4.238955497741699,
+ "step": 5060
+ },
+ {
+ "epoch": 70.29344978165939,
+ "grad_norm": 0.015436639077961445,
+ "learning_rate": 0.0006,
+ "loss": 4.232516288757324,
+ "step": 5061
+ },
+ {
+ "epoch": 70.30742358078602,
+ "grad_norm": 0.015660086646676064,
+ "learning_rate": 0.0006,
+ "loss": 4.260349273681641,
+ "step": 5062
+ },
+ {
+ "epoch": 70.32139737991267,
+ "grad_norm": 0.016095709055662155,
+ "learning_rate": 0.0006,
+ "loss": 4.168505668640137,
+ "step": 5063
+ },
+ {
+ "epoch": 70.3353711790393,
+ "grad_norm": 0.015001940540969372,
+ "learning_rate": 0.0006,
+ "loss": 4.161739349365234,
+ "step": 5064
+ },
+ {
+ "epoch": 70.34934497816595,
+ "grad_norm": 0.014668774791061878,
+ "learning_rate": 0.0006,
+ "loss": 4.181029796600342,
+ "step": 5065
+ },
+ {
+ "epoch": 70.36331877729258,
+ "grad_norm": 0.013266284950077534,
+ "learning_rate": 0.0006,
+ "loss": 4.3052873611450195,
+ "step": 5066
+ },
+ {
+ "epoch": 70.37729257641921,
+ "grad_norm": 0.015396138653159142,
+ "learning_rate": 0.0006,
+ "loss": 4.34413480758667,
+ "step": 5067
+ },
+ {
+ "epoch": 70.39126637554585,
+ "grad_norm": 0.015511523932218552,
+ "learning_rate": 0.0006,
+ "loss": 4.235037803649902,
+ "step": 5068
+ },
+ {
+ "epoch": 70.40524017467249,
+ "grad_norm": 0.014825738966464996,
+ "learning_rate": 0.0006,
+ "loss": 4.262448310852051,
+ "step": 5069
+ },
+ {
+ "epoch": 70.41921397379913,
+ "grad_norm": 0.015309952199459076,
+ "learning_rate": 0.0006,
+ "loss": 4.258711814880371,
+ "step": 5070
+ },
+ {
+ "epoch": 70.43318777292576,
+ "grad_norm": 0.017614055424928665,
+ "learning_rate": 0.0006,
+ "loss": 4.286028861999512,
+ "step": 5071
+ },
+ {
+ "epoch": 70.44716157205241,
+ "grad_norm": 0.0166039876639843,
+ "learning_rate": 0.0006,
+ "loss": 4.149197578430176,
+ "step": 5072
+ },
+ {
+ "epoch": 70.46113537117904,
+ "grad_norm": 0.016911419108510017,
+ "learning_rate": 0.0006,
+ "loss": 4.271981239318848,
+ "step": 5073
+ },
+ {
+ "epoch": 70.47510917030567,
+ "grad_norm": 0.016806820407509804,
+ "learning_rate": 0.0006,
+ "loss": 4.213315010070801,
+ "step": 5074
+ },
+ {
+ "epoch": 70.48908296943232,
+ "grad_norm": 0.015221742913126945,
+ "learning_rate": 0.0006,
+ "loss": 4.171091079711914,
+ "step": 5075
+ },
+ {
+ "epoch": 70.50305676855895,
+ "grad_norm": 0.014175347983837128,
+ "learning_rate": 0.0006,
+ "loss": 4.185309410095215,
+ "step": 5076
+ },
+ {
+ "epoch": 70.5170305676856,
+ "grad_norm": 0.013897798955440521,
+ "learning_rate": 0.0006,
+ "loss": 4.250119209289551,
+ "step": 5077
+ },
+ {
+ "epoch": 70.53100436681223,
+ "grad_norm": 0.014791003428399563,
+ "learning_rate": 0.0006,
+ "loss": 4.2504682540893555,
+ "step": 5078
+ },
+ {
+ "epoch": 70.54497816593886,
+ "grad_norm": 0.016700128093361855,
+ "learning_rate": 0.0006,
+ "loss": 4.235134124755859,
+ "step": 5079
+ },
+ {
+ "epoch": 70.5589519650655,
+ "grad_norm": 0.016679363325238228,
+ "learning_rate": 0.0006,
+ "loss": 4.32028865814209,
+ "step": 5080
+ },
+ {
+ "epoch": 70.57292576419214,
+ "grad_norm": 0.015117401257157326,
+ "learning_rate": 0.0006,
+ "loss": 4.210995674133301,
+ "step": 5081
+ },
+ {
+ "epoch": 70.58689956331878,
+ "grad_norm": 0.014441263861954212,
+ "learning_rate": 0.0006,
+ "loss": 4.27849006652832,
+ "step": 5082
+ },
+ {
+ "epoch": 70.60087336244541,
+ "grad_norm": 0.01523780357092619,
+ "learning_rate": 0.0006,
+ "loss": 4.197360992431641,
+ "step": 5083
+ },
+ {
+ "epoch": 70.61484716157206,
+ "grad_norm": 0.016201738268136978,
+ "learning_rate": 0.0006,
+ "loss": 4.28239107131958,
+ "step": 5084
+ },
+ {
+ "epoch": 70.62882096069869,
+ "grad_norm": 0.016003292053937912,
+ "learning_rate": 0.0006,
+ "loss": 4.209715843200684,
+ "step": 5085
+ },
+ {
+ "epoch": 70.64279475982532,
+ "grad_norm": 0.01472893450409174,
+ "learning_rate": 0.0006,
+ "loss": 4.169078826904297,
+ "step": 5086
+ },
+ {
+ "epoch": 70.65676855895197,
+ "grad_norm": 0.01604326069355011,
+ "learning_rate": 0.0006,
+ "loss": 4.090967178344727,
+ "step": 5087
+ },
+ {
+ "epoch": 70.6707423580786,
+ "grad_norm": 0.01598356105387211,
+ "learning_rate": 0.0006,
+ "loss": 4.16630220413208,
+ "step": 5088
+ },
+ {
+ "epoch": 70.68471615720524,
+ "grad_norm": 0.016988780349493027,
+ "learning_rate": 0.0006,
+ "loss": 4.117822170257568,
+ "step": 5089
+ },
+ {
+ "epoch": 70.69868995633188,
+ "grad_norm": 0.018582893535494804,
+ "learning_rate": 0.0006,
+ "loss": 4.192902088165283,
+ "step": 5090
+ },
+ {
+ "epoch": 70.71266375545852,
+ "grad_norm": 0.01793155074119568,
+ "learning_rate": 0.0006,
+ "loss": 4.314806938171387,
+ "step": 5091
+ },
+ {
+ "epoch": 70.72663755458515,
+ "grad_norm": 0.016286427155137062,
+ "learning_rate": 0.0006,
+ "loss": 4.2137556076049805,
+ "step": 5092
+ },
+ {
+ "epoch": 70.74061135371178,
+ "grad_norm": 0.014718376100063324,
+ "learning_rate": 0.0006,
+ "loss": 4.3073015213012695,
+ "step": 5093
+ },
+ {
+ "epoch": 70.75458515283843,
+ "grad_norm": 0.014675077050924301,
+ "learning_rate": 0.0006,
+ "loss": 4.160961151123047,
+ "step": 5094
+ },
+ {
+ "epoch": 70.76855895196506,
+ "grad_norm": 0.015550913289189339,
+ "learning_rate": 0.0006,
+ "loss": 4.322887420654297,
+ "step": 5095
+ },
+ {
+ "epoch": 70.78253275109171,
+ "grad_norm": 0.01453235000371933,
+ "learning_rate": 0.0006,
+ "loss": 4.2915873527526855,
+ "step": 5096
+ },
+ {
+ "epoch": 70.79650655021834,
+ "grad_norm": 0.01459164172410965,
+ "learning_rate": 0.0006,
+ "loss": 4.247288227081299,
+ "step": 5097
+ },
+ {
+ "epoch": 70.81048034934499,
+ "grad_norm": 0.015395080670714378,
+ "learning_rate": 0.0006,
+ "loss": 4.360955715179443,
+ "step": 5098
+ },
+ {
+ "epoch": 70.82445414847162,
+ "grad_norm": 0.01317683607339859,
+ "learning_rate": 0.0006,
+ "loss": 4.181969165802002,
+ "step": 5099
+ },
+ {
+ "epoch": 70.83842794759825,
+ "grad_norm": 0.013917932286858559,
+ "learning_rate": 0.0006,
+ "loss": 4.264626979827881,
+ "step": 5100
+ },
+ {
+ "epoch": 70.8524017467249,
+ "grad_norm": 0.015268695540726185,
+ "learning_rate": 0.0006,
+ "loss": 4.287857532501221,
+ "step": 5101
+ },
+ {
+ "epoch": 70.86637554585153,
+ "grad_norm": 0.015531329438090324,
+ "learning_rate": 0.0006,
+ "loss": 4.294188499450684,
+ "step": 5102
+ },
+ {
+ "epoch": 70.88034934497817,
+ "grad_norm": 0.015607129782438278,
+ "learning_rate": 0.0006,
+ "loss": 4.331040859222412,
+ "step": 5103
+ },
+ {
+ "epoch": 70.8943231441048,
+ "grad_norm": 0.015297149308025837,
+ "learning_rate": 0.0006,
+ "loss": 4.243650436401367,
+ "step": 5104
+ },
+ {
+ "epoch": 70.90829694323143,
+ "grad_norm": 0.014790025539696217,
+ "learning_rate": 0.0006,
+ "loss": 4.159859657287598,
+ "step": 5105
+ },
+ {
+ "epoch": 70.92227074235808,
+ "grad_norm": 0.013466274365782738,
+ "learning_rate": 0.0006,
+ "loss": 4.221992492675781,
+ "step": 5106
+ },
+ {
+ "epoch": 70.93624454148471,
+ "grad_norm": 0.01361898984760046,
+ "learning_rate": 0.0006,
+ "loss": 4.278751373291016,
+ "step": 5107
+ },
+ {
+ "epoch": 70.95021834061136,
+ "grad_norm": 0.01325751468539238,
+ "learning_rate": 0.0006,
+ "loss": 4.166001319885254,
+ "step": 5108
+ },
+ {
+ "epoch": 70.96419213973799,
+ "grad_norm": 0.012727087363600731,
+ "learning_rate": 0.0006,
+ "loss": 4.219997406005859,
+ "step": 5109
+ },
+ {
+ "epoch": 70.97816593886463,
+ "grad_norm": 0.013966867700219154,
+ "learning_rate": 0.0006,
+ "loss": 4.207491397857666,
+ "step": 5110
+ },
+ {
+ "epoch": 70.99213973799127,
+ "grad_norm": 0.015053587965667248,
+ "learning_rate": 0.0006,
+ "loss": 4.343245983123779,
+ "step": 5111
+ },
+ {
+ "epoch": 71.0,
+ "grad_norm": 0.017723368480801582,
+ "learning_rate": 0.0006,
+ "loss": 4.356122016906738,
+ "step": 5112
+ },
+ {
+ "epoch": 71.0,
+ "eval_loss": 4.598303318023682,
+ "eval_runtime": 56.7521,
+ "eval_samples_per_second": 43.029,
+ "eval_steps_per_second": 1.357,
+ "step": 5112
+ },
+ {
+ "epoch": 71.01397379912663,
+ "grad_norm": 0.018996279686689377,
+ "learning_rate": 0.0006,
+ "loss": 4.234772205352783,
+ "step": 5113
+ },
+ {
+ "epoch": 71.02794759825328,
+ "grad_norm": 0.02041555754840374,
+ "learning_rate": 0.0006,
+ "loss": 4.275816917419434,
+ "step": 5114
+ },
+ {
+ "epoch": 71.04192139737991,
+ "grad_norm": 0.021955285221338272,
+ "learning_rate": 0.0006,
+ "loss": 4.166038513183594,
+ "step": 5115
+ },
+ {
+ "epoch": 71.05589519650655,
+ "grad_norm": 0.021308740600943565,
+ "learning_rate": 0.0006,
+ "loss": 4.13063907623291,
+ "step": 5116
+ },
+ {
+ "epoch": 71.06986899563319,
+ "grad_norm": 0.01747054047882557,
+ "learning_rate": 0.0006,
+ "loss": 4.200522422790527,
+ "step": 5117
+ },
+ {
+ "epoch": 71.08384279475983,
+ "grad_norm": 0.018544532358646393,
+ "learning_rate": 0.0006,
+ "loss": 4.173915386199951,
+ "step": 5118
+ },
+ {
+ "epoch": 71.09781659388646,
+ "grad_norm": 0.0182361900806427,
+ "learning_rate": 0.0006,
+ "loss": 4.27392578125,
+ "step": 5119
+ },
+ {
+ "epoch": 71.1117903930131,
+ "grad_norm": 0.017062073573470116,
+ "learning_rate": 0.0006,
+ "loss": 4.219576358795166,
+ "step": 5120
+ },
+ {
+ "epoch": 71.12576419213974,
+ "grad_norm": 0.016289666295051575,
+ "learning_rate": 0.0006,
+ "loss": 4.133376598358154,
+ "step": 5121
+ },
+ {
+ "epoch": 71.13973799126637,
+ "grad_norm": 0.014864642173051834,
+ "learning_rate": 0.0006,
+ "loss": 4.168276786804199,
+ "step": 5122
+ },
+ {
+ "epoch": 71.15371179039302,
+ "grad_norm": 0.015309540554881096,
+ "learning_rate": 0.0006,
+ "loss": 4.175103187561035,
+ "step": 5123
+ },
+ {
+ "epoch": 71.16768558951965,
+ "grad_norm": 0.01556059904396534,
+ "learning_rate": 0.0006,
+ "loss": 4.289899826049805,
+ "step": 5124
+ },
+ {
+ "epoch": 71.18165938864628,
+ "grad_norm": 0.01479957066476345,
+ "learning_rate": 0.0006,
+ "loss": 4.341495037078857,
+ "step": 5125
+ },
+ {
+ "epoch": 71.19563318777293,
+ "grad_norm": 0.014762179926037788,
+ "learning_rate": 0.0006,
+ "loss": 4.252333641052246,
+ "step": 5126
+ },
+ {
+ "epoch": 71.20960698689956,
+ "grad_norm": 0.014897473156452179,
+ "learning_rate": 0.0006,
+ "loss": 4.2244181632995605,
+ "step": 5127
+ },
+ {
+ "epoch": 71.2235807860262,
+ "grad_norm": 0.012990163639187813,
+ "learning_rate": 0.0006,
+ "loss": 4.237288475036621,
+ "step": 5128
+ },
+ {
+ "epoch": 71.23755458515284,
+ "grad_norm": 0.012858416885137558,
+ "learning_rate": 0.0006,
+ "loss": 4.265738487243652,
+ "step": 5129
+ },
+ {
+ "epoch": 71.25152838427948,
+ "grad_norm": 0.013454788364470005,
+ "learning_rate": 0.0006,
+ "loss": 4.289131164550781,
+ "step": 5130
+ },
+ {
+ "epoch": 71.26550218340611,
+ "grad_norm": 0.013717238791286945,
+ "learning_rate": 0.0006,
+ "loss": 4.2844014167785645,
+ "step": 5131
+ },
+ {
+ "epoch": 71.27947598253274,
+ "grad_norm": 0.0135411461815238,
+ "learning_rate": 0.0006,
+ "loss": 4.139082908630371,
+ "step": 5132
+ },
+ {
+ "epoch": 71.29344978165939,
+ "grad_norm": 0.01406062301248312,
+ "learning_rate": 0.0006,
+ "loss": 4.226719856262207,
+ "step": 5133
+ },
+ {
+ "epoch": 71.30742358078602,
+ "grad_norm": 0.014027190394699574,
+ "learning_rate": 0.0006,
+ "loss": 4.191699981689453,
+ "step": 5134
+ },
+ {
+ "epoch": 71.32139737991267,
+ "grad_norm": 0.013809907250106335,
+ "learning_rate": 0.0006,
+ "loss": 4.242542266845703,
+ "step": 5135
+ },
+ {
+ "epoch": 71.3353711790393,
+ "grad_norm": 0.013507804833352566,
+ "learning_rate": 0.0006,
+ "loss": 4.068206310272217,
+ "step": 5136
+ },
+ {
+ "epoch": 71.34934497816595,
+ "grad_norm": 0.01432533748447895,
+ "learning_rate": 0.0006,
+ "loss": 4.378222942352295,
+ "step": 5137
+ },
+ {
+ "epoch": 71.36331877729258,
+ "grad_norm": 0.01753002591431141,
+ "learning_rate": 0.0006,
+ "loss": 4.174706935882568,
+ "step": 5138
+ },
+ {
+ "epoch": 71.37729257641921,
+ "grad_norm": 0.020048344507813454,
+ "learning_rate": 0.0006,
+ "loss": 4.19920015335083,
+ "step": 5139
+ },
+ {
+ "epoch": 71.39126637554585,
+ "grad_norm": 0.018551401793956757,
+ "learning_rate": 0.0006,
+ "loss": 4.175182342529297,
+ "step": 5140
+ },
+ {
+ "epoch": 71.40524017467249,
+ "grad_norm": 0.01598959229886532,
+ "learning_rate": 0.0006,
+ "loss": 4.2305169105529785,
+ "step": 5141
+ },
+ {
+ "epoch": 71.41921397379913,
+ "grad_norm": 0.016029560938477516,
+ "learning_rate": 0.0006,
+ "loss": 4.230090618133545,
+ "step": 5142
+ },
+ {
+ "epoch": 71.43318777292576,
+ "grad_norm": 0.01595657505095005,
+ "learning_rate": 0.0006,
+ "loss": 4.138402938842773,
+ "step": 5143
+ },
+ {
+ "epoch": 71.44716157205241,
+ "grad_norm": 0.01623230241239071,
+ "learning_rate": 0.0006,
+ "loss": 4.196106433868408,
+ "step": 5144
+ },
+ {
+ "epoch": 71.46113537117904,
+ "grad_norm": 0.014057597145438194,
+ "learning_rate": 0.0006,
+ "loss": 4.2599897384643555,
+ "step": 5145
+ },
+ {
+ "epoch": 71.47510917030567,
+ "grad_norm": 0.017161158844828606,
+ "learning_rate": 0.0006,
+ "loss": 4.183909893035889,
+ "step": 5146
+ },
+ {
+ "epoch": 71.48908296943232,
+ "grad_norm": 0.017038581892848015,
+ "learning_rate": 0.0006,
+ "loss": 4.373933792114258,
+ "step": 5147
+ },
+ {
+ "epoch": 71.50305676855895,
+ "grad_norm": 0.015851367264986038,
+ "learning_rate": 0.0006,
+ "loss": 4.2620086669921875,
+ "step": 5148
+ },
+ {
+ "epoch": 71.5170305676856,
+ "grad_norm": 0.014598303474485874,
+ "learning_rate": 0.0006,
+ "loss": 4.197151184082031,
+ "step": 5149
+ },
+ {
+ "epoch": 71.53100436681223,
+ "grad_norm": 0.01339802611619234,
+ "learning_rate": 0.0006,
+ "loss": 4.110706329345703,
+ "step": 5150
+ },
+ {
+ "epoch": 71.54497816593886,
+ "grad_norm": 0.015567775815725327,
+ "learning_rate": 0.0006,
+ "loss": 4.1210246086120605,
+ "step": 5151
+ },
+ {
+ "epoch": 71.5589519650655,
+ "grad_norm": 0.015616007149219513,
+ "learning_rate": 0.0006,
+ "loss": 4.129796028137207,
+ "step": 5152
+ },
+ {
+ "epoch": 71.57292576419214,
+ "grad_norm": 0.014198964461684227,
+ "learning_rate": 0.0006,
+ "loss": 4.193232536315918,
+ "step": 5153
+ },
+ {
+ "epoch": 71.58689956331878,
+ "grad_norm": 0.015226383693516254,
+ "learning_rate": 0.0006,
+ "loss": 4.223779678344727,
+ "step": 5154
+ },
+ {
+ "epoch": 71.60087336244541,
+ "grad_norm": 0.015610109083354473,
+ "learning_rate": 0.0006,
+ "loss": 4.223179817199707,
+ "step": 5155
+ },
+ {
+ "epoch": 71.61484716157206,
+ "grad_norm": 0.016034303233027458,
+ "learning_rate": 0.0006,
+ "loss": 4.196593284606934,
+ "step": 5156
+ },
+ {
+ "epoch": 71.62882096069869,
+ "grad_norm": 0.015648581087589264,
+ "learning_rate": 0.0006,
+ "loss": 4.341497421264648,
+ "step": 5157
+ },
+ {
+ "epoch": 71.64279475982532,
+ "grad_norm": 0.014117540791630745,
+ "learning_rate": 0.0006,
+ "loss": 4.194307327270508,
+ "step": 5158
+ },
+ {
+ "epoch": 71.65676855895197,
+ "grad_norm": 0.015612686984241009,
+ "learning_rate": 0.0006,
+ "loss": 4.290152549743652,
+ "step": 5159
+ },
+ {
+ "epoch": 71.6707423580786,
+ "grad_norm": 0.01722259446978569,
+ "learning_rate": 0.0006,
+ "loss": 4.258018493652344,
+ "step": 5160
+ },
+ {
+ "epoch": 71.68471615720524,
+ "grad_norm": 0.015923351049423218,
+ "learning_rate": 0.0006,
+ "loss": 4.21684455871582,
+ "step": 5161
+ },
+ {
+ "epoch": 71.69868995633188,
+ "grad_norm": 0.013663450255990028,
+ "learning_rate": 0.0006,
+ "loss": 4.175601959228516,
+ "step": 5162
+ },
+ {
+ "epoch": 71.71266375545852,
+ "grad_norm": 0.017075596377253532,
+ "learning_rate": 0.0006,
+ "loss": 4.205138683319092,
+ "step": 5163
+ },
+ {
+ "epoch": 71.72663755458515,
+ "grad_norm": 0.01723487675189972,
+ "learning_rate": 0.0006,
+ "loss": 4.08839225769043,
+ "step": 5164
+ },
+ {
+ "epoch": 71.74061135371178,
+ "grad_norm": 0.017355505377054214,
+ "learning_rate": 0.0006,
+ "loss": 4.165130615234375,
+ "step": 5165
+ },
+ {
+ "epoch": 71.75458515283843,
+ "grad_norm": 0.01868261583149433,
+ "learning_rate": 0.0006,
+ "loss": 4.16977071762085,
+ "step": 5166
+ },
+ {
+ "epoch": 71.76855895196506,
+ "grad_norm": 0.01897321827709675,
+ "learning_rate": 0.0006,
+ "loss": 4.173497200012207,
+ "step": 5167
+ },
+ {
+ "epoch": 71.78253275109171,
+ "grad_norm": 0.02045825682580471,
+ "learning_rate": 0.0006,
+ "loss": 4.176192283630371,
+ "step": 5168
+ },
+ {
+ "epoch": 71.79650655021834,
+ "grad_norm": 0.022895624861121178,
+ "learning_rate": 0.0006,
+ "loss": 4.163009166717529,
+ "step": 5169
+ },
+ {
+ "epoch": 71.81048034934499,
+ "grad_norm": 0.02062244527041912,
+ "learning_rate": 0.0006,
+ "loss": 4.156868934631348,
+ "step": 5170
+ },
+ {
+ "epoch": 71.82445414847162,
+ "grad_norm": 0.0176058579236269,
+ "learning_rate": 0.0006,
+ "loss": 4.154374122619629,
+ "step": 5171
+ },
+ {
+ "epoch": 71.83842794759825,
+ "grad_norm": 0.018049508333206177,
+ "learning_rate": 0.0006,
+ "loss": 4.296828269958496,
+ "step": 5172
+ },
+ {
+ "epoch": 71.8524017467249,
+ "grad_norm": 0.019466130062937737,
+ "learning_rate": 0.0006,
+ "loss": 4.19450569152832,
+ "step": 5173
+ },
+ {
+ "epoch": 71.86637554585153,
+ "grad_norm": 0.01838165521621704,
+ "learning_rate": 0.0006,
+ "loss": 4.11812686920166,
+ "step": 5174
+ },
+ {
+ "epoch": 71.88034934497817,
+ "grad_norm": 0.01608051173388958,
+ "learning_rate": 0.0006,
+ "loss": 4.240464210510254,
+ "step": 5175
+ },
+ {
+ "epoch": 71.8943231441048,
+ "grad_norm": 0.015807967633008957,
+ "learning_rate": 0.0006,
+ "loss": 4.275581359863281,
+ "step": 5176
+ },
+ {
+ "epoch": 71.90829694323143,
+ "grad_norm": 0.017366575077176094,
+ "learning_rate": 0.0006,
+ "loss": 4.064750671386719,
+ "step": 5177
+ },
+ {
+ "epoch": 71.92227074235808,
+ "grad_norm": 0.016241202130913734,
+ "learning_rate": 0.0006,
+ "loss": 4.203697681427002,
+ "step": 5178
+ },
+ {
+ "epoch": 71.93624454148471,
+ "grad_norm": 0.019105732440948486,
+ "learning_rate": 0.0006,
+ "loss": 4.140021800994873,
+ "step": 5179
+ },
+ {
+ "epoch": 71.95021834061136,
+ "grad_norm": 0.018492674455046654,
+ "learning_rate": 0.0006,
+ "loss": 4.2534074783325195,
+ "step": 5180
+ },
+ {
+ "epoch": 71.96419213973799,
+ "grad_norm": 0.019301170483231544,
+ "learning_rate": 0.0006,
+ "loss": 4.381392478942871,
+ "step": 5181
+ },
+ {
+ "epoch": 71.97816593886463,
+ "grad_norm": 0.019571911543607712,
+ "learning_rate": 0.0006,
+ "loss": 4.131056785583496,
+ "step": 5182
+ },
+ {
+ "epoch": 71.99213973799127,
+ "grad_norm": 0.01859748177230358,
+ "learning_rate": 0.0006,
+ "loss": 4.278930187225342,
+ "step": 5183
+ },
+ {
+ "epoch": 72.0,
+ "grad_norm": 0.02017994411289692,
+ "learning_rate": 0.0006,
+ "loss": 4.116464614868164,
+ "step": 5184
+ },
+ {
+ "epoch": 72.0,
+ "eval_loss": 4.60784912109375,
+ "eval_runtime": 56.3242,
+ "eval_samples_per_second": 43.356,
+ "eval_steps_per_second": 1.367,
+ "step": 5184
+ },
+ {
+ "epoch": 72.01397379912663,
+ "grad_norm": 0.030090296640992165,
+ "learning_rate": 0.0006,
+ "loss": 4.218588352203369,
+ "step": 5185
+ },
+ {
+ "epoch": 72.02794759825328,
+ "grad_norm": 1.330443024635315,
+ "learning_rate": 0.0006,
+ "loss": 4.417870044708252,
+ "step": 5186
+ },
+ {
+ "epoch": 72.04192139737991,
+ "grad_norm": 0.6991518139839172,
+ "learning_rate": 0.0006,
+ "loss": 8.414234161376953,
+ "step": 5187
+ },
+ {
+ "epoch": 72.05589519650655,
+ "grad_norm": 0.2631438374519348,
+ "learning_rate": 0.0006,
+ "loss": 7.41729736328125,
+ "step": 5188
+ },
+ {
+ "epoch": 72.06986899563319,
+ "grad_norm": 0.12391754984855652,
+ "learning_rate": 0.0006,
+ "loss": 7.209164619445801,
+ "step": 5189
+ },
+ {
+ "epoch": 72.08384279475983,
+ "grad_norm": 0.13519170880317688,
+ "learning_rate": 0.0006,
+ "loss": 6.943539619445801,
+ "step": 5190
+ },
+ {
+ "epoch": 72.09781659388646,
+ "grad_norm": 1.258439302444458,
+ "learning_rate": 0.0006,
+ "loss": 10.267714500427246,
+ "step": 5191
+ },
+ {
+ "epoch": 72.1117903930131,
+ "grad_norm": 1.0296908617019653,
+ "learning_rate": 0.0006,
+ "loss": 9.51768970489502,
+ "step": 5192
+ },
+ {
+ "epoch": 72.12576419213974,
+ "grad_norm": 0.12273016571998596,
+ "learning_rate": 0.0006,
+ "loss": 7.634477615356445,
+ "step": 5193
+ },
+ {
+ "epoch": 72.13973799126637,
+ "grad_norm": 0.06170198693871498,
+ "learning_rate": 0.0006,
+ "loss": 7.592115879058838,
+ "step": 5194
+ },
+ {
+ "epoch": 72.15371179039302,
+ "grad_norm": 0.06295108795166016,
+ "learning_rate": 0.0006,
+ "loss": 7.576314449310303,
+ "step": 5195
+ },
+ {
+ "epoch": 72.16768558951965,
+ "grad_norm": 0.0666993111371994,
+ "learning_rate": 0.0006,
+ "loss": 7.53862190246582,
+ "step": 5196
+ },
+ {
+ "epoch": 72.18165938864628,
+ "grad_norm": 0.052215222269296646,
+ "learning_rate": 0.0006,
+ "loss": 7.502930641174316,
+ "step": 5197
+ },
+ {
+ "epoch": 72.19563318777293,
+ "grad_norm": 0.04978961497545242,
+ "learning_rate": 0.0006,
+ "loss": 7.471630096435547,
+ "step": 5198
+ },
+ {
+ "epoch": 72.20960698689956,
+ "grad_norm": 0.07275046408176422,
+ "learning_rate": 0.0006,
+ "loss": 7.496924877166748,
+ "step": 5199
+ },
+ {
+ "epoch": 72.2235807860262,
+ "grad_norm": 0.04491819068789482,
+ "learning_rate": 0.0006,
+ "loss": 7.424494743347168,
+ "step": 5200
+ },
+ {
+ "epoch": 72.23755458515284,
+ "grad_norm": 0.043718695640563965,
+ "learning_rate": 0.0006,
+ "loss": 7.441090106964111,
+ "step": 5201
+ },
+ {
+ "epoch": 72.25152838427948,
+ "grad_norm": 0.04483497887849808,
+ "learning_rate": 0.0006,
+ "loss": 7.3753862380981445,
+ "step": 5202
+ },
+ {
+ "epoch": 72.26550218340611,
+ "grad_norm": 0.03463944420218468,
+ "learning_rate": 0.0006,
+ "loss": 7.375782012939453,
+ "step": 5203
+ },
+ {
+ "epoch": 72.27947598253274,
+ "grad_norm": 0.0397045724093914,
+ "learning_rate": 0.0006,
+ "loss": 7.358898639678955,
+ "step": 5204
+ },
+ {
+ "epoch": 72.29344978165939,
+ "grad_norm": 0.029339438304305077,
+ "learning_rate": 0.0006,
+ "loss": 7.3774871826171875,
+ "step": 5205
+ },
+ {
+ "epoch": 72.30742358078602,
+ "grad_norm": 0.025469699874520302,
+ "learning_rate": 0.0006,
+ "loss": 7.324203014373779,
+ "step": 5206
+ },
+ {
+ "epoch": 72.32139737991267,
+ "grad_norm": 0.030497785657644272,
+ "learning_rate": 0.0006,
+ "loss": 7.329351902008057,
+ "step": 5207
+ },
+ {
+ "epoch": 72.3353711790393,
+ "grad_norm": 0.03382078558206558,
+ "learning_rate": 0.0006,
+ "loss": 7.289940357208252,
+ "step": 5208
+ },
+ {
+ "epoch": 72.34934497816595,
+ "grad_norm": 0.019869405776262283,
+ "learning_rate": 0.0006,
+ "loss": 7.272582054138184,
+ "step": 5209
+ },
+ {
+ "epoch": 72.36331877729258,
+ "grad_norm": 0.02385753206908703,
+ "learning_rate": 0.0006,
+ "loss": 7.2654266357421875,
+ "step": 5210
+ },
+ {
+ "epoch": 72.37729257641921,
+ "grad_norm": 0.02794312685728073,
+ "learning_rate": 0.0006,
+ "loss": 7.250753402709961,
+ "step": 5211
+ },
+ {
+ "epoch": 72.39126637554585,
+ "grad_norm": 0.02445952221751213,
+ "learning_rate": 0.0006,
+ "loss": 7.293970108032227,
+ "step": 5212
+ },
+ {
+ "epoch": 72.40524017467249,
+ "grad_norm": 0.019660986959934235,
+ "learning_rate": 0.0006,
+ "loss": 7.1948771476745605,
+ "step": 5213
+ },
+ {
+ "epoch": 72.41921397379913,
+ "grad_norm": 0.024110861122608185,
+ "learning_rate": 0.0006,
+ "loss": 7.224499225616455,
+ "step": 5214
+ },
+ {
+ "epoch": 72.43318777292576,
+ "grad_norm": 0.021630609408020973,
+ "learning_rate": 0.0006,
+ "loss": 7.1736884117126465,
+ "step": 5215
+ },
+ {
+ "epoch": 72.44716157205241,
+ "grad_norm": 0.02185880020260811,
+ "learning_rate": 0.0006,
+ "loss": 7.167229652404785,
+ "step": 5216
+ },
+ {
+ "epoch": 72.46113537117904,
+ "grad_norm": 0.018913568928837776,
+ "learning_rate": 0.0006,
+ "loss": 7.196551322937012,
+ "step": 5217
+ },
+ {
+ "epoch": 72.47510917030567,
+ "grad_norm": 0.01802486553788185,
+ "learning_rate": 0.0006,
+ "loss": 7.144286632537842,
+ "step": 5218
+ },
+ {
+ "epoch": 72.48908296943232,
+ "grad_norm": 0.020386485382914543,
+ "learning_rate": 0.0006,
+ "loss": 7.196260929107666,
+ "step": 5219
+ },
+ {
+ "epoch": 72.50305676855895,
+ "grad_norm": 0.020077696070075035,
+ "learning_rate": 0.0006,
+ "loss": 7.138718605041504,
+ "step": 5220
+ },
+ {
+ "epoch": 72.5170305676856,
+ "grad_norm": 0.017224183306097984,
+ "learning_rate": 0.0006,
+ "loss": 7.119478702545166,
+ "step": 5221
+ },
+ {
+ "epoch": 72.53100436681223,
+ "grad_norm": 0.020847434177994728,
+ "learning_rate": 0.0006,
+ "loss": 7.068922996520996,
+ "step": 5222
+ },
+ {
+ "epoch": 72.54497816593886,
+ "grad_norm": 0.021764913573861122,
+ "learning_rate": 0.0006,
+ "loss": 7.071352958679199,
+ "step": 5223
+ },
+ {
+ "epoch": 72.5589519650655,
+ "grad_norm": 0.02697215974330902,
+ "learning_rate": 0.0006,
+ "loss": 7.033344268798828,
+ "step": 5224
+ },
+ {
+ "epoch": 72.57292576419214,
+ "grad_norm": 0.042392998933792114,
+ "learning_rate": 0.0006,
+ "loss": 7.0589919090271,
+ "step": 5225
+ },
+ {
+ "epoch": 72.58689956331878,
+ "grad_norm": 0.06616316735744476,
+ "learning_rate": 0.0006,
+ "loss": 7.060244083404541,
+ "step": 5226
+ },
+ {
+ "epoch": 72.60087336244541,
+ "grad_norm": 0.0828910544514656,
+ "learning_rate": 0.0006,
+ "loss": 7.057173728942871,
+ "step": 5227
+ },
+ {
+ "epoch": 72.61484716157206,
+ "grad_norm": 0.046767015010118484,
+ "learning_rate": 0.0006,
+ "loss": 7.010563850402832,
+ "step": 5228
+ },
+ {
+ "epoch": 72.62882096069869,
+ "grad_norm": 0.03831760957837105,
+ "learning_rate": 0.0006,
+ "loss": 6.9028825759887695,
+ "step": 5229
+ },
+ {
+ "epoch": 72.64279475982532,
+ "grad_norm": 0.05109567195177078,
+ "learning_rate": 0.0006,
+ "loss": 6.923257827758789,
+ "step": 5230
+ },
+ {
+ "epoch": 72.65676855895197,
+ "grad_norm": 0.02897934429347515,
+ "learning_rate": 0.0006,
+ "loss": 6.823014259338379,
+ "step": 5231
+ },
+ {
+ "epoch": 72.6707423580786,
+ "grad_norm": 0.03769192099571228,
+ "learning_rate": 0.0006,
+ "loss": 6.87416410446167,
+ "step": 5232
+ },
+ {
+ "epoch": 72.68471615720524,
+ "grad_norm": 0.03677637130022049,
+ "learning_rate": 0.0006,
+ "loss": 6.7649030685424805,
+ "step": 5233
+ },
+ {
+ "epoch": 72.69868995633188,
+ "grad_norm": 0.0371948778629303,
+ "learning_rate": 0.0006,
+ "loss": 6.749821662902832,
+ "step": 5234
+ },
+ {
+ "epoch": 72.71266375545852,
+ "grad_norm": 0.03057795763015747,
+ "learning_rate": 0.0006,
+ "loss": 6.748945236206055,
+ "step": 5235
+ },
+ {
+ "epoch": 72.72663755458515,
+ "grad_norm": 0.028109321370720863,
+ "learning_rate": 0.0006,
+ "loss": 6.663548946380615,
+ "step": 5236
+ },
+ {
+ "epoch": 72.74061135371178,
+ "grad_norm": 0.02442420832812786,
+ "learning_rate": 0.0006,
+ "loss": 6.636251449584961,
+ "step": 5237
+ },
+ {
+ "epoch": 72.75458515283843,
+ "grad_norm": 0.031841423362493515,
+ "learning_rate": 0.0006,
+ "loss": 6.606883525848389,
+ "step": 5238
+ },
+ {
+ "epoch": 72.76855895196506,
+ "grad_norm": 0.028100458905100822,
+ "learning_rate": 0.0006,
+ "loss": 6.603745937347412,
+ "step": 5239
+ },
+ {
+ "epoch": 72.78253275109171,
+ "grad_norm": 0.03222901374101639,
+ "learning_rate": 0.0006,
+ "loss": 6.463825225830078,
+ "step": 5240
+ },
+ {
+ "epoch": 72.79650655021834,
+ "grad_norm": 0.04766521975398064,
+ "learning_rate": 0.0006,
+ "loss": 6.43734073638916,
+ "step": 5241
+ },
+ {
+ "epoch": 72.81048034934499,
+ "grad_norm": 0.09058783948421478,
+ "learning_rate": 0.0006,
+ "loss": 6.448520183563232,
+ "step": 5242
+ },
+ {
+ "epoch": 72.82445414847162,
+ "grad_norm": 0.08995158225297928,
+ "learning_rate": 0.0006,
+ "loss": 6.522093296051025,
+ "step": 5243
+ },
+ {
+ "epoch": 72.83842794759825,
+ "grad_norm": 0.05491040274500847,
+ "learning_rate": 0.0006,
+ "loss": 6.37354040145874,
+ "step": 5244
+ },
+ {
+ "epoch": 72.8524017467249,
+ "grad_norm": 0.05198514088988304,
+ "learning_rate": 0.0006,
+ "loss": 6.362032413482666,
+ "step": 5245
+ },
+ {
+ "epoch": 72.86637554585153,
+ "grad_norm": 0.05464765429496765,
+ "learning_rate": 0.0006,
+ "loss": 6.326773643493652,
+ "step": 5246
+ },
+ {
+ "epoch": 72.88034934497817,
+ "grad_norm": 0.07337361574172974,
+ "learning_rate": 0.0006,
+ "loss": 6.297532081604004,
+ "step": 5247
+ },
+ {
+ "epoch": 72.8943231441048,
+ "grad_norm": 0.10061230510473251,
+ "learning_rate": 0.0006,
+ "loss": 6.359557628631592,
+ "step": 5248
+ },
+ {
+ "epoch": 72.90829694323143,
+ "grad_norm": 0.10005959123373032,
+ "learning_rate": 0.0006,
+ "loss": 6.422270774841309,
+ "step": 5249
+ },
+ {
+ "epoch": 72.92227074235808,
+ "grad_norm": 0.08989301323890686,
+ "learning_rate": 0.0006,
+ "loss": 6.289371490478516,
+ "step": 5250
+ },
+ {
+ "epoch": 72.93624454148471,
+ "grad_norm": 0.06371799111366272,
+ "learning_rate": 0.0006,
+ "loss": 6.199618339538574,
+ "step": 5251
+ },
+ {
+ "epoch": 72.95021834061136,
+ "grad_norm": 0.05932944267988205,
+ "learning_rate": 0.0006,
+ "loss": 6.111152648925781,
+ "step": 5252
+ },
+ {
+ "epoch": 72.96419213973799,
+ "grad_norm": 0.04735502973198891,
+ "learning_rate": 0.0006,
+ "loss": 6.026026248931885,
+ "step": 5253
+ },
+ {
+ "epoch": 72.97816593886463,
+ "grad_norm": 0.04270688816905022,
+ "learning_rate": 0.0006,
+ "loss": 6.01606559753418,
+ "step": 5254
+ },
+ {
+ "epoch": 72.99213973799127,
+ "grad_norm": 0.03774143010377884,
+ "learning_rate": 0.0006,
+ "loss": 5.8935441970825195,
+ "step": 5255
+ },
+ {
+ "epoch": 73.0,
+ "grad_norm": 0.04313631355762482,
+ "learning_rate": 0.0006,
+ "loss": 5.842061996459961,
+ "step": 5256
+ },
+ {
+ "epoch": 73.0,
+ "eval_loss": 5.9985785484313965,
+ "eval_runtime": 56.3891,
+ "eval_samples_per_second": 43.306,
+ "eval_steps_per_second": 1.366,
+ "step": 5256
+ },
+ {
+ "epoch": 73.01397379912663,
+ "grad_norm": 0.045038655400276184,
+ "learning_rate": 0.0006,
+ "loss": 5.8209662437438965,
+ "step": 5257
+ },
+ {
+ "epoch": 73.02794759825328,
+ "grad_norm": 0.03706910088658333,
+ "learning_rate": 0.0006,
+ "loss": 5.806189060211182,
+ "step": 5258
+ },
+ {
+ "epoch": 73.04192139737991,
+ "grad_norm": 0.04051685333251953,
+ "learning_rate": 0.0006,
+ "loss": 5.706869125366211,
+ "step": 5259
+ },
+ {
+ "epoch": 73.05589519650655,
+ "grad_norm": 0.04849153012037277,
+ "learning_rate": 0.0006,
+ "loss": 5.696385383605957,
+ "step": 5260
+ },
+ {
+ "epoch": 73.06986899563319,
+ "grad_norm": 0.06124403700232506,
+ "learning_rate": 0.0006,
+ "loss": 5.618223667144775,
+ "step": 5261
+ },
+ {
+ "epoch": 73.08384279475983,
+ "grad_norm": 0.07375937700271606,
+ "learning_rate": 0.0006,
+ "loss": 5.631045341491699,
+ "step": 5262
+ },
+ {
+ "epoch": 73.09781659388646,
+ "grad_norm": 0.07413844019174576,
+ "learning_rate": 0.0006,
+ "loss": 5.6532182693481445,
+ "step": 5263
+ },
+ {
+ "epoch": 73.1117903930131,
+ "grad_norm": 0.0633770152926445,
+ "learning_rate": 0.0006,
+ "loss": 5.57209587097168,
+ "step": 5264
+ },
+ {
+ "epoch": 73.12576419213974,
+ "grad_norm": 0.05436766892671585,
+ "learning_rate": 0.0006,
+ "loss": 5.53944730758667,
+ "step": 5265
+ },
+ {
+ "epoch": 73.13973799126637,
+ "grad_norm": 0.04639481008052826,
+ "learning_rate": 0.0006,
+ "loss": 5.502948760986328,
+ "step": 5266
+ },
+ {
+ "epoch": 73.15371179039302,
+ "grad_norm": 0.07556558400392532,
+ "learning_rate": 0.0006,
+ "loss": 5.429069995880127,
+ "step": 5267
+ },
+ {
+ "epoch": 73.16768558951965,
+ "grad_norm": 0.1127929836511612,
+ "learning_rate": 0.0006,
+ "loss": 5.545061111450195,
+ "step": 5268
+ },
+ {
+ "epoch": 73.18165938864628,
+ "grad_norm": 0.07620745897293091,
+ "learning_rate": 0.0006,
+ "loss": 5.507102966308594,
+ "step": 5269
+ },
+ {
+ "epoch": 73.19563318777293,
+ "grad_norm": 0.05076538026332855,
+ "learning_rate": 0.0006,
+ "loss": 5.329682350158691,
+ "step": 5270
+ },
+ {
+ "epoch": 73.20960698689956,
+ "grad_norm": 0.05407547950744629,
+ "learning_rate": 0.0006,
+ "loss": 5.3048553466796875,
+ "step": 5271
+ },
+ {
+ "epoch": 73.2235807860262,
+ "grad_norm": 0.059095583856105804,
+ "learning_rate": 0.0006,
+ "loss": 5.292568683624268,
+ "step": 5272
+ },
+ {
+ "epoch": 73.23755458515284,
+ "grad_norm": 0.06193305924534798,
+ "learning_rate": 0.0006,
+ "loss": 5.1870927810668945,
+ "step": 5273
+ },
+ {
+ "epoch": 73.25152838427948,
+ "grad_norm": 0.062261562794446945,
+ "learning_rate": 0.0006,
+ "loss": 5.256819248199463,
+ "step": 5274
+ },
+ {
+ "epoch": 73.26550218340611,
+ "grad_norm": 0.0732516348361969,
+ "learning_rate": 0.0006,
+ "loss": 5.256464004516602,
+ "step": 5275
+ },
+ {
+ "epoch": 73.27947598253274,
+ "grad_norm": 0.12177979946136475,
+ "learning_rate": 0.0006,
+ "loss": 5.284239768981934,
+ "step": 5276
+ },
+ {
+ "epoch": 73.29344978165939,
+ "grad_norm": 0.10459653288125992,
+ "learning_rate": 0.0006,
+ "loss": 5.269713401794434,
+ "step": 5277
+ },
+ {
+ "epoch": 73.30742358078602,
+ "grad_norm": 0.07854694873094559,
+ "learning_rate": 0.0006,
+ "loss": 5.3245391845703125,
+ "step": 5278
+ },
+ {
+ "epoch": 73.32139737991267,
+ "grad_norm": 0.06030094623565674,
+ "learning_rate": 0.0006,
+ "loss": 5.033702850341797,
+ "step": 5279
+ },
+ {
+ "epoch": 73.3353711790393,
+ "grad_norm": 0.05168461427092552,
+ "learning_rate": 0.0006,
+ "loss": 5.140530109405518,
+ "step": 5280
+ },
+ {
+ "epoch": 73.34934497816595,
+ "grad_norm": 0.041064050048589706,
+ "learning_rate": 0.0006,
+ "loss": 5.1085896492004395,
+ "step": 5281
+ },
+ {
+ "epoch": 73.36331877729258,
+ "grad_norm": 0.03788219019770622,
+ "learning_rate": 0.0006,
+ "loss": 5.045009613037109,
+ "step": 5282
+ },
+ {
+ "epoch": 73.37729257641921,
+ "grad_norm": 0.03851071000099182,
+ "learning_rate": 0.0006,
+ "loss": 5.055008888244629,
+ "step": 5283
+ },
+ {
+ "epoch": 73.39126637554585,
+ "grad_norm": 0.03448539599776268,
+ "learning_rate": 0.0006,
+ "loss": 5.10657262802124,
+ "step": 5284
+ },
+ {
+ "epoch": 73.40524017467249,
+ "grad_norm": 0.026618091389536858,
+ "learning_rate": 0.0006,
+ "loss": 5.0342912673950195,
+ "step": 5285
+ },
+ {
+ "epoch": 73.41921397379913,
+ "grad_norm": 0.03060489147901535,
+ "learning_rate": 0.0006,
+ "loss": 5.022700786590576,
+ "step": 5286
+ },
+ {
+ "epoch": 73.43318777292576,
+ "grad_norm": 0.03362000361084938,
+ "learning_rate": 0.0006,
+ "loss": 4.891584396362305,
+ "step": 5287
+ },
+ {
+ "epoch": 73.44716157205241,
+ "grad_norm": 0.03887505829334259,
+ "learning_rate": 0.0006,
+ "loss": 4.772258758544922,
+ "step": 5288
+ },
+ {
+ "epoch": 73.46113537117904,
+ "grad_norm": 0.07157225161790848,
+ "learning_rate": 0.0006,
+ "loss": 4.997142791748047,
+ "step": 5289
+ },
+ {
+ "epoch": 73.47510917030567,
+ "grad_norm": 0.10180624574422836,
+ "learning_rate": 0.0006,
+ "loss": 4.908722877502441,
+ "step": 5290
+ },
+ {
+ "epoch": 73.48908296943232,
+ "grad_norm": 0.06529122591018677,
+ "learning_rate": 0.0006,
+ "loss": 4.98001766204834,
+ "step": 5291
+ },
+ {
+ "epoch": 73.50305676855895,
+ "grad_norm": 0.055171865969896317,
+ "learning_rate": 0.0006,
+ "loss": 4.877438068389893,
+ "step": 5292
+ },
+ {
+ "epoch": 73.5170305676856,
+ "grad_norm": 0.0318770632147789,
+ "learning_rate": 0.0006,
+ "loss": 4.836934566497803,
+ "step": 5293
+ },
+ {
+ "epoch": 73.53100436681223,
+ "grad_norm": 0.042072564363479614,
+ "learning_rate": 0.0006,
+ "loss": 4.919489860534668,
+ "step": 5294
+ },
+ {
+ "epoch": 73.54497816593886,
+ "grad_norm": 0.029197754338383675,
+ "learning_rate": 0.0006,
+ "loss": 4.805579662322998,
+ "step": 5295
+ },
+ {
+ "epoch": 73.5589519650655,
+ "grad_norm": 0.03249497339129448,
+ "learning_rate": 0.0006,
+ "loss": 4.704806327819824,
+ "step": 5296
+ },
+ {
+ "epoch": 73.57292576419214,
+ "grad_norm": 0.025050945580005646,
+ "learning_rate": 0.0006,
+ "loss": 4.936512470245361,
+ "step": 5297
+ },
+ {
+ "epoch": 73.58689956331878,
+ "grad_norm": 0.029073381796479225,
+ "learning_rate": 0.0006,
+ "loss": 4.757592678070068,
+ "step": 5298
+ },
+ {
+ "epoch": 73.60087336244541,
+ "grad_norm": 0.03465264290571213,
+ "learning_rate": 0.0006,
+ "loss": 4.729528427124023,
+ "step": 5299
+ },
+ {
+ "epoch": 73.61484716157206,
+ "grad_norm": 0.054795749485492706,
+ "learning_rate": 0.0006,
+ "loss": 4.77252197265625,
+ "step": 5300
+ },
+ {
+ "epoch": 73.62882096069869,
+ "grad_norm": 0.18129196763038635,
+ "learning_rate": 0.0006,
+ "loss": 4.901700019836426,
+ "step": 5301
+ },
+ {
+ "epoch": 73.64279475982532,
+ "grad_norm": 0.23274236917495728,
+ "learning_rate": 0.0006,
+ "loss": 5.207432270050049,
+ "step": 5302
+ },
+ {
+ "epoch": 73.65676855895197,
+ "grad_norm": 0.18162772059440613,
+ "learning_rate": 0.0006,
+ "loss": 5.326632976531982,
+ "step": 5303
+ },
+ {
+ "epoch": 73.6707423580786,
+ "grad_norm": 0.11858413368463516,
+ "learning_rate": 0.0006,
+ "loss": 5.3730058670043945,
+ "step": 5304
+ },
+ {
+ "epoch": 73.68471615720524,
+ "grad_norm": 0.06031268090009689,
+ "learning_rate": 0.0006,
+ "loss": 5.330936431884766,
+ "step": 5305
+ },
+ {
+ "epoch": 73.69868995633188,
+ "grad_norm": 0.059304628521203995,
+ "learning_rate": 0.0006,
+ "loss": 5.280514717102051,
+ "step": 5306
+ },
+ {
+ "epoch": 73.71266375545852,
+ "grad_norm": 0.04702131822705269,
+ "learning_rate": 0.0006,
+ "loss": 5.288274765014648,
+ "step": 5307
+ },
+ {
+ "epoch": 73.72663755458515,
+ "grad_norm": 0.037792228162288666,
+ "learning_rate": 0.0006,
+ "loss": 5.173289775848389,
+ "step": 5308
+ },
+ {
+ "epoch": 73.74061135371178,
+ "grad_norm": 0.037763576954603195,
+ "learning_rate": 0.0006,
+ "loss": 5.084990978240967,
+ "step": 5309
+ },
+ {
+ "epoch": 73.75458515283843,
+ "grad_norm": 0.03337346389889717,
+ "learning_rate": 0.0006,
+ "loss": 5.121417045593262,
+ "step": 5310
+ },
+ {
+ "epoch": 73.76855895196506,
+ "grad_norm": 0.03184802085161209,
+ "learning_rate": 0.0006,
+ "loss": 5.208918571472168,
+ "step": 5311
+ },
+ {
+ "epoch": 73.78253275109171,
+ "grad_norm": 0.02832048572599888,
+ "learning_rate": 0.0006,
+ "loss": 4.923543453216553,
+ "step": 5312
+ },
+ {
+ "epoch": 73.79650655021834,
+ "grad_norm": 0.028297102078795433,
+ "learning_rate": 0.0006,
+ "loss": 4.944367408752441,
+ "step": 5313
+ },
+ {
+ "epoch": 73.81048034934499,
+ "grad_norm": 0.027701659128069878,
+ "learning_rate": 0.0006,
+ "loss": 4.858168601989746,
+ "step": 5314
+ },
+ {
+ "epoch": 73.82445414847162,
+ "grad_norm": 0.02214639261364937,
+ "learning_rate": 0.0006,
+ "loss": 4.887052536010742,
+ "step": 5315
+ },
+ {
+ "epoch": 73.83842794759825,
+ "grad_norm": 0.01995093561708927,
+ "learning_rate": 0.0006,
+ "loss": 4.885797500610352,
+ "step": 5316
+ },
+ {
+ "epoch": 73.8524017467249,
+ "grad_norm": 0.020473985001444817,
+ "learning_rate": 0.0006,
+ "loss": 4.917325019836426,
+ "step": 5317
+ },
+ {
+ "epoch": 73.86637554585153,
+ "grad_norm": 0.0192868635058403,
+ "learning_rate": 0.0006,
+ "loss": 4.801746845245361,
+ "step": 5318
+ },
+ {
+ "epoch": 73.88034934497817,
+ "grad_norm": 0.017108267173171043,
+ "learning_rate": 0.0006,
+ "loss": 4.893620491027832,
+ "step": 5319
+ },
+ {
+ "epoch": 73.8943231441048,
+ "grad_norm": 0.01718071475625038,
+ "learning_rate": 0.0006,
+ "loss": 4.940695285797119,
+ "step": 5320
+ },
+ {
+ "epoch": 73.90829694323143,
+ "grad_norm": 0.017821557819843292,
+ "learning_rate": 0.0006,
+ "loss": 4.767613410949707,
+ "step": 5321
+ },
+ {
+ "epoch": 73.92227074235808,
+ "grad_norm": 0.015262950211763382,
+ "learning_rate": 0.0006,
+ "loss": 4.910315036773682,
+ "step": 5322
+ },
+ {
+ "epoch": 73.93624454148471,
+ "grad_norm": 0.01769006811082363,
+ "learning_rate": 0.0006,
+ "loss": 4.7743449211120605,
+ "step": 5323
+ },
+ {
+ "epoch": 73.95021834061136,
+ "grad_norm": 0.014460788108408451,
+ "learning_rate": 0.0006,
+ "loss": 4.743494987487793,
+ "step": 5324
+ },
+ {
+ "epoch": 73.96419213973799,
+ "grad_norm": 0.01403993833810091,
+ "learning_rate": 0.0006,
+ "loss": 4.666640281677246,
+ "step": 5325
+ },
+ {
+ "epoch": 73.97816593886463,
+ "grad_norm": 0.013997809961438179,
+ "learning_rate": 0.0006,
+ "loss": 4.698747158050537,
+ "step": 5326
+ },
+ {
+ "epoch": 73.99213973799127,
+ "grad_norm": 0.014244182966649532,
+ "learning_rate": 0.0006,
+ "loss": 4.745250701904297,
+ "step": 5327
+ },
+ {
+ "epoch": 74.0,
+ "grad_norm": 0.015144186094403267,
+ "learning_rate": 0.0006,
+ "loss": 4.667946815490723,
+ "step": 5328
+ },
+ {
+ "epoch": 74.0,
+ "eval_loss": 5.005440711975098,
+ "eval_runtime": 56.7836,
+ "eval_samples_per_second": 43.005,
+ "eval_steps_per_second": 1.356,
+ "step": 5328
+ },
+ {
+ "epoch": 74.01397379912663,
+ "grad_norm": 0.020384902134537697,
+ "learning_rate": 0.0006,
+ "loss": 4.626955509185791,
+ "step": 5329
+ },
+ {
+ "epoch": 74.02794759825328,
+ "grad_norm": 0.03624720871448517,
+ "learning_rate": 0.0006,
+ "loss": 4.713998317718506,
+ "step": 5330
+ },
+ {
+ "epoch": 74.04192139737991,
+ "grad_norm": 0.05701754242181778,
+ "learning_rate": 0.0006,
+ "loss": 4.717874526977539,
+ "step": 5331
+ },
+ {
+ "epoch": 74.05589519650655,
+ "grad_norm": 0.03942621126770973,
+ "learning_rate": 0.0006,
+ "loss": 4.7026190757751465,
+ "step": 5332
+ },
+ {
+ "epoch": 74.06986899563319,
+ "grad_norm": 0.017803272232413292,
+ "learning_rate": 0.0006,
+ "loss": 4.702548027038574,
+ "step": 5333
+ },
+ {
+ "epoch": 74.08384279475983,
+ "grad_norm": 0.017024677246809006,
+ "learning_rate": 0.0006,
+ "loss": 4.69752311706543,
+ "step": 5334
+ },
+ {
+ "epoch": 74.09781659388646,
+ "grad_norm": 0.016467854380607605,
+ "learning_rate": 0.0006,
+ "loss": 4.624876976013184,
+ "step": 5335
+ },
+ {
+ "epoch": 74.1117903930131,
+ "grad_norm": 0.016569215804338455,
+ "learning_rate": 0.0006,
+ "loss": 4.645040512084961,
+ "step": 5336
+ },
+ {
+ "epoch": 74.12576419213974,
+ "grad_norm": 0.012872067280113697,
+ "learning_rate": 0.0006,
+ "loss": 4.64582633972168,
+ "step": 5337
+ },
+ {
+ "epoch": 74.13973799126637,
+ "grad_norm": 0.015689387917518616,
+ "learning_rate": 0.0006,
+ "loss": 4.596040725708008,
+ "step": 5338
+ },
+ {
+ "epoch": 74.15371179039302,
+ "grad_norm": 0.013082082383334637,
+ "learning_rate": 0.0006,
+ "loss": 4.648033142089844,
+ "step": 5339
+ },
+ {
+ "epoch": 74.16768558951965,
+ "grad_norm": 0.013728172518312931,
+ "learning_rate": 0.0006,
+ "loss": 4.657873153686523,
+ "step": 5340
+ },
+ {
+ "epoch": 74.18165938864628,
+ "grad_norm": 0.013120594434440136,
+ "learning_rate": 0.0006,
+ "loss": 4.696074485778809,
+ "step": 5341
+ },
+ {
+ "epoch": 74.19563318777293,
+ "grad_norm": 0.01431245356798172,
+ "learning_rate": 0.0006,
+ "loss": 4.5679121017456055,
+ "step": 5342
+ },
+ {
+ "epoch": 74.20960698689956,
+ "grad_norm": 0.013994511216878891,
+ "learning_rate": 0.0006,
+ "loss": 4.5236406326293945,
+ "step": 5343
+ },
+ {
+ "epoch": 74.2235807860262,
+ "grad_norm": 0.011971707455813885,
+ "learning_rate": 0.0006,
+ "loss": 4.626547813415527,
+ "step": 5344
+ },
+ {
+ "epoch": 74.23755458515284,
+ "grad_norm": 0.012805287726223469,
+ "learning_rate": 0.0006,
+ "loss": 4.556386947631836,
+ "step": 5345
+ },
+ {
+ "epoch": 74.25152838427948,
+ "grad_norm": 0.011694584973156452,
+ "learning_rate": 0.0006,
+ "loss": 4.656987190246582,
+ "step": 5346
+ },
+ {
+ "epoch": 74.26550218340611,
+ "grad_norm": 0.013396821916103363,
+ "learning_rate": 0.0006,
+ "loss": 4.526467323303223,
+ "step": 5347
+ },
+ {
+ "epoch": 74.27947598253274,
+ "grad_norm": 0.014398031868040562,
+ "learning_rate": 0.0006,
+ "loss": 4.595100402832031,
+ "step": 5348
+ },
+ {
+ "epoch": 74.29344978165939,
+ "grad_norm": 0.02165800705552101,
+ "learning_rate": 0.0006,
+ "loss": 4.460549354553223,
+ "step": 5349
+ },
+ {
+ "epoch": 74.30742358078602,
+ "grad_norm": 0.03276512026786804,
+ "learning_rate": 0.0006,
+ "loss": 4.498189449310303,
+ "step": 5350
+ },
+ {
+ "epoch": 74.32139737991267,
+ "grad_norm": 0.04616224020719528,
+ "learning_rate": 0.0006,
+ "loss": 4.5268659591674805,
+ "step": 5351
+ },
+ {
+ "epoch": 74.3353711790393,
+ "grad_norm": 0.03863786533474922,
+ "learning_rate": 0.0006,
+ "loss": 4.554747104644775,
+ "step": 5352
+ },
+ {
+ "epoch": 74.34934497816595,
+ "grad_norm": 0.01749865710735321,
+ "learning_rate": 0.0006,
+ "loss": 4.6656036376953125,
+ "step": 5353
+ },
+ {
+ "epoch": 74.36331877729258,
+ "grad_norm": 0.021814396604895592,
+ "learning_rate": 0.0006,
+ "loss": 4.607456207275391,
+ "step": 5354
+ },
+ {
+ "epoch": 74.37729257641921,
+ "grad_norm": 0.017018496990203857,
+ "learning_rate": 0.0006,
+ "loss": 4.54205322265625,
+ "step": 5355
+ },
+ {
+ "epoch": 74.39126637554585,
+ "grad_norm": 0.018297340720891953,
+ "learning_rate": 0.0006,
+ "loss": 4.608870506286621,
+ "step": 5356
+ },
+ {
+ "epoch": 74.40524017467249,
+ "grad_norm": 0.01473482046276331,
+ "learning_rate": 0.0006,
+ "loss": 4.384891510009766,
+ "step": 5357
+ },
+ {
+ "epoch": 74.41921397379913,
+ "grad_norm": 0.01579415239393711,
+ "learning_rate": 0.0006,
+ "loss": 4.588859558105469,
+ "step": 5358
+ },
+ {
+ "epoch": 74.43318777292576,
+ "grad_norm": 0.013906026259064674,
+ "learning_rate": 0.0006,
+ "loss": 4.525175094604492,
+ "step": 5359
+ },
+ {
+ "epoch": 74.44716157205241,
+ "grad_norm": 0.014163361862301826,
+ "learning_rate": 0.0006,
+ "loss": 4.495802879333496,
+ "step": 5360
+ },
+ {
+ "epoch": 74.46113537117904,
+ "grad_norm": 0.014009335078299046,
+ "learning_rate": 0.0006,
+ "loss": 4.5960869789123535,
+ "step": 5361
+ },
+ {
+ "epoch": 74.47510917030567,
+ "grad_norm": 0.012765258550643921,
+ "learning_rate": 0.0006,
+ "loss": 4.515678405761719,
+ "step": 5362
+ },
+ {
+ "epoch": 74.48908296943232,
+ "grad_norm": 0.014071918092668056,
+ "learning_rate": 0.0006,
+ "loss": 4.394025802612305,
+ "step": 5363
+ },
+ {
+ "epoch": 74.50305676855895,
+ "grad_norm": 0.01244515273720026,
+ "learning_rate": 0.0006,
+ "loss": 4.576985836029053,
+ "step": 5364
+ },
+ {
+ "epoch": 74.5170305676856,
+ "grad_norm": 0.012975298799574375,
+ "learning_rate": 0.0006,
+ "loss": 4.540657997131348,
+ "step": 5365
+ },
+ {
+ "epoch": 74.53100436681223,
+ "grad_norm": 0.012380947358906269,
+ "learning_rate": 0.0006,
+ "loss": 4.55731725692749,
+ "step": 5366
+ },
+ {
+ "epoch": 74.54497816593886,
+ "grad_norm": 0.012577617540955544,
+ "learning_rate": 0.0006,
+ "loss": 4.500924110412598,
+ "step": 5367
+ },
+ {
+ "epoch": 74.5589519650655,
+ "grad_norm": 0.012089254334568977,
+ "learning_rate": 0.0006,
+ "loss": 4.561795234680176,
+ "step": 5368
+ },
+ {
+ "epoch": 74.57292576419214,
+ "grad_norm": 0.012024011462926865,
+ "learning_rate": 0.0006,
+ "loss": 4.430635452270508,
+ "step": 5369
+ },
+ {
+ "epoch": 74.58689956331878,
+ "grad_norm": 0.012361960485577583,
+ "learning_rate": 0.0006,
+ "loss": 4.4664106369018555,
+ "step": 5370
+ },
+ {
+ "epoch": 74.60087336244541,
+ "grad_norm": 0.012149893678724766,
+ "learning_rate": 0.0006,
+ "loss": 4.522271156311035,
+ "step": 5371
+ },
+ {
+ "epoch": 74.61484716157206,
+ "grad_norm": 0.01166840922087431,
+ "learning_rate": 0.0006,
+ "loss": 4.3870463371276855,
+ "step": 5372
+ },
+ {
+ "epoch": 74.62882096069869,
+ "grad_norm": 0.01160407718271017,
+ "learning_rate": 0.0006,
+ "loss": 4.504255771636963,
+ "step": 5373
+ },
+ {
+ "epoch": 74.64279475982532,
+ "grad_norm": 0.011259177699685097,
+ "learning_rate": 0.0006,
+ "loss": 4.507267951965332,
+ "step": 5374
+ },
+ {
+ "epoch": 74.65676855895197,
+ "grad_norm": 0.012285715900361538,
+ "learning_rate": 0.0006,
+ "loss": 4.472966194152832,
+ "step": 5375
+ },
+ {
+ "epoch": 74.6707423580786,
+ "grad_norm": 0.011214440688490868,
+ "learning_rate": 0.0006,
+ "loss": 4.5840253829956055,
+ "step": 5376
+ },
+ {
+ "epoch": 74.68471615720524,
+ "grad_norm": 0.012394879944622517,
+ "learning_rate": 0.0006,
+ "loss": 4.417830944061279,
+ "step": 5377
+ },
+ {
+ "epoch": 74.69868995633188,
+ "grad_norm": 0.012360049411654472,
+ "learning_rate": 0.0006,
+ "loss": 4.535454750061035,
+ "step": 5378
+ },
+ {
+ "epoch": 74.71266375545852,
+ "grad_norm": 0.011961652897298336,
+ "learning_rate": 0.0006,
+ "loss": 4.403409957885742,
+ "step": 5379
+ },
+ {
+ "epoch": 74.72663755458515,
+ "grad_norm": 0.012078274972736835,
+ "learning_rate": 0.0006,
+ "loss": 4.448366641998291,
+ "step": 5380
+ },
+ {
+ "epoch": 74.74061135371178,
+ "grad_norm": 0.013239333406090736,
+ "learning_rate": 0.0006,
+ "loss": 4.543450355529785,
+ "step": 5381
+ },
+ {
+ "epoch": 74.75458515283843,
+ "grad_norm": 0.01476441603153944,
+ "learning_rate": 0.0006,
+ "loss": 4.488677978515625,
+ "step": 5382
+ },
+ {
+ "epoch": 74.76855895196506,
+ "grad_norm": 0.013968009501695633,
+ "learning_rate": 0.0006,
+ "loss": 4.603639602661133,
+ "step": 5383
+ },
+ {
+ "epoch": 74.78253275109171,
+ "grad_norm": 0.012322458438575268,
+ "learning_rate": 0.0006,
+ "loss": 4.637874603271484,
+ "step": 5384
+ },
+ {
+ "epoch": 74.79650655021834,
+ "grad_norm": 0.013352593407034874,
+ "learning_rate": 0.0006,
+ "loss": 4.474833965301514,
+ "step": 5385
+ },
+ {
+ "epoch": 74.81048034934499,
+ "grad_norm": 0.014671728946268559,
+ "learning_rate": 0.0006,
+ "loss": 4.516847610473633,
+ "step": 5386
+ },
+ {
+ "epoch": 74.82445414847162,
+ "grad_norm": 0.017163321375846863,
+ "learning_rate": 0.0006,
+ "loss": 4.462325096130371,
+ "step": 5387
+ },
+ {
+ "epoch": 74.83842794759825,
+ "grad_norm": 0.02012511156499386,
+ "learning_rate": 0.0006,
+ "loss": 4.355982780456543,
+ "step": 5388
+ },
+ {
+ "epoch": 74.8524017467249,
+ "grad_norm": 0.01881697215139866,
+ "learning_rate": 0.0006,
+ "loss": 4.451120376586914,
+ "step": 5389
+ },
+ {
+ "epoch": 74.86637554585153,
+ "grad_norm": 0.017279407009482384,
+ "learning_rate": 0.0006,
+ "loss": 4.461202621459961,
+ "step": 5390
+ },
+ {
+ "epoch": 74.88034934497817,
+ "grad_norm": 0.015470407903194427,
+ "learning_rate": 0.0006,
+ "loss": 4.4968671798706055,
+ "step": 5391
+ },
+ {
+ "epoch": 74.8943231441048,
+ "grad_norm": 0.01411783043295145,
+ "learning_rate": 0.0006,
+ "loss": 4.558697700500488,
+ "step": 5392
+ },
+ {
+ "epoch": 74.90829694323143,
+ "grad_norm": 0.017178812995553017,
+ "learning_rate": 0.0006,
+ "loss": 4.254153251647949,
+ "step": 5393
+ },
+ {
+ "epoch": 74.92227074235808,
+ "grad_norm": 0.019193004816770554,
+ "learning_rate": 0.0006,
+ "loss": 4.394747257232666,
+ "step": 5394
+ },
+ {
+ "epoch": 74.93624454148471,
+ "grad_norm": 0.018942881375551224,
+ "learning_rate": 0.0006,
+ "loss": 4.459539413452148,
+ "step": 5395
+ },
+ {
+ "epoch": 74.95021834061136,
+ "grad_norm": 0.01641223020851612,
+ "learning_rate": 0.0006,
+ "loss": 4.485820770263672,
+ "step": 5396
+ },
+ {
+ "epoch": 74.96419213973799,
+ "grad_norm": 0.014485753141343594,
+ "learning_rate": 0.0006,
+ "loss": 4.5068583488464355,
+ "step": 5397
+ },
+ {
+ "epoch": 74.97816593886463,
+ "grad_norm": 0.018242299556732178,
+ "learning_rate": 0.0006,
+ "loss": 4.508835792541504,
+ "step": 5398
+ },
+ {
+ "epoch": 74.99213973799127,
+ "grad_norm": 0.016491873189806938,
+ "learning_rate": 0.0006,
+ "loss": 4.4760870933532715,
+ "step": 5399
+ },
+ {
+ "epoch": 75.0,
+ "grad_norm": 0.014543353579938412,
+ "learning_rate": 0.0006,
+ "loss": 4.540217399597168,
+ "step": 5400
+ },
+ {
+ "epoch": 75.0,
+ "eval_loss": 4.7617573738098145,
+ "eval_runtime": 57.321,
+ "eval_samples_per_second": 42.602,
+ "eval_steps_per_second": 1.343,
+ "step": 5400
+ },
+ {
+ "epoch": 75.01397379912663,
+ "grad_norm": 0.014417114667594433,
+ "learning_rate": 0.0006,
+ "loss": 4.332956790924072,
+ "step": 5401
+ },
+ {
+ "epoch": 75.02794759825328,
+ "grad_norm": 0.01607639715075493,
+ "learning_rate": 0.0006,
+ "loss": 4.352453708648682,
+ "step": 5402
+ },
+ {
+ "epoch": 75.04192139737991,
+ "grad_norm": 0.015438318252563477,
+ "learning_rate": 0.0006,
+ "loss": 4.387139320373535,
+ "step": 5403
+ },
+ {
+ "epoch": 75.05589519650655,
+ "grad_norm": 0.015539007261395454,
+ "learning_rate": 0.0006,
+ "loss": 4.453171730041504,
+ "step": 5404
+ },
+ {
+ "epoch": 75.06986899563319,
+ "grad_norm": 0.017361903563141823,
+ "learning_rate": 0.0006,
+ "loss": 4.512143135070801,
+ "step": 5405
+ },
+ {
+ "epoch": 75.08384279475983,
+ "grad_norm": 0.01815599389374256,
+ "learning_rate": 0.0006,
+ "loss": 4.508997440338135,
+ "step": 5406
+ },
+ {
+ "epoch": 75.09781659388646,
+ "grad_norm": 0.01620607078075409,
+ "learning_rate": 0.0006,
+ "loss": 4.322444915771484,
+ "step": 5407
+ },
+ {
+ "epoch": 75.1117903930131,
+ "grad_norm": 0.014773265458643436,
+ "learning_rate": 0.0006,
+ "loss": 4.359251976013184,
+ "step": 5408
+ },
+ {
+ "epoch": 75.12576419213974,
+ "grad_norm": 0.014099023304879665,
+ "learning_rate": 0.0006,
+ "loss": 4.313380241394043,
+ "step": 5409
+ },
+ {
+ "epoch": 75.13973799126637,
+ "grad_norm": 0.014387735165655613,
+ "learning_rate": 0.0006,
+ "loss": 4.435023307800293,
+ "step": 5410
+ },
+ {
+ "epoch": 75.15371179039302,
+ "grad_norm": 0.014886162243783474,
+ "learning_rate": 0.0006,
+ "loss": 4.335939407348633,
+ "step": 5411
+ },
+ {
+ "epoch": 75.16768558951965,
+ "grad_norm": 0.015766102820634842,
+ "learning_rate": 0.0006,
+ "loss": 4.408674240112305,
+ "step": 5412
+ },
+ {
+ "epoch": 75.18165938864628,
+ "grad_norm": 0.016383284702897072,
+ "learning_rate": 0.0006,
+ "loss": 4.385453224182129,
+ "step": 5413
+ },
+ {
+ "epoch": 75.19563318777293,
+ "grad_norm": 0.013535384088754654,
+ "learning_rate": 0.0006,
+ "loss": 4.292785167694092,
+ "step": 5414
+ },
+ {
+ "epoch": 75.20960698689956,
+ "grad_norm": 0.01354036945849657,
+ "learning_rate": 0.0006,
+ "loss": 4.306252479553223,
+ "step": 5415
+ },
+ {
+ "epoch": 75.2235807860262,
+ "grad_norm": 0.015818078070878983,
+ "learning_rate": 0.0006,
+ "loss": 4.425487995147705,
+ "step": 5416
+ },
+ {
+ "epoch": 75.23755458515284,
+ "grad_norm": 0.014363568276166916,
+ "learning_rate": 0.0006,
+ "loss": 4.395087242126465,
+ "step": 5417
+ },
+ {
+ "epoch": 75.25152838427948,
+ "grad_norm": 0.013895470649003983,
+ "learning_rate": 0.0006,
+ "loss": 4.364090442657471,
+ "step": 5418
+ },
+ {
+ "epoch": 75.26550218340611,
+ "grad_norm": 0.014178668148815632,
+ "learning_rate": 0.0006,
+ "loss": 4.44619083404541,
+ "step": 5419
+ },
+ {
+ "epoch": 75.27947598253274,
+ "grad_norm": 0.0152514623478055,
+ "learning_rate": 0.0006,
+ "loss": 4.348445892333984,
+ "step": 5420
+ },
+ {
+ "epoch": 75.29344978165939,
+ "grad_norm": 0.015397852286696434,
+ "learning_rate": 0.0006,
+ "loss": 4.476832389831543,
+ "step": 5421
+ },
+ {
+ "epoch": 75.30742358078602,
+ "grad_norm": 0.018009863793849945,
+ "learning_rate": 0.0006,
+ "loss": 4.431162357330322,
+ "step": 5422
+ },
+ {
+ "epoch": 75.32139737991267,
+ "grad_norm": 0.017038367688655853,
+ "learning_rate": 0.0006,
+ "loss": 4.474787712097168,
+ "step": 5423
+ },
+ {
+ "epoch": 75.3353711790393,
+ "grad_norm": 0.013763288035988808,
+ "learning_rate": 0.0006,
+ "loss": 4.470419883728027,
+ "step": 5424
+ },
+ {
+ "epoch": 75.34934497816595,
+ "grad_norm": 0.013972110114991665,
+ "learning_rate": 0.0006,
+ "loss": 4.373147964477539,
+ "step": 5425
+ },
+ {
+ "epoch": 75.36331877729258,
+ "grad_norm": 0.017197875306010246,
+ "learning_rate": 0.0006,
+ "loss": 4.494963645935059,
+ "step": 5426
+ },
+ {
+ "epoch": 75.37729257641921,
+ "grad_norm": 0.01687047816812992,
+ "learning_rate": 0.0006,
+ "loss": 4.435934066772461,
+ "step": 5427
+ },
+ {
+ "epoch": 75.39126637554585,
+ "grad_norm": 0.01578272506594658,
+ "learning_rate": 0.0006,
+ "loss": 4.37918758392334,
+ "step": 5428
+ },
+ {
+ "epoch": 75.40524017467249,
+ "grad_norm": 0.015956247225403786,
+ "learning_rate": 0.0006,
+ "loss": 4.455668926239014,
+ "step": 5429
+ },
+ {
+ "epoch": 75.41921397379913,
+ "grad_norm": 0.013495155610144138,
+ "learning_rate": 0.0006,
+ "loss": 4.4271039962768555,
+ "step": 5430
+ },
+ {
+ "epoch": 75.43318777292576,
+ "grad_norm": 0.012851621024310589,
+ "learning_rate": 0.0006,
+ "loss": 4.213107109069824,
+ "step": 5431
+ },
+ {
+ "epoch": 75.44716157205241,
+ "grad_norm": 0.013351158238947392,
+ "learning_rate": 0.0006,
+ "loss": 4.3070783615112305,
+ "step": 5432
+ },
+ {
+ "epoch": 75.46113537117904,
+ "grad_norm": 0.014765217900276184,
+ "learning_rate": 0.0006,
+ "loss": 4.385021209716797,
+ "step": 5433
+ },
+ {
+ "epoch": 75.47510917030567,
+ "grad_norm": 0.016091614961624146,
+ "learning_rate": 0.0006,
+ "loss": 4.312644004821777,
+ "step": 5434
+ },
+ {
+ "epoch": 75.48908296943232,
+ "grad_norm": 0.01612013578414917,
+ "learning_rate": 0.0006,
+ "loss": 4.4482011795043945,
+ "step": 5435
+ },
+ {
+ "epoch": 75.50305676855895,
+ "grad_norm": 0.013157603330910206,
+ "learning_rate": 0.0006,
+ "loss": 4.427262306213379,
+ "step": 5436
+ },
+ {
+ "epoch": 75.5170305676856,
+ "grad_norm": 0.016039809212088585,
+ "learning_rate": 0.0006,
+ "loss": 4.396635055541992,
+ "step": 5437
+ },
+ {
+ "epoch": 75.53100436681223,
+ "grad_norm": 0.01794283092021942,
+ "learning_rate": 0.0006,
+ "loss": 4.476870059967041,
+ "step": 5438
+ },
+ {
+ "epoch": 75.54497816593886,
+ "grad_norm": 0.017030486837029457,
+ "learning_rate": 0.0006,
+ "loss": 4.361156463623047,
+ "step": 5439
+ },
+ {
+ "epoch": 75.5589519650655,
+ "grad_norm": 0.017915161326527596,
+ "learning_rate": 0.0006,
+ "loss": 4.286147117614746,
+ "step": 5440
+ },
+ {
+ "epoch": 75.57292576419214,
+ "grad_norm": 0.017514433711767197,
+ "learning_rate": 0.0006,
+ "loss": 4.394091606140137,
+ "step": 5441
+ },
+ {
+ "epoch": 75.58689956331878,
+ "grad_norm": 0.015128589235246181,
+ "learning_rate": 0.0006,
+ "loss": 4.383126735687256,
+ "step": 5442
+ },
+ {
+ "epoch": 75.60087336244541,
+ "grad_norm": 0.0202616099268198,
+ "learning_rate": 0.0006,
+ "loss": 4.336188316345215,
+ "step": 5443
+ },
+ {
+ "epoch": 75.61484716157206,
+ "grad_norm": 0.019978471100330353,
+ "learning_rate": 0.0006,
+ "loss": 4.461453437805176,
+ "step": 5444
+ },
+ {
+ "epoch": 75.62882096069869,
+ "grad_norm": 0.01738741435110569,
+ "learning_rate": 0.0006,
+ "loss": 4.371197700500488,
+ "step": 5445
+ },
+ {
+ "epoch": 75.64279475982532,
+ "grad_norm": 0.016157809644937515,
+ "learning_rate": 0.0006,
+ "loss": 4.3312835693359375,
+ "step": 5446
+ },
+ {
+ "epoch": 75.65676855895197,
+ "grad_norm": 0.016394613310694695,
+ "learning_rate": 0.0006,
+ "loss": 4.380190849304199,
+ "step": 5447
+ },
+ {
+ "epoch": 75.6707423580786,
+ "grad_norm": 0.014727243222296238,
+ "learning_rate": 0.0006,
+ "loss": 4.308028221130371,
+ "step": 5448
+ },
+ {
+ "epoch": 75.68471615720524,
+ "grad_norm": 0.01695958711206913,
+ "learning_rate": 0.0006,
+ "loss": 4.399503707885742,
+ "step": 5449
+ },
+ {
+ "epoch": 75.69868995633188,
+ "grad_norm": 0.015720795840024948,
+ "learning_rate": 0.0006,
+ "loss": 4.342249870300293,
+ "step": 5450
+ },
+ {
+ "epoch": 75.71266375545852,
+ "grad_norm": 0.01574772410094738,
+ "learning_rate": 0.0006,
+ "loss": 4.331503868103027,
+ "step": 5451
+ },
+ {
+ "epoch": 75.72663755458515,
+ "grad_norm": 0.014812648296356201,
+ "learning_rate": 0.0006,
+ "loss": 4.332413673400879,
+ "step": 5452
+ },
+ {
+ "epoch": 75.74061135371178,
+ "grad_norm": 0.012850426137447357,
+ "learning_rate": 0.0006,
+ "loss": 4.2884063720703125,
+ "step": 5453
+ },
+ {
+ "epoch": 75.75458515283843,
+ "grad_norm": 0.012876970693469048,
+ "learning_rate": 0.0006,
+ "loss": 4.399855613708496,
+ "step": 5454
+ },
+ {
+ "epoch": 75.76855895196506,
+ "grad_norm": 0.012958863750100136,
+ "learning_rate": 0.0006,
+ "loss": 4.411185264587402,
+ "step": 5455
+ },
+ {
+ "epoch": 75.78253275109171,
+ "grad_norm": 0.012197135016322136,
+ "learning_rate": 0.0006,
+ "loss": 4.272603988647461,
+ "step": 5456
+ },
+ {
+ "epoch": 75.79650655021834,
+ "grad_norm": 0.013483566232025623,
+ "learning_rate": 0.0006,
+ "loss": 4.3564348220825195,
+ "step": 5457
+ },
+ {
+ "epoch": 75.81048034934499,
+ "grad_norm": 0.012365429662168026,
+ "learning_rate": 0.0006,
+ "loss": 4.354825973510742,
+ "step": 5458
+ },
+ {
+ "epoch": 75.82445414847162,
+ "grad_norm": 0.012238141149282455,
+ "learning_rate": 0.0006,
+ "loss": 4.440438747406006,
+ "step": 5459
+ },
+ {
+ "epoch": 75.83842794759825,
+ "grad_norm": 0.013380305841565132,
+ "learning_rate": 0.0006,
+ "loss": 4.2813215255737305,
+ "step": 5460
+ },
+ {
+ "epoch": 75.8524017467249,
+ "grad_norm": 0.01443779468536377,
+ "learning_rate": 0.0006,
+ "loss": 4.289078235626221,
+ "step": 5461
+ },
+ {
+ "epoch": 75.86637554585153,
+ "grad_norm": 0.012646574527025223,
+ "learning_rate": 0.0006,
+ "loss": 4.374177932739258,
+ "step": 5462
+ },
+ {
+ "epoch": 75.88034934497817,
+ "grad_norm": 0.012403388507664204,
+ "learning_rate": 0.0006,
+ "loss": 4.343016147613525,
+ "step": 5463
+ },
+ {
+ "epoch": 75.8943231441048,
+ "grad_norm": 0.01319252047687769,
+ "learning_rate": 0.0006,
+ "loss": 4.309957027435303,
+ "step": 5464
+ },
+ {
+ "epoch": 75.90829694323143,
+ "grad_norm": 0.012597428634762764,
+ "learning_rate": 0.0006,
+ "loss": 4.403654098510742,
+ "step": 5465
+ },
+ {
+ "epoch": 75.92227074235808,
+ "grad_norm": 0.012724226340651512,
+ "learning_rate": 0.0006,
+ "loss": 4.326869010925293,
+ "step": 5466
+ },
+ {
+ "epoch": 75.93624454148471,
+ "grad_norm": 0.012698772363364697,
+ "learning_rate": 0.0006,
+ "loss": 4.268430233001709,
+ "step": 5467
+ },
+ {
+ "epoch": 75.95021834061136,
+ "grad_norm": 0.01376490481197834,
+ "learning_rate": 0.0006,
+ "loss": 4.42446231842041,
+ "step": 5468
+ },
+ {
+ "epoch": 75.96419213973799,
+ "grad_norm": 0.013175144791603088,
+ "learning_rate": 0.0006,
+ "loss": 4.374675750732422,
+ "step": 5469
+ },
+ {
+ "epoch": 75.97816593886463,
+ "grad_norm": 0.012731322087347507,
+ "learning_rate": 0.0006,
+ "loss": 4.372966766357422,
+ "step": 5470
+ },
+ {
+ "epoch": 75.99213973799127,
+ "grad_norm": 0.01360505260527134,
+ "learning_rate": 0.0006,
+ "loss": 4.3095903396606445,
+ "step": 5471
+ },
+ {
+ "epoch": 76.0,
+ "grad_norm": 0.014011901803314686,
+ "learning_rate": 0.0006,
+ "loss": 4.26954460144043,
+ "step": 5472
+ },
+ {
+ "epoch": 76.0,
+ "eval_loss": 4.689640998840332,
+ "eval_runtime": 56.9095,
+ "eval_samples_per_second": 42.91,
+ "eval_steps_per_second": 1.353,
+ "step": 5472
+ },
+ {
+ "epoch": 76.01397379912663,
+ "grad_norm": 0.015351055189967155,
+ "learning_rate": 0.0006,
+ "loss": 4.348268985748291,
+ "step": 5473
+ },
+ {
+ "epoch": 76.02794759825328,
+ "grad_norm": 0.017164217308163643,
+ "learning_rate": 0.0006,
+ "loss": 4.244204998016357,
+ "step": 5474
+ },
+ {
+ "epoch": 76.04192139737991,
+ "grad_norm": 0.021624932065606117,
+ "learning_rate": 0.0006,
+ "loss": 4.311168670654297,
+ "step": 5475
+ },
+ {
+ "epoch": 76.05589519650655,
+ "grad_norm": 0.028712468221783638,
+ "learning_rate": 0.0006,
+ "loss": 4.322287082672119,
+ "step": 5476
+ },
+ {
+ "epoch": 76.06986899563319,
+ "grad_norm": 0.027915487065911293,
+ "learning_rate": 0.0006,
+ "loss": 4.381207466125488,
+ "step": 5477
+ },
+ {
+ "epoch": 76.08384279475983,
+ "grad_norm": 0.017955303192138672,
+ "learning_rate": 0.0006,
+ "loss": 4.2635016441345215,
+ "step": 5478
+ },
+ {
+ "epoch": 76.09781659388646,
+ "grad_norm": 0.019113514572381973,
+ "learning_rate": 0.0006,
+ "loss": 4.399471282958984,
+ "step": 5479
+ },
+ {
+ "epoch": 76.1117903930131,
+ "grad_norm": 0.017502542585134506,
+ "learning_rate": 0.0006,
+ "loss": 4.321204662322998,
+ "step": 5480
+ },
+ {
+ "epoch": 76.12576419213974,
+ "grad_norm": 0.01852525770664215,
+ "learning_rate": 0.0006,
+ "loss": 4.217884063720703,
+ "step": 5481
+ },
+ {
+ "epoch": 76.13973799126637,
+ "grad_norm": 0.019836828112602234,
+ "learning_rate": 0.0006,
+ "loss": 4.278532028198242,
+ "step": 5482
+ },
+ {
+ "epoch": 76.15371179039302,
+ "grad_norm": 0.02101975679397583,
+ "learning_rate": 0.0006,
+ "loss": 4.246720314025879,
+ "step": 5483
+ },
+ {
+ "epoch": 76.16768558951965,
+ "grad_norm": 0.021202990785241127,
+ "learning_rate": 0.0006,
+ "loss": 4.517517566680908,
+ "step": 5484
+ },
+ {
+ "epoch": 76.18165938864628,
+ "grad_norm": 0.021428769454360008,
+ "learning_rate": 0.0006,
+ "loss": 4.3373517990112305,
+ "step": 5485
+ },
+ {
+ "epoch": 76.19563318777293,
+ "grad_norm": 0.018263190984725952,
+ "learning_rate": 0.0006,
+ "loss": 4.402856826782227,
+ "step": 5486
+ },
+ {
+ "epoch": 76.20960698689956,
+ "grad_norm": 0.019025536254048347,
+ "learning_rate": 0.0006,
+ "loss": 4.314227104187012,
+ "step": 5487
+ },
+ {
+ "epoch": 76.2235807860262,
+ "grad_norm": 0.020918630063533783,
+ "learning_rate": 0.0006,
+ "loss": 4.454369068145752,
+ "step": 5488
+ },
+ {
+ "epoch": 76.23755458515284,
+ "grad_norm": 0.019307617098093033,
+ "learning_rate": 0.0006,
+ "loss": 4.247509002685547,
+ "step": 5489
+ },
+ {
+ "epoch": 76.25152838427948,
+ "grad_norm": 0.014956331811845303,
+ "learning_rate": 0.0006,
+ "loss": 4.365175247192383,
+ "step": 5490
+ },
+ {
+ "epoch": 76.26550218340611,
+ "grad_norm": 0.017213059589266777,
+ "learning_rate": 0.0006,
+ "loss": 4.384005069732666,
+ "step": 5491
+ },
+ {
+ "epoch": 76.27947598253274,
+ "grad_norm": 0.017575597390532494,
+ "learning_rate": 0.0006,
+ "loss": 4.41934871673584,
+ "step": 5492
+ },
+ {
+ "epoch": 76.29344978165939,
+ "grad_norm": 0.015879256650805473,
+ "learning_rate": 0.0006,
+ "loss": 4.26675271987915,
+ "step": 5493
+ },
+ {
+ "epoch": 76.30742358078602,
+ "grad_norm": 0.01564522460103035,
+ "learning_rate": 0.0006,
+ "loss": 4.277339935302734,
+ "step": 5494
+ },
+ {
+ "epoch": 76.32139737991267,
+ "grad_norm": 0.016414586454629898,
+ "learning_rate": 0.0006,
+ "loss": 4.4138593673706055,
+ "step": 5495
+ },
+ {
+ "epoch": 76.3353711790393,
+ "grad_norm": 0.014732337556779385,
+ "learning_rate": 0.0006,
+ "loss": 4.281006336212158,
+ "step": 5496
+ },
+ {
+ "epoch": 76.34934497816595,
+ "grad_norm": 0.015461455099284649,
+ "learning_rate": 0.0006,
+ "loss": 4.263422966003418,
+ "step": 5497
+ },
+ {
+ "epoch": 76.36331877729258,
+ "grad_norm": 0.014529814012348652,
+ "learning_rate": 0.0006,
+ "loss": 4.36165189743042,
+ "step": 5498
+ },
+ {
+ "epoch": 76.37729257641921,
+ "grad_norm": 0.01469483319669962,
+ "learning_rate": 0.0006,
+ "loss": 4.309642791748047,
+ "step": 5499
+ },
+ {
+ "epoch": 76.39126637554585,
+ "grad_norm": 0.016490574926137924,
+ "learning_rate": 0.0006,
+ "loss": 4.288150787353516,
+ "step": 5500
+ },
+ {
+ "epoch": 76.40524017467249,
+ "grad_norm": 0.014593565836548805,
+ "learning_rate": 0.0006,
+ "loss": 4.271614074707031,
+ "step": 5501
+ },
+ {
+ "epoch": 76.41921397379913,
+ "grad_norm": 0.012796535156667233,
+ "learning_rate": 0.0006,
+ "loss": 4.420042037963867,
+ "step": 5502
+ },
+ {
+ "epoch": 76.43318777292576,
+ "grad_norm": 0.016086025163531303,
+ "learning_rate": 0.0006,
+ "loss": 4.289943695068359,
+ "step": 5503
+ },
+ {
+ "epoch": 76.44716157205241,
+ "grad_norm": 0.015439470298588276,
+ "learning_rate": 0.0006,
+ "loss": 4.3023271560668945,
+ "step": 5504
+ },
+ {
+ "epoch": 76.46113537117904,
+ "grad_norm": 0.015066367574036121,
+ "learning_rate": 0.0006,
+ "loss": 4.197081565856934,
+ "step": 5505
+ },
+ {
+ "epoch": 76.47510917030567,
+ "grad_norm": 0.016653263941407204,
+ "learning_rate": 0.0006,
+ "loss": 4.333785057067871,
+ "step": 5506
+ },
+ {
+ "epoch": 76.48908296943232,
+ "grad_norm": 0.014791973866522312,
+ "learning_rate": 0.0006,
+ "loss": 4.323811054229736,
+ "step": 5507
+ },
+ {
+ "epoch": 76.50305676855895,
+ "grad_norm": 0.014694827608764172,
+ "learning_rate": 0.0006,
+ "loss": 4.195852279663086,
+ "step": 5508
+ },
+ {
+ "epoch": 76.5170305676856,
+ "grad_norm": 0.015120396390557289,
+ "learning_rate": 0.0006,
+ "loss": 4.42379903793335,
+ "step": 5509
+ },
+ {
+ "epoch": 76.53100436681223,
+ "grad_norm": 0.014584025368094444,
+ "learning_rate": 0.0006,
+ "loss": 4.2702531814575195,
+ "step": 5510
+ },
+ {
+ "epoch": 76.54497816593886,
+ "grad_norm": 0.013981613330543041,
+ "learning_rate": 0.0006,
+ "loss": 4.383029937744141,
+ "step": 5511
+ },
+ {
+ "epoch": 76.5589519650655,
+ "grad_norm": 0.012930642813444138,
+ "learning_rate": 0.0006,
+ "loss": 4.296136856079102,
+ "step": 5512
+ },
+ {
+ "epoch": 76.57292576419214,
+ "grad_norm": 0.013060989789664745,
+ "learning_rate": 0.0006,
+ "loss": 4.346380233764648,
+ "step": 5513
+ },
+ {
+ "epoch": 76.58689956331878,
+ "grad_norm": 0.013209798373281956,
+ "learning_rate": 0.0006,
+ "loss": 4.316067218780518,
+ "step": 5514
+ },
+ {
+ "epoch": 76.60087336244541,
+ "grad_norm": 0.012423216365277767,
+ "learning_rate": 0.0006,
+ "loss": 4.370274543762207,
+ "step": 5515
+ },
+ {
+ "epoch": 76.61484716157206,
+ "grad_norm": 0.013205411843955517,
+ "learning_rate": 0.0006,
+ "loss": 4.429969310760498,
+ "step": 5516
+ },
+ {
+ "epoch": 76.62882096069869,
+ "grad_norm": 0.013049827888607979,
+ "learning_rate": 0.0006,
+ "loss": 4.378549098968506,
+ "step": 5517
+ },
+ {
+ "epoch": 76.64279475982532,
+ "grad_norm": 0.012075880542397499,
+ "learning_rate": 0.0006,
+ "loss": 4.348293304443359,
+ "step": 5518
+ },
+ {
+ "epoch": 76.65676855895197,
+ "grad_norm": 0.014932847581803799,
+ "learning_rate": 0.0006,
+ "loss": 4.282538890838623,
+ "step": 5519
+ },
+ {
+ "epoch": 76.6707423580786,
+ "grad_norm": 0.015615998767316341,
+ "learning_rate": 0.0006,
+ "loss": 4.274641036987305,
+ "step": 5520
+ },
+ {
+ "epoch": 76.68471615720524,
+ "grad_norm": 0.01424513477832079,
+ "learning_rate": 0.0006,
+ "loss": 4.243166923522949,
+ "step": 5521
+ },
+ {
+ "epoch": 76.69868995633188,
+ "grad_norm": 0.012775546871125698,
+ "learning_rate": 0.0006,
+ "loss": 4.335611820220947,
+ "step": 5522
+ },
+ {
+ "epoch": 76.71266375545852,
+ "grad_norm": 0.014461303129792213,
+ "learning_rate": 0.0006,
+ "loss": 4.246976375579834,
+ "step": 5523
+ },
+ {
+ "epoch": 76.72663755458515,
+ "grad_norm": 0.0145436255261302,
+ "learning_rate": 0.0006,
+ "loss": 4.275895118713379,
+ "step": 5524
+ },
+ {
+ "epoch": 76.74061135371178,
+ "grad_norm": 0.014055085368454456,
+ "learning_rate": 0.0006,
+ "loss": 4.331108093261719,
+ "step": 5525
+ },
+ {
+ "epoch": 76.75458515283843,
+ "grad_norm": 0.011933894827961922,
+ "learning_rate": 0.0006,
+ "loss": 4.293721675872803,
+ "step": 5526
+ },
+ {
+ "epoch": 76.76855895196506,
+ "grad_norm": 0.01277392078191042,
+ "learning_rate": 0.0006,
+ "loss": 4.312374114990234,
+ "step": 5527
+ },
+ {
+ "epoch": 76.78253275109171,
+ "grad_norm": 0.014227936044335365,
+ "learning_rate": 0.0006,
+ "loss": 4.238729000091553,
+ "step": 5528
+ },
+ {
+ "epoch": 76.79650655021834,
+ "grad_norm": 0.01488024927675724,
+ "learning_rate": 0.0006,
+ "loss": 4.24587345123291,
+ "step": 5529
+ },
+ {
+ "epoch": 76.81048034934499,
+ "grad_norm": 0.015341250225901604,
+ "learning_rate": 0.0006,
+ "loss": 4.257378578186035,
+ "step": 5530
+ },
+ {
+ "epoch": 76.82445414847162,
+ "grad_norm": 0.015179437585175037,
+ "learning_rate": 0.0006,
+ "loss": 4.2812910079956055,
+ "step": 5531
+ },
+ {
+ "epoch": 76.83842794759825,
+ "grad_norm": 0.012814724817872047,
+ "learning_rate": 0.0006,
+ "loss": 4.2309465408325195,
+ "step": 5532
+ },
+ {
+ "epoch": 76.8524017467249,
+ "grad_norm": 0.013810716569423676,
+ "learning_rate": 0.0006,
+ "loss": 4.325292110443115,
+ "step": 5533
+ },
+ {
+ "epoch": 76.86637554585153,
+ "grad_norm": 0.017468232661485672,
+ "learning_rate": 0.0006,
+ "loss": 4.243788719177246,
+ "step": 5534
+ },
+ {
+ "epoch": 76.88034934497817,
+ "grad_norm": 0.01739857904613018,
+ "learning_rate": 0.0006,
+ "loss": 4.345390796661377,
+ "step": 5535
+ },
+ {
+ "epoch": 76.8943231441048,
+ "grad_norm": 0.015329061076045036,
+ "learning_rate": 0.0006,
+ "loss": 4.415142059326172,
+ "step": 5536
+ },
+ {
+ "epoch": 76.90829694323143,
+ "grad_norm": 0.01598426327109337,
+ "learning_rate": 0.0006,
+ "loss": 4.299558639526367,
+ "step": 5537
+ },
+ {
+ "epoch": 76.92227074235808,
+ "grad_norm": 0.01621454581618309,
+ "learning_rate": 0.0006,
+ "loss": 4.284354209899902,
+ "step": 5538
+ },
+ {
+ "epoch": 76.93624454148471,
+ "grad_norm": 0.015956968069076538,
+ "learning_rate": 0.0006,
+ "loss": 4.293837547302246,
+ "step": 5539
+ },
+ {
+ "epoch": 76.95021834061136,
+ "grad_norm": 0.01776493526995182,
+ "learning_rate": 0.0006,
+ "loss": 4.229729175567627,
+ "step": 5540
+ },
+ {
+ "epoch": 76.96419213973799,
+ "grad_norm": 0.015317477285861969,
+ "learning_rate": 0.0006,
+ "loss": 4.3328447341918945,
+ "step": 5541
+ },
+ {
+ "epoch": 76.97816593886463,
+ "grad_norm": 0.015140251256525517,
+ "learning_rate": 0.0006,
+ "loss": 4.310840606689453,
+ "step": 5542
+ },
+ {
+ "epoch": 76.99213973799127,
+ "grad_norm": 0.015326812863349915,
+ "learning_rate": 0.0006,
+ "loss": 4.230832099914551,
+ "step": 5543
+ },
+ {
+ "epoch": 77.0,
+ "grad_norm": 0.01607763208448887,
+ "learning_rate": 0.0006,
+ "loss": 4.2236328125,
+ "step": 5544
+ },
+ {
+ "epoch": 77.0,
+ "eval_loss": 4.662221431732178,
+ "eval_runtime": 56.5194,
+ "eval_samples_per_second": 43.206,
+ "eval_steps_per_second": 1.362,
+ "step": 5544
+ },
+ {
+ "epoch": 77.01397379912663,
+ "grad_norm": 0.014839510433375835,
+ "learning_rate": 0.0006,
+ "loss": 4.254111289978027,
+ "step": 5545
+ },
+ {
+ "epoch": 77.02794759825328,
+ "grad_norm": 0.01519741676747799,
+ "learning_rate": 0.0006,
+ "loss": 4.1607184410095215,
+ "step": 5546
+ },
+ {
+ "epoch": 77.04192139737991,
+ "grad_norm": 0.015472603030502796,
+ "learning_rate": 0.0006,
+ "loss": 4.323320388793945,
+ "step": 5547
+ },
+ {
+ "epoch": 77.05589519650655,
+ "grad_norm": 0.017937948927283287,
+ "learning_rate": 0.0006,
+ "loss": 4.308490753173828,
+ "step": 5548
+ },
+ {
+ "epoch": 77.06986899563319,
+ "grad_norm": 0.020890841260552406,
+ "learning_rate": 0.0006,
+ "loss": 4.259101867675781,
+ "step": 5549
+ },
+ {
+ "epoch": 77.08384279475983,
+ "grad_norm": 0.01892857998609543,
+ "learning_rate": 0.0006,
+ "loss": 4.386970520019531,
+ "step": 5550
+ },
+ {
+ "epoch": 77.09781659388646,
+ "grad_norm": 0.016583630815148354,
+ "learning_rate": 0.0006,
+ "loss": 4.233604907989502,
+ "step": 5551
+ },
+ {
+ "epoch": 77.1117903930131,
+ "grad_norm": 0.016063468530774117,
+ "learning_rate": 0.0006,
+ "loss": 4.339544296264648,
+ "step": 5552
+ },
+ {
+ "epoch": 77.12576419213974,
+ "grad_norm": 0.018326779827475548,
+ "learning_rate": 0.0006,
+ "loss": 4.367204666137695,
+ "step": 5553
+ },
+ {
+ "epoch": 77.13973799126637,
+ "grad_norm": 0.021590933203697205,
+ "learning_rate": 0.0006,
+ "loss": 4.352685928344727,
+ "step": 5554
+ },
+ {
+ "epoch": 77.15371179039302,
+ "grad_norm": 0.019643642008304596,
+ "learning_rate": 0.0006,
+ "loss": 4.258535385131836,
+ "step": 5555
+ },
+ {
+ "epoch": 77.16768558951965,
+ "grad_norm": 0.018476149067282677,
+ "learning_rate": 0.0006,
+ "loss": 4.232568740844727,
+ "step": 5556
+ },
+ {
+ "epoch": 77.18165938864628,
+ "grad_norm": 0.01584405079483986,
+ "learning_rate": 0.0006,
+ "loss": 4.258371353149414,
+ "step": 5557
+ },
+ {
+ "epoch": 77.19563318777293,
+ "grad_norm": 0.01575201191008091,
+ "learning_rate": 0.0006,
+ "loss": 4.183218002319336,
+ "step": 5558
+ },
+ {
+ "epoch": 77.20960698689956,
+ "grad_norm": 0.016871824860572815,
+ "learning_rate": 0.0006,
+ "loss": 4.194563865661621,
+ "step": 5559
+ },
+ {
+ "epoch": 77.2235807860262,
+ "grad_norm": 0.02059333398938179,
+ "learning_rate": 0.0006,
+ "loss": 4.195281982421875,
+ "step": 5560
+ },
+ {
+ "epoch": 77.23755458515284,
+ "grad_norm": 0.022640453651547432,
+ "learning_rate": 0.0006,
+ "loss": 4.279615879058838,
+ "step": 5561
+ },
+ {
+ "epoch": 77.25152838427948,
+ "grad_norm": 0.020335959270596504,
+ "learning_rate": 0.0006,
+ "loss": 4.25040864944458,
+ "step": 5562
+ },
+ {
+ "epoch": 77.26550218340611,
+ "grad_norm": 0.020010793581604958,
+ "learning_rate": 0.0006,
+ "loss": 4.291116714477539,
+ "step": 5563
+ },
+ {
+ "epoch": 77.27947598253274,
+ "grad_norm": 0.02109045907855034,
+ "learning_rate": 0.0006,
+ "loss": 4.364088535308838,
+ "step": 5564
+ },
+ {
+ "epoch": 77.29344978165939,
+ "grad_norm": 0.019519446417689323,
+ "learning_rate": 0.0006,
+ "loss": 4.329166412353516,
+ "step": 5565
+ },
+ {
+ "epoch": 77.30742358078602,
+ "grad_norm": 0.01689804159104824,
+ "learning_rate": 0.0006,
+ "loss": 4.244176864624023,
+ "step": 5566
+ },
+ {
+ "epoch": 77.32139737991267,
+ "grad_norm": 0.01893681474030018,
+ "learning_rate": 0.0006,
+ "loss": 4.239068984985352,
+ "step": 5567
+ },
+ {
+ "epoch": 77.3353711790393,
+ "grad_norm": 0.015998193994164467,
+ "learning_rate": 0.0006,
+ "loss": 4.173398017883301,
+ "step": 5568
+ },
+ {
+ "epoch": 77.34934497816595,
+ "grad_norm": 0.013892280869185925,
+ "learning_rate": 0.0006,
+ "loss": 4.287654876708984,
+ "step": 5569
+ },
+ {
+ "epoch": 77.36331877729258,
+ "grad_norm": 0.017060643061995506,
+ "learning_rate": 0.0006,
+ "loss": 4.398447036743164,
+ "step": 5570
+ },
+ {
+ "epoch": 77.37729257641921,
+ "grad_norm": 0.015168849378824234,
+ "learning_rate": 0.0006,
+ "loss": 4.15665864944458,
+ "step": 5571
+ },
+ {
+ "epoch": 77.39126637554585,
+ "grad_norm": 0.014886523596942425,
+ "learning_rate": 0.0006,
+ "loss": 4.196673393249512,
+ "step": 5572
+ },
+ {
+ "epoch": 77.40524017467249,
+ "grad_norm": 0.016941796988248825,
+ "learning_rate": 0.0006,
+ "loss": 4.314562797546387,
+ "step": 5573
+ },
+ {
+ "epoch": 77.41921397379913,
+ "grad_norm": 0.015353812836110592,
+ "learning_rate": 0.0006,
+ "loss": 4.409579277038574,
+ "step": 5574
+ },
+ {
+ "epoch": 77.43318777292576,
+ "grad_norm": 0.015192852355539799,
+ "learning_rate": 0.0006,
+ "loss": 4.392045974731445,
+ "step": 5575
+ },
+ {
+ "epoch": 77.44716157205241,
+ "grad_norm": 0.014232569374144077,
+ "learning_rate": 0.0006,
+ "loss": 4.21639347076416,
+ "step": 5576
+ },
+ {
+ "epoch": 77.46113537117904,
+ "grad_norm": 0.014453226700425148,
+ "learning_rate": 0.0006,
+ "loss": 4.2802276611328125,
+ "step": 5577
+ },
+ {
+ "epoch": 77.47510917030567,
+ "grad_norm": 0.012881236150860786,
+ "learning_rate": 0.0006,
+ "loss": 4.375980854034424,
+ "step": 5578
+ },
+ {
+ "epoch": 77.48908296943232,
+ "grad_norm": 0.01583976298570633,
+ "learning_rate": 0.0006,
+ "loss": 4.264520645141602,
+ "step": 5579
+ },
+ {
+ "epoch": 77.50305676855895,
+ "grad_norm": 0.014171584509313107,
+ "learning_rate": 0.0006,
+ "loss": 4.225279808044434,
+ "step": 5580
+ },
+ {
+ "epoch": 77.5170305676856,
+ "grad_norm": 0.014715159311890602,
+ "learning_rate": 0.0006,
+ "loss": 4.369834899902344,
+ "step": 5581
+ },
+ {
+ "epoch": 77.53100436681223,
+ "grad_norm": 0.016554079949855804,
+ "learning_rate": 0.0006,
+ "loss": 4.2976179122924805,
+ "step": 5582
+ },
+ {
+ "epoch": 77.54497816593886,
+ "grad_norm": 0.015095747075974941,
+ "learning_rate": 0.0006,
+ "loss": 4.294407844543457,
+ "step": 5583
+ },
+ {
+ "epoch": 77.5589519650655,
+ "grad_norm": 0.012726361863315105,
+ "learning_rate": 0.0006,
+ "loss": 4.221401214599609,
+ "step": 5584
+ },
+ {
+ "epoch": 77.57292576419214,
+ "grad_norm": 0.014999309554696083,
+ "learning_rate": 0.0006,
+ "loss": 4.243929386138916,
+ "step": 5585
+ },
+ {
+ "epoch": 77.58689956331878,
+ "grad_norm": 0.016526037827134132,
+ "learning_rate": 0.0006,
+ "loss": 4.374139785766602,
+ "step": 5586
+ },
+ {
+ "epoch": 77.60087336244541,
+ "grad_norm": 0.01648339442908764,
+ "learning_rate": 0.0006,
+ "loss": 4.392425060272217,
+ "step": 5587
+ },
+ {
+ "epoch": 77.61484716157206,
+ "grad_norm": 0.015269213356077671,
+ "learning_rate": 0.0006,
+ "loss": 4.33986759185791,
+ "step": 5588
+ },
+ {
+ "epoch": 77.62882096069869,
+ "grad_norm": 0.01690507121384144,
+ "learning_rate": 0.0006,
+ "loss": 4.307476043701172,
+ "step": 5589
+ },
+ {
+ "epoch": 77.64279475982532,
+ "grad_norm": 0.016438685357570648,
+ "learning_rate": 0.0006,
+ "loss": 4.2406206130981445,
+ "step": 5590
+ },
+ {
+ "epoch": 77.65676855895197,
+ "grad_norm": 0.0157600287348032,
+ "learning_rate": 0.0006,
+ "loss": 4.314070224761963,
+ "step": 5591
+ },
+ {
+ "epoch": 77.6707423580786,
+ "grad_norm": 0.015127947553992271,
+ "learning_rate": 0.0006,
+ "loss": 4.22475528717041,
+ "step": 5592
+ },
+ {
+ "epoch": 77.68471615720524,
+ "grad_norm": 0.016632677987217903,
+ "learning_rate": 0.0006,
+ "loss": 4.26954984664917,
+ "step": 5593
+ },
+ {
+ "epoch": 77.69868995633188,
+ "grad_norm": 0.013150460086762905,
+ "learning_rate": 0.0006,
+ "loss": 4.369131088256836,
+ "step": 5594
+ },
+ {
+ "epoch": 77.71266375545852,
+ "grad_norm": 0.014513005502521992,
+ "learning_rate": 0.0006,
+ "loss": 4.3248796463012695,
+ "step": 5595
+ },
+ {
+ "epoch": 77.72663755458515,
+ "grad_norm": 0.014090991578996181,
+ "learning_rate": 0.0006,
+ "loss": 4.268381118774414,
+ "step": 5596
+ },
+ {
+ "epoch": 77.74061135371178,
+ "grad_norm": 0.01633097417652607,
+ "learning_rate": 0.0006,
+ "loss": 4.238439083099365,
+ "step": 5597
+ },
+ {
+ "epoch": 77.75458515283843,
+ "grad_norm": 0.01710933819413185,
+ "learning_rate": 0.0006,
+ "loss": 4.399697780609131,
+ "step": 5598
+ },
+ {
+ "epoch": 77.76855895196506,
+ "grad_norm": 0.019251767545938492,
+ "learning_rate": 0.0006,
+ "loss": 4.290450096130371,
+ "step": 5599
+ },
+ {
+ "epoch": 77.78253275109171,
+ "grad_norm": 0.01849171705543995,
+ "learning_rate": 0.0006,
+ "loss": 4.267734527587891,
+ "step": 5600
+ },
+ {
+ "epoch": 77.79650655021834,
+ "grad_norm": 0.014173115603625774,
+ "learning_rate": 0.0006,
+ "loss": 4.288799285888672,
+ "step": 5601
+ },
+ {
+ "epoch": 77.81048034934499,
+ "grad_norm": 0.014572718180716038,
+ "learning_rate": 0.0006,
+ "loss": 4.3790435791015625,
+ "step": 5602
+ },
+ {
+ "epoch": 77.82445414847162,
+ "grad_norm": 0.016375336796045303,
+ "learning_rate": 0.0006,
+ "loss": 4.231725692749023,
+ "step": 5603
+ },
+ {
+ "epoch": 77.83842794759825,
+ "grad_norm": 0.017326701432466507,
+ "learning_rate": 0.0006,
+ "loss": 4.211495399475098,
+ "step": 5604
+ },
+ {
+ "epoch": 77.8524017467249,
+ "grad_norm": 0.016399390995502472,
+ "learning_rate": 0.0006,
+ "loss": 4.386699199676514,
+ "step": 5605
+ },
+ {
+ "epoch": 77.86637554585153,
+ "grad_norm": 0.018269989639520645,
+ "learning_rate": 0.0006,
+ "loss": 4.254856109619141,
+ "step": 5606
+ },
+ {
+ "epoch": 77.88034934497817,
+ "grad_norm": 0.018443435430526733,
+ "learning_rate": 0.0006,
+ "loss": 4.391961097717285,
+ "step": 5607
+ },
+ {
+ "epoch": 77.8943231441048,
+ "grad_norm": 0.01784636080265045,
+ "learning_rate": 0.0006,
+ "loss": 4.273881912231445,
+ "step": 5608
+ },
+ {
+ "epoch": 77.90829694323143,
+ "grad_norm": 0.01782994158565998,
+ "learning_rate": 0.0006,
+ "loss": 4.26833963394165,
+ "step": 5609
+ },
+ {
+ "epoch": 77.92227074235808,
+ "grad_norm": 0.016260741278529167,
+ "learning_rate": 0.0006,
+ "loss": 4.306807994842529,
+ "step": 5610
+ },
+ {
+ "epoch": 77.93624454148471,
+ "grad_norm": 0.015482903458178043,
+ "learning_rate": 0.0006,
+ "loss": 4.187826633453369,
+ "step": 5611
+ },
+ {
+ "epoch": 77.95021834061136,
+ "grad_norm": 0.015193257480859756,
+ "learning_rate": 0.0006,
+ "loss": 4.286060333251953,
+ "step": 5612
+ },
+ {
+ "epoch": 77.96419213973799,
+ "grad_norm": 0.015110115520656109,
+ "learning_rate": 0.0006,
+ "loss": 4.2828192710876465,
+ "step": 5613
+ },
+ {
+ "epoch": 77.97816593886463,
+ "grad_norm": 0.015973486006259918,
+ "learning_rate": 0.0006,
+ "loss": 4.317939758300781,
+ "step": 5614
+ },
+ {
+ "epoch": 77.99213973799127,
+ "grad_norm": 0.017363594844937325,
+ "learning_rate": 0.0006,
+ "loss": 4.287721633911133,
+ "step": 5615
+ },
+ {
+ "epoch": 78.0,
+ "grad_norm": 0.015277642756700516,
+ "learning_rate": 0.0006,
+ "loss": 4.356618881225586,
+ "step": 5616
+ },
+ {
+ "epoch": 78.0,
+ "eval_loss": 4.642678737640381,
+ "eval_runtime": 56.7335,
+ "eval_samples_per_second": 43.043,
+ "eval_steps_per_second": 1.357,
+ "step": 5616
+ },
+ {
+ "epoch": 78.01397379912663,
+ "grad_norm": 0.016794051975011826,
+ "learning_rate": 0.0006,
+ "loss": 4.279930591583252,
+ "step": 5617
+ },
+ {
+ "epoch": 78.02794759825328,
+ "grad_norm": 0.01632673107087612,
+ "learning_rate": 0.0006,
+ "loss": 4.235943794250488,
+ "step": 5618
+ },
+ {
+ "epoch": 78.04192139737991,
+ "grad_norm": 0.01437876746058464,
+ "learning_rate": 0.0006,
+ "loss": 4.2542243003845215,
+ "step": 5619
+ },
+ {
+ "epoch": 78.05589519650655,
+ "grad_norm": 0.014993119053542614,
+ "learning_rate": 0.0006,
+ "loss": 4.262046813964844,
+ "step": 5620
+ },
+ {
+ "epoch": 78.06986899563319,
+ "grad_norm": 0.015721119940280914,
+ "learning_rate": 0.0006,
+ "loss": 4.173436164855957,
+ "step": 5621
+ },
+ {
+ "epoch": 78.08384279475983,
+ "grad_norm": 0.01334085687994957,
+ "learning_rate": 0.0006,
+ "loss": 4.322175979614258,
+ "step": 5622
+ },
+ {
+ "epoch": 78.09781659388646,
+ "grad_norm": 0.015561388805508614,
+ "learning_rate": 0.0006,
+ "loss": 4.242125988006592,
+ "step": 5623
+ },
+ {
+ "epoch": 78.1117903930131,
+ "grad_norm": 0.014418602921068668,
+ "learning_rate": 0.0006,
+ "loss": 4.266186714172363,
+ "step": 5624
+ },
+ {
+ "epoch": 78.12576419213974,
+ "grad_norm": 0.013739909045398235,
+ "learning_rate": 0.0006,
+ "loss": 4.231633186340332,
+ "step": 5625
+ },
+ {
+ "epoch": 78.13973799126637,
+ "grad_norm": 0.01402067206799984,
+ "learning_rate": 0.0006,
+ "loss": 4.188406944274902,
+ "step": 5626
+ },
+ {
+ "epoch": 78.15371179039302,
+ "grad_norm": 0.013639481738209724,
+ "learning_rate": 0.0006,
+ "loss": 4.1586594581604,
+ "step": 5627
+ },
+ {
+ "epoch": 78.16768558951965,
+ "grad_norm": 0.012858159840106964,
+ "learning_rate": 0.0006,
+ "loss": 4.360448360443115,
+ "step": 5628
+ },
+ {
+ "epoch": 78.18165938864628,
+ "grad_norm": 0.013089739717543125,
+ "learning_rate": 0.0006,
+ "loss": 4.344371795654297,
+ "step": 5629
+ },
+ {
+ "epoch": 78.19563318777293,
+ "grad_norm": 0.01382314320653677,
+ "learning_rate": 0.0006,
+ "loss": 4.20378303527832,
+ "step": 5630
+ },
+ {
+ "epoch": 78.20960698689956,
+ "grad_norm": 0.014206619001924992,
+ "learning_rate": 0.0006,
+ "loss": 4.239891529083252,
+ "step": 5631
+ },
+ {
+ "epoch": 78.2235807860262,
+ "grad_norm": 0.01472168043255806,
+ "learning_rate": 0.0006,
+ "loss": 4.282382488250732,
+ "step": 5632
+ },
+ {
+ "epoch": 78.23755458515284,
+ "grad_norm": 0.014237025752663612,
+ "learning_rate": 0.0006,
+ "loss": 4.241545677185059,
+ "step": 5633
+ },
+ {
+ "epoch": 78.25152838427948,
+ "grad_norm": 0.013231472112238407,
+ "learning_rate": 0.0006,
+ "loss": 4.184896469116211,
+ "step": 5634
+ },
+ {
+ "epoch": 78.26550218340611,
+ "grad_norm": 0.013538471423089504,
+ "learning_rate": 0.0006,
+ "loss": 4.325181007385254,
+ "step": 5635
+ },
+ {
+ "epoch": 78.27947598253274,
+ "grad_norm": 0.013541961088776588,
+ "learning_rate": 0.0006,
+ "loss": 4.368722915649414,
+ "step": 5636
+ },
+ {
+ "epoch": 78.29344978165939,
+ "grad_norm": 0.013446723110973835,
+ "learning_rate": 0.0006,
+ "loss": 4.301422595977783,
+ "step": 5637
+ },
+ {
+ "epoch": 78.30742358078602,
+ "grad_norm": 0.014753809198737144,
+ "learning_rate": 0.0006,
+ "loss": 4.259127616882324,
+ "step": 5638
+ },
+ {
+ "epoch": 78.32139737991267,
+ "grad_norm": 0.013256008736789227,
+ "learning_rate": 0.0006,
+ "loss": 4.284032821655273,
+ "step": 5639
+ },
+ {
+ "epoch": 78.3353711790393,
+ "grad_norm": 0.014272629283368587,
+ "learning_rate": 0.0006,
+ "loss": 4.234989166259766,
+ "step": 5640
+ },
+ {
+ "epoch": 78.34934497816595,
+ "grad_norm": 0.014756826683878899,
+ "learning_rate": 0.0006,
+ "loss": 4.143499374389648,
+ "step": 5641
+ },
+ {
+ "epoch": 78.36331877729258,
+ "grad_norm": 0.016549205407500267,
+ "learning_rate": 0.0006,
+ "loss": 4.311034202575684,
+ "step": 5642
+ },
+ {
+ "epoch": 78.37729257641921,
+ "grad_norm": 0.017735283821821213,
+ "learning_rate": 0.0006,
+ "loss": 4.246767520904541,
+ "step": 5643
+ },
+ {
+ "epoch": 78.39126637554585,
+ "grad_norm": 0.017370613291859627,
+ "learning_rate": 0.0006,
+ "loss": 4.2460832595825195,
+ "step": 5644
+ },
+ {
+ "epoch": 78.40524017467249,
+ "grad_norm": 0.014642846770584583,
+ "learning_rate": 0.0006,
+ "loss": 4.303811550140381,
+ "step": 5645
+ },
+ {
+ "epoch": 78.41921397379913,
+ "grad_norm": 0.013445726595818996,
+ "learning_rate": 0.0006,
+ "loss": 4.25921630859375,
+ "step": 5646
+ },
+ {
+ "epoch": 78.43318777292576,
+ "grad_norm": 0.014254065230488777,
+ "learning_rate": 0.0006,
+ "loss": 4.129446506500244,
+ "step": 5647
+ },
+ {
+ "epoch": 78.44716157205241,
+ "grad_norm": 0.01883019506931305,
+ "learning_rate": 0.0006,
+ "loss": 4.191716194152832,
+ "step": 5648
+ },
+ {
+ "epoch": 78.46113537117904,
+ "grad_norm": 0.0204437468200922,
+ "learning_rate": 0.0006,
+ "loss": 4.333085536956787,
+ "step": 5649
+ },
+ {
+ "epoch": 78.47510917030567,
+ "grad_norm": 0.01652318798005581,
+ "learning_rate": 0.0006,
+ "loss": 4.135462760925293,
+ "step": 5650
+ },
+ {
+ "epoch": 78.48908296943232,
+ "grad_norm": 0.012753551825881004,
+ "learning_rate": 0.0006,
+ "loss": 4.195051193237305,
+ "step": 5651
+ },
+ {
+ "epoch": 78.50305676855895,
+ "grad_norm": 0.01499118935316801,
+ "learning_rate": 0.0006,
+ "loss": 4.157092094421387,
+ "step": 5652
+ },
+ {
+ "epoch": 78.5170305676856,
+ "grad_norm": 0.015383945778012276,
+ "learning_rate": 0.0006,
+ "loss": 4.253818035125732,
+ "step": 5653
+ },
+ {
+ "epoch": 78.53100436681223,
+ "grad_norm": 0.015721332281827927,
+ "learning_rate": 0.0006,
+ "loss": 4.228775978088379,
+ "step": 5654
+ },
+ {
+ "epoch": 78.54497816593886,
+ "grad_norm": 0.014595463871955872,
+ "learning_rate": 0.0006,
+ "loss": 4.135197639465332,
+ "step": 5655
+ },
+ {
+ "epoch": 78.5589519650655,
+ "grad_norm": 0.015059644356369972,
+ "learning_rate": 0.0006,
+ "loss": 4.255614280700684,
+ "step": 5656
+ },
+ {
+ "epoch": 78.57292576419214,
+ "grad_norm": 0.013833344914019108,
+ "learning_rate": 0.0006,
+ "loss": 4.301250457763672,
+ "step": 5657
+ },
+ {
+ "epoch": 78.58689956331878,
+ "grad_norm": 0.013665376231074333,
+ "learning_rate": 0.0006,
+ "loss": 4.1734232902526855,
+ "step": 5658
+ },
+ {
+ "epoch": 78.60087336244541,
+ "grad_norm": 0.01539086177945137,
+ "learning_rate": 0.0006,
+ "loss": 4.221459865570068,
+ "step": 5659
+ },
+ {
+ "epoch": 78.61484716157206,
+ "grad_norm": 0.013313917443156242,
+ "learning_rate": 0.0006,
+ "loss": 4.255619049072266,
+ "step": 5660
+ },
+ {
+ "epoch": 78.62882096069869,
+ "grad_norm": 0.01447741687297821,
+ "learning_rate": 0.0006,
+ "loss": 4.298219680786133,
+ "step": 5661
+ },
+ {
+ "epoch": 78.64279475982532,
+ "grad_norm": 0.012796915136277676,
+ "learning_rate": 0.0006,
+ "loss": 4.2542405128479,
+ "step": 5662
+ },
+ {
+ "epoch": 78.65676855895197,
+ "grad_norm": 0.014058141969144344,
+ "learning_rate": 0.0006,
+ "loss": 4.284477233886719,
+ "step": 5663
+ },
+ {
+ "epoch": 78.6707423580786,
+ "grad_norm": 0.012705069966614246,
+ "learning_rate": 0.0006,
+ "loss": 4.2826666831970215,
+ "step": 5664
+ },
+ {
+ "epoch": 78.68471615720524,
+ "grad_norm": 0.013245890848338604,
+ "learning_rate": 0.0006,
+ "loss": 4.371612548828125,
+ "step": 5665
+ },
+ {
+ "epoch": 78.69868995633188,
+ "grad_norm": 0.014192677102982998,
+ "learning_rate": 0.0006,
+ "loss": 4.267344951629639,
+ "step": 5666
+ },
+ {
+ "epoch": 78.71266375545852,
+ "grad_norm": 0.015346777625381947,
+ "learning_rate": 0.0006,
+ "loss": 4.222797393798828,
+ "step": 5667
+ },
+ {
+ "epoch": 78.72663755458515,
+ "grad_norm": 0.014068621210753918,
+ "learning_rate": 0.0006,
+ "loss": 4.172964572906494,
+ "step": 5668
+ },
+ {
+ "epoch": 78.74061135371178,
+ "grad_norm": 0.014364945702254772,
+ "learning_rate": 0.0006,
+ "loss": 4.259319305419922,
+ "step": 5669
+ },
+ {
+ "epoch": 78.75458515283843,
+ "grad_norm": 0.017337387427687645,
+ "learning_rate": 0.0006,
+ "loss": 4.29533576965332,
+ "step": 5670
+ },
+ {
+ "epoch": 78.76855895196506,
+ "grad_norm": 0.017416715621948242,
+ "learning_rate": 0.0006,
+ "loss": 4.340364456176758,
+ "step": 5671
+ },
+ {
+ "epoch": 78.78253275109171,
+ "grad_norm": 0.015925846993923187,
+ "learning_rate": 0.0006,
+ "loss": 4.177454948425293,
+ "step": 5672
+ },
+ {
+ "epoch": 78.79650655021834,
+ "grad_norm": 0.013130443170666695,
+ "learning_rate": 0.0006,
+ "loss": 4.277454853057861,
+ "step": 5673
+ },
+ {
+ "epoch": 78.81048034934499,
+ "grad_norm": 0.014628687873482704,
+ "learning_rate": 0.0006,
+ "loss": 4.273757457733154,
+ "step": 5674
+ },
+ {
+ "epoch": 78.82445414847162,
+ "grad_norm": 0.01601659320294857,
+ "learning_rate": 0.0006,
+ "loss": 4.281774520874023,
+ "step": 5675
+ },
+ {
+ "epoch": 78.83842794759825,
+ "grad_norm": 0.015443485230207443,
+ "learning_rate": 0.0006,
+ "loss": 4.199937343597412,
+ "step": 5676
+ },
+ {
+ "epoch": 78.8524017467249,
+ "grad_norm": 0.013934246264398098,
+ "learning_rate": 0.0006,
+ "loss": 4.331369876861572,
+ "step": 5677
+ },
+ {
+ "epoch": 78.86637554585153,
+ "grad_norm": 0.01572083681821823,
+ "learning_rate": 0.0006,
+ "loss": 4.2828474044799805,
+ "step": 5678
+ },
+ {
+ "epoch": 78.88034934497817,
+ "grad_norm": 0.013444120064377785,
+ "learning_rate": 0.0006,
+ "loss": 4.183859825134277,
+ "step": 5679
+ },
+ {
+ "epoch": 78.8943231441048,
+ "grad_norm": 0.01591254211962223,
+ "learning_rate": 0.0006,
+ "loss": 4.295256614685059,
+ "step": 5680
+ },
+ {
+ "epoch": 78.90829694323143,
+ "grad_norm": 0.017556684091687202,
+ "learning_rate": 0.0006,
+ "loss": 4.132272720336914,
+ "step": 5681
+ },
+ {
+ "epoch": 78.92227074235808,
+ "grad_norm": 0.018184322863817215,
+ "learning_rate": 0.0006,
+ "loss": 4.251728057861328,
+ "step": 5682
+ },
+ {
+ "epoch": 78.93624454148471,
+ "grad_norm": 0.016207285225391388,
+ "learning_rate": 0.0006,
+ "loss": 4.205544471740723,
+ "step": 5683
+ },
+ {
+ "epoch": 78.95021834061136,
+ "grad_norm": 0.014478376135230064,
+ "learning_rate": 0.0006,
+ "loss": 4.182462215423584,
+ "step": 5684
+ },
+ {
+ "epoch": 78.96419213973799,
+ "grad_norm": 0.014255235902965069,
+ "learning_rate": 0.0006,
+ "loss": 4.253366947174072,
+ "step": 5685
+ },
+ {
+ "epoch": 78.97816593886463,
+ "grad_norm": 0.01524396613240242,
+ "learning_rate": 0.0006,
+ "loss": 4.221786022186279,
+ "step": 5686
+ },
+ {
+ "epoch": 78.99213973799127,
+ "grad_norm": 0.015057693235576153,
+ "learning_rate": 0.0006,
+ "loss": 4.208454608917236,
+ "step": 5687
+ },
+ {
+ "epoch": 79.0,
+ "grad_norm": 0.015478022396564484,
+ "learning_rate": 0.0006,
+ "loss": 4.336383819580078,
+ "step": 5688
+ },
+ {
+ "epoch": 79.0,
+ "eval_loss": 4.641749382019043,
+ "eval_runtime": 56.8167,
+ "eval_samples_per_second": 42.98,
+ "eval_steps_per_second": 1.355,
+ "step": 5688
+ },
+ {
+ "epoch": 79.01397379912663,
+ "grad_norm": 0.014143316075205803,
+ "learning_rate": 0.0006,
+ "loss": 4.260974884033203,
+ "step": 5689
+ },
+ {
+ "epoch": 79.02794759825328,
+ "grad_norm": 0.013120883144438267,
+ "learning_rate": 0.0006,
+ "loss": 4.139886856079102,
+ "step": 5690
+ },
+ {
+ "epoch": 79.04192139737991,
+ "grad_norm": 0.015019661746919155,
+ "learning_rate": 0.0006,
+ "loss": 4.267004013061523,
+ "step": 5691
+ },
+ {
+ "epoch": 79.05589519650655,
+ "grad_norm": 0.014891284517943859,
+ "learning_rate": 0.0006,
+ "loss": 4.205019950866699,
+ "step": 5692
+ },
+ {
+ "epoch": 79.06986899563319,
+ "grad_norm": 0.015527728945016861,
+ "learning_rate": 0.0006,
+ "loss": 4.1513566970825195,
+ "step": 5693
+ },
+ {
+ "epoch": 79.08384279475983,
+ "grad_norm": 0.017527181655168533,
+ "learning_rate": 0.0006,
+ "loss": 4.148828983306885,
+ "step": 5694
+ },
+ {
+ "epoch": 79.09781659388646,
+ "grad_norm": 0.01632206328213215,
+ "learning_rate": 0.0006,
+ "loss": 4.195789813995361,
+ "step": 5695
+ },
+ {
+ "epoch": 79.1117903930131,
+ "grad_norm": 0.01590503193438053,
+ "learning_rate": 0.0006,
+ "loss": 4.199155330657959,
+ "step": 5696
+ },
+ {
+ "epoch": 79.12576419213974,
+ "grad_norm": 0.016489997506141663,
+ "learning_rate": 0.0006,
+ "loss": 4.155154705047607,
+ "step": 5697
+ },
+ {
+ "epoch": 79.13973799126637,
+ "grad_norm": 0.0161368977278471,
+ "learning_rate": 0.0006,
+ "loss": 4.164384841918945,
+ "step": 5698
+ },
+ {
+ "epoch": 79.15371179039302,
+ "grad_norm": 0.018574809655547142,
+ "learning_rate": 0.0006,
+ "loss": 4.1898193359375,
+ "step": 5699
+ },
+ {
+ "epoch": 79.16768558951965,
+ "grad_norm": 0.018987983465194702,
+ "learning_rate": 0.0006,
+ "loss": 4.219719409942627,
+ "step": 5700
+ },
+ {
+ "epoch": 79.18165938864628,
+ "grad_norm": 0.01668606884777546,
+ "learning_rate": 0.0006,
+ "loss": 4.152734756469727,
+ "step": 5701
+ },
+ {
+ "epoch": 79.19563318777293,
+ "grad_norm": 0.014046299271285534,
+ "learning_rate": 0.0006,
+ "loss": 4.152964115142822,
+ "step": 5702
+ },
+ {
+ "epoch": 79.20960698689956,
+ "grad_norm": 0.01628032699227333,
+ "learning_rate": 0.0006,
+ "loss": 4.299479961395264,
+ "step": 5703
+ },
+ {
+ "epoch": 79.2235807860262,
+ "grad_norm": 0.017998795956373215,
+ "learning_rate": 0.0006,
+ "loss": 4.231203079223633,
+ "step": 5704
+ },
+ {
+ "epoch": 79.23755458515284,
+ "grad_norm": 0.014799647964537144,
+ "learning_rate": 0.0006,
+ "loss": 4.3530192375183105,
+ "step": 5705
+ },
+ {
+ "epoch": 79.25152838427948,
+ "grad_norm": 0.014640678651630878,
+ "learning_rate": 0.0006,
+ "loss": 4.287322998046875,
+ "step": 5706
+ },
+ {
+ "epoch": 79.26550218340611,
+ "grad_norm": 0.0172616895288229,
+ "learning_rate": 0.0006,
+ "loss": 4.148009777069092,
+ "step": 5707
+ },
+ {
+ "epoch": 79.27947598253274,
+ "grad_norm": 0.01811790093779564,
+ "learning_rate": 0.0006,
+ "loss": 4.269433975219727,
+ "step": 5708
+ },
+ {
+ "epoch": 79.29344978165939,
+ "grad_norm": 0.01764000579714775,
+ "learning_rate": 0.0006,
+ "loss": 4.292638778686523,
+ "step": 5709
+ },
+ {
+ "epoch": 79.30742358078602,
+ "grad_norm": 0.018030935898423195,
+ "learning_rate": 0.0006,
+ "loss": 4.267889976501465,
+ "step": 5710
+ },
+ {
+ "epoch": 79.32139737991267,
+ "grad_norm": 0.017860891297459602,
+ "learning_rate": 0.0006,
+ "loss": 4.211110591888428,
+ "step": 5711
+ },
+ {
+ "epoch": 79.3353711790393,
+ "grad_norm": 0.017295073717832565,
+ "learning_rate": 0.0006,
+ "loss": 4.224645614624023,
+ "step": 5712
+ },
+ {
+ "epoch": 79.34934497816595,
+ "grad_norm": 0.015891101211309433,
+ "learning_rate": 0.0006,
+ "loss": 4.224797248840332,
+ "step": 5713
+ },
+ {
+ "epoch": 79.36331877729258,
+ "grad_norm": 0.014680417254567146,
+ "learning_rate": 0.0006,
+ "loss": 4.182242393493652,
+ "step": 5714
+ },
+ {
+ "epoch": 79.37729257641921,
+ "grad_norm": 0.016621140763163567,
+ "learning_rate": 0.0006,
+ "loss": 4.186334609985352,
+ "step": 5715
+ },
+ {
+ "epoch": 79.39126637554585,
+ "grad_norm": 0.016207249835133553,
+ "learning_rate": 0.0006,
+ "loss": 4.277074813842773,
+ "step": 5716
+ },
+ {
+ "epoch": 79.40524017467249,
+ "grad_norm": 0.015375478193163872,
+ "learning_rate": 0.0006,
+ "loss": 4.154317855834961,
+ "step": 5717
+ },
+ {
+ "epoch": 79.41921397379913,
+ "grad_norm": 0.013453228399157524,
+ "learning_rate": 0.0006,
+ "loss": 4.265270233154297,
+ "step": 5718
+ },
+ {
+ "epoch": 79.43318777292576,
+ "grad_norm": 0.014312694780528545,
+ "learning_rate": 0.0006,
+ "loss": 4.232220649719238,
+ "step": 5719
+ },
+ {
+ "epoch": 79.44716157205241,
+ "grad_norm": 0.014759554527699947,
+ "learning_rate": 0.0006,
+ "loss": 4.272275924682617,
+ "step": 5720
+ },
+ {
+ "epoch": 79.46113537117904,
+ "grad_norm": 0.015885505825281143,
+ "learning_rate": 0.0006,
+ "loss": 4.2229719161987305,
+ "step": 5721
+ },
+ {
+ "epoch": 79.47510917030567,
+ "grad_norm": 0.016516581177711487,
+ "learning_rate": 0.0006,
+ "loss": 4.200624465942383,
+ "step": 5722
+ },
+ {
+ "epoch": 79.48908296943232,
+ "grad_norm": 0.018254555761814117,
+ "learning_rate": 0.0006,
+ "loss": 4.19648551940918,
+ "step": 5723
+ },
+ {
+ "epoch": 79.50305676855895,
+ "grad_norm": 0.01467090006917715,
+ "learning_rate": 0.0006,
+ "loss": 4.135190486907959,
+ "step": 5724
+ },
+ {
+ "epoch": 79.5170305676856,
+ "grad_norm": 0.013474797829985619,
+ "learning_rate": 0.0006,
+ "loss": 4.311361312866211,
+ "step": 5725
+ },
+ {
+ "epoch": 79.53100436681223,
+ "grad_norm": 0.01742764748632908,
+ "learning_rate": 0.0006,
+ "loss": 4.1845903396606445,
+ "step": 5726
+ },
+ {
+ "epoch": 79.54497816593886,
+ "grad_norm": 0.018786882981657982,
+ "learning_rate": 0.0006,
+ "loss": 4.163942337036133,
+ "step": 5727
+ },
+ {
+ "epoch": 79.5589519650655,
+ "grad_norm": 0.017831219360232353,
+ "learning_rate": 0.0006,
+ "loss": 4.236685276031494,
+ "step": 5728
+ },
+ {
+ "epoch": 79.57292576419214,
+ "grad_norm": 0.016334986314177513,
+ "learning_rate": 0.0006,
+ "loss": 4.255853652954102,
+ "step": 5729
+ },
+ {
+ "epoch": 79.58689956331878,
+ "grad_norm": 0.0138487434014678,
+ "learning_rate": 0.0006,
+ "loss": 4.227874279022217,
+ "step": 5730
+ },
+ {
+ "epoch": 79.60087336244541,
+ "grad_norm": 0.01531760673969984,
+ "learning_rate": 0.0006,
+ "loss": 4.221432685852051,
+ "step": 5731
+ },
+ {
+ "epoch": 79.61484716157206,
+ "grad_norm": 0.01603361964225769,
+ "learning_rate": 0.0006,
+ "loss": 4.278326511383057,
+ "step": 5732
+ },
+ {
+ "epoch": 79.62882096069869,
+ "grad_norm": 0.013149895705282688,
+ "learning_rate": 0.0006,
+ "loss": 4.2474517822265625,
+ "step": 5733
+ },
+ {
+ "epoch": 79.64279475982532,
+ "grad_norm": 0.013428786769509315,
+ "learning_rate": 0.0006,
+ "loss": 4.1701178550720215,
+ "step": 5734
+ },
+ {
+ "epoch": 79.65676855895197,
+ "grad_norm": 0.014305585995316505,
+ "learning_rate": 0.0006,
+ "loss": 4.250778675079346,
+ "step": 5735
+ },
+ {
+ "epoch": 79.6707423580786,
+ "grad_norm": 0.014544566161930561,
+ "learning_rate": 0.0006,
+ "loss": 4.243236541748047,
+ "step": 5736
+ },
+ {
+ "epoch": 79.68471615720524,
+ "grad_norm": 0.013524414040148258,
+ "learning_rate": 0.0006,
+ "loss": 4.255457878112793,
+ "step": 5737
+ },
+ {
+ "epoch": 79.69868995633188,
+ "grad_norm": 0.014750641770660877,
+ "learning_rate": 0.0006,
+ "loss": 4.202816486358643,
+ "step": 5738
+ },
+ {
+ "epoch": 79.71266375545852,
+ "grad_norm": 0.015090773813426495,
+ "learning_rate": 0.0006,
+ "loss": 4.3053178787231445,
+ "step": 5739
+ },
+ {
+ "epoch": 79.72663755458515,
+ "grad_norm": 0.01606169529259205,
+ "learning_rate": 0.0006,
+ "loss": 4.224017143249512,
+ "step": 5740
+ },
+ {
+ "epoch": 79.74061135371178,
+ "grad_norm": 0.016303520649671555,
+ "learning_rate": 0.0006,
+ "loss": 4.190760612487793,
+ "step": 5741
+ },
+ {
+ "epoch": 79.75458515283843,
+ "grad_norm": 0.01465672068297863,
+ "learning_rate": 0.0006,
+ "loss": 4.24506950378418,
+ "step": 5742
+ },
+ {
+ "epoch": 79.76855895196506,
+ "grad_norm": 0.012616301886737347,
+ "learning_rate": 0.0006,
+ "loss": 4.174039840698242,
+ "step": 5743
+ },
+ {
+ "epoch": 79.78253275109171,
+ "grad_norm": 0.01358284242451191,
+ "learning_rate": 0.0006,
+ "loss": 4.236380577087402,
+ "step": 5744
+ },
+ {
+ "epoch": 79.79650655021834,
+ "grad_norm": 0.01557177770882845,
+ "learning_rate": 0.0006,
+ "loss": 4.102962493896484,
+ "step": 5745
+ },
+ {
+ "epoch": 79.81048034934499,
+ "grad_norm": 0.015584317035973072,
+ "learning_rate": 0.0006,
+ "loss": 4.242603302001953,
+ "step": 5746
+ },
+ {
+ "epoch": 79.82445414847162,
+ "grad_norm": 0.012597334571182728,
+ "learning_rate": 0.0006,
+ "loss": 4.202791690826416,
+ "step": 5747
+ },
+ {
+ "epoch": 79.83842794759825,
+ "grad_norm": 0.013722898438572884,
+ "learning_rate": 0.0006,
+ "loss": 4.204328536987305,
+ "step": 5748
+ },
+ {
+ "epoch": 79.8524017467249,
+ "grad_norm": 0.014669225551187992,
+ "learning_rate": 0.0006,
+ "loss": 4.164331912994385,
+ "step": 5749
+ },
+ {
+ "epoch": 79.86637554585153,
+ "grad_norm": 0.015012740157544613,
+ "learning_rate": 0.0006,
+ "loss": 4.260565280914307,
+ "step": 5750
+ },
+ {
+ "epoch": 79.88034934497817,
+ "grad_norm": 0.016026340425014496,
+ "learning_rate": 0.0006,
+ "loss": 4.299723148345947,
+ "step": 5751
+ },
+ {
+ "epoch": 79.8943231441048,
+ "grad_norm": 0.01626049168407917,
+ "learning_rate": 0.0006,
+ "loss": 4.207252502441406,
+ "step": 5752
+ },
+ {
+ "epoch": 79.90829694323143,
+ "grad_norm": 0.015415354631841183,
+ "learning_rate": 0.0006,
+ "loss": 4.151934623718262,
+ "step": 5753
+ },
+ {
+ "epoch": 79.92227074235808,
+ "grad_norm": 0.014158242382109165,
+ "learning_rate": 0.0006,
+ "loss": 4.3461995124816895,
+ "step": 5754
+ },
+ {
+ "epoch": 79.93624454148471,
+ "grad_norm": 0.01585628278553486,
+ "learning_rate": 0.0006,
+ "loss": 4.334761619567871,
+ "step": 5755
+ },
+ {
+ "epoch": 79.95021834061136,
+ "grad_norm": 0.013710759580135345,
+ "learning_rate": 0.0006,
+ "loss": 4.180401802062988,
+ "step": 5756
+ },
+ {
+ "epoch": 79.96419213973799,
+ "grad_norm": 0.014159204438328743,
+ "learning_rate": 0.0006,
+ "loss": 4.322331428527832,
+ "step": 5757
+ },
+ {
+ "epoch": 79.97816593886463,
+ "grad_norm": 0.015153195708990097,
+ "learning_rate": 0.0006,
+ "loss": 4.287478923797607,
+ "step": 5758
+ },
+ {
+ "epoch": 79.99213973799127,
+ "grad_norm": 0.014100227504968643,
+ "learning_rate": 0.0006,
+ "loss": 4.231807708740234,
+ "step": 5759
+ },
+ {
+ "epoch": 80.0,
+ "grad_norm": 0.015248682349920273,
+ "learning_rate": 0.0006,
+ "loss": 4.261106014251709,
+ "step": 5760
+ },
+ {
+ "epoch": 80.0,
+ "eval_loss": 4.610264301300049,
+ "eval_runtime": 56.514,
+ "eval_samples_per_second": 43.211,
+ "eval_steps_per_second": 1.362,
+ "step": 5760
+ },
+ {
+ "epoch": 80.01397379912663,
+ "grad_norm": 0.014730271883308887,
+ "learning_rate": 0.0006,
+ "loss": 4.108556747436523,
+ "step": 5761
+ },
+ {
+ "epoch": 80.02794759825328,
+ "grad_norm": 0.014575716108083725,
+ "learning_rate": 0.0006,
+ "loss": 4.169658660888672,
+ "step": 5762
+ },
+ {
+ "epoch": 80.04192139737991,
+ "grad_norm": 0.014706958085298538,
+ "learning_rate": 0.0006,
+ "loss": 4.252662658691406,
+ "step": 5763
+ },
+ {
+ "epoch": 80.05589519650655,
+ "grad_norm": 0.01561224739998579,
+ "learning_rate": 0.0006,
+ "loss": 4.193892955780029,
+ "step": 5764
+ },
+ {
+ "epoch": 80.06986899563319,
+ "grad_norm": 0.014702324755489826,
+ "learning_rate": 0.0006,
+ "loss": 4.1889238357543945,
+ "step": 5765
+ },
+ {
+ "epoch": 80.08384279475983,
+ "grad_norm": 0.015565741807222366,
+ "learning_rate": 0.0006,
+ "loss": 4.243549823760986,
+ "step": 5766
+ },
+ {
+ "epoch": 80.09781659388646,
+ "grad_norm": 0.01647181063890457,
+ "learning_rate": 0.0006,
+ "loss": 4.275446891784668,
+ "step": 5767
+ },
+ {
+ "epoch": 80.1117903930131,
+ "grad_norm": 0.018177716061472893,
+ "learning_rate": 0.0006,
+ "loss": 4.12040376663208,
+ "step": 5768
+ },
+ {
+ "epoch": 80.12576419213974,
+ "grad_norm": 0.01745191030204296,
+ "learning_rate": 0.0006,
+ "loss": 4.140371322631836,
+ "step": 5769
+ },
+ {
+ "epoch": 80.13973799126637,
+ "grad_norm": 0.01568993367254734,
+ "learning_rate": 0.0006,
+ "loss": 4.258280277252197,
+ "step": 5770
+ },
+ {
+ "epoch": 80.15371179039302,
+ "grad_norm": 0.01636597141623497,
+ "learning_rate": 0.0006,
+ "loss": 4.179052352905273,
+ "step": 5771
+ },
+ {
+ "epoch": 80.16768558951965,
+ "grad_norm": 0.016839100047945976,
+ "learning_rate": 0.0006,
+ "loss": 4.127636432647705,
+ "step": 5772
+ },
+ {
+ "epoch": 80.18165938864628,
+ "grad_norm": 0.015471016988158226,
+ "learning_rate": 0.0006,
+ "loss": 4.181325912475586,
+ "step": 5773
+ },
+ {
+ "epoch": 80.19563318777293,
+ "grad_norm": 0.015836089849472046,
+ "learning_rate": 0.0006,
+ "loss": 4.142168998718262,
+ "step": 5774
+ },
+ {
+ "epoch": 80.20960698689956,
+ "grad_norm": 0.014782954007387161,
+ "learning_rate": 0.0006,
+ "loss": 4.308799743652344,
+ "step": 5775
+ },
+ {
+ "epoch": 80.2235807860262,
+ "grad_norm": 0.014040893875062466,
+ "learning_rate": 0.0006,
+ "loss": 4.216936111450195,
+ "step": 5776
+ },
+ {
+ "epoch": 80.23755458515284,
+ "grad_norm": 0.01420280709862709,
+ "learning_rate": 0.0006,
+ "loss": 4.235230445861816,
+ "step": 5777
+ },
+ {
+ "epoch": 80.25152838427948,
+ "grad_norm": 0.013341007754206657,
+ "learning_rate": 0.0006,
+ "loss": 4.369109153747559,
+ "step": 5778
+ },
+ {
+ "epoch": 80.26550218340611,
+ "grad_norm": 0.014208327978849411,
+ "learning_rate": 0.0006,
+ "loss": 4.219112396240234,
+ "step": 5779
+ },
+ {
+ "epoch": 80.27947598253274,
+ "grad_norm": 0.013825052417814732,
+ "learning_rate": 0.0006,
+ "loss": 4.1983489990234375,
+ "step": 5780
+ },
+ {
+ "epoch": 80.29344978165939,
+ "grad_norm": 0.012916515581309795,
+ "learning_rate": 0.0006,
+ "loss": 4.204582691192627,
+ "step": 5781
+ },
+ {
+ "epoch": 80.30742358078602,
+ "grad_norm": 0.013071243651211262,
+ "learning_rate": 0.0006,
+ "loss": 4.17643928527832,
+ "step": 5782
+ },
+ {
+ "epoch": 80.32139737991267,
+ "grad_norm": 0.015507174655795097,
+ "learning_rate": 0.0006,
+ "loss": 4.172976016998291,
+ "step": 5783
+ },
+ {
+ "epoch": 80.3353711790393,
+ "grad_norm": 0.014967243187129498,
+ "learning_rate": 0.0006,
+ "loss": 4.141168594360352,
+ "step": 5784
+ },
+ {
+ "epoch": 80.34934497816595,
+ "grad_norm": 0.015629924833774567,
+ "learning_rate": 0.0006,
+ "loss": 4.253526210784912,
+ "step": 5785
+ },
+ {
+ "epoch": 80.36331877729258,
+ "grad_norm": 0.01607954502105713,
+ "learning_rate": 0.0006,
+ "loss": 4.249395370483398,
+ "step": 5786
+ },
+ {
+ "epoch": 80.37729257641921,
+ "grad_norm": 0.01500706560909748,
+ "learning_rate": 0.0006,
+ "loss": 4.279012680053711,
+ "step": 5787
+ },
+ {
+ "epoch": 80.39126637554585,
+ "grad_norm": 0.014343900606036186,
+ "learning_rate": 0.0006,
+ "loss": 4.268509864807129,
+ "step": 5788
+ },
+ {
+ "epoch": 80.40524017467249,
+ "grad_norm": 0.014682197012007236,
+ "learning_rate": 0.0006,
+ "loss": 4.206334114074707,
+ "step": 5789
+ },
+ {
+ "epoch": 80.41921397379913,
+ "grad_norm": 0.014989163726568222,
+ "learning_rate": 0.0006,
+ "loss": 4.143601417541504,
+ "step": 5790
+ },
+ {
+ "epoch": 80.43318777292576,
+ "grad_norm": 0.015001608058810234,
+ "learning_rate": 0.0006,
+ "loss": 4.164111137390137,
+ "step": 5791
+ },
+ {
+ "epoch": 80.44716157205241,
+ "grad_norm": 0.015222722664475441,
+ "learning_rate": 0.0006,
+ "loss": 4.2369914054870605,
+ "step": 5792
+ },
+ {
+ "epoch": 80.46113537117904,
+ "grad_norm": 0.014126183465123177,
+ "learning_rate": 0.0006,
+ "loss": 4.2209930419921875,
+ "step": 5793
+ },
+ {
+ "epoch": 80.47510917030567,
+ "grad_norm": 0.013065959326922894,
+ "learning_rate": 0.0006,
+ "loss": 4.231184959411621,
+ "step": 5794
+ },
+ {
+ "epoch": 80.48908296943232,
+ "grad_norm": 0.013833962380886078,
+ "learning_rate": 0.0006,
+ "loss": 4.226434707641602,
+ "step": 5795
+ },
+ {
+ "epoch": 80.50305676855895,
+ "grad_norm": 0.01387125626206398,
+ "learning_rate": 0.0006,
+ "loss": 4.241328716278076,
+ "step": 5796
+ },
+ {
+ "epoch": 80.5170305676856,
+ "grad_norm": 0.014145063236355782,
+ "learning_rate": 0.0006,
+ "loss": 4.2268805503845215,
+ "step": 5797
+ },
+ {
+ "epoch": 80.53100436681223,
+ "grad_norm": 0.012990240007638931,
+ "learning_rate": 0.0006,
+ "loss": 4.1718339920043945,
+ "step": 5798
+ },
+ {
+ "epoch": 80.54497816593886,
+ "grad_norm": 0.014185411855578423,
+ "learning_rate": 0.0006,
+ "loss": 4.232397556304932,
+ "step": 5799
+ },
+ {
+ "epoch": 80.5589519650655,
+ "grad_norm": 0.014260523952543736,
+ "learning_rate": 0.0006,
+ "loss": 4.282734394073486,
+ "step": 5800
+ },
+ {
+ "epoch": 80.57292576419214,
+ "grad_norm": 0.012542120181024075,
+ "learning_rate": 0.0006,
+ "loss": 4.134551525115967,
+ "step": 5801
+ },
+ {
+ "epoch": 80.58689956331878,
+ "grad_norm": 0.013078657910227776,
+ "learning_rate": 0.0006,
+ "loss": 4.229815483093262,
+ "step": 5802
+ },
+ {
+ "epoch": 80.60087336244541,
+ "grad_norm": 0.012821056880056858,
+ "learning_rate": 0.0006,
+ "loss": 4.208278179168701,
+ "step": 5803
+ },
+ {
+ "epoch": 80.61484716157206,
+ "grad_norm": 0.012964180670678616,
+ "learning_rate": 0.0006,
+ "loss": 4.188477039337158,
+ "step": 5804
+ },
+ {
+ "epoch": 80.62882096069869,
+ "grad_norm": 0.01631746254861355,
+ "learning_rate": 0.0006,
+ "loss": 4.3214497566223145,
+ "step": 5805
+ },
+ {
+ "epoch": 80.64279475982532,
+ "grad_norm": 0.01726640947163105,
+ "learning_rate": 0.0006,
+ "loss": 4.279359817504883,
+ "step": 5806
+ },
+ {
+ "epoch": 80.65676855895197,
+ "grad_norm": 0.018778258934617043,
+ "learning_rate": 0.0006,
+ "loss": 4.294047832489014,
+ "step": 5807
+ },
+ {
+ "epoch": 80.6707423580786,
+ "grad_norm": 0.01796831749379635,
+ "learning_rate": 0.0006,
+ "loss": 4.314162254333496,
+ "step": 5808
+ },
+ {
+ "epoch": 80.68471615720524,
+ "grad_norm": 0.016495538875460625,
+ "learning_rate": 0.0006,
+ "loss": 4.273871421813965,
+ "step": 5809
+ },
+ {
+ "epoch": 80.69868995633188,
+ "grad_norm": 0.013590306043624878,
+ "learning_rate": 0.0006,
+ "loss": 4.14961051940918,
+ "step": 5810
+ },
+ {
+ "epoch": 80.71266375545852,
+ "grad_norm": 0.014670426957309246,
+ "learning_rate": 0.0006,
+ "loss": 4.211853981018066,
+ "step": 5811
+ },
+ {
+ "epoch": 80.72663755458515,
+ "grad_norm": 0.01641341857612133,
+ "learning_rate": 0.0006,
+ "loss": 4.211948871612549,
+ "step": 5812
+ },
+ {
+ "epoch": 80.74061135371178,
+ "grad_norm": 0.016365470364689827,
+ "learning_rate": 0.0006,
+ "loss": 4.200774192810059,
+ "step": 5813
+ },
+ {
+ "epoch": 80.75458515283843,
+ "grad_norm": 0.01613970287144184,
+ "learning_rate": 0.0006,
+ "loss": 4.288322448730469,
+ "step": 5814
+ },
+ {
+ "epoch": 80.76855895196506,
+ "grad_norm": 0.016742777079343796,
+ "learning_rate": 0.0006,
+ "loss": 4.23801326751709,
+ "step": 5815
+ },
+ {
+ "epoch": 80.78253275109171,
+ "grad_norm": 0.015351069159805775,
+ "learning_rate": 0.0006,
+ "loss": 4.263832092285156,
+ "step": 5816
+ },
+ {
+ "epoch": 80.79650655021834,
+ "grad_norm": 0.013435336761176586,
+ "learning_rate": 0.0006,
+ "loss": 4.160857200622559,
+ "step": 5817
+ },
+ {
+ "epoch": 80.81048034934499,
+ "grad_norm": 0.01474886853247881,
+ "learning_rate": 0.0006,
+ "loss": 4.216025352478027,
+ "step": 5818
+ },
+ {
+ "epoch": 80.82445414847162,
+ "grad_norm": 0.01586935855448246,
+ "learning_rate": 0.0006,
+ "loss": 4.256402969360352,
+ "step": 5819
+ },
+ {
+ "epoch": 80.83842794759825,
+ "grad_norm": 0.014596572145819664,
+ "learning_rate": 0.0006,
+ "loss": 4.142584800720215,
+ "step": 5820
+ },
+ {
+ "epoch": 80.8524017467249,
+ "grad_norm": 0.01501593366265297,
+ "learning_rate": 0.0006,
+ "loss": 4.250462055206299,
+ "step": 5821
+ },
+ {
+ "epoch": 80.86637554585153,
+ "grad_norm": 0.014773511327803135,
+ "learning_rate": 0.0006,
+ "loss": 4.253305912017822,
+ "step": 5822
+ },
+ {
+ "epoch": 80.88034934497817,
+ "grad_norm": 0.013246326707303524,
+ "learning_rate": 0.0006,
+ "loss": 4.170881271362305,
+ "step": 5823
+ },
+ {
+ "epoch": 80.8943231441048,
+ "grad_norm": 0.013013756833970547,
+ "learning_rate": 0.0006,
+ "loss": 4.264721870422363,
+ "step": 5824
+ },
+ {
+ "epoch": 80.90829694323143,
+ "grad_norm": 0.013155903667211533,
+ "learning_rate": 0.0006,
+ "loss": 4.274667739868164,
+ "step": 5825
+ },
+ {
+ "epoch": 80.92227074235808,
+ "grad_norm": 0.01448952779173851,
+ "learning_rate": 0.0006,
+ "loss": 4.24777889251709,
+ "step": 5826
+ },
+ {
+ "epoch": 80.93624454148471,
+ "grad_norm": 0.01415231078863144,
+ "learning_rate": 0.0006,
+ "loss": 4.22014856338501,
+ "step": 5827
+ },
+ {
+ "epoch": 80.95021834061136,
+ "grad_norm": 0.014796273782849312,
+ "learning_rate": 0.0006,
+ "loss": 4.22036600112915,
+ "step": 5828
+ },
+ {
+ "epoch": 80.96419213973799,
+ "grad_norm": 0.013005326502025127,
+ "learning_rate": 0.0006,
+ "loss": 4.276123046875,
+ "step": 5829
+ },
+ {
+ "epoch": 80.97816593886463,
+ "grad_norm": 0.012851177714765072,
+ "learning_rate": 0.0006,
+ "loss": 4.248575687408447,
+ "step": 5830
+ },
+ {
+ "epoch": 80.99213973799127,
+ "grad_norm": 0.013740002177655697,
+ "learning_rate": 0.0006,
+ "loss": 4.051140308380127,
+ "step": 5831
+ },
+ {
+ "epoch": 81.0,
+ "grad_norm": 0.015296104364097118,
+ "learning_rate": 0.0006,
+ "loss": 4.251155853271484,
+ "step": 5832
+ },
+ {
+ "epoch": 81.0,
+ "eval_loss": 4.594127655029297,
+ "eval_runtime": 56.0959,
+ "eval_samples_per_second": 43.533,
+ "eval_steps_per_second": 1.373,
+ "step": 5832
+ },
+ {
+ "epoch": 81.01397379912663,
+ "grad_norm": 0.015804558992385864,
+ "learning_rate": 0.0006,
+ "loss": 4.152894020080566,
+ "step": 5833
+ },
+ {
+ "epoch": 81.02794759825328,
+ "grad_norm": 0.017000854015350342,
+ "learning_rate": 0.0006,
+ "loss": 4.266067981719971,
+ "step": 5834
+ },
+ {
+ "epoch": 81.04192139737991,
+ "grad_norm": 0.016351008787751198,
+ "learning_rate": 0.0006,
+ "loss": 4.192703723907471,
+ "step": 5835
+ },
+ {
+ "epoch": 81.05589519650655,
+ "grad_norm": 0.014703895896673203,
+ "learning_rate": 0.0006,
+ "loss": 4.23554801940918,
+ "step": 5836
+ },
+ {
+ "epoch": 81.06986899563319,
+ "grad_norm": 0.01619596965610981,
+ "learning_rate": 0.0006,
+ "loss": 4.233000755310059,
+ "step": 5837
+ },
+ {
+ "epoch": 81.08384279475983,
+ "grad_norm": 0.01815493032336235,
+ "learning_rate": 0.0006,
+ "loss": 4.200490951538086,
+ "step": 5838
+ },
+ {
+ "epoch": 81.09781659388646,
+ "grad_norm": 0.018305106088519096,
+ "learning_rate": 0.0006,
+ "loss": 4.154242515563965,
+ "step": 5839
+ },
+ {
+ "epoch": 81.1117903930131,
+ "grad_norm": 0.0200856514275074,
+ "learning_rate": 0.0006,
+ "loss": 4.25820779800415,
+ "step": 5840
+ },
+ {
+ "epoch": 81.12576419213974,
+ "grad_norm": 0.02101440541446209,
+ "learning_rate": 0.0006,
+ "loss": 4.168209075927734,
+ "step": 5841
+ },
+ {
+ "epoch": 81.13973799126637,
+ "grad_norm": 0.022829681634902954,
+ "learning_rate": 0.0006,
+ "loss": 4.25560998916626,
+ "step": 5842
+ },
+ {
+ "epoch": 81.15371179039302,
+ "grad_norm": 0.0195082426071167,
+ "learning_rate": 0.0006,
+ "loss": 4.2035627365112305,
+ "step": 5843
+ },
+ {
+ "epoch": 81.16768558951965,
+ "grad_norm": 0.01716248318552971,
+ "learning_rate": 0.0006,
+ "loss": 4.11322021484375,
+ "step": 5844
+ },
+ {
+ "epoch": 81.18165938864628,
+ "grad_norm": 0.02034664712846279,
+ "learning_rate": 0.0006,
+ "loss": 4.179389953613281,
+ "step": 5845
+ },
+ {
+ "epoch": 81.19563318777293,
+ "grad_norm": 0.023578710854053497,
+ "learning_rate": 0.0006,
+ "loss": 4.265695571899414,
+ "step": 5846
+ },
+ {
+ "epoch": 81.20960698689956,
+ "grad_norm": 0.019244182854890823,
+ "learning_rate": 0.0006,
+ "loss": 4.292043209075928,
+ "step": 5847
+ },
+ {
+ "epoch": 81.2235807860262,
+ "grad_norm": 0.01577291637659073,
+ "learning_rate": 0.0006,
+ "loss": 4.250776290893555,
+ "step": 5848
+ },
+ {
+ "epoch": 81.23755458515284,
+ "grad_norm": 0.01617090217769146,
+ "learning_rate": 0.0006,
+ "loss": 4.193098068237305,
+ "step": 5849
+ },
+ {
+ "epoch": 81.25152838427948,
+ "grad_norm": 0.016090184450149536,
+ "learning_rate": 0.0006,
+ "loss": 4.112357139587402,
+ "step": 5850
+ },
+ {
+ "epoch": 81.26550218340611,
+ "grad_norm": 0.01441365759819746,
+ "learning_rate": 0.0006,
+ "loss": 4.114147186279297,
+ "step": 5851
+ },
+ {
+ "epoch": 81.27947598253274,
+ "grad_norm": 0.014122162945568562,
+ "learning_rate": 0.0006,
+ "loss": 4.276175022125244,
+ "step": 5852
+ },
+ {
+ "epoch": 81.29344978165939,
+ "grad_norm": 0.014859487302601337,
+ "learning_rate": 0.0006,
+ "loss": 4.094699859619141,
+ "step": 5853
+ },
+ {
+ "epoch": 81.30742358078602,
+ "grad_norm": 0.013624267652630806,
+ "learning_rate": 0.0006,
+ "loss": 4.2648773193359375,
+ "step": 5854
+ },
+ {
+ "epoch": 81.32139737991267,
+ "grad_norm": 0.01417913381010294,
+ "learning_rate": 0.0006,
+ "loss": 4.228567123413086,
+ "step": 5855
+ },
+ {
+ "epoch": 81.3353711790393,
+ "grad_norm": 0.014560805633664131,
+ "learning_rate": 0.0006,
+ "loss": 4.148930549621582,
+ "step": 5856
+ },
+ {
+ "epoch": 81.34934497816595,
+ "grad_norm": 0.013662154786288738,
+ "learning_rate": 0.0006,
+ "loss": 4.192107677459717,
+ "step": 5857
+ },
+ {
+ "epoch": 81.36331877729258,
+ "grad_norm": 0.013434980995953083,
+ "learning_rate": 0.0006,
+ "loss": 4.162052154541016,
+ "step": 5858
+ },
+ {
+ "epoch": 81.37729257641921,
+ "grad_norm": 0.01300742756575346,
+ "learning_rate": 0.0006,
+ "loss": 4.262385368347168,
+ "step": 5859
+ },
+ {
+ "epoch": 81.39126637554585,
+ "grad_norm": 0.012544351629912853,
+ "learning_rate": 0.0006,
+ "loss": 4.215371131896973,
+ "step": 5860
+ },
+ {
+ "epoch": 81.40524017467249,
+ "grad_norm": 0.013663645833730698,
+ "learning_rate": 0.0006,
+ "loss": 4.247312545776367,
+ "step": 5861
+ },
+ {
+ "epoch": 81.41921397379913,
+ "grad_norm": 0.013127296231687069,
+ "learning_rate": 0.0006,
+ "loss": 4.087685585021973,
+ "step": 5862
+ },
+ {
+ "epoch": 81.43318777292576,
+ "grad_norm": 0.014081938192248344,
+ "learning_rate": 0.0006,
+ "loss": 4.272648811340332,
+ "step": 5863
+ },
+ {
+ "epoch": 81.44716157205241,
+ "grad_norm": 0.015470641665160656,
+ "learning_rate": 0.0006,
+ "loss": 4.278578758239746,
+ "step": 5864
+ },
+ {
+ "epoch": 81.46113537117904,
+ "grad_norm": 0.016524532809853554,
+ "learning_rate": 0.0006,
+ "loss": 4.118268013000488,
+ "step": 5865
+ },
+ {
+ "epoch": 81.47510917030567,
+ "grad_norm": 0.016242722049355507,
+ "learning_rate": 0.0006,
+ "loss": 4.182405471801758,
+ "step": 5866
+ },
+ {
+ "epoch": 81.48908296943232,
+ "grad_norm": 0.01557017583400011,
+ "learning_rate": 0.0006,
+ "loss": 4.229862213134766,
+ "step": 5867
+ },
+ {
+ "epoch": 81.50305676855895,
+ "grad_norm": 0.01562870666384697,
+ "learning_rate": 0.0006,
+ "loss": 4.27754020690918,
+ "step": 5868
+ },
+ {
+ "epoch": 81.5170305676856,
+ "grad_norm": 0.018606074154376984,
+ "learning_rate": 0.0006,
+ "loss": 4.148681640625,
+ "step": 5869
+ },
+ {
+ "epoch": 81.53100436681223,
+ "grad_norm": 0.017740730196237564,
+ "learning_rate": 0.0006,
+ "loss": 4.321932792663574,
+ "step": 5870
+ },
+ {
+ "epoch": 81.54497816593886,
+ "grad_norm": 0.014517677947878838,
+ "learning_rate": 0.0006,
+ "loss": 4.22862434387207,
+ "step": 5871
+ },
+ {
+ "epoch": 81.5589519650655,
+ "grad_norm": 0.01389714702963829,
+ "learning_rate": 0.0006,
+ "loss": 4.198000907897949,
+ "step": 5872
+ },
+ {
+ "epoch": 81.57292576419214,
+ "grad_norm": 0.015348682180047035,
+ "learning_rate": 0.0006,
+ "loss": 4.293485164642334,
+ "step": 5873
+ },
+ {
+ "epoch": 81.58689956331878,
+ "grad_norm": 0.015965791419148445,
+ "learning_rate": 0.0006,
+ "loss": 4.23481559753418,
+ "step": 5874
+ },
+ {
+ "epoch": 81.60087336244541,
+ "grad_norm": 0.015662726014852524,
+ "learning_rate": 0.0006,
+ "loss": 4.158148765563965,
+ "step": 5875
+ },
+ {
+ "epoch": 81.61484716157206,
+ "grad_norm": 0.014402080327272415,
+ "learning_rate": 0.0006,
+ "loss": 4.132723808288574,
+ "step": 5876
+ },
+ {
+ "epoch": 81.62882096069869,
+ "grad_norm": 0.016846125945448875,
+ "learning_rate": 0.0006,
+ "loss": 4.1654462814331055,
+ "step": 5877
+ },
+ {
+ "epoch": 81.64279475982532,
+ "grad_norm": 0.0196097232401371,
+ "learning_rate": 0.0006,
+ "loss": 4.228450298309326,
+ "step": 5878
+ },
+ {
+ "epoch": 81.65676855895197,
+ "grad_norm": 0.016005294397473335,
+ "learning_rate": 0.0006,
+ "loss": 4.230222225189209,
+ "step": 5879
+ },
+ {
+ "epoch": 81.6707423580786,
+ "grad_norm": 0.014685059897601604,
+ "learning_rate": 0.0006,
+ "loss": 4.1367387771606445,
+ "step": 5880
+ },
+ {
+ "epoch": 81.68471615720524,
+ "grad_norm": 0.01647144928574562,
+ "learning_rate": 0.0006,
+ "loss": 4.096675395965576,
+ "step": 5881
+ },
+ {
+ "epoch": 81.69868995633188,
+ "grad_norm": 0.015978610143065453,
+ "learning_rate": 0.0006,
+ "loss": 4.232292175292969,
+ "step": 5882
+ },
+ {
+ "epoch": 81.71266375545852,
+ "grad_norm": 0.01510288193821907,
+ "learning_rate": 0.0006,
+ "loss": 4.274591445922852,
+ "step": 5883
+ },
+ {
+ "epoch": 81.72663755458515,
+ "grad_norm": 0.014942476525902748,
+ "learning_rate": 0.0006,
+ "loss": 4.161482810974121,
+ "step": 5884
+ },
+ {
+ "epoch": 81.74061135371178,
+ "grad_norm": 0.014595299027860165,
+ "learning_rate": 0.0006,
+ "loss": 4.204291343688965,
+ "step": 5885
+ },
+ {
+ "epoch": 81.75458515283843,
+ "grad_norm": 0.014952429570257664,
+ "learning_rate": 0.0006,
+ "loss": 4.242499351501465,
+ "step": 5886
+ },
+ {
+ "epoch": 81.76855895196506,
+ "grad_norm": 0.014761554077267647,
+ "learning_rate": 0.0006,
+ "loss": 4.20393180847168,
+ "step": 5887
+ },
+ {
+ "epoch": 81.78253275109171,
+ "grad_norm": 0.014749647118151188,
+ "learning_rate": 0.0006,
+ "loss": 4.227565765380859,
+ "step": 5888
+ },
+ {
+ "epoch": 81.79650655021834,
+ "grad_norm": 0.014384493231773376,
+ "learning_rate": 0.0006,
+ "loss": 4.282228946685791,
+ "step": 5889
+ },
+ {
+ "epoch": 81.81048034934499,
+ "grad_norm": 0.01260330155491829,
+ "learning_rate": 0.0006,
+ "loss": 4.2882866859436035,
+ "step": 5890
+ },
+ {
+ "epoch": 81.82445414847162,
+ "grad_norm": 0.014055311679840088,
+ "learning_rate": 0.0006,
+ "loss": 4.22351598739624,
+ "step": 5891
+ },
+ {
+ "epoch": 81.83842794759825,
+ "grad_norm": 0.013674065470695496,
+ "learning_rate": 0.0006,
+ "loss": 4.240716934204102,
+ "step": 5892
+ },
+ {
+ "epoch": 81.8524017467249,
+ "grad_norm": 0.01225196197628975,
+ "learning_rate": 0.0006,
+ "loss": 4.2958598136901855,
+ "step": 5893
+ },
+ {
+ "epoch": 81.86637554585153,
+ "grad_norm": 0.013538680970668793,
+ "learning_rate": 0.0006,
+ "loss": 4.24342679977417,
+ "step": 5894
+ },
+ {
+ "epoch": 81.88034934497817,
+ "grad_norm": 0.013151394203305244,
+ "learning_rate": 0.0006,
+ "loss": 4.081999778747559,
+ "step": 5895
+ },
+ {
+ "epoch": 81.8943231441048,
+ "grad_norm": 0.014260678552091122,
+ "learning_rate": 0.0006,
+ "loss": 4.156423568725586,
+ "step": 5896
+ },
+ {
+ "epoch": 81.90829694323143,
+ "grad_norm": 0.01329776830971241,
+ "learning_rate": 0.0006,
+ "loss": 4.1713995933532715,
+ "step": 5897
+ },
+ {
+ "epoch": 81.92227074235808,
+ "grad_norm": 0.014623276889324188,
+ "learning_rate": 0.0006,
+ "loss": 4.244891166687012,
+ "step": 5898
+ },
+ {
+ "epoch": 81.93624454148471,
+ "grad_norm": 0.015760324895381927,
+ "learning_rate": 0.0006,
+ "loss": 4.076528549194336,
+ "step": 5899
+ },
+ {
+ "epoch": 81.95021834061136,
+ "grad_norm": 0.014331766404211521,
+ "learning_rate": 0.0006,
+ "loss": 4.283037185668945,
+ "step": 5900
+ },
+ {
+ "epoch": 81.96419213973799,
+ "grad_norm": 0.012383842840790749,
+ "learning_rate": 0.0006,
+ "loss": 4.117308616638184,
+ "step": 5901
+ },
+ {
+ "epoch": 81.97816593886463,
+ "grad_norm": 0.01250660140067339,
+ "learning_rate": 0.0006,
+ "loss": 4.241250991821289,
+ "step": 5902
+ },
+ {
+ "epoch": 81.99213973799127,
+ "grad_norm": 0.012276900932192802,
+ "learning_rate": 0.0006,
+ "loss": 4.141091823577881,
+ "step": 5903
+ },
+ {
+ "epoch": 82.0,
+ "grad_norm": 0.01588737778365612,
+ "learning_rate": 0.0006,
+ "loss": 4.065532684326172,
+ "step": 5904
+ },
+ {
+ "epoch": 82.0,
+ "eval_loss": 4.612304210662842,
+ "eval_runtime": 57.5411,
+ "eval_samples_per_second": 42.439,
+ "eval_steps_per_second": 1.338,
+ "step": 5904
+ },
+ {
+ "epoch": 82.01397379912663,
+ "grad_norm": 0.017433544620871544,
+ "learning_rate": 0.0006,
+ "loss": 4.202497482299805,
+ "step": 5905
+ },
+ {
+ "epoch": 82.02794759825328,
+ "grad_norm": 0.017260415479540825,
+ "learning_rate": 0.0006,
+ "loss": 4.099949359893799,
+ "step": 5906
+ },
+ {
+ "epoch": 82.04192139737991,
+ "grad_norm": 0.01577908545732498,
+ "learning_rate": 0.0006,
+ "loss": 4.17806339263916,
+ "step": 5907
+ },
+ {
+ "epoch": 82.05589519650655,
+ "grad_norm": 0.016093818470835686,
+ "learning_rate": 0.0006,
+ "loss": 4.219491004943848,
+ "step": 5908
+ },
+ {
+ "epoch": 82.06986899563319,
+ "grad_norm": 0.015426302328705788,
+ "learning_rate": 0.0006,
+ "loss": 4.207982063293457,
+ "step": 5909
+ },
+ {
+ "epoch": 82.08384279475983,
+ "grad_norm": 0.01605442352592945,
+ "learning_rate": 0.0006,
+ "loss": 4.265942573547363,
+ "step": 5910
+ },
+ {
+ "epoch": 82.09781659388646,
+ "grad_norm": 0.01787225715816021,
+ "learning_rate": 0.0006,
+ "loss": 4.151535511016846,
+ "step": 5911
+ },
+ {
+ "epoch": 82.1117903930131,
+ "grad_norm": 0.01750164106488228,
+ "learning_rate": 0.0006,
+ "loss": 4.133655071258545,
+ "step": 5912
+ },
+ {
+ "epoch": 82.12576419213974,
+ "grad_norm": 0.01602957211434841,
+ "learning_rate": 0.0006,
+ "loss": 4.176972389221191,
+ "step": 5913
+ },
+ {
+ "epoch": 82.13973799126637,
+ "grad_norm": 0.01553812064230442,
+ "learning_rate": 0.0006,
+ "loss": 4.125692367553711,
+ "step": 5914
+ },
+ {
+ "epoch": 82.15371179039302,
+ "grad_norm": 0.013978508301079273,
+ "learning_rate": 0.0006,
+ "loss": 4.207271575927734,
+ "step": 5915
+ },
+ {
+ "epoch": 82.16768558951965,
+ "grad_norm": 0.013561118394136429,
+ "learning_rate": 0.0006,
+ "loss": 4.238930702209473,
+ "step": 5916
+ },
+ {
+ "epoch": 82.18165938864628,
+ "grad_norm": 0.013285507448017597,
+ "learning_rate": 0.0006,
+ "loss": 4.129464626312256,
+ "step": 5917
+ },
+ {
+ "epoch": 82.19563318777293,
+ "grad_norm": 0.013903839513659477,
+ "learning_rate": 0.0006,
+ "loss": 4.214109420776367,
+ "step": 5918
+ },
+ {
+ "epoch": 82.20960698689956,
+ "grad_norm": 0.014426837675273418,
+ "learning_rate": 0.0006,
+ "loss": 4.143196105957031,
+ "step": 5919
+ },
+ {
+ "epoch": 82.2235807860262,
+ "grad_norm": 0.015347165986895561,
+ "learning_rate": 0.0006,
+ "loss": 4.235439300537109,
+ "step": 5920
+ },
+ {
+ "epoch": 82.23755458515284,
+ "grad_norm": 0.014266039244830608,
+ "learning_rate": 0.0006,
+ "loss": 3.991617202758789,
+ "step": 5921
+ },
+ {
+ "epoch": 82.25152838427948,
+ "grad_norm": 0.013101593591272831,
+ "learning_rate": 0.0006,
+ "loss": 4.266017913818359,
+ "step": 5922
+ },
+ {
+ "epoch": 82.26550218340611,
+ "grad_norm": 0.013741881586611271,
+ "learning_rate": 0.0006,
+ "loss": 4.224879741668701,
+ "step": 5923
+ },
+ {
+ "epoch": 82.27947598253274,
+ "grad_norm": 0.014206945896148682,
+ "learning_rate": 0.0006,
+ "loss": 4.102272987365723,
+ "step": 5924
+ },
+ {
+ "epoch": 82.29344978165939,
+ "grad_norm": 0.013798723928630352,
+ "learning_rate": 0.0006,
+ "loss": 4.30418586730957,
+ "step": 5925
+ },
+ {
+ "epoch": 82.30742358078602,
+ "grad_norm": 0.015470122918486595,
+ "learning_rate": 0.0006,
+ "loss": 4.2513322830200195,
+ "step": 5926
+ },
+ {
+ "epoch": 82.32139737991267,
+ "grad_norm": 0.01655442826449871,
+ "learning_rate": 0.0006,
+ "loss": 4.223002910614014,
+ "step": 5927
+ },
+ {
+ "epoch": 82.3353711790393,
+ "grad_norm": 0.017170004546642303,
+ "learning_rate": 0.0006,
+ "loss": 4.253904819488525,
+ "step": 5928
+ },
+ {
+ "epoch": 82.34934497816595,
+ "grad_norm": 0.017595812678337097,
+ "learning_rate": 0.0006,
+ "loss": 4.184927940368652,
+ "step": 5929
+ },
+ {
+ "epoch": 82.36331877729258,
+ "grad_norm": 0.015073256567120552,
+ "learning_rate": 0.0006,
+ "loss": 4.202239036560059,
+ "step": 5930
+ },
+ {
+ "epoch": 82.37729257641921,
+ "grad_norm": 0.014310941100120544,
+ "learning_rate": 0.0006,
+ "loss": 4.124934673309326,
+ "step": 5931
+ },
+ {
+ "epoch": 82.39126637554585,
+ "grad_norm": 0.013671821914613247,
+ "learning_rate": 0.0006,
+ "loss": 4.138919353485107,
+ "step": 5932
+ },
+ {
+ "epoch": 82.40524017467249,
+ "grad_norm": 0.014103314839303493,
+ "learning_rate": 0.0006,
+ "loss": 4.211289405822754,
+ "step": 5933
+ },
+ {
+ "epoch": 82.41921397379913,
+ "grad_norm": 0.014272668398916721,
+ "learning_rate": 0.0006,
+ "loss": 4.06290340423584,
+ "step": 5934
+ },
+ {
+ "epoch": 82.43318777292576,
+ "grad_norm": 0.01387016475200653,
+ "learning_rate": 0.0006,
+ "loss": 4.101334571838379,
+ "step": 5935
+ },
+ {
+ "epoch": 82.44716157205241,
+ "grad_norm": 0.013318683952093124,
+ "learning_rate": 0.0006,
+ "loss": 4.206409931182861,
+ "step": 5936
+ },
+ {
+ "epoch": 82.46113537117904,
+ "grad_norm": 0.013781268149614334,
+ "learning_rate": 0.0006,
+ "loss": 4.189474105834961,
+ "step": 5937
+ },
+ {
+ "epoch": 82.47510917030567,
+ "grad_norm": 0.013195851817727089,
+ "learning_rate": 0.0006,
+ "loss": 4.276607036590576,
+ "step": 5938
+ },
+ {
+ "epoch": 82.48908296943232,
+ "grad_norm": 0.014358163811266422,
+ "learning_rate": 0.0006,
+ "loss": 4.203262805938721,
+ "step": 5939
+ },
+ {
+ "epoch": 82.50305676855895,
+ "grad_norm": 0.01403636671602726,
+ "learning_rate": 0.0006,
+ "loss": 4.079258441925049,
+ "step": 5940
+ },
+ {
+ "epoch": 82.5170305676856,
+ "grad_norm": 0.013272535055875778,
+ "learning_rate": 0.0006,
+ "loss": 4.19197940826416,
+ "step": 5941
+ },
+ {
+ "epoch": 82.53100436681223,
+ "grad_norm": 0.013787264935672283,
+ "learning_rate": 0.0006,
+ "loss": 4.12995719909668,
+ "step": 5942
+ },
+ {
+ "epoch": 82.54497816593886,
+ "grad_norm": 0.015020057559013367,
+ "learning_rate": 0.0006,
+ "loss": 4.148890495300293,
+ "step": 5943
+ },
+ {
+ "epoch": 82.5589519650655,
+ "grad_norm": 0.014199157245457172,
+ "learning_rate": 0.0006,
+ "loss": 4.201516151428223,
+ "step": 5944
+ },
+ {
+ "epoch": 82.57292576419214,
+ "grad_norm": 0.013827123679220676,
+ "learning_rate": 0.0006,
+ "loss": 4.240894317626953,
+ "step": 5945
+ },
+ {
+ "epoch": 82.58689956331878,
+ "grad_norm": 0.014767560176551342,
+ "learning_rate": 0.0006,
+ "loss": 4.173320770263672,
+ "step": 5946
+ },
+ {
+ "epoch": 82.60087336244541,
+ "grad_norm": 0.01612078957259655,
+ "learning_rate": 0.0006,
+ "loss": 4.266448020935059,
+ "step": 5947
+ },
+ {
+ "epoch": 82.61484716157206,
+ "grad_norm": 0.016706543043255806,
+ "learning_rate": 0.0006,
+ "loss": 4.229398250579834,
+ "step": 5948
+ },
+ {
+ "epoch": 82.62882096069869,
+ "grad_norm": 0.015630796551704407,
+ "learning_rate": 0.0006,
+ "loss": 4.181282997131348,
+ "step": 5949
+ },
+ {
+ "epoch": 82.64279475982532,
+ "grad_norm": 0.013103844597935677,
+ "learning_rate": 0.0006,
+ "loss": 4.309617042541504,
+ "step": 5950
+ },
+ {
+ "epoch": 82.65676855895197,
+ "grad_norm": 0.016136541962623596,
+ "learning_rate": 0.0006,
+ "loss": 4.224159240722656,
+ "step": 5951
+ },
+ {
+ "epoch": 82.6707423580786,
+ "grad_norm": 0.016160814091563225,
+ "learning_rate": 0.0006,
+ "loss": 4.267970085144043,
+ "step": 5952
+ },
+ {
+ "epoch": 82.68471615720524,
+ "grad_norm": 0.01672680303454399,
+ "learning_rate": 0.0006,
+ "loss": 4.230506896972656,
+ "step": 5953
+ },
+ {
+ "epoch": 82.69868995633188,
+ "grad_norm": 0.0159586351364851,
+ "learning_rate": 0.0006,
+ "loss": 4.159542083740234,
+ "step": 5954
+ },
+ {
+ "epoch": 82.71266375545852,
+ "grad_norm": 0.01461219135671854,
+ "learning_rate": 0.0006,
+ "loss": 4.27863073348999,
+ "step": 5955
+ },
+ {
+ "epoch": 82.72663755458515,
+ "grad_norm": 0.015088613145053387,
+ "learning_rate": 0.0006,
+ "loss": 4.272418975830078,
+ "step": 5956
+ },
+ {
+ "epoch": 82.74061135371178,
+ "grad_norm": 0.01594499684870243,
+ "learning_rate": 0.0006,
+ "loss": 4.271326065063477,
+ "step": 5957
+ },
+ {
+ "epoch": 82.75458515283843,
+ "grad_norm": 0.016648447141051292,
+ "learning_rate": 0.0006,
+ "loss": 4.141478538513184,
+ "step": 5958
+ },
+ {
+ "epoch": 82.76855895196506,
+ "grad_norm": 0.01540316641330719,
+ "learning_rate": 0.0006,
+ "loss": 4.111814498901367,
+ "step": 5959
+ },
+ {
+ "epoch": 82.78253275109171,
+ "grad_norm": 0.01634364202618599,
+ "learning_rate": 0.0006,
+ "loss": 4.3113112449646,
+ "step": 5960
+ },
+ {
+ "epoch": 82.79650655021834,
+ "grad_norm": 0.014768977649509907,
+ "learning_rate": 0.0006,
+ "loss": 4.184217929840088,
+ "step": 5961
+ },
+ {
+ "epoch": 82.81048034934499,
+ "grad_norm": 0.013453755527734756,
+ "learning_rate": 0.0006,
+ "loss": 4.247729301452637,
+ "step": 5962
+ },
+ {
+ "epoch": 82.82445414847162,
+ "grad_norm": 0.014731907285749912,
+ "learning_rate": 0.0006,
+ "loss": 4.352839469909668,
+ "step": 5963
+ },
+ {
+ "epoch": 82.83842794759825,
+ "grad_norm": 0.016047345474362373,
+ "learning_rate": 0.0006,
+ "loss": 4.257283687591553,
+ "step": 5964
+ },
+ {
+ "epoch": 82.8524017467249,
+ "grad_norm": 0.017141170799732208,
+ "learning_rate": 0.0006,
+ "loss": 4.22850227355957,
+ "step": 5965
+ },
+ {
+ "epoch": 82.86637554585153,
+ "grad_norm": 0.01370396837592125,
+ "learning_rate": 0.0006,
+ "loss": 4.1169257164001465,
+ "step": 5966
+ },
+ {
+ "epoch": 82.88034934497817,
+ "grad_norm": 0.012060081586241722,
+ "learning_rate": 0.0006,
+ "loss": 4.195242404937744,
+ "step": 5967
+ },
+ {
+ "epoch": 82.8943231441048,
+ "grad_norm": 0.015528378076851368,
+ "learning_rate": 0.0006,
+ "loss": 4.294939994812012,
+ "step": 5968
+ },
+ {
+ "epoch": 82.90829694323143,
+ "grad_norm": 0.018151041120290756,
+ "learning_rate": 0.0006,
+ "loss": 4.137192726135254,
+ "step": 5969
+ },
+ {
+ "epoch": 82.92227074235808,
+ "grad_norm": 0.017895735800266266,
+ "learning_rate": 0.0006,
+ "loss": 4.174036026000977,
+ "step": 5970
+ },
+ {
+ "epoch": 82.93624454148471,
+ "grad_norm": 0.015610386617481709,
+ "learning_rate": 0.0006,
+ "loss": 4.230136871337891,
+ "step": 5971
+ },
+ {
+ "epoch": 82.95021834061136,
+ "grad_norm": 0.01339928898960352,
+ "learning_rate": 0.0006,
+ "loss": 4.199074745178223,
+ "step": 5972
+ },
+ {
+ "epoch": 82.96419213973799,
+ "grad_norm": 0.015454214997589588,
+ "learning_rate": 0.0006,
+ "loss": 4.216109752655029,
+ "step": 5973
+ },
+ {
+ "epoch": 82.97816593886463,
+ "grad_norm": 0.014978132210671902,
+ "learning_rate": 0.0006,
+ "loss": 4.131511688232422,
+ "step": 5974
+ },
+ {
+ "epoch": 82.99213973799127,
+ "grad_norm": 0.015611122362315655,
+ "learning_rate": 0.0006,
+ "loss": 4.246518135070801,
+ "step": 5975
+ },
+ {
+ "epoch": 83.0,
+ "grad_norm": 0.016682764515280724,
+ "learning_rate": 0.0006,
+ "loss": 4.110680103302002,
+ "step": 5976
+ },
+ {
+ "epoch": 83.0,
+ "eval_loss": 4.641707897186279,
+ "eval_runtime": 57.2799,
+ "eval_samples_per_second": 42.633,
+ "eval_steps_per_second": 1.344,
+ "step": 5976
+ },
+ {
+ "epoch": 83.01397379912663,
+ "grad_norm": 0.015007982030510902,
+ "learning_rate": 0.0006,
+ "loss": 4.100912094116211,
+ "step": 5977
+ },
+ {
+ "epoch": 83.02794759825328,
+ "grad_norm": 0.017609436064958572,
+ "learning_rate": 0.0006,
+ "loss": 4.233399868011475,
+ "step": 5978
+ },
+ {
+ "epoch": 83.04192139737991,
+ "grad_norm": 0.019817272201180458,
+ "learning_rate": 0.0006,
+ "loss": 4.191173076629639,
+ "step": 5979
+ },
+ {
+ "epoch": 83.05589519650655,
+ "grad_norm": 0.017752140760421753,
+ "learning_rate": 0.0006,
+ "loss": 4.044437885284424,
+ "step": 5980
+ },
+ {
+ "epoch": 83.06986899563319,
+ "grad_norm": 0.015482760034501553,
+ "learning_rate": 0.0006,
+ "loss": 4.319931983947754,
+ "step": 5981
+ },
+ {
+ "epoch": 83.08384279475983,
+ "grad_norm": 0.015153988264501095,
+ "learning_rate": 0.0006,
+ "loss": 4.095277309417725,
+ "step": 5982
+ },
+ {
+ "epoch": 83.09781659388646,
+ "grad_norm": 0.017025282606482506,
+ "learning_rate": 0.0006,
+ "loss": 4.235110759735107,
+ "step": 5983
+ },
+ {
+ "epoch": 83.1117903930131,
+ "grad_norm": 0.01566586270928383,
+ "learning_rate": 0.0006,
+ "loss": 4.1934685707092285,
+ "step": 5984
+ },
+ {
+ "epoch": 83.12576419213974,
+ "grad_norm": 0.016419263556599617,
+ "learning_rate": 0.0006,
+ "loss": 4.234419822692871,
+ "step": 5985
+ },
+ {
+ "epoch": 83.13973799126637,
+ "grad_norm": 0.017132675275206566,
+ "learning_rate": 0.0006,
+ "loss": 4.221447944641113,
+ "step": 5986
+ },
+ {
+ "epoch": 83.15371179039302,
+ "grad_norm": 0.016138629987835884,
+ "learning_rate": 0.0006,
+ "loss": 4.199830055236816,
+ "step": 5987
+ },
+ {
+ "epoch": 83.16768558951965,
+ "grad_norm": 0.016210174188017845,
+ "learning_rate": 0.0006,
+ "loss": 4.081717491149902,
+ "step": 5988
+ },
+ {
+ "epoch": 83.18165938864628,
+ "grad_norm": 0.016055308282375336,
+ "learning_rate": 0.0006,
+ "loss": 4.247655391693115,
+ "step": 5989
+ },
+ {
+ "epoch": 83.19563318777293,
+ "grad_norm": 0.016132952645421028,
+ "learning_rate": 0.0006,
+ "loss": 4.262535572052002,
+ "step": 5990
+ },
+ {
+ "epoch": 83.20960698689956,
+ "grad_norm": 0.01779935695230961,
+ "learning_rate": 0.0006,
+ "loss": 4.153120040893555,
+ "step": 5991
+ },
+ {
+ "epoch": 83.2235807860262,
+ "grad_norm": 0.018360646441578865,
+ "learning_rate": 0.0006,
+ "loss": 4.232644081115723,
+ "step": 5992
+ },
+ {
+ "epoch": 83.23755458515284,
+ "grad_norm": 0.019629675894975662,
+ "learning_rate": 0.0006,
+ "loss": 3.9758734703063965,
+ "step": 5993
+ },
+ {
+ "epoch": 83.25152838427948,
+ "grad_norm": 0.017819996923208237,
+ "learning_rate": 0.0006,
+ "loss": 4.196932315826416,
+ "step": 5994
+ },
+ {
+ "epoch": 83.26550218340611,
+ "grad_norm": 0.01557681243866682,
+ "learning_rate": 0.0006,
+ "loss": 4.122654914855957,
+ "step": 5995
+ },
+ {
+ "epoch": 83.27947598253274,
+ "grad_norm": 0.017858853563666344,
+ "learning_rate": 0.0006,
+ "loss": 4.185330867767334,
+ "step": 5996
+ },
+ {
+ "epoch": 83.29344978165939,
+ "grad_norm": 0.018065953627228737,
+ "learning_rate": 0.0006,
+ "loss": 4.218471050262451,
+ "step": 5997
+ },
+ {
+ "epoch": 83.30742358078602,
+ "grad_norm": 0.017247939482331276,
+ "learning_rate": 0.0006,
+ "loss": 4.127076148986816,
+ "step": 5998
+ },
+ {
+ "epoch": 83.32139737991267,
+ "grad_norm": 0.017047952860593796,
+ "learning_rate": 0.0006,
+ "loss": 4.131356716156006,
+ "step": 5999
+ },
+ {
+ "epoch": 83.3353711790393,
+ "grad_norm": 0.015905629843473434,
+ "learning_rate": 0.0006,
+ "loss": 4.2239789962768555,
+ "step": 6000
+ },
+ {
+ "epoch": 83.34934497816595,
+ "grad_norm": 0.016522347927093506,
+ "learning_rate": 0.0006,
+ "loss": 4.054897785186768,
+ "step": 6001
+ },
+ {
+ "epoch": 83.36331877729258,
+ "grad_norm": 0.01622571051120758,
+ "learning_rate": 0.0006,
+ "loss": 4.284210681915283,
+ "step": 6002
+ },
+ {
+ "epoch": 83.37729257641921,
+ "grad_norm": 0.01640467159450054,
+ "learning_rate": 0.0006,
+ "loss": 4.268840789794922,
+ "step": 6003
+ },
+ {
+ "epoch": 83.39126637554585,
+ "grad_norm": 0.015698859468102455,
+ "learning_rate": 0.0006,
+ "loss": 4.179793357849121,
+ "step": 6004
+ },
+ {
+ "epoch": 83.40524017467249,
+ "grad_norm": 0.016694528982043266,
+ "learning_rate": 0.0006,
+ "loss": 4.132297515869141,
+ "step": 6005
+ },
+ {
+ "epoch": 83.41921397379913,
+ "grad_norm": 0.015345900319516659,
+ "learning_rate": 0.0006,
+ "loss": 4.090163230895996,
+ "step": 6006
+ },
+ {
+ "epoch": 83.43318777292576,
+ "grad_norm": 0.014722758904099464,
+ "learning_rate": 0.0006,
+ "loss": 4.179166793823242,
+ "step": 6007
+ },
+ {
+ "epoch": 83.44716157205241,
+ "grad_norm": 0.014884887263178825,
+ "learning_rate": 0.0006,
+ "loss": 4.137659072875977,
+ "step": 6008
+ },
+ {
+ "epoch": 83.46113537117904,
+ "grad_norm": 0.014041902497410774,
+ "learning_rate": 0.0006,
+ "loss": 4.1202073097229,
+ "step": 6009
+ },
+ {
+ "epoch": 83.47510917030567,
+ "grad_norm": 0.012799333781003952,
+ "learning_rate": 0.0006,
+ "loss": 4.092607498168945,
+ "step": 6010
+ },
+ {
+ "epoch": 83.48908296943232,
+ "grad_norm": 0.013272707350552082,
+ "learning_rate": 0.0006,
+ "loss": 4.132715225219727,
+ "step": 6011
+ },
+ {
+ "epoch": 83.50305676855895,
+ "grad_norm": 0.012934371829032898,
+ "learning_rate": 0.0006,
+ "loss": 4.230278968811035,
+ "step": 6012
+ },
+ {
+ "epoch": 83.5170305676856,
+ "grad_norm": 0.0120006762444973,
+ "learning_rate": 0.0006,
+ "loss": 4.140068531036377,
+ "step": 6013
+ },
+ {
+ "epoch": 83.53100436681223,
+ "grad_norm": 0.013616700656712055,
+ "learning_rate": 0.0006,
+ "loss": 4.20107364654541,
+ "step": 6014
+ },
+ {
+ "epoch": 83.54497816593886,
+ "grad_norm": 0.013957851566374302,
+ "learning_rate": 0.0006,
+ "loss": 4.124945163726807,
+ "step": 6015
+ },
+ {
+ "epoch": 83.5589519650655,
+ "grad_norm": 0.012762055732309818,
+ "learning_rate": 0.0006,
+ "loss": 4.022052764892578,
+ "step": 6016
+ },
+ {
+ "epoch": 83.57292576419214,
+ "grad_norm": 0.013080576434731483,
+ "learning_rate": 0.0006,
+ "loss": 4.103252410888672,
+ "step": 6017
+ },
+ {
+ "epoch": 83.58689956331878,
+ "grad_norm": 0.013957500457763672,
+ "learning_rate": 0.0006,
+ "loss": 4.235089302062988,
+ "step": 6018
+ },
+ {
+ "epoch": 83.60087336244541,
+ "grad_norm": 0.014707164838910103,
+ "learning_rate": 0.0006,
+ "loss": 4.191570281982422,
+ "step": 6019
+ },
+ {
+ "epoch": 83.61484716157206,
+ "grad_norm": 0.012807869352400303,
+ "learning_rate": 0.0006,
+ "loss": 4.294543266296387,
+ "step": 6020
+ },
+ {
+ "epoch": 83.62882096069869,
+ "grad_norm": 0.012027994729578495,
+ "learning_rate": 0.0006,
+ "loss": 4.13849401473999,
+ "step": 6021
+ },
+ {
+ "epoch": 83.64279475982532,
+ "grad_norm": 0.012147724628448486,
+ "learning_rate": 0.0006,
+ "loss": 4.127481937408447,
+ "step": 6022
+ },
+ {
+ "epoch": 83.65676855895197,
+ "grad_norm": 0.013241392560303211,
+ "learning_rate": 0.0006,
+ "loss": 4.218972682952881,
+ "step": 6023
+ },
+ {
+ "epoch": 83.6707423580786,
+ "grad_norm": 0.014071648940443993,
+ "learning_rate": 0.0006,
+ "loss": 4.096724033355713,
+ "step": 6024
+ },
+ {
+ "epoch": 83.68471615720524,
+ "grad_norm": 0.015212797559797764,
+ "learning_rate": 0.0006,
+ "loss": 4.27097225189209,
+ "step": 6025
+ },
+ {
+ "epoch": 83.69868995633188,
+ "grad_norm": 0.017647385597229004,
+ "learning_rate": 0.0006,
+ "loss": 4.209835529327393,
+ "step": 6026
+ },
+ {
+ "epoch": 83.71266375545852,
+ "grad_norm": 0.01709647662937641,
+ "learning_rate": 0.0006,
+ "loss": 4.042106628417969,
+ "step": 6027
+ },
+ {
+ "epoch": 83.72663755458515,
+ "grad_norm": 0.014530927874147892,
+ "learning_rate": 0.0006,
+ "loss": 4.181731700897217,
+ "step": 6028
+ },
+ {
+ "epoch": 83.74061135371178,
+ "grad_norm": 0.015356711111962795,
+ "learning_rate": 0.0006,
+ "loss": 4.314183235168457,
+ "step": 6029
+ },
+ {
+ "epoch": 83.75458515283843,
+ "grad_norm": 0.015500029549002647,
+ "learning_rate": 0.0006,
+ "loss": 4.06776762008667,
+ "step": 6030
+ },
+ {
+ "epoch": 83.76855895196506,
+ "grad_norm": 0.015351332724094391,
+ "learning_rate": 0.0006,
+ "loss": 4.185141563415527,
+ "step": 6031
+ },
+ {
+ "epoch": 83.78253275109171,
+ "grad_norm": 0.016678154468536377,
+ "learning_rate": 0.0006,
+ "loss": 4.179529666900635,
+ "step": 6032
+ },
+ {
+ "epoch": 83.79650655021834,
+ "grad_norm": 0.0156265702098608,
+ "learning_rate": 0.0006,
+ "loss": 4.238584041595459,
+ "step": 6033
+ },
+ {
+ "epoch": 83.81048034934499,
+ "grad_norm": 0.016418596729636192,
+ "learning_rate": 0.0006,
+ "loss": 4.1593523025512695,
+ "step": 6034
+ },
+ {
+ "epoch": 83.82445414847162,
+ "grad_norm": 0.016113586723804474,
+ "learning_rate": 0.0006,
+ "loss": 4.157470226287842,
+ "step": 6035
+ },
+ {
+ "epoch": 83.83842794759825,
+ "grad_norm": 0.014434042386710644,
+ "learning_rate": 0.0006,
+ "loss": 4.196226119995117,
+ "step": 6036
+ },
+ {
+ "epoch": 83.8524017467249,
+ "grad_norm": 0.012647042982280254,
+ "learning_rate": 0.0006,
+ "loss": 4.130589485168457,
+ "step": 6037
+ },
+ {
+ "epoch": 83.86637554585153,
+ "grad_norm": 0.014412877149879932,
+ "learning_rate": 0.0006,
+ "loss": 4.154845237731934,
+ "step": 6038
+ },
+ {
+ "epoch": 83.88034934497817,
+ "grad_norm": 0.014222027733922005,
+ "learning_rate": 0.0006,
+ "loss": 4.189724922180176,
+ "step": 6039
+ },
+ {
+ "epoch": 83.8943231441048,
+ "grad_norm": 0.013084255158901215,
+ "learning_rate": 0.0006,
+ "loss": 4.303338527679443,
+ "step": 6040
+ },
+ {
+ "epoch": 83.90829694323143,
+ "grad_norm": 0.012659232132136822,
+ "learning_rate": 0.0006,
+ "loss": 4.08192253112793,
+ "step": 6041
+ },
+ {
+ "epoch": 83.92227074235808,
+ "grad_norm": 0.01246276218444109,
+ "learning_rate": 0.0006,
+ "loss": 4.283324718475342,
+ "step": 6042
+ },
+ {
+ "epoch": 83.93624454148471,
+ "grad_norm": 0.01340118795633316,
+ "learning_rate": 0.0006,
+ "loss": 4.117908477783203,
+ "step": 6043
+ },
+ {
+ "epoch": 83.95021834061136,
+ "grad_norm": 0.014496722258627415,
+ "learning_rate": 0.0006,
+ "loss": 4.110318660736084,
+ "step": 6044
+ },
+ {
+ "epoch": 83.96419213973799,
+ "grad_norm": 0.013348298147320747,
+ "learning_rate": 0.0006,
+ "loss": 4.185139179229736,
+ "step": 6045
+ },
+ {
+ "epoch": 83.97816593886463,
+ "grad_norm": 0.013294277712702751,
+ "learning_rate": 0.0006,
+ "loss": 4.271645545959473,
+ "step": 6046
+ },
+ {
+ "epoch": 83.99213973799127,
+ "grad_norm": 0.014431594870984554,
+ "learning_rate": 0.0006,
+ "loss": 4.161383152008057,
+ "step": 6047
+ },
+ {
+ "epoch": 84.0,
+ "grad_norm": 0.016268927603960037,
+ "learning_rate": 0.0006,
+ "loss": 4.178324222564697,
+ "step": 6048
+ },
+ {
+ "epoch": 84.0,
+ "eval_loss": 4.6307549476623535,
+ "eval_runtime": 56.8521,
+ "eval_samples_per_second": 42.954,
+ "eval_steps_per_second": 1.354,
+ "step": 6048
+ },
+ {
+ "epoch": 84.01397379912663,
+ "grad_norm": 0.014864908531308174,
+ "learning_rate": 0.0006,
+ "loss": 4.121300220489502,
+ "step": 6049
+ },
+ {
+ "epoch": 84.02794759825328,
+ "grad_norm": 0.014863367192447186,
+ "learning_rate": 0.0006,
+ "loss": 4.1650543212890625,
+ "step": 6050
+ },
+ {
+ "epoch": 84.04192139737991,
+ "grad_norm": 0.014379646629095078,
+ "learning_rate": 0.0006,
+ "loss": 4.0201239585876465,
+ "step": 6051
+ },
+ {
+ "epoch": 84.05589519650655,
+ "grad_norm": 0.014871201477944851,
+ "learning_rate": 0.0006,
+ "loss": 4.0820698738098145,
+ "step": 6052
+ },
+ {
+ "epoch": 84.06986899563319,
+ "grad_norm": 0.013085179962217808,
+ "learning_rate": 0.0006,
+ "loss": 4.223984718322754,
+ "step": 6053
+ },
+ {
+ "epoch": 84.08384279475983,
+ "grad_norm": 0.012485949322581291,
+ "learning_rate": 0.0006,
+ "loss": 4.182206153869629,
+ "step": 6054
+ },
+ {
+ "epoch": 84.09781659388646,
+ "grad_norm": 0.013653291389346123,
+ "learning_rate": 0.0006,
+ "loss": 4.142557144165039,
+ "step": 6055
+ },
+ {
+ "epoch": 84.1117903930131,
+ "grad_norm": 0.013930532149970531,
+ "learning_rate": 0.0006,
+ "loss": 4.13972282409668,
+ "step": 6056
+ },
+ {
+ "epoch": 84.12576419213974,
+ "grad_norm": 0.01297379657626152,
+ "learning_rate": 0.0006,
+ "loss": 4.144664764404297,
+ "step": 6057
+ },
+ {
+ "epoch": 84.13973799126637,
+ "grad_norm": 0.013210492208600044,
+ "learning_rate": 0.0006,
+ "loss": 4.253764629364014,
+ "step": 6058
+ },
+ {
+ "epoch": 84.15371179039302,
+ "grad_norm": 0.014650176279246807,
+ "learning_rate": 0.0006,
+ "loss": 4.246866226196289,
+ "step": 6059
+ },
+ {
+ "epoch": 84.16768558951965,
+ "grad_norm": 0.01482292078435421,
+ "learning_rate": 0.0006,
+ "loss": 4.140187740325928,
+ "step": 6060
+ },
+ {
+ "epoch": 84.18165938864628,
+ "grad_norm": 0.01592552848160267,
+ "learning_rate": 0.0006,
+ "loss": 4.170705318450928,
+ "step": 6061
+ },
+ {
+ "epoch": 84.19563318777293,
+ "grad_norm": 0.017152156680822372,
+ "learning_rate": 0.0006,
+ "loss": 4.126092433929443,
+ "step": 6062
+ },
+ {
+ "epoch": 84.20960698689956,
+ "grad_norm": 0.015880905091762543,
+ "learning_rate": 0.0006,
+ "loss": 4.161938190460205,
+ "step": 6063
+ },
+ {
+ "epoch": 84.2235807860262,
+ "grad_norm": 0.015451471321284771,
+ "learning_rate": 0.0006,
+ "loss": 4.263461112976074,
+ "step": 6064
+ },
+ {
+ "epoch": 84.23755458515284,
+ "grad_norm": 0.014365996234118938,
+ "learning_rate": 0.0006,
+ "loss": 4.14488410949707,
+ "step": 6065
+ },
+ {
+ "epoch": 84.25152838427948,
+ "grad_norm": 0.016364749521017075,
+ "learning_rate": 0.0006,
+ "loss": 4.112489700317383,
+ "step": 6066
+ },
+ {
+ "epoch": 84.26550218340611,
+ "grad_norm": 0.01791754551231861,
+ "learning_rate": 0.0006,
+ "loss": 4.133513450622559,
+ "step": 6067
+ },
+ {
+ "epoch": 84.27947598253274,
+ "grad_norm": 0.016609208658337593,
+ "learning_rate": 0.0006,
+ "loss": 4.201829433441162,
+ "step": 6068
+ },
+ {
+ "epoch": 84.29344978165939,
+ "grad_norm": 0.017587820068001747,
+ "learning_rate": 0.0006,
+ "loss": 4.203151702880859,
+ "step": 6069
+ },
+ {
+ "epoch": 84.30742358078602,
+ "grad_norm": 0.017047742381691933,
+ "learning_rate": 0.0006,
+ "loss": 4.077008247375488,
+ "step": 6070
+ },
+ {
+ "epoch": 84.32139737991267,
+ "grad_norm": 0.01524646207690239,
+ "learning_rate": 0.0006,
+ "loss": 4.198038101196289,
+ "step": 6071
+ },
+ {
+ "epoch": 84.3353711790393,
+ "grad_norm": 0.014196421951055527,
+ "learning_rate": 0.0006,
+ "loss": 4.164245128631592,
+ "step": 6072
+ },
+ {
+ "epoch": 84.34934497816595,
+ "grad_norm": 0.016291776672005653,
+ "learning_rate": 0.0006,
+ "loss": 4.150152206420898,
+ "step": 6073
+ },
+ {
+ "epoch": 84.36331877729258,
+ "grad_norm": 0.015627920627593994,
+ "learning_rate": 0.0006,
+ "loss": 4.110006332397461,
+ "step": 6074
+ },
+ {
+ "epoch": 84.37729257641921,
+ "grad_norm": 0.013251487165689468,
+ "learning_rate": 0.0006,
+ "loss": 4.101463317871094,
+ "step": 6075
+ },
+ {
+ "epoch": 84.39126637554585,
+ "grad_norm": 0.013601605780422688,
+ "learning_rate": 0.0006,
+ "loss": 4.191105842590332,
+ "step": 6076
+ },
+ {
+ "epoch": 84.40524017467249,
+ "grad_norm": 0.014878741465508938,
+ "learning_rate": 0.0006,
+ "loss": 4.153046131134033,
+ "step": 6077
+ },
+ {
+ "epoch": 84.41921397379913,
+ "grad_norm": 0.014783897437155247,
+ "learning_rate": 0.0006,
+ "loss": 4.362000465393066,
+ "step": 6078
+ },
+ {
+ "epoch": 84.43318777292576,
+ "grad_norm": 0.014221332035958767,
+ "learning_rate": 0.0006,
+ "loss": 4.144364833831787,
+ "step": 6079
+ },
+ {
+ "epoch": 84.44716157205241,
+ "grad_norm": 0.014911034144461155,
+ "learning_rate": 0.0006,
+ "loss": 4.088857650756836,
+ "step": 6080
+ },
+ {
+ "epoch": 84.46113537117904,
+ "grad_norm": 0.014924770221114159,
+ "learning_rate": 0.0006,
+ "loss": 4.1677446365356445,
+ "step": 6081
+ },
+ {
+ "epoch": 84.47510917030567,
+ "grad_norm": 0.014394409023225307,
+ "learning_rate": 0.0006,
+ "loss": 4.087103843688965,
+ "step": 6082
+ },
+ {
+ "epoch": 84.48908296943232,
+ "grad_norm": 0.013926676474511623,
+ "learning_rate": 0.0006,
+ "loss": 3.9972851276397705,
+ "step": 6083
+ },
+ {
+ "epoch": 84.50305676855895,
+ "grad_norm": 0.013069954700767994,
+ "learning_rate": 0.0006,
+ "loss": 3.998237133026123,
+ "step": 6084
+ },
+ {
+ "epoch": 84.5170305676856,
+ "grad_norm": 0.012748011387884617,
+ "learning_rate": 0.0006,
+ "loss": 4.195858955383301,
+ "step": 6085
+ },
+ {
+ "epoch": 84.53100436681223,
+ "grad_norm": 0.012278077192604542,
+ "learning_rate": 0.0006,
+ "loss": 4.102165222167969,
+ "step": 6086
+ },
+ {
+ "epoch": 84.54497816593886,
+ "grad_norm": 0.014259163290262222,
+ "learning_rate": 0.0006,
+ "loss": 4.092750549316406,
+ "step": 6087
+ },
+ {
+ "epoch": 84.5589519650655,
+ "grad_norm": 0.014553683809936047,
+ "learning_rate": 0.0006,
+ "loss": 4.1412763595581055,
+ "step": 6088
+ },
+ {
+ "epoch": 84.57292576419214,
+ "grad_norm": 0.013605243526399136,
+ "learning_rate": 0.0006,
+ "loss": 4.192173480987549,
+ "step": 6089
+ },
+ {
+ "epoch": 84.58689956331878,
+ "grad_norm": 0.014259358868002892,
+ "learning_rate": 0.0006,
+ "loss": 4.356374740600586,
+ "step": 6090
+ },
+ {
+ "epoch": 84.60087336244541,
+ "grad_norm": 0.014166025444865227,
+ "learning_rate": 0.0006,
+ "loss": 4.165982246398926,
+ "step": 6091
+ },
+ {
+ "epoch": 84.61484716157206,
+ "grad_norm": 0.012198803946375847,
+ "learning_rate": 0.0006,
+ "loss": 4.287566184997559,
+ "step": 6092
+ },
+ {
+ "epoch": 84.62882096069869,
+ "grad_norm": 0.013794238679111004,
+ "learning_rate": 0.0006,
+ "loss": 4.199882984161377,
+ "step": 6093
+ },
+ {
+ "epoch": 84.64279475982532,
+ "grad_norm": 0.014757642522454262,
+ "learning_rate": 0.0006,
+ "loss": 4.116573333740234,
+ "step": 6094
+ },
+ {
+ "epoch": 84.65676855895197,
+ "grad_norm": 0.014591689221560955,
+ "learning_rate": 0.0006,
+ "loss": 4.240896224975586,
+ "step": 6095
+ },
+ {
+ "epoch": 84.6707423580786,
+ "grad_norm": 0.015217579901218414,
+ "learning_rate": 0.0006,
+ "loss": 4.257142066955566,
+ "step": 6096
+ },
+ {
+ "epoch": 84.68471615720524,
+ "grad_norm": 0.015371371991932392,
+ "learning_rate": 0.0006,
+ "loss": 4.181037425994873,
+ "step": 6097
+ },
+ {
+ "epoch": 84.69868995633188,
+ "grad_norm": 0.014412983320653439,
+ "learning_rate": 0.0006,
+ "loss": 4.2207794189453125,
+ "step": 6098
+ },
+ {
+ "epoch": 84.71266375545852,
+ "grad_norm": 0.013773671351373196,
+ "learning_rate": 0.0006,
+ "loss": 4.211123466491699,
+ "step": 6099
+ },
+ {
+ "epoch": 84.72663755458515,
+ "grad_norm": 0.012838046066462994,
+ "learning_rate": 0.0006,
+ "loss": 4.175459384918213,
+ "step": 6100
+ },
+ {
+ "epoch": 84.74061135371178,
+ "grad_norm": 0.014094929210841656,
+ "learning_rate": 0.0006,
+ "loss": 4.1724958419799805,
+ "step": 6101
+ },
+ {
+ "epoch": 84.75458515283843,
+ "grad_norm": 0.014573262073099613,
+ "learning_rate": 0.0006,
+ "loss": 4.15885066986084,
+ "step": 6102
+ },
+ {
+ "epoch": 84.76855895196506,
+ "grad_norm": 0.011964819394052029,
+ "learning_rate": 0.0006,
+ "loss": 3.9804892539978027,
+ "step": 6103
+ },
+ {
+ "epoch": 84.78253275109171,
+ "grad_norm": 0.014140539802610874,
+ "learning_rate": 0.0006,
+ "loss": 4.203708648681641,
+ "step": 6104
+ },
+ {
+ "epoch": 84.79650655021834,
+ "grad_norm": 0.016618940979242325,
+ "learning_rate": 0.0006,
+ "loss": 4.147684097290039,
+ "step": 6105
+ },
+ {
+ "epoch": 84.81048034934499,
+ "grad_norm": 0.015094385482370853,
+ "learning_rate": 0.0006,
+ "loss": 4.392223834991455,
+ "step": 6106
+ },
+ {
+ "epoch": 84.82445414847162,
+ "grad_norm": 0.014386717230081558,
+ "learning_rate": 0.0006,
+ "loss": 4.253332138061523,
+ "step": 6107
+ },
+ {
+ "epoch": 84.83842794759825,
+ "grad_norm": 0.013877755962312222,
+ "learning_rate": 0.0006,
+ "loss": 4.075056076049805,
+ "step": 6108
+ },
+ {
+ "epoch": 84.8524017467249,
+ "grad_norm": 0.014062155969440937,
+ "learning_rate": 0.0006,
+ "loss": 4.072048664093018,
+ "step": 6109
+ },
+ {
+ "epoch": 84.86637554585153,
+ "grad_norm": 0.014445619657635689,
+ "learning_rate": 0.0006,
+ "loss": 4.151832580566406,
+ "step": 6110
+ },
+ {
+ "epoch": 84.88034934497817,
+ "grad_norm": 0.015116363763809204,
+ "learning_rate": 0.0006,
+ "loss": 4.088900566101074,
+ "step": 6111
+ },
+ {
+ "epoch": 84.8943231441048,
+ "grad_norm": 0.014425474219024181,
+ "learning_rate": 0.0006,
+ "loss": 4.200483322143555,
+ "step": 6112
+ },
+ {
+ "epoch": 84.90829694323143,
+ "grad_norm": 0.01449573878198862,
+ "learning_rate": 0.0006,
+ "loss": 4.239742279052734,
+ "step": 6113
+ },
+ {
+ "epoch": 84.92227074235808,
+ "grad_norm": 0.015698261559009552,
+ "learning_rate": 0.0006,
+ "loss": 4.197391510009766,
+ "step": 6114
+ },
+ {
+ "epoch": 84.93624454148471,
+ "grad_norm": 0.014545733109116554,
+ "learning_rate": 0.0006,
+ "loss": 4.035508155822754,
+ "step": 6115
+ },
+ {
+ "epoch": 84.95021834061136,
+ "grad_norm": 0.012796483002603054,
+ "learning_rate": 0.0006,
+ "loss": 4.195248126983643,
+ "step": 6116
+ },
+ {
+ "epoch": 84.96419213973799,
+ "grad_norm": 0.01264485064893961,
+ "learning_rate": 0.0006,
+ "loss": 4.158118724822998,
+ "step": 6117
+ },
+ {
+ "epoch": 84.97816593886463,
+ "grad_norm": 0.013695276342332363,
+ "learning_rate": 0.0006,
+ "loss": 4.255288124084473,
+ "step": 6118
+ },
+ {
+ "epoch": 84.99213973799127,
+ "grad_norm": 0.013459564186632633,
+ "learning_rate": 0.0006,
+ "loss": 4.204586505889893,
+ "step": 6119
+ },
+ {
+ "epoch": 85.0,
+ "grad_norm": 0.014312652871012688,
+ "learning_rate": 0.0006,
+ "loss": 4.167879104614258,
+ "step": 6120
+ },
+ {
+ "epoch": 85.0,
+ "eval_loss": 4.608152389526367,
+ "eval_runtime": 57.1903,
+ "eval_samples_per_second": 42.7,
+ "eval_steps_per_second": 1.346,
+ "step": 6120
+ },
+ {
+ "epoch": 85.01397379912663,
+ "grad_norm": 0.013918368145823479,
+ "learning_rate": 0.0006,
+ "loss": 4.129991054534912,
+ "step": 6121
+ },
+ {
+ "epoch": 85.02794759825328,
+ "grad_norm": 0.01604384370148182,
+ "learning_rate": 0.0006,
+ "loss": 4.233926773071289,
+ "step": 6122
+ },
+ {
+ "epoch": 85.04192139737991,
+ "grad_norm": 0.016565652564167976,
+ "learning_rate": 0.0006,
+ "loss": 4.129604339599609,
+ "step": 6123
+ },
+ {
+ "epoch": 85.05589519650655,
+ "grad_norm": 0.0149751678109169,
+ "learning_rate": 0.0006,
+ "loss": 4.165839195251465,
+ "step": 6124
+ },
+ {
+ "epoch": 85.06986899563319,
+ "grad_norm": 0.013812461867928505,
+ "learning_rate": 0.0006,
+ "loss": 4.153792858123779,
+ "step": 6125
+ },
+ {
+ "epoch": 85.08384279475983,
+ "grad_norm": 0.014432745985686779,
+ "learning_rate": 0.0006,
+ "loss": 4.170895576477051,
+ "step": 6126
+ },
+ {
+ "epoch": 85.09781659388646,
+ "grad_norm": 0.015208868309855461,
+ "learning_rate": 0.0006,
+ "loss": 4.10639762878418,
+ "step": 6127
+ },
+ {
+ "epoch": 85.1117903930131,
+ "grad_norm": 0.013209272176027298,
+ "learning_rate": 0.0006,
+ "loss": 4.200364112854004,
+ "step": 6128
+ },
+ {
+ "epoch": 85.12576419213974,
+ "grad_norm": 0.014807002618908882,
+ "learning_rate": 0.0006,
+ "loss": 4.172666549682617,
+ "step": 6129
+ },
+ {
+ "epoch": 85.13973799126637,
+ "grad_norm": 0.015662945806980133,
+ "learning_rate": 0.0006,
+ "loss": 4.145219326019287,
+ "step": 6130
+ },
+ {
+ "epoch": 85.15371179039302,
+ "grad_norm": 0.014598497189581394,
+ "learning_rate": 0.0006,
+ "loss": 4.275618553161621,
+ "step": 6131
+ },
+ {
+ "epoch": 85.16768558951965,
+ "grad_norm": 0.013778576627373695,
+ "learning_rate": 0.0006,
+ "loss": 4.204716682434082,
+ "step": 6132
+ },
+ {
+ "epoch": 85.18165938864628,
+ "grad_norm": 0.015762265771627426,
+ "learning_rate": 0.0006,
+ "loss": 4.145524024963379,
+ "step": 6133
+ },
+ {
+ "epoch": 85.19563318777293,
+ "grad_norm": 0.015579195693135262,
+ "learning_rate": 0.0006,
+ "loss": 4.150625228881836,
+ "step": 6134
+ },
+ {
+ "epoch": 85.20960698689956,
+ "grad_norm": 0.016641216352581978,
+ "learning_rate": 0.0006,
+ "loss": 4.141819953918457,
+ "step": 6135
+ },
+ {
+ "epoch": 85.2235807860262,
+ "grad_norm": 0.01632271707057953,
+ "learning_rate": 0.0006,
+ "loss": 4.030620574951172,
+ "step": 6136
+ },
+ {
+ "epoch": 85.23755458515284,
+ "grad_norm": 0.015564918518066406,
+ "learning_rate": 0.0006,
+ "loss": 4.147753715515137,
+ "step": 6137
+ },
+ {
+ "epoch": 85.25152838427948,
+ "grad_norm": 0.0161129217594862,
+ "learning_rate": 0.0006,
+ "loss": 4.128524303436279,
+ "step": 6138
+ },
+ {
+ "epoch": 85.26550218340611,
+ "grad_norm": 0.017766550183296204,
+ "learning_rate": 0.0006,
+ "loss": 4.134856224060059,
+ "step": 6139
+ },
+ {
+ "epoch": 85.27947598253274,
+ "grad_norm": 0.01869054324924946,
+ "learning_rate": 0.0006,
+ "loss": 4.231565475463867,
+ "step": 6140
+ },
+ {
+ "epoch": 85.29344978165939,
+ "grad_norm": 0.01917010359466076,
+ "learning_rate": 0.0006,
+ "loss": 4.21175479888916,
+ "step": 6141
+ },
+ {
+ "epoch": 85.30742358078602,
+ "grad_norm": 0.018116485327482224,
+ "learning_rate": 0.0006,
+ "loss": 4.122644424438477,
+ "step": 6142
+ },
+ {
+ "epoch": 85.32139737991267,
+ "grad_norm": 0.015949107706546783,
+ "learning_rate": 0.0006,
+ "loss": 4.135458946228027,
+ "step": 6143
+ },
+ {
+ "epoch": 85.3353711790393,
+ "grad_norm": 0.014783683232963085,
+ "learning_rate": 0.0006,
+ "loss": 4.118994235992432,
+ "step": 6144
+ },
+ {
+ "epoch": 85.34934497816595,
+ "grad_norm": 0.015406692400574684,
+ "learning_rate": 0.0006,
+ "loss": 4.157268524169922,
+ "step": 6145
+ },
+ {
+ "epoch": 85.36331877729258,
+ "grad_norm": 0.013893205672502518,
+ "learning_rate": 0.0006,
+ "loss": 4.056166648864746,
+ "step": 6146
+ },
+ {
+ "epoch": 85.37729257641921,
+ "grad_norm": 0.014380923472344875,
+ "learning_rate": 0.0006,
+ "loss": 4.206905364990234,
+ "step": 6147
+ },
+ {
+ "epoch": 85.39126637554585,
+ "grad_norm": 0.015669258311390877,
+ "learning_rate": 0.0006,
+ "loss": 4.213881492614746,
+ "step": 6148
+ },
+ {
+ "epoch": 85.40524017467249,
+ "grad_norm": 0.01553428452461958,
+ "learning_rate": 0.0006,
+ "loss": 4.1814374923706055,
+ "step": 6149
+ },
+ {
+ "epoch": 85.41921397379913,
+ "grad_norm": 0.01697700284421444,
+ "learning_rate": 0.0006,
+ "loss": 4.083939552307129,
+ "step": 6150
+ },
+ {
+ "epoch": 85.43318777292576,
+ "grad_norm": 0.016639070585370064,
+ "learning_rate": 0.0006,
+ "loss": 4.279311180114746,
+ "step": 6151
+ },
+ {
+ "epoch": 85.44716157205241,
+ "grad_norm": 0.016285281628370285,
+ "learning_rate": 0.0006,
+ "loss": 4.244845390319824,
+ "step": 6152
+ },
+ {
+ "epoch": 85.46113537117904,
+ "grad_norm": 0.016071317717432976,
+ "learning_rate": 0.0006,
+ "loss": 4.089814186096191,
+ "step": 6153
+ },
+ {
+ "epoch": 85.47510917030567,
+ "grad_norm": 0.01640418916940689,
+ "learning_rate": 0.0006,
+ "loss": 4.168088912963867,
+ "step": 6154
+ },
+ {
+ "epoch": 85.48908296943232,
+ "grad_norm": 0.014781414531171322,
+ "learning_rate": 0.0006,
+ "loss": 4.196092128753662,
+ "step": 6155
+ },
+ {
+ "epoch": 85.50305676855895,
+ "grad_norm": 0.013393213972449303,
+ "learning_rate": 0.0006,
+ "loss": 4.193429946899414,
+ "step": 6156
+ },
+ {
+ "epoch": 85.5170305676856,
+ "grad_norm": 0.01467406190931797,
+ "learning_rate": 0.0006,
+ "loss": 4.216714859008789,
+ "step": 6157
+ },
+ {
+ "epoch": 85.53100436681223,
+ "grad_norm": 0.014950765296816826,
+ "learning_rate": 0.0006,
+ "loss": 4.093854904174805,
+ "step": 6158
+ },
+ {
+ "epoch": 85.54497816593886,
+ "grad_norm": 0.014655301347374916,
+ "learning_rate": 0.0006,
+ "loss": 4.09483528137207,
+ "step": 6159
+ },
+ {
+ "epoch": 85.5589519650655,
+ "grad_norm": 0.01425520796328783,
+ "learning_rate": 0.0006,
+ "loss": 4.090396881103516,
+ "step": 6160
+ },
+ {
+ "epoch": 85.57292576419214,
+ "grad_norm": 0.01403278298676014,
+ "learning_rate": 0.0006,
+ "loss": 4.088724136352539,
+ "step": 6161
+ },
+ {
+ "epoch": 85.58689956331878,
+ "grad_norm": 0.01364967692643404,
+ "learning_rate": 0.0006,
+ "loss": 4.095558166503906,
+ "step": 6162
+ },
+ {
+ "epoch": 85.60087336244541,
+ "grad_norm": 0.013085267506539822,
+ "learning_rate": 0.0006,
+ "loss": 4.163962364196777,
+ "step": 6163
+ },
+ {
+ "epoch": 85.61484716157206,
+ "grad_norm": 0.0136906448751688,
+ "learning_rate": 0.0006,
+ "loss": 4.197134017944336,
+ "step": 6164
+ },
+ {
+ "epoch": 85.62882096069869,
+ "grad_norm": 0.013472451828420162,
+ "learning_rate": 0.0006,
+ "loss": 4.1578369140625,
+ "step": 6165
+ },
+ {
+ "epoch": 85.64279475982532,
+ "grad_norm": 0.012329773046076298,
+ "learning_rate": 0.0006,
+ "loss": 4.231907844543457,
+ "step": 6166
+ },
+ {
+ "epoch": 85.65676855895197,
+ "grad_norm": 0.011977200396358967,
+ "learning_rate": 0.0006,
+ "loss": 4.1552958488464355,
+ "step": 6167
+ },
+ {
+ "epoch": 85.6707423580786,
+ "grad_norm": 0.013049956411123276,
+ "learning_rate": 0.0006,
+ "loss": 4.17852783203125,
+ "step": 6168
+ },
+ {
+ "epoch": 85.68471615720524,
+ "grad_norm": 0.011440741829574108,
+ "learning_rate": 0.0006,
+ "loss": 4.144646644592285,
+ "step": 6169
+ },
+ {
+ "epoch": 85.69868995633188,
+ "grad_norm": 0.012472936883568764,
+ "learning_rate": 0.0006,
+ "loss": 4.2306413650512695,
+ "step": 6170
+ },
+ {
+ "epoch": 85.71266375545852,
+ "grad_norm": 0.013015356846153736,
+ "learning_rate": 0.0006,
+ "loss": 4.193439960479736,
+ "step": 6171
+ },
+ {
+ "epoch": 85.72663755458515,
+ "grad_norm": 0.013832319527864456,
+ "learning_rate": 0.0006,
+ "loss": 4.18073034286499,
+ "step": 6172
+ },
+ {
+ "epoch": 85.74061135371178,
+ "grad_norm": 0.014613982290029526,
+ "learning_rate": 0.0006,
+ "loss": 4.223531246185303,
+ "step": 6173
+ },
+ {
+ "epoch": 85.75458515283843,
+ "grad_norm": 0.014209083281457424,
+ "learning_rate": 0.0006,
+ "loss": 4.142666816711426,
+ "step": 6174
+ },
+ {
+ "epoch": 85.76855895196506,
+ "grad_norm": 0.013396669179201126,
+ "learning_rate": 0.0006,
+ "loss": 4.185786247253418,
+ "step": 6175
+ },
+ {
+ "epoch": 85.78253275109171,
+ "grad_norm": 0.01364929135888815,
+ "learning_rate": 0.0006,
+ "loss": 4.125226974487305,
+ "step": 6176
+ },
+ {
+ "epoch": 85.79650655021834,
+ "grad_norm": 0.012561003677546978,
+ "learning_rate": 0.0006,
+ "loss": 4.084519386291504,
+ "step": 6177
+ },
+ {
+ "epoch": 85.81048034934499,
+ "grad_norm": 0.013508481904864311,
+ "learning_rate": 0.0006,
+ "loss": 4.193173885345459,
+ "step": 6178
+ },
+ {
+ "epoch": 85.82445414847162,
+ "grad_norm": 0.014154613018035889,
+ "learning_rate": 0.0006,
+ "loss": 4.257272720336914,
+ "step": 6179
+ },
+ {
+ "epoch": 85.83842794759825,
+ "grad_norm": 0.012478531338274479,
+ "learning_rate": 0.0006,
+ "loss": 4.161380767822266,
+ "step": 6180
+ },
+ {
+ "epoch": 85.8524017467249,
+ "grad_norm": 0.013468451797962189,
+ "learning_rate": 0.0006,
+ "loss": 4.121395111083984,
+ "step": 6181
+ },
+ {
+ "epoch": 85.86637554585153,
+ "grad_norm": 0.015155194327235222,
+ "learning_rate": 0.0006,
+ "loss": 4.242815971374512,
+ "step": 6182
+ },
+ {
+ "epoch": 85.88034934497817,
+ "grad_norm": 0.016223272308707237,
+ "learning_rate": 0.0006,
+ "loss": 4.176027297973633,
+ "step": 6183
+ },
+ {
+ "epoch": 85.8943231441048,
+ "grad_norm": 0.013741295784711838,
+ "learning_rate": 0.0006,
+ "loss": 4.131257057189941,
+ "step": 6184
+ },
+ {
+ "epoch": 85.90829694323143,
+ "grad_norm": 0.014982067979872227,
+ "learning_rate": 0.0006,
+ "loss": 4.237896919250488,
+ "step": 6185
+ },
+ {
+ "epoch": 85.92227074235808,
+ "grad_norm": 0.018197322264313698,
+ "learning_rate": 0.0006,
+ "loss": 4.154962539672852,
+ "step": 6186
+ },
+ {
+ "epoch": 85.93624454148471,
+ "grad_norm": 0.01971200481057167,
+ "learning_rate": 0.0006,
+ "loss": 4.16819429397583,
+ "step": 6187
+ },
+ {
+ "epoch": 85.95021834061136,
+ "grad_norm": 0.018008720129728317,
+ "learning_rate": 0.0006,
+ "loss": 4.146164417266846,
+ "step": 6188
+ },
+ {
+ "epoch": 85.96419213973799,
+ "grad_norm": 0.015859264880418777,
+ "learning_rate": 0.0006,
+ "loss": 4.208099365234375,
+ "step": 6189
+ },
+ {
+ "epoch": 85.97816593886463,
+ "grad_norm": 0.013846252113580704,
+ "learning_rate": 0.0006,
+ "loss": 4.1881184577941895,
+ "step": 6190
+ },
+ {
+ "epoch": 85.99213973799127,
+ "grad_norm": 0.018641939386725426,
+ "learning_rate": 0.0006,
+ "loss": 4.127632141113281,
+ "step": 6191
+ },
+ {
+ "epoch": 86.0,
+ "grad_norm": 0.02385227382183075,
+ "learning_rate": 0.0006,
+ "loss": 4.281900405883789,
+ "step": 6192
+ },
+ {
+ "epoch": 86.0,
+ "eval_loss": 4.625458717346191,
+ "eval_runtime": 57.0634,
+ "eval_samples_per_second": 42.795,
+ "eval_steps_per_second": 1.349,
+ "step": 6192
+ },
+ {
+ "epoch": 86.01397379912663,
+ "grad_norm": 0.019032124429941177,
+ "learning_rate": 0.0006,
+ "loss": 4.2163190841674805,
+ "step": 6193
+ },
+ {
+ "epoch": 86.02794759825328,
+ "grad_norm": 0.018049925565719604,
+ "learning_rate": 0.0006,
+ "loss": 4.1534037590026855,
+ "step": 6194
+ },
+ {
+ "epoch": 86.04192139737991,
+ "grad_norm": 0.019116360694169998,
+ "learning_rate": 0.0006,
+ "loss": 4.212411403656006,
+ "step": 6195
+ },
+ {
+ "epoch": 86.05589519650655,
+ "grad_norm": 0.018967336043715477,
+ "learning_rate": 0.0006,
+ "loss": 4.190918922424316,
+ "step": 6196
+ },
+ {
+ "epoch": 86.06986899563319,
+ "grad_norm": 0.01945611648261547,
+ "learning_rate": 0.0006,
+ "loss": 4.18459415435791,
+ "step": 6197
+ },
+ {
+ "epoch": 86.08384279475983,
+ "grad_norm": 0.018104076385498047,
+ "learning_rate": 0.0006,
+ "loss": 4.0942816734313965,
+ "step": 6198
+ },
+ {
+ "epoch": 86.09781659388646,
+ "grad_norm": 0.01815992221236229,
+ "learning_rate": 0.0006,
+ "loss": 4.011547565460205,
+ "step": 6199
+ },
+ {
+ "epoch": 86.1117903930131,
+ "grad_norm": 0.017574431374669075,
+ "learning_rate": 0.0006,
+ "loss": 4.051593780517578,
+ "step": 6200
+ },
+ {
+ "epoch": 86.12576419213974,
+ "grad_norm": 0.01516508124768734,
+ "learning_rate": 0.0006,
+ "loss": 4.239673614501953,
+ "step": 6201
+ },
+ {
+ "epoch": 86.13973799126637,
+ "grad_norm": 0.015689702704548836,
+ "learning_rate": 0.0006,
+ "loss": 4.1085734367370605,
+ "step": 6202
+ },
+ {
+ "epoch": 86.15371179039302,
+ "grad_norm": 0.01596900261938572,
+ "learning_rate": 0.0006,
+ "loss": 4.1509599685668945,
+ "step": 6203
+ },
+ {
+ "epoch": 86.16768558951965,
+ "grad_norm": 0.01550096832215786,
+ "learning_rate": 0.0006,
+ "loss": 4.160238265991211,
+ "step": 6204
+ },
+ {
+ "epoch": 86.18165938864628,
+ "grad_norm": 0.014601529575884342,
+ "learning_rate": 0.0006,
+ "loss": 4.029219150543213,
+ "step": 6205
+ },
+ {
+ "epoch": 86.19563318777293,
+ "grad_norm": 0.014006740413606167,
+ "learning_rate": 0.0006,
+ "loss": 4.117501258850098,
+ "step": 6206
+ },
+ {
+ "epoch": 86.20960698689956,
+ "grad_norm": 0.013191957026720047,
+ "learning_rate": 0.0006,
+ "loss": 4.252002716064453,
+ "step": 6207
+ },
+ {
+ "epoch": 86.2235807860262,
+ "grad_norm": 0.013031814247369766,
+ "learning_rate": 0.0006,
+ "loss": 4.209517478942871,
+ "step": 6208
+ },
+ {
+ "epoch": 86.23755458515284,
+ "grad_norm": 0.013605714775621891,
+ "learning_rate": 0.0006,
+ "loss": 4.176309585571289,
+ "step": 6209
+ },
+ {
+ "epoch": 86.25152838427948,
+ "grad_norm": 0.012753087095916271,
+ "learning_rate": 0.0006,
+ "loss": 4.097990036010742,
+ "step": 6210
+ },
+ {
+ "epoch": 86.26550218340611,
+ "grad_norm": 0.012940628454089165,
+ "learning_rate": 0.0006,
+ "loss": 4.085589408874512,
+ "step": 6211
+ },
+ {
+ "epoch": 86.27947598253274,
+ "grad_norm": 0.014286619611084461,
+ "learning_rate": 0.0006,
+ "loss": 4.199261665344238,
+ "step": 6212
+ },
+ {
+ "epoch": 86.29344978165939,
+ "grad_norm": 0.014643525704741478,
+ "learning_rate": 0.0006,
+ "loss": 4.23718786239624,
+ "step": 6213
+ },
+ {
+ "epoch": 86.30742358078602,
+ "grad_norm": 0.015182147733867168,
+ "learning_rate": 0.0006,
+ "loss": 4.219855308532715,
+ "step": 6214
+ },
+ {
+ "epoch": 86.32139737991267,
+ "grad_norm": 0.01530102826654911,
+ "learning_rate": 0.0006,
+ "loss": 4.190094947814941,
+ "step": 6215
+ },
+ {
+ "epoch": 86.3353711790393,
+ "grad_norm": 0.015492603182792664,
+ "learning_rate": 0.0006,
+ "loss": 4.153780460357666,
+ "step": 6216
+ },
+ {
+ "epoch": 86.34934497816595,
+ "grad_norm": 0.015339907258749008,
+ "learning_rate": 0.0006,
+ "loss": 4.182865619659424,
+ "step": 6217
+ },
+ {
+ "epoch": 86.36331877729258,
+ "grad_norm": 0.015576880425214767,
+ "learning_rate": 0.0006,
+ "loss": 4.196417331695557,
+ "step": 6218
+ },
+ {
+ "epoch": 86.37729257641921,
+ "grad_norm": 0.014805737882852554,
+ "learning_rate": 0.0006,
+ "loss": 4.265018463134766,
+ "step": 6219
+ },
+ {
+ "epoch": 86.39126637554585,
+ "grad_norm": 0.01597665622830391,
+ "learning_rate": 0.0006,
+ "loss": 4.2864837646484375,
+ "step": 6220
+ },
+ {
+ "epoch": 86.40524017467249,
+ "grad_norm": 0.01693643443286419,
+ "learning_rate": 0.0006,
+ "loss": 4.118192672729492,
+ "step": 6221
+ },
+ {
+ "epoch": 86.41921397379913,
+ "grad_norm": 0.016648951917886734,
+ "learning_rate": 0.0006,
+ "loss": 4.23712682723999,
+ "step": 6222
+ },
+ {
+ "epoch": 86.43318777292576,
+ "grad_norm": 0.014809766784310341,
+ "learning_rate": 0.0006,
+ "loss": 4.251391410827637,
+ "step": 6223
+ },
+ {
+ "epoch": 86.44716157205241,
+ "grad_norm": 0.014333373866975307,
+ "learning_rate": 0.0006,
+ "loss": 4.17243766784668,
+ "step": 6224
+ },
+ {
+ "epoch": 86.46113537117904,
+ "grad_norm": 0.013912711292505264,
+ "learning_rate": 0.0006,
+ "loss": 4.250836372375488,
+ "step": 6225
+ },
+ {
+ "epoch": 86.47510917030567,
+ "grad_norm": 0.014821341261267662,
+ "learning_rate": 0.0006,
+ "loss": 4.073995590209961,
+ "step": 6226
+ },
+ {
+ "epoch": 86.48908296943232,
+ "grad_norm": 0.016634922474622726,
+ "learning_rate": 0.0006,
+ "loss": 4.155867576599121,
+ "step": 6227
+ },
+ {
+ "epoch": 86.50305676855895,
+ "grad_norm": 0.01748695969581604,
+ "learning_rate": 0.0006,
+ "loss": 4.139344692230225,
+ "step": 6228
+ },
+ {
+ "epoch": 86.5170305676856,
+ "grad_norm": 0.01569363661110401,
+ "learning_rate": 0.0006,
+ "loss": 4.23923397064209,
+ "step": 6229
+ },
+ {
+ "epoch": 86.53100436681223,
+ "grad_norm": 0.014037121087312698,
+ "learning_rate": 0.0006,
+ "loss": 4.200334072113037,
+ "step": 6230
+ },
+ {
+ "epoch": 86.54497816593886,
+ "grad_norm": 0.01401279866695404,
+ "learning_rate": 0.0006,
+ "loss": 4.035246849060059,
+ "step": 6231
+ },
+ {
+ "epoch": 86.5589519650655,
+ "grad_norm": 0.014603731222450733,
+ "learning_rate": 0.0006,
+ "loss": 4.141385555267334,
+ "step": 6232
+ },
+ {
+ "epoch": 86.57292576419214,
+ "grad_norm": 0.015113704837858677,
+ "learning_rate": 0.0006,
+ "loss": 4.130824089050293,
+ "step": 6233
+ },
+ {
+ "epoch": 86.58689956331878,
+ "grad_norm": 0.01588696800172329,
+ "learning_rate": 0.0006,
+ "loss": 4.05495023727417,
+ "step": 6234
+ },
+ {
+ "epoch": 86.60087336244541,
+ "grad_norm": 0.01593640074133873,
+ "learning_rate": 0.0006,
+ "loss": 4.22935152053833,
+ "step": 6235
+ },
+ {
+ "epoch": 86.61484716157206,
+ "grad_norm": 0.014292445965111256,
+ "learning_rate": 0.0006,
+ "loss": 4.309839248657227,
+ "step": 6236
+ },
+ {
+ "epoch": 86.62882096069869,
+ "grad_norm": 0.015471610240638256,
+ "learning_rate": 0.0006,
+ "loss": 4.268278121948242,
+ "step": 6237
+ },
+ {
+ "epoch": 86.64279475982532,
+ "grad_norm": 0.016643544659018517,
+ "learning_rate": 0.0006,
+ "loss": 4.135109901428223,
+ "step": 6238
+ },
+ {
+ "epoch": 86.65676855895197,
+ "grad_norm": 0.014733029529452324,
+ "learning_rate": 0.0006,
+ "loss": 4.060197830200195,
+ "step": 6239
+ },
+ {
+ "epoch": 86.6707423580786,
+ "grad_norm": 0.014869216829538345,
+ "learning_rate": 0.0006,
+ "loss": 4.126917362213135,
+ "step": 6240
+ },
+ {
+ "epoch": 86.68471615720524,
+ "grad_norm": 0.01608203910291195,
+ "learning_rate": 0.0006,
+ "loss": 4.183716773986816,
+ "step": 6241
+ },
+ {
+ "epoch": 86.69868995633188,
+ "grad_norm": 0.016178175806999207,
+ "learning_rate": 0.0006,
+ "loss": 4.118875503540039,
+ "step": 6242
+ },
+ {
+ "epoch": 86.71266375545852,
+ "grad_norm": 0.014576597139239311,
+ "learning_rate": 0.0006,
+ "loss": 4.261541366577148,
+ "step": 6243
+ },
+ {
+ "epoch": 86.72663755458515,
+ "grad_norm": 0.014466420747339725,
+ "learning_rate": 0.0006,
+ "loss": 4.307284355163574,
+ "step": 6244
+ },
+ {
+ "epoch": 86.74061135371178,
+ "grad_norm": 0.014042644761502743,
+ "learning_rate": 0.0006,
+ "loss": 4.2729997634887695,
+ "step": 6245
+ },
+ {
+ "epoch": 86.75458515283843,
+ "grad_norm": 0.014589524827897549,
+ "learning_rate": 0.0006,
+ "loss": 4.072981834411621,
+ "step": 6246
+ },
+ {
+ "epoch": 86.76855895196506,
+ "grad_norm": 0.018765483051538467,
+ "learning_rate": 0.0006,
+ "loss": 4.264769554138184,
+ "step": 6247
+ },
+ {
+ "epoch": 86.78253275109171,
+ "grad_norm": 0.019397731870412827,
+ "learning_rate": 0.0006,
+ "loss": 4.183379173278809,
+ "step": 6248
+ },
+ {
+ "epoch": 86.79650655021834,
+ "grad_norm": 0.017525402829051018,
+ "learning_rate": 0.0006,
+ "loss": 4.1494245529174805,
+ "step": 6249
+ },
+ {
+ "epoch": 86.81048034934499,
+ "grad_norm": 0.016310200095176697,
+ "learning_rate": 0.0006,
+ "loss": 4.149104118347168,
+ "step": 6250
+ },
+ {
+ "epoch": 86.82445414847162,
+ "grad_norm": 0.01506423857063055,
+ "learning_rate": 0.0006,
+ "loss": 4.22091817855835,
+ "step": 6251
+ },
+ {
+ "epoch": 86.83842794759825,
+ "grad_norm": 0.014200284145772457,
+ "learning_rate": 0.0006,
+ "loss": 4.177569389343262,
+ "step": 6252
+ },
+ {
+ "epoch": 86.8524017467249,
+ "grad_norm": 0.015016552060842514,
+ "learning_rate": 0.0006,
+ "loss": 4.089859962463379,
+ "step": 6253
+ },
+ {
+ "epoch": 86.86637554585153,
+ "grad_norm": 0.016054971143603325,
+ "learning_rate": 0.0006,
+ "loss": 4.157166004180908,
+ "step": 6254
+ },
+ {
+ "epoch": 86.88034934497817,
+ "grad_norm": 0.014388521201908588,
+ "learning_rate": 0.0006,
+ "loss": 4.1437578201293945,
+ "step": 6255
+ },
+ {
+ "epoch": 86.8943231441048,
+ "grad_norm": 0.015705369412899017,
+ "learning_rate": 0.0006,
+ "loss": 4.045689582824707,
+ "step": 6256
+ },
+ {
+ "epoch": 86.90829694323143,
+ "grad_norm": 0.014192395843565464,
+ "learning_rate": 0.0006,
+ "loss": 4.098793029785156,
+ "step": 6257
+ },
+ {
+ "epoch": 86.92227074235808,
+ "grad_norm": 0.013751034624874592,
+ "learning_rate": 0.0006,
+ "loss": 3.989920139312744,
+ "step": 6258
+ },
+ {
+ "epoch": 86.93624454148471,
+ "grad_norm": 0.01582193188369274,
+ "learning_rate": 0.0006,
+ "loss": 4.283449172973633,
+ "step": 6259
+ },
+ {
+ "epoch": 86.95021834061136,
+ "grad_norm": 0.01476013008505106,
+ "learning_rate": 0.0006,
+ "loss": 4.221401691436768,
+ "step": 6260
+ },
+ {
+ "epoch": 86.96419213973799,
+ "grad_norm": 0.01346337515860796,
+ "learning_rate": 0.0006,
+ "loss": 4.05896520614624,
+ "step": 6261
+ },
+ {
+ "epoch": 86.97816593886463,
+ "grad_norm": 0.012899307534098625,
+ "learning_rate": 0.0006,
+ "loss": 4.159560203552246,
+ "step": 6262
+ },
+ {
+ "epoch": 86.99213973799127,
+ "grad_norm": 0.013150574639439583,
+ "learning_rate": 0.0006,
+ "loss": 4.217330455780029,
+ "step": 6263
+ },
+ {
+ "epoch": 87.0,
+ "grad_norm": 0.014940370805561543,
+ "learning_rate": 0.0006,
+ "loss": 4.1611833572387695,
+ "step": 6264
+ },
+ {
+ "epoch": 87.0,
+ "eval_loss": 4.666043758392334,
+ "eval_runtime": 56.5978,
+ "eval_samples_per_second": 43.147,
+ "eval_steps_per_second": 1.36,
+ "step": 6264
+ },
+ {
+ "epoch": 87.01397379912663,
+ "grad_norm": 0.013721180148422718,
+ "learning_rate": 0.0006,
+ "loss": 4.142454147338867,
+ "step": 6265
+ },
+ {
+ "epoch": 87.02794759825328,
+ "grad_norm": 0.014805255457758904,
+ "learning_rate": 0.0006,
+ "loss": 4.121677875518799,
+ "step": 6266
+ },
+ {
+ "epoch": 87.04192139737991,
+ "grad_norm": 0.015576514415442944,
+ "learning_rate": 0.0006,
+ "loss": 4.1188435554504395,
+ "step": 6267
+ },
+ {
+ "epoch": 87.05589519650655,
+ "grad_norm": 0.01605502888560295,
+ "learning_rate": 0.0006,
+ "loss": 4.255428314208984,
+ "step": 6268
+ },
+ {
+ "epoch": 87.06986899563319,
+ "grad_norm": 0.014933343045413494,
+ "learning_rate": 0.0006,
+ "loss": 4.122596740722656,
+ "step": 6269
+ },
+ {
+ "epoch": 87.08384279475983,
+ "grad_norm": 0.014630894176661968,
+ "learning_rate": 0.0006,
+ "loss": 4.148405075073242,
+ "step": 6270
+ },
+ {
+ "epoch": 87.09781659388646,
+ "grad_norm": 0.014211163856089115,
+ "learning_rate": 0.0006,
+ "loss": 4.058030128479004,
+ "step": 6271
+ },
+ {
+ "epoch": 87.1117903930131,
+ "grad_norm": 0.014060168527066708,
+ "learning_rate": 0.0006,
+ "loss": 4.146586894989014,
+ "step": 6272
+ },
+ {
+ "epoch": 87.12576419213974,
+ "grad_norm": 0.014740100130438805,
+ "learning_rate": 0.0006,
+ "loss": 4.077739715576172,
+ "step": 6273
+ },
+ {
+ "epoch": 87.13973799126637,
+ "grad_norm": 0.013970664702355862,
+ "learning_rate": 0.0006,
+ "loss": 4.038252830505371,
+ "step": 6274
+ },
+ {
+ "epoch": 87.15371179039302,
+ "grad_norm": 0.014399951323866844,
+ "learning_rate": 0.0006,
+ "loss": 4.143700122833252,
+ "step": 6275
+ },
+ {
+ "epoch": 87.16768558951965,
+ "grad_norm": 0.01496410183608532,
+ "learning_rate": 0.0006,
+ "loss": 4.105609893798828,
+ "step": 6276
+ },
+ {
+ "epoch": 87.18165938864628,
+ "grad_norm": 0.014287355355918407,
+ "learning_rate": 0.0006,
+ "loss": 4.173645973205566,
+ "step": 6277
+ },
+ {
+ "epoch": 87.19563318777293,
+ "grad_norm": 0.0141292130574584,
+ "learning_rate": 0.0006,
+ "loss": 4.068511486053467,
+ "step": 6278
+ },
+ {
+ "epoch": 87.20960698689956,
+ "grad_norm": 0.01393523346632719,
+ "learning_rate": 0.0006,
+ "loss": 4.060687065124512,
+ "step": 6279
+ },
+ {
+ "epoch": 87.2235807860262,
+ "grad_norm": 0.01431021187454462,
+ "learning_rate": 0.0006,
+ "loss": 4.219209671020508,
+ "step": 6280
+ },
+ {
+ "epoch": 87.23755458515284,
+ "grad_norm": 0.014856027439236641,
+ "learning_rate": 0.0006,
+ "loss": 4.146749973297119,
+ "step": 6281
+ },
+ {
+ "epoch": 87.25152838427948,
+ "grad_norm": 0.013677539303898811,
+ "learning_rate": 0.0006,
+ "loss": 3.9859797954559326,
+ "step": 6282
+ },
+ {
+ "epoch": 87.26550218340611,
+ "grad_norm": 0.014078967273235321,
+ "learning_rate": 0.0006,
+ "loss": 4.129668235778809,
+ "step": 6283
+ },
+ {
+ "epoch": 87.27947598253274,
+ "grad_norm": 0.015564394183456898,
+ "learning_rate": 0.0006,
+ "loss": 4.0007829666137695,
+ "step": 6284
+ },
+ {
+ "epoch": 87.29344978165939,
+ "grad_norm": 0.015077749267220497,
+ "learning_rate": 0.0006,
+ "loss": 4.112264156341553,
+ "step": 6285
+ },
+ {
+ "epoch": 87.30742358078602,
+ "grad_norm": 0.015341218560934067,
+ "learning_rate": 0.0006,
+ "loss": 4.112249374389648,
+ "step": 6286
+ },
+ {
+ "epoch": 87.32139737991267,
+ "grad_norm": 0.016399366781115532,
+ "learning_rate": 0.0006,
+ "loss": 4.176196098327637,
+ "step": 6287
+ },
+ {
+ "epoch": 87.3353711790393,
+ "grad_norm": 0.01599763333797455,
+ "learning_rate": 0.0006,
+ "loss": 4.163441181182861,
+ "step": 6288
+ },
+ {
+ "epoch": 87.34934497816595,
+ "grad_norm": 0.015439776703715324,
+ "learning_rate": 0.0006,
+ "loss": 4.11902379989624,
+ "step": 6289
+ },
+ {
+ "epoch": 87.36331877729258,
+ "grad_norm": 0.015909943729639053,
+ "learning_rate": 0.0006,
+ "loss": 4.16621732711792,
+ "step": 6290
+ },
+ {
+ "epoch": 87.37729257641921,
+ "grad_norm": 0.016710078343749046,
+ "learning_rate": 0.0006,
+ "loss": 4.044960975646973,
+ "step": 6291
+ },
+ {
+ "epoch": 87.39126637554585,
+ "grad_norm": 0.015883933752775192,
+ "learning_rate": 0.0006,
+ "loss": 4.065184593200684,
+ "step": 6292
+ },
+ {
+ "epoch": 87.40524017467249,
+ "grad_norm": 0.013912747614085674,
+ "learning_rate": 0.0006,
+ "loss": 4.102450370788574,
+ "step": 6293
+ },
+ {
+ "epoch": 87.41921397379913,
+ "grad_norm": 0.013170420192182064,
+ "learning_rate": 0.0006,
+ "loss": 4.234364032745361,
+ "step": 6294
+ },
+ {
+ "epoch": 87.43318777292576,
+ "grad_norm": 0.014043302275240421,
+ "learning_rate": 0.0006,
+ "loss": 4.129565238952637,
+ "step": 6295
+ },
+ {
+ "epoch": 87.44716157205241,
+ "grad_norm": 0.015010139904916286,
+ "learning_rate": 0.0006,
+ "loss": 4.166678428649902,
+ "step": 6296
+ },
+ {
+ "epoch": 87.46113537117904,
+ "grad_norm": 0.014721784740686417,
+ "learning_rate": 0.0006,
+ "loss": 4.182618618011475,
+ "step": 6297
+ },
+ {
+ "epoch": 87.47510917030567,
+ "grad_norm": 0.015713468194007874,
+ "learning_rate": 0.0006,
+ "loss": 4.103855133056641,
+ "step": 6298
+ },
+ {
+ "epoch": 87.48908296943232,
+ "grad_norm": 0.017204247415065765,
+ "learning_rate": 0.0006,
+ "loss": 4.140937328338623,
+ "step": 6299
+ },
+ {
+ "epoch": 87.50305676855895,
+ "grad_norm": 0.016856668516993523,
+ "learning_rate": 0.0006,
+ "loss": 4.147859573364258,
+ "step": 6300
+ },
+ {
+ "epoch": 87.5170305676856,
+ "grad_norm": 0.017205018550157547,
+ "learning_rate": 0.0006,
+ "loss": 4.1667327880859375,
+ "step": 6301
+ },
+ {
+ "epoch": 87.53100436681223,
+ "grad_norm": 0.014821539632976055,
+ "learning_rate": 0.0006,
+ "loss": 4.089935302734375,
+ "step": 6302
+ },
+ {
+ "epoch": 87.54497816593886,
+ "grad_norm": 0.01446484588086605,
+ "learning_rate": 0.0006,
+ "loss": 4.243657112121582,
+ "step": 6303
+ },
+ {
+ "epoch": 87.5589519650655,
+ "grad_norm": 0.014374484308063984,
+ "learning_rate": 0.0006,
+ "loss": 4.14356803894043,
+ "step": 6304
+ },
+ {
+ "epoch": 87.57292576419214,
+ "grad_norm": 0.01420109998434782,
+ "learning_rate": 0.0006,
+ "loss": 4.1273651123046875,
+ "step": 6305
+ },
+ {
+ "epoch": 87.58689956331878,
+ "grad_norm": 0.014810388907790184,
+ "learning_rate": 0.0006,
+ "loss": 4.222290515899658,
+ "step": 6306
+ },
+ {
+ "epoch": 87.60087336244541,
+ "grad_norm": 0.015826983377337456,
+ "learning_rate": 0.0006,
+ "loss": 4.137996196746826,
+ "step": 6307
+ },
+ {
+ "epoch": 87.61484716157206,
+ "grad_norm": 0.014457454904913902,
+ "learning_rate": 0.0006,
+ "loss": 4.0780744552612305,
+ "step": 6308
+ },
+ {
+ "epoch": 87.62882096069869,
+ "grad_norm": 0.014458604156970978,
+ "learning_rate": 0.0006,
+ "loss": 4.1287922859191895,
+ "step": 6309
+ },
+ {
+ "epoch": 87.64279475982532,
+ "grad_norm": 0.015282071195542812,
+ "learning_rate": 0.0006,
+ "loss": 4.189508438110352,
+ "step": 6310
+ },
+ {
+ "epoch": 87.65676855895197,
+ "grad_norm": 0.017536642029881477,
+ "learning_rate": 0.0006,
+ "loss": 4.070173263549805,
+ "step": 6311
+ },
+ {
+ "epoch": 87.6707423580786,
+ "grad_norm": 0.017303580418229103,
+ "learning_rate": 0.0006,
+ "loss": 4.119133949279785,
+ "step": 6312
+ },
+ {
+ "epoch": 87.68471615720524,
+ "grad_norm": 0.016148580238223076,
+ "learning_rate": 0.0006,
+ "loss": 3.994354486465454,
+ "step": 6313
+ },
+ {
+ "epoch": 87.69868995633188,
+ "grad_norm": 0.01706942543387413,
+ "learning_rate": 0.0006,
+ "loss": 4.183113098144531,
+ "step": 6314
+ },
+ {
+ "epoch": 87.71266375545852,
+ "grad_norm": 0.016971010714769363,
+ "learning_rate": 0.0006,
+ "loss": 4.104086875915527,
+ "step": 6315
+ },
+ {
+ "epoch": 87.72663755458515,
+ "grad_norm": 0.015516847372055054,
+ "learning_rate": 0.0006,
+ "loss": 4.2409772872924805,
+ "step": 6316
+ },
+ {
+ "epoch": 87.74061135371178,
+ "grad_norm": 0.015352269634604454,
+ "learning_rate": 0.0006,
+ "loss": 4.026427268981934,
+ "step": 6317
+ },
+ {
+ "epoch": 87.75458515283843,
+ "grad_norm": 0.017702415585517883,
+ "learning_rate": 0.0006,
+ "loss": 4.177840232849121,
+ "step": 6318
+ },
+ {
+ "epoch": 87.76855895196506,
+ "grad_norm": 0.01629541628062725,
+ "learning_rate": 0.0006,
+ "loss": 4.229039192199707,
+ "step": 6319
+ },
+ {
+ "epoch": 87.78253275109171,
+ "grad_norm": 0.01503711100667715,
+ "learning_rate": 0.0006,
+ "loss": 4.214350700378418,
+ "step": 6320
+ },
+ {
+ "epoch": 87.79650655021834,
+ "grad_norm": 0.014913883060216904,
+ "learning_rate": 0.0006,
+ "loss": 4.168880939483643,
+ "step": 6321
+ },
+ {
+ "epoch": 87.81048034934499,
+ "grad_norm": 0.014513909816741943,
+ "learning_rate": 0.0006,
+ "loss": 4.162267684936523,
+ "step": 6322
+ },
+ {
+ "epoch": 87.82445414847162,
+ "grad_norm": 0.014635873027145863,
+ "learning_rate": 0.0006,
+ "loss": 4.220067977905273,
+ "step": 6323
+ },
+ {
+ "epoch": 87.83842794759825,
+ "grad_norm": 0.017102796584367752,
+ "learning_rate": 0.0006,
+ "loss": 4.1493377685546875,
+ "step": 6324
+ },
+ {
+ "epoch": 87.8524017467249,
+ "grad_norm": 0.015533884055912495,
+ "learning_rate": 0.0006,
+ "loss": 4.0646209716796875,
+ "step": 6325
+ },
+ {
+ "epoch": 87.86637554585153,
+ "grad_norm": 0.014617550186812878,
+ "learning_rate": 0.0006,
+ "loss": 4.1390533447265625,
+ "step": 6326
+ },
+ {
+ "epoch": 87.88034934497817,
+ "grad_norm": 0.014142967760562897,
+ "learning_rate": 0.0006,
+ "loss": 4.133934020996094,
+ "step": 6327
+ },
+ {
+ "epoch": 87.8943231441048,
+ "grad_norm": 0.013182433322072029,
+ "learning_rate": 0.0006,
+ "loss": 4.125327110290527,
+ "step": 6328
+ },
+ {
+ "epoch": 87.90829694323143,
+ "grad_norm": 0.013928229920566082,
+ "learning_rate": 0.0006,
+ "loss": 4.142777442932129,
+ "step": 6329
+ },
+ {
+ "epoch": 87.92227074235808,
+ "grad_norm": 0.015337013639509678,
+ "learning_rate": 0.0006,
+ "loss": 4.140210151672363,
+ "step": 6330
+ },
+ {
+ "epoch": 87.93624454148471,
+ "grad_norm": 0.014873536303639412,
+ "learning_rate": 0.0006,
+ "loss": 4.102636337280273,
+ "step": 6331
+ },
+ {
+ "epoch": 87.95021834061136,
+ "grad_norm": 0.01387431938201189,
+ "learning_rate": 0.0006,
+ "loss": 4.149059295654297,
+ "step": 6332
+ },
+ {
+ "epoch": 87.96419213973799,
+ "grad_norm": 0.013234943151473999,
+ "learning_rate": 0.0006,
+ "loss": 4.1564459800720215,
+ "step": 6333
+ },
+ {
+ "epoch": 87.97816593886463,
+ "grad_norm": 0.013781017623841763,
+ "learning_rate": 0.0006,
+ "loss": 4.189839839935303,
+ "step": 6334
+ },
+ {
+ "epoch": 87.99213973799127,
+ "grad_norm": 0.013733026571571827,
+ "learning_rate": 0.0006,
+ "loss": 4.29277229309082,
+ "step": 6335
+ },
+ {
+ "epoch": 88.0,
+ "grad_norm": 0.014752327464520931,
+ "learning_rate": 0.0006,
+ "loss": 4.163867950439453,
+ "step": 6336
+ },
+ {
+ "epoch": 88.0,
+ "eval_loss": 4.535566329956055,
+ "eval_runtime": 56.4677,
+ "eval_samples_per_second": 43.246,
+ "eval_steps_per_second": 1.364,
+ "step": 6336
+ },
+ {
+ "epoch": 88.01397379912663,
+ "grad_norm": 0.013958732597529888,
+ "learning_rate": 0.0006,
+ "loss": 4.200681209564209,
+ "step": 6337
+ },
+ {
+ "epoch": 88.02794759825328,
+ "grad_norm": 0.015283545479178429,
+ "learning_rate": 0.0006,
+ "loss": 4.1736016273498535,
+ "step": 6338
+ },
+ {
+ "epoch": 88.04192139737991,
+ "grad_norm": 0.016192059963941574,
+ "learning_rate": 0.0006,
+ "loss": 4.040589809417725,
+ "step": 6339
+ },
+ {
+ "epoch": 88.05589519650655,
+ "grad_norm": 0.018096154555678368,
+ "learning_rate": 0.0006,
+ "loss": 4.146696090698242,
+ "step": 6340
+ },
+ {
+ "epoch": 88.06986899563319,
+ "grad_norm": 0.01797664538025856,
+ "learning_rate": 0.0006,
+ "loss": 4.007248878479004,
+ "step": 6341
+ },
+ {
+ "epoch": 88.08384279475983,
+ "grad_norm": 0.016275612637400627,
+ "learning_rate": 0.0006,
+ "loss": 4.193206787109375,
+ "step": 6342
+ },
+ {
+ "epoch": 88.09781659388646,
+ "grad_norm": 0.014210258610546589,
+ "learning_rate": 0.0006,
+ "loss": 4.064548492431641,
+ "step": 6343
+ },
+ {
+ "epoch": 88.1117903930131,
+ "grad_norm": 0.016085151582956314,
+ "learning_rate": 0.0006,
+ "loss": 4.003668785095215,
+ "step": 6344
+ },
+ {
+ "epoch": 88.12576419213974,
+ "grad_norm": 0.01930483989417553,
+ "learning_rate": 0.0006,
+ "loss": 4.236578941345215,
+ "step": 6345
+ },
+ {
+ "epoch": 88.13973799126637,
+ "grad_norm": 0.017140906304121017,
+ "learning_rate": 0.0006,
+ "loss": 3.9593005180358887,
+ "step": 6346
+ },
+ {
+ "epoch": 88.15371179039302,
+ "grad_norm": 0.01521560549736023,
+ "learning_rate": 0.0006,
+ "loss": 4.137644290924072,
+ "step": 6347
+ },
+ {
+ "epoch": 88.16768558951965,
+ "grad_norm": 0.015788141638040543,
+ "learning_rate": 0.0006,
+ "loss": 4.124845027923584,
+ "step": 6348
+ },
+ {
+ "epoch": 88.18165938864628,
+ "grad_norm": 0.01784597896039486,
+ "learning_rate": 0.0006,
+ "loss": 4.139750957489014,
+ "step": 6349
+ },
+ {
+ "epoch": 88.19563318777293,
+ "grad_norm": 0.016266558319330215,
+ "learning_rate": 0.0006,
+ "loss": 4.074851989746094,
+ "step": 6350
+ },
+ {
+ "epoch": 88.20960698689956,
+ "grad_norm": 0.014024483971297741,
+ "learning_rate": 0.0006,
+ "loss": 4.164689064025879,
+ "step": 6351
+ },
+ {
+ "epoch": 88.2235807860262,
+ "grad_norm": 0.015694420784711838,
+ "learning_rate": 0.0006,
+ "loss": 4.219979286193848,
+ "step": 6352
+ },
+ {
+ "epoch": 88.23755458515284,
+ "grad_norm": 0.014897859655320644,
+ "learning_rate": 0.0006,
+ "loss": 4.050738334655762,
+ "step": 6353
+ },
+ {
+ "epoch": 88.25152838427948,
+ "grad_norm": 0.014827374368906021,
+ "learning_rate": 0.0006,
+ "loss": 4.137135028839111,
+ "step": 6354
+ },
+ {
+ "epoch": 88.26550218340611,
+ "grad_norm": 0.01454879529774189,
+ "learning_rate": 0.0006,
+ "loss": 3.9980897903442383,
+ "step": 6355
+ },
+ {
+ "epoch": 88.27947598253274,
+ "grad_norm": 0.015623947605490685,
+ "learning_rate": 0.0006,
+ "loss": 4.226551055908203,
+ "step": 6356
+ },
+ {
+ "epoch": 88.29344978165939,
+ "grad_norm": 0.015527035109698772,
+ "learning_rate": 0.0006,
+ "loss": 4.11053466796875,
+ "step": 6357
+ },
+ {
+ "epoch": 88.30742358078602,
+ "grad_norm": 0.015242574736475945,
+ "learning_rate": 0.0006,
+ "loss": 4.105966567993164,
+ "step": 6358
+ },
+ {
+ "epoch": 88.32139737991267,
+ "grad_norm": 0.01375538855791092,
+ "learning_rate": 0.0006,
+ "loss": 4.148453712463379,
+ "step": 6359
+ },
+ {
+ "epoch": 88.3353711790393,
+ "grad_norm": 0.01595226116478443,
+ "learning_rate": 0.0006,
+ "loss": 4.095836162567139,
+ "step": 6360
+ },
+ {
+ "epoch": 88.34934497816595,
+ "grad_norm": 0.015252740122377872,
+ "learning_rate": 0.0006,
+ "loss": 4.136335849761963,
+ "step": 6361
+ },
+ {
+ "epoch": 88.36331877729258,
+ "grad_norm": 0.014405866153538227,
+ "learning_rate": 0.0006,
+ "loss": 4.084691047668457,
+ "step": 6362
+ },
+ {
+ "epoch": 88.37729257641921,
+ "grad_norm": 0.014755490235984325,
+ "learning_rate": 0.0006,
+ "loss": 4.095475673675537,
+ "step": 6363
+ },
+ {
+ "epoch": 88.39126637554585,
+ "grad_norm": 0.015339708887040615,
+ "learning_rate": 0.0006,
+ "loss": 4.163136959075928,
+ "step": 6364
+ },
+ {
+ "epoch": 88.40524017467249,
+ "grad_norm": 0.014667164534330368,
+ "learning_rate": 0.0006,
+ "loss": 4.062265396118164,
+ "step": 6365
+ },
+ {
+ "epoch": 88.41921397379913,
+ "grad_norm": 0.014588423073291779,
+ "learning_rate": 0.0006,
+ "loss": 4.13359260559082,
+ "step": 6366
+ },
+ {
+ "epoch": 88.43318777292576,
+ "grad_norm": 0.014005377888679504,
+ "learning_rate": 0.0006,
+ "loss": 4.219711780548096,
+ "step": 6367
+ },
+ {
+ "epoch": 88.44716157205241,
+ "grad_norm": 0.012769930064678192,
+ "learning_rate": 0.0006,
+ "loss": 4.114658355712891,
+ "step": 6368
+ },
+ {
+ "epoch": 88.46113537117904,
+ "grad_norm": 0.013632693327963352,
+ "learning_rate": 0.0006,
+ "loss": 4.0675225257873535,
+ "step": 6369
+ },
+ {
+ "epoch": 88.47510917030567,
+ "grad_norm": 0.013829804956912994,
+ "learning_rate": 0.0006,
+ "loss": 4.1125969886779785,
+ "step": 6370
+ },
+ {
+ "epoch": 88.48908296943232,
+ "grad_norm": 0.013882109895348549,
+ "learning_rate": 0.0006,
+ "loss": 4.1627092361450195,
+ "step": 6371
+ },
+ {
+ "epoch": 88.50305676855895,
+ "grad_norm": 0.014023828320205212,
+ "learning_rate": 0.0006,
+ "loss": 4.01038932800293,
+ "step": 6372
+ },
+ {
+ "epoch": 88.5170305676856,
+ "grad_norm": 0.014840706251561642,
+ "learning_rate": 0.0006,
+ "loss": 4.117877960205078,
+ "step": 6373
+ },
+ {
+ "epoch": 88.53100436681223,
+ "grad_norm": 0.014165930449962616,
+ "learning_rate": 0.0006,
+ "loss": 4.1864237785339355,
+ "step": 6374
+ },
+ {
+ "epoch": 88.54497816593886,
+ "grad_norm": 0.014390201307833195,
+ "learning_rate": 0.0006,
+ "loss": 4.1044602394104,
+ "step": 6375
+ },
+ {
+ "epoch": 88.5589519650655,
+ "grad_norm": 0.015037395060062408,
+ "learning_rate": 0.0006,
+ "loss": 4.101107120513916,
+ "step": 6376
+ },
+ {
+ "epoch": 88.57292576419214,
+ "grad_norm": 0.01671452447772026,
+ "learning_rate": 0.0006,
+ "loss": 4.170581817626953,
+ "step": 6377
+ },
+ {
+ "epoch": 88.58689956331878,
+ "grad_norm": 0.01993076130747795,
+ "learning_rate": 0.0006,
+ "loss": 4.261412620544434,
+ "step": 6378
+ },
+ {
+ "epoch": 88.60087336244541,
+ "grad_norm": 0.023721016943454742,
+ "learning_rate": 0.0006,
+ "loss": 4.199363708496094,
+ "step": 6379
+ },
+ {
+ "epoch": 88.61484716157206,
+ "grad_norm": 0.023726163432002068,
+ "learning_rate": 0.0006,
+ "loss": 4.061285018920898,
+ "step": 6380
+ },
+ {
+ "epoch": 88.62882096069869,
+ "grad_norm": 0.018352877348661423,
+ "learning_rate": 0.0006,
+ "loss": 4.0721282958984375,
+ "step": 6381
+ },
+ {
+ "epoch": 88.64279475982532,
+ "grad_norm": 0.01679830253124237,
+ "learning_rate": 0.0006,
+ "loss": 4.1717939376831055,
+ "step": 6382
+ },
+ {
+ "epoch": 88.65676855895197,
+ "grad_norm": 0.017193637788295746,
+ "learning_rate": 0.0006,
+ "loss": 4.230783939361572,
+ "step": 6383
+ },
+ {
+ "epoch": 88.6707423580786,
+ "grad_norm": 0.01581464521586895,
+ "learning_rate": 0.0006,
+ "loss": 4.029906272888184,
+ "step": 6384
+ },
+ {
+ "epoch": 88.68471615720524,
+ "grad_norm": 0.014049704186618328,
+ "learning_rate": 0.0006,
+ "loss": 4.192776679992676,
+ "step": 6385
+ },
+ {
+ "epoch": 88.69868995633188,
+ "grad_norm": 0.017214450985193253,
+ "learning_rate": 0.0006,
+ "loss": 4.248916149139404,
+ "step": 6386
+ },
+ {
+ "epoch": 88.71266375545852,
+ "grad_norm": 0.0173103716224432,
+ "learning_rate": 0.0006,
+ "loss": 4.124347686767578,
+ "step": 6387
+ },
+ {
+ "epoch": 88.72663755458515,
+ "grad_norm": 0.015745092183351517,
+ "learning_rate": 0.0006,
+ "loss": 4.192878246307373,
+ "step": 6388
+ },
+ {
+ "epoch": 88.74061135371178,
+ "grad_norm": 0.013750772923231125,
+ "learning_rate": 0.0006,
+ "loss": 4.1564483642578125,
+ "step": 6389
+ },
+ {
+ "epoch": 88.75458515283843,
+ "grad_norm": 0.014969858340919018,
+ "learning_rate": 0.0006,
+ "loss": 4.0980424880981445,
+ "step": 6390
+ },
+ {
+ "epoch": 88.76855895196506,
+ "grad_norm": 0.01287173479795456,
+ "learning_rate": 0.0006,
+ "loss": 4.145712852478027,
+ "step": 6391
+ },
+ {
+ "epoch": 88.78253275109171,
+ "grad_norm": 0.012716300785541534,
+ "learning_rate": 0.0006,
+ "loss": 4.214226722717285,
+ "step": 6392
+ },
+ {
+ "epoch": 88.79650655021834,
+ "grad_norm": 0.013855204917490482,
+ "learning_rate": 0.0006,
+ "loss": 4.1667256355285645,
+ "step": 6393
+ },
+ {
+ "epoch": 88.81048034934499,
+ "grad_norm": 0.013650864362716675,
+ "learning_rate": 0.0006,
+ "loss": 4.161942481994629,
+ "step": 6394
+ },
+ {
+ "epoch": 88.82445414847162,
+ "grad_norm": 0.013780126348137856,
+ "learning_rate": 0.0006,
+ "loss": 4.087825775146484,
+ "step": 6395
+ },
+ {
+ "epoch": 88.83842794759825,
+ "grad_norm": 0.014029948972165585,
+ "learning_rate": 0.0006,
+ "loss": 4.142813205718994,
+ "step": 6396
+ },
+ {
+ "epoch": 88.8524017467249,
+ "grad_norm": 0.01322692260146141,
+ "learning_rate": 0.0006,
+ "loss": 4.106776714324951,
+ "step": 6397
+ },
+ {
+ "epoch": 88.86637554585153,
+ "grad_norm": 0.013522760011255741,
+ "learning_rate": 0.0006,
+ "loss": 4.090996265411377,
+ "step": 6398
+ },
+ {
+ "epoch": 88.88034934497817,
+ "grad_norm": 0.014474684372544289,
+ "learning_rate": 0.0006,
+ "loss": 4.124624252319336,
+ "step": 6399
+ },
+ {
+ "epoch": 88.8943231441048,
+ "grad_norm": 0.014490422792732716,
+ "learning_rate": 0.0006,
+ "loss": 4.065431594848633,
+ "step": 6400
+ },
+ {
+ "epoch": 88.90829694323143,
+ "grad_norm": 0.015379710122942924,
+ "learning_rate": 0.0006,
+ "loss": 4.207939624786377,
+ "step": 6401
+ },
+ {
+ "epoch": 88.92227074235808,
+ "grad_norm": 0.015914618968963623,
+ "learning_rate": 0.0006,
+ "loss": 4.041152000427246,
+ "step": 6402
+ },
+ {
+ "epoch": 88.93624454148471,
+ "grad_norm": 0.01480048056691885,
+ "learning_rate": 0.0006,
+ "loss": 4.138754844665527,
+ "step": 6403
+ },
+ {
+ "epoch": 88.95021834061136,
+ "grad_norm": 0.013734851032495499,
+ "learning_rate": 0.0006,
+ "loss": 4.25323486328125,
+ "step": 6404
+ },
+ {
+ "epoch": 88.96419213973799,
+ "grad_norm": 0.01452311035245657,
+ "learning_rate": 0.0006,
+ "loss": 4.115228176116943,
+ "step": 6405
+ },
+ {
+ "epoch": 88.97816593886463,
+ "grad_norm": 0.016197681427001953,
+ "learning_rate": 0.0006,
+ "loss": 4.099531173706055,
+ "step": 6406
+ },
+ {
+ "epoch": 88.99213973799127,
+ "grad_norm": 0.017169523984193802,
+ "learning_rate": 0.0006,
+ "loss": 4.182981491088867,
+ "step": 6407
+ },
+ {
+ "epoch": 89.0,
+ "grad_norm": 0.017855646088719368,
+ "learning_rate": 0.0006,
+ "loss": 4.1470136642456055,
+ "step": 6408
+ },
+ {
+ "epoch": 89.0,
+ "eval_loss": 4.6035614013671875,
+ "eval_runtime": 57.0625,
+ "eval_samples_per_second": 42.795,
+ "eval_steps_per_second": 1.349,
+ "step": 6408
+ },
+ {
+ "epoch": 89.01397379912663,
+ "grad_norm": 0.014834101311862469,
+ "learning_rate": 0.0006,
+ "loss": 4.159243583679199,
+ "step": 6409
+ },
+ {
+ "epoch": 89.02794759825328,
+ "grad_norm": 0.015316772274672985,
+ "learning_rate": 0.0006,
+ "loss": 4.0709123611450195,
+ "step": 6410
+ },
+ {
+ "epoch": 89.04192139737991,
+ "grad_norm": 0.01576591096818447,
+ "learning_rate": 0.0006,
+ "loss": 4.0921196937561035,
+ "step": 6411
+ },
+ {
+ "epoch": 89.05589519650655,
+ "grad_norm": 0.015948045998811722,
+ "learning_rate": 0.0006,
+ "loss": 4.181877136230469,
+ "step": 6412
+ },
+ {
+ "epoch": 89.06986899563319,
+ "grad_norm": 0.016612380743026733,
+ "learning_rate": 0.0006,
+ "loss": 4.081249237060547,
+ "step": 6413
+ },
+ {
+ "epoch": 89.08384279475983,
+ "grad_norm": 0.017455685883760452,
+ "learning_rate": 0.0006,
+ "loss": 4.175990104675293,
+ "step": 6414
+ },
+ {
+ "epoch": 89.09781659388646,
+ "grad_norm": 0.015040521509945393,
+ "learning_rate": 0.0006,
+ "loss": 4.121378421783447,
+ "step": 6415
+ },
+ {
+ "epoch": 89.1117903930131,
+ "grad_norm": 0.014513209462165833,
+ "learning_rate": 0.0006,
+ "loss": 3.996107816696167,
+ "step": 6416
+ },
+ {
+ "epoch": 89.12576419213974,
+ "grad_norm": 0.015455638989806175,
+ "learning_rate": 0.0006,
+ "loss": 4.073675155639648,
+ "step": 6417
+ },
+ {
+ "epoch": 89.13973799126637,
+ "grad_norm": 0.015468129888176918,
+ "learning_rate": 0.0006,
+ "loss": 4.130242824554443,
+ "step": 6418
+ },
+ {
+ "epoch": 89.15371179039302,
+ "grad_norm": 0.015284297987818718,
+ "learning_rate": 0.0006,
+ "loss": 4.18555212020874,
+ "step": 6419
+ },
+ {
+ "epoch": 89.16768558951965,
+ "grad_norm": 0.01476874016225338,
+ "learning_rate": 0.0006,
+ "loss": 4.134175777435303,
+ "step": 6420
+ },
+ {
+ "epoch": 89.18165938864628,
+ "grad_norm": 0.015836201608181,
+ "learning_rate": 0.0006,
+ "loss": 4.094529151916504,
+ "step": 6421
+ },
+ {
+ "epoch": 89.19563318777293,
+ "grad_norm": 0.0161928478628397,
+ "learning_rate": 0.0006,
+ "loss": 4.12647819519043,
+ "step": 6422
+ },
+ {
+ "epoch": 89.20960698689956,
+ "grad_norm": 0.015680814161896706,
+ "learning_rate": 0.0006,
+ "loss": 3.9620392322540283,
+ "step": 6423
+ },
+ {
+ "epoch": 89.2235807860262,
+ "grad_norm": 0.01584681309759617,
+ "learning_rate": 0.0006,
+ "loss": 4.049304008483887,
+ "step": 6424
+ },
+ {
+ "epoch": 89.23755458515284,
+ "grad_norm": 0.01452907919883728,
+ "learning_rate": 0.0006,
+ "loss": 4.064027786254883,
+ "step": 6425
+ },
+ {
+ "epoch": 89.25152838427948,
+ "grad_norm": 0.014002552255988121,
+ "learning_rate": 0.0006,
+ "loss": 4.151862621307373,
+ "step": 6426
+ },
+ {
+ "epoch": 89.26550218340611,
+ "grad_norm": 0.014658625237643719,
+ "learning_rate": 0.0006,
+ "loss": 4.1672210693359375,
+ "step": 6427
+ },
+ {
+ "epoch": 89.27947598253274,
+ "grad_norm": 0.015021766535937786,
+ "learning_rate": 0.0006,
+ "loss": 3.9742586612701416,
+ "step": 6428
+ },
+ {
+ "epoch": 89.29344978165939,
+ "grad_norm": 0.014681321568787098,
+ "learning_rate": 0.0006,
+ "loss": 4.024159908294678,
+ "step": 6429
+ },
+ {
+ "epoch": 89.30742358078602,
+ "grad_norm": 0.014161549508571625,
+ "learning_rate": 0.0006,
+ "loss": 4.105640888214111,
+ "step": 6430
+ },
+ {
+ "epoch": 89.32139737991267,
+ "grad_norm": 0.014801453799009323,
+ "learning_rate": 0.0006,
+ "loss": 4.123049736022949,
+ "step": 6431
+ },
+ {
+ "epoch": 89.3353711790393,
+ "grad_norm": 0.014307181350886822,
+ "learning_rate": 0.0006,
+ "loss": 4.086954593658447,
+ "step": 6432
+ },
+ {
+ "epoch": 89.34934497816595,
+ "grad_norm": 0.013299369253218174,
+ "learning_rate": 0.0006,
+ "loss": 4.109533786773682,
+ "step": 6433
+ },
+ {
+ "epoch": 89.36331877729258,
+ "grad_norm": 0.014388669282197952,
+ "learning_rate": 0.0006,
+ "loss": 4.222370147705078,
+ "step": 6434
+ },
+ {
+ "epoch": 89.37729257641921,
+ "grad_norm": 0.014501879923045635,
+ "learning_rate": 0.0006,
+ "loss": 4.224178314208984,
+ "step": 6435
+ },
+ {
+ "epoch": 89.39126637554585,
+ "grad_norm": 0.014109021984040737,
+ "learning_rate": 0.0006,
+ "loss": 4.112702369689941,
+ "step": 6436
+ },
+ {
+ "epoch": 89.40524017467249,
+ "grad_norm": 0.013521002605557442,
+ "learning_rate": 0.0006,
+ "loss": 4.010124683380127,
+ "step": 6437
+ },
+ {
+ "epoch": 89.41921397379913,
+ "grad_norm": 0.013288144953548908,
+ "learning_rate": 0.0006,
+ "loss": 4.127592086791992,
+ "step": 6438
+ },
+ {
+ "epoch": 89.43318777292576,
+ "grad_norm": 0.013882486149668694,
+ "learning_rate": 0.0006,
+ "loss": 4.031871318817139,
+ "step": 6439
+ },
+ {
+ "epoch": 89.44716157205241,
+ "grad_norm": 0.01289335172623396,
+ "learning_rate": 0.0006,
+ "loss": 4.0757551193237305,
+ "step": 6440
+ },
+ {
+ "epoch": 89.46113537117904,
+ "grad_norm": 0.01356302760541439,
+ "learning_rate": 0.0006,
+ "loss": 4.075678825378418,
+ "step": 6441
+ },
+ {
+ "epoch": 89.47510917030567,
+ "grad_norm": 0.01447434164583683,
+ "learning_rate": 0.0006,
+ "loss": 4.093453407287598,
+ "step": 6442
+ },
+ {
+ "epoch": 89.48908296943232,
+ "grad_norm": 0.014443481341004372,
+ "learning_rate": 0.0006,
+ "loss": 4.075372219085693,
+ "step": 6443
+ },
+ {
+ "epoch": 89.50305676855895,
+ "grad_norm": 0.01314218994230032,
+ "learning_rate": 0.0006,
+ "loss": 4.171533584594727,
+ "step": 6444
+ },
+ {
+ "epoch": 89.5170305676856,
+ "grad_norm": 0.01348346658051014,
+ "learning_rate": 0.0006,
+ "loss": 4.227236270904541,
+ "step": 6445
+ },
+ {
+ "epoch": 89.53100436681223,
+ "grad_norm": 0.013370679691433907,
+ "learning_rate": 0.0006,
+ "loss": 3.973062515258789,
+ "step": 6446
+ },
+ {
+ "epoch": 89.54497816593886,
+ "grad_norm": 0.01410642359405756,
+ "learning_rate": 0.0006,
+ "loss": 4.231175422668457,
+ "step": 6447
+ },
+ {
+ "epoch": 89.5589519650655,
+ "grad_norm": 0.01419337373226881,
+ "learning_rate": 0.0006,
+ "loss": 4.145868301391602,
+ "step": 6448
+ },
+ {
+ "epoch": 89.57292576419214,
+ "grad_norm": 0.013691053725779057,
+ "learning_rate": 0.0006,
+ "loss": 4.175131797790527,
+ "step": 6449
+ },
+ {
+ "epoch": 89.58689956331878,
+ "grad_norm": 0.013714492321014404,
+ "learning_rate": 0.0006,
+ "loss": 4.251893997192383,
+ "step": 6450
+ },
+ {
+ "epoch": 89.60087336244541,
+ "grad_norm": 0.01393197476863861,
+ "learning_rate": 0.0006,
+ "loss": 4.047079086303711,
+ "step": 6451
+ },
+ {
+ "epoch": 89.61484716157206,
+ "grad_norm": 0.013957309536635876,
+ "learning_rate": 0.0006,
+ "loss": 4.128530502319336,
+ "step": 6452
+ },
+ {
+ "epoch": 89.62882096069869,
+ "grad_norm": 0.013346714898943901,
+ "learning_rate": 0.0006,
+ "loss": 4.2296929359436035,
+ "step": 6453
+ },
+ {
+ "epoch": 89.64279475982532,
+ "grad_norm": 0.014243829064071178,
+ "learning_rate": 0.0006,
+ "loss": 4.128233909606934,
+ "step": 6454
+ },
+ {
+ "epoch": 89.65676855895197,
+ "grad_norm": 0.01606106385588646,
+ "learning_rate": 0.0006,
+ "loss": 4.202202320098877,
+ "step": 6455
+ },
+ {
+ "epoch": 89.6707423580786,
+ "grad_norm": 0.016074348241090775,
+ "learning_rate": 0.0006,
+ "loss": 4.215323448181152,
+ "step": 6456
+ },
+ {
+ "epoch": 89.68471615720524,
+ "grad_norm": 0.01611054316163063,
+ "learning_rate": 0.0006,
+ "loss": 4.096758842468262,
+ "step": 6457
+ },
+ {
+ "epoch": 89.69868995633188,
+ "grad_norm": 0.013788217678666115,
+ "learning_rate": 0.0006,
+ "loss": 4.098560810089111,
+ "step": 6458
+ },
+ {
+ "epoch": 89.71266375545852,
+ "grad_norm": 0.014408969320356846,
+ "learning_rate": 0.0006,
+ "loss": 4.107398509979248,
+ "step": 6459
+ },
+ {
+ "epoch": 89.72663755458515,
+ "grad_norm": 0.014765267260372639,
+ "learning_rate": 0.0006,
+ "loss": 4.057718276977539,
+ "step": 6460
+ },
+ {
+ "epoch": 89.74061135371178,
+ "grad_norm": 0.017395280301570892,
+ "learning_rate": 0.0006,
+ "loss": 4.100795269012451,
+ "step": 6461
+ },
+ {
+ "epoch": 89.75458515283843,
+ "grad_norm": 0.016821149736642838,
+ "learning_rate": 0.0006,
+ "loss": 4.153731346130371,
+ "step": 6462
+ },
+ {
+ "epoch": 89.76855895196506,
+ "grad_norm": 0.015633124858140945,
+ "learning_rate": 0.0006,
+ "loss": 4.061585903167725,
+ "step": 6463
+ },
+ {
+ "epoch": 89.78253275109171,
+ "grad_norm": 0.01451497245579958,
+ "learning_rate": 0.0006,
+ "loss": 4.1624884605407715,
+ "step": 6464
+ },
+ {
+ "epoch": 89.79650655021834,
+ "grad_norm": 0.013868392445147038,
+ "learning_rate": 0.0006,
+ "loss": 4.178536415100098,
+ "step": 6465
+ },
+ {
+ "epoch": 89.81048034934499,
+ "grad_norm": 0.014259875752031803,
+ "learning_rate": 0.0006,
+ "loss": 4.073114395141602,
+ "step": 6466
+ },
+ {
+ "epoch": 89.82445414847162,
+ "grad_norm": 0.013651594519615173,
+ "learning_rate": 0.0006,
+ "loss": 4.050265312194824,
+ "step": 6467
+ },
+ {
+ "epoch": 89.83842794759825,
+ "grad_norm": 0.014063458889722824,
+ "learning_rate": 0.0006,
+ "loss": 4.2014689445495605,
+ "step": 6468
+ },
+ {
+ "epoch": 89.8524017467249,
+ "grad_norm": 0.014179578050971031,
+ "learning_rate": 0.0006,
+ "loss": 4.110550880432129,
+ "step": 6469
+ },
+ {
+ "epoch": 89.86637554585153,
+ "grad_norm": 0.013616854324936867,
+ "learning_rate": 0.0006,
+ "loss": 4.1776580810546875,
+ "step": 6470
+ },
+ {
+ "epoch": 89.88034934497817,
+ "grad_norm": 0.014351330697536469,
+ "learning_rate": 0.0006,
+ "loss": 4.058526992797852,
+ "step": 6471
+ },
+ {
+ "epoch": 89.8943231441048,
+ "grad_norm": 0.0141244325786829,
+ "learning_rate": 0.0006,
+ "loss": 4.068673610687256,
+ "step": 6472
+ },
+ {
+ "epoch": 89.90829694323143,
+ "grad_norm": 0.01552131213247776,
+ "learning_rate": 0.0006,
+ "loss": 4.086787223815918,
+ "step": 6473
+ },
+ {
+ "epoch": 89.92227074235808,
+ "grad_norm": 0.014614894054830074,
+ "learning_rate": 0.0006,
+ "loss": 4.2190656661987305,
+ "step": 6474
+ },
+ {
+ "epoch": 89.93624454148471,
+ "grad_norm": 0.014369758777320385,
+ "learning_rate": 0.0006,
+ "loss": 4.117021560668945,
+ "step": 6475
+ },
+ {
+ "epoch": 89.95021834061136,
+ "grad_norm": 0.015352046117186546,
+ "learning_rate": 0.0006,
+ "loss": 4.167664527893066,
+ "step": 6476
+ },
+ {
+ "epoch": 89.96419213973799,
+ "grad_norm": 0.014314834959805012,
+ "learning_rate": 0.0006,
+ "loss": 4.160182952880859,
+ "step": 6477
+ },
+ {
+ "epoch": 89.97816593886463,
+ "grad_norm": 0.013511341996490955,
+ "learning_rate": 0.0006,
+ "loss": 4.096034049987793,
+ "step": 6478
+ },
+ {
+ "epoch": 89.99213973799127,
+ "grad_norm": 0.01308482326567173,
+ "learning_rate": 0.0006,
+ "loss": 4.062689781188965,
+ "step": 6479
+ },
+ {
+ "epoch": 90.0,
+ "grad_norm": 0.01442806702107191,
+ "learning_rate": 0.0006,
+ "loss": 4.1955885887146,
+ "step": 6480
+ },
+ {
+ "epoch": 90.0,
+ "eval_loss": 4.644347667694092,
+ "eval_runtime": 56.8677,
+ "eval_samples_per_second": 42.942,
+ "eval_steps_per_second": 1.354,
+ "step": 6480
+ },
+ {
+ "epoch": 90.01397379912663,
+ "grad_norm": 0.014839007519185543,
+ "learning_rate": 0.0006,
+ "loss": 4.171868324279785,
+ "step": 6481
+ },
+ {
+ "epoch": 90.02794759825328,
+ "grad_norm": 0.01476900465786457,
+ "learning_rate": 0.0006,
+ "loss": 4.119754791259766,
+ "step": 6482
+ },
+ {
+ "epoch": 90.04192139737991,
+ "grad_norm": 0.015179301612079144,
+ "learning_rate": 0.0006,
+ "loss": 4.06740140914917,
+ "step": 6483
+ },
+ {
+ "epoch": 90.05589519650655,
+ "grad_norm": 0.015552980825304985,
+ "learning_rate": 0.0006,
+ "loss": 4.112268447875977,
+ "step": 6484
+ },
+ {
+ "epoch": 90.06986899563319,
+ "grad_norm": 0.014930234290659428,
+ "learning_rate": 0.0006,
+ "loss": 4.071866989135742,
+ "step": 6485
+ },
+ {
+ "epoch": 90.08384279475983,
+ "grad_norm": 0.014124148525297642,
+ "learning_rate": 0.0006,
+ "loss": 4.013731956481934,
+ "step": 6486
+ },
+ {
+ "epoch": 90.09781659388646,
+ "grad_norm": 0.014204252511262894,
+ "learning_rate": 0.0006,
+ "loss": 4.133620738983154,
+ "step": 6487
+ },
+ {
+ "epoch": 90.1117903930131,
+ "grad_norm": 0.014845543541014194,
+ "learning_rate": 0.0006,
+ "loss": 4.122956275939941,
+ "step": 6488
+ },
+ {
+ "epoch": 90.12576419213974,
+ "grad_norm": 0.0146744754165411,
+ "learning_rate": 0.0006,
+ "loss": 4.087121963500977,
+ "step": 6489
+ },
+ {
+ "epoch": 90.13973799126637,
+ "grad_norm": 0.016007190570235252,
+ "learning_rate": 0.0006,
+ "loss": 3.98740816116333,
+ "step": 6490
+ },
+ {
+ "epoch": 90.15371179039302,
+ "grad_norm": 0.018072661012411118,
+ "learning_rate": 0.0006,
+ "loss": 4.1155009269714355,
+ "step": 6491
+ },
+ {
+ "epoch": 90.16768558951965,
+ "grad_norm": 0.016868166625499725,
+ "learning_rate": 0.0006,
+ "loss": 4.123727321624756,
+ "step": 6492
+ },
+ {
+ "epoch": 90.18165938864628,
+ "grad_norm": 0.017091726884245872,
+ "learning_rate": 0.0006,
+ "loss": 4.1405229568481445,
+ "step": 6493
+ },
+ {
+ "epoch": 90.19563318777293,
+ "grad_norm": 0.018329855054616928,
+ "learning_rate": 0.0006,
+ "loss": 4.093900680541992,
+ "step": 6494
+ },
+ {
+ "epoch": 90.20960698689956,
+ "grad_norm": 0.018906457349658012,
+ "learning_rate": 0.0006,
+ "loss": 4.0552849769592285,
+ "step": 6495
+ },
+ {
+ "epoch": 90.2235807860262,
+ "grad_norm": 0.017275869846343994,
+ "learning_rate": 0.0006,
+ "loss": 4.1310625076293945,
+ "step": 6496
+ },
+ {
+ "epoch": 90.23755458515284,
+ "grad_norm": 0.015029897913336754,
+ "learning_rate": 0.0006,
+ "loss": 4.100166320800781,
+ "step": 6497
+ },
+ {
+ "epoch": 90.25152838427948,
+ "grad_norm": 0.014891199767589569,
+ "learning_rate": 0.0006,
+ "loss": 4.1314802169799805,
+ "step": 6498
+ },
+ {
+ "epoch": 90.26550218340611,
+ "grad_norm": 0.014877690002322197,
+ "learning_rate": 0.0006,
+ "loss": 4.076098442077637,
+ "step": 6499
+ },
+ {
+ "epoch": 90.27947598253274,
+ "grad_norm": 0.014670864678919315,
+ "learning_rate": 0.0006,
+ "loss": 4.066137313842773,
+ "step": 6500
+ },
+ {
+ "epoch": 90.29344978165939,
+ "grad_norm": 0.01444234699010849,
+ "learning_rate": 0.0006,
+ "loss": 4.013591766357422,
+ "step": 6501
+ },
+ {
+ "epoch": 90.30742358078602,
+ "grad_norm": 0.01339263841509819,
+ "learning_rate": 0.0006,
+ "loss": 4.117959499359131,
+ "step": 6502
+ },
+ {
+ "epoch": 90.32139737991267,
+ "grad_norm": 0.014753810130059719,
+ "learning_rate": 0.0006,
+ "loss": 4.042947769165039,
+ "step": 6503
+ },
+ {
+ "epoch": 90.3353711790393,
+ "grad_norm": 0.014947964809834957,
+ "learning_rate": 0.0006,
+ "loss": 4.112810134887695,
+ "step": 6504
+ },
+ {
+ "epoch": 90.34934497816595,
+ "grad_norm": 0.015155353583395481,
+ "learning_rate": 0.0006,
+ "loss": 4.213214874267578,
+ "step": 6505
+ },
+ {
+ "epoch": 90.36331877729258,
+ "grad_norm": 0.015502216294407845,
+ "learning_rate": 0.0006,
+ "loss": 4.127619743347168,
+ "step": 6506
+ },
+ {
+ "epoch": 90.37729257641921,
+ "grad_norm": 0.015377781353890896,
+ "learning_rate": 0.0006,
+ "loss": 4.135862350463867,
+ "step": 6507
+ },
+ {
+ "epoch": 90.39126637554585,
+ "grad_norm": 0.01662006787955761,
+ "learning_rate": 0.0006,
+ "loss": 4.076877593994141,
+ "step": 6508
+ },
+ {
+ "epoch": 90.40524017467249,
+ "grad_norm": 0.016324395313858986,
+ "learning_rate": 0.0006,
+ "loss": 4.214895725250244,
+ "step": 6509
+ },
+ {
+ "epoch": 90.41921397379913,
+ "grad_norm": 0.016099024564027786,
+ "learning_rate": 0.0006,
+ "loss": 4.139717102050781,
+ "step": 6510
+ },
+ {
+ "epoch": 90.43318777292576,
+ "grad_norm": 0.014403503388166428,
+ "learning_rate": 0.0006,
+ "loss": 4.040343284606934,
+ "step": 6511
+ },
+ {
+ "epoch": 90.44716157205241,
+ "grad_norm": 0.01413416862487793,
+ "learning_rate": 0.0006,
+ "loss": 4.087702751159668,
+ "step": 6512
+ },
+ {
+ "epoch": 90.46113537117904,
+ "grad_norm": 0.01603989489376545,
+ "learning_rate": 0.0006,
+ "loss": 4.173558712005615,
+ "step": 6513
+ },
+ {
+ "epoch": 90.47510917030567,
+ "grad_norm": 0.016846206039190292,
+ "learning_rate": 0.0006,
+ "loss": 4.1955461502075195,
+ "step": 6514
+ },
+ {
+ "epoch": 90.48908296943232,
+ "grad_norm": 0.01652812585234642,
+ "learning_rate": 0.0006,
+ "loss": 4.061838150024414,
+ "step": 6515
+ },
+ {
+ "epoch": 90.50305676855895,
+ "grad_norm": 0.015100528486073017,
+ "learning_rate": 0.0006,
+ "loss": 4.041808605194092,
+ "step": 6516
+ },
+ {
+ "epoch": 90.5170305676856,
+ "grad_norm": 0.013867822475731373,
+ "learning_rate": 0.0006,
+ "loss": 4.12662410736084,
+ "step": 6517
+ },
+ {
+ "epoch": 90.53100436681223,
+ "grad_norm": 0.013684107922017574,
+ "learning_rate": 0.0006,
+ "loss": 4.1254048347473145,
+ "step": 6518
+ },
+ {
+ "epoch": 90.54497816593886,
+ "grad_norm": 0.014200013130903244,
+ "learning_rate": 0.0006,
+ "loss": 4.183216094970703,
+ "step": 6519
+ },
+ {
+ "epoch": 90.5589519650655,
+ "grad_norm": 0.013707022182643414,
+ "learning_rate": 0.0006,
+ "loss": 4.07176399230957,
+ "step": 6520
+ },
+ {
+ "epoch": 90.57292576419214,
+ "grad_norm": 0.013996582478284836,
+ "learning_rate": 0.0006,
+ "loss": 4.006575584411621,
+ "step": 6521
+ },
+ {
+ "epoch": 90.58689956331878,
+ "grad_norm": 0.014886210672557354,
+ "learning_rate": 0.0006,
+ "loss": 4.118956565856934,
+ "step": 6522
+ },
+ {
+ "epoch": 90.60087336244541,
+ "grad_norm": 0.014022685587406158,
+ "learning_rate": 0.0006,
+ "loss": 3.96986722946167,
+ "step": 6523
+ },
+ {
+ "epoch": 90.61484716157206,
+ "grad_norm": 0.014547634869813919,
+ "learning_rate": 0.0006,
+ "loss": 4.122886657714844,
+ "step": 6524
+ },
+ {
+ "epoch": 90.62882096069869,
+ "grad_norm": 0.013682851567864418,
+ "learning_rate": 0.0006,
+ "loss": 3.988292694091797,
+ "step": 6525
+ },
+ {
+ "epoch": 90.64279475982532,
+ "grad_norm": 0.014016421511769295,
+ "learning_rate": 0.0006,
+ "loss": 4.110270977020264,
+ "step": 6526
+ },
+ {
+ "epoch": 90.65676855895197,
+ "grad_norm": 0.013392233289778233,
+ "learning_rate": 0.0006,
+ "loss": 4.065059661865234,
+ "step": 6527
+ },
+ {
+ "epoch": 90.6707423580786,
+ "grad_norm": 0.012107795104384422,
+ "learning_rate": 0.0006,
+ "loss": 4.034260272979736,
+ "step": 6528
+ },
+ {
+ "epoch": 90.68471615720524,
+ "grad_norm": 0.013198604807257652,
+ "learning_rate": 0.0006,
+ "loss": 4.162083625793457,
+ "step": 6529
+ },
+ {
+ "epoch": 90.69868995633188,
+ "grad_norm": 0.012380579486489296,
+ "learning_rate": 0.0006,
+ "loss": 4.040456295013428,
+ "step": 6530
+ },
+ {
+ "epoch": 90.71266375545852,
+ "grad_norm": 0.012784135527908802,
+ "learning_rate": 0.0006,
+ "loss": 4.0823798179626465,
+ "step": 6531
+ },
+ {
+ "epoch": 90.72663755458515,
+ "grad_norm": 0.012920220382511616,
+ "learning_rate": 0.0006,
+ "loss": 4.181064128875732,
+ "step": 6532
+ },
+ {
+ "epoch": 90.74061135371178,
+ "grad_norm": 0.01338814664632082,
+ "learning_rate": 0.0006,
+ "loss": 4.111660003662109,
+ "step": 6533
+ },
+ {
+ "epoch": 90.75458515283843,
+ "grad_norm": 0.013628040440380573,
+ "learning_rate": 0.0006,
+ "loss": 4.237852573394775,
+ "step": 6534
+ },
+ {
+ "epoch": 90.76855895196506,
+ "grad_norm": 0.013384854421019554,
+ "learning_rate": 0.0006,
+ "loss": 4.121862888336182,
+ "step": 6535
+ },
+ {
+ "epoch": 90.78253275109171,
+ "grad_norm": 0.014726929366588593,
+ "learning_rate": 0.0006,
+ "loss": 4.1695990562438965,
+ "step": 6536
+ },
+ {
+ "epoch": 90.79650655021834,
+ "grad_norm": 0.0161006897687912,
+ "learning_rate": 0.0006,
+ "loss": 4.17794132232666,
+ "step": 6537
+ },
+ {
+ "epoch": 90.81048034934499,
+ "grad_norm": 0.016761884093284607,
+ "learning_rate": 0.0006,
+ "loss": 4.093764781951904,
+ "step": 6538
+ },
+ {
+ "epoch": 90.82445414847162,
+ "grad_norm": 0.017246825620532036,
+ "learning_rate": 0.0006,
+ "loss": 3.9382331371307373,
+ "step": 6539
+ },
+ {
+ "epoch": 90.83842794759825,
+ "grad_norm": 0.018141290172934532,
+ "learning_rate": 0.0006,
+ "loss": 3.973609209060669,
+ "step": 6540
+ },
+ {
+ "epoch": 90.8524017467249,
+ "grad_norm": 0.018129250034689903,
+ "learning_rate": 0.0006,
+ "loss": 4.227414131164551,
+ "step": 6541
+ },
+ {
+ "epoch": 90.86637554585153,
+ "grad_norm": 0.018234573304653168,
+ "learning_rate": 0.0006,
+ "loss": 4.151870250701904,
+ "step": 6542
+ },
+ {
+ "epoch": 90.88034934497817,
+ "grad_norm": 0.01560172438621521,
+ "learning_rate": 0.0006,
+ "loss": 4.127397537231445,
+ "step": 6543
+ },
+ {
+ "epoch": 90.8943231441048,
+ "grad_norm": 0.014367681927978992,
+ "learning_rate": 0.0006,
+ "loss": 4.17820930480957,
+ "step": 6544
+ },
+ {
+ "epoch": 90.90829694323143,
+ "grad_norm": 0.016071319580078125,
+ "learning_rate": 0.0006,
+ "loss": 4.059854507446289,
+ "step": 6545
+ },
+ {
+ "epoch": 90.92227074235808,
+ "grad_norm": 0.01589169353246689,
+ "learning_rate": 0.0006,
+ "loss": 4.282748222351074,
+ "step": 6546
+ },
+ {
+ "epoch": 90.93624454148471,
+ "grad_norm": 0.013639772310853004,
+ "learning_rate": 0.0006,
+ "loss": 4.147368907928467,
+ "step": 6547
+ },
+ {
+ "epoch": 90.95021834061136,
+ "grad_norm": 0.014248427003622055,
+ "learning_rate": 0.0006,
+ "loss": 4.240248680114746,
+ "step": 6548
+ },
+ {
+ "epoch": 90.96419213973799,
+ "grad_norm": 0.014600320719182491,
+ "learning_rate": 0.0006,
+ "loss": 4.122123718261719,
+ "step": 6549
+ },
+ {
+ "epoch": 90.97816593886463,
+ "grad_norm": 0.014371675439178944,
+ "learning_rate": 0.0006,
+ "loss": 4.013302803039551,
+ "step": 6550
+ },
+ {
+ "epoch": 90.99213973799127,
+ "grad_norm": 0.013133526779711246,
+ "learning_rate": 0.0006,
+ "loss": 4.097136497497559,
+ "step": 6551
+ },
+ {
+ "epoch": 91.0,
+ "grad_norm": 0.015599353238940239,
+ "learning_rate": 0.0006,
+ "loss": 4.198554039001465,
+ "step": 6552
+ },
+ {
+ "epoch": 91.0,
+ "eval_loss": 4.603907585144043,
+ "eval_runtime": 56.9418,
+ "eval_samples_per_second": 42.886,
+ "eval_steps_per_second": 1.352,
+ "step": 6552
+ },
+ {
+ "epoch": 91.01397379912663,
+ "grad_norm": 0.015917908400297165,
+ "learning_rate": 0.0006,
+ "loss": 4.133877754211426,
+ "step": 6553
+ },
+ {
+ "epoch": 91.02794759825328,
+ "grad_norm": 0.015485800802707672,
+ "learning_rate": 0.0006,
+ "loss": 4.111617088317871,
+ "step": 6554
+ },
+ {
+ "epoch": 91.04192139737991,
+ "grad_norm": 0.012969603762030602,
+ "learning_rate": 0.0006,
+ "loss": 4.064828395843506,
+ "step": 6555
+ },
+ {
+ "epoch": 91.05589519650655,
+ "grad_norm": 0.014482097700238228,
+ "learning_rate": 0.0006,
+ "loss": 4.101710319519043,
+ "step": 6556
+ },
+ {
+ "epoch": 91.06986899563319,
+ "grad_norm": 0.017242740839719772,
+ "learning_rate": 0.0006,
+ "loss": 4.015342712402344,
+ "step": 6557
+ },
+ {
+ "epoch": 91.08384279475983,
+ "grad_norm": 0.017100440338253975,
+ "learning_rate": 0.0006,
+ "loss": 3.9801602363586426,
+ "step": 6558
+ },
+ {
+ "epoch": 91.09781659388646,
+ "grad_norm": 0.015969131141901016,
+ "learning_rate": 0.0006,
+ "loss": 4.049958229064941,
+ "step": 6559
+ },
+ {
+ "epoch": 91.1117903930131,
+ "grad_norm": 0.016018927097320557,
+ "learning_rate": 0.0006,
+ "loss": 4.094139099121094,
+ "step": 6560
+ },
+ {
+ "epoch": 91.12576419213974,
+ "grad_norm": 0.018131926655769348,
+ "learning_rate": 0.0006,
+ "loss": 3.9499974250793457,
+ "step": 6561
+ },
+ {
+ "epoch": 91.13973799126637,
+ "grad_norm": 0.019425220787525177,
+ "learning_rate": 0.0006,
+ "loss": 3.9910552501678467,
+ "step": 6562
+ },
+ {
+ "epoch": 91.15371179039302,
+ "grad_norm": 0.01908663846552372,
+ "learning_rate": 0.0006,
+ "loss": 4.032179832458496,
+ "step": 6563
+ },
+ {
+ "epoch": 91.16768558951965,
+ "grad_norm": 0.01871207356452942,
+ "learning_rate": 0.0006,
+ "loss": 4.112380027770996,
+ "step": 6564
+ },
+ {
+ "epoch": 91.18165938864628,
+ "grad_norm": 0.020867953076958656,
+ "learning_rate": 0.0006,
+ "loss": 4.06105899810791,
+ "step": 6565
+ },
+ {
+ "epoch": 91.19563318777293,
+ "grad_norm": 0.019053412601351738,
+ "learning_rate": 0.0006,
+ "loss": 3.992504596710205,
+ "step": 6566
+ },
+ {
+ "epoch": 91.20960698689956,
+ "grad_norm": 0.016084402799606323,
+ "learning_rate": 0.0006,
+ "loss": 4.100559234619141,
+ "step": 6567
+ },
+ {
+ "epoch": 91.2235807860262,
+ "grad_norm": 0.016857590526342392,
+ "learning_rate": 0.0006,
+ "loss": 4.1567277908325195,
+ "step": 6568
+ },
+ {
+ "epoch": 91.23755458515284,
+ "grad_norm": 0.01702957972884178,
+ "learning_rate": 0.0006,
+ "loss": 4.234771728515625,
+ "step": 6569
+ },
+ {
+ "epoch": 91.25152838427948,
+ "grad_norm": 0.016869494691491127,
+ "learning_rate": 0.0006,
+ "loss": 3.9931387901306152,
+ "step": 6570
+ },
+ {
+ "epoch": 91.26550218340611,
+ "grad_norm": 0.01696581020951271,
+ "learning_rate": 0.0006,
+ "loss": 4.184908866882324,
+ "step": 6571
+ },
+ {
+ "epoch": 91.27947598253274,
+ "grad_norm": 0.014982126653194427,
+ "learning_rate": 0.0006,
+ "loss": 4.036668300628662,
+ "step": 6572
+ },
+ {
+ "epoch": 91.29344978165939,
+ "grad_norm": 0.014363301917910576,
+ "learning_rate": 0.0006,
+ "loss": 4.074010848999023,
+ "step": 6573
+ },
+ {
+ "epoch": 91.30742358078602,
+ "grad_norm": 0.015492337755858898,
+ "learning_rate": 0.0006,
+ "loss": 4.127626419067383,
+ "step": 6574
+ },
+ {
+ "epoch": 91.32139737991267,
+ "grad_norm": 0.015375050716102123,
+ "learning_rate": 0.0006,
+ "loss": 4.159229755401611,
+ "step": 6575
+ },
+ {
+ "epoch": 91.3353711790393,
+ "grad_norm": 0.015068565495312214,
+ "learning_rate": 0.0006,
+ "loss": 4.172623634338379,
+ "step": 6576
+ },
+ {
+ "epoch": 91.34934497816595,
+ "grad_norm": 0.0140256742015481,
+ "learning_rate": 0.0006,
+ "loss": 4.207065105438232,
+ "step": 6577
+ },
+ {
+ "epoch": 91.36331877729258,
+ "grad_norm": 0.014220152981579304,
+ "learning_rate": 0.0006,
+ "loss": 4.131444931030273,
+ "step": 6578
+ },
+ {
+ "epoch": 91.37729257641921,
+ "grad_norm": 0.0143149234354496,
+ "learning_rate": 0.0006,
+ "loss": 4.065218448638916,
+ "step": 6579
+ },
+ {
+ "epoch": 91.39126637554585,
+ "grad_norm": 0.013149659149348736,
+ "learning_rate": 0.0006,
+ "loss": 4.106743335723877,
+ "step": 6580
+ },
+ {
+ "epoch": 91.40524017467249,
+ "grad_norm": 0.014550330117344856,
+ "learning_rate": 0.0006,
+ "loss": 4.123722553253174,
+ "step": 6581
+ },
+ {
+ "epoch": 91.41921397379913,
+ "grad_norm": 0.014611710794270039,
+ "learning_rate": 0.0006,
+ "loss": 4.079610824584961,
+ "step": 6582
+ },
+ {
+ "epoch": 91.43318777292576,
+ "grad_norm": 0.013369769789278507,
+ "learning_rate": 0.0006,
+ "loss": 4.133187770843506,
+ "step": 6583
+ },
+ {
+ "epoch": 91.44716157205241,
+ "grad_norm": 0.013243545778095722,
+ "learning_rate": 0.0006,
+ "loss": 4.042607307434082,
+ "step": 6584
+ },
+ {
+ "epoch": 91.46113537117904,
+ "grad_norm": 0.013388266786932945,
+ "learning_rate": 0.0006,
+ "loss": 4.14863395690918,
+ "step": 6585
+ },
+ {
+ "epoch": 91.47510917030567,
+ "grad_norm": 0.014333273284137249,
+ "learning_rate": 0.0006,
+ "loss": 4.109699726104736,
+ "step": 6586
+ },
+ {
+ "epoch": 91.48908296943232,
+ "grad_norm": 0.012742757797241211,
+ "learning_rate": 0.0006,
+ "loss": 4.126582145690918,
+ "step": 6587
+ },
+ {
+ "epoch": 91.50305676855895,
+ "grad_norm": 0.012880304828286171,
+ "learning_rate": 0.0006,
+ "loss": 4.205300331115723,
+ "step": 6588
+ },
+ {
+ "epoch": 91.5170305676856,
+ "grad_norm": 0.013896043412387371,
+ "learning_rate": 0.0006,
+ "loss": 4.132510185241699,
+ "step": 6589
+ },
+ {
+ "epoch": 91.53100436681223,
+ "grad_norm": 0.014361281879246235,
+ "learning_rate": 0.0006,
+ "loss": 4.1487932205200195,
+ "step": 6590
+ },
+ {
+ "epoch": 91.54497816593886,
+ "grad_norm": 0.015137741342186928,
+ "learning_rate": 0.0006,
+ "loss": 4.1564531326293945,
+ "step": 6591
+ },
+ {
+ "epoch": 91.5589519650655,
+ "grad_norm": 0.014660682529211044,
+ "learning_rate": 0.0006,
+ "loss": 4.056064128875732,
+ "step": 6592
+ },
+ {
+ "epoch": 91.57292576419214,
+ "grad_norm": 0.014888476580381393,
+ "learning_rate": 0.0006,
+ "loss": 4.0633344650268555,
+ "step": 6593
+ },
+ {
+ "epoch": 91.58689956331878,
+ "grad_norm": 0.012162433005869389,
+ "learning_rate": 0.0006,
+ "loss": 4.026293754577637,
+ "step": 6594
+ },
+ {
+ "epoch": 91.60087336244541,
+ "grad_norm": 0.012892271392047405,
+ "learning_rate": 0.0006,
+ "loss": 4.267086029052734,
+ "step": 6595
+ },
+ {
+ "epoch": 91.61484716157206,
+ "grad_norm": 0.01450759544968605,
+ "learning_rate": 0.0006,
+ "loss": 4.149359703063965,
+ "step": 6596
+ },
+ {
+ "epoch": 91.62882096069869,
+ "grad_norm": 0.01419843453913927,
+ "learning_rate": 0.0006,
+ "loss": 4.094183921813965,
+ "step": 6597
+ },
+ {
+ "epoch": 91.64279475982532,
+ "grad_norm": 0.014401616528630257,
+ "learning_rate": 0.0006,
+ "loss": 4.125172138214111,
+ "step": 6598
+ },
+ {
+ "epoch": 91.65676855895197,
+ "grad_norm": 0.013945513404905796,
+ "learning_rate": 0.0006,
+ "loss": 4.146623134613037,
+ "step": 6599
+ },
+ {
+ "epoch": 91.6707423580786,
+ "grad_norm": 0.013516047969460487,
+ "learning_rate": 0.0006,
+ "loss": 4.160526275634766,
+ "step": 6600
+ },
+ {
+ "epoch": 91.68471615720524,
+ "grad_norm": 0.013262378051877022,
+ "learning_rate": 0.0006,
+ "loss": 4.074923038482666,
+ "step": 6601
+ },
+ {
+ "epoch": 91.69868995633188,
+ "grad_norm": 0.014365669339895248,
+ "learning_rate": 0.0006,
+ "loss": 4.118094444274902,
+ "step": 6602
+ },
+ {
+ "epoch": 91.71266375545852,
+ "grad_norm": 0.016713734716176987,
+ "learning_rate": 0.0006,
+ "loss": 4.021799087524414,
+ "step": 6603
+ },
+ {
+ "epoch": 91.72663755458515,
+ "grad_norm": 0.01988680474460125,
+ "learning_rate": 0.0006,
+ "loss": 4.238436222076416,
+ "step": 6604
+ },
+ {
+ "epoch": 91.74061135371178,
+ "grad_norm": 0.022749733179807663,
+ "learning_rate": 0.0006,
+ "loss": 4.010470390319824,
+ "step": 6605
+ },
+ {
+ "epoch": 91.75458515283843,
+ "grad_norm": 0.022533461451530457,
+ "learning_rate": 0.0006,
+ "loss": 4.152201175689697,
+ "step": 6606
+ },
+ {
+ "epoch": 91.76855895196506,
+ "grad_norm": 0.02021082676947117,
+ "learning_rate": 0.0006,
+ "loss": 4.120572090148926,
+ "step": 6607
+ },
+ {
+ "epoch": 91.78253275109171,
+ "grad_norm": 0.016298197209835052,
+ "learning_rate": 0.0006,
+ "loss": 4.115815162658691,
+ "step": 6608
+ },
+ {
+ "epoch": 91.79650655021834,
+ "grad_norm": 0.017562786117196083,
+ "learning_rate": 0.0006,
+ "loss": 4.109724044799805,
+ "step": 6609
+ },
+ {
+ "epoch": 91.81048034934499,
+ "grad_norm": 0.018153153359889984,
+ "learning_rate": 0.0006,
+ "loss": 4.143759727478027,
+ "step": 6610
+ },
+ {
+ "epoch": 91.82445414847162,
+ "grad_norm": 0.014762306585907936,
+ "learning_rate": 0.0006,
+ "loss": 4.25596809387207,
+ "step": 6611
+ },
+ {
+ "epoch": 91.83842794759825,
+ "grad_norm": 0.015074343420565128,
+ "learning_rate": 0.0006,
+ "loss": 4.207723140716553,
+ "step": 6612
+ },
+ {
+ "epoch": 91.8524017467249,
+ "grad_norm": 0.015028579160571098,
+ "learning_rate": 0.0006,
+ "loss": 4.01950216293335,
+ "step": 6613
+ },
+ {
+ "epoch": 91.86637554585153,
+ "grad_norm": 0.013996795751154423,
+ "learning_rate": 0.0006,
+ "loss": 4.132997989654541,
+ "step": 6614
+ },
+ {
+ "epoch": 91.88034934497817,
+ "grad_norm": 0.014978236518800259,
+ "learning_rate": 0.0006,
+ "loss": 4.132402420043945,
+ "step": 6615
+ },
+ {
+ "epoch": 91.8943231441048,
+ "grad_norm": 0.013111517764627934,
+ "learning_rate": 0.0006,
+ "loss": 4.1223649978637695,
+ "step": 6616
+ },
+ {
+ "epoch": 91.90829694323143,
+ "grad_norm": 0.013872234150767326,
+ "learning_rate": 0.0006,
+ "loss": 4.096590042114258,
+ "step": 6617
+ },
+ {
+ "epoch": 91.92227074235808,
+ "grad_norm": 0.013642470352351665,
+ "learning_rate": 0.0006,
+ "loss": 4.236227035522461,
+ "step": 6618
+ },
+ {
+ "epoch": 91.93624454148471,
+ "grad_norm": 0.014684873633086681,
+ "learning_rate": 0.0006,
+ "loss": 4.037961959838867,
+ "step": 6619
+ },
+ {
+ "epoch": 91.95021834061136,
+ "grad_norm": 0.014498206786811352,
+ "learning_rate": 0.0006,
+ "loss": 4.043325424194336,
+ "step": 6620
+ },
+ {
+ "epoch": 91.96419213973799,
+ "grad_norm": 0.012768656946718693,
+ "learning_rate": 0.0006,
+ "loss": 4.138803482055664,
+ "step": 6621
+ },
+ {
+ "epoch": 91.97816593886463,
+ "grad_norm": 0.012155905365943909,
+ "learning_rate": 0.0006,
+ "loss": 4.163424968719482,
+ "step": 6622
+ },
+ {
+ "epoch": 91.99213973799127,
+ "grad_norm": 0.012584930285811424,
+ "learning_rate": 0.0006,
+ "loss": 4.086629867553711,
+ "step": 6623
+ },
+ {
+ "epoch": 92.0,
+ "grad_norm": 0.01578151248395443,
+ "learning_rate": 0.0006,
+ "loss": 4.108386993408203,
+ "step": 6624
+ },
+ {
+ "epoch": 92.0,
+ "eval_loss": 4.541166305541992,
+ "eval_runtime": 56.871,
+ "eval_samples_per_second": 42.939,
+ "eval_steps_per_second": 1.354,
+ "step": 6624
+ },
+ {
+ "epoch": 92.01397379912663,
+ "grad_norm": 0.016901545226573944,
+ "learning_rate": 0.0006,
+ "loss": 4.043694496154785,
+ "step": 6625
+ },
+ {
+ "epoch": 92.02794759825328,
+ "grad_norm": 0.016299039125442505,
+ "learning_rate": 0.0006,
+ "loss": 4.100504398345947,
+ "step": 6626
+ },
+ {
+ "epoch": 92.04192139737991,
+ "grad_norm": 0.013988777063786983,
+ "learning_rate": 0.0006,
+ "loss": 4.063943862915039,
+ "step": 6627
+ },
+ {
+ "epoch": 92.05589519650655,
+ "grad_norm": 0.01343096699565649,
+ "learning_rate": 0.0006,
+ "loss": 4.0321197509765625,
+ "step": 6628
+ },
+ {
+ "epoch": 92.06986899563319,
+ "grad_norm": 0.014343147166073322,
+ "learning_rate": 0.0006,
+ "loss": 4.095499038696289,
+ "step": 6629
+ },
+ {
+ "epoch": 92.08384279475983,
+ "grad_norm": 0.014989756047725677,
+ "learning_rate": 0.0006,
+ "loss": 4.122310638427734,
+ "step": 6630
+ },
+ {
+ "epoch": 92.09781659388646,
+ "grad_norm": 0.013929174281656742,
+ "learning_rate": 0.0006,
+ "loss": 4.02870512008667,
+ "step": 6631
+ },
+ {
+ "epoch": 92.1117903930131,
+ "grad_norm": 0.014139998704195023,
+ "learning_rate": 0.0006,
+ "loss": 4.180675506591797,
+ "step": 6632
+ },
+ {
+ "epoch": 92.12576419213974,
+ "grad_norm": 0.012920474633574486,
+ "learning_rate": 0.0006,
+ "loss": 4.111686706542969,
+ "step": 6633
+ },
+ {
+ "epoch": 92.13973799126637,
+ "grad_norm": 0.01291937381029129,
+ "learning_rate": 0.0006,
+ "loss": 3.958092212677002,
+ "step": 6634
+ },
+ {
+ "epoch": 92.15371179039302,
+ "grad_norm": 0.013250474818050861,
+ "learning_rate": 0.0006,
+ "loss": 4.075105667114258,
+ "step": 6635
+ },
+ {
+ "epoch": 92.16768558951965,
+ "grad_norm": 0.01321406476199627,
+ "learning_rate": 0.0006,
+ "loss": 4.072322368621826,
+ "step": 6636
+ },
+ {
+ "epoch": 92.18165938864628,
+ "grad_norm": 0.014002865180373192,
+ "learning_rate": 0.0006,
+ "loss": 4.061572074890137,
+ "step": 6637
+ },
+ {
+ "epoch": 92.19563318777293,
+ "grad_norm": 0.013876105658710003,
+ "learning_rate": 0.0006,
+ "loss": 4.099213600158691,
+ "step": 6638
+ },
+ {
+ "epoch": 92.20960698689956,
+ "grad_norm": 0.01464459951967001,
+ "learning_rate": 0.0006,
+ "loss": 4.1573638916015625,
+ "step": 6639
+ },
+ {
+ "epoch": 92.2235807860262,
+ "grad_norm": 0.015354706905782223,
+ "learning_rate": 0.0006,
+ "loss": 4.078714370727539,
+ "step": 6640
+ },
+ {
+ "epoch": 92.23755458515284,
+ "grad_norm": 0.015265689231455326,
+ "learning_rate": 0.0006,
+ "loss": 4.036652565002441,
+ "step": 6641
+ },
+ {
+ "epoch": 92.25152838427948,
+ "grad_norm": 0.016140855848789215,
+ "learning_rate": 0.0006,
+ "loss": 4.120541095733643,
+ "step": 6642
+ },
+ {
+ "epoch": 92.26550218340611,
+ "grad_norm": 0.017044052481651306,
+ "learning_rate": 0.0006,
+ "loss": 4.015849590301514,
+ "step": 6643
+ },
+ {
+ "epoch": 92.27947598253274,
+ "grad_norm": 0.01469327136874199,
+ "learning_rate": 0.0006,
+ "loss": 4.137286186218262,
+ "step": 6644
+ },
+ {
+ "epoch": 92.29344978165939,
+ "grad_norm": 0.014596942812204361,
+ "learning_rate": 0.0006,
+ "loss": 4.11309814453125,
+ "step": 6645
+ },
+ {
+ "epoch": 92.30742358078602,
+ "grad_norm": 0.014671559445559978,
+ "learning_rate": 0.0006,
+ "loss": 4.087950229644775,
+ "step": 6646
+ },
+ {
+ "epoch": 92.32139737991267,
+ "grad_norm": 0.015040507540106773,
+ "learning_rate": 0.0006,
+ "loss": 3.967053174972534,
+ "step": 6647
+ },
+ {
+ "epoch": 92.3353711790393,
+ "grad_norm": 0.01417640782892704,
+ "learning_rate": 0.0006,
+ "loss": 3.990346908569336,
+ "step": 6648
+ },
+ {
+ "epoch": 92.34934497816595,
+ "grad_norm": 0.013855187222361565,
+ "learning_rate": 0.0006,
+ "loss": 3.965132474899292,
+ "step": 6649
+ },
+ {
+ "epoch": 92.36331877729258,
+ "grad_norm": 0.014291216619312763,
+ "learning_rate": 0.0006,
+ "loss": 4.161717891693115,
+ "step": 6650
+ },
+ {
+ "epoch": 92.37729257641921,
+ "grad_norm": 0.014148853719234467,
+ "learning_rate": 0.0006,
+ "loss": 4.019951820373535,
+ "step": 6651
+ },
+ {
+ "epoch": 92.39126637554585,
+ "grad_norm": 0.01414303295314312,
+ "learning_rate": 0.0006,
+ "loss": 4.125976085662842,
+ "step": 6652
+ },
+ {
+ "epoch": 92.40524017467249,
+ "grad_norm": 0.01358176488429308,
+ "learning_rate": 0.0006,
+ "loss": 4.088022232055664,
+ "step": 6653
+ },
+ {
+ "epoch": 92.41921397379913,
+ "grad_norm": 0.014174354262650013,
+ "learning_rate": 0.0006,
+ "loss": 4.1137518882751465,
+ "step": 6654
+ },
+ {
+ "epoch": 92.43318777292576,
+ "grad_norm": 0.013756301254034042,
+ "learning_rate": 0.0006,
+ "loss": 4.0088725090026855,
+ "step": 6655
+ },
+ {
+ "epoch": 92.44716157205241,
+ "grad_norm": 0.014257770963013172,
+ "learning_rate": 0.0006,
+ "loss": 4.05248498916626,
+ "step": 6656
+ },
+ {
+ "epoch": 92.46113537117904,
+ "grad_norm": 0.015319934114813805,
+ "learning_rate": 0.0006,
+ "loss": 4.1375732421875,
+ "step": 6657
+ },
+ {
+ "epoch": 92.47510917030567,
+ "grad_norm": 0.014655457809567451,
+ "learning_rate": 0.0006,
+ "loss": 4.172857284545898,
+ "step": 6658
+ },
+ {
+ "epoch": 92.48908296943232,
+ "grad_norm": 0.014981928281486034,
+ "learning_rate": 0.0006,
+ "loss": 4.064846038818359,
+ "step": 6659
+ },
+ {
+ "epoch": 92.50305676855895,
+ "grad_norm": 0.015197387896478176,
+ "learning_rate": 0.0006,
+ "loss": 4.119397163391113,
+ "step": 6660
+ },
+ {
+ "epoch": 92.5170305676856,
+ "grad_norm": 0.016060935333371162,
+ "learning_rate": 0.0006,
+ "loss": 4.09212064743042,
+ "step": 6661
+ },
+ {
+ "epoch": 92.53100436681223,
+ "grad_norm": 0.014884199015796185,
+ "learning_rate": 0.0006,
+ "loss": 4.027483940124512,
+ "step": 6662
+ },
+ {
+ "epoch": 92.54497816593886,
+ "grad_norm": 0.01562194712460041,
+ "learning_rate": 0.0006,
+ "loss": 3.9722909927368164,
+ "step": 6663
+ },
+ {
+ "epoch": 92.5589519650655,
+ "grad_norm": 0.014801043085753918,
+ "learning_rate": 0.0006,
+ "loss": 4.048130989074707,
+ "step": 6664
+ },
+ {
+ "epoch": 92.57292576419214,
+ "grad_norm": 0.013954494148492813,
+ "learning_rate": 0.0006,
+ "loss": 4.050351619720459,
+ "step": 6665
+ },
+ {
+ "epoch": 92.58689956331878,
+ "grad_norm": 0.013292429968714714,
+ "learning_rate": 0.0006,
+ "loss": 4.245860576629639,
+ "step": 6666
+ },
+ {
+ "epoch": 92.60087336244541,
+ "grad_norm": 0.015737095847725868,
+ "learning_rate": 0.0006,
+ "loss": 4.154323101043701,
+ "step": 6667
+ },
+ {
+ "epoch": 92.61484716157206,
+ "grad_norm": 0.018242565914988518,
+ "learning_rate": 0.0006,
+ "loss": 4.118710517883301,
+ "step": 6668
+ },
+ {
+ "epoch": 92.62882096069869,
+ "grad_norm": 0.017666636034846306,
+ "learning_rate": 0.0006,
+ "loss": 4.1013875007629395,
+ "step": 6669
+ },
+ {
+ "epoch": 92.64279475982532,
+ "grad_norm": 0.015017223544418812,
+ "learning_rate": 0.0006,
+ "loss": 4.084001064300537,
+ "step": 6670
+ },
+ {
+ "epoch": 92.65676855895197,
+ "grad_norm": 0.01359238475561142,
+ "learning_rate": 0.0006,
+ "loss": 4.091456413269043,
+ "step": 6671
+ },
+ {
+ "epoch": 92.6707423580786,
+ "grad_norm": 0.014524881727993488,
+ "learning_rate": 0.0006,
+ "loss": 4.121245384216309,
+ "step": 6672
+ },
+ {
+ "epoch": 92.68471615720524,
+ "grad_norm": 0.016214480623602867,
+ "learning_rate": 0.0006,
+ "loss": 4.133184432983398,
+ "step": 6673
+ },
+ {
+ "epoch": 92.69868995633188,
+ "grad_norm": 0.0176643505692482,
+ "learning_rate": 0.0006,
+ "loss": 4.142827987670898,
+ "step": 6674
+ },
+ {
+ "epoch": 92.71266375545852,
+ "grad_norm": 0.017004093155264854,
+ "learning_rate": 0.0006,
+ "loss": 4.067666530609131,
+ "step": 6675
+ },
+ {
+ "epoch": 92.72663755458515,
+ "grad_norm": 0.014077397994697094,
+ "learning_rate": 0.0006,
+ "loss": 4.076465129852295,
+ "step": 6676
+ },
+ {
+ "epoch": 92.74061135371178,
+ "grad_norm": 0.015905627980828285,
+ "learning_rate": 0.0006,
+ "loss": 4.083726406097412,
+ "step": 6677
+ },
+ {
+ "epoch": 92.75458515283843,
+ "grad_norm": 0.016467567533254623,
+ "learning_rate": 0.0006,
+ "loss": 4.241217136383057,
+ "step": 6678
+ },
+ {
+ "epoch": 92.76855895196506,
+ "grad_norm": 0.01541806198656559,
+ "learning_rate": 0.0006,
+ "loss": 4.05365514755249,
+ "step": 6679
+ },
+ {
+ "epoch": 92.78253275109171,
+ "grad_norm": 0.015028242021799088,
+ "learning_rate": 0.0006,
+ "loss": 4.029950141906738,
+ "step": 6680
+ },
+ {
+ "epoch": 92.79650655021834,
+ "grad_norm": 0.014366233721375465,
+ "learning_rate": 0.0006,
+ "loss": 4.181772232055664,
+ "step": 6681
+ },
+ {
+ "epoch": 92.81048034934499,
+ "grad_norm": 0.014313179068267345,
+ "learning_rate": 0.0006,
+ "loss": 4.152658462524414,
+ "step": 6682
+ },
+ {
+ "epoch": 92.82445414847162,
+ "grad_norm": 0.016504302620887756,
+ "learning_rate": 0.0006,
+ "loss": 4.080532073974609,
+ "step": 6683
+ },
+ {
+ "epoch": 92.83842794759825,
+ "grad_norm": 0.01685677096247673,
+ "learning_rate": 0.0006,
+ "loss": 4.194268226623535,
+ "step": 6684
+ },
+ {
+ "epoch": 92.8524017467249,
+ "grad_norm": 0.018684150651097298,
+ "learning_rate": 0.0006,
+ "loss": 4.218497276306152,
+ "step": 6685
+ },
+ {
+ "epoch": 92.86637554585153,
+ "grad_norm": 0.01880732551217079,
+ "learning_rate": 0.0006,
+ "loss": 4.133942604064941,
+ "step": 6686
+ },
+ {
+ "epoch": 92.88034934497817,
+ "grad_norm": 0.016504917293787003,
+ "learning_rate": 0.0006,
+ "loss": 4.139945030212402,
+ "step": 6687
+ },
+ {
+ "epoch": 92.8943231441048,
+ "grad_norm": 0.016495356336236,
+ "learning_rate": 0.0006,
+ "loss": 4.098918914794922,
+ "step": 6688
+ },
+ {
+ "epoch": 92.90829694323143,
+ "grad_norm": 0.0170587245374918,
+ "learning_rate": 0.0006,
+ "loss": 4.025818824768066,
+ "step": 6689
+ },
+ {
+ "epoch": 92.92227074235808,
+ "grad_norm": 0.01718820631504059,
+ "learning_rate": 0.0006,
+ "loss": 4.125217437744141,
+ "step": 6690
+ },
+ {
+ "epoch": 92.93624454148471,
+ "grad_norm": 0.015789354220032692,
+ "learning_rate": 0.0006,
+ "loss": 4.150068283081055,
+ "step": 6691
+ },
+ {
+ "epoch": 92.95021834061136,
+ "grad_norm": 0.01587480492889881,
+ "learning_rate": 0.0006,
+ "loss": 4.194794654846191,
+ "step": 6692
+ },
+ {
+ "epoch": 92.96419213973799,
+ "grad_norm": 0.015669209882616997,
+ "learning_rate": 0.0006,
+ "loss": 4.0550336837768555,
+ "step": 6693
+ },
+ {
+ "epoch": 92.97816593886463,
+ "grad_norm": 0.01519355084747076,
+ "learning_rate": 0.0006,
+ "loss": 4.114232063293457,
+ "step": 6694
+ },
+ {
+ "epoch": 92.99213973799127,
+ "grad_norm": 0.014596261084079742,
+ "learning_rate": 0.0006,
+ "loss": 3.992065906524658,
+ "step": 6695
+ },
+ {
+ "epoch": 93.0,
+ "grad_norm": 0.016378438100218773,
+ "learning_rate": 0.0006,
+ "loss": 4.093693733215332,
+ "step": 6696
+ },
+ {
+ "epoch": 93.0,
+ "eval_loss": 4.541905879974365,
+ "eval_runtime": 57.0177,
+ "eval_samples_per_second": 42.829,
+ "eval_steps_per_second": 1.35,
+ "step": 6696
+ },
+ {
+ "epoch": 93.01397379912663,
+ "grad_norm": 0.01593143492937088,
+ "learning_rate": 0.0006,
+ "loss": 4.052624225616455,
+ "step": 6697
+ },
+ {
+ "epoch": 93.02794759825328,
+ "grad_norm": 0.016086837276816368,
+ "learning_rate": 0.0006,
+ "loss": 4.130306243896484,
+ "step": 6698
+ },
+ {
+ "epoch": 93.04192139737991,
+ "grad_norm": 0.016004323959350586,
+ "learning_rate": 0.0006,
+ "loss": 3.965893268585205,
+ "step": 6699
+ },
+ {
+ "epoch": 93.05589519650655,
+ "grad_norm": 0.017667284235358238,
+ "learning_rate": 0.0006,
+ "loss": 4.095634460449219,
+ "step": 6700
+ },
+ {
+ "epoch": 93.06986899563319,
+ "grad_norm": 0.01949942484498024,
+ "learning_rate": 0.0006,
+ "loss": 3.966041326522827,
+ "step": 6701
+ },
+ {
+ "epoch": 93.08384279475983,
+ "grad_norm": 0.021898971870541573,
+ "learning_rate": 0.0006,
+ "loss": 4.010378837585449,
+ "step": 6702
+ },
+ {
+ "epoch": 93.09781659388646,
+ "grad_norm": 0.019985264167189598,
+ "learning_rate": 0.0006,
+ "loss": 4.061577320098877,
+ "step": 6703
+ },
+ {
+ "epoch": 93.1117903930131,
+ "grad_norm": 0.016470855101943016,
+ "learning_rate": 0.0006,
+ "loss": 4.0601677894592285,
+ "step": 6704
+ },
+ {
+ "epoch": 93.12576419213974,
+ "grad_norm": 0.015600379556417465,
+ "learning_rate": 0.0006,
+ "loss": 4.0860700607299805,
+ "step": 6705
+ },
+ {
+ "epoch": 93.13973799126637,
+ "grad_norm": 0.017514333128929138,
+ "learning_rate": 0.0006,
+ "loss": 4.240778923034668,
+ "step": 6706
+ },
+ {
+ "epoch": 93.15371179039302,
+ "grad_norm": 0.01780967228114605,
+ "learning_rate": 0.0006,
+ "loss": 4.076499938964844,
+ "step": 6707
+ },
+ {
+ "epoch": 93.16768558951965,
+ "grad_norm": 0.016370438039302826,
+ "learning_rate": 0.0006,
+ "loss": 4.0677618980407715,
+ "step": 6708
+ },
+ {
+ "epoch": 93.18165938864628,
+ "grad_norm": 0.015205491334199905,
+ "learning_rate": 0.0006,
+ "loss": 4.004945755004883,
+ "step": 6709
+ },
+ {
+ "epoch": 93.19563318777293,
+ "grad_norm": 0.01537297572940588,
+ "learning_rate": 0.0006,
+ "loss": 4.121147155761719,
+ "step": 6710
+ },
+ {
+ "epoch": 93.20960698689956,
+ "grad_norm": 0.017343711107969284,
+ "learning_rate": 0.0006,
+ "loss": 4.103215217590332,
+ "step": 6711
+ },
+ {
+ "epoch": 93.2235807860262,
+ "grad_norm": 0.01618131436407566,
+ "learning_rate": 0.0006,
+ "loss": 4.083741664886475,
+ "step": 6712
+ },
+ {
+ "epoch": 93.23755458515284,
+ "grad_norm": 0.013340235687792301,
+ "learning_rate": 0.0006,
+ "loss": 4.055032730102539,
+ "step": 6713
+ },
+ {
+ "epoch": 93.25152838427948,
+ "grad_norm": 0.014084120281040668,
+ "learning_rate": 0.0006,
+ "loss": 4.005610942840576,
+ "step": 6714
+ },
+ {
+ "epoch": 93.26550218340611,
+ "grad_norm": 0.016104530543088913,
+ "learning_rate": 0.0006,
+ "loss": 4.091764450073242,
+ "step": 6715
+ },
+ {
+ "epoch": 93.27947598253274,
+ "grad_norm": 0.015232563950121403,
+ "learning_rate": 0.0006,
+ "loss": 4.105735778808594,
+ "step": 6716
+ },
+ {
+ "epoch": 93.29344978165939,
+ "grad_norm": 0.014003508724272251,
+ "learning_rate": 0.0006,
+ "loss": 4.05783224105835,
+ "step": 6717
+ },
+ {
+ "epoch": 93.30742358078602,
+ "grad_norm": 0.01614833064377308,
+ "learning_rate": 0.0006,
+ "loss": 4.124567031860352,
+ "step": 6718
+ },
+ {
+ "epoch": 93.32139737991267,
+ "grad_norm": 0.015010425820946693,
+ "learning_rate": 0.0006,
+ "loss": 4.07656192779541,
+ "step": 6719
+ },
+ {
+ "epoch": 93.3353711790393,
+ "grad_norm": 0.01444012951105833,
+ "learning_rate": 0.0006,
+ "loss": 4.122415542602539,
+ "step": 6720
+ },
+ {
+ "epoch": 93.34934497816595,
+ "grad_norm": 0.015715785324573517,
+ "learning_rate": 0.0006,
+ "loss": 4.070283889770508,
+ "step": 6721
+ },
+ {
+ "epoch": 93.36331877729258,
+ "grad_norm": 0.014976450242102146,
+ "learning_rate": 0.0006,
+ "loss": 4.087043762207031,
+ "step": 6722
+ },
+ {
+ "epoch": 93.37729257641921,
+ "grad_norm": 0.014868560247123241,
+ "learning_rate": 0.0006,
+ "loss": 3.9671802520751953,
+ "step": 6723
+ },
+ {
+ "epoch": 93.39126637554585,
+ "grad_norm": 0.013518013060092926,
+ "learning_rate": 0.0006,
+ "loss": 4.136781692504883,
+ "step": 6724
+ },
+ {
+ "epoch": 93.40524017467249,
+ "grad_norm": 0.013498986139893532,
+ "learning_rate": 0.0006,
+ "loss": 3.982309103012085,
+ "step": 6725
+ },
+ {
+ "epoch": 93.41921397379913,
+ "grad_norm": 0.013707910664379597,
+ "learning_rate": 0.0006,
+ "loss": 4.181885719299316,
+ "step": 6726
+ },
+ {
+ "epoch": 93.43318777292576,
+ "grad_norm": 0.013913207687437534,
+ "learning_rate": 0.0006,
+ "loss": 3.980070114135742,
+ "step": 6727
+ },
+ {
+ "epoch": 93.44716157205241,
+ "grad_norm": 0.013169893994927406,
+ "learning_rate": 0.0006,
+ "loss": 4.023380279541016,
+ "step": 6728
+ },
+ {
+ "epoch": 93.46113537117904,
+ "grad_norm": 0.014451324008405209,
+ "learning_rate": 0.0006,
+ "loss": 4.038372039794922,
+ "step": 6729
+ },
+ {
+ "epoch": 93.47510917030567,
+ "grad_norm": 0.014401541091501713,
+ "learning_rate": 0.0006,
+ "loss": 4.069009304046631,
+ "step": 6730
+ },
+ {
+ "epoch": 93.48908296943232,
+ "grad_norm": 0.013971361331641674,
+ "learning_rate": 0.0006,
+ "loss": 4.105923652648926,
+ "step": 6731
+ },
+ {
+ "epoch": 93.50305676855895,
+ "grad_norm": 0.013192536309361458,
+ "learning_rate": 0.0006,
+ "loss": 4.07828950881958,
+ "step": 6732
+ },
+ {
+ "epoch": 93.5170305676856,
+ "grad_norm": 0.013581741601228714,
+ "learning_rate": 0.0006,
+ "loss": 4.079203128814697,
+ "step": 6733
+ },
+ {
+ "epoch": 93.53100436681223,
+ "grad_norm": 0.01508194301277399,
+ "learning_rate": 0.0006,
+ "loss": 4.084749221801758,
+ "step": 6734
+ },
+ {
+ "epoch": 93.54497816593886,
+ "grad_norm": 0.0156723503023386,
+ "learning_rate": 0.0006,
+ "loss": 4.070216178894043,
+ "step": 6735
+ },
+ {
+ "epoch": 93.5589519650655,
+ "grad_norm": 0.014996538870036602,
+ "learning_rate": 0.0006,
+ "loss": 4.111318588256836,
+ "step": 6736
+ },
+ {
+ "epoch": 93.57292576419214,
+ "grad_norm": 0.012815269641578197,
+ "learning_rate": 0.0006,
+ "loss": 4.130414962768555,
+ "step": 6737
+ },
+ {
+ "epoch": 93.58689956331878,
+ "grad_norm": 0.012887579388916492,
+ "learning_rate": 0.0006,
+ "loss": 4.128446102142334,
+ "step": 6738
+ },
+ {
+ "epoch": 93.60087336244541,
+ "grad_norm": 0.013632671907544136,
+ "learning_rate": 0.0006,
+ "loss": 4.20977258682251,
+ "step": 6739
+ },
+ {
+ "epoch": 93.61484716157206,
+ "grad_norm": 0.014318895526230335,
+ "learning_rate": 0.0006,
+ "loss": 4.005739212036133,
+ "step": 6740
+ },
+ {
+ "epoch": 93.62882096069869,
+ "grad_norm": 0.014160127379000187,
+ "learning_rate": 0.0006,
+ "loss": 4.122544288635254,
+ "step": 6741
+ },
+ {
+ "epoch": 93.64279475982532,
+ "grad_norm": 0.01308423187583685,
+ "learning_rate": 0.0006,
+ "loss": 4.108774662017822,
+ "step": 6742
+ },
+ {
+ "epoch": 93.65676855895197,
+ "grad_norm": 0.014293737709522247,
+ "learning_rate": 0.0006,
+ "loss": 3.983156681060791,
+ "step": 6743
+ },
+ {
+ "epoch": 93.6707423580786,
+ "grad_norm": 0.015155135653913021,
+ "learning_rate": 0.0006,
+ "loss": 4.006117820739746,
+ "step": 6744
+ },
+ {
+ "epoch": 93.68471615720524,
+ "grad_norm": 0.014391839504241943,
+ "learning_rate": 0.0006,
+ "loss": 4.1658935546875,
+ "step": 6745
+ },
+ {
+ "epoch": 93.69868995633188,
+ "grad_norm": 0.014924845658242702,
+ "learning_rate": 0.0006,
+ "loss": 4.090868949890137,
+ "step": 6746
+ },
+ {
+ "epoch": 93.71266375545852,
+ "grad_norm": 0.014786627143621445,
+ "learning_rate": 0.0006,
+ "loss": 4.121716499328613,
+ "step": 6747
+ },
+ {
+ "epoch": 93.72663755458515,
+ "grad_norm": 0.016948824748396873,
+ "learning_rate": 0.0006,
+ "loss": 4.213179111480713,
+ "step": 6748
+ },
+ {
+ "epoch": 93.74061135371178,
+ "grad_norm": 0.01732620596885681,
+ "learning_rate": 0.0006,
+ "loss": 4.079215049743652,
+ "step": 6749
+ },
+ {
+ "epoch": 93.75458515283843,
+ "grad_norm": 0.016298124566674232,
+ "learning_rate": 0.0006,
+ "loss": 4.096126079559326,
+ "step": 6750
+ },
+ {
+ "epoch": 93.76855895196506,
+ "grad_norm": 0.015744732692837715,
+ "learning_rate": 0.0006,
+ "loss": 4.046300411224365,
+ "step": 6751
+ },
+ {
+ "epoch": 93.78253275109171,
+ "grad_norm": 0.016330812126398087,
+ "learning_rate": 0.0006,
+ "loss": 4.098011016845703,
+ "step": 6752
+ },
+ {
+ "epoch": 93.79650655021834,
+ "grad_norm": 0.01710600219666958,
+ "learning_rate": 0.0006,
+ "loss": 4.031402587890625,
+ "step": 6753
+ },
+ {
+ "epoch": 93.81048034934499,
+ "grad_norm": 0.01561805047094822,
+ "learning_rate": 0.0006,
+ "loss": 4.023388862609863,
+ "step": 6754
+ },
+ {
+ "epoch": 93.82445414847162,
+ "grad_norm": 0.013637243770062923,
+ "learning_rate": 0.0006,
+ "loss": 4.143238067626953,
+ "step": 6755
+ },
+ {
+ "epoch": 93.83842794759825,
+ "grad_norm": 0.013088075444102287,
+ "learning_rate": 0.0006,
+ "loss": 3.9083704948425293,
+ "step": 6756
+ },
+ {
+ "epoch": 93.8524017467249,
+ "grad_norm": 0.015893537551164627,
+ "learning_rate": 0.0006,
+ "loss": 4.102263450622559,
+ "step": 6757
+ },
+ {
+ "epoch": 93.86637554585153,
+ "grad_norm": 0.015213928185403347,
+ "learning_rate": 0.0006,
+ "loss": 4.111978530883789,
+ "step": 6758
+ },
+ {
+ "epoch": 93.88034934497817,
+ "grad_norm": 0.014543982222676277,
+ "learning_rate": 0.0006,
+ "loss": 4.05889892578125,
+ "step": 6759
+ },
+ {
+ "epoch": 93.8943231441048,
+ "grad_norm": 0.015104752965271473,
+ "learning_rate": 0.0006,
+ "loss": 4.136166572570801,
+ "step": 6760
+ },
+ {
+ "epoch": 93.90829694323143,
+ "grad_norm": 0.01505881268531084,
+ "learning_rate": 0.0006,
+ "loss": 4.061883449554443,
+ "step": 6761
+ },
+ {
+ "epoch": 93.92227074235808,
+ "grad_norm": 0.015248455107212067,
+ "learning_rate": 0.0006,
+ "loss": 4.109863758087158,
+ "step": 6762
+ },
+ {
+ "epoch": 93.93624454148471,
+ "grad_norm": 0.013651185669004917,
+ "learning_rate": 0.0006,
+ "loss": 4.061549663543701,
+ "step": 6763
+ },
+ {
+ "epoch": 93.95021834061136,
+ "grad_norm": 0.013806282542645931,
+ "learning_rate": 0.0006,
+ "loss": 4.205403804779053,
+ "step": 6764
+ },
+ {
+ "epoch": 93.96419213973799,
+ "grad_norm": 0.01538451761007309,
+ "learning_rate": 0.0006,
+ "loss": 4.14816951751709,
+ "step": 6765
+ },
+ {
+ "epoch": 93.97816593886463,
+ "grad_norm": 0.01602427288889885,
+ "learning_rate": 0.0006,
+ "loss": 4.000303268432617,
+ "step": 6766
+ },
+ {
+ "epoch": 93.99213973799127,
+ "grad_norm": 0.014558793976902962,
+ "learning_rate": 0.0006,
+ "loss": 4.096562385559082,
+ "step": 6767
+ },
+ {
+ "epoch": 94.0,
+ "grad_norm": 0.01613018475472927,
+ "learning_rate": 0.0006,
+ "loss": 4.262835502624512,
+ "step": 6768
+ },
+ {
+ "epoch": 94.0,
+ "eval_loss": 4.532126426696777,
+ "eval_runtime": 56.7592,
+ "eval_samples_per_second": 43.024,
+ "eval_steps_per_second": 1.357,
+ "step": 6768
+ },
+ {
+ "epoch": 94.01397379912663,
+ "grad_norm": 0.01614535041153431,
+ "learning_rate": 0.0006,
+ "loss": 4.073431491851807,
+ "step": 6769
+ },
+ {
+ "epoch": 94.02794759825328,
+ "grad_norm": 0.015986818820238113,
+ "learning_rate": 0.0006,
+ "loss": 4.070001602172852,
+ "step": 6770
+ },
+ {
+ "epoch": 94.04192139737991,
+ "grad_norm": 0.015678714960813522,
+ "learning_rate": 0.0006,
+ "loss": 4.06920051574707,
+ "step": 6771
+ },
+ {
+ "epoch": 94.05589519650655,
+ "grad_norm": 0.016593851149082184,
+ "learning_rate": 0.0006,
+ "loss": 4.095780849456787,
+ "step": 6772
+ },
+ {
+ "epoch": 94.06986899563319,
+ "grad_norm": 0.01780948042869568,
+ "learning_rate": 0.0006,
+ "loss": 4.092916488647461,
+ "step": 6773
+ },
+ {
+ "epoch": 94.08384279475983,
+ "grad_norm": 0.01755528897047043,
+ "learning_rate": 0.0006,
+ "loss": 3.9833455085754395,
+ "step": 6774
+ },
+ {
+ "epoch": 94.09781659388646,
+ "grad_norm": 0.01708311215043068,
+ "learning_rate": 0.0006,
+ "loss": 4.053336143493652,
+ "step": 6775
+ },
+ {
+ "epoch": 94.1117903930131,
+ "grad_norm": 0.017362922430038452,
+ "learning_rate": 0.0006,
+ "loss": 4.014343738555908,
+ "step": 6776
+ },
+ {
+ "epoch": 94.12576419213974,
+ "grad_norm": 0.01824064739048481,
+ "learning_rate": 0.0006,
+ "loss": 4.1215691566467285,
+ "step": 6777
+ },
+ {
+ "epoch": 94.13973799126637,
+ "grad_norm": 0.017743906006217003,
+ "learning_rate": 0.0006,
+ "loss": 4.091215133666992,
+ "step": 6778
+ },
+ {
+ "epoch": 94.15371179039302,
+ "grad_norm": 0.016585713252425194,
+ "learning_rate": 0.0006,
+ "loss": 4.013662338256836,
+ "step": 6779
+ },
+ {
+ "epoch": 94.16768558951965,
+ "grad_norm": 0.017397213727235794,
+ "learning_rate": 0.0006,
+ "loss": 4.070602893829346,
+ "step": 6780
+ },
+ {
+ "epoch": 94.18165938864628,
+ "grad_norm": 0.01733497530221939,
+ "learning_rate": 0.0006,
+ "loss": 4.1304426193237305,
+ "step": 6781
+ },
+ {
+ "epoch": 94.19563318777293,
+ "grad_norm": 0.015215300023555756,
+ "learning_rate": 0.0006,
+ "loss": 3.965014696121216,
+ "step": 6782
+ },
+ {
+ "epoch": 94.20960698689956,
+ "grad_norm": 0.015908699482679367,
+ "learning_rate": 0.0006,
+ "loss": 3.924964427947998,
+ "step": 6783
+ },
+ {
+ "epoch": 94.2235807860262,
+ "grad_norm": 0.014044968411326408,
+ "learning_rate": 0.0006,
+ "loss": 4.08673095703125,
+ "step": 6784
+ },
+ {
+ "epoch": 94.23755458515284,
+ "grad_norm": 0.01375142764300108,
+ "learning_rate": 0.0006,
+ "loss": 4.011466979980469,
+ "step": 6785
+ },
+ {
+ "epoch": 94.25152838427948,
+ "grad_norm": 0.015539257787168026,
+ "learning_rate": 0.0006,
+ "loss": 4.09976863861084,
+ "step": 6786
+ },
+ {
+ "epoch": 94.26550218340611,
+ "grad_norm": 0.014303367584943771,
+ "learning_rate": 0.0006,
+ "loss": 4.108369827270508,
+ "step": 6787
+ },
+ {
+ "epoch": 94.27947598253274,
+ "grad_norm": 0.01281928364187479,
+ "learning_rate": 0.0006,
+ "loss": 4.073737144470215,
+ "step": 6788
+ },
+ {
+ "epoch": 94.29344978165939,
+ "grad_norm": 0.013859959319233894,
+ "learning_rate": 0.0006,
+ "loss": 4.054823875427246,
+ "step": 6789
+ },
+ {
+ "epoch": 94.30742358078602,
+ "grad_norm": 0.013326860964298248,
+ "learning_rate": 0.0006,
+ "loss": 4.133251190185547,
+ "step": 6790
+ },
+ {
+ "epoch": 94.32139737991267,
+ "grad_norm": 0.013923076912760735,
+ "learning_rate": 0.0006,
+ "loss": 4.190706729888916,
+ "step": 6791
+ },
+ {
+ "epoch": 94.3353711790393,
+ "grad_norm": 0.01420847699046135,
+ "learning_rate": 0.0006,
+ "loss": 3.9336605072021484,
+ "step": 6792
+ },
+ {
+ "epoch": 94.34934497816595,
+ "grad_norm": 0.014607555232942104,
+ "learning_rate": 0.0006,
+ "loss": 4.128399848937988,
+ "step": 6793
+ },
+ {
+ "epoch": 94.36331877729258,
+ "grad_norm": 0.012862961739301682,
+ "learning_rate": 0.0006,
+ "loss": 4.0132527351379395,
+ "step": 6794
+ },
+ {
+ "epoch": 94.37729257641921,
+ "grad_norm": 0.012530217878520489,
+ "learning_rate": 0.0006,
+ "loss": 4.0384368896484375,
+ "step": 6795
+ },
+ {
+ "epoch": 94.39126637554585,
+ "grad_norm": 0.01286802627146244,
+ "learning_rate": 0.0006,
+ "loss": 3.95570707321167,
+ "step": 6796
+ },
+ {
+ "epoch": 94.40524017467249,
+ "grad_norm": 0.013450917787849903,
+ "learning_rate": 0.0006,
+ "loss": 4.016484260559082,
+ "step": 6797
+ },
+ {
+ "epoch": 94.41921397379913,
+ "grad_norm": 0.01393423043191433,
+ "learning_rate": 0.0006,
+ "loss": 4.124959468841553,
+ "step": 6798
+ },
+ {
+ "epoch": 94.43318777292576,
+ "grad_norm": 0.013077184557914734,
+ "learning_rate": 0.0006,
+ "loss": 4.029163837432861,
+ "step": 6799
+ },
+ {
+ "epoch": 94.44716157205241,
+ "grad_norm": 0.013577334582805634,
+ "learning_rate": 0.0006,
+ "loss": 4.22929573059082,
+ "step": 6800
+ },
+ {
+ "epoch": 94.46113537117904,
+ "grad_norm": 0.013930312357842922,
+ "learning_rate": 0.0006,
+ "loss": 4.024711608886719,
+ "step": 6801
+ },
+ {
+ "epoch": 94.47510917030567,
+ "grad_norm": 0.013746250420808792,
+ "learning_rate": 0.0006,
+ "loss": 3.9766428470611572,
+ "step": 6802
+ },
+ {
+ "epoch": 94.48908296943232,
+ "grad_norm": 0.013682983815670013,
+ "learning_rate": 0.0006,
+ "loss": 4.072981834411621,
+ "step": 6803
+ },
+ {
+ "epoch": 94.50305676855895,
+ "grad_norm": 0.01473842654377222,
+ "learning_rate": 0.0006,
+ "loss": 4.197623252868652,
+ "step": 6804
+ },
+ {
+ "epoch": 94.5170305676856,
+ "grad_norm": 0.014133758842945099,
+ "learning_rate": 0.0006,
+ "loss": 4.055549621582031,
+ "step": 6805
+ },
+ {
+ "epoch": 94.53100436681223,
+ "grad_norm": 0.0139645179733634,
+ "learning_rate": 0.0006,
+ "loss": 4.130135536193848,
+ "step": 6806
+ },
+ {
+ "epoch": 94.54497816593886,
+ "grad_norm": 0.013973365537822247,
+ "learning_rate": 0.0006,
+ "loss": 4.006230354309082,
+ "step": 6807
+ },
+ {
+ "epoch": 94.5589519650655,
+ "grad_norm": 0.014227126725018024,
+ "learning_rate": 0.0006,
+ "loss": 4.130459785461426,
+ "step": 6808
+ },
+ {
+ "epoch": 94.57292576419214,
+ "grad_norm": 0.013129506260156631,
+ "learning_rate": 0.0006,
+ "loss": 4.014974594116211,
+ "step": 6809
+ },
+ {
+ "epoch": 94.58689956331878,
+ "grad_norm": 0.013015891425311565,
+ "learning_rate": 0.0006,
+ "loss": 4.161370277404785,
+ "step": 6810
+ },
+ {
+ "epoch": 94.60087336244541,
+ "grad_norm": 0.013957147486507893,
+ "learning_rate": 0.0006,
+ "loss": 4.09519100189209,
+ "step": 6811
+ },
+ {
+ "epoch": 94.61484716157206,
+ "grad_norm": 0.015494572930037975,
+ "learning_rate": 0.0006,
+ "loss": 4.029949188232422,
+ "step": 6812
+ },
+ {
+ "epoch": 94.62882096069869,
+ "grad_norm": 0.016732869669795036,
+ "learning_rate": 0.0006,
+ "loss": 4.135669708251953,
+ "step": 6813
+ },
+ {
+ "epoch": 94.64279475982532,
+ "grad_norm": 0.01769322343170643,
+ "learning_rate": 0.0006,
+ "loss": 4.109872817993164,
+ "step": 6814
+ },
+ {
+ "epoch": 94.65676855895197,
+ "grad_norm": 0.015932537615299225,
+ "learning_rate": 0.0006,
+ "loss": 4.014595031738281,
+ "step": 6815
+ },
+ {
+ "epoch": 94.6707423580786,
+ "grad_norm": 0.015152990818023682,
+ "learning_rate": 0.0006,
+ "loss": 4.1983442306518555,
+ "step": 6816
+ },
+ {
+ "epoch": 94.68471615720524,
+ "grad_norm": 0.01590685173869133,
+ "learning_rate": 0.0006,
+ "loss": 4.119250774383545,
+ "step": 6817
+ },
+ {
+ "epoch": 94.69868995633188,
+ "grad_norm": 0.016018465161323547,
+ "learning_rate": 0.0006,
+ "loss": 4.048069953918457,
+ "step": 6818
+ },
+ {
+ "epoch": 94.71266375545852,
+ "grad_norm": 0.016621652990579605,
+ "learning_rate": 0.0006,
+ "loss": 4.043731689453125,
+ "step": 6819
+ },
+ {
+ "epoch": 94.72663755458515,
+ "grad_norm": 0.014535024762153625,
+ "learning_rate": 0.0006,
+ "loss": 4.1886820793151855,
+ "step": 6820
+ },
+ {
+ "epoch": 94.74061135371178,
+ "grad_norm": 0.014192107133567333,
+ "learning_rate": 0.0006,
+ "loss": 4.103938102722168,
+ "step": 6821
+ },
+ {
+ "epoch": 94.75458515283843,
+ "grad_norm": 0.0166022852063179,
+ "learning_rate": 0.0006,
+ "loss": 4.017818450927734,
+ "step": 6822
+ },
+ {
+ "epoch": 94.76855895196506,
+ "grad_norm": 0.016247155144810677,
+ "learning_rate": 0.0006,
+ "loss": 4.0038981437683105,
+ "step": 6823
+ },
+ {
+ "epoch": 94.78253275109171,
+ "grad_norm": 0.01593470573425293,
+ "learning_rate": 0.0006,
+ "loss": 4.043969631195068,
+ "step": 6824
+ },
+ {
+ "epoch": 94.79650655021834,
+ "grad_norm": 0.014334345236420631,
+ "learning_rate": 0.0006,
+ "loss": 4.097829341888428,
+ "step": 6825
+ },
+ {
+ "epoch": 94.81048034934499,
+ "grad_norm": 0.015581986866891384,
+ "learning_rate": 0.0006,
+ "loss": 4.148984432220459,
+ "step": 6826
+ },
+ {
+ "epoch": 94.82445414847162,
+ "grad_norm": 0.017519738525152206,
+ "learning_rate": 0.0006,
+ "loss": 4.245623588562012,
+ "step": 6827
+ },
+ {
+ "epoch": 94.83842794759825,
+ "grad_norm": 0.015155277214944363,
+ "learning_rate": 0.0006,
+ "loss": 4.031019687652588,
+ "step": 6828
+ },
+ {
+ "epoch": 94.8524017467249,
+ "grad_norm": 0.014969222247600555,
+ "learning_rate": 0.0006,
+ "loss": 4.036084175109863,
+ "step": 6829
+ },
+ {
+ "epoch": 94.86637554585153,
+ "grad_norm": 0.015987908467650414,
+ "learning_rate": 0.0006,
+ "loss": 4.08396053314209,
+ "step": 6830
+ },
+ {
+ "epoch": 94.88034934497817,
+ "grad_norm": 0.016324151307344437,
+ "learning_rate": 0.0006,
+ "loss": 4.036787033081055,
+ "step": 6831
+ },
+ {
+ "epoch": 94.8943231441048,
+ "grad_norm": 0.015957487747073174,
+ "learning_rate": 0.0006,
+ "loss": 4.147607803344727,
+ "step": 6832
+ },
+ {
+ "epoch": 94.90829694323143,
+ "grad_norm": 0.015286967158317566,
+ "learning_rate": 0.0006,
+ "loss": 4.184615135192871,
+ "step": 6833
+ },
+ {
+ "epoch": 94.92227074235808,
+ "grad_norm": 0.01338377594947815,
+ "learning_rate": 0.0006,
+ "loss": 4.014981269836426,
+ "step": 6834
+ },
+ {
+ "epoch": 94.93624454148471,
+ "grad_norm": 0.013997633941471577,
+ "learning_rate": 0.0006,
+ "loss": 4.021933078765869,
+ "step": 6835
+ },
+ {
+ "epoch": 94.95021834061136,
+ "grad_norm": 0.014572402462363243,
+ "learning_rate": 0.0006,
+ "loss": 4.0326948165893555,
+ "step": 6836
+ },
+ {
+ "epoch": 94.96419213973799,
+ "grad_norm": 0.014379586093127728,
+ "learning_rate": 0.0006,
+ "loss": 4.060697555541992,
+ "step": 6837
+ },
+ {
+ "epoch": 94.97816593886463,
+ "grad_norm": 0.013513042591512203,
+ "learning_rate": 0.0006,
+ "loss": 4.05123233795166,
+ "step": 6838
+ },
+ {
+ "epoch": 94.99213973799127,
+ "grad_norm": 0.012904108501970768,
+ "learning_rate": 0.0006,
+ "loss": 4.020852565765381,
+ "step": 6839
+ },
+ {
+ "epoch": 95.0,
+ "grad_norm": 0.014364209026098251,
+ "learning_rate": 0.0006,
+ "loss": 4.080837249755859,
+ "step": 6840
+ },
+ {
+ "epoch": 95.0,
+ "eval_loss": 4.608776569366455,
+ "eval_runtime": 56.6516,
+ "eval_samples_per_second": 43.106,
+ "eval_steps_per_second": 1.359,
+ "step": 6840
+ },
+ {
+ "epoch": 95.01397379912663,
+ "grad_norm": 0.01357510220259428,
+ "learning_rate": 0.0006,
+ "loss": 4.105312347412109,
+ "step": 6841
+ },
+ {
+ "epoch": 95.02794759825328,
+ "grad_norm": 0.014248257502913475,
+ "learning_rate": 0.0006,
+ "loss": 3.971667766571045,
+ "step": 6842
+ },
+ {
+ "epoch": 95.04192139737991,
+ "grad_norm": 0.014773483388125896,
+ "learning_rate": 0.0006,
+ "loss": 4.03553581237793,
+ "step": 6843
+ },
+ {
+ "epoch": 95.05589519650655,
+ "grad_norm": 0.014766111969947815,
+ "learning_rate": 0.0006,
+ "loss": 4.074707508087158,
+ "step": 6844
+ },
+ {
+ "epoch": 95.06986899563319,
+ "grad_norm": 0.014778018929064274,
+ "learning_rate": 0.0006,
+ "loss": 4.1083855628967285,
+ "step": 6845
+ },
+ {
+ "epoch": 95.08384279475983,
+ "grad_norm": 0.014466686174273491,
+ "learning_rate": 0.0006,
+ "loss": 4.054587364196777,
+ "step": 6846
+ },
+ {
+ "epoch": 95.09781659388646,
+ "grad_norm": 0.015195904299616814,
+ "learning_rate": 0.0006,
+ "loss": 4.142580032348633,
+ "step": 6847
+ },
+ {
+ "epoch": 95.1117903930131,
+ "grad_norm": 0.01530557032674551,
+ "learning_rate": 0.0006,
+ "loss": 4.062134265899658,
+ "step": 6848
+ },
+ {
+ "epoch": 95.12576419213974,
+ "grad_norm": 0.015203768387436867,
+ "learning_rate": 0.0006,
+ "loss": 4.082962989807129,
+ "step": 6849
+ },
+ {
+ "epoch": 95.13973799126637,
+ "grad_norm": 0.014906628057360649,
+ "learning_rate": 0.0006,
+ "loss": 4.033127784729004,
+ "step": 6850
+ },
+ {
+ "epoch": 95.15371179039302,
+ "grad_norm": 0.01553534809499979,
+ "learning_rate": 0.0006,
+ "loss": 4.051743030548096,
+ "step": 6851
+ },
+ {
+ "epoch": 95.16768558951965,
+ "grad_norm": 0.01607976108789444,
+ "learning_rate": 0.0006,
+ "loss": 4.038073539733887,
+ "step": 6852
+ },
+ {
+ "epoch": 95.18165938864628,
+ "grad_norm": 0.015776360407471657,
+ "learning_rate": 0.0006,
+ "loss": 4.013091087341309,
+ "step": 6853
+ },
+ {
+ "epoch": 95.19563318777293,
+ "grad_norm": 0.01623626798391342,
+ "learning_rate": 0.0006,
+ "loss": 4.12293815612793,
+ "step": 6854
+ },
+ {
+ "epoch": 95.20960698689956,
+ "grad_norm": 0.016887828707695007,
+ "learning_rate": 0.0006,
+ "loss": 4.120452404022217,
+ "step": 6855
+ },
+ {
+ "epoch": 95.2235807860262,
+ "grad_norm": 0.01561128068715334,
+ "learning_rate": 0.0006,
+ "loss": 4.071322441101074,
+ "step": 6856
+ },
+ {
+ "epoch": 95.23755458515284,
+ "grad_norm": 0.015679724514484406,
+ "learning_rate": 0.0006,
+ "loss": 4.065072059631348,
+ "step": 6857
+ },
+ {
+ "epoch": 95.25152838427948,
+ "grad_norm": 0.016506962478160858,
+ "learning_rate": 0.0006,
+ "loss": 4.0667924880981445,
+ "step": 6858
+ },
+ {
+ "epoch": 95.26550218340611,
+ "grad_norm": 0.015340839512646198,
+ "learning_rate": 0.0006,
+ "loss": 4.14047908782959,
+ "step": 6859
+ },
+ {
+ "epoch": 95.27947598253274,
+ "grad_norm": 0.013935413211584091,
+ "learning_rate": 0.0006,
+ "loss": 3.9688191413879395,
+ "step": 6860
+ },
+ {
+ "epoch": 95.29344978165939,
+ "grad_norm": 0.015223405323922634,
+ "learning_rate": 0.0006,
+ "loss": 4.101613521575928,
+ "step": 6861
+ },
+ {
+ "epoch": 95.30742358078602,
+ "grad_norm": 0.01562546379864216,
+ "learning_rate": 0.0006,
+ "loss": 4.128626823425293,
+ "step": 6862
+ },
+ {
+ "epoch": 95.32139737991267,
+ "grad_norm": 0.016159312799572945,
+ "learning_rate": 0.0006,
+ "loss": 4.080667018890381,
+ "step": 6863
+ },
+ {
+ "epoch": 95.3353711790393,
+ "grad_norm": 0.01724511943757534,
+ "learning_rate": 0.0006,
+ "loss": 4.08463191986084,
+ "step": 6864
+ },
+ {
+ "epoch": 95.34934497816595,
+ "grad_norm": 0.01592276245355606,
+ "learning_rate": 0.0006,
+ "loss": 4.127612113952637,
+ "step": 6865
+ },
+ {
+ "epoch": 95.36331877729258,
+ "grad_norm": 0.01744082197546959,
+ "learning_rate": 0.0006,
+ "loss": 4.034480094909668,
+ "step": 6866
+ },
+ {
+ "epoch": 95.37729257641921,
+ "grad_norm": 0.01774069108068943,
+ "learning_rate": 0.0006,
+ "loss": 4.091196537017822,
+ "step": 6867
+ },
+ {
+ "epoch": 95.39126637554585,
+ "grad_norm": 0.01653675176203251,
+ "learning_rate": 0.0006,
+ "loss": 4.032937526702881,
+ "step": 6868
+ },
+ {
+ "epoch": 95.40524017467249,
+ "grad_norm": 0.014933439902961254,
+ "learning_rate": 0.0006,
+ "loss": 3.9386508464813232,
+ "step": 6869
+ },
+ {
+ "epoch": 95.41921397379913,
+ "grad_norm": 0.014743257313966751,
+ "learning_rate": 0.0006,
+ "loss": 4.126462459564209,
+ "step": 6870
+ },
+ {
+ "epoch": 95.43318777292576,
+ "grad_norm": 0.014614643529057503,
+ "learning_rate": 0.0006,
+ "loss": 4.093838214874268,
+ "step": 6871
+ },
+ {
+ "epoch": 95.44716157205241,
+ "grad_norm": 0.01665084809064865,
+ "learning_rate": 0.0006,
+ "loss": 4.1471967697143555,
+ "step": 6872
+ },
+ {
+ "epoch": 95.46113537117904,
+ "grad_norm": 0.018056001514196396,
+ "learning_rate": 0.0006,
+ "loss": 4.1388349533081055,
+ "step": 6873
+ },
+ {
+ "epoch": 95.47510917030567,
+ "grad_norm": 0.02129282057285309,
+ "learning_rate": 0.0006,
+ "loss": 4.130726337432861,
+ "step": 6874
+ },
+ {
+ "epoch": 95.48908296943232,
+ "grad_norm": 0.022237276658415794,
+ "learning_rate": 0.0006,
+ "loss": 4.10750675201416,
+ "step": 6875
+ },
+ {
+ "epoch": 95.50305676855895,
+ "grad_norm": 0.021106960251927376,
+ "learning_rate": 0.0006,
+ "loss": 4.136451721191406,
+ "step": 6876
+ },
+ {
+ "epoch": 95.5170305676856,
+ "grad_norm": 0.019107872620224953,
+ "learning_rate": 0.0006,
+ "loss": 4.139644145965576,
+ "step": 6877
+ },
+ {
+ "epoch": 95.53100436681223,
+ "grad_norm": 0.018290145322680473,
+ "learning_rate": 0.0006,
+ "loss": 4.161884784698486,
+ "step": 6878
+ },
+ {
+ "epoch": 95.54497816593886,
+ "grad_norm": 0.018161306157708168,
+ "learning_rate": 0.0006,
+ "loss": 4.1252336502075195,
+ "step": 6879
+ },
+ {
+ "epoch": 95.5589519650655,
+ "grad_norm": 0.018432756885886192,
+ "learning_rate": 0.0006,
+ "loss": 4.0635986328125,
+ "step": 6880
+ },
+ {
+ "epoch": 95.57292576419214,
+ "grad_norm": 0.016592558473348618,
+ "learning_rate": 0.0006,
+ "loss": 4.037112236022949,
+ "step": 6881
+ },
+ {
+ "epoch": 95.58689956331878,
+ "grad_norm": 0.015283452346920967,
+ "learning_rate": 0.0006,
+ "loss": 4.1964826583862305,
+ "step": 6882
+ },
+ {
+ "epoch": 95.60087336244541,
+ "grad_norm": 0.015539180487394333,
+ "learning_rate": 0.0006,
+ "loss": 4.110917568206787,
+ "step": 6883
+ },
+ {
+ "epoch": 95.61484716157206,
+ "grad_norm": 0.016508113592863083,
+ "learning_rate": 0.0006,
+ "loss": 3.9705517292022705,
+ "step": 6884
+ },
+ {
+ "epoch": 95.62882096069869,
+ "grad_norm": 0.016635406762361526,
+ "learning_rate": 0.0006,
+ "loss": 4.025636196136475,
+ "step": 6885
+ },
+ {
+ "epoch": 95.64279475982532,
+ "grad_norm": 0.01688900962471962,
+ "learning_rate": 0.0006,
+ "loss": 4.01814079284668,
+ "step": 6886
+ },
+ {
+ "epoch": 95.65676855895197,
+ "grad_norm": 0.01614062674343586,
+ "learning_rate": 0.0006,
+ "loss": 4.098041534423828,
+ "step": 6887
+ },
+ {
+ "epoch": 95.6707423580786,
+ "grad_norm": 0.015679053962230682,
+ "learning_rate": 0.0006,
+ "loss": 4.077049255371094,
+ "step": 6888
+ },
+ {
+ "epoch": 95.68471615720524,
+ "grad_norm": 0.016617605462670326,
+ "learning_rate": 0.0006,
+ "loss": 4.10363245010376,
+ "step": 6889
+ },
+ {
+ "epoch": 95.69868995633188,
+ "grad_norm": 0.015966709703207016,
+ "learning_rate": 0.0006,
+ "loss": 4.100924491882324,
+ "step": 6890
+ },
+ {
+ "epoch": 95.71266375545852,
+ "grad_norm": 0.015768567100167274,
+ "learning_rate": 0.0006,
+ "loss": 4.063976287841797,
+ "step": 6891
+ },
+ {
+ "epoch": 95.72663755458515,
+ "grad_norm": 0.015800341963768005,
+ "learning_rate": 0.0006,
+ "loss": 4.008409023284912,
+ "step": 6892
+ },
+ {
+ "epoch": 95.74061135371178,
+ "grad_norm": 0.014288571663200855,
+ "learning_rate": 0.0006,
+ "loss": 4.139591217041016,
+ "step": 6893
+ },
+ {
+ "epoch": 95.75458515283843,
+ "grad_norm": 0.013743448071181774,
+ "learning_rate": 0.0006,
+ "loss": 3.967352867126465,
+ "step": 6894
+ },
+ {
+ "epoch": 95.76855895196506,
+ "grad_norm": 0.01387989055365324,
+ "learning_rate": 0.0006,
+ "loss": 4.127554893493652,
+ "step": 6895
+ },
+ {
+ "epoch": 95.78253275109171,
+ "grad_norm": 0.013910802081227303,
+ "learning_rate": 0.0006,
+ "loss": 4.130941390991211,
+ "step": 6896
+ },
+ {
+ "epoch": 95.79650655021834,
+ "grad_norm": 0.013936805538833141,
+ "learning_rate": 0.0006,
+ "loss": 4.138744831085205,
+ "step": 6897
+ },
+ {
+ "epoch": 95.81048034934499,
+ "grad_norm": 0.015135723166167736,
+ "learning_rate": 0.0006,
+ "loss": 4.131860256195068,
+ "step": 6898
+ },
+ {
+ "epoch": 95.82445414847162,
+ "grad_norm": 0.014921358786523342,
+ "learning_rate": 0.0006,
+ "loss": 4.1949462890625,
+ "step": 6899
+ },
+ {
+ "epoch": 95.83842794759825,
+ "grad_norm": 0.01401488296687603,
+ "learning_rate": 0.0006,
+ "loss": 4.181267738342285,
+ "step": 6900
+ },
+ {
+ "epoch": 95.8524017467249,
+ "grad_norm": 0.013691626489162445,
+ "learning_rate": 0.0006,
+ "loss": 4.113667964935303,
+ "step": 6901
+ },
+ {
+ "epoch": 95.86637554585153,
+ "grad_norm": 0.013890047557651997,
+ "learning_rate": 0.0006,
+ "loss": 4.163067817687988,
+ "step": 6902
+ },
+ {
+ "epoch": 95.88034934497817,
+ "grad_norm": 0.01385503076016903,
+ "learning_rate": 0.0006,
+ "loss": 4.076153755187988,
+ "step": 6903
+ },
+ {
+ "epoch": 95.8943231441048,
+ "grad_norm": 0.01383926346898079,
+ "learning_rate": 0.0006,
+ "loss": 4.102423667907715,
+ "step": 6904
+ },
+ {
+ "epoch": 95.90829694323143,
+ "grad_norm": 0.015961717814207077,
+ "learning_rate": 0.0006,
+ "loss": 4.1543779373168945,
+ "step": 6905
+ },
+ {
+ "epoch": 95.92227074235808,
+ "grad_norm": 0.016718363389372826,
+ "learning_rate": 0.0006,
+ "loss": 3.991607189178467,
+ "step": 6906
+ },
+ {
+ "epoch": 95.93624454148471,
+ "grad_norm": 0.01585792563855648,
+ "learning_rate": 0.0006,
+ "loss": 4.048587799072266,
+ "step": 6907
+ },
+ {
+ "epoch": 95.95021834061136,
+ "grad_norm": 0.015311713330447674,
+ "learning_rate": 0.0006,
+ "loss": 4.1070451736450195,
+ "step": 6908
+ },
+ {
+ "epoch": 95.96419213973799,
+ "grad_norm": 0.014521263539791107,
+ "learning_rate": 0.0006,
+ "loss": 4.087587356567383,
+ "step": 6909
+ },
+ {
+ "epoch": 95.97816593886463,
+ "grad_norm": 0.01433157455176115,
+ "learning_rate": 0.0006,
+ "loss": 4.139904499053955,
+ "step": 6910
+ },
+ {
+ "epoch": 95.99213973799127,
+ "grad_norm": 0.014475514180958271,
+ "learning_rate": 0.0006,
+ "loss": 4.182525634765625,
+ "step": 6911
+ },
+ {
+ "epoch": 96.0,
+ "grad_norm": 0.015245620161294937,
+ "learning_rate": 0.0006,
+ "loss": 4.1199870109558105,
+ "step": 6912
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 28800,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 400,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": true,
+ "should_log": false,
+ "should_save": false,
+ "should_training_stop": false
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 2.946828826878935e+18,
+ "train_batch_size": 8,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6912/training_args.bin b/runs/l2r50-i2-fulle-lm/checkpoint-6912/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..cad232831c222cc7f8b938b8bbe42f61e7d89c3a
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-6912/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:f7c40e3bfa2f5ac1450ca1243e3f2a895d44a06ed5995e6aefd97f78d229d8c8
+size 4792
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-72/chat_template.jinja b/runs/l2r50-i2-fulle-lm/checkpoint-72/chat_template.jinja
new file mode 100644
index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-72/chat_template.jinja
@@ -0,0 +1,85 @@
+{%- if tools %}
+ {{- '<|im_start|>system\n' }}
+ {%- if messages[0].role == 'system' %}
+ {{- messages[0].content + '\n\n' }}
+ {%- endif %}
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }}
+ {%- for tool in tools %}
+ {{- "\n" }}
+ {{- tool | tojson }}
+ {%- endfor %}
+ {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }}
+{%- else %}
+ {%- if messages[0].role == 'system' %}
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
+ {%- endif %}
+{%- endif %}
+{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
+{%- for message in messages[::-1] %}
+ {%- set index = (messages|length - 1) - loop.index0 %}
+ {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %}
+ {%- set ns.multi_step_tool = false %}
+ {%- set ns.last_query_index = index %}
+ {%- endif %}
+{%- endfor %}
+{%- for message in messages %}
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
+ {%- elif message.role == "assistant" %}
+ {%- set content = message.content %}
+ {%- set reasoning_content = '' %}
+ {%- if message.reasoning_content is defined and message.reasoning_content is not none %}
+ {%- set reasoning_content = message.reasoning_content %}
+ {%- else %}
+ {%- if '' in message.content %}
+ {%- set content = message.content.split('')[-1].lstrip('\n') %}
+ {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %}
+ {%- endif %}
+ {%- endif %}
+ {%- if loop.index0 > ns.last_query_index %}
+ {%- if loop.last or (not loop.last and reasoning_content) %}
+ {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- if message.tool_calls %}
+ {%- for tool_call in message.tool_calls %}
+ {%- if (loop.first and content) or (not loop.first) %}
+ {{- '\n' }}
+ {%- endif %}
+ {%- if tool_call.function %}
+ {%- set tool_call = tool_call.function %}
+ {%- endif %}
+ {{- '\n{"name": "' }}
+ {{- tool_call.name }}
+ {{- '", "arguments": ' }}
+ {%- if tool_call.arguments is string %}
+ {{- tool_call.arguments }}
+ {%- else %}
+ {{- tool_call.arguments | tojson }}
+ {%- endif %}
+ {{- '}\n' }}
+ {%- endfor %}
+ {%- endif %}
+ {{- '<|im_end|>\n' }}
+ {%- elif message.role == "tool" %}
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
+ {{- '<|im_start|>user' }}
+ {%- endif %}
+ {{- '\n\n' }}
+ {{- message.content }}
+ {{- '\n' }}
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
+ {{- '<|im_end|>\n' }}
+ {%- endif %}
+ {%- endif %}
+{%- endfor %}
+{%- if add_generation_prompt %}
+ {{- '<|im_start|>assistant\n' }}
+ {%- if enable_thinking is defined and enable_thinking is false %}
+ {{- '\n\n\n\n' }}
+ {%- endif %}
+{%- endif %}
\ No newline at end of file
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-72/config.json b/runs/l2r50-i2-fulle-lm/checkpoint-72/config.json
new file mode 100644
index 0000000000000000000000000000000000000000..22f250953c6ce8fc91ee4411856dd73041a415e2
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-72/config.json
@@ -0,0 +1,32 @@
+{
+ "architectures": [
+ "LlamaForCausalLM"
+ ],
+ "attention_bias": false,
+ "attention_dropout": 0.0,
+ "bos_token_id": null,
+ "dtype": "float32",
+ "eos_token_id": 151645,
+ "head_dim": 128,
+ "hidden_act": "silu",
+ "hidden_size": 512,
+ "initializer_range": 0.02,
+ "intermediate_size": 1536,
+ "max_position_embeddings": 2048,
+ "mlp_bias": false,
+ "model_type": "llama",
+ "num_attention_heads": 4,
+ "num_hidden_layers": 20,
+ "num_key_value_heads": 4,
+ "pad_token_id": 151645,
+ "pretraining_tp": 1,
+ "rms_norm_eps": 1e-06,
+ "rope_parameters": {
+ "rope_theta": 10000.0,
+ "rope_type": "default"
+ },
+ "tie_word_embeddings": true,
+ "transformers_version": "5.5.0",
+ "use_cache": false,
+ "vocab_size": 151673
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-72/generation_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-72/generation_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-72/generation_config.json
@@ -0,0 +1,11 @@
+{
+ "_from_model_config": true,
+ "eos_token_id": [
+ 151645
+ ],
+ "output_attentions": false,
+ "output_hidden_states": false,
+ "pad_token_id": 151645,
+ "transformers_version": "5.5.0",
+ "use_cache": true
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-72/model.safetensors b/runs/l2r50-i2-fulle-lm/checkpoint-72/model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..3cd56dfadca87c7ad5ef1d49ed08fdd77955d583
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-72/model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:2e92299b482a10bf724c8131c23f573f479a18d94a8335b3d22a1c327e7b4922
+size 583360328
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-72/optimizer.pt b/runs/l2r50-i2-fulle-lm/checkpoint-72/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..b7eeeb97cf14b488246532851ad85039b6f5910a
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-72/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:a049079dd03efbc2f0be2294231a6d522211af4dd67fad4631b8bbb7875df1c6
+size 1166833530
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-72/rng_state_0.pth b/runs/l2r50-i2-fulle-lm/checkpoint-72/rng_state_0.pth
new file mode 100644
index 0000000000000000000000000000000000000000..942953431b6b69250b3ac046cdf5cca437235031
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-72/rng_state_0.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:54c9019619b0417fc7c61a86a46149eb207b6bbd0181871dcec50cb454d3f367
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-72/rng_state_1.pth b/runs/l2r50-i2-fulle-lm/checkpoint-72/rng_state_1.pth
new file mode 100644
index 0000000000000000000000000000000000000000..ea26b0c2f5adce832b4d8762855e8be5382d63d7
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-72/rng_state_1.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:4dddd0ae188208a24e5ab9a9fc1928ad2f9275ce147a0a992232c03bba455215
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-72/rng_state_2.pth b/runs/l2r50-i2-fulle-lm/checkpoint-72/rng_state_2.pth
new file mode 100644
index 0000000000000000000000000000000000000000..409136377918f95867748fce9457de5970be4a6f
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-72/rng_state_2.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:8068dd390bcd57d32f1dfc5e792134a165f320ca31900b27252a73a53d7ec360
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-72/rng_state_3.pth b/runs/l2r50-i2-fulle-lm/checkpoint-72/rng_state_3.pth
new file mode 100644
index 0000000000000000000000000000000000000000..7eaf02283cfd4b5f41ba417c553b0f39b696450e
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-72/rng_state_3.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:6014b1c6c1718e10a75cbe0f7295e4d1975ef1aa122296fb2c3b54f4b1835c45
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-72/scheduler.pt b/runs/l2r50-i2-fulle-lm/checkpoint-72/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..43cf114590f1088fd438450226c42885d843da41
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-72/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:878ec53f5c2a524fc37af578b94a9025923faefea55f621be8e5643a79915d0f
+size 1064
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-72/tokenizer.json b/runs/l2r50-i2-fulle-lm/checkpoint-72/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..53786016f070e11184813f3c83101dc5e93e392e
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-72/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:c3c63fda1646a555448e8f03662542e04b11d103721319e4d9b267d7da03ceb8
+size 11424463
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-72/tokenizer_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-72/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..98e2fb9b2cc54785dc9d871b8e503f8b8f2defe7
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-72/tokenizer_config.json
@@ -0,0 +1,20 @@
+{
+ "add_prefix_space": false,
+ "backend": "tokenizers",
+ "bos_token": null,
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "<|im_end|>",
+ "errors": "replace",
+ "extra_special_tokens": [
+ "<|l2r_pred|>",
+ "<|r2l_pred|>",
+ "<|next_1_pred|>",
+ "<|next_2_pred|>"
+ ],
+ "is_local": false,
+ "model_max_length": 131072,
+ "pad_token": "<|im_end|>",
+ "split_special_tokens": false,
+ "tokenizer_class": "Qwen2Tokenizer",
+ "unk_token": null
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-72/trainer_state.json b/runs/l2r50-i2-fulle-lm/checkpoint-72/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..eee91baef0b1848c0b5219cb5ba7b8365d12789d
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-72/trainer_state.json
@@ -0,0 +1,538 @@
+{
+ "best_global_step": null,
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 1.0,
+ "eval_steps": 500,
+ "global_step": 72,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.013973799126637555,
+ "grad_norm": 0.1739506870508194,
+ "learning_rate": 0.0,
+ "loss": 11.99299430847168,
+ "step": 1
+ },
+ {
+ "epoch": 0.02794759825327511,
+ "grad_norm": 0.17216597497463226,
+ "learning_rate": 5.999999999999999e-06,
+ "loss": 11.988704681396484,
+ "step": 2
+ },
+ {
+ "epoch": 0.04192139737991266,
+ "grad_norm": 0.17459715902805328,
+ "learning_rate": 1.1999999999999999e-05,
+ "loss": 11.945581436157227,
+ "step": 3
+ },
+ {
+ "epoch": 0.05589519650655022,
+ "grad_norm": 0.17577414214611053,
+ "learning_rate": 1.7999999999999997e-05,
+ "loss": 11.859772682189941,
+ "step": 4
+ },
+ {
+ "epoch": 0.06986899563318777,
+ "grad_norm": 0.16163651645183563,
+ "learning_rate": 2.3999999999999997e-05,
+ "loss": 11.744415283203125,
+ "step": 5
+ },
+ {
+ "epoch": 0.08384279475982533,
+ "grad_norm": 0.14409932494163513,
+ "learning_rate": 2.9999999999999997e-05,
+ "loss": 11.618099212646484,
+ "step": 6
+ },
+ {
+ "epoch": 0.09781659388646288,
+ "grad_norm": 0.13079848885536194,
+ "learning_rate": 3.5999999999999994e-05,
+ "loss": 11.505935668945312,
+ "step": 7
+ },
+ {
+ "epoch": 0.11179039301310044,
+ "grad_norm": 0.11964613944292068,
+ "learning_rate": 4.2e-05,
+ "loss": 11.416532516479492,
+ "step": 8
+ },
+ {
+ "epoch": 0.125764192139738,
+ "grad_norm": 0.11345142871141434,
+ "learning_rate": 4.7999999999999994e-05,
+ "loss": 11.33206558227539,
+ "step": 9
+ },
+ {
+ "epoch": 0.13973799126637554,
+ "grad_norm": 0.11033952981233597,
+ "learning_rate": 5.399999999999999e-05,
+ "loss": 11.259052276611328,
+ "step": 10
+ },
+ {
+ "epoch": 0.1537117903930131,
+ "grad_norm": 0.10776489973068237,
+ "learning_rate": 5.9999999999999995e-05,
+ "loss": 11.203384399414062,
+ "step": 11
+ },
+ {
+ "epoch": 0.16768558951965065,
+ "grad_norm": 0.10721946507692337,
+ "learning_rate": 6.599999999999999e-05,
+ "loss": 11.153295516967773,
+ "step": 12
+ },
+ {
+ "epoch": 0.18165938864628822,
+ "grad_norm": 0.10673430562019348,
+ "learning_rate": 7.199999999999999e-05,
+ "loss": 11.110200881958008,
+ "step": 13
+ },
+ {
+ "epoch": 0.19563318777292577,
+ "grad_norm": 0.10630254447460175,
+ "learning_rate": 7.8e-05,
+ "loss": 11.067477226257324,
+ "step": 14
+ },
+ {
+ "epoch": 0.2096069868995633,
+ "grad_norm": 0.10631615668535233,
+ "learning_rate": 8.4e-05,
+ "loss": 11.022941589355469,
+ "step": 15
+ },
+ {
+ "epoch": 0.22358078602620088,
+ "grad_norm": 0.10609443485736847,
+ "learning_rate": 8.999999999999999e-05,
+ "loss": 10.970746040344238,
+ "step": 16
+ },
+ {
+ "epoch": 0.23755458515283842,
+ "grad_norm": 0.10648232698440552,
+ "learning_rate": 9.599999999999999e-05,
+ "loss": 10.91267204284668,
+ "step": 17
+ },
+ {
+ "epoch": 0.251528384279476,
+ "grad_norm": 0.10584589838981628,
+ "learning_rate": 0.000102,
+ "loss": 10.854547500610352,
+ "step": 18
+ },
+ {
+ "epoch": 0.26550218340611353,
+ "grad_norm": 0.10591720044612885,
+ "learning_rate": 0.00010799999999999998,
+ "loss": 10.788917541503906,
+ "step": 19
+ },
+ {
+ "epoch": 0.2794759825327511,
+ "grad_norm": 0.10611672699451447,
+ "learning_rate": 0.00011399999999999999,
+ "loss": 10.717556953430176,
+ "step": 20
+ },
+ {
+ "epoch": 0.2934497816593886,
+ "grad_norm": 0.10455359518527985,
+ "learning_rate": 0.00011999999999999999,
+ "loss": 10.650186538696289,
+ "step": 21
+ },
+ {
+ "epoch": 0.3074235807860262,
+ "grad_norm": 0.1047525629401207,
+ "learning_rate": 0.00012599999999999997,
+ "loss": 10.56808090209961,
+ "step": 22
+ },
+ {
+ "epoch": 0.32139737991266376,
+ "grad_norm": 0.10511605441570282,
+ "learning_rate": 0.00013199999999999998,
+ "loss": 10.484028816223145,
+ "step": 23
+ },
+ {
+ "epoch": 0.3353711790393013,
+ "grad_norm": 0.10467177629470825,
+ "learning_rate": 0.000138,
+ "loss": 10.399947166442871,
+ "step": 24
+ },
+ {
+ "epoch": 0.34934497816593885,
+ "grad_norm": 0.10432136058807373,
+ "learning_rate": 0.00014399999999999998,
+ "loss": 10.312714576721191,
+ "step": 25
+ },
+ {
+ "epoch": 0.36331877729257644,
+ "grad_norm": 0.10527262091636658,
+ "learning_rate": 0.00015,
+ "loss": 10.212261199951172,
+ "step": 26
+ },
+ {
+ "epoch": 0.377292576419214,
+ "grad_norm": 0.103610560297966,
+ "learning_rate": 0.000156,
+ "loss": 10.132034301757812,
+ "step": 27
+ },
+ {
+ "epoch": 0.39126637554585153,
+ "grad_norm": 0.10553059726953506,
+ "learning_rate": 0.000162,
+ "loss": 10.020434379577637,
+ "step": 28
+ },
+ {
+ "epoch": 0.4052401746724891,
+ "grad_norm": 0.1037299782037735,
+ "learning_rate": 0.000168,
+ "loss": 9.934466361999512,
+ "step": 29
+ },
+ {
+ "epoch": 0.4192139737991266,
+ "grad_norm": 0.1042739674448967,
+ "learning_rate": 0.00017399999999999997,
+ "loss": 9.822539329528809,
+ "step": 30
+ },
+ {
+ "epoch": 0.4331877729257642,
+ "grad_norm": 0.10300865024328232,
+ "learning_rate": 0.00017999999999999998,
+ "loss": 9.734994888305664,
+ "step": 31
+ },
+ {
+ "epoch": 0.44716157205240176,
+ "grad_norm": 0.10291654616594315,
+ "learning_rate": 0.000186,
+ "loss": 9.624505996704102,
+ "step": 32
+ },
+ {
+ "epoch": 0.4611353711790393,
+ "grad_norm": 0.10210851579904556,
+ "learning_rate": 0.00019199999999999998,
+ "loss": 9.518205642700195,
+ "step": 33
+ },
+ {
+ "epoch": 0.47510917030567684,
+ "grad_norm": 0.10222529619932175,
+ "learning_rate": 0.000198,
+ "loss": 9.412222862243652,
+ "step": 34
+ },
+ {
+ "epoch": 0.4890829694323144,
+ "grad_norm": 0.09992019087076187,
+ "learning_rate": 0.000204,
+ "loss": 9.326566696166992,
+ "step": 35
+ },
+ {
+ "epoch": 0.503056768558952,
+ "grad_norm": 0.09931084513664246,
+ "learning_rate": 0.00020999999999999998,
+ "loss": 9.210111618041992,
+ "step": 36
+ },
+ {
+ "epoch": 0.5170305676855895,
+ "grad_norm": 0.09928256273269653,
+ "learning_rate": 0.00021599999999999996,
+ "loss": 9.100493431091309,
+ "step": 37
+ },
+ {
+ "epoch": 0.5310043668122271,
+ "grad_norm": 0.09843816608190536,
+ "learning_rate": 0.00022199999999999998,
+ "loss": 8.996543884277344,
+ "step": 38
+ },
+ {
+ "epoch": 0.5449781659388646,
+ "grad_norm": 0.09825374186038971,
+ "learning_rate": 0.00022799999999999999,
+ "loss": 8.888991355895996,
+ "step": 39
+ },
+ {
+ "epoch": 0.5589519650655022,
+ "grad_norm": 0.09602385759353638,
+ "learning_rate": 0.000234,
+ "loss": 8.802907943725586,
+ "step": 40
+ },
+ {
+ "epoch": 0.5729257641921397,
+ "grad_norm": 0.09414064139127731,
+ "learning_rate": 0.00023999999999999998,
+ "loss": 8.715423583984375,
+ "step": 41
+ },
+ {
+ "epoch": 0.5868995633187772,
+ "grad_norm": 0.09273893386125565,
+ "learning_rate": 0.00024599999999999996,
+ "loss": 8.61374282836914,
+ "step": 42
+ },
+ {
+ "epoch": 0.6008733624454149,
+ "grad_norm": 0.09079599380493164,
+ "learning_rate": 0.00025199999999999995,
+ "loss": 8.502904891967773,
+ "step": 43
+ },
+ {
+ "epoch": 0.6148471615720524,
+ "grad_norm": 0.08605825901031494,
+ "learning_rate": 0.000258,
+ "loss": 8.453008651733398,
+ "step": 44
+ },
+ {
+ "epoch": 0.62882096069869,
+ "grad_norm": 0.08338996022939682,
+ "learning_rate": 0.00026399999999999997,
+ "loss": 8.365163803100586,
+ "step": 45
+ },
+ {
+ "epoch": 0.6427947598253275,
+ "grad_norm": 0.08150867372751236,
+ "learning_rate": 0.00027,
+ "loss": 8.285372734069824,
+ "step": 46
+ },
+ {
+ "epoch": 0.6567685589519651,
+ "grad_norm": 0.07794807851314545,
+ "learning_rate": 0.000276,
+ "loss": 8.185811996459961,
+ "step": 47
+ },
+ {
+ "epoch": 0.6707423580786026,
+ "grad_norm": 0.07592720538377762,
+ "learning_rate": 0.00028199999999999997,
+ "loss": 8.098482131958008,
+ "step": 48
+ },
+ {
+ "epoch": 0.6847161572052402,
+ "grad_norm": 0.0709897130727768,
+ "learning_rate": 0.00028799999999999995,
+ "loss": 8.042091369628906,
+ "step": 49
+ },
+ {
+ "epoch": 0.6986899563318777,
+ "grad_norm": 0.06582305580377579,
+ "learning_rate": 0.000294,
+ "loss": 7.972844123840332,
+ "step": 50
+ },
+ {
+ "epoch": 0.7126637554585152,
+ "grad_norm": 0.059452351182699203,
+ "learning_rate": 0.0003,
+ "loss": 7.924579620361328,
+ "step": 51
+ },
+ {
+ "epoch": 0.7266375545851529,
+ "grad_norm": 0.052426718175411224,
+ "learning_rate": 0.00030599999999999996,
+ "loss": 7.876336097717285,
+ "step": 52
+ },
+ {
+ "epoch": 0.7406113537117904,
+ "grad_norm": 0.04768150672316551,
+ "learning_rate": 0.000312,
+ "loss": 7.8102707862854,
+ "step": 53
+ },
+ {
+ "epoch": 0.754585152838428,
+ "grad_norm": 0.03970075398683548,
+ "learning_rate": 0.000318,
+ "loss": 7.791913986206055,
+ "step": 54
+ },
+ {
+ "epoch": 0.7685589519650655,
+ "grad_norm": 0.03559752553701401,
+ "learning_rate": 0.000324,
+ "loss": 7.756960868835449,
+ "step": 55
+ },
+ {
+ "epoch": 0.7825327510917031,
+ "grad_norm": 0.029321353882551193,
+ "learning_rate": 0.00033,
+ "loss": 7.724878787994385,
+ "step": 56
+ },
+ {
+ "epoch": 0.7965065502183406,
+ "grad_norm": 0.023485427722334862,
+ "learning_rate": 0.000336,
+ "loss": 7.72553014755249,
+ "step": 57
+ },
+ {
+ "epoch": 0.8104803493449781,
+ "grad_norm": 0.019319474697113037,
+ "learning_rate": 0.00034199999999999996,
+ "loss": 7.709493160247803,
+ "step": 58
+ },
+ {
+ "epoch": 0.8244541484716157,
+ "grad_norm": 0.016705691814422607,
+ "learning_rate": 0.00034799999999999995,
+ "loss": 7.6873297691345215,
+ "step": 59
+ },
+ {
+ "epoch": 0.8384279475982532,
+ "grad_norm": 0.017667734995484352,
+ "learning_rate": 0.00035399999999999993,
+ "loss": 7.677059173583984,
+ "step": 60
+ },
+ {
+ "epoch": 0.8524017467248908,
+ "grad_norm": 0.01671992614865303,
+ "learning_rate": 0.00035999999999999997,
+ "loss": 7.643512725830078,
+ "step": 61
+ },
+ {
+ "epoch": 0.8663755458515284,
+ "grad_norm": 0.012288172729313374,
+ "learning_rate": 0.00036599999999999995,
+ "loss": 7.658343315124512,
+ "step": 62
+ },
+ {
+ "epoch": 0.880349344978166,
+ "grad_norm": 0.012004776857793331,
+ "learning_rate": 0.000372,
+ "loss": 7.649759292602539,
+ "step": 63
+ },
+ {
+ "epoch": 0.8943231441048035,
+ "grad_norm": 0.012487445957958698,
+ "learning_rate": 0.00037799999999999997,
+ "loss": 7.631503105163574,
+ "step": 64
+ },
+ {
+ "epoch": 0.9082969432314411,
+ "grad_norm": 0.01489241048693657,
+ "learning_rate": 0.00038399999999999996,
+ "loss": 7.647110462188721,
+ "step": 65
+ },
+ {
+ "epoch": 0.9222707423580786,
+ "grad_norm": 0.01428980939090252,
+ "learning_rate": 0.00039,
+ "loss": 7.636983871459961,
+ "step": 66
+ },
+ {
+ "epoch": 0.9362445414847161,
+ "grad_norm": 0.010711144655942917,
+ "learning_rate": 0.000396,
+ "loss": 7.621130466461182,
+ "step": 67
+ },
+ {
+ "epoch": 0.9502183406113537,
+ "grad_norm": 0.010178886353969574,
+ "learning_rate": 0.000402,
+ "loss": 7.593924522399902,
+ "step": 68
+ },
+ {
+ "epoch": 0.9641921397379912,
+ "grad_norm": 0.010604874230921268,
+ "learning_rate": 0.000408,
+ "loss": 7.6015496253967285,
+ "step": 69
+ },
+ {
+ "epoch": 0.9781659388646288,
+ "grad_norm": 0.009751426056027412,
+ "learning_rate": 0.0004139999999999999,
+ "loss": 7.571964263916016,
+ "step": 70
+ },
+ {
+ "epoch": 0.9921397379912664,
+ "grad_norm": 0.009816481731832027,
+ "learning_rate": 0.00041999999999999996,
+ "loss": 7.57344913482666,
+ "step": 71
+ },
+ {
+ "epoch": 1.0,
+ "grad_norm": 0.008665637113153934,
+ "learning_rate": 0.00042599999999999995,
+ "loss": 7.579050064086914,
+ "step": 72
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 28800,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 400,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": true,
+ "should_log": false,
+ "should_save": false,
+ "should_training_stop": false
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 3.069613361332224e+16,
+ "train_batch_size": 8,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-72/training_args.bin b/runs/l2r50-i2-fulle-lm/checkpoint-72/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..cad232831c222cc7f8b938b8bbe42f61e7d89c3a
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-72/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:f7c40e3bfa2f5ac1450ca1243e3f2a895d44a06ed5995e6aefd97f78d229d8c8
+size 4792
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-7200/chat_template.jinja b/runs/l2r50-i2-fulle-lm/checkpoint-7200/chat_template.jinja
new file mode 100644
index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-7200/chat_template.jinja
@@ -0,0 +1,85 @@
+{%- if tools %}
+ {{- '<|im_start|>system\n' }}
+ {%- if messages[0].role == 'system' %}
+ {{- messages[0].content + '\n\n' }}
+ {%- endif %}
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }}
+ {%- for tool in tools %}
+ {{- "\n" }}
+ {{- tool | tojson }}
+ {%- endfor %}
+ {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }}
+{%- else %}
+ {%- if messages[0].role == 'system' %}
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
+ {%- endif %}
+{%- endif %}
+{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
+{%- for message in messages[::-1] %}
+ {%- set index = (messages|length - 1) - loop.index0 %}
+ {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %}
+ {%- set ns.multi_step_tool = false %}
+ {%- set ns.last_query_index = index %}
+ {%- endif %}
+{%- endfor %}
+{%- for message in messages %}
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
+ {%- elif message.role == "assistant" %}
+ {%- set content = message.content %}
+ {%- set reasoning_content = '' %}
+ {%- if message.reasoning_content is defined and message.reasoning_content is not none %}
+ {%- set reasoning_content = message.reasoning_content %}
+ {%- else %}
+ {%- if '' in message.content %}
+ {%- set content = message.content.split('')[-1].lstrip('\n') %}
+ {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %}
+ {%- endif %}
+ {%- endif %}
+ {%- if loop.index0 > ns.last_query_index %}
+ {%- if loop.last or (not loop.last and reasoning_content) %}
+ {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- if message.tool_calls %}
+ {%- for tool_call in message.tool_calls %}
+ {%- if (loop.first and content) or (not loop.first) %}
+ {{- '\n' }}
+ {%- endif %}
+ {%- if tool_call.function %}
+ {%- set tool_call = tool_call.function %}
+ {%- endif %}
+ {{- '\n{"name": "' }}
+ {{- tool_call.name }}
+ {{- '", "arguments": ' }}
+ {%- if tool_call.arguments is string %}
+ {{- tool_call.arguments }}
+ {%- else %}
+ {{- tool_call.arguments | tojson }}
+ {%- endif %}
+ {{- '}\n' }}
+ {%- endfor %}
+ {%- endif %}
+ {{- '<|im_end|>\n' }}
+ {%- elif message.role == "tool" %}
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
+ {{- '<|im_start|>user' }}
+ {%- endif %}
+ {{- '\n\n' }}
+ {{- message.content }}
+ {{- '\n' }}
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
+ {{- '<|im_end|>\n' }}
+ {%- endif %}
+ {%- endif %}
+{%- endfor %}
+{%- if add_generation_prompt %}
+ {{- '<|im_start|>assistant\n' }}
+ {%- if enable_thinking is defined and enable_thinking is false %}
+ {{- '\n\n\n\n' }}
+ {%- endif %}
+{%- endif %}
\ No newline at end of file
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-7200/config.json b/runs/l2r50-i2-fulle-lm/checkpoint-7200/config.json
new file mode 100644
index 0000000000000000000000000000000000000000..22f250953c6ce8fc91ee4411856dd73041a415e2
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-7200/config.json
@@ -0,0 +1,32 @@
+{
+ "architectures": [
+ "LlamaForCausalLM"
+ ],
+ "attention_bias": false,
+ "attention_dropout": 0.0,
+ "bos_token_id": null,
+ "dtype": "float32",
+ "eos_token_id": 151645,
+ "head_dim": 128,
+ "hidden_act": "silu",
+ "hidden_size": 512,
+ "initializer_range": 0.02,
+ "intermediate_size": 1536,
+ "max_position_embeddings": 2048,
+ "mlp_bias": false,
+ "model_type": "llama",
+ "num_attention_heads": 4,
+ "num_hidden_layers": 20,
+ "num_key_value_heads": 4,
+ "pad_token_id": 151645,
+ "pretraining_tp": 1,
+ "rms_norm_eps": 1e-06,
+ "rope_parameters": {
+ "rope_theta": 10000.0,
+ "rope_type": "default"
+ },
+ "tie_word_embeddings": true,
+ "transformers_version": "5.5.0",
+ "use_cache": false,
+ "vocab_size": 151673
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-7200/generation_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-7200/generation_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-7200/generation_config.json
@@ -0,0 +1,11 @@
+{
+ "_from_model_config": true,
+ "eos_token_id": [
+ 151645
+ ],
+ "output_attentions": false,
+ "output_hidden_states": false,
+ "pad_token_id": 151645,
+ "transformers_version": "5.5.0",
+ "use_cache": true
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-7200/model.safetensors b/runs/l2r50-i2-fulle-lm/checkpoint-7200/model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..11b1b00a9fb8630c6045e70e23bd9ee0ac04b991
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-7200/model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:992ea84d161c0ced7c124ed737b6e624e3398b7f1573ca2595bcdaf054443a17
+size 583360328
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-7200/optimizer.pt b/runs/l2r50-i2-fulle-lm/checkpoint-7200/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..cc7b60119cc8d73a8da7efe1ef589422eb496662
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-7200/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:acc9564263da743645154f940ed413297dcb67ed8abb28c36c62eb1619aebaa8
+size 1166833530
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-7200/rng_state_0.pth b/runs/l2r50-i2-fulle-lm/checkpoint-7200/rng_state_0.pth
new file mode 100644
index 0000000000000000000000000000000000000000..e5981255de81c81160f3351d04278830d3bb2507
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-7200/rng_state_0.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:0fc4fb416576a2293bd2b09eb738accb4d54bde927d92f2fe90be3a448c990e5
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-7200/rng_state_1.pth b/runs/l2r50-i2-fulle-lm/checkpoint-7200/rng_state_1.pth
new file mode 100644
index 0000000000000000000000000000000000000000..edd57664d15ac959bb114a9d87d65d0436eb2365
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-7200/rng_state_1.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:69ae21eb3a416db05955df400b5252da17a3f027ae01a9ee9dcf0c5b62137ce1
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-7200/rng_state_2.pth b/runs/l2r50-i2-fulle-lm/checkpoint-7200/rng_state_2.pth
new file mode 100644
index 0000000000000000000000000000000000000000..fe7b2d78121a938380d42815769910f27018d664
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-7200/rng_state_2.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:907ef0007bd49a0897d16fbbeecd39ead8005845e40b3dffeb613c567c7d36e0
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-7200/rng_state_3.pth b/runs/l2r50-i2-fulle-lm/checkpoint-7200/rng_state_3.pth
new file mode 100644
index 0000000000000000000000000000000000000000..48e04b3f4c6cb13a09a4fc89cc4a3af943e22dc5
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-7200/rng_state_3.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:8bbf8a37dcf5f2f1715be25d2aae1d740d39d750efa40b777134b7c66ef0bff1
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-7200/scheduler.pt b/runs/l2r50-i2-fulle-lm/checkpoint-7200/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..624c26e041c29e4eef4955201869827f6f6a567e
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-7200/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:3f84784070f59d3d6db22567a60409a64f07923bdf1c69e2acde1a156c945423
+size 1064
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-7200/tokenizer.json b/runs/l2r50-i2-fulle-lm/checkpoint-7200/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..53786016f070e11184813f3c83101dc5e93e392e
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-7200/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:c3c63fda1646a555448e8f03662542e04b11d103721319e4d9b267d7da03ceb8
+size 11424463
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-7200/tokenizer_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-7200/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..98e2fb9b2cc54785dc9d871b8e503f8b8f2defe7
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-7200/tokenizer_config.json
@@ -0,0 +1,20 @@
+{
+ "add_prefix_space": false,
+ "backend": "tokenizers",
+ "bos_token": null,
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "<|im_end|>",
+ "errors": "replace",
+ "extra_special_tokens": [
+ "<|l2r_pred|>",
+ "<|r2l_pred|>",
+ "<|next_1_pred|>",
+ "<|next_2_pred|>"
+ ],
+ "is_local": false,
+ "model_max_length": 131072,
+ "pad_token": "<|im_end|>",
+ "split_special_tokens": false,
+ "tokenizer_class": "Qwen2Tokenizer",
+ "unk_token": null
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-7200/trainer_state.json b/runs/l2r50-i2-fulle-lm/checkpoint-7200/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..6dad79039d68014c806eb6ddcc522fb63745620e
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-7200/trainer_state.json
@@ -0,0 +1,51226 @@
+{
+ "best_global_step": null,
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 100.0,
+ "eval_steps": 500,
+ "global_step": 7200,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.013973799126637555,
+ "grad_norm": 0.1739506870508194,
+ "learning_rate": 0.0,
+ "loss": 11.99299430847168,
+ "step": 1
+ },
+ {
+ "epoch": 0.02794759825327511,
+ "grad_norm": 0.17216597497463226,
+ "learning_rate": 5.999999999999999e-06,
+ "loss": 11.988704681396484,
+ "step": 2
+ },
+ {
+ "epoch": 0.04192139737991266,
+ "grad_norm": 0.17459715902805328,
+ "learning_rate": 1.1999999999999999e-05,
+ "loss": 11.945581436157227,
+ "step": 3
+ },
+ {
+ "epoch": 0.05589519650655022,
+ "grad_norm": 0.17577414214611053,
+ "learning_rate": 1.7999999999999997e-05,
+ "loss": 11.859772682189941,
+ "step": 4
+ },
+ {
+ "epoch": 0.06986899563318777,
+ "grad_norm": 0.16163651645183563,
+ "learning_rate": 2.3999999999999997e-05,
+ "loss": 11.744415283203125,
+ "step": 5
+ },
+ {
+ "epoch": 0.08384279475982533,
+ "grad_norm": 0.14409932494163513,
+ "learning_rate": 2.9999999999999997e-05,
+ "loss": 11.618099212646484,
+ "step": 6
+ },
+ {
+ "epoch": 0.09781659388646288,
+ "grad_norm": 0.13079848885536194,
+ "learning_rate": 3.5999999999999994e-05,
+ "loss": 11.505935668945312,
+ "step": 7
+ },
+ {
+ "epoch": 0.11179039301310044,
+ "grad_norm": 0.11964613944292068,
+ "learning_rate": 4.2e-05,
+ "loss": 11.416532516479492,
+ "step": 8
+ },
+ {
+ "epoch": 0.125764192139738,
+ "grad_norm": 0.11345142871141434,
+ "learning_rate": 4.7999999999999994e-05,
+ "loss": 11.33206558227539,
+ "step": 9
+ },
+ {
+ "epoch": 0.13973799126637554,
+ "grad_norm": 0.11033952981233597,
+ "learning_rate": 5.399999999999999e-05,
+ "loss": 11.259052276611328,
+ "step": 10
+ },
+ {
+ "epoch": 0.1537117903930131,
+ "grad_norm": 0.10776489973068237,
+ "learning_rate": 5.9999999999999995e-05,
+ "loss": 11.203384399414062,
+ "step": 11
+ },
+ {
+ "epoch": 0.16768558951965065,
+ "grad_norm": 0.10721946507692337,
+ "learning_rate": 6.599999999999999e-05,
+ "loss": 11.153295516967773,
+ "step": 12
+ },
+ {
+ "epoch": 0.18165938864628822,
+ "grad_norm": 0.10673430562019348,
+ "learning_rate": 7.199999999999999e-05,
+ "loss": 11.110200881958008,
+ "step": 13
+ },
+ {
+ "epoch": 0.19563318777292577,
+ "grad_norm": 0.10630254447460175,
+ "learning_rate": 7.8e-05,
+ "loss": 11.067477226257324,
+ "step": 14
+ },
+ {
+ "epoch": 0.2096069868995633,
+ "grad_norm": 0.10631615668535233,
+ "learning_rate": 8.4e-05,
+ "loss": 11.022941589355469,
+ "step": 15
+ },
+ {
+ "epoch": 0.22358078602620088,
+ "grad_norm": 0.10609443485736847,
+ "learning_rate": 8.999999999999999e-05,
+ "loss": 10.970746040344238,
+ "step": 16
+ },
+ {
+ "epoch": 0.23755458515283842,
+ "grad_norm": 0.10648232698440552,
+ "learning_rate": 9.599999999999999e-05,
+ "loss": 10.91267204284668,
+ "step": 17
+ },
+ {
+ "epoch": 0.251528384279476,
+ "grad_norm": 0.10584589838981628,
+ "learning_rate": 0.000102,
+ "loss": 10.854547500610352,
+ "step": 18
+ },
+ {
+ "epoch": 0.26550218340611353,
+ "grad_norm": 0.10591720044612885,
+ "learning_rate": 0.00010799999999999998,
+ "loss": 10.788917541503906,
+ "step": 19
+ },
+ {
+ "epoch": 0.2794759825327511,
+ "grad_norm": 0.10611672699451447,
+ "learning_rate": 0.00011399999999999999,
+ "loss": 10.717556953430176,
+ "step": 20
+ },
+ {
+ "epoch": 0.2934497816593886,
+ "grad_norm": 0.10455359518527985,
+ "learning_rate": 0.00011999999999999999,
+ "loss": 10.650186538696289,
+ "step": 21
+ },
+ {
+ "epoch": 0.3074235807860262,
+ "grad_norm": 0.1047525629401207,
+ "learning_rate": 0.00012599999999999997,
+ "loss": 10.56808090209961,
+ "step": 22
+ },
+ {
+ "epoch": 0.32139737991266376,
+ "grad_norm": 0.10511605441570282,
+ "learning_rate": 0.00013199999999999998,
+ "loss": 10.484028816223145,
+ "step": 23
+ },
+ {
+ "epoch": 0.3353711790393013,
+ "grad_norm": 0.10467177629470825,
+ "learning_rate": 0.000138,
+ "loss": 10.399947166442871,
+ "step": 24
+ },
+ {
+ "epoch": 0.34934497816593885,
+ "grad_norm": 0.10432136058807373,
+ "learning_rate": 0.00014399999999999998,
+ "loss": 10.312714576721191,
+ "step": 25
+ },
+ {
+ "epoch": 0.36331877729257644,
+ "grad_norm": 0.10527262091636658,
+ "learning_rate": 0.00015,
+ "loss": 10.212261199951172,
+ "step": 26
+ },
+ {
+ "epoch": 0.377292576419214,
+ "grad_norm": 0.103610560297966,
+ "learning_rate": 0.000156,
+ "loss": 10.132034301757812,
+ "step": 27
+ },
+ {
+ "epoch": 0.39126637554585153,
+ "grad_norm": 0.10553059726953506,
+ "learning_rate": 0.000162,
+ "loss": 10.020434379577637,
+ "step": 28
+ },
+ {
+ "epoch": 0.4052401746724891,
+ "grad_norm": 0.1037299782037735,
+ "learning_rate": 0.000168,
+ "loss": 9.934466361999512,
+ "step": 29
+ },
+ {
+ "epoch": 0.4192139737991266,
+ "grad_norm": 0.1042739674448967,
+ "learning_rate": 0.00017399999999999997,
+ "loss": 9.822539329528809,
+ "step": 30
+ },
+ {
+ "epoch": 0.4331877729257642,
+ "grad_norm": 0.10300865024328232,
+ "learning_rate": 0.00017999999999999998,
+ "loss": 9.734994888305664,
+ "step": 31
+ },
+ {
+ "epoch": 0.44716157205240176,
+ "grad_norm": 0.10291654616594315,
+ "learning_rate": 0.000186,
+ "loss": 9.624505996704102,
+ "step": 32
+ },
+ {
+ "epoch": 0.4611353711790393,
+ "grad_norm": 0.10210851579904556,
+ "learning_rate": 0.00019199999999999998,
+ "loss": 9.518205642700195,
+ "step": 33
+ },
+ {
+ "epoch": 0.47510917030567684,
+ "grad_norm": 0.10222529619932175,
+ "learning_rate": 0.000198,
+ "loss": 9.412222862243652,
+ "step": 34
+ },
+ {
+ "epoch": 0.4890829694323144,
+ "grad_norm": 0.09992019087076187,
+ "learning_rate": 0.000204,
+ "loss": 9.326566696166992,
+ "step": 35
+ },
+ {
+ "epoch": 0.503056768558952,
+ "grad_norm": 0.09931084513664246,
+ "learning_rate": 0.00020999999999999998,
+ "loss": 9.210111618041992,
+ "step": 36
+ },
+ {
+ "epoch": 0.5170305676855895,
+ "grad_norm": 0.09928256273269653,
+ "learning_rate": 0.00021599999999999996,
+ "loss": 9.100493431091309,
+ "step": 37
+ },
+ {
+ "epoch": 0.5310043668122271,
+ "grad_norm": 0.09843816608190536,
+ "learning_rate": 0.00022199999999999998,
+ "loss": 8.996543884277344,
+ "step": 38
+ },
+ {
+ "epoch": 0.5449781659388646,
+ "grad_norm": 0.09825374186038971,
+ "learning_rate": 0.00022799999999999999,
+ "loss": 8.888991355895996,
+ "step": 39
+ },
+ {
+ "epoch": 0.5589519650655022,
+ "grad_norm": 0.09602385759353638,
+ "learning_rate": 0.000234,
+ "loss": 8.802907943725586,
+ "step": 40
+ },
+ {
+ "epoch": 0.5729257641921397,
+ "grad_norm": 0.09414064139127731,
+ "learning_rate": 0.00023999999999999998,
+ "loss": 8.715423583984375,
+ "step": 41
+ },
+ {
+ "epoch": 0.5868995633187772,
+ "grad_norm": 0.09273893386125565,
+ "learning_rate": 0.00024599999999999996,
+ "loss": 8.61374282836914,
+ "step": 42
+ },
+ {
+ "epoch": 0.6008733624454149,
+ "grad_norm": 0.09079599380493164,
+ "learning_rate": 0.00025199999999999995,
+ "loss": 8.502904891967773,
+ "step": 43
+ },
+ {
+ "epoch": 0.6148471615720524,
+ "grad_norm": 0.08605825901031494,
+ "learning_rate": 0.000258,
+ "loss": 8.453008651733398,
+ "step": 44
+ },
+ {
+ "epoch": 0.62882096069869,
+ "grad_norm": 0.08338996022939682,
+ "learning_rate": 0.00026399999999999997,
+ "loss": 8.365163803100586,
+ "step": 45
+ },
+ {
+ "epoch": 0.6427947598253275,
+ "grad_norm": 0.08150867372751236,
+ "learning_rate": 0.00027,
+ "loss": 8.285372734069824,
+ "step": 46
+ },
+ {
+ "epoch": 0.6567685589519651,
+ "grad_norm": 0.07794807851314545,
+ "learning_rate": 0.000276,
+ "loss": 8.185811996459961,
+ "step": 47
+ },
+ {
+ "epoch": 0.6707423580786026,
+ "grad_norm": 0.07592720538377762,
+ "learning_rate": 0.00028199999999999997,
+ "loss": 8.098482131958008,
+ "step": 48
+ },
+ {
+ "epoch": 0.6847161572052402,
+ "grad_norm": 0.0709897130727768,
+ "learning_rate": 0.00028799999999999995,
+ "loss": 8.042091369628906,
+ "step": 49
+ },
+ {
+ "epoch": 0.6986899563318777,
+ "grad_norm": 0.06582305580377579,
+ "learning_rate": 0.000294,
+ "loss": 7.972844123840332,
+ "step": 50
+ },
+ {
+ "epoch": 0.7126637554585152,
+ "grad_norm": 0.059452351182699203,
+ "learning_rate": 0.0003,
+ "loss": 7.924579620361328,
+ "step": 51
+ },
+ {
+ "epoch": 0.7266375545851529,
+ "grad_norm": 0.052426718175411224,
+ "learning_rate": 0.00030599999999999996,
+ "loss": 7.876336097717285,
+ "step": 52
+ },
+ {
+ "epoch": 0.7406113537117904,
+ "grad_norm": 0.04768150672316551,
+ "learning_rate": 0.000312,
+ "loss": 7.8102707862854,
+ "step": 53
+ },
+ {
+ "epoch": 0.754585152838428,
+ "grad_norm": 0.03970075398683548,
+ "learning_rate": 0.000318,
+ "loss": 7.791913986206055,
+ "step": 54
+ },
+ {
+ "epoch": 0.7685589519650655,
+ "grad_norm": 0.03559752553701401,
+ "learning_rate": 0.000324,
+ "loss": 7.756960868835449,
+ "step": 55
+ },
+ {
+ "epoch": 0.7825327510917031,
+ "grad_norm": 0.029321353882551193,
+ "learning_rate": 0.00033,
+ "loss": 7.724878787994385,
+ "step": 56
+ },
+ {
+ "epoch": 0.7965065502183406,
+ "grad_norm": 0.023485427722334862,
+ "learning_rate": 0.000336,
+ "loss": 7.72553014755249,
+ "step": 57
+ },
+ {
+ "epoch": 0.8104803493449781,
+ "grad_norm": 0.019319474697113037,
+ "learning_rate": 0.00034199999999999996,
+ "loss": 7.709493160247803,
+ "step": 58
+ },
+ {
+ "epoch": 0.8244541484716157,
+ "grad_norm": 0.016705691814422607,
+ "learning_rate": 0.00034799999999999995,
+ "loss": 7.6873297691345215,
+ "step": 59
+ },
+ {
+ "epoch": 0.8384279475982532,
+ "grad_norm": 0.017667734995484352,
+ "learning_rate": 0.00035399999999999993,
+ "loss": 7.677059173583984,
+ "step": 60
+ },
+ {
+ "epoch": 0.8524017467248908,
+ "grad_norm": 0.01671992614865303,
+ "learning_rate": 0.00035999999999999997,
+ "loss": 7.643512725830078,
+ "step": 61
+ },
+ {
+ "epoch": 0.8663755458515284,
+ "grad_norm": 0.012288172729313374,
+ "learning_rate": 0.00036599999999999995,
+ "loss": 7.658343315124512,
+ "step": 62
+ },
+ {
+ "epoch": 0.880349344978166,
+ "grad_norm": 0.012004776857793331,
+ "learning_rate": 0.000372,
+ "loss": 7.649759292602539,
+ "step": 63
+ },
+ {
+ "epoch": 0.8943231441048035,
+ "grad_norm": 0.012487445957958698,
+ "learning_rate": 0.00037799999999999997,
+ "loss": 7.631503105163574,
+ "step": 64
+ },
+ {
+ "epoch": 0.9082969432314411,
+ "grad_norm": 0.01489241048693657,
+ "learning_rate": 0.00038399999999999996,
+ "loss": 7.647110462188721,
+ "step": 65
+ },
+ {
+ "epoch": 0.9222707423580786,
+ "grad_norm": 0.01428980939090252,
+ "learning_rate": 0.00039,
+ "loss": 7.636983871459961,
+ "step": 66
+ },
+ {
+ "epoch": 0.9362445414847161,
+ "grad_norm": 0.010711144655942917,
+ "learning_rate": 0.000396,
+ "loss": 7.621130466461182,
+ "step": 67
+ },
+ {
+ "epoch": 0.9502183406113537,
+ "grad_norm": 0.010178886353969574,
+ "learning_rate": 0.000402,
+ "loss": 7.593924522399902,
+ "step": 68
+ },
+ {
+ "epoch": 0.9641921397379912,
+ "grad_norm": 0.010604874230921268,
+ "learning_rate": 0.000408,
+ "loss": 7.6015496253967285,
+ "step": 69
+ },
+ {
+ "epoch": 0.9781659388646288,
+ "grad_norm": 0.009751426056027412,
+ "learning_rate": 0.0004139999999999999,
+ "loss": 7.571964263916016,
+ "step": 70
+ },
+ {
+ "epoch": 0.9921397379912664,
+ "grad_norm": 0.009816481731832027,
+ "learning_rate": 0.00041999999999999996,
+ "loss": 7.57344913482666,
+ "step": 71
+ },
+ {
+ "epoch": 1.0,
+ "grad_norm": 0.008665637113153934,
+ "learning_rate": 0.00042599999999999995,
+ "loss": 7.579050064086914,
+ "step": 72
+ },
+ {
+ "epoch": 1.0,
+ "eval_loss": 7.581204891204834,
+ "eval_runtime": 59.0091,
+ "eval_samples_per_second": 41.383,
+ "eval_steps_per_second": 1.305,
+ "step": 72
+ },
+ {
+ "epoch": 1.0139737991266375,
+ "grad_norm": 0.009684201329946518,
+ "learning_rate": 0.00043199999999999993,
+ "loss": 7.568594932556152,
+ "step": 73
+ },
+ {
+ "epoch": 1.027947598253275,
+ "grad_norm": 0.025783710181713104,
+ "learning_rate": 0.00043799999999999997,
+ "loss": 7.578948497772217,
+ "step": 74
+ },
+ {
+ "epoch": 1.0419213973799126,
+ "grad_norm": 0.0327819399535656,
+ "learning_rate": 0.00044399999999999995,
+ "loss": 7.548173904418945,
+ "step": 75
+ },
+ {
+ "epoch": 1.0558951965065502,
+ "grad_norm": 0.021604498848319054,
+ "learning_rate": 0.00045,
+ "loss": 7.552953720092773,
+ "step": 76
+ },
+ {
+ "epoch": 1.0698689956331877,
+ "grad_norm": 0.010230105370283127,
+ "learning_rate": 0.00045599999999999997,
+ "loss": 7.5213212966918945,
+ "step": 77
+ },
+ {
+ "epoch": 1.0838427947598253,
+ "grad_norm": 0.023315589874982834,
+ "learning_rate": 0.00046199999999999995,
+ "loss": 7.501640319824219,
+ "step": 78
+ },
+ {
+ "epoch": 1.0978165938864628,
+ "grad_norm": 0.01109944935888052,
+ "learning_rate": 0.000468,
+ "loss": 7.539519786834717,
+ "step": 79
+ },
+ {
+ "epoch": 1.1117903930131003,
+ "grad_norm": 0.026093758642673492,
+ "learning_rate": 0.000474,
+ "loss": 7.525928974151611,
+ "step": 80
+ },
+ {
+ "epoch": 1.125764192139738,
+ "grad_norm": 0.03274895250797272,
+ "learning_rate": 0.00047999999999999996,
+ "loss": 7.535602569580078,
+ "step": 81
+ },
+ {
+ "epoch": 1.1397379912663754,
+ "grad_norm": 0.01543602254241705,
+ "learning_rate": 0.000486,
+ "loss": 7.513683795928955,
+ "step": 82
+ },
+ {
+ "epoch": 1.1537117903930132,
+ "grad_norm": 0.028377249836921692,
+ "learning_rate": 0.0004919999999999999,
+ "loss": 7.49384880065918,
+ "step": 83
+ },
+ {
+ "epoch": 1.1676855895196507,
+ "grad_norm": 0.04121481627225876,
+ "learning_rate": 0.000498,
+ "loss": 7.488447189331055,
+ "step": 84
+ },
+ {
+ "epoch": 1.1816593886462883,
+ "grad_norm": 0.013282302767038345,
+ "learning_rate": 0.0005039999999999999,
+ "loss": 7.463010787963867,
+ "step": 85
+ },
+ {
+ "epoch": 1.1956331877729258,
+ "grad_norm": 0.03200364485383034,
+ "learning_rate": 0.0005099999999999999,
+ "loss": 7.473515033721924,
+ "step": 86
+ },
+ {
+ "epoch": 1.2096069868995634,
+ "grad_norm": 0.020768029615283012,
+ "learning_rate": 0.000516,
+ "loss": 7.464688777923584,
+ "step": 87
+ },
+ {
+ "epoch": 1.223580786026201,
+ "grad_norm": 0.021715788170695305,
+ "learning_rate": 0.000522,
+ "loss": 7.4835028648376465,
+ "step": 88
+ },
+ {
+ "epoch": 1.2375545851528384,
+ "grad_norm": 0.026255754753947258,
+ "learning_rate": 0.0005279999999999999,
+ "loss": 7.443392753601074,
+ "step": 89
+ },
+ {
+ "epoch": 1.251528384279476,
+ "grad_norm": 0.022325951606035233,
+ "learning_rate": 0.000534,
+ "loss": 7.457373142242432,
+ "step": 90
+ },
+ {
+ "epoch": 1.2655021834061135,
+ "grad_norm": 0.009270387701690197,
+ "learning_rate": 0.00054,
+ "loss": 7.455986976623535,
+ "step": 91
+ },
+ {
+ "epoch": 1.279475982532751,
+ "grad_norm": 0.02369544468820095,
+ "learning_rate": 0.0005459999999999999,
+ "loss": 7.442216873168945,
+ "step": 92
+ },
+ {
+ "epoch": 1.2934497816593886,
+ "grad_norm": 0.008474506437778473,
+ "learning_rate": 0.000552,
+ "loss": 7.4294915199279785,
+ "step": 93
+ },
+ {
+ "epoch": 1.3074235807860262,
+ "grad_norm": 0.017043249681591988,
+ "learning_rate": 0.000558,
+ "loss": 7.406932830810547,
+ "step": 94
+ },
+ {
+ "epoch": 1.3213973799126637,
+ "grad_norm": 0.011813412420451641,
+ "learning_rate": 0.0005639999999999999,
+ "loss": 7.415900230407715,
+ "step": 95
+ },
+ {
+ "epoch": 1.3353711790393012,
+ "grad_norm": 0.01086883433163166,
+ "learning_rate": 0.00057,
+ "loss": 7.452998638153076,
+ "step": 96
+ },
+ {
+ "epoch": 1.3493449781659388,
+ "grad_norm": 0.014700951054692268,
+ "learning_rate": 0.0005759999999999999,
+ "loss": 7.410249710083008,
+ "step": 97
+ },
+ {
+ "epoch": 1.3633187772925766,
+ "grad_norm": 0.014512976631522179,
+ "learning_rate": 0.0005819999999999999,
+ "loss": 7.429132461547852,
+ "step": 98
+ },
+ {
+ "epoch": 1.3772925764192139,
+ "grad_norm": 0.012519920244812965,
+ "learning_rate": 0.000588,
+ "loss": 7.430817604064941,
+ "step": 99
+ },
+ {
+ "epoch": 1.3912663755458516,
+ "grad_norm": 0.007132664322853088,
+ "learning_rate": 0.0005939999999999999,
+ "loss": 7.405234336853027,
+ "step": 100
+ },
+ {
+ "epoch": 1.405240174672489,
+ "grad_norm": 0.019489388912916183,
+ "learning_rate": 0.0006,
+ "loss": 7.427791595458984,
+ "step": 101
+ },
+ {
+ "epoch": 1.4192139737991267,
+ "grad_norm": 0.027252664789557457,
+ "learning_rate": 0.0006,
+ "loss": 7.41761589050293,
+ "step": 102
+ },
+ {
+ "epoch": 1.4331877729257643,
+ "grad_norm": 0.02501637488603592,
+ "learning_rate": 0.0006,
+ "loss": 7.416735649108887,
+ "step": 103
+ },
+ {
+ "epoch": 1.4471615720524018,
+ "grad_norm": 0.00989772193133831,
+ "learning_rate": 0.0006,
+ "loss": 7.3756632804870605,
+ "step": 104
+ },
+ {
+ "epoch": 1.4611353711790394,
+ "grad_norm": 0.01622949168086052,
+ "learning_rate": 0.0006,
+ "loss": 7.400811672210693,
+ "step": 105
+ },
+ {
+ "epoch": 1.475109170305677,
+ "grad_norm": 0.012407763861119747,
+ "learning_rate": 0.0006,
+ "loss": 7.37675666809082,
+ "step": 106
+ },
+ {
+ "epoch": 1.4890829694323144,
+ "grad_norm": 0.010353565216064453,
+ "learning_rate": 0.0006,
+ "loss": 7.4013142585754395,
+ "step": 107
+ },
+ {
+ "epoch": 1.503056768558952,
+ "grad_norm": 0.0120553532615304,
+ "learning_rate": 0.0006,
+ "loss": 7.379873275756836,
+ "step": 108
+ },
+ {
+ "epoch": 1.5170305676855895,
+ "grad_norm": 0.01610225811600685,
+ "learning_rate": 0.0006,
+ "loss": 7.371631145477295,
+ "step": 109
+ },
+ {
+ "epoch": 1.531004366812227,
+ "grad_norm": 0.01159879844635725,
+ "learning_rate": 0.0006,
+ "loss": 7.389007568359375,
+ "step": 110
+ },
+ {
+ "epoch": 1.5449781659388646,
+ "grad_norm": 0.008013816550374031,
+ "learning_rate": 0.0006,
+ "loss": 7.395920753479004,
+ "step": 111
+ },
+ {
+ "epoch": 1.5589519650655022,
+ "grad_norm": 0.016334809362888336,
+ "learning_rate": 0.0006,
+ "loss": 7.393121719360352,
+ "step": 112
+ },
+ {
+ "epoch": 1.5729257641921397,
+ "grad_norm": 0.018821721896529198,
+ "learning_rate": 0.0006,
+ "loss": 7.367543697357178,
+ "step": 113
+ },
+ {
+ "epoch": 1.5868995633187772,
+ "grad_norm": 0.014088994823396206,
+ "learning_rate": 0.0006,
+ "loss": 7.385537624359131,
+ "step": 114
+ },
+ {
+ "epoch": 1.600873362445415,
+ "grad_norm": 0.005864050704985857,
+ "learning_rate": 0.0006,
+ "loss": 7.365346908569336,
+ "step": 115
+ },
+ {
+ "epoch": 1.6148471615720523,
+ "grad_norm": 0.011745430529117584,
+ "learning_rate": 0.0006,
+ "loss": 7.390227317810059,
+ "step": 116
+ },
+ {
+ "epoch": 1.62882096069869,
+ "grad_norm": 0.017205573618412018,
+ "learning_rate": 0.0006,
+ "loss": 7.3558549880981445,
+ "step": 117
+ },
+ {
+ "epoch": 1.6427947598253274,
+ "grad_norm": 0.021256346255540848,
+ "learning_rate": 0.0006,
+ "loss": 7.377346992492676,
+ "step": 118
+ },
+ {
+ "epoch": 1.6567685589519652,
+ "grad_norm": 0.019335204735398293,
+ "learning_rate": 0.0006,
+ "loss": 7.346194267272949,
+ "step": 119
+ },
+ {
+ "epoch": 1.6707423580786025,
+ "grad_norm": 0.008134003728628159,
+ "learning_rate": 0.0006,
+ "loss": 7.3063578605651855,
+ "step": 120
+ },
+ {
+ "epoch": 1.6847161572052403,
+ "grad_norm": 0.017613831907510757,
+ "learning_rate": 0.0006,
+ "loss": 7.338171005249023,
+ "step": 121
+ },
+ {
+ "epoch": 1.6986899563318776,
+ "grad_norm": 0.00890734139829874,
+ "learning_rate": 0.0006,
+ "loss": 7.355384826660156,
+ "step": 122
+ },
+ {
+ "epoch": 1.7126637554585153,
+ "grad_norm": 0.013448155485093594,
+ "learning_rate": 0.0006,
+ "loss": 7.344296455383301,
+ "step": 123
+ },
+ {
+ "epoch": 1.726637554585153,
+ "grad_norm": 0.009241633117198944,
+ "learning_rate": 0.0006,
+ "loss": 7.3454060554504395,
+ "step": 124
+ },
+ {
+ "epoch": 1.7406113537117904,
+ "grad_norm": 0.018561311066150665,
+ "learning_rate": 0.0006,
+ "loss": 7.32348108291626,
+ "step": 125
+ },
+ {
+ "epoch": 1.754585152838428,
+ "grad_norm": 0.025667704641819,
+ "learning_rate": 0.0006,
+ "loss": 7.356142520904541,
+ "step": 126
+ },
+ {
+ "epoch": 1.7685589519650655,
+ "grad_norm": 0.036232151091098785,
+ "learning_rate": 0.0006,
+ "loss": 7.363539695739746,
+ "step": 127
+ },
+ {
+ "epoch": 1.782532751091703,
+ "grad_norm": 0.059483714401721954,
+ "learning_rate": 0.0006,
+ "loss": 7.402331352233887,
+ "step": 128
+ },
+ {
+ "epoch": 1.7965065502183406,
+ "grad_norm": 0.05365009233355522,
+ "learning_rate": 0.0006,
+ "loss": 7.40457820892334,
+ "step": 129
+ },
+ {
+ "epoch": 1.8104803493449781,
+ "grad_norm": 0.02738415263593197,
+ "learning_rate": 0.0006,
+ "loss": 7.3756208419799805,
+ "step": 130
+ },
+ {
+ "epoch": 1.8244541484716157,
+ "grad_norm": 0.035017091780900955,
+ "learning_rate": 0.0006,
+ "loss": 7.379263877868652,
+ "step": 131
+ },
+ {
+ "epoch": 1.8384279475982532,
+ "grad_norm": 0.02819863148033619,
+ "learning_rate": 0.0006,
+ "loss": 7.327673435211182,
+ "step": 132
+ },
+ {
+ "epoch": 1.8524017467248908,
+ "grad_norm": 0.017754238098859787,
+ "learning_rate": 0.0006,
+ "loss": 7.333420276641846,
+ "step": 133
+ },
+ {
+ "epoch": 1.8663755458515285,
+ "grad_norm": 0.024685733020305634,
+ "learning_rate": 0.0006,
+ "loss": 7.34493350982666,
+ "step": 134
+ },
+ {
+ "epoch": 1.8803493449781659,
+ "grad_norm": 0.019426532089710236,
+ "learning_rate": 0.0006,
+ "loss": 7.328448295593262,
+ "step": 135
+ },
+ {
+ "epoch": 1.8943231441048036,
+ "grad_norm": 0.02106391079723835,
+ "learning_rate": 0.0006,
+ "loss": 7.327917098999023,
+ "step": 136
+ },
+ {
+ "epoch": 1.908296943231441,
+ "grad_norm": 0.02056879736483097,
+ "learning_rate": 0.0006,
+ "loss": 7.3311991691589355,
+ "step": 137
+ },
+ {
+ "epoch": 1.9222707423580787,
+ "grad_norm": 0.016865141689777374,
+ "learning_rate": 0.0006,
+ "loss": 7.312766075134277,
+ "step": 138
+ },
+ {
+ "epoch": 1.936244541484716,
+ "grad_norm": 0.017549855634570122,
+ "learning_rate": 0.0006,
+ "loss": 7.31191349029541,
+ "step": 139
+ },
+ {
+ "epoch": 1.9502183406113538,
+ "grad_norm": 0.015930157154798508,
+ "learning_rate": 0.0006,
+ "loss": 7.330901622772217,
+ "step": 140
+ },
+ {
+ "epoch": 1.9641921397379911,
+ "grad_norm": 0.013283558189868927,
+ "learning_rate": 0.0006,
+ "loss": 7.298318386077881,
+ "step": 141
+ },
+ {
+ "epoch": 1.9781659388646289,
+ "grad_norm": 0.012023469433188438,
+ "learning_rate": 0.0006,
+ "loss": 7.282337188720703,
+ "step": 142
+ },
+ {
+ "epoch": 1.9921397379912664,
+ "grad_norm": 0.01459511648863554,
+ "learning_rate": 0.0006,
+ "loss": 7.282924175262451,
+ "step": 143
+ },
+ {
+ "epoch": 2.0,
+ "grad_norm": 0.017474235966801643,
+ "learning_rate": 0.0006,
+ "loss": 7.254333972930908,
+ "step": 144
+ },
+ {
+ "epoch": 2.0,
+ "eval_loss": 7.318708896636963,
+ "eval_runtime": 59.6555,
+ "eval_samples_per_second": 40.935,
+ "eval_steps_per_second": 1.291,
+ "step": 144
+ },
+ {
+ "epoch": 2.0139737991266378,
+ "grad_norm": 0.019757507368922234,
+ "learning_rate": 0.0006,
+ "loss": 7.30288553237915,
+ "step": 145
+ },
+ {
+ "epoch": 2.027947598253275,
+ "grad_norm": 0.014571248553693295,
+ "learning_rate": 0.0006,
+ "loss": 7.274681568145752,
+ "step": 146
+ },
+ {
+ "epoch": 2.041921397379913,
+ "grad_norm": 0.015744198113679886,
+ "learning_rate": 0.0006,
+ "loss": 7.278632640838623,
+ "step": 147
+ },
+ {
+ "epoch": 2.05589519650655,
+ "grad_norm": 0.028048858046531677,
+ "learning_rate": 0.0006,
+ "loss": 7.26558780670166,
+ "step": 148
+ },
+ {
+ "epoch": 2.069868995633188,
+ "grad_norm": 0.06547307223081589,
+ "learning_rate": 0.0006,
+ "loss": 7.32567834854126,
+ "step": 149
+ },
+ {
+ "epoch": 2.0838427947598253,
+ "grad_norm": 0.05969763547182083,
+ "learning_rate": 0.0006,
+ "loss": 7.395836353302002,
+ "step": 150
+ },
+ {
+ "epoch": 2.097816593886463,
+ "grad_norm": 0.032574281096458435,
+ "learning_rate": 0.0006,
+ "loss": 7.318150997161865,
+ "step": 151
+ },
+ {
+ "epoch": 2.1117903930131003,
+ "grad_norm": 0.029452962800860405,
+ "learning_rate": 0.0006,
+ "loss": 7.308066368103027,
+ "step": 152
+ },
+ {
+ "epoch": 2.125764192139738,
+ "grad_norm": 0.03161991387605667,
+ "learning_rate": 0.0006,
+ "loss": 7.306290626525879,
+ "step": 153
+ },
+ {
+ "epoch": 2.1397379912663754,
+ "grad_norm": 0.027096295729279518,
+ "learning_rate": 0.0006,
+ "loss": 7.297796249389648,
+ "step": 154
+ },
+ {
+ "epoch": 2.153711790393013,
+ "grad_norm": 0.028657056391239166,
+ "learning_rate": 0.0006,
+ "loss": 7.317424774169922,
+ "step": 155
+ },
+ {
+ "epoch": 2.1676855895196505,
+ "grad_norm": 0.02046995982527733,
+ "learning_rate": 0.0006,
+ "loss": 7.297986030578613,
+ "step": 156
+ },
+ {
+ "epoch": 2.1816593886462883,
+ "grad_norm": 0.02220369316637516,
+ "learning_rate": 0.0006,
+ "loss": 7.268237590789795,
+ "step": 157
+ },
+ {
+ "epoch": 2.1956331877729256,
+ "grad_norm": 0.018381357192993164,
+ "learning_rate": 0.0006,
+ "loss": 7.259415626525879,
+ "step": 158
+ },
+ {
+ "epoch": 2.2096069868995634,
+ "grad_norm": 0.0204413291066885,
+ "learning_rate": 0.0006,
+ "loss": 7.258173942565918,
+ "step": 159
+ },
+ {
+ "epoch": 2.2235807860262007,
+ "grad_norm": 0.015874288976192474,
+ "learning_rate": 0.0006,
+ "loss": 7.281834602355957,
+ "step": 160
+ },
+ {
+ "epoch": 2.2375545851528384,
+ "grad_norm": 0.01634068600833416,
+ "learning_rate": 0.0006,
+ "loss": 7.2469868659973145,
+ "step": 161
+ },
+ {
+ "epoch": 2.251528384279476,
+ "grad_norm": 0.01728747971355915,
+ "learning_rate": 0.0006,
+ "loss": 7.280995845794678,
+ "step": 162
+ },
+ {
+ "epoch": 2.2655021834061135,
+ "grad_norm": 0.016341188922524452,
+ "learning_rate": 0.0006,
+ "loss": 7.273357391357422,
+ "step": 163
+ },
+ {
+ "epoch": 2.279475982532751,
+ "grad_norm": 0.016125807538628578,
+ "learning_rate": 0.0006,
+ "loss": 7.273510932922363,
+ "step": 164
+ },
+ {
+ "epoch": 2.2934497816593886,
+ "grad_norm": 0.01629287749528885,
+ "learning_rate": 0.0006,
+ "loss": 7.267029762268066,
+ "step": 165
+ },
+ {
+ "epoch": 2.3074235807860264,
+ "grad_norm": 0.015296131372451782,
+ "learning_rate": 0.0006,
+ "loss": 7.211989402770996,
+ "step": 166
+ },
+ {
+ "epoch": 2.3213973799126637,
+ "grad_norm": 0.01709570363163948,
+ "learning_rate": 0.0006,
+ "loss": 7.242851734161377,
+ "step": 167
+ },
+ {
+ "epoch": 2.3353711790393015,
+ "grad_norm": 0.012552388943731785,
+ "learning_rate": 0.0006,
+ "loss": 7.241696834564209,
+ "step": 168
+ },
+ {
+ "epoch": 2.349344978165939,
+ "grad_norm": 0.014571291394531727,
+ "learning_rate": 0.0006,
+ "loss": 7.270681381225586,
+ "step": 169
+ },
+ {
+ "epoch": 2.3633187772925766,
+ "grad_norm": 0.013389437459409237,
+ "learning_rate": 0.0006,
+ "loss": 7.191600799560547,
+ "step": 170
+ },
+ {
+ "epoch": 2.377292576419214,
+ "grad_norm": 0.011895314790308475,
+ "learning_rate": 0.0006,
+ "loss": 7.213610649108887,
+ "step": 171
+ },
+ {
+ "epoch": 2.3912663755458516,
+ "grad_norm": 0.01345642190426588,
+ "learning_rate": 0.0006,
+ "loss": 7.193787574768066,
+ "step": 172
+ },
+ {
+ "epoch": 2.405240174672489,
+ "grad_norm": 0.0091794328764081,
+ "learning_rate": 0.0006,
+ "loss": 7.233364105224609,
+ "step": 173
+ },
+ {
+ "epoch": 2.4192139737991267,
+ "grad_norm": 0.013767086900770664,
+ "learning_rate": 0.0006,
+ "loss": 7.217272758483887,
+ "step": 174
+ },
+ {
+ "epoch": 2.433187772925764,
+ "grad_norm": 0.015134960412979126,
+ "learning_rate": 0.0006,
+ "loss": 7.192791938781738,
+ "step": 175
+ },
+ {
+ "epoch": 2.447161572052402,
+ "grad_norm": 0.01790648326277733,
+ "learning_rate": 0.0006,
+ "loss": 7.229467391967773,
+ "step": 176
+ },
+ {
+ "epoch": 2.461135371179039,
+ "grad_norm": 0.015362377278506756,
+ "learning_rate": 0.0006,
+ "loss": 7.180346488952637,
+ "step": 177
+ },
+ {
+ "epoch": 2.475109170305677,
+ "grad_norm": 0.010505498386919498,
+ "learning_rate": 0.0006,
+ "loss": 7.188833236694336,
+ "step": 178
+ },
+ {
+ "epoch": 2.489082969432314,
+ "grad_norm": 0.009129747748374939,
+ "learning_rate": 0.0006,
+ "loss": 7.203153133392334,
+ "step": 179
+ },
+ {
+ "epoch": 2.503056768558952,
+ "grad_norm": 0.01147883478552103,
+ "learning_rate": 0.0006,
+ "loss": 7.163824558258057,
+ "step": 180
+ },
+ {
+ "epoch": 2.5170305676855893,
+ "grad_norm": 0.011338942684233189,
+ "learning_rate": 0.0006,
+ "loss": 7.16287899017334,
+ "step": 181
+ },
+ {
+ "epoch": 2.531004366812227,
+ "grad_norm": 0.013277675025165081,
+ "learning_rate": 0.0006,
+ "loss": 7.155905723571777,
+ "step": 182
+ },
+ {
+ "epoch": 2.544978165938865,
+ "grad_norm": 0.015269347466528416,
+ "learning_rate": 0.0006,
+ "loss": 7.189371109008789,
+ "step": 183
+ },
+ {
+ "epoch": 2.558951965065502,
+ "grad_norm": 0.017905596643686295,
+ "learning_rate": 0.0006,
+ "loss": 7.177547931671143,
+ "step": 184
+ },
+ {
+ "epoch": 2.5729257641921395,
+ "grad_norm": 0.024431610479950905,
+ "learning_rate": 0.0006,
+ "loss": 7.192169189453125,
+ "step": 185
+ },
+ {
+ "epoch": 2.5868995633187772,
+ "grad_norm": 0.03656961768865585,
+ "learning_rate": 0.0006,
+ "loss": 7.124387741088867,
+ "step": 186
+ },
+ {
+ "epoch": 2.600873362445415,
+ "grad_norm": 0.04337048903107643,
+ "learning_rate": 0.0006,
+ "loss": 7.140231609344482,
+ "step": 187
+ },
+ {
+ "epoch": 2.6148471615720523,
+ "grad_norm": 0.04047228768467903,
+ "learning_rate": 0.0006,
+ "loss": 7.171927452087402,
+ "step": 188
+ },
+ {
+ "epoch": 2.62882096069869,
+ "grad_norm": 0.07825770974159241,
+ "learning_rate": 0.0006,
+ "loss": 7.207631587982178,
+ "step": 189
+ },
+ {
+ "epoch": 2.6427947598253274,
+ "grad_norm": 0.07118063420057297,
+ "learning_rate": 0.0006,
+ "loss": 7.209488391876221,
+ "step": 190
+ },
+ {
+ "epoch": 2.656768558951965,
+ "grad_norm": 0.05153293535113335,
+ "learning_rate": 0.0006,
+ "loss": 7.14793062210083,
+ "step": 191
+ },
+ {
+ "epoch": 2.6707423580786025,
+ "grad_norm": 0.03318839520215988,
+ "learning_rate": 0.0006,
+ "loss": 7.163010597229004,
+ "step": 192
+ },
+ {
+ "epoch": 2.6847161572052403,
+ "grad_norm": 0.029333539307117462,
+ "learning_rate": 0.0006,
+ "loss": 7.159955978393555,
+ "step": 193
+ },
+ {
+ "epoch": 2.6986899563318776,
+ "grad_norm": 0.029405880719423294,
+ "learning_rate": 0.0006,
+ "loss": 7.136178493499756,
+ "step": 194
+ },
+ {
+ "epoch": 2.7126637554585153,
+ "grad_norm": 0.037815988063812256,
+ "learning_rate": 0.0006,
+ "loss": 7.183186054229736,
+ "step": 195
+ },
+ {
+ "epoch": 2.726637554585153,
+ "grad_norm": 0.020307250320911407,
+ "learning_rate": 0.0006,
+ "loss": 7.133164882659912,
+ "step": 196
+ },
+ {
+ "epoch": 2.7406113537117904,
+ "grad_norm": 0.026048069819808006,
+ "learning_rate": 0.0006,
+ "loss": 7.124977111816406,
+ "step": 197
+ },
+ {
+ "epoch": 2.7545851528384278,
+ "grad_norm": 0.027316724881529808,
+ "learning_rate": 0.0006,
+ "loss": 7.07265567779541,
+ "step": 198
+ },
+ {
+ "epoch": 2.7685589519650655,
+ "grad_norm": 0.021629009395837784,
+ "learning_rate": 0.0006,
+ "loss": 7.150020599365234,
+ "step": 199
+ },
+ {
+ "epoch": 2.7825327510917033,
+ "grad_norm": 0.01605929434299469,
+ "learning_rate": 0.0006,
+ "loss": 7.10919189453125,
+ "step": 200
+ },
+ {
+ "epoch": 2.7965065502183406,
+ "grad_norm": 0.020800089463591576,
+ "learning_rate": 0.0006,
+ "loss": 7.132045269012451,
+ "step": 201
+ },
+ {
+ "epoch": 2.810480349344978,
+ "grad_norm": 0.019225774332880974,
+ "learning_rate": 0.0006,
+ "loss": 7.128364562988281,
+ "step": 202
+ },
+ {
+ "epoch": 2.8244541484716157,
+ "grad_norm": 0.01561372634023428,
+ "learning_rate": 0.0006,
+ "loss": 7.083669662475586,
+ "step": 203
+ },
+ {
+ "epoch": 2.8384279475982535,
+ "grad_norm": 0.021969519555568695,
+ "learning_rate": 0.0006,
+ "loss": 7.050841331481934,
+ "step": 204
+ },
+ {
+ "epoch": 2.8524017467248908,
+ "grad_norm": 0.021453695371747017,
+ "learning_rate": 0.0006,
+ "loss": 7.097085475921631,
+ "step": 205
+ },
+ {
+ "epoch": 2.8663755458515285,
+ "grad_norm": 0.022836022078990936,
+ "learning_rate": 0.0006,
+ "loss": 7.10405969619751,
+ "step": 206
+ },
+ {
+ "epoch": 2.880349344978166,
+ "grad_norm": 0.02430851384997368,
+ "learning_rate": 0.0006,
+ "loss": 7.045181751251221,
+ "step": 207
+ },
+ {
+ "epoch": 2.8943231441048036,
+ "grad_norm": 0.020166993141174316,
+ "learning_rate": 0.0006,
+ "loss": 7.068183898925781,
+ "step": 208
+ },
+ {
+ "epoch": 2.908296943231441,
+ "grad_norm": 0.0118505684658885,
+ "learning_rate": 0.0006,
+ "loss": 7.07132625579834,
+ "step": 209
+ },
+ {
+ "epoch": 2.9222707423580787,
+ "grad_norm": 0.023772427812218666,
+ "learning_rate": 0.0006,
+ "loss": 7.055238723754883,
+ "step": 210
+ },
+ {
+ "epoch": 2.936244541484716,
+ "grad_norm": 0.02350712940096855,
+ "learning_rate": 0.0006,
+ "loss": 7.055898666381836,
+ "step": 211
+ },
+ {
+ "epoch": 2.950218340611354,
+ "grad_norm": 0.017244907096028328,
+ "learning_rate": 0.0006,
+ "loss": 7.0099711418151855,
+ "step": 212
+ },
+ {
+ "epoch": 2.964192139737991,
+ "grad_norm": 0.021565575152635574,
+ "learning_rate": 0.0006,
+ "loss": 6.991751670837402,
+ "step": 213
+ },
+ {
+ "epoch": 2.978165938864629,
+ "grad_norm": 0.03241860866546631,
+ "learning_rate": 0.0006,
+ "loss": 7.005980491638184,
+ "step": 214
+ },
+ {
+ "epoch": 2.992139737991266,
+ "grad_norm": 0.04613726586103439,
+ "learning_rate": 0.0006,
+ "loss": 7.027859687805176,
+ "step": 215
+ },
+ {
+ "epoch": 3.0,
+ "grad_norm": 0.06404894590377808,
+ "learning_rate": 0.0006,
+ "loss": 7.123725891113281,
+ "step": 216
+ },
+ {
+ "epoch": 3.0,
+ "eval_loss": 7.070647716522217,
+ "eval_runtime": 58.5801,
+ "eval_samples_per_second": 41.687,
+ "eval_steps_per_second": 1.314,
+ "step": 216
+ },
+ {
+ "epoch": 3.0139737991266378,
+ "grad_norm": 0.05542432889342308,
+ "learning_rate": 0.0006,
+ "loss": 7.078299522399902,
+ "step": 217
+ },
+ {
+ "epoch": 3.027947598253275,
+ "grad_norm": 0.026484526693820953,
+ "learning_rate": 0.0006,
+ "loss": 6.961159706115723,
+ "step": 218
+ },
+ {
+ "epoch": 3.041921397379913,
+ "grad_norm": 0.028276391327381134,
+ "learning_rate": 0.0006,
+ "loss": 7.019961833953857,
+ "step": 219
+ },
+ {
+ "epoch": 3.05589519650655,
+ "grad_norm": 0.024486595764756203,
+ "learning_rate": 0.0006,
+ "loss": 7.039677619934082,
+ "step": 220
+ },
+ {
+ "epoch": 3.069868995633188,
+ "grad_norm": 0.017620893195271492,
+ "learning_rate": 0.0006,
+ "loss": 6.9270782470703125,
+ "step": 221
+ },
+ {
+ "epoch": 3.0838427947598253,
+ "grad_norm": 0.02547302283346653,
+ "learning_rate": 0.0006,
+ "loss": 6.924233436584473,
+ "step": 222
+ },
+ {
+ "epoch": 3.097816593886463,
+ "grad_norm": 0.02539009042084217,
+ "learning_rate": 0.0006,
+ "loss": 6.973979949951172,
+ "step": 223
+ },
+ {
+ "epoch": 3.1117903930131003,
+ "grad_norm": 0.03256227448582649,
+ "learning_rate": 0.0006,
+ "loss": 6.933725357055664,
+ "step": 224
+ },
+ {
+ "epoch": 3.125764192139738,
+ "grad_norm": 0.04651800915598869,
+ "learning_rate": 0.0006,
+ "loss": 6.951044082641602,
+ "step": 225
+ },
+ {
+ "epoch": 3.1397379912663754,
+ "grad_norm": 0.04822975769639015,
+ "learning_rate": 0.0006,
+ "loss": 6.9614410400390625,
+ "step": 226
+ },
+ {
+ "epoch": 3.153711790393013,
+ "grad_norm": 0.031165020540356636,
+ "learning_rate": 0.0006,
+ "loss": 6.91193151473999,
+ "step": 227
+ },
+ {
+ "epoch": 3.1676855895196505,
+ "grad_norm": 0.029754292219877243,
+ "learning_rate": 0.0006,
+ "loss": 6.91802978515625,
+ "step": 228
+ },
+ {
+ "epoch": 3.1816593886462883,
+ "grad_norm": 0.02193152718245983,
+ "learning_rate": 0.0006,
+ "loss": 6.928090572357178,
+ "step": 229
+ },
+ {
+ "epoch": 3.1956331877729256,
+ "grad_norm": 0.02428470179438591,
+ "learning_rate": 0.0006,
+ "loss": 6.90396785736084,
+ "step": 230
+ },
+ {
+ "epoch": 3.2096069868995634,
+ "grad_norm": 0.023374345153570175,
+ "learning_rate": 0.0006,
+ "loss": 6.885858535766602,
+ "step": 231
+ },
+ {
+ "epoch": 3.2235807860262007,
+ "grad_norm": 0.023405000567436218,
+ "learning_rate": 0.0006,
+ "loss": 6.8855767250061035,
+ "step": 232
+ },
+ {
+ "epoch": 3.2375545851528384,
+ "grad_norm": 0.017277134582400322,
+ "learning_rate": 0.0006,
+ "loss": 6.901449203491211,
+ "step": 233
+ },
+ {
+ "epoch": 3.251528384279476,
+ "grad_norm": 0.019496750086545944,
+ "learning_rate": 0.0006,
+ "loss": 6.885687828063965,
+ "step": 234
+ },
+ {
+ "epoch": 3.2655021834061135,
+ "grad_norm": 0.016893276944756508,
+ "learning_rate": 0.0006,
+ "loss": 6.872369766235352,
+ "step": 235
+ },
+ {
+ "epoch": 3.279475982532751,
+ "grad_norm": 0.02058715932071209,
+ "learning_rate": 0.0006,
+ "loss": 6.823280334472656,
+ "step": 236
+ },
+ {
+ "epoch": 3.2934497816593886,
+ "grad_norm": 0.015530755743384361,
+ "learning_rate": 0.0006,
+ "loss": 6.860499858856201,
+ "step": 237
+ },
+ {
+ "epoch": 3.3074235807860264,
+ "grad_norm": 0.01934926211833954,
+ "learning_rate": 0.0006,
+ "loss": 6.817448616027832,
+ "step": 238
+ },
+ {
+ "epoch": 3.3213973799126637,
+ "grad_norm": 0.017413552850484848,
+ "learning_rate": 0.0006,
+ "loss": 6.881955146789551,
+ "step": 239
+ },
+ {
+ "epoch": 3.3353711790393015,
+ "grad_norm": 0.026478875428438187,
+ "learning_rate": 0.0006,
+ "loss": 6.827404499053955,
+ "step": 240
+ },
+ {
+ "epoch": 3.349344978165939,
+ "grad_norm": 0.032611701637506485,
+ "learning_rate": 0.0006,
+ "loss": 6.78285551071167,
+ "step": 241
+ },
+ {
+ "epoch": 3.3633187772925766,
+ "grad_norm": 0.041900306940078735,
+ "learning_rate": 0.0006,
+ "loss": 6.783053874969482,
+ "step": 242
+ },
+ {
+ "epoch": 3.377292576419214,
+ "grad_norm": 0.04149497672915459,
+ "learning_rate": 0.0006,
+ "loss": 6.745220184326172,
+ "step": 243
+ },
+ {
+ "epoch": 3.3912663755458516,
+ "grad_norm": 0.023949826136231422,
+ "learning_rate": 0.0006,
+ "loss": 6.841533184051514,
+ "step": 244
+ },
+ {
+ "epoch": 3.405240174672489,
+ "grad_norm": 0.028601842001080513,
+ "learning_rate": 0.0006,
+ "loss": 6.852852821350098,
+ "step": 245
+ },
+ {
+ "epoch": 3.4192139737991267,
+ "grad_norm": 0.027365155518054962,
+ "learning_rate": 0.0006,
+ "loss": 6.74376106262207,
+ "step": 246
+ },
+ {
+ "epoch": 3.433187772925764,
+ "grad_norm": 0.02856568619608879,
+ "learning_rate": 0.0006,
+ "loss": 6.74906063079834,
+ "step": 247
+ },
+ {
+ "epoch": 3.447161572052402,
+ "grad_norm": 0.019172416999936104,
+ "learning_rate": 0.0006,
+ "loss": 6.809206008911133,
+ "step": 248
+ },
+ {
+ "epoch": 3.461135371179039,
+ "grad_norm": 0.02161746844649315,
+ "learning_rate": 0.0006,
+ "loss": 6.789888381958008,
+ "step": 249
+ },
+ {
+ "epoch": 3.475109170305677,
+ "grad_norm": 0.020273666828870773,
+ "learning_rate": 0.0006,
+ "loss": 6.73512077331543,
+ "step": 250
+ },
+ {
+ "epoch": 3.489082969432314,
+ "grad_norm": 0.022329283878207207,
+ "learning_rate": 0.0006,
+ "loss": 6.826148509979248,
+ "step": 251
+ },
+ {
+ "epoch": 3.503056768558952,
+ "grad_norm": 0.02284284122288227,
+ "learning_rate": 0.0006,
+ "loss": 6.748521327972412,
+ "step": 252
+ },
+ {
+ "epoch": 3.5170305676855893,
+ "grad_norm": 0.026368247345089912,
+ "learning_rate": 0.0006,
+ "loss": 6.645879745483398,
+ "step": 253
+ },
+ {
+ "epoch": 3.531004366812227,
+ "grad_norm": 0.03627244383096695,
+ "learning_rate": 0.0006,
+ "loss": 6.666064739227295,
+ "step": 254
+ },
+ {
+ "epoch": 3.544978165938865,
+ "grad_norm": 0.04140935093164444,
+ "learning_rate": 0.0006,
+ "loss": 6.712275981903076,
+ "step": 255
+ },
+ {
+ "epoch": 3.558951965065502,
+ "grad_norm": 0.0390462800860405,
+ "learning_rate": 0.0006,
+ "loss": 6.665900230407715,
+ "step": 256
+ },
+ {
+ "epoch": 3.5729257641921395,
+ "grad_norm": 0.05178583785891533,
+ "learning_rate": 0.0006,
+ "loss": 6.753298759460449,
+ "step": 257
+ },
+ {
+ "epoch": 3.5868995633187772,
+ "grad_norm": 0.054328061640262604,
+ "learning_rate": 0.0006,
+ "loss": 6.732633113861084,
+ "step": 258
+ },
+ {
+ "epoch": 3.600873362445415,
+ "grad_norm": 0.04590460658073425,
+ "learning_rate": 0.0006,
+ "loss": 6.65250301361084,
+ "step": 259
+ },
+ {
+ "epoch": 3.6148471615720523,
+ "grad_norm": 0.044331666082143784,
+ "learning_rate": 0.0006,
+ "loss": 6.68654203414917,
+ "step": 260
+ },
+ {
+ "epoch": 3.62882096069869,
+ "grad_norm": 0.05308730527758598,
+ "learning_rate": 0.0006,
+ "loss": 6.642482757568359,
+ "step": 261
+ },
+ {
+ "epoch": 3.6427947598253274,
+ "grad_norm": 0.03999016433954239,
+ "learning_rate": 0.0006,
+ "loss": 6.625227451324463,
+ "step": 262
+ },
+ {
+ "epoch": 3.656768558951965,
+ "grad_norm": 0.04088086634874344,
+ "learning_rate": 0.0006,
+ "loss": 6.67623233795166,
+ "step": 263
+ },
+ {
+ "epoch": 3.6707423580786025,
+ "grad_norm": 0.03615636005997658,
+ "learning_rate": 0.0006,
+ "loss": 6.6869001388549805,
+ "step": 264
+ },
+ {
+ "epoch": 3.6847161572052403,
+ "grad_norm": 0.02817360684275627,
+ "learning_rate": 0.0006,
+ "loss": 6.674601078033447,
+ "step": 265
+ },
+ {
+ "epoch": 3.6986899563318776,
+ "grad_norm": 0.032586123794317245,
+ "learning_rate": 0.0006,
+ "loss": 6.612994194030762,
+ "step": 266
+ },
+ {
+ "epoch": 3.7126637554585153,
+ "grad_norm": 0.031117867678403854,
+ "learning_rate": 0.0006,
+ "loss": 6.6244354248046875,
+ "step": 267
+ },
+ {
+ "epoch": 3.726637554585153,
+ "grad_norm": 0.028330687433481216,
+ "learning_rate": 0.0006,
+ "loss": 6.669557094573975,
+ "step": 268
+ },
+ {
+ "epoch": 3.7406113537117904,
+ "grad_norm": 0.021489202976226807,
+ "learning_rate": 0.0006,
+ "loss": 6.632035732269287,
+ "step": 269
+ },
+ {
+ "epoch": 3.7545851528384278,
+ "grad_norm": 0.024269424378871918,
+ "learning_rate": 0.0006,
+ "loss": 6.660867691040039,
+ "step": 270
+ },
+ {
+ "epoch": 3.7685589519650655,
+ "grad_norm": 0.02237873338162899,
+ "learning_rate": 0.0006,
+ "loss": 6.602587699890137,
+ "step": 271
+ },
+ {
+ "epoch": 3.7825327510917033,
+ "grad_norm": 0.026503929868340492,
+ "learning_rate": 0.0006,
+ "loss": 6.599200248718262,
+ "step": 272
+ },
+ {
+ "epoch": 3.7965065502183406,
+ "grad_norm": 0.03687124326825142,
+ "learning_rate": 0.0006,
+ "loss": 6.567296504974365,
+ "step": 273
+ },
+ {
+ "epoch": 3.810480349344978,
+ "grad_norm": 0.04708952456712723,
+ "learning_rate": 0.0006,
+ "loss": 6.621265411376953,
+ "step": 274
+ },
+ {
+ "epoch": 3.8244541484716157,
+ "grad_norm": 0.041846614331007004,
+ "learning_rate": 0.0006,
+ "loss": 6.550536155700684,
+ "step": 275
+ },
+ {
+ "epoch": 3.8384279475982535,
+ "grad_norm": 0.031249675899744034,
+ "learning_rate": 0.0006,
+ "loss": 6.54278564453125,
+ "step": 276
+ },
+ {
+ "epoch": 3.8524017467248908,
+ "grad_norm": 0.023082256317138672,
+ "learning_rate": 0.0006,
+ "loss": 6.544787406921387,
+ "step": 277
+ },
+ {
+ "epoch": 3.8663755458515285,
+ "grad_norm": 0.020995570346713066,
+ "learning_rate": 0.0006,
+ "loss": 6.518143653869629,
+ "step": 278
+ },
+ {
+ "epoch": 3.880349344978166,
+ "grad_norm": 0.020839324221014977,
+ "learning_rate": 0.0006,
+ "loss": 6.546667098999023,
+ "step": 279
+ },
+ {
+ "epoch": 3.8943231441048036,
+ "grad_norm": 0.018377432599663734,
+ "learning_rate": 0.0006,
+ "loss": 6.478307723999023,
+ "step": 280
+ },
+ {
+ "epoch": 3.908296943231441,
+ "grad_norm": 0.021713724359869957,
+ "learning_rate": 0.0006,
+ "loss": 6.485574245452881,
+ "step": 281
+ },
+ {
+ "epoch": 3.9222707423580787,
+ "grad_norm": 0.01359301246702671,
+ "learning_rate": 0.0006,
+ "loss": 6.570268630981445,
+ "step": 282
+ },
+ {
+ "epoch": 3.936244541484716,
+ "grad_norm": 0.016233332455158234,
+ "learning_rate": 0.0006,
+ "loss": 6.57150936126709,
+ "step": 283
+ },
+ {
+ "epoch": 3.950218340611354,
+ "grad_norm": 0.020687608048319817,
+ "learning_rate": 0.0006,
+ "loss": 6.527956008911133,
+ "step": 284
+ },
+ {
+ "epoch": 3.964192139737991,
+ "grad_norm": 0.02032964862883091,
+ "learning_rate": 0.0006,
+ "loss": 6.468169689178467,
+ "step": 285
+ },
+ {
+ "epoch": 3.978165938864629,
+ "grad_norm": 0.016676288098096848,
+ "learning_rate": 0.0006,
+ "loss": 6.448418617248535,
+ "step": 286
+ },
+ {
+ "epoch": 3.992139737991266,
+ "grad_norm": 0.0168539360165596,
+ "learning_rate": 0.0006,
+ "loss": 6.461180686950684,
+ "step": 287
+ },
+ {
+ "epoch": 4.0,
+ "grad_norm": 0.02666233666241169,
+ "learning_rate": 0.0006,
+ "loss": 6.344893932342529,
+ "step": 288
+ },
+ {
+ "epoch": 4.0,
+ "eval_loss": 6.5384979248046875,
+ "eval_runtime": 58.814,
+ "eval_samples_per_second": 41.521,
+ "eval_steps_per_second": 1.309,
+ "step": 288
+ },
+ {
+ "epoch": 4.013973799126638,
+ "grad_norm": 0.04154708981513977,
+ "learning_rate": 0.0006,
+ "loss": 6.41855525970459,
+ "step": 289
+ },
+ {
+ "epoch": 4.0279475982532755,
+ "grad_norm": 0.04895668104290962,
+ "learning_rate": 0.0006,
+ "loss": 6.499302864074707,
+ "step": 290
+ },
+ {
+ "epoch": 4.041921397379912,
+ "grad_norm": 0.034464482218027115,
+ "learning_rate": 0.0006,
+ "loss": 6.438611030578613,
+ "step": 291
+ },
+ {
+ "epoch": 4.05589519650655,
+ "grad_norm": 0.06845773756504059,
+ "learning_rate": 0.0006,
+ "loss": 6.538361072540283,
+ "step": 292
+ },
+ {
+ "epoch": 4.069868995633188,
+ "grad_norm": 0.053768858313560486,
+ "learning_rate": 0.0006,
+ "loss": 6.527524948120117,
+ "step": 293
+ },
+ {
+ "epoch": 4.083842794759826,
+ "grad_norm": 0.039037931710481644,
+ "learning_rate": 0.0006,
+ "loss": 6.503847122192383,
+ "step": 294
+ },
+ {
+ "epoch": 4.097816593886463,
+ "grad_norm": 0.03962196782231331,
+ "learning_rate": 0.0006,
+ "loss": 6.475178241729736,
+ "step": 295
+ },
+ {
+ "epoch": 4.1117903930131,
+ "grad_norm": 0.049101535230875015,
+ "learning_rate": 0.0006,
+ "loss": 6.474587917327881,
+ "step": 296
+ },
+ {
+ "epoch": 4.125764192139738,
+ "grad_norm": 0.02830282226204872,
+ "learning_rate": 0.0006,
+ "loss": 6.45065975189209,
+ "step": 297
+ },
+ {
+ "epoch": 4.139737991266376,
+ "grad_norm": 0.023153482005000114,
+ "learning_rate": 0.0006,
+ "loss": 6.5218706130981445,
+ "step": 298
+ },
+ {
+ "epoch": 4.153711790393013,
+ "grad_norm": 0.025064002722501755,
+ "learning_rate": 0.0006,
+ "loss": 6.456475257873535,
+ "step": 299
+ },
+ {
+ "epoch": 4.1676855895196505,
+ "grad_norm": 0.02357092685997486,
+ "learning_rate": 0.0006,
+ "loss": 6.485952377319336,
+ "step": 300
+ },
+ {
+ "epoch": 4.181659388646288,
+ "grad_norm": 0.021832305938005447,
+ "learning_rate": 0.0006,
+ "loss": 6.425506591796875,
+ "step": 301
+ },
+ {
+ "epoch": 4.195633187772926,
+ "grad_norm": 0.0227900929749012,
+ "learning_rate": 0.0006,
+ "loss": 6.439155578613281,
+ "step": 302
+ },
+ {
+ "epoch": 4.209606986899563,
+ "grad_norm": 0.015961607918143272,
+ "learning_rate": 0.0006,
+ "loss": 6.460862636566162,
+ "step": 303
+ },
+ {
+ "epoch": 4.223580786026201,
+ "grad_norm": 0.017876390367746353,
+ "learning_rate": 0.0006,
+ "loss": 6.44564151763916,
+ "step": 304
+ },
+ {
+ "epoch": 4.2375545851528384,
+ "grad_norm": 0.013703204691410065,
+ "learning_rate": 0.0006,
+ "loss": 6.418414115905762,
+ "step": 305
+ },
+ {
+ "epoch": 4.251528384279476,
+ "grad_norm": 0.015240306034684181,
+ "learning_rate": 0.0006,
+ "loss": 6.411238670349121,
+ "step": 306
+ },
+ {
+ "epoch": 4.265502183406113,
+ "grad_norm": 0.01587662845849991,
+ "learning_rate": 0.0006,
+ "loss": 6.395642280578613,
+ "step": 307
+ },
+ {
+ "epoch": 4.279475982532751,
+ "grad_norm": 0.014650700613856316,
+ "learning_rate": 0.0006,
+ "loss": 6.302793025970459,
+ "step": 308
+ },
+ {
+ "epoch": 4.293449781659389,
+ "grad_norm": 0.016533225774765015,
+ "learning_rate": 0.0006,
+ "loss": 6.380588054656982,
+ "step": 309
+ },
+ {
+ "epoch": 4.307423580786026,
+ "grad_norm": 0.019963741302490234,
+ "learning_rate": 0.0006,
+ "loss": 6.345336437225342,
+ "step": 310
+ },
+ {
+ "epoch": 4.321397379912664,
+ "grad_norm": 0.02237936295568943,
+ "learning_rate": 0.0006,
+ "loss": 6.387770652770996,
+ "step": 311
+ },
+ {
+ "epoch": 4.335371179039301,
+ "grad_norm": 0.026658328250050545,
+ "learning_rate": 0.0006,
+ "loss": 6.323662757873535,
+ "step": 312
+ },
+ {
+ "epoch": 4.349344978165939,
+ "grad_norm": 0.02852284163236618,
+ "learning_rate": 0.0006,
+ "loss": 6.382204055786133,
+ "step": 313
+ },
+ {
+ "epoch": 4.3633187772925766,
+ "grad_norm": 0.028815090656280518,
+ "learning_rate": 0.0006,
+ "loss": 6.402390480041504,
+ "step": 314
+ },
+ {
+ "epoch": 4.377292576419214,
+ "grad_norm": 0.029393529519438744,
+ "learning_rate": 0.0006,
+ "loss": 6.280278205871582,
+ "step": 315
+ },
+ {
+ "epoch": 4.391266375545851,
+ "grad_norm": 0.02186041697859764,
+ "learning_rate": 0.0006,
+ "loss": 6.302707672119141,
+ "step": 316
+ },
+ {
+ "epoch": 4.405240174672489,
+ "grad_norm": 0.02071218006312847,
+ "learning_rate": 0.0006,
+ "loss": 6.3365020751953125,
+ "step": 317
+ },
+ {
+ "epoch": 4.419213973799127,
+ "grad_norm": 0.02779117226600647,
+ "learning_rate": 0.0006,
+ "loss": 6.252157211303711,
+ "step": 318
+ },
+ {
+ "epoch": 4.4331877729257645,
+ "grad_norm": 0.03470654785633087,
+ "learning_rate": 0.0006,
+ "loss": 6.319070339202881,
+ "step": 319
+ },
+ {
+ "epoch": 4.447161572052401,
+ "grad_norm": 0.037067804485559464,
+ "learning_rate": 0.0006,
+ "loss": 6.314116477966309,
+ "step": 320
+ },
+ {
+ "epoch": 4.461135371179039,
+ "grad_norm": 0.032182756811380386,
+ "learning_rate": 0.0006,
+ "loss": 6.267298698425293,
+ "step": 321
+ },
+ {
+ "epoch": 4.475109170305677,
+ "grad_norm": 0.026305217295885086,
+ "learning_rate": 0.0006,
+ "loss": 6.331688404083252,
+ "step": 322
+ },
+ {
+ "epoch": 4.489082969432315,
+ "grad_norm": 0.027228357270359993,
+ "learning_rate": 0.0006,
+ "loss": 6.31139612197876,
+ "step": 323
+ },
+ {
+ "epoch": 4.503056768558952,
+ "grad_norm": 0.025781169533729553,
+ "learning_rate": 0.0006,
+ "loss": 6.194684982299805,
+ "step": 324
+ },
+ {
+ "epoch": 4.517030567685589,
+ "grad_norm": 0.029186977073550224,
+ "learning_rate": 0.0006,
+ "loss": 6.31306266784668,
+ "step": 325
+ },
+ {
+ "epoch": 4.531004366812227,
+ "grad_norm": 0.023608777672052383,
+ "learning_rate": 0.0006,
+ "loss": 6.285243034362793,
+ "step": 326
+ },
+ {
+ "epoch": 4.544978165938865,
+ "grad_norm": 0.02267594076693058,
+ "learning_rate": 0.0006,
+ "loss": 6.30342960357666,
+ "step": 327
+ },
+ {
+ "epoch": 4.558951965065502,
+ "grad_norm": 0.01926310732960701,
+ "learning_rate": 0.0006,
+ "loss": 6.280606269836426,
+ "step": 328
+ },
+ {
+ "epoch": 4.5729257641921395,
+ "grad_norm": 0.026484837755560875,
+ "learning_rate": 0.0006,
+ "loss": 6.237980365753174,
+ "step": 329
+ },
+ {
+ "epoch": 4.586899563318777,
+ "grad_norm": 0.026228854432702065,
+ "learning_rate": 0.0006,
+ "loss": 6.3018293380737305,
+ "step": 330
+ },
+ {
+ "epoch": 4.600873362445415,
+ "grad_norm": 0.022096967324614525,
+ "learning_rate": 0.0006,
+ "loss": 6.253863334655762,
+ "step": 331
+ },
+ {
+ "epoch": 4.614847161572053,
+ "grad_norm": 0.027223890647292137,
+ "learning_rate": 0.0006,
+ "loss": 6.166882038116455,
+ "step": 332
+ },
+ {
+ "epoch": 4.62882096069869,
+ "grad_norm": 0.03160124272108078,
+ "learning_rate": 0.0006,
+ "loss": 6.230309963226318,
+ "step": 333
+ },
+ {
+ "epoch": 4.642794759825327,
+ "grad_norm": 0.04110539332032204,
+ "learning_rate": 0.0006,
+ "loss": 6.2334818840026855,
+ "step": 334
+ },
+ {
+ "epoch": 4.656768558951965,
+ "grad_norm": 0.04355933889746666,
+ "learning_rate": 0.0006,
+ "loss": 6.160956382751465,
+ "step": 335
+ },
+ {
+ "epoch": 4.670742358078603,
+ "grad_norm": 0.03611086308956146,
+ "learning_rate": 0.0006,
+ "loss": 6.23399543762207,
+ "step": 336
+ },
+ {
+ "epoch": 4.68471615720524,
+ "grad_norm": 0.051811400800943375,
+ "learning_rate": 0.0006,
+ "loss": 6.241855144500732,
+ "step": 337
+ },
+ {
+ "epoch": 4.698689956331878,
+ "grad_norm": 0.057434193789958954,
+ "learning_rate": 0.0006,
+ "loss": 6.220993995666504,
+ "step": 338
+ },
+ {
+ "epoch": 4.712663755458515,
+ "grad_norm": 0.04420953616499901,
+ "learning_rate": 0.0006,
+ "loss": 6.300461292266846,
+ "step": 339
+ },
+ {
+ "epoch": 4.726637554585153,
+ "grad_norm": 0.03515457361936569,
+ "learning_rate": 0.0006,
+ "loss": 6.269349098205566,
+ "step": 340
+ },
+ {
+ "epoch": 4.74061135371179,
+ "grad_norm": 0.03222600743174553,
+ "learning_rate": 0.0006,
+ "loss": 6.248613357543945,
+ "step": 341
+ },
+ {
+ "epoch": 4.754585152838428,
+ "grad_norm": 0.0384046845138073,
+ "learning_rate": 0.0006,
+ "loss": 6.266979217529297,
+ "step": 342
+ },
+ {
+ "epoch": 4.7685589519650655,
+ "grad_norm": 0.0514867827296257,
+ "learning_rate": 0.0006,
+ "loss": 6.234103202819824,
+ "step": 343
+ },
+ {
+ "epoch": 4.782532751091703,
+ "grad_norm": 0.050747793167829514,
+ "learning_rate": 0.0006,
+ "loss": 6.322582244873047,
+ "step": 344
+ },
+ {
+ "epoch": 4.796506550218341,
+ "grad_norm": 0.040878988802433014,
+ "learning_rate": 0.0006,
+ "loss": 6.212509632110596,
+ "step": 345
+ },
+ {
+ "epoch": 4.810480349344978,
+ "grad_norm": 0.034655820578336716,
+ "learning_rate": 0.0006,
+ "loss": 6.265879154205322,
+ "step": 346
+ },
+ {
+ "epoch": 4.824454148471616,
+ "grad_norm": 0.04202224686741829,
+ "learning_rate": 0.0006,
+ "loss": 6.2896409034729,
+ "step": 347
+ },
+ {
+ "epoch": 4.8384279475982535,
+ "grad_norm": 0.03632277995347977,
+ "learning_rate": 0.0006,
+ "loss": 6.253917217254639,
+ "step": 348
+ },
+ {
+ "epoch": 4.85240174672489,
+ "grad_norm": 0.028645765036344528,
+ "learning_rate": 0.0006,
+ "loss": 6.207965850830078,
+ "step": 349
+ },
+ {
+ "epoch": 4.866375545851528,
+ "grad_norm": 0.025305170565843582,
+ "learning_rate": 0.0006,
+ "loss": 6.168068885803223,
+ "step": 350
+ },
+ {
+ "epoch": 4.880349344978166,
+ "grad_norm": 0.022137803956866264,
+ "learning_rate": 0.0006,
+ "loss": 6.235283374786377,
+ "step": 351
+ },
+ {
+ "epoch": 4.894323144104804,
+ "grad_norm": 0.01814538985490799,
+ "learning_rate": 0.0006,
+ "loss": 6.259641647338867,
+ "step": 352
+ },
+ {
+ "epoch": 4.908296943231441,
+ "grad_norm": 0.018465086817741394,
+ "learning_rate": 0.0006,
+ "loss": 6.111738681793213,
+ "step": 353
+ },
+ {
+ "epoch": 4.922270742358078,
+ "grad_norm": 0.018638933077454567,
+ "learning_rate": 0.0006,
+ "loss": 6.225642204284668,
+ "step": 354
+ },
+ {
+ "epoch": 4.936244541484716,
+ "grad_norm": 0.016461260616779327,
+ "learning_rate": 0.0006,
+ "loss": 6.140647888183594,
+ "step": 355
+ },
+ {
+ "epoch": 4.950218340611354,
+ "grad_norm": 0.015458385460078716,
+ "learning_rate": 0.0006,
+ "loss": 6.155224800109863,
+ "step": 356
+ },
+ {
+ "epoch": 4.964192139737992,
+ "grad_norm": 0.014934048056602478,
+ "learning_rate": 0.0006,
+ "loss": 6.101555824279785,
+ "step": 357
+ },
+ {
+ "epoch": 4.978165938864628,
+ "grad_norm": 0.013979545794427395,
+ "learning_rate": 0.0006,
+ "loss": 6.161718845367432,
+ "step": 358
+ },
+ {
+ "epoch": 4.992139737991266,
+ "grad_norm": 0.01640394888818264,
+ "learning_rate": 0.0006,
+ "loss": 6.055768966674805,
+ "step": 359
+ },
+ {
+ "epoch": 5.0,
+ "grad_norm": 0.015135680325329304,
+ "learning_rate": 0.0006,
+ "loss": 6.169695854187012,
+ "step": 360
+ },
+ {
+ "epoch": 5.0,
+ "eval_loss": 6.16408109664917,
+ "eval_runtime": 59.3649,
+ "eval_samples_per_second": 41.135,
+ "eval_steps_per_second": 1.297,
+ "step": 360
+ },
+ {
+ "epoch": 5.013973799126638,
+ "grad_norm": 0.012849048711359501,
+ "learning_rate": 0.0006,
+ "loss": 6.055830001831055,
+ "step": 361
+ },
+ {
+ "epoch": 5.0279475982532755,
+ "grad_norm": 0.018169576302170753,
+ "learning_rate": 0.0006,
+ "loss": 6.082655429840088,
+ "step": 362
+ },
+ {
+ "epoch": 5.041921397379912,
+ "grad_norm": 0.024293430149555206,
+ "learning_rate": 0.0006,
+ "loss": 6.109032154083252,
+ "step": 363
+ },
+ {
+ "epoch": 5.05589519650655,
+ "grad_norm": 0.03266787901520729,
+ "learning_rate": 0.0006,
+ "loss": 6.030613899230957,
+ "step": 364
+ },
+ {
+ "epoch": 5.069868995633188,
+ "grad_norm": 0.03572266176342964,
+ "learning_rate": 0.0006,
+ "loss": 6.050390720367432,
+ "step": 365
+ },
+ {
+ "epoch": 5.083842794759826,
+ "grad_norm": 0.032625213265419006,
+ "learning_rate": 0.0006,
+ "loss": 6.169313430786133,
+ "step": 366
+ },
+ {
+ "epoch": 5.097816593886463,
+ "grad_norm": 0.04051872715353966,
+ "learning_rate": 0.0006,
+ "loss": 6.10398006439209,
+ "step": 367
+ },
+ {
+ "epoch": 5.1117903930131,
+ "grad_norm": 0.0351213738322258,
+ "learning_rate": 0.0006,
+ "loss": 6.121973514556885,
+ "step": 368
+ },
+ {
+ "epoch": 5.125764192139738,
+ "grad_norm": 0.03523759916424751,
+ "learning_rate": 0.0006,
+ "loss": 6.0626654624938965,
+ "step": 369
+ },
+ {
+ "epoch": 5.139737991266376,
+ "grad_norm": 0.02706182189285755,
+ "learning_rate": 0.0006,
+ "loss": 6.049354076385498,
+ "step": 370
+ },
+ {
+ "epoch": 5.153711790393013,
+ "grad_norm": 0.024212589487433434,
+ "learning_rate": 0.0006,
+ "loss": 6.034826278686523,
+ "step": 371
+ },
+ {
+ "epoch": 5.1676855895196505,
+ "grad_norm": 0.024798082187771797,
+ "learning_rate": 0.0006,
+ "loss": 6.004058837890625,
+ "step": 372
+ },
+ {
+ "epoch": 5.181659388646288,
+ "grad_norm": 0.02325567416846752,
+ "learning_rate": 0.0006,
+ "loss": 5.986461639404297,
+ "step": 373
+ },
+ {
+ "epoch": 5.195633187772926,
+ "grad_norm": 0.019060570746660233,
+ "learning_rate": 0.0006,
+ "loss": 6.059736251831055,
+ "step": 374
+ },
+ {
+ "epoch": 5.209606986899563,
+ "grad_norm": 0.016822976991534233,
+ "learning_rate": 0.0006,
+ "loss": 6.072957515716553,
+ "step": 375
+ },
+ {
+ "epoch": 5.223580786026201,
+ "grad_norm": 0.017218658700585365,
+ "learning_rate": 0.0006,
+ "loss": 6.031739234924316,
+ "step": 376
+ },
+ {
+ "epoch": 5.2375545851528384,
+ "grad_norm": 0.014889383688569069,
+ "learning_rate": 0.0006,
+ "loss": 6.025674343109131,
+ "step": 377
+ },
+ {
+ "epoch": 5.251528384279476,
+ "grad_norm": 0.01708107627928257,
+ "learning_rate": 0.0006,
+ "loss": 6.090451240539551,
+ "step": 378
+ },
+ {
+ "epoch": 5.265502183406113,
+ "grad_norm": 0.020260317251086235,
+ "learning_rate": 0.0006,
+ "loss": 6.094257354736328,
+ "step": 379
+ },
+ {
+ "epoch": 5.279475982532751,
+ "grad_norm": 0.020110400393605232,
+ "learning_rate": 0.0006,
+ "loss": 6.021188735961914,
+ "step": 380
+ },
+ {
+ "epoch": 5.293449781659389,
+ "grad_norm": 0.02027830481529236,
+ "learning_rate": 0.0006,
+ "loss": 6.022156238555908,
+ "step": 381
+ },
+ {
+ "epoch": 5.307423580786026,
+ "grad_norm": 0.022747062146663666,
+ "learning_rate": 0.0006,
+ "loss": 6.011836051940918,
+ "step": 382
+ },
+ {
+ "epoch": 5.321397379912664,
+ "grad_norm": 0.01990230567753315,
+ "learning_rate": 0.0006,
+ "loss": 6.003054618835449,
+ "step": 383
+ },
+ {
+ "epoch": 5.335371179039301,
+ "grad_norm": 0.01596238650381565,
+ "learning_rate": 0.0006,
+ "loss": 5.991410732269287,
+ "step": 384
+ },
+ {
+ "epoch": 5.349344978165939,
+ "grad_norm": 0.015847105532884598,
+ "learning_rate": 0.0006,
+ "loss": 6.100622177124023,
+ "step": 385
+ },
+ {
+ "epoch": 5.3633187772925766,
+ "grad_norm": 0.016179397702217102,
+ "learning_rate": 0.0006,
+ "loss": 6.032659530639648,
+ "step": 386
+ },
+ {
+ "epoch": 5.377292576419214,
+ "grad_norm": 0.018256602808833122,
+ "learning_rate": 0.0006,
+ "loss": 5.961983680725098,
+ "step": 387
+ },
+ {
+ "epoch": 5.391266375545851,
+ "grad_norm": 0.02805912122130394,
+ "learning_rate": 0.0006,
+ "loss": 5.983541965484619,
+ "step": 388
+ },
+ {
+ "epoch": 5.405240174672489,
+ "grad_norm": 0.039082255214452744,
+ "learning_rate": 0.0006,
+ "loss": 6.021474361419678,
+ "step": 389
+ },
+ {
+ "epoch": 5.419213973799127,
+ "grad_norm": 0.036757659167051315,
+ "learning_rate": 0.0006,
+ "loss": 6.005046367645264,
+ "step": 390
+ },
+ {
+ "epoch": 5.4331877729257645,
+ "grad_norm": 0.035277366638183594,
+ "learning_rate": 0.0006,
+ "loss": 6.036296844482422,
+ "step": 391
+ },
+ {
+ "epoch": 5.447161572052401,
+ "grad_norm": 0.034404635429382324,
+ "learning_rate": 0.0006,
+ "loss": 5.994539260864258,
+ "step": 392
+ },
+ {
+ "epoch": 5.461135371179039,
+ "grad_norm": 0.0377206988632679,
+ "learning_rate": 0.0006,
+ "loss": 5.941534519195557,
+ "step": 393
+ },
+ {
+ "epoch": 5.475109170305677,
+ "grad_norm": 0.0389922633767128,
+ "learning_rate": 0.0006,
+ "loss": 5.983644485473633,
+ "step": 394
+ },
+ {
+ "epoch": 5.489082969432315,
+ "grad_norm": 0.04176778718829155,
+ "learning_rate": 0.0006,
+ "loss": 6.030922889709473,
+ "step": 395
+ },
+ {
+ "epoch": 5.503056768558952,
+ "grad_norm": 0.0337153784930706,
+ "learning_rate": 0.0006,
+ "loss": 5.961367130279541,
+ "step": 396
+ },
+ {
+ "epoch": 5.517030567685589,
+ "grad_norm": 0.036116816103458405,
+ "learning_rate": 0.0006,
+ "loss": 5.819280624389648,
+ "step": 397
+ },
+ {
+ "epoch": 5.531004366812227,
+ "grad_norm": 0.030724041163921356,
+ "learning_rate": 0.0006,
+ "loss": 5.906380653381348,
+ "step": 398
+ },
+ {
+ "epoch": 5.544978165938865,
+ "grad_norm": 0.030264202505350113,
+ "learning_rate": 0.0006,
+ "loss": 5.975833415985107,
+ "step": 399
+ },
+ {
+ "epoch": 5.558951965065502,
+ "grad_norm": 0.032096318900585175,
+ "learning_rate": 0.0006,
+ "loss": 6.001348972320557,
+ "step": 400
+ },
+ {
+ "epoch": 5.5729257641921395,
+ "grad_norm": 0.030579356476664543,
+ "learning_rate": 0.0006,
+ "loss": 5.928768634796143,
+ "step": 401
+ },
+ {
+ "epoch": 5.586899563318777,
+ "grad_norm": 0.028241973370313644,
+ "learning_rate": 0.0006,
+ "loss": 5.92582893371582,
+ "step": 402
+ },
+ {
+ "epoch": 5.600873362445415,
+ "grad_norm": 0.023086953908205032,
+ "learning_rate": 0.0006,
+ "loss": 5.892926216125488,
+ "step": 403
+ },
+ {
+ "epoch": 5.614847161572053,
+ "grad_norm": 0.02547992393374443,
+ "learning_rate": 0.0006,
+ "loss": 5.903195381164551,
+ "step": 404
+ },
+ {
+ "epoch": 5.62882096069869,
+ "grad_norm": 0.023089831694960594,
+ "learning_rate": 0.0006,
+ "loss": 5.912033557891846,
+ "step": 405
+ },
+ {
+ "epoch": 5.642794759825327,
+ "grad_norm": 0.025446368381381035,
+ "learning_rate": 0.0006,
+ "loss": 5.926138401031494,
+ "step": 406
+ },
+ {
+ "epoch": 5.656768558951965,
+ "grad_norm": 0.03748747706413269,
+ "learning_rate": 0.0006,
+ "loss": 5.913451194763184,
+ "step": 407
+ },
+ {
+ "epoch": 5.670742358078603,
+ "grad_norm": 0.03493810072541237,
+ "learning_rate": 0.0006,
+ "loss": 5.97633695602417,
+ "step": 408
+ },
+ {
+ "epoch": 5.68471615720524,
+ "grad_norm": 0.023131275549530983,
+ "learning_rate": 0.0006,
+ "loss": 5.899571418762207,
+ "step": 409
+ },
+ {
+ "epoch": 5.698689956331878,
+ "grad_norm": 0.02986014075577259,
+ "learning_rate": 0.0006,
+ "loss": 5.894903182983398,
+ "step": 410
+ },
+ {
+ "epoch": 5.712663755458515,
+ "grad_norm": 0.030171144753694534,
+ "learning_rate": 0.0006,
+ "loss": 5.876595497131348,
+ "step": 411
+ },
+ {
+ "epoch": 5.726637554585153,
+ "grad_norm": 0.029377546161413193,
+ "learning_rate": 0.0006,
+ "loss": 5.911637306213379,
+ "step": 412
+ },
+ {
+ "epoch": 5.74061135371179,
+ "grad_norm": 0.029829490929841995,
+ "learning_rate": 0.0006,
+ "loss": 5.819629669189453,
+ "step": 413
+ },
+ {
+ "epoch": 5.754585152838428,
+ "grad_norm": 0.024139299988746643,
+ "learning_rate": 0.0006,
+ "loss": 5.81728458404541,
+ "step": 414
+ },
+ {
+ "epoch": 5.7685589519650655,
+ "grad_norm": 0.021453432738780975,
+ "learning_rate": 0.0006,
+ "loss": 5.987326145172119,
+ "step": 415
+ },
+ {
+ "epoch": 5.782532751091703,
+ "grad_norm": 0.017942246049642563,
+ "learning_rate": 0.0006,
+ "loss": 5.893008232116699,
+ "step": 416
+ },
+ {
+ "epoch": 5.796506550218341,
+ "grad_norm": 0.019724637269973755,
+ "learning_rate": 0.0006,
+ "loss": 5.912441253662109,
+ "step": 417
+ },
+ {
+ "epoch": 5.810480349344978,
+ "grad_norm": 0.018597912043333054,
+ "learning_rate": 0.0006,
+ "loss": 5.860394477844238,
+ "step": 418
+ },
+ {
+ "epoch": 5.824454148471616,
+ "grad_norm": 0.016232412308454514,
+ "learning_rate": 0.0006,
+ "loss": 5.94991397857666,
+ "step": 419
+ },
+ {
+ "epoch": 5.8384279475982535,
+ "grad_norm": 0.01667204685509205,
+ "learning_rate": 0.0006,
+ "loss": 5.8506364822387695,
+ "step": 420
+ },
+ {
+ "epoch": 5.85240174672489,
+ "grad_norm": 0.014232151210308075,
+ "learning_rate": 0.0006,
+ "loss": 5.928424835205078,
+ "step": 421
+ },
+ {
+ "epoch": 5.866375545851528,
+ "grad_norm": 0.01570476032793522,
+ "learning_rate": 0.0006,
+ "loss": 5.953692436218262,
+ "step": 422
+ },
+ {
+ "epoch": 5.880349344978166,
+ "grad_norm": 0.01586179807782173,
+ "learning_rate": 0.0006,
+ "loss": 5.820014476776123,
+ "step": 423
+ },
+ {
+ "epoch": 5.894323144104804,
+ "grad_norm": 0.014729145914316177,
+ "learning_rate": 0.0006,
+ "loss": 5.860154628753662,
+ "step": 424
+ },
+ {
+ "epoch": 5.908296943231441,
+ "grad_norm": 0.013111459091305733,
+ "learning_rate": 0.0006,
+ "loss": 5.85988187789917,
+ "step": 425
+ },
+ {
+ "epoch": 5.922270742358078,
+ "grad_norm": 0.012937922962009907,
+ "learning_rate": 0.0006,
+ "loss": 5.756265640258789,
+ "step": 426
+ },
+ {
+ "epoch": 5.936244541484716,
+ "grad_norm": 0.01377453189343214,
+ "learning_rate": 0.0006,
+ "loss": 5.8333048820495605,
+ "step": 427
+ },
+ {
+ "epoch": 5.950218340611354,
+ "grad_norm": 0.014054795727133751,
+ "learning_rate": 0.0006,
+ "loss": 5.803333759307861,
+ "step": 428
+ },
+ {
+ "epoch": 5.964192139737992,
+ "grad_norm": 0.01674959622323513,
+ "learning_rate": 0.0006,
+ "loss": 5.8718485832214355,
+ "step": 429
+ },
+ {
+ "epoch": 5.978165938864628,
+ "grad_norm": 0.018733017146587372,
+ "learning_rate": 0.0006,
+ "loss": 5.875979900360107,
+ "step": 430
+ },
+ {
+ "epoch": 5.992139737991266,
+ "grad_norm": 0.02088983915746212,
+ "learning_rate": 0.0006,
+ "loss": 5.763634204864502,
+ "step": 431
+ },
+ {
+ "epoch": 6.0,
+ "grad_norm": 0.026705985888838768,
+ "learning_rate": 0.0006,
+ "loss": 5.6922478675842285,
+ "step": 432
+ },
+ {
+ "epoch": 6.0,
+ "eval_loss": 5.847958087921143,
+ "eval_runtime": 58.5451,
+ "eval_samples_per_second": 41.711,
+ "eval_steps_per_second": 1.315,
+ "step": 432
+ },
+ {
+ "epoch": 6.013973799126638,
+ "grad_norm": 0.03181544691324234,
+ "learning_rate": 0.0006,
+ "loss": 5.786368370056152,
+ "step": 433
+ },
+ {
+ "epoch": 6.0279475982532755,
+ "grad_norm": 0.03238112851977348,
+ "learning_rate": 0.0006,
+ "loss": 5.810310363769531,
+ "step": 434
+ },
+ {
+ "epoch": 6.041921397379912,
+ "grad_norm": 0.03916756808757782,
+ "learning_rate": 0.0006,
+ "loss": 5.7761993408203125,
+ "step": 435
+ },
+ {
+ "epoch": 6.05589519650655,
+ "grad_norm": 0.044009730219841,
+ "learning_rate": 0.0006,
+ "loss": 5.898112773895264,
+ "step": 436
+ },
+ {
+ "epoch": 6.069868995633188,
+ "grad_norm": 0.045235246419906616,
+ "learning_rate": 0.0006,
+ "loss": 5.876371383666992,
+ "step": 437
+ },
+ {
+ "epoch": 6.083842794759826,
+ "grad_norm": 0.050218384712934494,
+ "learning_rate": 0.0006,
+ "loss": 5.805103302001953,
+ "step": 438
+ },
+ {
+ "epoch": 6.097816593886463,
+ "grad_norm": 0.05444490164518356,
+ "learning_rate": 0.0006,
+ "loss": 5.826424598693848,
+ "step": 439
+ },
+ {
+ "epoch": 6.1117903930131,
+ "grad_norm": 0.04890982061624527,
+ "learning_rate": 0.0006,
+ "loss": 5.7735395431518555,
+ "step": 440
+ },
+ {
+ "epoch": 6.125764192139738,
+ "grad_norm": 0.05478639155626297,
+ "learning_rate": 0.0006,
+ "loss": 5.823677062988281,
+ "step": 441
+ },
+ {
+ "epoch": 6.139737991266376,
+ "grad_norm": 0.057562313973903656,
+ "learning_rate": 0.0006,
+ "loss": 5.889334678649902,
+ "step": 442
+ },
+ {
+ "epoch": 6.153711790393013,
+ "grad_norm": 0.06447703391313553,
+ "learning_rate": 0.0006,
+ "loss": 5.84621524810791,
+ "step": 443
+ },
+ {
+ "epoch": 6.1676855895196505,
+ "grad_norm": 0.048861872404813766,
+ "learning_rate": 0.0006,
+ "loss": 5.931595802307129,
+ "step": 444
+ },
+ {
+ "epoch": 6.181659388646288,
+ "grad_norm": 0.05521339178085327,
+ "learning_rate": 0.0006,
+ "loss": 5.8229780197143555,
+ "step": 445
+ },
+ {
+ "epoch": 6.195633187772926,
+ "grad_norm": 0.053666990250349045,
+ "learning_rate": 0.0006,
+ "loss": 5.879191875457764,
+ "step": 446
+ },
+ {
+ "epoch": 6.209606986899563,
+ "grad_norm": 0.0451025515794754,
+ "learning_rate": 0.0006,
+ "loss": 5.801628112792969,
+ "step": 447
+ },
+ {
+ "epoch": 6.223580786026201,
+ "grad_norm": 0.04291655868291855,
+ "learning_rate": 0.0006,
+ "loss": 5.914680480957031,
+ "step": 448
+ },
+ {
+ "epoch": 6.2375545851528384,
+ "grad_norm": 0.035102490335702896,
+ "learning_rate": 0.0006,
+ "loss": 5.843682765960693,
+ "step": 449
+ },
+ {
+ "epoch": 6.251528384279476,
+ "grad_norm": 0.03403995931148529,
+ "learning_rate": 0.0006,
+ "loss": 5.827154159545898,
+ "step": 450
+ },
+ {
+ "epoch": 6.265502183406113,
+ "grad_norm": 0.03444375470280647,
+ "learning_rate": 0.0006,
+ "loss": 5.800136566162109,
+ "step": 451
+ },
+ {
+ "epoch": 6.279475982532751,
+ "grad_norm": 0.03165159001946449,
+ "learning_rate": 0.0006,
+ "loss": 5.906252384185791,
+ "step": 452
+ },
+ {
+ "epoch": 6.293449781659389,
+ "grad_norm": 0.026153093203902245,
+ "learning_rate": 0.0006,
+ "loss": 5.7423200607299805,
+ "step": 453
+ },
+ {
+ "epoch": 6.307423580786026,
+ "grad_norm": 0.024357657879590988,
+ "learning_rate": 0.0006,
+ "loss": 5.701364517211914,
+ "step": 454
+ },
+ {
+ "epoch": 6.321397379912664,
+ "grad_norm": 0.021508049219846725,
+ "learning_rate": 0.0006,
+ "loss": 5.8328094482421875,
+ "step": 455
+ },
+ {
+ "epoch": 6.335371179039301,
+ "grad_norm": 0.017313921824097633,
+ "learning_rate": 0.0006,
+ "loss": 5.792145252227783,
+ "step": 456
+ },
+ {
+ "epoch": 6.349344978165939,
+ "grad_norm": 0.018563397228717804,
+ "learning_rate": 0.0006,
+ "loss": 5.732672214508057,
+ "step": 457
+ },
+ {
+ "epoch": 6.3633187772925766,
+ "grad_norm": 0.016568679362535477,
+ "learning_rate": 0.0006,
+ "loss": 5.81948184967041,
+ "step": 458
+ },
+ {
+ "epoch": 6.377292576419214,
+ "grad_norm": 0.014133770950138569,
+ "learning_rate": 0.0006,
+ "loss": 5.734982490539551,
+ "step": 459
+ },
+ {
+ "epoch": 6.391266375545851,
+ "grad_norm": 0.014114447869360447,
+ "learning_rate": 0.0006,
+ "loss": 5.70073127746582,
+ "step": 460
+ },
+ {
+ "epoch": 6.405240174672489,
+ "grad_norm": 0.012907393276691437,
+ "learning_rate": 0.0006,
+ "loss": 5.758626937866211,
+ "step": 461
+ },
+ {
+ "epoch": 6.419213973799127,
+ "grad_norm": 0.012673444114625454,
+ "learning_rate": 0.0006,
+ "loss": 5.775138854980469,
+ "step": 462
+ },
+ {
+ "epoch": 6.4331877729257645,
+ "grad_norm": 0.01311410591006279,
+ "learning_rate": 0.0006,
+ "loss": 5.70920991897583,
+ "step": 463
+ },
+ {
+ "epoch": 6.447161572052401,
+ "grad_norm": 0.012935544364154339,
+ "learning_rate": 0.0006,
+ "loss": 5.781479835510254,
+ "step": 464
+ },
+ {
+ "epoch": 6.461135371179039,
+ "grad_norm": 0.01439738366752863,
+ "learning_rate": 0.0006,
+ "loss": 5.691315650939941,
+ "step": 465
+ },
+ {
+ "epoch": 6.475109170305677,
+ "grad_norm": 0.01205186452716589,
+ "learning_rate": 0.0006,
+ "loss": 5.662499904632568,
+ "step": 466
+ },
+ {
+ "epoch": 6.489082969432315,
+ "grad_norm": 0.011575652286410332,
+ "learning_rate": 0.0006,
+ "loss": 5.6271772384643555,
+ "step": 467
+ },
+ {
+ "epoch": 6.503056768558952,
+ "grad_norm": 0.011672502383589745,
+ "learning_rate": 0.0006,
+ "loss": 5.761662006378174,
+ "step": 468
+ },
+ {
+ "epoch": 6.517030567685589,
+ "grad_norm": 0.011541751213371754,
+ "learning_rate": 0.0006,
+ "loss": 5.762078285217285,
+ "step": 469
+ },
+ {
+ "epoch": 6.531004366812227,
+ "grad_norm": 0.011396365240216255,
+ "learning_rate": 0.0006,
+ "loss": 5.67873477935791,
+ "step": 470
+ },
+ {
+ "epoch": 6.544978165938865,
+ "grad_norm": 0.010688499547541142,
+ "learning_rate": 0.0006,
+ "loss": 5.717051982879639,
+ "step": 471
+ },
+ {
+ "epoch": 6.558951965065502,
+ "grad_norm": 0.012224317528307438,
+ "learning_rate": 0.0006,
+ "loss": 5.707948684692383,
+ "step": 472
+ },
+ {
+ "epoch": 6.5729257641921395,
+ "grad_norm": 0.011856825090944767,
+ "learning_rate": 0.0006,
+ "loss": 5.70878791809082,
+ "step": 473
+ },
+ {
+ "epoch": 6.586899563318777,
+ "grad_norm": 0.01199064590036869,
+ "learning_rate": 0.0006,
+ "loss": 5.708697319030762,
+ "step": 474
+ },
+ {
+ "epoch": 6.600873362445415,
+ "grad_norm": 0.01219446212053299,
+ "learning_rate": 0.0006,
+ "loss": 5.6061577796936035,
+ "step": 475
+ },
+ {
+ "epoch": 6.614847161572053,
+ "grad_norm": 0.013397484086453915,
+ "learning_rate": 0.0006,
+ "loss": 5.624789714813232,
+ "step": 476
+ },
+ {
+ "epoch": 6.62882096069869,
+ "grad_norm": 0.01483136136084795,
+ "learning_rate": 0.0006,
+ "loss": 5.721141338348389,
+ "step": 477
+ },
+ {
+ "epoch": 6.642794759825327,
+ "grad_norm": 0.0194488987326622,
+ "learning_rate": 0.0006,
+ "loss": 5.604279518127441,
+ "step": 478
+ },
+ {
+ "epoch": 6.656768558951965,
+ "grad_norm": 0.02079896256327629,
+ "learning_rate": 0.0006,
+ "loss": 5.586322784423828,
+ "step": 479
+ },
+ {
+ "epoch": 6.670742358078603,
+ "grad_norm": 0.018996229395270348,
+ "learning_rate": 0.0006,
+ "loss": 5.5987772941589355,
+ "step": 480
+ },
+ {
+ "epoch": 6.68471615720524,
+ "grad_norm": 0.015779603272676468,
+ "learning_rate": 0.0006,
+ "loss": 5.588602066040039,
+ "step": 481
+ },
+ {
+ "epoch": 6.698689956331878,
+ "grad_norm": 0.015322973020374775,
+ "learning_rate": 0.0006,
+ "loss": 5.686467170715332,
+ "step": 482
+ },
+ {
+ "epoch": 6.712663755458515,
+ "grad_norm": 0.015282213687896729,
+ "learning_rate": 0.0006,
+ "loss": 5.664676666259766,
+ "step": 483
+ },
+ {
+ "epoch": 6.726637554585153,
+ "grad_norm": 0.016589272767305374,
+ "learning_rate": 0.0006,
+ "loss": 5.666739463806152,
+ "step": 484
+ },
+ {
+ "epoch": 6.74061135371179,
+ "grad_norm": 0.01811421848833561,
+ "learning_rate": 0.0006,
+ "loss": 5.647593021392822,
+ "step": 485
+ },
+ {
+ "epoch": 6.754585152838428,
+ "grad_norm": 0.019652029499411583,
+ "learning_rate": 0.0006,
+ "loss": 5.628726959228516,
+ "step": 486
+ },
+ {
+ "epoch": 6.7685589519650655,
+ "grad_norm": 0.02118665538728237,
+ "learning_rate": 0.0006,
+ "loss": 5.6580705642700195,
+ "step": 487
+ },
+ {
+ "epoch": 6.782532751091703,
+ "grad_norm": 0.02237832546234131,
+ "learning_rate": 0.0006,
+ "loss": 5.635931968688965,
+ "step": 488
+ },
+ {
+ "epoch": 6.796506550218341,
+ "grad_norm": 0.023897631093859673,
+ "learning_rate": 0.0006,
+ "loss": 5.517949104309082,
+ "step": 489
+ },
+ {
+ "epoch": 6.810480349344978,
+ "grad_norm": 0.02442227490246296,
+ "learning_rate": 0.0006,
+ "loss": 5.600021839141846,
+ "step": 490
+ },
+ {
+ "epoch": 6.824454148471616,
+ "grad_norm": 0.024675287306308746,
+ "learning_rate": 0.0006,
+ "loss": 5.6639909744262695,
+ "step": 491
+ },
+ {
+ "epoch": 6.8384279475982535,
+ "grad_norm": 0.030372392386198044,
+ "learning_rate": 0.0006,
+ "loss": 5.582888603210449,
+ "step": 492
+ },
+ {
+ "epoch": 6.85240174672489,
+ "grad_norm": 0.03237690031528473,
+ "learning_rate": 0.0006,
+ "loss": 5.573896884918213,
+ "step": 493
+ },
+ {
+ "epoch": 6.866375545851528,
+ "grad_norm": 0.03511589393019676,
+ "learning_rate": 0.0006,
+ "loss": 5.565878868103027,
+ "step": 494
+ },
+ {
+ "epoch": 6.880349344978166,
+ "grad_norm": 0.03883938118815422,
+ "learning_rate": 0.0006,
+ "loss": 5.58540678024292,
+ "step": 495
+ },
+ {
+ "epoch": 6.894323144104804,
+ "grad_norm": 0.04175502806901932,
+ "learning_rate": 0.0006,
+ "loss": 5.6095194816589355,
+ "step": 496
+ },
+ {
+ "epoch": 6.908296943231441,
+ "grad_norm": 0.04013441503047943,
+ "learning_rate": 0.0006,
+ "loss": 5.654401779174805,
+ "step": 497
+ },
+ {
+ "epoch": 6.922270742358078,
+ "grad_norm": 0.04618770629167557,
+ "learning_rate": 0.0006,
+ "loss": 5.6483473777771,
+ "step": 498
+ },
+ {
+ "epoch": 6.936244541484716,
+ "grad_norm": 0.04507308453321457,
+ "learning_rate": 0.0006,
+ "loss": 5.5856218338012695,
+ "step": 499
+ },
+ {
+ "epoch": 6.950218340611354,
+ "grad_norm": 0.03151383996009827,
+ "learning_rate": 0.0006,
+ "loss": 5.594086647033691,
+ "step": 500
+ },
+ {
+ "epoch": 6.964192139737992,
+ "grad_norm": 0.026883797720074654,
+ "learning_rate": 0.0006,
+ "loss": 5.607676029205322,
+ "step": 501
+ },
+ {
+ "epoch": 6.978165938864628,
+ "grad_norm": 0.02981836162507534,
+ "learning_rate": 0.0006,
+ "loss": 5.563666820526123,
+ "step": 502
+ },
+ {
+ "epoch": 6.992139737991266,
+ "grad_norm": 0.03233477845788002,
+ "learning_rate": 0.0006,
+ "loss": 5.667543411254883,
+ "step": 503
+ },
+ {
+ "epoch": 7.0,
+ "grad_norm": 0.029927456751465797,
+ "learning_rate": 0.0006,
+ "loss": 5.733482837677002,
+ "step": 504
+ },
+ {
+ "epoch": 7.0,
+ "eval_loss": 5.679076671600342,
+ "eval_runtime": 59.5965,
+ "eval_samples_per_second": 40.976,
+ "eval_steps_per_second": 1.292,
+ "step": 504
+ },
+ {
+ "epoch": 7.013973799126638,
+ "grad_norm": 0.027081597596406937,
+ "learning_rate": 0.0006,
+ "loss": 5.634487152099609,
+ "step": 505
+ },
+ {
+ "epoch": 7.0279475982532755,
+ "grad_norm": 0.031959421932697296,
+ "learning_rate": 0.0006,
+ "loss": 5.643294811248779,
+ "step": 506
+ },
+ {
+ "epoch": 7.041921397379912,
+ "grad_norm": 0.02802063524723053,
+ "learning_rate": 0.0006,
+ "loss": 5.573239326477051,
+ "step": 507
+ },
+ {
+ "epoch": 7.05589519650655,
+ "grad_norm": 0.032579705119132996,
+ "learning_rate": 0.0006,
+ "loss": 5.627150058746338,
+ "step": 508
+ },
+ {
+ "epoch": 7.069868995633188,
+ "grad_norm": 0.02503928542137146,
+ "learning_rate": 0.0006,
+ "loss": 5.6045308113098145,
+ "step": 509
+ },
+ {
+ "epoch": 7.083842794759826,
+ "grad_norm": 0.02421317622065544,
+ "learning_rate": 0.0006,
+ "loss": 5.522153377532959,
+ "step": 510
+ },
+ {
+ "epoch": 7.097816593886463,
+ "grad_norm": 0.021983426064252853,
+ "learning_rate": 0.0006,
+ "loss": 5.632939338684082,
+ "step": 511
+ },
+ {
+ "epoch": 7.1117903930131,
+ "grad_norm": 0.021933183073997498,
+ "learning_rate": 0.0006,
+ "loss": 5.592185974121094,
+ "step": 512
+ },
+ {
+ "epoch": 7.125764192139738,
+ "grad_norm": 0.02287031151354313,
+ "learning_rate": 0.0006,
+ "loss": 5.613700866699219,
+ "step": 513
+ },
+ {
+ "epoch": 7.139737991266376,
+ "grad_norm": 0.02281602844595909,
+ "learning_rate": 0.0006,
+ "loss": 5.551383018493652,
+ "step": 514
+ },
+ {
+ "epoch": 7.153711790393013,
+ "grad_norm": 0.021050360053777695,
+ "learning_rate": 0.0006,
+ "loss": 5.55275821685791,
+ "step": 515
+ },
+ {
+ "epoch": 7.1676855895196505,
+ "grad_norm": 0.018299926072359085,
+ "learning_rate": 0.0006,
+ "loss": 5.528225421905518,
+ "step": 516
+ },
+ {
+ "epoch": 7.181659388646288,
+ "grad_norm": 0.02024853602051735,
+ "learning_rate": 0.0006,
+ "loss": 5.522453784942627,
+ "step": 517
+ },
+ {
+ "epoch": 7.195633187772926,
+ "grad_norm": 0.02135239914059639,
+ "learning_rate": 0.0006,
+ "loss": 5.527522563934326,
+ "step": 518
+ },
+ {
+ "epoch": 7.209606986899563,
+ "grad_norm": 0.023394184187054634,
+ "learning_rate": 0.0006,
+ "loss": 5.4535017013549805,
+ "step": 519
+ },
+ {
+ "epoch": 7.223580786026201,
+ "grad_norm": 0.025952080264687538,
+ "learning_rate": 0.0006,
+ "loss": 5.5374674797058105,
+ "step": 520
+ },
+ {
+ "epoch": 7.2375545851528384,
+ "grad_norm": 0.022287718951702118,
+ "learning_rate": 0.0006,
+ "loss": 5.493753433227539,
+ "step": 521
+ },
+ {
+ "epoch": 7.251528384279476,
+ "grad_norm": 0.02018447406589985,
+ "learning_rate": 0.0006,
+ "loss": 5.510575294494629,
+ "step": 522
+ },
+ {
+ "epoch": 7.265502183406113,
+ "grad_norm": 0.02050507813692093,
+ "learning_rate": 0.0006,
+ "loss": 5.466026782989502,
+ "step": 523
+ },
+ {
+ "epoch": 7.279475982532751,
+ "grad_norm": 0.023688802495598793,
+ "learning_rate": 0.0006,
+ "loss": 5.485815525054932,
+ "step": 524
+ },
+ {
+ "epoch": 7.293449781659389,
+ "grad_norm": 0.021078843623399734,
+ "learning_rate": 0.0006,
+ "loss": 5.415581703186035,
+ "step": 525
+ },
+ {
+ "epoch": 7.307423580786026,
+ "grad_norm": 0.018266011029481888,
+ "learning_rate": 0.0006,
+ "loss": 5.402815818786621,
+ "step": 526
+ },
+ {
+ "epoch": 7.321397379912664,
+ "grad_norm": 0.019693493843078613,
+ "learning_rate": 0.0006,
+ "loss": 5.504674911499023,
+ "step": 527
+ },
+ {
+ "epoch": 7.335371179039301,
+ "grad_norm": 0.02031373605132103,
+ "learning_rate": 0.0006,
+ "loss": 5.503837585449219,
+ "step": 528
+ },
+ {
+ "epoch": 7.349344978165939,
+ "grad_norm": 0.017636626958847046,
+ "learning_rate": 0.0006,
+ "loss": 5.530580997467041,
+ "step": 529
+ },
+ {
+ "epoch": 7.3633187772925766,
+ "grad_norm": 0.01787244901061058,
+ "learning_rate": 0.0006,
+ "loss": 5.450218677520752,
+ "step": 530
+ },
+ {
+ "epoch": 7.377292576419214,
+ "grad_norm": 0.0170669574290514,
+ "learning_rate": 0.0006,
+ "loss": 5.481570243835449,
+ "step": 531
+ },
+ {
+ "epoch": 7.391266375545851,
+ "grad_norm": 0.01802165061235428,
+ "learning_rate": 0.0006,
+ "loss": 5.386394500732422,
+ "step": 532
+ },
+ {
+ "epoch": 7.405240174672489,
+ "grad_norm": 0.022949472069740295,
+ "learning_rate": 0.0006,
+ "loss": 5.424929618835449,
+ "step": 533
+ },
+ {
+ "epoch": 7.419213973799127,
+ "grad_norm": 0.034213390201330185,
+ "learning_rate": 0.0006,
+ "loss": 5.346663475036621,
+ "step": 534
+ },
+ {
+ "epoch": 7.4331877729257645,
+ "grad_norm": 0.03847593814134598,
+ "learning_rate": 0.0006,
+ "loss": 5.418482780456543,
+ "step": 535
+ },
+ {
+ "epoch": 7.447161572052401,
+ "grad_norm": 0.029393676668405533,
+ "learning_rate": 0.0006,
+ "loss": 5.456090927124023,
+ "step": 536
+ },
+ {
+ "epoch": 7.461135371179039,
+ "grad_norm": 0.03388667106628418,
+ "learning_rate": 0.0006,
+ "loss": 5.572983741760254,
+ "step": 537
+ },
+ {
+ "epoch": 7.475109170305677,
+ "grad_norm": 0.039690613746643066,
+ "learning_rate": 0.0006,
+ "loss": 5.414067268371582,
+ "step": 538
+ },
+ {
+ "epoch": 7.489082969432315,
+ "grad_norm": 0.03630776330828667,
+ "learning_rate": 0.0006,
+ "loss": 5.523739814758301,
+ "step": 539
+ },
+ {
+ "epoch": 7.503056768558952,
+ "grad_norm": 0.03356870263814926,
+ "learning_rate": 0.0006,
+ "loss": 5.444767951965332,
+ "step": 540
+ },
+ {
+ "epoch": 7.517030567685589,
+ "grad_norm": 0.030812304466962814,
+ "learning_rate": 0.0006,
+ "loss": 5.461244106292725,
+ "step": 541
+ },
+ {
+ "epoch": 7.531004366812227,
+ "grad_norm": 0.03319217637181282,
+ "learning_rate": 0.0006,
+ "loss": 5.502161026000977,
+ "step": 542
+ },
+ {
+ "epoch": 7.544978165938865,
+ "grad_norm": 0.032997481524944305,
+ "learning_rate": 0.0006,
+ "loss": 5.4162278175354,
+ "step": 543
+ },
+ {
+ "epoch": 7.558951965065502,
+ "grad_norm": 0.03364962339401245,
+ "learning_rate": 0.0006,
+ "loss": 5.415736198425293,
+ "step": 544
+ },
+ {
+ "epoch": 7.5729257641921395,
+ "grad_norm": 0.0344221405684948,
+ "learning_rate": 0.0006,
+ "loss": 5.541967868804932,
+ "step": 545
+ },
+ {
+ "epoch": 7.586899563318777,
+ "grad_norm": 0.029609503224492073,
+ "learning_rate": 0.0006,
+ "loss": 5.440614223480225,
+ "step": 546
+ },
+ {
+ "epoch": 7.600873362445415,
+ "grad_norm": 0.029055926948785782,
+ "learning_rate": 0.0006,
+ "loss": 5.462865352630615,
+ "step": 547
+ },
+ {
+ "epoch": 7.614847161572053,
+ "grad_norm": 0.02658848837018013,
+ "learning_rate": 0.0006,
+ "loss": 5.444467544555664,
+ "step": 548
+ },
+ {
+ "epoch": 7.62882096069869,
+ "grad_norm": 0.026276404038071632,
+ "learning_rate": 0.0006,
+ "loss": 5.462278842926025,
+ "step": 549
+ },
+ {
+ "epoch": 7.642794759825327,
+ "grad_norm": 0.0286889486014843,
+ "learning_rate": 0.0006,
+ "loss": 5.4849534034729,
+ "step": 550
+ },
+ {
+ "epoch": 7.656768558951965,
+ "grad_norm": 0.026849817484617233,
+ "learning_rate": 0.0006,
+ "loss": 5.431473731994629,
+ "step": 551
+ },
+ {
+ "epoch": 7.670742358078603,
+ "grad_norm": 0.02312396466732025,
+ "learning_rate": 0.0006,
+ "loss": 5.45462703704834,
+ "step": 552
+ },
+ {
+ "epoch": 7.68471615720524,
+ "grad_norm": 0.026197485625743866,
+ "learning_rate": 0.0006,
+ "loss": 5.452552318572998,
+ "step": 553
+ },
+ {
+ "epoch": 7.698689956331878,
+ "grad_norm": 0.02747255191206932,
+ "learning_rate": 0.0006,
+ "loss": 5.447169780731201,
+ "step": 554
+ },
+ {
+ "epoch": 7.712663755458515,
+ "grad_norm": 0.028123432770371437,
+ "learning_rate": 0.0006,
+ "loss": 5.394400596618652,
+ "step": 555
+ },
+ {
+ "epoch": 7.726637554585153,
+ "grad_norm": 0.02599870041012764,
+ "learning_rate": 0.0006,
+ "loss": 5.398341178894043,
+ "step": 556
+ },
+ {
+ "epoch": 7.74061135371179,
+ "grad_norm": 0.022171657532453537,
+ "learning_rate": 0.0006,
+ "loss": 5.368820667266846,
+ "step": 557
+ },
+ {
+ "epoch": 7.754585152838428,
+ "grad_norm": 0.022309480234980583,
+ "learning_rate": 0.0006,
+ "loss": 5.416772365570068,
+ "step": 558
+ },
+ {
+ "epoch": 7.7685589519650655,
+ "grad_norm": 0.024074165150523186,
+ "learning_rate": 0.0006,
+ "loss": 5.359002113342285,
+ "step": 559
+ },
+ {
+ "epoch": 7.782532751091703,
+ "grad_norm": 0.02636653557419777,
+ "learning_rate": 0.0006,
+ "loss": 5.484368324279785,
+ "step": 560
+ },
+ {
+ "epoch": 7.796506550218341,
+ "grad_norm": 0.02203752100467682,
+ "learning_rate": 0.0006,
+ "loss": 5.375498294830322,
+ "step": 561
+ },
+ {
+ "epoch": 7.810480349344978,
+ "grad_norm": 0.019924819469451904,
+ "learning_rate": 0.0006,
+ "loss": 5.334723472595215,
+ "step": 562
+ },
+ {
+ "epoch": 7.824454148471616,
+ "grad_norm": 0.018755966797471046,
+ "learning_rate": 0.0006,
+ "loss": 5.348781108856201,
+ "step": 563
+ },
+ {
+ "epoch": 7.8384279475982535,
+ "grad_norm": 0.016747845336794853,
+ "learning_rate": 0.0006,
+ "loss": 5.433079719543457,
+ "step": 564
+ },
+ {
+ "epoch": 7.85240174672489,
+ "grad_norm": 0.0170395877212286,
+ "learning_rate": 0.0006,
+ "loss": 5.346794128417969,
+ "step": 565
+ },
+ {
+ "epoch": 7.866375545851528,
+ "grad_norm": 0.016074176877737045,
+ "learning_rate": 0.0006,
+ "loss": 5.3277740478515625,
+ "step": 566
+ },
+ {
+ "epoch": 7.880349344978166,
+ "grad_norm": 0.014119806699454784,
+ "learning_rate": 0.0006,
+ "loss": 5.483551979064941,
+ "step": 567
+ },
+ {
+ "epoch": 7.894323144104804,
+ "grad_norm": 0.015270394273102283,
+ "learning_rate": 0.0006,
+ "loss": 5.378519535064697,
+ "step": 568
+ },
+ {
+ "epoch": 7.908296943231441,
+ "grad_norm": 0.015677539631724358,
+ "learning_rate": 0.0006,
+ "loss": 5.290210247039795,
+ "step": 569
+ },
+ {
+ "epoch": 7.922270742358078,
+ "grad_norm": 0.015930423513054848,
+ "learning_rate": 0.0006,
+ "loss": 5.427360534667969,
+ "step": 570
+ },
+ {
+ "epoch": 7.936244541484716,
+ "grad_norm": 0.016641564667224884,
+ "learning_rate": 0.0006,
+ "loss": 5.301599979400635,
+ "step": 571
+ },
+ {
+ "epoch": 7.950218340611354,
+ "grad_norm": 0.018293552100658417,
+ "learning_rate": 0.0006,
+ "loss": 5.277889251708984,
+ "step": 572
+ },
+ {
+ "epoch": 7.964192139737992,
+ "grad_norm": 0.017618799582123756,
+ "learning_rate": 0.0006,
+ "loss": 5.297194480895996,
+ "step": 573
+ },
+ {
+ "epoch": 7.978165938864628,
+ "grad_norm": 0.015034242533147335,
+ "learning_rate": 0.0006,
+ "loss": 5.42428731918335,
+ "step": 574
+ },
+ {
+ "epoch": 7.992139737991266,
+ "grad_norm": 0.01617511734366417,
+ "learning_rate": 0.0006,
+ "loss": 5.420044898986816,
+ "step": 575
+ },
+ {
+ "epoch": 8.0,
+ "grad_norm": 0.01837257295846939,
+ "learning_rate": 0.0006,
+ "loss": 5.240387916564941,
+ "step": 576
+ },
+ {
+ "epoch": 8.0,
+ "eval_loss": 5.423932075500488,
+ "eval_runtime": 58.0841,
+ "eval_samples_per_second": 42.042,
+ "eval_steps_per_second": 1.326,
+ "step": 576
+ },
+ {
+ "epoch": 8.013973799126637,
+ "grad_norm": 0.020953809842467308,
+ "learning_rate": 0.0006,
+ "loss": 5.4320783615112305,
+ "step": 577
+ },
+ {
+ "epoch": 8.027947598253276,
+ "grad_norm": 0.0227745920419693,
+ "learning_rate": 0.0006,
+ "loss": 5.310197830200195,
+ "step": 578
+ },
+ {
+ "epoch": 8.041921397379912,
+ "grad_norm": 0.021372603252530098,
+ "learning_rate": 0.0006,
+ "loss": 5.339468002319336,
+ "step": 579
+ },
+ {
+ "epoch": 8.055895196506551,
+ "grad_norm": 0.022401731461286545,
+ "learning_rate": 0.0006,
+ "loss": 5.279829978942871,
+ "step": 580
+ },
+ {
+ "epoch": 8.069868995633188,
+ "grad_norm": 0.023942379280924797,
+ "learning_rate": 0.0006,
+ "loss": 5.372323989868164,
+ "step": 581
+ },
+ {
+ "epoch": 8.083842794759825,
+ "grad_norm": 0.024380534887313843,
+ "learning_rate": 0.0006,
+ "loss": 5.253486633300781,
+ "step": 582
+ },
+ {
+ "epoch": 8.097816593886463,
+ "grad_norm": 0.03246625140309334,
+ "learning_rate": 0.0006,
+ "loss": 5.305774211883545,
+ "step": 583
+ },
+ {
+ "epoch": 8.1117903930131,
+ "grad_norm": 0.030399736016988754,
+ "learning_rate": 0.0006,
+ "loss": 5.370976448059082,
+ "step": 584
+ },
+ {
+ "epoch": 8.125764192139737,
+ "grad_norm": 0.027148913592100143,
+ "learning_rate": 0.0006,
+ "loss": 5.368010520935059,
+ "step": 585
+ },
+ {
+ "epoch": 8.139737991266376,
+ "grad_norm": 0.02925540879368782,
+ "learning_rate": 0.0006,
+ "loss": 5.303013801574707,
+ "step": 586
+ },
+ {
+ "epoch": 8.153711790393013,
+ "grad_norm": 0.02915453538298607,
+ "learning_rate": 0.0006,
+ "loss": 5.329667091369629,
+ "step": 587
+ },
+ {
+ "epoch": 8.167685589519651,
+ "grad_norm": 0.031732626259326935,
+ "learning_rate": 0.0006,
+ "loss": 5.3634934425354,
+ "step": 588
+ },
+ {
+ "epoch": 8.181659388646288,
+ "grad_norm": 0.036166731268167496,
+ "learning_rate": 0.0006,
+ "loss": 5.223980903625488,
+ "step": 589
+ },
+ {
+ "epoch": 8.195633187772925,
+ "grad_norm": 0.035654108971357346,
+ "learning_rate": 0.0006,
+ "loss": 5.361059188842773,
+ "step": 590
+ },
+ {
+ "epoch": 8.209606986899564,
+ "grad_norm": 0.0365324430167675,
+ "learning_rate": 0.0006,
+ "loss": 5.336151599884033,
+ "step": 591
+ },
+ {
+ "epoch": 8.2235807860262,
+ "grad_norm": 0.03578880429267883,
+ "learning_rate": 0.0006,
+ "loss": 5.434076309204102,
+ "step": 592
+ },
+ {
+ "epoch": 8.237554585152838,
+ "grad_norm": 0.035734012722969055,
+ "learning_rate": 0.0006,
+ "loss": 5.330893516540527,
+ "step": 593
+ },
+ {
+ "epoch": 8.251528384279476,
+ "grad_norm": 0.03196857124567032,
+ "learning_rate": 0.0006,
+ "loss": 5.306009769439697,
+ "step": 594
+ },
+ {
+ "epoch": 8.265502183406113,
+ "grad_norm": 0.032042086124420166,
+ "learning_rate": 0.0006,
+ "loss": 5.343267917633057,
+ "step": 595
+ },
+ {
+ "epoch": 8.279475982532752,
+ "grad_norm": 0.03160746395587921,
+ "learning_rate": 0.0006,
+ "loss": 5.2353715896606445,
+ "step": 596
+ },
+ {
+ "epoch": 8.293449781659389,
+ "grad_norm": 0.026689818128943443,
+ "learning_rate": 0.0006,
+ "loss": 5.254042625427246,
+ "step": 597
+ },
+ {
+ "epoch": 8.307423580786025,
+ "grad_norm": 0.02880188450217247,
+ "learning_rate": 0.0006,
+ "loss": 5.326833248138428,
+ "step": 598
+ },
+ {
+ "epoch": 8.321397379912664,
+ "grad_norm": 0.027516381815075874,
+ "learning_rate": 0.0006,
+ "loss": 5.342336177825928,
+ "step": 599
+ },
+ {
+ "epoch": 8.335371179039301,
+ "grad_norm": 0.027875030413269997,
+ "learning_rate": 0.0006,
+ "loss": 5.200719833374023,
+ "step": 600
+ },
+ {
+ "epoch": 8.34934497816594,
+ "grad_norm": 0.0268265288323164,
+ "learning_rate": 0.0006,
+ "loss": 5.249449729919434,
+ "step": 601
+ },
+ {
+ "epoch": 8.363318777292577,
+ "grad_norm": 0.024824608117341995,
+ "learning_rate": 0.0006,
+ "loss": 5.318952560424805,
+ "step": 602
+ },
+ {
+ "epoch": 8.377292576419213,
+ "grad_norm": 0.022990627214312553,
+ "learning_rate": 0.0006,
+ "loss": 5.244993209838867,
+ "step": 603
+ },
+ {
+ "epoch": 8.391266375545852,
+ "grad_norm": 0.022804604843258858,
+ "learning_rate": 0.0006,
+ "loss": 5.3236846923828125,
+ "step": 604
+ },
+ {
+ "epoch": 8.405240174672489,
+ "grad_norm": 0.02419872209429741,
+ "learning_rate": 0.0006,
+ "loss": 5.255486488342285,
+ "step": 605
+ },
+ {
+ "epoch": 8.419213973799126,
+ "grad_norm": 0.021952059119939804,
+ "learning_rate": 0.0006,
+ "loss": 5.3511552810668945,
+ "step": 606
+ },
+ {
+ "epoch": 8.433187772925764,
+ "grad_norm": 0.022375497967004776,
+ "learning_rate": 0.0006,
+ "loss": 5.294790267944336,
+ "step": 607
+ },
+ {
+ "epoch": 8.447161572052401,
+ "grad_norm": 0.019242815673351288,
+ "learning_rate": 0.0006,
+ "loss": 5.252618789672852,
+ "step": 608
+ },
+ {
+ "epoch": 8.46113537117904,
+ "grad_norm": 0.0173486340790987,
+ "learning_rate": 0.0006,
+ "loss": 5.30147647857666,
+ "step": 609
+ },
+ {
+ "epoch": 8.475109170305677,
+ "grad_norm": 0.01612604409456253,
+ "learning_rate": 0.0006,
+ "loss": 5.23960542678833,
+ "step": 610
+ },
+ {
+ "epoch": 8.489082969432314,
+ "grad_norm": 0.014293976128101349,
+ "learning_rate": 0.0006,
+ "loss": 5.119932651519775,
+ "step": 611
+ },
+ {
+ "epoch": 8.503056768558952,
+ "grad_norm": 0.014453536830842495,
+ "learning_rate": 0.0006,
+ "loss": 5.166794776916504,
+ "step": 612
+ },
+ {
+ "epoch": 8.51703056768559,
+ "grad_norm": 0.014685769565403461,
+ "learning_rate": 0.0006,
+ "loss": 5.264822959899902,
+ "step": 613
+ },
+ {
+ "epoch": 8.531004366812226,
+ "grad_norm": 0.01427740603685379,
+ "learning_rate": 0.0006,
+ "loss": 5.278433799743652,
+ "step": 614
+ },
+ {
+ "epoch": 8.544978165938865,
+ "grad_norm": 0.014864951372146606,
+ "learning_rate": 0.0006,
+ "loss": 5.214512348175049,
+ "step": 615
+ },
+ {
+ "epoch": 8.558951965065502,
+ "grad_norm": 0.01565164513885975,
+ "learning_rate": 0.0006,
+ "loss": 5.183066368103027,
+ "step": 616
+ },
+ {
+ "epoch": 8.57292576419214,
+ "grad_norm": 0.01779816672205925,
+ "learning_rate": 0.0006,
+ "loss": 5.163185119628906,
+ "step": 617
+ },
+ {
+ "epoch": 8.586899563318777,
+ "grad_norm": 0.017254827544093132,
+ "learning_rate": 0.0006,
+ "loss": 5.243555068969727,
+ "step": 618
+ },
+ {
+ "epoch": 8.600873362445414,
+ "grad_norm": 0.01650584116578102,
+ "learning_rate": 0.0006,
+ "loss": 5.234884738922119,
+ "step": 619
+ },
+ {
+ "epoch": 8.614847161572053,
+ "grad_norm": 0.017021115869283676,
+ "learning_rate": 0.0006,
+ "loss": 5.12271785736084,
+ "step": 620
+ },
+ {
+ "epoch": 8.62882096069869,
+ "grad_norm": 0.01773759163916111,
+ "learning_rate": 0.0006,
+ "loss": 5.242153167724609,
+ "step": 621
+ },
+ {
+ "epoch": 8.642794759825328,
+ "grad_norm": 0.015679042786359787,
+ "learning_rate": 0.0006,
+ "loss": 5.170779228210449,
+ "step": 622
+ },
+ {
+ "epoch": 8.656768558951965,
+ "grad_norm": 0.013760549947619438,
+ "learning_rate": 0.0006,
+ "loss": 5.238644599914551,
+ "step": 623
+ },
+ {
+ "epoch": 8.670742358078602,
+ "grad_norm": 0.014571522362530231,
+ "learning_rate": 0.0006,
+ "loss": 5.177056312561035,
+ "step": 624
+ },
+ {
+ "epoch": 8.68471615720524,
+ "grad_norm": 0.016209015622735023,
+ "learning_rate": 0.0006,
+ "loss": 5.163750648498535,
+ "step": 625
+ },
+ {
+ "epoch": 8.698689956331878,
+ "grad_norm": 0.016813941299915314,
+ "learning_rate": 0.0006,
+ "loss": 5.183428764343262,
+ "step": 626
+ },
+ {
+ "epoch": 8.712663755458514,
+ "grad_norm": 0.020985357463359833,
+ "learning_rate": 0.0006,
+ "loss": 5.229465007781982,
+ "step": 627
+ },
+ {
+ "epoch": 8.726637554585153,
+ "grad_norm": 0.02679632417857647,
+ "learning_rate": 0.0006,
+ "loss": 5.257368087768555,
+ "step": 628
+ },
+ {
+ "epoch": 8.74061135371179,
+ "grad_norm": 0.025756070390343666,
+ "learning_rate": 0.0006,
+ "loss": 5.253736972808838,
+ "step": 629
+ },
+ {
+ "epoch": 8.754585152838429,
+ "grad_norm": 0.02643490768969059,
+ "learning_rate": 0.0006,
+ "loss": 5.228633403778076,
+ "step": 630
+ },
+ {
+ "epoch": 8.768558951965066,
+ "grad_norm": 0.03303210437297821,
+ "learning_rate": 0.0006,
+ "loss": 5.163540363311768,
+ "step": 631
+ },
+ {
+ "epoch": 8.782532751091702,
+ "grad_norm": 0.030432431027293205,
+ "learning_rate": 0.0006,
+ "loss": 5.243169784545898,
+ "step": 632
+ },
+ {
+ "epoch": 8.796506550218341,
+ "grad_norm": 0.028631288558244705,
+ "learning_rate": 0.0006,
+ "loss": 5.166398048400879,
+ "step": 633
+ },
+ {
+ "epoch": 8.810480349344978,
+ "grad_norm": 0.026188310235738754,
+ "learning_rate": 0.0006,
+ "loss": 5.128122329711914,
+ "step": 634
+ },
+ {
+ "epoch": 8.824454148471617,
+ "grad_norm": 0.025308528915047646,
+ "learning_rate": 0.0006,
+ "loss": 5.211284637451172,
+ "step": 635
+ },
+ {
+ "epoch": 8.838427947598253,
+ "grad_norm": 0.020657729357481003,
+ "learning_rate": 0.0006,
+ "loss": 5.161575794219971,
+ "step": 636
+ },
+ {
+ "epoch": 8.85240174672489,
+ "grad_norm": 0.021901234984397888,
+ "learning_rate": 0.0006,
+ "loss": 5.20050048828125,
+ "step": 637
+ },
+ {
+ "epoch": 8.866375545851529,
+ "grad_norm": 0.021219318732619286,
+ "learning_rate": 0.0006,
+ "loss": 5.211699485778809,
+ "step": 638
+ },
+ {
+ "epoch": 8.880349344978166,
+ "grad_norm": 0.0200974028557539,
+ "learning_rate": 0.0006,
+ "loss": 5.092503547668457,
+ "step": 639
+ },
+ {
+ "epoch": 8.894323144104803,
+ "grad_norm": 0.023804882541298866,
+ "learning_rate": 0.0006,
+ "loss": 5.216068267822266,
+ "step": 640
+ },
+ {
+ "epoch": 8.908296943231441,
+ "grad_norm": 0.026088010519742966,
+ "learning_rate": 0.0006,
+ "loss": 5.155592918395996,
+ "step": 641
+ },
+ {
+ "epoch": 8.922270742358078,
+ "grad_norm": 0.02601276896893978,
+ "learning_rate": 0.0006,
+ "loss": 5.197238922119141,
+ "step": 642
+ },
+ {
+ "epoch": 8.936244541484717,
+ "grad_norm": 0.020387211814522743,
+ "learning_rate": 0.0006,
+ "loss": 5.1955437660217285,
+ "step": 643
+ },
+ {
+ "epoch": 8.950218340611354,
+ "grad_norm": 0.019214622676372528,
+ "learning_rate": 0.0006,
+ "loss": 5.206346035003662,
+ "step": 644
+ },
+ {
+ "epoch": 8.96419213973799,
+ "grad_norm": 0.019674314185976982,
+ "learning_rate": 0.0006,
+ "loss": 5.219857215881348,
+ "step": 645
+ },
+ {
+ "epoch": 8.97816593886463,
+ "grad_norm": 0.020318234339356422,
+ "learning_rate": 0.0006,
+ "loss": 5.191132545471191,
+ "step": 646
+ },
+ {
+ "epoch": 8.992139737991266,
+ "grad_norm": 0.021428676322102547,
+ "learning_rate": 0.0006,
+ "loss": 5.214695930480957,
+ "step": 647
+ },
+ {
+ "epoch": 9.0,
+ "grad_norm": 0.023235054686665535,
+ "learning_rate": 0.0006,
+ "loss": 5.229091644287109,
+ "step": 648
+ },
+ {
+ "epoch": 9.0,
+ "eval_loss": 5.289593696594238,
+ "eval_runtime": 56.9256,
+ "eval_samples_per_second": 42.898,
+ "eval_steps_per_second": 1.353,
+ "step": 648
+ },
+ {
+ "epoch": 9.013973799126637,
+ "grad_norm": 0.02320289984345436,
+ "learning_rate": 0.0006,
+ "loss": 4.983669281005859,
+ "step": 649
+ },
+ {
+ "epoch": 9.027947598253276,
+ "grad_norm": 0.018521282821893692,
+ "learning_rate": 0.0006,
+ "loss": 5.200575828552246,
+ "step": 650
+ },
+ {
+ "epoch": 9.041921397379912,
+ "grad_norm": 0.0180169939994812,
+ "learning_rate": 0.0006,
+ "loss": 5.148033618927002,
+ "step": 651
+ },
+ {
+ "epoch": 9.055895196506551,
+ "grad_norm": 0.019969860091805458,
+ "learning_rate": 0.0006,
+ "loss": 5.099125385284424,
+ "step": 652
+ },
+ {
+ "epoch": 9.069868995633188,
+ "grad_norm": 0.01643305830657482,
+ "learning_rate": 0.0006,
+ "loss": 5.092347145080566,
+ "step": 653
+ },
+ {
+ "epoch": 9.083842794759825,
+ "grad_norm": 0.016022363677620888,
+ "learning_rate": 0.0006,
+ "loss": 5.129279613494873,
+ "step": 654
+ },
+ {
+ "epoch": 9.097816593886463,
+ "grad_norm": 0.01657041162252426,
+ "learning_rate": 0.0006,
+ "loss": 5.159562587738037,
+ "step": 655
+ },
+ {
+ "epoch": 9.1117903930131,
+ "grad_norm": 0.017924228683114052,
+ "learning_rate": 0.0006,
+ "loss": 5.165497779846191,
+ "step": 656
+ },
+ {
+ "epoch": 9.125764192139737,
+ "grad_norm": 0.021592361852526665,
+ "learning_rate": 0.0006,
+ "loss": 5.204775810241699,
+ "step": 657
+ },
+ {
+ "epoch": 9.139737991266376,
+ "grad_norm": 0.026924120262265205,
+ "learning_rate": 0.0006,
+ "loss": 5.128296852111816,
+ "step": 658
+ },
+ {
+ "epoch": 9.153711790393013,
+ "grad_norm": 0.023156899958848953,
+ "learning_rate": 0.0006,
+ "loss": 5.1255340576171875,
+ "step": 659
+ },
+ {
+ "epoch": 9.167685589519651,
+ "grad_norm": 0.020670367404818535,
+ "learning_rate": 0.0006,
+ "loss": 5.128479957580566,
+ "step": 660
+ },
+ {
+ "epoch": 9.181659388646288,
+ "grad_norm": 0.025447173044085503,
+ "learning_rate": 0.0006,
+ "loss": 5.154821872711182,
+ "step": 661
+ },
+ {
+ "epoch": 9.195633187772925,
+ "grad_norm": 0.027940068393945694,
+ "learning_rate": 0.0006,
+ "loss": 5.143466472625732,
+ "step": 662
+ },
+ {
+ "epoch": 9.209606986899564,
+ "grad_norm": 0.026847844943404198,
+ "learning_rate": 0.0006,
+ "loss": 5.045665740966797,
+ "step": 663
+ },
+ {
+ "epoch": 9.2235807860262,
+ "grad_norm": 0.02340601570904255,
+ "learning_rate": 0.0006,
+ "loss": 5.21933126449585,
+ "step": 664
+ },
+ {
+ "epoch": 9.237554585152838,
+ "grad_norm": 0.02340371161699295,
+ "learning_rate": 0.0006,
+ "loss": 4.942949295043945,
+ "step": 665
+ },
+ {
+ "epoch": 9.251528384279476,
+ "grad_norm": 0.02221992425620556,
+ "learning_rate": 0.0006,
+ "loss": 5.197099208831787,
+ "step": 666
+ },
+ {
+ "epoch": 9.265502183406113,
+ "grad_norm": 0.023508677259087563,
+ "learning_rate": 0.0006,
+ "loss": 5.022897720336914,
+ "step": 667
+ },
+ {
+ "epoch": 9.279475982532752,
+ "grad_norm": 0.026120394468307495,
+ "learning_rate": 0.0006,
+ "loss": 5.087725639343262,
+ "step": 668
+ },
+ {
+ "epoch": 9.293449781659389,
+ "grad_norm": 0.027273228392004967,
+ "learning_rate": 0.0006,
+ "loss": 5.090963840484619,
+ "step": 669
+ },
+ {
+ "epoch": 9.307423580786025,
+ "grad_norm": 0.03241586685180664,
+ "learning_rate": 0.0006,
+ "loss": 5.047842979431152,
+ "step": 670
+ },
+ {
+ "epoch": 9.321397379912664,
+ "grad_norm": 0.030245667323470116,
+ "learning_rate": 0.0006,
+ "loss": 5.104803085327148,
+ "step": 671
+ },
+ {
+ "epoch": 9.335371179039301,
+ "grad_norm": 0.027698364108800888,
+ "learning_rate": 0.0006,
+ "loss": 5.0644097328186035,
+ "step": 672
+ },
+ {
+ "epoch": 9.34934497816594,
+ "grad_norm": 0.029692554846405983,
+ "learning_rate": 0.0006,
+ "loss": 5.111942291259766,
+ "step": 673
+ },
+ {
+ "epoch": 9.363318777292577,
+ "grad_norm": 0.03328656405210495,
+ "learning_rate": 0.0006,
+ "loss": 5.1574554443359375,
+ "step": 674
+ },
+ {
+ "epoch": 9.377292576419213,
+ "grad_norm": 0.031938180327415466,
+ "learning_rate": 0.0006,
+ "loss": 5.190371990203857,
+ "step": 675
+ },
+ {
+ "epoch": 9.391266375545852,
+ "grad_norm": 0.024858945980668068,
+ "learning_rate": 0.0006,
+ "loss": 5.121420860290527,
+ "step": 676
+ },
+ {
+ "epoch": 9.405240174672489,
+ "grad_norm": 0.023033203557133675,
+ "learning_rate": 0.0006,
+ "loss": 5.0974345207214355,
+ "step": 677
+ },
+ {
+ "epoch": 9.419213973799126,
+ "grad_norm": 0.022083545103669167,
+ "learning_rate": 0.0006,
+ "loss": 5.046319961547852,
+ "step": 678
+ },
+ {
+ "epoch": 9.433187772925764,
+ "grad_norm": 0.0228437427431345,
+ "learning_rate": 0.0006,
+ "loss": 5.119963645935059,
+ "step": 679
+ },
+ {
+ "epoch": 9.447161572052401,
+ "grad_norm": 0.0242351982742548,
+ "learning_rate": 0.0006,
+ "loss": 5.0444560050964355,
+ "step": 680
+ },
+ {
+ "epoch": 9.46113537117904,
+ "grad_norm": 0.02401185780763626,
+ "learning_rate": 0.0006,
+ "loss": 5.063238620758057,
+ "step": 681
+ },
+ {
+ "epoch": 9.475109170305677,
+ "grad_norm": 0.02560959942638874,
+ "learning_rate": 0.0006,
+ "loss": 5.035224914550781,
+ "step": 682
+ },
+ {
+ "epoch": 9.489082969432314,
+ "grad_norm": 0.024985626339912415,
+ "learning_rate": 0.0006,
+ "loss": 5.049284934997559,
+ "step": 683
+ },
+ {
+ "epoch": 9.503056768558952,
+ "grad_norm": 0.023178115487098694,
+ "learning_rate": 0.0006,
+ "loss": 5.2266621589660645,
+ "step": 684
+ },
+ {
+ "epoch": 9.51703056768559,
+ "grad_norm": 0.0215512253344059,
+ "learning_rate": 0.0006,
+ "loss": 5.1875715255737305,
+ "step": 685
+ },
+ {
+ "epoch": 9.531004366812226,
+ "grad_norm": 0.01662837713956833,
+ "learning_rate": 0.0006,
+ "loss": 5.137012481689453,
+ "step": 686
+ },
+ {
+ "epoch": 9.544978165938865,
+ "grad_norm": 0.01952764391899109,
+ "learning_rate": 0.0006,
+ "loss": 5.072500228881836,
+ "step": 687
+ },
+ {
+ "epoch": 9.558951965065502,
+ "grad_norm": 0.017521077767014503,
+ "learning_rate": 0.0006,
+ "loss": 5.045711994171143,
+ "step": 688
+ },
+ {
+ "epoch": 9.57292576419214,
+ "grad_norm": 0.017355265095829964,
+ "learning_rate": 0.0006,
+ "loss": 5.02881383895874,
+ "step": 689
+ },
+ {
+ "epoch": 9.586899563318777,
+ "grad_norm": 0.015587719157338142,
+ "learning_rate": 0.0006,
+ "loss": 5.132787704467773,
+ "step": 690
+ },
+ {
+ "epoch": 9.600873362445414,
+ "grad_norm": 0.017469845712184906,
+ "learning_rate": 0.0006,
+ "loss": 5.105447292327881,
+ "step": 691
+ },
+ {
+ "epoch": 9.614847161572053,
+ "grad_norm": 0.019630100578069687,
+ "learning_rate": 0.0006,
+ "loss": 5.032464981079102,
+ "step": 692
+ },
+ {
+ "epoch": 9.62882096069869,
+ "grad_norm": 0.01784994639456272,
+ "learning_rate": 0.0006,
+ "loss": 5.081345558166504,
+ "step": 693
+ },
+ {
+ "epoch": 9.642794759825328,
+ "grad_norm": 0.01872754842042923,
+ "learning_rate": 0.0006,
+ "loss": 4.9772114753723145,
+ "step": 694
+ },
+ {
+ "epoch": 9.656768558951965,
+ "grad_norm": 0.01999032311141491,
+ "learning_rate": 0.0006,
+ "loss": 5.089740753173828,
+ "step": 695
+ },
+ {
+ "epoch": 9.670742358078602,
+ "grad_norm": 0.01710767112672329,
+ "learning_rate": 0.0006,
+ "loss": 4.9869890213012695,
+ "step": 696
+ },
+ {
+ "epoch": 9.68471615720524,
+ "grad_norm": 0.015144161880016327,
+ "learning_rate": 0.0006,
+ "loss": 5.022564888000488,
+ "step": 697
+ },
+ {
+ "epoch": 9.698689956331878,
+ "grad_norm": 0.01618543453514576,
+ "learning_rate": 0.0006,
+ "loss": 5.035921573638916,
+ "step": 698
+ },
+ {
+ "epoch": 9.712663755458514,
+ "grad_norm": 0.014653387479484081,
+ "learning_rate": 0.0006,
+ "loss": 4.914456367492676,
+ "step": 699
+ },
+ {
+ "epoch": 9.726637554585153,
+ "grad_norm": 0.013434977270662785,
+ "learning_rate": 0.0006,
+ "loss": 5.063764572143555,
+ "step": 700
+ },
+ {
+ "epoch": 9.74061135371179,
+ "grad_norm": 0.015128864906728268,
+ "learning_rate": 0.0006,
+ "loss": 5.052367687225342,
+ "step": 701
+ },
+ {
+ "epoch": 9.754585152838429,
+ "grad_norm": 0.014210768043994904,
+ "learning_rate": 0.0006,
+ "loss": 5.016489028930664,
+ "step": 702
+ },
+ {
+ "epoch": 9.768558951965066,
+ "grad_norm": 0.015191680751740932,
+ "learning_rate": 0.0006,
+ "loss": 4.990176200866699,
+ "step": 703
+ },
+ {
+ "epoch": 9.782532751091702,
+ "grad_norm": 0.018285127356648445,
+ "learning_rate": 0.0006,
+ "loss": 5.015164375305176,
+ "step": 704
+ },
+ {
+ "epoch": 9.796506550218341,
+ "grad_norm": 0.0204079058021307,
+ "learning_rate": 0.0006,
+ "loss": 5.091142654418945,
+ "step": 705
+ },
+ {
+ "epoch": 9.810480349344978,
+ "grad_norm": 0.01980586163699627,
+ "learning_rate": 0.0006,
+ "loss": 5.03350830078125,
+ "step": 706
+ },
+ {
+ "epoch": 9.824454148471617,
+ "grad_norm": 0.02048729732632637,
+ "learning_rate": 0.0006,
+ "loss": 5.03611421585083,
+ "step": 707
+ },
+ {
+ "epoch": 9.838427947598253,
+ "grad_norm": 0.02058163844048977,
+ "learning_rate": 0.0006,
+ "loss": 4.9613447189331055,
+ "step": 708
+ },
+ {
+ "epoch": 9.85240174672489,
+ "grad_norm": 0.018004924058914185,
+ "learning_rate": 0.0006,
+ "loss": 4.966541767120361,
+ "step": 709
+ },
+ {
+ "epoch": 9.866375545851529,
+ "grad_norm": 0.0198152307420969,
+ "learning_rate": 0.0006,
+ "loss": 5.01750373840332,
+ "step": 710
+ },
+ {
+ "epoch": 9.880349344978166,
+ "grad_norm": 0.018532969057559967,
+ "learning_rate": 0.0006,
+ "loss": 5.097424507141113,
+ "step": 711
+ },
+ {
+ "epoch": 9.894323144104803,
+ "grad_norm": 0.018939971923828125,
+ "learning_rate": 0.0006,
+ "loss": 5.177223205566406,
+ "step": 712
+ },
+ {
+ "epoch": 9.908296943231441,
+ "grad_norm": 0.017969170585274696,
+ "learning_rate": 0.0006,
+ "loss": 5.118013858795166,
+ "step": 713
+ },
+ {
+ "epoch": 9.922270742358078,
+ "grad_norm": 0.018021270632743835,
+ "learning_rate": 0.0006,
+ "loss": 5.034787178039551,
+ "step": 714
+ },
+ {
+ "epoch": 9.936244541484717,
+ "grad_norm": 0.017009710893034935,
+ "learning_rate": 0.0006,
+ "loss": 5.003920555114746,
+ "step": 715
+ },
+ {
+ "epoch": 9.950218340611354,
+ "grad_norm": 0.017722049728035927,
+ "learning_rate": 0.0006,
+ "loss": 5.031939506530762,
+ "step": 716
+ },
+ {
+ "epoch": 9.96419213973799,
+ "grad_norm": 0.017134087160229683,
+ "learning_rate": 0.0006,
+ "loss": 5.093474388122559,
+ "step": 717
+ },
+ {
+ "epoch": 9.97816593886463,
+ "grad_norm": 0.018511613830924034,
+ "learning_rate": 0.0006,
+ "loss": 5.078604698181152,
+ "step": 718
+ },
+ {
+ "epoch": 9.992139737991266,
+ "grad_norm": 0.021206334233283997,
+ "learning_rate": 0.0006,
+ "loss": 5.052572727203369,
+ "step": 719
+ },
+ {
+ "epoch": 10.0,
+ "grad_norm": 0.02239062264561653,
+ "learning_rate": 0.0006,
+ "loss": 5.063547134399414,
+ "step": 720
+ },
+ {
+ "epoch": 10.0,
+ "eval_loss": 5.173044681549072,
+ "eval_runtime": 56.7287,
+ "eval_samples_per_second": 43.047,
+ "eval_steps_per_second": 1.357,
+ "step": 720
+ },
+ {
+ "epoch": 10.013973799126637,
+ "grad_norm": 0.020757969468832016,
+ "learning_rate": 0.0006,
+ "loss": 5.045648574829102,
+ "step": 721
+ },
+ {
+ "epoch": 10.027947598253276,
+ "grad_norm": 0.022697702050209045,
+ "learning_rate": 0.0006,
+ "loss": 4.993833065032959,
+ "step": 722
+ },
+ {
+ "epoch": 10.041921397379912,
+ "grad_norm": 0.02434331737458706,
+ "learning_rate": 0.0006,
+ "loss": 5.073219299316406,
+ "step": 723
+ },
+ {
+ "epoch": 10.055895196506551,
+ "grad_norm": 0.025499660521745682,
+ "learning_rate": 0.0006,
+ "loss": 5.019072532653809,
+ "step": 724
+ },
+ {
+ "epoch": 10.069868995633188,
+ "grad_norm": 0.029281053692102432,
+ "learning_rate": 0.0006,
+ "loss": 4.959851264953613,
+ "step": 725
+ },
+ {
+ "epoch": 10.083842794759825,
+ "grad_norm": 0.033878110349178314,
+ "learning_rate": 0.0006,
+ "loss": 5.132717609405518,
+ "step": 726
+ },
+ {
+ "epoch": 10.097816593886463,
+ "grad_norm": 0.03385764732956886,
+ "learning_rate": 0.0006,
+ "loss": 5.0482892990112305,
+ "step": 727
+ },
+ {
+ "epoch": 10.1117903930131,
+ "grad_norm": 0.03466648980975151,
+ "learning_rate": 0.0006,
+ "loss": 4.946855545043945,
+ "step": 728
+ },
+ {
+ "epoch": 10.125764192139737,
+ "grad_norm": 0.028546305373311043,
+ "learning_rate": 0.0006,
+ "loss": 5.027774333953857,
+ "step": 729
+ },
+ {
+ "epoch": 10.139737991266376,
+ "grad_norm": 0.030575869604945183,
+ "learning_rate": 0.0006,
+ "loss": 4.9768476486206055,
+ "step": 730
+ },
+ {
+ "epoch": 10.153711790393013,
+ "grad_norm": 0.03614303469657898,
+ "learning_rate": 0.0006,
+ "loss": 5.078113555908203,
+ "step": 731
+ },
+ {
+ "epoch": 10.167685589519651,
+ "grad_norm": 0.040991660207509995,
+ "learning_rate": 0.0006,
+ "loss": 5.033185958862305,
+ "step": 732
+ },
+ {
+ "epoch": 10.181659388646288,
+ "grad_norm": 0.04551699757575989,
+ "learning_rate": 0.0006,
+ "loss": 5.0820746421813965,
+ "step": 733
+ },
+ {
+ "epoch": 10.195633187772925,
+ "grad_norm": 0.040193989872932434,
+ "learning_rate": 0.0006,
+ "loss": 5.059700965881348,
+ "step": 734
+ },
+ {
+ "epoch": 10.209606986899564,
+ "grad_norm": 0.035851605236530304,
+ "learning_rate": 0.0006,
+ "loss": 5.142467498779297,
+ "step": 735
+ },
+ {
+ "epoch": 10.2235807860262,
+ "grad_norm": 0.034200169146060944,
+ "learning_rate": 0.0006,
+ "loss": 5.072312355041504,
+ "step": 736
+ },
+ {
+ "epoch": 10.237554585152838,
+ "grad_norm": 0.033621110022068024,
+ "learning_rate": 0.0006,
+ "loss": 5.08037805557251,
+ "step": 737
+ },
+ {
+ "epoch": 10.251528384279476,
+ "grad_norm": 0.037372369319200516,
+ "learning_rate": 0.0006,
+ "loss": 4.998425483703613,
+ "step": 738
+ },
+ {
+ "epoch": 10.265502183406113,
+ "grad_norm": 0.037044707685709,
+ "learning_rate": 0.0006,
+ "loss": 4.923993110656738,
+ "step": 739
+ },
+ {
+ "epoch": 10.279475982532752,
+ "grad_norm": 0.030454808846116066,
+ "learning_rate": 0.0006,
+ "loss": 5.0889997482299805,
+ "step": 740
+ },
+ {
+ "epoch": 10.293449781659389,
+ "grad_norm": 0.030969638377428055,
+ "learning_rate": 0.0006,
+ "loss": 4.984112739562988,
+ "step": 741
+ },
+ {
+ "epoch": 10.307423580786025,
+ "grad_norm": 0.0276983380317688,
+ "learning_rate": 0.0006,
+ "loss": 4.955412864685059,
+ "step": 742
+ },
+ {
+ "epoch": 10.321397379912664,
+ "grad_norm": 0.02716052532196045,
+ "learning_rate": 0.0006,
+ "loss": 5.0098876953125,
+ "step": 743
+ },
+ {
+ "epoch": 10.335371179039301,
+ "grad_norm": 0.02403552085161209,
+ "learning_rate": 0.0006,
+ "loss": 5.165160655975342,
+ "step": 744
+ },
+ {
+ "epoch": 10.34934497816594,
+ "grad_norm": 0.0257862601429224,
+ "learning_rate": 0.0006,
+ "loss": 5.001547813415527,
+ "step": 745
+ },
+ {
+ "epoch": 10.363318777292577,
+ "grad_norm": 0.02300378680229187,
+ "learning_rate": 0.0006,
+ "loss": 5.038888931274414,
+ "step": 746
+ },
+ {
+ "epoch": 10.377292576419213,
+ "grad_norm": 0.019766854122281075,
+ "learning_rate": 0.0006,
+ "loss": 4.924291610717773,
+ "step": 747
+ },
+ {
+ "epoch": 10.391266375545852,
+ "grad_norm": 0.01888395845890045,
+ "learning_rate": 0.0006,
+ "loss": 5.059971332550049,
+ "step": 748
+ },
+ {
+ "epoch": 10.405240174672489,
+ "grad_norm": 0.015577499754726887,
+ "learning_rate": 0.0006,
+ "loss": 4.987686634063721,
+ "step": 749
+ },
+ {
+ "epoch": 10.419213973799126,
+ "grad_norm": 0.016012346372008324,
+ "learning_rate": 0.0006,
+ "loss": 5.015157699584961,
+ "step": 750
+ },
+ {
+ "epoch": 10.433187772925764,
+ "grad_norm": 0.016277998685836792,
+ "learning_rate": 0.0006,
+ "loss": 5.004931449890137,
+ "step": 751
+ },
+ {
+ "epoch": 10.447161572052401,
+ "grad_norm": 0.017504561692476273,
+ "learning_rate": 0.0006,
+ "loss": 4.902827739715576,
+ "step": 752
+ },
+ {
+ "epoch": 10.46113537117904,
+ "grad_norm": 0.01569022797048092,
+ "learning_rate": 0.0006,
+ "loss": 4.954863548278809,
+ "step": 753
+ },
+ {
+ "epoch": 10.475109170305677,
+ "grad_norm": 0.014114422723650932,
+ "learning_rate": 0.0006,
+ "loss": 5.00656795501709,
+ "step": 754
+ },
+ {
+ "epoch": 10.489082969432314,
+ "grad_norm": 0.01573832333087921,
+ "learning_rate": 0.0006,
+ "loss": 4.978353023529053,
+ "step": 755
+ },
+ {
+ "epoch": 10.503056768558952,
+ "grad_norm": 0.016480036079883575,
+ "learning_rate": 0.0006,
+ "loss": 4.953327655792236,
+ "step": 756
+ },
+ {
+ "epoch": 10.51703056768559,
+ "grad_norm": 0.014008025638759136,
+ "learning_rate": 0.0006,
+ "loss": 5.050840377807617,
+ "step": 757
+ },
+ {
+ "epoch": 10.531004366812226,
+ "grad_norm": 0.013147998601198196,
+ "learning_rate": 0.0006,
+ "loss": 4.974964141845703,
+ "step": 758
+ },
+ {
+ "epoch": 10.544978165938865,
+ "grad_norm": 0.013807603158056736,
+ "learning_rate": 0.0006,
+ "loss": 4.927907943725586,
+ "step": 759
+ },
+ {
+ "epoch": 10.558951965065502,
+ "grad_norm": 0.013555224984884262,
+ "learning_rate": 0.0006,
+ "loss": 4.980690956115723,
+ "step": 760
+ },
+ {
+ "epoch": 10.57292576419214,
+ "grad_norm": 0.013045408762991428,
+ "learning_rate": 0.0006,
+ "loss": 4.9436116218566895,
+ "step": 761
+ },
+ {
+ "epoch": 10.586899563318777,
+ "grad_norm": 0.012858862057328224,
+ "learning_rate": 0.0006,
+ "loss": 4.9595770835876465,
+ "step": 762
+ },
+ {
+ "epoch": 10.600873362445414,
+ "grad_norm": 0.01468253880739212,
+ "learning_rate": 0.0006,
+ "loss": 4.90725040435791,
+ "step": 763
+ },
+ {
+ "epoch": 10.614847161572053,
+ "grad_norm": 0.013224679045379162,
+ "learning_rate": 0.0006,
+ "loss": 4.877615928649902,
+ "step": 764
+ },
+ {
+ "epoch": 10.62882096069869,
+ "grad_norm": 0.013555348850786686,
+ "learning_rate": 0.0006,
+ "loss": 4.9456610679626465,
+ "step": 765
+ },
+ {
+ "epoch": 10.642794759825328,
+ "grad_norm": 0.014387160539627075,
+ "learning_rate": 0.0006,
+ "loss": 4.9233717918396,
+ "step": 766
+ },
+ {
+ "epoch": 10.656768558951965,
+ "grad_norm": 0.014354088343679905,
+ "learning_rate": 0.0006,
+ "loss": 4.983511924743652,
+ "step": 767
+ },
+ {
+ "epoch": 10.670742358078602,
+ "grad_norm": 0.01332700252532959,
+ "learning_rate": 0.0006,
+ "loss": 4.934416770935059,
+ "step": 768
+ },
+ {
+ "epoch": 10.68471615720524,
+ "grad_norm": 0.011543313041329384,
+ "learning_rate": 0.0006,
+ "loss": 4.96388053894043,
+ "step": 769
+ },
+ {
+ "epoch": 10.698689956331878,
+ "grad_norm": 0.012330878525972366,
+ "learning_rate": 0.0006,
+ "loss": 4.946396350860596,
+ "step": 770
+ },
+ {
+ "epoch": 10.712663755458514,
+ "grad_norm": 0.01379779726266861,
+ "learning_rate": 0.0006,
+ "loss": 4.981019020080566,
+ "step": 771
+ },
+ {
+ "epoch": 10.726637554585153,
+ "grad_norm": 0.014612431637942791,
+ "learning_rate": 0.0006,
+ "loss": 4.894567966461182,
+ "step": 772
+ },
+ {
+ "epoch": 10.74061135371179,
+ "grad_norm": 0.013062899932265282,
+ "learning_rate": 0.0006,
+ "loss": 5.033038139343262,
+ "step": 773
+ },
+ {
+ "epoch": 10.754585152838429,
+ "grad_norm": 0.013898049481213093,
+ "learning_rate": 0.0006,
+ "loss": 4.823636054992676,
+ "step": 774
+ },
+ {
+ "epoch": 10.768558951965066,
+ "grad_norm": 0.01502019353210926,
+ "learning_rate": 0.0006,
+ "loss": 5.020156383514404,
+ "step": 775
+ },
+ {
+ "epoch": 10.782532751091702,
+ "grad_norm": 0.013810204342007637,
+ "learning_rate": 0.0006,
+ "loss": 4.935647010803223,
+ "step": 776
+ },
+ {
+ "epoch": 10.796506550218341,
+ "grad_norm": 0.013616513460874557,
+ "learning_rate": 0.0006,
+ "loss": 4.901486396789551,
+ "step": 777
+ },
+ {
+ "epoch": 10.810480349344978,
+ "grad_norm": 0.013601024635136127,
+ "learning_rate": 0.0006,
+ "loss": 4.931873321533203,
+ "step": 778
+ },
+ {
+ "epoch": 10.824454148471617,
+ "grad_norm": 0.013668350875377655,
+ "learning_rate": 0.0006,
+ "loss": 4.937256813049316,
+ "step": 779
+ },
+ {
+ "epoch": 10.838427947598253,
+ "grad_norm": 0.01581823080778122,
+ "learning_rate": 0.0006,
+ "loss": 4.9611406326293945,
+ "step": 780
+ },
+ {
+ "epoch": 10.85240174672489,
+ "grad_norm": 0.019478484988212585,
+ "learning_rate": 0.0006,
+ "loss": 4.9212236404418945,
+ "step": 781
+ },
+ {
+ "epoch": 10.866375545851529,
+ "grad_norm": 0.026921333745121956,
+ "learning_rate": 0.0006,
+ "loss": 4.88501501083374,
+ "step": 782
+ },
+ {
+ "epoch": 10.880349344978166,
+ "grad_norm": 0.03155914694070816,
+ "learning_rate": 0.0006,
+ "loss": 4.989967346191406,
+ "step": 783
+ },
+ {
+ "epoch": 10.894323144104803,
+ "grad_norm": 0.029689429327845573,
+ "learning_rate": 0.0006,
+ "loss": 4.936000823974609,
+ "step": 784
+ },
+ {
+ "epoch": 10.908296943231441,
+ "grad_norm": 0.026626063510775566,
+ "learning_rate": 0.0006,
+ "loss": 4.984785079956055,
+ "step": 785
+ },
+ {
+ "epoch": 10.922270742358078,
+ "grad_norm": 0.026168212294578552,
+ "learning_rate": 0.0006,
+ "loss": 4.908282279968262,
+ "step": 786
+ },
+ {
+ "epoch": 10.936244541484717,
+ "grad_norm": 0.02828158251941204,
+ "learning_rate": 0.0006,
+ "loss": 4.926390647888184,
+ "step": 787
+ },
+ {
+ "epoch": 10.950218340611354,
+ "grad_norm": 0.02649078331887722,
+ "learning_rate": 0.0006,
+ "loss": 5.00400972366333,
+ "step": 788
+ },
+ {
+ "epoch": 10.96419213973799,
+ "grad_norm": 0.0256856270134449,
+ "learning_rate": 0.0006,
+ "loss": 5.0317487716674805,
+ "step": 789
+ },
+ {
+ "epoch": 10.97816593886463,
+ "grad_norm": 0.024728883057832718,
+ "learning_rate": 0.0006,
+ "loss": 4.9299635887146,
+ "step": 790
+ },
+ {
+ "epoch": 10.992139737991266,
+ "grad_norm": 0.026241116225719452,
+ "learning_rate": 0.0006,
+ "loss": 4.947319984436035,
+ "step": 791
+ },
+ {
+ "epoch": 11.0,
+ "grad_norm": 0.023526743054389954,
+ "learning_rate": 0.0006,
+ "loss": 4.94429349899292,
+ "step": 792
+ },
+ {
+ "epoch": 11.0,
+ "eval_loss": 5.104345798492432,
+ "eval_runtime": 57.0955,
+ "eval_samples_per_second": 42.77,
+ "eval_steps_per_second": 1.349,
+ "step": 792
+ },
+ {
+ "epoch": 11.013973799126637,
+ "grad_norm": 0.0216788612306118,
+ "learning_rate": 0.0006,
+ "loss": 4.881950855255127,
+ "step": 793
+ },
+ {
+ "epoch": 11.027947598253276,
+ "grad_norm": 0.01830463856458664,
+ "learning_rate": 0.0006,
+ "loss": 4.971016883850098,
+ "step": 794
+ },
+ {
+ "epoch": 11.041921397379912,
+ "grad_norm": 0.02204521745443344,
+ "learning_rate": 0.0006,
+ "loss": 4.894623756408691,
+ "step": 795
+ },
+ {
+ "epoch": 11.055895196506551,
+ "grad_norm": 0.0210255216807127,
+ "learning_rate": 0.0006,
+ "loss": 4.834009170532227,
+ "step": 796
+ },
+ {
+ "epoch": 11.069868995633188,
+ "grad_norm": 0.019797448068857193,
+ "learning_rate": 0.0006,
+ "loss": 5.086714744567871,
+ "step": 797
+ },
+ {
+ "epoch": 11.083842794759825,
+ "grad_norm": 0.02036641724407673,
+ "learning_rate": 0.0006,
+ "loss": 4.920056343078613,
+ "step": 798
+ },
+ {
+ "epoch": 11.097816593886463,
+ "grad_norm": 0.01966066285967827,
+ "learning_rate": 0.0006,
+ "loss": 4.879859447479248,
+ "step": 799
+ },
+ {
+ "epoch": 11.1117903930131,
+ "grad_norm": 0.0202149897813797,
+ "learning_rate": 0.0006,
+ "loss": 4.933640480041504,
+ "step": 800
+ },
+ {
+ "epoch": 11.125764192139737,
+ "grad_norm": 0.01893039606511593,
+ "learning_rate": 0.0006,
+ "loss": 4.901970863342285,
+ "step": 801
+ },
+ {
+ "epoch": 11.139737991266376,
+ "grad_norm": 0.016040155664086342,
+ "learning_rate": 0.0006,
+ "loss": 4.905491828918457,
+ "step": 802
+ },
+ {
+ "epoch": 11.153711790393013,
+ "grad_norm": 0.01720350608229637,
+ "learning_rate": 0.0006,
+ "loss": 4.798816680908203,
+ "step": 803
+ },
+ {
+ "epoch": 11.167685589519651,
+ "grad_norm": 0.017002852633595467,
+ "learning_rate": 0.0006,
+ "loss": 4.851859092712402,
+ "step": 804
+ },
+ {
+ "epoch": 11.181659388646288,
+ "grad_norm": 0.015594680793583393,
+ "learning_rate": 0.0006,
+ "loss": 4.90328311920166,
+ "step": 805
+ },
+ {
+ "epoch": 11.195633187772925,
+ "grad_norm": 0.014904462732374668,
+ "learning_rate": 0.0006,
+ "loss": 4.904215335845947,
+ "step": 806
+ },
+ {
+ "epoch": 11.209606986899564,
+ "grad_norm": 0.01366228237748146,
+ "learning_rate": 0.0006,
+ "loss": 4.9813055992126465,
+ "step": 807
+ },
+ {
+ "epoch": 11.2235807860262,
+ "grad_norm": 0.014719638973474503,
+ "learning_rate": 0.0006,
+ "loss": 4.927306175231934,
+ "step": 808
+ },
+ {
+ "epoch": 11.237554585152838,
+ "grad_norm": 0.015016036108136177,
+ "learning_rate": 0.0006,
+ "loss": 4.874215126037598,
+ "step": 809
+ },
+ {
+ "epoch": 11.251528384279476,
+ "grad_norm": 0.014925424009561539,
+ "learning_rate": 0.0006,
+ "loss": 4.838257789611816,
+ "step": 810
+ },
+ {
+ "epoch": 11.265502183406113,
+ "grad_norm": 0.014335823245346546,
+ "learning_rate": 0.0006,
+ "loss": 4.940608024597168,
+ "step": 811
+ },
+ {
+ "epoch": 11.279475982532752,
+ "grad_norm": 0.014987512491643429,
+ "learning_rate": 0.0006,
+ "loss": 4.831531524658203,
+ "step": 812
+ },
+ {
+ "epoch": 11.293449781659389,
+ "grad_norm": 0.016087457537651062,
+ "learning_rate": 0.0006,
+ "loss": 4.863227367401123,
+ "step": 813
+ },
+ {
+ "epoch": 11.307423580786025,
+ "grad_norm": 0.01671300269663334,
+ "learning_rate": 0.0006,
+ "loss": 4.8535661697387695,
+ "step": 814
+ },
+ {
+ "epoch": 11.321397379912664,
+ "grad_norm": 0.015229142270982265,
+ "learning_rate": 0.0006,
+ "loss": 4.948635101318359,
+ "step": 815
+ },
+ {
+ "epoch": 11.335371179039301,
+ "grad_norm": 0.01532587967813015,
+ "learning_rate": 0.0006,
+ "loss": 4.9035797119140625,
+ "step": 816
+ },
+ {
+ "epoch": 11.34934497816594,
+ "grad_norm": 0.015715476125478745,
+ "learning_rate": 0.0006,
+ "loss": 4.894709587097168,
+ "step": 817
+ },
+ {
+ "epoch": 11.363318777292577,
+ "grad_norm": 0.014156977646052837,
+ "learning_rate": 0.0006,
+ "loss": 4.960206985473633,
+ "step": 818
+ },
+ {
+ "epoch": 11.377292576419213,
+ "grad_norm": 0.017076566815376282,
+ "learning_rate": 0.0006,
+ "loss": 5.005850791931152,
+ "step": 819
+ },
+ {
+ "epoch": 11.391266375545852,
+ "grad_norm": 0.0172084029763937,
+ "learning_rate": 0.0006,
+ "loss": 4.903355598449707,
+ "step": 820
+ },
+ {
+ "epoch": 11.405240174672489,
+ "grad_norm": 0.0180258359760046,
+ "learning_rate": 0.0006,
+ "loss": 4.842667579650879,
+ "step": 821
+ },
+ {
+ "epoch": 11.419213973799126,
+ "grad_norm": 0.022130120545625687,
+ "learning_rate": 0.0006,
+ "loss": 4.799970626831055,
+ "step": 822
+ },
+ {
+ "epoch": 11.433187772925764,
+ "grad_norm": 0.02384897693991661,
+ "learning_rate": 0.0006,
+ "loss": 4.860746383666992,
+ "step": 823
+ },
+ {
+ "epoch": 11.447161572052401,
+ "grad_norm": 0.021881572902202606,
+ "learning_rate": 0.0006,
+ "loss": 4.918107032775879,
+ "step": 824
+ },
+ {
+ "epoch": 11.46113537117904,
+ "grad_norm": 0.023085080087184906,
+ "learning_rate": 0.0006,
+ "loss": 4.878946304321289,
+ "step": 825
+ },
+ {
+ "epoch": 11.475109170305677,
+ "grad_norm": 0.02291012555360794,
+ "learning_rate": 0.0006,
+ "loss": 4.958887100219727,
+ "step": 826
+ },
+ {
+ "epoch": 11.489082969432314,
+ "grad_norm": 0.0199968870729208,
+ "learning_rate": 0.0006,
+ "loss": 4.80955696105957,
+ "step": 827
+ },
+ {
+ "epoch": 11.503056768558952,
+ "grad_norm": 0.018702229484915733,
+ "learning_rate": 0.0006,
+ "loss": 5.0276408195495605,
+ "step": 828
+ },
+ {
+ "epoch": 11.51703056768559,
+ "grad_norm": 0.017787311226129532,
+ "learning_rate": 0.0006,
+ "loss": 4.96018123626709,
+ "step": 829
+ },
+ {
+ "epoch": 11.531004366812226,
+ "grad_norm": 0.01704501546919346,
+ "learning_rate": 0.0006,
+ "loss": 4.914553165435791,
+ "step": 830
+ },
+ {
+ "epoch": 11.544978165938865,
+ "grad_norm": 0.017519677057862282,
+ "learning_rate": 0.0006,
+ "loss": 4.798098087310791,
+ "step": 831
+ },
+ {
+ "epoch": 11.558951965065502,
+ "grad_norm": 0.018511991947889328,
+ "learning_rate": 0.0006,
+ "loss": 4.817785263061523,
+ "step": 832
+ },
+ {
+ "epoch": 11.57292576419214,
+ "grad_norm": 0.01876644790172577,
+ "learning_rate": 0.0006,
+ "loss": 4.818995475769043,
+ "step": 833
+ },
+ {
+ "epoch": 11.586899563318777,
+ "grad_norm": 0.01753568835556507,
+ "learning_rate": 0.0006,
+ "loss": 4.811938285827637,
+ "step": 834
+ },
+ {
+ "epoch": 11.600873362445414,
+ "grad_norm": 0.017046531662344933,
+ "learning_rate": 0.0006,
+ "loss": 4.914674282073975,
+ "step": 835
+ },
+ {
+ "epoch": 11.614847161572053,
+ "grad_norm": 0.01802094094455242,
+ "learning_rate": 0.0006,
+ "loss": 4.993284225463867,
+ "step": 836
+ },
+ {
+ "epoch": 11.62882096069869,
+ "grad_norm": 0.019493764266371727,
+ "learning_rate": 0.0006,
+ "loss": 4.895628452301025,
+ "step": 837
+ },
+ {
+ "epoch": 11.642794759825328,
+ "grad_norm": 0.019461210817098618,
+ "learning_rate": 0.0006,
+ "loss": 4.869269371032715,
+ "step": 838
+ },
+ {
+ "epoch": 11.656768558951965,
+ "grad_norm": 0.01981567218899727,
+ "learning_rate": 0.0006,
+ "loss": 4.863052845001221,
+ "step": 839
+ },
+ {
+ "epoch": 11.670742358078602,
+ "grad_norm": 0.019894491881132126,
+ "learning_rate": 0.0006,
+ "loss": 4.8775787353515625,
+ "step": 840
+ },
+ {
+ "epoch": 11.68471615720524,
+ "grad_norm": 0.01875786855816841,
+ "learning_rate": 0.0006,
+ "loss": 4.817763328552246,
+ "step": 841
+ },
+ {
+ "epoch": 11.698689956331878,
+ "grad_norm": 0.020926695317029953,
+ "learning_rate": 0.0006,
+ "loss": 4.7470526695251465,
+ "step": 842
+ },
+ {
+ "epoch": 11.712663755458514,
+ "grad_norm": 0.021477103233337402,
+ "learning_rate": 0.0006,
+ "loss": 4.762682914733887,
+ "step": 843
+ },
+ {
+ "epoch": 11.726637554585153,
+ "grad_norm": 0.020628711208701134,
+ "learning_rate": 0.0006,
+ "loss": 4.807985782623291,
+ "step": 844
+ },
+ {
+ "epoch": 11.74061135371179,
+ "grad_norm": 0.021489612758159637,
+ "learning_rate": 0.0006,
+ "loss": 4.843091011047363,
+ "step": 845
+ },
+ {
+ "epoch": 11.754585152838429,
+ "grad_norm": 0.021194491535425186,
+ "learning_rate": 0.0006,
+ "loss": 4.8300251960754395,
+ "step": 846
+ },
+ {
+ "epoch": 11.768558951965066,
+ "grad_norm": 0.01833350397646427,
+ "learning_rate": 0.0006,
+ "loss": 5.000202178955078,
+ "step": 847
+ },
+ {
+ "epoch": 11.782532751091702,
+ "grad_norm": 0.018096931278705597,
+ "learning_rate": 0.0006,
+ "loss": 4.880648136138916,
+ "step": 848
+ },
+ {
+ "epoch": 11.796506550218341,
+ "grad_norm": 0.0197161752730608,
+ "learning_rate": 0.0006,
+ "loss": 4.788316249847412,
+ "step": 849
+ },
+ {
+ "epoch": 11.810480349344978,
+ "grad_norm": 0.0203181654214859,
+ "learning_rate": 0.0006,
+ "loss": 4.870944976806641,
+ "step": 850
+ },
+ {
+ "epoch": 11.824454148471617,
+ "grad_norm": 0.01908455230295658,
+ "learning_rate": 0.0006,
+ "loss": 4.890144348144531,
+ "step": 851
+ },
+ {
+ "epoch": 11.838427947598253,
+ "grad_norm": 0.021281801164150238,
+ "learning_rate": 0.0006,
+ "loss": 4.89578914642334,
+ "step": 852
+ },
+ {
+ "epoch": 11.85240174672489,
+ "grad_norm": 0.02124273031949997,
+ "learning_rate": 0.0006,
+ "loss": 4.883749008178711,
+ "step": 853
+ },
+ {
+ "epoch": 11.866375545851529,
+ "grad_norm": 0.020969398319721222,
+ "learning_rate": 0.0006,
+ "loss": 4.809142589569092,
+ "step": 854
+ },
+ {
+ "epoch": 11.880349344978166,
+ "grad_norm": 0.01967657171189785,
+ "learning_rate": 0.0006,
+ "loss": 4.821864604949951,
+ "step": 855
+ },
+ {
+ "epoch": 11.894323144104803,
+ "grad_norm": 0.020887333899736404,
+ "learning_rate": 0.0006,
+ "loss": 4.917985916137695,
+ "step": 856
+ },
+ {
+ "epoch": 11.908296943231441,
+ "grad_norm": 0.019203083589673042,
+ "learning_rate": 0.0006,
+ "loss": 4.824960708618164,
+ "step": 857
+ },
+ {
+ "epoch": 11.922270742358078,
+ "grad_norm": 0.020162401720881462,
+ "learning_rate": 0.0006,
+ "loss": 4.772617816925049,
+ "step": 858
+ },
+ {
+ "epoch": 11.936244541484717,
+ "grad_norm": 0.02251487225294113,
+ "learning_rate": 0.0006,
+ "loss": 4.735418319702148,
+ "step": 859
+ },
+ {
+ "epoch": 11.950218340611354,
+ "grad_norm": 0.020313527435064316,
+ "learning_rate": 0.0006,
+ "loss": 4.939233779907227,
+ "step": 860
+ },
+ {
+ "epoch": 11.96419213973799,
+ "grad_norm": 0.021635599434375763,
+ "learning_rate": 0.0006,
+ "loss": 4.861807346343994,
+ "step": 861
+ },
+ {
+ "epoch": 11.97816593886463,
+ "grad_norm": 0.02285311557352543,
+ "learning_rate": 0.0006,
+ "loss": 4.932732582092285,
+ "step": 862
+ },
+ {
+ "epoch": 11.992139737991266,
+ "grad_norm": 0.02072535641491413,
+ "learning_rate": 0.0006,
+ "loss": 4.796767711639404,
+ "step": 863
+ },
+ {
+ "epoch": 12.0,
+ "grad_norm": 0.020588304847478867,
+ "learning_rate": 0.0006,
+ "loss": 4.997537612915039,
+ "step": 864
+ },
+ {
+ "epoch": 12.0,
+ "eval_loss": 5.044651985168457,
+ "eval_runtime": 56.9145,
+ "eval_samples_per_second": 42.906,
+ "eval_steps_per_second": 1.353,
+ "step": 864
+ },
+ {
+ "epoch": 12.013973799126637,
+ "grad_norm": 0.017597166821360588,
+ "learning_rate": 0.0006,
+ "loss": 4.830524444580078,
+ "step": 865
+ },
+ {
+ "epoch": 12.027947598253276,
+ "grad_norm": 0.01716788113117218,
+ "learning_rate": 0.0006,
+ "loss": 4.894325256347656,
+ "step": 866
+ },
+ {
+ "epoch": 12.041921397379912,
+ "grad_norm": 0.017712457105517387,
+ "learning_rate": 0.0006,
+ "loss": 4.868297576904297,
+ "step": 867
+ },
+ {
+ "epoch": 12.055895196506551,
+ "grad_norm": 0.021486390382051468,
+ "learning_rate": 0.0006,
+ "loss": 4.922049522399902,
+ "step": 868
+ },
+ {
+ "epoch": 12.069868995633188,
+ "grad_norm": 0.024560654535889626,
+ "learning_rate": 0.0006,
+ "loss": 4.734241008758545,
+ "step": 869
+ },
+ {
+ "epoch": 12.083842794759825,
+ "grad_norm": 0.025870252400636673,
+ "learning_rate": 0.0006,
+ "loss": 4.868566989898682,
+ "step": 870
+ },
+ {
+ "epoch": 12.097816593886463,
+ "grad_norm": 0.0277389045804739,
+ "learning_rate": 0.0006,
+ "loss": 4.755006790161133,
+ "step": 871
+ },
+ {
+ "epoch": 12.1117903930131,
+ "grad_norm": 0.029100844636559486,
+ "learning_rate": 0.0006,
+ "loss": 4.909667015075684,
+ "step": 872
+ },
+ {
+ "epoch": 12.125764192139737,
+ "grad_norm": 0.031833067536354065,
+ "learning_rate": 0.0006,
+ "loss": 4.858674049377441,
+ "step": 873
+ },
+ {
+ "epoch": 12.139737991266376,
+ "grad_norm": 0.032709237188100815,
+ "learning_rate": 0.0006,
+ "loss": 4.823696136474609,
+ "step": 874
+ },
+ {
+ "epoch": 12.153711790393013,
+ "grad_norm": 0.02995380386710167,
+ "learning_rate": 0.0006,
+ "loss": 4.850235462188721,
+ "step": 875
+ },
+ {
+ "epoch": 12.167685589519651,
+ "grad_norm": 0.034098610281944275,
+ "learning_rate": 0.0006,
+ "loss": 4.8169989585876465,
+ "step": 876
+ },
+ {
+ "epoch": 12.181659388646288,
+ "grad_norm": 0.03377080336213112,
+ "learning_rate": 0.0006,
+ "loss": 4.90738582611084,
+ "step": 877
+ },
+ {
+ "epoch": 12.195633187772925,
+ "grad_norm": 0.029780825600028038,
+ "learning_rate": 0.0006,
+ "loss": 4.811163902282715,
+ "step": 878
+ },
+ {
+ "epoch": 12.209606986899564,
+ "grad_norm": 0.03118024580180645,
+ "learning_rate": 0.0006,
+ "loss": 4.861286640167236,
+ "step": 879
+ },
+ {
+ "epoch": 12.2235807860262,
+ "grad_norm": 0.033052004873752594,
+ "learning_rate": 0.0006,
+ "loss": 4.865021228790283,
+ "step": 880
+ },
+ {
+ "epoch": 12.237554585152838,
+ "grad_norm": 0.029070032760500908,
+ "learning_rate": 0.0006,
+ "loss": 4.947303771972656,
+ "step": 881
+ },
+ {
+ "epoch": 12.251528384279476,
+ "grad_norm": 0.028355715796351433,
+ "learning_rate": 0.0006,
+ "loss": 4.804648399353027,
+ "step": 882
+ },
+ {
+ "epoch": 12.265502183406113,
+ "grad_norm": 0.031257227063179016,
+ "learning_rate": 0.0006,
+ "loss": 4.757718086242676,
+ "step": 883
+ },
+ {
+ "epoch": 12.279475982532752,
+ "grad_norm": 0.031438373029232025,
+ "learning_rate": 0.0006,
+ "loss": 4.851268768310547,
+ "step": 884
+ },
+ {
+ "epoch": 12.293449781659389,
+ "grad_norm": 0.030492138117551804,
+ "learning_rate": 0.0006,
+ "loss": 4.7891035079956055,
+ "step": 885
+ },
+ {
+ "epoch": 12.307423580786025,
+ "grad_norm": 0.026115991175174713,
+ "learning_rate": 0.0006,
+ "loss": 4.86474609375,
+ "step": 886
+ },
+ {
+ "epoch": 12.321397379912664,
+ "grad_norm": 0.024184465408325195,
+ "learning_rate": 0.0006,
+ "loss": 4.860456943511963,
+ "step": 887
+ },
+ {
+ "epoch": 12.335371179039301,
+ "grad_norm": 0.025437982752919197,
+ "learning_rate": 0.0006,
+ "loss": 4.806251049041748,
+ "step": 888
+ },
+ {
+ "epoch": 12.34934497816594,
+ "grad_norm": 0.022390231490135193,
+ "learning_rate": 0.0006,
+ "loss": 4.807086944580078,
+ "step": 889
+ },
+ {
+ "epoch": 12.363318777292577,
+ "grad_norm": 0.020174294710159302,
+ "learning_rate": 0.0006,
+ "loss": 4.78292179107666,
+ "step": 890
+ },
+ {
+ "epoch": 12.377292576419213,
+ "grad_norm": 0.019734324887394905,
+ "learning_rate": 0.0006,
+ "loss": 4.784509658813477,
+ "step": 891
+ },
+ {
+ "epoch": 12.391266375545852,
+ "grad_norm": 0.016733523458242416,
+ "learning_rate": 0.0006,
+ "loss": 4.811755180358887,
+ "step": 892
+ },
+ {
+ "epoch": 12.405240174672489,
+ "grad_norm": 0.01570146717131138,
+ "learning_rate": 0.0006,
+ "loss": 4.720088005065918,
+ "step": 893
+ },
+ {
+ "epoch": 12.419213973799126,
+ "grad_norm": 0.01601393334567547,
+ "learning_rate": 0.0006,
+ "loss": 4.783543109893799,
+ "step": 894
+ },
+ {
+ "epoch": 12.433187772925764,
+ "grad_norm": 0.014035487547516823,
+ "learning_rate": 0.0006,
+ "loss": 4.848340034484863,
+ "step": 895
+ },
+ {
+ "epoch": 12.447161572052401,
+ "grad_norm": 0.015248487703502178,
+ "learning_rate": 0.0006,
+ "loss": 4.773287773132324,
+ "step": 896
+ },
+ {
+ "epoch": 12.46113537117904,
+ "grad_norm": 0.01532268151640892,
+ "learning_rate": 0.0006,
+ "loss": 4.799646854400635,
+ "step": 897
+ },
+ {
+ "epoch": 12.475109170305677,
+ "grad_norm": 0.013982534408569336,
+ "learning_rate": 0.0006,
+ "loss": 4.8090500831604,
+ "step": 898
+ },
+ {
+ "epoch": 12.489082969432314,
+ "grad_norm": 0.014081164263188839,
+ "learning_rate": 0.0006,
+ "loss": 4.751888751983643,
+ "step": 899
+ },
+ {
+ "epoch": 12.503056768558952,
+ "grad_norm": 0.01380133256316185,
+ "learning_rate": 0.0006,
+ "loss": 4.78144645690918,
+ "step": 900
+ },
+ {
+ "epoch": 12.51703056768559,
+ "grad_norm": 0.013555269688367844,
+ "learning_rate": 0.0006,
+ "loss": 4.779257774353027,
+ "step": 901
+ },
+ {
+ "epoch": 12.531004366812226,
+ "grad_norm": 0.013075701892375946,
+ "learning_rate": 0.0006,
+ "loss": 4.820705413818359,
+ "step": 902
+ },
+ {
+ "epoch": 12.544978165938865,
+ "grad_norm": 0.01276316400617361,
+ "learning_rate": 0.0006,
+ "loss": 4.836733818054199,
+ "step": 903
+ },
+ {
+ "epoch": 12.558951965065502,
+ "grad_norm": 0.013033892959356308,
+ "learning_rate": 0.0006,
+ "loss": 4.764178276062012,
+ "step": 904
+ },
+ {
+ "epoch": 12.57292576419214,
+ "grad_norm": 0.013996980153024197,
+ "learning_rate": 0.0006,
+ "loss": 4.767983436584473,
+ "step": 905
+ },
+ {
+ "epoch": 12.586899563318777,
+ "grad_norm": 0.015222184360027313,
+ "learning_rate": 0.0006,
+ "loss": 4.759721755981445,
+ "step": 906
+ },
+ {
+ "epoch": 12.600873362445414,
+ "grad_norm": 0.01623941957950592,
+ "learning_rate": 0.0006,
+ "loss": 4.849811553955078,
+ "step": 907
+ },
+ {
+ "epoch": 12.614847161572053,
+ "grad_norm": 0.015711182728409767,
+ "learning_rate": 0.0006,
+ "loss": 4.7829084396362305,
+ "step": 908
+ },
+ {
+ "epoch": 12.62882096069869,
+ "grad_norm": 0.014137467369437218,
+ "learning_rate": 0.0006,
+ "loss": 4.73321533203125,
+ "step": 909
+ },
+ {
+ "epoch": 12.642794759825328,
+ "grad_norm": 0.012738462537527084,
+ "learning_rate": 0.0006,
+ "loss": 4.870052337646484,
+ "step": 910
+ },
+ {
+ "epoch": 12.656768558951965,
+ "grad_norm": 0.013683130033314228,
+ "learning_rate": 0.0006,
+ "loss": 4.82658052444458,
+ "step": 911
+ },
+ {
+ "epoch": 12.670742358078602,
+ "grad_norm": 0.014409830793738365,
+ "learning_rate": 0.0006,
+ "loss": 4.707494735717773,
+ "step": 912
+ },
+ {
+ "epoch": 12.68471615720524,
+ "grad_norm": 0.013806809671223164,
+ "learning_rate": 0.0006,
+ "loss": 4.778928279876709,
+ "step": 913
+ },
+ {
+ "epoch": 12.698689956331878,
+ "grad_norm": 0.012857912108302116,
+ "learning_rate": 0.0006,
+ "loss": 4.890717506408691,
+ "step": 914
+ },
+ {
+ "epoch": 12.712663755458514,
+ "grad_norm": 0.013828758150339127,
+ "learning_rate": 0.0006,
+ "loss": 4.8357439041137695,
+ "step": 915
+ },
+ {
+ "epoch": 12.726637554585153,
+ "grad_norm": 0.01478351280093193,
+ "learning_rate": 0.0006,
+ "loss": 4.869986534118652,
+ "step": 916
+ },
+ {
+ "epoch": 12.74061135371179,
+ "grad_norm": 0.01326421182602644,
+ "learning_rate": 0.0006,
+ "loss": 4.830646514892578,
+ "step": 917
+ },
+ {
+ "epoch": 12.754585152838429,
+ "grad_norm": 0.013335433788597584,
+ "learning_rate": 0.0006,
+ "loss": 4.7786149978637695,
+ "step": 918
+ },
+ {
+ "epoch": 12.768558951965066,
+ "grad_norm": 0.014261147007346153,
+ "learning_rate": 0.0006,
+ "loss": 4.788693428039551,
+ "step": 919
+ },
+ {
+ "epoch": 12.782532751091702,
+ "grad_norm": 0.014274783432483673,
+ "learning_rate": 0.0006,
+ "loss": 4.77295446395874,
+ "step": 920
+ },
+ {
+ "epoch": 12.796506550218341,
+ "grad_norm": 0.015074674971401691,
+ "learning_rate": 0.0006,
+ "loss": 4.830754280090332,
+ "step": 921
+ },
+ {
+ "epoch": 12.810480349344978,
+ "grad_norm": 0.015599515289068222,
+ "learning_rate": 0.0006,
+ "loss": 4.837759017944336,
+ "step": 922
+ },
+ {
+ "epoch": 12.824454148471617,
+ "grad_norm": 0.014595242217183113,
+ "learning_rate": 0.0006,
+ "loss": 4.882965564727783,
+ "step": 923
+ },
+ {
+ "epoch": 12.838427947598253,
+ "grad_norm": 0.014951467514038086,
+ "learning_rate": 0.0006,
+ "loss": 4.709748268127441,
+ "step": 924
+ },
+ {
+ "epoch": 12.85240174672489,
+ "grad_norm": 0.01539295818656683,
+ "learning_rate": 0.0006,
+ "loss": 4.791160583496094,
+ "step": 925
+ },
+ {
+ "epoch": 12.866375545851529,
+ "grad_norm": 0.015987906605005264,
+ "learning_rate": 0.0006,
+ "loss": 4.661876678466797,
+ "step": 926
+ },
+ {
+ "epoch": 12.880349344978166,
+ "grad_norm": 0.016527341678738594,
+ "learning_rate": 0.0006,
+ "loss": 4.721952438354492,
+ "step": 927
+ },
+ {
+ "epoch": 12.894323144104803,
+ "grad_norm": 0.01637161523103714,
+ "learning_rate": 0.0006,
+ "loss": 4.704446792602539,
+ "step": 928
+ },
+ {
+ "epoch": 12.908296943231441,
+ "grad_norm": 0.015221050009131432,
+ "learning_rate": 0.0006,
+ "loss": 4.771907806396484,
+ "step": 929
+ },
+ {
+ "epoch": 12.922270742358078,
+ "grad_norm": 0.016242166981101036,
+ "learning_rate": 0.0006,
+ "loss": 4.806267738342285,
+ "step": 930
+ },
+ {
+ "epoch": 12.936244541484717,
+ "grad_norm": 0.01628357358276844,
+ "learning_rate": 0.0006,
+ "loss": 4.7252326011657715,
+ "step": 931
+ },
+ {
+ "epoch": 12.950218340611354,
+ "grad_norm": 0.017428133636713028,
+ "learning_rate": 0.0006,
+ "loss": 4.765689849853516,
+ "step": 932
+ },
+ {
+ "epoch": 12.96419213973799,
+ "grad_norm": 0.01870061084628105,
+ "learning_rate": 0.0006,
+ "loss": 4.814488887786865,
+ "step": 933
+ },
+ {
+ "epoch": 12.97816593886463,
+ "grad_norm": 0.019796060398221016,
+ "learning_rate": 0.0006,
+ "loss": 4.739185810089111,
+ "step": 934
+ },
+ {
+ "epoch": 12.992139737991266,
+ "grad_norm": 0.01985820196568966,
+ "learning_rate": 0.0006,
+ "loss": 4.810608863830566,
+ "step": 935
+ },
+ {
+ "epoch": 13.0,
+ "grad_norm": 0.020834477618336678,
+ "learning_rate": 0.0006,
+ "loss": 4.727858066558838,
+ "step": 936
+ },
+ {
+ "epoch": 13.0,
+ "eval_loss": 4.979116916656494,
+ "eval_runtime": 57.2877,
+ "eval_samples_per_second": 42.627,
+ "eval_steps_per_second": 1.344,
+ "step": 936
+ },
+ {
+ "epoch": 13.013973799126637,
+ "grad_norm": 0.02025519125163555,
+ "learning_rate": 0.0006,
+ "loss": 4.628875732421875,
+ "step": 937
+ },
+ {
+ "epoch": 13.027947598253276,
+ "grad_norm": 0.019596368074417114,
+ "learning_rate": 0.0006,
+ "loss": 4.694699287414551,
+ "step": 938
+ },
+ {
+ "epoch": 13.041921397379912,
+ "grad_norm": 0.020494934171438217,
+ "learning_rate": 0.0006,
+ "loss": 4.691821098327637,
+ "step": 939
+ },
+ {
+ "epoch": 13.055895196506551,
+ "grad_norm": 0.020121760666370392,
+ "learning_rate": 0.0006,
+ "loss": 4.71478271484375,
+ "step": 940
+ },
+ {
+ "epoch": 13.069868995633188,
+ "grad_norm": 0.021091405302286148,
+ "learning_rate": 0.0006,
+ "loss": 4.695915222167969,
+ "step": 941
+ },
+ {
+ "epoch": 13.083842794759825,
+ "grad_norm": 0.019787423312664032,
+ "learning_rate": 0.0006,
+ "loss": 4.8652448654174805,
+ "step": 942
+ },
+ {
+ "epoch": 13.097816593886463,
+ "grad_norm": 0.0185833927243948,
+ "learning_rate": 0.0006,
+ "loss": 4.7065887451171875,
+ "step": 943
+ },
+ {
+ "epoch": 13.1117903930131,
+ "grad_norm": 0.021220475435256958,
+ "learning_rate": 0.0006,
+ "loss": 4.741232395172119,
+ "step": 944
+ },
+ {
+ "epoch": 13.125764192139737,
+ "grad_norm": 0.02136622555553913,
+ "learning_rate": 0.0006,
+ "loss": 4.705967903137207,
+ "step": 945
+ },
+ {
+ "epoch": 13.139737991266376,
+ "grad_norm": 0.0191587433218956,
+ "learning_rate": 0.0006,
+ "loss": 4.758667945861816,
+ "step": 946
+ },
+ {
+ "epoch": 13.153711790393013,
+ "grad_norm": 0.018290022388100624,
+ "learning_rate": 0.0006,
+ "loss": 4.760666847229004,
+ "step": 947
+ },
+ {
+ "epoch": 13.167685589519651,
+ "grad_norm": 0.021704409271478653,
+ "learning_rate": 0.0006,
+ "loss": 4.750555992126465,
+ "step": 948
+ },
+ {
+ "epoch": 13.181659388646288,
+ "grad_norm": 0.021668918430805206,
+ "learning_rate": 0.0006,
+ "loss": 4.840903282165527,
+ "step": 949
+ },
+ {
+ "epoch": 13.195633187772925,
+ "grad_norm": 0.019497141242027283,
+ "learning_rate": 0.0006,
+ "loss": 4.738969326019287,
+ "step": 950
+ },
+ {
+ "epoch": 13.209606986899564,
+ "grad_norm": 0.018500596284866333,
+ "learning_rate": 0.0006,
+ "loss": 4.75154972076416,
+ "step": 951
+ },
+ {
+ "epoch": 13.2235807860262,
+ "grad_norm": 0.01872239261865616,
+ "learning_rate": 0.0006,
+ "loss": 4.811738014221191,
+ "step": 952
+ },
+ {
+ "epoch": 13.237554585152838,
+ "grad_norm": 0.020321547985076904,
+ "learning_rate": 0.0006,
+ "loss": 4.815384387969971,
+ "step": 953
+ },
+ {
+ "epoch": 13.251528384279476,
+ "grad_norm": 0.02283575013279915,
+ "learning_rate": 0.0006,
+ "loss": 4.74582052230835,
+ "step": 954
+ },
+ {
+ "epoch": 13.265502183406113,
+ "grad_norm": 0.02474398910999298,
+ "learning_rate": 0.0006,
+ "loss": 4.738375186920166,
+ "step": 955
+ },
+ {
+ "epoch": 13.279475982532752,
+ "grad_norm": 0.02065764181315899,
+ "learning_rate": 0.0006,
+ "loss": 4.746875286102295,
+ "step": 956
+ },
+ {
+ "epoch": 13.293449781659389,
+ "grad_norm": 0.020534520968794823,
+ "learning_rate": 0.0006,
+ "loss": 4.656914710998535,
+ "step": 957
+ },
+ {
+ "epoch": 13.307423580786025,
+ "grad_norm": 0.0206378772854805,
+ "learning_rate": 0.0006,
+ "loss": 4.842615604400635,
+ "step": 958
+ },
+ {
+ "epoch": 13.321397379912664,
+ "grad_norm": 0.017935654148459435,
+ "learning_rate": 0.0006,
+ "loss": 4.717827796936035,
+ "step": 959
+ },
+ {
+ "epoch": 13.335371179039301,
+ "grad_norm": 0.016958363354206085,
+ "learning_rate": 0.0006,
+ "loss": 4.681070327758789,
+ "step": 960
+ },
+ {
+ "epoch": 13.34934497816594,
+ "grad_norm": 0.015735412016510963,
+ "learning_rate": 0.0006,
+ "loss": 4.7957048416137695,
+ "step": 961
+ },
+ {
+ "epoch": 13.363318777292577,
+ "grad_norm": 0.01563699170947075,
+ "learning_rate": 0.0006,
+ "loss": 4.701192378997803,
+ "step": 962
+ },
+ {
+ "epoch": 13.377292576419213,
+ "grad_norm": 0.015207760035991669,
+ "learning_rate": 0.0006,
+ "loss": 4.73256254196167,
+ "step": 963
+ },
+ {
+ "epoch": 13.391266375545852,
+ "grad_norm": 0.015396883711218834,
+ "learning_rate": 0.0006,
+ "loss": 4.76412296295166,
+ "step": 964
+ },
+ {
+ "epoch": 13.405240174672489,
+ "grad_norm": 0.015690583735704422,
+ "learning_rate": 0.0006,
+ "loss": 4.791157245635986,
+ "step": 965
+ },
+ {
+ "epoch": 13.419213973799126,
+ "grad_norm": 0.01755724661052227,
+ "learning_rate": 0.0006,
+ "loss": 4.845666885375977,
+ "step": 966
+ },
+ {
+ "epoch": 13.433187772925764,
+ "grad_norm": 0.019443422555923462,
+ "learning_rate": 0.0006,
+ "loss": 4.5902299880981445,
+ "step": 967
+ },
+ {
+ "epoch": 13.447161572052401,
+ "grad_norm": 0.02118338644504547,
+ "learning_rate": 0.0006,
+ "loss": 4.740627288818359,
+ "step": 968
+ },
+ {
+ "epoch": 13.46113537117904,
+ "grad_norm": 0.02174839936196804,
+ "learning_rate": 0.0006,
+ "loss": 4.814560890197754,
+ "step": 969
+ },
+ {
+ "epoch": 13.475109170305677,
+ "grad_norm": 0.022860374301671982,
+ "learning_rate": 0.0006,
+ "loss": 4.66440486907959,
+ "step": 970
+ },
+ {
+ "epoch": 13.489082969432314,
+ "grad_norm": 0.024742677807807922,
+ "learning_rate": 0.0006,
+ "loss": 4.7451653480529785,
+ "step": 971
+ },
+ {
+ "epoch": 13.503056768558952,
+ "grad_norm": 0.027068907395005226,
+ "learning_rate": 0.0006,
+ "loss": 4.802584171295166,
+ "step": 972
+ },
+ {
+ "epoch": 13.51703056768559,
+ "grad_norm": 0.026316625997424126,
+ "learning_rate": 0.0006,
+ "loss": 4.842397689819336,
+ "step": 973
+ },
+ {
+ "epoch": 13.531004366812226,
+ "grad_norm": 0.024610962718725204,
+ "learning_rate": 0.0006,
+ "loss": 4.650721073150635,
+ "step": 974
+ },
+ {
+ "epoch": 13.544978165938865,
+ "grad_norm": 0.024122603237628937,
+ "learning_rate": 0.0006,
+ "loss": 4.6984357833862305,
+ "step": 975
+ },
+ {
+ "epoch": 13.558951965065502,
+ "grad_norm": 0.022699708119034767,
+ "learning_rate": 0.0006,
+ "loss": 4.793007850646973,
+ "step": 976
+ },
+ {
+ "epoch": 13.57292576419214,
+ "grad_norm": 0.021706534549593925,
+ "learning_rate": 0.0006,
+ "loss": 4.691362380981445,
+ "step": 977
+ },
+ {
+ "epoch": 13.586899563318777,
+ "grad_norm": 0.021363291889429092,
+ "learning_rate": 0.0006,
+ "loss": 4.76762580871582,
+ "step": 978
+ },
+ {
+ "epoch": 13.600873362445414,
+ "grad_norm": 0.023476677015423775,
+ "learning_rate": 0.0006,
+ "loss": 4.777202606201172,
+ "step": 979
+ },
+ {
+ "epoch": 13.614847161572053,
+ "grad_norm": 0.02457100711762905,
+ "learning_rate": 0.0006,
+ "loss": 4.590703010559082,
+ "step": 980
+ },
+ {
+ "epoch": 13.62882096069869,
+ "grad_norm": 0.02469099499285221,
+ "learning_rate": 0.0006,
+ "loss": 4.640185356140137,
+ "step": 981
+ },
+ {
+ "epoch": 13.642794759825328,
+ "grad_norm": 0.024914603680372238,
+ "learning_rate": 0.0006,
+ "loss": 4.772201061248779,
+ "step": 982
+ },
+ {
+ "epoch": 13.656768558951965,
+ "grad_norm": 0.025133173912763596,
+ "learning_rate": 0.0006,
+ "loss": 4.743337154388428,
+ "step": 983
+ },
+ {
+ "epoch": 13.670742358078602,
+ "grad_norm": 0.02034137211740017,
+ "learning_rate": 0.0006,
+ "loss": 4.752127647399902,
+ "step": 984
+ },
+ {
+ "epoch": 13.68471615720524,
+ "grad_norm": 0.019645661115646362,
+ "learning_rate": 0.0006,
+ "loss": 4.838453769683838,
+ "step": 985
+ },
+ {
+ "epoch": 13.698689956331878,
+ "grad_norm": 0.01900297962129116,
+ "learning_rate": 0.0006,
+ "loss": 4.804868221282959,
+ "step": 986
+ },
+ {
+ "epoch": 13.712663755458514,
+ "grad_norm": 0.020292341709136963,
+ "learning_rate": 0.0006,
+ "loss": 4.609743118286133,
+ "step": 987
+ },
+ {
+ "epoch": 13.726637554585153,
+ "grad_norm": 0.017014775425195694,
+ "learning_rate": 0.0006,
+ "loss": 4.740612983703613,
+ "step": 988
+ },
+ {
+ "epoch": 13.74061135371179,
+ "grad_norm": 0.015199844725430012,
+ "learning_rate": 0.0006,
+ "loss": 4.719701766967773,
+ "step": 989
+ },
+ {
+ "epoch": 13.754585152838429,
+ "grad_norm": 0.015125414356589317,
+ "learning_rate": 0.0006,
+ "loss": 4.686114311218262,
+ "step": 990
+ },
+ {
+ "epoch": 13.768558951965066,
+ "grad_norm": 0.014779540710151196,
+ "learning_rate": 0.0006,
+ "loss": 4.674347877502441,
+ "step": 991
+ },
+ {
+ "epoch": 13.782532751091702,
+ "grad_norm": 0.015255163423717022,
+ "learning_rate": 0.0006,
+ "loss": 4.6309309005737305,
+ "step": 992
+ },
+ {
+ "epoch": 13.796506550218341,
+ "grad_norm": 0.013692197389900684,
+ "learning_rate": 0.0006,
+ "loss": 4.698497295379639,
+ "step": 993
+ },
+ {
+ "epoch": 13.810480349344978,
+ "grad_norm": 0.014529623091220856,
+ "learning_rate": 0.0006,
+ "loss": 4.628443717956543,
+ "step": 994
+ },
+ {
+ "epoch": 13.824454148471617,
+ "grad_norm": 0.01576659083366394,
+ "learning_rate": 0.0006,
+ "loss": 4.788015842437744,
+ "step": 995
+ },
+ {
+ "epoch": 13.838427947598253,
+ "grad_norm": 0.015834277495741844,
+ "learning_rate": 0.0006,
+ "loss": 4.672703742980957,
+ "step": 996
+ },
+ {
+ "epoch": 13.85240174672489,
+ "grad_norm": 0.015539503656327724,
+ "learning_rate": 0.0006,
+ "loss": 4.731503486633301,
+ "step": 997
+ },
+ {
+ "epoch": 13.866375545851529,
+ "grad_norm": 0.013463746756315231,
+ "learning_rate": 0.0006,
+ "loss": 4.7006330490112305,
+ "step": 998
+ },
+ {
+ "epoch": 13.880349344978166,
+ "grad_norm": 0.016377726569771767,
+ "learning_rate": 0.0006,
+ "loss": 4.636074066162109,
+ "step": 999
+ },
+ {
+ "epoch": 13.894323144104803,
+ "grad_norm": 0.016660034656524658,
+ "learning_rate": 0.0006,
+ "loss": 4.742693901062012,
+ "step": 1000
+ },
+ {
+ "epoch": 13.908296943231441,
+ "grad_norm": 0.014902938157320023,
+ "learning_rate": 0.0006,
+ "loss": 4.769811630249023,
+ "step": 1001
+ },
+ {
+ "epoch": 13.922270742358078,
+ "grad_norm": 0.014883476309478283,
+ "learning_rate": 0.0006,
+ "loss": 4.731159210205078,
+ "step": 1002
+ },
+ {
+ "epoch": 13.936244541484717,
+ "grad_norm": 0.014494217932224274,
+ "learning_rate": 0.0006,
+ "loss": 4.659485816955566,
+ "step": 1003
+ },
+ {
+ "epoch": 13.950218340611354,
+ "grad_norm": 0.015426991507411003,
+ "learning_rate": 0.0006,
+ "loss": 4.755229949951172,
+ "step": 1004
+ },
+ {
+ "epoch": 13.96419213973799,
+ "grad_norm": 0.016648586839437485,
+ "learning_rate": 0.0006,
+ "loss": 4.74111270904541,
+ "step": 1005
+ },
+ {
+ "epoch": 13.97816593886463,
+ "grad_norm": 0.016980541869997978,
+ "learning_rate": 0.0006,
+ "loss": 4.74072265625,
+ "step": 1006
+ },
+ {
+ "epoch": 13.992139737991266,
+ "grad_norm": 0.018070152029395103,
+ "learning_rate": 0.0006,
+ "loss": 4.732975006103516,
+ "step": 1007
+ },
+ {
+ "epoch": 14.0,
+ "grad_norm": 0.019910665228962898,
+ "learning_rate": 0.0006,
+ "loss": 4.634697914123535,
+ "step": 1008
+ },
+ {
+ "epoch": 14.0,
+ "eval_loss": 4.850640296936035,
+ "eval_runtime": 57.3039,
+ "eval_samples_per_second": 42.615,
+ "eval_steps_per_second": 1.344,
+ "step": 1008
+ },
+ {
+ "epoch": 14.013973799126637,
+ "grad_norm": 0.01980053074657917,
+ "learning_rate": 0.0006,
+ "loss": 4.628744602203369,
+ "step": 1009
+ },
+ {
+ "epoch": 14.027947598253276,
+ "grad_norm": 0.02497623674571514,
+ "learning_rate": 0.0006,
+ "loss": 4.75930118560791,
+ "step": 1010
+ },
+ {
+ "epoch": 14.041921397379912,
+ "grad_norm": 0.02794097363948822,
+ "learning_rate": 0.0006,
+ "loss": 4.691340446472168,
+ "step": 1011
+ },
+ {
+ "epoch": 14.055895196506551,
+ "grad_norm": 0.023725535720586777,
+ "learning_rate": 0.0006,
+ "loss": 4.733664512634277,
+ "step": 1012
+ },
+ {
+ "epoch": 14.069868995633188,
+ "grad_norm": 0.022966474294662476,
+ "learning_rate": 0.0006,
+ "loss": 4.664808750152588,
+ "step": 1013
+ },
+ {
+ "epoch": 14.083842794759825,
+ "grad_norm": 0.02530563250184059,
+ "learning_rate": 0.0006,
+ "loss": 4.8047285079956055,
+ "step": 1014
+ },
+ {
+ "epoch": 14.097816593886463,
+ "grad_norm": 0.02491082064807415,
+ "learning_rate": 0.0006,
+ "loss": 4.667956352233887,
+ "step": 1015
+ },
+ {
+ "epoch": 14.1117903930131,
+ "grad_norm": 0.022110294550657272,
+ "learning_rate": 0.0006,
+ "loss": 4.666140556335449,
+ "step": 1016
+ },
+ {
+ "epoch": 14.125764192139737,
+ "grad_norm": 0.02065849117934704,
+ "learning_rate": 0.0006,
+ "loss": 4.581960678100586,
+ "step": 1017
+ },
+ {
+ "epoch": 14.139737991266376,
+ "grad_norm": 0.01941298507153988,
+ "learning_rate": 0.0006,
+ "loss": 4.614226341247559,
+ "step": 1018
+ },
+ {
+ "epoch": 14.153711790393013,
+ "grad_norm": 0.022441856563091278,
+ "learning_rate": 0.0006,
+ "loss": 4.638352394104004,
+ "step": 1019
+ },
+ {
+ "epoch": 14.167685589519651,
+ "grad_norm": 0.021519694477319717,
+ "learning_rate": 0.0006,
+ "loss": 4.700160980224609,
+ "step": 1020
+ },
+ {
+ "epoch": 14.181659388646288,
+ "grad_norm": 0.018956486135721207,
+ "learning_rate": 0.0006,
+ "loss": 4.649593353271484,
+ "step": 1021
+ },
+ {
+ "epoch": 14.195633187772925,
+ "grad_norm": 0.016681140288710594,
+ "learning_rate": 0.0006,
+ "loss": 4.603952407836914,
+ "step": 1022
+ },
+ {
+ "epoch": 14.209606986899564,
+ "grad_norm": 0.01788809522986412,
+ "learning_rate": 0.0006,
+ "loss": 4.669160842895508,
+ "step": 1023
+ },
+ {
+ "epoch": 14.2235807860262,
+ "grad_norm": 0.016047228127717972,
+ "learning_rate": 0.0006,
+ "loss": 4.6367950439453125,
+ "step": 1024
+ },
+ {
+ "epoch": 14.237554585152838,
+ "grad_norm": 0.01614239625632763,
+ "learning_rate": 0.0006,
+ "loss": 4.664529323577881,
+ "step": 1025
+ },
+ {
+ "epoch": 14.251528384279476,
+ "grad_norm": 0.016362549737095833,
+ "learning_rate": 0.0006,
+ "loss": 4.702373504638672,
+ "step": 1026
+ },
+ {
+ "epoch": 14.265502183406113,
+ "grad_norm": 0.015800409018993378,
+ "learning_rate": 0.0006,
+ "loss": 4.7594428062438965,
+ "step": 1027
+ },
+ {
+ "epoch": 14.279475982532752,
+ "grad_norm": 0.016730139032006264,
+ "learning_rate": 0.0006,
+ "loss": 4.677244186401367,
+ "step": 1028
+ },
+ {
+ "epoch": 14.293449781659389,
+ "grad_norm": 0.01661759801208973,
+ "learning_rate": 0.0006,
+ "loss": 4.633679389953613,
+ "step": 1029
+ },
+ {
+ "epoch": 14.307423580786025,
+ "grad_norm": 0.017673347145318985,
+ "learning_rate": 0.0006,
+ "loss": 4.649107456207275,
+ "step": 1030
+ },
+ {
+ "epoch": 14.321397379912664,
+ "grad_norm": 0.019243352115154266,
+ "learning_rate": 0.0006,
+ "loss": 4.611257553100586,
+ "step": 1031
+ },
+ {
+ "epoch": 14.335371179039301,
+ "grad_norm": 0.019817164167761803,
+ "learning_rate": 0.0006,
+ "loss": 4.588086128234863,
+ "step": 1032
+ },
+ {
+ "epoch": 14.34934497816594,
+ "grad_norm": 0.022292211651802063,
+ "learning_rate": 0.0006,
+ "loss": 4.783066749572754,
+ "step": 1033
+ },
+ {
+ "epoch": 14.363318777292577,
+ "grad_norm": 0.02140096202492714,
+ "learning_rate": 0.0006,
+ "loss": 4.604718208312988,
+ "step": 1034
+ },
+ {
+ "epoch": 14.377292576419213,
+ "grad_norm": 0.0187817495316267,
+ "learning_rate": 0.0006,
+ "loss": 4.721773147583008,
+ "step": 1035
+ },
+ {
+ "epoch": 14.391266375545852,
+ "grad_norm": 0.019879210740327835,
+ "learning_rate": 0.0006,
+ "loss": 4.789286136627197,
+ "step": 1036
+ },
+ {
+ "epoch": 14.405240174672489,
+ "grad_norm": 0.01885785162448883,
+ "learning_rate": 0.0006,
+ "loss": 4.669743537902832,
+ "step": 1037
+ },
+ {
+ "epoch": 14.419213973799126,
+ "grad_norm": 0.018324270844459534,
+ "learning_rate": 0.0006,
+ "loss": 4.65901517868042,
+ "step": 1038
+ },
+ {
+ "epoch": 14.433187772925764,
+ "grad_norm": 0.017664821818470955,
+ "learning_rate": 0.0006,
+ "loss": 4.667839527130127,
+ "step": 1039
+ },
+ {
+ "epoch": 14.447161572052401,
+ "grad_norm": 0.015702905133366585,
+ "learning_rate": 0.0006,
+ "loss": 4.64700984954834,
+ "step": 1040
+ },
+ {
+ "epoch": 14.46113537117904,
+ "grad_norm": 0.016674194484949112,
+ "learning_rate": 0.0006,
+ "loss": 4.694093227386475,
+ "step": 1041
+ },
+ {
+ "epoch": 14.475109170305677,
+ "grad_norm": 0.016869477927684784,
+ "learning_rate": 0.0006,
+ "loss": 4.619544982910156,
+ "step": 1042
+ },
+ {
+ "epoch": 14.489082969432314,
+ "grad_norm": 0.016557540744543076,
+ "learning_rate": 0.0006,
+ "loss": 4.568667411804199,
+ "step": 1043
+ },
+ {
+ "epoch": 14.503056768558952,
+ "grad_norm": 0.014500945806503296,
+ "learning_rate": 0.0006,
+ "loss": 4.561741828918457,
+ "step": 1044
+ },
+ {
+ "epoch": 14.51703056768559,
+ "grad_norm": 0.014228739775717258,
+ "learning_rate": 0.0006,
+ "loss": 4.654277324676514,
+ "step": 1045
+ },
+ {
+ "epoch": 14.531004366812226,
+ "grad_norm": 0.013894238509237766,
+ "learning_rate": 0.0006,
+ "loss": 4.5867719650268555,
+ "step": 1046
+ },
+ {
+ "epoch": 14.544978165938865,
+ "grad_norm": 0.01550068985670805,
+ "learning_rate": 0.0006,
+ "loss": 4.61168098449707,
+ "step": 1047
+ },
+ {
+ "epoch": 14.558951965065502,
+ "grad_norm": 0.015940772369503975,
+ "learning_rate": 0.0006,
+ "loss": 4.691247940063477,
+ "step": 1048
+ },
+ {
+ "epoch": 14.57292576419214,
+ "grad_norm": 0.01714535616338253,
+ "learning_rate": 0.0006,
+ "loss": 4.61183500289917,
+ "step": 1049
+ },
+ {
+ "epoch": 14.586899563318777,
+ "grad_norm": 0.017925038933753967,
+ "learning_rate": 0.0006,
+ "loss": 4.719022750854492,
+ "step": 1050
+ },
+ {
+ "epoch": 14.600873362445414,
+ "grad_norm": 0.018141552805900574,
+ "learning_rate": 0.0006,
+ "loss": 4.760281562805176,
+ "step": 1051
+ },
+ {
+ "epoch": 14.614847161572053,
+ "grad_norm": 0.017131328582763672,
+ "learning_rate": 0.0006,
+ "loss": 4.709680080413818,
+ "step": 1052
+ },
+ {
+ "epoch": 14.62882096069869,
+ "grad_norm": 0.018730048090219498,
+ "learning_rate": 0.0006,
+ "loss": 4.5994720458984375,
+ "step": 1053
+ },
+ {
+ "epoch": 14.642794759825328,
+ "grad_norm": 0.018894298002123833,
+ "learning_rate": 0.0006,
+ "loss": 4.669070720672607,
+ "step": 1054
+ },
+ {
+ "epoch": 14.656768558951965,
+ "grad_norm": 0.017643198370933533,
+ "learning_rate": 0.0006,
+ "loss": 4.641790390014648,
+ "step": 1055
+ },
+ {
+ "epoch": 14.670742358078602,
+ "grad_norm": 0.017647244036197662,
+ "learning_rate": 0.0006,
+ "loss": 4.653899192810059,
+ "step": 1056
+ },
+ {
+ "epoch": 14.68471615720524,
+ "grad_norm": 0.017304198816418648,
+ "learning_rate": 0.0006,
+ "loss": 4.7487473487854,
+ "step": 1057
+ },
+ {
+ "epoch": 14.698689956331878,
+ "grad_norm": 0.01638447679579258,
+ "learning_rate": 0.0006,
+ "loss": 4.71945858001709,
+ "step": 1058
+ },
+ {
+ "epoch": 14.712663755458514,
+ "grad_norm": 0.015892071649432182,
+ "learning_rate": 0.0006,
+ "loss": 4.744903564453125,
+ "step": 1059
+ },
+ {
+ "epoch": 14.726637554585153,
+ "grad_norm": 0.017420614138245583,
+ "learning_rate": 0.0006,
+ "loss": 4.781563758850098,
+ "step": 1060
+ },
+ {
+ "epoch": 14.74061135371179,
+ "grad_norm": 0.016498520970344543,
+ "learning_rate": 0.0006,
+ "loss": 4.622779846191406,
+ "step": 1061
+ },
+ {
+ "epoch": 14.754585152838429,
+ "grad_norm": 0.017482127994298935,
+ "learning_rate": 0.0006,
+ "loss": 4.62825870513916,
+ "step": 1062
+ },
+ {
+ "epoch": 14.768558951965066,
+ "grad_norm": 0.019018379971385002,
+ "learning_rate": 0.0006,
+ "loss": 4.693194389343262,
+ "step": 1063
+ },
+ {
+ "epoch": 14.782532751091702,
+ "grad_norm": 0.01993914321064949,
+ "learning_rate": 0.0006,
+ "loss": 4.742903709411621,
+ "step": 1064
+ },
+ {
+ "epoch": 14.796506550218341,
+ "grad_norm": 0.02301335707306862,
+ "learning_rate": 0.0006,
+ "loss": 4.613251686096191,
+ "step": 1065
+ },
+ {
+ "epoch": 14.810480349344978,
+ "grad_norm": 0.024088222533464432,
+ "learning_rate": 0.0006,
+ "loss": 4.594305992126465,
+ "step": 1066
+ },
+ {
+ "epoch": 14.824454148471617,
+ "grad_norm": 0.02493320405483246,
+ "learning_rate": 0.0006,
+ "loss": 4.700588226318359,
+ "step": 1067
+ },
+ {
+ "epoch": 14.838427947598253,
+ "grad_norm": 0.02501937560737133,
+ "learning_rate": 0.0006,
+ "loss": 4.601742744445801,
+ "step": 1068
+ },
+ {
+ "epoch": 14.85240174672489,
+ "grad_norm": 0.023490751162171364,
+ "learning_rate": 0.0006,
+ "loss": 4.589211463928223,
+ "step": 1069
+ },
+ {
+ "epoch": 14.866375545851529,
+ "grad_norm": 0.019695665687322617,
+ "learning_rate": 0.0006,
+ "loss": 4.635081768035889,
+ "step": 1070
+ },
+ {
+ "epoch": 14.880349344978166,
+ "grad_norm": 0.019276980310678482,
+ "learning_rate": 0.0006,
+ "loss": 4.737002372741699,
+ "step": 1071
+ },
+ {
+ "epoch": 14.894323144104803,
+ "grad_norm": 0.022430511191487312,
+ "learning_rate": 0.0006,
+ "loss": 4.808295726776123,
+ "step": 1072
+ },
+ {
+ "epoch": 14.908296943231441,
+ "grad_norm": 0.021543916314840317,
+ "learning_rate": 0.0006,
+ "loss": 4.559475421905518,
+ "step": 1073
+ },
+ {
+ "epoch": 14.922270742358078,
+ "grad_norm": 0.019885241985321045,
+ "learning_rate": 0.0006,
+ "loss": 4.647948741912842,
+ "step": 1074
+ },
+ {
+ "epoch": 14.936244541484717,
+ "grad_norm": 0.02115175500512123,
+ "learning_rate": 0.0006,
+ "loss": 4.680505752563477,
+ "step": 1075
+ },
+ {
+ "epoch": 14.950218340611354,
+ "grad_norm": 0.02156151458621025,
+ "learning_rate": 0.0006,
+ "loss": 4.673477649688721,
+ "step": 1076
+ },
+ {
+ "epoch": 14.96419213973799,
+ "grad_norm": 0.02184601128101349,
+ "learning_rate": 0.0006,
+ "loss": 4.644402027130127,
+ "step": 1077
+ },
+ {
+ "epoch": 14.97816593886463,
+ "grad_norm": 0.0211983360350132,
+ "learning_rate": 0.0006,
+ "loss": 4.821558475494385,
+ "step": 1078
+ },
+ {
+ "epoch": 14.992139737991266,
+ "grad_norm": 0.019636567682027817,
+ "learning_rate": 0.0006,
+ "loss": 4.578868865966797,
+ "step": 1079
+ },
+ {
+ "epoch": 15.0,
+ "grad_norm": 0.019997350871562958,
+ "learning_rate": 0.0006,
+ "loss": 4.691307544708252,
+ "step": 1080
+ },
+ {
+ "epoch": 15.0,
+ "eval_loss": 4.872478008270264,
+ "eval_runtime": 56.8513,
+ "eval_samples_per_second": 42.954,
+ "eval_steps_per_second": 1.354,
+ "step": 1080
+ },
+ {
+ "epoch": 15.013973799126637,
+ "grad_norm": 0.018830662593245506,
+ "learning_rate": 0.0006,
+ "loss": 4.761779308319092,
+ "step": 1081
+ },
+ {
+ "epoch": 15.027947598253276,
+ "grad_norm": 0.016876481473445892,
+ "learning_rate": 0.0006,
+ "loss": 4.61148738861084,
+ "step": 1082
+ },
+ {
+ "epoch": 15.041921397379912,
+ "grad_norm": 0.015550863929092884,
+ "learning_rate": 0.0006,
+ "loss": 4.676124095916748,
+ "step": 1083
+ },
+ {
+ "epoch": 15.055895196506551,
+ "grad_norm": 0.017178896814584732,
+ "learning_rate": 0.0006,
+ "loss": 4.588935852050781,
+ "step": 1084
+ },
+ {
+ "epoch": 15.069868995633188,
+ "grad_norm": 0.015465447679162025,
+ "learning_rate": 0.0006,
+ "loss": 4.605499267578125,
+ "step": 1085
+ },
+ {
+ "epoch": 15.083842794759825,
+ "grad_norm": 0.01606195978820324,
+ "learning_rate": 0.0006,
+ "loss": 4.723373889923096,
+ "step": 1086
+ },
+ {
+ "epoch": 15.097816593886463,
+ "grad_norm": 0.016521936282515526,
+ "learning_rate": 0.0006,
+ "loss": 4.646608352661133,
+ "step": 1087
+ },
+ {
+ "epoch": 15.1117903930131,
+ "grad_norm": 0.01589205674827099,
+ "learning_rate": 0.0006,
+ "loss": 4.498138427734375,
+ "step": 1088
+ },
+ {
+ "epoch": 15.125764192139737,
+ "grad_norm": 0.01524856686592102,
+ "learning_rate": 0.0006,
+ "loss": 4.594562530517578,
+ "step": 1089
+ },
+ {
+ "epoch": 15.139737991266376,
+ "grad_norm": 0.014714641496539116,
+ "learning_rate": 0.0006,
+ "loss": 4.630744934082031,
+ "step": 1090
+ },
+ {
+ "epoch": 15.153711790393013,
+ "grad_norm": 0.015333331190049648,
+ "learning_rate": 0.0006,
+ "loss": 4.682388782501221,
+ "step": 1091
+ },
+ {
+ "epoch": 15.167685589519651,
+ "grad_norm": 0.01550073828548193,
+ "learning_rate": 0.0006,
+ "loss": 4.692652702331543,
+ "step": 1092
+ },
+ {
+ "epoch": 15.181659388646288,
+ "grad_norm": 0.014783318154513836,
+ "learning_rate": 0.0006,
+ "loss": 4.6223835945129395,
+ "step": 1093
+ },
+ {
+ "epoch": 15.195633187772925,
+ "grad_norm": 0.014372751116752625,
+ "learning_rate": 0.0006,
+ "loss": 4.52687406539917,
+ "step": 1094
+ },
+ {
+ "epoch": 15.209606986899564,
+ "grad_norm": 0.014779130928218365,
+ "learning_rate": 0.0006,
+ "loss": 4.637009620666504,
+ "step": 1095
+ },
+ {
+ "epoch": 15.2235807860262,
+ "grad_norm": 0.015697073191404343,
+ "learning_rate": 0.0006,
+ "loss": 4.682981014251709,
+ "step": 1096
+ },
+ {
+ "epoch": 15.237554585152838,
+ "grad_norm": 0.016523664817214012,
+ "learning_rate": 0.0006,
+ "loss": 4.705722808837891,
+ "step": 1097
+ },
+ {
+ "epoch": 15.251528384279476,
+ "grad_norm": 0.017633996903896332,
+ "learning_rate": 0.0006,
+ "loss": 4.574686050415039,
+ "step": 1098
+ },
+ {
+ "epoch": 15.265502183406113,
+ "grad_norm": 0.019955076277256012,
+ "learning_rate": 0.0006,
+ "loss": 4.634450912475586,
+ "step": 1099
+ },
+ {
+ "epoch": 15.279475982532752,
+ "grad_norm": 0.02132618986070156,
+ "learning_rate": 0.0006,
+ "loss": 4.56110143661499,
+ "step": 1100
+ },
+ {
+ "epoch": 15.293449781659389,
+ "grad_norm": 0.0207593385130167,
+ "learning_rate": 0.0006,
+ "loss": 4.629396438598633,
+ "step": 1101
+ },
+ {
+ "epoch": 15.307423580786025,
+ "grad_norm": 0.017882540822029114,
+ "learning_rate": 0.0006,
+ "loss": 4.730963230133057,
+ "step": 1102
+ },
+ {
+ "epoch": 15.321397379912664,
+ "grad_norm": 0.01917876861989498,
+ "learning_rate": 0.0006,
+ "loss": 4.64585542678833,
+ "step": 1103
+ },
+ {
+ "epoch": 15.335371179039301,
+ "grad_norm": 0.01943974196910858,
+ "learning_rate": 0.0006,
+ "loss": 4.656181335449219,
+ "step": 1104
+ },
+ {
+ "epoch": 15.34934497816594,
+ "grad_norm": 0.019904715940356255,
+ "learning_rate": 0.0006,
+ "loss": 4.619329452514648,
+ "step": 1105
+ },
+ {
+ "epoch": 15.363318777292577,
+ "grad_norm": 0.019174346700310707,
+ "learning_rate": 0.0006,
+ "loss": 4.647892475128174,
+ "step": 1106
+ },
+ {
+ "epoch": 15.377292576419213,
+ "grad_norm": 0.018244758248329163,
+ "learning_rate": 0.0006,
+ "loss": 4.56166934967041,
+ "step": 1107
+ },
+ {
+ "epoch": 15.391266375545852,
+ "grad_norm": 0.018914205953478813,
+ "learning_rate": 0.0006,
+ "loss": 4.648063659667969,
+ "step": 1108
+ },
+ {
+ "epoch": 15.405240174672489,
+ "grad_norm": 0.01961071416735649,
+ "learning_rate": 0.0006,
+ "loss": 4.665525436401367,
+ "step": 1109
+ },
+ {
+ "epoch": 15.419213973799126,
+ "grad_norm": 0.01841311901807785,
+ "learning_rate": 0.0006,
+ "loss": 4.611342430114746,
+ "step": 1110
+ },
+ {
+ "epoch": 15.433187772925764,
+ "grad_norm": 0.01863427273929119,
+ "learning_rate": 0.0006,
+ "loss": 4.664256572723389,
+ "step": 1111
+ },
+ {
+ "epoch": 15.447161572052401,
+ "grad_norm": 0.020506877452135086,
+ "learning_rate": 0.0006,
+ "loss": 4.631237506866455,
+ "step": 1112
+ },
+ {
+ "epoch": 15.46113537117904,
+ "grad_norm": 0.02223833277821541,
+ "learning_rate": 0.0006,
+ "loss": 4.668654441833496,
+ "step": 1113
+ },
+ {
+ "epoch": 15.475109170305677,
+ "grad_norm": 0.023336702957749367,
+ "learning_rate": 0.0006,
+ "loss": 4.562833309173584,
+ "step": 1114
+ },
+ {
+ "epoch": 15.489082969432314,
+ "grad_norm": 0.023052413016557693,
+ "learning_rate": 0.0006,
+ "loss": 4.621532917022705,
+ "step": 1115
+ },
+ {
+ "epoch": 15.503056768558952,
+ "grad_norm": 0.023420924320816994,
+ "learning_rate": 0.0006,
+ "loss": 4.665465354919434,
+ "step": 1116
+ },
+ {
+ "epoch": 15.51703056768559,
+ "grad_norm": 0.024071281775832176,
+ "learning_rate": 0.0006,
+ "loss": 4.595980644226074,
+ "step": 1117
+ },
+ {
+ "epoch": 15.531004366812226,
+ "grad_norm": 0.023406682536005974,
+ "learning_rate": 0.0006,
+ "loss": 4.642635345458984,
+ "step": 1118
+ },
+ {
+ "epoch": 15.544978165938865,
+ "grad_norm": 0.02138647809624672,
+ "learning_rate": 0.0006,
+ "loss": 4.704513072967529,
+ "step": 1119
+ },
+ {
+ "epoch": 15.558951965065502,
+ "grad_norm": 0.0221384447067976,
+ "learning_rate": 0.0006,
+ "loss": 4.567503452301025,
+ "step": 1120
+ },
+ {
+ "epoch": 15.57292576419214,
+ "grad_norm": 0.020879851654171944,
+ "learning_rate": 0.0006,
+ "loss": 4.664057731628418,
+ "step": 1121
+ },
+ {
+ "epoch": 15.586899563318777,
+ "grad_norm": 0.021478967741131783,
+ "learning_rate": 0.0006,
+ "loss": 4.539627552032471,
+ "step": 1122
+ },
+ {
+ "epoch": 15.600873362445414,
+ "grad_norm": 0.022748656570911407,
+ "learning_rate": 0.0006,
+ "loss": 4.5654497146606445,
+ "step": 1123
+ },
+ {
+ "epoch": 15.614847161572053,
+ "grad_norm": 0.021870248019695282,
+ "learning_rate": 0.0006,
+ "loss": 4.56899356842041,
+ "step": 1124
+ },
+ {
+ "epoch": 15.62882096069869,
+ "grad_norm": 0.021711355075240135,
+ "learning_rate": 0.0006,
+ "loss": 4.657808303833008,
+ "step": 1125
+ },
+ {
+ "epoch": 15.642794759825328,
+ "grad_norm": 0.022810909897089005,
+ "learning_rate": 0.0006,
+ "loss": 4.6526689529418945,
+ "step": 1126
+ },
+ {
+ "epoch": 15.656768558951965,
+ "grad_norm": 0.022561343386769295,
+ "learning_rate": 0.0006,
+ "loss": 4.534887790679932,
+ "step": 1127
+ },
+ {
+ "epoch": 15.670742358078602,
+ "grad_norm": 0.019909802824258804,
+ "learning_rate": 0.0006,
+ "loss": 4.600790023803711,
+ "step": 1128
+ },
+ {
+ "epoch": 15.68471615720524,
+ "grad_norm": 0.020092325285077095,
+ "learning_rate": 0.0006,
+ "loss": 4.773830413818359,
+ "step": 1129
+ },
+ {
+ "epoch": 15.698689956331878,
+ "grad_norm": 0.018834445625543594,
+ "learning_rate": 0.0006,
+ "loss": 4.628868103027344,
+ "step": 1130
+ },
+ {
+ "epoch": 15.712663755458514,
+ "grad_norm": 0.017552798613905907,
+ "learning_rate": 0.0006,
+ "loss": 4.647500038146973,
+ "step": 1131
+ },
+ {
+ "epoch": 15.726637554585153,
+ "grad_norm": 0.01751251146197319,
+ "learning_rate": 0.0006,
+ "loss": 4.590545654296875,
+ "step": 1132
+ },
+ {
+ "epoch": 15.74061135371179,
+ "grad_norm": 0.01699099875986576,
+ "learning_rate": 0.0006,
+ "loss": 4.509699821472168,
+ "step": 1133
+ },
+ {
+ "epoch": 15.754585152838429,
+ "grad_norm": 0.015808649361133575,
+ "learning_rate": 0.0006,
+ "loss": 4.606992244720459,
+ "step": 1134
+ },
+ {
+ "epoch": 15.768558951965066,
+ "grad_norm": 0.01655443385243416,
+ "learning_rate": 0.0006,
+ "loss": 4.585219860076904,
+ "step": 1135
+ },
+ {
+ "epoch": 15.782532751091702,
+ "grad_norm": 0.01708163693547249,
+ "learning_rate": 0.0006,
+ "loss": 4.668147087097168,
+ "step": 1136
+ },
+ {
+ "epoch": 15.796506550218341,
+ "grad_norm": 0.017161913216114044,
+ "learning_rate": 0.0006,
+ "loss": 4.7170634269714355,
+ "step": 1137
+ },
+ {
+ "epoch": 15.810480349344978,
+ "grad_norm": 0.017262674868106842,
+ "learning_rate": 0.0006,
+ "loss": 4.572132110595703,
+ "step": 1138
+ },
+ {
+ "epoch": 15.824454148471617,
+ "grad_norm": 0.017713576555252075,
+ "learning_rate": 0.0006,
+ "loss": 4.702538967132568,
+ "step": 1139
+ },
+ {
+ "epoch": 15.838427947598253,
+ "grad_norm": 0.018730733543634415,
+ "learning_rate": 0.0006,
+ "loss": 4.629721641540527,
+ "step": 1140
+ },
+ {
+ "epoch": 15.85240174672489,
+ "grad_norm": 0.017960187047719955,
+ "learning_rate": 0.0006,
+ "loss": 4.661039352416992,
+ "step": 1141
+ },
+ {
+ "epoch": 15.866375545851529,
+ "grad_norm": 0.01750079356133938,
+ "learning_rate": 0.0006,
+ "loss": 4.503710746765137,
+ "step": 1142
+ },
+ {
+ "epoch": 15.880349344978166,
+ "grad_norm": 0.016630422323942184,
+ "learning_rate": 0.0006,
+ "loss": 4.6582136154174805,
+ "step": 1143
+ },
+ {
+ "epoch": 15.894323144104803,
+ "grad_norm": 0.015896832570433617,
+ "learning_rate": 0.0006,
+ "loss": 4.596693992614746,
+ "step": 1144
+ },
+ {
+ "epoch": 15.908296943231441,
+ "grad_norm": 0.01566407084465027,
+ "learning_rate": 0.0006,
+ "loss": 4.6337199211120605,
+ "step": 1145
+ },
+ {
+ "epoch": 15.922270742358078,
+ "grad_norm": 0.016823330894112587,
+ "learning_rate": 0.0006,
+ "loss": 4.686994552612305,
+ "step": 1146
+ },
+ {
+ "epoch": 15.936244541484717,
+ "grad_norm": 0.016208255663514137,
+ "learning_rate": 0.0006,
+ "loss": 4.6205549240112305,
+ "step": 1147
+ },
+ {
+ "epoch": 15.950218340611354,
+ "grad_norm": 0.0171707421541214,
+ "learning_rate": 0.0006,
+ "loss": 4.582968711853027,
+ "step": 1148
+ },
+ {
+ "epoch": 15.96419213973799,
+ "grad_norm": 0.017009446397423744,
+ "learning_rate": 0.0006,
+ "loss": 4.510274887084961,
+ "step": 1149
+ },
+ {
+ "epoch": 15.97816593886463,
+ "grad_norm": 0.01565811224281788,
+ "learning_rate": 0.0006,
+ "loss": 4.55834436416626,
+ "step": 1150
+ },
+ {
+ "epoch": 15.992139737991266,
+ "grad_norm": 0.015276037156581879,
+ "learning_rate": 0.0006,
+ "loss": 4.537870407104492,
+ "step": 1151
+ },
+ {
+ "epoch": 16.0,
+ "grad_norm": 0.017305882647633553,
+ "learning_rate": 0.0006,
+ "loss": 4.611305236816406,
+ "step": 1152
+ },
+ {
+ "epoch": 16.0,
+ "eval_loss": 4.775513172149658,
+ "eval_runtime": 56.7341,
+ "eval_samples_per_second": 43.043,
+ "eval_steps_per_second": 1.357,
+ "step": 1152
+ },
+ {
+ "epoch": 16.01397379912664,
+ "grad_norm": 0.016752436757087708,
+ "learning_rate": 0.0006,
+ "loss": 4.541312217712402,
+ "step": 1153
+ },
+ {
+ "epoch": 16.027947598253274,
+ "grad_norm": 0.01643134094774723,
+ "learning_rate": 0.0006,
+ "loss": 4.585402965545654,
+ "step": 1154
+ },
+ {
+ "epoch": 16.041921397379912,
+ "grad_norm": 0.01888720504939556,
+ "learning_rate": 0.0006,
+ "loss": 4.597825050354004,
+ "step": 1155
+ },
+ {
+ "epoch": 16.05589519650655,
+ "grad_norm": 0.022044632583856583,
+ "learning_rate": 0.0006,
+ "loss": 4.532772064208984,
+ "step": 1156
+ },
+ {
+ "epoch": 16.069868995633186,
+ "grad_norm": 0.022986987605690956,
+ "learning_rate": 0.0006,
+ "loss": 4.551351547241211,
+ "step": 1157
+ },
+ {
+ "epoch": 16.083842794759825,
+ "grad_norm": 0.020116791129112244,
+ "learning_rate": 0.0006,
+ "loss": 4.484149932861328,
+ "step": 1158
+ },
+ {
+ "epoch": 16.097816593886463,
+ "grad_norm": 0.021442247554659843,
+ "learning_rate": 0.0006,
+ "loss": 4.515316486358643,
+ "step": 1159
+ },
+ {
+ "epoch": 16.111790393013102,
+ "grad_norm": 0.020974772050976753,
+ "learning_rate": 0.0006,
+ "loss": 4.51606559753418,
+ "step": 1160
+ },
+ {
+ "epoch": 16.125764192139737,
+ "grad_norm": 0.021274864673614502,
+ "learning_rate": 0.0006,
+ "loss": 4.587329864501953,
+ "step": 1161
+ },
+ {
+ "epoch": 16.139737991266376,
+ "grad_norm": 0.022799784317612648,
+ "learning_rate": 0.0006,
+ "loss": 4.568149089813232,
+ "step": 1162
+ },
+ {
+ "epoch": 16.153711790393015,
+ "grad_norm": 0.021784933283925056,
+ "learning_rate": 0.0006,
+ "loss": 4.515336990356445,
+ "step": 1163
+ },
+ {
+ "epoch": 16.16768558951965,
+ "grad_norm": 0.024261271581053734,
+ "learning_rate": 0.0006,
+ "loss": 4.5127081871032715,
+ "step": 1164
+ },
+ {
+ "epoch": 16.18165938864629,
+ "grad_norm": 0.02248375490307808,
+ "learning_rate": 0.0006,
+ "loss": 4.754251003265381,
+ "step": 1165
+ },
+ {
+ "epoch": 16.195633187772927,
+ "grad_norm": 0.02136492356657982,
+ "learning_rate": 0.0006,
+ "loss": 4.58193826675415,
+ "step": 1166
+ },
+ {
+ "epoch": 16.209606986899562,
+ "grad_norm": 0.020727066323161125,
+ "learning_rate": 0.0006,
+ "loss": 4.564209938049316,
+ "step": 1167
+ },
+ {
+ "epoch": 16.2235807860262,
+ "grad_norm": 0.017789119854569435,
+ "learning_rate": 0.0006,
+ "loss": 4.698680877685547,
+ "step": 1168
+ },
+ {
+ "epoch": 16.23755458515284,
+ "grad_norm": 0.017603034153580666,
+ "learning_rate": 0.0006,
+ "loss": 4.501171588897705,
+ "step": 1169
+ },
+ {
+ "epoch": 16.251528384279474,
+ "grad_norm": 0.017345862463116646,
+ "learning_rate": 0.0006,
+ "loss": 4.593590259552002,
+ "step": 1170
+ },
+ {
+ "epoch": 16.265502183406113,
+ "grad_norm": 0.018047425895929337,
+ "learning_rate": 0.0006,
+ "loss": 4.597467422485352,
+ "step": 1171
+ },
+ {
+ "epoch": 16.27947598253275,
+ "grad_norm": 0.01947222836315632,
+ "learning_rate": 0.0006,
+ "loss": 4.5072221755981445,
+ "step": 1172
+ },
+ {
+ "epoch": 16.29344978165939,
+ "grad_norm": 0.019598500803112984,
+ "learning_rate": 0.0006,
+ "loss": 4.601351737976074,
+ "step": 1173
+ },
+ {
+ "epoch": 16.307423580786025,
+ "grad_norm": 0.020002514123916626,
+ "learning_rate": 0.0006,
+ "loss": 4.580431938171387,
+ "step": 1174
+ },
+ {
+ "epoch": 16.321397379912664,
+ "grad_norm": 0.0216530729085207,
+ "learning_rate": 0.0006,
+ "loss": 4.424681663513184,
+ "step": 1175
+ },
+ {
+ "epoch": 16.335371179039303,
+ "grad_norm": 0.01949235238134861,
+ "learning_rate": 0.0006,
+ "loss": 4.751883029937744,
+ "step": 1176
+ },
+ {
+ "epoch": 16.349344978165938,
+ "grad_norm": 0.017779016867280006,
+ "learning_rate": 0.0006,
+ "loss": 4.541490077972412,
+ "step": 1177
+ },
+ {
+ "epoch": 16.363318777292577,
+ "grad_norm": 0.01909121684730053,
+ "learning_rate": 0.0006,
+ "loss": 4.594141006469727,
+ "step": 1178
+ },
+ {
+ "epoch": 16.377292576419215,
+ "grad_norm": 0.019894849509000778,
+ "learning_rate": 0.0006,
+ "loss": 4.555070400238037,
+ "step": 1179
+ },
+ {
+ "epoch": 16.39126637554585,
+ "grad_norm": 0.01993032731115818,
+ "learning_rate": 0.0006,
+ "loss": 4.521401405334473,
+ "step": 1180
+ },
+ {
+ "epoch": 16.40524017467249,
+ "grad_norm": 0.0210788045078516,
+ "learning_rate": 0.0006,
+ "loss": 4.672842979431152,
+ "step": 1181
+ },
+ {
+ "epoch": 16.419213973799128,
+ "grad_norm": 0.02274598926305771,
+ "learning_rate": 0.0006,
+ "loss": 4.629504203796387,
+ "step": 1182
+ },
+ {
+ "epoch": 16.433187772925763,
+ "grad_norm": 0.020448120310902596,
+ "learning_rate": 0.0006,
+ "loss": 4.548209190368652,
+ "step": 1183
+ },
+ {
+ "epoch": 16.4471615720524,
+ "grad_norm": 0.017565643414855003,
+ "learning_rate": 0.0006,
+ "loss": 4.529054641723633,
+ "step": 1184
+ },
+ {
+ "epoch": 16.46113537117904,
+ "grad_norm": 0.019504237920045853,
+ "learning_rate": 0.0006,
+ "loss": 4.48182487487793,
+ "step": 1185
+ },
+ {
+ "epoch": 16.475109170305675,
+ "grad_norm": 0.016823742538690567,
+ "learning_rate": 0.0006,
+ "loss": 4.608696937561035,
+ "step": 1186
+ },
+ {
+ "epoch": 16.489082969432314,
+ "grad_norm": 0.01602894812822342,
+ "learning_rate": 0.0006,
+ "loss": 4.620062351226807,
+ "step": 1187
+ },
+ {
+ "epoch": 16.503056768558952,
+ "grad_norm": 0.016615642234683037,
+ "learning_rate": 0.0006,
+ "loss": 4.69378662109375,
+ "step": 1188
+ },
+ {
+ "epoch": 16.51703056768559,
+ "grad_norm": 0.018576730042696,
+ "learning_rate": 0.0006,
+ "loss": 4.537041664123535,
+ "step": 1189
+ },
+ {
+ "epoch": 16.531004366812226,
+ "grad_norm": 0.018611151725053787,
+ "learning_rate": 0.0006,
+ "loss": 4.496169090270996,
+ "step": 1190
+ },
+ {
+ "epoch": 16.544978165938865,
+ "grad_norm": 0.018622413277626038,
+ "learning_rate": 0.0006,
+ "loss": 4.557358264923096,
+ "step": 1191
+ },
+ {
+ "epoch": 16.558951965065503,
+ "grad_norm": 0.01823544315993786,
+ "learning_rate": 0.0006,
+ "loss": 4.730792999267578,
+ "step": 1192
+ },
+ {
+ "epoch": 16.57292576419214,
+ "grad_norm": 0.018796846270561218,
+ "learning_rate": 0.0006,
+ "loss": 4.781791687011719,
+ "step": 1193
+ },
+ {
+ "epoch": 16.586899563318777,
+ "grad_norm": 0.01728292554616928,
+ "learning_rate": 0.0006,
+ "loss": 4.515257835388184,
+ "step": 1194
+ },
+ {
+ "epoch": 16.600873362445416,
+ "grad_norm": 0.01699589006602764,
+ "learning_rate": 0.0006,
+ "loss": 4.5996904373168945,
+ "step": 1195
+ },
+ {
+ "epoch": 16.61484716157205,
+ "grad_norm": 0.015905970707535744,
+ "learning_rate": 0.0006,
+ "loss": 4.601301193237305,
+ "step": 1196
+ },
+ {
+ "epoch": 16.62882096069869,
+ "grad_norm": 0.01338571310043335,
+ "learning_rate": 0.0006,
+ "loss": 4.557708740234375,
+ "step": 1197
+ },
+ {
+ "epoch": 16.64279475982533,
+ "grad_norm": 0.015298855490982533,
+ "learning_rate": 0.0006,
+ "loss": 4.542287349700928,
+ "step": 1198
+ },
+ {
+ "epoch": 16.656768558951963,
+ "grad_norm": 0.014753367751836777,
+ "learning_rate": 0.0006,
+ "loss": 4.595787048339844,
+ "step": 1199
+ },
+ {
+ "epoch": 16.670742358078602,
+ "grad_norm": 0.015527483075857162,
+ "learning_rate": 0.0006,
+ "loss": 4.607529163360596,
+ "step": 1200
+ },
+ {
+ "epoch": 16.68471615720524,
+ "grad_norm": 0.015042162500321865,
+ "learning_rate": 0.0006,
+ "loss": 4.686626434326172,
+ "step": 1201
+ },
+ {
+ "epoch": 16.69868995633188,
+ "grad_norm": 0.0159393772482872,
+ "learning_rate": 0.0006,
+ "loss": 4.47656774520874,
+ "step": 1202
+ },
+ {
+ "epoch": 16.712663755458514,
+ "grad_norm": 0.01532787922769785,
+ "learning_rate": 0.0006,
+ "loss": 4.55678653717041,
+ "step": 1203
+ },
+ {
+ "epoch": 16.726637554585153,
+ "grad_norm": 0.016332164406776428,
+ "learning_rate": 0.0006,
+ "loss": 4.51441764831543,
+ "step": 1204
+ },
+ {
+ "epoch": 16.74061135371179,
+ "grad_norm": 0.018901726230978966,
+ "learning_rate": 0.0006,
+ "loss": 4.57227897644043,
+ "step": 1205
+ },
+ {
+ "epoch": 16.754585152838427,
+ "grad_norm": 0.018589433282613754,
+ "learning_rate": 0.0006,
+ "loss": 4.550070762634277,
+ "step": 1206
+ },
+ {
+ "epoch": 16.768558951965066,
+ "grad_norm": 0.01634034886956215,
+ "learning_rate": 0.0006,
+ "loss": 4.586910724639893,
+ "step": 1207
+ },
+ {
+ "epoch": 16.782532751091704,
+ "grad_norm": 0.017599433660507202,
+ "learning_rate": 0.0006,
+ "loss": 4.5891923904418945,
+ "step": 1208
+ },
+ {
+ "epoch": 16.79650655021834,
+ "grad_norm": 0.01878446340560913,
+ "learning_rate": 0.0006,
+ "loss": 4.57227087020874,
+ "step": 1209
+ },
+ {
+ "epoch": 16.810480349344978,
+ "grad_norm": 0.01803469844162464,
+ "learning_rate": 0.0006,
+ "loss": 4.583845138549805,
+ "step": 1210
+ },
+ {
+ "epoch": 16.824454148471617,
+ "grad_norm": 0.018505144864320755,
+ "learning_rate": 0.0006,
+ "loss": 4.561151504516602,
+ "step": 1211
+ },
+ {
+ "epoch": 16.83842794759825,
+ "grad_norm": 0.01920081488788128,
+ "learning_rate": 0.0006,
+ "loss": 4.506834030151367,
+ "step": 1212
+ },
+ {
+ "epoch": 16.85240174672489,
+ "grad_norm": 0.0177712831646204,
+ "learning_rate": 0.0006,
+ "loss": 4.557195663452148,
+ "step": 1213
+ },
+ {
+ "epoch": 16.86637554585153,
+ "grad_norm": 0.019113918766379356,
+ "learning_rate": 0.0006,
+ "loss": 4.651227951049805,
+ "step": 1214
+ },
+ {
+ "epoch": 16.880349344978168,
+ "grad_norm": 0.02012539468705654,
+ "learning_rate": 0.0006,
+ "loss": 4.4952239990234375,
+ "step": 1215
+ },
+ {
+ "epoch": 16.894323144104803,
+ "grad_norm": 0.019993646070361137,
+ "learning_rate": 0.0006,
+ "loss": 4.517433166503906,
+ "step": 1216
+ },
+ {
+ "epoch": 16.90829694323144,
+ "grad_norm": 0.020041441544890404,
+ "learning_rate": 0.0006,
+ "loss": 4.489407539367676,
+ "step": 1217
+ },
+ {
+ "epoch": 16.92227074235808,
+ "grad_norm": 0.0183611661195755,
+ "learning_rate": 0.0006,
+ "loss": 4.5435638427734375,
+ "step": 1218
+ },
+ {
+ "epoch": 16.936244541484715,
+ "grad_norm": 0.019138429313898087,
+ "learning_rate": 0.0006,
+ "loss": 4.545949459075928,
+ "step": 1219
+ },
+ {
+ "epoch": 16.950218340611354,
+ "grad_norm": 0.018554266542196274,
+ "learning_rate": 0.0006,
+ "loss": 4.579620361328125,
+ "step": 1220
+ },
+ {
+ "epoch": 16.964192139737992,
+ "grad_norm": 0.0193545650690794,
+ "learning_rate": 0.0006,
+ "loss": 4.588255405426025,
+ "step": 1221
+ },
+ {
+ "epoch": 16.978165938864628,
+ "grad_norm": 0.019389165565371513,
+ "learning_rate": 0.0006,
+ "loss": 4.520698547363281,
+ "step": 1222
+ },
+ {
+ "epoch": 16.992139737991266,
+ "grad_norm": 0.016601379960775375,
+ "learning_rate": 0.0006,
+ "loss": 4.4809064865112305,
+ "step": 1223
+ },
+ {
+ "epoch": 17.0,
+ "grad_norm": 0.01771303080022335,
+ "learning_rate": 0.0006,
+ "loss": 4.721122741699219,
+ "step": 1224
+ },
+ {
+ "epoch": 17.0,
+ "eval_loss": 4.7903666496276855,
+ "eval_runtime": 56.9826,
+ "eval_samples_per_second": 42.855,
+ "eval_steps_per_second": 1.351,
+ "step": 1224
+ },
+ {
+ "epoch": 17.01397379912664,
+ "grad_norm": 0.017720289528369904,
+ "learning_rate": 0.0006,
+ "loss": 4.523834705352783,
+ "step": 1225
+ },
+ {
+ "epoch": 17.027947598253274,
+ "grad_norm": 0.020137161016464233,
+ "learning_rate": 0.0006,
+ "loss": 4.574061870574951,
+ "step": 1226
+ },
+ {
+ "epoch": 17.041921397379912,
+ "grad_norm": 0.02002314105629921,
+ "learning_rate": 0.0006,
+ "loss": 4.512407302856445,
+ "step": 1227
+ },
+ {
+ "epoch": 17.05589519650655,
+ "grad_norm": 0.019628183916211128,
+ "learning_rate": 0.0006,
+ "loss": 4.502533435821533,
+ "step": 1228
+ },
+ {
+ "epoch": 17.069868995633186,
+ "grad_norm": 0.019808458164334297,
+ "learning_rate": 0.0006,
+ "loss": 4.463156700134277,
+ "step": 1229
+ },
+ {
+ "epoch": 17.083842794759825,
+ "grad_norm": 0.021250786259770393,
+ "learning_rate": 0.0006,
+ "loss": 4.622682094573975,
+ "step": 1230
+ },
+ {
+ "epoch": 17.097816593886463,
+ "grad_norm": 0.022031694650650024,
+ "learning_rate": 0.0006,
+ "loss": 4.451882839202881,
+ "step": 1231
+ },
+ {
+ "epoch": 17.111790393013102,
+ "grad_norm": 0.01898987777531147,
+ "learning_rate": 0.0006,
+ "loss": 4.529331684112549,
+ "step": 1232
+ },
+ {
+ "epoch": 17.125764192139737,
+ "grad_norm": 0.018749834969639778,
+ "learning_rate": 0.0006,
+ "loss": 4.435815334320068,
+ "step": 1233
+ },
+ {
+ "epoch": 17.139737991266376,
+ "grad_norm": 0.01896674931049347,
+ "learning_rate": 0.0006,
+ "loss": 4.484029293060303,
+ "step": 1234
+ },
+ {
+ "epoch": 17.153711790393015,
+ "grad_norm": 0.018286997452378273,
+ "learning_rate": 0.0006,
+ "loss": 4.523510932922363,
+ "step": 1235
+ },
+ {
+ "epoch": 17.16768558951965,
+ "grad_norm": 0.017273059114813805,
+ "learning_rate": 0.0006,
+ "loss": 4.451767444610596,
+ "step": 1236
+ },
+ {
+ "epoch": 17.18165938864629,
+ "grad_norm": 0.016499150544404984,
+ "learning_rate": 0.0006,
+ "loss": 4.5568647384643555,
+ "step": 1237
+ },
+ {
+ "epoch": 17.195633187772927,
+ "grad_norm": 0.015934469178318977,
+ "learning_rate": 0.0006,
+ "loss": 4.4639997482299805,
+ "step": 1238
+ },
+ {
+ "epoch": 17.209606986899562,
+ "grad_norm": 0.01714209094643593,
+ "learning_rate": 0.0006,
+ "loss": 4.547614097595215,
+ "step": 1239
+ },
+ {
+ "epoch": 17.2235807860262,
+ "grad_norm": 0.019167495891451836,
+ "learning_rate": 0.0006,
+ "loss": 4.571097373962402,
+ "step": 1240
+ },
+ {
+ "epoch": 17.23755458515284,
+ "grad_norm": 0.018276996910572052,
+ "learning_rate": 0.0006,
+ "loss": 4.489764213562012,
+ "step": 1241
+ },
+ {
+ "epoch": 17.251528384279474,
+ "grad_norm": 0.021296866238117218,
+ "learning_rate": 0.0006,
+ "loss": 4.526913642883301,
+ "step": 1242
+ },
+ {
+ "epoch": 17.265502183406113,
+ "grad_norm": 0.022606629878282547,
+ "learning_rate": 0.0006,
+ "loss": 4.521624565124512,
+ "step": 1243
+ },
+ {
+ "epoch": 17.27947598253275,
+ "grad_norm": 0.021204711869359016,
+ "learning_rate": 0.0006,
+ "loss": 4.500864028930664,
+ "step": 1244
+ },
+ {
+ "epoch": 17.29344978165939,
+ "grad_norm": 0.01988849975168705,
+ "learning_rate": 0.0006,
+ "loss": 4.45162296295166,
+ "step": 1245
+ },
+ {
+ "epoch": 17.307423580786025,
+ "grad_norm": 0.020268384367227554,
+ "learning_rate": 0.0006,
+ "loss": 4.536198616027832,
+ "step": 1246
+ },
+ {
+ "epoch": 17.321397379912664,
+ "grad_norm": 0.020185677334666252,
+ "learning_rate": 0.0006,
+ "loss": 4.440262794494629,
+ "step": 1247
+ },
+ {
+ "epoch": 17.335371179039303,
+ "grad_norm": 0.020106853917241096,
+ "learning_rate": 0.0006,
+ "loss": 4.475907325744629,
+ "step": 1248
+ },
+ {
+ "epoch": 17.349344978165938,
+ "grad_norm": 0.018883297219872475,
+ "learning_rate": 0.0006,
+ "loss": 4.640661716461182,
+ "step": 1249
+ },
+ {
+ "epoch": 17.363318777292577,
+ "grad_norm": 0.01852620206773281,
+ "learning_rate": 0.0006,
+ "loss": 4.578243255615234,
+ "step": 1250
+ },
+ {
+ "epoch": 17.377292576419215,
+ "grad_norm": 0.01886451430618763,
+ "learning_rate": 0.0006,
+ "loss": 4.546026229858398,
+ "step": 1251
+ },
+ {
+ "epoch": 17.39126637554585,
+ "grad_norm": 0.01777452416718006,
+ "learning_rate": 0.0006,
+ "loss": 4.521542549133301,
+ "step": 1252
+ },
+ {
+ "epoch": 17.40524017467249,
+ "grad_norm": 0.016626447439193726,
+ "learning_rate": 0.0006,
+ "loss": 4.52629280090332,
+ "step": 1253
+ },
+ {
+ "epoch": 17.419213973799128,
+ "grad_norm": 0.01599477231502533,
+ "learning_rate": 0.0006,
+ "loss": 4.4796037673950195,
+ "step": 1254
+ },
+ {
+ "epoch": 17.433187772925763,
+ "grad_norm": 0.017021458595991135,
+ "learning_rate": 0.0006,
+ "loss": 4.468813419342041,
+ "step": 1255
+ },
+ {
+ "epoch": 17.4471615720524,
+ "grad_norm": 0.016925333067774773,
+ "learning_rate": 0.0006,
+ "loss": 4.584898948669434,
+ "step": 1256
+ },
+ {
+ "epoch": 17.46113537117904,
+ "grad_norm": 0.01714998669922352,
+ "learning_rate": 0.0006,
+ "loss": 4.523544788360596,
+ "step": 1257
+ },
+ {
+ "epoch": 17.475109170305675,
+ "grad_norm": 0.01802288182079792,
+ "learning_rate": 0.0006,
+ "loss": 4.595581531524658,
+ "step": 1258
+ },
+ {
+ "epoch": 17.489082969432314,
+ "grad_norm": 0.0180289875715971,
+ "learning_rate": 0.0006,
+ "loss": 4.394440650939941,
+ "step": 1259
+ },
+ {
+ "epoch": 17.503056768558952,
+ "grad_norm": 0.018504245206713676,
+ "learning_rate": 0.0006,
+ "loss": 4.4794416427612305,
+ "step": 1260
+ },
+ {
+ "epoch": 17.51703056768559,
+ "grad_norm": 0.019973233342170715,
+ "learning_rate": 0.0006,
+ "loss": 4.4241437911987305,
+ "step": 1261
+ },
+ {
+ "epoch": 17.531004366812226,
+ "grad_norm": 0.020063083618879318,
+ "learning_rate": 0.0006,
+ "loss": 4.48068380355835,
+ "step": 1262
+ },
+ {
+ "epoch": 17.544978165938865,
+ "grad_norm": 0.02066640742123127,
+ "learning_rate": 0.0006,
+ "loss": 4.628993034362793,
+ "step": 1263
+ },
+ {
+ "epoch": 17.558951965065503,
+ "grad_norm": 0.02206416428089142,
+ "learning_rate": 0.0006,
+ "loss": 4.564790725708008,
+ "step": 1264
+ },
+ {
+ "epoch": 17.57292576419214,
+ "grad_norm": 0.020640335977077484,
+ "learning_rate": 0.0006,
+ "loss": 4.578705787658691,
+ "step": 1265
+ },
+ {
+ "epoch": 17.586899563318777,
+ "grad_norm": 0.019093405455350876,
+ "learning_rate": 0.0006,
+ "loss": 4.5023603439331055,
+ "step": 1266
+ },
+ {
+ "epoch": 17.600873362445416,
+ "grad_norm": 0.019459193572402,
+ "learning_rate": 0.0006,
+ "loss": 4.594531059265137,
+ "step": 1267
+ },
+ {
+ "epoch": 17.61484716157205,
+ "grad_norm": 0.02021171525120735,
+ "learning_rate": 0.0006,
+ "loss": 4.770299911499023,
+ "step": 1268
+ },
+ {
+ "epoch": 17.62882096069869,
+ "grad_norm": 0.021161191165447235,
+ "learning_rate": 0.0006,
+ "loss": 4.5380401611328125,
+ "step": 1269
+ },
+ {
+ "epoch": 17.64279475982533,
+ "grad_norm": 0.021013403311371803,
+ "learning_rate": 0.0006,
+ "loss": 4.455524921417236,
+ "step": 1270
+ },
+ {
+ "epoch": 17.656768558951963,
+ "grad_norm": 0.020336592569947243,
+ "learning_rate": 0.0006,
+ "loss": 4.533672332763672,
+ "step": 1271
+ },
+ {
+ "epoch": 17.670742358078602,
+ "grad_norm": 0.020159434527158737,
+ "learning_rate": 0.0006,
+ "loss": 4.596151351928711,
+ "step": 1272
+ },
+ {
+ "epoch": 17.68471615720524,
+ "grad_norm": 0.020432960242033005,
+ "learning_rate": 0.0006,
+ "loss": 4.55531644821167,
+ "step": 1273
+ },
+ {
+ "epoch": 17.69868995633188,
+ "grad_norm": 0.025466229766607285,
+ "learning_rate": 0.0006,
+ "loss": 4.566912651062012,
+ "step": 1274
+ },
+ {
+ "epoch": 17.712663755458514,
+ "grad_norm": 0.026537558063864708,
+ "learning_rate": 0.0006,
+ "loss": 4.526003837585449,
+ "step": 1275
+ },
+ {
+ "epoch": 17.726637554585153,
+ "grad_norm": 0.024397607892751694,
+ "learning_rate": 0.0006,
+ "loss": 4.541755676269531,
+ "step": 1276
+ },
+ {
+ "epoch": 17.74061135371179,
+ "grad_norm": 0.022062312811613083,
+ "learning_rate": 0.0006,
+ "loss": 4.6532793045043945,
+ "step": 1277
+ },
+ {
+ "epoch": 17.754585152838427,
+ "grad_norm": 0.02204282023012638,
+ "learning_rate": 0.0006,
+ "loss": 4.582825660705566,
+ "step": 1278
+ },
+ {
+ "epoch": 17.768558951965066,
+ "grad_norm": 0.020844416692852974,
+ "learning_rate": 0.0006,
+ "loss": 4.522336483001709,
+ "step": 1279
+ },
+ {
+ "epoch": 17.782532751091704,
+ "grad_norm": 0.019815703853964806,
+ "learning_rate": 0.0006,
+ "loss": 4.607398509979248,
+ "step": 1280
+ },
+ {
+ "epoch": 17.79650655021834,
+ "grad_norm": 0.016676954925060272,
+ "learning_rate": 0.0006,
+ "loss": 4.446206092834473,
+ "step": 1281
+ },
+ {
+ "epoch": 17.810480349344978,
+ "grad_norm": 0.0166452769190073,
+ "learning_rate": 0.0006,
+ "loss": 4.505931854248047,
+ "step": 1282
+ },
+ {
+ "epoch": 17.824454148471617,
+ "grad_norm": 0.015928907319903374,
+ "learning_rate": 0.0006,
+ "loss": 4.455267429351807,
+ "step": 1283
+ },
+ {
+ "epoch": 17.83842794759825,
+ "grad_norm": 0.015219560824334621,
+ "learning_rate": 0.0006,
+ "loss": 4.5623674392700195,
+ "step": 1284
+ },
+ {
+ "epoch": 17.85240174672489,
+ "grad_norm": 0.015125355683267117,
+ "learning_rate": 0.0006,
+ "loss": 4.554104804992676,
+ "step": 1285
+ },
+ {
+ "epoch": 17.86637554585153,
+ "grad_norm": 0.015612718649208546,
+ "learning_rate": 0.0006,
+ "loss": 4.484847068786621,
+ "step": 1286
+ },
+ {
+ "epoch": 17.880349344978168,
+ "grad_norm": 0.015215662308037281,
+ "learning_rate": 0.0006,
+ "loss": 4.4802093505859375,
+ "step": 1287
+ },
+ {
+ "epoch": 17.894323144104803,
+ "grad_norm": 0.01511684525758028,
+ "learning_rate": 0.0006,
+ "loss": 4.537298679351807,
+ "step": 1288
+ },
+ {
+ "epoch": 17.90829694323144,
+ "grad_norm": 0.013841322623193264,
+ "learning_rate": 0.0006,
+ "loss": 4.427488327026367,
+ "step": 1289
+ },
+ {
+ "epoch": 17.92227074235808,
+ "grad_norm": 0.015115964226424694,
+ "learning_rate": 0.0006,
+ "loss": 4.4547119140625,
+ "step": 1290
+ },
+ {
+ "epoch": 17.936244541484715,
+ "grad_norm": 0.016411006450653076,
+ "learning_rate": 0.0006,
+ "loss": 4.52994966506958,
+ "step": 1291
+ },
+ {
+ "epoch": 17.950218340611354,
+ "grad_norm": 0.017313152551651,
+ "learning_rate": 0.0006,
+ "loss": 4.5337653160095215,
+ "step": 1292
+ },
+ {
+ "epoch": 17.964192139737992,
+ "grad_norm": 0.016559796407818794,
+ "learning_rate": 0.0006,
+ "loss": 4.5277910232543945,
+ "step": 1293
+ },
+ {
+ "epoch": 17.978165938864628,
+ "grad_norm": 0.015854761004447937,
+ "learning_rate": 0.0006,
+ "loss": 4.6334309577941895,
+ "step": 1294
+ },
+ {
+ "epoch": 17.992139737991266,
+ "grad_norm": 0.015658238902688026,
+ "learning_rate": 0.0006,
+ "loss": 4.613372802734375,
+ "step": 1295
+ },
+ {
+ "epoch": 18.0,
+ "grad_norm": 0.0169155802577734,
+ "learning_rate": 0.0006,
+ "loss": 4.527356147766113,
+ "step": 1296
+ },
+ {
+ "epoch": 18.0,
+ "eval_loss": 4.764543056488037,
+ "eval_runtime": 56.5998,
+ "eval_samples_per_second": 43.145,
+ "eval_steps_per_second": 1.36,
+ "step": 1296
+ },
+ {
+ "epoch": 18.01397379912664,
+ "grad_norm": 0.015348963439464569,
+ "learning_rate": 0.0006,
+ "loss": 4.466423988342285,
+ "step": 1297
+ },
+ {
+ "epoch": 18.027947598253274,
+ "grad_norm": 0.01768588088452816,
+ "learning_rate": 0.0006,
+ "loss": 4.545933246612549,
+ "step": 1298
+ },
+ {
+ "epoch": 18.041921397379912,
+ "grad_norm": 0.017381226643919945,
+ "learning_rate": 0.0006,
+ "loss": 4.494060039520264,
+ "step": 1299
+ },
+ {
+ "epoch": 18.05589519650655,
+ "grad_norm": 0.019002335146069527,
+ "learning_rate": 0.0006,
+ "loss": 4.395200729370117,
+ "step": 1300
+ },
+ {
+ "epoch": 18.069868995633186,
+ "grad_norm": 0.01827932707965374,
+ "learning_rate": 0.0006,
+ "loss": 4.424848556518555,
+ "step": 1301
+ },
+ {
+ "epoch": 18.083842794759825,
+ "grad_norm": 0.016383031383156776,
+ "learning_rate": 0.0006,
+ "loss": 4.494722366333008,
+ "step": 1302
+ },
+ {
+ "epoch": 18.097816593886463,
+ "grad_norm": 0.016265157610177994,
+ "learning_rate": 0.0006,
+ "loss": 4.416522979736328,
+ "step": 1303
+ },
+ {
+ "epoch": 18.111790393013102,
+ "grad_norm": 0.01702386513352394,
+ "learning_rate": 0.0006,
+ "loss": 4.600103378295898,
+ "step": 1304
+ },
+ {
+ "epoch": 18.125764192139737,
+ "grad_norm": 0.018495498225092888,
+ "learning_rate": 0.0006,
+ "loss": 4.588401794433594,
+ "step": 1305
+ },
+ {
+ "epoch": 18.139737991266376,
+ "grad_norm": 0.017921684309840202,
+ "learning_rate": 0.0006,
+ "loss": 4.478075981140137,
+ "step": 1306
+ },
+ {
+ "epoch": 18.153711790393015,
+ "grad_norm": 0.01801278069615364,
+ "learning_rate": 0.0006,
+ "loss": 4.413186550140381,
+ "step": 1307
+ },
+ {
+ "epoch": 18.16768558951965,
+ "grad_norm": 0.017211275175213814,
+ "learning_rate": 0.0006,
+ "loss": 4.461341857910156,
+ "step": 1308
+ },
+ {
+ "epoch": 18.18165938864629,
+ "grad_norm": 0.017589328810572624,
+ "learning_rate": 0.0006,
+ "loss": 4.546969413757324,
+ "step": 1309
+ },
+ {
+ "epoch": 18.195633187772927,
+ "grad_norm": 0.01797698810696602,
+ "learning_rate": 0.0006,
+ "loss": 4.530797004699707,
+ "step": 1310
+ },
+ {
+ "epoch": 18.209606986899562,
+ "grad_norm": 0.020408619195222855,
+ "learning_rate": 0.0006,
+ "loss": 4.62999153137207,
+ "step": 1311
+ },
+ {
+ "epoch": 18.2235807860262,
+ "grad_norm": 0.01994038000702858,
+ "learning_rate": 0.0006,
+ "loss": 4.586811065673828,
+ "step": 1312
+ },
+ {
+ "epoch": 18.23755458515284,
+ "grad_norm": 0.019381340593099594,
+ "learning_rate": 0.0006,
+ "loss": 4.4769368171691895,
+ "step": 1313
+ },
+ {
+ "epoch": 18.251528384279474,
+ "grad_norm": 0.020484555512666702,
+ "learning_rate": 0.0006,
+ "loss": 4.524302959442139,
+ "step": 1314
+ },
+ {
+ "epoch": 18.265502183406113,
+ "grad_norm": 0.02426416426897049,
+ "learning_rate": 0.0006,
+ "loss": 4.561705589294434,
+ "step": 1315
+ },
+ {
+ "epoch": 18.27947598253275,
+ "grad_norm": 0.024748101830482483,
+ "learning_rate": 0.0006,
+ "loss": 4.4654316902160645,
+ "step": 1316
+ },
+ {
+ "epoch": 18.29344978165939,
+ "grad_norm": 0.020703328773379326,
+ "learning_rate": 0.0006,
+ "loss": 4.53033447265625,
+ "step": 1317
+ },
+ {
+ "epoch": 18.307423580786025,
+ "grad_norm": 0.01765458658337593,
+ "learning_rate": 0.0006,
+ "loss": 4.5154266357421875,
+ "step": 1318
+ },
+ {
+ "epoch": 18.321397379912664,
+ "grad_norm": 0.018646420910954475,
+ "learning_rate": 0.0006,
+ "loss": 4.5224714279174805,
+ "step": 1319
+ },
+ {
+ "epoch": 18.335371179039303,
+ "grad_norm": 0.01907474733889103,
+ "learning_rate": 0.0006,
+ "loss": 4.5055437088012695,
+ "step": 1320
+ },
+ {
+ "epoch": 18.349344978165938,
+ "grad_norm": 0.018179846927523613,
+ "learning_rate": 0.0006,
+ "loss": 4.462161064147949,
+ "step": 1321
+ },
+ {
+ "epoch": 18.363318777292577,
+ "grad_norm": 0.016697878018021584,
+ "learning_rate": 0.0006,
+ "loss": 4.473718643188477,
+ "step": 1322
+ },
+ {
+ "epoch": 18.377292576419215,
+ "grad_norm": 0.017504019662737846,
+ "learning_rate": 0.0006,
+ "loss": 4.474348068237305,
+ "step": 1323
+ },
+ {
+ "epoch": 18.39126637554585,
+ "grad_norm": 0.018889913335442543,
+ "learning_rate": 0.0006,
+ "loss": 4.4531636238098145,
+ "step": 1324
+ },
+ {
+ "epoch": 18.40524017467249,
+ "grad_norm": 0.019167251884937286,
+ "learning_rate": 0.0006,
+ "loss": 4.4278364181518555,
+ "step": 1325
+ },
+ {
+ "epoch": 18.419213973799128,
+ "grad_norm": 0.015895741060376167,
+ "learning_rate": 0.0006,
+ "loss": 4.474592208862305,
+ "step": 1326
+ },
+ {
+ "epoch": 18.433187772925763,
+ "grad_norm": 0.01531882956624031,
+ "learning_rate": 0.0006,
+ "loss": 4.403757572174072,
+ "step": 1327
+ },
+ {
+ "epoch": 18.4471615720524,
+ "grad_norm": 0.01565849967300892,
+ "learning_rate": 0.0006,
+ "loss": 4.507664680480957,
+ "step": 1328
+ },
+ {
+ "epoch": 18.46113537117904,
+ "grad_norm": 0.01602315716445446,
+ "learning_rate": 0.0006,
+ "loss": 4.5648908615112305,
+ "step": 1329
+ },
+ {
+ "epoch": 18.475109170305675,
+ "grad_norm": 0.016467366367578506,
+ "learning_rate": 0.0006,
+ "loss": 4.428762912750244,
+ "step": 1330
+ },
+ {
+ "epoch": 18.489082969432314,
+ "grad_norm": 0.01604127697646618,
+ "learning_rate": 0.0006,
+ "loss": 4.527771472930908,
+ "step": 1331
+ },
+ {
+ "epoch": 18.503056768558952,
+ "grad_norm": 0.018825586885213852,
+ "learning_rate": 0.0006,
+ "loss": 4.485386848449707,
+ "step": 1332
+ },
+ {
+ "epoch": 18.51703056768559,
+ "grad_norm": 0.019258007407188416,
+ "learning_rate": 0.0006,
+ "loss": 4.414972305297852,
+ "step": 1333
+ },
+ {
+ "epoch": 18.531004366812226,
+ "grad_norm": 0.018192021176218987,
+ "learning_rate": 0.0006,
+ "loss": 4.412166595458984,
+ "step": 1334
+ },
+ {
+ "epoch": 18.544978165938865,
+ "grad_norm": 0.016120698302984238,
+ "learning_rate": 0.0006,
+ "loss": 4.484317779541016,
+ "step": 1335
+ },
+ {
+ "epoch": 18.558951965065503,
+ "grad_norm": 0.016496622934937477,
+ "learning_rate": 0.0006,
+ "loss": 4.514986515045166,
+ "step": 1336
+ },
+ {
+ "epoch": 18.57292576419214,
+ "grad_norm": 0.01795116998255253,
+ "learning_rate": 0.0006,
+ "loss": 4.41473388671875,
+ "step": 1337
+ },
+ {
+ "epoch": 18.586899563318777,
+ "grad_norm": 0.019062276929616928,
+ "learning_rate": 0.0006,
+ "loss": 4.601996898651123,
+ "step": 1338
+ },
+ {
+ "epoch": 18.600873362445416,
+ "grad_norm": 0.017174631357192993,
+ "learning_rate": 0.0006,
+ "loss": 4.571218490600586,
+ "step": 1339
+ },
+ {
+ "epoch": 18.61484716157205,
+ "grad_norm": 0.014015606604516506,
+ "learning_rate": 0.0006,
+ "loss": 4.526606559753418,
+ "step": 1340
+ },
+ {
+ "epoch": 18.62882096069869,
+ "grad_norm": 0.015358510427176952,
+ "learning_rate": 0.0006,
+ "loss": 4.604286193847656,
+ "step": 1341
+ },
+ {
+ "epoch": 18.64279475982533,
+ "grad_norm": 0.01619391329586506,
+ "learning_rate": 0.0006,
+ "loss": 4.560883522033691,
+ "step": 1342
+ },
+ {
+ "epoch": 18.656768558951963,
+ "grad_norm": 0.015972964465618134,
+ "learning_rate": 0.0006,
+ "loss": 4.479755401611328,
+ "step": 1343
+ },
+ {
+ "epoch": 18.670742358078602,
+ "grad_norm": 0.016861025243997574,
+ "learning_rate": 0.0006,
+ "loss": 4.474471092224121,
+ "step": 1344
+ },
+ {
+ "epoch": 18.68471615720524,
+ "grad_norm": 0.01822078227996826,
+ "learning_rate": 0.0006,
+ "loss": 4.509489059448242,
+ "step": 1345
+ },
+ {
+ "epoch": 18.69868995633188,
+ "grad_norm": 0.019415007904171944,
+ "learning_rate": 0.0006,
+ "loss": 4.508941173553467,
+ "step": 1346
+ },
+ {
+ "epoch": 18.712663755458514,
+ "grad_norm": 0.019856126978993416,
+ "learning_rate": 0.0006,
+ "loss": 4.446670055389404,
+ "step": 1347
+ },
+ {
+ "epoch": 18.726637554585153,
+ "grad_norm": 0.01842520199716091,
+ "learning_rate": 0.0006,
+ "loss": 4.524422645568848,
+ "step": 1348
+ },
+ {
+ "epoch": 18.74061135371179,
+ "grad_norm": 0.017196480184793472,
+ "learning_rate": 0.0006,
+ "loss": 4.498466968536377,
+ "step": 1349
+ },
+ {
+ "epoch": 18.754585152838427,
+ "grad_norm": 0.01823010854423046,
+ "learning_rate": 0.0006,
+ "loss": 4.534684181213379,
+ "step": 1350
+ },
+ {
+ "epoch": 18.768558951965066,
+ "grad_norm": 0.020398065447807312,
+ "learning_rate": 0.0006,
+ "loss": 4.456494331359863,
+ "step": 1351
+ },
+ {
+ "epoch": 18.782532751091704,
+ "grad_norm": 0.022992262616753578,
+ "learning_rate": 0.0006,
+ "loss": 4.476374626159668,
+ "step": 1352
+ },
+ {
+ "epoch": 18.79650655021834,
+ "grad_norm": 0.02235420234501362,
+ "learning_rate": 0.0006,
+ "loss": 4.587776184082031,
+ "step": 1353
+ },
+ {
+ "epoch": 18.810480349344978,
+ "grad_norm": 0.02093626745045185,
+ "learning_rate": 0.0006,
+ "loss": 4.424172401428223,
+ "step": 1354
+ },
+ {
+ "epoch": 18.824454148471617,
+ "grad_norm": 0.01997302658855915,
+ "learning_rate": 0.0006,
+ "loss": 4.448066711425781,
+ "step": 1355
+ },
+ {
+ "epoch": 18.83842794759825,
+ "grad_norm": 0.018265459686517715,
+ "learning_rate": 0.0006,
+ "loss": 4.436610698699951,
+ "step": 1356
+ },
+ {
+ "epoch": 18.85240174672489,
+ "grad_norm": 0.01673833839595318,
+ "learning_rate": 0.0006,
+ "loss": 4.5502214431762695,
+ "step": 1357
+ },
+ {
+ "epoch": 18.86637554585153,
+ "grad_norm": 0.017908833920955658,
+ "learning_rate": 0.0006,
+ "loss": 4.550530433654785,
+ "step": 1358
+ },
+ {
+ "epoch": 18.880349344978168,
+ "grad_norm": 0.01796654611825943,
+ "learning_rate": 0.0006,
+ "loss": 4.531463146209717,
+ "step": 1359
+ },
+ {
+ "epoch": 18.894323144104803,
+ "grad_norm": 0.01646183803677559,
+ "learning_rate": 0.0006,
+ "loss": 4.498582363128662,
+ "step": 1360
+ },
+ {
+ "epoch": 18.90829694323144,
+ "grad_norm": 0.015430327504873276,
+ "learning_rate": 0.0006,
+ "loss": 4.52571439743042,
+ "step": 1361
+ },
+ {
+ "epoch": 18.92227074235808,
+ "grad_norm": 0.016527952626347542,
+ "learning_rate": 0.0006,
+ "loss": 4.547013759613037,
+ "step": 1362
+ },
+ {
+ "epoch": 18.936244541484715,
+ "grad_norm": 0.016263660043478012,
+ "learning_rate": 0.0006,
+ "loss": 4.468754291534424,
+ "step": 1363
+ },
+ {
+ "epoch": 18.950218340611354,
+ "grad_norm": 0.016347210854291916,
+ "learning_rate": 0.0006,
+ "loss": 4.450741767883301,
+ "step": 1364
+ },
+ {
+ "epoch": 18.964192139737992,
+ "grad_norm": 0.01701398566365242,
+ "learning_rate": 0.0006,
+ "loss": 4.534899711608887,
+ "step": 1365
+ },
+ {
+ "epoch": 18.978165938864628,
+ "grad_norm": 0.017693081870675087,
+ "learning_rate": 0.0006,
+ "loss": 4.460553169250488,
+ "step": 1366
+ },
+ {
+ "epoch": 18.992139737991266,
+ "grad_norm": 0.018746308982372284,
+ "learning_rate": 0.0006,
+ "loss": 4.538792610168457,
+ "step": 1367
+ },
+ {
+ "epoch": 19.0,
+ "grad_norm": 0.020099416375160217,
+ "learning_rate": 0.0006,
+ "loss": 4.449617385864258,
+ "step": 1368
+ },
+ {
+ "epoch": 19.0,
+ "eval_loss": 4.7756571769714355,
+ "eval_runtime": 57.1804,
+ "eval_samples_per_second": 42.707,
+ "eval_steps_per_second": 1.347,
+ "step": 1368
+ },
+ {
+ "epoch": 19.01397379912664,
+ "grad_norm": 0.018726933747529984,
+ "learning_rate": 0.0006,
+ "loss": 4.446059703826904,
+ "step": 1369
+ },
+ {
+ "epoch": 19.027947598253274,
+ "grad_norm": 0.01956087350845337,
+ "learning_rate": 0.0006,
+ "loss": 4.4520039558410645,
+ "step": 1370
+ },
+ {
+ "epoch": 19.041921397379912,
+ "grad_norm": 0.020092811435461044,
+ "learning_rate": 0.0006,
+ "loss": 4.543343544006348,
+ "step": 1371
+ },
+ {
+ "epoch": 19.05589519650655,
+ "grad_norm": 0.01843622885644436,
+ "learning_rate": 0.0006,
+ "loss": 4.535579204559326,
+ "step": 1372
+ },
+ {
+ "epoch": 19.069868995633186,
+ "grad_norm": 0.016989488154649734,
+ "learning_rate": 0.0006,
+ "loss": 4.52316951751709,
+ "step": 1373
+ },
+ {
+ "epoch": 19.083842794759825,
+ "grad_norm": 0.01805422455072403,
+ "learning_rate": 0.0006,
+ "loss": 4.5393266677856445,
+ "step": 1374
+ },
+ {
+ "epoch": 19.097816593886463,
+ "grad_norm": 0.01874629408121109,
+ "learning_rate": 0.0006,
+ "loss": 4.422059059143066,
+ "step": 1375
+ },
+ {
+ "epoch": 19.111790393013102,
+ "grad_norm": 0.01804221048951149,
+ "learning_rate": 0.0006,
+ "loss": 4.50508451461792,
+ "step": 1376
+ },
+ {
+ "epoch": 19.125764192139737,
+ "grad_norm": 0.01785474829375744,
+ "learning_rate": 0.0006,
+ "loss": 4.412599563598633,
+ "step": 1377
+ },
+ {
+ "epoch": 19.139737991266376,
+ "grad_norm": 0.01971791312098503,
+ "learning_rate": 0.0006,
+ "loss": 4.518318176269531,
+ "step": 1378
+ },
+ {
+ "epoch": 19.153711790393015,
+ "grad_norm": 0.02061633951961994,
+ "learning_rate": 0.0006,
+ "loss": 4.332393169403076,
+ "step": 1379
+ },
+ {
+ "epoch": 19.16768558951965,
+ "grad_norm": 0.020563429221510887,
+ "learning_rate": 0.0006,
+ "loss": 4.445863723754883,
+ "step": 1380
+ },
+ {
+ "epoch": 19.18165938864629,
+ "grad_norm": 0.019995713606476784,
+ "learning_rate": 0.0006,
+ "loss": 4.497312068939209,
+ "step": 1381
+ },
+ {
+ "epoch": 19.195633187772927,
+ "grad_norm": 0.019823070615530014,
+ "learning_rate": 0.0006,
+ "loss": 4.475085258483887,
+ "step": 1382
+ },
+ {
+ "epoch": 19.209606986899562,
+ "grad_norm": 0.019039355218410492,
+ "learning_rate": 0.0006,
+ "loss": 4.50432014465332,
+ "step": 1383
+ },
+ {
+ "epoch": 19.2235807860262,
+ "grad_norm": 0.017876973375678062,
+ "learning_rate": 0.0006,
+ "loss": 4.384788513183594,
+ "step": 1384
+ },
+ {
+ "epoch": 19.23755458515284,
+ "grad_norm": 0.018697699531912804,
+ "learning_rate": 0.0006,
+ "loss": 4.410778045654297,
+ "step": 1385
+ },
+ {
+ "epoch": 19.251528384279474,
+ "grad_norm": 0.018826806917786598,
+ "learning_rate": 0.0006,
+ "loss": 4.524034023284912,
+ "step": 1386
+ },
+ {
+ "epoch": 19.265502183406113,
+ "grad_norm": 0.01824226602911949,
+ "learning_rate": 0.0006,
+ "loss": 4.526393890380859,
+ "step": 1387
+ },
+ {
+ "epoch": 19.27947598253275,
+ "grad_norm": 0.01753208599984646,
+ "learning_rate": 0.0006,
+ "loss": 4.493703365325928,
+ "step": 1388
+ },
+ {
+ "epoch": 19.29344978165939,
+ "grad_norm": 0.01742432825267315,
+ "learning_rate": 0.0006,
+ "loss": 4.392831802368164,
+ "step": 1389
+ },
+ {
+ "epoch": 19.307423580786025,
+ "grad_norm": 0.017099356278777122,
+ "learning_rate": 0.0006,
+ "loss": 4.578668594360352,
+ "step": 1390
+ },
+ {
+ "epoch": 19.321397379912664,
+ "grad_norm": 0.017946023494005203,
+ "learning_rate": 0.0006,
+ "loss": 4.456718444824219,
+ "step": 1391
+ },
+ {
+ "epoch": 19.335371179039303,
+ "grad_norm": 0.017705943435430527,
+ "learning_rate": 0.0006,
+ "loss": 4.553682327270508,
+ "step": 1392
+ },
+ {
+ "epoch": 19.349344978165938,
+ "grad_norm": 0.016647523269057274,
+ "learning_rate": 0.0006,
+ "loss": 4.525016784667969,
+ "step": 1393
+ },
+ {
+ "epoch": 19.363318777292577,
+ "grad_norm": 0.015407416969537735,
+ "learning_rate": 0.0006,
+ "loss": 4.500231742858887,
+ "step": 1394
+ },
+ {
+ "epoch": 19.377292576419215,
+ "grad_norm": 0.017264019697904587,
+ "learning_rate": 0.0006,
+ "loss": 4.4615349769592285,
+ "step": 1395
+ },
+ {
+ "epoch": 19.39126637554585,
+ "grad_norm": 0.01871555484831333,
+ "learning_rate": 0.0006,
+ "loss": 4.281026840209961,
+ "step": 1396
+ },
+ {
+ "epoch": 19.40524017467249,
+ "grad_norm": 0.017361650243401527,
+ "learning_rate": 0.0006,
+ "loss": 4.479382514953613,
+ "step": 1397
+ },
+ {
+ "epoch": 19.419213973799128,
+ "grad_norm": 0.019781362265348434,
+ "learning_rate": 0.0006,
+ "loss": 4.402320861816406,
+ "step": 1398
+ },
+ {
+ "epoch": 19.433187772925763,
+ "grad_norm": 0.019043298438191414,
+ "learning_rate": 0.0006,
+ "loss": 4.476983070373535,
+ "step": 1399
+ },
+ {
+ "epoch": 19.4471615720524,
+ "grad_norm": 0.017137985676527023,
+ "learning_rate": 0.0006,
+ "loss": 4.418689250946045,
+ "step": 1400
+ },
+ {
+ "epoch": 19.46113537117904,
+ "grad_norm": 0.016481833532452583,
+ "learning_rate": 0.0006,
+ "loss": 4.579145431518555,
+ "step": 1401
+ },
+ {
+ "epoch": 19.475109170305675,
+ "grad_norm": 0.016507714986801147,
+ "learning_rate": 0.0006,
+ "loss": 4.519158363342285,
+ "step": 1402
+ },
+ {
+ "epoch": 19.489082969432314,
+ "grad_norm": 0.018092872574925423,
+ "learning_rate": 0.0006,
+ "loss": 4.513748645782471,
+ "step": 1403
+ },
+ {
+ "epoch": 19.503056768558952,
+ "grad_norm": 0.018153440207242966,
+ "learning_rate": 0.0006,
+ "loss": 4.645384311676025,
+ "step": 1404
+ },
+ {
+ "epoch": 19.51703056768559,
+ "grad_norm": 0.01657119207084179,
+ "learning_rate": 0.0006,
+ "loss": 4.399757385253906,
+ "step": 1405
+ },
+ {
+ "epoch": 19.531004366812226,
+ "grad_norm": 0.0156280305236578,
+ "learning_rate": 0.0006,
+ "loss": 4.467836380004883,
+ "step": 1406
+ },
+ {
+ "epoch": 19.544978165938865,
+ "grad_norm": 0.016838403418660164,
+ "learning_rate": 0.0006,
+ "loss": 4.529176235198975,
+ "step": 1407
+ },
+ {
+ "epoch": 19.558951965065503,
+ "grad_norm": 0.01743236929178238,
+ "learning_rate": 0.0006,
+ "loss": 4.440032482147217,
+ "step": 1408
+ },
+ {
+ "epoch": 19.57292576419214,
+ "grad_norm": 0.01553610060364008,
+ "learning_rate": 0.0006,
+ "loss": 4.425588607788086,
+ "step": 1409
+ },
+ {
+ "epoch": 19.586899563318777,
+ "grad_norm": 0.017259271815419197,
+ "learning_rate": 0.0006,
+ "loss": 4.405452728271484,
+ "step": 1410
+ },
+ {
+ "epoch": 19.600873362445416,
+ "grad_norm": 0.015568481758236885,
+ "learning_rate": 0.0006,
+ "loss": 4.444279193878174,
+ "step": 1411
+ },
+ {
+ "epoch": 19.61484716157205,
+ "grad_norm": 0.015331248752772808,
+ "learning_rate": 0.0006,
+ "loss": 4.451846122741699,
+ "step": 1412
+ },
+ {
+ "epoch": 19.62882096069869,
+ "grad_norm": 0.015667378902435303,
+ "learning_rate": 0.0006,
+ "loss": 4.533931255340576,
+ "step": 1413
+ },
+ {
+ "epoch": 19.64279475982533,
+ "grad_norm": 0.016177602112293243,
+ "learning_rate": 0.0006,
+ "loss": 4.467313289642334,
+ "step": 1414
+ },
+ {
+ "epoch": 19.656768558951963,
+ "grad_norm": 0.01733129471540451,
+ "learning_rate": 0.0006,
+ "loss": 4.496893882751465,
+ "step": 1415
+ },
+ {
+ "epoch": 19.670742358078602,
+ "grad_norm": 0.016539594158530235,
+ "learning_rate": 0.0006,
+ "loss": 4.401961326599121,
+ "step": 1416
+ },
+ {
+ "epoch": 19.68471615720524,
+ "grad_norm": 0.016058551147580147,
+ "learning_rate": 0.0006,
+ "loss": 4.503375053405762,
+ "step": 1417
+ },
+ {
+ "epoch": 19.69868995633188,
+ "grad_norm": 0.01715777814388275,
+ "learning_rate": 0.0006,
+ "loss": 4.3808369636535645,
+ "step": 1418
+ },
+ {
+ "epoch": 19.712663755458514,
+ "grad_norm": 0.01681119203567505,
+ "learning_rate": 0.0006,
+ "loss": 4.5469560623168945,
+ "step": 1419
+ },
+ {
+ "epoch": 19.726637554585153,
+ "grad_norm": 0.01634550653398037,
+ "learning_rate": 0.0006,
+ "loss": 4.328197956085205,
+ "step": 1420
+ },
+ {
+ "epoch": 19.74061135371179,
+ "grad_norm": 0.01549347210675478,
+ "learning_rate": 0.0006,
+ "loss": 4.534621238708496,
+ "step": 1421
+ },
+ {
+ "epoch": 19.754585152838427,
+ "grad_norm": 0.01654665172100067,
+ "learning_rate": 0.0006,
+ "loss": 4.4724225997924805,
+ "step": 1422
+ },
+ {
+ "epoch": 19.768558951965066,
+ "grad_norm": 0.017119025811553,
+ "learning_rate": 0.0006,
+ "loss": 4.3897247314453125,
+ "step": 1423
+ },
+ {
+ "epoch": 19.782532751091704,
+ "grad_norm": 0.01603223942220211,
+ "learning_rate": 0.0006,
+ "loss": 4.412731170654297,
+ "step": 1424
+ },
+ {
+ "epoch": 19.79650655021834,
+ "grad_norm": 0.016774067655205727,
+ "learning_rate": 0.0006,
+ "loss": 4.54707145690918,
+ "step": 1425
+ },
+ {
+ "epoch": 19.810480349344978,
+ "grad_norm": 0.016453025862574577,
+ "learning_rate": 0.0006,
+ "loss": 4.462501049041748,
+ "step": 1426
+ },
+ {
+ "epoch": 19.824454148471617,
+ "grad_norm": 0.01597629487514496,
+ "learning_rate": 0.0006,
+ "loss": 4.229758262634277,
+ "step": 1427
+ },
+ {
+ "epoch": 19.83842794759825,
+ "grad_norm": 0.017999034374952316,
+ "learning_rate": 0.0006,
+ "loss": 4.389263153076172,
+ "step": 1428
+ },
+ {
+ "epoch": 19.85240174672489,
+ "grad_norm": 0.02165898121893406,
+ "learning_rate": 0.0006,
+ "loss": 4.342130661010742,
+ "step": 1429
+ },
+ {
+ "epoch": 19.86637554585153,
+ "grad_norm": 0.02222822792828083,
+ "learning_rate": 0.0006,
+ "loss": 4.492661476135254,
+ "step": 1430
+ },
+ {
+ "epoch": 19.880349344978168,
+ "grad_norm": 0.019378066062927246,
+ "learning_rate": 0.0006,
+ "loss": 4.4614105224609375,
+ "step": 1431
+ },
+ {
+ "epoch": 19.894323144104803,
+ "grad_norm": 0.019908083602786064,
+ "learning_rate": 0.0006,
+ "loss": 4.370929718017578,
+ "step": 1432
+ },
+ {
+ "epoch": 19.90829694323144,
+ "grad_norm": 0.022296471521258354,
+ "learning_rate": 0.0006,
+ "loss": 4.438320159912109,
+ "step": 1433
+ },
+ {
+ "epoch": 19.92227074235808,
+ "grad_norm": 0.021810343489050865,
+ "learning_rate": 0.0006,
+ "loss": 4.462067127227783,
+ "step": 1434
+ },
+ {
+ "epoch": 19.936244541484715,
+ "grad_norm": 0.02129209227859974,
+ "learning_rate": 0.0006,
+ "loss": 4.541261196136475,
+ "step": 1435
+ },
+ {
+ "epoch": 19.950218340611354,
+ "grad_norm": 0.02079629711806774,
+ "learning_rate": 0.0006,
+ "loss": 4.36712646484375,
+ "step": 1436
+ },
+ {
+ "epoch": 19.964192139737992,
+ "grad_norm": 0.022501740604639053,
+ "learning_rate": 0.0006,
+ "loss": 4.5279083251953125,
+ "step": 1437
+ },
+ {
+ "epoch": 19.978165938864628,
+ "grad_norm": 0.019121676683425903,
+ "learning_rate": 0.0006,
+ "loss": 4.477789878845215,
+ "step": 1438
+ },
+ {
+ "epoch": 19.992139737991266,
+ "grad_norm": 0.017429588362574577,
+ "learning_rate": 0.0006,
+ "loss": 4.4023847579956055,
+ "step": 1439
+ },
+ {
+ "epoch": 20.0,
+ "grad_norm": 0.018857665359973907,
+ "learning_rate": 0.0006,
+ "loss": 4.494743347167969,
+ "step": 1440
+ },
+ {
+ "epoch": 20.0,
+ "eval_loss": 4.694555759429932,
+ "eval_runtime": 56.4146,
+ "eval_samples_per_second": 43.287,
+ "eval_steps_per_second": 1.365,
+ "step": 1440
+ },
+ {
+ "epoch": 20.01397379912664,
+ "grad_norm": 0.016505861654877663,
+ "learning_rate": 0.0006,
+ "loss": 4.506229877471924,
+ "step": 1441
+ },
+ {
+ "epoch": 20.027947598253274,
+ "grad_norm": 0.017571842297911644,
+ "learning_rate": 0.0006,
+ "loss": 4.384073257446289,
+ "step": 1442
+ },
+ {
+ "epoch": 20.041921397379912,
+ "grad_norm": 0.019398365169763565,
+ "learning_rate": 0.0006,
+ "loss": 4.482177734375,
+ "step": 1443
+ },
+ {
+ "epoch": 20.05589519650655,
+ "grad_norm": 0.02011062018573284,
+ "learning_rate": 0.0006,
+ "loss": 4.440773963928223,
+ "step": 1444
+ },
+ {
+ "epoch": 20.069868995633186,
+ "grad_norm": 0.019026953727006912,
+ "learning_rate": 0.0006,
+ "loss": 4.507168769836426,
+ "step": 1445
+ },
+ {
+ "epoch": 20.083842794759825,
+ "grad_norm": 0.020506031811237335,
+ "learning_rate": 0.0006,
+ "loss": 4.469079971313477,
+ "step": 1446
+ },
+ {
+ "epoch": 20.097816593886463,
+ "grad_norm": 0.019608162343502045,
+ "learning_rate": 0.0006,
+ "loss": 4.395927429199219,
+ "step": 1447
+ },
+ {
+ "epoch": 20.111790393013102,
+ "grad_norm": 0.019419124349951744,
+ "learning_rate": 0.0006,
+ "loss": 4.484235763549805,
+ "step": 1448
+ },
+ {
+ "epoch": 20.125764192139737,
+ "grad_norm": 0.021775128319859505,
+ "learning_rate": 0.0006,
+ "loss": 4.333173751831055,
+ "step": 1449
+ },
+ {
+ "epoch": 20.139737991266376,
+ "grad_norm": 0.025512272492051125,
+ "learning_rate": 0.0006,
+ "loss": 4.460362434387207,
+ "step": 1450
+ },
+ {
+ "epoch": 20.153711790393015,
+ "grad_norm": 0.025157401338219643,
+ "learning_rate": 0.0006,
+ "loss": 4.4361395835876465,
+ "step": 1451
+ },
+ {
+ "epoch": 20.16768558951965,
+ "grad_norm": 0.019862616434693336,
+ "learning_rate": 0.0006,
+ "loss": 4.426807403564453,
+ "step": 1452
+ },
+ {
+ "epoch": 20.18165938864629,
+ "grad_norm": 0.01932152360677719,
+ "learning_rate": 0.0006,
+ "loss": 4.375677108764648,
+ "step": 1453
+ },
+ {
+ "epoch": 20.195633187772927,
+ "grad_norm": 0.01760999858379364,
+ "learning_rate": 0.0006,
+ "loss": 4.42259407043457,
+ "step": 1454
+ },
+ {
+ "epoch": 20.209606986899562,
+ "grad_norm": 0.016998112201690674,
+ "learning_rate": 0.0006,
+ "loss": 4.481280326843262,
+ "step": 1455
+ },
+ {
+ "epoch": 20.2235807860262,
+ "grad_norm": 0.017199190333485603,
+ "learning_rate": 0.0006,
+ "loss": 4.496712684631348,
+ "step": 1456
+ },
+ {
+ "epoch": 20.23755458515284,
+ "grad_norm": 0.016833679750561714,
+ "learning_rate": 0.0006,
+ "loss": 4.40736722946167,
+ "step": 1457
+ },
+ {
+ "epoch": 20.251528384279474,
+ "grad_norm": 0.017546730116009712,
+ "learning_rate": 0.0006,
+ "loss": 4.4915266036987305,
+ "step": 1458
+ },
+ {
+ "epoch": 20.265502183406113,
+ "grad_norm": 0.016595078632235527,
+ "learning_rate": 0.0006,
+ "loss": 4.460815906524658,
+ "step": 1459
+ },
+ {
+ "epoch": 20.27947598253275,
+ "grad_norm": 0.017728568986058235,
+ "learning_rate": 0.0006,
+ "loss": 4.346048831939697,
+ "step": 1460
+ },
+ {
+ "epoch": 20.29344978165939,
+ "grad_norm": 0.01849197782576084,
+ "learning_rate": 0.0006,
+ "loss": 4.399681091308594,
+ "step": 1461
+ },
+ {
+ "epoch": 20.307423580786025,
+ "grad_norm": 0.018336588516831398,
+ "learning_rate": 0.0006,
+ "loss": 4.414183616638184,
+ "step": 1462
+ },
+ {
+ "epoch": 20.321397379912664,
+ "grad_norm": 0.018086519092321396,
+ "learning_rate": 0.0006,
+ "loss": 4.346586227416992,
+ "step": 1463
+ },
+ {
+ "epoch": 20.335371179039303,
+ "grad_norm": 0.019697299227118492,
+ "learning_rate": 0.0006,
+ "loss": 4.4420952796936035,
+ "step": 1464
+ },
+ {
+ "epoch": 20.349344978165938,
+ "grad_norm": 0.02198074571788311,
+ "learning_rate": 0.0006,
+ "loss": 4.588729381561279,
+ "step": 1465
+ },
+ {
+ "epoch": 20.363318777292577,
+ "grad_norm": 0.022813329473137856,
+ "learning_rate": 0.0006,
+ "loss": 4.49729585647583,
+ "step": 1466
+ },
+ {
+ "epoch": 20.377292576419215,
+ "grad_norm": 0.022006649523973465,
+ "learning_rate": 0.0006,
+ "loss": 4.475641250610352,
+ "step": 1467
+ },
+ {
+ "epoch": 20.39126637554585,
+ "grad_norm": 0.023188438266515732,
+ "learning_rate": 0.0006,
+ "loss": 4.404322624206543,
+ "step": 1468
+ },
+ {
+ "epoch": 20.40524017467249,
+ "grad_norm": 0.022226877510547638,
+ "learning_rate": 0.0006,
+ "loss": 4.50904655456543,
+ "step": 1469
+ },
+ {
+ "epoch": 20.419213973799128,
+ "grad_norm": 0.021082637831568718,
+ "learning_rate": 0.0006,
+ "loss": 4.339011192321777,
+ "step": 1470
+ },
+ {
+ "epoch": 20.433187772925763,
+ "grad_norm": 0.02013542130589485,
+ "learning_rate": 0.0006,
+ "loss": 4.516942024230957,
+ "step": 1471
+ },
+ {
+ "epoch": 20.4471615720524,
+ "grad_norm": 0.01898968778550625,
+ "learning_rate": 0.0006,
+ "loss": 4.315804481506348,
+ "step": 1472
+ },
+ {
+ "epoch": 20.46113537117904,
+ "grad_norm": 0.019450880587100983,
+ "learning_rate": 0.0006,
+ "loss": 4.4794921875,
+ "step": 1473
+ },
+ {
+ "epoch": 20.475109170305675,
+ "grad_norm": 0.018577802926301956,
+ "learning_rate": 0.0006,
+ "loss": 4.422323703765869,
+ "step": 1474
+ },
+ {
+ "epoch": 20.489082969432314,
+ "grad_norm": 0.01778412237763405,
+ "learning_rate": 0.0006,
+ "loss": 4.429789066314697,
+ "step": 1475
+ },
+ {
+ "epoch": 20.503056768558952,
+ "grad_norm": 0.017134476453065872,
+ "learning_rate": 0.0006,
+ "loss": 4.416459083557129,
+ "step": 1476
+ },
+ {
+ "epoch": 20.51703056768559,
+ "grad_norm": 0.017153557389974594,
+ "learning_rate": 0.0006,
+ "loss": 4.446170806884766,
+ "step": 1477
+ },
+ {
+ "epoch": 20.531004366812226,
+ "grad_norm": 0.014998532831668854,
+ "learning_rate": 0.0006,
+ "loss": 4.419447898864746,
+ "step": 1478
+ },
+ {
+ "epoch": 20.544978165938865,
+ "grad_norm": 0.015877850353717804,
+ "learning_rate": 0.0006,
+ "loss": 4.312338829040527,
+ "step": 1479
+ },
+ {
+ "epoch": 20.558951965065503,
+ "grad_norm": 0.01569489948451519,
+ "learning_rate": 0.0006,
+ "loss": 4.572905540466309,
+ "step": 1480
+ },
+ {
+ "epoch": 20.57292576419214,
+ "grad_norm": 0.01591474376618862,
+ "learning_rate": 0.0006,
+ "loss": 4.414907455444336,
+ "step": 1481
+ },
+ {
+ "epoch": 20.586899563318777,
+ "grad_norm": 0.015395903028547764,
+ "learning_rate": 0.0006,
+ "loss": 4.549131393432617,
+ "step": 1482
+ },
+ {
+ "epoch": 20.600873362445416,
+ "grad_norm": 0.015715358778834343,
+ "learning_rate": 0.0006,
+ "loss": 4.530194282531738,
+ "step": 1483
+ },
+ {
+ "epoch": 20.61484716157205,
+ "grad_norm": 0.016042975708842278,
+ "learning_rate": 0.0006,
+ "loss": 4.322149276733398,
+ "step": 1484
+ },
+ {
+ "epoch": 20.62882096069869,
+ "grad_norm": 0.01664186641573906,
+ "learning_rate": 0.0006,
+ "loss": 4.424866676330566,
+ "step": 1485
+ },
+ {
+ "epoch": 20.64279475982533,
+ "grad_norm": 0.01631534844636917,
+ "learning_rate": 0.0006,
+ "loss": 4.441752910614014,
+ "step": 1486
+ },
+ {
+ "epoch": 20.656768558951963,
+ "grad_norm": 0.014057058840990067,
+ "learning_rate": 0.0006,
+ "loss": 4.399982452392578,
+ "step": 1487
+ },
+ {
+ "epoch": 20.670742358078602,
+ "grad_norm": 0.015076273120939732,
+ "learning_rate": 0.0006,
+ "loss": 4.441324234008789,
+ "step": 1488
+ },
+ {
+ "epoch": 20.68471615720524,
+ "grad_norm": 0.014138277620077133,
+ "learning_rate": 0.0006,
+ "loss": 4.4312543869018555,
+ "step": 1489
+ },
+ {
+ "epoch": 20.69868995633188,
+ "grad_norm": 0.013762440532445908,
+ "learning_rate": 0.0006,
+ "loss": 4.387807846069336,
+ "step": 1490
+ },
+ {
+ "epoch": 20.712663755458514,
+ "grad_norm": 0.015058411285281181,
+ "learning_rate": 0.0006,
+ "loss": 4.594747543334961,
+ "step": 1491
+ },
+ {
+ "epoch": 20.726637554585153,
+ "grad_norm": 0.013827123679220676,
+ "learning_rate": 0.0006,
+ "loss": 4.434604167938232,
+ "step": 1492
+ },
+ {
+ "epoch": 20.74061135371179,
+ "grad_norm": 0.015315295197069645,
+ "learning_rate": 0.0006,
+ "loss": 4.588066101074219,
+ "step": 1493
+ },
+ {
+ "epoch": 20.754585152838427,
+ "grad_norm": 0.018678732216358185,
+ "learning_rate": 0.0006,
+ "loss": 4.3514204025268555,
+ "step": 1494
+ },
+ {
+ "epoch": 20.768558951965066,
+ "grad_norm": 0.019521350041031837,
+ "learning_rate": 0.0006,
+ "loss": 4.437658786773682,
+ "step": 1495
+ },
+ {
+ "epoch": 20.782532751091704,
+ "grad_norm": 0.018363403156399727,
+ "learning_rate": 0.0006,
+ "loss": 4.495002269744873,
+ "step": 1496
+ },
+ {
+ "epoch": 20.79650655021834,
+ "grad_norm": 0.016830725595355034,
+ "learning_rate": 0.0006,
+ "loss": 4.402471542358398,
+ "step": 1497
+ },
+ {
+ "epoch": 20.810480349344978,
+ "grad_norm": 0.01722072623670101,
+ "learning_rate": 0.0006,
+ "loss": 4.437653541564941,
+ "step": 1498
+ },
+ {
+ "epoch": 20.824454148471617,
+ "grad_norm": 0.017459599301218987,
+ "learning_rate": 0.0006,
+ "loss": 4.366603851318359,
+ "step": 1499
+ },
+ {
+ "epoch": 20.83842794759825,
+ "grad_norm": 0.018902219831943512,
+ "learning_rate": 0.0006,
+ "loss": 4.501561164855957,
+ "step": 1500
+ },
+ {
+ "epoch": 20.85240174672489,
+ "grad_norm": 0.020814096555113792,
+ "learning_rate": 0.0006,
+ "loss": 4.535295486450195,
+ "step": 1501
+ },
+ {
+ "epoch": 20.86637554585153,
+ "grad_norm": 0.018990976735949516,
+ "learning_rate": 0.0006,
+ "loss": 4.333580017089844,
+ "step": 1502
+ },
+ {
+ "epoch": 20.880349344978168,
+ "grad_norm": 0.01837880350649357,
+ "learning_rate": 0.0006,
+ "loss": 4.313388824462891,
+ "step": 1503
+ },
+ {
+ "epoch": 20.894323144104803,
+ "grad_norm": 0.019820360466837883,
+ "learning_rate": 0.0006,
+ "loss": 4.388157844543457,
+ "step": 1504
+ },
+ {
+ "epoch": 20.90829694323144,
+ "grad_norm": 0.020181143656373024,
+ "learning_rate": 0.0006,
+ "loss": 4.572224140167236,
+ "step": 1505
+ },
+ {
+ "epoch": 20.92227074235808,
+ "grad_norm": 0.021206077188253403,
+ "learning_rate": 0.0006,
+ "loss": 4.3127970695495605,
+ "step": 1506
+ },
+ {
+ "epoch": 20.936244541484715,
+ "grad_norm": 0.017962060868740082,
+ "learning_rate": 0.0006,
+ "loss": 4.384734630584717,
+ "step": 1507
+ },
+ {
+ "epoch": 20.950218340611354,
+ "grad_norm": 0.01737220585346222,
+ "learning_rate": 0.0006,
+ "loss": 4.4075751304626465,
+ "step": 1508
+ },
+ {
+ "epoch": 20.964192139737992,
+ "grad_norm": 0.01839268207550049,
+ "learning_rate": 0.0006,
+ "loss": 4.553779125213623,
+ "step": 1509
+ },
+ {
+ "epoch": 20.978165938864628,
+ "grad_norm": 0.020102640613913536,
+ "learning_rate": 0.0006,
+ "loss": 4.399398326873779,
+ "step": 1510
+ },
+ {
+ "epoch": 20.992139737991266,
+ "grad_norm": 0.018831463530659676,
+ "learning_rate": 0.0006,
+ "loss": 4.358499526977539,
+ "step": 1511
+ },
+ {
+ "epoch": 21.0,
+ "grad_norm": 0.018254734575748444,
+ "learning_rate": 0.0006,
+ "loss": 4.384119510650635,
+ "step": 1512
+ },
+ {
+ "epoch": 21.0,
+ "eval_loss": 4.7524237632751465,
+ "eval_runtime": 57.8812,
+ "eval_samples_per_second": 42.19,
+ "eval_steps_per_second": 1.33,
+ "step": 1512
+ },
+ {
+ "epoch": 21.01397379912664,
+ "grad_norm": 0.019728587940335274,
+ "learning_rate": 0.0006,
+ "loss": 4.32363224029541,
+ "step": 1513
+ },
+ {
+ "epoch": 21.027947598253274,
+ "grad_norm": 0.02216625027358532,
+ "learning_rate": 0.0006,
+ "loss": 4.436988353729248,
+ "step": 1514
+ },
+ {
+ "epoch": 21.041921397379912,
+ "grad_norm": 0.02227606810629368,
+ "learning_rate": 0.0006,
+ "loss": 4.406063556671143,
+ "step": 1515
+ },
+ {
+ "epoch": 21.05589519650655,
+ "grad_norm": 0.020187009125947952,
+ "learning_rate": 0.0006,
+ "loss": 4.324402332305908,
+ "step": 1516
+ },
+ {
+ "epoch": 21.069868995633186,
+ "grad_norm": 0.02095157466828823,
+ "learning_rate": 0.0006,
+ "loss": 4.456544876098633,
+ "step": 1517
+ },
+ {
+ "epoch": 21.083842794759825,
+ "grad_norm": 0.02146013453602791,
+ "learning_rate": 0.0006,
+ "loss": 4.368387222290039,
+ "step": 1518
+ },
+ {
+ "epoch": 21.097816593886463,
+ "grad_norm": 0.02162035182118416,
+ "learning_rate": 0.0006,
+ "loss": 4.48838996887207,
+ "step": 1519
+ },
+ {
+ "epoch": 21.111790393013102,
+ "grad_norm": 0.021745676174759865,
+ "learning_rate": 0.0006,
+ "loss": 4.30778694152832,
+ "step": 1520
+ },
+ {
+ "epoch": 21.125764192139737,
+ "grad_norm": 0.02173588052392006,
+ "learning_rate": 0.0006,
+ "loss": 4.433149814605713,
+ "step": 1521
+ },
+ {
+ "epoch": 21.139737991266376,
+ "grad_norm": 0.020950648933649063,
+ "learning_rate": 0.0006,
+ "loss": 4.520089149475098,
+ "step": 1522
+ },
+ {
+ "epoch": 21.153711790393015,
+ "grad_norm": 0.020684752613306046,
+ "learning_rate": 0.0006,
+ "loss": 4.372437477111816,
+ "step": 1523
+ },
+ {
+ "epoch": 21.16768558951965,
+ "grad_norm": 0.019472869113087654,
+ "learning_rate": 0.0006,
+ "loss": 4.437015533447266,
+ "step": 1524
+ },
+ {
+ "epoch": 21.18165938864629,
+ "grad_norm": 0.01846962980926037,
+ "learning_rate": 0.0006,
+ "loss": 4.449531555175781,
+ "step": 1525
+ },
+ {
+ "epoch": 21.195633187772927,
+ "grad_norm": 0.017377803102135658,
+ "learning_rate": 0.0006,
+ "loss": 4.458279132843018,
+ "step": 1526
+ },
+ {
+ "epoch": 21.209606986899562,
+ "grad_norm": 0.016002200543880463,
+ "learning_rate": 0.0006,
+ "loss": 4.266297817230225,
+ "step": 1527
+ },
+ {
+ "epoch": 21.2235807860262,
+ "grad_norm": 0.015575903467833996,
+ "learning_rate": 0.0006,
+ "loss": 4.342632293701172,
+ "step": 1528
+ },
+ {
+ "epoch": 21.23755458515284,
+ "grad_norm": 0.015634698793292046,
+ "learning_rate": 0.0006,
+ "loss": 4.338623523712158,
+ "step": 1529
+ },
+ {
+ "epoch": 21.251528384279474,
+ "grad_norm": 0.01449581328779459,
+ "learning_rate": 0.0006,
+ "loss": 4.354330062866211,
+ "step": 1530
+ },
+ {
+ "epoch": 21.265502183406113,
+ "grad_norm": 0.01487616915255785,
+ "learning_rate": 0.0006,
+ "loss": 4.486835479736328,
+ "step": 1531
+ },
+ {
+ "epoch": 21.27947598253275,
+ "grad_norm": 0.016346555203199387,
+ "learning_rate": 0.0006,
+ "loss": 4.407098770141602,
+ "step": 1532
+ },
+ {
+ "epoch": 21.29344978165939,
+ "grad_norm": 0.01682182401418686,
+ "learning_rate": 0.0006,
+ "loss": 4.436809539794922,
+ "step": 1533
+ },
+ {
+ "epoch": 21.307423580786025,
+ "grad_norm": 0.017704004421830177,
+ "learning_rate": 0.0006,
+ "loss": 4.3854780197143555,
+ "step": 1534
+ },
+ {
+ "epoch": 21.321397379912664,
+ "grad_norm": 0.018437420949339867,
+ "learning_rate": 0.0006,
+ "loss": 4.413283824920654,
+ "step": 1535
+ },
+ {
+ "epoch": 21.335371179039303,
+ "grad_norm": 0.016911081969738007,
+ "learning_rate": 0.0006,
+ "loss": 4.357306003570557,
+ "step": 1536
+ },
+ {
+ "epoch": 21.349344978165938,
+ "grad_norm": 0.01642223633825779,
+ "learning_rate": 0.0006,
+ "loss": 4.422322750091553,
+ "step": 1537
+ },
+ {
+ "epoch": 21.363318777292577,
+ "grad_norm": 0.01780877821147442,
+ "learning_rate": 0.0006,
+ "loss": 4.387620449066162,
+ "step": 1538
+ },
+ {
+ "epoch": 21.377292576419215,
+ "grad_norm": 0.01653975248336792,
+ "learning_rate": 0.0006,
+ "loss": 4.327267646789551,
+ "step": 1539
+ },
+ {
+ "epoch": 21.39126637554585,
+ "grad_norm": 0.01572156324982643,
+ "learning_rate": 0.0006,
+ "loss": 4.475979804992676,
+ "step": 1540
+ },
+ {
+ "epoch": 21.40524017467249,
+ "grad_norm": 0.017156066372990608,
+ "learning_rate": 0.0006,
+ "loss": 4.4011125564575195,
+ "step": 1541
+ },
+ {
+ "epoch": 21.419213973799128,
+ "grad_norm": 0.01788829080760479,
+ "learning_rate": 0.0006,
+ "loss": 4.4174089431762695,
+ "step": 1542
+ },
+ {
+ "epoch": 21.433187772925763,
+ "grad_norm": 0.01728985086083412,
+ "learning_rate": 0.0006,
+ "loss": 4.429244041442871,
+ "step": 1543
+ },
+ {
+ "epoch": 21.4471615720524,
+ "grad_norm": 0.016534466296434402,
+ "learning_rate": 0.0006,
+ "loss": 4.326068878173828,
+ "step": 1544
+ },
+ {
+ "epoch": 21.46113537117904,
+ "grad_norm": 0.016582584008574486,
+ "learning_rate": 0.0006,
+ "loss": 4.2975969314575195,
+ "step": 1545
+ },
+ {
+ "epoch": 21.475109170305675,
+ "grad_norm": 0.018034178763628006,
+ "learning_rate": 0.0006,
+ "loss": 4.501319408416748,
+ "step": 1546
+ },
+ {
+ "epoch": 21.489082969432314,
+ "grad_norm": 0.017827702686190605,
+ "learning_rate": 0.0006,
+ "loss": 4.502284526824951,
+ "step": 1547
+ },
+ {
+ "epoch": 21.503056768558952,
+ "grad_norm": 0.017182299867272377,
+ "learning_rate": 0.0006,
+ "loss": 4.380828857421875,
+ "step": 1548
+ },
+ {
+ "epoch": 21.51703056768559,
+ "grad_norm": 0.019101588055491447,
+ "learning_rate": 0.0006,
+ "loss": 4.31564998626709,
+ "step": 1549
+ },
+ {
+ "epoch": 21.531004366812226,
+ "grad_norm": 0.0204677302390337,
+ "learning_rate": 0.0006,
+ "loss": 4.3840131759643555,
+ "step": 1550
+ },
+ {
+ "epoch": 21.544978165938865,
+ "grad_norm": 0.0188993439078331,
+ "learning_rate": 0.0006,
+ "loss": 4.4257731437683105,
+ "step": 1551
+ },
+ {
+ "epoch": 21.558951965065503,
+ "grad_norm": 0.01730230636894703,
+ "learning_rate": 0.0006,
+ "loss": 4.445132255554199,
+ "step": 1552
+ },
+ {
+ "epoch": 21.57292576419214,
+ "grad_norm": 0.01801574043929577,
+ "learning_rate": 0.0006,
+ "loss": 4.427064895629883,
+ "step": 1553
+ },
+ {
+ "epoch": 21.586899563318777,
+ "grad_norm": 0.018417565152049065,
+ "learning_rate": 0.0006,
+ "loss": 4.319595813751221,
+ "step": 1554
+ },
+ {
+ "epoch": 21.600873362445416,
+ "grad_norm": 0.01683160290122032,
+ "learning_rate": 0.0006,
+ "loss": 4.349218368530273,
+ "step": 1555
+ },
+ {
+ "epoch": 21.61484716157205,
+ "grad_norm": 0.016973106190562248,
+ "learning_rate": 0.0006,
+ "loss": 4.375990390777588,
+ "step": 1556
+ },
+ {
+ "epoch": 21.62882096069869,
+ "grad_norm": 0.015688583254814148,
+ "learning_rate": 0.0006,
+ "loss": 4.408170223236084,
+ "step": 1557
+ },
+ {
+ "epoch": 21.64279475982533,
+ "grad_norm": 0.01573043316602707,
+ "learning_rate": 0.0006,
+ "loss": 4.416255950927734,
+ "step": 1558
+ },
+ {
+ "epoch": 21.656768558951963,
+ "grad_norm": 0.01744293048977852,
+ "learning_rate": 0.0006,
+ "loss": 4.477141380310059,
+ "step": 1559
+ },
+ {
+ "epoch": 21.670742358078602,
+ "grad_norm": 0.017503513023257256,
+ "learning_rate": 0.0006,
+ "loss": 4.3367486000061035,
+ "step": 1560
+ },
+ {
+ "epoch": 21.68471615720524,
+ "grad_norm": 0.016592150554060936,
+ "learning_rate": 0.0006,
+ "loss": 4.314785480499268,
+ "step": 1561
+ },
+ {
+ "epoch": 21.69868995633188,
+ "grad_norm": 0.015774084255099297,
+ "learning_rate": 0.0006,
+ "loss": 4.448209285736084,
+ "step": 1562
+ },
+ {
+ "epoch": 21.712663755458514,
+ "grad_norm": 0.015779150649905205,
+ "learning_rate": 0.0006,
+ "loss": 4.392690658569336,
+ "step": 1563
+ },
+ {
+ "epoch": 21.726637554585153,
+ "grad_norm": 0.015392648056149483,
+ "learning_rate": 0.0006,
+ "loss": 4.437741756439209,
+ "step": 1564
+ },
+ {
+ "epoch": 21.74061135371179,
+ "grad_norm": 0.01530836895108223,
+ "learning_rate": 0.0006,
+ "loss": 4.424595832824707,
+ "step": 1565
+ },
+ {
+ "epoch": 21.754585152838427,
+ "grad_norm": 0.015600779093801975,
+ "learning_rate": 0.0006,
+ "loss": 4.335417747497559,
+ "step": 1566
+ },
+ {
+ "epoch": 21.768558951965066,
+ "grad_norm": 0.014646518044173717,
+ "learning_rate": 0.0006,
+ "loss": 4.279081344604492,
+ "step": 1567
+ },
+ {
+ "epoch": 21.782532751091704,
+ "grad_norm": 0.014132864773273468,
+ "learning_rate": 0.0006,
+ "loss": 4.345950126647949,
+ "step": 1568
+ },
+ {
+ "epoch": 21.79650655021834,
+ "grad_norm": 0.01425724383443594,
+ "learning_rate": 0.0006,
+ "loss": 4.277588844299316,
+ "step": 1569
+ },
+ {
+ "epoch": 21.810480349344978,
+ "grad_norm": 0.015396572649478912,
+ "learning_rate": 0.0006,
+ "loss": 4.326196670532227,
+ "step": 1570
+ },
+ {
+ "epoch": 21.824454148471617,
+ "grad_norm": 0.01683669537305832,
+ "learning_rate": 0.0006,
+ "loss": 4.3294172286987305,
+ "step": 1571
+ },
+ {
+ "epoch": 21.83842794759825,
+ "grad_norm": 0.016365647315979004,
+ "learning_rate": 0.0006,
+ "loss": 4.523321628570557,
+ "step": 1572
+ },
+ {
+ "epoch": 21.85240174672489,
+ "grad_norm": 0.017142774537205696,
+ "learning_rate": 0.0006,
+ "loss": 4.475842475891113,
+ "step": 1573
+ },
+ {
+ "epoch": 21.86637554585153,
+ "grad_norm": 0.015950839966535568,
+ "learning_rate": 0.0006,
+ "loss": 4.397187232971191,
+ "step": 1574
+ },
+ {
+ "epoch": 21.880349344978168,
+ "grad_norm": 0.016180362552404404,
+ "learning_rate": 0.0006,
+ "loss": 4.414778709411621,
+ "step": 1575
+ },
+ {
+ "epoch": 21.894323144104803,
+ "grad_norm": 0.017615556716918945,
+ "learning_rate": 0.0006,
+ "loss": 4.348567962646484,
+ "step": 1576
+ },
+ {
+ "epoch": 21.90829694323144,
+ "grad_norm": 0.01896277628839016,
+ "learning_rate": 0.0006,
+ "loss": 4.431762218475342,
+ "step": 1577
+ },
+ {
+ "epoch": 21.92227074235808,
+ "grad_norm": 0.01990971714258194,
+ "learning_rate": 0.0006,
+ "loss": 4.431285858154297,
+ "step": 1578
+ },
+ {
+ "epoch": 21.936244541484715,
+ "grad_norm": 0.0201258547604084,
+ "learning_rate": 0.0006,
+ "loss": 4.309872627258301,
+ "step": 1579
+ },
+ {
+ "epoch": 21.950218340611354,
+ "grad_norm": 0.019620198756456375,
+ "learning_rate": 0.0006,
+ "loss": 4.484044551849365,
+ "step": 1580
+ },
+ {
+ "epoch": 21.964192139737992,
+ "grad_norm": 0.01771123707294464,
+ "learning_rate": 0.0006,
+ "loss": 4.368093013763428,
+ "step": 1581
+ },
+ {
+ "epoch": 21.978165938864628,
+ "grad_norm": 0.017501654103398323,
+ "learning_rate": 0.0006,
+ "loss": 4.372651100158691,
+ "step": 1582
+ },
+ {
+ "epoch": 21.992139737991266,
+ "grad_norm": 0.01693608984351158,
+ "learning_rate": 0.0006,
+ "loss": 4.379167556762695,
+ "step": 1583
+ },
+ {
+ "epoch": 22.0,
+ "grad_norm": 0.02002849616110325,
+ "learning_rate": 0.0006,
+ "loss": 4.329627990722656,
+ "step": 1584
+ },
+ {
+ "epoch": 22.0,
+ "eval_loss": 4.723696231842041,
+ "eval_runtime": 56.1758,
+ "eval_samples_per_second": 43.471,
+ "eval_steps_per_second": 1.371,
+ "step": 1584
+ },
+ {
+ "epoch": 22.01397379912664,
+ "grad_norm": 0.02030305378139019,
+ "learning_rate": 0.0006,
+ "loss": 4.355566024780273,
+ "step": 1585
+ },
+ {
+ "epoch": 22.027947598253274,
+ "grad_norm": 0.018638934940099716,
+ "learning_rate": 0.0006,
+ "loss": 4.314009189605713,
+ "step": 1586
+ },
+ {
+ "epoch": 22.041921397379912,
+ "grad_norm": 0.019917072728276253,
+ "learning_rate": 0.0006,
+ "loss": 4.367525577545166,
+ "step": 1587
+ },
+ {
+ "epoch": 22.05589519650655,
+ "grad_norm": 0.020785167813301086,
+ "learning_rate": 0.0006,
+ "loss": 4.375561714172363,
+ "step": 1588
+ },
+ {
+ "epoch": 22.069868995633186,
+ "grad_norm": 0.021214453503489494,
+ "learning_rate": 0.0006,
+ "loss": 4.418305397033691,
+ "step": 1589
+ },
+ {
+ "epoch": 22.083842794759825,
+ "grad_norm": 0.019051678478717804,
+ "learning_rate": 0.0006,
+ "loss": 4.215126991271973,
+ "step": 1590
+ },
+ {
+ "epoch": 22.097816593886463,
+ "grad_norm": 0.02323366142809391,
+ "learning_rate": 0.0006,
+ "loss": 4.306807994842529,
+ "step": 1591
+ },
+ {
+ "epoch": 22.111790393013102,
+ "grad_norm": 0.026186738163232803,
+ "learning_rate": 0.0006,
+ "loss": 4.408508777618408,
+ "step": 1592
+ },
+ {
+ "epoch": 22.125764192139737,
+ "grad_norm": 0.027361227199435234,
+ "learning_rate": 0.0006,
+ "loss": 4.329322814941406,
+ "step": 1593
+ },
+ {
+ "epoch": 22.139737991266376,
+ "grad_norm": 0.026564646512269974,
+ "learning_rate": 0.0006,
+ "loss": 4.346642017364502,
+ "step": 1594
+ },
+ {
+ "epoch": 22.153711790393015,
+ "grad_norm": 0.025073667988181114,
+ "learning_rate": 0.0006,
+ "loss": 4.3984198570251465,
+ "step": 1595
+ },
+ {
+ "epoch": 22.16768558951965,
+ "grad_norm": 0.02347794733941555,
+ "learning_rate": 0.0006,
+ "loss": 4.352538108825684,
+ "step": 1596
+ },
+ {
+ "epoch": 22.18165938864629,
+ "grad_norm": 0.02203468047082424,
+ "learning_rate": 0.0006,
+ "loss": 4.298148155212402,
+ "step": 1597
+ },
+ {
+ "epoch": 22.195633187772927,
+ "grad_norm": 0.020621957257390022,
+ "learning_rate": 0.0006,
+ "loss": 4.480309963226318,
+ "step": 1598
+ },
+ {
+ "epoch": 22.209606986899562,
+ "grad_norm": 0.023710055276751518,
+ "learning_rate": 0.0006,
+ "loss": 4.339317321777344,
+ "step": 1599
+ },
+ {
+ "epoch": 22.2235807860262,
+ "grad_norm": 0.0233482476323843,
+ "learning_rate": 0.0006,
+ "loss": 4.436896324157715,
+ "step": 1600
+ },
+ {
+ "epoch": 22.23755458515284,
+ "grad_norm": 0.020122403278946877,
+ "learning_rate": 0.0006,
+ "loss": 4.236002445220947,
+ "step": 1601
+ },
+ {
+ "epoch": 22.251528384279474,
+ "grad_norm": 0.01957266591489315,
+ "learning_rate": 0.0006,
+ "loss": 4.497328281402588,
+ "step": 1602
+ },
+ {
+ "epoch": 22.265502183406113,
+ "grad_norm": 0.018964895978569984,
+ "learning_rate": 0.0006,
+ "loss": 4.36360502243042,
+ "step": 1603
+ },
+ {
+ "epoch": 22.27947598253275,
+ "grad_norm": 0.018902769312262535,
+ "learning_rate": 0.0006,
+ "loss": 4.427093505859375,
+ "step": 1604
+ },
+ {
+ "epoch": 22.29344978165939,
+ "grad_norm": 0.018918119370937347,
+ "learning_rate": 0.0006,
+ "loss": 4.294950485229492,
+ "step": 1605
+ },
+ {
+ "epoch": 22.307423580786025,
+ "grad_norm": 0.019315775483846664,
+ "learning_rate": 0.0006,
+ "loss": 4.427952289581299,
+ "step": 1606
+ },
+ {
+ "epoch": 22.321397379912664,
+ "grad_norm": 0.018350522965192795,
+ "learning_rate": 0.0006,
+ "loss": 4.272945404052734,
+ "step": 1607
+ },
+ {
+ "epoch": 22.335371179039303,
+ "grad_norm": 0.01759534701704979,
+ "learning_rate": 0.0006,
+ "loss": 4.348023891448975,
+ "step": 1608
+ },
+ {
+ "epoch": 22.349344978165938,
+ "grad_norm": 0.01646057888865471,
+ "learning_rate": 0.0006,
+ "loss": 4.278518199920654,
+ "step": 1609
+ },
+ {
+ "epoch": 22.363318777292577,
+ "grad_norm": 0.015875060111284256,
+ "learning_rate": 0.0006,
+ "loss": 4.423527240753174,
+ "step": 1610
+ },
+ {
+ "epoch": 22.377292576419215,
+ "grad_norm": 0.01638529635965824,
+ "learning_rate": 0.0006,
+ "loss": 4.321723937988281,
+ "step": 1611
+ },
+ {
+ "epoch": 22.39126637554585,
+ "grad_norm": 0.016942061483860016,
+ "learning_rate": 0.0006,
+ "loss": 4.346164703369141,
+ "step": 1612
+ },
+ {
+ "epoch": 22.40524017467249,
+ "grad_norm": 0.015209621749818325,
+ "learning_rate": 0.0006,
+ "loss": 4.434340000152588,
+ "step": 1613
+ },
+ {
+ "epoch": 22.419213973799128,
+ "grad_norm": 0.014423009008169174,
+ "learning_rate": 0.0006,
+ "loss": 4.342560768127441,
+ "step": 1614
+ },
+ {
+ "epoch": 22.433187772925763,
+ "grad_norm": 0.014369525946676731,
+ "learning_rate": 0.0006,
+ "loss": 4.296586990356445,
+ "step": 1615
+ },
+ {
+ "epoch": 22.4471615720524,
+ "grad_norm": 0.01392540242522955,
+ "learning_rate": 0.0006,
+ "loss": 4.433176040649414,
+ "step": 1616
+ },
+ {
+ "epoch": 22.46113537117904,
+ "grad_norm": 0.013206899166107178,
+ "learning_rate": 0.0006,
+ "loss": 4.32962703704834,
+ "step": 1617
+ },
+ {
+ "epoch": 22.475109170305675,
+ "grad_norm": 0.013427077792584896,
+ "learning_rate": 0.0006,
+ "loss": 4.420437335968018,
+ "step": 1618
+ },
+ {
+ "epoch": 22.489082969432314,
+ "grad_norm": 0.014194196090102196,
+ "learning_rate": 0.0006,
+ "loss": 4.410478591918945,
+ "step": 1619
+ },
+ {
+ "epoch": 22.503056768558952,
+ "grad_norm": 0.01362364087253809,
+ "learning_rate": 0.0006,
+ "loss": 4.338932037353516,
+ "step": 1620
+ },
+ {
+ "epoch": 22.51703056768559,
+ "grad_norm": 0.014299589209258556,
+ "learning_rate": 0.0006,
+ "loss": 4.34089994430542,
+ "step": 1621
+ },
+ {
+ "epoch": 22.531004366812226,
+ "grad_norm": 0.014061295427381992,
+ "learning_rate": 0.0006,
+ "loss": 4.29201602935791,
+ "step": 1622
+ },
+ {
+ "epoch": 22.544978165938865,
+ "grad_norm": 0.013945229351520538,
+ "learning_rate": 0.0006,
+ "loss": 4.326052188873291,
+ "step": 1623
+ },
+ {
+ "epoch": 22.558951965065503,
+ "grad_norm": 0.01387068536132574,
+ "learning_rate": 0.0006,
+ "loss": 4.393521308898926,
+ "step": 1624
+ },
+ {
+ "epoch": 22.57292576419214,
+ "grad_norm": 0.013931898400187492,
+ "learning_rate": 0.0006,
+ "loss": 4.3739752769470215,
+ "step": 1625
+ },
+ {
+ "epoch": 22.586899563318777,
+ "grad_norm": 0.01497740764170885,
+ "learning_rate": 0.0006,
+ "loss": 4.3991618156433105,
+ "step": 1626
+ },
+ {
+ "epoch": 22.600873362445416,
+ "grad_norm": 0.015842905268073082,
+ "learning_rate": 0.0006,
+ "loss": 4.405340671539307,
+ "step": 1627
+ },
+ {
+ "epoch": 22.61484716157205,
+ "grad_norm": 0.01580633595585823,
+ "learning_rate": 0.0006,
+ "loss": 4.280102729797363,
+ "step": 1628
+ },
+ {
+ "epoch": 22.62882096069869,
+ "grad_norm": 0.016567382961511612,
+ "learning_rate": 0.0006,
+ "loss": 4.407657146453857,
+ "step": 1629
+ },
+ {
+ "epoch": 22.64279475982533,
+ "grad_norm": 0.01648622378706932,
+ "learning_rate": 0.0006,
+ "loss": 4.392590522766113,
+ "step": 1630
+ },
+ {
+ "epoch": 22.656768558951963,
+ "grad_norm": 0.017066307365894318,
+ "learning_rate": 0.0006,
+ "loss": 4.324862003326416,
+ "step": 1631
+ },
+ {
+ "epoch": 22.670742358078602,
+ "grad_norm": 0.01637062057852745,
+ "learning_rate": 0.0006,
+ "loss": 4.34598445892334,
+ "step": 1632
+ },
+ {
+ "epoch": 22.68471615720524,
+ "grad_norm": 0.018038960173726082,
+ "learning_rate": 0.0006,
+ "loss": 4.320711612701416,
+ "step": 1633
+ },
+ {
+ "epoch": 22.69868995633188,
+ "grad_norm": 0.018096428364515305,
+ "learning_rate": 0.0006,
+ "loss": 4.536527156829834,
+ "step": 1634
+ },
+ {
+ "epoch": 22.712663755458514,
+ "grad_norm": 0.017889291048049927,
+ "learning_rate": 0.0006,
+ "loss": 4.388131141662598,
+ "step": 1635
+ },
+ {
+ "epoch": 22.726637554585153,
+ "grad_norm": 0.018771521747112274,
+ "learning_rate": 0.0006,
+ "loss": 4.357668876647949,
+ "step": 1636
+ },
+ {
+ "epoch": 22.74061135371179,
+ "grad_norm": 0.02100873365998268,
+ "learning_rate": 0.0006,
+ "loss": 4.293972015380859,
+ "step": 1637
+ },
+ {
+ "epoch": 22.754585152838427,
+ "grad_norm": 0.017715785652399063,
+ "learning_rate": 0.0006,
+ "loss": 4.349849224090576,
+ "step": 1638
+ },
+ {
+ "epoch": 22.768558951965066,
+ "grad_norm": 0.015546333976089954,
+ "learning_rate": 0.0006,
+ "loss": 4.311728000640869,
+ "step": 1639
+ },
+ {
+ "epoch": 22.782532751091704,
+ "grad_norm": 0.01605990156531334,
+ "learning_rate": 0.0006,
+ "loss": 4.499123573303223,
+ "step": 1640
+ },
+ {
+ "epoch": 22.79650655021834,
+ "grad_norm": 0.016642142087221146,
+ "learning_rate": 0.0006,
+ "loss": 4.4319257736206055,
+ "step": 1641
+ },
+ {
+ "epoch": 22.810480349344978,
+ "grad_norm": 0.016120387241244316,
+ "learning_rate": 0.0006,
+ "loss": 4.4373650550842285,
+ "step": 1642
+ },
+ {
+ "epoch": 22.824454148471617,
+ "grad_norm": 0.015174890868365765,
+ "learning_rate": 0.0006,
+ "loss": 4.314082145690918,
+ "step": 1643
+ },
+ {
+ "epoch": 22.83842794759825,
+ "grad_norm": 0.01551714539527893,
+ "learning_rate": 0.0006,
+ "loss": 4.360363960266113,
+ "step": 1644
+ },
+ {
+ "epoch": 22.85240174672489,
+ "grad_norm": 0.014966742135584354,
+ "learning_rate": 0.0006,
+ "loss": 4.346253395080566,
+ "step": 1645
+ },
+ {
+ "epoch": 22.86637554585153,
+ "grad_norm": 0.014288028702139854,
+ "learning_rate": 0.0006,
+ "loss": 4.464366436004639,
+ "step": 1646
+ },
+ {
+ "epoch": 22.880349344978168,
+ "grad_norm": 0.015291682444512844,
+ "learning_rate": 0.0006,
+ "loss": 4.504101753234863,
+ "step": 1647
+ },
+ {
+ "epoch": 22.894323144104803,
+ "grad_norm": 0.014534426853060722,
+ "learning_rate": 0.0006,
+ "loss": 4.462172031402588,
+ "step": 1648
+ },
+ {
+ "epoch": 22.90829694323144,
+ "grad_norm": 0.014129819348454475,
+ "learning_rate": 0.0006,
+ "loss": 4.350126266479492,
+ "step": 1649
+ },
+ {
+ "epoch": 22.92227074235808,
+ "grad_norm": 0.014932668767869473,
+ "learning_rate": 0.0006,
+ "loss": 4.374392032623291,
+ "step": 1650
+ },
+ {
+ "epoch": 22.936244541484715,
+ "grad_norm": 0.014647853560745716,
+ "learning_rate": 0.0006,
+ "loss": 4.252264499664307,
+ "step": 1651
+ },
+ {
+ "epoch": 22.950218340611354,
+ "grad_norm": 0.013553360477089882,
+ "learning_rate": 0.0006,
+ "loss": 4.379729270935059,
+ "step": 1652
+ },
+ {
+ "epoch": 22.964192139737992,
+ "grad_norm": 0.014606194570660591,
+ "learning_rate": 0.0006,
+ "loss": 4.363372802734375,
+ "step": 1653
+ },
+ {
+ "epoch": 22.978165938864628,
+ "grad_norm": 0.015000631101429462,
+ "learning_rate": 0.0006,
+ "loss": 4.408048629760742,
+ "step": 1654
+ },
+ {
+ "epoch": 22.992139737991266,
+ "grad_norm": 0.015370816923677921,
+ "learning_rate": 0.0006,
+ "loss": 4.297451019287109,
+ "step": 1655
+ },
+ {
+ "epoch": 23.0,
+ "grad_norm": 0.01709144562482834,
+ "learning_rate": 0.0006,
+ "loss": 4.294776916503906,
+ "step": 1656
+ },
+ {
+ "epoch": 23.0,
+ "eval_loss": 4.678618907928467,
+ "eval_runtime": 56.1668,
+ "eval_samples_per_second": 43.478,
+ "eval_steps_per_second": 1.371,
+ "step": 1656
+ },
+ {
+ "epoch": 23.01397379912664,
+ "grad_norm": 0.01727290451526642,
+ "learning_rate": 0.0006,
+ "loss": 4.311701774597168,
+ "step": 1657
+ },
+ {
+ "epoch": 23.027947598253274,
+ "grad_norm": 0.020645759999752045,
+ "learning_rate": 0.0006,
+ "loss": 4.429897308349609,
+ "step": 1658
+ },
+ {
+ "epoch": 23.041921397379912,
+ "grad_norm": 0.02204226516187191,
+ "learning_rate": 0.0006,
+ "loss": 4.365151882171631,
+ "step": 1659
+ },
+ {
+ "epoch": 23.05589519650655,
+ "grad_norm": 0.023069269955158234,
+ "learning_rate": 0.0006,
+ "loss": 4.3710832595825195,
+ "step": 1660
+ },
+ {
+ "epoch": 23.069868995633186,
+ "grad_norm": 0.024151179939508438,
+ "learning_rate": 0.0006,
+ "loss": 4.403108596801758,
+ "step": 1661
+ },
+ {
+ "epoch": 23.083842794759825,
+ "grad_norm": 0.02511233650147915,
+ "learning_rate": 0.0006,
+ "loss": 4.414128303527832,
+ "step": 1662
+ },
+ {
+ "epoch": 23.097816593886463,
+ "grad_norm": 0.025802113115787506,
+ "learning_rate": 0.0006,
+ "loss": 4.388265609741211,
+ "step": 1663
+ },
+ {
+ "epoch": 23.111790393013102,
+ "grad_norm": 0.02377348765730858,
+ "learning_rate": 0.0006,
+ "loss": 4.357505798339844,
+ "step": 1664
+ },
+ {
+ "epoch": 23.125764192139737,
+ "grad_norm": 0.020876651629805565,
+ "learning_rate": 0.0006,
+ "loss": 4.23043966293335,
+ "step": 1665
+ },
+ {
+ "epoch": 23.139737991266376,
+ "grad_norm": 0.01959727331995964,
+ "learning_rate": 0.0006,
+ "loss": 4.280970096588135,
+ "step": 1666
+ },
+ {
+ "epoch": 23.153711790393015,
+ "grad_norm": 0.018678616732358932,
+ "learning_rate": 0.0006,
+ "loss": 4.346716403961182,
+ "step": 1667
+ },
+ {
+ "epoch": 23.16768558951965,
+ "grad_norm": 0.02108944021165371,
+ "learning_rate": 0.0006,
+ "loss": 4.277904510498047,
+ "step": 1668
+ },
+ {
+ "epoch": 23.18165938864629,
+ "grad_norm": 0.020221566781401634,
+ "learning_rate": 0.0006,
+ "loss": 4.367693901062012,
+ "step": 1669
+ },
+ {
+ "epoch": 23.195633187772927,
+ "grad_norm": 0.019599711522459984,
+ "learning_rate": 0.0006,
+ "loss": 4.257092475891113,
+ "step": 1670
+ },
+ {
+ "epoch": 23.209606986899562,
+ "grad_norm": 0.019833361729979515,
+ "learning_rate": 0.0006,
+ "loss": 4.540737628936768,
+ "step": 1671
+ },
+ {
+ "epoch": 23.2235807860262,
+ "grad_norm": 0.020352095365524292,
+ "learning_rate": 0.0006,
+ "loss": 4.360923767089844,
+ "step": 1672
+ },
+ {
+ "epoch": 23.23755458515284,
+ "grad_norm": 0.01939748041331768,
+ "learning_rate": 0.0006,
+ "loss": 4.393110275268555,
+ "step": 1673
+ },
+ {
+ "epoch": 23.251528384279474,
+ "grad_norm": 0.01884354278445244,
+ "learning_rate": 0.0006,
+ "loss": 4.383938789367676,
+ "step": 1674
+ },
+ {
+ "epoch": 23.265502183406113,
+ "grad_norm": 0.02008470520377159,
+ "learning_rate": 0.0006,
+ "loss": 4.346344947814941,
+ "step": 1675
+ },
+ {
+ "epoch": 23.27947598253275,
+ "grad_norm": 0.019065195694565773,
+ "learning_rate": 0.0006,
+ "loss": 4.36611795425415,
+ "step": 1676
+ },
+ {
+ "epoch": 23.29344978165939,
+ "grad_norm": 0.01938541606068611,
+ "learning_rate": 0.0006,
+ "loss": 4.4009599685668945,
+ "step": 1677
+ },
+ {
+ "epoch": 23.307423580786025,
+ "grad_norm": 0.019044723361730576,
+ "learning_rate": 0.0006,
+ "loss": 4.340169906616211,
+ "step": 1678
+ },
+ {
+ "epoch": 23.321397379912664,
+ "grad_norm": 0.018850378692150116,
+ "learning_rate": 0.0006,
+ "loss": 4.37674617767334,
+ "step": 1679
+ },
+ {
+ "epoch": 23.335371179039303,
+ "grad_norm": 0.019076095893979073,
+ "learning_rate": 0.0006,
+ "loss": 4.226874351501465,
+ "step": 1680
+ },
+ {
+ "epoch": 23.349344978165938,
+ "grad_norm": 0.017543038353323936,
+ "learning_rate": 0.0006,
+ "loss": 4.337751388549805,
+ "step": 1681
+ },
+ {
+ "epoch": 23.363318777292577,
+ "grad_norm": 0.017170218750834465,
+ "learning_rate": 0.0006,
+ "loss": 4.282512664794922,
+ "step": 1682
+ },
+ {
+ "epoch": 23.377292576419215,
+ "grad_norm": 0.01574796624481678,
+ "learning_rate": 0.0006,
+ "loss": 4.30638313293457,
+ "step": 1683
+ },
+ {
+ "epoch": 23.39126637554585,
+ "grad_norm": 0.015160846523940563,
+ "learning_rate": 0.0006,
+ "loss": 4.336716175079346,
+ "step": 1684
+ },
+ {
+ "epoch": 23.40524017467249,
+ "grad_norm": 0.01534526702016592,
+ "learning_rate": 0.0006,
+ "loss": 4.2174787521362305,
+ "step": 1685
+ },
+ {
+ "epoch": 23.419213973799128,
+ "grad_norm": 0.015289999544620514,
+ "learning_rate": 0.0006,
+ "loss": 4.272546291351318,
+ "step": 1686
+ },
+ {
+ "epoch": 23.433187772925763,
+ "grad_norm": 0.014894185587763786,
+ "learning_rate": 0.0006,
+ "loss": 4.296045303344727,
+ "step": 1687
+ },
+ {
+ "epoch": 23.4471615720524,
+ "grad_norm": 0.014869502745568752,
+ "learning_rate": 0.0006,
+ "loss": 4.425461769104004,
+ "step": 1688
+ },
+ {
+ "epoch": 23.46113537117904,
+ "grad_norm": 0.01474483497440815,
+ "learning_rate": 0.0006,
+ "loss": 4.361867427825928,
+ "step": 1689
+ },
+ {
+ "epoch": 23.475109170305675,
+ "grad_norm": 0.015299245715141296,
+ "learning_rate": 0.0006,
+ "loss": 4.364523410797119,
+ "step": 1690
+ },
+ {
+ "epoch": 23.489082969432314,
+ "grad_norm": 0.015201340429484844,
+ "learning_rate": 0.0006,
+ "loss": 4.285182952880859,
+ "step": 1691
+ },
+ {
+ "epoch": 23.503056768558952,
+ "grad_norm": 0.016057293862104416,
+ "learning_rate": 0.0006,
+ "loss": 4.284873962402344,
+ "step": 1692
+ },
+ {
+ "epoch": 23.51703056768559,
+ "grad_norm": 0.015871062874794006,
+ "learning_rate": 0.0006,
+ "loss": 4.326470851898193,
+ "step": 1693
+ },
+ {
+ "epoch": 23.531004366812226,
+ "grad_norm": 0.01586288772523403,
+ "learning_rate": 0.0006,
+ "loss": 4.324883460998535,
+ "step": 1694
+ },
+ {
+ "epoch": 23.544978165938865,
+ "grad_norm": 0.016092827543616295,
+ "learning_rate": 0.0006,
+ "loss": 4.3258867263793945,
+ "step": 1695
+ },
+ {
+ "epoch": 23.558951965065503,
+ "grad_norm": 0.01582016795873642,
+ "learning_rate": 0.0006,
+ "loss": 4.358892440795898,
+ "step": 1696
+ },
+ {
+ "epoch": 23.57292576419214,
+ "grad_norm": 0.01688452623784542,
+ "learning_rate": 0.0006,
+ "loss": 4.280842304229736,
+ "step": 1697
+ },
+ {
+ "epoch": 23.586899563318777,
+ "grad_norm": 0.015865886583924294,
+ "learning_rate": 0.0006,
+ "loss": 4.245567321777344,
+ "step": 1698
+ },
+ {
+ "epoch": 23.600873362445416,
+ "grad_norm": 0.015303662046790123,
+ "learning_rate": 0.0006,
+ "loss": 4.383112907409668,
+ "step": 1699
+ },
+ {
+ "epoch": 23.61484716157205,
+ "grad_norm": 0.015649516135454178,
+ "learning_rate": 0.0006,
+ "loss": 4.519082069396973,
+ "step": 1700
+ },
+ {
+ "epoch": 23.62882096069869,
+ "grad_norm": 0.0167120099067688,
+ "learning_rate": 0.0006,
+ "loss": 4.447847366333008,
+ "step": 1701
+ },
+ {
+ "epoch": 23.64279475982533,
+ "grad_norm": 0.016368716955184937,
+ "learning_rate": 0.0006,
+ "loss": 4.434186935424805,
+ "step": 1702
+ },
+ {
+ "epoch": 23.656768558951963,
+ "grad_norm": 0.017379503697156906,
+ "learning_rate": 0.0006,
+ "loss": 4.317594528198242,
+ "step": 1703
+ },
+ {
+ "epoch": 23.670742358078602,
+ "grad_norm": 0.018573811277747154,
+ "learning_rate": 0.0006,
+ "loss": 4.205793857574463,
+ "step": 1704
+ },
+ {
+ "epoch": 23.68471615720524,
+ "grad_norm": 0.01749090477824211,
+ "learning_rate": 0.0006,
+ "loss": 4.3985700607299805,
+ "step": 1705
+ },
+ {
+ "epoch": 23.69868995633188,
+ "grad_norm": 0.0165668074041605,
+ "learning_rate": 0.0006,
+ "loss": 4.329226970672607,
+ "step": 1706
+ },
+ {
+ "epoch": 23.712663755458514,
+ "grad_norm": 0.016913846135139465,
+ "learning_rate": 0.0006,
+ "loss": 4.388485908508301,
+ "step": 1707
+ },
+ {
+ "epoch": 23.726637554585153,
+ "grad_norm": 0.016813859343528748,
+ "learning_rate": 0.0006,
+ "loss": 4.3395915031433105,
+ "step": 1708
+ },
+ {
+ "epoch": 23.74061135371179,
+ "grad_norm": 0.01676975190639496,
+ "learning_rate": 0.0006,
+ "loss": 4.254745960235596,
+ "step": 1709
+ },
+ {
+ "epoch": 23.754585152838427,
+ "grad_norm": 0.016370350494980812,
+ "learning_rate": 0.0006,
+ "loss": 4.2612223625183105,
+ "step": 1710
+ },
+ {
+ "epoch": 23.768558951965066,
+ "grad_norm": 0.01696198247373104,
+ "learning_rate": 0.0006,
+ "loss": 4.327524662017822,
+ "step": 1711
+ },
+ {
+ "epoch": 23.782532751091704,
+ "grad_norm": 0.0189108457416296,
+ "learning_rate": 0.0006,
+ "loss": 4.325839519500732,
+ "step": 1712
+ },
+ {
+ "epoch": 23.79650655021834,
+ "grad_norm": 0.01862351782619953,
+ "learning_rate": 0.0006,
+ "loss": 4.333662509918213,
+ "step": 1713
+ },
+ {
+ "epoch": 23.810480349344978,
+ "grad_norm": 0.01655014418065548,
+ "learning_rate": 0.0006,
+ "loss": 4.39305305480957,
+ "step": 1714
+ },
+ {
+ "epoch": 23.824454148471617,
+ "grad_norm": 0.015899376943707466,
+ "learning_rate": 0.0006,
+ "loss": 4.390053749084473,
+ "step": 1715
+ },
+ {
+ "epoch": 23.83842794759825,
+ "grad_norm": 0.016519056633114815,
+ "learning_rate": 0.0006,
+ "loss": 4.255903244018555,
+ "step": 1716
+ },
+ {
+ "epoch": 23.85240174672489,
+ "grad_norm": 0.0167539119720459,
+ "learning_rate": 0.0006,
+ "loss": 4.255727767944336,
+ "step": 1717
+ },
+ {
+ "epoch": 23.86637554585153,
+ "grad_norm": 0.0171353816986084,
+ "learning_rate": 0.0006,
+ "loss": 4.376740455627441,
+ "step": 1718
+ },
+ {
+ "epoch": 23.880349344978168,
+ "grad_norm": 0.016719117760658264,
+ "learning_rate": 0.0006,
+ "loss": 4.3376688957214355,
+ "step": 1719
+ },
+ {
+ "epoch": 23.894323144104803,
+ "grad_norm": 0.01561494916677475,
+ "learning_rate": 0.0006,
+ "loss": 4.3345746994018555,
+ "step": 1720
+ },
+ {
+ "epoch": 23.90829694323144,
+ "grad_norm": 0.016303174197673798,
+ "learning_rate": 0.0006,
+ "loss": 4.332371234893799,
+ "step": 1721
+ },
+ {
+ "epoch": 23.92227074235808,
+ "grad_norm": 0.016722865402698517,
+ "learning_rate": 0.0006,
+ "loss": 4.267125129699707,
+ "step": 1722
+ },
+ {
+ "epoch": 23.936244541484715,
+ "grad_norm": 0.017072124406695366,
+ "learning_rate": 0.0006,
+ "loss": 4.322751998901367,
+ "step": 1723
+ },
+ {
+ "epoch": 23.950218340611354,
+ "grad_norm": 0.01699932850897312,
+ "learning_rate": 0.0006,
+ "loss": 4.378422737121582,
+ "step": 1724
+ },
+ {
+ "epoch": 23.964192139737992,
+ "grad_norm": 0.01742720976471901,
+ "learning_rate": 0.0006,
+ "loss": 4.345083236694336,
+ "step": 1725
+ },
+ {
+ "epoch": 23.978165938864628,
+ "grad_norm": 0.018797501921653748,
+ "learning_rate": 0.0006,
+ "loss": 4.290443420410156,
+ "step": 1726
+ },
+ {
+ "epoch": 23.992139737991266,
+ "grad_norm": 0.018090683966875076,
+ "learning_rate": 0.0006,
+ "loss": 4.377580642700195,
+ "step": 1727
+ },
+ {
+ "epoch": 24.0,
+ "grad_norm": 0.017305459827184677,
+ "learning_rate": 0.0006,
+ "loss": 4.364426612854004,
+ "step": 1728
+ },
+ {
+ "epoch": 24.0,
+ "eval_loss": 4.647762298583984,
+ "eval_runtime": 56.5402,
+ "eval_samples_per_second": 43.191,
+ "eval_steps_per_second": 1.362,
+ "step": 1728
+ },
+ {
+ "epoch": 24.01397379912664,
+ "grad_norm": 0.016493607312440872,
+ "learning_rate": 0.0006,
+ "loss": 4.252255916595459,
+ "step": 1729
+ },
+ {
+ "epoch": 24.027947598253274,
+ "grad_norm": 0.016529450193047523,
+ "learning_rate": 0.0006,
+ "loss": 4.352746963500977,
+ "step": 1730
+ },
+ {
+ "epoch": 24.041921397379912,
+ "grad_norm": 0.016940191388130188,
+ "learning_rate": 0.0006,
+ "loss": 4.301870822906494,
+ "step": 1731
+ },
+ {
+ "epoch": 24.05589519650655,
+ "grad_norm": 0.01752558909356594,
+ "learning_rate": 0.0006,
+ "loss": 4.2762451171875,
+ "step": 1732
+ },
+ {
+ "epoch": 24.069868995633186,
+ "grad_norm": 0.017112884670495987,
+ "learning_rate": 0.0006,
+ "loss": 4.2659149169921875,
+ "step": 1733
+ },
+ {
+ "epoch": 24.083842794759825,
+ "grad_norm": 0.01900586113333702,
+ "learning_rate": 0.0006,
+ "loss": 4.306058883666992,
+ "step": 1734
+ },
+ {
+ "epoch": 24.097816593886463,
+ "grad_norm": 0.019447680562734604,
+ "learning_rate": 0.0006,
+ "loss": 4.450128555297852,
+ "step": 1735
+ },
+ {
+ "epoch": 24.111790393013102,
+ "grad_norm": 0.01948186755180359,
+ "learning_rate": 0.0006,
+ "loss": 4.202869415283203,
+ "step": 1736
+ },
+ {
+ "epoch": 24.125764192139737,
+ "grad_norm": 0.01716383546590805,
+ "learning_rate": 0.0006,
+ "loss": 4.318876266479492,
+ "step": 1737
+ },
+ {
+ "epoch": 24.139737991266376,
+ "grad_norm": 0.017049988731741905,
+ "learning_rate": 0.0006,
+ "loss": 4.237238883972168,
+ "step": 1738
+ },
+ {
+ "epoch": 24.153711790393015,
+ "grad_norm": 0.015169923193752766,
+ "learning_rate": 0.0006,
+ "loss": 4.289426803588867,
+ "step": 1739
+ },
+ {
+ "epoch": 24.16768558951965,
+ "grad_norm": 0.014760667458176613,
+ "learning_rate": 0.0006,
+ "loss": 4.356924057006836,
+ "step": 1740
+ },
+ {
+ "epoch": 24.18165938864629,
+ "grad_norm": 0.015127750113606453,
+ "learning_rate": 0.0006,
+ "loss": 4.142916679382324,
+ "step": 1741
+ },
+ {
+ "epoch": 24.195633187772927,
+ "grad_norm": 0.01549555640667677,
+ "learning_rate": 0.0006,
+ "loss": 4.345563888549805,
+ "step": 1742
+ },
+ {
+ "epoch": 24.209606986899562,
+ "grad_norm": 0.016393091529607773,
+ "learning_rate": 0.0006,
+ "loss": 4.417495250701904,
+ "step": 1743
+ },
+ {
+ "epoch": 24.2235807860262,
+ "grad_norm": 0.01785266026854515,
+ "learning_rate": 0.0006,
+ "loss": 4.3553547859191895,
+ "step": 1744
+ },
+ {
+ "epoch": 24.23755458515284,
+ "grad_norm": 0.017574617639183998,
+ "learning_rate": 0.0006,
+ "loss": 4.241024494171143,
+ "step": 1745
+ },
+ {
+ "epoch": 24.251528384279474,
+ "grad_norm": 0.019060306251049042,
+ "learning_rate": 0.0006,
+ "loss": 4.3685150146484375,
+ "step": 1746
+ },
+ {
+ "epoch": 24.265502183406113,
+ "grad_norm": 0.0214633010327816,
+ "learning_rate": 0.0006,
+ "loss": 4.323866844177246,
+ "step": 1747
+ },
+ {
+ "epoch": 24.27947598253275,
+ "grad_norm": 0.021055003628134727,
+ "learning_rate": 0.0006,
+ "loss": 4.278757095336914,
+ "step": 1748
+ },
+ {
+ "epoch": 24.29344978165939,
+ "grad_norm": 0.018618572503328323,
+ "learning_rate": 0.0006,
+ "loss": 4.28849983215332,
+ "step": 1749
+ },
+ {
+ "epoch": 24.307423580786025,
+ "grad_norm": 0.018666435033082962,
+ "learning_rate": 0.0006,
+ "loss": 4.368590354919434,
+ "step": 1750
+ },
+ {
+ "epoch": 24.321397379912664,
+ "grad_norm": 0.020626146346330643,
+ "learning_rate": 0.0006,
+ "loss": 4.2951836585998535,
+ "step": 1751
+ },
+ {
+ "epoch": 24.335371179039303,
+ "grad_norm": 0.01929759792983532,
+ "learning_rate": 0.0006,
+ "loss": 4.3664703369140625,
+ "step": 1752
+ },
+ {
+ "epoch": 24.349344978165938,
+ "grad_norm": 0.020956002175807953,
+ "learning_rate": 0.0006,
+ "loss": 4.345308303833008,
+ "step": 1753
+ },
+ {
+ "epoch": 24.363318777292577,
+ "grad_norm": 0.019227957352995872,
+ "learning_rate": 0.0006,
+ "loss": 4.349849224090576,
+ "step": 1754
+ },
+ {
+ "epoch": 24.377292576419215,
+ "grad_norm": 0.015970097854733467,
+ "learning_rate": 0.0006,
+ "loss": 4.223012447357178,
+ "step": 1755
+ },
+ {
+ "epoch": 24.39126637554585,
+ "grad_norm": 0.015951646491885185,
+ "learning_rate": 0.0006,
+ "loss": 4.236420631408691,
+ "step": 1756
+ },
+ {
+ "epoch": 24.40524017467249,
+ "grad_norm": 0.016032515093684196,
+ "learning_rate": 0.0006,
+ "loss": 4.226899147033691,
+ "step": 1757
+ },
+ {
+ "epoch": 24.419213973799128,
+ "grad_norm": 0.01573132909834385,
+ "learning_rate": 0.0006,
+ "loss": 4.315916061401367,
+ "step": 1758
+ },
+ {
+ "epoch": 24.433187772925763,
+ "grad_norm": 0.015367005951702595,
+ "learning_rate": 0.0006,
+ "loss": 4.358939170837402,
+ "step": 1759
+ },
+ {
+ "epoch": 24.4471615720524,
+ "grad_norm": 0.01669907011091709,
+ "learning_rate": 0.0006,
+ "loss": 4.232147216796875,
+ "step": 1760
+ },
+ {
+ "epoch": 24.46113537117904,
+ "grad_norm": 0.018708152696490288,
+ "learning_rate": 0.0006,
+ "loss": 4.329852104187012,
+ "step": 1761
+ },
+ {
+ "epoch": 24.475109170305675,
+ "grad_norm": 0.019035987555980682,
+ "learning_rate": 0.0006,
+ "loss": 4.260340213775635,
+ "step": 1762
+ },
+ {
+ "epoch": 24.489082969432314,
+ "grad_norm": 0.01832144521176815,
+ "learning_rate": 0.0006,
+ "loss": 4.359543800354004,
+ "step": 1763
+ },
+ {
+ "epoch": 24.503056768558952,
+ "grad_norm": 0.015313294716179371,
+ "learning_rate": 0.0006,
+ "loss": 4.279088020324707,
+ "step": 1764
+ },
+ {
+ "epoch": 24.51703056768559,
+ "grad_norm": 0.015803923830389977,
+ "learning_rate": 0.0006,
+ "loss": 4.230682849884033,
+ "step": 1765
+ },
+ {
+ "epoch": 24.531004366812226,
+ "grad_norm": 0.016748374328017235,
+ "learning_rate": 0.0006,
+ "loss": 4.349661827087402,
+ "step": 1766
+ },
+ {
+ "epoch": 24.544978165938865,
+ "grad_norm": 0.015732428058981895,
+ "learning_rate": 0.0006,
+ "loss": 4.372441291809082,
+ "step": 1767
+ },
+ {
+ "epoch": 24.558951965065503,
+ "grad_norm": 0.016564829275012016,
+ "learning_rate": 0.0006,
+ "loss": 4.323090553283691,
+ "step": 1768
+ },
+ {
+ "epoch": 24.57292576419214,
+ "grad_norm": 0.01603010483086109,
+ "learning_rate": 0.0006,
+ "loss": 4.347898006439209,
+ "step": 1769
+ },
+ {
+ "epoch": 24.586899563318777,
+ "grad_norm": 0.01674809865653515,
+ "learning_rate": 0.0006,
+ "loss": 4.269165515899658,
+ "step": 1770
+ },
+ {
+ "epoch": 24.600873362445416,
+ "grad_norm": 0.017333930358290672,
+ "learning_rate": 0.0006,
+ "loss": 4.353298187255859,
+ "step": 1771
+ },
+ {
+ "epoch": 24.61484716157205,
+ "grad_norm": 0.017519580200314522,
+ "learning_rate": 0.0006,
+ "loss": 4.265199661254883,
+ "step": 1772
+ },
+ {
+ "epoch": 24.62882096069869,
+ "grad_norm": 0.017675306648015976,
+ "learning_rate": 0.0006,
+ "loss": 4.322596549987793,
+ "step": 1773
+ },
+ {
+ "epoch": 24.64279475982533,
+ "grad_norm": 0.017166534438729286,
+ "learning_rate": 0.0006,
+ "loss": 4.413900375366211,
+ "step": 1774
+ },
+ {
+ "epoch": 24.656768558951963,
+ "grad_norm": 0.016362346708774567,
+ "learning_rate": 0.0006,
+ "loss": 4.307236671447754,
+ "step": 1775
+ },
+ {
+ "epoch": 24.670742358078602,
+ "grad_norm": 0.016086872667074203,
+ "learning_rate": 0.0006,
+ "loss": 4.235316276550293,
+ "step": 1776
+ },
+ {
+ "epoch": 24.68471615720524,
+ "grad_norm": 0.01689537614583969,
+ "learning_rate": 0.0006,
+ "loss": 4.390591621398926,
+ "step": 1777
+ },
+ {
+ "epoch": 24.69868995633188,
+ "grad_norm": 0.016836725175380707,
+ "learning_rate": 0.0006,
+ "loss": 4.2990031242370605,
+ "step": 1778
+ },
+ {
+ "epoch": 24.712663755458514,
+ "grad_norm": 0.015627335757017136,
+ "learning_rate": 0.0006,
+ "loss": 4.166745185852051,
+ "step": 1779
+ },
+ {
+ "epoch": 24.726637554585153,
+ "grad_norm": 0.015007016249001026,
+ "learning_rate": 0.0006,
+ "loss": 4.295273303985596,
+ "step": 1780
+ },
+ {
+ "epoch": 24.74061135371179,
+ "grad_norm": 0.016084209084510803,
+ "learning_rate": 0.0006,
+ "loss": 4.342424392700195,
+ "step": 1781
+ },
+ {
+ "epoch": 24.754585152838427,
+ "grad_norm": 0.01595073565840721,
+ "learning_rate": 0.0006,
+ "loss": 4.3060102462768555,
+ "step": 1782
+ },
+ {
+ "epoch": 24.768558951965066,
+ "grad_norm": 0.016113461926579475,
+ "learning_rate": 0.0006,
+ "loss": 4.267908096313477,
+ "step": 1783
+ },
+ {
+ "epoch": 24.782532751091704,
+ "grad_norm": 0.01564556173980236,
+ "learning_rate": 0.0006,
+ "loss": 4.324642658233643,
+ "step": 1784
+ },
+ {
+ "epoch": 24.79650655021834,
+ "grad_norm": 0.01575268618762493,
+ "learning_rate": 0.0006,
+ "loss": 4.328181743621826,
+ "step": 1785
+ },
+ {
+ "epoch": 24.810480349344978,
+ "grad_norm": 0.016830824315547943,
+ "learning_rate": 0.0006,
+ "loss": 4.247906684875488,
+ "step": 1786
+ },
+ {
+ "epoch": 24.824454148471617,
+ "grad_norm": 0.017989547923207283,
+ "learning_rate": 0.0006,
+ "loss": 4.228281021118164,
+ "step": 1787
+ },
+ {
+ "epoch": 24.83842794759825,
+ "grad_norm": 0.016783086583018303,
+ "learning_rate": 0.0006,
+ "loss": 4.2880964279174805,
+ "step": 1788
+ },
+ {
+ "epoch": 24.85240174672489,
+ "grad_norm": 0.016467561945319176,
+ "learning_rate": 0.0006,
+ "loss": 4.36118745803833,
+ "step": 1789
+ },
+ {
+ "epoch": 24.86637554585153,
+ "grad_norm": 0.01780523732304573,
+ "learning_rate": 0.0006,
+ "loss": 4.172321319580078,
+ "step": 1790
+ },
+ {
+ "epoch": 24.880349344978168,
+ "grad_norm": 0.017102031037211418,
+ "learning_rate": 0.0006,
+ "loss": 4.352826118469238,
+ "step": 1791
+ },
+ {
+ "epoch": 24.894323144104803,
+ "grad_norm": 0.016409218311309814,
+ "learning_rate": 0.0006,
+ "loss": 4.212162017822266,
+ "step": 1792
+ },
+ {
+ "epoch": 24.90829694323144,
+ "grad_norm": 0.018660584464669228,
+ "learning_rate": 0.0006,
+ "loss": 4.346684455871582,
+ "step": 1793
+ },
+ {
+ "epoch": 24.92227074235808,
+ "grad_norm": 0.017389440909028053,
+ "learning_rate": 0.0006,
+ "loss": 4.197913646697998,
+ "step": 1794
+ },
+ {
+ "epoch": 24.936244541484715,
+ "grad_norm": 0.01596485823392868,
+ "learning_rate": 0.0006,
+ "loss": 4.370617866516113,
+ "step": 1795
+ },
+ {
+ "epoch": 24.950218340611354,
+ "grad_norm": 0.01709776371717453,
+ "learning_rate": 0.0006,
+ "loss": 4.313847064971924,
+ "step": 1796
+ },
+ {
+ "epoch": 24.964192139737992,
+ "grad_norm": 0.016571134328842163,
+ "learning_rate": 0.0006,
+ "loss": 4.1895856857299805,
+ "step": 1797
+ },
+ {
+ "epoch": 24.978165938864628,
+ "grad_norm": 0.016851048916578293,
+ "learning_rate": 0.0006,
+ "loss": 4.350772857666016,
+ "step": 1798
+ },
+ {
+ "epoch": 24.992139737991266,
+ "grad_norm": 0.017007440328598022,
+ "learning_rate": 0.0006,
+ "loss": 4.399008750915527,
+ "step": 1799
+ },
+ {
+ "epoch": 25.0,
+ "grad_norm": 0.01791212148964405,
+ "learning_rate": 0.0006,
+ "loss": 4.306121826171875,
+ "step": 1800
+ },
+ {
+ "epoch": 25.0,
+ "eval_loss": 4.636417865753174,
+ "eval_runtime": 56.8296,
+ "eval_samples_per_second": 42.971,
+ "eval_steps_per_second": 1.355,
+ "step": 1800
+ },
+ {
+ "epoch": 25.01397379912664,
+ "grad_norm": 0.01729617640376091,
+ "learning_rate": 0.0006,
+ "loss": 4.257018089294434,
+ "step": 1801
+ },
+ {
+ "epoch": 25.027947598253274,
+ "grad_norm": 0.017466643825173378,
+ "learning_rate": 0.0006,
+ "loss": 4.307605266571045,
+ "step": 1802
+ },
+ {
+ "epoch": 25.041921397379912,
+ "grad_norm": 0.018614279106259346,
+ "learning_rate": 0.0006,
+ "loss": 4.360141277313232,
+ "step": 1803
+ },
+ {
+ "epoch": 25.05589519650655,
+ "grad_norm": 0.02010771445930004,
+ "learning_rate": 0.0006,
+ "loss": 4.196428298950195,
+ "step": 1804
+ },
+ {
+ "epoch": 25.069868995633186,
+ "grad_norm": 0.02090228535234928,
+ "learning_rate": 0.0006,
+ "loss": 4.288947105407715,
+ "step": 1805
+ },
+ {
+ "epoch": 25.083842794759825,
+ "grad_norm": 0.02050255797803402,
+ "learning_rate": 0.0006,
+ "loss": 4.259582042694092,
+ "step": 1806
+ },
+ {
+ "epoch": 25.097816593886463,
+ "grad_norm": 0.021844089031219482,
+ "learning_rate": 0.0006,
+ "loss": 4.405138969421387,
+ "step": 1807
+ },
+ {
+ "epoch": 25.111790393013102,
+ "grad_norm": 0.022287018597126007,
+ "learning_rate": 0.0006,
+ "loss": 4.333793640136719,
+ "step": 1808
+ },
+ {
+ "epoch": 25.125764192139737,
+ "grad_norm": 0.02288329415023327,
+ "learning_rate": 0.0006,
+ "loss": 4.3910980224609375,
+ "step": 1809
+ },
+ {
+ "epoch": 25.139737991266376,
+ "grad_norm": 0.02373199723660946,
+ "learning_rate": 0.0006,
+ "loss": 4.32921028137207,
+ "step": 1810
+ },
+ {
+ "epoch": 25.153711790393015,
+ "grad_norm": 0.024960234761238098,
+ "learning_rate": 0.0006,
+ "loss": 4.295609474182129,
+ "step": 1811
+ },
+ {
+ "epoch": 25.16768558951965,
+ "grad_norm": 0.026743892580270767,
+ "learning_rate": 0.0006,
+ "loss": 4.315308094024658,
+ "step": 1812
+ },
+ {
+ "epoch": 25.18165938864629,
+ "grad_norm": 0.022138185799121857,
+ "learning_rate": 0.0006,
+ "loss": 4.360208511352539,
+ "step": 1813
+ },
+ {
+ "epoch": 25.195633187772927,
+ "grad_norm": 0.022120574489235878,
+ "learning_rate": 0.0006,
+ "loss": 4.1973772048950195,
+ "step": 1814
+ },
+ {
+ "epoch": 25.209606986899562,
+ "grad_norm": 0.02230502851307392,
+ "learning_rate": 0.0006,
+ "loss": 4.285558700561523,
+ "step": 1815
+ },
+ {
+ "epoch": 25.2235807860262,
+ "grad_norm": 0.022301986813545227,
+ "learning_rate": 0.0006,
+ "loss": 4.238343238830566,
+ "step": 1816
+ },
+ {
+ "epoch": 25.23755458515284,
+ "grad_norm": 0.020050447434186935,
+ "learning_rate": 0.0006,
+ "loss": 4.298235893249512,
+ "step": 1817
+ },
+ {
+ "epoch": 25.251528384279474,
+ "grad_norm": 0.01987724006175995,
+ "learning_rate": 0.0006,
+ "loss": 4.301384449005127,
+ "step": 1818
+ },
+ {
+ "epoch": 25.265502183406113,
+ "grad_norm": 0.020440855994820595,
+ "learning_rate": 0.0006,
+ "loss": 4.269218444824219,
+ "step": 1819
+ },
+ {
+ "epoch": 25.27947598253275,
+ "grad_norm": 0.01891893707215786,
+ "learning_rate": 0.0006,
+ "loss": 4.2401957511901855,
+ "step": 1820
+ },
+ {
+ "epoch": 25.29344978165939,
+ "grad_norm": 0.01903688907623291,
+ "learning_rate": 0.0006,
+ "loss": 4.139670372009277,
+ "step": 1821
+ },
+ {
+ "epoch": 25.307423580786025,
+ "grad_norm": 0.018856938928365707,
+ "learning_rate": 0.0006,
+ "loss": 4.233214378356934,
+ "step": 1822
+ },
+ {
+ "epoch": 25.321397379912664,
+ "grad_norm": 0.017791323363780975,
+ "learning_rate": 0.0006,
+ "loss": 4.2342023849487305,
+ "step": 1823
+ },
+ {
+ "epoch": 25.335371179039303,
+ "grad_norm": 0.017304565757513046,
+ "learning_rate": 0.0006,
+ "loss": 4.355198860168457,
+ "step": 1824
+ },
+ {
+ "epoch": 25.349344978165938,
+ "grad_norm": 0.01576206088066101,
+ "learning_rate": 0.0006,
+ "loss": 4.114147186279297,
+ "step": 1825
+ },
+ {
+ "epoch": 25.363318777292577,
+ "grad_norm": 0.015105154365301132,
+ "learning_rate": 0.0006,
+ "loss": 4.313795566558838,
+ "step": 1826
+ },
+ {
+ "epoch": 25.377292576419215,
+ "grad_norm": 0.01563587784767151,
+ "learning_rate": 0.0006,
+ "loss": 4.308282852172852,
+ "step": 1827
+ },
+ {
+ "epoch": 25.39126637554585,
+ "grad_norm": 0.015980906784534454,
+ "learning_rate": 0.0006,
+ "loss": 4.345149040222168,
+ "step": 1828
+ },
+ {
+ "epoch": 25.40524017467249,
+ "grad_norm": 0.01669284701347351,
+ "learning_rate": 0.0006,
+ "loss": 4.237611293792725,
+ "step": 1829
+ },
+ {
+ "epoch": 25.419213973799128,
+ "grad_norm": 0.016202064231038094,
+ "learning_rate": 0.0006,
+ "loss": 4.3072309494018555,
+ "step": 1830
+ },
+ {
+ "epoch": 25.433187772925763,
+ "grad_norm": 0.01478662807494402,
+ "learning_rate": 0.0006,
+ "loss": 4.276215076446533,
+ "step": 1831
+ },
+ {
+ "epoch": 25.4471615720524,
+ "grad_norm": 0.015168731100857258,
+ "learning_rate": 0.0006,
+ "loss": 4.37509822845459,
+ "step": 1832
+ },
+ {
+ "epoch": 25.46113537117904,
+ "grad_norm": 0.01648980937898159,
+ "learning_rate": 0.0006,
+ "loss": 4.349167346954346,
+ "step": 1833
+ },
+ {
+ "epoch": 25.475109170305675,
+ "grad_norm": 0.01708398200571537,
+ "learning_rate": 0.0006,
+ "loss": 4.346580505371094,
+ "step": 1834
+ },
+ {
+ "epoch": 25.489082969432314,
+ "grad_norm": 0.017370784655213356,
+ "learning_rate": 0.0006,
+ "loss": 4.299466133117676,
+ "step": 1835
+ },
+ {
+ "epoch": 25.503056768558952,
+ "grad_norm": 0.016248228028416634,
+ "learning_rate": 0.0006,
+ "loss": 4.184610366821289,
+ "step": 1836
+ },
+ {
+ "epoch": 25.51703056768559,
+ "grad_norm": 0.017106913030147552,
+ "learning_rate": 0.0006,
+ "loss": 4.366856098175049,
+ "step": 1837
+ },
+ {
+ "epoch": 25.531004366812226,
+ "grad_norm": 0.016071965917944908,
+ "learning_rate": 0.0006,
+ "loss": 4.297691345214844,
+ "step": 1838
+ },
+ {
+ "epoch": 25.544978165938865,
+ "grad_norm": 0.016451986506581306,
+ "learning_rate": 0.0006,
+ "loss": 4.229983329772949,
+ "step": 1839
+ },
+ {
+ "epoch": 25.558951965065503,
+ "grad_norm": 0.016195081174373627,
+ "learning_rate": 0.0006,
+ "loss": 4.259893417358398,
+ "step": 1840
+ },
+ {
+ "epoch": 25.57292576419214,
+ "grad_norm": 0.0165175162255764,
+ "learning_rate": 0.0006,
+ "loss": 4.3348236083984375,
+ "step": 1841
+ },
+ {
+ "epoch": 25.586899563318777,
+ "grad_norm": 0.01651517115533352,
+ "learning_rate": 0.0006,
+ "loss": 4.319541931152344,
+ "step": 1842
+ },
+ {
+ "epoch": 25.600873362445416,
+ "grad_norm": 0.016143690794706345,
+ "learning_rate": 0.0006,
+ "loss": 4.345652103424072,
+ "step": 1843
+ },
+ {
+ "epoch": 25.61484716157205,
+ "grad_norm": 0.015140696428716183,
+ "learning_rate": 0.0006,
+ "loss": 4.212047576904297,
+ "step": 1844
+ },
+ {
+ "epoch": 25.62882096069869,
+ "grad_norm": 0.014637443237006664,
+ "learning_rate": 0.0006,
+ "loss": 4.187453269958496,
+ "step": 1845
+ },
+ {
+ "epoch": 25.64279475982533,
+ "grad_norm": 0.014585614204406738,
+ "learning_rate": 0.0006,
+ "loss": 4.253936767578125,
+ "step": 1846
+ },
+ {
+ "epoch": 25.656768558951963,
+ "grad_norm": 0.015158289112150669,
+ "learning_rate": 0.0006,
+ "loss": 4.34612512588501,
+ "step": 1847
+ },
+ {
+ "epoch": 25.670742358078602,
+ "grad_norm": 0.015202849172055721,
+ "learning_rate": 0.0006,
+ "loss": 4.270634651184082,
+ "step": 1848
+ },
+ {
+ "epoch": 25.68471615720524,
+ "grad_norm": 0.015363575890660286,
+ "learning_rate": 0.0006,
+ "loss": 4.271080493927002,
+ "step": 1849
+ },
+ {
+ "epoch": 25.69868995633188,
+ "grad_norm": 0.014360226690769196,
+ "learning_rate": 0.0006,
+ "loss": 4.375033855438232,
+ "step": 1850
+ },
+ {
+ "epoch": 25.712663755458514,
+ "grad_norm": 0.014263181947171688,
+ "learning_rate": 0.0006,
+ "loss": 4.424138069152832,
+ "step": 1851
+ },
+ {
+ "epoch": 25.726637554585153,
+ "grad_norm": 0.014398688450455666,
+ "learning_rate": 0.0006,
+ "loss": 4.410154819488525,
+ "step": 1852
+ },
+ {
+ "epoch": 25.74061135371179,
+ "grad_norm": 0.016131488606333733,
+ "learning_rate": 0.0006,
+ "loss": 4.324014663696289,
+ "step": 1853
+ },
+ {
+ "epoch": 25.754585152838427,
+ "grad_norm": 0.017568735405802727,
+ "learning_rate": 0.0006,
+ "loss": 4.367861747741699,
+ "step": 1854
+ },
+ {
+ "epoch": 25.768558951965066,
+ "grad_norm": 0.017907511442899704,
+ "learning_rate": 0.0006,
+ "loss": 4.352965354919434,
+ "step": 1855
+ },
+ {
+ "epoch": 25.782532751091704,
+ "grad_norm": 0.016918016597628593,
+ "learning_rate": 0.0006,
+ "loss": 4.252397537231445,
+ "step": 1856
+ },
+ {
+ "epoch": 25.79650655021834,
+ "grad_norm": 0.017274610698223114,
+ "learning_rate": 0.0006,
+ "loss": 4.342309951782227,
+ "step": 1857
+ },
+ {
+ "epoch": 25.810480349344978,
+ "grad_norm": 0.017879825085401535,
+ "learning_rate": 0.0006,
+ "loss": 4.330987453460693,
+ "step": 1858
+ },
+ {
+ "epoch": 25.824454148471617,
+ "grad_norm": 0.01736099272966385,
+ "learning_rate": 0.0006,
+ "loss": 4.307255744934082,
+ "step": 1859
+ },
+ {
+ "epoch": 25.83842794759825,
+ "grad_norm": 0.016575824469327927,
+ "learning_rate": 0.0006,
+ "loss": 4.262016773223877,
+ "step": 1860
+ },
+ {
+ "epoch": 25.85240174672489,
+ "grad_norm": 0.015181140042841434,
+ "learning_rate": 0.0006,
+ "loss": 4.291863441467285,
+ "step": 1861
+ },
+ {
+ "epoch": 25.86637554585153,
+ "grad_norm": 0.015965687111020088,
+ "learning_rate": 0.0006,
+ "loss": 4.424836158752441,
+ "step": 1862
+ },
+ {
+ "epoch": 25.880349344978168,
+ "grad_norm": 0.01687219925224781,
+ "learning_rate": 0.0006,
+ "loss": 4.343531131744385,
+ "step": 1863
+ },
+ {
+ "epoch": 25.894323144104803,
+ "grad_norm": 0.015115504153072834,
+ "learning_rate": 0.0006,
+ "loss": 4.357808589935303,
+ "step": 1864
+ },
+ {
+ "epoch": 25.90829694323144,
+ "grad_norm": 0.014150998555123806,
+ "learning_rate": 0.0006,
+ "loss": 4.324550151824951,
+ "step": 1865
+ },
+ {
+ "epoch": 25.92227074235808,
+ "grad_norm": 0.014989510178565979,
+ "learning_rate": 0.0006,
+ "loss": 4.380428314208984,
+ "step": 1866
+ },
+ {
+ "epoch": 25.936244541484715,
+ "grad_norm": 0.01527960691601038,
+ "learning_rate": 0.0006,
+ "loss": 4.399374961853027,
+ "step": 1867
+ },
+ {
+ "epoch": 25.950218340611354,
+ "grad_norm": 0.015434633940458298,
+ "learning_rate": 0.0006,
+ "loss": 4.260984420776367,
+ "step": 1868
+ },
+ {
+ "epoch": 25.964192139737992,
+ "grad_norm": 0.015370228327810764,
+ "learning_rate": 0.0006,
+ "loss": 4.3340253829956055,
+ "step": 1869
+ },
+ {
+ "epoch": 25.978165938864628,
+ "grad_norm": 0.015348542481660843,
+ "learning_rate": 0.0006,
+ "loss": 4.369132995605469,
+ "step": 1870
+ },
+ {
+ "epoch": 25.992139737991266,
+ "grad_norm": 0.015354936942458153,
+ "learning_rate": 0.0006,
+ "loss": 4.381937026977539,
+ "step": 1871
+ },
+ {
+ "epoch": 26.0,
+ "grad_norm": 0.01796272210776806,
+ "learning_rate": 0.0006,
+ "loss": 4.286768436431885,
+ "step": 1872
+ },
+ {
+ "epoch": 26.0,
+ "eval_loss": 4.576984405517578,
+ "eval_runtime": 57.2091,
+ "eval_samples_per_second": 42.685,
+ "eval_steps_per_second": 1.346,
+ "step": 1872
+ },
+ {
+ "epoch": 26.01397379912664,
+ "grad_norm": 0.01748845726251602,
+ "learning_rate": 0.0006,
+ "loss": 4.337504863739014,
+ "step": 1873
+ },
+ {
+ "epoch": 26.027947598253274,
+ "grad_norm": 0.019244296476244926,
+ "learning_rate": 0.0006,
+ "loss": 4.264280796051025,
+ "step": 1874
+ },
+ {
+ "epoch": 26.041921397379912,
+ "grad_norm": 0.0215692650526762,
+ "learning_rate": 0.0006,
+ "loss": 4.350830078125,
+ "step": 1875
+ },
+ {
+ "epoch": 26.05589519650655,
+ "grad_norm": 0.022189252078533173,
+ "learning_rate": 0.0006,
+ "loss": 4.299098968505859,
+ "step": 1876
+ },
+ {
+ "epoch": 26.069868995633186,
+ "grad_norm": 0.022384043782949448,
+ "learning_rate": 0.0006,
+ "loss": 4.1185173988342285,
+ "step": 1877
+ },
+ {
+ "epoch": 26.083842794759825,
+ "grad_norm": 0.02076014317572117,
+ "learning_rate": 0.0006,
+ "loss": 4.297612190246582,
+ "step": 1878
+ },
+ {
+ "epoch": 26.097816593886463,
+ "grad_norm": 0.020843051373958588,
+ "learning_rate": 0.0006,
+ "loss": 4.294098854064941,
+ "step": 1879
+ },
+ {
+ "epoch": 26.111790393013102,
+ "grad_norm": 0.020654456689953804,
+ "learning_rate": 0.0006,
+ "loss": 4.267856597900391,
+ "step": 1880
+ },
+ {
+ "epoch": 26.125764192139737,
+ "grad_norm": 0.01860472559928894,
+ "learning_rate": 0.0006,
+ "loss": 4.284232139587402,
+ "step": 1881
+ },
+ {
+ "epoch": 26.139737991266376,
+ "grad_norm": 0.018719421699643135,
+ "learning_rate": 0.0006,
+ "loss": 4.277214527130127,
+ "step": 1882
+ },
+ {
+ "epoch": 26.153711790393015,
+ "grad_norm": 0.020903829485177994,
+ "learning_rate": 0.0006,
+ "loss": 4.254947662353516,
+ "step": 1883
+ },
+ {
+ "epoch": 26.16768558951965,
+ "grad_norm": 0.02020149491727352,
+ "learning_rate": 0.0006,
+ "loss": 4.2526936531066895,
+ "step": 1884
+ },
+ {
+ "epoch": 26.18165938864629,
+ "grad_norm": 0.0180392786860466,
+ "learning_rate": 0.0006,
+ "loss": 4.309922695159912,
+ "step": 1885
+ },
+ {
+ "epoch": 26.195633187772927,
+ "grad_norm": 0.018163125962018967,
+ "learning_rate": 0.0006,
+ "loss": 4.13860559463501,
+ "step": 1886
+ },
+ {
+ "epoch": 26.209606986899562,
+ "grad_norm": 0.01853291131556034,
+ "learning_rate": 0.0006,
+ "loss": 4.190926551818848,
+ "step": 1887
+ },
+ {
+ "epoch": 26.2235807860262,
+ "grad_norm": 0.01757677271962166,
+ "learning_rate": 0.0006,
+ "loss": 4.163999557495117,
+ "step": 1888
+ },
+ {
+ "epoch": 26.23755458515284,
+ "grad_norm": 0.01767992228269577,
+ "learning_rate": 0.0006,
+ "loss": 4.184430122375488,
+ "step": 1889
+ },
+ {
+ "epoch": 26.251528384279474,
+ "grad_norm": 0.018139973282814026,
+ "learning_rate": 0.0006,
+ "loss": 4.300766944885254,
+ "step": 1890
+ },
+ {
+ "epoch": 26.265502183406113,
+ "grad_norm": 0.020184461027383804,
+ "learning_rate": 0.0006,
+ "loss": 4.232537269592285,
+ "step": 1891
+ },
+ {
+ "epoch": 26.27947598253275,
+ "grad_norm": 0.02028856985270977,
+ "learning_rate": 0.0006,
+ "loss": 4.324883460998535,
+ "step": 1892
+ },
+ {
+ "epoch": 26.29344978165939,
+ "grad_norm": 0.017120616510510445,
+ "learning_rate": 0.0006,
+ "loss": 4.23493766784668,
+ "step": 1893
+ },
+ {
+ "epoch": 26.307423580786025,
+ "grad_norm": 0.016202978789806366,
+ "learning_rate": 0.0006,
+ "loss": 4.197602272033691,
+ "step": 1894
+ },
+ {
+ "epoch": 26.321397379912664,
+ "grad_norm": 0.017231421545147896,
+ "learning_rate": 0.0006,
+ "loss": 4.302001953125,
+ "step": 1895
+ },
+ {
+ "epoch": 26.335371179039303,
+ "grad_norm": 0.01725655607879162,
+ "learning_rate": 0.0006,
+ "loss": 4.356800079345703,
+ "step": 1896
+ },
+ {
+ "epoch": 26.349344978165938,
+ "grad_norm": 0.016657443717122078,
+ "learning_rate": 0.0006,
+ "loss": 4.266307830810547,
+ "step": 1897
+ },
+ {
+ "epoch": 26.363318777292577,
+ "grad_norm": 0.01616556942462921,
+ "learning_rate": 0.0006,
+ "loss": 4.307868003845215,
+ "step": 1898
+ },
+ {
+ "epoch": 26.377292576419215,
+ "grad_norm": 0.016621306538581848,
+ "learning_rate": 0.0006,
+ "loss": 4.172905921936035,
+ "step": 1899
+ },
+ {
+ "epoch": 26.39126637554585,
+ "grad_norm": 0.016243206337094307,
+ "learning_rate": 0.0006,
+ "loss": 4.162896156311035,
+ "step": 1900
+ },
+ {
+ "epoch": 26.40524017467249,
+ "grad_norm": 0.016663808375597,
+ "learning_rate": 0.0006,
+ "loss": 4.258232116699219,
+ "step": 1901
+ },
+ {
+ "epoch": 26.419213973799128,
+ "grad_norm": 0.01677478663623333,
+ "learning_rate": 0.0006,
+ "loss": 4.336090087890625,
+ "step": 1902
+ },
+ {
+ "epoch": 26.433187772925763,
+ "grad_norm": 0.01641303487122059,
+ "learning_rate": 0.0006,
+ "loss": 4.191476821899414,
+ "step": 1903
+ },
+ {
+ "epoch": 26.4471615720524,
+ "grad_norm": 0.016700323671102524,
+ "learning_rate": 0.0006,
+ "loss": 4.299427509307861,
+ "step": 1904
+ },
+ {
+ "epoch": 26.46113537117904,
+ "grad_norm": 0.017595035955309868,
+ "learning_rate": 0.0006,
+ "loss": 4.312828063964844,
+ "step": 1905
+ },
+ {
+ "epoch": 26.475109170305675,
+ "grad_norm": 0.020300284028053284,
+ "learning_rate": 0.0006,
+ "loss": 4.28331184387207,
+ "step": 1906
+ },
+ {
+ "epoch": 26.489082969432314,
+ "grad_norm": 0.02198321744799614,
+ "learning_rate": 0.0006,
+ "loss": 4.3280839920043945,
+ "step": 1907
+ },
+ {
+ "epoch": 26.503056768558952,
+ "grad_norm": 0.02003200724720955,
+ "learning_rate": 0.0006,
+ "loss": 4.325623989105225,
+ "step": 1908
+ },
+ {
+ "epoch": 26.51703056768559,
+ "grad_norm": 0.01870287023484707,
+ "learning_rate": 0.0006,
+ "loss": 4.251203536987305,
+ "step": 1909
+ },
+ {
+ "epoch": 26.531004366812226,
+ "grad_norm": 0.01835448481142521,
+ "learning_rate": 0.0006,
+ "loss": 4.335144996643066,
+ "step": 1910
+ },
+ {
+ "epoch": 26.544978165938865,
+ "grad_norm": 0.02046327292919159,
+ "learning_rate": 0.0006,
+ "loss": 4.238023281097412,
+ "step": 1911
+ },
+ {
+ "epoch": 26.558951965065503,
+ "grad_norm": 0.019771311432123184,
+ "learning_rate": 0.0006,
+ "loss": 4.353594779968262,
+ "step": 1912
+ },
+ {
+ "epoch": 26.57292576419214,
+ "grad_norm": 0.018090544268488884,
+ "learning_rate": 0.0006,
+ "loss": 4.304343223571777,
+ "step": 1913
+ },
+ {
+ "epoch": 26.586899563318777,
+ "grad_norm": 0.017489777877926826,
+ "learning_rate": 0.0006,
+ "loss": 4.204375267028809,
+ "step": 1914
+ },
+ {
+ "epoch": 26.600873362445416,
+ "grad_norm": 0.016785891726613045,
+ "learning_rate": 0.0006,
+ "loss": 4.286669731140137,
+ "step": 1915
+ },
+ {
+ "epoch": 26.61484716157205,
+ "grad_norm": 0.016382789239287376,
+ "learning_rate": 0.0006,
+ "loss": 4.347662448883057,
+ "step": 1916
+ },
+ {
+ "epoch": 26.62882096069869,
+ "grad_norm": 0.016581876203417778,
+ "learning_rate": 0.0006,
+ "loss": 4.2716193199157715,
+ "step": 1917
+ },
+ {
+ "epoch": 26.64279475982533,
+ "grad_norm": 0.015627184882760048,
+ "learning_rate": 0.0006,
+ "loss": 4.311489105224609,
+ "step": 1918
+ },
+ {
+ "epoch": 26.656768558951963,
+ "grad_norm": 0.015908483415842056,
+ "learning_rate": 0.0006,
+ "loss": 4.262099742889404,
+ "step": 1919
+ },
+ {
+ "epoch": 26.670742358078602,
+ "grad_norm": 0.015576236881315708,
+ "learning_rate": 0.0006,
+ "loss": 4.275871276855469,
+ "step": 1920
+ },
+ {
+ "epoch": 26.68471615720524,
+ "grad_norm": 0.014621064998209476,
+ "learning_rate": 0.0006,
+ "loss": 4.184260845184326,
+ "step": 1921
+ },
+ {
+ "epoch": 26.69868995633188,
+ "grad_norm": 0.014797811396420002,
+ "learning_rate": 0.0006,
+ "loss": 4.3412628173828125,
+ "step": 1922
+ },
+ {
+ "epoch": 26.712663755458514,
+ "grad_norm": 0.015610570088028908,
+ "learning_rate": 0.0006,
+ "loss": 4.181094646453857,
+ "step": 1923
+ },
+ {
+ "epoch": 26.726637554585153,
+ "grad_norm": 0.016048265621066093,
+ "learning_rate": 0.0006,
+ "loss": 4.365290641784668,
+ "step": 1924
+ },
+ {
+ "epoch": 26.74061135371179,
+ "grad_norm": 0.016604071483016014,
+ "learning_rate": 0.0006,
+ "loss": 4.2628912925720215,
+ "step": 1925
+ },
+ {
+ "epoch": 26.754585152838427,
+ "grad_norm": 0.017044221982359886,
+ "learning_rate": 0.0006,
+ "loss": 4.378512382507324,
+ "step": 1926
+ },
+ {
+ "epoch": 26.768558951965066,
+ "grad_norm": 0.01746384985744953,
+ "learning_rate": 0.0006,
+ "loss": 4.3888773918151855,
+ "step": 1927
+ },
+ {
+ "epoch": 26.782532751091704,
+ "grad_norm": 0.01572711206972599,
+ "learning_rate": 0.0006,
+ "loss": 4.189935684204102,
+ "step": 1928
+ },
+ {
+ "epoch": 26.79650655021834,
+ "grad_norm": 0.014976629987359047,
+ "learning_rate": 0.0006,
+ "loss": 4.280525207519531,
+ "step": 1929
+ },
+ {
+ "epoch": 26.810480349344978,
+ "grad_norm": 0.014804664999246597,
+ "learning_rate": 0.0006,
+ "loss": 4.270630836486816,
+ "step": 1930
+ },
+ {
+ "epoch": 26.824454148471617,
+ "grad_norm": 0.014641920104622841,
+ "learning_rate": 0.0006,
+ "loss": 4.319505214691162,
+ "step": 1931
+ },
+ {
+ "epoch": 26.83842794759825,
+ "grad_norm": 0.014928505755960941,
+ "learning_rate": 0.0006,
+ "loss": 4.180147647857666,
+ "step": 1932
+ },
+ {
+ "epoch": 26.85240174672489,
+ "grad_norm": 0.016130365431308746,
+ "learning_rate": 0.0006,
+ "loss": 4.260660648345947,
+ "step": 1933
+ },
+ {
+ "epoch": 26.86637554585153,
+ "grad_norm": 0.015642493963241577,
+ "learning_rate": 0.0006,
+ "loss": 4.304064750671387,
+ "step": 1934
+ },
+ {
+ "epoch": 26.880349344978168,
+ "grad_norm": 0.015988217666745186,
+ "learning_rate": 0.0006,
+ "loss": 4.2548017501831055,
+ "step": 1935
+ },
+ {
+ "epoch": 26.894323144104803,
+ "grad_norm": 0.016501398757100105,
+ "learning_rate": 0.0006,
+ "loss": 4.260058879852295,
+ "step": 1936
+ },
+ {
+ "epoch": 26.90829694323144,
+ "grad_norm": 0.01645040698349476,
+ "learning_rate": 0.0006,
+ "loss": 4.483081340789795,
+ "step": 1937
+ },
+ {
+ "epoch": 26.92227074235808,
+ "grad_norm": 0.016140863299369812,
+ "learning_rate": 0.0006,
+ "loss": 4.1776838302612305,
+ "step": 1938
+ },
+ {
+ "epoch": 26.936244541484715,
+ "grad_norm": 0.016681019216775894,
+ "learning_rate": 0.0006,
+ "loss": 4.224747657775879,
+ "step": 1939
+ },
+ {
+ "epoch": 26.950218340611354,
+ "grad_norm": 0.01650378853082657,
+ "learning_rate": 0.0006,
+ "loss": 4.305703163146973,
+ "step": 1940
+ },
+ {
+ "epoch": 26.964192139737992,
+ "grad_norm": 0.017051683738827705,
+ "learning_rate": 0.0006,
+ "loss": 4.285835266113281,
+ "step": 1941
+ },
+ {
+ "epoch": 26.978165938864628,
+ "grad_norm": 0.016894327476620674,
+ "learning_rate": 0.0006,
+ "loss": 4.318901538848877,
+ "step": 1942
+ },
+ {
+ "epoch": 26.992139737991266,
+ "grad_norm": 0.017313100397586823,
+ "learning_rate": 0.0006,
+ "loss": 4.281434535980225,
+ "step": 1943
+ },
+ {
+ "epoch": 27.0,
+ "grad_norm": 0.018450884148478508,
+ "learning_rate": 0.0006,
+ "loss": 4.335104942321777,
+ "step": 1944
+ },
+ {
+ "epoch": 27.0,
+ "eval_loss": 4.640687465667725,
+ "eval_runtime": 56.9624,
+ "eval_samples_per_second": 42.87,
+ "eval_steps_per_second": 1.352,
+ "step": 1944
+ },
+ {
+ "epoch": 27.01397379912664,
+ "grad_norm": 0.017755698412656784,
+ "learning_rate": 0.0006,
+ "loss": 4.278824329376221,
+ "step": 1945
+ },
+ {
+ "epoch": 27.027947598253274,
+ "grad_norm": 0.01724310778081417,
+ "learning_rate": 0.0006,
+ "loss": 4.311049461364746,
+ "step": 1946
+ },
+ {
+ "epoch": 27.041921397379912,
+ "grad_norm": 0.01690123789012432,
+ "learning_rate": 0.0006,
+ "loss": 4.261994361877441,
+ "step": 1947
+ },
+ {
+ "epoch": 27.05589519650655,
+ "grad_norm": 0.018406696617603302,
+ "learning_rate": 0.0006,
+ "loss": 4.1925153732299805,
+ "step": 1948
+ },
+ {
+ "epoch": 27.069868995633186,
+ "grad_norm": 0.019945867359638214,
+ "learning_rate": 0.0006,
+ "loss": 4.108537197113037,
+ "step": 1949
+ },
+ {
+ "epoch": 27.083842794759825,
+ "grad_norm": 0.020316628739237785,
+ "learning_rate": 0.0006,
+ "loss": 4.209043025970459,
+ "step": 1950
+ },
+ {
+ "epoch": 27.097816593886463,
+ "grad_norm": 0.019690370187163353,
+ "learning_rate": 0.0006,
+ "loss": 4.195789337158203,
+ "step": 1951
+ },
+ {
+ "epoch": 27.111790393013102,
+ "grad_norm": 0.01785232499241829,
+ "learning_rate": 0.0006,
+ "loss": 4.21466064453125,
+ "step": 1952
+ },
+ {
+ "epoch": 27.125764192139737,
+ "grad_norm": 0.0164767038077116,
+ "learning_rate": 0.0006,
+ "loss": 4.270247459411621,
+ "step": 1953
+ },
+ {
+ "epoch": 27.139737991266376,
+ "grad_norm": 0.018008455634117126,
+ "learning_rate": 0.0006,
+ "loss": 4.201042175292969,
+ "step": 1954
+ },
+ {
+ "epoch": 27.153711790393015,
+ "grad_norm": 0.01807340793311596,
+ "learning_rate": 0.0006,
+ "loss": 4.25289249420166,
+ "step": 1955
+ },
+ {
+ "epoch": 27.16768558951965,
+ "grad_norm": 0.016246909275650978,
+ "learning_rate": 0.0006,
+ "loss": 4.116283416748047,
+ "step": 1956
+ },
+ {
+ "epoch": 27.18165938864629,
+ "grad_norm": 0.017545776441693306,
+ "learning_rate": 0.0006,
+ "loss": 4.316999435424805,
+ "step": 1957
+ },
+ {
+ "epoch": 27.195633187772927,
+ "grad_norm": 0.017915375530719757,
+ "learning_rate": 0.0006,
+ "loss": 4.253859519958496,
+ "step": 1958
+ },
+ {
+ "epoch": 27.209606986899562,
+ "grad_norm": 0.018112223595380783,
+ "learning_rate": 0.0006,
+ "loss": 4.2541022300720215,
+ "step": 1959
+ },
+ {
+ "epoch": 27.2235807860262,
+ "grad_norm": 0.01846139505505562,
+ "learning_rate": 0.0006,
+ "loss": 4.20438289642334,
+ "step": 1960
+ },
+ {
+ "epoch": 27.23755458515284,
+ "grad_norm": 0.01943897269666195,
+ "learning_rate": 0.0006,
+ "loss": 4.2799577713012695,
+ "step": 1961
+ },
+ {
+ "epoch": 27.251528384279474,
+ "grad_norm": 0.018253348767757416,
+ "learning_rate": 0.0006,
+ "loss": 4.1720499992370605,
+ "step": 1962
+ },
+ {
+ "epoch": 27.265502183406113,
+ "grad_norm": 0.018292246386408806,
+ "learning_rate": 0.0006,
+ "loss": 4.28231143951416,
+ "step": 1963
+ },
+ {
+ "epoch": 27.27947598253275,
+ "grad_norm": 0.01943155936896801,
+ "learning_rate": 0.0006,
+ "loss": 4.314976692199707,
+ "step": 1964
+ },
+ {
+ "epoch": 27.29344978165939,
+ "grad_norm": 0.019064994528889656,
+ "learning_rate": 0.0006,
+ "loss": 4.25328254699707,
+ "step": 1965
+ },
+ {
+ "epoch": 27.307423580786025,
+ "grad_norm": 0.017423273995518684,
+ "learning_rate": 0.0006,
+ "loss": 4.195992469787598,
+ "step": 1966
+ },
+ {
+ "epoch": 27.321397379912664,
+ "grad_norm": 0.01677924580872059,
+ "learning_rate": 0.0006,
+ "loss": 4.260043144226074,
+ "step": 1967
+ },
+ {
+ "epoch": 27.335371179039303,
+ "grad_norm": 0.01720673404633999,
+ "learning_rate": 0.0006,
+ "loss": 4.222780704498291,
+ "step": 1968
+ },
+ {
+ "epoch": 27.349344978165938,
+ "grad_norm": 0.017300743609666824,
+ "learning_rate": 0.0006,
+ "loss": 4.109524250030518,
+ "step": 1969
+ },
+ {
+ "epoch": 27.363318777292577,
+ "grad_norm": 0.016065871343016624,
+ "learning_rate": 0.0006,
+ "loss": 4.36108922958374,
+ "step": 1970
+ },
+ {
+ "epoch": 27.377292576419215,
+ "grad_norm": 0.016626616939902306,
+ "learning_rate": 0.0006,
+ "loss": 4.220630645751953,
+ "step": 1971
+ },
+ {
+ "epoch": 27.39126637554585,
+ "grad_norm": 0.016482602804899216,
+ "learning_rate": 0.0006,
+ "loss": 4.344797134399414,
+ "step": 1972
+ },
+ {
+ "epoch": 27.40524017467249,
+ "grad_norm": 0.01727953553199768,
+ "learning_rate": 0.0006,
+ "loss": 4.296645164489746,
+ "step": 1973
+ },
+ {
+ "epoch": 27.419213973799128,
+ "grad_norm": 0.01522731315344572,
+ "learning_rate": 0.0006,
+ "loss": 4.223751068115234,
+ "step": 1974
+ },
+ {
+ "epoch": 27.433187772925763,
+ "grad_norm": 0.01581776700913906,
+ "learning_rate": 0.0006,
+ "loss": 4.351379871368408,
+ "step": 1975
+ },
+ {
+ "epoch": 27.4471615720524,
+ "grad_norm": 0.016394605860114098,
+ "learning_rate": 0.0006,
+ "loss": 4.2412919998168945,
+ "step": 1976
+ },
+ {
+ "epoch": 27.46113537117904,
+ "grad_norm": 0.01751169003546238,
+ "learning_rate": 0.0006,
+ "loss": 4.257606506347656,
+ "step": 1977
+ },
+ {
+ "epoch": 27.475109170305675,
+ "grad_norm": 0.017452310770750046,
+ "learning_rate": 0.0006,
+ "loss": 4.380134582519531,
+ "step": 1978
+ },
+ {
+ "epoch": 27.489082969432314,
+ "grad_norm": 0.017741898074746132,
+ "learning_rate": 0.0006,
+ "loss": 4.279632568359375,
+ "step": 1979
+ },
+ {
+ "epoch": 27.503056768558952,
+ "grad_norm": 0.017680590972304344,
+ "learning_rate": 0.0006,
+ "loss": 4.273406028747559,
+ "step": 1980
+ },
+ {
+ "epoch": 27.51703056768559,
+ "grad_norm": 0.01864718459546566,
+ "learning_rate": 0.0006,
+ "loss": 4.278583526611328,
+ "step": 1981
+ },
+ {
+ "epoch": 27.531004366812226,
+ "grad_norm": 0.019558541476726532,
+ "learning_rate": 0.0006,
+ "loss": 4.211919784545898,
+ "step": 1982
+ },
+ {
+ "epoch": 27.544978165938865,
+ "grad_norm": 0.019467400386929512,
+ "learning_rate": 0.0006,
+ "loss": 4.276822566986084,
+ "step": 1983
+ },
+ {
+ "epoch": 27.558951965065503,
+ "grad_norm": 0.01756172813475132,
+ "learning_rate": 0.0006,
+ "loss": 4.165379047393799,
+ "step": 1984
+ },
+ {
+ "epoch": 27.57292576419214,
+ "grad_norm": 0.017458263784646988,
+ "learning_rate": 0.0006,
+ "loss": 4.2992401123046875,
+ "step": 1985
+ },
+ {
+ "epoch": 27.586899563318777,
+ "grad_norm": 0.018089665099978447,
+ "learning_rate": 0.0006,
+ "loss": 4.243378639221191,
+ "step": 1986
+ },
+ {
+ "epoch": 27.600873362445416,
+ "grad_norm": 0.019535524770617485,
+ "learning_rate": 0.0006,
+ "loss": 4.321477890014648,
+ "step": 1987
+ },
+ {
+ "epoch": 27.61484716157205,
+ "grad_norm": 0.01997970975935459,
+ "learning_rate": 0.0006,
+ "loss": 4.123082637786865,
+ "step": 1988
+ },
+ {
+ "epoch": 27.62882096069869,
+ "grad_norm": 0.01913067139685154,
+ "learning_rate": 0.0006,
+ "loss": 4.289896488189697,
+ "step": 1989
+ },
+ {
+ "epoch": 27.64279475982533,
+ "grad_norm": 0.0169806070625782,
+ "learning_rate": 0.0006,
+ "loss": 4.247004985809326,
+ "step": 1990
+ },
+ {
+ "epoch": 27.656768558951963,
+ "grad_norm": 0.017315007746219635,
+ "learning_rate": 0.0006,
+ "loss": 4.2095441818237305,
+ "step": 1991
+ },
+ {
+ "epoch": 27.670742358078602,
+ "grad_norm": 0.015098759904503822,
+ "learning_rate": 0.0006,
+ "loss": 4.145560264587402,
+ "step": 1992
+ },
+ {
+ "epoch": 27.68471615720524,
+ "grad_norm": 0.016240514814853668,
+ "learning_rate": 0.0006,
+ "loss": 4.254927635192871,
+ "step": 1993
+ },
+ {
+ "epoch": 27.69868995633188,
+ "grad_norm": 0.01621108129620552,
+ "learning_rate": 0.0006,
+ "loss": 4.219725608825684,
+ "step": 1994
+ },
+ {
+ "epoch": 27.712663755458514,
+ "grad_norm": 0.015191009268164635,
+ "learning_rate": 0.0006,
+ "loss": 4.194345474243164,
+ "step": 1995
+ },
+ {
+ "epoch": 27.726637554585153,
+ "grad_norm": 0.014387983828783035,
+ "learning_rate": 0.0006,
+ "loss": 4.211467742919922,
+ "step": 1996
+ },
+ {
+ "epoch": 27.74061135371179,
+ "grad_norm": 0.015460561029613018,
+ "learning_rate": 0.0006,
+ "loss": 4.358700752258301,
+ "step": 1997
+ },
+ {
+ "epoch": 27.754585152838427,
+ "grad_norm": 0.015100076794624329,
+ "learning_rate": 0.0006,
+ "loss": 4.171223163604736,
+ "step": 1998
+ },
+ {
+ "epoch": 27.768558951965066,
+ "grad_norm": 0.015607891604304314,
+ "learning_rate": 0.0006,
+ "loss": 4.217874526977539,
+ "step": 1999
+ },
+ {
+ "epoch": 27.782532751091704,
+ "grad_norm": 0.015650689601898193,
+ "learning_rate": 0.0006,
+ "loss": 4.280780792236328,
+ "step": 2000
+ },
+ {
+ "epoch": 27.79650655021834,
+ "grad_norm": 0.01580994389951229,
+ "learning_rate": 0.0006,
+ "loss": 4.159212112426758,
+ "step": 2001
+ },
+ {
+ "epoch": 27.810480349344978,
+ "grad_norm": 0.016607413068413734,
+ "learning_rate": 0.0006,
+ "loss": 4.238853454589844,
+ "step": 2002
+ },
+ {
+ "epoch": 27.824454148471617,
+ "grad_norm": 0.017383860424160957,
+ "learning_rate": 0.0006,
+ "loss": 4.233944892883301,
+ "step": 2003
+ },
+ {
+ "epoch": 27.83842794759825,
+ "grad_norm": 0.014966659247875214,
+ "learning_rate": 0.0006,
+ "loss": 4.167392730712891,
+ "step": 2004
+ },
+ {
+ "epoch": 27.85240174672489,
+ "grad_norm": 0.014589878730475903,
+ "learning_rate": 0.0006,
+ "loss": 4.2390899658203125,
+ "step": 2005
+ },
+ {
+ "epoch": 27.86637554585153,
+ "grad_norm": 0.017179450020194054,
+ "learning_rate": 0.0006,
+ "loss": 4.365785598754883,
+ "step": 2006
+ },
+ {
+ "epoch": 27.880349344978168,
+ "grad_norm": 0.01893901452422142,
+ "learning_rate": 0.0006,
+ "loss": 4.289430141448975,
+ "step": 2007
+ },
+ {
+ "epoch": 27.894323144104803,
+ "grad_norm": 0.018619712442159653,
+ "learning_rate": 0.0006,
+ "loss": 4.310665130615234,
+ "step": 2008
+ },
+ {
+ "epoch": 27.90829694323144,
+ "grad_norm": 0.019180385395884514,
+ "learning_rate": 0.0006,
+ "loss": 4.27454137802124,
+ "step": 2009
+ },
+ {
+ "epoch": 27.92227074235808,
+ "grad_norm": 0.017811523750424385,
+ "learning_rate": 0.0006,
+ "loss": 4.312562465667725,
+ "step": 2010
+ },
+ {
+ "epoch": 27.936244541484715,
+ "grad_norm": 0.016583573073148727,
+ "learning_rate": 0.0006,
+ "loss": 4.202253341674805,
+ "step": 2011
+ },
+ {
+ "epoch": 27.950218340611354,
+ "grad_norm": 0.01658615842461586,
+ "learning_rate": 0.0006,
+ "loss": 4.183433532714844,
+ "step": 2012
+ },
+ {
+ "epoch": 27.964192139737992,
+ "grad_norm": 0.015327401459217072,
+ "learning_rate": 0.0006,
+ "loss": 4.2125139236450195,
+ "step": 2013
+ },
+ {
+ "epoch": 27.978165938864628,
+ "grad_norm": 0.015004601329565048,
+ "learning_rate": 0.0006,
+ "loss": 4.257392883300781,
+ "step": 2014
+ },
+ {
+ "epoch": 27.992139737991266,
+ "grad_norm": 0.015259161591529846,
+ "learning_rate": 0.0006,
+ "loss": 4.241818428039551,
+ "step": 2015
+ },
+ {
+ "epoch": 28.0,
+ "grad_norm": 0.016808776184916496,
+ "learning_rate": 0.0006,
+ "loss": 4.146241188049316,
+ "step": 2016
+ },
+ {
+ "epoch": 28.0,
+ "eval_loss": 4.663090705871582,
+ "eval_runtime": 56.4559,
+ "eval_samples_per_second": 43.255,
+ "eval_steps_per_second": 1.364,
+ "step": 2016
+ },
+ {
+ "epoch": 28.01397379912664,
+ "grad_norm": 0.015784382820129395,
+ "learning_rate": 0.0006,
+ "loss": 4.044191360473633,
+ "step": 2017
+ },
+ {
+ "epoch": 28.027947598253274,
+ "grad_norm": 0.015636710450053215,
+ "learning_rate": 0.0006,
+ "loss": 4.128754615783691,
+ "step": 2018
+ },
+ {
+ "epoch": 28.041921397379912,
+ "grad_norm": 0.015364975668489933,
+ "learning_rate": 0.0006,
+ "loss": 4.191835403442383,
+ "step": 2019
+ },
+ {
+ "epoch": 28.05589519650655,
+ "grad_norm": 0.015015700832009315,
+ "learning_rate": 0.0006,
+ "loss": 4.120022773742676,
+ "step": 2020
+ },
+ {
+ "epoch": 28.069868995633186,
+ "grad_norm": 0.016666799783706665,
+ "learning_rate": 0.0006,
+ "loss": 4.33906364440918,
+ "step": 2021
+ },
+ {
+ "epoch": 28.083842794759825,
+ "grad_norm": 0.01780104823410511,
+ "learning_rate": 0.0006,
+ "loss": 4.279458999633789,
+ "step": 2022
+ },
+ {
+ "epoch": 28.097816593886463,
+ "grad_norm": 0.01870882883667946,
+ "learning_rate": 0.0006,
+ "loss": 4.34728479385376,
+ "step": 2023
+ },
+ {
+ "epoch": 28.111790393013102,
+ "grad_norm": 0.018406381830573082,
+ "learning_rate": 0.0006,
+ "loss": 4.211311340332031,
+ "step": 2024
+ },
+ {
+ "epoch": 28.125764192139737,
+ "grad_norm": 0.017833448946475983,
+ "learning_rate": 0.0006,
+ "loss": 4.223254203796387,
+ "step": 2025
+ },
+ {
+ "epoch": 28.139737991266376,
+ "grad_norm": 0.017552688717842102,
+ "learning_rate": 0.0006,
+ "loss": 4.208823204040527,
+ "step": 2026
+ },
+ {
+ "epoch": 28.153711790393015,
+ "grad_norm": 0.018500229343771935,
+ "learning_rate": 0.0006,
+ "loss": 4.198025226593018,
+ "step": 2027
+ },
+ {
+ "epoch": 28.16768558951965,
+ "grad_norm": 0.01782156340777874,
+ "learning_rate": 0.0006,
+ "loss": 4.179978370666504,
+ "step": 2028
+ },
+ {
+ "epoch": 28.18165938864629,
+ "grad_norm": 0.019580967724323273,
+ "learning_rate": 0.0006,
+ "loss": 4.291114807128906,
+ "step": 2029
+ },
+ {
+ "epoch": 28.195633187772927,
+ "grad_norm": 0.020804036408662796,
+ "learning_rate": 0.0006,
+ "loss": 4.220607280731201,
+ "step": 2030
+ },
+ {
+ "epoch": 28.209606986899562,
+ "grad_norm": 0.019595693796873093,
+ "learning_rate": 0.0006,
+ "loss": 4.2905731201171875,
+ "step": 2031
+ },
+ {
+ "epoch": 28.2235807860262,
+ "grad_norm": 0.019247086718678474,
+ "learning_rate": 0.0006,
+ "loss": 4.2467360496521,
+ "step": 2032
+ },
+ {
+ "epoch": 28.23755458515284,
+ "grad_norm": 0.01958036608994007,
+ "learning_rate": 0.0006,
+ "loss": 4.286137580871582,
+ "step": 2033
+ },
+ {
+ "epoch": 28.251528384279474,
+ "grad_norm": 0.021354753524065018,
+ "learning_rate": 0.0006,
+ "loss": 4.234755039215088,
+ "step": 2034
+ },
+ {
+ "epoch": 28.265502183406113,
+ "grad_norm": 0.02206375077366829,
+ "learning_rate": 0.0006,
+ "loss": 4.149641036987305,
+ "step": 2035
+ },
+ {
+ "epoch": 28.27947598253275,
+ "grad_norm": 0.02201310358941555,
+ "learning_rate": 0.0006,
+ "loss": 4.1928629875183105,
+ "step": 2036
+ },
+ {
+ "epoch": 28.29344978165939,
+ "grad_norm": 0.01961950585246086,
+ "learning_rate": 0.0006,
+ "loss": 4.211404323577881,
+ "step": 2037
+ },
+ {
+ "epoch": 28.307423580786025,
+ "grad_norm": 0.019211795181035995,
+ "learning_rate": 0.0006,
+ "loss": 4.200976848602295,
+ "step": 2038
+ },
+ {
+ "epoch": 28.321397379912664,
+ "grad_norm": 0.01779370754957199,
+ "learning_rate": 0.0006,
+ "loss": 4.249148845672607,
+ "step": 2039
+ },
+ {
+ "epoch": 28.335371179039303,
+ "grad_norm": 0.018635908141732216,
+ "learning_rate": 0.0006,
+ "loss": 4.11806058883667,
+ "step": 2040
+ },
+ {
+ "epoch": 28.349344978165938,
+ "grad_norm": 0.019792694598436356,
+ "learning_rate": 0.0006,
+ "loss": 4.335224151611328,
+ "step": 2041
+ },
+ {
+ "epoch": 28.363318777292577,
+ "grad_norm": 0.021731717512011528,
+ "learning_rate": 0.0006,
+ "loss": 4.302778244018555,
+ "step": 2042
+ },
+ {
+ "epoch": 28.377292576419215,
+ "grad_norm": 0.020823447033762932,
+ "learning_rate": 0.0006,
+ "loss": 4.260610580444336,
+ "step": 2043
+ },
+ {
+ "epoch": 28.39126637554585,
+ "grad_norm": 0.019385412335395813,
+ "learning_rate": 0.0006,
+ "loss": 4.2247724533081055,
+ "step": 2044
+ },
+ {
+ "epoch": 28.40524017467249,
+ "grad_norm": 0.019539108499884605,
+ "learning_rate": 0.0006,
+ "loss": 4.236655235290527,
+ "step": 2045
+ },
+ {
+ "epoch": 28.419213973799128,
+ "grad_norm": 0.01942690648138523,
+ "learning_rate": 0.0006,
+ "loss": 4.251850605010986,
+ "step": 2046
+ },
+ {
+ "epoch": 28.433187772925763,
+ "grad_norm": 0.019667886197566986,
+ "learning_rate": 0.0006,
+ "loss": 4.222312927246094,
+ "step": 2047
+ },
+ {
+ "epoch": 28.4471615720524,
+ "grad_norm": 0.02029012329876423,
+ "learning_rate": 0.0006,
+ "loss": 4.246252059936523,
+ "step": 2048
+ },
+ {
+ "epoch": 28.46113537117904,
+ "grad_norm": 0.01784469373524189,
+ "learning_rate": 0.0006,
+ "loss": 4.192128658294678,
+ "step": 2049
+ },
+ {
+ "epoch": 28.475109170305675,
+ "grad_norm": 0.0160287544131279,
+ "learning_rate": 0.0006,
+ "loss": 4.208607196807861,
+ "step": 2050
+ },
+ {
+ "epoch": 28.489082969432314,
+ "grad_norm": 0.015348346903920174,
+ "learning_rate": 0.0006,
+ "loss": 4.236598014831543,
+ "step": 2051
+ },
+ {
+ "epoch": 28.503056768558952,
+ "grad_norm": 0.015198206529021263,
+ "learning_rate": 0.0006,
+ "loss": 4.235552787780762,
+ "step": 2052
+ },
+ {
+ "epoch": 28.51703056768559,
+ "grad_norm": 0.01523754745721817,
+ "learning_rate": 0.0006,
+ "loss": 4.1941728591918945,
+ "step": 2053
+ },
+ {
+ "epoch": 28.531004366812226,
+ "grad_norm": 0.015559614636003971,
+ "learning_rate": 0.0006,
+ "loss": 4.294757843017578,
+ "step": 2054
+ },
+ {
+ "epoch": 28.544978165938865,
+ "grad_norm": 0.01568358950316906,
+ "learning_rate": 0.0006,
+ "loss": 4.213065147399902,
+ "step": 2055
+ },
+ {
+ "epoch": 28.558951965065503,
+ "grad_norm": 0.015336516313254833,
+ "learning_rate": 0.0006,
+ "loss": 4.240293979644775,
+ "step": 2056
+ },
+ {
+ "epoch": 28.57292576419214,
+ "grad_norm": 0.015041496604681015,
+ "learning_rate": 0.0006,
+ "loss": 4.273397445678711,
+ "step": 2057
+ },
+ {
+ "epoch": 28.586899563318777,
+ "grad_norm": 0.01610128954052925,
+ "learning_rate": 0.0006,
+ "loss": 4.125369071960449,
+ "step": 2058
+ },
+ {
+ "epoch": 28.600873362445416,
+ "grad_norm": 0.015974976122379303,
+ "learning_rate": 0.0006,
+ "loss": 4.133459091186523,
+ "step": 2059
+ },
+ {
+ "epoch": 28.61484716157205,
+ "grad_norm": 0.015936799347400665,
+ "learning_rate": 0.0006,
+ "loss": 4.306667327880859,
+ "step": 2060
+ },
+ {
+ "epoch": 28.62882096069869,
+ "grad_norm": 0.01795247197151184,
+ "learning_rate": 0.0006,
+ "loss": 4.205156326293945,
+ "step": 2061
+ },
+ {
+ "epoch": 28.64279475982533,
+ "grad_norm": 0.019639814272522926,
+ "learning_rate": 0.0006,
+ "loss": 4.277614593505859,
+ "step": 2062
+ },
+ {
+ "epoch": 28.656768558951963,
+ "grad_norm": 0.01722300611436367,
+ "learning_rate": 0.0006,
+ "loss": 4.191032886505127,
+ "step": 2063
+ },
+ {
+ "epoch": 28.670742358078602,
+ "grad_norm": 0.01612633652985096,
+ "learning_rate": 0.0006,
+ "loss": 4.20717191696167,
+ "step": 2064
+ },
+ {
+ "epoch": 28.68471615720524,
+ "grad_norm": 0.01603279449045658,
+ "learning_rate": 0.0006,
+ "loss": 4.228884220123291,
+ "step": 2065
+ },
+ {
+ "epoch": 28.69868995633188,
+ "grad_norm": 0.016405927017331123,
+ "learning_rate": 0.0006,
+ "loss": 4.201292991638184,
+ "step": 2066
+ },
+ {
+ "epoch": 28.712663755458514,
+ "grad_norm": 0.016644183546304703,
+ "learning_rate": 0.0006,
+ "loss": 4.239282131195068,
+ "step": 2067
+ },
+ {
+ "epoch": 28.726637554585153,
+ "grad_norm": 0.0175587497651577,
+ "learning_rate": 0.0006,
+ "loss": 4.289038181304932,
+ "step": 2068
+ },
+ {
+ "epoch": 28.74061135371179,
+ "grad_norm": 0.018223397433757782,
+ "learning_rate": 0.0006,
+ "loss": 4.2266998291015625,
+ "step": 2069
+ },
+ {
+ "epoch": 28.754585152838427,
+ "grad_norm": 0.018680868670344353,
+ "learning_rate": 0.0006,
+ "loss": 4.348849296569824,
+ "step": 2070
+ },
+ {
+ "epoch": 28.768558951965066,
+ "grad_norm": 0.016726728528738022,
+ "learning_rate": 0.0006,
+ "loss": 4.292080879211426,
+ "step": 2071
+ },
+ {
+ "epoch": 28.782532751091704,
+ "grad_norm": 0.017386795952916145,
+ "learning_rate": 0.0006,
+ "loss": 4.208870887756348,
+ "step": 2072
+ },
+ {
+ "epoch": 28.79650655021834,
+ "grad_norm": 0.01865178905427456,
+ "learning_rate": 0.0006,
+ "loss": 4.054150581359863,
+ "step": 2073
+ },
+ {
+ "epoch": 28.810480349344978,
+ "grad_norm": 0.01776815392076969,
+ "learning_rate": 0.0006,
+ "loss": 4.202404499053955,
+ "step": 2074
+ },
+ {
+ "epoch": 28.824454148471617,
+ "grad_norm": 0.01643497310578823,
+ "learning_rate": 0.0006,
+ "loss": 4.131855010986328,
+ "step": 2075
+ },
+ {
+ "epoch": 28.83842794759825,
+ "grad_norm": 0.01574273779988289,
+ "learning_rate": 0.0006,
+ "loss": 4.296788692474365,
+ "step": 2076
+ },
+ {
+ "epoch": 28.85240174672489,
+ "grad_norm": 0.015613908879458904,
+ "learning_rate": 0.0006,
+ "loss": 4.210949897766113,
+ "step": 2077
+ },
+ {
+ "epoch": 28.86637554585153,
+ "grad_norm": 0.015543212182819843,
+ "learning_rate": 0.0006,
+ "loss": 4.218136310577393,
+ "step": 2078
+ },
+ {
+ "epoch": 28.880349344978168,
+ "grad_norm": 0.014681251719594002,
+ "learning_rate": 0.0006,
+ "loss": 4.19401741027832,
+ "step": 2079
+ },
+ {
+ "epoch": 28.894323144104803,
+ "grad_norm": 0.014393828809261322,
+ "learning_rate": 0.0006,
+ "loss": 4.198468208312988,
+ "step": 2080
+ },
+ {
+ "epoch": 28.90829694323144,
+ "grad_norm": 0.01580170728266239,
+ "learning_rate": 0.0006,
+ "loss": 4.224969863891602,
+ "step": 2081
+ },
+ {
+ "epoch": 28.92227074235808,
+ "grad_norm": 0.01484165620058775,
+ "learning_rate": 0.0006,
+ "loss": 4.12236213684082,
+ "step": 2082
+ },
+ {
+ "epoch": 28.936244541484715,
+ "grad_norm": 0.01497623696923256,
+ "learning_rate": 0.0006,
+ "loss": 4.24141788482666,
+ "step": 2083
+ },
+ {
+ "epoch": 28.950218340611354,
+ "grad_norm": 0.015080046840012074,
+ "learning_rate": 0.0006,
+ "loss": 4.284151077270508,
+ "step": 2084
+ },
+ {
+ "epoch": 28.964192139737992,
+ "grad_norm": 0.016207212582230568,
+ "learning_rate": 0.0006,
+ "loss": 4.298693656921387,
+ "step": 2085
+ },
+ {
+ "epoch": 28.978165938864628,
+ "grad_norm": 0.015474417246878147,
+ "learning_rate": 0.0006,
+ "loss": 4.2412261962890625,
+ "step": 2086
+ },
+ {
+ "epoch": 28.992139737991266,
+ "grad_norm": 0.01515351701527834,
+ "learning_rate": 0.0006,
+ "loss": 4.3054423332214355,
+ "step": 2087
+ },
+ {
+ "epoch": 29.0,
+ "grad_norm": 0.016411345452070236,
+ "learning_rate": 0.0006,
+ "loss": 4.104830741882324,
+ "step": 2088
+ },
+ {
+ "epoch": 29.0,
+ "eval_loss": 4.661567687988281,
+ "eval_runtime": 57.1244,
+ "eval_samples_per_second": 42.749,
+ "eval_steps_per_second": 1.348,
+ "step": 2088
+ },
+ {
+ "epoch": 29.01397379912664,
+ "grad_norm": 0.01665342226624489,
+ "learning_rate": 0.0006,
+ "loss": 4.311391830444336,
+ "step": 2089
+ },
+ {
+ "epoch": 29.027947598253274,
+ "grad_norm": 0.018053214997053146,
+ "learning_rate": 0.0006,
+ "loss": 4.135597229003906,
+ "step": 2090
+ },
+ {
+ "epoch": 29.041921397379912,
+ "grad_norm": 0.018213748931884766,
+ "learning_rate": 0.0006,
+ "loss": 4.303299903869629,
+ "step": 2091
+ },
+ {
+ "epoch": 29.05589519650655,
+ "grad_norm": 0.018460242077708244,
+ "learning_rate": 0.0006,
+ "loss": 4.20850944519043,
+ "step": 2092
+ },
+ {
+ "epoch": 29.069868995633186,
+ "grad_norm": 0.018365809693932533,
+ "learning_rate": 0.0006,
+ "loss": 4.275847434997559,
+ "step": 2093
+ },
+ {
+ "epoch": 29.083842794759825,
+ "grad_norm": 0.018346186727285385,
+ "learning_rate": 0.0006,
+ "loss": 4.202091217041016,
+ "step": 2094
+ },
+ {
+ "epoch": 29.097816593886463,
+ "grad_norm": 0.017920587211847305,
+ "learning_rate": 0.0006,
+ "loss": 4.182092666625977,
+ "step": 2095
+ },
+ {
+ "epoch": 29.111790393013102,
+ "grad_norm": 0.01812003180384636,
+ "learning_rate": 0.0006,
+ "loss": 4.259721755981445,
+ "step": 2096
+ },
+ {
+ "epoch": 29.125764192139737,
+ "grad_norm": 0.01700986735522747,
+ "learning_rate": 0.0006,
+ "loss": 4.067573070526123,
+ "step": 2097
+ },
+ {
+ "epoch": 29.139737991266376,
+ "grad_norm": 0.018069760873913765,
+ "learning_rate": 0.0006,
+ "loss": 4.234400749206543,
+ "step": 2098
+ },
+ {
+ "epoch": 29.153711790393015,
+ "grad_norm": 0.01905428245663643,
+ "learning_rate": 0.0006,
+ "loss": 4.16609001159668,
+ "step": 2099
+ },
+ {
+ "epoch": 29.16768558951965,
+ "grad_norm": 0.018711242824792862,
+ "learning_rate": 0.0006,
+ "loss": 4.249088764190674,
+ "step": 2100
+ },
+ {
+ "epoch": 29.18165938864629,
+ "grad_norm": 0.018721306696534157,
+ "learning_rate": 0.0006,
+ "loss": 4.14540958404541,
+ "step": 2101
+ },
+ {
+ "epoch": 29.195633187772927,
+ "grad_norm": 0.020139874890446663,
+ "learning_rate": 0.0006,
+ "loss": 4.049248695373535,
+ "step": 2102
+ },
+ {
+ "epoch": 29.209606986899562,
+ "grad_norm": 0.020849574357271194,
+ "learning_rate": 0.0006,
+ "loss": 4.235224723815918,
+ "step": 2103
+ },
+ {
+ "epoch": 29.2235807860262,
+ "grad_norm": 0.018693549558520317,
+ "learning_rate": 0.0006,
+ "loss": 4.232961177825928,
+ "step": 2104
+ },
+ {
+ "epoch": 29.23755458515284,
+ "grad_norm": 0.017889857292175293,
+ "learning_rate": 0.0006,
+ "loss": 4.171422004699707,
+ "step": 2105
+ },
+ {
+ "epoch": 29.251528384279474,
+ "grad_norm": 0.016690224409103394,
+ "learning_rate": 0.0006,
+ "loss": 4.244717597961426,
+ "step": 2106
+ },
+ {
+ "epoch": 29.265502183406113,
+ "grad_norm": 0.01708616502583027,
+ "learning_rate": 0.0006,
+ "loss": 4.125068664550781,
+ "step": 2107
+ },
+ {
+ "epoch": 29.27947598253275,
+ "grad_norm": 0.017928237095475197,
+ "learning_rate": 0.0006,
+ "loss": 4.208292007446289,
+ "step": 2108
+ },
+ {
+ "epoch": 29.29344978165939,
+ "grad_norm": 0.018250394612550735,
+ "learning_rate": 0.0006,
+ "loss": 4.118630886077881,
+ "step": 2109
+ },
+ {
+ "epoch": 29.307423580786025,
+ "grad_norm": 0.020160246640443802,
+ "learning_rate": 0.0006,
+ "loss": 4.206025123596191,
+ "step": 2110
+ },
+ {
+ "epoch": 29.321397379912664,
+ "grad_norm": 0.020045453682541847,
+ "learning_rate": 0.0006,
+ "loss": 4.2758684158325195,
+ "step": 2111
+ },
+ {
+ "epoch": 29.335371179039303,
+ "grad_norm": 0.019556907936930656,
+ "learning_rate": 0.0006,
+ "loss": 4.188272953033447,
+ "step": 2112
+ },
+ {
+ "epoch": 29.349344978165938,
+ "grad_norm": 0.019185200333595276,
+ "learning_rate": 0.0006,
+ "loss": 4.17451286315918,
+ "step": 2113
+ },
+ {
+ "epoch": 29.363318777292577,
+ "grad_norm": 0.01816629432141781,
+ "learning_rate": 0.0006,
+ "loss": 4.165246963500977,
+ "step": 2114
+ },
+ {
+ "epoch": 29.377292576419215,
+ "grad_norm": 0.01865146867930889,
+ "learning_rate": 0.0006,
+ "loss": 4.148184776306152,
+ "step": 2115
+ },
+ {
+ "epoch": 29.39126637554585,
+ "grad_norm": 0.018338464200496674,
+ "learning_rate": 0.0006,
+ "loss": 4.288936614990234,
+ "step": 2116
+ },
+ {
+ "epoch": 29.40524017467249,
+ "grad_norm": 0.018986187875270844,
+ "learning_rate": 0.0006,
+ "loss": 4.183379650115967,
+ "step": 2117
+ },
+ {
+ "epoch": 29.419213973799128,
+ "grad_norm": 0.018409637734293938,
+ "learning_rate": 0.0006,
+ "loss": 4.057574272155762,
+ "step": 2118
+ },
+ {
+ "epoch": 29.433187772925763,
+ "grad_norm": 0.016664542257785797,
+ "learning_rate": 0.0006,
+ "loss": 4.18517541885376,
+ "step": 2119
+ },
+ {
+ "epoch": 29.4471615720524,
+ "grad_norm": 0.016486089676618576,
+ "learning_rate": 0.0006,
+ "loss": 4.179367542266846,
+ "step": 2120
+ },
+ {
+ "epoch": 29.46113537117904,
+ "grad_norm": 0.01787523552775383,
+ "learning_rate": 0.0006,
+ "loss": 4.184126377105713,
+ "step": 2121
+ },
+ {
+ "epoch": 29.475109170305675,
+ "grad_norm": 0.016574613749980927,
+ "learning_rate": 0.0006,
+ "loss": 4.290170192718506,
+ "step": 2122
+ },
+ {
+ "epoch": 29.489082969432314,
+ "grad_norm": 0.01643376611173153,
+ "learning_rate": 0.0006,
+ "loss": 4.052913665771484,
+ "step": 2123
+ },
+ {
+ "epoch": 29.503056768558952,
+ "grad_norm": 0.016503969207406044,
+ "learning_rate": 0.0006,
+ "loss": 4.335184097290039,
+ "step": 2124
+ },
+ {
+ "epoch": 29.51703056768559,
+ "grad_norm": 0.017596479505300522,
+ "learning_rate": 0.0006,
+ "loss": 4.165539741516113,
+ "step": 2125
+ },
+ {
+ "epoch": 29.531004366812226,
+ "grad_norm": 0.017468582838773727,
+ "learning_rate": 0.0006,
+ "loss": 4.315964698791504,
+ "step": 2126
+ },
+ {
+ "epoch": 29.544978165938865,
+ "grad_norm": 0.01724618673324585,
+ "learning_rate": 0.0006,
+ "loss": 4.243681907653809,
+ "step": 2127
+ },
+ {
+ "epoch": 29.558951965065503,
+ "grad_norm": 0.018384616822004318,
+ "learning_rate": 0.0006,
+ "loss": 4.112447261810303,
+ "step": 2128
+ },
+ {
+ "epoch": 29.57292576419214,
+ "grad_norm": 0.01824074238538742,
+ "learning_rate": 0.0006,
+ "loss": 4.236065864562988,
+ "step": 2129
+ },
+ {
+ "epoch": 29.586899563318777,
+ "grad_norm": 0.01625971868634224,
+ "learning_rate": 0.0006,
+ "loss": 4.18143892288208,
+ "step": 2130
+ },
+ {
+ "epoch": 29.600873362445416,
+ "grad_norm": 0.015537494793534279,
+ "learning_rate": 0.0006,
+ "loss": 4.225147247314453,
+ "step": 2131
+ },
+ {
+ "epoch": 29.61484716157205,
+ "grad_norm": 0.015818974003195763,
+ "learning_rate": 0.0006,
+ "loss": 4.3100175857543945,
+ "step": 2132
+ },
+ {
+ "epoch": 29.62882096069869,
+ "grad_norm": 0.015909671783447266,
+ "learning_rate": 0.0006,
+ "loss": 4.169775009155273,
+ "step": 2133
+ },
+ {
+ "epoch": 29.64279475982533,
+ "grad_norm": 0.01769658550620079,
+ "learning_rate": 0.0006,
+ "loss": 4.278904914855957,
+ "step": 2134
+ },
+ {
+ "epoch": 29.656768558951963,
+ "grad_norm": 0.015641704201698303,
+ "learning_rate": 0.0006,
+ "loss": 4.2426228523254395,
+ "step": 2135
+ },
+ {
+ "epoch": 29.670742358078602,
+ "grad_norm": 0.01565658301115036,
+ "learning_rate": 0.0006,
+ "loss": 4.232687950134277,
+ "step": 2136
+ },
+ {
+ "epoch": 29.68471615720524,
+ "grad_norm": 0.016485940665006638,
+ "learning_rate": 0.0006,
+ "loss": 4.209839820861816,
+ "step": 2137
+ },
+ {
+ "epoch": 29.69868995633188,
+ "grad_norm": 0.017622729763388634,
+ "learning_rate": 0.0006,
+ "loss": 4.196017265319824,
+ "step": 2138
+ },
+ {
+ "epoch": 29.712663755458514,
+ "grad_norm": 0.018320564180612564,
+ "learning_rate": 0.0006,
+ "loss": 4.225683689117432,
+ "step": 2139
+ },
+ {
+ "epoch": 29.726637554585153,
+ "grad_norm": 0.018140794709324837,
+ "learning_rate": 0.0006,
+ "loss": 4.311944961547852,
+ "step": 2140
+ },
+ {
+ "epoch": 29.74061135371179,
+ "grad_norm": 0.016227100044488907,
+ "learning_rate": 0.0006,
+ "loss": 4.230257511138916,
+ "step": 2141
+ },
+ {
+ "epoch": 29.754585152838427,
+ "grad_norm": 0.015759488567709923,
+ "learning_rate": 0.0006,
+ "loss": 4.119174003601074,
+ "step": 2142
+ },
+ {
+ "epoch": 29.768558951965066,
+ "grad_norm": 0.01757766678929329,
+ "learning_rate": 0.0006,
+ "loss": 4.165826797485352,
+ "step": 2143
+ },
+ {
+ "epoch": 29.782532751091704,
+ "grad_norm": 0.019661923870444298,
+ "learning_rate": 0.0006,
+ "loss": 4.392203330993652,
+ "step": 2144
+ },
+ {
+ "epoch": 29.79650655021834,
+ "grad_norm": 0.019478755071759224,
+ "learning_rate": 0.0006,
+ "loss": 4.348371982574463,
+ "step": 2145
+ },
+ {
+ "epoch": 29.810480349344978,
+ "grad_norm": 0.019149569794535637,
+ "learning_rate": 0.0006,
+ "loss": 4.34104061126709,
+ "step": 2146
+ },
+ {
+ "epoch": 29.824454148471617,
+ "grad_norm": 0.017707521095871925,
+ "learning_rate": 0.0006,
+ "loss": 4.228226661682129,
+ "step": 2147
+ },
+ {
+ "epoch": 29.83842794759825,
+ "grad_norm": 0.015615378506481647,
+ "learning_rate": 0.0006,
+ "loss": 4.175541877746582,
+ "step": 2148
+ },
+ {
+ "epoch": 29.85240174672489,
+ "grad_norm": 0.016111129894852638,
+ "learning_rate": 0.0006,
+ "loss": 4.267936706542969,
+ "step": 2149
+ },
+ {
+ "epoch": 29.86637554585153,
+ "grad_norm": 0.01528779324144125,
+ "learning_rate": 0.0006,
+ "loss": 4.186014175415039,
+ "step": 2150
+ },
+ {
+ "epoch": 29.880349344978168,
+ "grad_norm": 0.014925646595656872,
+ "learning_rate": 0.0006,
+ "loss": 4.174860000610352,
+ "step": 2151
+ },
+ {
+ "epoch": 29.894323144104803,
+ "grad_norm": 0.016917673870921135,
+ "learning_rate": 0.0006,
+ "loss": 4.33908748626709,
+ "step": 2152
+ },
+ {
+ "epoch": 29.90829694323144,
+ "grad_norm": 0.016007075086236,
+ "learning_rate": 0.0006,
+ "loss": 4.266141414642334,
+ "step": 2153
+ },
+ {
+ "epoch": 29.92227074235808,
+ "grad_norm": 0.016178956255316734,
+ "learning_rate": 0.0006,
+ "loss": 4.271215915679932,
+ "step": 2154
+ },
+ {
+ "epoch": 29.936244541484715,
+ "grad_norm": 0.016180871054530144,
+ "learning_rate": 0.0006,
+ "loss": 4.278801441192627,
+ "step": 2155
+ },
+ {
+ "epoch": 29.950218340611354,
+ "grad_norm": 0.015335503034293652,
+ "learning_rate": 0.0006,
+ "loss": 4.194978713989258,
+ "step": 2156
+ },
+ {
+ "epoch": 29.964192139737992,
+ "grad_norm": 0.016580764204263687,
+ "learning_rate": 0.0006,
+ "loss": 4.290809631347656,
+ "step": 2157
+ },
+ {
+ "epoch": 29.978165938864628,
+ "grad_norm": 0.01623128168284893,
+ "learning_rate": 0.0006,
+ "loss": 4.205451488494873,
+ "step": 2158
+ },
+ {
+ "epoch": 29.992139737991266,
+ "grad_norm": 0.014904248528182507,
+ "learning_rate": 0.0006,
+ "loss": 4.24924373626709,
+ "step": 2159
+ },
+ {
+ "epoch": 30.0,
+ "grad_norm": 0.017089299857616425,
+ "learning_rate": 0.0006,
+ "loss": 4.222072601318359,
+ "step": 2160
+ },
+ {
+ "epoch": 30.0,
+ "eval_loss": 4.585411071777344,
+ "eval_runtime": 57.1164,
+ "eval_samples_per_second": 42.755,
+ "eval_steps_per_second": 1.348,
+ "step": 2160
+ },
+ {
+ "epoch": 30.01397379912664,
+ "grad_norm": 0.017226383090019226,
+ "learning_rate": 0.0006,
+ "loss": 4.19467830657959,
+ "step": 2161
+ },
+ {
+ "epoch": 30.027947598253274,
+ "grad_norm": 0.01635066606104374,
+ "learning_rate": 0.0006,
+ "loss": 4.176875591278076,
+ "step": 2162
+ },
+ {
+ "epoch": 30.041921397379912,
+ "grad_norm": 0.01735362410545349,
+ "learning_rate": 0.0006,
+ "loss": 4.13667106628418,
+ "step": 2163
+ },
+ {
+ "epoch": 30.05589519650655,
+ "grad_norm": 0.017047109082341194,
+ "learning_rate": 0.0006,
+ "loss": 4.136329650878906,
+ "step": 2164
+ },
+ {
+ "epoch": 30.069868995633186,
+ "grad_norm": 0.018116053193807602,
+ "learning_rate": 0.0006,
+ "loss": 4.228706359863281,
+ "step": 2165
+ },
+ {
+ "epoch": 30.083842794759825,
+ "grad_norm": 0.01791389286518097,
+ "learning_rate": 0.0006,
+ "loss": 4.1597747802734375,
+ "step": 2166
+ },
+ {
+ "epoch": 30.097816593886463,
+ "grad_norm": 0.017793502658605576,
+ "learning_rate": 0.0006,
+ "loss": 4.0807976722717285,
+ "step": 2167
+ },
+ {
+ "epoch": 30.111790393013102,
+ "grad_norm": 0.01827503927052021,
+ "learning_rate": 0.0006,
+ "loss": 4.170646667480469,
+ "step": 2168
+ },
+ {
+ "epoch": 30.125764192139737,
+ "grad_norm": 0.017228564247488976,
+ "learning_rate": 0.0006,
+ "loss": 4.146934509277344,
+ "step": 2169
+ },
+ {
+ "epoch": 30.139737991266376,
+ "grad_norm": 0.01757437363266945,
+ "learning_rate": 0.0006,
+ "loss": 4.1859130859375,
+ "step": 2170
+ },
+ {
+ "epoch": 30.153711790393015,
+ "grad_norm": 0.01701093092560768,
+ "learning_rate": 0.0006,
+ "loss": 4.126981258392334,
+ "step": 2171
+ },
+ {
+ "epoch": 30.16768558951965,
+ "grad_norm": 0.015220489352941513,
+ "learning_rate": 0.0006,
+ "loss": 4.164018630981445,
+ "step": 2172
+ },
+ {
+ "epoch": 30.18165938864629,
+ "grad_norm": 0.016254238784313202,
+ "learning_rate": 0.0006,
+ "loss": 4.258275508880615,
+ "step": 2173
+ },
+ {
+ "epoch": 30.195633187772927,
+ "grad_norm": 0.016609614714980125,
+ "learning_rate": 0.0006,
+ "loss": 4.222233772277832,
+ "step": 2174
+ },
+ {
+ "epoch": 30.209606986899562,
+ "grad_norm": 0.016531318426132202,
+ "learning_rate": 0.0006,
+ "loss": 4.144489288330078,
+ "step": 2175
+ },
+ {
+ "epoch": 30.2235807860262,
+ "grad_norm": 0.01567668654024601,
+ "learning_rate": 0.0006,
+ "loss": 4.181916236877441,
+ "step": 2176
+ },
+ {
+ "epoch": 30.23755458515284,
+ "grad_norm": 0.017195681110024452,
+ "learning_rate": 0.0006,
+ "loss": 4.189693927764893,
+ "step": 2177
+ },
+ {
+ "epoch": 30.251528384279474,
+ "grad_norm": 0.016882948577404022,
+ "learning_rate": 0.0006,
+ "loss": 4.21604061126709,
+ "step": 2178
+ },
+ {
+ "epoch": 30.265502183406113,
+ "grad_norm": 0.016261622309684753,
+ "learning_rate": 0.0006,
+ "loss": 4.227639198303223,
+ "step": 2179
+ },
+ {
+ "epoch": 30.27947598253275,
+ "grad_norm": 0.016411812976002693,
+ "learning_rate": 0.0006,
+ "loss": 4.130362510681152,
+ "step": 2180
+ },
+ {
+ "epoch": 30.29344978165939,
+ "grad_norm": 0.01732019893825054,
+ "learning_rate": 0.0006,
+ "loss": 4.230443000793457,
+ "step": 2181
+ },
+ {
+ "epoch": 30.307423580786025,
+ "grad_norm": 0.01825905591249466,
+ "learning_rate": 0.0006,
+ "loss": 4.175817489624023,
+ "step": 2182
+ },
+ {
+ "epoch": 30.321397379912664,
+ "grad_norm": 0.01887713558971882,
+ "learning_rate": 0.0006,
+ "loss": 4.169020652770996,
+ "step": 2183
+ },
+ {
+ "epoch": 30.335371179039303,
+ "grad_norm": 0.019391661509871483,
+ "learning_rate": 0.0006,
+ "loss": 4.176318645477295,
+ "step": 2184
+ },
+ {
+ "epoch": 30.349344978165938,
+ "grad_norm": 0.020840002223849297,
+ "learning_rate": 0.0006,
+ "loss": 4.174228668212891,
+ "step": 2185
+ },
+ {
+ "epoch": 30.363318777292577,
+ "grad_norm": 0.020193178206682205,
+ "learning_rate": 0.0006,
+ "loss": 4.127919673919678,
+ "step": 2186
+ },
+ {
+ "epoch": 30.377292576419215,
+ "grad_norm": 0.018449561670422554,
+ "learning_rate": 0.0006,
+ "loss": 4.229487895965576,
+ "step": 2187
+ },
+ {
+ "epoch": 30.39126637554585,
+ "grad_norm": 0.01711983233690262,
+ "learning_rate": 0.0006,
+ "loss": 4.201651573181152,
+ "step": 2188
+ },
+ {
+ "epoch": 30.40524017467249,
+ "grad_norm": 0.017293395474553108,
+ "learning_rate": 0.0006,
+ "loss": 4.168704032897949,
+ "step": 2189
+ },
+ {
+ "epoch": 30.419213973799128,
+ "grad_norm": 0.018526358529925346,
+ "learning_rate": 0.0006,
+ "loss": 4.171864032745361,
+ "step": 2190
+ },
+ {
+ "epoch": 30.433187772925763,
+ "grad_norm": 0.019287291914224625,
+ "learning_rate": 0.0006,
+ "loss": 4.039534568786621,
+ "step": 2191
+ },
+ {
+ "epoch": 30.4471615720524,
+ "grad_norm": 0.016902444884181023,
+ "learning_rate": 0.0006,
+ "loss": 4.282787799835205,
+ "step": 2192
+ },
+ {
+ "epoch": 30.46113537117904,
+ "grad_norm": 0.01656750962138176,
+ "learning_rate": 0.0006,
+ "loss": 4.213098526000977,
+ "step": 2193
+ },
+ {
+ "epoch": 30.475109170305675,
+ "grad_norm": 0.016805065795779228,
+ "learning_rate": 0.0006,
+ "loss": 4.1583476066589355,
+ "step": 2194
+ },
+ {
+ "epoch": 30.489082969432314,
+ "grad_norm": 0.017166638746857643,
+ "learning_rate": 0.0006,
+ "loss": 4.188634395599365,
+ "step": 2195
+ },
+ {
+ "epoch": 30.503056768558952,
+ "grad_norm": 0.01592106744647026,
+ "learning_rate": 0.0006,
+ "loss": 4.234043121337891,
+ "step": 2196
+ },
+ {
+ "epoch": 30.51703056768559,
+ "grad_norm": 0.015270989388227463,
+ "learning_rate": 0.0006,
+ "loss": 4.301420211791992,
+ "step": 2197
+ },
+ {
+ "epoch": 30.531004366812226,
+ "grad_norm": 0.015194724313914776,
+ "learning_rate": 0.0006,
+ "loss": 4.18703031539917,
+ "step": 2198
+ },
+ {
+ "epoch": 30.544978165938865,
+ "grad_norm": 0.015587746165692806,
+ "learning_rate": 0.0006,
+ "loss": 4.138928413391113,
+ "step": 2199
+ },
+ {
+ "epoch": 30.558951965065503,
+ "grad_norm": 0.016029570251703262,
+ "learning_rate": 0.0006,
+ "loss": 4.290863990783691,
+ "step": 2200
+ },
+ {
+ "epoch": 30.57292576419214,
+ "grad_norm": 0.015651309862732887,
+ "learning_rate": 0.0006,
+ "loss": 4.272062301635742,
+ "step": 2201
+ },
+ {
+ "epoch": 30.586899563318777,
+ "grad_norm": 0.015992306172847748,
+ "learning_rate": 0.0006,
+ "loss": 4.2487406730651855,
+ "step": 2202
+ },
+ {
+ "epoch": 30.600873362445416,
+ "grad_norm": 0.014969157986342907,
+ "learning_rate": 0.0006,
+ "loss": 4.115347862243652,
+ "step": 2203
+ },
+ {
+ "epoch": 30.61484716157205,
+ "grad_norm": 0.014639027416706085,
+ "learning_rate": 0.0006,
+ "loss": 4.2207489013671875,
+ "step": 2204
+ },
+ {
+ "epoch": 30.62882096069869,
+ "grad_norm": 0.015059413388371468,
+ "learning_rate": 0.0006,
+ "loss": 4.274044990539551,
+ "step": 2205
+ },
+ {
+ "epoch": 30.64279475982533,
+ "grad_norm": 0.0166641678661108,
+ "learning_rate": 0.0006,
+ "loss": 4.191373348236084,
+ "step": 2206
+ },
+ {
+ "epoch": 30.656768558951963,
+ "grad_norm": 0.017806345596909523,
+ "learning_rate": 0.0006,
+ "loss": 4.139800548553467,
+ "step": 2207
+ },
+ {
+ "epoch": 30.670742358078602,
+ "grad_norm": 0.018712421879172325,
+ "learning_rate": 0.0006,
+ "loss": 4.035965442657471,
+ "step": 2208
+ },
+ {
+ "epoch": 30.68471615720524,
+ "grad_norm": 0.018457243219017982,
+ "learning_rate": 0.0006,
+ "loss": 4.2579240798950195,
+ "step": 2209
+ },
+ {
+ "epoch": 30.69868995633188,
+ "grad_norm": 0.01920999586582184,
+ "learning_rate": 0.0006,
+ "loss": 4.131112098693848,
+ "step": 2210
+ },
+ {
+ "epoch": 30.712663755458514,
+ "grad_norm": 0.01729944534599781,
+ "learning_rate": 0.0006,
+ "loss": 4.254059314727783,
+ "step": 2211
+ },
+ {
+ "epoch": 30.726637554585153,
+ "grad_norm": 0.016914231702685356,
+ "learning_rate": 0.0006,
+ "loss": 4.233428955078125,
+ "step": 2212
+ },
+ {
+ "epoch": 30.74061135371179,
+ "grad_norm": 0.01913098618388176,
+ "learning_rate": 0.0006,
+ "loss": 4.23525333404541,
+ "step": 2213
+ },
+ {
+ "epoch": 30.754585152838427,
+ "grad_norm": 0.019150281324982643,
+ "learning_rate": 0.0006,
+ "loss": 4.213695049285889,
+ "step": 2214
+ },
+ {
+ "epoch": 30.768558951965066,
+ "grad_norm": 0.018947452306747437,
+ "learning_rate": 0.0006,
+ "loss": 4.151697158813477,
+ "step": 2215
+ },
+ {
+ "epoch": 30.782532751091704,
+ "grad_norm": 0.016058508306741714,
+ "learning_rate": 0.0006,
+ "loss": 4.220280170440674,
+ "step": 2216
+ },
+ {
+ "epoch": 30.79650655021834,
+ "grad_norm": 0.017488619312644005,
+ "learning_rate": 0.0006,
+ "loss": 4.154428482055664,
+ "step": 2217
+ },
+ {
+ "epoch": 30.810480349344978,
+ "grad_norm": 0.017355090007185936,
+ "learning_rate": 0.0006,
+ "loss": 4.256086349487305,
+ "step": 2218
+ },
+ {
+ "epoch": 30.824454148471617,
+ "grad_norm": 0.016636032611131668,
+ "learning_rate": 0.0006,
+ "loss": 4.227564334869385,
+ "step": 2219
+ },
+ {
+ "epoch": 30.83842794759825,
+ "grad_norm": 0.016273891553282738,
+ "learning_rate": 0.0006,
+ "loss": 4.200466156005859,
+ "step": 2220
+ },
+ {
+ "epoch": 30.85240174672489,
+ "grad_norm": 0.016939911991357803,
+ "learning_rate": 0.0006,
+ "loss": 4.1571221351623535,
+ "step": 2221
+ },
+ {
+ "epoch": 30.86637554585153,
+ "grad_norm": 0.018332522362470627,
+ "learning_rate": 0.0006,
+ "loss": 4.228398323059082,
+ "step": 2222
+ },
+ {
+ "epoch": 30.880349344978168,
+ "grad_norm": 0.019508114084601402,
+ "learning_rate": 0.0006,
+ "loss": 4.222960472106934,
+ "step": 2223
+ },
+ {
+ "epoch": 30.894323144104803,
+ "grad_norm": 0.018440278246998787,
+ "learning_rate": 0.0006,
+ "loss": 4.261051177978516,
+ "step": 2224
+ },
+ {
+ "epoch": 30.90829694323144,
+ "grad_norm": 0.018113229423761368,
+ "learning_rate": 0.0006,
+ "loss": 4.109235763549805,
+ "step": 2225
+ },
+ {
+ "epoch": 30.92227074235808,
+ "grad_norm": 0.015392083674669266,
+ "learning_rate": 0.0006,
+ "loss": 4.21950101852417,
+ "step": 2226
+ },
+ {
+ "epoch": 30.936244541484715,
+ "grad_norm": 0.01633290946483612,
+ "learning_rate": 0.0006,
+ "loss": 4.176053047180176,
+ "step": 2227
+ },
+ {
+ "epoch": 30.950218340611354,
+ "grad_norm": 0.016236383467912674,
+ "learning_rate": 0.0006,
+ "loss": 4.314309120178223,
+ "step": 2228
+ },
+ {
+ "epoch": 30.964192139737992,
+ "grad_norm": 0.015284373424947262,
+ "learning_rate": 0.0006,
+ "loss": 4.303133487701416,
+ "step": 2229
+ },
+ {
+ "epoch": 30.978165938864628,
+ "grad_norm": 0.015982897952198982,
+ "learning_rate": 0.0006,
+ "loss": 4.25739860534668,
+ "step": 2230
+ },
+ {
+ "epoch": 30.992139737991266,
+ "grad_norm": 0.016682198271155357,
+ "learning_rate": 0.0006,
+ "loss": 4.292210578918457,
+ "step": 2231
+ },
+ {
+ "epoch": 31.0,
+ "grad_norm": 0.017521467059850693,
+ "learning_rate": 0.0006,
+ "loss": 4.170637130737305,
+ "step": 2232
+ },
+ {
+ "epoch": 31.0,
+ "eval_loss": 4.642786979675293,
+ "eval_runtime": 57.029,
+ "eval_samples_per_second": 42.82,
+ "eval_steps_per_second": 1.35,
+ "step": 2232
+ },
+ {
+ "epoch": 31.01397379912664,
+ "grad_norm": 0.016607852652668953,
+ "learning_rate": 0.0006,
+ "loss": 4.106224536895752,
+ "step": 2233
+ },
+ {
+ "epoch": 31.027947598253274,
+ "grad_norm": 0.018448904156684875,
+ "learning_rate": 0.0006,
+ "loss": 4.224459648132324,
+ "step": 2234
+ },
+ {
+ "epoch": 31.041921397379912,
+ "grad_norm": 0.020481228828430176,
+ "learning_rate": 0.0006,
+ "loss": 4.305548667907715,
+ "step": 2235
+ },
+ {
+ "epoch": 31.05589519650655,
+ "grad_norm": 0.01999707892537117,
+ "learning_rate": 0.0006,
+ "loss": 4.253663539886475,
+ "step": 2236
+ },
+ {
+ "epoch": 31.069868995633186,
+ "grad_norm": 0.020428303629159927,
+ "learning_rate": 0.0006,
+ "loss": 4.166954040527344,
+ "step": 2237
+ },
+ {
+ "epoch": 31.083842794759825,
+ "grad_norm": 0.022608213126659393,
+ "learning_rate": 0.0006,
+ "loss": 4.172819137573242,
+ "step": 2238
+ },
+ {
+ "epoch": 31.097816593886463,
+ "grad_norm": 0.024669520556926727,
+ "learning_rate": 0.0006,
+ "loss": 4.106586456298828,
+ "step": 2239
+ },
+ {
+ "epoch": 31.111790393013102,
+ "grad_norm": 0.02247351035475731,
+ "learning_rate": 0.0006,
+ "loss": 4.160271644592285,
+ "step": 2240
+ },
+ {
+ "epoch": 31.125764192139737,
+ "grad_norm": 0.0207810141146183,
+ "learning_rate": 0.0006,
+ "loss": 4.272491931915283,
+ "step": 2241
+ },
+ {
+ "epoch": 31.139737991266376,
+ "grad_norm": 0.021576499566435814,
+ "learning_rate": 0.0006,
+ "loss": 4.171474456787109,
+ "step": 2242
+ },
+ {
+ "epoch": 31.153711790393015,
+ "grad_norm": 0.02337106689810753,
+ "learning_rate": 0.0006,
+ "loss": 4.094712734222412,
+ "step": 2243
+ },
+ {
+ "epoch": 31.16768558951965,
+ "grad_norm": 0.02219875156879425,
+ "learning_rate": 0.0006,
+ "loss": 4.1800994873046875,
+ "step": 2244
+ },
+ {
+ "epoch": 31.18165938864629,
+ "grad_norm": 0.022680504247546196,
+ "learning_rate": 0.0006,
+ "loss": 4.2054362297058105,
+ "step": 2245
+ },
+ {
+ "epoch": 31.195633187772927,
+ "grad_norm": 0.0208734143525362,
+ "learning_rate": 0.0006,
+ "loss": 4.06068229675293,
+ "step": 2246
+ },
+ {
+ "epoch": 31.209606986899562,
+ "grad_norm": 0.01811066083610058,
+ "learning_rate": 0.0006,
+ "loss": 4.2465128898620605,
+ "step": 2247
+ },
+ {
+ "epoch": 31.2235807860262,
+ "grad_norm": 0.018571637570858,
+ "learning_rate": 0.0006,
+ "loss": 4.191955089569092,
+ "step": 2248
+ },
+ {
+ "epoch": 31.23755458515284,
+ "grad_norm": 0.018598807975649834,
+ "learning_rate": 0.0006,
+ "loss": 4.200872421264648,
+ "step": 2249
+ },
+ {
+ "epoch": 31.251528384279474,
+ "grad_norm": 0.017577625811100006,
+ "learning_rate": 0.0006,
+ "loss": 4.148700714111328,
+ "step": 2250
+ },
+ {
+ "epoch": 31.265502183406113,
+ "grad_norm": 0.01817980408668518,
+ "learning_rate": 0.0006,
+ "loss": 4.256978988647461,
+ "step": 2251
+ },
+ {
+ "epoch": 31.27947598253275,
+ "grad_norm": 0.017983486875891685,
+ "learning_rate": 0.0006,
+ "loss": 4.175005912780762,
+ "step": 2252
+ },
+ {
+ "epoch": 31.29344978165939,
+ "grad_norm": 0.016751961782574654,
+ "learning_rate": 0.0006,
+ "loss": 4.237518310546875,
+ "step": 2253
+ },
+ {
+ "epoch": 31.307423580786025,
+ "grad_norm": 0.016407202929258347,
+ "learning_rate": 0.0006,
+ "loss": 4.197269439697266,
+ "step": 2254
+ },
+ {
+ "epoch": 31.321397379912664,
+ "grad_norm": 0.01578645594418049,
+ "learning_rate": 0.0006,
+ "loss": 4.179751396179199,
+ "step": 2255
+ },
+ {
+ "epoch": 31.335371179039303,
+ "grad_norm": 0.016001226380467415,
+ "learning_rate": 0.0006,
+ "loss": 4.08821964263916,
+ "step": 2256
+ },
+ {
+ "epoch": 31.349344978165938,
+ "grad_norm": 0.015122811309993267,
+ "learning_rate": 0.0006,
+ "loss": 4.047828674316406,
+ "step": 2257
+ },
+ {
+ "epoch": 31.363318777292577,
+ "grad_norm": 0.015399965457618237,
+ "learning_rate": 0.0006,
+ "loss": 4.1506524085998535,
+ "step": 2258
+ },
+ {
+ "epoch": 31.377292576419215,
+ "grad_norm": 0.01534605398774147,
+ "learning_rate": 0.0006,
+ "loss": 4.27952766418457,
+ "step": 2259
+ },
+ {
+ "epoch": 31.39126637554585,
+ "grad_norm": 0.015461381524801254,
+ "learning_rate": 0.0006,
+ "loss": 4.27102518081665,
+ "step": 2260
+ },
+ {
+ "epoch": 31.40524017467249,
+ "grad_norm": 0.016769153997302055,
+ "learning_rate": 0.0006,
+ "loss": 4.191592216491699,
+ "step": 2261
+ },
+ {
+ "epoch": 31.419213973799128,
+ "grad_norm": 0.016557959839701653,
+ "learning_rate": 0.0006,
+ "loss": 4.159705638885498,
+ "step": 2262
+ },
+ {
+ "epoch": 31.433187772925763,
+ "grad_norm": 0.01659569889307022,
+ "learning_rate": 0.0006,
+ "loss": 4.2136664390563965,
+ "step": 2263
+ },
+ {
+ "epoch": 31.4471615720524,
+ "grad_norm": 0.017101937904953957,
+ "learning_rate": 0.0006,
+ "loss": 4.2716498374938965,
+ "step": 2264
+ },
+ {
+ "epoch": 31.46113537117904,
+ "grad_norm": 0.01722985878586769,
+ "learning_rate": 0.0006,
+ "loss": 4.300534248352051,
+ "step": 2265
+ },
+ {
+ "epoch": 31.475109170305675,
+ "grad_norm": 0.016800185665488243,
+ "learning_rate": 0.0006,
+ "loss": 4.113240718841553,
+ "step": 2266
+ },
+ {
+ "epoch": 31.489082969432314,
+ "grad_norm": 0.01603875495493412,
+ "learning_rate": 0.0006,
+ "loss": 4.157105445861816,
+ "step": 2267
+ },
+ {
+ "epoch": 31.503056768558952,
+ "grad_norm": 0.015903932973742485,
+ "learning_rate": 0.0006,
+ "loss": 4.233669757843018,
+ "step": 2268
+ },
+ {
+ "epoch": 31.51703056768559,
+ "grad_norm": 0.0160983856767416,
+ "learning_rate": 0.0006,
+ "loss": 4.147704124450684,
+ "step": 2269
+ },
+ {
+ "epoch": 31.531004366812226,
+ "grad_norm": 0.01604730263352394,
+ "learning_rate": 0.0006,
+ "loss": 4.1465911865234375,
+ "step": 2270
+ },
+ {
+ "epoch": 31.544978165938865,
+ "grad_norm": 0.016205785796046257,
+ "learning_rate": 0.0006,
+ "loss": 4.106704235076904,
+ "step": 2271
+ },
+ {
+ "epoch": 31.558951965065503,
+ "grad_norm": 0.017008036375045776,
+ "learning_rate": 0.0006,
+ "loss": 4.191000938415527,
+ "step": 2272
+ },
+ {
+ "epoch": 31.57292576419214,
+ "grad_norm": 0.01699635200202465,
+ "learning_rate": 0.0006,
+ "loss": 4.187042236328125,
+ "step": 2273
+ },
+ {
+ "epoch": 31.586899563318777,
+ "grad_norm": 0.01692967116832733,
+ "learning_rate": 0.0006,
+ "loss": 4.208805561065674,
+ "step": 2274
+ },
+ {
+ "epoch": 31.600873362445416,
+ "grad_norm": 0.017608756199479103,
+ "learning_rate": 0.0006,
+ "loss": 4.200512886047363,
+ "step": 2275
+ },
+ {
+ "epoch": 31.61484716157205,
+ "grad_norm": 0.016590990126132965,
+ "learning_rate": 0.0006,
+ "loss": 4.194705009460449,
+ "step": 2276
+ },
+ {
+ "epoch": 31.62882096069869,
+ "grad_norm": 0.0170235987752676,
+ "learning_rate": 0.0006,
+ "loss": 4.210326194763184,
+ "step": 2277
+ },
+ {
+ "epoch": 31.64279475982533,
+ "grad_norm": 0.01604801043868065,
+ "learning_rate": 0.0006,
+ "loss": 4.196024417877197,
+ "step": 2278
+ },
+ {
+ "epoch": 31.656768558951963,
+ "grad_norm": 0.015694202855229378,
+ "learning_rate": 0.0006,
+ "loss": 4.178236484527588,
+ "step": 2279
+ },
+ {
+ "epoch": 31.670742358078602,
+ "grad_norm": 0.015473959036171436,
+ "learning_rate": 0.0006,
+ "loss": 4.259339809417725,
+ "step": 2280
+ },
+ {
+ "epoch": 31.68471615720524,
+ "grad_norm": 0.014689362607896328,
+ "learning_rate": 0.0006,
+ "loss": 3.9702324867248535,
+ "step": 2281
+ },
+ {
+ "epoch": 31.69868995633188,
+ "grad_norm": 0.015245123766362667,
+ "learning_rate": 0.0006,
+ "loss": 4.197025775909424,
+ "step": 2282
+ },
+ {
+ "epoch": 31.712663755458514,
+ "grad_norm": 0.016148347407579422,
+ "learning_rate": 0.0006,
+ "loss": 4.231856346130371,
+ "step": 2283
+ },
+ {
+ "epoch": 31.726637554585153,
+ "grad_norm": 0.015169818885624409,
+ "learning_rate": 0.0006,
+ "loss": 4.230169296264648,
+ "step": 2284
+ },
+ {
+ "epoch": 31.74061135371179,
+ "grad_norm": 0.014843013137578964,
+ "learning_rate": 0.0006,
+ "loss": 4.195633411407471,
+ "step": 2285
+ },
+ {
+ "epoch": 31.754585152838427,
+ "grad_norm": 0.015229340642690659,
+ "learning_rate": 0.0006,
+ "loss": 4.231280326843262,
+ "step": 2286
+ },
+ {
+ "epoch": 31.768558951965066,
+ "grad_norm": 0.01551708485931158,
+ "learning_rate": 0.0006,
+ "loss": 4.17064094543457,
+ "step": 2287
+ },
+ {
+ "epoch": 31.782532751091704,
+ "grad_norm": 0.015863575041294098,
+ "learning_rate": 0.0006,
+ "loss": 4.1579484939575195,
+ "step": 2288
+ },
+ {
+ "epoch": 31.79650655021834,
+ "grad_norm": 0.015118198469281197,
+ "learning_rate": 0.0006,
+ "loss": 4.1626787185668945,
+ "step": 2289
+ },
+ {
+ "epoch": 31.810480349344978,
+ "grad_norm": 0.015220776200294495,
+ "learning_rate": 0.0006,
+ "loss": 4.135746955871582,
+ "step": 2290
+ },
+ {
+ "epoch": 31.824454148471617,
+ "grad_norm": 0.014529784210026264,
+ "learning_rate": 0.0006,
+ "loss": 4.164931297302246,
+ "step": 2291
+ },
+ {
+ "epoch": 31.83842794759825,
+ "grad_norm": 0.015533296391367912,
+ "learning_rate": 0.0006,
+ "loss": 4.112464904785156,
+ "step": 2292
+ },
+ {
+ "epoch": 31.85240174672489,
+ "grad_norm": 0.016562335193157196,
+ "learning_rate": 0.0006,
+ "loss": 4.164898872375488,
+ "step": 2293
+ },
+ {
+ "epoch": 31.86637554585153,
+ "grad_norm": 0.016258113086223602,
+ "learning_rate": 0.0006,
+ "loss": 4.1979265213012695,
+ "step": 2294
+ },
+ {
+ "epoch": 31.880349344978168,
+ "grad_norm": 0.01558110024780035,
+ "learning_rate": 0.0006,
+ "loss": 4.174675941467285,
+ "step": 2295
+ },
+ {
+ "epoch": 31.894323144104803,
+ "grad_norm": 0.016412515193223953,
+ "learning_rate": 0.0006,
+ "loss": 4.217909336090088,
+ "step": 2296
+ },
+ {
+ "epoch": 31.90829694323144,
+ "grad_norm": 0.015118442475795746,
+ "learning_rate": 0.0006,
+ "loss": 4.137143611907959,
+ "step": 2297
+ },
+ {
+ "epoch": 31.92227074235808,
+ "grad_norm": 0.01506161317229271,
+ "learning_rate": 0.0006,
+ "loss": 4.196122169494629,
+ "step": 2298
+ },
+ {
+ "epoch": 31.936244541484715,
+ "grad_norm": 0.017165271565318108,
+ "learning_rate": 0.0006,
+ "loss": 4.14834451675415,
+ "step": 2299
+ },
+ {
+ "epoch": 31.950218340611354,
+ "grad_norm": 0.01696440577507019,
+ "learning_rate": 0.0006,
+ "loss": 4.218868732452393,
+ "step": 2300
+ },
+ {
+ "epoch": 31.964192139737992,
+ "grad_norm": 0.017349394038319588,
+ "learning_rate": 0.0006,
+ "loss": 4.2754011154174805,
+ "step": 2301
+ },
+ {
+ "epoch": 31.978165938864628,
+ "grad_norm": 0.018387921154499054,
+ "learning_rate": 0.0006,
+ "loss": 4.282623291015625,
+ "step": 2302
+ },
+ {
+ "epoch": 31.992139737991266,
+ "grad_norm": 0.018837926909327507,
+ "learning_rate": 0.0006,
+ "loss": 4.185425281524658,
+ "step": 2303
+ },
+ {
+ "epoch": 32.0,
+ "grad_norm": 0.018998464569449425,
+ "learning_rate": 0.0006,
+ "loss": 4.151291847229004,
+ "step": 2304
+ },
+ {
+ "epoch": 32.0,
+ "eval_loss": 4.522614002227783,
+ "eval_runtime": 57.0106,
+ "eval_samples_per_second": 42.834,
+ "eval_steps_per_second": 1.351,
+ "step": 2304
+ },
+ {
+ "epoch": 32.01397379912664,
+ "grad_norm": 0.019036833196878433,
+ "learning_rate": 0.0006,
+ "loss": 4.183202266693115,
+ "step": 2305
+ },
+ {
+ "epoch": 32.02794759825328,
+ "grad_norm": 0.02207397110760212,
+ "learning_rate": 0.0006,
+ "loss": 4.17875862121582,
+ "step": 2306
+ },
+ {
+ "epoch": 32.041921397379916,
+ "grad_norm": 0.022008279338479042,
+ "learning_rate": 0.0006,
+ "loss": 4.08042049407959,
+ "step": 2307
+ },
+ {
+ "epoch": 32.05589519650655,
+ "grad_norm": 0.022407017648220062,
+ "learning_rate": 0.0006,
+ "loss": 4.112189769744873,
+ "step": 2308
+ },
+ {
+ "epoch": 32.069868995633186,
+ "grad_norm": 0.021599747240543365,
+ "learning_rate": 0.0006,
+ "loss": 4.126455307006836,
+ "step": 2309
+ },
+ {
+ "epoch": 32.083842794759825,
+ "grad_norm": 0.020340966060757637,
+ "learning_rate": 0.0006,
+ "loss": 4.080538272857666,
+ "step": 2310
+ },
+ {
+ "epoch": 32.09781659388646,
+ "grad_norm": 0.020956892520189285,
+ "learning_rate": 0.0006,
+ "loss": 4.171943664550781,
+ "step": 2311
+ },
+ {
+ "epoch": 32.1117903930131,
+ "grad_norm": 0.019966667518019676,
+ "learning_rate": 0.0006,
+ "loss": 4.144007682800293,
+ "step": 2312
+ },
+ {
+ "epoch": 32.12576419213974,
+ "grad_norm": 0.018256952986121178,
+ "learning_rate": 0.0006,
+ "loss": 4.208107948303223,
+ "step": 2313
+ },
+ {
+ "epoch": 32.13973799126637,
+ "grad_norm": 0.017289170995354652,
+ "learning_rate": 0.0006,
+ "loss": 4.182024955749512,
+ "step": 2314
+ },
+ {
+ "epoch": 32.15371179039301,
+ "grad_norm": 0.017656700685620308,
+ "learning_rate": 0.0006,
+ "loss": 4.068373680114746,
+ "step": 2315
+ },
+ {
+ "epoch": 32.16768558951965,
+ "grad_norm": 0.017903361469507217,
+ "learning_rate": 0.0006,
+ "loss": 4.2345428466796875,
+ "step": 2316
+ },
+ {
+ "epoch": 32.18165938864629,
+ "grad_norm": 0.01785109005868435,
+ "learning_rate": 0.0006,
+ "loss": 4.219776153564453,
+ "step": 2317
+ },
+ {
+ "epoch": 32.19563318777293,
+ "grad_norm": 0.01887257769703865,
+ "learning_rate": 0.0006,
+ "loss": 4.139206409454346,
+ "step": 2318
+ },
+ {
+ "epoch": 32.209606986899566,
+ "grad_norm": 0.01924685761332512,
+ "learning_rate": 0.0006,
+ "loss": 4.121052265167236,
+ "step": 2319
+ },
+ {
+ "epoch": 32.223580786026204,
+ "grad_norm": 0.020574072375893593,
+ "learning_rate": 0.0006,
+ "loss": 4.13262414932251,
+ "step": 2320
+ },
+ {
+ "epoch": 32.237554585152836,
+ "grad_norm": 0.023849064484238625,
+ "learning_rate": 0.0006,
+ "loss": 4.239119529724121,
+ "step": 2321
+ },
+ {
+ "epoch": 32.251528384279474,
+ "grad_norm": 0.026239361613988876,
+ "learning_rate": 0.0006,
+ "loss": 4.248993873596191,
+ "step": 2322
+ },
+ {
+ "epoch": 32.26550218340611,
+ "grad_norm": 0.023980243131518364,
+ "learning_rate": 0.0006,
+ "loss": 4.159852504730225,
+ "step": 2323
+ },
+ {
+ "epoch": 32.27947598253275,
+ "grad_norm": 0.020442405715584755,
+ "learning_rate": 0.0006,
+ "loss": 4.133546829223633,
+ "step": 2324
+ },
+ {
+ "epoch": 32.29344978165939,
+ "grad_norm": 0.020661186426877975,
+ "learning_rate": 0.0006,
+ "loss": 4.078243732452393,
+ "step": 2325
+ },
+ {
+ "epoch": 32.30742358078603,
+ "grad_norm": 0.02046387456357479,
+ "learning_rate": 0.0006,
+ "loss": 4.172012805938721,
+ "step": 2326
+ },
+ {
+ "epoch": 32.32139737991266,
+ "grad_norm": 0.02089597098529339,
+ "learning_rate": 0.0006,
+ "loss": 4.0604705810546875,
+ "step": 2327
+ },
+ {
+ "epoch": 32.3353711790393,
+ "grad_norm": 0.01875591278076172,
+ "learning_rate": 0.0006,
+ "loss": 4.125911235809326,
+ "step": 2328
+ },
+ {
+ "epoch": 32.34934497816594,
+ "grad_norm": 0.018606796860694885,
+ "learning_rate": 0.0006,
+ "loss": 4.097362518310547,
+ "step": 2329
+ },
+ {
+ "epoch": 32.36331877729258,
+ "grad_norm": 0.01766626164317131,
+ "learning_rate": 0.0006,
+ "loss": 4.118616580963135,
+ "step": 2330
+ },
+ {
+ "epoch": 32.377292576419215,
+ "grad_norm": 0.017692960798740387,
+ "learning_rate": 0.0006,
+ "loss": 4.1457672119140625,
+ "step": 2331
+ },
+ {
+ "epoch": 32.391266375545854,
+ "grad_norm": 0.017619704827666283,
+ "learning_rate": 0.0006,
+ "loss": 4.175961494445801,
+ "step": 2332
+ },
+ {
+ "epoch": 32.40524017467249,
+ "grad_norm": 0.01784413494169712,
+ "learning_rate": 0.0006,
+ "loss": 4.051291465759277,
+ "step": 2333
+ },
+ {
+ "epoch": 32.419213973799124,
+ "grad_norm": 0.018508171662688255,
+ "learning_rate": 0.0006,
+ "loss": 4.0638556480407715,
+ "step": 2334
+ },
+ {
+ "epoch": 32.43318777292576,
+ "grad_norm": 0.01731436885893345,
+ "learning_rate": 0.0006,
+ "loss": 4.1805419921875,
+ "step": 2335
+ },
+ {
+ "epoch": 32.4471615720524,
+ "grad_norm": 0.015886498615145683,
+ "learning_rate": 0.0006,
+ "loss": 4.151026725769043,
+ "step": 2336
+ },
+ {
+ "epoch": 32.46113537117904,
+ "grad_norm": 0.016583820804953575,
+ "learning_rate": 0.0006,
+ "loss": 4.203604698181152,
+ "step": 2337
+ },
+ {
+ "epoch": 32.47510917030568,
+ "grad_norm": 0.0176097322255373,
+ "learning_rate": 0.0006,
+ "loss": 4.135159015655518,
+ "step": 2338
+ },
+ {
+ "epoch": 32.48908296943232,
+ "grad_norm": 0.016657903790473938,
+ "learning_rate": 0.0006,
+ "loss": 4.126681804656982,
+ "step": 2339
+ },
+ {
+ "epoch": 32.50305676855895,
+ "grad_norm": 0.016866130754351616,
+ "learning_rate": 0.0006,
+ "loss": 4.162341117858887,
+ "step": 2340
+ },
+ {
+ "epoch": 32.51703056768559,
+ "grad_norm": 0.016286680474877357,
+ "learning_rate": 0.0006,
+ "loss": 4.261801242828369,
+ "step": 2341
+ },
+ {
+ "epoch": 32.531004366812226,
+ "grad_norm": 0.01481365505605936,
+ "learning_rate": 0.0006,
+ "loss": 4.110016822814941,
+ "step": 2342
+ },
+ {
+ "epoch": 32.544978165938865,
+ "grad_norm": 0.016605013981461525,
+ "learning_rate": 0.0006,
+ "loss": 4.230024814605713,
+ "step": 2343
+ },
+ {
+ "epoch": 32.5589519650655,
+ "grad_norm": 0.01645175740122795,
+ "learning_rate": 0.0006,
+ "loss": 4.174612045288086,
+ "step": 2344
+ },
+ {
+ "epoch": 32.57292576419214,
+ "grad_norm": 0.014725026674568653,
+ "learning_rate": 0.0006,
+ "loss": 4.066371440887451,
+ "step": 2345
+ },
+ {
+ "epoch": 32.58689956331878,
+ "grad_norm": 0.015003837645053864,
+ "learning_rate": 0.0006,
+ "loss": 4.193343639373779,
+ "step": 2346
+ },
+ {
+ "epoch": 32.60087336244541,
+ "grad_norm": 0.015346302650868893,
+ "learning_rate": 0.0006,
+ "loss": 4.210659980773926,
+ "step": 2347
+ },
+ {
+ "epoch": 32.61484716157205,
+ "grad_norm": 0.015135962516069412,
+ "learning_rate": 0.0006,
+ "loss": 4.204618453979492,
+ "step": 2348
+ },
+ {
+ "epoch": 32.62882096069869,
+ "grad_norm": 0.01601956970989704,
+ "learning_rate": 0.0006,
+ "loss": 4.145564556121826,
+ "step": 2349
+ },
+ {
+ "epoch": 32.64279475982533,
+ "grad_norm": 0.01517851185053587,
+ "learning_rate": 0.0006,
+ "loss": 4.051673889160156,
+ "step": 2350
+ },
+ {
+ "epoch": 32.65676855895197,
+ "grad_norm": 0.014894292689859867,
+ "learning_rate": 0.0006,
+ "loss": 4.184818267822266,
+ "step": 2351
+ },
+ {
+ "epoch": 32.670742358078606,
+ "grad_norm": 0.015720834955573082,
+ "learning_rate": 0.0006,
+ "loss": 4.220036506652832,
+ "step": 2352
+ },
+ {
+ "epoch": 32.68471615720524,
+ "grad_norm": 0.01573433168232441,
+ "learning_rate": 0.0006,
+ "loss": 4.076773166656494,
+ "step": 2353
+ },
+ {
+ "epoch": 32.698689956331876,
+ "grad_norm": 0.015828052535653114,
+ "learning_rate": 0.0006,
+ "loss": 4.116329669952393,
+ "step": 2354
+ },
+ {
+ "epoch": 32.712663755458514,
+ "grad_norm": 0.015462259761989117,
+ "learning_rate": 0.0006,
+ "loss": 4.1155242919921875,
+ "step": 2355
+ },
+ {
+ "epoch": 32.72663755458515,
+ "grad_norm": 0.015763849020004272,
+ "learning_rate": 0.0006,
+ "loss": 4.169687747955322,
+ "step": 2356
+ },
+ {
+ "epoch": 32.74061135371179,
+ "grad_norm": 0.014935145154595375,
+ "learning_rate": 0.0006,
+ "loss": 4.2090911865234375,
+ "step": 2357
+ },
+ {
+ "epoch": 32.75458515283843,
+ "grad_norm": 0.015119184739887714,
+ "learning_rate": 0.0006,
+ "loss": 4.131838798522949,
+ "step": 2358
+ },
+ {
+ "epoch": 32.76855895196506,
+ "grad_norm": 0.014833525754511356,
+ "learning_rate": 0.0006,
+ "loss": 4.17618465423584,
+ "step": 2359
+ },
+ {
+ "epoch": 32.7825327510917,
+ "grad_norm": 0.01588359661400318,
+ "learning_rate": 0.0006,
+ "loss": 4.163723945617676,
+ "step": 2360
+ },
+ {
+ "epoch": 32.79650655021834,
+ "grad_norm": 0.0168188214302063,
+ "learning_rate": 0.0006,
+ "loss": 4.178189277648926,
+ "step": 2361
+ },
+ {
+ "epoch": 32.81048034934498,
+ "grad_norm": 0.01726095750927925,
+ "learning_rate": 0.0006,
+ "loss": 4.144223690032959,
+ "step": 2362
+ },
+ {
+ "epoch": 32.82445414847162,
+ "grad_norm": 0.01701410301029682,
+ "learning_rate": 0.0006,
+ "loss": 4.209896087646484,
+ "step": 2363
+ },
+ {
+ "epoch": 32.838427947598255,
+ "grad_norm": 0.01682264916598797,
+ "learning_rate": 0.0006,
+ "loss": 4.26223087310791,
+ "step": 2364
+ },
+ {
+ "epoch": 32.852401746724894,
+ "grad_norm": 0.015597456134855747,
+ "learning_rate": 0.0006,
+ "loss": 4.292664051055908,
+ "step": 2365
+ },
+ {
+ "epoch": 32.866375545851525,
+ "grad_norm": 0.015258579514920712,
+ "learning_rate": 0.0006,
+ "loss": 4.165006637573242,
+ "step": 2366
+ },
+ {
+ "epoch": 32.880349344978164,
+ "grad_norm": 0.01662200503051281,
+ "learning_rate": 0.0006,
+ "loss": 4.309786319732666,
+ "step": 2367
+ },
+ {
+ "epoch": 32.8943231441048,
+ "grad_norm": 0.015691161155700684,
+ "learning_rate": 0.0006,
+ "loss": 4.167202949523926,
+ "step": 2368
+ },
+ {
+ "epoch": 32.90829694323144,
+ "grad_norm": 0.014870637096464634,
+ "learning_rate": 0.0006,
+ "loss": 4.142415523529053,
+ "step": 2369
+ },
+ {
+ "epoch": 32.92227074235808,
+ "grad_norm": 0.01567252166569233,
+ "learning_rate": 0.0006,
+ "loss": 4.2312092781066895,
+ "step": 2370
+ },
+ {
+ "epoch": 32.93624454148472,
+ "grad_norm": 0.014948616735637188,
+ "learning_rate": 0.0006,
+ "loss": 4.1514129638671875,
+ "step": 2371
+ },
+ {
+ "epoch": 32.95021834061135,
+ "grad_norm": 0.014654500409960747,
+ "learning_rate": 0.0006,
+ "loss": 4.205010890960693,
+ "step": 2372
+ },
+ {
+ "epoch": 32.96419213973799,
+ "grad_norm": 0.015059148892760277,
+ "learning_rate": 0.0006,
+ "loss": 4.148405075073242,
+ "step": 2373
+ },
+ {
+ "epoch": 32.97816593886463,
+ "grad_norm": 0.015026576817035675,
+ "learning_rate": 0.0006,
+ "loss": 4.182535171508789,
+ "step": 2374
+ },
+ {
+ "epoch": 32.992139737991266,
+ "grad_norm": 0.016394320875406265,
+ "learning_rate": 0.0006,
+ "loss": 4.129051208496094,
+ "step": 2375
+ },
+ {
+ "epoch": 33.0,
+ "grad_norm": 0.01867043413221836,
+ "learning_rate": 0.0006,
+ "loss": 4.185556411743164,
+ "step": 2376
+ },
+ {
+ "epoch": 33.0,
+ "eval_loss": 4.640405654907227,
+ "eval_runtime": 56.5861,
+ "eval_samples_per_second": 43.155,
+ "eval_steps_per_second": 1.361,
+ "step": 2376
+ },
+ {
+ "epoch": 33.01397379912664,
+ "grad_norm": 0.018626727163791656,
+ "learning_rate": 0.0006,
+ "loss": 4.143031597137451,
+ "step": 2377
+ },
+ {
+ "epoch": 33.02794759825328,
+ "grad_norm": 0.019283363595604897,
+ "learning_rate": 0.0006,
+ "loss": 4.117845058441162,
+ "step": 2378
+ },
+ {
+ "epoch": 33.041921397379916,
+ "grad_norm": 0.018185822293162346,
+ "learning_rate": 0.0006,
+ "loss": 4.076862812042236,
+ "step": 2379
+ },
+ {
+ "epoch": 33.05589519650655,
+ "grad_norm": 0.019411196932196617,
+ "learning_rate": 0.0006,
+ "loss": 4.124156475067139,
+ "step": 2380
+ },
+ {
+ "epoch": 33.069868995633186,
+ "grad_norm": 0.019248466938734055,
+ "learning_rate": 0.0006,
+ "loss": 4.201735019683838,
+ "step": 2381
+ },
+ {
+ "epoch": 33.083842794759825,
+ "grad_norm": 0.018856720998883247,
+ "learning_rate": 0.0006,
+ "loss": 4.00296688079834,
+ "step": 2382
+ },
+ {
+ "epoch": 33.09781659388646,
+ "grad_norm": 0.018554294481873512,
+ "learning_rate": 0.0006,
+ "loss": 4.270577430725098,
+ "step": 2383
+ },
+ {
+ "epoch": 33.1117903930131,
+ "grad_norm": 0.019946224987506866,
+ "learning_rate": 0.0006,
+ "loss": 4.119589328765869,
+ "step": 2384
+ },
+ {
+ "epoch": 33.12576419213974,
+ "grad_norm": 0.021395670250058174,
+ "learning_rate": 0.0006,
+ "loss": 4.128847122192383,
+ "step": 2385
+ },
+ {
+ "epoch": 33.13973799126637,
+ "grad_norm": 0.02009674906730652,
+ "learning_rate": 0.0006,
+ "loss": 4.183228969573975,
+ "step": 2386
+ },
+ {
+ "epoch": 33.15371179039301,
+ "grad_norm": 0.01820223033428192,
+ "learning_rate": 0.0006,
+ "loss": 4.131601810455322,
+ "step": 2387
+ },
+ {
+ "epoch": 33.16768558951965,
+ "grad_norm": 0.018653156235814095,
+ "learning_rate": 0.0006,
+ "loss": 4.116430282592773,
+ "step": 2388
+ },
+ {
+ "epoch": 33.18165938864629,
+ "grad_norm": 0.016828063875436783,
+ "learning_rate": 0.0006,
+ "loss": 4.054663181304932,
+ "step": 2389
+ },
+ {
+ "epoch": 33.19563318777293,
+ "grad_norm": 0.017025409266352654,
+ "learning_rate": 0.0006,
+ "loss": 4.110044956207275,
+ "step": 2390
+ },
+ {
+ "epoch": 33.209606986899566,
+ "grad_norm": 0.016608452424407005,
+ "learning_rate": 0.0006,
+ "loss": 4.102668285369873,
+ "step": 2391
+ },
+ {
+ "epoch": 33.223580786026204,
+ "grad_norm": 0.01743577979505062,
+ "learning_rate": 0.0006,
+ "loss": 4.166791915893555,
+ "step": 2392
+ },
+ {
+ "epoch": 33.237554585152836,
+ "grad_norm": 0.017253734171390533,
+ "learning_rate": 0.0006,
+ "loss": 4.217832088470459,
+ "step": 2393
+ },
+ {
+ "epoch": 33.251528384279474,
+ "grad_norm": 0.017937595024704933,
+ "learning_rate": 0.0006,
+ "loss": 4.2173967361450195,
+ "step": 2394
+ },
+ {
+ "epoch": 33.26550218340611,
+ "grad_norm": 0.01985993981361389,
+ "learning_rate": 0.0006,
+ "loss": 4.17849588394165,
+ "step": 2395
+ },
+ {
+ "epoch": 33.27947598253275,
+ "grad_norm": 0.019243355840444565,
+ "learning_rate": 0.0006,
+ "loss": 4.086523056030273,
+ "step": 2396
+ },
+ {
+ "epoch": 33.29344978165939,
+ "grad_norm": 0.01924416422843933,
+ "learning_rate": 0.0006,
+ "loss": 4.128574371337891,
+ "step": 2397
+ },
+ {
+ "epoch": 33.30742358078603,
+ "grad_norm": 0.01983758620917797,
+ "learning_rate": 0.0006,
+ "loss": 4.061581611633301,
+ "step": 2398
+ },
+ {
+ "epoch": 33.32139737991266,
+ "grad_norm": 0.01930462196469307,
+ "learning_rate": 0.0006,
+ "loss": 4.091731071472168,
+ "step": 2399
+ },
+ {
+ "epoch": 33.3353711790393,
+ "grad_norm": 0.017988424748182297,
+ "learning_rate": 0.0006,
+ "loss": 4.155599117279053,
+ "step": 2400
+ },
+ {
+ "epoch": 33.34934497816594,
+ "grad_norm": 0.015194767154753208,
+ "learning_rate": 0.0006,
+ "loss": 4.144134044647217,
+ "step": 2401
+ },
+ {
+ "epoch": 33.36331877729258,
+ "grad_norm": 0.01536081824451685,
+ "learning_rate": 0.0006,
+ "loss": 4.127272129058838,
+ "step": 2402
+ },
+ {
+ "epoch": 33.377292576419215,
+ "grad_norm": 0.015431297942996025,
+ "learning_rate": 0.0006,
+ "loss": 4.3213019371032715,
+ "step": 2403
+ },
+ {
+ "epoch": 33.391266375545854,
+ "grad_norm": 0.015740280970931053,
+ "learning_rate": 0.0006,
+ "loss": 4.121467590332031,
+ "step": 2404
+ },
+ {
+ "epoch": 33.40524017467249,
+ "grad_norm": 0.015379384160041809,
+ "learning_rate": 0.0006,
+ "loss": 4.095216751098633,
+ "step": 2405
+ },
+ {
+ "epoch": 33.419213973799124,
+ "grad_norm": 0.01528073474764824,
+ "learning_rate": 0.0006,
+ "loss": 4.13047981262207,
+ "step": 2406
+ },
+ {
+ "epoch": 33.43318777292576,
+ "grad_norm": 0.015739979222416878,
+ "learning_rate": 0.0006,
+ "loss": 4.2533135414123535,
+ "step": 2407
+ },
+ {
+ "epoch": 33.4471615720524,
+ "grad_norm": 0.01519764307886362,
+ "learning_rate": 0.0006,
+ "loss": 4.131782054901123,
+ "step": 2408
+ },
+ {
+ "epoch": 33.46113537117904,
+ "grad_norm": 0.015425996854901314,
+ "learning_rate": 0.0006,
+ "loss": 4.02811336517334,
+ "step": 2409
+ },
+ {
+ "epoch": 33.47510917030568,
+ "grad_norm": 0.015378830023109913,
+ "learning_rate": 0.0006,
+ "loss": 4.08652925491333,
+ "step": 2410
+ },
+ {
+ "epoch": 33.48908296943232,
+ "grad_norm": 0.01473075058311224,
+ "learning_rate": 0.0006,
+ "loss": 4.214364528656006,
+ "step": 2411
+ },
+ {
+ "epoch": 33.50305676855895,
+ "grad_norm": 0.015141637064516544,
+ "learning_rate": 0.0006,
+ "loss": 4.155712127685547,
+ "step": 2412
+ },
+ {
+ "epoch": 33.51703056768559,
+ "grad_norm": 0.01641593500971794,
+ "learning_rate": 0.0006,
+ "loss": 4.097907066345215,
+ "step": 2413
+ },
+ {
+ "epoch": 33.531004366812226,
+ "grad_norm": 0.015182029455900192,
+ "learning_rate": 0.0006,
+ "loss": 4.238499641418457,
+ "step": 2414
+ },
+ {
+ "epoch": 33.544978165938865,
+ "grad_norm": 0.01643732376396656,
+ "learning_rate": 0.0006,
+ "loss": 4.0661492347717285,
+ "step": 2415
+ },
+ {
+ "epoch": 33.5589519650655,
+ "grad_norm": 0.015614228323101997,
+ "learning_rate": 0.0006,
+ "loss": 4.120795249938965,
+ "step": 2416
+ },
+ {
+ "epoch": 33.57292576419214,
+ "grad_norm": 0.016066187992691994,
+ "learning_rate": 0.0006,
+ "loss": 4.096465110778809,
+ "step": 2417
+ },
+ {
+ "epoch": 33.58689956331878,
+ "grad_norm": 0.017214342951774597,
+ "learning_rate": 0.0006,
+ "loss": 4.242628574371338,
+ "step": 2418
+ },
+ {
+ "epoch": 33.60087336244541,
+ "grad_norm": 0.01787586510181427,
+ "learning_rate": 0.0006,
+ "loss": 4.149589538574219,
+ "step": 2419
+ },
+ {
+ "epoch": 33.61484716157205,
+ "grad_norm": 0.019455241039395332,
+ "learning_rate": 0.0006,
+ "loss": 4.267796039581299,
+ "step": 2420
+ },
+ {
+ "epoch": 33.62882096069869,
+ "grad_norm": 0.01984674669802189,
+ "learning_rate": 0.0006,
+ "loss": 4.175220966339111,
+ "step": 2421
+ },
+ {
+ "epoch": 33.64279475982533,
+ "grad_norm": 0.021933717653155327,
+ "learning_rate": 0.0006,
+ "loss": 4.106647491455078,
+ "step": 2422
+ },
+ {
+ "epoch": 33.65676855895197,
+ "grad_norm": 0.02123589813709259,
+ "learning_rate": 0.0006,
+ "loss": 4.204996109008789,
+ "step": 2423
+ },
+ {
+ "epoch": 33.670742358078606,
+ "grad_norm": 0.019814589992165565,
+ "learning_rate": 0.0006,
+ "loss": 4.186171531677246,
+ "step": 2424
+ },
+ {
+ "epoch": 33.68471615720524,
+ "grad_norm": 0.02059728465974331,
+ "learning_rate": 0.0006,
+ "loss": 4.170509338378906,
+ "step": 2425
+ },
+ {
+ "epoch": 33.698689956331876,
+ "grad_norm": 0.018483608961105347,
+ "learning_rate": 0.0006,
+ "loss": 4.085585594177246,
+ "step": 2426
+ },
+ {
+ "epoch": 33.712663755458514,
+ "grad_norm": 0.018403852358460426,
+ "learning_rate": 0.0006,
+ "loss": 4.2551774978637695,
+ "step": 2427
+ },
+ {
+ "epoch": 33.72663755458515,
+ "grad_norm": 0.018653830513358116,
+ "learning_rate": 0.0006,
+ "loss": 4.091766357421875,
+ "step": 2428
+ },
+ {
+ "epoch": 33.74061135371179,
+ "grad_norm": 0.017692334949970245,
+ "learning_rate": 0.0006,
+ "loss": 4.100838661193848,
+ "step": 2429
+ },
+ {
+ "epoch": 33.75458515283843,
+ "grad_norm": 0.016029126942157745,
+ "learning_rate": 0.0006,
+ "loss": 4.1530680656433105,
+ "step": 2430
+ },
+ {
+ "epoch": 33.76855895196506,
+ "grad_norm": 0.016095684841275215,
+ "learning_rate": 0.0006,
+ "loss": 4.20544958114624,
+ "step": 2431
+ },
+ {
+ "epoch": 33.7825327510917,
+ "grad_norm": 0.016375798732042313,
+ "learning_rate": 0.0006,
+ "loss": 4.199177265167236,
+ "step": 2432
+ },
+ {
+ "epoch": 33.79650655021834,
+ "grad_norm": 0.01590786688029766,
+ "learning_rate": 0.0006,
+ "loss": 4.200799465179443,
+ "step": 2433
+ },
+ {
+ "epoch": 33.81048034934498,
+ "grad_norm": 0.014896097593009472,
+ "learning_rate": 0.0006,
+ "loss": 4.185850143432617,
+ "step": 2434
+ },
+ {
+ "epoch": 33.82445414847162,
+ "grad_norm": 0.01438513956964016,
+ "learning_rate": 0.0006,
+ "loss": 4.214962005615234,
+ "step": 2435
+ },
+ {
+ "epoch": 33.838427947598255,
+ "grad_norm": 0.014459339901804924,
+ "learning_rate": 0.0006,
+ "loss": 4.196972370147705,
+ "step": 2436
+ },
+ {
+ "epoch": 33.852401746724894,
+ "grad_norm": 0.014782169833779335,
+ "learning_rate": 0.0006,
+ "loss": 4.0885910987854,
+ "step": 2437
+ },
+ {
+ "epoch": 33.866375545851525,
+ "grad_norm": 0.01520370040088892,
+ "learning_rate": 0.0006,
+ "loss": 4.0323262214660645,
+ "step": 2438
+ },
+ {
+ "epoch": 33.880349344978164,
+ "grad_norm": 0.014056282117962837,
+ "learning_rate": 0.0006,
+ "loss": 4.1386518478393555,
+ "step": 2439
+ },
+ {
+ "epoch": 33.8943231441048,
+ "grad_norm": 0.013567990623414516,
+ "learning_rate": 0.0006,
+ "loss": 4.095754623413086,
+ "step": 2440
+ },
+ {
+ "epoch": 33.90829694323144,
+ "grad_norm": 0.013306716457009315,
+ "learning_rate": 0.0006,
+ "loss": 4.1623077392578125,
+ "step": 2441
+ },
+ {
+ "epoch": 33.92227074235808,
+ "grad_norm": 0.014047150500118732,
+ "learning_rate": 0.0006,
+ "loss": 4.142541885375977,
+ "step": 2442
+ },
+ {
+ "epoch": 33.93624454148472,
+ "grad_norm": 0.01377617847174406,
+ "learning_rate": 0.0006,
+ "loss": 4.268551826477051,
+ "step": 2443
+ },
+ {
+ "epoch": 33.95021834061135,
+ "grad_norm": 0.014671621844172478,
+ "learning_rate": 0.0006,
+ "loss": 4.198196887969971,
+ "step": 2444
+ },
+ {
+ "epoch": 33.96419213973799,
+ "grad_norm": 0.014806744642555714,
+ "learning_rate": 0.0006,
+ "loss": 4.057899475097656,
+ "step": 2445
+ },
+ {
+ "epoch": 33.97816593886463,
+ "grad_norm": 0.015734272077679634,
+ "learning_rate": 0.0006,
+ "loss": 4.163505554199219,
+ "step": 2446
+ },
+ {
+ "epoch": 33.992139737991266,
+ "grad_norm": 0.015250146389007568,
+ "learning_rate": 0.0006,
+ "loss": 4.076294898986816,
+ "step": 2447
+ },
+ {
+ "epoch": 34.0,
+ "grad_norm": 0.016733702272176743,
+ "learning_rate": 0.0006,
+ "loss": 4.138420104980469,
+ "step": 2448
+ },
+ {
+ "epoch": 34.0,
+ "eval_loss": 4.542503356933594,
+ "eval_runtime": 56.983,
+ "eval_samples_per_second": 42.855,
+ "eval_steps_per_second": 1.351,
+ "step": 2448
+ },
+ {
+ "epoch": 34.01397379912664,
+ "grad_norm": 0.015866613015532494,
+ "learning_rate": 0.0006,
+ "loss": 4.032081127166748,
+ "step": 2449
+ },
+ {
+ "epoch": 34.02794759825328,
+ "grad_norm": 0.01946570910513401,
+ "learning_rate": 0.0006,
+ "loss": 4.122320175170898,
+ "step": 2450
+ },
+ {
+ "epoch": 34.041921397379916,
+ "grad_norm": 0.020265938714146614,
+ "learning_rate": 0.0006,
+ "loss": 4.06460428237915,
+ "step": 2451
+ },
+ {
+ "epoch": 34.05589519650655,
+ "grad_norm": 0.018887171521782875,
+ "learning_rate": 0.0006,
+ "loss": 4.096906661987305,
+ "step": 2452
+ },
+ {
+ "epoch": 34.069868995633186,
+ "grad_norm": 0.017304692417383194,
+ "learning_rate": 0.0006,
+ "loss": 4.137261390686035,
+ "step": 2453
+ },
+ {
+ "epoch": 34.083842794759825,
+ "grad_norm": 0.017029082402586937,
+ "learning_rate": 0.0006,
+ "loss": 4.081112861633301,
+ "step": 2454
+ },
+ {
+ "epoch": 34.09781659388646,
+ "grad_norm": 0.016414813697338104,
+ "learning_rate": 0.0006,
+ "loss": 4.140783309936523,
+ "step": 2455
+ },
+ {
+ "epoch": 34.1117903930131,
+ "grad_norm": 0.018238281831145287,
+ "learning_rate": 0.0006,
+ "loss": 4.161116600036621,
+ "step": 2456
+ },
+ {
+ "epoch": 34.12576419213974,
+ "grad_norm": 0.019929232075810432,
+ "learning_rate": 0.0006,
+ "loss": 4.243722438812256,
+ "step": 2457
+ },
+ {
+ "epoch": 34.13973799126637,
+ "grad_norm": 0.019375786185264587,
+ "learning_rate": 0.0006,
+ "loss": 4.134873390197754,
+ "step": 2458
+ },
+ {
+ "epoch": 34.15371179039301,
+ "grad_norm": 0.018151137977838516,
+ "learning_rate": 0.0006,
+ "loss": 4.098745346069336,
+ "step": 2459
+ },
+ {
+ "epoch": 34.16768558951965,
+ "grad_norm": 0.017054934054613113,
+ "learning_rate": 0.0006,
+ "loss": 4.194662570953369,
+ "step": 2460
+ },
+ {
+ "epoch": 34.18165938864629,
+ "grad_norm": 0.018220555037260056,
+ "learning_rate": 0.0006,
+ "loss": 4.128866195678711,
+ "step": 2461
+ },
+ {
+ "epoch": 34.19563318777293,
+ "grad_norm": 0.01692879945039749,
+ "learning_rate": 0.0006,
+ "loss": 4.062477111816406,
+ "step": 2462
+ },
+ {
+ "epoch": 34.209606986899566,
+ "grad_norm": 0.01674199476838112,
+ "learning_rate": 0.0006,
+ "loss": 4.186149597167969,
+ "step": 2463
+ },
+ {
+ "epoch": 34.223580786026204,
+ "grad_norm": 0.016527455300092697,
+ "learning_rate": 0.0006,
+ "loss": 4.0734028816223145,
+ "step": 2464
+ },
+ {
+ "epoch": 34.237554585152836,
+ "grad_norm": 0.016448386013507843,
+ "learning_rate": 0.0006,
+ "loss": 4.098896026611328,
+ "step": 2465
+ },
+ {
+ "epoch": 34.251528384279474,
+ "grad_norm": 0.014644953422248363,
+ "learning_rate": 0.0006,
+ "loss": 4.1031904220581055,
+ "step": 2466
+ },
+ {
+ "epoch": 34.26550218340611,
+ "grad_norm": 0.016649341210722923,
+ "learning_rate": 0.0006,
+ "loss": 4.148791313171387,
+ "step": 2467
+ },
+ {
+ "epoch": 34.27947598253275,
+ "grad_norm": 0.017442714422941208,
+ "learning_rate": 0.0006,
+ "loss": 4.070799827575684,
+ "step": 2468
+ },
+ {
+ "epoch": 34.29344978165939,
+ "grad_norm": 0.017852013930678368,
+ "learning_rate": 0.0006,
+ "loss": 4.031924247741699,
+ "step": 2469
+ },
+ {
+ "epoch": 34.30742358078603,
+ "grad_norm": 0.016371270641684532,
+ "learning_rate": 0.0006,
+ "loss": 4.149420738220215,
+ "step": 2470
+ },
+ {
+ "epoch": 34.32139737991266,
+ "grad_norm": 0.015846814960241318,
+ "learning_rate": 0.0006,
+ "loss": 4.15427303314209,
+ "step": 2471
+ },
+ {
+ "epoch": 34.3353711790393,
+ "grad_norm": 0.017411163076758385,
+ "learning_rate": 0.0006,
+ "loss": 4.1304192543029785,
+ "step": 2472
+ },
+ {
+ "epoch": 34.34934497816594,
+ "grad_norm": 0.01780046708881855,
+ "learning_rate": 0.0006,
+ "loss": 4.191732406616211,
+ "step": 2473
+ },
+ {
+ "epoch": 34.36331877729258,
+ "grad_norm": 0.01958383060991764,
+ "learning_rate": 0.0006,
+ "loss": 4.068355083465576,
+ "step": 2474
+ },
+ {
+ "epoch": 34.377292576419215,
+ "grad_norm": 0.022173317149281502,
+ "learning_rate": 0.0006,
+ "loss": 4.182929039001465,
+ "step": 2475
+ },
+ {
+ "epoch": 34.391266375545854,
+ "grad_norm": 0.022389156743884087,
+ "learning_rate": 0.0006,
+ "loss": 4.110218048095703,
+ "step": 2476
+ },
+ {
+ "epoch": 34.40524017467249,
+ "grad_norm": 0.022478511556982994,
+ "learning_rate": 0.0006,
+ "loss": 4.095479488372803,
+ "step": 2477
+ },
+ {
+ "epoch": 34.419213973799124,
+ "grad_norm": 0.019504647701978683,
+ "learning_rate": 0.0006,
+ "loss": 4.101964950561523,
+ "step": 2478
+ },
+ {
+ "epoch": 34.43318777292576,
+ "grad_norm": 0.019378576427698135,
+ "learning_rate": 0.0006,
+ "loss": 4.14620304107666,
+ "step": 2479
+ },
+ {
+ "epoch": 34.4471615720524,
+ "grad_norm": 0.017993204295635223,
+ "learning_rate": 0.0006,
+ "loss": 4.183492660522461,
+ "step": 2480
+ },
+ {
+ "epoch": 34.46113537117904,
+ "grad_norm": 0.01814805157482624,
+ "learning_rate": 0.0006,
+ "loss": 4.151247978210449,
+ "step": 2481
+ },
+ {
+ "epoch": 34.47510917030568,
+ "grad_norm": 0.018622448667883873,
+ "learning_rate": 0.0006,
+ "loss": 4.020938396453857,
+ "step": 2482
+ },
+ {
+ "epoch": 34.48908296943232,
+ "grad_norm": 0.018360132351517677,
+ "learning_rate": 0.0006,
+ "loss": 4.064734935760498,
+ "step": 2483
+ },
+ {
+ "epoch": 34.50305676855895,
+ "grad_norm": 0.017715711146593094,
+ "learning_rate": 0.0006,
+ "loss": 4.185687065124512,
+ "step": 2484
+ },
+ {
+ "epoch": 34.51703056768559,
+ "grad_norm": 0.01714295893907547,
+ "learning_rate": 0.0006,
+ "loss": 4.1044721603393555,
+ "step": 2485
+ },
+ {
+ "epoch": 34.531004366812226,
+ "grad_norm": 0.016943803057074547,
+ "learning_rate": 0.0006,
+ "loss": 4.106837272644043,
+ "step": 2486
+ },
+ {
+ "epoch": 34.544978165938865,
+ "grad_norm": 0.0182423684746027,
+ "learning_rate": 0.0006,
+ "loss": 4.1041340827941895,
+ "step": 2487
+ },
+ {
+ "epoch": 34.5589519650655,
+ "grad_norm": 0.01636587455868721,
+ "learning_rate": 0.0006,
+ "loss": 3.9735546112060547,
+ "step": 2488
+ },
+ {
+ "epoch": 34.57292576419214,
+ "grad_norm": 0.014593546278774738,
+ "learning_rate": 0.0006,
+ "loss": 4.041980266571045,
+ "step": 2489
+ },
+ {
+ "epoch": 34.58689956331878,
+ "grad_norm": 0.015436965972185135,
+ "learning_rate": 0.0006,
+ "loss": 4.301148414611816,
+ "step": 2490
+ },
+ {
+ "epoch": 34.60087336244541,
+ "grad_norm": 0.015075269155204296,
+ "learning_rate": 0.0006,
+ "loss": 4.118527412414551,
+ "step": 2491
+ },
+ {
+ "epoch": 34.61484716157205,
+ "grad_norm": 0.01701514609158039,
+ "learning_rate": 0.0006,
+ "loss": 4.245968818664551,
+ "step": 2492
+ },
+ {
+ "epoch": 34.62882096069869,
+ "grad_norm": 0.018047887831926346,
+ "learning_rate": 0.0006,
+ "loss": 4.227261066436768,
+ "step": 2493
+ },
+ {
+ "epoch": 34.64279475982533,
+ "grad_norm": 0.018499085679650307,
+ "learning_rate": 0.0006,
+ "loss": 4.121360778808594,
+ "step": 2494
+ },
+ {
+ "epoch": 34.65676855895197,
+ "grad_norm": 0.018987711519002914,
+ "learning_rate": 0.0006,
+ "loss": 4.131564617156982,
+ "step": 2495
+ },
+ {
+ "epoch": 34.670742358078606,
+ "grad_norm": 0.01834476925432682,
+ "learning_rate": 0.0006,
+ "loss": 4.224941253662109,
+ "step": 2496
+ },
+ {
+ "epoch": 34.68471615720524,
+ "grad_norm": 0.017507942393422127,
+ "learning_rate": 0.0006,
+ "loss": 4.1394548416137695,
+ "step": 2497
+ },
+ {
+ "epoch": 34.698689956331876,
+ "grad_norm": 0.01680580899119377,
+ "learning_rate": 0.0006,
+ "loss": 4.1073431968688965,
+ "step": 2498
+ },
+ {
+ "epoch": 34.712663755458514,
+ "grad_norm": 0.015233514830470085,
+ "learning_rate": 0.0006,
+ "loss": 4.136994361877441,
+ "step": 2499
+ },
+ {
+ "epoch": 34.72663755458515,
+ "grad_norm": 0.015482917428016663,
+ "learning_rate": 0.0006,
+ "loss": 4.110719680786133,
+ "step": 2500
+ },
+ {
+ "epoch": 34.74061135371179,
+ "grad_norm": 0.015940461307764053,
+ "learning_rate": 0.0006,
+ "loss": 4.207851886749268,
+ "step": 2501
+ },
+ {
+ "epoch": 34.75458515283843,
+ "grad_norm": 0.015355597250163555,
+ "learning_rate": 0.0006,
+ "loss": 4.169469833374023,
+ "step": 2502
+ },
+ {
+ "epoch": 34.76855895196506,
+ "grad_norm": 0.015974000096321106,
+ "learning_rate": 0.0006,
+ "loss": 4.110061168670654,
+ "step": 2503
+ },
+ {
+ "epoch": 34.7825327510917,
+ "grad_norm": 0.016088927164673805,
+ "learning_rate": 0.0006,
+ "loss": 4.155452728271484,
+ "step": 2504
+ },
+ {
+ "epoch": 34.79650655021834,
+ "grad_norm": 0.016156120225787163,
+ "learning_rate": 0.0006,
+ "loss": 4.178092956542969,
+ "step": 2505
+ },
+ {
+ "epoch": 34.81048034934498,
+ "grad_norm": 0.014928647316992283,
+ "learning_rate": 0.0006,
+ "loss": 4.05457878112793,
+ "step": 2506
+ },
+ {
+ "epoch": 34.82445414847162,
+ "grad_norm": 0.015289328061044216,
+ "learning_rate": 0.0006,
+ "loss": 4.18869686126709,
+ "step": 2507
+ },
+ {
+ "epoch": 34.838427947598255,
+ "grad_norm": 0.015799039974808693,
+ "learning_rate": 0.0006,
+ "loss": 4.106073379516602,
+ "step": 2508
+ },
+ {
+ "epoch": 34.852401746724894,
+ "grad_norm": 0.015824418514966965,
+ "learning_rate": 0.0006,
+ "loss": 4.0634589195251465,
+ "step": 2509
+ },
+ {
+ "epoch": 34.866375545851525,
+ "grad_norm": 0.015321656130254269,
+ "learning_rate": 0.0006,
+ "loss": 4.093090057373047,
+ "step": 2510
+ },
+ {
+ "epoch": 34.880349344978164,
+ "grad_norm": 0.014380166307091713,
+ "learning_rate": 0.0006,
+ "loss": 4.1750640869140625,
+ "step": 2511
+ },
+ {
+ "epoch": 34.8943231441048,
+ "grad_norm": 0.014962567947804928,
+ "learning_rate": 0.0006,
+ "loss": 4.185277938842773,
+ "step": 2512
+ },
+ {
+ "epoch": 34.90829694323144,
+ "grad_norm": 0.014467543922364712,
+ "learning_rate": 0.0006,
+ "loss": 4.1600799560546875,
+ "step": 2513
+ },
+ {
+ "epoch": 34.92227074235808,
+ "grad_norm": 0.01545181218534708,
+ "learning_rate": 0.0006,
+ "loss": 4.096299648284912,
+ "step": 2514
+ },
+ {
+ "epoch": 34.93624454148472,
+ "grad_norm": 0.01605340465903282,
+ "learning_rate": 0.0006,
+ "loss": 4.203323841094971,
+ "step": 2515
+ },
+ {
+ "epoch": 34.95021834061135,
+ "grad_norm": 0.015203815884888172,
+ "learning_rate": 0.0006,
+ "loss": 4.201385021209717,
+ "step": 2516
+ },
+ {
+ "epoch": 34.96419213973799,
+ "grad_norm": 0.015149409882724285,
+ "learning_rate": 0.0006,
+ "loss": 4.088164329528809,
+ "step": 2517
+ },
+ {
+ "epoch": 34.97816593886463,
+ "grad_norm": 0.014885558746755123,
+ "learning_rate": 0.0006,
+ "loss": 4.073911666870117,
+ "step": 2518
+ },
+ {
+ "epoch": 34.992139737991266,
+ "grad_norm": 0.015003419481217861,
+ "learning_rate": 0.0006,
+ "loss": 4.157113552093506,
+ "step": 2519
+ },
+ {
+ "epoch": 35.0,
+ "grad_norm": 0.01641916297376156,
+ "learning_rate": 0.0006,
+ "loss": 3.970052719116211,
+ "step": 2520
+ },
+ {
+ "epoch": 35.0,
+ "eval_loss": 4.641692161560059,
+ "eval_runtime": 56.8836,
+ "eval_samples_per_second": 42.93,
+ "eval_steps_per_second": 1.354,
+ "step": 2520
+ },
+ {
+ "epoch": 35.01397379912664,
+ "grad_norm": 0.015894673764705658,
+ "learning_rate": 0.0006,
+ "loss": 4.104437351226807,
+ "step": 2521
+ },
+ {
+ "epoch": 35.02794759825328,
+ "grad_norm": 0.0173867866396904,
+ "learning_rate": 0.0006,
+ "loss": 4.091726303100586,
+ "step": 2522
+ },
+ {
+ "epoch": 35.041921397379916,
+ "grad_norm": 0.019272832199931145,
+ "learning_rate": 0.0006,
+ "loss": 4.217905044555664,
+ "step": 2523
+ },
+ {
+ "epoch": 35.05589519650655,
+ "grad_norm": 0.021820371970534325,
+ "learning_rate": 0.0006,
+ "loss": 4.111428260803223,
+ "step": 2524
+ },
+ {
+ "epoch": 35.069868995633186,
+ "grad_norm": 0.02340741828083992,
+ "learning_rate": 0.0006,
+ "loss": 4.115771293640137,
+ "step": 2525
+ },
+ {
+ "epoch": 35.083842794759825,
+ "grad_norm": 0.023760691285133362,
+ "learning_rate": 0.0006,
+ "loss": 4.132383346557617,
+ "step": 2526
+ },
+ {
+ "epoch": 35.09781659388646,
+ "grad_norm": 0.02073468267917633,
+ "learning_rate": 0.0006,
+ "loss": 4.272691249847412,
+ "step": 2527
+ },
+ {
+ "epoch": 35.1117903930131,
+ "grad_norm": 0.02197718247771263,
+ "learning_rate": 0.0006,
+ "loss": 4.105586528778076,
+ "step": 2528
+ },
+ {
+ "epoch": 35.12576419213974,
+ "grad_norm": 0.0200254675000906,
+ "learning_rate": 0.0006,
+ "loss": 3.968820095062256,
+ "step": 2529
+ },
+ {
+ "epoch": 35.13973799126637,
+ "grad_norm": 0.01929587684571743,
+ "learning_rate": 0.0006,
+ "loss": 4.045468807220459,
+ "step": 2530
+ },
+ {
+ "epoch": 35.15371179039301,
+ "grad_norm": 0.01886162906885147,
+ "learning_rate": 0.0006,
+ "loss": 4.185571670532227,
+ "step": 2531
+ },
+ {
+ "epoch": 35.16768558951965,
+ "grad_norm": 0.019789839163422585,
+ "learning_rate": 0.0006,
+ "loss": 3.9734551906585693,
+ "step": 2532
+ },
+ {
+ "epoch": 35.18165938864629,
+ "grad_norm": 0.0198142658919096,
+ "learning_rate": 0.0006,
+ "loss": 4.2662506103515625,
+ "step": 2533
+ },
+ {
+ "epoch": 35.19563318777293,
+ "grad_norm": 0.02087012305855751,
+ "learning_rate": 0.0006,
+ "loss": 4.175931930541992,
+ "step": 2534
+ },
+ {
+ "epoch": 35.209606986899566,
+ "grad_norm": 0.020294535905122757,
+ "learning_rate": 0.0006,
+ "loss": 4.060324668884277,
+ "step": 2535
+ },
+ {
+ "epoch": 35.223580786026204,
+ "grad_norm": 0.01908983662724495,
+ "learning_rate": 0.0006,
+ "loss": 4.217939853668213,
+ "step": 2536
+ },
+ {
+ "epoch": 35.237554585152836,
+ "grad_norm": 0.0178892333060503,
+ "learning_rate": 0.0006,
+ "loss": 4.190216541290283,
+ "step": 2537
+ },
+ {
+ "epoch": 35.251528384279474,
+ "grad_norm": 0.017892315983772278,
+ "learning_rate": 0.0006,
+ "loss": 4.067419052124023,
+ "step": 2538
+ },
+ {
+ "epoch": 35.26550218340611,
+ "grad_norm": 0.018556250259280205,
+ "learning_rate": 0.0006,
+ "loss": 4.105958461761475,
+ "step": 2539
+ },
+ {
+ "epoch": 35.27947598253275,
+ "grad_norm": 0.01944422349333763,
+ "learning_rate": 0.0006,
+ "loss": 4.155775547027588,
+ "step": 2540
+ },
+ {
+ "epoch": 35.29344978165939,
+ "grad_norm": 0.017333146184682846,
+ "learning_rate": 0.0006,
+ "loss": 4.244513511657715,
+ "step": 2541
+ },
+ {
+ "epoch": 35.30742358078603,
+ "grad_norm": 0.01639244332909584,
+ "learning_rate": 0.0006,
+ "loss": 4.167815685272217,
+ "step": 2542
+ },
+ {
+ "epoch": 35.32139737991266,
+ "grad_norm": 0.01610143668949604,
+ "learning_rate": 0.0006,
+ "loss": 4.121486663818359,
+ "step": 2543
+ },
+ {
+ "epoch": 35.3353711790393,
+ "grad_norm": 0.01656411588191986,
+ "learning_rate": 0.0006,
+ "loss": 4.1755499839782715,
+ "step": 2544
+ },
+ {
+ "epoch": 35.34934497816594,
+ "grad_norm": 0.015075297094881535,
+ "learning_rate": 0.0006,
+ "loss": 4.102992534637451,
+ "step": 2545
+ },
+ {
+ "epoch": 35.36331877729258,
+ "grad_norm": 0.01538170501589775,
+ "learning_rate": 0.0006,
+ "loss": 4.161603927612305,
+ "step": 2546
+ },
+ {
+ "epoch": 35.377292576419215,
+ "grad_norm": 0.01483312901109457,
+ "learning_rate": 0.0006,
+ "loss": 4.085355758666992,
+ "step": 2547
+ },
+ {
+ "epoch": 35.391266375545854,
+ "grad_norm": 0.014894738793373108,
+ "learning_rate": 0.0006,
+ "loss": 4.083585262298584,
+ "step": 2548
+ },
+ {
+ "epoch": 35.40524017467249,
+ "grad_norm": 0.01475780550390482,
+ "learning_rate": 0.0006,
+ "loss": 4.012645721435547,
+ "step": 2549
+ },
+ {
+ "epoch": 35.419213973799124,
+ "grad_norm": 0.014531653374433517,
+ "learning_rate": 0.0006,
+ "loss": 4.197113990783691,
+ "step": 2550
+ },
+ {
+ "epoch": 35.43318777292576,
+ "grad_norm": 0.015316493809223175,
+ "learning_rate": 0.0006,
+ "loss": 4.227970123291016,
+ "step": 2551
+ },
+ {
+ "epoch": 35.4471615720524,
+ "grad_norm": 0.014685571193695068,
+ "learning_rate": 0.0006,
+ "loss": 4.089415550231934,
+ "step": 2552
+ },
+ {
+ "epoch": 35.46113537117904,
+ "grad_norm": 0.015337035991251469,
+ "learning_rate": 0.0006,
+ "loss": 4.080583572387695,
+ "step": 2553
+ },
+ {
+ "epoch": 35.47510917030568,
+ "grad_norm": 0.016754208132624626,
+ "learning_rate": 0.0006,
+ "loss": 4.163935661315918,
+ "step": 2554
+ },
+ {
+ "epoch": 35.48908296943232,
+ "grad_norm": 0.017535798251628876,
+ "learning_rate": 0.0006,
+ "loss": 4.093027591705322,
+ "step": 2555
+ },
+ {
+ "epoch": 35.50305676855895,
+ "grad_norm": 0.017336975783109665,
+ "learning_rate": 0.0006,
+ "loss": 4.127961158752441,
+ "step": 2556
+ },
+ {
+ "epoch": 35.51703056768559,
+ "grad_norm": 0.016242152079939842,
+ "learning_rate": 0.0006,
+ "loss": 4.155593395233154,
+ "step": 2557
+ },
+ {
+ "epoch": 35.531004366812226,
+ "grad_norm": 0.015121322125196457,
+ "learning_rate": 0.0006,
+ "loss": 3.984004020690918,
+ "step": 2558
+ },
+ {
+ "epoch": 35.544978165938865,
+ "grad_norm": 0.015178884379565716,
+ "learning_rate": 0.0006,
+ "loss": 4.111564636230469,
+ "step": 2559
+ },
+ {
+ "epoch": 35.5589519650655,
+ "grad_norm": 0.014777131378650665,
+ "learning_rate": 0.0006,
+ "loss": 4.041402816772461,
+ "step": 2560
+ },
+ {
+ "epoch": 35.57292576419214,
+ "grad_norm": 0.015197676606476307,
+ "learning_rate": 0.0006,
+ "loss": 4.133706092834473,
+ "step": 2561
+ },
+ {
+ "epoch": 35.58689956331878,
+ "grad_norm": 0.015905972570180893,
+ "learning_rate": 0.0006,
+ "loss": 4.1069135665893555,
+ "step": 2562
+ },
+ {
+ "epoch": 35.60087336244541,
+ "grad_norm": 0.015942877158522606,
+ "learning_rate": 0.0006,
+ "loss": 4.0720624923706055,
+ "step": 2563
+ },
+ {
+ "epoch": 35.61484716157205,
+ "grad_norm": 0.016697583720088005,
+ "learning_rate": 0.0006,
+ "loss": 4.167448043823242,
+ "step": 2564
+ },
+ {
+ "epoch": 35.62882096069869,
+ "grad_norm": 0.015603577718138695,
+ "learning_rate": 0.0006,
+ "loss": 4.127327919006348,
+ "step": 2565
+ },
+ {
+ "epoch": 35.64279475982533,
+ "grad_norm": 0.017153460532426834,
+ "learning_rate": 0.0006,
+ "loss": 4.056543350219727,
+ "step": 2566
+ },
+ {
+ "epoch": 35.65676855895197,
+ "grad_norm": 0.016731958836317062,
+ "learning_rate": 0.0006,
+ "loss": 4.0681843757629395,
+ "step": 2567
+ },
+ {
+ "epoch": 35.670742358078606,
+ "grad_norm": 0.016167359426617622,
+ "learning_rate": 0.0006,
+ "loss": 4.174992084503174,
+ "step": 2568
+ },
+ {
+ "epoch": 35.68471615720524,
+ "grad_norm": 0.01615108922123909,
+ "learning_rate": 0.0006,
+ "loss": 4.1383161544799805,
+ "step": 2569
+ },
+ {
+ "epoch": 35.698689956331876,
+ "grad_norm": 0.015750816091895103,
+ "learning_rate": 0.0006,
+ "loss": 4.031595706939697,
+ "step": 2570
+ },
+ {
+ "epoch": 35.712663755458514,
+ "grad_norm": 0.015416189096868038,
+ "learning_rate": 0.0006,
+ "loss": 4.067498207092285,
+ "step": 2571
+ },
+ {
+ "epoch": 35.72663755458515,
+ "grad_norm": 0.01570792682468891,
+ "learning_rate": 0.0006,
+ "loss": 4.1338300704956055,
+ "step": 2572
+ },
+ {
+ "epoch": 35.74061135371179,
+ "grad_norm": 0.016367752104997635,
+ "learning_rate": 0.0006,
+ "loss": 4.152523040771484,
+ "step": 2573
+ },
+ {
+ "epoch": 35.75458515283843,
+ "grad_norm": 0.016349803656339645,
+ "learning_rate": 0.0006,
+ "loss": 4.101985454559326,
+ "step": 2574
+ },
+ {
+ "epoch": 35.76855895196506,
+ "grad_norm": 0.01663660630583763,
+ "learning_rate": 0.0006,
+ "loss": 4.190130710601807,
+ "step": 2575
+ },
+ {
+ "epoch": 35.7825327510917,
+ "grad_norm": 0.015599170699715614,
+ "learning_rate": 0.0006,
+ "loss": 4.0792436599731445,
+ "step": 2576
+ },
+ {
+ "epoch": 35.79650655021834,
+ "grad_norm": 0.015472427010536194,
+ "learning_rate": 0.0006,
+ "loss": 4.186446666717529,
+ "step": 2577
+ },
+ {
+ "epoch": 35.81048034934498,
+ "grad_norm": 0.015520439483225346,
+ "learning_rate": 0.0006,
+ "loss": 4.246635437011719,
+ "step": 2578
+ },
+ {
+ "epoch": 35.82445414847162,
+ "grad_norm": 0.016998281702399254,
+ "learning_rate": 0.0006,
+ "loss": 4.13435173034668,
+ "step": 2579
+ },
+ {
+ "epoch": 35.838427947598255,
+ "grad_norm": 0.016290361061692238,
+ "learning_rate": 0.0006,
+ "loss": 4.107428073883057,
+ "step": 2580
+ },
+ {
+ "epoch": 35.852401746724894,
+ "grad_norm": 0.0165516659617424,
+ "learning_rate": 0.0006,
+ "loss": 4.1612138748168945,
+ "step": 2581
+ },
+ {
+ "epoch": 35.866375545851525,
+ "grad_norm": 0.01736091636121273,
+ "learning_rate": 0.0006,
+ "loss": 4.1072869300842285,
+ "step": 2582
+ },
+ {
+ "epoch": 35.880349344978164,
+ "grad_norm": 0.01787031814455986,
+ "learning_rate": 0.0006,
+ "loss": 4.1076812744140625,
+ "step": 2583
+ },
+ {
+ "epoch": 35.8943231441048,
+ "grad_norm": 0.015437985770404339,
+ "learning_rate": 0.0006,
+ "loss": 4.097488880157471,
+ "step": 2584
+ },
+ {
+ "epoch": 35.90829694323144,
+ "grad_norm": 0.016306741163134575,
+ "learning_rate": 0.0006,
+ "loss": 4.105017185211182,
+ "step": 2585
+ },
+ {
+ "epoch": 35.92227074235808,
+ "grad_norm": 0.015723245218396187,
+ "learning_rate": 0.0006,
+ "loss": 4.205283164978027,
+ "step": 2586
+ },
+ {
+ "epoch": 35.93624454148472,
+ "grad_norm": 0.014626456424593925,
+ "learning_rate": 0.0006,
+ "loss": 4.086552619934082,
+ "step": 2587
+ },
+ {
+ "epoch": 35.95021834061135,
+ "grad_norm": 0.014348454773426056,
+ "learning_rate": 0.0006,
+ "loss": 4.171329975128174,
+ "step": 2588
+ },
+ {
+ "epoch": 35.96419213973799,
+ "grad_norm": 0.015598480589687824,
+ "learning_rate": 0.0006,
+ "loss": 4.140815258026123,
+ "step": 2589
+ },
+ {
+ "epoch": 35.97816593886463,
+ "grad_norm": 0.017126062884926796,
+ "learning_rate": 0.0006,
+ "loss": 4.157176971435547,
+ "step": 2590
+ },
+ {
+ "epoch": 35.992139737991266,
+ "grad_norm": 0.01860199309885502,
+ "learning_rate": 0.0006,
+ "loss": 4.252880096435547,
+ "step": 2591
+ },
+ {
+ "epoch": 36.0,
+ "grad_norm": 0.021780220791697502,
+ "learning_rate": 0.0006,
+ "loss": 4.249861717224121,
+ "step": 2592
+ },
+ {
+ "epoch": 36.0,
+ "eval_loss": 4.606466293334961,
+ "eval_runtime": 56.6217,
+ "eval_samples_per_second": 43.128,
+ "eval_steps_per_second": 1.36,
+ "step": 2592
+ },
+ {
+ "epoch": 36.01397379912664,
+ "grad_norm": 0.020892828702926636,
+ "learning_rate": 0.0006,
+ "loss": 4.102055549621582,
+ "step": 2593
+ },
+ {
+ "epoch": 36.02794759825328,
+ "grad_norm": 0.020306255668401718,
+ "learning_rate": 0.0006,
+ "loss": 4.184308052062988,
+ "step": 2594
+ },
+ {
+ "epoch": 36.041921397379916,
+ "grad_norm": 0.02025097794830799,
+ "learning_rate": 0.0006,
+ "loss": 3.9691858291625977,
+ "step": 2595
+ },
+ {
+ "epoch": 36.05589519650655,
+ "grad_norm": 0.020254382863640785,
+ "learning_rate": 0.0006,
+ "loss": 4.02957820892334,
+ "step": 2596
+ },
+ {
+ "epoch": 36.069868995633186,
+ "grad_norm": 0.019307196140289307,
+ "learning_rate": 0.0006,
+ "loss": 4.033181667327881,
+ "step": 2597
+ },
+ {
+ "epoch": 36.083842794759825,
+ "grad_norm": 0.0198476854711771,
+ "learning_rate": 0.0006,
+ "loss": 4.0816874504089355,
+ "step": 2598
+ },
+ {
+ "epoch": 36.09781659388646,
+ "grad_norm": 0.019916515797376633,
+ "learning_rate": 0.0006,
+ "loss": 4.164889812469482,
+ "step": 2599
+ },
+ {
+ "epoch": 36.1117903930131,
+ "grad_norm": 0.0189144778996706,
+ "learning_rate": 0.0006,
+ "loss": 4.038427352905273,
+ "step": 2600
+ },
+ {
+ "epoch": 36.12576419213974,
+ "grad_norm": 0.01754768192768097,
+ "learning_rate": 0.0006,
+ "loss": 4.010427474975586,
+ "step": 2601
+ },
+ {
+ "epoch": 36.13973799126637,
+ "grad_norm": 0.016862990334630013,
+ "learning_rate": 0.0006,
+ "loss": 4.035970211029053,
+ "step": 2602
+ },
+ {
+ "epoch": 36.15371179039301,
+ "grad_norm": 0.019616752862930298,
+ "learning_rate": 0.0006,
+ "loss": 4.079898834228516,
+ "step": 2603
+ },
+ {
+ "epoch": 36.16768558951965,
+ "grad_norm": 0.020131465047597885,
+ "learning_rate": 0.0006,
+ "loss": 4.144684791564941,
+ "step": 2604
+ },
+ {
+ "epoch": 36.18165938864629,
+ "grad_norm": 0.02072521299123764,
+ "learning_rate": 0.0006,
+ "loss": 4.013660430908203,
+ "step": 2605
+ },
+ {
+ "epoch": 36.19563318777293,
+ "grad_norm": 0.01924615167081356,
+ "learning_rate": 0.0006,
+ "loss": 4.071398735046387,
+ "step": 2606
+ },
+ {
+ "epoch": 36.209606986899566,
+ "grad_norm": 0.02013954520225525,
+ "learning_rate": 0.0006,
+ "loss": 4.127094268798828,
+ "step": 2607
+ },
+ {
+ "epoch": 36.223580786026204,
+ "grad_norm": 0.02042008936405182,
+ "learning_rate": 0.0006,
+ "loss": 4.154261589050293,
+ "step": 2608
+ },
+ {
+ "epoch": 36.237554585152836,
+ "grad_norm": 0.021998876705765724,
+ "learning_rate": 0.0006,
+ "loss": 4.204150199890137,
+ "step": 2609
+ },
+ {
+ "epoch": 36.251528384279474,
+ "grad_norm": 0.022946683689951897,
+ "learning_rate": 0.0006,
+ "loss": 4.013632774353027,
+ "step": 2610
+ },
+ {
+ "epoch": 36.26550218340611,
+ "grad_norm": 0.02061830274760723,
+ "learning_rate": 0.0006,
+ "loss": 4.171084403991699,
+ "step": 2611
+ },
+ {
+ "epoch": 36.27947598253275,
+ "grad_norm": 0.01883152686059475,
+ "learning_rate": 0.0006,
+ "loss": 4.149848461151123,
+ "step": 2612
+ },
+ {
+ "epoch": 36.29344978165939,
+ "grad_norm": 0.019945282489061356,
+ "learning_rate": 0.0006,
+ "loss": 4.174656867980957,
+ "step": 2613
+ },
+ {
+ "epoch": 36.30742358078603,
+ "grad_norm": 0.020751051604747772,
+ "learning_rate": 0.0006,
+ "loss": 4.080747127532959,
+ "step": 2614
+ },
+ {
+ "epoch": 36.32139737991266,
+ "grad_norm": 0.01879691146314144,
+ "learning_rate": 0.0006,
+ "loss": 4.057356357574463,
+ "step": 2615
+ },
+ {
+ "epoch": 36.3353711790393,
+ "grad_norm": 0.019037071615457535,
+ "learning_rate": 0.0006,
+ "loss": 4.204607009887695,
+ "step": 2616
+ },
+ {
+ "epoch": 36.34934497816594,
+ "grad_norm": 0.01872078701853752,
+ "learning_rate": 0.0006,
+ "loss": 4.050586700439453,
+ "step": 2617
+ },
+ {
+ "epoch": 36.36331877729258,
+ "grad_norm": 0.018092811107635498,
+ "learning_rate": 0.0006,
+ "loss": 4.102832794189453,
+ "step": 2618
+ },
+ {
+ "epoch": 36.377292576419215,
+ "grad_norm": 0.017010487616062164,
+ "learning_rate": 0.0006,
+ "loss": 4.191788673400879,
+ "step": 2619
+ },
+ {
+ "epoch": 36.391266375545854,
+ "grad_norm": 0.01691143773496151,
+ "learning_rate": 0.0006,
+ "loss": 4.160456657409668,
+ "step": 2620
+ },
+ {
+ "epoch": 36.40524017467249,
+ "grad_norm": 0.017046816647052765,
+ "learning_rate": 0.0006,
+ "loss": 4.073828220367432,
+ "step": 2621
+ },
+ {
+ "epoch": 36.419213973799124,
+ "grad_norm": 0.01707981713116169,
+ "learning_rate": 0.0006,
+ "loss": 4.0980095863342285,
+ "step": 2622
+ },
+ {
+ "epoch": 36.43318777292576,
+ "grad_norm": 0.01630452275276184,
+ "learning_rate": 0.0006,
+ "loss": 4.101321220397949,
+ "step": 2623
+ },
+ {
+ "epoch": 36.4471615720524,
+ "grad_norm": 0.01576915942132473,
+ "learning_rate": 0.0006,
+ "loss": 4.158249855041504,
+ "step": 2624
+ },
+ {
+ "epoch": 36.46113537117904,
+ "grad_norm": 0.016179712489247322,
+ "learning_rate": 0.0006,
+ "loss": 4.077852249145508,
+ "step": 2625
+ },
+ {
+ "epoch": 36.47510917030568,
+ "grad_norm": 0.01609169878065586,
+ "learning_rate": 0.0006,
+ "loss": 4.187317848205566,
+ "step": 2626
+ },
+ {
+ "epoch": 36.48908296943232,
+ "grad_norm": 0.017148254439234734,
+ "learning_rate": 0.0006,
+ "loss": 4.109071731567383,
+ "step": 2627
+ },
+ {
+ "epoch": 36.50305676855895,
+ "grad_norm": 0.015842624008655548,
+ "learning_rate": 0.0006,
+ "loss": 4.145096302032471,
+ "step": 2628
+ },
+ {
+ "epoch": 36.51703056768559,
+ "grad_norm": 0.014906194061040878,
+ "learning_rate": 0.0006,
+ "loss": 4.028440475463867,
+ "step": 2629
+ },
+ {
+ "epoch": 36.531004366812226,
+ "grad_norm": 0.01485038734972477,
+ "learning_rate": 0.0006,
+ "loss": 4.175201416015625,
+ "step": 2630
+ },
+ {
+ "epoch": 36.544978165938865,
+ "grad_norm": 0.015834150835871696,
+ "learning_rate": 0.0006,
+ "loss": 4.076434135437012,
+ "step": 2631
+ },
+ {
+ "epoch": 36.5589519650655,
+ "grad_norm": 0.015411358326673508,
+ "learning_rate": 0.0006,
+ "loss": 4.03912878036499,
+ "step": 2632
+ },
+ {
+ "epoch": 36.57292576419214,
+ "grad_norm": 0.014762846753001213,
+ "learning_rate": 0.0006,
+ "loss": 4.019827365875244,
+ "step": 2633
+ },
+ {
+ "epoch": 36.58689956331878,
+ "grad_norm": 0.015269497409462929,
+ "learning_rate": 0.0006,
+ "loss": 4.116957187652588,
+ "step": 2634
+ },
+ {
+ "epoch": 36.60087336244541,
+ "grad_norm": 0.015171617269515991,
+ "learning_rate": 0.0006,
+ "loss": 4.2069926261901855,
+ "step": 2635
+ },
+ {
+ "epoch": 36.61484716157205,
+ "grad_norm": 0.014770902693271637,
+ "learning_rate": 0.0006,
+ "loss": 4.092021942138672,
+ "step": 2636
+ },
+ {
+ "epoch": 36.62882096069869,
+ "grad_norm": 0.014957531355321407,
+ "learning_rate": 0.0006,
+ "loss": 4.16819953918457,
+ "step": 2637
+ },
+ {
+ "epoch": 36.64279475982533,
+ "grad_norm": 0.014774451963603497,
+ "learning_rate": 0.0006,
+ "loss": 4.046073913574219,
+ "step": 2638
+ },
+ {
+ "epoch": 36.65676855895197,
+ "grad_norm": 0.014320445246994495,
+ "learning_rate": 0.0006,
+ "loss": 3.977792501449585,
+ "step": 2639
+ },
+ {
+ "epoch": 36.670742358078606,
+ "grad_norm": 0.01413130946457386,
+ "learning_rate": 0.0006,
+ "loss": 4.130649566650391,
+ "step": 2640
+ },
+ {
+ "epoch": 36.68471615720524,
+ "grad_norm": 0.014336307533085346,
+ "learning_rate": 0.0006,
+ "loss": 4.161357402801514,
+ "step": 2641
+ },
+ {
+ "epoch": 36.698689956331876,
+ "grad_norm": 0.014516931027173996,
+ "learning_rate": 0.0006,
+ "loss": 4.14248514175415,
+ "step": 2642
+ },
+ {
+ "epoch": 36.712663755458514,
+ "grad_norm": 0.014377126470208168,
+ "learning_rate": 0.0006,
+ "loss": 4.062711715698242,
+ "step": 2643
+ },
+ {
+ "epoch": 36.72663755458515,
+ "grad_norm": 0.014332668855786324,
+ "learning_rate": 0.0006,
+ "loss": 4.06423807144165,
+ "step": 2644
+ },
+ {
+ "epoch": 36.74061135371179,
+ "grad_norm": 0.014030571095645428,
+ "learning_rate": 0.0006,
+ "loss": 4.178613662719727,
+ "step": 2645
+ },
+ {
+ "epoch": 36.75458515283843,
+ "grad_norm": 0.014376776292920113,
+ "learning_rate": 0.0006,
+ "loss": 4.073130130767822,
+ "step": 2646
+ },
+ {
+ "epoch": 36.76855895196506,
+ "grad_norm": 0.013827716931700706,
+ "learning_rate": 0.0006,
+ "loss": 4.043869972229004,
+ "step": 2647
+ },
+ {
+ "epoch": 36.7825327510917,
+ "grad_norm": 0.015548762865364552,
+ "learning_rate": 0.0006,
+ "loss": 4.134049415588379,
+ "step": 2648
+ },
+ {
+ "epoch": 36.79650655021834,
+ "grad_norm": 0.017041970044374466,
+ "learning_rate": 0.0006,
+ "loss": 4.205024242401123,
+ "step": 2649
+ },
+ {
+ "epoch": 36.81048034934498,
+ "grad_norm": 0.017638275399804115,
+ "learning_rate": 0.0006,
+ "loss": 4.199924468994141,
+ "step": 2650
+ },
+ {
+ "epoch": 36.82445414847162,
+ "grad_norm": 0.017771627753973007,
+ "learning_rate": 0.0006,
+ "loss": 4.15706729888916,
+ "step": 2651
+ },
+ {
+ "epoch": 36.838427947598255,
+ "grad_norm": 0.015933562070131302,
+ "learning_rate": 0.0006,
+ "loss": 4.220742225646973,
+ "step": 2652
+ },
+ {
+ "epoch": 36.852401746724894,
+ "grad_norm": 0.01668144389986992,
+ "learning_rate": 0.0006,
+ "loss": 4.161426544189453,
+ "step": 2653
+ },
+ {
+ "epoch": 36.866375545851525,
+ "grad_norm": 0.018014680594205856,
+ "learning_rate": 0.0006,
+ "loss": 4.183046817779541,
+ "step": 2654
+ },
+ {
+ "epoch": 36.880349344978164,
+ "grad_norm": 0.018808670341968536,
+ "learning_rate": 0.0006,
+ "loss": 4.105293273925781,
+ "step": 2655
+ },
+ {
+ "epoch": 36.8943231441048,
+ "grad_norm": 0.018003476783633232,
+ "learning_rate": 0.0006,
+ "loss": 4.103171348571777,
+ "step": 2656
+ },
+ {
+ "epoch": 36.90829694323144,
+ "grad_norm": 0.016865914687514305,
+ "learning_rate": 0.0006,
+ "loss": 4.125634670257568,
+ "step": 2657
+ },
+ {
+ "epoch": 36.92227074235808,
+ "grad_norm": 0.018371153622865677,
+ "learning_rate": 0.0006,
+ "loss": 4.192461013793945,
+ "step": 2658
+ },
+ {
+ "epoch": 36.93624454148472,
+ "grad_norm": 0.019045377150177956,
+ "learning_rate": 0.0006,
+ "loss": 4.198105812072754,
+ "step": 2659
+ },
+ {
+ "epoch": 36.95021834061135,
+ "grad_norm": 0.01594168320298195,
+ "learning_rate": 0.0006,
+ "loss": 4.169788360595703,
+ "step": 2660
+ },
+ {
+ "epoch": 36.96419213973799,
+ "grad_norm": 0.015645820647478104,
+ "learning_rate": 0.0006,
+ "loss": 4.047277450561523,
+ "step": 2661
+ },
+ {
+ "epoch": 36.97816593886463,
+ "grad_norm": 0.016313156113028526,
+ "learning_rate": 0.0006,
+ "loss": 4.087392807006836,
+ "step": 2662
+ },
+ {
+ "epoch": 36.992139737991266,
+ "grad_norm": 0.017050625756382942,
+ "learning_rate": 0.0006,
+ "loss": 4.0908002853393555,
+ "step": 2663
+ },
+ {
+ "epoch": 37.0,
+ "grad_norm": 0.017730968073010445,
+ "learning_rate": 0.0006,
+ "loss": 4.110957622528076,
+ "step": 2664
+ },
+ {
+ "epoch": 37.0,
+ "eval_loss": 4.563793182373047,
+ "eval_runtime": 56.5922,
+ "eval_samples_per_second": 43.151,
+ "eval_steps_per_second": 1.361,
+ "step": 2664
+ },
+ {
+ "epoch": 37.01397379912664,
+ "grad_norm": 0.015687910839915276,
+ "learning_rate": 0.0006,
+ "loss": 4.143589019775391,
+ "step": 2665
+ },
+ {
+ "epoch": 37.02794759825328,
+ "grad_norm": 0.017451366409659386,
+ "learning_rate": 0.0006,
+ "loss": 4.0326666831970215,
+ "step": 2666
+ },
+ {
+ "epoch": 37.041921397379916,
+ "grad_norm": 0.01909538544714451,
+ "learning_rate": 0.0006,
+ "loss": 3.9706249237060547,
+ "step": 2667
+ },
+ {
+ "epoch": 37.05589519650655,
+ "grad_norm": 0.01988379657268524,
+ "learning_rate": 0.0006,
+ "loss": 3.9531009197235107,
+ "step": 2668
+ },
+ {
+ "epoch": 37.069868995633186,
+ "grad_norm": 0.02035115472972393,
+ "learning_rate": 0.0006,
+ "loss": 3.962214469909668,
+ "step": 2669
+ },
+ {
+ "epoch": 37.083842794759825,
+ "grad_norm": 0.01952001079916954,
+ "learning_rate": 0.0006,
+ "loss": 4.182262897491455,
+ "step": 2670
+ },
+ {
+ "epoch": 37.09781659388646,
+ "grad_norm": 0.01854325644671917,
+ "learning_rate": 0.0006,
+ "loss": 4.127663612365723,
+ "step": 2671
+ },
+ {
+ "epoch": 37.1117903930131,
+ "grad_norm": 0.017757616937160492,
+ "learning_rate": 0.0006,
+ "loss": 4.100804805755615,
+ "step": 2672
+ },
+ {
+ "epoch": 37.12576419213974,
+ "grad_norm": 0.016712479293346405,
+ "learning_rate": 0.0006,
+ "loss": 4.026284217834473,
+ "step": 2673
+ },
+ {
+ "epoch": 37.13973799126637,
+ "grad_norm": 0.0168139711022377,
+ "learning_rate": 0.0006,
+ "loss": 4.183882713317871,
+ "step": 2674
+ },
+ {
+ "epoch": 37.15371179039301,
+ "grad_norm": 0.017060423269867897,
+ "learning_rate": 0.0006,
+ "loss": 4.131832599639893,
+ "step": 2675
+ },
+ {
+ "epoch": 37.16768558951965,
+ "grad_norm": 0.01726631447672844,
+ "learning_rate": 0.0006,
+ "loss": 3.9156057834625244,
+ "step": 2676
+ },
+ {
+ "epoch": 37.18165938864629,
+ "grad_norm": 0.016348622739315033,
+ "learning_rate": 0.0006,
+ "loss": 4.1731367111206055,
+ "step": 2677
+ },
+ {
+ "epoch": 37.19563318777293,
+ "grad_norm": 0.016640059649944305,
+ "learning_rate": 0.0006,
+ "loss": 4.10997200012207,
+ "step": 2678
+ },
+ {
+ "epoch": 37.209606986899566,
+ "grad_norm": 0.01601342298090458,
+ "learning_rate": 0.0006,
+ "loss": 4.081145286560059,
+ "step": 2679
+ },
+ {
+ "epoch": 37.223580786026204,
+ "grad_norm": 0.01735631190240383,
+ "learning_rate": 0.0006,
+ "loss": 4.202108860015869,
+ "step": 2680
+ },
+ {
+ "epoch": 37.237554585152836,
+ "grad_norm": 0.019164685159921646,
+ "learning_rate": 0.0006,
+ "loss": 4.104109764099121,
+ "step": 2681
+ },
+ {
+ "epoch": 37.251528384279474,
+ "grad_norm": 0.01881759613752365,
+ "learning_rate": 0.0006,
+ "loss": 4.145532131195068,
+ "step": 2682
+ },
+ {
+ "epoch": 37.26550218340611,
+ "grad_norm": 0.017656555399298668,
+ "learning_rate": 0.0006,
+ "loss": 4.012782096862793,
+ "step": 2683
+ },
+ {
+ "epoch": 37.27947598253275,
+ "grad_norm": 0.01624823734164238,
+ "learning_rate": 0.0006,
+ "loss": 4.001727104187012,
+ "step": 2684
+ },
+ {
+ "epoch": 37.29344978165939,
+ "grad_norm": 0.016973139718174934,
+ "learning_rate": 0.0006,
+ "loss": 4.095399856567383,
+ "step": 2685
+ },
+ {
+ "epoch": 37.30742358078603,
+ "grad_norm": 0.015503671951591969,
+ "learning_rate": 0.0006,
+ "loss": 4.034431457519531,
+ "step": 2686
+ },
+ {
+ "epoch": 37.32139737991266,
+ "grad_norm": 0.015370250679552555,
+ "learning_rate": 0.0006,
+ "loss": 4.075836181640625,
+ "step": 2687
+ },
+ {
+ "epoch": 37.3353711790393,
+ "grad_norm": 0.01564021222293377,
+ "learning_rate": 0.0006,
+ "loss": 4.125202178955078,
+ "step": 2688
+ },
+ {
+ "epoch": 37.34934497816594,
+ "grad_norm": 0.016717975959181786,
+ "learning_rate": 0.0006,
+ "loss": 4.153017044067383,
+ "step": 2689
+ },
+ {
+ "epoch": 37.36331877729258,
+ "grad_norm": 0.015263617038726807,
+ "learning_rate": 0.0006,
+ "loss": 4.021814823150635,
+ "step": 2690
+ },
+ {
+ "epoch": 37.377292576419215,
+ "grad_norm": 0.016497142612934113,
+ "learning_rate": 0.0006,
+ "loss": 4.069095611572266,
+ "step": 2691
+ },
+ {
+ "epoch": 37.391266375545854,
+ "grad_norm": 0.01669355481863022,
+ "learning_rate": 0.0006,
+ "loss": 4.174587726593018,
+ "step": 2692
+ },
+ {
+ "epoch": 37.40524017467249,
+ "grad_norm": 0.019695773720741272,
+ "learning_rate": 0.0006,
+ "loss": 4.0073041915893555,
+ "step": 2693
+ },
+ {
+ "epoch": 37.419213973799124,
+ "grad_norm": 0.023720644414424896,
+ "learning_rate": 0.0006,
+ "loss": 4.111948013305664,
+ "step": 2694
+ },
+ {
+ "epoch": 37.43318777292576,
+ "grad_norm": 0.02399350143969059,
+ "learning_rate": 0.0006,
+ "loss": 4.081164360046387,
+ "step": 2695
+ },
+ {
+ "epoch": 37.4471615720524,
+ "grad_norm": 0.019630271941423416,
+ "learning_rate": 0.0006,
+ "loss": 4.065731048583984,
+ "step": 2696
+ },
+ {
+ "epoch": 37.46113537117904,
+ "grad_norm": 0.019679008051753044,
+ "learning_rate": 0.0006,
+ "loss": 3.993852376937866,
+ "step": 2697
+ },
+ {
+ "epoch": 37.47510917030568,
+ "grad_norm": 0.020597299560904503,
+ "learning_rate": 0.0006,
+ "loss": 4.275555610656738,
+ "step": 2698
+ },
+ {
+ "epoch": 37.48908296943232,
+ "grad_norm": 0.01983889564871788,
+ "learning_rate": 0.0006,
+ "loss": 4.254708290100098,
+ "step": 2699
+ },
+ {
+ "epoch": 37.50305676855895,
+ "grad_norm": 0.019150668755173683,
+ "learning_rate": 0.0006,
+ "loss": 4.097885608673096,
+ "step": 2700
+ },
+ {
+ "epoch": 37.51703056768559,
+ "grad_norm": 0.018847715109586716,
+ "learning_rate": 0.0006,
+ "loss": 4.054081439971924,
+ "step": 2701
+ },
+ {
+ "epoch": 37.531004366812226,
+ "grad_norm": 0.019607288762927055,
+ "learning_rate": 0.0006,
+ "loss": 3.9891440868377686,
+ "step": 2702
+ },
+ {
+ "epoch": 37.544978165938865,
+ "grad_norm": 0.018863150849938393,
+ "learning_rate": 0.0006,
+ "loss": 4.084877014160156,
+ "step": 2703
+ },
+ {
+ "epoch": 37.5589519650655,
+ "grad_norm": 0.017340239137411118,
+ "learning_rate": 0.0006,
+ "loss": 4.113290786743164,
+ "step": 2704
+ },
+ {
+ "epoch": 37.57292576419214,
+ "grad_norm": 0.015628693625330925,
+ "learning_rate": 0.0006,
+ "loss": 4.1483001708984375,
+ "step": 2705
+ },
+ {
+ "epoch": 37.58689956331878,
+ "grad_norm": 0.016493283212184906,
+ "learning_rate": 0.0006,
+ "loss": 4.1040143966674805,
+ "step": 2706
+ },
+ {
+ "epoch": 37.60087336244541,
+ "grad_norm": 0.016732443124055862,
+ "learning_rate": 0.0006,
+ "loss": 4.07205057144165,
+ "step": 2707
+ },
+ {
+ "epoch": 37.61484716157205,
+ "grad_norm": 0.017176156863570213,
+ "learning_rate": 0.0006,
+ "loss": 4.1279754638671875,
+ "step": 2708
+ },
+ {
+ "epoch": 37.62882096069869,
+ "grad_norm": 0.015730608254671097,
+ "learning_rate": 0.0006,
+ "loss": 4.086658000946045,
+ "step": 2709
+ },
+ {
+ "epoch": 37.64279475982533,
+ "grad_norm": 0.016111090779304504,
+ "learning_rate": 0.0006,
+ "loss": 4.064335823059082,
+ "step": 2710
+ },
+ {
+ "epoch": 37.65676855895197,
+ "grad_norm": 0.01769379712641239,
+ "learning_rate": 0.0006,
+ "loss": 4.116047382354736,
+ "step": 2711
+ },
+ {
+ "epoch": 37.670742358078606,
+ "grad_norm": 0.01683102548122406,
+ "learning_rate": 0.0006,
+ "loss": 4.095986843109131,
+ "step": 2712
+ },
+ {
+ "epoch": 37.68471615720524,
+ "grad_norm": 0.015726450830698013,
+ "learning_rate": 0.0006,
+ "loss": 4.10775089263916,
+ "step": 2713
+ },
+ {
+ "epoch": 37.698689956331876,
+ "grad_norm": 0.014760235324501991,
+ "learning_rate": 0.0006,
+ "loss": 4.1234893798828125,
+ "step": 2714
+ },
+ {
+ "epoch": 37.712663755458514,
+ "grad_norm": 0.014933438040316105,
+ "learning_rate": 0.0006,
+ "loss": 4.079200744628906,
+ "step": 2715
+ },
+ {
+ "epoch": 37.72663755458515,
+ "grad_norm": 0.015723997727036476,
+ "learning_rate": 0.0006,
+ "loss": 4.07865047454834,
+ "step": 2716
+ },
+ {
+ "epoch": 37.74061135371179,
+ "grad_norm": 0.014539001509547234,
+ "learning_rate": 0.0006,
+ "loss": 4.036242485046387,
+ "step": 2717
+ },
+ {
+ "epoch": 37.75458515283843,
+ "grad_norm": 0.014844979159533978,
+ "learning_rate": 0.0006,
+ "loss": 4.1051483154296875,
+ "step": 2718
+ },
+ {
+ "epoch": 37.76855895196506,
+ "grad_norm": 0.015939844772219658,
+ "learning_rate": 0.0006,
+ "loss": 4.017253875732422,
+ "step": 2719
+ },
+ {
+ "epoch": 37.7825327510917,
+ "grad_norm": 0.014771555550396442,
+ "learning_rate": 0.0006,
+ "loss": 4.0694661140441895,
+ "step": 2720
+ },
+ {
+ "epoch": 37.79650655021834,
+ "grad_norm": 0.015777921304106712,
+ "learning_rate": 0.0006,
+ "loss": 3.947587490081787,
+ "step": 2721
+ },
+ {
+ "epoch": 37.81048034934498,
+ "grad_norm": 0.017451291903853416,
+ "learning_rate": 0.0006,
+ "loss": 4.093625068664551,
+ "step": 2722
+ },
+ {
+ "epoch": 37.82445414847162,
+ "grad_norm": 0.016829246655106544,
+ "learning_rate": 0.0006,
+ "loss": 4.09030818939209,
+ "step": 2723
+ },
+ {
+ "epoch": 37.838427947598255,
+ "grad_norm": 0.016984397545456886,
+ "learning_rate": 0.0006,
+ "loss": 4.140626907348633,
+ "step": 2724
+ },
+ {
+ "epoch": 37.852401746724894,
+ "grad_norm": 0.01732603646814823,
+ "learning_rate": 0.0006,
+ "loss": 4.128276824951172,
+ "step": 2725
+ },
+ {
+ "epoch": 37.866375545851525,
+ "grad_norm": 0.01744619384407997,
+ "learning_rate": 0.0006,
+ "loss": 4.103130340576172,
+ "step": 2726
+ },
+ {
+ "epoch": 37.880349344978164,
+ "grad_norm": 0.016331320628523827,
+ "learning_rate": 0.0006,
+ "loss": 4.037951469421387,
+ "step": 2727
+ },
+ {
+ "epoch": 37.8943231441048,
+ "grad_norm": 0.015919430181384087,
+ "learning_rate": 0.0006,
+ "loss": 4.121513366699219,
+ "step": 2728
+ },
+ {
+ "epoch": 37.90829694323144,
+ "grad_norm": 0.016153695061802864,
+ "learning_rate": 0.0006,
+ "loss": 4.099384307861328,
+ "step": 2729
+ },
+ {
+ "epoch": 37.92227074235808,
+ "grad_norm": 0.01628800667822361,
+ "learning_rate": 0.0006,
+ "loss": 4.092325687408447,
+ "step": 2730
+ },
+ {
+ "epoch": 37.93624454148472,
+ "grad_norm": 0.017285684123635292,
+ "learning_rate": 0.0006,
+ "loss": 3.9828906059265137,
+ "step": 2731
+ },
+ {
+ "epoch": 37.95021834061135,
+ "grad_norm": 0.016047554090619087,
+ "learning_rate": 0.0006,
+ "loss": 4.100419044494629,
+ "step": 2732
+ },
+ {
+ "epoch": 37.96419213973799,
+ "grad_norm": 0.01562986709177494,
+ "learning_rate": 0.0006,
+ "loss": 4.065832614898682,
+ "step": 2733
+ },
+ {
+ "epoch": 37.97816593886463,
+ "grad_norm": 0.01677163504064083,
+ "learning_rate": 0.0006,
+ "loss": 4.077518463134766,
+ "step": 2734
+ },
+ {
+ "epoch": 37.992139737991266,
+ "grad_norm": 0.01755492016673088,
+ "learning_rate": 0.0006,
+ "loss": 4.1076154708862305,
+ "step": 2735
+ },
+ {
+ "epoch": 38.0,
+ "grad_norm": 0.01929762214422226,
+ "learning_rate": 0.0006,
+ "loss": 4.09603214263916,
+ "step": 2736
+ },
+ {
+ "epoch": 38.0,
+ "eval_loss": 4.582442283630371,
+ "eval_runtime": 62.8649,
+ "eval_samples_per_second": 38.845,
+ "eval_steps_per_second": 1.225,
+ "step": 2736
+ },
+ {
+ "epoch": 38.01397379912664,
+ "grad_norm": 0.019129136577248573,
+ "learning_rate": 0.0006,
+ "loss": 4.095005989074707,
+ "step": 2737
+ },
+ {
+ "epoch": 38.02794759825328,
+ "grad_norm": 0.019167408347129822,
+ "learning_rate": 0.0006,
+ "loss": 4.054426670074463,
+ "step": 2738
+ },
+ {
+ "epoch": 38.041921397379916,
+ "grad_norm": 0.01789460889995098,
+ "learning_rate": 0.0006,
+ "loss": 4.0434417724609375,
+ "step": 2739
+ },
+ {
+ "epoch": 38.05589519650655,
+ "grad_norm": 0.015947027131915092,
+ "learning_rate": 0.0006,
+ "loss": 4.074516296386719,
+ "step": 2740
+ },
+ {
+ "epoch": 38.069868995633186,
+ "grad_norm": 0.016667252406477928,
+ "learning_rate": 0.0006,
+ "loss": 4.108545303344727,
+ "step": 2741
+ },
+ {
+ "epoch": 38.083842794759825,
+ "grad_norm": 0.016014399006962776,
+ "learning_rate": 0.0006,
+ "loss": 4.026925086975098,
+ "step": 2742
+ },
+ {
+ "epoch": 38.09781659388646,
+ "grad_norm": 0.015397695824503899,
+ "learning_rate": 0.0006,
+ "loss": 4.163970470428467,
+ "step": 2743
+ },
+ {
+ "epoch": 38.1117903930131,
+ "grad_norm": 0.01563289947807789,
+ "learning_rate": 0.0006,
+ "loss": 4.116578578948975,
+ "step": 2744
+ },
+ {
+ "epoch": 38.12576419213974,
+ "grad_norm": 0.015512318350374699,
+ "learning_rate": 0.0006,
+ "loss": 4.096311569213867,
+ "step": 2745
+ },
+ {
+ "epoch": 38.13973799126637,
+ "grad_norm": 0.015426642261445522,
+ "learning_rate": 0.0006,
+ "loss": 4.017065048217773,
+ "step": 2746
+ },
+ {
+ "epoch": 38.15371179039301,
+ "grad_norm": 0.015115435235202312,
+ "learning_rate": 0.0006,
+ "loss": 4.005434989929199,
+ "step": 2747
+ },
+ {
+ "epoch": 38.16768558951965,
+ "grad_norm": 0.015235894359648228,
+ "learning_rate": 0.0006,
+ "loss": 3.982743263244629,
+ "step": 2748
+ },
+ {
+ "epoch": 38.18165938864629,
+ "grad_norm": 0.016095101833343506,
+ "learning_rate": 0.0006,
+ "loss": 4.0849385261535645,
+ "step": 2749
+ },
+ {
+ "epoch": 38.19563318777293,
+ "grad_norm": 0.015201197937130928,
+ "learning_rate": 0.0006,
+ "loss": 4.103863716125488,
+ "step": 2750
+ },
+ {
+ "epoch": 38.209606986899566,
+ "grad_norm": 0.015687478706240654,
+ "learning_rate": 0.0006,
+ "loss": 4.12074089050293,
+ "step": 2751
+ },
+ {
+ "epoch": 38.223580786026204,
+ "grad_norm": 0.015201589092612267,
+ "learning_rate": 0.0006,
+ "loss": 4.056872367858887,
+ "step": 2752
+ },
+ {
+ "epoch": 38.237554585152836,
+ "grad_norm": 0.015154894441366196,
+ "learning_rate": 0.0006,
+ "loss": 4.089670181274414,
+ "step": 2753
+ },
+ {
+ "epoch": 38.251528384279474,
+ "grad_norm": 0.01586608588695526,
+ "learning_rate": 0.0006,
+ "loss": 4.155317306518555,
+ "step": 2754
+ },
+ {
+ "epoch": 38.26550218340611,
+ "grad_norm": 0.01609654724597931,
+ "learning_rate": 0.0006,
+ "loss": 3.995800495147705,
+ "step": 2755
+ },
+ {
+ "epoch": 38.27947598253275,
+ "grad_norm": 0.015495185740292072,
+ "learning_rate": 0.0006,
+ "loss": 3.995769739151001,
+ "step": 2756
+ },
+ {
+ "epoch": 38.29344978165939,
+ "grad_norm": 0.017042381688952446,
+ "learning_rate": 0.0006,
+ "loss": 4.132832050323486,
+ "step": 2757
+ },
+ {
+ "epoch": 38.30742358078603,
+ "grad_norm": 0.018706614151597023,
+ "learning_rate": 0.0006,
+ "loss": 4.038820743560791,
+ "step": 2758
+ },
+ {
+ "epoch": 38.32139737991266,
+ "grad_norm": 0.018316026777029037,
+ "learning_rate": 0.0006,
+ "loss": 3.972052574157715,
+ "step": 2759
+ },
+ {
+ "epoch": 38.3353711790393,
+ "grad_norm": 0.018375389277935028,
+ "learning_rate": 0.0006,
+ "loss": 4.152097702026367,
+ "step": 2760
+ },
+ {
+ "epoch": 38.34934497816594,
+ "grad_norm": 0.016879886388778687,
+ "learning_rate": 0.0006,
+ "loss": 4.020869731903076,
+ "step": 2761
+ },
+ {
+ "epoch": 38.36331877729258,
+ "grad_norm": 0.015728183090686798,
+ "learning_rate": 0.0006,
+ "loss": 4.008016586303711,
+ "step": 2762
+ },
+ {
+ "epoch": 38.377292576419215,
+ "grad_norm": 0.016147758811712265,
+ "learning_rate": 0.0006,
+ "loss": 4.035854339599609,
+ "step": 2763
+ },
+ {
+ "epoch": 38.391266375545854,
+ "grad_norm": 0.017774797976017,
+ "learning_rate": 0.0006,
+ "loss": 4.122506141662598,
+ "step": 2764
+ },
+ {
+ "epoch": 38.40524017467249,
+ "grad_norm": 0.016590069979429245,
+ "learning_rate": 0.0006,
+ "loss": 4.082847595214844,
+ "step": 2765
+ },
+ {
+ "epoch": 38.419213973799124,
+ "grad_norm": 0.016544528305530548,
+ "learning_rate": 0.0006,
+ "loss": 4.184270858764648,
+ "step": 2766
+ },
+ {
+ "epoch": 38.43318777292576,
+ "grad_norm": 0.017972925677895546,
+ "learning_rate": 0.0006,
+ "loss": 3.987558603286743,
+ "step": 2767
+ },
+ {
+ "epoch": 38.4471615720524,
+ "grad_norm": 0.018562227487564087,
+ "learning_rate": 0.0006,
+ "loss": 4.1193461418151855,
+ "step": 2768
+ },
+ {
+ "epoch": 38.46113537117904,
+ "grad_norm": 0.018896685913205147,
+ "learning_rate": 0.0006,
+ "loss": 4.057808876037598,
+ "step": 2769
+ },
+ {
+ "epoch": 38.47510917030568,
+ "grad_norm": 0.016793809831142426,
+ "learning_rate": 0.0006,
+ "loss": 4.069523334503174,
+ "step": 2770
+ },
+ {
+ "epoch": 38.48908296943232,
+ "grad_norm": 0.01698676310479641,
+ "learning_rate": 0.0006,
+ "loss": 3.997624158859253,
+ "step": 2771
+ },
+ {
+ "epoch": 38.50305676855895,
+ "grad_norm": 0.016784202307462692,
+ "learning_rate": 0.0006,
+ "loss": 4.174795627593994,
+ "step": 2772
+ },
+ {
+ "epoch": 38.51703056768559,
+ "grad_norm": 0.016785014420747757,
+ "learning_rate": 0.0006,
+ "loss": 3.999321222305298,
+ "step": 2773
+ },
+ {
+ "epoch": 38.531004366812226,
+ "grad_norm": 0.018615184351801872,
+ "learning_rate": 0.0006,
+ "loss": 4.191237926483154,
+ "step": 2774
+ },
+ {
+ "epoch": 38.544978165938865,
+ "grad_norm": 0.019090790301561356,
+ "learning_rate": 0.0006,
+ "loss": 4.106966018676758,
+ "step": 2775
+ },
+ {
+ "epoch": 38.5589519650655,
+ "grad_norm": 0.018298320472240448,
+ "learning_rate": 0.0006,
+ "loss": 4.063363552093506,
+ "step": 2776
+ },
+ {
+ "epoch": 38.57292576419214,
+ "grad_norm": 0.018541090190410614,
+ "learning_rate": 0.0006,
+ "loss": 4.101193428039551,
+ "step": 2777
+ },
+ {
+ "epoch": 38.58689956331878,
+ "grad_norm": 0.01849970407783985,
+ "learning_rate": 0.0006,
+ "loss": 4.071335792541504,
+ "step": 2778
+ },
+ {
+ "epoch": 38.60087336244541,
+ "grad_norm": 0.018309077247977257,
+ "learning_rate": 0.0006,
+ "loss": 4.245305061340332,
+ "step": 2779
+ },
+ {
+ "epoch": 38.61484716157205,
+ "grad_norm": 0.017681840807199478,
+ "learning_rate": 0.0006,
+ "loss": 4.081173419952393,
+ "step": 2780
+ },
+ {
+ "epoch": 38.62882096069869,
+ "grad_norm": 0.016286224126815796,
+ "learning_rate": 0.0006,
+ "loss": 4.024351596832275,
+ "step": 2781
+ },
+ {
+ "epoch": 38.64279475982533,
+ "grad_norm": 0.015837235376238823,
+ "learning_rate": 0.0006,
+ "loss": 4.028205394744873,
+ "step": 2782
+ },
+ {
+ "epoch": 38.65676855895197,
+ "grad_norm": 0.016062596812844276,
+ "learning_rate": 0.0006,
+ "loss": 4.131122589111328,
+ "step": 2783
+ },
+ {
+ "epoch": 38.670742358078606,
+ "grad_norm": 0.01711239665746689,
+ "learning_rate": 0.0006,
+ "loss": 4.124416351318359,
+ "step": 2784
+ },
+ {
+ "epoch": 38.68471615720524,
+ "grad_norm": 0.015304913744330406,
+ "learning_rate": 0.0006,
+ "loss": 4.056122779846191,
+ "step": 2785
+ },
+ {
+ "epoch": 38.698689956331876,
+ "grad_norm": 0.015971887856721878,
+ "learning_rate": 0.0006,
+ "loss": 4.132555961608887,
+ "step": 2786
+ },
+ {
+ "epoch": 38.712663755458514,
+ "grad_norm": 0.015787221491336823,
+ "learning_rate": 0.0006,
+ "loss": 4.051596164703369,
+ "step": 2787
+ },
+ {
+ "epoch": 38.72663755458515,
+ "grad_norm": 0.01605444774031639,
+ "learning_rate": 0.0006,
+ "loss": 4.191808700561523,
+ "step": 2788
+ },
+ {
+ "epoch": 38.74061135371179,
+ "grad_norm": 0.016383009031414986,
+ "learning_rate": 0.0006,
+ "loss": 4.0648417472839355,
+ "step": 2789
+ },
+ {
+ "epoch": 38.75458515283843,
+ "grad_norm": 0.016135554760694504,
+ "learning_rate": 0.0006,
+ "loss": 3.988563060760498,
+ "step": 2790
+ },
+ {
+ "epoch": 38.76855895196506,
+ "grad_norm": 0.016634158790111542,
+ "learning_rate": 0.0006,
+ "loss": 4.133493423461914,
+ "step": 2791
+ },
+ {
+ "epoch": 38.7825327510917,
+ "grad_norm": 0.01688210666179657,
+ "learning_rate": 0.0006,
+ "loss": 4.091277122497559,
+ "step": 2792
+ },
+ {
+ "epoch": 38.79650655021834,
+ "grad_norm": 0.016102122142910957,
+ "learning_rate": 0.0006,
+ "loss": 4.051219940185547,
+ "step": 2793
+ },
+ {
+ "epoch": 38.81048034934498,
+ "grad_norm": 0.016585038974881172,
+ "learning_rate": 0.0006,
+ "loss": 4.058587074279785,
+ "step": 2794
+ },
+ {
+ "epoch": 38.82445414847162,
+ "grad_norm": 0.017700951546430588,
+ "learning_rate": 0.0006,
+ "loss": 4.016030311584473,
+ "step": 2795
+ },
+ {
+ "epoch": 38.838427947598255,
+ "grad_norm": 0.017922203987836838,
+ "learning_rate": 0.0006,
+ "loss": 4.079977989196777,
+ "step": 2796
+ },
+ {
+ "epoch": 38.852401746724894,
+ "grad_norm": 0.017299890518188477,
+ "learning_rate": 0.0006,
+ "loss": 4.062984943389893,
+ "step": 2797
+ },
+ {
+ "epoch": 38.866375545851525,
+ "grad_norm": 0.018648581579327583,
+ "learning_rate": 0.0006,
+ "loss": 4.065717697143555,
+ "step": 2798
+ },
+ {
+ "epoch": 38.880349344978164,
+ "grad_norm": 0.019057920202612877,
+ "learning_rate": 0.0006,
+ "loss": 4.227892875671387,
+ "step": 2799
+ },
+ {
+ "epoch": 38.8943231441048,
+ "grad_norm": 0.018824422731995583,
+ "learning_rate": 0.0006,
+ "loss": 4.116058349609375,
+ "step": 2800
+ },
+ {
+ "epoch": 38.90829694323144,
+ "grad_norm": 0.019084978848695755,
+ "learning_rate": 0.0006,
+ "loss": 4.061677932739258,
+ "step": 2801
+ },
+ {
+ "epoch": 38.92227074235808,
+ "grad_norm": 0.017842981964349747,
+ "learning_rate": 0.0006,
+ "loss": 4.083813667297363,
+ "step": 2802
+ },
+ {
+ "epoch": 38.93624454148472,
+ "grad_norm": 0.01880858838558197,
+ "learning_rate": 0.0006,
+ "loss": 4.152561187744141,
+ "step": 2803
+ },
+ {
+ "epoch": 38.95021834061135,
+ "grad_norm": 0.017400087788701057,
+ "learning_rate": 0.0006,
+ "loss": 4.154635429382324,
+ "step": 2804
+ },
+ {
+ "epoch": 38.96419213973799,
+ "grad_norm": 0.017682507634162903,
+ "learning_rate": 0.0006,
+ "loss": 4.151128768920898,
+ "step": 2805
+ },
+ {
+ "epoch": 38.97816593886463,
+ "grad_norm": 0.015803052112460136,
+ "learning_rate": 0.0006,
+ "loss": 4.10498571395874,
+ "step": 2806
+ },
+ {
+ "epoch": 38.992139737991266,
+ "grad_norm": 0.01582511141896248,
+ "learning_rate": 0.0006,
+ "loss": 4.140159606933594,
+ "step": 2807
+ },
+ {
+ "epoch": 39.0,
+ "grad_norm": 0.01925666816532612,
+ "learning_rate": 0.0006,
+ "loss": 4.126394271850586,
+ "step": 2808
+ },
+ {
+ "epoch": 39.0,
+ "eval_loss": 4.615617275238037,
+ "eval_runtime": 56.8895,
+ "eval_samples_per_second": 42.925,
+ "eval_steps_per_second": 1.354,
+ "step": 2808
+ },
+ {
+ "epoch": 39.01397379912664,
+ "grad_norm": 0.01677853614091873,
+ "learning_rate": 0.0006,
+ "loss": 4.102395057678223,
+ "step": 2809
+ },
+ {
+ "epoch": 39.02794759825328,
+ "grad_norm": 0.017901137471199036,
+ "learning_rate": 0.0006,
+ "loss": 3.976609230041504,
+ "step": 2810
+ },
+ {
+ "epoch": 39.041921397379916,
+ "grad_norm": 0.017978668212890625,
+ "learning_rate": 0.0006,
+ "loss": 3.944416046142578,
+ "step": 2811
+ },
+ {
+ "epoch": 39.05589519650655,
+ "grad_norm": 0.019967492669820786,
+ "learning_rate": 0.0006,
+ "loss": 3.992234468460083,
+ "step": 2812
+ },
+ {
+ "epoch": 39.069868995633186,
+ "grad_norm": 0.018937544897198677,
+ "learning_rate": 0.0006,
+ "loss": 4.028573036193848,
+ "step": 2813
+ },
+ {
+ "epoch": 39.083842794759825,
+ "grad_norm": 0.016852952539920807,
+ "learning_rate": 0.0006,
+ "loss": 4.062442779541016,
+ "step": 2814
+ },
+ {
+ "epoch": 39.09781659388646,
+ "grad_norm": 0.016643647104501724,
+ "learning_rate": 0.0006,
+ "loss": 4.08871603012085,
+ "step": 2815
+ },
+ {
+ "epoch": 39.1117903930131,
+ "grad_norm": 0.016872866079211235,
+ "learning_rate": 0.0006,
+ "loss": 3.9514245986938477,
+ "step": 2816
+ },
+ {
+ "epoch": 39.12576419213974,
+ "grad_norm": 0.018045669421553612,
+ "learning_rate": 0.0006,
+ "loss": 4.038842678070068,
+ "step": 2817
+ },
+ {
+ "epoch": 39.13973799126637,
+ "grad_norm": 0.017853403463959694,
+ "learning_rate": 0.0006,
+ "loss": 3.93709135055542,
+ "step": 2818
+ },
+ {
+ "epoch": 39.15371179039301,
+ "grad_norm": 0.017507443204522133,
+ "learning_rate": 0.0006,
+ "loss": 3.98665714263916,
+ "step": 2819
+ },
+ {
+ "epoch": 39.16768558951965,
+ "grad_norm": 0.01766853965818882,
+ "learning_rate": 0.0006,
+ "loss": 3.937126636505127,
+ "step": 2820
+ },
+ {
+ "epoch": 39.18165938864629,
+ "grad_norm": 0.016660021618008614,
+ "learning_rate": 0.0006,
+ "loss": 3.8397014141082764,
+ "step": 2821
+ },
+ {
+ "epoch": 39.19563318777293,
+ "grad_norm": 0.017719309777021408,
+ "learning_rate": 0.0006,
+ "loss": 4.003579616546631,
+ "step": 2822
+ },
+ {
+ "epoch": 39.209606986899566,
+ "grad_norm": 0.018234414979815483,
+ "learning_rate": 0.0006,
+ "loss": 4.053859710693359,
+ "step": 2823
+ },
+ {
+ "epoch": 39.223580786026204,
+ "grad_norm": 0.018777024000883102,
+ "learning_rate": 0.0006,
+ "loss": 4.056182861328125,
+ "step": 2824
+ },
+ {
+ "epoch": 39.237554585152836,
+ "grad_norm": 0.017443502321839333,
+ "learning_rate": 0.0006,
+ "loss": 4.025068283081055,
+ "step": 2825
+ },
+ {
+ "epoch": 39.251528384279474,
+ "grad_norm": 0.016500331461429596,
+ "learning_rate": 0.0006,
+ "loss": 4.101631164550781,
+ "step": 2826
+ },
+ {
+ "epoch": 39.26550218340611,
+ "grad_norm": 0.016101544722914696,
+ "learning_rate": 0.0006,
+ "loss": 4.027082443237305,
+ "step": 2827
+ },
+ {
+ "epoch": 39.27947598253275,
+ "grad_norm": 0.015471171587705612,
+ "learning_rate": 0.0006,
+ "loss": 4.022909641265869,
+ "step": 2828
+ },
+ {
+ "epoch": 39.29344978165939,
+ "grad_norm": 0.015779748558998108,
+ "learning_rate": 0.0006,
+ "loss": 4.18796968460083,
+ "step": 2829
+ },
+ {
+ "epoch": 39.30742358078603,
+ "grad_norm": 0.015522679314017296,
+ "learning_rate": 0.0006,
+ "loss": 4.005777359008789,
+ "step": 2830
+ },
+ {
+ "epoch": 39.32139737991266,
+ "grad_norm": 0.015226504765450954,
+ "learning_rate": 0.0006,
+ "loss": 4.108416557312012,
+ "step": 2831
+ },
+ {
+ "epoch": 39.3353711790393,
+ "grad_norm": 0.015095778740942478,
+ "learning_rate": 0.0006,
+ "loss": 4.08029842376709,
+ "step": 2832
+ },
+ {
+ "epoch": 39.34934497816594,
+ "grad_norm": 0.015601031482219696,
+ "learning_rate": 0.0006,
+ "loss": 3.989492416381836,
+ "step": 2833
+ },
+ {
+ "epoch": 39.36331877729258,
+ "grad_norm": 0.014825052581727505,
+ "learning_rate": 0.0006,
+ "loss": 3.998520612716675,
+ "step": 2834
+ },
+ {
+ "epoch": 39.377292576419215,
+ "grad_norm": 0.014821100980043411,
+ "learning_rate": 0.0006,
+ "loss": 4.001777172088623,
+ "step": 2835
+ },
+ {
+ "epoch": 39.391266375545854,
+ "grad_norm": 0.015333754941821098,
+ "learning_rate": 0.0006,
+ "loss": 4.11985969543457,
+ "step": 2836
+ },
+ {
+ "epoch": 39.40524017467249,
+ "grad_norm": 0.015940699726343155,
+ "learning_rate": 0.0006,
+ "loss": 4.025373935699463,
+ "step": 2837
+ },
+ {
+ "epoch": 39.419213973799124,
+ "grad_norm": 0.015096287243068218,
+ "learning_rate": 0.0006,
+ "loss": 4.068551063537598,
+ "step": 2838
+ },
+ {
+ "epoch": 39.43318777292576,
+ "grad_norm": 0.01533548254519701,
+ "learning_rate": 0.0006,
+ "loss": 3.9800801277160645,
+ "step": 2839
+ },
+ {
+ "epoch": 39.4471615720524,
+ "grad_norm": 0.014648732729256153,
+ "learning_rate": 0.0006,
+ "loss": 3.9670190811157227,
+ "step": 2840
+ },
+ {
+ "epoch": 39.46113537117904,
+ "grad_norm": 0.015759671106934547,
+ "learning_rate": 0.0006,
+ "loss": 4.050443649291992,
+ "step": 2841
+ },
+ {
+ "epoch": 39.47510917030568,
+ "grad_norm": 0.016054073348641396,
+ "learning_rate": 0.0006,
+ "loss": 4.22734260559082,
+ "step": 2842
+ },
+ {
+ "epoch": 39.48908296943232,
+ "grad_norm": 0.0160264540463686,
+ "learning_rate": 0.0006,
+ "loss": 4.116174221038818,
+ "step": 2843
+ },
+ {
+ "epoch": 39.50305676855895,
+ "grad_norm": 0.015426610596477985,
+ "learning_rate": 0.0006,
+ "loss": 4.000178337097168,
+ "step": 2844
+ },
+ {
+ "epoch": 39.51703056768559,
+ "grad_norm": 0.01520642638206482,
+ "learning_rate": 0.0006,
+ "loss": 4.218410968780518,
+ "step": 2845
+ },
+ {
+ "epoch": 39.531004366812226,
+ "grad_norm": 0.015305949375033379,
+ "learning_rate": 0.0006,
+ "loss": 4.025891304016113,
+ "step": 2846
+ },
+ {
+ "epoch": 39.544978165938865,
+ "grad_norm": 0.015461564064025879,
+ "learning_rate": 0.0006,
+ "loss": 4.100950717926025,
+ "step": 2847
+ },
+ {
+ "epoch": 39.5589519650655,
+ "grad_norm": 0.01712625101208687,
+ "learning_rate": 0.0006,
+ "loss": 4.0620856285095215,
+ "step": 2848
+ },
+ {
+ "epoch": 39.57292576419214,
+ "grad_norm": 0.017133301123976707,
+ "learning_rate": 0.0006,
+ "loss": 4.097138404846191,
+ "step": 2849
+ },
+ {
+ "epoch": 39.58689956331878,
+ "grad_norm": 0.01683625392615795,
+ "learning_rate": 0.0006,
+ "loss": 4.084364891052246,
+ "step": 2850
+ },
+ {
+ "epoch": 39.60087336244541,
+ "grad_norm": 0.0165710486471653,
+ "learning_rate": 0.0006,
+ "loss": 4.042361259460449,
+ "step": 2851
+ },
+ {
+ "epoch": 39.61484716157205,
+ "grad_norm": 0.01660393364727497,
+ "learning_rate": 0.0006,
+ "loss": 4.1395487785339355,
+ "step": 2852
+ },
+ {
+ "epoch": 39.62882096069869,
+ "grad_norm": 0.017025845125317574,
+ "learning_rate": 0.0006,
+ "loss": 4.1207075119018555,
+ "step": 2853
+ },
+ {
+ "epoch": 39.64279475982533,
+ "grad_norm": 0.01660497486591339,
+ "learning_rate": 0.0006,
+ "loss": 4.103846073150635,
+ "step": 2854
+ },
+ {
+ "epoch": 39.65676855895197,
+ "grad_norm": 0.01607946865260601,
+ "learning_rate": 0.0006,
+ "loss": 4.189016342163086,
+ "step": 2855
+ },
+ {
+ "epoch": 39.670742358078606,
+ "grad_norm": 0.01614505983889103,
+ "learning_rate": 0.0006,
+ "loss": 4.079860687255859,
+ "step": 2856
+ },
+ {
+ "epoch": 39.68471615720524,
+ "grad_norm": 0.017085086554288864,
+ "learning_rate": 0.0006,
+ "loss": 3.9774293899536133,
+ "step": 2857
+ },
+ {
+ "epoch": 39.698689956331876,
+ "grad_norm": 0.01816520094871521,
+ "learning_rate": 0.0006,
+ "loss": 4.113666534423828,
+ "step": 2858
+ },
+ {
+ "epoch": 39.712663755458514,
+ "grad_norm": 0.016597526147961617,
+ "learning_rate": 0.0006,
+ "loss": 4.068242073059082,
+ "step": 2859
+ },
+ {
+ "epoch": 39.72663755458515,
+ "grad_norm": 0.01672235131263733,
+ "learning_rate": 0.0006,
+ "loss": 4.024444580078125,
+ "step": 2860
+ },
+ {
+ "epoch": 39.74061135371179,
+ "grad_norm": 0.020093556493520737,
+ "learning_rate": 0.0006,
+ "loss": 4.064124584197998,
+ "step": 2861
+ },
+ {
+ "epoch": 39.75458515283843,
+ "grad_norm": 0.023141304031014442,
+ "learning_rate": 0.0006,
+ "loss": 4.130439758300781,
+ "step": 2862
+ },
+ {
+ "epoch": 39.76855895196506,
+ "grad_norm": 0.02289372682571411,
+ "learning_rate": 0.0006,
+ "loss": 4.14794921875,
+ "step": 2863
+ },
+ {
+ "epoch": 39.7825327510917,
+ "grad_norm": 0.01919020712375641,
+ "learning_rate": 0.0006,
+ "loss": 4.123961925506592,
+ "step": 2864
+ },
+ {
+ "epoch": 39.79650655021834,
+ "grad_norm": 0.016932329162955284,
+ "learning_rate": 0.0006,
+ "loss": 4.0944132804870605,
+ "step": 2865
+ },
+ {
+ "epoch": 39.81048034934498,
+ "grad_norm": 0.016644535586237907,
+ "learning_rate": 0.0006,
+ "loss": 4.124269962310791,
+ "step": 2866
+ },
+ {
+ "epoch": 39.82445414847162,
+ "grad_norm": 0.016302216798067093,
+ "learning_rate": 0.0006,
+ "loss": 4.0267205238342285,
+ "step": 2867
+ },
+ {
+ "epoch": 39.838427947598255,
+ "grad_norm": 0.01673300378024578,
+ "learning_rate": 0.0006,
+ "loss": 4.10650634765625,
+ "step": 2868
+ },
+ {
+ "epoch": 39.852401746724894,
+ "grad_norm": 0.016581356525421143,
+ "learning_rate": 0.0006,
+ "loss": 4.1547956466674805,
+ "step": 2869
+ },
+ {
+ "epoch": 39.866375545851525,
+ "grad_norm": 0.01691797934472561,
+ "learning_rate": 0.0006,
+ "loss": 4.056936740875244,
+ "step": 2870
+ },
+ {
+ "epoch": 39.880349344978164,
+ "grad_norm": 0.01596726104617119,
+ "learning_rate": 0.0006,
+ "loss": 3.9891414642333984,
+ "step": 2871
+ },
+ {
+ "epoch": 39.8943231441048,
+ "grad_norm": 0.015615091659128666,
+ "learning_rate": 0.0006,
+ "loss": 4.137948036193848,
+ "step": 2872
+ },
+ {
+ "epoch": 39.90829694323144,
+ "grad_norm": 0.015703413635492325,
+ "learning_rate": 0.0006,
+ "loss": 3.9844930171966553,
+ "step": 2873
+ },
+ {
+ "epoch": 39.92227074235808,
+ "grad_norm": 0.015453094616532326,
+ "learning_rate": 0.0006,
+ "loss": 4.098443508148193,
+ "step": 2874
+ },
+ {
+ "epoch": 39.93624454148472,
+ "grad_norm": 0.01449158601462841,
+ "learning_rate": 0.0006,
+ "loss": 4.090852737426758,
+ "step": 2875
+ },
+ {
+ "epoch": 39.95021834061135,
+ "grad_norm": 0.01537515688687563,
+ "learning_rate": 0.0006,
+ "loss": 4.239612579345703,
+ "step": 2876
+ },
+ {
+ "epoch": 39.96419213973799,
+ "grad_norm": 0.015989921987056732,
+ "learning_rate": 0.0006,
+ "loss": 4.142073154449463,
+ "step": 2877
+ },
+ {
+ "epoch": 39.97816593886463,
+ "grad_norm": 0.015985611826181412,
+ "learning_rate": 0.0006,
+ "loss": 4.100076675415039,
+ "step": 2878
+ },
+ {
+ "epoch": 39.992139737991266,
+ "grad_norm": 0.015260833315551281,
+ "learning_rate": 0.0006,
+ "loss": 4.068576335906982,
+ "step": 2879
+ },
+ {
+ "epoch": 40.0,
+ "grad_norm": 0.018047941848635674,
+ "learning_rate": 0.0006,
+ "loss": 4.005878925323486,
+ "step": 2880
+ },
+ {
+ "epoch": 40.0,
+ "eval_loss": 4.622848987579346,
+ "eval_runtime": 56.9618,
+ "eval_samples_per_second": 42.871,
+ "eval_steps_per_second": 1.352,
+ "step": 2880
+ },
+ {
+ "epoch": 40.01397379912664,
+ "grad_norm": 0.015976838767528534,
+ "learning_rate": 0.0006,
+ "loss": 4.035689353942871,
+ "step": 2881
+ },
+ {
+ "epoch": 40.02794759825328,
+ "grad_norm": 0.015176220797002316,
+ "learning_rate": 0.0006,
+ "loss": 3.932197332382202,
+ "step": 2882
+ },
+ {
+ "epoch": 40.041921397379916,
+ "grad_norm": 0.016676228493452072,
+ "learning_rate": 0.0006,
+ "loss": 4.014171123504639,
+ "step": 2883
+ },
+ {
+ "epoch": 40.05589519650655,
+ "grad_norm": 0.01698281429708004,
+ "learning_rate": 0.0006,
+ "loss": 4.117893218994141,
+ "step": 2884
+ },
+ {
+ "epoch": 40.069868995633186,
+ "grad_norm": 0.017872486263513565,
+ "learning_rate": 0.0006,
+ "loss": 4.132119178771973,
+ "step": 2885
+ },
+ {
+ "epoch": 40.083842794759825,
+ "grad_norm": 0.017615554854273796,
+ "learning_rate": 0.0006,
+ "loss": 4.072000980377197,
+ "step": 2886
+ },
+ {
+ "epoch": 40.09781659388646,
+ "grad_norm": 0.017469394952058792,
+ "learning_rate": 0.0006,
+ "loss": 4.017726421356201,
+ "step": 2887
+ },
+ {
+ "epoch": 40.1117903930131,
+ "grad_norm": 0.01954762637615204,
+ "learning_rate": 0.0006,
+ "loss": 4.014443874359131,
+ "step": 2888
+ },
+ {
+ "epoch": 40.12576419213974,
+ "grad_norm": 0.018935853615403175,
+ "learning_rate": 0.0006,
+ "loss": 4.0042948722839355,
+ "step": 2889
+ },
+ {
+ "epoch": 40.13973799126637,
+ "grad_norm": 0.019382692873477936,
+ "learning_rate": 0.0006,
+ "loss": 4.06160831451416,
+ "step": 2890
+ },
+ {
+ "epoch": 40.15371179039301,
+ "grad_norm": 0.019677115604281425,
+ "learning_rate": 0.0006,
+ "loss": 4.13466215133667,
+ "step": 2891
+ },
+ {
+ "epoch": 40.16768558951965,
+ "grad_norm": 0.020365122705698013,
+ "learning_rate": 0.0006,
+ "loss": 4.095000267028809,
+ "step": 2892
+ },
+ {
+ "epoch": 40.18165938864629,
+ "grad_norm": 0.021842138841748238,
+ "learning_rate": 0.0006,
+ "loss": 3.9877853393554688,
+ "step": 2893
+ },
+ {
+ "epoch": 40.19563318777293,
+ "grad_norm": 0.02354881353676319,
+ "learning_rate": 0.0006,
+ "loss": 3.9430112838745117,
+ "step": 2894
+ },
+ {
+ "epoch": 40.209606986899566,
+ "grad_norm": 0.026602735742926598,
+ "learning_rate": 0.0006,
+ "loss": 4.012076377868652,
+ "step": 2895
+ },
+ {
+ "epoch": 40.223580786026204,
+ "grad_norm": 0.023976484313607216,
+ "learning_rate": 0.0006,
+ "loss": 3.9768595695495605,
+ "step": 2896
+ },
+ {
+ "epoch": 40.237554585152836,
+ "grad_norm": 0.02336616814136505,
+ "learning_rate": 0.0006,
+ "loss": 4.1096038818359375,
+ "step": 2897
+ },
+ {
+ "epoch": 40.251528384279474,
+ "grad_norm": 0.02426023595035076,
+ "learning_rate": 0.0006,
+ "loss": 4.157189846038818,
+ "step": 2898
+ },
+ {
+ "epoch": 40.26550218340611,
+ "grad_norm": 0.021572696045041084,
+ "learning_rate": 0.0006,
+ "loss": 3.979175090789795,
+ "step": 2899
+ },
+ {
+ "epoch": 40.27947598253275,
+ "grad_norm": 0.08300432562828064,
+ "learning_rate": 0.0006,
+ "loss": 4.092336654663086,
+ "step": 2900
+ },
+ {
+ "epoch": 40.29344978165939,
+ "grad_norm": 0.7323967814445496,
+ "learning_rate": 0.0006,
+ "loss": 5.337990760803223,
+ "step": 2901
+ },
+ {
+ "epoch": 40.30742358078603,
+ "grad_norm": 0.3363337814807892,
+ "learning_rate": 0.0006,
+ "loss": 6.40043830871582,
+ "step": 2902
+ },
+ {
+ "epoch": 40.32139737991266,
+ "grad_norm": 0.598686933517456,
+ "learning_rate": 0.0006,
+ "loss": 7.004765510559082,
+ "step": 2903
+ },
+ {
+ "epoch": 40.3353711790393,
+ "grad_norm": 0.5998364686965942,
+ "learning_rate": 0.0006,
+ "loss": 8.184747695922852,
+ "step": 2904
+ },
+ {
+ "epoch": 40.34934497816594,
+ "grad_norm": 1.408208966255188,
+ "learning_rate": 0.0006,
+ "loss": 8.88978385925293,
+ "step": 2905
+ },
+ {
+ "epoch": 40.36331877729258,
+ "grad_norm": 0.7829785943031311,
+ "learning_rate": 0.0006,
+ "loss": 11.13601303100586,
+ "step": 2906
+ },
+ {
+ "epoch": 40.377292576419215,
+ "grad_norm": 0.7769215106964111,
+ "learning_rate": 0.0006,
+ "loss": 9.90416145324707,
+ "step": 2907
+ },
+ {
+ "epoch": 40.391266375545854,
+ "grad_norm": 0.46508079767227173,
+ "learning_rate": 0.0006,
+ "loss": 9.647209167480469,
+ "step": 2908
+ },
+ {
+ "epoch": 40.40524017467249,
+ "grad_norm": 1.0228968858718872,
+ "learning_rate": 0.0006,
+ "loss": 11.685434341430664,
+ "step": 2909
+ },
+ {
+ "epoch": 40.419213973799124,
+ "grad_norm": 0.20553778111934662,
+ "learning_rate": 0.0006,
+ "loss": 8.621708869934082,
+ "step": 2910
+ },
+ {
+ "epoch": 40.43318777292576,
+ "grad_norm": 0.18466424942016602,
+ "learning_rate": 0.0006,
+ "loss": 8.590787887573242,
+ "step": 2911
+ },
+ {
+ "epoch": 40.4471615720524,
+ "grad_norm": 0.10458594560623169,
+ "learning_rate": 0.0006,
+ "loss": 8.324458122253418,
+ "step": 2912
+ },
+ {
+ "epoch": 40.46113537117904,
+ "grad_norm": 0.1096842959523201,
+ "learning_rate": 0.0006,
+ "loss": 8.300851821899414,
+ "step": 2913
+ },
+ {
+ "epoch": 40.47510917030568,
+ "grad_norm": 0.10501226782798767,
+ "learning_rate": 0.0006,
+ "loss": 8.21009635925293,
+ "step": 2914
+ },
+ {
+ "epoch": 40.48908296943232,
+ "grad_norm": 0.13704165816307068,
+ "learning_rate": 0.0006,
+ "loss": 8.141355514526367,
+ "step": 2915
+ },
+ {
+ "epoch": 40.50305676855895,
+ "grad_norm": 0.12129108607769012,
+ "learning_rate": 0.0006,
+ "loss": 8.074612617492676,
+ "step": 2916
+ },
+ {
+ "epoch": 40.51703056768559,
+ "grad_norm": 0.10028897225856781,
+ "learning_rate": 0.0006,
+ "loss": 8.00384521484375,
+ "step": 2917
+ },
+ {
+ "epoch": 40.531004366812226,
+ "grad_norm": 0.11764246970415115,
+ "learning_rate": 0.0006,
+ "loss": 8.078598022460938,
+ "step": 2918
+ },
+ {
+ "epoch": 40.544978165938865,
+ "grad_norm": 0.10081959515810013,
+ "learning_rate": 0.0006,
+ "loss": 8.008938789367676,
+ "step": 2919
+ },
+ {
+ "epoch": 40.5589519650655,
+ "grad_norm": 0.1003170907497406,
+ "learning_rate": 0.0006,
+ "loss": 7.962940216064453,
+ "step": 2920
+ },
+ {
+ "epoch": 40.57292576419214,
+ "grad_norm": 0.08763492107391357,
+ "learning_rate": 0.0006,
+ "loss": 7.909419536590576,
+ "step": 2921
+ },
+ {
+ "epoch": 40.58689956331878,
+ "grad_norm": 0.0636807233095169,
+ "learning_rate": 0.0006,
+ "loss": 7.878905773162842,
+ "step": 2922
+ },
+ {
+ "epoch": 40.60087336244541,
+ "grad_norm": 0.06005936488509178,
+ "learning_rate": 0.0006,
+ "loss": 7.850980758666992,
+ "step": 2923
+ },
+ {
+ "epoch": 40.61484716157205,
+ "grad_norm": 0.06415296345949173,
+ "learning_rate": 0.0006,
+ "loss": 7.830353260040283,
+ "step": 2924
+ },
+ {
+ "epoch": 40.62882096069869,
+ "grad_norm": 0.0732007846236229,
+ "learning_rate": 0.0006,
+ "loss": 7.830746650695801,
+ "step": 2925
+ },
+ {
+ "epoch": 40.64279475982533,
+ "grad_norm": 0.06628110259771347,
+ "learning_rate": 0.0006,
+ "loss": 7.82661771774292,
+ "step": 2926
+ },
+ {
+ "epoch": 40.65676855895197,
+ "grad_norm": 0.05544218048453331,
+ "learning_rate": 0.0006,
+ "loss": 7.814305782318115,
+ "step": 2927
+ },
+ {
+ "epoch": 40.670742358078606,
+ "grad_norm": 0.050327103585004807,
+ "learning_rate": 0.0006,
+ "loss": 7.764055252075195,
+ "step": 2928
+ },
+ {
+ "epoch": 40.68471615720524,
+ "grad_norm": 0.04918425530195236,
+ "learning_rate": 0.0006,
+ "loss": 7.744625091552734,
+ "step": 2929
+ },
+ {
+ "epoch": 40.698689956331876,
+ "grad_norm": 0.042862266302108765,
+ "learning_rate": 0.0006,
+ "loss": 7.764426231384277,
+ "step": 2930
+ },
+ {
+ "epoch": 40.712663755458514,
+ "grad_norm": 0.04338089004158974,
+ "learning_rate": 0.0006,
+ "loss": 7.712530136108398,
+ "step": 2931
+ },
+ {
+ "epoch": 40.72663755458515,
+ "grad_norm": 0.04127737879753113,
+ "learning_rate": 0.0006,
+ "loss": 7.742180347442627,
+ "step": 2932
+ },
+ {
+ "epoch": 40.74061135371179,
+ "grad_norm": 0.03358631581068039,
+ "learning_rate": 0.0006,
+ "loss": 7.714834213256836,
+ "step": 2933
+ },
+ {
+ "epoch": 40.75458515283843,
+ "grad_norm": 0.02845790982246399,
+ "learning_rate": 0.0006,
+ "loss": 7.7021026611328125,
+ "step": 2934
+ },
+ {
+ "epoch": 40.76855895196506,
+ "grad_norm": 0.03012419492006302,
+ "learning_rate": 0.0006,
+ "loss": 7.712377548217773,
+ "step": 2935
+ },
+ {
+ "epoch": 40.7825327510917,
+ "grad_norm": 0.032981179654598236,
+ "learning_rate": 0.0006,
+ "loss": 7.666385650634766,
+ "step": 2936
+ },
+ {
+ "epoch": 40.79650655021834,
+ "grad_norm": 0.03357074037194252,
+ "learning_rate": 0.0006,
+ "loss": 7.649097919464111,
+ "step": 2937
+ },
+ {
+ "epoch": 40.81048034934498,
+ "grad_norm": 0.030914708971977234,
+ "learning_rate": 0.0006,
+ "loss": 7.649196624755859,
+ "step": 2938
+ },
+ {
+ "epoch": 40.82445414847162,
+ "grad_norm": 0.026296131312847137,
+ "learning_rate": 0.0006,
+ "loss": 7.629964828491211,
+ "step": 2939
+ },
+ {
+ "epoch": 40.838427947598255,
+ "grad_norm": 0.026677245274186134,
+ "learning_rate": 0.0006,
+ "loss": 7.599973678588867,
+ "step": 2940
+ },
+ {
+ "epoch": 40.852401746724894,
+ "grad_norm": 0.023208219558000565,
+ "learning_rate": 0.0006,
+ "loss": 7.601624488830566,
+ "step": 2941
+ },
+ {
+ "epoch": 40.866375545851525,
+ "grad_norm": 0.020741475746035576,
+ "learning_rate": 0.0006,
+ "loss": 7.603902339935303,
+ "step": 2942
+ },
+ {
+ "epoch": 40.880349344978164,
+ "grad_norm": 0.019221751019358635,
+ "learning_rate": 0.0006,
+ "loss": 7.5535888671875,
+ "step": 2943
+ },
+ {
+ "epoch": 40.8943231441048,
+ "grad_norm": 0.019491951912641525,
+ "learning_rate": 0.0006,
+ "loss": 7.54597282409668,
+ "step": 2944
+ },
+ {
+ "epoch": 40.90829694323144,
+ "grad_norm": 0.019702572375535965,
+ "learning_rate": 0.0006,
+ "loss": 7.537567138671875,
+ "step": 2945
+ },
+ {
+ "epoch": 40.92227074235808,
+ "grad_norm": 0.01768800988793373,
+ "learning_rate": 0.0006,
+ "loss": 7.512800216674805,
+ "step": 2946
+ },
+ {
+ "epoch": 40.93624454148472,
+ "grad_norm": 0.01926138624548912,
+ "learning_rate": 0.0006,
+ "loss": 7.509932518005371,
+ "step": 2947
+ },
+ {
+ "epoch": 40.95021834061135,
+ "grad_norm": 0.019030006602406502,
+ "learning_rate": 0.0006,
+ "loss": 7.497983932495117,
+ "step": 2948
+ },
+ {
+ "epoch": 40.96419213973799,
+ "grad_norm": 0.017313841730356216,
+ "learning_rate": 0.0006,
+ "loss": 7.463518142700195,
+ "step": 2949
+ },
+ {
+ "epoch": 40.97816593886463,
+ "grad_norm": 0.01310762483626604,
+ "learning_rate": 0.0006,
+ "loss": 7.474668502807617,
+ "step": 2950
+ },
+ {
+ "epoch": 40.992139737991266,
+ "grad_norm": 0.013008282519876957,
+ "learning_rate": 0.0006,
+ "loss": 7.45982551574707,
+ "step": 2951
+ },
+ {
+ "epoch": 41.0,
+ "grad_norm": 0.01441363524645567,
+ "learning_rate": 0.0006,
+ "loss": 7.446529388427734,
+ "step": 2952
+ },
+ {
+ "epoch": 41.0,
+ "eval_loss": 7.486120223999023,
+ "eval_runtime": 70.4986,
+ "eval_samples_per_second": 34.639,
+ "eval_steps_per_second": 1.092,
+ "step": 2952
+ },
+ {
+ "epoch": 41.01397379912664,
+ "grad_norm": 0.014061855152249336,
+ "learning_rate": 0.0006,
+ "loss": 7.4692182540893555,
+ "step": 2953
+ },
+ {
+ "epoch": 41.02794759825328,
+ "grad_norm": 0.01587500050663948,
+ "learning_rate": 0.0006,
+ "loss": 7.4182024002075195,
+ "step": 2954
+ },
+ {
+ "epoch": 41.041921397379916,
+ "grad_norm": 0.014783984050154686,
+ "learning_rate": 0.0006,
+ "loss": 7.466778755187988,
+ "step": 2955
+ },
+ {
+ "epoch": 41.05589519650655,
+ "grad_norm": 0.011560854502022266,
+ "learning_rate": 0.0006,
+ "loss": 7.438002586364746,
+ "step": 2956
+ },
+ {
+ "epoch": 41.069868995633186,
+ "grad_norm": 0.00970957800745964,
+ "learning_rate": 0.0006,
+ "loss": 7.442694187164307,
+ "step": 2957
+ },
+ {
+ "epoch": 41.083842794759825,
+ "grad_norm": 0.009633459150791168,
+ "learning_rate": 0.0006,
+ "loss": 7.458884239196777,
+ "step": 2958
+ },
+ {
+ "epoch": 41.09781659388646,
+ "grad_norm": 0.010664350353181362,
+ "learning_rate": 0.0006,
+ "loss": 7.48378324508667,
+ "step": 2959
+ },
+ {
+ "epoch": 41.1117903930131,
+ "grad_norm": 0.011440591886639595,
+ "learning_rate": 0.0006,
+ "loss": 7.474484920501709,
+ "step": 2960
+ },
+ {
+ "epoch": 41.12576419213974,
+ "grad_norm": 0.01150908600538969,
+ "learning_rate": 0.0006,
+ "loss": 7.417902946472168,
+ "step": 2961
+ },
+ {
+ "epoch": 41.13973799126637,
+ "grad_norm": 0.010375697165727615,
+ "learning_rate": 0.0006,
+ "loss": 7.403279781341553,
+ "step": 2962
+ },
+ {
+ "epoch": 41.15371179039301,
+ "grad_norm": 0.009075703099370003,
+ "learning_rate": 0.0006,
+ "loss": 7.447347164154053,
+ "step": 2963
+ },
+ {
+ "epoch": 41.16768558951965,
+ "grad_norm": 0.008892177604138851,
+ "learning_rate": 0.0006,
+ "loss": 7.402681350708008,
+ "step": 2964
+ },
+ {
+ "epoch": 41.18165938864629,
+ "grad_norm": 0.009492084383964539,
+ "learning_rate": 0.0006,
+ "loss": 7.3871355056762695,
+ "step": 2965
+ },
+ {
+ "epoch": 41.19563318777293,
+ "grad_norm": 0.009139460511505604,
+ "learning_rate": 0.0006,
+ "loss": 7.419617176055908,
+ "step": 2966
+ },
+ {
+ "epoch": 41.209606986899566,
+ "grad_norm": 0.007959138602018356,
+ "learning_rate": 0.0006,
+ "loss": 7.374949932098389,
+ "step": 2967
+ },
+ {
+ "epoch": 41.223580786026204,
+ "grad_norm": 0.007571548223495483,
+ "learning_rate": 0.0006,
+ "loss": 7.4205241203308105,
+ "step": 2968
+ },
+ {
+ "epoch": 41.237554585152836,
+ "grad_norm": 0.007885400205850601,
+ "learning_rate": 0.0006,
+ "loss": 7.416543006896973,
+ "step": 2969
+ },
+ {
+ "epoch": 41.251528384279474,
+ "grad_norm": 0.008650158531963825,
+ "learning_rate": 0.0006,
+ "loss": 7.357326984405518,
+ "step": 2970
+ },
+ {
+ "epoch": 41.26550218340611,
+ "grad_norm": 0.007483262103050947,
+ "learning_rate": 0.0006,
+ "loss": 7.410891056060791,
+ "step": 2971
+ },
+ {
+ "epoch": 41.27947598253275,
+ "grad_norm": 0.00869277399033308,
+ "learning_rate": 0.0006,
+ "loss": 7.392127990722656,
+ "step": 2972
+ },
+ {
+ "epoch": 41.29344978165939,
+ "grad_norm": 0.00885078776627779,
+ "learning_rate": 0.0006,
+ "loss": 7.405449867248535,
+ "step": 2973
+ },
+ {
+ "epoch": 41.30742358078603,
+ "grad_norm": 0.00958226341754198,
+ "learning_rate": 0.0006,
+ "loss": 7.399868488311768,
+ "step": 2974
+ },
+ {
+ "epoch": 41.32139737991266,
+ "grad_norm": 0.006991599686443806,
+ "learning_rate": 0.0006,
+ "loss": 7.35369873046875,
+ "step": 2975
+ },
+ {
+ "epoch": 41.3353711790393,
+ "grad_norm": 0.007415142375975847,
+ "learning_rate": 0.0006,
+ "loss": 7.358353137969971,
+ "step": 2976
+ },
+ {
+ "epoch": 41.34934497816594,
+ "grad_norm": 0.007024371065199375,
+ "learning_rate": 0.0006,
+ "loss": 7.4070892333984375,
+ "step": 2977
+ },
+ {
+ "epoch": 41.36331877729258,
+ "grad_norm": 0.007234332151710987,
+ "learning_rate": 0.0006,
+ "loss": 7.379120349884033,
+ "step": 2978
+ },
+ {
+ "epoch": 41.377292576419215,
+ "grad_norm": 0.00798892229795456,
+ "learning_rate": 0.0006,
+ "loss": 7.365455627441406,
+ "step": 2979
+ },
+ {
+ "epoch": 41.391266375545854,
+ "grad_norm": 0.007303398102521896,
+ "learning_rate": 0.0006,
+ "loss": 7.341798782348633,
+ "step": 2980
+ },
+ {
+ "epoch": 41.40524017467249,
+ "grad_norm": 0.007119669578969479,
+ "learning_rate": 0.0006,
+ "loss": 7.363280296325684,
+ "step": 2981
+ },
+ {
+ "epoch": 41.419213973799124,
+ "grad_norm": 0.007457619532942772,
+ "learning_rate": 0.0006,
+ "loss": 7.39388370513916,
+ "step": 2982
+ },
+ {
+ "epoch": 41.43318777292576,
+ "grad_norm": 0.007905688136816025,
+ "learning_rate": 0.0006,
+ "loss": 7.380652904510498,
+ "step": 2983
+ },
+ {
+ "epoch": 41.4471615720524,
+ "grad_norm": 0.006212855689227581,
+ "learning_rate": 0.0006,
+ "loss": 7.398769378662109,
+ "step": 2984
+ },
+ {
+ "epoch": 41.46113537117904,
+ "grad_norm": 0.0070641543716192245,
+ "learning_rate": 0.0006,
+ "loss": 7.3503546714782715,
+ "step": 2985
+ },
+ {
+ "epoch": 41.47510917030568,
+ "grad_norm": 0.00787805300205946,
+ "learning_rate": 0.0006,
+ "loss": 7.347678184509277,
+ "step": 2986
+ },
+ {
+ "epoch": 41.48908296943232,
+ "grad_norm": 0.006760426796972752,
+ "learning_rate": 0.0006,
+ "loss": 7.380006790161133,
+ "step": 2987
+ },
+ {
+ "epoch": 41.50305676855895,
+ "grad_norm": 0.00772385997697711,
+ "learning_rate": 0.0006,
+ "loss": 7.363680362701416,
+ "step": 2988
+ },
+ {
+ "epoch": 41.51703056768559,
+ "grad_norm": 0.007373564876616001,
+ "learning_rate": 0.0006,
+ "loss": 7.333014488220215,
+ "step": 2989
+ },
+ {
+ "epoch": 41.531004366812226,
+ "grad_norm": 0.006717653013765812,
+ "learning_rate": 0.0006,
+ "loss": 7.360076427459717,
+ "step": 2990
+ },
+ {
+ "epoch": 41.544978165938865,
+ "grad_norm": 0.007191413082182407,
+ "learning_rate": 0.0006,
+ "loss": 7.397030830383301,
+ "step": 2991
+ },
+ {
+ "epoch": 41.5589519650655,
+ "grad_norm": 0.007831960916519165,
+ "learning_rate": 0.0006,
+ "loss": 7.369959831237793,
+ "step": 2992
+ },
+ {
+ "epoch": 41.57292576419214,
+ "grad_norm": 0.006947481073439121,
+ "learning_rate": 0.0006,
+ "loss": 7.333254337310791,
+ "step": 2993
+ },
+ {
+ "epoch": 41.58689956331878,
+ "grad_norm": 0.006243606563657522,
+ "learning_rate": 0.0006,
+ "loss": 7.327378273010254,
+ "step": 2994
+ },
+ {
+ "epoch": 41.60087336244541,
+ "grad_norm": 0.006298339460045099,
+ "learning_rate": 0.0006,
+ "loss": 7.323482036590576,
+ "step": 2995
+ },
+ {
+ "epoch": 41.61484716157205,
+ "grad_norm": 0.007587412837892771,
+ "learning_rate": 0.0006,
+ "loss": 7.313789367675781,
+ "step": 2996
+ },
+ {
+ "epoch": 41.62882096069869,
+ "grad_norm": 0.007140746805816889,
+ "learning_rate": 0.0006,
+ "loss": 7.356067657470703,
+ "step": 2997
+ },
+ {
+ "epoch": 41.64279475982533,
+ "grad_norm": 0.008315625600516796,
+ "learning_rate": 0.0006,
+ "loss": 7.3339643478393555,
+ "step": 2998
+ },
+ {
+ "epoch": 41.65676855895197,
+ "grad_norm": 0.006547102238982916,
+ "learning_rate": 0.0006,
+ "loss": 7.342191696166992,
+ "step": 2999
+ },
+ {
+ "epoch": 41.670742358078606,
+ "grad_norm": 0.006068122107535601,
+ "learning_rate": 0.0006,
+ "loss": 7.321015357971191,
+ "step": 3000
+ },
+ {
+ "epoch": 41.68471615720524,
+ "grad_norm": 0.007316991221159697,
+ "learning_rate": 0.0006,
+ "loss": 7.346490859985352,
+ "step": 3001
+ },
+ {
+ "epoch": 41.698689956331876,
+ "grad_norm": 0.0067388564348220825,
+ "learning_rate": 0.0006,
+ "loss": 7.340723991394043,
+ "step": 3002
+ },
+ {
+ "epoch": 41.712663755458514,
+ "grad_norm": 0.006839963141828775,
+ "learning_rate": 0.0006,
+ "loss": 7.329365253448486,
+ "step": 3003
+ },
+ {
+ "epoch": 41.72663755458515,
+ "grad_norm": 0.00805041566491127,
+ "learning_rate": 0.0006,
+ "loss": 7.298934459686279,
+ "step": 3004
+ },
+ {
+ "epoch": 41.74061135371179,
+ "grad_norm": 0.007197503466159105,
+ "learning_rate": 0.0006,
+ "loss": 7.331496238708496,
+ "step": 3005
+ },
+ {
+ "epoch": 41.75458515283843,
+ "grad_norm": 0.007220805156975985,
+ "learning_rate": 0.0006,
+ "loss": 7.317695140838623,
+ "step": 3006
+ },
+ {
+ "epoch": 41.76855895196506,
+ "grad_norm": 0.00874276738613844,
+ "learning_rate": 0.0006,
+ "loss": 7.272614479064941,
+ "step": 3007
+ },
+ {
+ "epoch": 41.7825327510917,
+ "grad_norm": 0.008386565372347832,
+ "learning_rate": 0.0006,
+ "loss": 7.295244216918945,
+ "step": 3008
+ },
+ {
+ "epoch": 41.79650655021834,
+ "grad_norm": 0.0077730282209813595,
+ "learning_rate": 0.0006,
+ "loss": 7.3395490646362305,
+ "step": 3009
+ },
+ {
+ "epoch": 41.81048034934498,
+ "grad_norm": 0.008441480807960033,
+ "learning_rate": 0.0006,
+ "loss": 7.319921493530273,
+ "step": 3010
+ },
+ {
+ "epoch": 41.82445414847162,
+ "grad_norm": 0.007398269604891539,
+ "learning_rate": 0.0006,
+ "loss": 7.3029465675354,
+ "step": 3011
+ },
+ {
+ "epoch": 41.838427947598255,
+ "grad_norm": 0.006878677289932966,
+ "learning_rate": 0.0006,
+ "loss": 7.318467140197754,
+ "step": 3012
+ },
+ {
+ "epoch": 41.852401746724894,
+ "grad_norm": 0.007616598624736071,
+ "learning_rate": 0.0006,
+ "loss": 7.3112077713012695,
+ "step": 3013
+ },
+ {
+ "epoch": 41.866375545851525,
+ "grad_norm": 0.011629699729382992,
+ "learning_rate": 0.0006,
+ "loss": 7.32523250579834,
+ "step": 3014
+ },
+ {
+ "epoch": 41.880349344978164,
+ "grad_norm": 0.011011249385774136,
+ "learning_rate": 0.0006,
+ "loss": 7.287247657775879,
+ "step": 3015
+ },
+ {
+ "epoch": 41.8943231441048,
+ "grad_norm": 0.010208295658230782,
+ "learning_rate": 0.0006,
+ "loss": 7.292117118835449,
+ "step": 3016
+ },
+ {
+ "epoch": 41.90829694323144,
+ "grad_norm": 0.008651095442473888,
+ "learning_rate": 0.0006,
+ "loss": 7.302201271057129,
+ "step": 3017
+ },
+ {
+ "epoch": 41.92227074235808,
+ "grad_norm": 0.006692920345813036,
+ "learning_rate": 0.0006,
+ "loss": 7.319568634033203,
+ "step": 3018
+ },
+ {
+ "epoch": 41.93624454148472,
+ "grad_norm": 0.010033348575234413,
+ "learning_rate": 0.0006,
+ "loss": 7.2738566398620605,
+ "step": 3019
+ },
+ {
+ "epoch": 41.95021834061135,
+ "grad_norm": 0.011790832504630089,
+ "learning_rate": 0.0006,
+ "loss": 7.298540115356445,
+ "step": 3020
+ },
+ {
+ "epoch": 41.96419213973799,
+ "grad_norm": 0.018622662872076035,
+ "learning_rate": 0.0006,
+ "loss": 7.267182350158691,
+ "step": 3021
+ },
+ {
+ "epoch": 41.97816593886463,
+ "grad_norm": 0.02520257607102394,
+ "learning_rate": 0.0006,
+ "loss": 7.261566162109375,
+ "step": 3022
+ },
+ {
+ "epoch": 41.992139737991266,
+ "grad_norm": 0.036332644522190094,
+ "learning_rate": 0.0006,
+ "loss": 7.270719051361084,
+ "step": 3023
+ },
+ {
+ "epoch": 42.0,
+ "grad_norm": 0.04717305675148964,
+ "learning_rate": 0.0006,
+ "loss": 7.293332099914551,
+ "step": 3024
+ },
+ {
+ "epoch": 42.0,
+ "eval_loss": 7.354280471801758,
+ "eval_runtime": 56.4353,
+ "eval_samples_per_second": 43.271,
+ "eval_steps_per_second": 1.364,
+ "step": 3024
+ },
+ {
+ "epoch": 42.01397379912664,
+ "grad_norm": 0.06661626696586609,
+ "learning_rate": 0.0006,
+ "loss": 7.338189601898193,
+ "step": 3025
+ },
+ {
+ "epoch": 42.02794759825328,
+ "grad_norm": 0.05452054366469383,
+ "learning_rate": 0.0006,
+ "loss": 7.316764831542969,
+ "step": 3026
+ },
+ {
+ "epoch": 42.041921397379916,
+ "grad_norm": 0.025638775900006294,
+ "learning_rate": 0.0006,
+ "loss": 7.34204626083374,
+ "step": 3027
+ },
+ {
+ "epoch": 42.05589519650655,
+ "grad_norm": 0.022398799657821655,
+ "learning_rate": 0.0006,
+ "loss": 7.262149333953857,
+ "step": 3028
+ },
+ {
+ "epoch": 42.069868995633186,
+ "grad_norm": 0.03713499382138252,
+ "learning_rate": 0.0006,
+ "loss": 7.27695369720459,
+ "step": 3029
+ },
+ {
+ "epoch": 42.083842794759825,
+ "grad_norm": 0.014949247241020203,
+ "learning_rate": 0.0006,
+ "loss": 7.296508312225342,
+ "step": 3030
+ },
+ {
+ "epoch": 42.09781659388646,
+ "grad_norm": 0.02703995071351528,
+ "learning_rate": 0.0006,
+ "loss": 7.258018493652344,
+ "step": 3031
+ },
+ {
+ "epoch": 42.1117903930131,
+ "grad_norm": 0.020352816209197044,
+ "learning_rate": 0.0006,
+ "loss": 7.241450309753418,
+ "step": 3032
+ },
+ {
+ "epoch": 42.12576419213974,
+ "grad_norm": 0.014578479342162609,
+ "learning_rate": 0.0006,
+ "loss": 7.257323265075684,
+ "step": 3033
+ },
+ {
+ "epoch": 42.13973799126637,
+ "grad_norm": 0.026053044945001602,
+ "learning_rate": 0.0006,
+ "loss": 7.254680633544922,
+ "step": 3034
+ },
+ {
+ "epoch": 42.15371179039301,
+ "grad_norm": 0.015120275318622589,
+ "learning_rate": 0.0006,
+ "loss": 7.261074066162109,
+ "step": 3035
+ },
+ {
+ "epoch": 42.16768558951965,
+ "grad_norm": 0.016822177916765213,
+ "learning_rate": 0.0006,
+ "loss": 7.279064178466797,
+ "step": 3036
+ },
+ {
+ "epoch": 42.18165938864629,
+ "grad_norm": 0.016367588192224503,
+ "learning_rate": 0.0006,
+ "loss": 7.251429557800293,
+ "step": 3037
+ },
+ {
+ "epoch": 42.19563318777293,
+ "grad_norm": 0.014350418001413345,
+ "learning_rate": 0.0006,
+ "loss": 7.259264945983887,
+ "step": 3038
+ },
+ {
+ "epoch": 42.209606986899566,
+ "grad_norm": 0.017078228294849396,
+ "learning_rate": 0.0006,
+ "loss": 7.227105140686035,
+ "step": 3039
+ },
+ {
+ "epoch": 42.223580786026204,
+ "grad_norm": 0.011759317480027676,
+ "learning_rate": 0.0006,
+ "loss": 7.275888442993164,
+ "step": 3040
+ },
+ {
+ "epoch": 42.237554585152836,
+ "grad_norm": 0.013019340112805367,
+ "learning_rate": 0.0006,
+ "loss": 7.2517008781433105,
+ "step": 3041
+ },
+ {
+ "epoch": 42.251528384279474,
+ "grad_norm": 0.01178441196680069,
+ "learning_rate": 0.0006,
+ "loss": 7.23576545715332,
+ "step": 3042
+ },
+ {
+ "epoch": 42.26550218340611,
+ "grad_norm": 0.014518891461193562,
+ "learning_rate": 0.0006,
+ "loss": 7.253895282745361,
+ "step": 3043
+ },
+ {
+ "epoch": 42.27947598253275,
+ "grad_norm": 0.018159950152039528,
+ "learning_rate": 0.0006,
+ "loss": 7.278058052062988,
+ "step": 3044
+ },
+ {
+ "epoch": 42.29344978165939,
+ "grad_norm": 0.007587796077132225,
+ "learning_rate": 0.0006,
+ "loss": 7.278578758239746,
+ "step": 3045
+ },
+ {
+ "epoch": 42.30742358078603,
+ "grad_norm": 0.015994979068636894,
+ "learning_rate": 0.0006,
+ "loss": 7.239841461181641,
+ "step": 3046
+ },
+ {
+ "epoch": 42.32139737991266,
+ "grad_norm": 0.017830168828368187,
+ "learning_rate": 0.0006,
+ "loss": 7.2443976402282715,
+ "step": 3047
+ },
+ {
+ "epoch": 42.3353711790393,
+ "grad_norm": 0.008738759905099869,
+ "learning_rate": 0.0006,
+ "loss": 7.232264518737793,
+ "step": 3048
+ },
+ {
+ "epoch": 42.34934497816594,
+ "grad_norm": 0.0175449438393116,
+ "learning_rate": 0.0006,
+ "loss": 7.228000640869141,
+ "step": 3049
+ },
+ {
+ "epoch": 42.36331877729258,
+ "grad_norm": 0.019611820578575134,
+ "learning_rate": 0.0006,
+ "loss": 7.234999179840088,
+ "step": 3050
+ },
+ {
+ "epoch": 42.377292576419215,
+ "grad_norm": 0.015302863903343678,
+ "learning_rate": 0.0006,
+ "loss": 7.242547988891602,
+ "step": 3051
+ },
+ {
+ "epoch": 42.391266375545854,
+ "grad_norm": 0.025919731706380844,
+ "learning_rate": 0.0006,
+ "loss": 7.216652870178223,
+ "step": 3052
+ },
+ {
+ "epoch": 42.40524017467249,
+ "grad_norm": 0.034587468951940536,
+ "learning_rate": 0.0006,
+ "loss": 7.245336532592773,
+ "step": 3053
+ },
+ {
+ "epoch": 42.419213973799124,
+ "grad_norm": 0.019198840484023094,
+ "learning_rate": 0.0006,
+ "loss": 7.256758689880371,
+ "step": 3054
+ },
+ {
+ "epoch": 42.43318777292576,
+ "grad_norm": 0.012316389009356499,
+ "learning_rate": 0.0006,
+ "loss": 7.209102630615234,
+ "step": 3055
+ },
+ {
+ "epoch": 42.4471615720524,
+ "grad_norm": 0.023512162268161774,
+ "learning_rate": 0.0006,
+ "loss": 7.250742435455322,
+ "step": 3056
+ },
+ {
+ "epoch": 42.46113537117904,
+ "grad_norm": 0.025457851588726044,
+ "learning_rate": 0.0006,
+ "loss": 7.263757705688477,
+ "step": 3057
+ },
+ {
+ "epoch": 42.47510917030568,
+ "grad_norm": 0.02192617766559124,
+ "learning_rate": 0.0006,
+ "loss": 7.246977806091309,
+ "step": 3058
+ },
+ {
+ "epoch": 42.48908296943232,
+ "grad_norm": 0.02541281282901764,
+ "learning_rate": 0.0006,
+ "loss": 7.232361316680908,
+ "step": 3059
+ },
+ {
+ "epoch": 42.50305676855895,
+ "grad_norm": 0.036353763192892075,
+ "learning_rate": 0.0006,
+ "loss": 7.242556571960449,
+ "step": 3060
+ },
+ {
+ "epoch": 42.51703056768559,
+ "grad_norm": 0.04555872827768326,
+ "learning_rate": 0.0006,
+ "loss": 7.228940963745117,
+ "step": 3061
+ },
+ {
+ "epoch": 42.531004366812226,
+ "grad_norm": 0.04563378915190697,
+ "learning_rate": 0.0006,
+ "loss": 7.235276222229004,
+ "step": 3062
+ },
+ {
+ "epoch": 42.544978165938865,
+ "grad_norm": 0.04275760427117348,
+ "learning_rate": 0.0006,
+ "loss": 7.238895416259766,
+ "step": 3063
+ },
+ {
+ "epoch": 42.5589519650655,
+ "grad_norm": 0.04535885155200958,
+ "learning_rate": 0.0006,
+ "loss": 7.205336570739746,
+ "step": 3064
+ },
+ {
+ "epoch": 42.57292576419214,
+ "grad_norm": 0.015534617006778717,
+ "learning_rate": 0.0006,
+ "loss": 7.238393783569336,
+ "step": 3065
+ },
+ {
+ "epoch": 42.58689956331878,
+ "grad_norm": 0.03286299481987953,
+ "learning_rate": 0.0006,
+ "loss": 7.216453552246094,
+ "step": 3066
+ },
+ {
+ "epoch": 42.60087336244541,
+ "grad_norm": 0.025453168898820877,
+ "learning_rate": 0.0006,
+ "loss": 7.203750133514404,
+ "step": 3067
+ },
+ {
+ "epoch": 42.61484716157205,
+ "grad_norm": 0.021501649171113968,
+ "learning_rate": 0.0006,
+ "loss": 7.241621971130371,
+ "step": 3068
+ },
+ {
+ "epoch": 42.62882096069869,
+ "grad_norm": 0.02743634767830372,
+ "learning_rate": 0.0006,
+ "loss": 7.22076940536499,
+ "step": 3069
+ },
+ {
+ "epoch": 42.64279475982533,
+ "grad_norm": 0.020890483632683754,
+ "learning_rate": 0.0006,
+ "loss": 7.218412399291992,
+ "step": 3070
+ },
+ {
+ "epoch": 42.65676855895197,
+ "grad_norm": 0.022499850019812584,
+ "learning_rate": 0.0006,
+ "loss": 7.217696189880371,
+ "step": 3071
+ },
+ {
+ "epoch": 42.670742358078606,
+ "grad_norm": 0.01653696969151497,
+ "learning_rate": 0.0006,
+ "loss": 7.156417369842529,
+ "step": 3072
+ },
+ {
+ "epoch": 42.68471615720524,
+ "grad_norm": 0.01834225095808506,
+ "learning_rate": 0.0006,
+ "loss": 7.211247444152832,
+ "step": 3073
+ },
+ {
+ "epoch": 42.698689956331876,
+ "grad_norm": 0.020137229934334755,
+ "learning_rate": 0.0006,
+ "loss": 7.232499599456787,
+ "step": 3074
+ },
+ {
+ "epoch": 42.712663755458514,
+ "grad_norm": 0.012671459466218948,
+ "learning_rate": 0.0006,
+ "loss": 7.172441482543945,
+ "step": 3075
+ },
+ {
+ "epoch": 42.72663755458515,
+ "grad_norm": 0.016803130507469177,
+ "learning_rate": 0.0006,
+ "loss": 7.152103424072266,
+ "step": 3076
+ },
+ {
+ "epoch": 42.74061135371179,
+ "grad_norm": 0.01623712107539177,
+ "learning_rate": 0.0006,
+ "loss": 7.187032699584961,
+ "step": 3077
+ },
+ {
+ "epoch": 42.75458515283843,
+ "grad_norm": 0.015938356518745422,
+ "learning_rate": 0.0006,
+ "loss": 7.171140670776367,
+ "step": 3078
+ },
+ {
+ "epoch": 42.76855895196506,
+ "grad_norm": 0.02580736204981804,
+ "learning_rate": 0.0006,
+ "loss": 7.204472541809082,
+ "step": 3079
+ },
+ {
+ "epoch": 42.7825327510917,
+ "grad_norm": 0.023372117429971695,
+ "learning_rate": 0.0006,
+ "loss": 7.165554523468018,
+ "step": 3080
+ },
+ {
+ "epoch": 42.79650655021834,
+ "grad_norm": 0.04440586268901825,
+ "learning_rate": 0.0006,
+ "loss": 7.169431209564209,
+ "step": 3081
+ },
+ {
+ "epoch": 42.81048034934498,
+ "grad_norm": 0.06797406077384949,
+ "learning_rate": 0.0006,
+ "loss": 7.168193340301514,
+ "step": 3082
+ },
+ {
+ "epoch": 42.82445414847162,
+ "grad_norm": 0.0914321169257164,
+ "learning_rate": 0.0006,
+ "loss": 7.204671859741211,
+ "step": 3083
+ },
+ {
+ "epoch": 42.838427947598255,
+ "grad_norm": 0.0506940595805645,
+ "learning_rate": 0.0006,
+ "loss": 7.15949821472168,
+ "step": 3084
+ },
+ {
+ "epoch": 42.852401746724894,
+ "grad_norm": 0.04819890484213829,
+ "learning_rate": 0.0006,
+ "loss": 7.192837238311768,
+ "step": 3085
+ },
+ {
+ "epoch": 42.866375545851525,
+ "grad_norm": 0.040044233202934265,
+ "learning_rate": 0.0006,
+ "loss": 7.1600165367126465,
+ "step": 3086
+ },
+ {
+ "epoch": 42.880349344978164,
+ "grad_norm": 0.02573299966752529,
+ "learning_rate": 0.0006,
+ "loss": 7.153427600860596,
+ "step": 3087
+ },
+ {
+ "epoch": 42.8943231441048,
+ "grad_norm": 0.046942535787820816,
+ "learning_rate": 0.0006,
+ "loss": 7.16767692565918,
+ "step": 3088
+ },
+ {
+ "epoch": 42.90829694323144,
+ "grad_norm": 0.02683161571621895,
+ "learning_rate": 0.0006,
+ "loss": 7.17144250869751,
+ "step": 3089
+ },
+ {
+ "epoch": 42.92227074235808,
+ "grad_norm": 0.05363674834370613,
+ "learning_rate": 0.0006,
+ "loss": 7.120128631591797,
+ "step": 3090
+ },
+ {
+ "epoch": 42.93624454148472,
+ "grad_norm": 0.023756247013807297,
+ "learning_rate": 0.0006,
+ "loss": 7.169703006744385,
+ "step": 3091
+ },
+ {
+ "epoch": 42.95021834061135,
+ "grad_norm": 0.04216783866286278,
+ "learning_rate": 0.0006,
+ "loss": 7.1206865310668945,
+ "step": 3092
+ },
+ {
+ "epoch": 42.96419213973799,
+ "grad_norm": 0.02219565585255623,
+ "learning_rate": 0.0006,
+ "loss": 7.157215118408203,
+ "step": 3093
+ },
+ {
+ "epoch": 42.97816593886463,
+ "grad_norm": 0.03527127206325531,
+ "learning_rate": 0.0006,
+ "loss": 7.166672706604004,
+ "step": 3094
+ },
+ {
+ "epoch": 42.992139737991266,
+ "grad_norm": 0.017748817801475525,
+ "learning_rate": 0.0006,
+ "loss": 7.154747009277344,
+ "step": 3095
+ },
+ {
+ "epoch": 43.0,
+ "grad_norm": 0.033165838569402695,
+ "learning_rate": 0.0006,
+ "loss": 7.150866508483887,
+ "step": 3096
+ },
+ {
+ "epoch": 43.0,
+ "eval_loss": 7.144403457641602,
+ "eval_runtime": 56.8853,
+ "eval_samples_per_second": 42.928,
+ "eval_steps_per_second": 1.354,
+ "step": 3096
+ },
+ {
+ "epoch": 43.01397379912664,
+ "grad_norm": 0.015199829824268818,
+ "learning_rate": 0.0006,
+ "loss": 7.122735023498535,
+ "step": 3097
+ },
+ {
+ "epoch": 43.02794759825328,
+ "grad_norm": 0.03220256045460701,
+ "learning_rate": 0.0006,
+ "loss": 7.143553733825684,
+ "step": 3098
+ },
+ {
+ "epoch": 43.041921397379916,
+ "grad_norm": 0.014687920920550823,
+ "learning_rate": 0.0006,
+ "loss": 7.080960750579834,
+ "step": 3099
+ },
+ {
+ "epoch": 43.05589519650655,
+ "grad_norm": 0.023522932082414627,
+ "learning_rate": 0.0006,
+ "loss": 7.137356758117676,
+ "step": 3100
+ },
+ {
+ "epoch": 43.069868995633186,
+ "grad_norm": 0.014808216132223606,
+ "learning_rate": 0.0006,
+ "loss": 7.09143590927124,
+ "step": 3101
+ },
+ {
+ "epoch": 43.083842794759825,
+ "grad_norm": 0.029339246451854706,
+ "learning_rate": 0.0006,
+ "loss": 7.122198104858398,
+ "step": 3102
+ },
+ {
+ "epoch": 43.09781659388646,
+ "grad_norm": 0.022953078150749207,
+ "learning_rate": 0.0006,
+ "loss": 7.092505931854248,
+ "step": 3103
+ },
+ {
+ "epoch": 43.1117903930131,
+ "grad_norm": 0.04590052738785744,
+ "learning_rate": 0.0006,
+ "loss": 7.113100051879883,
+ "step": 3104
+ },
+ {
+ "epoch": 43.12576419213974,
+ "grad_norm": 0.0524442233145237,
+ "learning_rate": 0.0006,
+ "loss": 7.147022247314453,
+ "step": 3105
+ },
+ {
+ "epoch": 43.13973799126637,
+ "grad_norm": 0.0315794013440609,
+ "learning_rate": 0.0006,
+ "loss": 7.094045162200928,
+ "step": 3106
+ },
+ {
+ "epoch": 43.15371179039301,
+ "grad_norm": 0.05112722888588905,
+ "learning_rate": 0.0006,
+ "loss": 7.054388523101807,
+ "step": 3107
+ },
+ {
+ "epoch": 43.16768558951965,
+ "grad_norm": 0.03146776929497719,
+ "learning_rate": 0.0006,
+ "loss": 7.116674423217773,
+ "step": 3108
+ },
+ {
+ "epoch": 43.18165938864629,
+ "grad_norm": 0.03391246497631073,
+ "learning_rate": 0.0006,
+ "loss": 7.150345802307129,
+ "step": 3109
+ },
+ {
+ "epoch": 43.19563318777293,
+ "grad_norm": 0.026957903057336807,
+ "learning_rate": 0.0006,
+ "loss": 7.058531284332275,
+ "step": 3110
+ },
+ {
+ "epoch": 43.209606986899566,
+ "grad_norm": 0.041532568633556366,
+ "learning_rate": 0.0006,
+ "loss": 7.085498809814453,
+ "step": 3111
+ },
+ {
+ "epoch": 43.223580786026204,
+ "grad_norm": 0.020001588389277458,
+ "learning_rate": 0.0006,
+ "loss": 7.131487846374512,
+ "step": 3112
+ },
+ {
+ "epoch": 43.237554585152836,
+ "grad_norm": 0.02268712781369686,
+ "learning_rate": 0.0006,
+ "loss": 7.0790863037109375,
+ "step": 3113
+ },
+ {
+ "epoch": 43.251528384279474,
+ "grad_norm": 0.023363925516605377,
+ "learning_rate": 0.0006,
+ "loss": 7.067816734313965,
+ "step": 3114
+ },
+ {
+ "epoch": 43.26550218340611,
+ "grad_norm": 0.019367508590221405,
+ "learning_rate": 0.0006,
+ "loss": 7.042041778564453,
+ "step": 3115
+ },
+ {
+ "epoch": 43.27947598253275,
+ "grad_norm": 0.02745090238749981,
+ "learning_rate": 0.0006,
+ "loss": 7.064733028411865,
+ "step": 3116
+ },
+ {
+ "epoch": 43.29344978165939,
+ "grad_norm": 0.02543802745640278,
+ "learning_rate": 0.0006,
+ "loss": 7.058948040008545,
+ "step": 3117
+ },
+ {
+ "epoch": 43.30742358078603,
+ "grad_norm": 0.018874092027544975,
+ "learning_rate": 0.0006,
+ "loss": 7.009946346282959,
+ "step": 3118
+ },
+ {
+ "epoch": 43.32139737991266,
+ "grad_norm": 0.024662822484970093,
+ "learning_rate": 0.0006,
+ "loss": 7.015913963317871,
+ "step": 3119
+ },
+ {
+ "epoch": 43.3353711790393,
+ "grad_norm": 0.02441416308283806,
+ "learning_rate": 0.0006,
+ "loss": 7.061622142791748,
+ "step": 3120
+ },
+ {
+ "epoch": 43.34934497816594,
+ "grad_norm": 0.031887322664260864,
+ "learning_rate": 0.0006,
+ "loss": 7.046883583068848,
+ "step": 3121
+ },
+ {
+ "epoch": 43.36331877729258,
+ "grad_norm": 0.021930739283561707,
+ "learning_rate": 0.0006,
+ "loss": 7.0309624671936035,
+ "step": 3122
+ },
+ {
+ "epoch": 43.377292576419215,
+ "grad_norm": 0.02722783200442791,
+ "learning_rate": 0.0006,
+ "loss": 7.072924613952637,
+ "step": 3123
+ },
+ {
+ "epoch": 43.391266375545854,
+ "grad_norm": 0.01483179721981287,
+ "learning_rate": 0.0006,
+ "loss": 7.028624534606934,
+ "step": 3124
+ },
+ {
+ "epoch": 43.40524017467249,
+ "grad_norm": 0.024793945252895355,
+ "learning_rate": 0.0006,
+ "loss": 7.001623153686523,
+ "step": 3125
+ },
+ {
+ "epoch": 43.419213973799124,
+ "grad_norm": 0.020148713141679764,
+ "learning_rate": 0.0006,
+ "loss": 6.9938812255859375,
+ "step": 3126
+ },
+ {
+ "epoch": 43.43318777292576,
+ "grad_norm": 0.02413850650191307,
+ "learning_rate": 0.0006,
+ "loss": 7.001243591308594,
+ "step": 3127
+ },
+ {
+ "epoch": 43.4471615720524,
+ "grad_norm": 0.028549203649163246,
+ "learning_rate": 0.0006,
+ "loss": 6.987671375274658,
+ "step": 3128
+ },
+ {
+ "epoch": 43.46113537117904,
+ "grad_norm": 0.025988785549998283,
+ "learning_rate": 0.0006,
+ "loss": 6.9551167488098145,
+ "step": 3129
+ },
+ {
+ "epoch": 43.47510917030568,
+ "grad_norm": 0.027140729129314423,
+ "learning_rate": 0.0006,
+ "loss": 6.9547810554504395,
+ "step": 3130
+ },
+ {
+ "epoch": 43.48908296943232,
+ "grad_norm": 0.030130082741379738,
+ "learning_rate": 0.0006,
+ "loss": 6.993186950683594,
+ "step": 3131
+ },
+ {
+ "epoch": 43.50305676855895,
+ "grad_norm": 0.025996994227170944,
+ "learning_rate": 0.0006,
+ "loss": 6.984624862670898,
+ "step": 3132
+ },
+ {
+ "epoch": 43.51703056768559,
+ "grad_norm": 0.02423008903861046,
+ "learning_rate": 0.0006,
+ "loss": 6.98813009262085,
+ "step": 3133
+ },
+ {
+ "epoch": 43.531004366812226,
+ "grad_norm": 0.029665837064385414,
+ "learning_rate": 0.0006,
+ "loss": 6.986610412597656,
+ "step": 3134
+ },
+ {
+ "epoch": 43.544978165938865,
+ "grad_norm": 0.028768595308065414,
+ "learning_rate": 0.0006,
+ "loss": 6.898797035217285,
+ "step": 3135
+ },
+ {
+ "epoch": 43.5589519650655,
+ "grad_norm": 0.03111208975315094,
+ "learning_rate": 0.0006,
+ "loss": 6.923741817474365,
+ "step": 3136
+ },
+ {
+ "epoch": 43.57292576419214,
+ "grad_norm": 0.0368187390267849,
+ "learning_rate": 0.0006,
+ "loss": 6.994343280792236,
+ "step": 3137
+ },
+ {
+ "epoch": 43.58689956331878,
+ "grad_norm": 0.04076763987541199,
+ "learning_rate": 0.0006,
+ "loss": 6.909458160400391,
+ "step": 3138
+ },
+ {
+ "epoch": 43.60087336244541,
+ "grad_norm": 0.03281061723828316,
+ "learning_rate": 0.0006,
+ "loss": 6.924851417541504,
+ "step": 3139
+ },
+ {
+ "epoch": 43.61484716157205,
+ "grad_norm": 0.025486988946795464,
+ "learning_rate": 0.0006,
+ "loss": 6.962926864624023,
+ "step": 3140
+ },
+ {
+ "epoch": 43.62882096069869,
+ "grad_norm": 0.04583485797047615,
+ "learning_rate": 0.0006,
+ "loss": 6.876555442810059,
+ "step": 3141
+ },
+ {
+ "epoch": 43.64279475982533,
+ "grad_norm": 0.05460785701870918,
+ "learning_rate": 0.0006,
+ "loss": 6.952932357788086,
+ "step": 3142
+ },
+ {
+ "epoch": 43.65676855895197,
+ "grad_norm": 0.040917906910181046,
+ "learning_rate": 0.0006,
+ "loss": 6.9624924659729,
+ "step": 3143
+ },
+ {
+ "epoch": 43.670742358078606,
+ "grad_norm": 0.04009668901562691,
+ "learning_rate": 0.0006,
+ "loss": 6.94033145904541,
+ "step": 3144
+ },
+ {
+ "epoch": 43.68471615720524,
+ "grad_norm": 0.04500625655055046,
+ "learning_rate": 0.0006,
+ "loss": 6.961839199066162,
+ "step": 3145
+ },
+ {
+ "epoch": 43.698689956331876,
+ "grad_norm": 0.03182258829474449,
+ "learning_rate": 0.0006,
+ "loss": 6.930580139160156,
+ "step": 3146
+ },
+ {
+ "epoch": 43.712663755458514,
+ "grad_norm": 0.053621504455804825,
+ "learning_rate": 0.0006,
+ "loss": 6.950235366821289,
+ "step": 3147
+ },
+ {
+ "epoch": 43.72663755458515,
+ "grad_norm": 0.031919289380311966,
+ "learning_rate": 0.0006,
+ "loss": 6.895111560821533,
+ "step": 3148
+ },
+ {
+ "epoch": 43.74061135371179,
+ "grad_norm": 0.042713891714811325,
+ "learning_rate": 0.0006,
+ "loss": 6.940885543823242,
+ "step": 3149
+ },
+ {
+ "epoch": 43.75458515283843,
+ "grad_norm": 0.02385653182864189,
+ "learning_rate": 0.0006,
+ "loss": 6.948801517486572,
+ "step": 3150
+ },
+ {
+ "epoch": 43.76855895196506,
+ "grad_norm": 0.04100663587450981,
+ "learning_rate": 0.0006,
+ "loss": 6.970411777496338,
+ "step": 3151
+ },
+ {
+ "epoch": 43.7825327510917,
+ "grad_norm": 0.02729583904147148,
+ "learning_rate": 0.0006,
+ "loss": 6.899379253387451,
+ "step": 3152
+ },
+ {
+ "epoch": 43.79650655021834,
+ "grad_norm": 0.03128620237112045,
+ "learning_rate": 0.0006,
+ "loss": 6.931319236755371,
+ "step": 3153
+ },
+ {
+ "epoch": 43.81048034934498,
+ "grad_norm": 0.025488659739494324,
+ "learning_rate": 0.0006,
+ "loss": 6.8486409187316895,
+ "step": 3154
+ },
+ {
+ "epoch": 43.82445414847162,
+ "grad_norm": 0.03679376840591431,
+ "learning_rate": 0.0006,
+ "loss": 6.840811252593994,
+ "step": 3155
+ },
+ {
+ "epoch": 43.838427947598255,
+ "grad_norm": 0.020578034222126007,
+ "learning_rate": 0.0006,
+ "loss": 6.9088616371154785,
+ "step": 3156
+ },
+ {
+ "epoch": 43.852401746724894,
+ "grad_norm": 0.02935841865837574,
+ "learning_rate": 0.0006,
+ "loss": 6.806676864624023,
+ "step": 3157
+ },
+ {
+ "epoch": 43.866375545851525,
+ "grad_norm": 0.023319244384765625,
+ "learning_rate": 0.0006,
+ "loss": 6.878457546234131,
+ "step": 3158
+ },
+ {
+ "epoch": 43.880349344978164,
+ "grad_norm": 0.02779589593410492,
+ "learning_rate": 0.0006,
+ "loss": 6.822447776794434,
+ "step": 3159
+ },
+ {
+ "epoch": 43.8943231441048,
+ "grad_norm": 0.01879825070500374,
+ "learning_rate": 0.0006,
+ "loss": 6.854121208190918,
+ "step": 3160
+ },
+ {
+ "epoch": 43.90829694323144,
+ "grad_norm": 0.023286139592528343,
+ "learning_rate": 0.0006,
+ "loss": 6.836827278137207,
+ "step": 3161
+ },
+ {
+ "epoch": 43.92227074235808,
+ "grad_norm": 0.030358297750353813,
+ "learning_rate": 0.0006,
+ "loss": 6.860037803649902,
+ "step": 3162
+ },
+ {
+ "epoch": 43.93624454148472,
+ "grad_norm": 0.029621105641126633,
+ "learning_rate": 0.0006,
+ "loss": 6.884025573730469,
+ "step": 3163
+ },
+ {
+ "epoch": 43.95021834061135,
+ "grad_norm": 0.029829824343323708,
+ "learning_rate": 0.0006,
+ "loss": 6.8702392578125,
+ "step": 3164
+ },
+ {
+ "epoch": 43.96419213973799,
+ "grad_norm": 0.036892473697662354,
+ "learning_rate": 0.0006,
+ "loss": 6.806652069091797,
+ "step": 3165
+ },
+ {
+ "epoch": 43.97816593886463,
+ "grad_norm": 0.03206709027290344,
+ "learning_rate": 0.0006,
+ "loss": 6.828495979309082,
+ "step": 3166
+ },
+ {
+ "epoch": 43.992139737991266,
+ "grad_norm": 0.02744000405073166,
+ "learning_rate": 0.0006,
+ "loss": 6.832002639770508,
+ "step": 3167
+ },
+ {
+ "epoch": 44.0,
+ "grad_norm": 0.02205660194158554,
+ "learning_rate": 0.0006,
+ "loss": 6.898828506469727,
+ "step": 3168
+ },
+ {
+ "epoch": 44.0,
+ "eval_loss": 6.8150553703308105,
+ "eval_runtime": 56.784,
+ "eval_samples_per_second": 43.005,
+ "eval_steps_per_second": 1.356,
+ "step": 3168
+ },
+ {
+ "epoch": 44.01397379912664,
+ "grad_norm": 0.030421260744333267,
+ "learning_rate": 0.0006,
+ "loss": 6.775057792663574,
+ "step": 3169
+ },
+ {
+ "epoch": 44.02794759825328,
+ "grad_norm": 0.0465611033141613,
+ "learning_rate": 0.0006,
+ "loss": 6.849390506744385,
+ "step": 3170
+ },
+ {
+ "epoch": 44.041921397379916,
+ "grad_norm": 0.05977972596883774,
+ "learning_rate": 0.0006,
+ "loss": 6.725576400756836,
+ "step": 3171
+ },
+ {
+ "epoch": 44.05589519650655,
+ "grad_norm": 0.05432697385549545,
+ "learning_rate": 0.0006,
+ "loss": 6.768024444580078,
+ "step": 3172
+ },
+ {
+ "epoch": 44.069868995633186,
+ "grad_norm": 0.04850967228412628,
+ "learning_rate": 0.0006,
+ "loss": 6.799457550048828,
+ "step": 3173
+ },
+ {
+ "epoch": 44.083842794759825,
+ "grad_norm": 0.06201786920428276,
+ "learning_rate": 0.0006,
+ "loss": 6.897234916687012,
+ "step": 3174
+ },
+ {
+ "epoch": 44.09781659388646,
+ "grad_norm": 0.04695591703057289,
+ "learning_rate": 0.0006,
+ "loss": 6.826470375061035,
+ "step": 3175
+ },
+ {
+ "epoch": 44.1117903930131,
+ "grad_norm": 0.03850164636969566,
+ "learning_rate": 0.0006,
+ "loss": 6.835901260375977,
+ "step": 3176
+ },
+ {
+ "epoch": 44.12576419213974,
+ "grad_norm": 0.04250155761837959,
+ "learning_rate": 0.0006,
+ "loss": 6.782310485839844,
+ "step": 3177
+ },
+ {
+ "epoch": 44.13973799126637,
+ "grad_norm": 0.041465479880571365,
+ "learning_rate": 0.0006,
+ "loss": 6.785732746124268,
+ "step": 3178
+ },
+ {
+ "epoch": 44.15371179039301,
+ "grad_norm": 0.02527807280421257,
+ "learning_rate": 0.0006,
+ "loss": 6.78061580657959,
+ "step": 3179
+ },
+ {
+ "epoch": 44.16768558951965,
+ "grad_norm": 0.03660374879837036,
+ "learning_rate": 0.0006,
+ "loss": 6.783992767333984,
+ "step": 3180
+ },
+ {
+ "epoch": 44.18165938864629,
+ "grad_norm": 0.027567964047193527,
+ "learning_rate": 0.0006,
+ "loss": 6.765413284301758,
+ "step": 3181
+ },
+ {
+ "epoch": 44.19563318777293,
+ "grad_norm": 0.024301424622535706,
+ "learning_rate": 0.0006,
+ "loss": 6.760368347167969,
+ "step": 3182
+ },
+ {
+ "epoch": 44.209606986899566,
+ "grad_norm": 0.0266821701079607,
+ "learning_rate": 0.0006,
+ "loss": 6.71417236328125,
+ "step": 3183
+ },
+ {
+ "epoch": 44.223580786026204,
+ "grad_norm": 0.028125090524554253,
+ "learning_rate": 0.0006,
+ "loss": 6.646126747131348,
+ "step": 3184
+ },
+ {
+ "epoch": 44.237554585152836,
+ "grad_norm": 0.0185737032443285,
+ "learning_rate": 0.0006,
+ "loss": 6.712944030761719,
+ "step": 3185
+ },
+ {
+ "epoch": 44.251528384279474,
+ "grad_norm": 0.025836674496531487,
+ "learning_rate": 0.0006,
+ "loss": 6.707030296325684,
+ "step": 3186
+ },
+ {
+ "epoch": 44.26550218340611,
+ "grad_norm": 0.029744870960712433,
+ "learning_rate": 0.0006,
+ "loss": 6.6271071434021,
+ "step": 3187
+ },
+ {
+ "epoch": 44.27947598253275,
+ "grad_norm": 0.03021426685154438,
+ "learning_rate": 0.0006,
+ "loss": 6.677826404571533,
+ "step": 3188
+ },
+ {
+ "epoch": 44.29344978165939,
+ "grad_norm": 0.032727014273405075,
+ "learning_rate": 0.0006,
+ "loss": 6.657764434814453,
+ "step": 3189
+ },
+ {
+ "epoch": 44.30742358078603,
+ "grad_norm": 0.03696778789162636,
+ "learning_rate": 0.0006,
+ "loss": 6.632883548736572,
+ "step": 3190
+ },
+ {
+ "epoch": 44.32139737991266,
+ "grad_norm": 0.034630827605724335,
+ "learning_rate": 0.0006,
+ "loss": 6.759583473205566,
+ "step": 3191
+ },
+ {
+ "epoch": 44.3353711790393,
+ "grad_norm": 0.030517758801579475,
+ "learning_rate": 0.0006,
+ "loss": 6.702268600463867,
+ "step": 3192
+ },
+ {
+ "epoch": 44.34934497816594,
+ "grad_norm": 0.04105190932750702,
+ "learning_rate": 0.0006,
+ "loss": 6.705792427062988,
+ "step": 3193
+ },
+ {
+ "epoch": 44.36331877729258,
+ "grad_norm": 0.06659174710512161,
+ "learning_rate": 0.0006,
+ "loss": 6.73862361907959,
+ "step": 3194
+ },
+ {
+ "epoch": 44.377292576419215,
+ "grad_norm": 0.08013619482517242,
+ "learning_rate": 0.0006,
+ "loss": 6.710282802581787,
+ "step": 3195
+ },
+ {
+ "epoch": 44.391266375545854,
+ "grad_norm": 0.049850963056087494,
+ "learning_rate": 0.0006,
+ "loss": 6.684974670410156,
+ "step": 3196
+ },
+ {
+ "epoch": 44.40524017467249,
+ "grad_norm": 0.061247408390045166,
+ "learning_rate": 0.0006,
+ "loss": 6.716034412384033,
+ "step": 3197
+ },
+ {
+ "epoch": 44.419213973799124,
+ "grad_norm": 0.05868857726454735,
+ "learning_rate": 0.0006,
+ "loss": 6.589107513427734,
+ "step": 3198
+ },
+ {
+ "epoch": 44.43318777292576,
+ "grad_norm": 0.048878561705350876,
+ "learning_rate": 0.0006,
+ "loss": 6.694855690002441,
+ "step": 3199
+ },
+ {
+ "epoch": 44.4471615720524,
+ "grad_norm": 0.028748324140906334,
+ "learning_rate": 0.0006,
+ "loss": 6.629969596862793,
+ "step": 3200
+ },
+ {
+ "epoch": 44.46113537117904,
+ "grad_norm": 0.04060306400060654,
+ "learning_rate": 0.0006,
+ "loss": 6.721431255340576,
+ "step": 3201
+ },
+ {
+ "epoch": 44.47510917030568,
+ "grad_norm": 0.03833824768662453,
+ "learning_rate": 0.0006,
+ "loss": 6.625029563903809,
+ "step": 3202
+ },
+ {
+ "epoch": 44.48908296943232,
+ "grad_norm": 0.03791997581720352,
+ "learning_rate": 0.0006,
+ "loss": 6.669187545776367,
+ "step": 3203
+ },
+ {
+ "epoch": 44.50305676855895,
+ "grad_norm": 0.03259684890508652,
+ "learning_rate": 0.0006,
+ "loss": 6.665643692016602,
+ "step": 3204
+ },
+ {
+ "epoch": 44.51703056768559,
+ "grad_norm": 0.040658626705408096,
+ "learning_rate": 0.0006,
+ "loss": 6.639796257019043,
+ "step": 3205
+ },
+ {
+ "epoch": 44.531004366812226,
+ "grad_norm": 0.03691193461418152,
+ "learning_rate": 0.0006,
+ "loss": 6.692244529724121,
+ "step": 3206
+ },
+ {
+ "epoch": 44.544978165938865,
+ "grad_norm": 0.02814812771975994,
+ "learning_rate": 0.0006,
+ "loss": 6.61978006362915,
+ "step": 3207
+ },
+ {
+ "epoch": 44.5589519650655,
+ "grad_norm": 0.022155074402689934,
+ "learning_rate": 0.0006,
+ "loss": 6.625905990600586,
+ "step": 3208
+ },
+ {
+ "epoch": 44.57292576419214,
+ "grad_norm": 0.030794885009527206,
+ "learning_rate": 0.0006,
+ "loss": 6.678500652313232,
+ "step": 3209
+ },
+ {
+ "epoch": 44.58689956331878,
+ "grad_norm": 0.03201768547296524,
+ "learning_rate": 0.0006,
+ "loss": 6.6015729904174805,
+ "step": 3210
+ },
+ {
+ "epoch": 44.60087336244541,
+ "grad_norm": 0.025138361379504204,
+ "learning_rate": 0.0006,
+ "loss": 6.568234443664551,
+ "step": 3211
+ },
+ {
+ "epoch": 44.61484716157205,
+ "grad_norm": 0.02969353087246418,
+ "learning_rate": 0.0006,
+ "loss": 6.588366508483887,
+ "step": 3212
+ },
+ {
+ "epoch": 44.62882096069869,
+ "grad_norm": 0.03136962279677391,
+ "learning_rate": 0.0006,
+ "loss": 6.479961395263672,
+ "step": 3213
+ },
+ {
+ "epoch": 44.64279475982533,
+ "grad_norm": 0.030178170651197433,
+ "learning_rate": 0.0006,
+ "loss": 6.622910976409912,
+ "step": 3214
+ },
+ {
+ "epoch": 44.65676855895197,
+ "grad_norm": 0.02682245895266533,
+ "learning_rate": 0.0006,
+ "loss": 6.597707748413086,
+ "step": 3215
+ },
+ {
+ "epoch": 44.670742358078606,
+ "grad_norm": 0.025154557079076767,
+ "learning_rate": 0.0006,
+ "loss": 6.630541801452637,
+ "step": 3216
+ },
+ {
+ "epoch": 44.68471615720524,
+ "grad_norm": 0.03564463183283806,
+ "learning_rate": 0.0006,
+ "loss": 6.542823791503906,
+ "step": 3217
+ },
+ {
+ "epoch": 44.698689956331876,
+ "grad_norm": 0.049477651715278625,
+ "learning_rate": 0.0006,
+ "loss": 6.646030426025391,
+ "step": 3218
+ },
+ {
+ "epoch": 44.712663755458514,
+ "grad_norm": 0.06536535918712616,
+ "learning_rate": 0.0006,
+ "loss": 6.595463275909424,
+ "step": 3219
+ },
+ {
+ "epoch": 44.72663755458515,
+ "grad_norm": 0.07710316777229309,
+ "learning_rate": 0.0006,
+ "loss": 6.55524206161499,
+ "step": 3220
+ },
+ {
+ "epoch": 44.74061135371179,
+ "grad_norm": 0.09571163356304169,
+ "learning_rate": 0.0006,
+ "loss": 6.659764289855957,
+ "step": 3221
+ },
+ {
+ "epoch": 44.75458515283843,
+ "grad_norm": 0.055076830089092255,
+ "learning_rate": 0.0006,
+ "loss": 6.658882141113281,
+ "step": 3222
+ },
+ {
+ "epoch": 44.76855895196506,
+ "grad_norm": 0.07671099156141281,
+ "learning_rate": 0.0006,
+ "loss": 6.605809211730957,
+ "step": 3223
+ },
+ {
+ "epoch": 44.7825327510917,
+ "grad_norm": 0.05295584350824356,
+ "learning_rate": 0.0006,
+ "loss": 6.641455173492432,
+ "step": 3224
+ },
+ {
+ "epoch": 44.79650655021834,
+ "grad_norm": 0.09128042310476303,
+ "learning_rate": 0.0006,
+ "loss": 6.691066741943359,
+ "step": 3225
+ },
+ {
+ "epoch": 44.81048034934498,
+ "grad_norm": 0.05362958461046219,
+ "learning_rate": 0.0006,
+ "loss": 6.573406219482422,
+ "step": 3226
+ },
+ {
+ "epoch": 44.82445414847162,
+ "grad_norm": 0.046910062432289124,
+ "learning_rate": 0.0006,
+ "loss": 6.550876617431641,
+ "step": 3227
+ },
+ {
+ "epoch": 44.838427947598255,
+ "grad_norm": 0.03494197130203247,
+ "learning_rate": 0.0006,
+ "loss": 6.540179252624512,
+ "step": 3228
+ },
+ {
+ "epoch": 44.852401746724894,
+ "grad_norm": 0.04716387391090393,
+ "learning_rate": 0.0006,
+ "loss": 6.490459442138672,
+ "step": 3229
+ },
+ {
+ "epoch": 44.866375545851525,
+ "grad_norm": 0.04190506786108017,
+ "learning_rate": 0.0006,
+ "loss": 6.509925365447998,
+ "step": 3230
+ },
+ {
+ "epoch": 44.880349344978164,
+ "grad_norm": 0.04749064892530441,
+ "learning_rate": 0.0006,
+ "loss": 6.478779315948486,
+ "step": 3231
+ },
+ {
+ "epoch": 44.8943231441048,
+ "grad_norm": 0.028002185747027397,
+ "learning_rate": 0.0006,
+ "loss": 6.528904914855957,
+ "step": 3232
+ },
+ {
+ "epoch": 44.90829694323144,
+ "grad_norm": 0.036473166197538376,
+ "learning_rate": 0.0006,
+ "loss": 6.549042701721191,
+ "step": 3233
+ },
+ {
+ "epoch": 44.92227074235808,
+ "grad_norm": 0.02543315477669239,
+ "learning_rate": 0.0006,
+ "loss": 6.498875141143799,
+ "step": 3234
+ },
+ {
+ "epoch": 44.93624454148472,
+ "grad_norm": 0.029906807467341423,
+ "learning_rate": 0.0006,
+ "loss": 6.453019142150879,
+ "step": 3235
+ },
+ {
+ "epoch": 44.95021834061135,
+ "grad_norm": 0.027231857180595398,
+ "learning_rate": 0.0006,
+ "loss": 6.428829193115234,
+ "step": 3236
+ },
+ {
+ "epoch": 44.96419213973799,
+ "grad_norm": 0.02506154775619507,
+ "learning_rate": 0.0006,
+ "loss": 6.420858383178711,
+ "step": 3237
+ },
+ {
+ "epoch": 44.97816593886463,
+ "grad_norm": 0.0269472673535347,
+ "learning_rate": 0.0006,
+ "loss": 6.484989643096924,
+ "step": 3238
+ },
+ {
+ "epoch": 44.992139737991266,
+ "grad_norm": 0.023410487920045853,
+ "learning_rate": 0.0006,
+ "loss": 6.454830169677734,
+ "step": 3239
+ },
+ {
+ "epoch": 45.0,
+ "grad_norm": 0.034175023436546326,
+ "learning_rate": 0.0006,
+ "loss": 6.430975437164307,
+ "step": 3240
+ },
+ {
+ "epoch": 45.0,
+ "eval_loss": 6.458896160125732,
+ "eval_runtime": 56.6934,
+ "eval_samples_per_second": 43.074,
+ "eval_steps_per_second": 1.358,
+ "step": 3240
+ },
+ {
+ "epoch": 45.01397379912664,
+ "grad_norm": 0.045031629502773285,
+ "learning_rate": 0.0006,
+ "loss": 6.4243011474609375,
+ "step": 3241
+ },
+ {
+ "epoch": 45.02794759825328,
+ "grad_norm": 0.04968734085559845,
+ "learning_rate": 0.0006,
+ "loss": 6.380091667175293,
+ "step": 3242
+ },
+ {
+ "epoch": 45.041921397379916,
+ "grad_norm": 0.047145187854766846,
+ "learning_rate": 0.0006,
+ "loss": 6.477178573608398,
+ "step": 3243
+ },
+ {
+ "epoch": 45.05589519650655,
+ "grad_norm": 0.04304948449134827,
+ "learning_rate": 0.0006,
+ "loss": 6.399726867675781,
+ "step": 3244
+ },
+ {
+ "epoch": 45.069868995633186,
+ "grad_norm": 0.04466010630130768,
+ "learning_rate": 0.0006,
+ "loss": 6.3415679931640625,
+ "step": 3245
+ },
+ {
+ "epoch": 45.083842794759825,
+ "grad_norm": 0.04305306077003479,
+ "learning_rate": 0.0006,
+ "loss": 6.3413310050964355,
+ "step": 3246
+ },
+ {
+ "epoch": 45.09781659388646,
+ "grad_norm": 0.043064575642347336,
+ "learning_rate": 0.0006,
+ "loss": 6.4838714599609375,
+ "step": 3247
+ },
+ {
+ "epoch": 45.1117903930131,
+ "grad_norm": 0.049503277987241745,
+ "learning_rate": 0.0006,
+ "loss": 6.38473653793335,
+ "step": 3248
+ },
+ {
+ "epoch": 45.12576419213974,
+ "grad_norm": 0.04613238573074341,
+ "learning_rate": 0.0006,
+ "loss": 6.370769500732422,
+ "step": 3249
+ },
+ {
+ "epoch": 45.13973799126637,
+ "grad_norm": 0.052119478583335876,
+ "learning_rate": 0.0006,
+ "loss": 6.362321853637695,
+ "step": 3250
+ },
+ {
+ "epoch": 45.15371179039301,
+ "grad_norm": 0.045110076665878296,
+ "learning_rate": 0.0006,
+ "loss": 6.369844913482666,
+ "step": 3251
+ },
+ {
+ "epoch": 45.16768558951965,
+ "grad_norm": 0.03798583149909973,
+ "learning_rate": 0.0006,
+ "loss": 6.324633598327637,
+ "step": 3252
+ },
+ {
+ "epoch": 45.18165938864629,
+ "grad_norm": 0.03744659945368767,
+ "learning_rate": 0.0006,
+ "loss": 6.349421501159668,
+ "step": 3253
+ },
+ {
+ "epoch": 45.19563318777293,
+ "grad_norm": 0.03767292574048042,
+ "learning_rate": 0.0006,
+ "loss": 6.334238052368164,
+ "step": 3254
+ },
+ {
+ "epoch": 45.209606986899566,
+ "grad_norm": 0.03577509894967079,
+ "learning_rate": 0.0006,
+ "loss": 6.3453826904296875,
+ "step": 3255
+ },
+ {
+ "epoch": 45.223580786026204,
+ "grad_norm": 0.03403240442276001,
+ "learning_rate": 0.0006,
+ "loss": 6.2836174964904785,
+ "step": 3256
+ },
+ {
+ "epoch": 45.237554585152836,
+ "grad_norm": 0.03517331928014755,
+ "learning_rate": 0.0006,
+ "loss": 6.281718730926514,
+ "step": 3257
+ },
+ {
+ "epoch": 45.251528384279474,
+ "grad_norm": 0.04409139230847359,
+ "learning_rate": 0.0006,
+ "loss": 6.320815086364746,
+ "step": 3258
+ },
+ {
+ "epoch": 45.26550218340611,
+ "grad_norm": 0.054750364273786545,
+ "learning_rate": 0.0006,
+ "loss": 6.343618392944336,
+ "step": 3259
+ },
+ {
+ "epoch": 45.27947598253275,
+ "grad_norm": 0.07612067461013794,
+ "learning_rate": 0.0006,
+ "loss": 6.310614585876465,
+ "step": 3260
+ },
+ {
+ "epoch": 45.29344978165939,
+ "grad_norm": 0.08438985049724579,
+ "learning_rate": 0.0006,
+ "loss": 6.2369184494018555,
+ "step": 3261
+ },
+ {
+ "epoch": 45.30742358078603,
+ "grad_norm": 0.0782201811671257,
+ "learning_rate": 0.0006,
+ "loss": 6.254452705383301,
+ "step": 3262
+ },
+ {
+ "epoch": 45.32139737991266,
+ "grad_norm": 0.09928355365991592,
+ "learning_rate": 0.0006,
+ "loss": 6.3112077713012695,
+ "step": 3263
+ },
+ {
+ "epoch": 45.3353711790393,
+ "grad_norm": 0.07233133912086487,
+ "learning_rate": 0.0006,
+ "loss": 6.408087253570557,
+ "step": 3264
+ },
+ {
+ "epoch": 45.34934497816594,
+ "grad_norm": 0.07478684186935425,
+ "learning_rate": 0.0006,
+ "loss": 6.3770036697387695,
+ "step": 3265
+ },
+ {
+ "epoch": 45.36331877729258,
+ "grad_norm": 0.06036846712231636,
+ "learning_rate": 0.0006,
+ "loss": 6.33865213394165,
+ "step": 3266
+ },
+ {
+ "epoch": 45.377292576419215,
+ "grad_norm": 0.05783458054065704,
+ "learning_rate": 0.0006,
+ "loss": 6.363203048706055,
+ "step": 3267
+ },
+ {
+ "epoch": 45.391266375545854,
+ "grad_norm": 0.05443909019231796,
+ "learning_rate": 0.0006,
+ "loss": 6.287069797515869,
+ "step": 3268
+ },
+ {
+ "epoch": 45.40524017467249,
+ "grad_norm": 0.0557909794151783,
+ "learning_rate": 0.0006,
+ "loss": 6.297784805297852,
+ "step": 3269
+ },
+ {
+ "epoch": 45.419213973799124,
+ "grad_norm": 0.03483372926712036,
+ "learning_rate": 0.0006,
+ "loss": 6.280734062194824,
+ "step": 3270
+ },
+ {
+ "epoch": 45.43318777292576,
+ "grad_norm": 0.0425543375313282,
+ "learning_rate": 0.0006,
+ "loss": 6.33565616607666,
+ "step": 3271
+ },
+ {
+ "epoch": 45.4471615720524,
+ "grad_norm": 0.03230159357190132,
+ "learning_rate": 0.0006,
+ "loss": 6.197698593139648,
+ "step": 3272
+ },
+ {
+ "epoch": 45.46113537117904,
+ "grad_norm": 0.036897290498018265,
+ "learning_rate": 0.0006,
+ "loss": 6.241239547729492,
+ "step": 3273
+ },
+ {
+ "epoch": 45.47510917030568,
+ "grad_norm": 0.031103838235139847,
+ "learning_rate": 0.0006,
+ "loss": 6.159435749053955,
+ "step": 3274
+ },
+ {
+ "epoch": 45.48908296943232,
+ "grad_norm": 0.027116449549794197,
+ "learning_rate": 0.0006,
+ "loss": 6.212225914001465,
+ "step": 3275
+ },
+ {
+ "epoch": 45.50305676855895,
+ "grad_norm": 0.029923981055617332,
+ "learning_rate": 0.0006,
+ "loss": 6.157269477844238,
+ "step": 3276
+ },
+ {
+ "epoch": 45.51703056768559,
+ "grad_norm": 0.02521538734436035,
+ "learning_rate": 0.0006,
+ "loss": 6.146918773651123,
+ "step": 3277
+ },
+ {
+ "epoch": 45.531004366812226,
+ "grad_norm": 0.02758920192718506,
+ "learning_rate": 0.0006,
+ "loss": 6.0446977615356445,
+ "step": 3278
+ },
+ {
+ "epoch": 45.544978165938865,
+ "grad_norm": 0.024328220635652542,
+ "learning_rate": 0.0006,
+ "loss": 6.099508285522461,
+ "step": 3279
+ },
+ {
+ "epoch": 45.5589519650655,
+ "grad_norm": 0.027425771579146385,
+ "learning_rate": 0.0006,
+ "loss": 6.084299087524414,
+ "step": 3280
+ },
+ {
+ "epoch": 45.57292576419214,
+ "grad_norm": 0.026927651837468147,
+ "learning_rate": 0.0006,
+ "loss": 6.146334171295166,
+ "step": 3281
+ },
+ {
+ "epoch": 45.58689956331878,
+ "grad_norm": 0.04054379090666771,
+ "learning_rate": 0.0006,
+ "loss": 6.1301493644714355,
+ "step": 3282
+ },
+ {
+ "epoch": 45.60087336244541,
+ "grad_norm": 0.04456290975213051,
+ "learning_rate": 0.0006,
+ "loss": 6.152511119842529,
+ "step": 3283
+ },
+ {
+ "epoch": 45.61484716157205,
+ "grad_norm": 0.052821505814790726,
+ "learning_rate": 0.0006,
+ "loss": 6.113544940948486,
+ "step": 3284
+ },
+ {
+ "epoch": 45.62882096069869,
+ "grad_norm": 0.07251739501953125,
+ "learning_rate": 0.0006,
+ "loss": 6.118298530578613,
+ "step": 3285
+ },
+ {
+ "epoch": 45.64279475982533,
+ "grad_norm": 0.09433556348085403,
+ "learning_rate": 0.0006,
+ "loss": 6.170897483825684,
+ "step": 3286
+ },
+ {
+ "epoch": 45.65676855895197,
+ "grad_norm": 0.057683493942022324,
+ "learning_rate": 0.0006,
+ "loss": 6.1747822761535645,
+ "step": 3287
+ },
+ {
+ "epoch": 45.670742358078606,
+ "grad_norm": 0.06363270431756973,
+ "learning_rate": 0.0006,
+ "loss": 6.1617889404296875,
+ "step": 3288
+ },
+ {
+ "epoch": 45.68471615720524,
+ "grad_norm": 0.07528169453144073,
+ "learning_rate": 0.0006,
+ "loss": 6.230060577392578,
+ "step": 3289
+ },
+ {
+ "epoch": 45.698689956331876,
+ "grad_norm": 0.07816170156002045,
+ "learning_rate": 0.0006,
+ "loss": 6.171257019042969,
+ "step": 3290
+ },
+ {
+ "epoch": 45.712663755458514,
+ "grad_norm": 0.05860164389014244,
+ "learning_rate": 0.0006,
+ "loss": 6.095484256744385,
+ "step": 3291
+ },
+ {
+ "epoch": 45.72663755458515,
+ "grad_norm": 0.04747429117560387,
+ "learning_rate": 0.0006,
+ "loss": 6.060067176818848,
+ "step": 3292
+ },
+ {
+ "epoch": 45.74061135371179,
+ "grad_norm": 0.04514474794268608,
+ "learning_rate": 0.0006,
+ "loss": 6.100218772888184,
+ "step": 3293
+ },
+ {
+ "epoch": 45.75458515283843,
+ "grad_norm": 0.05168681964278221,
+ "learning_rate": 0.0006,
+ "loss": 6.0621867179870605,
+ "step": 3294
+ },
+ {
+ "epoch": 45.76855895196506,
+ "grad_norm": 0.04665876924991608,
+ "learning_rate": 0.0006,
+ "loss": 6.118292808532715,
+ "step": 3295
+ },
+ {
+ "epoch": 45.7825327510917,
+ "grad_norm": 0.049374762922525406,
+ "learning_rate": 0.0006,
+ "loss": 6.098394393920898,
+ "step": 3296
+ },
+ {
+ "epoch": 45.79650655021834,
+ "grad_norm": 0.03793549910187721,
+ "learning_rate": 0.0006,
+ "loss": 6.069541931152344,
+ "step": 3297
+ },
+ {
+ "epoch": 45.81048034934498,
+ "grad_norm": 0.036362383514642715,
+ "learning_rate": 0.0006,
+ "loss": 6.1188435554504395,
+ "step": 3298
+ },
+ {
+ "epoch": 45.82445414847162,
+ "grad_norm": 0.03397097811102867,
+ "learning_rate": 0.0006,
+ "loss": 6.117589950561523,
+ "step": 3299
+ },
+ {
+ "epoch": 45.838427947598255,
+ "grad_norm": 0.030977753922343254,
+ "learning_rate": 0.0006,
+ "loss": 6.0236711502075195,
+ "step": 3300
+ },
+ {
+ "epoch": 45.852401746724894,
+ "grad_norm": 0.02800217643380165,
+ "learning_rate": 0.0006,
+ "loss": 6.034825325012207,
+ "step": 3301
+ },
+ {
+ "epoch": 45.866375545851525,
+ "grad_norm": 0.03129778802394867,
+ "learning_rate": 0.0006,
+ "loss": 5.980271339416504,
+ "step": 3302
+ },
+ {
+ "epoch": 45.880349344978164,
+ "grad_norm": 0.025886094197630882,
+ "learning_rate": 0.0006,
+ "loss": 5.966419219970703,
+ "step": 3303
+ },
+ {
+ "epoch": 45.8943231441048,
+ "grad_norm": 0.025712328031659126,
+ "learning_rate": 0.0006,
+ "loss": 5.938652515411377,
+ "step": 3304
+ },
+ {
+ "epoch": 45.90829694323144,
+ "grad_norm": 0.022508298978209496,
+ "learning_rate": 0.0006,
+ "loss": 5.982312202453613,
+ "step": 3305
+ },
+ {
+ "epoch": 45.92227074235808,
+ "grad_norm": 0.024880951270461082,
+ "learning_rate": 0.0006,
+ "loss": 5.931150436401367,
+ "step": 3306
+ },
+ {
+ "epoch": 45.93624454148472,
+ "grad_norm": 0.030007576569914818,
+ "learning_rate": 0.0006,
+ "loss": 5.936639308929443,
+ "step": 3307
+ },
+ {
+ "epoch": 45.95021834061135,
+ "grad_norm": 0.033314041793346405,
+ "learning_rate": 0.0006,
+ "loss": 5.944248676300049,
+ "step": 3308
+ },
+ {
+ "epoch": 45.96419213973799,
+ "grad_norm": 0.043078381568193436,
+ "learning_rate": 0.0006,
+ "loss": 5.971286773681641,
+ "step": 3309
+ },
+ {
+ "epoch": 45.97816593886463,
+ "grad_norm": 0.059296440333127975,
+ "learning_rate": 0.0006,
+ "loss": 5.910497665405273,
+ "step": 3310
+ },
+ {
+ "epoch": 45.992139737991266,
+ "grad_norm": 0.06232666224241257,
+ "learning_rate": 0.0006,
+ "loss": 5.929817199707031,
+ "step": 3311
+ },
+ {
+ "epoch": 46.0,
+ "grad_norm": 0.05303940549492836,
+ "learning_rate": 0.0006,
+ "loss": 5.910822868347168,
+ "step": 3312
+ },
+ {
+ "epoch": 46.0,
+ "eval_loss": 5.977851390838623,
+ "eval_runtime": 56.2435,
+ "eval_samples_per_second": 43.418,
+ "eval_steps_per_second": 1.369,
+ "step": 3312
+ },
+ {
+ "epoch": 46.01397379912664,
+ "grad_norm": 0.06943798810243607,
+ "learning_rate": 0.0006,
+ "loss": 5.854922294616699,
+ "step": 3313
+ },
+ {
+ "epoch": 46.02794759825328,
+ "grad_norm": 0.05236278846859932,
+ "learning_rate": 0.0006,
+ "loss": 5.876611232757568,
+ "step": 3314
+ },
+ {
+ "epoch": 46.041921397379916,
+ "grad_norm": 0.06240001320838928,
+ "learning_rate": 0.0006,
+ "loss": 5.878912448883057,
+ "step": 3315
+ },
+ {
+ "epoch": 46.05589519650655,
+ "grad_norm": 0.07401192933320999,
+ "learning_rate": 0.0006,
+ "loss": 5.985136032104492,
+ "step": 3316
+ },
+ {
+ "epoch": 46.069868995633186,
+ "grad_norm": 0.0826529711484909,
+ "learning_rate": 0.0006,
+ "loss": 5.939439296722412,
+ "step": 3317
+ },
+ {
+ "epoch": 46.083842794759825,
+ "grad_norm": 0.07999005168676376,
+ "learning_rate": 0.0006,
+ "loss": 5.982336044311523,
+ "step": 3318
+ },
+ {
+ "epoch": 46.09781659388646,
+ "grad_norm": 0.07014869153499603,
+ "learning_rate": 0.0006,
+ "loss": 5.896534442901611,
+ "step": 3319
+ },
+ {
+ "epoch": 46.1117903930131,
+ "grad_norm": 0.05795128643512726,
+ "learning_rate": 0.0006,
+ "loss": 5.843111991882324,
+ "step": 3320
+ },
+ {
+ "epoch": 46.12576419213974,
+ "grad_norm": 0.06374258548021317,
+ "learning_rate": 0.0006,
+ "loss": 5.78752326965332,
+ "step": 3321
+ },
+ {
+ "epoch": 46.13973799126637,
+ "grad_norm": 0.045524805784225464,
+ "learning_rate": 0.0006,
+ "loss": 5.86755895614624,
+ "step": 3322
+ },
+ {
+ "epoch": 46.15371179039301,
+ "grad_norm": 0.03773504123091698,
+ "learning_rate": 0.0006,
+ "loss": 5.83627462387085,
+ "step": 3323
+ },
+ {
+ "epoch": 46.16768558951965,
+ "grad_norm": 0.035815201699733734,
+ "learning_rate": 0.0006,
+ "loss": 5.798238277435303,
+ "step": 3324
+ },
+ {
+ "epoch": 46.18165938864629,
+ "grad_norm": 0.03756783902645111,
+ "learning_rate": 0.0006,
+ "loss": 5.845197677612305,
+ "step": 3325
+ },
+ {
+ "epoch": 46.19563318777293,
+ "grad_norm": 0.03569113463163376,
+ "learning_rate": 0.0006,
+ "loss": 5.888932704925537,
+ "step": 3326
+ },
+ {
+ "epoch": 46.209606986899566,
+ "grad_norm": 0.029467813670635223,
+ "learning_rate": 0.0006,
+ "loss": 5.792047500610352,
+ "step": 3327
+ },
+ {
+ "epoch": 46.223580786026204,
+ "grad_norm": 0.025698406621813774,
+ "learning_rate": 0.0006,
+ "loss": 5.783734321594238,
+ "step": 3328
+ },
+ {
+ "epoch": 46.237554585152836,
+ "grad_norm": 0.028426671400666237,
+ "learning_rate": 0.0006,
+ "loss": 5.762084007263184,
+ "step": 3329
+ },
+ {
+ "epoch": 46.251528384279474,
+ "grad_norm": 0.03099622018635273,
+ "learning_rate": 0.0006,
+ "loss": 5.719109058380127,
+ "step": 3330
+ },
+ {
+ "epoch": 46.26550218340611,
+ "grad_norm": 0.0310219656676054,
+ "learning_rate": 0.0006,
+ "loss": 5.737057209014893,
+ "step": 3331
+ },
+ {
+ "epoch": 46.27947598253275,
+ "grad_norm": 0.02713446132838726,
+ "learning_rate": 0.0006,
+ "loss": 5.758657455444336,
+ "step": 3332
+ },
+ {
+ "epoch": 46.29344978165939,
+ "grad_norm": 0.03099844604730606,
+ "learning_rate": 0.0006,
+ "loss": 5.742086410522461,
+ "step": 3333
+ },
+ {
+ "epoch": 46.30742358078603,
+ "grad_norm": 0.030258730053901672,
+ "learning_rate": 0.0006,
+ "loss": 5.767426490783691,
+ "step": 3334
+ },
+ {
+ "epoch": 46.32139737991266,
+ "grad_norm": 0.029034852981567383,
+ "learning_rate": 0.0006,
+ "loss": 5.651792049407959,
+ "step": 3335
+ },
+ {
+ "epoch": 46.3353711790393,
+ "grad_norm": 0.03138665854930878,
+ "learning_rate": 0.0006,
+ "loss": 5.664980888366699,
+ "step": 3336
+ },
+ {
+ "epoch": 46.34934497816594,
+ "grad_norm": 0.040546901524066925,
+ "learning_rate": 0.0006,
+ "loss": 5.665165901184082,
+ "step": 3337
+ },
+ {
+ "epoch": 46.36331877729258,
+ "grad_norm": 0.06395485252141953,
+ "learning_rate": 0.0006,
+ "loss": 5.7776713371276855,
+ "step": 3338
+ },
+ {
+ "epoch": 46.377292576419215,
+ "grad_norm": 0.10088056325912476,
+ "learning_rate": 0.0006,
+ "loss": 5.755645751953125,
+ "step": 3339
+ },
+ {
+ "epoch": 46.391266375545854,
+ "grad_norm": 0.09527938067913055,
+ "learning_rate": 0.0006,
+ "loss": 5.838809490203857,
+ "step": 3340
+ },
+ {
+ "epoch": 46.40524017467249,
+ "grad_norm": 0.0849648267030716,
+ "learning_rate": 0.0006,
+ "loss": 5.781932830810547,
+ "step": 3341
+ },
+ {
+ "epoch": 46.419213973799124,
+ "grad_norm": 0.10142666101455688,
+ "learning_rate": 0.0006,
+ "loss": 5.838970184326172,
+ "step": 3342
+ },
+ {
+ "epoch": 46.43318777292576,
+ "grad_norm": 0.42731454968452454,
+ "learning_rate": 0.0006,
+ "loss": 6.071902275085449,
+ "step": 3343
+ },
+ {
+ "epoch": 46.4471615720524,
+ "grad_norm": 0.13102488219738007,
+ "learning_rate": 0.0006,
+ "loss": 6.8635993003845215,
+ "step": 3344
+ },
+ {
+ "epoch": 46.46113537117904,
+ "grad_norm": 0.23691938817501068,
+ "learning_rate": 0.0006,
+ "loss": 7.313158988952637,
+ "step": 3345
+ },
+ {
+ "epoch": 46.47510917030568,
+ "grad_norm": 0.11670669168233871,
+ "learning_rate": 0.0006,
+ "loss": 7.240363121032715,
+ "step": 3346
+ },
+ {
+ "epoch": 46.48908296943232,
+ "grad_norm": 0.09985820949077606,
+ "learning_rate": 0.0006,
+ "loss": 7.113332271575928,
+ "step": 3347
+ },
+ {
+ "epoch": 46.50305676855895,
+ "grad_norm": 0.11406408250331879,
+ "learning_rate": 0.0006,
+ "loss": 7.051285743713379,
+ "step": 3348
+ },
+ {
+ "epoch": 46.51703056768559,
+ "grad_norm": 0.07900336384773254,
+ "learning_rate": 0.0006,
+ "loss": 7.014365196228027,
+ "step": 3349
+ },
+ {
+ "epoch": 46.531004366812226,
+ "grad_norm": 0.05380333960056305,
+ "learning_rate": 0.0006,
+ "loss": 6.8868818283081055,
+ "step": 3350
+ },
+ {
+ "epoch": 46.544978165938865,
+ "grad_norm": 0.05868449807167053,
+ "learning_rate": 0.0006,
+ "loss": 6.783599853515625,
+ "step": 3351
+ },
+ {
+ "epoch": 46.5589519650655,
+ "grad_norm": 0.04428785294294357,
+ "learning_rate": 0.0006,
+ "loss": 6.798511505126953,
+ "step": 3352
+ },
+ {
+ "epoch": 46.57292576419214,
+ "grad_norm": 0.055527277290821075,
+ "learning_rate": 0.0006,
+ "loss": 6.694507598876953,
+ "step": 3353
+ },
+ {
+ "epoch": 46.58689956331878,
+ "grad_norm": 0.03872432932257652,
+ "learning_rate": 0.0006,
+ "loss": 6.736631393432617,
+ "step": 3354
+ },
+ {
+ "epoch": 46.60087336244541,
+ "grad_norm": 0.04330779239535332,
+ "learning_rate": 0.0006,
+ "loss": 6.576806545257568,
+ "step": 3355
+ },
+ {
+ "epoch": 46.61484716157205,
+ "grad_norm": 0.036767736077308655,
+ "learning_rate": 0.0006,
+ "loss": 6.559871196746826,
+ "step": 3356
+ },
+ {
+ "epoch": 46.62882096069869,
+ "grad_norm": 0.039857108145952225,
+ "learning_rate": 0.0006,
+ "loss": 6.5749831199646,
+ "step": 3357
+ },
+ {
+ "epoch": 46.64279475982533,
+ "grad_norm": 0.048719972372055054,
+ "learning_rate": 0.0006,
+ "loss": 6.4915618896484375,
+ "step": 3358
+ },
+ {
+ "epoch": 46.65676855895197,
+ "grad_norm": 0.042849015444517136,
+ "learning_rate": 0.0006,
+ "loss": 6.413516044616699,
+ "step": 3359
+ },
+ {
+ "epoch": 46.670742358078606,
+ "grad_norm": 0.02871009334921837,
+ "learning_rate": 0.0006,
+ "loss": 6.320952415466309,
+ "step": 3360
+ },
+ {
+ "epoch": 46.68471615720524,
+ "grad_norm": 0.03193620592355728,
+ "learning_rate": 0.0006,
+ "loss": 6.4176435470581055,
+ "step": 3361
+ },
+ {
+ "epoch": 46.698689956331876,
+ "grad_norm": 0.027417024597525597,
+ "learning_rate": 0.0006,
+ "loss": 6.342319488525391,
+ "step": 3362
+ },
+ {
+ "epoch": 46.712663755458514,
+ "grad_norm": 0.03305937349796295,
+ "learning_rate": 0.0006,
+ "loss": 6.246112823486328,
+ "step": 3363
+ },
+ {
+ "epoch": 46.72663755458515,
+ "grad_norm": 0.031403183937072754,
+ "learning_rate": 0.0006,
+ "loss": 6.149503231048584,
+ "step": 3364
+ },
+ {
+ "epoch": 46.74061135371179,
+ "grad_norm": 0.031563788652420044,
+ "learning_rate": 0.0006,
+ "loss": 6.278808116912842,
+ "step": 3365
+ },
+ {
+ "epoch": 46.75458515283843,
+ "grad_norm": 0.028366589918732643,
+ "learning_rate": 0.0006,
+ "loss": 6.057036876678467,
+ "step": 3366
+ },
+ {
+ "epoch": 46.76855895196506,
+ "grad_norm": 0.023715803399682045,
+ "learning_rate": 0.0006,
+ "loss": 6.111324310302734,
+ "step": 3367
+ },
+ {
+ "epoch": 46.7825327510917,
+ "grad_norm": 0.026546811684966087,
+ "learning_rate": 0.0006,
+ "loss": 6.105669975280762,
+ "step": 3368
+ },
+ {
+ "epoch": 46.79650655021834,
+ "grad_norm": 0.023531019687652588,
+ "learning_rate": 0.0006,
+ "loss": 6.068072319030762,
+ "step": 3369
+ },
+ {
+ "epoch": 46.81048034934498,
+ "grad_norm": 0.024480128660798073,
+ "learning_rate": 0.0006,
+ "loss": 5.947896957397461,
+ "step": 3370
+ },
+ {
+ "epoch": 46.82445414847162,
+ "grad_norm": 0.02517310529947281,
+ "learning_rate": 0.0006,
+ "loss": 6.030037879943848,
+ "step": 3371
+ },
+ {
+ "epoch": 46.838427947598255,
+ "grad_norm": 0.026723403483629227,
+ "learning_rate": 0.0006,
+ "loss": 5.963776111602783,
+ "step": 3372
+ },
+ {
+ "epoch": 46.852401746724894,
+ "grad_norm": 0.03163504600524902,
+ "learning_rate": 0.0006,
+ "loss": 5.893466949462891,
+ "step": 3373
+ },
+ {
+ "epoch": 46.866375545851525,
+ "grad_norm": 0.04058683291077614,
+ "learning_rate": 0.0006,
+ "loss": 5.9202799797058105,
+ "step": 3374
+ },
+ {
+ "epoch": 46.880349344978164,
+ "grad_norm": 0.04964117333292961,
+ "learning_rate": 0.0006,
+ "loss": 5.9366559982299805,
+ "step": 3375
+ },
+ {
+ "epoch": 46.8943231441048,
+ "grad_norm": 0.055406033992767334,
+ "learning_rate": 0.0006,
+ "loss": 5.912230968475342,
+ "step": 3376
+ },
+ {
+ "epoch": 46.90829694323144,
+ "grad_norm": 0.04643625393509865,
+ "learning_rate": 0.0006,
+ "loss": 5.87260103225708,
+ "step": 3377
+ },
+ {
+ "epoch": 46.92227074235808,
+ "grad_norm": 0.029325876384973526,
+ "learning_rate": 0.0006,
+ "loss": 5.854763984680176,
+ "step": 3378
+ },
+ {
+ "epoch": 46.93624454148472,
+ "grad_norm": 0.03159382566809654,
+ "learning_rate": 0.0006,
+ "loss": 5.880882263183594,
+ "step": 3379
+ },
+ {
+ "epoch": 46.95021834061135,
+ "grad_norm": 0.030224012210965157,
+ "learning_rate": 0.0006,
+ "loss": 5.787360191345215,
+ "step": 3380
+ },
+ {
+ "epoch": 46.96419213973799,
+ "grad_norm": 0.028368130326271057,
+ "learning_rate": 0.0006,
+ "loss": 5.817195892333984,
+ "step": 3381
+ },
+ {
+ "epoch": 46.97816593886463,
+ "grad_norm": 0.029411399737000465,
+ "learning_rate": 0.0006,
+ "loss": 5.783403396606445,
+ "step": 3382
+ },
+ {
+ "epoch": 46.992139737991266,
+ "grad_norm": 0.029927462339401245,
+ "learning_rate": 0.0006,
+ "loss": 5.662444591522217,
+ "step": 3383
+ },
+ {
+ "epoch": 47.0,
+ "grad_norm": 0.02872961387038231,
+ "learning_rate": 0.0006,
+ "loss": 5.742781639099121,
+ "step": 3384
+ },
+ {
+ "epoch": 47.0,
+ "eval_loss": 5.870260238647461,
+ "eval_runtime": 57.3734,
+ "eval_samples_per_second": 42.563,
+ "eval_steps_per_second": 1.342,
+ "step": 3384
+ },
+ {
+ "epoch": 47.01397379912664,
+ "grad_norm": 0.028542131185531616,
+ "learning_rate": 0.0006,
+ "loss": 5.753256797790527,
+ "step": 3385
+ },
+ {
+ "epoch": 47.02794759825328,
+ "grad_norm": 0.03227119892835617,
+ "learning_rate": 0.0006,
+ "loss": 5.722246170043945,
+ "step": 3386
+ },
+ {
+ "epoch": 47.041921397379916,
+ "grad_norm": 0.031461723148822784,
+ "learning_rate": 0.0006,
+ "loss": 5.740520000457764,
+ "step": 3387
+ },
+ {
+ "epoch": 47.05589519650655,
+ "grad_norm": 0.03184220567345619,
+ "learning_rate": 0.0006,
+ "loss": 5.720890045166016,
+ "step": 3388
+ },
+ {
+ "epoch": 47.069868995633186,
+ "grad_norm": 0.030015580356121063,
+ "learning_rate": 0.0006,
+ "loss": 5.597814559936523,
+ "step": 3389
+ },
+ {
+ "epoch": 47.083842794759825,
+ "grad_norm": 0.029931779950857162,
+ "learning_rate": 0.0006,
+ "loss": 5.5412468910217285,
+ "step": 3390
+ },
+ {
+ "epoch": 47.09781659388646,
+ "grad_norm": 0.03206505626440048,
+ "learning_rate": 0.0006,
+ "loss": 5.6967597007751465,
+ "step": 3391
+ },
+ {
+ "epoch": 47.1117903930131,
+ "grad_norm": 0.03978161886334419,
+ "learning_rate": 0.0006,
+ "loss": 5.59661340713501,
+ "step": 3392
+ },
+ {
+ "epoch": 47.12576419213974,
+ "grad_norm": 0.04619080200791359,
+ "learning_rate": 0.0006,
+ "loss": 5.795788764953613,
+ "step": 3393
+ },
+ {
+ "epoch": 47.13973799126637,
+ "grad_norm": 0.049085382372140884,
+ "learning_rate": 0.0006,
+ "loss": 5.667335510253906,
+ "step": 3394
+ },
+ {
+ "epoch": 47.15371179039301,
+ "grad_norm": 0.04830858111381531,
+ "learning_rate": 0.0006,
+ "loss": 5.668806552886963,
+ "step": 3395
+ },
+ {
+ "epoch": 47.16768558951965,
+ "grad_norm": 0.0583178736269474,
+ "learning_rate": 0.0006,
+ "loss": 5.692285537719727,
+ "step": 3396
+ },
+ {
+ "epoch": 47.18165938864629,
+ "grad_norm": 0.05182253196835518,
+ "learning_rate": 0.0006,
+ "loss": 5.678645133972168,
+ "step": 3397
+ },
+ {
+ "epoch": 47.19563318777293,
+ "grad_norm": 0.04011905938386917,
+ "learning_rate": 0.0006,
+ "loss": 5.661550998687744,
+ "step": 3398
+ },
+ {
+ "epoch": 47.209606986899566,
+ "grad_norm": 0.03888486698269844,
+ "learning_rate": 0.0006,
+ "loss": 5.643001556396484,
+ "step": 3399
+ },
+ {
+ "epoch": 47.223580786026204,
+ "grad_norm": 0.030658265575766563,
+ "learning_rate": 0.0006,
+ "loss": 5.659090995788574,
+ "step": 3400
+ },
+ {
+ "epoch": 47.237554585152836,
+ "grad_norm": 0.03609256446361542,
+ "learning_rate": 0.0006,
+ "loss": 5.544501781463623,
+ "step": 3401
+ },
+ {
+ "epoch": 47.251528384279474,
+ "grad_norm": 0.033589769154787064,
+ "learning_rate": 0.0006,
+ "loss": 5.463878631591797,
+ "step": 3402
+ },
+ {
+ "epoch": 47.26550218340611,
+ "grad_norm": 0.035128287971019745,
+ "learning_rate": 0.0006,
+ "loss": 5.5680060386657715,
+ "step": 3403
+ },
+ {
+ "epoch": 47.27947598253275,
+ "grad_norm": 0.038684550672769547,
+ "learning_rate": 0.0006,
+ "loss": 5.587898254394531,
+ "step": 3404
+ },
+ {
+ "epoch": 47.29344978165939,
+ "grad_norm": 0.043566226959228516,
+ "learning_rate": 0.0006,
+ "loss": 5.580896854400635,
+ "step": 3405
+ },
+ {
+ "epoch": 47.30742358078603,
+ "grad_norm": 0.04428374394774437,
+ "learning_rate": 0.0006,
+ "loss": 5.603460311889648,
+ "step": 3406
+ },
+ {
+ "epoch": 47.32139737991266,
+ "grad_norm": 0.037368934601545334,
+ "learning_rate": 0.0006,
+ "loss": 5.564740180969238,
+ "step": 3407
+ },
+ {
+ "epoch": 47.3353711790393,
+ "grad_norm": 0.04092350974678993,
+ "learning_rate": 0.0006,
+ "loss": 5.584153652191162,
+ "step": 3408
+ },
+ {
+ "epoch": 47.34934497816594,
+ "grad_norm": 0.045031506568193436,
+ "learning_rate": 0.0006,
+ "loss": 5.601186275482178,
+ "step": 3409
+ },
+ {
+ "epoch": 47.36331877729258,
+ "grad_norm": 0.041830409318208694,
+ "learning_rate": 0.0006,
+ "loss": 5.512788772583008,
+ "step": 3410
+ },
+ {
+ "epoch": 47.377292576419215,
+ "grad_norm": 0.037633974105119705,
+ "learning_rate": 0.0006,
+ "loss": 5.503828525543213,
+ "step": 3411
+ },
+ {
+ "epoch": 47.391266375545854,
+ "grad_norm": 0.03473588079214096,
+ "learning_rate": 0.0006,
+ "loss": 5.570557594299316,
+ "step": 3412
+ },
+ {
+ "epoch": 47.40524017467249,
+ "grad_norm": 0.039325181394815445,
+ "learning_rate": 0.0006,
+ "loss": 5.427783966064453,
+ "step": 3413
+ },
+ {
+ "epoch": 47.419213973799124,
+ "grad_norm": 0.04034624621272087,
+ "learning_rate": 0.0006,
+ "loss": 5.572591304779053,
+ "step": 3414
+ },
+ {
+ "epoch": 47.43318777292576,
+ "grad_norm": 0.05245184153318405,
+ "learning_rate": 0.0006,
+ "loss": 5.505208969116211,
+ "step": 3415
+ },
+ {
+ "epoch": 47.4471615720524,
+ "grad_norm": 0.0568675734102726,
+ "learning_rate": 0.0006,
+ "loss": 5.589373588562012,
+ "step": 3416
+ },
+ {
+ "epoch": 47.46113537117904,
+ "grad_norm": 0.048336587846279144,
+ "learning_rate": 0.0006,
+ "loss": 5.522005081176758,
+ "step": 3417
+ },
+ {
+ "epoch": 47.47510917030568,
+ "grad_norm": 0.04194975271821022,
+ "learning_rate": 0.0006,
+ "loss": 5.387576103210449,
+ "step": 3418
+ },
+ {
+ "epoch": 47.48908296943232,
+ "grad_norm": 0.04019433259963989,
+ "learning_rate": 0.0006,
+ "loss": 5.522615432739258,
+ "step": 3419
+ },
+ {
+ "epoch": 47.50305676855895,
+ "grad_norm": 0.05709933117032051,
+ "learning_rate": 0.0006,
+ "loss": 5.420286178588867,
+ "step": 3420
+ },
+ {
+ "epoch": 47.51703056768559,
+ "grad_norm": 0.09130338579416275,
+ "learning_rate": 0.0006,
+ "loss": 5.484622478485107,
+ "step": 3421
+ },
+ {
+ "epoch": 47.531004366812226,
+ "grad_norm": 0.2757541835308075,
+ "learning_rate": 0.0006,
+ "loss": 5.821832656860352,
+ "step": 3422
+ },
+ {
+ "epoch": 47.544978165938865,
+ "grad_norm": 0.28292080760002136,
+ "learning_rate": 0.0006,
+ "loss": 6.579934120178223,
+ "step": 3423
+ },
+ {
+ "epoch": 47.5589519650655,
+ "grad_norm": 0.15293604135513306,
+ "learning_rate": 0.0006,
+ "loss": 6.478213310241699,
+ "step": 3424
+ },
+ {
+ "epoch": 47.57292576419214,
+ "grad_norm": 0.11284393817186356,
+ "learning_rate": 0.0006,
+ "loss": 6.4066386222839355,
+ "step": 3425
+ },
+ {
+ "epoch": 47.58689956331878,
+ "grad_norm": 0.10933073610067368,
+ "learning_rate": 0.0006,
+ "loss": 6.3306169509887695,
+ "step": 3426
+ },
+ {
+ "epoch": 47.60087336244541,
+ "grad_norm": 0.05837641656398773,
+ "learning_rate": 0.0006,
+ "loss": 6.250000953674316,
+ "step": 3427
+ },
+ {
+ "epoch": 47.61484716157205,
+ "grad_norm": 0.05987229570746422,
+ "learning_rate": 0.0006,
+ "loss": 6.152463436126709,
+ "step": 3428
+ },
+ {
+ "epoch": 47.62882096069869,
+ "grad_norm": 0.04064938426017761,
+ "learning_rate": 0.0006,
+ "loss": 6.117897033691406,
+ "step": 3429
+ },
+ {
+ "epoch": 47.64279475982533,
+ "grad_norm": 0.04565057530999184,
+ "learning_rate": 0.0006,
+ "loss": 6.042291641235352,
+ "step": 3430
+ },
+ {
+ "epoch": 47.65676855895197,
+ "grad_norm": 0.03475063294172287,
+ "learning_rate": 0.0006,
+ "loss": 6.0071868896484375,
+ "step": 3431
+ },
+ {
+ "epoch": 47.670742358078606,
+ "grad_norm": 0.03809220343828201,
+ "learning_rate": 0.0006,
+ "loss": 6.0128021240234375,
+ "step": 3432
+ },
+ {
+ "epoch": 47.68471615720524,
+ "grad_norm": 0.03307219594717026,
+ "learning_rate": 0.0006,
+ "loss": 5.894664287567139,
+ "step": 3433
+ },
+ {
+ "epoch": 47.698689956331876,
+ "grad_norm": 0.028824346140027046,
+ "learning_rate": 0.0006,
+ "loss": 5.914865493774414,
+ "step": 3434
+ },
+ {
+ "epoch": 47.712663755458514,
+ "grad_norm": 0.027923690155148506,
+ "learning_rate": 0.0006,
+ "loss": 5.851759910583496,
+ "step": 3435
+ },
+ {
+ "epoch": 47.72663755458515,
+ "grad_norm": 0.025492684915661812,
+ "learning_rate": 0.0006,
+ "loss": 5.749186038970947,
+ "step": 3436
+ },
+ {
+ "epoch": 47.74061135371179,
+ "grad_norm": 0.025105174630880356,
+ "learning_rate": 0.0006,
+ "loss": 5.803566932678223,
+ "step": 3437
+ },
+ {
+ "epoch": 47.75458515283843,
+ "grad_norm": 0.023298880085349083,
+ "learning_rate": 0.0006,
+ "loss": 5.76276159286499,
+ "step": 3438
+ },
+ {
+ "epoch": 47.76855895196506,
+ "grad_norm": 0.02468576282262802,
+ "learning_rate": 0.0006,
+ "loss": 5.833852767944336,
+ "step": 3439
+ },
+ {
+ "epoch": 47.7825327510917,
+ "grad_norm": 0.022730689495801926,
+ "learning_rate": 0.0006,
+ "loss": 5.736724376678467,
+ "step": 3440
+ },
+ {
+ "epoch": 47.79650655021834,
+ "grad_norm": 0.021335354074835777,
+ "learning_rate": 0.0006,
+ "loss": 5.644248008728027,
+ "step": 3441
+ },
+ {
+ "epoch": 47.81048034934498,
+ "grad_norm": 0.02288101427257061,
+ "learning_rate": 0.0006,
+ "loss": 5.7011942863464355,
+ "step": 3442
+ },
+ {
+ "epoch": 47.82445414847162,
+ "grad_norm": 0.01961585320532322,
+ "learning_rate": 0.0006,
+ "loss": 5.63215446472168,
+ "step": 3443
+ },
+ {
+ "epoch": 47.838427947598255,
+ "grad_norm": 0.017659205943346024,
+ "learning_rate": 0.0006,
+ "loss": 5.744845390319824,
+ "step": 3444
+ },
+ {
+ "epoch": 47.852401746724894,
+ "grad_norm": 0.02281784638762474,
+ "learning_rate": 0.0006,
+ "loss": 5.696993350982666,
+ "step": 3445
+ },
+ {
+ "epoch": 47.866375545851525,
+ "grad_norm": 0.016394242644309998,
+ "learning_rate": 0.0006,
+ "loss": 5.659870147705078,
+ "step": 3446
+ },
+ {
+ "epoch": 47.880349344978164,
+ "grad_norm": 0.01967368647456169,
+ "learning_rate": 0.0006,
+ "loss": 5.638072490692139,
+ "step": 3447
+ },
+ {
+ "epoch": 47.8943231441048,
+ "grad_norm": 0.017326267436146736,
+ "learning_rate": 0.0006,
+ "loss": 5.638836860656738,
+ "step": 3448
+ },
+ {
+ "epoch": 47.90829694323144,
+ "grad_norm": 0.017778463661670685,
+ "learning_rate": 0.0006,
+ "loss": 5.634466648101807,
+ "step": 3449
+ },
+ {
+ "epoch": 47.92227074235808,
+ "grad_norm": 0.017269758507609367,
+ "learning_rate": 0.0006,
+ "loss": 5.506277561187744,
+ "step": 3450
+ },
+ {
+ "epoch": 47.93624454148472,
+ "grad_norm": 0.01733289286494255,
+ "learning_rate": 0.0006,
+ "loss": 5.544455528259277,
+ "step": 3451
+ },
+ {
+ "epoch": 47.95021834061135,
+ "grad_norm": 0.017787795513868332,
+ "learning_rate": 0.0006,
+ "loss": 5.498943328857422,
+ "step": 3452
+ },
+ {
+ "epoch": 47.96419213973799,
+ "grad_norm": 0.015770932659506798,
+ "learning_rate": 0.0006,
+ "loss": 5.537227630615234,
+ "step": 3453
+ },
+ {
+ "epoch": 47.97816593886463,
+ "grad_norm": 0.016132809221744537,
+ "learning_rate": 0.0006,
+ "loss": 5.489081382751465,
+ "step": 3454
+ },
+ {
+ "epoch": 47.992139737991266,
+ "grad_norm": 0.01749664731323719,
+ "learning_rate": 0.0006,
+ "loss": 5.454263687133789,
+ "step": 3455
+ },
+ {
+ "epoch": 48.0,
+ "grad_norm": 0.016711296513676643,
+ "learning_rate": 0.0006,
+ "loss": 5.530783653259277,
+ "step": 3456
+ },
+ {
+ "epoch": 48.0,
+ "eval_loss": 5.596925258636475,
+ "eval_runtime": 56.9612,
+ "eval_samples_per_second": 42.871,
+ "eval_steps_per_second": 1.352,
+ "step": 3456
+ },
+ {
+ "epoch": 48.01397379912664,
+ "grad_norm": 0.01867641694843769,
+ "learning_rate": 0.0006,
+ "loss": 5.506853103637695,
+ "step": 3457
+ },
+ {
+ "epoch": 48.02794759825328,
+ "grad_norm": 0.023250792175531387,
+ "learning_rate": 0.0006,
+ "loss": 5.4074177742004395,
+ "step": 3458
+ },
+ {
+ "epoch": 48.041921397379916,
+ "grad_norm": 0.01970936730504036,
+ "learning_rate": 0.0006,
+ "loss": 5.508777618408203,
+ "step": 3459
+ },
+ {
+ "epoch": 48.05589519650655,
+ "grad_norm": 0.01879260502755642,
+ "learning_rate": 0.0006,
+ "loss": 5.412424087524414,
+ "step": 3460
+ },
+ {
+ "epoch": 48.069868995633186,
+ "grad_norm": 0.017420092597603798,
+ "learning_rate": 0.0006,
+ "loss": 5.43361759185791,
+ "step": 3461
+ },
+ {
+ "epoch": 48.083842794759825,
+ "grad_norm": 0.018707862123847008,
+ "learning_rate": 0.0006,
+ "loss": 5.476800918579102,
+ "step": 3462
+ },
+ {
+ "epoch": 48.09781659388646,
+ "grad_norm": 0.023185882717370987,
+ "learning_rate": 0.0006,
+ "loss": 5.465272426605225,
+ "step": 3463
+ },
+ {
+ "epoch": 48.1117903930131,
+ "grad_norm": 0.023570368066430092,
+ "learning_rate": 0.0006,
+ "loss": 5.391471862792969,
+ "step": 3464
+ },
+ {
+ "epoch": 48.12576419213974,
+ "grad_norm": 0.02651144750416279,
+ "learning_rate": 0.0006,
+ "loss": 5.422799110412598,
+ "step": 3465
+ },
+ {
+ "epoch": 48.13973799126637,
+ "grad_norm": 0.03316329047083855,
+ "learning_rate": 0.0006,
+ "loss": 5.3952202796936035,
+ "step": 3466
+ },
+ {
+ "epoch": 48.15371179039301,
+ "grad_norm": 0.04012278467416763,
+ "learning_rate": 0.0006,
+ "loss": 5.540853023529053,
+ "step": 3467
+ },
+ {
+ "epoch": 48.16768558951965,
+ "grad_norm": 0.04763203114271164,
+ "learning_rate": 0.0006,
+ "loss": 5.387890338897705,
+ "step": 3468
+ },
+ {
+ "epoch": 48.18165938864629,
+ "grad_norm": 0.05129165202379227,
+ "learning_rate": 0.0006,
+ "loss": 5.414722442626953,
+ "step": 3469
+ },
+ {
+ "epoch": 48.19563318777293,
+ "grad_norm": 0.04438071325421333,
+ "learning_rate": 0.0006,
+ "loss": 5.386244773864746,
+ "step": 3470
+ },
+ {
+ "epoch": 48.209606986899566,
+ "grad_norm": 0.040234919637441635,
+ "learning_rate": 0.0006,
+ "loss": 5.431617736816406,
+ "step": 3471
+ },
+ {
+ "epoch": 48.223580786026204,
+ "grad_norm": 0.03217008337378502,
+ "learning_rate": 0.0006,
+ "loss": 5.30168342590332,
+ "step": 3472
+ },
+ {
+ "epoch": 48.237554585152836,
+ "grad_norm": 0.03087974339723587,
+ "learning_rate": 0.0006,
+ "loss": 5.316389083862305,
+ "step": 3473
+ },
+ {
+ "epoch": 48.251528384279474,
+ "grad_norm": 0.02789347991347313,
+ "learning_rate": 0.0006,
+ "loss": 5.385970592498779,
+ "step": 3474
+ },
+ {
+ "epoch": 48.26550218340611,
+ "grad_norm": 0.024792097508907318,
+ "learning_rate": 0.0006,
+ "loss": 5.362329959869385,
+ "step": 3475
+ },
+ {
+ "epoch": 48.27947598253275,
+ "grad_norm": 0.025869522243738174,
+ "learning_rate": 0.0006,
+ "loss": 5.28400182723999,
+ "step": 3476
+ },
+ {
+ "epoch": 48.29344978165939,
+ "grad_norm": 0.021339882165193558,
+ "learning_rate": 0.0006,
+ "loss": 5.36669921875,
+ "step": 3477
+ },
+ {
+ "epoch": 48.30742358078603,
+ "grad_norm": 0.021462520584464073,
+ "learning_rate": 0.0006,
+ "loss": 5.311222076416016,
+ "step": 3478
+ },
+ {
+ "epoch": 48.32139737991266,
+ "grad_norm": 0.02193879708647728,
+ "learning_rate": 0.0006,
+ "loss": 5.291441917419434,
+ "step": 3479
+ },
+ {
+ "epoch": 48.3353711790393,
+ "grad_norm": 0.01963093690574169,
+ "learning_rate": 0.0006,
+ "loss": 5.342780113220215,
+ "step": 3480
+ },
+ {
+ "epoch": 48.34934497816594,
+ "grad_norm": 0.020634736865758896,
+ "learning_rate": 0.0006,
+ "loss": 5.144141674041748,
+ "step": 3481
+ },
+ {
+ "epoch": 48.36331877729258,
+ "grad_norm": 0.02126968279480934,
+ "learning_rate": 0.0006,
+ "loss": 5.232624530792236,
+ "step": 3482
+ },
+ {
+ "epoch": 48.377292576419215,
+ "grad_norm": 0.021203994750976562,
+ "learning_rate": 0.0006,
+ "loss": 5.290219783782959,
+ "step": 3483
+ },
+ {
+ "epoch": 48.391266375545854,
+ "grad_norm": 0.02193225920200348,
+ "learning_rate": 0.0006,
+ "loss": 5.29197359085083,
+ "step": 3484
+ },
+ {
+ "epoch": 48.40524017467249,
+ "grad_norm": 0.023174211382865906,
+ "learning_rate": 0.0006,
+ "loss": 5.204349517822266,
+ "step": 3485
+ },
+ {
+ "epoch": 48.419213973799124,
+ "grad_norm": 0.024161914363503456,
+ "learning_rate": 0.0006,
+ "loss": 5.339351654052734,
+ "step": 3486
+ },
+ {
+ "epoch": 48.43318777292576,
+ "grad_norm": 0.02221524342894554,
+ "learning_rate": 0.0006,
+ "loss": 5.263001441955566,
+ "step": 3487
+ },
+ {
+ "epoch": 48.4471615720524,
+ "grad_norm": 0.026632830500602722,
+ "learning_rate": 0.0006,
+ "loss": 5.221281051635742,
+ "step": 3488
+ },
+ {
+ "epoch": 48.46113537117904,
+ "grad_norm": 0.034923870116472244,
+ "learning_rate": 0.0006,
+ "loss": 5.296039581298828,
+ "step": 3489
+ },
+ {
+ "epoch": 48.47510917030568,
+ "grad_norm": 0.046184979379177094,
+ "learning_rate": 0.0006,
+ "loss": 5.212769985198975,
+ "step": 3490
+ },
+ {
+ "epoch": 48.48908296943232,
+ "grad_norm": 0.05154098570346832,
+ "learning_rate": 0.0006,
+ "loss": 5.337864875793457,
+ "step": 3491
+ },
+ {
+ "epoch": 48.50305676855895,
+ "grad_norm": 0.042967259883880615,
+ "learning_rate": 0.0006,
+ "loss": 5.267958641052246,
+ "step": 3492
+ },
+ {
+ "epoch": 48.51703056768559,
+ "grad_norm": 0.05356583371758461,
+ "learning_rate": 0.0006,
+ "loss": 5.1404218673706055,
+ "step": 3493
+ },
+ {
+ "epoch": 48.531004366812226,
+ "grad_norm": 0.08219032734632492,
+ "learning_rate": 0.0006,
+ "loss": 5.300687789916992,
+ "step": 3494
+ },
+ {
+ "epoch": 48.544978165938865,
+ "grad_norm": 0.19204089045524597,
+ "learning_rate": 0.0006,
+ "loss": 5.456025123596191,
+ "step": 3495
+ },
+ {
+ "epoch": 48.5589519650655,
+ "grad_norm": 0.25917696952819824,
+ "learning_rate": 0.0006,
+ "loss": 6.156761169433594,
+ "step": 3496
+ },
+ {
+ "epoch": 48.57292576419214,
+ "grad_norm": 0.1566283404827118,
+ "learning_rate": 0.0006,
+ "loss": 5.862346649169922,
+ "step": 3497
+ },
+ {
+ "epoch": 48.58689956331878,
+ "grad_norm": 0.1761486828327179,
+ "learning_rate": 0.0006,
+ "loss": 6.012795448303223,
+ "step": 3498
+ },
+ {
+ "epoch": 48.60087336244541,
+ "grad_norm": 0.09857621043920517,
+ "learning_rate": 0.0006,
+ "loss": 5.9248151779174805,
+ "step": 3499
+ },
+ {
+ "epoch": 48.61484716157205,
+ "grad_norm": 0.05968055501580238,
+ "learning_rate": 0.0006,
+ "loss": 5.780706882476807,
+ "step": 3500
+ },
+ {
+ "epoch": 48.62882096069869,
+ "grad_norm": 0.057433564215898514,
+ "learning_rate": 0.0006,
+ "loss": 5.645388603210449,
+ "step": 3501
+ },
+ {
+ "epoch": 48.64279475982533,
+ "grad_norm": 0.039659615606069565,
+ "learning_rate": 0.0006,
+ "loss": 5.695138931274414,
+ "step": 3502
+ },
+ {
+ "epoch": 48.65676855895197,
+ "grad_norm": 0.04019695147871971,
+ "learning_rate": 0.0006,
+ "loss": 5.672435283660889,
+ "step": 3503
+ },
+ {
+ "epoch": 48.670742358078606,
+ "grad_norm": 0.03033018670976162,
+ "learning_rate": 0.0006,
+ "loss": 5.682260513305664,
+ "step": 3504
+ },
+ {
+ "epoch": 48.68471615720524,
+ "grad_norm": 0.03215787187218666,
+ "learning_rate": 0.0006,
+ "loss": 5.604182243347168,
+ "step": 3505
+ },
+ {
+ "epoch": 48.698689956331876,
+ "grad_norm": 0.026865126565098763,
+ "learning_rate": 0.0006,
+ "loss": 5.64040470123291,
+ "step": 3506
+ },
+ {
+ "epoch": 48.712663755458514,
+ "grad_norm": 0.02831609919667244,
+ "learning_rate": 0.0006,
+ "loss": 5.552720069885254,
+ "step": 3507
+ },
+ {
+ "epoch": 48.72663755458515,
+ "grad_norm": 0.024226613342761993,
+ "learning_rate": 0.0006,
+ "loss": 5.556946277618408,
+ "step": 3508
+ },
+ {
+ "epoch": 48.74061135371179,
+ "grad_norm": 0.025393934920430183,
+ "learning_rate": 0.0006,
+ "loss": 5.476178169250488,
+ "step": 3509
+ },
+ {
+ "epoch": 48.75458515283843,
+ "grad_norm": 0.01984412968158722,
+ "learning_rate": 0.0006,
+ "loss": 5.471088409423828,
+ "step": 3510
+ },
+ {
+ "epoch": 48.76855895196506,
+ "grad_norm": 0.02035129815340042,
+ "learning_rate": 0.0006,
+ "loss": 5.438812732696533,
+ "step": 3511
+ },
+ {
+ "epoch": 48.7825327510917,
+ "grad_norm": 0.0208350270986557,
+ "learning_rate": 0.0006,
+ "loss": 5.381746292114258,
+ "step": 3512
+ },
+ {
+ "epoch": 48.79650655021834,
+ "grad_norm": 0.017783882096409798,
+ "learning_rate": 0.0006,
+ "loss": 5.351675510406494,
+ "step": 3513
+ },
+ {
+ "epoch": 48.81048034934498,
+ "grad_norm": 0.01671770215034485,
+ "learning_rate": 0.0006,
+ "loss": 5.329428672790527,
+ "step": 3514
+ },
+ {
+ "epoch": 48.82445414847162,
+ "grad_norm": 0.017891215160489082,
+ "learning_rate": 0.0006,
+ "loss": 5.212074279785156,
+ "step": 3515
+ },
+ {
+ "epoch": 48.838427947598255,
+ "grad_norm": 0.016847791150212288,
+ "learning_rate": 0.0006,
+ "loss": 5.299534797668457,
+ "step": 3516
+ },
+ {
+ "epoch": 48.852401746724894,
+ "grad_norm": 0.01593637838959694,
+ "learning_rate": 0.0006,
+ "loss": 5.30466890335083,
+ "step": 3517
+ },
+ {
+ "epoch": 48.866375545851525,
+ "grad_norm": 0.014713102020323277,
+ "learning_rate": 0.0006,
+ "loss": 5.294345378875732,
+ "step": 3518
+ },
+ {
+ "epoch": 48.880349344978164,
+ "grad_norm": 0.01433857623487711,
+ "learning_rate": 0.0006,
+ "loss": 5.2148542404174805,
+ "step": 3519
+ },
+ {
+ "epoch": 48.8943231441048,
+ "grad_norm": 0.014979379251599312,
+ "learning_rate": 0.0006,
+ "loss": 5.298582077026367,
+ "step": 3520
+ },
+ {
+ "epoch": 48.90829694323144,
+ "grad_norm": 0.013494192622601986,
+ "learning_rate": 0.0006,
+ "loss": 5.309783935546875,
+ "step": 3521
+ },
+ {
+ "epoch": 48.92227074235808,
+ "grad_norm": 0.014020012691617012,
+ "learning_rate": 0.0006,
+ "loss": 5.284879684448242,
+ "step": 3522
+ },
+ {
+ "epoch": 48.93624454148472,
+ "grad_norm": 0.01348058134317398,
+ "learning_rate": 0.0006,
+ "loss": 5.249116897583008,
+ "step": 3523
+ },
+ {
+ "epoch": 48.95021834061135,
+ "grad_norm": 0.013280218467116356,
+ "learning_rate": 0.0006,
+ "loss": 5.273489952087402,
+ "step": 3524
+ },
+ {
+ "epoch": 48.96419213973799,
+ "grad_norm": 0.012647110037505627,
+ "learning_rate": 0.0006,
+ "loss": 5.306652069091797,
+ "step": 3525
+ },
+ {
+ "epoch": 48.97816593886463,
+ "grad_norm": 0.012847724370658398,
+ "learning_rate": 0.0006,
+ "loss": 5.235745429992676,
+ "step": 3526
+ },
+ {
+ "epoch": 48.992139737991266,
+ "grad_norm": 0.011820166371762753,
+ "learning_rate": 0.0006,
+ "loss": 5.238091468811035,
+ "step": 3527
+ },
+ {
+ "epoch": 49.0,
+ "grad_norm": 0.01255585066974163,
+ "learning_rate": 0.0006,
+ "loss": 5.231295585632324,
+ "step": 3528
+ },
+ {
+ "epoch": 49.0,
+ "eval_loss": 5.394587993621826,
+ "eval_runtime": 56.3734,
+ "eval_samples_per_second": 43.318,
+ "eval_steps_per_second": 1.366,
+ "step": 3528
+ },
+ {
+ "epoch": 49.01397379912664,
+ "grad_norm": 0.011031972244381905,
+ "learning_rate": 0.0006,
+ "loss": 5.1354570388793945,
+ "step": 3529
+ },
+ {
+ "epoch": 49.02794759825328,
+ "grad_norm": 0.010631263256072998,
+ "learning_rate": 0.0006,
+ "loss": 5.177639961242676,
+ "step": 3530
+ },
+ {
+ "epoch": 49.041921397379916,
+ "grad_norm": 0.01133775431662798,
+ "learning_rate": 0.0006,
+ "loss": 5.283428192138672,
+ "step": 3531
+ },
+ {
+ "epoch": 49.05589519650655,
+ "grad_norm": 0.01154172234237194,
+ "learning_rate": 0.0006,
+ "loss": 5.206364154815674,
+ "step": 3532
+ },
+ {
+ "epoch": 49.069868995633186,
+ "grad_norm": 0.010192356072366238,
+ "learning_rate": 0.0006,
+ "loss": 5.2385735511779785,
+ "step": 3533
+ },
+ {
+ "epoch": 49.083842794759825,
+ "grad_norm": 0.012789200060069561,
+ "learning_rate": 0.0006,
+ "loss": 5.071800231933594,
+ "step": 3534
+ },
+ {
+ "epoch": 49.09781659388646,
+ "grad_norm": 0.011282820254564285,
+ "learning_rate": 0.0006,
+ "loss": 5.218020439147949,
+ "step": 3535
+ },
+ {
+ "epoch": 49.1117903930131,
+ "grad_norm": 0.012153448536992073,
+ "learning_rate": 0.0006,
+ "loss": 5.134698867797852,
+ "step": 3536
+ },
+ {
+ "epoch": 49.12576419213974,
+ "grad_norm": 0.011920543387532234,
+ "learning_rate": 0.0006,
+ "loss": 5.141620635986328,
+ "step": 3537
+ },
+ {
+ "epoch": 49.13973799126637,
+ "grad_norm": 0.010949329473078251,
+ "learning_rate": 0.0006,
+ "loss": 5.116259574890137,
+ "step": 3538
+ },
+ {
+ "epoch": 49.15371179039301,
+ "grad_norm": 0.012362448498606682,
+ "learning_rate": 0.0006,
+ "loss": 5.090177059173584,
+ "step": 3539
+ },
+ {
+ "epoch": 49.16768558951965,
+ "grad_norm": 0.011575071141123772,
+ "learning_rate": 0.0006,
+ "loss": 5.093441009521484,
+ "step": 3540
+ },
+ {
+ "epoch": 49.18165938864629,
+ "grad_norm": 0.011126236990094185,
+ "learning_rate": 0.0006,
+ "loss": 5.060235023498535,
+ "step": 3541
+ },
+ {
+ "epoch": 49.19563318777293,
+ "grad_norm": 0.0107401292771101,
+ "learning_rate": 0.0006,
+ "loss": 5.140704154968262,
+ "step": 3542
+ },
+ {
+ "epoch": 49.209606986899566,
+ "grad_norm": 0.011202906258404255,
+ "learning_rate": 0.0006,
+ "loss": 5.10855770111084,
+ "step": 3543
+ },
+ {
+ "epoch": 49.223580786026204,
+ "grad_norm": 0.011083454824984074,
+ "learning_rate": 0.0006,
+ "loss": 5.092751502990723,
+ "step": 3544
+ },
+ {
+ "epoch": 49.237554585152836,
+ "grad_norm": 0.009523800574243069,
+ "learning_rate": 0.0006,
+ "loss": 5.13733434677124,
+ "step": 3545
+ },
+ {
+ "epoch": 49.251528384279474,
+ "grad_norm": 0.011298353783786297,
+ "learning_rate": 0.0006,
+ "loss": 5.036048889160156,
+ "step": 3546
+ },
+ {
+ "epoch": 49.26550218340611,
+ "grad_norm": 0.010676483623683453,
+ "learning_rate": 0.0006,
+ "loss": 5.051901340484619,
+ "step": 3547
+ },
+ {
+ "epoch": 49.27947598253275,
+ "grad_norm": 0.011463399045169353,
+ "learning_rate": 0.0006,
+ "loss": 4.983170986175537,
+ "step": 3548
+ },
+ {
+ "epoch": 49.29344978165939,
+ "grad_norm": 0.010603402741253376,
+ "learning_rate": 0.0006,
+ "loss": 5.112311363220215,
+ "step": 3549
+ },
+ {
+ "epoch": 49.30742358078603,
+ "grad_norm": 0.010535670444369316,
+ "learning_rate": 0.0006,
+ "loss": 5.085227966308594,
+ "step": 3550
+ },
+ {
+ "epoch": 49.32139737991266,
+ "grad_norm": 0.011843928135931492,
+ "learning_rate": 0.0006,
+ "loss": 5.07575798034668,
+ "step": 3551
+ },
+ {
+ "epoch": 49.3353711790393,
+ "grad_norm": 0.010587952099740505,
+ "learning_rate": 0.0006,
+ "loss": 5.019237041473389,
+ "step": 3552
+ },
+ {
+ "epoch": 49.34934497816594,
+ "grad_norm": 0.012285659089684486,
+ "learning_rate": 0.0006,
+ "loss": 5.144603729248047,
+ "step": 3553
+ },
+ {
+ "epoch": 49.36331877729258,
+ "grad_norm": 0.013119431212544441,
+ "learning_rate": 0.0006,
+ "loss": 5.035256862640381,
+ "step": 3554
+ },
+ {
+ "epoch": 49.377292576419215,
+ "grad_norm": 0.012138537131249905,
+ "learning_rate": 0.0006,
+ "loss": 4.996667861938477,
+ "step": 3555
+ },
+ {
+ "epoch": 49.391266375545854,
+ "grad_norm": 0.012142804451286793,
+ "learning_rate": 0.0006,
+ "loss": 4.95058536529541,
+ "step": 3556
+ },
+ {
+ "epoch": 49.40524017467249,
+ "grad_norm": 0.012205676175653934,
+ "learning_rate": 0.0006,
+ "loss": 5.096931457519531,
+ "step": 3557
+ },
+ {
+ "epoch": 49.419213973799124,
+ "grad_norm": 0.012086744420230389,
+ "learning_rate": 0.0006,
+ "loss": 4.99384880065918,
+ "step": 3558
+ },
+ {
+ "epoch": 49.43318777292576,
+ "grad_norm": 0.012918103486299515,
+ "learning_rate": 0.0006,
+ "loss": 5.031294822692871,
+ "step": 3559
+ },
+ {
+ "epoch": 49.4471615720524,
+ "grad_norm": 0.012006713077425957,
+ "learning_rate": 0.0006,
+ "loss": 5.0547943115234375,
+ "step": 3560
+ },
+ {
+ "epoch": 49.46113537117904,
+ "grad_norm": 0.0120399734005332,
+ "learning_rate": 0.0006,
+ "loss": 5.014238357543945,
+ "step": 3561
+ },
+ {
+ "epoch": 49.47510917030568,
+ "grad_norm": 0.013306990265846252,
+ "learning_rate": 0.0006,
+ "loss": 4.972925186157227,
+ "step": 3562
+ },
+ {
+ "epoch": 49.48908296943232,
+ "grad_norm": 0.013878636993467808,
+ "learning_rate": 0.0006,
+ "loss": 5.079033851623535,
+ "step": 3563
+ },
+ {
+ "epoch": 49.50305676855895,
+ "grad_norm": 0.012454566545784473,
+ "learning_rate": 0.0006,
+ "loss": 4.93708610534668,
+ "step": 3564
+ },
+ {
+ "epoch": 49.51703056768559,
+ "grad_norm": 0.014075061306357384,
+ "learning_rate": 0.0006,
+ "loss": 4.965582370758057,
+ "step": 3565
+ },
+ {
+ "epoch": 49.531004366812226,
+ "grad_norm": 0.012525617145001888,
+ "learning_rate": 0.0006,
+ "loss": 4.968148708343506,
+ "step": 3566
+ },
+ {
+ "epoch": 49.544978165938865,
+ "grad_norm": 0.012418225407600403,
+ "learning_rate": 0.0006,
+ "loss": 5.06901741027832,
+ "step": 3567
+ },
+ {
+ "epoch": 49.5589519650655,
+ "grad_norm": 0.015034239739179611,
+ "learning_rate": 0.0006,
+ "loss": 4.984277725219727,
+ "step": 3568
+ },
+ {
+ "epoch": 49.57292576419214,
+ "grad_norm": 0.014329628087580204,
+ "learning_rate": 0.0006,
+ "loss": 4.951558589935303,
+ "step": 3569
+ },
+ {
+ "epoch": 49.58689956331878,
+ "grad_norm": 0.017769070342183113,
+ "learning_rate": 0.0006,
+ "loss": 4.966532230377197,
+ "step": 3570
+ },
+ {
+ "epoch": 49.60087336244541,
+ "grad_norm": 0.01909453794360161,
+ "learning_rate": 0.0006,
+ "loss": 4.919890880584717,
+ "step": 3571
+ },
+ {
+ "epoch": 49.61484716157205,
+ "grad_norm": 0.017519688233733177,
+ "learning_rate": 0.0006,
+ "loss": 5.031286716461182,
+ "step": 3572
+ },
+ {
+ "epoch": 49.62882096069869,
+ "grad_norm": 0.017143897712230682,
+ "learning_rate": 0.0006,
+ "loss": 4.969764709472656,
+ "step": 3573
+ },
+ {
+ "epoch": 49.64279475982533,
+ "grad_norm": 0.01688285917043686,
+ "learning_rate": 0.0006,
+ "loss": 5.0733442306518555,
+ "step": 3574
+ },
+ {
+ "epoch": 49.65676855895197,
+ "grad_norm": 0.018547311425209045,
+ "learning_rate": 0.0006,
+ "loss": 4.980863571166992,
+ "step": 3575
+ },
+ {
+ "epoch": 49.670742358078606,
+ "grad_norm": 0.018877137452363968,
+ "learning_rate": 0.0006,
+ "loss": 4.954646110534668,
+ "step": 3576
+ },
+ {
+ "epoch": 49.68471615720524,
+ "grad_norm": 0.018681600689888,
+ "learning_rate": 0.0006,
+ "loss": 4.959702014923096,
+ "step": 3577
+ },
+ {
+ "epoch": 49.698689956331876,
+ "grad_norm": 0.018947336822748184,
+ "learning_rate": 0.0006,
+ "loss": 5.070502758026123,
+ "step": 3578
+ },
+ {
+ "epoch": 49.712663755458514,
+ "grad_norm": 0.018817484378814697,
+ "learning_rate": 0.0006,
+ "loss": 5.056508541107178,
+ "step": 3579
+ },
+ {
+ "epoch": 49.72663755458515,
+ "grad_norm": 0.01931724324822426,
+ "learning_rate": 0.0006,
+ "loss": 4.922451972961426,
+ "step": 3580
+ },
+ {
+ "epoch": 49.74061135371179,
+ "grad_norm": 0.022047819569706917,
+ "learning_rate": 0.0006,
+ "loss": 5.053170204162598,
+ "step": 3581
+ },
+ {
+ "epoch": 49.75458515283843,
+ "grad_norm": 0.02279801294207573,
+ "learning_rate": 0.0006,
+ "loss": 4.8768439292907715,
+ "step": 3582
+ },
+ {
+ "epoch": 49.76855895196506,
+ "grad_norm": 0.020661601796746254,
+ "learning_rate": 0.0006,
+ "loss": 4.907928466796875,
+ "step": 3583
+ },
+ {
+ "epoch": 49.7825327510917,
+ "grad_norm": 0.020479509606957436,
+ "learning_rate": 0.0006,
+ "loss": 5.040694236755371,
+ "step": 3584
+ },
+ {
+ "epoch": 49.79650655021834,
+ "grad_norm": 0.021654162555933,
+ "learning_rate": 0.0006,
+ "loss": 5.018434524536133,
+ "step": 3585
+ },
+ {
+ "epoch": 49.81048034934498,
+ "grad_norm": 0.022486824542284012,
+ "learning_rate": 0.0006,
+ "loss": 5.086538314819336,
+ "step": 3586
+ },
+ {
+ "epoch": 49.82445414847162,
+ "grad_norm": 0.02141369879245758,
+ "learning_rate": 0.0006,
+ "loss": 5.035694122314453,
+ "step": 3587
+ },
+ {
+ "epoch": 49.838427947598255,
+ "grad_norm": 0.015926677733659744,
+ "learning_rate": 0.0006,
+ "loss": 4.905038833618164,
+ "step": 3588
+ },
+ {
+ "epoch": 49.852401746724894,
+ "grad_norm": 0.015775298699736595,
+ "learning_rate": 0.0006,
+ "loss": 5.003902435302734,
+ "step": 3589
+ },
+ {
+ "epoch": 49.866375545851525,
+ "grad_norm": 0.01872118003666401,
+ "learning_rate": 0.0006,
+ "loss": 4.891060829162598,
+ "step": 3590
+ },
+ {
+ "epoch": 49.880349344978164,
+ "grad_norm": 0.02066192403435707,
+ "learning_rate": 0.0006,
+ "loss": 5.002897262573242,
+ "step": 3591
+ },
+ {
+ "epoch": 49.8943231441048,
+ "grad_norm": 0.02348714880645275,
+ "learning_rate": 0.0006,
+ "loss": 4.9189958572387695,
+ "step": 3592
+ },
+ {
+ "epoch": 49.90829694323144,
+ "grad_norm": 0.020402103662490845,
+ "learning_rate": 0.0006,
+ "loss": 4.992661476135254,
+ "step": 3593
+ },
+ {
+ "epoch": 49.92227074235808,
+ "grad_norm": 0.020466072484850883,
+ "learning_rate": 0.0006,
+ "loss": 4.962968826293945,
+ "step": 3594
+ },
+ {
+ "epoch": 49.93624454148472,
+ "grad_norm": 0.01922314241528511,
+ "learning_rate": 0.0006,
+ "loss": 4.897730827331543,
+ "step": 3595
+ },
+ {
+ "epoch": 49.95021834061135,
+ "grad_norm": 0.017228754237294197,
+ "learning_rate": 0.0006,
+ "loss": 4.964842796325684,
+ "step": 3596
+ },
+ {
+ "epoch": 49.96419213973799,
+ "grad_norm": 0.016021743416786194,
+ "learning_rate": 0.0006,
+ "loss": 4.910758972167969,
+ "step": 3597
+ },
+ {
+ "epoch": 49.97816593886463,
+ "grad_norm": 0.017985103651881218,
+ "learning_rate": 0.0006,
+ "loss": 4.916072368621826,
+ "step": 3598
+ },
+ {
+ "epoch": 49.992139737991266,
+ "grad_norm": 0.018780240789055824,
+ "learning_rate": 0.0006,
+ "loss": 4.94658088684082,
+ "step": 3599
+ },
+ {
+ "epoch": 50.0,
+ "grad_norm": 0.01776314526796341,
+ "learning_rate": 0.0006,
+ "loss": 4.969357967376709,
+ "step": 3600
+ },
+ {
+ "epoch": 50.0,
+ "eval_loss": 5.141158580780029,
+ "eval_runtime": 56.2332,
+ "eval_samples_per_second": 43.426,
+ "eval_steps_per_second": 1.369,
+ "step": 3600
+ },
+ {
+ "epoch": 50.01397379912664,
+ "grad_norm": 0.016308696940541267,
+ "learning_rate": 0.0006,
+ "loss": 4.924664497375488,
+ "step": 3601
+ },
+ {
+ "epoch": 50.02794759825328,
+ "grad_norm": 0.018810199573636055,
+ "learning_rate": 0.0006,
+ "loss": 4.82293701171875,
+ "step": 3602
+ },
+ {
+ "epoch": 50.041921397379916,
+ "grad_norm": 0.02227160707116127,
+ "learning_rate": 0.0006,
+ "loss": 4.9612298011779785,
+ "step": 3603
+ },
+ {
+ "epoch": 50.05589519650655,
+ "grad_norm": 0.025266462936997414,
+ "learning_rate": 0.0006,
+ "loss": 4.963835716247559,
+ "step": 3604
+ },
+ {
+ "epoch": 50.069868995633186,
+ "grad_norm": 0.028588181361556053,
+ "learning_rate": 0.0006,
+ "loss": 4.992118835449219,
+ "step": 3605
+ },
+ {
+ "epoch": 50.083842794759825,
+ "grad_norm": 0.02739487960934639,
+ "learning_rate": 0.0006,
+ "loss": 4.915547847747803,
+ "step": 3606
+ },
+ {
+ "epoch": 50.09781659388646,
+ "grad_norm": 0.027000300586223602,
+ "learning_rate": 0.0006,
+ "loss": 4.883801460266113,
+ "step": 3607
+ },
+ {
+ "epoch": 50.1117903930131,
+ "grad_norm": 0.023468483239412308,
+ "learning_rate": 0.0006,
+ "loss": 5.006168365478516,
+ "step": 3608
+ },
+ {
+ "epoch": 50.12576419213974,
+ "grad_norm": 0.026753634214401245,
+ "learning_rate": 0.0006,
+ "loss": 4.92526912689209,
+ "step": 3609
+ },
+ {
+ "epoch": 50.13973799126637,
+ "grad_norm": 0.026816733181476593,
+ "learning_rate": 0.0006,
+ "loss": 4.847546100616455,
+ "step": 3610
+ },
+ {
+ "epoch": 50.15371179039301,
+ "grad_norm": 0.027362391352653503,
+ "learning_rate": 0.0006,
+ "loss": 4.862548351287842,
+ "step": 3611
+ },
+ {
+ "epoch": 50.16768558951965,
+ "grad_norm": 0.03450845181941986,
+ "learning_rate": 0.0006,
+ "loss": 4.963202953338623,
+ "step": 3612
+ },
+ {
+ "epoch": 50.18165938864629,
+ "grad_norm": 0.03801408037543297,
+ "learning_rate": 0.0006,
+ "loss": 4.770240306854248,
+ "step": 3613
+ },
+ {
+ "epoch": 50.19563318777293,
+ "grad_norm": 0.03579140082001686,
+ "learning_rate": 0.0006,
+ "loss": 4.797981262207031,
+ "step": 3614
+ },
+ {
+ "epoch": 50.209606986899566,
+ "grad_norm": 0.029393760487437248,
+ "learning_rate": 0.0006,
+ "loss": 4.798581600189209,
+ "step": 3615
+ },
+ {
+ "epoch": 50.223580786026204,
+ "grad_norm": 0.026111142709851265,
+ "learning_rate": 0.0006,
+ "loss": 4.990712642669678,
+ "step": 3616
+ },
+ {
+ "epoch": 50.237554585152836,
+ "grad_norm": 0.02213387005031109,
+ "learning_rate": 0.0006,
+ "loss": 4.865842342376709,
+ "step": 3617
+ },
+ {
+ "epoch": 50.251528384279474,
+ "grad_norm": 0.025450684130191803,
+ "learning_rate": 0.0006,
+ "loss": 4.903746128082275,
+ "step": 3618
+ },
+ {
+ "epoch": 50.26550218340611,
+ "grad_norm": 0.020915117114782333,
+ "learning_rate": 0.0006,
+ "loss": 4.951366424560547,
+ "step": 3619
+ },
+ {
+ "epoch": 50.27947598253275,
+ "grad_norm": 0.020972849801182747,
+ "learning_rate": 0.0006,
+ "loss": 4.783311367034912,
+ "step": 3620
+ },
+ {
+ "epoch": 50.29344978165939,
+ "grad_norm": 0.016956273466348648,
+ "learning_rate": 0.0006,
+ "loss": 4.955767631530762,
+ "step": 3621
+ },
+ {
+ "epoch": 50.30742358078603,
+ "grad_norm": 0.018045252189040184,
+ "learning_rate": 0.0006,
+ "loss": 4.869322776794434,
+ "step": 3622
+ },
+ {
+ "epoch": 50.32139737991266,
+ "grad_norm": 0.018199188634753227,
+ "learning_rate": 0.0006,
+ "loss": 4.95680046081543,
+ "step": 3623
+ },
+ {
+ "epoch": 50.3353711790393,
+ "grad_norm": 0.016688082367181778,
+ "learning_rate": 0.0006,
+ "loss": 4.783535957336426,
+ "step": 3624
+ },
+ {
+ "epoch": 50.34934497816594,
+ "grad_norm": 0.017540352419018745,
+ "learning_rate": 0.0006,
+ "loss": 4.8236083984375,
+ "step": 3625
+ },
+ {
+ "epoch": 50.36331877729258,
+ "grad_norm": 0.018054187297821045,
+ "learning_rate": 0.0006,
+ "loss": 4.764655113220215,
+ "step": 3626
+ },
+ {
+ "epoch": 50.377292576419215,
+ "grad_norm": 0.01741131953895092,
+ "learning_rate": 0.0006,
+ "loss": 4.875613689422607,
+ "step": 3627
+ },
+ {
+ "epoch": 50.391266375545854,
+ "grad_norm": 0.01682269014418125,
+ "learning_rate": 0.0006,
+ "loss": 4.818509101867676,
+ "step": 3628
+ },
+ {
+ "epoch": 50.40524017467249,
+ "grad_norm": 0.016652341932058334,
+ "learning_rate": 0.0006,
+ "loss": 4.974020481109619,
+ "step": 3629
+ },
+ {
+ "epoch": 50.419213973799124,
+ "grad_norm": 0.02035282365977764,
+ "learning_rate": 0.0006,
+ "loss": 4.893037796020508,
+ "step": 3630
+ },
+ {
+ "epoch": 50.43318777292576,
+ "grad_norm": 0.019907061010599136,
+ "learning_rate": 0.0006,
+ "loss": 4.848119735717773,
+ "step": 3631
+ },
+ {
+ "epoch": 50.4471615720524,
+ "grad_norm": 0.023861493915319443,
+ "learning_rate": 0.0006,
+ "loss": 4.819353103637695,
+ "step": 3632
+ },
+ {
+ "epoch": 50.46113537117904,
+ "grad_norm": 0.025254026055336,
+ "learning_rate": 0.0006,
+ "loss": 4.769312858581543,
+ "step": 3633
+ },
+ {
+ "epoch": 50.47510917030568,
+ "grad_norm": 0.026154937222599983,
+ "learning_rate": 0.0006,
+ "loss": 4.7586774826049805,
+ "step": 3634
+ },
+ {
+ "epoch": 50.48908296943232,
+ "grad_norm": 0.026342090219259262,
+ "learning_rate": 0.0006,
+ "loss": 4.861218452453613,
+ "step": 3635
+ },
+ {
+ "epoch": 50.50305676855895,
+ "grad_norm": 0.028169559314846992,
+ "learning_rate": 0.0006,
+ "loss": 4.818380355834961,
+ "step": 3636
+ },
+ {
+ "epoch": 50.51703056768559,
+ "grad_norm": 0.02820264920592308,
+ "learning_rate": 0.0006,
+ "loss": 4.892871856689453,
+ "step": 3637
+ },
+ {
+ "epoch": 50.531004366812226,
+ "grad_norm": 0.03345080465078354,
+ "learning_rate": 0.0006,
+ "loss": 4.811084747314453,
+ "step": 3638
+ },
+ {
+ "epoch": 50.544978165938865,
+ "grad_norm": 0.04005736857652664,
+ "learning_rate": 0.0006,
+ "loss": 4.765410900115967,
+ "step": 3639
+ },
+ {
+ "epoch": 50.5589519650655,
+ "grad_norm": 0.04333870857954025,
+ "learning_rate": 0.0006,
+ "loss": 4.956520080566406,
+ "step": 3640
+ },
+ {
+ "epoch": 50.57292576419214,
+ "grad_norm": 0.04216952994465828,
+ "learning_rate": 0.0006,
+ "loss": 4.793441295623779,
+ "step": 3641
+ },
+ {
+ "epoch": 50.58689956331878,
+ "grad_norm": 0.02936951443552971,
+ "learning_rate": 0.0006,
+ "loss": 4.8447957038879395,
+ "step": 3642
+ },
+ {
+ "epoch": 50.60087336244541,
+ "grad_norm": 0.026937026530504227,
+ "learning_rate": 0.0006,
+ "loss": 4.962215423583984,
+ "step": 3643
+ },
+ {
+ "epoch": 50.61484716157205,
+ "grad_norm": 0.02628890983760357,
+ "learning_rate": 0.0006,
+ "loss": 4.883061408996582,
+ "step": 3644
+ },
+ {
+ "epoch": 50.62882096069869,
+ "grad_norm": 0.025984834879636765,
+ "learning_rate": 0.0006,
+ "loss": 4.768662452697754,
+ "step": 3645
+ },
+ {
+ "epoch": 50.64279475982533,
+ "grad_norm": 0.023998012766242027,
+ "learning_rate": 0.0006,
+ "loss": 4.851325035095215,
+ "step": 3646
+ },
+ {
+ "epoch": 50.65676855895197,
+ "grad_norm": 0.02143184095621109,
+ "learning_rate": 0.0006,
+ "loss": 4.8274688720703125,
+ "step": 3647
+ },
+ {
+ "epoch": 50.670742358078606,
+ "grad_norm": 0.02488900162279606,
+ "learning_rate": 0.0006,
+ "loss": 4.831140518188477,
+ "step": 3648
+ },
+ {
+ "epoch": 50.68471615720524,
+ "grad_norm": 0.02330533228814602,
+ "learning_rate": 0.0006,
+ "loss": 4.850462913513184,
+ "step": 3649
+ },
+ {
+ "epoch": 50.698689956331876,
+ "grad_norm": 0.02121943049132824,
+ "learning_rate": 0.0006,
+ "loss": 4.870253086090088,
+ "step": 3650
+ },
+ {
+ "epoch": 50.712663755458514,
+ "grad_norm": 0.02080935798585415,
+ "learning_rate": 0.0006,
+ "loss": 4.798770904541016,
+ "step": 3651
+ },
+ {
+ "epoch": 50.72663755458515,
+ "grad_norm": 0.019047029316425323,
+ "learning_rate": 0.0006,
+ "loss": 4.8427300453186035,
+ "step": 3652
+ },
+ {
+ "epoch": 50.74061135371179,
+ "grad_norm": 0.01824103854596615,
+ "learning_rate": 0.0006,
+ "loss": 4.820980548858643,
+ "step": 3653
+ },
+ {
+ "epoch": 50.75458515283843,
+ "grad_norm": 0.018129967153072357,
+ "learning_rate": 0.0006,
+ "loss": 4.946732997894287,
+ "step": 3654
+ },
+ {
+ "epoch": 50.76855895196506,
+ "grad_norm": 0.01784815452992916,
+ "learning_rate": 0.0006,
+ "loss": 4.829516410827637,
+ "step": 3655
+ },
+ {
+ "epoch": 50.7825327510917,
+ "grad_norm": 0.020021555945277214,
+ "learning_rate": 0.0006,
+ "loss": 4.780428409576416,
+ "step": 3656
+ },
+ {
+ "epoch": 50.79650655021834,
+ "grad_norm": 0.01863253302872181,
+ "learning_rate": 0.0006,
+ "loss": 4.791567802429199,
+ "step": 3657
+ },
+ {
+ "epoch": 50.81048034934498,
+ "grad_norm": 0.019151683896780014,
+ "learning_rate": 0.0006,
+ "loss": 4.824892997741699,
+ "step": 3658
+ },
+ {
+ "epoch": 50.82445414847162,
+ "grad_norm": 0.01891648955643177,
+ "learning_rate": 0.0006,
+ "loss": 4.789605140686035,
+ "step": 3659
+ },
+ {
+ "epoch": 50.838427947598255,
+ "grad_norm": 0.016019510105252266,
+ "learning_rate": 0.0006,
+ "loss": 4.779555320739746,
+ "step": 3660
+ },
+ {
+ "epoch": 50.852401746724894,
+ "grad_norm": 0.016096491366624832,
+ "learning_rate": 0.0006,
+ "loss": 4.753842353820801,
+ "step": 3661
+ },
+ {
+ "epoch": 50.866375545851525,
+ "grad_norm": 0.01693476364016533,
+ "learning_rate": 0.0006,
+ "loss": 4.768477439880371,
+ "step": 3662
+ },
+ {
+ "epoch": 50.880349344978164,
+ "grad_norm": 0.01749575324356556,
+ "learning_rate": 0.0006,
+ "loss": 4.779143333435059,
+ "step": 3663
+ },
+ {
+ "epoch": 50.8943231441048,
+ "grad_norm": 0.016851043328642845,
+ "learning_rate": 0.0006,
+ "loss": 4.780387878417969,
+ "step": 3664
+ },
+ {
+ "epoch": 50.90829694323144,
+ "grad_norm": 0.019045202061533928,
+ "learning_rate": 0.0006,
+ "loss": 4.769913673400879,
+ "step": 3665
+ },
+ {
+ "epoch": 50.92227074235808,
+ "grad_norm": 0.01954123005270958,
+ "learning_rate": 0.0006,
+ "loss": 4.769009590148926,
+ "step": 3666
+ },
+ {
+ "epoch": 50.93624454148472,
+ "grad_norm": 0.022527096793055534,
+ "learning_rate": 0.0006,
+ "loss": 4.789215087890625,
+ "step": 3667
+ },
+ {
+ "epoch": 50.95021834061135,
+ "grad_norm": 0.02871634066104889,
+ "learning_rate": 0.0006,
+ "loss": 4.749727725982666,
+ "step": 3668
+ },
+ {
+ "epoch": 50.96419213973799,
+ "grad_norm": 0.04045053571462631,
+ "learning_rate": 0.0006,
+ "loss": 4.909038543701172,
+ "step": 3669
+ },
+ {
+ "epoch": 50.97816593886463,
+ "grad_norm": 0.04156879335641861,
+ "learning_rate": 0.0006,
+ "loss": 4.876831531524658,
+ "step": 3670
+ },
+ {
+ "epoch": 50.992139737991266,
+ "grad_norm": 0.02621397376060486,
+ "learning_rate": 0.0006,
+ "loss": 4.797173500061035,
+ "step": 3671
+ },
+ {
+ "epoch": 51.0,
+ "grad_norm": 0.03389597684144974,
+ "learning_rate": 0.0006,
+ "loss": 4.8428850173950195,
+ "step": 3672
+ },
+ {
+ "epoch": 51.0,
+ "eval_loss": 4.989261627197266,
+ "eval_runtime": 56.358,
+ "eval_samples_per_second": 43.33,
+ "eval_steps_per_second": 1.366,
+ "step": 3672
+ },
+ {
+ "epoch": 51.01397379912664,
+ "grad_norm": 0.03216972202062607,
+ "learning_rate": 0.0006,
+ "loss": 4.755102634429932,
+ "step": 3673
+ },
+ {
+ "epoch": 51.02794759825328,
+ "grad_norm": 0.03190374746918678,
+ "learning_rate": 0.0006,
+ "loss": 4.720424652099609,
+ "step": 3674
+ },
+ {
+ "epoch": 51.041921397379916,
+ "grad_norm": 0.031727299094200134,
+ "learning_rate": 0.0006,
+ "loss": 4.803277969360352,
+ "step": 3675
+ },
+ {
+ "epoch": 51.05589519650655,
+ "grad_norm": 0.02473079040646553,
+ "learning_rate": 0.0006,
+ "loss": 4.761087417602539,
+ "step": 3676
+ },
+ {
+ "epoch": 51.069868995633186,
+ "grad_norm": 0.020360209047794342,
+ "learning_rate": 0.0006,
+ "loss": 4.753884792327881,
+ "step": 3677
+ },
+ {
+ "epoch": 51.083842794759825,
+ "grad_norm": 0.02078923024237156,
+ "learning_rate": 0.0006,
+ "loss": 4.647130489349365,
+ "step": 3678
+ },
+ {
+ "epoch": 51.09781659388646,
+ "grad_norm": 0.019387200474739075,
+ "learning_rate": 0.0006,
+ "loss": 4.74135160446167,
+ "step": 3679
+ },
+ {
+ "epoch": 51.1117903930131,
+ "grad_norm": 0.01874949410557747,
+ "learning_rate": 0.0006,
+ "loss": 4.795329570770264,
+ "step": 3680
+ },
+ {
+ "epoch": 51.12576419213974,
+ "grad_norm": 0.019331837072968483,
+ "learning_rate": 0.0006,
+ "loss": 4.897967338562012,
+ "step": 3681
+ },
+ {
+ "epoch": 51.13973799126637,
+ "grad_norm": 0.017066050320863724,
+ "learning_rate": 0.0006,
+ "loss": 4.825351715087891,
+ "step": 3682
+ },
+ {
+ "epoch": 51.15371179039301,
+ "grad_norm": 0.01757810264825821,
+ "learning_rate": 0.0006,
+ "loss": 4.719335079193115,
+ "step": 3683
+ },
+ {
+ "epoch": 51.16768558951965,
+ "grad_norm": 0.018174266442656517,
+ "learning_rate": 0.0006,
+ "loss": 4.732680320739746,
+ "step": 3684
+ },
+ {
+ "epoch": 51.18165938864629,
+ "grad_norm": 0.018497727811336517,
+ "learning_rate": 0.0006,
+ "loss": 4.731050491333008,
+ "step": 3685
+ },
+ {
+ "epoch": 51.19563318777293,
+ "grad_norm": 0.0186338871717453,
+ "learning_rate": 0.0006,
+ "loss": 4.748035430908203,
+ "step": 3686
+ },
+ {
+ "epoch": 51.209606986899566,
+ "grad_norm": 0.018531711772084236,
+ "learning_rate": 0.0006,
+ "loss": 4.9007463455200195,
+ "step": 3687
+ },
+ {
+ "epoch": 51.223580786026204,
+ "grad_norm": 0.018221354112029076,
+ "learning_rate": 0.0006,
+ "loss": 4.84077262878418,
+ "step": 3688
+ },
+ {
+ "epoch": 51.237554585152836,
+ "grad_norm": 0.0203656405210495,
+ "learning_rate": 0.0006,
+ "loss": 4.73009729385376,
+ "step": 3689
+ },
+ {
+ "epoch": 51.251528384279474,
+ "grad_norm": 0.01964893378317356,
+ "learning_rate": 0.0006,
+ "loss": 4.772583961486816,
+ "step": 3690
+ },
+ {
+ "epoch": 51.26550218340611,
+ "grad_norm": 0.019606510177254677,
+ "learning_rate": 0.0006,
+ "loss": 4.599812030792236,
+ "step": 3691
+ },
+ {
+ "epoch": 51.27947598253275,
+ "grad_norm": 0.020501255989074707,
+ "learning_rate": 0.0006,
+ "loss": 4.809898853302002,
+ "step": 3692
+ },
+ {
+ "epoch": 51.29344978165939,
+ "grad_norm": 0.01870838552713394,
+ "learning_rate": 0.0006,
+ "loss": 4.741898059844971,
+ "step": 3693
+ },
+ {
+ "epoch": 51.30742358078603,
+ "grad_norm": 0.01802559196949005,
+ "learning_rate": 0.0006,
+ "loss": 4.746562480926514,
+ "step": 3694
+ },
+ {
+ "epoch": 51.32139737991266,
+ "grad_norm": 0.02053048275411129,
+ "learning_rate": 0.0006,
+ "loss": 4.69932746887207,
+ "step": 3695
+ },
+ {
+ "epoch": 51.3353711790393,
+ "grad_norm": 0.02661825157701969,
+ "learning_rate": 0.0006,
+ "loss": 4.68748664855957,
+ "step": 3696
+ },
+ {
+ "epoch": 51.34934497816594,
+ "grad_norm": 0.028269147500395775,
+ "learning_rate": 0.0006,
+ "loss": 4.731565475463867,
+ "step": 3697
+ },
+ {
+ "epoch": 51.36331877729258,
+ "grad_norm": 0.02650964818894863,
+ "learning_rate": 0.0006,
+ "loss": 4.733857154846191,
+ "step": 3698
+ },
+ {
+ "epoch": 51.377292576419215,
+ "grad_norm": 0.023696066811680794,
+ "learning_rate": 0.0006,
+ "loss": 4.812362194061279,
+ "step": 3699
+ },
+ {
+ "epoch": 51.391266375545854,
+ "grad_norm": 0.02365880273282528,
+ "learning_rate": 0.0006,
+ "loss": 4.733426094055176,
+ "step": 3700
+ },
+ {
+ "epoch": 51.40524017467249,
+ "grad_norm": 0.026249447837471962,
+ "learning_rate": 0.0006,
+ "loss": 4.715622901916504,
+ "step": 3701
+ },
+ {
+ "epoch": 51.419213973799124,
+ "grad_norm": 0.03301899880170822,
+ "learning_rate": 0.0006,
+ "loss": 4.7223663330078125,
+ "step": 3702
+ },
+ {
+ "epoch": 51.43318777292576,
+ "grad_norm": 0.044686660170555115,
+ "learning_rate": 0.0006,
+ "loss": 4.8121161460876465,
+ "step": 3703
+ },
+ {
+ "epoch": 51.4471615720524,
+ "grad_norm": 0.04733522981405258,
+ "learning_rate": 0.0006,
+ "loss": 4.841394424438477,
+ "step": 3704
+ },
+ {
+ "epoch": 51.46113537117904,
+ "grad_norm": 0.0395209975540638,
+ "learning_rate": 0.0006,
+ "loss": 4.773837089538574,
+ "step": 3705
+ },
+ {
+ "epoch": 51.47510917030568,
+ "grad_norm": 0.032790958881378174,
+ "learning_rate": 0.0006,
+ "loss": 4.833066463470459,
+ "step": 3706
+ },
+ {
+ "epoch": 51.48908296943232,
+ "grad_norm": 0.030578427016735077,
+ "learning_rate": 0.0006,
+ "loss": 4.685190200805664,
+ "step": 3707
+ },
+ {
+ "epoch": 51.50305676855895,
+ "grad_norm": 0.031619083136320114,
+ "learning_rate": 0.0006,
+ "loss": 4.740378379821777,
+ "step": 3708
+ },
+ {
+ "epoch": 51.51703056768559,
+ "grad_norm": 0.02385825850069523,
+ "learning_rate": 0.0006,
+ "loss": 4.812911510467529,
+ "step": 3709
+ },
+ {
+ "epoch": 51.531004366812226,
+ "grad_norm": 0.023523425683379173,
+ "learning_rate": 0.0006,
+ "loss": 4.669580459594727,
+ "step": 3710
+ },
+ {
+ "epoch": 51.544978165938865,
+ "grad_norm": 0.021316280588507652,
+ "learning_rate": 0.0006,
+ "loss": 4.76888370513916,
+ "step": 3711
+ },
+ {
+ "epoch": 51.5589519650655,
+ "grad_norm": 0.024663880467414856,
+ "learning_rate": 0.0006,
+ "loss": 4.702592372894287,
+ "step": 3712
+ },
+ {
+ "epoch": 51.57292576419214,
+ "grad_norm": 0.024173280224204063,
+ "learning_rate": 0.0006,
+ "loss": 4.710604667663574,
+ "step": 3713
+ },
+ {
+ "epoch": 51.58689956331878,
+ "grad_norm": 0.02754114381968975,
+ "learning_rate": 0.0006,
+ "loss": 4.746356964111328,
+ "step": 3714
+ },
+ {
+ "epoch": 51.60087336244541,
+ "grad_norm": 0.027903622016310692,
+ "learning_rate": 0.0006,
+ "loss": 4.7546491622924805,
+ "step": 3715
+ },
+ {
+ "epoch": 51.61484716157205,
+ "grad_norm": 0.024871617555618286,
+ "learning_rate": 0.0006,
+ "loss": 4.605746269226074,
+ "step": 3716
+ },
+ {
+ "epoch": 51.62882096069869,
+ "grad_norm": 0.021239031106233597,
+ "learning_rate": 0.0006,
+ "loss": 4.6942458152771,
+ "step": 3717
+ },
+ {
+ "epoch": 51.64279475982533,
+ "grad_norm": 0.01925622671842575,
+ "learning_rate": 0.0006,
+ "loss": 4.593558311462402,
+ "step": 3718
+ },
+ {
+ "epoch": 51.65676855895197,
+ "grad_norm": 0.01956283673644066,
+ "learning_rate": 0.0006,
+ "loss": 4.631172180175781,
+ "step": 3719
+ },
+ {
+ "epoch": 51.670742358078606,
+ "grad_norm": 0.02269871160387993,
+ "learning_rate": 0.0006,
+ "loss": 4.729236602783203,
+ "step": 3720
+ },
+ {
+ "epoch": 51.68471615720524,
+ "grad_norm": 0.02253916673362255,
+ "learning_rate": 0.0006,
+ "loss": 4.7191033363342285,
+ "step": 3721
+ },
+ {
+ "epoch": 51.698689956331876,
+ "grad_norm": 0.026231050491333008,
+ "learning_rate": 0.0006,
+ "loss": 4.811850547790527,
+ "step": 3722
+ },
+ {
+ "epoch": 51.712663755458514,
+ "grad_norm": 0.02558821812272072,
+ "learning_rate": 0.0006,
+ "loss": 4.681466102600098,
+ "step": 3723
+ },
+ {
+ "epoch": 51.72663755458515,
+ "grad_norm": 0.02124209702014923,
+ "learning_rate": 0.0006,
+ "loss": 4.693664073944092,
+ "step": 3724
+ },
+ {
+ "epoch": 51.74061135371179,
+ "grad_norm": 0.016638319939374924,
+ "learning_rate": 0.0006,
+ "loss": 4.771440029144287,
+ "step": 3725
+ },
+ {
+ "epoch": 51.75458515283843,
+ "grad_norm": 0.018048347905278206,
+ "learning_rate": 0.0006,
+ "loss": 4.643252372741699,
+ "step": 3726
+ },
+ {
+ "epoch": 51.76855895196506,
+ "grad_norm": 0.019904717803001404,
+ "learning_rate": 0.0006,
+ "loss": 4.66530704498291,
+ "step": 3727
+ },
+ {
+ "epoch": 51.7825327510917,
+ "grad_norm": 0.0194614976644516,
+ "learning_rate": 0.0006,
+ "loss": 4.711943626403809,
+ "step": 3728
+ },
+ {
+ "epoch": 51.79650655021834,
+ "grad_norm": 0.019175555557012558,
+ "learning_rate": 0.0006,
+ "loss": 4.721931457519531,
+ "step": 3729
+ },
+ {
+ "epoch": 51.81048034934498,
+ "grad_norm": 0.02093447372317314,
+ "learning_rate": 0.0006,
+ "loss": 4.633059024810791,
+ "step": 3730
+ },
+ {
+ "epoch": 51.82445414847162,
+ "grad_norm": 0.01930283196270466,
+ "learning_rate": 0.0006,
+ "loss": 4.659373760223389,
+ "step": 3731
+ },
+ {
+ "epoch": 51.838427947598255,
+ "grad_norm": 0.015350676141679287,
+ "learning_rate": 0.0006,
+ "loss": 4.763708114624023,
+ "step": 3732
+ },
+ {
+ "epoch": 51.852401746724894,
+ "grad_norm": 0.014139448292553425,
+ "learning_rate": 0.0006,
+ "loss": 4.753891468048096,
+ "step": 3733
+ },
+ {
+ "epoch": 51.866375545851525,
+ "grad_norm": 0.016215872019529343,
+ "learning_rate": 0.0006,
+ "loss": 4.658967018127441,
+ "step": 3734
+ },
+ {
+ "epoch": 51.880349344978164,
+ "grad_norm": 0.017328619956970215,
+ "learning_rate": 0.0006,
+ "loss": 4.696579933166504,
+ "step": 3735
+ },
+ {
+ "epoch": 51.8943231441048,
+ "grad_norm": 0.018103396520018578,
+ "learning_rate": 0.0006,
+ "loss": 4.627464771270752,
+ "step": 3736
+ },
+ {
+ "epoch": 51.90829694323144,
+ "grad_norm": 0.01715235412120819,
+ "learning_rate": 0.0006,
+ "loss": 4.7368011474609375,
+ "step": 3737
+ },
+ {
+ "epoch": 51.92227074235808,
+ "grad_norm": 0.014790094457566738,
+ "learning_rate": 0.0006,
+ "loss": 4.629441738128662,
+ "step": 3738
+ },
+ {
+ "epoch": 51.93624454148472,
+ "grad_norm": 0.017114296555519104,
+ "learning_rate": 0.0006,
+ "loss": 4.761256694793701,
+ "step": 3739
+ },
+ {
+ "epoch": 51.95021834061135,
+ "grad_norm": 0.019009845331311226,
+ "learning_rate": 0.0006,
+ "loss": 4.789155006408691,
+ "step": 3740
+ },
+ {
+ "epoch": 51.96419213973799,
+ "grad_norm": 0.019696692004799843,
+ "learning_rate": 0.0006,
+ "loss": 4.721269607543945,
+ "step": 3741
+ },
+ {
+ "epoch": 51.97816593886463,
+ "grad_norm": 0.020542822778224945,
+ "learning_rate": 0.0006,
+ "loss": 4.687038898468018,
+ "step": 3742
+ },
+ {
+ "epoch": 51.992139737991266,
+ "grad_norm": 0.021267160773277283,
+ "learning_rate": 0.0006,
+ "loss": 4.747860908508301,
+ "step": 3743
+ },
+ {
+ "epoch": 52.0,
+ "grad_norm": 0.025049567222595215,
+ "learning_rate": 0.0006,
+ "loss": 4.8473100662231445,
+ "step": 3744
+ },
+ {
+ "epoch": 52.0,
+ "eval_loss": 4.943043231964111,
+ "eval_runtime": 56.4929,
+ "eval_samples_per_second": 43.227,
+ "eval_steps_per_second": 1.363,
+ "step": 3744
+ },
+ {
+ "epoch": 52.01397379912664,
+ "grad_norm": 0.03054986707866192,
+ "learning_rate": 0.0006,
+ "loss": 4.856931686401367,
+ "step": 3745
+ },
+ {
+ "epoch": 52.02794759825328,
+ "grad_norm": 0.03838375583291054,
+ "learning_rate": 0.0006,
+ "loss": 4.699807167053223,
+ "step": 3746
+ },
+ {
+ "epoch": 52.041921397379916,
+ "grad_norm": 0.03383488208055496,
+ "learning_rate": 0.0006,
+ "loss": 4.6626787185668945,
+ "step": 3747
+ },
+ {
+ "epoch": 52.05589519650655,
+ "grad_norm": 0.0201561376452446,
+ "learning_rate": 0.0006,
+ "loss": 4.621831893920898,
+ "step": 3748
+ },
+ {
+ "epoch": 52.069868995633186,
+ "grad_norm": 0.028749438002705574,
+ "learning_rate": 0.0006,
+ "loss": 4.682540416717529,
+ "step": 3749
+ },
+ {
+ "epoch": 52.083842794759825,
+ "grad_norm": 0.02681174874305725,
+ "learning_rate": 0.0006,
+ "loss": 4.629997253417969,
+ "step": 3750
+ },
+ {
+ "epoch": 52.09781659388646,
+ "grad_norm": 0.024063318967819214,
+ "learning_rate": 0.0006,
+ "loss": 4.573021411895752,
+ "step": 3751
+ },
+ {
+ "epoch": 52.1117903930131,
+ "grad_norm": 0.02324330434203148,
+ "learning_rate": 0.0006,
+ "loss": 4.739465236663818,
+ "step": 3752
+ },
+ {
+ "epoch": 52.12576419213974,
+ "grad_norm": 0.0225426834076643,
+ "learning_rate": 0.0006,
+ "loss": 4.7335333824157715,
+ "step": 3753
+ },
+ {
+ "epoch": 52.13973799126637,
+ "grad_norm": 0.019593428820371628,
+ "learning_rate": 0.0006,
+ "loss": 4.704958438873291,
+ "step": 3754
+ },
+ {
+ "epoch": 52.15371179039301,
+ "grad_norm": 0.020777082070708275,
+ "learning_rate": 0.0006,
+ "loss": 4.638696670532227,
+ "step": 3755
+ },
+ {
+ "epoch": 52.16768558951965,
+ "grad_norm": 0.021525604650378227,
+ "learning_rate": 0.0006,
+ "loss": 4.643706798553467,
+ "step": 3756
+ },
+ {
+ "epoch": 52.18165938864629,
+ "grad_norm": 0.018943654373288155,
+ "learning_rate": 0.0006,
+ "loss": 4.5173211097717285,
+ "step": 3757
+ },
+ {
+ "epoch": 52.19563318777293,
+ "grad_norm": 0.01856580562889576,
+ "learning_rate": 0.0006,
+ "loss": 4.678788661956787,
+ "step": 3758
+ },
+ {
+ "epoch": 52.209606986899566,
+ "grad_norm": 0.019417259842157364,
+ "learning_rate": 0.0006,
+ "loss": 4.587244987487793,
+ "step": 3759
+ },
+ {
+ "epoch": 52.223580786026204,
+ "grad_norm": 0.01911911740899086,
+ "learning_rate": 0.0006,
+ "loss": 4.688503265380859,
+ "step": 3760
+ },
+ {
+ "epoch": 52.237554585152836,
+ "grad_norm": 0.021017879247665405,
+ "learning_rate": 0.0006,
+ "loss": 4.65829610824585,
+ "step": 3761
+ },
+ {
+ "epoch": 52.251528384279474,
+ "grad_norm": 0.02048875205218792,
+ "learning_rate": 0.0006,
+ "loss": 4.624055862426758,
+ "step": 3762
+ },
+ {
+ "epoch": 52.26550218340611,
+ "grad_norm": 0.01932772807776928,
+ "learning_rate": 0.0006,
+ "loss": 4.556562900543213,
+ "step": 3763
+ },
+ {
+ "epoch": 52.27947598253275,
+ "grad_norm": 0.01871614344418049,
+ "learning_rate": 0.0006,
+ "loss": 4.581201076507568,
+ "step": 3764
+ },
+ {
+ "epoch": 52.29344978165939,
+ "grad_norm": 0.017253106459975243,
+ "learning_rate": 0.0006,
+ "loss": 4.605219841003418,
+ "step": 3765
+ },
+ {
+ "epoch": 52.30742358078603,
+ "grad_norm": 0.01831238530576229,
+ "learning_rate": 0.0006,
+ "loss": 4.759900093078613,
+ "step": 3766
+ },
+ {
+ "epoch": 52.32139737991266,
+ "grad_norm": 0.017946207895874977,
+ "learning_rate": 0.0006,
+ "loss": 4.5598344802856445,
+ "step": 3767
+ },
+ {
+ "epoch": 52.3353711790393,
+ "grad_norm": 0.01680106669664383,
+ "learning_rate": 0.0006,
+ "loss": 4.605034828186035,
+ "step": 3768
+ },
+ {
+ "epoch": 52.34934497816594,
+ "grad_norm": 0.021482540294528008,
+ "learning_rate": 0.0006,
+ "loss": 4.620411396026611,
+ "step": 3769
+ },
+ {
+ "epoch": 52.36331877729258,
+ "grad_norm": 0.02210252732038498,
+ "learning_rate": 0.0006,
+ "loss": 4.667577743530273,
+ "step": 3770
+ },
+ {
+ "epoch": 52.377292576419215,
+ "grad_norm": 0.02237485907971859,
+ "learning_rate": 0.0006,
+ "loss": 4.628812789916992,
+ "step": 3771
+ },
+ {
+ "epoch": 52.391266375545854,
+ "grad_norm": 0.02196541801095009,
+ "learning_rate": 0.0006,
+ "loss": 4.572988510131836,
+ "step": 3772
+ },
+ {
+ "epoch": 52.40524017467249,
+ "grad_norm": 0.01930229924619198,
+ "learning_rate": 0.0006,
+ "loss": 4.588339805603027,
+ "step": 3773
+ },
+ {
+ "epoch": 52.419213973799124,
+ "grad_norm": 0.02222434990108013,
+ "learning_rate": 0.0006,
+ "loss": 4.766482353210449,
+ "step": 3774
+ },
+ {
+ "epoch": 52.43318777292576,
+ "grad_norm": 0.028806306421756744,
+ "learning_rate": 0.0006,
+ "loss": 4.617892742156982,
+ "step": 3775
+ },
+ {
+ "epoch": 52.4471615720524,
+ "grad_norm": 0.03306042402982712,
+ "learning_rate": 0.0006,
+ "loss": 4.665334701538086,
+ "step": 3776
+ },
+ {
+ "epoch": 52.46113537117904,
+ "grad_norm": 0.0358956903219223,
+ "learning_rate": 0.0006,
+ "loss": 4.566668510437012,
+ "step": 3777
+ },
+ {
+ "epoch": 52.47510917030568,
+ "grad_norm": 0.030406497418880463,
+ "learning_rate": 0.0006,
+ "loss": 4.513612270355225,
+ "step": 3778
+ },
+ {
+ "epoch": 52.48908296943232,
+ "grad_norm": 0.030295712873339653,
+ "learning_rate": 0.0006,
+ "loss": 4.640189170837402,
+ "step": 3779
+ },
+ {
+ "epoch": 52.50305676855895,
+ "grad_norm": 0.024981003254652023,
+ "learning_rate": 0.0006,
+ "loss": 4.57558012008667,
+ "step": 3780
+ },
+ {
+ "epoch": 52.51703056768559,
+ "grad_norm": 0.024990713223814964,
+ "learning_rate": 0.0006,
+ "loss": 4.708589553833008,
+ "step": 3781
+ },
+ {
+ "epoch": 52.531004366812226,
+ "grad_norm": 0.025034775957465172,
+ "learning_rate": 0.0006,
+ "loss": 4.70468282699585,
+ "step": 3782
+ },
+ {
+ "epoch": 52.544978165938865,
+ "grad_norm": 0.028243759647011757,
+ "learning_rate": 0.0006,
+ "loss": 4.712060451507568,
+ "step": 3783
+ },
+ {
+ "epoch": 52.5589519650655,
+ "grad_norm": 0.026984544470906258,
+ "learning_rate": 0.0006,
+ "loss": 4.505004405975342,
+ "step": 3784
+ },
+ {
+ "epoch": 52.57292576419214,
+ "grad_norm": 0.050843529403209686,
+ "learning_rate": 0.0006,
+ "loss": 4.593271255493164,
+ "step": 3785
+ },
+ {
+ "epoch": 52.58689956331878,
+ "grad_norm": 0.25422802567481995,
+ "learning_rate": 0.0006,
+ "loss": 4.7178802490234375,
+ "step": 3786
+ },
+ {
+ "epoch": 52.60087336244541,
+ "grad_norm": 0.6078406572341919,
+ "learning_rate": 0.0006,
+ "loss": 6.144372940063477,
+ "step": 3787
+ },
+ {
+ "epoch": 52.61484716157205,
+ "grad_norm": 0.25933101773262024,
+ "learning_rate": 0.0006,
+ "loss": 5.971083641052246,
+ "step": 3788
+ },
+ {
+ "epoch": 52.62882096069869,
+ "grad_norm": 0.294657826423645,
+ "learning_rate": 0.0006,
+ "loss": 6.204090118408203,
+ "step": 3789
+ },
+ {
+ "epoch": 52.64279475982533,
+ "grad_norm": 0.23641522228717804,
+ "learning_rate": 0.0006,
+ "loss": 6.869574546813965,
+ "step": 3790
+ },
+ {
+ "epoch": 52.65676855895197,
+ "grad_norm": 0.23717311024665833,
+ "learning_rate": 0.0006,
+ "loss": 6.739034175872803,
+ "step": 3791
+ },
+ {
+ "epoch": 52.670742358078606,
+ "grad_norm": 0.10300362855195999,
+ "learning_rate": 0.0006,
+ "loss": 6.49191951751709,
+ "step": 3792
+ },
+ {
+ "epoch": 52.68471615720524,
+ "grad_norm": 0.1484968513250351,
+ "learning_rate": 0.0006,
+ "loss": 6.6021270751953125,
+ "step": 3793
+ },
+ {
+ "epoch": 52.698689956331876,
+ "grad_norm": 0.08052115142345428,
+ "learning_rate": 0.0006,
+ "loss": 6.432796478271484,
+ "step": 3794
+ },
+ {
+ "epoch": 52.712663755458514,
+ "grad_norm": 0.06869253516197205,
+ "learning_rate": 0.0006,
+ "loss": 6.212784767150879,
+ "step": 3795
+ },
+ {
+ "epoch": 52.72663755458515,
+ "grad_norm": 0.05589582398533821,
+ "learning_rate": 0.0006,
+ "loss": 6.168288707733154,
+ "step": 3796
+ },
+ {
+ "epoch": 52.74061135371179,
+ "grad_norm": 0.04840940609574318,
+ "learning_rate": 0.0006,
+ "loss": 6.163174629211426,
+ "step": 3797
+ },
+ {
+ "epoch": 52.75458515283843,
+ "grad_norm": 0.04578601196408272,
+ "learning_rate": 0.0006,
+ "loss": 6.038390636444092,
+ "step": 3798
+ },
+ {
+ "epoch": 52.76855895196506,
+ "grad_norm": 0.04102053865790367,
+ "learning_rate": 0.0006,
+ "loss": 5.929539680480957,
+ "step": 3799
+ },
+ {
+ "epoch": 52.7825327510917,
+ "grad_norm": 0.038398485630750656,
+ "learning_rate": 0.0006,
+ "loss": 6.007378101348877,
+ "step": 3800
+ },
+ {
+ "epoch": 52.79650655021834,
+ "grad_norm": 0.03466642275452614,
+ "learning_rate": 0.0006,
+ "loss": 5.767291069030762,
+ "step": 3801
+ },
+ {
+ "epoch": 52.81048034934498,
+ "grad_norm": 0.03328905999660492,
+ "learning_rate": 0.0006,
+ "loss": 5.785032272338867,
+ "step": 3802
+ },
+ {
+ "epoch": 52.82445414847162,
+ "grad_norm": 0.03186681121587753,
+ "learning_rate": 0.0006,
+ "loss": 5.6942057609558105,
+ "step": 3803
+ },
+ {
+ "epoch": 52.838427947598255,
+ "grad_norm": 0.030553551390767097,
+ "learning_rate": 0.0006,
+ "loss": 5.680774688720703,
+ "step": 3804
+ },
+ {
+ "epoch": 52.852401746724894,
+ "grad_norm": 0.02578839473426342,
+ "learning_rate": 0.0006,
+ "loss": 5.703461647033691,
+ "step": 3805
+ },
+ {
+ "epoch": 52.866375545851525,
+ "grad_norm": 0.026261312887072563,
+ "learning_rate": 0.0006,
+ "loss": 5.642549514770508,
+ "step": 3806
+ },
+ {
+ "epoch": 52.880349344978164,
+ "grad_norm": 0.02782430872321129,
+ "learning_rate": 0.0006,
+ "loss": 5.563597679138184,
+ "step": 3807
+ },
+ {
+ "epoch": 52.8943231441048,
+ "grad_norm": 0.024467799812555313,
+ "learning_rate": 0.0006,
+ "loss": 5.583395957946777,
+ "step": 3808
+ },
+ {
+ "epoch": 52.90829694323144,
+ "grad_norm": 0.022602897137403488,
+ "learning_rate": 0.0006,
+ "loss": 5.50742769241333,
+ "step": 3809
+ },
+ {
+ "epoch": 52.92227074235808,
+ "grad_norm": 0.024292029440402985,
+ "learning_rate": 0.0006,
+ "loss": 5.367804527282715,
+ "step": 3810
+ },
+ {
+ "epoch": 52.93624454148472,
+ "grad_norm": 0.021062295883893967,
+ "learning_rate": 0.0006,
+ "loss": 5.423705101013184,
+ "step": 3811
+ },
+ {
+ "epoch": 52.95021834061135,
+ "grad_norm": 0.022226469591259956,
+ "learning_rate": 0.0006,
+ "loss": 5.344039440155029,
+ "step": 3812
+ },
+ {
+ "epoch": 52.96419213973799,
+ "grad_norm": 0.023691752925515175,
+ "learning_rate": 0.0006,
+ "loss": 5.2581892013549805,
+ "step": 3813
+ },
+ {
+ "epoch": 52.97816593886463,
+ "grad_norm": 0.02181999757885933,
+ "learning_rate": 0.0006,
+ "loss": 5.2710113525390625,
+ "step": 3814
+ },
+ {
+ "epoch": 52.992139737991266,
+ "grad_norm": 0.025295857340097427,
+ "learning_rate": 0.0006,
+ "loss": 5.278355598449707,
+ "step": 3815
+ },
+ {
+ "epoch": 53.0,
+ "grad_norm": 0.02216828241944313,
+ "learning_rate": 0.0006,
+ "loss": 5.127945899963379,
+ "step": 3816
+ },
+ {
+ "epoch": 53.0,
+ "eval_loss": 5.3411736488342285,
+ "eval_runtime": 56.2837,
+ "eval_samples_per_second": 43.387,
+ "eval_steps_per_second": 1.368,
+ "step": 3816
+ },
+ {
+ "epoch": 53.01397379912664,
+ "grad_norm": 0.022496026009321213,
+ "learning_rate": 0.0006,
+ "loss": 5.19655704498291,
+ "step": 3817
+ },
+ {
+ "epoch": 53.02794759825328,
+ "grad_norm": 0.03201168775558472,
+ "learning_rate": 0.0006,
+ "loss": 5.094621181488037,
+ "step": 3818
+ },
+ {
+ "epoch": 53.041921397379916,
+ "grad_norm": 0.02683178335428238,
+ "learning_rate": 0.0006,
+ "loss": 5.19899845123291,
+ "step": 3819
+ },
+ {
+ "epoch": 53.05589519650655,
+ "grad_norm": 0.022659169510006905,
+ "learning_rate": 0.0006,
+ "loss": 5.184103012084961,
+ "step": 3820
+ },
+ {
+ "epoch": 53.069868995633186,
+ "grad_norm": 0.024784443899989128,
+ "learning_rate": 0.0006,
+ "loss": 5.124286651611328,
+ "step": 3821
+ },
+ {
+ "epoch": 53.083842794759825,
+ "grad_norm": 0.028710370883345604,
+ "learning_rate": 0.0006,
+ "loss": 5.130426406860352,
+ "step": 3822
+ },
+ {
+ "epoch": 53.09781659388646,
+ "grad_norm": 0.02366088517010212,
+ "learning_rate": 0.0006,
+ "loss": 5.1453399658203125,
+ "step": 3823
+ },
+ {
+ "epoch": 53.1117903930131,
+ "grad_norm": 0.01813938096165657,
+ "learning_rate": 0.0006,
+ "loss": 5.05271053314209,
+ "step": 3824
+ },
+ {
+ "epoch": 53.12576419213974,
+ "grad_norm": 0.02051435597240925,
+ "learning_rate": 0.0006,
+ "loss": 5.023038864135742,
+ "step": 3825
+ },
+ {
+ "epoch": 53.13973799126637,
+ "grad_norm": 0.02185754105448723,
+ "learning_rate": 0.0006,
+ "loss": 4.8950018882751465,
+ "step": 3826
+ },
+ {
+ "epoch": 53.15371179039301,
+ "grad_norm": 0.017223743721842766,
+ "learning_rate": 0.0006,
+ "loss": 5.078621864318848,
+ "step": 3827
+ },
+ {
+ "epoch": 53.16768558951965,
+ "grad_norm": 0.01832217164337635,
+ "learning_rate": 0.0006,
+ "loss": 4.921341419219971,
+ "step": 3828
+ },
+ {
+ "epoch": 53.18165938864629,
+ "grad_norm": 0.02114005573093891,
+ "learning_rate": 0.0006,
+ "loss": 4.997865676879883,
+ "step": 3829
+ },
+ {
+ "epoch": 53.19563318777293,
+ "grad_norm": 0.021426431834697723,
+ "learning_rate": 0.0006,
+ "loss": 4.877785682678223,
+ "step": 3830
+ },
+ {
+ "epoch": 53.209606986899566,
+ "grad_norm": 0.021327560767531395,
+ "learning_rate": 0.0006,
+ "loss": 4.847512245178223,
+ "step": 3831
+ },
+ {
+ "epoch": 53.223580786026204,
+ "grad_norm": 0.027825530618429184,
+ "learning_rate": 0.0006,
+ "loss": 5.004823207855225,
+ "step": 3832
+ },
+ {
+ "epoch": 53.237554585152836,
+ "grad_norm": 0.03756481409072876,
+ "learning_rate": 0.0006,
+ "loss": 4.817290306091309,
+ "step": 3833
+ },
+ {
+ "epoch": 53.251528384279474,
+ "grad_norm": 0.04933721572160721,
+ "learning_rate": 0.0006,
+ "loss": 4.863317489624023,
+ "step": 3834
+ },
+ {
+ "epoch": 53.26550218340611,
+ "grad_norm": 0.04588694870471954,
+ "learning_rate": 0.0006,
+ "loss": 4.864742755889893,
+ "step": 3835
+ },
+ {
+ "epoch": 53.27947598253275,
+ "grad_norm": 0.02728264033794403,
+ "learning_rate": 0.0006,
+ "loss": 4.9068756103515625,
+ "step": 3836
+ },
+ {
+ "epoch": 53.29344978165939,
+ "grad_norm": 0.036240726709365845,
+ "learning_rate": 0.0006,
+ "loss": 4.902045249938965,
+ "step": 3837
+ },
+ {
+ "epoch": 53.30742358078603,
+ "grad_norm": 0.02703195810317993,
+ "learning_rate": 0.0006,
+ "loss": 4.908980369567871,
+ "step": 3838
+ },
+ {
+ "epoch": 53.32139737991266,
+ "grad_norm": 0.024894049391150475,
+ "learning_rate": 0.0006,
+ "loss": 4.794848442077637,
+ "step": 3839
+ },
+ {
+ "epoch": 53.3353711790393,
+ "grad_norm": 0.02389993891119957,
+ "learning_rate": 0.0006,
+ "loss": 4.8234734535217285,
+ "step": 3840
+ },
+ {
+ "epoch": 53.34934497816594,
+ "grad_norm": 0.02486400306224823,
+ "learning_rate": 0.0006,
+ "loss": 4.851463794708252,
+ "step": 3841
+ },
+ {
+ "epoch": 53.36331877729258,
+ "grad_norm": 0.02613813802599907,
+ "learning_rate": 0.0006,
+ "loss": 4.860444068908691,
+ "step": 3842
+ },
+ {
+ "epoch": 53.377292576419215,
+ "grad_norm": 0.0229744054377079,
+ "learning_rate": 0.0006,
+ "loss": 4.709606170654297,
+ "step": 3843
+ },
+ {
+ "epoch": 53.391266375545854,
+ "grad_norm": 0.02123042568564415,
+ "learning_rate": 0.0006,
+ "loss": 4.654613494873047,
+ "step": 3844
+ },
+ {
+ "epoch": 53.40524017467249,
+ "grad_norm": 0.017352338880300522,
+ "learning_rate": 0.0006,
+ "loss": 4.769761085510254,
+ "step": 3845
+ },
+ {
+ "epoch": 53.419213973799124,
+ "grad_norm": 0.019808724522590637,
+ "learning_rate": 0.0006,
+ "loss": 4.98514461517334,
+ "step": 3846
+ },
+ {
+ "epoch": 53.43318777292576,
+ "grad_norm": 0.018776152282953262,
+ "learning_rate": 0.0006,
+ "loss": 4.749990463256836,
+ "step": 3847
+ },
+ {
+ "epoch": 53.4471615720524,
+ "grad_norm": 0.02017521858215332,
+ "learning_rate": 0.0006,
+ "loss": 4.757381439208984,
+ "step": 3848
+ },
+ {
+ "epoch": 53.46113537117904,
+ "grad_norm": 0.019509416073560715,
+ "learning_rate": 0.0006,
+ "loss": 4.8111572265625,
+ "step": 3849
+ },
+ {
+ "epoch": 53.47510917030568,
+ "grad_norm": 0.015543580986559391,
+ "learning_rate": 0.0006,
+ "loss": 4.7501678466796875,
+ "step": 3850
+ },
+ {
+ "epoch": 53.48908296943232,
+ "grad_norm": 0.017941651865839958,
+ "learning_rate": 0.0006,
+ "loss": 4.701838493347168,
+ "step": 3851
+ },
+ {
+ "epoch": 53.50305676855895,
+ "grad_norm": 0.01678212359547615,
+ "learning_rate": 0.0006,
+ "loss": 4.787798881530762,
+ "step": 3852
+ },
+ {
+ "epoch": 53.51703056768559,
+ "grad_norm": 0.019050654023885727,
+ "learning_rate": 0.0006,
+ "loss": 4.754859924316406,
+ "step": 3853
+ },
+ {
+ "epoch": 53.531004366812226,
+ "grad_norm": 0.017406078055500984,
+ "learning_rate": 0.0006,
+ "loss": 4.752772808074951,
+ "step": 3854
+ },
+ {
+ "epoch": 53.544978165938865,
+ "grad_norm": 0.01773344725370407,
+ "learning_rate": 0.0006,
+ "loss": 4.651812553405762,
+ "step": 3855
+ },
+ {
+ "epoch": 53.5589519650655,
+ "grad_norm": 0.017221346497535706,
+ "learning_rate": 0.0006,
+ "loss": 4.792244911193848,
+ "step": 3856
+ },
+ {
+ "epoch": 53.57292576419214,
+ "grad_norm": 0.01722758449614048,
+ "learning_rate": 0.0006,
+ "loss": 4.7902960777282715,
+ "step": 3857
+ },
+ {
+ "epoch": 53.58689956331878,
+ "grad_norm": 0.020445192232728004,
+ "learning_rate": 0.0006,
+ "loss": 4.67392110824585,
+ "step": 3858
+ },
+ {
+ "epoch": 53.60087336244541,
+ "grad_norm": 0.02267114631831646,
+ "learning_rate": 0.0006,
+ "loss": 4.796089172363281,
+ "step": 3859
+ },
+ {
+ "epoch": 53.61484716157205,
+ "grad_norm": 0.027129942551255226,
+ "learning_rate": 0.0006,
+ "loss": 4.774303913116455,
+ "step": 3860
+ },
+ {
+ "epoch": 53.62882096069869,
+ "grad_norm": 0.025454839691519737,
+ "learning_rate": 0.0006,
+ "loss": 4.708813190460205,
+ "step": 3861
+ },
+ {
+ "epoch": 53.64279475982533,
+ "grad_norm": 0.019248757511377335,
+ "learning_rate": 0.0006,
+ "loss": 4.760467529296875,
+ "step": 3862
+ },
+ {
+ "epoch": 53.65676855895197,
+ "grad_norm": 0.01791176199913025,
+ "learning_rate": 0.0006,
+ "loss": 4.789203643798828,
+ "step": 3863
+ },
+ {
+ "epoch": 53.670742358078606,
+ "grad_norm": 0.016306612640619278,
+ "learning_rate": 0.0006,
+ "loss": 4.732605934143066,
+ "step": 3864
+ },
+ {
+ "epoch": 53.68471615720524,
+ "grad_norm": 0.017836976796388626,
+ "learning_rate": 0.0006,
+ "loss": 4.691871643066406,
+ "step": 3865
+ },
+ {
+ "epoch": 53.698689956331876,
+ "grad_norm": 0.015102861449122429,
+ "learning_rate": 0.0006,
+ "loss": 4.724156856536865,
+ "step": 3866
+ },
+ {
+ "epoch": 53.712663755458514,
+ "grad_norm": 0.017458142712712288,
+ "learning_rate": 0.0006,
+ "loss": 4.73524808883667,
+ "step": 3867
+ },
+ {
+ "epoch": 53.72663755458515,
+ "grad_norm": 0.01591932773590088,
+ "learning_rate": 0.0006,
+ "loss": 4.705888748168945,
+ "step": 3868
+ },
+ {
+ "epoch": 53.74061135371179,
+ "grad_norm": 0.01713593676686287,
+ "learning_rate": 0.0006,
+ "loss": 4.644516944885254,
+ "step": 3869
+ },
+ {
+ "epoch": 53.75458515283843,
+ "grad_norm": 0.01905885338783264,
+ "learning_rate": 0.0006,
+ "loss": 4.698416233062744,
+ "step": 3870
+ },
+ {
+ "epoch": 53.76855895196506,
+ "grad_norm": 0.01766437292098999,
+ "learning_rate": 0.0006,
+ "loss": 4.6796555519104,
+ "step": 3871
+ },
+ {
+ "epoch": 53.7825327510917,
+ "grad_norm": 0.01552243810147047,
+ "learning_rate": 0.0006,
+ "loss": 4.78138542175293,
+ "step": 3872
+ },
+ {
+ "epoch": 53.79650655021834,
+ "grad_norm": 0.014191331341862679,
+ "learning_rate": 0.0006,
+ "loss": 4.703588485717773,
+ "step": 3873
+ },
+ {
+ "epoch": 53.81048034934498,
+ "grad_norm": 0.015872279182076454,
+ "learning_rate": 0.0006,
+ "loss": 4.705611228942871,
+ "step": 3874
+ },
+ {
+ "epoch": 53.82445414847162,
+ "grad_norm": 0.01787560060620308,
+ "learning_rate": 0.0006,
+ "loss": 4.8227949142456055,
+ "step": 3875
+ },
+ {
+ "epoch": 53.838427947598255,
+ "grad_norm": 0.0158209390938282,
+ "learning_rate": 0.0006,
+ "loss": 4.645227432250977,
+ "step": 3876
+ },
+ {
+ "epoch": 53.852401746724894,
+ "grad_norm": 0.014362678863108158,
+ "learning_rate": 0.0006,
+ "loss": 4.7225422859191895,
+ "step": 3877
+ },
+ {
+ "epoch": 53.866375545851525,
+ "grad_norm": 0.014899161644279957,
+ "learning_rate": 0.0006,
+ "loss": 4.685914039611816,
+ "step": 3878
+ },
+ {
+ "epoch": 53.880349344978164,
+ "grad_norm": 0.01628505438566208,
+ "learning_rate": 0.0006,
+ "loss": 4.674065589904785,
+ "step": 3879
+ },
+ {
+ "epoch": 53.8943231441048,
+ "grad_norm": 0.01873331144452095,
+ "learning_rate": 0.0006,
+ "loss": 4.811744213104248,
+ "step": 3880
+ },
+ {
+ "epoch": 53.90829694323144,
+ "grad_norm": 0.02229069173336029,
+ "learning_rate": 0.0006,
+ "loss": 4.721514701843262,
+ "step": 3881
+ },
+ {
+ "epoch": 53.92227074235808,
+ "grad_norm": 0.028012916445732117,
+ "learning_rate": 0.0006,
+ "loss": 4.773603439331055,
+ "step": 3882
+ },
+ {
+ "epoch": 53.93624454148472,
+ "grad_norm": 0.036322783678770065,
+ "learning_rate": 0.0006,
+ "loss": 4.787134170532227,
+ "step": 3883
+ },
+ {
+ "epoch": 53.95021834061135,
+ "grad_norm": 0.048149142414331436,
+ "learning_rate": 0.0006,
+ "loss": 4.736196517944336,
+ "step": 3884
+ },
+ {
+ "epoch": 53.96419213973799,
+ "grad_norm": 0.051918916404247284,
+ "learning_rate": 0.0006,
+ "loss": 4.686728000640869,
+ "step": 3885
+ },
+ {
+ "epoch": 53.97816593886463,
+ "grad_norm": 0.0481342189013958,
+ "learning_rate": 0.0006,
+ "loss": 4.733730792999268,
+ "step": 3886
+ },
+ {
+ "epoch": 53.992139737991266,
+ "grad_norm": 0.045424818992614746,
+ "learning_rate": 0.0006,
+ "loss": 4.748723983764648,
+ "step": 3887
+ },
+ {
+ "epoch": 54.0,
+ "grad_norm": 0.04153316840529442,
+ "learning_rate": 0.0006,
+ "loss": 4.715220928192139,
+ "step": 3888
+ },
+ {
+ "epoch": 54.0,
+ "eval_loss": 4.964215278625488,
+ "eval_runtime": 56.7757,
+ "eval_samples_per_second": 43.011,
+ "eval_steps_per_second": 1.356,
+ "step": 3888
+ },
+ {
+ "epoch": 54.01397379912664,
+ "grad_norm": 0.03459750860929489,
+ "learning_rate": 0.0006,
+ "loss": 4.661557674407959,
+ "step": 3889
+ },
+ {
+ "epoch": 54.02794759825328,
+ "grad_norm": 0.025409208610653877,
+ "learning_rate": 0.0006,
+ "loss": 4.7065582275390625,
+ "step": 3890
+ },
+ {
+ "epoch": 54.041921397379916,
+ "grad_norm": 0.03202550485730171,
+ "learning_rate": 0.0006,
+ "loss": 4.751070976257324,
+ "step": 3891
+ },
+ {
+ "epoch": 54.05589519650655,
+ "grad_norm": 0.02968018501996994,
+ "learning_rate": 0.0006,
+ "loss": 4.635388374328613,
+ "step": 3892
+ },
+ {
+ "epoch": 54.069868995633186,
+ "grad_norm": 0.02757406234741211,
+ "learning_rate": 0.0006,
+ "loss": 4.706214427947998,
+ "step": 3893
+ },
+ {
+ "epoch": 54.083842794759825,
+ "grad_norm": 0.025223620235919952,
+ "learning_rate": 0.0006,
+ "loss": 4.703792572021484,
+ "step": 3894
+ },
+ {
+ "epoch": 54.09781659388646,
+ "grad_norm": 0.02431778982281685,
+ "learning_rate": 0.0006,
+ "loss": 4.658116340637207,
+ "step": 3895
+ },
+ {
+ "epoch": 54.1117903930131,
+ "grad_norm": 0.02792172320187092,
+ "learning_rate": 0.0006,
+ "loss": 4.638895034790039,
+ "step": 3896
+ },
+ {
+ "epoch": 54.12576419213974,
+ "grad_norm": 0.032253511250019073,
+ "learning_rate": 0.0006,
+ "loss": 4.693902492523193,
+ "step": 3897
+ },
+ {
+ "epoch": 54.13973799126637,
+ "grad_norm": 0.027044931426644325,
+ "learning_rate": 0.0006,
+ "loss": 4.575805187225342,
+ "step": 3898
+ },
+ {
+ "epoch": 54.15371179039301,
+ "grad_norm": 0.023270361125469208,
+ "learning_rate": 0.0006,
+ "loss": 4.67645263671875,
+ "step": 3899
+ },
+ {
+ "epoch": 54.16768558951965,
+ "grad_norm": 0.01929079182446003,
+ "learning_rate": 0.0006,
+ "loss": 4.68062686920166,
+ "step": 3900
+ },
+ {
+ "epoch": 54.18165938864629,
+ "grad_norm": 0.020343473181128502,
+ "learning_rate": 0.0006,
+ "loss": 4.708086967468262,
+ "step": 3901
+ },
+ {
+ "epoch": 54.19563318777293,
+ "grad_norm": 0.0211793202906847,
+ "learning_rate": 0.0006,
+ "loss": 4.677323341369629,
+ "step": 3902
+ },
+ {
+ "epoch": 54.209606986899566,
+ "grad_norm": 0.020341217517852783,
+ "learning_rate": 0.0006,
+ "loss": 4.760937690734863,
+ "step": 3903
+ },
+ {
+ "epoch": 54.223580786026204,
+ "grad_norm": 0.019421054050326347,
+ "learning_rate": 0.0006,
+ "loss": 4.508834362030029,
+ "step": 3904
+ },
+ {
+ "epoch": 54.237554585152836,
+ "grad_norm": 0.016701679676771164,
+ "learning_rate": 0.0006,
+ "loss": 4.664399147033691,
+ "step": 3905
+ },
+ {
+ "epoch": 54.251528384279474,
+ "grad_norm": 0.016561470925807953,
+ "learning_rate": 0.0006,
+ "loss": 4.742990493774414,
+ "step": 3906
+ },
+ {
+ "epoch": 54.26550218340611,
+ "grad_norm": 0.017446085810661316,
+ "learning_rate": 0.0006,
+ "loss": 4.63038444519043,
+ "step": 3907
+ },
+ {
+ "epoch": 54.27947598253275,
+ "grad_norm": 0.014340086840093136,
+ "learning_rate": 0.0006,
+ "loss": 4.688791275024414,
+ "step": 3908
+ },
+ {
+ "epoch": 54.29344978165939,
+ "grad_norm": 0.013580698519945145,
+ "learning_rate": 0.0006,
+ "loss": 4.625302314758301,
+ "step": 3909
+ },
+ {
+ "epoch": 54.30742358078603,
+ "grad_norm": 0.014931539073586464,
+ "learning_rate": 0.0006,
+ "loss": 4.6167497634887695,
+ "step": 3910
+ },
+ {
+ "epoch": 54.32139737991266,
+ "grad_norm": 0.013492150232195854,
+ "learning_rate": 0.0006,
+ "loss": 4.657521724700928,
+ "step": 3911
+ },
+ {
+ "epoch": 54.3353711790393,
+ "grad_norm": 0.015151413157582283,
+ "learning_rate": 0.0006,
+ "loss": 4.634521484375,
+ "step": 3912
+ },
+ {
+ "epoch": 54.34934497816594,
+ "grad_norm": 0.01302040833979845,
+ "learning_rate": 0.0006,
+ "loss": 4.718928337097168,
+ "step": 3913
+ },
+ {
+ "epoch": 54.36331877729258,
+ "grad_norm": 0.013433529995381832,
+ "learning_rate": 0.0006,
+ "loss": 4.584803581237793,
+ "step": 3914
+ },
+ {
+ "epoch": 54.377292576419215,
+ "grad_norm": 0.012987039051949978,
+ "learning_rate": 0.0006,
+ "loss": 4.6393585205078125,
+ "step": 3915
+ },
+ {
+ "epoch": 54.391266375545854,
+ "grad_norm": 0.013695075176656246,
+ "learning_rate": 0.0006,
+ "loss": 4.556313514709473,
+ "step": 3916
+ },
+ {
+ "epoch": 54.40524017467249,
+ "grad_norm": 0.013718255795538425,
+ "learning_rate": 0.0006,
+ "loss": 4.576223850250244,
+ "step": 3917
+ },
+ {
+ "epoch": 54.419213973799124,
+ "grad_norm": 0.013699112460017204,
+ "learning_rate": 0.0006,
+ "loss": 4.63889217376709,
+ "step": 3918
+ },
+ {
+ "epoch": 54.43318777292576,
+ "grad_norm": 0.013516264036297798,
+ "learning_rate": 0.0006,
+ "loss": 4.669532299041748,
+ "step": 3919
+ },
+ {
+ "epoch": 54.4471615720524,
+ "grad_norm": 0.013597294688224792,
+ "learning_rate": 0.0006,
+ "loss": 4.6644287109375,
+ "step": 3920
+ },
+ {
+ "epoch": 54.46113537117904,
+ "grad_norm": 0.013185849413275719,
+ "learning_rate": 0.0006,
+ "loss": 4.636900901794434,
+ "step": 3921
+ },
+ {
+ "epoch": 54.47510917030568,
+ "grad_norm": 0.014311990700662136,
+ "learning_rate": 0.0006,
+ "loss": 4.6047587394714355,
+ "step": 3922
+ },
+ {
+ "epoch": 54.48908296943232,
+ "grad_norm": 0.012426027096807957,
+ "learning_rate": 0.0006,
+ "loss": 4.6161394119262695,
+ "step": 3923
+ },
+ {
+ "epoch": 54.50305676855895,
+ "grad_norm": 0.01283493172377348,
+ "learning_rate": 0.0006,
+ "loss": 4.584190368652344,
+ "step": 3924
+ },
+ {
+ "epoch": 54.51703056768559,
+ "grad_norm": 0.01339943427592516,
+ "learning_rate": 0.0006,
+ "loss": 4.624184608459473,
+ "step": 3925
+ },
+ {
+ "epoch": 54.531004366812226,
+ "grad_norm": 0.012859724462032318,
+ "learning_rate": 0.0006,
+ "loss": 4.6174635887146,
+ "step": 3926
+ },
+ {
+ "epoch": 54.544978165938865,
+ "grad_norm": 0.014420023187994957,
+ "learning_rate": 0.0006,
+ "loss": 4.536544322967529,
+ "step": 3927
+ },
+ {
+ "epoch": 54.5589519650655,
+ "grad_norm": 0.014237185008823872,
+ "learning_rate": 0.0006,
+ "loss": 4.681514739990234,
+ "step": 3928
+ },
+ {
+ "epoch": 54.57292576419214,
+ "grad_norm": 0.013340328820049763,
+ "learning_rate": 0.0006,
+ "loss": 4.579689979553223,
+ "step": 3929
+ },
+ {
+ "epoch": 54.58689956331878,
+ "grad_norm": 0.013894530944526196,
+ "learning_rate": 0.0006,
+ "loss": 4.638958930969238,
+ "step": 3930
+ },
+ {
+ "epoch": 54.60087336244541,
+ "grad_norm": 0.016181141138076782,
+ "learning_rate": 0.0006,
+ "loss": 4.573508262634277,
+ "step": 3931
+ },
+ {
+ "epoch": 54.61484716157205,
+ "grad_norm": 0.013221992179751396,
+ "learning_rate": 0.0006,
+ "loss": 4.6229047775268555,
+ "step": 3932
+ },
+ {
+ "epoch": 54.62882096069869,
+ "grad_norm": 0.013241561129689217,
+ "learning_rate": 0.0006,
+ "loss": 4.641152858734131,
+ "step": 3933
+ },
+ {
+ "epoch": 54.64279475982533,
+ "grad_norm": 0.016237538307905197,
+ "learning_rate": 0.0006,
+ "loss": 4.6421217918396,
+ "step": 3934
+ },
+ {
+ "epoch": 54.65676855895197,
+ "grad_norm": 0.020866507664322853,
+ "learning_rate": 0.0006,
+ "loss": 4.646058082580566,
+ "step": 3935
+ },
+ {
+ "epoch": 54.670742358078606,
+ "grad_norm": 0.022182893007993698,
+ "learning_rate": 0.0006,
+ "loss": 4.554501533508301,
+ "step": 3936
+ },
+ {
+ "epoch": 54.68471615720524,
+ "grad_norm": 0.02254868485033512,
+ "learning_rate": 0.0006,
+ "loss": 4.563869953155518,
+ "step": 3937
+ },
+ {
+ "epoch": 54.698689956331876,
+ "grad_norm": 0.019730940461158752,
+ "learning_rate": 0.0006,
+ "loss": 4.592165946960449,
+ "step": 3938
+ },
+ {
+ "epoch": 54.712663755458514,
+ "grad_norm": 0.015146561898291111,
+ "learning_rate": 0.0006,
+ "loss": 4.6228742599487305,
+ "step": 3939
+ },
+ {
+ "epoch": 54.72663755458515,
+ "grad_norm": 0.01824716106057167,
+ "learning_rate": 0.0006,
+ "loss": 4.716236114501953,
+ "step": 3940
+ },
+ {
+ "epoch": 54.74061135371179,
+ "grad_norm": 0.017730550840497017,
+ "learning_rate": 0.0006,
+ "loss": 4.650232315063477,
+ "step": 3941
+ },
+ {
+ "epoch": 54.75458515283843,
+ "grad_norm": 0.014875595457851887,
+ "learning_rate": 0.0006,
+ "loss": 4.6377458572387695,
+ "step": 3942
+ },
+ {
+ "epoch": 54.76855895196506,
+ "grad_norm": 0.01613209955394268,
+ "learning_rate": 0.0006,
+ "loss": 4.5768537521362305,
+ "step": 3943
+ },
+ {
+ "epoch": 54.7825327510917,
+ "grad_norm": 0.017311280593276024,
+ "learning_rate": 0.0006,
+ "loss": 4.625744342803955,
+ "step": 3944
+ },
+ {
+ "epoch": 54.79650655021834,
+ "grad_norm": 0.018770085647702217,
+ "learning_rate": 0.0006,
+ "loss": 4.613787651062012,
+ "step": 3945
+ },
+ {
+ "epoch": 54.81048034934498,
+ "grad_norm": 0.0165257528424263,
+ "learning_rate": 0.0006,
+ "loss": 4.5991315841674805,
+ "step": 3946
+ },
+ {
+ "epoch": 54.82445414847162,
+ "grad_norm": 0.017433857545256615,
+ "learning_rate": 0.0006,
+ "loss": 4.513872146606445,
+ "step": 3947
+ },
+ {
+ "epoch": 54.838427947598255,
+ "grad_norm": 0.020126905292272568,
+ "learning_rate": 0.0006,
+ "loss": 4.616185188293457,
+ "step": 3948
+ },
+ {
+ "epoch": 54.852401746724894,
+ "grad_norm": 0.01819358579814434,
+ "learning_rate": 0.0006,
+ "loss": 4.535691261291504,
+ "step": 3949
+ },
+ {
+ "epoch": 54.866375545851525,
+ "grad_norm": 0.01899927295744419,
+ "learning_rate": 0.0006,
+ "loss": 4.595199108123779,
+ "step": 3950
+ },
+ {
+ "epoch": 54.880349344978164,
+ "grad_norm": 0.02005566656589508,
+ "learning_rate": 0.0006,
+ "loss": 4.633289337158203,
+ "step": 3951
+ },
+ {
+ "epoch": 54.8943231441048,
+ "grad_norm": 0.021570220589637756,
+ "learning_rate": 0.0006,
+ "loss": 4.592503547668457,
+ "step": 3952
+ },
+ {
+ "epoch": 54.90829694323144,
+ "grad_norm": 0.01883506216108799,
+ "learning_rate": 0.0006,
+ "loss": 4.6597795486450195,
+ "step": 3953
+ },
+ {
+ "epoch": 54.92227074235808,
+ "grad_norm": 0.018070021644234657,
+ "learning_rate": 0.0006,
+ "loss": 4.601474285125732,
+ "step": 3954
+ },
+ {
+ "epoch": 54.93624454148472,
+ "grad_norm": 0.018090378493070602,
+ "learning_rate": 0.0006,
+ "loss": 4.569846153259277,
+ "step": 3955
+ },
+ {
+ "epoch": 54.95021834061135,
+ "grad_norm": 0.018751563504338264,
+ "learning_rate": 0.0006,
+ "loss": 4.646064281463623,
+ "step": 3956
+ },
+ {
+ "epoch": 54.96419213973799,
+ "grad_norm": 0.017512254416942596,
+ "learning_rate": 0.0006,
+ "loss": 4.606670379638672,
+ "step": 3957
+ },
+ {
+ "epoch": 54.97816593886463,
+ "grad_norm": 0.018771110102534294,
+ "learning_rate": 0.0006,
+ "loss": 4.604315757751465,
+ "step": 3958
+ },
+ {
+ "epoch": 54.992139737991266,
+ "grad_norm": 0.01611141301691532,
+ "learning_rate": 0.0006,
+ "loss": 4.643245220184326,
+ "step": 3959
+ },
+ {
+ "epoch": 55.0,
+ "grad_norm": 0.0156728383153677,
+ "learning_rate": 0.0006,
+ "loss": 4.6739044189453125,
+ "step": 3960
+ },
+ {
+ "epoch": 55.0,
+ "eval_loss": 4.851032257080078,
+ "eval_runtime": 56.8699,
+ "eval_samples_per_second": 42.94,
+ "eval_steps_per_second": 1.354,
+ "step": 3960
+ },
+ {
+ "epoch": 55.01397379912664,
+ "grad_norm": 0.016214672476053238,
+ "learning_rate": 0.0006,
+ "loss": 4.571749687194824,
+ "step": 3961
+ },
+ {
+ "epoch": 55.02794759825328,
+ "grad_norm": 0.01649155467748642,
+ "learning_rate": 0.0006,
+ "loss": 4.553592205047607,
+ "step": 3962
+ },
+ {
+ "epoch": 55.041921397379916,
+ "grad_norm": 0.016793379560112953,
+ "learning_rate": 0.0006,
+ "loss": 4.650628566741943,
+ "step": 3963
+ },
+ {
+ "epoch": 55.05589519650655,
+ "grad_norm": 0.0191937405616045,
+ "learning_rate": 0.0006,
+ "loss": 4.620532989501953,
+ "step": 3964
+ },
+ {
+ "epoch": 55.069868995633186,
+ "grad_norm": 0.020976556465029716,
+ "learning_rate": 0.0006,
+ "loss": 4.6329545974731445,
+ "step": 3965
+ },
+ {
+ "epoch": 55.083842794759825,
+ "grad_norm": 0.023450477048754692,
+ "learning_rate": 0.0006,
+ "loss": 4.60828971862793,
+ "step": 3966
+ },
+ {
+ "epoch": 55.09781659388646,
+ "grad_norm": 0.026778312399983406,
+ "learning_rate": 0.0006,
+ "loss": 4.617481708526611,
+ "step": 3967
+ },
+ {
+ "epoch": 55.1117903930131,
+ "grad_norm": 0.02333568036556244,
+ "learning_rate": 0.0006,
+ "loss": 4.580722332000732,
+ "step": 3968
+ },
+ {
+ "epoch": 55.12576419213974,
+ "grad_norm": 0.018491310998797417,
+ "learning_rate": 0.0006,
+ "loss": 4.590953350067139,
+ "step": 3969
+ },
+ {
+ "epoch": 55.13973799126637,
+ "grad_norm": 0.020339803770184517,
+ "learning_rate": 0.0006,
+ "loss": 4.518374443054199,
+ "step": 3970
+ },
+ {
+ "epoch": 55.15371179039301,
+ "grad_norm": 0.01759706623852253,
+ "learning_rate": 0.0006,
+ "loss": 4.560522079467773,
+ "step": 3971
+ },
+ {
+ "epoch": 55.16768558951965,
+ "grad_norm": 0.015875980257987976,
+ "learning_rate": 0.0006,
+ "loss": 4.4795708656311035,
+ "step": 3972
+ },
+ {
+ "epoch": 55.18165938864629,
+ "grad_norm": 0.01650134287774563,
+ "learning_rate": 0.0006,
+ "loss": 4.57435941696167,
+ "step": 3973
+ },
+ {
+ "epoch": 55.19563318777293,
+ "grad_norm": 0.01734331250190735,
+ "learning_rate": 0.0006,
+ "loss": 4.661007881164551,
+ "step": 3974
+ },
+ {
+ "epoch": 55.209606986899566,
+ "grad_norm": 0.01651303470134735,
+ "learning_rate": 0.0006,
+ "loss": 4.511148452758789,
+ "step": 3975
+ },
+ {
+ "epoch": 55.223580786026204,
+ "grad_norm": 0.016189761459827423,
+ "learning_rate": 0.0006,
+ "loss": 4.592350006103516,
+ "step": 3976
+ },
+ {
+ "epoch": 55.237554585152836,
+ "grad_norm": 0.01833222061395645,
+ "learning_rate": 0.0006,
+ "loss": 4.506649971008301,
+ "step": 3977
+ },
+ {
+ "epoch": 55.251528384279474,
+ "grad_norm": 0.02433410845696926,
+ "learning_rate": 0.0006,
+ "loss": 4.49500036239624,
+ "step": 3978
+ },
+ {
+ "epoch": 55.26550218340611,
+ "grad_norm": 0.0314289927482605,
+ "learning_rate": 0.0006,
+ "loss": 4.507624626159668,
+ "step": 3979
+ },
+ {
+ "epoch": 55.27947598253275,
+ "grad_norm": 0.03013679012656212,
+ "learning_rate": 0.0006,
+ "loss": 4.533641815185547,
+ "step": 3980
+ },
+ {
+ "epoch": 55.29344978165939,
+ "grad_norm": 0.024524061009287834,
+ "learning_rate": 0.0006,
+ "loss": 4.5721116065979,
+ "step": 3981
+ },
+ {
+ "epoch": 55.30742358078603,
+ "grad_norm": 0.02640336938202381,
+ "learning_rate": 0.0006,
+ "loss": 4.652521133422852,
+ "step": 3982
+ },
+ {
+ "epoch": 55.32139737991266,
+ "grad_norm": 0.021309101954102516,
+ "learning_rate": 0.0006,
+ "loss": 4.600375175476074,
+ "step": 3983
+ },
+ {
+ "epoch": 55.3353711790393,
+ "grad_norm": 0.023008549585938454,
+ "learning_rate": 0.0006,
+ "loss": 4.53914737701416,
+ "step": 3984
+ },
+ {
+ "epoch": 55.34934497816594,
+ "grad_norm": 0.022491468116641045,
+ "learning_rate": 0.0006,
+ "loss": 4.661163330078125,
+ "step": 3985
+ },
+ {
+ "epoch": 55.36331877729258,
+ "grad_norm": 0.01904081553220749,
+ "learning_rate": 0.0006,
+ "loss": 4.5426177978515625,
+ "step": 3986
+ },
+ {
+ "epoch": 55.377292576419215,
+ "grad_norm": 0.019229518249630928,
+ "learning_rate": 0.0006,
+ "loss": 4.606066703796387,
+ "step": 3987
+ },
+ {
+ "epoch": 55.391266375545854,
+ "grad_norm": 0.01781383901834488,
+ "learning_rate": 0.0006,
+ "loss": 4.54642391204834,
+ "step": 3988
+ },
+ {
+ "epoch": 55.40524017467249,
+ "grad_norm": 0.0174600500613451,
+ "learning_rate": 0.0006,
+ "loss": 4.711891174316406,
+ "step": 3989
+ },
+ {
+ "epoch": 55.419213973799124,
+ "grad_norm": 0.018154339864850044,
+ "learning_rate": 0.0006,
+ "loss": 4.527122974395752,
+ "step": 3990
+ },
+ {
+ "epoch": 55.43318777292576,
+ "grad_norm": 0.01914648339152336,
+ "learning_rate": 0.0006,
+ "loss": 4.641975402832031,
+ "step": 3991
+ },
+ {
+ "epoch": 55.4471615720524,
+ "grad_norm": 0.017118625342845917,
+ "learning_rate": 0.0006,
+ "loss": 4.511425971984863,
+ "step": 3992
+ },
+ {
+ "epoch": 55.46113537117904,
+ "grad_norm": 0.01551489345729351,
+ "learning_rate": 0.0006,
+ "loss": 4.551141738891602,
+ "step": 3993
+ },
+ {
+ "epoch": 55.47510917030568,
+ "grad_norm": 0.015072065405547619,
+ "learning_rate": 0.0006,
+ "loss": 4.651115894317627,
+ "step": 3994
+ },
+ {
+ "epoch": 55.48908296943232,
+ "grad_norm": 0.017010986804962158,
+ "learning_rate": 0.0006,
+ "loss": 4.583282470703125,
+ "step": 3995
+ },
+ {
+ "epoch": 55.50305676855895,
+ "grad_norm": 0.01764831133186817,
+ "learning_rate": 0.0006,
+ "loss": 4.593364238739014,
+ "step": 3996
+ },
+ {
+ "epoch": 55.51703056768559,
+ "grad_norm": 0.01762884296476841,
+ "learning_rate": 0.0006,
+ "loss": 4.45671272277832,
+ "step": 3997
+ },
+ {
+ "epoch": 55.531004366812226,
+ "grad_norm": 0.01600457727909088,
+ "learning_rate": 0.0006,
+ "loss": 4.511260032653809,
+ "step": 3998
+ },
+ {
+ "epoch": 55.544978165938865,
+ "grad_norm": 0.015706021338701248,
+ "learning_rate": 0.0006,
+ "loss": 4.47593355178833,
+ "step": 3999
+ },
+ {
+ "epoch": 55.5589519650655,
+ "grad_norm": 0.017718996852636337,
+ "learning_rate": 0.0006,
+ "loss": 4.610543251037598,
+ "step": 4000
+ },
+ {
+ "epoch": 55.57292576419214,
+ "grad_norm": 0.015073757618665695,
+ "learning_rate": 0.0006,
+ "loss": 4.466620445251465,
+ "step": 4001
+ },
+ {
+ "epoch": 55.58689956331878,
+ "grad_norm": 0.01789247989654541,
+ "learning_rate": 0.0006,
+ "loss": 4.503409385681152,
+ "step": 4002
+ },
+ {
+ "epoch": 55.60087336244541,
+ "grad_norm": 0.022329598665237427,
+ "learning_rate": 0.0006,
+ "loss": 4.62099552154541,
+ "step": 4003
+ },
+ {
+ "epoch": 55.61484716157205,
+ "grad_norm": 0.02343829534947872,
+ "learning_rate": 0.0006,
+ "loss": 4.620813369750977,
+ "step": 4004
+ },
+ {
+ "epoch": 55.62882096069869,
+ "grad_norm": 0.023279687389731407,
+ "learning_rate": 0.0006,
+ "loss": 4.545765399932861,
+ "step": 4005
+ },
+ {
+ "epoch": 55.64279475982533,
+ "grad_norm": 0.020817464217543602,
+ "learning_rate": 0.0006,
+ "loss": 4.561470031738281,
+ "step": 4006
+ },
+ {
+ "epoch": 55.65676855895197,
+ "grad_norm": 0.020270058885216713,
+ "learning_rate": 0.0006,
+ "loss": 4.5875020027160645,
+ "step": 4007
+ },
+ {
+ "epoch": 55.670742358078606,
+ "grad_norm": 0.021085843443870544,
+ "learning_rate": 0.0006,
+ "loss": 4.504560470581055,
+ "step": 4008
+ },
+ {
+ "epoch": 55.68471615720524,
+ "grad_norm": 0.018507730215787888,
+ "learning_rate": 0.0006,
+ "loss": 4.520644664764404,
+ "step": 4009
+ },
+ {
+ "epoch": 55.698689956331876,
+ "grad_norm": 0.020571516826748848,
+ "learning_rate": 0.0006,
+ "loss": 4.539643287658691,
+ "step": 4010
+ },
+ {
+ "epoch": 55.712663755458514,
+ "grad_norm": 0.021484510973095894,
+ "learning_rate": 0.0006,
+ "loss": 4.628833293914795,
+ "step": 4011
+ },
+ {
+ "epoch": 55.72663755458515,
+ "grad_norm": 0.017693819478154182,
+ "learning_rate": 0.0006,
+ "loss": 4.6023783683776855,
+ "step": 4012
+ },
+ {
+ "epoch": 55.74061135371179,
+ "grad_norm": 0.020186059176921844,
+ "learning_rate": 0.0006,
+ "loss": 4.634884834289551,
+ "step": 4013
+ },
+ {
+ "epoch": 55.75458515283843,
+ "grad_norm": 0.01835780404508114,
+ "learning_rate": 0.0006,
+ "loss": 4.577791213989258,
+ "step": 4014
+ },
+ {
+ "epoch": 55.76855895196506,
+ "grad_norm": 0.018162988126277924,
+ "learning_rate": 0.0006,
+ "loss": 4.641597270965576,
+ "step": 4015
+ },
+ {
+ "epoch": 55.7825327510917,
+ "grad_norm": 0.018909504637122154,
+ "learning_rate": 0.0006,
+ "loss": 4.487447738647461,
+ "step": 4016
+ },
+ {
+ "epoch": 55.79650655021834,
+ "grad_norm": 0.018972251564264297,
+ "learning_rate": 0.0006,
+ "loss": 4.590317726135254,
+ "step": 4017
+ },
+ {
+ "epoch": 55.81048034934498,
+ "grad_norm": 0.01793607883155346,
+ "learning_rate": 0.0006,
+ "loss": 4.4994611740112305,
+ "step": 4018
+ },
+ {
+ "epoch": 55.82445414847162,
+ "grad_norm": 0.02123514749109745,
+ "learning_rate": 0.0006,
+ "loss": 4.479006767272949,
+ "step": 4019
+ },
+ {
+ "epoch": 55.838427947598255,
+ "grad_norm": 0.021246599033474922,
+ "learning_rate": 0.0006,
+ "loss": 4.4738640785217285,
+ "step": 4020
+ },
+ {
+ "epoch": 55.852401746724894,
+ "grad_norm": 0.020131602883338928,
+ "learning_rate": 0.0006,
+ "loss": 4.600308418273926,
+ "step": 4021
+ },
+ {
+ "epoch": 55.866375545851525,
+ "grad_norm": 0.019159460440278053,
+ "learning_rate": 0.0006,
+ "loss": 4.540908336639404,
+ "step": 4022
+ },
+ {
+ "epoch": 55.880349344978164,
+ "grad_norm": 0.023114705458283424,
+ "learning_rate": 0.0006,
+ "loss": 4.641247749328613,
+ "step": 4023
+ },
+ {
+ "epoch": 55.8943231441048,
+ "grad_norm": 0.02296135015785694,
+ "learning_rate": 0.0006,
+ "loss": 4.444624423980713,
+ "step": 4024
+ },
+ {
+ "epoch": 55.90829694323144,
+ "grad_norm": 0.02204655110836029,
+ "learning_rate": 0.0006,
+ "loss": 4.506349563598633,
+ "step": 4025
+ },
+ {
+ "epoch": 55.92227074235808,
+ "grad_norm": 0.019338225945830345,
+ "learning_rate": 0.0006,
+ "loss": 4.616047382354736,
+ "step": 4026
+ },
+ {
+ "epoch": 55.93624454148472,
+ "grad_norm": 0.020357884466648102,
+ "learning_rate": 0.0006,
+ "loss": 4.48170280456543,
+ "step": 4027
+ },
+ {
+ "epoch": 55.95021834061135,
+ "grad_norm": 0.02426772564649582,
+ "learning_rate": 0.0006,
+ "loss": 4.576877117156982,
+ "step": 4028
+ },
+ {
+ "epoch": 55.96419213973799,
+ "grad_norm": 0.023351816460490227,
+ "learning_rate": 0.0006,
+ "loss": 4.458885192871094,
+ "step": 4029
+ },
+ {
+ "epoch": 55.97816593886463,
+ "grad_norm": 0.019892243668437004,
+ "learning_rate": 0.0006,
+ "loss": 4.520229816436768,
+ "step": 4030
+ },
+ {
+ "epoch": 55.992139737991266,
+ "grad_norm": 0.018331089988350868,
+ "learning_rate": 0.0006,
+ "loss": 4.432201862335205,
+ "step": 4031
+ },
+ {
+ "epoch": 56.0,
+ "grad_norm": 0.021300828084349632,
+ "learning_rate": 0.0006,
+ "loss": 4.617072582244873,
+ "step": 4032
+ },
+ {
+ "epoch": 56.0,
+ "eval_loss": 4.811229705810547,
+ "eval_runtime": 56.4846,
+ "eval_samples_per_second": 43.233,
+ "eval_steps_per_second": 1.363,
+ "step": 4032
+ },
+ {
+ "epoch": 56.01397379912664,
+ "grad_norm": 0.02199401892721653,
+ "learning_rate": 0.0006,
+ "loss": 4.513965606689453,
+ "step": 4033
+ },
+ {
+ "epoch": 56.02794759825328,
+ "grad_norm": 0.022109858691692352,
+ "learning_rate": 0.0006,
+ "loss": 4.557568550109863,
+ "step": 4034
+ },
+ {
+ "epoch": 56.041921397379916,
+ "grad_norm": 0.020103694871068,
+ "learning_rate": 0.0006,
+ "loss": 4.569512844085693,
+ "step": 4035
+ },
+ {
+ "epoch": 56.05589519650655,
+ "grad_norm": 0.01898941583931446,
+ "learning_rate": 0.0006,
+ "loss": 4.623138904571533,
+ "step": 4036
+ },
+ {
+ "epoch": 56.069868995633186,
+ "grad_norm": 0.02227010577917099,
+ "learning_rate": 0.0006,
+ "loss": 4.433316230773926,
+ "step": 4037
+ },
+ {
+ "epoch": 56.083842794759825,
+ "grad_norm": 0.024683045223355293,
+ "learning_rate": 0.0006,
+ "loss": 4.458296775817871,
+ "step": 4038
+ },
+ {
+ "epoch": 56.09781659388646,
+ "grad_norm": 0.02900184690952301,
+ "learning_rate": 0.0006,
+ "loss": 4.531060218811035,
+ "step": 4039
+ },
+ {
+ "epoch": 56.1117903930131,
+ "grad_norm": 0.025557566434144974,
+ "learning_rate": 0.0006,
+ "loss": 4.56685733795166,
+ "step": 4040
+ },
+ {
+ "epoch": 56.12576419213974,
+ "grad_norm": 0.021414656192064285,
+ "learning_rate": 0.0006,
+ "loss": 4.469037055969238,
+ "step": 4041
+ },
+ {
+ "epoch": 56.13973799126637,
+ "grad_norm": 0.023262616246938705,
+ "learning_rate": 0.0006,
+ "loss": 4.493876934051514,
+ "step": 4042
+ },
+ {
+ "epoch": 56.15371179039301,
+ "grad_norm": 0.023011326789855957,
+ "learning_rate": 0.0006,
+ "loss": 4.4572906494140625,
+ "step": 4043
+ },
+ {
+ "epoch": 56.16768558951965,
+ "grad_norm": 0.02177615463733673,
+ "learning_rate": 0.0006,
+ "loss": 4.615389823913574,
+ "step": 4044
+ },
+ {
+ "epoch": 56.18165938864629,
+ "grad_norm": 0.021527713164687157,
+ "learning_rate": 0.0006,
+ "loss": 4.613762855529785,
+ "step": 4045
+ },
+ {
+ "epoch": 56.19563318777293,
+ "grad_norm": 0.01982325315475464,
+ "learning_rate": 0.0006,
+ "loss": 4.53670597076416,
+ "step": 4046
+ },
+ {
+ "epoch": 56.209606986899566,
+ "grad_norm": 0.022484654560685158,
+ "learning_rate": 0.0006,
+ "loss": 4.605565071105957,
+ "step": 4047
+ },
+ {
+ "epoch": 56.223580786026204,
+ "grad_norm": 0.02346140146255493,
+ "learning_rate": 0.0006,
+ "loss": 4.6265950202941895,
+ "step": 4048
+ },
+ {
+ "epoch": 56.237554585152836,
+ "grad_norm": 0.01876714453101158,
+ "learning_rate": 0.0006,
+ "loss": 4.481525421142578,
+ "step": 4049
+ },
+ {
+ "epoch": 56.251528384279474,
+ "grad_norm": 0.01926851086318493,
+ "learning_rate": 0.0006,
+ "loss": 4.630258083343506,
+ "step": 4050
+ },
+ {
+ "epoch": 56.26550218340611,
+ "grad_norm": 0.020858589559793472,
+ "learning_rate": 0.0006,
+ "loss": 4.5265913009643555,
+ "step": 4051
+ },
+ {
+ "epoch": 56.27947598253275,
+ "grad_norm": 0.01889280416071415,
+ "learning_rate": 0.0006,
+ "loss": 4.581275463104248,
+ "step": 4052
+ },
+ {
+ "epoch": 56.29344978165939,
+ "grad_norm": 0.01801002025604248,
+ "learning_rate": 0.0006,
+ "loss": 4.498557090759277,
+ "step": 4053
+ },
+ {
+ "epoch": 56.30742358078603,
+ "grad_norm": 0.02252180315554142,
+ "learning_rate": 0.0006,
+ "loss": 4.565267086029053,
+ "step": 4054
+ },
+ {
+ "epoch": 56.32139737991266,
+ "grad_norm": 0.02306070365011692,
+ "learning_rate": 0.0006,
+ "loss": 4.3591389656066895,
+ "step": 4055
+ },
+ {
+ "epoch": 56.3353711790393,
+ "grad_norm": 0.020292513072490692,
+ "learning_rate": 0.0006,
+ "loss": 4.407517433166504,
+ "step": 4056
+ },
+ {
+ "epoch": 56.34934497816594,
+ "grad_norm": 0.019154123961925507,
+ "learning_rate": 0.0006,
+ "loss": 4.482261657714844,
+ "step": 4057
+ },
+ {
+ "epoch": 56.36331877729258,
+ "grad_norm": 0.017338937148451805,
+ "learning_rate": 0.0006,
+ "loss": 4.4463582038879395,
+ "step": 4058
+ },
+ {
+ "epoch": 56.377292576419215,
+ "grad_norm": 0.01730559952557087,
+ "learning_rate": 0.0006,
+ "loss": 4.542474746704102,
+ "step": 4059
+ },
+ {
+ "epoch": 56.391266375545854,
+ "grad_norm": 0.017029426991939545,
+ "learning_rate": 0.0006,
+ "loss": 4.552488327026367,
+ "step": 4060
+ },
+ {
+ "epoch": 56.40524017467249,
+ "grad_norm": 0.01745498552918434,
+ "learning_rate": 0.0006,
+ "loss": 4.52928352355957,
+ "step": 4061
+ },
+ {
+ "epoch": 56.419213973799124,
+ "grad_norm": 0.0168515145778656,
+ "learning_rate": 0.0006,
+ "loss": 4.4516801834106445,
+ "step": 4062
+ },
+ {
+ "epoch": 56.43318777292576,
+ "grad_norm": 0.018545417115092278,
+ "learning_rate": 0.0006,
+ "loss": 4.569811820983887,
+ "step": 4063
+ },
+ {
+ "epoch": 56.4471615720524,
+ "grad_norm": 0.02227838523685932,
+ "learning_rate": 0.0006,
+ "loss": 4.412224769592285,
+ "step": 4064
+ },
+ {
+ "epoch": 56.46113537117904,
+ "grad_norm": 0.023338504135608673,
+ "learning_rate": 0.0006,
+ "loss": 4.490184307098389,
+ "step": 4065
+ },
+ {
+ "epoch": 56.47510917030568,
+ "grad_norm": 0.02257535606622696,
+ "learning_rate": 0.0006,
+ "loss": 4.479152679443359,
+ "step": 4066
+ },
+ {
+ "epoch": 56.48908296943232,
+ "grad_norm": 0.022737598046660423,
+ "learning_rate": 0.0006,
+ "loss": 4.418235778808594,
+ "step": 4067
+ },
+ {
+ "epoch": 56.50305676855895,
+ "grad_norm": 0.024603573605418205,
+ "learning_rate": 0.0006,
+ "loss": 4.45781135559082,
+ "step": 4068
+ },
+ {
+ "epoch": 56.51703056768559,
+ "grad_norm": 0.024953778833150864,
+ "learning_rate": 0.0006,
+ "loss": 4.593907833099365,
+ "step": 4069
+ },
+ {
+ "epoch": 56.531004366812226,
+ "grad_norm": 0.02195645309984684,
+ "learning_rate": 0.0006,
+ "loss": 4.467643737792969,
+ "step": 4070
+ },
+ {
+ "epoch": 56.544978165938865,
+ "grad_norm": 0.017126578837633133,
+ "learning_rate": 0.0006,
+ "loss": 4.555484771728516,
+ "step": 4071
+ },
+ {
+ "epoch": 56.5589519650655,
+ "grad_norm": 0.019482605159282684,
+ "learning_rate": 0.0006,
+ "loss": 4.430915832519531,
+ "step": 4072
+ },
+ {
+ "epoch": 56.57292576419214,
+ "grad_norm": 0.0219392292201519,
+ "learning_rate": 0.0006,
+ "loss": 4.593365669250488,
+ "step": 4073
+ },
+ {
+ "epoch": 56.58689956331878,
+ "grad_norm": 0.02420145645737648,
+ "learning_rate": 0.0006,
+ "loss": 4.491467475891113,
+ "step": 4074
+ },
+ {
+ "epoch": 56.60087336244541,
+ "grad_norm": 0.021523723378777504,
+ "learning_rate": 0.0006,
+ "loss": 4.463634490966797,
+ "step": 4075
+ },
+ {
+ "epoch": 56.61484716157205,
+ "grad_norm": 0.020242048427462578,
+ "learning_rate": 0.0006,
+ "loss": 4.577968597412109,
+ "step": 4076
+ },
+ {
+ "epoch": 56.62882096069869,
+ "grad_norm": 0.017771173268556595,
+ "learning_rate": 0.0006,
+ "loss": 4.568269729614258,
+ "step": 4077
+ },
+ {
+ "epoch": 56.64279475982533,
+ "grad_norm": 0.01598544418811798,
+ "learning_rate": 0.0006,
+ "loss": 4.476841926574707,
+ "step": 4078
+ },
+ {
+ "epoch": 56.65676855895197,
+ "grad_norm": 0.0168446097522974,
+ "learning_rate": 0.0006,
+ "loss": 4.480965614318848,
+ "step": 4079
+ },
+ {
+ "epoch": 56.670742358078606,
+ "grad_norm": 0.014948660507798195,
+ "learning_rate": 0.0006,
+ "loss": 4.535400390625,
+ "step": 4080
+ },
+ {
+ "epoch": 56.68471615720524,
+ "grad_norm": 0.014899153262376785,
+ "learning_rate": 0.0006,
+ "loss": 4.476048946380615,
+ "step": 4081
+ },
+ {
+ "epoch": 56.698689956331876,
+ "grad_norm": 0.014051870442926884,
+ "learning_rate": 0.0006,
+ "loss": 4.533884048461914,
+ "step": 4082
+ },
+ {
+ "epoch": 56.712663755458514,
+ "grad_norm": 0.014708572067320347,
+ "learning_rate": 0.0006,
+ "loss": 4.586690902709961,
+ "step": 4083
+ },
+ {
+ "epoch": 56.72663755458515,
+ "grad_norm": 0.014797299169003963,
+ "learning_rate": 0.0006,
+ "loss": 4.446291446685791,
+ "step": 4084
+ },
+ {
+ "epoch": 56.74061135371179,
+ "grad_norm": 0.015256541781127453,
+ "learning_rate": 0.0006,
+ "loss": 4.553572654724121,
+ "step": 4085
+ },
+ {
+ "epoch": 56.75458515283843,
+ "grad_norm": 0.016042491421103477,
+ "learning_rate": 0.0006,
+ "loss": 4.507801055908203,
+ "step": 4086
+ },
+ {
+ "epoch": 56.76855895196506,
+ "grad_norm": 0.018835173919796944,
+ "learning_rate": 0.0006,
+ "loss": 4.393911838531494,
+ "step": 4087
+ },
+ {
+ "epoch": 56.7825327510917,
+ "grad_norm": 0.021994033828377724,
+ "learning_rate": 0.0006,
+ "loss": 4.543368339538574,
+ "step": 4088
+ },
+ {
+ "epoch": 56.79650655021834,
+ "grad_norm": 0.0230990182608366,
+ "learning_rate": 0.0006,
+ "loss": 4.507420539855957,
+ "step": 4089
+ },
+ {
+ "epoch": 56.81048034934498,
+ "grad_norm": 0.018818842247128487,
+ "learning_rate": 0.0006,
+ "loss": 4.499941349029541,
+ "step": 4090
+ },
+ {
+ "epoch": 56.82445414847162,
+ "grad_norm": 0.018257278949022293,
+ "learning_rate": 0.0006,
+ "loss": 4.380057334899902,
+ "step": 4091
+ },
+ {
+ "epoch": 56.838427947598255,
+ "grad_norm": 0.020073018968105316,
+ "learning_rate": 0.0006,
+ "loss": 4.572811603546143,
+ "step": 4092
+ },
+ {
+ "epoch": 56.852401746724894,
+ "grad_norm": 0.02078658528625965,
+ "learning_rate": 0.0006,
+ "loss": 4.446835517883301,
+ "step": 4093
+ },
+ {
+ "epoch": 56.866375545851525,
+ "grad_norm": 0.01793118566274643,
+ "learning_rate": 0.0006,
+ "loss": 4.517854690551758,
+ "step": 4094
+ },
+ {
+ "epoch": 56.880349344978164,
+ "grad_norm": 0.01608281210064888,
+ "learning_rate": 0.0006,
+ "loss": 4.506482124328613,
+ "step": 4095
+ },
+ {
+ "epoch": 56.8943231441048,
+ "grad_norm": 0.017405368387699127,
+ "learning_rate": 0.0006,
+ "loss": 4.439718246459961,
+ "step": 4096
+ },
+ {
+ "epoch": 56.90829694323144,
+ "grad_norm": 0.017930863425135612,
+ "learning_rate": 0.0006,
+ "loss": 4.409365653991699,
+ "step": 4097
+ },
+ {
+ "epoch": 56.92227074235808,
+ "grad_norm": 0.015757670626044273,
+ "learning_rate": 0.0006,
+ "loss": 4.505547523498535,
+ "step": 4098
+ },
+ {
+ "epoch": 56.93624454148472,
+ "grad_norm": 0.01580345816910267,
+ "learning_rate": 0.0006,
+ "loss": 4.462774753570557,
+ "step": 4099
+ },
+ {
+ "epoch": 56.95021834061135,
+ "grad_norm": 0.01659180410206318,
+ "learning_rate": 0.0006,
+ "loss": 4.49345588684082,
+ "step": 4100
+ },
+ {
+ "epoch": 56.96419213973799,
+ "grad_norm": 0.01877259835600853,
+ "learning_rate": 0.0006,
+ "loss": 4.574443340301514,
+ "step": 4101
+ },
+ {
+ "epoch": 56.97816593886463,
+ "grad_norm": 0.01977839693427086,
+ "learning_rate": 0.0006,
+ "loss": 4.574034690856934,
+ "step": 4102
+ },
+ {
+ "epoch": 56.992139737991266,
+ "grad_norm": 0.020627597346901894,
+ "learning_rate": 0.0006,
+ "loss": 4.50969123840332,
+ "step": 4103
+ },
+ {
+ "epoch": 57.0,
+ "grad_norm": 0.025128807872533798,
+ "learning_rate": 0.0006,
+ "loss": 4.401662826538086,
+ "step": 4104
+ },
+ {
+ "epoch": 57.0,
+ "eval_loss": 4.779252052307129,
+ "eval_runtime": 56.8104,
+ "eval_samples_per_second": 42.985,
+ "eval_steps_per_second": 1.355,
+ "step": 4104
+ },
+ {
+ "epoch": 57.01397379912664,
+ "grad_norm": 0.027712570503354073,
+ "learning_rate": 0.0006,
+ "loss": 4.472955703735352,
+ "step": 4105
+ },
+ {
+ "epoch": 57.02794759825328,
+ "grad_norm": 0.02164197526872158,
+ "learning_rate": 0.0006,
+ "loss": 4.387596130371094,
+ "step": 4106
+ },
+ {
+ "epoch": 57.041921397379916,
+ "grad_norm": 0.017972752451896667,
+ "learning_rate": 0.0006,
+ "loss": 4.4326276779174805,
+ "step": 4107
+ },
+ {
+ "epoch": 57.05589519650655,
+ "grad_norm": 0.020811019465327263,
+ "learning_rate": 0.0006,
+ "loss": 4.452951431274414,
+ "step": 4108
+ },
+ {
+ "epoch": 57.069868995633186,
+ "grad_norm": 0.019913259893655777,
+ "learning_rate": 0.0006,
+ "loss": 4.448301315307617,
+ "step": 4109
+ },
+ {
+ "epoch": 57.083842794759825,
+ "grad_norm": 0.017409102991223335,
+ "learning_rate": 0.0006,
+ "loss": 4.4108357429504395,
+ "step": 4110
+ },
+ {
+ "epoch": 57.09781659388646,
+ "grad_norm": 0.019374269992113113,
+ "learning_rate": 0.0006,
+ "loss": 4.398412227630615,
+ "step": 4111
+ },
+ {
+ "epoch": 57.1117903930131,
+ "grad_norm": 0.015521776862442493,
+ "learning_rate": 0.0006,
+ "loss": 4.508594989776611,
+ "step": 4112
+ },
+ {
+ "epoch": 57.12576419213974,
+ "grad_norm": 0.017201920971274376,
+ "learning_rate": 0.0006,
+ "loss": 4.449558734893799,
+ "step": 4113
+ },
+ {
+ "epoch": 57.13973799126637,
+ "grad_norm": 0.018062541261315346,
+ "learning_rate": 0.0006,
+ "loss": 4.517874240875244,
+ "step": 4114
+ },
+ {
+ "epoch": 57.15371179039301,
+ "grad_norm": 0.015846073627471924,
+ "learning_rate": 0.0006,
+ "loss": 4.427501678466797,
+ "step": 4115
+ },
+ {
+ "epoch": 57.16768558951965,
+ "grad_norm": 0.01849362626671791,
+ "learning_rate": 0.0006,
+ "loss": 4.433181285858154,
+ "step": 4116
+ },
+ {
+ "epoch": 57.18165938864629,
+ "grad_norm": 0.01948804222047329,
+ "learning_rate": 0.0006,
+ "loss": 4.446122646331787,
+ "step": 4117
+ },
+ {
+ "epoch": 57.19563318777293,
+ "grad_norm": 0.017354389652609825,
+ "learning_rate": 0.0006,
+ "loss": 4.431153774261475,
+ "step": 4118
+ },
+ {
+ "epoch": 57.209606986899566,
+ "grad_norm": 0.015817373991012573,
+ "learning_rate": 0.0006,
+ "loss": 4.448566913604736,
+ "step": 4119
+ },
+ {
+ "epoch": 57.223580786026204,
+ "grad_norm": 0.01635785959661007,
+ "learning_rate": 0.0006,
+ "loss": 4.408031463623047,
+ "step": 4120
+ },
+ {
+ "epoch": 57.237554585152836,
+ "grad_norm": 0.014528338797390461,
+ "learning_rate": 0.0006,
+ "loss": 4.541668891906738,
+ "step": 4121
+ },
+ {
+ "epoch": 57.251528384279474,
+ "grad_norm": 0.01589174196124077,
+ "learning_rate": 0.0006,
+ "loss": 4.513199329376221,
+ "step": 4122
+ },
+ {
+ "epoch": 57.26550218340611,
+ "grad_norm": 0.016093695536255836,
+ "learning_rate": 0.0006,
+ "loss": 4.516178131103516,
+ "step": 4123
+ },
+ {
+ "epoch": 57.27947598253275,
+ "grad_norm": 0.01788361556828022,
+ "learning_rate": 0.0006,
+ "loss": 4.524880409240723,
+ "step": 4124
+ },
+ {
+ "epoch": 57.29344978165939,
+ "grad_norm": 0.01591576263308525,
+ "learning_rate": 0.0006,
+ "loss": 4.44764518737793,
+ "step": 4125
+ },
+ {
+ "epoch": 57.30742358078603,
+ "grad_norm": 0.015452570281922817,
+ "learning_rate": 0.0006,
+ "loss": 4.46844482421875,
+ "step": 4126
+ },
+ {
+ "epoch": 57.32139737991266,
+ "grad_norm": 0.016404492780566216,
+ "learning_rate": 0.0006,
+ "loss": 4.443678855895996,
+ "step": 4127
+ },
+ {
+ "epoch": 57.3353711790393,
+ "grad_norm": 0.019967148080468178,
+ "learning_rate": 0.0006,
+ "loss": 4.430230140686035,
+ "step": 4128
+ },
+ {
+ "epoch": 57.34934497816594,
+ "grad_norm": 0.02594015561044216,
+ "learning_rate": 0.0006,
+ "loss": 4.47061824798584,
+ "step": 4129
+ },
+ {
+ "epoch": 57.36331877729258,
+ "grad_norm": 0.031096890568733215,
+ "learning_rate": 0.0006,
+ "loss": 4.462275981903076,
+ "step": 4130
+ },
+ {
+ "epoch": 57.377292576419215,
+ "grad_norm": 0.02834898792207241,
+ "learning_rate": 0.0006,
+ "loss": 4.41143798828125,
+ "step": 4131
+ },
+ {
+ "epoch": 57.391266375545854,
+ "grad_norm": 0.023621153086423874,
+ "learning_rate": 0.0006,
+ "loss": 4.562548637390137,
+ "step": 4132
+ },
+ {
+ "epoch": 57.40524017467249,
+ "grad_norm": 0.02277296595275402,
+ "learning_rate": 0.0006,
+ "loss": 4.456247329711914,
+ "step": 4133
+ },
+ {
+ "epoch": 57.419213973799124,
+ "grad_norm": 0.019868768751621246,
+ "learning_rate": 0.0006,
+ "loss": 4.524300575256348,
+ "step": 4134
+ },
+ {
+ "epoch": 57.43318777292576,
+ "grad_norm": 0.01887078955769539,
+ "learning_rate": 0.0006,
+ "loss": 4.540097236633301,
+ "step": 4135
+ },
+ {
+ "epoch": 57.4471615720524,
+ "grad_norm": 0.024268802255392075,
+ "learning_rate": 0.0006,
+ "loss": 4.397233963012695,
+ "step": 4136
+ },
+ {
+ "epoch": 57.46113537117904,
+ "grad_norm": 0.025739721953868866,
+ "learning_rate": 0.0006,
+ "loss": 4.354194641113281,
+ "step": 4137
+ },
+ {
+ "epoch": 57.47510917030568,
+ "grad_norm": 0.02442290261387825,
+ "learning_rate": 0.0006,
+ "loss": 4.303822994232178,
+ "step": 4138
+ },
+ {
+ "epoch": 57.48908296943232,
+ "grad_norm": 0.022215209901332855,
+ "learning_rate": 0.0006,
+ "loss": 4.4870452880859375,
+ "step": 4139
+ },
+ {
+ "epoch": 57.50305676855895,
+ "grad_norm": 0.01789158768951893,
+ "learning_rate": 0.0006,
+ "loss": 4.343809127807617,
+ "step": 4140
+ },
+ {
+ "epoch": 57.51703056768559,
+ "grad_norm": 0.020381739363074303,
+ "learning_rate": 0.0006,
+ "loss": 4.53312873840332,
+ "step": 4141
+ },
+ {
+ "epoch": 57.531004366812226,
+ "grad_norm": 0.019585080444812775,
+ "learning_rate": 0.0006,
+ "loss": 4.510162353515625,
+ "step": 4142
+ },
+ {
+ "epoch": 57.544978165938865,
+ "grad_norm": 0.01985127478837967,
+ "learning_rate": 0.0006,
+ "loss": 4.3654022216796875,
+ "step": 4143
+ },
+ {
+ "epoch": 57.5589519650655,
+ "grad_norm": 0.01962905190885067,
+ "learning_rate": 0.0006,
+ "loss": 4.560830116271973,
+ "step": 4144
+ },
+ {
+ "epoch": 57.57292576419214,
+ "grad_norm": 0.02679629437625408,
+ "learning_rate": 0.0006,
+ "loss": 4.300804138183594,
+ "step": 4145
+ },
+ {
+ "epoch": 57.58689956331878,
+ "grad_norm": 0.039830856025218964,
+ "learning_rate": 0.0006,
+ "loss": 4.54150915145874,
+ "step": 4146
+ },
+ {
+ "epoch": 57.60087336244541,
+ "grad_norm": 0.049527477473020554,
+ "learning_rate": 0.0006,
+ "loss": 4.405792236328125,
+ "step": 4147
+ },
+ {
+ "epoch": 57.61484716157205,
+ "grad_norm": 0.042966946959495544,
+ "learning_rate": 0.0006,
+ "loss": 4.500643730163574,
+ "step": 4148
+ },
+ {
+ "epoch": 57.62882096069869,
+ "grad_norm": 0.08999177813529968,
+ "learning_rate": 0.0006,
+ "loss": 4.504790782928467,
+ "step": 4149
+ },
+ {
+ "epoch": 57.64279475982533,
+ "grad_norm": 0.44693124294281006,
+ "learning_rate": 0.0006,
+ "loss": 4.845344066619873,
+ "step": 4150
+ },
+ {
+ "epoch": 57.65676855895197,
+ "grad_norm": 0.7143144011497498,
+ "learning_rate": 0.0006,
+ "loss": 6.907845497131348,
+ "step": 4151
+ },
+ {
+ "epoch": 57.670742358078606,
+ "grad_norm": 0.25239595770835876,
+ "learning_rate": 0.0006,
+ "loss": 7.082579612731934,
+ "step": 4152
+ },
+ {
+ "epoch": 57.68471615720524,
+ "grad_norm": 0.26550790667533875,
+ "learning_rate": 0.0006,
+ "loss": 7.50180196762085,
+ "step": 4153
+ },
+ {
+ "epoch": 57.698689956331876,
+ "grad_norm": 0.14735378324985504,
+ "learning_rate": 0.0006,
+ "loss": 7.214590072631836,
+ "step": 4154
+ },
+ {
+ "epoch": 57.712663755458514,
+ "grad_norm": 0.17685334384441376,
+ "learning_rate": 0.0006,
+ "loss": 6.861320495605469,
+ "step": 4155
+ },
+ {
+ "epoch": 57.72663755458515,
+ "grad_norm": 0.18231339752674103,
+ "learning_rate": 0.0006,
+ "loss": 6.791839599609375,
+ "step": 4156
+ },
+ {
+ "epoch": 57.74061135371179,
+ "grad_norm": 0.09230359643697739,
+ "learning_rate": 0.0006,
+ "loss": 6.760223388671875,
+ "step": 4157
+ },
+ {
+ "epoch": 57.75458515283843,
+ "grad_norm": 0.06918249279260635,
+ "learning_rate": 0.0006,
+ "loss": 6.532999515533447,
+ "step": 4158
+ },
+ {
+ "epoch": 57.76855895196506,
+ "grad_norm": 0.0901007279753685,
+ "learning_rate": 0.0006,
+ "loss": 6.52907133102417,
+ "step": 4159
+ },
+ {
+ "epoch": 57.7825327510917,
+ "grad_norm": 0.05678229779005051,
+ "learning_rate": 0.0006,
+ "loss": 6.295290470123291,
+ "step": 4160
+ },
+ {
+ "epoch": 57.79650655021834,
+ "grad_norm": 0.04450210556387901,
+ "learning_rate": 0.0006,
+ "loss": 6.14900541305542,
+ "step": 4161
+ },
+ {
+ "epoch": 57.81048034934498,
+ "grad_norm": 0.049191512167453766,
+ "learning_rate": 0.0006,
+ "loss": 6.19650411605835,
+ "step": 4162
+ },
+ {
+ "epoch": 57.82445414847162,
+ "grad_norm": 0.03908967226743698,
+ "learning_rate": 0.0006,
+ "loss": 6.054948806762695,
+ "step": 4163
+ },
+ {
+ "epoch": 57.838427947598255,
+ "grad_norm": 0.04023387283086777,
+ "learning_rate": 0.0006,
+ "loss": 6.034873962402344,
+ "step": 4164
+ },
+ {
+ "epoch": 57.852401746724894,
+ "grad_norm": 0.03304268419742584,
+ "learning_rate": 0.0006,
+ "loss": 5.769852161407471,
+ "step": 4165
+ },
+ {
+ "epoch": 57.866375545851525,
+ "grad_norm": 0.03432450816035271,
+ "learning_rate": 0.0006,
+ "loss": 5.788785457611084,
+ "step": 4166
+ },
+ {
+ "epoch": 57.880349344978164,
+ "grad_norm": 0.029965840280056,
+ "learning_rate": 0.0006,
+ "loss": 5.815243721008301,
+ "step": 4167
+ },
+ {
+ "epoch": 57.8943231441048,
+ "grad_norm": 0.03637698292732239,
+ "learning_rate": 0.0006,
+ "loss": 5.7204060554504395,
+ "step": 4168
+ },
+ {
+ "epoch": 57.90829694323144,
+ "grad_norm": 0.03477517142891884,
+ "learning_rate": 0.0006,
+ "loss": 5.605930328369141,
+ "step": 4169
+ },
+ {
+ "epoch": 57.92227074235808,
+ "grad_norm": 0.05061859264969826,
+ "learning_rate": 0.0006,
+ "loss": 5.547054767608643,
+ "step": 4170
+ },
+ {
+ "epoch": 57.93624454148472,
+ "grad_norm": 0.06299655884504318,
+ "learning_rate": 0.0006,
+ "loss": 5.552166938781738,
+ "step": 4171
+ },
+ {
+ "epoch": 57.95021834061135,
+ "grad_norm": 0.04215948283672333,
+ "learning_rate": 0.0006,
+ "loss": 5.375457763671875,
+ "step": 4172
+ },
+ {
+ "epoch": 57.96419213973799,
+ "grad_norm": 0.0348566472530365,
+ "learning_rate": 0.0006,
+ "loss": 5.418606281280518,
+ "step": 4173
+ },
+ {
+ "epoch": 57.97816593886463,
+ "grad_norm": 0.03807530924677849,
+ "learning_rate": 0.0006,
+ "loss": 5.403156280517578,
+ "step": 4174
+ },
+ {
+ "epoch": 57.992139737991266,
+ "grad_norm": 0.02989993803203106,
+ "learning_rate": 0.0006,
+ "loss": 5.355951309204102,
+ "step": 4175
+ },
+ {
+ "epoch": 58.0,
+ "grad_norm": 0.034300558269023895,
+ "learning_rate": 0.0006,
+ "loss": 5.44920539855957,
+ "step": 4176
+ },
+ {
+ "epoch": 58.0,
+ "eval_loss": 5.416731834411621,
+ "eval_runtime": 56.7934,
+ "eval_samples_per_second": 42.998,
+ "eval_steps_per_second": 1.356,
+ "step": 4176
+ },
+ {
+ "epoch": 58.01397379912664,
+ "grad_norm": 0.033961448818445206,
+ "learning_rate": 0.0006,
+ "loss": 5.250707626342773,
+ "step": 4177
+ },
+ {
+ "epoch": 58.02794759825328,
+ "grad_norm": 0.027343502268195152,
+ "learning_rate": 0.0006,
+ "loss": 5.149238109588623,
+ "step": 4178
+ },
+ {
+ "epoch": 58.041921397379916,
+ "grad_norm": 0.03033650480210781,
+ "learning_rate": 0.0006,
+ "loss": 5.072463512420654,
+ "step": 4179
+ },
+ {
+ "epoch": 58.05589519650655,
+ "grad_norm": 0.028262868523597717,
+ "learning_rate": 0.0006,
+ "loss": 5.109417915344238,
+ "step": 4180
+ },
+ {
+ "epoch": 58.069868995633186,
+ "grad_norm": 0.02373034693300724,
+ "learning_rate": 0.0006,
+ "loss": 5.082464218139648,
+ "step": 4181
+ },
+ {
+ "epoch": 58.083842794759825,
+ "grad_norm": 0.025879688560962677,
+ "learning_rate": 0.0006,
+ "loss": 5.056792259216309,
+ "step": 4182
+ },
+ {
+ "epoch": 58.09781659388646,
+ "grad_norm": 0.026252275332808495,
+ "learning_rate": 0.0006,
+ "loss": 5.064239978790283,
+ "step": 4183
+ },
+ {
+ "epoch": 58.1117903930131,
+ "grad_norm": 0.028177792206406593,
+ "learning_rate": 0.0006,
+ "loss": 4.922310829162598,
+ "step": 4184
+ },
+ {
+ "epoch": 58.12576419213974,
+ "grad_norm": 0.02630820870399475,
+ "learning_rate": 0.0006,
+ "loss": 4.992884635925293,
+ "step": 4185
+ },
+ {
+ "epoch": 58.13973799126637,
+ "grad_norm": 0.024890903383493423,
+ "learning_rate": 0.0006,
+ "loss": 4.910036563873291,
+ "step": 4186
+ },
+ {
+ "epoch": 58.15371179039301,
+ "grad_norm": 0.02822992391884327,
+ "learning_rate": 0.0006,
+ "loss": 4.866640090942383,
+ "step": 4187
+ },
+ {
+ "epoch": 58.16768558951965,
+ "grad_norm": 0.03253607079386711,
+ "learning_rate": 0.0006,
+ "loss": 4.9112958908081055,
+ "step": 4188
+ },
+ {
+ "epoch": 58.18165938864629,
+ "grad_norm": 0.03765944764018059,
+ "learning_rate": 0.0006,
+ "loss": 4.774468421936035,
+ "step": 4189
+ },
+ {
+ "epoch": 58.19563318777293,
+ "grad_norm": 0.039874881505966187,
+ "learning_rate": 0.0006,
+ "loss": 4.831678867340088,
+ "step": 4190
+ },
+ {
+ "epoch": 58.209606986899566,
+ "grad_norm": 0.034595925360918045,
+ "learning_rate": 0.0006,
+ "loss": 4.8243794441223145,
+ "step": 4191
+ },
+ {
+ "epoch": 58.223580786026204,
+ "grad_norm": 0.04840189591050148,
+ "learning_rate": 0.0006,
+ "loss": 4.854156494140625,
+ "step": 4192
+ },
+ {
+ "epoch": 58.237554585152836,
+ "grad_norm": 0.06358753889799118,
+ "learning_rate": 0.0006,
+ "loss": 4.836430549621582,
+ "step": 4193
+ },
+ {
+ "epoch": 58.251528384279474,
+ "grad_norm": 0.0507555715739727,
+ "learning_rate": 0.0006,
+ "loss": 4.88357400894165,
+ "step": 4194
+ },
+ {
+ "epoch": 58.26550218340611,
+ "grad_norm": 0.04384690150618553,
+ "learning_rate": 0.0006,
+ "loss": 4.774049282073975,
+ "step": 4195
+ },
+ {
+ "epoch": 58.27947598253275,
+ "grad_norm": 0.03246928006410599,
+ "learning_rate": 0.0006,
+ "loss": 4.686605930328369,
+ "step": 4196
+ },
+ {
+ "epoch": 58.29344978165939,
+ "grad_norm": 0.028878789395093918,
+ "learning_rate": 0.0006,
+ "loss": 4.796070575714111,
+ "step": 4197
+ },
+ {
+ "epoch": 58.30742358078603,
+ "grad_norm": 0.028175361454486847,
+ "learning_rate": 0.0006,
+ "loss": 4.728490829467773,
+ "step": 4198
+ },
+ {
+ "epoch": 58.32139737991266,
+ "grad_norm": 0.029428014531731606,
+ "learning_rate": 0.0006,
+ "loss": 4.7159624099731445,
+ "step": 4199
+ },
+ {
+ "epoch": 58.3353711790393,
+ "grad_norm": 0.031227827072143555,
+ "learning_rate": 0.0006,
+ "loss": 4.702625274658203,
+ "step": 4200
+ },
+ {
+ "epoch": 58.34934497816594,
+ "grad_norm": 0.028843428939580917,
+ "learning_rate": 0.0006,
+ "loss": 4.811799049377441,
+ "step": 4201
+ },
+ {
+ "epoch": 58.36331877729258,
+ "grad_norm": 0.026780391111969948,
+ "learning_rate": 0.0006,
+ "loss": 4.81662654876709,
+ "step": 4202
+ },
+ {
+ "epoch": 58.377292576419215,
+ "grad_norm": 0.019699757918715477,
+ "learning_rate": 0.0006,
+ "loss": 4.713685989379883,
+ "step": 4203
+ },
+ {
+ "epoch": 58.391266375545854,
+ "grad_norm": 0.024389250203967094,
+ "learning_rate": 0.0006,
+ "loss": 4.72354793548584,
+ "step": 4204
+ },
+ {
+ "epoch": 58.40524017467249,
+ "grad_norm": 0.02218274027109146,
+ "learning_rate": 0.0006,
+ "loss": 4.674125671386719,
+ "step": 4205
+ },
+ {
+ "epoch": 58.419213973799124,
+ "grad_norm": 0.02041488140821457,
+ "learning_rate": 0.0006,
+ "loss": 4.7909698486328125,
+ "step": 4206
+ },
+ {
+ "epoch": 58.43318777292576,
+ "grad_norm": 0.018561935052275658,
+ "learning_rate": 0.0006,
+ "loss": 4.808770179748535,
+ "step": 4207
+ },
+ {
+ "epoch": 58.4471615720524,
+ "grad_norm": 0.01708846725523472,
+ "learning_rate": 0.0006,
+ "loss": 4.701626777648926,
+ "step": 4208
+ },
+ {
+ "epoch": 58.46113537117904,
+ "grad_norm": 0.0222734697163105,
+ "learning_rate": 0.0006,
+ "loss": 4.505410194396973,
+ "step": 4209
+ },
+ {
+ "epoch": 58.47510917030568,
+ "grad_norm": 0.03197057917714119,
+ "learning_rate": 0.0006,
+ "loss": 4.610360145568848,
+ "step": 4210
+ },
+ {
+ "epoch": 58.48908296943232,
+ "grad_norm": 0.06854520738124847,
+ "learning_rate": 0.0006,
+ "loss": 4.625649452209473,
+ "step": 4211
+ },
+ {
+ "epoch": 58.50305676855895,
+ "grad_norm": 0.09981521219015121,
+ "learning_rate": 0.0006,
+ "loss": 4.832150459289551,
+ "step": 4212
+ },
+ {
+ "epoch": 58.51703056768559,
+ "grad_norm": 0.046356040984392166,
+ "learning_rate": 0.0006,
+ "loss": 4.5641632080078125,
+ "step": 4213
+ },
+ {
+ "epoch": 58.531004366812226,
+ "grad_norm": 0.03882667422294617,
+ "learning_rate": 0.0006,
+ "loss": 4.666974067687988,
+ "step": 4214
+ },
+ {
+ "epoch": 58.544978165938865,
+ "grad_norm": 0.033203233033418655,
+ "learning_rate": 0.0006,
+ "loss": 4.629413604736328,
+ "step": 4215
+ },
+ {
+ "epoch": 58.5589519650655,
+ "grad_norm": 0.02806735597550869,
+ "learning_rate": 0.0006,
+ "loss": 4.763245582580566,
+ "step": 4216
+ },
+ {
+ "epoch": 58.57292576419214,
+ "grad_norm": 0.026728278025984764,
+ "learning_rate": 0.0006,
+ "loss": 4.629597187042236,
+ "step": 4217
+ },
+ {
+ "epoch": 58.58689956331878,
+ "grad_norm": 0.024380959570407867,
+ "learning_rate": 0.0006,
+ "loss": 4.633614540100098,
+ "step": 4218
+ },
+ {
+ "epoch": 58.60087336244541,
+ "grad_norm": 0.022312544286251068,
+ "learning_rate": 0.0006,
+ "loss": 4.671379566192627,
+ "step": 4219
+ },
+ {
+ "epoch": 58.61484716157205,
+ "grad_norm": 0.022049568593502045,
+ "learning_rate": 0.0006,
+ "loss": 4.661557197570801,
+ "step": 4220
+ },
+ {
+ "epoch": 58.62882096069869,
+ "grad_norm": 0.02119818702340126,
+ "learning_rate": 0.0006,
+ "loss": 4.688562393188477,
+ "step": 4221
+ },
+ {
+ "epoch": 58.64279475982533,
+ "grad_norm": 0.019769301638007164,
+ "learning_rate": 0.0006,
+ "loss": 4.607748031616211,
+ "step": 4222
+ },
+ {
+ "epoch": 58.65676855895197,
+ "grad_norm": 0.02124079130589962,
+ "learning_rate": 0.0006,
+ "loss": 4.570640563964844,
+ "step": 4223
+ },
+ {
+ "epoch": 58.670742358078606,
+ "grad_norm": 0.016344040632247925,
+ "learning_rate": 0.0006,
+ "loss": 4.676113128662109,
+ "step": 4224
+ },
+ {
+ "epoch": 58.68471615720524,
+ "grad_norm": 0.01735256239771843,
+ "learning_rate": 0.0006,
+ "loss": 4.62088680267334,
+ "step": 4225
+ },
+ {
+ "epoch": 58.698689956331876,
+ "grad_norm": 0.019323458895087242,
+ "learning_rate": 0.0006,
+ "loss": 4.579405307769775,
+ "step": 4226
+ },
+ {
+ "epoch": 58.712663755458514,
+ "grad_norm": 0.01760544814169407,
+ "learning_rate": 0.0006,
+ "loss": 4.625407695770264,
+ "step": 4227
+ },
+ {
+ "epoch": 58.72663755458515,
+ "grad_norm": 0.015593250282108784,
+ "learning_rate": 0.0006,
+ "loss": 4.6326117515563965,
+ "step": 4228
+ },
+ {
+ "epoch": 58.74061135371179,
+ "grad_norm": 0.014357523061335087,
+ "learning_rate": 0.0006,
+ "loss": 4.6869401931762695,
+ "step": 4229
+ },
+ {
+ "epoch": 58.75458515283843,
+ "grad_norm": 0.01604357920587063,
+ "learning_rate": 0.0006,
+ "loss": 4.690784931182861,
+ "step": 4230
+ },
+ {
+ "epoch": 58.76855895196506,
+ "grad_norm": 0.01673845760524273,
+ "learning_rate": 0.0006,
+ "loss": 4.513410568237305,
+ "step": 4231
+ },
+ {
+ "epoch": 58.7825327510917,
+ "grad_norm": 0.013657779432833195,
+ "learning_rate": 0.0006,
+ "loss": 4.637057781219482,
+ "step": 4232
+ },
+ {
+ "epoch": 58.79650655021834,
+ "grad_norm": 0.013373114168643951,
+ "learning_rate": 0.0006,
+ "loss": 4.719967842102051,
+ "step": 4233
+ },
+ {
+ "epoch": 58.81048034934498,
+ "grad_norm": 0.016345568001270294,
+ "learning_rate": 0.0006,
+ "loss": 4.604109764099121,
+ "step": 4234
+ },
+ {
+ "epoch": 58.82445414847162,
+ "grad_norm": 0.01573033444583416,
+ "learning_rate": 0.0006,
+ "loss": 4.6471428871154785,
+ "step": 4235
+ },
+ {
+ "epoch": 58.838427947598255,
+ "grad_norm": 0.01517449039965868,
+ "learning_rate": 0.0006,
+ "loss": 4.722684860229492,
+ "step": 4236
+ },
+ {
+ "epoch": 58.852401746724894,
+ "grad_norm": 0.0145102022215724,
+ "learning_rate": 0.0006,
+ "loss": 4.662420272827148,
+ "step": 4237
+ },
+ {
+ "epoch": 58.866375545851525,
+ "grad_norm": 0.01457217987626791,
+ "learning_rate": 0.0006,
+ "loss": 4.585169792175293,
+ "step": 4238
+ },
+ {
+ "epoch": 58.880349344978164,
+ "grad_norm": 0.01406773366034031,
+ "learning_rate": 0.0006,
+ "loss": 4.5413665771484375,
+ "step": 4239
+ },
+ {
+ "epoch": 58.8943231441048,
+ "grad_norm": 0.012653871439397335,
+ "learning_rate": 0.0006,
+ "loss": 4.553636074066162,
+ "step": 4240
+ },
+ {
+ "epoch": 58.90829694323144,
+ "grad_norm": 0.01261200476437807,
+ "learning_rate": 0.0006,
+ "loss": 4.534193992614746,
+ "step": 4241
+ },
+ {
+ "epoch": 58.92227074235808,
+ "grad_norm": 0.012955864891409874,
+ "learning_rate": 0.0006,
+ "loss": 4.697248458862305,
+ "step": 4242
+ },
+ {
+ "epoch": 58.93624454148472,
+ "grad_norm": 0.012134749442338943,
+ "learning_rate": 0.0006,
+ "loss": 4.557095527648926,
+ "step": 4243
+ },
+ {
+ "epoch": 58.95021834061135,
+ "grad_norm": 0.012164677493274212,
+ "learning_rate": 0.0006,
+ "loss": 4.623579025268555,
+ "step": 4244
+ },
+ {
+ "epoch": 58.96419213973799,
+ "grad_norm": 0.013802947476506233,
+ "learning_rate": 0.0006,
+ "loss": 4.562838077545166,
+ "step": 4245
+ },
+ {
+ "epoch": 58.97816593886463,
+ "grad_norm": 0.011420476250350475,
+ "learning_rate": 0.0006,
+ "loss": 4.62265682220459,
+ "step": 4246
+ },
+ {
+ "epoch": 58.992139737991266,
+ "grad_norm": 0.01169576682150364,
+ "learning_rate": 0.0006,
+ "loss": 4.727449417114258,
+ "step": 4247
+ },
+ {
+ "epoch": 59.0,
+ "grad_norm": 0.013577710837125778,
+ "learning_rate": 0.0006,
+ "loss": 4.536325454711914,
+ "step": 4248
+ },
+ {
+ "epoch": 59.0,
+ "eval_loss": 4.86527681350708,
+ "eval_runtime": 60.4463,
+ "eval_samples_per_second": 40.399,
+ "eval_steps_per_second": 1.274,
+ "step": 4248
+ },
+ {
+ "epoch": 59.01397379912664,
+ "grad_norm": 0.015251655131578445,
+ "learning_rate": 0.0006,
+ "loss": 4.634598731994629,
+ "step": 4249
+ },
+ {
+ "epoch": 59.02794759825328,
+ "grad_norm": 0.017414938658475876,
+ "learning_rate": 0.0006,
+ "loss": 4.521329879760742,
+ "step": 4250
+ },
+ {
+ "epoch": 59.041921397379916,
+ "grad_norm": 0.01656588353216648,
+ "learning_rate": 0.0006,
+ "loss": 4.446192264556885,
+ "step": 4251
+ },
+ {
+ "epoch": 59.05589519650655,
+ "grad_norm": 0.012038925662636757,
+ "learning_rate": 0.0006,
+ "loss": 4.523855209350586,
+ "step": 4252
+ },
+ {
+ "epoch": 59.069868995633186,
+ "grad_norm": 0.014403033070266247,
+ "learning_rate": 0.0006,
+ "loss": 4.561079978942871,
+ "step": 4253
+ },
+ {
+ "epoch": 59.083842794759825,
+ "grad_norm": 0.016243597492575645,
+ "learning_rate": 0.0006,
+ "loss": 4.673602104187012,
+ "step": 4254
+ },
+ {
+ "epoch": 59.09781659388646,
+ "grad_norm": 0.016492877155542374,
+ "learning_rate": 0.0006,
+ "loss": 4.564824104309082,
+ "step": 4255
+ },
+ {
+ "epoch": 59.1117903930131,
+ "grad_norm": 0.015980370342731476,
+ "learning_rate": 0.0006,
+ "loss": 4.339249134063721,
+ "step": 4256
+ },
+ {
+ "epoch": 59.12576419213974,
+ "grad_norm": 0.019687356427311897,
+ "learning_rate": 0.0006,
+ "loss": 4.481834888458252,
+ "step": 4257
+ },
+ {
+ "epoch": 59.13973799126637,
+ "grad_norm": 0.02983798086643219,
+ "learning_rate": 0.0006,
+ "loss": 4.4495697021484375,
+ "step": 4258
+ },
+ {
+ "epoch": 59.15371179039301,
+ "grad_norm": 0.04600981995463371,
+ "learning_rate": 0.0006,
+ "loss": 4.567838668823242,
+ "step": 4259
+ },
+ {
+ "epoch": 59.16768558951965,
+ "grad_norm": 0.04852001741528511,
+ "learning_rate": 0.0006,
+ "loss": 4.594701766967773,
+ "step": 4260
+ },
+ {
+ "epoch": 59.18165938864629,
+ "grad_norm": 0.03429504856467247,
+ "learning_rate": 0.0006,
+ "loss": 4.573735237121582,
+ "step": 4261
+ },
+ {
+ "epoch": 59.19563318777293,
+ "grad_norm": 0.029345160350203514,
+ "learning_rate": 0.0006,
+ "loss": 4.488258361816406,
+ "step": 4262
+ },
+ {
+ "epoch": 59.209606986899566,
+ "grad_norm": 0.01961623504757881,
+ "learning_rate": 0.0006,
+ "loss": 4.545222282409668,
+ "step": 4263
+ },
+ {
+ "epoch": 59.223580786026204,
+ "grad_norm": 0.020406130701303482,
+ "learning_rate": 0.0006,
+ "loss": 4.471898078918457,
+ "step": 4264
+ },
+ {
+ "epoch": 59.237554585152836,
+ "grad_norm": 0.020500272512435913,
+ "learning_rate": 0.0006,
+ "loss": 4.538337707519531,
+ "step": 4265
+ },
+ {
+ "epoch": 59.251528384279474,
+ "grad_norm": 0.020727017894387245,
+ "learning_rate": 0.0006,
+ "loss": 4.437284469604492,
+ "step": 4266
+ },
+ {
+ "epoch": 59.26550218340611,
+ "grad_norm": 0.03395998477935791,
+ "learning_rate": 0.0006,
+ "loss": 4.459124565124512,
+ "step": 4267
+ },
+ {
+ "epoch": 59.27947598253275,
+ "grad_norm": 0.04635605961084366,
+ "learning_rate": 0.0006,
+ "loss": 4.520228385925293,
+ "step": 4268
+ },
+ {
+ "epoch": 59.29344978165939,
+ "grad_norm": 0.04253006726503372,
+ "learning_rate": 0.0006,
+ "loss": 4.578673362731934,
+ "step": 4269
+ },
+ {
+ "epoch": 59.30742358078603,
+ "grad_norm": 0.02403509058058262,
+ "learning_rate": 0.0006,
+ "loss": 4.484639644622803,
+ "step": 4270
+ },
+ {
+ "epoch": 59.32139737991266,
+ "grad_norm": 0.02254190482199192,
+ "learning_rate": 0.0006,
+ "loss": 4.511314868927002,
+ "step": 4271
+ },
+ {
+ "epoch": 59.3353711790393,
+ "grad_norm": 0.021585190668702126,
+ "learning_rate": 0.0006,
+ "loss": 4.530239105224609,
+ "step": 4272
+ },
+ {
+ "epoch": 59.34934497816594,
+ "grad_norm": 0.021210316568613052,
+ "learning_rate": 0.0006,
+ "loss": 4.593514442443848,
+ "step": 4273
+ },
+ {
+ "epoch": 59.36331877729258,
+ "grad_norm": 0.019317107275128365,
+ "learning_rate": 0.0006,
+ "loss": 4.5591535568237305,
+ "step": 4274
+ },
+ {
+ "epoch": 59.377292576419215,
+ "grad_norm": 0.018059708178043365,
+ "learning_rate": 0.0006,
+ "loss": 4.4499592781066895,
+ "step": 4275
+ },
+ {
+ "epoch": 59.391266375545854,
+ "grad_norm": 0.017847446724772453,
+ "learning_rate": 0.0006,
+ "loss": 4.473999500274658,
+ "step": 4276
+ },
+ {
+ "epoch": 59.40524017467249,
+ "grad_norm": 0.014944756403565407,
+ "learning_rate": 0.0006,
+ "loss": 4.5934529304504395,
+ "step": 4277
+ },
+ {
+ "epoch": 59.419213973799124,
+ "grad_norm": 0.01663832925260067,
+ "learning_rate": 0.0006,
+ "loss": 4.504262924194336,
+ "step": 4278
+ },
+ {
+ "epoch": 59.43318777292576,
+ "grad_norm": 0.015571796335279942,
+ "learning_rate": 0.0006,
+ "loss": 4.512310981750488,
+ "step": 4279
+ },
+ {
+ "epoch": 59.4471615720524,
+ "grad_norm": 0.015896400436758995,
+ "learning_rate": 0.0006,
+ "loss": 4.452714920043945,
+ "step": 4280
+ },
+ {
+ "epoch": 59.46113537117904,
+ "grad_norm": 0.016010284423828125,
+ "learning_rate": 0.0006,
+ "loss": 4.6188063621521,
+ "step": 4281
+ },
+ {
+ "epoch": 59.47510917030568,
+ "grad_norm": 0.015075593255460262,
+ "learning_rate": 0.0006,
+ "loss": 4.477978229522705,
+ "step": 4282
+ },
+ {
+ "epoch": 59.48908296943232,
+ "grad_norm": 0.014698950573801994,
+ "learning_rate": 0.0006,
+ "loss": 4.544955730438232,
+ "step": 4283
+ },
+ {
+ "epoch": 59.50305676855895,
+ "grad_norm": 0.0136245833709836,
+ "learning_rate": 0.0006,
+ "loss": 4.530460357666016,
+ "step": 4284
+ },
+ {
+ "epoch": 59.51703056768559,
+ "grad_norm": 0.013118310831487179,
+ "learning_rate": 0.0006,
+ "loss": 4.614217758178711,
+ "step": 4285
+ },
+ {
+ "epoch": 59.531004366812226,
+ "grad_norm": 0.01295961532741785,
+ "learning_rate": 0.0006,
+ "loss": 4.43156099319458,
+ "step": 4286
+ },
+ {
+ "epoch": 59.544978165938865,
+ "grad_norm": 0.012429811991751194,
+ "learning_rate": 0.0006,
+ "loss": 4.517333030700684,
+ "step": 4287
+ },
+ {
+ "epoch": 59.5589519650655,
+ "grad_norm": 0.01261830423027277,
+ "learning_rate": 0.0006,
+ "loss": 4.569780349731445,
+ "step": 4288
+ },
+ {
+ "epoch": 59.57292576419214,
+ "grad_norm": 0.012394499965012074,
+ "learning_rate": 0.0006,
+ "loss": 4.5970869064331055,
+ "step": 4289
+ },
+ {
+ "epoch": 59.58689956331878,
+ "grad_norm": 0.01290759164839983,
+ "learning_rate": 0.0006,
+ "loss": 4.437100410461426,
+ "step": 4290
+ },
+ {
+ "epoch": 59.60087336244541,
+ "grad_norm": 0.012042169459164143,
+ "learning_rate": 0.0006,
+ "loss": 4.567404747009277,
+ "step": 4291
+ },
+ {
+ "epoch": 59.61484716157205,
+ "grad_norm": 0.012338520959019661,
+ "learning_rate": 0.0006,
+ "loss": 4.476193428039551,
+ "step": 4292
+ },
+ {
+ "epoch": 59.62882096069869,
+ "grad_norm": 0.01321258582174778,
+ "learning_rate": 0.0006,
+ "loss": 4.566680431365967,
+ "step": 4293
+ },
+ {
+ "epoch": 59.64279475982533,
+ "grad_norm": 0.013478122651576996,
+ "learning_rate": 0.0006,
+ "loss": 4.487398147583008,
+ "step": 4294
+ },
+ {
+ "epoch": 59.65676855895197,
+ "grad_norm": 0.01537603884935379,
+ "learning_rate": 0.0006,
+ "loss": 4.449337005615234,
+ "step": 4295
+ },
+ {
+ "epoch": 59.670742358078606,
+ "grad_norm": 0.017140565440058708,
+ "learning_rate": 0.0006,
+ "loss": 4.5174360275268555,
+ "step": 4296
+ },
+ {
+ "epoch": 59.68471615720524,
+ "grad_norm": 0.017931461334228516,
+ "learning_rate": 0.0006,
+ "loss": 4.522104740142822,
+ "step": 4297
+ },
+ {
+ "epoch": 59.698689956331876,
+ "grad_norm": 0.018037918955087662,
+ "learning_rate": 0.0006,
+ "loss": 4.496928691864014,
+ "step": 4298
+ },
+ {
+ "epoch": 59.712663755458514,
+ "grad_norm": 0.01684504561126232,
+ "learning_rate": 0.0006,
+ "loss": 4.5332417488098145,
+ "step": 4299
+ },
+ {
+ "epoch": 59.72663755458515,
+ "grad_norm": 0.016268383711576462,
+ "learning_rate": 0.0006,
+ "loss": 4.468893051147461,
+ "step": 4300
+ },
+ {
+ "epoch": 59.74061135371179,
+ "grad_norm": 0.014859385788440704,
+ "learning_rate": 0.0006,
+ "loss": 4.551092147827148,
+ "step": 4301
+ },
+ {
+ "epoch": 59.75458515283843,
+ "grad_norm": 0.013589047826826572,
+ "learning_rate": 0.0006,
+ "loss": 4.470579147338867,
+ "step": 4302
+ },
+ {
+ "epoch": 59.76855895196506,
+ "grad_norm": 0.015704551711678505,
+ "learning_rate": 0.0006,
+ "loss": 4.330068111419678,
+ "step": 4303
+ },
+ {
+ "epoch": 59.7825327510917,
+ "grad_norm": 0.017869921401143074,
+ "learning_rate": 0.0006,
+ "loss": 4.428255081176758,
+ "step": 4304
+ },
+ {
+ "epoch": 59.79650655021834,
+ "grad_norm": 0.019828537479043007,
+ "learning_rate": 0.0006,
+ "loss": 4.496023178100586,
+ "step": 4305
+ },
+ {
+ "epoch": 59.81048034934498,
+ "grad_norm": 0.015856236219406128,
+ "learning_rate": 0.0006,
+ "loss": 4.427152156829834,
+ "step": 4306
+ },
+ {
+ "epoch": 59.82445414847162,
+ "grad_norm": 0.013349304907023907,
+ "learning_rate": 0.0006,
+ "loss": 4.417333602905273,
+ "step": 4307
+ },
+ {
+ "epoch": 59.838427947598255,
+ "grad_norm": 0.014111637137830257,
+ "learning_rate": 0.0006,
+ "loss": 4.4507575035095215,
+ "step": 4308
+ },
+ {
+ "epoch": 59.852401746724894,
+ "grad_norm": 0.015186597593128681,
+ "learning_rate": 0.0006,
+ "loss": 4.375116348266602,
+ "step": 4309
+ },
+ {
+ "epoch": 59.866375545851525,
+ "grad_norm": 0.01604730449616909,
+ "learning_rate": 0.0006,
+ "loss": 4.604518890380859,
+ "step": 4310
+ },
+ {
+ "epoch": 59.880349344978164,
+ "grad_norm": 0.015308464877307415,
+ "learning_rate": 0.0006,
+ "loss": 4.507232189178467,
+ "step": 4311
+ },
+ {
+ "epoch": 59.8943231441048,
+ "grad_norm": 0.014017206616699696,
+ "learning_rate": 0.0006,
+ "loss": 4.398487091064453,
+ "step": 4312
+ },
+ {
+ "epoch": 59.90829694323144,
+ "grad_norm": 0.012911394238471985,
+ "learning_rate": 0.0006,
+ "loss": 4.567872047424316,
+ "step": 4313
+ },
+ {
+ "epoch": 59.92227074235808,
+ "grad_norm": 0.012775142677128315,
+ "learning_rate": 0.0006,
+ "loss": 4.562105655670166,
+ "step": 4314
+ },
+ {
+ "epoch": 59.93624454148472,
+ "grad_norm": 0.014750408008694649,
+ "learning_rate": 0.0006,
+ "loss": 4.456263542175293,
+ "step": 4315
+ },
+ {
+ "epoch": 59.95021834061135,
+ "grad_norm": 0.016726011410355568,
+ "learning_rate": 0.0006,
+ "loss": 4.360722064971924,
+ "step": 4316
+ },
+ {
+ "epoch": 59.96419213973799,
+ "grad_norm": 0.01734776981174946,
+ "learning_rate": 0.0006,
+ "loss": 4.572544574737549,
+ "step": 4317
+ },
+ {
+ "epoch": 59.97816593886463,
+ "grad_norm": 0.01822415366768837,
+ "learning_rate": 0.0006,
+ "loss": 4.490501880645752,
+ "step": 4318
+ },
+ {
+ "epoch": 59.992139737991266,
+ "grad_norm": 0.014638577587902546,
+ "learning_rate": 0.0006,
+ "loss": 4.440951824188232,
+ "step": 4319
+ },
+ {
+ "epoch": 60.0,
+ "grad_norm": 0.01255202479660511,
+ "learning_rate": 0.0006,
+ "loss": 4.559123516082764,
+ "step": 4320
+ },
+ {
+ "epoch": 60.0,
+ "eval_loss": 4.741089344024658,
+ "eval_runtime": 56.8997,
+ "eval_samples_per_second": 42.918,
+ "eval_steps_per_second": 1.353,
+ "step": 4320
+ },
+ {
+ "epoch": 60.01397379912664,
+ "grad_norm": 0.016548393294215202,
+ "learning_rate": 0.0006,
+ "loss": 4.442024230957031,
+ "step": 4321
+ },
+ {
+ "epoch": 60.02794759825328,
+ "grad_norm": 0.019759127870202065,
+ "learning_rate": 0.0006,
+ "loss": 4.394334316253662,
+ "step": 4322
+ },
+ {
+ "epoch": 60.041921397379916,
+ "grad_norm": 0.018325170502066612,
+ "learning_rate": 0.0006,
+ "loss": 4.318270683288574,
+ "step": 4323
+ },
+ {
+ "epoch": 60.05589519650655,
+ "grad_norm": 0.015289463102817535,
+ "learning_rate": 0.0006,
+ "loss": 4.318228721618652,
+ "step": 4324
+ },
+ {
+ "epoch": 60.069868995633186,
+ "grad_norm": 0.015424872748553753,
+ "learning_rate": 0.0006,
+ "loss": 4.3605499267578125,
+ "step": 4325
+ },
+ {
+ "epoch": 60.083842794759825,
+ "grad_norm": 0.014085669070482254,
+ "learning_rate": 0.0006,
+ "loss": 4.359585762023926,
+ "step": 4326
+ },
+ {
+ "epoch": 60.09781659388646,
+ "grad_norm": 0.014004210010170937,
+ "learning_rate": 0.0006,
+ "loss": 4.447783470153809,
+ "step": 4327
+ },
+ {
+ "epoch": 60.1117903930131,
+ "grad_norm": 0.013514582999050617,
+ "learning_rate": 0.0006,
+ "loss": 4.4850239753723145,
+ "step": 4328
+ },
+ {
+ "epoch": 60.12576419213974,
+ "grad_norm": 0.013049394823610783,
+ "learning_rate": 0.0006,
+ "loss": 4.528675556182861,
+ "step": 4329
+ },
+ {
+ "epoch": 60.13973799126637,
+ "grad_norm": 0.01378058921545744,
+ "learning_rate": 0.0006,
+ "loss": 4.444823741912842,
+ "step": 4330
+ },
+ {
+ "epoch": 60.15371179039301,
+ "grad_norm": 0.013752233237028122,
+ "learning_rate": 0.0006,
+ "loss": 4.5205206871032715,
+ "step": 4331
+ },
+ {
+ "epoch": 60.16768558951965,
+ "grad_norm": 0.013431803323328495,
+ "learning_rate": 0.0006,
+ "loss": 4.547320365905762,
+ "step": 4332
+ },
+ {
+ "epoch": 60.18165938864629,
+ "grad_norm": 0.014544196426868439,
+ "learning_rate": 0.0006,
+ "loss": 4.517143249511719,
+ "step": 4333
+ },
+ {
+ "epoch": 60.19563318777293,
+ "grad_norm": 0.01612176187336445,
+ "learning_rate": 0.0006,
+ "loss": 4.517232894897461,
+ "step": 4334
+ },
+ {
+ "epoch": 60.209606986899566,
+ "grad_norm": 0.01803239807486534,
+ "learning_rate": 0.0006,
+ "loss": 4.377574920654297,
+ "step": 4335
+ },
+ {
+ "epoch": 60.223580786026204,
+ "grad_norm": 0.014871489256620407,
+ "learning_rate": 0.0006,
+ "loss": 4.538285255432129,
+ "step": 4336
+ },
+ {
+ "epoch": 60.237554585152836,
+ "grad_norm": 0.013927377760410309,
+ "learning_rate": 0.0006,
+ "loss": 4.386175632476807,
+ "step": 4337
+ },
+ {
+ "epoch": 60.251528384279474,
+ "grad_norm": 0.014757219702005386,
+ "learning_rate": 0.0006,
+ "loss": 4.528553009033203,
+ "step": 4338
+ },
+ {
+ "epoch": 60.26550218340611,
+ "grad_norm": 0.01433873176574707,
+ "learning_rate": 0.0006,
+ "loss": 4.476929187774658,
+ "step": 4339
+ },
+ {
+ "epoch": 60.27947598253275,
+ "grad_norm": 0.015768803656101227,
+ "learning_rate": 0.0006,
+ "loss": 4.50970458984375,
+ "step": 4340
+ },
+ {
+ "epoch": 60.29344978165939,
+ "grad_norm": 0.015651429072022438,
+ "learning_rate": 0.0006,
+ "loss": 4.427133083343506,
+ "step": 4341
+ },
+ {
+ "epoch": 60.30742358078603,
+ "grad_norm": 0.01673000678420067,
+ "learning_rate": 0.0006,
+ "loss": 4.447723865509033,
+ "step": 4342
+ },
+ {
+ "epoch": 60.32139737991266,
+ "grad_norm": 0.018481194972991943,
+ "learning_rate": 0.0006,
+ "loss": 4.559832572937012,
+ "step": 4343
+ },
+ {
+ "epoch": 60.3353711790393,
+ "grad_norm": 0.016731027513742447,
+ "learning_rate": 0.0006,
+ "loss": 4.470893859863281,
+ "step": 4344
+ },
+ {
+ "epoch": 60.34934497816594,
+ "grad_norm": 0.01504011545330286,
+ "learning_rate": 0.0006,
+ "loss": 4.483014106750488,
+ "step": 4345
+ },
+ {
+ "epoch": 60.36331877729258,
+ "grad_norm": 0.01602456159889698,
+ "learning_rate": 0.0006,
+ "loss": 4.548428058624268,
+ "step": 4346
+ },
+ {
+ "epoch": 60.377292576419215,
+ "grad_norm": 0.016027364879846573,
+ "learning_rate": 0.0006,
+ "loss": 4.426779747009277,
+ "step": 4347
+ },
+ {
+ "epoch": 60.391266375545854,
+ "grad_norm": 0.01873827911913395,
+ "learning_rate": 0.0006,
+ "loss": 4.353686332702637,
+ "step": 4348
+ },
+ {
+ "epoch": 60.40524017467249,
+ "grad_norm": 0.01907406374812126,
+ "learning_rate": 0.0006,
+ "loss": 4.502945423126221,
+ "step": 4349
+ },
+ {
+ "epoch": 60.419213973799124,
+ "grad_norm": 0.016083354130387306,
+ "learning_rate": 0.0006,
+ "loss": 4.431731700897217,
+ "step": 4350
+ },
+ {
+ "epoch": 60.43318777292576,
+ "grad_norm": 0.015293709933757782,
+ "learning_rate": 0.0006,
+ "loss": 4.480112552642822,
+ "step": 4351
+ },
+ {
+ "epoch": 60.4471615720524,
+ "grad_norm": 0.01688101328909397,
+ "learning_rate": 0.0006,
+ "loss": 4.50853157043457,
+ "step": 4352
+ },
+ {
+ "epoch": 60.46113537117904,
+ "grad_norm": 0.02094118855893612,
+ "learning_rate": 0.0006,
+ "loss": 4.359930038452148,
+ "step": 4353
+ },
+ {
+ "epoch": 60.47510917030568,
+ "grad_norm": 0.018188107758760452,
+ "learning_rate": 0.0006,
+ "loss": 4.549592018127441,
+ "step": 4354
+ },
+ {
+ "epoch": 60.48908296943232,
+ "grad_norm": 0.01462319865822792,
+ "learning_rate": 0.0006,
+ "loss": 4.49467658996582,
+ "step": 4355
+ },
+ {
+ "epoch": 60.50305676855895,
+ "grad_norm": 0.017508579418063164,
+ "learning_rate": 0.0006,
+ "loss": 4.398318767547607,
+ "step": 4356
+ },
+ {
+ "epoch": 60.51703056768559,
+ "grad_norm": 0.018200315535068512,
+ "learning_rate": 0.0006,
+ "loss": 4.44169282913208,
+ "step": 4357
+ },
+ {
+ "epoch": 60.531004366812226,
+ "grad_norm": 0.01717468351125717,
+ "learning_rate": 0.0006,
+ "loss": 4.481400489807129,
+ "step": 4358
+ },
+ {
+ "epoch": 60.544978165938865,
+ "grad_norm": 0.019329151138663292,
+ "learning_rate": 0.0006,
+ "loss": 4.578839302062988,
+ "step": 4359
+ },
+ {
+ "epoch": 60.5589519650655,
+ "grad_norm": 0.016109555959701538,
+ "learning_rate": 0.0006,
+ "loss": 4.461043357849121,
+ "step": 4360
+ },
+ {
+ "epoch": 60.57292576419214,
+ "grad_norm": 0.01494457945227623,
+ "learning_rate": 0.0006,
+ "loss": 4.505096912384033,
+ "step": 4361
+ },
+ {
+ "epoch": 60.58689956331878,
+ "grad_norm": 0.01729399710893631,
+ "learning_rate": 0.0006,
+ "loss": 4.366747856140137,
+ "step": 4362
+ },
+ {
+ "epoch": 60.60087336244541,
+ "grad_norm": 0.01705821603536606,
+ "learning_rate": 0.0006,
+ "loss": 4.42534065246582,
+ "step": 4363
+ },
+ {
+ "epoch": 60.61484716157205,
+ "grad_norm": 0.018391354009509087,
+ "learning_rate": 0.0006,
+ "loss": 4.350294589996338,
+ "step": 4364
+ },
+ {
+ "epoch": 60.62882096069869,
+ "grad_norm": 0.016989829018712044,
+ "learning_rate": 0.0006,
+ "loss": 4.45905876159668,
+ "step": 4365
+ },
+ {
+ "epoch": 60.64279475982533,
+ "grad_norm": 0.017192212864756584,
+ "learning_rate": 0.0006,
+ "loss": 4.526487350463867,
+ "step": 4366
+ },
+ {
+ "epoch": 60.65676855895197,
+ "grad_norm": 0.018091315403580666,
+ "learning_rate": 0.0006,
+ "loss": 4.309333324432373,
+ "step": 4367
+ },
+ {
+ "epoch": 60.670742358078606,
+ "grad_norm": 0.014051459729671478,
+ "learning_rate": 0.0006,
+ "loss": 4.338171005249023,
+ "step": 4368
+ },
+ {
+ "epoch": 60.68471615720524,
+ "grad_norm": 0.013166418299078941,
+ "learning_rate": 0.0006,
+ "loss": 4.409758567810059,
+ "step": 4369
+ },
+ {
+ "epoch": 60.698689956331876,
+ "grad_norm": 0.014260428957641125,
+ "learning_rate": 0.0006,
+ "loss": 4.524338245391846,
+ "step": 4370
+ },
+ {
+ "epoch": 60.712663755458514,
+ "grad_norm": 0.017019255086779594,
+ "learning_rate": 0.0006,
+ "loss": 4.387436866760254,
+ "step": 4371
+ },
+ {
+ "epoch": 60.72663755458515,
+ "grad_norm": 0.017070675268769264,
+ "learning_rate": 0.0006,
+ "loss": 4.437321662902832,
+ "step": 4372
+ },
+ {
+ "epoch": 60.74061135371179,
+ "grad_norm": 0.015197164379060268,
+ "learning_rate": 0.0006,
+ "loss": 4.371205806732178,
+ "step": 4373
+ },
+ {
+ "epoch": 60.75458515283843,
+ "grad_norm": 0.014791185967624187,
+ "learning_rate": 0.0006,
+ "loss": 4.44439697265625,
+ "step": 4374
+ },
+ {
+ "epoch": 60.76855895196506,
+ "grad_norm": 0.016458261758089066,
+ "learning_rate": 0.0006,
+ "loss": 4.445584297180176,
+ "step": 4375
+ },
+ {
+ "epoch": 60.7825327510917,
+ "grad_norm": 0.017706342041492462,
+ "learning_rate": 0.0006,
+ "loss": 4.348708152770996,
+ "step": 4376
+ },
+ {
+ "epoch": 60.79650655021834,
+ "grad_norm": 0.017561476677656174,
+ "learning_rate": 0.0006,
+ "loss": 4.521929740905762,
+ "step": 4377
+ },
+ {
+ "epoch": 60.81048034934498,
+ "grad_norm": 0.016557445749640465,
+ "learning_rate": 0.0006,
+ "loss": 4.545806884765625,
+ "step": 4378
+ },
+ {
+ "epoch": 60.82445414847162,
+ "grad_norm": 0.016953222453594208,
+ "learning_rate": 0.0006,
+ "loss": 4.4230875968933105,
+ "step": 4379
+ },
+ {
+ "epoch": 60.838427947598255,
+ "grad_norm": 0.017094967886805534,
+ "learning_rate": 0.0006,
+ "loss": 4.423985481262207,
+ "step": 4380
+ },
+ {
+ "epoch": 60.852401746724894,
+ "grad_norm": 0.014787515625357628,
+ "learning_rate": 0.0006,
+ "loss": 4.39473819732666,
+ "step": 4381
+ },
+ {
+ "epoch": 60.866375545851525,
+ "grad_norm": 0.016208263114094734,
+ "learning_rate": 0.0006,
+ "loss": 4.3810038566589355,
+ "step": 4382
+ },
+ {
+ "epoch": 60.880349344978164,
+ "grad_norm": 0.017844321206212044,
+ "learning_rate": 0.0006,
+ "loss": 4.35086727142334,
+ "step": 4383
+ },
+ {
+ "epoch": 60.8943231441048,
+ "grad_norm": 0.017539294436573982,
+ "learning_rate": 0.0006,
+ "loss": 4.460260391235352,
+ "step": 4384
+ },
+ {
+ "epoch": 60.90829694323144,
+ "grad_norm": 0.01612042263150215,
+ "learning_rate": 0.0006,
+ "loss": 4.370500564575195,
+ "step": 4385
+ },
+ {
+ "epoch": 60.92227074235808,
+ "grad_norm": 0.015381601639091969,
+ "learning_rate": 0.0006,
+ "loss": 4.435868263244629,
+ "step": 4386
+ },
+ {
+ "epoch": 60.93624454148472,
+ "grad_norm": 0.01603585295379162,
+ "learning_rate": 0.0006,
+ "loss": 4.434070587158203,
+ "step": 4387
+ },
+ {
+ "epoch": 60.95021834061135,
+ "grad_norm": 0.016268130391836166,
+ "learning_rate": 0.0006,
+ "loss": 4.404343128204346,
+ "step": 4388
+ },
+ {
+ "epoch": 60.96419213973799,
+ "grad_norm": 0.014980032108724117,
+ "learning_rate": 0.0006,
+ "loss": 4.399213790893555,
+ "step": 4389
+ },
+ {
+ "epoch": 60.97816593886463,
+ "grad_norm": 0.016899267211556435,
+ "learning_rate": 0.0006,
+ "loss": 4.508069038391113,
+ "step": 4390
+ },
+ {
+ "epoch": 60.992139737991266,
+ "grad_norm": 0.013756908476352692,
+ "learning_rate": 0.0006,
+ "loss": 4.541858673095703,
+ "step": 4391
+ },
+ {
+ "epoch": 61.0,
+ "grad_norm": 0.01578413136303425,
+ "learning_rate": 0.0006,
+ "loss": 4.543513298034668,
+ "step": 4392
+ },
+ {
+ "epoch": 61.0,
+ "eval_loss": 4.756770610809326,
+ "eval_runtime": 56.4572,
+ "eval_samples_per_second": 43.254,
+ "eval_steps_per_second": 1.364,
+ "step": 4392
+ },
+ {
+ "epoch": 61.01397379912664,
+ "grad_norm": 0.013387506827712059,
+ "learning_rate": 0.0006,
+ "loss": 4.475942611694336,
+ "step": 4393
+ },
+ {
+ "epoch": 61.02794759825328,
+ "grad_norm": 0.014278904534876347,
+ "learning_rate": 0.0006,
+ "loss": 4.442403793334961,
+ "step": 4394
+ },
+ {
+ "epoch": 61.041921397379916,
+ "grad_norm": 0.014359553344547749,
+ "learning_rate": 0.0006,
+ "loss": 4.509463310241699,
+ "step": 4395
+ },
+ {
+ "epoch": 61.05589519650655,
+ "grad_norm": 0.01718050241470337,
+ "learning_rate": 0.0006,
+ "loss": 4.354458808898926,
+ "step": 4396
+ },
+ {
+ "epoch": 61.069868995633186,
+ "grad_norm": 0.015988312661647797,
+ "learning_rate": 0.0006,
+ "loss": 4.484508514404297,
+ "step": 4397
+ },
+ {
+ "epoch": 61.083842794759825,
+ "grad_norm": 0.015466735698282719,
+ "learning_rate": 0.0006,
+ "loss": 4.326908111572266,
+ "step": 4398
+ },
+ {
+ "epoch": 61.09781659388646,
+ "grad_norm": 0.014541332609951496,
+ "learning_rate": 0.0006,
+ "loss": 4.405823230743408,
+ "step": 4399
+ },
+ {
+ "epoch": 61.1117903930131,
+ "grad_norm": 0.014803987927734852,
+ "learning_rate": 0.0006,
+ "loss": 4.375386714935303,
+ "step": 4400
+ },
+ {
+ "epoch": 61.12576419213974,
+ "grad_norm": 0.013962093740701675,
+ "learning_rate": 0.0006,
+ "loss": 4.300838470458984,
+ "step": 4401
+ },
+ {
+ "epoch": 61.13973799126637,
+ "grad_norm": 0.015213954262435436,
+ "learning_rate": 0.0006,
+ "loss": 4.388306617736816,
+ "step": 4402
+ },
+ {
+ "epoch": 61.15371179039301,
+ "grad_norm": 0.018965506926178932,
+ "learning_rate": 0.0006,
+ "loss": 4.388340950012207,
+ "step": 4403
+ },
+ {
+ "epoch": 61.16768558951965,
+ "grad_norm": 0.01958410255610943,
+ "learning_rate": 0.0006,
+ "loss": 4.35012149810791,
+ "step": 4404
+ },
+ {
+ "epoch": 61.18165938864629,
+ "grad_norm": 0.01938101276755333,
+ "learning_rate": 0.0006,
+ "loss": 4.368372917175293,
+ "step": 4405
+ },
+ {
+ "epoch": 61.19563318777293,
+ "grad_norm": 0.016556516289711,
+ "learning_rate": 0.0006,
+ "loss": 4.486198425292969,
+ "step": 4406
+ },
+ {
+ "epoch": 61.209606986899566,
+ "grad_norm": 0.017031289637088776,
+ "learning_rate": 0.0006,
+ "loss": 4.421676158905029,
+ "step": 4407
+ },
+ {
+ "epoch": 61.223580786026204,
+ "grad_norm": 0.018023479729890823,
+ "learning_rate": 0.0006,
+ "loss": 4.38751220703125,
+ "step": 4408
+ },
+ {
+ "epoch": 61.237554585152836,
+ "grad_norm": 0.0178135447204113,
+ "learning_rate": 0.0006,
+ "loss": 4.336528301239014,
+ "step": 4409
+ },
+ {
+ "epoch": 61.251528384279474,
+ "grad_norm": 0.016782190650701523,
+ "learning_rate": 0.0006,
+ "loss": 4.386693954467773,
+ "step": 4410
+ },
+ {
+ "epoch": 61.26550218340611,
+ "grad_norm": 0.015494848601520061,
+ "learning_rate": 0.0006,
+ "loss": 4.308948040008545,
+ "step": 4411
+ },
+ {
+ "epoch": 61.27947598253275,
+ "grad_norm": 0.015002378262579441,
+ "learning_rate": 0.0006,
+ "loss": 4.314206123352051,
+ "step": 4412
+ },
+ {
+ "epoch": 61.29344978165939,
+ "grad_norm": 0.015798920765519142,
+ "learning_rate": 0.0006,
+ "loss": 4.502900123596191,
+ "step": 4413
+ },
+ {
+ "epoch": 61.30742358078603,
+ "grad_norm": 0.015444336459040642,
+ "learning_rate": 0.0006,
+ "loss": 4.473160743713379,
+ "step": 4414
+ },
+ {
+ "epoch": 61.32139737991266,
+ "grad_norm": 0.017025096341967583,
+ "learning_rate": 0.0006,
+ "loss": 4.460621356964111,
+ "step": 4415
+ },
+ {
+ "epoch": 61.3353711790393,
+ "grad_norm": 0.019502537325024605,
+ "learning_rate": 0.0006,
+ "loss": 4.385551452636719,
+ "step": 4416
+ },
+ {
+ "epoch": 61.34934497816594,
+ "grad_norm": 0.018529541790485382,
+ "learning_rate": 0.0006,
+ "loss": 4.425390720367432,
+ "step": 4417
+ },
+ {
+ "epoch": 61.36331877729258,
+ "grad_norm": 0.013155706226825714,
+ "learning_rate": 0.0006,
+ "loss": 4.467020034790039,
+ "step": 4418
+ },
+ {
+ "epoch": 61.377292576419215,
+ "grad_norm": 0.01518090907484293,
+ "learning_rate": 0.0006,
+ "loss": 4.369664192199707,
+ "step": 4419
+ },
+ {
+ "epoch": 61.391266375545854,
+ "grad_norm": 0.01626511849462986,
+ "learning_rate": 0.0006,
+ "loss": 4.463231086730957,
+ "step": 4420
+ },
+ {
+ "epoch": 61.40524017467249,
+ "grad_norm": 0.01701558195054531,
+ "learning_rate": 0.0006,
+ "loss": 4.43708610534668,
+ "step": 4421
+ },
+ {
+ "epoch": 61.419213973799124,
+ "grad_norm": 0.01761862449347973,
+ "learning_rate": 0.0006,
+ "loss": 4.3314208984375,
+ "step": 4422
+ },
+ {
+ "epoch": 61.43318777292576,
+ "grad_norm": 0.017084648832678795,
+ "learning_rate": 0.0006,
+ "loss": 4.39676570892334,
+ "step": 4423
+ },
+ {
+ "epoch": 61.4471615720524,
+ "grad_norm": 0.01667078025639057,
+ "learning_rate": 0.0006,
+ "loss": 4.3789873123168945,
+ "step": 4424
+ },
+ {
+ "epoch": 61.46113537117904,
+ "grad_norm": 0.01877436414361,
+ "learning_rate": 0.0006,
+ "loss": 4.415275573730469,
+ "step": 4425
+ },
+ {
+ "epoch": 61.47510917030568,
+ "grad_norm": 0.020250486209988594,
+ "learning_rate": 0.0006,
+ "loss": 4.478635787963867,
+ "step": 4426
+ },
+ {
+ "epoch": 61.48908296943232,
+ "grad_norm": 0.020244868472218513,
+ "learning_rate": 0.0006,
+ "loss": 4.296854019165039,
+ "step": 4427
+ },
+ {
+ "epoch": 61.50305676855895,
+ "grad_norm": 0.018723875284194946,
+ "learning_rate": 0.0006,
+ "loss": 4.3760881423950195,
+ "step": 4428
+ },
+ {
+ "epoch": 61.51703056768559,
+ "grad_norm": 0.015138164162635803,
+ "learning_rate": 0.0006,
+ "loss": 4.39974308013916,
+ "step": 4429
+ },
+ {
+ "epoch": 61.531004366812226,
+ "grad_norm": 0.015111138112843037,
+ "learning_rate": 0.0006,
+ "loss": 4.34659481048584,
+ "step": 4430
+ },
+ {
+ "epoch": 61.544978165938865,
+ "grad_norm": 0.0164666585624218,
+ "learning_rate": 0.0006,
+ "loss": 4.415737152099609,
+ "step": 4431
+ },
+ {
+ "epoch": 61.5589519650655,
+ "grad_norm": 0.020891312509775162,
+ "learning_rate": 0.0006,
+ "loss": 4.300319671630859,
+ "step": 4432
+ },
+ {
+ "epoch": 61.57292576419214,
+ "grad_norm": 0.02089156024158001,
+ "learning_rate": 0.0006,
+ "loss": 4.373029708862305,
+ "step": 4433
+ },
+ {
+ "epoch": 61.58689956331878,
+ "grad_norm": 0.020114559680223465,
+ "learning_rate": 0.0006,
+ "loss": 4.434499740600586,
+ "step": 4434
+ },
+ {
+ "epoch": 61.60087336244541,
+ "grad_norm": 0.01791365258395672,
+ "learning_rate": 0.0006,
+ "loss": 4.466872215270996,
+ "step": 4435
+ },
+ {
+ "epoch": 61.61484716157205,
+ "grad_norm": 0.01879458874464035,
+ "learning_rate": 0.0006,
+ "loss": 4.5176520347595215,
+ "step": 4436
+ },
+ {
+ "epoch": 61.62882096069869,
+ "grad_norm": 0.016371937468647957,
+ "learning_rate": 0.0006,
+ "loss": 4.3507585525512695,
+ "step": 4437
+ },
+ {
+ "epoch": 61.64279475982533,
+ "grad_norm": 0.01648143120110035,
+ "learning_rate": 0.0006,
+ "loss": 4.398778438568115,
+ "step": 4438
+ },
+ {
+ "epoch": 61.65676855895197,
+ "grad_norm": 0.017240576446056366,
+ "learning_rate": 0.0006,
+ "loss": 4.513933181762695,
+ "step": 4439
+ },
+ {
+ "epoch": 61.670742358078606,
+ "grad_norm": 0.01734284684062004,
+ "learning_rate": 0.0006,
+ "loss": 4.315343856811523,
+ "step": 4440
+ },
+ {
+ "epoch": 61.68471615720524,
+ "grad_norm": 0.01639772206544876,
+ "learning_rate": 0.0006,
+ "loss": 4.4641218185424805,
+ "step": 4441
+ },
+ {
+ "epoch": 61.698689956331876,
+ "grad_norm": 0.0164708960801363,
+ "learning_rate": 0.0006,
+ "loss": 4.439968585968018,
+ "step": 4442
+ },
+ {
+ "epoch": 61.712663755458514,
+ "grad_norm": 0.01757933758199215,
+ "learning_rate": 0.0006,
+ "loss": 4.408327102661133,
+ "step": 4443
+ },
+ {
+ "epoch": 61.72663755458515,
+ "grad_norm": 0.01789589412510395,
+ "learning_rate": 0.0006,
+ "loss": 4.478257179260254,
+ "step": 4444
+ },
+ {
+ "epoch": 61.74061135371179,
+ "grad_norm": 0.01822184957563877,
+ "learning_rate": 0.0006,
+ "loss": 4.412046909332275,
+ "step": 4445
+ },
+ {
+ "epoch": 61.75458515283843,
+ "grad_norm": 0.018624618649482727,
+ "learning_rate": 0.0006,
+ "loss": 4.427122116088867,
+ "step": 4446
+ },
+ {
+ "epoch": 61.76855895196506,
+ "grad_norm": 0.018674887716770172,
+ "learning_rate": 0.0006,
+ "loss": 4.429147720336914,
+ "step": 4447
+ },
+ {
+ "epoch": 61.7825327510917,
+ "grad_norm": 0.015317119657993317,
+ "learning_rate": 0.0006,
+ "loss": 4.2808451652526855,
+ "step": 4448
+ },
+ {
+ "epoch": 61.79650655021834,
+ "grad_norm": 0.015400845557451248,
+ "learning_rate": 0.0006,
+ "loss": 4.420193672180176,
+ "step": 4449
+ },
+ {
+ "epoch": 61.81048034934498,
+ "grad_norm": 0.016743650659918785,
+ "learning_rate": 0.0006,
+ "loss": 4.498254776000977,
+ "step": 4450
+ },
+ {
+ "epoch": 61.82445414847162,
+ "grad_norm": 0.014330781064927578,
+ "learning_rate": 0.0006,
+ "loss": 4.445384979248047,
+ "step": 4451
+ },
+ {
+ "epoch": 61.838427947598255,
+ "grad_norm": 0.015046479180455208,
+ "learning_rate": 0.0006,
+ "loss": 4.368041038513184,
+ "step": 4452
+ },
+ {
+ "epoch": 61.852401746724894,
+ "grad_norm": 0.01630168780684471,
+ "learning_rate": 0.0006,
+ "loss": 4.460142135620117,
+ "step": 4453
+ },
+ {
+ "epoch": 61.866375545851525,
+ "grad_norm": 0.01637885719537735,
+ "learning_rate": 0.0006,
+ "loss": 4.416606426239014,
+ "step": 4454
+ },
+ {
+ "epoch": 61.880349344978164,
+ "grad_norm": 0.016785888001322746,
+ "learning_rate": 0.0006,
+ "loss": 4.452545642852783,
+ "step": 4455
+ },
+ {
+ "epoch": 61.8943231441048,
+ "grad_norm": 0.015044555068016052,
+ "learning_rate": 0.0006,
+ "loss": 4.427334785461426,
+ "step": 4456
+ },
+ {
+ "epoch": 61.90829694323144,
+ "grad_norm": 0.013699211180210114,
+ "learning_rate": 0.0006,
+ "loss": 4.408761024475098,
+ "step": 4457
+ },
+ {
+ "epoch": 61.92227074235808,
+ "grad_norm": 0.014021803624927998,
+ "learning_rate": 0.0006,
+ "loss": 4.543607711791992,
+ "step": 4458
+ },
+ {
+ "epoch": 61.93624454148472,
+ "grad_norm": 0.015438460744917393,
+ "learning_rate": 0.0006,
+ "loss": 4.313363075256348,
+ "step": 4459
+ },
+ {
+ "epoch": 61.95021834061135,
+ "grad_norm": 0.017515428364276886,
+ "learning_rate": 0.0006,
+ "loss": 4.310412406921387,
+ "step": 4460
+ },
+ {
+ "epoch": 61.96419213973799,
+ "grad_norm": 0.016683880239725113,
+ "learning_rate": 0.0006,
+ "loss": 4.3378448486328125,
+ "step": 4461
+ },
+ {
+ "epoch": 61.97816593886463,
+ "grad_norm": 0.019384048879146576,
+ "learning_rate": 0.0006,
+ "loss": 4.450568675994873,
+ "step": 4462
+ },
+ {
+ "epoch": 61.992139737991266,
+ "grad_norm": 0.020651493221521378,
+ "learning_rate": 0.0006,
+ "loss": 4.453197479248047,
+ "step": 4463
+ },
+ {
+ "epoch": 62.0,
+ "grad_norm": 0.020991764962673187,
+ "learning_rate": 0.0006,
+ "loss": 4.441527366638184,
+ "step": 4464
+ },
+ {
+ "epoch": 62.0,
+ "eval_loss": 4.700193881988525,
+ "eval_runtime": 56.6098,
+ "eval_samples_per_second": 43.137,
+ "eval_steps_per_second": 1.36,
+ "step": 4464
+ },
+ {
+ "epoch": 62.01397379912664,
+ "grad_norm": 0.01811373233795166,
+ "learning_rate": 0.0006,
+ "loss": 4.37346076965332,
+ "step": 4465
+ },
+ {
+ "epoch": 62.02794759825328,
+ "grad_norm": 0.01667127199470997,
+ "learning_rate": 0.0006,
+ "loss": 4.237151145935059,
+ "step": 4466
+ },
+ {
+ "epoch": 62.041921397379916,
+ "grad_norm": 0.014971534721553326,
+ "learning_rate": 0.0006,
+ "loss": 4.367053031921387,
+ "step": 4467
+ },
+ {
+ "epoch": 62.05589519650655,
+ "grad_norm": 0.015655893832445145,
+ "learning_rate": 0.0006,
+ "loss": 4.437426567077637,
+ "step": 4468
+ },
+ {
+ "epoch": 62.069868995633186,
+ "grad_norm": 0.014935447834432125,
+ "learning_rate": 0.0006,
+ "loss": 4.3635125160217285,
+ "step": 4469
+ },
+ {
+ "epoch": 62.083842794759825,
+ "grad_norm": 0.015363802202045918,
+ "learning_rate": 0.0006,
+ "loss": 4.40492057800293,
+ "step": 4470
+ },
+ {
+ "epoch": 62.09781659388646,
+ "grad_norm": 0.0159201230853796,
+ "learning_rate": 0.0006,
+ "loss": 4.392058849334717,
+ "step": 4471
+ },
+ {
+ "epoch": 62.1117903930131,
+ "grad_norm": 0.014598241075873375,
+ "learning_rate": 0.0006,
+ "loss": 4.369944095611572,
+ "step": 4472
+ },
+ {
+ "epoch": 62.12576419213974,
+ "grad_norm": 0.015394841320812702,
+ "learning_rate": 0.0006,
+ "loss": 4.425097942352295,
+ "step": 4473
+ },
+ {
+ "epoch": 62.13973799126637,
+ "grad_norm": 0.015913628041744232,
+ "learning_rate": 0.0006,
+ "loss": 4.420596122741699,
+ "step": 4474
+ },
+ {
+ "epoch": 62.15371179039301,
+ "grad_norm": 0.01575258933007717,
+ "learning_rate": 0.0006,
+ "loss": 4.32290506362915,
+ "step": 4475
+ },
+ {
+ "epoch": 62.16768558951965,
+ "grad_norm": 0.0155344782397151,
+ "learning_rate": 0.0006,
+ "loss": 4.273068428039551,
+ "step": 4476
+ },
+ {
+ "epoch": 62.18165938864629,
+ "grad_norm": 0.014947040006518364,
+ "learning_rate": 0.0006,
+ "loss": 4.384868621826172,
+ "step": 4477
+ },
+ {
+ "epoch": 62.19563318777293,
+ "grad_norm": 0.01557591650635004,
+ "learning_rate": 0.0006,
+ "loss": 4.458410263061523,
+ "step": 4478
+ },
+ {
+ "epoch": 62.209606986899566,
+ "grad_norm": 0.015182198956608772,
+ "learning_rate": 0.0006,
+ "loss": 4.300738334655762,
+ "step": 4479
+ },
+ {
+ "epoch": 62.223580786026204,
+ "grad_norm": 0.01634717360138893,
+ "learning_rate": 0.0006,
+ "loss": 4.436875820159912,
+ "step": 4480
+ },
+ {
+ "epoch": 62.237554585152836,
+ "grad_norm": 0.018813718110322952,
+ "learning_rate": 0.0006,
+ "loss": 4.454665184020996,
+ "step": 4481
+ },
+ {
+ "epoch": 62.251528384279474,
+ "grad_norm": 0.020965630188584328,
+ "learning_rate": 0.0006,
+ "loss": 4.357341766357422,
+ "step": 4482
+ },
+ {
+ "epoch": 62.26550218340611,
+ "grad_norm": 0.020738353952765465,
+ "learning_rate": 0.0006,
+ "loss": 4.456644058227539,
+ "step": 4483
+ },
+ {
+ "epoch": 62.27947598253275,
+ "grad_norm": 0.018813546746969223,
+ "learning_rate": 0.0006,
+ "loss": 4.381941795349121,
+ "step": 4484
+ },
+ {
+ "epoch": 62.29344978165939,
+ "grad_norm": 0.015431761741638184,
+ "learning_rate": 0.0006,
+ "loss": 4.41126823425293,
+ "step": 4485
+ },
+ {
+ "epoch": 62.30742358078603,
+ "grad_norm": 0.016335977241396904,
+ "learning_rate": 0.0006,
+ "loss": 4.419488906860352,
+ "step": 4486
+ },
+ {
+ "epoch": 62.32139737991266,
+ "grad_norm": 0.01832406409084797,
+ "learning_rate": 0.0006,
+ "loss": 4.579026222229004,
+ "step": 4487
+ },
+ {
+ "epoch": 62.3353711790393,
+ "grad_norm": 0.01790742017328739,
+ "learning_rate": 0.0006,
+ "loss": 4.289964199066162,
+ "step": 4488
+ },
+ {
+ "epoch": 62.34934497816594,
+ "grad_norm": 0.017266202718019485,
+ "learning_rate": 0.0006,
+ "loss": 4.415959358215332,
+ "step": 4489
+ },
+ {
+ "epoch": 62.36331877729258,
+ "grad_norm": 0.015976132825016975,
+ "learning_rate": 0.0006,
+ "loss": 4.358587265014648,
+ "step": 4490
+ },
+ {
+ "epoch": 62.377292576419215,
+ "grad_norm": 0.016873881220817566,
+ "learning_rate": 0.0006,
+ "loss": 4.354607105255127,
+ "step": 4491
+ },
+ {
+ "epoch": 62.391266375545854,
+ "grad_norm": 0.01476745493710041,
+ "learning_rate": 0.0006,
+ "loss": 4.384904384613037,
+ "step": 4492
+ },
+ {
+ "epoch": 62.40524017467249,
+ "grad_norm": 0.013810782693326473,
+ "learning_rate": 0.0006,
+ "loss": 4.459230899810791,
+ "step": 4493
+ },
+ {
+ "epoch": 62.419213973799124,
+ "grad_norm": 0.015150421299040318,
+ "learning_rate": 0.0006,
+ "loss": 4.369471549987793,
+ "step": 4494
+ },
+ {
+ "epoch": 62.43318777292576,
+ "grad_norm": 0.014539610594511032,
+ "learning_rate": 0.0006,
+ "loss": 4.3564982414245605,
+ "step": 4495
+ },
+ {
+ "epoch": 62.4471615720524,
+ "grad_norm": 0.015554017387330532,
+ "learning_rate": 0.0006,
+ "loss": 4.414379119873047,
+ "step": 4496
+ },
+ {
+ "epoch": 62.46113537117904,
+ "grad_norm": 0.016608357429504395,
+ "learning_rate": 0.0006,
+ "loss": 4.311585426330566,
+ "step": 4497
+ },
+ {
+ "epoch": 62.47510917030568,
+ "grad_norm": 0.01885881833732128,
+ "learning_rate": 0.0006,
+ "loss": 4.306629180908203,
+ "step": 4498
+ },
+ {
+ "epoch": 62.48908296943232,
+ "grad_norm": 0.022128663957118988,
+ "learning_rate": 0.0006,
+ "loss": 4.345483779907227,
+ "step": 4499
+ },
+ {
+ "epoch": 62.50305676855895,
+ "grad_norm": 0.023401018232107162,
+ "learning_rate": 0.0006,
+ "loss": 4.398204326629639,
+ "step": 4500
+ },
+ {
+ "epoch": 62.51703056768559,
+ "grad_norm": 0.022080009803175926,
+ "learning_rate": 0.0006,
+ "loss": 4.375027179718018,
+ "step": 4501
+ },
+ {
+ "epoch": 62.531004366812226,
+ "grad_norm": 0.020004864782094955,
+ "learning_rate": 0.0006,
+ "loss": 4.33494758605957,
+ "step": 4502
+ },
+ {
+ "epoch": 62.544978165938865,
+ "grad_norm": 0.01786700263619423,
+ "learning_rate": 0.0006,
+ "loss": 4.4428300857543945,
+ "step": 4503
+ },
+ {
+ "epoch": 62.5589519650655,
+ "grad_norm": 0.01885388419032097,
+ "learning_rate": 0.0006,
+ "loss": 4.4138946533203125,
+ "step": 4504
+ },
+ {
+ "epoch": 62.57292576419214,
+ "grad_norm": 0.019890939816832542,
+ "learning_rate": 0.0006,
+ "loss": 4.278632164001465,
+ "step": 4505
+ },
+ {
+ "epoch": 62.58689956331878,
+ "grad_norm": 0.02109031192958355,
+ "learning_rate": 0.0006,
+ "loss": 4.3756327629089355,
+ "step": 4506
+ },
+ {
+ "epoch": 62.60087336244541,
+ "grad_norm": 0.020813899114727974,
+ "learning_rate": 0.0006,
+ "loss": 4.410213470458984,
+ "step": 4507
+ },
+ {
+ "epoch": 62.61484716157205,
+ "grad_norm": 0.021655689924955368,
+ "learning_rate": 0.0006,
+ "loss": 4.30633020401001,
+ "step": 4508
+ },
+ {
+ "epoch": 62.62882096069869,
+ "grad_norm": 0.018155649304389954,
+ "learning_rate": 0.0006,
+ "loss": 4.407284736633301,
+ "step": 4509
+ },
+ {
+ "epoch": 62.64279475982533,
+ "grad_norm": 0.017765069380402565,
+ "learning_rate": 0.0006,
+ "loss": 4.321715354919434,
+ "step": 4510
+ },
+ {
+ "epoch": 62.65676855895197,
+ "grad_norm": 0.017724554985761642,
+ "learning_rate": 0.0006,
+ "loss": 4.493371963500977,
+ "step": 4511
+ },
+ {
+ "epoch": 62.670742358078606,
+ "grad_norm": 0.01492092851549387,
+ "learning_rate": 0.0006,
+ "loss": 4.443025588989258,
+ "step": 4512
+ },
+ {
+ "epoch": 62.68471615720524,
+ "grad_norm": 0.01615324057638645,
+ "learning_rate": 0.0006,
+ "loss": 4.3289594650268555,
+ "step": 4513
+ },
+ {
+ "epoch": 62.698689956331876,
+ "grad_norm": 0.016569366678595543,
+ "learning_rate": 0.0006,
+ "loss": 4.369720458984375,
+ "step": 4514
+ },
+ {
+ "epoch": 62.712663755458514,
+ "grad_norm": 0.016509901732206345,
+ "learning_rate": 0.0006,
+ "loss": 4.292316436767578,
+ "step": 4515
+ },
+ {
+ "epoch": 62.72663755458515,
+ "grad_norm": 0.014270003885030746,
+ "learning_rate": 0.0006,
+ "loss": 4.349854469299316,
+ "step": 4516
+ },
+ {
+ "epoch": 62.74061135371179,
+ "grad_norm": 0.014235340990126133,
+ "learning_rate": 0.0006,
+ "loss": 4.473727226257324,
+ "step": 4517
+ },
+ {
+ "epoch": 62.75458515283843,
+ "grad_norm": 0.015163225121796131,
+ "learning_rate": 0.0006,
+ "loss": 4.452878475189209,
+ "step": 4518
+ },
+ {
+ "epoch": 62.76855895196506,
+ "grad_norm": 0.015748055651783943,
+ "learning_rate": 0.0006,
+ "loss": 4.521087169647217,
+ "step": 4519
+ },
+ {
+ "epoch": 62.7825327510917,
+ "grad_norm": 0.01661108434200287,
+ "learning_rate": 0.0006,
+ "loss": 4.443885326385498,
+ "step": 4520
+ },
+ {
+ "epoch": 62.79650655021834,
+ "grad_norm": 0.01610301062464714,
+ "learning_rate": 0.0006,
+ "loss": 4.312990188598633,
+ "step": 4521
+ },
+ {
+ "epoch": 62.81048034934498,
+ "grad_norm": 0.017007583752274513,
+ "learning_rate": 0.0006,
+ "loss": 4.408937454223633,
+ "step": 4522
+ },
+ {
+ "epoch": 62.82445414847162,
+ "grad_norm": 0.01749461516737938,
+ "learning_rate": 0.0006,
+ "loss": 4.346066951751709,
+ "step": 4523
+ },
+ {
+ "epoch": 62.838427947598255,
+ "grad_norm": 0.018243493512272835,
+ "learning_rate": 0.0006,
+ "loss": 4.357658386230469,
+ "step": 4524
+ },
+ {
+ "epoch": 62.852401746724894,
+ "grad_norm": 0.017499985173344612,
+ "learning_rate": 0.0006,
+ "loss": 4.374061584472656,
+ "step": 4525
+ },
+ {
+ "epoch": 62.866375545851525,
+ "grad_norm": 0.014536325819790363,
+ "learning_rate": 0.0006,
+ "loss": 4.189321994781494,
+ "step": 4526
+ },
+ {
+ "epoch": 62.880349344978164,
+ "grad_norm": 0.014728891663253307,
+ "learning_rate": 0.0006,
+ "loss": 4.425910949707031,
+ "step": 4527
+ },
+ {
+ "epoch": 62.8943231441048,
+ "grad_norm": 0.015013772062957287,
+ "learning_rate": 0.0006,
+ "loss": 4.337319850921631,
+ "step": 4528
+ },
+ {
+ "epoch": 62.90829694323144,
+ "grad_norm": 0.01626027375459671,
+ "learning_rate": 0.0006,
+ "loss": 4.330339431762695,
+ "step": 4529
+ },
+ {
+ "epoch": 62.92227074235808,
+ "grad_norm": 0.014773444272577763,
+ "learning_rate": 0.0006,
+ "loss": 4.4188737869262695,
+ "step": 4530
+ },
+ {
+ "epoch": 62.93624454148472,
+ "grad_norm": 0.01422932744026184,
+ "learning_rate": 0.0006,
+ "loss": 4.355986595153809,
+ "step": 4531
+ },
+ {
+ "epoch": 62.95021834061135,
+ "grad_norm": 0.012170110829174519,
+ "learning_rate": 0.0006,
+ "loss": 4.405722618103027,
+ "step": 4532
+ },
+ {
+ "epoch": 62.96419213973799,
+ "grad_norm": 0.013233253732323647,
+ "learning_rate": 0.0006,
+ "loss": 4.377286911010742,
+ "step": 4533
+ },
+ {
+ "epoch": 62.97816593886463,
+ "grad_norm": 0.014497299678623676,
+ "learning_rate": 0.0006,
+ "loss": 4.508796691894531,
+ "step": 4534
+ },
+ {
+ "epoch": 62.992139737991266,
+ "grad_norm": 0.014718741178512573,
+ "learning_rate": 0.0006,
+ "loss": 4.370919704437256,
+ "step": 4535
+ },
+ {
+ "epoch": 63.0,
+ "grad_norm": 0.015833502635359764,
+ "learning_rate": 0.0006,
+ "loss": 4.258753776550293,
+ "step": 4536
+ },
+ {
+ "epoch": 63.0,
+ "eval_loss": 4.714163303375244,
+ "eval_runtime": 57.0269,
+ "eval_samples_per_second": 42.822,
+ "eval_steps_per_second": 1.35,
+ "step": 4536
+ },
+ {
+ "epoch": 63.01397379912664,
+ "grad_norm": 0.014277510344982147,
+ "learning_rate": 0.0006,
+ "loss": 4.3112640380859375,
+ "step": 4537
+ },
+ {
+ "epoch": 63.02794759825328,
+ "grad_norm": 0.015377935022115707,
+ "learning_rate": 0.0006,
+ "loss": 4.342782974243164,
+ "step": 4538
+ },
+ {
+ "epoch": 63.041921397379916,
+ "grad_norm": 0.015406244434416294,
+ "learning_rate": 0.0006,
+ "loss": 4.474056720733643,
+ "step": 4539
+ },
+ {
+ "epoch": 63.05589519650655,
+ "grad_norm": 0.018121499568223953,
+ "learning_rate": 0.0006,
+ "loss": 4.397963523864746,
+ "step": 4540
+ },
+ {
+ "epoch": 63.069868995633186,
+ "grad_norm": 0.022634636610746384,
+ "learning_rate": 0.0006,
+ "loss": 4.306631088256836,
+ "step": 4541
+ },
+ {
+ "epoch": 63.083842794759825,
+ "grad_norm": 0.026448970660567284,
+ "learning_rate": 0.0006,
+ "loss": 4.283175468444824,
+ "step": 4542
+ },
+ {
+ "epoch": 63.09781659388646,
+ "grad_norm": 0.024104537442326546,
+ "learning_rate": 0.0006,
+ "loss": 4.315826416015625,
+ "step": 4543
+ },
+ {
+ "epoch": 63.1117903930131,
+ "grad_norm": 0.019040480256080627,
+ "learning_rate": 0.0006,
+ "loss": 4.44057035446167,
+ "step": 4544
+ },
+ {
+ "epoch": 63.12576419213974,
+ "grad_norm": 0.02006000280380249,
+ "learning_rate": 0.0006,
+ "loss": 4.405603408813477,
+ "step": 4545
+ },
+ {
+ "epoch": 63.13973799126637,
+ "grad_norm": 0.019980071112513542,
+ "learning_rate": 0.0006,
+ "loss": 4.393726348876953,
+ "step": 4546
+ },
+ {
+ "epoch": 63.15371179039301,
+ "grad_norm": 0.01641865260899067,
+ "learning_rate": 0.0006,
+ "loss": 4.272546768188477,
+ "step": 4547
+ },
+ {
+ "epoch": 63.16768558951965,
+ "grad_norm": 0.01890353485941887,
+ "learning_rate": 0.0006,
+ "loss": 4.382566928863525,
+ "step": 4548
+ },
+ {
+ "epoch": 63.18165938864629,
+ "grad_norm": 0.018370937556028366,
+ "learning_rate": 0.0006,
+ "loss": 4.423953533172607,
+ "step": 4549
+ },
+ {
+ "epoch": 63.19563318777293,
+ "grad_norm": 0.014985294081270695,
+ "learning_rate": 0.0006,
+ "loss": 4.308627128601074,
+ "step": 4550
+ },
+ {
+ "epoch": 63.209606986899566,
+ "grad_norm": 0.017513327300548553,
+ "learning_rate": 0.0006,
+ "loss": 4.248819351196289,
+ "step": 4551
+ },
+ {
+ "epoch": 63.223580786026204,
+ "grad_norm": 0.019464161247015,
+ "learning_rate": 0.0006,
+ "loss": 4.359729766845703,
+ "step": 4552
+ },
+ {
+ "epoch": 63.237554585152836,
+ "grad_norm": 0.018678035587072372,
+ "learning_rate": 0.0006,
+ "loss": 4.3299665451049805,
+ "step": 4553
+ },
+ {
+ "epoch": 63.251528384279474,
+ "grad_norm": 0.017909523099660873,
+ "learning_rate": 0.0006,
+ "loss": 4.246147155761719,
+ "step": 4554
+ },
+ {
+ "epoch": 63.26550218340611,
+ "grad_norm": 0.01747293211519718,
+ "learning_rate": 0.0006,
+ "loss": 4.41165018081665,
+ "step": 4555
+ },
+ {
+ "epoch": 63.27947598253275,
+ "grad_norm": 0.018670853227376938,
+ "learning_rate": 0.0006,
+ "loss": 4.409883499145508,
+ "step": 4556
+ },
+ {
+ "epoch": 63.29344978165939,
+ "grad_norm": 0.0189529862254858,
+ "learning_rate": 0.0006,
+ "loss": 4.418654918670654,
+ "step": 4557
+ },
+ {
+ "epoch": 63.30742358078603,
+ "grad_norm": 0.018394414335489273,
+ "learning_rate": 0.0006,
+ "loss": 4.38355827331543,
+ "step": 4558
+ },
+ {
+ "epoch": 63.32139737991266,
+ "grad_norm": 0.018317820504307747,
+ "learning_rate": 0.0006,
+ "loss": 4.456843376159668,
+ "step": 4559
+ },
+ {
+ "epoch": 63.3353711790393,
+ "grad_norm": 0.019225597381591797,
+ "learning_rate": 0.0006,
+ "loss": 4.232264518737793,
+ "step": 4560
+ },
+ {
+ "epoch": 63.34934497816594,
+ "grad_norm": 0.016663681715726852,
+ "learning_rate": 0.0006,
+ "loss": 4.393517017364502,
+ "step": 4561
+ },
+ {
+ "epoch": 63.36331877729258,
+ "grad_norm": 0.014975365251302719,
+ "learning_rate": 0.0006,
+ "loss": 4.375183582305908,
+ "step": 4562
+ },
+ {
+ "epoch": 63.377292576419215,
+ "grad_norm": 0.01771732047200203,
+ "learning_rate": 0.0006,
+ "loss": 4.3073410987854,
+ "step": 4563
+ },
+ {
+ "epoch": 63.391266375545854,
+ "grad_norm": 0.01663549244403839,
+ "learning_rate": 0.0006,
+ "loss": 4.311003684997559,
+ "step": 4564
+ },
+ {
+ "epoch": 63.40524017467249,
+ "grad_norm": 0.016301177442073822,
+ "learning_rate": 0.0006,
+ "loss": 4.326807022094727,
+ "step": 4565
+ },
+ {
+ "epoch": 63.419213973799124,
+ "grad_norm": 0.016781188547611237,
+ "learning_rate": 0.0006,
+ "loss": 4.306562423706055,
+ "step": 4566
+ },
+ {
+ "epoch": 63.43318777292576,
+ "grad_norm": 0.01767745055258274,
+ "learning_rate": 0.0006,
+ "loss": 4.367082595825195,
+ "step": 4567
+ },
+ {
+ "epoch": 63.4471615720524,
+ "grad_norm": 0.016728565096855164,
+ "learning_rate": 0.0006,
+ "loss": 4.327400207519531,
+ "step": 4568
+ },
+ {
+ "epoch": 63.46113537117904,
+ "grad_norm": 0.016289403662085533,
+ "learning_rate": 0.0006,
+ "loss": 4.392712116241455,
+ "step": 4569
+ },
+ {
+ "epoch": 63.47510917030568,
+ "grad_norm": 0.016233716160058975,
+ "learning_rate": 0.0006,
+ "loss": 4.354372024536133,
+ "step": 4570
+ },
+ {
+ "epoch": 63.48908296943232,
+ "grad_norm": 0.016547974199056625,
+ "learning_rate": 0.0006,
+ "loss": 4.309050559997559,
+ "step": 4571
+ },
+ {
+ "epoch": 63.50305676855895,
+ "grad_norm": 0.018466509878635406,
+ "learning_rate": 0.0006,
+ "loss": 4.318345069885254,
+ "step": 4572
+ },
+ {
+ "epoch": 63.51703056768559,
+ "grad_norm": 0.017685720697045326,
+ "learning_rate": 0.0006,
+ "loss": 4.399726867675781,
+ "step": 4573
+ },
+ {
+ "epoch": 63.531004366812226,
+ "grad_norm": 0.015801234170794487,
+ "learning_rate": 0.0006,
+ "loss": 4.328495502471924,
+ "step": 4574
+ },
+ {
+ "epoch": 63.544978165938865,
+ "grad_norm": 0.0188873540610075,
+ "learning_rate": 0.0006,
+ "loss": 4.253244400024414,
+ "step": 4575
+ },
+ {
+ "epoch": 63.5589519650655,
+ "grad_norm": 0.01868271268904209,
+ "learning_rate": 0.0006,
+ "loss": 4.463746070861816,
+ "step": 4576
+ },
+ {
+ "epoch": 63.57292576419214,
+ "grad_norm": 0.01844949834048748,
+ "learning_rate": 0.0006,
+ "loss": 4.400300025939941,
+ "step": 4577
+ },
+ {
+ "epoch": 63.58689956331878,
+ "grad_norm": 0.015256480313837528,
+ "learning_rate": 0.0006,
+ "loss": 4.445844650268555,
+ "step": 4578
+ },
+ {
+ "epoch": 63.60087336244541,
+ "grad_norm": 0.015305760316550732,
+ "learning_rate": 0.0006,
+ "loss": 4.317327499389648,
+ "step": 4579
+ },
+ {
+ "epoch": 63.61484716157205,
+ "grad_norm": 0.015200769528746605,
+ "learning_rate": 0.0006,
+ "loss": 4.25112247467041,
+ "step": 4580
+ },
+ {
+ "epoch": 63.62882096069869,
+ "grad_norm": 0.015919538214802742,
+ "learning_rate": 0.0006,
+ "loss": 4.277937889099121,
+ "step": 4581
+ },
+ {
+ "epoch": 63.64279475982533,
+ "grad_norm": 0.01572665013372898,
+ "learning_rate": 0.0006,
+ "loss": 4.302237510681152,
+ "step": 4582
+ },
+ {
+ "epoch": 63.65676855895197,
+ "grad_norm": 0.016428731381893158,
+ "learning_rate": 0.0006,
+ "loss": 4.36762809753418,
+ "step": 4583
+ },
+ {
+ "epoch": 63.670742358078606,
+ "grad_norm": 0.01706082746386528,
+ "learning_rate": 0.0006,
+ "loss": 4.236526966094971,
+ "step": 4584
+ },
+ {
+ "epoch": 63.68471615720524,
+ "grad_norm": 0.017850136384367943,
+ "learning_rate": 0.0006,
+ "loss": 4.264129638671875,
+ "step": 4585
+ },
+ {
+ "epoch": 63.698689956331876,
+ "grad_norm": 0.016079438850283623,
+ "learning_rate": 0.0006,
+ "loss": 4.449799060821533,
+ "step": 4586
+ },
+ {
+ "epoch": 63.712663755458514,
+ "grad_norm": 0.014996383339166641,
+ "learning_rate": 0.0006,
+ "loss": 4.404574394226074,
+ "step": 4587
+ },
+ {
+ "epoch": 63.72663755458515,
+ "grad_norm": 0.015910431742668152,
+ "learning_rate": 0.0006,
+ "loss": 4.348101615905762,
+ "step": 4588
+ },
+ {
+ "epoch": 63.74061135371179,
+ "grad_norm": 0.01630021072924137,
+ "learning_rate": 0.0006,
+ "loss": 4.3193817138671875,
+ "step": 4589
+ },
+ {
+ "epoch": 63.75458515283843,
+ "grad_norm": 0.014571773819625378,
+ "learning_rate": 0.0006,
+ "loss": 4.374402046203613,
+ "step": 4590
+ },
+ {
+ "epoch": 63.76855895196506,
+ "grad_norm": 0.01662345975637436,
+ "learning_rate": 0.0006,
+ "loss": 4.329075813293457,
+ "step": 4591
+ },
+ {
+ "epoch": 63.7825327510917,
+ "grad_norm": 0.016604198142886162,
+ "learning_rate": 0.0006,
+ "loss": 4.436126232147217,
+ "step": 4592
+ },
+ {
+ "epoch": 63.79650655021834,
+ "grad_norm": 0.016578884795308113,
+ "learning_rate": 0.0006,
+ "loss": 4.361468315124512,
+ "step": 4593
+ },
+ {
+ "epoch": 63.81048034934498,
+ "grad_norm": 0.016339825466275215,
+ "learning_rate": 0.0006,
+ "loss": 4.449892044067383,
+ "step": 4594
+ },
+ {
+ "epoch": 63.82445414847162,
+ "grad_norm": 0.016805455088615417,
+ "learning_rate": 0.0006,
+ "loss": 4.415866851806641,
+ "step": 4595
+ },
+ {
+ "epoch": 63.838427947598255,
+ "grad_norm": 0.017366213724017143,
+ "learning_rate": 0.0006,
+ "loss": 4.361730575561523,
+ "step": 4596
+ },
+ {
+ "epoch": 63.852401746724894,
+ "grad_norm": 0.015599003992974758,
+ "learning_rate": 0.0006,
+ "loss": 4.363637924194336,
+ "step": 4597
+ },
+ {
+ "epoch": 63.866375545851525,
+ "grad_norm": 0.015674632042646408,
+ "learning_rate": 0.0006,
+ "loss": 4.332045078277588,
+ "step": 4598
+ },
+ {
+ "epoch": 63.880349344978164,
+ "grad_norm": 0.015350443311035633,
+ "learning_rate": 0.0006,
+ "loss": 4.320063591003418,
+ "step": 4599
+ },
+ {
+ "epoch": 63.8943231441048,
+ "grad_norm": 0.015388146974146366,
+ "learning_rate": 0.0006,
+ "loss": 4.398199081420898,
+ "step": 4600
+ },
+ {
+ "epoch": 63.90829694323144,
+ "grad_norm": 0.01450140681117773,
+ "learning_rate": 0.0006,
+ "loss": 4.244076728820801,
+ "step": 4601
+ },
+ {
+ "epoch": 63.92227074235808,
+ "grad_norm": 0.01637491025030613,
+ "learning_rate": 0.0006,
+ "loss": 4.213549613952637,
+ "step": 4602
+ },
+ {
+ "epoch": 63.93624454148472,
+ "grad_norm": 0.0159814041107893,
+ "learning_rate": 0.0006,
+ "loss": 4.421775817871094,
+ "step": 4603
+ },
+ {
+ "epoch": 63.95021834061135,
+ "grad_norm": 0.017386827617883682,
+ "learning_rate": 0.0006,
+ "loss": 4.340023994445801,
+ "step": 4604
+ },
+ {
+ "epoch": 63.96419213973799,
+ "grad_norm": 0.019679656252264977,
+ "learning_rate": 0.0006,
+ "loss": 4.352534294128418,
+ "step": 4605
+ },
+ {
+ "epoch": 63.97816593886463,
+ "grad_norm": 0.022391516715288162,
+ "learning_rate": 0.0006,
+ "loss": 4.3538312911987305,
+ "step": 4606
+ },
+ {
+ "epoch": 63.992139737991266,
+ "grad_norm": 0.023281006142497063,
+ "learning_rate": 0.0006,
+ "loss": 4.432089328765869,
+ "step": 4607
+ },
+ {
+ "epoch": 64.0,
+ "grad_norm": 0.019916288554668427,
+ "learning_rate": 0.0006,
+ "loss": 4.466937065124512,
+ "step": 4608
+ },
+ {
+ "epoch": 64.0,
+ "eval_loss": 4.659008502960205,
+ "eval_runtime": 56.0019,
+ "eval_samples_per_second": 43.606,
+ "eval_steps_per_second": 1.375,
+ "step": 4608
+ },
+ {
+ "epoch": 64.01397379912663,
+ "grad_norm": 0.017351187765598297,
+ "learning_rate": 0.0006,
+ "loss": 4.294129371643066,
+ "step": 4609
+ },
+ {
+ "epoch": 64.02794759825328,
+ "grad_norm": 0.02405865117907524,
+ "learning_rate": 0.0006,
+ "loss": 4.266717910766602,
+ "step": 4610
+ },
+ {
+ "epoch": 64.04192139737991,
+ "grad_norm": 0.022158386185765266,
+ "learning_rate": 0.0006,
+ "loss": 4.395932197570801,
+ "step": 4611
+ },
+ {
+ "epoch": 64.05589519650655,
+ "grad_norm": 0.019238505512475967,
+ "learning_rate": 0.0006,
+ "loss": 4.266704082489014,
+ "step": 4612
+ },
+ {
+ "epoch": 64.06986899563319,
+ "grad_norm": 0.021670425310730934,
+ "learning_rate": 0.0006,
+ "loss": 4.3160552978515625,
+ "step": 4613
+ },
+ {
+ "epoch": 64.08384279475983,
+ "grad_norm": 0.02076641097664833,
+ "learning_rate": 0.0006,
+ "loss": 4.415626525878906,
+ "step": 4614
+ },
+ {
+ "epoch": 64.09781659388646,
+ "grad_norm": 0.018223078921437263,
+ "learning_rate": 0.0006,
+ "loss": 4.292387008666992,
+ "step": 4615
+ },
+ {
+ "epoch": 64.1117903930131,
+ "grad_norm": 0.017401937395334244,
+ "learning_rate": 0.0006,
+ "loss": 4.424524307250977,
+ "step": 4616
+ },
+ {
+ "epoch": 64.12576419213974,
+ "grad_norm": 0.01663743332028389,
+ "learning_rate": 0.0006,
+ "loss": 4.350032806396484,
+ "step": 4617
+ },
+ {
+ "epoch": 64.13973799126637,
+ "grad_norm": 0.018716629594564438,
+ "learning_rate": 0.0006,
+ "loss": 4.309157371520996,
+ "step": 4618
+ },
+ {
+ "epoch": 64.15371179039302,
+ "grad_norm": 0.019569825381040573,
+ "learning_rate": 0.0006,
+ "loss": 4.358292579650879,
+ "step": 4619
+ },
+ {
+ "epoch": 64.16768558951965,
+ "grad_norm": 0.018994422629475594,
+ "learning_rate": 0.0006,
+ "loss": 4.3338823318481445,
+ "step": 4620
+ },
+ {
+ "epoch": 64.18165938864628,
+ "grad_norm": 0.017690688371658325,
+ "learning_rate": 0.0006,
+ "loss": 4.414468765258789,
+ "step": 4621
+ },
+ {
+ "epoch": 64.19563318777293,
+ "grad_norm": 0.0170609001070261,
+ "learning_rate": 0.0006,
+ "loss": 4.4337568283081055,
+ "step": 4622
+ },
+ {
+ "epoch": 64.20960698689956,
+ "grad_norm": 0.016984131187200546,
+ "learning_rate": 0.0006,
+ "loss": 4.304839134216309,
+ "step": 4623
+ },
+ {
+ "epoch": 64.2235807860262,
+ "grad_norm": 0.01619867794215679,
+ "learning_rate": 0.0006,
+ "loss": 4.389623165130615,
+ "step": 4624
+ },
+ {
+ "epoch": 64.23755458515284,
+ "grad_norm": 0.015428110957145691,
+ "learning_rate": 0.0006,
+ "loss": 4.277515888214111,
+ "step": 4625
+ },
+ {
+ "epoch": 64.25152838427948,
+ "grad_norm": 0.015909532085061073,
+ "learning_rate": 0.0006,
+ "loss": 4.266029357910156,
+ "step": 4626
+ },
+ {
+ "epoch": 64.26550218340611,
+ "grad_norm": 0.014296560548245907,
+ "learning_rate": 0.0006,
+ "loss": 4.286067485809326,
+ "step": 4627
+ },
+ {
+ "epoch": 64.27947598253274,
+ "grad_norm": 0.014065971598029137,
+ "learning_rate": 0.0006,
+ "loss": 4.353528022766113,
+ "step": 4628
+ },
+ {
+ "epoch": 64.29344978165939,
+ "grad_norm": 0.014386293478310108,
+ "learning_rate": 0.0006,
+ "loss": 4.318715572357178,
+ "step": 4629
+ },
+ {
+ "epoch": 64.30742358078602,
+ "grad_norm": 0.014001123607158661,
+ "learning_rate": 0.0006,
+ "loss": 4.362497329711914,
+ "step": 4630
+ },
+ {
+ "epoch": 64.32139737991267,
+ "grad_norm": 0.013408979400992393,
+ "learning_rate": 0.0006,
+ "loss": 4.417397499084473,
+ "step": 4631
+ },
+ {
+ "epoch": 64.3353711790393,
+ "grad_norm": 0.014960093423724174,
+ "learning_rate": 0.0006,
+ "loss": 4.359710693359375,
+ "step": 4632
+ },
+ {
+ "epoch": 64.34934497816595,
+ "grad_norm": 0.014658777043223381,
+ "learning_rate": 0.0006,
+ "loss": 4.318822383880615,
+ "step": 4633
+ },
+ {
+ "epoch": 64.36331877729258,
+ "grad_norm": 0.013766744174063206,
+ "learning_rate": 0.0006,
+ "loss": 4.265679836273193,
+ "step": 4634
+ },
+ {
+ "epoch": 64.37729257641921,
+ "grad_norm": 0.013821255415678024,
+ "learning_rate": 0.0006,
+ "loss": 4.3452911376953125,
+ "step": 4635
+ },
+ {
+ "epoch": 64.39126637554585,
+ "grad_norm": 0.013541066087782383,
+ "learning_rate": 0.0006,
+ "loss": 4.429254531860352,
+ "step": 4636
+ },
+ {
+ "epoch": 64.40524017467249,
+ "grad_norm": 0.013985645025968552,
+ "learning_rate": 0.0006,
+ "loss": 4.304452896118164,
+ "step": 4637
+ },
+ {
+ "epoch": 64.41921397379913,
+ "grad_norm": 0.015079808421432972,
+ "learning_rate": 0.0006,
+ "loss": 4.296828746795654,
+ "step": 4638
+ },
+ {
+ "epoch": 64.43318777292576,
+ "grad_norm": 0.014896944165229797,
+ "learning_rate": 0.0006,
+ "loss": 4.299156665802002,
+ "step": 4639
+ },
+ {
+ "epoch": 64.44716157205241,
+ "grad_norm": 0.013770041987299919,
+ "learning_rate": 0.0006,
+ "loss": 4.351439952850342,
+ "step": 4640
+ },
+ {
+ "epoch": 64.46113537117904,
+ "grad_norm": 0.01531178504228592,
+ "learning_rate": 0.0006,
+ "loss": 4.373715877532959,
+ "step": 4641
+ },
+ {
+ "epoch": 64.47510917030567,
+ "grad_norm": 0.014024450443685055,
+ "learning_rate": 0.0006,
+ "loss": 4.368053436279297,
+ "step": 4642
+ },
+ {
+ "epoch": 64.48908296943232,
+ "grad_norm": 0.013864024542272091,
+ "learning_rate": 0.0006,
+ "loss": 4.378564834594727,
+ "step": 4643
+ },
+ {
+ "epoch": 64.50305676855895,
+ "grad_norm": 0.013925445266067982,
+ "learning_rate": 0.0006,
+ "loss": 4.261663436889648,
+ "step": 4644
+ },
+ {
+ "epoch": 64.5170305676856,
+ "grad_norm": 0.015011338517069817,
+ "learning_rate": 0.0006,
+ "loss": 4.337285995483398,
+ "step": 4645
+ },
+ {
+ "epoch": 64.53100436681223,
+ "grad_norm": 0.016941986978054047,
+ "learning_rate": 0.0006,
+ "loss": 4.336877822875977,
+ "step": 4646
+ },
+ {
+ "epoch": 64.54497816593886,
+ "grad_norm": 0.01719909906387329,
+ "learning_rate": 0.0006,
+ "loss": 4.149813175201416,
+ "step": 4647
+ },
+ {
+ "epoch": 64.5589519650655,
+ "grad_norm": 0.01608402095735073,
+ "learning_rate": 0.0006,
+ "loss": 4.191280364990234,
+ "step": 4648
+ },
+ {
+ "epoch": 64.57292576419214,
+ "grad_norm": 0.015284820459783077,
+ "learning_rate": 0.0006,
+ "loss": 4.35953426361084,
+ "step": 4649
+ },
+ {
+ "epoch": 64.58689956331878,
+ "grad_norm": 0.014855682849884033,
+ "learning_rate": 0.0006,
+ "loss": 4.308401107788086,
+ "step": 4650
+ },
+ {
+ "epoch": 64.60087336244541,
+ "grad_norm": 0.017553111538290977,
+ "learning_rate": 0.0006,
+ "loss": 4.279720306396484,
+ "step": 4651
+ },
+ {
+ "epoch": 64.61484716157206,
+ "grad_norm": 0.019506998360157013,
+ "learning_rate": 0.0006,
+ "loss": 4.319524765014648,
+ "step": 4652
+ },
+ {
+ "epoch": 64.62882096069869,
+ "grad_norm": 0.020293425768613815,
+ "learning_rate": 0.0006,
+ "loss": 4.340523719787598,
+ "step": 4653
+ },
+ {
+ "epoch": 64.64279475982532,
+ "grad_norm": 0.01664556749165058,
+ "learning_rate": 0.0006,
+ "loss": 4.296785831451416,
+ "step": 4654
+ },
+ {
+ "epoch": 64.65676855895197,
+ "grad_norm": 0.012592585757374763,
+ "learning_rate": 0.0006,
+ "loss": 4.464595317840576,
+ "step": 4655
+ },
+ {
+ "epoch": 64.6707423580786,
+ "grad_norm": 0.01549589168280363,
+ "learning_rate": 0.0006,
+ "loss": 4.275053024291992,
+ "step": 4656
+ },
+ {
+ "epoch": 64.68471615720524,
+ "grad_norm": 0.01705172471702099,
+ "learning_rate": 0.0006,
+ "loss": 4.388431549072266,
+ "step": 4657
+ },
+ {
+ "epoch": 64.69868995633188,
+ "grad_norm": 0.015438582748174667,
+ "learning_rate": 0.0006,
+ "loss": 4.312070846557617,
+ "step": 4658
+ },
+ {
+ "epoch": 64.71266375545852,
+ "grad_norm": 0.014448083005845547,
+ "learning_rate": 0.0006,
+ "loss": 4.237599849700928,
+ "step": 4659
+ },
+ {
+ "epoch": 64.72663755458515,
+ "grad_norm": 0.015267071314156055,
+ "learning_rate": 0.0006,
+ "loss": 4.29893684387207,
+ "step": 4660
+ },
+ {
+ "epoch": 64.74061135371178,
+ "grad_norm": 0.016615847125649452,
+ "learning_rate": 0.0006,
+ "loss": 4.432273864746094,
+ "step": 4661
+ },
+ {
+ "epoch": 64.75458515283843,
+ "grad_norm": 0.016584362834692,
+ "learning_rate": 0.0006,
+ "loss": 4.202266216278076,
+ "step": 4662
+ },
+ {
+ "epoch": 64.76855895196506,
+ "grad_norm": 0.018313994631171227,
+ "learning_rate": 0.0006,
+ "loss": 4.367966651916504,
+ "step": 4663
+ },
+ {
+ "epoch": 64.78253275109171,
+ "grad_norm": 0.017946481704711914,
+ "learning_rate": 0.0006,
+ "loss": 4.205172538757324,
+ "step": 4664
+ },
+ {
+ "epoch": 64.79650655021834,
+ "grad_norm": 0.017744354903697968,
+ "learning_rate": 0.0006,
+ "loss": 4.295661926269531,
+ "step": 4665
+ },
+ {
+ "epoch": 64.81048034934499,
+ "grad_norm": 0.016103824600577354,
+ "learning_rate": 0.0006,
+ "loss": 4.276543617248535,
+ "step": 4666
+ },
+ {
+ "epoch": 64.82445414847162,
+ "grad_norm": 0.014590012840926647,
+ "learning_rate": 0.0006,
+ "loss": 4.414247035980225,
+ "step": 4667
+ },
+ {
+ "epoch": 64.83842794759825,
+ "grad_norm": 0.01782386749982834,
+ "learning_rate": 0.0006,
+ "loss": 4.203220367431641,
+ "step": 4668
+ },
+ {
+ "epoch": 64.8524017467249,
+ "grad_norm": 0.022122733294963837,
+ "learning_rate": 0.0006,
+ "loss": 4.38078498840332,
+ "step": 4669
+ },
+ {
+ "epoch": 64.86637554585153,
+ "grad_norm": 0.0250293780118227,
+ "learning_rate": 0.0006,
+ "loss": 4.3310136795043945,
+ "step": 4670
+ },
+ {
+ "epoch": 64.88034934497817,
+ "grad_norm": 0.022438062354922295,
+ "learning_rate": 0.0006,
+ "loss": 4.428631782531738,
+ "step": 4671
+ },
+ {
+ "epoch": 64.8943231441048,
+ "grad_norm": 0.018518192693591118,
+ "learning_rate": 0.0006,
+ "loss": 4.428250312805176,
+ "step": 4672
+ },
+ {
+ "epoch": 64.90829694323143,
+ "grad_norm": 0.018105225637555122,
+ "learning_rate": 0.0006,
+ "loss": 4.336209297180176,
+ "step": 4673
+ },
+ {
+ "epoch": 64.92227074235808,
+ "grad_norm": 0.017564399167895317,
+ "learning_rate": 0.0006,
+ "loss": 4.416011810302734,
+ "step": 4674
+ },
+ {
+ "epoch": 64.93624454148471,
+ "grad_norm": 0.018747031688690186,
+ "learning_rate": 0.0006,
+ "loss": 4.419938087463379,
+ "step": 4675
+ },
+ {
+ "epoch": 64.95021834061136,
+ "grad_norm": 0.018686484545469284,
+ "learning_rate": 0.0006,
+ "loss": 4.347185134887695,
+ "step": 4676
+ },
+ {
+ "epoch": 64.96419213973799,
+ "grad_norm": 0.019329898059368134,
+ "learning_rate": 0.0006,
+ "loss": 4.343749046325684,
+ "step": 4677
+ },
+ {
+ "epoch": 64.97816593886463,
+ "grad_norm": 0.018517782911658287,
+ "learning_rate": 0.0006,
+ "loss": 4.264604568481445,
+ "step": 4678
+ },
+ {
+ "epoch": 64.99213973799127,
+ "grad_norm": 0.016579125076532364,
+ "learning_rate": 0.0006,
+ "loss": 4.303423881530762,
+ "step": 4679
+ },
+ {
+ "epoch": 65.0,
+ "grad_norm": 0.01617567613720894,
+ "learning_rate": 0.0006,
+ "loss": 4.326556205749512,
+ "step": 4680
+ },
+ {
+ "epoch": 65.0,
+ "eval_loss": 4.655026912689209,
+ "eval_runtime": 57.1299,
+ "eval_samples_per_second": 42.745,
+ "eval_steps_per_second": 1.348,
+ "step": 4680
+ },
+ {
+ "epoch": 65.01397379912663,
+ "grad_norm": 0.01607249490916729,
+ "learning_rate": 0.0006,
+ "loss": 4.276607036590576,
+ "step": 4681
+ },
+ {
+ "epoch": 65.02794759825328,
+ "grad_norm": 0.016779497265815735,
+ "learning_rate": 0.0006,
+ "loss": 4.412464618682861,
+ "step": 4682
+ },
+ {
+ "epoch": 65.04192139737991,
+ "grad_norm": 0.018161555752158165,
+ "learning_rate": 0.0006,
+ "loss": 4.199329376220703,
+ "step": 4683
+ },
+ {
+ "epoch": 65.05589519650655,
+ "grad_norm": 0.017508309334516525,
+ "learning_rate": 0.0006,
+ "loss": 4.217567443847656,
+ "step": 4684
+ },
+ {
+ "epoch": 65.06986899563319,
+ "grad_norm": 0.015716979280114174,
+ "learning_rate": 0.0006,
+ "loss": 4.322023391723633,
+ "step": 4685
+ },
+ {
+ "epoch": 65.08384279475983,
+ "grad_norm": 0.015113635919988155,
+ "learning_rate": 0.0006,
+ "loss": 4.333515167236328,
+ "step": 4686
+ },
+ {
+ "epoch": 65.09781659388646,
+ "grad_norm": 0.01437560748308897,
+ "learning_rate": 0.0006,
+ "loss": 4.295609474182129,
+ "step": 4687
+ },
+ {
+ "epoch": 65.1117903930131,
+ "grad_norm": 0.013074862770736217,
+ "learning_rate": 0.0006,
+ "loss": 4.314120769500732,
+ "step": 4688
+ },
+ {
+ "epoch": 65.12576419213974,
+ "grad_norm": 0.012908284552395344,
+ "learning_rate": 0.0006,
+ "loss": 4.216890811920166,
+ "step": 4689
+ },
+ {
+ "epoch": 65.13973799126637,
+ "grad_norm": 0.013001754879951477,
+ "learning_rate": 0.0006,
+ "loss": 4.287717342376709,
+ "step": 4690
+ },
+ {
+ "epoch": 65.15371179039302,
+ "grad_norm": 0.013694263994693756,
+ "learning_rate": 0.0006,
+ "loss": 4.333837985992432,
+ "step": 4691
+ },
+ {
+ "epoch": 65.16768558951965,
+ "grad_norm": 0.014461833983659744,
+ "learning_rate": 0.0006,
+ "loss": 4.323545455932617,
+ "step": 4692
+ },
+ {
+ "epoch": 65.18165938864628,
+ "grad_norm": 0.013670675456523895,
+ "learning_rate": 0.0006,
+ "loss": 4.3313493728637695,
+ "step": 4693
+ },
+ {
+ "epoch": 65.19563318777293,
+ "grad_norm": 0.013094350695610046,
+ "learning_rate": 0.0006,
+ "loss": 4.23858118057251,
+ "step": 4694
+ },
+ {
+ "epoch": 65.20960698689956,
+ "grad_norm": 0.013996944762766361,
+ "learning_rate": 0.0006,
+ "loss": 4.428705215454102,
+ "step": 4695
+ },
+ {
+ "epoch": 65.2235807860262,
+ "grad_norm": 0.013295048847794533,
+ "learning_rate": 0.0006,
+ "loss": 4.238950252532959,
+ "step": 4696
+ },
+ {
+ "epoch": 65.23755458515284,
+ "grad_norm": 0.013380616903305054,
+ "learning_rate": 0.0006,
+ "loss": 4.273185729980469,
+ "step": 4697
+ },
+ {
+ "epoch": 65.25152838427948,
+ "grad_norm": 0.015062890015542507,
+ "learning_rate": 0.0006,
+ "loss": 4.29472017288208,
+ "step": 4698
+ },
+ {
+ "epoch": 65.26550218340611,
+ "grad_norm": 0.017317263409495354,
+ "learning_rate": 0.0006,
+ "loss": 4.337954998016357,
+ "step": 4699
+ },
+ {
+ "epoch": 65.27947598253274,
+ "grad_norm": 0.017698992043733597,
+ "learning_rate": 0.0006,
+ "loss": 4.336437225341797,
+ "step": 4700
+ },
+ {
+ "epoch": 65.29344978165939,
+ "grad_norm": 0.01628575474023819,
+ "learning_rate": 0.0006,
+ "loss": 4.302372455596924,
+ "step": 4701
+ },
+ {
+ "epoch": 65.30742358078602,
+ "grad_norm": 0.014531854540109634,
+ "learning_rate": 0.0006,
+ "loss": 4.256450653076172,
+ "step": 4702
+ },
+ {
+ "epoch": 65.32139737991267,
+ "grad_norm": 0.014549105428159237,
+ "learning_rate": 0.0006,
+ "loss": 4.302240371704102,
+ "step": 4703
+ },
+ {
+ "epoch": 65.3353711790393,
+ "grad_norm": 0.014881663955748081,
+ "learning_rate": 0.0006,
+ "loss": 4.3316121101379395,
+ "step": 4704
+ },
+ {
+ "epoch": 65.34934497816595,
+ "grad_norm": 0.014513437636196613,
+ "learning_rate": 0.0006,
+ "loss": 4.441212177276611,
+ "step": 4705
+ },
+ {
+ "epoch": 65.36331877729258,
+ "grad_norm": 0.014122087508440018,
+ "learning_rate": 0.0006,
+ "loss": 4.343296527862549,
+ "step": 4706
+ },
+ {
+ "epoch": 65.37729257641921,
+ "grad_norm": 0.014817348681390285,
+ "learning_rate": 0.0006,
+ "loss": 4.195674896240234,
+ "step": 4707
+ },
+ {
+ "epoch": 65.39126637554585,
+ "grad_norm": 0.01787056028842926,
+ "learning_rate": 0.0006,
+ "loss": 4.3507914543151855,
+ "step": 4708
+ },
+ {
+ "epoch": 65.40524017467249,
+ "grad_norm": 0.015295136719942093,
+ "learning_rate": 0.0006,
+ "loss": 4.234372138977051,
+ "step": 4709
+ },
+ {
+ "epoch": 65.41921397379913,
+ "grad_norm": 0.014324234798550606,
+ "learning_rate": 0.0006,
+ "loss": 4.2907867431640625,
+ "step": 4710
+ },
+ {
+ "epoch": 65.43318777292576,
+ "grad_norm": 0.016090670600533485,
+ "learning_rate": 0.0006,
+ "loss": 4.177244186401367,
+ "step": 4711
+ },
+ {
+ "epoch": 65.44716157205241,
+ "grad_norm": 0.0174604132771492,
+ "learning_rate": 0.0006,
+ "loss": 4.335577964782715,
+ "step": 4712
+ },
+ {
+ "epoch": 65.46113537117904,
+ "grad_norm": 0.016731899231672287,
+ "learning_rate": 0.0006,
+ "loss": 4.30197811126709,
+ "step": 4713
+ },
+ {
+ "epoch": 65.47510917030567,
+ "grad_norm": 0.017423735931515694,
+ "learning_rate": 0.0006,
+ "loss": 4.3528852462768555,
+ "step": 4714
+ },
+ {
+ "epoch": 65.48908296943232,
+ "grad_norm": 0.01780976541340351,
+ "learning_rate": 0.0006,
+ "loss": 4.185036659240723,
+ "step": 4715
+ },
+ {
+ "epoch": 65.50305676855895,
+ "grad_norm": 0.019977200776338577,
+ "learning_rate": 0.0006,
+ "loss": 4.328766345977783,
+ "step": 4716
+ },
+ {
+ "epoch": 65.5170305676856,
+ "grad_norm": 0.020284287631511688,
+ "learning_rate": 0.0006,
+ "loss": 4.285494804382324,
+ "step": 4717
+ },
+ {
+ "epoch": 65.53100436681223,
+ "grad_norm": 0.019168002530932426,
+ "learning_rate": 0.0006,
+ "loss": 4.328560829162598,
+ "step": 4718
+ },
+ {
+ "epoch": 65.54497816593886,
+ "grad_norm": 0.019217776134610176,
+ "learning_rate": 0.0006,
+ "loss": 4.189998626708984,
+ "step": 4719
+ },
+ {
+ "epoch": 65.5589519650655,
+ "grad_norm": 0.018960488960146904,
+ "learning_rate": 0.0006,
+ "loss": 4.4995012283325195,
+ "step": 4720
+ },
+ {
+ "epoch": 65.57292576419214,
+ "grad_norm": 0.01941288262605667,
+ "learning_rate": 0.0006,
+ "loss": 4.26570987701416,
+ "step": 4721
+ },
+ {
+ "epoch": 65.58689956331878,
+ "grad_norm": 0.020263448357582092,
+ "learning_rate": 0.0006,
+ "loss": 4.411645889282227,
+ "step": 4722
+ },
+ {
+ "epoch": 65.60087336244541,
+ "grad_norm": 0.02144799567759037,
+ "learning_rate": 0.0006,
+ "loss": 4.163675308227539,
+ "step": 4723
+ },
+ {
+ "epoch": 65.61484716157206,
+ "grad_norm": 0.019790446385741234,
+ "learning_rate": 0.0006,
+ "loss": 4.2361063957214355,
+ "step": 4724
+ },
+ {
+ "epoch": 65.62882096069869,
+ "grad_norm": 0.01829805225133896,
+ "learning_rate": 0.0006,
+ "loss": 4.449810028076172,
+ "step": 4725
+ },
+ {
+ "epoch": 65.64279475982532,
+ "grad_norm": 0.01571105420589447,
+ "learning_rate": 0.0006,
+ "loss": 4.281406879425049,
+ "step": 4726
+ },
+ {
+ "epoch": 65.65676855895197,
+ "grad_norm": 0.016477147117257118,
+ "learning_rate": 0.0006,
+ "loss": 4.293408393859863,
+ "step": 4727
+ },
+ {
+ "epoch": 65.6707423580786,
+ "grad_norm": 0.016766032204031944,
+ "learning_rate": 0.0006,
+ "loss": 4.378252983093262,
+ "step": 4728
+ },
+ {
+ "epoch": 65.68471615720524,
+ "grad_norm": 0.0159933939576149,
+ "learning_rate": 0.0006,
+ "loss": 4.3065667152404785,
+ "step": 4729
+ },
+ {
+ "epoch": 65.69868995633188,
+ "grad_norm": 0.01658209227025509,
+ "learning_rate": 0.0006,
+ "loss": 4.174026966094971,
+ "step": 4730
+ },
+ {
+ "epoch": 65.71266375545852,
+ "grad_norm": 0.015081380493938923,
+ "learning_rate": 0.0006,
+ "loss": 4.360222339630127,
+ "step": 4731
+ },
+ {
+ "epoch": 65.72663755458515,
+ "grad_norm": 0.015110890381038189,
+ "learning_rate": 0.0006,
+ "loss": 4.218810558319092,
+ "step": 4732
+ },
+ {
+ "epoch": 65.74061135371178,
+ "grad_norm": 0.015300953760743141,
+ "learning_rate": 0.0006,
+ "loss": 4.331979274749756,
+ "step": 4733
+ },
+ {
+ "epoch": 65.75458515283843,
+ "grad_norm": 0.01715504191815853,
+ "learning_rate": 0.0006,
+ "loss": 4.349112033843994,
+ "step": 4734
+ },
+ {
+ "epoch": 65.76855895196506,
+ "grad_norm": 0.015870366245508194,
+ "learning_rate": 0.0006,
+ "loss": 4.329850196838379,
+ "step": 4735
+ },
+ {
+ "epoch": 65.78253275109171,
+ "grad_norm": 0.01541937980800867,
+ "learning_rate": 0.0006,
+ "loss": 4.3400163650512695,
+ "step": 4736
+ },
+ {
+ "epoch": 65.79650655021834,
+ "grad_norm": 0.01738942787051201,
+ "learning_rate": 0.0006,
+ "loss": 4.220798492431641,
+ "step": 4737
+ },
+ {
+ "epoch": 65.81048034934499,
+ "grad_norm": 0.01685990020632744,
+ "learning_rate": 0.0006,
+ "loss": 4.177238464355469,
+ "step": 4738
+ },
+ {
+ "epoch": 65.82445414847162,
+ "grad_norm": 0.016183465719223022,
+ "learning_rate": 0.0006,
+ "loss": 4.327216625213623,
+ "step": 4739
+ },
+ {
+ "epoch": 65.83842794759825,
+ "grad_norm": 0.01677694357931614,
+ "learning_rate": 0.0006,
+ "loss": 4.27766227722168,
+ "step": 4740
+ },
+ {
+ "epoch": 65.8524017467249,
+ "grad_norm": 0.01645066775381565,
+ "learning_rate": 0.0006,
+ "loss": 4.288347244262695,
+ "step": 4741
+ },
+ {
+ "epoch": 65.86637554585153,
+ "grad_norm": 0.018034635111689568,
+ "learning_rate": 0.0006,
+ "loss": 4.292458534240723,
+ "step": 4742
+ },
+ {
+ "epoch": 65.88034934497817,
+ "grad_norm": 0.017973346635699272,
+ "learning_rate": 0.0006,
+ "loss": 4.3117265701293945,
+ "step": 4743
+ },
+ {
+ "epoch": 65.8943231441048,
+ "grad_norm": 0.016605669632554054,
+ "learning_rate": 0.0006,
+ "loss": 4.280615329742432,
+ "step": 4744
+ },
+ {
+ "epoch": 65.90829694323143,
+ "grad_norm": 0.016298074275255203,
+ "learning_rate": 0.0006,
+ "loss": 4.301933288574219,
+ "step": 4745
+ },
+ {
+ "epoch": 65.92227074235808,
+ "grad_norm": 0.015383216552436352,
+ "learning_rate": 0.0006,
+ "loss": 4.352383136749268,
+ "step": 4746
+ },
+ {
+ "epoch": 65.93624454148471,
+ "grad_norm": 0.016222182661294937,
+ "learning_rate": 0.0006,
+ "loss": 4.305725574493408,
+ "step": 4747
+ },
+ {
+ "epoch": 65.95021834061136,
+ "grad_norm": 0.01691329851746559,
+ "learning_rate": 0.0006,
+ "loss": 4.29502010345459,
+ "step": 4748
+ },
+ {
+ "epoch": 65.96419213973799,
+ "grad_norm": 0.017637677490711212,
+ "learning_rate": 0.0006,
+ "loss": 4.3133039474487305,
+ "step": 4749
+ },
+ {
+ "epoch": 65.97816593886463,
+ "grad_norm": 0.021362047642469406,
+ "learning_rate": 0.0006,
+ "loss": 4.458834648132324,
+ "step": 4750
+ },
+ {
+ "epoch": 65.99213973799127,
+ "grad_norm": 0.023553011938929558,
+ "learning_rate": 0.0006,
+ "loss": 4.298727989196777,
+ "step": 4751
+ },
+ {
+ "epoch": 66.0,
+ "grad_norm": 0.02154950052499771,
+ "learning_rate": 0.0006,
+ "loss": 4.403392791748047,
+ "step": 4752
+ },
+ {
+ "epoch": 66.0,
+ "eval_loss": 4.724340438842773,
+ "eval_runtime": 56.9555,
+ "eval_samples_per_second": 42.876,
+ "eval_steps_per_second": 1.352,
+ "step": 4752
+ },
+ {
+ "epoch": 66.01397379912663,
+ "grad_norm": 0.01916784606873989,
+ "learning_rate": 0.0006,
+ "loss": 4.22931432723999,
+ "step": 4753
+ },
+ {
+ "epoch": 66.02794759825328,
+ "grad_norm": 0.017492054030299187,
+ "learning_rate": 0.0006,
+ "loss": 4.271522521972656,
+ "step": 4754
+ },
+ {
+ "epoch": 66.04192139737991,
+ "grad_norm": 0.018408095464110374,
+ "learning_rate": 0.0006,
+ "loss": 4.337705135345459,
+ "step": 4755
+ },
+ {
+ "epoch": 66.05589519650655,
+ "grad_norm": 0.019220108166337013,
+ "learning_rate": 0.0006,
+ "loss": 4.280340194702148,
+ "step": 4756
+ },
+ {
+ "epoch": 66.06986899563319,
+ "grad_norm": 0.01773735322058201,
+ "learning_rate": 0.0006,
+ "loss": 4.390201568603516,
+ "step": 4757
+ },
+ {
+ "epoch": 66.08384279475983,
+ "grad_norm": 0.01679125428199768,
+ "learning_rate": 0.0006,
+ "loss": 4.299282073974609,
+ "step": 4758
+ },
+ {
+ "epoch": 66.09781659388646,
+ "grad_norm": 0.01479059923440218,
+ "learning_rate": 0.0006,
+ "loss": 4.472630500793457,
+ "step": 4759
+ },
+ {
+ "epoch": 66.1117903930131,
+ "grad_norm": 0.015130104497075081,
+ "learning_rate": 0.0006,
+ "loss": 4.301194190979004,
+ "step": 4760
+ },
+ {
+ "epoch": 66.12576419213974,
+ "grad_norm": 0.0158979594707489,
+ "learning_rate": 0.0006,
+ "loss": 4.375751972198486,
+ "step": 4761
+ },
+ {
+ "epoch": 66.13973799126637,
+ "grad_norm": 0.014432408846914768,
+ "learning_rate": 0.0006,
+ "loss": 4.279788017272949,
+ "step": 4762
+ },
+ {
+ "epoch": 66.15371179039302,
+ "grad_norm": 0.013026784174144268,
+ "learning_rate": 0.0006,
+ "loss": 4.255035400390625,
+ "step": 4763
+ },
+ {
+ "epoch": 66.16768558951965,
+ "grad_norm": 0.014422653242945671,
+ "learning_rate": 0.0006,
+ "loss": 4.261832237243652,
+ "step": 4764
+ },
+ {
+ "epoch": 66.18165938864628,
+ "grad_norm": 0.014613019302487373,
+ "learning_rate": 0.0006,
+ "loss": 4.382184028625488,
+ "step": 4765
+ },
+ {
+ "epoch": 66.19563318777293,
+ "grad_norm": 0.014989894814789295,
+ "learning_rate": 0.0006,
+ "loss": 4.40528678894043,
+ "step": 4766
+ },
+ {
+ "epoch": 66.20960698689956,
+ "grad_norm": 0.018190350383520126,
+ "learning_rate": 0.0006,
+ "loss": 4.302414894104004,
+ "step": 4767
+ },
+ {
+ "epoch": 66.2235807860262,
+ "grad_norm": 0.020887890830636024,
+ "learning_rate": 0.0006,
+ "loss": 4.289920806884766,
+ "step": 4768
+ },
+ {
+ "epoch": 66.23755458515284,
+ "grad_norm": 0.02376876398921013,
+ "learning_rate": 0.0006,
+ "loss": 4.297855377197266,
+ "step": 4769
+ },
+ {
+ "epoch": 66.25152838427948,
+ "grad_norm": 0.02152269519865513,
+ "learning_rate": 0.0006,
+ "loss": 4.1972832679748535,
+ "step": 4770
+ },
+ {
+ "epoch": 66.26550218340611,
+ "grad_norm": 0.015572802163660526,
+ "learning_rate": 0.0006,
+ "loss": 4.307487487792969,
+ "step": 4771
+ },
+ {
+ "epoch": 66.27947598253274,
+ "grad_norm": 0.015228749252855778,
+ "learning_rate": 0.0006,
+ "loss": 4.103250503540039,
+ "step": 4772
+ },
+ {
+ "epoch": 66.29344978165939,
+ "grad_norm": 0.016925504431128502,
+ "learning_rate": 0.0006,
+ "loss": 4.247461318969727,
+ "step": 4773
+ },
+ {
+ "epoch": 66.30742358078602,
+ "grad_norm": 0.01613711006939411,
+ "learning_rate": 0.0006,
+ "loss": 4.4699273109436035,
+ "step": 4774
+ },
+ {
+ "epoch": 66.32139737991267,
+ "grad_norm": 0.018332140520215034,
+ "learning_rate": 0.0006,
+ "loss": 4.22969388961792,
+ "step": 4775
+ },
+ {
+ "epoch": 66.3353711790393,
+ "grad_norm": 0.019575074315071106,
+ "learning_rate": 0.0006,
+ "loss": 4.29397439956665,
+ "step": 4776
+ },
+ {
+ "epoch": 66.34934497816595,
+ "grad_norm": 0.018177419900894165,
+ "learning_rate": 0.0006,
+ "loss": 4.221899509429932,
+ "step": 4777
+ },
+ {
+ "epoch": 66.36331877729258,
+ "grad_norm": 0.014292889274656773,
+ "learning_rate": 0.0006,
+ "loss": 4.325457572937012,
+ "step": 4778
+ },
+ {
+ "epoch": 66.37729257641921,
+ "grad_norm": 0.015098796226084232,
+ "learning_rate": 0.0006,
+ "loss": 4.364837169647217,
+ "step": 4779
+ },
+ {
+ "epoch": 66.39126637554585,
+ "grad_norm": 0.013992635533213615,
+ "learning_rate": 0.0006,
+ "loss": 4.299082279205322,
+ "step": 4780
+ },
+ {
+ "epoch": 66.40524017467249,
+ "grad_norm": 0.01504511572420597,
+ "learning_rate": 0.0006,
+ "loss": 4.230877876281738,
+ "step": 4781
+ },
+ {
+ "epoch": 66.41921397379913,
+ "grad_norm": 0.014875639230012894,
+ "learning_rate": 0.0006,
+ "loss": 4.368405818939209,
+ "step": 4782
+ },
+ {
+ "epoch": 66.43318777292576,
+ "grad_norm": 0.014306887984275818,
+ "learning_rate": 0.0006,
+ "loss": 4.389184474945068,
+ "step": 4783
+ },
+ {
+ "epoch": 66.44716157205241,
+ "grad_norm": 0.012959563173353672,
+ "learning_rate": 0.0006,
+ "loss": 4.37934684753418,
+ "step": 4784
+ },
+ {
+ "epoch": 66.46113537117904,
+ "grad_norm": 0.013206703588366508,
+ "learning_rate": 0.0006,
+ "loss": 4.442799091339111,
+ "step": 4785
+ },
+ {
+ "epoch": 66.47510917030567,
+ "grad_norm": 0.014848909340798855,
+ "learning_rate": 0.0006,
+ "loss": 4.203481674194336,
+ "step": 4786
+ },
+ {
+ "epoch": 66.48908296943232,
+ "grad_norm": 0.015637682750821114,
+ "learning_rate": 0.0006,
+ "loss": 4.27552604675293,
+ "step": 4787
+ },
+ {
+ "epoch": 66.50305676855895,
+ "grad_norm": 0.016017816960811615,
+ "learning_rate": 0.0006,
+ "loss": 4.296402931213379,
+ "step": 4788
+ },
+ {
+ "epoch": 66.5170305676856,
+ "grad_norm": 0.01577814109623432,
+ "learning_rate": 0.0006,
+ "loss": 4.301733016967773,
+ "step": 4789
+ },
+ {
+ "epoch": 66.53100436681223,
+ "grad_norm": 0.015293175354599953,
+ "learning_rate": 0.0006,
+ "loss": 4.369877815246582,
+ "step": 4790
+ },
+ {
+ "epoch": 66.54497816593886,
+ "grad_norm": 0.014262278564274311,
+ "learning_rate": 0.0006,
+ "loss": 4.3314738273620605,
+ "step": 4791
+ },
+ {
+ "epoch": 66.5589519650655,
+ "grad_norm": 0.015123482793569565,
+ "learning_rate": 0.0006,
+ "loss": 4.313591957092285,
+ "step": 4792
+ },
+ {
+ "epoch": 66.57292576419214,
+ "grad_norm": 0.015458766371011734,
+ "learning_rate": 0.0006,
+ "loss": 4.3045854568481445,
+ "step": 4793
+ },
+ {
+ "epoch": 66.58689956331878,
+ "grad_norm": 0.017217233777046204,
+ "learning_rate": 0.0006,
+ "loss": 4.269944190979004,
+ "step": 4794
+ },
+ {
+ "epoch": 66.60087336244541,
+ "grad_norm": 0.01962282694876194,
+ "learning_rate": 0.0006,
+ "loss": 4.271735191345215,
+ "step": 4795
+ },
+ {
+ "epoch": 66.61484716157206,
+ "grad_norm": 0.01917066052556038,
+ "learning_rate": 0.0006,
+ "loss": 4.288168430328369,
+ "step": 4796
+ },
+ {
+ "epoch": 66.62882096069869,
+ "grad_norm": 0.018211524933576584,
+ "learning_rate": 0.0006,
+ "loss": 4.280660629272461,
+ "step": 4797
+ },
+ {
+ "epoch": 66.64279475982532,
+ "grad_norm": 0.01936524175107479,
+ "learning_rate": 0.0006,
+ "loss": 4.3710551261901855,
+ "step": 4798
+ },
+ {
+ "epoch": 66.65676855895197,
+ "grad_norm": 0.01882351003587246,
+ "learning_rate": 0.0006,
+ "loss": 4.342750549316406,
+ "step": 4799
+ },
+ {
+ "epoch": 66.6707423580786,
+ "grad_norm": 0.017149122431874275,
+ "learning_rate": 0.0006,
+ "loss": 4.32134485244751,
+ "step": 4800
+ },
+ {
+ "epoch": 66.68471615720524,
+ "grad_norm": 0.01699327677488327,
+ "learning_rate": 0.0006,
+ "loss": 4.248893737792969,
+ "step": 4801
+ },
+ {
+ "epoch": 66.69868995633188,
+ "grad_norm": 0.016873255372047424,
+ "learning_rate": 0.0006,
+ "loss": 4.277350425720215,
+ "step": 4802
+ },
+ {
+ "epoch": 66.71266375545852,
+ "grad_norm": 0.016592837870121002,
+ "learning_rate": 0.0006,
+ "loss": 4.221665859222412,
+ "step": 4803
+ },
+ {
+ "epoch": 66.72663755458515,
+ "grad_norm": 0.015887131914496422,
+ "learning_rate": 0.0006,
+ "loss": 4.328040599822998,
+ "step": 4804
+ },
+ {
+ "epoch": 66.74061135371178,
+ "grad_norm": 0.014057286083698273,
+ "learning_rate": 0.0006,
+ "loss": 4.3294548988342285,
+ "step": 4805
+ },
+ {
+ "epoch": 66.75458515283843,
+ "grad_norm": 0.013135905377566814,
+ "learning_rate": 0.0006,
+ "loss": 4.254383563995361,
+ "step": 4806
+ },
+ {
+ "epoch": 66.76855895196506,
+ "grad_norm": 0.013632557354867458,
+ "learning_rate": 0.0006,
+ "loss": 4.342815399169922,
+ "step": 4807
+ },
+ {
+ "epoch": 66.78253275109171,
+ "grad_norm": 0.015512688085436821,
+ "learning_rate": 0.0006,
+ "loss": 4.297835350036621,
+ "step": 4808
+ },
+ {
+ "epoch": 66.79650655021834,
+ "grad_norm": 0.01621832884848118,
+ "learning_rate": 0.0006,
+ "loss": 4.314773082733154,
+ "step": 4809
+ },
+ {
+ "epoch": 66.81048034934499,
+ "grad_norm": 0.016196032986044884,
+ "learning_rate": 0.0006,
+ "loss": 4.29209566116333,
+ "step": 4810
+ },
+ {
+ "epoch": 66.82445414847162,
+ "grad_norm": 0.015463353134691715,
+ "learning_rate": 0.0006,
+ "loss": 4.262795448303223,
+ "step": 4811
+ },
+ {
+ "epoch": 66.83842794759825,
+ "grad_norm": 0.016143787652254105,
+ "learning_rate": 0.0006,
+ "loss": 4.274284362792969,
+ "step": 4812
+ },
+ {
+ "epoch": 66.8524017467249,
+ "grad_norm": 0.01610061153769493,
+ "learning_rate": 0.0006,
+ "loss": 4.337160587310791,
+ "step": 4813
+ },
+ {
+ "epoch": 66.86637554585153,
+ "grad_norm": 0.01582838036119938,
+ "learning_rate": 0.0006,
+ "loss": 4.192723274230957,
+ "step": 4814
+ },
+ {
+ "epoch": 66.88034934497817,
+ "grad_norm": 0.015081540681421757,
+ "learning_rate": 0.0006,
+ "loss": 4.370708465576172,
+ "step": 4815
+ },
+ {
+ "epoch": 66.8943231441048,
+ "grad_norm": 0.015929723158478737,
+ "learning_rate": 0.0006,
+ "loss": 4.340244293212891,
+ "step": 4816
+ },
+ {
+ "epoch": 66.90829694323143,
+ "grad_norm": 0.01784711331129074,
+ "learning_rate": 0.0006,
+ "loss": 4.260406970977783,
+ "step": 4817
+ },
+ {
+ "epoch": 66.92227074235808,
+ "grad_norm": 0.017434703186154366,
+ "learning_rate": 0.0006,
+ "loss": 4.286900043487549,
+ "step": 4818
+ },
+ {
+ "epoch": 66.93624454148471,
+ "grad_norm": 0.016262540593743324,
+ "learning_rate": 0.0006,
+ "loss": 4.292142391204834,
+ "step": 4819
+ },
+ {
+ "epoch": 66.95021834061136,
+ "grad_norm": 0.016122829169034958,
+ "learning_rate": 0.0006,
+ "loss": 4.291485786437988,
+ "step": 4820
+ },
+ {
+ "epoch": 66.96419213973799,
+ "grad_norm": 0.014361178502440453,
+ "learning_rate": 0.0006,
+ "loss": 4.275425910949707,
+ "step": 4821
+ },
+ {
+ "epoch": 66.97816593886463,
+ "grad_norm": 0.01373987551778555,
+ "learning_rate": 0.0006,
+ "loss": 4.223120212554932,
+ "step": 4822
+ },
+ {
+ "epoch": 66.99213973799127,
+ "grad_norm": 0.015687376260757446,
+ "learning_rate": 0.0006,
+ "loss": 4.355381011962891,
+ "step": 4823
+ },
+ {
+ "epoch": 67.0,
+ "grad_norm": 0.01998465321958065,
+ "learning_rate": 0.0006,
+ "loss": 4.264053821563721,
+ "step": 4824
+ },
+ {
+ "epoch": 67.0,
+ "eval_loss": 4.662398338317871,
+ "eval_runtime": 56.7928,
+ "eval_samples_per_second": 42.998,
+ "eval_steps_per_second": 1.356,
+ "step": 4824
+ },
+ {
+ "epoch": 67.01397379912663,
+ "grad_norm": 0.021736836060881615,
+ "learning_rate": 0.0006,
+ "loss": 4.243354797363281,
+ "step": 4825
+ },
+ {
+ "epoch": 67.02794759825328,
+ "grad_norm": 0.021476037800312042,
+ "learning_rate": 0.0006,
+ "loss": 4.218207836151123,
+ "step": 4826
+ },
+ {
+ "epoch": 67.04192139737991,
+ "grad_norm": 0.017188388854265213,
+ "learning_rate": 0.0006,
+ "loss": 4.18048095703125,
+ "step": 4827
+ },
+ {
+ "epoch": 67.05589519650655,
+ "grad_norm": 0.017293719574809074,
+ "learning_rate": 0.0006,
+ "loss": 4.239145278930664,
+ "step": 4828
+ },
+ {
+ "epoch": 67.06986899563319,
+ "grad_norm": 0.017383819445967674,
+ "learning_rate": 0.0006,
+ "loss": 4.114765167236328,
+ "step": 4829
+ },
+ {
+ "epoch": 67.08384279475983,
+ "grad_norm": 0.01651529036462307,
+ "learning_rate": 0.0006,
+ "loss": 4.351023197174072,
+ "step": 4830
+ },
+ {
+ "epoch": 67.09781659388646,
+ "grad_norm": 0.015893027186393738,
+ "learning_rate": 0.0006,
+ "loss": 4.261432647705078,
+ "step": 4831
+ },
+ {
+ "epoch": 67.1117903930131,
+ "grad_norm": 0.013381517492234707,
+ "learning_rate": 0.0006,
+ "loss": 4.172064781188965,
+ "step": 4832
+ },
+ {
+ "epoch": 67.12576419213974,
+ "grad_norm": 0.014642654918134212,
+ "learning_rate": 0.0006,
+ "loss": 4.356171131134033,
+ "step": 4833
+ },
+ {
+ "epoch": 67.13973799126637,
+ "grad_norm": 0.01425282284617424,
+ "learning_rate": 0.0006,
+ "loss": 4.303495407104492,
+ "step": 4834
+ },
+ {
+ "epoch": 67.15371179039302,
+ "grad_norm": 0.014574599452316761,
+ "learning_rate": 0.0006,
+ "loss": 4.2850847244262695,
+ "step": 4835
+ },
+ {
+ "epoch": 67.16768558951965,
+ "grad_norm": 0.012676282785832882,
+ "learning_rate": 0.0006,
+ "loss": 4.297845840454102,
+ "step": 4836
+ },
+ {
+ "epoch": 67.18165938864628,
+ "grad_norm": 0.013948136940598488,
+ "learning_rate": 0.0006,
+ "loss": 4.324530601501465,
+ "step": 4837
+ },
+ {
+ "epoch": 67.19563318777293,
+ "grad_norm": 0.014195773750543594,
+ "learning_rate": 0.0006,
+ "loss": 4.31236457824707,
+ "step": 4838
+ },
+ {
+ "epoch": 67.20960698689956,
+ "grad_norm": 0.013631382025778294,
+ "learning_rate": 0.0006,
+ "loss": 4.086362838745117,
+ "step": 4839
+ },
+ {
+ "epoch": 67.2235807860262,
+ "grad_norm": 0.012798587791621685,
+ "learning_rate": 0.0006,
+ "loss": 4.328471660614014,
+ "step": 4840
+ },
+ {
+ "epoch": 67.23755458515284,
+ "grad_norm": 0.014443321153521538,
+ "learning_rate": 0.0006,
+ "loss": 4.278019905090332,
+ "step": 4841
+ },
+ {
+ "epoch": 67.25152838427948,
+ "grad_norm": 0.015355420298874378,
+ "learning_rate": 0.0006,
+ "loss": 4.333611011505127,
+ "step": 4842
+ },
+ {
+ "epoch": 67.26550218340611,
+ "grad_norm": 0.016487296670675278,
+ "learning_rate": 0.0006,
+ "loss": 4.351336479187012,
+ "step": 4843
+ },
+ {
+ "epoch": 67.27947598253274,
+ "grad_norm": 0.017422541975975037,
+ "learning_rate": 0.0006,
+ "loss": 4.386096477508545,
+ "step": 4844
+ },
+ {
+ "epoch": 67.29344978165939,
+ "grad_norm": 0.015790918841958046,
+ "learning_rate": 0.0006,
+ "loss": 4.301377773284912,
+ "step": 4845
+ },
+ {
+ "epoch": 67.30742358078602,
+ "grad_norm": 0.016427364200353622,
+ "learning_rate": 0.0006,
+ "loss": 4.2838945388793945,
+ "step": 4846
+ },
+ {
+ "epoch": 67.32139737991267,
+ "grad_norm": 0.017050761729478836,
+ "learning_rate": 0.0006,
+ "loss": 4.282417297363281,
+ "step": 4847
+ },
+ {
+ "epoch": 67.3353711790393,
+ "grad_norm": 0.015006253495812416,
+ "learning_rate": 0.0006,
+ "loss": 4.352236270904541,
+ "step": 4848
+ },
+ {
+ "epoch": 67.34934497816595,
+ "grad_norm": 0.012996627949178219,
+ "learning_rate": 0.0006,
+ "loss": 4.309242248535156,
+ "step": 4849
+ },
+ {
+ "epoch": 67.36331877729258,
+ "grad_norm": 0.01367615070194006,
+ "learning_rate": 0.0006,
+ "loss": 4.211801528930664,
+ "step": 4850
+ },
+ {
+ "epoch": 67.37729257641921,
+ "grad_norm": 0.014775963500142097,
+ "learning_rate": 0.0006,
+ "loss": 4.360859394073486,
+ "step": 4851
+ },
+ {
+ "epoch": 67.39126637554585,
+ "grad_norm": 0.015854304656386375,
+ "learning_rate": 0.0006,
+ "loss": 4.211830139160156,
+ "step": 4852
+ },
+ {
+ "epoch": 67.40524017467249,
+ "grad_norm": 0.016888894140720367,
+ "learning_rate": 0.0006,
+ "loss": 4.216549873352051,
+ "step": 4853
+ },
+ {
+ "epoch": 67.41921397379913,
+ "grad_norm": 0.0170894768089056,
+ "learning_rate": 0.0006,
+ "loss": 4.263486385345459,
+ "step": 4854
+ },
+ {
+ "epoch": 67.43318777292576,
+ "grad_norm": 0.015829868614673615,
+ "learning_rate": 0.0006,
+ "loss": 4.360196113586426,
+ "step": 4855
+ },
+ {
+ "epoch": 67.44716157205241,
+ "grad_norm": 0.015220202505588531,
+ "learning_rate": 0.0006,
+ "loss": 4.243794918060303,
+ "step": 4856
+ },
+ {
+ "epoch": 67.46113537117904,
+ "grad_norm": 0.01413388829678297,
+ "learning_rate": 0.0006,
+ "loss": 4.286624908447266,
+ "step": 4857
+ },
+ {
+ "epoch": 67.47510917030567,
+ "grad_norm": 0.01426626369357109,
+ "learning_rate": 0.0006,
+ "loss": 4.257316589355469,
+ "step": 4858
+ },
+ {
+ "epoch": 67.48908296943232,
+ "grad_norm": 0.015957845374941826,
+ "learning_rate": 0.0006,
+ "loss": 4.300443649291992,
+ "step": 4859
+ },
+ {
+ "epoch": 67.50305676855895,
+ "grad_norm": 0.015499051660299301,
+ "learning_rate": 0.0006,
+ "loss": 4.278162002563477,
+ "step": 4860
+ },
+ {
+ "epoch": 67.5170305676856,
+ "grad_norm": 0.015145624987781048,
+ "learning_rate": 0.0006,
+ "loss": 4.267950534820557,
+ "step": 4861
+ },
+ {
+ "epoch": 67.53100436681223,
+ "grad_norm": 0.014610803686082363,
+ "learning_rate": 0.0006,
+ "loss": 4.247499942779541,
+ "step": 4862
+ },
+ {
+ "epoch": 67.54497816593886,
+ "grad_norm": 0.014860909432172775,
+ "learning_rate": 0.0006,
+ "loss": 4.263200759887695,
+ "step": 4863
+ },
+ {
+ "epoch": 67.5589519650655,
+ "grad_norm": 0.01718984916806221,
+ "learning_rate": 0.0006,
+ "loss": 4.290019989013672,
+ "step": 4864
+ },
+ {
+ "epoch": 67.57292576419214,
+ "grad_norm": 0.01558963768184185,
+ "learning_rate": 0.0006,
+ "loss": 4.223816871643066,
+ "step": 4865
+ },
+ {
+ "epoch": 67.58689956331878,
+ "grad_norm": 0.014927188865840435,
+ "learning_rate": 0.0006,
+ "loss": 4.363586902618408,
+ "step": 4866
+ },
+ {
+ "epoch": 67.60087336244541,
+ "grad_norm": 0.01573595404624939,
+ "learning_rate": 0.0006,
+ "loss": 4.254765510559082,
+ "step": 4867
+ },
+ {
+ "epoch": 67.61484716157206,
+ "grad_norm": 0.015618929639458656,
+ "learning_rate": 0.0006,
+ "loss": 4.25072717666626,
+ "step": 4868
+ },
+ {
+ "epoch": 67.62882096069869,
+ "grad_norm": 0.01605590060353279,
+ "learning_rate": 0.0006,
+ "loss": 4.25208854675293,
+ "step": 4869
+ },
+ {
+ "epoch": 67.64279475982532,
+ "grad_norm": 0.01804601214826107,
+ "learning_rate": 0.0006,
+ "loss": 4.351354122161865,
+ "step": 4870
+ },
+ {
+ "epoch": 67.65676855895197,
+ "grad_norm": 0.020595047622919083,
+ "learning_rate": 0.0006,
+ "loss": 4.193732738494873,
+ "step": 4871
+ },
+ {
+ "epoch": 67.6707423580786,
+ "grad_norm": 0.023037966340780258,
+ "learning_rate": 0.0006,
+ "loss": 4.27754020690918,
+ "step": 4872
+ },
+ {
+ "epoch": 67.68471615720524,
+ "grad_norm": 0.020680895075201988,
+ "learning_rate": 0.0006,
+ "loss": 4.314730167388916,
+ "step": 4873
+ },
+ {
+ "epoch": 67.69868995633188,
+ "grad_norm": 0.01887434534728527,
+ "learning_rate": 0.0006,
+ "loss": 4.252305030822754,
+ "step": 4874
+ },
+ {
+ "epoch": 67.71266375545852,
+ "grad_norm": 0.016333596780896187,
+ "learning_rate": 0.0006,
+ "loss": 4.146123886108398,
+ "step": 4875
+ },
+ {
+ "epoch": 67.72663755458515,
+ "grad_norm": 0.015688810497522354,
+ "learning_rate": 0.0006,
+ "loss": 4.2814154624938965,
+ "step": 4876
+ },
+ {
+ "epoch": 67.74061135371178,
+ "grad_norm": 0.016270726919174194,
+ "learning_rate": 0.0006,
+ "loss": 4.323200225830078,
+ "step": 4877
+ },
+ {
+ "epoch": 67.75458515283843,
+ "grad_norm": 0.016824988648295403,
+ "learning_rate": 0.0006,
+ "loss": 4.269501686096191,
+ "step": 4878
+ },
+ {
+ "epoch": 67.76855895196506,
+ "grad_norm": 0.01730562001466751,
+ "learning_rate": 0.0006,
+ "loss": 4.187993049621582,
+ "step": 4879
+ },
+ {
+ "epoch": 67.78253275109171,
+ "grad_norm": 0.01546655036509037,
+ "learning_rate": 0.0006,
+ "loss": 4.340415000915527,
+ "step": 4880
+ },
+ {
+ "epoch": 67.79650655021834,
+ "grad_norm": 0.014173958450555801,
+ "learning_rate": 0.0006,
+ "loss": 4.217419624328613,
+ "step": 4881
+ },
+ {
+ "epoch": 67.81048034934499,
+ "grad_norm": 0.016810474917292595,
+ "learning_rate": 0.0006,
+ "loss": 4.352302074432373,
+ "step": 4882
+ },
+ {
+ "epoch": 67.82445414847162,
+ "grad_norm": 0.0175373163074255,
+ "learning_rate": 0.0006,
+ "loss": 4.20377254486084,
+ "step": 4883
+ },
+ {
+ "epoch": 67.83842794759825,
+ "grad_norm": 0.015734676271677017,
+ "learning_rate": 0.0006,
+ "loss": 4.147495746612549,
+ "step": 4884
+ },
+ {
+ "epoch": 67.8524017467249,
+ "grad_norm": 0.015091817826032639,
+ "learning_rate": 0.0006,
+ "loss": 4.272252082824707,
+ "step": 4885
+ },
+ {
+ "epoch": 67.86637554585153,
+ "grad_norm": 0.015000219456851482,
+ "learning_rate": 0.0006,
+ "loss": 4.302773952484131,
+ "step": 4886
+ },
+ {
+ "epoch": 67.88034934497817,
+ "grad_norm": 0.013827184215188026,
+ "learning_rate": 0.0006,
+ "loss": 4.337242126464844,
+ "step": 4887
+ },
+ {
+ "epoch": 67.8943231441048,
+ "grad_norm": 0.015308346599340439,
+ "learning_rate": 0.0006,
+ "loss": 4.140895366668701,
+ "step": 4888
+ },
+ {
+ "epoch": 67.90829694323143,
+ "grad_norm": 0.015356178395450115,
+ "learning_rate": 0.0006,
+ "loss": 4.312541484832764,
+ "step": 4889
+ },
+ {
+ "epoch": 67.92227074235808,
+ "grad_norm": 0.0137353939935565,
+ "learning_rate": 0.0006,
+ "loss": 4.173184394836426,
+ "step": 4890
+ },
+ {
+ "epoch": 67.93624454148471,
+ "grad_norm": 0.01332368515431881,
+ "learning_rate": 0.0006,
+ "loss": 4.191199779510498,
+ "step": 4891
+ },
+ {
+ "epoch": 67.95021834061136,
+ "grad_norm": 0.01488130260258913,
+ "learning_rate": 0.0006,
+ "loss": 4.331752300262451,
+ "step": 4892
+ },
+ {
+ "epoch": 67.96419213973799,
+ "grad_norm": 0.015449733473360538,
+ "learning_rate": 0.0006,
+ "loss": 4.2385454177856445,
+ "step": 4893
+ },
+ {
+ "epoch": 67.97816593886463,
+ "grad_norm": 0.01556103490293026,
+ "learning_rate": 0.0006,
+ "loss": 4.262353897094727,
+ "step": 4894
+ },
+ {
+ "epoch": 67.99213973799127,
+ "grad_norm": 0.014613849110901356,
+ "learning_rate": 0.0006,
+ "loss": 4.223945617675781,
+ "step": 4895
+ },
+ {
+ "epoch": 68.0,
+ "grad_norm": 0.015323462896049023,
+ "learning_rate": 0.0006,
+ "loss": 4.317168712615967,
+ "step": 4896
+ },
+ {
+ "epoch": 68.0,
+ "eval_loss": 4.671603679656982,
+ "eval_runtime": 56.7317,
+ "eval_samples_per_second": 43.045,
+ "eval_steps_per_second": 1.357,
+ "step": 4896
+ },
+ {
+ "epoch": 68.01397379912663,
+ "grad_norm": 0.01910698413848877,
+ "learning_rate": 0.0006,
+ "loss": 4.206167221069336,
+ "step": 4897
+ },
+ {
+ "epoch": 68.02794759825328,
+ "grad_norm": 0.025112485513091087,
+ "learning_rate": 0.0006,
+ "loss": 4.384706974029541,
+ "step": 4898
+ },
+ {
+ "epoch": 68.04192139737991,
+ "grad_norm": 0.02618720941245556,
+ "learning_rate": 0.0006,
+ "loss": 4.321046829223633,
+ "step": 4899
+ },
+ {
+ "epoch": 68.05589519650655,
+ "grad_norm": 0.020853538066148758,
+ "learning_rate": 0.0006,
+ "loss": 4.217638969421387,
+ "step": 4900
+ },
+ {
+ "epoch": 68.06986899563319,
+ "grad_norm": 0.020025717094540596,
+ "learning_rate": 0.0006,
+ "loss": 4.306824684143066,
+ "step": 4901
+ },
+ {
+ "epoch": 68.08384279475983,
+ "grad_norm": 0.023868650197982788,
+ "learning_rate": 0.0006,
+ "loss": 4.327455997467041,
+ "step": 4902
+ },
+ {
+ "epoch": 68.09781659388646,
+ "grad_norm": 0.022279081866145134,
+ "learning_rate": 0.0006,
+ "loss": 4.400699615478516,
+ "step": 4903
+ },
+ {
+ "epoch": 68.1117903930131,
+ "grad_norm": 0.018965687602758408,
+ "learning_rate": 0.0006,
+ "loss": 4.19097900390625,
+ "step": 4904
+ },
+ {
+ "epoch": 68.12576419213974,
+ "grad_norm": 0.01936378888785839,
+ "learning_rate": 0.0006,
+ "loss": 4.206711769104004,
+ "step": 4905
+ },
+ {
+ "epoch": 68.13973799126637,
+ "grad_norm": 0.01811872608959675,
+ "learning_rate": 0.0006,
+ "loss": 4.204622268676758,
+ "step": 4906
+ },
+ {
+ "epoch": 68.15371179039302,
+ "grad_norm": 0.018530402332544327,
+ "learning_rate": 0.0006,
+ "loss": 4.289602279663086,
+ "step": 4907
+ },
+ {
+ "epoch": 68.16768558951965,
+ "grad_norm": 0.01755962148308754,
+ "learning_rate": 0.0006,
+ "loss": 4.317145347595215,
+ "step": 4908
+ },
+ {
+ "epoch": 68.18165938864628,
+ "grad_norm": 0.020413702353835106,
+ "learning_rate": 0.0006,
+ "loss": 4.2882843017578125,
+ "step": 4909
+ },
+ {
+ "epoch": 68.19563318777293,
+ "grad_norm": 0.020064299926161766,
+ "learning_rate": 0.0006,
+ "loss": 4.248260021209717,
+ "step": 4910
+ },
+ {
+ "epoch": 68.20960698689956,
+ "grad_norm": 0.017415888607501984,
+ "learning_rate": 0.0006,
+ "loss": 4.370295524597168,
+ "step": 4911
+ },
+ {
+ "epoch": 68.2235807860262,
+ "grad_norm": 0.015005858615040779,
+ "learning_rate": 0.0006,
+ "loss": 4.319567680358887,
+ "step": 4912
+ },
+ {
+ "epoch": 68.23755458515284,
+ "grad_norm": 0.016349777579307556,
+ "learning_rate": 0.0006,
+ "loss": 4.301776885986328,
+ "step": 4913
+ },
+ {
+ "epoch": 68.25152838427948,
+ "grad_norm": 0.015779560431838036,
+ "learning_rate": 0.0006,
+ "loss": 4.271171569824219,
+ "step": 4914
+ },
+ {
+ "epoch": 68.26550218340611,
+ "grad_norm": 0.013237644918262959,
+ "learning_rate": 0.0006,
+ "loss": 4.324908256530762,
+ "step": 4915
+ },
+ {
+ "epoch": 68.27947598253274,
+ "grad_norm": 0.014076225459575653,
+ "learning_rate": 0.0006,
+ "loss": 4.177175998687744,
+ "step": 4916
+ },
+ {
+ "epoch": 68.29344978165939,
+ "grad_norm": 0.013906744308769703,
+ "learning_rate": 0.0006,
+ "loss": 4.11588191986084,
+ "step": 4917
+ },
+ {
+ "epoch": 68.30742358078602,
+ "grad_norm": 0.013939064927399158,
+ "learning_rate": 0.0006,
+ "loss": 4.183963775634766,
+ "step": 4918
+ },
+ {
+ "epoch": 68.32139737991267,
+ "grad_norm": 0.014572170563042164,
+ "learning_rate": 0.0006,
+ "loss": 4.19586181640625,
+ "step": 4919
+ },
+ {
+ "epoch": 68.3353711790393,
+ "grad_norm": 0.014085350558161736,
+ "learning_rate": 0.0006,
+ "loss": 4.170892238616943,
+ "step": 4920
+ },
+ {
+ "epoch": 68.34934497816595,
+ "grad_norm": 0.014357863925397396,
+ "learning_rate": 0.0006,
+ "loss": 4.321155548095703,
+ "step": 4921
+ },
+ {
+ "epoch": 68.36331877729258,
+ "grad_norm": 0.01461675763130188,
+ "learning_rate": 0.0006,
+ "loss": 4.293334007263184,
+ "step": 4922
+ },
+ {
+ "epoch": 68.37729257641921,
+ "grad_norm": 0.014828486368060112,
+ "learning_rate": 0.0006,
+ "loss": 4.1439714431762695,
+ "step": 4923
+ },
+ {
+ "epoch": 68.39126637554585,
+ "grad_norm": 0.013730363920331001,
+ "learning_rate": 0.0006,
+ "loss": 4.287973880767822,
+ "step": 4924
+ },
+ {
+ "epoch": 68.40524017467249,
+ "grad_norm": 0.013852203264832497,
+ "learning_rate": 0.0006,
+ "loss": 4.248824119567871,
+ "step": 4925
+ },
+ {
+ "epoch": 68.41921397379913,
+ "grad_norm": 0.014005862176418304,
+ "learning_rate": 0.0006,
+ "loss": 4.293768882751465,
+ "step": 4926
+ },
+ {
+ "epoch": 68.43318777292576,
+ "grad_norm": 0.014435851015150547,
+ "learning_rate": 0.0006,
+ "loss": 4.204813003540039,
+ "step": 4927
+ },
+ {
+ "epoch": 68.44716157205241,
+ "grad_norm": 0.015305934473872185,
+ "learning_rate": 0.0006,
+ "loss": 4.329364776611328,
+ "step": 4928
+ },
+ {
+ "epoch": 68.46113537117904,
+ "grad_norm": 0.014356814324855804,
+ "learning_rate": 0.0006,
+ "loss": 4.264206409454346,
+ "step": 4929
+ },
+ {
+ "epoch": 68.47510917030567,
+ "grad_norm": 0.01432347297668457,
+ "learning_rate": 0.0006,
+ "loss": 4.29440450668335,
+ "step": 4930
+ },
+ {
+ "epoch": 68.48908296943232,
+ "grad_norm": 0.012672432698309422,
+ "learning_rate": 0.0006,
+ "loss": 4.258024215698242,
+ "step": 4931
+ },
+ {
+ "epoch": 68.50305676855895,
+ "grad_norm": 0.014209977351129055,
+ "learning_rate": 0.0006,
+ "loss": 4.216668128967285,
+ "step": 4932
+ },
+ {
+ "epoch": 68.5170305676856,
+ "grad_norm": 0.014606663957238197,
+ "learning_rate": 0.0006,
+ "loss": 4.294184684753418,
+ "step": 4933
+ },
+ {
+ "epoch": 68.53100436681223,
+ "grad_norm": 0.014986990951001644,
+ "learning_rate": 0.0006,
+ "loss": 4.382591247558594,
+ "step": 4934
+ },
+ {
+ "epoch": 68.54497816593886,
+ "grad_norm": 0.013998783193528652,
+ "learning_rate": 0.0006,
+ "loss": 4.195143222808838,
+ "step": 4935
+ },
+ {
+ "epoch": 68.5589519650655,
+ "grad_norm": 0.015011102892458439,
+ "learning_rate": 0.0006,
+ "loss": 4.313785076141357,
+ "step": 4936
+ },
+ {
+ "epoch": 68.57292576419214,
+ "grad_norm": 0.01641209051012993,
+ "learning_rate": 0.0006,
+ "loss": 4.322803020477295,
+ "step": 4937
+ },
+ {
+ "epoch": 68.58689956331878,
+ "grad_norm": 0.016207464039325714,
+ "learning_rate": 0.0006,
+ "loss": 4.311431884765625,
+ "step": 4938
+ },
+ {
+ "epoch": 68.60087336244541,
+ "grad_norm": 0.01747811958193779,
+ "learning_rate": 0.0006,
+ "loss": 4.310845375061035,
+ "step": 4939
+ },
+ {
+ "epoch": 68.61484716157206,
+ "grad_norm": 0.019475717097520828,
+ "learning_rate": 0.0006,
+ "loss": 4.338624954223633,
+ "step": 4940
+ },
+ {
+ "epoch": 68.62882096069869,
+ "grad_norm": 0.018641719594597816,
+ "learning_rate": 0.0006,
+ "loss": 4.34913969039917,
+ "step": 4941
+ },
+ {
+ "epoch": 68.64279475982532,
+ "grad_norm": 0.018376432359218597,
+ "learning_rate": 0.0006,
+ "loss": 4.28464412689209,
+ "step": 4942
+ },
+ {
+ "epoch": 68.65676855895197,
+ "grad_norm": 0.017095010727643967,
+ "learning_rate": 0.0006,
+ "loss": 4.217957496643066,
+ "step": 4943
+ },
+ {
+ "epoch": 68.6707423580786,
+ "grad_norm": 0.016321983188390732,
+ "learning_rate": 0.0006,
+ "loss": 4.220042705535889,
+ "step": 4944
+ },
+ {
+ "epoch": 68.68471615720524,
+ "grad_norm": 0.017128000035881996,
+ "learning_rate": 0.0006,
+ "loss": 4.293500900268555,
+ "step": 4945
+ },
+ {
+ "epoch": 68.69868995633188,
+ "grad_norm": 0.015592411160469055,
+ "learning_rate": 0.0006,
+ "loss": 4.195798873901367,
+ "step": 4946
+ },
+ {
+ "epoch": 68.71266375545852,
+ "grad_norm": 0.016607079654932022,
+ "learning_rate": 0.0006,
+ "loss": 4.2654337882995605,
+ "step": 4947
+ },
+ {
+ "epoch": 68.72663755458515,
+ "grad_norm": 0.01545762736350298,
+ "learning_rate": 0.0006,
+ "loss": 4.303828239440918,
+ "step": 4948
+ },
+ {
+ "epoch": 68.74061135371178,
+ "grad_norm": 0.014328422956168652,
+ "learning_rate": 0.0006,
+ "loss": 4.29379940032959,
+ "step": 4949
+ },
+ {
+ "epoch": 68.75458515283843,
+ "grad_norm": 0.016516804695129395,
+ "learning_rate": 0.0006,
+ "loss": 4.2317214012146,
+ "step": 4950
+ },
+ {
+ "epoch": 68.76855895196506,
+ "grad_norm": 0.016324171796441078,
+ "learning_rate": 0.0006,
+ "loss": 4.278043746948242,
+ "step": 4951
+ },
+ {
+ "epoch": 68.78253275109171,
+ "grad_norm": 0.015511159785091877,
+ "learning_rate": 0.0006,
+ "loss": 4.188244342803955,
+ "step": 4952
+ },
+ {
+ "epoch": 68.79650655021834,
+ "grad_norm": 0.01676103286445141,
+ "learning_rate": 0.0006,
+ "loss": 4.2169976234436035,
+ "step": 4953
+ },
+ {
+ "epoch": 68.81048034934499,
+ "grad_norm": 0.015428312122821808,
+ "learning_rate": 0.0006,
+ "loss": 4.4004716873168945,
+ "step": 4954
+ },
+ {
+ "epoch": 68.82445414847162,
+ "grad_norm": 0.0162410456687212,
+ "learning_rate": 0.0006,
+ "loss": 4.282752513885498,
+ "step": 4955
+ },
+ {
+ "epoch": 68.83842794759825,
+ "grad_norm": 0.017578862607479095,
+ "learning_rate": 0.0006,
+ "loss": 4.098004341125488,
+ "step": 4956
+ },
+ {
+ "epoch": 68.8524017467249,
+ "grad_norm": 0.01805749163031578,
+ "learning_rate": 0.0006,
+ "loss": 4.277988433837891,
+ "step": 4957
+ },
+ {
+ "epoch": 68.86637554585153,
+ "grad_norm": 0.016728045418858528,
+ "learning_rate": 0.0006,
+ "loss": 4.29766321182251,
+ "step": 4958
+ },
+ {
+ "epoch": 68.88034934497817,
+ "grad_norm": 0.014899657107889652,
+ "learning_rate": 0.0006,
+ "loss": 4.329736709594727,
+ "step": 4959
+ },
+ {
+ "epoch": 68.8943231441048,
+ "grad_norm": 0.013441305607557297,
+ "learning_rate": 0.0006,
+ "loss": 4.2873215675354,
+ "step": 4960
+ },
+ {
+ "epoch": 68.90829694323143,
+ "grad_norm": 0.01554128248244524,
+ "learning_rate": 0.0006,
+ "loss": 4.278327941894531,
+ "step": 4961
+ },
+ {
+ "epoch": 68.92227074235808,
+ "grad_norm": 0.01806696318089962,
+ "learning_rate": 0.0006,
+ "loss": 4.38275146484375,
+ "step": 4962
+ },
+ {
+ "epoch": 68.93624454148471,
+ "grad_norm": 0.017025409266352654,
+ "learning_rate": 0.0006,
+ "loss": 4.210447311401367,
+ "step": 4963
+ },
+ {
+ "epoch": 68.95021834061136,
+ "grad_norm": 0.01597420684993267,
+ "learning_rate": 0.0006,
+ "loss": 4.309865951538086,
+ "step": 4964
+ },
+ {
+ "epoch": 68.96419213973799,
+ "grad_norm": 0.015835624188184738,
+ "learning_rate": 0.0006,
+ "loss": 4.225904941558838,
+ "step": 4965
+ },
+ {
+ "epoch": 68.97816593886463,
+ "grad_norm": 0.018170492723584175,
+ "learning_rate": 0.0006,
+ "loss": 4.248478889465332,
+ "step": 4966
+ },
+ {
+ "epoch": 68.99213973799127,
+ "grad_norm": 0.017372991889715195,
+ "learning_rate": 0.0006,
+ "loss": 4.283486843109131,
+ "step": 4967
+ },
+ {
+ "epoch": 69.0,
+ "grad_norm": 0.016305390745401382,
+ "learning_rate": 0.0006,
+ "loss": 4.273636817932129,
+ "step": 4968
+ },
+ {
+ "epoch": 69.0,
+ "eval_loss": 4.66873836517334,
+ "eval_runtime": 57.0791,
+ "eval_samples_per_second": 42.783,
+ "eval_steps_per_second": 1.349,
+ "step": 4968
+ },
+ {
+ "epoch": 69.01397379912663,
+ "grad_norm": 0.016567355021834373,
+ "learning_rate": 0.0006,
+ "loss": 4.269988059997559,
+ "step": 4969
+ },
+ {
+ "epoch": 69.02794759825328,
+ "grad_norm": 0.017476852983236313,
+ "learning_rate": 0.0006,
+ "loss": 4.2518720626831055,
+ "step": 4970
+ },
+ {
+ "epoch": 69.04192139737991,
+ "grad_norm": 0.017070338129997253,
+ "learning_rate": 0.0006,
+ "loss": 4.238024711608887,
+ "step": 4971
+ },
+ {
+ "epoch": 69.05589519650655,
+ "grad_norm": 0.01639373227953911,
+ "learning_rate": 0.0006,
+ "loss": 4.188558101654053,
+ "step": 4972
+ },
+ {
+ "epoch": 69.06986899563319,
+ "grad_norm": 0.01856113411486149,
+ "learning_rate": 0.0006,
+ "loss": 4.339556694030762,
+ "step": 4973
+ },
+ {
+ "epoch": 69.08384279475983,
+ "grad_norm": 0.01762578636407852,
+ "learning_rate": 0.0006,
+ "loss": 4.141518592834473,
+ "step": 4974
+ },
+ {
+ "epoch": 69.09781659388646,
+ "grad_norm": 0.015437949448823929,
+ "learning_rate": 0.0006,
+ "loss": 4.225461959838867,
+ "step": 4975
+ },
+ {
+ "epoch": 69.1117903930131,
+ "grad_norm": 0.015293503180146217,
+ "learning_rate": 0.0006,
+ "loss": 4.290884971618652,
+ "step": 4976
+ },
+ {
+ "epoch": 69.12576419213974,
+ "grad_norm": 0.016259880736470222,
+ "learning_rate": 0.0006,
+ "loss": 4.121919631958008,
+ "step": 4977
+ },
+ {
+ "epoch": 69.13973799126637,
+ "grad_norm": 0.017225584015250206,
+ "learning_rate": 0.0006,
+ "loss": 4.136716842651367,
+ "step": 4978
+ },
+ {
+ "epoch": 69.15371179039302,
+ "grad_norm": 0.01699289306998253,
+ "learning_rate": 0.0006,
+ "loss": 4.193601608276367,
+ "step": 4979
+ },
+ {
+ "epoch": 69.16768558951965,
+ "grad_norm": 0.015613161027431488,
+ "learning_rate": 0.0006,
+ "loss": 4.288212299346924,
+ "step": 4980
+ },
+ {
+ "epoch": 69.18165938864628,
+ "grad_norm": 0.01643933728337288,
+ "learning_rate": 0.0006,
+ "loss": 4.324707984924316,
+ "step": 4981
+ },
+ {
+ "epoch": 69.19563318777293,
+ "grad_norm": 0.01648905873298645,
+ "learning_rate": 0.0006,
+ "loss": 4.287806510925293,
+ "step": 4982
+ },
+ {
+ "epoch": 69.20960698689956,
+ "grad_norm": 0.015444950200617313,
+ "learning_rate": 0.0006,
+ "loss": 4.200622081756592,
+ "step": 4983
+ },
+ {
+ "epoch": 69.2235807860262,
+ "grad_norm": 0.016034258529543877,
+ "learning_rate": 0.0006,
+ "loss": 4.2193827629089355,
+ "step": 4984
+ },
+ {
+ "epoch": 69.23755458515284,
+ "grad_norm": 0.017984388396143913,
+ "learning_rate": 0.0006,
+ "loss": 4.18237829208374,
+ "step": 4985
+ },
+ {
+ "epoch": 69.25152838427948,
+ "grad_norm": 0.0195505041629076,
+ "learning_rate": 0.0006,
+ "loss": 4.216727256774902,
+ "step": 4986
+ },
+ {
+ "epoch": 69.26550218340611,
+ "grad_norm": 0.017779147252440453,
+ "learning_rate": 0.0006,
+ "loss": 4.278628826141357,
+ "step": 4987
+ },
+ {
+ "epoch": 69.27947598253274,
+ "grad_norm": 0.017119476571679115,
+ "learning_rate": 0.0006,
+ "loss": 4.2409138679504395,
+ "step": 4988
+ },
+ {
+ "epoch": 69.29344978165939,
+ "grad_norm": 0.016840364784002304,
+ "learning_rate": 0.0006,
+ "loss": 4.207575798034668,
+ "step": 4989
+ },
+ {
+ "epoch": 69.30742358078602,
+ "grad_norm": 0.014593573287129402,
+ "learning_rate": 0.0006,
+ "loss": 4.164905548095703,
+ "step": 4990
+ },
+ {
+ "epoch": 69.32139737991267,
+ "grad_norm": 0.013922630809247494,
+ "learning_rate": 0.0006,
+ "loss": 4.2696990966796875,
+ "step": 4991
+ },
+ {
+ "epoch": 69.3353711790393,
+ "grad_norm": 0.01577564887702465,
+ "learning_rate": 0.0006,
+ "loss": 4.201589107513428,
+ "step": 4992
+ },
+ {
+ "epoch": 69.34934497816595,
+ "grad_norm": 0.016392096877098083,
+ "learning_rate": 0.0006,
+ "loss": 4.248831748962402,
+ "step": 4993
+ },
+ {
+ "epoch": 69.36331877729258,
+ "grad_norm": 0.015961311757564545,
+ "learning_rate": 0.0006,
+ "loss": 4.181260108947754,
+ "step": 4994
+ },
+ {
+ "epoch": 69.37729257641921,
+ "grad_norm": 0.015350551344454288,
+ "learning_rate": 0.0006,
+ "loss": 4.230352878570557,
+ "step": 4995
+ },
+ {
+ "epoch": 69.39126637554585,
+ "grad_norm": 0.015870466828346252,
+ "learning_rate": 0.0006,
+ "loss": 4.298305511474609,
+ "step": 4996
+ },
+ {
+ "epoch": 69.40524017467249,
+ "grad_norm": 0.013716642744839191,
+ "learning_rate": 0.0006,
+ "loss": 4.220139503479004,
+ "step": 4997
+ },
+ {
+ "epoch": 69.41921397379913,
+ "grad_norm": 0.014431777410209179,
+ "learning_rate": 0.0006,
+ "loss": 4.288130760192871,
+ "step": 4998
+ },
+ {
+ "epoch": 69.43318777292576,
+ "grad_norm": 0.01426647324115038,
+ "learning_rate": 0.0006,
+ "loss": 4.295767784118652,
+ "step": 4999
+ },
+ {
+ "epoch": 69.44716157205241,
+ "grad_norm": 0.014286615885794163,
+ "learning_rate": 0.0006,
+ "loss": 4.38106632232666,
+ "step": 5000
+ },
+ {
+ "epoch": 69.46113537117904,
+ "grad_norm": 0.017990585416555405,
+ "learning_rate": 0.0006,
+ "loss": 4.1593017578125,
+ "step": 5001
+ },
+ {
+ "epoch": 69.47510917030567,
+ "grad_norm": 0.018846049904823303,
+ "learning_rate": 0.0006,
+ "loss": 4.216091632843018,
+ "step": 5002
+ },
+ {
+ "epoch": 69.48908296943232,
+ "grad_norm": 0.018179181963205338,
+ "learning_rate": 0.0006,
+ "loss": 4.240332126617432,
+ "step": 5003
+ },
+ {
+ "epoch": 69.50305676855895,
+ "grad_norm": 0.016693448647856712,
+ "learning_rate": 0.0006,
+ "loss": 4.256624221801758,
+ "step": 5004
+ },
+ {
+ "epoch": 69.5170305676856,
+ "grad_norm": 0.01718989387154579,
+ "learning_rate": 0.0006,
+ "loss": 4.300940990447998,
+ "step": 5005
+ },
+ {
+ "epoch": 69.53100436681223,
+ "grad_norm": 0.020236829295754433,
+ "learning_rate": 0.0006,
+ "loss": 4.289146423339844,
+ "step": 5006
+ },
+ {
+ "epoch": 69.54497816593886,
+ "grad_norm": 0.024860285222530365,
+ "learning_rate": 0.0006,
+ "loss": 4.216216087341309,
+ "step": 5007
+ },
+ {
+ "epoch": 69.5589519650655,
+ "grad_norm": 0.02559969387948513,
+ "learning_rate": 0.0006,
+ "loss": 4.262890815734863,
+ "step": 5008
+ },
+ {
+ "epoch": 69.57292576419214,
+ "grad_norm": 0.019180957227945328,
+ "learning_rate": 0.0006,
+ "loss": 4.230309009552002,
+ "step": 5009
+ },
+ {
+ "epoch": 69.58689956331878,
+ "grad_norm": 0.020442591980099678,
+ "learning_rate": 0.0006,
+ "loss": 4.291759490966797,
+ "step": 5010
+ },
+ {
+ "epoch": 69.60087336244541,
+ "grad_norm": 0.0212619137018919,
+ "learning_rate": 0.0006,
+ "loss": 4.3529839515686035,
+ "step": 5011
+ },
+ {
+ "epoch": 69.61484716157206,
+ "grad_norm": 0.02067175693809986,
+ "learning_rate": 0.0006,
+ "loss": 4.33197021484375,
+ "step": 5012
+ },
+ {
+ "epoch": 69.62882096069869,
+ "grad_norm": 0.01802872307598591,
+ "learning_rate": 0.0006,
+ "loss": 4.2101969718933105,
+ "step": 5013
+ },
+ {
+ "epoch": 69.64279475982532,
+ "grad_norm": 0.01787702925503254,
+ "learning_rate": 0.0006,
+ "loss": 4.270336627960205,
+ "step": 5014
+ },
+ {
+ "epoch": 69.65676855895197,
+ "grad_norm": 0.01586683839559555,
+ "learning_rate": 0.0006,
+ "loss": 4.222309589385986,
+ "step": 5015
+ },
+ {
+ "epoch": 69.6707423580786,
+ "grad_norm": 0.01717999391257763,
+ "learning_rate": 0.0006,
+ "loss": 4.253704071044922,
+ "step": 5016
+ },
+ {
+ "epoch": 69.68471615720524,
+ "grad_norm": 0.01594279147684574,
+ "learning_rate": 0.0006,
+ "loss": 4.344658851623535,
+ "step": 5017
+ },
+ {
+ "epoch": 69.69868995633188,
+ "grad_norm": 0.014419942162930965,
+ "learning_rate": 0.0006,
+ "loss": 4.2144880294799805,
+ "step": 5018
+ },
+ {
+ "epoch": 69.71266375545852,
+ "grad_norm": 0.015824224799871445,
+ "learning_rate": 0.0006,
+ "loss": 4.226252555847168,
+ "step": 5019
+ },
+ {
+ "epoch": 69.72663755458515,
+ "grad_norm": 0.013329979032278061,
+ "learning_rate": 0.0006,
+ "loss": 4.238587379455566,
+ "step": 5020
+ },
+ {
+ "epoch": 69.74061135371178,
+ "grad_norm": 0.013745979405939579,
+ "learning_rate": 0.0006,
+ "loss": 4.317209243774414,
+ "step": 5021
+ },
+ {
+ "epoch": 69.75458515283843,
+ "grad_norm": 0.014064167626202106,
+ "learning_rate": 0.0006,
+ "loss": 4.316982746124268,
+ "step": 5022
+ },
+ {
+ "epoch": 69.76855895196506,
+ "grad_norm": 0.014611301012337208,
+ "learning_rate": 0.0006,
+ "loss": 4.212728500366211,
+ "step": 5023
+ },
+ {
+ "epoch": 69.78253275109171,
+ "grad_norm": 0.014947416260838509,
+ "learning_rate": 0.0006,
+ "loss": 4.319550514221191,
+ "step": 5024
+ },
+ {
+ "epoch": 69.79650655021834,
+ "grad_norm": 0.014780526980757713,
+ "learning_rate": 0.0006,
+ "loss": 4.1888933181762695,
+ "step": 5025
+ },
+ {
+ "epoch": 69.81048034934499,
+ "grad_norm": 0.01611459068953991,
+ "learning_rate": 0.0006,
+ "loss": 4.239692211151123,
+ "step": 5026
+ },
+ {
+ "epoch": 69.82445414847162,
+ "grad_norm": 0.015395848080515862,
+ "learning_rate": 0.0006,
+ "loss": 4.308794021606445,
+ "step": 5027
+ },
+ {
+ "epoch": 69.83842794759825,
+ "grad_norm": 0.016113396733999252,
+ "learning_rate": 0.0006,
+ "loss": 4.328035354614258,
+ "step": 5028
+ },
+ {
+ "epoch": 69.8524017467249,
+ "grad_norm": 0.016386888921260834,
+ "learning_rate": 0.0006,
+ "loss": 4.272579193115234,
+ "step": 5029
+ },
+ {
+ "epoch": 69.86637554585153,
+ "grad_norm": 0.013675205409526825,
+ "learning_rate": 0.0006,
+ "loss": 4.226552963256836,
+ "step": 5030
+ },
+ {
+ "epoch": 69.88034934497817,
+ "grad_norm": 0.01472668256610632,
+ "learning_rate": 0.0006,
+ "loss": 4.224167346954346,
+ "step": 5031
+ },
+ {
+ "epoch": 69.8943231441048,
+ "grad_norm": 0.014958228915929794,
+ "learning_rate": 0.0006,
+ "loss": 4.232619285583496,
+ "step": 5032
+ },
+ {
+ "epoch": 69.90829694323143,
+ "grad_norm": 0.012926832772791386,
+ "learning_rate": 0.0006,
+ "loss": 4.27107048034668,
+ "step": 5033
+ },
+ {
+ "epoch": 69.92227074235808,
+ "grad_norm": 0.013138801790773869,
+ "learning_rate": 0.0006,
+ "loss": 4.288477420806885,
+ "step": 5034
+ },
+ {
+ "epoch": 69.93624454148471,
+ "grad_norm": 0.013415461406111717,
+ "learning_rate": 0.0006,
+ "loss": 4.2033257484436035,
+ "step": 5035
+ },
+ {
+ "epoch": 69.95021834061136,
+ "grad_norm": 0.013807233422994614,
+ "learning_rate": 0.0006,
+ "loss": 4.274764537811279,
+ "step": 5036
+ },
+ {
+ "epoch": 69.96419213973799,
+ "grad_norm": 0.013929193839430809,
+ "learning_rate": 0.0006,
+ "loss": 4.1854352951049805,
+ "step": 5037
+ },
+ {
+ "epoch": 69.97816593886463,
+ "grad_norm": 0.015497001819312572,
+ "learning_rate": 0.0006,
+ "loss": 4.226498126983643,
+ "step": 5038
+ },
+ {
+ "epoch": 69.99213973799127,
+ "grad_norm": 0.01874048449099064,
+ "learning_rate": 0.0006,
+ "loss": 4.304137706756592,
+ "step": 5039
+ },
+ {
+ "epoch": 70.0,
+ "grad_norm": 0.021745791658759117,
+ "learning_rate": 0.0006,
+ "loss": 4.206809997558594,
+ "step": 5040
+ },
+ {
+ "epoch": 70.0,
+ "eval_loss": 4.675539970397949,
+ "eval_runtime": 56.9034,
+ "eval_samples_per_second": 42.915,
+ "eval_steps_per_second": 1.353,
+ "step": 5040
+ },
+ {
+ "epoch": 70.01397379912663,
+ "grad_norm": 0.02124839462339878,
+ "learning_rate": 0.0006,
+ "loss": 4.198199272155762,
+ "step": 5041
+ },
+ {
+ "epoch": 70.02794759825328,
+ "grad_norm": 0.019432688131928444,
+ "learning_rate": 0.0006,
+ "loss": 4.187261581420898,
+ "step": 5042
+ },
+ {
+ "epoch": 70.04192139737991,
+ "grad_norm": 0.01769290305674076,
+ "learning_rate": 0.0006,
+ "loss": 4.155284881591797,
+ "step": 5043
+ },
+ {
+ "epoch": 70.05589519650655,
+ "grad_norm": 0.01751520484685898,
+ "learning_rate": 0.0006,
+ "loss": 4.386857986450195,
+ "step": 5044
+ },
+ {
+ "epoch": 70.06986899563319,
+ "grad_norm": 0.018138699233531952,
+ "learning_rate": 0.0006,
+ "loss": 4.117342472076416,
+ "step": 5045
+ },
+ {
+ "epoch": 70.08384279475983,
+ "grad_norm": 0.016331054270267487,
+ "learning_rate": 0.0006,
+ "loss": 4.225295066833496,
+ "step": 5046
+ },
+ {
+ "epoch": 70.09781659388646,
+ "grad_norm": 0.01658729277551174,
+ "learning_rate": 0.0006,
+ "loss": 4.143835067749023,
+ "step": 5047
+ },
+ {
+ "epoch": 70.1117903930131,
+ "grad_norm": 0.016367698088288307,
+ "learning_rate": 0.0006,
+ "loss": 4.1469268798828125,
+ "step": 5048
+ },
+ {
+ "epoch": 70.12576419213974,
+ "grad_norm": 0.015151629224419594,
+ "learning_rate": 0.0006,
+ "loss": 4.117629051208496,
+ "step": 5049
+ },
+ {
+ "epoch": 70.13973799126637,
+ "grad_norm": 0.016434503719210625,
+ "learning_rate": 0.0006,
+ "loss": 4.107848644256592,
+ "step": 5050
+ },
+ {
+ "epoch": 70.15371179039302,
+ "grad_norm": 0.017837485298514366,
+ "learning_rate": 0.0006,
+ "loss": 4.10837459564209,
+ "step": 5051
+ },
+ {
+ "epoch": 70.16768558951965,
+ "grad_norm": 0.0165961142629385,
+ "learning_rate": 0.0006,
+ "loss": 4.169414043426514,
+ "step": 5052
+ },
+ {
+ "epoch": 70.18165938864628,
+ "grad_norm": 0.0157014112919569,
+ "learning_rate": 0.0006,
+ "loss": 4.0629706382751465,
+ "step": 5053
+ },
+ {
+ "epoch": 70.19563318777293,
+ "grad_norm": 0.01581883803009987,
+ "learning_rate": 0.0006,
+ "loss": 4.220012664794922,
+ "step": 5054
+ },
+ {
+ "epoch": 70.20960698689956,
+ "grad_norm": 0.015285917557775974,
+ "learning_rate": 0.0006,
+ "loss": 4.275337219238281,
+ "step": 5055
+ },
+ {
+ "epoch": 70.2235807860262,
+ "grad_norm": 0.014984549954533577,
+ "learning_rate": 0.0006,
+ "loss": 4.284764289855957,
+ "step": 5056
+ },
+ {
+ "epoch": 70.23755458515284,
+ "grad_norm": 0.015172012150287628,
+ "learning_rate": 0.0006,
+ "loss": 4.234498023986816,
+ "step": 5057
+ },
+ {
+ "epoch": 70.25152838427948,
+ "grad_norm": 0.016846632584929466,
+ "learning_rate": 0.0006,
+ "loss": 4.181689262390137,
+ "step": 5058
+ },
+ {
+ "epoch": 70.26550218340611,
+ "grad_norm": 0.01920711249113083,
+ "learning_rate": 0.0006,
+ "loss": 4.235280513763428,
+ "step": 5059
+ },
+ {
+ "epoch": 70.27947598253274,
+ "grad_norm": 0.017765291035175323,
+ "learning_rate": 0.0006,
+ "loss": 4.238955497741699,
+ "step": 5060
+ },
+ {
+ "epoch": 70.29344978165939,
+ "grad_norm": 0.015436639077961445,
+ "learning_rate": 0.0006,
+ "loss": 4.232516288757324,
+ "step": 5061
+ },
+ {
+ "epoch": 70.30742358078602,
+ "grad_norm": 0.015660086646676064,
+ "learning_rate": 0.0006,
+ "loss": 4.260349273681641,
+ "step": 5062
+ },
+ {
+ "epoch": 70.32139737991267,
+ "grad_norm": 0.016095709055662155,
+ "learning_rate": 0.0006,
+ "loss": 4.168505668640137,
+ "step": 5063
+ },
+ {
+ "epoch": 70.3353711790393,
+ "grad_norm": 0.015001940540969372,
+ "learning_rate": 0.0006,
+ "loss": 4.161739349365234,
+ "step": 5064
+ },
+ {
+ "epoch": 70.34934497816595,
+ "grad_norm": 0.014668774791061878,
+ "learning_rate": 0.0006,
+ "loss": 4.181029796600342,
+ "step": 5065
+ },
+ {
+ "epoch": 70.36331877729258,
+ "grad_norm": 0.013266284950077534,
+ "learning_rate": 0.0006,
+ "loss": 4.3052873611450195,
+ "step": 5066
+ },
+ {
+ "epoch": 70.37729257641921,
+ "grad_norm": 0.015396138653159142,
+ "learning_rate": 0.0006,
+ "loss": 4.34413480758667,
+ "step": 5067
+ },
+ {
+ "epoch": 70.39126637554585,
+ "grad_norm": 0.015511523932218552,
+ "learning_rate": 0.0006,
+ "loss": 4.235037803649902,
+ "step": 5068
+ },
+ {
+ "epoch": 70.40524017467249,
+ "grad_norm": 0.014825738966464996,
+ "learning_rate": 0.0006,
+ "loss": 4.262448310852051,
+ "step": 5069
+ },
+ {
+ "epoch": 70.41921397379913,
+ "grad_norm": 0.015309952199459076,
+ "learning_rate": 0.0006,
+ "loss": 4.258711814880371,
+ "step": 5070
+ },
+ {
+ "epoch": 70.43318777292576,
+ "grad_norm": 0.017614055424928665,
+ "learning_rate": 0.0006,
+ "loss": 4.286028861999512,
+ "step": 5071
+ },
+ {
+ "epoch": 70.44716157205241,
+ "grad_norm": 0.0166039876639843,
+ "learning_rate": 0.0006,
+ "loss": 4.149197578430176,
+ "step": 5072
+ },
+ {
+ "epoch": 70.46113537117904,
+ "grad_norm": 0.016911419108510017,
+ "learning_rate": 0.0006,
+ "loss": 4.271981239318848,
+ "step": 5073
+ },
+ {
+ "epoch": 70.47510917030567,
+ "grad_norm": 0.016806820407509804,
+ "learning_rate": 0.0006,
+ "loss": 4.213315010070801,
+ "step": 5074
+ },
+ {
+ "epoch": 70.48908296943232,
+ "grad_norm": 0.015221742913126945,
+ "learning_rate": 0.0006,
+ "loss": 4.171091079711914,
+ "step": 5075
+ },
+ {
+ "epoch": 70.50305676855895,
+ "grad_norm": 0.014175347983837128,
+ "learning_rate": 0.0006,
+ "loss": 4.185309410095215,
+ "step": 5076
+ },
+ {
+ "epoch": 70.5170305676856,
+ "grad_norm": 0.013897798955440521,
+ "learning_rate": 0.0006,
+ "loss": 4.250119209289551,
+ "step": 5077
+ },
+ {
+ "epoch": 70.53100436681223,
+ "grad_norm": 0.014791003428399563,
+ "learning_rate": 0.0006,
+ "loss": 4.2504682540893555,
+ "step": 5078
+ },
+ {
+ "epoch": 70.54497816593886,
+ "grad_norm": 0.016700128093361855,
+ "learning_rate": 0.0006,
+ "loss": 4.235134124755859,
+ "step": 5079
+ },
+ {
+ "epoch": 70.5589519650655,
+ "grad_norm": 0.016679363325238228,
+ "learning_rate": 0.0006,
+ "loss": 4.32028865814209,
+ "step": 5080
+ },
+ {
+ "epoch": 70.57292576419214,
+ "grad_norm": 0.015117401257157326,
+ "learning_rate": 0.0006,
+ "loss": 4.210995674133301,
+ "step": 5081
+ },
+ {
+ "epoch": 70.58689956331878,
+ "grad_norm": 0.014441263861954212,
+ "learning_rate": 0.0006,
+ "loss": 4.27849006652832,
+ "step": 5082
+ },
+ {
+ "epoch": 70.60087336244541,
+ "grad_norm": 0.01523780357092619,
+ "learning_rate": 0.0006,
+ "loss": 4.197360992431641,
+ "step": 5083
+ },
+ {
+ "epoch": 70.61484716157206,
+ "grad_norm": 0.016201738268136978,
+ "learning_rate": 0.0006,
+ "loss": 4.28239107131958,
+ "step": 5084
+ },
+ {
+ "epoch": 70.62882096069869,
+ "grad_norm": 0.016003292053937912,
+ "learning_rate": 0.0006,
+ "loss": 4.209715843200684,
+ "step": 5085
+ },
+ {
+ "epoch": 70.64279475982532,
+ "grad_norm": 0.01472893450409174,
+ "learning_rate": 0.0006,
+ "loss": 4.169078826904297,
+ "step": 5086
+ },
+ {
+ "epoch": 70.65676855895197,
+ "grad_norm": 0.01604326069355011,
+ "learning_rate": 0.0006,
+ "loss": 4.090967178344727,
+ "step": 5087
+ },
+ {
+ "epoch": 70.6707423580786,
+ "grad_norm": 0.01598356105387211,
+ "learning_rate": 0.0006,
+ "loss": 4.16630220413208,
+ "step": 5088
+ },
+ {
+ "epoch": 70.68471615720524,
+ "grad_norm": 0.016988780349493027,
+ "learning_rate": 0.0006,
+ "loss": 4.117822170257568,
+ "step": 5089
+ },
+ {
+ "epoch": 70.69868995633188,
+ "grad_norm": 0.018582893535494804,
+ "learning_rate": 0.0006,
+ "loss": 4.192902088165283,
+ "step": 5090
+ },
+ {
+ "epoch": 70.71266375545852,
+ "grad_norm": 0.01793155074119568,
+ "learning_rate": 0.0006,
+ "loss": 4.314806938171387,
+ "step": 5091
+ },
+ {
+ "epoch": 70.72663755458515,
+ "grad_norm": 0.016286427155137062,
+ "learning_rate": 0.0006,
+ "loss": 4.2137556076049805,
+ "step": 5092
+ },
+ {
+ "epoch": 70.74061135371178,
+ "grad_norm": 0.014718376100063324,
+ "learning_rate": 0.0006,
+ "loss": 4.3073015213012695,
+ "step": 5093
+ },
+ {
+ "epoch": 70.75458515283843,
+ "grad_norm": 0.014675077050924301,
+ "learning_rate": 0.0006,
+ "loss": 4.160961151123047,
+ "step": 5094
+ },
+ {
+ "epoch": 70.76855895196506,
+ "grad_norm": 0.015550913289189339,
+ "learning_rate": 0.0006,
+ "loss": 4.322887420654297,
+ "step": 5095
+ },
+ {
+ "epoch": 70.78253275109171,
+ "grad_norm": 0.01453235000371933,
+ "learning_rate": 0.0006,
+ "loss": 4.2915873527526855,
+ "step": 5096
+ },
+ {
+ "epoch": 70.79650655021834,
+ "grad_norm": 0.01459164172410965,
+ "learning_rate": 0.0006,
+ "loss": 4.247288227081299,
+ "step": 5097
+ },
+ {
+ "epoch": 70.81048034934499,
+ "grad_norm": 0.015395080670714378,
+ "learning_rate": 0.0006,
+ "loss": 4.360955715179443,
+ "step": 5098
+ },
+ {
+ "epoch": 70.82445414847162,
+ "grad_norm": 0.01317683607339859,
+ "learning_rate": 0.0006,
+ "loss": 4.181969165802002,
+ "step": 5099
+ },
+ {
+ "epoch": 70.83842794759825,
+ "grad_norm": 0.013917932286858559,
+ "learning_rate": 0.0006,
+ "loss": 4.264626979827881,
+ "step": 5100
+ },
+ {
+ "epoch": 70.8524017467249,
+ "grad_norm": 0.015268695540726185,
+ "learning_rate": 0.0006,
+ "loss": 4.287857532501221,
+ "step": 5101
+ },
+ {
+ "epoch": 70.86637554585153,
+ "grad_norm": 0.015531329438090324,
+ "learning_rate": 0.0006,
+ "loss": 4.294188499450684,
+ "step": 5102
+ },
+ {
+ "epoch": 70.88034934497817,
+ "grad_norm": 0.015607129782438278,
+ "learning_rate": 0.0006,
+ "loss": 4.331040859222412,
+ "step": 5103
+ },
+ {
+ "epoch": 70.8943231441048,
+ "grad_norm": 0.015297149308025837,
+ "learning_rate": 0.0006,
+ "loss": 4.243650436401367,
+ "step": 5104
+ },
+ {
+ "epoch": 70.90829694323143,
+ "grad_norm": 0.014790025539696217,
+ "learning_rate": 0.0006,
+ "loss": 4.159859657287598,
+ "step": 5105
+ },
+ {
+ "epoch": 70.92227074235808,
+ "grad_norm": 0.013466274365782738,
+ "learning_rate": 0.0006,
+ "loss": 4.221992492675781,
+ "step": 5106
+ },
+ {
+ "epoch": 70.93624454148471,
+ "grad_norm": 0.01361898984760046,
+ "learning_rate": 0.0006,
+ "loss": 4.278751373291016,
+ "step": 5107
+ },
+ {
+ "epoch": 70.95021834061136,
+ "grad_norm": 0.01325751468539238,
+ "learning_rate": 0.0006,
+ "loss": 4.166001319885254,
+ "step": 5108
+ },
+ {
+ "epoch": 70.96419213973799,
+ "grad_norm": 0.012727087363600731,
+ "learning_rate": 0.0006,
+ "loss": 4.219997406005859,
+ "step": 5109
+ },
+ {
+ "epoch": 70.97816593886463,
+ "grad_norm": 0.013966867700219154,
+ "learning_rate": 0.0006,
+ "loss": 4.207491397857666,
+ "step": 5110
+ },
+ {
+ "epoch": 70.99213973799127,
+ "grad_norm": 0.015053587965667248,
+ "learning_rate": 0.0006,
+ "loss": 4.343245983123779,
+ "step": 5111
+ },
+ {
+ "epoch": 71.0,
+ "grad_norm": 0.017723368480801582,
+ "learning_rate": 0.0006,
+ "loss": 4.356122016906738,
+ "step": 5112
+ },
+ {
+ "epoch": 71.0,
+ "eval_loss": 4.598303318023682,
+ "eval_runtime": 56.7521,
+ "eval_samples_per_second": 43.029,
+ "eval_steps_per_second": 1.357,
+ "step": 5112
+ },
+ {
+ "epoch": 71.01397379912663,
+ "grad_norm": 0.018996279686689377,
+ "learning_rate": 0.0006,
+ "loss": 4.234772205352783,
+ "step": 5113
+ },
+ {
+ "epoch": 71.02794759825328,
+ "grad_norm": 0.02041555754840374,
+ "learning_rate": 0.0006,
+ "loss": 4.275816917419434,
+ "step": 5114
+ },
+ {
+ "epoch": 71.04192139737991,
+ "grad_norm": 0.021955285221338272,
+ "learning_rate": 0.0006,
+ "loss": 4.166038513183594,
+ "step": 5115
+ },
+ {
+ "epoch": 71.05589519650655,
+ "grad_norm": 0.021308740600943565,
+ "learning_rate": 0.0006,
+ "loss": 4.13063907623291,
+ "step": 5116
+ },
+ {
+ "epoch": 71.06986899563319,
+ "grad_norm": 0.01747054047882557,
+ "learning_rate": 0.0006,
+ "loss": 4.200522422790527,
+ "step": 5117
+ },
+ {
+ "epoch": 71.08384279475983,
+ "grad_norm": 0.018544532358646393,
+ "learning_rate": 0.0006,
+ "loss": 4.173915386199951,
+ "step": 5118
+ },
+ {
+ "epoch": 71.09781659388646,
+ "grad_norm": 0.0182361900806427,
+ "learning_rate": 0.0006,
+ "loss": 4.27392578125,
+ "step": 5119
+ },
+ {
+ "epoch": 71.1117903930131,
+ "grad_norm": 0.017062073573470116,
+ "learning_rate": 0.0006,
+ "loss": 4.219576358795166,
+ "step": 5120
+ },
+ {
+ "epoch": 71.12576419213974,
+ "grad_norm": 0.016289666295051575,
+ "learning_rate": 0.0006,
+ "loss": 4.133376598358154,
+ "step": 5121
+ },
+ {
+ "epoch": 71.13973799126637,
+ "grad_norm": 0.014864642173051834,
+ "learning_rate": 0.0006,
+ "loss": 4.168276786804199,
+ "step": 5122
+ },
+ {
+ "epoch": 71.15371179039302,
+ "grad_norm": 0.015309540554881096,
+ "learning_rate": 0.0006,
+ "loss": 4.175103187561035,
+ "step": 5123
+ },
+ {
+ "epoch": 71.16768558951965,
+ "grad_norm": 0.01556059904396534,
+ "learning_rate": 0.0006,
+ "loss": 4.289899826049805,
+ "step": 5124
+ },
+ {
+ "epoch": 71.18165938864628,
+ "grad_norm": 0.01479957066476345,
+ "learning_rate": 0.0006,
+ "loss": 4.341495037078857,
+ "step": 5125
+ },
+ {
+ "epoch": 71.19563318777293,
+ "grad_norm": 0.014762179926037788,
+ "learning_rate": 0.0006,
+ "loss": 4.252333641052246,
+ "step": 5126
+ },
+ {
+ "epoch": 71.20960698689956,
+ "grad_norm": 0.014897473156452179,
+ "learning_rate": 0.0006,
+ "loss": 4.2244181632995605,
+ "step": 5127
+ },
+ {
+ "epoch": 71.2235807860262,
+ "grad_norm": 0.012990163639187813,
+ "learning_rate": 0.0006,
+ "loss": 4.237288475036621,
+ "step": 5128
+ },
+ {
+ "epoch": 71.23755458515284,
+ "grad_norm": 0.012858416885137558,
+ "learning_rate": 0.0006,
+ "loss": 4.265738487243652,
+ "step": 5129
+ },
+ {
+ "epoch": 71.25152838427948,
+ "grad_norm": 0.013454788364470005,
+ "learning_rate": 0.0006,
+ "loss": 4.289131164550781,
+ "step": 5130
+ },
+ {
+ "epoch": 71.26550218340611,
+ "grad_norm": 0.013717238791286945,
+ "learning_rate": 0.0006,
+ "loss": 4.2844014167785645,
+ "step": 5131
+ },
+ {
+ "epoch": 71.27947598253274,
+ "grad_norm": 0.0135411461815238,
+ "learning_rate": 0.0006,
+ "loss": 4.139082908630371,
+ "step": 5132
+ },
+ {
+ "epoch": 71.29344978165939,
+ "grad_norm": 0.01406062301248312,
+ "learning_rate": 0.0006,
+ "loss": 4.226719856262207,
+ "step": 5133
+ },
+ {
+ "epoch": 71.30742358078602,
+ "grad_norm": 0.014027190394699574,
+ "learning_rate": 0.0006,
+ "loss": 4.191699981689453,
+ "step": 5134
+ },
+ {
+ "epoch": 71.32139737991267,
+ "grad_norm": 0.013809907250106335,
+ "learning_rate": 0.0006,
+ "loss": 4.242542266845703,
+ "step": 5135
+ },
+ {
+ "epoch": 71.3353711790393,
+ "grad_norm": 0.013507804833352566,
+ "learning_rate": 0.0006,
+ "loss": 4.068206310272217,
+ "step": 5136
+ },
+ {
+ "epoch": 71.34934497816595,
+ "grad_norm": 0.01432533748447895,
+ "learning_rate": 0.0006,
+ "loss": 4.378222942352295,
+ "step": 5137
+ },
+ {
+ "epoch": 71.36331877729258,
+ "grad_norm": 0.01753002591431141,
+ "learning_rate": 0.0006,
+ "loss": 4.174706935882568,
+ "step": 5138
+ },
+ {
+ "epoch": 71.37729257641921,
+ "grad_norm": 0.020048344507813454,
+ "learning_rate": 0.0006,
+ "loss": 4.19920015335083,
+ "step": 5139
+ },
+ {
+ "epoch": 71.39126637554585,
+ "grad_norm": 0.018551401793956757,
+ "learning_rate": 0.0006,
+ "loss": 4.175182342529297,
+ "step": 5140
+ },
+ {
+ "epoch": 71.40524017467249,
+ "grad_norm": 0.01598959229886532,
+ "learning_rate": 0.0006,
+ "loss": 4.2305169105529785,
+ "step": 5141
+ },
+ {
+ "epoch": 71.41921397379913,
+ "grad_norm": 0.016029560938477516,
+ "learning_rate": 0.0006,
+ "loss": 4.230090618133545,
+ "step": 5142
+ },
+ {
+ "epoch": 71.43318777292576,
+ "grad_norm": 0.01595657505095005,
+ "learning_rate": 0.0006,
+ "loss": 4.138402938842773,
+ "step": 5143
+ },
+ {
+ "epoch": 71.44716157205241,
+ "grad_norm": 0.01623230241239071,
+ "learning_rate": 0.0006,
+ "loss": 4.196106433868408,
+ "step": 5144
+ },
+ {
+ "epoch": 71.46113537117904,
+ "grad_norm": 0.014057597145438194,
+ "learning_rate": 0.0006,
+ "loss": 4.2599897384643555,
+ "step": 5145
+ },
+ {
+ "epoch": 71.47510917030567,
+ "grad_norm": 0.017161158844828606,
+ "learning_rate": 0.0006,
+ "loss": 4.183909893035889,
+ "step": 5146
+ },
+ {
+ "epoch": 71.48908296943232,
+ "grad_norm": 0.017038581892848015,
+ "learning_rate": 0.0006,
+ "loss": 4.373933792114258,
+ "step": 5147
+ },
+ {
+ "epoch": 71.50305676855895,
+ "grad_norm": 0.015851367264986038,
+ "learning_rate": 0.0006,
+ "loss": 4.2620086669921875,
+ "step": 5148
+ },
+ {
+ "epoch": 71.5170305676856,
+ "grad_norm": 0.014598303474485874,
+ "learning_rate": 0.0006,
+ "loss": 4.197151184082031,
+ "step": 5149
+ },
+ {
+ "epoch": 71.53100436681223,
+ "grad_norm": 0.01339802611619234,
+ "learning_rate": 0.0006,
+ "loss": 4.110706329345703,
+ "step": 5150
+ },
+ {
+ "epoch": 71.54497816593886,
+ "grad_norm": 0.015567775815725327,
+ "learning_rate": 0.0006,
+ "loss": 4.1210246086120605,
+ "step": 5151
+ },
+ {
+ "epoch": 71.5589519650655,
+ "grad_norm": 0.015616007149219513,
+ "learning_rate": 0.0006,
+ "loss": 4.129796028137207,
+ "step": 5152
+ },
+ {
+ "epoch": 71.57292576419214,
+ "grad_norm": 0.014198964461684227,
+ "learning_rate": 0.0006,
+ "loss": 4.193232536315918,
+ "step": 5153
+ },
+ {
+ "epoch": 71.58689956331878,
+ "grad_norm": 0.015226383693516254,
+ "learning_rate": 0.0006,
+ "loss": 4.223779678344727,
+ "step": 5154
+ },
+ {
+ "epoch": 71.60087336244541,
+ "grad_norm": 0.015610109083354473,
+ "learning_rate": 0.0006,
+ "loss": 4.223179817199707,
+ "step": 5155
+ },
+ {
+ "epoch": 71.61484716157206,
+ "grad_norm": 0.016034303233027458,
+ "learning_rate": 0.0006,
+ "loss": 4.196593284606934,
+ "step": 5156
+ },
+ {
+ "epoch": 71.62882096069869,
+ "grad_norm": 0.015648581087589264,
+ "learning_rate": 0.0006,
+ "loss": 4.341497421264648,
+ "step": 5157
+ },
+ {
+ "epoch": 71.64279475982532,
+ "grad_norm": 0.014117540791630745,
+ "learning_rate": 0.0006,
+ "loss": 4.194307327270508,
+ "step": 5158
+ },
+ {
+ "epoch": 71.65676855895197,
+ "grad_norm": 0.015612686984241009,
+ "learning_rate": 0.0006,
+ "loss": 4.290152549743652,
+ "step": 5159
+ },
+ {
+ "epoch": 71.6707423580786,
+ "grad_norm": 0.01722259446978569,
+ "learning_rate": 0.0006,
+ "loss": 4.258018493652344,
+ "step": 5160
+ },
+ {
+ "epoch": 71.68471615720524,
+ "grad_norm": 0.015923351049423218,
+ "learning_rate": 0.0006,
+ "loss": 4.21684455871582,
+ "step": 5161
+ },
+ {
+ "epoch": 71.69868995633188,
+ "grad_norm": 0.013663450255990028,
+ "learning_rate": 0.0006,
+ "loss": 4.175601959228516,
+ "step": 5162
+ },
+ {
+ "epoch": 71.71266375545852,
+ "grad_norm": 0.017075596377253532,
+ "learning_rate": 0.0006,
+ "loss": 4.205138683319092,
+ "step": 5163
+ },
+ {
+ "epoch": 71.72663755458515,
+ "grad_norm": 0.01723487675189972,
+ "learning_rate": 0.0006,
+ "loss": 4.08839225769043,
+ "step": 5164
+ },
+ {
+ "epoch": 71.74061135371178,
+ "grad_norm": 0.017355505377054214,
+ "learning_rate": 0.0006,
+ "loss": 4.165130615234375,
+ "step": 5165
+ },
+ {
+ "epoch": 71.75458515283843,
+ "grad_norm": 0.01868261583149433,
+ "learning_rate": 0.0006,
+ "loss": 4.16977071762085,
+ "step": 5166
+ },
+ {
+ "epoch": 71.76855895196506,
+ "grad_norm": 0.01897321827709675,
+ "learning_rate": 0.0006,
+ "loss": 4.173497200012207,
+ "step": 5167
+ },
+ {
+ "epoch": 71.78253275109171,
+ "grad_norm": 0.02045825682580471,
+ "learning_rate": 0.0006,
+ "loss": 4.176192283630371,
+ "step": 5168
+ },
+ {
+ "epoch": 71.79650655021834,
+ "grad_norm": 0.022895624861121178,
+ "learning_rate": 0.0006,
+ "loss": 4.163009166717529,
+ "step": 5169
+ },
+ {
+ "epoch": 71.81048034934499,
+ "grad_norm": 0.02062244527041912,
+ "learning_rate": 0.0006,
+ "loss": 4.156868934631348,
+ "step": 5170
+ },
+ {
+ "epoch": 71.82445414847162,
+ "grad_norm": 0.0176058579236269,
+ "learning_rate": 0.0006,
+ "loss": 4.154374122619629,
+ "step": 5171
+ },
+ {
+ "epoch": 71.83842794759825,
+ "grad_norm": 0.018049508333206177,
+ "learning_rate": 0.0006,
+ "loss": 4.296828269958496,
+ "step": 5172
+ },
+ {
+ "epoch": 71.8524017467249,
+ "grad_norm": 0.019466130062937737,
+ "learning_rate": 0.0006,
+ "loss": 4.19450569152832,
+ "step": 5173
+ },
+ {
+ "epoch": 71.86637554585153,
+ "grad_norm": 0.01838165521621704,
+ "learning_rate": 0.0006,
+ "loss": 4.11812686920166,
+ "step": 5174
+ },
+ {
+ "epoch": 71.88034934497817,
+ "grad_norm": 0.01608051173388958,
+ "learning_rate": 0.0006,
+ "loss": 4.240464210510254,
+ "step": 5175
+ },
+ {
+ "epoch": 71.8943231441048,
+ "grad_norm": 0.015807967633008957,
+ "learning_rate": 0.0006,
+ "loss": 4.275581359863281,
+ "step": 5176
+ },
+ {
+ "epoch": 71.90829694323143,
+ "grad_norm": 0.017366575077176094,
+ "learning_rate": 0.0006,
+ "loss": 4.064750671386719,
+ "step": 5177
+ },
+ {
+ "epoch": 71.92227074235808,
+ "grad_norm": 0.016241202130913734,
+ "learning_rate": 0.0006,
+ "loss": 4.203697681427002,
+ "step": 5178
+ },
+ {
+ "epoch": 71.93624454148471,
+ "grad_norm": 0.019105732440948486,
+ "learning_rate": 0.0006,
+ "loss": 4.140021800994873,
+ "step": 5179
+ },
+ {
+ "epoch": 71.95021834061136,
+ "grad_norm": 0.018492674455046654,
+ "learning_rate": 0.0006,
+ "loss": 4.2534074783325195,
+ "step": 5180
+ },
+ {
+ "epoch": 71.96419213973799,
+ "grad_norm": 0.019301170483231544,
+ "learning_rate": 0.0006,
+ "loss": 4.381392478942871,
+ "step": 5181
+ },
+ {
+ "epoch": 71.97816593886463,
+ "grad_norm": 0.019571911543607712,
+ "learning_rate": 0.0006,
+ "loss": 4.131056785583496,
+ "step": 5182
+ },
+ {
+ "epoch": 71.99213973799127,
+ "grad_norm": 0.01859748177230358,
+ "learning_rate": 0.0006,
+ "loss": 4.278930187225342,
+ "step": 5183
+ },
+ {
+ "epoch": 72.0,
+ "grad_norm": 0.02017994411289692,
+ "learning_rate": 0.0006,
+ "loss": 4.116464614868164,
+ "step": 5184
+ },
+ {
+ "epoch": 72.0,
+ "eval_loss": 4.60784912109375,
+ "eval_runtime": 56.3242,
+ "eval_samples_per_second": 43.356,
+ "eval_steps_per_second": 1.367,
+ "step": 5184
+ },
+ {
+ "epoch": 72.01397379912663,
+ "grad_norm": 0.030090296640992165,
+ "learning_rate": 0.0006,
+ "loss": 4.218588352203369,
+ "step": 5185
+ },
+ {
+ "epoch": 72.02794759825328,
+ "grad_norm": 1.330443024635315,
+ "learning_rate": 0.0006,
+ "loss": 4.417870044708252,
+ "step": 5186
+ },
+ {
+ "epoch": 72.04192139737991,
+ "grad_norm": 0.6991518139839172,
+ "learning_rate": 0.0006,
+ "loss": 8.414234161376953,
+ "step": 5187
+ },
+ {
+ "epoch": 72.05589519650655,
+ "grad_norm": 0.2631438374519348,
+ "learning_rate": 0.0006,
+ "loss": 7.41729736328125,
+ "step": 5188
+ },
+ {
+ "epoch": 72.06986899563319,
+ "grad_norm": 0.12391754984855652,
+ "learning_rate": 0.0006,
+ "loss": 7.209164619445801,
+ "step": 5189
+ },
+ {
+ "epoch": 72.08384279475983,
+ "grad_norm": 0.13519170880317688,
+ "learning_rate": 0.0006,
+ "loss": 6.943539619445801,
+ "step": 5190
+ },
+ {
+ "epoch": 72.09781659388646,
+ "grad_norm": 1.258439302444458,
+ "learning_rate": 0.0006,
+ "loss": 10.267714500427246,
+ "step": 5191
+ },
+ {
+ "epoch": 72.1117903930131,
+ "grad_norm": 1.0296908617019653,
+ "learning_rate": 0.0006,
+ "loss": 9.51768970489502,
+ "step": 5192
+ },
+ {
+ "epoch": 72.12576419213974,
+ "grad_norm": 0.12273016571998596,
+ "learning_rate": 0.0006,
+ "loss": 7.634477615356445,
+ "step": 5193
+ },
+ {
+ "epoch": 72.13973799126637,
+ "grad_norm": 0.06170198693871498,
+ "learning_rate": 0.0006,
+ "loss": 7.592115879058838,
+ "step": 5194
+ },
+ {
+ "epoch": 72.15371179039302,
+ "grad_norm": 0.06295108795166016,
+ "learning_rate": 0.0006,
+ "loss": 7.576314449310303,
+ "step": 5195
+ },
+ {
+ "epoch": 72.16768558951965,
+ "grad_norm": 0.0666993111371994,
+ "learning_rate": 0.0006,
+ "loss": 7.53862190246582,
+ "step": 5196
+ },
+ {
+ "epoch": 72.18165938864628,
+ "grad_norm": 0.052215222269296646,
+ "learning_rate": 0.0006,
+ "loss": 7.502930641174316,
+ "step": 5197
+ },
+ {
+ "epoch": 72.19563318777293,
+ "grad_norm": 0.04978961497545242,
+ "learning_rate": 0.0006,
+ "loss": 7.471630096435547,
+ "step": 5198
+ },
+ {
+ "epoch": 72.20960698689956,
+ "grad_norm": 0.07275046408176422,
+ "learning_rate": 0.0006,
+ "loss": 7.496924877166748,
+ "step": 5199
+ },
+ {
+ "epoch": 72.2235807860262,
+ "grad_norm": 0.04491819068789482,
+ "learning_rate": 0.0006,
+ "loss": 7.424494743347168,
+ "step": 5200
+ },
+ {
+ "epoch": 72.23755458515284,
+ "grad_norm": 0.043718695640563965,
+ "learning_rate": 0.0006,
+ "loss": 7.441090106964111,
+ "step": 5201
+ },
+ {
+ "epoch": 72.25152838427948,
+ "grad_norm": 0.04483497887849808,
+ "learning_rate": 0.0006,
+ "loss": 7.3753862380981445,
+ "step": 5202
+ },
+ {
+ "epoch": 72.26550218340611,
+ "grad_norm": 0.03463944420218468,
+ "learning_rate": 0.0006,
+ "loss": 7.375782012939453,
+ "step": 5203
+ },
+ {
+ "epoch": 72.27947598253274,
+ "grad_norm": 0.0397045724093914,
+ "learning_rate": 0.0006,
+ "loss": 7.358898639678955,
+ "step": 5204
+ },
+ {
+ "epoch": 72.29344978165939,
+ "grad_norm": 0.029339438304305077,
+ "learning_rate": 0.0006,
+ "loss": 7.3774871826171875,
+ "step": 5205
+ },
+ {
+ "epoch": 72.30742358078602,
+ "grad_norm": 0.025469699874520302,
+ "learning_rate": 0.0006,
+ "loss": 7.324203014373779,
+ "step": 5206
+ },
+ {
+ "epoch": 72.32139737991267,
+ "grad_norm": 0.030497785657644272,
+ "learning_rate": 0.0006,
+ "loss": 7.329351902008057,
+ "step": 5207
+ },
+ {
+ "epoch": 72.3353711790393,
+ "grad_norm": 0.03382078558206558,
+ "learning_rate": 0.0006,
+ "loss": 7.289940357208252,
+ "step": 5208
+ },
+ {
+ "epoch": 72.34934497816595,
+ "grad_norm": 0.019869405776262283,
+ "learning_rate": 0.0006,
+ "loss": 7.272582054138184,
+ "step": 5209
+ },
+ {
+ "epoch": 72.36331877729258,
+ "grad_norm": 0.02385753206908703,
+ "learning_rate": 0.0006,
+ "loss": 7.2654266357421875,
+ "step": 5210
+ },
+ {
+ "epoch": 72.37729257641921,
+ "grad_norm": 0.02794312685728073,
+ "learning_rate": 0.0006,
+ "loss": 7.250753402709961,
+ "step": 5211
+ },
+ {
+ "epoch": 72.39126637554585,
+ "grad_norm": 0.02445952221751213,
+ "learning_rate": 0.0006,
+ "loss": 7.293970108032227,
+ "step": 5212
+ },
+ {
+ "epoch": 72.40524017467249,
+ "grad_norm": 0.019660986959934235,
+ "learning_rate": 0.0006,
+ "loss": 7.1948771476745605,
+ "step": 5213
+ },
+ {
+ "epoch": 72.41921397379913,
+ "grad_norm": 0.024110861122608185,
+ "learning_rate": 0.0006,
+ "loss": 7.224499225616455,
+ "step": 5214
+ },
+ {
+ "epoch": 72.43318777292576,
+ "grad_norm": 0.021630609408020973,
+ "learning_rate": 0.0006,
+ "loss": 7.1736884117126465,
+ "step": 5215
+ },
+ {
+ "epoch": 72.44716157205241,
+ "grad_norm": 0.02185880020260811,
+ "learning_rate": 0.0006,
+ "loss": 7.167229652404785,
+ "step": 5216
+ },
+ {
+ "epoch": 72.46113537117904,
+ "grad_norm": 0.018913568928837776,
+ "learning_rate": 0.0006,
+ "loss": 7.196551322937012,
+ "step": 5217
+ },
+ {
+ "epoch": 72.47510917030567,
+ "grad_norm": 0.01802486553788185,
+ "learning_rate": 0.0006,
+ "loss": 7.144286632537842,
+ "step": 5218
+ },
+ {
+ "epoch": 72.48908296943232,
+ "grad_norm": 0.020386485382914543,
+ "learning_rate": 0.0006,
+ "loss": 7.196260929107666,
+ "step": 5219
+ },
+ {
+ "epoch": 72.50305676855895,
+ "grad_norm": 0.020077696070075035,
+ "learning_rate": 0.0006,
+ "loss": 7.138718605041504,
+ "step": 5220
+ },
+ {
+ "epoch": 72.5170305676856,
+ "grad_norm": 0.017224183306097984,
+ "learning_rate": 0.0006,
+ "loss": 7.119478702545166,
+ "step": 5221
+ },
+ {
+ "epoch": 72.53100436681223,
+ "grad_norm": 0.020847434177994728,
+ "learning_rate": 0.0006,
+ "loss": 7.068922996520996,
+ "step": 5222
+ },
+ {
+ "epoch": 72.54497816593886,
+ "grad_norm": 0.021764913573861122,
+ "learning_rate": 0.0006,
+ "loss": 7.071352958679199,
+ "step": 5223
+ },
+ {
+ "epoch": 72.5589519650655,
+ "grad_norm": 0.02697215974330902,
+ "learning_rate": 0.0006,
+ "loss": 7.033344268798828,
+ "step": 5224
+ },
+ {
+ "epoch": 72.57292576419214,
+ "grad_norm": 0.042392998933792114,
+ "learning_rate": 0.0006,
+ "loss": 7.0589919090271,
+ "step": 5225
+ },
+ {
+ "epoch": 72.58689956331878,
+ "grad_norm": 0.06616316735744476,
+ "learning_rate": 0.0006,
+ "loss": 7.060244083404541,
+ "step": 5226
+ },
+ {
+ "epoch": 72.60087336244541,
+ "grad_norm": 0.0828910544514656,
+ "learning_rate": 0.0006,
+ "loss": 7.057173728942871,
+ "step": 5227
+ },
+ {
+ "epoch": 72.61484716157206,
+ "grad_norm": 0.046767015010118484,
+ "learning_rate": 0.0006,
+ "loss": 7.010563850402832,
+ "step": 5228
+ },
+ {
+ "epoch": 72.62882096069869,
+ "grad_norm": 0.03831760957837105,
+ "learning_rate": 0.0006,
+ "loss": 6.9028825759887695,
+ "step": 5229
+ },
+ {
+ "epoch": 72.64279475982532,
+ "grad_norm": 0.05109567195177078,
+ "learning_rate": 0.0006,
+ "loss": 6.923257827758789,
+ "step": 5230
+ },
+ {
+ "epoch": 72.65676855895197,
+ "grad_norm": 0.02897934429347515,
+ "learning_rate": 0.0006,
+ "loss": 6.823014259338379,
+ "step": 5231
+ },
+ {
+ "epoch": 72.6707423580786,
+ "grad_norm": 0.03769192099571228,
+ "learning_rate": 0.0006,
+ "loss": 6.87416410446167,
+ "step": 5232
+ },
+ {
+ "epoch": 72.68471615720524,
+ "grad_norm": 0.03677637130022049,
+ "learning_rate": 0.0006,
+ "loss": 6.7649030685424805,
+ "step": 5233
+ },
+ {
+ "epoch": 72.69868995633188,
+ "grad_norm": 0.0371948778629303,
+ "learning_rate": 0.0006,
+ "loss": 6.749821662902832,
+ "step": 5234
+ },
+ {
+ "epoch": 72.71266375545852,
+ "grad_norm": 0.03057795763015747,
+ "learning_rate": 0.0006,
+ "loss": 6.748945236206055,
+ "step": 5235
+ },
+ {
+ "epoch": 72.72663755458515,
+ "grad_norm": 0.028109321370720863,
+ "learning_rate": 0.0006,
+ "loss": 6.663548946380615,
+ "step": 5236
+ },
+ {
+ "epoch": 72.74061135371178,
+ "grad_norm": 0.02442420832812786,
+ "learning_rate": 0.0006,
+ "loss": 6.636251449584961,
+ "step": 5237
+ },
+ {
+ "epoch": 72.75458515283843,
+ "grad_norm": 0.031841423362493515,
+ "learning_rate": 0.0006,
+ "loss": 6.606883525848389,
+ "step": 5238
+ },
+ {
+ "epoch": 72.76855895196506,
+ "grad_norm": 0.028100458905100822,
+ "learning_rate": 0.0006,
+ "loss": 6.603745937347412,
+ "step": 5239
+ },
+ {
+ "epoch": 72.78253275109171,
+ "grad_norm": 0.03222901374101639,
+ "learning_rate": 0.0006,
+ "loss": 6.463825225830078,
+ "step": 5240
+ },
+ {
+ "epoch": 72.79650655021834,
+ "grad_norm": 0.04766521975398064,
+ "learning_rate": 0.0006,
+ "loss": 6.43734073638916,
+ "step": 5241
+ },
+ {
+ "epoch": 72.81048034934499,
+ "grad_norm": 0.09058783948421478,
+ "learning_rate": 0.0006,
+ "loss": 6.448520183563232,
+ "step": 5242
+ },
+ {
+ "epoch": 72.82445414847162,
+ "grad_norm": 0.08995158225297928,
+ "learning_rate": 0.0006,
+ "loss": 6.522093296051025,
+ "step": 5243
+ },
+ {
+ "epoch": 72.83842794759825,
+ "grad_norm": 0.05491040274500847,
+ "learning_rate": 0.0006,
+ "loss": 6.37354040145874,
+ "step": 5244
+ },
+ {
+ "epoch": 72.8524017467249,
+ "grad_norm": 0.05198514088988304,
+ "learning_rate": 0.0006,
+ "loss": 6.362032413482666,
+ "step": 5245
+ },
+ {
+ "epoch": 72.86637554585153,
+ "grad_norm": 0.05464765429496765,
+ "learning_rate": 0.0006,
+ "loss": 6.326773643493652,
+ "step": 5246
+ },
+ {
+ "epoch": 72.88034934497817,
+ "grad_norm": 0.07337361574172974,
+ "learning_rate": 0.0006,
+ "loss": 6.297532081604004,
+ "step": 5247
+ },
+ {
+ "epoch": 72.8943231441048,
+ "grad_norm": 0.10061230510473251,
+ "learning_rate": 0.0006,
+ "loss": 6.359557628631592,
+ "step": 5248
+ },
+ {
+ "epoch": 72.90829694323143,
+ "grad_norm": 0.10005959123373032,
+ "learning_rate": 0.0006,
+ "loss": 6.422270774841309,
+ "step": 5249
+ },
+ {
+ "epoch": 72.92227074235808,
+ "grad_norm": 0.08989301323890686,
+ "learning_rate": 0.0006,
+ "loss": 6.289371490478516,
+ "step": 5250
+ },
+ {
+ "epoch": 72.93624454148471,
+ "grad_norm": 0.06371799111366272,
+ "learning_rate": 0.0006,
+ "loss": 6.199618339538574,
+ "step": 5251
+ },
+ {
+ "epoch": 72.95021834061136,
+ "grad_norm": 0.05932944267988205,
+ "learning_rate": 0.0006,
+ "loss": 6.111152648925781,
+ "step": 5252
+ },
+ {
+ "epoch": 72.96419213973799,
+ "grad_norm": 0.04735502973198891,
+ "learning_rate": 0.0006,
+ "loss": 6.026026248931885,
+ "step": 5253
+ },
+ {
+ "epoch": 72.97816593886463,
+ "grad_norm": 0.04270688816905022,
+ "learning_rate": 0.0006,
+ "loss": 6.01606559753418,
+ "step": 5254
+ },
+ {
+ "epoch": 72.99213973799127,
+ "grad_norm": 0.03774143010377884,
+ "learning_rate": 0.0006,
+ "loss": 5.8935441970825195,
+ "step": 5255
+ },
+ {
+ "epoch": 73.0,
+ "grad_norm": 0.04313631355762482,
+ "learning_rate": 0.0006,
+ "loss": 5.842061996459961,
+ "step": 5256
+ },
+ {
+ "epoch": 73.0,
+ "eval_loss": 5.9985785484313965,
+ "eval_runtime": 56.3891,
+ "eval_samples_per_second": 43.306,
+ "eval_steps_per_second": 1.366,
+ "step": 5256
+ },
+ {
+ "epoch": 73.01397379912663,
+ "grad_norm": 0.045038655400276184,
+ "learning_rate": 0.0006,
+ "loss": 5.8209662437438965,
+ "step": 5257
+ },
+ {
+ "epoch": 73.02794759825328,
+ "grad_norm": 0.03706910088658333,
+ "learning_rate": 0.0006,
+ "loss": 5.806189060211182,
+ "step": 5258
+ },
+ {
+ "epoch": 73.04192139737991,
+ "grad_norm": 0.04051685333251953,
+ "learning_rate": 0.0006,
+ "loss": 5.706869125366211,
+ "step": 5259
+ },
+ {
+ "epoch": 73.05589519650655,
+ "grad_norm": 0.04849153012037277,
+ "learning_rate": 0.0006,
+ "loss": 5.696385383605957,
+ "step": 5260
+ },
+ {
+ "epoch": 73.06986899563319,
+ "grad_norm": 0.06124403700232506,
+ "learning_rate": 0.0006,
+ "loss": 5.618223667144775,
+ "step": 5261
+ },
+ {
+ "epoch": 73.08384279475983,
+ "grad_norm": 0.07375937700271606,
+ "learning_rate": 0.0006,
+ "loss": 5.631045341491699,
+ "step": 5262
+ },
+ {
+ "epoch": 73.09781659388646,
+ "grad_norm": 0.07413844019174576,
+ "learning_rate": 0.0006,
+ "loss": 5.6532182693481445,
+ "step": 5263
+ },
+ {
+ "epoch": 73.1117903930131,
+ "grad_norm": 0.0633770152926445,
+ "learning_rate": 0.0006,
+ "loss": 5.57209587097168,
+ "step": 5264
+ },
+ {
+ "epoch": 73.12576419213974,
+ "grad_norm": 0.05436766892671585,
+ "learning_rate": 0.0006,
+ "loss": 5.53944730758667,
+ "step": 5265
+ },
+ {
+ "epoch": 73.13973799126637,
+ "grad_norm": 0.04639481008052826,
+ "learning_rate": 0.0006,
+ "loss": 5.502948760986328,
+ "step": 5266
+ },
+ {
+ "epoch": 73.15371179039302,
+ "grad_norm": 0.07556558400392532,
+ "learning_rate": 0.0006,
+ "loss": 5.429069995880127,
+ "step": 5267
+ },
+ {
+ "epoch": 73.16768558951965,
+ "grad_norm": 0.1127929836511612,
+ "learning_rate": 0.0006,
+ "loss": 5.545061111450195,
+ "step": 5268
+ },
+ {
+ "epoch": 73.18165938864628,
+ "grad_norm": 0.07620745897293091,
+ "learning_rate": 0.0006,
+ "loss": 5.507102966308594,
+ "step": 5269
+ },
+ {
+ "epoch": 73.19563318777293,
+ "grad_norm": 0.05076538026332855,
+ "learning_rate": 0.0006,
+ "loss": 5.329682350158691,
+ "step": 5270
+ },
+ {
+ "epoch": 73.20960698689956,
+ "grad_norm": 0.05407547950744629,
+ "learning_rate": 0.0006,
+ "loss": 5.3048553466796875,
+ "step": 5271
+ },
+ {
+ "epoch": 73.2235807860262,
+ "grad_norm": 0.059095583856105804,
+ "learning_rate": 0.0006,
+ "loss": 5.292568683624268,
+ "step": 5272
+ },
+ {
+ "epoch": 73.23755458515284,
+ "grad_norm": 0.06193305924534798,
+ "learning_rate": 0.0006,
+ "loss": 5.1870927810668945,
+ "step": 5273
+ },
+ {
+ "epoch": 73.25152838427948,
+ "grad_norm": 0.062261562794446945,
+ "learning_rate": 0.0006,
+ "loss": 5.256819248199463,
+ "step": 5274
+ },
+ {
+ "epoch": 73.26550218340611,
+ "grad_norm": 0.0732516348361969,
+ "learning_rate": 0.0006,
+ "loss": 5.256464004516602,
+ "step": 5275
+ },
+ {
+ "epoch": 73.27947598253274,
+ "grad_norm": 0.12177979946136475,
+ "learning_rate": 0.0006,
+ "loss": 5.284239768981934,
+ "step": 5276
+ },
+ {
+ "epoch": 73.29344978165939,
+ "grad_norm": 0.10459653288125992,
+ "learning_rate": 0.0006,
+ "loss": 5.269713401794434,
+ "step": 5277
+ },
+ {
+ "epoch": 73.30742358078602,
+ "grad_norm": 0.07854694873094559,
+ "learning_rate": 0.0006,
+ "loss": 5.3245391845703125,
+ "step": 5278
+ },
+ {
+ "epoch": 73.32139737991267,
+ "grad_norm": 0.06030094623565674,
+ "learning_rate": 0.0006,
+ "loss": 5.033702850341797,
+ "step": 5279
+ },
+ {
+ "epoch": 73.3353711790393,
+ "grad_norm": 0.05168461427092552,
+ "learning_rate": 0.0006,
+ "loss": 5.140530109405518,
+ "step": 5280
+ },
+ {
+ "epoch": 73.34934497816595,
+ "grad_norm": 0.041064050048589706,
+ "learning_rate": 0.0006,
+ "loss": 5.1085896492004395,
+ "step": 5281
+ },
+ {
+ "epoch": 73.36331877729258,
+ "grad_norm": 0.03788219019770622,
+ "learning_rate": 0.0006,
+ "loss": 5.045009613037109,
+ "step": 5282
+ },
+ {
+ "epoch": 73.37729257641921,
+ "grad_norm": 0.03851071000099182,
+ "learning_rate": 0.0006,
+ "loss": 5.055008888244629,
+ "step": 5283
+ },
+ {
+ "epoch": 73.39126637554585,
+ "grad_norm": 0.03448539599776268,
+ "learning_rate": 0.0006,
+ "loss": 5.10657262802124,
+ "step": 5284
+ },
+ {
+ "epoch": 73.40524017467249,
+ "grad_norm": 0.026618091389536858,
+ "learning_rate": 0.0006,
+ "loss": 5.0342912673950195,
+ "step": 5285
+ },
+ {
+ "epoch": 73.41921397379913,
+ "grad_norm": 0.03060489147901535,
+ "learning_rate": 0.0006,
+ "loss": 5.022700786590576,
+ "step": 5286
+ },
+ {
+ "epoch": 73.43318777292576,
+ "grad_norm": 0.03362000361084938,
+ "learning_rate": 0.0006,
+ "loss": 4.891584396362305,
+ "step": 5287
+ },
+ {
+ "epoch": 73.44716157205241,
+ "grad_norm": 0.03887505829334259,
+ "learning_rate": 0.0006,
+ "loss": 4.772258758544922,
+ "step": 5288
+ },
+ {
+ "epoch": 73.46113537117904,
+ "grad_norm": 0.07157225161790848,
+ "learning_rate": 0.0006,
+ "loss": 4.997142791748047,
+ "step": 5289
+ },
+ {
+ "epoch": 73.47510917030567,
+ "grad_norm": 0.10180624574422836,
+ "learning_rate": 0.0006,
+ "loss": 4.908722877502441,
+ "step": 5290
+ },
+ {
+ "epoch": 73.48908296943232,
+ "grad_norm": 0.06529122591018677,
+ "learning_rate": 0.0006,
+ "loss": 4.98001766204834,
+ "step": 5291
+ },
+ {
+ "epoch": 73.50305676855895,
+ "grad_norm": 0.055171865969896317,
+ "learning_rate": 0.0006,
+ "loss": 4.877438068389893,
+ "step": 5292
+ },
+ {
+ "epoch": 73.5170305676856,
+ "grad_norm": 0.0318770632147789,
+ "learning_rate": 0.0006,
+ "loss": 4.836934566497803,
+ "step": 5293
+ },
+ {
+ "epoch": 73.53100436681223,
+ "grad_norm": 0.042072564363479614,
+ "learning_rate": 0.0006,
+ "loss": 4.919489860534668,
+ "step": 5294
+ },
+ {
+ "epoch": 73.54497816593886,
+ "grad_norm": 0.029197754338383675,
+ "learning_rate": 0.0006,
+ "loss": 4.805579662322998,
+ "step": 5295
+ },
+ {
+ "epoch": 73.5589519650655,
+ "grad_norm": 0.03249497339129448,
+ "learning_rate": 0.0006,
+ "loss": 4.704806327819824,
+ "step": 5296
+ },
+ {
+ "epoch": 73.57292576419214,
+ "grad_norm": 0.025050945580005646,
+ "learning_rate": 0.0006,
+ "loss": 4.936512470245361,
+ "step": 5297
+ },
+ {
+ "epoch": 73.58689956331878,
+ "grad_norm": 0.029073381796479225,
+ "learning_rate": 0.0006,
+ "loss": 4.757592678070068,
+ "step": 5298
+ },
+ {
+ "epoch": 73.60087336244541,
+ "grad_norm": 0.03465264290571213,
+ "learning_rate": 0.0006,
+ "loss": 4.729528427124023,
+ "step": 5299
+ },
+ {
+ "epoch": 73.61484716157206,
+ "grad_norm": 0.054795749485492706,
+ "learning_rate": 0.0006,
+ "loss": 4.77252197265625,
+ "step": 5300
+ },
+ {
+ "epoch": 73.62882096069869,
+ "grad_norm": 0.18129196763038635,
+ "learning_rate": 0.0006,
+ "loss": 4.901700019836426,
+ "step": 5301
+ },
+ {
+ "epoch": 73.64279475982532,
+ "grad_norm": 0.23274236917495728,
+ "learning_rate": 0.0006,
+ "loss": 5.207432270050049,
+ "step": 5302
+ },
+ {
+ "epoch": 73.65676855895197,
+ "grad_norm": 0.18162772059440613,
+ "learning_rate": 0.0006,
+ "loss": 5.326632976531982,
+ "step": 5303
+ },
+ {
+ "epoch": 73.6707423580786,
+ "grad_norm": 0.11858413368463516,
+ "learning_rate": 0.0006,
+ "loss": 5.3730058670043945,
+ "step": 5304
+ },
+ {
+ "epoch": 73.68471615720524,
+ "grad_norm": 0.06031268090009689,
+ "learning_rate": 0.0006,
+ "loss": 5.330936431884766,
+ "step": 5305
+ },
+ {
+ "epoch": 73.69868995633188,
+ "grad_norm": 0.059304628521203995,
+ "learning_rate": 0.0006,
+ "loss": 5.280514717102051,
+ "step": 5306
+ },
+ {
+ "epoch": 73.71266375545852,
+ "grad_norm": 0.04702131822705269,
+ "learning_rate": 0.0006,
+ "loss": 5.288274765014648,
+ "step": 5307
+ },
+ {
+ "epoch": 73.72663755458515,
+ "grad_norm": 0.037792228162288666,
+ "learning_rate": 0.0006,
+ "loss": 5.173289775848389,
+ "step": 5308
+ },
+ {
+ "epoch": 73.74061135371178,
+ "grad_norm": 0.037763576954603195,
+ "learning_rate": 0.0006,
+ "loss": 5.084990978240967,
+ "step": 5309
+ },
+ {
+ "epoch": 73.75458515283843,
+ "grad_norm": 0.03337346389889717,
+ "learning_rate": 0.0006,
+ "loss": 5.121417045593262,
+ "step": 5310
+ },
+ {
+ "epoch": 73.76855895196506,
+ "grad_norm": 0.03184802085161209,
+ "learning_rate": 0.0006,
+ "loss": 5.208918571472168,
+ "step": 5311
+ },
+ {
+ "epoch": 73.78253275109171,
+ "grad_norm": 0.02832048572599888,
+ "learning_rate": 0.0006,
+ "loss": 4.923543453216553,
+ "step": 5312
+ },
+ {
+ "epoch": 73.79650655021834,
+ "grad_norm": 0.028297102078795433,
+ "learning_rate": 0.0006,
+ "loss": 4.944367408752441,
+ "step": 5313
+ },
+ {
+ "epoch": 73.81048034934499,
+ "grad_norm": 0.027701659128069878,
+ "learning_rate": 0.0006,
+ "loss": 4.858168601989746,
+ "step": 5314
+ },
+ {
+ "epoch": 73.82445414847162,
+ "grad_norm": 0.02214639261364937,
+ "learning_rate": 0.0006,
+ "loss": 4.887052536010742,
+ "step": 5315
+ },
+ {
+ "epoch": 73.83842794759825,
+ "grad_norm": 0.01995093561708927,
+ "learning_rate": 0.0006,
+ "loss": 4.885797500610352,
+ "step": 5316
+ },
+ {
+ "epoch": 73.8524017467249,
+ "grad_norm": 0.020473985001444817,
+ "learning_rate": 0.0006,
+ "loss": 4.917325019836426,
+ "step": 5317
+ },
+ {
+ "epoch": 73.86637554585153,
+ "grad_norm": 0.0192868635058403,
+ "learning_rate": 0.0006,
+ "loss": 4.801746845245361,
+ "step": 5318
+ },
+ {
+ "epoch": 73.88034934497817,
+ "grad_norm": 0.017108267173171043,
+ "learning_rate": 0.0006,
+ "loss": 4.893620491027832,
+ "step": 5319
+ },
+ {
+ "epoch": 73.8943231441048,
+ "grad_norm": 0.01718071475625038,
+ "learning_rate": 0.0006,
+ "loss": 4.940695285797119,
+ "step": 5320
+ },
+ {
+ "epoch": 73.90829694323143,
+ "grad_norm": 0.017821557819843292,
+ "learning_rate": 0.0006,
+ "loss": 4.767613410949707,
+ "step": 5321
+ },
+ {
+ "epoch": 73.92227074235808,
+ "grad_norm": 0.015262950211763382,
+ "learning_rate": 0.0006,
+ "loss": 4.910315036773682,
+ "step": 5322
+ },
+ {
+ "epoch": 73.93624454148471,
+ "grad_norm": 0.01769006811082363,
+ "learning_rate": 0.0006,
+ "loss": 4.7743449211120605,
+ "step": 5323
+ },
+ {
+ "epoch": 73.95021834061136,
+ "grad_norm": 0.014460788108408451,
+ "learning_rate": 0.0006,
+ "loss": 4.743494987487793,
+ "step": 5324
+ },
+ {
+ "epoch": 73.96419213973799,
+ "grad_norm": 0.01403993833810091,
+ "learning_rate": 0.0006,
+ "loss": 4.666640281677246,
+ "step": 5325
+ },
+ {
+ "epoch": 73.97816593886463,
+ "grad_norm": 0.013997809961438179,
+ "learning_rate": 0.0006,
+ "loss": 4.698747158050537,
+ "step": 5326
+ },
+ {
+ "epoch": 73.99213973799127,
+ "grad_norm": 0.014244182966649532,
+ "learning_rate": 0.0006,
+ "loss": 4.745250701904297,
+ "step": 5327
+ },
+ {
+ "epoch": 74.0,
+ "grad_norm": 0.015144186094403267,
+ "learning_rate": 0.0006,
+ "loss": 4.667946815490723,
+ "step": 5328
+ },
+ {
+ "epoch": 74.0,
+ "eval_loss": 5.005440711975098,
+ "eval_runtime": 56.7836,
+ "eval_samples_per_second": 43.005,
+ "eval_steps_per_second": 1.356,
+ "step": 5328
+ },
+ {
+ "epoch": 74.01397379912663,
+ "grad_norm": 0.020384902134537697,
+ "learning_rate": 0.0006,
+ "loss": 4.626955509185791,
+ "step": 5329
+ },
+ {
+ "epoch": 74.02794759825328,
+ "grad_norm": 0.03624720871448517,
+ "learning_rate": 0.0006,
+ "loss": 4.713998317718506,
+ "step": 5330
+ },
+ {
+ "epoch": 74.04192139737991,
+ "grad_norm": 0.05701754242181778,
+ "learning_rate": 0.0006,
+ "loss": 4.717874526977539,
+ "step": 5331
+ },
+ {
+ "epoch": 74.05589519650655,
+ "grad_norm": 0.03942621126770973,
+ "learning_rate": 0.0006,
+ "loss": 4.7026190757751465,
+ "step": 5332
+ },
+ {
+ "epoch": 74.06986899563319,
+ "grad_norm": 0.017803272232413292,
+ "learning_rate": 0.0006,
+ "loss": 4.702548027038574,
+ "step": 5333
+ },
+ {
+ "epoch": 74.08384279475983,
+ "grad_norm": 0.017024677246809006,
+ "learning_rate": 0.0006,
+ "loss": 4.69752311706543,
+ "step": 5334
+ },
+ {
+ "epoch": 74.09781659388646,
+ "grad_norm": 0.016467854380607605,
+ "learning_rate": 0.0006,
+ "loss": 4.624876976013184,
+ "step": 5335
+ },
+ {
+ "epoch": 74.1117903930131,
+ "grad_norm": 0.016569215804338455,
+ "learning_rate": 0.0006,
+ "loss": 4.645040512084961,
+ "step": 5336
+ },
+ {
+ "epoch": 74.12576419213974,
+ "grad_norm": 0.012872067280113697,
+ "learning_rate": 0.0006,
+ "loss": 4.64582633972168,
+ "step": 5337
+ },
+ {
+ "epoch": 74.13973799126637,
+ "grad_norm": 0.015689387917518616,
+ "learning_rate": 0.0006,
+ "loss": 4.596040725708008,
+ "step": 5338
+ },
+ {
+ "epoch": 74.15371179039302,
+ "grad_norm": 0.013082082383334637,
+ "learning_rate": 0.0006,
+ "loss": 4.648033142089844,
+ "step": 5339
+ },
+ {
+ "epoch": 74.16768558951965,
+ "grad_norm": 0.013728172518312931,
+ "learning_rate": 0.0006,
+ "loss": 4.657873153686523,
+ "step": 5340
+ },
+ {
+ "epoch": 74.18165938864628,
+ "grad_norm": 0.013120594434440136,
+ "learning_rate": 0.0006,
+ "loss": 4.696074485778809,
+ "step": 5341
+ },
+ {
+ "epoch": 74.19563318777293,
+ "grad_norm": 0.01431245356798172,
+ "learning_rate": 0.0006,
+ "loss": 4.5679121017456055,
+ "step": 5342
+ },
+ {
+ "epoch": 74.20960698689956,
+ "grad_norm": 0.013994511216878891,
+ "learning_rate": 0.0006,
+ "loss": 4.5236406326293945,
+ "step": 5343
+ },
+ {
+ "epoch": 74.2235807860262,
+ "grad_norm": 0.011971707455813885,
+ "learning_rate": 0.0006,
+ "loss": 4.626547813415527,
+ "step": 5344
+ },
+ {
+ "epoch": 74.23755458515284,
+ "grad_norm": 0.012805287726223469,
+ "learning_rate": 0.0006,
+ "loss": 4.556386947631836,
+ "step": 5345
+ },
+ {
+ "epoch": 74.25152838427948,
+ "grad_norm": 0.011694584973156452,
+ "learning_rate": 0.0006,
+ "loss": 4.656987190246582,
+ "step": 5346
+ },
+ {
+ "epoch": 74.26550218340611,
+ "grad_norm": 0.013396821916103363,
+ "learning_rate": 0.0006,
+ "loss": 4.526467323303223,
+ "step": 5347
+ },
+ {
+ "epoch": 74.27947598253274,
+ "grad_norm": 0.014398031868040562,
+ "learning_rate": 0.0006,
+ "loss": 4.595100402832031,
+ "step": 5348
+ },
+ {
+ "epoch": 74.29344978165939,
+ "grad_norm": 0.02165800705552101,
+ "learning_rate": 0.0006,
+ "loss": 4.460549354553223,
+ "step": 5349
+ },
+ {
+ "epoch": 74.30742358078602,
+ "grad_norm": 0.03276512026786804,
+ "learning_rate": 0.0006,
+ "loss": 4.498189449310303,
+ "step": 5350
+ },
+ {
+ "epoch": 74.32139737991267,
+ "grad_norm": 0.04616224020719528,
+ "learning_rate": 0.0006,
+ "loss": 4.5268659591674805,
+ "step": 5351
+ },
+ {
+ "epoch": 74.3353711790393,
+ "grad_norm": 0.03863786533474922,
+ "learning_rate": 0.0006,
+ "loss": 4.554747104644775,
+ "step": 5352
+ },
+ {
+ "epoch": 74.34934497816595,
+ "grad_norm": 0.01749865710735321,
+ "learning_rate": 0.0006,
+ "loss": 4.6656036376953125,
+ "step": 5353
+ },
+ {
+ "epoch": 74.36331877729258,
+ "grad_norm": 0.021814396604895592,
+ "learning_rate": 0.0006,
+ "loss": 4.607456207275391,
+ "step": 5354
+ },
+ {
+ "epoch": 74.37729257641921,
+ "grad_norm": 0.017018496990203857,
+ "learning_rate": 0.0006,
+ "loss": 4.54205322265625,
+ "step": 5355
+ },
+ {
+ "epoch": 74.39126637554585,
+ "grad_norm": 0.018297340720891953,
+ "learning_rate": 0.0006,
+ "loss": 4.608870506286621,
+ "step": 5356
+ },
+ {
+ "epoch": 74.40524017467249,
+ "grad_norm": 0.01473482046276331,
+ "learning_rate": 0.0006,
+ "loss": 4.384891510009766,
+ "step": 5357
+ },
+ {
+ "epoch": 74.41921397379913,
+ "grad_norm": 0.01579415239393711,
+ "learning_rate": 0.0006,
+ "loss": 4.588859558105469,
+ "step": 5358
+ },
+ {
+ "epoch": 74.43318777292576,
+ "grad_norm": 0.013906026259064674,
+ "learning_rate": 0.0006,
+ "loss": 4.525175094604492,
+ "step": 5359
+ },
+ {
+ "epoch": 74.44716157205241,
+ "grad_norm": 0.014163361862301826,
+ "learning_rate": 0.0006,
+ "loss": 4.495802879333496,
+ "step": 5360
+ },
+ {
+ "epoch": 74.46113537117904,
+ "grad_norm": 0.014009335078299046,
+ "learning_rate": 0.0006,
+ "loss": 4.5960869789123535,
+ "step": 5361
+ },
+ {
+ "epoch": 74.47510917030567,
+ "grad_norm": 0.012765258550643921,
+ "learning_rate": 0.0006,
+ "loss": 4.515678405761719,
+ "step": 5362
+ },
+ {
+ "epoch": 74.48908296943232,
+ "grad_norm": 0.014071918092668056,
+ "learning_rate": 0.0006,
+ "loss": 4.394025802612305,
+ "step": 5363
+ },
+ {
+ "epoch": 74.50305676855895,
+ "grad_norm": 0.01244515273720026,
+ "learning_rate": 0.0006,
+ "loss": 4.576985836029053,
+ "step": 5364
+ },
+ {
+ "epoch": 74.5170305676856,
+ "grad_norm": 0.012975298799574375,
+ "learning_rate": 0.0006,
+ "loss": 4.540657997131348,
+ "step": 5365
+ },
+ {
+ "epoch": 74.53100436681223,
+ "grad_norm": 0.012380947358906269,
+ "learning_rate": 0.0006,
+ "loss": 4.55731725692749,
+ "step": 5366
+ },
+ {
+ "epoch": 74.54497816593886,
+ "grad_norm": 0.012577617540955544,
+ "learning_rate": 0.0006,
+ "loss": 4.500924110412598,
+ "step": 5367
+ },
+ {
+ "epoch": 74.5589519650655,
+ "grad_norm": 0.012089254334568977,
+ "learning_rate": 0.0006,
+ "loss": 4.561795234680176,
+ "step": 5368
+ },
+ {
+ "epoch": 74.57292576419214,
+ "grad_norm": 0.012024011462926865,
+ "learning_rate": 0.0006,
+ "loss": 4.430635452270508,
+ "step": 5369
+ },
+ {
+ "epoch": 74.58689956331878,
+ "grad_norm": 0.012361960485577583,
+ "learning_rate": 0.0006,
+ "loss": 4.4664106369018555,
+ "step": 5370
+ },
+ {
+ "epoch": 74.60087336244541,
+ "grad_norm": 0.012149893678724766,
+ "learning_rate": 0.0006,
+ "loss": 4.522271156311035,
+ "step": 5371
+ },
+ {
+ "epoch": 74.61484716157206,
+ "grad_norm": 0.01166840922087431,
+ "learning_rate": 0.0006,
+ "loss": 4.3870463371276855,
+ "step": 5372
+ },
+ {
+ "epoch": 74.62882096069869,
+ "grad_norm": 0.01160407718271017,
+ "learning_rate": 0.0006,
+ "loss": 4.504255771636963,
+ "step": 5373
+ },
+ {
+ "epoch": 74.64279475982532,
+ "grad_norm": 0.011259177699685097,
+ "learning_rate": 0.0006,
+ "loss": 4.507267951965332,
+ "step": 5374
+ },
+ {
+ "epoch": 74.65676855895197,
+ "grad_norm": 0.012285715900361538,
+ "learning_rate": 0.0006,
+ "loss": 4.472966194152832,
+ "step": 5375
+ },
+ {
+ "epoch": 74.6707423580786,
+ "grad_norm": 0.011214440688490868,
+ "learning_rate": 0.0006,
+ "loss": 4.5840253829956055,
+ "step": 5376
+ },
+ {
+ "epoch": 74.68471615720524,
+ "grad_norm": 0.012394879944622517,
+ "learning_rate": 0.0006,
+ "loss": 4.417830944061279,
+ "step": 5377
+ },
+ {
+ "epoch": 74.69868995633188,
+ "grad_norm": 0.012360049411654472,
+ "learning_rate": 0.0006,
+ "loss": 4.535454750061035,
+ "step": 5378
+ },
+ {
+ "epoch": 74.71266375545852,
+ "grad_norm": 0.011961652897298336,
+ "learning_rate": 0.0006,
+ "loss": 4.403409957885742,
+ "step": 5379
+ },
+ {
+ "epoch": 74.72663755458515,
+ "grad_norm": 0.012078274972736835,
+ "learning_rate": 0.0006,
+ "loss": 4.448366641998291,
+ "step": 5380
+ },
+ {
+ "epoch": 74.74061135371178,
+ "grad_norm": 0.013239333406090736,
+ "learning_rate": 0.0006,
+ "loss": 4.543450355529785,
+ "step": 5381
+ },
+ {
+ "epoch": 74.75458515283843,
+ "grad_norm": 0.01476441603153944,
+ "learning_rate": 0.0006,
+ "loss": 4.488677978515625,
+ "step": 5382
+ },
+ {
+ "epoch": 74.76855895196506,
+ "grad_norm": 0.013968009501695633,
+ "learning_rate": 0.0006,
+ "loss": 4.603639602661133,
+ "step": 5383
+ },
+ {
+ "epoch": 74.78253275109171,
+ "grad_norm": 0.012322458438575268,
+ "learning_rate": 0.0006,
+ "loss": 4.637874603271484,
+ "step": 5384
+ },
+ {
+ "epoch": 74.79650655021834,
+ "grad_norm": 0.013352593407034874,
+ "learning_rate": 0.0006,
+ "loss": 4.474833965301514,
+ "step": 5385
+ },
+ {
+ "epoch": 74.81048034934499,
+ "grad_norm": 0.014671728946268559,
+ "learning_rate": 0.0006,
+ "loss": 4.516847610473633,
+ "step": 5386
+ },
+ {
+ "epoch": 74.82445414847162,
+ "grad_norm": 0.017163321375846863,
+ "learning_rate": 0.0006,
+ "loss": 4.462325096130371,
+ "step": 5387
+ },
+ {
+ "epoch": 74.83842794759825,
+ "grad_norm": 0.02012511156499386,
+ "learning_rate": 0.0006,
+ "loss": 4.355982780456543,
+ "step": 5388
+ },
+ {
+ "epoch": 74.8524017467249,
+ "grad_norm": 0.01881697215139866,
+ "learning_rate": 0.0006,
+ "loss": 4.451120376586914,
+ "step": 5389
+ },
+ {
+ "epoch": 74.86637554585153,
+ "grad_norm": 0.017279407009482384,
+ "learning_rate": 0.0006,
+ "loss": 4.461202621459961,
+ "step": 5390
+ },
+ {
+ "epoch": 74.88034934497817,
+ "grad_norm": 0.015470407903194427,
+ "learning_rate": 0.0006,
+ "loss": 4.4968671798706055,
+ "step": 5391
+ },
+ {
+ "epoch": 74.8943231441048,
+ "grad_norm": 0.01411783043295145,
+ "learning_rate": 0.0006,
+ "loss": 4.558697700500488,
+ "step": 5392
+ },
+ {
+ "epoch": 74.90829694323143,
+ "grad_norm": 0.017178812995553017,
+ "learning_rate": 0.0006,
+ "loss": 4.254153251647949,
+ "step": 5393
+ },
+ {
+ "epoch": 74.92227074235808,
+ "grad_norm": 0.019193004816770554,
+ "learning_rate": 0.0006,
+ "loss": 4.394747257232666,
+ "step": 5394
+ },
+ {
+ "epoch": 74.93624454148471,
+ "grad_norm": 0.018942881375551224,
+ "learning_rate": 0.0006,
+ "loss": 4.459539413452148,
+ "step": 5395
+ },
+ {
+ "epoch": 74.95021834061136,
+ "grad_norm": 0.01641223020851612,
+ "learning_rate": 0.0006,
+ "loss": 4.485820770263672,
+ "step": 5396
+ },
+ {
+ "epoch": 74.96419213973799,
+ "grad_norm": 0.014485753141343594,
+ "learning_rate": 0.0006,
+ "loss": 4.5068583488464355,
+ "step": 5397
+ },
+ {
+ "epoch": 74.97816593886463,
+ "grad_norm": 0.018242299556732178,
+ "learning_rate": 0.0006,
+ "loss": 4.508835792541504,
+ "step": 5398
+ },
+ {
+ "epoch": 74.99213973799127,
+ "grad_norm": 0.016491873189806938,
+ "learning_rate": 0.0006,
+ "loss": 4.4760870933532715,
+ "step": 5399
+ },
+ {
+ "epoch": 75.0,
+ "grad_norm": 0.014543353579938412,
+ "learning_rate": 0.0006,
+ "loss": 4.540217399597168,
+ "step": 5400
+ },
+ {
+ "epoch": 75.0,
+ "eval_loss": 4.7617573738098145,
+ "eval_runtime": 57.321,
+ "eval_samples_per_second": 42.602,
+ "eval_steps_per_second": 1.343,
+ "step": 5400
+ },
+ {
+ "epoch": 75.01397379912663,
+ "grad_norm": 0.014417114667594433,
+ "learning_rate": 0.0006,
+ "loss": 4.332956790924072,
+ "step": 5401
+ },
+ {
+ "epoch": 75.02794759825328,
+ "grad_norm": 0.01607639715075493,
+ "learning_rate": 0.0006,
+ "loss": 4.352453708648682,
+ "step": 5402
+ },
+ {
+ "epoch": 75.04192139737991,
+ "grad_norm": 0.015438318252563477,
+ "learning_rate": 0.0006,
+ "loss": 4.387139320373535,
+ "step": 5403
+ },
+ {
+ "epoch": 75.05589519650655,
+ "grad_norm": 0.015539007261395454,
+ "learning_rate": 0.0006,
+ "loss": 4.453171730041504,
+ "step": 5404
+ },
+ {
+ "epoch": 75.06986899563319,
+ "grad_norm": 0.017361903563141823,
+ "learning_rate": 0.0006,
+ "loss": 4.512143135070801,
+ "step": 5405
+ },
+ {
+ "epoch": 75.08384279475983,
+ "grad_norm": 0.01815599389374256,
+ "learning_rate": 0.0006,
+ "loss": 4.508997440338135,
+ "step": 5406
+ },
+ {
+ "epoch": 75.09781659388646,
+ "grad_norm": 0.01620607078075409,
+ "learning_rate": 0.0006,
+ "loss": 4.322444915771484,
+ "step": 5407
+ },
+ {
+ "epoch": 75.1117903930131,
+ "grad_norm": 0.014773265458643436,
+ "learning_rate": 0.0006,
+ "loss": 4.359251976013184,
+ "step": 5408
+ },
+ {
+ "epoch": 75.12576419213974,
+ "grad_norm": 0.014099023304879665,
+ "learning_rate": 0.0006,
+ "loss": 4.313380241394043,
+ "step": 5409
+ },
+ {
+ "epoch": 75.13973799126637,
+ "grad_norm": 0.014387735165655613,
+ "learning_rate": 0.0006,
+ "loss": 4.435023307800293,
+ "step": 5410
+ },
+ {
+ "epoch": 75.15371179039302,
+ "grad_norm": 0.014886162243783474,
+ "learning_rate": 0.0006,
+ "loss": 4.335939407348633,
+ "step": 5411
+ },
+ {
+ "epoch": 75.16768558951965,
+ "grad_norm": 0.015766102820634842,
+ "learning_rate": 0.0006,
+ "loss": 4.408674240112305,
+ "step": 5412
+ },
+ {
+ "epoch": 75.18165938864628,
+ "grad_norm": 0.016383284702897072,
+ "learning_rate": 0.0006,
+ "loss": 4.385453224182129,
+ "step": 5413
+ },
+ {
+ "epoch": 75.19563318777293,
+ "grad_norm": 0.013535384088754654,
+ "learning_rate": 0.0006,
+ "loss": 4.292785167694092,
+ "step": 5414
+ },
+ {
+ "epoch": 75.20960698689956,
+ "grad_norm": 0.01354036945849657,
+ "learning_rate": 0.0006,
+ "loss": 4.306252479553223,
+ "step": 5415
+ },
+ {
+ "epoch": 75.2235807860262,
+ "grad_norm": 0.015818078070878983,
+ "learning_rate": 0.0006,
+ "loss": 4.425487995147705,
+ "step": 5416
+ },
+ {
+ "epoch": 75.23755458515284,
+ "grad_norm": 0.014363568276166916,
+ "learning_rate": 0.0006,
+ "loss": 4.395087242126465,
+ "step": 5417
+ },
+ {
+ "epoch": 75.25152838427948,
+ "grad_norm": 0.013895470649003983,
+ "learning_rate": 0.0006,
+ "loss": 4.364090442657471,
+ "step": 5418
+ },
+ {
+ "epoch": 75.26550218340611,
+ "grad_norm": 0.014178668148815632,
+ "learning_rate": 0.0006,
+ "loss": 4.44619083404541,
+ "step": 5419
+ },
+ {
+ "epoch": 75.27947598253274,
+ "grad_norm": 0.0152514623478055,
+ "learning_rate": 0.0006,
+ "loss": 4.348445892333984,
+ "step": 5420
+ },
+ {
+ "epoch": 75.29344978165939,
+ "grad_norm": 0.015397852286696434,
+ "learning_rate": 0.0006,
+ "loss": 4.476832389831543,
+ "step": 5421
+ },
+ {
+ "epoch": 75.30742358078602,
+ "grad_norm": 0.018009863793849945,
+ "learning_rate": 0.0006,
+ "loss": 4.431162357330322,
+ "step": 5422
+ },
+ {
+ "epoch": 75.32139737991267,
+ "grad_norm": 0.017038367688655853,
+ "learning_rate": 0.0006,
+ "loss": 4.474787712097168,
+ "step": 5423
+ },
+ {
+ "epoch": 75.3353711790393,
+ "grad_norm": 0.013763288035988808,
+ "learning_rate": 0.0006,
+ "loss": 4.470419883728027,
+ "step": 5424
+ },
+ {
+ "epoch": 75.34934497816595,
+ "grad_norm": 0.013972110114991665,
+ "learning_rate": 0.0006,
+ "loss": 4.373147964477539,
+ "step": 5425
+ },
+ {
+ "epoch": 75.36331877729258,
+ "grad_norm": 0.017197875306010246,
+ "learning_rate": 0.0006,
+ "loss": 4.494963645935059,
+ "step": 5426
+ },
+ {
+ "epoch": 75.37729257641921,
+ "grad_norm": 0.01687047816812992,
+ "learning_rate": 0.0006,
+ "loss": 4.435934066772461,
+ "step": 5427
+ },
+ {
+ "epoch": 75.39126637554585,
+ "grad_norm": 0.01578272506594658,
+ "learning_rate": 0.0006,
+ "loss": 4.37918758392334,
+ "step": 5428
+ },
+ {
+ "epoch": 75.40524017467249,
+ "grad_norm": 0.015956247225403786,
+ "learning_rate": 0.0006,
+ "loss": 4.455668926239014,
+ "step": 5429
+ },
+ {
+ "epoch": 75.41921397379913,
+ "grad_norm": 0.013495155610144138,
+ "learning_rate": 0.0006,
+ "loss": 4.4271039962768555,
+ "step": 5430
+ },
+ {
+ "epoch": 75.43318777292576,
+ "grad_norm": 0.012851621024310589,
+ "learning_rate": 0.0006,
+ "loss": 4.213107109069824,
+ "step": 5431
+ },
+ {
+ "epoch": 75.44716157205241,
+ "grad_norm": 0.013351158238947392,
+ "learning_rate": 0.0006,
+ "loss": 4.3070783615112305,
+ "step": 5432
+ },
+ {
+ "epoch": 75.46113537117904,
+ "grad_norm": 0.014765217900276184,
+ "learning_rate": 0.0006,
+ "loss": 4.385021209716797,
+ "step": 5433
+ },
+ {
+ "epoch": 75.47510917030567,
+ "grad_norm": 0.016091614961624146,
+ "learning_rate": 0.0006,
+ "loss": 4.312644004821777,
+ "step": 5434
+ },
+ {
+ "epoch": 75.48908296943232,
+ "grad_norm": 0.01612013578414917,
+ "learning_rate": 0.0006,
+ "loss": 4.4482011795043945,
+ "step": 5435
+ },
+ {
+ "epoch": 75.50305676855895,
+ "grad_norm": 0.013157603330910206,
+ "learning_rate": 0.0006,
+ "loss": 4.427262306213379,
+ "step": 5436
+ },
+ {
+ "epoch": 75.5170305676856,
+ "grad_norm": 0.016039809212088585,
+ "learning_rate": 0.0006,
+ "loss": 4.396635055541992,
+ "step": 5437
+ },
+ {
+ "epoch": 75.53100436681223,
+ "grad_norm": 0.01794283092021942,
+ "learning_rate": 0.0006,
+ "loss": 4.476870059967041,
+ "step": 5438
+ },
+ {
+ "epoch": 75.54497816593886,
+ "grad_norm": 0.017030486837029457,
+ "learning_rate": 0.0006,
+ "loss": 4.361156463623047,
+ "step": 5439
+ },
+ {
+ "epoch": 75.5589519650655,
+ "grad_norm": 0.017915161326527596,
+ "learning_rate": 0.0006,
+ "loss": 4.286147117614746,
+ "step": 5440
+ },
+ {
+ "epoch": 75.57292576419214,
+ "grad_norm": 0.017514433711767197,
+ "learning_rate": 0.0006,
+ "loss": 4.394091606140137,
+ "step": 5441
+ },
+ {
+ "epoch": 75.58689956331878,
+ "grad_norm": 0.015128589235246181,
+ "learning_rate": 0.0006,
+ "loss": 4.383126735687256,
+ "step": 5442
+ },
+ {
+ "epoch": 75.60087336244541,
+ "grad_norm": 0.0202616099268198,
+ "learning_rate": 0.0006,
+ "loss": 4.336188316345215,
+ "step": 5443
+ },
+ {
+ "epoch": 75.61484716157206,
+ "grad_norm": 0.019978471100330353,
+ "learning_rate": 0.0006,
+ "loss": 4.461453437805176,
+ "step": 5444
+ },
+ {
+ "epoch": 75.62882096069869,
+ "grad_norm": 0.01738741435110569,
+ "learning_rate": 0.0006,
+ "loss": 4.371197700500488,
+ "step": 5445
+ },
+ {
+ "epoch": 75.64279475982532,
+ "grad_norm": 0.016157809644937515,
+ "learning_rate": 0.0006,
+ "loss": 4.3312835693359375,
+ "step": 5446
+ },
+ {
+ "epoch": 75.65676855895197,
+ "grad_norm": 0.016394613310694695,
+ "learning_rate": 0.0006,
+ "loss": 4.380190849304199,
+ "step": 5447
+ },
+ {
+ "epoch": 75.6707423580786,
+ "grad_norm": 0.014727243222296238,
+ "learning_rate": 0.0006,
+ "loss": 4.308028221130371,
+ "step": 5448
+ },
+ {
+ "epoch": 75.68471615720524,
+ "grad_norm": 0.01695958711206913,
+ "learning_rate": 0.0006,
+ "loss": 4.399503707885742,
+ "step": 5449
+ },
+ {
+ "epoch": 75.69868995633188,
+ "grad_norm": 0.015720795840024948,
+ "learning_rate": 0.0006,
+ "loss": 4.342249870300293,
+ "step": 5450
+ },
+ {
+ "epoch": 75.71266375545852,
+ "grad_norm": 0.01574772410094738,
+ "learning_rate": 0.0006,
+ "loss": 4.331503868103027,
+ "step": 5451
+ },
+ {
+ "epoch": 75.72663755458515,
+ "grad_norm": 0.014812648296356201,
+ "learning_rate": 0.0006,
+ "loss": 4.332413673400879,
+ "step": 5452
+ },
+ {
+ "epoch": 75.74061135371178,
+ "grad_norm": 0.012850426137447357,
+ "learning_rate": 0.0006,
+ "loss": 4.2884063720703125,
+ "step": 5453
+ },
+ {
+ "epoch": 75.75458515283843,
+ "grad_norm": 0.012876970693469048,
+ "learning_rate": 0.0006,
+ "loss": 4.399855613708496,
+ "step": 5454
+ },
+ {
+ "epoch": 75.76855895196506,
+ "grad_norm": 0.012958863750100136,
+ "learning_rate": 0.0006,
+ "loss": 4.411185264587402,
+ "step": 5455
+ },
+ {
+ "epoch": 75.78253275109171,
+ "grad_norm": 0.012197135016322136,
+ "learning_rate": 0.0006,
+ "loss": 4.272603988647461,
+ "step": 5456
+ },
+ {
+ "epoch": 75.79650655021834,
+ "grad_norm": 0.013483566232025623,
+ "learning_rate": 0.0006,
+ "loss": 4.3564348220825195,
+ "step": 5457
+ },
+ {
+ "epoch": 75.81048034934499,
+ "grad_norm": 0.012365429662168026,
+ "learning_rate": 0.0006,
+ "loss": 4.354825973510742,
+ "step": 5458
+ },
+ {
+ "epoch": 75.82445414847162,
+ "grad_norm": 0.012238141149282455,
+ "learning_rate": 0.0006,
+ "loss": 4.440438747406006,
+ "step": 5459
+ },
+ {
+ "epoch": 75.83842794759825,
+ "grad_norm": 0.013380305841565132,
+ "learning_rate": 0.0006,
+ "loss": 4.2813215255737305,
+ "step": 5460
+ },
+ {
+ "epoch": 75.8524017467249,
+ "grad_norm": 0.01443779468536377,
+ "learning_rate": 0.0006,
+ "loss": 4.289078235626221,
+ "step": 5461
+ },
+ {
+ "epoch": 75.86637554585153,
+ "grad_norm": 0.012646574527025223,
+ "learning_rate": 0.0006,
+ "loss": 4.374177932739258,
+ "step": 5462
+ },
+ {
+ "epoch": 75.88034934497817,
+ "grad_norm": 0.012403388507664204,
+ "learning_rate": 0.0006,
+ "loss": 4.343016147613525,
+ "step": 5463
+ },
+ {
+ "epoch": 75.8943231441048,
+ "grad_norm": 0.01319252047687769,
+ "learning_rate": 0.0006,
+ "loss": 4.309957027435303,
+ "step": 5464
+ },
+ {
+ "epoch": 75.90829694323143,
+ "grad_norm": 0.012597428634762764,
+ "learning_rate": 0.0006,
+ "loss": 4.403654098510742,
+ "step": 5465
+ },
+ {
+ "epoch": 75.92227074235808,
+ "grad_norm": 0.012724226340651512,
+ "learning_rate": 0.0006,
+ "loss": 4.326869010925293,
+ "step": 5466
+ },
+ {
+ "epoch": 75.93624454148471,
+ "grad_norm": 0.012698772363364697,
+ "learning_rate": 0.0006,
+ "loss": 4.268430233001709,
+ "step": 5467
+ },
+ {
+ "epoch": 75.95021834061136,
+ "grad_norm": 0.01376490481197834,
+ "learning_rate": 0.0006,
+ "loss": 4.42446231842041,
+ "step": 5468
+ },
+ {
+ "epoch": 75.96419213973799,
+ "grad_norm": 0.013175144791603088,
+ "learning_rate": 0.0006,
+ "loss": 4.374675750732422,
+ "step": 5469
+ },
+ {
+ "epoch": 75.97816593886463,
+ "grad_norm": 0.012731322087347507,
+ "learning_rate": 0.0006,
+ "loss": 4.372966766357422,
+ "step": 5470
+ },
+ {
+ "epoch": 75.99213973799127,
+ "grad_norm": 0.01360505260527134,
+ "learning_rate": 0.0006,
+ "loss": 4.3095903396606445,
+ "step": 5471
+ },
+ {
+ "epoch": 76.0,
+ "grad_norm": 0.014011901803314686,
+ "learning_rate": 0.0006,
+ "loss": 4.26954460144043,
+ "step": 5472
+ },
+ {
+ "epoch": 76.0,
+ "eval_loss": 4.689640998840332,
+ "eval_runtime": 56.9095,
+ "eval_samples_per_second": 42.91,
+ "eval_steps_per_second": 1.353,
+ "step": 5472
+ },
+ {
+ "epoch": 76.01397379912663,
+ "grad_norm": 0.015351055189967155,
+ "learning_rate": 0.0006,
+ "loss": 4.348268985748291,
+ "step": 5473
+ },
+ {
+ "epoch": 76.02794759825328,
+ "grad_norm": 0.017164217308163643,
+ "learning_rate": 0.0006,
+ "loss": 4.244204998016357,
+ "step": 5474
+ },
+ {
+ "epoch": 76.04192139737991,
+ "grad_norm": 0.021624932065606117,
+ "learning_rate": 0.0006,
+ "loss": 4.311168670654297,
+ "step": 5475
+ },
+ {
+ "epoch": 76.05589519650655,
+ "grad_norm": 0.028712468221783638,
+ "learning_rate": 0.0006,
+ "loss": 4.322287082672119,
+ "step": 5476
+ },
+ {
+ "epoch": 76.06986899563319,
+ "grad_norm": 0.027915487065911293,
+ "learning_rate": 0.0006,
+ "loss": 4.381207466125488,
+ "step": 5477
+ },
+ {
+ "epoch": 76.08384279475983,
+ "grad_norm": 0.017955303192138672,
+ "learning_rate": 0.0006,
+ "loss": 4.2635016441345215,
+ "step": 5478
+ },
+ {
+ "epoch": 76.09781659388646,
+ "grad_norm": 0.019113514572381973,
+ "learning_rate": 0.0006,
+ "loss": 4.399471282958984,
+ "step": 5479
+ },
+ {
+ "epoch": 76.1117903930131,
+ "grad_norm": 0.017502542585134506,
+ "learning_rate": 0.0006,
+ "loss": 4.321204662322998,
+ "step": 5480
+ },
+ {
+ "epoch": 76.12576419213974,
+ "grad_norm": 0.01852525770664215,
+ "learning_rate": 0.0006,
+ "loss": 4.217884063720703,
+ "step": 5481
+ },
+ {
+ "epoch": 76.13973799126637,
+ "grad_norm": 0.019836828112602234,
+ "learning_rate": 0.0006,
+ "loss": 4.278532028198242,
+ "step": 5482
+ },
+ {
+ "epoch": 76.15371179039302,
+ "grad_norm": 0.02101975679397583,
+ "learning_rate": 0.0006,
+ "loss": 4.246720314025879,
+ "step": 5483
+ },
+ {
+ "epoch": 76.16768558951965,
+ "grad_norm": 0.021202990785241127,
+ "learning_rate": 0.0006,
+ "loss": 4.517517566680908,
+ "step": 5484
+ },
+ {
+ "epoch": 76.18165938864628,
+ "grad_norm": 0.021428769454360008,
+ "learning_rate": 0.0006,
+ "loss": 4.3373517990112305,
+ "step": 5485
+ },
+ {
+ "epoch": 76.19563318777293,
+ "grad_norm": 0.018263190984725952,
+ "learning_rate": 0.0006,
+ "loss": 4.402856826782227,
+ "step": 5486
+ },
+ {
+ "epoch": 76.20960698689956,
+ "grad_norm": 0.019025536254048347,
+ "learning_rate": 0.0006,
+ "loss": 4.314227104187012,
+ "step": 5487
+ },
+ {
+ "epoch": 76.2235807860262,
+ "grad_norm": 0.020918630063533783,
+ "learning_rate": 0.0006,
+ "loss": 4.454369068145752,
+ "step": 5488
+ },
+ {
+ "epoch": 76.23755458515284,
+ "grad_norm": 0.019307617098093033,
+ "learning_rate": 0.0006,
+ "loss": 4.247509002685547,
+ "step": 5489
+ },
+ {
+ "epoch": 76.25152838427948,
+ "grad_norm": 0.014956331811845303,
+ "learning_rate": 0.0006,
+ "loss": 4.365175247192383,
+ "step": 5490
+ },
+ {
+ "epoch": 76.26550218340611,
+ "grad_norm": 0.017213059589266777,
+ "learning_rate": 0.0006,
+ "loss": 4.384005069732666,
+ "step": 5491
+ },
+ {
+ "epoch": 76.27947598253274,
+ "grad_norm": 0.017575597390532494,
+ "learning_rate": 0.0006,
+ "loss": 4.41934871673584,
+ "step": 5492
+ },
+ {
+ "epoch": 76.29344978165939,
+ "grad_norm": 0.015879256650805473,
+ "learning_rate": 0.0006,
+ "loss": 4.26675271987915,
+ "step": 5493
+ },
+ {
+ "epoch": 76.30742358078602,
+ "grad_norm": 0.01564522460103035,
+ "learning_rate": 0.0006,
+ "loss": 4.277339935302734,
+ "step": 5494
+ },
+ {
+ "epoch": 76.32139737991267,
+ "grad_norm": 0.016414586454629898,
+ "learning_rate": 0.0006,
+ "loss": 4.4138593673706055,
+ "step": 5495
+ },
+ {
+ "epoch": 76.3353711790393,
+ "grad_norm": 0.014732337556779385,
+ "learning_rate": 0.0006,
+ "loss": 4.281006336212158,
+ "step": 5496
+ },
+ {
+ "epoch": 76.34934497816595,
+ "grad_norm": 0.015461455099284649,
+ "learning_rate": 0.0006,
+ "loss": 4.263422966003418,
+ "step": 5497
+ },
+ {
+ "epoch": 76.36331877729258,
+ "grad_norm": 0.014529814012348652,
+ "learning_rate": 0.0006,
+ "loss": 4.36165189743042,
+ "step": 5498
+ },
+ {
+ "epoch": 76.37729257641921,
+ "grad_norm": 0.01469483319669962,
+ "learning_rate": 0.0006,
+ "loss": 4.309642791748047,
+ "step": 5499
+ },
+ {
+ "epoch": 76.39126637554585,
+ "grad_norm": 0.016490574926137924,
+ "learning_rate": 0.0006,
+ "loss": 4.288150787353516,
+ "step": 5500
+ },
+ {
+ "epoch": 76.40524017467249,
+ "grad_norm": 0.014593565836548805,
+ "learning_rate": 0.0006,
+ "loss": 4.271614074707031,
+ "step": 5501
+ },
+ {
+ "epoch": 76.41921397379913,
+ "grad_norm": 0.012796535156667233,
+ "learning_rate": 0.0006,
+ "loss": 4.420042037963867,
+ "step": 5502
+ },
+ {
+ "epoch": 76.43318777292576,
+ "grad_norm": 0.016086025163531303,
+ "learning_rate": 0.0006,
+ "loss": 4.289943695068359,
+ "step": 5503
+ },
+ {
+ "epoch": 76.44716157205241,
+ "grad_norm": 0.015439470298588276,
+ "learning_rate": 0.0006,
+ "loss": 4.3023271560668945,
+ "step": 5504
+ },
+ {
+ "epoch": 76.46113537117904,
+ "grad_norm": 0.015066367574036121,
+ "learning_rate": 0.0006,
+ "loss": 4.197081565856934,
+ "step": 5505
+ },
+ {
+ "epoch": 76.47510917030567,
+ "grad_norm": 0.016653263941407204,
+ "learning_rate": 0.0006,
+ "loss": 4.333785057067871,
+ "step": 5506
+ },
+ {
+ "epoch": 76.48908296943232,
+ "grad_norm": 0.014791973866522312,
+ "learning_rate": 0.0006,
+ "loss": 4.323811054229736,
+ "step": 5507
+ },
+ {
+ "epoch": 76.50305676855895,
+ "grad_norm": 0.014694827608764172,
+ "learning_rate": 0.0006,
+ "loss": 4.195852279663086,
+ "step": 5508
+ },
+ {
+ "epoch": 76.5170305676856,
+ "grad_norm": 0.015120396390557289,
+ "learning_rate": 0.0006,
+ "loss": 4.42379903793335,
+ "step": 5509
+ },
+ {
+ "epoch": 76.53100436681223,
+ "grad_norm": 0.014584025368094444,
+ "learning_rate": 0.0006,
+ "loss": 4.2702531814575195,
+ "step": 5510
+ },
+ {
+ "epoch": 76.54497816593886,
+ "grad_norm": 0.013981613330543041,
+ "learning_rate": 0.0006,
+ "loss": 4.383029937744141,
+ "step": 5511
+ },
+ {
+ "epoch": 76.5589519650655,
+ "grad_norm": 0.012930642813444138,
+ "learning_rate": 0.0006,
+ "loss": 4.296136856079102,
+ "step": 5512
+ },
+ {
+ "epoch": 76.57292576419214,
+ "grad_norm": 0.013060989789664745,
+ "learning_rate": 0.0006,
+ "loss": 4.346380233764648,
+ "step": 5513
+ },
+ {
+ "epoch": 76.58689956331878,
+ "grad_norm": 0.013209798373281956,
+ "learning_rate": 0.0006,
+ "loss": 4.316067218780518,
+ "step": 5514
+ },
+ {
+ "epoch": 76.60087336244541,
+ "grad_norm": 0.012423216365277767,
+ "learning_rate": 0.0006,
+ "loss": 4.370274543762207,
+ "step": 5515
+ },
+ {
+ "epoch": 76.61484716157206,
+ "grad_norm": 0.013205411843955517,
+ "learning_rate": 0.0006,
+ "loss": 4.429969310760498,
+ "step": 5516
+ },
+ {
+ "epoch": 76.62882096069869,
+ "grad_norm": 0.013049827888607979,
+ "learning_rate": 0.0006,
+ "loss": 4.378549098968506,
+ "step": 5517
+ },
+ {
+ "epoch": 76.64279475982532,
+ "grad_norm": 0.012075880542397499,
+ "learning_rate": 0.0006,
+ "loss": 4.348293304443359,
+ "step": 5518
+ },
+ {
+ "epoch": 76.65676855895197,
+ "grad_norm": 0.014932847581803799,
+ "learning_rate": 0.0006,
+ "loss": 4.282538890838623,
+ "step": 5519
+ },
+ {
+ "epoch": 76.6707423580786,
+ "grad_norm": 0.015615998767316341,
+ "learning_rate": 0.0006,
+ "loss": 4.274641036987305,
+ "step": 5520
+ },
+ {
+ "epoch": 76.68471615720524,
+ "grad_norm": 0.01424513477832079,
+ "learning_rate": 0.0006,
+ "loss": 4.243166923522949,
+ "step": 5521
+ },
+ {
+ "epoch": 76.69868995633188,
+ "grad_norm": 0.012775546871125698,
+ "learning_rate": 0.0006,
+ "loss": 4.335611820220947,
+ "step": 5522
+ },
+ {
+ "epoch": 76.71266375545852,
+ "grad_norm": 0.014461303129792213,
+ "learning_rate": 0.0006,
+ "loss": 4.246976375579834,
+ "step": 5523
+ },
+ {
+ "epoch": 76.72663755458515,
+ "grad_norm": 0.0145436255261302,
+ "learning_rate": 0.0006,
+ "loss": 4.275895118713379,
+ "step": 5524
+ },
+ {
+ "epoch": 76.74061135371178,
+ "grad_norm": 0.014055085368454456,
+ "learning_rate": 0.0006,
+ "loss": 4.331108093261719,
+ "step": 5525
+ },
+ {
+ "epoch": 76.75458515283843,
+ "grad_norm": 0.011933894827961922,
+ "learning_rate": 0.0006,
+ "loss": 4.293721675872803,
+ "step": 5526
+ },
+ {
+ "epoch": 76.76855895196506,
+ "grad_norm": 0.01277392078191042,
+ "learning_rate": 0.0006,
+ "loss": 4.312374114990234,
+ "step": 5527
+ },
+ {
+ "epoch": 76.78253275109171,
+ "grad_norm": 0.014227936044335365,
+ "learning_rate": 0.0006,
+ "loss": 4.238729000091553,
+ "step": 5528
+ },
+ {
+ "epoch": 76.79650655021834,
+ "grad_norm": 0.01488024927675724,
+ "learning_rate": 0.0006,
+ "loss": 4.24587345123291,
+ "step": 5529
+ },
+ {
+ "epoch": 76.81048034934499,
+ "grad_norm": 0.015341250225901604,
+ "learning_rate": 0.0006,
+ "loss": 4.257378578186035,
+ "step": 5530
+ },
+ {
+ "epoch": 76.82445414847162,
+ "grad_norm": 0.015179437585175037,
+ "learning_rate": 0.0006,
+ "loss": 4.2812910079956055,
+ "step": 5531
+ },
+ {
+ "epoch": 76.83842794759825,
+ "grad_norm": 0.012814724817872047,
+ "learning_rate": 0.0006,
+ "loss": 4.2309465408325195,
+ "step": 5532
+ },
+ {
+ "epoch": 76.8524017467249,
+ "grad_norm": 0.013810716569423676,
+ "learning_rate": 0.0006,
+ "loss": 4.325292110443115,
+ "step": 5533
+ },
+ {
+ "epoch": 76.86637554585153,
+ "grad_norm": 0.017468232661485672,
+ "learning_rate": 0.0006,
+ "loss": 4.243788719177246,
+ "step": 5534
+ },
+ {
+ "epoch": 76.88034934497817,
+ "grad_norm": 0.01739857904613018,
+ "learning_rate": 0.0006,
+ "loss": 4.345390796661377,
+ "step": 5535
+ },
+ {
+ "epoch": 76.8943231441048,
+ "grad_norm": 0.015329061076045036,
+ "learning_rate": 0.0006,
+ "loss": 4.415142059326172,
+ "step": 5536
+ },
+ {
+ "epoch": 76.90829694323143,
+ "grad_norm": 0.01598426327109337,
+ "learning_rate": 0.0006,
+ "loss": 4.299558639526367,
+ "step": 5537
+ },
+ {
+ "epoch": 76.92227074235808,
+ "grad_norm": 0.01621454581618309,
+ "learning_rate": 0.0006,
+ "loss": 4.284354209899902,
+ "step": 5538
+ },
+ {
+ "epoch": 76.93624454148471,
+ "grad_norm": 0.015956968069076538,
+ "learning_rate": 0.0006,
+ "loss": 4.293837547302246,
+ "step": 5539
+ },
+ {
+ "epoch": 76.95021834061136,
+ "grad_norm": 0.01776493526995182,
+ "learning_rate": 0.0006,
+ "loss": 4.229729175567627,
+ "step": 5540
+ },
+ {
+ "epoch": 76.96419213973799,
+ "grad_norm": 0.015317477285861969,
+ "learning_rate": 0.0006,
+ "loss": 4.3328447341918945,
+ "step": 5541
+ },
+ {
+ "epoch": 76.97816593886463,
+ "grad_norm": 0.015140251256525517,
+ "learning_rate": 0.0006,
+ "loss": 4.310840606689453,
+ "step": 5542
+ },
+ {
+ "epoch": 76.99213973799127,
+ "grad_norm": 0.015326812863349915,
+ "learning_rate": 0.0006,
+ "loss": 4.230832099914551,
+ "step": 5543
+ },
+ {
+ "epoch": 77.0,
+ "grad_norm": 0.01607763208448887,
+ "learning_rate": 0.0006,
+ "loss": 4.2236328125,
+ "step": 5544
+ },
+ {
+ "epoch": 77.0,
+ "eval_loss": 4.662221431732178,
+ "eval_runtime": 56.5194,
+ "eval_samples_per_second": 43.206,
+ "eval_steps_per_second": 1.362,
+ "step": 5544
+ },
+ {
+ "epoch": 77.01397379912663,
+ "grad_norm": 0.014839510433375835,
+ "learning_rate": 0.0006,
+ "loss": 4.254111289978027,
+ "step": 5545
+ },
+ {
+ "epoch": 77.02794759825328,
+ "grad_norm": 0.01519741676747799,
+ "learning_rate": 0.0006,
+ "loss": 4.1607184410095215,
+ "step": 5546
+ },
+ {
+ "epoch": 77.04192139737991,
+ "grad_norm": 0.015472603030502796,
+ "learning_rate": 0.0006,
+ "loss": 4.323320388793945,
+ "step": 5547
+ },
+ {
+ "epoch": 77.05589519650655,
+ "grad_norm": 0.017937948927283287,
+ "learning_rate": 0.0006,
+ "loss": 4.308490753173828,
+ "step": 5548
+ },
+ {
+ "epoch": 77.06986899563319,
+ "grad_norm": 0.020890841260552406,
+ "learning_rate": 0.0006,
+ "loss": 4.259101867675781,
+ "step": 5549
+ },
+ {
+ "epoch": 77.08384279475983,
+ "grad_norm": 0.01892857998609543,
+ "learning_rate": 0.0006,
+ "loss": 4.386970520019531,
+ "step": 5550
+ },
+ {
+ "epoch": 77.09781659388646,
+ "grad_norm": 0.016583630815148354,
+ "learning_rate": 0.0006,
+ "loss": 4.233604907989502,
+ "step": 5551
+ },
+ {
+ "epoch": 77.1117903930131,
+ "grad_norm": 0.016063468530774117,
+ "learning_rate": 0.0006,
+ "loss": 4.339544296264648,
+ "step": 5552
+ },
+ {
+ "epoch": 77.12576419213974,
+ "grad_norm": 0.018326779827475548,
+ "learning_rate": 0.0006,
+ "loss": 4.367204666137695,
+ "step": 5553
+ },
+ {
+ "epoch": 77.13973799126637,
+ "grad_norm": 0.021590933203697205,
+ "learning_rate": 0.0006,
+ "loss": 4.352685928344727,
+ "step": 5554
+ },
+ {
+ "epoch": 77.15371179039302,
+ "grad_norm": 0.019643642008304596,
+ "learning_rate": 0.0006,
+ "loss": 4.258535385131836,
+ "step": 5555
+ },
+ {
+ "epoch": 77.16768558951965,
+ "grad_norm": 0.018476149067282677,
+ "learning_rate": 0.0006,
+ "loss": 4.232568740844727,
+ "step": 5556
+ },
+ {
+ "epoch": 77.18165938864628,
+ "grad_norm": 0.01584405079483986,
+ "learning_rate": 0.0006,
+ "loss": 4.258371353149414,
+ "step": 5557
+ },
+ {
+ "epoch": 77.19563318777293,
+ "grad_norm": 0.01575201191008091,
+ "learning_rate": 0.0006,
+ "loss": 4.183218002319336,
+ "step": 5558
+ },
+ {
+ "epoch": 77.20960698689956,
+ "grad_norm": 0.016871824860572815,
+ "learning_rate": 0.0006,
+ "loss": 4.194563865661621,
+ "step": 5559
+ },
+ {
+ "epoch": 77.2235807860262,
+ "grad_norm": 0.02059333398938179,
+ "learning_rate": 0.0006,
+ "loss": 4.195281982421875,
+ "step": 5560
+ },
+ {
+ "epoch": 77.23755458515284,
+ "grad_norm": 0.022640453651547432,
+ "learning_rate": 0.0006,
+ "loss": 4.279615879058838,
+ "step": 5561
+ },
+ {
+ "epoch": 77.25152838427948,
+ "grad_norm": 0.020335959270596504,
+ "learning_rate": 0.0006,
+ "loss": 4.25040864944458,
+ "step": 5562
+ },
+ {
+ "epoch": 77.26550218340611,
+ "grad_norm": 0.020010793581604958,
+ "learning_rate": 0.0006,
+ "loss": 4.291116714477539,
+ "step": 5563
+ },
+ {
+ "epoch": 77.27947598253274,
+ "grad_norm": 0.02109045907855034,
+ "learning_rate": 0.0006,
+ "loss": 4.364088535308838,
+ "step": 5564
+ },
+ {
+ "epoch": 77.29344978165939,
+ "grad_norm": 0.019519446417689323,
+ "learning_rate": 0.0006,
+ "loss": 4.329166412353516,
+ "step": 5565
+ },
+ {
+ "epoch": 77.30742358078602,
+ "grad_norm": 0.01689804159104824,
+ "learning_rate": 0.0006,
+ "loss": 4.244176864624023,
+ "step": 5566
+ },
+ {
+ "epoch": 77.32139737991267,
+ "grad_norm": 0.01893681474030018,
+ "learning_rate": 0.0006,
+ "loss": 4.239068984985352,
+ "step": 5567
+ },
+ {
+ "epoch": 77.3353711790393,
+ "grad_norm": 0.015998193994164467,
+ "learning_rate": 0.0006,
+ "loss": 4.173398017883301,
+ "step": 5568
+ },
+ {
+ "epoch": 77.34934497816595,
+ "grad_norm": 0.013892280869185925,
+ "learning_rate": 0.0006,
+ "loss": 4.287654876708984,
+ "step": 5569
+ },
+ {
+ "epoch": 77.36331877729258,
+ "grad_norm": 0.017060643061995506,
+ "learning_rate": 0.0006,
+ "loss": 4.398447036743164,
+ "step": 5570
+ },
+ {
+ "epoch": 77.37729257641921,
+ "grad_norm": 0.015168849378824234,
+ "learning_rate": 0.0006,
+ "loss": 4.15665864944458,
+ "step": 5571
+ },
+ {
+ "epoch": 77.39126637554585,
+ "grad_norm": 0.014886523596942425,
+ "learning_rate": 0.0006,
+ "loss": 4.196673393249512,
+ "step": 5572
+ },
+ {
+ "epoch": 77.40524017467249,
+ "grad_norm": 0.016941796988248825,
+ "learning_rate": 0.0006,
+ "loss": 4.314562797546387,
+ "step": 5573
+ },
+ {
+ "epoch": 77.41921397379913,
+ "grad_norm": 0.015353812836110592,
+ "learning_rate": 0.0006,
+ "loss": 4.409579277038574,
+ "step": 5574
+ },
+ {
+ "epoch": 77.43318777292576,
+ "grad_norm": 0.015192852355539799,
+ "learning_rate": 0.0006,
+ "loss": 4.392045974731445,
+ "step": 5575
+ },
+ {
+ "epoch": 77.44716157205241,
+ "grad_norm": 0.014232569374144077,
+ "learning_rate": 0.0006,
+ "loss": 4.21639347076416,
+ "step": 5576
+ },
+ {
+ "epoch": 77.46113537117904,
+ "grad_norm": 0.014453226700425148,
+ "learning_rate": 0.0006,
+ "loss": 4.2802276611328125,
+ "step": 5577
+ },
+ {
+ "epoch": 77.47510917030567,
+ "grad_norm": 0.012881236150860786,
+ "learning_rate": 0.0006,
+ "loss": 4.375980854034424,
+ "step": 5578
+ },
+ {
+ "epoch": 77.48908296943232,
+ "grad_norm": 0.01583976298570633,
+ "learning_rate": 0.0006,
+ "loss": 4.264520645141602,
+ "step": 5579
+ },
+ {
+ "epoch": 77.50305676855895,
+ "grad_norm": 0.014171584509313107,
+ "learning_rate": 0.0006,
+ "loss": 4.225279808044434,
+ "step": 5580
+ },
+ {
+ "epoch": 77.5170305676856,
+ "grad_norm": 0.014715159311890602,
+ "learning_rate": 0.0006,
+ "loss": 4.369834899902344,
+ "step": 5581
+ },
+ {
+ "epoch": 77.53100436681223,
+ "grad_norm": 0.016554079949855804,
+ "learning_rate": 0.0006,
+ "loss": 4.2976179122924805,
+ "step": 5582
+ },
+ {
+ "epoch": 77.54497816593886,
+ "grad_norm": 0.015095747075974941,
+ "learning_rate": 0.0006,
+ "loss": 4.294407844543457,
+ "step": 5583
+ },
+ {
+ "epoch": 77.5589519650655,
+ "grad_norm": 0.012726361863315105,
+ "learning_rate": 0.0006,
+ "loss": 4.221401214599609,
+ "step": 5584
+ },
+ {
+ "epoch": 77.57292576419214,
+ "grad_norm": 0.014999309554696083,
+ "learning_rate": 0.0006,
+ "loss": 4.243929386138916,
+ "step": 5585
+ },
+ {
+ "epoch": 77.58689956331878,
+ "grad_norm": 0.016526037827134132,
+ "learning_rate": 0.0006,
+ "loss": 4.374139785766602,
+ "step": 5586
+ },
+ {
+ "epoch": 77.60087336244541,
+ "grad_norm": 0.01648339442908764,
+ "learning_rate": 0.0006,
+ "loss": 4.392425060272217,
+ "step": 5587
+ },
+ {
+ "epoch": 77.61484716157206,
+ "grad_norm": 0.015269213356077671,
+ "learning_rate": 0.0006,
+ "loss": 4.33986759185791,
+ "step": 5588
+ },
+ {
+ "epoch": 77.62882096069869,
+ "grad_norm": 0.01690507121384144,
+ "learning_rate": 0.0006,
+ "loss": 4.307476043701172,
+ "step": 5589
+ },
+ {
+ "epoch": 77.64279475982532,
+ "grad_norm": 0.016438685357570648,
+ "learning_rate": 0.0006,
+ "loss": 4.2406206130981445,
+ "step": 5590
+ },
+ {
+ "epoch": 77.65676855895197,
+ "grad_norm": 0.0157600287348032,
+ "learning_rate": 0.0006,
+ "loss": 4.314070224761963,
+ "step": 5591
+ },
+ {
+ "epoch": 77.6707423580786,
+ "grad_norm": 0.015127947553992271,
+ "learning_rate": 0.0006,
+ "loss": 4.22475528717041,
+ "step": 5592
+ },
+ {
+ "epoch": 77.68471615720524,
+ "grad_norm": 0.016632677987217903,
+ "learning_rate": 0.0006,
+ "loss": 4.26954984664917,
+ "step": 5593
+ },
+ {
+ "epoch": 77.69868995633188,
+ "grad_norm": 0.013150460086762905,
+ "learning_rate": 0.0006,
+ "loss": 4.369131088256836,
+ "step": 5594
+ },
+ {
+ "epoch": 77.71266375545852,
+ "grad_norm": 0.014513005502521992,
+ "learning_rate": 0.0006,
+ "loss": 4.3248796463012695,
+ "step": 5595
+ },
+ {
+ "epoch": 77.72663755458515,
+ "grad_norm": 0.014090991578996181,
+ "learning_rate": 0.0006,
+ "loss": 4.268381118774414,
+ "step": 5596
+ },
+ {
+ "epoch": 77.74061135371178,
+ "grad_norm": 0.01633097417652607,
+ "learning_rate": 0.0006,
+ "loss": 4.238439083099365,
+ "step": 5597
+ },
+ {
+ "epoch": 77.75458515283843,
+ "grad_norm": 0.01710933819413185,
+ "learning_rate": 0.0006,
+ "loss": 4.399697780609131,
+ "step": 5598
+ },
+ {
+ "epoch": 77.76855895196506,
+ "grad_norm": 0.019251767545938492,
+ "learning_rate": 0.0006,
+ "loss": 4.290450096130371,
+ "step": 5599
+ },
+ {
+ "epoch": 77.78253275109171,
+ "grad_norm": 0.01849171705543995,
+ "learning_rate": 0.0006,
+ "loss": 4.267734527587891,
+ "step": 5600
+ },
+ {
+ "epoch": 77.79650655021834,
+ "grad_norm": 0.014173115603625774,
+ "learning_rate": 0.0006,
+ "loss": 4.288799285888672,
+ "step": 5601
+ },
+ {
+ "epoch": 77.81048034934499,
+ "grad_norm": 0.014572718180716038,
+ "learning_rate": 0.0006,
+ "loss": 4.3790435791015625,
+ "step": 5602
+ },
+ {
+ "epoch": 77.82445414847162,
+ "grad_norm": 0.016375336796045303,
+ "learning_rate": 0.0006,
+ "loss": 4.231725692749023,
+ "step": 5603
+ },
+ {
+ "epoch": 77.83842794759825,
+ "grad_norm": 0.017326701432466507,
+ "learning_rate": 0.0006,
+ "loss": 4.211495399475098,
+ "step": 5604
+ },
+ {
+ "epoch": 77.8524017467249,
+ "grad_norm": 0.016399390995502472,
+ "learning_rate": 0.0006,
+ "loss": 4.386699199676514,
+ "step": 5605
+ },
+ {
+ "epoch": 77.86637554585153,
+ "grad_norm": 0.018269989639520645,
+ "learning_rate": 0.0006,
+ "loss": 4.254856109619141,
+ "step": 5606
+ },
+ {
+ "epoch": 77.88034934497817,
+ "grad_norm": 0.018443435430526733,
+ "learning_rate": 0.0006,
+ "loss": 4.391961097717285,
+ "step": 5607
+ },
+ {
+ "epoch": 77.8943231441048,
+ "grad_norm": 0.01784636080265045,
+ "learning_rate": 0.0006,
+ "loss": 4.273881912231445,
+ "step": 5608
+ },
+ {
+ "epoch": 77.90829694323143,
+ "grad_norm": 0.01782994158565998,
+ "learning_rate": 0.0006,
+ "loss": 4.26833963394165,
+ "step": 5609
+ },
+ {
+ "epoch": 77.92227074235808,
+ "grad_norm": 0.016260741278529167,
+ "learning_rate": 0.0006,
+ "loss": 4.306807994842529,
+ "step": 5610
+ },
+ {
+ "epoch": 77.93624454148471,
+ "grad_norm": 0.015482903458178043,
+ "learning_rate": 0.0006,
+ "loss": 4.187826633453369,
+ "step": 5611
+ },
+ {
+ "epoch": 77.95021834061136,
+ "grad_norm": 0.015193257480859756,
+ "learning_rate": 0.0006,
+ "loss": 4.286060333251953,
+ "step": 5612
+ },
+ {
+ "epoch": 77.96419213973799,
+ "grad_norm": 0.015110115520656109,
+ "learning_rate": 0.0006,
+ "loss": 4.2828192710876465,
+ "step": 5613
+ },
+ {
+ "epoch": 77.97816593886463,
+ "grad_norm": 0.015973486006259918,
+ "learning_rate": 0.0006,
+ "loss": 4.317939758300781,
+ "step": 5614
+ },
+ {
+ "epoch": 77.99213973799127,
+ "grad_norm": 0.017363594844937325,
+ "learning_rate": 0.0006,
+ "loss": 4.287721633911133,
+ "step": 5615
+ },
+ {
+ "epoch": 78.0,
+ "grad_norm": 0.015277642756700516,
+ "learning_rate": 0.0006,
+ "loss": 4.356618881225586,
+ "step": 5616
+ },
+ {
+ "epoch": 78.0,
+ "eval_loss": 4.642678737640381,
+ "eval_runtime": 56.7335,
+ "eval_samples_per_second": 43.043,
+ "eval_steps_per_second": 1.357,
+ "step": 5616
+ },
+ {
+ "epoch": 78.01397379912663,
+ "grad_norm": 0.016794051975011826,
+ "learning_rate": 0.0006,
+ "loss": 4.279930591583252,
+ "step": 5617
+ },
+ {
+ "epoch": 78.02794759825328,
+ "grad_norm": 0.01632673107087612,
+ "learning_rate": 0.0006,
+ "loss": 4.235943794250488,
+ "step": 5618
+ },
+ {
+ "epoch": 78.04192139737991,
+ "grad_norm": 0.01437876746058464,
+ "learning_rate": 0.0006,
+ "loss": 4.2542243003845215,
+ "step": 5619
+ },
+ {
+ "epoch": 78.05589519650655,
+ "grad_norm": 0.014993119053542614,
+ "learning_rate": 0.0006,
+ "loss": 4.262046813964844,
+ "step": 5620
+ },
+ {
+ "epoch": 78.06986899563319,
+ "grad_norm": 0.015721119940280914,
+ "learning_rate": 0.0006,
+ "loss": 4.173436164855957,
+ "step": 5621
+ },
+ {
+ "epoch": 78.08384279475983,
+ "grad_norm": 0.01334085687994957,
+ "learning_rate": 0.0006,
+ "loss": 4.322175979614258,
+ "step": 5622
+ },
+ {
+ "epoch": 78.09781659388646,
+ "grad_norm": 0.015561388805508614,
+ "learning_rate": 0.0006,
+ "loss": 4.242125988006592,
+ "step": 5623
+ },
+ {
+ "epoch": 78.1117903930131,
+ "grad_norm": 0.014418602921068668,
+ "learning_rate": 0.0006,
+ "loss": 4.266186714172363,
+ "step": 5624
+ },
+ {
+ "epoch": 78.12576419213974,
+ "grad_norm": 0.013739909045398235,
+ "learning_rate": 0.0006,
+ "loss": 4.231633186340332,
+ "step": 5625
+ },
+ {
+ "epoch": 78.13973799126637,
+ "grad_norm": 0.01402067206799984,
+ "learning_rate": 0.0006,
+ "loss": 4.188406944274902,
+ "step": 5626
+ },
+ {
+ "epoch": 78.15371179039302,
+ "grad_norm": 0.013639481738209724,
+ "learning_rate": 0.0006,
+ "loss": 4.1586594581604,
+ "step": 5627
+ },
+ {
+ "epoch": 78.16768558951965,
+ "grad_norm": 0.012858159840106964,
+ "learning_rate": 0.0006,
+ "loss": 4.360448360443115,
+ "step": 5628
+ },
+ {
+ "epoch": 78.18165938864628,
+ "grad_norm": 0.013089739717543125,
+ "learning_rate": 0.0006,
+ "loss": 4.344371795654297,
+ "step": 5629
+ },
+ {
+ "epoch": 78.19563318777293,
+ "grad_norm": 0.01382314320653677,
+ "learning_rate": 0.0006,
+ "loss": 4.20378303527832,
+ "step": 5630
+ },
+ {
+ "epoch": 78.20960698689956,
+ "grad_norm": 0.014206619001924992,
+ "learning_rate": 0.0006,
+ "loss": 4.239891529083252,
+ "step": 5631
+ },
+ {
+ "epoch": 78.2235807860262,
+ "grad_norm": 0.01472168043255806,
+ "learning_rate": 0.0006,
+ "loss": 4.282382488250732,
+ "step": 5632
+ },
+ {
+ "epoch": 78.23755458515284,
+ "grad_norm": 0.014237025752663612,
+ "learning_rate": 0.0006,
+ "loss": 4.241545677185059,
+ "step": 5633
+ },
+ {
+ "epoch": 78.25152838427948,
+ "grad_norm": 0.013231472112238407,
+ "learning_rate": 0.0006,
+ "loss": 4.184896469116211,
+ "step": 5634
+ },
+ {
+ "epoch": 78.26550218340611,
+ "grad_norm": 0.013538471423089504,
+ "learning_rate": 0.0006,
+ "loss": 4.325181007385254,
+ "step": 5635
+ },
+ {
+ "epoch": 78.27947598253274,
+ "grad_norm": 0.013541961088776588,
+ "learning_rate": 0.0006,
+ "loss": 4.368722915649414,
+ "step": 5636
+ },
+ {
+ "epoch": 78.29344978165939,
+ "grad_norm": 0.013446723110973835,
+ "learning_rate": 0.0006,
+ "loss": 4.301422595977783,
+ "step": 5637
+ },
+ {
+ "epoch": 78.30742358078602,
+ "grad_norm": 0.014753809198737144,
+ "learning_rate": 0.0006,
+ "loss": 4.259127616882324,
+ "step": 5638
+ },
+ {
+ "epoch": 78.32139737991267,
+ "grad_norm": 0.013256008736789227,
+ "learning_rate": 0.0006,
+ "loss": 4.284032821655273,
+ "step": 5639
+ },
+ {
+ "epoch": 78.3353711790393,
+ "grad_norm": 0.014272629283368587,
+ "learning_rate": 0.0006,
+ "loss": 4.234989166259766,
+ "step": 5640
+ },
+ {
+ "epoch": 78.34934497816595,
+ "grad_norm": 0.014756826683878899,
+ "learning_rate": 0.0006,
+ "loss": 4.143499374389648,
+ "step": 5641
+ },
+ {
+ "epoch": 78.36331877729258,
+ "grad_norm": 0.016549205407500267,
+ "learning_rate": 0.0006,
+ "loss": 4.311034202575684,
+ "step": 5642
+ },
+ {
+ "epoch": 78.37729257641921,
+ "grad_norm": 0.017735283821821213,
+ "learning_rate": 0.0006,
+ "loss": 4.246767520904541,
+ "step": 5643
+ },
+ {
+ "epoch": 78.39126637554585,
+ "grad_norm": 0.017370613291859627,
+ "learning_rate": 0.0006,
+ "loss": 4.2460832595825195,
+ "step": 5644
+ },
+ {
+ "epoch": 78.40524017467249,
+ "grad_norm": 0.014642846770584583,
+ "learning_rate": 0.0006,
+ "loss": 4.303811550140381,
+ "step": 5645
+ },
+ {
+ "epoch": 78.41921397379913,
+ "grad_norm": 0.013445726595818996,
+ "learning_rate": 0.0006,
+ "loss": 4.25921630859375,
+ "step": 5646
+ },
+ {
+ "epoch": 78.43318777292576,
+ "grad_norm": 0.014254065230488777,
+ "learning_rate": 0.0006,
+ "loss": 4.129446506500244,
+ "step": 5647
+ },
+ {
+ "epoch": 78.44716157205241,
+ "grad_norm": 0.01883019506931305,
+ "learning_rate": 0.0006,
+ "loss": 4.191716194152832,
+ "step": 5648
+ },
+ {
+ "epoch": 78.46113537117904,
+ "grad_norm": 0.0204437468200922,
+ "learning_rate": 0.0006,
+ "loss": 4.333085536956787,
+ "step": 5649
+ },
+ {
+ "epoch": 78.47510917030567,
+ "grad_norm": 0.01652318798005581,
+ "learning_rate": 0.0006,
+ "loss": 4.135462760925293,
+ "step": 5650
+ },
+ {
+ "epoch": 78.48908296943232,
+ "grad_norm": 0.012753551825881004,
+ "learning_rate": 0.0006,
+ "loss": 4.195051193237305,
+ "step": 5651
+ },
+ {
+ "epoch": 78.50305676855895,
+ "grad_norm": 0.01499118935316801,
+ "learning_rate": 0.0006,
+ "loss": 4.157092094421387,
+ "step": 5652
+ },
+ {
+ "epoch": 78.5170305676856,
+ "grad_norm": 0.015383945778012276,
+ "learning_rate": 0.0006,
+ "loss": 4.253818035125732,
+ "step": 5653
+ },
+ {
+ "epoch": 78.53100436681223,
+ "grad_norm": 0.015721332281827927,
+ "learning_rate": 0.0006,
+ "loss": 4.228775978088379,
+ "step": 5654
+ },
+ {
+ "epoch": 78.54497816593886,
+ "grad_norm": 0.014595463871955872,
+ "learning_rate": 0.0006,
+ "loss": 4.135197639465332,
+ "step": 5655
+ },
+ {
+ "epoch": 78.5589519650655,
+ "grad_norm": 0.015059644356369972,
+ "learning_rate": 0.0006,
+ "loss": 4.255614280700684,
+ "step": 5656
+ },
+ {
+ "epoch": 78.57292576419214,
+ "grad_norm": 0.013833344914019108,
+ "learning_rate": 0.0006,
+ "loss": 4.301250457763672,
+ "step": 5657
+ },
+ {
+ "epoch": 78.58689956331878,
+ "grad_norm": 0.013665376231074333,
+ "learning_rate": 0.0006,
+ "loss": 4.1734232902526855,
+ "step": 5658
+ },
+ {
+ "epoch": 78.60087336244541,
+ "grad_norm": 0.01539086177945137,
+ "learning_rate": 0.0006,
+ "loss": 4.221459865570068,
+ "step": 5659
+ },
+ {
+ "epoch": 78.61484716157206,
+ "grad_norm": 0.013313917443156242,
+ "learning_rate": 0.0006,
+ "loss": 4.255619049072266,
+ "step": 5660
+ },
+ {
+ "epoch": 78.62882096069869,
+ "grad_norm": 0.01447741687297821,
+ "learning_rate": 0.0006,
+ "loss": 4.298219680786133,
+ "step": 5661
+ },
+ {
+ "epoch": 78.64279475982532,
+ "grad_norm": 0.012796915136277676,
+ "learning_rate": 0.0006,
+ "loss": 4.2542405128479,
+ "step": 5662
+ },
+ {
+ "epoch": 78.65676855895197,
+ "grad_norm": 0.014058141969144344,
+ "learning_rate": 0.0006,
+ "loss": 4.284477233886719,
+ "step": 5663
+ },
+ {
+ "epoch": 78.6707423580786,
+ "grad_norm": 0.012705069966614246,
+ "learning_rate": 0.0006,
+ "loss": 4.2826666831970215,
+ "step": 5664
+ },
+ {
+ "epoch": 78.68471615720524,
+ "grad_norm": 0.013245890848338604,
+ "learning_rate": 0.0006,
+ "loss": 4.371612548828125,
+ "step": 5665
+ },
+ {
+ "epoch": 78.69868995633188,
+ "grad_norm": 0.014192677102982998,
+ "learning_rate": 0.0006,
+ "loss": 4.267344951629639,
+ "step": 5666
+ },
+ {
+ "epoch": 78.71266375545852,
+ "grad_norm": 0.015346777625381947,
+ "learning_rate": 0.0006,
+ "loss": 4.222797393798828,
+ "step": 5667
+ },
+ {
+ "epoch": 78.72663755458515,
+ "grad_norm": 0.014068621210753918,
+ "learning_rate": 0.0006,
+ "loss": 4.172964572906494,
+ "step": 5668
+ },
+ {
+ "epoch": 78.74061135371178,
+ "grad_norm": 0.014364945702254772,
+ "learning_rate": 0.0006,
+ "loss": 4.259319305419922,
+ "step": 5669
+ },
+ {
+ "epoch": 78.75458515283843,
+ "grad_norm": 0.017337387427687645,
+ "learning_rate": 0.0006,
+ "loss": 4.29533576965332,
+ "step": 5670
+ },
+ {
+ "epoch": 78.76855895196506,
+ "grad_norm": 0.017416715621948242,
+ "learning_rate": 0.0006,
+ "loss": 4.340364456176758,
+ "step": 5671
+ },
+ {
+ "epoch": 78.78253275109171,
+ "grad_norm": 0.015925846993923187,
+ "learning_rate": 0.0006,
+ "loss": 4.177454948425293,
+ "step": 5672
+ },
+ {
+ "epoch": 78.79650655021834,
+ "grad_norm": 0.013130443170666695,
+ "learning_rate": 0.0006,
+ "loss": 4.277454853057861,
+ "step": 5673
+ },
+ {
+ "epoch": 78.81048034934499,
+ "grad_norm": 0.014628687873482704,
+ "learning_rate": 0.0006,
+ "loss": 4.273757457733154,
+ "step": 5674
+ },
+ {
+ "epoch": 78.82445414847162,
+ "grad_norm": 0.01601659320294857,
+ "learning_rate": 0.0006,
+ "loss": 4.281774520874023,
+ "step": 5675
+ },
+ {
+ "epoch": 78.83842794759825,
+ "grad_norm": 0.015443485230207443,
+ "learning_rate": 0.0006,
+ "loss": 4.199937343597412,
+ "step": 5676
+ },
+ {
+ "epoch": 78.8524017467249,
+ "grad_norm": 0.013934246264398098,
+ "learning_rate": 0.0006,
+ "loss": 4.331369876861572,
+ "step": 5677
+ },
+ {
+ "epoch": 78.86637554585153,
+ "grad_norm": 0.01572083681821823,
+ "learning_rate": 0.0006,
+ "loss": 4.2828474044799805,
+ "step": 5678
+ },
+ {
+ "epoch": 78.88034934497817,
+ "grad_norm": 0.013444120064377785,
+ "learning_rate": 0.0006,
+ "loss": 4.183859825134277,
+ "step": 5679
+ },
+ {
+ "epoch": 78.8943231441048,
+ "grad_norm": 0.01591254211962223,
+ "learning_rate": 0.0006,
+ "loss": 4.295256614685059,
+ "step": 5680
+ },
+ {
+ "epoch": 78.90829694323143,
+ "grad_norm": 0.017556684091687202,
+ "learning_rate": 0.0006,
+ "loss": 4.132272720336914,
+ "step": 5681
+ },
+ {
+ "epoch": 78.92227074235808,
+ "grad_norm": 0.018184322863817215,
+ "learning_rate": 0.0006,
+ "loss": 4.251728057861328,
+ "step": 5682
+ },
+ {
+ "epoch": 78.93624454148471,
+ "grad_norm": 0.016207285225391388,
+ "learning_rate": 0.0006,
+ "loss": 4.205544471740723,
+ "step": 5683
+ },
+ {
+ "epoch": 78.95021834061136,
+ "grad_norm": 0.014478376135230064,
+ "learning_rate": 0.0006,
+ "loss": 4.182462215423584,
+ "step": 5684
+ },
+ {
+ "epoch": 78.96419213973799,
+ "grad_norm": 0.014255235902965069,
+ "learning_rate": 0.0006,
+ "loss": 4.253366947174072,
+ "step": 5685
+ },
+ {
+ "epoch": 78.97816593886463,
+ "grad_norm": 0.01524396613240242,
+ "learning_rate": 0.0006,
+ "loss": 4.221786022186279,
+ "step": 5686
+ },
+ {
+ "epoch": 78.99213973799127,
+ "grad_norm": 0.015057693235576153,
+ "learning_rate": 0.0006,
+ "loss": 4.208454608917236,
+ "step": 5687
+ },
+ {
+ "epoch": 79.0,
+ "grad_norm": 0.015478022396564484,
+ "learning_rate": 0.0006,
+ "loss": 4.336383819580078,
+ "step": 5688
+ },
+ {
+ "epoch": 79.0,
+ "eval_loss": 4.641749382019043,
+ "eval_runtime": 56.8167,
+ "eval_samples_per_second": 42.98,
+ "eval_steps_per_second": 1.355,
+ "step": 5688
+ },
+ {
+ "epoch": 79.01397379912663,
+ "grad_norm": 0.014143316075205803,
+ "learning_rate": 0.0006,
+ "loss": 4.260974884033203,
+ "step": 5689
+ },
+ {
+ "epoch": 79.02794759825328,
+ "grad_norm": 0.013120883144438267,
+ "learning_rate": 0.0006,
+ "loss": 4.139886856079102,
+ "step": 5690
+ },
+ {
+ "epoch": 79.04192139737991,
+ "grad_norm": 0.015019661746919155,
+ "learning_rate": 0.0006,
+ "loss": 4.267004013061523,
+ "step": 5691
+ },
+ {
+ "epoch": 79.05589519650655,
+ "grad_norm": 0.014891284517943859,
+ "learning_rate": 0.0006,
+ "loss": 4.205019950866699,
+ "step": 5692
+ },
+ {
+ "epoch": 79.06986899563319,
+ "grad_norm": 0.015527728945016861,
+ "learning_rate": 0.0006,
+ "loss": 4.1513566970825195,
+ "step": 5693
+ },
+ {
+ "epoch": 79.08384279475983,
+ "grad_norm": 0.017527181655168533,
+ "learning_rate": 0.0006,
+ "loss": 4.148828983306885,
+ "step": 5694
+ },
+ {
+ "epoch": 79.09781659388646,
+ "grad_norm": 0.01632206328213215,
+ "learning_rate": 0.0006,
+ "loss": 4.195789813995361,
+ "step": 5695
+ },
+ {
+ "epoch": 79.1117903930131,
+ "grad_norm": 0.01590503193438053,
+ "learning_rate": 0.0006,
+ "loss": 4.199155330657959,
+ "step": 5696
+ },
+ {
+ "epoch": 79.12576419213974,
+ "grad_norm": 0.016489997506141663,
+ "learning_rate": 0.0006,
+ "loss": 4.155154705047607,
+ "step": 5697
+ },
+ {
+ "epoch": 79.13973799126637,
+ "grad_norm": 0.0161368977278471,
+ "learning_rate": 0.0006,
+ "loss": 4.164384841918945,
+ "step": 5698
+ },
+ {
+ "epoch": 79.15371179039302,
+ "grad_norm": 0.018574809655547142,
+ "learning_rate": 0.0006,
+ "loss": 4.1898193359375,
+ "step": 5699
+ },
+ {
+ "epoch": 79.16768558951965,
+ "grad_norm": 0.018987983465194702,
+ "learning_rate": 0.0006,
+ "loss": 4.219719409942627,
+ "step": 5700
+ },
+ {
+ "epoch": 79.18165938864628,
+ "grad_norm": 0.01668606884777546,
+ "learning_rate": 0.0006,
+ "loss": 4.152734756469727,
+ "step": 5701
+ },
+ {
+ "epoch": 79.19563318777293,
+ "grad_norm": 0.014046299271285534,
+ "learning_rate": 0.0006,
+ "loss": 4.152964115142822,
+ "step": 5702
+ },
+ {
+ "epoch": 79.20960698689956,
+ "grad_norm": 0.01628032699227333,
+ "learning_rate": 0.0006,
+ "loss": 4.299479961395264,
+ "step": 5703
+ },
+ {
+ "epoch": 79.2235807860262,
+ "grad_norm": 0.017998795956373215,
+ "learning_rate": 0.0006,
+ "loss": 4.231203079223633,
+ "step": 5704
+ },
+ {
+ "epoch": 79.23755458515284,
+ "grad_norm": 0.014799647964537144,
+ "learning_rate": 0.0006,
+ "loss": 4.3530192375183105,
+ "step": 5705
+ },
+ {
+ "epoch": 79.25152838427948,
+ "grad_norm": 0.014640678651630878,
+ "learning_rate": 0.0006,
+ "loss": 4.287322998046875,
+ "step": 5706
+ },
+ {
+ "epoch": 79.26550218340611,
+ "grad_norm": 0.0172616895288229,
+ "learning_rate": 0.0006,
+ "loss": 4.148009777069092,
+ "step": 5707
+ },
+ {
+ "epoch": 79.27947598253274,
+ "grad_norm": 0.01811790093779564,
+ "learning_rate": 0.0006,
+ "loss": 4.269433975219727,
+ "step": 5708
+ },
+ {
+ "epoch": 79.29344978165939,
+ "grad_norm": 0.01764000579714775,
+ "learning_rate": 0.0006,
+ "loss": 4.292638778686523,
+ "step": 5709
+ },
+ {
+ "epoch": 79.30742358078602,
+ "grad_norm": 0.018030935898423195,
+ "learning_rate": 0.0006,
+ "loss": 4.267889976501465,
+ "step": 5710
+ },
+ {
+ "epoch": 79.32139737991267,
+ "grad_norm": 0.017860891297459602,
+ "learning_rate": 0.0006,
+ "loss": 4.211110591888428,
+ "step": 5711
+ },
+ {
+ "epoch": 79.3353711790393,
+ "grad_norm": 0.017295073717832565,
+ "learning_rate": 0.0006,
+ "loss": 4.224645614624023,
+ "step": 5712
+ },
+ {
+ "epoch": 79.34934497816595,
+ "grad_norm": 0.015891101211309433,
+ "learning_rate": 0.0006,
+ "loss": 4.224797248840332,
+ "step": 5713
+ },
+ {
+ "epoch": 79.36331877729258,
+ "grad_norm": 0.014680417254567146,
+ "learning_rate": 0.0006,
+ "loss": 4.182242393493652,
+ "step": 5714
+ },
+ {
+ "epoch": 79.37729257641921,
+ "grad_norm": 0.016621140763163567,
+ "learning_rate": 0.0006,
+ "loss": 4.186334609985352,
+ "step": 5715
+ },
+ {
+ "epoch": 79.39126637554585,
+ "grad_norm": 0.016207249835133553,
+ "learning_rate": 0.0006,
+ "loss": 4.277074813842773,
+ "step": 5716
+ },
+ {
+ "epoch": 79.40524017467249,
+ "grad_norm": 0.015375478193163872,
+ "learning_rate": 0.0006,
+ "loss": 4.154317855834961,
+ "step": 5717
+ },
+ {
+ "epoch": 79.41921397379913,
+ "grad_norm": 0.013453228399157524,
+ "learning_rate": 0.0006,
+ "loss": 4.265270233154297,
+ "step": 5718
+ },
+ {
+ "epoch": 79.43318777292576,
+ "grad_norm": 0.014312694780528545,
+ "learning_rate": 0.0006,
+ "loss": 4.232220649719238,
+ "step": 5719
+ },
+ {
+ "epoch": 79.44716157205241,
+ "grad_norm": 0.014759554527699947,
+ "learning_rate": 0.0006,
+ "loss": 4.272275924682617,
+ "step": 5720
+ },
+ {
+ "epoch": 79.46113537117904,
+ "grad_norm": 0.015885505825281143,
+ "learning_rate": 0.0006,
+ "loss": 4.2229719161987305,
+ "step": 5721
+ },
+ {
+ "epoch": 79.47510917030567,
+ "grad_norm": 0.016516581177711487,
+ "learning_rate": 0.0006,
+ "loss": 4.200624465942383,
+ "step": 5722
+ },
+ {
+ "epoch": 79.48908296943232,
+ "grad_norm": 0.018254555761814117,
+ "learning_rate": 0.0006,
+ "loss": 4.19648551940918,
+ "step": 5723
+ },
+ {
+ "epoch": 79.50305676855895,
+ "grad_norm": 0.01467090006917715,
+ "learning_rate": 0.0006,
+ "loss": 4.135190486907959,
+ "step": 5724
+ },
+ {
+ "epoch": 79.5170305676856,
+ "grad_norm": 0.013474797829985619,
+ "learning_rate": 0.0006,
+ "loss": 4.311361312866211,
+ "step": 5725
+ },
+ {
+ "epoch": 79.53100436681223,
+ "grad_norm": 0.01742764748632908,
+ "learning_rate": 0.0006,
+ "loss": 4.1845903396606445,
+ "step": 5726
+ },
+ {
+ "epoch": 79.54497816593886,
+ "grad_norm": 0.018786882981657982,
+ "learning_rate": 0.0006,
+ "loss": 4.163942337036133,
+ "step": 5727
+ },
+ {
+ "epoch": 79.5589519650655,
+ "grad_norm": 0.017831219360232353,
+ "learning_rate": 0.0006,
+ "loss": 4.236685276031494,
+ "step": 5728
+ },
+ {
+ "epoch": 79.57292576419214,
+ "grad_norm": 0.016334986314177513,
+ "learning_rate": 0.0006,
+ "loss": 4.255853652954102,
+ "step": 5729
+ },
+ {
+ "epoch": 79.58689956331878,
+ "grad_norm": 0.0138487434014678,
+ "learning_rate": 0.0006,
+ "loss": 4.227874279022217,
+ "step": 5730
+ },
+ {
+ "epoch": 79.60087336244541,
+ "grad_norm": 0.01531760673969984,
+ "learning_rate": 0.0006,
+ "loss": 4.221432685852051,
+ "step": 5731
+ },
+ {
+ "epoch": 79.61484716157206,
+ "grad_norm": 0.01603361964225769,
+ "learning_rate": 0.0006,
+ "loss": 4.278326511383057,
+ "step": 5732
+ },
+ {
+ "epoch": 79.62882096069869,
+ "grad_norm": 0.013149895705282688,
+ "learning_rate": 0.0006,
+ "loss": 4.2474517822265625,
+ "step": 5733
+ },
+ {
+ "epoch": 79.64279475982532,
+ "grad_norm": 0.013428786769509315,
+ "learning_rate": 0.0006,
+ "loss": 4.1701178550720215,
+ "step": 5734
+ },
+ {
+ "epoch": 79.65676855895197,
+ "grad_norm": 0.014305585995316505,
+ "learning_rate": 0.0006,
+ "loss": 4.250778675079346,
+ "step": 5735
+ },
+ {
+ "epoch": 79.6707423580786,
+ "grad_norm": 0.014544566161930561,
+ "learning_rate": 0.0006,
+ "loss": 4.243236541748047,
+ "step": 5736
+ },
+ {
+ "epoch": 79.68471615720524,
+ "grad_norm": 0.013524414040148258,
+ "learning_rate": 0.0006,
+ "loss": 4.255457878112793,
+ "step": 5737
+ },
+ {
+ "epoch": 79.69868995633188,
+ "grad_norm": 0.014750641770660877,
+ "learning_rate": 0.0006,
+ "loss": 4.202816486358643,
+ "step": 5738
+ },
+ {
+ "epoch": 79.71266375545852,
+ "grad_norm": 0.015090773813426495,
+ "learning_rate": 0.0006,
+ "loss": 4.3053178787231445,
+ "step": 5739
+ },
+ {
+ "epoch": 79.72663755458515,
+ "grad_norm": 0.01606169529259205,
+ "learning_rate": 0.0006,
+ "loss": 4.224017143249512,
+ "step": 5740
+ },
+ {
+ "epoch": 79.74061135371178,
+ "grad_norm": 0.016303520649671555,
+ "learning_rate": 0.0006,
+ "loss": 4.190760612487793,
+ "step": 5741
+ },
+ {
+ "epoch": 79.75458515283843,
+ "grad_norm": 0.01465672068297863,
+ "learning_rate": 0.0006,
+ "loss": 4.24506950378418,
+ "step": 5742
+ },
+ {
+ "epoch": 79.76855895196506,
+ "grad_norm": 0.012616301886737347,
+ "learning_rate": 0.0006,
+ "loss": 4.174039840698242,
+ "step": 5743
+ },
+ {
+ "epoch": 79.78253275109171,
+ "grad_norm": 0.01358284242451191,
+ "learning_rate": 0.0006,
+ "loss": 4.236380577087402,
+ "step": 5744
+ },
+ {
+ "epoch": 79.79650655021834,
+ "grad_norm": 0.01557177770882845,
+ "learning_rate": 0.0006,
+ "loss": 4.102962493896484,
+ "step": 5745
+ },
+ {
+ "epoch": 79.81048034934499,
+ "grad_norm": 0.015584317035973072,
+ "learning_rate": 0.0006,
+ "loss": 4.242603302001953,
+ "step": 5746
+ },
+ {
+ "epoch": 79.82445414847162,
+ "grad_norm": 0.012597334571182728,
+ "learning_rate": 0.0006,
+ "loss": 4.202791690826416,
+ "step": 5747
+ },
+ {
+ "epoch": 79.83842794759825,
+ "grad_norm": 0.013722898438572884,
+ "learning_rate": 0.0006,
+ "loss": 4.204328536987305,
+ "step": 5748
+ },
+ {
+ "epoch": 79.8524017467249,
+ "grad_norm": 0.014669225551187992,
+ "learning_rate": 0.0006,
+ "loss": 4.164331912994385,
+ "step": 5749
+ },
+ {
+ "epoch": 79.86637554585153,
+ "grad_norm": 0.015012740157544613,
+ "learning_rate": 0.0006,
+ "loss": 4.260565280914307,
+ "step": 5750
+ },
+ {
+ "epoch": 79.88034934497817,
+ "grad_norm": 0.016026340425014496,
+ "learning_rate": 0.0006,
+ "loss": 4.299723148345947,
+ "step": 5751
+ },
+ {
+ "epoch": 79.8943231441048,
+ "grad_norm": 0.01626049168407917,
+ "learning_rate": 0.0006,
+ "loss": 4.207252502441406,
+ "step": 5752
+ },
+ {
+ "epoch": 79.90829694323143,
+ "grad_norm": 0.015415354631841183,
+ "learning_rate": 0.0006,
+ "loss": 4.151934623718262,
+ "step": 5753
+ },
+ {
+ "epoch": 79.92227074235808,
+ "grad_norm": 0.014158242382109165,
+ "learning_rate": 0.0006,
+ "loss": 4.3461995124816895,
+ "step": 5754
+ },
+ {
+ "epoch": 79.93624454148471,
+ "grad_norm": 0.01585628278553486,
+ "learning_rate": 0.0006,
+ "loss": 4.334761619567871,
+ "step": 5755
+ },
+ {
+ "epoch": 79.95021834061136,
+ "grad_norm": 0.013710759580135345,
+ "learning_rate": 0.0006,
+ "loss": 4.180401802062988,
+ "step": 5756
+ },
+ {
+ "epoch": 79.96419213973799,
+ "grad_norm": 0.014159204438328743,
+ "learning_rate": 0.0006,
+ "loss": 4.322331428527832,
+ "step": 5757
+ },
+ {
+ "epoch": 79.97816593886463,
+ "grad_norm": 0.015153195708990097,
+ "learning_rate": 0.0006,
+ "loss": 4.287478923797607,
+ "step": 5758
+ },
+ {
+ "epoch": 79.99213973799127,
+ "grad_norm": 0.014100227504968643,
+ "learning_rate": 0.0006,
+ "loss": 4.231807708740234,
+ "step": 5759
+ },
+ {
+ "epoch": 80.0,
+ "grad_norm": 0.015248682349920273,
+ "learning_rate": 0.0006,
+ "loss": 4.261106014251709,
+ "step": 5760
+ },
+ {
+ "epoch": 80.0,
+ "eval_loss": 4.610264301300049,
+ "eval_runtime": 56.514,
+ "eval_samples_per_second": 43.211,
+ "eval_steps_per_second": 1.362,
+ "step": 5760
+ },
+ {
+ "epoch": 80.01397379912663,
+ "grad_norm": 0.014730271883308887,
+ "learning_rate": 0.0006,
+ "loss": 4.108556747436523,
+ "step": 5761
+ },
+ {
+ "epoch": 80.02794759825328,
+ "grad_norm": 0.014575716108083725,
+ "learning_rate": 0.0006,
+ "loss": 4.169658660888672,
+ "step": 5762
+ },
+ {
+ "epoch": 80.04192139737991,
+ "grad_norm": 0.014706958085298538,
+ "learning_rate": 0.0006,
+ "loss": 4.252662658691406,
+ "step": 5763
+ },
+ {
+ "epoch": 80.05589519650655,
+ "grad_norm": 0.01561224739998579,
+ "learning_rate": 0.0006,
+ "loss": 4.193892955780029,
+ "step": 5764
+ },
+ {
+ "epoch": 80.06986899563319,
+ "grad_norm": 0.014702324755489826,
+ "learning_rate": 0.0006,
+ "loss": 4.1889238357543945,
+ "step": 5765
+ },
+ {
+ "epoch": 80.08384279475983,
+ "grad_norm": 0.015565741807222366,
+ "learning_rate": 0.0006,
+ "loss": 4.243549823760986,
+ "step": 5766
+ },
+ {
+ "epoch": 80.09781659388646,
+ "grad_norm": 0.01647181063890457,
+ "learning_rate": 0.0006,
+ "loss": 4.275446891784668,
+ "step": 5767
+ },
+ {
+ "epoch": 80.1117903930131,
+ "grad_norm": 0.018177716061472893,
+ "learning_rate": 0.0006,
+ "loss": 4.12040376663208,
+ "step": 5768
+ },
+ {
+ "epoch": 80.12576419213974,
+ "grad_norm": 0.01745191030204296,
+ "learning_rate": 0.0006,
+ "loss": 4.140371322631836,
+ "step": 5769
+ },
+ {
+ "epoch": 80.13973799126637,
+ "grad_norm": 0.01568993367254734,
+ "learning_rate": 0.0006,
+ "loss": 4.258280277252197,
+ "step": 5770
+ },
+ {
+ "epoch": 80.15371179039302,
+ "grad_norm": 0.01636597141623497,
+ "learning_rate": 0.0006,
+ "loss": 4.179052352905273,
+ "step": 5771
+ },
+ {
+ "epoch": 80.16768558951965,
+ "grad_norm": 0.016839100047945976,
+ "learning_rate": 0.0006,
+ "loss": 4.127636432647705,
+ "step": 5772
+ },
+ {
+ "epoch": 80.18165938864628,
+ "grad_norm": 0.015471016988158226,
+ "learning_rate": 0.0006,
+ "loss": 4.181325912475586,
+ "step": 5773
+ },
+ {
+ "epoch": 80.19563318777293,
+ "grad_norm": 0.015836089849472046,
+ "learning_rate": 0.0006,
+ "loss": 4.142168998718262,
+ "step": 5774
+ },
+ {
+ "epoch": 80.20960698689956,
+ "grad_norm": 0.014782954007387161,
+ "learning_rate": 0.0006,
+ "loss": 4.308799743652344,
+ "step": 5775
+ },
+ {
+ "epoch": 80.2235807860262,
+ "grad_norm": 0.014040893875062466,
+ "learning_rate": 0.0006,
+ "loss": 4.216936111450195,
+ "step": 5776
+ },
+ {
+ "epoch": 80.23755458515284,
+ "grad_norm": 0.01420280709862709,
+ "learning_rate": 0.0006,
+ "loss": 4.235230445861816,
+ "step": 5777
+ },
+ {
+ "epoch": 80.25152838427948,
+ "grad_norm": 0.013341007754206657,
+ "learning_rate": 0.0006,
+ "loss": 4.369109153747559,
+ "step": 5778
+ },
+ {
+ "epoch": 80.26550218340611,
+ "grad_norm": 0.014208327978849411,
+ "learning_rate": 0.0006,
+ "loss": 4.219112396240234,
+ "step": 5779
+ },
+ {
+ "epoch": 80.27947598253274,
+ "grad_norm": 0.013825052417814732,
+ "learning_rate": 0.0006,
+ "loss": 4.1983489990234375,
+ "step": 5780
+ },
+ {
+ "epoch": 80.29344978165939,
+ "grad_norm": 0.012916515581309795,
+ "learning_rate": 0.0006,
+ "loss": 4.204582691192627,
+ "step": 5781
+ },
+ {
+ "epoch": 80.30742358078602,
+ "grad_norm": 0.013071243651211262,
+ "learning_rate": 0.0006,
+ "loss": 4.17643928527832,
+ "step": 5782
+ },
+ {
+ "epoch": 80.32139737991267,
+ "grad_norm": 0.015507174655795097,
+ "learning_rate": 0.0006,
+ "loss": 4.172976016998291,
+ "step": 5783
+ },
+ {
+ "epoch": 80.3353711790393,
+ "grad_norm": 0.014967243187129498,
+ "learning_rate": 0.0006,
+ "loss": 4.141168594360352,
+ "step": 5784
+ },
+ {
+ "epoch": 80.34934497816595,
+ "grad_norm": 0.015629924833774567,
+ "learning_rate": 0.0006,
+ "loss": 4.253526210784912,
+ "step": 5785
+ },
+ {
+ "epoch": 80.36331877729258,
+ "grad_norm": 0.01607954502105713,
+ "learning_rate": 0.0006,
+ "loss": 4.249395370483398,
+ "step": 5786
+ },
+ {
+ "epoch": 80.37729257641921,
+ "grad_norm": 0.01500706560909748,
+ "learning_rate": 0.0006,
+ "loss": 4.279012680053711,
+ "step": 5787
+ },
+ {
+ "epoch": 80.39126637554585,
+ "grad_norm": 0.014343900606036186,
+ "learning_rate": 0.0006,
+ "loss": 4.268509864807129,
+ "step": 5788
+ },
+ {
+ "epoch": 80.40524017467249,
+ "grad_norm": 0.014682197012007236,
+ "learning_rate": 0.0006,
+ "loss": 4.206334114074707,
+ "step": 5789
+ },
+ {
+ "epoch": 80.41921397379913,
+ "grad_norm": 0.014989163726568222,
+ "learning_rate": 0.0006,
+ "loss": 4.143601417541504,
+ "step": 5790
+ },
+ {
+ "epoch": 80.43318777292576,
+ "grad_norm": 0.015001608058810234,
+ "learning_rate": 0.0006,
+ "loss": 4.164111137390137,
+ "step": 5791
+ },
+ {
+ "epoch": 80.44716157205241,
+ "grad_norm": 0.015222722664475441,
+ "learning_rate": 0.0006,
+ "loss": 4.2369914054870605,
+ "step": 5792
+ },
+ {
+ "epoch": 80.46113537117904,
+ "grad_norm": 0.014126183465123177,
+ "learning_rate": 0.0006,
+ "loss": 4.2209930419921875,
+ "step": 5793
+ },
+ {
+ "epoch": 80.47510917030567,
+ "grad_norm": 0.013065959326922894,
+ "learning_rate": 0.0006,
+ "loss": 4.231184959411621,
+ "step": 5794
+ },
+ {
+ "epoch": 80.48908296943232,
+ "grad_norm": 0.013833962380886078,
+ "learning_rate": 0.0006,
+ "loss": 4.226434707641602,
+ "step": 5795
+ },
+ {
+ "epoch": 80.50305676855895,
+ "grad_norm": 0.01387125626206398,
+ "learning_rate": 0.0006,
+ "loss": 4.241328716278076,
+ "step": 5796
+ },
+ {
+ "epoch": 80.5170305676856,
+ "grad_norm": 0.014145063236355782,
+ "learning_rate": 0.0006,
+ "loss": 4.2268805503845215,
+ "step": 5797
+ },
+ {
+ "epoch": 80.53100436681223,
+ "grad_norm": 0.012990240007638931,
+ "learning_rate": 0.0006,
+ "loss": 4.1718339920043945,
+ "step": 5798
+ },
+ {
+ "epoch": 80.54497816593886,
+ "grad_norm": 0.014185411855578423,
+ "learning_rate": 0.0006,
+ "loss": 4.232397556304932,
+ "step": 5799
+ },
+ {
+ "epoch": 80.5589519650655,
+ "grad_norm": 0.014260523952543736,
+ "learning_rate": 0.0006,
+ "loss": 4.282734394073486,
+ "step": 5800
+ },
+ {
+ "epoch": 80.57292576419214,
+ "grad_norm": 0.012542120181024075,
+ "learning_rate": 0.0006,
+ "loss": 4.134551525115967,
+ "step": 5801
+ },
+ {
+ "epoch": 80.58689956331878,
+ "grad_norm": 0.013078657910227776,
+ "learning_rate": 0.0006,
+ "loss": 4.229815483093262,
+ "step": 5802
+ },
+ {
+ "epoch": 80.60087336244541,
+ "grad_norm": 0.012821056880056858,
+ "learning_rate": 0.0006,
+ "loss": 4.208278179168701,
+ "step": 5803
+ },
+ {
+ "epoch": 80.61484716157206,
+ "grad_norm": 0.012964180670678616,
+ "learning_rate": 0.0006,
+ "loss": 4.188477039337158,
+ "step": 5804
+ },
+ {
+ "epoch": 80.62882096069869,
+ "grad_norm": 0.01631746254861355,
+ "learning_rate": 0.0006,
+ "loss": 4.3214497566223145,
+ "step": 5805
+ },
+ {
+ "epoch": 80.64279475982532,
+ "grad_norm": 0.01726640947163105,
+ "learning_rate": 0.0006,
+ "loss": 4.279359817504883,
+ "step": 5806
+ },
+ {
+ "epoch": 80.65676855895197,
+ "grad_norm": 0.018778258934617043,
+ "learning_rate": 0.0006,
+ "loss": 4.294047832489014,
+ "step": 5807
+ },
+ {
+ "epoch": 80.6707423580786,
+ "grad_norm": 0.01796831749379635,
+ "learning_rate": 0.0006,
+ "loss": 4.314162254333496,
+ "step": 5808
+ },
+ {
+ "epoch": 80.68471615720524,
+ "grad_norm": 0.016495538875460625,
+ "learning_rate": 0.0006,
+ "loss": 4.273871421813965,
+ "step": 5809
+ },
+ {
+ "epoch": 80.69868995633188,
+ "grad_norm": 0.013590306043624878,
+ "learning_rate": 0.0006,
+ "loss": 4.14961051940918,
+ "step": 5810
+ },
+ {
+ "epoch": 80.71266375545852,
+ "grad_norm": 0.014670426957309246,
+ "learning_rate": 0.0006,
+ "loss": 4.211853981018066,
+ "step": 5811
+ },
+ {
+ "epoch": 80.72663755458515,
+ "grad_norm": 0.01641341857612133,
+ "learning_rate": 0.0006,
+ "loss": 4.211948871612549,
+ "step": 5812
+ },
+ {
+ "epoch": 80.74061135371178,
+ "grad_norm": 0.016365470364689827,
+ "learning_rate": 0.0006,
+ "loss": 4.200774192810059,
+ "step": 5813
+ },
+ {
+ "epoch": 80.75458515283843,
+ "grad_norm": 0.01613970287144184,
+ "learning_rate": 0.0006,
+ "loss": 4.288322448730469,
+ "step": 5814
+ },
+ {
+ "epoch": 80.76855895196506,
+ "grad_norm": 0.016742777079343796,
+ "learning_rate": 0.0006,
+ "loss": 4.23801326751709,
+ "step": 5815
+ },
+ {
+ "epoch": 80.78253275109171,
+ "grad_norm": 0.015351069159805775,
+ "learning_rate": 0.0006,
+ "loss": 4.263832092285156,
+ "step": 5816
+ },
+ {
+ "epoch": 80.79650655021834,
+ "grad_norm": 0.013435336761176586,
+ "learning_rate": 0.0006,
+ "loss": 4.160857200622559,
+ "step": 5817
+ },
+ {
+ "epoch": 80.81048034934499,
+ "grad_norm": 0.01474886853247881,
+ "learning_rate": 0.0006,
+ "loss": 4.216025352478027,
+ "step": 5818
+ },
+ {
+ "epoch": 80.82445414847162,
+ "grad_norm": 0.01586935855448246,
+ "learning_rate": 0.0006,
+ "loss": 4.256402969360352,
+ "step": 5819
+ },
+ {
+ "epoch": 80.83842794759825,
+ "grad_norm": 0.014596572145819664,
+ "learning_rate": 0.0006,
+ "loss": 4.142584800720215,
+ "step": 5820
+ },
+ {
+ "epoch": 80.8524017467249,
+ "grad_norm": 0.01501593366265297,
+ "learning_rate": 0.0006,
+ "loss": 4.250462055206299,
+ "step": 5821
+ },
+ {
+ "epoch": 80.86637554585153,
+ "grad_norm": 0.014773511327803135,
+ "learning_rate": 0.0006,
+ "loss": 4.253305912017822,
+ "step": 5822
+ },
+ {
+ "epoch": 80.88034934497817,
+ "grad_norm": 0.013246326707303524,
+ "learning_rate": 0.0006,
+ "loss": 4.170881271362305,
+ "step": 5823
+ },
+ {
+ "epoch": 80.8943231441048,
+ "grad_norm": 0.013013756833970547,
+ "learning_rate": 0.0006,
+ "loss": 4.264721870422363,
+ "step": 5824
+ },
+ {
+ "epoch": 80.90829694323143,
+ "grad_norm": 0.013155903667211533,
+ "learning_rate": 0.0006,
+ "loss": 4.274667739868164,
+ "step": 5825
+ },
+ {
+ "epoch": 80.92227074235808,
+ "grad_norm": 0.01448952779173851,
+ "learning_rate": 0.0006,
+ "loss": 4.24777889251709,
+ "step": 5826
+ },
+ {
+ "epoch": 80.93624454148471,
+ "grad_norm": 0.01415231078863144,
+ "learning_rate": 0.0006,
+ "loss": 4.22014856338501,
+ "step": 5827
+ },
+ {
+ "epoch": 80.95021834061136,
+ "grad_norm": 0.014796273782849312,
+ "learning_rate": 0.0006,
+ "loss": 4.22036600112915,
+ "step": 5828
+ },
+ {
+ "epoch": 80.96419213973799,
+ "grad_norm": 0.013005326502025127,
+ "learning_rate": 0.0006,
+ "loss": 4.276123046875,
+ "step": 5829
+ },
+ {
+ "epoch": 80.97816593886463,
+ "grad_norm": 0.012851177714765072,
+ "learning_rate": 0.0006,
+ "loss": 4.248575687408447,
+ "step": 5830
+ },
+ {
+ "epoch": 80.99213973799127,
+ "grad_norm": 0.013740002177655697,
+ "learning_rate": 0.0006,
+ "loss": 4.051140308380127,
+ "step": 5831
+ },
+ {
+ "epoch": 81.0,
+ "grad_norm": 0.015296104364097118,
+ "learning_rate": 0.0006,
+ "loss": 4.251155853271484,
+ "step": 5832
+ },
+ {
+ "epoch": 81.0,
+ "eval_loss": 4.594127655029297,
+ "eval_runtime": 56.0959,
+ "eval_samples_per_second": 43.533,
+ "eval_steps_per_second": 1.373,
+ "step": 5832
+ },
+ {
+ "epoch": 81.01397379912663,
+ "grad_norm": 0.015804558992385864,
+ "learning_rate": 0.0006,
+ "loss": 4.152894020080566,
+ "step": 5833
+ },
+ {
+ "epoch": 81.02794759825328,
+ "grad_norm": 0.017000854015350342,
+ "learning_rate": 0.0006,
+ "loss": 4.266067981719971,
+ "step": 5834
+ },
+ {
+ "epoch": 81.04192139737991,
+ "grad_norm": 0.016351008787751198,
+ "learning_rate": 0.0006,
+ "loss": 4.192703723907471,
+ "step": 5835
+ },
+ {
+ "epoch": 81.05589519650655,
+ "grad_norm": 0.014703895896673203,
+ "learning_rate": 0.0006,
+ "loss": 4.23554801940918,
+ "step": 5836
+ },
+ {
+ "epoch": 81.06986899563319,
+ "grad_norm": 0.01619596965610981,
+ "learning_rate": 0.0006,
+ "loss": 4.233000755310059,
+ "step": 5837
+ },
+ {
+ "epoch": 81.08384279475983,
+ "grad_norm": 0.01815493032336235,
+ "learning_rate": 0.0006,
+ "loss": 4.200490951538086,
+ "step": 5838
+ },
+ {
+ "epoch": 81.09781659388646,
+ "grad_norm": 0.018305106088519096,
+ "learning_rate": 0.0006,
+ "loss": 4.154242515563965,
+ "step": 5839
+ },
+ {
+ "epoch": 81.1117903930131,
+ "grad_norm": 0.0200856514275074,
+ "learning_rate": 0.0006,
+ "loss": 4.25820779800415,
+ "step": 5840
+ },
+ {
+ "epoch": 81.12576419213974,
+ "grad_norm": 0.02101440541446209,
+ "learning_rate": 0.0006,
+ "loss": 4.168209075927734,
+ "step": 5841
+ },
+ {
+ "epoch": 81.13973799126637,
+ "grad_norm": 0.022829681634902954,
+ "learning_rate": 0.0006,
+ "loss": 4.25560998916626,
+ "step": 5842
+ },
+ {
+ "epoch": 81.15371179039302,
+ "grad_norm": 0.0195082426071167,
+ "learning_rate": 0.0006,
+ "loss": 4.2035627365112305,
+ "step": 5843
+ },
+ {
+ "epoch": 81.16768558951965,
+ "grad_norm": 0.01716248318552971,
+ "learning_rate": 0.0006,
+ "loss": 4.11322021484375,
+ "step": 5844
+ },
+ {
+ "epoch": 81.18165938864628,
+ "grad_norm": 0.02034664712846279,
+ "learning_rate": 0.0006,
+ "loss": 4.179389953613281,
+ "step": 5845
+ },
+ {
+ "epoch": 81.19563318777293,
+ "grad_norm": 0.023578710854053497,
+ "learning_rate": 0.0006,
+ "loss": 4.265695571899414,
+ "step": 5846
+ },
+ {
+ "epoch": 81.20960698689956,
+ "grad_norm": 0.019244182854890823,
+ "learning_rate": 0.0006,
+ "loss": 4.292043209075928,
+ "step": 5847
+ },
+ {
+ "epoch": 81.2235807860262,
+ "grad_norm": 0.01577291637659073,
+ "learning_rate": 0.0006,
+ "loss": 4.250776290893555,
+ "step": 5848
+ },
+ {
+ "epoch": 81.23755458515284,
+ "grad_norm": 0.01617090217769146,
+ "learning_rate": 0.0006,
+ "loss": 4.193098068237305,
+ "step": 5849
+ },
+ {
+ "epoch": 81.25152838427948,
+ "grad_norm": 0.016090184450149536,
+ "learning_rate": 0.0006,
+ "loss": 4.112357139587402,
+ "step": 5850
+ },
+ {
+ "epoch": 81.26550218340611,
+ "grad_norm": 0.01441365759819746,
+ "learning_rate": 0.0006,
+ "loss": 4.114147186279297,
+ "step": 5851
+ },
+ {
+ "epoch": 81.27947598253274,
+ "grad_norm": 0.014122162945568562,
+ "learning_rate": 0.0006,
+ "loss": 4.276175022125244,
+ "step": 5852
+ },
+ {
+ "epoch": 81.29344978165939,
+ "grad_norm": 0.014859487302601337,
+ "learning_rate": 0.0006,
+ "loss": 4.094699859619141,
+ "step": 5853
+ },
+ {
+ "epoch": 81.30742358078602,
+ "grad_norm": 0.013624267652630806,
+ "learning_rate": 0.0006,
+ "loss": 4.2648773193359375,
+ "step": 5854
+ },
+ {
+ "epoch": 81.32139737991267,
+ "grad_norm": 0.01417913381010294,
+ "learning_rate": 0.0006,
+ "loss": 4.228567123413086,
+ "step": 5855
+ },
+ {
+ "epoch": 81.3353711790393,
+ "grad_norm": 0.014560805633664131,
+ "learning_rate": 0.0006,
+ "loss": 4.148930549621582,
+ "step": 5856
+ },
+ {
+ "epoch": 81.34934497816595,
+ "grad_norm": 0.013662154786288738,
+ "learning_rate": 0.0006,
+ "loss": 4.192107677459717,
+ "step": 5857
+ },
+ {
+ "epoch": 81.36331877729258,
+ "grad_norm": 0.013434980995953083,
+ "learning_rate": 0.0006,
+ "loss": 4.162052154541016,
+ "step": 5858
+ },
+ {
+ "epoch": 81.37729257641921,
+ "grad_norm": 0.01300742756575346,
+ "learning_rate": 0.0006,
+ "loss": 4.262385368347168,
+ "step": 5859
+ },
+ {
+ "epoch": 81.39126637554585,
+ "grad_norm": 0.012544351629912853,
+ "learning_rate": 0.0006,
+ "loss": 4.215371131896973,
+ "step": 5860
+ },
+ {
+ "epoch": 81.40524017467249,
+ "grad_norm": 0.013663645833730698,
+ "learning_rate": 0.0006,
+ "loss": 4.247312545776367,
+ "step": 5861
+ },
+ {
+ "epoch": 81.41921397379913,
+ "grad_norm": 0.013127296231687069,
+ "learning_rate": 0.0006,
+ "loss": 4.087685585021973,
+ "step": 5862
+ },
+ {
+ "epoch": 81.43318777292576,
+ "grad_norm": 0.014081938192248344,
+ "learning_rate": 0.0006,
+ "loss": 4.272648811340332,
+ "step": 5863
+ },
+ {
+ "epoch": 81.44716157205241,
+ "grad_norm": 0.015470641665160656,
+ "learning_rate": 0.0006,
+ "loss": 4.278578758239746,
+ "step": 5864
+ },
+ {
+ "epoch": 81.46113537117904,
+ "grad_norm": 0.016524532809853554,
+ "learning_rate": 0.0006,
+ "loss": 4.118268013000488,
+ "step": 5865
+ },
+ {
+ "epoch": 81.47510917030567,
+ "grad_norm": 0.016242722049355507,
+ "learning_rate": 0.0006,
+ "loss": 4.182405471801758,
+ "step": 5866
+ },
+ {
+ "epoch": 81.48908296943232,
+ "grad_norm": 0.01557017583400011,
+ "learning_rate": 0.0006,
+ "loss": 4.229862213134766,
+ "step": 5867
+ },
+ {
+ "epoch": 81.50305676855895,
+ "grad_norm": 0.01562870666384697,
+ "learning_rate": 0.0006,
+ "loss": 4.27754020690918,
+ "step": 5868
+ },
+ {
+ "epoch": 81.5170305676856,
+ "grad_norm": 0.018606074154376984,
+ "learning_rate": 0.0006,
+ "loss": 4.148681640625,
+ "step": 5869
+ },
+ {
+ "epoch": 81.53100436681223,
+ "grad_norm": 0.017740730196237564,
+ "learning_rate": 0.0006,
+ "loss": 4.321932792663574,
+ "step": 5870
+ },
+ {
+ "epoch": 81.54497816593886,
+ "grad_norm": 0.014517677947878838,
+ "learning_rate": 0.0006,
+ "loss": 4.22862434387207,
+ "step": 5871
+ },
+ {
+ "epoch": 81.5589519650655,
+ "grad_norm": 0.01389714702963829,
+ "learning_rate": 0.0006,
+ "loss": 4.198000907897949,
+ "step": 5872
+ },
+ {
+ "epoch": 81.57292576419214,
+ "grad_norm": 0.015348682180047035,
+ "learning_rate": 0.0006,
+ "loss": 4.293485164642334,
+ "step": 5873
+ },
+ {
+ "epoch": 81.58689956331878,
+ "grad_norm": 0.015965791419148445,
+ "learning_rate": 0.0006,
+ "loss": 4.23481559753418,
+ "step": 5874
+ },
+ {
+ "epoch": 81.60087336244541,
+ "grad_norm": 0.015662726014852524,
+ "learning_rate": 0.0006,
+ "loss": 4.158148765563965,
+ "step": 5875
+ },
+ {
+ "epoch": 81.61484716157206,
+ "grad_norm": 0.014402080327272415,
+ "learning_rate": 0.0006,
+ "loss": 4.132723808288574,
+ "step": 5876
+ },
+ {
+ "epoch": 81.62882096069869,
+ "grad_norm": 0.016846125945448875,
+ "learning_rate": 0.0006,
+ "loss": 4.1654462814331055,
+ "step": 5877
+ },
+ {
+ "epoch": 81.64279475982532,
+ "grad_norm": 0.0196097232401371,
+ "learning_rate": 0.0006,
+ "loss": 4.228450298309326,
+ "step": 5878
+ },
+ {
+ "epoch": 81.65676855895197,
+ "grad_norm": 0.016005294397473335,
+ "learning_rate": 0.0006,
+ "loss": 4.230222225189209,
+ "step": 5879
+ },
+ {
+ "epoch": 81.6707423580786,
+ "grad_norm": 0.014685059897601604,
+ "learning_rate": 0.0006,
+ "loss": 4.1367387771606445,
+ "step": 5880
+ },
+ {
+ "epoch": 81.68471615720524,
+ "grad_norm": 0.01647144928574562,
+ "learning_rate": 0.0006,
+ "loss": 4.096675395965576,
+ "step": 5881
+ },
+ {
+ "epoch": 81.69868995633188,
+ "grad_norm": 0.015978610143065453,
+ "learning_rate": 0.0006,
+ "loss": 4.232292175292969,
+ "step": 5882
+ },
+ {
+ "epoch": 81.71266375545852,
+ "grad_norm": 0.01510288193821907,
+ "learning_rate": 0.0006,
+ "loss": 4.274591445922852,
+ "step": 5883
+ },
+ {
+ "epoch": 81.72663755458515,
+ "grad_norm": 0.014942476525902748,
+ "learning_rate": 0.0006,
+ "loss": 4.161482810974121,
+ "step": 5884
+ },
+ {
+ "epoch": 81.74061135371178,
+ "grad_norm": 0.014595299027860165,
+ "learning_rate": 0.0006,
+ "loss": 4.204291343688965,
+ "step": 5885
+ },
+ {
+ "epoch": 81.75458515283843,
+ "grad_norm": 0.014952429570257664,
+ "learning_rate": 0.0006,
+ "loss": 4.242499351501465,
+ "step": 5886
+ },
+ {
+ "epoch": 81.76855895196506,
+ "grad_norm": 0.014761554077267647,
+ "learning_rate": 0.0006,
+ "loss": 4.20393180847168,
+ "step": 5887
+ },
+ {
+ "epoch": 81.78253275109171,
+ "grad_norm": 0.014749647118151188,
+ "learning_rate": 0.0006,
+ "loss": 4.227565765380859,
+ "step": 5888
+ },
+ {
+ "epoch": 81.79650655021834,
+ "grad_norm": 0.014384493231773376,
+ "learning_rate": 0.0006,
+ "loss": 4.282228946685791,
+ "step": 5889
+ },
+ {
+ "epoch": 81.81048034934499,
+ "grad_norm": 0.01260330155491829,
+ "learning_rate": 0.0006,
+ "loss": 4.2882866859436035,
+ "step": 5890
+ },
+ {
+ "epoch": 81.82445414847162,
+ "grad_norm": 0.014055311679840088,
+ "learning_rate": 0.0006,
+ "loss": 4.22351598739624,
+ "step": 5891
+ },
+ {
+ "epoch": 81.83842794759825,
+ "grad_norm": 0.013674065470695496,
+ "learning_rate": 0.0006,
+ "loss": 4.240716934204102,
+ "step": 5892
+ },
+ {
+ "epoch": 81.8524017467249,
+ "grad_norm": 0.01225196197628975,
+ "learning_rate": 0.0006,
+ "loss": 4.2958598136901855,
+ "step": 5893
+ },
+ {
+ "epoch": 81.86637554585153,
+ "grad_norm": 0.013538680970668793,
+ "learning_rate": 0.0006,
+ "loss": 4.24342679977417,
+ "step": 5894
+ },
+ {
+ "epoch": 81.88034934497817,
+ "grad_norm": 0.013151394203305244,
+ "learning_rate": 0.0006,
+ "loss": 4.081999778747559,
+ "step": 5895
+ },
+ {
+ "epoch": 81.8943231441048,
+ "grad_norm": 0.014260678552091122,
+ "learning_rate": 0.0006,
+ "loss": 4.156423568725586,
+ "step": 5896
+ },
+ {
+ "epoch": 81.90829694323143,
+ "grad_norm": 0.01329776830971241,
+ "learning_rate": 0.0006,
+ "loss": 4.1713995933532715,
+ "step": 5897
+ },
+ {
+ "epoch": 81.92227074235808,
+ "grad_norm": 0.014623276889324188,
+ "learning_rate": 0.0006,
+ "loss": 4.244891166687012,
+ "step": 5898
+ },
+ {
+ "epoch": 81.93624454148471,
+ "grad_norm": 0.015760324895381927,
+ "learning_rate": 0.0006,
+ "loss": 4.076528549194336,
+ "step": 5899
+ },
+ {
+ "epoch": 81.95021834061136,
+ "grad_norm": 0.014331766404211521,
+ "learning_rate": 0.0006,
+ "loss": 4.283037185668945,
+ "step": 5900
+ },
+ {
+ "epoch": 81.96419213973799,
+ "grad_norm": 0.012383842840790749,
+ "learning_rate": 0.0006,
+ "loss": 4.117308616638184,
+ "step": 5901
+ },
+ {
+ "epoch": 81.97816593886463,
+ "grad_norm": 0.01250660140067339,
+ "learning_rate": 0.0006,
+ "loss": 4.241250991821289,
+ "step": 5902
+ },
+ {
+ "epoch": 81.99213973799127,
+ "grad_norm": 0.012276900932192802,
+ "learning_rate": 0.0006,
+ "loss": 4.141091823577881,
+ "step": 5903
+ },
+ {
+ "epoch": 82.0,
+ "grad_norm": 0.01588737778365612,
+ "learning_rate": 0.0006,
+ "loss": 4.065532684326172,
+ "step": 5904
+ },
+ {
+ "epoch": 82.0,
+ "eval_loss": 4.612304210662842,
+ "eval_runtime": 57.5411,
+ "eval_samples_per_second": 42.439,
+ "eval_steps_per_second": 1.338,
+ "step": 5904
+ },
+ {
+ "epoch": 82.01397379912663,
+ "grad_norm": 0.017433544620871544,
+ "learning_rate": 0.0006,
+ "loss": 4.202497482299805,
+ "step": 5905
+ },
+ {
+ "epoch": 82.02794759825328,
+ "grad_norm": 0.017260415479540825,
+ "learning_rate": 0.0006,
+ "loss": 4.099949359893799,
+ "step": 5906
+ },
+ {
+ "epoch": 82.04192139737991,
+ "grad_norm": 0.01577908545732498,
+ "learning_rate": 0.0006,
+ "loss": 4.17806339263916,
+ "step": 5907
+ },
+ {
+ "epoch": 82.05589519650655,
+ "grad_norm": 0.016093818470835686,
+ "learning_rate": 0.0006,
+ "loss": 4.219491004943848,
+ "step": 5908
+ },
+ {
+ "epoch": 82.06986899563319,
+ "grad_norm": 0.015426302328705788,
+ "learning_rate": 0.0006,
+ "loss": 4.207982063293457,
+ "step": 5909
+ },
+ {
+ "epoch": 82.08384279475983,
+ "grad_norm": 0.01605442352592945,
+ "learning_rate": 0.0006,
+ "loss": 4.265942573547363,
+ "step": 5910
+ },
+ {
+ "epoch": 82.09781659388646,
+ "grad_norm": 0.01787225715816021,
+ "learning_rate": 0.0006,
+ "loss": 4.151535511016846,
+ "step": 5911
+ },
+ {
+ "epoch": 82.1117903930131,
+ "grad_norm": 0.01750164106488228,
+ "learning_rate": 0.0006,
+ "loss": 4.133655071258545,
+ "step": 5912
+ },
+ {
+ "epoch": 82.12576419213974,
+ "grad_norm": 0.01602957211434841,
+ "learning_rate": 0.0006,
+ "loss": 4.176972389221191,
+ "step": 5913
+ },
+ {
+ "epoch": 82.13973799126637,
+ "grad_norm": 0.01553812064230442,
+ "learning_rate": 0.0006,
+ "loss": 4.125692367553711,
+ "step": 5914
+ },
+ {
+ "epoch": 82.15371179039302,
+ "grad_norm": 0.013978508301079273,
+ "learning_rate": 0.0006,
+ "loss": 4.207271575927734,
+ "step": 5915
+ },
+ {
+ "epoch": 82.16768558951965,
+ "grad_norm": 0.013561118394136429,
+ "learning_rate": 0.0006,
+ "loss": 4.238930702209473,
+ "step": 5916
+ },
+ {
+ "epoch": 82.18165938864628,
+ "grad_norm": 0.013285507448017597,
+ "learning_rate": 0.0006,
+ "loss": 4.129464626312256,
+ "step": 5917
+ },
+ {
+ "epoch": 82.19563318777293,
+ "grad_norm": 0.013903839513659477,
+ "learning_rate": 0.0006,
+ "loss": 4.214109420776367,
+ "step": 5918
+ },
+ {
+ "epoch": 82.20960698689956,
+ "grad_norm": 0.014426837675273418,
+ "learning_rate": 0.0006,
+ "loss": 4.143196105957031,
+ "step": 5919
+ },
+ {
+ "epoch": 82.2235807860262,
+ "grad_norm": 0.015347165986895561,
+ "learning_rate": 0.0006,
+ "loss": 4.235439300537109,
+ "step": 5920
+ },
+ {
+ "epoch": 82.23755458515284,
+ "grad_norm": 0.014266039244830608,
+ "learning_rate": 0.0006,
+ "loss": 3.991617202758789,
+ "step": 5921
+ },
+ {
+ "epoch": 82.25152838427948,
+ "grad_norm": 0.013101593591272831,
+ "learning_rate": 0.0006,
+ "loss": 4.266017913818359,
+ "step": 5922
+ },
+ {
+ "epoch": 82.26550218340611,
+ "grad_norm": 0.013741881586611271,
+ "learning_rate": 0.0006,
+ "loss": 4.224879741668701,
+ "step": 5923
+ },
+ {
+ "epoch": 82.27947598253274,
+ "grad_norm": 0.014206945896148682,
+ "learning_rate": 0.0006,
+ "loss": 4.102272987365723,
+ "step": 5924
+ },
+ {
+ "epoch": 82.29344978165939,
+ "grad_norm": 0.013798723928630352,
+ "learning_rate": 0.0006,
+ "loss": 4.30418586730957,
+ "step": 5925
+ },
+ {
+ "epoch": 82.30742358078602,
+ "grad_norm": 0.015470122918486595,
+ "learning_rate": 0.0006,
+ "loss": 4.2513322830200195,
+ "step": 5926
+ },
+ {
+ "epoch": 82.32139737991267,
+ "grad_norm": 0.01655442826449871,
+ "learning_rate": 0.0006,
+ "loss": 4.223002910614014,
+ "step": 5927
+ },
+ {
+ "epoch": 82.3353711790393,
+ "grad_norm": 0.017170004546642303,
+ "learning_rate": 0.0006,
+ "loss": 4.253904819488525,
+ "step": 5928
+ },
+ {
+ "epoch": 82.34934497816595,
+ "grad_norm": 0.017595812678337097,
+ "learning_rate": 0.0006,
+ "loss": 4.184927940368652,
+ "step": 5929
+ },
+ {
+ "epoch": 82.36331877729258,
+ "grad_norm": 0.015073256567120552,
+ "learning_rate": 0.0006,
+ "loss": 4.202239036560059,
+ "step": 5930
+ },
+ {
+ "epoch": 82.37729257641921,
+ "grad_norm": 0.014310941100120544,
+ "learning_rate": 0.0006,
+ "loss": 4.124934673309326,
+ "step": 5931
+ },
+ {
+ "epoch": 82.39126637554585,
+ "grad_norm": 0.013671821914613247,
+ "learning_rate": 0.0006,
+ "loss": 4.138919353485107,
+ "step": 5932
+ },
+ {
+ "epoch": 82.40524017467249,
+ "grad_norm": 0.014103314839303493,
+ "learning_rate": 0.0006,
+ "loss": 4.211289405822754,
+ "step": 5933
+ },
+ {
+ "epoch": 82.41921397379913,
+ "grad_norm": 0.014272668398916721,
+ "learning_rate": 0.0006,
+ "loss": 4.06290340423584,
+ "step": 5934
+ },
+ {
+ "epoch": 82.43318777292576,
+ "grad_norm": 0.01387016475200653,
+ "learning_rate": 0.0006,
+ "loss": 4.101334571838379,
+ "step": 5935
+ },
+ {
+ "epoch": 82.44716157205241,
+ "grad_norm": 0.013318683952093124,
+ "learning_rate": 0.0006,
+ "loss": 4.206409931182861,
+ "step": 5936
+ },
+ {
+ "epoch": 82.46113537117904,
+ "grad_norm": 0.013781268149614334,
+ "learning_rate": 0.0006,
+ "loss": 4.189474105834961,
+ "step": 5937
+ },
+ {
+ "epoch": 82.47510917030567,
+ "grad_norm": 0.013195851817727089,
+ "learning_rate": 0.0006,
+ "loss": 4.276607036590576,
+ "step": 5938
+ },
+ {
+ "epoch": 82.48908296943232,
+ "grad_norm": 0.014358163811266422,
+ "learning_rate": 0.0006,
+ "loss": 4.203262805938721,
+ "step": 5939
+ },
+ {
+ "epoch": 82.50305676855895,
+ "grad_norm": 0.01403636671602726,
+ "learning_rate": 0.0006,
+ "loss": 4.079258441925049,
+ "step": 5940
+ },
+ {
+ "epoch": 82.5170305676856,
+ "grad_norm": 0.013272535055875778,
+ "learning_rate": 0.0006,
+ "loss": 4.19197940826416,
+ "step": 5941
+ },
+ {
+ "epoch": 82.53100436681223,
+ "grad_norm": 0.013787264935672283,
+ "learning_rate": 0.0006,
+ "loss": 4.12995719909668,
+ "step": 5942
+ },
+ {
+ "epoch": 82.54497816593886,
+ "grad_norm": 0.015020057559013367,
+ "learning_rate": 0.0006,
+ "loss": 4.148890495300293,
+ "step": 5943
+ },
+ {
+ "epoch": 82.5589519650655,
+ "grad_norm": 0.014199157245457172,
+ "learning_rate": 0.0006,
+ "loss": 4.201516151428223,
+ "step": 5944
+ },
+ {
+ "epoch": 82.57292576419214,
+ "grad_norm": 0.013827123679220676,
+ "learning_rate": 0.0006,
+ "loss": 4.240894317626953,
+ "step": 5945
+ },
+ {
+ "epoch": 82.58689956331878,
+ "grad_norm": 0.014767560176551342,
+ "learning_rate": 0.0006,
+ "loss": 4.173320770263672,
+ "step": 5946
+ },
+ {
+ "epoch": 82.60087336244541,
+ "grad_norm": 0.01612078957259655,
+ "learning_rate": 0.0006,
+ "loss": 4.266448020935059,
+ "step": 5947
+ },
+ {
+ "epoch": 82.61484716157206,
+ "grad_norm": 0.016706543043255806,
+ "learning_rate": 0.0006,
+ "loss": 4.229398250579834,
+ "step": 5948
+ },
+ {
+ "epoch": 82.62882096069869,
+ "grad_norm": 0.015630796551704407,
+ "learning_rate": 0.0006,
+ "loss": 4.181282997131348,
+ "step": 5949
+ },
+ {
+ "epoch": 82.64279475982532,
+ "grad_norm": 0.013103844597935677,
+ "learning_rate": 0.0006,
+ "loss": 4.309617042541504,
+ "step": 5950
+ },
+ {
+ "epoch": 82.65676855895197,
+ "grad_norm": 0.016136541962623596,
+ "learning_rate": 0.0006,
+ "loss": 4.224159240722656,
+ "step": 5951
+ },
+ {
+ "epoch": 82.6707423580786,
+ "grad_norm": 0.016160814091563225,
+ "learning_rate": 0.0006,
+ "loss": 4.267970085144043,
+ "step": 5952
+ },
+ {
+ "epoch": 82.68471615720524,
+ "grad_norm": 0.01672680303454399,
+ "learning_rate": 0.0006,
+ "loss": 4.230506896972656,
+ "step": 5953
+ },
+ {
+ "epoch": 82.69868995633188,
+ "grad_norm": 0.0159586351364851,
+ "learning_rate": 0.0006,
+ "loss": 4.159542083740234,
+ "step": 5954
+ },
+ {
+ "epoch": 82.71266375545852,
+ "grad_norm": 0.01461219135671854,
+ "learning_rate": 0.0006,
+ "loss": 4.27863073348999,
+ "step": 5955
+ },
+ {
+ "epoch": 82.72663755458515,
+ "grad_norm": 0.015088613145053387,
+ "learning_rate": 0.0006,
+ "loss": 4.272418975830078,
+ "step": 5956
+ },
+ {
+ "epoch": 82.74061135371178,
+ "grad_norm": 0.01594499684870243,
+ "learning_rate": 0.0006,
+ "loss": 4.271326065063477,
+ "step": 5957
+ },
+ {
+ "epoch": 82.75458515283843,
+ "grad_norm": 0.016648447141051292,
+ "learning_rate": 0.0006,
+ "loss": 4.141478538513184,
+ "step": 5958
+ },
+ {
+ "epoch": 82.76855895196506,
+ "grad_norm": 0.01540316641330719,
+ "learning_rate": 0.0006,
+ "loss": 4.111814498901367,
+ "step": 5959
+ },
+ {
+ "epoch": 82.78253275109171,
+ "grad_norm": 0.01634364202618599,
+ "learning_rate": 0.0006,
+ "loss": 4.3113112449646,
+ "step": 5960
+ },
+ {
+ "epoch": 82.79650655021834,
+ "grad_norm": 0.014768977649509907,
+ "learning_rate": 0.0006,
+ "loss": 4.184217929840088,
+ "step": 5961
+ },
+ {
+ "epoch": 82.81048034934499,
+ "grad_norm": 0.013453755527734756,
+ "learning_rate": 0.0006,
+ "loss": 4.247729301452637,
+ "step": 5962
+ },
+ {
+ "epoch": 82.82445414847162,
+ "grad_norm": 0.014731907285749912,
+ "learning_rate": 0.0006,
+ "loss": 4.352839469909668,
+ "step": 5963
+ },
+ {
+ "epoch": 82.83842794759825,
+ "grad_norm": 0.016047345474362373,
+ "learning_rate": 0.0006,
+ "loss": 4.257283687591553,
+ "step": 5964
+ },
+ {
+ "epoch": 82.8524017467249,
+ "grad_norm": 0.017141170799732208,
+ "learning_rate": 0.0006,
+ "loss": 4.22850227355957,
+ "step": 5965
+ },
+ {
+ "epoch": 82.86637554585153,
+ "grad_norm": 0.01370396837592125,
+ "learning_rate": 0.0006,
+ "loss": 4.1169257164001465,
+ "step": 5966
+ },
+ {
+ "epoch": 82.88034934497817,
+ "grad_norm": 0.012060081586241722,
+ "learning_rate": 0.0006,
+ "loss": 4.195242404937744,
+ "step": 5967
+ },
+ {
+ "epoch": 82.8943231441048,
+ "grad_norm": 0.015528378076851368,
+ "learning_rate": 0.0006,
+ "loss": 4.294939994812012,
+ "step": 5968
+ },
+ {
+ "epoch": 82.90829694323143,
+ "grad_norm": 0.018151041120290756,
+ "learning_rate": 0.0006,
+ "loss": 4.137192726135254,
+ "step": 5969
+ },
+ {
+ "epoch": 82.92227074235808,
+ "grad_norm": 0.017895735800266266,
+ "learning_rate": 0.0006,
+ "loss": 4.174036026000977,
+ "step": 5970
+ },
+ {
+ "epoch": 82.93624454148471,
+ "grad_norm": 0.015610386617481709,
+ "learning_rate": 0.0006,
+ "loss": 4.230136871337891,
+ "step": 5971
+ },
+ {
+ "epoch": 82.95021834061136,
+ "grad_norm": 0.01339928898960352,
+ "learning_rate": 0.0006,
+ "loss": 4.199074745178223,
+ "step": 5972
+ },
+ {
+ "epoch": 82.96419213973799,
+ "grad_norm": 0.015454214997589588,
+ "learning_rate": 0.0006,
+ "loss": 4.216109752655029,
+ "step": 5973
+ },
+ {
+ "epoch": 82.97816593886463,
+ "grad_norm": 0.014978132210671902,
+ "learning_rate": 0.0006,
+ "loss": 4.131511688232422,
+ "step": 5974
+ },
+ {
+ "epoch": 82.99213973799127,
+ "grad_norm": 0.015611122362315655,
+ "learning_rate": 0.0006,
+ "loss": 4.246518135070801,
+ "step": 5975
+ },
+ {
+ "epoch": 83.0,
+ "grad_norm": 0.016682764515280724,
+ "learning_rate": 0.0006,
+ "loss": 4.110680103302002,
+ "step": 5976
+ },
+ {
+ "epoch": 83.0,
+ "eval_loss": 4.641707897186279,
+ "eval_runtime": 57.2799,
+ "eval_samples_per_second": 42.633,
+ "eval_steps_per_second": 1.344,
+ "step": 5976
+ },
+ {
+ "epoch": 83.01397379912663,
+ "grad_norm": 0.015007982030510902,
+ "learning_rate": 0.0006,
+ "loss": 4.100912094116211,
+ "step": 5977
+ },
+ {
+ "epoch": 83.02794759825328,
+ "grad_norm": 0.017609436064958572,
+ "learning_rate": 0.0006,
+ "loss": 4.233399868011475,
+ "step": 5978
+ },
+ {
+ "epoch": 83.04192139737991,
+ "grad_norm": 0.019817272201180458,
+ "learning_rate": 0.0006,
+ "loss": 4.191173076629639,
+ "step": 5979
+ },
+ {
+ "epoch": 83.05589519650655,
+ "grad_norm": 0.017752140760421753,
+ "learning_rate": 0.0006,
+ "loss": 4.044437885284424,
+ "step": 5980
+ },
+ {
+ "epoch": 83.06986899563319,
+ "grad_norm": 0.015482760034501553,
+ "learning_rate": 0.0006,
+ "loss": 4.319931983947754,
+ "step": 5981
+ },
+ {
+ "epoch": 83.08384279475983,
+ "grad_norm": 0.015153988264501095,
+ "learning_rate": 0.0006,
+ "loss": 4.095277309417725,
+ "step": 5982
+ },
+ {
+ "epoch": 83.09781659388646,
+ "grad_norm": 0.017025282606482506,
+ "learning_rate": 0.0006,
+ "loss": 4.235110759735107,
+ "step": 5983
+ },
+ {
+ "epoch": 83.1117903930131,
+ "grad_norm": 0.01566586270928383,
+ "learning_rate": 0.0006,
+ "loss": 4.1934685707092285,
+ "step": 5984
+ },
+ {
+ "epoch": 83.12576419213974,
+ "grad_norm": 0.016419263556599617,
+ "learning_rate": 0.0006,
+ "loss": 4.234419822692871,
+ "step": 5985
+ },
+ {
+ "epoch": 83.13973799126637,
+ "grad_norm": 0.017132675275206566,
+ "learning_rate": 0.0006,
+ "loss": 4.221447944641113,
+ "step": 5986
+ },
+ {
+ "epoch": 83.15371179039302,
+ "grad_norm": 0.016138629987835884,
+ "learning_rate": 0.0006,
+ "loss": 4.199830055236816,
+ "step": 5987
+ },
+ {
+ "epoch": 83.16768558951965,
+ "grad_norm": 0.016210174188017845,
+ "learning_rate": 0.0006,
+ "loss": 4.081717491149902,
+ "step": 5988
+ },
+ {
+ "epoch": 83.18165938864628,
+ "grad_norm": 0.016055308282375336,
+ "learning_rate": 0.0006,
+ "loss": 4.247655391693115,
+ "step": 5989
+ },
+ {
+ "epoch": 83.19563318777293,
+ "grad_norm": 0.016132952645421028,
+ "learning_rate": 0.0006,
+ "loss": 4.262535572052002,
+ "step": 5990
+ },
+ {
+ "epoch": 83.20960698689956,
+ "grad_norm": 0.01779935695230961,
+ "learning_rate": 0.0006,
+ "loss": 4.153120040893555,
+ "step": 5991
+ },
+ {
+ "epoch": 83.2235807860262,
+ "grad_norm": 0.018360646441578865,
+ "learning_rate": 0.0006,
+ "loss": 4.232644081115723,
+ "step": 5992
+ },
+ {
+ "epoch": 83.23755458515284,
+ "grad_norm": 0.019629675894975662,
+ "learning_rate": 0.0006,
+ "loss": 3.9758734703063965,
+ "step": 5993
+ },
+ {
+ "epoch": 83.25152838427948,
+ "grad_norm": 0.017819996923208237,
+ "learning_rate": 0.0006,
+ "loss": 4.196932315826416,
+ "step": 5994
+ },
+ {
+ "epoch": 83.26550218340611,
+ "grad_norm": 0.01557681243866682,
+ "learning_rate": 0.0006,
+ "loss": 4.122654914855957,
+ "step": 5995
+ },
+ {
+ "epoch": 83.27947598253274,
+ "grad_norm": 0.017858853563666344,
+ "learning_rate": 0.0006,
+ "loss": 4.185330867767334,
+ "step": 5996
+ },
+ {
+ "epoch": 83.29344978165939,
+ "grad_norm": 0.018065953627228737,
+ "learning_rate": 0.0006,
+ "loss": 4.218471050262451,
+ "step": 5997
+ },
+ {
+ "epoch": 83.30742358078602,
+ "grad_norm": 0.017247939482331276,
+ "learning_rate": 0.0006,
+ "loss": 4.127076148986816,
+ "step": 5998
+ },
+ {
+ "epoch": 83.32139737991267,
+ "grad_norm": 0.017047952860593796,
+ "learning_rate": 0.0006,
+ "loss": 4.131356716156006,
+ "step": 5999
+ },
+ {
+ "epoch": 83.3353711790393,
+ "grad_norm": 0.015905629843473434,
+ "learning_rate": 0.0006,
+ "loss": 4.2239789962768555,
+ "step": 6000
+ },
+ {
+ "epoch": 83.34934497816595,
+ "grad_norm": 0.016522347927093506,
+ "learning_rate": 0.0006,
+ "loss": 4.054897785186768,
+ "step": 6001
+ },
+ {
+ "epoch": 83.36331877729258,
+ "grad_norm": 0.01622571051120758,
+ "learning_rate": 0.0006,
+ "loss": 4.284210681915283,
+ "step": 6002
+ },
+ {
+ "epoch": 83.37729257641921,
+ "grad_norm": 0.01640467159450054,
+ "learning_rate": 0.0006,
+ "loss": 4.268840789794922,
+ "step": 6003
+ },
+ {
+ "epoch": 83.39126637554585,
+ "grad_norm": 0.015698859468102455,
+ "learning_rate": 0.0006,
+ "loss": 4.179793357849121,
+ "step": 6004
+ },
+ {
+ "epoch": 83.40524017467249,
+ "grad_norm": 0.016694528982043266,
+ "learning_rate": 0.0006,
+ "loss": 4.132297515869141,
+ "step": 6005
+ },
+ {
+ "epoch": 83.41921397379913,
+ "grad_norm": 0.015345900319516659,
+ "learning_rate": 0.0006,
+ "loss": 4.090163230895996,
+ "step": 6006
+ },
+ {
+ "epoch": 83.43318777292576,
+ "grad_norm": 0.014722758904099464,
+ "learning_rate": 0.0006,
+ "loss": 4.179166793823242,
+ "step": 6007
+ },
+ {
+ "epoch": 83.44716157205241,
+ "grad_norm": 0.014884887263178825,
+ "learning_rate": 0.0006,
+ "loss": 4.137659072875977,
+ "step": 6008
+ },
+ {
+ "epoch": 83.46113537117904,
+ "grad_norm": 0.014041902497410774,
+ "learning_rate": 0.0006,
+ "loss": 4.1202073097229,
+ "step": 6009
+ },
+ {
+ "epoch": 83.47510917030567,
+ "grad_norm": 0.012799333781003952,
+ "learning_rate": 0.0006,
+ "loss": 4.092607498168945,
+ "step": 6010
+ },
+ {
+ "epoch": 83.48908296943232,
+ "grad_norm": 0.013272707350552082,
+ "learning_rate": 0.0006,
+ "loss": 4.132715225219727,
+ "step": 6011
+ },
+ {
+ "epoch": 83.50305676855895,
+ "grad_norm": 0.012934371829032898,
+ "learning_rate": 0.0006,
+ "loss": 4.230278968811035,
+ "step": 6012
+ },
+ {
+ "epoch": 83.5170305676856,
+ "grad_norm": 0.0120006762444973,
+ "learning_rate": 0.0006,
+ "loss": 4.140068531036377,
+ "step": 6013
+ },
+ {
+ "epoch": 83.53100436681223,
+ "grad_norm": 0.013616700656712055,
+ "learning_rate": 0.0006,
+ "loss": 4.20107364654541,
+ "step": 6014
+ },
+ {
+ "epoch": 83.54497816593886,
+ "grad_norm": 0.013957851566374302,
+ "learning_rate": 0.0006,
+ "loss": 4.124945163726807,
+ "step": 6015
+ },
+ {
+ "epoch": 83.5589519650655,
+ "grad_norm": 0.012762055732309818,
+ "learning_rate": 0.0006,
+ "loss": 4.022052764892578,
+ "step": 6016
+ },
+ {
+ "epoch": 83.57292576419214,
+ "grad_norm": 0.013080576434731483,
+ "learning_rate": 0.0006,
+ "loss": 4.103252410888672,
+ "step": 6017
+ },
+ {
+ "epoch": 83.58689956331878,
+ "grad_norm": 0.013957500457763672,
+ "learning_rate": 0.0006,
+ "loss": 4.235089302062988,
+ "step": 6018
+ },
+ {
+ "epoch": 83.60087336244541,
+ "grad_norm": 0.014707164838910103,
+ "learning_rate": 0.0006,
+ "loss": 4.191570281982422,
+ "step": 6019
+ },
+ {
+ "epoch": 83.61484716157206,
+ "grad_norm": 0.012807869352400303,
+ "learning_rate": 0.0006,
+ "loss": 4.294543266296387,
+ "step": 6020
+ },
+ {
+ "epoch": 83.62882096069869,
+ "grad_norm": 0.012027994729578495,
+ "learning_rate": 0.0006,
+ "loss": 4.13849401473999,
+ "step": 6021
+ },
+ {
+ "epoch": 83.64279475982532,
+ "grad_norm": 0.012147724628448486,
+ "learning_rate": 0.0006,
+ "loss": 4.127481937408447,
+ "step": 6022
+ },
+ {
+ "epoch": 83.65676855895197,
+ "grad_norm": 0.013241392560303211,
+ "learning_rate": 0.0006,
+ "loss": 4.218972682952881,
+ "step": 6023
+ },
+ {
+ "epoch": 83.6707423580786,
+ "grad_norm": 0.014071648940443993,
+ "learning_rate": 0.0006,
+ "loss": 4.096724033355713,
+ "step": 6024
+ },
+ {
+ "epoch": 83.68471615720524,
+ "grad_norm": 0.015212797559797764,
+ "learning_rate": 0.0006,
+ "loss": 4.27097225189209,
+ "step": 6025
+ },
+ {
+ "epoch": 83.69868995633188,
+ "grad_norm": 0.017647385597229004,
+ "learning_rate": 0.0006,
+ "loss": 4.209835529327393,
+ "step": 6026
+ },
+ {
+ "epoch": 83.71266375545852,
+ "grad_norm": 0.01709647662937641,
+ "learning_rate": 0.0006,
+ "loss": 4.042106628417969,
+ "step": 6027
+ },
+ {
+ "epoch": 83.72663755458515,
+ "grad_norm": 0.014530927874147892,
+ "learning_rate": 0.0006,
+ "loss": 4.181731700897217,
+ "step": 6028
+ },
+ {
+ "epoch": 83.74061135371178,
+ "grad_norm": 0.015356711111962795,
+ "learning_rate": 0.0006,
+ "loss": 4.314183235168457,
+ "step": 6029
+ },
+ {
+ "epoch": 83.75458515283843,
+ "grad_norm": 0.015500029549002647,
+ "learning_rate": 0.0006,
+ "loss": 4.06776762008667,
+ "step": 6030
+ },
+ {
+ "epoch": 83.76855895196506,
+ "grad_norm": 0.015351332724094391,
+ "learning_rate": 0.0006,
+ "loss": 4.185141563415527,
+ "step": 6031
+ },
+ {
+ "epoch": 83.78253275109171,
+ "grad_norm": 0.016678154468536377,
+ "learning_rate": 0.0006,
+ "loss": 4.179529666900635,
+ "step": 6032
+ },
+ {
+ "epoch": 83.79650655021834,
+ "grad_norm": 0.0156265702098608,
+ "learning_rate": 0.0006,
+ "loss": 4.238584041595459,
+ "step": 6033
+ },
+ {
+ "epoch": 83.81048034934499,
+ "grad_norm": 0.016418596729636192,
+ "learning_rate": 0.0006,
+ "loss": 4.1593523025512695,
+ "step": 6034
+ },
+ {
+ "epoch": 83.82445414847162,
+ "grad_norm": 0.016113586723804474,
+ "learning_rate": 0.0006,
+ "loss": 4.157470226287842,
+ "step": 6035
+ },
+ {
+ "epoch": 83.83842794759825,
+ "grad_norm": 0.014434042386710644,
+ "learning_rate": 0.0006,
+ "loss": 4.196226119995117,
+ "step": 6036
+ },
+ {
+ "epoch": 83.8524017467249,
+ "grad_norm": 0.012647042982280254,
+ "learning_rate": 0.0006,
+ "loss": 4.130589485168457,
+ "step": 6037
+ },
+ {
+ "epoch": 83.86637554585153,
+ "grad_norm": 0.014412877149879932,
+ "learning_rate": 0.0006,
+ "loss": 4.154845237731934,
+ "step": 6038
+ },
+ {
+ "epoch": 83.88034934497817,
+ "grad_norm": 0.014222027733922005,
+ "learning_rate": 0.0006,
+ "loss": 4.189724922180176,
+ "step": 6039
+ },
+ {
+ "epoch": 83.8943231441048,
+ "grad_norm": 0.013084255158901215,
+ "learning_rate": 0.0006,
+ "loss": 4.303338527679443,
+ "step": 6040
+ },
+ {
+ "epoch": 83.90829694323143,
+ "grad_norm": 0.012659232132136822,
+ "learning_rate": 0.0006,
+ "loss": 4.08192253112793,
+ "step": 6041
+ },
+ {
+ "epoch": 83.92227074235808,
+ "grad_norm": 0.01246276218444109,
+ "learning_rate": 0.0006,
+ "loss": 4.283324718475342,
+ "step": 6042
+ },
+ {
+ "epoch": 83.93624454148471,
+ "grad_norm": 0.01340118795633316,
+ "learning_rate": 0.0006,
+ "loss": 4.117908477783203,
+ "step": 6043
+ },
+ {
+ "epoch": 83.95021834061136,
+ "grad_norm": 0.014496722258627415,
+ "learning_rate": 0.0006,
+ "loss": 4.110318660736084,
+ "step": 6044
+ },
+ {
+ "epoch": 83.96419213973799,
+ "grad_norm": 0.013348298147320747,
+ "learning_rate": 0.0006,
+ "loss": 4.185139179229736,
+ "step": 6045
+ },
+ {
+ "epoch": 83.97816593886463,
+ "grad_norm": 0.013294277712702751,
+ "learning_rate": 0.0006,
+ "loss": 4.271645545959473,
+ "step": 6046
+ },
+ {
+ "epoch": 83.99213973799127,
+ "grad_norm": 0.014431594870984554,
+ "learning_rate": 0.0006,
+ "loss": 4.161383152008057,
+ "step": 6047
+ },
+ {
+ "epoch": 84.0,
+ "grad_norm": 0.016268927603960037,
+ "learning_rate": 0.0006,
+ "loss": 4.178324222564697,
+ "step": 6048
+ },
+ {
+ "epoch": 84.0,
+ "eval_loss": 4.6307549476623535,
+ "eval_runtime": 56.8521,
+ "eval_samples_per_second": 42.954,
+ "eval_steps_per_second": 1.354,
+ "step": 6048
+ },
+ {
+ "epoch": 84.01397379912663,
+ "grad_norm": 0.014864908531308174,
+ "learning_rate": 0.0006,
+ "loss": 4.121300220489502,
+ "step": 6049
+ },
+ {
+ "epoch": 84.02794759825328,
+ "grad_norm": 0.014863367192447186,
+ "learning_rate": 0.0006,
+ "loss": 4.1650543212890625,
+ "step": 6050
+ },
+ {
+ "epoch": 84.04192139737991,
+ "grad_norm": 0.014379646629095078,
+ "learning_rate": 0.0006,
+ "loss": 4.0201239585876465,
+ "step": 6051
+ },
+ {
+ "epoch": 84.05589519650655,
+ "grad_norm": 0.014871201477944851,
+ "learning_rate": 0.0006,
+ "loss": 4.0820698738098145,
+ "step": 6052
+ },
+ {
+ "epoch": 84.06986899563319,
+ "grad_norm": 0.013085179962217808,
+ "learning_rate": 0.0006,
+ "loss": 4.223984718322754,
+ "step": 6053
+ },
+ {
+ "epoch": 84.08384279475983,
+ "grad_norm": 0.012485949322581291,
+ "learning_rate": 0.0006,
+ "loss": 4.182206153869629,
+ "step": 6054
+ },
+ {
+ "epoch": 84.09781659388646,
+ "grad_norm": 0.013653291389346123,
+ "learning_rate": 0.0006,
+ "loss": 4.142557144165039,
+ "step": 6055
+ },
+ {
+ "epoch": 84.1117903930131,
+ "grad_norm": 0.013930532149970531,
+ "learning_rate": 0.0006,
+ "loss": 4.13972282409668,
+ "step": 6056
+ },
+ {
+ "epoch": 84.12576419213974,
+ "grad_norm": 0.01297379657626152,
+ "learning_rate": 0.0006,
+ "loss": 4.144664764404297,
+ "step": 6057
+ },
+ {
+ "epoch": 84.13973799126637,
+ "grad_norm": 0.013210492208600044,
+ "learning_rate": 0.0006,
+ "loss": 4.253764629364014,
+ "step": 6058
+ },
+ {
+ "epoch": 84.15371179039302,
+ "grad_norm": 0.014650176279246807,
+ "learning_rate": 0.0006,
+ "loss": 4.246866226196289,
+ "step": 6059
+ },
+ {
+ "epoch": 84.16768558951965,
+ "grad_norm": 0.01482292078435421,
+ "learning_rate": 0.0006,
+ "loss": 4.140187740325928,
+ "step": 6060
+ },
+ {
+ "epoch": 84.18165938864628,
+ "grad_norm": 0.01592552848160267,
+ "learning_rate": 0.0006,
+ "loss": 4.170705318450928,
+ "step": 6061
+ },
+ {
+ "epoch": 84.19563318777293,
+ "grad_norm": 0.017152156680822372,
+ "learning_rate": 0.0006,
+ "loss": 4.126092433929443,
+ "step": 6062
+ },
+ {
+ "epoch": 84.20960698689956,
+ "grad_norm": 0.015880905091762543,
+ "learning_rate": 0.0006,
+ "loss": 4.161938190460205,
+ "step": 6063
+ },
+ {
+ "epoch": 84.2235807860262,
+ "grad_norm": 0.015451471321284771,
+ "learning_rate": 0.0006,
+ "loss": 4.263461112976074,
+ "step": 6064
+ },
+ {
+ "epoch": 84.23755458515284,
+ "grad_norm": 0.014365996234118938,
+ "learning_rate": 0.0006,
+ "loss": 4.14488410949707,
+ "step": 6065
+ },
+ {
+ "epoch": 84.25152838427948,
+ "grad_norm": 0.016364749521017075,
+ "learning_rate": 0.0006,
+ "loss": 4.112489700317383,
+ "step": 6066
+ },
+ {
+ "epoch": 84.26550218340611,
+ "grad_norm": 0.01791754551231861,
+ "learning_rate": 0.0006,
+ "loss": 4.133513450622559,
+ "step": 6067
+ },
+ {
+ "epoch": 84.27947598253274,
+ "grad_norm": 0.016609208658337593,
+ "learning_rate": 0.0006,
+ "loss": 4.201829433441162,
+ "step": 6068
+ },
+ {
+ "epoch": 84.29344978165939,
+ "grad_norm": 0.017587820068001747,
+ "learning_rate": 0.0006,
+ "loss": 4.203151702880859,
+ "step": 6069
+ },
+ {
+ "epoch": 84.30742358078602,
+ "grad_norm": 0.017047742381691933,
+ "learning_rate": 0.0006,
+ "loss": 4.077008247375488,
+ "step": 6070
+ },
+ {
+ "epoch": 84.32139737991267,
+ "grad_norm": 0.01524646207690239,
+ "learning_rate": 0.0006,
+ "loss": 4.198038101196289,
+ "step": 6071
+ },
+ {
+ "epoch": 84.3353711790393,
+ "grad_norm": 0.014196421951055527,
+ "learning_rate": 0.0006,
+ "loss": 4.164245128631592,
+ "step": 6072
+ },
+ {
+ "epoch": 84.34934497816595,
+ "grad_norm": 0.016291776672005653,
+ "learning_rate": 0.0006,
+ "loss": 4.150152206420898,
+ "step": 6073
+ },
+ {
+ "epoch": 84.36331877729258,
+ "grad_norm": 0.015627920627593994,
+ "learning_rate": 0.0006,
+ "loss": 4.110006332397461,
+ "step": 6074
+ },
+ {
+ "epoch": 84.37729257641921,
+ "grad_norm": 0.013251487165689468,
+ "learning_rate": 0.0006,
+ "loss": 4.101463317871094,
+ "step": 6075
+ },
+ {
+ "epoch": 84.39126637554585,
+ "grad_norm": 0.013601605780422688,
+ "learning_rate": 0.0006,
+ "loss": 4.191105842590332,
+ "step": 6076
+ },
+ {
+ "epoch": 84.40524017467249,
+ "grad_norm": 0.014878741465508938,
+ "learning_rate": 0.0006,
+ "loss": 4.153046131134033,
+ "step": 6077
+ },
+ {
+ "epoch": 84.41921397379913,
+ "grad_norm": 0.014783897437155247,
+ "learning_rate": 0.0006,
+ "loss": 4.362000465393066,
+ "step": 6078
+ },
+ {
+ "epoch": 84.43318777292576,
+ "grad_norm": 0.014221332035958767,
+ "learning_rate": 0.0006,
+ "loss": 4.144364833831787,
+ "step": 6079
+ },
+ {
+ "epoch": 84.44716157205241,
+ "grad_norm": 0.014911034144461155,
+ "learning_rate": 0.0006,
+ "loss": 4.088857650756836,
+ "step": 6080
+ },
+ {
+ "epoch": 84.46113537117904,
+ "grad_norm": 0.014924770221114159,
+ "learning_rate": 0.0006,
+ "loss": 4.1677446365356445,
+ "step": 6081
+ },
+ {
+ "epoch": 84.47510917030567,
+ "grad_norm": 0.014394409023225307,
+ "learning_rate": 0.0006,
+ "loss": 4.087103843688965,
+ "step": 6082
+ },
+ {
+ "epoch": 84.48908296943232,
+ "grad_norm": 0.013926676474511623,
+ "learning_rate": 0.0006,
+ "loss": 3.9972851276397705,
+ "step": 6083
+ },
+ {
+ "epoch": 84.50305676855895,
+ "grad_norm": 0.013069954700767994,
+ "learning_rate": 0.0006,
+ "loss": 3.998237133026123,
+ "step": 6084
+ },
+ {
+ "epoch": 84.5170305676856,
+ "grad_norm": 0.012748011387884617,
+ "learning_rate": 0.0006,
+ "loss": 4.195858955383301,
+ "step": 6085
+ },
+ {
+ "epoch": 84.53100436681223,
+ "grad_norm": 0.012278077192604542,
+ "learning_rate": 0.0006,
+ "loss": 4.102165222167969,
+ "step": 6086
+ },
+ {
+ "epoch": 84.54497816593886,
+ "grad_norm": 0.014259163290262222,
+ "learning_rate": 0.0006,
+ "loss": 4.092750549316406,
+ "step": 6087
+ },
+ {
+ "epoch": 84.5589519650655,
+ "grad_norm": 0.014553683809936047,
+ "learning_rate": 0.0006,
+ "loss": 4.1412763595581055,
+ "step": 6088
+ },
+ {
+ "epoch": 84.57292576419214,
+ "grad_norm": 0.013605243526399136,
+ "learning_rate": 0.0006,
+ "loss": 4.192173480987549,
+ "step": 6089
+ },
+ {
+ "epoch": 84.58689956331878,
+ "grad_norm": 0.014259358868002892,
+ "learning_rate": 0.0006,
+ "loss": 4.356374740600586,
+ "step": 6090
+ },
+ {
+ "epoch": 84.60087336244541,
+ "grad_norm": 0.014166025444865227,
+ "learning_rate": 0.0006,
+ "loss": 4.165982246398926,
+ "step": 6091
+ },
+ {
+ "epoch": 84.61484716157206,
+ "grad_norm": 0.012198803946375847,
+ "learning_rate": 0.0006,
+ "loss": 4.287566184997559,
+ "step": 6092
+ },
+ {
+ "epoch": 84.62882096069869,
+ "grad_norm": 0.013794238679111004,
+ "learning_rate": 0.0006,
+ "loss": 4.199882984161377,
+ "step": 6093
+ },
+ {
+ "epoch": 84.64279475982532,
+ "grad_norm": 0.014757642522454262,
+ "learning_rate": 0.0006,
+ "loss": 4.116573333740234,
+ "step": 6094
+ },
+ {
+ "epoch": 84.65676855895197,
+ "grad_norm": 0.014591689221560955,
+ "learning_rate": 0.0006,
+ "loss": 4.240896224975586,
+ "step": 6095
+ },
+ {
+ "epoch": 84.6707423580786,
+ "grad_norm": 0.015217579901218414,
+ "learning_rate": 0.0006,
+ "loss": 4.257142066955566,
+ "step": 6096
+ },
+ {
+ "epoch": 84.68471615720524,
+ "grad_norm": 0.015371371991932392,
+ "learning_rate": 0.0006,
+ "loss": 4.181037425994873,
+ "step": 6097
+ },
+ {
+ "epoch": 84.69868995633188,
+ "grad_norm": 0.014412983320653439,
+ "learning_rate": 0.0006,
+ "loss": 4.2207794189453125,
+ "step": 6098
+ },
+ {
+ "epoch": 84.71266375545852,
+ "grad_norm": 0.013773671351373196,
+ "learning_rate": 0.0006,
+ "loss": 4.211123466491699,
+ "step": 6099
+ },
+ {
+ "epoch": 84.72663755458515,
+ "grad_norm": 0.012838046066462994,
+ "learning_rate": 0.0006,
+ "loss": 4.175459384918213,
+ "step": 6100
+ },
+ {
+ "epoch": 84.74061135371178,
+ "grad_norm": 0.014094929210841656,
+ "learning_rate": 0.0006,
+ "loss": 4.1724958419799805,
+ "step": 6101
+ },
+ {
+ "epoch": 84.75458515283843,
+ "grad_norm": 0.014573262073099613,
+ "learning_rate": 0.0006,
+ "loss": 4.15885066986084,
+ "step": 6102
+ },
+ {
+ "epoch": 84.76855895196506,
+ "grad_norm": 0.011964819394052029,
+ "learning_rate": 0.0006,
+ "loss": 3.9804892539978027,
+ "step": 6103
+ },
+ {
+ "epoch": 84.78253275109171,
+ "grad_norm": 0.014140539802610874,
+ "learning_rate": 0.0006,
+ "loss": 4.203708648681641,
+ "step": 6104
+ },
+ {
+ "epoch": 84.79650655021834,
+ "grad_norm": 0.016618940979242325,
+ "learning_rate": 0.0006,
+ "loss": 4.147684097290039,
+ "step": 6105
+ },
+ {
+ "epoch": 84.81048034934499,
+ "grad_norm": 0.015094385482370853,
+ "learning_rate": 0.0006,
+ "loss": 4.392223834991455,
+ "step": 6106
+ },
+ {
+ "epoch": 84.82445414847162,
+ "grad_norm": 0.014386717230081558,
+ "learning_rate": 0.0006,
+ "loss": 4.253332138061523,
+ "step": 6107
+ },
+ {
+ "epoch": 84.83842794759825,
+ "grad_norm": 0.013877755962312222,
+ "learning_rate": 0.0006,
+ "loss": 4.075056076049805,
+ "step": 6108
+ },
+ {
+ "epoch": 84.8524017467249,
+ "grad_norm": 0.014062155969440937,
+ "learning_rate": 0.0006,
+ "loss": 4.072048664093018,
+ "step": 6109
+ },
+ {
+ "epoch": 84.86637554585153,
+ "grad_norm": 0.014445619657635689,
+ "learning_rate": 0.0006,
+ "loss": 4.151832580566406,
+ "step": 6110
+ },
+ {
+ "epoch": 84.88034934497817,
+ "grad_norm": 0.015116363763809204,
+ "learning_rate": 0.0006,
+ "loss": 4.088900566101074,
+ "step": 6111
+ },
+ {
+ "epoch": 84.8943231441048,
+ "grad_norm": 0.014425474219024181,
+ "learning_rate": 0.0006,
+ "loss": 4.200483322143555,
+ "step": 6112
+ },
+ {
+ "epoch": 84.90829694323143,
+ "grad_norm": 0.01449573878198862,
+ "learning_rate": 0.0006,
+ "loss": 4.239742279052734,
+ "step": 6113
+ },
+ {
+ "epoch": 84.92227074235808,
+ "grad_norm": 0.015698261559009552,
+ "learning_rate": 0.0006,
+ "loss": 4.197391510009766,
+ "step": 6114
+ },
+ {
+ "epoch": 84.93624454148471,
+ "grad_norm": 0.014545733109116554,
+ "learning_rate": 0.0006,
+ "loss": 4.035508155822754,
+ "step": 6115
+ },
+ {
+ "epoch": 84.95021834061136,
+ "grad_norm": 0.012796483002603054,
+ "learning_rate": 0.0006,
+ "loss": 4.195248126983643,
+ "step": 6116
+ },
+ {
+ "epoch": 84.96419213973799,
+ "grad_norm": 0.01264485064893961,
+ "learning_rate": 0.0006,
+ "loss": 4.158118724822998,
+ "step": 6117
+ },
+ {
+ "epoch": 84.97816593886463,
+ "grad_norm": 0.013695276342332363,
+ "learning_rate": 0.0006,
+ "loss": 4.255288124084473,
+ "step": 6118
+ },
+ {
+ "epoch": 84.99213973799127,
+ "grad_norm": 0.013459564186632633,
+ "learning_rate": 0.0006,
+ "loss": 4.204586505889893,
+ "step": 6119
+ },
+ {
+ "epoch": 85.0,
+ "grad_norm": 0.014312652871012688,
+ "learning_rate": 0.0006,
+ "loss": 4.167879104614258,
+ "step": 6120
+ },
+ {
+ "epoch": 85.0,
+ "eval_loss": 4.608152389526367,
+ "eval_runtime": 57.1903,
+ "eval_samples_per_second": 42.7,
+ "eval_steps_per_second": 1.346,
+ "step": 6120
+ },
+ {
+ "epoch": 85.01397379912663,
+ "grad_norm": 0.013918368145823479,
+ "learning_rate": 0.0006,
+ "loss": 4.129991054534912,
+ "step": 6121
+ },
+ {
+ "epoch": 85.02794759825328,
+ "grad_norm": 0.01604384370148182,
+ "learning_rate": 0.0006,
+ "loss": 4.233926773071289,
+ "step": 6122
+ },
+ {
+ "epoch": 85.04192139737991,
+ "grad_norm": 0.016565652564167976,
+ "learning_rate": 0.0006,
+ "loss": 4.129604339599609,
+ "step": 6123
+ },
+ {
+ "epoch": 85.05589519650655,
+ "grad_norm": 0.0149751678109169,
+ "learning_rate": 0.0006,
+ "loss": 4.165839195251465,
+ "step": 6124
+ },
+ {
+ "epoch": 85.06986899563319,
+ "grad_norm": 0.013812461867928505,
+ "learning_rate": 0.0006,
+ "loss": 4.153792858123779,
+ "step": 6125
+ },
+ {
+ "epoch": 85.08384279475983,
+ "grad_norm": 0.014432745985686779,
+ "learning_rate": 0.0006,
+ "loss": 4.170895576477051,
+ "step": 6126
+ },
+ {
+ "epoch": 85.09781659388646,
+ "grad_norm": 0.015208868309855461,
+ "learning_rate": 0.0006,
+ "loss": 4.10639762878418,
+ "step": 6127
+ },
+ {
+ "epoch": 85.1117903930131,
+ "grad_norm": 0.013209272176027298,
+ "learning_rate": 0.0006,
+ "loss": 4.200364112854004,
+ "step": 6128
+ },
+ {
+ "epoch": 85.12576419213974,
+ "grad_norm": 0.014807002618908882,
+ "learning_rate": 0.0006,
+ "loss": 4.172666549682617,
+ "step": 6129
+ },
+ {
+ "epoch": 85.13973799126637,
+ "grad_norm": 0.015662945806980133,
+ "learning_rate": 0.0006,
+ "loss": 4.145219326019287,
+ "step": 6130
+ },
+ {
+ "epoch": 85.15371179039302,
+ "grad_norm": 0.014598497189581394,
+ "learning_rate": 0.0006,
+ "loss": 4.275618553161621,
+ "step": 6131
+ },
+ {
+ "epoch": 85.16768558951965,
+ "grad_norm": 0.013778576627373695,
+ "learning_rate": 0.0006,
+ "loss": 4.204716682434082,
+ "step": 6132
+ },
+ {
+ "epoch": 85.18165938864628,
+ "grad_norm": 0.015762265771627426,
+ "learning_rate": 0.0006,
+ "loss": 4.145524024963379,
+ "step": 6133
+ },
+ {
+ "epoch": 85.19563318777293,
+ "grad_norm": 0.015579195693135262,
+ "learning_rate": 0.0006,
+ "loss": 4.150625228881836,
+ "step": 6134
+ },
+ {
+ "epoch": 85.20960698689956,
+ "grad_norm": 0.016641216352581978,
+ "learning_rate": 0.0006,
+ "loss": 4.141819953918457,
+ "step": 6135
+ },
+ {
+ "epoch": 85.2235807860262,
+ "grad_norm": 0.01632271707057953,
+ "learning_rate": 0.0006,
+ "loss": 4.030620574951172,
+ "step": 6136
+ },
+ {
+ "epoch": 85.23755458515284,
+ "grad_norm": 0.015564918518066406,
+ "learning_rate": 0.0006,
+ "loss": 4.147753715515137,
+ "step": 6137
+ },
+ {
+ "epoch": 85.25152838427948,
+ "grad_norm": 0.0161129217594862,
+ "learning_rate": 0.0006,
+ "loss": 4.128524303436279,
+ "step": 6138
+ },
+ {
+ "epoch": 85.26550218340611,
+ "grad_norm": 0.017766550183296204,
+ "learning_rate": 0.0006,
+ "loss": 4.134856224060059,
+ "step": 6139
+ },
+ {
+ "epoch": 85.27947598253274,
+ "grad_norm": 0.01869054324924946,
+ "learning_rate": 0.0006,
+ "loss": 4.231565475463867,
+ "step": 6140
+ },
+ {
+ "epoch": 85.29344978165939,
+ "grad_norm": 0.01917010359466076,
+ "learning_rate": 0.0006,
+ "loss": 4.21175479888916,
+ "step": 6141
+ },
+ {
+ "epoch": 85.30742358078602,
+ "grad_norm": 0.018116485327482224,
+ "learning_rate": 0.0006,
+ "loss": 4.122644424438477,
+ "step": 6142
+ },
+ {
+ "epoch": 85.32139737991267,
+ "grad_norm": 0.015949107706546783,
+ "learning_rate": 0.0006,
+ "loss": 4.135458946228027,
+ "step": 6143
+ },
+ {
+ "epoch": 85.3353711790393,
+ "grad_norm": 0.014783683232963085,
+ "learning_rate": 0.0006,
+ "loss": 4.118994235992432,
+ "step": 6144
+ },
+ {
+ "epoch": 85.34934497816595,
+ "grad_norm": 0.015406692400574684,
+ "learning_rate": 0.0006,
+ "loss": 4.157268524169922,
+ "step": 6145
+ },
+ {
+ "epoch": 85.36331877729258,
+ "grad_norm": 0.013893205672502518,
+ "learning_rate": 0.0006,
+ "loss": 4.056166648864746,
+ "step": 6146
+ },
+ {
+ "epoch": 85.37729257641921,
+ "grad_norm": 0.014380923472344875,
+ "learning_rate": 0.0006,
+ "loss": 4.206905364990234,
+ "step": 6147
+ },
+ {
+ "epoch": 85.39126637554585,
+ "grad_norm": 0.015669258311390877,
+ "learning_rate": 0.0006,
+ "loss": 4.213881492614746,
+ "step": 6148
+ },
+ {
+ "epoch": 85.40524017467249,
+ "grad_norm": 0.01553428452461958,
+ "learning_rate": 0.0006,
+ "loss": 4.1814374923706055,
+ "step": 6149
+ },
+ {
+ "epoch": 85.41921397379913,
+ "grad_norm": 0.01697700284421444,
+ "learning_rate": 0.0006,
+ "loss": 4.083939552307129,
+ "step": 6150
+ },
+ {
+ "epoch": 85.43318777292576,
+ "grad_norm": 0.016639070585370064,
+ "learning_rate": 0.0006,
+ "loss": 4.279311180114746,
+ "step": 6151
+ },
+ {
+ "epoch": 85.44716157205241,
+ "grad_norm": 0.016285281628370285,
+ "learning_rate": 0.0006,
+ "loss": 4.244845390319824,
+ "step": 6152
+ },
+ {
+ "epoch": 85.46113537117904,
+ "grad_norm": 0.016071317717432976,
+ "learning_rate": 0.0006,
+ "loss": 4.089814186096191,
+ "step": 6153
+ },
+ {
+ "epoch": 85.47510917030567,
+ "grad_norm": 0.01640418916940689,
+ "learning_rate": 0.0006,
+ "loss": 4.168088912963867,
+ "step": 6154
+ },
+ {
+ "epoch": 85.48908296943232,
+ "grad_norm": 0.014781414531171322,
+ "learning_rate": 0.0006,
+ "loss": 4.196092128753662,
+ "step": 6155
+ },
+ {
+ "epoch": 85.50305676855895,
+ "grad_norm": 0.013393213972449303,
+ "learning_rate": 0.0006,
+ "loss": 4.193429946899414,
+ "step": 6156
+ },
+ {
+ "epoch": 85.5170305676856,
+ "grad_norm": 0.01467406190931797,
+ "learning_rate": 0.0006,
+ "loss": 4.216714859008789,
+ "step": 6157
+ },
+ {
+ "epoch": 85.53100436681223,
+ "grad_norm": 0.014950765296816826,
+ "learning_rate": 0.0006,
+ "loss": 4.093854904174805,
+ "step": 6158
+ },
+ {
+ "epoch": 85.54497816593886,
+ "grad_norm": 0.014655301347374916,
+ "learning_rate": 0.0006,
+ "loss": 4.09483528137207,
+ "step": 6159
+ },
+ {
+ "epoch": 85.5589519650655,
+ "grad_norm": 0.01425520796328783,
+ "learning_rate": 0.0006,
+ "loss": 4.090396881103516,
+ "step": 6160
+ },
+ {
+ "epoch": 85.57292576419214,
+ "grad_norm": 0.01403278298676014,
+ "learning_rate": 0.0006,
+ "loss": 4.088724136352539,
+ "step": 6161
+ },
+ {
+ "epoch": 85.58689956331878,
+ "grad_norm": 0.01364967692643404,
+ "learning_rate": 0.0006,
+ "loss": 4.095558166503906,
+ "step": 6162
+ },
+ {
+ "epoch": 85.60087336244541,
+ "grad_norm": 0.013085267506539822,
+ "learning_rate": 0.0006,
+ "loss": 4.163962364196777,
+ "step": 6163
+ },
+ {
+ "epoch": 85.61484716157206,
+ "grad_norm": 0.0136906448751688,
+ "learning_rate": 0.0006,
+ "loss": 4.197134017944336,
+ "step": 6164
+ },
+ {
+ "epoch": 85.62882096069869,
+ "grad_norm": 0.013472451828420162,
+ "learning_rate": 0.0006,
+ "loss": 4.1578369140625,
+ "step": 6165
+ },
+ {
+ "epoch": 85.64279475982532,
+ "grad_norm": 0.012329773046076298,
+ "learning_rate": 0.0006,
+ "loss": 4.231907844543457,
+ "step": 6166
+ },
+ {
+ "epoch": 85.65676855895197,
+ "grad_norm": 0.011977200396358967,
+ "learning_rate": 0.0006,
+ "loss": 4.1552958488464355,
+ "step": 6167
+ },
+ {
+ "epoch": 85.6707423580786,
+ "grad_norm": 0.013049956411123276,
+ "learning_rate": 0.0006,
+ "loss": 4.17852783203125,
+ "step": 6168
+ },
+ {
+ "epoch": 85.68471615720524,
+ "grad_norm": 0.011440741829574108,
+ "learning_rate": 0.0006,
+ "loss": 4.144646644592285,
+ "step": 6169
+ },
+ {
+ "epoch": 85.69868995633188,
+ "grad_norm": 0.012472936883568764,
+ "learning_rate": 0.0006,
+ "loss": 4.2306413650512695,
+ "step": 6170
+ },
+ {
+ "epoch": 85.71266375545852,
+ "grad_norm": 0.013015356846153736,
+ "learning_rate": 0.0006,
+ "loss": 4.193439960479736,
+ "step": 6171
+ },
+ {
+ "epoch": 85.72663755458515,
+ "grad_norm": 0.013832319527864456,
+ "learning_rate": 0.0006,
+ "loss": 4.18073034286499,
+ "step": 6172
+ },
+ {
+ "epoch": 85.74061135371178,
+ "grad_norm": 0.014613982290029526,
+ "learning_rate": 0.0006,
+ "loss": 4.223531246185303,
+ "step": 6173
+ },
+ {
+ "epoch": 85.75458515283843,
+ "grad_norm": 0.014209083281457424,
+ "learning_rate": 0.0006,
+ "loss": 4.142666816711426,
+ "step": 6174
+ },
+ {
+ "epoch": 85.76855895196506,
+ "grad_norm": 0.013396669179201126,
+ "learning_rate": 0.0006,
+ "loss": 4.185786247253418,
+ "step": 6175
+ },
+ {
+ "epoch": 85.78253275109171,
+ "grad_norm": 0.01364929135888815,
+ "learning_rate": 0.0006,
+ "loss": 4.125226974487305,
+ "step": 6176
+ },
+ {
+ "epoch": 85.79650655021834,
+ "grad_norm": 0.012561003677546978,
+ "learning_rate": 0.0006,
+ "loss": 4.084519386291504,
+ "step": 6177
+ },
+ {
+ "epoch": 85.81048034934499,
+ "grad_norm": 0.013508481904864311,
+ "learning_rate": 0.0006,
+ "loss": 4.193173885345459,
+ "step": 6178
+ },
+ {
+ "epoch": 85.82445414847162,
+ "grad_norm": 0.014154613018035889,
+ "learning_rate": 0.0006,
+ "loss": 4.257272720336914,
+ "step": 6179
+ },
+ {
+ "epoch": 85.83842794759825,
+ "grad_norm": 0.012478531338274479,
+ "learning_rate": 0.0006,
+ "loss": 4.161380767822266,
+ "step": 6180
+ },
+ {
+ "epoch": 85.8524017467249,
+ "grad_norm": 0.013468451797962189,
+ "learning_rate": 0.0006,
+ "loss": 4.121395111083984,
+ "step": 6181
+ },
+ {
+ "epoch": 85.86637554585153,
+ "grad_norm": 0.015155194327235222,
+ "learning_rate": 0.0006,
+ "loss": 4.242815971374512,
+ "step": 6182
+ },
+ {
+ "epoch": 85.88034934497817,
+ "grad_norm": 0.016223272308707237,
+ "learning_rate": 0.0006,
+ "loss": 4.176027297973633,
+ "step": 6183
+ },
+ {
+ "epoch": 85.8943231441048,
+ "grad_norm": 0.013741295784711838,
+ "learning_rate": 0.0006,
+ "loss": 4.131257057189941,
+ "step": 6184
+ },
+ {
+ "epoch": 85.90829694323143,
+ "grad_norm": 0.014982067979872227,
+ "learning_rate": 0.0006,
+ "loss": 4.237896919250488,
+ "step": 6185
+ },
+ {
+ "epoch": 85.92227074235808,
+ "grad_norm": 0.018197322264313698,
+ "learning_rate": 0.0006,
+ "loss": 4.154962539672852,
+ "step": 6186
+ },
+ {
+ "epoch": 85.93624454148471,
+ "grad_norm": 0.01971200481057167,
+ "learning_rate": 0.0006,
+ "loss": 4.16819429397583,
+ "step": 6187
+ },
+ {
+ "epoch": 85.95021834061136,
+ "grad_norm": 0.018008720129728317,
+ "learning_rate": 0.0006,
+ "loss": 4.146164417266846,
+ "step": 6188
+ },
+ {
+ "epoch": 85.96419213973799,
+ "grad_norm": 0.015859264880418777,
+ "learning_rate": 0.0006,
+ "loss": 4.208099365234375,
+ "step": 6189
+ },
+ {
+ "epoch": 85.97816593886463,
+ "grad_norm": 0.013846252113580704,
+ "learning_rate": 0.0006,
+ "loss": 4.1881184577941895,
+ "step": 6190
+ },
+ {
+ "epoch": 85.99213973799127,
+ "grad_norm": 0.018641939386725426,
+ "learning_rate": 0.0006,
+ "loss": 4.127632141113281,
+ "step": 6191
+ },
+ {
+ "epoch": 86.0,
+ "grad_norm": 0.02385227382183075,
+ "learning_rate": 0.0006,
+ "loss": 4.281900405883789,
+ "step": 6192
+ },
+ {
+ "epoch": 86.0,
+ "eval_loss": 4.625458717346191,
+ "eval_runtime": 57.0634,
+ "eval_samples_per_second": 42.795,
+ "eval_steps_per_second": 1.349,
+ "step": 6192
+ },
+ {
+ "epoch": 86.01397379912663,
+ "grad_norm": 0.019032124429941177,
+ "learning_rate": 0.0006,
+ "loss": 4.2163190841674805,
+ "step": 6193
+ },
+ {
+ "epoch": 86.02794759825328,
+ "grad_norm": 0.018049925565719604,
+ "learning_rate": 0.0006,
+ "loss": 4.1534037590026855,
+ "step": 6194
+ },
+ {
+ "epoch": 86.04192139737991,
+ "grad_norm": 0.019116360694169998,
+ "learning_rate": 0.0006,
+ "loss": 4.212411403656006,
+ "step": 6195
+ },
+ {
+ "epoch": 86.05589519650655,
+ "grad_norm": 0.018967336043715477,
+ "learning_rate": 0.0006,
+ "loss": 4.190918922424316,
+ "step": 6196
+ },
+ {
+ "epoch": 86.06986899563319,
+ "grad_norm": 0.01945611648261547,
+ "learning_rate": 0.0006,
+ "loss": 4.18459415435791,
+ "step": 6197
+ },
+ {
+ "epoch": 86.08384279475983,
+ "grad_norm": 0.018104076385498047,
+ "learning_rate": 0.0006,
+ "loss": 4.0942816734313965,
+ "step": 6198
+ },
+ {
+ "epoch": 86.09781659388646,
+ "grad_norm": 0.01815992221236229,
+ "learning_rate": 0.0006,
+ "loss": 4.011547565460205,
+ "step": 6199
+ },
+ {
+ "epoch": 86.1117903930131,
+ "grad_norm": 0.017574431374669075,
+ "learning_rate": 0.0006,
+ "loss": 4.051593780517578,
+ "step": 6200
+ },
+ {
+ "epoch": 86.12576419213974,
+ "grad_norm": 0.01516508124768734,
+ "learning_rate": 0.0006,
+ "loss": 4.239673614501953,
+ "step": 6201
+ },
+ {
+ "epoch": 86.13973799126637,
+ "grad_norm": 0.015689702704548836,
+ "learning_rate": 0.0006,
+ "loss": 4.1085734367370605,
+ "step": 6202
+ },
+ {
+ "epoch": 86.15371179039302,
+ "grad_norm": 0.01596900261938572,
+ "learning_rate": 0.0006,
+ "loss": 4.1509599685668945,
+ "step": 6203
+ },
+ {
+ "epoch": 86.16768558951965,
+ "grad_norm": 0.01550096832215786,
+ "learning_rate": 0.0006,
+ "loss": 4.160238265991211,
+ "step": 6204
+ },
+ {
+ "epoch": 86.18165938864628,
+ "grad_norm": 0.014601529575884342,
+ "learning_rate": 0.0006,
+ "loss": 4.029219150543213,
+ "step": 6205
+ },
+ {
+ "epoch": 86.19563318777293,
+ "grad_norm": 0.014006740413606167,
+ "learning_rate": 0.0006,
+ "loss": 4.117501258850098,
+ "step": 6206
+ },
+ {
+ "epoch": 86.20960698689956,
+ "grad_norm": 0.013191957026720047,
+ "learning_rate": 0.0006,
+ "loss": 4.252002716064453,
+ "step": 6207
+ },
+ {
+ "epoch": 86.2235807860262,
+ "grad_norm": 0.013031814247369766,
+ "learning_rate": 0.0006,
+ "loss": 4.209517478942871,
+ "step": 6208
+ },
+ {
+ "epoch": 86.23755458515284,
+ "grad_norm": 0.013605714775621891,
+ "learning_rate": 0.0006,
+ "loss": 4.176309585571289,
+ "step": 6209
+ },
+ {
+ "epoch": 86.25152838427948,
+ "grad_norm": 0.012753087095916271,
+ "learning_rate": 0.0006,
+ "loss": 4.097990036010742,
+ "step": 6210
+ },
+ {
+ "epoch": 86.26550218340611,
+ "grad_norm": 0.012940628454089165,
+ "learning_rate": 0.0006,
+ "loss": 4.085589408874512,
+ "step": 6211
+ },
+ {
+ "epoch": 86.27947598253274,
+ "grad_norm": 0.014286619611084461,
+ "learning_rate": 0.0006,
+ "loss": 4.199261665344238,
+ "step": 6212
+ },
+ {
+ "epoch": 86.29344978165939,
+ "grad_norm": 0.014643525704741478,
+ "learning_rate": 0.0006,
+ "loss": 4.23718786239624,
+ "step": 6213
+ },
+ {
+ "epoch": 86.30742358078602,
+ "grad_norm": 0.015182147733867168,
+ "learning_rate": 0.0006,
+ "loss": 4.219855308532715,
+ "step": 6214
+ },
+ {
+ "epoch": 86.32139737991267,
+ "grad_norm": 0.01530102826654911,
+ "learning_rate": 0.0006,
+ "loss": 4.190094947814941,
+ "step": 6215
+ },
+ {
+ "epoch": 86.3353711790393,
+ "grad_norm": 0.015492603182792664,
+ "learning_rate": 0.0006,
+ "loss": 4.153780460357666,
+ "step": 6216
+ },
+ {
+ "epoch": 86.34934497816595,
+ "grad_norm": 0.015339907258749008,
+ "learning_rate": 0.0006,
+ "loss": 4.182865619659424,
+ "step": 6217
+ },
+ {
+ "epoch": 86.36331877729258,
+ "grad_norm": 0.015576880425214767,
+ "learning_rate": 0.0006,
+ "loss": 4.196417331695557,
+ "step": 6218
+ },
+ {
+ "epoch": 86.37729257641921,
+ "grad_norm": 0.014805737882852554,
+ "learning_rate": 0.0006,
+ "loss": 4.265018463134766,
+ "step": 6219
+ },
+ {
+ "epoch": 86.39126637554585,
+ "grad_norm": 0.01597665622830391,
+ "learning_rate": 0.0006,
+ "loss": 4.2864837646484375,
+ "step": 6220
+ },
+ {
+ "epoch": 86.40524017467249,
+ "grad_norm": 0.01693643443286419,
+ "learning_rate": 0.0006,
+ "loss": 4.118192672729492,
+ "step": 6221
+ },
+ {
+ "epoch": 86.41921397379913,
+ "grad_norm": 0.016648951917886734,
+ "learning_rate": 0.0006,
+ "loss": 4.23712682723999,
+ "step": 6222
+ },
+ {
+ "epoch": 86.43318777292576,
+ "grad_norm": 0.014809766784310341,
+ "learning_rate": 0.0006,
+ "loss": 4.251391410827637,
+ "step": 6223
+ },
+ {
+ "epoch": 86.44716157205241,
+ "grad_norm": 0.014333373866975307,
+ "learning_rate": 0.0006,
+ "loss": 4.17243766784668,
+ "step": 6224
+ },
+ {
+ "epoch": 86.46113537117904,
+ "grad_norm": 0.013912711292505264,
+ "learning_rate": 0.0006,
+ "loss": 4.250836372375488,
+ "step": 6225
+ },
+ {
+ "epoch": 86.47510917030567,
+ "grad_norm": 0.014821341261267662,
+ "learning_rate": 0.0006,
+ "loss": 4.073995590209961,
+ "step": 6226
+ },
+ {
+ "epoch": 86.48908296943232,
+ "grad_norm": 0.016634922474622726,
+ "learning_rate": 0.0006,
+ "loss": 4.155867576599121,
+ "step": 6227
+ },
+ {
+ "epoch": 86.50305676855895,
+ "grad_norm": 0.01748695969581604,
+ "learning_rate": 0.0006,
+ "loss": 4.139344692230225,
+ "step": 6228
+ },
+ {
+ "epoch": 86.5170305676856,
+ "grad_norm": 0.01569363661110401,
+ "learning_rate": 0.0006,
+ "loss": 4.23923397064209,
+ "step": 6229
+ },
+ {
+ "epoch": 86.53100436681223,
+ "grad_norm": 0.014037121087312698,
+ "learning_rate": 0.0006,
+ "loss": 4.200334072113037,
+ "step": 6230
+ },
+ {
+ "epoch": 86.54497816593886,
+ "grad_norm": 0.01401279866695404,
+ "learning_rate": 0.0006,
+ "loss": 4.035246849060059,
+ "step": 6231
+ },
+ {
+ "epoch": 86.5589519650655,
+ "grad_norm": 0.014603731222450733,
+ "learning_rate": 0.0006,
+ "loss": 4.141385555267334,
+ "step": 6232
+ },
+ {
+ "epoch": 86.57292576419214,
+ "grad_norm": 0.015113704837858677,
+ "learning_rate": 0.0006,
+ "loss": 4.130824089050293,
+ "step": 6233
+ },
+ {
+ "epoch": 86.58689956331878,
+ "grad_norm": 0.01588696800172329,
+ "learning_rate": 0.0006,
+ "loss": 4.05495023727417,
+ "step": 6234
+ },
+ {
+ "epoch": 86.60087336244541,
+ "grad_norm": 0.01593640074133873,
+ "learning_rate": 0.0006,
+ "loss": 4.22935152053833,
+ "step": 6235
+ },
+ {
+ "epoch": 86.61484716157206,
+ "grad_norm": 0.014292445965111256,
+ "learning_rate": 0.0006,
+ "loss": 4.309839248657227,
+ "step": 6236
+ },
+ {
+ "epoch": 86.62882096069869,
+ "grad_norm": 0.015471610240638256,
+ "learning_rate": 0.0006,
+ "loss": 4.268278121948242,
+ "step": 6237
+ },
+ {
+ "epoch": 86.64279475982532,
+ "grad_norm": 0.016643544659018517,
+ "learning_rate": 0.0006,
+ "loss": 4.135109901428223,
+ "step": 6238
+ },
+ {
+ "epoch": 86.65676855895197,
+ "grad_norm": 0.014733029529452324,
+ "learning_rate": 0.0006,
+ "loss": 4.060197830200195,
+ "step": 6239
+ },
+ {
+ "epoch": 86.6707423580786,
+ "grad_norm": 0.014869216829538345,
+ "learning_rate": 0.0006,
+ "loss": 4.126917362213135,
+ "step": 6240
+ },
+ {
+ "epoch": 86.68471615720524,
+ "grad_norm": 0.01608203910291195,
+ "learning_rate": 0.0006,
+ "loss": 4.183716773986816,
+ "step": 6241
+ },
+ {
+ "epoch": 86.69868995633188,
+ "grad_norm": 0.016178175806999207,
+ "learning_rate": 0.0006,
+ "loss": 4.118875503540039,
+ "step": 6242
+ },
+ {
+ "epoch": 86.71266375545852,
+ "grad_norm": 0.014576597139239311,
+ "learning_rate": 0.0006,
+ "loss": 4.261541366577148,
+ "step": 6243
+ },
+ {
+ "epoch": 86.72663755458515,
+ "grad_norm": 0.014466420747339725,
+ "learning_rate": 0.0006,
+ "loss": 4.307284355163574,
+ "step": 6244
+ },
+ {
+ "epoch": 86.74061135371178,
+ "grad_norm": 0.014042644761502743,
+ "learning_rate": 0.0006,
+ "loss": 4.2729997634887695,
+ "step": 6245
+ },
+ {
+ "epoch": 86.75458515283843,
+ "grad_norm": 0.014589524827897549,
+ "learning_rate": 0.0006,
+ "loss": 4.072981834411621,
+ "step": 6246
+ },
+ {
+ "epoch": 86.76855895196506,
+ "grad_norm": 0.018765483051538467,
+ "learning_rate": 0.0006,
+ "loss": 4.264769554138184,
+ "step": 6247
+ },
+ {
+ "epoch": 86.78253275109171,
+ "grad_norm": 0.019397731870412827,
+ "learning_rate": 0.0006,
+ "loss": 4.183379173278809,
+ "step": 6248
+ },
+ {
+ "epoch": 86.79650655021834,
+ "grad_norm": 0.017525402829051018,
+ "learning_rate": 0.0006,
+ "loss": 4.1494245529174805,
+ "step": 6249
+ },
+ {
+ "epoch": 86.81048034934499,
+ "grad_norm": 0.016310200095176697,
+ "learning_rate": 0.0006,
+ "loss": 4.149104118347168,
+ "step": 6250
+ },
+ {
+ "epoch": 86.82445414847162,
+ "grad_norm": 0.01506423857063055,
+ "learning_rate": 0.0006,
+ "loss": 4.22091817855835,
+ "step": 6251
+ },
+ {
+ "epoch": 86.83842794759825,
+ "grad_norm": 0.014200284145772457,
+ "learning_rate": 0.0006,
+ "loss": 4.177569389343262,
+ "step": 6252
+ },
+ {
+ "epoch": 86.8524017467249,
+ "grad_norm": 0.015016552060842514,
+ "learning_rate": 0.0006,
+ "loss": 4.089859962463379,
+ "step": 6253
+ },
+ {
+ "epoch": 86.86637554585153,
+ "grad_norm": 0.016054971143603325,
+ "learning_rate": 0.0006,
+ "loss": 4.157166004180908,
+ "step": 6254
+ },
+ {
+ "epoch": 86.88034934497817,
+ "grad_norm": 0.014388521201908588,
+ "learning_rate": 0.0006,
+ "loss": 4.1437578201293945,
+ "step": 6255
+ },
+ {
+ "epoch": 86.8943231441048,
+ "grad_norm": 0.015705369412899017,
+ "learning_rate": 0.0006,
+ "loss": 4.045689582824707,
+ "step": 6256
+ },
+ {
+ "epoch": 86.90829694323143,
+ "grad_norm": 0.014192395843565464,
+ "learning_rate": 0.0006,
+ "loss": 4.098793029785156,
+ "step": 6257
+ },
+ {
+ "epoch": 86.92227074235808,
+ "grad_norm": 0.013751034624874592,
+ "learning_rate": 0.0006,
+ "loss": 3.989920139312744,
+ "step": 6258
+ },
+ {
+ "epoch": 86.93624454148471,
+ "grad_norm": 0.01582193188369274,
+ "learning_rate": 0.0006,
+ "loss": 4.283449172973633,
+ "step": 6259
+ },
+ {
+ "epoch": 86.95021834061136,
+ "grad_norm": 0.01476013008505106,
+ "learning_rate": 0.0006,
+ "loss": 4.221401691436768,
+ "step": 6260
+ },
+ {
+ "epoch": 86.96419213973799,
+ "grad_norm": 0.01346337515860796,
+ "learning_rate": 0.0006,
+ "loss": 4.05896520614624,
+ "step": 6261
+ },
+ {
+ "epoch": 86.97816593886463,
+ "grad_norm": 0.012899307534098625,
+ "learning_rate": 0.0006,
+ "loss": 4.159560203552246,
+ "step": 6262
+ },
+ {
+ "epoch": 86.99213973799127,
+ "grad_norm": 0.013150574639439583,
+ "learning_rate": 0.0006,
+ "loss": 4.217330455780029,
+ "step": 6263
+ },
+ {
+ "epoch": 87.0,
+ "grad_norm": 0.014940370805561543,
+ "learning_rate": 0.0006,
+ "loss": 4.1611833572387695,
+ "step": 6264
+ },
+ {
+ "epoch": 87.0,
+ "eval_loss": 4.666043758392334,
+ "eval_runtime": 56.5978,
+ "eval_samples_per_second": 43.147,
+ "eval_steps_per_second": 1.36,
+ "step": 6264
+ },
+ {
+ "epoch": 87.01397379912663,
+ "grad_norm": 0.013721180148422718,
+ "learning_rate": 0.0006,
+ "loss": 4.142454147338867,
+ "step": 6265
+ },
+ {
+ "epoch": 87.02794759825328,
+ "grad_norm": 0.014805255457758904,
+ "learning_rate": 0.0006,
+ "loss": 4.121677875518799,
+ "step": 6266
+ },
+ {
+ "epoch": 87.04192139737991,
+ "grad_norm": 0.015576514415442944,
+ "learning_rate": 0.0006,
+ "loss": 4.1188435554504395,
+ "step": 6267
+ },
+ {
+ "epoch": 87.05589519650655,
+ "grad_norm": 0.01605502888560295,
+ "learning_rate": 0.0006,
+ "loss": 4.255428314208984,
+ "step": 6268
+ },
+ {
+ "epoch": 87.06986899563319,
+ "grad_norm": 0.014933343045413494,
+ "learning_rate": 0.0006,
+ "loss": 4.122596740722656,
+ "step": 6269
+ },
+ {
+ "epoch": 87.08384279475983,
+ "grad_norm": 0.014630894176661968,
+ "learning_rate": 0.0006,
+ "loss": 4.148405075073242,
+ "step": 6270
+ },
+ {
+ "epoch": 87.09781659388646,
+ "grad_norm": 0.014211163856089115,
+ "learning_rate": 0.0006,
+ "loss": 4.058030128479004,
+ "step": 6271
+ },
+ {
+ "epoch": 87.1117903930131,
+ "grad_norm": 0.014060168527066708,
+ "learning_rate": 0.0006,
+ "loss": 4.146586894989014,
+ "step": 6272
+ },
+ {
+ "epoch": 87.12576419213974,
+ "grad_norm": 0.014740100130438805,
+ "learning_rate": 0.0006,
+ "loss": 4.077739715576172,
+ "step": 6273
+ },
+ {
+ "epoch": 87.13973799126637,
+ "grad_norm": 0.013970664702355862,
+ "learning_rate": 0.0006,
+ "loss": 4.038252830505371,
+ "step": 6274
+ },
+ {
+ "epoch": 87.15371179039302,
+ "grad_norm": 0.014399951323866844,
+ "learning_rate": 0.0006,
+ "loss": 4.143700122833252,
+ "step": 6275
+ },
+ {
+ "epoch": 87.16768558951965,
+ "grad_norm": 0.01496410183608532,
+ "learning_rate": 0.0006,
+ "loss": 4.105609893798828,
+ "step": 6276
+ },
+ {
+ "epoch": 87.18165938864628,
+ "grad_norm": 0.014287355355918407,
+ "learning_rate": 0.0006,
+ "loss": 4.173645973205566,
+ "step": 6277
+ },
+ {
+ "epoch": 87.19563318777293,
+ "grad_norm": 0.0141292130574584,
+ "learning_rate": 0.0006,
+ "loss": 4.068511486053467,
+ "step": 6278
+ },
+ {
+ "epoch": 87.20960698689956,
+ "grad_norm": 0.01393523346632719,
+ "learning_rate": 0.0006,
+ "loss": 4.060687065124512,
+ "step": 6279
+ },
+ {
+ "epoch": 87.2235807860262,
+ "grad_norm": 0.01431021187454462,
+ "learning_rate": 0.0006,
+ "loss": 4.219209671020508,
+ "step": 6280
+ },
+ {
+ "epoch": 87.23755458515284,
+ "grad_norm": 0.014856027439236641,
+ "learning_rate": 0.0006,
+ "loss": 4.146749973297119,
+ "step": 6281
+ },
+ {
+ "epoch": 87.25152838427948,
+ "grad_norm": 0.013677539303898811,
+ "learning_rate": 0.0006,
+ "loss": 3.9859797954559326,
+ "step": 6282
+ },
+ {
+ "epoch": 87.26550218340611,
+ "grad_norm": 0.014078967273235321,
+ "learning_rate": 0.0006,
+ "loss": 4.129668235778809,
+ "step": 6283
+ },
+ {
+ "epoch": 87.27947598253274,
+ "grad_norm": 0.015564394183456898,
+ "learning_rate": 0.0006,
+ "loss": 4.0007829666137695,
+ "step": 6284
+ },
+ {
+ "epoch": 87.29344978165939,
+ "grad_norm": 0.015077749267220497,
+ "learning_rate": 0.0006,
+ "loss": 4.112264156341553,
+ "step": 6285
+ },
+ {
+ "epoch": 87.30742358078602,
+ "grad_norm": 0.015341218560934067,
+ "learning_rate": 0.0006,
+ "loss": 4.112249374389648,
+ "step": 6286
+ },
+ {
+ "epoch": 87.32139737991267,
+ "grad_norm": 0.016399366781115532,
+ "learning_rate": 0.0006,
+ "loss": 4.176196098327637,
+ "step": 6287
+ },
+ {
+ "epoch": 87.3353711790393,
+ "grad_norm": 0.01599763333797455,
+ "learning_rate": 0.0006,
+ "loss": 4.163441181182861,
+ "step": 6288
+ },
+ {
+ "epoch": 87.34934497816595,
+ "grad_norm": 0.015439776703715324,
+ "learning_rate": 0.0006,
+ "loss": 4.11902379989624,
+ "step": 6289
+ },
+ {
+ "epoch": 87.36331877729258,
+ "grad_norm": 0.015909943729639053,
+ "learning_rate": 0.0006,
+ "loss": 4.16621732711792,
+ "step": 6290
+ },
+ {
+ "epoch": 87.37729257641921,
+ "grad_norm": 0.016710078343749046,
+ "learning_rate": 0.0006,
+ "loss": 4.044960975646973,
+ "step": 6291
+ },
+ {
+ "epoch": 87.39126637554585,
+ "grad_norm": 0.015883933752775192,
+ "learning_rate": 0.0006,
+ "loss": 4.065184593200684,
+ "step": 6292
+ },
+ {
+ "epoch": 87.40524017467249,
+ "grad_norm": 0.013912747614085674,
+ "learning_rate": 0.0006,
+ "loss": 4.102450370788574,
+ "step": 6293
+ },
+ {
+ "epoch": 87.41921397379913,
+ "grad_norm": 0.013170420192182064,
+ "learning_rate": 0.0006,
+ "loss": 4.234364032745361,
+ "step": 6294
+ },
+ {
+ "epoch": 87.43318777292576,
+ "grad_norm": 0.014043302275240421,
+ "learning_rate": 0.0006,
+ "loss": 4.129565238952637,
+ "step": 6295
+ },
+ {
+ "epoch": 87.44716157205241,
+ "grad_norm": 0.015010139904916286,
+ "learning_rate": 0.0006,
+ "loss": 4.166678428649902,
+ "step": 6296
+ },
+ {
+ "epoch": 87.46113537117904,
+ "grad_norm": 0.014721784740686417,
+ "learning_rate": 0.0006,
+ "loss": 4.182618618011475,
+ "step": 6297
+ },
+ {
+ "epoch": 87.47510917030567,
+ "grad_norm": 0.015713468194007874,
+ "learning_rate": 0.0006,
+ "loss": 4.103855133056641,
+ "step": 6298
+ },
+ {
+ "epoch": 87.48908296943232,
+ "grad_norm": 0.017204247415065765,
+ "learning_rate": 0.0006,
+ "loss": 4.140937328338623,
+ "step": 6299
+ },
+ {
+ "epoch": 87.50305676855895,
+ "grad_norm": 0.016856668516993523,
+ "learning_rate": 0.0006,
+ "loss": 4.147859573364258,
+ "step": 6300
+ },
+ {
+ "epoch": 87.5170305676856,
+ "grad_norm": 0.017205018550157547,
+ "learning_rate": 0.0006,
+ "loss": 4.1667327880859375,
+ "step": 6301
+ },
+ {
+ "epoch": 87.53100436681223,
+ "grad_norm": 0.014821539632976055,
+ "learning_rate": 0.0006,
+ "loss": 4.089935302734375,
+ "step": 6302
+ },
+ {
+ "epoch": 87.54497816593886,
+ "grad_norm": 0.01446484588086605,
+ "learning_rate": 0.0006,
+ "loss": 4.243657112121582,
+ "step": 6303
+ },
+ {
+ "epoch": 87.5589519650655,
+ "grad_norm": 0.014374484308063984,
+ "learning_rate": 0.0006,
+ "loss": 4.14356803894043,
+ "step": 6304
+ },
+ {
+ "epoch": 87.57292576419214,
+ "grad_norm": 0.01420109998434782,
+ "learning_rate": 0.0006,
+ "loss": 4.1273651123046875,
+ "step": 6305
+ },
+ {
+ "epoch": 87.58689956331878,
+ "grad_norm": 0.014810388907790184,
+ "learning_rate": 0.0006,
+ "loss": 4.222290515899658,
+ "step": 6306
+ },
+ {
+ "epoch": 87.60087336244541,
+ "grad_norm": 0.015826983377337456,
+ "learning_rate": 0.0006,
+ "loss": 4.137996196746826,
+ "step": 6307
+ },
+ {
+ "epoch": 87.61484716157206,
+ "grad_norm": 0.014457454904913902,
+ "learning_rate": 0.0006,
+ "loss": 4.0780744552612305,
+ "step": 6308
+ },
+ {
+ "epoch": 87.62882096069869,
+ "grad_norm": 0.014458604156970978,
+ "learning_rate": 0.0006,
+ "loss": 4.1287922859191895,
+ "step": 6309
+ },
+ {
+ "epoch": 87.64279475982532,
+ "grad_norm": 0.015282071195542812,
+ "learning_rate": 0.0006,
+ "loss": 4.189508438110352,
+ "step": 6310
+ },
+ {
+ "epoch": 87.65676855895197,
+ "grad_norm": 0.017536642029881477,
+ "learning_rate": 0.0006,
+ "loss": 4.070173263549805,
+ "step": 6311
+ },
+ {
+ "epoch": 87.6707423580786,
+ "grad_norm": 0.017303580418229103,
+ "learning_rate": 0.0006,
+ "loss": 4.119133949279785,
+ "step": 6312
+ },
+ {
+ "epoch": 87.68471615720524,
+ "grad_norm": 0.016148580238223076,
+ "learning_rate": 0.0006,
+ "loss": 3.994354486465454,
+ "step": 6313
+ },
+ {
+ "epoch": 87.69868995633188,
+ "grad_norm": 0.01706942543387413,
+ "learning_rate": 0.0006,
+ "loss": 4.183113098144531,
+ "step": 6314
+ },
+ {
+ "epoch": 87.71266375545852,
+ "grad_norm": 0.016971010714769363,
+ "learning_rate": 0.0006,
+ "loss": 4.104086875915527,
+ "step": 6315
+ },
+ {
+ "epoch": 87.72663755458515,
+ "grad_norm": 0.015516847372055054,
+ "learning_rate": 0.0006,
+ "loss": 4.2409772872924805,
+ "step": 6316
+ },
+ {
+ "epoch": 87.74061135371178,
+ "grad_norm": 0.015352269634604454,
+ "learning_rate": 0.0006,
+ "loss": 4.026427268981934,
+ "step": 6317
+ },
+ {
+ "epoch": 87.75458515283843,
+ "grad_norm": 0.017702415585517883,
+ "learning_rate": 0.0006,
+ "loss": 4.177840232849121,
+ "step": 6318
+ },
+ {
+ "epoch": 87.76855895196506,
+ "grad_norm": 0.01629541628062725,
+ "learning_rate": 0.0006,
+ "loss": 4.229039192199707,
+ "step": 6319
+ },
+ {
+ "epoch": 87.78253275109171,
+ "grad_norm": 0.01503711100667715,
+ "learning_rate": 0.0006,
+ "loss": 4.214350700378418,
+ "step": 6320
+ },
+ {
+ "epoch": 87.79650655021834,
+ "grad_norm": 0.014913883060216904,
+ "learning_rate": 0.0006,
+ "loss": 4.168880939483643,
+ "step": 6321
+ },
+ {
+ "epoch": 87.81048034934499,
+ "grad_norm": 0.014513909816741943,
+ "learning_rate": 0.0006,
+ "loss": 4.162267684936523,
+ "step": 6322
+ },
+ {
+ "epoch": 87.82445414847162,
+ "grad_norm": 0.014635873027145863,
+ "learning_rate": 0.0006,
+ "loss": 4.220067977905273,
+ "step": 6323
+ },
+ {
+ "epoch": 87.83842794759825,
+ "grad_norm": 0.017102796584367752,
+ "learning_rate": 0.0006,
+ "loss": 4.1493377685546875,
+ "step": 6324
+ },
+ {
+ "epoch": 87.8524017467249,
+ "grad_norm": 0.015533884055912495,
+ "learning_rate": 0.0006,
+ "loss": 4.0646209716796875,
+ "step": 6325
+ },
+ {
+ "epoch": 87.86637554585153,
+ "grad_norm": 0.014617550186812878,
+ "learning_rate": 0.0006,
+ "loss": 4.1390533447265625,
+ "step": 6326
+ },
+ {
+ "epoch": 87.88034934497817,
+ "grad_norm": 0.014142967760562897,
+ "learning_rate": 0.0006,
+ "loss": 4.133934020996094,
+ "step": 6327
+ },
+ {
+ "epoch": 87.8943231441048,
+ "grad_norm": 0.013182433322072029,
+ "learning_rate": 0.0006,
+ "loss": 4.125327110290527,
+ "step": 6328
+ },
+ {
+ "epoch": 87.90829694323143,
+ "grad_norm": 0.013928229920566082,
+ "learning_rate": 0.0006,
+ "loss": 4.142777442932129,
+ "step": 6329
+ },
+ {
+ "epoch": 87.92227074235808,
+ "grad_norm": 0.015337013639509678,
+ "learning_rate": 0.0006,
+ "loss": 4.140210151672363,
+ "step": 6330
+ },
+ {
+ "epoch": 87.93624454148471,
+ "grad_norm": 0.014873536303639412,
+ "learning_rate": 0.0006,
+ "loss": 4.102636337280273,
+ "step": 6331
+ },
+ {
+ "epoch": 87.95021834061136,
+ "grad_norm": 0.01387431938201189,
+ "learning_rate": 0.0006,
+ "loss": 4.149059295654297,
+ "step": 6332
+ },
+ {
+ "epoch": 87.96419213973799,
+ "grad_norm": 0.013234943151473999,
+ "learning_rate": 0.0006,
+ "loss": 4.1564459800720215,
+ "step": 6333
+ },
+ {
+ "epoch": 87.97816593886463,
+ "grad_norm": 0.013781017623841763,
+ "learning_rate": 0.0006,
+ "loss": 4.189839839935303,
+ "step": 6334
+ },
+ {
+ "epoch": 87.99213973799127,
+ "grad_norm": 0.013733026571571827,
+ "learning_rate": 0.0006,
+ "loss": 4.29277229309082,
+ "step": 6335
+ },
+ {
+ "epoch": 88.0,
+ "grad_norm": 0.014752327464520931,
+ "learning_rate": 0.0006,
+ "loss": 4.163867950439453,
+ "step": 6336
+ },
+ {
+ "epoch": 88.0,
+ "eval_loss": 4.535566329956055,
+ "eval_runtime": 56.4677,
+ "eval_samples_per_second": 43.246,
+ "eval_steps_per_second": 1.364,
+ "step": 6336
+ },
+ {
+ "epoch": 88.01397379912663,
+ "grad_norm": 0.013958732597529888,
+ "learning_rate": 0.0006,
+ "loss": 4.200681209564209,
+ "step": 6337
+ },
+ {
+ "epoch": 88.02794759825328,
+ "grad_norm": 0.015283545479178429,
+ "learning_rate": 0.0006,
+ "loss": 4.1736016273498535,
+ "step": 6338
+ },
+ {
+ "epoch": 88.04192139737991,
+ "grad_norm": 0.016192059963941574,
+ "learning_rate": 0.0006,
+ "loss": 4.040589809417725,
+ "step": 6339
+ },
+ {
+ "epoch": 88.05589519650655,
+ "grad_norm": 0.018096154555678368,
+ "learning_rate": 0.0006,
+ "loss": 4.146696090698242,
+ "step": 6340
+ },
+ {
+ "epoch": 88.06986899563319,
+ "grad_norm": 0.01797664538025856,
+ "learning_rate": 0.0006,
+ "loss": 4.007248878479004,
+ "step": 6341
+ },
+ {
+ "epoch": 88.08384279475983,
+ "grad_norm": 0.016275612637400627,
+ "learning_rate": 0.0006,
+ "loss": 4.193206787109375,
+ "step": 6342
+ },
+ {
+ "epoch": 88.09781659388646,
+ "grad_norm": 0.014210258610546589,
+ "learning_rate": 0.0006,
+ "loss": 4.064548492431641,
+ "step": 6343
+ },
+ {
+ "epoch": 88.1117903930131,
+ "grad_norm": 0.016085151582956314,
+ "learning_rate": 0.0006,
+ "loss": 4.003668785095215,
+ "step": 6344
+ },
+ {
+ "epoch": 88.12576419213974,
+ "grad_norm": 0.01930483989417553,
+ "learning_rate": 0.0006,
+ "loss": 4.236578941345215,
+ "step": 6345
+ },
+ {
+ "epoch": 88.13973799126637,
+ "grad_norm": 0.017140906304121017,
+ "learning_rate": 0.0006,
+ "loss": 3.9593005180358887,
+ "step": 6346
+ },
+ {
+ "epoch": 88.15371179039302,
+ "grad_norm": 0.01521560549736023,
+ "learning_rate": 0.0006,
+ "loss": 4.137644290924072,
+ "step": 6347
+ },
+ {
+ "epoch": 88.16768558951965,
+ "grad_norm": 0.015788141638040543,
+ "learning_rate": 0.0006,
+ "loss": 4.124845027923584,
+ "step": 6348
+ },
+ {
+ "epoch": 88.18165938864628,
+ "grad_norm": 0.01784597896039486,
+ "learning_rate": 0.0006,
+ "loss": 4.139750957489014,
+ "step": 6349
+ },
+ {
+ "epoch": 88.19563318777293,
+ "grad_norm": 0.016266558319330215,
+ "learning_rate": 0.0006,
+ "loss": 4.074851989746094,
+ "step": 6350
+ },
+ {
+ "epoch": 88.20960698689956,
+ "grad_norm": 0.014024483971297741,
+ "learning_rate": 0.0006,
+ "loss": 4.164689064025879,
+ "step": 6351
+ },
+ {
+ "epoch": 88.2235807860262,
+ "grad_norm": 0.015694420784711838,
+ "learning_rate": 0.0006,
+ "loss": 4.219979286193848,
+ "step": 6352
+ },
+ {
+ "epoch": 88.23755458515284,
+ "grad_norm": 0.014897859655320644,
+ "learning_rate": 0.0006,
+ "loss": 4.050738334655762,
+ "step": 6353
+ },
+ {
+ "epoch": 88.25152838427948,
+ "grad_norm": 0.014827374368906021,
+ "learning_rate": 0.0006,
+ "loss": 4.137135028839111,
+ "step": 6354
+ },
+ {
+ "epoch": 88.26550218340611,
+ "grad_norm": 0.01454879529774189,
+ "learning_rate": 0.0006,
+ "loss": 3.9980897903442383,
+ "step": 6355
+ },
+ {
+ "epoch": 88.27947598253274,
+ "grad_norm": 0.015623947605490685,
+ "learning_rate": 0.0006,
+ "loss": 4.226551055908203,
+ "step": 6356
+ },
+ {
+ "epoch": 88.29344978165939,
+ "grad_norm": 0.015527035109698772,
+ "learning_rate": 0.0006,
+ "loss": 4.11053466796875,
+ "step": 6357
+ },
+ {
+ "epoch": 88.30742358078602,
+ "grad_norm": 0.015242574736475945,
+ "learning_rate": 0.0006,
+ "loss": 4.105966567993164,
+ "step": 6358
+ },
+ {
+ "epoch": 88.32139737991267,
+ "grad_norm": 0.01375538855791092,
+ "learning_rate": 0.0006,
+ "loss": 4.148453712463379,
+ "step": 6359
+ },
+ {
+ "epoch": 88.3353711790393,
+ "grad_norm": 0.01595226116478443,
+ "learning_rate": 0.0006,
+ "loss": 4.095836162567139,
+ "step": 6360
+ },
+ {
+ "epoch": 88.34934497816595,
+ "grad_norm": 0.015252740122377872,
+ "learning_rate": 0.0006,
+ "loss": 4.136335849761963,
+ "step": 6361
+ },
+ {
+ "epoch": 88.36331877729258,
+ "grad_norm": 0.014405866153538227,
+ "learning_rate": 0.0006,
+ "loss": 4.084691047668457,
+ "step": 6362
+ },
+ {
+ "epoch": 88.37729257641921,
+ "grad_norm": 0.014755490235984325,
+ "learning_rate": 0.0006,
+ "loss": 4.095475673675537,
+ "step": 6363
+ },
+ {
+ "epoch": 88.39126637554585,
+ "grad_norm": 0.015339708887040615,
+ "learning_rate": 0.0006,
+ "loss": 4.163136959075928,
+ "step": 6364
+ },
+ {
+ "epoch": 88.40524017467249,
+ "grad_norm": 0.014667164534330368,
+ "learning_rate": 0.0006,
+ "loss": 4.062265396118164,
+ "step": 6365
+ },
+ {
+ "epoch": 88.41921397379913,
+ "grad_norm": 0.014588423073291779,
+ "learning_rate": 0.0006,
+ "loss": 4.13359260559082,
+ "step": 6366
+ },
+ {
+ "epoch": 88.43318777292576,
+ "grad_norm": 0.014005377888679504,
+ "learning_rate": 0.0006,
+ "loss": 4.219711780548096,
+ "step": 6367
+ },
+ {
+ "epoch": 88.44716157205241,
+ "grad_norm": 0.012769930064678192,
+ "learning_rate": 0.0006,
+ "loss": 4.114658355712891,
+ "step": 6368
+ },
+ {
+ "epoch": 88.46113537117904,
+ "grad_norm": 0.013632693327963352,
+ "learning_rate": 0.0006,
+ "loss": 4.0675225257873535,
+ "step": 6369
+ },
+ {
+ "epoch": 88.47510917030567,
+ "grad_norm": 0.013829804956912994,
+ "learning_rate": 0.0006,
+ "loss": 4.1125969886779785,
+ "step": 6370
+ },
+ {
+ "epoch": 88.48908296943232,
+ "grad_norm": 0.013882109895348549,
+ "learning_rate": 0.0006,
+ "loss": 4.1627092361450195,
+ "step": 6371
+ },
+ {
+ "epoch": 88.50305676855895,
+ "grad_norm": 0.014023828320205212,
+ "learning_rate": 0.0006,
+ "loss": 4.01038932800293,
+ "step": 6372
+ },
+ {
+ "epoch": 88.5170305676856,
+ "grad_norm": 0.014840706251561642,
+ "learning_rate": 0.0006,
+ "loss": 4.117877960205078,
+ "step": 6373
+ },
+ {
+ "epoch": 88.53100436681223,
+ "grad_norm": 0.014165930449962616,
+ "learning_rate": 0.0006,
+ "loss": 4.1864237785339355,
+ "step": 6374
+ },
+ {
+ "epoch": 88.54497816593886,
+ "grad_norm": 0.014390201307833195,
+ "learning_rate": 0.0006,
+ "loss": 4.1044602394104,
+ "step": 6375
+ },
+ {
+ "epoch": 88.5589519650655,
+ "grad_norm": 0.015037395060062408,
+ "learning_rate": 0.0006,
+ "loss": 4.101107120513916,
+ "step": 6376
+ },
+ {
+ "epoch": 88.57292576419214,
+ "grad_norm": 0.01671452447772026,
+ "learning_rate": 0.0006,
+ "loss": 4.170581817626953,
+ "step": 6377
+ },
+ {
+ "epoch": 88.58689956331878,
+ "grad_norm": 0.01993076130747795,
+ "learning_rate": 0.0006,
+ "loss": 4.261412620544434,
+ "step": 6378
+ },
+ {
+ "epoch": 88.60087336244541,
+ "grad_norm": 0.023721016943454742,
+ "learning_rate": 0.0006,
+ "loss": 4.199363708496094,
+ "step": 6379
+ },
+ {
+ "epoch": 88.61484716157206,
+ "grad_norm": 0.023726163432002068,
+ "learning_rate": 0.0006,
+ "loss": 4.061285018920898,
+ "step": 6380
+ },
+ {
+ "epoch": 88.62882096069869,
+ "grad_norm": 0.018352877348661423,
+ "learning_rate": 0.0006,
+ "loss": 4.0721282958984375,
+ "step": 6381
+ },
+ {
+ "epoch": 88.64279475982532,
+ "grad_norm": 0.01679830253124237,
+ "learning_rate": 0.0006,
+ "loss": 4.1717939376831055,
+ "step": 6382
+ },
+ {
+ "epoch": 88.65676855895197,
+ "grad_norm": 0.017193637788295746,
+ "learning_rate": 0.0006,
+ "loss": 4.230783939361572,
+ "step": 6383
+ },
+ {
+ "epoch": 88.6707423580786,
+ "grad_norm": 0.01581464521586895,
+ "learning_rate": 0.0006,
+ "loss": 4.029906272888184,
+ "step": 6384
+ },
+ {
+ "epoch": 88.68471615720524,
+ "grad_norm": 0.014049704186618328,
+ "learning_rate": 0.0006,
+ "loss": 4.192776679992676,
+ "step": 6385
+ },
+ {
+ "epoch": 88.69868995633188,
+ "grad_norm": 0.017214450985193253,
+ "learning_rate": 0.0006,
+ "loss": 4.248916149139404,
+ "step": 6386
+ },
+ {
+ "epoch": 88.71266375545852,
+ "grad_norm": 0.0173103716224432,
+ "learning_rate": 0.0006,
+ "loss": 4.124347686767578,
+ "step": 6387
+ },
+ {
+ "epoch": 88.72663755458515,
+ "grad_norm": 0.015745092183351517,
+ "learning_rate": 0.0006,
+ "loss": 4.192878246307373,
+ "step": 6388
+ },
+ {
+ "epoch": 88.74061135371178,
+ "grad_norm": 0.013750772923231125,
+ "learning_rate": 0.0006,
+ "loss": 4.1564483642578125,
+ "step": 6389
+ },
+ {
+ "epoch": 88.75458515283843,
+ "grad_norm": 0.014969858340919018,
+ "learning_rate": 0.0006,
+ "loss": 4.0980424880981445,
+ "step": 6390
+ },
+ {
+ "epoch": 88.76855895196506,
+ "grad_norm": 0.01287173479795456,
+ "learning_rate": 0.0006,
+ "loss": 4.145712852478027,
+ "step": 6391
+ },
+ {
+ "epoch": 88.78253275109171,
+ "grad_norm": 0.012716300785541534,
+ "learning_rate": 0.0006,
+ "loss": 4.214226722717285,
+ "step": 6392
+ },
+ {
+ "epoch": 88.79650655021834,
+ "grad_norm": 0.013855204917490482,
+ "learning_rate": 0.0006,
+ "loss": 4.1667256355285645,
+ "step": 6393
+ },
+ {
+ "epoch": 88.81048034934499,
+ "grad_norm": 0.013650864362716675,
+ "learning_rate": 0.0006,
+ "loss": 4.161942481994629,
+ "step": 6394
+ },
+ {
+ "epoch": 88.82445414847162,
+ "grad_norm": 0.013780126348137856,
+ "learning_rate": 0.0006,
+ "loss": 4.087825775146484,
+ "step": 6395
+ },
+ {
+ "epoch": 88.83842794759825,
+ "grad_norm": 0.014029948972165585,
+ "learning_rate": 0.0006,
+ "loss": 4.142813205718994,
+ "step": 6396
+ },
+ {
+ "epoch": 88.8524017467249,
+ "grad_norm": 0.01322692260146141,
+ "learning_rate": 0.0006,
+ "loss": 4.106776714324951,
+ "step": 6397
+ },
+ {
+ "epoch": 88.86637554585153,
+ "grad_norm": 0.013522760011255741,
+ "learning_rate": 0.0006,
+ "loss": 4.090996265411377,
+ "step": 6398
+ },
+ {
+ "epoch": 88.88034934497817,
+ "grad_norm": 0.014474684372544289,
+ "learning_rate": 0.0006,
+ "loss": 4.124624252319336,
+ "step": 6399
+ },
+ {
+ "epoch": 88.8943231441048,
+ "grad_norm": 0.014490422792732716,
+ "learning_rate": 0.0006,
+ "loss": 4.065431594848633,
+ "step": 6400
+ },
+ {
+ "epoch": 88.90829694323143,
+ "grad_norm": 0.015379710122942924,
+ "learning_rate": 0.0006,
+ "loss": 4.207939624786377,
+ "step": 6401
+ },
+ {
+ "epoch": 88.92227074235808,
+ "grad_norm": 0.015914618968963623,
+ "learning_rate": 0.0006,
+ "loss": 4.041152000427246,
+ "step": 6402
+ },
+ {
+ "epoch": 88.93624454148471,
+ "grad_norm": 0.01480048056691885,
+ "learning_rate": 0.0006,
+ "loss": 4.138754844665527,
+ "step": 6403
+ },
+ {
+ "epoch": 88.95021834061136,
+ "grad_norm": 0.013734851032495499,
+ "learning_rate": 0.0006,
+ "loss": 4.25323486328125,
+ "step": 6404
+ },
+ {
+ "epoch": 88.96419213973799,
+ "grad_norm": 0.01452311035245657,
+ "learning_rate": 0.0006,
+ "loss": 4.115228176116943,
+ "step": 6405
+ },
+ {
+ "epoch": 88.97816593886463,
+ "grad_norm": 0.016197681427001953,
+ "learning_rate": 0.0006,
+ "loss": 4.099531173706055,
+ "step": 6406
+ },
+ {
+ "epoch": 88.99213973799127,
+ "grad_norm": 0.017169523984193802,
+ "learning_rate": 0.0006,
+ "loss": 4.182981491088867,
+ "step": 6407
+ },
+ {
+ "epoch": 89.0,
+ "grad_norm": 0.017855646088719368,
+ "learning_rate": 0.0006,
+ "loss": 4.1470136642456055,
+ "step": 6408
+ },
+ {
+ "epoch": 89.0,
+ "eval_loss": 4.6035614013671875,
+ "eval_runtime": 57.0625,
+ "eval_samples_per_second": 42.795,
+ "eval_steps_per_second": 1.349,
+ "step": 6408
+ },
+ {
+ "epoch": 89.01397379912663,
+ "grad_norm": 0.014834101311862469,
+ "learning_rate": 0.0006,
+ "loss": 4.159243583679199,
+ "step": 6409
+ },
+ {
+ "epoch": 89.02794759825328,
+ "grad_norm": 0.015316772274672985,
+ "learning_rate": 0.0006,
+ "loss": 4.0709123611450195,
+ "step": 6410
+ },
+ {
+ "epoch": 89.04192139737991,
+ "grad_norm": 0.01576591096818447,
+ "learning_rate": 0.0006,
+ "loss": 4.0921196937561035,
+ "step": 6411
+ },
+ {
+ "epoch": 89.05589519650655,
+ "grad_norm": 0.015948045998811722,
+ "learning_rate": 0.0006,
+ "loss": 4.181877136230469,
+ "step": 6412
+ },
+ {
+ "epoch": 89.06986899563319,
+ "grad_norm": 0.016612380743026733,
+ "learning_rate": 0.0006,
+ "loss": 4.081249237060547,
+ "step": 6413
+ },
+ {
+ "epoch": 89.08384279475983,
+ "grad_norm": 0.017455685883760452,
+ "learning_rate": 0.0006,
+ "loss": 4.175990104675293,
+ "step": 6414
+ },
+ {
+ "epoch": 89.09781659388646,
+ "grad_norm": 0.015040521509945393,
+ "learning_rate": 0.0006,
+ "loss": 4.121378421783447,
+ "step": 6415
+ },
+ {
+ "epoch": 89.1117903930131,
+ "grad_norm": 0.014513209462165833,
+ "learning_rate": 0.0006,
+ "loss": 3.996107816696167,
+ "step": 6416
+ },
+ {
+ "epoch": 89.12576419213974,
+ "grad_norm": 0.015455638989806175,
+ "learning_rate": 0.0006,
+ "loss": 4.073675155639648,
+ "step": 6417
+ },
+ {
+ "epoch": 89.13973799126637,
+ "grad_norm": 0.015468129888176918,
+ "learning_rate": 0.0006,
+ "loss": 4.130242824554443,
+ "step": 6418
+ },
+ {
+ "epoch": 89.15371179039302,
+ "grad_norm": 0.015284297987818718,
+ "learning_rate": 0.0006,
+ "loss": 4.18555212020874,
+ "step": 6419
+ },
+ {
+ "epoch": 89.16768558951965,
+ "grad_norm": 0.01476874016225338,
+ "learning_rate": 0.0006,
+ "loss": 4.134175777435303,
+ "step": 6420
+ },
+ {
+ "epoch": 89.18165938864628,
+ "grad_norm": 0.015836201608181,
+ "learning_rate": 0.0006,
+ "loss": 4.094529151916504,
+ "step": 6421
+ },
+ {
+ "epoch": 89.19563318777293,
+ "grad_norm": 0.0161928478628397,
+ "learning_rate": 0.0006,
+ "loss": 4.12647819519043,
+ "step": 6422
+ },
+ {
+ "epoch": 89.20960698689956,
+ "grad_norm": 0.015680814161896706,
+ "learning_rate": 0.0006,
+ "loss": 3.9620392322540283,
+ "step": 6423
+ },
+ {
+ "epoch": 89.2235807860262,
+ "grad_norm": 0.01584681309759617,
+ "learning_rate": 0.0006,
+ "loss": 4.049304008483887,
+ "step": 6424
+ },
+ {
+ "epoch": 89.23755458515284,
+ "grad_norm": 0.01452907919883728,
+ "learning_rate": 0.0006,
+ "loss": 4.064027786254883,
+ "step": 6425
+ },
+ {
+ "epoch": 89.25152838427948,
+ "grad_norm": 0.014002552255988121,
+ "learning_rate": 0.0006,
+ "loss": 4.151862621307373,
+ "step": 6426
+ },
+ {
+ "epoch": 89.26550218340611,
+ "grad_norm": 0.014658625237643719,
+ "learning_rate": 0.0006,
+ "loss": 4.1672210693359375,
+ "step": 6427
+ },
+ {
+ "epoch": 89.27947598253274,
+ "grad_norm": 0.015021766535937786,
+ "learning_rate": 0.0006,
+ "loss": 3.9742586612701416,
+ "step": 6428
+ },
+ {
+ "epoch": 89.29344978165939,
+ "grad_norm": 0.014681321568787098,
+ "learning_rate": 0.0006,
+ "loss": 4.024159908294678,
+ "step": 6429
+ },
+ {
+ "epoch": 89.30742358078602,
+ "grad_norm": 0.014161549508571625,
+ "learning_rate": 0.0006,
+ "loss": 4.105640888214111,
+ "step": 6430
+ },
+ {
+ "epoch": 89.32139737991267,
+ "grad_norm": 0.014801453799009323,
+ "learning_rate": 0.0006,
+ "loss": 4.123049736022949,
+ "step": 6431
+ },
+ {
+ "epoch": 89.3353711790393,
+ "grad_norm": 0.014307181350886822,
+ "learning_rate": 0.0006,
+ "loss": 4.086954593658447,
+ "step": 6432
+ },
+ {
+ "epoch": 89.34934497816595,
+ "grad_norm": 0.013299369253218174,
+ "learning_rate": 0.0006,
+ "loss": 4.109533786773682,
+ "step": 6433
+ },
+ {
+ "epoch": 89.36331877729258,
+ "grad_norm": 0.014388669282197952,
+ "learning_rate": 0.0006,
+ "loss": 4.222370147705078,
+ "step": 6434
+ },
+ {
+ "epoch": 89.37729257641921,
+ "grad_norm": 0.014501879923045635,
+ "learning_rate": 0.0006,
+ "loss": 4.224178314208984,
+ "step": 6435
+ },
+ {
+ "epoch": 89.39126637554585,
+ "grad_norm": 0.014109021984040737,
+ "learning_rate": 0.0006,
+ "loss": 4.112702369689941,
+ "step": 6436
+ },
+ {
+ "epoch": 89.40524017467249,
+ "grad_norm": 0.013521002605557442,
+ "learning_rate": 0.0006,
+ "loss": 4.010124683380127,
+ "step": 6437
+ },
+ {
+ "epoch": 89.41921397379913,
+ "grad_norm": 0.013288144953548908,
+ "learning_rate": 0.0006,
+ "loss": 4.127592086791992,
+ "step": 6438
+ },
+ {
+ "epoch": 89.43318777292576,
+ "grad_norm": 0.013882486149668694,
+ "learning_rate": 0.0006,
+ "loss": 4.031871318817139,
+ "step": 6439
+ },
+ {
+ "epoch": 89.44716157205241,
+ "grad_norm": 0.01289335172623396,
+ "learning_rate": 0.0006,
+ "loss": 4.0757551193237305,
+ "step": 6440
+ },
+ {
+ "epoch": 89.46113537117904,
+ "grad_norm": 0.01356302760541439,
+ "learning_rate": 0.0006,
+ "loss": 4.075678825378418,
+ "step": 6441
+ },
+ {
+ "epoch": 89.47510917030567,
+ "grad_norm": 0.01447434164583683,
+ "learning_rate": 0.0006,
+ "loss": 4.093453407287598,
+ "step": 6442
+ },
+ {
+ "epoch": 89.48908296943232,
+ "grad_norm": 0.014443481341004372,
+ "learning_rate": 0.0006,
+ "loss": 4.075372219085693,
+ "step": 6443
+ },
+ {
+ "epoch": 89.50305676855895,
+ "grad_norm": 0.01314218994230032,
+ "learning_rate": 0.0006,
+ "loss": 4.171533584594727,
+ "step": 6444
+ },
+ {
+ "epoch": 89.5170305676856,
+ "grad_norm": 0.01348346658051014,
+ "learning_rate": 0.0006,
+ "loss": 4.227236270904541,
+ "step": 6445
+ },
+ {
+ "epoch": 89.53100436681223,
+ "grad_norm": 0.013370679691433907,
+ "learning_rate": 0.0006,
+ "loss": 3.973062515258789,
+ "step": 6446
+ },
+ {
+ "epoch": 89.54497816593886,
+ "grad_norm": 0.01410642359405756,
+ "learning_rate": 0.0006,
+ "loss": 4.231175422668457,
+ "step": 6447
+ },
+ {
+ "epoch": 89.5589519650655,
+ "grad_norm": 0.01419337373226881,
+ "learning_rate": 0.0006,
+ "loss": 4.145868301391602,
+ "step": 6448
+ },
+ {
+ "epoch": 89.57292576419214,
+ "grad_norm": 0.013691053725779057,
+ "learning_rate": 0.0006,
+ "loss": 4.175131797790527,
+ "step": 6449
+ },
+ {
+ "epoch": 89.58689956331878,
+ "grad_norm": 0.013714492321014404,
+ "learning_rate": 0.0006,
+ "loss": 4.251893997192383,
+ "step": 6450
+ },
+ {
+ "epoch": 89.60087336244541,
+ "grad_norm": 0.01393197476863861,
+ "learning_rate": 0.0006,
+ "loss": 4.047079086303711,
+ "step": 6451
+ },
+ {
+ "epoch": 89.61484716157206,
+ "grad_norm": 0.013957309536635876,
+ "learning_rate": 0.0006,
+ "loss": 4.128530502319336,
+ "step": 6452
+ },
+ {
+ "epoch": 89.62882096069869,
+ "grad_norm": 0.013346714898943901,
+ "learning_rate": 0.0006,
+ "loss": 4.2296929359436035,
+ "step": 6453
+ },
+ {
+ "epoch": 89.64279475982532,
+ "grad_norm": 0.014243829064071178,
+ "learning_rate": 0.0006,
+ "loss": 4.128233909606934,
+ "step": 6454
+ },
+ {
+ "epoch": 89.65676855895197,
+ "grad_norm": 0.01606106385588646,
+ "learning_rate": 0.0006,
+ "loss": 4.202202320098877,
+ "step": 6455
+ },
+ {
+ "epoch": 89.6707423580786,
+ "grad_norm": 0.016074348241090775,
+ "learning_rate": 0.0006,
+ "loss": 4.215323448181152,
+ "step": 6456
+ },
+ {
+ "epoch": 89.68471615720524,
+ "grad_norm": 0.01611054316163063,
+ "learning_rate": 0.0006,
+ "loss": 4.096758842468262,
+ "step": 6457
+ },
+ {
+ "epoch": 89.69868995633188,
+ "grad_norm": 0.013788217678666115,
+ "learning_rate": 0.0006,
+ "loss": 4.098560810089111,
+ "step": 6458
+ },
+ {
+ "epoch": 89.71266375545852,
+ "grad_norm": 0.014408969320356846,
+ "learning_rate": 0.0006,
+ "loss": 4.107398509979248,
+ "step": 6459
+ },
+ {
+ "epoch": 89.72663755458515,
+ "grad_norm": 0.014765267260372639,
+ "learning_rate": 0.0006,
+ "loss": 4.057718276977539,
+ "step": 6460
+ },
+ {
+ "epoch": 89.74061135371178,
+ "grad_norm": 0.017395280301570892,
+ "learning_rate": 0.0006,
+ "loss": 4.100795269012451,
+ "step": 6461
+ },
+ {
+ "epoch": 89.75458515283843,
+ "grad_norm": 0.016821149736642838,
+ "learning_rate": 0.0006,
+ "loss": 4.153731346130371,
+ "step": 6462
+ },
+ {
+ "epoch": 89.76855895196506,
+ "grad_norm": 0.015633124858140945,
+ "learning_rate": 0.0006,
+ "loss": 4.061585903167725,
+ "step": 6463
+ },
+ {
+ "epoch": 89.78253275109171,
+ "grad_norm": 0.01451497245579958,
+ "learning_rate": 0.0006,
+ "loss": 4.1624884605407715,
+ "step": 6464
+ },
+ {
+ "epoch": 89.79650655021834,
+ "grad_norm": 0.013868392445147038,
+ "learning_rate": 0.0006,
+ "loss": 4.178536415100098,
+ "step": 6465
+ },
+ {
+ "epoch": 89.81048034934499,
+ "grad_norm": 0.014259875752031803,
+ "learning_rate": 0.0006,
+ "loss": 4.073114395141602,
+ "step": 6466
+ },
+ {
+ "epoch": 89.82445414847162,
+ "grad_norm": 0.013651594519615173,
+ "learning_rate": 0.0006,
+ "loss": 4.050265312194824,
+ "step": 6467
+ },
+ {
+ "epoch": 89.83842794759825,
+ "grad_norm": 0.014063458889722824,
+ "learning_rate": 0.0006,
+ "loss": 4.2014689445495605,
+ "step": 6468
+ },
+ {
+ "epoch": 89.8524017467249,
+ "grad_norm": 0.014179578050971031,
+ "learning_rate": 0.0006,
+ "loss": 4.110550880432129,
+ "step": 6469
+ },
+ {
+ "epoch": 89.86637554585153,
+ "grad_norm": 0.013616854324936867,
+ "learning_rate": 0.0006,
+ "loss": 4.1776580810546875,
+ "step": 6470
+ },
+ {
+ "epoch": 89.88034934497817,
+ "grad_norm": 0.014351330697536469,
+ "learning_rate": 0.0006,
+ "loss": 4.058526992797852,
+ "step": 6471
+ },
+ {
+ "epoch": 89.8943231441048,
+ "grad_norm": 0.0141244325786829,
+ "learning_rate": 0.0006,
+ "loss": 4.068673610687256,
+ "step": 6472
+ },
+ {
+ "epoch": 89.90829694323143,
+ "grad_norm": 0.01552131213247776,
+ "learning_rate": 0.0006,
+ "loss": 4.086787223815918,
+ "step": 6473
+ },
+ {
+ "epoch": 89.92227074235808,
+ "grad_norm": 0.014614894054830074,
+ "learning_rate": 0.0006,
+ "loss": 4.2190656661987305,
+ "step": 6474
+ },
+ {
+ "epoch": 89.93624454148471,
+ "grad_norm": 0.014369758777320385,
+ "learning_rate": 0.0006,
+ "loss": 4.117021560668945,
+ "step": 6475
+ },
+ {
+ "epoch": 89.95021834061136,
+ "grad_norm": 0.015352046117186546,
+ "learning_rate": 0.0006,
+ "loss": 4.167664527893066,
+ "step": 6476
+ },
+ {
+ "epoch": 89.96419213973799,
+ "grad_norm": 0.014314834959805012,
+ "learning_rate": 0.0006,
+ "loss": 4.160182952880859,
+ "step": 6477
+ },
+ {
+ "epoch": 89.97816593886463,
+ "grad_norm": 0.013511341996490955,
+ "learning_rate": 0.0006,
+ "loss": 4.096034049987793,
+ "step": 6478
+ },
+ {
+ "epoch": 89.99213973799127,
+ "grad_norm": 0.01308482326567173,
+ "learning_rate": 0.0006,
+ "loss": 4.062689781188965,
+ "step": 6479
+ },
+ {
+ "epoch": 90.0,
+ "grad_norm": 0.01442806702107191,
+ "learning_rate": 0.0006,
+ "loss": 4.1955885887146,
+ "step": 6480
+ },
+ {
+ "epoch": 90.0,
+ "eval_loss": 4.644347667694092,
+ "eval_runtime": 56.8677,
+ "eval_samples_per_second": 42.942,
+ "eval_steps_per_second": 1.354,
+ "step": 6480
+ },
+ {
+ "epoch": 90.01397379912663,
+ "grad_norm": 0.014839007519185543,
+ "learning_rate": 0.0006,
+ "loss": 4.171868324279785,
+ "step": 6481
+ },
+ {
+ "epoch": 90.02794759825328,
+ "grad_norm": 0.01476900465786457,
+ "learning_rate": 0.0006,
+ "loss": 4.119754791259766,
+ "step": 6482
+ },
+ {
+ "epoch": 90.04192139737991,
+ "grad_norm": 0.015179301612079144,
+ "learning_rate": 0.0006,
+ "loss": 4.06740140914917,
+ "step": 6483
+ },
+ {
+ "epoch": 90.05589519650655,
+ "grad_norm": 0.015552980825304985,
+ "learning_rate": 0.0006,
+ "loss": 4.112268447875977,
+ "step": 6484
+ },
+ {
+ "epoch": 90.06986899563319,
+ "grad_norm": 0.014930234290659428,
+ "learning_rate": 0.0006,
+ "loss": 4.071866989135742,
+ "step": 6485
+ },
+ {
+ "epoch": 90.08384279475983,
+ "grad_norm": 0.014124148525297642,
+ "learning_rate": 0.0006,
+ "loss": 4.013731956481934,
+ "step": 6486
+ },
+ {
+ "epoch": 90.09781659388646,
+ "grad_norm": 0.014204252511262894,
+ "learning_rate": 0.0006,
+ "loss": 4.133620738983154,
+ "step": 6487
+ },
+ {
+ "epoch": 90.1117903930131,
+ "grad_norm": 0.014845543541014194,
+ "learning_rate": 0.0006,
+ "loss": 4.122956275939941,
+ "step": 6488
+ },
+ {
+ "epoch": 90.12576419213974,
+ "grad_norm": 0.0146744754165411,
+ "learning_rate": 0.0006,
+ "loss": 4.087121963500977,
+ "step": 6489
+ },
+ {
+ "epoch": 90.13973799126637,
+ "grad_norm": 0.016007190570235252,
+ "learning_rate": 0.0006,
+ "loss": 3.98740816116333,
+ "step": 6490
+ },
+ {
+ "epoch": 90.15371179039302,
+ "grad_norm": 0.018072661012411118,
+ "learning_rate": 0.0006,
+ "loss": 4.1155009269714355,
+ "step": 6491
+ },
+ {
+ "epoch": 90.16768558951965,
+ "grad_norm": 0.016868166625499725,
+ "learning_rate": 0.0006,
+ "loss": 4.123727321624756,
+ "step": 6492
+ },
+ {
+ "epoch": 90.18165938864628,
+ "grad_norm": 0.017091726884245872,
+ "learning_rate": 0.0006,
+ "loss": 4.1405229568481445,
+ "step": 6493
+ },
+ {
+ "epoch": 90.19563318777293,
+ "grad_norm": 0.018329855054616928,
+ "learning_rate": 0.0006,
+ "loss": 4.093900680541992,
+ "step": 6494
+ },
+ {
+ "epoch": 90.20960698689956,
+ "grad_norm": 0.018906457349658012,
+ "learning_rate": 0.0006,
+ "loss": 4.0552849769592285,
+ "step": 6495
+ },
+ {
+ "epoch": 90.2235807860262,
+ "grad_norm": 0.017275869846343994,
+ "learning_rate": 0.0006,
+ "loss": 4.1310625076293945,
+ "step": 6496
+ },
+ {
+ "epoch": 90.23755458515284,
+ "grad_norm": 0.015029897913336754,
+ "learning_rate": 0.0006,
+ "loss": 4.100166320800781,
+ "step": 6497
+ },
+ {
+ "epoch": 90.25152838427948,
+ "grad_norm": 0.014891199767589569,
+ "learning_rate": 0.0006,
+ "loss": 4.1314802169799805,
+ "step": 6498
+ },
+ {
+ "epoch": 90.26550218340611,
+ "grad_norm": 0.014877690002322197,
+ "learning_rate": 0.0006,
+ "loss": 4.076098442077637,
+ "step": 6499
+ },
+ {
+ "epoch": 90.27947598253274,
+ "grad_norm": 0.014670864678919315,
+ "learning_rate": 0.0006,
+ "loss": 4.066137313842773,
+ "step": 6500
+ },
+ {
+ "epoch": 90.29344978165939,
+ "grad_norm": 0.01444234699010849,
+ "learning_rate": 0.0006,
+ "loss": 4.013591766357422,
+ "step": 6501
+ },
+ {
+ "epoch": 90.30742358078602,
+ "grad_norm": 0.01339263841509819,
+ "learning_rate": 0.0006,
+ "loss": 4.117959499359131,
+ "step": 6502
+ },
+ {
+ "epoch": 90.32139737991267,
+ "grad_norm": 0.014753810130059719,
+ "learning_rate": 0.0006,
+ "loss": 4.042947769165039,
+ "step": 6503
+ },
+ {
+ "epoch": 90.3353711790393,
+ "grad_norm": 0.014947964809834957,
+ "learning_rate": 0.0006,
+ "loss": 4.112810134887695,
+ "step": 6504
+ },
+ {
+ "epoch": 90.34934497816595,
+ "grad_norm": 0.015155353583395481,
+ "learning_rate": 0.0006,
+ "loss": 4.213214874267578,
+ "step": 6505
+ },
+ {
+ "epoch": 90.36331877729258,
+ "grad_norm": 0.015502216294407845,
+ "learning_rate": 0.0006,
+ "loss": 4.127619743347168,
+ "step": 6506
+ },
+ {
+ "epoch": 90.37729257641921,
+ "grad_norm": 0.015377781353890896,
+ "learning_rate": 0.0006,
+ "loss": 4.135862350463867,
+ "step": 6507
+ },
+ {
+ "epoch": 90.39126637554585,
+ "grad_norm": 0.01662006787955761,
+ "learning_rate": 0.0006,
+ "loss": 4.076877593994141,
+ "step": 6508
+ },
+ {
+ "epoch": 90.40524017467249,
+ "grad_norm": 0.016324395313858986,
+ "learning_rate": 0.0006,
+ "loss": 4.214895725250244,
+ "step": 6509
+ },
+ {
+ "epoch": 90.41921397379913,
+ "grad_norm": 0.016099024564027786,
+ "learning_rate": 0.0006,
+ "loss": 4.139717102050781,
+ "step": 6510
+ },
+ {
+ "epoch": 90.43318777292576,
+ "grad_norm": 0.014403503388166428,
+ "learning_rate": 0.0006,
+ "loss": 4.040343284606934,
+ "step": 6511
+ },
+ {
+ "epoch": 90.44716157205241,
+ "grad_norm": 0.01413416862487793,
+ "learning_rate": 0.0006,
+ "loss": 4.087702751159668,
+ "step": 6512
+ },
+ {
+ "epoch": 90.46113537117904,
+ "grad_norm": 0.01603989489376545,
+ "learning_rate": 0.0006,
+ "loss": 4.173558712005615,
+ "step": 6513
+ },
+ {
+ "epoch": 90.47510917030567,
+ "grad_norm": 0.016846206039190292,
+ "learning_rate": 0.0006,
+ "loss": 4.1955461502075195,
+ "step": 6514
+ },
+ {
+ "epoch": 90.48908296943232,
+ "grad_norm": 0.01652812585234642,
+ "learning_rate": 0.0006,
+ "loss": 4.061838150024414,
+ "step": 6515
+ },
+ {
+ "epoch": 90.50305676855895,
+ "grad_norm": 0.015100528486073017,
+ "learning_rate": 0.0006,
+ "loss": 4.041808605194092,
+ "step": 6516
+ },
+ {
+ "epoch": 90.5170305676856,
+ "grad_norm": 0.013867822475731373,
+ "learning_rate": 0.0006,
+ "loss": 4.12662410736084,
+ "step": 6517
+ },
+ {
+ "epoch": 90.53100436681223,
+ "grad_norm": 0.013684107922017574,
+ "learning_rate": 0.0006,
+ "loss": 4.1254048347473145,
+ "step": 6518
+ },
+ {
+ "epoch": 90.54497816593886,
+ "grad_norm": 0.014200013130903244,
+ "learning_rate": 0.0006,
+ "loss": 4.183216094970703,
+ "step": 6519
+ },
+ {
+ "epoch": 90.5589519650655,
+ "grad_norm": 0.013707022182643414,
+ "learning_rate": 0.0006,
+ "loss": 4.07176399230957,
+ "step": 6520
+ },
+ {
+ "epoch": 90.57292576419214,
+ "grad_norm": 0.013996582478284836,
+ "learning_rate": 0.0006,
+ "loss": 4.006575584411621,
+ "step": 6521
+ },
+ {
+ "epoch": 90.58689956331878,
+ "grad_norm": 0.014886210672557354,
+ "learning_rate": 0.0006,
+ "loss": 4.118956565856934,
+ "step": 6522
+ },
+ {
+ "epoch": 90.60087336244541,
+ "grad_norm": 0.014022685587406158,
+ "learning_rate": 0.0006,
+ "loss": 3.96986722946167,
+ "step": 6523
+ },
+ {
+ "epoch": 90.61484716157206,
+ "grad_norm": 0.014547634869813919,
+ "learning_rate": 0.0006,
+ "loss": 4.122886657714844,
+ "step": 6524
+ },
+ {
+ "epoch": 90.62882096069869,
+ "grad_norm": 0.013682851567864418,
+ "learning_rate": 0.0006,
+ "loss": 3.988292694091797,
+ "step": 6525
+ },
+ {
+ "epoch": 90.64279475982532,
+ "grad_norm": 0.014016421511769295,
+ "learning_rate": 0.0006,
+ "loss": 4.110270977020264,
+ "step": 6526
+ },
+ {
+ "epoch": 90.65676855895197,
+ "grad_norm": 0.013392233289778233,
+ "learning_rate": 0.0006,
+ "loss": 4.065059661865234,
+ "step": 6527
+ },
+ {
+ "epoch": 90.6707423580786,
+ "grad_norm": 0.012107795104384422,
+ "learning_rate": 0.0006,
+ "loss": 4.034260272979736,
+ "step": 6528
+ },
+ {
+ "epoch": 90.68471615720524,
+ "grad_norm": 0.013198604807257652,
+ "learning_rate": 0.0006,
+ "loss": 4.162083625793457,
+ "step": 6529
+ },
+ {
+ "epoch": 90.69868995633188,
+ "grad_norm": 0.012380579486489296,
+ "learning_rate": 0.0006,
+ "loss": 4.040456295013428,
+ "step": 6530
+ },
+ {
+ "epoch": 90.71266375545852,
+ "grad_norm": 0.012784135527908802,
+ "learning_rate": 0.0006,
+ "loss": 4.0823798179626465,
+ "step": 6531
+ },
+ {
+ "epoch": 90.72663755458515,
+ "grad_norm": 0.012920220382511616,
+ "learning_rate": 0.0006,
+ "loss": 4.181064128875732,
+ "step": 6532
+ },
+ {
+ "epoch": 90.74061135371178,
+ "grad_norm": 0.01338814664632082,
+ "learning_rate": 0.0006,
+ "loss": 4.111660003662109,
+ "step": 6533
+ },
+ {
+ "epoch": 90.75458515283843,
+ "grad_norm": 0.013628040440380573,
+ "learning_rate": 0.0006,
+ "loss": 4.237852573394775,
+ "step": 6534
+ },
+ {
+ "epoch": 90.76855895196506,
+ "grad_norm": 0.013384854421019554,
+ "learning_rate": 0.0006,
+ "loss": 4.121862888336182,
+ "step": 6535
+ },
+ {
+ "epoch": 90.78253275109171,
+ "grad_norm": 0.014726929366588593,
+ "learning_rate": 0.0006,
+ "loss": 4.1695990562438965,
+ "step": 6536
+ },
+ {
+ "epoch": 90.79650655021834,
+ "grad_norm": 0.0161006897687912,
+ "learning_rate": 0.0006,
+ "loss": 4.17794132232666,
+ "step": 6537
+ },
+ {
+ "epoch": 90.81048034934499,
+ "grad_norm": 0.016761884093284607,
+ "learning_rate": 0.0006,
+ "loss": 4.093764781951904,
+ "step": 6538
+ },
+ {
+ "epoch": 90.82445414847162,
+ "grad_norm": 0.017246825620532036,
+ "learning_rate": 0.0006,
+ "loss": 3.9382331371307373,
+ "step": 6539
+ },
+ {
+ "epoch": 90.83842794759825,
+ "grad_norm": 0.018141290172934532,
+ "learning_rate": 0.0006,
+ "loss": 3.973609209060669,
+ "step": 6540
+ },
+ {
+ "epoch": 90.8524017467249,
+ "grad_norm": 0.018129250034689903,
+ "learning_rate": 0.0006,
+ "loss": 4.227414131164551,
+ "step": 6541
+ },
+ {
+ "epoch": 90.86637554585153,
+ "grad_norm": 0.018234573304653168,
+ "learning_rate": 0.0006,
+ "loss": 4.151870250701904,
+ "step": 6542
+ },
+ {
+ "epoch": 90.88034934497817,
+ "grad_norm": 0.01560172438621521,
+ "learning_rate": 0.0006,
+ "loss": 4.127397537231445,
+ "step": 6543
+ },
+ {
+ "epoch": 90.8943231441048,
+ "grad_norm": 0.014367681927978992,
+ "learning_rate": 0.0006,
+ "loss": 4.17820930480957,
+ "step": 6544
+ },
+ {
+ "epoch": 90.90829694323143,
+ "grad_norm": 0.016071319580078125,
+ "learning_rate": 0.0006,
+ "loss": 4.059854507446289,
+ "step": 6545
+ },
+ {
+ "epoch": 90.92227074235808,
+ "grad_norm": 0.01589169353246689,
+ "learning_rate": 0.0006,
+ "loss": 4.282748222351074,
+ "step": 6546
+ },
+ {
+ "epoch": 90.93624454148471,
+ "grad_norm": 0.013639772310853004,
+ "learning_rate": 0.0006,
+ "loss": 4.147368907928467,
+ "step": 6547
+ },
+ {
+ "epoch": 90.95021834061136,
+ "grad_norm": 0.014248427003622055,
+ "learning_rate": 0.0006,
+ "loss": 4.240248680114746,
+ "step": 6548
+ },
+ {
+ "epoch": 90.96419213973799,
+ "grad_norm": 0.014600320719182491,
+ "learning_rate": 0.0006,
+ "loss": 4.122123718261719,
+ "step": 6549
+ },
+ {
+ "epoch": 90.97816593886463,
+ "grad_norm": 0.014371675439178944,
+ "learning_rate": 0.0006,
+ "loss": 4.013302803039551,
+ "step": 6550
+ },
+ {
+ "epoch": 90.99213973799127,
+ "grad_norm": 0.013133526779711246,
+ "learning_rate": 0.0006,
+ "loss": 4.097136497497559,
+ "step": 6551
+ },
+ {
+ "epoch": 91.0,
+ "grad_norm": 0.015599353238940239,
+ "learning_rate": 0.0006,
+ "loss": 4.198554039001465,
+ "step": 6552
+ },
+ {
+ "epoch": 91.0,
+ "eval_loss": 4.603907585144043,
+ "eval_runtime": 56.9418,
+ "eval_samples_per_second": 42.886,
+ "eval_steps_per_second": 1.352,
+ "step": 6552
+ },
+ {
+ "epoch": 91.01397379912663,
+ "grad_norm": 0.015917908400297165,
+ "learning_rate": 0.0006,
+ "loss": 4.133877754211426,
+ "step": 6553
+ },
+ {
+ "epoch": 91.02794759825328,
+ "grad_norm": 0.015485800802707672,
+ "learning_rate": 0.0006,
+ "loss": 4.111617088317871,
+ "step": 6554
+ },
+ {
+ "epoch": 91.04192139737991,
+ "grad_norm": 0.012969603762030602,
+ "learning_rate": 0.0006,
+ "loss": 4.064828395843506,
+ "step": 6555
+ },
+ {
+ "epoch": 91.05589519650655,
+ "grad_norm": 0.014482097700238228,
+ "learning_rate": 0.0006,
+ "loss": 4.101710319519043,
+ "step": 6556
+ },
+ {
+ "epoch": 91.06986899563319,
+ "grad_norm": 0.017242740839719772,
+ "learning_rate": 0.0006,
+ "loss": 4.015342712402344,
+ "step": 6557
+ },
+ {
+ "epoch": 91.08384279475983,
+ "grad_norm": 0.017100440338253975,
+ "learning_rate": 0.0006,
+ "loss": 3.9801602363586426,
+ "step": 6558
+ },
+ {
+ "epoch": 91.09781659388646,
+ "grad_norm": 0.015969131141901016,
+ "learning_rate": 0.0006,
+ "loss": 4.049958229064941,
+ "step": 6559
+ },
+ {
+ "epoch": 91.1117903930131,
+ "grad_norm": 0.016018927097320557,
+ "learning_rate": 0.0006,
+ "loss": 4.094139099121094,
+ "step": 6560
+ },
+ {
+ "epoch": 91.12576419213974,
+ "grad_norm": 0.018131926655769348,
+ "learning_rate": 0.0006,
+ "loss": 3.9499974250793457,
+ "step": 6561
+ },
+ {
+ "epoch": 91.13973799126637,
+ "grad_norm": 0.019425220787525177,
+ "learning_rate": 0.0006,
+ "loss": 3.9910552501678467,
+ "step": 6562
+ },
+ {
+ "epoch": 91.15371179039302,
+ "grad_norm": 0.01908663846552372,
+ "learning_rate": 0.0006,
+ "loss": 4.032179832458496,
+ "step": 6563
+ },
+ {
+ "epoch": 91.16768558951965,
+ "grad_norm": 0.01871207356452942,
+ "learning_rate": 0.0006,
+ "loss": 4.112380027770996,
+ "step": 6564
+ },
+ {
+ "epoch": 91.18165938864628,
+ "grad_norm": 0.020867953076958656,
+ "learning_rate": 0.0006,
+ "loss": 4.06105899810791,
+ "step": 6565
+ },
+ {
+ "epoch": 91.19563318777293,
+ "grad_norm": 0.019053412601351738,
+ "learning_rate": 0.0006,
+ "loss": 3.992504596710205,
+ "step": 6566
+ },
+ {
+ "epoch": 91.20960698689956,
+ "grad_norm": 0.016084402799606323,
+ "learning_rate": 0.0006,
+ "loss": 4.100559234619141,
+ "step": 6567
+ },
+ {
+ "epoch": 91.2235807860262,
+ "grad_norm": 0.016857590526342392,
+ "learning_rate": 0.0006,
+ "loss": 4.1567277908325195,
+ "step": 6568
+ },
+ {
+ "epoch": 91.23755458515284,
+ "grad_norm": 0.01702957972884178,
+ "learning_rate": 0.0006,
+ "loss": 4.234771728515625,
+ "step": 6569
+ },
+ {
+ "epoch": 91.25152838427948,
+ "grad_norm": 0.016869494691491127,
+ "learning_rate": 0.0006,
+ "loss": 3.9931387901306152,
+ "step": 6570
+ },
+ {
+ "epoch": 91.26550218340611,
+ "grad_norm": 0.01696581020951271,
+ "learning_rate": 0.0006,
+ "loss": 4.184908866882324,
+ "step": 6571
+ },
+ {
+ "epoch": 91.27947598253274,
+ "grad_norm": 0.014982126653194427,
+ "learning_rate": 0.0006,
+ "loss": 4.036668300628662,
+ "step": 6572
+ },
+ {
+ "epoch": 91.29344978165939,
+ "grad_norm": 0.014363301917910576,
+ "learning_rate": 0.0006,
+ "loss": 4.074010848999023,
+ "step": 6573
+ },
+ {
+ "epoch": 91.30742358078602,
+ "grad_norm": 0.015492337755858898,
+ "learning_rate": 0.0006,
+ "loss": 4.127626419067383,
+ "step": 6574
+ },
+ {
+ "epoch": 91.32139737991267,
+ "grad_norm": 0.015375050716102123,
+ "learning_rate": 0.0006,
+ "loss": 4.159229755401611,
+ "step": 6575
+ },
+ {
+ "epoch": 91.3353711790393,
+ "grad_norm": 0.015068565495312214,
+ "learning_rate": 0.0006,
+ "loss": 4.172623634338379,
+ "step": 6576
+ },
+ {
+ "epoch": 91.34934497816595,
+ "grad_norm": 0.0140256742015481,
+ "learning_rate": 0.0006,
+ "loss": 4.207065105438232,
+ "step": 6577
+ },
+ {
+ "epoch": 91.36331877729258,
+ "grad_norm": 0.014220152981579304,
+ "learning_rate": 0.0006,
+ "loss": 4.131444931030273,
+ "step": 6578
+ },
+ {
+ "epoch": 91.37729257641921,
+ "grad_norm": 0.0143149234354496,
+ "learning_rate": 0.0006,
+ "loss": 4.065218448638916,
+ "step": 6579
+ },
+ {
+ "epoch": 91.39126637554585,
+ "grad_norm": 0.013149659149348736,
+ "learning_rate": 0.0006,
+ "loss": 4.106743335723877,
+ "step": 6580
+ },
+ {
+ "epoch": 91.40524017467249,
+ "grad_norm": 0.014550330117344856,
+ "learning_rate": 0.0006,
+ "loss": 4.123722553253174,
+ "step": 6581
+ },
+ {
+ "epoch": 91.41921397379913,
+ "grad_norm": 0.014611710794270039,
+ "learning_rate": 0.0006,
+ "loss": 4.079610824584961,
+ "step": 6582
+ },
+ {
+ "epoch": 91.43318777292576,
+ "grad_norm": 0.013369769789278507,
+ "learning_rate": 0.0006,
+ "loss": 4.133187770843506,
+ "step": 6583
+ },
+ {
+ "epoch": 91.44716157205241,
+ "grad_norm": 0.013243545778095722,
+ "learning_rate": 0.0006,
+ "loss": 4.042607307434082,
+ "step": 6584
+ },
+ {
+ "epoch": 91.46113537117904,
+ "grad_norm": 0.013388266786932945,
+ "learning_rate": 0.0006,
+ "loss": 4.14863395690918,
+ "step": 6585
+ },
+ {
+ "epoch": 91.47510917030567,
+ "grad_norm": 0.014333273284137249,
+ "learning_rate": 0.0006,
+ "loss": 4.109699726104736,
+ "step": 6586
+ },
+ {
+ "epoch": 91.48908296943232,
+ "grad_norm": 0.012742757797241211,
+ "learning_rate": 0.0006,
+ "loss": 4.126582145690918,
+ "step": 6587
+ },
+ {
+ "epoch": 91.50305676855895,
+ "grad_norm": 0.012880304828286171,
+ "learning_rate": 0.0006,
+ "loss": 4.205300331115723,
+ "step": 6588
+ },
+ {
+ "epoch": 91.5170305676856,
+ "grad_norm": 0.013896043412387371,
+ "learning_rate": 0.0006,
+ "loss": 4.132510185241699,
+ "step": 6589
+ },
+ {
+ "epoch": 91.53100436681223,
+ "grad_norm": 0.014361281879246235,
+ "learning_rate": 0.0006,
+ "loss": 4.1487932205200195,
+ "step": 6590
+ },
+ {
+ "epoch": 91.54497816593886,
+ "grad_norm": 0.015137741342186928,
+ "learning_rate": 0.0006,
+ "loss": 4.1564531326293945,
+ "step": 6591
+ },
+ {
+ "epoch": 91.5589519650655,
+ "grad_norm": 0.014660682529211044,
+ "learning_rate": 0.0006,
+ "loss": 4.056064128875732,
+ "step": 6592
+ },
+ {
+ "epoch": 91.57292576419214,
+ "grad_norm": 0.014888476580381393,
+ "learning_rate": 0.0006,
+ "loss": 4.0633344650268555,
+ "step": 6593
+ },
+ {
+ "epoch": 91.58689956331878,
+ "grad_norm": 0.012162433005869389,
+ "learning_rate": 0.0006,
+ "loss": 4.026293754577637,
+ "step": 6594
+ },
+ {
+ "epoch": 91.60087336244541,
+ "grad_norm": 0.012892271392047405,
+ "learning_rate": 0.0006,
+ "loss": 4.267086029052734,
+ "step": 6595
+ },
+ {
+ "epoch": 91.61484716157206,
+ "grad_norm": 0.01450759544968605,
+ "learning_rate": 0.0006,
+ "loss": 4.149359703063965,
+ "step": 6596
+ },
+ {
+ "epoch": 91.62882096069869,
+ "grad_norm": 0.01419843453913927,
+ "learning_rate": 0.0006,
+ "loss": 4.094183921813965,
+ "step": 6597
+ },
+ {
+ "epoch": 91.64279475982532,
+ "grad_norm": 0.014401616528630257,
+ "learning_rate": 0.0006,
+ "loss": 4.125172138214111,
+ "step": 6598
+ },
+ {
+ "epoch": 91.65676855895197,
+ "grad_norm": 0.013945513404905796,
+ "learning_rate": 0.0006,
+ "loss": 4.146623134613037,
+ "step": 6599
+ },
+ {
+ "epoch": 91.6707423580786,
+ "grad_norm": 0.013516047969460487,
+ "learning_rate": 0.0006,
+ "loss": 4.160526275634766,
+ "step": 6600
+ },
+ {
+ "epoch": 91.68471615720524,
+ "grad_norm": 0.013262378051877022,
+ "learning_rate": 0.0006,
+ "loss": 4.074923038482666,
+ "step": 6601
+ },
+ {
+ "epoch": 91.69868995633188,
+ "grad_norm": 0.014365669339895248,
+ "learning_rate": 0.0006,
+ "loss": 4.118094444274902,
+ "step": 6602
+ },
+ {
+ "epoch": 91.71266375545852,
+ "grad_norm": 0.016713734716176987,
+ "learning_rate": 0.0006,
+ "loss": 4.021799087524414,
+ "step": 6603
+ },
+ {
+ "epoch": 91.72663755458515,
+ "grad_norm": 0.01988680474460125,
+ "learning_rate": 0.0006,
+ "loss": 4.238436222076416,
+ "step": 6604
+ },
+ {
+ "epoch": 91.74061135371178,
+ "grad_norm": 0.022749733179807663,
+ "learning_rate": 0.0006,
+ "loss": 4.010470390319824,
+ "step": 6605
+ },
+ {
+ "epoch": 91.75458515283843,
+ "grad_norm": 0.022533461451530457,
+ "learning_rate": 0.0006,
+ "loss": 4.152201175689697,
+ "step": 6606
+ },
+ {
+ "epoch": 91.76855895196506,
+ "grad_norm": 0.02021082676947117,
+ "learning_rate": 0.0006,
+ "loss": 4.120572090148926,
+ "step": 6607
+ },
+ {
+ "epoch": 91.78253275109171,
+ "grad_norm": 0.016298197209835052,
+ "learning_rate": 0.0006,
+ "loss": 4.115815162658691,
+ "step": 6608
+ },
+ {
+ "epoch": 91.79650655021834,
+ "grad_norm": 0.017562786117196083,
+ "learning_rate": 0.0006,
+ "loss": 4.109724044799805,
+ "step": 6609
+ },
+ {
+ "epoch": 91.81048034934499,
+ "grad_norm": 0.018153153359889984,
+ "learning_rate": 0.0006,
+ "loss": 4.143759727478027,
+ "step": 6610
+ },
+ {
+ "epoch": 91.82445414847162,
+ "grad_norm": 0.014762306585907936,
+ "learning_rate": 0.0006,
+ "loss": 4.25596809387207,
+ "step": 6611
+ },
+ {
+ "epoch": 91.83842794759825,
+ "grad_norm": 0.015074343420565128,
+ "learning_rate": 0.0006,
+ "loss": 4.207723140716553,
+ "step": 6612
+ },
+ {
+ "epoch": 91.8524017467249,
+ "grad_norm": 0.015028579160571098,
+ "learning_rate": 0.0006,
+ "loss": 4.01950216293335,
+ "step": 6613
+ },
+ {
+ "epoch": 91.86637554585153,
+ "grad_norm": 0.013996795751154423,
+ "learning_rate": 0.0006,
+ "loss": 4.132997989654541,
+ "step": 6614
+ },
+ {
+ "epoch": 91.88034934497817,
+ "grad_norm": 0.014978236518800259,
+ "learning_rate": 0.0006,
+ "loss": 4.132402420043945,
+ "step": 6615
+ },
+ {
+ "epoch": 91.8943231441048,
+ "grad_norm": 0.013111517764627934,
+ "learning_rate": 0.0006,
+ "loss": 4.1223649978637695,
+ "step": 6616
+ },
+ {
+ "epoch": 91.90829694323143,
+ "grad_norm": 0.013872234150767326,
+ "learning_rate": 0.0006,
+ "loss": 4.096590042114258,
+ "step": 6617
+ },
+ {
+ "epoch": 91.92227074235808,
+ "grad_norm": 0.013642470352351665,
+ "learning_rate": 0.0006,
+ "loss": 4.236227035522461,
+ "step": 6618
+ },
+ {
+ "epoch": 91.93624454148471,
+ "grad_norm": 0.014684873633086681,
+ "learning_rate": 0.0006,
+ "loss": 4.037961959838867,
+ "step": 6619
+ },
+ {
+ "epoch": 91.95021834061136,
+ "grad_norm": 0.014498206786811352,
+ "learning_rate": 0.0006,
+ "loss": 4.043325424194336,
+ "step": 6620
+ },
+ {
+ "epoch": 91.96419213973799,
+ "grad_norm": 0.012768656946718693,
+ "learning_rate": 0.0006,
+ "loss": 4.138803482055664,
+ "step": 6621
+ },
+ {
+ "epoch": 91.97816593886463,
+ "grad_norm": 0.012155905365943909,
+ "learning_rate": 0.0006,
+ "loss": 4.163424968719482,
+ "step": 6622
+ },
+ {
+ "epoch": 91.99213973799127,
+ "grad_norm": 0.012584930285811424,
+ "learning_rate": 0.0006,
+ "loss": 4.086629867553711,
+ "step": 6623
+ },
+ {
+ "epoch": 92.0,
+ "grad_norm": 0.01578151248395443,
+ "learning_rate": 0.0006,
+ "loss": 4.108386993408203,
+ "step": 6624
+ },
+ {
+ "epoch": 92.0,
+ "eval_loss": 4.541166305541992,
+ "eval_runtime": 56.871,
+ "eval_samples_per_second": 42.939,
+ "eval_steps_per_second": 1.354,
+ "step": 6624
+ },
+ {
+ "epoch": 92.01397379912663,
+ "grad_norm": 0.016901545226573944,
+ "learning_rate": 0.0006,
+ "loss": 4.043694496154785,
+ "step": 6625
+ },
+ {
+ "epoch": 92.02794759825328,
+ "grad_norm": 0.016299039125442505,
+ "learning_rate": 0.0006,
+ "loss": 4.100504398345947,
+ "step": 6626
+ },
+ {
+ "epoch": 92.04192139737991,
+ "grad_norm": 0.013988777063786983,
+ "learning_rate": 0.0006,
+ "loss": 4.063943862915039,
+ "step": 6627
+ },
+ {
+ "epoch": 92.05589519650655,
+ "grad_norm": 0.01343096699565649,
+ "learning_rate": 0.0006,
+ "loss": 4.0321197509765625,
+ "step": 6628
+ },
+ {
+ "epoch": 92.06986899563319,
+ "grad_norm": 0.014343147166073322,
+ "learning_rate": 0.0006,
+ "loss": 4.095499038696289,
+ "step": 6629
+ },
+ {
+ "epoch": 92.08384279475983,
+ "grad_norm": 0.014989756047725677,
+ "learning_rate": 0.0006,
+ "loss": 4.122310638427734,
+ "step": 6630
+ },
+ {
+ "epoch": 92.09781659388646,
+ "grad_norm": 0.013929174281656742,
+ "learning_rate": 0.0006,
+ "loss": 4.02870512008667,
+ "step": 6631
+ },
+ {
+ "epoch": 92.1117903930131,
+ "grad_norm": 0.014139998704195023,
+ "learning_rate": 0.0006,
+ "loss": 4.180675506591797,
+ "step": 6632
+ },
+ {
+ "epoch": 92.12576419213974,
+ "grad_norm": 0.012920474633574486,
+ "learning_rate": 0.0006,
+ "loss": 4.111686706542969,
+ "step": 6633
+ },
+ {
+ "epoch": 92.13973799126637,
+ "grad_norm": 0.01291937381029129,
+ "learning_rate": 0.0006,
+ "loss": 3.958092212677002,
+ "step": 6634
+ },
+ {
+ "epoch": 92.15371179039302,
+ "grad_norm": 0.013250474818050861,
+ "learning_rate": 0.0006,
+ "loss": 4.075105667114258,
+ "step": 6635
+ },
+ {
+ "epoch": 92.16768558951965,
+ "grad_norm": 0.01321406476199627,
+ "learning_rate": 0.0006,
+ "loss": 4.072322368621826,
+ "step": 6636
+ },
+ {
+ "epoch": 92.18165938864628,
+ "grad_norm": 0.014002865180373192,
+ "learning_rate": 0.0006,
+ "loss": 4.061572074890137,
+ "step": 6637
+ },
+ {
+ "epoch": 92.19563318777293,
+ "grad_norm": 0.013876105658710003,
+ "learning_rate": 0.0006,
+ "loss": 4.099213600158691,
+ "step": 6638
+ },
+ {
+ "epoch": 92.20960698689956,
+ "grad_norm": 0.01464459951967001,
+ "learning_rate": 0.0006,
+ "loss": 4.1573638916015625,
+ "step": 6639
+ },
+ {
+ "epoch": 92.2235807860262,
+ "grad_norm": 0.015354706905782223,
+ "learning_rate": 0.0006,
+ "loss": 4.078714370727539,
+ "step": 6640
+ },
+ {
+ "epoch": 92.23755458515284,
+ "grad_norm": 0.015265689231455326,
+ "learning_rate": 0.0006,
+ "loss": 4.036652565002441,
+ "step": 6641
+ },
+ {
+ "epoch": 92.25152838427948,
+ "grad_norm": 0.016140855848789215,
+ "learning_rate": 0.0006,
+ "loss": 4.120541095733643,
+ "step": 6642
+ },
+ {
+ "epoch": 92.26550218340611,
+ "grad_norm": 0.017044052481651306,
+ "learning_rate": 0.0006,
+ "loss": 4.015849590301514,
+ "step": 6643
+ },
+ {
+ "epoch": 92.27947598253274,
+ "grad_norm": 0.01469327136874199,
+ "learning_rate": 0.0006,
+ "loss": 4.137286186218262,
+ "step": 6644
+ },
+ {
+ "epoch": 92.29344978165939,
+ "grad_norm": 0.014596942812204361,
+ "learning_rate": 0.0006,
+ "loss": 4.11309814453125,
+ "step": 6645
+ },
+ {
+ "epoch": 92.30742358078602,
+ "grad_norm": 0.014671559445559978,
+ "learning_rate": 0.0006,
+ "loss": 4.087950229644775,
+ "step": 6646
+ },
+ {
+ "epoch": 92.32139737991267,
+ "grad_norm": 0.015040507540106773,
+ "learning_rate": 0.0006,
+ "loss": 3.967053174972534,
+ "step": 6647
+ },
+ {
+ "epoch": 92.3353711790393,
+ "grad_norm": 0.01417640782892704,
+ "learning_rate": 0.0006,
+ "loss": 3.990346908569336,
+ "step": 6648
+ },
+ {
+ "epoch": 92.34934497816595,
+ "grad_norm": 0.013855187222361565,
+ "learning_rate": 0.0006,
+ "loss": 3.965132474899292,
+ "step": 6649
+ },
+ {
+ "epoch": 92.36331877729258,
+ "grad_norm": 0.014291216619312763,
+ "learning_rate": 0.0006,
+ "loss": 4.161717891693115,
+ "step": 6650
+ },
+ {
+ "epoch": 92.37729257641921,
+ "grad_norm": 0.014148853719234467,
+ "learning_rate": 0.0006,
+ "loss": 4.019951820373535,
+ "step": 6651
+ },
+ {
+ "epoch": 92.39126637554585,
+ "grad_norm": 0.01414303295314312,
+ "learning_rate": 0.0006,
+ "loss": 4.125976085662842,
+ "step": 6652
+ },
+ {
+ "epoch": 92.40524017467249,
+ "grad_norm": 0.01358176488429308,
+ "learning_rate": 0.0006,
+ "loss": 4.088022232055664,
+ "step": 6653
+ },
+ {
+ "epoch": 92.41921397379913,
+ "grad_norm": 0.014174354262650013,
+ "learning_rate": 0.0006,
+ "loss": 4.1137518882751465,
+ "step": 6654
+ },
+ {
+ "epoch": 92.43318777292576,
+ "grad_norm": 0.013756301254034042,
+ "learning_rate": 0.0006,
+ "loss": 4.0088725090026855,
+ "step": 6655
+ },
+ {
+ "epoch": 92.44716157205241,
+ "grad_norm": 0.014257770963013172,
+ "learning_rate": 0.0006,
+ "loss": 4.05248498916626,
+ "step": 6656
+ },
+ {
+ "epoch": 92.46113537117904,
+ "grad_norm": 0.015319934114813805,
+ "learning_rate": 0.0006,
+ "loss": 4.1375732421875,
+ "step": 6657
+ },
+ {
+ "epoch": 92.47510917030567,
+ "grad_norm": 0.014655457809567451,
+ "learning_rate": 0.0006,
+ "loss": 4.172857284545898,
+ "step": 6658
+ },
+ {
+ "epoch": 92.48908296943232,
+ "grad_norm": 0.014981928281486034,
+ "learning_rate": 0.0006,
+ "loss": 4.064846038818359,
+ "step": 6659
+ },
+ {
+ "epoch": 92.50305676855895,
+ "grad_norm": 0.015197387896478176,
+ "learning_rate": 0.0006,
+ "loss": 4.119397163391113,
+ "step": 6660
+ },
+ {
+ "epoch": 92.5170305676856,
+ "grad_norm": 0.016060935333371162,
+ "learning_rate": 0.0006,
+ "loss": 4.09212064743042,
+ "step": 6661
+ },
+ {
+ "epoch": 92.53100436681223,
+ "grad_norm": 0.014884199015796185,
+ "learning_rate": 0.0006,
+ "loss": 4.027483940124512,
+ "step": 6662
+ },
+ {
+ "epoch": 92.54497816593886,
+ "grad_norm": 0.01562194712460041,
+ "learning_rate": 0.0006,
+ "loss": 3.9722909927368164,
+ "step": 6663
+ },
+ {
+ "epoch": 92.5589519650655,
+ "grad_norm": 0.014801043085753918,
+ "learning_rate": 0.0006,
+ "loss": 4.048130989074707,
+ "step": 6664
+ },
+ {
+ "epoch": 92.57292576419214,
+ "grad_norm": 0.013954494148492813,
+ "learning_rate": 0.0006,
+ "loss": 4.050351619720459,
+ "step": 6665
+ },
+ {
+ "epoch": 92.58689956331878,
+ "grad_norm": 0.013292429968714714,
+ "learning_rate": 0.0006,
+ "loss": 4.245860576629639,
+ "step": 6666
+ },
+ {
+ "epoch": 92.60087336244541,
+ "grad_norm": 0.015737095847725868,
+ "learning_rate": 0.0006,
+ "loss": 4.154323101043701,
+ "step": 6667
+ },
+ {
+ "epoch": 92.61484716157206,
+ "grad_norm": 0.018242565914988518,
+ "learning_rate": 0.0006,
+ "loss": 4.118710517883301,
+ "step": 6668
+ },
+ {
+ "epoch": 92.62882096069869,
+ "grad_norm": 0.017666636034846306,
+ "learning_rate": 0.0006,
+ "loss": 4.1013875007629395,
+ "step": 6669
+ },
+ {
+ "epoch": 92.64279475982532,
+ "grad_norm": 0.015017223544418812,
+ "learning_rate": 0.0006,
+ "loss": 4.084001064300537,
+ "step": 6670
+ },
+ {
+ "epoch": 92.65676855895197,
+ "grad_norm": 0.01359238475561142,
+ "learning_rate": 0.0006,
+ "loss": 4.091456413269043,
+ "step": 6671
+ },
+ {
+ "epoch": 92.6707423580786,
+ "grad_norm": 0.014524881727993488,
+ "learning_rate": 0.0006,
+ "loss": 4.121245384216309,
+ "step": 6672
+ },
+ {
+ "epoch": 92.68471615720524,
+ "grad_norm": 0.016214480623602867,
+ "learning_rate": 0.0006,
+ "loss": 4.133184432983398,
+ "step": 6673
+ },
+ {
+ "epoch": 92.69868995633188,
+ "grad_norm": 0.0176643505692482,
+ "learning_rate": 0.0006,
+ "loss": 4.142827987670898,
+ "step": 6674
+ },
+ {
+ "epoch": 92.71266375545852,
+ "grad_norm": 0.017004093155264854,
+ "learning_rate": 0.0006,
+ "loss": 4.067666530609131,
+ "step": 6675
+ },
+ {
+ "epoch": 92.72663755458515,
+ "grad_norm": 0.014077397994697094,
+ "learning_rate": 0.0006,
+ "loss": 4.076465129852295,
+ "step": 6676
+ },
+ {
+ "epoch": 92.74061135371178,
+ "grad_norm": 0.015905627980828285,
+ "learning_rate": 0.0006,
+ "loss": 4.083726406097412,
+ "step": 6677
+ },
+ {
+ "epoch": 92.75458515283843,
+ "grad_norm": 0.016467567533254623,
+ "learning_rate": 0.0006,
+ "loss": 4.241217136383057,
+ "step": 6678
+ },
+ {
+ "epoch": 92.76855895196506,
+ "grad_norm": 0.01541806198656559,
+ "learning_rate": 0.0006,
+ "loss": 4.05365514755249,
+ "step": 6679
+ },
+ {
+ "epoch": 92.78253275109171,
+ "grad_norm": 0.015028242021799088,
+ "learning_rate": 0.0006,
+ "loss": 4.029950141906738,
+ "step": 6680
+ },
+ {
+ "epoch": 92.79650655021834,
+ "grad_norm": 0.014366233721375465,
+ "learning_rate": 0.0006,
+ "loss": 4.181772232055664,
+ "step": 6681
+ },
+ {
+ "epoch": 92.81048034934499,
+ "grad_norm": 0.014313179068267345,
+ "learning_rate": 0.0006,
+ "loss": 4.152658462524414,
+ "step": 6682
+ },
+ {
+ "epoch": 92.82445414847162,
+ "grad_norm": 0.016504302620887756,
+ "learning_rate": 0.0006,
+ "loss": 4.080532073974609,
+ "step": 6683
+ },
+ {
+ "epoch": 92.83842794759825,
+ "grad_norm": 0.01685677096247673,
+ "learning_rate": 0.0006,
+ "loss": 4.194268226623535,
+ "step": 6684
+ },
+ {
+ "epoch": 92.8524017467249,
+ "grad_norm": 0.018684150651097298,
+ "learning_rate": 0.0006,
+ "loss": 4.218497276306152,
+ "step": 6685
+ },
+ {
+ "epoch": 92.86637554585153,
+ "grad_norm": 0.01880732551217079,
+ "learning_rate": 0.0006,
+ "loss": 4.133942604064941,
+ "step": 6686
+ },
+ {
+ "epoch": 92.88034934497817,
+ "grad_norm": 0.016504917293787003,
+ "learning_rate": 0.0006,
+ "loss": 4.139945030212402,
+ "step": 6687
+ },
+ {
+ "epoch": 92.8943231441048,
+ "grad_norm": 0.016495356336236,
+ "learning_rate": 0.0006,
+ "loss": 4.098918914794922,
+ "step": 6688
+ },
+ {
+ "epoch": 92.90829694323143,
+ "grad_norm": 0.0170587245374918,
+ "learning_rate": 0.0006,
+ "loss": 4.025818824768066,
+ "step": 6689
+ },
+ {
+ "epoch": 92.92227074235808,
+ "grad_norm": 0.01718820631504059,
+ "learning_rate": 0.0006,
+ "loss": 4.125217437744141,
+ "step": 6690
+ },
+ {
+ "epoch": 92.93624454148471,
+ "grad_norm": 0.015789354220032692,
+ "learning_rate": 0.0006,
+ "loss": 4.150068283081055,
+ "step": 6691
+ },
+ {
+ "epoch": 92.95021834061136,
+ "grad_norm": 0.01587480492889881,
+ "learning_rate": 0.0006,
+ "loss": 4.194794654846191,
+ "step": 6692
+ },
+ {
+ "epoch": 92.96419213973799,
+ "grad_norm": 0.015669209882616997,
+ "learning_rate": 0.0006,
+ "loss": 4.0550336837768555,
+ "step": 6693
+ },
+ {
+ "epoch": 92.97816593886463,
+ "grad_norm": 0.01519355084747076,
+ "learning_rate": 0.0006,
+ "loss": 4.114232063293457,
+ "step": 6694
+ },
+ {
+ "epoch": 92.99213973799127,
+ "grad_norm": 0.014596261084079742,
+ "learning_rate": 0.0006,
+ "loss": 3.992065906524658,
+ "step": 6695
+ },
+ {
+ "epoch": 93.0,
+ "grad_norm": 0.016378438100218773,
+ "learning_rate": 0.0006,
+ "loss": 4.093693733215332,
+ "step": 6696
+ },
+ {
+ "epoch": 93.0,
+ "eval_loss": 4.541905879974365,
+ "eval_runtime": 57.0177,
+ "eval_samples_per_second": 42.829,
+ "eval_steps_per_second": 1.35,
+ "step": 6696
+ },
+ {
+ "epoch": 93.01397379912663,
+ "grad_norm": 0.01593143492937088,
+ "learning_rate": 0.0006,
+ "loss": 4.052624225616455,
+ "step": 6697
+ },
+ {
+ "epoch": 93.02794759825328,
+ "grad_norm": 0.016086837276816368,
+ "learning_rate": 0.0006,
+ "loss": 4.130306243896484,
+ "step": 6698
+ },
+ {
+ "epoch": 93.04192139737991,
+ "grad_norm": 0.016004323959350586,
+ "learning_rate": 0.0006,
+ "loss": 3.965893268585205,
+ "step": 6699
+ },
+ {
+ "epoch": 93.05589519650655,
+ "grad_norm": 0.017667284235358238,
+ "learning_rate": 0.0006,
+ "loss": 4.095634460449219,
+ "step": 6700
+ },
+ {
+ "epoch": 93.06986899563319,
+ "grad_norm": 0.01949942484498024,
+ "learning_rate": 0.0006,
+ "loss": 3.966041326522827,
+ "step": 6701
+ },
+ {
+ "epoch": 93.08384279475983,
+ "grad_norm": 0.021898971870541573,
+ "learning_rate": 0.0006,
+ "loss": 4.010378837585449,
+ "step": 6702
+ },
+ {
+ "epoch": 93.09781659388646,
+ "grad_norm": 0.019985264167189598,
+ "learning_rate": 0.0006,
+ "loss": 4.061577320098877,
+ "step": 6703
+ },
+ {
+ "epoch": 93.1117903930131,
+ "grad_norm": 0.016470855101943016,
+ "learning_rate": 0.0006,
+ "loss": 4.0601677894592285,
+ "step": 6704
+ },
+ {
+ "epoch": 93.12576419213974,
+ "grad_norm": 0.015600379556417465,
+ "learning_rate": 0.0006,
+ "loss": 4.0860700607299805,
+ "step": 6705
+ },
+ {
+ "epoch": 93.13973799126637,
+ "grad_norm": 0.017514333128929138,
+ "learning_rate": 0.0006,
+ "loss": 4.240778923034668,
+ "step": 6706
+ },
+ {
+ "epoch": 93.15371179039302,
+ "grad_norm": 0.01780967228114605,
+ "learning_rate": 0.0006,
+ "loss": 4.076499938964844,
+ "step": 6707
+ },
+ {
+ "epoch": 93.16768558951965,
+ "grad_norm": 0.016370438039302826,
+ "learning_rate": 0.0006,
+ "loss": 4.0677618980407715,
+ "step": 6708
+ },
+ {
+ "epoch": 93.18165938864628,
+ "grad_norm": 0.015205491334199905,
+ "learning_rate": 0.0006,
+ "loss": 4.004945755004883,
+ "step": 6709
+ },
+ {
+ "epoch": 93.19563318777293,
+ "grad_norm": 0.01537297572940588,
+ "learning_rate": 0.0006,
+ "loss": 4.121147155761719,
+ "step": 6710
+ },
+ {
+ "epoch": 93.20960698689956,
+ "grad_norm": 0.017343711107969284,
+ "learning_rate": 0.0006,
+ "loss": 4.103215217590332,
+ "step": 6711
+ },
+ {
+ "epoch": 93.2235807860262,
+ "grad_norm": 0.01618131436407566,
+ "learning_rate": 0.0006,
+ "loss": 4.083741664886475,
+ "step": 6712
+ },
+ {
+ "epoch": 93.23755458515284,
+ "grad_norm": 0.013340235687792301,
+ "learning_rate": 0.0006,
+ "loss": 4.055032730102539,
+ "step": 6713
+ },
+ {
+ "epoch": 93.25152838427948,
+ "grad_norm": 0.014084120281040668,
+ "learning_rate": 0.0006,
+ "loss": 4.005610942840576,
+ "step": 6714
+ },
+ {
+ "epoch": 93.26550218340611,
+ "grad_norm": 0.016104530543088913,
+ "learning_rate": 0.0006,
+ "loss": 4.091764450073242,
+ "step": 6715
+ },
+ {
+ "epoch": 93.27947598253274,
+ "grad_norm": 0.015232563950121403,
+ "learning_rate": 0.0006,
+ "loss": 4.105735778808594,
+ "step": 6716
+ },
+ {
+ "epoch": 93.29344978165939,
+ "grad_norm": 0.014003508724272251,
+ "learning_rate": 0.0006,
+ "loss": 4.05783224105835,
+ "step": 6717
+ },
+ {
+ "epoch": 93.30742358078602,
+ "grad_norm": 0.01614833064377308,
+ "learning_rate": 0.0006,
+ "loss": 4.124567031860352,
+ "step": 6718
+ },
+ {
+ "epoch": 93.32139737991267,
+ "grad_norm": 0.015010425820946693,
+ "learning_rate": 0.0006,
+ "loss": 4.07656192779541,
+ "step": 6719
+ },
+ {
+ "epoch": 93.3353711790393,
+ "grad_norm": 0.01444012951105833,
+ "learning_rate": 0.0006,
+ "loss": 4.122415542602539,
+ "step": 6720
+ },
+ {
+ "epoch": 93.34934497816595,
+ "grad_norm": 0.015715785324573517,
+ "learning_rate": 0.0006,
+ "loss": 4.070283889770508,
+ "step": 6721
+ },
+ {
+ "epoch": 93.36331877729258,
+ "grad_norm": 0.014976450242102146,
+ "learning_rate": 0.0006,
+ "loss": 4.087043762207031,
+ "step": 6722
+ },
+ {
+ "epoch": 93.37729257641921,
+ "grad_norm": 0.014868560247123241,
+ "learning_rate": 0.0006,
+ "loss": 3.9671802520751953,
+ "step": 6723
+ },
+ {
+ "epoch": 93.39126637554585,
+ "grad_norm": 0.013518013060092926,
+ "learning_rate": 0.0006,
+ "loss": 4.136781692504883,
+ "step": 6724
+ },
+ {
+ "epoch": 93.40524017467249,
+ "grad_norm": 0.013498986139893532,
+ "learning_rate": 0.0006,
+ "loss": 3.982309103012085,
+ "step": 6725
+ },
+ {
+ "epoch": 93.41921397379913,
+ "grad_norm": 0.013707910664379597,
+ "learning_rate": 0.0006,
+ "loss": 4.181885719299316,
+ "step": 6726
+ },
+ {
+ "epoch": 93.43318777292576,
+ "grad_norm": 0.013913207687437534,
+ "learning_rate": 0.0006,
+ "loss": 3.980070114135742,
+ "step": 6727
+ },
+ {
+ "epoch": 93.44716157205241,
+ "grad_norm": 0.013169893994927406,
+ "learning_rate": 0.0006,
+ "loss": 4.023380279541016,
+ "step": 6728
+ },
+ {
+ "epoch": 93.46113537117904,
+ "grad_norm": 0.014451324008405209,
+ "learning_rate": 0.0006,
+ "loss": 4.038372039794922,
+ "step": 6729
+ },
+ {
+ "epoch": 93.47510917030567,
+ "grad_norm": 0.014401541091501713,
+ "learning_rate": 0.0006,
+ "loss": 4.069009304046631,
+ "step": 6730
+ },
+ {
+ "epoch": 93.48908296943232,
+ "grad_norm": 0.013971361331641674,
+ "learning_rate": 0.0006,
+ "loss": 4.105923652648926,
+ "step": 6731
+ },
+ {
+ "epoch": 93.50305676855895,
+ "grad_norm": 0.013192536309361458,
+ "learning_rate": 0.0006,
+ "loss": 4.07828950881958,
+ "step": 6732
+ },
+ {
+ "epoch": 93.5170305676856,
+ "grad_norm": 0.013581741601228714,
+ "learning_rate": 0.0006,
+ "loss": 4.079203128814697,
+ "step": 6733
+ },
+ {
+ "epoch": 93.53100436681223,
+ "grad_norm": 0.01508194301277399,
+ "learning_rate": 0.0006,
+ "loss": 4.084749221801758,
+ "step": 6734
+ },
+ {
+ "epoch": 93.54497816593886,
+ "grad_norm": 0.0156723503023386,
+ "learning_rate": 0.0006,
+ "loss": 4.070216178894043,
+ "step": 6735
+ },
+ {
+ "epoch": 93.5589519650655,
+ "grad_norm": 0.014996538870036602,
+ "learning_rate": 0.0006,
+ "loss": 4.111318588256836,
+ "step": 6736
+ },
+ {
+ "epoch": 93.57292576419214,
+ "grad_norm": 0.012815269641578197,
+ "learning_rate": 0.0006,
+ "loss": 4.130414962768555,
+ "step": 6737
+ },
+ {
+ "epoch": 93.58689956331878,
+ "grad_norm": 0.012887579388916492,
+ "learning_rate": 0.0006,
+ "loss": 4.128446102142334,
+ "step": 6738
+ },
+ {
+ "epoch": 93.60087336244541,
+ "grad_norm": 0.013632671907544136,
+ "learning_rate": 0.0006,
+ "loss": 4.20977258682251,
+ "step": 6739
+ },
+ {
+ "epoch": 93.61484716157206,
+ "grad_norm": 0.014318895526230335,
+ "learning_rate": 0.0006,
+ "loss": 4.005739212036133,
+ "step": 6740
+ },
+ {
+ "epoch": 93.62882096069869,
+ "grad_norm": 0.014160127379000187,
+ "learning_rate": 0.0006,
+ "loss": 4.122544288635254,
+ "step": 6741
+ },
+ {
+ "epoch": 93.64279475982532,
+ "grad_norm": 0.01308423187583685,
+ "learning_rate": 0.0006,
+ "loss": 4.108774662017822,
+ "step": 6742
+ },
+ {
+ "epoch": 93.65676855895197,
+ "grad_norm": 0.014293737709522247,
+ "learning_rate": 0.0006,
+ "loss": 3.983156681060791,
+ "step": 6743
+ },
+ {
+ "epoch": 93.6707423580786,
+ "grad_norm": 0.015155135653913021,
+ "learning_rate": 0.0006,
+ "loss": 4.006117820739746,
+ "step": 6744
+ },
+ {
+ "epoch": 93.68471615720524,
+ "grad_norm": 0.014391839504241943,
+ "learning_rate": 0.0006,
+ "loss": 4.1658935546875,
+ "step": 6745
+ },
+ {
+ "epoch": 93.69868995633188,
+ "grad_norm": 0.014924845658242702,
+ "learning_rate": 0.0006,
+ "loss": 4.090868949890137,
+ "step": 6746
+ },
+ {
+ "epoch": 93.71266375545852,
+ "grad_norm": 0.014786627143621445,
+ "learning_rate": 0.0006,
+ "loss": 4.121716499328613,
+ "step": 6747
+ },
+ {
+ "epoch": 93.72663755458515,
+ "grad_norm": 0.016948824748396873,
+ "learning_rate": 0.0006,
+ "loss": 4.213179111480713,
+ "step": 6748
+ },
+ {
+ "epoch": 93.74061135371178,
+ "grad_norm": 0.01732620596885681,
+ "learning_rate": 0.0006,
+ "loss": 4.079215049743652,
+ "step": 6749
+ },
+ {
+ "epoch": 93.75458515283843,
+ "grad_norm": 0.016298124566674232,
+ "learning_rate": 0.0006,
+ "loss": 4.096126079559326,
+ "step": 6750
+ },
+ {
+ "epoch": 93.76855895196506,
+ "grad_norm": 0.015744732692837715,
+ "learning_rate": 0.0006,
+ "loss": 4.046300411224365,
+ "step": 6751
+ },
+ {
+ "epoch": 93.78253275109171,
+ "grad_norm": 0.016330812126398087,
+ "learning_rate": 0.0006,
+ "loss": 4.098011016845703,
+ "step": 6752
+ },
+ {
+ "epoch": 93.79650655021834,
+ "grad_norm": 0.01710600219666958,
+ "learning_rate": 0.0006,
+ "loss": 4.031402587890625,
+ "step": 6753
+ },
+ {
+ "epoch": 93.81048034934499,
+ "grad_norm": 0.01561805047094822,
+ "learning_rate": 0.0006,
+ "loss": 4.023388862609863,
+ "step": 6754
+ },
+ {
+ "epoch": 93.82445414847162,
+ "grad_norm": 0.013637243770062923,
+ "learning_rate": 0.0006,
+ "loss": 4.143238067626953,
+ "step": 6755
+ },
+ {
+ "epoch": 93.83842794759825,
+ "grad_norm": 0.013088075444102287,
+ "learning_rate": 0.0006,
+ "loss": 3.9083704948425293,
+ "step": 6756
+ },
+ {
+ "epoch": 93.8524017467249,
+ "grad_norm": 0.015893537551164627,
+ "learning_rate": 0.0006,
+ "loss": 4.102263450622559,
+ "step": 6757
+ },
+ {
+ "epoch": 93.86637554585153,
+ "grad_norm": 0.015213928185403347,
+ "learning_rate": 0.0006,
+ "loss": 4.111978530883789,
+ "step": 6758
+ },
+ {
+ "epoch": 93.88034934497817,
+ "grad_norm": 0.014543982222676277,
+ "learning_rate": 0.0006,
+ "loss": 4.05889892578125,
+ "step": 6759
+ },
+ {
+ "epoch": 93.8943231441048,
+ "grad_norm": 0.015104752965271473,
+ "learning_rate": 0.0006,
+ "loss": 4.136166572570801,
+ "step": 6760
+ },
+ {
+ "epoch": 93.90829694323143,
+ "grad_norm": 0.01505881268531084,
+ "learning_rate": 0.0006,
+ "loss": 4.061883449554443,
+ "step": 6761
+ },
+ {
+ "epoch": 93.92227074235808,
+ "grad_norm": 0.015248455107212067,
+ "learning_rate": 0.0006,
+ "loss": 4.109863758087158,
+ "step": 6762
+ },
+ {
+ "epoch": 93.93624454148471,
+ "grad_norm": 0.013651185669004917,
+ "learning_rate": 0.0006,
+ "loss": 4.061549663543701,
+ "step": 6763
+ },
+ {
+ "epoch": 93.95021834061136,
+ "grad_norm": 0.013806282542645931,
+ "learning_rate": 0.0006,
+ "loss": 4.205403804779053,
+ "step": 6764
+ },
+ {
+ "epoch": 93.96419213973799,
+ "grad_norm": 0.01538451761007309,
+ "learning_rate": 0.0006,
+ "loss": 4.14816951751709,
+ "step": 6765
+ },
+ {
+ "epoch": 93.97816593886463,
+ "grad_norm": 0.01602427288889885,
+ "learning_rate": 0.0006,
+ "loss": 4.000303268432617,
+ "step": 6766
+ },
+ {
+ "epoch": 93.99213973799127,
+ "grad_norm": 0.014558793976902962,
+ "learning_rate": 0.0006,
+ "loss": 4.096562385559082,
+ "step": 6767
+ },
+ {
+ "epoch": 94.0,
+ "grad_norm": 0.01613018475472927,
+ "learning_rate": 0.0006,
+ "loss": 4.262835502624512,
+ "step": 6768
+ },
+ {
+ "epoch": 94.0,
+ "eval_loss": 4.532126426696777,
+ "eval_runtime": 56.7592,
+ "eval_samples_per_second": 43.024,
+ "eval_steps_per_second": 1.357,
+ "step": 6768
+ },
+ {
+ "epoch": 94.01397379912663,
+ "grad_norm": 0.01614535041153431,
+ "learning_rate": 0.0006,
+ "loss": 4.073431491851807,
+ "step": 6769
+ },
+ {
+ "epoch": 94.02794759825328,
+ "grad_norm": 0.015986818820238113,
+ "learning_rate": 0.0006,
+ "loss": 4.070001602172852,
+ "step": 6770
+ },
+ {
+ "epoch": 94.04192139737991,
+ "grad_norm": 0.015678714960813522,
+ "learning_rate": 0.0006,
+ "loss": 4.06920051574707,
+ "step": 6771
+ },
+ {
+ "epoch": 94.05589519650655,
+ "grad_norm": 0.016593851149082184,
+ "learning_rate": 0.0006,
+ "loss": 4.095780849456787,
+ "step": 6772
+ },
+ {
+ "epoch": 94.06986899563319,
+ "grad_norm": 0.01780948042869568,
+ "learning_rate": 0.0006,
+ "loss": 4.092916488647461,
+ "step": 6773
+ },
+ {
+ "epoch": 94.08384279475983,
+ "grad_norm": 0.01755528897047043,
+ "learning_rate": 0.0006,
+ "loss": 3.9833455085754395,
+ "step": 6774
+ },
+ {
+ "epoch": 94.09781659388646,
+ "grad_norm": 0.01708311215043068,
+ "learning_rate": 0.0006,
+ "loss": 4.053336143493652,
+ "step": 6775
+ },
+ {
+ "epoch": 94.1117903930131,
+ "grad_norm": 0.017362922430038452,
+ "learning_rate": 0.0006,
+ "loss": 4.014343738555908,
+ "step": 6776
+ },
+ {
+ "epoch": 94.12576419213974,
+ "grad_norm": 0.01824064739048481,
+ "learning_rate": 0.0006,
+ "loss": 4.1215691566467285,
+ "step": 6777
+ },
+ {
+ "epoch": 94.13973799126637,
+ "grad_norm": 0.017743906006217003,
+ "learning_rate": 0.0006,
+ "loss": 4.091215133666992,
+ "step": 6778
+ },
+ {
+ "epoch": 94.15371179039302,
+ "grad_norm": 0.016585713252425194,
+ "learning_rate": 0.0006,
+ "loss": 4.013662338256836,
+ "step": 6779
+ },
+ {
+ "epoch": 94.16768558951965,
+ "grad_norm": 0.017397213727235794,
+ "learning_rate": 0.0006,
+ "loss": 4.070602893829346,
+ "step": 6780
+ },
+ {
+ "epoch": 94.18165938864628,
+ "grad_norm": 0.01733497530221939,
+ "learning_rate": 0.0006,
+ "loss": 4.1304426193237305,
+ "step": 6781
+ },
+ {
+ "epoch": 94.19563318777293,
+ "grad_norm": 0.015215300023555756,
+ "learning_rate": 0.0006,
+ "loss": 3.965014696121216,
+ "step": 6782
+ },
+ {
+ "epoch": 94.20960698689956,
+ "grad_norm": 0.015908699482679367,
+ "learning_rate": 0.0006,
+ "loss": 3.924964427947998,
+ "step": 6783
+ },
+ {
+ "epoch": 94.2235807860262,
+ "grad_norm": 0.014044968411326408,
+ "learning_rate": 0.0006,
+ "loss": 4.08673095703125,
+ "step": 6784
+ },
+ {
+ "epoch": 94.23755458515284,
+ "grad_norm": 0.01375142764300108,
+ "learning_rate": 0.0006,
+ "loss": 4.011466979980469,
+ "step": 6785
+ },
+ {
+ "epoch": 94.25152838427948,
+ "grad_norm": 0.015539257787168026,
+ "learning_rate": 0.0006,
+ "loss": 4.09976863861084,
+ "step": 6786
+ },
+ {
+ "epoch": 94.26550218340611,
+ "grad_norm": 0.014303367584943771,
+ "learning_rate": 0.0006,
+ "loss": 4.108369827270508,
+ "step": 6787
+ },
+ {
+ "epoch": 94.27947598253274,
+ "grad_norm": 0.01281928364187479,
+ "learning_rate": 0.0006,
+ "loss": 4.073737144470215,
+ "step": 6788
+ },
+ {
+ "epoch": 94.29344978165939,
+ "grad_norm": 0.013859959319233894,
+ "learning_rate": 0.0006,
+ "loss": 4.054823875427246,
+ "step": 6789
+ },
+ {
+ "epoch": 94.30742358078602,
+ "grad_norm": 0.013326860964298248,
+ "learning_rate": 0.0006,
+ "loss": 4.133251190185547,
+ "step": 6790
+ },
+ {
+ "epoch": 94.32139737991267,
+ "grad_norm": 0.013923076912760735,
+ "learning_rate": 0.0006,
+ "loss": 4.190706729888916,
+ "step": 6791
+ },
+ {
+ "epoch": 94.3353711790393,
+ "grad_norm": 0.01420847699046135,
+ "learning_rate": 0.0006,
+ "loss": 3.9336605072021484,
+ "step": 6792
+ },
+ {
+ "epoch": 94.34934497816595,
+ "grad_norm": 0.014607555232942104,
+ "learning_rate": 0.0006,
+ "loss": 4.128399848937988,
+ "step": 6793
+ },
+ {
+ "epoch": 94.36331877729258,
+ "grad_norm": 0.012862961739301682,
+ "learning_rate": 0.0006,
+ "loss": 4.0132527351379395,
+ "step": 6794
+ },
+ {
+ "epoch": 94.37729257641921,
+ "grad_norm": 0.012530217878520489,
+ "learning_rate": 0.0006,
+ "loss": 4.0384368896484375,
+ "step": 6795
+ },
+ {
+ "epoch": 94.39126637554585,
+ "grad_norm": 0.01286802627146244,
+ "learning_rate": 0.0006,
+ "loss": 3.95570707321167,
+ "step": 6796
+ },
+ {
+ "epoch": 94.40524017467249,
+ "grad_norm": 0.013450917787849903,
+ "learning_rate": 0.0006,
+ "loss": 4.016484260559082,
+ "step": 6797
+ },
+ {
+ "epoch": 94.41921397379913,
+ "grad_norm": 0.01393423043191433,
+ "learning_rate": 0.0006,
+ "loss": 4.124959468841553,
+ "step": 6798
+ },
+ {
+ "epoch": 94.43318777292576,
+ "grad_norm": 0.013077184557914734,
+ "learning_rate": 0.0006,
+ "loss": 4.029163837432861,
+ "step": 6799
+ },
+ {
+ "epoch": 94.44716157205241,
+ "grad_norm": 0.013577334582805634,
+ "learning_rate": 0.0006,
+ "loss": 4.22929573059082,
+ "step": 6800
+ },
+ {
+ "epoch": 94.46113537117904,
+ "grad_norm": 0.013930312357842922,
+ "learning_rate": 0.0006,
+ "loss": 4.024711608886719,
+ "step": 6801
+ },
+ {
+ "epoch": 94.47510917030567,
+ "grad_norm": 0.013746250420808792,
+ "learning_rate": 0.0006,
+ "loss": 3.9766428470611572,
+ "step": 6802
+ },
+ {
+ "epoch": 94.48908296943232,
+ "grad_norm": 0.013682983815670013,
+ "learning_rate": 0.0006,
+ "loss": 4.072981834411621,
+ "step": 6803
+ },
+ {
+ "epoch": 94.50305676855895,
+ "grad_norm": 0.01473842654377222,
+ "learning_rate": 0.0006,
+ "loss": 4.197623252868652,
+ "step": 6804
+ },
+ {
+ "epoch": 94.5170305676856,
+ "grad_norm": 0.014133758842945099,
+ "learning_rate": 0.0006,
+ "loss": 4.055549621582031,
+ "step": 6805
+ },
+ {
+ "epoch": 94.53100436681223,
+ "grad_norm": 0.0139645179733634,
+ "learning_rate": 0.0006,
+ "loss": 4.130135536193848,
+ "step": 6806
+ },
+ {
+ "epoch": 94.54497816593886,
+ "grad_norm": 0.013973365537822247,
+ "learning_rate": 0.0006,
+ "loss": 4.006230354309082,
+ "step": 6807
+ },
+ {
+ "epoch": 94.5589519650655,
+ "grad_norm": 0.014227126725018024,
+ "learning_rate": 0.0006,
+ "loss": 4.130459785461426,
+ "step": 6808
+ },
+ {
+ "epoch": 94.57292576419214,
+ "grad_norm": 0.013129506260156631,
+ "learning_rate": 0.0006,
+ "loss": 4.014974594116211,
+ "step": 6809
+ },
+ {
+ "epoch": 94.58689956331878,
+ "grad_norm": 0.013015891425311565,
+ "learning_rate": 0.0006,
+ "loss": 4.161370277404785,
+ "step": 6810
+ },
+ {
+ "epoch": 94.60087336244541,
+ "grad_norm": 0.013957147486507893,
+ "learning_rate": 0.0006,
+ "loss": 4.09519100189209,
+ "step": 6811
+ },
+ {
+ "epoch": 94.61484716157206,
+ "grad_norm": 0.015494572930037975,
+ "learning_rate": 0.0006,
+ "loss": 4.029949188232422,
+ "step": 6812
+ },
+ {
+ "epoch": 94.62882096069869,
+ "grad_norm": 0.016732869669795036,
+ "learning_rate": 0.0006,
+ "loss": 4.135669708251953,
+ "step": 6813
+ },
+ {
+ "epoch": 94.64279475982532,
+ "grad_norm": 0.01769322343170643,
+ "learning_rate": 0.0006,
+ "loss": 4.109872817993164,
+ "step": 6814
+ },
+ {
+ "epoch": 94.65676855895197,
+ "grad_norm": 0.015932537615299225,
+ "learning_rate": 0.0006,
+ "loss": 4.014595031738281,
+ "step": 6815
+ },
+ {
+ "epoch": 94.6707423580786,
+ "grad_norm": 0.015152990818023682,
+ "learning_rate": 0.0006,
+ "loss": 4.1983442306518555,
+ "step": 6816
+ },
+ {
+ "epoch": 94.68471615720524,
+ "grad_norm": 0.01590685173869133,
+ "learning_rate": 0.0006,
+ "loss": 4.119250774383545,
+ "step": 6817
+ },
+ {
+ "epoch": 94.69868995633188,
+ "grad_norm": 0.016018465161323547,
+ "learning_rate": 0.0006,
+ "loss": 4.048069953918457,
+ "step": 6818
+ },
+ {
+ "epoch": 94.71266375545852,
+ "grad_norm": 0.016621652990579605,
+ "learning_rate": 0.0006,
+ "loss": 4.043731689453125,
+ "step": 6819
+ },
+ {
+ "epoch": 94.72663755458515,
+ "grad_norm": 0.014535024762153625,
+ "learning_rate": 0.0006,
+ "loss": 4.1886820793151855,
+ "step": 6820
+ },
+ {
+ "epoch": 94.74061135371178,
+ "grad_norm": 0.014192107133567333,
+ "learning_rate": 0.0006,
+ "loss": 4.103938102722168,
+ "step": 6821
+ },
+ {
+ "epoch": 94.75458515283843,
+ "grad_norm": 0.0166022852063179,
+ "learning_rate": 0.0006,
+ "loss": 4.017818450927734,
+ "step": 6822
+ },
+ {
+ "epoch": 94.76855895196506,
+ "grad_norm": 0.016247155144810677,
+ "learning_rate": 0.0006,
+ "loss": 4.0038981437683105,
+ "step": 6823
+ },
+ {
+ "epoch": 94.78253275109171,
+ "grad_norm": 0.01593470573425293,
+ "learning_rate": 0.0006,
+ "loss": 4.043969631195068,
+ "step": 6824
+ },
+ {
+ "epoch": 94.79650655021834,
+ "grad_norm": 0.014334345236420631,
+ "learning_rate": 0.0006,
+ "loss": 4.097829341888428,
+ "step": 6825
+ },
+ {
+ "epoch": 94.81048034934499,
+ "grad_norm": 0.015581986866891384,
+ "learning_rate": 0.0006,
+ "loss": 4.148984432220459,
+ "step": 6826
+ },
+ {
+ "epoch": 94.82445414847162,
+ "grad_norm": 0.017519738525152206,
+ "learning_rate": 0.0006,
+ "loss": 4.245623588562012,
+ "step": 6827
+ },
+ {
+ "epoch": 94.83842794759825,
+ "grad_norm": 0.015155277214944363,
+ "learning_rate": 0.0006,
+ "loss": 4.031019687652588,
+ "step": 6828
+ },
+ {
+ "epoch": 94.8524017467249,
+ "grad_norm": 0.014969222247600555,
+ "learning_rate": 0.0006,
+ "loss": 4.036084175109863,
+ "step": 6829
+ },
+ {
+ "epoch": 94.86637554585153,
+ "grad_norm": 0.015987908467650414,
+ "learning_rate": 0.0006,
+ "loss": 4.08396053314209,
+ "step": 6830
+ },
+ {
+ "epoch": 94.88034934497817,
+ "grad_norm": 0.016324151307344437,
+ "learning_rate": 0.0006,
+ "loss": 4.036787033081055,
+ "step": 6831
+ },
+ {
+ "epoch": 94.8943231441048,
+ "grad_norm": 0.015957487747073174,
+ "learning_rate": 0.0006,
+ "loss": 4.147607803344727,
+ "step": 6832
+ },
+ {
+ "epoch": 94.90829694323143,
+ "grad_norm": 0.015286967158317566,
+ "learning_rate": 0.0006,
+ "loss": 4.184615135192871,
+ "step": 6833
+ },
+ {
+ "epoch": 94.92227074235808,
+ "grad_norm": 0.01338377594947815,
+ "learning_rate": 0.0006,
+ "loss": 4.014981269836426,
+ "step": 6834
+ },
+ {
+ "epoch": 94.93624454148471,
+ "grad_norm": 0.013997633941471577,
+ "learning_rate": 0.0006,
+ "loss": 4.021933078765869,
+ "step": 6835
+ },
+ {
+ "epoch": 94.95021834061136,
+ "grad_norm": 0.014572402462363243,
+ "learning_rate": 0.0006,
+ "loss": 4.0326948165893555,
+ "step": 6836
+ },
+ {
+ "epoch": 94.96419213973799,
+ "grad_norm": 0.014379586093127728,
+ "learning_rate": 0.0006,
+ "loss": 4.060697555541992,
+ "step": 6837
+ },
+ {
+ "epoch": 94.97816593886463,
+ "grad_norm": 0.013513042591512203,
+ "learning_rate": 0.0006,
+ "loss": 4.05123233795166,
+ "step": 6838
+ },
+ {
+ "epoch": 94.99213973799127,
+ "grad_norm": 0.012904108501970768,
+ "learning_rate": 0.0006,
+ "loss": 4.020852565765381,
+ "step": 6839
+ },
+ {
+ "epoch": 95.0,
+ "grad_norm": 0.014364209026098251,
+ "learning_rate": 0.0006,
+ "loss": 4.080837249755859,
+ "step": 6840
+ },
+ {
+ "epoch": 95.0,
+ "eval_loss": 4.608776569366455,
+ "eval_runtime": 56.6516,
+ "eval_samples_per_second": 43.106,
+ "eval_steps_per_second": 1.359,
+ "step": 6840
+ },
+ {
+ "epoch": 95.01397379912663,
+ "grad_norm": 0.01357510220259428,
+ "learning_rate": 0.0006,
+ "loss": 4.105312347412109,
+ "step": 6841
+ },
+ {
+ "epoch": 95.02794759825328,
+ "grad_norm": 0.014248257502913475,
+ "learning_rate": 0.0006,
+ "loss": 3.971667766571045,
+ "step": 6842
+ },
+ {
+ "epoch": 95.04192139737991,
+ "grad_norm": 0.014773483388125896,
+ "learning_rate": 0.0006,
+ "loss": 4.03553581237793,
+ "step": 6843
+ },
+ {
+ "epoch": 95.05589519650655,
+ "grad_norm": 0.014766111969947815,
+ "learning_rate": 0.0006,
+ "loss": 4.074707508087158,
+ "step": 6844
+ },
+ {
+ "epoch": 95.06986899563319,
+ "grad_norm": 0.014778018929064274,
+ "learning_rate": 0.0006,
+ "loss": 4.1083855628967285,
+ "step": 6845
+ },
+ {
+ "epoch": 95.08384279475983,
+ "grad_norm": 0.014466686174273491,
+ "learning_rate": 0.0006,
+ "loss": 4.054587364196777,
+ "step": 6846
+ },
+ {
+ "epoch": 95.09781659388646,
+ "grad_norm": 0.015195904299616814,
+ "learning_rate": 0.0006,
+ "loss": 4.142580032348633,
+ "step": 6847
+ },
+ {
+ "epoch": 95.1117903930131,
+ "grad_norm": 0.01530557032674551,
+ "learning_rate": 0.0006,
+ "loss": 4.062134265899658,
+ "step": 6848
+ },
+ {
+ "epoch": 95.12576419213974,
+ "grad_norm": 0.015203768387436867,
+ "learning_rate": 0.0006,
+ "loss": 4.082962989807129,
+ "step": 6849
+ },
+ {
+ "epoch": 95.13973799126637,
+ "grad_norm": 0.014906628057360649,
+ "learning_rate": 0.0006,
+ "loss": 4.033127784729004,
+ "step": 6850
+ },
+ {
+ "epoch": 95.15371179039302,
+ "grad_norm": 0.01553534809499979,
+ "learning_rate": 0.0006,
+ "loss": 4.051743030548096,
+ "step": 6851
+ },
+ {
+ "epoch": 95.16768558951965,
+ "grad_norm": 0.01607976108789444,
+ "learning_rate": 0.0006,
+ "loss": 4.038073539733887,
+ "step": 6852
+ },
+ {
+ "epoch": 95.18165938864628,
+ "grad_norm": 0.015776360407471657,
+ "learning_rate": 0.0006,
+ "loss": 4.013091087341309,
+ "step": 6853
+ },
+ {
+ "epoch": 95.19563318777293,
+ "grad_norm": 0.01623626798391342,
+ "learning_rate": 0.0006,
+ "loss": 4.12293815612793,
+ "step": 6854
+ },
+ {
+ "epoch": 95.20960698689956,
+ "grad_norm": 0.016887828707695007,
+ "learning_rate": 0.0006,
+ "loss": 4.120452404022217,
+ "step": 6855
+ },
+ {
+ "epoch": 95.2235807860262,
+ "grad_norm": 0.01561128068715334,
+ "learning_rate": 0.0006,
+ "loss": 4.071322441101074,
+ "step": 6856
+ },
+ {
+ "epoch": 95.23755458515284,
+ "grad_norm": 0.015679724514484406,
+ "learning_rate": 0.0006,
+ "loss": 4.065072059631348,
+ "step": 6857
+ },
+ {
+ "epoch": 95.25152838427948,
+ "grad_norm": 0.016506962478160858,
+ "learning_rate": 0.0006,
+ "loss": 4.0667924880981445,
+ "step": 6858
+ },
+ {
+ "epoch": 95.26550218340611,
+ "grad_norm": 0.015340839512646198,
+ "learning_rate": 0.0006,
+ "loss": 4.14047908782959,
+ "step": 6859
+ },
+ {
+ "epoch": 95.27947598253274,
+ "grad_norm": 0.013935413211584091,
+ "learning_rate": 0.0006,
+ "loss": 3.9688191413879395,
+ "step": 6860
+ },
+ {
+ "epoch": 95.29344978165939,
+ "grad_norm": 0.015223405323922634,
+ "learning_rate": 0.0006,
+ "loss": 4.101613521575928,
+ "step": 6861
+ },
+ {
+ "epoch": 95.30742358078602,
+ "grad_norm": 0.01562546379864216,
+ "learning_rate": 0.0006,
+ "loss": 4.128626823425293,
+ "step": 6862
+ },
+ {
+ "epoch": 95.32139737991267,
+ "grad_norm": 0.016159312799572945,
+ "learning_rate": 0.0006,
+ "loss": 4.080667018890381,
+ "step": 6863
+ },
+ {
+ "epoch": 95.3353711790393,
+ "grad_norm": 0.01724511943757534,
+ "learning_rate": 0.0006,
+ "loss": 4.08463191986084,
+ "step": 6864
+ },
+ {
+ "epoch": 95.34934497816595,
+ "grad_norm": 0.01592276245355606,
+ "learning_rate": 0.0006,
+ "loss": 4.127612113952637,
+ "step": 6865
+ },
+ {
+ "epoch": 95.36331877729258,
+ "grad_norm": 0.01744082197546959,
+ "learning_rate": 0.0006,
+ "loss": 4.034480094909668,
+ "step": 6866
+ },
+ {
+ "epoch": 95.37729257641921,
+ "grad_norm": 0.01774069108068943,
+ "learning_rate": 0.0006,
+ "loss": 4.091196537017822,
+ "step": 6867
+ },
+ {
+ "epoch": 95.39126637554585,
+ "grad_norm": 0.01653675176203251,
+ "learning_rate": 0.0006,
+ "loss": 4.032937526702881,
+ "step": 6868
+ },
+ {
+ "epoch": 95.40524017467249,
+ "grad_norm": 0.014933439902961254,
+ "learning_rate": 0.0006,
+ "loss": 3.9386508464813232,
+ "step": 6869
+ },
+ {
+ "epoch": 95.41921397379913,
+ "grad_norm": 0.014743257313966751,
+ "learning_rate": 0.0006,
+ "loss": 4.126462459564209,
+ "step": 6870
+ },
+ {
+ "epoch": 95.43318777292576,
+ "grad_norm": 0.014614643529057503,
+ "learning_rate": 0.0006,
+ "loss": 4.093838214874268,
+ "step": 6871
+ },
+ {
+ "epoch": 95.44716157205241,
+ "grad_norm": 0.01665084809064865,
+ "learning_rate": 0.0006,
+ "loss": 4.1471967697143555,
+ "step": 6872
+ },
+ {
+ "epoch": 95.46113537117904,
+ "grad_norm": 0.018056001514196396,
+ "learning_rate": 0.0006,
+ "loss": 4.1388349533081055,
+ "step": 6873
+ },
+ {
+ "epoch": 95.47510917030567,
+ "grad_norm": 0.02129282057285309,
+ "learning_rate": 0.0006,
+ "loss": 4.130726337432861,
+ "step": 6874
+ },
+ {
+ "epoch": 95.48908296943232,
+ "grad_norm": 0.022237276658415794,
+ "learning_rate": 0.0006,
+ "loss": 4.10750675201416,
+ "step": 6875
+ },
+ {
+ "epoch": 95.50305676855895,
+ "grad_norm": 0.021106960251927376,
+ "learning_rate": 0.0006,
+ "loss": 4.136451721191406,
+ "step": 6876
+ },
+ {
+ "epoch": 95.5170305676856,
+ "grad_norm": 0.019107872620224953,
+ "learning_rate": 0.0006,
+ "loss": 4.139644145965576,
+ "step": 6877
+ },
+ {
+ "epoch": 95.53100436681223,
+ "grad_norm": 0.018290145322680473,
+ "learning_rate": 0.0006,
+ "loss": 4.161884784698486,
+ "step": 6878
+ },
+ {
+ "epoch": 95.54497816593886,
+ "grad_norm": 0.018161306157708168,
+ "learning_rate": 0.0006,
+ "loss": 4.1252336502075195,
+ "step": 6879
+ },
+ {
+ "epoch": 95.5589519650655,
+ "grad_norm": 0.018432756885886192,
+ "learning_rate": 0.0006,
+ "loss": 4.0635986328125,
+ "step": 6880
+ },
+ {
+ "epoch": 95.57292576419214,
+ "grad_norm": 0.016592558473348618,
+ "learning_rate": 0.0006,
+ "loss": 4.037112236022949,
+ "step": 6881
+ },
+ {
+ "epoch": 95.58689956331878,
+ "grad_norm": 0.015283452346920967,
+ "learning_rate": 0.0006,
+ "loss": 4.1964826583862305,
+ "step": 6882
+ },
+ {
+ "epoch": 95.60087336244541,
+ "grad_norm": 0.015539180487394333,
+ "learning_rate": 0.0006,
+ "loss": 4.110917568206787,
+ "step": 6883
+ },
+ {
+ "epoch": 95.61484716157206,
+ "grad_norm": 0.016508113592863083,
+ "learning_rate": 0.0006,
+ "loss": 3.9705517292022705,
+ "step": 6884
+ },
+ {
+ "epoch": 95.62882096069869,
+ "grad_norm": 0.016635406762361526,
+ "learning_rate": 0.0006,
+ "loss": 4.025636196136475,
+ "step": 6885
+ },
+ {
+ "epoch": 95.64279475982532,
+ "grad_norm": 0.01688900962471962,
+ "learning_rate": 0.0006,
+ "loss": 4.01814079284668,
+ "step": 6886
+ },
+ {
+ "epoch": 95.65676855895197,
+ "grad_norm": 0.01614062674343586,
+ "learning_rate": 0.0006,
+ "loss": 4.098041534423828,
+ "step": 6887
+ },
+ {
+ "epoch": 95.6707423580786,
+ "grad_norm": 0.015679053962230682,
+ "learning_rate": 0.0006,
+ "loss": 4.077049255371094,
+ "step": 6888
+ },
+ {
+ "epoch": 95.68471615720524,
+ "grad_norm": 0.016617605462670326,
+ "learning_rate": 0.0006,
+ "loss": 4.10363245010376,
+ "step": 6889
+ },
+ {
+ "epoch": 95.69868995633188,
+ "grad_norm": 0.015966709703207016,
+ "learning_rate": 0.0006,
+ "loss": 4.100924491882324,
+ "step": 6890
+ },
+ {
+ "epoch": 95.71266375545852,
+ "grad_norm": 0.015768567100167274,
+ "learning_rate": 0.0006,
+ "loss": 4.063976287841797,
+ "step": 6891
+ },
+ {
+ "epoch": 95.72663755458515,
+ "grad_norm": 0.015800341963768005,
+ "learning_rate": 0.0006,
+ "loss": 4.008409023284912,
+ "step": 6892
+ },
+ {
+ "epoch": 95.74061135371178,
+ "grad_norm": 0.014288571663200855,
+ "learning_rate": 0.0006,
+ "loss": 4.139591217041016,
+ "step": 6893
+ },
+ {
+ "epoch": 95.75458515283843,
+ "grad_norm": 0.013743448071181774,
+ "learning_rate": 0.0006,
+ "loss": 3.967352867126465,
+ "step": 6894
+ },
+ {
+ "epoch": 95.76855895196506,
+ "grad_norm": 0.01387989055365324,
+ "learning_rate": 0.0006,
+ "loss": 4.127554893493652,
+ "step": 6895
+ },
+ {
+ "epoch": 95.78253275109171,
+ "grad_norm": 0.013910802081227303,
+ "learning_rate": 0.0006,
+ "loss": 4.130941390991211,
+ "step": 6896
+ },
+ {
+ "epoch": 95.79650655021834,
+ "grad_norm": 0.013936805538833141,
+ "learning_rate": 0.0006,
+ "loss": 4.138744831085205,
+ "step": 6897
+ },
+ {
+ "epoch": 95.81048034934499,
+ "grad_norm": 0.015135723166167736,
+ "learning_rate": 0.0006,
+ "loss": 4.131860256195068,
+ "step": 6898
+ },
+ {
+ "epoch": 95.82445414847162,
+ "grad_norm": 0.014921358786523342,
+ "learning_rate": 0.0006,
+ "loss": 4.1949462890625,
+ "step": 6899
+ },
+ {
+ "epoch": 95.83842794759825,
+ "grad_norm": 0.01401488296687603,
+ "learning_rate": 0.0006,
+ "loss": 4.181267738342285,
+ "step": 6900
+ },
+ {
+ "epoch": 95.8524017467249,
+ "grad_norm": 0.013691626489162445,
+ "learning_rate": 0.0006,
+ "loss": 4.113667964935303,
+ "step": 6901
+ },
+ {
+ "epoch": 95.86637554585153,
+ "grad_norm": 0.013890047557651997,
+ "learning_rate": 0.0006,
+ "loss": 4.163067817687988,
+ "step": 6902
+ },
+ {
+ "epoch": 95.88034934497817,
+ "grad_norm": 0.01385503076016903,
+ "learning_rate": 0.0006,
+ "loss": 4.076153755187988,
+ "step": 6903
+ },
+ {
+ "epoch": 95.8943231441048,
+ "grad_norm": 0.01383926346898079,
+ "learning_rate": 0.0006,
+ "loss": 4.102423667907715,
+ "step": 6904
+ },
+ {
+ "epoch": 95.90829694323143,
+ "grad_norm": 0.015961717814207077,
+ "learning_rate": 0.0006,
+ "loss": 4.1543779373168945,
+ "step": 6905
+ },
+ {
+ "epoch": 95.92227074235808,
+ "grad_norm": 0.016718363389372826,
+ "learning_rate": 0.0006,
+ "loss": 3.991607189178467,
+ "step": 6906
+ },
+ {
+ "epoch": 95.93624454148471,
+ "grad_norm": 0.01585792563855648,
+ "learning_rate": 0.0006,
+ "loss": 4.048587799072266,
+ "step": 6907
+ },
+ {
+ "epoch": 95.95021834061136,
+ "grad_norm": 0.015311713330447674,
+ "learning_rate": 0.0006,
+ "loss": 4.1070451736450195,
+ "step": 6908
+ },
+ {
+ "epoch": 95.96419213973799,
+ "grad_norm": 0.014521263539791107,
+ "learning_rate": 0.0006,
+ "loss": 4.087587356567383,
+ "step": 6909
+ },
+ {
+ "epoch": 95.97816593886463,
+ "grad_norm": 0.01433157455176115,
+ "learning_rate": 0.0006,
+ "loss": 4.139904499053955,
+ "step": 6910
+ },
+ {
+ "epoch": 95.99213973799127,
+ "grad_norm": 0.014475514180958271,
+ "learning_rate": 0.0006,
+ "loss": 4.182525634765625,
+ "step": 6911
+ },
+ {
+ "epoch": 96.0,
+ "grad_norm": 0.015245620161294937,
+ "learning_rate": 0.0006,
+ "loss": 4.1199870109558105,
+ "step": 6912
+ },
+ {
+ "epoch": 96.0,
+ "eval_loss": 4.58721399307251,
+ "eval_runtime": 56.395,
+ "eval_samples_per_second": 43.302,
+ "eval_steps_per_second": 1.365,
+ "step": 6912
+ },
+ {
+ "epoch": 96.01397379912663,
+ "grad_norm": 0.014325885102152824,
+ "learning_rate": 0.0006,
+ "loss": 4.025214195251465,
+ "step": 6913
+ },
+ {
+ "epoch": 96.02794759825328,
+ "grad_norm": 0.01678297482430935,
+ "learning_rate": 0.0006,
+ "loss": 3.998842716217041,
+ "step": 6914
+ },
+ {
+ "epoch": 96.04192139737991,
+ "grad_norm": 0.017044762149453163,
+ "learning_rate": 0.0006,
+ "loss": 4.122270584106445,
+ "step": 6915
+ },
+ {
+ "epoch": 96.05589519650655,
+ "grad_norm": 0.017583075910806656,
+ "learning_rate": 0.0006,
+ "loss": 3.9782891273498535,
+ "step": 6916
+ },
+ {
+ "epoch": 96.06986899563319,
+ "grad_norm": 0.0179787315428257,
+ "learning_rate": 0.0006,
+ "loss": 4.030402660369873,
+ "step": 6917
+ },
+ {
+ "epoch": 96.08384279475983,
+ "grad_norm": 0.0159553661942482,
+ "learning_rate": 0.0006,
+ "loss": 3.949040651321411,
+ "step": 6918
+ },
+ {
+ "epoch": 96.09781659388646,
+ "grad_norm": 0.01554026547819376,
+ "learning_rate": 0.0006,
+ "loss": 4.0429534912109375,
+ "step": 6919
+ },
+ {
+ "epoch": 96.1117903930131,
+ "grad_norm": 0.016099359840154648,
+ "learning_rate": 0.0006,
+ "loss": 4.037537097930908,
+ "step": 6920
+ },
+ {
+ "epoch": 96.12576419213974,
+ "grad_norm": 0.01676846854388714,
+ "learning_rate": 0.0006,
+ "loss": 4.022023677825928,
+ "step": 6921
+ },
+ {
+ "epoch": 96.13973799126637,
+ "grad_norm": 0.019358212128281593,
+ "learning_rate": 0.0006,
+ "loss": 4.048679351806641,
+ "step": 6922
+ },
+ {
+ "epoch": 96.15371179039302,
+ "grad_norm": 0.020390303805470467,
+ "learning_rate": 0.0006,
+ "loss": 4.144132614135742,
+ "step": 6923
+ },
+ {
+ "epoch": 96.16768558951965,
+ "grad_norm": 0.01884407177567482,
+ "learning_rate": 0.0006,
+ "loss": 4.128715515136719,
+ "step": 6924
+ },
+ {
+ "epoch": 96.18165938864628,
+ "grad_norm": 0.015098451636731625,
+ "learning_rate": 0.0006,
+ "loss": 4.128741264343262,
+ "step": 6925
+ },
+ {
+ "epoch": 96.19563318777293,
+ "grad_norm": 0.0148141635581851,
+ "learning_rate": 0.0006,
+ "loss": 4.054587364196777,
+ "step": 6926
+ },
+ {
+ "epoch": 96.20960698689956,
+ "grad_norm": 0.015594304539263248,
+ "learning_rate": 0.0006,
+ "loss": 4.073890209197998,
+ "step": 6927
+ },
+ {
+ "epoch": 96.2235807860262,
+ "grad_norm": 0.0139383003115654,
+ "learning_rate": 0.0006,
+ "loss": 4.115614891052246,
+ "step": 6928
+ },
+ {
+ "epoch": 96.23755458515284,
+ "grad_norm": 0.0150460135191679,
+ "learning_rate": 0.0006,
+ "loss": 4.040738105773926,
+ "step": 6929
+ },
+ {
+ "epoch": 96.25152838427948,
+ "grad_norm": 0.013879122212529182,
+ "learning_rate": 0.0006,
+ "loss": 3.9292469024658203,
+ "step": 6930
+ },
+ {
+ "epoch": 96.26550218340611,
+ "grad_norm": 0.015546537935733795,
+ "learning_rate": 0.0006,
+ "loss": 4.111361503601074,
+ "step": 6931
+ },
+ {
+ "epoch": 96.27947598253274,
+ "grad_norm": 0.016172925010323524,
+ "learning_rate": 0.0006,
+ "loss": 4.041029930114746,
+ "step": 6932
+ },
+ {
+ "epoch": 96.29344978165939,
+ "grad_norm": 0.015113193541765213,
+ "learning_rate": 0.0006,
+ "loss": 4.157950401306152,
+ "step": 6933
+ },
+ {
+ "epoch": 96.30742358078602,
+ "grad_norm": 0.014663247391581535,
+ "learning_rate": 0.0006,
+ "loss": 4.020952224731445,
+ "step": 6934
+ },
+ {
+ "epoch": 96.32139737991267,
+ "grad_norm": 0.015092858113348484,
+ "learning_rate": 0.0006,
+ "loss": 4.065241813659668,
+ "step": 6935
+ },
+ {
+ "epoch": 96.3353711790393,
+ "grad_norm": 0.013199460692703724,
+ "learning_rate": 0.0006,
+ "loss": 4.062674522399902,
+ "step": 6936
+ },
+ {
+ "epoch": 96.34934497816595,
+ "grad_norm": 0.015065675601363182,
+ "learning_rate": 0.0006,
+ "loss": 4.038748741149902,
+ "step": 6937
+ },
+ {
+ "epoch": 96.36331877729258,
+ "grad_norm": 0.014270043931901455,
+ "learning_rate": 0.0006,
+ "loss": 4.073092460632324,
+ "step": 6938
+ },
+ {
+ "epoch": 96.37729257641921,
+ "grad_norm": 0.01423644833266735,
+ "learning_rate": 0.0006,
+ "loss": 4.238361835479736,
+ "step": 6939
+ },
+ {
+ "epoch": 96.39126637554585,
+ "grad_norm": 0.01380224246531725,
+ "learning_rate": 0.0006,
+ "loss": 4.137866973876953,
+ "step": 6940
+ },
+ {
+ "epoch": 96.40524017467249,
+ "grad_norm": 0.012464135885238647,
+ "learning_rate": 0.0006,
+ "loss": 3.9527339935302734,
+ "step": 6941
+ },
+ {
+ "epoch": 96.41921397379913,
+ "grad_norm": 0.013094968162477016,
+ "learning_rate": 0.0006,
+ "loss": 4.106327056884766,
+ "step": 6942
+ },
+ {
+ "epoch": 96.43318777292576,
+ "grad_norm": 0.013918030075728893,
+ "learning_rate": 0.0006,
+ "loss": 4.072292327880859,
+ "step": 6943
+ },
+ {
+ "epoch": 96.44716157205241,
+ "grad_norm": 0.013835827820003033,
+ "learning_rate": 0.0006,
+ "loss": 4.094313621520996,
+ "step": 6944
+ },
+ {
+ "epoch": 96.46113537117904,
+ "grad_norm": 0.013890027068555355,
+ "learning_rate": 0.0006,
+ "loss": 3.947615623474121,
+ "step": 6945
+ },
+ {
+ "epoch": 96.47510917030567,
+ "grad_norm": 0.012371144257485867,
+ "learning_rate": 0.0006,
+ "loss": 4.038928985595703,
+ "step": 6946
+ },
+ {
+ "epoch": 96.48908296943232,
+ "grad_norm": 0.012236416339874268,
+ "learning_rate": 0.0006,
+ "loss": 3.9835925102233887,
+ "step": 6947
+ },
+ {
+ "epoch": 96.50305676855895,
+ "grad_norm": 0.012866619043052197,
+ "learning_rate": 0.0006,
+ "loss": 4.076041221618652,
+ "step": 6948
+ },
+ {
+ "epoch": 96.5170305676856,
+ "grad_norm": 0.013425248675048351,
+ "learning_rate": 0.0006,
+ "loss": 4.086307525634766,
+ "step": 6949
+ },
+ {
+ "epoch": 96.53100436681223,
+ "grad_norm": 0.014234314672648907,
+ "learning_rate": 0.0006,
+ "loss": 4.087897300720215,
+ "step": 6950
+ },
+ {
+ "epoch": 96.54497816593886,
+ "grad_norm": 0.013670518063008785,
+ "learning_rate": 0.0006,
+ "loss": 4.088400840759277,
+ "step": 6951
+ },
+ {
+ "epoch": 96.5589519650655,
+ "grad_norm": 0.013920796103775501,
+ "learning_rate": 0.0006,
+ "loss": 4.026619911193848,
+ "step": 6952
+ },
+ {
+ "epoch": 96.57292576419214,
+ "grad_norm": 0.013115637935698032,
+ "learning_rate": 0.0006,
+ "loss": 4.141090393066406,
+ "step": 6953
+ },
+ {
+ "epoch": 96.58689956331878,
+ "grad_norm": 0.013466455042362213,
+ "learning_rate": 0.0006,
+ "loss": 4.048084259033203,
+ "step": 6954
+ },
+ {
+ "epoch": 96.60087336244541,
+ "grad_norm": 0.013038012199103832,
+ "learning_rate": 0.0006,
+ "loss": 4.034649848937988,
+ "step": 6955
+ },
+ {
+ "epoch": 96.61484716157206,
+ "grad_norm": 0.012926512397825718,
+ "learning_rate": 0.0006,
+ "loss": 4.138445854187012,
+ "step": 6956
+ },
+ {
+ "epoch": 96.62882096069869,
+ "grad_norm": 0.013271111994981766,
+ "learning_rate": 0.0006,
+ "loss": 4.240598201751709,
+ "step": 6957
+ },
+ {
+ "epoch": 96.64279475982532,
+ "grad_norm": 0.013092855922877789,
+ "learning_rate": 0.0006,
+ "loss": 4.091870307922363,
+ "step": 6958
+ },
+ {
+ "epoch": 96.65676855895197,
+ "grad_norm": 0.013304130174219608,
+ "learning_rate": 0.0006,
+ "loss": 4.048708915710449,
+ "step": 6959
+ },
+ {
+ "epoch": 96.6707423580786,
+ "grad_norm": 0.01289574895054102,
+ "learning_rate": 0.0006,
+ "loss": 4.083846092224121,
+ "step": 6960
+ },
+ {
+ "epoch": 96.68471615720524,
+ "grad_norm": 0.012587146833539009,
+ "learning_rate": 0.0006,
+ "loss": 4.1233015060424805,
+ "step": 6961
+ },
+ {
+ "epoch": 96.69868995633188,
+ "grad_norm": 0.012312786653637886,
+ "learning_rate": 0.0006,
+ "loss": 4.0938262939453125,
+ "step": 6962
+ },
+ {
+ "epoch": 96.71266375545852,
+ "grad_norm": 0.014632516540586948,
+ "learning_rate": 0.0006,
+ "loss": 4.0226545333862305,
+ "step": 6963
+ },
+ {
+ "epoch": 96.72663755458515,
+ "grad_norm": 0.015110369771718979,
+ "learning_rate": 0.0006,
+ "loss": 4.052829742431641,
+ "step": 6964
+ },
+ {
+ "epoch": 96.74061135371178,
+ "grad_norm": 0.015997016802430153,
+ "learning_rate": 0.0006,
+ "loss": 4.126044750213623,
+ "step": 6965
+ },
+ {
+ "epoch": 96.75458515283843,
+ "grad_norm": 0.016428956761956215,
+ "learning_rate": 0.0006,
+ "loss": 4.126011848449707,
+ "step": 6966
+ },
+ {
+ "epoch": 96.76855895196506,
+ "grad_norm": 0.01638922467827797,
+ "learning_rate": 0.0006,
+ "loss": 4.1207685470581055,
+ "step": 6967
+ },
+ {
+ "epoch": 96.78253275109171,
+ "grad_norm": 0.01387663371860981,
+ "learning_rate": 0.0006,
+ "loss": 3.9403412342071533,
+ "step": 6968
+ },
+ {
+ "epoch": 96.79650655021834,
+ "grad_norm": 0.014120660722255707,
+ "learning_rate": 0.0006,
+ "loss": 4.202868461608887,
+ "step": 6969
+ },
+ {
+ "epoch": 96.81048034934499,
+ "grad_norm": 0.014854361303150654,
+ "learning_rate": 0.0006,
+ "loss": 4.015185832977295,
+ "step": 6970
+ },
+ {
+ "epoch": 96.82445414847162,
+ "grad_norm": 0.014359727501869202,
+ "learning_rate": 0.0006,
+ "loss": 4.119946479797363,
+ "step": 6971
+ },
+ {
+ "epoch": 96.83842794759825,
+ "grad_norm": 0.01564539223909378,
+ "learning_rate": 0.0006,
+ "loss": 3.9208037853240967,
+ "step": 6972
+ },
+ {
+ "epoch": 96.8524017467249,
+ "grad_norm": 0.015814676880836487,
+ "learning_rate": 0.0006,
+ "loss": 4.056029319763184,
+ "step": 6973
+ },
+ {
+ "epoch": 96.86637554585153,
+ "grad_norm": 0.016342822462320328,
+ "learning_rate": 0.0006,
+ "loss": 4.160576343536377,
+ "step": 6974
+ },
+ {
+ "epoch": 96.88034934497817,
+ "grad_norm": 0.016163306310772896,
+ "learning_rate": 0.0006,
+ "loss": 3.94950008392334,
+ "step": 6975
+ },
+ {
+ "epoch": 96.8943231441048,
+ "grad_norm": 0.018308300524950027,
+ "learning_rate": 0.0006,
+ "loss": 4.096719741821289,
+ "step": 6976
+ },
+ {
+ "epoch": 96.90829694323143,
+ "grad_norm": 0.01985356956720352,
+ "learning_rate": 0.0006,
+ "loss": 4.052276611328125,
+ "step": 6977
+ },
+ {
+ "epoch": 96.92227074235808,
+ "grad_norm": 0.01949184387922287,
+ "learning_rate": 0.0006,
+ "loss": 4.077932357788086,
+ "step": 6978
+ },
+ {
+ "epoch": 96.93624454148471,
+ "grad_norm": 0.018964678049087524,
+ "learning_rate": 0.0006,
+ "loss": 4.019469738006592,
+ "step": 6979
+ },
+ {
+ "epoch": 96.95021834061136,
+ "grad_norm": 0.019365238025784492,
+ "learning_rate": 0.0006,
+ "loss": 4.032814979553223,
+ "step": 6980
+ },
+ {
+ "epoch": 96.96419213973799,
+ "grad_norm": 0.01766037568449974,
+ "learning_rate": 0.0006,
+ "loss": 4.061336517333984,
+ "step": 6981
+ },
+ {
+ "epoch": 96.97816593886463,
+ "grad_norm": 0.018002403900027275,
+ "learning_rate": 0.0006,
+ "loss": 4.054131507873535,
+ "step": 6982
+ },
+ {
+ "epoch": 96.99213973799127,
+ "grad_norm": 0.01888604648411274,
+ "learning_rate": 0.0006,
+ "loss": 3.9295921325683594,
+ "step": 6983
+ },
+ {
+ "epoch": 97.0,
+ "grad_norm": 0.01609404757618904,
+ "learning_rate": 0.0006,
+ "loss": 4.037106513977051,
+ "step": 6984
+ },
+ {
+ "epoch": 97.0,
+ "eval_loss": 4.5986433029174805,
+ "eval_runtime": 56.5822,
+ "eval_samples_per_second": 43.158,
+ "eval_steps_per_second": 1.361,
+ "step": 6984
+ },
+ {
+ "epoch": 97.01397379912663,
+ "grad_norm": 0.015459981746971607,
+ "learning_rate": 0.0006,
+ "loss": 4.06088924407959,
+ "step": 6985
+ },
+ {
+ "epoch": 97.02794759825328,
+ "grad_norm": 0.016333693638443947,
+ "learning_rate": 0.0006,
+ "loss": 3.9525675773620605,
+ "step": 6986
+ },
+ {
+ "epoch": 97.04192139737991,
+ "grad_norm": 0.0157504603266716,
+ "learning_rate": 0.0006,
+ "loss": 4.143257141113281,
+ "step": 6987
+ },
+ {
+ "epoch": 97.05589519650655,
+ "grad_norm": 0.015252048149704933,
+ "learning_rate": 0.0006,
+ "loss": 4.158165454864502,
+ "step": 6988
+ },
+ {
+ "epoch": 97.06986899563319,
+ "grad_norm": 0.015041683800518513,
+ "learning_rate": 0.0006,
+ "loss": 4.030884742736816,
+ "step": 6989
+ },
+ {
+ "epoch": 97.08384279475983,
+ "grad_norm": 0.014693699777126312,
+ "learning_rate": 0.0006,
+ "loss": 3.8566339015960693,
+ "step": 6990
+ },
+ {
+ "epoch": 97.09781659388646,
+ "grad_norm": 0.0141681469976902,
+ "learning_rate": 0.0006,
+ "loss": 4.001909255981445,
+ "step": 6991
+ },
+ {
+ "epoch": 97.1117903930131,
+ "grad_norm": 0.013231917284429073,
+ "learning_rate": 0.0006,
+ "loss": 3.995041608810425,
+ "step": 6992
+ },
+ {
+ "epoch": 97.12576419213974,
+ "grad_norm": 0.014617119915783405,
+ "learning_rate": 0.0006,
+ "loss": 4.012053489685059,
+ "step": 6993
+ },
+ {
+ "epoch": 97.13973799126637,
+ "grad_norm": 0.014983531087636948,
+ "learning_rate": 0.0006,
+ "loss": 4.084125518798828,
+ "step": 6994
+ },
+ {
+ "epoch": 97.15371179039302,
+ "grad_norm": 0.013970102183520794,
+ "learning_rate": 0.0006,
+ "loss": 4.001402378082275,
+ "step": 6995
+ },
+ {
+ "epoch": 97.16768558951965,
+ "grad_norm": 0.0134900426492095,
+ "learning_rate": 0.0006,
+ "loss": 4.007713794708252,
+ "step": 6996
+ },
+ {
+ "epoch": 97.18165938864628,
+ "grad_norm": 0.013806418515741825,
+ "learning_rate": 0.0006,
+ "loss": 3.9524781703948975,
+ "step": 6997
+ },
+ {
+ "epoch": 97.19563318777293,
+ "grad_norm": 0.014304264448583126,
+ "learning_rate": 0.0006,
+ "loss": 3.9666526317596436,
+ "step": 6998
+ },
+ {
+ "epoch": 97.20960698689956,
+ "grad_norm": 0.016392510384321213,
+ "learning_rate": 0.0006,
+ "loss": 3.9915683269500732,
+ "step": 6999
+ },
+ {
+ "epoch": 97.2235807860262,
+ "grad_norm": 0.017044387757778168,
+ "learning_rate": 0.0006,
+ "loss": 4.11865234375,
+ "step": 7000
+ },
+ {
+ "epoch": 97.23755458515284,
+ "grad_norm": 0.016325082629919052,
+ "learning_rate": 0.0006,
+ "loss": 4.0636820793151855,
+ "step": 7001
+ },
+ {
+ "epoch": 97.25152838427948,
+ "grad_norm": 0.014634544961154461,
+ "learning_rate": 0.0006,
+ "loss": 3.967127799987793,
+ "step": 7002
+ },
+ {
+ "epoch": 97.26550218340611,
+ "grad_norm": 0.014043385162949562,
+ "learning_rate": 0.0006,
+ "loss": 4.072741985321045,
+ "step": 7003
+ },
+ {
+ "epoch": 97.27947598253274,
+ "grad_norm": 0.01376431342214346,
+ "learning_rate": 0.0006,
+ "loss": 4.07547664642334,
+ "step": 7004
+ },
+ {
+ "epoch": 97.29344978165939,
+ "grad_norm": 0.013959003612399101,
+ "learning_rate": 0.0006,
+ "loss": 3.9862523078918457,
+ "step": 7005
+ },
+ {
+ "epoch": 97.30742358078602,
+ "grad_norm": 0.01420508325099945,
+ "learning_rate": 0.0006,
+ "loss": 4.085540294647217,
+ "step": 7006
+ },
+ {
+ "epoch": 97.32139737991267,
+ "grad_norm": 0.01613876223564148,
+ "learning_rate": 0.0006,
+ "loss": 4.025545120239258,
+ "step": 7007
+ },
+ {
+ "epoch": 97.3353711790393,
+ "grad_norm": 0.017000367864966393,
+ "learning_rate": 0.0006,
+ "loss": 4.047660827636719,
+ "step": 7008
+ },
+ {
+ "epoch": 97.34934497816595,
+ "grad_norm": 0.01646680012345314,
+ "learning_rate": 0.0006,
+ "loss": 4.087928771972656,
+ "step": 7009
+ },
+ {
+ "epoch": 97.36331877729258,
+ "grad_norm": 0.015728263184428215,
+ "learning_rate": 0.0006,
+ "loss": 4.022225856781006,
+ "step": 7010
+ },
+ {
+ "epoch": 97.37729257641921,
+ "grad_norm": 0.015262200497090816,
+ "learning_rate": 0.0006,
+ "loss": 3.944697618484497,
+ "step": 7011
+ },
+ {
+ "epoch": 97.39126637554585,
+ "grad_norm": 0.016452079638838768,
+ "learning_rate": 0.0006,
+ "loss": 3.957644462585449,
+ "step": 7012
+ },
+ {
+ "epoch": 97.40524017467249,
+ "grad_norm": 0.018871447071433067,
+ "learning_rate": 0.0006,
+ "loss": 4.017784118652344,
+ "step": 7013
+ },
+ {
+ "epoch": 97.41921397379913,
+ "grad_norm": 0.01902632787823677,
+ "learning_rate": 0.0006,
+ "loss": 3.973320484161377,
+ "step": 7014
+ },
+ {
+ "epoch": 97.43318777292576,
+ "grad_norm": 0.016632359474897385,
+ "learning_rate": 0.0006,
+ "loss": 4.047779083251953,
+ "step": 7015
+ },
+ {
+ "epoch": 97.44716157205241,
+ "grad_norm": 0.01676122285425663,
+ "learning_rate": 0.0006,
+ "loss": 4.149074554443359,
+ "step": 7016
+ },
+ {
+ "epoch": 97.46113537117904,
+ "grad_norm": 0.01764233037829399,
+ "learning_rate": 0.0006,
+ "loss": 3.92502760887146,
+ "step": 7017
+ },
+ {
+ "epoch": 97.47510917030567,
+ "grad_norm": 0.01836075820028782,
+ "learning_rate": 0.0006,
+ "loss": 4.173796653747559,
+ "step": 7018
+ },
+ {
+ "epoch": 97.48908296943232,
+ "grad_norm": 0.017320366576313972,
+ "learning_rate": 0.0006,
+ "loss": 4.007115364074707,
+ "step": 7019
+ },
+ {
+ "epoch": 97.50305676855895,
+ "grad_norm": 0.015976276248693466,
+ "learning_rate": 0.0006,
+ "loss": 4.165112495422363,
+ "step": 7020
+ },
+ {
+ "epoch": 97.5170305676856,
+ "grad_norm": 0.0173528753221035,
+ "learning_rate": 0.0006,
+ "loss": 4.158927917480469,
+ "step": 7021
+ },
+ {
+ "epoch": 97.53100436681223,
+ "grad_norm": 0.01881110668182373,
+ "learning_rate": 0.0006,
+ "loss": 4.069684982299805,
+ "step": 7022
+ },
+ {
+ "epoch": 97.54497816593886,
+ "grad_norm": 0.01849045790731907,
+ "learning_rate": 0.0006,
+ "loss": 4.085758209228516,
+ "step": 7023
+ },
+ {
+ "epoch": 97.5589519650655,
+ "grad_norm": 0.016648925840854645,
+ "learning_rate": 0.0006,
+ "loss": 4.047055721282959,
+ "step": 7024
+ },
+ {
+ "epoch": 97.57292576419214,
+ "grad_norm": 0.014700878411531448,
+ "learning_rate": 0.0006,
+ "loss": 4.0177507400512695,
+ "step": 7025
+ },
+ {
+ "epoch": 97.58689956331878,
+ "grad_norm": 0.014601817354559898,
+ "learning_rate": 0.0006,
+ "loss": 4.070394039154053,
+ "step": 7026
+ },
+ {
+ "epoch": 97.60087336244541,
+ "grad_norm": 0.016391467303037643,
+ "learning_rate": 0.0006,
+ "loss": 3.957401990890503,
+ "step": 7027
+ },
+ {
+ "epoch": 97.61484716157206,
+ "grad_norm": 0.015401630662381649,
+ "learning_rate": 0.0006,
+ "loss": 4.040555000305176,
+ "step": 7028
+ },
+ {
+ "epoch": 97.62882096069869,
+ "grad_norm": 0.014783227816224098,
+ "learning_rate": 0.0006,
+ "loss": 4.039798259735107,
+ "step": 7029
+ },
+ {
+ "epoch": 97.64279475982532,
+ "grad_norm": 0.014704087749123573,
+ "learning_rate": 0.0006,
+ "loss": 4.059834003448486,
+ "step": 7030
+ },
+ {
+ "epoch": 97.65676855895197,
+ "grad_norm": 0.01375004556030035,
+ "learning_rate": 0.0006,
+ "loss": 4.084839344024658,
+ "step": 7031
+ },
+ {
+ "epoch": 97.6707423580786,
+ "grad_norm": 0.01403744611889124,
+ "learning_rate": 0.0006,
+ "loss": 4.060206890106201,
+ "step": 7032
+ },
+ {
+ "epoch": 97.68471615720524,
+ "grad_norm": 0.012827585451304913,
+ "learning_rate": 0.0006,
+ "loss": 4.187256813049316,
+ "step": 7033
+ },
+ {
+ "epoch": 97.69868995633188,
+ "grad_norm": 0.013072148896753788,
+ "learning_rate": 0.0006,
+ "loss": 4.136262893676758,
+ "step": 7034
+ },
+ {
+ "epoch": 97.71266375545852,
+ "grad_norm": 0.013496107421815395,
+ "learning_rate": 0.0006,
+ "loss": 3.9419479370117188,
+ "step": 7035
+ },
+ {
+ "epoch": 97.72663755458515,
+ "grad_norm": 0.014445106498897076,
+ "learning_rate": 0.0006,
+ "loss": 3.9494845867156982,
+ "step": 7036
+ },
+ {
+ "epoch": 97.74061135371178,
+ "grad_norm": 0.014507179148495197,
+ "learning_rate": 0.0006,
+ "loss": 4.05567741394043,
+ "step": 7037
+ },
+ {
+ "epoch": 97.75458515283843,
+ "grad_norm": 0.015422643162310123,
+ "learning_rate": 0.0006,
+ "loss": 4.103999137878418,
+ "step": 7038
+ },
+ {
+ "epoch": 97.76855895196506,
+ "grad_norm": 0.01515049859881401,
+ "learning_rate": 0.0006,
+ "loss": 4.094231605529785,
+ "step": 7039
+ },
+ {
+ "epoch": 97.78253275109171,
+ "grad_norm": 0.013850145973265171,
+ "learning_rate": 0.0006,
+ "loss": 4.084870338439941,
+ "step": 7040
+ },
+ {
+ "epoch": 97.79650655021834,
+ "grad_norm": 0.013332060538232327,
+ "learning_rate": 0.0006,
+ "loss": 4.193096160888672,
+ "step": 7041
+ },
+ {
+ "epoch": 97.81048034934499,
+ "grad_norm": 0.014778176322579384,
+ "learning_rate": 0.0006,
+ "loss": 4.0440216064453125,
+ "step": 7042
+ },
+ {
+ "epoch": 97.82445414847162,
+ "grad_norm": 0.013498589396476746,
+ "learning_rate": 0.0006,
+ "loss": 4.088800430297852,
+ "step": 7043
+ },
+ {
+ "epoch": 97.83842794759825,
+ "grad_norm": 0.014520619064569473,
+ "learning_rate": 0.0006,
+ "loss": 4.156583786010742,
+ "step": 7044
+ },
+ {
+ "epoch": 97.8524017467249,
+ "grad_norm": 0.014858689159154892,
+ "learning_rate": 0.0006,
+ "loss": 4.113174915313721,
+ "step": 7045
+ },
+ {
+ "epoch": 97.86637554585153,
+ "grad_norm": 0.015516536310315132,
+ "learning_rate": 0.0006,
+ "loss": 4.07958984375,
+ "step": 7046
+ },
+ {
+ "epoch": 97.88034934497817,
+ "grad_norm": 0.016619030386209488,
+ "learning_rate": 0.0006,
+ "loss": 4.03786563873291,
+ "step": 7047
+ },
+ {
+ "epoch": 97.8943231441048,
+ "grad_norm": 0.015789631754159927,
+ "learning_rate": 0.0006,
+ "loss": 4.065053462982178,
+ "step": 7048
+ },
+ {
+ "epoch": 97.90829694323143,
+ "grad_norm": 0.015848442912101746,
+ "learning_rate": 0.0006,
+ "loss": 4.183546543121338,
+ "step": 7049
+ },
+ {
+ "epoch": 97.92227074235808,
+ "grad_norm": 0.014621895737946033,
+ "learning_rate": 0.0006,
+ "loss": 4.024744987487793,
+ "step": 7050
+ },
+ {
+ "epoch": 97.93624454148471,
+ "grad_norm": 0.015793966129422188,
+ "learning_rate": 0.0006,
+ "loss": 3.9715819358825684,
+ "step": 7051
+ },
+ {
+ "epoch": 97.95021834061136,
+ "grad_norm": 0.014637742191553116,
+ "learning_rate": 0.0006,
+ "loss": 4.061223983764648,
+ "step": 7052
+ },
+ {
+ "epoch": 97.96419213973799,
+ "grad_norm": 0.0143470773473382,
+ "learning_rate": 0.0006,
+ "loss": 4.00929069519043,
+ "step": 7053
+ },
+ {
+ "epoch": 97.97816593886463,
+ "grad_norm": 0.014167622663080692,
+ "learning_rate": 0.0006,
+ "loss": 3.982048749923706,
+ "step": 7054
+ },
+ {
+ "epoch": 97.99213973799127,
+ "grad_norm": 0.01527823694050312,
+ "learning_rate": 0.0006,
+ "loss": 4.063419818878174,
+ "step": 7055
+ },
+ {
+ "epoch": 98.0,
+ "grad_norm": 0.015410988591611385,
+ "learning_rate": 0.0006,
+ "loss": 4.058195114135742,
+ "step": 7056
+ },
+ {
+ "epoch": 98.0,
+ "eval_loss": 4.5309038162231445,
+ "eval_runtime": 56.6931,
+ "eval_samples_per_second": 43.074,
+ "eval_steps_per_second": 1.358,
+ "step": 7056
+ },
+ {
+ "epoch": 98.01397379912663,
+ "grad_norm": 0.013760345987975597,
+ "learning_rate": 0.0006,
+ "loss": 4.059067249298096,
+ "step": 7057
+ },
+ {
+ "epoch": 98.02794759825328,
+ "grad_norm": 0.013799279928207397,
+ "learning_rate": 0.0006,
+ "loss": 3.975419044494629,
+ "step": 7058
+ },
+ {
+ "epoch": 98.04192139737991,
+ "grad_norm": 0.014793423935770988,
+ "learning_rate": 0.0006,
+ "loss": 4.163051605224609,
+ "step": 7059
+ },
+ {
+ "epoch": 98.05589519650655,
+ "grad_norm": 0.015820559114217758,
+ "learning_rate": 0.0006,
+ "loss": 4.032471656799316,
+ "step": 7060
+ },
+ {
+ "epoch": 98.06986899563319,
+ "grad_norm": 0.014252204447984695,
+ "learning_rate": 0.0006,
+ "loss": 4.038850784301758,
+ "step": 7061
+ },
+ {
+ "epoch": 98.08384279475983,
+ "grad_norm": 0.01272337045520544,
+ "learning_rate": 0.0006,
+ "loss": 4.054554462432861,
+ "step": 7062
+ },
+ {
+ "epoch": 98.09781659388646,
+ "grad_norm": 0.01498495228588581,
+ "learning_rate": 0.0006,
+ "loss": 4.009291648864746,
+ "step": 7063
+ },
+ {
+ "epoch": 98.1117903930131,
+ "grad_norm": 0.015501809306442738,
+ "learning_rate": 0.0006,
+ "loss": 3.999412775039673,
+ "step": 7064
+ },
+ {
+ "epoch": 98.12576419213974,
+ "grad_norm": 0.015108918771147728,
+ "learning_rate": 0.0006,
+ "loss": 4.015168190002441,
+ "step": 7065
+ },
+ {
+ "epoch": 98.13973799126637,
+ "grad_norm": 0.01489692460745573,
+ "learning_rate": 0.0006,
+ "loss": 4.163906097412109,
+ "step": 7066
+ },
+ {
+ "epoch": 98.15371179039302,
+ "grad_norm": 0.014246871694922447,
+ "learning_rate": 0.0006,
+ "loss": 3.993173122406006,
+ "step": 7067
+ },
+ {
+ "epoch": 98.16768558951965,
+ "grad_norm": 0.015163703821599483,
+ "learning_rate": 0.0006,
+ "loss": 3.908207893371582,
+ "step": 7068
+ },
+ {
+ "epoch": 98.18165938864628,
+ "grad_norm": 0.0141439288854599,
+ "learning_rate": 0.0006,
+ "loss": 4.064990997314453,
+ "step": 7069
+ },
+ {
+ "epoch": 98.19563318777293,
+ "grad_norm": 0.013251710683107376,
+ "learning_rate": 0.0006,
+ "loss": 4.042545318603516,
+ "step": 7070
+ },
+ {
+ "epoch": 98.20960698689956,
+ "grad_norm": 0.01428674440830946,
+ "learning_rate": 0.0006,
+ "loss": 4.041232585906982,
+ "step": 7071
+ },
+ {
+ "epoch": 98.2235807860262,
+ "grad_norm": 0.012818179093301296,
+ "learning_rate": 0.0006,
+ "loss": 3.9160139560699463,
+ "step": 7072
+ },
+ {
+ "epoch": 98.23755458515284,
+ "grad_norm": 0.012890368700027466,
+ "learning_rate": 0.0006,
+ "loss": 3.9687235355377197,
+ "step": 7073
+ },
+ {
+ "epoch": 98.25152838427948,
+ "grad_norm": 0.014132961630821228,
+ "learning_rate": 0.0006,
+ "loss": 4.066800117492676,
+ "step": 7074
+ },
+ {
+ "epoch": 98.26550218340611,
+ "grad_norm": 0.013544383458793163,
+ "learning_rate": 0.0006,
+ "loss": 4.111777305603027,
+ "step": 7075
+ },
+ {
+ "epoch": 98.27947598253274,
+ "grad_norm": 0.014414685778319836,
+ "learning_rate": 0.0006,
+ "loss": 4.109633445739746,
+ "step": 7076
+ },
+ {
+ "epoch": 98.29344978165939,
+ "grad_norm": 0.015201127156615257,
+ "learning_rate": 0.0006,
+ "loss": 4.033087253570557,
+ "step": 7077
+ },
+ {
+ "epoch": 98.30742358078602,
+ "grad_norm": 0.016936076804995537,
+ "learning_rate": 0.0006,
+ "loss": 4.10897970199585,
+ "step": 7078
+ },
+ {
+ "epoch": 98.32139737991267,
+ "grad_norm": 0.01556481048464775,
+ "learning_rate": 0.0006,
+ "loss": 3.968942165374756,
+ "step": 7079
+ },
+ {
+ "epoch": 98.3353711790393,
+ "grad_norm": 0.015971971675753593,
+ "learning_rate": 0.0006,
+ "loss": 4.024805068969727,
+ "step": 7080
+ },
+ {
+ "epoch": 98.34934497816595,
+ "grad_norm": 0.016742313280701637,
+ "learning_rate": 0.0006,
+ "loss": 4.037862777709961,
+ "step": 7081
+ },
+ {
+ "epoch": 98.36331877729258,
+ "grad_norm": 0.015753688290715218,
+ "learning_rate": 0.0006,
+ "loss": 4.015701770782471,
+ "step": 7082
+ },
+ {
+ "epoch": 98.37729257641921,
+ "grad_norm": 0.014211287721991539,
+ "learning_rate": 0.0006,
+ "loss": 4.068171501159668,
+ "step": 7083
+ },
+ {
+ "epoch": 98.39126637554585,
+ "grad_norm": 0.013903875835239887,
+ "learning_rate": 0.0006,
+ "loss": 4.052516937255859,
+ "step": 7084
+ },
+ {
+ "epoch": 98.40524017467249,
+ "grad_norm": 0.015590954571962357,
+ "learning_rate": 0.0006,
+ "loss": 4.056125164031982,
+ "step": 7085
+ },
+ {
+ "epoch": 98.41921397379913,
+ "grad_norm": 0.016314011067152023,
+ "learning_rate": 0.0006,
+ "loss": 4.014211177825928,
+ "step": 7086
+ },
+ {
+ "epoch": 98.43318777292576,
+ "grad_norm": 0.017208732664585114,
+ "learning_rate": 0.0006,
+ "loss": 4.0615339279174805,
+ "step": 7087
+ },
+ {
+ "epoch": 98.44716157205241,
+ "grad_norm": 0.017213109880685806,
+ "learning_rate": 0.0006,
+ "loss": 3.9580445289611816,
+ "step": 7088
+ },
+ {
+ "epoch": 98.46113537117904,
+ "grad_norm": 0.016858434304594994,
+ "learning_rate": 0.0006,
+ "loss": 3.9932594299316406,
+ "step": 7089
+ },
+ {
+ "epoch": 98.47510917030567,
+ "grad_norm": 0.014381436631083488,
+ "learning_rate": 0.0006,
+ "loss": 4.159034729003906,
+ "step": 7090
+ },
+ {
+ "epoch": 98.48908296943232,
+ "grad_norm": 0.013713826425373554,
+ "learning_rate": 0.0006,
+ "loss": 4.055055141448975,
+ "step": 7091
+ },
+ {
+ "epoch": 98.50305676855895,
+ "grad_norm": 0.014764957129955292,
+ "learning_rate": 0.0006,
+ "loss": 4.059102535247803,
+ "step": 7092
+ },
+ {
+ "epoch": 98.5170305676856,
+ "grad_norm": 0.014522310346364975,
+ "learning_rate": 0.0006,
+ "loss": 4.093504905700684,
+ "step": 7093
+ },
+ {
+ "epoch": 98.53100436681223,
+ "grad_norm": 0.013484988361597061,
+ "learning_rate": 0.0006,
+ "loss": 4.079945087432861,
+ "step": 7094
+ },
+ {
+ "epoch": 98.54497816593886,
+ "grad_norm": 0.013231529854238033,
+ "learning_rate": 0.0006,
+ "loss": 4.046164512634277,
+ "step": 7095
+ },
+ {
+ "epoch": 98.5589519650655,
+ "grad_norm": 0.01392320916056633,
+ "learning_rate": 0.0006,
+ "loss": 3.975592613220215,
+ "step": 7096
+ },
+ {
+ "epoch": 98.57292576419214,
+ "grad_norm": 0.012323490343987942,
+ "learning_rate": 0.0006,
+ "loss": 3.9167721271514893,
+ "step": 7097
+ },
+ {
+ "epoch": 98.58689956331878,
+ "grad_norm": 0.012565829791128635,
+ "learning_rate": 0.0006,
+ "loss": 4.085109710693359,
+ "step": 7098
+ },
+ {
+ "epoch": 98.60087336244541,
+ "grad_norm": 0.013498359359800816,
+ "learning_rate": 0.0006,
+ "loss": 4.102974891662598,
+ "step": 7099
+ },
+ {
+ "epoch": 98.61484716157206,
+ "grad_norm": 0.014182298444211483,
+ "learning_rate": 0.0006,
+ "loss": 4.041601657867432,
+ "step": 7100
+ },
+ {
+ "epoch": 98.62882096069869,
+ "grad_norm": 0.013925517909228802,
+ "learning_rate": 0.0006,
+ "loss": 4.115788459777832,
+ "step": 7101
+ },
+ {
+ "epoch": 98.64279475982532,
+ "grad_norm": 0.03483536094427109,
+ "learning_rate": 0.0006,
+ "loss": 4.105683326721191,
+ "step": 7102
+ },
+ {
+ "epoch": 98.65676855895197,
+ "grad_norm": 0.014278477989137173,
+ "learning_rate": 0.0006,
+ "loss": 4.04194974899292,
+ "step": 7103
+ },
+ {
+ "epoch": 98.6707423580786,
+ "grad_norm": 0.015309838578104973,
+ "learning_rate": 0.0006,
+ "loss": 4.094045162200928,
+ "step": 7104
+ },
+ {
+ "epoch": 98.68471615720524,
+ "grad_norm": 0.015831857919692993,
+ "learning_rate": 0.0006,
+ "loss": 3.9989206790924072,
+ "step": 7105
+ },
+ {
+ "epoch": 98.69868995633188,
+ "grad_norm": 0.014071783982217312,
+ "learning_rate": 0.0006,
+ "loss": 4.0480852127075195,
+ "step": 7106
+ },
+ {
+ "epoch": 98.71266375545852,
+ "grad_norm": 0.014099884778261185,
+ "learning_rate": 0.0006,
+ "loss": 4.0519633293151855,
+ "step": 7107
+ },
+ {
+ "epoch": 98.72663755458515,
+ "grad_norm": 0.0143047496676445,
+ "learning_rate": 0.0006,
+ "loss": 3.9782276153564453,
+ "step": 7108
+ },
+ {
+ "epoch": 98.74061135371178,
+ "grad_norm": 0.013741862028837204,
+ "learning_rate": 0.0006,
+ "loss": 4.035103797912598,
+ "step": 7109
+ },
+ {
+ "epoch": 98.75458515283843,
+ "grad_norm": 0.014014648273587227,
+ "learning_rate": 0.0006,
+ "loss": 4.09281063079834,
+ "step": 7110
+ },
+ {
+ "epoch": 98.76855895196506,
+ "grad_norm": 0.013909446075558662,
+ "learning_rate": 0.0006,
+ "loss": 4.033784866333008,
+ "step": 7111
+ },
+ {
+ "epoch": 98.78253275109171,
+ "grad_norm": 0.013188562355935574,
+ "learning_rate": 0.0006,
+ "loss": 4.095016956329346,
+ "step": 7112
+ },
+ {
+ "epoch": 98.79650655021834,
+ "grad_norm": 0.013292612507939339,
+ "learning_rate": 0.0006,
+ "loss": 4.000730514526367,
+ "step": 7113
+ },
+ {
+ "epoch": 98.81048034934499,
+ "grad_norm": 0.013189040124416351,
+ "learning_rate": 0.0006,
+ "loss": 4.087883949279785,
+ "step": 7114
+ },
+ {
+ "epoch": 98.82445414847162,
+ "grad_norm": 0.013953711837530136,
+ "learning_rate": 0.0006,
+ "loss": 4.051388740539551,
+ "step": 7115
+ },
+ {
+ "epoch": 98.83842794759825,
+ "grad_norm": 0.014403220266103745,
+ "learning_rate": 0.0006,
+ "loss": 3.9786248207092285,
+ "step": 7116
+ },
+ {
+ "epoch": 98.8524017467249,
+ "grad_norm": 0.013747540302574635,
+ "learning_rate": 0.0006,
+ "loss": 3.9938254356384277,
+ "step": 7117
+ },
+ {
+ "epoch": 98.86637554585153,
+ "grad_norm": 0.014865634962916374,
+ "learning_rate": 0.0006,
+ "loss": 4.075680255889893,
+ "step": 7118
+ },
+ {
+ "epoch": 98.88034934497817,
+ "grad_norm": 0.015241672284901142,
+ "learning_rate": 0.0006,
+ "loss": 3.955733299255371,
+ "step": 7119
+ },
+ {
+ "epoch": 98.8943231441048,
+ "grad_norm": 0.013307510875165462,
+ "learning_rate": 0.0006,
+ "loss": 3.94956374168396,
+ "step": 7120
+ },
+ {
+ "epoch": 98.90829694323143,
+ "grad_norm": 0.013500694185495377,
+ "learning_rate": 0.0006,
+ "loss": 4.120021343231201,
+ "step": 7121
+ },
+ {
+ "epoch": 98.92227074235808,
+ "grad_norm": 0.015430806204676628,
+ "learning_rate": 0.0006,
+ "loss": 4.108362674713135,
+ "step": 7122
+ },
+ {
+ "epoch": 98.93624454148471,
+ "grad_norm": 0.015447405166924,
+ "learning_rate": 0.0006,
+ "loss": 4.040914058685303,
+ "step": 7123
+ },
+ {
+ "epoch": 98.95021834061136,
+ "grad_norm": 0.014040131121873856,
+ "learning_rate": 0.0006,
+ "loss": 4.086688995361328,
+ "step": 7124
+ },
+ {
+ "epoch": 98.96419213973799,
+ "grad_norm": 0.0136969443410635,
+ "learning_rate": 0.0006,
+ "loss": 4.083828926086426,
+ "step": 7125
+ },
+ {
+ "epoch": 98.97816593886463,
+ "grad_norm": 0.01413224171847105,
+ "learning_rate": 0.0006,
+ "loss": 4.142146110534668,
+ "step": 7126
+ },
+ {
+ "epoch": 98.99213973799127,
+ "grad_norm": 0.012932452373206615,
+ "learning_rate": 0.0006,
+ "loss": 4.1193342208862305,
+ "step": 7127
+ },
+ {
+ "epoch": 99.0,
+ "grad_norm": 0.01586783304810524,
+ "learning_rate": 0.0006,
+ "loss": 4.191825866699219,
+ "step": 7128
+ },
+ {
+ "epoch": 99.0,
+ "eval_loss": 4.550753116607666,
+ "eval_runtime": 56.4315,
+ "eval_samples_per_second": 43.274,
+ "eval_steps_per_second": 1.364,
+ "step": 7128
+ },
+ {
+ "epoch": 99.01397379912663,
+ "grad_norm": 0.016847660765051842,
+ "learning_rate": 0.0006,
+ "loss": 3.9874870777130127,
+ "step": 7129
+ },
+ {
+ "epoch": 99.02794759825328,
+ "grad_norm": 0.016426295042037964,
+ "learning_rate": 0.0006,
+ "loss": 4.108767032623291,
+ "step": 7130
+ },
+ {
+ "epoch": 99.04192139737991,
+ "grad_norm": 0.01687243953347206,
+ "learning_rate": 0.0006,
+ "loss": 3.9658308029174805,
+ "step": 7131
+ },
+ {
+ "epoch": 99.05589519650655,
+ "grad_norm": 0.016468465328216553,
+ "learning_rate": 0.0006,
+ "loss": 3.91381573677063,
+ "step": 7132
+ },
+ {
+ "epoch": 99.06986899563319,
+ "grad_norm": 0.013777337968349457,
+ "learning_rate": 0.0006,
+ "loss": 3.9960741996765137,
+ "step": 7133
+ },
+ {
+ "epoch": 99.08384279475983,
+ "grad_norm": 0.014061779715120792,
+ "learning_rate": 0.0006,
+ "loss": 4.069479942321777,
+ "step": 7134
+ },
+ {
+ "epoch": 99.09781659388646,
+ "grad_norm": 0.01513410173356533,
+ "learning_rate": 0.0006,
+ "loss": 4.145164489746094,
+ "step": 7135
+ },
+ {
+ "epoch": 99.1117903930131,
+ "grad_norm": 0.01567850261926651,
+ "learning_rate": 0.0006,
+ "loss": 4.016868591308594,
+ "step": 7136
+ },
+ {
+ "epoch": 99.12576419213974,
+ "grad_norm": 0.01662704534828663,
+ "learning_rate": 0.0006,
+ "loss": 3.966663360595703,
+ "step": 7137
+ },
+ {
+ "epoch": 99.13973799126637,
+ "grad_norm": 0.01690712384879589,
+ "learning_rate": 0.0006,
+ "loss": 4.003656387329102,
+ "step": 7138
+ },
+ {
+ "epoch": 99.15371179039302,
+ "grad_norm": 0.01541030965745449,
+ "learning_rate": 0.0006,
+ "loss": 3.9990856647491455,
+ "step": 7139
+ },
+ {
+ "epoch": 99.16768558951965,
+ "grad_norm": 0.014438044279813766,
+ "learning_rate": 0.0006,
+ "loss": 4.089632034301758,
+ "step": 7140
+ },
+ {
+ "epoch": 99.18165938864628,
+ "grad_norm": 0.01470417995005846,
+ "learning_rate": 0.0006,
+ "loss": 4.018681049346924,
+ "step": 7141
+ },
+ {
+ "epoch": 99.19563318777293,
+ "grad_norm": 0.015567616559565067,
+ "learning_rate": 0.0006,
+ "loss": 4.067337989807129,
+ "step": 7142
+ },
+ {
+ "epoch": 99.20960698689956,
+ "grad_norm": 0.01784040965139866,
+ "learning_rate": 0.0006,
+ "loss": 4.083293914794922,
+ "step": 7143
+ },
+ {
+ "epoch": 99.2235807860262,
+ "grad_norm": 0.017351210117340088,
+ "learning_rate": 0.0006,
+ "loss": 4.115224361419678,
+ "step": 7144
+ },
+ {
+ "epoch": 99.23755458515284,
+ "grad_norm": 0.014561209827661514,
+ "learning_rate": 0.0006,
+ "loss": 4.1001057624816895,
+ "step": 7145
+ },
+ {
+ "epoch": 99.25152838427948,
+ "grad_norm": 0.013986383564770222,
+ "learning_rate": 0.0006,
+ "loss": 3.9758031368255615,
+ "step": 7146
+ },
+ {
+ "epoch": 99.26550218340611,
+ "grad_norm": 0.014053978025913239,
+ "learning_rate": 0.0006,
+ "loss": 4.1420369148254395,
+ "step": 7147
+ },
+ {
+ "epoch": 99.27947598253274,
+ "grad_norm": 0.014878278598189354,
+ "learning_rate": 0.0006,
+ "loss": 4.018189430236816,
+ "step": 7148
+ },
+ {
+ "epoch": 99.29344978165939,
+ "grad_norm": 0.014634845778346062,
+ "learning_rate": 0.0006,
+ "loss": 4.062045574188232,
+ "step": 7149
+ },
+ {
+ "epoch": 99.30742358078602,
+ "grad_norm": 0.014713970944285393,
+ "learning_rate": 0.0006,
+ "loss": 4.017163276672363,
+ "step": 7150
+ },
+ {
+ "epoch": 99.32139737991267,
+ "grad_norm": 0.01589985005557537,
+ "learning_rate": 0.0006,
+ "loss": 4.093014717102051,
+ "step": 7151
+ },
+ {
+ "epoch": 99.3353711790393,
+ "grad_norm": 0.016232378780841827,
+ "learning_rate": 0.0006,
+ "loss": 4.029753684997559,
+ "step": 7152
+ },
+ {
+ "epoch": 99.34934497816595,
+ "grad_norm": 0.016340898349881172,
+ "learning_rate": 0.0006,
+ "loss": 4.056434631347656,
+ "step": 7153
+ },
+ {
+ "epoch": 99.36331877729258,
+ "grad_norm": 0.01587427407503128,
+ "learning_rate": 0.0006,
+ "loss": 3.932020425796509,
+ "step": 7154
+ },
+ {
+ "epoch": 99.37729257641921,
+ "grad_norm": 0.014977093786001205,
+ "learning_rate": 0.0006,
+ "loss": 4.015125751495361,
+ "step": 7155
+ },
+ {
+ "epoch": 99.39126637554585,
+ "grad_norm": 0.014813411049544811,
+ "learning_rate": 0.0006,
+ "loss": 4.088035583496094,
+ "step": 7156
+ },
+ {
+ "epoch": 99.40524017467249,
+ "grad_norm": 0.014407774433493614,
+ "learning_rate": 0.0006,
+ "loss": 3.9979054927825928,
+ "step": 7157
+ },
+ {
+ "epoch": 99.41921397379913,
+ "grad_norm": 0.013969994150102139,
+ "learning_rate": 0.0006,
+ "loss": 4.180194854736328,
+ "step": 7158
+ },
+ {
+ "epoch": 99.43318777292576,
+ "grad_norm": 0.0159455519169569,
+ "learning_rate": 0.0006,
+ "loss": 4.049341201782227,
+ "step": 7159
+ },
+ {
+ "epoch": 99.44716157205241,
+ "grad_norm": 0.016426580026745796,
+ "learning_rate": 0.0006,
+ "loss": 4.01624059677124,
+ "step": 7160
+ },
+ {
+ "epoch": 99.46113537117904,
+ "grad_norm": 0.014921349473297596,
+ "learning_rate": 0.0006,
+ "loss": 4.003138542175293,
+ "step": 7161
+ },
+ {
+ "epoch": 99.47510917030567,
+ "grad_norm": 0.014607159420847893,
+ "learning_rate": 0.0006,
+ "loss": 3.993896484375,
+ "step": 7162
+ },
+ {
+ "epoch": 99.48908296943232,
+ "grad_norm": 0.014638705179095268,
+ "learning_rate": 0.0006,
+ "loss": 4.015478134155273,
+ "step": 7163
+ },
+ {
+ "epoch": 99.50305676855895,
+ "grad_norm": 0.01346661988645792,
+ "learning_rate": 0.0006,
+ "loss": 3.9999020099639893,
+ "step": 7164
+ },
+ {
+ "epoch": 99.5170305676856,
+ "grad_norm": 0.013515726663172245,
+ "learning_rate": 0.0006,
+ "loss": 4.045872688293457,
+ "step": 7165
+ },
+ {
+ "epoch": 99.53100436681223,
+ "grad_norm": 0.013773965649306774,
+ "learning_rate": 0.0006,
+ "loss": 4.094198703765869,
+ "step": 7166
+ },
+ {
+ "epoch": 99.54497816593886,
+ "grad_norm": 0.014282289892435074,
+ "learning_rate": 0.0006,
+ "loss": 4.091470718383789,
+ "step": 7167
+ },
+ {
+ "epoch": 99.5589519650655,
+ "grad_norm": 0.015272347256541252,
+ "learning_rate": 0.0006,
+ "loss": 4.026535987854004,
+ "step": 7168
+ },
+ {
+ "epoch": 99.57292576419214,
+ "grad_norm": 0.015618950128555298,
+ "learning_rate": 0.0006,
+ "loss": 4.006448745727539,
+ "step": 7169
+ },
+ {
+ "epoch": 99.58689956331878,
+ "grad_norm": 0.015366379171609879,
+ "learning_rate": 0.0006,
+ "loss": 4.025970458984375,
+ "step": 7170
+ },
+ {
+ "epoch": 99.60087336244541,
+ "grad_norm": 0.013261756859719753,
+ "learning_rate": 0.0006,
+ "loss": 3.980401039123535,
+ "step": 7171
+ },
+ {
+ "epoch": 99.61484716157206,
+ "grad_norm": 0.013438165187835693,
+ "learning_rate": 0.0006,
+ "loss": 4.0106096267700195,
+ "step": 7172
+ },
+ {
+ "epoch": 99.62882096069869,
+ "grad_norm": 0.015591824427247047,
+ "learning_rate": 0.0006,
+ "loss": 4.0458784103393555,
+ "step": 7173
+ },
+ {
+ "epoch": 99.64279475982532,
+ "grad_norm": 0.014863507822155952,
+ "learning_rate": 0.0006,
+ "loss": 4.014740943908691,
+ "step": 7174
+ },
+ {
+ "epoch": 99.65676855895197,
+ "grad_norm": 0.012508790008723736,
+ "learning_rate": 0.0006,
+ "loss": 4.022995948791504,
+ "step": 7175
+ },
+ {
+ "epoch": 99.6707423580786,
+ "grad_norm": 0.013107656501233578,
+ "learning_rate": 0.0006,
+ "loss": 4.085544109344482,
+ "step": 7176
+ },
+ {
+ "epoch": 99.68471615720524,
+ "grad_norm": 0.013550374656915665,
+ "learning_rate": 0.0006,
+ "loss": 4.1239237785339355,
+ "step": 7177
+ },
+ {
+ "epoch": 99.69868995633188,
+ "grad_norm": 0.012947923503816128,
+ "learning_rate": 0.0006,
+ "loss": 4.01287317276001,
+ "step": 7178
+ },
+ {
+ "epoch": 99.71266375545852,
+ "grad_norm": 0.015236291103065014,
+ "learning_rate": 0.0006,
+ "loss": 4.075140476226807,
+ "step": 7179
+ },
+ {
+ "epoch": 99.72663755458515,
+ "grad_norm": 0.014862542040646076,
+ "learning_rate": 0.0006,
+ "loss": 4.111439228057861,
+ "step": 7180
+ },
+ {
+ "epoch": 99.74061135371178,
+ "grad_norm": 0.013883531093597412,
+ "learning_rate": 0.0006,
+ "loss": 3.9860856533050537,
+ "step": 7181
+ },
+ {
+ "epoch": 99.75458515283843,
+ "grad_norm": 0.01507870014756918,
+ "learning_rate": 0.0006,
+ "loss": 4.096504211425781,
+ "step": 7182
+ },
+ {
+ "epoch": 99.76855895196506,
+ "grad_norm": 0.015378627926111221,
+ "learning_rate": 0.0006,
+ "loss": 4.034113883972168,
+ "step": 7183
+ },
+ {
+ "epoch": 99.78253275109171,
+ "grad_norm": 0.013696327805519104,
+ "learning_rate": 0.0006,
+ "loss": 4.026954650878906,
+ "step": 7184
+ },
+ {
+ "epoch": 99.79650655021834,
+ "grad_norm": 0.013519738800823689,
+ "learning_rate": 0.0006,
+ "loss": 4.040035247802734,
+ "step": 7185
+ },
+ {
+ "epoch": 99.81048034934499,
+ "grad_norm": 0.015677500516176224,
+ "learning_rate": 0.0006,
+ "loss": 4.057739734649658,
+ "step": 7186
+ },
+ {
+ "epoch": 99.82445414847162,
+ "grad_norm": 0.014739335514605045,
+ "learning_rate": 0.0006,
+ "loss": 4.022522449493408,
+ "step": 7187
+ },
+ {
+ "epoch": 99.83842794759825,
+ "grad_norm": 0.014079970307648182,
+ "learning_rate": 0.0006,
+ "loss": 3.9286317825317383,
+ "step": 7188
+ },
+ {
+ "epoch": 99.8524017467249,
+ "grad_norm": 0.014899874106049538,
+ "learning_rate": 0.0006,
+ "loss": 4.081925868988037,
+ "step": 7189
+ },
+ {
+ "epoch": 99.86637554585153,
+ "grad_norm": 0.014767617918550968,
+ "learning_rate": 0.0006,
+ "loss": 4.029382705688477,
+ "step": 7190
+ },
+ {
+ "epoch": 99.88034934497817,
+ "grad_norm": 0.014914393424987793,
+ "learning_rate": 0.0006,
+ "loss": 4.056460857391357,
+ "step": 7191
+ },
+ {
+ "epoch": 99.8943231441048,
+ "grad_norm": 0.01720450073480606,
+ "learning_rate": 0.0006,
+ "loss": 4.0184478759765625,
+ "step": 7192
+ },
+ {
+ "epoch": 99.90829694323143,
+ "grad_norm": 0.016545087099075317,
+ "learning_rate": 0.0006,
+ "loss": 3.9638831615448,
+ "step": 7193
+ },
+ {
+ "epoch": 99.92227074235808,
+ "grad_norm": 0.014489716850221157,
+ "learning_rate": 0.0006,
+ "loss": 4.0267653465271,
+ "step": 7194
+ },
+ {
+ "epoch": 99.93624454148471,
+ "grad_norm": 0.01438206061720848,
+ "learning_rate": 0.0006,
+ "loss": 4.212850570678711,
+ "step": 7195
+ },
+ {
+ "epoch": 99.95021834061136,
+ "grad_norm": 0.017636865377426147,
+ "learning_rate": 0.0006,
+ "loss": 3.942308187484741,
+ "step": 7196
+ },
+ {
+ "epoch": 99.96419213973799,
+ "grad_norm": 0.019075388088822365,
+ "learning_rate": 0.0006,
+ "loss": 4.026660919189453,
+ "step": 7197
+ },
+ {
+ "epoch": 99.97816593886463,
+ "grad_norm": 0.016569023951888084,
+ "learning_rate": 0.0006,
+ "loss": 4.018840789794922,
+ "step": 7198
+ },
+ {
+ "epoch": 99.99213973799127,
+ "grad_norm": 0.014842798933386803,
+ "learning_rate": 0.0006,
+ "loss": 4.041904449462891,
+ "step": 7199
+ },
+ {
+ "epoch": 100.0,
+ "grad_norm": 0.0167684406042099,
+ "learning_rate": 0.0006,
+ "loss": 3.9079346656799316,
+ "step": 7200
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 28800,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 400,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": true,
+ "should_log": false,
+ "should_save": false,
+ "should_training_stop": false
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 3.069613361332224e+18,
+ "train_batch_size": 8,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-7200/training_args.bin b/runs/l2r50-i2-fulle-lm/checkpoint-7200/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..cad232831c222cc7f8b938b8bbe42f61e7d89c3a
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-7200/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:f7c40e3bfa2f5ac1450ca1243e3f2a895d44a06ed5995e6aefd97f78d229d8c8
+size 4792
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-7488/chat_template.jinja b/runs/l2r50-i2-fulle-lm/checkpoint-7488/chat_template.jinja
new file mode 100644
index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-7488/chat_template.jinja
@@ -0,0 +1,85 @@
+{%- if tools %}
+ {{- '<|im_start|>system\n' }}
+ {%- if messages[0].role == 'system' %}
+ {{- messages[0].content + '\n\n' }}
+ {%- endif %}
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }}
+ {%- for tool in tools %}
+ {{- "\n" }}
+ {{- tool | tojson }}
+ {%- endfor %}
+ {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }}
+{%- else %}
+ {%- if messages[0].role == 'system' %}
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
+ {%- endif %}
+{%- endif %}
+{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
+{%- for message in messages[::-1] %}
+ {%- set index = (messages|length - 1) - loop.index0 %}
+ {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %}
+ {%- set ns.multi_step_tool = false %}
+ {%- set ns.last_query_index = index %}
+ {%- endif %}
+{%- endfor %}
+{%- for message in messages %}
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
+ {%- elif message.role == "assistant" %}
+ {%- set content = message.content %}
+ {%- set reasoning_content = '' %}
+ {%- if message.reasoning_content is defined and message.reasoning_content is not none %}
+ {%- set reasoning_content = message.reasoning_content %}
+ {%- else %}
+ {%- if '' in message.content %}
+ {%- set content = message.content.split('')[-1].lstrip('\n') %}
+ {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %}
+ {%- endif %}
+ {%- endif %}
+ {%- if loop.index0 > ns.last_query_index %}
+ {%- if loop.last or (not loop.last and reasoning_content) %}
+ {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- if message.tool_calls %}
+ {%- for tool_call in message.tool_calls %}
+ {%- if (loop.first and content) or (not loop.first) %}
+ {{- '\n' }}
+ {%- endif %}
+ {%- if tool_call.function %}
+ {%- set tool_call = tool_call.function %}
+ {%- endif %}
+ {{- '\n{"name": "' }}
+ {{- tool_call.name }}
+ {{- '", "arguments": ' }}
+ {%- if tool_call.arguments is string %}
+ {{- tool_call.arguments }}
+ {%- else %}
+ {{- tool_call.arguments | tojson }}
+ {%- endif %}
+ {{- '}\n' }}
+ {%- endfor %}
+ {%- endif %}
+ {{- '<|im_end|>\n' }}
+ {%- elif message.role == "tool" %}
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
+ {{- '<|im_start|>user' }}
+ {%- endif %}
+ {{- '\n\n' }}
+ {{- message.content }}
+ {{- '\n' }}
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
+ {{- '<|im_end|>\n' }}
+ {%- endif %}
+ {%- endif %}
+{%- endfor %}
+{%- if add_generation_prompt %}
+ {{- '<|im_start|>assistant\n' }}
+ {%- if enable_thinking is defined and enable_thinking is false %}
+ {{- '\n\n\n\n' }}
+ {%- endif %}
+{%- endif %}
\ No newline at end of file
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-7488/config.json b/runs/l2r50-i2-fulle-lm/checkpoint-7488/config.json
new file mode 100644
index 0000000000000000000000000000000000000000..22f250953c6ce8fc91ee4411856dd73041a415e2
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-7488/config.json
@@ -0,0 +1,32 @@
+{
+ "architectures": [
+ "LlamaForCausalLM"
+ ],
+ "attention_bias": false,
+ "attention_dropout": 0.0,
+ "bos_token_id": null,
+ "dtype": "float32",
+ "eos_token_id": 151645,
+ "head_dim": 128,
+ "hidden_act": "silu",
+ "hidden_size": 512,
+ "initializer_range": 0.02,
+ "intermediate_size": 1536,
+ "max_position_embeddings": 2048,
+ "mlp_bias": false,
+ "model_type": "llama",
+ "num_attention_heads": 4,
+ "num_hidden_layers": 20,
+ "num_key_value_heads": 4,
+ "pad_token_id": 151645,
+ "pretraining_tp": 1,
+ "rms_norm_eps": 1e-06,
+ "rope_parameters": {
+ "rope_theta": 10000.0,
+ "rope_type": "default"
+ },
+ "tie_word_embeddings": true,
+ "transformers_version": "5.5.0",
+ "use_cache": false,
+ "vocab_size": 151673
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-7488/generation_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-7488/generation_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-7488/generation_config.json
@@ -0,0 +1,11 @@
+{
+ "_from_model_config": true,
+ "eos_token_id": [
+ 151645
+ ],
+ "output_attentions": false,
+ "output_hidden_states": false,
+ "pad_token_id": 151645,
+ "transformers_version": "5.5.0",
+ "use_cache": true
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-7488/model.safetensors b/runs/l2r50-i2-fulle-lm/checkpoint-7488/model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..f144376bd538aaf7ae17f67f42915e51e25e1606
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-7488/model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:a3ec22916dffc70ec6616d3654e31d00a4a10c6ca1d1ebb5120714499328f0e1
+size 583360328
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-7488/optimizer.pt b/runs/l2r50-i2-fulle-lm/checkpoint-7488/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..5fbee4a581363ebb84c0136b3b253f02e01f4a89
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-7488/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:03d1a157ba5bd43a017408e321b80745d9c02dbb41a6dba2efb76d6c945344aa
+size 1166833530
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-7488/rng_state_0.pth b/runs/l2r50-i2-fulle-lm/checkpoint-7488/rng_state_0.pth
new file mode 100644
index 0000000000000000000000000000000000000000..61c0b7ef05f2f915a817eff75c694be3671b38f8
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-7488/rng_state_0.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:46f7bd2ee110b8044cafb2a9f1de5db77692de8efefbdc19f5143ad386e6245d
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-7488/rng_state_1.pth b/runs/l2r50-i2-fulle-lm/checkpoint-7488/rng_state_1.pth
new file mode 100644
index 0000000000000000000000000000000000000000..fb2ee3987752f036205b24ae6005def818f4545a
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-7488/rng_state_1.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:4d2ee4a11604a3be729b14d3e61f4e3d57002cafd565766464aec5477f1200da
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-7488/rng_state_2.pth b/runs/l2r50-i2-fulle-lm/checkpoint-7488/rng_state_2.pth
new file mode 100644
index 0000000000000000000000000000000000000000..e6c968a3d5ae020d0a1140c25257d1455b0fc9da
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-7488/rng_state_2.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:0ebce020dff1b60aa525ffeb3690e9655bea876e28d9e589bb23a37836707c3d
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-7488/rng_state_3.pth b/runs/l2r50-i2-fulle-lm/checkpoint-7488/rng_state_3.pth
new file mode 100644
index 0000000000000000000000000000000000000000..f1667e253147f06eace35355fed65f57719cd3f2
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-7488/rng_state_3.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:b2b882c43396f99a07747905c96e2976f1484c5177ccc0805b42fcfc137980d0
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-7488/scheduler.pt b/runs/l2r50-i2-fulle-lm/checkpoint-7488/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..132f9886976cefcb2664879521dcc5aa82c41d3f
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-7488/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:af32b492fedaa58d69ad948e745403317a9384ef5bb7e3fec0e073cc81d818a2
+size 1064
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-7488/tokenizer.json b/runs/l2r50-i2-fulle-lm/checkpoint-7488/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..53786016f070e11184813f3c83101dc5e93e392e
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-7488/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:c3c63fda1646a555448e8f03662542e04b11d103721319e4d9b267d7da03ceb8
+size 11424463
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-7488/tokenizer_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-7488/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..98e2fb9b2cc54785dc9d871b8e503f8b8f2defe7
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-7488/tokenizer_config.json
@@ -0,0 +1,20 @@
+{
+ "add_prefix_space": false,
+ "backend": "tokenizers",
+ "bos_token": null,
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "<|im_end|>",
+ "errors": "replace",
+ "extra_special_tokens": [
+ "<|l2r_pred|>",
+ "<|r2l_pred|>",
+ "<|next_1_pred|>",
+ "<|next_2_pred|>"
+ ],
+ "is_local": false,
+ "model_max_length": 131072,
+ "pad_token": "<|im_end|>",
+ "split_special_tokens": false,
+ "tokenizer_class": "Qwen2Tokenizer",
+ "unk_token": null
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-7488/trainer_state.json b/runs/l2r50-i2-fulle-lm/checkpoint-7488/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..69f2093714aae16a5aa9d2abe0d7bdb2cd3bd779
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-7488/trainer_state.json
@@ -0,0 +1,53274 @@
+{
+ "best_global_step": null,
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 104.0,
+ "eval_steps": 500,
+ "global_step": 7488,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.013973799126637555,
+ "grad_norm": 0.1739506870508194,
+ "learning_rate": 0.0,
+ "loss": 11.99299430847168,
+ "step": 1
+ },
+ {
+ "epoch": 0.02794759825327511,
+ "grad_norm": 0.17216597497463226,
+ "learning_rate": 5.999999999999999e-06,
+ "loss": 11.988704681396484,
+ "step": 2
+ },
+ {
+ "epoch": 0.04192139737991266,
+ "grad_norm": 0.17459715902805328,
+ "learning_rate": 1.1999999999999999e-05,
+ "loss": 11.945581436157227,
+ "step": 3
+ },
+ {
+ "epoch": 0.05589519650655022,
+ "grad_norm": 0.17577414214611053,
+ "learning_rate": 1.7999999999999997e-05,
+ "loss": 11.859772682189941,
+ "step": 4
+ },
+ {
+ "epoch": 0.06986899563318777,
+ "grad_norm": 0.16163651645183563,
+ "learning_rate": 2.3999999999999997e-05,
+ "loss": 11.744415283203125,
+ "step": 5
+ },
+ {
+ "epoch": 0.08384279475982533,
+ "grad_norm": 0.14409932494163513,
+ "learning_rate": 2.9999999999999997e-05,
+ "loss": 11.618099212646484,
+ "step": 6
+ },
+ {
+ "epoch": 0.09781659388646288,
+ "grad_norm": 0.13079848885536194,
+ "learning_rate": 3.5999999999999994e-05,
+ "loss": 11.505935668945312,
+ "step": 7
+ },
+ {
+ "epoch": 0.11179039301310044,
+ "grad_norm": 0.11964613944292068,
+ "learning_rate": 4.2e-05,
+ "loss": 11.416532516479492,
+ "step": 8
+ },
+ {
+ "epoch": 0.125764192139738,
+ "grad_norm": 0.11345142871141434,
+ "learning_rate": 4.7999999999999994e-05,
+ "loss": 11.33206558227539,
+ "step": 9
+ },
+ {
+ "epoch": 0.13973799126637554,
+ "grad_norm": 0.11033952981233597,
+ "learning_rate": 5.399999999999999e-05,
+ "loss": 11.259052276611328,
+ "step": 10
+ },
+ {
+ "epoch": 0.1537117903930131,
+ "grad_norm": 0.10776489973068237,
+ "learning_rate": 5.9999999999999995e-05,
+ "loss": 11.203384399414062,
+ "step": 11
+ },
+ {
+ "epoch": 0.16768558951965065,
+ "grad_norm": 0.10721946507692337,
+ "learning_rate": 6.599999999999999e-05,
+ "loss": 11.153295516967773,
+ "step": 12
+ },
+ {
+ "epoch": 0.18165938864628822,
+ "grad_norm": 0.10673430562019348,
+ "learning_rate": 7.199999999999999e-05,
+ "loss": 11.110200881958008,
+ "step": 13
+ },
+ {
+ "epoch": 0.19563318777292577,
+ "grad_norm": 0.10630254447460175,
+ "learning_rate": 7.8e-05,
+ "loss": 11.067477226257324,
+ "step": 14
+ },
+ {
+ "epoch": 0.2096069868995633,
+ "grad_norm": 0.10631615668535233,
+ "learning_rate": 8.4e-05,
+ "loss": 11.022941589355469,
+ "step": 15
+ },
+ {
+ "epoch": 0.22358078602620088,
+ "grad_norm": 0.10609443485736847,
+ "learning_rate": 8.999999999999999e-05,
+ "loss": 10.970746040344238,
+ "step": 16
+ },
+ {
+ "epoch": 0.23755458515283842,
+ "grad_norm": 0.10648232698440552,
+ "learning_rate": 9.599999999999999e-05,
+ "loss": 10.91267204284668,
+ "step": 17
+ },
+ {
+ "epoch": 0.251528384279476,
+ "grad_norm": 0.10584589838981628,
+ "learning_rate": 0.000102,
+ "loss": 10.854547500610352,
+ "step": 18
+ },
+ {
+ "epoch": 0.26550218340611353,
+ "grad_norm": 0.10591720044612885,
+ "learning_rate": 0.00010799999999999998,
+ "loss": 10.788917541503906,
+ "step": 19
+ },
+ {
+ "epoch": 0.2794759825327511,
+ "grad_norm": 0.10611672699451447,
+ "learning_rate": 0.00011399999999999999,
+ "loss": 10.717556953430176,
+ "step": 20
+ },
+ {
+ "epoch": 0.2934497816593886,
+ "grad_norm": 0.10455359518527985,
+ "learning_rate": 0.00011999999999999999,
+ "loss": 10.650186538696289,
+ "step": 21
+ },
+ {
+ "epoch": 0.3074235807860262,
+ "grad_norm": 0.1047525629401207,
+ "learning_rate": 0.00012599999999999997,
+ "loss": 10.56808090209961,
+ "step": 22
+ },
+ {
+ "epoch": 0.32139737991266376,
+ "grad_norm": 0.10511605441570282,
+ "learning_rate": 0.00013199999999999998,
+ "loss": 10.484028816223145,
+ "step": 23
+ },
+ {
+ "epoch": 0.3353711790393013,
+ "grad_norm": 0.10467177629470825,
+ "learning_rate": 0.000138,
+ "loss": 10.399947166442871,
+ "step": 24
+ },
+ {
+ "epoch": 0.34934497816593885,
+ "grad_norm": 0.10432136058807373,
+ "learning_rate": 0.00014399999999999998,
+ "loss": 10.312714576721191,
+ "step": 25
+ },
+ {
+ "epoch": 0.36331877729257644,
+ "grad_norm": 0.10527262091636658,
+ "learning_rate": 0.00015,
+ "loss": 10.212261199951172,
+ "step": 26
+ },
+ {
+ "epoch": 0.377292576419214,
+ "grad_norm": 0.103610560297966,
+ "learning_rate": 0.000156,
+ "loss": 10.132034301757812,
+ "step": 27
+ },
+ {
+ "epoch": 0.39126637554585153,
+ "grad_norm": 0.10553059726953506,
+ "learning_rate": 0.000162,
+ "loss": 10.020434379577637,
+ "step": 28
+ },
+ {
+ "epoch": 0.4052401746724891,
+ "grad_norm": 0.1037299782037735,
+ "learning_rate": 0.000168,
+ "loss": 9.934466361999512,
+ "step": 29
+ },
+ {
+ "epoch": 0.4192139737991266,
+ "grad_norm": 0.1042739674448967,
+ "learning_rate": 0.00017399999999999997,
+ "loss": 9.822539329528809,
+ "step": 30
+ },
+ {
+ "epoch": 0.4331877729257642,
+ "grad_norm": 0.10300865024328232,
+ "learning_rate": 0.00017999999999999998,
+ "loss": 9.734994888305664,
+ "step": 31
+ },
+ {
+ "epoch": 0.44716157205240176,
+ "grad_norm": 0.10291654616594315,
+ "learning_rate": 0.000186,
+ "loss": 9.624505996704102,
+ "step": 32
+ },
+ {
+ "epoch": 0.4611353711790393,
+ "grad_norm": 0.10210851579904556,
+ "learning_rate": 0.00019199999999999998,
+ "loss": 9.518205642700195,
+ "step": 33
+ },
+ {
+ "epoch": 0.47510917030567684,
+ "grad_norm": 0.10222529619932175,
+ "learning_rate": 0.000198,
+ "loss": 9.412222862243652,
+ "step": 34
+ },
+ {
+ "epoch": 0.4890829694323144,
+ "grad_norm": 0.09992019087076187,
+ "learning_rate": 0.000204,
+ "loss": 9.326566696166992,
+ "step": 35
+ },
+ {
+ "epoch": 0.503056768558952,
+ "grad_norm": 0.09931084513664246,
+ "learning_rate": 0.00020999999999999998,
+ "loss": 9.210111618041992,
+ "step": 36
+ },
+ {
+ "epoch": 0.5170305676855895,
+ "grad_norm": 0.09928256273269653,
+ "learning_rate": 0.00021599999999999996,
+ "loss": 9.100493431091309,
+ "step": 37
+ },
+ {
+ "epoch": 0.5310043668122271,
+ "grad_norm": 0.09843816608190536,
+ "learning_rate": 0.00022199999999999998,
+ "loss": 8.996543884277344,
+ "step": 38
+ },
+ {
+ "epoch": 0.5449781659388646,
+ "grad_norm": 0.09825374186038971,
+ "learning_rate": 0.00022799999999999999,
+ "loss": 8.888991355895996,
+ "step": 39
+ },
+ {
+ "epoch": 0.5589519650655022,
+ "grad_norm": 0.09602385759353638,
+ "learning_rate": 0.000234,
+ "loss": 8.802907943725586,
+ "step": 40
+ },
+ {
+ "epoch": 0.5729257641921397,
+ "grad_norm": 0.09414064139127731,
+ "learning_rate": 0.00023999999999999998,
+ "loss": 8.715423583984375,
+ "step": 41
+ },
+ {
+ "epoch": 0.5868995633187772,
+ "grad_norm": 0.09273893386125565,
+ "learning_rate": 0.00024599999999999996,
+ "loss": 8.61374282836914,
+ "step": 42
+ },
+ {
+ "epoch": 0.6008733624454149,
+ "grad_norm": 0.09079599380493164,
+ "learning_rate": 0.00025199999999999995,
+ "loss": 8.502904891967773,
+ "step": 43
+ },
+ {
+ "epoch": 0.6148471615720524,
+ "grad_norm": 0.08605825901031494,
+ "learning_rate": 0.000258,
+ "loss": 8.453008651733398,
+ "step": 44
+ },
+ {
+ "epoch": 0.62882096069869,
+ "grad_norm": 0.08338996022939682,
+ "learning_rate": 0.00026399999999999997,
+ "loss": 8.365163803100586,
+ "step": 45
+ },
+ {
+ "epoch": 0.6427947598253275,
+ "grad_norm": 0.08150867372751236,
+ "learning_rate": 0.00027,
+ "loss": 8.285372734069824,
+ "step": 46
+ },
+ {
+ "epoch": 0.6567685589519651,
+ "grad_norm": 0.07794807851314545,
+ "learning_rate": 0.000276,
+ "loss": 8.185811996459961,
+ "step": 47
+ },
+ {
+ "epoch": 0.6707423580786026,
+ "grad_norm": 0.07592720538377762,
+ "learning_rate": 0.00028199999999999997,
+ "loss": 8.098482131958008,
+ "step": 48
+ },
+ {
+ "epoch": 0.6847161572052402,
+ "grad_norm": 0.0709897130727768,
+ "learning_rate": 0.00028799999999999995,
+ "loss": 8.042091369628906,
+ "step": 49
+ },
+ {
+ "epoch": 0.6986899563318777,
+ "grad_norm": 0.06582305580377579,
+ "learning_rate": 0.000294,
+ "loss": 7.972844123840332,
+ "step": 50
+ },
+ {
+ "epoch": 0.7126637554585152,
+ "grad_norm": 0.059452351182699203,
+ "learning_rate": 0.0003,
+ "loss": 7.924579620361328,
+ "step": 51
+ },
+ {
+ "epoch": 0.7266375545851529,
+ "grad_norm": 0.052426718175411224,
+ "learning_rate": 0.00030599999999999996,
+ "loss": 7.876336097717285,
+ "step": 52
+ },
+ {
+ "epoch": 0.7406113537117904,
+ "grad_norm": 0.04768150672316551,
+ "learning_rate": 0.000312,
+ "loss": 7.8102707862854,
+ "step": 53
+ },
+ {
+ "epoch": 0.754585152838428,
+ "grad_norm": 0.03970075398683548,
+ "learning_rate": 0.000318,
+ "loss": 7.791913986206055,
+ "step": 54
+ },
+ {
+ "epoch": 0.7685589519650655,
+ "grad_norm": 0.03559752553701401,
+ "learning_rate": 0.000324,
+ "loss": 7.756960868835449,
+ "step": 55
+ },
+ {
+ "epoch": 0.7825327510917031,
+ "grad_norm": 0.029321353882551193,
+ "learning_rate": 0.00033,
+ "loss": 7.724878787994385,
+ "step": 56
+ },
+ {
+ "epoch": 0.7965065502183406,
+ "grad_norm": 0.023485427722334862,
+ "learning_rate": 0.000336,
+ "loss": 7.72553014755249,
+ "step": 57
+ },
+ {
+ "epoch": 0.8104803493449781,
+ "grad_norm": 0.019319474697113037,
+ "learning_rate": 0.00034199999999999996,
+ "loss": 7.709493160247803,
+ "step": 58
+ },
+ {
+ "epoch": 0.8244541484716157,
+ "grad_norm": 0.016705691814422607,
+ "learning_rate": 0.00034799999999999995,
+ "loss": 7.6873297691345215,
+ "step": 59
+ },
+ {
+ "epoch": 0.8384279475982532,
+ "grad_norm": 0.017667734995484352,
+ "learning_rate": 0.00035399999999999993,
+ "loss": 7.677059173583984,
+ "step": 60
+ },
+ {
+ "epoch": 0.8524017467248908,
+ "grad_norm": 0.01671992614865303,
+ "learning_rate": 0.00035999999999999997,
+ "loss": 7.643512725830078,
+ "step": 61
+ },
+ {
+ "epoch": 0.8663755458515284,
+ "grad_norm": 0.012288172729313374,
+ "learning_rate": 0.00036599999999999995,
+ "loss": 7.658343315124512,
+ "step": 62
+ },
+ {
+ "epoch": 0.880349344978166,
+ "grad_norm": 0.012004776857793331,
+ "learning_rate": 0.000372,
+ "loss": 7.649759292602539,
+ "step": 63
+ },
+ {
+ "epoch": 0.8943231441048035,
+ "grad_norm": 0.012487445957958698,
+ "learning_rate": 0.00037799999999999997,
+ "loss": 7.631503105163574,
+ "step": 64
+ },
+ {
+ "epoch": 0.9082969432314411,
+ "grad_norm": 0.01489241048693657,
+ "learning_rate": 0.00038399999999999996,
+ "loss": 7.647110462188721,
+ "step": 65
+ },
+ {
+ "epoch": 0.9222707423580786,
+ "grad_norm": 0.01428980939090252,
+ "learning_rate": 0.00039,
+ "loss": 7.636983871459961,
+ "step": 66
+ },
+ {
+ "epoch": 0.9362445414847161,
+ "grad_norm": 0.010711144655942917,
+ "learning_rate": 0.000396,
+ "loss": 7.621130466461182,
+ "step": 67
+ },
+ {
+ "epoch": 0.9502183406113537,
+ "grad_norm": 0.010178886353969574,
+ "learning_rate": 0.000402,
+ "loss": 7.593924522399902,
+ "step": 68
+ },
+ {
+ "epoch": 0.9641921397379912,
+ "grad_norm": 0.010604874230921268,
+ "learning_rate": 0.000408,
+ "loss": 7.6015496253967285,
+ "step": 69
+ },
+ {
+ "epoch": 0.9781659388646288,
+ "grad_norm": 0.009751426056027412,
+ "learning_rate": 0.0004139999999999999,
+ "loss": 7.571964263916016,
+ "step": 70
+ },
+ {
+ "epoch": 0.9921397379912664,
+ "grad_norm": 0.009816481731832027,
+ "learning_rate": 0.00041999999999999996,
+ "loss": 7.57344913482666,
+ "step": 71
+ },
+ {
+ "epoch": 1.0,
+ "grad_norm": 0.008665637113153934,
+ "learning_rate": 0.00042599999999999995,
+ "loss": 7.579050064086914,
+ "step": 72
+ },
+ {
+ "epoch": 1.0,
+ "eval_loss": 7.581204891204834,
+ "eval_runtime": 59.0091,
+ "eval_samples_per_second": 41.383,
+ "eval_steps_per_second": 1.305,
+ "step": 72
+ },
+ {
+ "epoch": 1.0139737991266375,
+ "grad_norm": 0.009684201329946518,
+ "learning_rate": 0.00043199999999999993,
+ "loss": 7.568594932556152,
+ "step": 73
+ },
+ {
+ "epoch": 1.027947598253275,
+ "grad_norm": 0.025783710181713104,
+ "learning_rate": 0.00043799999999999997,
+ "loss": 7.578948497772217,
+ "step": 74
+ },
+ {
+ "epoch": 1.0419213973799126,
+ "grad_norm": 0.0327819399535656,
+ "learning_rate": 0.00044399999999999995,
+ "loss": 7.548173904418945,
+ "step": 75
+ },
+ {
+ "epoch": 1.0558951965065502,
+ "grad_norm": 0.021604498848319054,
+ "learning_rate": 0.00045,
+ "loss": 7.552953720092773,
+ "step": 76
+ },
+ {
+ "epoch": 1.0698689956331877,
+ "grad_norm": 0.010230105370283127,
+ "learning_rate": 0.00045599999999999997,
+ "loss": 7.5213212966918945,
+ "step": 77
+ },
+ {
+ "epoch": 1.0838427947598253,
+ "grad_norm": 0.023315589874982834,
+ "learning_rate": 0.00046199999999999995,
+ "loss": 7.501640319824219,
+ "step": 78
+ },
+ {
+ "epoch": 1.0978165938864628,
+ "grad_norm": 0.01109944935888052,
+ "learning_rate": 0.000468,
+ "loss": 7.539519786834717,
+ "step": 79
+ },
+ {
+ "epoch": 1.1117903930131003,
+ "grad_norm": 0.026093758642673492,
+ "learning_rate": 0.000474,
+ "loss": 7.525928974151611,
+ "step": 80
+ },
+ {
+ "epoch": 1.125764192139738,
+ "grad_norm": 0.03274895250797272,
+ "learning_rate": 0.00047999999999999996,
+ "loss": 7.535602569580078,
+ "step": 81
+ },
+ {
+ "epoch": 1.1397379912663754,
+ "grad_norm": 0.01543602254241705,
+ "learning_rate": 0.000486,
+ "loss": 7.513683795928955,
+ "step": 82
+ },
+ {
+ "epoch": 1.1537117903930132,
+ "grad_norm": 0.028377249836921692,
+ "learning_rate": 0.0004919999999999999,
+ "loss": 7.49384880065918,
+ "step": 83
+ },
+ {
+ "epoch": 1.1676855895196507,
+ "grad_norm": 0.04121481627225876,
+ "learning_rate": 0.000498,
+ "loss": 7.488447189331055,
+ "step": 84
+ },
+ {
+ "epoch": 1.1816593886462883,
+ "grad_norm": 0.013282302767038345,
+ "learning_rate": 0.0005039999999999999,
+ "loss": 7.463010787963867,
+ "step": 85
+ },
+ {
+ "epoch": 1.1956331877729258,
+ "grad_norm": 0.03200364485383034,
+ "learning_rate": 0.0005099999999999999,
+ "loss": 7.473515033721924,
+ "step": 86
+ },
+ {
+ "epoch": 1.2096069868995634,
+ "grad_norm": 0.020768029615283012,
+ "learning_rate": 0.000516,
+ "loss": 7.464688777923584,
+ "step": 87
+ },
+ {
+ "epoch": 1.223580786026201,
+ "grad_norm": 0.021715788170695305,
+ "learning_rate": 0.000522,
+ "loss": 7.4835028648376465,
+ "step": 88
+ },
+ {
+ "epoch": 1.2375545851528384,
+ "grad_norm": 0.026255754753947258,
+ "learning_rate": 0.0005279999999999999,
+ "loss": 7.443392753601074,
+ "step": 89
+ },
+ {
+ "epoch": 1.251528384279476,
+ "grad_norm": 0.022325951606035233,
+ "learning_rate": 0.000534,
+ "loss": 7.457373142242432,
+ "step": 90
+ },
+ {
+ "epoch": 1.2655021834061135,
+ "grad_norm": 0.009270387701690197,
+ "learning_rate": 0.00054,
+ "loss": 7.455986976623535,
+ "step": 91
+ },
+ {
+ "epoch": 1.279475982532751,
+ "grad_norm": 0.02369544468820095,
+ "learning_rate": 0.0005459999999999999,
+ "loss": 7.442216873168945,
+ "step": 92
+ },
+ {
+ "epoch": 1.2934497816593886,
+ "grad_norm": 0.008474506437778473,
+ "learning_rate": 0.000552,
+ "loss": 7.4294915199279785,
+ "step": 93
+ },
+ {
+ "epoch": 1.3074235807860262,
+ "grad_norm": 0.017043249681591988,
+ "learning_rate": 0.000558,
+ "loss": 7.406932830810547,
+ "step": 94
+ },
+ {
+ "epoch": 1.3213973799126637,
+ "grad_norm": 0.011813412420451641,
+ "learning_rate": 0.0005639999999999999,
+ "loss": 7.415900230407715,
+ "step": 95
+ },
+ {
+ "epoch": 1.3353711790393012,
+ "grad_norm": 0.01086883433163166,
+ "learning_rate": 0.00057,
+ "loss": 7.452998638153076,
+ "step": 96
+ },
+ {
+ "epoch": 1.3493449781659388,
+ "grad_norm": 0.014700951054692268,
+ "learning_rate": 0.0005759999999999999,
+ "loss": 7.410249710083008,
+ "step": 97
+ },
+ {
+ "epoch": 1.3633187772925766,
+ "grad_norm": 0.014512976631522179,
+ "learning_rate": 0.0005819999999999999,
+ "loss": 7.429132461547852,
+ "step": 98
+ },
+ {
+ "epoch": 1.3772925764192139,
+ "grad_norm": 0.012519920244812965,
+ "learning_rate": 0.000588,
+ "loss": 7.430817604064941,
+ "step": 99
+ },
+ {
+ "epoch": 1.3912663755458516,
+ "grad_norm": 0.007132664322853088,
+ "learning_rate": 0.0005939999999999999,
+ "loss": 7.405234336853027,
+ "step": 100
+ },
+ {
+ "epoch": 1.405240174672489,
+ "grad_norm": 0.019489388912916183,
+ "learning_rate": 0.0006,
+ "loss": 7.427791595458984,
+ "step": 101
+ },
+ {
+ "epoch": 1.4192139737991267,
+ "grad_norm": 0.027252664789557457,
+ "learning_rate": 0.0006,
+ "loss": 7.41761589050293,
+ "step": 102
+ },
+ {
+ "epoch": 1.4331877729257643,
+ "grad_norm": 0.02501637488603592,
+ "learning_rate": 0.0006,
+ "loss": 7.416735649108887,
+ "step": 103
+ },
+ {
+ "epoch": 1.4471615720524018,
+ "grad_norm": 0.00989772193133831,
+ "learning_rate": 0.0006,
+ "loss": 7.3756632804870605,
+ "step": 104
+ },
+ {
+ "epoch": 1.4611353711790394,
+ "grad_norm": 0.01622949168086052,
+ "learning_rate": 0.0006,
+ "loss": 7.400811672210693,
+ "step": 105
+ },
+ {
+ "epoch": 1.475109170305677,
+ "grad_norm": 0.012407763861119747,
+ "learning_rate": 0.0006,
+ "loss": 7.37675666809082,
+ "step": 106
+ },
+ {
+ "epoch": 1.4890829694323144,
+ "grad_norm": 0.010353565216064453,
+ "learning_rate": 0.0006,
+ "loss": 7.4013142585754395,
+ "step": 107
+ },
+ {
+ "epoch": 1.503056768558952,
+ "grad_norm": 0.0120553532615304,
+ "learning_rate": 0.0006,
+ "loss": 7.379873275756836,
+ "step": 108
+ },
+ {
+ "epoch": 1.5170305676855895,
+ "grad_norm": 0.01610225811600685,
+ "learning_rate": 0.0006,
+ "loss": 7.371631145477295,
+ "step": 109
+ },
+ {
+ "epoch": 1.531004366812227,
+ "grad_norm": 0.01159879844635725,
+ "learning_rate": 0.0006,
+ "loss": 7.389007568359375,
+ "step": 110
+ },
+ {
+ "epoch": 1.5449781659388646,
+ "grad_norm": 0.008013816550374031,
+ "learning_rate": 0.0006,
+ "loss": 7.395920753479004,
+ "step": 111
+ },
+ {
+ "epoch": 1.5589519650655022,
+ "grad_norm": 0.016334809362888336,
+ "learning_rate": 0.0006,
+ "loss": 7.393121719360352,
+ "step": 112
+ },
+ {
+ "epoch": 1.5729257641921397,
+ "grad_norm": 0.018821721896529198,
+ "learning_rate": 0.0006,
+ "loss": 7.367543697357178,
+ "step": 113
+ },
+ {
+ "epoch": 1.5868995633187772,
+ "grad_norm": 0.014088994823396206,
+ "learning_rate": 0.0006,
+ "loss": 7.385537624359131,
+ "step": 114
+ },
+ {
+ "epoch": 1.600873362445415,
+ "grad_norm": 0.005864050704985857,
+ "learning_rate": 0.0006,
+ "loss": 7.365346908569336,
+ "step": 115
+ },
+ {
+ "epoch": 1.6148471615720523,
+ "grad_norm": 0.011745430529117584,
+ "learning_rate": 0.0006,
+ "loss": 7.390227317810059,
+ "step": 116
+ },
+ {
+ "epoch": 1.62882096069869,
+ "grad_norm": 0.017205573618412018,
+ "learning_rate": 0.0006,
+ "loss": 7.3558549880981445,
+ "step": 117
+ },
+ {
+ "epoch": 1.6427947598253274,
+ "grad_norm": 0.021256346255540848,
+ "learning_rate": 0.0006,
+ "loss": 7.377346992492676,
+ "step": 118
+ },
+ {
+ "epoch": 1.6567685589519652,
+ "grad_norm": 0.019335204735398293,
+ "learning_rate": 0.0006,
+ "loss": 7.346194267272949,
+ "step": 119
+ },
+ {
+ "epoch": 1.6707423580786025,
+ "grad_norm": 0.008134003728628159,
+ "learning_rate": 0.0006,
+ "loss": 7.3063578605651855,
+ "step": 120
+ },
+ {
+ "epoch": 1.6847161572052403,
+ "grad_norm": 0.017613831907510757,
+ "learning_rate": 0.0006,
+ "loss": 7.338171005249023,
+ "step": 121
+ },
+ {
+ "epoch": 1.6986899563318776,
+ "grad_norm": 0.00890734139829874,
+ "learning_rate": 0.0006,
+ "loss": 7.355384826660156,
+ "step": 122
+ },
+ {
+ "epoch": 1.7126637554585153,
+ "grad_norm": 0.013448155485093594,
+ "learning_rate": 0.0006,
+ "loss": 7.344296455383301,
+ "step": 123
+ },
+ {
+ "epoch": 1.726637554585153,
+ "grad_norm": 0.009241633117198944,
+ "learning_rate": 0.0006,
+ "loss": 7.3454060554504395,
+ "step": 124
+ },
+ {
+ "epoch": 1.7406113537117904,
+ "grad_norm": 0.018561311066150665,
+ "learning_rate": 0.0006,
+ "loss": 7.32348108291626,
+ "step": 125
+ },
+ {
+ "epoch": 1.754585152838428,
+ "grad_norm": 0.025667704641819,
+ "learning_rate": 0.0006,
+ "loss": 7.356142520904541,
+ "step": 126
+ },
+ {
+ "epoch": 1.7685589519650655,
+ "grad_norm": 0.036232151091098785,
+ "learning_rate": 0.0006,
+ "loss": 7.363539695739746,
+ "step": 127
+ },
+ {
+ "epoch": 1.782532751091703,
+ "grad_norm": 0.059483714401721954,
+ "learning_rate": 0.0006,
+ "loss": 7.402331352233887,
+ "step": 128
+ },
+ {
+ "epoch": 1.7965065502183406,
+ "grad_norm": 0.05365009233355522,
+ "learning_rate": 0.0006,
+ "loss": 7.40457820892334,
+ "step": 129
+ },
+ {
+ "epoch": 1.8104803493449781,
+ "grad_norm": 0.02738415263593197,
+ "learning_rate": 0.0006,
+ "loss": 7.3756208419799805,
+ "step": 130
+ },
+ {
+ "epoch": 1.8244541484716157,
+ "grad_norm": 0.035017091780900955,
+ "learning_rate": 0.0006,
+ "loss": 7.379263877868652,
+ "step": 131
+ },
+ {
+ "epoch": 1.8384279475982532,
+ "grad_norm": 0.02819863148033619,
+ "learning_rate": 0.0006,
+ "loss": 7.327673435211182,
+ "step": 132
+ },
+ {
+ "epoch": 1.8524017467248908,
+ "grad_norm": 0.017754238098859787,
+ "learning_rate": 0.0006,
+ "loss": 7.333420276641846,
+ "step": 133
+ },
+ {
+ "epoch": 1.8663755458515285,
+ "grad_norm": 0.024685733020305634,
+ "learning_rate": 0.0006,
+ "loss": 7.34493350982666,
+ "step": 134
+ },
+ {
+ "epoch": 1.8803493449781659,
+ "grad_norm": 0.019426532089710236,
+ "learning_rate": 0.0006,
+ "loss": 7.328448295593262,
+ "step": 135
+ },
+ {
+ "epoch": 1.8943231441048036,
+ "grad_norm": 0.02106391079723835,
+ "learning_rate": 0.0006,
+ "loss": 7.327917098999023,
+ "step": 136
+ },
+ {
+ "epoch": 1.908296943231441,
+ "grad_norm": 0.02056879736483097,
+ "learning_rate": 0.0006,
+ "loss": 7.3311991691589355,
+ "step": 137
+ },
+ {
+ "epoch": 1.9222707423580787,
+ "grad_norm": 0.016865141689777374,
+ "learning_rate": 0.0006,
+ "loss": 7.312766075134277,
+ "step": 138
+ },
+ {
+ "epoch": 1.936244541484716,
+ "grad_norm": 0.017549855634570122,
+ "learning_rate": 0.0006,
+ "loss": 7.31191349029541,
+ "step": 139
+ },
+ {
+ "epoch": 1.9502183406113538,
+ "grad_norm": 0.015930157154798508,
+ "learning_rate": 0.0006,
+ "loss": 7.330901622772217,
+ "step": 140
+ },
+ {
+ "epoch": 1.9641921397379911,
+ "grad_norm": 0.013283558189868927,
+ "learning_rate": 0.0006,
+ "loss": 7.298318386077881,
+ "step": 141
+ },
+ {
+ "epoch": 1.9781659388646289,
+ "grad_norm": 0.012023469433188438,
+ "learning_rate": 0.0006,
+ "loss": 7.282337188720703,
+ "step": 142
+ },
+ {
+ "epoch": 1.9921397379912664,
+ "grad_norm": 0.01459511648863554,
+ "learning_rate": 0.0006,
+ "loss": 7.282924175262451,
+ "step": 143
+ },
+ {
+ "epoch": 2.0,
+ "grad_norm": 0.017474235966801643,
+ "learning_rate": 0.0006,
+ "loss": 7.254333972930908,
+ "step": 144
+ },
+ {
+ "epoch": 2.0,
+ "eval_loss": 7.318708896636963,
+ "eval_runtime": 59.6555,
+ "eval_samples_per_second": 40.935,
+ "eval_steps_per_second": 1.291,
+ "step": 144
+ },
+ {
+ "epoch": 2.0139737991266378,
+ "grad_norm": 0.019757507368922234,
+ "learning_rate": 0.0006,
+ "loss": 7.30288553237915,
+ "step": 145
+ },
+ {
+ "epoch": 2.027947598253275,
+ "grad_norm": 0.014571248553693295,
+ "learning_rate": 0.0006,
+ "loss": 7.274681568145752,
+ "step": 146
+ },
+ {
+ "epoch": 2.041921397379913,
+ "grad_norm": 0.015744198113679886,
+ "learning_rate": 0.0006,
+ "loss": 7.278632640838623,
+ "step": 147
+ },
+ {
+ "epoch": 2.05589519650655,
+ "grad_norm": 0.028048858046531677,
+ "learning_rate": 0.0006,
+ "loss": 7.26558780670166,
+ "step": 148
+ },
+ {
+ "epoch": 2.069868995633188,
+ "grad_norm": 0.06547307223081589,
+ "learning_rate": 0.0006,
+ "loss": 7.32567834854126,
+ "step": 149
+ },
+ {
+ "epoch": 2.0838427947598253,
+ "grad_norm": 0.05969763547182083,
+ "learning_rate": 0.0006,
+ "loss": 7.395836353302002,
+ "step": 150
+ },
+ {
+ "epoch": 2.097816593886463,
+ "grad_norm": 0.032574281096458435,
+ "learning_rate": 0.0006,
+ "loss": 7.318150997161865,
+ "step": 151
+ },
+ {
+ "epoch": 2.1117903930131003,
+ "grad_norm": 0.029452962800860405,
+ "learning_rate": 0.0006,
+ "loss": 7.308066368103027,
+ "step": 152
+ },
+ {
+ "epoch": 2.125764192139738,
+ "grad_norm": 0.03161991387605667,
+ "learning_rate": 0.0006,
+ "loss": 7.306290626525879,
+ "step": 153
+ },
+ {
+ "epoch": 2.1397379912663754,
+ "grad_norm": 0.027096295729279518,
+ "learning_rate": 0.0006,
+ "loss": 7.297796249389648,
+ "step": 154
+ },
+ {
+ "epoch": 2.153711790393013,
+ "grad_norm": 0.028657056391239166,
+ "learning_rate": 0.0006,
+ "loss": 7.317424774169922,
+ "step": 155
+ },
+ {
+ "epoch": 2.1676855895196505,
+ "grad_norm": 0.02046995982527733,
+ "learning_rate": 0.0006,
+ "loss": 7.297986030578613,
+ "step": 156
+ },
+ {
+ "epoch": 2.1816593886462883,
+ "grad_norm": 0.02220369316637516,
+ "learning_rate": 0.0006,
+ "loss": 7.268237590789795,
+ "step": 157
+ },
+ {
+ "epoch": 2.1956331877729256,
+ "grad_norm": 0.018381357192993164,
+ "learning_rate": 0.0006,
+ "loss": 7.259415626525879,
+ "step": 158
+ },
+ {
+ "epoch": 2.2096069868995634,
+ "grad_norm": 0.0204413291066885,
+ "learning_rate": 0.0006,
+ "loss": 7.258173942565918,
+ "step": 159
+ },
+ {
+ "epoch": 2.2235807860262007,
+ "grad_norm": 0.015874288976192474,
+ "learning_rate": 0.0006,
+ "loss": 7.281834602355957,
+ "step": 160
+ },
+ {
+ "epoch": 2.2375545851528384,
+ "grad_norm": 0.01634068600833416,
+ "learning_rate": 0.0006,
+ "loss": 7.2469868659973145,
+ "step": 161
+ },
+ {
+ "epoch": 2.251528384279476,
+ "grad_norm": 0.01728747971355915,
+ "learning_rate": 0.0006,
+ "loss": 7.280995845794678,
+ "step": 162
+ },
+ {
+ "epoch": 2.2655021834061135,
+ "grad_norm": 0.016341188922524452,
+ "learning_rate": 0.0006,
+ "loss": 7.273357391357422,
+ "step": 163
+ },
+ {
+ "epoch": 2.279475982532751,
+ "grad_norm": 0.016125807538628578,
+ "learning_rate": 0.0006,
+ "loss": 7.273510932922363,
+ "step": 164
+ },
+ {
+ "epoch": 2.2934497816593886,
+ "grad_norm": 0.01629287749528885,
+ "learning_rate": 0.0006,
+ "loss": 7.267029762268066,
+ "step": 165
+ },
+ {
+ "epoch": 2.3074235807860264,
+ "grad_norm": 0.015296131372451782,
+ "learning_rate": 0.0006,
+ "loss": 7.211989402770996,
+ "step": 166
+ },
+ {
+ "epoch": 2.3213973799126637,
+ "grad_norm": 0.01709570363163948,
+ "learning_rate": 0.0006,
+ "loss": 7.242851734161377,
+ "step": 167
+ },
+ {
+ "epoch": 2.3353711790393015,
+ "grad_norm": 0.012552388943731785,
+ "learning_rate": 0.0006,
+ "loss": 7.241696834564209,
+ "step": 168
+ },
+ {
+ "epoch": 2.349344978165939,
+ "grad_norm": 0.014571291394531727,
+ "learning_rate": 0.0006,
+ "loss": 7.270681381225586,
+ "step": 169
+ },
+ {
+ "epoch": 2.3633187772925766,
+ "grad_norm": 0.013389437459409237,
+ "learning_rate": 0.0006,
+ "loss": 7.191600799560547,
+ "step": 170
+ },
+ {
+ "epoch": 2.377292576419214,
+ "grad_norm": 0.011895314790308475,
+ "learning_rate": 0.0006,
+ "loss": 7.213610649108887,
+ "step": 171
+ },
+ {
+ "epoch": 2.3912663755458516,
+ "grad_norm": 0.01345642190426588,
+ "learning_rate": 0.0006,
+ "loss": 7.193787574768066,
+ "step": 172
+ },
+ {
+ "epoch": 2.405240174672489,
+ "grad_norm": 0.0091794328764081,
+ "learning_rate": 0.0006,
+ "loss": 7.233364105224609,
+ "step": 173
+ },
+ {
+ "epoch": 2.4192139737991267,
+ "grad_norm": 0.013767086900770664,
+ "learning_rate": 0.0006,
+ "loss": 7.217272758483887,
+ "step": 174
+ },
+ {
+ "epoch": 2.433187772925764,
+ "grad_norm": 0.015134960412979126,
+ "learning_rate": 0.0006,
+ "loss": 7.192791938781738,
+ "step": 175
+ },
+ {
+ "epoch": 2.447161572052402,
+ "grad_norm": 0.01790648326277733,
+ "learning_rate": 0.0006,
+ "loss": 7.229467391967773,
+ "step": 176
+ },
+ {
+ "epoch": 2.461135371179039,
+ "grad_norm": 0.015362377278506756,
+ "learning_rate": 0.0006,
+ "loss": 7.180346488952637,
+ "step": 177
+ },
+ {
+ "epoch": 2.475109170305677,
+ "grad_norm": 0.010505498386919498,
+ "learning_rate": 0.0006,
+ "loss": 7.188833236694336,
+ "step": 178
+ },
+ {
+ "epoch": 2.489082969432314,
+ "grad_norm": 0.009129747748374939,
+ "learning_rate": 0.0006,
+ "loss": 7.203153133392334,
+ "step": 179
+ },
+ {
+ "epoch": 2.503056768558952,
+ "grad_norm": 0.01147883478552103,
+ "learning_rate": 0.0006,
+ "loss": 7.163824558258057,
+ "step": 180
+ },
+ {
+ "epoch": 2.5170305676855893,
+ "grad_norm": 0.011338942684233189,
+ "learning_rate": 0.0006,
+ "loss": 7.16287899017334,
+ "step": 181
+ },
+ {
+ "epoch": 2.531004366812227,
+ "grad_norm": 0.013277675025165081,
+ "learning_rate": 0.0006,
+ "loss": 7.155905723571777,
+ "step": 182
+ },
+ {
+ "epoch": 2.544978165938865,
+ "grad_norm": 0.015269347466528416,
+ "learning_rate": 0.0006,
+ "loss": 7.189371109008789,
+ "step": 183
+ },
+ {
+ "epoch": 2.558951965065502,
+ "grad_norm": 0.017905596643686295,
+ "learning_rate": 0.0006,
+ "loss": 7.177547931671143,
+ "step": 184
+ },
+ {
+ "epoch": 2.5729257641921395,
+ "grad_norm": 0.024431610479950905,
+ "learning_rate": 0.0006,
+ "loss": 7.192169189453125,
+ "step": 185
+ },
+ {
+ "epoch": 2.5868995633187772,
+ "grad_norm": 0.03656961768865585,
+ "learning_rate": 0.0006,
+ "loss": 7.124387741088867,
+ "step": 186
+ },
+ {
+ "epoch": 2.600873362445415,
+ "grad_norm": 0.04337048903107643,
+ "learning_rate": 0.0006,
+ "loss": 7.140231609344482,
+ "step": 187
+ },
+ {
+ "epoch": 2.6148471615720523,
+ "grad_norm": 0.04047228768467903,
+ "learning_rate": 0.0006,
+ "loss": 7.171927452087402,
+ "step": 188
+ },
+ {
+ "epoch": 2.62882096069869,
+ "grad_norm": 0.07825770974159241,
+ "learning_rate": 0.0006,
+ "loss": 7.207631587982178,
+ "step": 189
+ },
+ {
+ "epoch": 2.6427947598253274,
+ "grad_norm": 0.07118063420057297,
+ "learning_rate": 0.0006,
+ "loss": 7.209488391876221,
+ "step": 190
+ },
+ {
+ "epoch": 2.656768558951965,
+ "grad_norm": 0.05153293535113335,
+ "learning_rate": 0.0006,
+ "loss": 7.14793062210083,
+ "step": 191
+ },
+ {
+ "epoch": 2.6707423580786025,
+ "grad_norm": 0.03318839520215988,
+ "learning_rate": 0.0006,
+ "loss": 7.163010597229004,
+ "step": 192
+ },
+ {
+ "epoch": 2.6847161572052403,
+ "grad_norm": 0.029333539307117462,
+ "learning_rate": 0.0006,
+ "loss": 7.159955978393555,
+ "step": 193
+ },
+ {
+ "epoch": 2.6986899563318776,
+ "grad_norm": 0.029405880719423294,
+ "learning_rate": 0.0006,
+ "loss": 7.136178493499756,
+ "step": 194
+ },
+ {
+ "epoch": 2.7126637554585153,
+ "grad_norm": 0.037815988063812256,
+ "learning_rate": 0.0006,
+ "loss": 7.183186054229736,
+ "step": 195
+ },
+ {
+ "epoch": 2.726637554585153,
+ "grad_norm": 0.020307250320911407,
+ "learning_rate": 0.0006,
+ "loss": 7.133164882659912,
+ "step": 196
+ },
+ {
+ "epoch": 2.7406113537117904,
+ "grad_norm": 0.026048069819808006,
+ "learning_rate": 0.0006,
+ "loss": 7.124977111816406,
+ "step": 197
+ },
+ {
+ "epoch": 2.7545851528384278,
+ "grad_norm": 0.027316724881529808,
+ "learning_rate": 0.0006,
+ "loss": 7.07265567779541,
+ "step": 198
+ },
+ {
+ "epoch": 2.7685589519650655,
+ "grad_norm": 0.021629009395837784,
+ "learning_rate": 0.0006,
+ "loss": 7.150020599365234,
+ "step": 199
+ },
+ {
+ "epoch": 2.7825327510917033,
+ "grad_norm": 0.01605929434299469,
+ "learning_rate": 0.0006,
+ "loss": 7.10919189453125,
+ "step": 200
+ },
+ {
+ "epoch": 2.7965065502183406,
+ "grad_norm": 0.020800089463591576,
+ "learning_rate": 0.0006,
+ "loss": 7.132045269012451,
+ "step": 201
+ },
+ {
+ "epoch": 2.810480349344978,
+ "grad_norm": 0.019225774332880974,
+ "learning_rate": 0.0006,
+ "loss": 7.128364562988281,
+ "step": 202
+ },
+ {
+ "epoch": 2.8244541484716157,
+ "grad_norm": 0.01561372634023428,
+ "learning_rate": 0.0006,
+ "loss": 7.083669662475586,
+ "step": 203
+ },
+ {
+ "epoch": 2.8384279475982535,
+ "grad_norm": 0.021969519555568695,
+ "learning_rate": 0.0006,
+ "loss": 7.050841331481934,
+ "step": 204
+ },
+ {
+ "epoch": 2.8524017467248908,
+ "grad_norm": 0.021453695371747017,
+ "learning_rate": 0.0006,
+ "loss": 7.097085475921631,
+ "step": 205
+ },
+ {
+ "epoch": 2.8663755458515285,
+ "grad_norm": 0.022836022078990936,
+ "learning_rate": 0.0006,
+ "loss": 7.10405969619751,
+ "step": 206
+ },
+ {
+ "epoch": 2.880349344978166,
+ "grad_norm": 0.02430851384997368,
+ "learning_rate": 0.0006,
+ "loss": 7.045181751251221,
+ "step": 207
+ },
+ {
+ "epoch": 2.8943231441048036,
+ "grad_norm": 0.020166993141174316,
+ "learning_rate": 0.0006,
+ "loss": 7.068183898925781,
+ "step": 208
+ },
+ {
+ "epoch": 2.908296943231441,
+ "grad_norm": 0.0118505684658885,
+ "learning_rate": 0.0006,
+ "loss": 7.07132625579834,
+ "step": 209
+ },
+ {
+ "epoch": 2.9222707423580787,
+ "grad_norm": 0.023772427812218666,
+ "learning_rate": 0.0006,
+ "loss": 7.055238723754883,
+ "step": 210
+ },
+ {
+ "epoch": 2.936244541484716,
+ "grad_norm": 0.02350712940096855,
+ "learning_rate": 0.0006,
+ "loss": 7.055898666381836,
+ "step": 211
+ },
+ {
+ "epoch": 2.950218340611354,
+ "grad_norm": 0.017244907096028328,
+ "learning_rate": 0.0006,
+ "loss": 7.0099711418151855,
+ "step": 212
+ },
+ {
+ "epoch": 2.964192139737991,
+ "grad_norm": 0.021565575152635574,
+ "learning_rate": 0.0006,
+ "loss": 6.991751670837402,
+ "step": 213
+ },
+ {
+ "epoch": 2.978165938864629,
+ "grad_norm": 0.03241860866546631,
+ "learning_rate": 0.0006,
+ "loss": 7.005980491638184,
+ "step": 214
+ },
+ {
+ "epoch": 2.992139737991266,
+ "grad_norm": 0.04613726586103439,
+ "learning_rate": 0.0006,
+ "loss": 7.027859687805176,
+ "step": 215
+ },
+ {
+ "epoch": 3.0,
+ "grad_norm": 0.06404894590377808,
+ "learning_rate": 0.0006,
+ "loss": 7.123725891113281,
+ "step": 216
+ },
+ {
+ "epoch": 3.0,
+ "eval_loss": 7.070647716522217,
+ "eval_runtime": 58.5801,
+ "eval_samples_per_second": 41.687,
+ "eval_steps_per_second": 1.314,
+ "step": 216
+ },
+ {
+ "epoch": 3.0139737991266378,
+ "grad_norm": 0.05542432889342308,
+ "learning_rate": 0.0006,
+ "loss": 7.078299522399902,
+ "step": 217
+ },
+ {
+ "epoch": 3.027947598253275,
+ "grad_norm": 0.026484526693820953,
+ "learning_rate": 0.0006,
+ "loss": 6.961159706115723,
+ "step": 218
+ },
+ {
+ "epoch": 3.041921397379913,
+ "grad_norm": 0.028276391327381134,
+ "learning_rate": 0.0006,
+ "loss": 7.019961833953857,
+ "step": 219
+ },
+ {
+ "epoch": 3.05589519650655,
+ "grad_norm": 0.024486595764756203,
+ "learning_rate": 0.0006,
+ "loss": 7.039677619934082,
+ "step": 220
+ },
+ {
+ "epoch": 3.069868995633188,
+ "grad_norm": 0.017620893195271492,
+ "learning_rate": 0.0006,
+ "loss": 6.9270782470703125,
+ "step": 221
+ },
+ {
+ "epoch": 3.0838427947598253,
+ "grad_norm": 0.02547302283346653,
+ "learning_rate": 0.0006,
+ "loss": 6.924233436584473,
+ "step": 222
+ },
+ {
+ "epoch": 3.097816593886463,
+ "grad_norm": 0.02539009042084217,
+ "learning_rate": 0.0006,
+ "loss": 6.973979949951172,
+ "step": 223
+ },
+ {
+ "epoch": 3.1117903930131003,
+ "grad_norm": 0.03256227448582649,
+ "learning_rate": 0.0006,
+ "loss": 6.933725357055664,
+ "step": 224
+ },
+ {
+ "epoch": 3.125764192139738,
+ "grad_norm": 0.04651800915598869,
+ "learning_rate": 0.0006,
+ "loss": 6.951044082641602,
+ "step": 225
+ },
+ {
+ "epoch": 3.1397379912663754,
+ "grad_norm": 0.04822975769639015,
+ "learning_rate": 0.0006,
+ "loss": 6.9614410400390625,
+ "step": 226
+ },
+ {
+ "epoch": 3.153711790393013,
+ "grad_norm": 0.031165020540356636,
+ "learning_rate": 0.0006,
+ "loss": 6.91193151473999,
+ "step": 227
+ },
+ {
+ "epoch": 3.1676855895196505,
+ "grad_norm": 0.029754292219877243,
+ "learning_rate": 0.0006,
+ "loss": 6.91802978515625,
+ "step": 228
+ },
+ {
+ "epoch": 3.1816593886462883,
+ "grad_norm": 0.02193152718245983,
+ "learning_rate": 0.0006,
+ "loss": 6.928090572357178,
+ "step": 229
+ },
+ {
+ "epoch": 3.1956331877729256,
+ "grad_norm": 0.02428470179438591,
+ "learning_rate": 0.0006,
+ "loss": 6.90396785736084,
+ "step": 230
+ },
+ {
+ "epoch": 3.2096069868995634,
+ "grad_norm": 0.023374345153570175,
+ "learning_rate": 0.0006,
+ "loss": 6.885858535766602,
+ "step": 231
+ },
+ {
+ "epoch": 3.2235807860262007,
+ "grad_norm": 0.023405000567436218,
+ "learning_rate": 0.0006,
+ "loss": 6.8855767250061035,
+ "step": 232
+ },
+ {
+ "epoch": 3.2375545851528384,
+ "grad_norm": 0.017277134582400322,
+ "learning_rate": 0.0006,
+ "loss": 6.901449203491211,
+ "step": 233
+ },
+ {
+ "epoch": 3.251528384279476,
+ "grad_norm": 0.019496750086545944,
+ "learning_rate": 0.0006,
+ "loss": 6.885687828063965,
+ "step": 234
+ },
+ {
+ "epoch": 3.2655021834061135,
+ "grad_norm": 0.016893276944756508,
+ "learning_rate": 0.0006,
+ "loss": 6.872369766235352,
+ "step": 235
+ },
+ {
+ "epoch": 3.279475982532751,
+ "grad_norm": 0.02058715932071209,
+ "learning_rate": 0.0006,
+ "loss": 6.823280334472656,
+ "step": 236
+ },
+ {
+ "epoch": 3.2934497816593886,
+ "grad_norm": 0.015530755743384361,
+ "learning_rate": 0.0006,
+ "loss": 6.860499858856201,
+ "step": 237
+ },
+ {
+ "epoch": 3.3074235807860264,
+ "grad_norm": 0.01934926211833954,
+ "learning_rate": 0.0006,
+ "loss": 6.817448616027832,
+ "step": 238
+ },
+ {
+ "epoch": 3.3213973799126637,
+ "grad_norm": 0.017413552850484848,
+ "learning_rate": 0.0006,
+ "loss": 6.881955146789551,
+ "step": 239
+ },
+ {
+ "epoch": 3.3353711790393015,
+ "grad_norm": 0.026478875428438187,
+ "learning_rate": 0.0006,
+ "loss": 6.827404499053955,
+ "step": 240
+ },
+ {
+ "epoch": 3.349344978165939,
+ "grad_norm": 0.032611701637506485,
+ "learning_rate": 0.0006,
+ "loss": 6.78285551071167,
+ "step": 241
+ },
+ {
+ "epoch": 3.3633187772925766,
+ "grad_norm": 0.041900306940078735,
+ "learning_rate": 0.0006,
+ "loss": 6.783053874969482,
+ "step": 242
+ },
+ {
+ "epoch": 3.377292576419214,
+ "grad_norm": 0.04149497672915459,
+ "learning_rate": 0.0006,
+ "loss": 6.745220184326172,
+ "step": 243
+ },
+ {
+ "epoch": 3.3912663755458516,
+ "grad_norm": 0.023949826136231422,
+ "learning_rate": 0.0006,
+ "loss": 6.841533184051514,
+ "step": 244
+ },
+ {
+ "epoch": 3.405240174672489,
+ "grad_norm": 0.028601842001080513,
+ "learning_rate": 0.0006,
+ "loss": 6.852852821350098,
+ "step": 245
+ },
+ {
+ "epoch": 3.4192139737991267,
+ "grad_norm": 0.027365155518054962,
+ "learning_rate": 0.0006,
+ "loss": 6.74376106262207,
+ "step": 246
+ },
+ {
+ "epoch": 3.433187772925764,
+ "grad_norm": 0.02856568619608879,
+ "learning_rate": 0.0006,
+ "loss": 6.74906063079834,
+ "step": 247
+ },
+ {
+ "epoch": 3.447161572052402,
+ "grad_norm": 0.019172416999936104,
+ "learning_rate": 0.0006,
+ "loss": 6.809206008911133,
+ "step": 248
+ },
+ {
+ "epoch": 3.461135371179039,
+ "grad_norm": 0.02161746844649315,
+ "learning_rate": 0.0006,
+ "loss": 6.789888381958008,
+ "step": 249
+ },
+ {
+ "epoch": 3.475109170305677,
+ "grad_norm": 0.020273666828870773,
+ "learning_rate": 0.0006,
+ "loss": 6.73512077331543,
+ "step": 250
+ },
+ {
+ "epoch": 3.489082969432314,
+ "grad_norm": 0.022329283878207207,
+ "learning_rate": 0.0006,
+ "loss": 6.826148509979248,
+ "step": 251
+ },
+ {
+ "epoch": 3.503056768558952,
+ "grad_norm": 0.02284284122288227,
+ "learning_rate": 0.0006,
+ "loss": 6.748521327972412,
+ "step": 252
+ },
+ {
+ "epoch": 3.5170305676855893,
+ "grad_norm": 0.026368247345089912,
+ "learning_rate": 0.0006,
+ "loss": 6.645879745483398,
+ "step": 253
+ },
+ {
+ "epoch": 3.531004366812227,
+ "grad_norm": 0.03627244383096695,
+ "learning_rate": 0.0006,
+ "loss": 6.666064739227295,
+ "step": 254
+ },
+ {
+ "epoch": 3.544978165938865,
+ "grad_norm": 0.04140935093164444,
+ "learning_rate": 0.0006,
+ "loss": 6.712275981903076,
+ "step": 255
+ },
+ {
+ "epoch": 3.558951965065502,
+ "grad_norm": 0.0390462800860405,
+ "learning_rate": 0.0006,
+ "loss": 6.665900230407715,
+ "step": 256
+ },
+ {
+ "epoch": 3.5729257641921395,
+ "grad_norm": 0.05178583785891533,
+ "learning_rate": 0.0006,
+ "loss": 6.753298759460449,
+ "step": 257
+ },
+ {
+ "epoch": 3.5868995633187772,
+ "grad_norm": 0.054328061640262604,
+ "learning_rate": 0.0006,
+ "loss": 6.732633113861084,
+ "step": 258
+ },
+ {
+ "epoch": 3.600873362445415,
+ "grad_norm": 0.04590460658073425,
+ "learning_rate": 0.0006,
+ "loss": 6.65250301361084,
+ "step": 259
+ },
+ {
+ "epoch": 3.6148471615720523,
+ "grad_norm": 0.044331666082143784,
+ "learning_rate": 0.0006,
+ "loss": 6.68654203414917,
+ "step": 260
+ },
+ {
+ "epoch": 3.62882096069869,
+ "grad_norm": 0.05308730527758598,
+ "learning_rate": 0.0006,
+ "loss": 6.642482757568359,
+ "step": 261
+ },
+ {
+ "epoch": 3.6427947598253274,
+ "grad_norm": 0.03999016433954239,
+ "learning_rate": 0.0006,
+ "loss": 6.625227451324463,
+ "step": 262
+ },
+ {
+ "epoch": 3.656768558951965,
+ "grad_norm": 0.04088086634874344,
+ "learning_rate": 0.0006,
+ "loss": 6.67623233795166,
+ "step": 263
+ },
+ {
+ "epoch": 3.6707423580786025,
+ "grad_norm": 0.03615636005997658,
+ "learning_rate": 0.0006,
+ "loss": 6.6869001388549805,
+ "step": 264
+ },
+ {
+ "epoch": 3.6847161572052403,
+ "grad_norm": 0.02817360684275627,
+ "learning_rate": 0.0006,
+ "loss": 6.674601078033447,
+ "step": 265
+ },
+ {
+ "epoch": 3.6986899563318776,
+ "grad_norm": 0.032586123794317245,
+ "learning_rate": 0.0006,
+ "loss": 6.612994194030762,
+ "step": 266
+ },
+ {
+ "epoch": 3.7126637554585153,
+ "grad_norm": 0.031117867678403854,
+ "learning_rate": 0.0006,
+ "loss": 6.6244354248046875,
+ "step": 267
+ },
+ {
+ "epoch": 3.726637554585153,
+ "grad_norm": 0.028330687433481216,
+ "learning_rate": 0.0006,
+ "loss": 6.669557094573975,
+ "step": 268
+ },
+ {
+ "epoch": 3.7406113537117904,
+ "grad_norm": 0.021489202976226807,
+ "learning_rate": 0.0006,
+ "loss": 6.632035732269287,
+ "step": 269
+ },
+ {
+ "epoch": 3.7545851528384278,
+ "grad_norm": 0.024269424378871918,
+ "learning_rate": 0.0006,
+ "loss": 6.660867691040039,
+ "step": 270
+ },
+ {
+ "epoch": 3.7685589519650655,
+ "grad_norm": 0.02237873338162899,
+ "learning_rate": 0.0006,
+ "loss": 6.602587699890137,
+ "step": 271
+ },
+ {
+ "epoch": 3.7825327510917033,
+ "grad_norm": 0.026503929868340492,
+ "learning_rate": 0.0006,
+ "loss": 6.599200248718262,
+ "step": 272
+ },
+ {
+ "epoch": 3.7965065502183406,
+ "grad_norm": 0.03687124326825142,
+ "learning_rate": 0.0006,
+ "loss": 6.567296504974365,
+ "step": 273
+ },
+ {
+ "epoch": 3.810480349344978,
+ "grad_norm": 0.04708952456712723,
+ "learning_rate": 0.0006,
+ "loss": 6.621265411376953,
+ "step": 274
+ },
+ {
+ "epoch": 3.8244541484716157,
+ "grad_norm": 0.041846614331007004,
+ "learning_rate": 0.0006,
+ "loss": 6.550536155700684,
+ "step": 275
+ },
+ {
+ "epoch": 3.8384279475982535,
+ "grad_norm": 0.031249675899744034,
+ "learning_rate": 0.0006,
+ "loss": 6.54278564453125,
+ "step": 276
+ },
+ {
+ "epoch": 3.8524017467248908,
+ "grad_norm": 0.023082256317138672,
+ "learning_rate": 0.0006,
+ "loss": 6.544787406921387,
+ "step": 277
+ },
+ {
+ "epoch": 3.8663755458515285,
+ "grad_norm": 0.020995570346713066,
+ "learning_rate": 0.0006,
+ "loss": 6.518143653869629,
+ "step": 278
+ },
+ {
+ "epoch": 3.880349344978166,
+ "grad_norm": 0.020839324221014977,
+ "learning_rate": 0.0006,
+ "loss": 6.546667098999023,
+ "step": 279
+ },
+ {
+ "epoch": 3.8943231441048036,
+ "grad_norm": 0.018377432599663734,
+ "learning_rate": 0.0006,
+ "loss": 6.478307723999023,
+ "step": 280
+ },
+ {
+ "epoch": 3.908296943231441,
+ "grad_norm": 0.021713724359869957,
+ "learning_rate": 0.0006,
+ "loss": 6.485574245452881,
+ "step": 281
+ },
+ {
+ "epoch": 3.9222707423580787,
+ "grad_norm": 0.01359301246702671,
+ "learning_rate": 0.0006,
+ "loss": 6.570268630981445,
+ "step": 282
+ },
+ {
+ "epoch": 3.936244541484716,
+ "grad_norm": 0.016233332455158234,
+ "learning_rate": 0.0006,
+ "loss": 6.57150936126709,
+ "step": 283
+ },
+ {
+ "epoch": 3.950218340611354,
+ "grad_norm": 0.020687608048319817,
+ "learning_rate": 0.0006,
+ "loss": 6.527956008911133,
+ "step": 284
+ },
+ {
+ "epoch": 3.964192139737991,
+ "grad_norm": 0.02032964862883091,
+ "learning_rate": 0.0006,
+ "loss": 6.468169689178467,
+ "step": 285
+ },
+ {
+ "epoch": 3.978165938864629,
+ "grad_norm": 0.016676288098096848,
+ "learning_rate": 0.0006,
+ "loss": 6.448418617248535,
+ "step": 286
+ },
+ {
+ "epoch": 3.992139737991266,
+ "grad_norm": 0.0168539360165596,
+ "learning_rate": 0.0006,
+ "loss": 6.461180686950684,
+ "step": 287
+ },
+ {
+ "epoch": 4.0,
+ "grad_norm": 0.02666233666241169,
+ "learning_rate": 0.0006,
+ "loss": 6.344893932342529,
+ "step": 288
+ },
+ {
+ "epoch": 4.0,
+ "eval_loss": 6.5384979248046875,
+ "eval_runtime": 58.814,
+ "eval_samples_per_second": 41.521,
+ "eval_steps_per_second": 1.309,
+ "step": 288
+ },
+ {
+ "epoch": 4.013973799126638,
+ "grad_norm": 0.04154708981513977,
+ "learning_rate": 0.0006,
+ "loss": 6.41855525970459,
+ "step": 289
+ },
+ {
+ "epoch": 4.0279475982532755,
+ "grad_norm": 0.04895668104290962,
+ "learning_rate": 0.0006,
+ "loss": 6.499302864074707,
+ "step": 290
+ },
+ {
+ "epoch": 4.041921397379912,
+ "grad_norm": 0.034464482218027115,
+ "learning_rate": 0.0006,
+ "loss": 6.438611030578613,
+ "step": 291
+ },
+ {
+ "epoch": 4.05589519650655,
+ "grad_norm": 0.06845773756504059,
+ "learning_rate": 0.0006,
+ "loss": 6.538361072540283,
+ "step": 292
+ },
+ {
+ "epoch": 4.069868995633188,
+ "grad_norm": 0.053768858313560486,
+ "learning_rate": 0.0006,
+ "loss": 6.527524948120117,
+ "step": 293
+ },
+ {
+ "epoch": 4.083842794759826,
+ "grad_norm": 0.039037931710481644,
+ "learning_rate": 0.0006,
+ "loss": 6.503847122192383,
+ "step": 294
+ },
+ {
+ "epoch": 4.097816593886463,
+ "grad_norm": 0.03962196782231331,
+ "learning_rate": 0.0006,
+ "loss": 6.475178241729736,
+ "step": 295
+ },
+ {
+ "epoch": 4.1117903930131,
+ "grad_norm": 0.049101535230875015,
+ "learning_rate": 0.0006,
+ "loss": 6.474587917327881,
+ "step": 296
+ },
+ {
+ "epoch": 4.125764192139738,
+ "grad_norm": 0.02830282226204872,
+ "learning_rate": 0.0006,
+ "loss": 6.45065975189209,
+ "step": 297
+ },
+ {
+ "epoch": 4.139737991266376,
+ "grad_norm": 0.023153482005000114,
+ "learning_rate": 0.0006,
+ "loss": 6.5218706130981445,
+ "step": 298
+ },
+ {
+ "epoch": 4.153711790393013,
+ "grad_norm": 0.025064002722501755,
+ "learning_rate": 0.0006,
+ "loss": 6.456475257873535,
+ "step": 299
+ },
+ {
+ "epoch": 4.1676855895196505,
+ "grad_norm": 0.02357092685997486,
+ "learning_rate": 0.0006,
+ "loss": 6.485952377319336,
+ "step": 300
+ },
+ {
+ "epoch": 4.181659388646288,
+ "grad_norm": 0.021832305938005447,
+ "learning_rate": 0.0006,
+ "loss": 6.425506591796875,
+ "step": 301
+ },
+ {
+ "epoch": 4.195633187772926,
+ "grad_norm": 0.0227900929749012,
+ "learning_rate": 0.0006,
+ "loss": 6.439155578613281,
+ "step": 302
+ },
+ {
+ "epoch": 4.209606986899563,
+ "grad_norm": 0.015961607918143272,
+ "learning_rate": 0.0006,
+ "loss": 6.460862636566162,
+ "step": 303
+ },
+ {
+ "epoch": 4.223580786026201,
+ "grad_norm": 0.017876390367746353,
+ "learning_rate": 0.0006,
+ "loss": 6.44564151763916,
+ "step": 304
+ },
+ {
+ "epoch": 4.2375545851528384,
+ "grad_norm": 0.013703204691410065,
+ "learning_rate": 0.0006,
+ "loss": 6.418414115905762,
+ "step": 305
+ },
+ {
+ "epoch": 4.251528384279476,
+ "grad_norm": 0.015240306034684181,
+ "learning_rate": 0.0006,
+ "loss": 6.411238670349121,
+ "step": 306
+ },
+ {
+ "epoch": 4.265502183406113,
+ "grad_norm": 0.01587662845849991,
+ "learning_rate": 0.0006,
+ "loss": 6.395642280578613,
+ "step": 307
+ },
+ {
+ "epoch": 4.279475982532751,
+ "grad_norm": 0.014650700613856316,
+ "learning_rate": 0.0006,
+ "loss": 6.302793025970459,
+ "step": 308
+ },
+ {
+ "epoch": 4.293449781659389,
+ "grad_norm": 0.016533225774765015,
+ "learning_rate": 0.0006,
+ "loss": 6.380588054656982,
+ "step": 309
+ },
+ {
+ "epoch": 4.307423580786026,
+ "grad_norm": 0.019963741302490234,
+ "learning_rate": 0.0006,
+ "loss": 6.345336437225342,
+ "step": 310
+ },
+ {
+ "epoch": 4.321397379912664,
+ "grad_norm": 0.02237936295568943,
+ "learning_rate": 0.0006,
+ "loss": 6.387770652770996,
+ "step": 311
+ },
+ {
+ "epoch": 4.335371179039301,
+ "grad_norm": 0.026658328250050545,
+ "learning_rate": 0.0006,
+ "loss": 6.323662757873535,
+ "step": 312
+ },
+ {
+ "epoch": 4.349344978165939,
+ "grad_norm": 0.02852284163236618,
+ "learning_rate": 0.0006,
+ "loss": 6.382204055786133,
+ "step": 313
+ },
+ {
+ "epoch": 4.3633187772925766,
+ "grad_norm": 0.028815090656280518,
+ "learning_rate": 0.0006,
+ "loss": 6.402390480041504,
+ "step": 314
+ },
+ {
+ "epoch": 4.377292576419214,
+ "grad_norm": 0.029393529519438744,
+ "learning_rate": 0.0006,
+ "loss": 6.280278205871582,
+ "step": 315
+ },
+ {
+ "epoch": 4.391266375545851,
+ "grad_norm": 0.02186041697859764,
+ "learning_rate": 0.0006,
+ "loss": 6.302707672119141,
+ "step": 316
+ },
+ {
+ "epoch": 4.405240174672489,
+ "grad_norm": 0.02071218006312847,
+ "learning_rate": 0.0006,
+ "loss": 6.3365020751953125,
+ "step": 317
+ },
+ {
+ "epoch": 4.419213973799127,
+ "grad_norm": 0.02779117226600647,
+ "learning_rate": 0.0006,
+ "loss": 6.252157211303711,
+ "step": 318
+ },
+ {
+ "epoch": 4.4331877729257645,
+ "grad_norm": 0.03470654785633087,
+ "learning_rate": 0.0006,
+ "loss": 6.319070339202881,
+ "step": 319
+ },
+ {
+ "epoch": 4.447161572052401,
+ "grad_norm": 0.037067804485559464,
+ "learning_rate": 0.0006,
+ "loss": 6.314116477966309,
+ "step": 320
+ },
+ {
+ "epoch": 4.461135371179039,
+ "grad_norm": 0.032182756811380386,
+ "learning_rate": 0.0006,
+ "loss": 6.267298698425293,
+ "step": 321
+ },
+ {
+ "epoch": 4.475109170305677,
+ "grad_norm": 0.026305217295885086,
+ "learning_rate": 0.0006,
+ "loss": 6.331688404083252,
+ "step": 322
+ },
+ {
+ "epoch": 4.489082969432315,
+ "grad_norm": 0.027228357270359993,
+ "learning_rate": 0.0006,
+ "loss": 6.31139612197876,
+ "step": 323
+ },
+ {
+ "epoch": 4.503056768558952,
+ "grad_norm": 0.025781169533729553,
+ "learning_rate": 0.0006,
+ "loss": 6.194684982299805,
+ "step": 324
+ },
+ {
+ "epoch": 4.517030567685589,
+ "grad_norm": 0.029186977073550224,
+ "learning_rate": 0.0006,
+ "loss": 6.31306266784668,
+ "step": 325
+ },
+ {
+ "epoch": 4.531004366812227,
+ "grad_norm": 0.023608777672052383,
+ "learning_rate": 0.0006,
+ "loss": 6.285243034362793,
+ "step": 326
+ },
+ {
+ "epoch": 4.544978165938865,
+ "grad_norm": 0.02267594076693058,
+ "learning_rate": 0.0006,
+ "loss": 6.30342960357666,
+ "step": 327
+ },
+ {
+ "epoch": 4.558951965065502,
+ "grad_norm": 0.01926310732960701,
+ "learning_rate": 0.0006,
+ "loss": 6.280606269836426,
+ "step": 328
+ },
+ {
+ "epoch": 4.5729257641921395,
+ "grad_norm": 0.026484837755560875,
+ "learning_rate": 0.0006,
+ "loss": 6.237980365753174,
+ "step": 329
+ },
+ {
+ "epoch": 4.586899563318777,
+ "grad_norm": 0.026228854432702065,
+ "learning_rate": 0.0006,
+ "loss": 6.3018293380737305,
+ "step": 330
+ },
+ {
+ "epoch": 4.600873362445415,
+ "grad_norm": 0.022096967324614525,
+ "learning_rate": 0.0006,
+ "loss": 6.253863334655762,
+ "step": 331
+ },
+ {
+ "epoch": 4.614847161572053,
+ "grad_norm": 0.027223890647292137,
+ "learning_rate": 0.0006,
+ "loss": 6.166882038116455,
+ "step": 332
+ },
+ {
+ "epoch": 4.62882096069869,
+ "grad_norm": 0.03160124272108078,
+ "learning_rate": 0.0006,
+ "loss": 6.230309963226318,
+ "step": 333
+ },
+ {
+ "epoch": 4.642794759825327,
+ "grad_norm": 0.04110539332032204,
+ "learning_rate": 0.0006,
+ "loss": 6.2334818840026855,
+ "step": 334
+ },
+ {
+ "epoch": 4.656768558951965,
+ "grad_norm": 0.04355933889746666,
+ "learning_rate": 0.0006,
+ "loss": 6.160956382751465,
+ "step": 335
+ },
+ {
+ "epoch": 4.670742358078603,
+ "grad_norm": 0.03611086308956146,
+ "learning_rate": 0.0006,
+ "loss": 6.23399543762207,
+ "step": 336
+ },
+ {
+ "epoch": 4.68471615720524,
+ "grad_norm": 0.051811400800943375,
+ "learning_rate": 0.0006,
+ "loss": 6.241855144500732,
+ "step": 337
+ },
+ {
+ "epoch": 4.698689956331878,
+ "grad_norm": 0.057434193789958954,
+ "learning_rate": 0.0006,
+ "loss": 6.220993995666504,
+ "step": 338
+ },
+ {
+ "epoch": 4.712663755458515,
+ "grad_norm": 0.04420953616499901,
+ "learning_rate": 0.0006,
+ "loss": 6.300461292266846,
+ "step": 339
+ },
+ {
+ "epoch": 4.726637554585153,
+ "grad_norm": 0.03515457361936569,
+ "learning_rate": 0.0006,
+ "loss": 6.269349098205566,
+ "step": 340
+ },
+ {
+ "epoch": 4.74061135371179,
+ "grad_norm": 0.03222600743174553,
+ "learning_rate": 0.0006,
+ "loss": 6.248613357543945,
+ "step": 341
+ },
+ {
+ "epoch": 4.754585152838428,
+ "grad_norm": 0.0384046845138073,
+ "learning_rate": 0.0006,
+ "loss": 6.266979217529297,
+ "step": 342
+ },
+ {
+ "epoch": 4.7685589519650655,
+ "grad_norm": 0.0514867827296257,
+ "learning_rate": 0.0006,
+ "loss": 6.234103202819824,
+ "step": 343
+ },
+ {
+ "epoch": 4.782532751091703,
+ "grad_norm": 0.050747793167829514,
+ "learning_rate": 0.0006,
+ "loss": 6.322582244873047,
+ "step": 344
+ },
+ {
+ "epoch": 4.796506550218341,
+ "grad_norm": 0.040878988802433014,
+ "learning_rate": 0.0006,
+ "loss": 6.212509632110596,
+ "step": 345
+ },
+ {
+ "epoch": 4.810480349344978,
+ "grad_norm": 0.034655820578336716,
+ "learning_rate": 0.0006,
+ "loss": 6.265879154205322,
+ "step": 346
+ },
+ {
+ "epoch": 4.824454148471616,
+ "grad_norm": 0.04202224686741829,
+ "learning_rate": 0.0006,
+ "loss": 6.2896409034729,
+ "step": 347
+ },
+ {
+ "epoch": 4.8384279475982535,
+ "grad_norm": 0.03632277995347977,
+ "learning_rate": 0.0006,
+ "loss": 6.253917217254639,
+ "step": 348
+ },
+ {
+ "epoch": 4.85240174672489,
+ "grad_norm": 0.028645765036344528,
+ "learning_rate": 0.0006,
+ "loss": 6.207965850830078,
+ "step": 349
+ },
+ {
+ "epoch": 4.866375545851528,
+ "grad_norm": 0.025305170565843582,
+ "learning_rate": 0.0006,
+ "loss": 6.168068885803223,
+ "step": 350
+ },
+ {
+ "epoch": 4.880349344978166,
+ "grad_norm": 0.022137803956866264,
+ "learning_rate": 0.0006,
+ "loss": 6.235283374786377,
+ "step": 351
+ },
+ {
+ "epoch": 4.894323144104804,
+ "grad_norm": 0.01814538985490799,
+ "learning_rate": 0.0006,
+ "loss": 6.259641647338867,
+ "step": 352
+ },
+ {
+ "epoch": 4.908296943231441,
+ "grad_norm": 0.018465086817741394,
+ "learning_rate": 0.0006,
+ "loss": 6.111738681793213,
+ "step": 353
+ },
+ {
+ "epoch": 4.922270742358078,
+ "grad_norm": 0.018638933077454567,
+ "learning_rate": 0.0006,
+ "loss": 6.225642204284668,
+ "step": 354
+ },
+ {
+ "epoch": 4.936244541484716,
+ "grad_norm": 0.016461260616779327,
+ "learning_rate": 0.0006,
+ "loss": 6.140647888183594,
+ "step": 355
+ },
+ {
+ "epoch": 4.950218340611354,
+ "grad_norm": 0.015458385460078716,
+ "learning_rate": 0.0006,
+ "loss": 6.155224800109863,
+ "step": 356
+ },
+ {
+ "epoch": 4.964192139737992,
+ "grad_norm": 0.014934048056602478,
+ "learning_rate": 0.0006,
+ "loss": 6.101555824279785,
+ "step": 357
+ },
+ {
+ "epoch": 4.978165938864628,
+ "grad_norm": 0.013979545794427395,
+ "learning_rate": 0.0006,
+ "loss": 6.161718845367432,
+ "step": 358
+ },
+ {
+ "epoch": 4.992139737991266,
+ "grad_norm": 0.01640394888818264,
+ "learning_rate": 0.0006,
+ "loss": 6.055768966674805,
+ "step": 359
+ },
+ {
+ "epoch": 5.0,
+ "grad_norm": 0.015135680325329304,
+ "learning_rate": 0.0006,
+ "loss": 6.169695854187012,
+ "step": 360
+ },
+ {
+ "epoch": 5.0,
+ "eval_loss": 6.16408109664917,
+ "eval_runtime": 59.3649,
+ "eval_samples_per_second": 41.135,
+ "eval_steps_per_second": 1.297,
+ "step": 360
+ },
+ {
+ "epoch": 5.013973799126638,
+ "grad_norm": 0.012849048711359501,
+ "learning_rate": 0.0006,
+ "loss": 6.055830001831055,
+ "step": 361
+ },
+ {
+ "epoch": 5.0279475982532755,
+ "grad_norm": 0.018169576302170753,
+ "learning_rate": 0.0006,
+ "loss": 6.082655429840088,
+ "step": 362
+ },
+ {
+ "epoch": 5.041921397379912,
+ "grad_norm": 0.024293430149555206,
+ "learning_rate": 0.0006,
+ "loss": 6.109032154083252,
+ "step": 363
+ },
+ {
+ "epoch": 5.05589519650655,
+ "grad_norm": 0.03266787901520729,
+ "learning_rate": 0.0006,
+ "loss": 6.030613899230957,
+ "step": 364
+ },
+ {
+ "epoch": 5.069868995633188,
+ "grad_norm": 0.03572266176342964,
+ "learning_rate": 0.0006,
+ "loss": 6.050390720367432,
+ "step": 365
+ },
+ {
+ "epoch": 5.083842794759826,
+ "grad_norm": 0.032625213265419006,
+ "learning_rate": 0.0006,
+ "loss": 6.169313430786133,
+ "step": 366
+ },
+ {
+ "epoch": 5.097816593886463,
+ "grad_norm": 0.04051872715353966,
+ "learning_rate": 0.0006,
+ "loss": 6.10398006439209,
+ "step": 367
+ },
+ {
+ "epoch": 5.1117903930131,
+ "grad_norm": 0.0351213738322258,
+ "learning_rate": 0.0006,
+ "loss": 6.121973514556885,
+ "step": 368
+ },
+ {
+ "epoch": 5.125764192139738,
+ "grad_norm": 0.03523759916424751,
+ "learning_rate": 0.0006,
+ "loss": 6.0626654624938965,
+ "step": 369
+ },
+ {
+ "epoch": 5.139737991266376,
+ "grad_norm": 0.02706182189285755,
+ "learning_rate": 0.0006,
+ "loss": 6.049354076385498,
+ "step": 370
+ },
+ {
+ "epoch": 5.153711790393013,
+ "grad_norm": 0.024212589487433434,
+ "learning_rate": 0.0006,
+ "loss": 6.034826278686523,
+ "step": 371
+ },
+ {
+ "epoch": 5.1676855895196505,
+ "grad_norm": 0.024798082187771797,
+ "learning_rate": 0.0006,
+ "loss": 6.004058837890625,
+ "step": 372
+ },
+ {
+ "epoch": 5.181659388646288,
+ "grad_norm": 0.02325567416846752,
+ "learning_rate": 0.0006,
+ "loss": 5.986461639404297,
+ "step": 373
+ },
+ {
+ "epoch": 5.195633187772926,
+ "grad_norm": 0.019060570746660233,
+ "learning_rate": 0.0006,
+ "loss": 6.059736251831055,
+ "step": 374
+ },
+ {
+ "epoch": 5.209606986899563,
+ "grad_norm": 0.016822976991534233,
+ "learning_rate": 0.0006,
+ "loss": 6.072957515716553,
+ "step": 375
+ },
+ {
+ "epoch": 5.223580786026201,
+ "grad_norm": 0.017218658700585365,
+ "learning_rate": 0.0006,
+ "loss": 6.031739234924316,
+ "step": 376
+ },
+ {
+ "epoch": 5.2375545851528384,
+ "grad_norm": 0.014889383688569069,
+ "learning_rate": 0.0006,
+ "loss": 6.025674343109131,
+ "step": 377
+ },
+ {
+ "epoch": 5.251528384279476,
+ "grad_norm": 0.01708107627928257,
+ "learning_rate": 0.0006,
+ "loss": 6.090451240539551,
+ "step": 378
+ },
+ {
+ "epoch": 5.265502183406113,
+ "grad_norm": 0.020260317251086235,
+ "learning_rate": 0.0006,
+ "loss": 6.094257354736328,
+ "step": 379
+ },
+ {
+ "epoch": 5.279475982532751,
+ "grad_norm": 0.020110400393605232,
+ "learning_rate": 0.0006,
+ "loss": 6.021188735961914,
+ "step": 380
+ },
+ {
+ "epoch": 5.293449781659389,
+ "grad_norm": 0.02027830481529236,
+ "learning_rate": 0.0006,
+ "loss": 6.022156238555908,
+ "step": 381
+ },
+ {
+ "epoch": 5.307423580786026,
+ "grad_norm": 0.022747062146663666,
+ "learning_rate": 0.0006,
+ "loss": 6.011836051940918,
+ "step": 382
+ },
+ {
+ "epoch": 5.321397379912664,
+ "grad_norm": 0.01990230567753315,
+ "learning_rate": 0.0006,
+ "loss": 6.003054618835449,
+ "step": 383
+ },
+ {
+ "epoch": 5.335371179039301,
+ "grad_norm": 0.01596238650381565,
+ "learning_rate": 0.0006,
+ "loss": 5.991410732269287,
+ "step": 384
+ },
+ {
+ "epoch": 5.349344978165939,
+ "grad_norm": 0.015847105532884598,
+ "learning_rate": 0.0006,
+ "loss": 6.100622177124023,
+ "step": 385
+ },
+ {
+ "epoch": 5.3633187772925766,
+ "grad_norm": 0.016179397702217102,
+ "learning_rate": 0.0006,
+ "loss": 6.032659530639648,
+ "step": 386
+ },
+ {
+ "epoch": 5.377292576419214,
+ "grad_norm": 0.018256602808833122,
+ "learning_rate": 0.0006,
+ "loss": 5.961983680725098,
+ "step": 387
+ },
+ {
+ "epoch": 5.391266375545851,
+ "grad_norm": 0.02805912122130394,
+ "learning_rate": 0.0006,
+ "loss": 5.983541965484619,
+ "step": 388
+ },
+ {
+ "epoch": 5.405240174672489,
+ "grad_norm": 0.039082255214452744,
+ "learning_rate": 0.0006,
+ "loss": 6.021474361419678,
+ "step": 389
+ },
+ {
+ "epoch": 5.419213973799127,
+ "grad_norm": 0.036757659167051315,
+ "learning_rate": 0.0006,
+ "loss": 6.005046367645264,
+ "step": 390
+ },
+ {
+ "epoch": 5.4331877729257645,
+ "grad_norm": 0.035277366638183594,
+ "learning_rate": 0.0006,
+ "loss": 6.036296844482422,
+ "step": 391
+ },
+ {
+ "epoch": 5.447161572052401,
+ "grad_norm": 0.034404635429382324,
+ "learning_rate": 0.0006,
+ "loss": 5.994539260864258,
+ "step": 392
+ },
+ {
+ "epoch": 5.461135371179039,
+ "grad_norm": 0.0377206988632679,
+ "learning_rate": 0.0006,
+ "loss": 5.941534519195557,
+ "step": 393
+ },
+ {
+ "epoch": 5.475109170305677,
+ "grad_norm": 0.0389922633767128,
+ "learning_rate": 0.0006,
+ "loss": 5.983644485473633,
+ "step": 394
+ },
+ {
+ "epoch": 5.489082969432315,
+ "grad_norm": 0.04176778718829155,
+ "learning_rate": 0.0006,
+ "loss": 6.030922889709473,
+ "step": 395
+ },
+ {
+ "epoch": 5.503056768558952,
+ "grad_norm": 0.0337153784930706,
+ "learning_rate": 0.0006,
+ "loss": 5.961367130279541,
+ "step": 396
+ },
+ {
+ "epoch": 5.517030567685589,
+ "grad_norm": 0.036116816103458405,
+ "learning_rate": 0.0006,
+ "loss": 5.819280624389648,
+ "step": 397
+ },
+ {
+ "epoch": 5.531004366812227,
+ "grad_norm": 0.030724041163921356,
+ "learning_rate": 0.0006,
+ "loss": 5.906380653381348,
+ "step": 398
+ },
+ {
+ "epoch": 5.544978165938865,
+ "grad_norm": 0.030264202505350113,
+ "learning_rate": 0.0006,
+ "loss": 5.975833415985107,
+ "step": 399
+ },
+ {
+ "epoch": 5.558951965065502,
+ "grad_norm": 0.032096318900585175,
+ "learning_rate": 0.0006,
+ "loss": 6.001348972320557,
+ "step": 400
+ },
+ {
+ "epoch": 5.5729257641921395,
+ "grad_norm": 0.030579356476664543,
+ "learning_rate": 0.0006,
+ "loss": 5.928768634796143,
+ "step": 401
+ },
+ {
+ "epoch": 5.586899563318777,
+ "grad_norm": 0.028241973370313644,
+ "learning_rate": 0.0006,
+ "loss": 5.92582893371582,
+ "step": 402
+ },
+ {
+ "epoch": 5.600873362445415,
+ "grad_norm": 0.023086953908205032,
+ "learning_rate": 0.0006,
+ "loss": 5.892926216125488,
+ "step": 403
+ },
+ {
+ "epoch": 5.614847161572053,
+ "grad_norm": 0.02547992393374443,
+ "learning_rate": 0.0006,
+ "loss": 5.903195381164551,
+ "step": 404
+ },
+ {
+ "epoch": 5.62882096069869,
+ "grad_norm": 0.023089831694960594,
+ "learning_rate": 0.0006,
+ "loss": 5.912033557891846,
+ "step": 405
+ },
+ {
+ "epoch": 5.642794759825327,
+ "grad_norm": 0.025446368381381035,
+ "learning_rate": 0.0006,
+ "loss": 5.926138401031494,
+ "step": 406
+ },
+ {
+ "epoch": 5.656768558951965,
+ "grad_norm": 0.03748747706413269,
+ "learning_rate": 0.0006,
+ "loss": 5.913451194763184,
+ "step": 407
+ },
+ {
+ "epoch": 5.670742358078603,
+ "grad_norm": 0.03493810072541237,
+ "learning_rate": 0.0006,
+ "loss": 5.97633695602417,
+ "step": 408
+ },
+ {
+ "epoch": 5.68471615720524,
+ "grad_norm": 0.023131275549530983,
+ "learning_rate": 0.0006,
+ "loss": 5.899571418762207,
+ "step": 409
+ },
+ {
+ "epoch": 5.698689956331878,
+ "grad_norm": 0.02986014075577259,
+ "learning_rate": 0.0006,
+ "loss": 5.894903182983398,
+ "step": 410
+ },
+ {
+ "epoch": 5.712663755458515,
+ "grad_norm": 0.030171144753694534,
+ "learning_rate": 0.0006,
+ "loss": 5.876595497131348,
+ "step": 411
+ },
+ {
+ "epoch": 5.726637554585153,
+ "grad_norm": 0.029377546161413193,
+ "learning_rate": 0.0006,
+ "loss": 5.911637306213379,
+ "step": 412
+ },
+ {
+ "epoch": 5.74061135371179,
+ "grad_norm": 0.029829490929841995,
+ "learning_rate": 0.0006,
+ "loss": 5.819629669189453,
+ "step": 413
+ },
+ {
+ "epoch": 5.754585152838428,
+ "grad_norm": 0.024139299988746643,
+ "learning_rate": 0.0006,
+ "loss": 5.81728458404541,
+ "step": 414
+ },
+ {
+ "epoch": 5.7685589519650655,
+ "grad_norm": 0.021453432738780975,
+ "learning_rate": 0.0006,
+ "loss": 5.987326145172119,
+ "step": 415
+ },
+ {
+ "epoch": 5.782532751091703,
+ "grad_norm": 0.017942246049642563,
+ "learning_rate": 0.0006,
+ "loss": 5.893008232116699,
+ "step": 416
+ },
+ {
+ "epoch": 5.796506550218341,
+ "grad_norm": 0.019724637269973755,
+ "learning_rate": 0.0006,
+ "loss": 5.912441253662109,
+ "step": 417
+ },
+ {
+ "epoch": 5.810480349344978,
+ "grad_norm": 0.018597912043333054,
+ "learning_rate": 0.0006,
+ "loss": 5.860394477844238,
+ "step": 418
+ },
+ {
+ "epoch": 5.824454148471616,
+ "grad_norm": 0.016232412308454514,
+ "learning_rate": 0.0006,
+ "loss": 5.94991397857666,
+ "step": 419
+ },
+ {
+ "epoch": 5.8384279475982535,
+ "grad_norm": 0.01667204685509205,
+ "learning_rate": 0.0006,
+ "loss": 5.8506364822387695,
+ "step": 420
+ },
+ {
+ "epoch": 5.85240174672489,
+ "grad_norm": 0.014232151210308075,
+ "learning_rate": 0.0006,
+ "loss": 5.928424835205078,
+ "step": 421
+ },
+ {
+ "epoch": 5.866375545851528,
+ "grad_norm": 0.01570476032793522,
+ "learning_rate": 0.0006,
+ "loss": 5.953692436218262,
+ "step": 422
+ },
+ {
+ "epoch": 5.880349344978166,
+ "grad_norm": 0.01586179807782173,
+ "learning_rate": 0.0006,
+ "loss": 5.820014476776123,
+ "step": 423
+ },
+ {
+ "epoch": 5.894323144104804,
+ "grad_norm": 0.014729145914316177,
+ "learning_rate": 0.0006,
+ "loss": 5.860154628753662,
+ "step": 424
+ },
+ {
+ "epoch": 5.908296943231441,
+ "grad_norm": 0.013111459091305733,
+ "learning_rate": 0.0006,
+ "loss": 5.85988187789917,
+ "step": 425
+ },
+ {
+ "epoch": 5.922270742358078,
+ "grad_norm": 0.012937922962009907,
+ "learning_rate": 0.0006,
+ "loss": 5.756265640258789,
+ "step": 426
+ },
+ {
+ "epoch": 5.936244541484716,
+ "grad_norm": 0.01377453189343214,
+ "learning_rate": 0.0006,
+ "loss": 5.8333048820495605,
+ "step": 427
+ },
+ {
+ "epoch": 5.950218340611354,
+ "grad_norm": 0.014054795727133751,
+ "learning_rate": 0.0006,
+ "loss": 5.803333759307861,
+ "step": 428
+ },
+ {
+ "epoch": 5.964192139737992,
+ "grad_norm": 0.01674959622323513,
+ "learning_rate": 0.0006,
+ "loss": 5.8718485832214355,
+ "step": 429
+ },
+ {
+ "epoch": 5.978165938864628,
+ "grad_norm": 0.018733017146587372,
+ "learning_rate": 0.0006,
+ "loss": 5.875979900360107,
+ "step": 430
+ },
+ {
+ "epoch": 5.992139737991266,
+ "grad_norm": 0.02088983915746212,
+ "learning_rate": 0.0006,
+ "loss": 5.763634204864502,
+ "step": 431
+ },
+ {
+ "epoch": 6.0,
+ "grad_norm": 0.026705985888838768,
+ "learning_rate": 0.0006,
+ "loss": 5.6922478675842285,
+ "step": 432
+ },
+ {
+ "epoch": 6.0,
+ "eval_loss": 5.847958087921143,
+ "eval_runtime": 58.5451,
+ "eval_samples_per_second": 41.711,
+ "eval_steps_per_second": 1.315,
+ "step": 432
+ },
+ {
+ "epoch": 6.013973799126638,
+ "grad_norm": 0.03181544691324234,
+ "learning_rate": 0.0006,
+ "loss": 5.786368370056152,
+ "step": 433
+ },
+ {
+ "epoch": 6.0279475982532755,
+ "grad_norm": 0.03238112851977348,
+ "learning_rate": 0.0006,
+ "loss": 5.810310363769531,
+ "step": 434
+ },
+ {
+ "epoch": 6.041921397379912,
+ "grad_norm": 0.03916756808757782,
+ "learning_rate": 0.0006,
+ "loss": 5.7761993408203125,
+ "step": 435
+ },
+ {
+ "epoch": 6.05589519650655,
+ "grad_norm": 0.044009730219841,
+ "learning_rate": 0.0006,
+ "loss": 5.898112773895264,
+ "step": 436
+ },
+ {
+ "epoch": 6.069868995633188,
+ "grad_norm": 0.045235246419906616,
+ "learning_rate": 0.0006,
+ "loss": 5.876371383666992,
+ "step": 437
+ },
+ {
+ "epoch": 6.083842794759826,
+ "grad_norm": 0.050218384712934494,
+ "learning_rate": 0.0006,
+ "loss": 5.805103302001953,
+ "step": 438
+ },
+ {
+ "epoch": 6.097816593886463,
+ "grad_norm": 0.05444490164518356,
+ "learning_rate": 0.0006,
+ "loss": 5.826424598693848,
+ "step": 439
+ },
+ {
+ "epoch": 6.1117903930131,
+ "grad_norm": 0.04890982061624527,
+ "learning_rate": 0.0006,
+ "loss": 5.7735395431518555,
+ "step": 440
+ },
+ {
+ "epoch": 6.125764192139738,
+ "grad_norm": 0.05478639155626297,
+ "learning_rate": 0.0006,
+ "loss": 5.823677062988281,
+ "step": 441
+ },
+ {
+ "epoch": 6.139737991266376,
+ "grad_norm": 0.057562313973903656,
+ "learning_rate": 0.0006,
+ "loss": 5.889334678649902,
+ "step": 442
+ },
+ {
+ "epoch": 6.153711790393013,
+ "grad_norm": 0.06447703391313553,
+ "learning_rate": 0.0006,
+ "loss": 5.84621524810791,
+ "step": 443
+ },
+ {
+ "epoch": 6.1676855895196505,
+ "grad_norm": 0.048861872404813766,
+ "learning_rate": 0.0006,
+ "loss": 5.931595802307129,
+ "step": 444
+ },
+ {
+ "epoch": 6.181659388646288,
+ "grad_norm": 0.05521339178085327,
+ "learning_rate": 0.0006,
+ "loss": 5.8229780197143555,
+ "step": 445
+ },
+ {
+ "epoch": 6.195633187772926,
+ "grad_norm": 0.053666990250349045,
+ "learning_rate": 0.0006,
+ "loss": 5.879191875457764,
+ "step": 446
+ },
+ {
+ "epoch": 6.209606986899563,
+ "grad_norm": 0.0451025515794754,
+ "learning_rate": 0.0006,
+ "loss": 5.801628112792969,
+ "step": 447
+ },
+ {
+ "epoch": 6.223580786026201,
+ "grad_norm": 0.04291655868291855,
+ "learning_rate": 0.0006,
+ "loss": 5.914680480957031,
+ "step": 448
+ },
+ {
+ "epoch": 6.2375545851528384,
+ "grad_norm": 0.035102490335702896,
+ "learning_rate": 0.0006,
+ "loss": 5.843682765960693,
+ "step": 449
+ },
+ {
+ "epoch": 6.251528384279476,
+ "grad_norm": 0.03403995931148529,
+ "learning_rate": 0.0006,
+ "loss": 5.827154159545898,
+ "step": 450
+ },
+ {
+ "epoch": 6.265502183406113,
+ "grad_norm": 0.03444375470280647,
+ "learning_rate": 0.0006,
+ "loss": 5.800136566162109,
+ "step": 451
+ },
+ {
+ "epoch": 6.279475982532751,
+ "grad_norm": 0.03165159001946449,
+ "learning_rate": 0.0006,
+ "loss": 5.906252384185791,
+ "step": 452
+ },
+ {
+ "epoch": 6.293449781659389,
+ "grad_norm": 0.026153093203902245,
+ "learning_rate": 0.0006,
+ "loss": 5.7423200607299805,
+ "step": 453
+ },
+ {
+ "epoch": 6.307423580786026,
+ "grad_norm": 0.024357657879590988,
+ "learning_rate": 0.0006,
+ "loss": 5.701364517211914,
+ "step": 454
+ },
+ {
+ "epoch": 6.321397379912664,
+ "grad_norm": 0.021508049219846725,
+ "learning_rate": 0.0006,
+ "loss": 5.8328094482421875,
+ "step": 455
+ },
+ {
+ "epoch": 6.335371179039301,
+ "grad_norm": 0.017313921824097633,
+ "learning_rate": 0.0006,
+ "loss": 5.792145252227783,
+ "step": 456
+ },
+ {
+ "epoch": 6.349344978165939,
+ "grad_norm": 0.018563397228717804,
+ "learning_rate": 0.0006,
+ "loss": 5.732672214508057,
+ "step": 457
+ },
+ {
+ "epoch": 6.3633187772925766,
+ "grad_norm": 0.016568679362535477,
+ "learning_rate": 0.0006,
+ "loss": 5.81948184967041,
+ "step": 458
+ },
+ {
+ "epoch": 6.377292576419214,
+ "grad_norm": 0.014133770950138569,
+ "learning_rate": 0.0006,
+ "loss": 5.734982490539551,
+ "step": 459
+ },
+ {
+ "epoch": 6.391266375545851,
+ "grad_norm": 0.014114447869360447,
+ "learning_rate": 0.0006,
+ "loss": 5.70073127746582,
+ "step": 460
+ },
+ {
+ "epoch": 6.405240174672489,
+ "grad_norm": 0.012907393276691437,
+ "learning_rate": 0.0006,
+ "loss": 5.758626937866211,
+ "step": 461
+ },
+ {
+ "epoch": 6.419213973799127,
+ "grad_norm": 0.012673444114625454,
+ "learning_rate": 0.0006,
+ "loss": 5.775138854980469,
+ "step": 462
+ },
+ {
+ "epoch": 6.4331877729257645,
+ "grad_norm": 0.01311410591006279,
+ "learning_rate": 0.0006,
+ "loss": 5.70920991897583,
+ "step": 463
+ },
+ {
+ "epoch": 6.447161572052401,
+ "grad_norm": 0.012935544364154339,
+ "learning_rate": 0.0006,
+ "loss": 5.781479835510254,
+ "step": 464
+ },
+ {
+ "epoch": 6.461135371179039,
+ "grad_norm": 0.01439738366752863,
+ "learning_rate": 0.0006,
+ "loss": 5.691315650939941,
+ "step": 465
+ },
+ {
+ "epoch": 6.475109170305677,
+ "grad_norm": 0.01205186452716589,
+ "learning_rate": 0.0006,
+ "loss": 5.662499904632568,
+ "step": 466
+ },
+ {
+ "epoch": 6.489082969432315,
+ "grad_norm": 0.011575652286410332,
+ "learning_rate": 0.0006,
+ "loss": 5.6271772384643555,
+ "step": 467
+ },
+ {
+ "epoch": 6.503056768558952,
+ "grad_norm": 0.011672502383589745,
+ "learning_rate": 0.0006,
+ "loss": 5.761662006378174,
+ "step": 468
+ },
+ {
+ "epoch": 6.517030567685589,
+ "grad_norm": 0.011541751213371754,
+ "learning_rate": 0.0006,
+ "loss": 5.762078285217285,
+ "step": 469
+ },
+ {
+ "epoch": 6.531004366812227,
+ "grad_norm": 0.011396365240216255,
+ "learning_rate": 0.0006,
+ "loss": 5.67873477935791,
+ "step": 470
+ },
+ {
+ "epoch": 6.544978165938865,
+ "grad_norm": 0.010688499547541142,
+ "learning_rate": 0.0006,
+ "loss": 5.717051982879639,
+ "step": 471
+ },
+ {
+ "epoch": 6.558951965065502,
+ "grad_norm": 0.012224317528307438,
+ "learning_rate": 0.0006,
+ "loss": 5.707948684692383,
+ "step": 472
+ },
+ {
+ "epoch": 6.5729257641921395,
+ "grad_norm": 0.011856825090944767,
+ "learning_rate": 0.0006,
+ "loss": 5.70878791809082,
+ "step": 473
+ },
+ {
+ "epoch": 6.586899563318777,
+ "grad_norm": 0.01199064590036869,
+ "learning_rate": 0.0006,
+ "loss": 5.708697319030762,
+ "step": 474
+ },
+ {
+ "epoch": 6.600873362445415,
+ "grad_norm": 0.01219446212053299,
+ "learning_rate": 0.0006,
+ "loss": 5.6061577796936035,
+ "step": 475
+ },
+ {
+ "epoch": 6.614847161572053,
+ "grad_norm": 0.013397484086453915,
+ "learning_rate": 0.0006,
+ "loss": 5.624789714813232,
+ "step": 476
+ },
+ {
+ "epoch": 6.62882096069869,
+ "grad_norm": 0.01483136136084795,
+ "learning_rate": 0.0006,
+ "loss": 5.721141338348389,
+ "step": 477
+ },
+ {
+ "epoch": 6.642794759825327,
+ "grad_norm": 0.0194488987326622,
+ "learning_rate": 0.0006,
+ "loss": 5.604279518127441,
+ "step": 478
+ },
+ {
+ "epoch": 6.656768558951965,
+ "grad_norm": 0.02079896256327629,
+ "learning_rate": 0.0006,
+ "loss": 5.586322784423828,
+ "step": 479
+ },
+ {
+ "epoch": 6.670742358078603,
+ "grad_norm": 0.018996229395270348,
+ "learning_rate": 0.0006,
+ "loss": 5.5987772941589355,
+ "step": 480
+ },
+ {
+ "epoch": 6.68471615720524,
+ "grad_norm": 0.015779603272676468,
+ "learning_rate": 0.0006,
+ "loss": 5.588602066040039,
+ "step": 481
+ },
+ {
+ "epoch": 6.698689956331878,
+ "grad_norm": 0.015322973020374775,
+ "learning_rate": 0.0006,
+ "loss": 5.686467170715332,
+ "step": 482
+ },
+ {
+ "epoch": 6.712663755458515,
+ "grad_norm": 0.015282213687896729,
+ "learning_rate": 0.0006,
+ "loss": 5.664676666259766,
+ "step": 483
+ },
+ {
+ "epoch": 6.726637554585153,
+ "grad_norm": 0.016589272767305374,
+ "learning_rate": 0.0006,
+ "loss": 5.666739463806152,
+ "step": 484
+ },
+ {
+ "epoch": 6.74061135371179,
+ "grad_norm": 0.01811421848833561,
+ "learning_rate": 0.0006,
+ "loss": 5.647593021392822,
+ "step": 485
+ },
+ {
+ "epoch": 6.754585152838428,
+ "grad_norm": 0.019652029499411583,
+ "learning_rate": 0.0006,
+ "loss": 5.628726959228516,
+ "step": 486
+ },
+ {
+ "epoch": 6.7685589519650655,
+ "grad_norm": 0.02118665538728237,
+ "learning_rate": 0.0006,
+ "loss": 5.6580705642700195,
+ "step": 487
+ },
+ {
+ "epoch": 6.782532751091703,
+ "grad_norm": 0.02237832546234131,
+ "learning_rate": 0.0006,
+ "loss": 5.635931968688965,
+ "step": 488
+ },
+ {
+ "epoch": 6.796506550218341,
+ "grad_norm": 0.023897631093859673,
+ "learning_rate": 0.0006,
+ "loss": 5.517949104309082,
+ "step": 489
+ },
+ {
+ "epoch": 6.810480349344978,
+ "grad_norm": 0.02442227490246296,
+ "learning_rate": 0.0006,
+ "loss": 5.600021839141846,
+ "step": 490
+ },
+ {
+ "epoch": 6.824454148471616,
+ "grad_norm": 0.024675287306308746,
+ "learning_rate": 0.0006,
+ "loss": 5.6639909744262695,
+ "step": 491
+ },
+ {
+ "epoch": 6.8384279475982535,
+ "grad_norm": 0.030372392386198044,
+ "learning_rate": 0.0006,
+ "loss": 5.582888603210449,
+ "step": 492
+ },
+ {
+ "epoch": 6.85240174672489,
+ "grad_norm": 0.03237690031528473,
+ "learning_rate": 0.0006,
+ "loss": 5.573896884918213,
+ "step": 493
+ },
+ {
+ "epoch": 6.866375545851528,
+ "grad_norm": 0.03511589393019676,
+ "learning_rate": 0.0006,
+ "loss": 5.565878868103027,
+ "step": 494
+ },
+ {
+ "epoch": 6.880349344978166,
+ "grad_norm": 0.03883938118815422,
+ "learning_rate": 0.0006,
+ "loss": 5.58540678024292,
+ "step": 495
+ },
+ {
+ "epoch": 6.894323144104804,
+ "grad_norm": 0.04175502806901932,
+ "learning_rate": 0.0006,
+ "loss": 5.6095194816589355,
+ "step": 496
+ },
+ {
+ "epoch": 6.908296943231441,
+ "grad_norm": 0.04013441503047943,
+ "learning_rate": 0.0006,
+ "loss": 5.654401779174805,
+ "step": 497
+ },
+ {
+ "epoch": 6.922270742358078,
+ "grad_norm": 0.04618770629167557,
+ "learning_rate": 0.0006,
+ "loss": 5.6483473777771,
+ "step": 498
+ },
+ {
+ "epoch": 6.936244541484716,
+ "grad_norm": 0.04507308453321457,
+ "learning_rate": 0.0006,
+ "loss": 5.5856218338012695,
+ "step": 499
+ },
+ {
+ "epoch": 6.950218340611354,
+ "grad_norm": 0.03151383996009827,
+ "learning_rate": 0.0006,
+ "loss": 5.594086647033691,
+ "step": 500
+ },
+ {
+ "epoch": 6.964192139737992,
+ "grad_norm": 0.026883797720074654,
+ "learning_rate": 0.0006,
+ "loss": 5.607676029205322,
+ "step": 501
+ },
+ {
+ "epoch": 6.978165938864628,
+ "grad_norm": 0.02981836162507534,
+ "learning_rate": 0.0006,
+ "loss": 5.563666820526123,
+ "step": 502
+ },
+ {
+ "epoch": 6.992139737991266,
+ "grad_norm": 0.03233477845788002,
+ "learning_rate": 0.0006,
+ "loss": 5.667543411254883,
+ "step": 503
+ },
+ {
+ "epoch": 7.0,
+ "grad_norm": 0.029927456751465797,
+ "learning_rate": 0.0006,
+ "loss": 5.733482837677002,
+ "step": 504
+ },
+ {
+ "epoch": 7.0,
+ "eval_loss": 5.679076671600342,
+ "eval_runtime": 59.5965,
+ "eval_samples_per_second": 40.976,
+ "eval_steps_per_second": 1.292,
+ "step": 504
+ },
+ {
+ "epoch": 7.013973799126638,
+ "grad_norm": 0.027081597596406937,
+ "learning_rate": 0.0006,
+ "loss": 5.634487152099609,
+ "step": 505
+ },
+ {
+ "epoch": 7.0279475982532755,
+ "grad_norm": 0.031959421932697296,
+ "learning_rate": 0.0006,
+ "loss": 5.643294811248779,
+ "step": 506
+ },
+ {
+ "epoch": 7.041921397379912,
+ "grad_norm": 0.02802063524723053,
+ "learning_rate": 0.0006,
+ "loss": 5.573239326477051,
+ "step": 507
+ },
+ {
+ "epoch": 7.05589519650655,
+ "grad_norm": 0.032579705119132996,
+ "learning_rate": 0.0006,
+ "loss": 5.627150058746338,
+ "step": 508
+ },
+ {
+ "epoch": 7.069868995633188,
+ "grad_norm": 0.02503928542137146,
+ "learning_rate": 0.0006,
+ "loss": 5.6045308113098145,
+ "step": 509
+ },
+ {
+ "epoch": 7.083842794759826,
+ "grad_norm": 0.02421317622065544,
+ "learning_rate": 0.0006,
+ "loss": 5.522153377532959,
+ "step": 510
+ },
+ {
+ "epoch": 7.097816593886463,
+ "grad_norm": 0.021983426064252853,
+ "learning_rate": 0.0006,
+ "loss": 5.632939338684082,
+ "step": 511
+ },
+ {
+ "epoch": 7.1117903930131,
+ "grad_norm": 0.021933183073997498,
+ "learning_rate": 0.0006,
+ "loss": 5.592185974121094,
+ "step": 512
+ },
+ {
+ "epoch": 7.125764192139738,
+ "grad_norm": 0.02287031151354313,
+ "learning_rate": 0.0006,
+ "loss": 5.613700866699219,
+ "step": 513
+ },
+ {
+ "epoch": 7.139737991266376,
+ "grad_norm": 0.02281602844595909,
+ "learning_rate": 0.0006,
+ "loss": 5.551383018493652,
+ "step": 514
+ },
+ {
+ "epoch": 7.153711790393013,
+ "grad_norm": 0.021050360053777695,
+ "learning_rate": 0.0006,
+ "loss": 5.55275821685791,
+ "step": 515
+ },
+ {
+ "epoch": 7.1676855895196505,
+ "grad_norm": 0.018299926072359085,
+ "learning_rate": 0.0006,
+ "loss": 5.528225421905518,
+ "step": 516
+ },
+ {
+ "epoch": 7.181659388646288,
+ "grad_norm": 0.02024853602051735,
+ "learning_rate": 0.0006,
+ "loss": 5.522453784942627,
+ "step": 517
+ },
+ {
+ "epoch": 7.195633187772926,
+ "grad_norm": 0.02135239914059639,
+ "learning_rate": 0.0006,
+ "loss": 5.527522563934326,
+ "step": 518
+ },
+ {
+ "epoch": 7.209606986899563,
+ "grad_norm": 0.023394184187054634,
+ "learning_rate": 0.0006,
+ "loss": 5.4535017013549805,
+ "step": 519
+ },
+ {
+ "epoch": 7.223580786026201,
+ "grad_norm": 0.025952080264687538,
+ "learning_rate": 0.0006,
+ "loss": 5.5374674797058105,
+ "step": 520
+ },
+ {
+ "epoch": 7.2375545851528384,
+ "grad_norm": 0.022287718951702118,
+ "learning_rate": 0.0006,
+ "loss": 5.493753433227539,
+ "step": 521
+ },
+ {
+ "epoch": 7.251528384279476,
+ "grad_norm": 0.02018447406589985,
+ "learning_rate": 0.0006,
+ "loss": 5.510575294494629,
+ "step": 522
+ },
+ {
+ "epoch": 7.265502183406113,
+ "grad_norm": 0.02050507813692093,
+ "learning_rate": 0.0006,
+ "loss": 5.466026782989502,
+ "step": 523
+ },
+ {
+ "epoch": 7.279475982532751,
+ "grad_norm": 0.023688802495598793,
+ "learning_rate": 0.0006,
+ "loss": 5.485815525054932,
+ "step": 524
+ },
+ {
+ "epoch": 7.293449781659389,
+ "grad_norm": 0.021078843623399734,
+ "learning_rate": 0.0006,
+ "loss": 5.415581703186035,
+ "step": 525
+ },
+ {
+ "epoch": 7.307423580786026,
+ "grad_norm": 0.018266011029481888,
+ "learning_rate": 0.0006,
+ "loss": 5.402815818786621,
+ "step": 526
+ },
+ {
+ "epoch": 7.321397379912664,
+ "grad_norm": 0.019693493843078613,
+ "learning_rate": 0.0006,
+ "loss": 5.504674911499023,
+ "step": 527
+ },
+ {
+ "epoch": 7.335371179039301,
+ "grad_norm": 0.02031373605132103,
+ "learning_rate": 0.0006,
+ "loss": 5.503837585449219,
+ "step": 528
+ },
+ {
+ "epoch": 7.349344978165939,
+ "grad_norm": 0.017636626958847046,
+ "learning_rate": 0.0006,
+ "loss": 5.530580997467041,
+ "step": 529
+ },
+ {
+ "epoch": 7.3633187772925766,
+ "grad_norm": 0.01787244901061058,
+ "learning_rate": 0.0006,
+ "loss": 5.450218677520752,
+ "step": 530
+ },
+ {
+ "epoch": 7.377292576419214,
+ "grad_norm": 0.0170669574290514,
+ "learning_rate": 0.0006,
+ "loss": 5.481570243835449,
+ "step": 531
+ },
+ {
+ "epoch": 7.391266375545851,
+ "grad_norm": 0.01802165061235428,
+ "learning_rate": 0.0006,
+ "loss": 5.386394500732422,
+ "step": 532
+ },
+ {
+ "epoch": 7.405240174672489,
+ "grad_norm": 0.022949472069740295,
+ "learning_rate": 0.0006,
+ "loss": 5.424929618835449,
+ "step": 533
+ },
+ {
+ "epoch": 7.419213973799127,
+ "grad_norm": 0.034213390201330185,
+ "learning_rate": 0.0006,
+ "loss": 5.346663475036621,
+ "step": 534
+ },
+ {
+ "epoch": 7.4331877729257645,
+ "grad_norm": 0.03847593814134598,
+ "learning_rate": 0.0006,
+ "loss": 5.418482780456543,
+ "step": 535
+ },
+ {
+ "epoch": 7.447161572052401,
+ "grad_norm": 0.029393676668405533,
+ "learning_rate": 0.0006,
+ "loss": 5.456090927124023,
+ "step": 536
+ },
+ {
+ "epoch": 7.461135371179039,
+ "grad_norm": 0.03388667106628418,
+ "learning_rate": 0.0006,
+ "loss": 5.572983741760254,
+ "step": 537
+ },
+ {
+ "epoch": 7.475109170305677,
+ "grad_norm": 0.039690613746643066,
+ "learning_rate": 0.0006,
+ "loss": 5.414067268371582,
+ "step": 538
+ },
+ {
+ "epoch": 7.489082969432315,
+ "grad_norm": 0.03630776330828667,
+ "learning_rate": 0.0006,
+ "loss": 5.523739814758301,
+ "step": 539
+ },
+ {
+ "epoch": 7.503056768558952,
+ "grad_norm": 0.03356870263814926,
+ "learning_rate": 0.0006,
+ "loss": 5.444767951965332,
+ "step": 540
+ },
+ {
+ "epoch": 7.517030567685589,
+ "grad_norm": 0.030812304466962814,
+ "learning_rate": 0.0006,
+ "loss": 5.461244106292725,
+ "step": 541
+ },
+ {
+ "epoch": 7.531004366812227,
+ "grad_norm": 0.03319217637181282,
+ "learning_rate": 0.0006,
+ "loss": 5.502161026000977,
+ "step": 542
+ },
+ {
+ "epoch": 7.544978165938865,
+ "grad_norm": 0.032997481524944305,
+ "learning_rate": 0.0006,
+ "loss": 5.4162278175354,
+ "step": 543
+ },
+ {
+ "epoch": 7.558951965065502,
+ "grad_norm": 0.03364962339401245,
+ "learning_rate": 0.0006,
+ "loss": 5.415736198425293,
+ "step": 544
+ },
+ {
+ "epoch": 7.5729257641921395,
+ "grad_norm": 0.0344221405684948,
+ "learning_rate": 0.0006,
+ "loss": 5.541967868804932,
+ "step": 545
+ },
+ {
+ "epoch": 7.586899563318777,
+ "grad_norm": 0.029609503224492073,
+ "learning_rate": 0.0006,
+ "loss": 5.440614223480225,
+ "step": 546
+ },
+ {
+ "epoch": 7.600873362445415,
+ "grad_norm": 0.029055926948785782,
+ "learning_rate": 0.0006,
+ "loss": 5.462865352630615,
+ "step": 547
+ },
+ {
+ "epoch": 7.614847161572053,
+ "grad_norm": 0.02658848837018013,
+ "learning_rate": 0.0006,
+ "loss": 5.444467544555664,
+ "step": 548
+ },
+ {
+ "epoch": 7.62882096069869,
+ "grad_norm": 0.026276404038071632,
+ "learning_rate": 0.0006,
+ "loss": 5.462278842926025,
+ "step": 549
+ },
+ {
+ "epoch": 7.642794759825327,
+ "grad_norm": 0.0286889486014843,
+ "learning_rate": 0.0006,
+ "loss": 5.4849534034729,
+ "step": 550
+ },
+ {
+ "epoch": 7.656768558951965,
+ "grad_norm": 0.026849817484617233,
+ "learning_rate": 0.0006,
+ "loss": 5.431473731994629,
+ "step": 551
+ },
+ {
+ "epoch": 7.670742358078603,
+ "grad_norm": 0.02312396466732025,
+ "learning_rate": 0.0006,
+ "loss": 5.45462703704834,
+ "step": 552
+ },
+ {
+ "epoch": 7.68471615720524,
+ "grad_norm": 0.026197485625743866,
+ "learning_rate": 0.0006,
+ "loss": 5.452552318572998,
+ "step": 553
+ },
+ {
+ "epoch": 7.698689956331878,
+ "grad_norm": 0.02747255191206932,
+ "learning_rate": 0.0006,
+ "loss": 5.447169780731201,
+ "step": 554
+ },
+ {
+ "epoch": 7.712663755458515,
+ "grad_norm": 0.028123432770371437,
+ "learning_rate": 0.0006,
+ "loss": 5.394400596618652,
+ "step": 555
+ },
+ {
+ "epoch": 7.726637554585153,
+ "grad_norm": 0.02599870041012764,
+ "learning_rate": 0.0006,
+ "loss": 5.398341178894043,
+ "step": 556
+ },
+ {
+ "epoch": 7.74061135371179,
+ "grad_norm": 0.022171657532453537,
+ "learning_rate": 0.0006,
+ "loss": 5.368820667266846,
+ "step": 557
+ },
+ {
+ "epoch": 7.754585152838428,
+ "grad_norm": 0.022309480234980583,
+ "learning_rate": 0.0006,
+ "loss": 5.416772365570068,
+ "step": 558
+ },
+ {
+ "epoch": 7.7685589519650655,
+ "grad_norm": 0.024074165150523186,
+ "learning_rate": 0.0006,
+ "loss": 5.359002113342285,
+ "step": 559
+ },
+ {
+ "epoch": 7.782532751091703,
+ "grad_norm": 0.02636653557419777,
+ "learning_rate": 0.0006,
+ "loss": 5.484368324279785,
+ "step": 560
+ },
+ {
+ "epoch": 7.796506550218341,
+ "grad_norm": 0.02203752100467682,
+ "learning_rate": 0.0006,
+ "loss": 5.375498294830322,
+ "step": 561
+ },
+ {
+ "epoch": 7.810480349344978,
+ "grad_norm": 0.019924819469451904,
+ "learning_rate": 0.0006,
+ "loss": 5.334723472595215,
+ "step": 562
+ },
+ {
+ "epoch": 7.824454148471616,
+ "grad_norm": 0.018755966797471046,
+ "learning_rate": 0.0006,
+ "loss": 5.348781108856201,
+ "step": 563
+ },
+ {
+ "epoch": 7.8384279475982535,
+ "grad_norm": 0.016747845336794853,
+ "learning_rate": 0.0006,
+ "loss": 5.433079719543457,
+ "step": 564
+ },
+ {
+ "epoch": 7.85240174672489,
+ "grad_norm": 0.0170395877212286,
+ "learning_rate": 0.0006,
+ "loss": 5.346794128417969,
+ "step": 565
+ },
+ {
+ "epoch": 7.866375545851528,
+ "grad_norm": 0.016074176877737045,
+ "learning_rate": 0.0006,
+ "loss": 5.3277740478515625,
+ "step": 566
+ },
+ {
+ "epoch": 7.880349344978166,
+ "grad_norm": 0.014119806699454784,
+ "learning_rate": 0.0006,
+ "loss": 5.483551979064941,
+ "step": 567
+ },
+ {
+ "epoch": 7.894323144104804,
+ "grad_norm": 0.015270394273102283,
+ "learning_rate": 0.0006,
+ "loss": 5.378519535064697,
+ "step": 568
+ },
+ {
+ "epoch": 7.908296943231441,
+ "grad_norm": 0.015677539631724358,
+ "learning_rate": 0.0006,
+ "loss": 5.290210247039795,
+ "step": 569
+ },
+ {
+ "epoch": 7.922270742358078,
+ "grad_norm": 0.015930423513054848,
+ "learning_rate": 0.0006,
+ "loss": 5.427360534667969,
+ "step": 570
+ },
+ {
+ "epoch": 7.936244541484716,
+ "grad_norm": 0.016641564667224884,
+ "learning_rate": 0.0006,
+ "loss": 5.301599979400635,
+ "step": 571
+ },
+ {
+ "epoch": 7.950218340611354,
+ "grad_norm": 0.018293552100658417,
+ "learning_rate": 0.0006,
+ "loss": 5.277889251708984,
+ "step": 572
+ },
+ {
+ "epoch": 7.964192139737992,
+ "grad_norm": 0.017618799582123756,
+ "learning_rate": 0.0006,
+ "loss": 5.297194480895996,
+ "step": 573
+ },
+ {
+ "epoch": 7.978165938864628,
+ "grad_norm": 0.015034242533147335,
+ "learning_rate": 0.0006,
+ "loss": 5.42428731918335,
+ "step": 574
+ },
+ {
+ "epoch": 7.992139737991266,
+ "grad_norm": 0.01617511734366417,
+ "learning_rate": 0.0006,
+ "loss": 5.420044898986816,
+ "step": 575
+ },
+ {
+ "epoch": 8.0,
+ "grad_norm": 0.01837257295846939,
+ "learning_rate": 0.0006,
+ "loss": 5.240387916564941,
+ "step": 576
+ },
+ {
+ "epoch": 8.0,
+ "eval_loss": 5.423932075500488,
+ "eval_runtime": 58.0841,
+ "eval_samples_per_second": 42.042,
+ "eval_steps_per_second": 1.326,
+ "step": 576
+ },
+ {
+ "epoch": 8.013973799126637,
+ "grad_norm": 0.020953809842467308,
+ "learning_rate": 0.0006,
+ "loss": 5.4320783615112305,
+ "step": 577
+ },
+ {
+ "epoch": 8.027947598253276,
+ "grad_norm": 0.0227745920419693,
+ "learning_rate": 0.0006,
+ "loss": 5.310197830200195,
+ "step": 578
+ },
+ {
+ "epoch": 8.041921397379912,
+ "grad_norm": 0.021372603252530098,
+ "learning_rate": 0.0006,
+ "loss": 5.339468002319336,
+ "step": 579
+ },
+ {
+ "epoch": 8.055895196506551,
+ "grad_norm": 0.022401731461286545,
+ "learning_rate": 0.0006,
+ "loss": 5.279829978942871,
+ "step": 580
+ },
+ {
+ "epoch": 8.069868995633188,
+ "grad_norm": 0.023942379280924797,
+ "learning_rate": 0.0006,
+ "loss": 5.372323989868164,
+ "step": 581
+ },
+ {
+ "epoch": 8.083842794759825,
+ "grad_norm": 0.024380534887313843,
+ "learning_rate": 0.0006,
+ "loss": 5.253486633300781,
+ "step": 582
+ },
+ {
+ "epoch": 8.097816593886463,
+ "grad_norm": 0.03246625140309334,
+ "learning_rate": 0.0006,
+ "loss": 5.305774211883545,
+ "step": 583
+ },
+ {
+ "epoch": 8.1117903930131,
+ "grad_norm": 0.030399736016988754,
+ "learning_rate": 0.0006,
+ "loss": 5.370976448059082,
+ "step": 584
+ },
+ {
+ "epoch": 8.125764192139737,
+ "grad_norm": 0.027148913592100143,
+ "learning_rate": 0.0006,
+ "loss": 5.368010520935059,
+ "step": 585
+ },
+ {
+ "epoch": 8.139737991266376,
+ "grad_norm": 0.02925540879368782,
+ "learning_rate": 0.0006,
+ "loss": 5.303013801574707,
+ "step": 586
+ },
+ {
+ "epoch": 8.153711790393013,
+ "grad_norm": 0.02915453538298607,
+ "learning_rate": 0.0006,
+ "loss": 5.329667091369629,
+ "step": 587
+ },
+ {
+ "epoch": 8.167685589519651,
+ "grad_norm": 0.031732626259326935,
+ "learning_rate": 0.0006,
+ "loss": 5.3634934425354,
+ "step": 588
+ },
+ {
+ "epoch": 8.181659388646288,
+ "grad_norm": 0.036166731268167496,
+ "learning_rate": 0.0006,
+ "loss": 5.223980903625488,
+ "step": 589
+ },
+ {
+ "epoch": 8.195633187772925,
+ "grad_norm": 0.035654108971357346,
+ "learning_rate": 0.0006,
+ "loss": 5.361059188842773,
+ "step": 590
+ },
+ {
+ "epoch": 8.209606986899564,
+ "grad_norm": 0.0365324430167675,
+ "learning_rate": 0.0006,
+ "loss": 5.336151599884033,
+ "step": 591
+ },
+ {
+ "epoch": 8.2235807860262,
+ "grad_norm": 0.03578880429267883,
+ "learning_rate": 0.0006,
+ "loss": 5.434076309204102,
+ "step": 592
+ },
+ {
+ "epoch": 8.237554585152838,
+ "grad_norm": 0.035734012722969055,
+ "learning_rate": 0.0006,
+ "loss": 5.330893516540527,
+ "step": 593
+ },
+ {
+ "epoch": 8.251528384279476,
+ "grad_norm": 0.03196857124567032,
+ "learning_rate": 0.0006,
+ "loss": 5.306009769439697,
+ "step": 594
+ },
+ {
+ "epoch": 8.265502183406113,
+ "grad_norm": 0.032042086124420166,
+ "learning_rate": 0.0006,
+ "loss": 5.343267917633057,
+ "step": 595
+ },
+ {
+ "epoch": 8.279475982532752,
+ "grad_norm": 0.03160746395587921,
+ "learning_rate": 0.0006,
+ "loss": 5.2353715896606445,
+ "step": 596
+ },
+ {
+ "epoch": 8.293449781659389,
+ "grad_norm": 0.026689818128943443,
+ "learning_rate": 0.0006,
+ "loss": 5.254042625427246,
+ "step": 597
+ },
+ {
+ "epoch": 8.307423580786025,
+ "grad_norm": 0.02880188450217247,
+ "learning_rate": 0.0006,
+ "loss": 5.326833248138428,
+ "step": 598
+ },
+ {
+ "epoch": 8.321397379912664,
+ "grad_norm": 0.027516381815075874,
+ "learning_rate": 0.0006,
+ "loss": 5.342336177825928,
+ "step": 599
+ },
+ {
+ "epoch": 8.335371179039301,
+ "grad_norm": 0.027875030413269997,
+ "learning_rate": 0.0006,
+ "loss": 5.200719833374023,
+ "step": 600
+ },
+ {
+ "epoch": 8.34934497816594,
+ "grad_norm": 0.0268265288323164,
+ "learning_rate": 0.0006,
+ "loss": 5.249449729919434,
+ "step": 601
+ },
+ {
+ "epoch": 8.363318777292577,
+ "grad_norm": 0.024824608117341995,
+ "learning_rate": 0.0006,
+ "loss": 5.318952560424805,
+ "step": 602
+ },
+ {
+ "epoch": 8.377292576419213,
+ "grad_norm": 0.022990627214312553,
+ "learning_rate": 0.0006,
+ "loss": 5.244993209838867,
+ "step": 603
+ },
+ {
+ "epoch": 8.391266375545852,
+ "grad_norm": 0.022804604843258858,
+ "learning_rate": 0.0006,
+ "loss": 5.3236846923828125,
+ "step": 604
+ },
+ {
+ "epoch": 8.405240174672489,
+ "grad_norm": 0.02419872209429741,
+ "learning_rate": 0.0006,
+ "loss": 5.255486488342285,
+ "step": 605
+ },
+ {
+ "epoch": 8.419213973799126,
+ "grad_norm": 0.021952059119939804,
+ "learning_rate": 0.0006,
+ "loss": 5.3511552810668945,
+ "step": 606
+ },
+ {
+ "epoch": 8.433187772925764,
+ "grad_norm": 0.022375497967004776,
+ "learning_rate": 0.0006,
+ "loss": 5.294790267944336,
+ "step": 607
+ },
+ {
+ "epoch": 8.447161572052401,
+ "grad_norm": 0.019242815673351288,
+ "learning_rate": 0.0006,
+ "loss": 5.252618789672852,
+ "step": 608
+ },
+ {
+ "epoch": 8.46113537117904,
+ "grad_norm": 0.0173486340790987,
+ "learning_rate": 0.0006,
+ "loss": 5.30147647857666,
+ "step": 609
+ },
+ {
+ "epoch": 8.475109170305677,
+ "grad_norm": 0.01612604409456253,
+ "learning_rate": 0.0006,
+ "loss": 5.23960542678833,
+ "step": 610
+ },
+ {
+ "epoch": 8.489082969432314,
+ "grad_norm": 0.014293976128101349,
+ "learning_rate": 0.0006,
+ "loss": 5.119932651519775,
+ "step": 611
+ },
+ {
+ "epoch": 8.503056768558952,
+ "grad_norm": 0.014453536830842495,
+ "learning_rate": 0.0006,
+ "loss": 5.166794776916504,
+ "step": 612
+ },
+ {
+ "epoch": 8.51703056768559,
+ "grad_norm": 0.014685769565403461,
+ "learning_rate": 0.0006,
+ "loss": 5.264822959899902,
+ "step": 613
+ },
+ {
+ "epoch": 8.531004366812226,
+ "grad_norm": 0.01427740603685379,
+ "learning_rate": 0.0006,
+ "loss": 5.278433799743652,
+ "step": 614
+ },
+ {
+ "epoch": 8.544978165938865,
+ "grad_norm": 0.014864951372146606,
+ "learning_rate": 0.0006,
+ "loss": 5.214512348175049,
+ "step": 615
+ },
+ {
+ "epoch": 8.558951965065502,
+ "grad_norm": 0.01565164513885975,
+ "learning_rate": 0.0006,
+ "loss": 5.183066368103027,
+ "step": 616
+ },
+ {
+ "epoch": 8.57292576419214,
+ "grad_norm": 0.01779816672205925,
+ "learning_rate": 0.0006,
+ "loss": 5.163185119628906,
+ "step": 617
+ },
+ {
+ "epoch": 8.586899563318777,
+ "grad_norm": 0.017254827544093132,
+ "learning_rate": 0.0006,
+ "loss": 5.243555068969727,
+ "step": 618
+ },
+ {
+ "epoch": 8.600873362445414,
+ "grad_norm": 0.01650584116578102,
+ "learning_rate": 0.0006,
+ "loss": 5.234884738922119,
+ "step": 619
+ },
+ {
+ "epoch": 8.614847161572053,
+ "grad_norm": 0.017021115869283676,
+ "learning_rate": 0.0006,
+ "loss": 5.12271785736084,
+ "step": 620
+ },
+ {
+ "epoch": 8.62882096069869,
+ "grad_norm": 0.01773759163916111,
+ "learning_rate": 0.0006,
+ "loss": 5.242153167724609,
+ "step": 621
+ },
+ {
+ "epoch": 8.642794759825328,
+ "grad_norm": 0.015679042786359787,
+ "learning_rate": 0.0006,
+ "loss": 5.170779228210449,
+ "step": 622
+ },
+ {
+ "epoch": 8.656768558951965,
+ "grad_norm": 0.013760549947619438,
+ "learning_rate": 0.0006,
+ "loss": 5.238644599914551,
+ "step": 623
+ },
+ {
+ "epoch": 8.670742358078602,
+ "grad_norm": 0.014571522362530231,
+ "learning_rate": 0.0006,
+ "loss": 5.177056312561035,
+ "step": 624
+ },
+ {
+ "epoch": 8.68471615720524,
+ "grad_norm": 0.016209015622735023,
+ "learning_rate": 0.0006,
+ "loss": 5.163750648498535,
+ "step": 625
+ },
+ {
+ "epoch": 8.698689956331878,
+ "grad_norm": 0.016813941299915314,
+ "learning_rate": 0.0006,
+ "loss": 5.183428764343262,
+ "step": 626
+ },
+ {
+ "epoch": 8.712663755458514,
+ "grad_norm": 0.020985357463359833,
+ "learning_rate": 0.0006,
+ "loss": 5.229465007781982,
+ "step": 627
+ },
+ {
+ "epoch": 8.726637554585153,
+ "grad_norm": 0.02679632417857647,
+ "learning_rate": 0.0006,
+ "loss": 5.257368087768555,
+ "step": 628
+ },
+ {
+ "epoch": 8.74061135371179,
+ "grad_norm": 0.025756070390343666,
+ "learning_rate": 0.0006,
+ "loss": 5.253736972808838,
+ "step": 629
+ },
+ {
+ "epoch": 8.754585152838429,
+ "grad_norm": 0.02643490768969059,
+ "learning_rate": 0.0006,
+ "loss": 5.228633403778076,
+ "step": 630
+ },
+ {
+ "epoch": 8.768558951965066,
+ "grad_norm": 0.03303210437297821,
+ "learning_rate": 0.0006,
+ "loss": 5.163540363311768,
+ "step": 631
+ },
+ {
+ "epoch": 8.782532751091702,
+ "grad_norm": 0.030432431027293205,
+ "learning_rate": 0.0006,
+ "loss": 5.243169784545898,
+ "step": 632
+ },
+ {
+ "epoch": 8.796506550218341,
+ "grad_norm": 0.028631288558244705,
+ "learning_rate": 0.0006,
+ "loss": 5.166398048400879,
+ "step": 633
+ },
+ {
+ "epoch": 8.810480349344978,
+ "grad_norm": 0.026188310235738754,
+ "learning_rate": 0.0006,
+ "loss": 5.128122329711914,
+ "step": 634
+ },
+ {
+ "epoch": 8.824454148471617,
+ "grad_norm": 0.025308528915047646,
+ "learning_rate": 0.0006,
+ "loss": 5.211284637451172,
+ "step": 635
+ },
+ {
+ "epoch": 8.838427947598253,
+ "grad_norm": 0.020657729357481003,
+ "learning_rate": 0.0006,
+ "loss": 5.161575794219971,
+ "step": 636
+ },
+ {
+ "epoch": 8.85240174672489,
+ "grad_norm": 0.021901234984397888,
+ "learning_rate": 0.0006,
+ "loss": 5.20050048828125,
+ "step": 637
+ },
+ {
+ "epoch": 8.866375545851529,
+ "grad_norm": 0.021219318732619286,
+ "learning_rate": 0.0006,
+ "loss": 5.211699485778809,
+ "step": 638
+ },
+ {
+ "epoch": 8.880349344978166,
+ "grad_norm": 0.0200974028557539,
+ "learning_rate": 0.0006,
+ "loss": 5.092503547668457,
+ "step": 639
+ },
+ {
+ "epoch": 8.894323144104803,
+ "grad_norm": 0.023804882541298866,
+ "learning_rate": 0.0006,
+ "loss": 5.216068267822266,
+ "step": 640
+ },
+ {
+ "epoch": 8.908296943231441,
+ "grad_norm": 0.026088010519742966,
+ "learning_rate": 0.0006,
+ "loss": 5.155592918395996,
+ "step": 641
+ },
+ {
+ "epoch": 8.922270742358078,
+ "grad_norm": 0.02601276896893978,
+ "learning_rate": 0.0006,
+ "loss": 5.197238922119141,
+ "step": 642
+ },
+ {
+ "epoch": 8.936244541484717,
+ "grad_norm": 0.020387211814522743,
+ "learning_rate": 0.0006,
+ "loss": 5.1955437660217285,
+ "step": 643
+ },
+ {
+ "epoch": 8.950218340611354,
+ "grad_norm": 0.019214622676372528,
+ "learning_rate": 0.0006,
+ "loss": 5.206346035003662,
+ "step": 644
+ },
+ {
+ "epoch": 8.96419213973799,
+ "grad_norm": 0.019674314185976982,
+ "learning_rate": 0.0006,
+ "loss": 5.219857215881348,
+ "step": 645
+ },
+ {
+ "epoch": 8.97816593886463,
+ "grad_norm": 0.020318234339356422,
+ "learning_rate": 0.0006,
+ "loss": 5.191132545471191,
+ "step": 646
+ },
+ {
+ "epoch": 8.992139737991266,
+ "grad_norm": 0.021428676322102547,
+ "learning_rate": 0.0006,
+ "loss": 5.214695930480957,
+ "step": 647
+ },
+ {
+ "epoch": 9.0,
+ "grad_norm": 0.023235054686665535,
+ "learning_rate": 0.0006,
+ "loss": 5.229091644287109,
+ "step": 648
+ },
+ {
+ "epoch": 9.0,
+ "eval_loss": 5.289593696594238,
+ "eval_runtime": 56.9256,
+ "eval_samples_per_second": 42.898,
+ "eval_steps_per_second": 1.353,
+ "step": 648
+ },
+ {
+ "epoch": 9.013973799126637,
+ "grad_norm": 0.02320289984345436,
+ "learning_rate": 0.0006,
+ "loss": 4.983669281005859,
+ "step": 649
+ },
+ {
+ "epoch": 9.027947598253276,
+ "grad_norm": 0.018521282821893692,
+ "learning_rate": 0.0006,
+ "loss": 5.200575828552246,
+ "step": 650
+ },
+ {
+ "epoch": 9.041921397379912,
+ "grad_norm": 0.0180169939994812,
+ "learning_rate": 0.0006,
+ "loss": 5.148033618927002,
+ "step": 651
+ },
+ {
+ "epoch": 9.055895196506551,
+ "grad_norm": 0.019969860091805458,
+ "learning_rate": 0.0006,
+ "loss": 5.099125385284424,
+ "step": 652
+ },
+ {
+ "epoch": 9.069868995633188,
+ "grad_norm": 0.01643305830657482,
+ "learning_rate": 0.0006,
+ "loss": 5.092347145080566,
+ "step": 653
+ },
+ {
+ "epoch": 9.083842794759825,
+ "grad_norm": 0.016022363677620888,
+ "learning_rate": 0.0006,
+ "loss": 5.129279613494873,
+ "step": 654
+ },
+ {
+ "epoch": 9.097816593886463,
+ "grad_norm": 0.01657041162252426,
+ "learning_rate": 0.0006,
+ "loss": 5.159562587738037,
+ "step": 655
+ },
+ {
+ "epoch": 9.1117903930131,
+ "grad_norm": 0.017924228683114052,
+ "learning_rate": 0.0006,
+ "loss": 5.165497779846191,
+ "step": 656
+ },
+ {
+ "epoch": 9.125764192139737,
+ "grad_norm": 0.021592361852526665,
+ "learning_rate": 0.0006,
+ "loss": 5.204775810241699,
+ "step": 657
+ },
+ {
+ "epoch": 9.139737991266376,
+ "grad_norm": 0.026924120262265205,
+ "learning_rate": 0.0006,
+ "loss": 5.128296852111816,
+ "step": 658
+ },
+ {
+ "epoch": 9.153711790393013,
+ "grad_norm": 0.023156899958848953,
+ "learning_rate": 0.0006,
+ "loss": 5.1255340576171875,
+ "step": 659
+ },
+ {
+ "epoch": 9.167685589519651,
+ "grad_norm": 0.020670367404818535,
+ "learning_rate": 0.0006,
+ "loss": 5.128479957580566,
+ "step": 660
+ },
+ {
+ "epoch": 9.181659388646288,
+ "grad_norm": 0.025447173044085503,
+ "learning_rate": 0.0006,
+ "loss": 5.154821872711182,
+ "step": 661
+ },
+ {
+ "epoch": 9.195633187772925,
+ "grad_norm": 0.027940068393945694,
+ "learning_rate": 0.0006,
+ "loss": 5.143466472625732,
+ "step": 662
+ },
+ {
+ "epoch": 9.209606986899564,
+ "grad_norm": 0.026847844943404198,
+ "learning_rate": 0.0006,
+ "loss": 5.045665740966797,
+ "step": 663
+ },
+ {
+ "epoch": 9.2235807860262,
+ "grad_norm": 0.02340601570904255,
+ "learning_rate": 0.0006,
+ "loss": 5.21933126449585,
+ "step": 664
+ },
+ {
+ "epoch": 9.237554585152838,
+ "grad_norm": 0.02340371161699295,
+ "learning_rate": 0.0006,
+ "loss": 4.942949295043945,
+ "step": 665
+ },
+ {
+ "epoch": 9.251528384279476,
+ "grad_norm": 0.02221992425620556,
+ "learning_rate": 0.0006,
+ "loss": 5.197099208831787,
+ "step": 666
+ },
+ {
+ "epoch": 9.265502183406113,
+ "grad_norm": 0.023508677259087563,
+ "learning_rate": 0.0006,
+ "loss": 5.022897720336914,
+ "step": 667
+ },
+ {
+ "epoch": 9.279475982532752,
+ "grad_norm": 0.026120394468307495,
+ "learning_rate": 0.0006,
+ "loss": 5.087725639343262,
+ "step": 668
+ },
+ {
+ "epoch": 9.293449781659389,
+ "grad_norm": 0.027273228392004967,
+ "learning_rate": 0.0006,
+ "loss": 5.090963840484619,
+ "step": 669
+ },
+ {
+ "epoch": 9.307423580786025,
+ "grad_norm": 0.03241586685180664,
+ "learning_rate": 0.0006,
+ "loss": 5.047842979431152,
+ "step": 670
+ },
+ {
+ "epoch": 9.321397379912664,
+ "grad_norm": 0.030245667323470116,
+ "learning_rate": 0.0006,
+ "loss": 5.104803085327148,
+ "step": 671
+ },
+ {
+ "epoch": 9.335371179039301,
+ "grad_norm": 0.027698364108800888,
+ "learning_rate": 0.0006,
+ "loss": 5.0644097328186035,
+ "step": 672
+ },
+ {
+ "epoch": 9.34934497816594,
+ "grad_norm": 0.029692554846405983,
+ "learning_rate": 0.0006,
+ "loss": 5.111942291259766,
+ "step": 673
+ },
+ {
+ "epoch": 9.363318777292577,
+ "grad_norm": 0.03328656405210495,
+ "learning_rate": 0.0006,
+ "loss": 5.1574554443359375,
+ "step": 674
+ },
+ {
+ "epoch": 9.377292576419213,
+ "grad_norm": 0.031938180327415466,
+ "learning_rate": 0.0006,
+ "loss": 5.190371990203857,
+ "step": 675
+ },
+ {
+ "epoch": 9.391266375545852,
+ "grad_norm": 0.024858945980668068,
+ "learning_rate": 0.0006,
+ "loss": 5.121420860290527,
+ "step": 676
+ },
+ {
+ "epoch": 9.405240174672489,
+ "grad_norm": 0.023033203557133675,
+ "learning_rate": 0.0006,
+ "loss": 5.0974345207214355,
+ "step": 677
+ },
+ {
+ "epoch": 9.419213973799126,
+ "grad_norm": 0.022083545103669167,
+ "learning_rate": 0.0006,
+ "loss": 5.046319961547852,
+ "step": 678
+ },
+ {
+ "epoch": 9.433187772925764,
+ "grad_norm": 0.0228437427431345,
+ "learning_rate": 0.0006,
+ "loss": 5.119963645935059,
+ "step": 679
+ },
+ {
+ "epoch": 9.447161572052401,
+ "grad_norm": 0.0242351982742548,
+ "learning_rate": 0.0006,
+ "loss": 5.0444560050964355,
+ "step": 680
+ },
+ {
+ "epoch": 9.46113537117904,
+ "grad_norm": 0.02401185780763626,
+ "learning_rate": 0.0006,
+ "loss": 5.063238620758057,
+ "step": 681
+ },
+ {
+ "epoch": 9.475109170305677,
+ "grad_norm": 0.02560959942638874,
+ "learning_rate": 0.0006,
+ "loss": 5.035224914550781,
+ "step": 682
+ },
+ {
+ "epoch": 9.489082969432314,
+ "grad_norm": 0.024985626339912415,
+ "learning_rate": 0.0006,
+ "loss": 5.049284934997559,
+ "step": 683
+ },
+ {
+ "epoch": 9.503056768558952,
+ "grad_norm": 0.023178115487098694,
+ "learning_rate": 0.0006,
+ "loss": 5.2266621589660645,
+ "step": 684
+ },
+ {
+ "epoch": 9.51703056768559,
+ "grad_norm": 0.0215512253344059,
+ "learning_rate": 0.0006,
+ "loss": 5.1875715255737305,
+ "step": 685
+ },
+ {
+ "epoch": 9.531004366812226,
+ "grad_norm": 0.01662837713956833,
+ "learning_rate": 0.0006,
+ "loss": 5.137012481689453,
+ "step": 686
+ },
+ {
+ "epoch": 9.544978165938865,
+ "grad_norm": 0.01952764391899109,
+ "learning_rate": 0.0006,
+ "loss": 5.072500228881836,
+ "step": 687
+ },
+ {
+ "epoch": 9.558951965065502,
+ "grad_norm": 0.017521077767014503,
+ "learning_rate": 0.0006,
+ "loss": 5.045711994171143,
+ "step": 688
+ },
+ {
+ "epoch": 9.57292576419214,
+ "grad_norm": 0.017355265095829964,
+ "learning_rate": 0.0006,
+ "loss": 5.02881383895874,
+ "step": 689
+ },
+ {
+ "epoch": 9.586899563318777,
+ "grad_norm": 0.015587719157338142,
+ "learning_rate": 0.0006,
+ "loss": 5.132787704467773,
+ "step": 690
+ },
+ {
+ "epoch": 9.600873362445414,
+ "grad_norm": 0.017469845712184906,
+ "learning_rate": 0.0006,
+ "loss": 5.105447292327881,
+ "step": 691
+ },
+ {
+ "epoch": 9.614847161572053,
+ "grad_norm": 0.019630100578069687,
+ "learning_rate": 0.0006,
+ "loss": 5.032464981079102,
+ "step": 692
+ },
+ {
+ "epoch": 9.62882096069869,
+ "grad_norm": 0.01784994639456272,
+ "learning_rate": 0.0006,
+ "loss": 5.081345558166504,
+ "step": 693
+ },
+ {
+ "epoch": 9.642794759825328,
+ "grad_norm": 0.01872754842042923,
+ "learning_rate": 0.0006,
+ "loss": 4.9772114753723145,
+ "step": 694
+ },
+ {
+ "epoch": 9.656768558951965,
+ "grad_norm": 0.01999032311141491,
+ "learning_rate": 0.0006,
+ "loss": 5.089740753173828,
+ "step": 695
+ },
+ {
+ "epoch": 9.670742358078602,
+ "grad_norm": 0.01710767112672329,
+ "learning_rate": 0.0006,
+ "loss": 4.9869890213012695,
+ "step": 696
+ },
+ {
+ "epoch": 9.68471615720524,
+ "grad_norm": 0.015144161880016327,
+ "learning_rate": 0.0006,
+ "loss": 5.022564888000488,
+ "step": 697
+ },
+ {
+ "epoch": 9.698689956331878,
+ "grad_norm": 0.01618543453514576,
+ "learning_rate": 0.0006,
+ "loss": 5.035921573638916,
+ "step": 698
+ },
+ {
+ "epoch": 9.712663755458514,
+ "grad_norm": 0.014653387479484081,
+ "learning_rate": 0.0006,
+ "loss": 4.914456367492676,
+ "step": 699
+ },
+ {
+ "epoch": 9.726637554585153,
+ "grad_norm": 0.013434977270662785,
+ "learning_rate": 0.0006,
+ "loss": 5.063764572143555,
+ "step": 700
+ },
+ {
+ "epoch": 9.74061135371179,
+ "grad_norm": 0.015128864906728268,
+ "learning_rate": 0.0006,
+ "loss": 5.052367687225342,
+ "step": 701
+ },
+ {
+ "epoch": 9.754585152838429,
+ "grad_norm": 0.014210768043994904,
+ "learning_rate": 0.0006,
+ "loss": 5.016489028930664,
+ "step": 702
+ },
+ {
+ "epoch": 9.768558951965066,
+ "grad_norm": 0.015191680751740932,
+ "learning_rate": 0.0006,
+ "loss": 4.990176200866699,
+ "step": 703
+ },
+ {
+ "epoch": 9.782532751091702,
+ "grad_norm": 0.018285127356648445,
+ "learning_rate": 0.0006,
+ "loss": 5.015164375305176,
+ "step": 704
+ },
+ {
+ "epoch": 9.796506550218341,
+ "grad_norm": 0.0204079058021307,
+ "learning_rate": 0.0006,
+ "loss": 5.091142654418945,
+ "step": 705
+ },
+ {
+ "epoch": 9.810480349344978,
+ "grad_norm": 0.01980586163699627,
+ "learning_rate": 0.0006,
+ "loss": 5.03350830078125,
+ "step": 706
+ },
+ {
+ "epoch": 9.824454148471617,
+ "grad_norm": 0.02048729732632637,
+ "learning_rate": 0.0006,
+ "loss": 5.03611421585083,
+ "step": 707
+ },
+ {
+ "epoch": 9.838427947598253,
+ "grad_norm": 0.02058163844048977,
+ "learning_rate": 0.0006,
+ "loss": 4.9613447189331055,
+ "step": 708
+ },
+ {
+ "epoch": 9.85240174672489,
+ "grad_norm": 0.018004924058914185,
+ "learning_rate": 0.0006,
+ "loss": 4.966541767120361,
+ "step": 709
+ },
+ {
+ "epoch": 9.866375545851529,
+ "grad_norm": 0.0198152307420969,
+ "learning_rate": 0.0006,
+ "loss": 5.01750373840332,
+ "step": 710
+ },
+ {
+ "epoch": 9.880349344978166,
+ "grad_norm": 0.018532969057559967,
+ "learning_rate": 0.0006,
+ "loss": 5.097424507141113,
+ "step": 711
+ },
+ {
+ "epoch": 9.894323144104803,
+ "grad_norm": 0.018939971923828125,
+ "learning_rate": 0.0006,
+ "loss": 5.177223205566406,
+ "step": 712
+ },
+ {
+ "epoch": 9.908296943231441,
+ "grad_norm": 0.017969170585274696,
+ "learning_rate": 0.0006,
+ "loss": 5.118013858795166,
+ "step": 713
+ },
+ {
+ "epoch": 9.922270742358078,
+ "grad_norm": 0.018021270632743835,
+ "learning_rate": 0.0006,
+ "loss": 5.034787178039551,
+ "step": 714
+ },
+ {
+ "epoch": 9.936244541484717,
+ "grad_norm": 0.017009710893034935,
+ "learning_rate": 0.0006,
+ "loss": 5.003920555114746,
+ "step": 715
+ },
+ {
+ "epoch": 9.950218340611354,
+ "grad_norm": 0.017722049728035927,
+ "learning_rate": 0.0006,
+ "loss": 5.031939506530762,
+ "step": 716
+ },
+ {
+ "epoch": 9.96419213973799,
+ "grad_norm": 0.017134087160229683,
+ "learning_rate": 0.0006,
+ "loss": 5.093474388122559,
+ "step": 717
+ },
+ {
+ "epoch": 9.97816593886463,
+ "grad_norm": 0.018511613830924034,
+ "learning_rate": 0.0006,
+ "loss": 5.078604698181152,
+ "step": 718
+ },
+ {
+ "epoch": 9.992139737991266,
+ "grad_norm": 0.021206334233283997,
+ "learning_rate": 0.0006,
+ "loss": 5.052572727203369,
+ "step": 719
+ },
+ {
+ "epoch": 10.0,
+ "grad_norm": 0.02239062264561653,
+ "learning_rate": 0.0006,
+ "loss": 5.063547134399414,
+ "step": 720
+ },
+ {
+ "epoch": 10.0,
+ "eval_loss": 5.173044681549072,
+ "eval_runtime": 56.7287,
+ "eval_samples_per_second": 43.047,
+ "eval_steps_per_second": 1.357,
+ "step": 720
+ },
+ {
+ "epoch": 10.013973799126637,
+ "grad_norm": 0.020757969468832016,
+ "learning_rate": 0.0006,
+ "loss": 5.045648574829102,
+ "step": 721
+ },
+ {
+ "epoch": 10.027947598253276,
+ "grad_norm": 0.022697702050209045,
+ "learning_rate": 0.0006,
+ "loss": 4.993833065032959,
+ "step": 722
+ },
+ {
+ "epoch": 10.041921397379912,
+ "grad_norm": 0.02434331737458706,
+ "learning_rate": 0.0006,
+ "loss": 5.073219299316406,
+ "step": 723
+ },
+ {
+ "epoch": 10.055895196506551,
+ "grad_norm": 0.025499660521745682,
+ "learning_rate": 0.0006,
+ "loss": 5.019072532653809,
+ "step": 724
+ },
+ {
+ "epoch": 10.069868995633188,
+ "grad_norm": 0.029281053692102432,
+ "learning_rate": 0.0006,
+ "loss": 4.959851264953613,
+ "step": 725
+ },
+ {
+ "epoch": 10.083842794759825,
+ "grad_norm": 0.033878110349178314,
+ "learning_rate": 0.0006,
+ "loss": 5.132717609405518,
+ "step": 726
+ },
+ {
+ "epoch": 10.097816593886463,
+ "grad_norm": 0.03385764732956886,
+ "learning_rate": 0.0006,
+ "loss": 5.0482892990112305,
+ "step": 727
+ },
+ {
+ "epoch": 10.1117903930131,
+ "grad_norm": 0.03466648980975151,
+ "learning_rate": 0.0006,
+ "loss": 4.946855545043945,
+ "step": 728
+ },
+ {
+ "epoch": 10.125764192139737,
+ "grad_norm": 0.028546305373311043,
+ "learning_rate": 0.0006,
+ "loss": 5.027774333953857,
+ "step": 729
+ },
+ {
+ "epoch": 10.139737991266376,
+ "grad_norm": 0.030575869604945183,
+ "learning_rate": 0.0006,
+ "loss": 4.9768476486206055,
+ "step": 730
+ },
+ {
+ "epoch": 10.153711790393013,
+ "grad_norm": 0.03614303469657898,
+ "learning_rate": 0.0006,
+ "loss": 5.078113555908203,
+ "step": 731
+ },
+ {
+ "epoch": 10.167685589519651,
+ "grad_norm": 0.040991660207509995,
+ "learning_rate": 0.0006,
+ "loss": 5.033185958862305,
+ "step": 732
+ },
+ {
+ "epoch": 10.181659388646288,
+ "grad_norm": 0.04551699757575989,
+ "learning_rate": 0.0006,
+ "loss": 5.0820746421813965,
+ "step": 733
+ },
+ {
+ "epoch": 10.195633187772925,
+ "grad_norm": 0.040193989872932434,
+ "learning_rate": 0.0006,
+ "loss": 5.059700965881348,
+ "step": 734
+ },
+ {
+ "epoch": 10.209606986899564,
+ "grad_norm": 0.035851605236530304,
+ "learning_rate": 0.0006,
+ "loss": 5.142467498779297,
+ "step": 735
+ },
+ {
+ "epoch": 10.2235807860262,
+ "grad_norm": 0.034200169146060944,
+ "learning_rate": 0.0006,
+ "loss": 5.072312355041504,
+ "step": 736
+ },
+ {
+ "epoch": 10.237554585152838,
+ "grad_norm": 0.033621110022068024,
+ "learning_rate": 0.0006,
+ "loss": 5.08037805557251,
+ "step": 737
+ },
+ {
+ "epoch": 10.251528384279476,
+ "grad_norm": 0.037372369319200516,
+ "learning_rate": 0.0006,
+ "loss": 4.998425483703613,
+ "step": 738
+ },
+ {
+ "epoch": 10.265502183406113,
+ "grad_norm": 0.037044707685709,
+ "learning_rate": 0.0006,
+ "loss": 4.923993110656738,
+ "step": 739
+ },
+ {
+ "epoch": 10.279475982532752,
+ "grad_norm": 0.030454808846116066,
+ "learning_rate": 0.0006,
+ "loss": 5.0889997482299805,
+ "step": 740
+ },
+ {
+ "epoch": 10.293449781659389,
+ "grad_norm": 0.030969638377428055,
+ "learning_rate": 0.0006,
+ "loss": 4.984112739562988,
+ "step": 741
+ },
+ {
+ "epoch": 10.307423580786025,
+ "grad_norm": 0.0276983380317688,
+ "learning_rate": 0.0006,
+ "loss": 4.955412864685059,
+ "step": 742
+ },
+ {
+ "epoch": 10.321397379912664,
+ "grad_norm": 0.02716052532196045,
+ "learning_rate": 0.0006,
+ "loss": 5.0098876953125,
+ "step": 743
+ },
+ {
+ "epoch": 10.335371179039301,
+ "grad_norm": 0.02403552085161209,
+ "learning_rate": 0.0006,
+ "loss": 5.165160655975342,
+ "step": 744
+ },
+ {
+ "epoch": 10.34934497816594,
+ "grad_norm": 0.0257862601429224,
+ "learning_rate": 0.0006,
+ "loss": 5.001547813415527,
+ "step": 745
+ },
+ {
+ "epoch": 10.363318777292577,
+ "grad_norm": 0.02300378680229187,
+ "learning_rate": 0.0006,
+ "loss": 5.038888931274414,
+ "step": 746
+ },
+ {
+ "epoch": 10.377292576419213,
+ "grad_norm": 0.019766854122281075,
+ "learning_rate": 0.0006,
+ "loss": 4.924291610717773,
+ "step": 747
+ },
+ {
+ "epoch": 10.391266375545852,
+ "grad_norm": 0.01888395845890045,
+ "learning_rate": 0.0006,
+ "loss": 5.059971332550049,
+ "step": 748
+ },
+ {
+ "epoch": 10.405240174672489,
+ "grad_norm": 0.015577499754726887,
+ "learning_rate": 0.0006,
+ "loss": 4.987686634063721,
+ "step": 749
+ },
+ {
+ "epoch": 10.419213973799126,
+ "grad_norm": 0.016012346372008324,
+ "learning_rate": 0.0006,
+ "loss": 5.015157699584961,
+ "step": 750
+ },
+ {
+ "epoch": 10.433187772925764,
+ "grad_norm": 0.016277998685836792,
+ "learning_rate": 0.0006,
+ "loss": 5.004931449890137,
+ "step": 751
+ },
+ {
+ "epoch": 10.447161572052401,
+ "grad_norm": 0.017504561692476273,
+ "learning_rate": 0.0006,
+ "loss": 4.902827739715576,
+ "step": 752
+ },
+ {
+ "epoch": 10.46113537117904,
+ "grad_norm": 0.01569022797048092,
+ "learning_rate": 0.0006,
+ "loss": 4.954863548278809,
+ "step": 753
+ },
+ {
+ "epoch": 10.475109170305677,
+ "grad_norm": 0.014114422723650932,
+ "learning_rate": 0.0006,
+ "loss": 5.00656795501709,
+ "step": 754
+ },
+ {
+ "epoch": 10.489082969432314,
+ "grad_norm": 0.01573832333087921,
+ "learning_rate": 0.0006,
+ "loss": 4.978353023529053,
+ "step": 755
+ },
+ {
+ "epoch": 10.503056768558952,
+ "grad_norm": 0.016480036079883575,
+ "learning_rate": 0.0006,
+ "loss": 4.953327655792236,
+ "step": 756
+ },
+ {
+ "epoch": 10.51703056768559,
+ "grad_norm": 0.014008025638759136,
+ "learning_rate": 0.0006,
+ "loss": 5.050840377807617,
+ "step": 757
+ },
+ {
+ "epoch": 10.531004366812226,
+ "grad_norm": 0.013147998601198196,
+ "learning_rate": 0.0006,
+ "loss": 4.974964141845703,
+ "step": 758
+ },
+ {
+ "epoch": 10.544978165938865,
+ "grad_norm": 0.013807603158056736,
+ "learning_rate": 0.0006,
+ "loss": 4.927907943725586,
+ "step": 759
+ },
+ {
+ "epoch": 10.558951965065502,
+ "grad_norm": 0.013555224984884262,
+ "learning_rate": 0.0006,
+ "loss": 4.980690956115723,
+ "step": 760
+ },
+ {
+ "epoch": 10.57292576419214,
+ "grad_norm": 0.013045408762991428,
+ "learning_rate": 0.0006,
+ "loss": 4.9436116218566895,
+ "step": 761
+ },
+ {
+ "epoch": 10.586899563318777,
+ "grad_norm": 0.012858862057328224,
+ "learning_rate": 0.0006,
+ "loss": 4.9595770835876465,
+ "step": 762
+ },
+ {
+ "epoch": 10.600873362445414,
+ "grad_norm": 0.01468253880739212,
+ "learning_rate": 0.0006,
+ "loss": 4.90725040435791,
+ "step": 763
+ },
+ {
+ "epoch": 10.614847161572053,
+ "grad_norm": 0.013224679045379162,
+ "learning_rate": 0.0006,
+ "loss": 4.877615928649902,
+ "step": 764
+ },
+ {
+ "epoch": 10.62882096069869,
+ "grad_norm": 0.013555348850786686,
+ "learning_rate": 0.0006,
+ "loss": 4.9456610679626465,
+ "step": 765
+ },
+ {
+ "epoch": 10.642794759825328,
+ "grad_norm": 0.014387160539627075,
+ "learning_rate": 0.0006,
+ "loss": 4.9233717918396,
+ "step": 766
+ },
+ {
+ "epoch": 10.656768558951965,
+ "grad_norm": 0.014354088343679905,
+ "learning_rate": 0.0006,
+ "loss": 4.983511924743652,
+ "step": 767
+ },
+ {
+ "epoch": 10.670742358078602,
+ "grad_norm": 0.01332700252532959,
+ "learning_rate": 0.0006,
+ "loss": 4.934416770935059,
+ "step": 768
+ },
+ {
+ "epoch": 10.68471615720524,
+ "grad_norm": 0.011543313041329384,
+ "learning_rate": 0.0006,
+ "loss": 4.96388053894043,
+ "step": 769
+ },
+ {
+ "epoch": 10.698689956331878,
+ "grad_norm": 0.012330878525972366,
+ "learning_rate": 0.0006,
+ "loss": 4.946396350860596,
+ "step": 770
+ },
+ {
+ "epoch": 10.712663755458514,
+ "grad_norm": 0.01379779726266861,
+ "learning_rate": 0.0006,
+ "loss": 4.981019020080566,
+ "step": 771
+ },
+ {
+ "epoch": 10.726637554585153,
+ "grad_norm": 0.014612431637942791,
+ "learning_rate": 0.0006,
+ "loss": 4.894567966461182,
+ "step": 772
+ },
+ {
+ "epoch": 10.74061135371179,
+ "grad_norm": 0.013062899932265282,
+ "learning_rate": 0.0006,
+ "loss": 5.033038139343262,
+ "step": 773
+ },
+ {
+ "epoch": 10.754585152838429,
+ "grad_norm": 0.013898049481213093,
+ "learning_rate": 0.0006,
+ "loss": 4.823636054992676,
+ "step": 774
+ },
+ {
+ "epoch": 10.768558951965066,
+ "grad_norm": 0.01502019353210926,
+ "learning_rate": 0.0006,
+ "loss": 5.020156383514404,
+ "step": 775
+ },
+ {
+ "epoch": 10.782532751091702,
+ "grad_norm": 0.013810204342007637,
+ "learning_rate": 0.0006,
+ "loss": 4.935647010803223,
+ "step": 776
+ },
+ {
+ "epoch": 10.796506550218341,
+ "grad_norm": 0.013616513460874557,
+ "learning_rate": 0.0006,
+ "loss": 4.901486396789551,
+ "step": 777
+ },
+ {
+ "epoch": 10.810480349344978,
+ "grad_norm": 0.013601024635136127,
+ "learning_rate": 0.0006,
+ "loss": 4.931873321533203,
+ "step": 778
+ },
+ {
+ "epoch": 10.824454148471617,
+ "grad_norm": 0.013668350875377655,
+ "learning_rate": 0.0006,
+ "loss": 4.937256813049316,
+ "step": 779
+ },
+ {
+ "epoch": 10.838427947598253,
+ "grad_norm": 0.01581823080778122,
+ "learning_rate": 0.0006,
+ "loss": 4.9611406326293945,
+ "step": 780
+ },
+ {
+ "epoch": 10.85240174672489,
+ "grad_norm": 0.019478484988212585,
+ "learning_rate": 0.0006,
+ "loss": 4.9212236404418945,
+ "step": 781
+ },
+ {
+ "epoch": 10.866375545851529,
+ "grad_norm": 0.026921333745121956,
+ "learning_rate": 0.0006,
+ "loss": 4.88501501083374,
+ "step": 782
+ },
+ {
+ "epoch": 10.880349344978166,
+ "grad_norm": 0.03155914694070816,
+ "learning_rate": 0.0006,
+ "loss": 4.989967346191406,
+ "step": 783
+ },
+ {
+ "epoch": 10.894323144104803,
+ "grad_norm": 0.029689429327845573,
+ "learning_rate": 0.0006,
+ "loss": 4.936000823974609,
+ "step": 784
+ },
+ {
+ "epoch": 10.908296943231441,
+ "grad_norm": 0.026626063510775566,
+ "learning_rate": 0.0006,
+ "loss": 4.984785079956055,
+ "step": 785
+ },
+ {
+ "epoch": 10.922270742358078,
+ "grad_norm": 0.026168212294578552,
+ "learning_rate": 0.0006,
+ "loss": 4.908282279968262,
+ "step": 786
+ },
+ {
+ "epoch": 10.936244541484717,
+ "grad_norm": 0.02828158251941204,
+ "learning_rate": 0.0006,
+ "loss": 4.926390647888184,
+ "step": 787
+ },
+ {
+ "epoch": 10.950218340611354,
+ "grad_norm": 0.02649078331887722,
+ "learning_rate": 0.0006,
+ "loss": 5.00400972366333,
+ "step": 788
+ },
+ {
+ "epoch": 10.96419213973799,
+ "grad_norm": 0.0256856270134449,
+ "learning_rate": 0.0006,
+ "loss": 5.0317487716674805,
+ "step": 789
+ },
+ {
+ "epoch": 10.97816593886463,
+ "grad_norm": 0.024728883057832718,
+ "learning_rate": 0.0006,
+ "loss": 4.9299635887146,
+ "step": 790
+ },
+ {
+ "epoch": 10.992139737991266,
+ "grad_norm": 0.026241116225719452,
+ "learning_rate": 0.0006,
+ "loss": 4.947319984436035,
+ "step": 791
+ },
+ {
+ "epoch": 11.0,
+ "grad_norm": 0.023526743054389954,
+ "learning_rate": 0.0006,
+ "loss": 4.94429349899292,
+ "step": 792
+ },
+ {
+ "epoch": 11.0,
+ "eval_loss": 5.104345798492432,
+ "eval_runtime": 57.0955,
+ "eval_samples_per_second": 42.77,
+ "eval_steps_per_second": 1.349,
+ "step": 792
+ },
+ {
+ "epoch": 11.013973799126637,
+ "grad_norm": 0.0216788612306118,
+ "learning_rate": 0.0006,
+ "loss": 4.881950855255127,
+ "step": 793
+ },
+ {
+ "epoch": 11.027947598253276,
+ "grad_norm": 0.01830463856458664,
+ "learning_rate": 0.0006,
+ "loss": 4.971016883850098,
+ "step": 794
+ },
+ {
+ "epoch": 11.041921397379912,
+ "grad_norm": 0.02204521745443344,
+ "learning_rate": 0.0006,
+ "loss": 4.894623756408691,
+ "step": 795
+ },
+ {
+ "epoch": 11.055895196506551,
+ "grad_norm": 0.0210255216807127,
+ "learning_rate": 0.0006,
+ "loss": 4.834009170532227,
+ "step": 796
+ },
+ {
+ "epoch": 11.069868995633188,
+ "grad_norm": 0.019797448068857193,
+ "learning_rate": 0.0006,
+ "loss": 5.086714744567871,
+ "step": 797
+ },
+ {
+ "epoch": 11.083842794759825,
+ "grad_norm": 0.02036641724407673,
+ "learning_rate": 0.0006,
+ "loss": 4.920056343078613,
+ "step": 798
+ },
+ {
+ "epoch": 11.097816593886463,
+ "grad_norm": 0.01966066285967827,
+ "learning_rate": 0.0006,
+ "loss": 4.879859447479248,
+ "step": 799
+ },
+ {
+ "epoch": 11.1117903930131,
+ "grad_norm": 0.0202149897813797,
+ "learning_rate": 0.0006,
+ "loss": 4.933640480041504,
+ "step": 800
+ },
+ {
+ "epoch": 11.125764192139737,
+ "grad_norm": 0.01893039606511593,
+ "learning_rate": 0.0006,
+ "loss": 4.901970863342285,
+ "step": 801
+ },
+ {
+ "epoch": 11.139737991266376,
+ "grad_norm": 0.016040155664086342,
+ "learning_rate": 0.0006,
+ "loss": 4.905491828918457,
+ "step": 802
+ },
+ {
+ "epoch": 11.153711790393013,
+ "grad_norm": 0.01720350608229637,
+ "learning_rate": 0.0006,
+ "loss": 4.798816680908203,
+ "step": 803
+ },
+ {
+ "epoch": 11.167685589519651,
+ "grad_norm": 0.017002852633595467,
+ "learning_rate": 0.0006,
+ "loss": 4.851859092712402,
+ "step": 804
+ },
+ {
+ "epoch": 11.181659388646288,
+ "grad_norm": 0.015594680793583393,
+ "learning_rate": 0.0006,
+ "loss": 4.90328311920166,
+ "step": 805
+ },
+ {
+ "epoch": 11.195633187772925,
+ "grad_norm": 0.014904462732374668,
+ "learning_rate": 0.0006,
+ "loss": 4.904215335845947,
+ "step": 806
+ },
+ {
+ "epoch": 11.209606986899564,
+ "grad_norm": 0.01366228237748146,
+ "learning_rate": 0.0006,
+ "loss": 4.9813055992126465,
+ "step": 807
+ },
+ {
+ "epoch": 11.2235807860262,
+ "grad_norm": 0.014719638973474503,
+ "learning_rate": 0.0006,
+ "loss": 4.927306175231934,
+ "step": 808
+ },
+ {
+ "epoch": 11.237554585152838,
+ "grad_norm": 0.015016036108136177,
+ "learning_rate": 0.0006,
+ "loss": 4.874215126037598,
+ "step": 809
+ },
+ {
+ "epoch": 11.251528384279476,
+ "grad_norm": 0.014925424009561539,
+ "learning_rate": 0.0006,
+ "loss": 4.838257789611816,
+ "step": 810
+ },
+ {
+ "epoch": 11.265502183406113,
+ "grad_norm": 0.014335823245346546,
+ "learning_rate": 0.0006,
+ "loss": 4.940608024597168,
+ "step": 811
+ },
+ {
+ "epoch": 11.279475982532752,
+ "grad_norm": 0.014987512491643429,
+ "learning_rate": 0.0006,
+ "loss": 4.831531524658203,
+ "step": 812
+ },
+ {
+ "epoch": 11.293449781659389,
+ "grad_norm": 0.016087457537651062,
+ "learning_rate": 0.0006,
+ "loss": 4.863227367401123,
+ "step": 813
+ },
+ {
+ "epoch": 11.307423580786025,
+ "grad_norm": 0.01671300269663334,
+ "learning_rate": 0.0006,
+ "loss": 4.8535661697387695,
+ "step": 814
+ },
+ {
+ "epoch": 11.321397379912664,
+ "grad_norm": 0.015229142270982265,
+ "learning_rate": 0.0006,
+ "loss": 4.948635101318359,
+ "step": 815
+ },
+ {
+ "epoch": 11.335371179039301,
+ "grad_norm": 0.01532587967813015,
+ "learning_rate": 0.0006,
+ "loss": 4.9035797119140625,
+ "step": 816
+ },
+ {
+ "epoch": 11.34934497816594,
+ "grad_norm": 0.015715476125478745,
+ "learning_rate": 0.0006,
+ "loss": 4.894709587097168,
+ "step": 817
+ },
+ {
+ "epoch": 11.363318777292577,
+ "grad_norm": 0.014156977646052837,
+ "learning_rate": 0.0006,
+ "loss": 4.960206985473633,
+ "step": 818
+ },
+ {
+ "epoch": 11.377292576419213,
+ "grad_norm": 0.017076566815376282,
+ "learning_rate": 0.0006,
+ "loss": 5.005850791931152,
+ "step": 819
+ },
+ {
+ "epoch": 11.391266375545852,
+ "grad_norm": 0.0172084029763937,
+ "learning_rate": 0.0006,
+ "loss": 4.903355598449707,
+ "step": 820
+ },
+ {
+ "epoch": 11.405240174672489,
+ "grad_norm": 0.0180258359760046,
+ "learning_rate": 0.0006,
+ "loss": 4.842667579650879,
+ "step": 821
+ },
+ {
+ "epoch": 11.419213973799126,
+ "grad_norm": 0.022130120545625687,
+ "learning_rate": 0.0006,
+ "loss": 4.799970626831055,
+ "step": 822
+ },
+ {
+ "epoch": 11.433187772925764,
+ "grad_norm": 0.02384897693991661,
+ "learning_rate": 0.0006,
+ "loss": 4.860746383666992,
+ "step": 823
+ },
+ {
+ "epoch": 11.447161572052401,
+ "grad_norm": 0.021881572902202606,
+ "learning_rate": 0.0006,
+ "loss": 4.918107032775879,
+ "step": 824
+ },
+ {
+ "epoch": 11.46113537117904,
+ "grad_norm": 0.023085080087184906,
+ "learning_rate": 0.0006,
+ "loss": 4.878946304321289,
+ "step": 825
+ },
+ {
+ "epoch": 11.475109170305677,
+ "grad_norm": 0.02291012555360794,
+ "learning_rate": 0.0006,
+ "loss": 4.958887100219727,
+ "step": 826
+ },
+ {
+ "epoch": 11.489082969432314,
+ "grad_norm": 0.0199968870729208,
+ "learning_rate": 0.0006,
+ "loss": 4.80955696105957,
+ "step": 827
+ },
+ {
+ "epoch": 11.503056768558952,
+ "grad_norm": 0.018702229484915733,
+ "learning_rate": 0.0006,
+ "loss": 5.0276408195495605,
+ "step": 828
+ },
+ {
+ "epoch": 11.51703056768559,
+ "grad_norm": 0.017787311226129532,
+ "learning_rate": 0.0006,
+ "loss": 4.96018123626709,
+ "step": 829
+ },
+ {
+ "epoch": 11.531004366812226,
+ "grad_norm": 0.01704501546919346,
+ "learning_rate": 0.0006,
+ "loss": 4.914553165435791,
+ "step": 830
+ },
+ {
+ "epoch": 11.544978165938865,
+ "grad_norm": 0.017519677057862282,
+ "learning_rate": 0.0006,
+ "loss": 4.798098087310791,
+ "step": 831
+ },
+ {
+ "epoch": 11.558951965065502,
+ "grad_norm": 0.018511991947889328,
+ "learning_rate": 0.0006,
+ "loss": 4.817785263061523,
+ "step": 832
+ },
+ {
+ "epoch": 11.57292576419214,
+ "grad_norm": 0.01876644790172577,
+ "learning_rate": 0.0006,
+ "loss": 4.818995475769043,
+ "step": 833
+ },
+ {
+ "epoch": 11.586899563318777,
+ "grad_norm": 0.01753568835556507,
+ "learning_rate": 0.0006,
+ "loss": 4.811938285827637,
+ "step": 834
+ },
+ {
+ "epoch": 11.600873362445414,
+ "grad_norm": 0.017046531662344933,
+ "learning_rate": 0.0006,
+ "loss": 4.914674282073975,
+ "step": 835
+ },
+ {
+ "epoch": 11.614847161572053,
+ "grad_norm": 0.01802094094455242,
+ "learning_rate": 0.0006,
+ "loss": 4.993284225463867,
+ "step": 836
+ },
+ {
+ "epoch": 11.62882096069869,
+ "grad_norm": 0.019493764266371727,
+ "learning_rate": 0.0006,
+ "loss": 4.895628452301025,
+ "step": 837
+ },
+ {
+ "epoch": 11.642794759825328,
+ "grad_norm": 0.019461210817098618,
+ "learning_rate": 0.0006,
+ "loss": 4.869269371032715,
+ "step": 838
+ },
+ {
+ "epoch": 11.656768558951965,
+ "grad_norm": 0.01981567218899727,
+ "learning_rate": 0.0006,
+ "loss": 4.863052845001221,
+ "step": 839
+ },
+ {
+ "epoch": 11.670742358078602,
+ "grad_norm": 0.019894491881132126,
+ "learning_rate": 0.0006,
+ "loss": 4.8775787353515625,
+ "step": 840
+ },
+ {
+ "epoch": 11.68471615720524,
+ "grad_norm": 0.01875786855816841,
+ "learning_rate": 0.0006,
+ "loss": 4.817763328552246,
+ "step": 841
+ },
+ {
+ "epoch": 11.698689956331878,
+ "grad_norm": 0.020926695317029953,
+ "learning_rate": 0.0006,
+ "loss": 4.7470526695251465,
+ "step": 842
+ },
+ {
+ "epoch": 11.712663755458514,
+ "grad_norm": 0.021477103233337402,
+ "learning_rate": 0.0006,
+ "loss": 4.762682914733887,
+ "step": 843
+ },
+ {
+ "epoch": 11.726637554585153,
+ "grad_norm": 0.020628711208701134,
+ "learning_rate": 0.0006,
+ "loss": 4.807985782623291,
+ "step": 844
+ },
+ {
+ "epoch": 11.74061135371179,
+ "grad_norm": 0.021489612758159637,
+ "learning_rate": 0.0006,
+ "loss": 4.843091011047363,
+ "step": 845
+ },
+ {
+ "epoch": 11.754585152838429,
+ "grad_norm": 0.021194491535425186,
+ "learning_rate": 0.0006,
+ "loss": 4.8300251960754395,
+ "step": 846
+ },
+ {
+ "epoch": 11.768558951965066,
+ "grad_norm": 0.01833350397646427,
+ "learning_rate": 0.0006,
+ "loss": 5.000202178955078,
+ "step": 847
+ },
+ {
+ "epoch": 11.782532751091702,
+ "grad_norm": 0.018096931278705597,
+ "learning_rate": 0.0006,
+ "loss": 4.880648136138916,
+ "step": 848
+ },
+ {
+ "epoch": 11.796506550218341,
+ "grad_norm": 0.0197161752730608,
+ "learning_rate": 0.0006,
+ "loss": 4.788316249847412,
+ "step": 849
+ },
+ {
+ "epoch": 11.810480349344978,
+ "grad_norm": 0.0203181654214859,
+ "learning_rate": 0.0006,
+ "loss": 4.870944976806641,
+ "step": 850
+ },
+ {
+ "epoch": 11.824454148471617,
+ "grad_norm": 0.01908455230295658,
+ "learning_rate": 0.0006,
+ "loss": 4.890144348144531,
+ "step": 851
+ },
+ {
+ "epoch": 11.838427947598253,
+ "grad_norm": 0.021281801164150238,
+ "learning_rate": 0.0006,
+ "loss": 4.89578914642334,
+ "step": 852
+ },
+ {
+ "epoch": 11.85240174672489,
+ "grad_norm": 0.02124273031949997,
+ "learning_rate": 0.0006,
+ "loss": 4.883749008178711,
+ "step": 853
+ },
+ {
+ "epoch": 11.866375545851529,
+ "grad_norm": 0.020969398319721222,
+ "learning_rate": 0.0006,
+ "loss": 4.809142589569092,
+ "step": 854
+ },
+ {
+ "epoch": 11.880349344978166,
+ "grad_norm": 0.01967657171189785,
+ "learning_rate": 0.0006,
+ "loss": 4.821864604949951,
+ "step": 855
+ },
+ {
+ "epoch": 11.894323144104803,
+ "grad_norm": 0.020887333899736404,
+ "learning_rate": 0.0006,
+ "loss": 4.917985916137695,
+ "step": 856
+ },
+ {
+ "epoch": 11.908296943231441,
+ "grad_norm": 0.019203083589673042,
+ "learning_rate": 0.0006,
+ "loss": 4.824960708618164,
+ "step": 857
+ },
+ {
+ "epoch": 11.922270742358078,
+ "grad_norm": 0.020162401720881462,
+ "learning_rate": 0.0006,
+ "loss": 4.772617816925049,
+ "step": 858
+ },
+ {
+ "epoch": 11.936244541484717,
+ "grad_norm": 0.02251487225294113,
+ "learning_rate": 0.0006,
+ "loss": 4.735418319702148,
+ "step": 859
+ },
+ {
+ "epoch": 11.950218340611354,
+ "grad_norm": 0.020313527435064316,
+ "learning_rate": 0.0006,
+ "loss": 4.939233779907227,
+ "step": 860
+ },
+ {
+ "epoch": 11.96419213973799,
+ "grad_norm": 0.021635599434375763,
+ "learning_rate": 0.0006,
+ "loss": 4.861807346343994,
+ "step": 861
+ },
+ {
+ "epoch": 11.97816593886463,
+ "grad_norm": 0.02285311557352543,
+ "learning_rate": 0.0006,
+ "loss": 4.932732582092285,
+ "step": 862
+ },
+ {
+ "epoch": 11.992139737991266,
+ "grad_norm": 0.02072535641491413,
+ "learning_rate": 0.0006,
+ "loss": 4.796767711639404,
+ "step": 863
+ },
+ {
+ "epoch": 12.0,
+ "grad_norm": 0.020588304847478867,
+ "learning_rate": 0.0006,
+ "loss": 4.997537612915039,
+ "step": 864
+ },
+ {
+ "epoch": 12.0,
+ "eval_loss": 5.044651985168457,
+ "eval_runtime": 56.9145,
+ "eval_samples_per_second": 42.906,
+ "eval_steps_per_second": 1.353,
+ "step": 864
+ },
+ {
+ "epoch": 12.013973799126637,
+ "grad_norm": 0.017597166821360588,
+ "learning_rate": 0.0006,
+ "loss": 4.830524444580078,
+ "step": 865
+ },
+ {
+ "epoch": 12.027947598253276,
+ "grad_norm": 0.01716788113117218,
+ "learning_rate": 0.0006,
+ "loss": 4.894325256347656,
+ "step": 866
+ },
+ {
+ "epoch": 12.041921397379912,
+ "grad_norm": 0.017712457105517387,
+ "learning_rate": 0.0006,
+ "loss": 4.868297576904297,
+ "step": 867
+ },
+ {
+ "epoch": 12.055895196506551,
+ "grad_norm": 0.021486390382051468,
+ "learning_rate": 0.0006,
+ "loss": 4.922049522399902,
+ "step": 868
+ },
+ {
+ "epoch": 12.069868995633188,
+ "grad_norm": 0.024560654535889626,
+ "learning_rate": 0.0006,
+ "loss": 4.734241008758545,
+ "step": 869
+ },
+ {
+ "epoch": 12.083842794759825,
+ "grad_norm": 0.025870252400636673,
+ "learning_rate": 0.0006,
+ "loss": 4.868566989898682,
+ "step": 870
+ },
+ {
+ "epoch": 12.097816593886463,
+ "grad_norm": 0.0277389045804739,
+ "learning_rate": 0.0006,
+ "loss": 4.755006790161133,
+ "step": 871
+ },
+ {
+ "epoch": 12.1117903930131,
+ "grad_norm": 0.029100844636559486,
+ "learning_rate": 0.0006,
+ "loss": 4.909667015075684,
+ "step": 872
+ },
+ {
+ "epoch": 12.125764192139737,
+ "grad_norm": 0.031833067536354065,
+ "learning_rate": 0.0006,
+ "loss": 4.858674049377441,
+ "step": 873
+ },
+ {
+ "epoch": 12.139737991266376,
+ "grad_norm": 0.032709237188100815,
+ "learning_rate": 0.0006,
+ "loss": 4.823696136474609,
+ "step": 874
+ },
+ {
+ "epoch": 12.153711790393013,
+ "grad_norm": 0.02995380386710167,
+ "learning_rate": 0.0006,
+ "loss": 4.850235462188721,
+ "step": 875
+ },
+ {
+ "epoch": 12.167685589519651,
+ "grad_norm": 0.034098610281944275,
+ "learning_rate": 0.0006,
+ "loss": 4.8169989585876465,
+ "step": 876
+ },
+ {
+ "epoch": 12.181659388646288,
+ "grad_norm": 0.03377080336213112,
+ "learning_rate": 0.0006,
+ "loss": 4.90738582611084,
+ "step": 877
+ },
+ {
+ "epoch": 12.195633187772925,
+ "grad_norm": 0.029780825600028038,
+ "learning_rate": 0.0006,
+ "loss": 4.811163902282715,
+ "step": 878
+ },
+ {
+ "epoch": 12.209606986899564,
+ "grad_norm": 0.03118024580180645,
+ "learning_rate": 0.0006,
+ "loss": 4.861286640167236,
+ "step": 879
+ },
+ {
+ "epoch": 12.2235807860262,
+ "grad_norm": 0.033052004873752594,
+ "learning_rate": 0.0006,
+ "loss": 4.865021228790283,
+ "step": 880
+ },
+ {
+ "epoch": 12.237554585152838,
+ "grad_norm": 0.029070032760500908,
+ "learning_rate": 0.0006,
+ "loss": 4.947303771972656,
+ "step": 881
+ },
+ {
+ "epoch": 12.251528384279476,
+ "grad_norm": 0.028355715796351433,
+ "learning_rate": 0.0006,
+ "loss": 4.804648399353027,
+ "step": 882
+ },
+ {
+ "epoch": 12.265502183406113,
+ "grad_norm": 0.031257227063179016,
+ "learning_rate": 0.0006,
+ "loss": 4.757718086242676,
+ "step": 883
+ },
+ {
+ "epoch": 12.279475982532752,
+ "grad_norm": 0.031438373029232025,
+ "learning_rate": 0.0006,
+ "loss": 4.851268768310547,
+ "step": 884
+ },
+ {
+ "epoch": 12.293449781659389,
+ "grad_norm": 0.030492138117551804,
+ "learning_rate": 0.0006,
+ "loss": 4.7891035079956055,
+ "step": 885
+ },
+ {
+ "epoch": 12.307423580786025,
+ "grad_norm": 0.026115991175174713,
+ "learning_rate": 0.0006,
+ "loss": 4.86474609375,
+ "step": 886
+ },
+ {
+ "epoch": 12.321397379912664,
+ "grad_norm": 0.024184465408325195,
+ "learning_rate": 0.0006,
+ "loss": 4.860456943511963,
+ "step": 887
+ },
+ {
+ "epoch": 12.335371179039301,
+ "grad_norm": 0.025437982752919197,
+ "learning_rate": 0.0006,
+ "loss": 4.806251049041748,
+ "step": 888
+ },
+ {
+ "epoch": 12.34934497816594,
+ "grad_norm": 0.022390231490135193,
+ "learning_rate": 0.0006,
+ "loss": 4.807086944580078,
+ "step": 889
+ },
+ {
+ "epoch": 12.363318777292577,
+ "grad_norm": 0.020174294710159302,
+ "learning_rate": 0.0006,
+ "loss": 4.78292179107666,
+ "step": 890
+ },
+ {
+ "epoch": 12.377292576419213,
+ "grad_norm": 0.019734324887394905,
+ "learning_rate": 0.0006,
+ "loss": 4.784509658813477,
+ "step": 891
+ },
+ {
+ "epoch": 12.391266375545852,
+ "grad_norm": 0.016733523458242416,
+ "learning_rate": 0.0006,
+ "loss": 4.811755180358887,
+ "step": 892
+ },
+ {
+ "epoch": 12.405240174672489,
+ "grad_norm": 0.01570146717131138,
+ "learning_rate": 0.0006,
+ "loss": 4.720088005065918,
+ "step": 893
+ },
+ {
+ "epoch": 12.419213973799126,
+ "grad_norm": 0.01601393334567547,
+ "learning_rate": 0.0006,
+ "loss": 4.783543109893799,
+ "step": 894
+ },
+ {
+ "epoch": 12.433187772925764,
+ "grad_norm": 0.014035487547516823,
+ "learning_rate": 0.0006,
+ "loss": 4.848340034484863,
+ "step": 895
+ },
+ {
+ "epoch": 12.447161572052401,
+ "grad_norm": 0.015248487703502178,
+ "learning_rate": 0.0006,
+ "loss": 4.773287773132324,
+ "step": 896
+ },
+ {
+ "epoch": 12.46113537117904,
+ "grad_norm": 0.01532268151640892,
+ "learning_rate": 0.0006,
+ "loss": 4.799646854400635,
+ "step": 897
+ },
+ {
+ "epoch": 12.475109170305677,
+ "grad_norm": 0.013982534408569336,
+ "learning_rate": 0.0006,
+ "loss": 4.8090500831604,
+ "step": 898
+ },
+ {
+ "epoch": 12.489082969432314,
+ "grad_norm": 0.014081164263188839,
+ "learning_rate": 0.0006,
+ "loss": 4.751888751983643,
+ "step": 899
+ },
+ {
+ "epoch": 12.503056768558952,
+ "grad_norm": 0.01380133256316185,
+ "learning_rate": 0.0006,
+ "loss": 4.78144645690918,
+ "step": 900
+ },
+ {
+ "epoch": 12.51703056768559,
+ "grad_norm": 0.013555269688367844,
+ "learning_rate": 0.0006,
+ "loss": 4.779257774353027,
+ "step": 901
+ },
+ {
+ "epoch": 12.531004366812226,
+ "grad_norm": 0.013075701892375946,
+ "learning_rate": 0.0006,
+ "loss": 4.820705413818359,
+ "step": 902
+ },
+ {
+ "epoch": 12.544978165938865,
+ "grad_norm": 0.01276316400617361,
+ "learning_rate": 0.0006,
+ "loss": 4.836733818054199,
+ "step": 903
+ },
+ {
+ "epoch": 12.558951965065502,
+ "grad_norm": 0.013033892959356308,
+ "learning_rate": 0.0006,
+ "loss": 4.764178276062012,
+ "step": 904
+ },
+ {
+ "epoch": 12.57292576419214,
+ "grad_norm": 0.013996980153024197,
+ "learning_rate": 0.0006,
+ "loss": 4.767983436584473,
+ "step": 905
+ },
+ {
+ "epoch": 12.586899563318777,
+ "grad_norm": 0.015222184360027313,
+ "learning_rate": 0.0006,
+ "loss": 4.759721755981445,
+ "step": 906
+ },
+ {
+ "epoch": 12.600873362445414,
+ "grad_norm": 0.01623941957950592,
+ "learning_rate": 0.0006,
+ "loss": 4.849811553955078,
+ "step": 907
+ },
+ {
+ "epoch": 12.614847161572053,
+ "grad_norm": 0.015711182728409767,
+ "learning_rate": 0.0006,
+ "loss": 4.7829084396362305,
+ "step": 908
+ },
+ {
+ "epoch": 12.62882096069869,
+ "grad_norm": 0.014137467369437218,
+ "learning_rate": 0.0006,
+ "loss": 4.73321533203125,
+ "step": 909
+ },
+ {
+ "epoch": 12.642794759825328,
+ "grad_norm": 0.012738462537527084,
+ "learning_rate": 0.0006,
+ "loss": 4.870052337646484,
+ "step": 910
+ },
+ {
+ "epoch": 12.656768558951965,
+ "grad_norm": 0.013683130033314228,
+ "learning_rate": 0.0006,
+ "loss": 4.82658052444458,
+ "step": 911
+ },
+ {
+ "epoch": 12.670742358078602,
+ "grad_norm": 0.014409830793738365,
+ "learning_rate": 0.0006,
+ "loss": 4.707494735717773,
+ "step": 912
+ },
+ {
+ "epoch": 12.68471615720524,
+ "grad_norm": 0.013806809671223164,
+ "learning_rate": 0.0006,
+ "loss": 4.778928279876709,
+ "step": 913
+ },
+ {
+ "epoch": 12.698689956331878,
+ "grad_norm": 0.012857912108302116,
+ "learning_rate": 0.0006,
+ "loss": 4.890717506408691,
+ "step": 914
+ },
+ {
+ "epoch": 12.712663755458514,
+ "grad_norm": 0.013828758150339127,
+ "learning_rate": 0.0006,
+ "loss": 4.8357439041137695,
+ "step": 915
+ },
+ {
+ "epoch": 12.726637554585153,
+ "grad_norm": 0.01478351280093193,
+ "learning_rate": 0.0006,
+ "loss": 4.869986534118652,
+ "step": 916
+ },
+ {
+ "epoch": 12.74061135371179,
+ "grad_norm": 0.01326421182602644,
+ "learning_rate": 0.0006,
+ "loss": 4.830646514892578,
+ "step": 917
+ },
+ {
+ "epoch": 12.754585152838429,
+ "grad_norm": 0.013335433788597584,
+ "learning_rate": 0.0006,
+ "loss": 4.7786149978637695,
+ "step": 918
+ },
+ {
+ "epoch": 12.768558951965066,
+ "grad_norm": 0.014261147007346153,
+ "learning_rate": 0.0006,
+ "loss": 4.788693428039551,
+ "step": 919
+ },
+ {
+ "epoch": 12.782532751091702,
+ "grad_norm": 0.014274783432483673,
+ "learning_rate": 0.0006,
+ "loss": 4.77295446395874,
+ "step": 920
+ },
+ {
+ "epoch": 12.796506550218341,
+ "grad_norm": 0.015074674971401691,
+ "learning_rate": 0.0006,
+ "loss": 4.830754280090332,
+ "step": 921
+ },
+ {
+ "epoch": 12.810480349344978,
+ "grad_norm": 0.015599515289068222,
+ "learning_rate": 0.0006,
+ "loss": 4.837759017944336,
+ "step": 922
+ },
+ {
+ "epoch": 12.824454148471617,
+ "grad_norm": 0.014595242217183113,
+ "learning_rate": 0.0006,
+ "loss": 4.882965564727783,
+ "step": 923
+ },
+ {
+ "epoch": 12.838427947598253,
+ "grad_norm": 0.014951467514038086,
+ "learning_rate": 0.0006,
+ "loss": 4.709748268127441,
+ "step": 924
+ },
+ {
+ "epoch": 12.85240174672489,
+ "grad_norm": 0.01539295818656683,
+ "learning_rate": 0.0006,
+ "loss": 4.791160583496094,
+ "step": 925
+ },
+ {
+ "epoch": 12.866375545851529,
+ "grad_norm": 0.015987906605005264,
+ "learning_rate": 0.0006,
+ "loss": 4.661876678466797,
+ "step": 926
+ },
+ {
+ "epoch": 12.880349344978166,
+ "grad_norm": 0.016527341678738594,
+ "learning_rate": 0.0006,
+ "loss": 4.721952438354492,
+ "step": 927
+ },
+ {
+ "epoch": 12.894323144104803,
+ "grad_norm": 0.01637161523103714,
+ "learning_rate": 0.0006,
+ "loss": 4.704446792602539,
+ "step": 928
+ },
+ {
+ "epoch": 12.908296943231441,
+ "grad_norm": 0.015221050009131432,
+ "learning_rate": 0.0006,
+ "loss": 4.771907806396484,
+ "step": 929
+ },
+ {
+ "epoch": 12.922270742358078,
+ "grad_norm": 0.016242166981101036,
+ "learning_rate": 0.0006,
+ "loss": 4.806267738342285,
+ "step": 930
+ },
+ {
+ "epoch": 12.936244541484717,
+ "grad_norm": 0.01628357358276844,
+ "learning_rate": 0.0006,
+ "loss": 4.7252326011657715,
+ "step": 931
+ },
+ {
+ "epoch": 12.950218340611354,
+ "grad_norm": 0.017428133636713028,
+ "learning_rate": 0.0006,
+ "loss": 4.765689849853516,
+ "step": 932
+ },
+ {
+ "epoch": 12.96419213973799,
+ "grad_norm": 0.01870061084628105,
+ "learning_rate": 0.0006,
+ "loss": 4.814488887786865,
+ "step": 933
+ },
+ {
+ "epoch": 12.97816593886463,
+ "grad_norm": 0.019796060398221016,
+ "learning_rate": 0.0006,
+ "loss": 4.739185810089111,
+ "step": 934
+ },
+ {
+ "epoch": 12.992139737991266,
+ "grad_norm": 0.01985820196568966,
+ "learning_rate": 0.0006,
+ "loss": 4.810608863830566,
+ "step": 935
+ },
+ {
+ "epoch": 13.0,
+ "grad_norm": 0.020834477618336678,
+ "learning_rate": 0.0006,
+ "loss": 4.727858066558838,
+ "step": 936
+ },
+ {
+ "epoch": 13.0,
+ "eval_loss": 4.979116916656494,
+ "eval_runtime": 57.2877,
+ "eval_samples_per_second": 42.627,
+ "eval_steps_per_second": 1.344,
+ "step": 936
+ },
+ {
+ "epoch": 13.013973799126637,
+ "grad_norm": 0.02025519125163555,
+ "learning_rate": 0.0006,
+ "loss": 4.628875732421875,
+ "step": 937
+ },
+ {
+ "epoch": 13.027947598253276,
+ "grad_norm": 0.019596368074417114,
+ "learning_rate": 0.0006,
+ "loss": 4.694699287414551,
+ "step": 938
+ },
+ {
+ "epoch": 13.041921397379912,
+ "grad_norm": 0.020494934171438217,
+ "learning_rate": 0.0006,
+ "loss": 4.691821098327637,
+ "step": 939
+ },
+ {
+ "epoch": 13.055895196506551,
+ "grad_norm": 0.020121760666370392,
+ "learning_rate": 0.0006,
+ "loss": 4.71478271484375,
+ "step": 940
+ },
+ {
+ "epoch": 13.069868995633188,
+ "grad_norm": 0.021091405302286148,
+ "learning_rate": 0.0006,
+ "loss": 4.695915222167969,
+ "step": 941
+ },
+ {
+ "epoch": 13.083842794759825,
+ "grad_norm": 0.019787423312664032,
+ "learning_rate": 0.0006,
+ "loss": 4.8652448654174805,
+ "step": 942
+ },
+ {
+ "epoch": 13.097816593886463,
+ "grad_norm": 0.0185833927243948,
+ "learning_rate": 0.0006,
+ "loss": 4.7065887451171875,
+ "step": 943
+ },
+ {
+ "epoch": 13.1117903930131,
+ "grad_norm": 0.021220475435256958,
+ "learning_rate": 0.0006,
+ "loss": 4.741232395172119,
+ "step": 944
+ },
+ {
+ "epoch": 13.125764192139737,
+ "grad_norm": 0.02136622555553913,
+ "learning_rate": 0.0006,
+ "loss": 4.705967903137207,
+ "step": 945
+ },
+ {
+ "epoch": 13.139737991266376,
+ "grad_norm": 0.0191587433218956,
+ "learning_rate": 0.0006,
+ "loss": 4.758667945861816,
+ "step": 946
+ },
+ {
+ "epoch": 13.153711790393013,
+ "grad_norm": 0.018290022388100624,
+ "learning_rate": 0.0006,
+ "loss": 4.760666847229004,
+ "step": 947
+ },
+ {
+ "epoch": 13.167685589519651,
+ "grad_norm": 0.021704409271478653,
+ "learning_rate": 0.0006,
+ "loss": 4.750555992126465,
+ "step": 948
+ },
+ {
+ "epoch": 13.181659388646288,
+ "grad_norm": 0.021668918430805206,
+ "learning_rate": 0.0006,
+ "loss": 4.840903282165527,
+ "step": 949
+ },
+ {
+ "epoch": 13.195633187772925,
+ "grad_norm": 0.019497141242027283,
+ "learning_rate": 0.0006,
+ "loss": 4.738969326019287,
+ "step": 950
+ },
+ {
+ "epoch": 13.209606986899564,
+ "grad_norm": 0.018500596284866333,
+ "learning_rate": 0.0006,
+ "loss": 4.75154972076416,
+ "step": 951
+ },
+ {
+ "epoch": 13.2235807860262,
+ "grad_norm": 0.01872239261865616,
+ "learning_rate": 0.0006,
+ "loss": 4.811738014221191,
+ "step": 952
+ },
+ {
+ "epoch": 13.237554585152838,
+ "grad_norm": 0.020321547985076904,
+ "learning_rate": 0.0006,
+ "loss": 4.815384387969971,
+ "step": 953
+ },
+ {
+ "epoch": 13.251528384279476,
+ "grad_norm": 0.02283575013279915,
+ "learning_rate": 0.0006,
+ "loss": 4.74582052230835,
+ "step": 954
+ },
+ {
+ "epoch": 13.265502183406113,
+ "grad_norm": 0.02474398910999298,
+ "learning_rate": 0.0006,
+ "loss": 4.738375186920166,
+ "step": 955
+ },
+ {
+ "epoch": 13.279475982532752,
+ "grad_norm": 0.02065764181315899,
+ "learning_rate": 0.0006,
+ "loss": 4.746875286102295,
+ "step": 956
+ },
+ {
+ "epoch": 13.293449781659389,
+ "grad_norm": 0.020534520968794823,
+ "learning_rate": 0.0006,
+ "loss": 4.656914710998535,
+ "step": 957
+ },
+ {
+ "epoch": 13.307423580786025,
+ "grad_norm": 0.0206378772854805,
+ "learning_rate": 0.0006,
+ "loss": 4.842615604400635,
+ "step": 958
+ },
+ {
+ "epoch": 13.321397379912664,
+ "grad_norm": 0.017935654148459435,
+ "learning_rate": 0.0006,
+ "loss": 4.717827796936035,
+ "step": 959
+ },
+ {
+ "epoch": 13.335371179039301,
+ "grad_norm": 0.016958363354206085,
+ "learning_rate": 0.0006,
+ "loss": 4.681070327758789,
+ "step": 960
+ },
+ {
+ "epoch": 13.34934497816594,
+ "grad_norm": 0.015735412016510963,
+ "learning_rate": 0.0006,
+ "loss": 4.7957048416137695,
+ "step": 961
+ },
+ {
+ "epoch": 13.363318777292577,
+ "grad_norm": 0.01563699170947075,
+ "learning_rate": 0.0006,
+ "loss": 4.701192378997803,
+ "step": 962
+ },
+ {
+ "epoch": 13.377292576419213,
+ "grad_norm": 0.015207760035991669,
+ "learning_rate": 0.0006,
+ "loss": 4.73256254196167,
+ "step": 963
+ },
+ {
+ "epoch": 13.391266375545852,
+ "grad_norm": 0.015396883711218834,
+ "learning_rate": 0.0006,
+ "loss": 4.76412296295166,
+ "step": 964
+ },
+ {
+ "epoch": 13.405240174672489,
+ "grad_norm": 0.015690583735704422,
+ "learning_rate": 0.0006,
+ "loss": 4.791157245635986,
+ "step": 965
+ },
+ {
+ "epoch": 13.419213973799126,
+ "grad_norm": 0.01755724661052227,
+ "learning_rate": 0.0006,
+ "loss": 4.845666885375977,
+ "step": 966
+ },
+ {
+ "epoch": 13.433187772925764,
+ "grad_norm": 0.019443422555923462,
+ "learning_rate": 0.0006,
+ "loss": 4.5902299880981445,
+ "step": 967
+ },
+ {
+ "epoch": 13.447161572052401,
+ "grad_norm": 0.02118338644504547,
+ "learning_rate": 0.0006,
+ "loss": 4.740627288818359,
+ "step": 968
+ },
+ {
+ "epoch": 13.46113537117904,
+ "grad_norm": 0.02174839936196804,
+ "learning_rate": 0.0006,
+ "loss": 4.814560890197754,
+ "step": 969
+ },
+ {
+ "epoch": 13.475109170305677,
+ "grad_norm": 0.022860374301671982,
+ "learning_rate": 0.0006,
+ "loss": 4.66440486907959,
+ "step": 970
+ },
+ {
+ "epoch": 13.489082969432314,
+ "grad_norm": 0.024742677807807922,
+ "learning_rate": 0.0006,
+ "loss": 4.7451653480529785,
+ "step": 971
+ },
+ {
+ "epoch": 13.503056768558952,
+ "grad_norm": 0.027068907395005226,
+ "learning_rate": 0.0006,
+ "loss": 4.802584171295166,
+ "step": 972
+ },
+ {
+ "epoch": 13.51703056768559,
+ "grad_norm": 0.026316625997424126,
+ "learning_rate": 0.0006,
+ "loss": 4.842397689819336,
+ "step": 973
+ },
+ {
+ "epoch": 13.531004366812226,
+ "grad_norm": 0.024610962718725204,
+ "learning_rate": 0.0006,
+ "loss": 4.650721073150635,
+ "step": 974
+ },
+ {
+ "epoch": 13.544978165938865,
+ "grad_norm": 0.024122603237628937,
+ "learning_rate": 0.0006,
+ "loss": 4.6984357833862305,
+ "step": 975
+ },
+ {
+ "epoch": 13.558951965065502,
+ "grad_norm": 0.022699708119034767,
+ "learning_rate": 0.0006,
+ "loss": 4.793007850646973,
+ "step": 976
+ },
+ {
+ "epoch": 13.57292576419214,
+ "grad_norm": 0.021706534549593925,
+ "learning_rate": 0.0006,
+ "loss": 4.691362380981445,
+ "step": 977
+ },
+ {
+ "epoch": 13.586899563318777,
+ "grad_norm": 0.021363291889429092,
+ "learning_rate": 0.0006,
+ "loss": 4.76762580871582,
+ "step": 978
+ },
+ {
+ "epoch": 13.600873362445414,
+ "grad_norm": 0.023476677015423775,
+ "learning_rate": 0.0006,
+ "loss": 4.777202606201172,
+ "step": 979
+ },
+ {
+ "epoch": 13.614847161572053,
+ "grad_norm": 0.02457100711762905,
+ "learning_rate": 0.0006,
+ "loss": 4.590703010559082,
+ "step": 980
+ },
+ {
+ "epoch": 13.62882096069869,
+ "grad_norm": 0.02469099499285221,
+ "learning_rate": 0.0006,
+ "loss": 4.640185356140137,
+ "step": 981
+ },
+ {
+ "epoch": 13.642794759825328,
+ "grad_norm": 0.024914603680372238,
+ "learning_rate": 0.0006,
+ "loss": 4.772201061248779,
+ "step": 982
+ },
+ {
+ "epoch": 13.656768558951965,
+ "grad_norm": 0.025133173912763596,
+ "learning_rate": 0.0006,
+ "loss": 4.743337154388428,
+ "step": 983
+ },
+ {
+ "epoch": 13.670742358078602,
+ "grad_norm": 0.02034137211740017,
+ "learning_rate": 0.0006,
+ "loss": 4.752127647399902,
+ "step": 984
+ },
+ {
+ "epoch": 13.68471615720524,
+ "grad_norm": 0.019645661115646362,
+ "learning_rate": 0.0006,
+ "loss": 4.838453769683838,
+ "step": 985
+ },
+ {
+ "epoch": 13.698689956331878,
+ "grad_norm": 0.01900297962129116,
+ "learning_rate": 0.0006,
+ "loss": 4.804868221282959,
+ "step": 986
+ },
+ {
+ "epoch": 13.712663755458514,
+ "grad_norm": 0.020292341709136963,
+ "learning_rate": 0.0006,
+ "loss": 4.609743118286133,
+ "step": 987
+ },
+ {
+ "epoch": 13.726637554585153,
+ "grad_norm": 0.017014775425195694,
+ "learning_rate": 0.0006,
+ "loss": 4.740612983703613,
+ "step": 988
+ },
+ {
+ "epoch": 13.74061135371179,
+ "grad_norm": 0.015199844725430012,
+ "learning_rate": 0.0006,
+ "loss": 4.719701766967773,
+ "step": 989
+ },
+ {
+ "epoch": 13.754585152838429,
+ "grad_norm": 0.015125414356589317,
+ "learning_rate": 0.0006,
+ "loss": 4.686114311218262,
+ "step": 990
+ },
+ {
+ "epoch": 13.768558951965066,
+ "grad_norm": 0.014779540710151196,
+ "learning_rate": 0.0006,
+ "loss": 4.674347877502441,
+ "step": 991
+ },
+ {
+ "epoch": 13.782532751091702,
+ "grad_norm": 0.015255163423717022,
+ "learning_rate": 0.0006,
+ "loss": 4.6309309005737305,
+ "step": 992
+ },
+ {
+ "epoch": 13.796506550218341,
+ "grad_norm": 0.013692197389900684,
+ "learning_rate": 0.0006,
+ "loss": 4.698497295379639,
+ "step": 993
+ },
+ {
+ "epoch": 13.810480349344978,
+ "grad_norm": 0.014529623091220856,
+ "learning_rate": 0.0006,
+ "loss": 4.628443717956543,
+ "step": 994
+ },
+ {
+ "epoch": 13.824454148471617,
+ "grad_norm": 0.01576659083366394,
+ "learning_rate": 0.0006,
+ "loss": 4.788015842437744,
+ "step": 995
+ },
+ {
+ "epoch": 13.838427947598253,
+ "grad_norm": 0.015834277495741844,
+ "learning_rate": 0.0006,
+ "loss": 4.672703742980957,
+ "step": 996
+ },
+ {
+ "epoch": 13.85240174672489,
+ "grad_norm": 0.015539503656327724,
+ "learning_rate": 0.0006,
+ "loss": 4.731503486633301,
+ "step": 997
+ },
+ {
+ "epoch": 13.866375545851529,
+ "grad_norm": 0.013463746756315231,
+ "learning_rate": 0.0006,
+ "loss": 4.7006330490112305,
+ "step": 998
+ },
+ {
+ "epoch": 13.880349344978166,
+ "grad_norm": 0.016377726569771767,
+ "learning_rate": 0.0006,
+ "loss": 4.636074066162109,
+ "step": 999
+ },
+ {
+ "epoch": 13.894323144104803,
+ "grad_norm": 0.016660034656524658,
+ "learning_rate": 0.0006,
+ "loss": 4.742693901062012,
+ "step": 1000
+ },
+ {
+ "epoch": 13.908296943231441,
+ "grad_norm": 0.014902938157320023,
+ "learning_rate": 0.0006,
+ "loss": 4.769811630249023,
+ "step": 1001
+ },
+ {
+ "epoch": 13.922270742358078,
+ "grad_norm": 0.014883476309478283,
+ "learning_rate": 0.0006,
+ "loss": 4.731159210205078,
+ "step": 1002
+ },
+ {
+ "epoch": 13.936244541484717,
+ "grad_norm": 0.014494217932224274,
+ "learning_rate": 0.0006,
+ "loss": 4.659485816955566,
+ "step": 1003
+ },
+ {
+ "epoch": 13.950218340611354,
+ "grad_norm": 0.015426991507411003,
+ "learning_rate": 0.0006,
+ "loss": 4.755229949951172,
+ "step": 1004
+ },
+ {
+ "epoch": 13.96419213973799,
+ "grad_norm": 0.016648586839437485,
+ "learning_rate": 0.0006,
+ "loss": 4.74111270904541,
+ "step": 1005
+ },
+ {
+ "epoch": 13.97816593886463,
+ "grad_norm": 0.016980541869997978,
+ "learning_rate": 0.0006,
+ "loss": 4.74072265625,
+ "step": 1006
+ },
+ {
+ "epoch": 13.992139737991266,
+ "grad_norm": 0.018070152029395103,
+ "learning_rate": 0.0006,
+ "loss": 4.732975006103516,
+ "step": 1007
+ },
+ {
+ "epoch": 14.0,
+ "grad_norm": 0.019910665228962898,
+ "learning_rate": 0.0006,
+ "loss": 4.634697914123535,
+ "step": 1008
+ },
+ {
+ "epoch": 14.0,
+ "eval_loss": 4.850640296936035,
+ "eval_runtime": 57.3039,
+ "eval_samples_per_second": 42.615,
+ "eval_steps_per_second": 1.344,
+ "step": 1008
+ },
+ {
+ "epoch": 14.013973799126637,
+ "grad_norm": 0.01980053074657917,
+ "learning_rate": 0.0006,
+ "loss": 4.628744602203369,
+ "step": 1009
+ },
+ {
+ "epoch": 14.027947598253276,
+ "grad_norm": 0.02497623674571514,
+ "learning_rate": 0.0006,
+ "loss": 4.75930118560791,
+ "step": 1010
+ },
+ {
+ "epoch": 14.041921397379912,
+ "grad_norm": 0.02794097363948822,
+ "learning_rate": 0.0006,
+ "loss": 4.691340446472168,
+ "step": 1011
+ },
+ {
+ "epoch": 14.055895196506551,
+ "grad_norm": 0.023725535720586777,
+ "learning_rate": 0.0006,
+ "loss": 4.733664512634277,
+ "step": 1012
+ },
+ {
+ "epoch": 14.069868995633188,
+ "grad_norm": 0.022966474294662476,
+ "learning_rate": 0.0006,
+ "loss": 4.664808750152588,
+ "step": 1013
+ },
+ {
+ "epoch": 14.083842794759825,
+ "grad_norm": 0.02530563250184059,
+ "learning_rate": 0.0006,
+ "loss": 4.8047285079956055,
+ "step": 1014
+ },
+ {
+ "epoch": 14.097816593886463,
+ "grad_norm": 0.02491082064807415,
+ "learning_rate": 0.0006,
+ "loss": 4.667956352233887,
+ "step": 1015
+ },
+ {
+ "epoch": 14.1117903930131,
+ "grad_norm": 0.022110294550657272,
+ "learning_rate": 0.0006,
+ "loss": 4.666140556335449,
+ "step": 1016
+ },
+ {
+ "epoch": 14.125764192139737,
+ "grad_norm": 0.02065849117934704,
+ "learning_rate": 0.0006,
+ "loss": 4.581960678100586,
+ "step": 1017
+ },
+ {
+ "epoch": 14.139737991266376,
+ "grad_norm": 0.01941298507153988,
+ "learning_rate": 0.0006,
+ "loss": 4.614226341247559,
+ "step": 1018
+ },
+ {
+ "epoch": 14.153711790393013,
+ "grad_norm": 0.022441856563091278,
+ "learning_rate": 0.0006,
+ "loss": 4.638352394104004,
+ "step": 1019
+ },
+ {
+ "epoch": 14.167685589519651,
+ "grad_norm": 0.021519694477319717,
+ "learning_rate": 0.0006,
+ "loss": 4.700160980224609,
+ "step": 1020
+ },
+ {
+ "epoch": 14.181659388646288,
+ "grad_norm": 0.018956486135721207,
+ "learning_rate": 0.0006,
+ "loss": 4.649593353271484,
+ "step": 1021
+ },
+ {
+ "epoch": 14.195633187772925,
+ "grad_norm": 0.016681140288710594,
+ "learning_rate": 0.0006,
+ "loss": 4.603952407836914,
+ "step": 1022
+ },
+ {
+ "epoch": 14.209606986899564,
+ "grad_norm": 0.01788809522986412,
+ "learning_rate": 0.0006,
+ "loss": 4.669160842895508,
+ "step": 1023
+ },
+ {
+ "epoch": 14.2235807860262,
+ "grad_norm": 0.016047228127717972,
+ "learning_rate": 0.0006,
+ "loss": 4.6367950439453125,
+ "step": 1024
+ },
+ {
+ "epoch": 14.237554585152838,
+ "grad_norm": 0.01614239625632763,
+ "learning_rate": 0.0006,
+ "loss": 4.664529323577881,
+ "step": 1025
+ },
+ {
+ "epoch": 14.251528384279476,
+ "grad_norm": 0.016362549737095833,
+ "learning_rate": 0.0006,
+ "loss": 4.702373504638672,
+ "step": 1026
+ },
+ {
+ "epoch": 14.265502183406113,
+ "grad_norm": 0.015800409018993378,
+ "learning_rate": 0.0006,
+ "loss": 4.7594428062438965,
+ "step": 1027
+ },
+ {
+ "epoch": 14.279475982532752,
+ "grad_norm": 0.016730139032006264,
+ "learning_rate": 0.0006,
+ "loss": 4.677244186401367,
+ "step": 1028
+ },
+ {
+ "epoch": 14.293449781659389,
+ "grad_norm": 0.01661759801208973,
+ "learning_rate": 0.0006,
+ "loss": 4.633679389953613,
+ "step": 1029
+ },
+ {
+ "epoch": 14.307423580786025,
+ "grad_norm": 0.017673347145318985,
+ "learning_rate": 0.0006,
+ "loss": 4.649107456207275,
+ "step": 1030
+ },
+ {
+ "epoch": 14.321397379912664,
+ "grad_norm": 0.019243352115154266,
+ "learning_rate": 0.0006,
+ "loss": 4.611257553100586,
+ "step": 1031
+ },
+ {
+ "epoch": 14.335371179039301,
+ "grad_norm": 0.019817164167761803,
+ "learning_rate": 0.0006,
+ "loss": 4.588086128234863,
+ "step": 1032
+ },
+ {
+ "epoch": 14.34934497816594,
+ "grad_norm": 0.022292211651802063,
+ "learning_rate": 0.0006,
+ "loss": 4.783066749572754,
+ "step": 1033
+ },
+ {
+ "epoch": 14.363318777292577,
+ "grad_norm": 0.02140096202492714,
+ "learning_rate": 0.0006,
+ "loss": 4.604718208312988,
+ "step": 1034
+ },
+ {
+ "epoch": 14.377292576419213,
+ "grad_norm": 0.0187817495316267,
+ "learning_rate": 0.0006,
+ "loss": 4.721773147583008,
+ "step": 1035
+ },
+ {
+ "epoch": 14.391266375545852,
+ "grad_norm": 0.019879210740327835,
+ "learning_rate": 0.0006,
+ "loss": 4.789286136627197,
+ "step": 1036
+ },
+ {
+ "epoch": 14.405240174672489,
+ "grad_norm": 0.01885785162448883,
+ "learning_rate": 0.0006,
+ "loss": 4.669743537902832,
+ "step": 1037
+ },
+ {
+ "epoch": 14.419213973799126,
+ "grad_norm": 0.018324270844459534,
+ "learning_rate": 0.0006,
+ "loss": 4.65901517868042,
+ "step": 1038
+ },
+ {
+ "epoch": 14.433187772925764,
+ "grad_norm": 0.017664821818470955,
+ "learning_rate": 0.0006,
+ "loss": 4.667839527130127,
+ "step": 1039
+ },
+ {
+ "epoch": 14.447161572052401,
+ "grad_norm": 0.015702905133366585,
+ "learning_rate": 0.0006,
+ "loss": 4.64700984954834,
+ "step": 1040
+ },
+ {
+ "epoch": 14.46113537117904,
+ "grad_norm": 0.016674194484949112,
+ "learning_rate": 0.0006,
+ "loss": 4.694093227386475,
+ "step": 1041
+ },
+ {
+ "epoch": 14.475109170305677,
+ "grad_norm": 0.016869477927684784,
+ "learning_rate": 0.0006,
+ "loss": 4.619544982910156,
+ "step": 1042
+ },
+ {
+ "epoch": 14.489082969432314,
+ "grad_norm": 0.016557540744543076,
+ "learning_rate": 0.0006,
+ "loss": 4.568667411804199,
+ "step": 1043
+ },
+ {
+ "epoch": 14.503056768558952,
+ "grad_norm": 0.014500945806503296,
+ "learning_rate": 0.0006,
+ "loss": 4.561741828918457,
+ "step": 1044
+ },
+ {
+ "epoch": 14.51703056768559,
+ "grad_norm": 0.014228739775717258,
+ "learning_rate": 0.0006,
+ "loss": 4.654277324676514,
+ "step": 1045
+ },
+ {
+ "epoch": 14.531004366812226,
+ "grad_norm": 0.013894238509237766,
+ "learning_rate": 0.0006,
+ "loss": 4.5867719650268555,
+ "step": 1046
+ },
+ {
+ "epoch": 14.544978165938865,
+ "grad_norm": 0.01550068985670805,
+ "learning_rate": 0.0006,
+ "loss": 4.61168098449707,
+ "step": 1047
+ },
+ {
+ "epoch": 14.558951965065502,
+ "grad_norm": 0.015940772369503975,
+ "learning_rate": 0.0006,
+ "loss": 4.691247940063477,
+ "step": 1048
+ },
+ {
+ "epoch": 14.57292576419214,
+ "grad_norm": 0.01714535616338253,
+ "learning_rate": 0.0006,
+ "loss": 4.61183500289917,
+ "step": 1049
+ },
+ {
+ "epoch": 14.586899563318777,
+ "grad_norm": 0.017925038933753967,
+ "learning_rate": 0.0006,
+ "loss": 4.719022750854492,
+ "step": 1050
+ },
+ {
+ "epoch": 14.600873362445414,
+ "grad_norm": 0.018141552805900574,
+ "learning_rate": 0.0006,
+ "loss": 4.760281562805176,
+ "step": 1051
+ },
+ {
+ "epoch": 14.614847161572053,
+ "grad_norm": 0.017131328582763672,
+ "learning_rate": 0.0006,
+ "loss": 4.709680080413818,
+ "step": 1052
+ },
+ {
+ "epoch": 14.62882096069869,
+ "grad_norm": 0.018730048090219498,
+ "learning_rate": 0.0006,
+ "loss": 4.5994720458984375,
+ "step": 1053
+ },
+ {
+ "epoch": 14.642794759825328,
+ "grad_norm": 0.018894298002123833,
+ "learning_rate": 0.0006,
+ "loss": 4.669070720672607,
+ "step": 1054
+ },
+ {
+ "epoch": 14.656768558951965,
+ "grad_norm": 0.017643198370933533,
+ "learning_rate": 0.0006,
+ "loss": 4.641790390014648,
+ "step": 1055
+ },
+ {
+ "epoch": 14.670742358078602,
+ "grad_norm": 0.017647244036197662,
+ "learning_rate": 0.0006,
+ "loss": 4.653899192810059,
+ "step": 1056
+ },
+ {
+ "epoch": 14.68471615720524,
+ "grad_norm": 0.017304198816418648,
+ "learning_rate": 0.0006,
+ "loss": 4.7487473487854,
+ "step": 1057
+ },
+ {
+ "epoch": 14.698689956331878,
+ "grad_norm": 0.01638447679579258,
+ "learning_rate": 0.0006,
+ "loss": 4.71945858001709,
+ "step": 1058
+ },
+ {
+ "epoch": 14.712663755458514,
+ "grad_norm": 0.015892071649432182,
+ "learning_rate": 0.0006,
+ "loss": 4.744903564453125,
+ "step": 1059
+ },
+ {
+ "epoch": 14.726637554585153,
+ "grad_norm": 0.017420614138245583,
+ "learning_rate": 0.0006,
+ "loss": 4.781563758850098,
+ "step": 1060
+ },
+ {
+ "epoch": 14.74061135371179,
+ "grad_norm": 0.016498520970344543,
+ "learning_rate": 0.0006,
+ "loss": 4.622779846191406,
+ "step": 1061
+ },
+ {
+ "epoch": 14.754585152838429,
+ "grad_norm": 0.017482127994298935,
+ "learning_rate": 0.0006,
+ "loss": 4.62825870513916,
+ "step": 1062
+ },
+ {
+ "epoch": 14.768558951965066,
+ "grad_norm": 0.019018379971385002,
+ "learning_rate": 0.0006,
+ "loss": 4.693194389343262,
+ "step": 1063
+ },
+ {
+ "epoch": 14.782532751091702,
+ "grad_norm": 0.01993914321064949,
+ "learning_rate": 0.0006,
+ "loss": 4.742903709411621,
+ "step": 1064
+ },
+ {
+ "epoch": 14.796506550218341,
+ "grad_norm": 0.02301335707306862,
+ "learning_rate": 0.0006,
+ "loss": 4.613251686096191,
+ "step": 1065
+ },
+ {
+ "epoch": 14.810480349344978,
+ "grad_norm": 0.024088222533464432,
+ "learning_rate": 0.0006,
+ "loss": 4.594305992126465,
+ "step": 1066
+ },
+ {
+ "epoch": 14.824454148471617,
+ "grad_norm": 0.02493320405483246,
+ "learning_rate": 0.0006,
+ "loss": 4.700588226318359,
+ "step": 1067
+ },
+ {
+ "epoch": 14.838427947598253,
+ "grad_norm": 0.02501937560737133,
+ "learning_rate": 0.0006,
+ "loss": 4.601742744445801,
+ "step": 1068
+ },
+ {
+ "epoch": 14.85240174672489,
+ "grad_norm": 0.023490751162171364,
+ "learning_rate": 0.0006,
+ "loss": 4.589211463928223,
+ "step": 1069
+ },
+ {
+ "epoch": 14.866375545851529,
+ "grad_norm": 0.019695665687322617,
+ "learning_rate": 0.0006,
+ "loss": 4.635081768035889,
+ "step": 1070
+ },
+ {
+ "epoch": 14.880349344978166,
+ "grad_norm": 0.019276980310678482,
+ "learning_rate": 0.0006,
+ "loss": 4.737002372741699,
+ "step": 1071
+ },
+ {
+ "epoch": 14.894323144104803,
+ "grad_norm": 0.022430511191487312,
+ "learning_rate": 0.0006,
+ "loss": 4.808295726776123,
+ "step": 1072
+ },
+ {
+ "epoch": 14.908296943231441,
+ "grad_norm": 0.021543916314840317,
+ "learning_rate": 0.0006,
+ "loss": 4.559475421905518,
+ "step": 1073
+ },
+ {
+ "epoch": 14.922270742358078,
+ "grad_norm": 0.019885241985321045,
+ "learning_rate": 0.0006,
+ "loss": 4.647948741912842,
+ "step": 1074
+ },
+ {
+ "epoch": 14.936244541484717,
+ "grad_norm": 0.02115175500512123,
+ "learning_rate": 0.0006,
+ "loss": 4.680505752563477,
+ "step": 1075
+ },
+ {
+ "epoch": 14.950218340611354,
+ "grad_norm": 0.02156151458621025,
+ "learning_rate": 0.0006,
+ "loss": 4.673477649688721,
+ "step": 1076
+ },
+ {
+ "epoch": 14.96419213973799,
+ "grad_norm": 0.02184601128101349,
+ "learning_rate": 0.0006,
+ "loss": 4.644402027130127,
+ "step": 1077
+ },
+ {
+ "epoch": 14.97816593886463,
+ "grad_norm": 0.0211983360350132,
+ "learning_rate": 0.0006,
+ "loss": 4.821558475494385,
+ "step": 1078
+ },
+ {
+ "epoch": 14.992139737991266,
+ "grad_norm": 0.019636567682027817,
+ "learning_rate": 0.0006,
+ "loss": 4.578868865966797,
+ "step": 1079
+ },
+ {
+ "epoch": 15.0,
+ "grad_norm": 0.019997350871562958,
+ "learning_rate": 0.0006,
+ "loss": 4.691307544708252,
+ "step": 1080
+ },
+ {
+ "epoch": 15.0,
+ "eval_loss": 4.872478008270264,
+ "eval_runtime": 56.8513,
+ "eval_samples_per_second": 42.954,
+ "eval_steps_per_second": 1.354,
+ "step": 1080
+ },
+ {
+ "epoch": 15.013973799126637,
+ "grad_norm": 0.018830662593245506,
+ "learning_rate": 0.0006,
+ "loss": 4.761779308319092,
+ "step": 1081
+ },
+ {
+ "epoch": 15.027947598253276,
+ "grad_norm": 0.016876481473445892,
+ "learning_rate": 0.0006,
+ "loss": 4.61148738861084,
+ "step": 1082
+ },
+ {
+ "epoch": 15.041921397379912,
+ "grad_norm": 0.015550863929092884,
+ "learning_rate": 0.0006,
+ "loss": 4.676124095916748,
+ "step": 1083
+ },
+ {
+ "epoch": 15.055895196506551,
+ "grad_norm": 0.017178896814584732,
+ "learning_rate": 0.0006,
+ "loss": 4.588935852050781,
+ "step": 1084
+ },
+ {
+ "epoch": 15.069868995633188,
+ "grad_norm": 0.015465447679162025,
+ "learning_rate": 0.0006,
+ "loss": 4.605499267578125,
+ "step": 1085
+ },
+ {
+ "epoch": 15.083842794759825,
+ "grad_norm": 0.01606195978820324,
+ "learning_rate": 0.0006,
+ "loss": 4.723373889923096,
+ "step": 1086
+ },
+ {
+ "epoch": 15.097816593886463,
+ "grad_norm": 0.016521936282515526,
+ "learning_rate": 0.0006,
+ "loss": 4.646608352661133,
+ "step": 1087
+ },
+ {
+ "epoch": 15.1117903930131,
+ "grad_norm": 0.01589205674827099,
+ "learning_rate": 0.0006,
+ "loss": 4.498138427734375,
+ "step": 1088
+ },
+ {
+ "epoch": 15.125764192139737,
+ "grad_norm": 0.01524856686592102,
+ "learning_rate": 0.0006,
+ "loss": 4.594562530517578,
+ "step": 1089
+ },
+ {
+ "epoch": 15.139737991266376,
+ "grad_norm": 0.014714641496539116,
+ "learning_rate": 0.0006,
+ "loss": 4.630744934082031,
+ "step": 1090
+ },
+ {
+ "epoch": 15.153711790393013,
+ "grad_norm": 0.015333331190049648,
+ "learning_rate": 0.0006,
+ "loss": 4.682388782501221,
+ "step": 1091
+ },
+ {
+ "epoch": 15.167685589519651,
+ "grad_norm": 0.01550073828548193,
+ "learning_rate": 0.0006,
+ "loss": 4.692652702331543,
+ "step": 1092
+ },
+ {
+ "epoch": 15.181659388646288,
+ "grad_norm": 0.014783318154513836,
+ "learning_rate": 0.0006,
+ "loss": 4.6223835945129395,
+ "step": 1093
+ },
+ {
+ "epoch": 15.195633187772925,
+ "grad_norm": 0.014372751116752625,
+ "learning_rate": 0.0006,
+ "loss": 4.52687406539917,
+ "step": 1094
+ },
+ {
+ "epoch": 15.209606986899564,
+ "grad_norm": 0.014779130928218365,
+ "learning_rate": 0.0006,
+ "loss": 4.637009620666504,
+ "step": 1095
+ },
+ {
+ "epoch": 15.2235807860262,
+ "grad_norm": 0.015697073191404343,
+ "learning_rate": 0.0006,
+ "loss": 4.682981014251709,
+ "step": 1096
+ },
+ {
+ "epoch": 15.237554585152838,
+ "grad_norm": 0.016523664817214012,
+ "learning_rate": 0.0006,
+ "loss": 4.705722808837891,
+ "step": 1097
+ },
+ {
+ "epoch": 15.251528384279476,
+ "grad_norm": 0.017633996903896332,
+ "learning_rate": 0.0006,
+ "loss": 4.574686050415039,
+ "step": 1098
+ },
+ {
+ "epoch": 15.265502183406113,
+ "grad_norm": 0.019955076277256012,
+ "learning_rate": 0.0006,
+ "loss": 4.634450912475586,
+ "step": 1099
+ },
+ {
+ "epoch": 15.279475982532752,
+ "grad_norm": 0.02132618986070156,
+ "learning_rate": 0.0006,
+ "loss": 4.56110143661499,
+ "step": 1100
+ },
+ {
+ "epoch": 15.293449781659389,
+ "grad_norm": 0.0207593385130167,
+ "learning_rate": 0.0006,
+ "loss": 4.629396438598633,
+ "step": 1101
+ },
+ {
+ "epoch": 15.307423580786025,
+ "grad_norm": 0.017882540822029114,
+ "learning_rate": 0.0006,
+ "loss": 4.730963230133057,
+ "step": 1102
+ },
+ {
+ "epoch": 15.321397379912664,
+ "grad_norm": 0.01917876861989498,
+ "learning_rate": 0.0006,
+ "loss": 4.64585542678833,
+ "step": 1103
+ },
+ {
+ "epoch": 15.335371179039301,
+ "grad_norm": 0.01943974196910858,
+ "learning_rate": 0.0006,
+ "loss": 4.656181335449219,
+ "step": 1104
+ },
+ {
+ "epoch": 15.34934497816594,
+ "grad_norm": 0.019904715940356255,
+ "learning_rate": 0.0006,
+ "loss": 4.619329452514648,
+ "step": 1105
+ },
+ {
+ "epoch": 15.363318777292577,
+ "grad_norm": 0.019174346700310707,
+ "learning_rate": 0.0006,
+ "loss": 4.647892475128174,
+ "step": 1106
+ },
+ {
+ "epoch": 15.377292576419213,
+ "grad_norm": 0.018244758248329163,
+ "learning_rate": 0.0006,
+ "loss": 4.56166934967041,
+ "step": 1107
+ },
+ {
+ "epoch": 15.391266375545852,
+ "grad_norm": 0.018914205953478813,
+ "learning_rate": 0.0006,
+ "loss": 4.648063659667969,
+ "step": 1108
+ },
+ {
+ "epoch": 15.405240174672489,
+ "grad_norm": 0.01961071416735649,
+ "learning_rate": 0.0006,
+ "loss": 4.665525436401367,
+ "step": 1109
+ },
+ {
+ "epoch": 15.419213973799126,
+ "grad_norm": 0.01841311901807785,
+ "learning_rate": 0.0006,
+ "loss": 4.611342430114746,
+ "step": 1110
+ },
+ {
+ "epoch": 15.433187772925764,
+ "grad_norm": 0.01863427273929119,
+ "learning_rate": 0.0006,
+ "loss": 4.664256572723389,
+ "step": 1111
+ },
+ {
+ "epoch": 15.447161572052401,
+ "grad_norm": 0.020506877452135086,
+ "learning_rate": 0.0006,
+ "loss": 4.631237506866455,
+ "step": 1112
+ },
+ {
+ "epoch": 15.46113537117904,
+ "grad_norm": 0.02223833277821541,
+ "learning_rate": 0.0006,
+ "loss": 4.668654441833496,
+ "step": 1113
+ },
+ {
+ "epoch": 15.475109170305677,
+ "grad_norm": 0.023336702957749367,
+ "learning_rate": 0.0006,
+ "loss": 4.562833309173584,
+ "step": 1114
+ },
+ {
+ "epoch": 15.489082969432314,
+ "grad_norm": 0.023052413016557693,
+ "learning_rate": 0.0006,
+ "loss": 4.621532917022705,
+ "step": 1115
+ },
+ {
+ "epoch": 15.503056768558952,
+ "grad_norm": 0.023420924320816994,
+ "learning_rate": 0.0006,
+ "loss": 4.665465354919434,
+ "step": 1116
+ },
+ {
+ "epoch": 15.51703056768559,
+ "grad_norm": 0.024071281775832176,
+ "learning_rate": 0.0006,
+ "loss": 4.595980644226074,
+ "step": 1117
+ },
+ {
+ "epoch": 15.531004366812226,
+ "grad_norm": 0.023406682536005974,
+ "learning_rate": 0.0006,
+ "loss": 4.642635345458984,
+ "step": 1118
+ },
+ {
+ "epoch": 15.544978165938865,
+ "grad_norm": 0.02138647809624672,
+ "learning_rate": 0.0006,
+ "loss": 4.704513072967529,
+ "step": 1119
+ },
+ {
+ "epoch": 15.558951965065502,
+ "grad_norm": 0.0221384447067976,
+ "learning_rate": 0.0006,
+ "loss": 4.567503452301025,
+ "step": 1120
+ },
+ {
+ "epoch": 15.57292576419214,
+ "grad_norm": 0.020879851654171944,
+ "learning_rate": 0.0006,
+ "loss": 4.664057731628418,
+ "step": 1121
+ },
+ {
+ "epoch": 15.586899563318777,
+ "grad_norm": 0.021478967741131783,
+ "learning_rate": 0.0006,
+ "loss": 4.539627552032471,
+ "step": 1122
+ },
+ {
+ "epoch": 15.600873362445414,
+ "grad_norm": 0.022748656570911407,
+ "learning_rate": 0.0006,
+ "loss": 4.5654497146606445,
+ "step": 1123
+ },
+ {
+ "epoch": 15.614847161572053,
+ "grad_norm": 0.021870248019695282,
+ "learning_rate": 0.0006,
+ "loss": 4.56899356842041,
+ "step": 1124
+ },
+ {
+ "epoch": 15.62882096069869,
+ "grad_norm": 0.021711355075240135,
+ "learning_rate": 0.0006,
+ "loss": 4.657808303833008,
+ "step": 1125
+ },
+ {
+ "epoch": 15.642794759825328,
+ "grad_norm": 0.022810909897089005,
+ "learning_rate": 0.0006,
+ "loss": 4.6526689529418945,
+ "step": 1126
+ },
+ {
+ "epoch": 15.656768558951965,
+ "grad_norm": 0.022561343386769295,
+ "learning_rate": 0.0006,
+ "loss": 4.534887790679932,
+ "step": 1127
+ },
+ {
+ "epoch": 15.670742358078602,
+ "grad_norm": 0.019909802824258804,
+ "learning_rate": 0.0006,
+ "loss": 4.600790023803711,
+ "step": 1128
+ },
+ {
+ "epoch": 15.68471615720524,
+ "grad_norm": 0.020092325285077095,
+ "learning_rate": 0.0006,
+ "loss": 4.773830413818359,
+ "step": 1129
+ },
+ {
+ "epoch": 15.698689956331878,
+ "grad_norm": 0.018834445625543594,
+ "learning_rate": 0.0006,
+ "loss": 4.628868103027344,
+ "step": 1130
+ },
+ {
+ "epoch": 15.712663755458514,
+ "grad_norm": 0.017552798613905907,
+ "learning_rate": 0.0006,
+ "loss": 4.647500038146973,
+ "step": 1131
+ },
+ {
+ "epoch": 15.726637554585153,
+ "grad_norm": 0.01751251146197319,
+ "learning_rate": 0.0006,
+ "loss": 4.590545654296875,
+ "step": 1132
+ },
+ {
+ "epoch": 15.74061135371179,
+ "grad_norm": 0.01699099875986576,
+ "learning_rate": 0.0006,
+ "loss": 4.509699821472168,
+ "step": 1133
+ },
+ {
+ "epoch": 15.754585152838429,
+ "grad_norm": 0.015808649361133575,
+ "learning_rate": 0.0006,
+ "loss": 4.606992244720459,
+ "step": 1134
+ },
+ {
+ "epoch": 15.768558951965066,
+ "grad_norm": 0.01655443385243416,
+ "learning_rate": 0.0006,
+ "loss": 4.585219860076904,
+ "step": 1135
+ },
+ {
+ "epoch": 15.782532751091702,
+ "grad_norm": 0.01708163693547249,
+ "learning_rate": 0.0006,
+ "loss": 4.668147087097168,
+ "step": 1136
+ },
+ {
+ "epoch": 15.796506550218341,
+ "grad_norm": 0.017161913216114044,
+ "learning_rate": 0.0006,
+ "loss": 4.7170634269714355,
+ "step": 1137
+ },
+ {
+ "epoch": 15.810480349344978,
+ "grad_norm": 0.017262674868106842,
+ "learning_rate": 0.0006,
+ "loss": 4.572132110595703,
+ "step": 1138
+ },
+ {
+ "epoch": 15.824454148471617,
+ "grad_norm": 0.017713576555252075,
+ "learning_rate": 0.0006,
+ "loss": 4.702538967132568,
+ "step": 1139
+ },
+ {
+ "epoch": 15.838427947598253,
+ "grad_norm": 0.018730733543634415,
+ "learning_rate": 0.0006,
+ "loss": 4.629721641540527,
+ "step": 1140
+ },
+ {
+ "epoch": 15.85240174672489,
+ "grad_norm": 0.017960187047719955,
+ "learning_rate": 0.0006,
+ "loss": 4.661039352416992,
+ "step": 1141
+ },
+ {
+ "epoch": 15.866375545851529,
+ "grad_norm": 0.01750079356133938,
+ "learning_rate": 0.0006,
+ "loss": 4.503710746765137,
+ "step": 1142
+ },
+ {
+ "epoch": 15.880349344978166,
+ "grad_norm": 0.016630422323942184,
+ "learning_rate": 0.0006,
+ "loss": 4.6582136154174805,
+ "step": 1143
+ },
+ {
+ "epoch": 15.894323144104803,
+ "grad_norm": 0.015896832570433617,
+ "learning_rate": 0.0006,
+ "loss": 4.596693992614746,
+ "step": 1144
+ },
+ {
+ "epoch": 15.908296943231441,
+ "grad_norm": 0.01566407084465027,
+ "learning_rate": 0.0006,
+ "loss": 4.6337199211120605,
+ "step": 1145
+ },
+ {
+ "epoch": 15.922270742358078,
+ "grad_norm": 0.016823330894112587,
+ "learning_rate": 0.0006,
+ "loss": 4.686994552612305,
+ "step": 1146
+ },
+ {
+ "epoch": 15.936244541484717,
+ "grad_norm": 0.016208255663514137,
+ "learning_rate": 0.0006,
+ "loss": 4.6205549240112305,
+ "step": 1147
+ },
+ {
+ "epoch": 15.950218340611354,
+ "grad_norm": 0.0171707421541214,
+ "learning_rate": 0.0006,
+ "loss": 4.582968711853027,
+ "step": 1148
+ },
+ {
+ "epoch": 15.96419213973799,
+ "grad_norm": 0.017009446397423744,
+ "learning_rate": 0.0006,
+ "loss": 4.510274887084961,
+ "step": 1149
+ },
+ {
+ "epoch": 15.97816593886463,
+ "grad_norm": 0.01565811224281788,
+ "learning_rate": 0.0006,
+ "loss": 4.55834436416626,
+ "step": 1150
+ },
+ {
+ "epoch": 15.992139737991266,
+ "grad_norm": 0.015276037156581879,
+ "learning_rate": 0.0006,
+ "loss": 4.537870407104492,
+ "step": 1151
+ },
+ {
+ "epoch": 16.0,
+ "grad_norm": 0.017305882647633553,
+ "learning_rate": 0.0006,
+ "loss": 4.611305236816406,
+ "step": 1152
+ },
+ {
+ "epoch": 16.0,
+ "eval_loss": 4.775513172149658,
+ "eval_runtime": 56.7341,
+ "eval_samples_per_second": 43.043,
+ "eval_steps_per_second": 1.357,
+ "step": 1152
+ },
+ {
+ "epoch": 16.01397379912664,
+ "grad_norm": 0.016752436757087708,
+ "learning_rate": 0.0006,
+ "loss": 4.541312217712402,
+ "step": 1153
+ },
+ {
+ "epoch": 16.027947598253274,
+ "grad_norm": 0.01643134094774723,
+ "learning_rate": 0.0006,
+ "loss": 4.585402965545654,
+ "step": 1154
+ },
+ {
+ "epoch": 16.041921397379912,
+ "grad_norm": 0.01888720504939556,
+ "learning_rate": 0.0006,
+ "loss": 4.597825050354004,
+ "step": 1155
+ },
+ {
+ "epoch": 16.05589519650655,
+ "grad_norm": 0.022044632583856583,
+ "learning_rate": 0.0006,
+ "loss": 4.532772064208984,
+ "step": 1156
+ },
+ {
+ "epoch": 16.069868995633186,
+ "grad_norm": 0.022986987605690956,
+ "learning_rate": 0.0006,
+ "loss": 4.551351547241211,
+ "step": 1157
+ },
+ {
+ "epoch": 16.083842794759825,
+ "grad_norm": 0.020116791129112244,
+ "learning_rate": 0.0006,
+ "loss": 4.484149932861328,
+ "step": 1158
+ },
+ {
+ "epoch": 16.097816593886463,
+ "grad_norm": 0.021442247554659843,
+ "learning_rate": 0.0006,
+ "loss": 4.515316486358643,
+ "step": 1159
+ },
+ {
+ "epoch": 16.111790393013102,
+ "grad_norm": 0.020974772050976753,
+ "learning_rate": 0.0006,
+ "loss": 4.51606559753418,
+ "step": 1160
+ },
+ {
+ "epoch": 16.125764192139737,
+ "grad_norm": 0.021274864673614502,
+ "learning_rate": 0.0006,
+ "loss": 4.587329864501953,
+ "step": 1161
+ },
+ {
+ "epoch": 16.139737991266376,
+ "grad_norm": 0.022799784317612648,
+ "learning_rate": 0.0006,
+ "loss": 4.568149089813232,
+ "step": 1162
+ },
+ {
+ "epoch": 16.153711790393015,
+ "grad_norm": 0.021784933283925056,
+ "learning_rate": 0.0006,
+ "loss": 4.515336990356445,
+ "step": 1163
+ },
+ {
+ "epoch": 16.16768558951965,
+ "grad_norm": 0.024261271581053734,
+ "learning_rate": 0.0006,
+ "loss": 4.5127081871032715,
+ "step": 1164
+ },
+ {
+ "epoch": 16.18165938864629,
+ "grad_norm": 0.02248375490307808,
+ "learning_rate": 0.0006,
+ "loss": 4.754251003265381,
+ "step": 1165
+ },
+ {
+ "epoch": 16.195633187772927,
+ "grad_norm": 0.02136492356657982,
+ "learning_rate": 0.0006,
+ "loss": 4.58193826675415,
+ "step": 1166
+ },
+ {
+ "epoch": 16.209606986899562,
+ "grad_norm": 0.020727066323161125,
+ "learning_rate": 0.0006,
+ "loss": 4.564209938049316,
+ "step": 1167
+ },
+ {
+ "epoch": 16.2235807860262,
+ "grad_norm": 0.017789119854569435,
+ "learning_rate": 0.0006,
+ "loss": 4.698680877685547,
+ "step": 1168
+ },
+ {
+ "epoch": 16.23755458515284,
+ "grad_norm": 0.017603034153580666,
+ "learning_rate": 0.0006,
+ "loss": 4.501171588897705,
+ "step": 1169
+ },
+ {
+ "epoch": 16.251528384279474,
+ "grad_norm": 0.017345862463116646,
+ "learning_rate": 0.0006,
+ "loss": 4.593590259552002,
+ "step": 1170
+ },
+ {
+ "epoch": 16.265502183406113,
+ "grad_norm": 0.018047425895929337,
+ "learning_rate": 0.0006,
+ "loss": 4.597467422485352,
+ "step": 1171
+ },
+ {
+ "epoch": 16.27947598253275,
+ "grad_norm": 0.01947222836315632,
+ "learning_rate": 0.0006,
+ "loss": 4.5072221755981445,
+ "step": 1172
+ },
+ {
+ "epoch": 16.29344978165939,
+ "grad_norm": 0.019598500803112984,
+ "learning_rate": 0.0006,
+ "loss": 4.601351737976074,
+ "step": 1173
+ },
+ {
+ "epoch": 16.307423580786025,
+ "grad_norm": 0.020002514123916626,
+ "learning_rate": 0.0006,
+ "loss": 4.580431938171387,
+ "step": 1174
+ },
+ {
+ "epoch": 16.321397379912664,
+ "grad_norm": 0.0216530729085207,
+ "learning_rate": 0.0006,
+ "loss": 4.424681663513184,
+ "step": 1175
+ },
+ {
+ "epoch": 16.335371179039303,
+ "grad_norm": 0.01949235238134861,
+ "learning_rate": 0.0006,
+ "loss": 4.751883029937744,
+ "step": 1176
+ },
+ {
+ "epoch": 16.349344978165938,
+ "grad_norm": 0.017779016867280006,
+ "learning_rate": 0.0006,
+ "loss": 4.541490077972412,
+ "step": 1177
+ },
+ {
+ "epoch": 16.363318777292577,
+ "grad_norm": 0.01909121684730053,
+ "learning_rate": 0.0006,
+ "loss": 4.594141006469727,
+ "step": 1178
+ },
+ {
+ "epoch": 16.377292576419215,
+ "grad_norm": 0.019894849509000778,
+ "learning_rate": 0.0006,
+ "loss": 4.555070400238037,
+ "step": 1179
+ },
+ {
+ "epoch": 16.39126637554585,
+ "grad_norm": 0.01993032731115818,
+ "learning_rate": 0.0006,
+ "loss": 4.521401405334473,
+ "step": 1180
+ },
+ {
+ "epoch": 16.40524017467249,
+ "grad_norm": 0.0210788045078516,
+ "learning_rate": 0.0006,
+ "loss": 4.672842979431152,
+ "step": 1181
+ },
+ {
+ "epoch": 16.419213973799128,
+ "grad_norm": 0.02274598926305771,
+ "learning_rate": 0.0006,
+ "loss": 4.629504203796387,
+ "step": 1182
+ },
+ {
+ "epoch": 16.433187772925763,
+ "grad_norm": 0.020448120310902596,
+ "learning_rate": 0.0006,
+ "loss": 4.548209190368652,
+ "step": 1183
+ },
+ {
+ "epoch": 16.4471615720524,
+ "grad_norm": 0.017565643414855003,
+ "learning_rate": 0.0006,
+ "loss": 4.529054641723633,
+ "step": 1184
+ },
+ {
+ "epoch": 16.46113537117904,
+ "grad_norm": 0.019504237920045853,
+ "learning_rate": 0.0006,
+ "loss": 4.48182487487793,
+ "step": 1185
+ },
+ {
+ "epoch": 16.475109170305675,
+ "grad_norm": 0.016823742538690567,
+ "learning_rate": 0.0006,
+ "loss": 4.608696937561035,
+ "step": 1186
+ },
+ {
+ "epoch": 16.489082969432314,
+ "grad_norm": 0.01602894812822342,
+ "learning_rate": 0.0006,
+ "loss": 4.620062351226807,
+ "step": 1187
+ },
+ {
+ "epoch": 16.503056768558952,
+ "grad_norm": 0.016615642234683037,
+ "learning_rate": 0.0006,
+ "loss": 4.69378662109375,
+ "step": 1188
+ },
+ {
+ "epoch": 16.51703056768559,
+ "grad_norm": 0.018576730042696,
+ "learning_rate": 0.0006,
+ "loss": 4.537041664123535,
+ "step": 1189
+ },
+ {
+ "epoch": 16.531004366812226,
+ "grad_norm": 0.018611151725053787,
+ "learning_rate": 0.0006,
+ "loss": 4.496169090270996,
+ "step": 1190
+ },
+ {
+ "epoch": 16.544978165938865,
+ "grad_norm": 0.018622413277626038,
+ "learning_rate": 0.0006,
+ "loss": 4.557358264923096,
+ "step": 1191
+ },
+ {
+ "epoch": 16.558951965065503,
+ "grad_norm": 0.01823544315993786,
+ "learning_rate": 0.0006,
+ "loss": 4.730792999267578,
+ "step": 1192
+ },
+ {
+ "epoch": 16.57292576419214,
+ "grad_norm": 0.018796846270561218,
+ "learning_rate": 0.0006,
+ "loss": 4.781791687011719,
+ "step": 1193
+ },
+ {
+ "epoch": 16.586899563318777,
+ "grad_norm": 0.01728292554616928,
+ "learning_rate": 0.0006,
+ "loss": 4.515257835388184,
+ "step": 1194
+ },
+ {
+ "epoch": 16.600873362445416,
+ "grad_norm": 0.01699589006602764,
+ "learning_rate": 0.0006,
+ "loss": 4.5996904373168945,
+ "step": 1195
+ },
+ {
+ "epoch": 16.61484716157205,
+ "grad_norm": 0.015905970707535744,
+ "learning_rate": 0.0006,
+ "loss": 4.601301193237305,
+ "step": 1196
+ },
+ {
+ "epoch": 16.62882096069869,
+ "grad_norm": 0.01338571310043335,
+ "learning_rate": 0.0006,
+ "loss": 4.557708740234375,
+ "step": 1197
+ },
+ {
+ "epoch": 16.64279475982533,
+ "grad_norm": 0.015298855490982533,
+ "learning_rate": 0.0006,
+ "loss": 4.542287349700928,
+ "step": 1198
+ },
+ {
+ "epoch": 16.656768558951963,
+ "grad_norm": 0.014753367751836777,
+ "learning_rate": 0.0006,
+ "loss": 4.595787048339844,
+ "step": 1199
+ },
+ {
+ "epoch": 16.670742358078602,
+ "grad_norm": 0.015527483075857162,
+ "learning_rate": 0.0006,
+ "loss": 4.607529163360596,
+ "step": 1200
+ },
+ {
+ "epoch": 16.68471615720524,
+ "grad_norm": 0.015042162500321865,
+ "learning_rate": 0.0006,
+ "loss": 4.686626434326172,
+ "step": 1201
+ },
+ {
+ "epoch": 16.69868995633188,
+ "grad_norm": 0.0159393772482872,
+ "learning_rate": 0.0006,
+ "loss": 4.47656774520874,
+ "step": 1202
+ },
+ {
+ "epoch": 16.712663755458514,
+ "grad_norm": 0.01532787922769785,
+ "learning_rate": 0.0006,
+ "loss": 4.55678653717041,
+ "step": 1203
+ },
+ {
+ "epoch": 16.726637554585153,
+ "grad_norm": 0.016332164406776428,
+ "learning_rate": 0.0006,
+ "loss": 4.51441764831543,
+ "step": 1204
+ },
+ {
+ "epoch": 16.74061135371179,
+ "grad_norm": 0.018901726230978966,
+ "learning_rate": 0.0006,
+ "loss": 4.57227897644043,
+ "step": 1205
+ },
+ {
+ "epoch": 16.754585152838427,
+ "grad_norm": 0.018589433282613754,
+ "learning_rate": 0.0006,
+ "loss": 4.550070762634277,
+ "step": 1206
+ },
+ {
+ "epoch": 16.768558951965066,
+ "grad_norm": 0.01634034886956215,
+ "learning_rate": 0.0006,
+ "loss": 4.586910724639893,
+ "step": 1207
+ },
+ {
+ "epoch": 16.782532751091704,
+ "grad_norm": 0.017599433660507202,
+ "learning_rate": 0.0006,
+ "loss": 4.5891923904418945,
+ "step": 1208
+ },
+ {
+ "epoch": 16.79650655021834,
+ "grad_norm": 0.01878446340560913,
+ "learning_rate": 0.0006,
+ "loss": 4.57227087020874,
+ "step": 1209
+ },
+ {
+ "epoch": 16.810480349344978,
+ "grad_norm": 0.01803469844162464,
+ "learning_rate": 0.0006,
+ "loss": 4.583845138549805,
+ "step": 1210
+ },
+ {
+ "epoch": 16.824454148471617,
+ "grad_norm": 0.018505144864320755,
+ "learning_rate": 0.0006,
+ "loss": 4.561151504516602,
+ "step": 1211
+ },
+ {
+ "epoch": 16.83842794759825,
+ "grad_norm": 0.01920081488788128,
+ "learning_rate": 0.0006,
+ "loss": 4.506834030151367,
+ "step": 1212
+ },
+ {
+ "epoch": 16.85240174672489,
+ "grad_norm": 0.0177712831646204,
+ "learning_rate": 0.0006,
+ "loss": 4.557195663452148,
+ "step": 1213
+ },
+ {
+ "epoch": 16.86637554585153,
+ "grad_norm": 0.019113918766379356,
+ "learning_rate": 0.0006,
+ "loss": 4.651227951049805,
+ "step": 1214
+ },
+ {
+ "epoch": 16.880349344978168,
+ "grad_norm": 0.02012539468705654,
+ "learning_rate": 0.0006,
+ "loss": 4.4952239990234375,
+ "step": 1215
+ },
+ {
+ "epoch": 16.894323144104803,
+ "grad_norm": 0.019993646070361137,
+ "learning_rate": 0.0006,
+ "loss": 4.517433166503906,
+ "step": 1216
+ },
+ {
+ "epoch": 16.90829694323144,
+ "grad_norm": 0.020041441544890404,
+ "learning_rate": 0.0006,
+ "loss": 4.489407539367676,
+ "step": 1217
+ },
+ {
+ "epoch": 16.92227074235808,
+ "grad_norm": 0.0183611661195755,
+ "learning_rate": 0.0006,
+ "loss": 4.5435638427734375,
+ "step": 1218
+ },
+ {
+ "epoch": 16.936244541484715,
+ "grad_norm": 0.019138429313898087,
+ "learning_rate": 0.0006,
+ "loss": 4.545949459075928,
+ "step": 1219
+ },
+ {
+ "epoch": 16.950218340611354,
+ "grad_norm": 0.018554266542196274,
+ "learning_rate": 0.0006,
+ "loss": 4.579620361328125,
+ "step": 1220
+ },
+ {
+ "epoch": 16.964192139737992,
+ "grad_norm": 0.0193545650690794,
+ "learning_rate": 0.0006,
+ "loss": 4.588255405426025,
+ "step": 1221
+ },
+ {
+ "epoch": 16.978165938864628,
+ "grad_norm": 0.019389165565371513,
+ "learning_rate": 0.0006,
+ "loss": 4.520698547363281,
+ "step": 1222
+ },
+ {
+ "epoch": 16.992139737991266,
+ "grad_norm": 0.016601379960775375,
+ "learning_rate": 0.0006,
+ "loss": 4.4809064865112305,
+ "step": 1223
+ },
+ {
+ "epoch": 17.0,
+ "grad_norm": 0.01771303080022335,
+ "learning_rate": 0.0006,
+ "loss": 4.721122741699219,
+ "step": 1224
+ },
+ {
+ "epoch": 17.0,
+ "eval_loss": 4.7903666496276855,
+ "eval_runtime": 56.9826,
+ "eval_samples_per_second": 42.855,
+ "eval_steps_per_second": 1.351,
+ "step": 1224
+ },
+ {
+ "epoch": 17.01397379912664,
+ "grad_norm": 0.017720289528369904,
+ "learning_rate": 0.0006,
+ "loss": 4.523834705352783,
+ "step": 1225
+ },
+ {
+ "epoch": 17.027947598253274,
+ "grad_norm": 0.020137161016464233,
+ "learning_rate": 0.0006,
+ "loss": 4.574061870574951,
+ "step": 1226
+ },
+ {
+ "epoch": 17.041921397379912,
+ "grad_norm": 0.02002314105629921,
+ "learning_rate": 0.0006,
+ "loss": 4.512407302856445,
+ "step": 1227
+ },
+ {
+ "epoch": 17.05589519650655,
+ "grad_norm": 0.019628183916211128,
+ "learning_rate": 0.0006,
+ "loss": 4.502533435821533,
+ "step": 1228
+ },
+ {
+ "epoch": 17.069868995633186,
+ "grad_norm": 0.019808458164334297,
+ "learning_rate": 0.0006,
+ "loss": 4.463156700134277,
+ "step": 1229
+ },
+ {
+ "epoch": 17.083842794759825,
+ "grad_norm": 0.021250786259770393,
+ "learning_rate": 0.0006,
+ "loss": 4.622682094573975,
+ "step": 1230
+ },
+ {
+ "epoch": 17.097816593886463,
+ "grad_norm": 0.022031694650650024,
+ "learning_rate": 0.0006,
+ "loss": 4.451882839202881,
+ "step": 1231
+ },
+ {
+ "epoch": 17.111790393013102,
+ "grad_norm": 0.01898987777531147,
+ "learning_rate": 0.0006,
+ "loss": 4.529331684112549,
+ "step": 1232
+ },
+ {
+ "epoch": 17.125764192139737,
+ "grad_norm": 0.018749834969639778,
+ "learning_rate": 0.0006,
+ "loss": 4.435815334320068,
+ "step": 1233
+ },
+ {
+ "epoch": 17.139737991266376,
+ "grad_norm": 0.01896674931049347,
+ "learning_rate": 0.0006,
+ "loss": 4.484029293060303,
+ "step": 1234
+ },
+ {
+ "epoch": 17.153711790393015,
+ "grad_norm": 0.018286997452378273,
+ "learning_rate": 0.0006,
+ "loss": 4.523510932922363,
+ "step": 1235
+ },
+ {
+ "epoch": 17.16768558951965,
+ "grad_norm": 0.017273059114813805,
+ "learning_rate": 0.0006,
+ "loss": 4.451767444610596,
+ "step": 1236
+ },
+ {
+ "epoch": 17.18165938864629,
+ "grad_norm": 0.016499150544404984,
+ "learning_rate": 0.0006,
+ "loss": 4.5568647384643555,
+ "step": 1237
+ },
+ {
+ "epoch": 17.195633187772927,
+ "grad_norm": 0.015934469178318977,
+ "learning_rate": 0.0006,
+ "loss": 4.4639997482299805,
+ "step": 1238
+ },
+ {
+ "epoch": 17.209606986899562,
+ "grad_norm": 0.01714209094643593,
+ "learning_rate": 0.0006,
+ "loss": 4.547614097595215,
+ "step": 1239
+ },
+ {
+ "epoch": 17.2235807860262,
+ "grad_norm": 0.019167495891451836,
+ "learning_rate": 0.0006,
+ "loss": 4.571097373962402,
+ "step": 1240
+ },
+ {
+ "epoch": 17.23755458515284,
+ "grad_norm": 0.018276996910572052,
+ "learning_rate": 0.0006,
+ "loss": 4.489764213562012,
+ "step": 1241
+ },
+ {
+ "epoch": 17.251528384279474,
+ "grad_norm": 0.021296866238117218,
+ "learning_rate": 0.0006,
+ "loss": 4.526913642883301,
+ "step": 1242
+ },
+ {
+ "epoch": 17.265502183406113,
+ "grad_norm": 0.022606629878282547,
+ "learning_rate": 0.0006,
+ "loss": 4.521624565124512,
+ "step": 1243
+ },
+ {
+ "epoch": 17.27947598253275,
+ "grad_norm": 0.021204711869359016,
+ "learning_rate": 0.0006,
+ "loss": 4.500864028930664,
+ "step": 1244
+ },
+ {
+ "epoch": 17.29344978165939,
+ "grad_norm": 0.01988849975168705,
+ "learning_rate": 0.0006,
+ "loss": 4.45162296295166,
+ "step": 1245
+ },
+ {
+ "epoch": 17.307423580786025,
+ "grad_norm": 0.020268384367227554,
+ "learning_rate": 0.0006,
+ "loss": 4.536198616027832,
+ "step": 1246
+ },
+ {
+ "epoch": 17.321397379912664,
+ "grad_norm": 0.020185677334666252,
+ "learning_rate": 0.0006,
+ "loss": 4.440262794494629,
+ "step": 1247
+ },
+ {
+ "epoch": 17.335371179039303,
+ "grad_norm": 0.020106853917241096,
+ "learning_rate": 0.0006,
+ "loss": 4.475907325744629,
+ "step": 1248
+ },
+ {
+ "epoch": 17.349344978165938,
+ "grad_norm": 0.018883297219872475,
+ "learning_rate": 0.0006,
+ "loss": 4.640661716461182,
+ "step": 1249
+ },
+ {
+ "epoch": 17.363318777292577,
+ "grad_norm": 0.01852620206773281,
+ "learning_rate": 0.0006,
+ "loss": 4.578243255615234,
+ "step": 1250
+ },
+ {
+ "epoch": 17.377292576419215,
+ "grad_norm": 0.01886451430618763,
+ "learning_rate": 0.0006,
+ "loss": 4.546026229858398,
+ "step": 1251
+ },
+ {
+ "epoch": 17.39126637554585,
+ "grad_norm": 0.01777452416718006,
+ "learning_rate": 0.0006,
+ "loss": 4.521542549133301,
+ "step": 1252
+ },
+ {
+ "epoch": 17.40524017467249,
+ "grad_norm": 0.016626447439193726,
+ "learning_rate": 0.0006,
+ "loss": 4.52629280090332,
+ "step": 1253
+ },
+ {
+ "epoch": 17.419213973799128,
+ "grad_norm": 0.01599477231502533,
+ "learning_rate": 0.0006,
+ "loss": 4.4796037673950195,
+ "step": 1254
+ },
+ {
+ "epoch": 17.433187772925763,
+ "grad_norm": 0.017021458595991135,
+ "learning_rate": 0.0006,
+ "loss": 4.468813419342041,
+ "step": 1255
+ },
+ {
+ "epoch": 17.4471615720524,
+ "grad_norm": 0.016925333067774773,
+ "learning_rate": 0.0006,
+ "loss": 4.584898948669434,
+ "step": 1256
+ },
+ {
+ "epoch": 17.46113537117904,
+ "grad_norm": 0.01714998669922352,
+ "learning_rate": 0.0006,
+ "loss": 4.523544788360596,
+ "step": 1257
+ },
+ {
+ "epoch": 17.475109170305675,
+ "grad_norm": 0.01802288182079792,
+ "learning_rate": 0.0006,
+ "loss": 4.595581531524658,
+ "step": 1258
+ },
+ {
+ "epoch": 17.489082969432314,
+ "grad_norm": 0.0180289875715971,
+ "learning_rate": 0.0006,
+ "loss": 4.394440650939941,
+ "step": 1259
+ },
+ {
+ "epoch": 17.503056768558952,
+ "grad_norm": 0.018504245206713676,
+ "learning_rate": 0.0006,
+ "loss": 4.4794416427612305,
+ "step": 1260
+ },
+ {
+ "epoch": 17.51703056768559,
+ "grad_norm": 0.019973233342170715,
+ "learning_rate": 0.0006,
+ "loss": 4.4241437911987305,
+ "step": 1261
+ },
+ {
+ "epoch": 17.531004366812226,
+ "grad_norm": 0.020063083618879318,
+ "learning_rate": 0.0006,
+ "loss": 4.48068380355835,
+ "step": 1262
+ },
+ {
+ "epoch": 17.544978165938865,
+ "grad_norm": 0.02066640742123127,
+ "learning_rate": 0.0006,
+ "loss": 4.628993034362793,
+ "step": 1263
+ },
+ {
+ "epoch": 17.558951965065503,
+ "grad_norm": 0.02206416428089142,
+ "learning_rate": 0.0006,
+ "loss": 4.564790725708008,
+ "step": 1264
+ },
+ {
+ "epoch": 17.57292576419214,
+ "grad_norm": 0.020640335977077484,
+ "learning_rate": 0.0006,
+ "loss": 4.578705787658691,
+ "step": 1265
+ },
+ {
+ "epoch": 17.586899563318777,
+ "grad_norm": 0.019093405455350876,
+ "learning_rate": 0.0006,
+ "loss": 4.5023603439331055,
+ "step": 1266
+ },
+ {
+ "epoch": 17.600873362445416,
+ "grad_norm": 0.019459193572402,
+ "learning_rate": 0.0006,
+ "loss": 4.594531059265137,
+ "step": 1267
+ },
+ {
+ "epoch": 17.61484716157205,
+ "grad_norm": 0.02021171525120735,
+ "learning_rate": 0.0006,
+ "loss": 4.770299911499023,
+ "step": 1268
+ },
+ {
+ "epoch": 17.62882096069869,
+ "grad_norm": 0.021161191165447235,
+ "learning_rate": 0.0006,
+ "loss": 4.5380401611328125,
+ "step": 1269
+ },
+ {
+ "epoch": 17.64279475982533,
+ "grad_norm": 0.021013403311371803,
+ "learning_rate": 0.0006,
+ "loss": 4.455524921417236,
+ "step": 1270
+ },
+ {
+ "epoch": 17.656768558951963,
+ "grad_norm": 0.020336592569947243,
+ "learning_rate": 0.0006,
+ "loss": 4.533672332763672,
+ "step": 1271
+ },
+ {
+ "epoch": 17.670742358078602,
+ "grad_norm": 0.020159434527158737,
+ "learning_rate": 0.0006,
+ "loss": 4.596151351928711,
+ "step": 1272
+ },
+ {
+ "epoch": 17.68471615720524,
+ "grad_norm": 0.020432960242033005,
+ "learning_rate": 0.0006,
+ "loss": 4.55531644821167,
+ "step": 1273
+ },
+ {
+ "epoch": 17.69868995633188,
+ "grad_norm": 0.025466229766607285,
+ "learning_rate": 0.0006,
+ "loss": 4.566912651062012,
+ "step": 1274
+ },
+ {
+ "epoch": 17.712663755458514,
+ "grad_norm": 0.026537558063864708,
+ "learning_rate": 0.0006,
+ "loss": 4.526003837585449,
+ "step": 1275
+ },
+ {
+ "epoch": 17.726637554585153,
+ "grad_norm": 0.024397607892751694,
+ "learning_rate": 0.0006,
+ "loss": 4.541755676269531,
+ "step": 1276
+ },
+ {
+ "epoch": 17.74061135371179,
+ "grad_norm": 0.022062312811613083,
+ "learning_rate": 0.0006,
+ "loss": 4.6532793045043945,
+ "step": 1277
+ },
+ {
+ "epoch": 17.754585152838427,
+ "grad_norm": 0.02204282023012638,
+ "learning_rate": 0.0006,
+ "loss": 4.582825660705566,
+ "step": 1278
+ },
+ {
+ "epoch": 17.768558951965066,
+ "grad_norm": 0.020844416692852974,
+ "learning_rate": 0.0006,
+ "loss": 4.522336483001709,
+ "step": 1279
+ },
+ {
+ "epoch": 17.782532751091704,
+ "grad_norm": 0.019815703853964806,
+ "learning_rate": 0.0006,
+ "loss": 4.607398509979248,
+ "step": 1280
+ },
+ {
+ "epoch": 17.79650655021834,
+ "grad_norm": 0.016676954925060272,
+ "learning_rate": 0.0006,
+ "loss": 4.446206092834473,
+ "step": 1281
+ },
+ {
+ "epoch": 17.810480349344978,
+ "grad_norm": 0.0166452769190073,
+ "learning_rate": 0.0006,
+ "loss": 4.505931854248047,
+ "step": 1282
+ },
+ {
+ "epoch": 17.824454148471617,
+ "grad_norm": 0.015928907319903374,
+ "learning_rate": 0.0006,
+ "loss": 4.455267429351807,
+ "step": 1283
+ },
+ {
+ "epoch": 17.83842794759825,
+ "grad_norm": 0.015219560824334621,
+ "learning_rate": 0.0006,
+ "loss": 4.5623674392700195,
+ "step": 1284
+ },
+ {
+ "epoch": 17.85240174672489,
+ "grad_norm": 0.015125355683267117,
+ "learning_rate": 0.0006,
+ "loss": 4.554104804992676,
+ "step": 1285
+ },
+ {
+ "epoch": 17.86637554585153,
+ "grad_norm": 0.015612718649208546,
+ "learning_rate": 0.0006,
+ "loss": 4.484847068786621,
+ "step": 1286
+ },
+ {
+ "epoch": 17.880349344978168,
+ "grad_norm": 0.015215662308037281,
+ "learning_rate": 0.0006,
+ "loss": 4.4802093505859375,
+ "step": 1287
+ },
+ {
+ "epoch": 17.894323144104803,
+ "grad_norm": 0.01511684525758028,
+ "learning_rate": 0.0006,
+ "loss": 4.537298679351807,
+ "step": 1288
+ },
+ {
+ "epoch": 17.90829694323144,
+ "grad_norm": 0.013841322623193264,
+ "learning_rate": 0.0006,
+ "loss": 4.427488327026367,
+ "step": 1289
+ },
+ {
+ "epoch": 17.92227074235808,
+ "grad_norm": 0.015115964226424694,
+ "learning_rate": 0.0006,
+ "loss": 4.4547119140625,
+ "step": 1290
+ },
+ {
+ "epoch": 17.936244541484715,
+ "grad_norm": 0.016411006450653076,
+ "learning_rate": 0.0006,
+ "loss": 4.52994966506958,
+ "step": 1291
+ },
+ {
+ "epoch": 17.950218340611354,
+ "grad_norm": 0.017313152551651,
+ "learning_rate": 0.0006,
+ "loss": 4.5337653160095215,
+ "step": 1292
+ },
+ {
+ "epoch": 17.964192139737992,
+ "grad_norm": 0.016559796407818794,
+ "learning_rate": 0.0006,
+ "loss": 4.5277910232543945,
+ "step": 1293
+ },
+ {
+ "epoch": 17.978165938864628,
+ "grad_norm": 0.015854761004447937,
+ "learning_rate": 0.0006,
+ "loss": 4.6334309577941895,
+ "step": 1294
+ },
+ {
+ "epoch": 17.992139737991266,
+ "grad_norm": 0.015658238902688026,
+ "learning_rate": 0.0006,
+ "loss": 4.613372802734375,
+ "step": 1295
+ },
+ {
+ "epoch": 18.0,
+ "grad_norm": 0.0169155802577734,
+ "learning_rate": 0.0006,
+ "loss": 4.527356147766113,
+ "step": 1296
+ },
+ {
+ "epoch": 18.0,
+ "eval_loss": 4.764543056488037,
+ "eval_runtime": 56.5998,
+ "eval_samples_per_second": 43.145,
+ "eval_steps_per_second": 1.36,
+ "step": 1296
+ },
+ {
+ "epoch": 18.01397379912664,
+ "grad_norm": 0.015348963439464569,
+ "learning_rate": 0.0006,
+ "loss": 4.466423988342285,
+ "step": 1297
+ },
+ {
+ "epoch": 18.027947598253274,
+ "grad_norm": 0.01768588088452816,
+ "learning_rate": 0.0006,
+ "loss": 4.545933246612549,
+ "step": 1298
+ },
+ {
+ "epoch": 18.041921397379912,
+ "grad_norm": 0.017381226643919945,
+ "learning_rate": 0.0006,
+ "loss": 4.494060039520264,
+ "step": 1299
+ },
+ {
+ "epoch": 18.05589519650655,
+ "grad_norm": 0.019002335146069527,
+ "learning_rate": 0.0006,
+ "loss": 4.395200729370117,
+ "step": 1300
+ },
+ {
+ "epoch": 18.069868995633186,
+ "grad_norm": 0.01827932707965374,
+ "learning_rate": 0.0006,
+ "loss": 4.424848556518555,
+ "step": 1301
+ },
+ {
+ "epoch": 18.083842794759825,
+ "grad_norm": 0.016383031383156776,
+ "learning_rate": 0.0006,
+ "loss": 4.494722366333008,
+ "step": 1302
+ },
+ {
+ "epoch": 18.097816593886463,
+ "grad_norm": 0.016265157610177994,
+ "learning_rate": 0.0006,
+ "loss": 4.416522979736328,
+ "step": 1303
+ },
+ {
+ "epoch": 18.111790393013102,
+ "grad_norm": 0.01702386513352394,
+ "learning_rate": 0.0006,
+ "loss": 4.600103378295898,
+ "step": 1304
+ },
+ {
+ "epoch": 18.125764192139737,
+ "grad_norm": 0.018495498225092888,
+ "learning_rate": 0.0006,
+ "loss": 4.588401794433594,
+ "step": 1305
+ },
+ {
+ "epoch": 18.139737991266376,
+ "grad_norm": 0.017921684309840202,
+ "learning_rate": 0.0006,
+ "loss": 4.478075981140137,
+ "step": 1306
+ },
+ {
+ "epoch": 18.153711790393015,
+ "grad_norm": 0.01801278069615364,
+ "learning_rate": 0.0006,
+ "loss": 4.413186550140381,
+ "step": 1307
+ },
+ {
+ "epoch": 18.16768558951965,
+ "grad_norm": 0.017211275175213814,
+ "learning_rate": 0.0006,
+ "loss": 4.461341857910156,
+ "step": 1308
+ },
+ {
+ "epoch": 18.18165938864629,
+ "grad_norm": 0.017589328810572624,
+ "learning_rate": 0.0006,
+ "loss": 4.546969413757324,
+ "step": 1309
+ },
+ {
+ "epoch": 18.195633187772927,
+ "grad_norm": 0.01797698810696602,
+ "learning_rate": 0.0006,
+ "loss": 4.530797004699707,
+ "step": 1310
+ },
+ {
+ "epoch": 18.209606986899562,
+ "grad_norm": 0.020408619195222855,
+ "learning_rate": 0.0006,
+ "loss": 4.62999153137207,
+ "step": 1311
+ },
+ {
+ "epoch": 18.2235807860262,
+ "grad_norm": 0.01994038000702858,
+ "learning_rate": 0.0006,
+ "loss": 4.586811065673828,
+ "step": 1312
+ },
+ {
+ "epoch": 18.23755458515284,
+ "grad_norm": 0.019381340593099594,
+ "learning_rate": 0.0006,
+ "loss": 4.4769368171691895,
+ "step": 1313
+ },
+ {
+ "epoch": 18.251528384279474,
+ "grad_norm": 0.020484555512666702,
+ "learning_rate": 0.0006,
+ "loss": 4.524302959442139,
+ "step": 1314
+ },
+ {
+ "epoch": 18.265502183406113,
+ "grad_norm": 0.02426416426897049,
+ "learning_rate": 0.0006,
+ "loss": 4.561705589294434,
+ "step": 1315
+ },
+ {
+ "epoch": 18.27947598253275,
+ "grad_norm": 0.024748101830482483,
+ "learning_rate": 0.0006,
+ "loss": 4.4654316902160645,
+ "step": 1316
+ },
+ {
+ "epoch": 18.29344978165939,
+ "grad_norm": 0.020703328773379326,
+ "learning_rate": 0.0006,
+ "loss": 4.53033447265625,
+ "step": 1317
+ },
+ {
+ "epoch": 18.307423580786025,
+ "grad_norm": 0.01765458658337593,
+ "learning_rate": 0.0006,
+ "loss": 4.5154266357421875,
+ "step": 1318
+ },
+ {
+ "epoch": 18.321397379912664,
+ "grad_norm": 0.018646420910954475,
+ "learning_rate": 0.0006,
+ "loss": 4.5224714279174805,
+ "step": 1319
+ },
+ {
+ "epoch": 18.335371179039303,
+ "grad_norm": 0.01907474733889103,
+ "learning_rate": 0.0006,
+ "loss": 4.5055437088012695,
+ "step": 1320
+ },
+ {
+ "epoch": 18.349344978165938,
+ "grad_norm": 0.018179846927523613,
+ "learning_rate": 0.0006,
+ "loss": 4.462161064147949,
+ "step": 1321
+ },
+ {
+ "epoch": 18.363318777292577,
+ "grad_norm": 0.016697878018021584,
+ "learning_rate": 0.0006,
+ "loss": 4.473718643188477,
+ "step": 1322
+ },
+ {
+ "epoch": 18.377292576419215,
+ "grad_norm": 0.017504019662737846,
+ "learning_rate": 0.0006,
+ "loss": 4.474348068237305,
+ "step": 1323
+ },
+ {
+ "epoch": 18.39126637554585,
+ "grad_norm": 0.018889913335442543,
+ "learning_rate": 0.0006,
+ "loss": 4.4531636238098145,
+ "step": 1324
+ },
+ {
+ "epoch": 18.40524017467249,
+ "grad_norm": 0.019167251884937286,
+ "learning_rate": 0.0006,
+ "loss": 4.4278364181518555,
+ "step": 1325
+ },
+ {
+ "epoch": 18.419213973799128,
+ "grad_norm": 0.015895741060376167,
+ "learning_rate": 0.0006,
+ "loss": 4.474592208862305,
+ "step": 1326
+ },
+ {
+ "epoch": 18.433187772925763,
+ "grad_norm": 0.01531882956624031,
+ "learning_rate": 0.0006,
+ "loss": 4.403757572174072,
+ "step": 1327
+ },
+ {
+ "epoch": 18.4471615720524,
+ "grad_norm": 0.01565849967300892,
+ "learning_rate": 0.0006,
+ "loss": 4.507664680480957,
+ "step": 1328
+ },
+ {
+ "epoch": 18.46113537117904,
+ "grad_norm": 0.01602315716445446,
+ "learning_rate": 0.0006,
+ "loss": 4.5648908615112305,
+ "step": 1329
+ },
+ {
+ "epoch": 18.475109170305675,
+ "grad_norm": 0.016467366367578506,
+ "learning_rate": 0.0006,
+ "loss": 4.428762912750244,
+ "step": 1330
+ },
+ {
+ "epoch": 18.489082969432314,
+ "grad_norm": 0.01604127697646618,
+ "learning_rate": 0.0006,
+ "loss": 4.527771472930908,
+ "step": 1331
+ },
+ {
+ "epoch": 18.503056768558952,
+ "grad_norm": 0.018825586885213852,
+ "learning_rate": 0.0006,
+ "loss": 4.485386848449707,
+ "step": 1332
+ },
+ {
+ "epoch": 18.51703056768559,
+ "grad_norm": 0.019258007407188416,
+ "learning_rate": 0.0006,
+ "loss": 4.414972305297852,
+ "step": 1333
+ },
+ {
+ "epoch": 18.531004366812226,
+ "grad_norm": 0.018192021176218987,
+ "learning_rate": 0.0006,
+ "loss": 4.412166595458984,
+ "step": 1334
+ },
+ {
+ "epoch": 18.544978165938865,
+ "grad_norm": 0.016120698302984238,
+ "learning_rate": 0.0006,
+ "loss": 4.484317779541016,
+ "step": 1335
+ },
+ {
+ "epoch": 18.558951965065503,
+ "grad_norm": 0.016496622934937477,
+ "learning_rate": 0.0006,
+ "loss": 4.514986515045166,
+ "step": 1336
+ },
+ {
+ "epoch": 18.57292576419214,
+ "grad_norm": 0.01795116998255253,
+ "learning_rate": 0.0006,
+ "loss": 4.41473388671875,
+ "step": 1337
+ },
+ {
+ "epoch": 18.586899563318777,
+ "grad_norm": 0.019062276929616928,
+ "learning_rate": 0.0006,
+ "loss": 4.601996898651123,
+ "step": 1338
+ },
+ {
+ "epoch": 18.600873362445416,
+ "grad_norm": 0.017174631357192993,
+ "learning_rate": 0.0006,
+ "loss": 4.571218490600586,
+ "step": 1339
+ },
+ {
+ "epoch": 18.61484716157205,
+ "grad_norm": 0.014015606604516506,
+ "learning_rate": 0.0006,
+ "loss": 4.526606559753418,
+ "step": 1340
+ },
+ {
+ "epoch": 18.62882096069869,
+ "grad_norm": 0.015358510427176952,
+ "learning_rate": 0.0006,
+ "loss": 4.604286193847656,
+ "step": 1341
+ },
+ {
+ "epoch": 18.64279475982533,
+ "grad_norm": 0.01619391329586506,
+ "learning_rate": 0.0006,
+ "loss": 4.560883522033691,
+ "step": 1342
+ },
+ {
+ "epoch": 18.656768558951963,
+ "grad_norm": 0.015972964465618134,
+ "learning_rate": 0.0006,
+ "loss": 4.479755401611328,
+ "step": 1343
+ },
+ {
+ "epoch": 18.670742358078602,
+ "grad_norm": 0.016861025243997574,
+ "learning_rate": 0.0006,
+ "loss": 4.474471092224121,
+ "step": 1344
+ },
+ {
+ "epoch": 18.68471615720524,
+ "grad_norm": 0.01822078227996826,
+ "learning_rate": 0.0006,
+ "loss": 4.509489059448242,
+ "step": 1345
+ },
+ {
+ "epoch": 18.69868995633188,
+ "grad_norm": 0.019415007904171944,
+ "learning_rate": 0.0006,
+ "loss": 4.508941173553467,
+ "step": 1346
+ },
+ {
+ "epoch": 18.712663755458514,
+ "grad_norm": 0.019856126978993416,
+ "learning_rate": 0.0006,
+ "loss": 4.446670055389404,
+ "step": 1347
+ },
+ {
+ "epoch": 18.726637554585153,
+ "grad_norm": 0.01842520199716091,
+ "learning_rate": 0.0006,
+ "loss": 4.524422645568848,
+ "step": 1348
+ },
+ {
+ "epoch": 18.74061135371179,
+ "grad_norm": 0.017196480184793472,
+ "learning_rate": 0.0006,
+ "loss": 4.498466968536377,
+ "step": 1349
+ },
+ {
+ "epoch": 18.754585152838427,
+ "grad_norm": 0.01823010854423046,
+ "learning_rate": 0.0006,
+ "loss": 4.534684181213379,
+ "step": 1350
+ },
+ {
+ "epoch": 18.768558951965066,
+ "grad_norm": 0.020398065447807312,
+ "learning_rate": 0.0006,
+ "loss": 4.456494331359863,
+ "step": 1351
+ },
+ {
+ "epoch": 18.782532751091704,
+ "grad_norm": 0.022992262616753578,
+ "learning_rate": 0.0006,
+ "loss": 4.476374626159668,
+ "step": 1352
+ },
+ {
+ "epoch": 18.79650655021834,
+ "grad_norm": 0.02235420234501362,
+ "learning_rate": 0.0006,
+ "loss": 4.587776184082031,
+ "step": 1353
+ },
+ {
+ "epoch": 18.810480349344978,
+ "grad_norm": 0.02093626745045185,
+ "learning_rate": 0.0006,
+ "loss": 4.424172401428223,
+ "step": 1354
+ },
+ {
+ "epoch": 18.824454148471617,
+ "grad_norm": 0.01997302658855915,
+ "learning_rate": 0.0006,
+ "loss": 4.448066711425781,
+ "step": 1355
+ },
+ {
+ "epoch": 18.83842794759825,
+ "grad_norm": 0.018265459686517715,
+ "learning_rate": 0.0006,
+ "loss": 4.436610698699951,
+ "step": 1356
+ },
+ {
+ "epoch": 18.85240174672489,
+ "grad_norm": 0.01673833839595318,
+ "learning_rate": 0.0006,
+ "loss": 4.5502214431762695,
+ "step": 1357
+ },
+ {
+ "epoch": 18.86637554585153,
+ "grad_norm": 0.017908833920955658,
+ "learning_rate": 0.0006,
+ "loss": 4.550530433654785,
+ "step": 1358
+ },
+ {
+ "epoch": 18.880349344978168,
+ "grad_norm": 0.01796654611825943,
+ "learning_rate": 0.0006,
+ "loss": 4.531463146209717,
+ "step": 1359
+ },
+ {
+ "epoch": 18.894323144104803,
+ "grad_norm": 0.01646183803677559,
+ "learning_rate": 0.0006,
+ "loss": 4.498582363128662,
+ "step": 1360
+ },
+ {
+ "epoch": 18.90829694323144,
+ "grad_norm": 0.015430327504873276,
+ "learning_rate": 0.0006,
+ "loss": 4.52571439743042,
+ "step": 1361
+ },
+ {
+ "epoch": 18.92227074235808,
+ "grad_norm": 0.016527952626347542,
+ "learning_rate": 0.0006,
+ "loss": 4.547013759613037,
+ "step": 1362
+ },
+ {
+ "epoch": 18.936244541484715,
+ "grad_norm": 0.016263660043478012,
+ "learning_rate": 0.0006,
+ "loss": 4.468754291534424,
+ "step": 1363
+ },
+ {
+ "epoch": 18.950218340611354,
+ "grad_norm": 0.016347210854291916,
+ "learning_rate": 0.0006,
+ "loss": 4.450741767883301,
+ "step": 1364
+ },
+ {
+ "epoch": 18.964192139737992,
+ "grad_norm": 0.01701398566365242,
+ "learning_rate": 0.0006,
+ "loss": 4.534899711608887,
+ "step": 1365
+ },
+ {
+ "epoch": 18.978165938864628,
+ "grad_norm": 0.017693081870675087,
+ "learning_rate": 0.0006,
+ "loss": 4.460553169250488,
+ "step": 1366
+ },
+ {
+ "epoch": 18.992139737991266,
+ "grad_norm": 0.018746308982372284,
+ "learning_rate": 0.0006,
+ "loss": 4.538792610168457,
+ "step": 1367
+ },
+ {
+ "epoch": 19.0,
+ "grad_norm": 0.020099416375160217,
+ "learning_rate": 0.0006,
+ "loss": 4.449617385864258,
+ "step": 1368
+ },
+ {
+ "epoch": 19.0,
+ "eval_loss": 4.7756571769714355,
+ "eval_runtime": 57.1804,
+ "eval_samples_per_second": 42.707,
+ "eval_steps_per_second": 1.347,
+ "step": 1368
+ },
+ {
+ "epoch": 19.01397379912664,
+ "grad_norm": 0.018726933747529984,
+ "learning_rate": 0.0006,
+ "loss": 4.446059703826904,
+ "step": 1369
+ },
+ {
+ "epoch": 19.027947598253274,
+ "grad_norm": 0.01956087350845337,
+ "learning_rate": 0.0006,
+ "loss": 4.4520039558410645,
+ "step": 1370
+ },
+ {
+ "epoch": 19.041921397379912,
+ "grad_norm": 0.020092811435461044,
+ "learning_rate": 0.0006,
+ "loss": 4.543343544006348,
+ "step": 1371
+ },
+ {
+ "epoch": 19.05589519650655,
+ "grad_norm": 0.01843622885644436,
+ "learning_rate": 0.0006,
+ "loss": 4.535579204559326,
+ "step": 1372
+ },
+ {
+ "epoch": 19.069868995633186,
+ "grad_norm": 0.016989488154649734,
+ "learning_rate": 0.0006,
+ "loss": 4.52316951751709,
+ "step": 1373
+ },
+ {
+ "epoch": 19.083842794759825,
+ "grad_norm": 0.01805422455072403,
+ "learning_rate": 0.0006,
+ "loss": 4.5393266677856445,
+ "step": 1374
+ },
+ {
+ "epoch": 19.097816593886463,
+ "grad_norm": 0.01874629408121109,
+ "learning_rate": 0.0006,
+ "loss": 4.422059059143066,
+ "step": 1375
+ },
+ {
+ "epoch": 19.111790393013102,
+ "grad_norm": 0.01804221048951149,
+ "learning_rate": 0.0006,
+ "loss": 4.50508451461792,
+ "step": 1376
+ },
+ {
+ "epoch": 19.125764192139737,
+ "grad_norm": 0.01785474829375744,
+ "learning_rate": 0.0006,
+ "loss": 4.412599563598633,
+ "step": 1377
+ },
+ {
+ "epoch": 19.139737991266376,
+ "grad_norm": 0.01971791312098503,
+ "learning_rate": 0.0006,
+ "loss": 4.518318176269531,
+ "step": 1378
+ },
+ {
+ "epoch": 19.153711790393015,
+ "grad_norm": 0.02061633951961994,
+ "learning_rate": 0.0006,
+ "loss": 4.332393169403076,
+ "step": 1379
+ },
+ {
+ "epoch": 19.16768558951965,
+ "grad_norm": 0.020563429221510887,
+ "learning_rate": 0.0006,
+ "loss": 4.445863723754883,
+ "step": 1380
+ },
+ {
+ "epoch": 19.18165938864629,
+ "grad_norm": 0.019995713606476784,
+ "learning_rate": 0.0006,
+ "loss": 4.497312068939209,
+ "step": 1381
+ },
+ {
+ "epoch": 19.195633187772927,
+ "grad_norm": 0.019823070615530014,
+ "learning_rate": 0.0006,
+ "loss": 4.475085258483887,
+ "step": 1382
+ },
+ {
+ "epoch": 19.209606986899562,
+ "grad_norm": 0.019039355218410492,
+ "learning_rate": 0.0006,
+ "loss": 4.50432014465332,
+ "step": 1383
+ },
+ {
+ "epoch": 19.2235807860262,
+ "grad_norm": 0.017876973375678062,
+ "learning_rate": 0.0006,
+ "loss": 4.384788513183594,
+ "step": 1384
+ },
+ {
+ "epoch": 19.23755458515284,
+ "grad_norm": 0.018697699531912804,
+ "learning_rate": 0.0006,
+ "loss": 4.410778045654297,
+ "step": 1385
+ },
+ {
+ "epoch": 19.251528384279474,
+ "grad_norm": 0.018826806917786598,
+ "learning_rate": 0.0006,
+ "loss": 4.524034023284912,
+ "step": 1386
+ },
+ {
+ "epoch": 19.265502183406113,
+ "grad_norm": 0.01824226602911949,
+ "learning_rate": 0.0006,
+ "loss": 4.526393890380859,
+ "step": 1387
+ },
+ {
+ "epoch": 19.27947598253275,
+ "grad_norm": 0.01753208599984646,
+ "learning_rate": 0.0006,
+ "loss": 4.493703365325928,
+ "step": 1388
+ },
+ {
+ "epoch": 19.29344978165939,
+ "grad_norm": 0.01742432825267315,
+ "learning_rate": 0.0006,
+ "loss": 4.392831802368164,
+ "step": 1389
+ },
+ {
+ "epoch": 19.307423580786025,
+ "grad_norm": 0.017099356278777122,
+ "learning_rate": 0.0006,
+ "loss": 4.578668594360352,
+ "step": 1390
+ },
+ {
+ "epoch": 19.321397379912664,
+ "grad_norm": 0.017946023494005203,
+ "learning_rate": 0.0006,
+ "loss": 4.456718444824219,
+ "step": 1391
+ },
+ {
+ "epoch": 19.335371179039303,
+ "grad_norm": 0.017705943435430527,
+ "learning_rate": 0.0006,
+ "loss": 4.553682327270508,
+ "step": 1392
+ },
+ {
+ "epoch": 19.349344978165938,
+ "grad_norm": 0.016647523269057274,
+ "learning_rate": 0.0006,
+ "loss": 4.525016784667969,
+ "step": 1393
+ },
+ {
+ "epoch": 19.363318777292577,
+ "grad_norm": 0.015407416969537735,
+ "learning_rate": 0.0006,
+ "loss": 4.500231742858887,
+ "step": 1394
+ },
+ {
+ "epoch": 19.377292576419215,
+ "grad_norm": 0.017264019697904587,
+ "learning_rate": 0.0006,
+ "loss": 4.4615349769592285,
+ "step": 1395
+ },
+ {
+ "epoch": 19.39126637554585,
+ "grad_norm": 0.01871555484831333,
+ "learning_rate": 0.0006,
+ "loss": 4.281026840209961,
+ "step": 1396
+ },
+ {
+ "epoch": 19.40524017467249,
+ "grad_norm": 0.017361650243401527,
+ "learning_rate": 0.0006,
+ "loss": 4.479382514953613,
+ "step": 1397
+ },
+ {
+ "epoch": 19.419213973799128,
+ "grad_norm": 0.019781362265348434,
+ "learning_rate": 0.0006,
+ "loss": 4.402320861816406,
+ "step": 1398
+ },
+ {
+ "epoch": 19.433187772925763,
+ "grad_norm": 0.019043298438191414,
+ "learning_rate": 0.0006,
+ "loss": 4.476983070373535,
+ "step": 1399
+ },
+ {
+ "epoch": 19.4471615720524,
+ "grad_norm": 0.017137985676527023,
+ "learning_rate": 0.0006,
+ "loss": 4.418689250946045,
+ "step": 1400
+ },
+ {
+ "epoch": 19.46113537117904,
+ "grad_norm": 0.016481833532452583,
+ "learning_rate": 0.0006,
+ "loss": 4.579145431518555,
+ "step": 1401
+ },
+ {
+ "epoch": 19.475109170305675,
+ "grad_norm": 0.016507714986801147,
+ "learning_rate": 0.0006,
+ "loss": 4.519158363342285,
+ "step": 1402
+ },
+ {
+ "epoch": 19.489082969432314,
+ "grad_norm": 0.018092872574925423,
+ "learning_rate": 0.0006,
+ "loss": 4.513748645782471,
+ "step": 1403
+ },
+ {
+ "epoch": 19.503056768558952,
+ "grad_norm": 0.018153440207242966,
+ "learning_rate": 0.0006,
+ "loss": 4.645384311676025,
+ "step": 1404
+ },
+ {
+ "epoch": 19.51703056768559,
+ "grad_norm": 0.01657119207084179,
+ "learning_rate": 0.0006,
+ "loss": 4.399757385253906,
+ "step": 1405
+ },
+ {
+ "epoch": 19.531004366812226,
+ "grad_norm": 0.0156280305236578,
+ "learning_rate": 0.0006,
+ "loss": 4.467836380004883,
+ "step": 1406
+ },
+ {
+ "epoch": 19.544978165938865,
+ "grad_norm": 0.016838403418660164,
+ "learning_rate": 0.0006,
+ "loss": 4.529176235198975,
+ "step": 1407
+ },
+ {
+ "epoch": 19.558951965065503,
+ "grad_norm": 0.01743236929178238,
+ "learning_rate": 0.0006,
+ "loss": 4.440032482147217,
+ "step": 1408
+ },
+ {
+ "epoch": 19.57292576419214,
+ "grad_norm": 0.01553610060364008,
+ "learning_rate": 0.0006,
+ "loss": 4.425588607788086,
+ "step": 1409
+ },
+ {
+ "epoch": 19.586899563318777,
+ "grad_norm": 0.017259271815419197,
+ "learning_rate": 0.0006,
+ "loss": 4.405452728271484,
+ "step": 1410
+ },
+ {
+ "epoch": 19.600873362445416,
+ "grad_norm": 0.015568481758236885,
+ "learning_rate": 0.0006,
+ "loss": 4.444279193878174,
+ "step": 1411
+ },
+ {
+ "epoch": 19.61484716157205,
+ "grad_norm": 0.015331248752772808,
+ "learning_rate": 0.0006,
+ "loss": 4.451846122741699,
+ "step": 1412
+ },
+ {
+ "epoch": 19.62882096069869,
+ "grad_norm": 0.015667378902435303,
+ "learning_rate": 0.0006,
+ "loss": 4.533931255340576,
+ "step": 1413
+ },
+ {
+ "epoch": 19.64279475982533,
+ "grad_norm": 0.016177602112293243,
+ "learning_rate": 0.0006,
+ "loss": 4.467313289642334,
+ "step": 1414
+ },
+ {
+ "epoch": 19.656768558951963,
+ "grad_norm": 0.01733129471540451,
+ "learning_rate": 0.0006,
+ "loss": 4.496893882751465,
+ "step": 1415
+ },
+ {
+ "epoch": 19.670742358078602,
+ "grad_norm": 0.016539594158530235,
+ "learning_rate": 0.0006,
+ "loss": 4.401961326599121,
+ "step": 1416
+ },
+ {
+ "epoch": 19.68471615720524,
+ "grad_norm": 0.016058551147580147,
+ "learning_rate": 0.0006,
+ "loss": 4.503375053405762,
+ "step": 1417
+ },
+ {
+ "epoch": 19.69868995633188,
+ "grad_norm": 0.01715777814388275,
+ "learning_rate": 0.0006,
+ "loss": 4.3808369636535645,
+ "step": 1418
+ },
+ {
+ "epoch": 19.712663755458514,
+ "grad_norm": 0.01681119203567505,
+ "learning_rate": 0.0006,
+ "loss": 4.5469560623168945,
+ "step": 1419
+ },
+ {
+ "epoch": 19.726637554585153,
+ "grad_norm": 0.01634550653398037,
+ "learning_rate": 0.0006,
+ "loss": 4.328197956085205,
+ "step": 1420
+ },
+ {
+ "epoch": 19.74061135371179,
+ "grad_norm": 0.01549347210675478,
+ "learning_rate": 0.0006,
+ "loss": 4.534621238708496,
+ "step": 1421
+ },
+ {
+ "epoch": 19.754585152838427,
+ "grad_norm": 0.01654665172100067,
+ "learning_rate": 0.0006,
+ "loss": 4.4724225997924805,
+ "step": 1422
+ },
+ {
+ "epoch": 19.768558951965066,
+ "grad_norm": 0.017119025811553,
+ "learning_rate": 0.0006,
+ "loss": 4.3897247314453125,
+ "step": 1423
+ },
+ {
+ "epoch": 19.782532751091704,
+ "grad_norm": 0.01603223942220211,
+ "learning_rate": 0.0006,
+ "loss": 4.412731170654297,
+ "step": 1424
+ },
+ {
+ "epoch": 19.79650655021834,
+ "grad_norm": 0.016774067655205727,
+ "learning_rate": 0.0006,
+ "loss": 4.54707145690918,
+ "step": 1425
+ },
+ {
+ "epoch": 19.810480349344978,
+ "grad_norm": 0.016453025862574577,
+ "learning_rate": 0.0006,
+ "loss": 4.462501049041748,
+ "step": 1426
+ },
+ {
+ "epoch": 19.824454148471617,
+ "grad_norm": 0.01597629487514496,
+ "learning_rate": 0.0006,
+ "loss": 4.229758262634277,
+ "step": 1427
+ },
+ {
+ "epoch": 19.83842794759825,
+ "grad_norm": 0.017999034374952316,
+ "learning_rate": 0.0006,
+ "loss": 4.389263153076172,
+ "step": 1428
+ },
+ {
+ "epoch": 19.85240174672489,
+ "grad_norm": 0.02165898121893406,
+ "learning_rate": 0.0006,
+ "loss": 4.342130661010742,
+ "step": 1429
+ },
+ {
+ "epoch": 19.86637554585153,
+ "grad_norm": 0.02222822792828083,
+ "learning_rate": 0.0006,
+ "loss": 4.492661476135254,
+ "step": 1430
+ },
+ {
+ "epoch": 19.880349344978168,
+ "grad_norm": 0.019378066062927246,
+ "learning_rate": 0.0006,
+ "loss": 4.4614105224609375,
+ "step": 1431
+ },
+ {
+ "epoch": 19.894323144104803,
+ "grad_norm": 0.019908083602786064,
+ "learning_rate": 0.0006,
+ "loss": 4.370929718017578,
+ "step": 1432
+ },
+ {
+ "epoch": 19.90829694323144,
+ "grad_norm": 0.022296471521258354,
+ "learning_rate": 0.0006,
+ "loss": 4.438320159912109,
+ "step": 1433
+ },
+ {
+ "epoch": 19.92227074235808,
+ "grad_norm": 0.021810343489050865,
+ "learning_rate": 0.0006,
+ "loss": 4.462067127227783,
+ "step": 1434
+ },
+ {
+ "epoch": 19.936244541484715,
+ "grad_norm": 0.02129209227859974,
+ "learning_rate": 0.0006,
+ "loss": 4.541261196136475,
+ "step": 1435
+ },
+ {
+ "epoch": 19.950218340611354,
+ "grad_norm": 0.02079629711806774,
+ "learning_rate": 0.0006,
+ "loss": 4.36712646484375,
+ "step": 1436
+ },
+ {
+ "epoch": 19.964192139737992,
+ "grad_norm": 0.022501740604639053,
+ "learning_rate": 0.0006,
+ "loss": 4.5279083251953125,
+ "step": 1437
+ },
+ {
+ "epoch": 19.978165938864628,
+ "grad_norm": 0.019121676683425903,
+ "learning_rate": 0.0006,
+ "loss": 4.477789878845215,
+ "step": 1438
+ },
+ {
+ "epoch": 19.992139737991266,
+ "grad_norm": 0.017429588362574577,
+ "learning_rate": 0.0006,
+ "loss": 4.4023847579956055,
+ "step": 1439
+ },
+ {
+ "epoch": 20.0,
+ "grad_norm": 0.018857665359973907,
+ "learning_rate": 0.0006,
+ "loss": 4.494743347167969,
+ "step": 1440
+ },
+ {
+ "epoch": 20.0,
+ "eval_loss": 4.694555759429932,
+ "eval_runtime": 56.4146,
+ "eval_samples_per_second": 43.287,
+ "eval_steps_per_second": 1.365,
+ "step": 1440
+ },
+ {
+ "epoch": 20.01397379912664,
+ "grad_norm": 0.016505861654877663,
+ "learning_rate": 0.0006,
+ "loss": 4.506229877471924,
+ "step": 1441
+ },
+ {
+ "epoch": 20.027947598253274,
+ "grad_norm": 0.017571842297911644,
+ "learning_rate": 0.0006,
+ "loss": 4.384073257446289,
+ "step": 1442
+ },
+ {
+ "epoch": 20.041921397379912,
+ "grad_norm": 0.019398365169763565,
+ "learning_rate": 0.0006,
+ "loss": 4.482177734375,
+ "step": 1443
+ },
+ {
+ "epoch": 20.05589519650655,
+ "grad_norm": 0.02011062018573284,
+ "learning_rate": 0.0006,
+ "loss": 4.440773963928223,
+ "step": 1444
+ },
+ {
+ "epoch": 20.069868995633186,
+ "grad_norm": 0.019026953727006912,
+ "learning_rate": 0.0006,
+ "loss": 4.507168769836426,
+ "step": 1445
+ },
+ {
+ "epoch": 20.083842794759825,
+ "grad_norm": 0.020506031811237335,
+ "learning_rate": 0.0006,
+ "loss": 4.469079971313477,
+ "step": 1446
+ },
+ {
+ "epoch": 20.097816593886463,
+ "grad_norm": 0.019608162343502045,
+ "learning_rate": 0.0006,
+ "loss": 4.395927429199219,
+ "step": 1447
+ },
+ {
+ "epoch": 20.111790393013102,
+ "grad_norm": 0.019419124349951744,
+ "learning_rate": 0.0006,
+ "loss": 4.484235763549805,
+ "step": 1448
+ },
+ {
+ "epoch": 20.125764192139737,
+ "grad_norm": 0.021775128319859505,
+ "learning_rate": 0.0006,
+ "loss": 4.333173751831055,
+ "step": 1449
+ },
+ {
+ "epoch": 20.139737991266376,
+ "grad_norm": 0.025512272492051125,
+ "learning_rate": 0.0006,
+ "loss": 4.460362434387207,
+ "step": 1450
+ },
+ {
+ "epoch": 20.153711790393015,
+ "grad_norm": 0.025157401338219643,
+ "learning_rate": 0.0006,
+ "loss": 4.4361395835876465,
+ "step": 1451
+ },
+ {
+ "epoch": 20.16768558951965,
+ "grad_norm": 0.019862616434693336,
+ "learning_rate": 0.0006,
+ "loss": 4.426807403564453,
+ "step": 1452
+ },
+ {
+ "epoch": 20.18165938864629,
+ "grad_norm": 0.01932152360677719,
+ "learning_rate": 0.0006,
+ "loss": 4.375677108764648,
+ "step": 1453
+ },
+ {
+ "epoch": 20.195633187772927,
+ "grad_norm": 0.01760999858379364,
+ "learning_rate": 0.0006,
+ "loss": 4.42259407043457,
+ "step": 1454
+ },
+ {
+ "epoch": 20.209606986899562,
+ "grad_norm": 0.016998112201690674,
+ "learning_rate": 0.0006,
+ "loss": 4.481280326843262,
+ "step": 1455
+ },
+ {
+ "epoch": 20.2235807860262,
+ "grad_norm": 0.017199190333485603,
+ "learning_rate": 0.0006,
+ "loss": 4.496712684631348,
+ "step": 1456
+ },
+ {
+ "epoch": 20.23755458515284,
+ "grad_norm": 0.016833679750561714,
+ "learning_rate": 0.0006,
+ "loss": 4.40736722946167,
+ "step": 1457
+ },
+ {
+ "epoch": 20.251528384279474,
+ "grad_norm": 0.017546730116009712,
+ "learning_rate": 0.0006,
+ "loss": 4.4915266036987305,
+ "step": 1458
+ },
+ {
+ "epoch": 20.265502183406113,
+ "grad_norm": 0.016595078632235527,
+ "learning_rate": 0.0006,
+ "loss": 4.460815906524658,
+ "step": 1459
+ },
+ {
+ "epoch": 20.27947598253275,
+ "grad_norm": 0.017728568986058235,
+ "learning_rate": 0.0006,
+ "loss": 4.346048831939697,
+ "step": 1460
+ },
+ {
+ "epoch": 20.29344978165939,
+ "grad_norm": 0.01849197782576084,
+ "learning_rate": 0.0006,
+ "loss": 4.399681091308594,
+ "step": 1461
+ },
+ {
+ "epoch": 20.307423580786025,
+ "grad_norm": 0.018336588516831398,
+ "learning_rate": 0.0006,
+ "loss": 4.414183616638184,
+ "step": 1462
+ },
+ {
+ "epoch": 20.321397379912664,
+ "grad_norm": 0.018086519092321396,
+ "learning_rate": 0.0006,
+ "loss": 4.346586227416992,
+ "step": 1463
+ },
+ {
+ "epoch": 20.335371179039303,
+ "grad_norm": 0.019697299227118492,
+ "learning_rate": 0.0006,
+ "loss": 4.4420952796936035,
+ "step": 1464
+ },
+ {
+ "epoch": 20.349344978165938,
+ "grad_norm": 0.02198074571788311,
+ "learning_rate": 0.0006,
+ "loss": 4.588729381561279,
+ "step": 1465
+ },
+ {
+ "epoch": 20.363318777292577,
+ "grad_norm": 0.022813329473137856,
+ "learning_rate": 0.0006,
+ "loss": 4.49729585647583,
+ "step": 1466
+ },
+ {
+ "epoch": 20.377292576419215,
+ "grad_norm": 0.022006649523973465,
+ "learning_rate": 0.0006,
+ "loss": 4.475641250610352,
+ "step": 1467
+ },
+ {
+ "epoch": 20.39126637554585,
+ "grad_norm": 0.023188438266515732,
+ "learning_rate": 0.0006,
+ "loss": 4.404322624206543,
+ "step": 1468
+ },
+ {
+ "epoch": 20.40524017467249,
+ "grad_norm": 0.022226877510547638,
+ "learning_rate": 0.0006,
+ "loss": 4.50904655456543,
+ "step": 1469
+ },
+ {
+ "epoch": 20.419213973799128,
+ "grad_norm": 0.021082637831568718,
+ "learning_rate": 0.0006,
+ "loss": 4.339011192321777,
+ "step": 1470
+ },
+ {
+ "epoch": 20.433187772925763,
+ "grad_norm": 0.02013542130589485,
+ "learning_rate": 0.0006,
+ "loss": 4.516942024230957,
+ "step": 1471
+ },
+ {
+ "epoch": 20.4471615720524,
+ "grad_norm": 0.01898968778550625,
+ "learning_rate": 0.0006,
+ "loss": 4.315804481506348,
+ "step": 1472
+ },
+ {
+ "epoch": 20.46113537117904,
+ "grad_norm": 0.019450880587100983,
+ "learning_rate": 0.0006,
+ "loss": 4.4794921875,
+ "step": 1473
+ },
+ {
+ "epoch": 20.475109170305675,
+ "grad_norm": 0.018577802926301956,
+ "learning_rate": 0.0006,
+ "loss": 4.422323703765869,
+ "step": 1474
+ },
+ {
+ "epoch": 20.489082969432314,
+ "grad_norm": 0.01778412237763405,
+ "learning_rate": 0.0006,
+ "loss": 4.429789066314697,
+ "step": 1475
+ },
+ {
+ "epoch": 20.503056768558952,
+ "grad_norm": 0.017134476453065872,
+ "learning_rate": 0.0006,
+ "loss": 4.416459083557129,
+ "step": 1476
+ },
+ {
+ "epoch": 20.51703056768559,
+ "grad_norm": 0.017153557389974594,
+ "learning_rate": 0.0006,
+ "loss": 4.446170806884766,
+ "step": 1477
+ },
+ {
+ "epoch": 20.531004366812226,
+ "grad_norm": 0.014998532831668854,
+ "learning_rate": 0.0006,
+ "loss": 4.419447898864746,
+ "step": 1478
+ },
+ {
+ "epoch": 20.544978165938865,
+ "grad_norm": 0.015877850353717804,
+ "learning_rate": 0.0006,
+ "loss": 4.312338829040527,
+ "step": 1479
+ },
+ {
+ "epoch": 20.558951965065503,
+ "grad_norm": 0.01569489948451519,
+ "learning_rate": 0.0006,
+ "loss": 4.572905540466309,
+ "step": 1480
+ },
+ {
+ "epoch": 20.57292576419214,
+ "grad_norm": 0.01591474376618862,
+ "learning_rate": 0.0006,
+ "loss": 4.414907455444336,
+ "step": 1481
+ },
+ {
+ "epoch": 20.586899563318777,
+ "grad_norm": 0.015395903028547764,
+ "learning_rate": 0.0006,
+ "loss": 4.549131393432617,
+ "step": 1482
+ },
+ {
+ "epoch": 20.600873362445416,
+ "grad_norm": 0.015715358778834343,
+ "learning_rate": 0.0006,
+ "loss": 4.530194282531738,
+ "step": 1483
+ },
+ {
+ "epoch": 20.61484716157205,
+ "grad_norm": 0.016042975708842278,
+ "learning_rate": 0.0006,
+ "loss": 4.322149276733398,
+ "step": 1484
+ },
+ {
+ "epoch": 20.62882096069869,
+ "grad_norm": 0.01664186641573906,
+ "learning_rate": 0.0006,
+ "loss": 4.424866676330566,
+ "step": 1485
+ },
+ {
+ "epoch": 20.64279475982533,
+ "grad_norm": 0.01631534844636917,
+ "learning_rate": 0.0006,
+ "loss": 4.441752910614014,
+ "step": 1486
+ },
+ {
+ "epoch": 20.656768558951963,
+ "grad_norm": 0.014057058840990067,
+ "learning_rate": 0.0006,
+ "loss": 4.399982452392578,
+ "step": 1487
+ },
+ {
+ "epoch": 20.670742358078602,
+ "grad_norm": 0.015076273120939732,
+ "learning_rate": 0.0006,
+ "loss": 4.441324234008789,
+ "step": 1488
+ },
+ {
+ "epoch": 20.68471615720524,
+ "grad_norm": 0.014138277620077133,
+ "learning_rate": 0.0006,
+ "loss": 4.4312543869018555,
+ "step": 1489
+ },
+ {
+ "epoch": 20.69868995633188,
+ "grad_norm": 0.013762440532445908,
+ "learning_rate": 0.0006,
+ "loss": 4.387807846069336,
+ "step": 1490
+ },
+ {
+ "epoch": 20.712663755458514,
+ "grad_norm": 0.015058411285281181,
+ "learning_rate": 0.0006,
+ "loss": 4.594747543334961,
+ "step": 1491
+ },
+ {
+ "epoch": 20.726637554585153,
+ "grad_norm": 0.013827123679220676,
+ "learning_rate": 0.0006,
+ "loss": 4.434604167938232,
+ "step": 1492
+ },
+ {
+ "epoch": 20.74061135371179,
+ "grad_norm": 0.015315295197069645,
+ "learning_rate": 0.0006,
+ "loss": 4.588066101074219,
+ "step": 1493
+ },
+ {
+ "epoch": 20.754585152838427,
+ "grad_norm": 0.018678732216358185,
+ "learning_rate": 0.0006,
+ "loss": 4.3514204025268555,
+ "step": 1494
+ },
+ {
+ "epoch": 20.768558951965066,
+ "grad_norm": 0.019521350041031837,
+ "learning_rate": 0.0006,
+ "loss": 4.437658786773682,
+ "step": 1495
+ },
+ {
+ "epoch": 20.782532751091704,
+ "grad_norm": 0.018363403156399727,
+ "learning_rate": 0.0006,
+ "loss": 4.495002269744873,
+ "step": 1496
+ },
+ {
+ "epoch": 20.79650655021834,
+ "grad_norm": 0.016830725595355034,
+ "learning_rate": 0.0006,
+ "loss": 4.402471542358398,
+ "step": 1497
+ },
+ {
+ "epoch": 20.810480349344978,
+ "grad_norm": 0.01722072623670101,
+ "learning_rate": 0.0006,
+ "loss": 4.437653541564941,
+ "step": 1498
+ },
+ {
+ "epoch": 20.824454148471617,
+ "grad_norm": 0.017459599301218987,
+ "learning_rate": 0.0006,
+ "loss": 4.366603851318359,
+ "step": 1499
+ },
+ {
+ "epoch": 20.83842794759825,
+ "grad_norm": 0.018902219831943512,
+ "learning_rate": 0.0006,
+ "loss": 4.501561164855957,
+ "step": 1500
+ },
+ {
+ "epoch": 20.85240174672489,
+ "grad_norm": 0.020814096555113792,
+ "learning_rate": 0.0006,
+ "loss": 4.535295486450195,
+ "step": 1501
+ },
+ {
+ "epoch": 20.86637554585153,
+ "grad_norm": 0.018990976735949516,
+ "learning_rate": 0.0006,
+ "loss": 4.333580017089844,
+ "step": 1502
+ },
+ {
+ "epoch": 20.880349344978168,
+ "grad_norm": 0.01837880350649357,
+ "learning_rate": 0.0006,
+ "loss": 4.313388824462891,
+ "step": 1503
+ },
+ {
+ "epoch": 20.894323144104803,
+ "grad_norm": 0.019820360466837883,
+ "learning_rate": 0.0006,
+ "loss": 4.388157844543457,
+ "step": 1504
+ },
+ {
+ "epoch": 20.90829694323144,
+ "grad_norm": 0.020181143656373024,
+ "learning_rate": 0.0006,
+ "loss": 4.572224140167236,
+ "step": 1505
+ },
+ {
+ "epoch": 20.92227074235808,
+ "grad_norm": 0.021206077188253403,
+ "learning_rate": 0.0006,
+ "loss": 4.3127970695495605,
+ "step": 1506
+ },
+ {
+ "epoch": 20.936244541484715,
+ "grad_norm": 0.017962060868740082,
+ "learning_rate": 0.0006,
+ "loss": 4.384734630584717,
+ "step": 1507
+ },
+ {
+ "epoch": 20.950218340611354,
+ "grad_norm": 0.01737220585346222,
+ "learning_rate": 0.0006,
+ "loss": 4.4075751304626465,
+ "step": 1508
+ },
+ {
+ "epoch": 20.964192139737992,
+ "grad_norm": 0.01839268207550049,
+ "learning_rate": 0.0006,
+ "loss": 4.553779125213623,
+ "step": 1509
+ },
+ {
+ "epoch": 20.978165938864628,
+ "grad_norm": 0.020102640613913536,
+ "learning_rate": 0.0006,
+ "loss": 4.399398326873779,
+ "step": 1510
+ },
+ {
+ "epoch": 20.992139737991266,
+ "grad_norm": 0.018831463530659676,
+ "learning_rate": 0.0006,
+ "loss": 4.358499526977539,
+ "step": 1511
+ },
+ {
+ "epoch": 21.0,
+ "grad_norm": 0.018254734575748444,
+ "learning_rate": 0.0006,
+ "loss": 4.384119510650635,
+ "step": 1512
+ },
+ {
+ "epoch": 21.0,
+ "eval_loss": 4.7524237632751465,
+ "eval_runtime": 57.8812,
+ "eval_samples_per_second": 42.19,
+ "eval_steps_per_second": 1.33,
+ "step": 1512
+ },
+ {
+ "epoch": 21.01397379912664,
+ "grad_norm": 0.019728587940335274,
+ "learning_rate": 0.0006,
+ "loss": 4.32363224029541,
+ "step": 1513
+ },
+ {
+ "epoch": 21.027947598253274,
+ "grad_norm": 0.02216625027358532,
+ "learning_rate": 0.0006,
+ "loss": 4.436988353729248,
+ "step": 1514
+ },
+ {
+ "epoch": 21.041921397379912,
+ "grad_norm": 0.02227606810629368,
+ "learning_rate": 0.0006,
+ "loss": 4.406063556671143,
+ "step": 1515
+ },
+ {
+ "epoch": 21.05589519650655,
+ "grad_norm": 0.020187009125947952,
+ "learning_rate": 0.0006,
+ "loss": 4.324402332305908,
+ "step": 1516
+ },
+ {
+ "epoch": 21.069868995633186,
+ "grad_norm": 0.02095157466828823,
+ "learning_rate": 0.0006,
+ "loss": 4.456544876098633,
+ "step": 1517
+ },
+ {
+ "epoch": 21.083842794759825,
+ "grad_norm": 0.02146013453602791,
+ "learning_rate": 0.0006,
+ "loss": 4.368387222290039,
+ "step": 1518
+ },
+ {
+ "epoch": 21.097816593886463,
+ "grad_norm": 0.02162035182118416,
+ "learning_rate": 0.0006,
+ "loss": 4.48838996887207,
+ "step": 1519
+ },
+ {
+ "epoch": 21.111790393013102,
+ "grad_norm": 0.021745676174759865,
+ "learning_rate": 0.0006,
+ "loss": 4.30778694152832,
+ "step": 1520
+ },
+ {
+ "epoch": 21.125764192139737,
+ "grad_norm": 0.02173588052392006,
+ "learning_rate": 0.0006,
+ "loss": 4.433149814605713,
+ "step": 1521
+ },
+ {
+ "epoch": 21.139737991266376,
+ "grad_norm": 0.020950648933649063,
+ "learning_rate": 0.0006,
+ "loss": 4.520089149475098,
+ "step": 1522
+ },
+ {
+ "epoch": 21.153711790393015,
+ "grad_norm": 0.020684752613306046,
+ "learning_rate": 0.0006,
+ "loss": 4.372437477111816,
+ "step": 1523
+ },
+ {
+ "epoch": 21.16768558951965,
+ "grad_norm": 0.019472869113087654,
+ "learning_rate": 0.0006,
+ "loss": 4.437015533447266,
+ "step": 1524
+ },
+ {
+ "epoch": 21.18165938864629,
+ "grad_norm": 0.01846962980926037,
+ "learning_rate": 0.0006,
+ "loss": 4.449531555175781,
+ "step": 1525
+ },
+ {
+ "epoch": 21.195633187772927,
+ "grad_norm": 0.017377803102135658,
+ "learning_rate": 0.0006,
+ "loss": 4.458279132843018,
+ "step": 1526
+ },
+ {
+ "epoch": 21.209606986899562,
+ "grad_norm": 0.016002200543880463,
+ "learning_rate": 0.0006,
+ "loss": 4.266297817230225,
+ "step": 1527
+ },
+ {
+ "epoch": 21.2235807860262,
+ "grad_norm": 0.015575903467833996,
+ "learning_rate": 0.0006,
+ "loss": 4.342632293701172,
+ "step": 1528
+ },
+ {
+ "epoch": 21.23755458515284,
+ "grad_norm": 0.015634698793292046,
+ "learning_rate": 0.0006,
+ "loss": 4.338623523712158,
+ "step": 1529
+ },
+ {
+ "epoch": 21.251528384279474,
+ "grad_norm": 0.01449581328779459,
+ "learning_rate": 0.0006,
+ "loss": 4.354330062866211,
+ "step": 1530
+ },
+ {
+ "epoch": 21.265502183406113,
+ "grad_norm": 0.01487616915255785,
+ "learning_rate": 0.0006,
+ "loss": 4.486835479736328,
+ "step": 1531
+ },
+ {
+ "epoch": 21.27947598253275,
+ "grad_norm": 0.016346555203199387,
+ "learning_rate": 0.0006,
+ "loss": 4.407098770141602,
+ "step": 1532
+ },
+ {
+ "epoch": 21.29344978165939,
+ "grad_norm": 0.01682182401418686,
+ "learning_rate": 0.0006,
+ "loss": 4.436809539794922,
+ "step": 1533
+ },
+ {
+ "epoch": 21.307423580786025,
+ "grad_norm": 0.017704004421830177,
+ "learning_rate": 0.0006,
+ "loss": 4.3854780197143555,
+ "step": 1534
+ },
+ {
+ "epoch": 21.321397379912664,
+ "grad_norm": 0.018437420949339867,
+ "learning_rate": 0.0006,
+ "loss": 4.413283824920654,
+ "step": 1535
+ },
+ {
+ "epoch": 21.335371179039303,
+ "grad_norm": 0.016911081969738007,
+ "learning_rate": 0.0006,
+ "loss": 4.357306003570557,
+ "step": 1536
+ },
+ {
+ "epoch": 21.349344978165938,
+ "grad_norm": 0.01642223633825779,
+ "learning_rate": 0.0006,
+ "loss": 4.422322750091553,
+ "step": 1537
+ },
+ {
+ "epoch": 21.363318777292577,
+ "grad_norm": 0.01780877821147442,
+ "learning_rate": 0.0006,
+ "loss": 4.387620449066162,
+ "step": 1538
+ },
+ {
+ "epoch": 21.377292576419215,
+ "grad_norm": 0.01653975248336792,
+ "learning_rate": 0.0006,
+ "loss": 4.327267646789551,
+ "step": 1539
+ },
+ {
+ "epoch": 21.39126637554585,
+ "grad_norm": 0.01572156324982643,
+ "learning_rate": 0.0006,
+ "loss": 4.475979804992676,
+ "step": 1540
+ },
+ {
+ "epoch": 21.40524017467249,
+ "grad_norm": 0.017156066372990608,
+ "learning_rate": 0.0006,
+ "loss": 4.4011125564575195,
+ "step": 1541
+ },
+ {
+ "epoch": 21.419213973799128,
+ "grad_norm": 0.01788829080760479,
+ "learning_rate": 0.0006,
+ "loss": 4.4174089431762695,
+ "step": 1542
+ },
+ {
+ "epoch": 21.433187772925763,
+ "grad_norm": 0.01728985086083412,
+ "learning_rate": 0.0006,
+ "loss": 4.429244041442871,
+ "step": 1543
+ },
+ {
+ "epoch": 21.4471615720524,
+ "grad_norm": 0.016534466296434402,
+ "learning_rate": 0.0006,
+ "loss": 4.326068878173828,
+ "step": 1544
+ },
+ {
+ "epoch": 21.46113537117904,
+ "grad_norm": 0.016582584008574486,
+ "learning_rate": 0.0006,
+ "loss": 4.2975969314575195,
+ "step": 1545
+ },
+ {
+ "epoch": 21.475109170305675,
+ "grad_norm": 0.018034178763628006,
+ "learning_rate": 0.0006,
+ "loss": 4.501319408416748,
+ "step": 1546
+ },
+ {
+ "epoch": 21.489082969432314,
+ "grad_norm": 0.017827702686190605,
+ "learning_rate": 0.0006,
+ "loss": 4.502284526824951,
+ "step": 1547
+ },
+ {
+ "epoch": 21.503056768558952,
+ "grad_norm": 0.017182299867272377,
+ "learning_rate": 0.0006,
+ "loss": 4.380828857421875,
+ "step": 1548
+ },
+ {
+ "epoch": 21.51703056768559,
+ "grad_norm": 0.019101588055491447,
+ "learning_rate": 0.0006,
+ "loss": 4.31564998626709,
+ "step": 1549
+ },
+ {
+ "epoch": 21.531004366812226,
+ "grad_norm": 0.0204677302390337,
+ "learning_rate": 0.0006,
+ "loss": 4.3840131759643555,
+ "step": 1550
+ },
+ {
+ "epoch": 21.544978165938865,
+ "grad_norm": 0.0188993439078331,
+ "learning_rate": 0.0006,
+ "loss": 4.4257731437683105,
+ "step": 1551
+ },
+ {
+ "epoch": 21.558951965065503,
+ "grad_norm": 0.01730230636894703,
+ "learning_rate": 0.0006,
+ "loss": 4.445132255554199,
+ "step": 1552
+ },
+ {
+ "epoch": 21.57292576419214,
+ "grad_norm": 0.01801574043929577,
+ "learning_rate": 0.0006,
+ "loss": 4.427064895629883,
+ "step": 1553
+ },
+ {
+ "epoch": 21.586899563318777,
+ "grad_norm": 0.018417565152049065,
+ "learning_rate": 0.0006,
+ "loss": 4.319595813751221,
+ "step": 1554
+ },
+ {
+ "epoch": 21.600873362445416,
+ "grad_norm": 0.01683160290122032,
+ "learning_rate": 0.0006,
+ "loss": 4.349218368530273,
+ "step": 1555
+ },
+ {
+ "epoch": 21.61484716157205,
+ "grad_norm": 0.016973106190562248,
+ "learning_rate": 0.0006,
+ "loss": 4.375990390777588,
+ "step": 1556
+ },
+ {
+ "epoch": 21.62882096069869,
+ "grad_norm": 0.015688583254814148,
+ "learning_rate": 0.0006,
+ "loss": 4.408170223236084,
+ "step": 1557
+ },
+ {
+ "epoch": 21.64279475982533,
+ "grad_norm": 0.01573043316602707,
+ "learning_rate": 0.0006,
+ "loss": 4.416255950927734,
+ "step": 1558
+ },
+ {
+ "epoch": 21.656768558951963,
+ "grad_norm": 0.01744293048977852,
+ "learning_rate": 0.0006,
+ "loss": 4.477141380310059,
+ "step": 1559
+ },
+ {
+ "epoch": 21.670742358078602,
+ "grad_norm": 0.017503513023257256,
+ "learning_rate": 0.0006,
+ "loss": 4.3367486000061035,
+ "step": 1560
+ },
+ {
+ "epoch": 21.68471615720524,
+ "grad_norm": 0.016592150554060936,
+ "learning_rate": 0.0006,
+ "loss": 4.314785480499268,
+ "step": 1561
+ },
+ {
+ "epoch": 21.69868995633188,
+ "grad_norm": 0.015774084255099297,
+ "learning_rate": 0.0006,
+ "loss": 4.448209285736084,
+ "step": 1562
+ },
+ {
+ "epoch": 21.712663755458514,
+ "grad_norm": 0.015779150649905205,
+ "learning_rate": 0.0006,
+ "loss": 4.392690658569336,
+ "step": 1563
+ },
+ {
+ "epoch": 21.726637554585153,
+ "grad_norm": 0.015392648056149483,
+ "learning_rate": 0.0006,
+ "loss": 4.437741756439209,
+ "step": 1564
+ },
+ {
+ "epoch": 21.74061135371179,
+ "grad_norm": 0.01530836895108223,
+ "learning_rate": 0.0006,
+ "loss": 4.424595832824707,
+ "step": 1565
+ },
+ {
+ "epoch": 21.754585152838427,
+ "grad_norm": 0.015600779093801975,
+ "learning_rate": 0.0006,
+ "loss": 4.335417747497559,
+ "step": 1566
+ },
+ {
+ "epoch": 21.768558951965066,
+ "grad_norm": 0.014646518044173717,
+ "learning_rate": 0.0006,
+ "loss": 4.279081344604492,
+ "step": 1567
+ },
+ {
+ "epoch": 21.782532751091704,
+ "grad_norm": 0.014132864773273468,
+ "learning_rate": 0.0006,
+ "loss": 4.345950126647949,
+ "step": 1568
+ },
+ {
+ "epoch": 21.79650655021834,
+ "grad_norm": 0.01425724383443594,
+ "learning_rate": 0.0006,
+ "loss": 4.277588844299316,
+ "step": 1569
+ },
+ {
+ "epoch": 21.810480349344978,
+ "grad_norm": 0.015396572649478912,
+ "learning_rate": 0.0006,
+ "loss": 4.326196670532227,
+ "step": 1570
+ },
+ {
+ "epoch": 21.824454148471617,
+ "grad_norm": 0.01683669537305832,
+ "learning_rate": 0.0006,
+ "loss": 4.3294172286987305,
+ "step": 1571
+ },
+ {
+ "epoch": 21.83842794759825,
+ "grad_norm": 0.016365647315979004,
+ "learning_rate": 0.0006,
+ "loss": 4.523321628570557,
+ "step": 1572
+ },
+ {
+ "epoch": 21.85240174672489,
+ "grad_norm": 0.017142774537205696,
+ "learning_rate": 0.0006,
+ "loss": 4.475842475891113,
+ "step": 1573
+ },
+ {
+ "epoch": 21.86637554585153,
+ "grad_norm": 0.015950839966535568,
+ "learning_rate": 0.0006,
+ "loss": 4.397187232971191,
+ "step": 1574
+ },
+ {
+ "epoch": 21.880349344978168,
+ "grad_norm": 0.016180362552404404,
+ "learning_rate": 0.0006,
+ "loss": 4.414778709411621,
+ "step": 1575
+ },
+ {
+ "epoch": 21.894323144104803,
+ "grad_norm": 0.017615556716918945,
+ "learning_rate": 0.0006,
+ "loss": 4.348567962646484,
+ "step": 1576
+ },
+ {
+ "epoch": 21.90829694323144,
+ "grad_norm": 0.01896277628839016,
+ "learning_rate": 0.0006,
+ "loss": 4.431762218475342,
+ "step": 1577
+ },
+ {
+ "epoch": 21.92227074235808,
+ "grad_norm": 0.01990971714258194,
+ "learning_rate": 0.0006,
+ "loss": 4.431285858154297,
+ "step": 1578
+ },
+ {
+ "epoch": 21.936244541484715,
+ "grad_norm": 0.0201258547604084,
+ "learning_rate": 0.0006,
+ "loss": 4.309872627258301,
+ "step": 1579
+ },
+ {
+ "epoch": 21.950218340611354,
+ "grad_norm": 0.019620198756456375,
+ "learning_rate": 0.0006,
+ "loss": 4.484044551849365,
+ "step": 1580
+ },
+ {
+ "epoch": 21.964192139737992,
+ "grad_norm": 0.01771123707294464,
+ "learning_rate": 0.0006,
+ "loss": 4.368093013763428,
+ "step": 1581
+ },
+ {
+ "epoch": 21.978165938864628,
+ "grad_norm": 0.017501654103398323,
+ "learning_rate": 0.0006,
+ "loss": 4.372651100158691,
+ "step": 1582
+ },
+ {
+ "epoch": 21.992139737991266,
+ "grad_norm": 0.01693608984351158,
+ "learning_rate": 0.0006,
+ "loss": 4.379167556762695,
+ "step": 1583
+ },
+ {
+ "epoch": 22.0,
+ "grad_norm": 0.02002849616110325,
+ "learning_rate": 0.0006,
+ "loss": 4.329627990722656,
+ "step": 1584
+ },
+ {
+ "epoch": 22.0,
+ "eval_loss": 4.723696231842041,
+ "eval_runtime": 56.1758,
+ "eval_samples_per_second": 43.471,
+ "eval_steps_per_second": 1.371,
+ "step": 1584
+ },
+ {
+ "epoch": 22.01397379912664,
+ "grad_norm": 0.02030305378139019,
+ "learning_rate": 0.0006,
+ "loss": 4.355566024780273,
+ "step": 1585
+ },
+ {
+ "epoch": 22.027947598253274,
+ "grad_norm": 0.018638934940099716,
+ "learning_rate": 0.0006,
+ "loss": 4.314009189605713,
+ "step": 1586
+ },
+ {
+ "epoch": 22.041921397379912,
+ "grad_norm": 0.019917072728276253,
+ "learning_rate": 0.0006,
+ "loss": 4.367525577545166,
+ "step": 1587
+ },
+ {
+ "epoch": 22.05589519650655,
+ "grad_norm": 0.020785167813301086,
+ "learning_rate": 0.0006,
+ "loss": 4.375561714172363,
+ "step": 1588
+ },
+ {
+ "epoch": 22.069868995633186,
+ "grad_norm": 0.021214453503489494,
+ "learning_rate": 0.0006,
+ "loss": 4.418305397033691,
+ "step": 1589
+ },
+ {
+ "epoch": 22.083842794759825,
+ "grad_norm": 0.019051678478717804,
+ "learning_rate": 0.0006,
+ "loss": 4.215126991271973,
+ "step": 1590
+ },
+ {
+ "epoch": 22.097816593886463,
+ "grad_norm": 0.02323366142809391,
+ "learning_rate": 0.0006,
+ "loss": 4.306807994842529,
+ "step": 1591
+ },
+ {
+ "epoch": 22.111790393013102,
+ "grad_norm": 0.026186738163232803,
+ "learning_rate": 0.0006,
+ "loss": 4.408508777618408,
+ "step": 1592
+ },
+ {
+ "epoch": 22.125764192139737,
+ "grad_norm": 0.027361227199435234,
+ "learning_rate": 0.0006,
+ "loss": 4.329322814941406,
+ "step": 1593
+ },
+ {
+ "epoch": 22.139737991266376,
+ "grad_norm": 0.026564646512269974,
+ "learning_rate": 0.0006,
+ "loss": 4.346642017364502,
+ "step": 1594
+ },
+ {
+ "epoch": 22.153711790393015,
+ "grad_norm": 0.025073667988181114,
+ "learning_rate": 0.0006,
+ "loss": 4.3984198570251465,
+ "step": 1595
+ },
+ {
+ "epoch": 22.16768558951965,
+ "grad_norm": 0.02347794733941555,
+ "learning_rate": 0.0006,
+ "loss": 4.352538108825684,
+ "step": 1596
+ },
+ {
+ "epoch": 22.18165938864629,
+ "grad_norm": 0.02203468047082424,
+ "learning_rate": 0.0006,
+ "loss": 4.298148155212402,
+ "step": 1597
+ },
+ {
+ "epoch": 22.195633187772927,
+ "grad_norm": 0.020621957257390022,
+ "learning_rate": 0.0006,
+ "loss": 4.480309963226318,
+ "step": 1598
+ },
+ {
+ "epoch": 22.209606986899562,
+ "grad_norm": 0.023710055276751518,
+ "learning_rate": 0.0006,
+ "loss": 4.339317321777344,
+ "step": 1599
+ },
+ {
+ "epoch": 22.2235807860262,
+ "grad_norm": 0.0233482476323843,
+ "learning_rate": 0.0006,
+ "loss": 4.436896324157715,
+ "step": 1600
+ },
+ {
+ "epoch": 22.23755458515284,
+ "grad_norm": 0.020122403278946877,
+ "learning_rate": 0.0006,
+ "loss": 4.236002445220947,
+ "step": 1601
+ },
+ {
+ "epoch": 22.251528384279474,
+ "grad_norm": 0.01957266591489315,
+ "learning_rate": 0.0006,
+ "loss": 4.497328281402588,
+ "step": 1602
+ },
+ {
+ "epoch": 22.265502183406113,
+ "grad_norm": 0.018964895978569984,
+ "learning_rate": 0.0006,
+ "loss": 4.36360502243042,
+ "step": 1603
+ },
+ {
+ "epoch": 22.27947598253275,
+ "grad_norm": 0.018902769312262535,
+ "learning_rate": 0.0006,
+ "loss": 4.427093505859375,
+ "step": 1604
+ },
+ {
+ "epoch": 22.29344978165939,
+ "grad_norm": 0.018918119370937347,
+ "learning_rate": 0.0006,
+ "loss": 4.294950485229492,
+ "step": 1605
+ },
+ {
+ "epoch": 22.307423580786025,
+ "grad_norm": 0.019315775483846664,
+ "learning_rate": 0.0006,
+ "loss": 4.427952289581299,
+ "step": 1606
+ },
+ {
+ "epoch": 22.321397379912664,
+ "grad_norm": 0.018350522965192795,
+ "learning_rate": 0.0006,
+ "loss": 4.272945404052734,
+ "step": 1607
+ },
+ {
+ "epoch": 22.335371179039303,
+ "grad_norm": 0.01759534701704979,
+ "learning_rate": 0.0006,
+ "loss": 4.348023891448975,
+ "step": 1608
+ },
+ {
+ "epoch": 22.349344978165938,
+ "grad_norm": 0.01646057888865471,
+ "learning_rate": 0.0006,
+ "loss": 4.278518199920654,
+ "step": 1609
+ },
+ {
+ "epoch": 22.363318777292577,
+ "grad_norm": 0.015875060111284256,
+ "learning_rate": 0.0006,
+ "loss": 4.423527240753174,
+ "step": 1610
+ },
+ {
+ "epoch": 22.377292576419215,
+ "grad_norm": 0.01638529635965824,
+ "learning_rate": 0.0006,
+ "loss": 4.321723937988281,
+ "step": 1611
+ },
+ {
+ "epoch": 22.39126637554585,
+ "grad_norm": 0.016942061483860016,
+ "learning_rate": 0.0006,
+ "loss": 4.346164703369141,
+ "step": 1612
+ },
+ {
+ "epoch": 22.40524017467249,
+ "grad_norm": 0.015209621749818325,
+ "learning_rate": 0.0006,
+ "loss": 4.434340000152588,
+ "step": 1613
+ },
+ {
+ "epoch": 22.419213973799128,
+ "grad_norm": 0.014423009008169174,
+ "learning_rate": 0.0006,
+ "loss": 4.342560768127441,
+ "step": 1614
+ },
+ {
+ "epoch": 22.433187772925763,
+ "grad_norm": 0.014369525946676731,
+ "learning_rate": 0.0006,
+ "loss": 4.296586990356445,
+ "step": 1615
+ },
+ {
+ "epoch": 22.4471615720524,
+ "grad_norm": 0.01392540242522955,
+ "learning_rate": 0.0006,
+ "loss": 4.433176040649414,
+ "step": 1616
+ },
+ {
+ "epoch": 22.46113537117904,
+ "grad_norm": 0.013206899166107178,
+ "learning_rate": 0.0006,
+ "loss": 4.32962703704834,
+ "step": 1617
+ },
+ {
+ "epoch": 22.475109170305675,
+ "grad_norm": 0.013427077792584896,
+ "learning_rate": 0.0006,
+ "loss": 4.420437335968018,
+ "step": 1618
+ },
+ {
+ "epoch": 22.489082969432314,
+ "grad_norm": 0.014194196090102196,
+ "learning_rate": 0.0006,
+ "loss": 4.410478591918945,
+ "step": 1619
+ },
+ {
+ "epoch": 22.503056768558952,
+ "grad_norm": 0.01362364087253809,
+ "learning_rate": 0.0006,
+ "loss": 4.338932037353516,
+ "step": 1620
+ },
+ {
+ "epoch": 22.51703056768559,
+ "grad_norm": 0.014299589209258556,
+ "learning_rate": 0.0006,
+ "loss": 4.34089994430542,
+ "step": 1621
+ },
+ {
+ "epoch": 22.531004366812226,
+ "grad_norm": 0.014061295427381992,
+ "learning_rate": 0.0006,
+ "loss": 4.29201602935791,
+ "step": 1622
+ },
+ {
+ "epoch": 22.544978165938865,
+ "grad_norm": 0.013945229351520538,
+ "learning_rate": 0.0006,
+ "loss": 4.326052188873291,
+ "step": 1623
+ },
+ {
+ "epoch": 22.558951965065503,
+ "grad_norm": 0.01387068536132574,
+ "learning_rate": 0.0006,
+ "loss": 4.393521308898926,
+ "step": 1624
+ },
+ {
+ "epoch": 22.57292576419214,
+ "grad_norm": 0.013931898400187492,
+ "learning_rate": 0.0006,
+ "loss": 4.3739752769470215,
+ "step": 1625
+ },
+ {
+ "epoch": 22.586899563318777,
+ "grad_norm": 0.01497740764170885,
+ "learning_rate": 0.0006,
+ "loss": 4.3991618156433105,
+ "step": 1626
+ },
+ {
+ "epoch": 22.600873362445416,
+ "grad_norm": 0.015842905268073082,
+ "learning_rate": 0.0006,
+ "loss": 4.405340671539307,
+ "step": 1627
+ },
+ {
+ "epoch": 22.61484716157205,
+ "grad_norm": 0.01580633595585823,
+ "learning_rate": 0.0006,
+ "loss": 4.280102729797363,
+ "step": 1628
+ },
+ {
+ "epoch": 22.62882096069869,
+ "grad_norm": 0.016567382961511612,
+ "learning_rate": 0.0006,
+ "loss": 4.407657146453857,
+ "step": 1629
+ },
+ {
+ "epoch": 22.64279475982533,
+ "grad_norm": 0.01648622378706932,
+ "learning_rate": 0.0006,
+ "loss": 4.392590522766113,
+ "step": 1630
+ },
+ {
+ "epoch": 22.656768558951963,
+ "grad_norm": 0.017066307365894318,
+ "learning_rate": 0.0006,
+ "loss": 4.324862003326416,
+ "step": 1631
+ },
+ {
+ "epoch": 22.670742358078602,
+ "grad_norm": 0.01637062057852745,
+ "learning_rate": 0.0006,
+ "loss": 4.34598445892334,
+ "step": 1632
+ },
+ {
+ "epoch": 22.68471615720524,
+ "grad_norm": 0.018038960173726082,
+ "learning_rate": 0.0006,
+ "loss": 4.320711612701416,
+ "step": 1633
+ },
+ {
+ "epoch": 22.69868995633188,
+ "grad_norm": 0.018096428364515305,
+ "learning_rate": 0.0006,
+ "loss": 4.536527156829834,
+ "step": 1634
+ },
+ {
+ "epoch": 22.712663755458514,
+ "grad_norm": 0.017889291048049927,
+ "learning_rate": 0.0006,
+ "loss": 4.388131141662598,
+ "step": 1635
+ },
+ {
+ "epoch": 22.726637554585153,
+ "grad_norm": 0.018771521747112274,
+ "learning_rate": 0.0006,
+ "loss": 4.357668876647949,
+ "step": 1636
+ },
+ {
+ "epoch": 22.74061135371179,
+ "grad_norm": 0.02100873365998268,
+ "learning_rate": 0.0006,
+ "loss": 4.293972015380859,
+ "step": 1637
+ },
+ {
+ "epoch": 22.754585152838427,
+ "grad_norm": 0.017715785652399063,
+ "learning_rate": 0.0006,
+ "loss": 4.349849224090576,
+ "step": 1638
+ },
+ {
+ "epoch": 22.768558951965066,
+ "grad_norm": 0.015546333976089954,
+ "learning_rate": 0.0006,
+ "loss": 4.311728000640869,
+ "step": 1639
+ },
+ {
+ "epoch": 22.782532751091704,
+ "grad_norm": 0.01605990156531334,
+ "learning_rate": 0.0006,
+ "loss": 4.499123573303223,
+ "step": 1640
+ },
+ {
+ "epoch": 22.79650655021834,
+ "grad_norm": 0.016642142087221146,
+ "learning_rate": 0.0006,
+ "loss": 4.4319257736206055,
+ "step": 1641
+ },
+ {
+ "epoch": 22.810480349344978,
+ "grad_norm": 0.016120387241244316,
+ "learning_rate": 0.0006,
+ "loss": 4.4373650550842285,
+ "step": 1642
+ },
+ {
+ "epoch": 22.824454148471617,
+ "grad_norm": 0.015174890868365765,
+ "learning_rate": 0.0006,
+ "loss": 4.314082145690918,
+ "step": 1643
+ },
+ {
+ "epoch": 22.83842794759825,
+ "grad_norm": 0.01551714539527893,
+ "learning_rate": 0.0006,
+ "loss": 4.360363960266113,
+ "step": 1644
+ },
+ {
+ "epoch": 22.85240174672489,
+ "grad_norm": 0.014966742135584354,
+ "learning_rate": 0.0006,
+ "loss": 4.346253395080566,
+ "step": 1645
+ },
+ {
+ "epoch": 22.86637554585153,
+ "grad_norm": 0.014288028702139854,
+ "learning_rate": 0.0006,
+ "loss": 4.464366436004639,
+ "step": 1646
+ },
+ {
+ "epoch": 22.880349344978168,
+ "grad_norm": 0.015291682444512844,
+ "learning_rate": 0.0006,
+ "loss": 4.504101753234863,
+ "step": 1647
+ },
+ {
+ "epoch": 22.894323144104803,
+ "grad_norm": 0.014534426853060722,
+ "learning_rate": 0.0006,
+ "loss": 4.462172031402588,
+ "step": 1648
+ },
+ {
+ "epoch": 22.90829694323144,
+ "grad_norm": 0.014129819348454475,
+ "learning_rate": 0.0006,
+ "loss": 4.350126266479492,
+ "step": 1649
+ },
+ {
+ "epoch": 22.92227074235808,
+ "grad_norm": 0.014932668767869473,
+ "learning_rate": 0.0006,
+ "loss": 4.374392032623291,
+ "step": 1650
+ },
+ {
+ "epoch": 22.936244541484715,
+ "grad_norm": 0.014647853560745716,
+ "learning_rate": 0.0006,
+ "loss": 4.252264499664307,
+ "step": 1651
+ },
+ {
+ "epoch": 22.950218340611354,
+ "grad_norm": 0.013553360477089882,
+ "learning_rate": 0.0006,
+ "loss": 4.379729270935059,
+ "step": 1652
+ },
+ {
+ "epoch": 22.964192139737992,
+ "grad_norm": 0.014606194570660591,
+ "learning_rate": 0.0006,
+ "loss": 4.363372802734375,
+ "step": 1653
+ },
+ {
+ "epoch": 22.978165938864628,
+ "grad_norm": 0.015000631101429462,
+ "learning_rate": 0.0006,
+ "loss": 4.408048629760742,
+ "step": 1654
+ },
+ {
+ "epoch": 22.992139737991266,
+ "grad_norm": 0.015370816923677921,
+ "learning_rate": 0.0006,
+ "loss": 4.297451019287109,
+ "step": 1655
+ },
+ {
+ "epoch": 23.0,
+ "grad_norm": 0.01709144562482834,
+ "learning_rate": 0.0006,
+ "loss": 4.294776916503906,
+ "step": 1656
+ },
+ {
+ "epoch": 23.0,
+ "eval_loss": 4.678618907928467,
+ "eval_runtime": 56.1668,
+ "eval_samples_per_second": 43.478,
+ "eval_steps_per_second": 1.371,
+ "step": 1656
+ },
+ {
+ "epoch": 23.01397379912664,
+ "grad_norm": 0.01727290451526642,
+ "learning_rate": 0.0006,
+ "loss": 4.311701774597168,
+ "step": 1657
+ },
+ {
+ "epoch": 23.027947598253274,
+ "grad_norm": 0.020645759999752045,
+ "learning_rate": 0.0006,
+ "loss": 4.429897308349609,
+ "step": 1658
+ },
+ {
+ "epoch": 23.041921397379912,
+ "grad_norm": 0.02204226516187191,
+ "learning_rate": 0.0006,
+ "loss": 4.365151882171631,
+ "step": 1659
+ },
+ {
+ "epoch": 23.05589519650655,
+ "grad_norm": 0.023069269955158234,
+ "learning_rate": 0.0006,
+ "loss": 4.3710832595825195,
+ "step": 1660
+ },
+ {
+ "epoch": 23.069868995633186,
+ "grad_norm": 0.024151179939508438,
+ "learning_rate": 0.0006,
+ "loss": 4.403108596801758,
+ "step": 1661
+ },
+ {
+ "epoch": 23.083842794759825,
+ "grad_norm": 0.02511233650147915,
+ "learning_rate": 0.0006,
+ "loss": 4.414128303527832,
+ "step": 1662
+ },
+ {
+ "epoch": 23.097816593886463,
+ "grad_norm": 0.025802113115787506,
+ "learning_rate": 0.0006,
+ "loss": 4.388265609741211,
+ "step": 1663
+ },
+ {
+ "epoch": 23.111790393013102,
+ "grad_norm": 0.02377348765730858,
+ "learning_rate": 0.0006,
+ "loss": 4.357505798339844,
+ "step": 1664
+ },
+ {
+ "epoch": 23.125764192139737,
+ "grad_norm": 0.020876651629805565,
+ "learning_rate": 0.0006,
+ "loss": 4.23043966293335,
+ "step": 1665
+ },
+ {
+ "epoch": 23.139737991266376,
+ "grad_norm": 0.01959727331995964,
+ "learning_rate": 0.0006,
+ "loss": 4.280970096588135,
+ "step": 1666
+ },
+ {
+ "epoch": 23.153711790393015,
+ "grad_norm": 0.018678616732358932,
+ "learning_rate": 0.0006,
+ "loss": 4.346716403961182,
+ "step": 1667
+ },
+ {
+ "epoch": 23.16768558951965,
+ "grad_norm": 0.02108944021165371,
+ "learning_rate": 0.0006,
+ "loss": 4.277904510498047,
+ "step": 1668
+ },
+ {
+ "epoch": 23.18165938864629,
+ "grad_norm": 0.020221566781401634,
+ "learning_rate": 0.0006,
+ "loss": 4.367693901062012,
+ "step": 1669
+ },
+ {
+ "epoch": 23.195633187772927,
+ "grad_norm": 0.019599711522459984,
+ "learning_rate": 0.0006,
+ "loss": 4.257092475891113,
+ "step": 1670
+ },
+ {
+ "epoch": 23.209606986899562,
+ "grad_norm": 0.019833361729979515,
+ "learning_rate": 0.0006,
+ "loss": 4.540737628936768,
+ "step": 1671
+ },
+ {
+ "epoch": 23.2235807860262,
+ "grad_norm": 0.020352095365524292,
+ "learning_rate": 0.0006,
+ "loss": 4.360923767089844,
+ "step": 1672
+ },
+ {
+ "epoch": 23.23755458515284,
+ "grad_norm": 0.01939748041331768,
+ "learning_rate": 0.0006,
+ "loss": 4.393110275268555,
+ "step": 1673
+ },
+ {
+ "epoch": 23.251528384279474,
+ "grad_norm": 0.01884354278445244,
+ "learning_rate": 0.0006,
+ "loss": 4.383938789367676,
+ "step": 1674
+ },
+ {
+ "epoch": 23.265502183406113,
+ "grad_norm": 0.02008470520377159,
+ "learning_rate": 0.0006,
+ "loss": 4.346344947814941,
+ "step": 1675
+ },
+ {
+ "epoch": 23.27947598253275,
+ "grad_norm": 0.019065195694565773,
+ "learning_rate": 0.0006,
+ "loss": 4.36611795425415,
+ "step": 1676
+ },
+ {
+ "epoch": 23.29344978165939,
+ "grad_norm": 0.01938541606068611,
+ "learning_rate": 0.0006,
+ "loss": 4.4009599685668945,
+ "step": 1677
+ },
+ {
+ "epoch": 23.307423580786025,
+ "grad_norm": 0.019044723361730576,
+ "learning_rate": 0.0006,
+ "loss": 4.340169906616211,
+ "step": 1678
+ },
+ {
+ "epoch": 23.321397379912664,
+ "grad_norm": 0.018850378692150116,
+ "learning_rate": 0.0006,
+ "loss": 4.37674617767334,
+ "step": 1679
+ },
+ {
+ "epoch": 23.335371179039303,
+ "grad_norm": 0.019076095893979073,
+ "learning_rate": 0.0006,
+ "loss": 4.226874351501465,
+ "step": 1680
+ },
+ {
+ "epoch": 23.349344978165938,
+ "grad_norm": 0.017543038353323936,
+ "learning_rate": 0.0006,
+ "loss": 4.337751388549805,
+ "step": 1681
+ },
+ {
+ "epoch": 23.363318777292577,
+ "grad_norm": 0.017170218750834465,
+ "learning_rate": 0.0006,
+ "loss": 4.282512664794922,
+ "step": 1682
+ },
+ {
+ "epoch": 23.377292576419215,
+ "grad_norm": 0.01574796624481678,
+ "learning_rate": 0.0006,
+ "loss": 4.30638313293457,
+ "step": 1683
+ },
+ {
+ "epoch": 23.39126637554585,
+ "grad_norm": 0.015160846523940563,
+ "learning_rate": 0.0006,
+ "loss": 4.336716175079346,
+ "step": 1684
+ },
+ {
+ "epoch": 23.40524017467249,
+ "grad_norm": 0.01534526702016592,
+ "learning_rate": 0.0006,
+ "loss": 4.2174787521362305,
+ "step": 1685
+ },
+ {
+ "epoch": 23.419213973799128,
+ "grad_norm": 0.015289999544620514,
+ "learning_rate": 0.0006,
+ "loss": 4.272546291351318,
+ "step": 1686
+ },
+ {
+ "epoch": 23.433187772925763,
+ "grad_norm": 0.014894185587763786,
+ "learning_rate": 0.0006,
+ "loss": 4.296045303344727,
+ "step": 1687
+ },
+ {
+ "epoch": 23.4471615720524,
+ "grad_norm": 0.014869502745568752,
+ "learning_rate": 0.0006,
+ "loss": 4.425461769104004,
+ "step": 1688
+ },
+ {
+ "epoch": 23.46113537117904,
+ "grad_norm": 0.01474483497440815,
+ "learning_rate": 0.0006,
+ "loss": 4.361867427825928,
+ "step": 1689
+ },
+ {
+ "epoch": 23.475109170305675,
+ "grad_norm": 0.015299245715141296,
+ "learning_rate": 0.0006,
+ "loss": 4.364523410797119,
+ "step": 1690
+ },
+ {
+ "epoch": 23.489082969432314,
+ "grad_norm": 0.015201340429484844,
+ "learning_rate": 0.0006,
+ "loss": 4.285182952880859,
+ "step": 1691
+ },
+ {
+ "epoch": 23.503056768558952,
+ "grad_norm": 0.016057293862104416,
+ "learning_rate": 0.0006,
+ "loss": 4.284873962402344,
+ "step": 1692
+ },
+ {
+ "epoch": 23.51703056768559,
+ "grad_norm": 0.015871062874794006,
+ "learning_rate": 0.0006,
+ "loss": 4.326470851898193,
+ "step": 1693
+ },
+ {
+ "epoch": 23.531004366812226,
+ "grad_norm": 0.01586288772523403,
+ "learning_rate": 0.0006,
+ "loss": 4.324883460998535,
+ "step": 1694
+ },
+ {
+ "epoch": 23.544978165938865,
+ "grad_norm": 0.016092827543616295,
+ "learning_rate": 0.0006,
+ "loss": 4.3258867263793945,
+ "step": 1695
+ },
+ {
+ "epoch": 23.558951965065503,
+ "grad_norm": 0.01582016795873642,
+ "learning_rate": 0.0006,
+ "loss": 4.358892440795898,
+ "step": 1696
+ },
+ {
+ "epoch": 23.57292576419214,
+ "grad_norm": 0.01688452623784542,
+ "learning_rate": 0.0006,
+ "loss": 4.280842304229736,
+ "step": 1697
+ },
+ {
+ "epoch": 23.586899563318777,
+ "grad_norm": 0.015865886583924294,
+ "learning_rate": 0.0006,
+ "loss": 4.245567321777344,
+ "step": 1698
+ },
+ {
+ "epoch": 23.600873362445416,
+ "grad_norm": 0.015303662046790123,
+ "learning_rate": 0.0006,
+ "loss": 4.383112907409668,
+ "step": 1699
+ },
+ {
+ "epoch": 23.61484716157205,
+ "grad_norm": 0.015649516135454178,
+ "learning_rate": 0.0006,
+ "loss": 4.519082069396973,
+ "step": 1700
+ },
+ {
+ "epoch": 23.62882096069869,
+ "grad_norm": 0.0167120099067688,
+ "learning_rate": 0.0006,
+ "loss": 4.447847366333008,
+ "step": 1701
+ },
+ {
+ "epoch": 23.64279475982533,
+ "grad_norm": 0.016368716955184937,
+ "learning_rate": 0.0006,
+ "loss": 4.434186935424805,
+ "step": 1702
+ },
+ {
+ "epoch": 23.656768558951963,
+ "grad_norm": 0.017379503697156906,
+ "learning_rate": 0.0006,
+ "loss": 4.317594528198242,
+ "step": 1703
+ },
+ {
+ "epoch": 23.670742358078602,
+ "grad_norm": 0.018573811277747154,
+ "learning_rate": 0.0006,
+ "loss": 4.205793857574463,
+ "step": 1704
+ },
+ {
+ "epoch": 23.68471615720524,
+ "grad_norm": 0.01749090477824211,
+ "learning_rate": 0.0006,
+ "loss": 4.3985700607299805,
+ "step": 1705
+ },
+ {
+ "epoch": 23.69868995633188,
+ "grad_norm": 0.0165668074041605,
+ "learning_rate": 0.0006,
+ "loss": 4.329226970672607,
+ "step": 1706
+ },
+ {
+ "epoch": 23.712663755458514,
+ "grad_norm": 0.016913846135139465,
+ "learning_rate": 0.0006,
+ "loss": 4.388485908508301,
+ "step": 1707
+ },
+ {
+ "epoch": 23.726637554585153,
+ "grad_norm": 0.016813859343528748,
+ "learning_rate": 0.0006,
+ "loss": 4.3395915031433105,
+ "step": 1708
+ },
+ {
+ "epoch": 23.74061135371179,
+ "grad_norm": 0.01676975190639496,
+ "learning_rate": 0.0006,
+ "loss": 4.254745960235596,
+ "step": 1709
+ },
+ {
+ "epoch": 23.754585152838427,
+ "grad_norm": 0.016370350494980812,
+ "learning_rate": 0.0006,
+ "loss": 4.2612223625183105,
+ "step": 1710
+ },
+ {
+ "epoch": 23.768558951965066,
+ "grad_norm": 0.01696198247373104,
+ "learning_rate": 0.0006,
+ "loss": 4.327524662017822,
+ "step": 1711
+ },
+ {
+ "epoch": 23.782532751091704,
+ "grad_norm": 0.0189108457416296,
+ "learning_rate": 0.0006,
+ "loss": 4.325839519500732,
+ "step": 1712
+ },
+ {
+ "epoch": 23.79650655021834,
+ "grad_norm": 0.01862351782619953,
+ "learning_rate": 0.0006,
+ "loss": 4.333662509918213,
+ "step": 1713
+ },
+ {
+ "epoch": 23.810480349344978,
+ "grad_norm": 0.01655014418065548,
+ "learning_rate": 0.0006,
+ "loss": 4.39305305480957,
+ "step": 1714
+ },
+ {
+ "epoch": 23.824454148471617,
+ "grad_norm": 0.015899376943707466,
+ "learning_rate": 0.0006,
+ "loss": 4.390053749084473,
+ "step": 1715
+ },
+ {
+ "epoch": 23.83842794759825,
+ "grad_norm": 0.016519056633114815,
+ "learning_rate": 0.0006,
+ "loss": 4.255903244018555,
+ "step": 1716
+ },
+ {
+ "epoch": 23.85240174672489,
+ "grad_norm": 0.0167539119720459,
+ "learning_rate": 0.0006,
+ "loss": 4.255727767944336,
+ "step": 1717
+ },
+ {
+ "epoch": 23.86637554585153,
+ "grad_norm": 0.0171353816986084,
+ "learning_rate": 0.0006,
+ "loss": 4.376740455627441,
+ "step": 1718
+ },
+ {
+ "epoch": 23.880349344978168,
+ "grad_norm": 0.016719117760658264,
+ "learning_rate": 0.0006,
+ "loss": 4.3376688957214355,
+ "step": 1719
+ },
+ {
+ "epoch": 23.894323144104803,
+ "grad_norm": 0.01561494916677475,
+ "learning_rate": 0.0006,
+ "loss": 4.3345746994018555,
+ "step": 1720
+ },
+ {
+ "epoch": 23.90829694323144,
+ "grad_norm": 0.016303174197673798,
+ "learning_rate": 0.0006,
+ "loss": 4.332371234893799,
+ "step": 1721
+ },
+ {
+ "epoch": 23.92227074235808,
+ "grad_norm": 0.016722865402698517,
+ "learning_rate": 0.0006,
+ "loss": 4.267125129699707,
+ "step": 1722
+ },
+ {
+ "epoch": 23.936244541484715,
+ "grad_norm": 0.017072124406695366,
+ "learning_rate": 0.0006,
+ "loss": 4.322751998901367,
+ "step": 1723
+ },
+ {
+ "epoch": 23.950218340611354,
+ "grad_norm": 0.01699932850897312,
+ "learning_rate": 0.0006,
+ "loss": 4.378422737121582,
+ "step": 1724
+ },
+ {
+ "epoch": 23.964192139737992,
+ "grad_norm": 0.01742720976471901,
+ "learning_rate": 0.0006,
+ "loss": 4.345083236694336,
+ "step": 1725
+ },
+ {
+ "epoch": 23.978165938864628,
+ "grad_norm": 0.018797501921653748,
+ "learning_rate": 0.0006,
+ "loss": 4.290443420410156,
+ "step": 1726
+ },
+ {
+ "epoch": 23.992139737991266,
+ "grad_norm": 0.018090683966875076,
+ "learning_rate": 0.0006,
+ "loss": 4.377580642700195,
+ "step": 1727
+ },
+ {
+ "epoch": 24.0,
+ "grad_norm": 0.017305459827184677,
+ "learning_rate": 0.0006,
+ "loss": 4.364426612854004,
+ "step": 1728
+ },
+ {
+ "epoch": 24.0,
+ "eval_loss": 4.647762298583984,
+ "eval_runtime": 56.5402,
+ "eval_samples_per_second": 43.191,
+ "eval_steps_per_second": 1.362,
+ "step": 1728
+ },
+ {
+ "epoch": 24.01397379912664,
+ "grad_norm": 0.016493607312440872,
+ "learning_rate": 0.0006,
+ "loss": 4.252255916595459,
+ "step": 1729
+ },
+ {
+ "epoch": 24.027947598253274,
+ "grad_norm": 0.016529450193047523,
+ "learning_rate": 0.0006,
+ "loss": 4.352746963500977,
+ "step": 1730
+ },
+ {
+ "epoch": 24.041921397379912,
+ "grad_norm": 0.016940191388130188,
+ "learning_rate": 0.0006,
+ "loss": 4.301870822906494,
+ "step": 1731
+ },
+ {
+ "epoch": 24.05589519650655,
+ "grad_norm": 0.01752558909356594,
+ "learning_rate": 0.0006,
+ "loss": 4.2762451171875,
+ "step": 1732
+ },
+ {
+ "epoch": 24.069868995633186,
+ "grad_norm": 0.017112884670495987,
+ "learning_rate": 0.0006,
+ "loss": 4.2659149169921875,
+ "step": 1733
+ },
+ {
+ "epoch": 24.083842794759825,
+ "grad_norm": 0.01900586113333702,
+ "learning_rate": 0.0006,
+ "loss": 4.306058883666992,
+ "step": 1734
+ },
+ {
+ "epoch": 24.097816593886463,
+ "grad_norm": 0.019447680562734604,
+ "learning_rate": 0.0006,
+ "loss": 4.450128555297852,
+ "step": 1735
+ },
+ {
+ "epoch": 24.111790393013102,
+ "grad_norm": 0.01948186755180359,
+ "learning_rate": 0.0006,
+ "loss": 4.202869415283203,
+ "step": 1736
+ },
+ {
+ "epoch": 24.125764192139737,
+ "grad_norm": 0.01716383546590805,
+ "learning_rate": 0.0006,
+ "loss": 4.318876266479492,
+ "step": 1737
+ },
+ {
+ "epoch": 24.139737991266376,
+ "grad_norm": 0.017049988731741905,
+ "learning_rate": 0.0006,
+ "loss": 4.237238883972168,
+ "step": 1738
+ },
+ {
+ "epoch": 24.153711790393015,
+ "grad_norm": 0.015169923193752766,
+ "learning_rate": 0.0006,
+ "loss": 4.289426803588867,
+ "step": 1739
+ },
+ {
+ "epoch": 24.16768558951965,
+ "grad_norm": 0.014760667458176613,
+ "learning_rate": 0.0006,
+ "loss": 4.356924057006836,
+ "step": 1740
+ },
+ {
+ "epoch": 24.18165938864629,
+ "grad_norm": 0.015127750113606453,
+ "learning_rate": 0.0006,
+ "loss": 4.142916679382324,
+ "step": 1741
+ },
+ {
+ "epoch": 24.195633187772927,
+ "grad_norm": 0.01549555640667677,
+ "learning_rate": 0.0006,
+ "loss": 4.345563888549805,
+ "step": 1742
+ },
+ {
+ "epoch": 24.209606986899562,
+ "grad_norm": 0.016393091529607773,
+ "learning_rate": 0.0006,
+ "loss": 4.417495250701904,
+ "step": 1743
+ },
+ {
+ "epoch": 24.2235807860262,
+ "grad_norm": 0.01785266026854515,
+ "learning_rate": 0.0006,
+ "loss": 4.3553547859191895,
+ "step": 1744
+ },
+ {
+ "epoch": 24.23755458515284,
+ "grad_norm": 0.017574617639183998,
+ "learning_rate": 0.0006,
+ "loss": 4.241024494171143,
+ "step": 1745
+ },
+ {
+ "epoch": 24.251528384279474,
+ "grad_norm": 0.019060306251049042,
+ "learning_rate": 0.0006,
+ "loss": 4.3685150146484375,
+ "step": 1746
+ },
+ {
+ "epoch": 24.265502183406113,
+ "grad_norm": 0.0214633010327816,
+ "learning_rate": 0.0006,
+ "loss": 4.323866844177246,
+ "step": 1747
+ },
+ {
+ "epoch": 24.27947598253275,
+ "grad_norm": 0.021055003628134727,
+ "learning_rate": 0.0006,
+ "loss": 4.278757095336914,
+ "step": 1748
+ },
+ {
+ "epoch": 24.29344978165939,
+ "grad_norm": 0.018618572503328323,
+ "learning_rate": 0.0006,
+ "loss": 4.28849983215332,
+ "step": 1749
+ },
+ {
+ "epoch": 24.307423580786025,
+ "grad_norm": 0.018666435033082962,
+ "learning_rate": 0.0006,
+ "loss": 4.368590354919434,
+ "step": 1750
+ },
+ {
+ "epoch": 24.321397379912664,
+ "grad_norm": 0.020626146346330643,
+ "learning_rate": 0.0006,
+ "loss": 4.2951836585998535,
+ "step": 1751
+ },
+ {
+ "epoch": 24.335371179039303,
+ "grad_norm": 0.01929759792983532,
+ "learning_rate": 0.0006,
+ "loss": 4.3664703369140625,
+ "step": 1752
+ },
+ {
+ "epoch": 24.349344978165938,
+ "grad_norm": 0.020956002175807953,
+ "learning_rate": 0.0006,
+ "loss": 4.345308303833008,
+ "step": 1753
+ },
+ {
+ "epoch": 24.363318777292577,
+ "grad_norm": 0.019227957352995872,
+ "learning_rate": 0.0006,
+ "loss": 4.349849224090576,
+ "step": 1754
+ },
+ {
+ "epoch": 24.377292576419215,
+ "grad_norm": 0.015970097854733467,
+ "learning_rate": 0.0006,
+ "loss": 4.223012447357178,
+ "step": 1755
+ },
+ {
+ "epoch": 24.39126637554585,
+ "grad_norm": 0.015951646491885185,
+ "learning_rate": 0.0006,
+ "loss": 4.236420631408691,
+ "step": 1756
+ },
+ {
+ "epoch": 24.40524017467249,
+ "grad_norm": 0.016032515093684196,
+ "learning_rate": 0.0006,
+ "loss": 4.226899147033691,
+ "step": 1757
+ },
+ {
+ "epoch": 24.419213973799128,
+ "grad_norm": 0.01573132909834385,
+ "learning_rate": 0.0006,
+ "loss": 4.315916061401367,
+ "step": 1758
+ },
+ {
+ "epoch": 24.433187772925763,
+ "grad_norm": 0.015367005951702595,
+ "learning_rate": 0.0006,
+ "loss": 4.358939170837402,
+ "step": 1759
+ },
+ {
+ "epoch": 24.4471615720524,
+ "grad_norm": 0.01669907011091709,
+ "learning_rate": 0.0006,
+ "loss": 4.232147216796875,
+ "step": 1760
+ },
+ {
+ "epoch": 24.46113537117904,
+ "grad_norm": 0.018708152696490288,
+ "learning_rate": 0.0006,
+ "loss": 4.329852104187012,
+ "step": 1761
+ },
+ {
+ "epoch": 24.475109170305675,
+ "grad_norm": 0.019035987555980682,
+ "learning_rate": 0.0006,
+ "loss": 4.260340213775635,
+ "step": 1762
+ },
+ {
+ "epoch": 24.489082969432314,
+ "grad_norm": 0.01832144521176815,
+ "learning_rate": 0.0006,
+ "loss": 4.359543800354004,
+ "step": 1763
+ },
+ {
+ "epoch": 24.503056768558952,
+ "grad_norm": 0.015313294716179371,
+ "learning_rate": 0.0006,
+ "loss": 4.279088020324707,
+ "step": 1764
+ },
+ {
+ "epoch": 24.51703056768559,
+ "grad_norm": 0.015803923830389977,
+ "learning_rate": 0.0006,
+ "loss": 4.230682849884033,
+ "step": 1765
+ },
+ {
+ "epoch": 24.531004366812226,
+ "grad_norm": 0.016748374328017235,
+ "learning_rate": 0.0006,
+ "loss": 4.349661827087402,
+ "step": 1766
+ },
+ {
+ "epoch": 24.544978165938865,
+ "grad_norm": 0.015732428058981895,
+ "learning_rate": 0.0006,
+ "loss": 4.372441291809082,
+ "step": 1767
+ },
+ {
+ "epoch": 24.558951965065503,
+ "grad_norm": 0.016564829275012016,
+ "learning_rate": 0.0006,
+ "loss": 4.323090553283691,
+ "step": 1768
+ },
+ {
+ "epoch": 24.57292576419214,
+ "grad_norm": 0.01603010483086109,
+ "learning_rate": 0.0006,
+ "loss": 4.347898006439209,
+ "step": 1769
+ },
+ {
+ "epoch": 24.586899563318777,
+ "grad_norm": 0.01674809865653515,
+ "learning_rate": 0.0006,
+ "loss": 4.269165515899658,
+ "step": 1770
+ },
+ {
+ "epoch": 24.600873362445416,
+ "grad_norm": 0.017333930358290672,
+ "learning_rate": 0.0006,
+ "loss": 4.353298187255859,
+ "step": 1771
+ },
+ {
+ "epoch": 24.61484716157205,
+ "grad_norm": 0.017519580200314522,
+ "learning_rate": 0.0006,
+ "loss": 4.265199661254883,
+ "step": 1772
+ },
+ {
+ "epoch": 24.62882096069869,
+ "grad_norm": 0.017675306648015976,
+ "learning_rate": 0.0006,
+ "loss": 4.322596549987793,
+ "step": 1773
+ },
+ {
+ "epoch": 24.64279475982533,
+ "grad_norm": 0.017166534438729286,
+ "learning_rate": 0.0006,
+ "loss": 4.413900375366211,
+ "step": 1774
+ },
+ {
+ "epoch": 24.656768558951963,
+ "grad_norm": 0.016362346708774567,
+ "learning_rate": 0.0006,
+ "loss": 4.307236671447754,
+ "step": 1775
+ },
+ {
+ "epoch": 24.670742358078602,
+ "grad_norm": 0.016086872667074203,
+ "learning_rate": 0.0006,
+ "loss": 4.235316276550293,
+ "step": 1776
+ },
+ {
+ "epoch": 24.68471615720524,
+ "grad_norm": 0.01689537614583969,
+ "learning_rate": 0.0006,
+ "loss": 4.390591621398926,
+ "step": 1777
+ },
+ {
+ "epoch": 24.69868995633188,
+ "grad_norm": 0.016836725175380707,
+ "learning_rate": 0.0006,
+ "loss": 4.2990031242370605,
+ "step": 1778
+ },
+ {
+ "epoch": 24.712663755458514,
+ "grad_norm": 0.015627335757017136,
+ "learning_rate": 0.0006,
+ "loss": 4.166745185852051,
+ "step": 1779
+ },
+ {
+ "epoch": 24.726637554585153,
+ "grad_norm": 0.015007016249001026,
+ "learning_rate": 0.0006,
+ "loss": 4.295273303985596,
+ "step": 1780
+ },
+ {
+ "epoch": 24.74061135371179,
+ "grad_norm": 0.016084209084510803,
+ "learning_rate": 0.0006,
+ "loss": 4.342424392700195,
+ "step": 1781
+ },
+ {
+ "epoch": 24.754585152838427,
+ "grad_norm": 0.01595073565840721,
+ "learning_rate": 0.0006,
+ "loss": 4.3060102462768555,
+ "step": 1782
+ },
+ {
+ "epoch": 24.768558951965066,
+ "grad_norm": 0.016113461926579475,
+ "learning_rate": 0.0006,
+ "loss": 4.267908096313477,
+ "step": 1783
+ },
+ {
+ "epoch": 24.782532751091704,
+ "grad_norm": 0.01564556173980236,
+ "learning_rate": 0.0006,
+ "loss": 4.324642658233643,
+ "step": 1784
+ },
+ {
+ "epoch": 24.79650655021834,
+ "grad_norm": 0.01575268618762493,
+ "learning_rate": 0.0006,
+ "loss": 4.328181743621826,
+ "step": 1785
+ },
+ {
+ "epoch": 24.810480349344978,
+ "grad_norm": 0.016830824315547943,
+ "learning_rate": 0.0006,
+ "loss": 4.247906684875488,
+ "step": 1786
+ },
+ {
+ "epoch": 24.824454148471617,
+ "grad_norm": 0.017989547923207283,
+ "learning_rate": 0.0006,
+ "loss": 4.228281021118164,
+ "step": 1787
+ },
+ {
+ "epoch": 24.83842794759825,
+ "grad_norm": 0.016783086583018303,
+ "learning_rate": 0.0006,
+ "loss": 4.2880964279174805,
+ "step": 1788
+ },
+ {
+ "epoch": 24.85240174672489,
+ "grad_norm": 0.016467561945319176,
+ "learning_rate": 0.0006,
+ "loss": 4.36118745803833,
+ "step": 1789
+ },
+ {
+ "epoch": 24.86637554585153,
+ "grad_norm": 0.01780523732304573,
+ "learning_rate": 0.0006,
+ "loss": 4.172321319580078,
+ "step": 1790
+ },
+ {
+ "epoch": 24.880349344978168,
+ "grad_norm": 0.017102031037211418,
+ "learning_rate": 0.0006,
+ "loss": 4.352826118469238,
+ "step": 1791
+ },
+ {
+ "epoch": 24.894323144104803,
+ "grad_norm": 0.016409218311309814,
+ "learning_rate": 0.0006,
+ "loss": 4.212162017822266,
+ "step": 1792
+ },
+ {
+ "epoch": 24.90829694323144,
+ "grad_norm": 0.018660584464669228,
+ "learning_rate": 0.0006,
+ "loss": 4.346684455871582,
+ "step": 1793
+ },
+ {
+ "epoch": 24.92227074235808,
+ "grad_norm": 0.017389440909028053,
+ "learning_rate": 0.0006,
+ "loss": 4.197913646697998,
+ "step": 1794
+ },
+ {
+ "epoch": 24.936244541484715,
+ "grad_norm": 0.01596485823392868,
+ "learning_rate": 0.0006,
+ "loss": 4.370617866516113,
+ "step": 1795
+ },
+ {
+ "epoch": 24.950218340611354,
+ "grad_norm": 0.01709776371717453,
+ "learning_rate": 0.0006,
+ "loss": 4.313847064971924,
+ "step": 1796
+ },
+ {
+ "epoch": 24.964192139737992,
+ "grad_norm": 0.016571134328842163,
+ "learning_rate": 0.0006,
+ "loss": 4.1895856857299805,
+ "step": 1797
+ },
+ {
+ "epoch": 24.978165938864628,
+ "grad_norm": 0.016851048916578293,
+ "learning_rate": 0.0006,
+ "loss": 4.350772857666016,
+ "step": 1798
+ },
+ {
+ "epoch": 24.992139737991266,
+ "grad_norm": 0.017007440328598022,
+ "learning_rate": 0.0006,
+ "loss": 4.399008750915527,
+ "step": 1799
+ },
+ {
+ "epoch": 25.0,
+ "grad_norm": 0.01791212148964405,
+ "learning_rate": 0.0006,
+ "loss": 4.306121826171875,
+ "step": 1800
+ },
+ {
+ "epoch": 25.0,
+ "eval_loss": 4.636417865753174,
+ "eval_runtime": 56.8296,
+ "eval_samples_per_second": 42.971,
+ "eval_steps_per_second": 1.355,
+ "step": 1800
+ },
+ {
+ "epoch": 25.01397379912664,
+ "grad_norm": 0.01729617640376091,
+ "learning_rate": 0.0006,
+ "loss": 4.257018089294434,
+ "step": 1801
+ },
+ {
+ "epoch": 25.027947598253274,
+ "grad_norm": 0.017466643825173378,
+ "learning_rate": 0.0006,
+ "loss": 4.307605266571045,
+ "step": 1802
+ },
+ {
+ "epoch": 25.041921397379912,
+ "grad_norm": 0.018614279106259346,
+ "learning_rate": 0.0006,
+ "loss": 4.360141277313232,
+ "step": 1803
+ },
+ {
+ "epoch": 25.05589519650655,
+ "grad_norm": 0.02010771445930004,
+ "learning_rate": 0.0006,
+ "loss": 4.196428298950195,
+ "step": 1804
+ },
+ {
+ "epoch": 25.069868995633186,
+ "grad_norm": 0.02090228535234928,
+ "learning_rate": 0.0006,
+ "loss": 4.288947105407715,
+ "step": 1805
+ },
+ {
+ "epoch": 25.083842794759825,
+ "grad_norm": 0.02050255797803402,
+ "learning_rate": 0.0006,
+ "loss": 4.259582042694092,
+ "step": 1806
+ },
+ {
+ "epoch": 25.097816593886463,
+ "grad_norm": 0.021844089031219482,
+ "learning_rate": 0.0006,
+ "loss": 4.405138969421387,
+ "step": 1807
+ },
+ {
+ "epoch": 25.111790393013102,
+ "grad_norm": 0.022287018597126007,
+ "learning_rate": 0.0006,
+ "loss": 4.333793640136719,
+ "step": 1808
+ },
+ {
+ "epoch": 25.125764192139737,
+ "grad_norm": 0.02288329415023327,
+ "learning_rate": 0.0006,
+ "loss": 4.3910980224609375,
+ "step": 1809
+ },
+ {
+ "epoch": 25.139737991266376,
+ "grad_norm": 0.02373199723660946,
+ "learning_rate": 0.0006,
+ "loss": 4.32921028137207,
+ "step": 1810
+ },
+ {
+ "epoch": 25.153711790393015,
+ "grad_norm": 0.024960234761238098,
+ "learning_rate": 0.0006,
+ "loss": 4.295609474182129,
+ "step": 1811
+ },
+ {
+ "epoch": 25.16768558951965,
+ "grad_norm": 0.026743892580270767,
+ "learning_rate": 0.0006,
+ "loss": 4.315308094024658,
+ "step": 1812
+ },
+ {
+ "epoch": 25.18165938864629,
+ "grad_norm": 0.022138185799121857,
+ "learning_rate": 0.0006,
+ "loss": 4.360208511352539,
+ "step": 1813
+ },
+ {
+ "epoch": 25.195633187772927,
+ "grad_norm": 0.022120574489235878,
+ "learning_rate": 0.0006,
+ "loss": 4.1973772048950195,
+ "step": 1814
+ },
+ {
+ "epoch": 25.209606986899562,
+ "grad_norm": 0.02230502851307392,
+ "learning_rate": 0.0006,
+ "loss": 4.285558700561523,
+ "step": 1815
+ },
+ {
+ "epoch": 25.2235807860262,
+ "grad_norm": 0.022301986813545227,
+ "learning_rate": 0.0006,
+ "loss": 4.238343238830566,
+ "step": 1816
+ },
+ {
+ "epoch": 25.23755458515284,
+ "grad_norm": 0.020050447434186935,
+ "learning_rate": 0.0006,
+ "loss": 4.298235893249512,
+ "step": 1817
+ },
+ {
+ "epoch": 25.251528384279474,
+ "grad_norm": 0.01987724006175995,
+ "learning_rate": 0.0006,
+ "loss": 4.301384449005127,
+ "step": 1818
+ },
+ {
+ "epoch": 25.265502183406113,
+ "grad_norm": 0.020440855994820595,
+ "learning_rate": 0.0006,
+ "loss": 4.269218444824219,
+ "step": 1819
+ },
+ {
+ "epoch": 25.27947598253275,
+ "grad_norm": 0.01891893707215786,
+ "learning_rate": 0.0006,
+ "loss": 4.2401957511901855,
+ "step": 1820
+ },
+ {
+ "epoch": 25.29344978165939,
+ "grad_norm": 0.01903688907623291,
+ "learning_rate": 0.0006,
+ "loss": 4.139670372009277,
+ "step": 1821
+ },
+ {
+ "epoch": 25.307423580786025,
+ "grad_norm": 0.018856938928365707,
+ "learning_rate": 0.0006,
+ "loss": 4.233214378356934,
+ "step": 1822
+ },
+ {
+ "epoch": 25.321397379912664,
+ "grad_norm": 0.017791323363780975,
+ "learning_rate": 0.0006,
+ "loss": 4.2342023849487305,
+ "step": 1823
+ },
+ {
+ "epoch": 25.335371179039303,
+ "grad_norm": 0.017304565757513046,
+ "learning_rate": 0.0006,
+ "loss": 4.355198860168457,
+ "step": 1824
+ },
+ {
+ "epoch": 25.349344978165938,
+ "grad_norm": 0.01576206088066101,
+ "learning_rate": 0.0006,
+ "loss": 4.114147186279297,
+ "step": 1825
+ },
+ {
+ "epoch": 25.363318777292577,
+ "grad_norm": 0.015105154365301132,
+ "learning_rate": 0.0006,
+ "loss": 4.313795566558838,
+ "step": 1826
+ },
+ {
+ "epoch": 25.377292576419215,
+ "grad_norm": 0.01563587784767151,
+ "learning_rate": 0.0006,
+ "loss": 4.308282852172852,
+ "step": 1827
+ },
+ {
+ "epoch": 25.39126637554585,
+ "grad_norm": 0.015980906784534454,
+ "learning_rate": 0.0006,
+ "loss": 4.345149040222168,
+ "step": 1828
+ },
+ {
+ "epoch": 25.40524017467249,
+ "grad_norm": 0.01669284701347351,
+ "learning_rate": 0.0006,
+ "loss": 4.237611293792725,
+ "step": 1829
+ },
+ {
+ "epoch": 25.419213973799128,
+ "grad_norm": 0.016202064231038094,
+ "learning_rate": 0.0006,
+ "loss": 4.3072309494018555,
+ "step": 1830
+ },
+ {
+ "epoch": 25.433187772925763,
+ "grad_norm": 0.01478662807494402,
+ "learning_rate": 0.0006,
+ "loss": 4.276215076446533,
+ "step": 1831
+ },
+ {
+ "epoch": 25.4471615720524,
+ "grad_norm": 0.015168731100857258,
+ "learning_rate": 0.0006,
+ "loss": 4.37509822845459,
+ "step": 1832
+ },
+ {
+ "epoch": 25.46113537117904,
+ "grad_norm": 0.01648980937898159,
+ "learning_rate": 0.0006,
+ "loss": 4.349167346954346,
+ "step": 1833
+ },
+ {
+ "epoch": 25.475109170305675,
+ "grad_norm": 0.01708398200571537,
+ "learning_rate": 0.0006,
+ "loss": 4.346580505371094,
+ "step": 1834
+ },
+ {
+ "epoch": 25.489082969432314,
+ "grad_norm": 0.017370784655213356,
+ "learning_rate": 0.0006,
+ "loss": 4.299466133117676,
+ "step": 1835
+ },
+ {
+ "epoch": 25.503056768558952,
+ "grad_norm": 0.016248228028416634,
+ "learning_rate": 0.0006,
+ "loss": 4.184610366821289,
+ "step": 1836
+ },
+ {
+ "epoch": 25.51703056768559,
+ "grad_norm": 0.017106913030147552,
+ "learning_rate": 0.0006,
+ "loss": 4.366856098175049,
+ "step": 1837
+ },
+ {
+ "epoch": 25.531004366812226,
+ "grad_norm": 0.016071965917944908,
+ "learning_rate": 0.0006,
+ "loss": 4.297691345214844,
+ "step": 1838
+ },
+ {
+ "epoch": 25.544978165938865,
+ "grad_norm": 0.016451986506581306,
+ "learning_rate": 0.0006,
+ "loss": 4.229983329772949,
+ "step": 1839
+ },
+ {
+ "epoch": 25.558951965065503,
+ "grad_norm": 0.016195081174373627,
+ "learning_rate": 0.0006,
+ "loss": 4.259893417358398,
+ "step": 1840
+ },
+ {
+ "epoch": 25.57292576419214,
+ "grad_norm": 0.0165175162255764,
+ "learning_rate": 0.0006,
+ "loss": 4.3348236083984375,
+ "step": 1841
+ },
+ {
+ "epoch": 25.586899563318777,
+ "grad_norm": 0.01651517115533352,
+ "learning_rate": 0.0006,
+ "loss": 4.319541931152344,
+ "step": 1842
+ },
+ {
+ "epoch": 25.600873362445416,
+ "grad_norm": 0.016143690794706345,
+ "learning_rate": 0.0006,
+ "loss": 4.345652103424072,
+ "step": 1843
+ },
+ {
+ "epoch": 25.61484716157205,
+ "grad_norm": 0.015140696428716183,
+ "learning_rate": 0.0006,
+ "loss": 4.212047576904297,
+ "step": 1844
+ },
+ {
+ "epoch": 25.62882096069869,
+ "grad_norm": 0.014637443237006664,
+ "learning_rate": 0.0006,
+ "loss": 4.187453269958496,
+ "step": 1845
+ },
+ {
+ "epoch": 25.64279475982533,
+ "grad_norm": 0.014585614204406738,
+ "learning_rate": 0.0006,
+ "loss": 4.253936767578125,
+ "step": 1846
+ },
+ {
+ "epoch": 25.656768558951963,
+ "grad_norm": 0.015158289112150669,
+ "learning_rate": 0.0006,
+ "loss": 4.34612512588501,
+ "step": 1847
+ },
+ {
+ "epoch": 25.670742358078602,
+ "grad_norm": 0.015202849172055721,
+ "learning_rate": 0.0006,
+ "loss": 4.270634651184082,
+ "step": 1848
+ },
+ {
+ "epoch": 25.68471615720524,
+ "grad_norm": 0.015363575890660286,
+ "learning_rate": 0.0006,
+ "loss": 4.271080493927002,
+ "step": 1849
+ },
+ {
+ "epoch": 25.69868995633188,
+ "grad_norm": 0.014360226690769196,
+ "learning_rate": 0.0006,
+ "loss": 4.375033855438232,
+ "step": 1850
+ },
+ {
+ "epoch": 25.712663755458514,
+ "grad_norm": 0.014263181947171688,
+ "learning_rate": 0.0006,
+ "loss": 4.424138069152832,
+ "step": 1851
+ },
+ {
+ "epoch": 25.726637554585153,
+ "grad_norm": 0.014398688450455666,
+ "learning_rate": 0.0006,
+ "loss": 4.410154819488525,
+ "step": 1852
+ },
+ {
+ "epoch": 25.74061135371179,
+ "grad_norm": 0.016131488606333733,
+ "learning_rate": 0.0006,
+ "loss": 4.324014663696289,
+ "step": 1853
+ },
+ {
+ "epoch": 25.754585152838427,
+ "grad_norm": 0.017568735405802727,
+ "learning_rate": 0.0006,
+ "loss": 4.367861747741699,
+ "step": 1854
+ },
+ {
+ "epoch": 25.768558951965066,
+ "grad_norm": 0.017907511442899704,
+ "learning_rate": 0.0006,
+ "loss": 4.352965354919434,
+ "step": 1855
+ },
+ {
+ "epoch": 25.782532751091704,
+ "grad_norm": 0.016918016597628593,
+ "learning_rate": 0.0006,
+ "loss": 4.252397537231445,
+ "step": 1856
+ },
+ {
+ "epoch": 25.79650655021834,
+ "grad_norm": 0.017274610698223114,
+ "learning_rate": 0.0006,
+ "loss": 4.342309951782227,
+ "step": 1857
+ },
+ {
+ "epoch": 25.810480349344978,
+ "grad_norm": 0.017879825085401535,
+ "learning_rate": 0.0006,
+ "loss": 4.330987453460693,
+ "step": 1858
+ },
+ {
+ "epoch": 25.824454148471617,
+ "grad_norm": 0.01736099272966385,
+ "learning_rate": 0.0006,
+ "loss": 4.307255744934082,
+ "step": 1859
+ },
+ {
+ "epoch": 25.83842794759825,
+ "grad_norm": 0.016575824469327927,
+ "learning_rate": 0.0006,
+ "loss": 4.262016773223877,
+ "step": 1860
+ },
+ {
+ "epoch": 25.85240174672489,
+ "grad_norm": 0.015181140042841434,
+ "learning_rate": 0.0006,
+ "loss": 4.291863441467285,
+ "step": 1861
+ },
+ {
+ "epoch": 25.86637554585153,
+ "grad_norm": 0.015965687111020088,
+ "learning_rate": 0.0006,
+ "loss": 4.424836158752441,
+ "step": 1862
+ },
+ {
+ "epoch": 25.880349344978168,
+ "grad_norm": 0.01687219925224781,
+ "learning_rate": 0.0006,
+ "loss": 4.343531131744385,
+ "step": 1863
+ },
+ {
+ "epoch": 25.894323144104803,
+ "grad_norm": 0.015115504153072834,
+ "learning_rate": 0.0006,
+ "loss": 4.357808589935303,
+ "step": 1864
+ },
+ {
+ "epoch": 25.90829694323144,
+ "grad_norm": 0.014150998555123806,
+ "learning_rate": 0.0006,
+ "loss": 4.324550151824951,
+ "step": 1865
+ },
+ {
+ "epoch": 25.92227074235808,
+ "grad_norm": 0.014989510178565979,
+ "learning_rate": 0.0006,
+ "loss": 4.380428314208984,
+ "step": 1866
+ },
+ {
+ "epoch": 25.936244541484715,
+ "grad_norm": 0.01527960691601038,
+ "learning_rate": 0.0006,
+ "loss": 4.399374961853027,
+ "step": 1867
+ },
+ {
+ "epoch": 25.950218340611354,
+ "grad_norm": 0.015434633940458298,
+ "learning_rate": 0.0006,
+ "loss": 4.260984420776367,
+ "step": 1868
+ },
+ {
+ "epoch": 25.964192139737992,
+ "grad_norm": 0.015370228327810764,
+ "learning_rate": 0.0006,
+ "loss": 4.3340253829956055,
+ "step": 1869
+ },
+ {
+ "epoch": 25.978165938864628,
+ "grad_norm": 0.015348542481660843,
+ "learning_rate": 0.0006,
+ "loss": 4.369132995605469,
+ "step": 1870
+ },
+ {
+ "epoch": 25.992139737991266,
+ "grad_norm": 0.015354936942458153,
+ "learning_rate": 0.0006,
+ "loss": 4.381937026977539,
+ "step": 1871
+ },
+ {
+ "epoch": 26.0,
+ "grad_norm": 0.01796272210776806,
+ "learning_rate": 0.0006,
+ "loss": 4.286768436431885,
+ "step": 1872
+ },
+ {
+ "epoch": 26.0,
+ "eval_loss": 4.576984405517578,
+ "eval_runtime": 57.2091,
+ "eval_samples_per_second": 42.685,
+ "eval_steps_per_second": 1.346,
+ "step": 1872
+ },
+ {
+ "epoch": 26.01397379912664,
+ "grad_norm": 0.01748845726251602,
+ "learning_rate": 0.0006,
+ "loss": 4.337504863739014,
+ "step": 1873
+ },
+ {
+ "epoch": 26.027947598253274,
+ "grad_norm": 0.019244296476244926,
+ "learning_rate": 0.0006,
+ "loss": 4.264280796051025,
+ "step": 1874
+ },
+ {
+ "epoch": 26.041921397379912,
+ "grad_norm": 0.0215692650526762,
+ "learning_rate": 0.0006,
+ "loss": 4.350830078125,
+ "step": 1875
+ },
+ {
+ "epoch": 26.05589519650655,
+ "grad_norm": 0.022189252078533173,
+ "learning_rate": 0.0006,
+ "loss": 4.299098968505859,
+ "step": 1876
+ },
+ {
+ "epoch": 26.069868995633186,
+ "grad_norm": 0.022384043782949448,
+ "learning_rate": 0.0006,
+ "loss": 4.1185173988342285,
+ "step": 1877
+ },
+ {
+ "epoch": 26.083842794759825,
+ "grad_norm": 0.02076014317572117,
+ "learning_rate": 0.0006,
+ "loss": 4.297612190246582,
+ "step": 1878
+ },
+ {
+ "epoch": 26.097816593886463,
+ "grad_norm": 0.020843051373958588,
+ "learning_rate": 0.0006,
+ "loss": 4.294098854064941,
+ "step": 1879
+ },
+ {
+ "epoch": 26.111790393013102,
+ "grad_norm": 0.020654456689953804,
+ "learning_rate": 0.0006,
+ "loss": 4.267856597900391,
+ "step": 1880
+ },
+ {
+ "epoch": 26.125764192139737,
+ "grad_norm": 0.01860472559928894,
+ "learning_rate": 0.0006,
+ "loss": 4.284232139587402,
+ "step": 1881
+ },
+ {
+ "epoch": 26.139737991266376,
+ "grad_norm": 0.018719421699643135,
+ "learning_rate": 0.0006,
+ "loss": 4.277214527130127,
+ "step": 1882
+ },
+ {
+ "epoch": 26.153711790393015,
+ "grad_norm": 0.020903829485177994,
+ "learning_rate": 0.0006,
+ "loss": 4.254947662353516,
+ "step": 1883
+ },
+ {
+ "epoch": 26.16768558951965,
+ "grad_norm": 0.02020149491727352,
+ "learning_rate": 0.0006,
+ "loss": 4.2526936531066895,
+ "step": 1884
+ },
+ {
+ "epoch": 26.18165938864629,
+ "grad_norm": 0.0180392786860466,
+ "learning_rate": 0.0006,
+ "loss": 4.309922695159912,
+ "step": 1885
+ },
+ {
+ "epoch": 26.195633187772927,
+ "grad_norm": 0.018163125962018967,
+ "learning_rate": 0.0006,
+ "loss": 4.13860559463501,
+ "step": 1886
+ },
+ {
+ "epoch": 26.209606986899562,
+ "grad_norm": 0.01853291131556034,
+ "learning_rate": 0.0006,
+ "loss": 4.190926551818848,
+ "step": 1887
+ },
+ {
+ "epoch": 26.2235807860262,
+ "grad_norm": 0.01757677271962166,
+ "learning_rate": 0.0006,
+ "loss": 4.163999557495117,
+ "step": 1888
+ },
+ {
+ "epoch": 26.23755458515284,
+ "grad_norm": 0.01767992228269577,
+ "learning_rate": 0.0006,
+ "loss": 4.184430122375488,
+ "step": 1889
+ },
+ {
+ "epoch": 26.251528384279474,
+ "grad_norm": 0.018139973282814026,
+ "learning_rate": 0.0006,
+ "loss": 4.300766944885254,
+ "step": 1890
+ },
+ {
+ "epoch": 26.265502183406113,
+ "grad_norm": 0.020184461027383804,
+ "learning_rate": 0.0006,
+ "loss": 4.232537269592285,
+ "step": 1891
+ },
+ {
+ "epoch": 26.27947598253275,
+ "grad_norm": 0.02028856985270977,
+ "learning_rate": 0.0006,
+ "loss": 4.324883460998535,
+ "step": 1892
+ },
+ {
+ "epoch": 26.29344978165939,
+ "grad_norm": 0.017120616510510445,
+ "learning_rate": 0.0006,
+ "loss": 4.23493766784668,
+ "step": 1893
+ },
+ {
+ "epoch": 26.307423580786025,
+ "grad_norm": 0.016202978789806366,
+ "learning_rate": 0.0006,
+ "loss": 4.197602272033691,
+ "step": 1894
+ },
+ {
+ "epoch": 26.321397379912664,
+ "grad_norm": 0.017231421545147896,
+ "learning_rate": 0.0006,
+ "loss": 4.302001953125,
+ "step": 1895
+ },
+ {
+ "epoch": 26.335371179039303,
+ "grad_norm": 0.01725655607879162,
+ "learning_rate": 0.0006,
+ "loss": 4.356800079345703,
+ "step": 1896
+ },
+ {
+ "epoch": 26.349344978165938,
+ "grad_norm": 0.016657443717122078,
+ "learning_rate": 0.0006,
+ "loss": 4.266307830810547,
+ "step": 1897
+ },
+ {
+ "epoch": 26.363318777292577,
+ "grad_norm": 0.01616556942462921,
+ "learning_rate": 0.0006,
+ "loss": 4.307868003845215,
+ "step": 1898
+ },
+ {
+ "epoch": 26.377292576419215,
+ "grad_norm": 0.016621306538581848,
+ "learning_rate": 0.0006,
+ "loss": 4.172905921936035,
+ "step": 1899
+ },
+ {
+ "epoch": 26.39126637554585,
+ "grad_norm": 0.016243206337094307,
+ "learning_rate": 0.0006,
+ "loss": 4.162896156311035,
+ "step": 1900
+ },
+ {
+ "epoch": 26.40524017467249,
+ "grad_norm": 0.016663808375597,
+ "learning_rate": 0.0006,
+ "loss": 4.258232116699219,
+ "step": 1901
+ },
+ {
+ "epoch": 26.419213973799128,
+ "grad_norm": 0.01677478663623333,
+ "learning_rate": 0.0006,
+ "loss": 4.336090087890625,
+ "step": 1902
+ },
+ {
+ "epoch": 26.433187772925763,
+ "grad_norm": 0.01641303487122059,
+ "learning_rate": 0.0006,
+ "loss": 4.191476821899414,
+ "step": 1903
+ },
+ {
+ "epoch": 26.4471615720524,
+ "grad_norm": 0.016700323671102524,
+ "learning_rate": 0.0006,
+ "loss": 4.299427509307861,
+ "step": 1904
+ },
+ {
+ "epoch": 26.46113537117904,
+ "grad_norm": 0.017595035955309868,
+ "learning_rate": 0.0006,
+ "loss": 4.312828063964844,
+ "step": 1905
+ },
+ {
+ "epoch": 26.475109170305675,
+ "grad_norm": 0.020300284028053284,
+ "learning_rate": 0.0006,
+ "loss": 4.28331184387207,
+ "step": 1906
+ },
+ {
+ "epoch": 26.489082969432314,
+ "grad_norm": 0.02198321744799614,
+ "learning_rate": 0.0006,
+ "loss": 4.3280839920043945,
+ "step": 1907
+ },
+ {
+ "epoch": 26.503056768558952,
+ "grad_norm": 0.02003200724720955,
+ "learning_rate": 0.0006,
+ "loss": 4.325623989105225,
+ "step": 1908
+ },
+ {
+ "epoch": 26.51703056768559,
+ "grad_norm": 0.01870287023484707,
+ "learning_rate": 0.0006,
+ "loss": 4.251203536987305,
+ "step": 1909
+ },
+ {
+ "epoch": 26.531004366812226,
+ "grad_norm": 0.01835448481142521,
+ "learning_rate": 0.0006,
+ "loss": 4.335144996643066,
+ "step": 1910
+ },
+ {
+ "epoch": 26.544978165938865,
+ "grad_norm": 0.02046327292919159,
+ "learning_rate": 0.0006,
+ "loss": 4.238023281097412,
+ "step": 1911
+ },
+ {
+ "epoch": 26.558951965065503,
+ "grad_norm": 0.019771311432123184,
+ "learning_rate": 0.0006,
+ "loss": 4.353594779968262,
+ "step": 1912
+ },
+ {
+ "epoch": 26.57292576419214,
+ "grad_norm": 0.018090544268488884,
+ "learning_rate": 0.0006,
+ "loss": 4.304343223571777,
+ "step": 1913
+ },
+ {
+ "epoch": 26.586899563318777,
+ "grad_norm": 0.017489777877926826,
+ "learning_rate": 0.0006,
+ "loss": 4.204375267028809,
+ "step": 1914
+ },
+ {
+ "epoch": 26.600873362445416,
+ "grad_norm": 0.016785891726613045,
+ "learning_rate": 0.0006,
+ "loss": 4.286669731140137,
+ "step": 1915
+ },
+ {
+ "epoch": 26.61484716157205,
+ "grad_norm": 0.016382789239287376,
+ "learning_rate": 0.0006,
+ "loss": 4.347662448883057,
+ "step": 1916
+ },
+ {
+ "epoch": 26.62882096069869,
+ "grad_norm": 0.016581876203417778,
+ "learning_rate": 0.0006,
+ "loss": 4.2716193199157715,
+ "step": 1917
+ },
+ {
+ "epoch": 26.64279475982533,
+ "grad_norm": 0.015627184882760048,
+ "learning_rate": 0.0006,
+ "loss": 4.311489105224609,
+ "step": 1918
+ },
+ {
+ "epoch": 26.656768558951963,
+ "grad_norm": 0.015908483415842056,
+ "learning_rate": 0.0006,
+ "loss": 4.262099742889404,
+ "step": 1919
+ },
+ {
+ "epoch": 26.670742358078602,
+ "grad_norm": 0.015576236881315708,
+ "learning_rate": 0.0006,
+ "loss": 4.275871276855469,
+ "step": 1920
+ },
+ {
+ "epoch": 26.68471615720524,
+ "grad_norm": 0.014621064998209476,
+ "learning_rate": 0.0006,
+ "loss": 4.184260845184326,
+ "step": 1921
+ },
+ {
+ "epoch": 26.69868995633188,
+ "grad_norm": 0.014797811396420002,
+ "learning_rate": 0.0006,
+ "loss": 4.3412628173828125,
+ "step": 1922
+ },
+ {
+ "epoch": 26.712663755458514,
+ "grad_norm": 0.015610570088028908,
+ "learning_rate": 0.0006,
+ "loss": 4.181094646453857,
+ "step": 1923
+ },
+ {
+ "epoch": 26.726637554585153,
+ "grad_norm": 0.016048265621066093,
+ "learning_rate": 0.0006,
+ "loss": 4.365290641784668,
+ "step": 1924
+ },
+ {
+ "epoch": 26.74061135371179,
+ "grad_norm": 0.016604071483016014,
+ "learning_rate": 0.0006,
+ "loss": 4.2628912925720215,
+ "step": 1925
+ },
+ {
+ "epoch": 26.754585152838427,
+ "grad_norm": 0.017044221982359886,
+ "learning_rate": 0.0006,
+ "loss": 4.378512382507324,
+ "step": 1926
+ },
+ {
+ "epoch": 26.768558951965066,
+ "grad_norm": 0.01746384985744953,
+ "learning_rate": 0.0006,
+ "loss": 4.3888773918151855,
+ "step": 1927
+ },
+ {
+ "epoch": 26.782532751091704,
+ "grad_norm": 0.01572711206972599,
+ "learning_rate": 0.0006,
+ "loss": 4.189935684204102,
+ "step": 1928
+ },
+ {
+ "epoch": 26.79650655021834,
+ "grad_norm": 0.014976629987359047,
+ "learning_rate": 0.0006,
+ "loss": 4.280525207519531,
+ "step": 1929
+ },
+ {
+ "epoch": 26.810480349344978,
+ "grad_norm": 0.014804664999246597,
+ "learning_rate": 0.0006,
+ "loss": 4.270630836486816,
+ "step": 1930
+ },
+ {
+ "epoch": 26.824454148471617,
+ "grad_norm": 0.014641920104622841,
+ "learning_rate": 0.0006,
+ "loss": 4.319505214691162,
+ "step": 1931
+ },
+ {
+ "epoch": 26.83842794759825,
+ "grad_norm": 0.014928505755960941,
+ "learning_rate": 0.0006,
+ "loss": 4.180147647857666,
+ "step": 1932
+ },
+ {
+ "epoch": 26.85240174672489,
+ "grad_norm": 0.016130365431308746,
+ "learning_rate": 0.0006,
+ "loss": 4.260660648345947,
+ "step": 1933
+ },
+ {
+ "epoch": 26.86637554585153,
+ "grad_norm": 0.015642493963241577,
+ "learning_rate": 0.0006,
+ "loss": 4.304064750671387,
+ "step": 1934
+ },
+ {
+ "epoch": 26.880349344978168,
+ "grad_norm": 0.015988217666745186,
+ "learning_rate": 0.0006,
+ "loss": 4.2548017501831055,
+ "step": 1935
+ },
+ {
+ "epoch": 26.894323144104803,
+ "grad_norm": 0.016501398757100105,
+ "learning_rate": 0.0006,
+ "loss": 4.260058879852295,
+ "step": 1936
+ },
+ {
+ "epoch": 26.90829694323144,
+ "grad_norm": 0.01645040698349476,
+ "learning_rate": 0.0006,
+ "loss": 4.483081340789795,
+ "step": 1937
+ },
+ {
+ "epoch": 26.92227074235808,
+ "grad_norm": 0.016140863299369812,
+ "learning_rate": 0.0006,
+ "loss": 4.1776838302612305,
+ "step": 1938
+ },
+ {
+ "epoch": 26.936244541484715,
+ "grad_norm": 0.016681019216775894,
+ "learning_rate": 0.0006,
+ "loss": 4.224747657775879,
+ "step": 1939
+ },
+ {
+ "epoch": 26.950218340611354,
+ "grad_norm": 0.01650378853082657,
+ "learning_rate": 0.0006,
+ "loss": 4.305703163146973,
+ "step": 1940
+ },
+ {
+ "epoch": 26.964192139737992,
+ "grad_norm": 0.017051683738827705,
+ "learning_rate": 0.0006,
+ "loss": 4.285835266113281,
+ "step": 1941
+ },
+ {
+ "epoch": 26.978165938864628,
+ "grad_norm": 0.016894327476620674,
+ "learning_rate": 0.0006,
+ "loss": 4.318901538848877,
+ "step": 1942
+ },
+ {
+ "epoch": 26.992139737991266,
+ "grad_norm": 0.017313100397586823,
+ "learning_rate": 0.0006,
+ "loss": 4.281434535980225,
+ "step": 1943
+ },
+ {
+ "epoch": 27.0,
+ "grad_norm": 0.018450884148478508,
+ "learning_rate": 0.0006,
+ "loss": 4.335104942321777,
+ "step": 1944
+ },
+ {
+ "epoch": 27.0,
+ "eval_loss": 4.640687465667725,
+ "eval_runtime": 56.9624,
+ "eval_samples_per_second": 42.87,
+ "eval_steps_per_second": 1.352,
+ "step": 1944
+ },
+ {
+ "epoch": 27.01397379912664,
+ "grad_norm": 0.017755698412656784,
+ "learning_rate": 0.0006,
+ "loss": 4.278824329376221,
+ "step": 1945
+ },
+ {
+ "epoch": 27.027947598253274,
+ "grad_norm": 0.01724310778081417,
+ "learning_rate": 0.0006,
+ "loss": 4.311049461364746,
+ "step": 1946
+ },
+ {
+ "epoch": 27.041921397379912,
+ "grad_norm": 0.01690123789012432,
+ "learning_rate": 0.0006,
+ "loss": 4.261994361877441,
+ "step": 1947
+ },
+ {
+ "epoch": 27.05589519650655,
+ "grad_norm": 0.018406696617603302,
+ "learning_rate": 0.0006,
+ "loss": 4.1925153732299805,
+ "step": 1948
+ },
+ {
+ "epoch": 27.069868995633186,
+ "grad_norm": 0.019945867359638214,
+ "learning_rate": 0.0006,
+ "loss": 4.108537197113037,
+ "step": 1949
+ },
+ {
+ "epoch": 27.083842794759825,
+ "grad_norm": 0.020316628739237785,
+ "learning_rate": 0.0006,
+ "loss": 4.209043025970459,
+ "step": 1950
+ },
+ {
+ "epoch": 27.097816593886463,
+ "grad_norm": 0.019690370187163353,
+ "learning_rate": 0.0006,
+ "loss": 4.195789337158203,
+ "step": 1951
+ },
+ {
+ "epoch": 27.111790393013102,
+ "grad_norm": 0.01785232499241829,
+ "learning_rate": 0.0006,
+ "loss": 4.21466064453125,
+ "step": 1952
+ },
+ {
+ "epoch": 27.125764192139737,
+ "grad_norm": 0.0164767038077116,
+ "learning_rate": 0.0006,
+ "loss": 4.270247459411621,
+ "step": 1953
+ },
+ {
+ "epoch": 27.139737991266376,
+ "grad_norm": 0.018008455634117126,
+ "learning_rate": 0.0006,
+ "loss": 4.201042175292969,
+ "step": 1954
+ },
+ {
+ "epoch": 27.153711790393015,
+ "grad_norm": 0.01807340793311596,
+ "learning_rate": 0.0006,
+ "loss": 4.25289249420166,
+ "step": 1955
+ },
+ {
+ "epoch": 27.16768558951965,
+ "grad_norm": 0.016246909275650978,
+ "learning_rate": 0.0006,
+ "loss": 4.116283416748047,
+ "step": 1956
+ },
+ {
+ "epoch": 27.18165938864629,
+ "grad_norm": 0.017545776441693306,
+ "learning_rate": 0.0006,
+ "loss": 4.316999435424805,
+ "step": 1957
+ },
+ {
+ "epoch": 27.195633187772927,
+ "grad_norm": 0.017915375530719757,
+ "learning_rate": 0.0006,
+ "loss": 4.253859519958496,
+ "step": 1958
+ },
+ {
+ "epoch": 27.209606986899562,
+ "grad_norm": 0.018112223595380783,
+ "learning_rate": 0.0006,
+ "loss": 4.2541022300720215,
+ "step": 1959
+ },
+ {
+ "epoch": 27.2235807860262,
+ "grad_norm": 0.01846139505505562,
+ "learning_rate": 0.0006,
+ "loss": 4.20438289642334,
+ "step": 1960
+ },
+ {
+ "epoch": 27.23755458515284,
+ "grad_norm": 0.01943897269666195,
+ "learning_rate": 0.0006,
+ "loss": 4.2799577713012695,
+ "step": 1961
+ },
+ {
+ "epoch": 27.251528384279474,
+ "grad_norm": 0.018253348767757416,
+ "learning_rate": 0.0006,
+ "loss": 4.1720499992370605,
+ "step": 1962
+ },
+ {
+ "epoch": 27.265502183406113,
+ "grad_norm": 0.018292246386408806,
+ "learning_rate": 0.0006,
+ "loss": 4.28231143951416,
+ "step": 1963
+ },
+ {
+ "epoch": 27.27947598253275,
+ "grad_norm": 0.01943155936896801,
+ "learning_rate": 0.0006,
+ "loss": 4.314976692199707,
+ "step": 1964
+ },
+ {
+ "epoch": 27.29344978165939,
+ "grad_norm": 0.019064994528889656,
+ "learning_rate": 0.0006,
+ "loss": 4.25328254699707,
+ "step": 1965
+ },
+ {
+ "epoch": 27.307423580786025,
+ "grad_norm": 0.017423273995518684,
+ "learning_rate": 0.0006,
+ "loss": 4.195992469787598,
+ "step": 1966
+ },
+ {
+ "epoch": 27.321397379912664,
+ "grad_norm": 0.01677924580872059,
+ "learning_rate": 0.0006,
+ "loss": 4.260043144226074,
+ "step": 1967
+ },
+ {
+ "epoch": 27.335371179039303,
+ "grad_norm": 0.01720673404633999,
+ "learning_rate": 0.0006,
+ "loss": 4.222780704498291,
+ "step": 1968
+ },
+ {
+ "epoch": 27.349344978165938,
+ "grad_norm": 0.017300743609666824,
+ "learning_rate": 0.0006,
+ "loss": 4.109524250030518,
+ "step": 1969
+ },
+ {
+ "epoch": 27.363318777292577,
+ "grad_norm": 0.016065871343016624,
+ "learning_rate": 0.0006,
+ "loss": 4.36108922958374,
+ "step": 1970
+ },
+ {
+ "epoch": 27.377292576419215,
+ "grad_norm": 0.016626616939902306,
+ "learning_rate": 0.0006,
+ "loss": 4.220630645751953,
+ "step": 1971
+ },
+ {
+ "epoch": 27.39126637554585,
+ "grad_norm": 0.016482602804899216,
+ "learning_rate": 0.0006,
+ "loss": 4.344797134399414,
+ "step": 1972
+ },
+ {
+ "epoch": 27.40524017467249,
+ "grad_norm": 0.01727953553199768,
+ "learning_rate": 0.0006,
+ "loss": 4.296645164489746,
+ "step": 1973
+ },
+ {
+ "epoch": 27.419213973799128,
+ "grad_norm": 0.01522731315344572,
+ "learning_rate": 0.0006,
+ "loss": 4.223751068115234,
+ "step": 1974
+ },
+ {
+ "epoch": 27.433187772925763,
+ "grad_norm": 0.01581776700913906,
+ "learning_rate": 0.0006,
+ "loss": 4.351379871368408,
+ "step": 1975
+ },
+ {
+ "epoch": 27.4471615720524,
+ "grad_norm": 0.016394605860114098,
+ "learning_rate": 0.0006,
+ "loss": 4.2412919998168945,
+ "step": 1976
+ },
+ {
+ "epoch": 27.46113537117904,
+ "grad_norm": 0.01751169003546238,
+ "learning_rate": 0.0006,
+ "loss": 4.257606506347656,
+ "step": 1977
+ },
+ {
+ "epoch": 27.475109170305675,
+ "grad_norm": 0.017452310770750046,
+ "learning_rate": 0.0006,
+ "loss": 4.380134582519531,
+ "step": 1978
+ },
+ {
+ "epoch": 27.489082969432314,
+ "grad_norm": 0.017741898074746132,
+ "learning_rate": 0.0006,
+ "loss": 4.279632568359375,
+ "step": 1979
+ },
+ {
+ "epoch": 27.503056768558952,
+ "grad_norm": 0.017680590972304344,
+ "learning_rate": 0.0006,
+ "loss": 4.273406028747559,
+ "step": 1980
+ },
+ {
+ "epoch": 27.51703056768559,
+ "grad_norm": 0.01864718459546566,
+ "learning_rate": 0.0006,
+ "loss": 4.278583526611328,
+ "step": 1981
+ },
+ {
+ "epoch": 27.531004366812226,
+ "grad_norm": 0.019558541476726532,
+ "learning_rate": 0.0006,
+ "loss": 4.211919784545898,
+ "step": 1982
+ },
+ {
+ "epoch": 27.544978165938865,
+ "grad_norm": 0.019467400386929512,
+ "learning_rate": 0.0006,
+ "loss": 4.276822566986084,
+ "step": 1983
+ },
+ {
+ "epoch": 27.558951965065503,
+ "grad_norm": 0.01756172813475132,
+ "learning_rate": 0.0006,
+ "loss": 4.165379047393799,
+ "step": 1984
+ },
+ {
+ "epoch": 27.57292576419214,
+ "grad_norm": 0.017458263784646988,
+ "learning_rate": 0.0006,
+ "loss": 4.2992401123046875,
+ "step": 1985
+ },
+ {
+ "epoch": 27.586899563318777,
+ "grad_norm": 0.018089665099978447,
+ "learning_rate": 0.0006,
+ "loss": 4.243378639221191,
+ "step": 1986
+ },
+ {
+ "epoch": 27.600873362445416,
+ "grad_norm": 0.019535524770617485,
+ "learning_rate": 0.0006,
+ "loss": 4.321477890014648,
+ "step": 1987
+ },
+ {
+ "epoch": 27.61484716157205,
+ "grad_norm": 0.01997970975935459,
+ "learning_rate": 0.0006,
+ "loss": 4.123082637786865,
+ "step": 1988
+ },
+ {
+ "epoch": 27.62882096069869,
+ "grad_norm": 0.01913067139685154,
+ "learning_rate": 0.0006,
+ "loss": 4.289896488189697,
+ "step": 1989
+ },
+ {
+ "epoch": 27.64279475982533,
+ "grad_norm": 0.0169806070625782,
+ "learning_rate": 0.0006,
+ "loss": 4.247004985809326,
+ "step": 1990
+ },
+ {
+ "epoch": 27.656768558951963,
+ "grad_norm": 0.017315007746219635,
+ "learning_rate": 0.0006,
+ "loss": 4.2095441818237305,
+ "step": 1991
+ },
+ {
+ "epoch": 27.670742358078602,
+ "grad_norm": 0.015098759904503822,
+ "learning_rate": 0.0006,
+ "loss": 4.145560264587402,
+ "step": 1992
+ },
+ {
+ "epoch": 27.68471615720524,
+ "grad_norm": 0.016240514814853668,
+ "learning_rate": 0.0006,
+ "loss": 4.254927635192871,
+ "step": 1993
+ },
+ {
+ "epoch": 27.69868995633188,
+ "grad_norm": 0.01621108129620552,
+ "learning_rate": 0.0006,
+ "loss": 4.219725608825684,
+ "step": 1994
+ },
+ {
+ "epoch": 27.712663755458514,
+ "grad_norm": 0.015191009268164635,
+ "learning_rate": 0.0006,
+ "loss": 4.194345474243164,
+ "step": 1995
+ },
+ {
+ "epoch": 27.726637554585153,
+ "grad_norm": 0.014387983828783035,
+ "learning_rate": 0.0006,
+ "loss": 4.211467742919922,
+ "step": 1996
+ },
+ {
+ "epoch": 27.74061135371179,
+ "grad_norm": 0.015460561029613018,
+ "learning_rate": 0.0006,
+ "loss": 4.358700752258301,
+ "step": 1997
+ },
+ {
+ "epoch": 27.754585152838427,
+ "grad_norm": 0.015100076794624329,
+ "learning_rate": 0.0006,
+ "loss": 4.171223163604736,
+ "step": 1998
+ },
+ {
+ "epoch": 27.768558951965066,
+ "grad_norm": 0.015607891604304314,
+ "learning_rate": 0.0006,
+ "loss": 4.217874526977539,
+ "step": 1999
+ },
+ {
+ "epoch": 27.782532751091704,
+ "grad_norm": 0.015650689601898193,
+ "learning_rate": 0.0006,
+ "loss": 4.280780792236328,
+ "step": 2000
+ },
+ {
+ "epoch": 27.79650655021834,
+ "grad_norm": 0.01580994389951229,
+ "learning_rate": 0.0006,
+ "loss": 4.159212112426758,
+ "step": 2001
+ },
+ {
+ "epoch": 27.810480349344978,
+ "grad_norm": 0.016607413068413734,
+ "learning_rate": 0.0006,
+ "loss": 4.238853454589844,
+ "step": 2002
+ },
+ {
+ "epoch": 27.824454148471617,
+ "grad_norm": 0.017383860424160957,
+ "learning_rate": 0.0006,
+ "loss": 4.233944892883301,
+ "step": 2003
+ },
+ {
+ "epoch": 27.83842794759825,
+ "grad_norm": 0.014966659247875214,
+ "learning_rate": 0.0006,
+ "loss": 4.167392730712891,
+ "step": 2004
+ },
+ {
+ "epoch": 27.85240174672489,
+ "grad_norm": 0.014589878730475903,
+ "learning_rate": 0.0006,
+ "loss": 4.2390899658203125,
+ "step": 2005
+ },
+ {
+ "epoch": 27.86637554585153,
+ "grad_norm": 0.017179450020194054,
+ "learning_rate": 0.0006,
+ "loss": 4.365785598754883,
+ "step": 2006
+ },
+ {
+ "epoch": 27.880349344978168,
+ "grad_norm": 0.01893901452422142,
+ "learning_rate": 0.0006,
+ "loss": 4.289430141448975,
+ "step": 2007
+ },
+ {
+ "epoch": 27.894323144104803,
+ "grad_norm": 0.018619712442159653,
+ "learning_rate": 0.0006,
+ "loss": 4.310665130615234,
+ "step": 2008
+ },
+ {
+ "epoch": 27.90829694323144,
+ "grad_norm": 0.019180385395884514,
+ "learning_rate": 0.0006,
+ "loss": 4.27454137802124,
+ "step": 2009
+ },
+ {
+ "epoch": 27.92227074235808,
+ "grad_norm": 0.017811523750424385,
+ "learning_rate": 0.0006,
+ "loss": 4.312562465667725,
+ "step": 2010
+ },
+ {
+ "epoch": 27.936244541484715,
+ "grad_norm": 0.016583573073148727,
+ "learning_rate": 0.0006,
+ "loss": 4.202253341674805,
+ "step": 2011
+ },
+ {
+ "epoch": 27.950218340611354,
+ "grad_norm": 0.01658615842461586,
+ "learning_rate": 0.0006,
+ "loss": 4.183433532714844,
+ "step": 2012
+ },
+ {
+ "epoch": 27.964192139737992,
+ "grad_norm": 0.015327401459217072,
+ "learning_rate": 0.0006,
+ "loss": 4.2125139236450195,
+ "step": 2013
+ },
+ {
+ "epoch": 27.978165938864628,
+ "grad_norm": 0.015004601329565048,
+ "learning_rate": 0.0006,
+ "loss": 4.257392883300781,
+ "step": 2014
+ },
+ {
+ "epoch": 27.992139737991266,
+ "grad_norm": 0.015259161591529846,
+ "learning_rate": 0.0006,
+ "loss": 4.241818428039551,
+ "step": 2015
+ },
+ {
+ "epoch": 28.0,
+ "grad_norm": 0.016808776184916496,
+ "learning_rate": 0.0006,
+ "loss": 4.146241188049316,
+ "step": 2016
+ },
+ {
+ "epoch": 28.0,
+ "eval_loss": 4.663090705871582,
+ "eval_runtime": 56.4559,
+ "eval_samples_per_second": 43.255,
+ "eval_steps_per_second": 1.364,
+ "step": 2016
+ },
+ {
+ "epoch": 28.01397379912664,
+ "grad_norm": 0.015784382820129395,
+ "learning_rate": 0.0006,
+ "loss": 4.044191360473633,
+ "step": 2017
+ },
+ {
+ "epoch": 28.027947598253274,
+ "grad_norm": 0.015636710450053215,
+ "learning_rate": 0.0006,
+ "loss": 4.128754615783691,
+ "step": 2018
+ },
+ {
+ "epoch": 28.041921397379912,
+ "grad_norm": 0.015364975668489933,
+ "learning_rate": 0.0006,
+ "loss": 4.191835403442383,
+ "step": 2019
+ },
+ {
+ "epoch": 28.05589519650655,
+ "grad_norm": 0.015015700832009315,
+ "learning_rate": 0.0006,
+ "loss": 4.120022773742676,
+ "step": 2020
+ },
+ {
+ "epoch": 28.069868995633186,
+ "grad_norm": 0.016666799783706665,
+ "learning_rate": 0.0006,
+ "loss": 4.33906364440918,
+ "step": 2021
+ },
+ {
+ "epoch": 28.083842794759825,
+ "grad_norm": 0.01780104823410511,
+ "learning_rate": 0.0006,
+ "loss": 4.279458999633789,
+ "step": 2022
+ },
+ {
+ "epoch": 28.097816593886463,
+ "grad_norm": 0.01870882883667946,
+ "learning_rate": 0.0006,
+ "loss": 4.34728479385376,
+ "step": 2023
+ },
+ {
+ "epoch": 28.111790393013102,
+ "grad_norm": 0.018406381830573082,
+ "learning_rate": 0.0006,
+ "loss": 4.211311340332031,
+ "step": 2024
+ },
+ {
+ "epoch": 28.125764192139737,
+ "grad_norm": 0.017833448946475983,
+ "learning_rate": 0.0006,
+ "loss": 4.223254203796387,
+ "step": 2025
+ },
+ {
+ "epoch": 28.139737991266376,
+ "grad_norm": 0.017552688717842102,
+ "learning_rate": 0.0006,
+ "loss": 4.208823204040527,
+ "step": 2026
+ },
+ {
+ "epoch": 28.153711790393015,
+ "grad_norm": 0.018500229343771935,
+ "learning_rate": 0.0006,
+ "loss": 4.198025226593018,
+ "step": 2027
+ },
+ {
+ "epoch": 28.16768558951965,
+ "grad_norm": 0.01782156340777874,
+ "learning_rate": 0.0006,
+ "loss": 4.179978370666504,
+ "step": 2028
+ },
+ {
+ "epoch": 28.18165938864629,
+ "grad_norm": 0.019580967724323273,
+ "learning_rate": 0.0006,
+ "loss": 4.291114807128906,
+ "step": 2029
+ },
+ {
+ "epoch": 28.195633187772927,
+ "grad_norm": 0.020804036408662796,
+ "learning_rate": 0.0006,
+ "loss": 4.220607280731201,
+ "step": 2030
+ },
+ {
+ "epoch": 28.209606986899562,
+ "grad_norm": 0.019595693796873093,
+ "learning_rate": 0.0006,
+ "loss": 4.2905731201171875,
+ "step": 2031
+ },
+ {
+ "epoch": 28.2235807860262,
+ "grad_norm": 0.019247086718678474,
+ "learning_rate": 0.0006,
+ "loss": 4.2467360496521,
+ "step": 2032
+ },
+ {
+ "epoch": 28.23755458515284,
+ "grad_norm": 0.01958036608994007,
+ "learning_rate": 0.0006,
+ "loss": 4.286137580871582,
+ "step": 2033
+ },
+ {
+ "epoch": 28.251528384279474,
+ "grad_norm": 0.021354753524065018,
+ "learning_rate": 0.0006,
+ "loss": 4.234755039215088,
+ "step": 2034
+ },
+ {
+ "epoch": 28.265502183406113,
+ "grad_norm": 0.02206375077366829,
+ "learning_rate": 0.0006,
+ "loss": 4.149641036987305,
+ "step": 2035
+ },
+ {
+ "epoch": 28.27947598253275,
+ "grad_norm": 0.02201310358941555,
+ "learning_rate": 0.0006,
+ "loss": 4.1928629875183105,
+ "step": 2036
+ },
+ {
+ "epoch": 28.29344978165939,
+ "grad_norm": 0.01961950585246086,
+ "learning_rate": 0.0006,
+ "loss": 4.211404323577881,
+ "step": 2037
+ },
+ {
+ "epoch": 28.307423580786025,
+ "grad_norm": 0.019211795181035995,
+ "learning_rate": 0.0006,
+ "loss": 4.200976848602295,
+ "step": 2038
+ },
+ {
+ "epoch": 28.321397379912664,
+ "grad_norm": 0.01779370754957199,
+ "learning_rate": 0.0006,
+ "loss": 4.249148845672607,
+ "step": 2039
+ },
+ {
+ "epoch": 28.335371179039303,
+ "grad_norm": 0.018635908141732216,
+ "learning_rate": 0.0006,
+ "loss": 4.11806058883667,
+ "step": 2040
+ },
+ {
+ "epoch": 28.349344978165938,
+ "grad_norm": 0.019792694598436356,
+ "learning_rate": 0.0006,
+ "loss": 4.335224151611328,
+ "step": 2041
+ },
+ {
+ "epoch": 28.363318777292577,
+ "grad_norm": 0.021731717512011528,
+ "learning_rate": 0.0006,
+ "loss": 4.302778244018555,
+ "step": 2042
+ },
+ {
+ "epoch": 28.377292576419215,
+ "grad_norm": 0.020823447033762932,
+ "learning_rate": 0.0006,
+ "loss": 4.260610580444336,
+ "step": 2043
+ },
+ {
+ "epoch": 28.39126637554585,
+ "grad_norm": 0.019385412335395813,
+ "learning_rate": 0.0006,
+ "loss": 4.2247724533081055,
+ "step": 2044
+ },
+ {
+ "epoch": 28.40524017467249,
+ "grad_norm": 0.019539108499884605,
+ "learning_rate": 0.0006,
+ "loss": 4.236655235290527,
+ "step": 2045
+ },
+ {
+ "epoch": 28.419213973799128,
+ "grad_norm": 0.01942690648138523,
+ "learning_rate": 0.0006,
+ "loss": 4.251850605010986,
+ "step": 2046
+ },
+ {
+ "epoch": 28.433187772925763,
+ "grad_norm": 0.019667886197566986,
+ "learning_rate": 0.0006,
+ "loss": 4.222312927246094,
+ "step": 2047
+ },
+ {
+ "epoch": 28.4471615720524,
+ "grad_norm": 0.02029012329876423,
+ "learning_rate": 0.0006,
+ "loss": 4.246252059936523,
+ "step": 2048
+ },
+ {
+ "epoch": 28.46113537117904,
+ "grad_norm": 0.01784469373524189,
+ "learning_rate": 0.0006,
+ "loss": 4.192128658294678,
+ "step": 2049
+ },
+ {
+ "epoch": 28.475109170305675,
+ "grad_norm": 0.0160287544131279,
+ "learning_rate": 0.0006,
+ "loss": 4.208607196807861,
+ "step": 2050
+ },
+ {
+ "epoch": 28.489082969432314,
+ "grad_norm": 0.015348346903920174,
+ "learning_rate": 0.0006,
+ "loss": 4.236598014831543,
+ "step": 2051
+ },
+ {
+ "epoch": 28.503056768558952,
+ "grad_norm": 0.015198206529021263,
+ "learning_rate": 0.0006,
+ "loss": 4.235552787780762,
+ "step": 2052
+ },
+ {
+ "epoch": 28.51703056768559,
+ "grad_norm": 0.01523754745721817,
+ "learning_rate": 0.0006,
+ "loss": 4.1941728591918945,
+ "step": 2053
+ },
+ {
+ "epoch": 28.531004366812226,
+ "grad_norm": 0.015559614636003971,
+ "learning_rate": 0.0006,
+ "loss": 4.294757843017578,
+ "step": 2054
+ },
+ {
+ "epoch": 28.544978165938865,
+ "grad_norm": 0.01568358950316906,
+ "learning_rate": 0.0006,
+ "loss": 4.213065147399902,
+ "step": 2055
+ },
+ {
+ "epoch": 28.558951965065503,
+ "grad_norm": 0.015336516313254833,
+ "learning_rate": 0.0006,
+ "loss": 4.240293979644775,
+ "step": 2056
+ },
+ {
+ "epoch": 28.57292576419214,
+ "grad_norm": 0.015041496604681015,
+ "learning_rate": 0.0006,
+ "loss": 4.273397445678711,
+ "step": 2057
+ },
+ {
+ "epoch": 28.586899563318777,
+ "grad_norm": 0.01610128954052925,
+ "learning_rate": 0.0006,
+ "loss": 4.125369071960449,
+ "step": 2058
+ },
+ {
+ "epoch": 28.600873362445416,
+ "grad_norm": 0.015974976122379303,
+ "learning_rate": 0.0006,
+ "loss": 4.133459091186523,
+ "step": 2059
+ },
+ {
+ "epoch": 28.61484716157205,
+ "grad_norm": 0.015936799347400665,
+ "learning_rate": 0.0006,
+ "loss": 4.306667327880859,
+ "step": 2060
+ },
+ {
+ "epoch": 28.62882096069869,
+ "grad_norm": 0.01795247197151184,
+ "learning_rate": 0.0006,
+ "loss": 4.205156326293945,
+ "step": 2061
+ },
+ {
+ "epoch": 28.64279475982533,
+ "grad_norm": 0.019639814272522926,
+ "learning_rate": 0.0006,
+ "loss": 4.277614593505859,
+ "step": 2062
+ },
+ {
+ "epoch": 28.656768558951963,
+ "grad_norm": 0.01722300611436367,
+ "learning_rate": 0.0006,
+ "loss": 4.191032886505127,
+ "step": 2063
+ },
+ {
+ "epoch": 28.670742358078602,
+ "grad_norm": 0.01612633652985096,
+ "learning_rate": 0.0006,
+ "loss": 4.20717191696167,
+ "step": 2064
+ },
+ {
+ "epoch": 28.68471615720524,
+ "grad_norm": 0.01603279449045658,
+ "learning_rate": 0.0006,
+ "loss": 4.228884220123291,
+ "step": 2065
+ },
+ {
+ "epoch": 28.69868995633188,
+ "grad_norm": 0.016405927017331123,
+ "learning_rate": 0.0006,
+ "loss": 4.201292991638184,
+ "step": 2066
+ },
+ {
+ "epoch": 28.712663755458514,
+ "grad_norm": 0.016644183546304703,
+ "learning_rate": 0.0006,
+ "loss": 4.239282131195068,
+ "step": 2067
+ },
+ {
+ "epoch": 28.726637554585153,
+ "grad_norm": 0.0175587497651577,
+ "learning_rate": 0.0006,
+ "loss": 4.289038181304932,
+ "step": 2068
+ },
+ {
+ "epoch": 28.74061135371179,
+ "grad_norm": 0.018223397433757782,
+ "learning_rate": 0.0006,
+ "loss": 4.2266998291015625,
+ "step": 2069
+ },
+ {
+ "epoch": 28.754585152838427,
+ "grad_norm": 0.018680868670344353,
+ "learning_rate": 0.0006,
+ "loss": 4.348849296569824,
+ "step": 2070
+ },
+ {
+ "epoch": 28.768558951965066,
+ "grad_norm": 0.016726728528738022,
+ "learning_rate": 0.0006,
+ "loss": 4.292080879211426,
+ "step": 2071
+ },
+ {
+ "epoch": 28.782532751091704,
+ "grad_norm": 0.017386795952916145,
+ "learning_rate": 0.0006,
+ "loss": 4.208870887756348,
+ "step": 2072
+ },
+ {
+ "epoch": 28.79650655021834,
+ "grad_norm": 0.01865178905427456,
+ "learning_rate": 0.0006,
+ "loss": 4.054150581359863,
+ "step": 2073
+ },
+ {
+ "epoch": 28.810480349344978,
+ "grad_norm": 0.01776815392076969,
+ "learning_rate": 0.0006,
+ "loss": 4.202404499053955,
+ "step": 2074
+ },
+ {
+ "epoch": 28.824454148471617,
+ "grad_norm": 0.01643497310578823,
+ "learning_rate": 0.0006,
+ "loss": 4.131855010986328,
+ "step": 2075
+ },
+ {
+ "epoch": 28.83842794759825,
+ "grad_norm": 0.01574273779988289,
+ "learning_rate": 0.0006,
+ "loss": 4.296788692474365,
+ "step": 2076
+ },
+ {
+ "epoch": 28.85240174672489,
+ "grad_norm": 0.015613908879458904,
+ "learning_rate": 0.0006,
+ "loss": 4.210949897766113,
+ "step": 2077
+ },
+ {
+ "epoch": 28.86637554585153,
+ "grad_norm": 0.015543212182819843,
+ "learning_rate": 0.0006,
+ "loss": 4.218136310577393,
+ "step": 2078
+ },
+ {
+ "epoch": 28.880349344978168,
+ "grad_norm": 0.014681251719594002,
+ "learning_rate": 0.0006,
+ "loss": 4.19401741027832,
+ "step": 2079
+ },
+ {
+ "epoch": 28.894323144104803,
+ "grad_norm": 0.014393828809261322,
+ "learning_rate": 0.0006,
+ "loss": 4.198468208312988,
+ "step": 2080
+ },
+ {
+ "epoch": 28.90829694323144,
+ "grad_norm": 0.01580170728266239,
+ "learning_rate": 0.0006,
+ "loss": 4.224969863891602,
+ "step": 2081
+ },
+ {
+ "epoch": 28.92227074235808,
+ "grad_norm": 0.01484165620058775,
+ "learning_rate": 0.0006,
+ "loss": 4.12236213684082,
+ "step": 2082
+ },
+ {
+ "epoch": 28.936244541484715,
+ "grad_norm": 0.01497623696923256,
+ "learning_rate": 0.0006,
+ "loss": 4.24141788482666,
+ "step": 2083
+ },
+ {
+ "epoch": 28.950218340611354,
+ "grad_norm": 0.015080046840012074,
+ "learning_rate": 0.0006,
+ "loss": 4.284151077270508,
+ "step": 2084
+ },
+ {
+ "epoch": 28.964192139737992,
+ "grad_norm": 0.016207212582230568,
+ "learning_rate": 0.0006,
+ "loss": 4.298693656921387,
+ "step": 2085
+ },
+ {
+ "epoch": 28.978165938864628,
+ "grad_norm": 0.015474417246878147,
+ "learning_rate": 0.0006,
+ "loss": 4.2412261962890625,
+ "step": 2086
+ },
+ {
+ "epoch": 28.992139737991266,
+ "grad_norm": 0.01515351701527834,
+ "learning_rate": 0.0006,
+ "loss": 4.3054423332214355,
+ "step": 2087
+ },
+ {
+ "epoch": 29.0,
+ "grad_norm": 0.016411345452070236,
+ "learning_rate": 0.0006,
+ "loss": 4.104830741882324,
+ "step": 2088
+ },
+ {
+ "epoch": 29.0,
+ "eval_loss": 4.661567687988281,
+ "eval_runtime": 57.1244,
+ "eval_samples_per_second": 42.749,
+ "eval_steps_per_second": 1.348,
+ "step": 2088
+ },
+ {
+ "epoch": 29.01397379912664,
+ "grad_norm": 0.01665342226624489,
+ "learning_rate": 0.0006,
+ "loss": 4.311391830444336,
+ "step": 2089
+ },
+ {
+ "epoch": 29.027947598253274,
+ "grad_norm": 0.018053214997053146,
+ "learning_rate": 0.0006,
+ "loss": 4.135597229003906,
+ "step": 2090
+ },
+ {
+ "epoch": 29.041921397379912,
+ "grad_norm": 0.018213748931884766,
+ "learning_rate": 0.0006,
+ "loss": 4.303299903869629,
+ "step": 2091
+ },
+ {
+ "epoch": 29.05589519650655,
+ "grad_norm": 0.018460242077708244,
+ "learning_rate": 0.0006,
+ "loss": 4.20850944519043,
+ "step": 2092
+ },
+ {
+ "epoch": 29.069868995633186,
+ "grad_norm": 0.018365809693932533,
+ "learning_rate": 0.0006,
+ "loss": 4.275847434997559,
+ "step": 2093
+ },
+ {
+ "epoch": 29.083842794759825,
+ "grad_norm": 0.018346186727285385,
+ "learning_rate": 0.0006,
+ "loss": 4.202091217041016,
+ "step": 2094
+ },
+ {
+ "epoch": 29.097816593886463,
+ "grad_norm": 0.017920587211847305,
+ "learning_rate": 0.0006,
+ "loss": 4.182092666625977,
+ "step": 2095
+ },
+ {
+ "epoch": 29.111790393013102,
+ "grad_norm": 0.01812003180384636,
+ "learning_rate": 0.0006,
+ "loss": 4.259721755981445,
+ "step": 2096
+ },
+ {
+ "epoch": 29.125764192139737,
+ "grad_norm": 0.01700986735522747,
+ "learning_rate": 0.0006,
+ "loss": 4.067573070526123,
+ "step": 2097
+ },
+ {
+ "epoch": 29.139737991266376,
+ "grad_norm": 0.018069760873913765,
+ "learning_rate": 0.0006,
+ "loss": 4.234400749206543,
+ "step": 2098
+ },
+ {
+ "epoch": 29.153711790393015,
+ "grad_norm": 0.01905428245663643,
+ "learning_rate": 0.0006,
+ "loss": 4.16609001159668,
+ "step": 2099
+ },
+ {
+ "epoch": 29.16768558951965,
+ "grad_norm": 0.018711242824792862,
+ "learning_rate": 0.0006,
+ "loss": 4.249088764190674,
+ "step": 2100
+ },
+ {
+ "epoch": 29.18165938864629,
+ "grad_norm": 0.018721306696534157,
+ "learning_rate": 0.0006,
+ "loss": 4.14540958404541,
+ "step": 2101
+ },
+ {
+ "epoch": 29.195633187772927,
+ "grad_norm": 0.020139874890446663,
+ "learning_rate": 0.0006,
+ "loss": 4.049248695373535,
+ "step": 2102
+ },
+ {
+ "epoch": 29.209606986899562,
+ "grad_norm": 0.020849574357271194,
+ "learning_rate": 0.0006,
+ "loss": 4.235224723815918,
+ "step": 2103
+ },
+ {
+ "epoch": 29.2235807860262,
+ "grad_norm": 0.018693549558520317,
+ "learning_rate": 0.0006,
+ "loss": 4.232961177825928,
+ "step": 2104
+ },
+ {
+ "epoch": 29.23755458515284,
+ "grad_norm": 0.017889857292175293,
+ "learning_rate": 0.0006,
+ "loss": 4.171422004699707,
+ "step": 2105
+ },
+ {
+ "epoch": 29.251528384279474,
+ "grad_norm": 0.016690224409103394,
+ "learning_rate": 0.0006,
+ "loss": 4.244717597961426,
+ "step": 2106
+ },
+ {
+ "epoch": 29.265502183406113,
+ "grad_norm": 0.01708616502583027,
+ "learning_rate": 0.0006,
+ "loss": 4.125068664550781,
+ "step": 2107
+ },
+ {
+ "epoch": 29.27947598253275,
+ "grad_norm": 0.017928237095475197,
+ "learning_rate": 0.0006,
+ "loss": 4.208292007446289,
+ "step": 2108
+ },
+ {
+ "epoch": 29.29344978165939,
+ "grad_norm": 0.018250394612550735,
+ "learning_rate": 0.0006,
+ "loss": 4.118630886077881,
+ "step": 2109
+ },
+ {
+ "epoch": 29.307423580786025,
+ "grad_norm": 0.020160246640443802,
+ "learning_rate": 0.0006,
+ "loss": 4.206025123596191,
+ "step": 2110
+ },
+ {
+ "epoch": 29.321397379912664,
+ "grad_norm": 0.020045453682541847,
+ "learning_rate": 0.0006,
+ "loss": 4.2758684158325195,
+ "step": 2111
+ },
+ {
+ "epoch": 29.335371179039303,
+ "grad_norm": 0.019556907936930656,
+ "learning_rate": 0.0006,
+ "loss": 4.188272953033447,
+ "step": 2112
+ },
+ {
+ "epoch": 29.349344978165938,
+ "grad_norm": 0.019185200333595276,
+ "learning_rate": 0.0006,
+ "loss": 4.17451286315918,
+ "step": 2113
+ },
+ {
+ "epoch": 29.363318777292577,
+ "grad_norm": 0.01816629432141781,
+ "learning_rate": 0.0006,
+ "loss": 4.165246963500977,
+ "step": 2114
+ },
+ {
+ "epoch": 29.377292576419215,
+ "grad_norm": 0.01865146867930889,
+ "learning_rate": 0.0006,
+ "loss": 4.148184776306152,
+ "step": 2115
+ },
+ {
+ "epoch": 29.39126637554585,
+ "grad_norm": 0.018338464200496674,
+ "learning_rate": 0.0006,
+ "loss": 4.288936614990234,
+ "step": 2116
+ },
+ {
+ "epoch": 29.40524017467249,
+ "grad_norm": 0.018986187875270844,
+ "learning_rate": 0.0006,
+ "loss": 4.183379650115967,
+ "step": 2117
+ },
+ {
+ "epoch": 29.419213973799128,
+ "grad_norm": 0.018409637734293938,
+ "learning_rate": 0.0006,
+ "loss": 4.057574272155762,
+ "step": 2118
+ },
+ {
+ "epoch": 29.433187772925763,
+ "grad_norm": 0.016664542257785797,
+ "learning_rate": 0.0006,
+ "loss": 4.18517541885376,
+ "step": 2119
+ },
+ {
+ "epoch": 29.4471615720524,
+ "grad_norm": 0.016486089676618576,
+ "learning_rate": 0.0006,
+ "loss": 4.179367542266846,
+ "step": 2120
+ },
+ {
+ "epoch": 29.46113537117904,
+ "grad_norm": 0.01787523552775383,
+ "learning_rate": 0.0006,
+ "loss": 4.184126377105713,
+ "step": 2121
+ },
+ {
+ "epoch": 29.475109170305675,
+ "grad_norm": 0.016574613749980927,
+ "learning_rate": 0.0006,
+ "loss": 4.290170192718506,
+ "step": 2122
+ },
+ {
+ "epoch": 29.489082969432314,
+ "grad_norm": 0.01643376611173153,
+ "learning_rate": 0.0006,
+ "loss": 4.052913665771484,
+ "step": 2123
+ },
+ {
+ "epoch": 29.503056768558952,
+ "grad_norm": 0.016503969207406044,
+ "learning_rate": 0.0006,
+ "loss": 4.335184097290039,
+ "step": 2124
+ },
+ {
+ "epoch": 29.51703056768559,
+ "grad_norm": 0.017596479505300522,
+ "learning_rate": 0.0006,
+ "loss": 4.165539741516113,
+ "step": 2125
+ },
+ {
+ "epoch": 29.531004366812226,
+ "grad_norm": 0.017468582838773727,
+ "learning_rate": 0.0006,
+ "loss": 4.315964698791504,
+ "step": 2126
+ },
+ {
+ "epoch": 29.544978165938865,
+ "grad_norm": 0.01724618673324585,
+ "learning_rate": 0.0006,
+ "loss": 4.243681907653809,
+ "step": 2127
+ },
+ {
+ "epoch": 29.558951965065503,
+ "grad_norm": 0.018384616822004318,
+ "learning_rate": 0.0006,
+ "loss": 4.112447261810303,
+ "step": 2128
+ },
+ {
+ "epoch": 29.57292576419214,
+ "grad_norm": 0.01824074238538742,
+ "learning_rate": 0.0006,
+ "loss": 4.236065864562988,
+ "step": 2129
+ },
+ {
+ "epoch": 29.586899563318777,
+ "grad_norm": 0.01625971868634224,
+ "learning_rate": 0.0006,
+ "loss": 4.18143892288208,
+ "step": 2130
+ },
+ {
+ "epoch": 29.600873362445416,
+ "grad_norm": 0.015537494793534279,
+ "learning_rate": 0.0006,
+ "loss": 4.225147247314453,
+ "step": 2131
+ },
+ {
+ "epoch": 29.61484716157205,
+ "grad_norm": 0.015818974003195763,
+ "learning_rate": 0.0006,
+ "loss": 4.3100175857543945,
+ "step": 2132
+ },
+ {
+ "epoch": 29.62882096069869,
+ "grad_norm": 0.015909671783447266,
+ "learning_rate": 0.0006,
+ "loss": 4.169775009155273,
+ "step": 2133
+ },
+ {
+ "epoch": 29.64279475982533,
+ "grad_norm": 0.01769658550620079,
+ "learning_rate": 0.0006,
+ "loss": 4.278904914855957,
+ "step": 2134
+ },
+ {
+ "epoch": 29.656768558951963,
+ "grad_norm": 0.015641704201698303,
+ "learning_rate": 0.0006,
+ "loss": 4.2426228523254395,
+ "step": 2135
+ },
+ {
+ "epoch": 29.670742358078602,
+ "grad_norm": 0.01565658301115036,
+ "learning_rate": 0.0006,
+ "loss": 4.232687950134277,
+ "step": 2136
+ },
+ {
+ "epoch": 29.68471615720524,
+ "grad_norm": 0.016485940665006638,
+ "learning_rate": 0.0006,
+ "loss": 4.209839820861816,
+ "step": 2137
+ },
+ {
+ "epoch": 29.69868995633188,
+ "grad_norm": 0.017622729763388634,
+ "learning_rate": 0.0006,
+ "loss": 4.196017265319824,
+ "step": 2138
+ },
+ {
+ "epoch": 29.712663755458514,
+ "grad_norm": 0.018320564180612564,
+ "learning_rate": 0.0006,
+ "loss": 4.225683689117432,
+ "step": 2139
+ },
+ {
+ "epoch": 29.726637554585153,
+ "grad_norm": 0.018140794709324837,
+ "learning_rate": 0.0006,
+ "loss": 4.311944961547852,
+ "step": 2140
+ },
+ {
+ "epoch": 29.74061135371179,
+ "grad_norm": 0.016227100044488907,
+ "learning_rate": 0.0006,
+ "loss": 4.230257511138916,
+ "step": 2141
+ },
+ {
+ "epoch": 29.754585152838427,
+ "grad_norm": 0.015759488567709923,
+ "learning_rate": 0.0006,
+ "loss": 4.119174003601074,
+ "step": 2142
+ },
+ {
+ "epoch": 29.768558951965066,
+ "grad_norm": 0.01757766678929329,
+ "learning_rate": 0.0006,
+ "loss": 4.165826797485352,
+ "step": 2143
+ },
+ {
+ "epoch": 29.782532751091704,
+ "grad_norm": 0.019661923870444298,
+ "learning_rate": 0.0006,
+ "loss": 4.392203330993652,
+ "step": 2144
+ },
+ {
+ "epoch": 29.79650655021834,
+ "grad_norm": 0.019478755071759224,
+ "learning_rate": 0.0006,
+ "loss": 4.348371982574463,
+ "step": 2145
+ },
+ {
+ "epoch": 29.810480349344978,
+ "grad_norm": 0.019149569794535637,
+ "learning_rate": 0.0006,
+ "loss": 4.34104061126709,
+ "step": 2146
+ },
+ {
+ "epoch": 29.824454148471617,
+ "grad_norm": 0.017707521095871925,
+ "learning_rate": 0.0006,
+ "loss": 4.228226661682129,
+ "step": 2147
+ },
+ {
+ "epoch": 29.83842794759825,
+ "grad_norm": 0.015615378506481647,
+ "learning_rate": 0.0006,
+ "loss": 4.175541877746582,
+ "step": 2148
+ },
+ {
+ "epoch": 29.85240174672489,
+ "grad_norm": 0.016111129894852638,
+ "learning_rate": 0.0006,
+ "loss": 4.267936706542969,
+ "step": 2149
+ },
+ {
+ "epoch": 29.86637554585153,
+ "grad_norm": 0.01528779324144125,
+ "learning_rate": 0.0006,
+ "loss": 4.186014175415039,
+ "step": 2150
+ },
+ {
+ "epoch": 29.880349344978168,
+ "grad_norm": 0.014925646595656872,
+ "learning_rate": 0.0006,
+ "loss": 4.174860000610352,
+ "step": 2151
+ },
+ {
+ "epoch": 29.894323144104803,
+ "grad_norm": 0.016917673870921135,
+ "learning_rate": 0.0006,
+ "loss": 4.33908748626709,
+ "step": 2152
+ },
+ {
+ "epoch": 29.90829694323144,
+ "grad_norm": 0.016007075086236,
+ "learning_rate": 0.0006,
+ "loss": 4.266141414642334,
+ "step": 2153
+ },
+ {
+ "epoch": 29.92227074235808,
+ "grad_norm": 0.016178956255316734,
+ "learning_rate": 0.0006,
+ "loss": 4.271215915679932,
+ "step": 2154
+ },
+ {
+ "epoch": 29.936244541484715,
+ "grad_norm": 0.016180871054530144,
+ "learning_rate": 0.0006,
+ "loss": 4.278801441192627,
+ "step": 2155
+ },
+ {
+ "epoch": 29.950218340611354,
+ "grad_norm": 0.015335503034293652,
+ "learning_rate": 0.0006,
+ "loss": 4.194978713989258,
+ "step": 2156
+ },
+ {
+ "epoch": 29.964192139737992,
+ "grad_norm": 0.016580764204263687,
+ "learning_rate": 0.0006,
+ "loss": 4.290809631347656,
+ "step": 2157
+ },
+ {
+ "epoch": 29.978165938864628,
+ "grad_norm": 0.01623128168284893,
+ "learning_rate": 0.0006,
+ "loss": 4.205451488494873,
+ "step": 2158
+ },
+ {
+ "epoch": 29.992139737991266,
+ "grad_norm": 0.014904248528182507,
+ "learning_rate": 0.0006,
+ "loss": 4.24924373626709,
+ "step": 2159
+ },
+ {
+ "epoch": 30.0,
+ "grad_norm": 0.017089299857616425,
+ "learning_rate": 0.0006,
+ "loss": 4.222072601318359,
+ "step": 2160
+ },
+ {
+ "epoch": 30.0,
+ "eval_loss": 4.585411071777344,
+ "eval_runtime": 57.1164,
+ "eval_samples_per_second": 42.755,
+ "eval_steps_per_second": 1.348,
+ "step": 2160
+ },
+ {
+ "epoch": 30.01397379912664,
+ "grad_norm": 0.017226383090019226,
+ "learning_rate": 0.0006,
+ "loss": 4.19467830657959,
+ "step": 2161
+ },
+ {
+ "epoch": 30.027947598253274,
+ "grad_norm": 0.01635066606104374,
+ "learning_rate": 0.0006,
+ "loss": 4.176875591278076,
+ "step": 2162
+ },
+ {
+ "epoch": 30.041921397379912,
+ "grad_norm": 0.01735362410545349,
+ "learning_rate": 0.0006,
+ "loss": 4.13667106628418,
+ "step": 2163
+ },
+ {
+ "epoch": 30.05589519650655,
+ "grad_norm": 0.017047109082341194,
+ "learning_rate": 0.0006,
+ "loss": 4.136329650878906,
+ "step": 2164
+ },
+ {
+ "epoch": 30.069868995633186,
+ "grad_norm": 0.018116053193807602,
+ "learning_rate": 0.0006,
+ "loss": 4.228706359863281,
+ "step": 2165
+ },
+ {
+ "epoch": 30.083842794759825,
+ "grad_norm": 0.01791389286518097,
+ "learning_rate": 0.0006,
+ "loss": 4.1597747802734375,
+ "step": 2166
+ },
+ {
+ "epoch": 30.097816593886463,
+ "grad_norm": 0.017793502658605576,
+ "learning_rate": 0.0006,
+ "loss": 4.0807976722717285,
+ "step": 2167
+ },
+ {
+ "epoch": 30.111790393013102,
+ "grad_norm": 0.01827503927052021,
+ "learning_rate": 0.0006,
+ "loss": 4.170646667480469,
+ "step": 2168
+ },
+ {
+ "epoch": 30.125764192139737,
+ "grad_norm": 0.017228564247488976,
+ "learning_rate": 0.0006,
+ "loss": 4.146934509277344,
+ "step": 2169
+ },
+ {
+ "epoch": 30.139737991266376,
+ "grad_norm": 0.01757437363266945,
+ "learning_rate": 0.0006,
+ "loss": 4.1859130859375,
+ "step": 2170
+ },
+ {
+ "epoch": 30.153711790393015,
+ "grad_norm": 0.01701093092560768,
+ "learning_rate": 0.0006,
+ "loss": 4.126981258392334,
+ "step": 2171
+ },
+ {
+ "epoch": 30.16768558951965,
+ "grad_norm": 0.015220489352941513,
+ "learning_rate": 0.0006,
+ "loss": 4.164018630981445,
+ "step": 2172
+ },
+ {
+ "epoch": 30.18165938864629,
+ "grad_norm": 0.016254238784313202,
+ "learning_rate": 0.0006,
+ "loss": 4.258275508880615,
+ "step": 2173
+ },
+ {
+ "epoch": 30.195633187772927,
+ "grad_norm": 0.016609614714980125,
+ "learning_rate": 0.0006,
+ "loss": 4.222233772277832,
+ "step": 2174
+ },
+ {
+ "epoch": 30.209606986899562,
+ "grad_norm": 0.016531318426132202,
+ "learning_rate": 0.0006,
+ "loss": 4.144489288330078,
+ "step": 2175
+ },
+ {
+ "epoch": 30.2235807860262,
+ "grad_norm": 0.01567668654024601,
+ "learning_rate": 0.0006,
+ "loss": 4.181916236877441,
+ "step": 2176
+ },
+ {
+ "epoch": 30.23755458515284,
+ "grad_norm": 0.017195681110024452,
+ "learning_rate": 0.0006,
+ "loss": 4.189693927764893,
+ "step": 2177
+ },
+ {
+ "epoch": 30.251528384279474,
+ "grad_norm": 0.016882948577404022,
+ "learning_rate": 0.0006,
+ "loss": 4.21604061126709,
+ "step": 2178
+ },
+ {
+ "epoch": 30.265502183406113,
+ "grad_norm": 0.016261622309684753,
+ "learning_rate": 0.0006,
+ "loss": 4.227639198303223,
+ "step": 2179
+ },
+ {
+ "epoch": 30.27947598253275,
+ "grad_norm": 0.016411812976002693,
+ "learning_rate": 0.0006,
+ "loss": 4.130362510681152,
+ "step": 2180
+ },
+ {
+ "epoch": 30.29344978165939,
+ "grad_norm": 0.01732019893825054,
+ "learning_rate": 0.0006,
+ "loss": 4.230443000793457,
+ "step": 2181
+ },
+ {
+ "epoch": 30.307423580786025,
+ "grad_norm": 0.01825905591249466,
+ "learning_rate": 0.0006,
+ "loss": 4.175817489624023,
+ "step": 2182
+ },
+ {
+ "epoch": 30.321397379912664,
+ "grad_norm": 0.01887713558971882,
+ "learning_rate": 0.0006,
+ "loss": 4.169020652770996,
+ "step": 2183
+ },
+ {
+ "epoch": 30.335371179039303,
+ "grad_norm": 0.019391661509871483,
+ "learning_rate": 0.0006,
+ "loss": 4.176318645477295,
+ "step": 2184
+ },
+ {
+ "epoch": 30.349344978165938,
+ "grad_norm": 0.020840002223849297,
+ "learning_rate": 0.0006,
+ "loss": 4.174228668212891,
+ "step": 2185
+ },
+ {
+ "epoch": 30.363318777292577,
+ "grad_norm": 0.020193178206682205,
+ "learning_rate": 0.0006,
+ "loss": 4.127919673919678,
+ "step": 2186
+ },
+ {
+ "epoch": 30.377292576419215,
+ "grad_norm": 0.018449561670422554,
+ "learning_rate": 0.0006,
+ "loss": 4.229487895965576,
+ "step": 2187
+ },
+ {
+ "epoch": 30.39126637554585,
+ "grad_norm": 0.01711983233690262,
+ "learning_rate": 0.0006,
+ "loss": 4.201651573181152,
+ "step": 2188
+ },
+ {
+ "epoch": 30.40524017467249,
+ "grad_norm": 0.017293395474553108,
+ "learning_rate": 0.0006,
+ "loss": 4.168704032897949,
+ "step": 2189
+ },
+ {
+ "epoch": 30.419213973799128,
+ "grad_norm": 0.018526358529925346,
+ "learning_rate": 0.0006,
+ "loss": 4.171864032745361,
+ "step": 2190
+ },
+ {
+ "epoch": 30.433187772925763,
+ "grad_norm": 0.019287291914224625,
+ "learning_rate": 0.0006,
+ "loss": 4.039534568786621,
+ "step": 2191
+ },
+ {
+ "epoch": 30.4471615720524,
+ "grad_norm": 0.016902444884181023,
+ "learning_rate": 0.0006,
+ "loss": 4.282787799835205,
+ "step": 2192
+ },
+ {
+ "epoch": 30.46113537117904,
+ "grad_norm": 0.01656750962138176,
+ "learning_rate": 0.0006,
+ "loss": 4.213098526000977,
+ "step": 2193
+ },
+ {
+ "epoch": 30.475109170305675,
+ "grad_norm": 0.016805065795779228,
+ "learning_rate": 0.0006,
+ "loss": 4.1583476066589355,
+ "step": 2194
+ },
+ {
+ "epoch": 30.489082969432314,
+ "grad_norm": 0.017166638746857643,
+ "learning_rate": 0.0006,
+ "loss": 4.188634395599365,
+ "step": 2195
+ },
+ {
+ "epoch": 30.503056768558952,
+ "grad_norm": 0.01592106744647026,
+ "learning_rate": 0.0006,
+ "loss": 4.234043121337891,
+ "step": 2196
+ },
+ {
+ "epoch": 30.51703056768559,
+ "grad_norm": 0.015270989388227463,
+ "learning_rate": 0.0006,
+ "loss": 4.301420211791992,
+ "step": 2197
+ },
+ {
+ "epoch": 30.531004366812226,
+ "grad_norm": 0.015194724313914776,
+ "learning_rate": 0.0006,
+ "loss": 4.18703031539917,
+ "step": 2198
+ },
+ {
+ "epoch": 30.544978165938865,
+ "grad_norm": 0.015587746165692806,
+ "learning_rate": 0.0006,
+ "loss": 4.138928413391113,
+ "step": 2199
+ },
+ {
+ "epoch": 30.558951965065503,
+ "grad_norm": 0.016029570251703262,
+ "learning_rate": 0.0006,
+ "loss": 4.290863990783691,
+ "step": 2200
+ },
+ {
+ "epoch": 30.57292576419214,
+ "grad_norm": 0.015651309862732887,
+ "learning_rate": 0.0006,
+ "loss": 4.272062301635742,
+ "step": 2201
+ },
+ {
+ "epoch": 30.586899563318777,
+ "grad_norm": 0.015992306172847748,
+ "learning_rate": 0.0006,
+ "loss": 4.2487406730651855,
+ "step": 2202
+ },
+ {
+ "epoch": 30.600873362445416,
+ "grad_norm": 0.014969157986342907,
+ "learning_rate": 0.0006,
+ "loss": 4.115347862243652,
+ "step": 2203
+ },
+ {
+ "epoch": 30.61484716157205,
+ "grad_norm": 0.014639027416706085,
+ "learning_rate": 0.0006,
+ "loss": 4.2207489013671875,
+ "step": 2204
+ },
+ {
+ "epoch": 30.62882096069869,
+ "grad_norm": 0.015059413388371468,
+ "learning_rate": 0.0006,
+ "loss": 4.274044990539551,
+ "step": 2205
+ },
+ {
+ "epoch": 30.64279475982533,
+ "grad_norm": 0.0166641678661108,
+ "learning_rate": 0.0006,
+ "loss": 4.191373348236084,
+ "step": 2206
+ },
+ {
+ "epoch": 30.656768558951963,
+ "grad_norm": 0.017806345596909523,
+ "learning_rate": 0.0006,
+ "loss": 4.139800548553467,
+ "step": 2207
+ },
+ {
+ "epoch": 30.670742358078602,
+ "grad_norm": 0.018712421879172325,
+ "learning_rate": 0.0006,
+ "loss": 4.035965442657471,
+ "step": 2208
+ },
+ {
+ "epoch": 30.68471615720524,
+ "grad_norm": 0.018457243219017982,
+ "learning_rate": 0.0006,
+ "loss": 4.2579240798950195,
+ "step": 2209
+ },
+ {
+ "epoch": 30.69868995633188,
+ "grad_norm": 0.01920999586582184,
+ "learning_rate": 0.0006,
+ "loss": 4.131112098693848,
+ "step": 2210
+ },
+ {
+ "epoch": 30.712663755458514,
+ "grad_norm": 0.01729944534599781,
+ "learning_rate": 0.0006,
+ "loss": 4.254059314727783,
+ "step": 2211
+ },
+ {
+ "epoch": 30.726637554585153,
+ "grad_norm": 0.016914231702685356,
+ "learning_rate": 0.0006,
+ "loss": 4.233428955078125,
+ "step": 2212
+ },
+ {
+ "epoch": 30.74061135371179,
+ "grad_norm": 0.01913098618388176,
+ "learning_rate": 0.0006,
+ "loss": 4.23525333404541,
+ "step": 2213
+ },
+ {
+ "epoch": 30.754585152838427,
+ "grad_norm": 0.019150281324982643,
+ "learning_rate": 0.0006,
+ "loss": 4.213695049285889,
+ "step": 2214
+ },
+ {
+ "epoch": 30.768558951965066,
+ "grad_norm": 0.018947452306747437,
+ "learning_rate": 0.0006,
+ "loss": 4.151697158813477,
+ "step": 2215
+ },
+ {
+ "epoch": 30.782532751091704,
+ "grad_norm": 0.016058508306741714,
+ "learning_rate": 0.0006,
+ "loss": 4.220280170440674,
+ "step": 2216
+ },
+ {
+ "epoch": 30.79650655021834,
+ "grad_norm": 0.017488619312644005,
+ "learning_rate": 0.0006,
+ "loss": 4.154428482055664,
+ "step": 2217
+ },
+ {
+ "epoch": 30.810480349344978,
+ "grad_norm": 0.017355090007185936,
+ "learning_rate": 0.0006,
+ "loss": 4.256086349487305,
+ "step": 2218
+ },
+ {
+ "epoch": 30.824454148471617,
+ "grad_norm": 0.016636032611131668,
+ "learning_rate": 0.0006,
+ "loss": 4.227564334869385,
+ "step": 2219
+ },
+ {
+ "epoch": 30.83842794759825,
+ "grad_norm": 0.016273891553282738,
+ "learning_rate": 0.0006,
+ "loss": 4.200466156005859,
+ "step": 2220
+ },
+ {
+ "epoch": 30.85240174672489,
+ "grad_norm": 0.016939911991357803,
+ "learning_rate": 0.0006,
+ "loss": 4.1571221351623535,
+ "step": 2221
+ },
+ {
+ "epoch": 30.86637554585153,
+ "grad_norm": 0.018332522362470627,
+ "learning_rate": 0.0006,
+ "loss": 4.228398323059082,
+ "step": 2222
+ },
+ {
+ "epoch": 30.880349344978168,
+ "grad_norm": 0.019508114084601402,
+ "learning_rate": 0.0006,
+ "loss": 4.222960472106934,
+ "step": 2223
+ },
+ {
+ "epoch": 30.894323144104803,
+ "grad_norm": 0.018440278246998787,
+ "learning_rate": 0.0006,
+ "loss": 4.261051177978516,
+ "step": 2224
+ },
+ {
+ "epoch": 30.90829694323144,
+ "grad_norm": 0.018113229423761368,
+ "learning_rate": 0.0006,
+ "loss": 4.109235763549805,
+ "step": 2225
+ },
+ {
+ "epoch": 30.92227074235808,
+ "grad_norm": 0.015392083674669266,
+ "learning_rate": 0.0006,
+ "loss": 4.21950101852417,
+ "step": 2226
+ },
+ {
+ "epoch": 30.936244541484715,
+ "grad_norm": 0.01633290946483612,
+ "learning_rate": 0.0006,
+ "loss": 4.176053047180176,
+ "step": 2227
+ },
+ {
+ "epoch": 30.950218340611354,
+ "grad_norm": 0.016236383467912674,
+ "learning_rate": 0.0006,
+ "loss": 4.314309120178223,
+ "step": 2228
+ },
+ {
+ "epoch": 30.964192139737992,
+ "grad_norm": 0.015284373424947262,
+ "learning_rate": 0.0006,
+ "loss": 4.303133487701416,
+ "step": 2229
+ },
+ {
+ "epoch": 30.978165938864628,
+ "grad_norm": 0.015982897952198982,
+ "learning_rate": 0.0006,
+ "loss": 4.25739860534668,
+ "step": 2230
+ },
+ {
+ "epoch": 30.992139737991266,
+ "grad_norm": 0.016682198271155357,
+ "learning_rate": 0.0006,
+ "loss": 4.292210578918457,
+ "step": 2231
+ },
+ {
+ "epoch": 31.0,
+ "grad_norm": 0.017521467059850693,
+ "learning_rate": 0.0006,
+ "loss": 4.170637130737305,
+ "step": 2232
+ },
+ {
+ "epoch": 31.0,
+ "eval_loss": 4.642786979675293,
+ "eval_runtime": 57.029,
+ "eval_samples_per_second": 42.82,
+ "eval_steps_per_second": 1.35,
+ "step": 2232
+ },
+ {
+ "epoch": 31.01397379912664,
+ "grad_norm": 0.016607852652668953,
+ "learning_rate": 0.0006,
+ "loss": 4.106224536895752,
+ "step": 2233
+ },
+ {
+ "epoch": 31.027947598253274,
+ "grad_norm": 0.018448904156684875,
+ "learning_rate": 0.0006,
+ "loss": 4.224459648132324,
+ "step": 2234
+ },
+ {
+ "epoch": 31.041921397379912,
+ "grad_norm": 0.020481228828430176,
+ "learning_rate": 0.0006,
+ "loss": 4.305548667907715,
+ "step": 2235
+ },
+ {
+ "epoch": 31.05589519650655,
+ "grad_norm": 0.01999707892537117,
+ "learning_rate": 0.0006,
+ "loss": 4.253663539886475,
+ "step": 2236
+ },
+ {
+ "epoch": 31.069868995633186,
+ "grad_norm": 0.020428303629159927,
+ "learning_rate": 0.0006,
+ "loss": 4.166954040527344,
+ "step": 2237
+ },
+ {
+ "epoch": 31.083842794759825,
+ "grad_norm": 0.022608213126659393,
+ "learning_rate": 0.0006,
+ "loss": 4.172819137573242,
+ "step": 2238
+ },
+ {
+ "epoch": 31.097816593886463,
+ "grad_norm": 0.024669520556926727,
+ "learning_rate": 0.0006,
+ "loss": 4.106586456298828,
+ "step": 2239
+ },
+ {
+ "epoch": 31.111790393013102,
+ "grad_norm": 0.02247351035475731,
+ "learning_rate": 0.0006,
+ "loss": 4.160271644592285,
+ "step": 2240
+ },
+ {
+ "epoch": 31.125764192139737,
+ "grad_norm": 0.0207810141146183,
+ "learning_rate": 0.0006,
+ "loss": 4.272491931915283,
+ "step": 2241
+ },
+ {
+ "epoch": 31.139737991266376,
+ "grad_norm": 0.021576499566435814,
+ "learning_rate": 0.0006,
+ "loss": 4.171474456787109,
+ "step": 2242
+ },
+ {
+ "epoch": 31.153711790393015,
+ "grad_norm": 0.02337106689810753,
+ "learning_rate": 0.0006,
+ "loss": 4.094712734222412,
+ "step": 2243
+ },
+ {
+ "epoch": 31.16768558951965,
+ "grad_norm": 0.02219875156879425,
+ "learning_rate": 0.0006,
+ "loss": 4.1800994873046875,
+ "step": 2244
+ },
+ {
+ "epoch": 31.18165938864629,
+ "grad_norm": 0.022680504247546196,
+ "learning_rate": 0.0006,
+ "loss": 4.2054362297058105,
+ "step": 2245
+ },
+ {
+ "epoch": 31.195633187772927,
+ "grad_norm": 0.0208734143525362,
+ "learning_rate": 0.0006,
+ "loss": 4.06068229675293,
+ "step": 2246
+ },
+ {
+ "epoch": 31.209606986899562,
+ "grad_norm": 0.01811066083610058,
+ "learning_rate": 0.0006,
+ "loss": 4.2465128898620605,
+ "step": 2247
+ },
+ {
+ "epoch": 31.2235807860262,
+ "grad_norm": 0.018571637570858,
+ "learning_rate": 0.0006,
+ "loss": 4.191955089569092,
+ "step": 2248
+ },
+ {
+ "epoch": 31.23755458515284,
+ "grad_norm": 0.018598807975649834,
+ "learning_rate": 0.0006,
+ "loss": 4.200872421264648,
+ "step": 2249
+ },
+ {
+ "epoch": 31.251528384279474,
+ "grad_norm": 0.017577625811100006,
+ "learning_rate": 0.0006,
+ "loss": 4.148700714111328,
+ "step": 2250
+ },
+ {
+ "epoch": 31.265502183406113,
+ "grad_norm": 0.01817980408668518,
+ "learning_rate": 0.0006,
+ "loss": 4.256978988647461,
+ "step": 2251
+ },
+ {
+ "epoch": 31.27947598253275,
+ "grad_norm": 0.017983486875891685,
+ "learning_rate": 0.0006,
+ "loss": 4.175005912780762,
+ "step": 2252
+ },
+ {
+ "epoch": 31.29344978165939,
+ "grad_norm": 0.016751961782574654,
+ "learning_rate": 0.0006,
+ "loss": 4.237518310546875,
+ "step": 2253
+ },
+ {
+ "epoch": 31.307423580786025,
+ "grad_norm": 0.016407202929258347,
+ "learning_rate": 0.0006,
+ "loss": 4.197269439697266,
+ "step": 2254
+ },
+ {
+ "epoch": 31.321397379912664,
+ "grad_norm": 0.01578645594418049,
+ "learning_rate": 0.0006,
+ "loss": 4.179751396179199,
+ "step": 2255
+ },
+ {
+ "epoch": 31.335371179039303,
+ "grad_norm": 0.016001226380467415,
+ "learning_rate": 0.0006,
+ "loss": 4.08821964263916,
+ "step": 2256
+ },
+ {
+ "epoch": 31.349344978165938,
+ "grad_norm": 0.015122811309993267,
+ "learning_rate": 0.0006,
+ "loss": 4.047828674316406,
+ "step": 2257
+ },
+ {
+ "epoch": 31.363318777292577,
+ "grad_norm": 0.015399965457618237,
+ "learning_rate": 0.0006,
+ "loss": 4.1506524085998535,
+ "step": 2258
+ },
+ {
+ "epoch": 31.377292576419215,
+ "grad_norm": 0.01534605398774147,
+ "learning_rate": 0.0006,
+ "loss": 4.27952766418457,
+ "step": 2259
+ },
+ {
+ "epoch": 31.39126637554585,
+ "grad_norm": 0.015461381524801254,
+ "learning_rate": 0.0006,
+ "loss": 4.27102518081665,
+ "step": 2260
+ },
+ {
+ "epoch": 31.40524017467249,
+ "grad_norm": 0.016769153997302055,
+ "learning_rate": 0.0006,
+ "loss": 4.191592216491699,
+ "step": 2261
+ },
+ {
+ "epoch": 31.419213973799128,
+ "grad_norm": 0.016557959839701653,
+ "learning_rate": 0.0006,
+ "loss": 4.159705638885498,
+ "step": 2262
+ },
+ {
+ "epoch": 31.433187772925763,
+ "grad_norm": 0.01659569889307022,
+ "learning_rate": 0.0006,
+ "loss": 4.2136664390563965,
+ "step": 2263
+ },
+ {
+ "epoch": 31.4471615720524,
+ "grad_norm": 0.017101937904953957,
+ "learning_rate": 0.0006,
+ "loss": 4.2716498374938965,
+ "step": 2264
+ },
+ {
+ "epoch": 31.46113537117904,
+ "grad_norm": 0.01722985878586769,
+ "learning_rate": 0.0006,
+ "loss": 4.300534248352051,
+ "step": 2265
+ },
+ {
+ "epoch": 31.475109170305675,
+ "grad_norm": 0.016800185665488243,
+ "learning_rate": 0.0006,
+ "loss": 4.113240718841553,
+ "step": 2266
+ },
+ {
+ "epoch": 31.489082969432314,
+ "grad_norm": 0.01603875495493412,
+ "learning_rate": 0.0006,
+ "loss": 4.157105445861816,
+ "step": 2267
+ },
+ {
+ "epoch": 31.503056768558952,
+ "grad_norm": 0.015903932973742485,
+ "learning_rate": 0.0006,
+ "loss": 4.233669757843018,
+ "step": 2268
+ },
+ {
+ "epoch": 31.51703056768559,
+ "grad_norm": 0.0160983856767416,
+ "learning_rate": 0.0006,
+ "loss": 4.147704124450684,
+ "step": 2269
+ },
+ {
+ "epoch": 31.531004366812226,
+ "grad_norm": 0.01604730263352394,
+ "learning_rate": 0.0006,
+ "loss": 4.1465911865234375,
+ "step": 2270
+ },
+ {
+ "epoch": 31.544978165938865,
+ "grad_norm": 0.016205785796046257,
+ "learning_rate": 0.0006,
+ "loss": 4.106704235076904,
+ "step": 2271
+ },
+ {
+ "epoch": 31.558951965065503,
+ "grad_norm": 0.017008036375045776,
+ "learning_rate": 0.0006,
+ "loss": 4.191000938415527,
+ "step": 2272
+ },
+ {
+ "epoch": 31.57292576419214,
+ "grad_norm": 0.01699635200202465,
+ "learning_rate": 0.0006,
+ "loss": 4.187042236328125,
+ "step": 2273
+ },
+ {
+ "epoch": 31.586899563318777,
+ "grad_norm": 0.01692967116832733,
+ "learning_rate": 0.0006,
+ "loss": 4.208805561065674,
+ "step": 2274
+ },
+ {
+ "epoch": 31.600873362445416,
+ "grad_norm": 0.017608756199479103,
+ "learning_rate": 0.0006,
+ "loss": 4.200512886047363,
+ "step": 2275
+ },
+ {
+ "epoch": 31.61484716157205,
+ "grad_norm": 0.016590990126132965,
+ "learning_rate": 0.0006,
+ "loss": 4.194705009460449,
+ "step": 2276
+ },
+ {
+ "epoch": 31.62882096069869,
+ "grad_norm": 0.0170235987752676,
+ "learning_rate": 0.0006,
+ "loss": 4.210326194763184,
+ "step": 2277
+ },
+ {
+ "epoch": 31.64279475982533,
+ "grad_norm": 0.01604801043868065,
+ "learning_rate": 0.0006,
+ "loss": 4.196024417877197,
+ "step": 2278
+ },
+ {
+ "epoch": 31.656768558951963,
+ "grad_norm": 0.015694202855229378,
+ "learning_rate": 0.0006,
+ "loss": 4.178236484527588,
+ "step": 2279
+ },
+ {
+ "epoch": 31.670742358078602,
+ "grad_norm": 0.015473959036171436,
+ "learning_rate": 0.0006,
+ "loss": 4.259339809417725,
+ "step": 2280
+ },
+ {
+ "epoch": 31.68471615720524,
+ "grad_norm": 0.014689362607896328,
+ "learning_rate": 0.0006,
+ "loss": 3.9702324867248535,
+ "step": 2281
+ },
+ {
+ "epoch": 31.69868995633188,
+ "grad_norm": 0.015245123766362667,
+ "learning_rate": 0.0006,
+ "loss": 4.197025775909424,
+ "step": 2282
+ },
+ {
+ "epoch": 31.712663755458514,
+ "grad_norm": 0.016148347407579422,
+ "learning_rate": 0.0006,
+ "loss": 4.231856346130371,
+ "step": 2283
+ },
+ {
+ "epoch": 31.726637554585153,
+ "grad_norm": 0.015169818885624409,
+ "learning_rate": 0.0006,
+ "loss": 4.230169296264648,
+ "step": 2284
+ },
+ {
+ "epoch": 31.74061135371179,
+ "grad_norm": 0.014843013137578964,
+ "learning_rate": 0.0006,
+ "loss": 4.195633411407471,
+ "step": 2285
+ },
+ {
+ "epoch": 31.754585152838427,
+ "grad_norm": 0.015229340642690659,
+ "learning_rate": 0.0006,
+ "loss": 4.231280326843262,
+ "step": 2286
+ },
+ {
+ "epoch": 31.768558951965066,
+ "grad_norm": 0.01551708485931158,
+ "learning_rate": 0.0006,
+ "loss": 4.17064094543457,
+ "step": 2287
+ },
+ {
+ "epoch": 31.782532751091704,
+ "grad_norm": 0.015863575041294098,
+ "learning_rate": 0.0006,
+ "loss": 4.1579484939575195,
+ "step": 2288
+ },
+ {
+ "epoch": 31.79650655021834,
+ "grad_norm": 0.015118198469281197,
+ "learning_rate": 0.0006,
+ "loss": 4.1626787185668945,
+ "step": 2289
+ },
+ {
+ "epoch": 31.810480349344978,
+ "grad_norm": 0.015220776200294495,
+ "learning_rate": 0.0006,
+ "loss": 4.135746955871582,
+ "step": 2290
+ },
+ {
+ "epoch": 31.824454148471617,
+ "grad_norm": 0.014529784210026264,
+ "learning_rate": 0.0006,
+ "loss": 4.164931297302246,
+ "step": 2291
+ },
+ {
+ "epoch": 31.83842794759825,
+ "grad_norm": 0.015533296391367912,
+ "learning_rate": 0.0006,
+ "loss": 4.112464904785156,
+ "step": 2292
+ },
+ {
+ "epoch": 31.85240174672489,
+ "grad_norm": 0.016562335193157196,
+ "learning_rate": 0.0006,
+ "loss": 4.164898872375488,
+ "step": 2293
+ },
+ {
+ "epoch": 31.86637554585153,
+ "grad_norm": 0.016258113086223602,
+ "learning_rate": 0.0006,
+ "loss": 4.1979265213012695,
+ "step": 2294
+ },
+ {
+ "epoch": 31.880349344978168,
+ "grad_norm": 0.01558110024780035,
+ "learning_rate": 0.0006,
+ "loss": 4.174675941467285,
+ "step": 2295
+ },
+ {
+ "epoch": 31.894323144104803,
+ "grad_norm": 0.016412515193223953,
+ "learning_rate": 0.0006,
+ "loss": 4.217909336090088,
+ "step": 2296
+ },
+ {
+ "epoch": 31.90829694323144,
+ "grad_norm": 0.015118442475795746,
+ "learning_rate": 0.0006,
+ "loss": 4.137143611907959,
+ "step": 2297
+ },
+ {
+ "epoch": 31.92227074235808,
+ "grad_norm": 0.01506161317229271,
+ "learning_rate": 0.0006,
+ "loss": 4.196122169494629,
+ "step": 2298
+ },
+ {
+ "epoch": 31.936244541484715,
+ "grad_norm": 0.017165271565318108,
+ "learning_rate": 0.0006,
+ "loss": 4.14834451675415,
+ "step": 2299
+ },
+ {
+ "epoch": 31.950218340611354,
+ "grad_norm": 0.01696440577507019,
+ "learning_rate": 0.0006,
+ "loss": 4.218868732452393,
+ "step": 2300
+ },
+ {
+ "epoch": 31.964192139737992,
+ "grad_norm": 0.017349394038319588,
+ "learning_rate": 0.0006,
+ "loss": 4.2754011154174805,
+ "step": 2301
+ },
+ {
+ "epoch": 31.978165938864628,
+ "grad_norm": 0.018387921154499054,
+ "learning_rate": 0.0006,
+ "loss": 4.282623291015625,
+ "step": 2302
+ },
+ {
+ "epoch": 31.992139737991266,
+ "grad_norm": 0.018837926909327507,
+ "learning_rate": 0.0006,
+ "loss": 4.185425281524658,
+ "step": 2303
+ },
+ {
+ "epoch": 32.0,
+ "grad_norm": 0.018998464569449425,
+ "learning_rate": 0.0006,
+ "loss": 4.151291847229004,
+ "step": 2304
+ },
+ {
+ "epoch": 32.0,
+ "eval_loss": 4.522614002227783,
+ "eval_runtime": 57.0106,
+ "eval_samples_per_second": 42.834,
+ "eval_steps_per_second": 1.351,
+ "step": 2304
+ },
+ {
+ "epoch": 32.01397379912664,
+ "grad_norm": 0.019036833196878433,
+ "learning_rate": 0.0006,
+ "loss": 4.183202266693115,
+ "step": 2305
+ },
+ {
+ "epoch": 32.02794759825328,
+ "grad_norm": 0.02207397110760212,
+ "learning_rate": 0.0006,
+ "loss": 4.17875862121582,
+ "step": 2306
+ },
+ {
+ "epoch": 32.041921397379916,
+ "grad_norm": 0.022008279338479042,
+ "learning_rate": 0.0006,
+ "loss": 4.08042049407959,
+ "step": 2307
+ },
+ {
+ "epoch": 32.05589519650655,
+ "grad_norm": 0.022407017648220062,
+ "learning_rate": 0.0006,
+ "loss": 4.112189769744873,
+ "step": 2308
+ },
+ {
+ "epoch": 32.069868995633186,
+ "grad_norm": 0.021599747240543365,
+ "learning_rate": 0.0006,
+ "loss": 4.126455307006836,
+ "step": 2309
+ },
+ {
+ "epoch": 32.083842794759825,
+ "grad_norm": 0.020340966060757637,
+ "learning_rate": 0.0006,
+ "loss": 4.080538272857666,
+ "step": 2310
+ },
+ {
+ "epoch": 32.09781659388646,
+ "grad_norm": 0.020956892520189285,
+ "learning_rate": 0.0006,
+ "loss": 4.171943664550781,
+ "step": 2311
+ },
+ {
+ "epoch": 32.1117903930131,
+ "grad_norm": 0.019966667518019676,
+ "learning_rate": 0.0006,
+ "loss": 4.144007682800293,
+ "step": 2312
+ },
+ {
+ "epoch": 32.12576419213974,
+ "grad_norm": 0.018256952986121178,
+ "learning_rate": 0.0006,
+ "loss": 4.208107948303223,
+ "step": 2313
+ },
+ {
+ "epoch": 32.13973799126637,
+ "grad_norm": 0.017289170995354652,
+ "learning_rate": 0.0006,
+ "loss": 4.182024955749512,
+ "step": 2314
+ },
+ {
+ "epoch": 32.15371179039301,
+ "grad_norm": 0.017656700685620308,
+ "learning_rate": 0.0006,
+ "loss": 4.068373680114746,
+ "step": 2315
+ },
+ {
+ "epoch": 32.16768558951965,
+ "grad_norm": 0.017903361469507217,
+ "learning_rate": 0.0006,
+ "loss": 4.2345428466796875,
+ "step": 2316
+ },
+ {
+ "epoch": 32.18165938864629,
+ "grad_norm": 0.01785109005868435,
+ "learning_rate": 0.0006,
+ "loss": 4.219776153564453,
+ "step": 2317
+ },
+ {
+ "epoch": 32.19563318777293,
+ "grad_norm": 0.01887257769703865,
+ "learning_rate": 0.0006,
+ "loss": 4.139206409454346,
+ "step": 2318
+ },
+ {
+ "epoch": 32.209606986899566,
+ "grad_norm": 0.01924685761332512,
+ "learning_rate": 0.0006,
+ "loss": 4.121052265167236,
+ "step": 2319
+ },
+ {
+ "epoch": 32.223580786026204,
+ "grad_norm": 0.020574072375893593,
+ "learning_rate": 0.0006,
+ "loss": 4.13262414932251,
+ "step": 2320
+ },
+ {
+ "epoch": 32.237554585152836,
+ "grad_norm": 0.023849064484238625,
+ "learning_rate": 0.0006,
+ "loss": 4.239119529724121,
+ "step": 2321
+ },
+ {
+ "epoch": 32.251528384279474,
+ "grad_norm": 0.026239361613988876,
+ "learning_rate": 0.0006,
+ "loss": 4.248993873596191,
+ "step": 2322
+ },
+ {
+ "epoch": 32.26550218340611,
+ "grad_norm": 0.023980243131518364,
+ "learning_rate": 0.0006,
+ "loss": 4.159852504730225,
+ "step": 2323
+ },
+ {
+ "epoch": 32.27947598253275,
+ "grad_norm": 0.020442405715584755,
+ "learning_rate": 0.0006,
+ "loss": 4.133546829223633,
+ "step": 2324
+ },
+ {
+ "epoch": 32.29344978165939,
+ "grad_norm": 0.020661186426877975,
+ "learning_rate": 0.0006,
+ "loss": 4.078243732452393,
+ "step": 2325
+ },
+ {
+ "epoch": 32.30742358078603,
+ "grad_norm": 0.02046387456357479,
+ "learning_rate": 0.0006,
+ "loss": 4.172012805938721,
+ "step": 2326
+ },
+ {
+ "epoch": 32.32139737991266,
+ "grad_norm": 0.02089597098529339,
+ "learning_rate": 0.0006,
+ "loss": 4.0604705810546875,
+ "step": 2327
+ },
+ {
+ "epoch": 32.3353711790393,
+ "grad_norm": 0.01875591278076172,
+ "learning_rate": 0.0006,
+ "loss": 4.125911235809326,
+ "step": 2328
+ },
+ {
+ "epoch": 32.34934497816594,
+ "grad_norm": 0.018606796860694885,
+ "learning_rate": 0.0006,
+ "loss": 4.097362518310547,
+ "step": 2329
+ },
+ {
+ "epoch": 32.36331877729258,
+ "grad_norm": 0.01766626164317131,
+ "learning_rate": 0.0006,
+ "loss": 4.118616580963135,
+ "step": 2330
+ },
+ {
+ "epoch": 32.377292576419215,
+ "grad_norm": 0.017692960798740387,
+ "learning_rate": 0.0006,
+ "loss": 4.1457672119140625,
+ "step": 2331
+ },
+ {
+ "epoch": 32.391266375545854,
+ "grad_norm": 0.017619704827666283,
+ "learning_rate": 0.0006,
+ "loss": 4.175961494445801,
+ "step": 2332
+ },
+ {
+ "epoch": 32.40524017467249,
+ "grad_norm": 0.01784413494169712,
+ "learning_rate": 0.0006,
+ "loss": 4.051291465759277,
+ "step": 2333
+ },
+ {
+ "epoch": 32.419213973799124,
+ "grad_norm": 0.018508171662688255,
+ "learning_rate": 0.0006,
+ "loss": 4.0638556480407715,
+ "step": 2334
+ },
+ {
+ "epoch": 32.43318777292576,
+ "grad_norm": 0.01731436885893345,
+ "learning_rate": 0.0006,
+ "loss": 4.1805419921875,
+ "step": 2335
+ },
+ {
+ "epoch": 32.4471615720524,
+ "grad_norm": 0.015886498615145683,
+ "learning_rate": 0.0006,
+ "loss": 4.151026725769043,
+ "step": 2336
+ },
+ {
+ "epoch": 32.46113537117904,
+ "grad_norm": 0.016583820804953575,
+ "learning_rate": 0.0006,
+ "loss": 4.203604698181152,
+ "step": 2337
+ },
+ {
+ "epoch": 32.47510917030568,
+ "grad_norm": 0.0176097322255373,
+ "learning_rate": 0.0006,
+ "loss": 4.135159015655518,
+ "step": 2338
+ },
+ {
+ "epoch": 32.48908296943232,
+ "grad_norm": 0.016657903790473938,
+ "learning_rate": 0.0006,
+ "loss": 4.126681804656982,
+ "step": 2339
+ },
+ {
+ "epoch": 32.50305676855895,
+ "grad_norm": 0.016866130754351616,
+ "learning_rate": 0.0006,
+ "loss": 4.162341117858887,
+ "step": 2340
+ },
+ {
+ "epoch": 32.51703056768559,
+ "grad_norm": 0.016286680474877357,
+ "learning_rate": 0.0006,
+ "loss": 4.261801242828369,
+ "step": 2341
+ },
+ {
+ "epoch": 32.531004366812226,
+ "grad_norm": 0.01481365505605936,
+ "learning_rate": 0.0006,
+ "loss": 4.110016822814941,
+ "step": 2342
+ },
+ {
+ "epoch": 32.544978165938865,
+ "grad_norm": 0.016605013981461525,
+ "learning_rate": 0.0006,
+ "loss": 4.230024814605713,
+ "step": 2343
+ },
+ {
+ "epoch": 32.5589519650655,
+ "grad_norm": 0.01645175740122795,
+ "learning_rate": 0.0006,
+ "loss": 4.174612045288086,
+ "step": 2344
+ },
+ {
+ "epoch": 32.57292576419214,
+ "grad_norm": 0.014725026674568653,
+ "learning_rate": 0.0006,
+ "loss": 4.066371440887451,
+ "step": 2345
+ },
+ {
+ "epoch": 32.58689956331878,
+ "grad_norm": 0.015003837645053864,
+ "learning_rate": 0.0006,
+ "loss": 4.193343639373779,
+ "step": 2346
+ },
+ {
+ "epoch": 32.60087336244541,
+ "grad_norm": 0.015346302650868893,
+ "learning_rate": 0.0006,
+ "loss": 4.210659980773926,
+ "step": 2347
+ },
+ {
+ "epoch": 32.61484716157205,
+ "grad_norm": 0.015135962516069412,
+ "learning_rate": 0.0006,
+ "loss": 4.204618453979492,
+ "step": 2348
+ },
+ {
+ "epoch": 32.62882096069869,
+ "grad_norm": 0.01601956970989704,
+ "learning_rate": 0.0006,
+ "loss": 4.145564556121826,
+ "step": 2349
+ },
+ {
+ "epoch": 32.64279475982533,
+ "grad_norm": 0.01517851185053587,
+ "learning_rate": 0.0006,
+ "loss": 4.051673889160156,
+ "step": 2350
+ },
+ {
+ "epoch": 32.65676855895197,
+ "grad_norm": 0.014894292689859867,
+ "learning_rate": 0.0006,
+ "loss": 4.184818267822266,
+ "step": 2351
+ },
+ {
+ "epoch": 32.670742358078606,
+ "grad_norm": 0.015720834955573082,
+ "learning_rate": 0.0006,
+ "loss": 4.220036506652832,
+ "step": 2352
+ },
+ {
+ "epoch": 32.68471615720524,
+ "grad_norm": 0.01573433168232441,
+ "learning_rate": 0.0006,
+ "loss": 4.076773166656494,
+ "step": 2353
+ },
+ {
+ "epoch": 32.698689956331876,
+ "grad_norm": 0.015828052535653114,
+ "learning_rate": 0.0006,
+ "loss": 4.116329669952393,
+ "step": 2354
+ },
+ {
+ "epoch": 32.712663755458514,
+ "grad_norm": 0.015462259761989117,
+ "learning_rate": 0.0006,
+ "loss": 4.1155242919921875,
+ "step": 2355
+ },
+ {
+ "epoch": 32.72663755458515,
+ "grad_norm": 0.015763849020004272,
+ "learning_rate": 0.0006,
+ "loss": 4.169687747955322,
+ "step": 2356
+ },
+ {
+ "epoch": 32.74061135371179,
+ "grad_norm": 0.014935145154595375,
+ "learning_rate": 0.0006,
+ "loss": 4.2090911865234375,
+ "step": 2357
+ },
+ {
+ "epoch": 32.75458515283843,
+ "grad_norm": 0.015119184739887714,
+ "learning_rate": 0.0006,
+ "loss": 4.131838798522949,
+ "step": 2358
+ },
+ {
+ "epoch": 32.76855895196506,
+ "grad_norm": 0.014833525754511356,
+ "learning_rate": 0.0006,
+ "loss": 4.17618465423584,
+ "step": 2359
+ },
+ {
+ "epoch": 32.7825327510917,
+ "grad_norm": 0.01588359661400318,
+ "learning_rate": 0.0006,
+ "loss": 4.163723945617676,
+ "step": 2360
+ },
+ {
+ "epoch": 32.79650655021834,
+ "grad_norm": 0.0168188214302063,
+ "learning_rate": 0.0006,
+ "loss": 4.178189277648926,
+ "step": 2361
+ },
+ {
+ "epoch": 32.81048034934498,
+ "grad_norm": 0.01726095750927925,
+ "learning_rate": 0.0006,
+ "loss": 4.144223690032959,
+ "step": 2362
+ },
+ {
+ "epoch": 32.82445414847162,
+ "grad_norm": 0.01701410301029682,
+ "learning_rate": 0.0006,
+ "loss": 4.209896087646484,
+ "step": 2363
+ },
+ {
+ "epoch": 32.838427947598255,
+ "grad_norm": 0.01682264916598797,
+ "learning_rate": 0.0006,
+ "loss": 4.26223087310791,
+ "step": 2364
+ },
+ {
+ "epoch": 32.852401746724894,
+ "grad_norm": 0.015597456134855747,
+ "learning_rate": 0.0006,
+ "loss": 4.292664051055908,
+ "step": 2365
+ },
+ {
+ "epoch": 32.866375545851525,
+ "grad_norm": 0.015258579514920712,
+ "learning_rate": 0.0006,
+ "loss": 4.165006637573242,
+ "step": 2366
+ },
+ {
+ "epoch": 32.880349344978164,
+ "grad_norm": 0.01662200503051281,
+ "learning_rate": 0.0006,
+ "loss": 4.309786319732666,
+ "step": 2367
+ },
+ {
+ "epoch": 32.8943231441048,
+ "grad_norm": 0.015691161155700684,
+ "learning_rate": 0.0006,
+ "loss": 4.167202949523926,
+ "step": 2368
+ },
+ {
+ "epoch": 32.90829694323144,
+ "grad_norm": 0.014870637096464634,
+ "learning_rate": 0.0006,
+ "loss": 4.142415523529053,
+ "step": 2369
+ },
+ {
+ "epoch": 32.92227074235808,
+ "grad_norm": 0.01567252166569233,
+ "learning_rate": 0.0006,
+ "loss": 4.2312092781066895,
+ "step": 2370
+ },
+ {
+ "epoch": 32.93624454148472,
+ "grad_norm": 0.014948616735637188,
+ "learning_rate": 0.0006,
+ "loss": 4.1514129638671875,
+ "step": 2371
+ },
+ {
+ "epoch": 32.95021834061135,
+ "grad_norm": 0.014654500409960747,
+ "learning_rate": 0.0006,
+ "loss": 4.205010890960693,
+ "step": 2372
+ },
+ {
+ "epoch": 32.96419213973799,
+ "grad_norm": 0.015059148892760277,
+ "learning_rate": 0.0006,
+ "loss": 4.148405075073242,
+ "step": 2373
+ },
+ {
+ "epoch": 32.97816593886463,
+ "grad_norm": 0.015026576817035675,
+ "learning_rate": 0.0006,
+ "loss": 4.182535171508789,
+ "step": 2374
+ },
+ {
+ "epoch": 32.992139737991266,
+ "grad_norm": 0.016394320875406265,
+ "learning_rate": 0.0006,
+ "loss": 4.129051208496094,
+ "step": 2375
+ },
+ {
+ "epoch": 33.0,
+ "grad_norm": 0.01867043413221836,
+ "learning_rate": 0.0006,
+ "loss": 4.185556411743164,
+ "step": 2376
+ },
+ {
+ "epoch": 33.0,
+ "eval_loss": 4.640405654907227,
+ "eval_runtime": 56.5861,
+ "eval_samples_per_second": 43.155,
+ "eval_steps_per_second": 1.361,
+ "step": 2376
+ },
+ {
+ "epoch": 33.01397379912664,
+ "grad_norm": 0.018626727163791656,
+ "learning_rate": 0.0006,
+ "loss": 4.143031597137451,
+ "step": 2377
+ },
+ {
+ "epoch": 33.02794759825328,
+ "grad_norm": 0.019283363595604897,
+ "learning_rate": 0.0006,
+ "loss": 4.117845058441162,
+ "step": 2378
+ },
+ {
+ "epoch": 33.041921397379916,
+ "grad_norm": 0.018185822293162346,
+ "learning_rate": 0.0006,
+ "loss": 4.076862812042236,
+ "step": 2379
+ },
+ {
+ "epoch": 33.05589519650655,
+ "grad_norm": 0.019411196932196617,
+ "learning_rate": 0.0006,
+ "loss": 4.124156475067139,
+ "step": 2380
+ },
+ {
+ "epoch": 33.069868995633186,
+ "grad_norm": 0.019248466938734055,
+ "learning_rate": 0.0006,
+ "loss": 4.201735019683838,
+ "step": 2381
+ },
+ {
+ "epoch": 33.083842794759825,
+ "grad_norm": 0.018856720998883247,
+ "learning_rate": 0.0006,
+ "loss": 4.00296688079834,
+ "step": 2382
+ },
+ {
+ "epoch": 33.09781659388646,
+ "grad_norm": 0.018554294481873512,
+ "learning_rate": 0.0006,
+ "loss": 4.270577430725098,
+ "step": 2383
+ },
+ {
+ "epoch": 33.1117903930131,
+ "grad_norm": 0.019946224987506866,
+ "learning_rate": 0.0006,
+ "loss": 4.119589328765869,
+ "step": 2384
+ },
+ {
+ "epoch": 33.12576419213974,
+ "grad_norm": 0.021395670250058174,
+ "learning_rate": 0.0006,
+ "loss": 4.128847122192383,
+ "step": 2385
+ },
+ {
+ "epoch": 33.13973799126637,
+ "grad_norm": 0.02009674906730652,
+ "learning_rate": 0.0006,
+ "loss": 4.183228969573975,
+ "step": 2386
+ },
+ {
+ "epoch": 33.15371179039301,
+ "grad_norm": 0.01820223033428192,
+ "learning_rate": 0.0006,
+ "loss": 4.131601810455322,
+ "step": 2387
+ },
+ {
+ "epoch": 33.16768558951965,
+ "grad_norm": 0.018653156235814095,
+ "learning_rate": 0.0006,
+ "loss": 4.116430282592773,
+ "step": 2388
+ },
+ {
+ "epoch": 33.18165938864629,
+ "grad_norm": 0.016828063875436783,
+ "learning_rate": 0.0006,
+ "loss": 4.054663181304932,
+ "step": 2389
+ },
+ {
+ "epoch": 33.19563318777293,
+ "grad_norm": 0.017025409266352654,
+ "learning_rate": 0.0006,
+ "loss": 4.110044956207275,
+ "step": 2390
+ },
+ {
+ "epoch": 33.209606986899566,
+ "grad_norm": 0.016608452424407005,
+ "learning_rate": 0.0006,
+ "loss": 4.102668285369873,
+ "step": 2391
+ },
+ {
+ "epoch": 33.223580786026204,
+ "grad_norm": 0.01743577979505062,
+ "learning_rate": 0.0006,
+ "loss": 4.166791915893555,
+ "step": 2392
+ },
+ {
+ "epoch": 33.237554585152836,
+ "grad_norm": 0.017253734171390533,
+ "learning_rate": 0.0006,
+ "loss": 4.217832088470459,
+ "step": 2393
+ },
+ {
+ "epoch": 33.251528384279474,
+ "grad_norm": 0.017937595024704933,
+ "learning_rate": 0.0006,
+ "loss": 4.2173967361450195,
+ "step": 2394
+ },
+ {
+ "epoch": 33.26550218340611,
+ "grad_norm": 0.01985993981361389,
+ "learning_rate": 0.0006,
+ "loss": 4.17849588394165,
+ "step": 2395
+ },
+ {
+ "epoch": 33.27947598253275,
+ "grad_norm": 0.019243355840444565,
+ "learning_rate": 0.0006,
+ "loss": 4.086523056030273,
+ "step": 2396
+ },
+ {
+ "epoch": 33.29344978165939,
+ "grad_norm": 0.01924416422843933,
+ "learning_rate": 0.0006,
+ "loss": 4.128574371337891,
+ "step": 2397
+ },
+ {
+ "epoch": 33.30742358078603,
+ "grad_norm": 0.01983758620917797,
+ "learning_rate": 0.0006,
+ "loss": 4.061581611633301,
+ "step": 2398
+ },
+ {
+ "epoch": 33.32139737991266,
+ "grad_norm": 0.01930462196469307,
+ "learning_rate": 0.0006,
+ "loss": 4.091731071472168,
+ "step": 2399
+ },
+ {
+ "epoch": 33.3353711790393,
+ "grad_norm": 0.017988424748182297,
+ "learning_rate": 0.0006,
+ "loss": 4.155599117279053,
+ "step": 2400
+ },
+ {
+ "epoch": 33.34934497816594,
+ "grad_norm": 0.015194767154753208,
+ "learning_rate": 0.0006,
+ "loss": 4.144134044647217,
+ "step": 2401
+ },
+ {
+ "epoch": 33.36331877729258,
+ "grad_norm": 0.01536081824451685,
+ "learning_rate": 0.0006,
+ "loss": 4.127272129058838,
+ "step": 2402
+ },
+ {
+ "epoch": 33.377292576419215,
+ "grad_norm": 0.015431297942996025,
+ "learning_rate": 0.0006,
+ "loss": 4.3213019371032715,
+ "step": 2403
+ },
+ {
+ "epoch": 33.391266375545854,
+ "grad_norm": 0.015740280970931053,
+ "learning_rate": 0.0006,
+ "loss": 4.121467590332031,
+ "step": 2404
+ },
+ {
+ "epoch": 33.40524017467249,
+ "grad_norm": 0.015379384160041809,
+ "learning_rate": 0.0006,
+ "loss": 4.095216751098633,
+ "step": 2405
+ },
+ {
+ "epoch": 33.419213973799124,
+ "grad_norm": 0.01528073474764824,
+ "learning_rate": 0.0006,
+ "loss": 4.13047981262207,
+ "step": 2406
+ },
+ {
+ "epoch": 33.43318777292576,
+ "grad_norm": 0.015739979222416878,
+ "learning_rate": 0.0006,
+ "loss": 4.2533135414123535,
+ "step": 2407
+ },
+ {
+ "epoch": 33.4471615720524,
+ "grad_norm": 0.01519764307886362,
+ "learning_rate": 0.0006,
+ "loss": 4.131782054901123,
+ "step": 2408
+ },
+ {
+ "epoch": 33.46113537117904,
+ "grad_norm": 0.015425996854901314,
+ "learning_rate": 0.0006,
+ "loss": 4.02811336517334,
+ "step": 2409
+ },
+ {
+ "epoch": 33.47510917030568,
+ "grad_norm": 0.015378830023109913,
+ "learning_rate": 0.0006,
+ "loss": 4.08652925491333,
+ "step": 2410
+ },
+ {
+ "epoch": 33.48908296943232,
+ "grad_norm": 0.01473075058311224,
+ "learning_rate": 0.0006,
+ "loss": 4.214364528656006,
+ "step": 2411
+ },
+ {
+ "epoch": 33.50305676855895,
+ "grad_norm": 0.015141637064516544,
+ "learning_rate": 0.0006,
+ "loss": 4.155712127685547,
+ "step": 2412
+ },
+ {
+ "epoch": 33.51703056768559,
+ "grad_norm": 0.01641593500971794,
+ "learning_rate": 0.0006,
+ "loss": 4.097907066345215,
+ "step": 2413
+ },
+ {
+ "epoch": 33.531004366812226,
+ "grad_norm": 0.015182029455900192,
+ "learning_rate": 0.0006,
+ "loss": 4.238499641418457,
+ "step": 2414
+ },
+ {
+ "epoch": 33.544978165938865,
+ "grad_norm": 0.01643732376396656,
+ "learning_rate": 0.0006,
+ "loss": 4.0661492347717285,
+ "step": 2415
+ },
+ {
+ "epoch": 33.5589519650655,
+ "grad_norm": 0.015614228323101997,
+ "learning_rate": 0.0006,
+ "loss": 4.120795249938965,
+ "step": 2416
+ },
+ {
+ "epoch": 33.57292576419214,
+ "grad_norm": 0.016066187992691994,
+ "learning_rate": 0.0006,
+ "loss": 4.096465110778809,
+ "step": 2417
+ },
+ {
+ "epoch": 33.58689956331878,
+ "grad_norm": 0.017214342951774597,
+ "learning_rate": 0.0006,
+ "loss": 4.242628574371338,
+ "step": 2418
+ },
+ {
+ "epoch": 33.60087336244541,
+ "grad_norm": 0.01787586510181427,
+ "learning_rate": 0.0006,
+ "loss": 4.149589538574219,
+ "step": 2419
+ },
+ {
+ "epoch": 33.61484716157205,
+ "grad_norm": 0.019455241039395332,
+ "learning_rate": 0.0006,
+ "loss": 4.267796039581299,
+ "step": 2420
+ },
+ {
+ "epoch": 33.62882096069869,
+ "grad_norm": 0.01984674669802189,
+ "learning_rate": 0.0006,
+ "loss": 4.175220966339111,
+ "step": 2421
+ },
+ {
+ "epoch": 33.64279475982533,
+ "grad_norm": 0.021933717653155327,
+ "learning_rate": 0.0006,
+ "loss": 4.106647491455078,
+ "step": 2422
+ },
+ {
+ "epoch": 33.65676855895197,
+ "grad_norm": 0.02123589813709259,
+ "learning_rate": 0.0006,
+ "loss": 4.204996109008789,
+ "step": 2423
+ },
+ {
+ "epoch": 33.670742358078606,
+ "grad_norm": 0.019814589992165565,
+ "learning_rate": 0.0006,
+ "loss": 4.186171531677246,
+ "step": 2424
+ },
+ {
+ "epoch": 33.68471615720524,
+ "grad_norm": 0.02059728465974331,
+ "learning_rate": 0.0006,
+ "loss": 4.170509338378906,
+ "step": 2425
+ },
+ {
+ "epoch": 33.698689956331876,
+ "grad_norm": 0.018483608961105347,
+ "learning_rate": 0.0006,
+ "loss": 4.085585594177246,
+ "step": 2426
+ },
+ {
+ "epoch": 33.712663755458514,
+ "grad_norm": 0.018403852358460426,
+ "learning_rate": 0.0006,
+ "loss": 4.2551774978637695,
+ "step": 2427
+ },
+ {
+ "epoch": 33.72663755458515,
+ "grad_norm": 0.018653830513358116,
+ "learning_rate": 0.0006,
+ "loss": 4.091766357421875,
+ "step": 2428
+ },
+ {
+ "epoch": 33.74061135371179,
+ "grad_norm": 0.017692334949970245,
+ "learning_rate": 0.0006,
+ "loss": 4.100838661193848,
+ "step": 2429
+ },
+ {
+ "epoch": 33.75458515283843,
+ "grad_norm": 0.016029126942157745,
+ "learning_rate": 0.0006,
+ "loss": 4.1530680656433105,
+ "step": 2430
+ },
+ {
+ "epoch": 33.76855895196506,
+ "grad_norm": 0.016095684841275215,
+ "learning_rate": 0.0006,
+ "loss": 4.20544958114624,
+ "step": 2431
+ },
+ {
+ "epoch": 33.7825327510917,
+ "grad_norm": 0.016375798732042313,
+ "learning_rate": 0.0006,
+ "loss": 4.199177265167236,
+ "step": 2432
+ },
+ {
+ "epoch": 33.79650655021834,
+ "grad_norm": 0.01590786688029766,
+ "learning_rate": 0.0006,
+ "loss": 4.200799465179443,
+ "step": 2433
+ },
+ {
+ "epoch": 33.81048034934498,
+ "grad_norm": 0.014896097593009472,
+ "learning_rate": 0.0006,
+ "loss": 4.185850143432617,
+ "step": 2434
+ },
+ {
+ "epoch": 33.82445414847162,
+ "grad_norm": 0.01438513956964016,
+ "learning_rate": 0.0006,
+ "loss": 4.214962005615234,
+ "step": 2435
+ },
+ {
+ "epoch": 33.838427947598255,
+ "grad_norm": 0.014459339901804924,
+ "learning_rate": 0.0006,
+ "loss": 4.196972370147705,
+ "step": 2436
+ },
+ {
+ "epoch": 33.852401746724894,
+ "grad_norm": 0.014782169833779335,
+ "learning_rate": 0.0006,
+ "loss": 4.0885910987854,
+ "step": 2437
+ },
+ {
+ "epoch": 33.866375545851525,
+ "grad_norm": 0.01520370040088892,
+ "learning_rate": 0.0006,
+ "loss": 4.0323262214660645,
+ "step": 2438
+ },
+ {
+ "epoch": 33.880349344978164,
+ "grad_norm": 0.014056282117962837,
+ "learning_rate": 0.0006,
+ "loss": 4.1386518478393555,
+ "step": 2439
+ },
+ {
+ "epoch": 33.8943231441048,
+ "grad_norm": 0.013567990623414516,
+ "learning_rate": 0.0006,
+ "loss": 4.095754623413086,
+ "step": 2440
+ },
+ {
+ "epoch": 33.90829694323144,
+ "grad_norm": 0.013306716457009315,
+ "learning_rate": 0.0006,
+ "loss": 4.1623077392578125,
+ "step": 2441
+ },
+ {
+ "epoch": 33.92227074235808,
+ "grad_norm": 0.014047150500118732,
+ "learning_rate": 0.0006,
+ "loss": 4.142541885375977,
+ "step": 2442
+ },
+ {
+ "epoch": 33.93624454148472,
+ "grad_norm": 0.01377617847174406,
+ "learning_rate": 0.0006,
+ "loss": 4.268551826477051,
+ "step": 2443
+ },
+ {
+ "epoch": 33.95021834061135,
+ "grad_norm": 0.014671621844172478,
+ "learning_rate": 0.0006,
+ "loss": 4.198196887969971,
+ "step": 2444
+ },
+ {
+ "epoch": 33.96419213973799,
+ "grad_norm": 0.014806744642555714,
+ "learning_rate": 0.0006,
+ "loss": 4.057899475097656,
+ "step": 2445
+ },
+ {
+ "epoch": 33.97816593886463,
+ "grad_norm": 0.015734272077679634,
+ "learning_rate": 0.0006,
+ "loss": 4.163505554199219,
+ "step": 2446
+ },
+ {
+ "epoch": 33.992139737991266,
+ "grad_norm": 0.015250146389007568,
+ "learning_rate": 0.0006,
+ "loss": 4.076294898986816,
+ "step": 2447
+ },
+ {
+ "epoch": 34.0,
+ "grad_norm": 0.016733702272176743,
+ "learning_rate": 0.0006,
+ "loss": 4.138420104980469,
+ "step": 2448
+ },
+ {
+ "epoch": 34.0,
+ "eval_loss": 4.542503356933594,
+ "eval_runtime": 56.983,
+ "eval_samples_per_second": 42.855,
+ "eval_steps_per_second": 1.351,
+ "step": 2448
+ },
+ {
+ "epoch": 34.01397379912664,
+ "grad_norm": 0.015866613015532494,
+ "learning_rate": 0.0006,
+ "loss": 4.032081127166748,
+ "step": 2449
+ },
+ {
+ "epoch": 34.02794759825328,
+ "grad_norm": 0.01946570910513401,
+ "learning_rate": 0.0006,
+ "loss": 4.122320175170898,
+ "step": 2450
+ },
+ {
+ "epoch": 34.041921397379916,
+ "grad_norm": 0.020265938714146614,
+ "learning_rate": 0.0006,
+ "loss": 4.06460428237915,
+ "step": 2451
+ },
+ {
+ "epoch": 34.05589519650655,
+ "grad_norm": 0.018887171521782875,
+ "learning_rate": 0.0006,
+ "loss": 4.096906661987305,
+ "step": 2452
+ },
+ {
+ "epoch": 34.069868995633186,
+ "grad_norm": 0.017304692417383194,
+ "learning_rate": 0.0006,
+ "loss": 4.137261390686035,
+ "step": 2453
+ },
+ {
+ "epoch": 34.083842794759825,
+ "grad_norm": 0.017029082402586937,
+ "learning_rate": 0.0006,
+ "loss": 4.081112861633301,
+ "step": 2454
+ },
+ {
+ "epoch": 34.09781659388646,
+ "grad_norm": 0.016414813697338104,
+ "learning_rate": 0.0006,
+ "loss": 4.140783309936523,
+ "step": 2455
+ },
+ {
+ "epoch": 34.1117903930131,
+ "grad_norm": 0.018238281831145287,
+ "learning_rate": 0.0006,
+ "loss": 4.161116600036621,
+ "step": 2456
+ },
+ {
+ "epoch": 34.12576419213974,
+ "grad_norm": 0.019929232075810432,
+ "learning_rate": 0.0006,
+ "loss": 4.243722438812256,
+ "step": 2457
+ },
+ {
+ "epoch": 34.13973799126637,
+ "grad_norm": 0.019375786185264587,
+ "learning_rate": 0.0006,
+ "loss": 4.134873390197754,
+ "step": 2458
+ },
+ {
+ "epoch": 34.15371179039301,
+ "grad_norm": 0.018151137977838516,
+ "learning_rate": 0.0006,
+ "loss": 4.098745346069336,
+ "step": 2459
+ },
+ {
+ "epoch": 34.16768558951965,
+ "grad_norm": 0.017054934054613113,
+ "learning_rate": 0.0006,
+ "loss": 4.194662570953369,
+ "step": 2460
+ },
+ {
+ "epoch": 34.18165938864629,
+ "grad_norm": 0.018220555037260056,
+ "learning_rate": 0.0006,
+ "loss": 4.128866195678711,
+ "step": 2461
+ },
+ {
+ "epoch": 34.19563318777293,
+ "grad_norm": 0.01692879945039749,
+ "learning_rate": 0.0006,
+ "loss": 4.062477111816406,
+ "step": 2462
+ },
+ {
+ "epoch": 34.209606986899566,
+ "grad_norm": 0.01674199476838112,
+ "learning_rate": 0.0006,
+ "loss": 4.186149597167969,
+ "step": 2463
+ },
+ {
+ "epoch": 34.223580786026204,
+ "grad_norm": 0.016527455300092697,
+ "learning_rate": 0.0006,
+ "loss": 4.0734028816223145,
+ "step": 2464
+ },
+ {
+ "epoch": 34.237554585152836,
+ "grad_norm": 0.016448386013507843,
+ "learning_rate": 0.0006,
+ "loss": 4.098896026611328,
+ "step": 2465
+ },
+ {
+ "epoch": 34.251528384279474,
+ "grad_norm": 0.014644953422248363,
+ "learning_rate": 0.0006,
+ "loss": 4.1031904220581055,
+ "step": 2466
+ },
+ {
+ "epoch": 34.26550218340611,
+ "grad_norm": 0.016649341210722923,
+ "learning_rate": 0.0006,
+ "loss": 4.148791313171387,
+ "step": 2467
+ },
+ {
+ "epoch": 34.27947598253275,
+ "grad_norm": 0.017442714422941208,
+ "learning_rate": 0.0006,
+ "loss": 4.070799827575684,
+ "step": 2468
+ },
+ {
+ "epoch": 34.29344978165939,
+ "grad_norm": 0.017852013930678368,
+ "learning_rate": 0.0006,
+ "loss": 4.031924247741699,
+ "step": 2469
+ },
+ {
+ "epoch": 34.30742358078603,
+ "grad_norm": 0.016371270641684532,
+ "learning_rate": 0.0006,
+ "loss": 4.149420738220215,
+ "step": 2470
+ },
+ {
+ "epoch": 34.32139737991266,
+ "grad_norm": 0.015846814960241318,
+ "learning_rate": 0.0006,
+ "loss": 4.15427303314209,
+ "step": 2471
+ },
+ {
+ "epoch": 34.3353711790393,
+ "grad_norm": 0.017411163076758385,
+ "learning_rate": 0.0006,
+ "loss": 4.1304192543029785,
+ "step": 2472
+ },
+ {
+ "epoch": 34.34934497816594,
+ "grad_norm": 0.01780046708881855,
+ "learning_rate": 0.0006,
+ "loss": 4.191732406616211,
+ "step": 2473
+ },
+ {
+ "epoch": 34.36331877729258,
+ "grad_norm": 0.01958383060991764,
+ "learning_rate": 0.0006,
+ "loss": 4.068355083465576,
+ "step": 2474
+ },
+ {
+ "epoch": 34.377292576419215,
+ "grad_norm": 0.022173317149281502,
+ "learning_rate": 0.0006,
+ "loss": 4.182929039001465,
+ "step": 2475
+ },
+ {
+ "epoch": 34.391266375545854,
+ "grad_norm": 0.022389156743884087,
+ "learning_rate": 0.0006,
+ "loss": 4.110218048095703,
+ "step": 2476
+ },
+ {
+ "epoch": 34.40524017467249,
+ "grad_norm": 0.022478511556982994,
+ "learning_rate": 0.0006,
+ "loss": 4.095479488372803,
+ "step": 2477
+ },
+ {
+ "epoch": 34.419213973799124,
+ "grad_norm": 0.019504647701978683,
+ "learning_rate": 0.0006,
+ "loss": 4.101964950561523,
+ "step": 2478
+ },
+ {
+ "epoch": 34.43318777292576,
+ "grad_norm": 0.019378576427698135,
+ "learning_rate": 0.0006,
+ "loss": 4.14620304107666,
+ "step": 2479
+ },
+ {
+ "epoch": 34.4471615720524,
+ "grad_norm": 0.017993204295635223,
+ "learning_rate": 0.0006,
+ "loss": 4.183492660522461,
+ "step": 2480
+ },
+ {
+ "epoch": 34.46113537117904,
+ "grad_norm": 0.01814805157482624,
+ "learning_rate": 0.0006,
+ "loss": 4.151247978210449,
+ "step": 2481
+ },
+ {
+ "epoch": 34.47510917030568,
+ "grad_norm": 0.018622448667883873,
+ "learning_rate": 0.0006,
+ "loss": 4.020938396453857,
+ "step": 2482
+ },
+ {
+ "epoch": 34.48908296943232,
+ "grad_norm": 0.018360132351517677,
+ "learning_rate": 0.0006,
+ "loss": 4.064734935760498,
+ "step": 2483
+ },
+ {
+ "epoch": 34.50305676855895,
+ "grad_norm": 0.017715711146593094,
+ "learning_rate": 0.0006,
+ "loss": 4.185687065124512,
+ "step": 2484
+ },
+ {
+ "epoch": 34.51703056768559,
+ "grad_norm": 0.01714295893907547,
+ "learning_rate": 0.0006,
+ "loss": 4.1044721603393555,
+ "step": 2485
+ },
+ {
+ "epoch": 34.531004366812226,
+ "grad_norm": 0.016943803057074547,
+ "learning_rate": 0.0006,
+ "loss": 4.106837272644043,
+ "step": 2486
+ },
+ {
+ "epoch": 34.544978165938865,
+ "grad_norm": 0.0182423684746027,
+ "learning_rate": 0.0006,
+ "loss": 4.1041340827941895,
+ "step": 2487
+ },
+ {
+ "epoch": 34.5589519650655,
+ "grad_norm": 0.01636587455868721,
+ "learning_rate": 0.0006,
+ "loss": 3.9735546112060547,
+ "step": 2488
+ },
+ {
+ "epoch": 34.57292576419214,
+ "grad_norm": 0.014593546278774738,
+ "learning_rate": 0.0006,
+ "loss": 4.041980266571045,
+ "step": 2489
+ },
+ {
+ "epoch": 34.58689956331878,
+ "grad_norm": 0.015436965972185135,
+ "learning_rate": 0.0006,
+ "loss": 4.301148414611816,
+ "step": 2490
+ },
+ {
+ "epoch": 34.60087336244541,
+ "grad_norm": 0.015075269155204296,
+ "learning_rate": 0.0006,
+ "loss": 4.118527412414551,
+ "step": 2491
+ },
+ {
+ "epoch": 34.61484716157205,
+ "grad_norm": 0.01701514609158039,
+ "learning_rate": 0.0006,
+ "loss": 4.245968818664551,
+ "step": 2492
+ },
+ {
+ "epoch": 34.62882096069869,
+ "grad_norm": 0.018047887831926346,
+ "learning_rate": 0.0006,
+ "loss": 4.227261066436768,
+ "step": 2493
+ },
+ {
+ "epoch": 34.64279475982533,
+ "grad_norm": 0.018499085679650307,
+ "learning_rate": 0.0006,
+ "loss": 4.121360778808594,
+ "step": 2494
+ },
+ {
+ "epoch": 34.65676855895197,
+ "grad_norm": 0.018987711519002914,
+ "learning_rate": 0.0006,
+ "loss": 4.131564617156982,
+ "step": 2495
+ },
+ {
+ "epoch": 34.670742358078606,
+ "grad_norm": 0.01834476925432682,
+ "learning_rate": 0.0006,
+ "loss": 4.224941253662109,
+ "step": 2496
+ },
+ {
+ "epoch": 34.68471615720524,
+ "grad_norm": 0.017507942393422127,
+ "learning_rate": 0.0006,
+ "loss": 4.1394548416137695,
+ "step": 2497
+ },
+ {
+ "epoch": 34.698689956331876,
+ "grad_norm": 0.01680580899119377,
+ "learning_rate": 0.0006,
+ "loss": 4.1073431968688965,
+ "step": 2498
+ },
+ {
+ "epoch": 34.712663755458514,
+ "grad_norm": 0.015233514830470085,
+ "learning_rate": 0.0006,
+ "loss": 4.136994361877441,
+ "step": 2499
+ },
+ {
+ "epoch": 34.72663755458515,
+ "grad_norm": 0.015482917428016663,
+ "learning_rate": 0.0006,
+ "loss": 4.110719680786133,
+ "step": 2500
+ },
+ {
+ "epoch": 34.74061135371179,
+ "grad_norm": 0.015940461307764053,
+ "learning_rate": 0.0006,
+ "loss": 4.207851886749268,
+ "step": 2501
+ },
+ {
+ "epoch": 34.75458515283843,
+ "grad_norm": 0.015355597250163555,
+ "learning_rate": 0.0006,
+ "loss": 4.169469833374023,
+ "step": 2502
+ },
+ {
+ "epoch": 34.76855895196506,
+ "grad_norm": 0.015974000096321106,
+ "learning_rate": 0.0006,
+ "loss": 4.110061168670654,
+ "step": 2503
+ },
+ {
+ "epoch": 34.7825327510917,
+ "grad_norm": 0.016088927164673805,
+ "learning_rate": 0.0006,
+ "loss": 4.155452728271484,
+ "step": 2504
+ },
+ {
+ "epoch": 34.79650655021834,
+ "grad_norm": 0.016156120225787163,
+ "learning_rate": 0.0006,
+ "loss": 4.178092956542969,
+ "step": 2505
+ },
+ {
+ "epoch": 34.81048034934498,
+ "grad_norm": 0.014928647316992283,
+ "learning_rate": 0.0006,
+ "loss": 4.05457878112793,
+ "step": 2506
+ },
+ {
+ "epoch": 34.82445414847162,
+ "grad_norm": 0.015289328061044216,
+ "learning_rate": 0.0006,
+ "loss": 4.18869686126709,
+ "step": 2507
+ },
+ {
+ "epoch": 34.838427947598255,
+ "grad_norm": 0.015799039974808693,
+ "learning_rate": 0.0006,
+ "loss": 4.106073379516602,
+ "step": 2508
+ },
+ {
+ "epoch": 34.852401746724894,
+ "grad_norm": 0.015824418514966965,
+ "learning_rate": 0.0006,
+ "loss": 4.0634589195251465,
+ "step": 2509
+ },
+ {
+ "epoch": 34.866375545851525,
+ "grad_norm": 0.015321656130254269,
+ "learning_rate": 0.0006,
+ "loss": 4.093090057373047,
+ "step": 2510
+ },
+ {
+ "epoch": 34.880349344978164,
+ "grad_norm": 0.014380166307091713,
+ "learning_rate": 0.0006,
+ "loss": 4.1750640869140625,
+ "step": 2511
+ },
+ {
+ "epoch": 34.8943231441048,
+ "grad_norm": 0.014962567947804928,
+ "learning_rate": 0.0006,
+ "loss": 4.185277938842773,
+ "step": 2512
+ },
+ {
+ "epoch": 34.90829694323144,
+ "grad_norm": 0.014467543922364712,
+ "learning_rate": 0.0006,
+ "loss": 4.1600799560546875,
+ "step": 2513
+ },
+ {
+ "epoch": 34.92227074235808,
+ "grad_norm": 0.01545181218534708,
+ "learning_rate": 0.0006,
+ "loss": 4.096299648284912,
+ "step": 2514
+ },
+ {
+ "epoch": 34.93624454148472,
+ "grad_norm": 0.01605340465903282,
+ "learning_rate": 0.0006,
+ "loss": 4.203323841094971,
+ "step": 2515
+ },
+ {
+ "epoch": 34.95021834061135,
+ "grad_norm": 0.015203815884888172,
+ "learning_rate": 0.0006,
+ "loss": 4.201385021209717,
+ "step": 2516
+ },
+ {
+ "epoch": 34.96419213973799,
+ "grad_norm": 0.015149409882724285,
+ "learning_rate": 0.0006,
+ "loss": 4.088164329528809,
+ "step": 2517
+ },
+ {
+ "epoch": 34.97816593886463,
+ "grad_norm": 0.014885558746755123,
+ "learning_rate": 0.0006,
+ "loss": 4.073911666870117,
+ "step": 2518
+ },
+ {
+ "epoch": 34.992139737991266,
+ "grad_norm": 0.015003419481217861,
+ "learning_rate": 0.0006,
+ "loss": 4.157113552093506,
+ "step": 2519
+ },
+ {
+ "epoch": 35.0,
+ "grad_norm": 0.01641916297376156,
+ "learning_rate": 0.0006,
+ "loss": 3.970052719116211,
+ "step": 2520
+ },
+ {
+ "epoch": 35.0,
+ "eval_loss": 4.641692161560059,
+ "eval_runtime": 56.8836,
+ "eval_samples_per_second": 42.93,
+ "eval_steps_per_second": 1.354,
+ "step": 2520
+ },
+ {
+ "epoch": 35.01397379912664,
+ "grad_norm": 0.015894673764705658,
+ "learning_rate": 0.0006,
+ "loss": 4.104437351226807,
+ "step": 2521
+ },
+ {
+ "epoch": 35.02794759825328,
+ "grad_norm": 0.0173867866396904,
+ "learning_rate": 0.0006,
+ "loss": 4.091726303100586,
+ "step": 2522
+ },
+ {
+ "epoch": 35.041921397379916,
+ "grad_norm": 0.019272832199931145,
+ "learning_rate": 0.0006,
+ "loss": 4.217905044555664,
+ "step": 2523
+ },
+ {
+ "epoch": 35.05589519650655,
+ "grad_norm": 0.021820371970534325,
+ "learning_rate": 0.0006,
+ "loss": 4.111428260803223,
+ "step": 2524
+ },
+ {
+ "epoch": 35.069868995633186,
+ "grad_norm": 0.02340741828083992,
+ "learning_rate": 0.0006,
+ "loss": 4.115771293640137,
+ "step": 2525
+ },
+ {
+ "epoch": 35.083842794759825,
+ "grad_norm": 0.023760691285133362,
+ "learning_rate": 0.0006,
+ "loss": 4.132383346557617,
+ "step": 2526
+ },
+ {
+ "epoch": 35.09781659388646,
+ "grad_norm": 0.02073468267917633,
+ "learning_rate": 0.0006,
+ "loss": 4.272691249847412,
+ "step": 2527
+ },
+ {
+ "epoch": 35.1117903930131,
+ "grad_norm": 0.02197718247771263,
+ "learning_rate": 0.0006,
+ "loss": 4.105586528778076,
+ "step": 2528
+ },
+ {
+ "epoch": 35.12576419213974,
+ "grad_norm": 0.0200254675000906,
+ "learning_rate": 0.0006,
+ "loss": 3.968820095062256,
+ "step": 2529
+ },
+ {
+ "epoch": 35.13973799126637,
+ "grad_norm": 0.01929587684571743,
+ "learning_rate": 0.0006,
+ "loss": 4.045468807220459,
+ "step": 2530
+ },
+ {
+ "epoch": 35.15371179039301,
+ "grad_norm": 0.01886162906885147,
+ "learning_rate": 0.0006,
+ "loss": 4.185571670532227,
+ "step": 2531
+ },
+ {
+ "epoch": 35.16768558951965,
+ "grad_norm": 0.019789839163422585,
+ "learning_rate": 0.0006,
+ "loss": 3.9734551906585693,
+ "step": 2532
+ },
+ {
+ "epoch": 35.18165938864629,
+ "grad_norm": 0.0198142658919096,
+ "learning_rate": 0.0006,
+ "loss": 4.2662506103515625,
+ "step": 2533
+ },
+ {
+ "epoch": 35.19563318777293,
+ "grad_norm": 0.02087012305855751,
+ "learning_rate": 0.0006,
+ "loss": 4.175931930541992,
+ "step": 2534
+ },
+ {
+ "epoch": 35.209606986899566,
+ "grad_norm": 0.020294535905122757,
+ "learning_rate": 0.0006,
+ "loss": 4.060324668884277,
+ "step": 2535
+ },
+ {
+ "epoch": 35.223580786026204,
+ "grad_norm": 0.01908983662724495,
+ "learning_rate": 0.0006,
+ "loss": 4.217939853668213,
+ "step": 2536
+ },
+ {
+ "epoch": 35.237554585152836,
+ "grad_norm": 0.0178892333060503,
+ "learning_rate": 0.0006,
+ "loss": 4.190216541290283,
+ "step": 2537
+ },
+ {
+ "epoch": 35.251528384279474,
+ "grad_norm": 0.017892315983772278,
+ "learning_rate": 0.0006,
+ "loss": 4.067419052124023,
+ "step": 2538
+ },
+ {
+ "epoch": 35.26550218340611,
+ "grad_norm": 0.018556250259280205,
+ "learning_rate": 0.0006,
+ "loss": 4.105958461761475,
+ "step": 2539
+ },
+ {
+ "epoch": 35.27947598253275,
+ "grad_norm": 0.01944422349333763,
+ "learning_rate": 0.0006,
+ "loss": 4.155775547027588,
+ "step": 2540
+ },
+ {
+ "epoch": 35.29344978165939,
+ "grad_norm": 0.017333146184682846,
+ "learning_rate": 0.0006,
+ "loss": 4.244513511657715,
+ "step": 2541
+ },
+ {
+ "epoch": 35.30742358078603,
+ "grad_norm": 0.01639244332909584,
+ "learning_rate": 0.0006,
+ "loss": 4.167815685272217,
+ "step": 2542
+ },
+ {
+ "epoch": 35.32139737991266,
+ "grad_norm": 0.01610143668949604,
+ "learning_rate": 0.0006,
+ "loss": 4.121486663818359,
+ "step": 2543
+ },
+ {
+ "epoch": 35.3353711790393,
+ "grad_norm": 0.01656411588191986,
+ "learning_rate": 0.0006,
+ "loss": 4.1755499839782715,
+ "step": 2544
+ },
+ {
+ "epoch": 35.34934497816594,
+ "grad_norm": 0.015075297094881535,
+ "learning_rate": 0.0006,
+ "loss": 4.102992534637451,
+ "step": 2545
+ },
+ {
+ "epoch": 35.36331877729258,
+ "grad_norm": 0.01538170501589775,
+ "learning_rate": 0.0006,
+ "loss": 4.161603927612305,
+ "step": 2546
+ },
+ {
+ "epoch": 35.377292576419215,
+ "grad_norm": 0.01483312901109457,
+ "learning_rate": 0.0006,
+ "loss": 4.085355758666992,
+ "step": 2547
+ },
+ {
+ "epoch": 35.391266375545854,
+ "grad_norm": 0.014894738793373108,
+ "learning_rate": 0.0006,
+ "loss": 4.083585262298584,
+ "step": 2548
+ },
+ {
+ "epoch": 35.40524017467249,
+ "grad_norm": 0.01475780550390482,
+ "learning_rate": 0.0006,
+ "loss": 4.012645721435547,
+ "step": 2549
+ },
+ {
+ "epoch": 35.419213973799124,
+ "grad_norm": 0.014531653374433517,
+ "learning_rate": 0.0006,
+ "loss": 4.197113990783691,
+ "step": 2550
+ },
+ {
+ "epoch": 35.43318777292576,
+ "grad_norm": 0.015316493809223175,
+ "learning_rate": 0.0006,
+ "loss": 4.227970123291016,
+ "step": 2551
+ },
+ {
+ "epoch": 35.4471615720524,
+ "grad_norm": 0.014685571193695068,
+ "learning_rate": 0.0006,
+ "loss": 4.089415550231934,
+ "step": 2552
+ },
+ {
+ "epoch": 35.46113537117904,
+ "grad_norm": 0.015337035991251469,
+ "learning_rate": 0.0006,
+ "loss": 4.080583572387695,
+ "step": 2553
+ },
+ {
+ "epoch": 35.47510917030568,
+ "grad_norm": 0.016754208132624626,
+ "learning_rate": 0.0006,
+ "loss": 4.163935661315918,
+ "step": 2554
+ },
+ {
+ "epoch": 35.48908296943232,
+ "grad_norm": 0.017535798251628876,
+ "learning_rate": 0.0006,
+ "loss": 4.093027591705322,
+ "step": 2555
+ },
+ {
+ "epoch": 35.50305676855895,
+ "grad_norm": 0.017336975783109665,
+ "learning_rate": 0.0006,
+ "loss": 4.127961158752441,
+ "step": 2556
+ },
+ {
+ "epoch": 35.51703056768559,
+ "grad_norm": 0.016242152079939842,
+ "learning_rate": 0.0006,
+ "loss": 4.155593395233154,
+ "step": 2557
+ },
+ {
+ "epoch": 35.531004366812226,
+ "grad_norm": 0.015121322125196457,
+ "learning_rate": 0.0006,
+ "loss": 3.984004020690918,
+ "step": 2558
+ },
+ {
+ "epoch": 35.544978165938865,
+ "grad_norm": 0.015178884379565716,
+ "learning_rate": 0.0006,
+ "loss": 4.111564636230469,
+ "step": 2559
+ },
+ {
+ "epoch": 35.5589519650655,
+ "grad_norm": 0.014777131378650665,
+ "learning_rate": 0.0006,
+ "loss": 4.041402816772461,
+ "step": 2560
+ },
+ {
+ "epoch": 35.57292576419214,
+ "grad_norm": 0.015197676606476307,
+ "learning_rate": 0.0006,
+ "loss": 4.133706092834473,
+ "step": 2561
+ },
+ {
+ "epoch": 35.58689956331878,
+ "grad_norm": 0.015905972570180893,
+ "learning_rate": 0.0006,
+ "loss": 4.1069135665893555,
+ "step": 2562
+ },
+ {
+ "epoch": 35.60087336244541,
+ "grad_norm": 0.015942877158522606,
+ "learning_rate": 0.0006,
+ "loss": 4.0720624923706055,
+ "step": 2563
+ },
+ {
+ "epoch": 35.61484716157205,
+ "grad_norm": 0.016697583720088005,
+ "learning_rate": 0.0006,
+ "loss": 4.167448043823242,
+ "step": 2564
+ },
+ {
+ "epoch": 35.62882096069869,
+ "grad_norm": 0.015603577718138695,
+ "learning_rate": 0.0006,
+ "loss": 4.127327919006348,
+ "step": 2565
+ },
+ {
+ "epoch": 35.64279475982533,
+ "grad_norm": 0.017153460532426834,
+ "learning_rate": 0.0006,
+ "loss": 4.056543350219727,
+ "step": 2566
+ },
+ {
+ "epoch": 35.65676855895197,
+ "grad_norm": 0.016731958836317062,
+ "learning_rate": 0.0006,
+ "loss": 4.0681843757629395,
+ "step": 2567
+ },
+ {
+ "epoch": 35.670742358078606,
+ "grad_norm": 0.016167359426617622,
+ "learning_rate": 0.0006,
+ "loss": 4.174992084503174,
+ "step": 2568
+ },
+ {
+ "epoch": 35.68471615720524,
+ "grad_norm": 0.01615108922123909,
+ "learning_rate": 0.0006,
+ "loss": 4.1383161544799805,
+ "step": 2569
+ },
+ {
+ "epoch": 35.698689956331876,
+ "grad_norm": 0.015750816091895103,
+ "learning_rate": 0.0006,
+ "loss": 4.031595706939697,
+ "step": 2570
+ },
+ {
+ "epoch": 35.712663755458514,
+ "grad_norm": 0.015416189096868038,
+ "learning_rate": 0.0006,
+ "loss": 4.067498207092285,
+ "step": 2571
+ },
+ {
+ "epoch": 35.72663755458515,
+ "grad_norm": 0.01570792682468891,
+ "learning_rate": 0.0006,
+ "loss": 4.1338300704956055,
+ "step": 2572
+ },
+ {
+ "epoch": 35.74061135371179,
+ "grad_norm": 0.016367752104997635,
+ "learning_rate": 0.0006,
+ "loss": 4.152523040771484,
+ "step": 2573
+ },
+ {
+ "epoch": 35.75458515283843,
+ "grad_norm": 0.016349803656339645,
+ "learning_rate": 0.0006,
+ "loss": 4.101985454559326,
+ "step": 2574
+ },
+ {
+ "epoch": 35.76855895196506,
+ "grad_norm": 0.01663660630583763,
+ "learning_rate": 0.0006,
+ "loss": 4.190130710601807,
+ "step": 2575
+ },
+ {
+ "epoch": 35.7825327510917,
+ "grad_norm": 0.015599170699715614,
+ "learning_rate": 0.0006,
+ "loss": 4.0792436599731445,
+ "step": 2576
+ },
+ {
+ "epoch": 35.79650655021834,
+ "grad_norm": 0.015472427010536194,
+ "learning_rate": 0.0006,
+ "loss": 4.186446666717529,
+ "step": 2577
+ },
+ {
+ "epoch": 35.81048034934498,
+ "grad_norm": 0.015520439483225346,
+ "learning_rate": 0.0006,
+ "loss": 4.246635437011719,
+ "step": 2578
+ },
+ {
+ "epoch": 35.82445414847162,
+ "grad_norm": 0.016998281702399254,
+ "learning_rate": 0.0006,
+ "loss": 4.13435173034668,
+ "step": 2579
+ },
+ {
+ "epoch": 35.838427947598255,
+ "grad_norm": 0.016290361061692238,
+ "learning_rate": 0.0006,
+ "loss": 4.107428073883057,
+ "step": 2580
+ },
+ {
+ "epoch": 35.852401746724894,
+ "grad_norm": 0.0165516659617424,
+ "learning_rate": 0.0006,
+ "loss": 4.1612138748168945,
+ "step": 2581
+ },
+ {
+ "epoch": 35.866375545851525,
+ "grad_norm": 0.01736091636121273,
+ "learning_rate": 0.0006,
+ "loss": 4.1072869300842285,
+ "step": 2582
+ },
+ {
+ "epoch": 35.880349344978164,
+ "grad_norm": 0.01787031814455986,
+ "learning_rate": 0.0006,
+ "loss": 4.1076812744140625,
+ "step": 2583
+ },
+ {
+ "epoch": 35.8943231441048,
+ "grad_norm": 0.015437985770404339,
+ "learning_rate": 0.0006,
+ "loss": 4.097488880157471,
+ "step": 2584
+ },
+ {
+ "epoch": 35.90829694323144,
+ "grad_norm": 0.016306741163134575,
+ "learning_rate": 0.0006,
+ "loss": 4.105017185211182,
+ "step": 2585
+ },
+ {
+ "epoch": 35.92227074235808,
+ "grad_norm": 0.015723245218396187,
+ "learning_rate": 0.0006,
+ "loss": 4.205283164978027,
+ "step": 2586
+ },
+ {
+ "epoch": 35.93624454148472,
+ "grad_norm": 0.014626456424593925,
+ "learning_rate": 0.0006,
+ "loss": 4.086552619934082,
+ "step": 2587
+ },
+ {
+ "epoch": 35.95021834061135,
+ "grad_norm": 0.014348454773426056,
+ "learning_rate": 0.0006,
+ "loss": 4.171329975128174,
+ "step": 2588
+ },
+ {
+ "epoch": 35.96419213973799,
+ "grad_norm": 0.015598480589687824,
+ "learning_rate": 0.0006,
+ "loss": 4.140815258026123,
+ "step": 2589
+ },
+ {
+ "epoch": 35.97816593886463,
+ "grad_norm": 0.017126062884926796,
+ "learning_rate": 0.0006,
+ "loss": 4.157176971435547,
+ "step": 2590
+ },
+ {
+ "epoch": 35.992139737991266,
+ "grad_norm": 0.01860199309885502,
+ "learning_rate": 0.0006,
+ "loss": 4.252880096435547,
+ "step": 2591
+ },
+ {
+ "epoch": 36.0,
+ "grad_norm": 0.021780220791697502,
+ "learning_rate": 0.0006,
+ "loss": 4.249861717224121,
+ "step": 2592
+ },
+ {
+ "epoch": 36.0,
+ "eval_loss": 4.606466293334961,
+ "eval_runtime": 56.6217,
+ "eval_samples_per_second": 43.128,
+ "eval_steps_per_second": 1.36,
+ "step": 2592
+ },
+ {
+ "epoch": 36.01397379912664,
+ "grad_norm": 0.020892828702926636,
+ "learning_rate": 0.0006,
+ "loss": 4.102055549621582,
+ "step": 2593
+ },
+ {
+ "epoch": 36.02794759825328,
+ "grad_norm": 0.020306255668401718,
+ "learning_rate": 0.0006,
+ "loss": 4.184308052062988,
+ "step": 2594
+ },
+ {
+ "epoch": 36.041921397379916,
+ "grad_norm": 0.02025097794830799,
+ "learning_rate": 0.0006,
+ "loss": 3.9691858291625977,
+ "step": 2595
+ },
+ {
+ "epoch": 36.05589519650655,
+ "grad_norm": 0.020254382863640785,
+ "learning_rate": 0.0006,
+ "loss": 4.02957820892334,
+ "step": 2596
+ },
+ {
+ "epoch": 36.069868995633186,
+ "grad_norm": 0.019307196140289307,
+ "learning_rate": 0.0006,
+ "loss": 4.033181667327881,
+ "step": 2597
+ },
+ {
+ "epoch": 36.083842794759825,
+ "grad_norm": 0.0198476854711771,
+ "learning_rate": 0.0006,
+ "loss": 4.0816874504089355,
+ "step": 2598
+ },
+ {
+ "epoch": 36.09781659388646,
+ "grad_norm": 0.019916515797376633,
+ "learning_rate": 0.0006,
+ "loss": 4.164889812469482,
+ "step": 2599
+ },
+ {
+ "epoch": 36.1117903930131,
+ "grad_norm": 0.0189144778996706,
+ "learning_rate": 0.0006,
+ "loss": 4.038427352905273,
+ "step": 2600
+ },
+ {
+ "epoch": 36.12576419213974,
+ "grad_norm": 0.01754768192768097,
+ "learning_rate": 0.0006,
+ "loss": 4.010427474975586,
+ "step": 2601
+ },
+ {
+ "epoch": 36.13973799126637,
+ "grad_norm": 0.016862990334630013,
+ "learning_rate": 0.0006,
+ "loss": 4.035970211029053,
+ "step": 2602
+ },
+ {
+ "epoch": 36.15371179039301,
+ "grad_norm": 0.019616752862930298,
+ "learning_rate": 0.0006,
+ "loss": 4.079898834228516,
+ "step": 2603
+ },
+ {
+ "epoch": 36.16768558951965,
+ "grad_norm": 0.020131465047597885,
+ "learning_rate": 0.0006,
+ "loss": 4.144684791564941,
+ "step": 2604
+ },
+ {
+ "epoch": 36.18165938864629,
+ "grad_norm": 0.02072521299123764,
+ "learning_rate": 0.0006,
+ "loss": 4.013660430908203,
+ "step": 2605
+ },
+ {
+ "epoch": 36.19563318777293,
+ "grad_norm": 0.01924615167081356,
+ "learning_rate": 0.0006,
+ "loss": 4.071398735046387,
+ "step": 2606
+ },
+ {
+ "epoch": 36.209606986899566,
+ "grad_norm": 0.02013954520225525,
+ "learning_rate": 0.0006,
+ "loss": 4.127094268798828,
+ "step": 2607
+ },
+ {
+ "epoch": 36.223580786026204,
+ "grad_norm": 0.02042008936405182,
+ "learning_rate": 0.0006,
+ "loss": 4.154261589050293,
+ "step": 2608
+ },
+ {
+ "epoch": 36.237554585152836,
+ "grad_norm": 0.021998876705765724,
+ "learning_rate": 0.0006,
+ "loss": 4.204150199890137,
+ "step": 2609
+ },
+ {
+ "epoch": 36.251528384279474,
+ "grad_norm": 0.022946683689951897,
+ "learning_rate": 0.0006,
+ "loss": 4.013632774353027,
+ "step": 2610
+ },
+ {
+ "epoch": 36.26550218340611,
+ "grad_norm": 0.02061830274760723,
+ "learning_rate": 0.0006,
+ "loss": 4.171084403991699,
+ "step": 2611
+ },
+ {
+ "epoch": 36.27947598253275,
+ "grad_norm": 0.01883152686059475,
+ "learning_rate": 0.0006,
+ "loss": 4.149848461151123,
+ "step": 2612
+ },
+ {
+ "epoch": 36.29344978165939,
+ "grad_norm": 0.019945282489061356,
+ "learning_rate": 0.0006,
+ "loss": 4.174656867980957,
+ "step": 2613
+ },
+ {
+ "epoch": 36.30742358078603,
+ "grad_norm": 0.020751051604747772,
+ "learning_rate": 0.0006,
+ "loss": 4.080747127532959,
+ "step": 2614
+ },
+ {
+ "epoch": 36.32139737991266,
+ "grad_norm": 0.01879691146314144,
+ "learning_rate": 0.0006,
+ "loss": 4.057356357574463,
+ "step": 2615
+ },
+ {
+ "epoch": 36.3353711790393,
+ "grad_norm": 0.019037071615457535,
+ "learning_rate": 0.0006,
+ "loss": 4.204607009887695,
+ "step": 2616
+ },
+ {
+ "epoch": 36.34934497816594,
+ "grad_norm": 0.01872078701853752,
+ "learning_rate": 0.0006,
+ "loss": 4.050586700439453,
+ "step": 2617
+ },
+ {
+ "epoch": 36.36331877729258,
+ "grad_norm": 0.018092811107635498,
+ "learning_rate": 0.0006,
+ "loss": 4.102832794189453,
+ "step": 2618
+ },
+ {
+ "epoch": 36.377292576419215,
+ "grad_norm": 0.017010487616062164,
+ "learning_rate": 0.0006,
+ "loss": 4.191788673400879,
+ "step": 2619
+ },
+ {
+ "epoch": 36.391266375545854,
+ "grad_norm": 0.01691143773496151,
+ "learning_rate": 0.0006,
+ "loss": 4.160456657409668,
+ "step": 2620
+ },
+ {
+ "epoch": 36.40524017467249,
+ "grad_norm": 0.017046816647052765,
+ "learning_rate": 0.0006,
+ "loss": 4.073828220367432,
+ "step": 2621
+ },
+ {
+ "epoch": 36.419213973799124,
+ "grad_norm": 0.01707981713116169,
+ "learning_rate": 0.0006,
+ "loss": 4.0980095863342285,
+ "step": 2622
+ },
+ {
+ "epoch": 36.43318777292576,
+ "grad_norm": 0.01630452275276184,
+ "learning_rate": 0.0006,
+ "loss": 4.101321220397949,
+ "step": 2623
+ },
+ {
+ "epoch": 36.4471615720524,
+ "grad_norm": 0.01576915942132473,
+ "learning_rate": 0.0006,
+ "loss": 4.158249855041504,
+ "step": 2624
+ },
+ {
+ "epoch": 36.46113537117904,
+ "grad_norm": 0.016179712489247322,
+ "learning_rate": 0.0006,
+ "loss": 4.077852249145508,
+ "step": 2625
+ },
+ {
+ "epoch": 36.47510917030568,
+ "grad_norm": 0.01609169878065586,
+ "learning_rate": 0.0006,
+ "loss": 4.187317848205566,
+ "step": 2626
+ },
+ {
+ "epoch": 36.48908296943232,
+ "grad_norm": 0.017148254439234734,
+ "learning_rate": 0.0006,
+ "loss": 4.109071731567383,
+ "step": 2627
+ },
+ {
+ "epoch": 36.50305676855895,
+ "grad_norm": 0.015842624008655548,
+ "learning_rate": 0.0006,
+ "loss": 4.145096302032471,
+ "step": 2628
+ },
+ {
+ "epoch": 36.51703056768559,
+ "grad_norm": 0.014906194061040878,
+ "learning_rate": 0.0006,
+ "loss": 4.028440475463867,
+ "step": 2629
+ },
+ {
+ "epoch": 36.531004366812226,
+ "grad_norm": 0.01485038734972477,
+ "learning_rate": 0.0006,
+ "loss": 4.175201416015625,
+ "step": 2630
+ },
+ {
+ "epoch": 36.544978165938865,
+ "grad_norm": 0.015834150835871696,
+ "learning_rate": 0.0006,
+ "loss": 4.076434135437012,
+ "step": 2631
+ },
+ {
+ "epoch": 36.5589519650655,
+ "grad_norm": 0.015411358326673508,
+ "learning_rate": 0.0006,
+ "loss": 4.03912878036499,
+ "step": 2632
+ },
+ {
+ "epoch": 36.57292576419214,
+ "grad_norm": 0.014762846753001213,
+ "learning_rate": 0.0006,
+ "loss": 4.019827365875244,
+ "step": 2633
+ },
+ {
+ "epoch": 36.58689956331878,
+ "grad_norm": 0.015269497409462929,
+ "learning_rate": 0.0006,
+ "loss": 4.116957187652588,
+ "step": 2634
+ },
+ {
+ "epoch": 36.60087336244541,
+ "grad_norm": 0.015171617269515991,
+ "learning_rate": 0.0006,
+ "loss": 4.2069926261901855,
+ "step": 2635
+ },
+ {
+ "epoch": 36.61484716157205,
+ "grad_norm": 0.014770902693271637,
+ "learning_rate": 0.0006,
+ "loss": 4.092021942138672,
+ "step": 2636
+ },
+ {
+ "epoch": 36.62882096069869,
+ "grad_norm": 0.014957531355321407,
+ "learning_rate": 0.0006,
+ "loss": 4.16819953918457,
+ "step": 2637
+ },
+ {
+ "epoch": 36.64279475982533,
+ "grad_norm": 0.014774451963603497,
+ "learning_rate": 0.0006,
+ "loss": 4.046073913574219,
+ "step": 2638
+ },
+ {
+ "epoch": 36.65676855895197,
+ "grad_norm": 0.014320445246994495,
+ "learning_rate": 0.0006,
+ "loss": 3.977792501449585,
+ "step": 2639
+ },
+ {
+ "epoch": 36.670742358078606,
+ "grad_norm": 0.01413130946457386,
+ "learning_rate": 0.0006,
+ "loss": 4.130649566650391,
+ "step": 2640
+ },
+ {
+ "epoch": 36.68471615720524,
+ "grad_norm": 0.014336307533085346,
+ "learning_rate": 0.0006,
+ "loss": 4.161357402801514,
+ "step": 2641
+ },
+ {
+ "epoch": 36.698689956331876,
+ "grad_norm": 0.014516931027173996,
+ "learning_rate": 0.0006,
+ "loss": 4.14248514175415,
+ "step": 2642
+ },
+ {
+ "epoch": 36.712663755458514,
+ "grad_norm": 0.014377126470208168,
+ "learning_rate": 0.0006,
+ "loss": 4.062711715698242,
+ "step": 2643
+ },
+ {
+ "epoch": 36.72663755458515,
+ "grad_norm": 0.014332668855786324,
+ "learning_rate": 0.0006,
+ "loss": 4.06423807144165,
+ "step": 2644
+ },
+ {
+ "epoch": 36.74061135371179,
+ "grad_norm": 0.014030571095645428,
+ "learning_rate": 0.0006,
+ "loss": 4.178613662719727,
+ "step": 2645
+ },
+ {
+ "epoch": 36.75458515283843,
+ "grad_norm": 0.014376776292920113,
+ "learning_rate": 0.0006,
+ "loss": 4.073130130767822,
+ "step": 2646
+ },
+ {
+ "epoch": 36.76855895196506,
+ "grad_norm": 0.013827716931700706,
+ "learning_rate": 0.0006,
+ "loss": 4.043869972229004,
+ "step": 2647
+ },
+ {
+ "epoch": 36.7825327510917,
+ "grad_norm": 0.015548762865364552,
+ "learning_rate": 0.0006,
+ "loss": 4.134049415588379,
+ "step": 2648
+ },
+ {
+ "epoch": 36.79650655021834,
+ "grad_norm": 0.017041970044374466,
+ "learning_rate": 0.0006,
+ "loss": 4.205024242401123,
+ "step": 2649
+ },
+ {
+ "epoch": 36.81048034934498,
+ "grad_norm": 0.017638275399804115,
+ "learning_rate": 0.0006,
+ "loss": 4.199924468994141,
+ "step": 2650
+ },
+ {
+ "epoch": 36.82445414847162,
+ "grad_norm": 0.017771627753973007,
+ "learning_rate": 0.0006,
+ "loss": 4.15706729888916,
+ "step": 2651
+ },
+ {
+ "epoch": 36.838427947598255,
+ "grad_norm": 0.015933562070131302,
+ "learning_rate": 0.0006,
+ "loss": 4.220742225646973,
+ "step": 2652
+ },
+ {
+ "epoch": 36.852401746724894,
+ "grad_norm": 0.01668144389986992,
+ "learning_rate": 0.0006,
+ "loss": 4.161426544189453,
+ "step": 2653
+ },
+ {
+ "epoch": 36.866375545851525,
+ "grad_norm": 0.018014680594205856,
+ "learning_rate": 0.0006,
+ "loss": 4.183046817779541,
+ "step": 2654
+ },
+ {
+ "epoch": 36.880349344978164,
+ "grad_norm": 0.018808670341968536,
+ "learning_rate": 0.0006,
+ "loss": 4.105293273925781,
+ "step": 2655
+ },
+ {
+ "epoch": 36.8943231441048,
+ "grad_norm": 0.018003476783633232,
+ "learning_rate": 0.0006,
+ "loss": 4.103171348571777,
+ "step": 2656
+ },
+ {
+ "epoch": 36.90829694323144,
+ "grad_norm": 0.016865914687514305,
+ "learning_rate": 0.0006,
+ "loss": 4.125634670257568,
+ "step": 2657
+ },
+ {
+ "epoch": 36.92227074235808,
+ "grad_norm": 0.018371153622865677,
+ "learning_rate": 0.0006,
+ "loss": 4.192461013793945,
+ "step": 2658
+ },
+ {
+ "epoch": 36.93624454148472,
+ "grad_norm": 0.019045377150177956,
+ "learning_rate": 0.0006,
+ "loss": 4.198105812072754,
+ "step": 2659
+ },
+ {
+ "epoch": 36.95021834061135,
+ "grad_norm": 0.01594168320298195,
+ "learning_rate": 0.0006,
+ "loss": 4.169788360595703,
+ "step": 2660
+ },
+ {
+ "epoch": 36.96419213973799,
+ "grad_norm": 0.015645820647478104,
+ "learning_rate": 0.0006,
+ "loss": 4.047277450561523,
+ "step": 2661
+ },
+ {
+ "epoch": 36.97816593886463,
+ "grad_norm": 0.016313156113028526,
+ "learning_rate": 0.0006,
+ "loss": 4.087392807006836,
+ "step": 2662
+ },
+ {
+ "epoch": 36.992139737991266,
+ "grad_norm": 0.017050625756382942,
+ "learning_rate": 0.0006,
+ "loss": 4.0908002853393555,
+ "step": 2663
+ },
+ {
+ "epoch": 37.0,
+ "grad_norm": 0.017730968073010445,
+ "learning_rate": 0.0006,
+ "loss": 4.110957622528076,
+ "step": 2664
+ },
+ {
+ "epoch": 37.0,
+ "eval_loss": 4.563793182373047,
+ "eval_runtime": 56.5922,
+ "eval_samples_per_second": 43.151,
+ "eval_steps_per_second": 1.361,
+ "step": 2664
+ },
+ {
+ "epoch": 37.01397379912664,
+ "grad_norm": 0.015687910839915276,
+ "learning_rate": 0.0006,
+ "loss": 4.143589019775391,
+ "step": 2665
+ },
+ {
+ "epoch": 37.02794759825328,
+ "grad_norm": 0.017451366409659386,
+ "learning_rate": 0.0006,
+ "loss": 4.0326666831970215,
+ "step": 2666
+ },
+ {
+ "epoch": 37.041921397379916,
+ "grad_norm": 0.01909538544714451,
+ "learning_rate": 0.0006,
+ "loss": 3.9706249237060547,
+ "step": 2667
+ },
+ {
+ "epoch": 37.05589519650655,
+ "grad_norm": 0.01988379657268524,
+ "learning_rate": 0.0006,
+ "loss": 3.9531009197235107,
+ "step": 2668
+ },
+ {
+ "epoch": 37.069868995633186,
+ "grad_norm": 0.02035115472972393,
+ "learning_rate": 0.0006,
+ "loss": 3.962214469909668,
+ "step": 2669
+ },
+ {
+ "epoch": 37.083842794759825,
+ "grad_norm": 0.01952001079916954,
+ "learning_rate": 0.0006,
+ "loss": 4.182262897491455,
+ "step": 2670
+ },
+ {
+ "epoch": 37.09781659388646,
+ "grad_norm": 0.01854325644671917,
+ "learning_rate": 0.0006,
+ "loss": 4.127663612365723,
+ "step": 2671
+ },
+ {
+ "epoch": 37.1117903930131,
+ "grad_norm": 0.017757616937160492,
+ "learning_rate": 0.0006,
+ "loss": 4.100804805755615,
+ "step": 2672
+ },
+ {
+ "epoch": 37.12576419213974,
+ "grad_norm": 0.016712479293346405,
+ "learning_rate": 0.0006,
+ "loss": 4.026284217834473,
+ "step": 2673
+ },
+ {
+ "epoch": 37.13973799126637,
+ "grad_norm": 0.0168139711022377,
+ "learning_rate": 0.0006,
+ "loss": 4.183882713317871,
+ "step": 2674
+ },
+ {
+ "epoch": 37.15371179039301,
+ "grad_norm": 0.017060423269867897,
+ "learning_rate": 0.0006,
+ "loss": 4.131832599639893,
+ "step": 2675
+ },
+ {
+ "epoch": 37.16768558951965,
+ "grad_norm": 0.01726631447672844,
+ "learning_rate": 0.0006,
+ "loss": 3.9156057834625244,
+ "step": 2676
+ },
+ {
+ "epoch": 37.18165938864629,
+ "grad_norm": 0.016348622739315033,
+ "learning_rate": 0.0006,
+ "loss": 4.1731367111206055,
+ "step": 2677
+ },
+ {
+ "epoch": 37.19563318777293,
+ "grad_norm": 0.016640059649944305,
+ "learning_rate": 0.0006,
+ "loss": 4.10997200012207,
+ "step": 2678
+ },
+ {
+ "epoch": 37.209606986899566,
+ "grad_norm": 0.01601342298090458,
+ "learning_rate": 0.0006,
+ "loss": 4.081145286560059,
+ "step": 2679
+ },
+ {
+ "epoch": 37.223580786026204,
+ "grad_norm": 0.01735631190240383,
+ "learning_rate": 0.0006,
+ "loss": 4.202108860015869,
+ "step": 2680
+ },
+ {
+ "epoch": 37.237554585152836,
+ "grad_norm": 0.019164685159921646,
+ "learning_rate": 0.0006,
+ "loss": 4.104109764099121,
+ "step": 2681
+ },
+ {
+ "epoch": 37.251528384279474,
+ "grad_norm": 0.01881759613752365,
+ "learning_rate": 0.0006,
+ "loss": 4.145532131195068,
+ "step": 2682
+ },
+ {
+ "epoch": 37.26550218340611,
+ "grad_norm": 0.017656555399298668,
+ "learning_rate": 0.0006,
+ "loss": 4.012782096862793,
+ "step": 2683
+ },
+ {
+ "epoch": 37.27947598253275,
+ "grad_norm": 0.01624823734164238,
+ "learning_rate": 0.0006,
+ "loss": 4.001727104187012,
+ "step": 2684
+ },
+ {
+ "epoch": 37.29344978165939,
+ "grad_norm": 0.016973139718174934,
+ "learning_rate": 0.0006,
+ "loss": 4.095399856567383,
+ "step": 2685
+ },
+ {
+ "epoch": 37.30742358078603,
+ "grad_norm": 0.015503671951591969,
+ "learning_rate": 0.0006,
+ "loss": 4.034431457519531,
+ "step": 2686
+ },
+ {
+ "epoch": 37.32139737991266,
+ "grad_norm": 0.015370250679552555,
+ "learning_rate": 0.0006,
+ "loss": 4.075836181640625,
+ "step": 2687
+ },
+ {
+ "epoch": 37.3353711790393,
+ "grad_norm": 0.01564021222293377,
+ "learning_rate": 0.0006,
+ "loss": 4.125202178955078,
+ "step": 2688
+ },
+ {
+ "epoch": 37.34934497816594,
+ "grad_norm": 0.016717975959181786,
+ "learning_rate": 0.0006,
+ "loss": 4.153017044067383,
+ "step": 2689
+ },
+ {
+ "epoch": 37.36331877729258,
+ "grad_norm": 0.015263617038726807,
+ "learning_rate": 0.0006,
+ "loss": 4.021814823150635,
+ "step": 2690
+ },
+ {
+ "epoch": 37.377292576419215,
+ "grad_norm": 0.016497142612934113,
+ "learning_rate": 0.0006,
+ "loss": 4.069095611572266,
+ "step": 2691
+ },
+ {
+ "epoch": 37.391266375545854,
+ "grad_norm": 0.01669355481863022,
+ "learning_rate": 0.0006,
+ "loss": 4.174587726593018,
+ "step": 2692
+ },
+ {
+ "epoch": 37.40524017467249,
+ "grad_norm": 0.019695773720741272,
+ "learning_rate": 0.0006,
+ "loss": 4.0073041915893555,
+ "step": 2693
+ },
+ {
+ "epoch": 37.419213973799124,
+ "grad_norm": 0.023720644414424896,
+ "learning_rate": 0.0006,
+ "loss": 4.111948013305664,
+ "step": 2694
+ },
+ {
+ "epoch": 37.43318777292576,
+ "grad_norm": 0.02399350143969059,
+ "learning_rate": 0.0006,
+ "loss": 4.081164360046387,
+ "step": 2695
+ },
+ {
+ "epoch": 37.4471615720524,
+ "grad_norm": 0.019630271941423416,
+ "learning_rate": 0.0006,
+ "loss": 4.065731048583984,
+ "step": 2696
+ },
+ {
+ "epoch": 37.46113537117904,
+ "grad_norm": 0.019679008051753044,
+ "learning_rate": 0.0006,
+ "loss": 3.993852376937866,
+ "step": 2697
+ },
+ {
+ "epoch": 37.47510917030568,
+ "grad_norm": 0.020597299560904503,
+ "learning_rate": 0.0006,
+ "loss": 4.275555610656738,
+ "step": 2698
+ },
+ {
+ "epoch": 37.48908296943232,
+ "grad_norm": 0.01983889564871788,
+ "learning_rate": 0.0006,
+ "loss": 4.254708290100098,
+ "step": 2699
+ },
+ {
+ "epoch": 37.50305676855895,
+ "grad_norm": 0.019150668755173683,
+ "learning_rate": 0.0006,
+ "loss": 4.097885608673096,
+ "step": 2700
+ },
+ {
+ "epoch": 37.51703056768559,
+ "grad_norm": 0.018847715109586716,
+ "learning_rate": 0.0006,
+ "loss": 4.054081439971924,
+ "step": 2701
+ },
+ {
+ "epoch": 37.531004366812226,
+ "grad_norm": 0.019607288762927055,
+ "learning_rate": 0.0006,
+ "loss": 3.9891440868377686,
+ "step": 2702
+ },
+ {
+ "epoch": 37.544978165938865,
+ "grad_norm": 0.018863150849938393,
+ "learning_rate": 0.0006,
+ "loss": 4.084877014160156,
+ "step": 2703
+ },
+ {
+ "epoch": 37.5589519650655,
+ "grad_norm": 0.017340239137411118,
+ "learning_rate": 0.0006,
+ "loss": 4.113290786743164,
+ "step": 2704
+ },
+ {
+ "epoch": 37.57292576419214,
+ "grad_norm": 0.015628693625330925,
+ "learning_rate": 0.0006,
+ "loss": 4.1483001708984375,
+ "step": 2705
+ },
+ {
+ "epoch": 37.58689956331878,
+ "grad_norm": 0.016493283212184906,
+ "learning_rate": 0.0006,
+ "loss": 4.1040143966674805,
+ "step": 2706
+ },
+ {
+ "epoch": 37.60087336244541,
+ "grad_norm": 0.016732443124055862,
+ "learning_rate": 0.0006,
+ "loss": 4.07205057144165,
+ "step": 2707
+ },
+ {
+ "epoch": 37.61484716157205,
+ "grad_norm": 0.017176156863570213,
+ "learning_rate": 0.0006,
+ "loss": 4.1279754638671875,
+ "step": 2708
+ },
+ {
+ "epoch": 37.62882096069869,
+ "grad_norm": 0.015730608254671097,
+ "learning_rate": 0.0006,
+ "loss": 4.086658000946045,
+ "step": 2709
+ },
+ {
+ "epoch": 37.64279475982533,
+ "grad_norm": 0.016111090779304504,
+ "learning_rate": 0.0006,
+ "loss": 4.064335823059082,
+ "step": 2710
+ },
+ {
+ "epoch": 37.65676855895197,
+ "grad_norm": 0.01769379712641239,
+ "learning_rate": 0.0006,
+ "loss": 4.116047382354736,
+ "step": 2711
+ },
+ {
+ "epoch": 37.670742358078606,
+ "grad_norm": 0.01683102548122406,
+ "learning_rate": 0.0006,
+ "loss": 4.095986843109131,
+ "step": 2712
+ },
+ {
+ "epoch": 37.68471615720524,
+ "grad_norm": 0.015726450830698013,
+ "learning_rate": 0.0006,
+ "loss": 4.10775089263916,
+ "step": 2713
+ },
+ {
+ "epoch": 37.698689956331876,
+ "grad_norm": 0.014760235324501991,
+ "learning_rate": 0.0006,
+ "loss": 4.1234893798828125,
+ "step": 2714
+ },
+ {
+ "epoch": 37.712663755458514,
+ "grad_norm": 0.014933438040316105,
+ "learning_rate": 0.0006,
+ "loss": 4.079200744628906,
+ "step": 2715
+ },
+ {
+ "epoch": 37.72663755458515,
+ "grad_norm": 0.015723997727036476,
+ "learning_rate": 0.0006,
+ "loss": 4.07865047454834,
+ "step": 2716
+ },
+ {
+ "epoch": 37.74061135371179,
+ "grad_norm": 0.014539001509547234,
+ "learning_rate": 0.0006,
+ "loss": 4.036242485046387,
+ "step": 2717
+ },
+ {
+ "epoch": 37.75458515283843,
+ "grad_norm": 0.014844979159533978,
+ "learning_rate": 0.0006,
+ "loss": 4.1051483154296875,
+ "step": 2718
+ },
+ {
+ "epoch": 37.76855895196506,
+ "grad_norm": 0.015939844772219658,
+ "learning_rate": 0.0006,
+ "loss": 4.017253875732422,
+ "step": 2719
+ },
+ {
+ "epoch": 37.7825327510917,
+ "grad_norm": 0.014771555550396442,
+ "learning_rate": 0.0006,
+ "loss": 4.0694661140441895,
+ "step": 2720
+ },
+ {
+ "epoch": 37.79650655021834,
+ "grad_norm": 0.015777921304106712,
+ "learning_rate": 0.0006,
+ "loss": 3.947587490081787,
+ "step": 2721
+ },
+ {
+ "epoch": 37.81048034934498,
+ "grad_norm": 0.017451291903853416,
+ "learning_rate": 0.0006,
+ "loss": 4.093625068664551,
+ "step": 2722
+ },
+ {
+ "epoch": 37.82445414847162,
+ "grad_norm": 0.016829246655106544,
+ "learning_rate": 0.0006,
+ "loss": 4.09030818939209,
+ "step": 2723
+ },
+ {
+ "epoch": 37.838427947598255,
+ "grad_norm": 0.016984397545456886,
+ "learning_rate": 0.0006,
+ "loss": 4.140626907348633,
+ "step": 2724
+ },
+ {
+ "epoch": 37.852401746724894,
+ "grad_norm": 0.01732603646814823,
+ "learning_rate": 0.0006,
+ "loss": 4.128276824951172,
+ "step": 2725
+ },
+ {
+ "epoch": 37.866375545851525,
+ "grad_norm": 0.01744619384407997,
+ "learning_rate": 0.0006,
+ "loss": 4.103130340576172,
+ "step": 2726
+ },
+ {
+ "epoch": 37.880349344978164,
+ "grad_norm": 0.016331320628523827,
+ "learning_rate": 0.0006,
+ "loss": 4.037951469421387,
+ "step": 2727
+ },
+ {
+ "epoch": 37.8943231441048,
+ "grad_norm": 0.015919430181384087,
+ "learning_rate": 0.0006,
+ "loss": 4.121513366699219,
+ "step": 2728
+ },
+ {
+ "epoch": 37.90829694323144,
+ "grad_norm": 0.016153695061802864,
+ "learning_rate": 0.0006,
+ "loss": 4.099384307861328,
+ "step": 2729
+ },
+ {
+ "epoch": 37.92227074235808,
+ "grad_norm": 0.01628800667822361,
+ "learning_rate": 0.0006,
+ "loss": 4.092325687408447,
+ "step": 2730
+ },
+ {
+ "epoch": 37.93624454148472,
+ "grad_norm": 0.017285684123635292,
+ "learning_rate": 0.0006,
+ "loss": 3.9828906059265137,
+ "step": 2731
+ },
+ {
+ "epoch": 37.95021834061135,
+ "grad_norm": 0.016047554090619087,
+ "learning_rate": 0.0006,
+ "loss": 4.100419044494629,
+ "step": 2732
+ },
+ {
+ "epoch": 37.96419213973799,
+ "grad_norm": 0.01562986709177494,
+ "learning_rate": 0.0006,
+ "loss": 4.065832614898682,
+ "step": 2733
+ },
+ {
+ "epoch": 37.97816593886463,
+ "grad_norm": 0.01677163504064083,
+ "learning_rate": 0.0006,
+ "loss": 4.077518463134766,
+ "step": 2734
+ },
+ {
+ "epoch": 37.992139737991266,
+ "grad_norm": 0.01755492016673088,
+ "learning_rate": 0.0006,
+ "loss": 4.1076154708862305,
+ "step": 2735
+ },
+ {
+ "epoch": 38.0,
+ "grad_norm": 0.01929762214422226,
+ "learning_rate": 0.0006,
+ "loss": 4.09603214263916,
+ "step": 2736
+ },
+ {
+ "epoch": 38.0,
+ "eval_loss": 4.582442283630371,
+ "eval_runtime": 62.8649,
+ "eval_samples_per_second": 38.845,
+ "eval_steps_per_second": 1.225,
+ "step": 2736
+ },
+ {
+ "epoch": 38.01397379912664,
+ "grad_norm": 0.019129136577248573,
+ "learning_rate": 0.0006,
+ "loss": 4.095005989074707,
+ "step": 2737
+ },
+ {
+ "epoch": 38.02794759825328,
+ "grad_norm": 0.019167408347129822,
+ "learning_rate": 0.0006,
+ "loss": 4.054426670074463,
+ "step": 2738
+ },
+ {
+ "epoch": 38.041921397379916,
+ "grad_norm": 0.01789460889995098,
+ "learning_rate": 0.0006,
+ "loss": 4.0434417724609375,
+ "step": 2739
+ },
+ {
+ "epoch": 38.05589519650655,
+ "grad_norm": 0.015947027131915092,
+ "learning_rate": 0.0006,
+ "loss": 4.074516296386719,
+ "step": 2740
+ },
+ {
+ "epoch": 38.069868995633186,
+ "grad_norm": 0.016667252406477928,
+ "learning_rate": 0.0006,
+ "loss": 4.108545303344727,
+ "step": 2741
+ },
+ {
+ "epoch": 38.083842794759825,
+ "grad_norm": 0.016014399006962776,
+ "learning_rate": 0.0006,
+ "loss": 4.026925086975098,
+ "step": 2742
+ },
+ {
+ "epoch": 38.09781659388646,
+ "grad_norm": 0.015397695824503899,
+ "learning_rate": 0.0006,
+ "loss": 4.163970470428467,
+ "step": 2743
+ },
+ {
+ "epoch": 38.1117903930131,
+ "grad_norm": 0.01563289947807789,
+ "learning_rate": 0.0006,
+ "loss": 4.116578578948975,
+ "step": 2744
+ },
+ {
+ "epoch": 38.12576419213974,
+ "grad_norm": 0.015512318350374699,
+ "learning_rate": 0.0006,
+ "loss": 4.096311569213867,
+ "step": 2745
+ },
+ {
+ "epoch": 38.13973799126637,
+ "grad_norm": 0.015426642261445522,
+ "learning_rate": 0.0006,
+ "loss": 4.017065048217773,
+ "step": 2746
+ },
+ {
+ "epoch": 38.15371179039301,
+ "grad_norm": 0.015115435235202312,
+ "learning_rate": 0.0006,
+ "loss": 4.005434989929199,
+ "step": 2747
+ },
+ {
+ "epoch": 38.16768558951965,
+ "grad_norm": 0.015235894359648228,
+ "learning_rate": 0.0006,
+ "loss": 3.982743263244629,
+ "step": 2748
+ },
+ {
+ "epoch": 38.18165938864629,
+ "grad_norm": 0.016095101833343506,
+ "learning_rate": 0.0006,
+ "loss": 4.0849385261535645,
+ "step": 2749
+ },
+ {
+ "epoch": 38.19563318777293,
+ "grad_norm": 0.015201197937130928,
+ "learning_rate": 0.0006,
+ "loss": 4.103863716125488,
+ "step": 2750
+ },
+ {
+ "epoch": 38.209606986899566,
+ "grad_norm": 0.015687478706240654,
+ "learning_rate": 0.0006,
+ "loss": 4.12074089050293,
+ "step": 2751
+ },
+ {
+ "epoch": 38.223580786026204,
+ "grad_norm": 0.015201589092612267,
+ "learning_rate": 0.0006,
+ "loss": 4.056872367858887,
+ "step": 2752
+ },
+ {
+ "epoch": 38.237554585152836,
+ "grad_norm": 0.015154894441366196,
+ "learning_rate": 0.0006,
+ "loss": 4.089670181274414,
+ "step": 2753
+ },
+ {
+ "epoch": 38.251528384279474,
+ "grad_norm": 0.01586608588695526,
+ "learning_rate": 0.0006,
+ "loss": 4.155317306518555,
+ "step": 2754
+ },
+ {
+ "epoch": 38.26550218340611,
+ "grad_norm": 0.01609654724597931,
+ "learning_rate": 0.0006,
+ "loss": 3.995800495147705,
+ "step": 2755
+ },
+ {
+ "epoch": 38.27947598253275,
+ "grad_norm": 0.015495185740292072,
+ "learning_rate": 0.0006,
+ "loss": 3.995769739151001,
+ "step": 2756
+ },
+ {
+ "epoch": 38.29344978165939,
+ "grad_norm": 0.017042381688952446,
+ "learning_rate": 0.0006,
+ "loss": 4.132832050323486,
+ "step": 2757
+ },
+ {
+ "epoch": 38.30742358078603,
+ "grad_norm": 0.018706614151597023,
+ "learning_rate": 0.0006,
+ "loss": 4.038820743560791,
+ "step": 2758
+ },
+ {
+ "epoch": 38.32139737991266,
+ "grad_norm": 0.018316026777029037,
+ "learning_rate": 0.0006,
+ "loss": 3.972052574157715,
+ "step": 2759
+ },
+ {
+ "epoch": 38.3353711790393,
+ "grad_norm": 0.018375389277935028,
+ "learning_rate": 0.0006,
+ "loss": 4.152097702026367,
+ "step": 2760
+ },
+ {
+ "epoch": 38.34934497816594,
+ "grad_norm": 0.016879886388778687,
+ "learning_rate": 0.0006,
+ "loss": 4.020869731903076,
+ "step": 2761
+ },
+ {
+ "epoch": 38.36331877729258,
+ "grad_norm": 0.015728183090686798,
+ "learning_rate": 0.0006,
+ "loss": 4.008016586303711,
+ "step": 2762
+ },
+ {
+ "epoch": 38.377292576419215,
+ "grad_norm": 0.016147758811712265,
+ "learning_rate": 0.0006,
+ "loss": 4.035854339599609,
+ "step": 2763
+ },
+ {
+ "epoch": 38.391266375545854,
+ "grad_norm": 0.017774797976017,
+ "learning_rate": 0.0006,
+ "loss": 4.122506141662598,
+ "step": 2764
+ },
+ {
+ "epoch": 38.40524017467249,
+ "grad_norm": 0.016590069979429245,
+ "learning_rate": 0.0006,
+ "loss": 4.082847595214844,
+ "step": 2765
+ },
+ {
+ "epoch": 38.419213973799124,
+ "grad_norm": 0.016544528305530548,
+ "learning_rate": 0.0006,
+ "loss": 4.184270858764648,
+ "step": 2766
+ },
+ {
+ "epoch": 38.43318777292576,
+ "grad_norm": 0.017972925677895546,
+ "learning_rate": 0.0006,
+ "loss": 3.987558603286743,
+ "step": 2767
+ },
+ {
+ "epoch": 38.4471615720524,
+ "grad_norm": 0.018562227487564087,
+ "learning_rate": 0.0006,
+ "loss": 4.1193461418151855,
+ "step": 2768
+ },
+ {
+ "epoch": 38.46113537117904,
+ "grad_norm": 0.018896685913205147,
+ "learning_rate": 0.0006,
+ "loss": 4.057808876037598,
+ "step": 2769
+ },
+ {
+ "epoch": 38.47510917030568,
+ "grad_norm": 0.016793809831142426,
+ "learning_rate": 0.0006,
+ "loss": 4.069523334503174,
+ "step": 2770
+ },
+ {
+ "epoch": 38.48908296943232,
+ "grad_norm": 0.01698676310479641,
+ "learning_rate": 0.0006,
+ "loss": 3.997624158859253,
+ "step": 2771
+ },
+ {
+ "epoch": 38.50305676855895,
+ "grad_norm": 0.016784202307462692,
+ "learning_rate": 0.0006,
+ "loss": 4.174795627593994,
+ "step": 2772
+ },
+ {
+ "epoch": 38.51703056768559,
+ "grad_norm": 0.016785014420747757,
+ "learning_rate": 0.0006,
+ "loss": 3.999321222305298,
+ "step": 2773
+ },
+ {
+ "epoch": 38.531004366812226,
+ "grad_norm": 0.018615184351801872,
+ "learning_rate": 0.0006,
+ "loss": 4.191237926483154,
+ "step": 2774
+ },
+ {
+ "epoch": 38.544978165938865,
+ "grad_norm": 0.019090790301561356,
+ "learning_rate": 0.0006,
+ "loss": 4.106966018676758,
+ "step": 2775
+ },
+ {
+ "epoch": 38.5589519650655,
+ "grad_norm": 0.018298320472240448,
+ "learning_rate": 0.0006,
+ "loss": 4.063363552093506,
+ "step": 2776
+ },
+ {
+ "epoch": 38.57292576419214,
+ "grad_norm": 0.018541090190410614,
+ "learning_rate": 0.0006,
+ "loss": 4.101193428039551,
+ "step": 2777
+ },
+ {
+ "epoch": 38.58689956331878,
+ "grad_norm": 0.01849970407783985,
+ "learning_rate": 0.0006,
+ "loss": 4.071335792541504,
+ "step": 2778
+ },
+ {
+ "epoch": 38.60087336244541,
+ "grad_norm": 0.018309077247977257,
+ "learning_rate": 0.0006,
+ "loss": 4.245305061340332,
+ "step": 2779
+ },
+ {
+ "epoch": 38.61484716157205,
+ "grad_norm": 0.017681840807199478,
+ "learning_rate": 0.0006,
+ "loss": 4.081173419952393,
+ "step": 2780
+ },
+ {
+ "epoch": 38.62882096069869,
+ "grad_norm": 0.016286224126815796,
+ "learning_rate": 0.0006,
+ "loss": 4.024351596832275,
+ "step": 2781
+ },
+ {
+ "epoch": 38.64279475982533,
+ "grad_norm": 0.015837235376238823,
+ "learning_rate": 0.0006,
+ "loss": 4.028205394744873,
+ "step": 2782
+ },
+ {
+ "epoch": 38.65676855895197,
+ "grad_norm": 0.016062596812844276,
+ "learning_rate": 0.0006,
+ "loss": 4.131122589111328,
+ "step": 2783
+ },
+ {
+ "epoch": 38.670742358078606,
+ "grad_norm": 0.01711239665746689,
+ "learning_rate": 0.0006,
+ "loss": 4.124416351318359,
+ "step": 2784
+ },
+ {
+ "epoch": 38.68471615720524,
+ "grad_norm": 0.015304913744330406,
+ "learning_rate": 0.0006,
+ "loss": 4.056122779846191,
+ "step": 2785
+ },
+ {
+ "epoch": 38.698689956331876,
+ "grad_norm": 0.015971887856721878,
+ "learning_rate": 0.0006,
+ "loss": 4.132555961608887,
+ "step": 2786
+ },
+ {
+ "epoch": 38.712663755458514,
+ "grad_norm": 0.015787221491336823,
+ "learning_rate": 0.0006,
+ "loss": 4.051596164703369,
+ "step": 2787
+ },
+ {
+ "epoch": 38.72663755458515,
+ "grad_norm": 0.01605444774031639,
+ "learning_rate": 0.0006,
+ "loss": 4.191808700561523,
+ "step": 2788
+ },
+ {
+ "epoch": 38.74061135371179,
+ "grad_norm": 0.016383009031414986,
+ "learning_rate": 0.0006,
+ "loss": 4.0648417472839355,
+ "step": 2789
+ },
+ {
+ "epoch": 38.75458515283843,
+ "grad_norm": 0.016135554760694504,
+ "learning_rate": 0.0006,
+ "loss": 3.988563060760498,
+ "step": 2790
+ },
+ {
+ "epoch": 38.76855895196506,
+ "grad_norm": 0.016634158790111542,
+ "learning_rate": 0.0006,
+ "loss": 4.133493423461914,
+ "step": 2791
+ },
+ {
+ "epoch": 38.7825327510917,
+ "grad_norm": 0.01688210666179657,
+ "learning_rate": 0.0006,
+ "loss": 4.091277122497559,
+ "step": 2792
+ },
+ {
+ "epoch": 38.79650655021834,
+ "grad_norm": 0.016102122142910957,
+ "learning_rate": 0.0006,
+ "loss": 4.051219940185547,
+ "step": 2793
+ },
+ {
+ "epoch": 38.81048034934498,
+ "grad_norm": 0.016585038974881172,
+ "learning_rate": 0.0006,
+ "loss": 4.058587074279785,
+ "step": 2794
+ },
+ {
+ "epoch": 38.82445414847162,
+ "grad_norm": 0.017700951546430588,
+ "learning_rate": 0.0006,
+ "loss": 4.016030311584473,
+ "step": 2795
+ },
+ {
+ "epoch": 38.838427947598255,
+ "grad_norm": 0.017922203987836838,
+ "learning_rate": 0.0006,
+ "loss": 4.079977989196777,
+ "step": 2796
+ },
+ {
+ "epoch": 38.852401746724894,
+ "grad_norm": 0.017299890518188477,
+ "learning_rate": 0.0006,
+ "loss": 4.062984943389893,
+ "step": 2797
+ },
+ {
+ "epoch": 38.866375545851525,
+ "grad_norm": 0.018648581579327583,
+ "learning_rate": 0.0006,
+ "loss": 4.065717697143555,
+ "step": 2798
+ },
+ {
+ "epoch": 38.880349344978164,
+ "grad_norm": 0.019057920202612877,
+ "learning_rate": 0.0006,
+ "loss": 4.227892875671387,
+ "step": 2799
+ },
+ {
+ "epoch": 38.8943231441048,
+ "grad_norm": 0.018824422731995583,
+ "learning_rate": 0.0006,
+ "loss": 4.116058349609375,
+ "step": 2800
+ },
+ {
+ "epoch": 38.90829694323144,
+ "grad_norm": 0.019084978848695755,
+ "learning_rate": 0.0006,
+ "loss": 4.061677932739258,
+ "step": 2801
+ },
+ {
+ "epoch": 38.92227074235808,
+ "grad_norm": 0.017842981964349747,
+ "learning_rate": 0.0006,
+ "loss": 4.083813667297363,
+ "step": 2802
+ },
+ {
+ "epoch": 38.93624454148472,
+ "grad_norm": 0.01880858838558197,
+ "learning_rate": 0.0006,
+ "loss": 4.152561187744141,
+ "step": 2803
+ },
+ {
+ "epoch": 38.95021834061135,
+ "grad_norm": 0.017400087788701057,
+ "learning_rate": 0.0006,
+ "loss": 4.154635429382324,
+ "step": 2804
+ },
+ {
+ "epoch": 38.96419213973799,
+ "grad_norm": 0.017682507634162903,
+ "learning_rate": 0.0006,
+ "loss": 4.151128768920898,
+ "step": 2805
+ },
+ {
+ "epoch": 38.97816593886463,
+ "grad_norm": 0.015803052112460136,
+ "learning_rate": 0.0006,
+ "loss": 4.10498571395874,
+ "step": 2806
+ },
+ {
+ "epoch": 38.992139737991266,
+ "grad_norm": 0.01582511141896248,
+ "learning_rate": 0.0006,
+ "loss": 4.140159606933594,
+ "step": 2807
+ },
+ {
+ "epoch": 39.0,
+ "grad_norm": 0.01925666816532612,
+ "learning_rate": 0.0006,
+ "loss": 4.126394271850586,
+ "step": 2808
+ },
+ {
+ "epoch": 39.0,
+ "eval_loss": 4.615617275238037,
+ "eval_runtime": 56.8895,
+ "eval_samples_per_second": 42.925,
+ "eval_steps_per_second": 1.354,
+ "step": 2808
+ },
+ {
+ "epoch": 39.01397379912664,
+ "grad_norm": 0.01677853614091873,
+ "learning_rate": 0.0006,
+ "loss": 4.102395057678223,
+ "step": 2809
+ },
+ {
+ "epoch": 39.02794759825328,
+ "grad_norm": 0.017901137471199036,
+ "learning_rate": 0.0006,
+ "loss": 3.976609230041504,
+ "step": 2810
+ },
+ {
+ "epoch": 39.041921397379916,
+ "grad_norm": 0.017978668212890625,
+ "learning_rate": 0.0006,
+ "loss": 3.944416046142578,
+ "step": 2811
+ },
+ {
+ "epoch": 39.05589519650655,
+ "grad_norm": 0.019967492669820786,
+ "learning_rate": 0.0006,
+ "loss": 3.992234468460083,
+ "step": 2812
+ },
+ {
+ "epoch": 39.069868995633186,
+ "grad_norm": 0.018937544897198677,
+ "learning_rate": 0.0006,
+ "loss": 4.028573036193848,
+ "step": 2813
+ },
+ {
+ "epoch": 39.083842794759825,
+ "grad_norm": 0.016852952539920807,
+ "learning_rate": 0.0006,
+ "loss": 4.062442779541016,
+ "step": 2814
+ },
+ {
+ "epoch": 39.09781659388646,
+ "grad_norm": 0.016643647104501724,
+ "learning_rate": 0.0006,
+ "loss": 4.08871603012085,
+ "step": 2815
+ },
+ {
+ "epoch": 39.1117903930131,
+ "grad_norm": 0.016872866079211235,
+ "learning_rate": 0.0006,
+ "loss": 3.9514245986938477,
+ "step": 2816
+ },
+ {
+ "epoch": 39.12576419213974,
+ "grad_norm": 0.018045669421553612,
+ "learning_rate": 0.0006,
+ "loss": 4.038842678070068,
+ "step": 2817
+ },
+ {
+ "epoch": 39.13973799126637,
+ "grad_norm": 0.017853403463959694,
+ "learning_rate": 0.0006,
+ "loss": 3.93709135055542,
+ "step": 2818
+ },
+ {
+ "epoch": 39.15371179039301,
+ "grad_norm": 0.017507443204522133,
+ "learning_rate": 0.0006,
+ "loss": 3.98665714263916,
+ "step": 2819
+ },
+ {
+ "epoch": 39.16768558951965,
+ "grad_norm": 0.01766853965818882,
+ "learning_rate": 0.0006,
+ "loss": 3.937126636505127,
+ "step": 2820
+ },
+ {
+ "epoch": 39.18165938864629,
+ "grad_norm": 0.016660021618008614,
+ "learning_rate": 0.0006,
+ "loss": 3.8397014141082764,
+ "step": 2821
+ },
+ {
+ "epoch": 39.19563318777293,
+ "grad_norm": 0.017719309777021408,
+ "learning_rate": 0.0006,
+ "loss": 4.003579616546631,
+ "step": 2822
+ },
+ {
+ "epoch": 39.209606986899566,
+ "grad_norm": 0.018234414979815483,
+ "learning_rate": 0.0006,
+ "loss": 4.053859710693359,
+ "step": 2823
+ },
+ {
+ "epoch": 39.223580786026204,
+ "grad_norm": 0.018777024000883102,
+ "learning_rate": 0.0006,
+ "loss": 4.056182861328125,
+ "step": 2824
+ },
+ {
+ "epoch": 39.237554585152836,
+ "grad_norm": 0.017443502321839333,
+ "learning_rate": 0.0006,
+ "loss": 4.025068283081055,
+ "step": 2825
+ },
+ {
+ "epoch": 39.251528384279474,
+ "grad_norm": 0.016500331461429596,
+ "learning_rate": 0.0006,
+ "loss": 4.101631164550781,
+ "step": 2826
+ },
+ {
+ "epoch": 39.26550218340611,
+ "grad_norm": 0.016101544722914696,
+ "learning_rate": 0.0006,
+ "loss": 4.027082443237305,
+ "step": 2827
+ },
+ {
+ "epoch": 39.27947598253275,
+ "grad_norm": 0.015471171587705612,
+ "learning_rate": 0.0006,
+ "loss": 4.022909641265869,
+ "step": 2828
+ },
+ {
+ "epoch": 39.29344978165939,
+ "grad_norm": 0.015779748558998108,
+ "learning_rate": 0.0006,
+ "loss": 4.18796968460083,
+ "step": 2829
+ },
+ {
+ "epoch": 39.30742358078603,
+ "grad_norm": 0.015522679314017296,
+ "learning_rate": 0.0006,
+ "loss": 4.005777359008789,
+ "step": 2830
+ },
+ {
+ "epoch": 39.32139737991266,
+ "grad_norm": 0.015226504765450954,
+ "learning_rate": 0.0006,
+ "loss": 4.108416557312012,
+ "step": 2831
+ },
+ {
+ "epoch": 39.3353711790393,
+ "grad_norm": 0.015095778740942478,
+ "learning_rate": 0.0006,
+ "loss": 4.08029842376709,
+ "step": 2832
+ },
+ {
+ "epoch": 39.34934497816594,
+ "grad_norm": 0.015601031482219696,
+ "learning_rate": 0.0006,
+ "loss": 3.989492416381836,
+ "step": 2833
+ },
+ {
+ "epoch": 39.36331877729258,
+ "grad_norm": 0.014825052581727505,
+ "learning_rate": 0.0006,
+ "loss": 3.998520612716675,
+ "step": 2834
+ },
+ {
+ "epoch": 39.377292576419215,
+ "grad_norm": 0.014821100980043411,
+ "learning_rate": 0.0006,
+ "loss": 4.001777172088623,
+ "step": 2835
+ },
+ {
+ "epoch": 39.391266375545854,
+ "grad_norm": 0.015333754941821098,
+ "learning_rate": 0.0006,
+ "loss": 4.11985969543457,
+ "step": 2836
+ },
+ {
+ "epoch": 39.40524017467249,
+ "grad_norm": 0.015940699726343155,
+ "learning_rate": 0.0006,
+ "loss": 4.025373935699463,
+ "step": 2837
+ },
+ {
+ "epoch": 39.419213973799124,
+ "grad_norm": 0.015096287243068218,
+ "learning_rate": 0.0006,
+ "loss": 4.068551063537598,
+ "step": 2838
+ },
+ {
+ "epoch": 39.43318777292576,
+ "grad_norm": 0.01533548254519701,
+ "learning_rate": 0.0006,
+ "loss": 3.9800801277160645,
+ "step": 2839
+ },
+ {
+ "epoch": 39.4471615720524,
+ "grad_norm": 0.014648732729256153,
+ "learning_rate": 0.0006,
+ "loss": 3.9670190811157227,
+ "step": 2840
+ },
+ {
+ "epoch": 39.46113537117904,
+ "grad_norm": 0.015759671106934547,
+ "learning_rate": 0.0006,
+ "loss": 4.050443649291992,
+ "step": 2841
+ },
+ {
+ "epoch": 39.47510917030568,
+ "grad_norm": 0.016054073348641396,
+ "learning_rate": 0.0006,
+ "loss": 4.22734260559082,
+ "step": 2842
+ },
+ {
+ "epoch": 39.48908296943232,
+ "grad_norm": 0.0160264540463686,
+ "learning_rate": 0.0006,
+ "loss": 4.116174221038818,
+ "step": 2843
+ },
+ {
+ "epoch": 39.50305676855895,
+ "grad_norm": 0.015426610596477985,
+ "learning_rate": 0.0006,
+ "loss": 4.000178337097168,
+ "step": 2844
+ },
+ {
+ "epoch": 39.51703056768559,
+ "grad_norm": 0.01520642638206482,
+ "learning_rate": 0.0006,
+ "loss": 4.218410968780518,
+ "step": 2845
+ },
+ {
+ "epoch": 39.531004366812226,
+ "grad_norm": 0.015305949375033379,
+ "learning_rate": 0.0006,
+ "loss": 4.025891304016113,
+ "step": 2846
+ },
+ {
+ "epoch": 39.544978165938865,
+ "grad_norm": 0.015461564064025879,
+ "learning_rate": 0.0006,
+ "loss": 4.100950717926025,
+ "step": 2847
+ },
+ {
+ "epoch": 39.5589519650655,
+ "grad_norm": 0.01712625101208687,
+ "learning_rate": 0.0006,
+ "loss": 4.0620856285095215,
+ "step": 2848
+ },
+ {
+ "epoch": 39.57292576419214,
+ "grad_norm": 0.017133301123976707,
+ "learning_rate": 0.0006,
+ "loss": 4.097138404846191,
+ "step": 2849
+ },
+ {
+ "epoch": 39.58689956331878,
+ "grad_norm": 0.01683625392615795,
+ "learning_rate": 0.0006,
+ "loss": 4.084364891052246,
+ "step": 2850
+ },
+ {
+ "epoch": 39.60087336244541,
+ "grad_norm": 0.0165710486471653,
+ "learning_rate": 0.0006,
+ "loss": 4.042361259460449,
+ "step": 2851
+ },
+ {
+ "epoch": 39.61484716157205,
+ "grad_norm": 0.01660393364727497,
+ "learning_rate": 0.0006,
+ "loss": 4.1395487785339355,
+ "step": 2852
+ },
+ {
+ "epoch": 39.62882096069869,
+ "grad_norm": 0.017025845125317574,
+ "learning_rate": 0.0006,
+ "loss": 4.1207075119018555,
+ "step": 2853
+ },
+ {
+ "epoch": 39.64279475982533,
+ "grad_norm": 0.01660497486591339,
+ "learning_rate": 0.0006,
+ "loss": 4.103846073150635,
+ "step": 2854
+ },
+ {
+ "epoch": 39.65676855895197,
+ "grad_norm": 0.01607946865260601,
+ "learning_rate": 0.0006,
+ "loss": 4.189016342163086,
+ "step": 2855
+ },
+ {
+ "epoch": 39.670742358078606,
+ "grad_norm": 0.01614505983889103,
+ "learning_rate": 0.0006,
+ "loss": 4.079860687255859,
+ "step": 2856
+ },
+ {
+ "epoch": 39.68471615720524,
+ "grad_norm": 0.017085086554288864,
+ "learning_rate": 0.0006,
+ "loss": 3.9774293899536133,
+ "step": 2857
+ },
+ {
+ "epoch": 39.698689956331876,
+ "grad_norm": 0.01816520094871521,
+ "learning_rate": 0.0006,
+ "loss": 4.113666534423828,
+ "step": 2858
+ },
+ {
+ "epoch": 39.712663755458514,
+ "grad_norm": 0.016597526147961617,
+ "learning_rate": 0.0006,
+ "loss": 4.068242073059082,
+ "step": 2859
+ },
+ {
+ "epoch": 39.72663755458515,
+ "grad_norm": 0.01672235131263733,
+ "learning_rate": 0.0006,
+ "loss": 4.024444580078125,
+ "step": 2860
+ },
+ {
+ "epoch": 39.74061135371179,
+ "grad_norm": 0.020093556493520737,
+ "learning_rate": 0.0006,
+ "loss": 4.064124584197998,
+ "step": 2861
+ },
+ {
+ "epoch": 39.75458515283843,
+ "grad_norm": 0.023141304031014442,
+ "learning_rate": 0.0006,
+ "loss": 4.130439758300781,
+ "step": 2862
+ },
+ {
+ "epoch": 39.76855895196506,
+ "grad_norm": 0.02289372682571411,
+ "learning_rate": 0.0006,
+ "loss": 4.14794921875,
+ "step": 2863
+ },
+ {
+ "epoch": 39.7825327510917,
+ "grad_norm": 0.01919020712375641,
+ "learning_rate": 0.0006,
+ "loss": 4.123961925506592,
+ "step": 2864
+ },
+ {
+ "epoch": 39.79650655021834,
+ "grad_norm": 0.016932329162955284,
+ "learning_rate": 0.0006,
+ "loss": 4.0944132804870605,
+ "step": 2865
+ },
+ {
+ "epoch": 39.81048034934498,
+ "grad_norm": 0.016644535586237907,
+ "learning_rate": 0.0006,
+ "loss": 4.124269962310791,
+ "step": 2866
+ },
+ {
+ "epoch": 39.82445414847162,
+ "grad_norm": 0.016302216798067093,
+ "learning_rate": 0.0006,
+ "loss": 4.0267205238342285,
+ "step": 2867
+ },
+ {
+ "epoch": 39.838427947598255,
+ "grad_norm": 0.01673300378024578,
+ "learning_rate": 0.0006,
+ "loss": 4.10650634765625,
+ "step": 2868
+ },
+ {
+ "epoch": 39.852401746724894,
+ "grad_norm": 0.016581356525421143,
+ "learning_rate": 0.0006,
+ "loss": 4.1547956466674805,
+ "step": 2869
+ },
+ {
+ "epoch": 39.866375545851525,
+ "grad_norm": 0.01691797934472561,
+ "learning_rate": 0.0006,
+ "loss": 4.056936740875244,
+ "step": 2870
+ },
+ {
+ "epoch": 39.880349344978164,
+ "grad_norm": 0.01596726104617119,
+ "learning_rate": 0.0006,
+ "loss": 3.9891414642333984,
+ "step": 2871
+ },
+ {
+ "epoch": 39.8943231441048,
+ "grad_norm": 0.015615091659128666,
+ "learning_rate": 0.0006,
+ "loss": 4.137948036193848,
+ "step": 2872
+ },
+ {
+ "epoch": 39.90829694323144,
+ "grad_norm": 0.015703413635492325,
+ "learning_rate": 0.0006,
+ "loss": 3.9844930171966553,
+ "step": 2873
+ },
+ {
+ "epoch": 39.92227074235808,
+ "grad_norm": 0.015453094616532326,
+ "learning_rate": 0.0006,
+ "loss": 4.098443508148193,
+ "step": 2874
+ },
+ {
+ "epoch": 39.93624454148472,
+ "grad_norm": 0.01449158601462841,
+ "learning_rate": 0.0006,
+ "loss": 4.090852737426758,
+ "step": 2875
+ },
+ {
+ "epoch": 39.95021834061135,
+ "grad_norm": 0.01537515688687563,
+ "learning_rate": 0.0006,
+ "loss": 4.239612579345703,
+ "step": 2876
+ },
+ {
+ "epoch": 39.96419213973799,
+ "grad_norm": 0.015989921987056732,
+ "learning_rate": 0.0006,
+ "loss": 4.142073154449463,
+ "step": 2877
+ },
+ {
+ "epoch": 39.97816593886463,
+ "grad_norm": 0.015985611826181412,
+ "learning_rate": 0.0006,
+ "loss": 4.100076675415039,
+ "step": 2878
+ },
+ {
+ "epoch": 39.992139737991266,
+ "grad_norm": 0.015260833315551281,
+ "learning_rate": 0.0006,
+ "loss": 4.068576335906982,
+ "step": 2879
+ },
+ {
+ "epoch": 40.0,
+ "grad_norm": 0.018047941848635674,
+ "learning_rate": 0.0006,
+ "loss": 4.005878925323486,
+ "step": 2880
+ },
+ {
+ "epoch": 40.0,
+ "eval_loss": 4.622848987579346,
+ "eval_runtime": 56.9618,
+ "eval_samples_per_second": 42.871,
+ "eval_steps_per_second": 1.352,
+ "step": 2880
+ },
+ {
+ "epoch": 40.01397379912664,
+ "grad_norm": 0.015976838767528534,
+ "learning_rate": 0.0006,
+ "loss": 4.035689353942871,
+ "step": 2881
+ },
+ {
+ "epoch": 40.02794759825328,
+ "grad_norm": 0.015176220797002316,
+ "learning_rate": 0.0006,
+ "loss": 3.932197332382202,
+ "step": 2882
+ },
+ {
+ "epoch": 40.041921397379916,
+ "grad_norm": 0.016676228493452072,
+ "learning_rate": 0.0006,
+ "loss": 4.014171123504639,
+ "step": 2883
+ },
+ {
+ "epoch": 40.05589519650655,
+ "grad_norm": 0.01698281429708004,
+ "learning_rate": 0.0006,
+ "loss": 4.117893218994141,
+ "step": 2884
+ },
+ {
+ "epoch": 40.069868995633186,
+ "grad_norm": 0.017872486263513565,
+ "learning_rate": 0.0006,
+ "loss": 4.132119178771973,
+ "step": 2885
+ },
+ {
+ "epoch": 40.083842794759825,
+ "grad_norm": 0.017615554854273796,
+ "learning_rate": 0.0006,
+ "loss": 4.072000980377197,
+ "step": 2886
+ },
+ {
+ "epoch": 40.09781659388646,
+ "grad_norm": 0.017469394952058792,
+ "learning_rate": 0.0006,
+ "loss": 4.017726421356201,
+ "step": 2887
+ },
+ {
+ "epoch": 40.1117903930131,
+ "grad_norm": 0.01954762637615204,
+ "learning_rate": 0.0006,
+ "loss": 4.014443874359131,
+ "step": 2888
+ },
+ {
+ "epoch": 40.12576419213974,
+ "grad_norm": 0.018935853615403175,
+ "learning_rate": 0.0006,
+ "loss": 4.0042948722839355,
+ "step": 2889
+ },
+ {
+ "epoch": 40.13973799126637,
+ "grad_norm": 0.019382692873477936,
+ "learning_rate": 0.0006,
+ "loss": 4.06160831451416,
+ "step": 2890
+ },
+ {
+ "epoch": 40.15371179039301,
+ "grad_norm": 0.019677115604281425,
+ "learning_rate": 0.0006,
+ "loss": 4.13466215133667,
+ "step": 2891
+ },
+ {
+ "epoch": 40.16768558951965,
+ "grad_norm": 0.020365122705698013,
+ "learning_rate": 0.0006,
+ "loss": 4.095000267028809,
+ "step": 2892
+ },
+ {
+ "epoch": 40.18165938864629,
+ "grad_norm": 0.021842138841748238,
+ "learning_rate": 0.0006,
+ "loss": 3.9877853393554688,
+ "step": 2893
+ },
+ {
+ "epoch": 40.19563318777293,
+ "grad_norm": 0.02354881353676319,
+ "learning_rate": 0.0006,
+ "loss": 3.9430112838745117,
+ "step": 2894
+ },
+ {
+ "epoch": 40.209606986899566,
+ "grad_norm": 0.026602735742926598,
+ "learning_rate": 0.0006,
+ "loss": 4.012076377868652,
+ "step": 2895
+ },
+ {
+ "epoch": 40.223580786026204,
+ "grad_norm": 0.023976484313607216,
+ "learning_rate": 0.0006,
+ "loss": 3.9768595695495605,
+ "step": 2896
+ },
+ {
+ "epoch": 40.237554585152836,
+ "grad_norm": 0.02336616814136505,
+ "learning_rate": 0.0006,
+ "loss": 4.1096038818359375,
+ "step": 2897
+ },
+ {
+ "epoch": 40.251528384279474,
+ "grad_norm": 0.02426023595035076,
+ "learning_rate": 0.0006,
+ "loss": 4.157189846038818,
+ "step": 2898
+ },
+ {
+ "epoch": 40.26550218340611,
+ "grad_norm": 0.021572696045041084,
+ "learning_rate": 0.0006,
+ "loss": 3.979175090789795,
+ "step": 2899
+ },
+ {
+ "epoch": 40.27947598253275,
+ "grad_norm": 0.08300432562828064,
+ "learning_rate": 0.0006,
+ "loss": 4.092336654663086,
+ "step": 2900
+ },
+ {
+ "epoch": 40.29344978165939,
+ "grad_norm": 0.7323967814445496,
+ "learning_rate": 0.0006,
+ "loss": 5.337990760803223,
+ "step": 2901
+ },
+ {
+ "epoch": 40.30742358078603,
+ "grad_norm": 0.3363337814807892,
+ "learning_rate": 0.0006,
+ "loss": 6.40043830871582,
+ "step": 2902
+ },
+ {
+ "epoch": 40.32139737991266,
+ "grad_norm": 0.598686933517456,
+ "learning_rate": 0.0006,
+ "loss": 7.004765510559082,
+ "step": 2903
+ },
+ {
+ "epoch": 40.3353711790393,
+ "grad_norm": 0.5998364686965942,
+ "learning_rate": 0.0006,
+ "loss": 8.184747695922852,
+ "step": 2904
+ },
+ {
+ "epoch": 40.34934497816594,
+ "grad_norm": 1.408208966255188,
+ "learning_rate": 0.0006,
+ "loss": 8.88978385925293,
+ "step": 2905
+ },
+ {
+ "epoch": 40.36331877729258,
+ "grad_norm": 0.7829785943031311,
+ "learning_rate": 0.0006,
+ "loss": 11.13601303100586,
+ "step": 2906
+ },
+ {
+ "epoch": 40.377292576419215,
+ "grad_norm": 0.7769215106964111,
+ "learning_rate": 0.0006,
+ "loss": 9.90416145324707,
+ "step": 2907
+ },
+ {
+ "epoch": 40.391266375545854,
+ "grad_norm": 0.46508079767227173,
+ "learning_rate": 0.0006,
+ "loss": 9.647209167480469,
+ "step": 2908
+ },
+ {
+ "epoch": 40.40524017467249,
+ "grad_norm": 1.0228968858718872,
+ "learning_rate": 0.0006,
+ "loss": 11.685434341430664,
+ "step": 2909
+ },
+ {
+ "epoch": 40.419213973799124,
+ "grad_norm": 0.20553778111934662,
+ "learning_rate": 0.0006,
+ "loss": 8.621708869934082,
+ "step": 2910
+ },
+ {
+ "epoch": 40.43318777292576,
+ "grad_norm": 0.18466424942016602,
+ "learning_rate": 0.0006,
+ "loss": 8.590787887573242,
+ "step": 2911
+ },
+ {
+ "epoch": 40.4471615720524,
+ "grad_norm": 0.10458594560623169,
+ "learning_rate": 0.0006,
+ "loss": 8.324458122253418,
+ "step": 2912
+ },
+ {
+ "epoch": 40.46113537117904,
+ "grad_norm": 0.1096842959523201,
+ "learning_rate": 0.0006,
+ "loss": 8.300851821899414,
+ "step": 2913
+ },
+ {
+ "epoch": 40.47510917030568,
+ "grad_norm": 0.10501226782798767,
+ "learning_rate": 0.0006,
+ "loss": 8.21009635925293,
+ "step": 2914
+ },
+ {
+ "epoch": 40.48908296943232,
+ "grad_norm": 0.13704165816307068,
+ "learning_rate": 0.0006,
+ "loss": 8.141355514526367,
+ "step": 2915
+ },
+ {
+ "epoch": 40.50305676855895,
+ "grad_norm": 0.12129108607769012,
+ "learning_rate": 0.0006,
+ "loss": 8.074612617492676,
+ "step": 2916
+ },
+ {
+ "epoch": 40.51703056768559,
+ "grad_norm": 0.10028897225856781,
+ "learning_rate": 0.0006,
+ "loss": 8.00384521484375,
+ "step": 2917
+ },
+ {
+ "epoch": 40.531004366812226,
+ "grad_norm": 0.11764246970415115,
+ "learning_rate": 0.0006,
+ "loss": 8.078598022460938,
+ "step": 2918
+ },
+ {
+ "epoch": 40.544978165938865,
+ "grad_norm": 0.10081959515810013,
+ "learning_rate": 0.0006,
+ "loss": 8.008938789367676,
+ "step": 2919
+ },
+ {
+ "epoch": 40.5589519650655,
+ "grad_norm": 0.1003170907497406,
+ "learning_rate": 0.0006,
+ "loss": 7.962940216064453,
+ "step": 2920
+ },
+ {
+ "epoch": 40.57292576419214,
+ "grad_norm": 0.08763492107391357,
+ "learning_rate": 0.0006,
+ "loss": 7.909419536590576,
+ "step": 2921
+ },
+ {
+ "epoch": 40.58689956331878,
+ "grad_norm": 0.0636807233095169,
+ "learning_rate": 0.0006,
+ "loss": 7.878905773162842,
+ "step": 2922
+ },
+ {
+ "epoch": 40.60087336244541,
+ "grad_norm": 0.06005936488509178,
+ "learning_rate": 0.0006,
+ "loss": 7.850980758666992,
+ "step": 2923
+ },
+ {
+ "epoch": 40.61484716157205,
+ "grad_norm": 0.06415296345949173,
+ "learning_rate": 0.0006,
+ "loss": 7.830353260040283,
+ "step": 2924
+ },
+ {
+ "epoch": 40.62882096069869,
+ "grad_norm": 0.0732007846236229,
+ "learning_rate": 0.0006,
+ "loss": 7.830746650695801,
+ "step": 2925
+ },
+ {
+ "epoch": 40.64279475982533,
+ "grad_norm": 0.06628110259771347,
+ "learning_rate": 0.0006,
+ "loss": 7.82661771774292,
+ "step": 2926
+ },
+ {
+ "epoch": 40.65676855895197,
+ "grad_norm": 0.05544218048453331,
+ "learning_rate": 0.0006,
+ "loss": 7.814305782318115,
+ "step": 2927
+ },
+ {
+ "epoch": 40.670742358078606,
+ "grad_norm": 0.050327103585004807,
+ "learning_rate": 0.0006,
+ "loss": 7.764055252075195,
+ "step": 2928
+ },
+ {
+ "epoch": 40.68471615720524,
+ "grad_norm": 0.04918425530195236,
+ "learning_rate": 0.0006,
+ "loss": 7.744625091552734,
+ "step": 2929
+ },
+ {
+ "epoch": 40.698689956331876,
+ "grad_norm": 0.042862266302108765,
+ "learning_rate": 0.0006,
+ "loss": 7.764426231384277,
+ "step": 2930
+ },
+ {
+ "epoch": 40.712663755458514,
+ "grad_norm": 0.04338089004158974,
+ "learning_rate": 0.0006,
+ "loss": 7.712530136108398,
+ "step": 2931
+ },
+ {
+ "epoch": 40.72663755458515,
+ "grad_norm": 0.04127737879753113,
+ "learning_rate": 0.0006,
+ "loss": 7.742180347442627,
+ "step": 2932
+ },
+ {
+ "epoch": 40.74061135371179,
+ "grad_norm": 0.03358631581068039,
+ "learning_rate": 0.0006,
+ "loss": 7.714834213256836,
+ "step": 2933
+ },
+ {
+ "epoch": 40.75458515283843,
+ "grad_norm": 0.02845790982246399,
+ "learning_rate": 0.0006,
+ "loss": 7.7021026611328125,
+ "step": 2934
+ },
+ {
+ "epoch": 40.76855895196506,
+ "grad_norm": 0.03012419492006302,
+ "learning_rate": 0.0006,
+ "loss": 7.712377548217773,
+ "step": 2935
+ },
+ {
+ "epoch": 40.7825327510917,
+ "grad_norm": 0.032981179654598236,
+ "learning_rate": 0.0006,
+ "loss": 7.666385650634766,
+ "step": 2936
+ },
+ {
+ "epoch": 40.79650655021834,
+ "grad_norm": 0.03357074037194252,
+ "learning_rate": 0.0006,
+ "loss": 7.649097919464111,
+ "step": 2937
+ },
+ {
+ "epoch": 40.81048034934498,
+ "grad_norm": 0.030914708971977234,
+ "learning_rate": 0.0006,
+ "loss": 7.649196624755859,
+ "step": 2938
+ },
+ {
+ "epoch": 40.82445414847162,
+ "grad_norm": 0.026296131312847137,
+ "learning_rate": 0.0006,
+ "loss": 7.629964828491211,
+ "step": 2939
+ },
+ {
+ "epoch": 40.838427947598255,
+ "grad_norm": 0.026677245274186134,
+ "learning_rate": 0.0006,
+ "loss": 7.599973678588867,
+ "step": 2940
+ },
+ {
+ "epoch": 40.852401746724894,
+ "grad_norm": 0.023208219558000565,
+ "learning_rate": 0.0006,
+ "loss": 7.601624488830566,
+ "step": 2941
+ },
+ {
+ "epoch": 40.866375545851525,
+ "grad_norm": 0.020741475746035576,
+ "learning_rate": 0.0006,
+ "loss": 7.603902339935303,
+ "step": 2942
+ },
+ {
+ "epoch": 40.880349344978164,
+ "grad_norm": 0.019221751019358635,
+ "learning_rate": 0.0006,
+ "loss": 7.5535888671875,
+ "step": 2943
+ },
+ {
+ "epoch": 40.8943231441048,
+ "grad_norm": 0.019491951912641525,
+ "learning_rate": 0.0006,
+ "loss": 7.54597282409668,
+ "step": 2944
+ },
+ {
+ "epoch": 40.90829694323144,
+ "grad_norm": 0.019702572375535965,
+ "learning_rate": 0.0006,
+ "loss": 7.537567138671875,
+ "step": 2945
+ },
+ {
+ "epoch": 40.92227074235808,
+ "grad_norm": 0.01768800988793373,
+ "learning_rate": 0.0006,
+ "loss": 7.512800216674805,
+ "step": 2946
+ },
+ {
+ "epoch": 40.93624454148472,
+ "grad_norm": 0.01926138624548912,
+ "learning_rate": 0.0006,
+ "loss": 7.509932518005371,
+ "step": 2947
+ },
+ {
+ "epoch": 40.95021834061135,
+ "grad_norm": 0.019030006602406502,
+ "learning_rate": 0.0006,
+ "loss": 7.497983932495117,
+ "step": 2948
+ },
+ {
+ "epoch": 40.96419213973799,
+ "grad_norm": 0.017313841730356216,
+ "learning_rate": 0.0006,
+ "loss": 7.463518142700195,
+ "step": 2949
+ },
+ {
+ "epoch": 40.97816593886463,
+ "grad_norm": 0.01310762483626604,
+ "learning_rate": 0.0006,
+ "loss": 7.474668502807617,
+ "step": 2950
+ },
+ {
+ "epoch": 40.992139737991266,
+ "grad_norm": 0.013008282519876957,
+ "learning_rate": 0.0006,
+ "loss": 7.45982551574707,
+ "step": 2951
+ },
+ {
+ "epoch": 41.0,
+ "grad_norm": 0.01441363524645567,
+ "learning_rate": 0.0006,
+ "loss": 7.446529388427734,
+ "step": 2952
+ },
+ {
+ "epoch": 41.0,
+ "eval_loss": 7.486120223999023,
+ "eval_runtime": 70.4986,
+ "eval_samples_per_second": 34.639,
+ "eval_steps_per_second": 1.092,
+ "step": 2952
+ },
+ {
+ "epoch": 41.01397379912664,
+ "grad_norm": 0.014061855152249336,
+ "learning_rate": 0.0006,
+ "loss": 7.4692182540893555,
+ "step": 2953
+ },
+ {
+ "epoch": 41.02794759825328,
+ "grad_norm": 0.01587500050663948,
+ "learning_rate": 0.0006,
+ "loss": 7.4182024002075195,
+ "step": 2954
+ },
+ {
+ "epoch": 41.041921397379916,
+ "grad_norm": 0.014783984050154686,
+ "learning_rate": 0.0006,
+ "loss": 7.466778755187988,
+ "step": 2955
+ },
+ {
+ "epoch": 41.05589519650655,
+ "grad_norm": 0.011560854502022266,
+ "learning_rate": 0.0006,
+ "loss": 7.438002586364746,
+ "step": 2956
+ },
+ {
+ "epoch": 41.069868995633186,
+ "grad_norm": 0.00970957800745964,
+ "learning_rate": 0.0006,
+ "loss": 7.442694187164307,
+ "step": 2957
+ },
+ {
+ "epoch": 41.083842794759825,
+ "grad_norm": 0.009633459150791168,
+ "learning_rate": 0.0006,
+ "loss": 7.458884239196777,
+ "step": 2958
+ },
+ {
+ "epoch": 41.09781659388646,
+ "grad_norm": 0.010664350353181362,
+ "learning_rate": 0.0006,
+ "loss": 7.48378324508667,
+ "step": 2959
+ },
+ {
+ "epoch": 41.1117903930131,
+ "grad_norm": 0.011440591886639595,
+ "learning_rate": 0.0006,
+ "loss": 7.474484920501709,
+ "step": 2960
+ },
+ {
+ "epoch": 41.12576419213974,
+ "grad_norm": 0.01150908600538969,
+ "learning_rate": 0.0006,
+ "loss": 7.417902946472168,
+ "step": 2961
+ },
+ {
+ "epoch": 41.13973799126637,
+ "grad_norm": 0.010375697165727615,
+ "learning_rate": 0.0006,
+ "loss": 7.403279781341553,
+ "step": 2962
+ },
+ {
+ "epoch": 41.15371179039301,
+ "grad_norm": 0.009075703099370003,
+ "learning_rate": 0.0006,
+ "loss": 7.447347164154053,
+ "step": 2963
+ },
+ {
+ "epoch": 41.16768558951965,
+ "grad_norm": 0.008892177604138851,
+ "learning_rate": 0.0006,
+ "loss": 7.402681350708008,
+ "step": 2964
+ },
+ {
+ "epoch": 41.18165938864629,
+ "grad_norm": 0.009492084383964539,
+ "learning_rate": 0.0006,
+ "loss": 7.3871355056762695,
+ "step": 2965
+ },
+ {
+ "epoch": 41.19563318777293,
+ "grad_norm": 0.009139460511505604,
+ "learning_rate": 0.0006,
+ "loss": 7.419617176055908,
+ "step": 2966
+ },
+ {
+ "epoch": 41.209606986899566,
+ "grad_norm": 0.007959138602018356,
+ "learning_rate": 0.0006,
+ "loss": 7.374949932098389,
+ "step": 2967
+ },
+ {
+ "epoch": 41.223580786026204,
+ "grad_norm": 0.007571548223495483,
+ "learning_rate": 0.0006,
+ "loss": 7.4205241203308105,
+ "step": 2968
+ },
+ {
+ "epoch": 41.237554585152836,
+ "grad_norm": 0.007885400205850601,
+ "learning_rate": 0.0006,
+ "loss": 7.416543006896973,
+ "step": 2969
+ },
+ {
+ "epoch": 41.251528384279474,
+ "grad_norm": 0.008650158531963825,
+ "learning_rate": 0.0006,
+ "loss": 7.357326984405518,
+ "step": 2970
+ },
+ {
+ "epoch": 41.26550218340611,
+ "grad_norm": 0.007483262103050947,
+ "learning_rate": 0.0006,
+ "loss": 7.410891056060791,
+ "step": 2971
+ },
+ {
+ "epoch": 41.27947598253275,
+ "grad_norm": 0.00869277399033308,
+ "learning_rate": 0.0006,
+ "loss": 7.392127990722656,
+ "step": 2972
+ },
+ {
+ "epoch": 41.29344978165939,
+ "grad_norm": 0.00885078776627779,
+ "learning_rate": 0.0006,
+ "loss": 7.405449867248535,
+ "step": 2973
+ },
+ {
+ "epoch": 41.30742358078603,
+ "grad_norm": 0.00958226341754198,
+ "learning_rate": 0.0006,
+ "loss": 7.399868488311768,
+ "step": 2974
+ },
+ {
+ "epoch": 41.32139737991266,
+ "grad_norm": 0.006991599686443806,
+ "learning_rate": 0.0006,
+ "loss": 7.35369873046875,
+ "step": 2975
+ },
+ {
+ "epoch": 41.3353711790393,
+ "grad_norm": 0.007415142375975847,
+ "learning_rate": 0.0006,
+ "loss": 7.358353137969971,
+ "step": 2976
+ },
+ {
+ "epoch": 41.34934497816594,
+ "grad_norm": 0.007024371065199375,
+ "learning_rate": 0.0006,
+ "loss": 7.4070892333984375,
+ "step": 2977
+ },
+ {
+ "epoch": 41.36331877729258,
+ "grad_norm": 0.007234332151710987,
+ "learning_rate": 0.0006,
+ "loss": 7.379120349884033,
+ "step": 2978
+ },
+ {
+ "epoch": 41.377292576419215,
+ "grad_norm": 0.00798892229795456,
+ "learning_rate": 0.0006,
+ "loss": 7.365455627441406,
+ "step": 2979
+ },
+ {
+ "epoch": 41.391266375545854,
+ "grad_norm": 0.007303398102521896,
+ "learning_rate": 0.0006,
+ "loss": 7.341798782348633,
+ "step": 2980
+ },
+ {
+ "epoch": 41.40524017467249,
+ "grad_norm": 0.007119669578969479,
+ "learning_rate": 0.0006,
+ "loss": 7.363280296325684,
+ "step": 2981
+ },
+ {
+ "epoch": 41.419213973799124,
+ "grad_norm": 0.007457619532942772,
+ "learning_rate": 0.0006,
+ "loss": 7.39388370513916,
+ "step": 2982
+ },
+ {
+ "epoch": 41.43318777292576,
+ "grad_norm": 0.007905688136816025,
+ "learning_rate": 0.0006,
+ "loss": 7.380652904510498,
+ "step": 2983
+ },
+ {
+ "epoch": 41.4471615720524,
+ "grad_norm": 0.006212855689227581,
+ "learning_rate": 0.0006,
+ "loss": 7.398769378662109,
+ "step": 2984
+ },
+ {
+ "epoch": 41.46113537117904,
+ "grad_norm": 0.0070641543716192245,
+ "learning_rate": 0.0006,
+ "loss": 7.3503546714782715,
+ "step": 2985
+ },
+ {
+ "epoch": 41.47510917030568,
+ "grad_norm": 0.00787805300205946,
+ "learning_rate": 0.0006,
+ "loss": 7.347678184509277,
+ "step": 2986
+ },
+ {
+ "epoch": 41.48908296943232,
+ "grad_norm": 0.006760426796972752,
+ "learning_rate": 0.0006,
+ "loss": 7.380006790161133,
+ "step": 2987
+ },
+ {
+ "epoch": 41.50305676855895,
+ "grad_norm": 0.00772385997697711,
+ "learning_rate": 0.0006,
+ "loss": 7.363680362701416,
+ "step": 2988
+ },
+ {
+ "epoch": 41.51703056768559,
+ "grad_norm": 0.007373564876616001,
+ "learning_rate": 0.0006,
+ "loss": 7.333014488220215,
+ "step": 2989
+ },
+ {
+ "epoch": 41.531004366812226,
+ "grad_norm": 0.006717653013765812,
+ "learning_rate": 0.0006,
+ "loss": 7.360076427459717,
+ "step": 2990
+ },
+ {
+ "epoch": 41.544978165938865,
+ "grad_norm": 0.007191413082182407,
+ "learning_rate": 0.0006,
+ "loss": 7.397030830383301,
+ "step": 2991
+ },
+ {
+ "epoch": 41.5589519650655,
+ "grad_norm": 0.007831960916519165,
+ "learning_rate": 0.0006,
+ "loss": 7.369959831237793,
+ "step": 2992
+ },
+ {
+ "epoch": 41.57292576419214,
+ "grad_norm": 0.006947481073439121,
+ "learning_rate": 0.0006,
+ "loss": 7.333254337310791,
+ "step": 2993
+ },
+ {
+ "epoch": 41.58689956331878,
+ "grad_norm": 0.006243606563657522,
+ "learning_rate": 0.0006,
+ "loss": 7.327378273010254,
+ "step": 2994
+ },
+ {
+ "epoch": 41.60087336244541,
+ "grad_norm": 0.006298339460045099,
+ "learning_rate": 0.0006,
+ "loss": 7.323482036590576,
+ "step": 2995
+ },
+ {
+ "epoch": 41.61484716157205,
+ "grad_norm": 0.007587412837892771,
+ "learning_rate": 0.0006,
+ "loss": 7.313789367675781,
+ "step": 2996
+ },
+ {
+ "epoch": 41.62882096069869,
+ "grad_norm": 0.007140746805816889,
+ "learning_rate": 0.0006,
+ "loss": 7.356067657470703,
+ "step": 2997
+ },
+ {
+ "epoch": 41.64279475982533,
+ "grad_norm": 0.008315625600516796,
+ "learning_rate": 0.0006,
+ "loss": 7.3339643478393555,
+ "step": 2998
+ },
+ {
+ "epoch": 41.65676855895197,
+ "grad_norm": 0.006547102238982916,
+ "learning_rate": 0.0006,
+ "loss": 7.342191696166992,
+ "step": 2999
+ },
+ {
+ "epoch": 41.670742358078606,
+ "grad_norm": 0.006068122107535601,
+ "learning_rate": 0.0006,
+ "loss": 7.321015357971191,
+ "step": 3000
+ },
+ {
+ "epoch": 41.68471615720524,
+ "grad_norm": 0.007316991221159697,
+ "learning_rate": 0.0006,
+ "loss": 7.346490859985352,
+ "step": 3001
+ },
+ {
+ "epoch": 41.698689956331876,
+ "grad_norm": 0.0067388564348220825,
+ "learning_rate": 0.0006,
+ "loss": 7.340723991394043,
+ "step": 3002
+ },
+ {
+ "epoch": 41.712663755458514,
+ "grad_norm": 0.006839963141828775,
+ "learning_rate": 0.0006,
+ "loss": 7.329365253448486,
+ "step": 3003
+ },
+ {
+ "epoch": 41.72663755458515,
+ "grad_norm": 0.00805041566491127,
+ "learning_rate": 0.0006,
+ "loss": 7.298934459686279,
+ "step": 3004
+ },
+ {
+ "epoch": 41.74061135371179,
+ "grad_norm": 0.007197503466159105,
+ "learning_rate": 0.0006,
+ "loss": 7.331496238708496,
+ "step": 3005
+ },
+ {
+ "epoch": 41.75458515283843,
+ "grad_norm": 0.007220805156975985,
+ "learning_rate": 0.0006,
+ "loss": 7.317695140838623,
+ "step": 3006
+ },
+ {
+ "epoch": 41.76855895196506,
+ "grad_norm": 0.00874276738613844,
+ "learning_rate": 0.0006,
+ "loss": 7.272614479064941,
+ "step": 3007
+ },
+ {
+ "epoch": 41.7825327510917,
+ "grad_norm": 0.008386565372347832,
+ "learning_rate": 0.0006,
+ "loss": 7.295244216918945,
+ "step": 3008
+ },
+ {
+ "epoch": 41.79650655021834,
+ "grad_norm": 0.0077730282209813595,
+ "learning_rate": 0.0006,
+ "loss": 7.3395490646362305,
+ "step": 3009
+ },
+ {
+ "epoch": 41.81048034934498,
+ "grad_norm": 0.008441480807960033,
+ "learning_rate": 0.0006,
+ "loss": 7.319921493530273,
+ "step": 3010
+ },
+ {
+ "epoch": 41.82445414847162,
+ "grad_norm": 0.007398269604891539,
+ "learning_rate": 0.0006,
+ "loss": 7.3029465675354,
+ "step": 3011
+ },
+ {
+ "epoch": 41.838427947598255,
+ "grad_norm": 0.006878677289932966,
+ "learning_rate": 0.0006,
+ "loss": 7.318467140197754,
+ "step": 3012
+ },
+ {
+ "epoch": 41.852401746724894,
+ "grad_norm": 0.007616598624736071,
+ "learning_rate": 0.0006,
+ "loss": 7.3112077713012695,
+ "step": 3013
+ },
+ {
+ "epoch": 41.866375545851525,
+ "grad_norm": 0.011629699729382992,
+ "learning_rate": 0.0006,
+ "loss": 7.32523250579834,
+ "step": 3014
+ },
+ {
+ "epoch": 41.880349344978164,
+ "grad_norm": 0.011011249385774136,
+ "learning_rate": 0.0006,
+ "loss": 7.287247657775879,
+ "step": 3015
+ },
+ {
+ "epoch": 41.8943231441048,
+ "grad_norm": 0.010208295658230782,
+ "learning_rate": 0.0006,
+ "loss": 7.292117118835449,
+ "step": 3016
+ },
+ {
+ "epoch": 41.90829694323144,
+ "grad_norm": 0.008651095442473888,
+ "learning_rate": 0.0006,
+ "loss": 7.302201271057129,
+ "step": 3017
+ },
+ {
+ "epoch": 41.92227074235808,
+ "grad_norm": 0.006692920345813036,
+ "learning_rate": 0.0006,
+ "loss": 7.319568634033203,
+ "step": 3018
+ },
+ {
+ "epoch": 41.93624454148472,
+ "grad_norm": 0.010033348575234413,
+ "learning_rate": 0.0006,
+ "loss": 7.2738566398620605,
+ "step": 3019
+ },
+ {
+ "epoch": 41.95021834061135,
+ "grad_norm": 0.011790832504630089,
+ "learning_rate": 0.0006,
+ "loss": 7.298540115356445,
+ "step": 3020
+ },
+ {
+ "epoch": 41.96419213973799,
+ "grad_norm": 0.018622662872076035,
+ "learning_rate": 0.0006,
+ "loss": 7.267182350158691,
+ "step": 3021
+ },
+ {
+ "epoch": 41.97816593886463,
+ "grad_norm": 0.02520257607102394,
+ "learning_rate": 0.0006,
+ "loss": 7.261566162109375,
+ "step": 3022
+ },
+ {
+ "epoch": 41.992139737991266,
+ "grad_norm": 0.036332644522190094,
+ "learning_rate": 0.0006,
+ "loss": 7.270719051361084,
+ "step": 3023
+ },
+ {
+ "epoch": 42.0,
+ "grad_norm": 0.04717305675148964,
+ "learning_rate": 0.0006,
+ "loss": 7.293332099914551,
+ "step": 3024
+ },
+ {
+ "epoch": 42.0,
+ "eval_loss": 7.354280471801758,
+ "eval_runtime": 56.4353,
+ "eval_samples_per_second": 43.271,
+ "eval_steps_per_second": 1.364,
+ "step": 3024
+ },
+ {
+ "epoch": 42.01397379912664,
+ "grad_norm": 0.06661626696586609,
+ "learning_rate": 0.0006,
+ "loss": 7.338189601898193,
+ "step": 3025
+ },
+ {
+ "epoch": 42.02794759825328,
+ "grad_norm": 0.05452054366469383,
+ "learning_rate": 0.0006,
+ "loss": 7.316764831542969,
+ "step": 3026
+ },
+ {
+ "epoch": 42.041921397379916,
+ "grad_norm": 0.025638775900006294,
+ "learning_rate": 0.0006,
+ "loss": 7.34204626083374,
+ "step": 3027
+ },
+ {
+ "epoch": 42.05589519650655,
+ "grad_norm": 0.022398799657821655,
+ "learning_rate": 0.0006,
+ "loss": 7.262149333953857,
+ "step": 3028
+ },
+ {
+ "epoch": 42.069868995633186,
+ "grad_norm": 0.03713499382138252,
+ "learning_rate": 0.0006,
+ "loss": 7.27695369720459,
+ "step": 3029
+ },
+ {
+ "epoch": 42.083842794759825,
+ "grad_norm": 0.014949247241020203,
+ "learning_rate": 0.0006,
+ "loss": 7.296508312225342,
+ "step": 3030
+ },
+ {
+ "epoch": 42.09781659388646,
+ "grad_norm": 0.02703995071351528,
+ "learning_rate": 0.0006,
+ "loss": 7.258018493652344,
+ "step": 3031
+ },
+ {
+ "epoch": 42.1117903930131,
+ "grad_norm": 0.020352816209197044,
+ "learning_rate": 0.0006,
+ "loss": 7.241450309753418,
+ "step": 3032
+ },
+ {
+ "epoch": 42.12576419213974,
+ "grad_norm": 0.014578479342162609,
+ "learning_rate": 0.0006,
+ "loss": 7.257323265075684,
+ "step": 3033
+ },
+ {
+ "epoch": 42.13973799126637,
+ "grad_norm": 0.026053044945001602,
+ "learning_rate": 0.0006,
+ "loss": 7.254680633544922,
+ "step": 3034
+ },
+ {
+ "epoch": 42.15371179039301,
+ "grad_norm": 0.015120275318622589,
+ "learning_rate": 0.0006,
+ "loss": 7.261074066162109,
+ "step": 3035
+ },
+ {
+ "epoch": 42.16768558951965,
+ "grad_norm": 0.016822177916765213,
+ "learning_rate": 0.0006,
+ "loss": 7.279064178466797,
+ "step": 3036
+ },
+ {
+ "epoch": 42.18165938864629,
+ "grad_norm": 0.016367588192224503,
+ "learning_rate": 0.0006,
+ "loss": 7.251429557800293,
+ "step": 3037
+ },
+ {
+ "epoch": 42.19563318777293,
+ "grad_norm": 0.014350418001413345,
+ "learning_rate": 0.0006,
+ "loss": 7.259264945983887,
+ "step": 3038
+ },
+ {
+ "epoch": 42.209606986899566,
+ "grad_norm": 0.017078228294849396,
+ "learning_rate": 0.0006,
+ "loss": 7.227105140686035,
+ "step": 3039
+ },
+ {
+ "epoch": 42.223580786026204,
+ "grad_norm": 0.011759317480027676,
+ "learning_rate": 0.0006,
+ "loss": 7.275888442993164,
+ "step": 3040
+ },
+ {
+ "epoch": 42.237554585152836,
+ "grad_norm": 0.013019340112805367,
+ "learning_rate": 0.0006,
+ "loss": 7.2517008781433105,
+ "step": 3041
+ },
+ {
+ "epoch": 42.251528384279474,
+ "grad_norm": 0.01178441196680069,
+ "learning_rate": 0.0006,
+ "loss": 7.23576545715332,
+ "step": 3042
+ },
+ {
+ "epoch": 42.26550218340611,
+ "grad_norm": 0.014518891461193562,
+ "learning_rate": 0.0006,
+ "loss": 7.253895282745361,
+ "step": 3043
+ },
+ {
+ "epoch": 42.27947598253275,
+ "grad_norm": 0.018159950152039528,
+ "learning_rate": 0.0006,
+ "loss": 7.278058052062988,
+ "step": 3044
+ },
+ {
+ "epoch": 42.29344978165939,
+ "grad_norm": 0.007587796077132225,
+ "learning_rate": 0.0006,
+ "loss": 7.278578758239746,
+ "step": 3045
+ },
+ {
+ "epoch": 42.30742358078603,
+ "grad_norm": 0.015994979068636894,
+ "learning_rate": 0.0006,
+ "loss": 7.239841461181641,
+ "step": 3046
+ },
+ {
+ "epoch": 42.32139737991266,
+ "grad_norm": 0.017830168828368187,
+ "learning_rate": 0.0006,
+ "loss": 7.2443976402282715,
+ "step": 3047
+ },
+ {
+ "epoch": 42.3353711790393,
+ "grad_norm": 0.008738759905099869,
+ "learning_rate": 0.0006,
+ "loss": 7.232264518737793,
+ "step": 3048
+ },
+ {
+ "epoch": 42.34934497816594,
+ "grad_norm": 0.0175449438393116,
+ "learning_rate": 0.0006,
+ "loss": 7.228000640869141,
+ "step": 3049
+ },
+ {
+ "epoch": 42.36331877729258,
+ "grad_norm": 0.019611820578575134,
+ "learning_rate": 0.0006,
+ "loss": 7.234999179840088,
+ "step": 3050
+ },
+ {
+ "epoch": 42.377292576419215,
+ "grad_norm": 0.015302863903343678,
+ "learning_rate": 0.0006,
+ "loss": 7.242547988891602,
+ "step": 3051
+ },
+ {
+ "epoch": 42.391266375545854,
+ "grad_norm": 0.025919731706380844,
+ "learning_rate": 0.0006,
+ "loss": 7.216652870178223,
+ "step": 3052
+ },
+ {
+ "epoch": 42.40524017467249,
+ "grad_norm": 0.034587468951940536,
+ "learning_rate": 0.0006,
+ "loss": 7.245336532592773,
+ "step": 3053
+ },
+ {
+ "epoch": 42.419213973799124,
+ "grad_norm": 0.019198840484023094,
+ "learning_rate": 0.0006,
+ "loss": 7.256758689880371,
+ "step": 3054
+ },
+ {
+ "epoch": 42.43318777292576,
+ "grad_norm": 0.012316389009356499,
+ "learning_rate": 0.0006,
+ "loss": 7.209102630615234,
+ "step": 3055
+ },
+ {
+ "epoch": 42.4471615720524,
+ "grad_norm": 0.023512162268161774,
+ "learning_rate": 0.0006,
+ "loss": 7.250742435455322,
+ "step": 3056
+ },
+ {
+ "epoch": 42.46113537117904,
+ "grad_norm": 0.025457851588726044,
+ "learning_rate": 0.0006,
+ "loss": 7.263757705688477,
+ "step": 3057
+ },
+ {
+ "epoch": 42.47510917030568,
+ "grad_norm": 0.02192617766559124,
+ "learning_rate": 0.0006,
+ "loss": 7.246977806091309,
+ "step": 3058
+ },
+ {
+ "epoch": 42.48908296943232,
+ "grad_norm": 0.02541281282901764,
+ "learning_rate": 0.0006,
+ "loss": 7.232361316680908,
+ "step": 3059
+ },
+ {
+ "epoch": 42.50305676855895,
+ "grad_norm": 0.036353763192892075,
+ "learning_rate": 0.0006,
+ "loss": 7.242556571960449,
+ "step": 3060
+ },
+ {
+ "epoch": 42.51703056768559,
+ "grad_norm": 0.04555872827768326,
+ "learning_rate": 0.0006,
+ "loss": 7.228940963745117,
+ "step": 3061
+ },
+ {
+ "epoch": 42.531004366812226,
+ "grad_norm": 0.04563378915190697,
+ "learning_rate": 0.0006,
+ "loss": 7.235276222229004,
+ "step": 3062
+ },
+ {
+ "epoch": 42.544978165938865,
+ "grad_norm": 0.04275760427117348,
+ "learning_rate": 0.0006,
+ "loss": 7.238895416259766,
+ "step": 3063
+ },
+ {
+ "epoch": 42.5589519650655,
+ "grad_norm": 0.04535885155200958,
+ "learning_rate": 0.0006,
+ "loss": 7.205336570739746,
+ "step": 3064
+ },
+ {
+ "epoch": 42.57292576419214,
+ "grad_norm": 0.015534617006778717,
+ "learning_rate": 0.0006,
+ "loss": 7.238393783569336,
+ "step": 3065
+ },
+ {
+ "epoch": 42.58689956331878,
+ "grad_norm": 0.03286299481987953,
+ "learning_rate": 0.0006,
+ "loss": 7.216453552246094,
+ "step": 3066
+ },
+ {
+ "epoch": 42.60087336244541,
+ "grad_norm": 0.025453168898820877,
+ "learning_rate": 0.0006,
+ "loss": 7.203750133514404,
+ "step": 3067
+ },
+ {
+ "epoch": 42.61484716157205,
+ "grad_norm": 0.021501649171113968,
+ "learning_rate": 0.0006,
+ "loss": 7.241621971130371,
+ "step": 3068
+ },
+ {
+ "epoch": 42.62882096069869,
+ "grad_norm": 0.02743634767830372,
+ "learning_rate": 0.0006,
+ "loss": 7.22076940536499,
+ "step": 3069
+ },
+ {
+ "epoch": 42.64279475982533,
+ "grad_norm": 0.020890483632683754,
+ "learning_rate": 0.0006,
+ "loss": 7.218412399291992,
+ "step": 3070
+ },
+ {
+ "epoch": 42.65676855895197,
+ "grad_norm": 0.022499850019812584,
+ "learning_rate": 0.0006,
+ "loss": 7.217696189880371,
+ "step": 3071
+ },
+ {
+ "epoch": 42.670742358078606,
+ "grad_norm": 0.01653696969151497,
+ "learning_rate": 0.0006,
+ "loss": 7.156417369842529,
+ "step": 3072
+ },
+ {
+ "epoch": 42.68471615720524,
+ "grad_norm": 0.01834225095808506,
+ "learning_rate": 0.0006,
+ "loss": 7.211247444152832,
+ "step": 3073
+ },
+ {
+ "epoch": 42.698689956331876,
+ "grad_norm": 0.020137229934334755,
+ "learning_rate": 0.0006,
+ "loss": 7.232499599456787,
+ "step": 3074
+ },
+ {
+ "epoch": 42.712663755458514,
+ "grad_norm": 0.012671459466218948,
+ "learning_rate": 0.0006,
+ "loss": 7.172441482543945,
+ "step": 3075
+ },
+ {
+ "epoch": 42.72663755458515,
+ "grad_norm": 0.016803130507469177,
+ "learning_rate": 0.0006,
+ "loss": 7.152103424072266,
+ "step": 3076
+ },
+ {
+ "epoch": 42.74061135371179,
+ "grad_norm": 0.01623712107539177,
+ "learning_rate": 0.0006,
+ "loss": 7.187032699584961,
+ "step": 3077
+ },
+ {
+ "epoch": 42.75458515283843,
+ "grad_norm": 0.015938356518745422,
+ "learning_rate": 0.0006,
+ "loss": 7.171140670776367,
+ "step": 3078
+ },
+ {
+ "epoch": 42.76855895196506,
+ "grad_norm": 0.02580736204981804,
+ "learning_rate": 0.0006,
+ "loss": 7.204472541809082,
+ "step": 3079
+ },
+ {
+ "epoch": 42.7825327510917,
+ "grad_norm": 0.023372117429971695,
+ "learning_rate": 0.0006,
+ "loss": 7.165554523468018,
+ "step": 3080
+ },
+ {
+ "epoch": 42.79650655021834,
+ "grad_norm": 0.04440586268901825,
+ "learning_rate": 0.0006,
+ "loss": 7.169431209564209,
+ "step": 3081
+ },
+ {
+ "epoch": 42.81048034934498,
+ "grad_norm": 0.06797406077384949,
+ "learning_rate": 0.0006,
+ "loss": 7.168193340301514,
+ "step": 3082
+ },
+ {
+ "epoch": 42.82445414847162,
+ "grad_norm": 0.0914321169257164,
+ "learning_rate": 0.0006,
+ "loss": 7.204671859741211,
+ "step": 3083
+ },
+ {
+ "epoch": 42.838427947598255,
+ "grad_norm": 0.0506940595805645,
+ "learning_rate": 0.0006,
+ "loss": 7.15949821472168,
+ "step": 3084
+ },
+ {
+ "epoch": 42.852401746724894,
+ "grad_norm": 0.04819890484213829,
+ "learning_rate": 0.0006,
+ "loss": 7.192837238311768,
+ "step": 3085
+ },
+ {
+ "epoch": 42.866375545851525,
+ "grad_norm": 0.040044233202934265,
+ "learning_rate": 0.0006,
+ "loss": 7.1600165367126465,
+ "step": 3086
+ },
+ {
+ "epoch": 42.880349344978164,
+ "grad_norm": 0.02573299966752529,
+ "learning_rate": 0.0006,
+ "loss": 7.153427600860596,
+ "step": 3087
+ },
+ {
+ "epoch": 42.8943231441048,
+ "grad_norm": 0.046942535787820816,
+ "learning_rate": 0.0006,
+ "loss": 7.16767692565918,
+ "step": 3088
+ },
+ {
+ "epoch": 42.90829694323144,
+ "grad_norm": 0.02683161571621895,
+ "learning_rate": 0.0006,
+ "loss": 7.17144250869751,
+ "step": 3089
+ },
+ {
+ "epoch": 42.92227074235808,
+ "grad_norm": 0.05363674834370613,
+ "learning_rate": 0.0006,
+ "loss": 7.120128631591797,
+ "step": 3090
+ },
+ {
+ "epoch": 42.93624454148472,
+ "grad_norm": 0.023756247013807297,
+ "learning_rate": 0.0006,
+ "loss": 7.169703006744385,
+ "step": 3091
+ },
+ {
+ "epoch": 42.95021834061135,
+ "grad_norm": 0.04216783866286278,
+ "learning_rate": 0.0006,
+ "loss": 7.1206865310668945,
+ "step": 3092
+ },
+ {
+ "epoch": 42.96419213973799,
+ "grad_norm": 0.02219565585255623,
+ "learning_rate": 0.0006,
+ "loss": 7.157215118408203,
+ "step": 3093
+ },
+ {
+ "epoch": 42.97816593886463,
+ "grad_norm": 0.03527127206325531,
+ "learning_rate": 0.0006,
+ "loss": 7.166672706604004,
+ "step": 3094
+ },
+ {
+ "epoch": 42.992139737991266,
+ "grad_norm": 0.017748817801475525,
+ "learning_rate": 0.0006,
+ "loss": 7.154747009277344,
+ "step": 3095
+ },
+ {
+ "epoch": 43.0,
+ "grad_norm": 0.033165838569402695,
+ "learning_rate": 0.0006,
+ "loss": 7.150866508483887,
+ "step": 3096
+ },
+ {
+ "epoch": 43.0,
+ "eval_loss": 7.144403457641602,
+ "eval_runtime": 56.8853,
+ "eval_samples_per_second": 42.928,
+ "eval_steps_per_second": 1.354,
+ "step": 3096
+ },
+ {
+ "epoch": 43.01397379912664,
+ "grad_norm": 0.015199829824268818,
+ "learning_rate": 0.0006,
+ "loss": 7.122735023498535,
+ "step": 3097
+ },
+ {
+ "epoch": 43.02794759825328,
+ "grad_norm": 0.03220256045460701,
+ "learning_rate": 0.0006,
+ "loss": 7.143553733825684,
+ "step": 3098
+ },
+ {
+ "epoch": 43.041921397379916,
+ "grad_norm": 0.014687920920550823,
+ "learning_rate": 0.0006,
+ "loss": 7.080960750579834,
+ "step": 3099
+ },
+ {
+ "epoch": 43.05589519650655,
+ "grad_norm": 0.023522932082414627,
+ "learning_rate": 0.0006,
+ "loss": 7.137356758117676,
+ "step": 3100
+ },
+ {
+ "epoch": 43.069868995633186,
+ "grad_norm": 0.014808216132223606,
+ "learning_rate": 0.0006,
+ "loss": 7.09143590927124,
+ "step": 3101
+ },
+ {
+ "epoch": 43.083842794759825,
+ "grad_norm": 0.029339246451854706,
+ "learning_rate": 0.0006,
+ "loss": 7.122198104858398,
+ "step": 3102
+ },
+ {
+ "epoch": 43.09781659388646,
+ "grad_norm": 0.022953078150749207,
+ "learning_rate": 0.0006,
+ "loss": 7.092505931854248,
+ "step": 3103
+ },
+ {
+ "epoch": 43.1117903930131,
+ "grad_norm": 0.04590052738785744,
+ "learning_rate": 0.0006,
+ "loss": 7.113100051879883,
+ "step": 3104
+ },
+ {
+ "epoch": 43.12576419213974,
+ "grad_norm": 0.0524442233145237,
+ "learning_rate": 0.0006,
+ "loss": 7.147022247314453,
+ "step": 3105
+ },
+ {
+ "epoch": 43.13973799126637,
+ "grad_norm": 0.0315794013440609,
+ "learning_rate": 0.0006,
+ "loss": 7.094045162200928,
+ "step": 3106
+ },
+ {
+ "epoch": 43.15371179039301,
+ "grad_norm": 0.05112722888588905,
+ "learning_rate": 0.0006,
+ "loss": 7.054388523101807,
+ "step": 3107
+ },
+ {
+ "epoch": 43.16768558951965,
+ "grad_norm": 0.03146776929497719,
+ "learning_rate": 0.0006,
+ "loss": 7.116674423217773,
+ "step": 3108
+ },
+ {
+ "epoch": 43.18165938864629,
+ "grad_norm": 0.03391246497631073,
+ "learning_rate": 0.0006,
+ "loss": 7.150345802307129,
+ "step": 3109
+ },
+ {
+ "epoch": 43.19563318777293,
+ "grad_norm": 0.026957903057336807,
+ "learning_rate": 0.0006,
+ "loss": 7.058531284332275,
+ "step": 3110
+ },
+ {
+ "epoch": 43.209606986899566,
+ "grad_norm": 0.041532568633556366,
+ "learning_rate": 0.0006,
+ "loss": 7.085498809814453,
+ "step": 3111
+ },
+ {
+ "epoch": 43.223580786026204,
+ "grad_norm": 0.020001588389277458,
+ "learning_rate": 0.0006,
+ "loss": 7.131487846374512,
+ "step": 3112
+ },
+ {
+ "epoch": 43.237554585152836,
+ "grad_norm": 0.02268712781369686,
+ "learning_rate": 0.0006,
+ "loss": 7.0790863037109375,
+ "step": 3113
+ },
+ {
+ "epoch": 43.251528384279474,
+ "grad_norm": 0.023363925516605377,
+ "learning_rate": 0.0006,
+ "loss": 7.067816734313965,
+ "step": 3114
+ },
+ {
+ "epoch": 43.26550218340611,
+ "grad_norm": 0.019367508590221405,
+ "learning_rate": 0.0006,
+ "loss": 7.042041778564453,
+ "step": 3115
+ },
+ {
+ "epoch": 43.27947598253275,
+ "grad_norm": 0.02745090238749981,
+ "learning_rate": 0.0006,
+ "loss": 7.064733028411865,
+ "step": 3116
+ },
+ {
+ "epoch": 43.29344978165939,
+ "grad_norm": 0.02543802745640278,
+ "learning_rate": 0.0006,
+ "loss": 7.058948040008545,
+ "step": 3117
+ },
+ {
+ "epoch": 43.30742358078603,
+ "grad_norm": 0.018874092027544975,
+ "learning_rate": 0.0006,
+ "loss": 7.009946346282959,
+ "step": 3118
+ },
+ {
+ "epoch": 43.32139737991266,
+ "grad_norm": 0.024662822484970093,
+ "learning_rate": 0.0006,
+ "loss": 7.015913963317871,
+ "step": 3119
+ },
+ {
+ "epoch": 43.3353711790393,
+ "grad_norm": 0.02441416308283806,
+ "learning_rate": 0.0006,
+ "loss": 7.061622142791748,
+ "step": 3120
+ },
+ {
+ "epoch": 43.34934497816594,
+ "grad_norm": 0.031887322664260864,
+ "learning_rate": 0.0006,
+ "loss": 7.046883583068848,
+ "step": 3121
+ },
+ {
+ "epoch": 43.36331877729258,
+ "grad_norm": 0.021930739283561707,
+ "learning_rate": 0.0006,
+ "loss": 7.0309624671936035,
+ "step": 3122
+ },
+ {
+ "epoch": 43.377292576419215,
+ "grad_norm": 0.02722783200442791,
+ "learning_rate": 0.0006,
+ "loss": 7.072924613952637,
+ "step": 3123
+ },
+ {
+ "epoch": 43.391266375545854,
+ "grad_norm": 0.01483179721981287,
+ "learning_rate": 0.0006,
+ "loss": 7.028624534606934,
+ "step": 3124
+ },
+ {
+ "epoch": 43.40524017467249,
+ "grad_norm": 0.024793945252895355,
+ "learning_rate": 0.0006,
+ "loss": 7.001623153686523,
+ "step": 3125
+ },
+ {
+ "epoch": 43.419213973799124,
+ "grad_norm": 0.020148713141679764,
+ "learning_rate": 0.0006,
+ "loss": 6.9938812255859375,
+ "step": 3126
+ },
+ {
+ "epoch": 43.43318777292576,
+ "grad_norm": 0.02413850650191307,
+ "learning_rate": 0.0006,
+ "loss": 7.001243591308594,
+ "step": 3127
+ },
+ {
+ "epoch": 43.4471615720524,
+ "grad_norm": 0.028549203649163246,
+ "learning_rate": 0.0006,
+ "loss": 6.987671375274658,
+ "step": 3128
+ },
+ {
+ "epoch": 43.46113537117904,
+ "grad_norm": 0.025988785549998283,
+ "learning_rate": 0.0006,
+ "loss": 6.9551167488098145,
+ "step": 3129
+ },
+ {
+ "epoch": 43.47510917030568,
+ "grad_norm": 0.027140729129314423,
+ "learning_rate": 0.0006,
+ "loss": 6.9547810554504395,
+ "step": 3130
+ },
+ {
+ "epoch": 43.48908296943232,
+ "grad_norm": 0.030130082741379738,
+ "learning_rate": 0.0006,
+ "loss": 6.993186950683594,
+ "step": 3131
+ },
+ {
+ "epoch": 43.50305676855895,
+ "grad_norm": 0.025996994227170944,
+ "learning_rate": 0.0006,
+ "loss": 6.984624862670898,
+ "step": 3132
+ },
+ {
+ "epoch": 43.51703056768559,
+ "grad_norm": 0.02423008903861046,
+ "learning_rate": 0.0006,
+ "loss": 6.98813009262085,
+ "step": 3133
+ },
+ {
+ "epoch": 43.531004366812226,
+ "grad_norm": 0.029665837064385414,
+ "learning_rate": 0.0006,
+ "loss": 6.986610412597656,
+ "step": 3134
+ },
+ {
+ "epoch": 43.544978165938865,
+ "grad_norm": 0.028768595308065414,
+ "learning_rate": 0.0006,
+ "loss": 6.898797035217285,
+ "step": 3135
+ },
+ {
+ "epoch": 43.5589519650655,
+ "grad_norm": 0.03111208975315094,
+ "learning_rate": 0.0006,
+ "loss": 6.923741817474365,
+ "step": 3136
+ },
+ {
+ "epoch": 43.57292576419214,
+ "grad_norm": 0.0368187390267849,
+ "learning_rate": 0.0006,
+ "loss": 6.994343280792236,
+ "step": 3137
+ },
+ {
+ "epoch": 43.58689956331878,
+ "grad_norm": 0.04076763987541199,
+ "learning_rate": 0.0006,
+ "loss": 6.909458160400391,
+ "step": 3138
+ },
+ {
+ "epoch": 43.60087336244541,
+ "grad_norm": 0.03281061723828316,
+ "learning_rate": 0.0006,
+ "loss": 6.924851417541504,
+ "step": 3139
+ },
+ {
+ "epoch": 43.61484716157205,
+ "grad_norm": 0.025486988946795464,
+ "learning_rate": 0.0006,
+ "loss": 6.962926864624023,
+ "step": 3140
+ },
+ {
+ "epoch": 43.62882096069869,
+ "grad_norm": 0.04583485797047615,
+ "learning_rate": 0.0006,
+ "loss": 6.876555442810059,
+ "step": 3141
+ },
+ {
+ "epoch": 43.64279475982533,
+ "grad_norm": 0.05460785701870918,
+ "learning_rate": 0.0006,
+ "loss": 6.952932357788086,
+ "step": 3142
+ },
+ {
+ "epoch": 43.65676855895197,
+ "grad_norm": 0.040917906910181046,
+ "learning_rate": 0.0006,
+ "loss": 6.9624924659729,
+ "step": 3143
+ },
+ {
+ "epoch": 43.670742358078606,
+ "grad_norm": 0.04009668901562691,
+ "learning_rate": 0.0006,
+ "loss": 6.94033145904541,
+ "step": 3144
+ },
+ {
+ "epoch": 43.68471615720524,
+ "grad_norm": 0.04500625655055046,
+ "learning_rate": 0.0006,
+ "loss": 6.961839199066162,
+ "step": 3145
+ },
+ {
+ "epoch": 43.698689956331876,
+ "grad_norm": 0.03182258829474449,
+ "learning_rate": 0.0006,
+ "loss": 6.930580139160156,
+ "step": 3146
+ },
+ {
+ "epoch": 43.712663755458514,
+ "grad_norm": 0.053621504455804825,
+ "learning_rate": 0.0006,
+ "loss": 6.950235366821289,
+ "step": 3147
+ },
+ {
+ "epoch": 43.72663755458515,
+ "grad_norm": 0.031919289380311966,
+ "learning_rate": 0.0006,
+ "loss": 6.895111560821533,
+ "step": 3148
+ },
+ {
+ "epoch": 43.74061135371179,
+ "grad_norm": 0.042713891714811325,
+ "learning_rate": 0.0006,
+ "loss": 6.940885543823242,
+ "step": 3149
+ },
+ {
+ "epoch": 43.75458515283843,
+ "grad_norm": 0.02385653182864189,
+ "learning_rate": 0.0006,
+ "loss": 6.948801517486572,
+ "step": 3150
+ },
+ {
+ "epoch": 43.76855895196506,
+ "grad_norm": 0.04100663587450981,
+ "learning_rate": 0.0006,
+ "loss": 6.970411777496338,
+ "step": 3151
+ },
+ {
+ "epoch": 43.7825327510917,
+ "grad_norm": 0.02729583904147148,
+ "learning_rate": 0.0006,
+ "loss": 6.899379253387451,
+ "step": 3152
+ },
+ {
+ "epoch": 43.79650655021834,
+ "grad_norm": 0.03128620237112045,
+ "learning_rate": 0.0006,
+ "loss": 6.931319236755371,
+ "step": 3153
+ },
+ {
+ "epoch": 43.81048034934498,
+ "grad_norm": 0.025488659739494324,
+ "learning_rate": 0.0006,
+ "loss": 6.8486409187316895,
+ "step": 3154
+ },
+ {
+ "epoch": 43.82445414847162,
+ "grad_norm": 0.03679376840591431,
+ "learning_rate": 0.0006,
+ "loss": 6.840811252593994,
+ "step": 3155
+ },
+ {
+ "epoch": 43.838427947598255,
+ "grad_norm": 0.020578034222126007,
+ "learning_rate": 0.0006,
+ "loss": 6.9088616371154785,
+ "step": 3156
+ },
+ {
+ "epoch": 43.852401746724894,
+ "grad_norm": 0.02935841865837574,
+ "learning_rate": 0.0006,
+ "loss": 6.806676864624023,
+ "step": 3157
+ },
+ {
+ "epoch": 43.866375545851525,
+ "grad_norm": 0.023319244384765625,
+ "learning_rate": 0.0006,
+ "loss": 6.878457546234131,
+ "step": 3158
+ },
+ {
+ "epoch": 43.880349344978164,
+ "grad_norm": 0.02779589593410492,
+ "learning_rate": 0.0006,
+ "loss": 6.822447776794434,
+ "step": 3159
+ },
+ {
+ "epoch": 43.8943231441048,
+ "grad_norm": 0.01879825070500374,
+ "learning_rate": 0.0006,
+ "loss": 6.854121208190918,
+ "step": 3160
+ },
+ {
+ "epoch": 43.90829694323144,
+ "grad_norm": 0.023286139592528343,
+ "learning_rate": 0.0006,
+ "loss": 6.836827278137207,
+ "step": 3161
+ },
+ {
+ "epoch": 43.92227074235808,
+ "grad_norm": 0.030358297750353813,
+ "learning_rate": 0.0006,
+ "loss": 6.860037803649902,
+ "step": 3162
+ },
+ {
+ "epoch": 43.93624454148472,
+ "grad_norm": 0.029621105641126633,
+ "learning_rate": 0.0006,
+ "loss": 6.884025573730469,
+ "step": 3163
+ },
+ {
+ "epoch": 43.95021834061135,
+ "grad_norm": 0.029829824343323708,
+ "learning_rate": 0.0006,
+ "loss": 6.8702392578125,
+ "step": 3164
+ },
+ {
+ "epoch": 43.96419213973799,
+ "grad_norm": 0.036892473697662354,
+ "learning_rate": 0.0006,
+ "loss": 6.806652069091797,
+ "step": 3165
+ },
+ {
+ "epoch": 43.97816593886463,
+ "grad_norm": 0.03206709027290344,
+ "learning_rate": 0.0006,
+ "loss": 6.828495979309082,
+ "step": 3166
+ },
+ {
+ "epoch": 43.992139737991266,
+ "grad_norm": 0.02744000405073166,
+ "learning_rate": 0.0006,
+ "loss": 6.832002639770508,
+ "step": 3167
+ },
+ {
+ "epoch": 44.0,
+ "grad_norm": 0.02205660194158554,
+ "learning_rate": 0.0006,
+ "loss": 6.898828506469727,
+ "step": 3168
+ },
+ {
+ "epoch": 44.0,
+ "eval_loss": 6.8150553703308105,
+ "eval_runtime": 56.784,
+ "eval_samples_per_second": 43.005,
+ "eval_steps_per_second": 1.356,
+ "step": 3168
+ },
+ {
+ "epoch": 44.01397379912664,
+ "grad_norm": 0.030421260744333267,
+ "learning_rate": 0.0006,
+ "loss": 6.775057792663574,
+ "step": 3169
+ },
+ {
+ "epoch": 44.02794759825328,
+ "grad_norm": 0.0465611033141613,
+ "learning_rate": 0.0006,
+ "loss": 6.849390506744385,
+ "step": 3170
+ },
+ {
+ "epoch": 44.041921397379916,
+ "grad_norm": 0.05977972596883774,
+ "learning_rate": 0.0006,
+ "loss": 6.725576400756836,
+ "step": 3171
+ },
+ {
+ "epoch": 44.05589519650655,
+ "grad_norm": 0.05432697385549545,
+ "learning_rate": 0.0006,
+ "loss": 6.768024444580078,
+ "step": 3172
+ },
+ {
+ "epoch": 44.069868995633186,
+ "grad_norm": 0.04850967228412628,
+ "learning_rate": 0.0006,
+ "loss": 6.799457550048828,
+ "step": 3173
+ },
+ {
+ "epoch": 44.083842794759825,
+ "grad_norm": 0.06201786920428276,
+ "learning_rate": 0.0006,
+ "loss": 6.897234916687012,
+ "step": 3174
+ },
+ {
+ "epoch": 44.09781659388646,
+ "grad_norm": 0.04695591703057289,
+ "learning_rate": 0.0006,
+ "loss": 6.826470375061035,
+ "step": 3175
+ },
+ {
+ "epoch": 44.1117903930131,
+ "grad_norm": 0.03850164636969566,
+ "learning_rate": 0.0006,
+ "loss": 6.835901260375977,
+ "step": 3176
+ },
+ {
+ "epoch": 44.12576419213974,
+ "grad_norm": 0.04250155761837959,
+ "learning_rate": 0.0006,
+ "loss": 6.782310485839844,
+ "step": 3177
+ },
+ {
+ "epoch": 44.13973799126637,
+ "grad_norm": 0.041465479880571365,
+ "learning_rate": 0.0006,
+ "loss": 6.785732746124268,
+ "step": 3178
+ },
+ {
+ "epoch": 44.15371179039301,
+ "grad_norm": 0.02527807280421257,
+ "learning_rate": 0.0006,
+ "loss": 6.78061580657959,
+ "step": 3179
+ },
+ {
+ "epoch": 44.16768558951965,
+ "grad_norm": 0.03660374879837036,
+ "learning_rate": 0.0006,
+ "loss": 6.783992767333984,
+ "step": 3180
+ },
+ {
+ "epoch": 44.18165938864629,
+ "grad_norm": 0.027567964047193527,
+ "learning_rate": 0.0006,
+ "loss": 6.765413284301758,
+ "step": 3181
+ },
+ {
+ "epoch": 44.19563318777293,
+ "grad_norm": 0.024301424622535706,
+ "learning_rate": 0.0006,
+ "loss": 6.760368347167969,
+ "step": 3182
+ },
+ {
+ "epoch": 44.209606986899566,
+ "grad_norm": 0.0266821701079607,
+ "learning_rate": 0.0006,
+ "loss": 6.71417236328125,
+ "step": 3183
+ },
+ {
+ "epoch": 44.223580786026204,
+ "grad_norm": 0.028125090524554253,
+ "learning_rate": 0.0006,
+ "loss": 6.646126747131348,
+ "step": 3184
+ },
+ {
+ "epoch": 44.237554585152836,
+ "grad_norm": 0.0185737032443285,
+ "learning_rate": 0.0006,
+ "loss": 6.712944030761719,
+ "step": 3185
+ },
+ {
+ "epoch": 44.251528384279474,
+ "grad_norm": 0.025836674496531487,
+ "learning_rate": 0.0006,
+ "loss": 6.707030296325684,
+ "step": 3186
+ },
+ {
+ "epoch": 44.26550218340611,
+ "grad_norm": 0.029744870960712433,
+ "learning_rate": 0.0006,
+ "loss": 6.6271071434021,
+ "step": 3187
+ },
+ {
+ "epoch": 44.27947598253275,
+ "grad_norm": 0.03021426685154438,
+ "learning_rate": 0.0006,
+ "loss": 6.677826404571533,
+ "step": 3188
+ },
+ {
+ "epoch": 44.29344978165939,
+ "grad_norm": 0.032727014273405075,
+ "learning_rate": 0.0006,
+ "loss": 6.657764434814453,
+ "step": 3189
+ },
+ {
+ "epoch": 44.30742358078603,
+ "grad_norm": 0.03696778789162636,
+ "learning_rate": 0.0006,
+ "loss": 6.632883548736572,
+ "step": 3190
+ },
+ {
+ "epoch": 44.32139737991266,
+ "grad_norm": 0.034630827605724335,
+ "learning_rate": 0.0006,
+ "loss": 6.759583473205566,
+ "step": 3191
+ },
+ {
+ "epoch": 44.3353711790393,
+ "grad_norm": 0.030517758801579475,
+ "learning_rate": 0.0006,
+ "loss": 6.702268600463867,
+ "step": 3192
+ },
+ {
+ "epoch": 44.34934497816594,
+ "grad_norm": 0.04105190932750702,
+ "learning_rate": 0.0006,
+ "loss": 6.705792427062988,
+ "step": 3193
+ },
+ {
+ "epoch": 44.36331877729258,
+ "grad_norm": 0.06659174710512161,
+ "learning_rate": 0.0006,
+ "loss": 6.73862361907959,
+ "step": 3194
+ },
+ {
+ "epoch": 44.377292576419215,
+ "grad_norm": 0.08013619482517242,
+ "learning_rate": 0.0006,
+ "loss": 6.710282802581787,
+ "step": 3195
+ },
+ {
+ "epoch": 44.391266375545854,
+ "grad_norm": 0.049850963056087494,
+ "learning_rate": 0.0006,
+ "loss": 6.684974670410156,
+ "step": 3196
+ },
+ {
+ "epoch": 44.40524017467249,
+ "grad_norm": 0.061247408390045166,
+ "learning_rate": 0.0006,
+ "loss": 6.716034412384033,
+ "step": 3197
+ },
+ {
+ "epoch": 44.419213973799124,
+ "grad_norm": 0.05868857726454735,
+ "learning_rate": 0.0006,
+ "loss": 6.589107513427734,
+ "step": 3198
+ },
+ {
+ "epoch": 44.43318777292576,
+ "grad_norm": 0.048878561705350876,
+ "learning_rate": 0.0006,
+ "loss": 6.694855690002441,
+ "step": 3199
+ },
+ {
+ "epoch": 44.4471615720524,
+ "grad_norm": 0.028748324140906334,
+ "learning_rate": 0.0006,
+ "loss": 6.629969596862793,
+ "step": 3200
+ },
+ {
+ "epoch": 44.46113537117904,
+ "grad_norm": 0.04060306400060654,
+ "learning_rate": 0.0006,
+ "loss": 6.721431255340576,
+ "step": 3201
+ },
+ {
+ "epoch": 44.47510917030568,
+ "grad_norm": 0.03833824768662453,
+ "learning_rate": 0.0006,
+ "loss": 6.625029563903809,
+ "step": 3202
+ },
+ {
+ "epoch": 44.48908296943232,
+ "grad_norm": 0.03791997581720352,
+ "learning_rate": 0.0006,
+ "loss": 6.669187545776367,
+ "step": 3203
+ },
+ {
+ "epoch": 44.50305676855895,
+ "grad_norm": 0.03259684890508652,
+ "learning_rate": 0.0006,
+ "loss": 6.665643692016602,
+ "step": 3204
+ },
+ {
+ "epoch": 44.51703056768559,
+ "grad_norm": 0.040658626705408096,
+ "learning_rate": 0.0006,
+ "loss": 6.639796257019043,
+ "step": 3205
+ },
+ {
+ "epoch": 44.531004366812226,
+ "grad_norm": 0.03691193461418152,
+ "learning_rate": 0.0006,
+ "loss": 6.692244529724121,
+ "step": 3206
+ },
+ {
+ "epoch": 44.544978165938865,
+ "grad_norm": 0.02814812771975994,
+ "learning_rate": 0.0006,
+ "loss": 6.61978006362915,
+ "step": 3207
+ },
+ {
+ "epoch": 44.5589519650655,
+ "grad_norm": 0.022155074402689934,
+ "learning_rate": 0.0006,
+ "loss": 6.625905990600586,
+ "step": 3208
+ },
+ {
+ "epoch": 44.57292576419214,
+ "grad_norm": 0.030794885009527206,
+ "learning_rate": 0.0006,
+ "loss": 6.678500652313232,
+ "step": 3209
+ },
+ {
+ "epoch": 44.58689956331878,
+ "grad_norm": 0.03201768547296524,
+ "learning_rate": 0.0006,
+ "loss": 6.6015729904174805,
+ "step": 3210
+ },
+ {
+ "epoch": 44.60087336244541,
+ "grad_norm": 0.025138361379504204,
+ "learning_rate": 0.0006,
+ "loss": 6.568234443664551,
+ "step": 3211
+ },
+ {
+ "epoch": 44.61484716157205,
+ "grad_norm": 0.02969353087246418,
+ "learning_rate": 0.0006,
+ "loss": 6.588366508483887,
+ "step": 3212
+ },
+ {
+ "epoch": 44.62882096069869,
+ "grad_norm": 0.03136962279677391,
+ "learning_rate": 0.0006,
+ "loss": 6.479961395263672,
+ "step": 3213
+ },
+ {
+ "epoch": 44.64279475982533,
+ "grad_norm": 0.030178170651197433,
+ "learning_rate": 0.0006,
+ "loss": 6.622910976409912,
+ "step": 3214
+ },
+ {
+ "epoch": 44.65676855895197,
+ "grad_norm": 0.02682245895266533,
+ "learning_rate": 0.0006,
+ "loss": 6.597707748413086,
+ "step": 3215
+ },
+ {
+ "epoch": 44.670742358078606,
+ "grad_norm": 0.025154557079076767,
+ "learning_rate": 0.0006,
+ "loss": 6.630541801452637,
+ "step": 3216
+ },
+ {
+ "epoch": 44.68471615720524,
+ "grad_norm": 0.03564463183283806,
+ "learning_rate": 0.0006,
+ "loss": 6.542823791503906,
+ "step": 3217
+ },
+ {
+ "epoch": 44.698689956331876,
+ "grad_norm": 0.049477651715278625,
+ "learning_rate": 0.0006,
+ "loss": 6.646030426025391,
+ "step": 3218
+ },
+ {
+ "epoch": 44.712663755458514,
+ "grad_norm": 0.06536535918712616,
+ "learning_rate": 0.0006,
+ "loss": 6.595463275909424,
+ "step": 3219
+ },
+ {
+ "epoch": 44.72663755458515,
+ "grad_norm": 0.07710316777229309,
+ "learning_rate": 0.0006,
+ "loss": 6.55524206161499,
+ "step": 3220
+ },
+ {
+ "epoch": 44.74061135371179,
+ "grad_norm": 0.09571163356304169,
+ "learning_rate": 0.0006,
+ "loss": 6.659764289855957,
+ "step": 3221
+ },
+ {
+ "epoch": 44.75458515283843,
+ "grad_norm": 0.055076830089092255,
+ "learning_rate": 0.0006,
+ "loss": 6.658882141113281,
+ "step": 3222
+ },
+ {
+ "epoch": 44.76855895196506,
+ "grad_norm": 0.07671099156141281,
+ "learning_rate": 0.0006,
+ "loss": 6.605809211730957,
+ "step": 3223
+ },
+ {
+ "epoch": 44.7825327510917,
+ "grad_norm": 0.05295584350824356,
+ "learning_rate": 0.0006,
+ "loss": 6.641455173492432,
+ "step": 3224
+ },
+ {
+ "epoch": 44.79650655021834,
+ "grad_norm": 0.09128042310476303,
+ "learning_rate": 0.0006,
+ "loss": 6.691066741943359,
+ "step": 3225
+ },
+ {
+ "epoch": 44.81048034934498,
+ "grad_norm": 0.05362958461046219,
+ "learning_rate": 0.0006,
+ "loss": 6.573406219482422,
+ "step": 3226
+ },
+ {
+ "epoch": 44.82445414847162,
+ "grad_norm": 0.046910062432289124,
+ "learning_rate": 0.0006,
+ "loss": 6.550876617431641,
+ "step": 3227
+ },
+ {
+ "epoch": 44.838427947598255,
+ "grad_norm": 0.03494197130203247,
+ "learning_rate": 0.0006,
+ "loss": 6.540179252624512,
+ "step": 3228
+ },
+ {
+ "epoch": 44.852401746724894,
+ "grad_norm": 0.04716387391090393,
+ "learning_rate": 0.0006,
+ "loss": 6.490459442138672,
+ "step": 3229
+ },
+ {
+ "epoch": 44.866375545851525,
+ "grad_norm": 0.04190506786108017,
+ "learning_rate": 0.0006,
+ "loss": 6.509925365447998,
+ "step": 3230
+ },
+ {
+ "epoch": 44.880349344978164,
+ "grad_norm": 0.04749064892530441,
+ "learning_rate": 0.0006,
+ "loss": 6.478779315948486,
+ "step": 3231
+ },
+ {
+ "epoch": 44.8943231441048,
+ "grad_norm": 0.028002185747027397,
+ "learning_rate": 0.0006,
+ "loss": 6.528904914855957,
+ "step": 3232
+ },
+ {
+ "epoch": 44.90829694323144,
+ "grad_norm": 0.036473166197538376,
+ "learning_rate": 0.0006,
+ "loss": 6.549042701721191,
+ "step": 3233
+ },
+ {
+ "epoch": 44.92227074235808,
+ "grad_norm": 0.02543315477669239,
+ "learning_rate": 0.0006,
+ "loss": 6.498875141143799,
+ "step": 3234
+ },
+ {
+ "epoch": 44.93624454148472,
+ "grad_norm": 0.029906807467341423,
+ "learning_rate": 0.0006,
+ "loss": 6.453019142150879,
+ "step": 3235
+ },
+ {
+ "epoch": 44.95021834061135,
+ "grad_norm": 0.027231857180595398,
+ "learning_rate": 0.0006,
+ "loss": 6.428829193115234,
+ "step": 3236
+ },
+ {
+ "epoch": 44.96419213973799,
+ "grad_norm": 0.02506154775619507,
+ "learning_rate": 0.0006,
+ "loss": 6.420858383178711,
+ "step": 3237
+ },
+ {
+ "epoch": 44.97816593886463,
+ "grad_norm": 0.0269472673535347,
+ "learning_rate": 0.0006,
+ "loss": 6.484989643096924,
+ "step": 3238
+ },
+ {
+ "epoch": 44.992139737991266,
+ "grad_norm": 0.023410487920045853,
+ "learning_rate": 0.0006,
+ "loss": 6.454830169677734,
+ "step": 3239
+ },
+ {
+ "epoch": 45.0,
+ "grad_norm": 0.034175023436546326,
+ "learning_rate": 0.0006,
+ "loss": 6.430975437164307,
+ "step": 3240
+ },
+ {
+ "epoch": 45.0,
+ "eval_loss": 6.458896160125732,
+ "eval_runtime": 56.6934,
+ "eval_samples_per_second": 43.074,
+ "eval_steps_per_second": 1.358,
+ "step": 3240
+ },
+ {
+ "epoch": 45.01397379912664,
+ "grad_norm": 0.045031629502773285,
+ "learning_rate": 0.0006,
+ "loss": 6.4243011474609375,
+ "step": 3241
+ },
+ {
+ "epoch": 45.02794759825328,
+ "grad_norm": 0.04968734085559845,
+ "learning_rate": 0.0006,
+ "loss": 6.380091667175293,
+ "step": 3242
+ },
+ {
+ "epoch": 45.041921397379916,
+ "grad_norm": 0.047145187854766846,
+ "learning_rate": 0.0006,
+ "loss": 6.477178573608398,
+ "step": 3243
+ },
+ {
+ "epoch": 45.05589519650655,
+ "grad_norm": 0.04304948449134827,
+ "learning_rate": 0.0006,
+ "loss": 6.399726867675781,
+ "step": 3244
+ },
+ {
+ "epoch": 45.069868995633186,
+ "grad_norm": 0.04466010630130768,
+ "learning_rate": 0.0006,
+ "loss": 6.3415679931640625,
+ "step": 3245
+ },
+ {
+ "epoch": 45.083842794759825,
+ "grad_norm": 0.04305306077003479,
+ "learning_rate": 0.0006,
+ "loss": 6.3413310050964355,
+ "step": 3246
+ },
+ {
+ "epoch": 45.09781659388646,
+ "grad_norm": 0.043064575642347336,
+ "learning_rate": 0.0006,
+ "loss": 6.4838714599609375,
+ "step": 3247
+ },
+ {
+ "epoch": 45.1117903930131,
+ "grad_norm": 0.049503277987241745,
+ "learning_rate": 0.0006,
+ "loss": 6.38473653793335,
+ "step": 3248
+ },
+ {
+ "epoch": 45.12576419213974,
+ "grad_norm": 0.04613238573074341,
+ "learning_rate": 0.0006,
+ "loss": 6.370769500732422,
+ "step": 3249
+ },
+ {
+ "epoch": 45.13973799126637,
+ "grad_norm": 0.052119478583335876,
+ "learning_rate": 0.0006,
+ "loss": 6.362321853637695,
+ "step": 3250
+ },
+ {
+ "epoch": 45.15371179039301,
+ "grad_norm": 0.045110076665878296,
+ "learning_rate": 0.0006,
+ "loss": 6.369844913482666,
+ "step": 3251
+ },
+ {
+ "epoch": 45.16768558951965,
+ "grad_norm": 0.03798583149909973,
+ "learning_rate": 0.0006,
+ "loss": 6.324633598327637,
+ "step": 3252
+ },
+ {
+ "epoch": 45.18165938864629,
+ "grad_norm": 0.03744659945368767,
+ "learning_rate": 0.0006,
+ "loss": 6.349421501159668,
+ "step": 3253
+ },
+ {
+ "epoch": 45.19563318777293,
+ "grad_norm": 0.03767292574048042,
+ "learning_rate": 0.0006,
+ "loss": 6.334238052368164,
+ "step": 3254
+ },
+ {
+ "epoch": 45.209606986899566,
+ "grad_norm": 0.03577509894967079,
+ "learning_rate": 0.0006,
+ "loss": 6.3453826904296875,
+ "step": 3255
+ },
+ {
+ "epoch": 45.223580786026204,
+ "grad_norm": 0.03403240442276001,
+ "learning_rate": 0.0006,
+ "loss": 6.2836174964904785,
+ "step": 3256
+ },
+ {
+ "epoch": 45.237554585152836,
+ "grad_norm": 0.03517331928014755,
+ "learning_rate": 0.0006,
+ "loss": 6.281718730926514,
+ "step": 3257
+ },
+ {
+ "epoch": 45.251528384279474,
+ "grad_norm": 0.04409139230847359,
+ "learning_rate": 0.0006,
+ "loss": 6.320815086364746,
+ "step": 3258
+ },
+ {
+ "epoch": 45.26550218340611,
+ "grad_norm": 0.054750364273786545,
+ "learning_rate": 0.0006,
+ "loss": 6.343618392944336,
+ "step": 3259
+ },
+ {
+ "epoch": 45.27947598253275,
+ "grad_norm": 0.07612067461013794,
+ "learning_rate": 0.0006,
+ "loss": 6.310614585876465,
+ "step": 3260
+ },
+ {
+ "epoch": 45.29344978165939,
+ "grad_norm": 0.08438985049724579,
+ "learning_rate": 0.0006,
+ "loss": 6.2369184494018555,
+ "step": 3261
+ },
+ {
+ "epoch": 45.30742358078603,
+ "grad_norm": 0.0782201811671257,
+ "learning_rate": 0.0006,
+ "loss": 6.254452705383301,
+ "step": 3262
+ },
+ {
+ "epoch": 45.32139737991266,
+ "grad_norm": 0.09928355365991592,
+ "learning_rate": 0.0006,
+ "loss": 6.3112077713012695,
+ "step": 3263
+ },
+ {
+ "epoch": 45.3353711790393,
+ "grad_norm": 0.07233133912086487,
+ "learning_rate": 0.0006,
+ "loss": 6.408087253570557,
+ "step": 3264
+ },
+ {
+ "epoch": 45.34934497816594,
+ "grad_norm": 0.07478684186935425,
+ "learning_rate": 0.0006,
+ "loss": 6.3770036697387695,
+ "step": 3265
+ },
+ {
+ "epoch": 45.36331877729258,
+ "grad_norm": 0.06036846712231636,
+ "learning_rate": 0.0006,
+ "loss": 6.33865213394165,
+ "step": 3266
+ },
+ {
+ "epoch": 45.377292576419215,
+ "grad_norm": 0.05783458054065704,
+ "learning_rate": 0.0006,
+ "loss": 6.363203048706055,
+ "step": 3267
+ },
+ {
+ "epoch": 45.391266375545854,
+ "grad_norm": 0.05443909019231796,
+ "learning_rate": 0.0006,
+ "loss": 6.287069797515869,
+ "step": 3268
+ },
+ {
+ "epoch": 45.40524017467249,
+ "grad_norm": 0.0557909794151783,
+ "learning_rate": 0.0006,
+ "loss": 6.297784805297852,
+ "step": 3269
+ },
+ {
+ "epoch": 45.419213973799124,
+ "grad_norm": 0.03483372926712036,
+ "learning_rate": 0.0006,
+ "loss": 6.280734062194824,
+ "step": 3270
+ },
+ {
+ "epoch": 45.43318777292576,
+ "grad_norm": 0.0425543375313282,
+ "learning_rate": 0.0006,
+ "loss": 6.33565616607666,
+ "step": 3271
+ },
+ {
+ "epoch": 45.4471615720524,
+ "grad_norm": 0.03230159357190132,
+ "learning_rate": 0.0006,
+ "loss": 6.197698593139648,
+ "step": 3272
+ },
+ {
+ "epoch": 45.46113537117904,
+ "grad_norm": 0.036897290498018265,
+ "learning_rate": 0.0006,
+ "loss": 6.241239547729492,
+ "step": 3273
+ },
+ {
+ "epoch": 45.47510917030568,
+ "grad_norm": 0.031103838235139847,
+ "learning_rate": 0.0006,
+ "loss": 6.159435749053955,
+ "step": 3274
+ },
+ {
+ "epoch": 45.48908296943232,
+ "grad_norm": 0.027116449549794197,
+ "learning_rate": 0.0006,
+ "loss": 6.212225914001465,
+ "step": 3275
+ },
+ {
+ "epoch": 45.50305676855895,
+ "grad_norm": 0.029923981055617332,
+ "learning_rate": 0.0006,
+ "loss": 6.157269477844238,
+ "step": 3276
+ },
+ {
+ "epoch": 45.51703056768559,
+ "grad_norm": 0.02521538734436035,
+ "learning_rate": 0.0006,
+ "loss": 6.146918773651123,
+ "step": 3277
+ },
+ {
+ "epoch": 45.531004366812226,
+ "grad_norm": 0.02758920192718506,
+ "learning_rate": 0.0006,
+ "loss": 6.0446977615356445,
+ "step": 3278
+ },
+ {
+ "epoch": 45.544978165938865,
+ "grad_norm": 0.024328220635652542,
+ "learning_rate": 0.0006,
+ "loss": 6.099508285522461,
+ "step": 3279
+ },
+ {
+ "epoch": 45.5589519650655,
+ "grad_norm": 0.027425771579146385,
+ "learning_rate": 0.0006,
+ "loss": 6.084299087524414,
+ "step": 3280
+ },
+ {
+ "epoch": 45.57292576419214,
+ "grad_norm": 0.026927651837468147,
+ "learning_rate": 0.0006,
+ "loss": 6.146334171295166,
+ "step": 3281
+ },
+ {
+ "epoch": 45.58689956331878,
+ "grad_norm": 0.04054379090666771,
+ "learning_rate": 0.0006,
+ "loss": 6.1301493644714355,
+ "step": 3282
+ },
+ {
+ "epoch": 45.60087336244541,
+ "grad_norm": 0.04456290975213051,
+ "learning_rate": 0.0006,
+ "loss": 6.152511119842529,
+ "step": 3283
+ },
+ {
+ "epoch": 45.61484716157205,
+ "grad_norm": 0.052821505814790726,
+ "learning_rate": 0.0006,
+ "loss": 6.113544940948486,
+ "step": 3284
+ },
+ {
+ "epoch": 45.62882096069869,
+ "grad_norm": 0.07251739501953125,
+ "learning_rate": 0.0006,
+ "loss": 6.118298530578613,
+ "step": 3285
+ },
+ {
+ "epoch": 45.64279475982533,
+ "grad_norm": 0.09433556348085403,
+ "learning_rate": 0.0006,
+ "loss": 6.170897483825684,
+ "step": 3286
+ },
+ {
+ "epoch": 45.65676855895197,
+ "grad_norm": 0.057683493942022324,
+ "learning_rate": 0.0006,
+ "loss": 6.1747822761535645,
+ "step": 3287
+ },
+ {
+ "epoch": 45.670742358078606,
+ "grad_norm": 0.06363270431756973,
+ "learning_rate": 0.0006,
+ "loss": 6.1617889404296875,
+ "step": 3288
+ },
+ {
+ "epoch": 45.68471615720524,
+ "grad_norm": 0.07528169453144073,
+ "learning_rate": 0.0006,
+ "loss": 6.230060577392578,
+ "step": 3289
+ },
+ {
+ "epoch": 45.698689956331876,
+ "grad_norm": 0.07816170156002045,
+ "learning_rate": 0.0006,
+ "loss": 6.171257019042969,
+ "step": 3290
+ },
+ {
+ "epoch": 45.712663755458514,
+ "grad_norm": 0.05860164389014244,
+ "learning_rate": 0.0006,
+ "loss": 6.095484256744385,
+ "step": 3291
+ },
+ {
+ "epoch": 45.72663755458515,
+ "grad_norm": 0.04747429117560387,
+ "learning_rate": 0.0006,
+ "loss": 6.060067176818848,
+ "step": 3292
+ },
+ {
+ "epoch": 45.74061135371179,
+ "grad_norm": 0.04514474794268608,
+ "learning_rate": 0.0006,
+ "loss": 6.100218772888184,
+ "step": 3293
+ },
+ {
+ "epoch": 45.75458515283843,
+ "grad_norm": 0.05168681964278221,
+ "learning_rate": 0.0006,
+ "loss": 6.0621867179870605,
+ "step": 3294
+ },
+ {
+ "epoch": 45.76855895196506,
+ "grad_norm": 0.04665876924991608,
+ "learning_rate": 0.0006,
+ "loss": 6.118292808532715,
+ "step": 3295
+ },
+ {
+ "epoch": 45.7825327510917,
+ "grad_norm": 0.049374762922525406,
+ "learning_rate": 0.0006,
+ "loss": 6.098394393920898,
+ "step": 3296
+ },
+ {
+ "epoch": 45.79650655021834,
+ "grad_norm": 0.03793549910187721,
+ "learning_rate": 0.0006,
+ "loss": 6.069541931152344,
+ "step": 3297
+ },
+ {
+ "epoch": 45.81048034934498,
+ "grad_norm": 0.036362383514642715,
+ "learning_rate": 0.0006,
+ "loss": 6.1188435554504395,
+ "step": 3298
+ },
+ {
+ "epoch": 45.82445414847162,
+ "grad_norm": 0.03397097811102867,
+ "learning_rate": 0.0006,
+ "loss": 6.117589950561523,
+ "step": 3299
+ },
+ {
+ "epoch": 45.838427947598255,
+ "grad_norm": 0.030977753922343254,
+ "learning_rate": 0.0006,
+ "loss": 6.0236711502075195,
+ "step": 3300
+ },
+ {
+ "epoch": 45.852401746724894,
+ "grad_norm": 0.02800217643380165,
+ "learning_rate": 0.0006,
+ "loss": 6.034825325012207,
+ "step": 3301
+ },
+ {
+ "epoch": 45.866375545851525,
+ "grad_norm": 0.03129778802394867,
+ "learning_rate": 0.0006,
+ "loss": 5.980271339416504,
+ "step": 3302
+ },
+ {
+ "epoch": 45.880349344978164,
+ "grad_norm": 0.025886094197630882,
+ "learning_rate": 0.0006,
+ "loss": 5.966419219970703,
+ "step": 3303
+ },
+ {
+ "epoch": 45.8943231441048,
+ "grad_norm": 0.025712328031659126,
+ "learning_rate": 0.0006,
+ "loss": 5.938652515411377,
+ "step": 3304
+ },
+ {
+ "epoch": 45.90829694323144,
+ "grad_norm": 0.022508298978209496,
+ "learning_rate": 0.0006,
+ "loss": 5.982312202453613,
+ "step": 3305
+ },
+ {
+ "epoch": 45.92227074235808,
+ "grad_norm": 0.024880951270461082,
+ "learning_rate": 0.0006,
+ "loss": 5.931150436401367,
+ "step": 3306
+ },
+ {
+ "epoch": 45.93624454148472,
+ "grad_norm": 0.030007576569914818,
+ "learning_rate": 0.0006,
+ "loss": 5.936639308929443,
+ "step": 3307
+ },
+ {
+ "epoch": 45.95021834061135,
+ "grad_norm": 0.033314041793346405,
+ "learning_rate": 0.0006,
+ "loss": 5.944248676300049,
+ "step": 3308
+ },
+ {
+ "epoch": 45.96419213973799,
+ "grad_norm": 0.043078381568193436,
+ "learning_rate": 0.0006,
+ "loss": 5.971286773681641,
+ "step": 3309
+ },
+ {
+ "epoch": 45.97816593886463,
+ "grad_norm": 0.059296440333127975,
+ "learning_rate": 0.0006,
+ "loss": 5.910497665405273,
+ "step": 3310
+ },
+ {
+ "epoch": 45.992139737991266,
+ "grad_norm": 0.06232666224241257,
+ "learning_rate": 0.0006,
+ "loss": 5.929817199707031,
+ "step": 3311
+ },
+ {
+ "epoch": 46.0,
+ "grad_norm": 0.05303940549492836,
+ "learning_rate": 0.0006,
+ "loss": 5.910822868347168,
+ "step": 3312
+ },
+ {
+ "epoch": 46.0,
+ "eval_loss": 5.977851390838623,
+ "eval_runtime": 56.2435,
+ "eval_samples_per_second": 43.418,
+ "eval_steps_per_second": 1.369,
+ "step": 3312
+ },
+ {
+ "epoch": 46.01397379912664,
+ "grad_norm": 0.06943798810243607,
+ "learning_rate": 0.0006,
+ "loss": 5.854922294616699,
+ "step": 3313
+ },
+ {
+ "epoch": 46.02794759825328,
+ "grad_norm": 0.05236278846859932,
+ "learning_rate": 0.0006,
+ "loss": 5.876611232757568,
+ "step": 3314
+ },
+ {
+ "epoch": 46.041921397379916,
+ "grad_norm": 0.06240001320838928,
+ "learning_rate": 0.0006,
+ "loss": 5.878912448883057,
+ "step": 3315
+ },
+ {
+ "epoch": 46.05589519650655,
+ "grad_norm": 0.07401192933320999,
+ "learning_rate": 0.0006,
+ "loss": 5.985136032104492,
+ "step": 3316
+ },
+ {
+ "epoch": 46.069868995633186,
+ "grad_norm": 0.0826529711484909,
+ "learning_rate": 0.0006,
+ "loss": 5.939439296722412,
+ "step": 3317
+ },
+ {
+ "epoch": 46.083842794759825,
+ "grad_norm": 0.07999005168676376,
+ "learning_rate": 0.0006,
+ "loss": 5.982336044311523,
+ "step": 3318
+ },
+ {
+ "epoch": 46.09781659388646,
+ "grad_norm": 0.07014869153499603,
+ "learning_rate": 0.0006,
+ "loss": 5.896534442901611,
+ "step": 3319
+ },
+ {
+ "epoch": 46.1117903930131,
+ "grad_norm": 0.05795128643512726,
+ "learning_rate": 0.0006,
+ "loss": 5.843111991882324,
+ "step": 3320
+ },
+ {
+ "epoch": 46.12576419213974,
+ "grad_norm": 0.06374258548021317,
+ "learning_rate": 0.0006,
+ "loss": 5.78752326965332,
+ "step": 3321
+ },
+ {
+ "epoch": 46.13973799126637,
+ "grad_norm": 0.045524805784225464,
+ "learning_rate": 0.0006,
+ "loss": 5.86755895614624,
+ "step": 3322
+ },
+ {
+ "epoch": 46.15371179039301,
+ "grad_norm": 0.03773504123091698,
+ "learning_rate": 0.0006,
+ "loss": 5.83627462387085,
+ "step": 3323
+ },
+ {
+ "epoch": 46.16768558951965,
+ "grad_norm": 0.035815201699733734,
+ "learning_rate": 0.0006,
+ "loss": 5.798238277435303,
+ "step": 3324
+ },
+ {
+ "epoch": 46.18165938864629,
+ "grad_norm": 0.03756783902645111,
+ "learning_rate": 0.0006,
+ "loss": 5.845197677612305,
+ "step": 3325
+ },
+ {
+ "epoch": 46.19563318777293,
+ "grad_norm": 0.03569113463163376,
+ "learning_rate": 0.0006,
+ "loss": 5.888932704925537,
+ "step": 3326
+ },
+ {
+ "epoch": 46.209606986899566,
+ "grad_norm": 0.029467813670635223,
+ "learning_rate": 0.0006,
+ "loss": 5.792047500610352,
+ "step": 3327
+ },
+ {
+ "epoch": 46.223580786026204,
+ "grad_norm": 0.025698406621813774,
+ "learning_rate": 0.0006,
+ "loss": 5.783734321594238,
+ "step": 3328
+ },
+ {
+ "epoch": 46.237554585152836,
+ "grad_norm": 0.028426671400666237,
+ "learning_rate": 0.0006,
+ "loss": 5.762084007263184,
+ "step": 3329
+ },
+ {
+ "epoch": 46.251528384279474,
+ "grad_norm": 0.03099622018635273,
+ "learning_rate": 0.0006,
+ "loss": 5.719109058380127,
+ "step": 3330
+ },
+ {
+ "epoch": 46.26550218340611,
+ "grad_norm": 0.0310219656676054,
+ "learning_rate": 0.0006,
+ "loss": 5.737057209014893,
+ "step": 3331
+ },
+ {
+ "epoch": 46.27947598253275,
+ "grad_norm": 0.02713446132838726,
+ "learning_rate": 0.0006,
+ "loss": 5.758657455444336,
+ "step": 3332
+ },
+ {
+ "epoch": 46.29344978165939,
+ "grad_norm": 0.03099844604730606,
+ "learning_rate": 0.0006,
+ "loss": 5.742086410522461,
+ "step": 3333
+ },
+ {
+ "epoch": 46.30742358078603,
+ "grad_norm": 0.030258730053901672,
+ "learning_rate": 0.0006,
+ "loss": 5.767426490783691,
+ "step": 3334
+ },
+ {
+ "epoch": 46.32139737991266,
+ "grad_norm": 0.029034852981567383,
+ "learning_rate": 0.0006,
+ "loss": 5.651792049407959,
+ "step": 3335
+ },
+ {
+ "epoch": 46.3353711790393,
+ "grad_norm": 0.03138665854930878,
+ "learning_rate": 0.0006,
+ "loss": 5.664980888366699,
+ "step": 3336
+ },
+ {
+ "epoch": 46.34934497816594,
+ "grad_norm": 0.040546901524066925,
+ "learning_rate": 0.0006,
+ "loss": 5.665165901184082,
+ "step": 3337
+ },
+ {
+ "epoch": 46.36331877729258,
+ "grad_norm": 0.06395485252141953,
+ "learning_rate": 0.0006,
+ "loss": 5.7776713371276855,
+ "step": 3338
+ },
+ {
+ "epoch": 46.377292576419215,
+ "grad_norm": 0.10088056325912476,
+ "learning_rate": 0.0006,
+ "loss": 5.755645751953125,
+ "step": 3339
+ },
+ {
+ "epoch": 46.391266375545854,
+ "grad_norm": 0.09527938067913055,
+ "learning_rate": 0.0006,
+ "loss": 5.838809490203857,
+ "step": 3340
+ },
+ {
+ "epoch": 46.40524017467249,
+ "grad_norm": 0.0849648267030716,
+ "learning_rate": 0.0006,
+ "loss": 5.781932830810547,
+ "step": 3341
+ },
+ {
+ "epoch": 46.419213973799124,
+ "grad_norm": 0.10142666101455688,
+ "learning_rate": 0.0006,
+ "loss": 5.838970184326172,
+ "step": 3342
+ },
+ {
+ "epoch": 46.43318777292576,
+ "grad_norm": 0.42731454968452454,
+ "learning_rate": 0.0006,
+ "loss": 6.071902275085449,
+ "step": 3343
+ },
+ {
+ "epoch": 46.4471615720524,
+ "grad_norm": 0.13102488219738007,
+ "learning_rate": 0.0006,
+ "loss": 6.8635993003845215,
+ "step": 3344
+ },
+ {
+ "epoch": 46.46113537117904,
+ "grad_norm": 0.23691938817501068,
+ "learning_rate": 0.0006,
+ "loss": 7.313158988952637,
+ "step": 3345
+ },
+ {
+ "epoch": 46.47510917030568,
+ "grad_norm": 0.11670669168233871,
+ "learning_rate": 0.0006,
+ "loss": 7.240363121032715,
+ "step": 3346
+ },
+ {
+ "epoch": 46.48908296943232,
+ "grad_norm": 0.09985820949077606,
+ "learning_rate": 0.0006,
+ "loss": 7.113332271575928,
+ "step": 3347
+ },
+ {
+ "epoch": 46.50305676855895,
+ "grad_norm": 0.11406408250331879,
+ "learning_rate": 0.0006,
+ "loss": 7.051285743713379,
+ "step": 3348
+ },
+ {
+ "epoch": 46.51703056768559,
+ "grad_norm": 0.07900336384773254,
+ "learning_rate": 0.0006,
+ "loss": 7.014365196228027,
+ "step": 3349
+ },
+ {
+ "epoch": 46.531004366812226,
+ "grad_norm": 0.05380333960056305,
+ "learning_rate": 0.0006,
+ "loss": 6.8868818283081055,
+ "step": 3350
+ },
+ {
+ "epoch": 46.544978165938865,
+ "grad_norm": 0.05868449807167053,
+ "learning_rate": 0.0006,
+ "loss": 6.783599853515625,
+ "step": 3351
+ },
+ {
+ "epoch": 46.5589519650655,
+ "grad_norm": 0.04428785294294357,
+ "learning_rate": 0.0006,
+ "loss": 6.798511505126953,
+ "step": 3352
+ },
+ {
+ "epoch": 46.57292576419214,
+ "grad_norm": 0.055527277290821075,
+ "learning_rate": 0.0006,
+ "loss": 6.694507598876953,
+ "step": 3353
+ },
+ {
+ "epoch": 46.58689956331878,
+ "grad_norm": 0.03872432932257652,
+ "learning_rate": 0.0006,
+ "loss": 6.736631393432617,
+ "step": 3354
+ },
+ {
+ "epoch": 46.60087336244541,
+ "grad_norm": 0.04330779239535332,
+ "learning_rate": 0.0006,
+ "loss": 6.576806545257568,
+ "step": 3355
+ },
+ {
+ "epoch": 46.61484716157205,
+ "grad_norm": 0.036767736077308655,
+ "learning_rate": 0.0006,
+ "loss": 6.559871196746826,
+ "step": 3356
+ },
+ {
+ "epoch": 46.62882096069869,
+ "grad_norm": 0.039857108145952225,
+ "learning_rate": 0.0006,
+ "loss": 6.5749831199646,
+ "step": 3357
+ },
+ {
+ "epoch": 46.64279475982533,
+ "grad_norm": 0.048719972372055054,
+ "learning_rate": 0.0006,
+ "loss": 6.4915618896484375,
+ "step": 3358
+ },
+ {
+ "epoch": 46.65676855895197,
+ "grad_norm": 0.042849015444517136,
+ "learning_rate": 0.0006,
+ "loss": 6.413516044616699,
+ "step": 3359
+ },
+ {
+ "epoch": 46.670742358078606,
+ "grad_norm": 0.02871009334921837,
+ "learning_rate": 0.0006,
+ "loss": 6.320952415466309,
+ "step": 3360
+ },
+ {
+ "epoch": 46.68471615720524,
+ "grad_norm": 0.03193620592355728,
+ "learning_rate": 0.0006,
+ "loss": 6.4176435470581055,
+ "step": 3361
+ },
+ {
+ "epoch": 46.698689956331876,
+ "grad_norm": 0.027417024597525597,
+ "learning_rate": 0.0006,
+ "loss": 6.342319488525391,
+ "step": 3362
+ },
+ {
+ "epoch": 46.712663755458514,
+ "grad_norm": 0.03305937349796295,
+ "learning_rate": 0.0006,
+ "loss": 6.246112823486328,
+ "step": 3363
+ },
+ {
+ "epoch": 46.72663755458515,
+ "grad_norm": 0.031403183937072754,
+ "learning_rate": 0.0006,
+ "loss": 6.149503231048584,
+ "step": 3364
+ },
+ {
+ "epoch": 46.74061135371179,
+ "grad_norm": 0.031563788652420044,
+ "learning_rate": 0.0006,
+ "loss": 6.278808116912842,
+ "step": 3365
+ },
+ {
+ "epoch": 46.75458515283843,
+ "grad_norm": 0.028366589918732643,
+ "learning_rate": 0.0006,
+ "loss": 6.057036876678467,
+ "step": 3366
+ },
+ {
+ "epoch": 46.76855895196506,
+ "grad_norm": 0.023715803399682045,
+ "learning_rate": 0.0006,
+ "loss": 6.111324310302734,
+ "step": 3367
+ },
+ {
+ "epoch": 46.7825327510917,
+ "grad_norm": 0.026546811684966087,
+ "learning_rate": 0.0006,
+ "loss": 6.105669975280762,
+ "step": 3368
+ },
+ {
+ "epoch": 46.79650655021834,
+ "grad_norm": 0.023531019687652588,
+ "learning_rate": 0.0006,
+ "loss": 6.068072319030762,
+ "step": 3369
+ },
+ {
+ "epoch": 46.81048034934498,
+ "grad_norm": 0.024480128660798073,
+ "learning_rate": 0.0006,
+ "loss": 5.947896957397461,
+ "step": 3370
+ },
+ {
+ "epoch": 46.82445414847162,
+ "grad_norm": 0.02517310529947281,
+ "learning_rate": 0.0006,
+ "loss": 6.030037879943848,
+ "step": 3371
+ },
+ {
+ "epoch": 46.838427947598255,
+ "grad_norm": 0.026723403483629227,
+ "learning_rate": 0.0006,
+ "loss": 5.963776111602783,
+ "step": 3372
+ },
+ {
+ "epoch": 46.852401746724894,
+ "grad_norm": 0.03163504600524902,
+ "learning_rate": 0.0006,
+ "loss": 5.893466949462891,
+ "step": 3373
+ },
+ {
+ "epoch": 46.866375545851525,
+ "grad_norm": 0.04058683291077614,
+ "learning_rate": 0.0006,
+ "loss": 5.9202799797058105,
+ "step": 3374
+ },
+ {
+ "epoch": 46.880349344978164,
+ "grad_norm": 0.04964117333292961,
+ "learning_rate": 0.0006,
+ "loss": 5.9366559982299805,
+ "step": 3375
+ },
+ {
+ "epoch": 46.8943231441048,
+ "grad_norm": 0.055406033992767334,
+ "learning_rate": 0.0006,
+ "loss": 5.912230968475342,
+ "step": 3376
+ },
+ {
+ "epoch": 46.90829694323144,
+ "grad_norm": 0.04643625393509865,
+ "learning_rate": 0.0006,
+ "loss": 5.87260103225708,
+ "step": 3377
+ },
+ {
+ "epoch": 46.92227074235808,
+ "grad_norm": 0.029325876384973526,
+ "learning_rate": 0.0006,
+ "loss": 5.854763984680176,
+ "step": 3378
+ },
+ {
+ "epoch": 46.93624454148472,
+ "grad_norm": 0.03159382566809654,
+ "learning_rate": 0.0006,
+ "loss": 5.880882263183594,
+ "step": 3379
+ },
+ {
+ "epoch": 46.95021834061135,
+ "grad_norm": 0.030224012210965157,
+ "learning_rate": 0.0006,
+ "loss": 5.787360191345215,
+ "step": 3380
+ },
+ {
+ "epoch": 46.96419213973799,
+ "grad_norm": 0.028368130326271057,
+ "learning_rate": 0.0006,
+ "loss": 5.817195892333984,
+ "step": 3381
+ },
+ {
+ "epoch": 46.97816593886463,
+ "grad_norm": 0.029411399737000465,
+ "learning_rate": 0.0006,
+ "loss": 5.783403396606445,
+ "step": 3382
+ },
+ {
+ "epoch": 46.992139737991266,
+ "grad_norm": 0.029927462339401245,
+ "learning_rate": 0.0006,
+ "loss": 5.662444591522217,
+ "step": 3383
+ },
+ {
+ "epoch": 47.0,
+ "grad_norm": 0.02872961387038231,
+ "learning_rate": 0.0006,
+ "loss": 5.742781639099121,
+ "step": 3384
+ },
+ {
+ "epoch": 47.0,
+ "eval_loss": 5.870260238647461,
+ "eval_runtime": 57.3734,
+ "eval_samples_per_second": 42.563,
+ "eval_steps_per_second": 1.342,
+ "step": 3384
+ },
+ {
+ "epoch": 47.01397379912664,
+ "grad_norm": 0.028542131185531616,
+ "learning_rate": 0.0006,
+ "loss": 5.753256797790527,
+ "step": 3385
+ },
+ {
+ "epoch": 47.02794759825328,
+ "grad_norm": 0.03227119892835617,
+ "learning_rate": 0.0006,
+ "loss": 5.722246170043945,
+ "step": 3386
+ },
+ {
+ "epoch": 47.041921397379916,
+ "grad_norm": 0.031461723148822784,
+ "learning_rate": 0.0006,
+ "loss": 5.740520000457764,
+ "step": 3387
+ },
+ {
+ "epoch": 47.05589519650655,
+ "grad_norm": 0.03184220567345619,
+ "learning_rate": 0.0006,
+ "loss": 5.720890045166016,
+ "step": 3388
+ },
+ {
+ "epoch": 47.069868995633186,
+ "grad_norm": 0.030015580356121063,
+ "learning_rate": 0.0006,
+ "loss": 5.597814559936523,
+ "step": 3389
+ },
+ {
+ "epoch": 47.083842794759825,
+ "grad_norm": 0.029931779950857162,
+ "learning_rate": 0.0006,
+ "loss": 5.5412468910217285,
+ "step": 3390
+ },
+ {
+ "epoch": 47.09781659388646,
+ "grad_norm": 0.03206505626440048,
+ "learning_rate": 0.0006,
+ "loss": 5.6967597007751465,
+ "step": 3391
+ },
+ {
+ "epoch": 47.1117903930131,
+ "grad_norm": 0.03978161886334419,
+ "learning_rate": 0.0006,
+ "loss": 5.59661340713501,
+ "step": 3392
+ },
+ {
+ "epoch": 47.12576419213974,
+ "grad_norm": 0.04619080200791359,
+ "learning_rate": 0.0006,
+ "loss": 5.795788764953613,
+ "step": 3393
+ },
+ {
+ "epoch": 47.13973799126637,
+ "grad_norm": 0.049085382372140884,
+ "learning_rate": 0.0006,
+ "loss": 5.667335510253906,
+ "step": 3394
+ },
+ {
+ "epoch": 47.15371179039301,
+ "grad_norm": 0.04830858111381531,
+ "learning_rate": 0.0006,
+ "loss": 5.668806552886963,
+ "step": 3395
+ },
+ {
+ "epoch": 47.16768558951965,
+ "grad_norm": 0.0583178736269474,
+ "learning_rate": 0.0006,
+ "loss": 5.692285537719727,
+ "step": 3396
+ },
+ {
+ "epoch": 47.18165938864629,
+ "grad_norm": 0.05182253196835518,
+ "learning_rate": 0.0006,
+ "loss": 5.678645133972168,
+ "step": 3397
+ },
+ {
+ "epoch": 47.19563318777293,
+ "grad_norm": 0.04011905938386917,
+ "learning_rate": 0.0006,
+ "loss": 5.661550998687744,
+ "step": 3398
+ },
+ {
+ "epoch": 47.209606986899566,
+ "grad_norm": 0.03888486698269844,
+ "learning_rate": 0.0006,
+ "loss": 5.643001556396484,
+ "step": 3399
+ },
+ {
+ "epoch": 47.223580786026204,
+ "grad_norm": 0.030658265575766563,
+ "learning_rate": 0.0006,
+ "loss": 5.659090995788574,
+ "step": 3400
+ },
+ {
+ "epoch": 47.237554585152836,
+ "grad_norm": 0.03609256446361542,
+ "learning_rate": 0.0006,
+ "loss": 5.544501781463623,
+ "step": 3401
+ },
+ {
+ "epoch": 47.251528384279474,
+ "grad_norm": 0.033589769154787064,
+ "learning_rate": 0.0006,
+ "loss": 5.463878631591797,
+ "step": 3402
+ },
+ {
+ "epoch": 47.26550218340611,
+ "grad_norm": 0.035128287971019745,
+ "learning_rate": 0.0006,
+ "loss": 5.5680060386657715,
+ "step": 3403
+ },
+ {
+ "epoch": 47.27947598253275,
+ "grad_norm": 0.038684550672769547,
+ "learning_rate": 0.0006,
+ "loss": 5.587898254394531,
+ "step": 3404
+ },
+ {
+ "epoch": 47.29344978165939,
+ "grad_norm": 0.043566226959228516,
+ "learning_rate": 0.0006,
+ "loss": 5.580896854400635,
+ "step": 3405
+ },
+ {
+ "epoch": 47.30742358078603,
+ "grad_norm": 0.04428374394774437,
+ "learning_rate": 0.0006,
+ "loss": 5.603460311889648,
+ "step": 3406
+ },
+ {
+ "epoch": 47.32139737991266,
+ "grad_norm": 0.037368934601545334,
+ "learning_rate": 0.0006,
+ "loss": 5.564740180969238,
+ "step": 3407
+ },
+ {
+ "epoch": 47.3353711790393,
+ "grad_norm": 0.04092350974678993,
+ "learning_rate": 0.0006,
+ "loss": 5.584153652191162,
+ "step": 3408
+ },
+ {
+ "epoch": 47.34934497816594,
+ "grad_norm": 0.045031506568193436,
+ "learning_rate": 0.0006,
+ "loss": 5.601186275482178,
+ "step": 3409
+ },
+ {
+ "epoch": 47.36331877729258,
+ "grad_norm": 0.041830409318208694,
+ "learning_rate": 0.0006,
+ "loss": 5.512788772583008,
+ "step": 3410
+ },
+ {
+ "epoch": 47.377292576419215,
+ "grad_norm": 0.037633974105119705,
+ "learning_rate": 0.0006,
+ "loss": 5.503828525543213,
+ "step": 3411
+ },
+ {
+ "epoch": 47.391266375545854,
+ "grad_norm": 0.03473588079214096,
+ "learning_rate": 0.0006,
+ "loss": 5.570557594299316,
+ "step": 3412
+ },
+ {
+ "epoch": 47.40524017467249,
+ "grad_norm": 0.039325181394815445,
+ "learning_rate": 0.0006,
+ "loss": 5.427783966064453,
+ "step": 3413
+ },
+ {
+ "epoch": 47.419213973799124,
+ "grad_norm": 0.04034624621272087,
+ "learning_rate": 0.0006,
+ "loss": 5.572591304779053,
+ "step": 3414
+ },
+ {
+ "epoch": 47.43318777292576,
+ "grad_norm": 0.05245184153318405,
+ "learning_rate": 0.0006,
+ "loss": 5.505208969116211,
+ "step": 3415
+ },
+ {
+ "epoch": 47.4471615720524,
+ "grad_norm": 0.0568675734102726,
+ "learning_rate": 0.0006,
+ "loss": 5.589373588562012,
+ "step": 3416
+ },
+ {
+ "epoch": 47.46113537117904,
+ "grad_norm": 0.048336587846279144,
+ "learning_rate": 0.0006,
+ "loss": 5.522005081176758,
+ "step": 3417
+ },
+ {
+ "epoch": 47.47510917030568,
+ "grad_norm": 0.04194975271821022,
+ "learning_rate": 0.0006,
+ "loss": 5.387576103210449,
+ "step": 3418
+ },
+ {
+ "epoch": 47.48908296943232,
+ "grad_norm": 0.04019433259963989,
+ "learning_rate": 0.0006,
+ "loss": 5.522615432739258,
+ "step": 3419
+ },
+ {
+ "epoch": 47.50305676855895,
+ "grad_norm": 0.05709933117032051,
+ "learning_rate": 0.0006,
+ "loss": 5.420286178588867,
+ "step": 3420
+ },
+ {
+ "epoch": 47.51703056768559,
+ "grad_norm": 0.09130338579416275,
+ "learning_rate": 0.0006,
+ "loss": 5.484622478485107,
+ "step": 3421
+ },
+ {
+ "epoch": 47.531004366812226,
+ "grad_norm": 0.2757541835308075,
+ "learning_rate": 0.0006,
+ "loss": 5.821832656860352,
+ "step": 3422
+ },
+ {
+ "epoch": 47.544978165938865,
+ "grad_norm": 0.28292080760002136,
+ "learning_rate": 0.0006,
+ "loss": 6.579934120178223,
+ "step": 3423
+ },
+ {
+ "epoch": 47.5589519650655,
+ "grad_norm": 0.15293604135513306,
+ "learning_rate": 0.0006,
+ "loss": 6.478213310241699,
+ "step": 3424
+ },
+ {
+ "epoch": 47.57292576419214,
+ "grad_norm": 0.11284393817186356,
+ "learning_rate": 0.0006,
+ "loss": 6.4066386222839355,
+ "step": 3425
+ },
+ {
+ "epoch": 47.58689956331878,
+ "grad_norm": 0.10933073610067368,
+ "learning_rate": 0.0006,
+ "loss": 6.3306169509887695,
+ "step": 3426
+ },
+ {
+ "epoch": 47.60087336244541,
+ "grad_norm": 0.05837641656398773,
+ "learning_rate": 0.0006,
+ "loss": 6.250000953674316,
+ "step": 3427
+ },
+ {
+ "epoch": 47.61484716157205,
+ "grad_norm": 0.05987229570746422,
+ "learning_rate": 0.0006,
+ "loss": 6.152463436126709,
+ "step": 3428
+ },
+ {
+ "epoch": 47.62882096069869,
+ "grad_norm": 0.04064938426017761,
+ "learning_rate": 0.0006,
+ "loss": 6.117897033691406,
+ "step": 3429
+ },
+ {
+ "epoch": 47.64279475982533,
+ "grad_norm": 0.04565057530999184,
+ "learning_rate": 0.0006,
+ "loss": 6.042291641235352,
+ "step": 3430
+ },
+ {
+ "epoch": 47.65676855895197,
+ "grad_norm": 0.03475063294172287,
+ "learning_rate": 0.0006,
+ "loss": 6.0071868896484375,
+ "step": 3431
+ },
+ {
+ "epoch": 47.670742358078606,
+ "grad_norm": 0.03809220343828201,
+ "learning_rate": 0.0006,
+ "loss": 6.0128021240234375,
+ "step": 3432
+ },
+ {
+ "epoch": 47.68471615720524,
+ "grad_norm": 0.03307219594717026,
+ "learning_rate": 0.0006,
+ "loss": 5.894664287567139,
+ "step": 3433
+ },
+ {
+ "epoch": 47.698689956331876,
+ "grad_norm": 0.028824346140027046,
+ "learning_rate": 0.0006,
+ "loss": 5.914865493774414,
+ "step": 3434
+ },
+ {
+ "epoch": 47.712663755458514,
+ "grad_norm": 0.027923690155148506,
+ "learning_rate": 0.0006,
+ "loss": 5.851759910583496,
+ "step": 3435
+ },
+ {
+ "epoch": 47.72663755458515,
+ "grad_norm": 0.025492684915661812,
+ "learning_rate": 0.0006,
+ "loss": 5.749186038970947,
+ "step": 3436
+ },
+ {
+ "epoch": 47.74061135371179,
+ "grad_norm": 0.025105174630880356,
+ "learning_rate": 0.0006,
+ "loss": 5.803566932678223,
+ "step": 3437
+ },
+ {
+ "epoch": 47.75458515283843,
+ "grad_norm": 0.023298880085349083,
+ "learning_rate": 0.0006,
+ "loss": 5.76276159286499,
+ "step": 3438
+ },
+ {
+ "epoch": 47.76855895196506,
+ "grad_norm": 0.02468576282262802,
+ "learning_rate": 0.0006,
+ "loss": 5.833852767944336,
+ "step": 3439
+ },
+ {
+ "epoch": 47.7825327510917,
+ "grad_norm": 0.022730689495801926,
+ "learning_rate": 0.0006,
+ "loss": 5.736724376678467,
+ "step": 3440
+ },
+ {
+ "epoch": 47.79650655021834,
+ "grad_norm": 0.021335354074835777,
+ "learning_rate": 0.0006,
+ "loss": 5.644248008728027,
+ "step": 3441
+ },
+ {
+ "epoch": 47.81048034934498,
+ "grad_norm": 0.02288101427257061,
+ "learning_rate": 0.0006,
+ "loss": 5.7011942863464355,
+ "step": 3442
+ },
+ {
+ "epoch": 47.82445414847162,
+ "grad_norm": 0.01961585320532322,
+ "learning_rate": 0.0006,
+ "loss": 5.63215446472168,
+ "step": 3443
+ },
+ {
+ "epoch": 47.838427947598255,
+ "grad_norm": 0.017659205943346024,
+ "learning_rate": 0.0006,
+ "loss": 5.744845390319824,
+ "step": 3444
+ },
+ {
+ "epoch": 47.852401746724894,
+ "grad_norm": 0.02281784638762474,
+ "learning_rate": 0.0006,
+ "loss": 5.696993350982666,
+ "step": 3445
+ },
+ {
+ "epoch": 47.866375545851525,
+ "grad_norm": 0.016394242644309998,
+ "learning_rate": 0.0006,
+ "loss": 5.659870147705078,
+ "step": 3446
+ },
+ {
+ "epoch": 47.880349344978164,
+ "grad_norm": 0.01967368647456169,
+ "learning_rate": 0.0006,
+ "loss": 5.638072490692139,
+ "step": 3447
+ },
+ {
+ "epoch": 47.8943231441048,
+ "grad_norm": 0.017326267436146736,
+ "learning_rate": 0.0006,
+ "loss": 5.638836860656738,
+ "step": 3448
+ },
+ {
+ "epoch": 47.90829694323144,
+ "grad_norm": 0.017778463661670685,
+ "learning_rate": 0.0006,
+ "loss": 5.634466648101807,
+ "step": 3449
+ },
+ {
+ "epoch": 47.92227074235808,
+ "grad_norm": 0.017269758507609367,
+ "learning_rate": 0.0006,
+ "loss": 5.506277561187744,
+ "step": 3450
+ },
+ {
+ "epoch": 47.93624454148472,
+ "grad_norm": 0.01733289286494255,
+ "learning_rate": 0.0006,
+ "loss": 5.544455528259277,
+ "step": 3451
+ },
+ {
+ "epoch": 47.95021834061135,
+ "grad_norm": 0.017787795513868332,
+ "learning_rate": 0.0006,
+ "loss": 5.498943328857422,
+ "step": 3452
+ },
+ {
+ "epoch": 47.96419213973799,
+ "grad_norm": 0.015770932659506798,
+ "learning_rate": 0.0006,
+ "loss": 5.537227630615234,
+ "step": 3453
+ },
+ {
+ "epoch": 47.97816593886463,
+ "grad_norm": 0.016132809221744537,
+ "learning_rate": 0.0006,
+ "loss": 5.489081382751465,
+ "step": 3454
+ },
+ {
+ "epoch": 47.992139737991266,
+ "grad_norm": 0.01749664731323719,
+ "learning_rate": 0.0006,
+ "loss": 5.454263687133789,
+ "step": 3455
+ },
+ {
+ "epoch": 48.0,
+ "grad_norm": 0.016711296513676643,
+ "learning_rate": 0.0006,
+ "loss": 5.530783653259277,
+ "step": 3456
+ },
+ {
+ "epoch": 48.0,
+ "eval_loss": 5.596925258636475,
+ "eval_runtime": 56.9612,
+ "eval_samples_per_second": 42.871,
+ "eval_steps_per_second": 1.352,
+ "step": 3456
+ },
+ {
+ "epoch": 48.01397379912664,
+ "grad_norm": 0.01867641694843769,
+ "learning_rate": 0.0006,
+ "loss": 5.506853103637695,
+ "step": 3457
+ },
+ {
+ "epoch": 48.02794759825328,
+ "grad_norm": 0.023250792175531387,
+ "learning_rate": 0.0006,
+ "loss": 5.4074177742004395,
+ "step": 3458
+ },
+ {
+ "epoch": 48.041921397379916,
+ "grad_norm": 0.01970936730504036,
+ "learning_rate": 0.0006,
+ "loss": 5.508777618408203,
+ "step": 3459
+ },
+ {
+ "epoch": 48.05589519650655,
+ "grad_norm": 0.01879260502755642,
+ "learning_rate": 0.0006,
+ "loss": 5.412424087524414,
+ "step": 3460
+ },
+ {
+ "epoch": 48.069868995633186,
+ "grad_norm": 0.017420092597603798,
+ "learning_rate": 0.0006,
+ "loss": 5.43361759185791,
+ "step": 3461
+ },
+ {
+ "epoch": 48.083842794759825,
+ "grad_norm": 0.018707862123847008,
+ "learning_rate": 0.0006,
+ "loss": 5.476800918579102,
+ "step": 3462
+ },
+ {
+ "epoch": 48.09781659388646,
+ "grad_norm": 0.023185882717370987,
+ "learning_rate": 0.0006,
+ "loss": 5.465272426605225,
+ "step": 3463
+ },
+ {
+ "epoch": 48.1117903930131,
+ "grad_norm": 0.023570368066430092,
+ "learning_rate": 0.0006,
+ "loss": 5.391471862792969,
+ "step": 3464
+ },
+ {
+ "epoch": 48.12576419213974,
+ "grad_norm": 0.02651144750416279,
+ "learning_rate": 0.0006,
+ "loss": 5.422799110412598,
+ "step": 3465
+ },
+ {
+ "epoch": 48.13973799126637,
+ "grad_norm": 0.03316329047083855,
+ "learning_rate": 0.0006,
+ "loss": 5.3952202796936035,
+ "step": 3466
+ },
+ {
+ "epoch": 48.15371179039301,
+ "grad_norm": 0.04012278467416763,
+ "learning_rate": 0.0006,
+ "loss": 5.540853023529053,
+ "step": 3467
+ },
+ {
+ "epoch": 48.16768558951965,
+ "grad_norm": 0.04763203114271164,
+ "learning_rate": 0.0006,
+ "loss": 5.387890338897705,
+ "step": 3468
+ },
+ {
+ "epoch": 48.18165938864629,
+ "grad_norm": 0.05129165202379227,
+ "learning_rate": 0.0006,
+ "loss": 5.414722442626953,
+ "step": 3469
+ },
+ {
+ "epoch": 48.19563318777293,
+ "grad_norm": 0.04438071325421333,
+ "learning_rate": 0.0006,
+ "loss": 5.386244773864746,
+ "step": 3470
+ },
+ {
+ "epoch": 48.209606986899566,
+ "grad_norm": 0.040234919637441635,
+ "learning_rate": 0.0006,
+ "loss": 5.431617736816406,
+ "step": 3471
+ },
+ {
+ "epoch": 48.223580786026204,
+ "grad_norm": 0.03217008337378502,
+ "learning_rate": 0.0006,
+ "loss": 5.30168342590332,
+ "step": 3472
+ },
+ {
+ "epoch": 48.237554585152836,
+ "grad_norm": 0.03087974339723587,
+ "learning_rate": 0.0006,
+ "loss": 5.316389083862305,
+ "step": 3473
+ },
+ {
+ "epoch": 48.251528384279474,
+ "grad_norm": 0.02789347991347313,
+ "learning_rate": 0.0006,
+ "loss": 5.385970592498779,
+ "step": 3474
+ },
+ {
+ "epoch": 48.26550218340611,
+ "grad_norm": 0.024792097508907318,
+ "learning_rate": 0.0006,
+ "loss": 5.362329959869385,
+ "step": 3475
+ },
+ {
+ "epoch": 48.27947598253275,
+ "grad_norm": 0.025869522243738174,
+ "learning_rate": 0.0006,
+ "loss": 5.28400182723999,
+ "step": 3476
+ },
+ {
+ "epoch": 48.29344978165939,
+ "grad_norm": 0.021339882165193558,
+ "learning_rate": 0.0006,
+ "loss": 5.36669921875,
+ "step": 3477
+ },
+ {
+ "epoch": 48.30742358078603,
+ "grad_norm": 0.021462520584464073,
+ "learning_rate": 0.0006,
+ "loss": 5.311222076416016,
+ "step": 3478
+ },
+ {
+ "epoch": 48.32139737991266,
+ "grad_norm": 0.02193879708647728,
+ "learning_rate": 0.0006,
+ "loss": 5.291441917419434,
+ "step": 3479
+ },
+ {
+ "epoch": 48.3353711790393,
+ "grad_norm": 0.01963093690574169,
+ "learning_rate": 0.0006,
+ "loss": 5.342780113220215,
+ "step": 3480
+ },
+ {
+ "epoch": 48.34934497816594,
+ "grad_norm": 0.020634736865758896,
+ "learning_rate": 0.0006,
+ "loss": 5.144141674041748,
+ "step": 3481
+ },
+ {
+ "epoch": 48.36331877729258,
+ "grad_norm": 0.02126968279480934,
+ "learning_rate": 0.0006,
+ "loss": 5.232624530792236,
+ "step": 3482
+ },
+ {
+ "epoch": 48.377292576419215,
+ "grad_norm": 0.021203994750976562,
+ "learning_rate": 0.0006,
+ "loss": 5.290219783782959,
+ "step": 3483
+ },
+ {
+ "epoch": 48.391266375545854,
+ "grad_norm": 0.02193225920200348,
+ "learning_rate": 0.0006,
+ "loss": 5.29197359085083,
+ "step": 3484
+ },
+ {
+ "epoch": 48.40524017467249,
+ "grad_norm": 0.023174211382865906,
+ "learning_rate": 0.0006,
+ "loss": 5.204349517822266,
+ "step": 3485
+ },
+ {
+ "epoch": 48.419213973799124,
+ "grad_norm": 0.024161914363503456,
+ "learning_rate": 0.0006,
+ "loss": 5.339351654052734,
+ "step": 3486
+ },
+ {
+ "epoch": 48.43318777292576,
+ "grad_norm": 0.02221524342894554,
+ "learning_rate": 0.0006,
+ "loss": 5.263001441955566,
+ "step": 3487
+ },
+ {
+ "epoch": 48.4471615720524,
+ "grad_norm": 0.026632830500602722,
+ "learning_rate": 0.0006,
+ "loss": 5.221281051635742,
+ "step": 3488
+ },
+ {
+ "epoch": 48.46113537117904,
+ "grad_norm": 0.034923870116472244,
+ "learning_rate": 0.0006,
+ "loss": 5.296039581298828,
+ "step": 3489
+ },
+ {
+ "epoch": 48.47510917030568,
+ "grad_norm": 0.046184979379177094,
+ "learning_rate": 0.0006,
+ "loss": 5.212769985198975,
+ "step": 3490
+ },
+ {
+ "epoch": 48.48908296943232,
+ "grad_norm": 0.05154098570346832,
+ "learning_rate": 0.0006,
+ "loss": 5.337864875793457,
+ "step": 3491
+ },
+ {
+ "epoch": 48.50305676855895,
+ "grad_norm": 0.042967259883880615,
+ "learning_rate": 0.0006,
+ "loss": 5.267958641052246,
+ "step": 3492
+ },
+ {
+ "epoch": 48.51703056768559,
+ "grad_norm": 0.05356583371758461,
+ "learning_rate": 0.0006,
+ "loss": 5.1404218673706055,
+ "step": 3493
+ },
+ {
+ "epoch": 48.531004366812226,
+ "grad_norm": 0.08219032734632492,
+ "learning_rate": 0.0006,
+ "loss": 5.300687789916992,
+ "step": 3494
+ },
+ {
+ "epoch": 48.544978165938865,
+ "grad_norm": 0.19204089045524597,
+ "learning_rate": 0.0006,
+ "loss": 5.456025123596191,
+ "step": 3495
+ },
+ {
+ "epoch": 48.5589519650655,
+ "grad_norm": 0.25917696952819824,
+ "learning_rate": 0.0006,
+ "loss": 6.156761169433594,
+ "step": 3496
+ },
+ {
+ "epoch": 48.57292576419214,
+ "grad_norm": 0.1566283404827118,
+ "learning_rate": 0.0006,
+ "loss": 5.862346649169922,
+ "step": 3497
+ },
+ {
+ "epoch": 48.58689956331878,
+ "grad_norm": 0.1761486828327179,
+ "learning_rate": 0.0006,
+ "loss": 6.012795448303223,
+ "step": 3498
+ },
+ {
+ "epoch": 48.60087336244541,
+ "grad_norm": 0.09857621043920517,
+ "learning_rate": 0.0006,
+ "loss": 5.9248151779174805,
+ "step": 3499
+ },
+ {
+ "epoch": 48.61484716157205,
+ "grad_norm": 0.05968055501580238,
+ "learning_rate": 0.0006,
+ "loss": 5.780706882476807,
+ "step": 3500
+ },
+ {
+ "epoch": 48.62882096069869,
+ "grad_norm": 0.057433564215898514,
+ "learning_rate": 0.0006,
+ "loss": 5.645388603210449,
+ "step": 3501
+ },
+ {
+ "epoch": 48.64279475982533,
+ "grad_norm": 0.039659615606069565,
+ "learning_rate": 0.0006,
+ "loss": 5.695138931274414,
+ "step": 3502
+ },
+ {
+ "epoch": 48.65676855895197,
+ "grad_norm": 0.04019695147871971,
+ "learning_rate": 0.0006,
+ "loss": 5.672435283660889,
+ "step": 3503
+ },
+ {
+ "epoch": 48.670742358078606,
+ "grad_norm": 0.03033018670976162,
+ "learning_rate": 0.0006,
+ "loss": 5.682260513305664,
+ "step": 3504
+ },
+ {
+ "epoch": 48.68471615720524,
+ "grad_norm": 0.03215787187218666,
+ "learning_rate": 0.0006,
+ "loss": 5.604182243347168,
+ "step": 3505
+ },
+ {
+ "epoch": 48.698689956331876,
+ "grad_norm": 0.026865126565098763,
+ "learning_rate": 0.0006,
+ "loss": 5.64040470123291,
+ "step": 3506
+ },
+ {
+ "epoch": 48.712663755458514,
+ "grad_norm": 0.02831609919667244,
+ "learning_rate": 0.0006,
+ "loss": 5.552720069885254,
+ "step": 3507
+ },
+ {
+ "epoch": 48.72663755458515,
+ "grad_norm": 0.024226613342761993,
+ "learning_rate": 0.0006,
+ "loss": 5.556946277618408,
+ "step": 3508
+ },
+ {
+ "epoch": 48.74061135371179,
+ "grad_norm": 0.025393934920430183,
+ "learning_rate": 0.0006,
+ "loss": 5.476178169250488,
+ "step": 3509
+ },
+ {
+ "epoch": 48.75458515283843,
+ "grad_norm": 0.01984412968158722,
+ "learning_rate": 0.0006,
+ "loss": 5.471088409423828,
+ "step": 3510
+ },
+ {
+ "epoch": 48.76855895196506,
+ "grad_norm": 0.02035129815340042,
+ "learning_rate": 0.0006,
+ "loss": 5.438812732696533,
+ "step": 3511
+ },
+ {
+ "epoch": 48.7825327510917,
+ "grad_norm": 0.0208350270986557,
+ "learning_rate": 0.0006,
+ "loss": 5.381746292114258,
+ "step": 3512
+ },
+ {
+ "epoch": 48.79650655021834,
+ "grad_norm": 0.017783882096409798,
+ "learning_rate": 0.0006,
+ "loss": 5.351675510406494,
+ "step": 3513
+ },
+ {
+ "epoch": 48.81048034934498,
+ "grad_norm": 0.01671770215034485,
+ "learning_rate": 0.0006,
+ "loss": 5.329428672790527,
+ "step": 3514
+ },
+ {
+ "epoch": 48.82445414847162,
+ "grad_norm": 0.017891215160489082,
+ "learning_rate": 0.0006,
+ "loss": 5.212074279785156,
+ "step": 3515
+ },
+ {
+ "epoch": 48.838427947598255,
+ "grad_norm": 0.016847791150212288,
+ "learning_rate": 0.0006,
+ "loss": 5.299534797668457,
+ "step": 3516
+ },
+ {
+ "epoch": 48.852401746724894,
+ "grad_norm": 0.01593637838959694,
+ "learning_rate": 0.0006,
+ "loss": 5.30466890335083,
+ "step": 3517
+ },
+ {
+ "epoch": 48.866375545851525,
+ "grad_norm": 0.014713102020323277,
+ "learning_rate": 0.0006,
+ "loss": 5.294345378875732,
+ "step": 3518
+ },
+ {
+ "epoch": 48.880349344978164,
+ "grad_norm": 0.01433857623487711,
+ "learning_rate": 0.0006,
+ "loss": 5.2148542404174805,
+ "step": 3519
+ },
+ {
+ "epoch": 48.8943231441048,
+ "grad_norm": 0.014979379251599312,
+ "learning_rate": 0.0006,
+ "loss": 5.298582077026367,
+ "step": 3520
+ },
+ {
+ "epoch": 48.90829694323144,
+ "grad_norm": 0.013494192622601986,
+ "learning_rate": 0.0006,
+ "loss": 5.309783935546875,
+ "step": 3521
+ },
+ {
+ "epoch": 48.92227074235808,
+ "grad_norm": 0.014020012691617012,
+ "learning_rate": 0.0006,
+ "loss": 5.284879684448242,
+ "step": 3522
+ },
+ {
+ "epoch": 48.93624454148472,
+ "grad_norm": 0.01348058134317398,
+ "learning_rate": 0.0006,
+ "loss": 5.249116897583008,
+ "step": 3523
+ },
+ {
+ "epoch": 48.95021834061135,
+ "grad_norm": 0.013280218467116356,
+ "learning_rate": 0.0006,
+ "loss": 5.273489952087402,
+ "step": 3524
+ },
+ {
+ "epoch": 48.96419213973799,
+ "grad_norm": 0.012647110037505627,
+ "learning_rate": 0.0006,
+ "loss": 5.306652069091797,
+ "step": 3525
+ },
+ {
+ "epoch": 48.97816593886463,
+ "grad_norm": 0.012847724370658398,
+ "learning_rate": 0.0006,
+ "loss": 5.235745429992676,
+ "step": 3526
+ },
+ {
+ "epoch": 48.992139737991266,
+ "grad_norm": 0.011820166371762753,
+ "learning_rate": 0.0006,
+ "loss": 5.238091468811035,
+ "step": 3527
+ },
+ {
+ "epoch": 49.0,
+ "grad_norm": 0.01255585066974163,
+ "learning_rate": 0.0006,
+ "loss": 5.231295585632324,
+ "step": 3528
+ },
+ {
+ "epoch": 49.0,
+ "eval_loss": 5.394587993621826,
+ "eval_runtime": 56.3734,
+ "eval_samples_per_second": 43.318,
+ "eval_steps_per_second": 1.366,
+ "step": 3528
+ },
+ {
+ "epoch": 49.01397379912664,
+ "grad_norm": 0.011031972244381905,
+ "learning_rate": 0.0006,
+ "loss": 5.1354570388793945,
+ "step": 3529
+ },
+ {
+ "epoch": 49.02794759825328,
+ "grad_norm": 0.010631263256072998,
+ "learning_rate": 0.0006,
+ "loss": 5.177639961242676,
+ "step": 3530
+ },
+ {
+ "epoch": 49.041921397379916,
+ "grad_norm": 0.01133775431662798,
+ "learning_rate": 0.0006,
+ "loss": 5.283428192138672,
+ "step": 3531
+ },
+ {
+ "epoch": 49.05589519650655,
+ "grad_norm": 0.01154172234237194,
+ "learning_rate": 0.0006,
+ "loss": 5.206364154815674,
+ "step": 3532
+ },
+ {
+ "epoch": 49.069868995633186,
+ "grad_norm": 0.010192356072366238,
+ "learning_rate": 0.0006,
+ "loss": 5.2385735511779785,
+ "step": 3533
+ },
+ {
+ "epoch": 49.083842794759825,
+ "grad_norm": 0.012789200060069561,
+ "learning_rate": 0.0006,
+ "loss": 5.071800231933594,
+ "step": 3534
+ },
+ {
+ "epoch": 49.09781659388646,
+ "grad_norm": 0.011282820254564285,
+ "learning_rate": 0.0006,
+ "loss": 5.218020439147949,
+ "step": 3535
+ },
+ {
+ "epoch": 49.1117903930131,
+ "grad_norm": 0.012153448536992073,
+ "learning_rate": 0.0006,
+ "loss": 5.134698867797852,
+ "step": 3536
+ },
+ {
+ "epoch": 49.12576419213974,
+ "grad_norm": 0.011920543387532234,
+ "learning_rate": 0.0006,
+ "loss": 5.141620635986328,
+ "step": 3537
+ },
+ {
+ "epoch": 49.13973799126637,
+ "grad_norm": 0.010949329473078251,
+ "learning_rate": 0.0006,
+ "loss": 5.116259574890137,
+ "step": 3538
+ },
+ {
+ "epoch": 49.15371179039301,
+ "grad_norm": 0.012362448498606682,
+ "learning_rate": 0.0006,
+ "loss": 5.090177059173584,
+ "step": 3539
+ },
+ {
+ "epoch": 49.16768558951965,
+ "grad_norm": 0.011575071141123772,
+ "learning_rate": 0.0006,
+ "loss": 5.093441009521484,
+ "step": 3540
+ },
+ {
+ "epoch": 49.18165938864629,
+ "grad_norm": 0.011126236990094185,
+ "learning_rate": 0.0006,
+ "loss": 5.060235023498535,
+ "step": 3541
+ },
+ {
+ "epoch": 49.19563318777293,
+ "grad_norm": 0.0107401292771101,
+ "learning_rate": 0.0006,
+ "loss": 5.140704154968262,
+ "step": 3542
+ },
+ {
+ "epoch": 49.209606986899566,
+ "grad_norm": 0.011202906258404255,
+ "learning_rate": 0.0006,
+ "loss": 5.10855770111084,
+ "step": 3543
+ },
+ {
+ "epoch": 49.223580786026204,
+ "grad_norm": 0.011083454824984074,
+ "learning_rate": 0.0006,
+ "loss": 5.092751502990723,
+ "step": 3544
+ },
+ {
+ "epoch": 49.237554585152836,
+ "grad_norm": 0.009523800574243069,
+ "learning_rate": 0.0006,
+ "loss": 5.13733434677124,
+ "step": 3545
+ },
+ {
+ "epoch": 49.251528384279474,
+ "grad_norm": 0.011298353783786297,
+ "learning_rate": 0.0006,
+ "loss": 5.036048889160156,
+ "step": 3546
+ },
+ {
+ "epoch": 49.26550218340611,
+ "grad_norm": 0.010676483623683453,
+ "learning_rate": 0.0006,
+ "loss": 5.051901340484619,
+ "step": 3547
+ },
+ {
+ "epoch": 49.27947598253275,
+ "grad_norm": 0.011463399045169353,
+ "learning_rate": 0.0006,
+ "loss": 4.983170986175537,
+ "step": 3548
+ },
+ {
+ "epoch": 49.29344978165939,
+ "grad_norm": 0.010603402741253376,
+ "learning_rate": 0.0006,
+ "loss": 5.112311363220215,
+ "step": 3549
+ },
+ {
+ "epoch": 49.30742358078603,
+ "grad_norm": 0.010535670444369316,
+ "learning_rate": 0.0006,
+ "loss": 5.085227966308594,
+ "step": 3550
+ },
+ {
+ "epoch": 49.32139737991266,
+ "grad_norm": 0.011843928135931492,
+ "learning_rate": 0.0006,
+ "loss": 5.07575798034668,
+ "step": 3551
+ },
+ {
+ "epoch": 49.3353711790393,
+ "grad_norm": 0.010587952099740505,
+ "learning_rate": 0.0006,
+ "loss": 5.019237041473389,
+ "step": 3552
+ },
+ {
+ "epoch": 49.34934497816594,
+ "grad_norm": 0.012285659089684486,
+ "learning_rate": 0.0006,
+ "loss": 5.144603729248047,
+ "step": 3553
+ },
+ {
+ "epoch": 49.36331877729258,
+ "grad_norm": 0.013119431212544441,
+ "learning_rate": 0.0006,
+ "loss": 5.035256862640381,
+ "step": 3554
+ },
+ {
+ "epoch": 49.377292576419215,
+ "grad_norm": 0.012138537131249905,
+ "learning_rate": 0.0006,
+ "loss": 4.996667861938477,
+ "step": 3555
+ },
+ {
+ "epoch": 49.391266375545854,
+ "grad_norm": 0.012142804451286793,
+ "learning_rate": 0.0006,
+ "loss": 4.95058536529541,
+ "step": 3556
+ },
+ {
+ "epoch": 49.40524017467249,
+ "grad_norm": 0.012205676175653934,
+ "learning_rate": 0.0006,
+ "loss": 5.096931457519531,
+ "step": 3557
+ },
+ {
+ "epoch": 49.419213973799124,
+ "grad_norm": 0.012086744420230389,
+ "learning_rate": 0.0006,
+ "loss": 4.99384880065918,
+ "step": 3558
+ },
+ {
+ "epoch": 49.43318777292576,
+ "grad_norm": 0.012918103486299515,
+ "learning_rate": 0.0006,
+ "loss": 5.031294822692871,
+ "step": 3559
+ },
+ {
+ "epoch": 49.4471615720524,
+ "grad_norm": 0.012006713077425957,
+ "learning_rate": 0.0006,
+ "loss": 5.0547943115234375,
+ "step": 3560
+ },
+ {
+ "epoch": 49.46113537117904,
+ "grad_norm": 0.0120399734005332,
+ "learning_rate": 0.0006,
+ "loss": 5.014238357543945,
+ "step": 3561
+ },
+ {
+ "epoch": 49.47510917030568,
+ "grad_norm": 0.013306990265846252,
+ "learning_rate": 0.0006,
+ "loss": 4.972925186157227,
+ "step": 3562
+ },
+ {
+ "epoch": 49.48908296943232,
+ "grad_norm": 0.013878636993467808,
+ "learning_rate": 0.0006,
+ "loss": 5.079033851623535,
+ "step": 3563
+ },
+ {
+ "epoch": 49.50305676855895,
+ "grad_norm": 0.012454566545784473,
+ "learning_rate": 0.0006,
+ "loss": 4.93708610534668,
+ "step": 3564
+ },
+ {
+ "epoch": 49.51703056768559,
+ "grad_norm": 0.014075061306357384,
+ "learning_rate": 0.0006,
+ "loss": 4.965582370758057,
+ "step": 3565
+ },
+ {
+ "epoch": 49.531004366812226,
+ "grad_norm": 0.012525617145001888,
+ "learning_rate": 0.0006,
+ "loss": 4.968148708343506,
+ "step": 3566
+ },
+ {
+ "epoch": 49.544978165938865,
+ "grad_norm": 0.012418225407600403,
+ "learning_rate": 0.0006,
+ "loss": 5.06901741027832,
+ "step": 3567
+ },
+ {
+ "epoch": 49.5589519650655,
+ "grad_norm": 0.015034239739179611,
+ "learning_rate": 0.0006,
+ "loss": 4.984277725219727,
+ "step": 3568
+ },
+ {
+ "epoch": 49.57292576419214,
+ "grad_norm": 0.014329628087580204,
+ "learning_rate": 0.0006,
+ "loss": 4.951558589935303,
+ "step": 3569
+ },
+ {
+ "epoch": 49.58689956331878,
+ "grad_norm": 0.017769070342183113,
+ "learning_rate": 0.0006,
+ "loss": 4.966532230377197,
+ "step": 3570
+ },
+ {
+ "epoch": 49.60087336244541,
+ "grad_norm": 0.01909453794360161,
+ "learning_rate": 0.0006,
+ "loss": 4.919890880584717,
+ "step": 3571
+ },
+ {
+ "epoch": 49.61484716157205,
+ "grad_norm": 0.017519688233733177,
+ "learning_rate": 0.0006,
+ "loss": 5.031286716461182,
+ "step": 3572
+ },
+ {
+ "epoch": 49.62882096069869,
+ "grad_norm": 0.017143897712230682,
+ "learning_rate": 0.0006,
+ "loss": 4.969764709472656,
+ "step": 3573
+ },
+ {
+ "epoch": 49.64279475982533,
+ "grad_norm": 0.01688285917043686,
+ "learning_rate": 0.0006,
+ "loss": 5.0733442306518555,
+ "step": 3574
+ },
+ {
+ "epoch": 49.65676855895197,
+ "grad_norm": 0.018547311425209045,
+ "learning_rate": 0.0006,
+ "loss": 4.980863571166992,
+ "step": 3575
+ },
+ {
+ "epoch": 49.670742358078606,
+ "grad_norm": 0.018877137452363968,
+ "learning_rate": 0.0006,
+ "loss": 4.954646110534668,
+ "step": 3576
+ },
+ {
+ "epoch": 49.68471615720524,
+ "grad_norm": 0.018681600689888,
+ "learning_rate": 0.0006,
+ "loss": 4.959702014923096,
+ "step": 3577
+ },
+ {
+ "epoch": 49.698689956331876,
+ "grad_norm": 0.018947336822748184,
+ "learning_rate": 0.0006,
+ "loss": 5.070502758026123,
+ "step": 3578
+ },
+ {
+ "epoch": 49.712663755458514,
+ "grad_norm": 0.018817484378814697,
+ "learning_rate": 0.0006,
+ "loss": 5.056508541107178,
+ "step": 3579
+ },
+ {
+ "epoch": 49.72663755458515,
+ "grad_norm": 0.01931724324822426,
+ "learning_rate": 0.0006,
+ "loss": 4.922451972961426,
+ "step": 3580
+ },
+ {
+ "epoch": 49.74061135371179,
+ "grad_norm": 0.022047819569706917,
+ "learning_rate": 0.0006,
+ "loss": 5.053170204162598,
+ "step": 3581
+ },
+ {
+ "epoch": 49.75458515283843,
+ "grad_norm": 0.02279801294207573,
+ "learning_rate": 0.0006,
+ "loss": 4.8768439292907715,
+ "step": 3582
+ },
+ {
+ "epoch": 49.76855895196506,
+ "grad_norm": 0.020661601796746254,
+ "learning_rate": 0.0006,
+ "loss": 4.907928466796875,
+ "step": 3583
+ },
+ {
+ "epoch": 49.7825327510917,
+ "grad_norm": 0.020479509606957436,
+ "learning_rate": 0.0006,
+ "loss": 5.040694236755371,
+ "step": 3584
+ },
+ {
+ "epoch": 49.79650655021834,
+ "grad_norm": 0.021654162555933,
+ "learning_rate": 0.0006,
+ "loss": 5.018434524536133,
+ "step": 3585
+ },
+ {
+ "epoch": 49.81048034934498,
+ "grad_norm": 0.022486824542284012,
+ "learning_rate": 0.0006,
+ "loss": 5.086538314819336,
+ "step": 3586
+ },
+ {
+ "epoch": 49.82445414847162,
+ "grad_norm": 0.02141369879245758,
+ "learning_rate": 0.0006,
+ "loss": 5.035694122314453,
+ "step": 3587
+ },
+ {
+ "epoch": 49.838427947598255,
+ "grad_norm": 0.015926677733659744,
+ "learning_rate": 0.0006,
+ "loss": 4.905038833618164,
+ "step": 3588
+ },
+ {
+ "epoch": 49.852401746724894,
+ "grad_norm": 0.015775298699736595,
+ "learning_rate": 0.0006,
+ "loss": 5.003902435302734,
+ "step": 3589
+ },
+ {
+ "epoch": 49.866375545851525,
+ "grad_norm": 0.01872118003666401,
+ "learning_rate": 0.0006,
+ "loss": 4.891060829162598,
+ "step": 3590
+ },
+ {
+ "epoch": 49.880349344978164,
+ "grad_norm": 0.02066192403435707,
+ "learning_rate": 0.0006,
+ "loss": 5.002897262573242,
+ "step": 3591
+ },
+ {
+ "epoch": 49.8943231441048,
+ "grad_norm": 0.02348714880645275,
+ "learning_rate": 0.0006,
+ "loss": 4.9189958572387695,
+ "step": 3592
+ },
+ {
+ "epoch": 49.90829694323144,
+ "grad_norm": 0.020402103662490845,
+ "learning_rate": 0.0006,
+ "loss": 4.992661476135254,
+ "step": 3593
+ },
+ {
+ "epoch": 49.92227074235808,
+ "grad_norm": 0.020466072484850883,
+ "learning_rate": 0.0006,
+ "loss": 4.962968826293945,
+ "step": 3594
+ },
+ {
+ "epoch": 49.93624454148472,
+ "grad_norm": 0.01922314241528511,
+ "learning_rate": 0.0006,
+ "loss": 4.897730827331543,
+ "step": 3595
+ },
+ {
+ "epoch": 49.95021834061135,
+ "grad_norm": 0.017228754237294197,
+ "learning_rate": 0.0006,
+ "loss": 4.964842796325684,
+ "step": 3596
+ },
+ {
+ "epoch": 49.96419213973799,
+ "grad_norm": 0.016021743416786194,
+ "learning_rate": 0.0006,
+ "loss": 4.910758972167969,
+ "step": 3597
+ },
+ {
+ "epoch": 49.97816593886463,
+ "grad_norm": 0.017985103651881218,
+ "learning_rate": 0.0006,
+ "loss": 4.916072368621826,
+ "step": 3598
+ },
+ {
+ "epoch": 49.992139737991266,
+ "grad_norm": 0.018780240789055824,
+ "learning_rate": 0.0006,
+ "loss": 4.94658088684082,
+ "step": 3599
+ },
+ {
+ "epoch": 50.0,
+ "grad_norm": 0.01776314526796341,
+ "learning_rate": 0.0006,
+ "loss": 4.969357967376709,
+ "step": 3600
+ },
+ {
+ "epoch": 50.0,
+ "eval_loss": 5.141158580780029,
+ "eval_runtime": 56.2332,
+ "eval_samples_per_second": 43.426,
+ "eval_steps_per_second": 1.369,
+ "step": 3600
+ },
+ {
+ "epoch": 50.01397379912664,
+ "grad_norm": 0.016308696940541267,
+ "learning_rate": 0.0006,
+ "loss": 4.924664497375488,
+ "step": 3601
+ },
+ {
+ "epoch": 50.02794759825328,
+ "grad_norm": 0.018810199573636055,
+ "learning_rate": 0.0006,
+ "loss": 4.82293701171875,
+ "step": 3602
+ },
+ {
+ "epoch": 50.041921397379916,
+ "grad_norm": 0.02227160707116127,
+ "learning_rate": 0.0006,
+ "loss": 4.9612298011779785,
+ "step": 3603
+ },
+ {
+ "epoch": 50.05589519650655,
+ "grad_norm": 0.025266462936997414,
+ "learning_rate": 0.0006,
+ "loss": 4.963835716247559,
+ "step": 3604
+ },
+ {
+ "epoch": 50.069868995633186,
+ "grad_norm": 0.028588181361556053,
+ "learning_rate": 0.0006,
+ "loss": 4.992118835449219,
+ "step": 3605
+ },
+ {
+ "epoch": 50.083842794759825,
+ "grad_norm": 0.02739487960934639,
+ "learning_rate": 0.0006,
+ "loss": 4.915547847747803,
+ "step": 3606
+ },
+ {
+ "epoch": 50.09781659388646,
+ "grad_norm": 0.027000300586223602,
+ "learning_rate": 0.0006,
+ "loss": 4.883801460266113,
+ "step": 3607
+ },
+ {
+ "epoch": 50.1117903930131,
+ "grad_norm": 0.023468483239412308,
+ "learning_rate": 0.0006,
+ "loss": 5.006168365478516,
+ "step": 3608
+ },
+ {
+ "epoch": 50.12576419213974,
+ "grad_norm": 0.026753634214401245,
+ "learning_rate": 0.0006,
+ "loss": 4.92526912689209,
+ "step": 3609
+ },
+ {
+ "epoch": 50.13973799126637,
+ "grad_norm": 0.026816733181476593,
+ "learning_rate": 0.0006,
+ "loss": 4.847546100616455,
+ "step": 3610
+ },
+ {
+ "epoch": 50.15371179039301,
+ "grad_norm": 0.027362391352653503,
+ "learning_rate": 0.0006,
+ "loss": 4.862548351287842,
+ "step": 3611
+ },
+ {
+ "epoch": 50.16768558951965,
+ "grad_norm": 0.03450845181941986,
+ "learning_rate": 0.0006,
+ "loss": 4.963202953338623,
+ "step": 3612
+ },
+ {
+ "epoch": 50.18165938864629,
+ "grad_norm": 0.03801408037543297,
+ "learning_rate": 0.0006,
+ "loss": 4.770240306854248,
+ "step": 3613
+ },
+ {
+ "epoch": 50.19563318777293,
+ "grad_norm": 0.03579140082001686,
+ "learning_rate": 0.0006,
+ "loss": 4.797981262207031,
+ "step": 3614
+ },
+ {
+ "epoch": 50.209606986899566,
+ "grad_norm": 0.029393760487437248,
+ "learning_rate": 0.0006,
+ "loss": 4.798581600189209,
+ "step": 3615
+ },
+ {
+ "epoch": 50.223580786026204,
+ "grad_norm": 0.026111142709851265,
+ "learning_rate": 0.0006,
+ "loss": 4.990712642669678,
+ "step": 3616
+ },
+ {
+ "epoch": 50.237554585152836,
+ "grad_norm": 0.02213387005031109,
+ "learning_rate": 0.0006,
+ "loss": 4.865842342376709,
+ "step": 3617
+ },
+ {
+ "epoch": 50.251528384279474,
+ "grad_norm": 0.025450684130191803,
+ "learning_rate": 0.0006,
+ "loss": 4.903746128082275,
+ "step": 3618
+ },
+ {
+ "epoch": 50.26550218340611,
+ "grad_norm": 0.020915117114782333,
+ "learning_rate": 0.0006,
+ "loss": 4.951366424560547,
+ "step": 3619
+ },
+ {
+ "epoch": 50.27947598253275,
+ "grad_norm": 0.020972849801182747,
+ "learning_rate": 0.0006,
+ "loss": 4.783311367034912,
+ "step": 3620
+ },
+ {
+ "epoch": 50.29344978165939,
+ "grad_norm": 0.016956273466348648,
+ "learning_rate": 0.0006,
+ "loss": 4.955767631530762,
+ "step": 3621
+ },
+ {
+ "epoch": 50.30742358078603,
+ "grad_norm": 0.018045252189040184,
+ "learning_rate": 0.0006,
+ "loss": 4.869322776794434,
+ "step": 3622
+ },
+ {
+ "epoch": 50.32139737991266,
+ "grad_norm": 0.018199188634753227,
+ "learning_rate": 0.0006,
+ "loss": 4.95680046081543,
+ "step": 3623
+ },
+ {
+ "epoch": 50.3353711790393,
+ "grad_norm": 0.016688082367181778,
+ "learning_rate": 0.0006,
+ "loss": 4.783535957336426,
+ "step": 3624
+ },
+ {
+ "epoch": 50.34934497816594,
+ "grad_norm": 0.017540352419018745,
+ "learning_rate": 0.0006,
+ "loss": 4.8236083984375,
+ "step": 3625
+ },
+ {
+ "epoch": 50.36331877729258,
+ "grad_norm": 0.018054187297821045,
+ "learning_rate": 0.0006,
+ "loss": 4.764655113220215,
+ "step": 3626
+ },
+ {
+ "epoch": 50.377292576419215,
+ "grad_norm": 0.01741131953895092,
+ "learning_rate": 0.0006,
+ "loss": 4.875613689422607,
+ "step": 3627
+ },
+ {
+ "epoch": 50.391266375545854,
+ "grad_norm": 0.01682269014418125,
+ "learning_rate": 0.0006,
+ "loss": 4.818509101867676,
+ "step": 3628
+ },
+ {
+ "epoch": 50.40524017467249,
+ "grad_norm": 0.016652341932058334,
+ "learning_rate": 0.0006,
+ "loss": 4.974020481109619,
+ "step": 3629
+ },
+ {
+ "epoch": 50.419213973799124,
+ "grad_norm": 0.02035282365977764,
+ "learning_rate": 0.0006,
+ "loss": 4.893037796020508,
+ "step": 3630
+ },
+ {
+ "epoch": 50.43318777292576,
+ "grad_norm": 0.019907061010599136,
+ "learning_rate": 0.0006,
+ "loss": 4.848119735717773,
+ "step": 3631
+ },
+ {
+ "epoch": 50.4471615720524,
+ "grad_norm": 0.023861493915319443,
+ "learning_rate": 0.0006,
+ "loss": 4.819353103637695,
+ "step": 3632
+ },
+ {
+ "epoch": 50.46113537117904,
+ "grad_norm": 0.025254026055336,
+ "learning_rate": 0.0006,
+ "loss": 4.769312858581543,
+ "step": 3633
+ },
+ {
+ "epoch": 50.47510917030568,
+ "grad_norm": 0.026154937222599983,
+ "learning_rate": 0.0006,
+ "loss": 4.7586774826049805,
+ "step": 3634
+ },
+ {
+ "epoch": 50.48908296943232,
+ "grad_norm": 0.026342090219259262,
+ "learning_rate": 0.0006,
+ "loss": 4.861218452453613,
+ "step": 3635
+ },
+ {
+ "epoch": 50.50305676855895,
+ "grad_norm": 0.028169559314846992,
+ "learning_rate": 0.0006,
+ "loss": 4.818380355834961,
+ "step": 3636
+ },
+ {
+ "epoch": 50.51703056768559,
+ "grad_norm": 0.02820264920592308,
+ "learning_rate": 0.0006,
+ "loss": 4.892871856689453,
+ "step": 3637
+ },
+ {
+ "epoch": 50.531004366812226,
+ "grad_norm": 0.03345080465078354,
+ "learning_rate": 0.0006,
+ "loss": 4.811084747314453,
+ "step": 3638
+ },
+ {
+ "epoch": 50.544978165938865,
+ "grad_norm": 0.04005736857652664,
+ "learning_rate": 0.0006,
+ "loss": 4.765410900115967,
+ "step": 3639
+ },
+ {
+ "epoch": 50.5589519650655,
+ "grad_norm": 0.04333870857954025,
+ "learning_rate": 0.0006,
+ "loss": 4.956520080566406,
+ "step": 3640
+ },
+ {
+ "epoch": 50.57292576419214,
+ "grad_norm": 0.04216952994465828,
+ "learning_rate": 0.0006,
+ "loss": 4.793441295623779,
+ "step": 3641
+ },
+ {
+ "epoch": 50.58689956331878,
+ "grad_norm": 0.02936951443552971,
+ "learning_rate": 0.0006,
+ "loss": 4.8447957038879395,
+ "step": 3642
+ },
+ {
+ "epoch": 50.60087336244541,
+ "grad_norm": 0.026937026530504227,
+ "learning_rate": 0.0006,
+ "loss": 4.962215423583984,
+ "step": 3643
+ },
+ {
+ "epoch": 50.61484716157205,
+ "grad_norm": 0.02628890983760357,
+ "learning_rate": 0.0006,
+ "loss": 4.883061408996582,
+ "step": 3644
+ },
+ {
+ "epoch": 50.62882096069869,
+ "grad_norm": 0.025984834879636765,
+ "learning_rate": 0.0006,
+ "loss": 4.768662452697754,
+ "step": 3645
+ },
+ {
+ "epoch": 50.64279475982533,
+ "grad_norm": 0.023998012766242027,
+ "learning_rate": 0.0006,
+ "loss": 4.851325035095215,
+ "step": 3646
+ },
+ {
+ "epoch": 50.65676855895197,
+ "grad_norm": 0.02143184095621109,
+ "learning_rate": 0.0006,
+ "loss": 4.8274688720703125,
+ "step": 3647
+ },
+ {
+ "epoch": 50.670742358078606,
+ "grad_norm": 0.02488900162279606,
+ "learning_rate": 0.0006,
+ "loss": 4.831140518188477,
+ "step": 3648
+ },
+ {
+ "epoch": 50.68471615720524,
+ "grad_norm": 0.02330533228814602,
+ "learning_rate": 0.0006,
+ "loss": 4.850462913513184,
+ "step": 3649
+ },
+ {
+ "epoch": 50.698689956331876,
+ "grad_norm": 0.02121943049132824,
+ "learning_rate": 0.0006,
+ "loss": 4.870253086090088,
+ "step": 3650
+ },
+ {
+ "epoch": 50.712663755458514,
+ "grad_norm": 0.02080935798585415,
+ "learning_rate": 0.0006,
+ "loss": 4.798770904541016,
+ "step": 3651
+ },
+ {
+ "epoch": 50.72663755458515,
+ "grad_norm": 0.019047029316425323,
+ "learning_rate": 0.0006,
+ "loss": 4.8427300453186035,
+ "step": 3652
+ },
+ {
+ "epoch": 50.74061135371179,
+ "grad_norm": 0.01824103854596615,
+ "learning_rate": 0.0006,
+ "loss": 4.820980548858643,
+ "step": 3653
+ },
+ {
+ "epoch": 50.75458515283843,
+ "grad_norm": 0.018129967153072357,
+ "learning_rate": 0.0006,
+ "loss": 4.946732997894287,
+ "step": 3654
+ },
+ {
+ "epoch": 50.76855895196506,
+ "grad_norm": 0.01784815452992916,
+ "learning_rate": 0.0006,
+ "loss": 4.829516410827637,
+ "step": 3655
+ },
+ {
+ "epoch": 50.7825327510917,
+ "grad_norm": 0.020021555945277214,
+ "learning_rate": 0.0006,
+ "loss": 4.780428409576416,
+ "step": 3656
+ },
+ {
+ "epoch": 50.79650655021834,
+ "grad_norm": 0.01863253302872181,
+ "learning_rate": 0.0006,
+ "loss": 4.791567802429199,
+ "step": 3657
+ },
+ {
+ "epoch": 50.81048034934498,
+ "grad_norm": 0.019151683896780014,
+ "learning_rate": 0.0006,
+ "loss": 4.824892997741699,
+ "step": 3658
+ },
+ {
+ "epoch": 50.82445414847162,
+ "grad_norm": 0.01891648955643177,
+ "learning_rate": 0.0006,
+ "loss": 4.789605140686035,
+ "step": 3659
+ },
+ {
+ "epoch": 50.838427947598255,
+ "grad_norm": 0.016019510105252266,
+ "learning_rate": 0.0006,
+ "loss": 4.779555320739746,
+ "step": 3660
+ },
+ {
+ "epoch": 50.852401746724894,
+ "grad_norm": 0.016096491366624832,
+ "learning_rate": 0.0006,
+ "loss": 4.753842353820801,
+ "step": 3661
+ },
+ {
+ "epoch": 50.866375545851525,
+ "grad_norm": 0.01693476364016533,
+ "learning_rate": 0.0006,
+ "loss": 4.768477439880371,
+ "step": 3662
+ },
+ {
+ "epoch": 50.880349344978164,
+ "grad_norm": 0.01749575324356556,
+ "learning_rate": 0.0006,
+ "loss": 4.779143333435059,
+ "step": 3663
+ },
+ {
+ "epoch": 50.8943231441048,
+ "grad_norm": 0.016851043328642845,
+ "learning_rate": 0.0006,
+ "loss": 4.780387878417969,
+ "step": 3664
+ },
+ {
+ "epoch": 50.90829694323144,
+ "grad_norm": 0.019045202061533928,
+ "learning_rate": 0.0006,
+ "loss": 4.769913673400879,
+ "step": 3665
+ },
+ {
+ "epoch": 50.92227074235808,
+ "grad_norm": 0.01954123005270958,
+ "learning_rate": 0.0006,
+ "loss": 4.769009590148926,
+ "step": 3666
+ },
+ {
+ "epoch": 50.93624454148472,
+ "grad_norm": 0.022527096793055534,
+ "learning_rate": 0.0006,
+ "loss": 4.789215087890625,
+ "step": 3667
+ },
+ {
+ "epoch": 50.95021834061135,
+ "grad_norm": 0.02871634066104889,
+ "learning_rate": 0.0006,
+ "loss": 4.749727725982666,
+ "step": 3668
+ },
+ {
+ "epoch": 50.96419213973799,
+ "grad_norm": 0.04045053571462631,
+ "learning_rate": 0.0006,
+ "loss": 4.909038543701172,
+ "step": 3669
+ },
+ {
+ "epoch": 50.97816593886463,
+ "grad_norm": 0.04156879335641861,
+ "learning_rate": 0.0006,
+ "loss": 4.876831531524658,
+ "step": 3670
+ },
+ {
+ "epoch": 50.992139737991266,
+ "grad_norm": 0.02621397376060486,
+ "learning_rate": 0.0006,
+ "loss": 4.797173500061035,
+ "step": 3671
+ },
+ {
+ "epoch": 51.0,
+ "grad_norm": 0.03389597684144974,
+ "learning_rate": 0.0006,
+ "loss": 4.8428850173950195,
+ "step": 3672
+ },
+ {
+ "epoch": 51.0,
+ "eval_loss": 4.989261627197266,
+ "eval_runtime": 56.358,
+ "eval_samples_per_second": 43.33,
+ "eval_steps_per_second": 1.366,
+ "step": 3672
+ },
+ {
+ "epoch": 51.01397379912664,
+ "grad_norm": 0.03216972202062607,
+ "learning_rate": 0.0006,
+ "loss": 4.755102634429932,
+ "step": 3673
+ },
+ {
+ "epoch": 51.02794759825328,
+ "grad_norm": 0.03190374746918678,
+ "learning_rate": 0.0006,
+ "loss": 4.720424652099609,
+ "step": 3674
+ },
+ {
+ "epoch": 51.041921397379916,
+ "grad_norm": 0.031727299094200134,
+ "learning_rate": 0.0006,
+ "loss": 4.803277969360352,
+ "step": 3675
+ },
+ {
+ "epoch": 51.05589519650655,
+ "grad_norm": 0.02473079040646553,
+ "learning_rate": 0.0006,
+ "loss": 4.761087417602539,
+ "step": 3676
+ },
+ {
+ "epoch": 51.069868995633186,
+ "grad_norm": 0.020360209047794342,
+ "learning_rate": 0.0006,
+ "loss": 4.753884792327881,
+ "step": 3677
+ },
+ {
+ "epoch": 51.083842794759825,
+ "grad_norm": 0.02078923024237156,
+ "learning_rate": 0.0006,
+ "loss": 4.647130489349365,
+ "step": 3678
+ },
+ {
+ "epoch": 51.09781659388646,
+ "grad_norm": 0.019387200474739075,
+ "learning_rate": 0.0006,
+ "loss": 4.74135160446167,
+ "step": 3679
+ },
+ {
+ "epoch": 51.1117903930131,
+ "grad_norm": 0.01874949410557747,
+ "learning_rate": 0.0006,
+ "loss": 4.795329570770264,
+ "step": 3680
+ },
+ {
+ "epoch": 51.12576419213974,
+ "grad_norm": 0.019331837072968483,
+ "learning_rate": 0.0006,
+ "loss": 4.897967338562012,
+ "step": 3681
+ },
+ {
+ "epoch": 51.13973799126637,
+ "grad_norm": 0.017066050320863724,
+ "learning_rate": 0.0006,
+ "loss": 4.825351715087891,
+ "step": 3682
+ },
+ {
+ "epoch": 51.15371179039301,
+ "grad_norm": 0.01757810264825821,
+ "learning_rate": 0.0006,
+ "loss": 4.719335079193115,
+ "step": 3683
+ },
+ {
+ "epoch": 51.16768558951965,
+ "grad_norm": 0.018174266442656517,
+ "learning_rate": 0.0006,
+ "loss": 4.732680320739746,
+ "step": 3684
+ },
+ {
+ "epoch": 51.18165938864629,
+ "grad_norm": 0.018497727811336517,
+ "learning_rate": 0.0006,
+ "loss": 4.731050491333008,
+ "step": 3685
+ },
+ {
+ "epoch": 51.19563318777293,
+ "grad_norm": 0.0186338871717453,
+ "learning_rate": 0.0006,
+ "loss": 4.748035430908203,
+ "step": 3686
+ },
+ {
+ "epoch": 51.209606986899566,
+ "grad_norm": 0.018531711772084236,
+ "learning_rate": 0.0006,
+ "loss": 4.9007463455200195,
+ "step": 3687
+ },
+ {
+ "epoch": 51.223580786026204,
+ "grad_norm": 0.018221354112029076,
+ "learning_rate": 0.0006,
+ "loss": 4.84077262878418,
+ "step": 3688
+ },
+ {
+ "epoch": 51.237554585152836,
+ "grad_norm": 0.0203656405210495,
+ "learning_rate": 0.0006,
+ "loss": 4.73009729385376,
+ "step": 3689
+ },
+ {
+ "epoch": 51.251528384279474,
+ "grad_norm": 0.01964893378317356,
+ "learning_rate": 0.0006,
+ "loss": 4.772583961486816,
+ "step": 3690
+ },
+ {
+ "epoch": 51.26550218340611,
+ "grad_norm": 0.019606510177254677,
+ "learning_rate": 0.0006,
+ "loss": 4.599812030792236,
+ "step": 3691
+ },
+ {
+ "epoch": 51.27947598253275,
+ "grad_norm": 0.020501255989074707,
+ "learning_rate": 0.0006,
+ "loss": 4.809898853302002,
+ "step": 3692
+ },
+ {
+ "epoch": 51.29344978165939,
+ "grad_norm": 0.01870838552713394,
+ "learning_rate": 0.0006,
+ "loss": 4.741898059844971,
+ "step": 3693
+ },
+ {
+ "epoch": 51.30742358078603,
+ "grad_norm": 0.01802559196949005,
+ "learning_rate": 0.0006,
+ "loss": 4.746562480926514,
+ "step": 3694
+ },
+ {
+ "epoch": 51.32139737991266,
+ "grad_norm": 0.02053048275411129,
+ "learning_rate": 0.0006,
+ "loss": 4.69932746887207,
+ "step": 3695
+ },
+ {
+ "epoch": 51.3353711790393,
+ "grad_norm": 0.02661825157701969,
+ "learning_rate": 0.0006,
+ "loss": 4.68748664855957,
+ "step": 3696
+ },
+ {
+ "epoch": 51.34934497816594,
+ "grad_norm": 0.028269147500395775,
+ "learning_rate": 0.0006,
+ "loss": 4.731565475463867,
+ "step": 3697
+ },
+ {
+ "epoch": 51.36331877729258,
+ "grad_norm": 0.02650964818894863,
+ "learning_rate": 0.0006,
+ "loss": 4.733857154846191,
+ "step": 3698
+ },
+ {
+ "epoch": 51.377292576419215,
+ "grad_norm": 0.023696066811680794,
+ "learning_rate": 0.0006,
+ "loss": 4.812362194061279,
+ "step": 3699
+ },
+ {
+ "epoch": 51.391266375545854,
+ "grad_norm": 0.02365880273282528,
+ "learning_rate": 0.0006,
+ "loss": 4.733426094055176,
+ "step": 3700
+ },
+ {
+ "epoch": 51.40524017467249,
+ "grad_norm": 0.026249447837471962,
+ "learning_rate": 0.0006,
+ "loss": 4.715622901916504,
+ "step": 3701
+ },
+ {
+ "epoch": 51.419213973799124,
+ "grad_norm": 0.03301899880170822,
+ "learning_rate": 0.0006,
+ "loss": 4.7223663330078125,
+ "step": 3702
+ },
+ {
+ "epoch": 51.43318777292576,
+ "grad_norm": 0.044686660170555115,
+ "learning_rate": 0.0006,
+ "loss": 4.8121161460876465,
+ "step": 3703
+ },
+ {
+ "epoch": 51.4471615720524,
+ "grad_norm": 0.04733522981405258,
+ "learning_rate": 0.0006,
+ "loss": 4.841394424438477,
+ "step": 3704
+ },
+ {
+ "epoch": 51.46113537117904,
+ "grad_norm": 0.0395209975540638,
+ "learning_rate": 0.0006,
+ "loss": 4.773837089538574,
+ "step": 3705
+ },
+ {
+ "epoch": 51.47510917030568,
+ "grad_norm": 0.032790958881378174,
+ "learning_rate": 0.0006,
+ "loss": 4.833066463470459,
+ "step": 3706
+ },
+ {
+ "epoch": 51.48908296943232,
+ "grad_norm": 0.030578427016735077,
+ "learning_rate": 0.0006,
+ "loss": 4.685190200805664,
+ "step": 3707
+ },
+ {
+ "epoch": 51.50305676855895,
+ "grad_norm": 0.031619083136320114,
+ "learning_rate": 0.0006,
+ "loss": 4.740378379821777,
+ "step": 3708
+ },
+ {
+ "epoch": 51.51703056768559,
+ "grad_norm": 0.02385825850069523,
+ "learning_rate": 0.0006,
+ "loss": 4.812911510467529,
+ "step": 3709
+ },
+ {
+ "epoch": 51.531004366812226,
+ "grad_norm": 0.023523425683379173,
+ "learning_rate": 0.0006,
+ "loss": 4.669580459594727,
+ "step": 3710
+ },
+ {
+ "epoch": 51.544978165938865,
+ "grad_norm": 0.021316280588507652,
+ "learning_rate": 0.0006,
+ "loss": 4.76888370513916,
+ "step": 3711
+ },
+ {
+ "epoch": 51.5589519650655,
+ "grad_norm": 0.024663880467414856,
+ "learning_rate": 0.0006,
+ "loss": 4.702592372894287,
+ "step": 3712
+ },
+ {
+ "epoch": 51.57292576419214,
+ "grad_norm": 0.024173280224204063,
+ "learning_rate": 0.0006,
+ "loss": 4.710604667663574,
+ "step": 3713
+ },
+ {
+ "epoch": 51.58689956331878,
+ "grad_norm": 0.02754114381968975,
+ "learning_rate": 0.0006,
+ "loss": 4.746356964111328,
+ "step": 3714
+ },
+ {
+ "epoch": 51.60087336244541,
+ "grad_norm": 0.027903622016310692,
+ "learning_rate": 0.0006,
+ "loss": 4.7546491622924805,
+ "step": 3715
+ },
+ {
+ "epoch": 51.61484716157205,
+ "grad_norm": 0.024871617555618286,
+ "learning_rate": 0.0006,
+ "loss": 4.605746269226074,
+ "step": 3716
+ },
+ {
+ "epoch": 51.62882096069869,
+ "grad_norm": 0.021239031106233597,
+ "learning_rate": 0.0006,
+ "loss": 4.6942458152771,
+ "step": 3717
+ },
+ {
+ "epoch": 51.64279475982533,
+ "grad_norm": 0.01925622671842575,
+ "learning_rate": 0.0006,
+ "loss": 4.593558311462402,
+ "step": 3718
+ },
+ {
+ "epoch": 51.65676855895197,
+ "grad_norm": 0.01956283673644066,
+ "learning_rate": 0.0006,
+ "loss": 4.631172180175781,
+ "step": 3719
+ },
+ {
+ "epoch": 51.670742358078606,
+ "grad_norm": 0.02269871160387993,
+ "learning_rate": 0.0006,
+ "loss": 4.729236602783203,
+ "step": 3720
+ },
+ {
+ "epoch": 51.68471615720524,
+ "grad_norm": 0.02253916673362255,
+ "learning_rate": 0.0006,
+ "loss": 4.7191033363342285,
+ "step": 3721
+ },
+ {
+ "epoch": 51.698689956331876,
+ "grad_norm": 0.026231050491333008,
+ "learning_rate": 0.0006,
+ "loss": 4.811850547790527,
+ "step": 3722
+ },
+ {
+ "epoch": 51.712663755458514,
+ "grad_norm": 0.02558821812272072,
+ "learning_rate": 0.0006,
+ "loss": 4.681466102600098,
+ "step": 3723
+ },
+ {
+ "epoch": 51.72663755458515,
+ "grad_norm": 0.02124209702014923,
+ "learning_rate": 0.0006,
+ "loss": 4.693664073944092,
+ "step": 3724
+ },
+ {
+ "epoch": 51.74061135371179,
+ "grad_norm": 0.016638319939374924,
+ "learning_rate": 0.0006,
+ "loss": 4.771440029144287,
+ "step": 3725
+ },
+ {
+ "epoch": 51.75458515283843,
+ "grad_norm": 0.018048347905278206,
+ "learning_rate": 0.0006,
+ "loss": 4.643252372741699,
+ "step": 3726
+ },
+ {
+ "epoch": 51.76855895196506,
+ "grad_norm": 0.019904717803001404,
+ "learning_rate": 0.0006,
+ "loss": 4.66530704498291,
+ "step": 3727
+ },
+ {
+ "epoch": 51.7825327510917,
+ "grad_norm": 0.0194614976644516,
+ "learning_rate": 0.0006,
+ "loss": 4.711943626403809,
+ "step": 3728
+ },
+ {
+ "epoch": 51.79650655021834,
+ "grad_norm": 0.019175555557012558,
+ "learning_rate": 0.0006,
+ "loss": 4.721931457519531,
+ "step": 3729
+ },
+ {
+ "epoch": 51.81048034934498,
+ "grad_norm": 0.02093447372317314,
+ "learning_rate": 0.0006,
+ "loss": 4.633059024810791,
+ "step": 3730
+ },
+ {
+ "epoch": 51.82445414847162,
+ "grad_norm": 0.01930283196270466,
+ "learning_rate": 0.0006,
+ "loss": 4.659373760223389,
+ "step": 3731
+ },
+ {
+ "epoch": 51.838427947598255,
+ "grad_norm": 0.015350676141679287,
+ "learning_rate": 0.0006,
+ "loss": 4.763708114624023,
+ "step": 3732
+ },
+ {
+ "epoch": 51.852401746724894,
+ "grad_norm": 0.014139448292553425,
+ "learning_rate": 0.0006,
+ "loss": 4.753891468048096,
+ "step": 3733
+ },
+ {
+ "epoch": 51.866375545851525,
+ "grad_norm": 0.016215872019529343,
+ "learning_rate": 0.0006,
+ "loss": 4.658967018127441,
+ "step": 3734
+ },
+ {
+ "epoch": 51.880349344978164,
+ "grad_norm": 0.017328619956970215,
+ "learning_rate": 0.0006,
+ "loss": 4.696579933166504,
+ "step": 3735
+ },
+ {
+ "epoch": 51.8943231441048,
+ "grad_norm": 0.018103396520018578,
+ "learning_rate": 0.0006,
+ "loss": 4.627464771270752,
+ "step": 3736
+ },
+ {
+ "epoch": 51.90829694323144,
+ "grad_norm": 0.01715235412120819,
+ "learning_rate": 0.0006,
+ "loss": 4.7368011474609375,
+ "step": 3737
+ },
+ {
+ "epoch": 51.92227074235808,
+ "grad_norm": 0.014790094457566738,
+ "learning_rate": 0.0006,
+ "loss": 4.629441738128662,
+ "step": 3738
+ },
+ {
+ "epoch": 51.93624454148472,
+ "grad_norm": 0.017114296555519104,
+ "learning_rate": 0.0006,
+ "loss": 4.761256694793701,
+ "step": 3739
+ },
+ {
+ "epoch": 51.95021834061135,
+ "grad_norm": 0.019009845331311226,
+ "learning_rate": 0.0006,
+ "loss": 4.789155006408691,
+ "step": 3740
+ },
+ {
+ "epoch": 51.96419213973799,
+ "grad_norm": 0.019696692004799843,
+ "learning_rate": 0.0006,
+ "loss": 4.721269607543945,
+ "step": 3741
+ },
+ {
+ "epoch": 51.97816593886463,
+ "grad_norm": 0.020542822778224945,
+ "learning_rate": 0.0006,
+ "loss": 4.687038898468018,
+ "step": 3742
+ },
+ {
+ "epoch": 51.992139737991266,
+ "grad_norm": 0.021267160773277283,
+ "learning_rate": 0.0006,
+ "loss": 4.747860908508301,
+ "step": 3743
+ },
+ {
+ "epoch": 52.0,
+ "grad_norm": 0.025049567222595215,
+ "learning_rate": 0.0006,
+ "loss": 4.8473100662231445,
+ "step": 3744
+ },
+ {
+ "epoch": 52.0,
+ "eval_loss": 4.943043231964111,
+ "eval_runtime": 56.4929,
+ "eval_samples_per_second": 43.227,
+ "eval_steps_per_second": 1.363,
+ "step": 3744
+ },
+ {
+ "epoch": 52.01397379912664,
+ "grad_norm": 0.03054986707866192,
+ "learning_rate": 0.0006,
+ "loss": 4.856931686401367,
+ "step": 3745
+ },
+ {
+ "epoch": 52.02794759825328,
+ "grad_norm": 0.03838375583291054,
+ "learning_rate": 0.0006,
+ "loss": 4.699807167053223,
+ "step": 3746
+ },
+ {
+ "epoch": 52.041921397379916,
+ "grad_norm": 0.03383488208055496,
+ "learning_rate": 0.0006,
+ "loss": 4.6626787185668945,
+ "step": 3747
+ },
+ {
+ "epoch": 52.05589519650655,
+ "grad_norm": 0.0201561376452446,
+ "learning_rate": 0.0006,
+ "loss": 4.621831893920898,
+ "step": 3748
+ },
+ {
+ "epoch": 52.069868995633186,
+ "grad_norm": 0.028749438002705574,
+ "learning_rate": 0.0006,
+ "loss": 4.682540416717529,
+ "step": 3749
+ },
+ {
+ "epoch": 52.083842794759825,
+ "grad_norm": 0.02681174874305725,
+ "learning_rate": 0.0006,
+ "loss": 4.629997253417969,
+ "step": 3750
+ },
+ {
+ "epoch": 52.09781659388646,
+ "grad_norm": 0.024063318967819214,
+ "learning_rate": 0.0006,
+ "loss": 4.573021411895752,
+ "step": 3751
+ },
+ {
+ "epoch": 52.1117903930131,
+ "grad_norm": 0.02324330434203148,
+ "learning_rate": 0.0006,
+ "loss": 4.739465236663818,
+ "step": 3752
+ },
+ {
+ "epoch": 52.12576419213974,
+ "grad_norm": 0.0225426834076643,
+ "learning_rate": 0.0006,
+ "loss": 4.7335333824157715,
+ "step": 3753
+ },
+ {
+ "epoch": 52.13973799126637,
+ "grad_norm": 0.019593428820371628,
+ "learning_rate": 0.0006,
+ "loss": 4.704958438873291,
+ "step": 3754
+ },
+ {
+ "epoch": 52.15371179039301,
+ "grad_norm": 0.020777082070708275,
+ "learning_rate": 0.0006,
+ "loss": 4.638696670532227,
+ "step": 3755
+ },
+ {
+ "epoch": 52.16768558951965,
+ "grad_norm": 0.021525604650378227,
+ "learning_rate": 0.0006,
+ "loss": 4.643706798553467,
+ "step": 3756
+ },
+ {
+ "epoch": 52.18165938864629,
+ "grad_norm": 0.018943654373288155,
+ "learning_rate": 0.0006,
+ "loss": 4.5173211097717285,
+ "step": 3757
+ },
+ {
+ "epoch": 52.19563318777293,
+ "grad_norm": 0.01856580562889576,
+ "learning_rate": 0.0006,
+ "loss": 4.678788661956787,
+ "step": 3758
+ },
+ {
+ "epoch": 52.209606986899566,
+ "grad_norm": 0.019417259842157364,
+ "learning_rate": 0.0006,
+ "loss": 4.587244987487793,
+ "step": 3759
+ },
+ {
+ "epoch": 52.223580786026204,
+ "grad_norm": 0.01911911740899086,
+ "learning_rate": 0.0006,
+ "loss": 4.688503265380859,
+ "step": 3760
+ },
+ {
+ "epoch": 52.237554585152836,
+ "grad_norm": 0.021017879247665405,
+ "learning_rate": 0.0006,
+ "loss": 4.65829610824585,
+ "step": 3761
+ },
+ {
+ "epoch": 52.251528384279474,
+ "grad_norm": 0.02048875205218792,
+ "learning_rate": 0.0006,
+ "loss": 4.624055862426758,
+ "step": 3762
+ },
+ {
+ "epoch": 52.26550218340611,
+ "grad_norm": 0.01932772807776928,
+ "learning_rate": 0.0006,
+ "loss": 4.556562900543213,
+ "step": 3763
+ },
+ {
+ "epoch": 52.27947598253275,
+ "grad_norm": 0.01871614344418049,
+ "learning_rate": 0.0006,
+ "loss": 4.581201076507568,
+ "step": 3764
+ },
+ {
+ "epoch": 52.29344978165939,
+ "grad_norm": 0.017253106459975243,
+ "learning_rate": 0.0006,
+ "loss": 4.605219841003418,
+ "step": 3765
+ },
+ {
+ "epoch": 52.30742358078603,
+ "grad_norm": 0.01831238530576229,
+ "learning_rate": 0.0006,
+ "loss": 4.759900093078613,
+ "step": 3766
+ },
+ {
+ "epoch": 52.32139737991266,
+ "grad_norm": 0.017946207895874977,
+ "learning_rate": 0.0006,
+ "loss": 4.5598344802856445,
+ "step": 3767
+ },
+ {
+ "epoch": 52.3353711790393,
+ "grad_norm": 0.01680106669664383,
+ "learning_rate": 0.0006,
+ "loss": 4.605034828186035,
+ "step": 3768
+ },
+ {
+ "epoch": 52.34934497816594,
+ "grad_norm": 0.021482540294528008,
+ "learning_rate": 0.0006,
+ "loss": 4.620411396026611,
+ "step": 3769
+ },
+ {
+ "epoch": 52.36331877729258,
+ "grad_norm": 0.02210252732038498,
+ "learning_rate": 0.0006,
+ "loss": 4.667577743530273,
+ "step": 3770
+ },
+ {
+ "epoch": 52.377292576419215,
+ "grad_norm": 0.02237485907971859,
+ "learning_rate": 0.0006,
+ "loss": 4.628812789916992,
+ "step": 3771
+ },
+ {
+ "epoch": 52.391266375545854,
+ "grad_norm": 0.02196541801095009,
+ "learning_rate": 0.0006,
+ "loss": 4.572988510131836,
+ "step": 3772
+ },
+ {
+ "epoch": 52.40524017467249,
+ "grad_norm": 0.01930229924619198,
+ "learning_rate": 0.0006,
+ "loss": 4.588339805603027,
+ "step": 3773
+ },
+ {
+ "epoch": 52.419213973799124,
+ "grad_norm": 0.02222434990108013,
+ "learning_rate": 0.0006,
+ "loss": 4.766482353210449,
+ "step": 3774
+ },
+ {
+ "epoch": 52.43318777292576,
+ "grad_norm": 0.028806306421756744,
+ "learning_rate": 0.0006,
+ "loss": 4.617892742156982,
+ "step": 3775
+ },
+ {
+ "epoch": 52.4471615720524,
+ "grad_norm": 0.03306042402982712,
+ "learning_rate": 0.0006,
+ "loss": 4.665334701538086,
+ "step": 3776
+ },
+ {
+ "epoch": 52.46113537117904,
+ "grad_norm": 0.0358956903219223,
+ "learning_rate": 0.0006,
+ "loss": 4.566668510437012,
+ "step": 3777
+ },
+ {
+ "epoch": 52.47510917030568,
+ "grad_norm": 0.030406497418880463,
+ "learning_rate": 0.0006,
+ "loss": 4.513612270355225,
+ "step": 3778
+ },
+ {
+ "epoch": 52.48908296943232,
+ "grad_norm": 0.030295712873339653,
+ "learning_rate": 0.0006,
+ "loss": 4.640189170837402,
+ "step": 3779
+ },
+ {
+ "epoch": 52.50305676855895,
+ "grad_norm": 0.024981003254652023,
+ "learning_rate": 0.0006,
+ "loss": 4.57558012008667,
+ "step": 3780
+ },
+ {
+ "epoch": 52.51703056768559,
+ "grad_norm": 0.024990713223814964,
+ "learning_rate": 0.0006,
+ "loss": 4.708589553833008,
+ "step": 3781
+ },
+ {
+ "epoch": 52.531004366812226,
+ "grad_norm": 0.025034775957465172,
+ "learning_rate": 0.0006,
+ "loss": 4.70468282699585,
+ "step": 3782
+ },
+ {
+ "epoch": 52.544978165938865,
+ "grad_norm": 0.028243759647011757,
+ "learning_rate": 0.0006,
+ "loss": 4.712060451507568,
+ "step": 3783
+ },
+ {
+ "epoch": 52.5589519650655,
+ "grad_norm": 0.026984544470906258,
+ "learning_rate": 0.0006,
+ "loss": 4.505004405975342,
+ "step": 3784
+ },
+ {
+ "epoch": 52.57292576419214,
+ "grad_norm": 0.050843529403209686,
+ "learning_rate": 0.0006,
+ "loss": 4.593271255493164,
+ "step": 3785
+ },
+ {
+ "epoch": 52.58689956331878,
+ "grad_norm": 0.25422802567481995,
+ "learning_rate": 0.0006,
+ "loss": 4.7178802490234375,
+ "step": 3786
+ },
+ {
+ "epoch": 52.60087336244541,
+ "grad_norm": 0.6078406572341919,
+ "learning_rate": 0.0006,
+ "loss": 6.144372940063477,
+ "step": 3787
+ },
+ {
+ "epoch": 52.61484716157205,
+ "grad_norm": 0.25933101773262024,
+ "learning_rate": 0.0006,
+ "loss": 5.971083641052246,
+ "step": 3788
+ },
+ {
+ "epoch": 52.62882096069869,
+ "grad_norm": 0.294657826423645,
+ "learning_rate": 0.0006,
+ "loss": 6.204090118408203,
+ "step": 3789
+ },
+ {
+ "epoch": 52.64279475982533,
+ "grad_norm": 0.23641522228717804,
+ "learning_rate": 0.0006,
+ "loss": 6.869574546813965,
+ "step": 3790
+ },
+ {
+ "epoch": 52.65676855895197,
+ "grad_norm": 0.23717311024665833,
+ "learning_rate": 0.0006,
+ "loss": 6.739034175872803,
+ "step": 3791
+ },
+ {
+ "epoch": 52.670742358078606,
+ "grad_norm": 0.10300362855195999,
+ "learning_rate": 0.0006,
+ "loss": 6.49191951751709,
+ "step": 3792
+ },
+ {
+ "epoch": 52.68471615720524,
+ "grad_norm": 0.1484968513250351,
+ "learning_rate": 0.0006,
+ "loss": 6.6021270751953125,
+ "step": 3793
+ },
+ {
+ "epoch": 52.698689956331876,
+ "grad_norm": 0.08052115142345428,
+ "learning_rate": 0.0006,
+ "loss": 6.432796478271484,
+ "step": 3794
+ },
+ {
+ "epoch": 52.712663755458514,
+ "grad_norm": 0.06869253516197205,
+ "learning_rate": 0.0006,
+ "loss": 6.212784767150879,
+ "step": 3795
+ },
+ {
+ "epoch": 52.72663755458515,
+ "grad_norm": 0.05589582398533821,
+ "learning_rate": 0.0006,
+ "loss": 6.168288707733154,
+ "step": 3796
+ },
+ {
+ "epoch": 52.74061135371179,
+ "grad_norm": 0.04840940609574318,
+ "learning_rate": 0.0006,
+ "loss": 6.163174629211426,
+ "step": 3797
+ },
+ {
+ "epoch": 52.75458515283843,
+ "grad_norm": 0.04578601196408272,
+ "learning_rate": 0.0006,
+ "loss": 6.038390636444092,
+ "step": 3798
+ },
+ {
+ "epoch": 52.76855895196506,
+ "grad_norm": 0.04102053865790367,
+ "learning_rate": 0.0006,
+ "loss": 5.929539680480957,
+ "step": 3799
+ },
+ {
+ "epoch": 52.7825327510917,
+ "grad_norm": 0.038398485630750656,
+ "learning_rate": 0.0006,
+ "loss": 6.007378101348877,
+ "step": 3800
+ },
+ {
+ "epoch": 52.79650655021834,
+ "grad_norm": 0.03466642275452614,
+ "learning_rate": 0.0006,
+ "loss": 5.767291069030762,
+ "step": 3801
+ },
+ {
+ "epoch": 52.81048034934498,
+ "grad_norm": 0.03328905999660492,
+ "learning_rate": 0.0006,
+ "loss": 5.785032272338867,
+ "step": 3802
+ },
+ {
+ "epoch": 52.82445414847162,
+ "grad_norm": 0.03186681121587753,
+ "learning_rate": 0.0006,
+ "loss": 5.6942057609558105,
+ "step": 3803
+ },
+ {
+ "epoch": 52.838427947598255,
+ "grad_norm": 0.030553551390767097,
+ "learning_rate": 0.0006,
+ "loss": 5.680774688720703,
+ "step": 3804
+ },
+ {
+ "epoch": 52.852401746724894,
+ "grad_norm": 0.02578839473426342,
+ "learning_rate": 0.0006,
+ "loss": 5.703461647033691,
+ "step": 3805
+ },
+ {
+ "epoch": 52.866375545851525,
+ "grad_norm": 0.026261312887072563,
+ "learning_rate": 0.0006,
+ "loss": 5.642549514770508,
+ "step": 3806
+ },
+ {
+ "epoch": 52.880349344978164,
+ "grad_norm": 0.02782430872321129,
+ "learning_rate": 0.0006,
+ "loss": 5.563597679138184,
+ "step": 3807
+ },
+ {
+ "epoch": 52.8943231441048,
+ "grad_norm": 0.024467799812555313,
+ "learning_rate": 0.0006,
+ "loss": 5.583395957946777,
+ "step": 3808
+ },
+ {
+ "epoch": 52.90829694323144,
+ "grad_norm": 0.022602897137403488,
+ "learning_rate": 0.0006,
+ "loss": 5.50742769241333,
+ "step": 3809
+ },
+ {
+ "epoch": 52.92227074235808,
+ "grad_norm": 0.024292029440402985,
+ "learning_rate": 0.0006,
+ "loss": 5.367804527282715,
+ "step": 3810
+ },
+ {
+ "epoch": 52.93624454148472,
+ "grad_norm": 0.021062295883893967,
+ "learning_rate": 0.0006,
+ "loss": 5.423705101013184,
+ "step": 3811
+ },
+ {
+ "epoch": 52.95021834061135,
+ "grad_norm": 0.022226469591259956,
+ "learning_rate": 0.0006,
+ "loss": 5.344039440155029,
+ "step": 3812
+ },
+ {
+ "epoch": 52.96419213973799,
+ "grad_norm": 0.023691752925515175,
+ "learning_rate": 0.0006,
+ "loss": 5.2581892013549805,
+ "step": 3813
+ },
+ {
+ "epoch": 52.97816593886463,
+ "grad_norm": 0.02181999757885933,
+ "learning_rate": 0.0006,
+ "loss": 5.2710113525390625,
+ "step": 3814
+ },
+ {
+ "epoch": 52.992139737991266,
+ "grad_norm": 0.025295857340097427,
+ "learning_rate": 0.0006,
+ "loss": 5.278355598449707,
+ "step": 3815
+ },
+ {
+ "epoch": 53.0,
+ "grad_norm": 0.02216828241944313,
+ "learning_rate": 0.0006,
+ "loss": 5.127945899963379,
+ "step": 3816
+ },
+ {
+ "epoch": 53.0,
+ "eval_loss": 5.3411736488342285,
+ "eval_runtime": 56.2837,
+ "eval_samples_per_second": 43.387,
+ "eval_steps_per_second": 1.368,
+ "step": 3816
+ },
+ {
+ "epoch": 53.01397379912664,
+ "grad_norm": 0.022496026009321213,
+ "learning_rate": 0.0006,
+ "loss": 5.19655704498291,
+ "step": 3817
+ },
+ {
+ "epoch": 53.02794759825328,
+ "grad_norm": 0.03201168775558472,
+ "learning_rate": 0.0006,
+ "loss": 5.094621181488037,
+ "step": 3818
+ },
+ {
+ "epoch": 53.041921397379916,
+ "grad_norm": 0.02683178335428238,
+ "learning_rate": 0.0006,
+ "loss": 5.19899845123291,
+ "step": 3819
+ },
+ {
+ "epoch": 53.05589519650655,
+ "grad_norm": 0.022659169510006905,
+ "learning_rate": 0.0006,
+ "loss": 5.184103012084961,
+ "step": 3820
+ },
+ {
+ "epoch": 53.069868995633186,
+ "grad_norm": 0.024784443899989128,
+ "learning_rate": 0.0006,
+ "loss": 5.124286651611328,
+ "step": 3821
+ },
+ {
+ "epoch": 53.083842794759825,
+ "grad_norm": 0.028710370883345604,
+ "learning_rate": 0.0006,
+ "loss": 5.130426406860352,
+ "step": 3822
+ },
+ {
+ "epoch": 53.09781659388646,
+ "grad_norm": 0.02366088517010212,
+ "learning_rate": 0.0006,
+ "loss": 5.1453399658203125,
+ "step": 3823
+ },
+ {
+ "epoch": 53.1117903930131,
+ "grad_norm": 0.01813938096165657,
+ "learning_rate": 0.0006,
+ "loss": 5.05271053314209,
+ "step": 3824
+ },
+ {
+ "epoch": 53.12576419213974,
+ "grad_norm": 0.02051435597240925,
+ "learning_rate": 0.0006,
+ "loss": 5.023038864135742,
+ "step": 3825
+ },
+ {
+ "epoch": 53.13973799126637,
+ "grad_norm": 0.02185754105448723,
+ "learning_rate": 0.0006,
+ "loss": 4.8950018882751465,
+ "step": 3826
+ },
+ {
+ "epoch": 53.15371179039301,
+ "grad_norm": 0.017223743721842766,
+ "learning_rate": 0.0006,
+ "loss": 5.078621864318848,
+ "step": 3827
+ },
+ {
+ "epoch": 53.16768558951965,
+ "grad_norm": 0.01832217164337635,
+ "learning_rate": 0.0006,
+ "loss": 4.921341419219971,
+ "step": 3828
+ },
+ {
+ "epoch": 53.18165938864629,
+ "grad_norm": 0.02114005573093891,
+ "learning_rate": 0.0006,
+ "loss": 4.997865676879883,
+ "step": 3829
+ },
+ {
+ "epoch": 53.19563318777293,
+ "grad_norm": 0.021426431834697723,
+ "learning_rate": 0.0006,
+ "loss": 4.877785682678223,
+ "step": 3830
+ },
+ {
+ "epoch": 53.209606986899566,
+ "grad_norm": 0.021327560767531395,
+ "learning_rate": 0.0006,
+ "loss": 4.847512245178223,
+ "step": 3831
+ },
+ {
+ "epoch": 53.223580786026204,
+ "grad_norm": 0.027825530618429184,
+ "learning_rate": 0.0006,
+ "loss": 5.004823207855225,
+ "step": 3832
+ },
+ {
+ "epoch": 53.237554585152836,
+ "grad_norm": 0.03756481409072876,
+ "learning_rate": 0.0006,
+ "loss": 4.817290306091309,
+ "step": 3833
+ },
+ {
+ "epoch": 53.251528384279474,
+ "grad_norm": 0.04933721572160721,
+ "learning_rate": 0.0006,
+ "loss": 4.863317489624023,
+ "step": 3834
+ },
+ {
+ "epoch": 53.26550218340611,
+ "grad_norm": 0.04588694870471954,
+ "learning_rate": 0.0006,
+ "loss": 4.864742755889893,
+ "step": 3835
+ },
+ {
+ "epoch": 53.27947598253275,
+ "grad_norm": 0.02728264033794403,
+ "learning_rate": 0.0006,
+ "loss": 4.9068756103515625,
+ "step": 3836
+ },
+ {
+ "epoch": 53.29344978165939,
+ "grad_norm": 0.036240726709365845,
+ "learning_rate": 0.0006,
+ "loss": 4.902045249938965,
+ "step": 3837
+ },
+ {
+ "epoch": 53.30742358078603,
+ "grad_norm": 0.02703195810317993,
+ "learning_rate": 0.0006,
+ "loss": 4.908980369567871,
+ "step": 3838
+ },
+ {
+ "epoch": 53.32139737991266,
+ "grad_norm": 0.024894049391150475,
+ "learning_rate": 0.0006,
+ "loss": 4.794848442077637,
+ "step": 3839
+ },
+ {
+ "epoch": 53.3353711790393,
+ "grad_norm": 0.02389993891119957,
+ "learning_rate": 0.0006,
+ "loss": 4.8234734535217285,
+ "step": 3840
+ },
+ {
+ "epoch": 53.34934497816594,
+ "grad_norm": 0.02486400306224823,
+ "learning_rate": 0.0006,
+ "loss": 4.851463794708252,
+ "step": 3841
+ },
+ {
+ "epoch": 53.36331877729258,
+ "grad_norm": 0.02613813802599907,
+ "learning_rate": 0.0006,
+ "loss": 4.860444068908691,
+ "step": 3842
+ },
+ {
+ "epoch": 53.377292576419215,
+ "grad_norm": 0.0229744054377079,
+ "learning_rate": 0.0006,
+ "loss": 4.709606170654297,
+ "step": 3843
+ },
+ {
+ "epoch": 53.391266375545854,
+ "grad_norm": 0.02123042568564415,
+ "learning_rate": 0.0006,
+ "loss": 4.654613494873047,
+ "step": 3844
+ },
+ {
+ "epoch": 53.40524017467249,
+ "grad_norm": 0.017352338880300522,
+ "learning_rate": 0.0006,
+ "loss": 4.769761085510254,
+ "step": 3845
+ },
+ {
+ "epoch": 53.419213973799124,
+ "grad_norm": 0.019808724522590637,
+ "learning_rate": 0.0006,
+ "loss": 4.98514461517334,
+ "step": 3846
+ },
+ {
+ "epoch": 53.43318777292576,
+ "grad_norm": 0.018776152282953262,
+ "learning_rate": 0.0006,
+ "loss": 4.749990463256836,
+ "step": 3847
+ },
+ {
+ "epoch": 53.4471615720524,
+ "grad_norm": 0.02017521858215332,
+ "learning_rate": 0.0006,
+ "loss": 4.757381439208984,
+ "step": 3848
+ },
+ {
+ "epoch": 53.46113537117904,
+ "grad_norm": 0.019509416073560715,
+ "learning_rate": 0.0006,
+ "loss": 4.8111572265625,
+ "step": 3849
+ },
+ {
+ "epoch": 53.47510917030568,
+ "grad_norm": 0.015543580986559391,
+ "learning_rate": 0.0006,
+ "loss": 4.7501678466796875,
+ "step": 3850
+ },
+ {
+ "epoch": 53.48908296943232,
+ "grad_norm": 0.017941651865839958,
+ "learning_rate": 0.0006,
+ "loss": 4.701838493347168,
+ "step": 3851
+ },
+ {
+ "epoch": 53.50305676855895,
+ "grad_norm": 0.01678212359547615,
+ "learning_rate": 0.0006,
+ "loss": 4.787798881530762,
+ "step": 3852
+ },
+ {
+ "epoch": 53.51703056768559,
+ "grad_norm": 0.019050654023885727,
+ "learning_rate": 0.0006,
+ "loss": 4.754859924316406,
+ "step": 3853
+ },
+ {
+ "epoch": 53.531004366812226,
+ "grad_norm": 0.017406078055500984,
+ "learning_rate": 0.0006,
+ "loss": 4.752772808074951,
+ "step": 3854
+ },
+ {
+ "epoch": 53.544978165938865,
+ "grad_norm": 0.01773344725370407,
+ "learning_rate": 0.0006,
+ "loss": 4.651812553405762,
+ "step": 3855
+ },
+ {
+ "epoch": 53.5589519650655,
+ "grad_norm": 0.017221346497535706,
+ "learning_rate": 0.0006,
+ "loss": 4.792244911193848,
+ "step": 3856
+ },
+ {
+ "epoch": 53.57292576419214,
+ "grad_norm": 0.01722758449614048,
+ "learning_rate": 0.0006,
+ "loss": 4.7902960777282715,
+ "step": 3857
+ },
+ {
+ "epoch": 53.58689956331878,
+ "grad_norm": 0.020445192232728004,
+ "learning_rate": 0.0006,
+ "loss": 4.67392110824585,
+ "step": 3858
+ },
+ {
+ "epoch": 53.60087336244541,
+ "grad_norm": 0.02267114631831646,
+ "learning_rate": 0.0006,
+ "loss": 4.796089172363281,
+ "step": 3859
+ },
+ {
+ "epoch": 53.61484716157205,
+ "grad_norm": 0.027129942551255226,
+ "learning_rate": 0.0006,
+ "loss": 4.774303913116455,
+ "step": 3860
+ },
+ {
+ "epoch": 53.62882096069869,
+ "grad_norm": 0.025454839691519737,
+ "learning_rate": 0.0006,
+ "loss": 4.708813190460205,
+ "step": 3861
+ },
+ {
+ "epoch": 53.64279475982533,
+ "grad_norm": 0.019248757511377335,
+ "learning_rate": 0.0006,
+ "loss": 4.760467529296875,
+ "step": 3862
+ },
+ {
+ "epoch": 53.65676855895197,
+ "grad_norm": 0.01791176199913025,
+ "learning_rate": 0.0006,
+ "loss": 4.789203643798828,
+ "step": 3863
+ },
+ {
+ "epoch": 53.670742358078606,
+ "grad_norm": 0.016306612640619278,
+ "learning_rate": 0.0006,
+ "loss": 4.732605934143066,
+ "step": 3864
+ },
+ {
+ "epoch": 53.68471615720524,
+ "grad_norm": 0.017836976796388626,
+ "learning_rate": 0.0006,
+ "loss": 4.691871643066406,
+ "step": 3865
+ },
+ {
+ "epoch": 53.698689956331876,
+ "grad_norm": 0.015102861449122429,
+ "learning_rate": 0.0006,
+ "loss": 4.724156856536865,
+ "step": 3866
+ },
+ {
+ "epoch": 53.712663755458514,
+ "grad_norm": 0.017458142712712288,
+ "learning_rate": 0.0006,
+ "loss": 4.73524808883667,
+ "step": 3867
+ },
+ {
+ "epoch": 53.72663755458515,
+ "grad_norm": 0.01591932773590088,
+ "learning_rate": 0.0006,
+ "loss": 4.705888748168945,
+ "step": 3868
+ },
+ {
+ "epoch": 53.74061135371179,
+ "grad_norm": 0.01713593676686287,
+ "learning_rate": 0.0006,
+ "loss": 4.644516944885254,
+ "step": 3869
+ },
+ {
+ "epoch": 53.75458515283843,
+ "grad_norm": 0.01905885338783264,
+ "learning_rate": 0.0006,
+ "loss": 4.698416233062744,
+ "step": 3870
+ },
+ {
+ "epoch": 53.76855895196506,
+ "grad_norm": 0.01766437292098999,
+ "learning_rate": 0.0006,
+ "loss": 4.6796555519104,
+ "step": 3871
+ },
+ {
+ "epoch": 53.7825327510917,
+ "grad_norm": 0.01552243810147047,
+ "learning_rate": 0.0006,
+ "loss": 4.78138542175293,
+ "step": 3872
+ },
+ {
+ "epoch": 53.79650655021834,
+ "grad_norm": 0.014191331341862679,
+ "learning_rate": 0.0006,
+ "loss": 4.703588485717773,
+ "step": 3873
+ },
+ {
+ "epoch": 53.81048034934498,
+ "grad_norm": 0.015872279182076454,
+ "learning_rate": 0.0006,
+ "loss": 4.705611228942871,
+ "step": 3874
+ },
+ {
+ "epoch": 53.82445414847162,
+ "grad_norm": 0.01787560060620308,
+ "learning_rate": 0.0006,
+ "loss": 4.8227949142456055,
+ "step": 3875
+ },
+ {
+ "epoch": 53.838427947598255,
+ "grad_norm": 0.0158209390938282,
+ "learning_rate": 0.0006,
+ "loss": 4.645227432250977,
+ "step": 3876
+ },
+ {
+ "epoch": 53.852401746724894,
+ "grad_norm": 0.014362678863108158,
+ "learning_rate": 0.0006,
+ "loss": 4.7225422859191895,
+ "step": 3877
+ },
+ {
+ "epoch": 53.866375545851525,
+ "grad_norm": 0.014899161644279957,
+ "learning_rate": 0.0006,
+ "loss": 4.685914039611816,
+ "step": 3878
+ },
+ {
+ "epoch": 53.880349344978164,
+ "grad_norm": 0.01628505438566208,
+ "learning_rate": 0.0006,
+ "loss": 4.674065589904785,
+ "step": 3879
+ },
+ {
+ "epoch": 53.8943231441048,
+ "grad_norm": 0.01873331144452095,
+ "learning_rate": 0.0006,
+ "loss": 4.811744213104248,
+ "step": 3880
+ },
+ {
+ "epoch": 53.90829694323144,
+ "grad_norm": 0.02229069173336029,
+ "learning_rate": 0.0006,
+ "loss": 4.721514701843262,
+ "step": 3881
+ },
+ {
+ "epoch": 53.92227074235808,
+ "grad_norm": 0.028012916445732117,
+ "learning_rate": 0.0006,
+ "loss": 4.773603439331055,
+ "step": 3882
+ },
+ {
+ "epoch": 53.93624454148472,
+ "grad_norm": 0.036322783678770065,
+ "learning_rate": 0.0006,
+ "loss": 4.787134170532227,
+ "step": 3883
+ },
+ {
+ "epoch": 53.95021834061135,
+ "grad_norm": 0.048149142414331436,
+ "learning_rate": 0.0006,
+ "loss": 4.736196517944336,
+ "step": 3884
+ },
+ {
+ "epoch": 53.96419213973799,
+ "grad_norm": 0.051918916404247284,
+ "learning_rate": 0.0006,
+ "loss": 4.686728000640869,
+ "step": 3885
+ },
+ {
+ "epoch": 53.97816593886463,
+ "grad_norm": 0.0481342189013958,
+ "learning_rate": 0.0006,
+ "loss": 4.733730792999268,
+ "step": 3886
+ },
+ {
+ "epoch": 53.992139737991266,
+ "grad_norm": 0.045424818992614746,
+ "learning_rate": 0.0006,
+ "loss": 4.748723983764648,
+ "step": 3887
+ },
+ {
+ "epoch": 54.0,
+ "grad_norm": 0.04153316840529442,
+ "learning_rate": 0.0006,
+ "loss": 4.715220928192139,
+ "step": 3888
+ },
+ {
+ "epoch": 54.0,
+ "eval_loss": 4.964215278625488,
+ "eval_runtime": 56.7757,
+ "eval_samples_per_second": 43.011,
+ "eval_steps_per_second": 1.356,
+ "step": 3888
+ },
+ {
+ "epoch": 54.01397379912664,
+ "grad_norm": 0.03459750860929489,
+ "learning_rate": 0.0006,
+ "loss": 4.661557674407959,
+ "step": 3889
+ },
+ {
+ "epoch": 54.02794759825328,
+ "grad_norm": 0.025409208610653877,
+ "learning_rate": 0.0006,
+ "loss": 4.7065582275390625,
+ "step": 3890
+ },
+ {
+ "epoch": 54.041921397379916,
+ "grad_norm": 0.03202550485730171,
+ "learning_rate": 0.0006,
+ "loss": 4.751070976257324,
+ "step": 3891
+ },
+ {
+ "epoch": 54.05589519650655,
+ "grad_norm": 0.02968018501996994,
+ "learning_rate": 0.0006,
+ "loss": 4.635388374328613,
+ "step": 3892
+ },
+ {
+ "epoch": 54.069868995633186,
+ "grad_norm": 0.02757406234741211,
+ "learning_rate": 0.0006,
+ "loss": 4.706214427947998,
+ "step": 3893
+ },
+ {
+ "epoch": 54.083842794759825,
+ "grad_norm": 0.025223620235919952,
+ "learning_rate": 0.0006,
+ "loss": 4.703792572021484,
+ "step": 3894
+ },
+ {
+ "epoch": 54.09781659388646,
+ "grad_norm": 0.02431778982281685,
+ "learning_rate": 0.0006,
+ "loss": 4.658116340637207,
+ "step": 3895
+ },
+ {
+ "epoch": 54.1117903930131,
+ "grad_norm": 0.02792172320187092,
+ "learning_rate": 0.0006,
+ "loss": 4.638895034790039,
+ "step": 3896
+ },
+ {
+ "epoch": 54.12576419213974,
+ "grad_norm": 0.032253511250019073,
+ "learning_rate": 0.0006,
+ "loss": 4.693902492523193,
+ "step": 3897
+ },
+ {
+ "epoch": 54.13973799126637,
+ "grad_norm": 0.027044931426644325,
+ "learning_rate": 0.0006,
+ "loss": 4.575805187225342,
+ "step": 3898
+ },
+ {
+ "epoch": 54.15371179039301,
+ "grad_norm": 0.023270361125469208,
+ "learning_rate": 0.0006,
+ "loss": 4.67645263671875,
+ "step": 3899
+ },
+ {
+ "epoch": 54.16768558951965,
+ "grad_norm": 0.01929079182446003,
+ "learning_rate": 0.0006,
+ "loss": 4.68062686920166,
+ "step": 3900
+ },
+ {
+ "epoch": 54.18165938864629,
+ "grad_norm": 0.020343473181128502,
+ "learning_rate": 0.0006,
+ "loss": 4.708086967468262,
+ "step": 3901
+ },
+ {
+ "epoch": 54.19563318777293,
+ "grad_norm": 0.0211793202906847,
+ "learning_rate": 0.0006,
+ "loss": 4.677323341369629,
+ "step": 3902
+ },
+ {
+ "epoch": 54.209606986899566,
+ "grad_norm": 0.020341217517852783,
+ "learning_rate": 0.0006,
+ "loss": 4.760937690734863,
+ "step": 3903
+ },
+ {
+ "epoch": 54.223580786026204,
+ "grad_norm": 0.019421054050326347,
+ "learning_rate": 0.0006,
+ "loss": 4.508834362030029,
+ "step": 3904
+ },
+ {
+ "epoch": 54.237554585152836,
+ "grad_norm": 0.016701679676771164,
+ "learning_rate": 0.0006,
+ "loss": 4.664399147033691,
+ "step": 3905
+ },
+ {
+ "epoch": 54.251528384279474,
+ "grad_norm": 0.016561470925807953,
+ "learning_rate": 0.0006,
+ "loss": 4.742990493774414,
+ "step": 3906
+ },
+ {
+ "epoch": 54.26550218340611,
+ "grad_norm": 0.017446085810661316,
+ "learning_rate": 0.0006,
+ "loss": 4.63038444519043,
+ "step": 3907
+ },
+ {
+ "epoch": 54.27947598253275,
+ "grad_norm": 0.014340086840093136,
+ "learning_rate": 0.0006,
+ "loss": 4.688791275024414,
+ "step": 3908
+ },
+ {
+ "epoch": 54.29344978165939,
+ "grad_norm": 0.013580698519945145,
+ "learning_rate": 0.0006,
+ "loss": 4.625302314758301,
+ "step": 3909
+ },
+ {
+ "epoch": 54.30742358078603,
+ "grad_norm": 0.014931539073586464,
+ "learning_rate": 0.0006,
+ "loss": 4.6167497634887695,
+ "step": 3910
+ },
+ {
+ "epoch": 54.32139737991266,
+ "grad_norm": 0.013492150232195854,
+ "learning_rate": 0.0006,
+ "loss": 4.657521724700928,
+ "step": 3911
+ },
+ {
+ "epoch": 54.3353711790393,
+ "grad_norm": 0.015151413157582283,
+ "learning_rate": 0.0006,
+ "loss": 4.634521484375,
+ "step": 3912
+ },
+ {
+ "epoch": 54.34934497816594,
+ "grad_norm": 0.01302040833979845,
+ "learning_rate": 0.0006,
+ "loss": 4.718928337097168,
+ "step": 3913
+ },
+ {
+ "epoch": 54.36331877729258,
+ "grad_norm": 0.013433529995381832,
+ "learning_rate": 0.0006,
+ "loss": 4.584803581237793,
+ "step": 3914
+ },
+ {
+ "epoch": 54.377292576419215,
+ "grad_norm": 0.012987039051949978,
+ "learning_rate": 0.0006,
+ "loss": 4.6393585205078125,
+ "step": 3915
+ },
+ {
+ "epoch": 54.391266375545854,
+ "grad_norm": 0.013695075176656246,
+ "learning_rate": 0.0006,
+ "loss": 4.556313514709473,
+ "step": 3916
+ },
+ {
+ "epoch": 54.40524017467249,
+ "grad_norm": 0.013718255795538425,
+ "learning_rate": 0.0006,
+ "loss": 4.576223850250244,
+ "step": 3917
+ },
+ {
+ "epoch": 54.419213973799124,
+ "grad_norm": 0.013699112460017204,
+ "learning_rate": 0.0006,
+ "loss": 4.63889217376709,
+ "step": 3918
+ },
+ {
+ "epoch": 54.43318777292576,
+ "grad_norm": 0.013516264036297798,
+ "learning_rate": 0.0006,
+ "loss": 4.669532299041748,
+ "step": 3919
+ },
+ {
+ "epoch": 54.4471615720524,
+ "grad_norm": 0.013597294688224792,
+ "learning_rate": 0.0006,
+ "loss": 4.6644287109375,
+ "step": 3920
+ },
+ {
+ "epoch": 54.46113537117904,
+ "grad_norm": 0.013185849413275719,
+ "learning_rate": 0.0006,
+ "loss": 4.636900901794434,
+ "step": 3921
+ },
+ {
+ "epoch": 54.47510917030568,
+ "grad_norm": 0.014311990700662136,
+ "learning_rate": 0.0006,
+ "loss": 4.6047587394714355,
+ "step": 3922
+ },
+ {
+ "epoch": 54.48908296943232,
+ "grad_norm": 0.012426027096807957,
+ "learning_rate": 0.0006,
+ "loss": 4.6161394119262695,
+ "step": 3923
+ },
+ {
+ "epoch": 54.50305676855895,
+ "grad_norm": 0.01283493172377348,
+ "learning_rate": 0.0006,
+ "loss": 4.584190368652344,
+ "step": 3924
+ },
+ {
+ "epoch": 54.51703056768559,
+ "grad_norm": 0.01339943427592516,
+ "learning_rate": 0.0006,
+ "loss": 4.624184608459473,
+ "step": 3925
+ },
+ {
+ "epoch": 54.531004366812226,
+ "grad_norm": 0.012859724462032318,
+ "learning_rate": 0.0006,
+ "loss": 4.6174635887146,
+ "step": 3926
+ },
+ {
+ "epoch": 54.544978165938865,
+ "grad_norm": 0.014420023187994957,
+ "learning_rate": 0.0006,
+ "loss": 4.536544322967529,
+ "step": 3927
+ },
+ {
+ "epoch": 54.5589519650655,
+ "grad_norm": 0.014237185008823872,
+ "learning_rate": 0.0006,
+ "loss": 4.681514739990234,
+ "step": 3928
+ },
+ {
+ "epoch": 54.57292576419214,
+ "grad_norm": 0.013340328820049763,
+ "learning_rate": 0.0006,
+ "loss": 4.579689979553223,
+ "step": 3929
+ },
+ {
+ "epoch": 54.58689956331878,
+ "grad_norm": 0.013894530944526196,
+ "learning_rate": 0.0006,
+ "loss": 4.638958930969238,
+ "step": 3930
+ },
+ {
+ "epoch": 54.60087336244541,
+ "grad_norm": 0.016181141138076782,
+ "learning_rate": 0.0006,
+ "loss": 4.573508262634277,
+ "step": 3931
+ },
+ {
+ "epoch": 54.61484716157205,
+ "grad_norm": 0.013221992179751396,
+ "learning_rate": 0.0006,
+ "loss": 4.6229047775268555,
+ "step": 3932
+ },
+ {
+ "epoch": 54.62882096069869,
+ "grad_norm": 0.013241561129689217,
+ "learning_rate": 0.0006,
+ "loss": 4.641152858734131,
+ "step": 3933
+ },
+ {
+ "epoch": 54.64279475982533,
+ "grad_norm": 0.016237538307905197,
+ "learning_rate": 0.0006,
+ "loss": 4.6421217918396,
+ "step": 3934
+ },
+ {
+ "epoch": 54.65676855895197,
+ "grad_norm": 0.020866507664322853,
+ "learning_rate": 0.0006,
+ "loss": 4.646058082580566,
+ "step": 3935
+ },
+ {
+ "epoch": 54.670742358078606,
+ "grad_norm": 0.022182893007993698,
+ "learning_rate": 0.0006,
+ "loss": 4.554501533508301,
+ "step": 3936
+ },
+ {
+ "epoch": 54.68471615720524,
+ "grad_norm": 0.02254868485033512,
+ "learning_rate": 0.0006,
+ "loss": 4.563869953155518,
+ "step": 3937
+ },
+ {
+ "epoch": 54.698689956331876,
+ "grad_norm": 0.019730940461158752,
+ "learning_rate": 0.0006,
+ "loss": 4.592165946960449,
+ "step": 3938
+ },
+ {
+ "epoch": 54.712663755458514,
+ "grad_norm": 0.015146561898291111,
+ "learning_rate": 0.0006,
+ "loss": 4.6228742599487305,
+ "step": 3939
+ },
+ {
+ "epoch": 54.72663755458515,
+ "grad_norm": 0.01824716106057167,
+ "learning_rate": 0.0006,
+ "loss": 4.716236114501953,
+ "step": 3940
+ },
+ {
+ "epoch": 54.74061135371179,
+ "grad_norm": 0.017730550840497017,
+ "learning_rate": 0.0006,
+ "loss": 4.650232315063477,
+ "step": 3941
+ },
+ {
+ "epoch": 54.75458515283843,
+ "grad_norm": 0.014875595457851887,
+ "learning_rate": 0.0006,
+ "loss": 4.6377458572387695,
+ "step": 3942
+ },
+ {
+ "epoch": 54.76855895196506,
+ "grad_norm": 0.01613209955394268,
+ "learning_rate": 0.0006,
+ "loss": 4.5768537521362305,
+ "step": 3943
+ },
+ {
+ "epoch": 54.7825327510917,
+ "grad_norm": 0.017311280593276024,
+ "learning_rate": 0.0006,
+ "loss": 4.625744342803955,
+ "step": 3944
+ },
+ {
+ "epoch": 54.79650655021834,
+ "grad_norm": 0.018770085647702217,
+ "learning_rate": 0.0006,
+ "loss": 4.613787651062012,
+ "step": 3945
+ },
+ {
+ "epoch": 54.81048034934498,
+ "grad_norm": 0.0165257528424263,
+ "learning_rate": 0.0006,
+ "loss": 4.5991315841674805,
+ "step": 3946
+ },
+ {
+ "epoch": 54.82445414847162,
+ "grad_norm": 0.017433857545256615,
+ "learning_rate": 0.0006,
+ "loss": 4.513872146606445,
+ "step": 3947
+ },
+ {
+ "epoch": 54.838427947598255,
+ "grad_norm": 0.020126905292272568,
+ "learning_rate": 0.0006,
+ "loss": 4.616185188293457,
+ "step": 3948
+ },
+ {
+ "epoch": 54.852401746724894,
+ "grad_norm": 0.01819358579814434,
+ "learning_rate": 0.0006,
+ "loss": 4.535691261291504,
+ "step": 3949
+ },
+ {
+ "epoch": 54.866375545851525,
+ "grad_norm": 0.01899927295744419,
+ "learning_rate": 0.0006,
+ "loss": 4.595199108123779,
+ "step": 3950
+ },
+ {
+ "epoch": 54.880349344978164,
+ "grad_norm": 0.02005566656589508,
+ "learning_rate": 0.0006,
+ "loss": 4.633289337158203,
+ "step": 3951
+ },
+ {
+ "epoch": 54.8943231441048,
+ "grad_norm": 0.021570220589637756,
+ "learning_rate": 0.0006,
+ "loss": 4.592503547668457,
+ "step": 3952
+ },
+ {
+ "epoch": 54.90829694323144,
+ "grad_norm": 0.01883506216108799,
+ "learning_rate": 0.0006,
+ "loss": 4.6597795486450195,
+ "step": 3953
+ },
+ {
+ "epoch": 54.92227074235808,
+ "grad_norm": 0.018070021644234657,
+ "learning_rate": 0.0006,
+ "loss": 4.601474285125732,
+ "step": 3954
+ },
+ {
+ "epoch": 54.93624454148472,
+ "grad_norm": 0.018090378493070602,
+ "learning_rate": 0.0006,
+ "loss": 4.569846153259277,
+ "step": 3955
+ },
+ {
+ "epoch": 54.95021834061135,
+ "grad_norm": 0.018751563504338264,
+ "learning_rate": 0.0006,
+ "loss": 4.646064281463623,
+ "step": 3956
+ },
+ {
+ "epoch": 54.96419213973799,
+ "grad_norm": 0.017512254416942596,
+ "learning_rate": 0.0006,
+ "loss": 4.606670379638672,
+ "step": 3957
+ },
+ {
+ "epoch": 54.97816593886463,
+ "grad_norm": 0.018771110102534294,
+ "learning_rate": 0.0006,
+ "loss": 4.604315757751465,
+ "step": 3958
+ },
+ {
+ "epoch": 54.992139737991266,
+ "grad_norm": 0.01611141301691532,
+ "learning_rate": 0.0006,
+ "loss": 4.643245220184326,
+ "step": 3959
+ },
+ {
+ "epoch": 55.0,
+ "grad_norm": 0.0156728383153677,
+ "learning_rate": 0.0006,
+ "loss": 4.6739044189453125,
+ "step": 3960
+ },
+ {
+ "epoch": 55.0,
+ "eval_loss": 4.851032257080078,
+ "eval_runtime": 56.8699,
+ "eval_samples_per_second": 42.94,
+ "eval_steps_per_second": 1.354,
+ "step": 3960
+ },
+ {
+ "epoch": 55.01397379912664,
+ "grad_norm": 0.016214672476053238,
+ "learning_rate": 0.0006,
+ "loss": 4.571749687194824,
+ "step": 3961
+ },
+ {
+ "epoch": 55.02794759825328,
+ "grad_norm": 0.01649155467748642,
+ "learning_rate": 0.0006,
+ "loss": 4.553592205047607,
+ "step": 3962
+ },
+ {
+ "epoch": 55.041921397379916,
+ "grad_norm": 0.016793379560112953,
+ "learning_rate": 0.0006,
+ "loss": 4.650628566741943,
+ "step": 3963
+ },
+ {
+ "epoch": 55.05589519650655,
+ "grad_norm": 0.0191937405616045,
+ "learning_rate": 0.0006,
+ "loss": 4.620532989501953,
+ "step": 3964
+ },
+ {
+ "epoch": 55.069868995633186,
+ "grad_norm": 0.020976556465029716,
+ "learning_rate": 0.0006,
+ "loss": 4.6329545974731445,
+ "step": 3965
+ },
+ {
+ "epoch": 55.083842794759825,
+ "grad_norm": 0.023450477048754692,
+ "learning_rate": 0.0006,
+ "loss": 4.60828971862793,
+ "step": 3966
+ },
+ {
+ "epoch": 55.09781659388646,
+ "grad_norm": 0.026778312399983406,
+ "learning_rate": 0.0006,
+ "loss": 4.617481708526611,
+ "step": 3967
+ },
+ {
+ "epoch": 55.1117903930131,
+ "grad_norm": 0.02333568036556244,
+ "learning_rate": 0.0006,
+ "loss": 4.580722332000732,
+ "step": 3968
+ },
+ {
+ "epoch": 55.12576419213974,
+ "grad_norm": 0.018491310998797417,
+ "learning_rate": 0.0006,
+ "loss": 4.590953350067139,
+ "step": 3969
+ },
+ {
+ "epoch": 55.13973799126637,
+ "grad_norm": 0.020339803770184517,
+ "learning_rate": 0.0006,
+ "loss": 4.518374443054199,
+ "step": 3970
+ },
+ {
+ "epoch": 55.15371179039301,
+ "grad_norm": 0.01759706623852253,
+ "learning_rate": 0.0006,
+ "loss": 4.560522079467773,
+ "step": 3971
+ },
+ {
+ "epoch": 55.16768558951965,
+ "grad_norm": 0.015875980257987976,
+ "learning_rate": 0.0006,
+ "loss": 4.4795708656311035,
+ "step": 3972
+ },
+ {
+ "epoch": 55.18165938864629,
+ "grad_norm": 0.01650134287774563,
+ "learning_rate": 0.0006,
+ "loss": 4.57435941696167,
+ "step": 3973
+ },
+ {
+ "epoch": 55.19563318777293,
+ "grad_norm": 0.01734331250190735,
+ "learning_rate": 0.0006,
+ "loss": 4.661007881164551,
+ "step": 3974
+ },
+ {
+ "epoch": 55.209606986899566,
+ "grad_norm": 0.01651303470134735,
+ "learning_rate": 0.0006,
+ "loss": 4.511148452758789,
+ "step": 3975
+ },
+ {
+ "epoch": 55.223580786026204,
+ "grad_norm": 0.016189761459827423,
+ "learning_rate": 0.0006,
+ "loss": 4.592350006103516,
+ "step": 3976
+ },
+ {
+ "epoch": 55.237554585152836,
+ "grad_norm": 0.01833222061395645,
+ "learning_rate": 0.0006,
+ "loss": 4.506649971008301,
+ "step": 3977
+ },
+ {
+ "epoch": 55.251528384279474,
+ "grad_norm": 0.02433410845696926,
+ "learning_rate": 0.0006,
+ "loss": 4.49500036239624,
+ "step": 3978
+ },
+ {
+ "epoch": 55.26550218340611,
+ "grad_norm": 0.0314289927482605,
+ "learning_rate": 0.0006,
+ "loss": 4.507624626159668,
+ "step": 3979
+ },
+ {
+ "epoch": 55.27947598253275,
+ "grad_norm": 0.03013679012656212,
+ "learning_rate": 0.0006,
+ "loss": 4.533641815185547,
+ "step": 3980
+ },
+ {
+ "epoch": 55.29344978165939,
+ "grad_norm": 0.024524061009287834,
+ "learning_rate": 0.0006,
+ "loss": 4.5721116065979,
+ "step": 3981
+ },
+ {
+ "epoch": 55.30742358078603,
+ "grad_norm": 0.02640336938202381,
+ "learning_rate": 0.0006,
+ "loss": 4.652521133422852,
+ "step": 3982
+ },
+ {
+ "epoch": 55.32139737991266,
+ "grad_norm": 0.021309101954102516,
+ "learning_rate": 0.0006,
+ "loss": 4.600375175476074,
+ "step": 3983
+ },
+ {
+ "epoch": 55.3353711790393,
+ "grad_norm": 0.023008549585938454,
+ "learning_rate": 0.0006,
+ "loss": 4.53914737701416,
+ "step": 3984
+ },
+ {
+ "epoch": 55.34934497816594,
+ "grad_norm": 0.022491468116641045,
+ "learning_rate": 0.0006,
+ "loss": 4.661163330078125,
+ "step": 3985
+ },
+ {
+ "epoch": 55.36331877729258,
+ "grad_norm": 0.01904081553220749,
+ "learning_rate": 0.0006,
+ "loss": 4.5426177978515625,
+ "step": 3986
+ },
+ {
+ "epoch": 55.377292576419215,
+ "grad_norm": 0.019229518249630928,
+ "learning_rate": 0.0006,
+ "loss": 4.606066703796387,
+ "step": 3987
+ },
+ {
+ "epoch": 55.391266375545854,
+ "grad_norm": 0.01781383901834488,
+ "learning_rate": 0.0006,
+ "loss": 4.54642391204834,
+ "step": 3988
+ },
+ {
+ "epoch": 55.40524017467249,
+ "grad_norm": 0.0174600500613451,
+ "learning_rate": 0.0006,
+ "loss": 4.711891174316406,
+ "step": 3989
+ },
+ {
+ "epoch": 55.419213973799124,
+ "grad_norm": 0.018154339864850044,
+ "learning_rate": 0.0006,
+ "loss": 4.527122974395752,
+ "step": 3990
+ },
+ {
+ "epoch": 55.43318777292576,
+ "grad_norm": 0.01914648339152336,
+ "learning_rate": 0.0006,
+ "loss": 4.641975402832031,
+ "step": 3991
+ },
+ {
+ "epoch": 55.4471615720524,
+ "grad_norm": 0.017118625342845917,
+ "learning_rate": 0.0006,
+ "loss": 4.511425971984863,
+ "step": 3992
+ },
+ {
+ "epoch": 55.46113537117904,
+ "grad_norm": 0.01551489345729351,
+ "learning_rate": 0.0006,
+ "loss": 4.551141738891602,
+ "step": 3993
+ },
+ {
+ "epoch": 55.47510917030568,
+ "grad_norm": 0.015072065405547619,
+ "learning_rate": 0.0006,
+ "loss": 4.651115894317627,
+ "step": 3994
+ },
+ {
+ "epoch": 55.48908296943232,
+ "grad_norm": 0.017010986804962158,
+ "learning_rate": 0.0006,
+ "loss": 4.583282470703125,
+ "step": 3995
+ },
+ {
+ "epoch": 55.50305676855895,
+ "grad_norm": 0.01764831133186817,
+ "learning_rate": 0.0006,
+ "loss": 4.593364238739014,
+ "step": 3996
+ },
+ {
+ "epoch": 55.51703056768559,
+ "grad_norm": 0.01762884296476841,
+ "learning_rate": 0.0006,
+ "loss": 4.45671272277832,
+ "step": 3997
+ },
+ {
+ "epoch": 55.531004366812226,
+ "grad_norm": 0.01600457727909088,
+ "learning_rate": 0.0006,
+ "loss": 4.511260032653809,
+ "step": 3998
+ },
+ {
+ "epoch": 55.544978165938865,
+ "grad_norm": 0.015706021338701248,
+ "learning_rate": 0.0006,
+ "loss": 4.47593355178833,
+ "step": 3999
+ },
+ {
+ "epoch": 55.5589519650655,
+ "grad_norm": 0.017718996852636337,
+ "learning_rate": 0.0006,
+ "loss": 4.610543251037598,
+ "step": 4000
+ },
+ {
+ "epoch": 55.57292576419214,
+ "grad_norm": 0.015073757618665695,
+ "learning_rate": 0.0006,
+ "loss": 4.466620445251465,
+ "step": 4001
+ },
+ {
+ "epoch": 55.58689956331878,
+ "grad_norm": 0.01789247989654541,
+ "learning_rate": 0.0006,
+ "loss": 4.503409385681152,
+ "step": 4002
+ },
+ {
+ "epoch": 55.60087336244541,
+ "grad_norm": 0.022329598665237427,
+ "learning_rate": 0.0006,
+ "loss": 4.62099552154541,
+ "step": 4003
+ },
+ {
+ "epoch": 55.61484716157205,
+ "grad_norm": 0.02343829534947872,
+ "learning_rate": 0.0006,
+ "loss": 4.620813369750977,
+ "step": 4004
+ },
+ {
+ "epoch": 55.62882096069869,
+ "grad_norm": 0.023279687389731407,
+ "learning_rate": 0.0006,
+ "loss": 4.545765399932861,
+ "step": 4005
+ },
+ {
+ "epoch": 55.64279475982533,
+ "grad_norm": 0.020817464217543602,
+ "learning_rate": 0.0006,
+ "loss": 4.561470031738281,
+ "step": 4006
+ },
+ {
+ "epoch": 55.65676855895197,
+ "grad_norm": 0.020270058885216713,
+ "learning_rate": 0.0006,
+ "loss": 4.5875020027160645,
+ "step": 4007
+ },
+ {
+ "epoch": 55.670742358078606,
+ "grad_norm": 0.021085843443870544,
+ "learning_rate": 0.0006,
+ "loss": 4.504560470581055,
+ "step": 4008
+ },
+ {
+ "epoch": 55.68471615720524,
+ "grad_norm": 0.018507730215787888,
+ "learning_rate": 0.0006,
+ "loss": 4.520644664764404,
+ "step": 4009
+ },
+ {
+ "epoch": 55.698689956331876,
+ "grad_norm": 0.020571516826748848,
+ "learning_rate": 0.0006,
+ "loss": 4.539643287658691,
+ "step": 4010
+ },
+ {
+ "epoch": 55.712663755458514,
+ "grad_norm": 0.021484510973095894,
+ "learning_rate": 0.0006,
+ "loss": 4.628833293914795,
+ "step": 4011
+ },
+ {
+ "epoch": 55.72663755458515,
+ "grad_norm": 0.017693819478154182,
+ "learning_rate": 0.0006,
+ "loss": 4.6023783683776855,
+ "step": 4012
+ },
+ {
+ "epoch": 55.74061135371179,
+ "grad_norm": 0.020186059176921844,
+ "learning_rate": 0.0006,
+ "loss": 4.634884834289551,
+ "step": 4013
+ },
+ {
+ "epoch": 55.75458515283843,
+ "grad_norm": 0.01835780404508114,
+ "learning_rate": 0.0006,
+ "loss": 4.577791213989258,
+ "step": 4014
+ },
+ {
+ "epoch": 55.76855895196506,
+ "grad_norm": 0.018162988126277924,
+ "learning_rate": 0.0006,
+ "loss": 4.641597270965576,
+ "step": 4015
+ },
+ {
+ "epoch": 55.7825327510917,
+ "grad_norm": 0.018909504637122154,
+ "learning_rate": 0.0006,
+ "loss": 4.487447738647461,
+ "step": 4016
+ },
+ {
+ "epoch": 55.79650655021834,
+ "grad_norm": 0.018972251564264297,
+ "learning_rate": 0.0006,
+ "loss": 4.590317726135254,
+ "step": 4017
+ },
+ {
+ "epoch": 55.81048034934498,
+ "grad_norm": 0.01793607883155346,
+ "learning_rate": 0.0006,
+ "loss": 4.4994611740112305,
+ "step": 4018
+ },
+ {
+ "epoch": 55.82445414847162,
+ "grad_norm": 0.02123514749109745,
+ "learning_rate": 0.0006,
+ "loss": 4.479006767272949,
+ "step": 4019
+ },
+ {
+ "epoch": 55.838427947598255,
+ "grad_norm": 0.021246599033474922,
+ "learning_rate": 0.0006,
+ "loss": 4.4738640785217285,
+ "step": 4020
+ },
+ {
+ "epoch": 55.852401746724894,
+ "grad_norm": 0.020131602883338928,
+ "learning_rate": 0.0006,
+ "loss": 4.600308418273926,
+ "step": 4021
+ },
+ {
+ "epoch": 55.866375545851525,
+ "grad_norm": 0.019159460440278053,
+ "learning_rate": 0.0006,
+ "loss": 4.540908336639404,
+ "step": 4022
+ },
+ {
+ "epoch": 55.880349344978164,
+ "grad_norm": 0.023114705458283424,
+ "learning_rate": 0.0006,
+ "loss": 4.641247749328613,
+ "step": 4023
+ },
+ {
+ "epoch": 55.8943231441048,
+ "grad_norm": 0.02296135015785694,
+ "learning_rate": 0.0006,
+ "loss": 4.444624423980713,
+ "step": 4024
+ },
+ {
+ "epoch": 55.90829694323144,
+ "grad_norm": 0.02204655110836029,
+ "learning_rate": 0.0006,
+ "loss": 4.506349563598633,
+ "step": 4025
+ },
+ {
+ "epoch": 55.92227074235808,
+ "grad_norm": 0.019338225945830345,
+ "learning_rate": 0.0006,
+ "loss": 4.616047382354736,
+ "step": 4026
+ },
+ {
+ "epoch": 55.93624454148472,
+ "grad_norm": 0.020357884466648102,
+ "learning_rate": 0.0006,
+ "loss": 4.48170280456543,
+ "step": 4027
+ },
+ {
+ "epoch": 55.95021834061135,
+ "grad_norm": 0.02426772564649582,
+ "learning_rate": 0.0006,
+ "loss": 4.576877117156982,
+ "step": 4028
+ },
+ {
+ "epoch": 55.96419213973799,
+ "grad_norm": 0.023351816460490227,
+ "learning_rate": 0.0006,
+ "loss": 4.458885192871094,
+ "step": 4029
+ },
+ {
+ "epoch": 55.97816593886463,
+ "grad_norm": 0.019892243668437004,
+ "learning_rate": 0.0006,
+ "loss": 4.520229816436768,
+ "step": 4030
+ },
+ {
+ "epoch": 55.992139737991266,
+ "grad_norm": 0.018331089988350868,
+ "learning_rate": 0.0006,
+ "loss": 4.432201862335205,
+ "step": 4031
+ },
+ {
+ "epoch": 56.0,
+ "grad_norm": 0.021300828084349632,
+ "learning_rate": 0.0006,
+ "loss": 4.617072582244873,
+ "step": 4032
+ },
+ {
+ "epoch": 56.0,
+ "eval_loss": 4.811229705810547,
+ "eval_runtime": 56.4846,
+ "eval_samples_per_second": 43.233,
+ "eval_steps_per_second": 1.363,
+ "step": 4032
+ },
+ {
+ "epoch": 56.01397379912664,
+ "grad_norm": 0.02199401892721653,
+ "learning_rate": 0.0006,
+ "loss": 4.513965606689453,
+ "step": 4033
+ },
+ {
+ "epoch": 56.02794759825328,
+ "grad_norm": 0.022109858691692352,
+ "learning_rate": 0.0006,
+ "loss": 4.557568550109863,
+ "step": 4034
+ },
+ {
+ "epoch": 56.041921397379916,
+ "grad_norm": 0.020103694871068,
+ "learning_rate": 0.0006,
+ "loss": 4.569512844085693,
+ "step": 4035
+ },
+ {
+ "epoch": 56.05589519650655,
+ "grad_norm": 0.01898941583931446,
+ "learning_rate": 0.0006,
+ "loss": 4.623138904571533,
+ "step": 4036
+ },
+ {
+ "epoch": 56.069868995633186,
+ "grad_norm": 0.02227010577917099,
+ "learning_rate": 0.0006,
+ "loss": 4.433316230773926,
+ "step": 4037
+ },
+ {
+ "epoch": 56.083842794759825,
+ "grad_norm": 0.024683045223355293,
+ "learning_rate": 0.0006,
+ "loss": 4.458296775817871,
+ "step": 4038
+ },
+ {
+ "epoch": 56.09781659388646,
+ "grad_norm": 0.02900184690952301,
+ "learning_rate": 0.0006,
+ "loss": 4.531060218811035,
+ "step": 4039
+ },
+ {
+ "epoch": 56.1117903930131,
+ "grad_norm": 0.025557566434144974,
+ "learning_rate": 0.0006,
+ "loss": 4.56685733795166,
+ "step": 4040
+ },
+ {
+ "epoch": 56.12576419213974,
+ "grad_norm": 0.021414656192064285,
+ "learning_rate": 0.0006,
+ "loss": 4.469037055969238,
+ "step": 4041
+ },
+ {
+ "epoch": 56.13973799126637,
+ "grad_norm": 0.023262616246938705,
+ "learning_rate": 0.0006,
+ "loss": 4.493876934051514,
+ "step": 4042
+ },
+ {
+ "epoch": 56.15371179039301,
+ "grad_norm": 0.023011326789855957,
+ "learning_rate": 0.0006,
+ "loss": 4.4572906494140625,
+ "step": 4043
+ },
+ {
+ "epoch": 56.16768558951965,
+ "grad_norm": 0.02177615463733673,
+ "learning_rate": 0.0006,
+ "loss": 4.615389823913574,
+ "step": 4044
+ },
+ {
+ "epoch": 56.18165938864629,
+ "grad_norm": 0.021527713164687157,
+ "learning_rate": 0.0006,
+ "loss": 4.613762855529785,
+ "step": 4045
+ },
+ {
+ "epoch": 56.19563318777293,
+ "grad_norm": 0.01982325315475464,
+ "learning_rate": 0.0006,
+ "loss": 4.53670597076416,
+ "step": 4046
+ },
+ {
+ "epoch": 56.209606986899566,
+ "grad_norm": 0.022484654560685158,
+ "learning_rate": 0.0006,
+ "loss": 4.605565071105957,
+ "step": 4047
+ },
+ {
+ "epoch": 56.223580786026204,
+ "grad_norm": 0.02346140146255493,
+ "learning_rate": 0.0006,
+ "loss": 4.6265950202941895,
+ "step": 4048
+ },
+ {
+ "epoch": 56.237554585152836,
+ "grad_norm": 0.01876714453101158,
+ "learning_rate": 0.0006,
+ "loss": 4.481525421142578,
+ "step": 4049
+ },
+ {
+ "epoch": 56.251528384279474,
+ "grad_norm": 0.01926851086318493,
+ "learning_rate": 0.0006,
+ "loss": 4.630258083343506,
+ "step": 4050
+ },
+ {
+ "epoch": 56.26550218340611,
+ "grad_norm": 0.020858589559793472,
+ "learning_rate": 0.0006,
+ "loss": 4.5265913009643555,
+ "step": 4051
+ },
+ {
+ "epoch": 56.27947598253275,
+ "grad_norm": 0.01889280416071415,
+ "learning_rate": 0.0006,
+ "loss": 4.581275463104248,
+ "step": 4052
+ },
+ {
+ "epoch": 56.29344978165939,
+ "grad_norm": 0.01801002025604248,
+ "learning_rate": 0.0006,
+ "loss": 4.498557090759277,
+ "step": 4053
+ },
+ {
+ "epoch": 56.30742358078603,
+ "grad_norm": 0.02252180315554142,
+ "learning_rate": 0.0006,
+ "loss": 4.565267086029053,
+ "step": 4054
+ },
+ {
+ "epoch": 56.32139737991266,
+ "grad_norm": 0.02306070365011692,
+ "learning_rate": 0.0006,
+ "loss": 4.3591389656066895,
+ "step": 4055
+ },
+ {
+ "epoch": 56.3353711790393,
+ "grad_norm": 0.020292513072490692,
+ "learning_rate": 0.0006,
+ "loss": 4.407517433166504,
+ "step": 4056
+ },
+ {
+ "epoch": 56.34934497816594,
+ "grad_norm": 0.019154123961925507,
+ "learning_rate": 0.0006,
+ "loss": 4.482261657714844,
+ "step": 4057
+ },
+ {
+ "epoch": 56.36331877729258,
+ "grad_norm": 0.017338937148451805,
+ "learning_rate": 0.0006,
+ "loss": 4.4463582038879395,
+ "step": 4058
+ },
+ {
+ "epoch": 56.377292576419215,
+ "grad_norm": 0.01730559952557087,
+ "learning_rate": 0.0006,
+ "loss": 4.542474746704102,
+ "step": 4059
+ },
+ {
+ "epoch": 56.391266375545854,
+ "grad_norm": 0.017029426991939545,
+ "learning_rate": 0.0006,
+ "loss": 4.552488327026367,
+ "step": 4060
+ },
+ {
+ "epoch": 56.40524017467249,
+ "grad_norm": 0.01745498552918434,
+ "learning_rate": 0.0006,
+ "loss": 4.52928352355957,
+ "step": 4061
+ },
+ {
+ "epoch": 56.419213973799124,
+ "grad_norm": 0.0168515145778656,
+ "learning_rate": 0.0006,
+ "loss": 4.4516801834106445,
+ "step": 4062
+ },
+ {
+ "epoch": 56.43318777292576,
+ "grad_norm": 0.018545417115092278,
+ "learning_rate": 0.0006,
+ "loss": 4.569811820983887,
+ "step": 4063
+ },
+ {
+ "epoch": 56.4471615720524,
+ "grad_norm": 0.02227838523685932,
+ "learning_rate": 0.0006,
+ "loss": 4.412224769592285,
+ "step": 4064
+ },
+ {
+ "epoch": 56.46113537117904,
+ "grad_norm": 0.023338504135608673,
+ "learning_rate": 0.0006,
+ "loss": 4.490184307098389,
+ "step": 4065
+ },
+ {
+ "epoch": 56.47510917030568,
+ "grad_norm": 0.02257535606622696,
+ "learning_rate": 0.0006,
+ "loss": 4.479152679443359,
+ "step": 4066
+ },
+ {
+ "epoch": 56.48908296943232,
+ "grad_norm": 0.022737598046660423,
+ "learning_rate": 0.0006,
+ "loss": 4.418235778808594,
+ "step": 4067
+ },
+ {
+ "epoch": 56.50305676855895,
+ "grad_norm": 0.024603573605418205,
+ "learning_rate": 0.0006,
+ "loss": 4.45781135559082,
+ "step": 4068
+ },
+ {
+ "epoch": 56.51703056768559,
+ "grad_norm": 0.024953778833150864,
+ "learning_rate": 0.0006,
+ "loss": 4.593907833099365,
+ "step": 4069
+ },
+ {
+ "epoch": 56.531004366812226,
+ "grad_norm": 0.02195645309984684,
+ "learning_rate": 0.0006,
+ "loss": 4.467643737792969,
+ "step": 4070
+ },
+ {
+ "epoch": 56.544978165938865,
+ "grad_norm": 0.017126578837633133,
+ "learning_rate": 0.0006,
+ "loss": 4.555484771728516,
+ "step": 4071
+ },
+ {
+ "epoch": 56.5589519650655,
+ "grad_norm": 0.019482605159282684,
+ "learning_rate": 0.0006,
+ "loss": 4.430915832519531,
+ "step": 4072
+ },
+ {
+ "epoch": 56.57292576419214,
+ "grad_norm": 0.0219392292201519,
+ "learning_rate": 0.0006,
+ "loss": 4.593365669250488,
+ "step": 4073
+ },
+ {
+ "epoch": 56.58689956331878,
+ "grad_norm": 0.02420145645737648,
+ "learning_rate": 0.0006,
+ "loss": 4.491467475891113,
+ "step": 4074
+ },
+ {
+ "epoch": 56.60087336244541,
+ "grad_norm": 0.021523723378777504,
+ "learning_rate": 0.0006,
+ "loss": 4.463634490966797,
+ "step": 4075
+ },
+ {
+ "epoch": 56.61484716157205,
+ "grad_norm": 0.020242048427462578,
+ "learning_rate": 0.0006,
+ "loss": 4.577968597412109,
+ "step": 4076
+ },
+ {
+ "epoch": 56.62882096069869,
+ "grad_norm": 0.017771173268556595,
+ "learning_rate": 0.0006,
+ "loss": 4.568269729614258,
+ "step": 4077
+ },
+ {
+ "epoch": 56.64279475982533,
+ "grad_norm": 0.01598544418811798,
+ "learning_rate": 0.0006,
+ "loss": 4.476841926574707,
+ "step": 4078
+ },
+ {
+ "epoch": 56.65676855895197,
+ "grad_norm": 0.0168446097522974,
+ "learning_rate": 0.0006,
+ "loss": 4.480965614318848,
+ "step": 4079
+ },
+ {
+ "epoch": 56.670742358078606,
+ "grad_norm": 0.014948660507798195,
+ "learning_rate": 0.0006,
+ "loss": 4.535400390625,
+ "step": 4080
+ },
+ {
+ "epoch": 56.68471615720524,
+ "grad_norm": 0.014899153262376785,
+ "learning_rate": 0.0006,
+ "loss": 4.476048946380615,
+ "step": 4081
+ },
+ {
+ "epoch": 56.698689956331876,
+ "grad_norm": 0.014051870442926884,
+ "learning_rate": 0.0006,
+ "loss": 4.533884048461914,
+ "step": 4082
+ },
+ {
+ "epoch": 56.712663755458514,
+ "grad_norm": 0.014708572067320347,
+ "learning_rate": 0.0006,
+ "loss": 4.586690902709961,
+ "step": 4083
+ },
+ {
+ "epoch": 56.72663755458515,
+ "grad_norm": 0.014797299169003963,
+ "learning_rate": 0.0006,
+ "loss": 4.446291446685791,
+ "step": 4084
+ },
+ {
+ "epoch": 56.74061135371179,
+ "grad_norm": 0.015256541781127453,
+ "learning_rate": 0.0006,
+ "loss": 4.553572654724121,
+ "step": 4085
+ },
+ {
+ "epoch": 56.75458515283843,
+ "grad_norm": 0.016042491421103477,
+ "learning_rate": 0.0006,
+ "loss": 4.507801055908203,
+ "step": 4086
+ },
+ {
+ "epoch": 56.76855895196506,
+ "grad_norm": 0.018835173919796944,
+ "learning_rate": 0.0006,
+ "loss": 4.393911838531494,
+ "step": 4087
+ },
+ {
+ "epoch": 56.7825327510917,
+ "grad_norm": 0.021994033828377724,
+ "learning_rate": 0.0006,
+ "loss": 4.543368339538574,
+ "step": 4088
+ },
+ {
+ "epoch": 56.79650655021834,
+ "grad_norm": 0.0230990182608366,
+ "learning_rate": 0.0006,
+ "loss": 4.507420539855957,
+ "step": 4089
+ },
+ {
+ "epoch": 56.81048034934498,
+ "grad_norm": 0.018818842247128487,
+ "learning_rate": 0.0006,
+ "loss": 4.499941349029541,
+ "step": 4090
+ },
+ {
+ "epoch": 56.82445414847162,
+ "grad_norm": 0.018257278949022293,
+ "learning_rate": 0.0006,
+ "loss": 4.380057334899902,
+ "step": 4091
+ },
+ {
+ "epoch": 56.838427947598255,
+ "grad_norm": 0.020073018968105316,
+ "learning_rate": 0.0006,
+ "loss": 4.572811603546143,
+ "step": 4092
+ },
+ {
+ "epoch": 56.852401746724894,
+ "grad_norm": 0.02078658528625965,
+ "learning_rate": 0.0006,
+ "loss": 4.446835517883301,
+ "step": 4093
+ },
+ {
+ "epoch": 56.866375545851525,
+ "grad_norm": 0.01793118566274643,
+ "learning_rate": 0.0006,
+ "loss": 4.517854690551758,
+ "step": 4094
+ },
+ {
+ "epoch": 56.880349344978164,
+ "grad_norm": 0.01608281210064888,
+ "learning_rate": 0.0006,
+ "loss": 4.506482124328613,
+ "step": 4095
+ },
+ {
+ "epoch": 56.8943231441048,
+ "grad_norm": 0.017405368387699127,
+ "learning_rate": 0.0006,
+ "loss": 4.439718246459961,
+ "step": 4096
+ },
+ {
+ "epoch": 56.90829694323144,
+ "grad_norm": 0.017930863425135612,
+ "learning_rate": 0.0006,
+ "loss": 4.409365653991699,
+ "step": 4097
+ },
+ {
+ "epoch": 56.92227074235808,
+ "grad_norm": 0.015757670626044273,
+ "learning_rate": 0.0006,
+ "loss": 4.505547523498535,
+ "step": 4098
+ },
+ {
+ "epoch": 56.93624454148472,
+ "grad_norm": 0.01580345816910267,
+ "learning_rate": 0.0006,
+ "loss": 4.462774753570557,
+ "step": 4099
+ },
+ {
+ "epoch": 56.95021834061135,
+ "grad_norm": 0.01659180410206318,
+ "learning_rate": 0.0006,
+ "loss": 4.49345588684082,
+ "step": 4100
+ },
+ {
+ "epoch": 56.96419213973799,
+ "grad_norm": 0.01877259835600853,
+ "learning_rate": 0.0006,
+ "loss": 4.574443340301514,
+ "step": 4101
+ },
+ {
+ "epoch": 56.97816593886463,
+ "grad_norm": 0.01977839693427086,
+ "learning_rate": 0.0006,
+ "loss": 4.574034690856934,
+ "step": 4102
+ },
+ {
+ "epoch": 56.992139737991266,
+ "grad_norm": 0.020627597346901894,
+ "learning_rate": 0.0006,
+ "loss": 4.50969123840332,
+ "step": 4103
+ },
+ {
+ "epoch": 57.0,
+ "grad_norm": 0.025128807872533798,
+ "learning_rate": 0.0006,
+ "loss": 4.401662826538086,
+ "step": 4104
+ },
+ {
+ "epoch": 57.0,
+ "eval_loss": 4.779252052307129,
+ "eval_runtime": 56.8104,
+ "eval_samples_per_second": 42.985,
+ "eval_steps_per_second": 1.355,
+ "step": 4104
+ },
+ {
+ "epoch": 57.01397379912664,
+ "grad_norm": 0.027712570503354073,
+ "learning_rate": 0.0006,
+ "loss": 4.472955703735352,
+ "step": 4105
+ },
+ {
+ "epoch": 57.02794759825328,
+ "grad_norm": 0.02164197526872158,
+ "learning_rate": 0.0006,
+ "loss": 4.387596130371094,
+ "step": 4106
+ },
+ {
+ "epoch": 57.041921397379916,
+ "grad_norm": 0.017972752451896667,
+ "learning_rate": 0.0006,
+ "loss": 4.4326276779174805,
+ "step": 4107
+ },
+ {
+ "epoch": 57.05589519650655,
+ "grad_norm": 0.020811019465327263,
+ "learning_rate": 0.0006,
+ "loss": 4.452951431274414,
+ "step": 4108
+ },
+ {
+ "epoch": 57.069868995633186,
+ "grad_norm": 0.019913259893655777,
+ "learning_rate": 0.0006,
+ "loss": 4.448301315307617,
+ "step": 4109
+ },
+ {
+ "epoch": 57.083842794759825,
+ "grad_norm": 0.017409102991223335,
+ "learning_rate": 0.0006,
+ "loss": 4.4108357429504395,
+ "step": 4110
+ },
+ {
+ "epoch": 57.09781659388646,
+ "grad_norm": 0.019374269992113113,
+ "learning_rate": 0.0006,
+ "loss": 4.398412227630615,
+ "step": 4111
+ },
+ {
+ "epoch": 57.1117903930131,
+ "grad_norm": 0.015521776862442493,
+ "learning_rate": 0.0006,
+ "loss": 4.508594989776611,
+ "step": 4112
+ },
+ {
+ "epoch": 57.12576419213974,
+ "grad_norm": 0.017201920971274376,
+ "learning_rate": 0.0006,
+ "loss": 4.449558734893799,
+ "step": 4113
+ },
+ {
+ "epoch": 57.13973799126637,
+ "grad_norm": 0.018062541261315346,
+ "learning_rate": 0.0006,
+ "loss": 4.517874240875244,
+ "step": 4114
+ },
+ {
+ "epoch": 57.15371179039301,
+ "grad_norm": 0.015846073627471924,
+ "learning_rate": 0.0006,
+ "loss": 4.427501678466797,
+ "step": 4115
+ },
+ {
+ "epoch": 57.16768558951965,
+ "grad_norm": 0.01849362626671791,
+ "learning_rate": 0.0006,
+ "loss": 4.433181285858154,
+ "step": 4116
+ },
+ {
+ "epoch": 57.18165938864629,
+ "grad_norm": 0.01948804222047329,
+ "learning_rate": 0.0006,
+ "loss": 4.446122646331787,
+ "step": 4117
+ },
+ {
+ "epoch": 57.19563318777293,
+ "grad_norm": 0.017354389652609825,
+ "learning_rate": 0.0006,
+ "loss": 4.431153774261475,
+ "step": 4118
+ },
+ {
+ "epoch": 57.209606986899566,
+ "grad_norm": 0.015817373991012573,
+ "learning_rate": 0.0006,
+ "loss": 4.448566913604736,
+ "step": 4119
+ },
+ {
+ "epoch": 57.223580786026204,
+ "grad_norm": 0.01635785959661007,
+ "learning_rate": 0.0006,
+ "loss": 4.408031463623047,
+ "step": 4120
+ },
+ {
+ "epoch": 57.237554585152836,
+ "grad_norm": 0.014528338797390461,
+ "learning_rate": 0.0006,
+ "loss": 4.541668891906738,
+ "step": 4121
+ },
+ {
+ "epoch": 57.251528384279474,
+ "grad_norm": 0.01589174196124077,
+ "learning_rate": 0.0006,
+ "loss": 4.513199329376221,
+ "step": 4122
+ },
+ {
+ "epoch": 57.26550218340611,
+ "grad_norm": 0.016093695536255836,
+ "learning_rate": 0.0006,
+ "loss": 4.516178131103516,
+ "step": 4123
+ },
+ {
+ "epoch": 57.27947598253275,
+ "grad_norm": 0.01788361556828022,
+ "learning_rate": 0.0006,
+ "loss": 4.524880409240723,
+ "step": 4124
+ },
+ {
+ "epoch": 57.29344978165939,
+ "grad_norm": 0.01591576263308525,
+ "learning_rate": 0.0006,
+ "loss": 4.44764518737793,
+ "step": 4125
+ },
+ {
+ "epoch": 57.30742358078603,
+ "grad_norm": 0.015452570281922817,
+ "learning_rate": 0.0006,
+ "loss": 4.46844482421875,
+ "step": 4126
+ },
+ {
+ "epoch": 57.32139737991266,
+ "grad_norm": 0.016404492780566216,
+ "learning_rate": 0.0006,
+ "loss": 4.443678855895996,
+ "step": 4127
+ },
+ {
+ "epoch": 57.3353711790393,
+ "grad_norm": 0.019967148080468178,
+ "learning_rate": 0.0006,
+ "loss": 4.430230140686035,
+ "step": 4128
+ },
+ {
+ "epoch": 57.34934497816594,
+ "grad_norm": 0.02594015561044216,
+ "learning_rate": 0.0006,
+ "loss": 4.47061824798584,
+ "step": 4129
+ },
+ {
+ "epoch": 57.36331877729258,
+ "grad_norm": 0.031096890568733215,
+ "learning_rate": 0.0006,
+ "loss": 4.462275981903076,
+ "step": 4130
+ },
+ {
+ "epoch": 57.377292576419215,
+ "grad_norm": 0.02834898792207241,
+ "learning_rate": 0.0006,
+ "loss": 4.41143798828125,
+ "step": 4131
+ },
+ {
+ "epoch": 57.391266375545854,
+ "grad_norm": 0.023621153086423874,
+ "learning_rate": 0.0006,
+ "loss": 4.562548637390137,
+ "step": 4132
+ },
+ {
+ "epoch": 57.40524017467249,
+ "grad_norm": 0.02277296595275402,
+ "learning_rate": 0.0006,
+ "loss": 4.456247329711914,
+ "step": 4133
+ },
+ {
+ "epoch": 57.419213973799124,
+ "grad_norm": 0.019868768751621246,
+ "learning_rate": 0.0006,
+ "loss": 4.524300575256348,
+ "step": 4134
+ },
+ {
+ "epoch": 57.43318777292576,
+ "grad_norm": 0.01887078955769539,
+ "learning_rate": 0.0006,
+ "loss": 4.540097236633301,
+ "step": 4135
+ },
+ {
+ "epoch": 57.4471615720524,
+ "grad_norm": 0.024268802255392075,
+ "learning_rate": 0.0006,
+ "loss": 4.397233963012695,
+ "step": 4136
+ },
+ {
+ "epoch": 57.46113537117904,
+ "grad_norm": 0.025739721953868866,
+ "learning_rate": 0.0006,
+ "loss": 4.354194641113281,
+ "step": 4137
+ },
+ {
+ "epoch": 57.47510917030568,
+ "grad_norm": 0.02442290261387825,
+ "learning_rate": 0.0006,
+ "loss": 4.303822994232178,
+ "step": 4138
+ },
+ {
+ "epoch": 57.48908296943232,
+ "grad_norm": 0.022215209901332855,
+ "learning_rate": 0.0006,
+ "loss": 4.4870452880859375,
+ "step": 4139
+ },
+ {
+ "epoch": 57.50305676855895,
+ "grad_norm": 0.01789158768951893,
+ "learning_rate": 0.0006,
+ "loss": 4.343809127807617,
+ "step": 4140
+ },
+ {
+ "epoch": 57.51703056768559,
+ "grad_norm": 0.020381739363074303,
+ "learning_rate": 0.0006,
+ "loss": 4.53312873840332,
+ "step": 4141
+ },
+ {
+ "epoch": 57.531004366812226,
+ "grad_norm": 0.019585080444812775,
+ "learning_rate": 0.0006,
+ "loss": 4.510162353515625,
+ "step": 4142
+ },
+ {
+ "epoch": 57.544978165938865,
+ "grad_norm": 0.01985127478837967,
+ "learning_rate": 0.0006,
+ "loss": 4.3654022216796875,
+ "step": 4143
+ },
+ {
+ "epoch": 57.5589519650655,
+ "grad_norm": 0.01962905190885067,
+ "learning_rate": 0.0006,
+ "loss": 4.560830116271973,
+ "step": 4144
+ },
+ {
+ "epoch": 57.57292576419214,
+ "grad_norm": 0.02679629437625408,
+ "learning_rate": 0.0006,
+ "loss": 4.300804138183594,
+ "step": 4145
+ },
+ {
+ "epoch": 57.58689956331878,
+ "grad_norm": 0.039830856025218964,
+ "learning_rate": 0.0006,
+ "loss": 4.54150915145874,
+ "step": 4146
+ },
+ {
+ "epoch": 57.60087336244541,
+ "grad_norm": 0.049527477473020554,
+ "learning_rate": 0.0006,
+ "loss": 4.405792236328125,
+ "step": 4147
+ },
+ {
+ "epoch": 57.61484716157205,
+ "grad_norm": 0.042966946959495544,
+ "learning_rate": 0.0006,
+ "loss": 4.500643730163574,
+ "step": 4148
+ },
+ {
+ "epoch": 57.62882096069869,
+ "grad_norm": 0.08999177813529968,
+ "learning_rate": 0.0006,
+ "loss": 4.504790782928467,
+ "step": 4149
+ },
+ {
+ "epoch": 57.64279475982533,
+ "grad_norm": 0.44693124294281006,
+ "learning_rate": 0.0006,
+ "loss": 4.845344066619873,
+ "step": 4150
+ },
+ {
+ "epoch": 57.65676855895197,
+ "grad_norm": 0.7143144011497498,
+ "learning_rate": 0.0006,
+ "loss": 6.907845497131348,
+ "step": 4151
+ },
+ {
+ "epoch": 57.670742358078606,
+ "grad_norm": 0.25239595770835876,
+ "learning_rate": 0.0006,
+ "loss": 7.082579612731934,
+ "step": 4152
+ },
+ {
+ "epoch": 57.68471615720524,
+ "grad_norm": 0.26550790667533875,
+ "learning_rate": 0.0006,
+ "loss": 7.50180196762085,
+ "step": 4153
+ },
+ {
+ "epoch": 57.698689956331876,
+ "grad_norm": 0.14735378324985504,
+ "learning_rate": 0.0006,
+ "loss": 7.214590072631836,
+ "step": 4154
+ },
+ {
+ "epoch": 57.712663755458514,
+ "grad_norm": 0.17685334384441376,
+ "learning_rate": 0.0006,
+ "loss": 6.861320495605469,
+ "step": 4155
+ },
+ {
+ "epoch": 57.72663755458515,
+ "grad_norm": 0.18231339752674103,
+ "learning_rate": 0.0006,
+ "loss": 6.791839599609375,
+ "step": 4156
+ },
+ {
+ "epoch": 57.74061135371179,
+ "grad_norm": 0.09230359643697739,
+ "learning_rate": 0.0006,
+ "loss": 6.760223388671875,
+ "step": 4157
+ },
+ {
+ "epoch": 57.75458515283843,
+ "grad_norm": 0.06918249279260635,
+ "learning_rate": 0.0006,
+ "loss": 6.532999515533447,
+ "step": 4158
+ },
+ {
+ "epoch": 57.76855895196506,
+ "grad_norm": 0.0901007279753685,
+ "learning_rate": 0.0006,
+ "loss": 6.52907133102417,
+ "step": 4159
+ },
+ {
+ "epoch": 57.7825327510917,
+ "grad_norm": 0.05678229779005051,
+ "learning_rate": 0.0006,
+ "loss": 6.295290470123291,
+ "step": 4160
+ },
+ {
+ "epoch": 57.79650655021834,
+ "grad_norm": 0.04450210556387901,
+ "learning_rate": 0.0006,
+ "loss": 6.14900541305542,
+ "step": 4161
+ },
+ {
+ "epoch": 57.81048034934498,
+ "grad_norm": 0.049191512167453766,
+ "learning_rate": 0.0006,
+ "loss": 6.19650411605835,
+ "step": 4162
+ },
+ {
+ "epoch": 57.82445414847162,
+ "grad_norm": 0.03908967226743698,
+ "learning_rate": 0.0006,
+ "loss": 6.054948806762695,
+ "step": 4163
+ },
+ {
+ "epoch": 57.838427947598255,
+ "grad_norm": 0.04023387283086777,
+ "learning_rate": 0.0006,
+ "loss": 6.034873962402344,
+ "step": 4164
+ },
+ {
+ "epoch": 57.852401746724894,
+ "grad_norm": 0.03304268419742584,
+ "learning_rate": 0.0006,
+ "loss": 5.769852161407471,
+ "step": 4165
+ },
+ {
+ "epoch": 57.866375545851525,
+ "grad_norm": 0.03432450816035271,
+ "learning_rate": 0.0006,
+ "loss": 5.788785457611084,
+ "step": 4166
+ },
+ {
+ "epoch": 57.880349344978164,
+ "grad_norm": 0.029965840280056,
+ "learning_rate": 0.0006,
+ "loss": 5.815243721008301,
+ "step": 4167
+ },
+ {
+ "epoch": 57.8943231441048,
+ "grad_norm": 0.03637698292732239,
+ "learning_rate": 0.0006,
+ "loss": 5.7204060554504395,
+ "step": 4168
+ },
+ {
+ "epoch": 57.90829694323144,
+ "grad_norm": 0.03477517142891884,
+ "learning_rate": 0.0006,
+ "loss": 5.605930328369141,
+ "step": 4169
+ },
+ {
+ "epoch": 57.92227074235808,
+ "grad_norm": 0.05061859264969826,
+ "learning_rate": 0.0006,
+ "loss": 5.547054767608643,
+ "step": 4170
+ },
+ {
+ "epoch": 57.93624454148472,
+ "grad_norm": 0.06299655884504318,
+ "learning_rate": 0.0006,
+ "loss": 5.552166938781738,
+ "step": 4171
+ },
+ {
+ "epoch": 57.95021834061135,
+ "grad_norm": 0.04215948283672333,
+ "learning_rate": 0.0006,
+ "loss": 5.375457763671875,
+ "step": 4172
+ },
+ {
+ "epoch": 57.96419213973799,
+ "grad_norm": 0.0348566472530365,
+ "learning_rate": 0.0006,
+ "loss": 5.418606281280518,
+ "step": 4173
+ },
+ {
+ "epoch": 57.97816593886463,
+ "grad_norm": 0.03807530924677849,
+ "learning_rate": 0.0006,
+ "loss": 5.403156280517578,
+ "step": 4174
+ },
+ {
+ "epoch": 57.992139737991266,
+ "grad_norm": 0.02989993803203106,
+ "learning_rate": 0.0006,
+ "loss": 5.355951309204102,
+ "step": 4175
+ },
+ {
+ "epoch": 58.0,
+ "grad_norm": 0.034300558269023895,
+ "learning_rate": 0.0006,
+ "loss": 5.44920539855957,
+ "step": 4176
+ },
+ {
+ "epoch": 58.0,
+ "eval_loss": 5.416731834411621,
+ "eval_runtime": 56.7934,
+ "eval_samples_per_second": 42.998,
+ "eval_steps_per_second": 1.356,
+ "step": 4176
+ },
+ {
+ "epoch": 58.01397379912664,
+ "grad_norm": 0.033961448818445206,
+ "learning_rate": 0.0006,
+ "loss": 5.250707626342773,
+ "step": 4177
+ },
+ {
+ "epoch": 58.02794759825328,
+ "grad_norm": 0.027343502268195152,
+ "learning_rate": 0.0006,
+ "loss": 5.149238109588623,
+ "step": 4178
+ },
+ {
+ "epoch": 58.041921397379916,
+ "grad_norm": 0.03033650480210781,
+ "learning_rate": 0.0006,
+ "loss": 5.072463512420654,
+ "step": 4179
+ },
+ {
+ "epoch": 58.05589519650655,
+ "grad_norm": 0.028262868523597717,
+ "learning_rate": 0.0006,
+ "loss": 5.109417915344238,
+ "step": 4180
+ },
+ {
+ "epoch": 58.069868995633186,
+ "grad_norm": 0.02373034693300724,
+ "learning_rate": 0.0006,
+ "loss": 5.082464218139648,
+ "step": 4181
+ },
+ {
+ "epoch": 58.083842794759825,
+ "grad_norm": 0.025879688560962677,
+ "learning_rate": 0.0006,
+ "loss": 5.056792259216309,
+ "step": 4182
+ },
+ {
+ "epoch": 58.09781659388646,
+ "grad_norm": 0.026252275332808495,
+ "learning_rate": 0.0006,
+ "loss": 5.064239978790283,
+ "step": 4183
+ },
+ {
+ "epoch": 58.1117903930131,
+ "grad_norm": 0.028177792206406593,
+ "learning_rate": 0.0006,
+ "loss": 4.922310829162598,
+ "step": 4184
+ },
+ {
+ "epoch": 58.12576419213974,
+ "grad_norm": 0.02630820870399475,
+ "learning_rate": 0.0006,
+ "loss": 4.992884635925293,
+ "step": 4185
+ },
+ {
+ "epoch": 58.13973799126637,
+ "grad_norm": 0.024890903383493423,
+ "learning_rate": 0.0006,
+ "loss": 4.910036563873291,
+ "step": 4186
+ },
+ {
+ "epoch": 58.15371179039301,
+ "grad_norm": 0.02822992391884327,
+ "learning_rate": 0.0006,
+ "loss": 4.866640090942383,
+ "step": 4187
+ },
+ {
+ "epoch": 58.16768558951965,
+ "grad_norm": 0.03253607079386711,
+ "learning_rate": 0.0006,
+ "loss": 4.9112958908081055,
+ "step": 4188
+ },
+ {
+ "epoch": 58.18165938864629,
+ "grad_norm": 0.03765944764018059,
+ "learning_rate": 0.0006,
+ "loss": 4.774468421936035,
+ "step": 4189
+ },
+ {
+ "epoch": 58.19563318777293,
+ "grad_norm": 0.039874881505966187,
+ "learning_rate": 0.0006,
+ "loss": 4.831678867340088,
+ "step": 4190
+ },
+ {
+ "epoch": 58.209606986899566,
+ "grad_norm": 0.034595925360918045,
+ "learning_rate": 0.0006,
+ "loss": 4.8243794441223145,
+ "step": 4191
+ },
+ {
+ "epoch": 58.223580786026204,
+ "grad_norm": 0.04840189591050148,
+ "learning_rate": 0.0006,
+ "loss": 4.854156494140625,
+ "step": 4192
+ },
+ {
+ "epoch": 58.237554585152836,
+ "grad_norm": 0.06358753889799118,
+ "learning_rate": 0.0006,
+ "loss": 4.836430549621582,
+ "step": 4193
+ },
+ {
+ "epoch": 58.251528384279474,
+ "grad_norm": 0.0507555715739727,
+ "learning_rate": 0.0006,
+ "loss": 4.88357400894165,
+ "step": 4194
+ },
+ {
+ "epoch": 58.26550218340611,
+ "grad_norm": 0.04384690150618553,
+ "learning_rate": 0.0006,
+ "loss": 4.774049282073975,
+ "step": 4195
+ },
+ {
+ "epoch": 58.27947598253275,
+ "grad_norm": 0.03246928006410599,
+ "learning_rate": 0.0006,
+ "loss": 4.686605930328369,
+ "step": 4196
+ },
+ {
+ "epoch": 58.29344978165939,
+ "grad_norm": 0.028878789395093918,
+ "learning_rate": 0.0006,
+ "loss": 4.796070575714111,
+ "step": 4197
+ },
+ {
+ "epoch": 58.30742358078603,
+ "grad_norm": 0.028175361454486847,
+ "learning_rate": 0.0006,
+ "loss": 4.728490829467773,
+ "step": 4198
+ },
+ {
+ "epoch": 58.32139737991266,
+ "grad_norm": 0.029428014531731606,
+ "learning_rate": 0.0006,
+ "loss": 4.7159624099731445,
+ "step": 4199
+ },
+ {
+ "epoch": 58.3353711790393,
+ "grad_norm": 0.031227827072143555,
+ "learning_rate": 0.0006,
+ "loss": 4.702625274658203,
+ "step": 4200
+ },
+ {
+ "epoch": 58.34934497816594,
+ "grad_norm": 0.028843428939580917,
+ "learning_rate": 0.0006,
+ "loss": 4.811799049377441,
+ "step": 4201
+ },
+ {
+ "epoch": 58.36331877729258,
+ "grad_norm": 0.026780391111969948,
+ "learning_rate": 0.0006,
+ "loss": 4.81662654876709,
+ "step": 4202
+ },
+ {
+ "epoch": 58.377292576419215,
+ "grad_norm": 0.019699757918715477,
+ "learning_rate": 0.0006,
+ "loss": 4.713685989379883,
+ "step": 4203
+ },
+ {
+ "epoch": 58.391266375545854,
+ "grad_norm": 0.024389250203967094,
+ "learning_rate": 0.0006,
+ "loss": 4.72354793548584,
+ "step": 4204
+ },
+ {
+ "epoch": 58.40524017467249,
+ "grad_norm": 0.02218274027109146,
+ "learning_rate": 0.0006,
+ "loss": 4.674125671386719,
+ "step": 4205
+ },
+ {
+ "epoch": 58.419213973799124,
+ "grad_norm": 0.02041488140821457,
+ "learning_rate": 0.0006,
+ "loss": 4.7909698486328125,
+ "step": 4206
+ },
+ {
+ "epoch": 58.43318777292576,
+ "grad_norm": 0.018561935052275658,
+ "learning_rate": 0.0006,
+ "loss": 4.808770179748535,
+ "step": 4207
+ },
+ {
+ "epoch": 58.4471615720524,
+ "grad_norm": 0.01708846725523472,
+ "learning_rate": 0.0006,
+ "loss": 4.701626777648926,
+ "step": 4208
+ },
+ {
+ "epoch": 58.46113537117904,
+ "grad_norm": 0.0222734697163105,
+ "learning_rate": 0.0006,
+ "loss": 4.505410194396973,
+ "step": 4209
+ },
+ {
+ "epoch": 58.47510917030568,
+ "grad_norm": 0.03197057917714119,
+ "learning_rate": 0.0006,
+ "loss": 4.610360145568848,
+ "step": 4210
+ },
+ {
+ "epoch": 58.48908296943232,
+ "grad_norm": 0.06854520738124847,
+ "learning_rate": 0.0006,
+ "loss": 4.625649452209473,
+ "step": 4211
+ },
+ {
+ "epoch": 58.50305676855895,
+ "grad_norm": 0.09981521219015121,
+ "learning_rate": 0.0006,
+ "loss": 4.832150459289551,
+ "step": 4212
+ },
+ {
+ "epoch": 58.51703056768559,
+ "grad_norm": 0.046356040984392166,
+ "learning_rate": 0.0006,
+ "loss": 4.5641632080078125,
+ "step": 4213
+ },
+ {
+ "epoch": 58.531004366812226,
+ "grad_norm": 0.03882667422294617,
+ "learning_rate": 0.0006,
+ "loss": 4.666974067687988,
+ "step": 4214
+ },
+ {
+ "epoch": 58.544978165938865,
+ "grad_norm": 0.033203233033418655,
+ "learning_rate": 0.0006,
+ "loss": 4.629413604736328,
+ "step": 4215
+ },
+ {
+ "epoch": 58.5589519650655,
+ "grad_norm": 0.02806735597550869,
+ "learning_rate": 0.0006,
+ "loss": 4.763245582580566,
+ "step": 4216
+ },
+ {
+ "epoch": 58.57292576419214,
+ "grad_norm": 0.026728278025984764,
+ "learning_rate": 0.0006,
+ "loss": 4.629597187042236,
+ "step": 4217
+ },
+ {
+ "epoch": 58.58689956331878,
+ "grad_norm": 0.024380959570407867,
+ "learning_rate": 0.0006,
+ "loss": 4.633614540100098,
+ "step": 4218
+ },
+ {
+ "epoch": 58.60087336244541,
+ "grad_norm": 0.022312544286251068,
+ "learning_rate": 0.0006,
+ "loss": 4.671379566192627,
+ "step": 4219
+ },
+ {
+ "epoch": 58.61484716157205,
+ "grad_norm": 0.022049568593502045,
+ "learning_rate": 0.0006,
+ "loss": 4.661557197570801,
+ "step": 4220
+ },
+ {
+ "epoch": 58.62882096069869,
+ "grad_norm": 0.02119818702340126,
+ "learning_rate": 0.0006,
+ "loss": 4.688562393188477,
+ "step": 4221
+ },
+ {
+ "epoch": 58.64279475982533,
+ "grad_norm": 0.019769301638007164,
+ "learning_rate": 0.0006,
+ "loss": 4.607748031616211,
+ "step": 4222
+ },
+ {
+ "epoch": 58.65676855895197,
+ "grad_norm": 0.02124079130589962,
+ "learning_rate": 0.0006,
+ "loss": 4.570640563964844,
+ "step": 4223
+ },
+ {
+ "epoch": 58.670742358078606,
+ "grad_norm": 0.016344040632247925,
+ "learning_rate": 0.0006,
+ "loss": 4.676113128662109,
+ "step": 4224
+ },
+ {
+ "epoch": 58.68471615720524,
+ "grad_norm": 0.01735256239771843,
+ "learning_rate": 0.0006,
+ "loss": 4.62088680267334,
+ "step": 4225
+ },
+ {
+ "epoch": 58.698689956331876,
+ "grad_norm": 0.019323458895087242,
+ "learning_rate": 0.0006,
+ "loss": 4.579405307769775,
+ "step": 4226
+ },
+ {
+ "epoch": 58.712663755458514,
+ "grad_norm": 0.01760544814169407,
+ "learning_rate": 0.0006,
+ "loss": 4.625407695770264,
+ "step": 4227
+ },
+ {
+ "epoch": 58.72663755458515,
+ "grad_norm": 0.015593250282108784,
+ "learning_rate": 0.0006,
+ "loss": 4.6326117515563965,
+ "step": 4228
+ },
+ {
+ "epoch": 58.74061135371179,
+ "grad_norm": 0.014357523061335087,
+ "learning_rate": 0.0006,
+ "loss": 4.6869401931762695,
+ "step": 4229
+ },
+ {
+ "epoch": 58.75458515283843,
+ "grad_norm": 0.01604357920587063,
+ "learning_rate": 0.0006,
+ "loss": 4.690784931182861,
+ "step": 4230
+ },
+ {
+ "epoch": 58.76855895196506,
+ "grad_norm": 0.01673845760524273,
+ "learning_rate": 0.0006,
+ "loss": 4.513410568237305,
+ "step": 4231
+ },
+ {
+ "epoch": 58.7825327510917,
+ "grad_norm": 0.013657779432833195,
+ "learning_rate": 0.0006,
+ "loss": 4.637057781219482,
+ "step": 4232
+ },
+ {
+ "epoch": 58.79650655021834,
+ "grad_norm": 0.013373114168643951,
+ "learning_rate": 0.0006,
+ "loss": 4.719967842102051,
+ "step": 4233
+ },
+ {
+ "epoch": 58.81048034934498,
+ "grad_norm": 0.016345568001270294,
+ "learning_rate": 0.0006,
+ "loss": 4.604109764099121,
+ "step": 4234
+ },
+ {
+ "epoch": 58.82445414847162,
+ "grad_norm": 0.01573033444583416,
+ "learning_rate": 0.0006,
+ "loss": 4.6471428871154785,
+ "step": 4235
+ },
+ {
+ "epoch": 58.838427947598255,
+ "grad_norm": 0.01517449039965868,
+ "learning_rate": 0.0006,
+ "loss": 4.722684860229492,
+ "step": 4236
+ },
+ {
+ "epoch": 58.852401746724894,
+ "grad_norm": 0.0145102022215724,
+ "learning_rate": 0.0006,
+ "loss": 4.662420272827148,
+ "step": 4237
+ },
+ {
+ "epoch": 58.866375545851525,
+ "grad_norm": 0.01457217987626791,
+ "learning_rate": 0.0006,
+ "loss": 4.585169792175293,
+ "step": 4238
+ },
+ {
+ "epoch": 58.880349344978164,
+ "grad_norm": 0.01406773366034031,
+ "learning_rate": 0.0006,
+ "loss": 4.5413665771484375,
+ "step": 4239
+ },
+ {
+ "epoch": 58.8943231441048,
+ "grad_norm": 0.012653871439397335,
+ "learning_rate": 0.0006,
+ "loss": 4.553636074066162,
+ "step": 4240
+ },
+ {
+ "epoch": 58.90829694323144,
+ "grad_norm": 0.01261200476437807,
+ "learning_rate": 0.0006,
+ "loss": 4.534193992614746,
+ "step": 4241
+ },
+ {
+ "epoch": 58.92227074235808,
+ "grad_norm": 0.012955864891409874,
+ "learning_rate": 0.0006,
+ "loss": 4.697248458862305,
+ "step": 4242
+ },
+ {
+ "epoch": 58.93624454148472,
+ "grad_norm": 0.012134749442338943,
+ "learning_rate": 0.0006,
+ "loss": 4.557095527648926,
+ "step": 4243
+ },
+ {
+ "epoch": 58.95021834061135,
+ "grad_norm": 0.012164677493274212,
+ "learning_rate": 0.0006,
+ "loss": 4.623579025268555,
+ "step": 4244
+ },
+ {
+ "epoch": 58.96419213973799,
+ "grad_norm": 0.013802947476506233,
+ "learning_rate": 0.0006,
+ "loss": 4.562838077545166,
+ "step": 4245
+ },
+ {
+ "epoch": 58.97816593886463,
+ "grad_norm": 0.011420476250350475,
+ "learning_rate": 0.0006,
+ "loss": 4.62265682220459,
+ "step": 4246
+ },
+ {
+ "epoch": 58.992139737991266,
+ "grad_norm": 0.01169576682150364,
+ "learning_rate": 0.0006,
+ "loss": 4.727449417114258,
+ "step": 4247
+ },
+ {
+ "epoch": 59.0,
+ "grad_norm": 0.013577710837125778,
+ "learning_rate": 0.0006,
+ "loss": 4.536325454711914,
+ "step": 4248
+ },
+ {
+ "epoch": 59.0,
+ "eval_loss": 4.86527681350708,
+ "eval_runtime": 60.4463,
+ "eval_samples_per_second": 40.399,
+ "eval_steps_per_second": 1.274,
+ "step": 4248
+ },
+ {
+ "epoch": 59.01397379912664,
+ "grad_norm": 0.015251655131578445,
+ "learning_rate": 0.0006,
+ "loss": 4.634598731994629,
+ "step": 4249
+ },
+ {
+ "epoch": 59.02794759825328,
+ "grad_norm": 0.017414938658475876,
+ "learning_rate": 0.0006,
+ "loss": 4.521329879760742,
+ "step": 4250
+ },
+ {
+ "epoch": 59.041921397379916,
+ "grad_norm": 0.01656588353216648,
+ "learning_rate": 0.0006,
+ "loss": 4.446192264556885,
+ "step": 4251
+ },
+ {
+ "epoch": 59.05589519650655,
+ "grad_norm": 0.012038925662636757,
+ "learning_rate": 0.0006,
+ "loss": 4.523855209350586,
+ "step": 4252
+ },
+ {
+ "epoch": 59.069868995633186,
+ "grad_norm": 0.014403033070266247,
+ "learning_rate": 0.0006,
+ "loss": 4.561079978942871,
+ "step": 4253
+ },
+ {
+ "epoch": 59.083842794759825,
+ "grad_norm": 0.016243597492575645,
+ "learning_rate": 0.0006,
+ "loss": 4.673602104187012,
+ "step": 4254
+ },
+ {
+ "epoch": 59.09781659388646,
+ "grad_norm": 0.016492877155542374,
+ "learning_rate": 0.0006,
+ "loss": 4.564824104309082,
+ "step": 4255
+ },
+ {
+ "epoch": 59.1117903930131,
+ "grad_norm": 0.015980370342731476,
+ "learning_rate": 0.0006,
+ "loss": 4.339249134063721,
+ "step": 4256
+ },
+ {
+ "epoch": 59.12576419213974,
+ "grad_norm": 0.019687356427311897,
+ "learning_rate": 0.0006,
+ "loss": 4.481834888458252,
+ "step": 4257
+ },
+ {
+ "epoch": 59.13973799126637,
+ "grad_norm": 0.02983798086643219,
+ "learning_rate": 0.0006,
+ "loss": 4.4495697021484375,
+ "step": 4258
+ },
+ {
+ "epoch": 59.15371179039301,
+ "grad_norm": 0.04600981995463371,
+ "learning_rate": 0.0006,
+ "loss": 4.567838668823242,
+ "step": 4259
+ },
+ {
+ "epoch": 59.16768558951965,
+ "grad_norm": 0.04852001741528511,
+ "learning_rate": 0.0006,
+ "loss": 4.594701766967773,
+ "step": 4260
+ },
+ {
+ "epoch": 59.18165938864629,
+ "grad_norm": 0.03429504856467247,
+ "learning_rate": 0.0006,
+ "loss": 4.573735237121582,
+ "step": 4261
+ },
+ {
+ "epoch": 59.19563318777293,
+ "grad_norm": 0.029345160350203514,
+ "learning_rate": 0.0006,
+ "loss": 4.488258361816406,
+ "step": 4262
+ },
+ {
+ "epoch": 59.209606986899566,
+ "grad_norm": 0.01961623504757881,
+ "learning_rate": 0.0006,
+ "loss": 4.545222282409668,
+ "step": 4263
+ },
+ {
+ "epoch": 59.223580786026204,
+ "grad_norm": 0.020406130701303482,
+ "learning_rate": 0.0006,
+ "loss": 4.471898078918457,
+ "step": 4264
+ },
+ {
+ "epoch": 59.237554585152836,
+ "grad_norm": 0.020500272512435913,
+ "learning_rate": 0.0006,
+ "loss": 4.538337707519531,
+ "step": 4265
+ },
+ {
+ "epoch": 59.251528384279474,
+ "grad_norm": 0.020727017894387245,
+ "learning_rate": 0.0006,
+ "loss": 4.437284469604492,
+ "step": 4266
+ },
+ {
+ "epoch": 59.26550218340611,
+ "grad_norm": 0.03395998477935791,
+ "learning_rate": 0.0006,
+ "loss": 4.459124565124512,
+ "step": 4267
+ },
+ {
+ "epoch": 59.27947598253275,
+ "grad_norm": 0.04635605961084366,
+ "learning_rate": 0.0006,
+ "loss": 4.520228385925293,
+ "step": 4268
+ },
+ {
+ "epoch": 59.29344978165939,
+ "grad_norm": 0.04253006726503372,
+ "learning_rate": 0.0006,
+ "loss": 4.578673362731934,
+ "step": 4269
+ },
+ {
+ "epoch": 59.30742358078603,
+ "grad_norm": 0.02403509058058262,
+ "learning_rate": 0.0006,
+ "loss": 4.484639644622803,
+ "step": 4270
+ },
+ {
+ "epoch": 59.32139737991266,
+ "grad_norm": 0.02254190482199192,
+ "learning_rate": 0.0006,
+ "loss": 4.511314868927002,
+ "step": 4271
+ },
+ {
+ "epoch": 59.3353711790393,
+ "grad_norm": 0.021585190668702126,
+ "learning_rate": 0.0006,
+ "loss": 4.530239105224609,
+ "step": 4272
+ },
+ {
+ "epoch": 59.34934497816594,
+ "grad_norm": 0.021210316568613052,
+ "learning_rate": 0.0006,
+ "loss": 4.593514442443848,
+ "step": 4273
+ },
+ {
+ "epoch": 59.36331877729258,
+ "grad_norm": 0.019317107275128365,
+ "learning_rate": 0.0006,
+ "loss": 4.5591535568237305,
+ "step": 4274
+ },
+ {
+ "epoch": 59.377292576419215,
+ "grad_norm": 0.018059708178043365,
+ "learning_rate": 0.0006,
+ "loss": 4.4499592781066895,
+ "step": 4275
+ },
+ {
+ "epoch": 59.391266375545854,
+ "grad_norm": 0.017847446724772453,
+ "learning_rate": 0.0006,
+ "loss": 4.473999500274658,
+ "step": 4276
+ },
+ {
+ "epoch": 59.40524017467249,
+ "grad_norm": 0.014944756403565407,
+ "learning_rate": 0.0006,
+ "loss": 4.5934529304504395,
+ "step": 4277
+ },
+ {
+ "epoch": 59.419213973799124,
+ "grad_norm": 0.01663832925260067,
+ "learning_rate": 0.0006,
+ "loss": 4.504262924194336,
+ "step": 4278
+ },
+ {
+ "epoch": 59.43318777292576,
+ "grad_norm": 0.015571796335279942,
+ "learning_rate": 0.0006,
+ "loss": 4.512310981750488,
+ "step": 4279
+ },
+ {
+ "epoch": 59.4471615720524,
+ "grad_norm": 0.015896400436758995,
+ "learning_rate": 0.0006,
+ "loss": 4.452714920043945,
+ "step": 4280
+ },
+ {
+ "epoch": 59.46113537117904,
+ "grad_norm": 0.016010284423828125,
+ "learning_rate": 0.0006,
+ "loss": 4.6188063621521,
+ "step": 4281
+ },
+ {
+ "epoch": 59.47510917030568,
+ "grad_norm": 0.015075593255460262,
+ "learning_rate": 0.0006,
+ "loss": 4.477978229522705,
+ "step": 4282
+ },
+ {
+ "epoch": 59.48908296943232,
+ "grad_norm": 0.014698950573801994,
+ "learning_rate": 0.0006,
+ "loss": 4.544955730438232,
+ "step": 4283
+ },
+ {
+ "epoch": 59.50305676855895,
+ "grad_norm": 0.0136245833709836,
+ "learning_rate": 0.0006,
+ "loss": 4.530460357666016,
+ "step": 4284
+ },
+ {
+ "epoch": 59.51703056768559,
+ "grad_norm": 0.013118310831487179,
+ "learning_rate": 0.0006,
+ "loss": 4.614217758178711,
+ "step": 4285
+ },
+ {
+ "epoch": 59.531004366812226,
+ "grad_norm": 0.01295961532741785,
+ "learning_rate": 0.0006,
+ "loss": 4.43156099319458,
+ "step": 4286
+ },
+ {
+ "epoch": 59.544978165938865,
+ "grad_norm": 0.012429811991751194,
+ "learning_rate": 0.0006,
+ "loss": 4.517333030700684,
+ "step": 4287
+ },
+ {
+ "epoch": 59.5589519650655,
+ "grad_norm": 0.01261830423027277,
+ "learning_rate": 0.0006,
+ "loss": 4.569780349731445,
+ "step": 4288
+ },
+ {
+ "epoch": 59.57292576419214,
+ "grad_norm": 0.012394499965012074,
+ "learning_rate": 0.0006,
+ "loss": 4.5970869064331055,
+ "step": 4289
+ },
+ {
+ "epoch": 59.58689956331878,
+ "grad_norm": 0.01290759164839983,
+ "learning_rate": 0.0006,
+ "loss": 4.437100410461426,
+ "step": 4290
+ },
+ {
+ "epoch": 59.60087336244541,
+ "grad_norm": 0.012042169459164143,
+ "learning_rate": 0.0006,
+ "loss": 4.567404747009277,
+ "step": 4291
+ },
+ {
+ "epoch": 59.61484716157205,
+ "grad_norm": 0.012338520959019661,
+ "learning_rate": 0.0006,
+ "loss": 4.476193428039551,
+ "step": 4292
+ },
+ {
+ "epoch": 59.62882096069869,
+ "grad_norm": 0.01321258582174778,
+ "learning_rate": 0.0006,
+ "loss": 4.566680431365967,
+ "step": 4293
+ },
+ {
+ "epoch": 59.64279475982533,
+ "grad_norm": 0.013478122651576996,
+ "learning_rate": 0.0006,
+ "loss": 4.487398147583008,
+ "step": 4294
+ },
+ {
+ "epoch": 59.65676855895197,
+ "grad_norm": 0.01537603884935379,
+ "learning_rate": 0.0006,
+ "loss": 4.449337005615234,
+ "step": 4295
+ },
+ {
+ "epoch": 59.670742358078606,
+ "grad_norm": 0.017140565440058708,
+ "learning_rate": 0.0006,
+ "loss": 4.5174360275268555,
+ "step": 4296
+ },
+ {
+ "epoch": 59.68471615720524,
+ "grad_norm": 0.017931461334228516,
+ "learning_rate": 0.0006,
+ "loss": 4.522104740142822,
+ "step": 4297
+ },
+ {
+ "epoch": 59.698689956331876,
+ "grad_norm": 0.018037918955087662,
+ "learning_rate": 0.0006,
+ "loss": 4.496928691864014,
+ "step": 4298
+ },
+ {
+ "epoch": 59.712663755458514,
+ "grad_norm": 0.01684504561126232,
+ "learning_rate": 0.0006,
+ "loss": 4.5332417488098145,
+ "step": 4299
+ },
+ {
+ "epoch": 59.72663755458515,
+ "grad_norm": 0.016268383711576462,
+ "learning_rate": 0.0006,
+ "loss": 4.468893051147461,
+ "step": 4300
+ },
+ {
+ "epoch": 59.74061135371179,
+ "grad_norm": 0.014859385788440704,
+ "learning_rate": 0.0006,
+ "loss": 4.551092147827148,
+ "step": 4301
+ },
+ {
+ "epoch": 59.75458515283843,
+ "grad_norm": 0.013589047826826572,
+ "learning_rate": 0.0006,
+ "loss": 4.470579147338867,
+ "step": 4302
+ },
+ {
+ "epoch": 59.76855895196506,
+ "grad_norm": 0.015704551711678505,
+ "learning_rate": 0.0006,
+ "loss": 4.330068111419678,
+ "step": 4303
+ },
+ {
+ "epoch": 59.7825327510917,
+ "grad_norm": 0.017869921401143074,
+ "learning_rate": 0.0006,
+ "loss": 4.428255081176758,
+ "step": 4304
+ },
+ {
+ "epoch": 59.79650655021834,
+ "grad_norm": 0.019828537479043007,
+ "learning_rate": 0.0006,
+ "loss": 4.496023178100586,
+ "step": 4305
+ },
+ {
+ "epoch": 59.81048034934498,
+ "grad_norm": 0.015856236219406128,
+ "learning_rate": 0.0006,
+ "loss": 4.427152156829834,
+ "step": 4306
+ },
+ {
+ "epoch": 59.82445414847162,
+ "grad_norm": 0.013349304907023907,
+ "learning_rate": 0.0006,
+ "loss": 4.417333602905273,
+ "step": 4307
+ },
+ {
+ "epoch": 59.838427947598255,
+ "grad_norm": 0.014111637137830257,
+ "learning_rate": 0.0006,
+ "loss": 4.4507575035095215,
+ "step": 4308
+ },
+ {
+ "epoch": 59.852401746724894,
+ "grad_norm": 0.015186597593128681,
+ "learning_rate": 0.0006,
+ "loss": 4.375116348266602,
+ "step": 4309
+ },
+ {
+ "epoch": 59.866375545851525,
+ "grad_norm": 0.01604730449616909,
+ "learning_rate": 0.0006,
+ "loss": 4.604518890380859,
+ "step": 4310
+ },
+ {
+ "epoch": 59.880349344978164,
+ "grad_norm": 0.015308464877307415,
+ "learning_rate": 0.0006,
+ "loss": 4.507232189178467,
+ "step": 4311
+ },
+ {
+ "epoch": 59.8943231441048,
+ "grad_norm": 0.014017206616699696,
+ "learning_rate": 0.0006,
+ "loss": 4.398487091064453,
+ "step": 4312
+ },
+ {
+ "epoch": 59.90829694323144,
+ "grad_norm": 0.012911394238471985,
+ "learning_rate": 0.0006,
+ "loss": 4.567872047424316,
+ "step": 4313
+ },
+ {
+ "epoch": 59.92227074235808,
+ "grad_norm": 0.012775142677128315,
+ "learning_rate": 0.0006,
+ "loss": 4.562105655670166,
+ "step": 4314
+ },
+ {
+ "epoch": 59.93624454148472,
+ "grad_norm": 0.014750408008694649,
+ "learning_rate": 0.0006,
+ "loss": 4.456263542175293,
+ "step": 4315
+ },
+ {
+ "epoch": 59.95021834061135,
+ "grad_norm": 0.016726011410355568,
+ "learning_rate": 0.0006,
+ "loss": 4.360722064971924,
+ "step": 4316
+ },
+ {
+ "epoch": 59.96419213973799,
+ "grad_norm": 0.01734776981174946,
+ "learning_rate": 0.0006,
+ "loss": 4.572544574737549,
+ "step": 4317
+ },
+ {
+ "epoch": 59.97816593886463,
+ "grad_norm": 0.01822415366768837,
+ "learning_rate": 0.0006,
+ "loss": 4.490501880645752,
+ "step": 4318
+ },
+ {
+ "epoch": 59.992139737991266,
+ "grad_norm": 0.014638577587902546,
+ "learning_rate": 0.0006,
+ "loss": 4.440951824188232,
+ "step": 4319
+ },
+ {
+ "epoch": 60.0,
+ "grad_norm": 0.01255202479660511,
+ "learning_rate": 0.0006,
+ "loss": 4.559123516082764,
+ "step": 4320
+ },
+ {
+ "epoch": 60.0,
+ "eval_loss": 4.741089344024658,
+ "eval_runtime": 56.8997,
+ "eval_samples_per_second": 42.918,
+ "eval_steps_per_second": 1.353,
+ "step": 4320
+ },
+ {
+ "epoch": 60.01397379912664,
+ "grad_norm": 0.016548393294215202,
+ "learning_rate": 0.0006,
+ "loss": 4.442024230957031,
+ "step": 4321
+ },
+ {
+ "epoch": 60.02794759825328,
+ "grad_norm": 0.019759127870202065,
+ "learning_rate": 0.0006,
+ "loss": 4.394334316253662,
+ "step": 4322
+ },
+ {
+ "epoch": 60.041921397379916,
+ "grad_norm": 0.018325170502066612,
+ "learning_rate": 0.0006,
+ "loss": 4.318270683288574,
+ "step": 4323
+ },
+ {
+ "epoch": 60.05589519650655,
+ "grad_norm": 0.015289463102817535,
+ "learning_rate": 0.0006,
+ "loss": 4.318228721618652,
+ "step": 4324
+ },
+ {
+ "epoch": 60.069868995633186,
+ "grad_norm": 0.015424872748553753,
+ "learning_rate": 0.0006,
+ "loss": 4.3605499267578125,
+ "step": 4325
+ },
+ {
+ "epoch": 60.083842794759825,
+ "grad_norm": 0.014085669070482254,
+ "learning_rate": 0.0006,
+ "loss": 4.359585762023926,
+ "step": 4326
+ },
+ {
+ "epoch": 60.09781659388646,
+ "grad_norm": 0.014004210010170937,
+ "learning_rate": 0.0006,
+ "loss": 4.447783470153809,
+ "step": 4327
+ },
+ {
+ "epoch": 60.1117903930131,
+ "grad_norm": 0.013514582999050617,
+ "learning_rate": 0.0006,
+ "loss": 4.4850239753723145,
+ "step": 4328
+ },
+ {
+ "epoch": 60.12576419213974,
+ "grad_norm": 0.013049394823610783,
+ "learning_rate": 0.0006,
+ "loss": 4.528675556182861,
+ "step": 4329
+ },
+ {
+ "epoch": 60.13973799126637,
+ "grad_norm": 0.01378058921545744,
+ "learning_rate": 0.0006,
+ "loss": 4.444823741912842,
+ "step": 4330
+ },
+ {
+ "epoch": 60.15371179039301,
+ "grad_norm": 0.013752233237028122,
+ "learning_rate": 0.0006,
+ "loss": 4.5205206871032715,
+ "step": 4331
+ },
+ {
+ "epoch": 60.16768558951965,
+ "grad_norm": 0.013431803323328495,
+ "learning_rate": 0.0006,
+ "loss": 4.547320365905762,
+ "step": 4332
+ },
+ {
+ "epoch": 60.18165938864629,
+ "grad_norm": 0.014544196426868439,
+ "learning_rate": 0.0006,
+ "loss": 4.517143249511719,
+ "step": 4333
+ },
+ {
+ "epoch": 60.19563318777293,
+ "grad_norm": 0.01612176187336445,
+ "learning_rate": 0.0006,
+ "loss": 4.517232894897461,
+ "step": 4334
+ },
+ {
+ "epoch": 60.209606986899566,
+ "grad_norm": 0.01803239807486534,
+ "learning_rate": 0.0006,
+ "loss": 4.377574920654297,
+ "step": 4335
+ },
+ {
+ "epoch": 60.223580786026204,
+ "grad_norm": 0.014871489256620407,
+ "learning_rate": 0.0006,
+ "loss": 4.538285255432129,
+ "step": 4336
+ },
+ {
+ "epoch": 60.237554585152836,
+ "grad_norm": 0.013927377760410309,
+ "learning_rate": 0.0006,
+ "loss": 4.386175632476807,
+ "step": 4337
+ },
+ {
+ "epoch": 60.251528384279474,
+ "grad_norm": 0.014757219702005386,
+ "learning_rate": 0.0006,
+ "loss": 4.528553009033203,
+ "step": 4338
+ },
+ {
+ "epoch": 60.26550218340611,
+ "grad_norm": 0.01433873176574707,
+ "learning_rate": 0.0006,
+ "loss": 4.476929187774658,
+ "step": 4339
+ },
+ {
+ "epoch": 60.27947598253275,
+ "grad_norm": 0.015768803656101227,
+ "learning_rate": 0.0006,
+ "loss": 4.50970458984375,
+ "step": 4340
+ },
+ {
+ "epoch": 60.29344978165939,
+ "grad_norm": 0.015651429072022438,
+ "learning_rate": 0.0006,
+ "loss": 4.427133083343506,
+ "step": 4341
+ },
+ {
+ "epoch": 60.30742358078603,
+ "grad_norm": 0.01673000678420067,
+ "learning_rate": 0.0006,
+ "loss": 4.447723865509033,
+ "step": 4342
+ },
+ {
+ "epoch": 60.32139737991266,
+ "grad_norm": 0.018481194972991943,
+ "learning_rate": 0.0006,
+ "loss": 4.559832572937012,
+ "step": 4343
+ },
+ {
+ "epoch": 60.3353711790393,
+ "grad_norm": 0.016731027513742447,
+ "learning_rate": 0.0006,
+ "loss": 4.470893859863281,
+ "step": 4344
+ },
+ {
+ "epoch": 60.34934497816594,
+ "grad_norm": 0.01504011545330286,
+ "learning_rate": 0.0006,
+ "loss": 4.483014106750488,
+ "step": 4345
+ },
+ {
+ "epoch": 60.36331877729258,
+ "grad_norm": 0.01602456159889698,
+ "learning_rate": 0.0006,
+ "loss": 4.548428058624268,
+ "step": 4346
+ },
+ {
+ "epoch": 60.377292576419215,
+ "grad_norm": 0.016027364879846573,
+ "learning_rate": 0.0006,
+ "loss": 4.426779747009277,
+ "step": 4347
+ },
+ {
+ "epoch": 60.391266375545854,
+ "grad_norm": 0.01873827911913395,
+ "learning_rate": 0.0006,
+ "loss": 4.353686332702637,
+ "step": 4348
+ },
+ {
+ "epoch": 60.40524017467249,
+ "grad_norm": 0.01907406374812126,
+ "learning_rate": 0.0006,
+ "loss": 4.502945423126221,
+ "step": 4349
+ },
+ {
+ "epoch": 60.419213973799124,
+ "grad_norm": 0.016083354130387306,
+ "learning_rate": 0.0006,
+ "loss": 4.431731700897217,
+ "step": 4350
+ },
+ {
+ "epoch": 60.43318777292576,
+ "grad_norm": 0.015293709933757782,
+ "learning_rate": 0.0006,
+ "loss": 4.480112552642822,
+ "step": 4351
+ },
+ {
+ "epoch": 60.4471615720524,
+ "grad_norm": 0.01688101328909397,
+ "learning_rate": 0.0006,
+ "loss": 4.50853157043457,
+ "step": 4352
+ },
+ {
+ "epoch": 60.46113537117904,
+ "grad_norm": 0.02094118855893612,
+ "learning_rate": 0.0006,
+ "loss": 4.359930038452148,
+ "step": 4353
+ },
+ {
+ "epoch": 60.47510917030568,
+ "grad_norm": 0.018188107758760452,
+ "learning_rate": 0.0006,
+ "loss": 4.549592018127441,
+ "step": 4354
+ },
+ {
+ "epoch": 60.48908296943232,
+ "grad_norm": 0.01462319865822792,
+ "learning_rate": 0.0006,
+ "loss": 4.49467658996582,
+ "step": 4355
+ },
+ {
+ "epoch": 60.50305676855895,
+ "grad_norm": 0.017508579418063164,
+ "learning_rate": 0.0006,
+ "loss": 4.398318767547607,
+ "step": 4356
+ },
+ {
+ "epoch": 60.51703056768559,
+ "grad_norm": 0.018200315535068512,
+ "learning_rate": 0.0006,
+ "loss": 4.44169282913208,
+ "step": 4357
+ },
+ {
+ "epoch": 60.531004366812226,
+ "grad_norm": 0.01717468351125717,
+ "learning_rate": 0.0006,
+ "loss": 4.481400489807129,
+ "step": 4358
+ },
+ {
+ "epoch": 60.544978165938865,
+ "grad_norm": 0.019329151138663292,
+ "learning_rate": 0.0006,
+ "loss": 4.578839302062988,
+ "step": 4359
+ },
+ {
+ "epoch": 60.5589519650655,
+ "grad_norm": 0.016109555959701538,
+ "learning_rate": 0.0006,
+ "loss": 4.461043357849121,
+ "step": 4360
+ },
+ {
+ "epoch": 60.57292576419214,
+ "grad_norm": 0.01494457945227623,
+ "learning_rate": 0.0006,
+ "loss": 4.505096912384033,
+ "step": 4361
+ },
+ {
+ "epoch": 60.58689956331878,
+ "grad_norm": 0.01729399710893631,
+ "learning_rate": 0.0006,
+ "loss": 4.366747856140137,
+ "step": 4362
+ },
+ {
+ "epoch": 60.60087336244541,
+ "grad_norm": 0.01705821603536606,
+ "learning_rate": 0.0006,
+ "loss": 4.42534065246582,
+ "step": 4363
+ },
+ {
+ "epoch": 60.61484716157205,
+ "grad_norm": 0.018391354009509087,
+ "learning_rate": 0.0006,
+ "loss": 4.350294589996338,
+ "step": 4364
+ },
+ {
+ "epoch": 60.62882096069869,
+ "grad_norm": 0.016989829018712044,
+ "learning_rate": 0.0006,
+ "loss": 4.45905876159668,
+ "step": 4365
+ },
+ {
+ "epoch": 60.64279475982533,
+ "grad_norm": 0.017192212864756584,
+ "learning_rate": 0.0006,
+ "loss": 4.526487350463867,
+ "step": 4366
+ },
+ {
+ "epoch": 60.65676855895197,
+ "grad_norm": 0.018091315403580666,
+ "learning_rate": 0.0006,
+ "loss": 4.309333324432373,
+ "step": 4367
+ },
+ {
+ "epoch": 60.670742358078606,
+ "grad_norm": 0.014051459729671478,
+ "learning_rate": 0.0006,
+ "loss": 4.338171005249023,
+ "step": 4368
+ },
+ {
+ "epoch": 60.68471615720524,
+ "grad_norm": 0.013166418299078941,
+ "learning_rate": 0.0006,
+ "loss": 4.409758567810059,
+ "step": 4369
+ },
+ {
+ "epoch": 60.698689956331876,
+ "grad_norm": 0.014260428957641125,
+ "learning_rate": 0.0006,
+ "loss": 4.524338245391846,
+ "step": 4370
+ },
+ {
+ "epoch": 60.712663755458514,
+ "grad_norm": 0.017019255086779594,
+ "learning_rate": 0.0006,
+ "loss": 4.387436866760254,
+ "step": 4371
+ },
+ {
+ "epoch": 60.72663755458515,
+ "grad_norm": 0.017070675268769264,
+ "learning_rate": 0.0006,
+ "loss": 4.437321662902832,
+ "step": 4372
+ },
+ {
+ "epoch": 60.74061135371179,
+ "grad_norm": 0.015197164379060268,
+ "learning_rate": 0.0006,
+ "loss": 4.371205806732178,
+ "step": 4373
+ },
+ {
+ "epoch": 60.75458515283843,
+ "grad_norm": 0.014791185967624187,
+ "learning_rate": 0.0006,
+ "loss": 4.44439697265625,
+ "step": 4374
+ },
+ {
+ "epoch": 60.76855895196506,
+ "grad_norm": 0.016458261758089066,
+ "learning_rate": 0.0006,
+ "loss": 4.445584297180176,
+ "step": 4375
+ },
+ {
+ "epoch": 60.7825327510917,
+ "grad_norm": 0.017706342041492462,
+ "learning_rate": 0.0006,
+ "loss": 4.348708152770996,
+ "step": 4376
+ },
+ {
+ "epoch": 60.79650655021834,
+ "grad_norm": 0.017561476677656174,
+ "learning_rate": 0.0006,
+ "loss": 4.521929740905762,
+ "step": 4377
+ },
+ {
+ "epoch": 60.81048034934498,
+ "grad_norm": 0.016557445749640465,
+ "learning_rate": 0.0006,
+ "loss": 4.545806884765625,
+ "step": 4378
+ },
+ {
+ "epoch": 60.82445414847162,
+ "grad_norm": 0.016953222453594208,
+ "learning_rate": 0.0006,
+ "loss": 4.4230875968933105,
+ "step": 4379
+ },
+ {
+ "epoch": 60.838427947598255,
+ "grad_norm": 0.017094967886805534,
+ "learning_rate": 0.0006,
+ "loss": 4.423985481262207,
+ "step": 4380
+ },
+ {
+ "epoch": 60.852401746724894,
+ "grad_norm": 0.014787515625357628,
+ "learning_rate": 0.0006,
+ "loss": 4.39473819732666,
+ "step": 4381
+ },
+ {
+ "epoch": 60.866375545851525,
+ "grad_norm": 0.016208263114094734,
+ "learning_rate": 0.0006,
+ "loss": 4.3810038566589355,
+ "step": 4382
+ },
+ {
+ "epoch": 60.880349344978164,
+ "grad_norm": 0.017844321206212044,
+ "learning_rate": 0.0006,
+ "loss": 4.35086727142334,
+ "step": 4383
+ },
+ {
+ "epoch": 60.8943231441048,
+ "grad_norm": 0.017539294436573982,
+ "learning_rate": 0.0006,
+ "loss": 4.460260391235352,
+ "step": 4384
+ },
+ {
+ "epoch": 60.90829694323144,
+ "grad_norm": 0.01612042263150215,
+ "learning_rate": 0.0006,
+ "loss": 4.370500564575195,
+ "step": 4385
+ },
+ {
+ "epoch": 60.92227074235808,
+ "grad_norm": 0.015381601639091969,
+ "learning_rate": 0.0006,
+ "loss": 4.435868263244629,
+ "step": 4386
+ },
+ {
+ "epoch": 60.93624454148472,
+ "grad_norm": 0.01603585295379162,
+ "learning_rate": 0.0006,
+ "loss": 4.434070587158203,
+ "step": 4387
+ },
+ {
+ "epoch": 60.95021834061135,
+ "grad_norm": 0.016268130391836166,
+ "learning_rate": 0.0006,
+ "loss": 4.404343128204346,
+ "step": 4388
+ },
+ {
+ "epoch": 60.96419213973799,
+ "grad_norm": 0.014980032108724117,
+ "learning_rate": 0.0006,
+ "loss": 4.399213790893555,
+ "step": 4389
+ },
+ {
+ "epoch": 60.97816593886463,
+ "grad_norm": 0.016899267211556435,
+ "learning_rate": 0.0006,
+ "loss": 4.508069038391113,
+ "step": 4390
+ },
+ {
+ "epoch": 60.992139737991266,
+ "grad_norm": 0.013756908476352692,
+ "learning_rate": 0.0006,
+ "loss": 4.541858673095703,
+ "step": 4391
+ },
+ {
+ "epoch": 61.0,
+ "grad_norm": 0.01578413136303425,
+ "learning_rate": 0.0006,
+ "loss": 4.543513298034668,
+ "step": 4392
+ },
+ {
+ "epoch": 61.0,
+ "eval_loss": 4.756770610809326,
+ "eval_runtime": 56.4572,
+ "eval_samples_per_second": 43.254,
+ "eval_steps_per_second": 1.364,
+ "step": 4392
+ },
+ {
+ "epoch": 61.01397379912664,
+ "grad_norm": 0.013387506827712059,
+ "learning_rate": 0.0006,
+ "loss": 4.475942611694336,
+ "step": 4393
+ },
+ {
+ "epoch": 61.02794759825328,
+ "grad_norm": 0.014278904534876347,
+ "learning_rate": 0.0006,
+ "loss": 4.442403793334961,
+ "step": 4394
+ },
+ {
+ "epoch": 61.041921397379916,
+ "grad_norm": 0.014359553344547749,
+ "learning_rate": 0.0006,
+ "loss": 4.509463310241699,
+ "step": 4395
+ },
+ {
+ "epoch": 61.05589519650655,
+ "grad_norm": 0.01718050241470337,
+ "learning_rate": 0.0006,
+ "loss": 4.354458808898926,
+ "step": 4396
+ },
+ {
+ "epoch": 61.069868995633186,
+ "grad_norm": 0.015988312661647797,
+ "learning_rate": 0.0006,
+ "loss": 4.484508514404297,
+ "step": 4397
+ },
+ {
+ "epoch": 61.083842794759825,
+ "grad_norm": 0.015466735698282719,
+ "learning_rate": 0.0006,
+ "loss": 4.326908111572266,
+ "step": 4398
+ },
+ {
+ "epoch": 61.09781659388646,
+ "grad_norm": 0.014541332609951496,
+ "learning_rate": 0.0006,
+ "loss": 4.405823230743408,
+ "step": 4399
+ },
+ {
+ "epoch": 61.1117903930131,
+ "grad_norm": 0.014803987927734852,
+ "learning_rate": 0.0006,
+ "loss": 4.375386714935303,
+ "step": 4400
+ },
+ {
+ "epoch": 61.12576419213974,
+ "grad_norm": 0.013962093740701675,
+ "learning_rate": 0.0006,
+ "loss": 4.300838470458984,
+ "step": 4401
+ },
+ {
+ "epoch": 61.13973799126637,
+ "grad_norm": 0.015213954262435436,
+ "learning_rate": 0.0006,
+ "loss": 4.388306617736816,
+ "step": 4402
+ },
+ {
+ "epoch": 61.15371179039301,
+ "grad_norm": 0.018965506926178932,
+ "learning_rate": 0.0006,
+ "loss": 4.388340950012207,
+ "step": 4403
+ },
+ {
+ "epoch": 61.16768558951965,
+ "grad_norm": 0.01958410255610943,
+ "learning_rate": 0.0006,
+ "loss": 4.35012149810791,
+ "step": 4404
+ },
+ {
+ "epoch": 61.18165938864629,
+ "grad_norm": 0.01938101276755333,
+ "learning_rate": 0.0006,
+ "loss": 4.368372917175293,
+ "step": 4405
+ },
+ {
+ "epoch": 61.19563318777293,
+ "grad_norm": 0.016556516289711,
+ "learning_rate": 0.0006,
+ "loss": 4.486198425292969,
+ "step": 4406
+ },
+ {
+ "epoch": 61.209606986899566,
+ "grad_norm": 0.017031289637088776,
+ "learning_rate": 0.0006,
+ "loss": 4.421676158905029,
+ "step": 4407
+ },
+ {
+ "epoch": 61.223580786026204,
+ "grad_norm": 0.018023479729890823,
+ "learning_rate": 0.0006,
+ "loss": 4.38751220703125,
+ "step": 4408
+ },
+ {
+ "epoch": 61.237554585152836,
+ "grad_norm": 0.0178135447204113,
+ "learning_rate": 0.0006,
+ "loss": 4.336528301239014,
+ "step": 4409
+ },
+ {
+ "epoch": 61.251528384279474,
+ "grad_norm": 0.016782190650701523,
+ "learning_rate": 0.0006,
+ "loss": 4.386693954467773,
+ "step": 4410
+ },
+ {
+ "epoch": 61.26550218340611,
+ "grad_norm": 0.015494848601520061,
+ "learning_rate": 0.0006,
+ "loss": 4.308948040008545,
+ "step": 4411
+ },
+ {
+ "epoch": 61.27947598253275,
+ "grad_norm": 0.015002378262579441,
+ "learning_rate": 0.0006,
+ "loss": 4.314206123352051,
+ "step": 4412
+ },
+ {
+ "epoch": 61.29344978165939,
+ "grad_norm": 0.015798920765519142,
+ "learning_rate": 0.0006,
+ "loss": 4.502900123596191,
+ "step": 4413
+ },
+ {
+ "epoch": 61.30742358078603,
+ "grad_norm": 0.015444336459040642,
+ "learning_rate": 0.0006,
+ "loss": 4.473160743713379,
+ "step": 4414
+ },
+ {
+ "epoch": 61.32139737991266,
+ "grad_norm": 0.017025096341967583,
+ "learning_rate": 0.0006,
+ "loss": 4.460621356964111,
+ "step": 4415
+ },
+ {
+ "epoch": 61.3353711790393,
+ "grad_norm": 0.019502537325024605,
+ "learning_rate": 0.0006,
+ "loss": 4.385551452636719,
+ "step": 4416
+ },
+ {
+ "epoch": 61.34934497816594,
+ "grad_norm": 0.018529541790485382,
+ "learning_rate": 0.0006,
+ "loss": 4.425390720367432,
+ "step": 4417
+ },
+ {
+ "epoch": 61.36331877729258,
+ "grad_norm": 0.013155706226825714,
+ "learning_rate": 0.0006,
+ "loss": 4.467020034790039,
+ "step": 4418
+ },
+ {
+ "epoch": 61.377292576419215,
+ "grad_norm": 0.01518090907484293,
+ "learning_rate": 0.0006,
+ "loss": 4.369664192199707,
+ "step": 4419
+ },
+ {
+ "epoch": 61.391266375545854,
+ "grad_norm": 0.01626511849462986,
+ "learning_rate": 0.0006,
+ "loss": 4.463231086730957,
+ "step": 4420
+ },
+ {
+ "epoch": 61.40524017467249,
+ "grad_norm": 0.01701558195054531,
+ "learning_rate": 0.0006,
+ "loss": 4.43708610534668,
+ "step": 4421
+ },
+ {
+ "epoch": 61.419213973799124,
+ "grad_norm": 0.01761862449347973,
+ "learning_rate": 0.0006,
+ "loss": 4.3314208984375,
+ "step": 4422
+ },
+ {
+ "epoch": 61.43318777292576,
+ "grad_norm": 0.017084648832678795,
+ "learning_rate": 0.0006,
+ "loss": 4.39676570892334,
+ "step": 4423
+ },
+ {
+ "epoch": 61.4471615720524,
+ "grad_norm": 0.01667078025639057,
+ "learning_rate": 0.0006,
+ "loss": 4.3789873123168945,
+ "step": 4424
+ },
+ {
+ "epoch": 61.46113537117904,
+ "grad_norm": 0.01877436414361,
+ "learning_rate": 0.0006,
+ "loss": 4.415275573730469,
+ "step": 4425
+ },
+ {
+ "epoch": 61.47510917030568,
+ "grad_norm": 0.020250486209988594,
+ "learning_rate": 0.0006,
+ "loss": 4.478635787963867,
+ "step": 4426
+ },
+ {
+ "epoch": 61.48908296943232,
+ "grad_norm": 0.020244868472218513,
+ "learning_rate": 0.0006,
+ "loss": 4.296854019165039,
+ "step": 4427
+ },
+ {
+ "epoch": 61.50305676855895,
+ "grad_norm": 0.018723875284194946,
+ "learning_rate": 0.0006,
+ "loss": 4.3760881423950195,
+ "step": 4428
+ },
+ {
+ "epoch": 61.51703056768559,
+ "grad_norm": 0.015138164162635803,
+ "learning_rate": 0.0006,
+ "loss": 4.39974308013916,
+ "step": 4429
+ },
+ {
+ "epoch": 61.531004366812226,
+ "grad_norm": 0.015111138112843037,
+ "learning_rate": 0.0006,
+ "loss": 4.34659481048584,
+ "step": 4430
+ },
+ {
+ "epoch": 61.544978165938865,
+ "grad_norm": 0.0164666585624218,
+ "learning_rate": 0.0006,
+ "loss": 4.415737152099609,
+ "step": 4431
+ },
+ {
+ "epoch": 61.5589519650655,
+ "grad_norm": 0.020891312509775162,
+ "learning_rate": 0.0006,
+ "loss": 4.300319671630859,
+ "step": 4432
+ },
+ {
+ "epoch": 61.57292576419214,
+ "grad_norm": 0.02089156024158001,
+ "learning_rate": 0.0006,
+ "loss": 4.373029708862305,
+ "step": 4433
+ },
+ {
+ "epoch": 61.58689956331878,
+ "grad_norm": 0.020114559680223465,
+ "learning_rate": 0.0006,
+ "loss": 4.434499740600586,
+ "step": 4434
+ },
+ {
+ "epoch": 61.60087336244541,
+ "grad_norm": 0.01791365258395672,
+ "learning_rate": 0.0006,
+ "loss": 4.466872215270996,
+ "step": 4435
+ },
+ {
+ "epoch": 61.61484716157205,
+ "grad_norm": 0.01879458874464035,
+ "learning_rate": 0.0006,
+ "loss": 4.5176520347595215,
+ "step": 4436
+ },
+ {
+ "epoch": 61.62882096069869,
+ "grad_norm": 0.016371937468647957,
+ "learning_rate": 0.0006,
+ "loss": 4.3507585525512695,
+ "step": 4437
+ },
+ {
+ "epoch": 61.64279475982533,
+ "grad_norm": 0.01648143120110035,
+ "learning_rate": 0.0006,
+ "loss": 4.398778438568115,
+ "step": 4438
+ },
+ {
+ "epoch": 61.65676855895197,
+ "grad_norm": 0.017240576446056366,
+ "learning_rate": 0.0006,
+ "loss": 4.513933181762695,
+ "step": 4439
+ },
+ {
+ "epoch": 61.670742358078606,
+ "grad_norm": 0.01734284684062004,
+ "learning_rate": 0.0006,
+ "loss": 4.315343856811523,
+ "step": 4440
+ },
+ {
+ "epoch": 61.68471615720524,
+ "grad_norm": 0.01639772206544876,
+ "learning_rate": 0.0006,
+ "loss": 4.4641218185424805,
+ "step": 4441
+ },
+ {
+ "epoch": 61.698689956331876,
+ "grad_norm": 0.0164708960801363,
+ "learning_rate": 0.0006,
+ "loss": 4.439968585968018,
+ "step": 4442
+ },
+ {
+ "epoch": 61.712663755458514,
+ "grad_norm": 0.01757933758199215,
+ "learning_rate": 0.0006,
+ "loss": 4.408327102661133,
+ "step": 4443
+ },
+ {
+ "epoch": 61.72663755458515,
+ "grad_norm": 0.01789589412510395,
+ "learning_rate": 0.0006,
+ "loss": 4.478257179260254,
+ "step": 4444
+ },
+ {
+ "epoch": 61.74061135371179,
+ "grad_norm": 0.01822184957563877,
+ "learning_rate": 0.0006,
+ "loss": 4.412046909332275,
+ "step": 4445
+ },
+ {
+ "epoch": 61.75458515283843,
+ "grad_norm": 0.018624618649482727,
+ "learning_rate": 0.0006,
+ "loss": 4.427122116088867,
+ "step": 4446
+ },
+ {
+ "epoch": 61.76855895196506,
+ "grad_norm": 0.018674887716770172,
+ "learning_rate": 0.0006,
+ "loss": 4.429147720336914,
+ "step": 4447
+ },
+ {
+ "epoch": 61.7825327510917,
+ "grad_norm": 0.015317119657993317,
+ "learning_rate": 0.0006,
+ "loss": 4.2808451652526855,
+ "step": 4448
+ },
+ {
+ "epoch": 61.79650655021834,
+ "grad_norm": 0.015400845557451248,
+ "learning_rate": 0.0006,
+ "loss": 4.420193672180176,
+ "step": 4449
+ },
+ {
+ "epoch": 61.81048034934498,
+ "grad_norm": 0.016743650659918785,
+ "learning_rate": 0.0006,
+ "loss": 4.498254776000977,
+ "step": 4450
+ },
+ {
+ "epoch": 61.82445414847162,
+ "grad_norm": 0.014330781064927578,
+ "learning_rate": 0.0006,
+ "loss": 4.445384979248047,
+ "step": 4451
+ },
+ {
+ "epoch": 61.838427947598255,
+ "grad_norm": 0.015046479180455208,
+ "learning_rate": 0.0006,
+ "loss": 4.368041038513184,
+ "step": 4452
+ },
+ {
+ "epoch": 61.852401746724894,
+ "grad_norm": 0.01630168780684471,
+ "learning_rate": 0.0006,
+ "loss": 4.460142135620117,
+ "step": 4453
+ },
+ {
+ "epoch": 61.866375545851525,
+ "grad_norm": 0.01637885719537735,
+ "learning_rate": 0.0006,
+ "loss": 4.416606426239014,
+ "step": 4454
+ },
+ {
+ "epoch": 61.880349344978164,
+ "grad_norm": 0.016785888001322746,
+ "learning_rate": 0.0006,
+ "loss": 4.452545642852783,
+ "step": 4455
+ },
+ {
+ "epoch": 61.8943231441048,
+ "grad_norm": 0.015044555068016052,
+ "learning_rate": 0.0006,
+ "loss": 4.427334785461426,
+ "step": 4456
+ },
+ {
+ "epoch": 61.90829694323144,
+ "grad_norm": 0.013699211180210114,
+ "learning_rate": 0.0006,
+ "loss": 4.408761024475098,
+ "step": 4457
+ },
+ {
+ "epoch": 61.92227074235808,
+ "grad_norm": 0.014021803624927998,
+ "learning_rate": 0.0006,
+ "loss": 4.543607711791992,
+ "step": 4458
+ },
+ {
+ "epoch": 61.93624454148472,
+ "grad_norm": 0.015438460744917393,
+ "learning_rate": 0.0006,
+ "loss": 4.313363075256348,
+ "step": 4459
+ },
+ {
+ "epoch": 61.95021834061135,
+ "grad_norm": 0.017515428364276886,
+ "learning_rate": 0.0006,
+ "loss": 4.310412406921387,
+ "step": 4460
+ },
+ {
+ "epoch": 61.96419213973799,
+ "grad_norm": 0.016683880239725113,
+ "learning_rate": 0.0006,
+ "loss": 4.3378448486328125,
+ "step": 4461
+ },
+ {
+ "epoch": 61.97816593886463,
+ "grad_norm": 0.019384048879146576,
+ "learning_rate": 0.0006,
+ "loss": 4.450568675994873,
+ "step": 4462
+ },
+ {
+ "epoch": 61.992139737991266,
+ "grad_norm": 0.020651493221521378,
+ "learning_rate": 0.0006,
+ "loss": 4.453197479248047,
+ "step": 4463
+ },
+ {
+ "epoch": 62.0,
+ "grad_norm": 0.020991764962673187,
+ "learning_rate": 0.0006,
+ "loss": 4.441527366638184,
+ "step": 4464
+ },
+ {
+ "epoch": 62.0,
+ "eval_loss": 4.700193881988525,
+ "eval_runtime": 56.6098,
+ "eval_samples_per_second": 43.137,
+ "eval_steps_per_second": 1.36,
+ "step": 4464
+ },
+ {
+ "epoch": 62.01397379912664,
+ "grad_norm": 0.01811373233795166,
+ "learning_rate": 0.0006,
+ "loss": 4.37346076965332,
+ "step": 4465
+ },
+ {
+ "epoch": 62.02794759825328,
+ "grad_norm": 0.01667127199470997,
+ "learning_rate": 0.0006,
+ "loss": 4.237151145935059,
+ "step": 4466
+ },
+ {
+ "epoch": 62.041921397379916,
+ "grad_norm": 0.014971534721553326,
+ "learning_rate": 0.0006,
+ "loss": 4.367053031921387,
+ "step": 4467
+ },
+ {
+ "epoch": 62.05589519650655,
+ "grad_norm": 0.015655893832445145,
+ "learning_rate": 0.0006,
+ "loss": 4.437426567077637,
+ "step": 4468
+ },
+ {
+ "epoch": 62.069868995633186,
+ "grad_norm": 0.014935447834432125,
+ "learning_rate": 0.0006,
+ "loss": 4.3635125160217285,
+ "step": 4469
+ },
+ {
+ "epoch": 62.083842794759825,
+ "grad_norm": 0.015363802202045918,
+ "learning_rate": 0.0006,
+ "loss": 4.40492057800293,
+ "step": 4470
+ },
+ {
+ "epoch": 62.09781659388646,
+ "grad_norm": 0.0159201230853796,
+ "learning_rate": 0.0006,
+ "loss": 4.392058849334717,
+ "step": 4471
+ },
+ {
+ "epoch": 62.1117903930131,
+ "grad_norm": 0.014598241075873375,
+ "learning_rate": 0.0006,
+ "loss": 4.369944095611572,
+ "step": 4472
+ },
+ {
+ "epoch": 62.12576419213974,
+ "grad_norm": 0.015394841320812702,
+ "learning_rate": 0.0006,
+ "loss": 4.425097942352295,
+ "step": 4473
+ },
+ {
+ "epoch": 62.13973799126637,
+ "grad_norm": 0.015913628041744232,
+ "learning_rate": 0.0006,
+ "loss": 4.420596122741699,
+ "step": 4474
+ },
+ {
+ "epoch": 62.15371179039301,
+ "grad_norm": 0.01575258933007717,
+ "learning_rate": 0.0006,
+ "loss": 4.32290506362915,
+ "step": 4475
+ },
+ {
+ "epoch": 62.16768558951965,
+ "grad_norm": 0.0155344782397151,
+ "learning_rate": 0.0006,
+ "loss": 4.273068428039551,
+ "step": 4476
+ },
+ {
+ "epoch": 62.18165938864629,
+ "grad_norm": 0.014947040006518364,
+ "learning_rate": 0.0006,
+ "loss": 4.384868621826172,
+ "step": 4477
+ },
+ {
+ "epoch": 62.19563318777293,
+ "grad_norm": 0.01557591650635004,
+ "learning_rate": 0.0006,
+ "loss": 4.458410263061523,
+ "step": 4478
+ },
+ {
+ "epoch": 62.209606986899566,
+ "grad_norm": 0.015182198956608772,
+ "learning_rate": 0.0006,
+ "loss": 4.300738334655762,
+ "step": 4479
+ },
+ {
+ "epoch": 62.223580786026204,
+ "grad_norm": 0.01634717360138893,
+ "learning_rate": 0.0006,
+ "loss": 4.436875820159912,
+ "step": 4480
+ },
+ {
+ "epoch": 62.237554585152836,
+ "grad_norm": 0.018813718110322952,
+ "learning_rate": 0.0006,
+ "loss": 4.454665184020996,
+ "step": 4481
+ },
+ {
+ "epoch": 62.251528384279474,
+ "grad_norm": 0.020965630188584328,
+ "learning_rate": 0.0006,
+ "loss": 4.357341766357422,
+ "step": 4482
+ },
+ {
+ "epoch": 62.26550218340611,
+ "grad_norm": 0.020738353952765465,
+ "learning_rate": 0.0006,
+ "loss": 4.456644058227539,
+ "step": 4483
+ },
+ {
+ "epoch": 62.27947598253275,
+ "grad_norm": 0.018813546746969223,
+ "learning_rate": 0.0006,
+ "loss": 4.381941795349121,
+ "step": 4484
+ },
+ {
+ "epoch": 62.29344978165939,
+ "grad_norm": 0.015431761741638184,
+ "learning_rate": 0.0006,
+ "loss": 4.41126823425293,
+ "step": 4485
+ },
+ {
+ "epoch": 62.30742358078603,
+ "grad_norm": 0.016335977241396904,
+ "learning_rate": 0.0006,
+ "loss": 4.419488906860352,
+ "step": 4486
+ },
+ {
+ "epoch": 62.32139737991266,
+ "grad_norm": 0.01832406409084797,
+ "learning_rate": 0.0006,
+ "loss": 4.579026222229004,
+ "step": 4487
+ },
+ {
+ "epoch": 62.3353711790393,
+ "grad_norm": 0.01790742017328739,
+ "learning_rate": 0.0006,
+ "loss": 4.289964199066162,
+ "step": 4488
+ },
+ {
+ "epoch": 62.34934497816594,
+ "grad_norm": 0.017266202718019485,
+ "learning_rate": 0.0006,
+ "loss": 4.415959358215332,
+ "step": 4489
+ },
+ {
+ "epoch": 62.36331877729258,
+ "grad_norm": 0.015976132825016975,
+ "learning_rate": 0.0006,
+ "loss": 4.358587265014648,
+ "step": 4490
+ },
+ {
+ "epoch": 62.377292576419215,
+ "grad_norm": 0.016873881220817566,
+ "learning_rate": 0.0006,
+ "loss": 4.354607105255127,
+ "step": 4491
+ },
+ {
+ "epoch": 62.391266375545854,
+ "grad_norm": 0.01476745493710041,
+ "learning_rate": 0.0006,
+ "loss": 4.384904384613037,
+ "step": 4492
+ },
+ {
+ "epoch": 62.40524017467249,
+ "grad_norm": 0.013810782693326473,
+ "learning_rate": 0.0006,
+ "loss": 4.459230899810791,
+ "step": 4493
+ },
+ {
+ "epoch": 62.419213973799124,
+ "grad_norm": 0.015150421299040318,
+ "learning_rate": 0.0006,
+ "loss": 4.369471549987793,
+ "step": 4494
+ },
+ {
+ "epoch": 62.43318777292576,
+ "grad_norm": 0.014539610594511032,
+ "learning_rate": 0.0006,
+ "loss": 4.3564982414245605,
+ "step": 4495
+ },
+ {
+ "epoch": 62.4471615720524,
+ "grad_norm": 0.015554017387330532,
+ "learning_rate": 0.0006,
+ "loss": 4.414379119873047,
+ "step": 4496
+ },
+ {
+ "epoch": 62.46113537117904,
+ "grad_norm": 0.016608357429504395,
+ "learning_rate": 0.0006,
+ "loss": 4.311585426330566,
+ "step": 4497
+ },
+ {
+ "epoch": 62.47510917030568,
+ "grad_norm": 0.01885881833732128,
+ "learning_rate": 0.0006,
+ "loss": 4.306629180908203,
+ "step": 4498
+ },
+ {
+ "epoch": 62.48908296943232,
+ "grad_norm": 0.022128663957118988,
+ "learning_rate": 0.0006,
+ "loss": 4.345483779907227,
+ "step": 4499
+ },
+ {
+ "epoch": 62.50305676855895,
+ "grad_norm": 0.023401018232107162,
+ "learning_rate": 0.0006,
+ "loss": 4.398204326629639,
+ "step": 4500
+ },
+ {
+ "epoch": 62.51703056768559,
+ "grad_norm": 0.022080009803175926,
+ "learning_rate": 0.0006,
+ "loss": 4.375027179718018,
+ "step": 4501
+ },
+ {
+ "epoch": 62.531004366812226,
+ "grad_norm": 0.020004864782094955,
+ "learning_rate": 0.0006,
+ "loss": 4.33494758605957,
+ "step": 4502
+ },
+ {
+ "epoch": 62.544978165938865,
+ "grad_norm": 0.01786700263619423,
+ "learning_rate": 0.0006,
+ "loss": 4.4428300857543945,
+ "step": 4503
+ },
+ {
+ "epoch": 62.5589519650655,
+ "grad_norm": 0.01885388419032097,
+ "learning_rate": 0.0006,
+ "loss": 4.4138946533203125,
+ "step": 4504
+ },
+ {
+ "epoch": 62.57292576419214,
+ "grad_norm": 0.019890939816832542,
+ "learning_rate": 0.0006,
+ "loss": 4.278632164001465,
+ "step": 4505
+ },
+ {
+ "epoch": 62.58689956331878,
+ "grad_norm": 0.02109031192958355,
+ "learning_rate": 0.0006,
+ "loss": 4.3756327629089355,
+ "step": 4506
+ },
+ {
+ "epoch": 62.60087336244541,
+ "grad_norm": 0.020813899114727974,
+ "learning_rate": 0.0006,
+ "loss": 4.410213470458984,
+ "step": 4507
+ },
+ {
+ "epoch": 62.61484716157205,
+ "grad_norm": 0.021655689924955368,
+ "learning_rate": 0.0006,
+ "loss": 4.30633020401001,
+ "step": 4508
+ },
+ {
+ "epoch": 62.62882096069869,
+ "grad_norm": 0.018155649304389954,
+ "learning_rate": 0.0006,
+ "loss": 4.407284736633301,
+ "step": 4509
+ },
+ {
+ "epoch": 62.64279475982533,
+ "grad_norm": 0.017765069380402565,
+ "learning_rate": 0.0006,
+ "loss": 4.321715354919434,
+ "step": 4510
+ },
+ {
+ "epoch": 62.65676855895197,
+ "grad_norm": 0.017724554985761642,
+ "learning_rate": 0.0006,
+ "loss": 4.493371963500977,
+ "step": 4511
+ },
+ {
+ "epoch": 62.670742358078606,
+ "grad_norm": 0.01492092851549387,
+ "learning_rate": 0.0006,
+ "loss": 4.443025588989258,
+ "step": 4512
+ },
+ {
+ "epoch": 62.68471615720524,
+ "grad_norm": 0.01615324057638645,
+ "learning_rate": 0.0006,
+ "loss": 4.3289594650268555,
+ "step": 4513
+ },
+ {
+ "epoch": 62.698689956331876,
+ "grad_norm": 0.016569366678595543,
+ "learning_rate": 0.0006,
+ "loss": 4.369720458984375,
+ "step": 4514
+ },
+ {
+ "epoch": 62.712663755458514,
+ "grad_norm": 0.016509901732206345,
+ "learning_rate": 0.0006,
+ "loss": 4.292316436767578,
+ "step": 4515
+ },
+ {
+ "epoch": 62.72663755458515,
+ "grad_norm": 0.014270003885030746,
+ "learning_rate": 0.0006,
+ "loss": 4.349854469299316,
+ "step": 4516
+ },
+ {
+ "epoch": 62.74061135371179,
+ "grad_norm": 0.014235340990126133,
+ "learning_rate": 0.0006,
+ "loss": 4.473727226257324,
+ "step": 4517
+ },
+ {
+ "epoch": 62.75458515283843,
+ "grad_norm": 0.015163225121796131,
+ "learning_rate": 0.0006,
+ "loss": 4.452878475189209,
+ "step": 4518
+ },
+ {
+ "epoch": 62.76855895196506,
+ "grad_norm": 0.015748055651783943,
+ "learning_rate": 0.0006,
+ "loss": 4.521087169647217,
+ "step": 4519
+ },
+ {
+ "epoch": 62.7825327510917,
+ "grad_norm": 0.01661108434200287,
+ "learning_rate": 0.0006,
+ "loss": 4.443885326385498,
+ "step": 4520
+ },
+ {
+ "epoch": 62.79650655021834,
+ "grad_norm": 0.01610301062464714,
+ "learning_rate": 0.0006,
+ "loss": 4.312990188598633,
+ "step": 4521
+ },
+ {
+ "epoch": 62.81048034934498,
+ "grad_norm": 0.017007583752274513,
+ "learning_rate": 0.0006,
+ "loss": 4.408937454223633,
+ "step": 4522
+ },
+ {
+ "epoch": 62.82445414847162,
+ "grad_norm": 0.01749461516737938,
+ "learning_rate": 0.0006,
+ "loss": 4.346066951751709,
+ "step": 4523
+ },
+ {
+ "epoch": 62.838427947598255,
+ "grad_norm": 0.018243493512272835,
+ "learning_rate": 0.0006,
+ "loss": 4.357658386230469,
+ "step": 4524
+ },
+ {
+ "epoch": 62.852401746724894,
+ "grad_norm": 0.017499985173344612,
+ "learning_rate": 0.0006,
+ "loss": 4.374061584472656,
+ "step": 4525
+ },
+ {
+ "epoch": 62.866375545851525,
+ "grad_norm": 0.014536325819790363,
+ "learning_rate": 0.0006,
+ "loss": 4.189321994781494,
+ "step": 4526
+ },
+ {
+ "epoch": 62.880349344978164,
+ "grad_norm": 0.014728891663253307,
+ "learning_rate": 0.0006,
+ "loss": 4.425910949707031,
+ "step": 4527
+ },
+ {
+ "epoch": 62.8943231441048,
+ "grad_norm": 0.015013772062957287,
+ "learning_rate": 0.0006,
+ "loss": 4.337319850921631,
+ "step": 4528
+ },
+ {
+ "epoch": 62.90829694323144,
+ "grad_norm": 0.01626027375459671,
+ "learning_rate": 0.0006,
+ "loss": 4.330339431762695,
+ "step": 4529
+ },
+ {
+ "epoch": 62.92227074235808,
+ "grad_norm": 0.014773444272577763,
+ "learning_rate": 0.0006,
+ "loss": 4.4188737869262695,
+ "step": 4530
+ },
+ {
+ "epoch": 62.93624454148472,
+ "grad_norm": 0.01422932744026184,
+ "learning_rate": 0.0006,
+ "loss": 4.355986595153809,
+ "step": 4531
+ },
+ {
+ "epoch": 62.95021834061135,
+ "grad_norm": 0.012170110829174519,
+ "learning_rate": 0.0006,
+ "loss": 4.405722618103027,
+ "step": 4532
+ },
+ {
+ "epoch": 62.96419213973799,
+ "grad_norm": 0.013233253732323647,
+ "learning_rate": 0.0006,
+ "loss": 4.377286911010742,
+ "step": 4533
+ },
+ {
+ "epoch": 62.97816593886463,
+ "grad_norm": 0.014497299678623676,
+ "learning_rate": 0.0006,
+ "loss": 4.508796691894531,
+ "step": 4534
+ },
+ {
+ "epoch": 62.992139737991266,
+ "grad_norm": 0.014718741178512573,
+ "learning_rate": 0.0006,
+ "loss": 4.370919704437256,
+ "step": 4535
+ },
+ {
+ "epoch": 63.0,
+ "grad_norm": 0.015833502635359764,
+ "learning_rate": 0.0006,
+ "loss": 4.258753776550293,
+ "step": 4536
+ },
+ {
+ "epoch": 63.0,
+ "eval_loss": 4.714163303375244,
+ "eval_runtime": 57.0269,
+ "eval_samples_per_second": 42.822,
+ "eval_steps_per_second": 1.35,
+ "step": 4536
+ },
+ {
+ "epoch": 63.01397379912664,
+ "grad_norm": 0.014277510344982147,
+ "learning_rate": 0.0006,
+ "loss": 4.3112640380859375,
+ "step": 4537
+ },
+ {
+ "epoch": 63.02794759825328,
+ "grad_norm": 0.015377935022115707,
+ "learning_rate": 0.0006,
+ "loss": 4.342782974243164,
+ "step": 4538
+ },
+ {
+ "epoch": 63.041921397379916,
+ "grad_norm": 0.015406244434416294,
+ "learning_rate": 0.0006,
+ "loss": 4.474056720733643,
+ "step": 4539
+ },
+ {
+ "epoch": 63.05589519650655,
+ "grad_norm": 0.018121499568223953,
+ "learning_rate": 0.0006,
+ "loss": 4.397963523864746,
+ "step": 4540
+ },
+ {
+ "epoch": 63.069868995633186,
+ "grad_norm": 0.022634636610746384,
+ "learning_rate": 0.0006,
+ "loss": 4.306631088256836,
+ "step": 4541
+ },
+ {
+ "epoch": 63.083842794759825,
+ "grad_norm": 0.026448970660567284,
+ "learning_rate": 0.0006,
+ "loss": 4.283175468444824,
+ "step": 4542
+ },
+ {
+ "epoch": 63.09781659388646,
+ "grad_norm": 0.024104537442326546,
+ "learning_rate": 0.0006,
+ "loss": 4.315826416015625,
+ "step": 4543
+ },
+ {
+ "epoch": 63.1117903930131,
+ "grad_norm": 0.019040480256080627,
+ "learning_rate": 0.0006,
+ "loss": 4.44057035446167,
+ "step": 4544
+ },
+ {
+ "epoch": 63.12576419213974,
+ "grad_norm": 0.02006000280380249,
+ "learning_rate": 0.0006,
+ "loss": 4.405603408813477,
+ "step": 4545
+ },
+ {
+ "epoch": 63.13973799126637,
+ "grad_norm": 0.019980071112513542,
+ "learning_rate": 0.0006,
+ "loss": 4.393726348876953,
+ "step": 4546
+ },
+ {
+ "epoch": 63.15371179039301,
+ "grad_norm": 0.01641865260899067,
+ "learning_rate": 0.0006,
+ "loss": 4.272546768188477,
+ "step": 4547
+ },
+ {
+ "epoch": 63.16768558951965,
+ "grad_norm": 0.01890353485941887,
+ "learning_rate": 0.0006,
+ "loss": 4.382566928863525,
+ "step": 4548
+ },
+ {
+ "epoch": 63.18165938864629,
+ "grad_norm": 0.018370937556028366,
+ "learning_rate": 0.0006,
+ "loss": 4.423953533172607,
+ "step": 4549
+ },
+ {
+ "epoch": 63.19563318777293,
+ "grad_norm": 0.014985294081270695,
+ "learning_rate": 0.0006,
+ "loss": 4.308627128601074,
+ "step": 4550
+ },
+ {
+ "epoch": 63.209606986899566,
+ "grad_norm": 0.017513327300548553,
+ "learning_rate": 0.0006,
+ "loss": 4.248819351196289,
+ "step": 4551
+ },
+ {
+ "epoch": 63.223580786026204,
+ "grad_norm": 0.019464161247015,
+ "learning_rate": 0.0006,
+ "loss": 4.359729766845703,
+ "step": 4552
+ },
+ {
+ "epoch": 63.237554585152836,
+ "grad_norm": 0.018678035587072372,
+ "learning_rate": 0.0006,
+ "loss": 4.3299665451049805,
+ "step": 4553
+ },
+ {
+ "epoch": 63.251528384279474,
+ "grad_norm": 0.017909523099660873,
+ "learning_rate": 0.0006,
+ "loss": 4.246147155761719,
+ "step": 4554
+ },
+ {
+ "epoch": 63.26550218340611,
+ "grad_norm": 0.01747293211519718,
+ "learning_rate": 0.0006,
+ "loss": 4.41165018081665,
+ "step": 4555
+ },
+ {
+ "epoch": 63.27947598253275,
+ "grad_norm": 0.018670853227376938,
+ "learning_rate": 0.0006,
+ "loss": 4.409883499145508,
+ "step": 4556
+ },
+ {
+ "epoch": 63.29344978165939,
+ "grad_norm": 0.0189529862254858,
+ "learning_rate": 0.0006,
+ "loss": 4.418654918670654,
+ "step": 4557
+ },
+ {
+ "epoch": 63.30742358078603,
+ "grad_norm": 0.018394414335489273,
+ "learning_rate": 0.0006,
+ "loss": 4.38355827331543,
+ "step": 4558
+ },
+ {
+ "epoch": 63.32139737991266,
+ "grad_norm": 0.018317820504307747,
+ "learning_rate": 0.0006,
+ "loss": 4.456843376159668,
+ "step": 4559
+ },
+ {
+ "epoch": 63.3353711790393,
+ "grad_norm": 0.019225597381591797,
+ "learning_rate": 0.0006,
+ "loss": 4.232264518737793,
+ "step": 4560
+ },
+ {
+ "epoch": 63.34934497816594,
+ "grad_norm": 0.016663681715726852,
+ "learning_rate": 0.0006,
+ "loss": 4.393517017364502,
+ "step": 4561
+ },
+ {
+ "epoch": 63.36331877729258,
+ "grad_norm": 0.014975365251302719,
+ "learning_rate": 0.0006,
+ "loss": 4.375183582305908,
+ "step": 4562
+ },
+ {
+ "epoch": 63.377292576419215,
+ "grad_norm": 0.01771732047200203,
+ "learning_rate": 0.0006,
+ "loss": 4.3073410987854,
+ "step": 4563
+ },
+ {
+ "epoch": 63.391266375545854,
+ "grad_norm": 0.01663549244403839,
+ "learning_rate": 0.0006,
+ "loss": 4.311003684997559,
+ "step": 4564
+ },
+ {
+ "epoch": 63.40524017467249,
+ "grad_norm": 0.016301177442073822,
+ "learning_rate": 0.0006,
+ "loss": 4.326807022094727,
+ "step": 4565
+ },
+ {
+ "epoch": 63.419213973799124,
+ "grad_norm": 0.016781188547611237,
+ "learning_rate": 0.0006,
+ "loss": 4.306562423706055,
+ "step": 4566
+ },
+ {
+ "epoch": 63.43318777292576,
+ "grad_norm": 0.01767745055258274,
+ "learning_rate": 0.0006,
+ "loss": 4.367082595825195,
+ "step": 4567
+ },
+ {
+ "epoch": 63.4471615720524,
+ "grad_norm": 0.016728565096855164,
+ "learning_rate": 0.0006,
+ "loss": 4.327400207519531,
+ "step": 4568
+ },
+ {
+ "epoch": 63.46113537117904,
+ "grad_norm": 0.016289403662085533,
+ "learning_rate": 0.0006,
+ "loss": 4.392712116241455,
+ "step": 4569
+ },
+ {
+ "epoch": 63.47510917030568,
+ "grad_norm": 0.016233716160058975,
+ "learning_rate": 0.0006,
+ "loss": 4.354372024536133,
+ "step": 4570
+ },
+ {
+ "epoch": 63.48908296943232,
+ "grad_norm": 0.016547974199056625,
+ "learning_rate": 0.0006,
+ "loss": 4.309050559997559,
+ "step": 4571
+ },
+ {
+ "epoch": 63.50305676855895,
+ "grad_norm": 0.018466509878635406,
+ "learning_rate": 0.0006,
+ "loss": 4.318345069885254,
+ "step": 4572
+ },
+ {
+ "epoch": 63.51703056768559,
+ "grad_norm": 0.017685720697045326,
+ "learning_rate": 0.0006,
+ "loss": 4.399726867675781,
+ "step": 4573
+ },
+ {
+ "epoch": 63.531004366812226,
+ "grad_norm": 0.015801234170794487,
+ "learning_rate": 0.0006,
+ "loss": 4.328495502471924,
+ "step": 4574
+ },
+ {
+ "epoch": 63.544978165938865,
+ "grad_norm": 0.0188873540610075,
+ "learning_rate": 0.0006,
+ "loss": 4.253244400024414,
+ "step": 4575
+ },
+ {
+ "epoch": 63.5589519650655,
+ "grad_norm": 0.01868271268904209,
+ "learning_rate": 0.0006,
+ "loss": 4.463746070861816,
+ "step": 4576
+ },
+ {
+ "epoch": 63.57292576419214,
+ "grad_norm": 0.01844949834048748,
+ "learning_rate": 0.0006,
+ "loss": 4.400300025939941,
+ "step": 4577
+ },
+ {
+ "epoch": 63.58689956331878,
+ "grad_norm": 0.015256480313837528,
+ "learning_rate": 0.0006,
+ "loss": 4.445844650268555,
+ "step": 4578
+ },
+ {
+ "epoch": 63.60087336244541,
+ "grad_norm": 0.015305760316550732,
+ "learning_rate": 0.0006,
+ "loss": 4.317327499389648,
+ "step": 4579
+ },
+ {
+ "epoch": 63.61484716157205,
+ "grad_norm": 0.015200769528746605,
+ "learning_rate": 0.0006,
+ "loss": 4.25112247467041,
+ "step": 4580
+ },
+ {
+ "epoch": 63.62882096069869,
+ "grad_norm": 0.015919538214802742,
+ "learning_rate": 0.0006,
+ "loss": 4.277937889099121,
+ "step": 4581
+ },
+ {
+ "epoch": 63.64279475982533,
+ "grad_norm": 0.01572665013372898,
+ "learning_rate": 0.0006,
+ "loss": 4.302237510681152,
+ "step": 4582
+ },
+ {
+ "epoch": 63.65676855895197,
+ "grad_norm": 0.016428731381893158,
+ "learning_rate": 0.0006,
+ "loss": 4.36762809753418,
+ "step": 4583
+ },
+ {
+ "epoch": 63.670742358078606,
+ "grad_norm": 0.01706082746386528,
+ "learning_rate": 0.0006,
+ "loss": 4.236526966094971,
+ "step": 4584
+ },
+ {
+ "epoch": 63.68471615720524,
+ "grad_norm": 0.017850136384367943,
+ "learning_rate": 0.0006,
+ "loss": 4.264129638671875,
+ "step": 4585
+ },
+ {
+ "epoch": 63.698689956331876,
+ "grad_norm": 0.016079438850283623,
+ "learning_rate": 0.0006,
+ "loss": 4.449799060821533,
+ "step": 4586
+ },
+ {
+ "epoch": 63.712663755458514,
+ "grad_norm": 0.014996383339166641,
+ "learning_rate": 0.0006,
+ "loss": 4.404574394226074,
+ "step": 4587
+ },
+ {
+ "epoch": 63.72663755458515,
+ "grad_norm": 0.015910431742668152,
+ "learning_rate": 0.0006,
+ "loss": 4.348101615905762,
+ "step": 4588
+ },
+ {
+ "epoch": 63.74061135371179,
+ "grad_norm": 0.01630021072924137,
+ "learning_rate": 0.0006,
+ "loss": 4.3193817138671875,
+ "step": 4589
+ },
+ {
+ "epoch": 63.75458515283843,
+ "grad_norm": 0.014571773819625378,
+ "learning_rate": 0.0006,
+ "loss": 4.374402046203613,
+ "step": 4590
+ },
+ {
+ "epoch": 63.76855895196506,
+ "grad_norm": 0.01662345975637436,
+ "learning_rate": 0.0006,
+ "loss": 4.329075813293457,
+ "step": 4591
+ },
+ {
+ "epoch": 63.7825327510917,
+ "grad_norm": 0.016604198142886162,
+ "learning_rate": 0.0006,
+ "loss": 4.436126232147217,
+ "step": 4592
+ },
+ {
+ "epoch": 63.79650655021834,
+ "grad_norm": 0.016578884795308113,
+ "learning_rate": 0.0006,
+ "loss": 4.361468315124512,
+ "step": 4593
+ },
+ {
+ "epoch": 63.81048034934498,
+ "grad_norm": 0.016339825466275215,
+ "learning_rate": 0.0006,
+ "loss": 4.449892044067383,
+ "step": 4594
+ },
+ {
+ "epoch": 63.82445414847162,
+ "grad_norm": 0.016805455088615417,
+ "learning_rate": 0.0006,
+ "loss": 4.415866851806641,
+ "step": 4595
+ },
+ {
+ "epoch": 63.838427947598255,
+ "grad_norm": 0.017366213724017143,
+ "learning_rate": 0.0006,
+ "loss": 4.361730575561523,
+ "step": 4596
+ },
+ {
+ "epoch": 63.852401746724894,
+ "grad_norm": 0.015599003992974758,
+ "learning_rate": 0.0006,
+ "loss": 4.363637924194336,
+ "step": 4597
+ },
+ {
+ "epoch": 63.866375545851525,
+ "grad_norm": 0.015674632042646408,
+ "learning_rate": 0.0006,
+ "loss": 4.332045078277588,
+ "step": 4598
+ },
+ {
+ "epoch": 63.880349344978164,
+ "grad_norm": 0.015350443311035633,
+ "learning_rate": 0.0006,
+ "loss": 4.320063591003418,
+ "step": 4599
+ },
+ {
+ "epoch": 63.8943231441048,
+ "grad_norm": 0.015388146974146366,
+ "learning_rate": 0.0006,
+ "loss": 4.398199081420898,
+ "step": 4600
+ },
+ {
+ "epoch": 63.90829694323144,
+ "grad_norm": 0.01450140681117773,
+ "learning_rate": 0.0006,
+ "loss": 4.244076728820801,
+ "step": 4601
+ },
+ {
+ "epoch": 63.92227074235808,
+ "grad_norm": 0.01637491025030613,
+ "learning_rate": 0.0006,
+ "loss": 4.213549613952637,
+ "step": 4602
+ },
+ {
+ "epoch": 63.93624454148472,
+ "grad_norm": 0.0159814041107893,
+ "learning_rate": 0.0006,
+ "loss": 4.421775817871094,
+ "step": 4603
+ },
+ {
+ "epoch": 63.95021834061135,
+ "grad_norm": 0.017386827617883682,
+ "learning_rate": 0.0006,
+ "loss": 4.340023994445801,
+ "step": 4604
+ },
+ {
+ "epoch": 63.96419213973799,
+ "grad_norm": 0.019679656252264977,
+ "learning_rate": 0.0006,
+ "loss": 4.352534294128418,
+ "step": 4605
+ },
+ {
+ "epoch": 63.97816593886463,
+ "grad_norm": 0.022391516715288162,
+ "learning_rate": 0.0006,
+ "loss": 4.3538312911987305,
+ "step": 4606
+ },
+ {
+ "epoch": 63.992139737991266,
+ "grad_norm": 0.023281006142497063,
+ "learning_rate": 0.0006,
+ "loss": 4.432089328765869,
+ "step": 4607
+ },
+ {
+ "epoch": 64.0,
+ "grad_norm": 0.019916288554668427,
+ "learning_rate": 0.0006,
+ "loss": 4.466937065124512,
+ "step": 4608
+ },
+ {
+ "epoch": 64.0,
+ "eval_loss": 4.659008502960205,
+ "eval_runtime": 56.0019,
+ "eval_samples_per_second": 43.606,
+ "eval_steps_per_second": 1.375,
+ "step": 4608
+ },
+ {
+ "epoch": 64.01397379912663,
+ "grad_norm": 0.017351187765598297,
+ "learning_rate": 0.0006,
+ "loss": 4.294129371643066,
+ "step": 4609
+ },
+ {
+ "epoch": 64.02794759825328,
+ "grad_norm": 0.02405865117907524,
+ "learning_rate": 0.0006,
+ "loss": 4.266717910766602,
+ "step": 4610
+ },
+ {
+ "epoch": 64.04192139737991,
+ "grad_norm": 0.022158386185765266,
+ "learning_rate": 0.0006,
+ "loss": 4.395932197570801,
+ "step": 4611
+ },
+ {
+ "epoch": 64.05589519650655,
+ "grad_norm": 0.019238505512475967,
+ "learning_rate": 0.0006,
+ "loss": 4.266704082489014,
+ "step": 4612
+ },
+ {
+ "epoch": 64.06986899563319,
+ "grad_norm": 0.021670425310730934,
+ "learning_rate": 0.0006,
+ "loss": 4.3160552978515625,
+ "step": 4613
+ },
+ {
+ "epoch": 64.08384279475983,
+ "grad_norm": 0.02076641097664833,
+ "learning_rate": 0.0006,
+ "loss": 4.415626525878906,
+ "step": 4614
+ },
+ {
+ "epoch": 64.09781659388646,
+ "grad_norm": 0.018223078921437263,
+ "learning_rate": 0.0006,
+ "loss": 4.292387008666992,
+ "step": 4615
+ },
+ {
+ "epoch": 64.1117903930131,
+ "grad_norm": 0.017401937395334244,
+ "learning_rate": 0.0006,
+ "loss": 4.424524307250977,
+ "step": 4616
+ },
+ {
+ "epoch": 64.12576419213974,
+ "grad_norm": 0.01663743332028389,
+ "learning_rate": 0.0006,
+ "loss": 4.350032806396484,
+ "step": 4617
+ },
+ {
+ "epoch": 64.13973799126637,
+ "grad_norm": 0.018716629594564438,
+ "learning_rate": 0.0006,
+ "loss": 4.309157371520996,
+ "step": 4618
+ },
+ {
+ "epoch": 64.15371179039302,
+ "grad_norm": 0.019569825381040573,
+ "learning_rate": 0.0006,
+ "loss": 4.358292579650879,
+ "step": 4619
+ },
+ {
+ "epoch": 64.16768558951965,
+ "grad_norm": 0.018994422629475594,
+ "learning_rate": 0.0006,
+ "loss": 4.3338823318481445,
+ "step": 4620
+ },
+ {
+ "epoch": 64.18165938864628,
+ "grad_norm": 0.017690688371658325,
+ "learning_rate": 0.0006,
+ "loss": 4.414468765258789,
+ "step": 4621
+ },
+ {
+ "epoch": 64.19563318777293,
+ "grad_norm": 0.0170609001070261,
+ "learning_rate": 0.0006,
+ "loss": 4.4337568283081055,
+ "step": 4622
+ },
+ {
+ "epoch": 64.20960698689956,
+ "grad_norm": 0.016984131187200546,
+ "learning_rate": 0.0006,
+ "loss": 4.304839134216309,
+ "step": 4623
+ },
+ {
+ "epoch": 64.2235807860262,
+ "grad_norm": 0.01619867794215679,
+ "learning_rate": 0.0006,
+ "loss": 4.389623165130615,
+ "step": 4624
+ },
+ {
+ "epoch": 64.23755458515284,
+ "grad_norm": 0.015428110957145691,
+ "learning_rate": 0.0006,
+ "loss": 4.277515888214111,
+ "step": 4625
+ },
+ {
+ "epoch": 64.25152838427948,
+ "grad_norm": 0.015909532085061073,
+ "learning_rate": 0.0006,
+ "loss": 4.266029357910156,
+ "step": 4626
+ },
+ {
+ "epoch": 64.26550218340611,
+ "grad_norm": 0.014296560548245907,
+ "learning_rate": 0.0006,
+ "loss": 4.286067485809326,
+ "step": 4627
+ },
+ {
+ "epoch": 64.27947598253274,
+ "grad_norm": 0.014065971598029137,
+ "learning_rate": 0.0006,
+ "loss": 4.353528022766113,
+ "step": 4628
+ },
+ {
+ "epoch": 64.29344978165939,
+ "grad_norm": 0.014386293478310108,
+ "learning_rate": 0.0006,
+ "loss": 4.318715572357178,
+ "step": 4629
+ },
+ {
+ "epoch": 64.30742358078602,
+ "grad_norm": 0.014001123607158661,
+ "learning_rate": 0.0006,
+ "loss": 4.362497329711914,
+ "step": 4630
+ },
+ {
+ "epoch": 64.32139737991267,
+ "grad_norm": 0.013408979400992393,
+ "learning_rate": 0.0006,
+ "loss": 4.417397499084473,
+ "step": 4631
+ },
+ {
+ "epoch": 64.3353711790393,
+ "grad_norm": 0.014960093423724174,
+ "learning_rate": 0.0006,
+ "loss": 4.359710693359375,
+ "step": 4632
+ },
+ {
+ "epoch": 64.34934497816595,
+ "grad_norm": 0.014658777043223381,
+ "learning_rate": 0.0006,
+ "loss": 4.318822383880615,
+ "step": 4633
+ },
+ {
+ "epoch": 64.36331877729258,
+ "grad_norm": 0.013766744174063206,
+ "learning_rate": 0.0006,
+ "loss": 4.265679836273193,
+ "step": 4634
+ },
+ {
+ "epoch": 64.37729257641921,
+ "grad_norm": 0.013821255415678024,
+ "learning_rate": 0.0006,
+ "loss": 4.3452911376953125,
+ "step": 4635
+ },
+ {
+ "epoch": 64.39126637554585,
+ "grad_norm": 0.013541066087782383,
+ "learning_rate": 0.0006,
+ "loss": 4.429254531860352,
+ "step": 4636
+ },
+ {
+ "epoch": 64.40524017467249,
+ "grad_norm": 0.013985645025968552,
+ "learning_rate": 0.0006,
+ "loss": 4.304452896118164,
+ "step": 4637
+ },
+ {
+ "epoch": 64.41921397379913,
+ "grad_norm": 0.015079808421432972,
+ "learning_rate": 0.0006,
+ "loss": 4.296828746795654,
+ "step": 4638
+ },
+ {
+ "epoch": 64.43318777292576,
+ "grad_norm": 0.014896944165229797,
+ "learning_rate": 0.0006,
+ "loss": 4.299156665802002,
+ "step": 4639
+ },
+ {
+ "epoch": 64.44716157205241,
+ "grad_norm": 0.013770041987299919,
+ "learning_rate": 0.0006,
+ "loss": 4.351439952850342,
+ "step": 4640
+ },
+ {
+ "epoch": 64.46113537117904,
+ "grad_norm": 0.01531178504228592,
+ "learning_rate": 0.0006,
+ "loss": 4.373715877532959,
+ "step": 4641
+ },
+ {
+ "epoch": 64.47510917030567,
+ "grad_norm": 0.014024450443685055,
+ "learning_rate": 0.0006,
+ "loss": 4.368053436279297,
+ "step": 4642
+ },
+ {
+ "epoch": 64.48908296943232,
+ "grad_norm": 0.013864024542272091,
+ "learning_rate": 0.0006,
+ "loss": 4.378564834594727,
+ "step": 4643
+ },
+ {
+ "epoch": 64.50305676855895,
+ "grad_norm": 0.013925445266067982,
+ "learning_rate": 0.0006,
+ "loss": 4.261663436889648,
+ "step": 4644
+ },
+ {
+ "epoch": 64.5170305676856,
+ "grad_norm": 0.015011338517069817,
+ "learning_rate": 0.0006,
+ "loss": 4.337285995483398,
+ "step": 4645
+ },
+ {
+ "epoch": 64.53100436681223,
+ "grad_norm": 0.016941986978054047,
+ "learning_rate": 0.0006,
+ "loss": 4.336877822875977,
+ "step": 4646
+ },
+ {
+ "epoch": 64.54497816593886,
+ "grad_norm": 0.01719909906387329,
+ "learning_rate": 0.0006,
+ "loss": 4.149813175201416,
+ "step": 4647
+ },
+ {
+ "epoch": 64.5589519650655,
+ "grad_norm": 0.01608402095735073,
+ "learning_rate": 0.0006,
+ "loss": 4.191280364990234,
+ "step": 4648
+ },
+ {
+ "epoch": 64.57292576419214,
+ "grad_norm": 0.015284820459783077,
+ "learning_rate": 0.0006,
+ "loss": 4.35953426361084,
+ "step": 4649
+ },
+ {
+ "epoch": 64.58689956331878,
+ "grad_norm": 0.014855682849884033,
+ "learning_rate": 0.0006,
+ "loss": 4.308401107788086,
+ "step": 4650
+ },
+ {
+ "epoch": 64.60087336244541,
+ "grad_norm": 0.017553111538290977,
+ "learning_rate": 0.0006,
+ "loss": 4.279720306396484,
+ "step": 4651
+ },
+ {
+ "epoch": 64.61484716157206,
+ "grad_norm": 0.019506998360157013,
+ "learning_rate": 0.0006,
+ "loss": 4.319524765014648,
+ "step": 4652
+ },
+ {
+ "epoch": 64.62882096069869,
+ "grad_norm": 0.020293425768613815,
+ "learning_rate": 0.0006,
+ "loss": 4.340523719787598,
+ "step": 4653
+ },
+ {
+ "epoch": 64.64279475982532,
+ "grad_norm": 0.01664556749165058,
+ "learning_rate": 0.0006,
+ "loss": 4.296785831451416,
+ "step": 4654
+ },
+ {
+ "epoch": 64.65676855895197,
+ "grad_norm": 0.012592585757374763,
+ "learning_rate": 0.0006,
+ "loss": 4.464595317840576,
+ "step": 4655
+ },
+ {
+ "epoch": 64.6707423580786,
+ "grad_norm": 0.01549589168280363,
+ "learning_rate": 0.0006,
+ "loss": 4.275053024291992,
+ "step": 4656
+ },
+ {
+ "epoch": 64.68471615720524,
+ "grad_norm": 0.01705172471702099,
+ "learning_rate": 0.0006,
+ "loss": 4.388431549072266,
+ "step": 4657
+ },
+ {
+ "epoch": 64.69868995633188,
+ "grad_norm": 0.015438582748174667,
+ "learning_rate": 0.0006,
+ "loss": 4.312070846557617,
+ "step": 4658
+ },
+ {
+ "epoch": 64.71266375545852,
+ "grad_norm": 0.014448083005845547,
+ "learning_rate": 0.0006,
+ "loss": 4.237599849700928,
+ "step": 4659
+ },
+ {
+ "epoch": 64.72663755458515,
+ "grad_norm": 0.015267071314156055,
+ "learning_rate": 0.0006,
+ "loss": 4.29893684387207,
+ "step": 4660
+ },
+ {
+ "epoch": 64.74061135371178,
+ "grad_norm": 0.016615847125649452,
+ "learning_rate": 0.0006,
+ "loss": 4.432273864746094,
+ "step": 4661
+ },
+ {
+ "epoch": 64.75458515283843,
+ "grad_norm": 0.016584362834692,
+ "learning_rate": 0.0006,
+ "loss": 4.202266216278076,
+ "step": 4662
+ },
+ {
+ "epoch": 64.76855895196506,
+ "grad_norm": 0.018313994631171227,
+ "learning_rate": 0.0006,
+ "loss": 4.367966651916504,
+ "step": 4663
+ },
+ {
+ "epoch": 64.78253275109171,
+ "grad_norm": 0.017946481704711914,
+ "learning_rate": 0.0006,
+ "loss": 4.205172538757324,
+ "step": 4664
+ },
+ {
+ "epoch": 64.79650655021834,
+ "grad_norm": 0.017744354903697968,
+ "learning_rate": 0.0006,
+ "loss": 4.295661926269531,
+ "step": 4665
+ },
+ {
+ "epoch": 64.81048034934499,
+ "grad_norm": 0.016103824600577354,
+ "learning_rate": 0.0006,
+ "loss": 4.276543617248535,
+ "step": 4666
+ },
+ {
+ "epoch": 64.82445414847162,
+ "grad_norm": 0.014590012840926647,
+ "learning_rate": 0.0006,
+ "loss": 4.414247035980225,
+ "step": 4667
+ },
+ {
+ "epoch": 64.83842794759825,
+ "grad_norm": 0.01782386749982834,
+ "learning_rate": 0.0006,
+ "loss": 4.203220367431641,
+ "step": 4668
+ },
+ {
+ "epoch": 64.8524017467249,
+ "grad_norm": 0.022122733294963837,
+ "learning_rate": 0.0006,
+ "loss": 4.38078498840332,
+ "step": 4669
+ },
+ {
+ "epoch": 64.86637554585153,
+ "grad_norm": 0.0250293780118227,
+ "learning_rate": 0.0006,
+ "loss": 4.3310136795043945,
+ "step": 4670
+ },
+ {
+ "epoch": 64.88034934497817,
+ "grad_norm": 0.022438062354922295,
+ "learning_rate": 0.0006,
+ "loss": 4.428631782531738,
+ "step": 4671
+ },
+ {
+ "epoch": 64.8943231441048,
+ "grad_norm": 0.018518192693591118,
+ "learning_rate": 0.0006,
+ "loss": 4.428250312805176,
+ "step": 4672
+ },
+ {
+ "epoch": 64.90829694323143,
+ "grad_norm": 0.018105225637555122,
+ "learning_rate": 0.0006,
+ "loss": 4.336209297180176,
+ "step": 4673
+ },
+ {
+ "epoch": 64.92227074235808,
+ "grad_norm": 0.017564399167895317,
+ "learning_rate": 0.0006,
+ "loss": 4.416011810302734,
+ "step": 4674
+ },
+ {
+ "epoch": 64.93624454148471,
+ "grad_norm": 0.018747031688690186,
+ "learning_rate": 0.0006,
+ "loss": 4.419938087463379,
+ "step": 4675
+ },
+ {
+ "epoch": 64.95021834061136,
+ "grad_norm": 0.018686484545469284,
+ "learning_rate": 0.0006,
+ "loss": 4.347185134887695,
+ "step": 4676
+ },
+ {
+ "epoch": 64.96419213973799,
+ "grad_norm": 0.019329898059368134,
+ "learning_rate": 0.0006,
+ "loss": 4.343749046325684,
+ "step": 4677
+ },
+ {
+ "epoch": 64.97816593886463,
+ "grad_norm": 0.018517782911658287,
+ "learning_rate": 0.0006,
+ "loss": 4.264604568481445,
+ "step": 4678
+ },
+ {
+ "epoch": 64.99213973799127,
+ "grad_norm": 0.016579125076532364,
+ "learning_rate": 0.0006,
+ "loss": 4.303423881530762,
+ "step": 4679
+ },
+ {
+ "epoch": 65.0,
+ "grad_norm": 0.01617567613720894,
+ "learning_rate": 0.0006,
+ "loss": 4.326556205749512,
+ "step": 4680
+ },
+ {
+ "epoch": 65.0,
+ "eval_loss": 4.655026912689209,
+ "eval_runtime": 57.1299,
+ "eval_samples_per_second": 42.745,
+ "eval_steps_per_second": 1.348,
+ "step": 4680
+ },
+ {
+ "epoch": 65.01397379912663,
+ "grad_norm": 0.01607249490916729,
+ "learning_rate": 0.0006,
+ "loss": 4.276607036590576,
+ "step": 4681
+ },
+ {
+ "epoch": 65.02794759825328,
+ "grad_norm": 0.016779497265815735,
+ "learning_rate": 0.0006,
+ "loss": 4.412464618682861,
+ "step": 4682
+ },
+ {
+ "epoch": 65.04192139737991,
+ "grad_norm": 0.018161555752158165,
+ "learning_rate": 0.0006,
+ "loss": 4.199329376220703,
+ "step": 4683
+ },
+ {
+ "epoch": 65.05589519650655,
+ "grad_norm": 0.017508309334516525,
+ "learning_rate": 0.0006,
+ "loss": 4.217567443847656,
+ "step": 4684
+ },
+ {
+ "epoch": 65.06986899563319,
+ "grad_norm": 0.015716979280114174,
+ "learning_rate": 0.0006,
+ "loss": 4.322023391723633,
+ "step": 4685
+ },
+ {
+ "epoch": 65.08384279475983,
+ "grad_norm": 0.015113635919988155,
+ "learning_rate": 0.0006,
+ "loss": 4.333515167236328,
+ "step": 4686
+ },
+ {
+ "epoch": 65.09781659388646,
+ "grad_norm": 0.01437560748308897,
+ "learning_rate": 0.0006,
+ "loss": 4.295609474182129,
+ "step": 4687
+ },
+ {
+ "epoch": 65.1117903930131,
+ "grad_norm": 0.013074862770736217,
+ "learning_rate": 0.0006,
+ "loss": 4.314120769500732,
+ "step": 4688
+ },
+ {
+ "epoch": 65.12576419213974,
+ "grad_norm": 0.012908284552395344,
+ "learning_rate": 0.0006,
+ "loss": 4.216890811920166,
+ "step": 4689
+ },
+ {
+ "epoch": 65.13973799126637,
+ "grad_norm": 0.013001754879951477,
+ "learning_rate": 0.0006,
+ "loss": 4.287717342376709,
+ "step": 4690
+ },
+ {
+ "epoch": 65.15371179039302,
+ "grad_norm": 0.013694263994693756,
+ "learning_rate": 0.0006,
+ "loss": 4.333837985992432,
+ "step": 4691
+ },
+ {
+ "epoch": 65.16768558951965,
+ "grad_norm": 0.014461833983659744,
+ "learning_rate": 0.0006,
+ "loss": 4.323545455932617,
+ "step": 4692
+ },
+ {
+ "epoch": 65.18165938864628,
+ "grad_norm": 0.013670675456523895,
+ "learning_rate": 0.0006,
+ "loss": 4.3313493728637695,
+ "step": 4693
+ },
+ {
+ "epoch": 65.19563318777293,
+ "grad_norm": 0.013094350695610046,
+ "learning_rate": 0.0006,
+ "loss": 4.23858118057251,
+ "step": 4694
+ },
+ {
+ "epoch": 65.20960698689956,
+ "grad_norm": 0.013996944762766361,
+ "learning_rate": 0.0006,
+ "loss": 4.428705215454102,
+ "step": 4695
+ },
+ {
+ "epoch": 65.2235807860262,
+ "grad_norm": 0.013295048847794533,
+ "learning_rate": 0.0006,
+ "loss": 4.238950252532959,
+ "step": 4696
+ },
+ {
+ "epoch": 65.23755458515284,
+ "grad_norm": 0.013380616903305054,
+ "learning_rate": 0.0006,
+ "loss": 4.273185729980469,
+ "step": 4697
+ },
+ {
+ "epoch": 65.25152838427948,
+ "grad_norm": 0.015062890015542507,
+ "learning_rate": 0.0006,
+ "loss": 4.29472017288208,
+ "step": 4698
+ },
+ {
+ "epoch": 65.26550218340611,
+ "grad_norm": 0.017317263409495354,
+ "learning_rate": 0.0006,
+ "loss": 4.337954998016357,
+ "step": 4699
+ },
+ {
+ "epoch": 65.27947598253274,
+ "grad_norm": 0.017698992043733597,
+ "learning_rate": 0.0006,
+ "loss": 4.336437225341797,
+ "step": 4700
+ },
+ {
+ "epoch": 65.29344978165939,
+ "grad_norm": 0.01628575474023819,
+ "learning_rate": 0.0006,
+ "loss": 4.302372455596924,
+ "step": 4701
+ },
+ {
+ "epoch": 65.30742358078602,
+ "grad_norm": 0.014531854540109634,
+ "learning_rate": 0.0006,
+ "loss": 4.256450653076172,
+ "step": 4702
+ },
+ {
+ "epoch": 65.32139737991267,
+ "grad_norm": 0.014549105428159237,
+ "learning_rate": 0.0006,
+ "loss": 4.302240371704102,
+ "step": 4703
+ },
+ {
+ "epoch": 65.3353711790393,
+ "grad_norm": 0.014881663955748081,
+ "learning_rate": 0.0006,
+ "loss": 4.3316121101379395,
+ "step": 4704
+ },
+ {
+ "epoch": 65.34934497816595,
+ "grad_norm": 0.014513437636196613,
+ "learning_rate": 0.0006,
+ "loss": 4.441212177276611,
+ "step": 4705
+ },
+ {
+ "epoch": 65.36331877729258,
+ "grad_norm": 0.014122087508440018,
+ "learning_rate": 0.0006,
+ "loss": 4.343296527862549,
+ "step": 4706
+ },
+ {
+ "epoch": 65.37729257641921,
+ "grad_norm": 0.014817348681390285,
+ "learning_rate": 0.0006,
+ "loss": 4.195674896240234,
+ "step": 4707
+ },
+ {
+ "epoch": 65.39126637554585,
+ "grad_norm": 0.01787056028842926,
+ "learning_rate": 0.0006,
+ "loss": 4.3507914543151855,
+ "step": 4708
+ },
+ {
+ "epoch": 65.40524017467249,
+ "grad_norm": 0.015295136719942093,
+ "learning_rate": 0.0006,
+ "loss": 4.234372138977051,
+ "step": 4709
+ },
+ {
+ "epoch": 65.41921397379913,
+ "grad_norm": 0.014324234798550606,
+ "learning_rate": 0.0006,
+ "loss": 4.2907867431640625,
+ "step": 4710
+ },
+ {
+ "epoch": 65.43318777292576,
+ "grad_norm": 0.016090670600533485,
+ "learning_rate": 0.0006,
+ "loss": 4.177244186401367,
+ "step": 4711
+ },
+ {
+ "epoch": 65.44716157205241,
+ "grad_norm": 0.0174604132771492,
+ "learning_rate": 0.0006,
+ "loss": 4.335577964782715,
+ "step": 4712
+ },
+ {
+ "epoch": 65.46113537117904,
+ "grad_norm": 0.016731899231672287,
+ "learning_rate": 0.0006,
+ "loss": 4.30197811126709,
+ "step": 4713
+ },
+ {
+ "epoch": 65.47510917030567,
+ "grad_norm": 0.017423735931515694,
+ "learning_rate": 0.0006,
+ "loss": 4.3528852462768555,
+ "step": 4714
+ },
+ {
+ "epoch": 65.48908296943232,
+ "grad_norm": 0.01780976541340351,
+ "learning_rate": 0.0006,
+ "loss": 4.185036659240723,
+ "step": 4715
+ },
+ {
+ "epoch": 65.50305676855895,
+ "grad_norm": 0.019977200776338577,
+ "learning_rate": 0.0006,
+ "loss": 4.328766345977783,
+ "step": 4716
+ },
+ {
+ "epoch": 65.5170305676856,
+ "grad_norm": 0.020284287631511688,
+ "learning_rate": 0.0006,
+ "loss": 4.285494804382324,
+ "step": 4717
+ },
+ {
+ "epoch": 65.53100436681223,
+ "grad_norm": 0.019168002530932426,
+ "learning_rate": 0.0006,
+ "loss": 4.328560829162598,
+ "step": 4718
+ },
+ {
+ "epoch": 65.54497816593886,
+ "grad_norm": 0.019217776134610176,
+ "learning_rate": 0.0006,
+ "loss": 4.189998626708984,
+ "step": 4719
+ },
+ {
+ "epoch": 65.5589519650655,
+ "grad_norm": 0.018960488960146904,
+ "learning_rate": 0.0006,
+ "loss": 4.4995012283325195,
+ "step": 4720
+ },
+ {
+ "epoch": 65.57292576419214,
+ "grad_norm": 0.01941288262605667,
+ "learning_rate": 0.0006,
+ "loss": 4.26570987701416,
+ "step": 4721
+ },
+ {
+ "epoch": 65.58689956331878,
+ "grad_norm": 0.020263448357582092,
+ "learning_rate": 0.0006,
+ "loss": 4.411645889282227,
+ "step": 4722
+ },
+ {
+ "epoch": 65.60087336244541,
+ "grad_norm": 0.02144799567759037,
+ "learning_rate": 0.0006,
+ "loss": 4.163675308227539,
+ "step": 4723
+ },
+ {
+ "epoch": 65.61484716157206,
+ "grad_norm": 0.019790446385741234,
+ "learning_rate": 0.0006,
+ "loss": 4.2361063957214355,
+ "step": 4724
+ },
+ {
+ "epoch": 65.62882096069869,
+ "grad_norm": 0.01829805225133896,
+ "learning_rate": 0.0006,
+ "loss": 4.449810028076172,
+ "step": 4725
+ },
+ {
+ "epoch": 65.64279475982532,
+ "grad_norm": 0.01571105420589447,
+ "learning_rate": 0.0006,
+ "loss": 4.281406879425049,
+ "step": 4726
+ },
+ {
+ "epoch": 65.65676855895197,
+ "grad_norm": 0.016477147117257118,
+ "learning_rate": 0.0006,
+ "loss": 4.293408393859863,
+ "step": 4727
+ },
+ {
+ "epoch": 65.6707423580786,
+ "grad_norm": 0.016766032204031944,
+ "learning_rate": 0.0006,
+ "loss": 4.378252983093262,
+ "step": 4728
+ },
+ {
+ "epoch": 65.68471615720524,
+ "grad_norm": 0.0159933939576149,
+ "learning_rate": 0.0006,
+ "loss": 4.3065667152404785,
+ "step": 4729
+ },
+ {
+ "epoch": 65.69868995633188,
+ "grad_norm": 0.01658209227025509,
+ "learning_rate": 0.0006,
+ "loss": 4.174026966094971,
+ "step": 4730
+ },
+ {
+ "epoch": 65.71266375545852,
+ "grad_norm": 0.015081380493938923,
+ "learning_rate": 0.0006,
+ "loss": 4.360222339630127,
+ "step": 4731
+ },
+ {
+ "epoch": 65.72663755458515,
+ "grad_norm": 0.015110890381038189,
+ "learning_rate": 0.0006,
+ "loss": 4.218810558319092,
+ "step": 4732
+ },
+ {
+ "epoch": 65.74061135371178,
+ "grad_norm": 0.015300953760743141,
+ "learning_rate": 0.0006,
+ "loss": 4.331979274749756,
+ "step": 4733
+ },
+ {
+ "epoch": 65.75458515283843,
+ "grad_norm": 0.01715504191815853,
+ "learning_rate": 0.0006,
+ "loss": 4.349112033843994,
+ "step": 4734
+ },
+ {
+ "epoch": 65.76855895196506,
+ "grad_norm": 0.015870366245508194,
+ "learning_rate": 0.0006,
+ "loss": 4.329850196838379,
+ "step": 4735
+ },
+ {
+ "epoch": 65.78253275109171,
+ "grad_norm": 0.01541937980800867,
+ "learning_rate": 0.0006,
+ "loss": 4.3400163650512695,
+ "step": 4736
+ },
+ {
+ "epoch": 65.79650655021834,
+ "grad_norm": 0.01738942787051201,
+ "learning_rate": 0.0006,
+ "loss": 4.220798492431641,
+ "step": 4737
+ },
+ {
+ "epoch": 65.81048034934499,
+ "grad_norm": 0.01685990020632744,
+ "learning_rate": 0.0006,
+ "loss": 4.177238464355469,
+ "step": 4738
+ },
+ {
+ "epoch": 65.82445414847162,
+ "grad_norm": 0.016183465719223022,
+ "learning_rate": 0.0006,
+ "loss": 4.327216625213623,
+ "step": 4739
+ },
+ {
+ "epoch": 65.83842794759825,
+ "grad_norm": 0.01677694357931614,
+ "learning_rate": 0.0006,
+ "loss": 4.27766227722168,
+ "step": 4740
+ },
+ {
+ "epoch": 65.8524017467249,
+ "grad_norm": 0.01645066775381565,
+ "learning_rate": 0.0006,
+ "loss": 4.288347244262695,
+ "step": 4741
+ },
+ {
+ "epoch": 65.86637554585153,
+ "grad_norm": 0.018034635111689568,
+ "learning_rate": 0.0006,
+ "loss": 4.292458534240723,
+ "step": 4742
+ },
+ {
+ "epoch": 65.88034934497817,
+ "grad_norm": 0.017973346635699272,
+ "learning_rate": 0.0006,
+ "loss": 4.3117265701293945,
+ "step": 4743
+ },
+ {
+ "epoch": 65.8943231441048,
+ "grad_norm": 0.016605669632554054,
+ "learning_rate": 0.0006,
+ "loss": 4.280615329742432,
+ "step": 4744
+ },
+ {
+ "epoch": 65.90829694323143,
+ "grad_norm": 0.016298074275255203,
+ "learning_rate": 0.0006,
+ "loss": 4.301933288574219,
+ "step": 4745
+ },
+ {
+ "epoch": 65.92227074235808,
+ "grad_norm": 0.015383216552436352,
+ "learning_rate": 0.0006,
+ "loss": 4.352383136749268,
+ "step": 4746
+ },
+ {
+ "epoch": 65.93624454148471,
+ "grad_norm": 0.016222182661294937,
+ "learning_rate": 0.0006,
+ "loss": 4.305725574493408,
+ "step": 4747
+ },
+ {
+ "epoch": 65.95021834061136,
+ "grad_norm": 0.01691329851746559,
+ "learning_rate": 0.0006,
+ "loss": 4.29502010345459,
+ "step": 4748
+ },
+ {
+ "epoch": 65.96419213973799,
+ "grad_norm": 0.017637677490711212,
+ "learning_rate": 0.0006,
+ "loss": 4.3133039474487305,
+ "step": 4749
+ },
+ {
+ "epoch": 65.97816593886463,
+ "grad_norm": 0.021362047642469406,
+ "learning_rate": 0.0006,
+ "loss": 4.458834648132324,
+ "step": 4750
+ },
+ {
+ "epoch": 65.99213973799127,
+ "grad_norm": 0.023553011938929558,
+ "learning_rate": 0.0006,
+ "loss": 4.298727989196777,
+ "step": 4751
+ },
+ {
+ "epoch": 66.0,
+ "grad_norm": 0.02154950052499771,
+ "learning_rate": 0.0006,
+ "loss": 4.403392791748047,
+ "step": 4752
+ },
+ {
+ "epoch": 66.0,
+ "eval_loss": 4.724340438842773,
+ "eval_runtime": 56.9555,
+ "eval_samples_per_second": 42.876,
+ "eval_steps_per_second": 1.352,
+ "step": 4752
+ },
+ {
+ "epoch": 66.01397379912663,
+ "grad_norm": 0.01916784606873989,
+ "learning_rate": 0.0006,
+ "loss": 4.22931432723999,
+ "step": 4753
+ },
+ {
+ "epoch": 66.02794759825328,
+ "grad_norm": 0.017492054030299187,
+ "learning_rate": 0.0006,
+ "loss": 4.271522521972656,
+ "step": 4754
+ },
+ {
+ "epoch": 66.04192139737991,
+ "grad_norm": 0.018408095464110374,
+ "learning_rate": 0.0006,
+ "loss": 4.337705135345459,
+ "step": 4755
+ },
+ {
+ "epoch": 66.05589519650655,
+ "grad_norm": 0.019220108166337013,
+ "learning_rate": 0.0006,
+ "loss": 4.280340194702148,
+ "step": 4756
+ },
+ {
+ "epoch": 66.06986899563319,
+ "grad_norm": 0.01773735322058201,
+ "learning_rate": 0.0006,
+ "loss": 4.390201568603516,
+ "step": 4757
+ },
+ {
+ "epoch": 66.08384279475983,
+ "grad_norm": 0.01679125428199768,
+ "learning_rate": 0.0006,
+ "loss": 4.299282073974609,
+ "step": 4758
+ },
+ {
+ "epoch": 66.09781659388646,
+ "grad_norm": 0.01479059923440218,
+ "learning_rate": 0.0006,
+ "loss": 4.472630500793457,
+ "step": 4759
+ },
+ {
+ "epoch": 66.1117903930131,
+ "grad_norm": 0.015130104497075081,
+ "learning_rate": 0.0006,
+ "loss": 4.301194190979004,
+ "step": 4760
+ },
+ {
+ "epoch": 66.12576419213974,
+ "grad_norm": 0.0158979594707489,
+ "learning_rate": 0.0006,
+ "loss": 4.375751972198486,
+ "step": 4761
+ },
+ {
+ "epoch": 66.13973799126637,
+ "grad_norm": 0.014432408846914768,
+ "learning_rate": 0.0006,
+ "loss": 4.279788017272949,
+ "step": 4762
+ },
+ {
+ "epoch": 66.15371179039302,
+ "grad_norm": 0.013026784174144268,
+ "learning_rate": 0.0006,
+ "loss": 4.255035400390625,
+ "step": 4763
+ },
+ {
+ "epoch": 66.16768558951965,
+ "grad_norm": 0.014422653242945671,
+ "learning_rate": 0.0006,
+ "loss": 4.261832237243652,
+ "step": 4764
+ },
+ {
+ "epoch": 66.18165938864628,
+ "grad_norm": 0.014613019302487373,
+ "learning_rate": 0.0006,
+ "loss": 4.382184028625488,
+ "step": 4765
+ },
+ {
+ "epoch": 66.19563318777293,
+ "grad_norm": 0.014989894814789295,
+ "learning_rate": 0.0006,
+ "loss": 4.40528678894043,
+ "step": 4766
+ },
+ {
+ "epoch": 66.20960698689956,
+ "grad_norm": 0.018190350383520126,
+ "learning_rate": 0.0006,
+ "loss": 4.302414894104004,
+ "step": 4767
+ },
+ {
+ "epoch": 66.2235807860262,
+ "grad_norm": 0.020887890830636024,
+ "learning_rate": 0.0006,
+ "loss": 4.289920806884766,
+ "step": 4768
+ },
+ {
+ "epoch": 66.23755458515284,
+ "grad_norm": 0.02376876398921013,
+ "learning_rate": 0.0006,
+ "loss": 4.297855377197266,
+ "step": 4769
+ },
+ {
+ "epoch": 66.25152838427948,
+ "grad_norm": 0.02152269519865513,
+ "learning_rate": 0.0006,
+ "loss": 4.1972832679748535,
+ "step": 4770
+ },
+ {
+ "epoch": 66.26550218340611,
+ "grad_norm": 0.015572802163660526,
+ "learning_rate": 0.0006,
+ "loss": 4.307487487792969,
+ "step": 4771
+ },
+ {
+ "epoch": 66.27947598253274,
+ "grad_norm": 0.015228749252855778,
+ "learning_rate": 0.0006,
+ "loss": 4.103250503540039,
+ "step": 4772
+ },
+ {
+ "epoch": 66.29344978165939,
+ "grad_norm": 0.016925504431128502,
+ "learning_rate": 0.0006,
+ "loss": 4.247461318969727,
+ "step": 4773
+ },
+ {
+ "epoch": 66.30742358078602,
+ "grad_norm": 0.01613711006939411,
+ "learning_rate": 0.0006,
+ "loss": 4.4699273109436035,
+ "step": 4774
+ },
+ {
+ "epoch": 66.32139737991267,
+ "grad_norm": 0.018332140520215034,
+ "learning_rate": 0.0006,
+ "loss": 4.22969388961792,
+ "step": 4775
+ },
+ {
+ "epoch": 66.3353711790393,
+ "grad_norm": 0.019575074315071106,
+ "learning_rate": 0.0006,
+ "loss": 4.29397439956665,
+ "step": 4776
+ },
+ {
+ "epoch": 66.34934497816595,
+ "grad_norm": 0.018177419900894165,
+ "learning_rate": 0.0006,
+ "loss": 4.221899509429932,
+ "step": 4777
+ },
+ {
+ "epoch": 66.36331877729258,
+ "grad_norm": 0.014292889274656773,
+ "learning_rate": 0.0006,
+ "loss": 4.325457572937012,
+ "step": 4778
+ },
+ {
+ "epoch": 66.37729257641921,
+ "grad_norm": 0.015098796226084232,
+ "learning_rate": 0.0006,
+ "loss": 4.364837169647217,
+ "step": 4779
+ },
+ {
+ "epoch": 66.39126637554585,
+ "grad_norm": 0.013992635533213615,
+ "learning_rate": 0.0006,
+ "loss": 4.299082279205322,
+ "step": 4780
+ },
+ {
+ "epoch": 66.40524017467249,
+ "grad_norm": 0.01504511572420597,
+ "learning_rate": 0.0006,
+ "loss": 4.230877876281738,
+ "step": 4781
+ },
+ {
+ "epoch": 66.41921397379913,
+ "grad_norm": 0.014875639230012894,
+ "learning_rate": 0.0006,
+ "loss": 4.368405818939209,
+ "step": 4782
+ },
+ {
+ "epoch": 66.43318777292576,
+ "grad_norm": 0.014306887984275818,
+ "learning_rate": 0.0006,
+ "loss": 4.389184474945068,
+ "step": 4783
+ },
+ {
+ "epoch": 66.44716157205241,
+ "grad_norm": 0.012959563173353672,
+ "learning_rate": 0.0006,
+ "loss": 4.37934684753418,
+ "step": 4784
+ },
+ {
+ "epoch": 66.46113537117904,
+ "grad_norm": 0.013206703588366508,
+ "learning_rate": 0.0006,
+ "loss": 4.442799091339111,
+ "step": 4785
+ },
+ {
+ "epoch": 66.47510917030567,
+ "grad_norm": 0.014848909340798855,
+ "learning_rate": 0.0006,
+ "loss": 4.203481674194336,
+ "step": 4786
+ },
+ {
+ "epoch": 66.48908296943232,
+ "grad_norm": 0.015637682750821114,
+ "learning_rate": 0.0006,
+ "loss": 4.27552604675293,
+ "step": 4787
+ },
+ {
+ "epoch": 66.50305676855895,
+ "grad_norm": 0.016017816960811615,
+ "learning_rate": 0.0006,
+ "loss": 4.296402931213379,
+ "step": 4788
+ },
+ {
+ "epoch": 66.5170305676856,
+ "grad_norm": 0.01577814109623432,
+ "learning_rate": 0.0006,
+ "loss": 4.301733016967773,
+ "step": 4789
+ },
+ {
+ "epoch": 66.53100436681223,
+ "grad_norm": 0.015293175354599953,
+ "learning_rate": 0.0006,
+ "loss": 4.369877815246582,
+ "step": 4790
+ },
+ {
+ "epoch": 66.54497816593886,
+ "grad_norm": 0.014262278564274311,
+ "learning_rate": 0.0006,
+ "loss": 4.3314738273620605,
+ "step": 4791
+ },
+ {
+ "epoch": 66.5589519650655,
+ "grad_norm": 0.015123482793569565,
+ "learning_rate": 0.0006,
+ "loss": 4.313591957092285,
+ "step": 4792
+ },
+ {
+ "epoch": 66.57292576419214,
+ "grad_norm": 0.015458766371011734,
+ "learning_rate": 0.0006,
+ "loss": 4.3045854568481445,
+ "step": 4793
+ },
+ {
+ "epoch": 66.58689956331878,
+ "grad_norm": 0.017217233777046204,
+ "learning_rate": 0.0006,
+ "loss": 4.269944190979004,
+ "step": 4794
+ },
+ {
+ "epoch": 66.60087336244541,
+ "grad_norm": 0.01962282694876194,
+ "learning_rate": 0.0006,
+ "loss": 4.271735191345215,
+ "step": 4795
+ },
+ {
+ "epoch": 66.61484716157206,
+ "grad_norm": 0.01917066052556038,
+ "learning_rate": 0.0006,
+ "loss": 4.288168430328369,
+ "step": 4796
+ },
+ {
+ "epoch": 66.62882096069869,
+ "grad_norm": 0.018211524933576584,
+ "learning_rate": 0.0006,
+ "loss": 4.280660629272461,
+ "step": 4797
+ },
+ {
+ "epoch": 66.64279475982532,
+ "grad_norm": 0.01936524175107479,
+ "learning_rate": 0.0006,
+ "loss": 4.3710551261901855,
+ "step": 4798
+ },
+ {
+ "epoch": 66.65676855895197,
+ "grad_norm": 0.01882351003587246,
+ "learning_rate": 0.0006,
+ "loss": 4.342750549316406,
+ "step": 4799
+ },
+ {
+ "epoch": 66.6707423580786,
+ "grad_norm": 0.017149122431874275,
+ "learning_rate": 0.0006,
+ "loss": 4.32134485244751,
+ "step": 4800
+ },
+ {
+ "epoch": 66.68471615720524,
+ "grad_norm": 0.01699327677488327,
+ "learning_rate": 0.0006,
+ "loss": 4.248893737792969,
+ "step": 4801
+ },
+ {
+ "epoch": 66.69868995633188,
+ "grad_norm": 0.016873255372047424,
+ "learning_rate": 0.0006,
+ "loss": 4.277350425720215,
+ "step": 4802
+ },
+ {
+ "epoch": 66.71266375545852,
+ "grad_norm": 0.016592837870121002,
+ "learning_rate": 0.0006,
+ "loss": 4.221665859222412,
+ "step": 4803
+ },
+ {
+ "epoch": 66.72663755458515,
+ "grad_norm": 0.015887131914496422,
+ "learning_rate": 0.0006,
+ "loss": 4.328040599822998,
+ "step": 4804
+ },
+ {
+ "epoch": 66.74061135371178,
+ "grad_norm": 0.014057286083698273,
+ "learning_rate": 0.0006,
+ "loss": 4.3294548988342285,
+ "step": 4805
+ },
+ {
+ "epoch": 66.75458515283843,
+ "grad_norm": 0.013135905377566814,
+ "learning_rate": 0.0006,
+ "loss": 4.254383563995361,
+ "step": 4806
+ },
+ {
+ "epoch": 66.76855895196506,
+ "grad_norm": 0.013632557354867458,
+ "learning_rate": 0.0006,
+ "loss": 4.342815399169922,
+ "step": 4807
+ },
+ {
+ "epoch": 66.78253275109171,
+ "grad_norm": 0.015512688085436821,
+ "learning_rate": 0.0006,
+ "loss": 4.297835350036621,
+ "step": 4808
+ },
+ {
+ "epoch": 66.79650655021834,
+ "grad_norm": 0.01621832884848118,
+ "learning_rate": 0.0006,
+ "loss": 4.314773082733154,
+ "step": 4809
+ },
+ {
+ "epoch": 66.81048034934499,
+ "grad_norm": 0.016196032986044884,
+ "learning_rate": 0.0006,
+ "loss": 4.29209566116333,
+ "step": 4810
+ },
+ {
+ "epoch": 66.82445414847162,
+ "grad_norm": 0.015463353134691715,
+ "learning_rate": 0.0006,
+ "loss": 4.262795448303223,
+ "step": 4811
+ },
+ {
+ "epoch": 66.83842794759825,
+ "grad_norm": 0.016143787652254105,
+ "learning_rate": 0.0006,
+ "loss": 4.274284362792969,
+ "step": 4812
+ },
+ {
+ "epoch": 66.8524017467249,
+ "grad_norm": 0.01610061153769493,
+ "learning_rate": 0.0006,
+ "loss": 4.337160587310791,
+ "step": 4813
+ },
+ {
+ "epoch": 66.86637554585153,
+ "grad_norm": 0.01582838036119938,
+ "learning_rate": 0.0006,
+ "loss": 4.192723274230957,
+ "step": 4814
+ },
+ {
+ "epoch": 66.88034934497817,
+ "grad_norm": 0.015081540681421757,
+ "learning_rate": 0.0006,
+ "loss": 4.370708465576172,
+ "step": 4815
+ },
+ {
+ "epoch": 66.8943231441048,
+ "grad_norm": 0.015929723158478737,
+ "learning_rate": 0.0006,
+ "loss": 4.340244293212891,
+ "step": 4816
+ },
+ {
+ "epoch": 66.90829694323143,
+ "grad_norm": 0.01784711331129074,
+ "learning_rate": 0.0006,
+ "loss": 4.260406970977783,
+ "step": 4817
+ },
+ {
+ "epoch": 66.92227074235808,
+ "grad_norm": 0.017434703186154366,
+ "learning_rate": 0.0006,
+ "loss": 4.286900043487549,
+ "step": 4818
+ },
+ {
+ "epoch": 66.93624454148471,
+ "grad_norm": 0.016262540593743324,
+ "learning_rate": 0.0006,
+ "loss": 4.292142391204834,
+ "step": 4819
+ },
+ {
+ "epoch": 66.95021834061136,
+ "grad_norm": 0.016122829169034958,
+ "learning_rate": 0.0006,
+ "loss": 4.291485786437988,
+ "step": 4820
+ },
+ {
+ "epoch": 66.96419213973799,
+ "grad_norm": 0.014361178502440453,
+ "learning_rate": 0.0006,
+ "loss": 4.275425910949707,
+ "step": 4821
+ },
+ {
+ "epoch": 66.97816593886463,
+ "grad_norm": 0.01373987551778555,
+ "learning_rate": 0.0006,
+ "loss": 4.223120212554932,
+ "step": 4822
+ },
+ {
+ "epoch": 66.99213973799127,
+ "grad_norm": 0.015687376260757446,
+ "learning_rate": 0.0006,
+ "loss": 4.355381011962891,
+ "step": 4823
+ },
+ {
+ "epoch": 67.0,
+ "grad_norm": 0.01998465321958065,
+ "learning_rate": 0.0006,
+ "loss": 4.264053821563721,
+ "step": 4824
+ },
+ {
+ "epoch": 67.0,
+ "eval_loss": 4.662398338317871,
+ "eval_runtime": 56.7928,
+ "eval_samples_per_second": 42.998,
+ "eval_steps_per_second": 1.356,
+ "step": 4824
+ },
+ {
+ "epoch": 67.01397379912663,
+ "grad_norm": 0.021736836060881615,
+ "learning_rate": 0.0006,
+ "loss": 4.243354797363281,
+ "step": 4825
+ },
+ {
+ "epoch": 67.02794759825328,
+ "grad_norm": 0.021476037800312042,
+ "learning_rate": 0.0006,
+ "loss": 4.218207836151123,
+ "step": 4826
+ },
+ {
+ "epoch": 67.04192139737991,
+ "grad_norm": 0.017188388854265213,
+ "learning_rate": 0.0006,
+ "loss": 4.18048095703125,
+ "step": 4827
+ },
+ {
+ "epoch": 67.05589519650655,
+ "grad_norm": 0.017293719574809074,
+ "learning_rate": 0.0006,
+ "loss": 4.239145278930664,
+ "step": 4828
+ },
+ {
+ "epoch": 67.06986899563319,
+ "grad_norm": 0.017383819445967674,
+ "learning_rate": 0.0006,
+ "loss": 4.114765167236328,
+ "step": 4829
+ },
+ {
+ "epoch": 67.08384279475983,
+ "grad_norm": 0.01651529036462307,
+ "learning_rate": 0.0006,
+ "loss": 4.351023197174072,
+ "step": 4830
+ },
+ {
+ "epoch": 67.09781659388646,
+ "grad_norm": 0.015893027186393738,
+ "learning_rate": 0.0006,
+ "loss": 4.261432647705078,
+ "step": 4831
+ },
+ {
+ "epoch": 67.1117903930131,
+ "grad_norm": 0.013381517492234707,
+ "learning_rate": 0.0006,
+ "loss": 4.172064781188965,
+ "step": 4832
+ },
+ {
+ "epoch": 67.12576419213974,
+ "grad_norm": 0.014642654918134212,
+ "learning_rate": 0.0006,
+ "loss": 4.356171131134033,
+ "step": 4833
+ },
+ {
+ "epoch": 67.13973799126637,
+ "grad_norm": 0.01425282284617424,
+ "learning_rate": 0.0006,
+ "loss": 4.303495407104492,
+ "step": 4834
+ },
+ {
+ "epoch": 67.15371179039302,
+ "grad_norm": 0.014574599452316761,
+ "learning_rate": 0.0006,
+ "loss": 4.2850847244262695,
+ "step": 4835
+ },
+ {
+ "epoch": 67.16768558951965,
+ "grad_norm": 0.012676282785832882,
+ "learning_rate": 0.0006,
+ "loss": 4.297845840454102,
+ "step": 4836
+ },
+ {
+ "epoch": 67.18165938864628,
+ "grad_norm": 0.013948136940598488,
+ "learning_rate": 0.0006,
+ "loss": 4.324530601501465,
+ "step": 4837
+ },
+ {
+ "epoch": 67.19563318777293,
+ "grad_norm": 0.014195773750543594,
+ "learning_rate": 0.0006,
+ "loss": 4.31236457824707,
+ "step": 4838
+ },
+ {
+ "epoch": 67.20960698689956,
+ "grad_norm": 0.013631382025778294,
+ "learning_rate": 0.0006,
+ "loss": 4.086362838745117,
+ "step": 4839
+ },
+ {
+ "epoch": 67.2235807860262,
+ "grad_norm": 0.012798587791621685,
+ "learning_rate": 0.0006,
+ "loss": 4.328471660614014,
+ "step": 4840
+ },
+ {
+ "epoch": 67.23755458515284,
+ "grad_norm": 0.014443321153521538,
+ "learning_rate": 0.0006,
+ "loss": 4.278019905090332,
+ "step": 4841
+ },
+ {
+ "epoch": 67.25152838427948,
+ "grad_norm": 0.015355420298874378,
+ "learning_rate": 0.0006,
+ "loss": 4.333611011505127,
+ "step": 4842
+ },
+ {
+ "epoch": 67.26550218340611,
+ "grad_norm": 0.016487296670675278,
+ "learning_rate": 0.0006,
+ "loss": 4.351336479187012,
+ "step": 4843
+ },
+ {
+ "epoch": 67.27947598253274,
+ "grad_norm": 0.017422541975975037,
+ "learning_rate": 0.0006,
+ "loss": 4.386096477508545,
+ "step": 4844
+ },
+ {
+ "epoch": 67.29344978165939,
+ "grad_norm": 0.015790918841958046,
+ "learning_rate": 0.0006,
+ "loss": 4.301377773284912,
+ "step": 4845
+ },
+ {
+ "epoch": 67.30742358078602,
+ "grad_norm": 0.016427364200353622,
+ "learning_rate": 0.0006,
+ "loss": 4.2838945388793945,
+ "step": 4846
+ },
+ {
+ "epoch": 67.32139737991267,
+ "grad_norm": 0.017050761729478836,
+ "learning_rate": 0.0006,
+ "loss": 4.282417297363281,
+ "step": 4847
+ },
+ {
+ "epoch": 67.3353711790393,
+ "grad_norm": 0.015006253495812416,
+ "learning_rate": 0.0006,
+ "loss": 4.352236270904541,
+ "step": 4848
+ },
+ {
+ "epoch": 67.34934497816595,
+ "grad_norm": 0.012996627949178219,
+ "learning_rate": 0.0006,
+ "loss": 4.309242248535156,
+ "step": 4849
+ },
+ {
+ "epoch": 67.36331877729258,
+ "grad_norm": 0.01367615070194006,
+ "learning_rate": 0.0006,
+ "loss": 4.211801528930664,
+ "step": 4850
+ },
+ {
+ "epoch": 67.37729257641921,
+ "grad_norm": 0.014775963500142097,
+ "learning_rate": 0.0006,
+ "loss": 4.360859394073486,
+ "step": 4851
+ },
+ {
+ "epoch": 67.39126637554585,
+ "grad_norm": 0.015854304656386375,
+ "learning_rate": 0.0006,
+ "loss": 4.211830139160156,
+ "step": 4852
+ },
+ {
+ "epoch": 67.40524017467249,
+ "grad_norm": 0.016888894140720367,
+ "learning_rate": 0.0006,
+ "loss": 4.216549873352051,
+ "step": 4853
+ },
+ {
+ "epoch": 67.41921397379913,
+ "grad_norm": 0.0170894768089056,
+ "learning_rate": 0.0006,
+ "loss": 4.263486385345459,
+ "step": 4854
+ },
+ {
+ "epoch": 67.43318777292576,
+ "grad_norm": 0.015829868614673615,
+ "learning_rate": 0.0006,
+ "loss": 4.360196113586426,
+ "step": 4855
+ },
+ {
+ "epoch": 67.44716157205241,
+ "grad_norm": 0.015220202505588531,
+ "learning_rate": 0.0006,
+ "loss": 4.243794918060303,
+ "step": 4856
+ },
+ {
+ "epoch": 67.46113537117904,
+ "grad_norm": 0.01413388829678297,
+ "learning_rate": 0.0006,
+ "loss": 4.286624908447266,
+ "step": 4857
+ },
+ {
+ "epoch": 67.47510917030567,
+ "grad_norm": 0.01426626369357109,
+ "learning_rate": 0.0006,
+ "loss": 4.257316589355469,
+ "step": 4858
+ },
+ {
+ "epoch": 67.48908296943232,
+ "grad_norm": 0.015957845374941826,
+ "learning_rate": 0.0006,
+ "loss": 4.300443649291992,
+ "step": 4859
+ },
+ {
+ "epoch": 67.50305676855895,
+ "grad_norm": 0.015499051660299301,
+ "learning_rate": 0.0006,
+ "loss": 4.278162002563477,
+ "step": 4860
+ },
+ {
+ "epoch": 67.5170305676856,
+ "grad_norm": 0.015145624987781048,
+ "learning_rate": 0.0006,
+ "loss": 4.267950534820557,
+ "step": 4861
+ },
+ {
+ "epoch": 67.53100436681223,
+ "grad_norm": 0.014610803686082363,
+ "learning_rate": 0.0006,
+ "loss": 4.247499942779541,
+ "step": 4862
+ },
+ {
+ "epoch": 67.54497816593886,
+ "grad_norm": 0.014860909432172775,
+ "learning_rate": 0.0006,
+ "loss": 4.263200759887695,
+ "step": 4863
+ },
+ {
+ "epoch": 67.5589519650655,
+ "grad_norm": 0.01718984916806221,
+ "learning_rate": 0.0006,
+ "loss": 4.290019989013672,
+ "step": 4864
+ },
+ {
+ "epoch": 67.57292576419214,
+ "grad_norm": 0.01558963768184185,
+ "learning_rate": 0.0006,
+ "loss": 4.223816871643066,
+ "step": 4865
+ },
+ {
+ "epoch": 67.58689956331878,
+ "grad_norm": 0.014927188865840435,
+ "learning_rate": 0.0006,
+ "loss": 4.363586902618408,
+ "step": 4866
+ },
+ {
+ "epoch": 67.60087336244541,
+ "grad_norm": 0.01573595404624939,
+ "learning_rate": 0.0006,
+ "loss": 4.254765510559082,
+ "step": 4867
+ },
+ {
+ "epoch": 67.61484716157206,
+ "grad_norm": 0.015618929639458656,
+ "learning_rate": 0.0006,
+ "loss": 4.25072717666626,
+ "step": 4868
+ },
+ {
+ "epoch": 67.62882096069869,
+ "grad_norm": 0.01605590060353279,
+ "learning_rate": 0.0006,
+ "loss": 4.25208854675293,
+ "step": 4869
+ },
+ {
+ "epoch": 67.64279475982532,
+ "grad_norm": 0.01804601214826107,
+ "learning_rate": 0.0006,
+ "loss": 4.351354122161865,
+ "step": 4870
+ },
+ {
+ "epoch": 67.65676855895197,
+ "grad_norm": 0.020595047622919083,
+ "learning_rate": 0.0006,
+ "loss": 4.193732738494873,
+ "step": 4871
+ },
+ {
+ "epoch": 67.6707423580786,
+ "grad_norm": 0.023037966340780258,
+ "learning_rate": 0.0006,
+ "loss": 4.27754020690918,
+ "step": 4872
+ },
+ {
+ "epoch": 67.68471615720524,
+ "grad_norm": 0.020680895075201988,
+ "learning_rate": 0.0006,
+ "loss": 4.314730167388916,
+ "step": 4873
+ },
+ {
+ "epoch": 67.69868995633188,
+ "grad_norm": 0.01887434534728527,
+ "learning_rate": 0.0006,
+ "loss": 4.252305030822754,
+ "step": 4874
+ },
+ {
+ "epoch": 67.71266375545852,
+ "grad_norm": 0.016333596780896187,
+ "learning_rate": 0.0006,
+ "loss": 4.146123886108398,
+ "step": 4875
+ },
+ {
+ "epoch": 67.72663755458515,
+ "grad_norm": 0.015688810497522354,
+ "learning_rate": 0.0006,
+ "loss": 4.2814154624938965,
+ "step": 4876
+ },
+ {
+ "epoch": 67.74061135371178,
+ "grad_norm": 0.016270726919174194,
+ "learning_rate": 0.0006,
+ "loss": 4.323200225830078,
+ "step": 4877
+ },
+ {
+ "epoch": 67.75458515283843,
+ "grad_norm": 0.016824988648295403,
+ "learning_rate": 0.0006,
+ "loss": 4.269501686096191,
+ "step": 4878
+ },
+ {
+ "epoch": 67.76855895196506,
+ "grad_norm": 0.01730562001466751,
+ "learning_rate": 0.0006,
+ "loss": 4.187993049621582,
+ "step": 4879
+ },
+ {
+ "epoch": 67.78253275109171,
+ "grad_norm": 0.01546655036509037,
+ "learning_rate": 0.0006,
+ "loss": 4.340415000915527,
+ "step": 4880
+ },
+ {
+ "epoch": 67.79650655021834,
+ "grad_norm": 0.014173958450555801,
+ "learning_rate": 0.0006,
+ "loss": 4.217419624328613,
+ "step": 4881
+ },
+ {
+ "epoch": 67.81048034934499,
+ "grad_norm": 0.016810474917292595,
+ "learning_rate": 0.0006,
+ "loss": 4.352302074432373,
+ "step": 4882
+ },
+ {
+ "epoch": 67.82445414847162,
+ "grad_norm": 0.0175373163074255,
+ "learning_rate": 0.0006,
+ "loss": 4.20377254486084,
+ "step": 4883
+ },
+ {
+ "epoch": 67.83842794759825,
+ "grad_norm": 0.015734676271677017,
+ "learning_rate": 0.0006,
+ "loss": 4.147495746612549,
+ "step": 4884
+ },
+ {
+ "epoch": 67.8524017467249,
+ "grad_norm": 0.015091817826032639,
+ "learning_rate": 0.0006,
+ "loss": 4.272252082824707,
+ "step": 4885
+ },
+ {
+ "epoch": 67.86637554585153,
+ "grad_norm": 0.015000219456851482,
+ "learning_rate": 0.0006,
+ "loss": 4.302773952484131,
+ "step": 4886
+ },
+ {
+ "epoch": 67.88034934497817,
+ "grad_norm": 0.013827184215188026,
+ "learning_rate": 0.0006,
+ "loss": 4.337242126464844,
+ "step": 4887
+ },
+ {
+ "epoch": 67.8943231441048,
+ "grad_norm": 0.015308346599340439,
+ "learning_rate": 0.0006,
+ "loss": 4.140895366668701,
+ "step": 4888
+ },
+ {
+ "epoch": 67.90829694323143,
+ "grad_norm": 0.015356178395450115,
+ "learning_rate": 0.0006,
+ "loss": 4.312541484832764,
+ "step": 4889
+ },
+ {
+ "epoch": 67.92227074235808,
+ "grad_norm": 0.0137353939935565,
+ "learning_rate": 0.0006,
+ "loss": 4.173184394836426,
+ "step": 4890
+ },
+ {
+ "epoch": 67.93624454148471,
+ "grad_norm": 0.01332368515431881,
+ "learning_rate": 0.0006,
+ "loss": 4.191199779510498,
+ "step": 4891
+ },
+ {
+ "epoch": 67.95021834061136,
+ "grad_norm": 0.01488130260258913,
+ "learning_rate": 0.0006,
+ "loss": 4.331752300262451,
+ "step": 4892
+ },
+ {
+ "epoch": 67.96419213973799,
+ "grad_norm": 0.015449733473360538,
+ "learning_rate": 0.0006,
+ "loss": 4.2385454177856445,
+ "step": 4893
+ },
+ {
+ "epoch": 67.97816593886463,
+ "grad_norm": 0.01556103490293026,
+ "learning_rate": 0.0006,
+ "loss": 4.262353897094727,
+ "step": 4894
+ },
+ {
+ "epoch": 67.99213973799127,
+ "grad_norm": 0.014613849110901356,
+ "learning_rate": 0.0006,
+ "loss": 4.223945617675781,
+ "step": 4895
+ },
+ {
+ "epoch": 68.0,
+ "grad_norm": 0.015323462896049023,
+ "learning_rate": 0.0006,
+ "loss": 4.317168712615967,
+ "step": 4896
+ },
+ {
+ "epoch": 68.0,
+ "eval_loss": 4.671603679656982,
+ "eval_runtime": 56.7317,
+ "eval_samples_per_second": 43.045,
+ "eval_steps_per_second": 1.357,
+ "step": 4896
+ },
+ {
+ "epoch": 68.01397379912663,
+ "grad_norm": 0.01910698413848877,
+ "learning_rate": 0.0006,
+ "loss": 4.206167221069336,
+ "step": 4897
+ },
+ {
+ "epoch": 68.02794759825328,
+ "grad_norm": 0.025112485513091087,
+ "learning_rate": 0.0006,
+ "loss": 4.384706974029541,
+ "step": 4898
+ },
+ {
+ "epoch": 68.04192139737991,
+ "grad_norm": 0.02618720941245556,
+ "learning_rate": 0.0006,
+ "loss": 4.321046829223633,
+ "step": 4899
+ },
+ {
+ "epoch": 68.05589519650655,
+ "grad_norm": 0.020853538066148758,
+ "learning_rate": 0.0006,
+ "loss": 4.217638969421387,
+ "step": 4900
+ },
+ {
+ "epoch": 68.06986899563319,
+ "grad_norm": 0.020025717094540596,
+ "learning_rate": 0.0006,
+ "loss": 4.306824684143066,
+ "step": 4901
+ },
+ {
+ "epoch": 68.08384279475983,
+ "grad_norm": 0.023868650197982788,
+ "learning_rate": 0.0006,
+ "loss": 4.327455997467041,
+ "step": 4902
+ },
+ {
+ "epoch": 68.09781659388646,
+ "grad_norm": 0.022279081866145134,
+ "learning_rate": 0.0006,
+ "loss": 4.400699615478516,
+ "step": 4903
+ },
+ {
+ "epoch": 68.1117903930131,
+ "grad_norm": 0.018965687602758408,
+ "learning_rate": 0.0006,
+ "loss": 4.19097900390625,
+ "step": 4904
+ },
+ {
+ "epoch": 68.12576419213974,
+ "grad_norm": 0.01936378888785839,
+ "learning_rate": 0.0006,
+ "loss": 4.206711769104004,
+ "step": 4905
+ },
+ {
+ "epoch": 68.13973799126637,
+ "grad_norm": 0.01811872608959675,
+ "learning_rate": 0.0006,
+ "loss": 4.204622268676758,
+ "step": 4906
+ },
+ {
+ "epoch": 68.15371179039302,
+ "grad_norm": 0.018530402332544327,
+ "learning_rate": 0.0006,
+ "loss": 4.289602279663086,
+ "step": 4907
+ },
+ {
+ "epoch": 68.16768558951965,
+ "grad_norm": 0.01755962148308754,
+ "learning_rate": 0.0006,
+ "loss": 4.317145347595215,
+ "step": 4908
+ },
+ {
+ "epoch": 68.18165938864628,
+ "grad_norm": 0.020413702353835106,
+ "learning_rate": 0.0006,
+ "loss": 4.2882843017578125,
+ "step": 4909
+ },
+ {
+ "epoch": 68.19563318777293,
+ "grad_norm": 0.020064299926161766,
+ "learning_rate": 0.0006,
+ "loss": 4.248260021209717,
+ "step": 4910
+ },
+ {
+ "epoch": 68.20960698689956,
+ "grad_norm": 0.017415888607501984,
+ "learning_rate": 0.0006,
+ "loss": 4.370295524597168,
+ "step": 4911
+ },
+ {
+ "epoch": 68.2235807860262,
+ "grad_norm": 0.015005858615040779,
+ "learning_rate": 0.0006,
+ "loss": 4.319567680358887,
+ "step": 4912
+ },
+ {
+ "epoch": 68.23755458515284,
+ "grad_norm": 0.016349777579307556,
+ "learning_rate": 0.0006,
+ "loss": 4.301776885986328,
+ "step": 4913
+ },
+ {
+ "epoch": 68.25152838427948,
+ "grad_norm": 0.015779560431838036,
+ "learning_rate": 0.0006,
+ "loss": 4.271171569824219,
+ "step": 4914
+ },
+ {
+ "epoch": 68.26550218340611,
+ "grad_norm": 0.013237644918262959,
+ "learning_rate": 0.0006,
+ "loss": 4.324908256530762,
+ "step": 4915
+ },
+ {
+ "epoch": 68.27947598253274,
+ "grad_norm": 0.014076225459575653,
+ "learning_rate": 0.0006,
+ "loss": 4.177175998687744,
+ "step": 4916
+ },
+ {
+ "epoch": 68.29344978165939,
+ "grad_norm": 0.013906744308769703,
+ "learning_rate": 0.0006,
+ "loss": 4.11588191986084,
+ "step": 4917
+ },
+ {
+ "epoch": 68.30742358078602,
+ "grad_norm": 0.013939064927399158,
+ "learning_rate": 0.0006,
+ "loss": 4.183963775634766,
+ "step": 4918
+ },
+ {
+ "epoch": 68.32139737991267,
+ "grad_norm": 0.014572170563042164,
+ "learning_rate": 0.0006,
+ "loss": 4.19586181640625,
+ "step": 4919
+ },
+ {
+ "epoch": 68.3353711790393,
+ "grad_norm": 0.014085350558161736,
+ "learning_rate": 0.0006,
+ "loss": 4.170892238616943,
+ "step": 4920
+ },
+ {
+ "epoch": 68.34934497816595,
+ "grad_norm": 0.014357863925397396,
+ "learning_rate": 0.0006,
+ "loss": 4.321155548095703,
+ "step": 4921
+ },
+ {
+ "epoch": 68.36331877729258,
+ "grad_norm": 0.01461675763130188,
+ "learning_rate": 0.0006,
+ "loss": 4.293334007263184,
+ "step": 4922
+ },
+ {
+ "epoch": 68.37729257641921,
+ "grad_norm": 0.014828486368060112,
+ "learning_rate": 0.0006,
+ "loss": 4.1439714431762695,
+ "step": 4923
+ },
+ {
+ "epoch": 68.39126637554585,
+ "grad_norm": 0.013730363920331001,
+ "learning_rate": 0.0006,
+ "loss": 4.287973880767822,
+ "step": 4924
+ },
+ {
+ "epoch": 68.40524017467249,
+ "grad_norm": 0.013852203264832497,
+ "learning_rate": 0.0006,
+ "loss": 4.248824119567871,
+ "step": 4925
+ },
+ {
+ "epoch": 68.41921397379913,
+ "grad_norm": 0.014005862176418304,
+ "learning_rate": 0.0006,
+ "loss": 4.293768882751465,
+ "step": 4926
+ },
+ {
+ "epoch": 68.43318777292576,
+ "grad_norm": 0.014435851015150547,
+ "learning_rate": 0.0006,
+ "loss": 4.204813003540039,
+ "step": 4927
+ },
+ {
+ "epoch": 68.44716157205241,
+ "grad_norm": 0.015305934473872185,
+ "learning_rate": 0.0006,
+ "loss": 4.329364776611328,
+ "step": 4928
+ },
+ {
+ "epoch": 68.46113537117904,
+ "grad_norm": 0.014356814324855804,
+ "learning_rate": 0.0006,
+ "loss": 4.264206409454346,
+ "step": 4929
+ },
+ {
+ "epoch": 68.47510917030567,
+ "grad_norm": 0.01432347297668457,
+ "learning_rate": 0.0006,
+ "loss": 4.29440450668335,
+ "step": 4930
+ },
+ {
+ "epoch": 68.48908296943232,
+ "grad_norm": 0.012672432698309422,
+ "learning_rate": 0.0006,
+ "loss": 4.258024215698242,
+ "step": 4931
+ },
+ {
+ "epoch": 68.50305676855895,
+ "grad_norm": 0.014209977351129055,
+ "learning_rate": 0.0006,
+ "loss": 4.216668128967285,
+ "step": 4932
+ },
+ {
+ "epoch": 68.5170305676856,
+ "grad_norm": 0.014606663957238197,
+ "learning_rate": 0.0006,
+ "loss": 4.294184684753418,
+ "step": 4933
+ },
+ {
+ "epoch": 68.53100436681223,
+ "grad_norm": 0.014986990951001644,
+ "learning_rate": 0.0006,
+ "loss": 4.382591247558594,
+ "step": 4934
+ },
+ {
+ "epoch": 68.54497816593886,
+ "grad_norm": 0.013998783193528652,
+ "learning_rate": 0.0006,
+ "loss": 4.195143222808838,
+ "step": 4935
+ },
+ {
+ "epoch": 68.5589519650655,
+ "grad_norm": 0.015011102892458439,
+ "learning_rate": 0.0006,
+ "loss": 4.313785076141357,
+ "step": 4936
+ },
+ {
+ "epoch": 68.57292576419214,
+ "grad_norm": 0.01641209051012993,
+ "learning_rate": 0.0006,
+ "loss": 4.322803020477295,
+ "step": 4937
+ },
+ {
+ "epoch": 68.58689956331878,
+ "grad_norm": 0.016207464039325714,
+ "learning_rate": 0.0006,
+ "loss": 4.311431884765625,
+ "step": 4938
+ },
+ {
+ "epoch": 68.60087336244541,
+ "grad_norm": 0.01747811958193779,
+ "learning_rate": 0.0006,
+ "loss": 4.310845375061035,
+ "step": 4939
+ },
+ {
+ "epoch": 68.61484716157206,
+ "grad_norm": 0.019475717097520828,
+ "learning_rate": 0.0006,
+ "loss": 4.338624954223633,
+ "step": 4940
+ },
+ {
+ "epoch": 68.62882096069869,
+ "grad_norm": 0.018641719594597816,
+ "learning_rate": 0.0006,
+ "loss": 4.34913969039917,
+ "step": 4941
+ },
+ {
+ "epoch": 68.64279475982532,
+ "grad_norm": 0.018376432359218597,
+ "learning_rate": 0.0006,
+ "loss": 4.28464412689209,
+ "step": 4942
+ },
+ {
+ "epoch": 68.65676855895197,
+ "grad_norm": 0.017095010727643967,
+ "learning_rate": 0.0006,
+ "loss": 4.217957496643066,
+ "step": 4943
+ },
+ {
+ "epoch": 68.6707423580786,
+ "grad_norm": 0.016321983188390732,
+ "learning_rate": 0.0006,
+ "loss": 4.220042705535889,
+ "step": 4944
+ },
+ {
+ "epoch": 68.68471615720524,
+ "grad_norm": 0.017128000035881996,
+ "learning_rate": 0.0006,
+ "loss": 4.293500900268555,
+ "step": 4945
+ },
+ {
+ "epoch": 68.69868995633188,
+ "grad_norm": 0.015592411160469055,
+ "learning_rate": 0.0006,
+ "loss": 4.195798873901367,
+ "step": 4946
+ },
+ {
+ "epoch": 68.71266375545852,
+ "grad_norm": 0.016607079654932022,
+ "learning_rate": 0.0006,
+ "loss": 4.2654337882995605,
+ "step": 4947
+ },
+ {
+ "epoch": 68.72663755458515,
+ "grad_norm": 0.01545762736350298,
+ "learning_rate": 0.0006,
+ "loss": 4.303828239440918,
+ "step": 4948
+ },
+ {
+ "epoch": 68.74061135371178,
+ "grad_norm": 0.014328422956168652,
+ "learning_rate": 0.0006,
+ "loss": 4.29379940032959,
+ "step": 4949
+ },
+ {
+ "epoch": 68.75458515283843,
+ "grad_norm": 0.016516804695129395,
+ "learning_rate": 0.0006,
+ "loss": 4.2317214012146,
+ "step": 4950
+ },
+ {
+ "epoch": 68.76855895196506,
+ "grad_norm": 0.016324171796441078,
+ "learning_rate": 0.0006,
+ "loss": 4.278043746948242,
+ "step": 4951
+ },
+ {
+ "epoch": 68.78253275109171,
+ "grad_norm": 0.015511159785091877,
+ "learning_rate": 0.0006,
+ "loss": 4.188244342803955,
+ "step": 4952
+ },
+ {
+ "epoch": 68.79650655021834,
+ "grad_norm": 0.01676103286445141,
+ "learning_rate": 0.0006,
+ "loss": 4.2169976234436035,
+ "step": 4953
+ },
+ {
+ "epoch": 68.81048034934499,
+ "grad_norm": 0.015428312122821808,
+ "learning_rate": 0.0006,
+ "loss": 4.4004716873168945,
+ "step": 4954
+ },
+ {
+ "epoch": 68.82445414847162,
+ "grad_norm": 0.0162410456687212,
+ "learning_rate": 0.0006,
+ "loss": 4.282752513885498,
+ "step": 4955
+ },
+ {
+ "epoch": 68.83842794759825,
+ "grad_norm": 0.017578862607479095,
+ "learning_rate": 0.0006,
+ "loss": 4.098004341125488,
+ "step": 4956
+ },
+ {
+ "epoch": 68.8524017467249,
+ "grad_norm": 0.01805749163031578,
+ "learning_rate": 0.0006,
+ "loss": 4.277988433837891,
+ "step": 4957
+ },
+ {
+ "epoch": 68.86637554585153,
+ "grad_norm": 0.016728045418858528,
+ "learning_rate": 0.0006,
+ "loss": 4.29766321182251,
+ "step": 4958
+ },
+ {
+ "epoch": 68.88034934497817,
+ "grad_norm": 0.014899657107889652,
+ "learning_rate": 0.0006,
+ "loss": 4.329736709594727,
+ "step": 4959
+ },
+ {
+ "epoch": 68.8943231441048,
+ "grad_norm": 0.013441305607557297,
+ "learning_rate": 0.0006,
+ "loss": 4.2873215675354,
+ "step": 4960
+ },
+ {
+ "epoch": 68.90829694323143,
+ "grad_norm": 0.01554128248244524,
+ "learning_rate": 0.0006,
+ "loss": 4.278327941894531,
+ "step": 4961
+ },
+ {
+ "epoch": 68.92227074235808,
+ "grad_norm": 0.01806696318089962,
+ "learning_rate": 0.0006,
+ "loss": 4.38275146484375,
+ "step": 4962
+ },
+ {
+ "epoch": 68.93624454148471,
+ "grad_norm": 0.017025409266352654,
+ "learning_rate": 0.0006,
+ "loss": 4.210447311401367,
+ "step": 4963
+ },
+ {
+ "epoch": 68.95021834061136,
+ "grad_norm": 0.01597420684993267,
+ "learning_rate": 0.0006,
+ "loss": 4.309865951538086,
+ "step": 4964
+ },
+ {
+ "epoch": 68.96419213973799,
+ "grad_norm": 0.015835624188184738,
+ "learning_rate": 0.0006,
+ "loss": 4.225904941558838,
+ "step": 4965
+ },
+ {
+ "epoch": 68.97816593886463,
+ "grad_norm": 0.018170492723584175,
+ "learning_rate": 0.0006,
+ "loss": 4.248478889465332,
+ "step": 4966
+ },
+ {
+ "epoch": 68.99213973799127,
+ "grad_norm": 0.017372991889715195,
+ "learning_rate": 0.0006,
+ "loss": 4.283486843109131,
+ "step": 4967
+ },
+ {
+ "epoch": 69.0,
+ "grad_norm": 0.016305390745401382,
+ "learning_rate": 0.0006,
+ "loss": 4.273636817932129,
+ "step": 4968
+ },
+ {
+ "epoch": 69.0,
+ "eval_loss": 4.66873836517334,
+ "eval_runtime": 57.0791,
+ "eval_samples_per_second": 42.783,
+ "eval_steps_per_second": 1.349,
+ "step": 4968
+ },
+ {
+ "epoch": 69.01397379912663,
+ "grad_norm": 0.016567355021834373,
+ "learning_rate": 0.0006,
+ "loss": 4.269988059997559,
+ "step": 4969
+ },
+ {
+ "epoch": 69.02794759825328,
+ "grad_norm": 0.017476852983236313,
+ "learning_rate": 0.0006,
+ "loss": 4.2518720626831055,
+ "step": 4970
+ },
+ {
+ "epoch": 69.04192139737991,
+ "grad_norm": 0.017070338129997253,
+ "learning_rate": 0.0006,
+ "loss": 4.238024711608887,
+ "step": 4971
+ },
+ {
+ "epoch": 69.05589519650655,
+ "grad_norm": 0.01639373227953911,
+ "learning_rate": 0.0006,
+ "loss": 4.188558101654053,
+ "step": 4972
+ },
+ {
+ "epoch": 69.06986899563319,
+ "grad_norm": 0.01856113411486149,
+ "learning_rate": 0.0006,
+ "loss": 4.339556694030762,
+ "step": 4973
+ },
+ {
+ "epoch": 69.08384279475983,
+ "grad_norm": 0.01762578636407852,
+ "learning_rate": 0.0006,
+ "loss": 4.141518592834473,
+ "step": 4974
+ },
+ {
+ "epoch": 69.09781659388646,
+ "grad_norm": 0.015437949448823929,
+ "learning_rate": 0.0006,
+ "loss": 4.225461959838867,
+ "step": 4975
+ },
+ {
+ "epoch": 69.1117903930131,
+ "grad_norm": 0.015293503180146217,
+ "learning_rate": 0.0006,
+ "loss": 4.290884971618652,
+ "step": 4976
+ },
+ {
+ "epoch": 69.12576419213974,
+ "grad_norm": 0.016259880736470222,
+ "learning_rate": 0.0006,
+ "loss": 4.121919631958008,
+ "step": 4977
+ },
+ {
+ "epoch": 69.13973799126637,
+ "grad_norm": 0.017225584015250206,
+ "learning_rate": 0.0006,
+ "loss": 4.136716842651367,
+ "step": 4978
+ },
+ {
+ "epoch": 69.15371179039302,
+ "grad_norm": 0.01699289306998253,
+ "learning_rate": 0.0006,
+ "loss": 4.193601608276367,
+ "step": 4979
+ },
+ {
+ "epoch": 69.16768558951965,
+ "grad_norm": 0.015613161027431488,
+ "learning_rate": 0.0006,
+ "loss": 4.288212299346924,
+ "step": 4980
+ },
+ {
+ "epoch": 69.18165938864628,
+ "grad_norm": 0.01643933728337288,
+ "learning_rate": 0.0006,
+ "loss": 4.324707984924316,
+ "step": 4981
+ },
+ {
+ "epoch": 69.19563318777293,
+ "grad_norm": 0.01648905873298645,
+ "learning_rate": 0.0006,
+ "loss": 4.287806510925293,
+ "step": 4982
+ },
+ {
+ "epoch": 69.20960698689956,
+ "grad_norm": 0.015444950200617313,
+ "learning_rate": 0.0006,
+ "loss": 4.200622081756592,
+ "step": 4983
+ },
+ {
+ "epoch": 69.2235807860262,
+ "grad_norm": 0.016034258529543877,
+ "learning_rate": 0.0006,
+ "loss": 4.2193827629089355,
+ "step": 4984
+ },
+ {
+ "epoch": 69.23755458515284,
+ "grad_norm": 0.017984388396143913,
+ "learning_rate": 0.0006,
+ "loss": 4.18237829208374,
+ "step": 4985
+ },
+ {
+ "epoch": 69.25152838427948,
+ "grad_norm": 0.0195505041629076,
+ "learning_rate": 0.0006,
+ "loss": 4.216727256774902,
+ "step": 4986
+ },
+ {
+ "epoch": 69.26550218340611,
+ "grad_norm": 0.017779147252440453,
+ "learning_rate": 0.0006,
+ "loss": 4.278628826141357,
+ "step": 4987
+ },
+ {
+ "epoch": 69.27947598253274,
+ "grad_norm": 0.017119476571679115,
+ "learning_rate": 0.0006,
+ "loss": 4.2409138679504395,
+ "step": 4988
+ },
+ {
+ "epoch": 69.29344978165939,
+ "grad_norm": 0.016840364784002304,
+ "learning_rate": 0.0006,
+ "loss": 4.207575798034668,
+ "step": 4989
+ },
+ {
+ "epoch": 69.30742358078602,
+ "grad_norm": 0.014593573287129402,
+ "learning_rate": 0.0006,
+ "loss": 4.164905548095703,
+ "step": 4990
+ },
+ {
+ "epoch": 69.32139737991267,
+ "grad_norm": 0.013922630809247494,
+ "learning_rate": 0.0006,
+ "loss": 4.2696990966796875,
+ "step": 4991
+ },
+ {
+ "epoch": 69.3353711790393,
+ "grad_norm": 0.01577564887702465,
+ "learning_rate": 0.0006,
+ "loss": 4.201589107513428,
+ "step": 4992
+ },
+ {
+ "epoch": 69.34934497816595,
+ "grad_norm": 0.016392096877098083,
+ "learning_rate": 0.0006,
+ "loss": 4.248831748962402,
+ "step": 4993
+ },
+ {
+ "epoch": 69.36331877729258,
+ "grad_norm": 0.015961311757564545,
+ "learning_rate": 0.0006,
+ "loss": 4.181260108947754,
+ "step": 4994
+ },
+ {
+ "epoch": 69.37729257641921,
+ "grad_norm": 0.015350551344454288,
+ "learning_rate": 0.0006,
+ "loss": 4.230352878570557,
+ "step": 4995
+ },
+ {
+ "epoch": 69.39126637554585,
+ "grad_norm": 0.015870466828346252,
+ "learning_rate": 0.0006,
+ "loss": 4.298305511474609,
+ "step": 4996
+ },
+ {
+ "epoch": 69.40524017467249,
+ "grad_norm": 0.013716642744839191,
+ "learning_rate": 0.0006,
+ "loss": 4.220139503479004,
+ "step": 4997
+ },
+ {
+ "epoch": 69.41921397379913,
+ "grad_norm": 0.014431777410209179,
+ "learning_rate": 0.0006,
+ "loss": 4.288130760192871,
+ "step": 4998
+ },
+ {
+ "epoch": 69.43318777292576,
+ "grad_norm": 0.01426647324115038,
+ "learning_rate": 0.0006,
+ "loss": 4.295767784118652,
+ "step": 4999
+ },
+ {
+ "epoch": 69.44716157205241,
+ "grad_norm": 0.014286615885794163,
+ "learning_rate": 0.0006,
+ "loss": 4.38106632232666,
+ "step": 5000
+ },
+ {
+ "epoch": 69.46113537117904,
+ "grad_norm": 0.017990585416555405,
+ "learning_rate": 0.0006,
+ "loss": 4.1593017578125,
+ "step": 5001
+ },
+ {
+ "epoch": 69.47510917030567,
+ "grad_norm": 0.018846049904823303,
+ "learning_rate": 0.0006,
+ "loss": 4.216091632843018,
+ "step": 5002
+ },
+ {
+ "epoch": 69.48908296943232,
+ "grad_norm": 0.018179181963205338,
+ "learning_rate": 0.0006,
+ "loss": 4.240332126617432,
+ "step": 5003
+ },
+ {
+ "epoch": 69.50305676855895,
+ "grad_norm": 0.016693448647856712,
+ "learning_rate": 0.0006,
+ "loss": 4.256624221801758,
+ "step": 5004
+ },
+ {
+ "epoch": 69.5170305676856,
+ "grad_norm": 0.01718989387154579,
+ "learning_rate": 0.0006,
+ "loss": 4.300940990447998,
+ "step": 5005
+ },
+ {
+ "epoch": 69.53100436681223,
+ "grad_norm": 0.020236829295754433,
+ "learning_rate": 0.0006,
+ "loss": 4.289146423339844,
+ "step": 5006
+ },
+ {
+ "epoch": 69.54497816593886,
+ "grad_norm": 0.024860285222530365,
+ "learning_rate": 0.0006,
+ "loss": 4.216216087341309,
+ "step": 5007
+ },
+ {
+ "epoch": 69.5589519650655,
+ "grad_norm": 0.02559969387948513,
+ "learning_rate": 0.0006,
+ "loss": 4.262890815734863,
+ "step": 5008
+ },
+ {
+ "epoch": 69.57292576419214,
+ "grad_norm": 0.019180957227945328,
+ "learning_rate": 0.0006,
+ "loss": 4.230309009552002,
+ "step": 5009
+ },
+ {
+ "epoch": 69.58689956331878,
+ "grad_norm": 0.020442591980099678,
+ "learning_rate": 0.0006,
+ "loss": 4.291759490966797,
+ "step": 5010
+ },
+ {
+ "epoch": 69.60087336244541,
+ "grad_norm": 0.0212619137018919,
+ "learning_rate": 0.0006,
+ "loss": 4.3529839515686035,
+ "step": 5011
+ },
+ {
+ "epoch": 69.61484716157206,
+ "grad_norm": 0.02067175693809986,
+ "learning_rate": 0.0006,
+ "loss": 4.33197021484375,
+ "step": 5012
+ },
+ {
+ "epoch": 69.62882096069869,
+ "grad_norm": 0.01802872307598591,
+ "learning_rate": 0.0006,
+ "loss": 4.2101969718933105,
+ "step": 5013
+ },
+ {
+ "epoch": 69.64279475982532,
+ "grad_norm": 0.01787702925503254,
+ "learning_rate": 0.0006,
+ "loss": 4.270336627960205,
+ "step": 5014
+ },
+ {
+ "epoch": 69.65676855895197,
+ "grad_norm": 0.01586683839559555,
+ "learning_rate": 0.0006,
+ "loss": 4.222309589385986,
+ "step": 5015
+ },
+ {
+ "epoch": 69.6707423580786,
+ "grad_norm": 0.01717999391257763,
+ "learning_rate": 0.0006,
+ "loss": 4.253704071044922,
+ "step": 5016
+ },
+ {
+ "epoch": 69.68471615720524,
+ "grad_norm": 0.01594279147684574,
+ "learning_rate": 0.0006,
+ "loss": 4.344658851623535,
+ "step": 5017
+ },
+ {
+ "epoch": 69.69868995633188,
+ "grad_norm": 0.014419942162930965,
+ "learning_rate": 0.0006,
+ "loss": 4.2144880294799805,
+ "step": 5018
+ },
+ {
+ "epoch": 69.71266375545852,
+ "grad_norm": 0.015824224799871445,
+ "learning_rate": 0.0006,
+ "loss": 4.226252555847168,
+ "step": 5019
+ },
+ {
+ "epoch": 69.72663755458515,
+ "grad_norm": 0.013329979032278061,
+ "learning_rate": 0.0006,
+ "loss": 4.238587379455566,
+ "step": 5020
+ },
+ {
+ "epoch": 69.74061135371178,
+ "grad_norm": 0.013745979405939579,
+ "learning_rate": 0.0006,
+ "loss": 4.317209243774414,
+ "step": 5021
+ },
+ {
+ "epoch": 69.75458515283843,
+ "grad_norm": 0.014064167626202106,
+ "learning_rate": 0.0006,
+ "loss": 4.316982746124268,
+ "step": 5022
+ },
+ {
+ "epoch": 69.76855895196506,
+ "grad_norm": 0.014611301012337208,
+ "learning_rate": 0.0006,
+ "loss": 4.212728500366211,
+ "step": 5023
+ },
+ {
+ "epoch": 69.78253275109171,
+ "grad_norm": 0.014947416260838509,
+ "learning_rate": 0.0006,
+ "loss": 4.319550514221191,
+ "step": 5024
+ },
+ {
+ "epoch": 69.79650655021834,
+ "grad_norm": 0.014780526980757713,
+ "learning_rate": 0.0006,
+ "loss": 4.1888933181762695,
+ "step": 5025
+ },
+ {
+ "epoch": 69.81048034934499,
+ "grad_norm": 0.01611459068953991,
+ "learning_rate": 0.0006,
+ "loss": 4.239692211151123,
+ "step": 5026
+ },
+ {
+ "epoch": 69.82445414847162,
+ "grad_norm": 0.015395848080515862,
+ "learning_rate": 0.0006,
+ "loss": 4.308794021606445,
+ "step": 5027
+ },
+ {
+ "epoch": 69.83842794759825,
+ "grad_norm": 0.016113396733999252,
+ "learning_rate": 0.0006,
+ "loss": 4.328035354614258,
+ "step": 5028
+ },
+ {
+ "epoch": 69.8524017467249,
+ "grad_norm": 0.016386888921260834,
+ "learning_rate": 0.0006,
+ "loss": 4.272579193115234,
+ "step": 5029
+ },
+ {
+ "epoch": 69.86637554585153,
+ "grad_norm": 0.013675205409526825,
+ "learning_rate": 0.0006,
+ "loss": 4.226552963256836,
+ "step": 5030
+ },
+ {
+ "epoch": 69.88034934497817,
+ "grad_norm": 0.01472668256610632,
+ "learning_rate": 0.0006,
+ "loss": 4.224167346954346,
+ "step": 5031
+ },
+ {
+ "epoch": 69.8943231441048,
+ "grad_norm": 0.014958228915929794,
+ "learning_rate": 0.0006,
+ "loss": 4.232619285583496,
+ "step": 5032
+ },
+ {
+ "epoch": 69.90829694323143,
+ "grad_norm": 0.012926832772791386,
+ "learning_rate": 0.0006,
+ "loss": 4.27107048034668,
+ "step": 5033
+ },
+ {
+ "epoch": 69.92227074235808,
+ "grad_norm": 0.013138801790773869,
+ "learning_rate": 0.0006,
+ "loss": 4.288477420806885,
+ "step": 5034
+ },
+ {
+ "epoch": 69.93624454148471,
+ "grad_norm": 0.013415461406111717,
+ "learning_rate": 0.0006,
+ "loss": 4.2033257484436035,
+ "step": 5035
+ },
+ {
+ "epoch": 69.95021834061136,
+ "grad_norm": 0.013807233422994614,
+ "learning_rate": 0.0006,
+ "loss": 4.274764537811279,
+ "step": 5036
+ },
+ {
+ "epoch": 69.96419213973799,
+ "grad_norm": 0.013929193839430809,
+ "learning_rate": 0.0006,
+ "loss": 4.1854352951049805,
+ "step": 5037
+ },
+ {
+ "epoch": 69.97816593886463,
+ "grad_norm": 0.015497001819312572,
+ "learning_rate": 0.0006,
+ "loss": 4.226498126983643,
+ "step": 5038
+ },
+ {
+ "epoch": 69.99213973799127,
+ "grad_norm": 0.01874048449099064,
+ "learning_rate": 0.0006,
+ "loss": 4.304137706756592,
+ "step": 5039
+ },
+ {
+ "epoch": 70.0,
+ "grad_norm": 0.021745791658759117,
+ "learning_rate": 0.0006,
+ "loss": 4.206809997558594,
+ "step": 5040
+ },
+ {
+ "epoch": 70.0,
+ "eval_loss": 4.675539970397949,
+ "eval_runtime": 56.9034,
+ "eval_samples_per_second": 42.915,
+ "eval_steps_per_second": 1.353,
+ "step": 5040
+ },
+ {
+ "epoch": 70.01397379912663,
+ "grad_norm": 0.02124839462339878,
+ "learning_rate": 0.0006,
+ "loss": 4.198199272155762,
+ "step": 5041
+ },
+ {
+ "epoch": 70.02794759825328,
+ "grad_norm": 0.019432688131928444,
+ "learning_rate": 0.0006,
+ "loss": 4.187261581420898,
+ "step": 5042
+ },
+ {
+ "epoch": 70.04192139737991,
+ "grad_norm": 0.01769290305674076,
+ "learning_rate": 0.0006,
+ "loss": 4.155284881591797,
+ "step": 5043
+ },
+ {
+ "epoch": 70.05589519650655,
+ "grad_norm": 0.01751520484685898,
+ "learning_rate": 0.0006,
+ "loss": 4.386857986450195,
+ "step": 5044
+ },
+ {
+ "epoch": 70.06986899563319,
+ "grad_norm": 0.018138699233531952,
+ "learning_rate": 0.0006,
+ "loss": 4.117342472076416,
+ "step": 5045
+ },
+ {
+ "epoch": 70.08384279475983,
+ "grad_norm": 0.016331054270267487,
+ "learning_rate": 0.0006,
+ "loss": 4.225295066833496,
+ "step": 5046
+ },
+ {
+ "epoch": 70.09781659388646,
+ "grad_norm": 0.01658729277551174,
+ "learning_rate": 0.0006,
+ "loss": 4.143835067749023,
+ "step": 5047
+ },
+ {
+ "epoch": 70.1117903930131,
+ "grad_norm": 0.016367698088288307,
+ "learning_rate": 0.0006,
+ "loss": 4.1469268798828125,
+ "step": 5048
+ },
+ {
+ "epoch": 70.12576419213974,
+ "grad_norm": 0.015151629224419594,
+ "learning_rate": 0.0006,
+ "loss": 4.117629051208496,
+ "step": 5049
+ },
+ {
+ "epoch": 70.13973799126637,
+ "grad_norm": 0.016434503719210625,
+ "learning_rate": 0.0006,
+ "loss": 4.107848644256592,
+ "step": 5050
+ },
+ {
+ "epoch": 70.15371179039302,
+ "grad_norm": 0.017837485298514366,
+ "learning_rate": 0.0006,
+ "loss": 4.10837459564209,
+ "step": 5051
+ },
+ {
+ "epoch": 70.16768558951965,
+ "grad_norm": 0.0165961142629385,
+ "learning_rate": 0.0006,
+ "loss": 4.169414043426514,
+ "step": 5052
+ },
+ {
+ "epoch": 70.18165938864628,
+ "grad_norm": 0.0157014112919569,
+ "learning_rate": 0.0006,
+ "loss": 4.0629706382751465,
+ "step": 5053
+ },
+ {
+ "epoch": 70.19563318777293,
+ "grad_norm": 0.01581883803009987,
+ "learning_rate": 0.0006,
+ "loss": 4.220012664794922,
+ "step": 5054
+ },
+ {
+ "epoch": 70.20960698689956,
+ "grad_norm": 0.015285917557775974,
+ "learning_rate": 0.0006,
+ "loss": 4.275337219238281,
+ "step": 5055
+ },
+ {
+ "epoch": 70.2235807860262,
+ "grad_norm": 0.014984549954533577,
+ "learning_rate": 0.0006,
+ "loss": 4.284764289855957,
+ "step": 5056
+ },
+ {
+ "epoch": 70.23755458515284,
+ "grad_norm": 0.015172012150287628,
+ "learning_rate": 0.0006,
+ "loss": 4.234498023986816,
+ "step": 5057
+ },
+ {
+ "epoch": 70.25152838427948,
+ "grad_norm": 0.016846632584929466,
+ "learning_rate": 0.0006,
+ "loss": 4.181689262390137,
+ "step": 5058
+ },
+ {
+ "epoch": 70.26550218340611,
+ "grad_norm": 0.01920711249113083,
+ "learning_rate": 0.0006,
+ "loss": 4.235280513763428,
+ "step": 5059
+ },
+ {
+ "epoch": 70.27947598253274,
+ "grad_norm": 0.017765291035175323,
+ "learning_rate": 0.0006,
+ "loss": 4.238955497741699,
+ "step": 5060
+ },
+ {
+ "epoch": 70.29344978165939,
+ "grad_norm": 0.015436639077961445,
+ "learning_rate": 0.0006,
+ "loss": 4.232516288757324,
+ "step": 5061
+ },
+ {
+ "epoch": 70.30742358078602,
+ "grad_norm": 0.015660086646676064,
+ "learning_rate": 0.0006,
+ "loss": 4.260349273681641,
+ "step": 5062
+ },
+ {
+ "epoch": 70.32139737991267,
+ "grad_norm": 0.016095709055662155,
+ "learning_rate": 0.0006,
+ "loss": 4.168505668640137,
+ "step": 5063
+ },
+ {
+ "epoch": 70.3353711790393,
+ "grad_norm": 0.015001940540969372,
+ "learning_rate": 0.0006,
+ "loss": 4.161739349365234,
+ "step": 5064
+ },
+ {
+ "epoch": 70.34934497816595,
+ "grad_norm": 0.014668774791061878,
+ "learning_rate": 0.0006,
+ "loss": 4.181029796600342,
+ "step": 5065
+ },
+ {
+ "epoch": 70.36331877729258,
+ "grad_norm": 0.013266284950077534,
+ "learning_rate": 0.0006,
+ "loss": 4.3052873611450195,
+ "step": 5066
+ },
+ {
+ "epoch": 70.37729257641921,
+ "grad_norm": 0.015396138653159142,
+ "learning_rate": 0.0006,
+ "loss": 4.34413480758667,
+ "step": 5067
+ },
+ {
+ "epoch": 70.39126637554585,
+ "grad_norm": 0.015511523932218552,
+ "learning_rate": 0.0006,
+ "loss": 4.235037803649902,
+ "step": 5068
+ },
+ {
+ "epoch": 70.40524017467249,
+ "grad_norm": 0.014825738966464996,
+ "learning_rate": 0.0006,
+ "loss": 4.262448310852051,
+ "step": 5069
+ },
+ {
+ "epoch": 70.41921397379913,
+ "grad_norm": 0.015309952199459076,
+ "learning_rate": 0.0006,
+ "loss": 4.258711814880371,
+ "step": 5070
+ },
+ {
+ "epoch": 70.43318777292576,
+ "grad_norm": 0.017614055424928665,
+ "learning_rate": 0.0006,
+ "loss": 4.286028861999512,
+ "step": 5071
+ },
+ {
+ "epoch": 70.44716157205241,
+ "grad_norm": 0.0166039876639843,
+ "learning_rate": 0.0006,
+ "loss": 4.149197578430176,
+ "step": 5072
+ },
+ {
+ "epoch": 70.46113537117904,
+ "grad_norm": 0.016911419108510017,
+ "learning_rate": 0.0006,
+ "loss": 4.271981239318848,
+ "step": 5073
+ },
+ {
+ "epoch": 70.47510917030567,
+ "grad_norm": 0.016806820407509804,
+ "learning_rate": 0.0006,
+ "loss": 4.213315010070801,
+ "step": 5074
+ },
+ {
+ "epoch": 70.48908296943232,
+ "grad_norm": 0.015221742913126945,
+ "learning_rate": 0.0006,
+ "loss": 4.171091079711914,
+ "step": 5075
+ },
+ {
+ "epoch": 70.50305676855895,
+ "grad_norm": 0.014175347983837128,
+ "learning_rate": 0.0006,
+ "loss": 4.185309410095215,
+ "step": 5076
+ },
+ {
+ "epoch": 70.5170305676856,
+ "grad_norm": 0.013897798955440521,
+ "learning_rate": 0.0006,
+ "loss": 4.250119209289551,
+ "step": 5077
+ },
+ {
+ "epoch": 70.53100436681223,
+ "grad_norm": 0.014791003428399563,
+ "learning_rate": 0.0006,
+ "loss": 4.2504682540893555,
+ "step": 5078
+ },
+ {
+ "epoch": 70.54497816593886,
+ "grad_norm": 0.016700128093361855,
+ "learning_rate": 0.0006,
+ "loss": 4.235134124755859,
+ "step": 5079
+ },
+ {
+ "epoch": 70.5589519650655,
+ "grad_norm": 0.016679363325238228,
+ "learning_rate": 0.0006,
+ "loss": 4.32028865814209,
+ "step": 5080
+ },
+ {
+ "epoch": 70.57292576419214,
+ "grad_norm": 0.015117401257157326,
+ "learning_rate": 0.0006,
+ "loss": 4.210995674133301,
+ "step": 5081
+ },
+ {
+ "epoch": 70.58689956331878,
+ "grad_norm": 0.014441263861954212,
+ "learning_rate": 0.0006,
+ "loss": 4.27849006652832,
+ "step": 5082
+ },
+ {
+ "epoch": 70.60087336244541,
+ "grad_norm": 0.01523780357092619,
+ "learning_rate": 0.0006,
+ "loss": 4.197360992431641,
+ "step": 5083
+ },
+ {
+ "epoch": 70.61484716157206,
+ "grad_norm": 0.016201738268136978,
+ "learning_rate": 0.0006,
+ "loss": 4.28239107131958,
+ "step": 5084
+ },
+ {
+ "epoch": 70.62882096069869,
+ "grad_norm": 0.016003292053937912,
+ "learning_rate": 0.0006,
+ "loss": 4.209715843200684,
+ "step": 5085
+ },
+ {
+ "epoch": 70.64279475982532,
+ "grad_norm": 0.01472893450409174,
+ "learning_rate": 0.0006,
+ "loss": 4.169078826904297,
+ "step": 5086
+ },
+ {
+ "epoch": 70.65676855895197,
+ "grad_norm": 0.01604326069355011,
+ "learning_rate": 0.0006,
+ "loss": 4.090967178344727,
+ "step": 5087
+ },
+ {
+ "epoch": 70.6707423580786,
+ "grad_norm": 0.01598356105387211,
+ "learning_rate": 0.0006,
+ "loss": 4.16630220413208,
+ "step": 5088
+ },
+ {
+ "epoch": 70.68471615720524,
+ "grad_norm": 0.016988780349493027,
+ "learning_rate": 0.0006,
+ "loss": 4.117822170257568,
+ "step": 5089
+ },
+ {
+ "epoch": 70.69868995633188,
+ "grad_norm": 0.018582893535494804,
+ "learning_rate": 0.0006,
+ "loss": 4.192902088165283,
+ "step": 5090
+ },
+ {
+ "epoch": 70.71266375545852,
+ "grad_norm": 0.01793155074119568,
+ "learning_rate": 0.0006,
+ "loss": 4.314806938171387,
+ "step": 5091
+ },
+ {
+ "epoch": 70.72663755458515,
+ "grad_norm": 0.016286427155137062,
+ "learning_rate": 0.0006,
+ "loss": 4.2137556076049805,
+ "step": 5092
+ },
+ {
+ "epoch": 70.74061135371178,
+ "grad_norm": 0.014718376100063324,
+ "learning_rate": 0.0006,
+ "loss": 4.3073015213012695,
+ "step": 5093
+ },
+ {
+ "epoch": 70.75458515283843,
+ "grad_norm": 0.014675077050924301,
+ "learning_rate": 0.0006,
+ "loss": 4.160961151123047,
+ "step": 5094
+ },
+ {
+ "epoch": 70.76855895196506,
+ "grad_norm": 0.015550913289189339,
+ "learning_rate": 0.0006,
+ "loss": 4.322887420654297,
+ "step": 5095
+ },
+ {
+ "epoch": 70.78253275109171,
+ "grad_norm": 0.01453235000371933,
+ "learning_rate": 0.0006,
+ "loss": 4.2915873527526855,
+ "step": 5096
+ },
+ {
+ "epoch": 70.79650655021834,
+ "grad_norm": 0.01459164172410965,
+ "learning_rate": 0.0006,
+ "loss": 4.247288227081299,
+ "step": 5097
+ },
+ {
+ "epoch": 70.81048034934499,
+ "grad_norm": 0.015395080670714378,
+ "learning_rate": 0.0006,
+ "loss": 4.360955715179443,
+ "step": 5098
+ },
+ {
+ "epoch": 70.82445414847162,
+ "grad_norm": 0.01317683607339859,
+ "learning_rate": 0.0006,
+ "loss": 4.181969165802002,
+ "step": 5099
+ },
+ {
+ "epoch": 70.83842794759825,
+ "grad_norm": 0.013917932286858559,
+ "learning_rate": 0.0006,
+ "loss": 4.264626979827881,
+ "step": 5100
+ },
+ {
+ "epoch": 70.8524017467249,
+ "grad_norm": 0.015268695540726185,
+ "learning_rate": 0.0006,
+ "loss": 4.287857532501221,
+ "step": 5101
+ },
+ {
+ "epoch": 70.86637554585153,
+ "grad_norm": 0.015531329438090324,
+ "learning_rate": 0.0006,
+ "loss": 4.294188499450684,
+ "step": 5102
+ },
+ {
+ "epoch": 70.88034934497817,
+ "grad_norm": 0.015607129782438278,
+ "learning_rate": 0.0006,
+ "loss": 4.331040859222412,
+ "step": 5103
+ },
+ {
+ "epoch": 70.8943231441048,
+ "grad_norm": 0.015297149308025837,
+ "learning_rate": 0.0006,
+ "loss": 4.243650436401367,
+ "step": 5104
+ },
+ {
+ "epoch": 70.90829694323143,
+ "grad_norm": 0.014790025539696217,
+ "learning_rate": 0.0006,
+ "loss": 4.159859657287598,
+ "step": 5105
+ },
+ {
+ "epoch": 70.92227074235808,
+ "grad_norm": 0.013466274365782738,
+ "learning_rate": 0.0006,
+ "loss": 4.221992492675781,
+ "step": 5106
+ },
+ {
+ "epoch": 70.93624454148471,
+ "grad_norm": 0.01361898984760046,
+ "learning_rate": 0.0006,
+ "loss": 4.278751373291016,
+ "step": 5107
+ },
+ {
+ "epoch": 70.95021834061136,
+ "grad_norm": 0.01325751468539238,
+ "learning_rate": 0.0006,
+ "loss": 4.166001319885254,
+ "step": 5108
+ },
+ {
+ "epoch": 70.96419213973799,
+ "grad_norm": 0.012727087363600731,
+ "learning_rate": 0.0006,
+ "loss": 4.219997406005859,
+ "step": 5109
+ },
+ {
+ "epoch": 70.97816593886463,
+ "grad_norm": 0.013966867700219154,
+ "learning_rate": 0.0006,
+ "loss": 4.207491397857666,
+ "step": 5110
+ },
+ {
+ "epoch": 70.99213973799127,
+ "grad_norm": 0.015053587965667248,
+ "learning_rate": 0.0006,
+ "loss": 4.343245983123779,
+ "step": 5111
+ },
+ {
+ "epoch": 71.0,
+ "grad_norm": 0.017723368480801582,
+ "learning_rate": 0.0006,
+ "loss": 4.356122016906738,
+ "step": 5112
+ },
+ {
+ "epoch": 71.0,
+ "eval_loss": 4.598303318023682,
+ "eval_runtime": 56.7521,
+ "eval_samples_per_second": 43.029,
+ "eval_steps_per_second": 1.357,
+ "step": 5112
+ },
+ {
+ "epoch": 71.01397379912663,
+ "grad_norm": 0.018996279686689377,
+ "learning_rate": 0.0006,
+ "loss": 4.234772205352783,
+ "step": 5113
+ },
+ {
+ "epoch": 71.02794759825328,
+ "grad_norm": 0.02041555754840374,
+ "learning_rate": 0.0006,
+ "loss": 4.275816917419434,
+ "step": 5114
+ },
+ {
+ "epoch": 71.04192139737991,
+ "grad_norm": 0.021955285221338272,
+ "learning_rate": 0.0006,
+ "loss": 4.166038513183594,
+ "step": 5115
+ },
+ {
+ "epoch": 71.05589519650655,
+ "grad_norm": 0.021308740600943565,
+ "learning_rate": 0.0006,
+ "loss": 4.13063907623291,
+ "step": 5116
+ },
+ {
+ "epoch": 71.06986899563319,
+ "grad_norm": 0.01747054047882557,
+ "learning_rate": 0.0006,
+ "loss": 4.200522422790527,
+ "step": 5117
+ },
+ {
+ "epoch": 71.08384279475983,
+ "grad_norm": 0.018544532358646393,
+ "learning_rate": 0.0006,
+ "loss": 4.173915386199951,
+ "step": 5118
+ },
+ {
+ "epoch": 71.09781659388646,
+ "grad_norm": 0.0182361900806427,
+ "learning_rate": 0.0006,
+ "loss": 4.27392578125,
+ "step": 5119
+ },
+ {
+ "epoch": 71.1117903930131,
+ "grad_norm": 0.017062073573470116,
+ "learning_rate": 0.0006,
+ "loss": 4.219576358795166,
+ "step": 5120
+ },
+ {
+ "epoch": 71.12576419213974,
+ "grad_norm": 0.016289666295051575,
+ "learning_rate": 0.0006,
+ "loss": 4.133376598358154,
+ "step": 5121
+ },
+ {
+ "epoch": 71.13973799126637,
+ "grad_norm": 0.014864642173051834,
+ "learning_rate": 0.0006,
+ "loss": 4.168276786804199,
+ "step": 5122
+ },
+ {
+ "epoch": 71.15371179039302,
+ "grad_norm": 0.015309540554881096,
+ "learning_rate": 0.0006,
+ "loss": 4.175103187561035,
+ "step": 5123
+ },
+ {
+ "epoch": 71.16768558951965,
+ "grad_norm": 0.01556059904396534,
+ "learning_rate": 0.0006,
+ "loss": 4.289899826049805,
+ "step": 5124
+ },
+ {
+ "epoch": 71.18165938864628,
+ "grad_norm": 0.01479957066476345,
+ "learning_rate": 0.0006,
+ "loss": 4.341495037078857,
+ "step": 5125
+ },
+ {
+ "epoch": 71.19563318777293,
+ "grad_norm": 0.014762179926037788,
+ "learning_rate": 0.0006,
+ "loss": 4.252333641052246,
+ "step": 5126
+ },
+ {
+ "epoch": 71.20960698689956,
+ "grad_norm": 0.014897473156452179,
+ "learning_rate": 0.0006,
+ "loss": 4.2244181632995605,
+ "step": 5127
+ },
+ {
+ "epoch": 71.2235807860262,
+ "grad_norm": 0.012990163639187813,
+ "learning_rate": 0.0006,
+ "loss": 4.237288475036621,
+ "step": 5128
+ },
+ {
+ "epoch": 71.23755458515284,
+ "grad_norm": 0.012858416885137558,
+ "learning_rate": 0.0006,
+ "loss": 4.265738487243652,
+ "step": 5129
+ },
+ {
+ "epoch": 71.25152838427948,
+ "grad_norm": 0.013454788364470005,
+ "learning_rate": 0.0006,
+ "loss": 4.289131164550781,
+ "step": 5130
+ },
+ {
+ "epoch": 71.26550218340611,
+ "grad_norm": 0.013717238791286945,
+ "learning_rate": 0.0006,
+ "loss": 4.2844014167785645,
+ "step": 5131
+ },
+ {
+ "epoch": 71.27947598253274,
+ "grad_norm": 0.0135411461815238,
+ "learning_rate": 0.0006,
+ "loss": 4.139082908630371,
+ "step": 5132
+ },
+ {
+ "epoch": 71.29344978165939,
+ "grad_norm": 0.01406062301248312,
+ "learning_rate": 0.0006,
+ "loss": 4.226719856262207,
+ "step": 5133
+ },
+ {
+ "epoch": 71.30742358078602,
+ "grad_norm": 0.014027190394699574,
+ "learning_rate": 0.0006,
+ "loss": 4.191699981689453,
+ "step": 5134
+ },
+ {
+ "epoch": 71.32139737991267,
+ "grad_norm": 0.013809907250106335,
+ "learning_rate": 0.0006,
+ "loss": 4.242542266845703,
+ "step": 5135
+ },
+ {
+ "epoch": 71.3353711790393,
+ "grad_norm": 0.013507804833352566,
+ "learning_rate": 0.0006,
+ "loss": 4.068206310272217,
+ "step": 5136
+ },
+ {
+ "epoch": 71.34934497816595,
+ "grad_norm": 0.01432533748447895,
+ "learning_rate": 0.0006,
+ "loss": 4.378222942352295,
+ "step": 5137
+ },
+ {
+ "epoch": 71.36331877729258,
+ "grad_norm": 0.01753002591431141,
+ "learning_rate": 0.0006,
+ "loss": 4.174706935882568,
+ "step": 5138
+ },
+ {
+ "epoch": 71.37729257641921,
+ "grad_norm": 0.020048344507813454,
+ "learning_rate": 0.0006,
+ "loss": 4.19920015335083,
+ "step": 5139
+ },
+ {
+ "epoch": 71.39126637554585,
+ "grad_norm": 0.018551401793956757,
+ "learning_rate": 0.0006,
+ "loss": 4.175182342529297,
+ "step": 5140
+ },
+ {
+ "epoch": 71.40524017467249,
+ "grad_norm": 0.01598959229886532,
+ "learning_rate": 0.0006,
+ "loss": 4.2305169105529785,
+ "step": 5141
+ },
+ {
+ "epoch": 71.41921397379913,
+ "grad_norm": 0.016029560938477516,
+ "learning_rate": 0.0006,
+ "loss": 4.230090618133545,
+ "step": 5142
+ },
+ {
+ "epoch": 71.43318777292576,
+ "grad_norm": 0.01595657505095005,
+ "learning_rate": 0.0006,
+ "loss": 4.138402938842773,
+ "step": 5143
+ },
+ {
+ "epoch": 71.44716157205241,
+ "grad_norm": 0.01623230241239071,
+ "learning_rate": 0.0006,
+ "loss": 4.196106433868408,
+ "step": 5144
+ },
+ {
+ "epoch": 71.46113537117904,
+ "grad_norm": 0.014057597145438194,
+ "learning_rate": 0.0006,
+ "loss": 4.2599897384643555,
+ "step": 5145
+ },
+ {
+ "epoch": 71.47510917030567,
+ "grad_norm": 0.017161158844828606,
+ "learning_rate": 0.0006,
+ "loss": 4.183909893035889,
+ "step": 5146
+ },
+ {
+ "epoch": 71.48908296943232,
+ "grad_norm": 0.017038581892848015,
+ "learning_rate": 0.0006,
+ "loss": 4.373933792114258,
+ "step": 5147
+ },
+ {
+ "epoch": 71.50305676855895,
+ "grad_norm": 0.015851367264986038,
+ "learning_rate": 0.0006,
+ "loss": 4.2620086669921875,
+ "step": 5148
+ },
+ {
+ "epoch": 71.5170305676856,
+ "grad_norm": 0.014598303474485874,
+ "learning_rate": 0.0006,
+ "loss": 4.197151184082031,
+ "step": 5149
+ },
+ {
+ "epoch": 71.53100436681223,
+ "grad_norm": 0.01339802611619234,
+ "learning_rate": 0.0006,
+ "loss": 4.110706329345703,
+ "step": 5150
+ },
+ {
+ "epoch": 71.54497816593886,
+ "grad_norm": 0.015567775815725327,
+ "learning_rate": 0.0006,
+ "loss": 4.1210246086120605,
+ "step": 5151
+ },
+ {
+ "epoch": 71.5589519650655,
+ "grad_norm": 0.015616007149219513,
+ "learning_rate": 0.0006,
+ "loss": 4.129796028137207,
+ "step": 5152
+ },
+ {
+ "epoch": 71.57292576419214,
+ "grad_norm": 0.014198964461684227,
+ "learning_rate": 0.0006,
+ "loss": 4.193232536315918,
+ "step": 5153
+ },
+ {
+ "epoch": 71.58689956331878,
+ "grad_norm": 0.015226383693516254,
+ "learning_rate": 0.0006,
+ "loss": 4.223779678344727,
+ "step": 5154
+ },
+ {
+ "epoch": 71.60087336244541,
+ "grad_norm": 0.015610109083354473,
+ "learning_rate": 0.0006,
+ "loss": 4.223179817199707,
+ "step": 5155
+ },
+ {
+ "epoch": 71.61484716157206,
+ "grad_norm": 0.016034303233027458,
+ "learning_rate": 0.0006,
+ "loss": 4.196593284606934,
+ "step": 5156
+ },
+ {
+ "epoch": 71.62882096069869,
+ "grad_norm": 0.015648581087589264,
+ "learning_rate": 0.0006,
+ "loss": 4.341497421264648,
+ "step": 5157
+ },
+ {
+ "epoch": 71.64279475982532,
+ "grad_norm": 0.014117540791630745,
+ "learning_rate": 0.0006,
+ "loss": 4.194307327270508,
+ "step": 5158
+ },
+ {
+ "epoch": 71.65676855895197,
+ "grad_norm": 0.015612686984241009,
+ "learning_rate": 0.0006,
+ "loss": 4.290152549743652,
+ "step": 5159
+ },
+ {
+ "epoch": 71.6707423580786,
+ "grad_norm": 0.01722259446978569,
+ "learning_rate": 0.0006,
+ "loss": 4.258018493652344,
+ "step": 5160
+ },
+ {
+ "epoch": 71.68471615720524,
+ "grad_norm": 0.015923351049423218,
+ "learning_rate": 0.0006,
+ "loss": 4.21684455871582,
+ "step": 5161
+ },
+ {
+ "epoch": 71.69868995633188,
+ "grad_norm": 0.013663450255990028,
+ "learning_rate": 0.0006,
+ "loss": 4.175601959228516,
+ "step": 5162
+ },
+ {
+ "epoch": 71.71266375545852,
+ "grad_norm": 0.017075596377253532,
+ "learning_rate": 0.0006,
+ "loss": 4.205138683319092,
+ "step": 5163
+ },
+ {
+ "epoch": 71.72663755458515,
+ "grad_norm": 0.01723487675189972,
+ "learning_rate": 0.0006,
+ "loss": 4.08839225769043,
+ "step": 5164
+ },
+ {
+ "epoch": 71.74061135371178,
+ "grad_norm": 0.017355505377054214,
+ "learning_rate": 0.0006,
+ "loss": 4.165130615234375,
+ "step": 5165
+ },
+ {
+ "epoch": 71.75458515283843,
+ "grad_norm": 0.01868261583149433,
+ "learning_rate": 0.0006,
+ "loss": 4.16977071762085,
+ "step": 5166
+ },
+ {
+ "epoch": 71.76855895196506,
+ "grad_norm": 0.01897321827709675,
+ "learning_rate": 0.0006,
+ "loss": 4.173497200012207,
+ "step": 5167
+ },
+ {
+ "epoch": 71.78253275109171,
+ "grad_norm": 0.02045825682580471,
+ "learning_rate": 0.0006,
+ "loss": 4.176192283630371,
+ "step": 5168
+ },
+ {
+ "epoch": 71.79650655021834,
+ "grad_norm": 0.022895624861121178,
+ "learning_rate": 0.0006,
+ "loss": 4.163009166717529,
+ "step": 5169
+ },
+ {
+ "epoch": 71.81048034934499,
+ "grad_norm": 0.02062244527041912,
+ "learning_rate": 0.0006,
+ "loss": 4.156868934631348,
+ "step": 5170
+ },
+ {
+ "epoch": 71.82445414847162,
+ "grad_norm": 0.0176058579236269,
+ "learning_rate": 0.0006,
+ "loss": 4.154374122619629,
+ "step": 5171
+ },
+ {
+ "epoch": 71.83842794759825,
+ "grad_norm": 0.018049508333206177,
+ "learning_rate": 0.0006,
+ "loss": 4.296828269958496,
+ "step": 5172
+ },
+ {
+ "epoch": 71.8524017467249,
+ "grad_norm": 0.019466130062937737,
+ "learning_rate": 0.0006,
+ "loss": 4.19450569152832,
+ "step": 5173
+ },
+ {
+ "epoch": 71.86637554585153,
+ "grad_norm": 0.01838165521621704,
+ "learning_rate": 0.0006,
+ "loss": 4.11812686920166,
+ "step": 5174
+ },
+ {
+ "epoch": 71.88034934497817,
+ "grad_norm": 0.01608051173388958,
+ "learning_rate": 0.0006,
+ "loss": 4.240464210510254,
+ "step": 5175
+ },
+ {
+ "epoch": 71.8943231441048,
+ "grad_norm": 0.015807967633008957,
+ "learning_rate": 0.0006,
+ "loss": 4.275581359863281,
+ "step": 5176
+ },
+ {
+ "epoch": 71.90829694323143,
+ "grad_norm": 0.017366575077176094,
+ "learning_rate": 0.0006,
+ "loss": 4.064750671386719,
+ "step": 5177
+ },
+ {
+ "epoch": 71.92227074235808,
+ "grad_norm": 0.016241202130913734,
+ "learning_rate": 0.0006,
+ "loss": 4.203697681427002,
+ "step": 5178
+ },
+ {
+ "epoch": 71.93624454148471,
+ "grad_norm": 0.019105732440948486,
+ "learning_rate": 0.0006,
+ "loss": 4.140021800994873,
+ "step": 5179
+ },
+ {
+ "epoch": 71.95021834061136,
+ "grad_norm": 0.018492674455046654,
+ "learning_rate": 0.0006,
+ "loss": 4.2534074783325195,
+ "step": 5180
+ },
+ {
+ "epoch": 71.96419213973799,
+ "grad_norm": 0.019301170483231544,
+ "learning_rate": 0.0006,
+ "loss": 4.381392478942871,
+ "step": 5181
+ },
+ {
+ "epoch": 71.97816593886463,
+ "grad_norm": 0.019571911543607712,
+ "learning_rate": 0.0006,
+ "loss": 4.131056785583496,
+ "step": 5182
+ },
+ {
+ "epoch": 71.99213973799127,
+ "grad_norm": 0.01859748177230358,
+ "learning_rate": 0.0006,
+ "loss": 4.278930187225342,
+ "step": 5183
+ },
+ {
+ "epoch": 72.0,
+ "grad_norm": 0.02017994411289692,
+ "learning_rate": 0.0006,
+ "loss": 4.116464614868164,
+ "step": 5184
+ },
+ {
+ "epoch": 72.0,
+ "eval_loss": 4.60784912109375,
+ "eval_runtime": 56.3242,
+ "eval_samples_per_second": 43.356,
+ "eval_steps_per_second": 1.367,
+ "step": 5184
+ },
+ {
+ "epoch": 72.01397379912663,
+ "grad_norm": 0.030090296640992165,
+ "learning_rate": 0.0006,
+ "loss": 4.218588352203369,
+ "step": 5185
+ },
+ {
+ "epoch": 72.02794759825328,
+ "grad_norm": 1.330443024635315,
+ "learning_rate": 0.0006,
+ "loss": 4.417870044708252,
+ "step": 5186
+ },
+ {
+ "epoch": 72.04192139737991,
+ "grad_norm": 0.6991518139839172,
+ "learning_rate": 0.0006,
+ "loss": 8.414234161376953,
+ "step": 5187
+ },
+ {
+ "epoch": 72.05589519650655,
+ "grad_norm": 0.2631438374519348,
+ "learning_rate": 0.0006,
+ "loss": 7.41729736328125,
+ "step": 5188
+ },
+ {
+ "epoch": 72.06986899563319,
+ "grad_norm": 0.12391754984855652,
+ "learning_rate": 0.0006,
+ "loss": 7.209164619445801,
+ "step": 5189
+ },
+ {
+ "epoch": 72.08384279475983,
+ "grad_norm": 0.13519170880317688,
+ "learning_rate": 0.0006,
+ "loss": 6.943539619445801,
+ "step": 5190
+ },
+ {
+ "epoch": 72.09781659388646,
+ "grad_norm": 1.258439302444458,
+ "learning_rate": 0.0006,
+ "loss": 10.267714500427246,
+ "step": 5191
+ },
+ {
+ "epoch": 72.1117903930131,
+ "grad_norm": 1.0296908617019653,
+ "learning_rate": 0.0006,
+ "loss": 9.51768970489502,
+ "step": 5192
+ },
+ {
+ "epoch": 72.12576419213974,
+ "grad_norm": 0.12273016571998596,
+ "learning_rate": 0.0006,
+ "loss": 7.634477615356445,
+ "step": 5193
+ },
+ {
+ "epoch": 72.13973799126637,
+ "grad_norm": 0.06170198693871498,
+ "learning_rate": 0.0006,
+ "loss": 7.592115879058838,
+ "step": 5194
+ },
+ {
+ "epoch": 72.15371179039302,
+ "grad_norm": 0.06295108795166016,
+ "learning_rate": 0.0006,
+ "loss": 7.576314449310303,
+ "step": 5195
+ },
+ {
+ "epoch": 72.16768558951965,
+ "grad_norm": 0.0666993111371994,
+ "learning_rate": 0.0006,
+ "loss": 7.53862190246582,
+ "step": 5196
+ },
+ {
+ "epoch": 72.18165938864628,
+ "grad_norm": 0.052215222269296646,
+ "learning_rate": 0.0006,
+ "loss": 7.502930641174316,
+ "step": 5197
+ },
+ {
+ "epoch": 72.19563318777293,
+ "grad_norm": 0.04978961497545242,
+ "learning_rate": 0.0006,
+ "loss": 7.471630096435547,
+ "step": 5198
+ },
+ {
+ "epoch": 72.20960698689956,
+ "grad_norm": 0.07275046408176422,
+ "learning_rate": 0.0006,
+ "loss": 7.496924877166748,
+ "step": 5199
+ },
+ {
+ "epoch": 72.2235807860262,
+ "grad_norm": 0.04491819068789482,
+ "learning_rate": 0.0006,
+ "loss": 7.424494743347168,
+ "step": 5200
+ },
+ {
+ "epoch": 72.23755458515284,
+ "grad_norm": 0.043718695640563965,
+ "learning_rate": 0.0006,
+ "loss": 7.441090106964111,
+ "step": 5201
+ },
+ {
+ "epoch": 72.25152838427948,
+ "grad_norm": 0.04483497887849808,
+ "learning_rate": 0.0006,
+ "loss": 7.3753862380981445,
+ "step": 5202
+ },
+ {
+ "epoch": 72.26550218340611,
+ "grad_norm": 0.03463944420218468,
+ "learning_rate": 0.0006,
+ "loss": 7.375782012939453,
+ "step": 5203
+ },
+ {
+ "epoch": 72.27947598253274,
+ "grad_norm": 0.0397045724093914,
+ "learning_rate": 0.0006,
+ "loss": 7.358898639678955,
+ "step": 5204
+ },
+ {
+ "epoch": 72.29344978165939,
+ "grad_norm": 0.029339438304305077,
+ "learning_rate": 0.0006,
+ "loss": 7.3774871826171875,
+ "step": 5205
+ },
+ {
+ "epoch": 72.30742358078602,
+ "grad_norm": 0.025469699874520302,
+ "learning_rate": 0.0006,
+ "loss": 7.324203014373779,
+ "step": 5206
+ },
+ {
+ "epoch": 72.32139737991267,
+ "grad_norm": 0.030497785657644272,
+ "learning_rate": 0.0006,
+ "loss": 7.329351902008057,
+ "step": 5207
+ },
+ {
+ "epoch": 72.3353711790393,
+ "grad_norm": 0.03382078558206558,
+ "learning_rate": 0.0006,
+ "loss": 7.289940357208252,
+ "step": 5208
+ },
+ {
+ "epoch": 72.34934497816595,
+ "grad_norm": 0.019869405776262283,
+ "learning_rate": 0.0006,
+ "loss": 7.272582054138184,
+ "step": 5209
+ },
+ {
+ "epoch": 72.36331877729258,
+ "grad_norm": 0.02385753206908703,
+ "learning_rate": 0.0006,
+ "loss": 7.2654266357421875,
+ "step": 5210
+ },
+ {
+ "epoch": 72.37729257641921,
+ "grad_norm": 0.02794312685728073,
+ "learning_rate": 0.0006,
+ "loss": 7.250753402709961,
+ "step": 5211
+ },
+ {
+ "epoch": 72.39126637554585,
+ "grad_norm": 0.02445952221751213,
+ "learning_rate": 0.0006,
+ "loss": 7.293970108032227,
+ "step": 5212
+ },
+ {
+ "epoch": 72.40524017467249,
+ "grad_norm": 0.019660986959934235,
+ "learning_rate": 0.0006,
+ "loss": 7.1948771476745605,
+ "step": 5213
+ },
+ {
+ "epoch": 72.41921397379913,
+ "grad_norm": 0.024110861122608185,
+ "learning_rate": 0.0006,
+ "loss": 7.224499225616455,
+ "step": 5214
+ },
+ {
+ "epoch": 72.43318777292576,
+ "grad_norm": 0.021630609408020973,
+ "learning_rate": 0.0006,
+ "loss": 7.1736884117126465,
+ "step": 5215
+ },
+ {
+ "epoch": 72.44716157205241,
+ "grad_norm": 0.02185880020260811,
+ "learning_rate": 0.0006,
+ "loss": 7.167229652404785,
+ "step": 5216
+ },
+ {
+ "epoch": 72.46113537117904,
+ "grad_norm": 0.018913568928837776,
+ "learning_rate": 0.0006,
+ "loss": 7.196551322937012,
+ "step": 5217
+ },
+ {
+ "epoch": 72.47510917030567,
+ "grad_norm": 0.01802486553788185,
+ "learning_rate": 0.0006,
+ "loss": 7.144286632537842,
+ "step": 5218
+ },
+ {
+ "epoch": 72.48908296943232,
+ "grad_norm": 0.020386485382914543,
+ "learning_rate": 0.0006,
+ "loss": 7.196260929107666,
+ "step": 5219
+ },
+ {
+ "epoch": 72.50305676855895,
+ "grad_norm": 0.020077696070075035,
+ "learning_rate": 0.0006,
+ "loss": 7.138718605041504,
+ "step": 5220
+ },
+ {
+ "epoch": 72.5170305676856,
+ "grad_norm": 0.017224183306097984,
+ "learning_rate": 0.0006,
+ "loss": 7.119478702545166,
+ "step": 5221
+ },
+ {
+ "epoch": 72.53100436681223,
+ "grad_norm": 0.020847434177994728,
+ "learning_rate": 0.0006,
+ "loss": 7.068922996520996,
+ "step": 5222
+ },
+ {
+ "epoch": 72.54497816593886,
+ "grad_norm": 0.021764913573861122,
+ "learning_rate": 0.0006,
+ "loss": 7.071352958679199,
+ "step": 5223
+ },
+ {
+ "epoch": 72.5589519650655,
+ "grad_norm": 0.02697215974330902,
+ "learning_rate": 0.0006,
+ "loss": 7.033344268798828,
+ "step": 5224
+ },
+ {
+ "epoch": 72.57292576419214,
+ "grad_norm": 0.042392998933792114,
+ "learning_rate": 0.0006,
+ "loss": 7.0589919090271,
+ "step": 5225
+ },
+ {
+ "epoch": 72.58689956331878,
+ "grad_norm": 0.06616316735744476,
+ "learning_rate": 0.0006,
+ "loss": 7.060244083404541,
+ "step": 5226
+ },
+ {
+ "epoch": 72.60087336244541,
+ "grad_norm": 0.0828910544514656,
+ "learning_rate": 0.0006,
+ "loss": 7.057173728942871,
+ "step": 5227
+ },
+ {
+ "epoch": 72.61484716157206,
+ "grad_norm": 0.046767015010118484,
+ "learning_rate": 0.0006,
+ "loss": 7.010563850402832,
+ "step": 5228
+ },
+ {
+ "epoch": 72.62882096069869,
+ "grad_norm": 0.03831760957837105,
+ "learning_rate": 0.0006,
+ "loss": 6.9028825759887695,
+ "step": 5229
+ },
+ {
+ "epoch": 72.64279475982532,
+ "grad_norm": 0.05109567195177078,
+ "learning_rate": 0.0006,
+ "loss": 6.923257827758789,
+ "step": 5230
+ },
+ {
+ "epoch": 72.65676855895197,
+ "grad_norm": 0.02897934429347515,
+ "learning_rate": 0.0006,
+ "loss": 6.823014259338379,
+ "step": 5231
+ },
+ {
+ "epoch": 72.6707423580786,
+ "grad_norm": 0.03769192099571228,
+ "learning_rate": 0.0006,
+ "loss": 6.87416410446167,
+ "step": 5232
+ },
+ {
+ "epoch": 72.68471615720524,
+ "grad_norm": 0.03677637130022049,
+ "learning_rate": 0.0006,
+ "loss": 6.7649030685424805,
+ "step": 5233
+ },
+ {
+ "epoch": 72.69868995633188,
+ "grad_norm": 0.0371948778629303,
+ "learning_rate": 0.0006,
+ "loss": 6.749821662902832,
+ "step": 5234
+ },
+ {
+ "epoch": 72.71266375545852,
+ "grad_norm": 0.03057795763015747,
+ "learning_rate": 0.0006,
+ "loss": 6.748945236206055,
+ "step": 5235
+ },
+ {
+ "epoch": 72.72663755458515,
+ "grad_norm": 0.028109321370720863,
+ "learning_rate": 0.0006,
+ "loss": 6.663548946380615,
+ "step": 5236
+ },
+ {
+ "epoch": 72.74061135371178,
+ "grad_norm": 0.02442420832812786,
+ "learning_rate": 0.0006,
+ "loss": 6.636251449584961,
+ "step": 5237
+ },
+ {
+ "epoch": 72.75458515283843,
+ "grad_norm": 0.031841423362493515,
+ "learning_rate": 0.0006,
+ "loss": 6.606883525848389,
+ "step": 5238
+ },
+ {
+ "epoch": 72.76855895196506,
+ "grad_norm": 0.028100458905100822,
+ "learning_rate": 0.0006,
+ "loss": 6.603745937347412,
+ "step": 5239
+ },
+ {
+ "epoch": 72.78253275109171,
+ "grad_norm": 0.03222901374101639,
+ "learning_rate": 0.0006,
+ "loss": 6.463825225830078,
+ "step": 5240
+ },
+ {
+ "epoch": 72.79650655021834,
+ "grad_norm": 0.04766521975398064,
+ "learning_rate": 0.0006,
+ "loss": 6.43734073638916,
+ "step": 5241
+ },
+ {
+ "epoch": 72.81048034934499,
+ "grad_norm": 0.09058783948421478,
+ "learning_rate": 0.0006,
+ "loss": 6.448520183563232,
+ "step": 5242
+ },
+ {
+ "epoch": 72.82445414847162,
+ "grad_norm": 0.08995158225297928,
+ "learning_rate": 0.0006,
+ "loss": 6.522093296051025,
+ "step": 5243
+ },
+ {
+ "epoch": 72.83842794759825,
+ "grad_norm": 0.05491040274500847,
+ "learning_rate": 0.0006,
+ "loss": 6.37354040145874,
+ "step": 5244
+ },
+ {
+ "epoch": 72.8524017467249,
+ "grad_norm": 0.05198514088988304,
+ "learning_rate": 0.0006,
+ "loss": 6.362032413482666,
+ "step": 5245
+ },
+ {
+ "epoch": 72.86637554585153,
+ "grad_norm": 0.05464765429496765,
+ "learning_rate": 0.0006,
+ "loss": 6.326773643493652,
+ "step": 5246
+ },
+ {
+ "epoch": 72.88034934497817,
+ "grad_norm": 0.07337361574172974,
+ "learning_rate": 0.0006,
+ "loss": 6.297532081604004,
+ "step": 5247
+ },
+ {
+ "epoch": 72.8943231441048,
+ "grad_norm": 0.10061230510473251,
+ "learning_rate": 0.0006,
+ "loss": 6.359557628631592,
+ "step": 5248
+ },
+ {
+ "epoch": 72.90829694323143,
+ "grad_norm": 0.10005959123373032,
+ "learning_rate": 0.0006,
+ "loss": 6.422270774841309,
+ "step": 5249
+ },
+ {
+ "epoch": 72.92227074235808,
+ "grad_norm": 0.08989301323890686,
+ "learning_rate": 0.0006,
+ "loss": 6.289371490478516,
+ "step": 5250
+ },
+ {
+ "epoch": 72.93624454148471,
+ "grad_norm": 0.06371799111366272,
+ "learning_rate": 0.0006,
+ "loss": 6.199618339538574,
+ "step": 5251
+ },
+ {
+ "epoch": 72.95021834061136,
+ "grad_norm": 0.05932944267988205,
+ "learning_rate": 0.0006,
+ "loss": 6.111152648925781,
+ "step": 5252
+ },
+ {
+ "epoch": 72.96419213973799,
+ "grad_norm": 0.04735502973198891,
+ "learning_rate": 0.0006,
+ "loss": 6.026026248931885,
+ "step": 5253
+ },
+ {
+ "epoch": 72.97816593886463,
+ "grad_norm": 0.04270688816905022,
+ "learning_rate": 0.0006,
+ "loss": 6.01606559753418,
+ "step": 5254
+ },
+ {
+ "epoch": 72.99213973799127,
+ "grad_norm": 0.03774143010377884,
+ "learning_rate": 0.0006,
+ "loss": 5.8935441970825195,
+ "step": 5255
+ },
+ {
+ "epoch": 73.0,
+ "grad_norm": 0.04313631355762482,
+ "learning_rate": 0.0006,
+ "loss": 5.842061996459961,
+ "step": 5256
+ },
+ {
+ "epoch": 73.0,
+ "eval_loss": 5.9985785484313965,
+ "eval_runtime": 56.3891,
+ "eval_samples_per_second": 43.306,
+ "eval_steps_per_second": 1.366,
+ "step": 5256
+ },
+ {
+ "epoch": 73.01397379912663,
+ "grad_norm": 0.045038655400276184,
+ "learning_rate": 0.0006,
+ "loss": 5.8209662437438965,
+ "step": 5257
+ },
+ {
+ "epoch": 73.02794759825328,
+ "grad_norm": 0.03706910088658333,
+ "learning_rate": 0.0006,
+ "loss": 5.806189060211182,
+ "step": 5258
+ },
+ {
+ "epoch": 73.04192139737991,
+ "grad_norm": 0.04051685333251953,
+ "learning_rate": 0.0006,
+ "loss": 5.706869125366211,
+ "step": 5259
+ },
+ {
+ "epoch": 73.05589519650655,
+ "grad_norm": 0.04849153012037277,
+ "learning_rate": 0.0006,
+ "loss": 5.696385383605957,
+ "step": 5260
+ },
+ {
+ "epoch": 73.06986899563319,
+ "grad_norm": 0.06124403700232506,
+ "learning_rate": 0.0006,
+ "loss": 5.618223667144775,
+ "step": 5261
+ },
+ {
+ "epoch": 73.08384279475983,
+ "grad_norm": 0.07375937700271606,
+ "learning_rate": 0.0006,
+ "loss": 5.631045341491699,
+ "step": 5262
+ },
+ {
+ "epoch": 73.09781659388646,
+ "grad_norm": 0.07413844019174576,
+ "learning_rate": 0.0006,
+ "loss": 5.6532182693481445,
+ "step": 5263
+ },
+ {
+ "epoch": 73.1117903930131,
+ "grad_norm": 0.0633770152926445,
+ "learning_rate": 0.0006,
+ "loss": 5.57209587097168,
+ "step": 5264
+ },
+ {
+ "epoch": 73.12576419213974,
+ "grad_norm": 0.05436766892671585,
+ "learning_rate": 0.0006,
+ "loss": 5.53944730758667,
+ "step": 5265
+ },
+ {
+ "epoch": 73.13973799126637,
+ "grad_norm": 0.04639481008052826,
+ "learning_rate": 0.0006,
+ "loss": 5.502948760986328,
+ "step": 5266
+ },
+ {
+ "epoch": 73.15371179039302,
+ "grad_norm": 0.07556558400392532,
+ "learning_rate": 0.0006,
+ "loss": 5.429069995880127,
+ "step": 5267
+ },
+ {
+ "epoch": 73.16768558951965,
+ "grad_norm": 0.1127929836511612,
+ "learning_rate": 0.0006,
+ "loss": 5.545061111450195,
+ "step": 5268
+ },
+ {
+ "epoch": 73.18165938864628,
+ "grad_norm": 0.07620745897293091,
+ "learning_rate": 0.0006,
+ "loss": 5.507102966308594,
+ "step": 5269
+ },
+ {
+ "epoch": 73.19563318777293,
+ "grad_norm": 0.05076538026332855,
+ "learning_rate": 0.0006,
+ "loss": 5.329682350158691,
+ "step": 5270
+ },
+ {
+ "epoch": 73.20960698689956,
+ "grad_norm": 0.05407547950744629,
+ "learning_rate": 0.0006,
+ "loss": 5.3048553466796875,
+ "step": 5271
+ },
+ {
+ "epoch": 73.2235807860262,
+ "grad_norm": 0.059095583856105804,
+ "learning_rate": 0.0006,
+ "loss": 5.292568683624268,
+ "step": 5272
+ },
+ {
+ "epoch": 73.23755458515284,
+ "grad_norm": 0.06193305924534798,
+ "learning_rate": 0.0006,
+ "loss": 5.1870927810668945,
+ "step": 5273
+ },
+ {
+ "epoch": 73.25152838427948,
+ "grad_norm": 0.062261562794446945,
+ "learning_rate": 0.0006,
+ "loss": 5.256819248199463,
+ "step": 5274
+ },
+ {
+ "epoch": 73.26550218340611,
+ "grad_norm": 0.0732516348361969,
+ "learning_rate": 0.0006,
+ "loss": 5.256464004516602,
+ "step": 5275
+ },
+ {
+ "epoch": 73.27947598253274,
+ "grad_norm": 0.12177979946136475,
+ "learning_rate": 0.0006,
+ "loss": 5.284239768981934,
+ "step": 5276
+ },
+ {
+ "epoch": 73.29344978165939,
+ "grad_norm": 0.10459653288125992,
+ "learning_rate": 0.0006,
+ "loss": 5.269713401794434,
+ "step": 5277
+ },
+ {
+ "epoch": 73.30742358078602,
+ "grad_norm": 0.07854694873094559,
+ "learning_rate": 0.0006,
+ "loss": 5.3245391845703125,
+ "step": 5278
+ },
+ {
+ "epoch": 73.32139737991267,
+ "grad_norm": 0.06030094623565674,
+ "learning_rate": 0.0006,
+ "loss": 5.033702850341797,
+ "step": 5279
+ },
+ {
+ "epoch": 73.3353711790393,
+ "grad_norm": 0.05168461427092552,
+ "learning_rate": 0.0006,
+ "loss": 5.140530109405518,
+ "step": 5280
+ },
+ {
+ "epoch": 73.34934497816595,
+ "grad_norm": 0.041064050048589706,
+ "learning_rate": 0.0006,
+ "loss": 5.1085896492004395,
+ "step": 5281
+ },
+ {
+ "epoch": 73.36331877729258,
+ "grad_norm": 0.03788219019770622,
+ "learning_rate": 0.0006,
+ "loss": 5.045009613037109,
+ "step": 5282
+ },
+ {
+ "epoch": 73.37729257641921,
+ "grad_norm": 0.03851071000099182,
+ "learning_rate": 0.0006,
+ "loss": 5.055008888244629,
+ "step": 5283
+ },
+ {
+ "epoch": 73.39126637554585,
+ "grad_norm": 0.03448539599776268,
+ "learning_rate": 0.0006,
+ "loss": 5.10657262802124,
+ "step": 5284
+ },
+ {
+ "epoch": 73.40524017467249,
+ "grad_norm": 0.026618091389536858,
+ "learning_rate": 0.0006,
+ "loss": 5.0342912673950195,
+ "step": 5285
+ },
+ {
+ "epoch": 73.41921397379913,
+ "grad_norm": 0.03060489147901535,
+ "learning_rate": 0.0006,
+ "loss": 5.022700786590576,
+ "step": 5286
+ },
+ {
+ "epoch": 73.43318777292576,
+ "grad_norm": 0.03362000361084938,
+ "learning_rate": 0.0006,
+ "loss": 4.891584396362305,
+ "step": 5287
+ },
+ {
+ "epoch": 73.44716157205241,
+ "grad_norm": 0.03887505829334259,
+ "learning_rate": 0.0006,
+ "loss": 4.772258758544922,
+ "step": 5288
+ },
+ {
+ "epoch": 73.46113537117904,
+ "grad_norm": 0.07157225161790848,
+ "learning_rate": 0.0006,
+ "loss": 4.997142791748047,
+ "step": 5289
+ },
+ {
+ "epoch": 73.47510917030567,
+ "grad_norm": 0.10180624574422836,
+ "learning_rate": 0.0006,
+ "loss": 4.908722877502441,
+ "step": 5290
+ },
+ {
+ "epoch": 73.48908296943232,
+ "grad_norm": 0.06529122591018677,
+ "learning_rate": 0.0006,
+ "loss": 4.98001766204834,
+ "step": 5291
+ },
+ {
+ "epoch": 73.50305676855895,
+ "grad_norm": 0.055171865969896317,
+ "learning_rate": 0.0006,
+ "loss": 4.877438068389893,
+ "step": 5292
+ },
+ {
+ "epoch": 73.5170305676856,
+ "grad_norm": 0.0318770632147789,
+ "learning_rate": 0.0006,
+ "loss": 4.836934566497803,
+ "step": 5293
+ },
+ {
+ "epoch": 73.53100436681223,
+ "grad_norm": 0.042072564363479614,
+ "learning_rate": 0.0006,
+ "loss": 4.919489860534668,
+ "step": 5294
+ },
+ {
+ "epoch": 73.54497816593886,
+ "grad_norm": 0.029197754338383675,
+ "learning_rate": 0.0006,
+ "loss": 4.805579662322998,
+ "step": 5295
+ },
+ {
+ "epoch": 73.5589519650655,
+ "grad_norm": 0.03249497339129448,
+ "learning_rate": 0.0006,
+ "loss": 4.704806327819824,
+ "step": 5296
+ },
+ {
+ "epoch": 73.57292576419214,
+ "grad_norm": 0.025050945580005646,
+ "learning_rate": 0.0006,
+ "loss": 4.936512470245361,
+ "step": 5297
+ },
+ {
+ "epoch": 73.58689956331878,
+ "grad_norm": 0.029073381796479225,
+ "learning_rate": 0.0006,
+ "loss": 4.757592678070068,
+ "step": 5298
+ },
+ {
+ "epoch": 73.60087336244541,
+ "grad_norm": 0.03465264290571213,
+ "learning_rate": 0.0006,
+ "loss": 4.729528427124023,
+ "step": 5299
+ },
+ {
+ "epoch": 73.61484716157206,
+ "grad_norm": 0.054795749485492706,
+ "learning_rate": 0.0006,
+ "loss": 4.77252197265625,
+ "step": 5300
+ },
+ {
+ "epoch": 73.62882096069869,
+ "grad_norm": 0.18129196763038635,
+ "learning_rate": 0.0006,
+ "loss": 4.901700019836426,
+ "step": 5301
+ },
+ {
+ "epoch": 73.64279475982532,
+ "grad_norm": 0.23274236917495728,
+ "learning_rate": 0.0006,
+ "loss": 5.207432270050049,
+ "step": 5302
+ },
+ {
+ "epoch": 73.65676855895197,
+ "grad_norm": 0.18162772059440613,
+ "learning_rate": 0.0006,
+ "loss": 5.326632976531982,
+ "step": 5303
+ },
+ {
+ "epoch": 73.6707423580786,
+ "grad_norm": 0.11858413368463516,
+ "learning_rate": 0.0006,
+ "loss": 5.3730058670043945,
+ "step": 5304
+ },
+ {
+ "epoch": 73.68471615720524,
+ "grad_norm": 0.06031268090009689,
+ "learning_rate": 0.0006,
+ "loss": 5.330936431884766,
+ "step": 5305
+ },
+ {
+ "epoch": 73.69868995633188,
+ "grad_norm": 0.059304628521203995,
+ "learning_rate": 0.0006,
+ "loss": 5.280514717102051,
+ "step": 5306
+ },
+ {
+ "epoch": 73.71266375545852,
+ "grad_norm": 0.04702131822705269,
+ "learning_rate": 0.0006,
+ "loss": 5.288274765014648,
+ "step": 5307
+ },
+ {
+ "epoch": 73.72663755458515,
+ "grad_norm": 0.037792228162288666,
+ "learning_rate": 0.0006,
+ "loss": 5.173289775848389,
+ "step": 5308
+ },
+ {
+ "epoch": 73.74061135371178,
+ "grad_norm": 0.037763576954603195,
+ "learning_rate": 0.0006,
+ "loss": 5.084990978240967,
+ "step": 5309
+ },
+ {
+ "epoch": 73.75458515283843,
+ "grad_norm": 0.03337346389889717,
+ "learning_rate": 0.0006,
+ "loss": 5.121417045593262,
+ "step": 5310
+ },
+ {
+ "epoch": 73.76855895196506,
+ "grad_norm": 0.03184802085161209,
+ "learning_rate": 0.0006,
+ "loss": 5.208918571472168,
+ "step": 5311
+ },
+ {
+ "epoch": 73.78253275109171,
+ "grad_norm": 0.02832048572599888,
+ "learning_rate": 0.0006,
+ "loss": 4.923543453216553,
+ "step": 5312
+ },
+ {
+ "epoch": 73.79650655021834,
+ "grad_norm": 0.028297102078795433,
+ "learning_rate": 0.0006,
+ "loss": 4.944367408752441,
+ "step": 5313
+ },
+ {
+ "epoch": 73.81048034934499,
+ "grad_norm": 0.027701659128069878,
+ "learning_rate": 0.0006,
+ "loss": 4.858168601989746,
+ "step": 5314
+ },
+ {
+ "epoch": 73.82445414847162,
+ "grad_norm": 0.02214639261364937,
+ "learning_rate": 0.0006,
+ "loss": 4.887052536010742,
+ "step": 5315
+ },
+ {
+ "epoch": 73.83842794759825,
+ "grad_norm": 0.01995093561708927,
+ "learning_rate": 0.0006,
+ "loss": 4.885797500610352,
+ "step": 5316
+ },
+ {
+ "epoch": 73.8524017467249,
+ "grad_norm": 0.020473985001444817,
+ "learning_rate": 0.0006,
+ "loss": 4.917325019836426,
+ "step": 5317
+ },
+ {
+ "epoch": 73.86637554585153,
+ "grad_norm": 0.0192868635058403,
+ "learning_rate": 0.0006,
+ "loss": 4.801746845245361,
+ "step": 5318
+ },
+ {
+ "epoch": 73.88034934497817,
+ "grad_norm": 0.017108267173171043,
+ "learning_rate": 0.0006,
+ "loss": 4.893620491027832,
+ "step": 5319
+ },
+ {
+ "epoch": 73.8943231441048,
+ "grad_norm": 0.01718071475625038,
+ "learning_rate": 0.0006,
+ "loss": 4.940695285797119,
+ "step": 5320
+ },
+ {
+ "epoch": 73.90829694323143,
+ "grad_norm": 0.017821557819843292,
+ "learning_rate": 0.0006,
+ "loss": 4.767613410949707,
+ "step": 5321
+ },
+ {
+ "epoch": 73.92227074235808,
+ "grad_norm": 0.015262950211763382,
+ "learning_rate": 0.0006,
+ "loss": 4.910315036773682,
+ "step": 5322
+ },
+ {
+ "epoch": 73.93624454148471,
+ "grad_norm": 0.01769006811082363,
+ "learning_rate": 0.0006,
+ "loss": 4.7743449211120605,
+ "step": 5323
+ },
+ {
+ "epoch": 73.95021834061136,
+ "grad_norm": 0.014460788108408451,
+ "learning_rate": 0.0006,
+ "loss": 4.743494987487793,
+ "step": 5324
+ },
+ {
+ "epoch": 73.96419213973799,
+ "grad_norm": 0.01403993833810091,
+ "learning_rate": 0.0006,
+ "loss": 4.666640281677246,
+ "step": 5325
+ },
+ {
+ "epoch": 73.97816593886463,
+ "grad_norm": 0.013997809961438179,
+ "learning_rate": 0.0006,
+ "loss": 4.698747158050537,
+ "step": 5326
+ },
+ {
+ "epoch": 73.99213973799127,
+ "grad_norm": 0.014244182966649532,
+ "learning_rate": 0.0006,
+ "loss": 4.745250701904297,
+ "step": 5327
+ },
+ {
+ "epoch": 74.0,
+ "grad_norm": 0.015144186094403267,
+ "learning_rate": 0.0006,
+ "loss": 4.667946815490723,
+ "step": 5328
+ },
+ {
+ "epoch": 74.0,
+ "eval_loss": 5.005440711975098,
+ "eval_runtime": 56.7836,
+ "eval_samples_per_second": 43.005,
+ "eval_steps_per_second": 1.356,
+ "step": 5328
+ },
+ {
+ "epoch": 74.01397379912663,
+ "grad_norm": 0.020384902134537697,
+ "learning_rate": 0.0006,
+ "loss": 4.626955509185791,
+ "step": 5329
+ },
+ {
+ "epoch": 74.02794759825328,
+ "grad_norm": 0.03624720871448517,
+ "learning_rate": 0.0006,
+ "loss": 4.713998317718506,
+ "step": 5330
+ },
+ {
+ "epoch": 74.04192139737991,
+ "grad_norm": 0.05701754242181778,
+ "learning_rate": 0.0006,
+ "loss": 4.717874526977539,
+ "step": 5331
+ },
+ {
+ "epoch": 74.05589519650655,
+ "grad_norm": 0.03942621126770973,
+ "learning_rate": 0.0006,
+ "loss": 4.7026190757751465,
+ "step": 5332
+ },
+ {
+ "epoch": 74.06986899563319,
+ "grad_norm": 0.017803272232413292,
+ "learning_rate": 0.0006,
+ "loss": 4.702548027038574,
+ "step": 5333
+ },
+ {
+ "epoch": 74.08384279475983,
+ "grad_norm": 0.017024677246809006,
+ "learning_rate": 0.0006,
+ "loss": 4.69752311706543,
+ "step": 5334
+ },
+ {
+ "epoch": 74.09781659388646,
+ "grad_norm": 0.016467854380607605,
+ "learning_rate": 0.0006,
+ "loss": 4.624876976013184,
+ "step": 5335
+ },
+ {
+ "epoch": 74.1117903930131,
+ "grad_norm": 0.016569215804338455,
+ "learning_rate": 0.0006,
+ "loss": 4.645040512084961,
+ "step": 5336
+ },
+ {
+ "epoch": 74.12576419213974,
+ "grad_norm": 0.012872067280113697,
+ "learning_rate": 0.0006,
+ "loss": 4.64582633972168,
+ "step": 5337
+ },
+ {
+ "epoch": 74.13973799126637,
+ "grad_norm": 0.015689387917518616,
+ "learning_rate": 0.0006,
+ "loss": 4.596040725708008,
+ "step": 5338
+ },
+ {
+ "epoch": 74.15371179039302,
+ "grad_norm": 0.013082082383334637,
+ "learning_rate": 0.0006,
+ "loss": 4.648033142089844,
+ "step": 5339
+ },
+ {
+ "epoch": 74.16768558951965,
+ "grad_norm": 0.013728172518312931,
+ "learning_rate": 0.0006,
+ "loss": 4.657873153686523,
+ "step": 5340
+ },
+ {
+ "epoch": 74.18165938864628,
+ "grad_norm": 0.013120594434440136,
+ "learning_rate": 0.0006,
+ "loss": 4.696074485778809,
+ "step": 5341
+ },
+ {
+ "epoch": 74.19563318777293,
+ "grad_norm": 0.01431245356798172,
+ "learning_rate": 0.0006,
+ "loss": 4.5679121017456055,
+ "step": 5342
+ },
+ {
+ "epoch": 74.20960698689956,
+ "grad_norm": 0.013994511216878891,
+ "learning_rate": 0.0006,
+ "loss": 4.5236406326293945,
+ "step": 5343
+ },
+ {
+ "epoch": 74.2235807860262,
+ "grad_norm": 0.011971707455813885,
+ "learning_rate": 0.0006,
+ "loss": 4.626547813415527,
+ "step": 5344
+ },
+ {
+ "epoch": 74.23755458515284,
+ "grad_norm": 0.012805287726223469,
+ "learning_rate": 0.0006,
+ "loss": 4.556386947631836,
+ "step": 5345
+ },
+ {
+ "epoch": 74.25152838427948,
+ "grad_norm": 0.011694584973156452,
+ "learning_rate": 0.0006,
+ "loss": 4.656987190246582,
+ "step": 5346
+ },
+ {
+ "epoch": 74.26550218340611,
+ "grad_norm": 0.013396821916103363,
+ "learning_rate": 0.0006,
+ "loss": 4.526467323303223,
+ "step": 5347
+ },
+ {
+ "epoch": 74.27947598253274,
+ "grad_norm": 0.014398031868040562,
+ "learning_rate": 0.0006,
+ "loss": 4.595100402832031,
+ "step": 5348
+ },
+ {
+ "epoch": 74.29344978165939,
+ "grad_norm": 0.02165800705552101,
+ "learning_rate": 0.0006,
+ "loss": 4.460549354553223,
+ "step": 5349
+ },
+ {
+ "epoch": 74.30742358078602,
+ "grad_norm": 0.03276512026786804,
+ "learning_rate": 0.0006,
+ "loss": 4.498189449310303,
+ "step": 5350
+ },
+ {
+ "epoch": 74.32139737991267,
+ "grad_norm": 0.04616224020719528,
+ "learning_rate": 0.0006,
+ "loss": 4.5268659591674805,
+ "step": 5351
+ },
+ {
+ "epoch": 74.3353711790393,
+ "grad_norm": 0.03863786533474922,
+ "learning_rate": 0.0006,
+ "loss": 4.554747104644775,
+ "step": 5352
+ },
+ {
+ "epoch": 74.34934497816595,
+ "grad_norm": 0.01749865710735321,
+ "learning_rate": 0.0006,
+ "loss": 4.6656036376953125,
+ "step": 5353
+ },
+ {
+ "epoch": 74.36331877729258,
+ "grad_norm": 0.021814396604895592,
+ "learning_rate": 0.0006,
+ "loss": 4.607456207275391,
+ "step": 5354
+ },
+ {
+ "epoch": 74.37729257641921,
+ "grad_norm": 0.017018496990203857,
+ "learning_rate": 0.0006,
+ "loss": 4.54205322265625,
+ "step": 5355
+ },
+ {
+ "epoch": 74.39126637554585,
+ "grad_norm": 0.018297340720891953,
+ "learning_rate": 0.0006,
+ "loss": 4.608870506286621,
+ "step": 5356
+ },
+ {
+ "epoch": 74.40524017467249,
+ "grad_norm": 0.01473482046276331,
+ "learning_rate": 0.0006,
+ "loss": 4.384891510009766,
+ "step": 5357
+ },
+ {
+ "epoch": 74.41921397379913,
+ "grad_norm": 0.01579415239393711,
+ "learning_rate": 0.0006,
+ "loss": 4.588859558105469,
+ "step": 5358
+ },
+ {
+ "epoch": 74.43318777292576,
+ "grad_norm": 0.013906026259064674,
+ "learning_rate": 0.0006,
+ "loss": 4.525175094604492,
+ "step": 5359
+ },
+ {
+ "epoch": 74.44716157205241,
+ "grad_norm": 0.014163361862301826,
+ "learning_rate": 0.0006,
+ "loss": 4.495802879333496,
+ "step": 5360
+ },
+ {
+ "epoch": 74.46113537117904,
+ "grad_norm": 0.014009335078299046,
+ "learning_rate": 0.0006,
+ "loss": 4.5960869789123535,
+ "step": 5361
+ },
+ {
+ "epoch": 74.47510917030567,
+ "grad_norm": 0.012765258550643921,
+ "learning_rate": 0.0006,
+ "loss": 4.515678405761719,
+ "step": 5362
+ },
+ {
+ "epoch": 74.48908296943232,
+ "grad_norm": 0.014071918092668056,
+ "learning_rate": 0.0006,
+ "loss": 4.394025802612305,
+ "step": 5363
+ },
+ {
+ "epoch": 74.50305676855895,
+ "grad_norm": 0.01244515273720026,
+ "learning_rate": 0.0006,
+ "loss": 4.576985836029053,
+ "step": 5364
+ },
+ {
+ "epoch": 74.5170305676856,
+ "grad_norm": 0.012975298799574375,
+ "learning_rate": 0.0006,
+ "loss": 4.540657997131348,
+ "step": 5365
+ },
+ {
+ "epoch": 74.53100436681223,
+ "grad_norm": 0.012380947358906269,
+ "learning_rate": 0.0006,
+ "loss": 4.55731725692749,
+ "step": 5366
+ },
+ {
+ "epoch": 74.54497816593886,
+ "grad_norm": 0.012577617540955544,
+ "learning_rate": 0.0006,
+ "loss": 4.500924110412598,
+ "step": 5367
+ },
+ {
+ "epoch": 74.5589519650655,
+ "grad_norm": 0.012089254334568977,
+ "learning_rate": 0.0006,
+ "loss": 4.561795234680176,
+ "step": 5368
+ },
+ {
+ "epoch": 74.57292576419214,
+ "grad_norm": 0.012024011462926865,
+ "learning_rate": 0.0006,
+ "loss": 4.430635452270508,
+ "step": 5369
+ },
+ {
+ "epoch": 74.58689956331878,
+ "grad_norm": 0.012361960485577583,
+ "learning_rate": 0.0006,
+ "loss": 4.4664106369018555,
+ "step": 5370
+ },
+ {
+ "epoch": 74.60087336244541,
+ "grad_norm": 0.012149893678724766,
+ "learning_rate": 0.0006,
+ "loss": 4.522271156311035,
+ "step": 5371
+ },
+ {
+ "epoch": 74.61484716157206,
+ "grad_norm": 0.01166840922087431,
+ "learning_rate": 0.0006,
+ "loss": 4.3870463371276855,
+ "step": 5372
+ },
+ {
+ "epoch": 74.62882096069869,
+ "grad_norm": 0.01160407718271017,
+ "learning_rate": 0.0006,
+ "loss": 4.504255771636963,
+ "step": 5373
+ },
+ {
+ "epoch": 74.64279475982532,
+ "grad_norm": 0.011259177699685097,
+ "learning_rate": 0.0006,
+ "loss": 4.507267951965332,
+ "step": 5374
+ },
+ {
+ "epoch": 74.65676855895197,
+ "grad_norm": 0.012285715900361538,
+ "learning_rate": 0.0006,
+ "loss": 4.472966194152832,
+ "step": 5375
+ },
+ {
+ "epoch": 74.6707423580786,
+ "grad_norm": 0.011214440688490868,
+ "learning_rate": 0.0006,
+ "loss": 4.5840253829956055,
+ "step": 5376
+ },
+ {
+ "epoch": 74.68471615720524,
+ "grad_norm": 0.012394879944622517,
+ "learning_rate": 0.0006,
+ "loss": 4.417830944061279,
+ "step": 5377
+ },
+ {
+ "epoch": 74.69868995633188,
+ "grad_norm": 0.012360049411654472,
+ "learning_rate": 0.0006,
+ "loss": 4.535454750061035,
+ "step": 5378
+ },
+ {
+ "epoch": 74.71266375545852,
+ "grad_norm": 0.011961652897298336,
+ "learning_rate": 0.0006,
+ "loss": 4.403409957885742,
+ "step": 5379
+ },
+ {
+ "epoch": 74.72663755458515,
+ "grad_norm": 0.012078274972736835,
+ "learning_rate": 0.0006,
+ "loss": 4.448366641998291,
+ "step": 5380
+ },
+ {
+ "epoch": 74.74061135371178,
+ "grad_norm": 0.013239333406090736,
+ "learning_rate": 0.0006,
+ "loss": 4.543450355529785,
+ "step": 5381
+ },
+ {
+ "epoch": 74.75458515283843,
+ "grad_norm": 0.01476441603153944,
+ "learning_rate": 0.0006,
+ "loss": 4.488677978515625,
+ "step": 5382
+ },
+ {
+ "epoch": 74.76855895196506,
+ "grad_norm": 0.013968009501695633,
+ "learning_rate": 0.0006,
+ "loss": 4.603639602661133,
+ "step": 5383
+ },
+ {
+ "epoch": 74.78253275109171,
+ "grad_norm": 0.012322458438575268,
+ "learning_rate": 0.0006,
+ "loss": 4.637874603271484,
+ "step": 5384
+ },
+ {
+ "epoch": 74.79650655021834,
+ "grad_norm": 0.013352593407034874,
+ "learning_rate": 0.0006,
+ "loss": 4.474833965301514,
+ "step": 5385
+ },
+ {
+ "epoch": 74.81048034934499,
+ "grad_norm": 0.014671728946268559,
+ "learning_rate": 0.0006,
+ "loss": 4.516847610473633,
+ "step": 5386
+ },
+ {
+ "epoch": 74.82445414847162,
+ "grad_norm": 0.017163321375846863,
+ "learning_rate": 0.0006,
+ "loss": 4.462325096130371,
+ "step": 5387
+ },
+ {
+ "epoch": 74.83842794759825,
+ "grad_norm": 0.02012511156499386,
+ "learning_rate": 0.0006,
+ "loss": 4.355982780456543,
+ "step": 5388
+ },
+ {
+ "epoch": 74.8524017467249,
+ "grad_norm": 0.01881697215139866,
+ "learning_rate": 0.0006,
+ "loss": 4.451120376586914,
+ "step": 5389
+ },
+ {
+ "epoch": 74.86637554585153,
+ "grad_norm": 0.017279407009482384,
+ "learning_rate": 0.0006,
+ "loss": 4.461202621459961,
+ "step": 5390
+ },
+ {
+ "epoch": 74.88034934497817,
+ "grad_norm": 0.015470407903194427,
+ "learning_rate": 0.0006,
+ "loss": 4.4968671798706055,
+ "step": 5391
+ },
+ {
+ "epoch": 74.8943231441048,
+ "grad_norm": 0.01411783043295145,
+ "learning_rate": 0.0006,
+ "loss": 4.558697700500488,
+ "step": 5392
+ },
+ {
+ "epoch": 74.90829694323143,
+ "grad_norm": 0.017178812995553017,
+ "learning_rate": 0.0006,
+ "loss": 4.254153251647949,
+ "step": 5393
+ },
+ {
+ "epoch": 74.92227074235808,
+ "grad_norm": 0.019193004816770554,
+ "learning_rate": 0.0006,
+ "loss": 4.394747257232666,
+ "step": 5394
+ },
+ {
+ "epoch": 74.93624454148471,
+ "grad_norm": 0.018942881375551224,
+ "learning_rate": 0.0006,
+ "loss": 4.459539413452148,
+ "step": 5395
+ },
+ {
+ "epoch": 74.95021834061136,
+ "grad_norm": 0.01641223020851612,
+ "learning_rate": 0.0006,
+ "loss": 4.485820770263672,
+ "step": 5396
+ },
+ {
+ "epoch": 74.96419213973799,
+ "grad_norm": 0.014485753141343594,
+ "learning_rate": 0.0006,
+ "loss": 4.5068583488464355,
+ "step": 5397
+ },
+ {
+ "epoch": 74.97816593886463,
+ "grad_norm": 0.018242299556732178,
+ "learning_rate": 0.0006,
+ "loss": 4.508835792541504,
+ "step": 5398
+ },
+ {
+ "epoch": 74.99213973799127,
+ "grad_norm": 0.016491873189806938,
+ "learning_rate": 0.0006,
+ "loss": 4.4760870933532715,
+ "step": 5399
+ },
+ {
+ "epoch": 75.0,
+ "grad_norm": 0.014543353579938412,
+ "learning_rate": 0.0006,
+ "loss": 4.540217399597168,
+ "step": 5400
+ },
+ {
+ "epoch": 75.0,
+ "eval_loss": 4.7617573738098145,
+ "eval_runtime": 57.321,
+ "eval_samples_per_second": 42.602,
+ "eval_steps_per_second": 1.343,
+ "step": 5400
+ },
+ {
+ "epoch": 75.01397379912663,
+ "grad_norm": 0.014417114667594433,
+ "learning_rate": 0.0006,
+ "loss": 4.332956790924072,
+ "step": 5401
+ },
+ {
+ "epoch": 75.02794759825328,
+ "grad_norm": 0.01607639715075493,
+ "learning_rate": 0.0006,
+ "loss": 4.352453708648682,
+ "step": 5402
+ },
+ {
+ "epoch": 75.04192139737991,
+ "grad_norm": 0.015438318252563477,
+ "learning_rate": 0.0006,
+ "loss": 4.387139320373535,
+ "step": 5403
+ },
+ {
+ "epoch": 75.05589519650655,
+ "grad_norm": 0.015539007261395454,
+ "learning_rate": 0.0006,
+ "loss": 4.453171730041504,
+ "step": 5404
+ },
+ {
+ "epoch": 75.06986899563319,
+ "grad_norm": 0.017361903563141823,
+ "learning_rate": 0.0006,
+ "loss": 4.512143135070801,
+ "step": 5405
+ },
+ {
+ "epoch": 75.08384279475983,
+ "grad_norm": 0.01815599389374256,
+ "learning_rate": 0.0006,
+ "loss": 4.508997440338135,
+ "step": 5406
+ },
+ {
+ "epoch": 75.09781659388646,
+ "grad_norm": 0.01620607078075409,
+ "learning_rate": 0.0006,
+ "loss": 4.322444915771484,
+ "step": 5407
+ },
+ {
+ "epoch": 75.1117903930131,
+ "grad_norm": 0.014773265458643436,
+ "learning_rate": 0.0006,
+ "loss": 4.359251976013184,
+ "step": 5408
+ },
+ {
+ "epoch": 75.12576419213974,
+ "grad_norm": 0.014099023304879665,
+ "learning_rate": 0.0006,
+ "loss": 4.313380241394043,
+ "step": 5409
+ },
+ {
+ "epoch": 75.13973799126637,
+ "grad_norm": 0.014387735165655613,
+ "learning_rate": 0.0006,
+ "loss": 4.435023307800293,
+ "step": 5410
+ },
+ {
+ "epoch": 75.15371179039302,
+ "grad_norm": 0.014886162243783474,
+ "learning_rate": 0.0006,
+ "loss": 4.335939407348633,
+ "step": 5411
+ },
+ {
+ "epoch": 75.16768558951965,
+ "grad_norm": 0.015766102820634842,
+ "learning_rate": 0.0006,
+ "loss": 4.408674240112305,
+ "step": 5412
+ },
+ {
+ "epoch": 75.18165938864628,
+ "grad_norm": 0.016383284702897072,
+ "learning_rate": 0.0006,
+ "loss": 4.385453224182129,
+ "step": 5413
+ },
+ {
+ "epoch": 75.19563318777293,
+ "grad_norm": 0.013535384088754654,
+ "learning_rate": 0.0006,
+ "loss": 4.292785167694092,
+ "step": 5414
+ },
+ {
+ "epoch": 75.20960698689956,
+ "grad_norm": 0.01354036945849657,
+ "learning_rate": 0.0006,
+ "loss": 4.306252479553223,
+ "step": 5415
+ },
+ {
+ "epoch": 75.2235807860262,
+ "grad_norm": 0.015818078070878983,
+ "learning_rate": 0.0006,
+ "loss": 4.425487995147705,
+ "step": 5416
+ },
+ {
+ "epoch": 75.23755458515284,
+ "grad_norm": 0.014363568276166916,
+ "learning_rate": 0.0006,
+ "loss": 4.395087242126465,
+ "step": 5417
+ },
+ {
+ "epoch": 75.25152838427948,
+ "grad_norm": 0.013895470649003983,
+ "learning_rate": 0.0006,
+ "loss": 4.364090442657471,
+ "step": 5418
+ },
+ {
+ "epoch": 75.26550218340611,
+ "grad_norm": 0.014178668148815632,
+ "learning_rate": 0.0006,
+ "loss": 4.44619083404541,
+ "step": 5419
+ },
+ {
+ "epoch": 75.27947598253274,
+ "grad_norm": 0.0152514623478055,
+ "learning_rate": 0.0006,
+ "loss": 4.348445892333984,
+ "step": 5420
+ },
+ {
+ "epoch": 75.29344978165939,
+ "grad_norm": 0.015397852286696434,
+ "learning_rate": 0.0006,
+ "loss": 4.476832389831543,
+ "step": 5421
+ },
+ {
+ "epoch": 75.30742358078602,
+ "grad_norm": 0.018009863793849945,
+ "learning_rate": 0.0006,
+ "loss": 4.431162357330322,
+ "step": 5422
+ },
+ {
+ "epoch": 75.32139737991267,
+ "grad_norm": 0.017038367688655853,
+ "learning_rate": 0.0006,
+ "loss": 4.474787712097168,
+ "step": 5423
+ },
+ {
+ "epoch": 75.3353711790393,
+ "grad_norm": 0.013763288035988808,
+ "learning_rate": 0.0006,
+ "loss": 4.470419883728027,
+ "step": 5424
+ },
+ {
+ "epoch": 75.34934497816595,
+ "grad_norm": 0.013972110114991665,
+ "learning_rate": 0.0006,
+ "loss": 4.373147964477539,
+ "step": 5425
+ },
+ {
+ "epoch": 75.36331877729258,
+ "grad_norm": 0.017197875306010246,
+ "learning_rate": 0.0006,
+ "loss": 4.494963645935059,
+ "step": 5426
+ },
+ {
+ "epoch": 75.37729257641921,
+ "grad_norm": 0.01687047816812992,
+ "learning_rate": 0.0006,
+ "loss": 4.435934066772461,
+ "step": 5427
+ },
+ {
+ "epoch": 75.39126637554585,
+ "grad_norm": 0.01578272506594658,
+ "learning_rate": 0.0006,
+ "loss": 4.37918758392334,
+ "step": 5428
+ },
+ {
+ "epoch": 75.40524017467249,
+ "grad_norm": 0.015956247225403786,
+ "learning_rate": 0.0006,
+ "loss": 4.455668926239014,
+ "step": 5429
+ },
+ {
+ "epoch": 75.41921397379913,
+ "grad_norm": 0.013495155610144138,
+ "learning_rate": 0.0006,
+ "loss": 4.4271039962768555,
+ "step": 5430
+ },
+ {
+ "epoch": 75.43318777292576,
+ "grad_norm": 0.012851621024310589,
+ "learning_rate": 0.0006,
+ "loss": 4.213107109069824,
+ "step": 5431
+ },
+ {
+ "epoch": 75.44716157205241,
+ "grad_norm": 0.013351158238947392,
+ "learning_rate": 0.0006,
+ "loss": 4.3070783615112305,
+ "step": 5432
+ },
+ {
+ "epoch": 75.46113537117904,
+ "grad_norm": 0.014765217900276184,
+ "learning_rate": 0.0006,
+ "loss": 4.385021209716797,
+ "step": 5433
+ },
+ {
+ "epoch": 75.47510917030567,
+ "grad_norm": 0.016091614961624146,
+ "learning_rate": 0.0006,
+ "loss": 4.312644004821777,
+ "step": 5434
+ },
+ {
+ "epoch": 75.48908296943232,
+ "grad_norm": 0.01612013578414917,
+ "learning_rate": 0.0006,
+ "loss": 4.4482011795043945,
+ "step": 5435
+ },
+ {
+ "epoch": 75.50305676855895,
+ "grad_norm": 0.013157603330910206,
+ "learning_rate": 0.0006,
+ "loss": 4.427262306213379,
+ "step": 5436
+ },
+ {
+ "epoch": 75.5170305676856,
+ "grad_norm": 0.016039809212088585,
+ "learning_rate": 0.0006,
+ "loss": 4.396635055541992,
+ "step": 5437
+ },
+ {
+ "epoch": 75.53100436681223,
+ "grad_norm": 0.01794283092021942,
+ "learning_rate": 0.0006,
+ "loss": 4.476870059967041,
+ "step": 5438
+ },
+ {
+ "epoch": 75.54497816593886,
+ "grad_norm": 0.017030486837029457,
+ "learning_rate": 0.0006,
+ "loss": 4.361156463623047,
+ "step": 5439
+ },
+ {
+ "epoch": 75.5589519650655,
+ "grad_norm": 0.017915161326527596,
+ "learning_rate": 0.0006,
+ "loss": 4.286147117614746,
+ "step": 5440
+ },
+ {
+ "epoch": 75.57292576419214,
+ "grad_norm": 0.017514433711767197,
+ "learning_rate": 0.0006,
+ "loss": 4.394091606140137,
+ "step": 5441
+ },
+ {
+ "epoch": 75.58689956331878,
+ "grad_norm": 0.015128589235246181,
+ "learning_rate": 0.0006,
+ "loss": 4.383126735687256,
+ "step": 5442
+ },
+ {
+ "epoch": 75.60087336244541,
+ "grad_norm": 0.0202616099268198,
+ "learning_rate": 0.0006,
+ "loss": 4.336188316345215,
+ "step": 5443
+ },
+ {
+ "epoch": 75.61484716157206,
+ "grad_norm": 0.019978471100330353,
+ "learning_rate": 0.0006,
+ "loss": 4.461453437805176,
+ "step": 5444
+ },
+ {
+ "epoch": 75.62882096069869,
+ "grad_norm": 0.01738741435110569,
+ "learning_rate": 0.0006,
+ "loss": 4.371197700500488,
+ "step": 5445
+ },
+ {
+ "epoch": 75.64279475982532,
+ "grad_norm": 0.016157809644937515,
+ "learning_rate": 0.0006,
+ "loss": 4.3312835693359375,
+ "step": 5446
+ },
+ {
+ "epoch": 75.65676855895197,
+ "grad_norm": 0.016394613310694695,
+ "learning_rate": 0.0006,
+ "loss": 4.380190849304199,
+ "step": 5447
+ },
+ {
+ "epoch": 75.6707423580786,
+ "grad_norm": 0.014727243222296238,
+ "learning_rate": 0.0006,
+ "loss": 4.308028221130371,
+ "step": 5448
+ },
+ {
+ "epoch": 75.68471615720524,
+ "grad_norm": 0.01695958711206913,
+ "learning_rate": 0.0006,
+ "loss": 4.399503707885742,
+ "step": 5449
+ },
+ {
+ "epoch": 75.69868995633188,
+ "grad_norm": 0.015720795840024948,
+ "learning_rate": 0.0006,
+ "loss": 4.342249870300293,
+ "step": 5450
+ },
+ {
+ "epoch": 75.71266375545852,
+ "grad_norm": 0.01574772410094738,
+ "learning_rate": 0.0006,
+ "loss": 4.331503868103027,
+ "step": 5451
+ },
+ {
+ "epoch": 75.72663755458515,
+ "grad_norm": 0.014812648296356201,
+ "learning_rate": 0.0006,
+ "loss": 4.332413673400879,
+ "step": 5452
+ },
+ {
+ "epoch": 75.74061135371178,
+ "grad_norm": 0.012850426137447357,
+ "learning_rate": 0.0006,
+ "loss": 4.2884063720703125,
+ "step": 5453
+ },
+ {
+ "epoch": 75.75458515283843,
+ "grad_norm": 0.012876970693469048,
+ "learning_rate": 0.0006,
+ "loss": 4.399855613708496,
+ "step": 5454
+ },
+ {
+ "epoch": 75.76855895196506,
+ "grad_norm": 0.012958863750100136,
+ "learning_rate": 0.0006,
+ "loss": 4.411185264587402,
+ "step": 5455
+ },
+ {
+ "epoch": 75.78253275109171,
+ "grad_norm": 0.012197135016322136,
+ "learning_rate": 0.0006,
+ "loss": 4.272603988647461,
+ "step": 5456
+ },
+ {
+ "epoch": 75.79650655021834,
+ "grad_norm": 0.013483566232025623,
+ "learning_rate": 0.0006,
+ "loss": 4.3564348220825195,
+ "step": 5457
+ },
+ {
+ "epoch": 75.81048034934499,
+ "grad_norm": 0.012365429662168026,
+ "learning_rate": 0.0006,
+ "loss": 4.354825973510742,
+ "step": 5458
+ },
+ {
+ "epoch": 75.82445414847162,
+ "grad_norm": 0.012238141149282455,
+ "learning_rate": 0.0006,
+ "loss": 4.440438747406006,
+ "step": 5459
+ },
+ {
+ "epoch": 75.83842794759825,
+ "grad_norm": 0.013380305841565132,
+ "learning_rate": 0.0006,
+ "loss": 4.2813215255737305,
+ "step": 5460
+ },
+ {
+ "epoch": 75.8524017467249,
+ "grad_norm": 0.01443779468536377,
+ "learning_rate": 0.0006,
+ "loss": 4.289078235626221,
+ "step": 5461
+ },
+ {
+ "epoch": 75.86637554585153,
+ "grad_norm": 0.012646574527025223,
+ "learning_rate": 0.0006,
+ "loss": 4.374177932739258,
+ "step": 5462
+ },
+ {
+ "epoch": 75.88034934497817,
+ "grad_norm": 0.012403388507664204,
+ "learning_rate": 0.0006,
+ "loss": 4.343016147613525,
+ "step": 5463
+ },
+ {
+ "epoch": 75.8943231441048,
+ "grad_norm": 0.01319252047687769,
+ "learning_rate": 0.0006,
+ "loss": 4.309957027435303,
+ "step": 5464
+ },
+ {
+ "epoch": 75.90829694323143,
+ "grad_norm": 0.012597428634762764,
+ "learning_rate": 0.0006,
+ "loss": 4.403654098510742,
+ "step": 5465
+ },
+ {
+ "epoch": 75.92227074235808,
+ "grad_norm": 0.012724226340651512,
+ "learning_rate": 0.0006,
+ "loss": 4.326869010925293,
+ "step": 5466
+ },
+ {
+ "epoch": 75.93624454148471,
+ "grad_norm": 0.012698772363364697,
+ "learning_rate": 0.0006,
+ "loss": 4.268430233001709,
+ "step": 5467
+ },
+ {
+ "epoch": 75.95021834061136,
+ "grad_norm": 0.01376490481197834,
+ "learning_rate": 0.0006,
+ "loss": 4.42446231842041,
+ "step": 5468
+ },
+ {
+ "epoch": 75.96419213973799,
+ "grad_norm": 0.013175144791603088,
+ "learning_rate": 0.0006,
+ "loss": 4.374675750732422,
+ "step": 5469
+ },
+ {
+ "epoch": 75.97816593886463,
+ "grad_norm": 0.012731322087347507,
+ "learning_rate": 0.0006,
+ "loss": 4.372966766357422,
+ "step": 5470
+ },
+ {
+ "epoch": 75.99213973799127,
+ "grad_norm": 0.01360505260527134,
+ "learning_rate": 0.0006,
+ "loss": 4.3095903396606445,
+ "step": 5471
+ },
+ {
+ "epoch": 76.0,
+ "grad_norm": 0.014011901803314686,
+ "learning_rate": 0.0006,
+ "loss": 4.26954460144043,
+ "step": 5472
+ },
+ {
+ "epoch": 76.0,
+ "eval_loss": 4.689640998840332,
+ "eval_runtime": 56.9095,
+ "eval_samples_per_second": 42.91,
+ "eval_steps_per_second": 1.353,
+ "step": 5472
+ },
+ {
+ "epoch": 76.01397379912663,
+ "grad_norm": 0.015351055189967155,
+ "learning_rate": 0.0006,
+ "loss": 4.348268985748291,
+ "step": 5473
+ },
+ {
+ "epoch": 76.02794759825328,
+ "grad_norm": 0.017164217308163643,
+ "learning_rate": 0.0006,
+ "loss": 4.244204998016357,
+ "step": 5474
+ },
+ {
+ "epoch": 76.04192139737991,
+ "grad_norm": 0.021624932065606117,
+ "learning_rate": 0.0006,
+ "loss": 4.311168670654297,
+ "step": 5475
+ },
+ {
+ "epoch": 76.05589519650655,
+ "grad_norm": 0.028712468221783638,
+ "learning_rate": 0.0006,
+ "loss": 4.322287082672119,
+ "step": 5476
+ },
+ {
+ "epoch": 76.06986899563319,
+ "grad_norm": 0.027915487065911293,
+ "learning_rate": 0.0006,
+ "loss": 4.381207466125488,
+ "step": 5477
+ },
+ {
+ "epoch": 76.08384279475983,
+ "grad_norm": 0.017955303192138672,
+ "learning_rate": 0.0006,
+ "loss": 4.2635016441345215,
+ "step": 5478
+ },
+ {
+ "epoch": 76.09781659388646,
+ "grad_norm": 0.019113514572381973,
+ "learning_rate": 0.0006,
+ "loss": 4.399471282958984,
+ "step": 5479
+ },
+ {
+ "epoch": 76.1117903930131,
+ "grad_norm": 0.017502542585134506,
+ "learning_rate": 0.0006,
+ "loss": 4.321204662322998,
+ "step": 5480
+ },
+ {
+ "epoch": 76.12576419213974,
+ "grad_norm": 0.01852525770664215,
+ "learning_rate": 0.0006,
+ "loss": 4.217884063720703,
+ "step": 5481
+ },
+ {
+ "epoch": 76.13973799126637,
+ "grad_norm": 0.019836828112602234,
+ "learning_rate": 0.0006,
+ "loss": 4.278532028198242,
+ "step": 5482
+ },
+ {
+ "epoch": 76.15371179039302,
+ "grad_norm": 0.02101975679397583,
+ "learning_rate": 0.0006,
+ "loss": 4.246720314025879,
+ "step": 5483
+ },
+ {
+ "epoch": 76.16768558951965,
+ "grad_norm": 0.021202990785241127,
+ "learning_rate": 0.0006,
+ "loss": 4.517517566680908,
+ "step": 5484
+ },
+ {
+ "epoch": 76.18165938864628,
+ "grad_norm": 0.021428769454360008,
+ "learning_rate": 0.0006,
+ "loss": 4.3373517990112305,
+ "step": 5485
+ },
+ {
+ "epoch": 76.19563318777293,
+ "grad_norm": 0.018263190984725952,
+ "learning_rate": 0.0006,
+ "loss": 4.402856826782227,
+ "step": 5486
+ },
+ {
+ "epoch": 76.20960698689956,
+ "grad_norm": 0.019025536254048347,
+ "learning_rate": 0.0006,
+ "loss": 4.314227104187012,
+ "step": 5487
+ },
+ {
+ "epoch": 76.2235807860262,
+ "grad_norm": 0.020918630063533783,
+ "learning_rate": 0.0006,
+ "loss": 4.454369068145752,
+ "step": 5488
+ },
+ {
+ "epoch": 76.23755458515284,
+ "grad_norm": 0.019307617098093033,
+ "learning_rate": 0.0006,
+ "loss": 4.247509002685547,
+ "step": 5489
+ },
+ {
+ "epoch": 76.25152838427948,
+ "grad_norm": 0.014956331811845303,
+ "learning_rate": 0.0006,
+ "loss": 4.365175247192383,
+ "step": 5490
+ },
+ {
+ "epoch": 76.26550218340611,
+ "grad_norm": 0.017213059589266777,
+ "learning_rate": 0.0006,
+ "loss": 4.384005069732666,
+ "step": 5491
+ },
+ {
+ "epoch": 76.27947598253274,
+ "grad_norm": 0.017575597390532494,
+ "learning_rate": 0.0006,
+ "loss": 4.41934871673584,
+ "step": 5492
+ },
+ {
+ "epoch": 76.29344978165939,
+ "grad_norm": 0.015879256650805473,
+ "learning_rate": 0.0006,
+ "loss": 4.26675271987915,
+ "step": 5493
+ },
+ {
+ "epoch": 76.30742358078602,
+ "grad_norm": 0.01564522460103035,
+ "learning_rate": 0.0006,
+ "loss": 4.277339935302734,
+ "step": 5494
+ },
+ {
+ "epoch": 76.32139737991267,
+ "grad_norm": 0.016414586454629898,
+ "learning_rate": 0.0006,
+ "loss": 4.4138593673706055,
+ "step": 5495
+ },
+ {
+ "epoch": 76.3353711790393,
+ "grad_norm": 0.014732337556779385,
+ "learning_rate": 0.0006,
+ "loss": 4.281006336212158,
+ "step": 5496
+ },
+ {
+ "epoch": 76.34934497816595,
+ "grad_norm": 0.015461455099284649,
+ "learning_rate": 0.0006,
+ "loss": 4.263422966003418,
+ "step": 5497
+ },
+ {
+ "epoch": 76.36331877729258,
+ "grad_norm": 0.014529814012348652,
+ "learning_rate": 0.0006,
+ "loss": 4.36165189743042,
+ "step": 5498
+ },
+ {
+ "epoch": 76.37729257641921,
+ "grad_norm": 0.01469483319669962,
+ "learning_rate": 0.0006,
+ "loss": 4.309642791748047,
+ "step": 5499
+ },
+ {
+ "epoch": 76.39126637554585,
+ "grad_norm": 0.016490574926137924,
+ "learning_rate": 0.0006,
+ "loss": 4.288150787353516,
+ "step": 5500
+ },
+ {
+ "epoch": 76.40524017467249,
+ "grad_norm": 0.014593565836548805,
+ "learning_rate": 0.0006,
+ "loss": 4.271614074707031,
+ "step": 5501
+ },
+ {
+ "epoch": 76.41921397379913,
+ "grad_norm": 0.012796535156667233,
+ "learning_rate": 0.0006,
+ "loss": 4.420042037963867,
+ "step": 5502
+ },
+ {
+ "epoch": 76.43318777292576,
+ "grad_norm": 0.016086025163531303,
+ "learning_rate": 0.0006,
+ "loss": 4.289943695068359,
+ "step": 5503
+ },
+ {
+ "epoch": 76.44716157205241,
+ "grad_norm": 0.015439470298588276,
+ "learning_rate": 0.0006,
+ "loss": 4.3023271560668945,
+ "step": 5504
+ },
+ {
+ "epoch": 76.46113537117904,
+ "grad_norm": 0.015066367574036121,
+ "learning_rate": 0.0006,
+ "loss": 4.197081565856934,
+ "step": 5505
+ },
+ {
+ "epoch": 76.47510917030567,
+ "grad_norm": 0.016653263941407204,
+ "learning_rate": 0.0006,
+ "loss": 4.333785057067871,
+ "step": 5506
+ },
+ {
+ "epoch": 76.48908296943232,
+ "grad_norm": 0.014791973866522312,
+ "learning_rate": 0.0006,
+ "loss": 4.323811054229736,
+ "step": 5507
+ },
+ {
+ "epoch": 76.50305676855895,
+ "grad_norm": 0.014694827608764172,
+ "learning_rate": 0.0006,
+ "loss": 4.195852279663086,
+ "step": 5508
+ },
+ {
+ "epoch": 76.5170305676856,
+ "grad_norm": 0.015120396390557289,
+ "learning_rate": 0.0006,
+ "loss": 4.42379903793335,
+ "step": 5509
+ },
+ {
+ "epoch": 76.53100436681223,
+ "grad_norm": 0.014584025368094444,
+ "learning_rate": 0.0006,
+ "loss": 4.2702531814575195,
+ "step": 5510
+ },
+ {
+ "epoch": 76.54497816593886,
+ "grad_norm": 0.013981613330543041,
+ "learning_rate": 0.0006,
+ "loss": 4.383029937744141,
+ "step": 5511
+ },
+ {
+ "epoch": 76.5589519650655,
+ "grad_norm": 0.012930642813444138,
+ "learning_rate": 0.0006,
+ "loss": 4.296136856079102,
+ "step": 5512
+ },
+ {
+ "epoch": 76.57292576419214,
+ "grad_norm": 0.013060989789664745,
+ "learning_rate": 0.0006,
+ "loss": 4.346380233764648,
+ "step": 5513
+ },
+ {
+ "epoch": 76.58689956331878,
+ "grad_norm": 0.013209798373281956,
+ "learning_rate": 0.0006,
+ "loss": 4.316067218780518,
+ "step": 5514
+ },
+ {
+ "epoch": 76.60087336244541,
+ "grad_norm": 0.012423216365277767,
+ "learning_rate": 0.0006,
+ "loss": 4.370274543762207,
+ "step": 5515
+ },
+ {
+ "epoch": 76.61484716157206,
+ "grad_norm": 0.013205411843955517,
+ "learning_rate": 0.0006,
+ "loss": 4.429969310760498,
+ "step": 5516
+ },
+ {
+ "epoch": 76.62882096069869,
+ "grad_norm": 0.013049827888607979,
+ "learning_rate": 0.0006,
+ "loss": 4.378549098968506,
+ "step": 5517
+ },
+ {
+ "epoch": 76.64279475982532,
+ "grad_norm": 0.012075880542397499,
+ "learning_rate": 0.0006,
+ "loss": 4.348293304443359,
+ "step": 5518
+ },
+ {
+ "epoch": 76.65676855895197,
+ "grad_norm": 0.014932847581803799,
+ "learning_rate": 0.0006,
+ "loss": 4.282538890838623,
+ "step": 5519
+ },
+ {
+ "epoch": 76.6707423580786,
+ "grad_norm": 0.015615998767316341,
+ "learning_rate": 0.0006,
+ "loss": 4.274641036987305,
+ "step": 5520
+ },
+ {
+ "epoch": 76.68471615720524,
+ "grad_norm": 0.01424513477832079,
+ "learning_rate": 0.0006,
+ "loss": 4.243166923522949,
+ "step": 5521
+ },
+ {
+ "epoch": 76.69868995633188,
+ "grad_norm": 0.012775546871125698,
+ "learning_rate": 0.0006,
+ "loss": 4.335611820220947,
+ "step": 5522
+ },
+ {
+ "epoch": 76.71266375545852,
+ "grad_norm": 0.014461303129792213,
+ "learning_rate": 0.0006,
+ "loss": 4.246976375579834,
+ "step": 5523
+ },
+ {
+ "epoch": 76.72663755458515,
+ "grad_norm": 0.0145436255261302,
+ "learning_rate": 0.0006,
+ "loss": 4.275895118713379,
+ "step": 5524
+ },
+ {
+ "epoch": 76.74061135371178,
+ "grad_norm": 0.014055085368454456,
+ "learning_rate": 0.0006,
+ "loss": 4.331108093261719,
+ "step": 5525
+ },
+ {
+ "epoch": 76.75458515283843,
+ "grad_norm": 0.011933894827961922,
+ "learning_rate": 0.0006,
+ "loss": 4.293721675872803,
+ "step": 5526
+ },
+ {
+ "epoch": 76.76855895196506,
+ "grad_norm": 0.01277392078191042,
+ "learning_rate": 0.0006,
+ "loss": 4.312374114990234,
+ "step": 5527
+ },
+ {
+ "epoch": 76.78253275109171,
+ "grad_norm": 0.014227936044335365,
+ "learning_rate": 0.0006,
+ "loss": 4.238729000091553,
+ "step": 5528
+ },
+ {
+ "epoch": 76.79650655021834,
+ "grad_norm": 0.01488024927675724,
+ "learning_rate": 0.0006,
+ "loss": 4.24587345123291,
+ "step": 5529
+ },
+ {
+ "epoch": 76.81048034934499,
+ "grad_norm": 0.015341250225901604,
+ "learning_rate": 0.0006,
+ "loss": 4.257378578186035,
+ "step": 5530
+ },
+ {
+ "epoch": 76.82445414847162,
+ "grad_norm": 0.015179437585175037,
+ "learning_rate": 0.0006,
+ "loss": 4.2812910079956055,
+ "step": 5531
+ },
+ {
+ "epoch": 76.83842794759825,
+ "grad_norm": 0.012814724817872047,
+ "learning_rate": 0.0006,
+ "loss": 4.2309465408325195,
+ "step": 5532
+ },
+ {
+ "epoch": 76.8524017467249,
+ "grad_norm": 0.013810716569423676,
+ "learning_rate": 0.0006,
+ "loss": 4.325292110443115,
+ "step": 5533
+ },
+ {
+ "epoch": 76.86637554585153,
+ "grad_norm": 0.017468232661485672,
+ "learning_rate": 0.0006,
+ "loss": 4.243788719177246,
+ "step": 5534
+ },
+ {
+ "epoch": 76.88034934497817,
+ "grad_norm": 0.01739857904613018,
+ "learning_rate": 0.0006,
+ "loss": 4.345390796661377,
+ "step": 5535
+ },
+ {
+ "epoch": 76.8943231441048,
+ "grad_norm": 0.015329061076045036,
+ "learning_rate": 0.0006,
+ "loss": 4.415142059326172,
+ "step": 5536
+ },
+ {
+ "epoch": 76.90829694323143,
+ "grad_norm": 0.01598426327109337,
+ "learning_rate": 0.0006,
+ "loss": 4.299558639526367,
+ "step": 5537
+ },
+ {
+ "epoch": 76.92227074235808,
+ "grad_norm": 0.01621454581618309,
+ "learning_rate": 0.0006,
+ "loss": 4.284354209899902,
+ "step": 5538
+ },
+ {
+ "epoch": 76.93624454148471,
+ "grad_norm": 0.015956968069076538,
+ "learning_rate": 0.0006,
+ "loss": 4.293837547302246,
+ "step": 5539
+ },
+ {
+ "epoch": 76.95021834061136,
+ "grad_norm": 0.01776493526995182,
+ "learning_rate": 0.0006,
+ "loss": 4.229729175567627,
+ "step": 5540
+ },
+ {
+ "epoch": 76.96419213973799,
+ "grad_norm": 0.015317477285861969,
+ "learning_rate": 0.0006,
+ "loss": 4.3328447341918945,
+ "step": 5541
+ },
+ {
+ "epoch": 76.97816593886463,
+ "grad_norm": 0.015140251256525517,
+ "learning_rate": 0.0006,
+ "loss": 4.310840606689453,
+ "step": 5542
+ },
+ {
+ "epoch": 76.99213973799127,
+ "grad_norm": 0.015326812863349915,
+ "learning_rate": 0.0006,
+ "loss": 4.230832099914551,
+ "step": 5543
+ },
+ {
+ "epoch": 77.0,
+ "grad_norm": 0.01607763208448887,
+ "learning_rate": 0.0006,
+ "loss": 4.2236328125,
+ "step": 5544
+ },
+ {
+ "epoch": 77.0,
+ "eval_loss": 4.662221431732178,
+ "eval_runtime": 56.5194,
+ "eval_samples_per_second": 43.206,
+ "eval_steps_per_second": 1.362,
+ "step": 5544
+ },
+ {
+ "epoch": 77.01397379912663,
+ "grad_norm": 0.014839510433375835,
+ "learning_rate": 0.0006,
+ "loss": 4.254111289978027,
+ "step": 5545
+ },
+ {
+ "epoch": 77.02794759825328,
+ "grad_norm": 0.01519741676747799,
+ "learning_rate": 0.0006,
+ "loss": 4.1607184410095215,
+ "step": 5546
+ },
+ {
+ "epoch": 77.04192139737991,
+ "grad_norm": 0.015472603030502796,
+ "learning_rate": 0.0006,
+ "loss": 4.323320388793945,
+ "step": 5547
+ },
+ {
+ "epoch": 77.05589519650655,
+ "grad_norm": 0.017937948927283287,
+ "learning_rate": 0.0006,
+ "loss": 4.308490753173828,
+ "step": 5548
+ },
+ {
+ "epoch": 77.06986899563319,
+ "grad_norm": 0.020890841260552406,
+ "learning_rate": 0.0006,
+ "loss": 4.259101867675781,
+ "step": 5549
+ },
+ {
+ "epoch": 77.08384279475983,
+ "grad_norm": 0.01892857998609543,
+ "learning_rate": 0.0006,
+ "loss": 4.386970520019531,
+ "step": 5550
+ },
+ {
+ "epoch": 77.09781659388646,
+ "grad_norm": 0.016583630815148354,
+ "learning_rate": 0.0006,
+ "loss": 4.233604907989502,
+ "step": 5551
+ },
+ {
+ "epoch": 77.1117903930131,
+ "grad_norm": 0.016063468530774117,
+ "learning_rate": 0.0006,
+ "loss": 4.339544296264648,
+ "step": 5552
+ },
+ {
+ "epoch": 77.12576419213974,
+ "grad_norm": 0.018326779827475548,
+ "learning_rate": 0.0006,
+ "loss": 4.367204666137695,
+ "step": 5553
+ },
+ {
+ "epoch": 77.13973799126637,
+ "grad_norm": 0.021590933203697205,
+ "learning_rate": 0.0006,
+ "loss": 4.352685928344727,
+ "step": 5554
+ },
+ {
+ "epoch": 77.15371179039302,
+ "grad_norm": 0.019643642008304596,
+ "learning_rate": 0.0006,
+ "loss": 4.258535385131836,
+ "step": 5555
+ },
+ {
+ "epoch": 77.16768558951965,
+ "grad_norm": 0.018476149067282677,
+ "learning_rate": 0.0006,
+ "loss": 4.232568740844727,
+ "step": 5556
+ },
+ {
+ "epoch": 77.18165938864628,
+ "grad_norm": 0.01584405079483986,
+ "learning_rate": 0.0006,
+ "loss": 4.258371353149414,
+ "step": 5557
+ },
+ {
+ "epoch": 77.19563318777293,
+ "grad_norm": 0.01575201191008091,
+ "learning_rate": 0.0006,
+ "loss": 4.183218002319336,
+ "step": 5558
+ },
+ {
+ "epoch": 77.20960698689956,
+ "grad_norm": 0.016871824860572815,
+ "learning_rate": 0.0006,
+ "loss": 4.194563865661621,
+ "step": 5559
+ },
+ {
+ "epoch": 77.2235807860262,
+ "grad_norm": 0.02059333398938179,
+ "learning_rate": 0.0006,
+ "loss": 4.195281982421875,
+ "step": 5560
+ },
+ {
+ "epoch": 77.23755458515284,
+ "grad_norm": 0.022640453651547432,
+ "learning_rate": 0.0006,
+ "loss": 4.279615879058838,
+ "step": 5561
+ },
+ {
+ "epoch": 77.25152838427948,
+ "grad_norm": 0.020335959270596504,
+ "learning_rate": 0.0006,
+ "loss": 4.25040864944458,
+ "step": 5562
+ },
+ {
+ "epoch": 77.26550218340611,
+ "grad_norm": 0.020010793581604958,
+ "learning_rate": 0.0006,
+ "loss": 4.291116714477539,
+ "step": 5563
+ },
+ {
+ "epoch": 77.27947598253274,
+ "grad_norm": 0.02109045907855034,
+ "learning_rate": 0.0006,
+ "loss": 4.364088535308838,
+ "step": 5564
+ },
+ {
+ "epoch": 77.29344978165939,
+ "grad_norm": 0.019519446417689323,
+ "learning_rate": 0.0006,
+ "loss": 4.329166412353516,
+ "step": 5565
+ },
+ {
+ "epoch": 77.30742358078602,
+ "grad_norm": 0.01689804159104824,
+ "learning_rate": 0.0006,
+ "loss": 4.244176864624023,
+ "step": 5566
+ },
+ {
+ "epoch": 77.32139737991267,
+ "grad_norm": 0.01893681474030018,
+ "learning_rate": 0.0006,
+ "loss": 4.239068984985352,
+ "step": 5567
+ },
+ {
+ "epoch": 77.3353711790393,
+ "grad_norm": 0.015998193994164467,
+ "learning_rate": 0.0006,
+ "loss": 4.173398017883301,
+ "step": 5568
+ },
+ {
+ "epoch": 77.34934497816595,
+ "grad_norm": 0.013892280869185925,
+ "learning_rate": 0.0006,
+ "loss": 4.287654876708984,
+ "step": 5569
+ },
+ {
+ "epoch": 77.36331877729258,
+ "grad_norm": 0.017060643061995506,
+ "learning_rate": 0.0006,
+ "loss": 4.398447036743164,
+ "step": 5570
+ },
+ {
+ "epoch": 77.37729257641921,
+ "grad_norm": 0.015168849378824234,
+ "learning_rate": 0.0006,
+ "loss": 4.15665864944458,
+ "step": 5571
+ },
+ {
+ "epoch": 77.39126637554585,
+ "grad_norm": 0.014886523596942425,
+ "learning_rate": 0.0006,
+ "loss": 4.196673393249512,
+ "step": 5572
+ },
+ {
+ "epoch": 77.40524017467249,
+ "grad_norm": 0.016941796988248825,
+ "learning_rate": 0.0006,
+ "loss": 4.314562797546387,
+ "step": 5573
+ },
+ {
+ "epoch": 77.41921397379913,
+ "grad_norm": 0.015353812836110592,
+ "learning_rate": 0.0006,
+ "loss": 4.409579277038574,
+ "step": 5574
+ },
+ {
+ "epoch": 77.43318777292576,
+ "grad_norm": 0.015192852355539799,
+ "learning_rate": 0.0006,
+ "loss": 4.392045974731445,
+ "step": 5575
+ },
+ {
+ "epoch": 77.44716157205241,
+ "grad_norm": 0.014232569374144077,
+ "learning_rate": 0.0006,
+ "loss": 4.21639347076416,
+ "step": 5576
+ },
+ {
+ "epoch": 77.46113537117904,
+ "grad_norm": 0.014453226700425148,
+ "learning_rate": 0.0006,
+ "loss": 4.2802276611328125,
+ "step": 5577
+ },
+ {
+ "epoch": 77.47510917030567,
+ "grad_norm": 0.012881236150860786,
+ "learning_rate": 0.0006,
+ "loss": 4.375980854034424,
+ "step": 5578
+ },
+ {
+ "epoch": 77.48908296943232,
+ "grad_norm": 0.01583976298570633,
+ "learning_rate": 0.0006,
+ "loss": 4.264520645141602,
+ "step": 5579
+ },
+ {
+ "epoch": 77.50305676855895,
+ "grad_norm": 0.014171584509313107,
+ "learning_rate": 0.0006,
+ "loss": 4.225279808044434,
+ "step": 5580
+ },
+ {
+ "epoch": 77.5170305676856,
+ "grad_norm": 0.014715159311890602,
+ "learning_rate": 0.0006,
+ "loss": 4.369834899902344,
+ "step": 5581
+ },
+ {
+ "epoch": 77.53100436681223,
+ "grad_norm": 0.016554079949855804,
+ "learning_rate": 0.0006,
+ "loss": 4.2976179122924805,
+ "step": 5582
+ },
+ {
+ "epoch": 77.54497816593886,
+ "grad_norm": 0.015095747075974941,
+ "learning_rate": 0.0006,
+ "loss": 4.294407844543457,
+ "step": 5583
+ },
+ {
+ "epoch": 77.5589519650655,
+ "grad_norm": 0.012726361863315105,
+ "learning_rate": 0.0006,
+ "loss": 4.221401214599609,
+ "step": 5584
+ },
+ {
+ "epoch": 77.57292576419214,
+ "grad_norm": 0.014999309554696083,
+ "learning_rate": 0.0006,
+ "loss": 4.243929386138916,
+ "step": 5585
+ },
+ {
+ "epoch": 77.58689956331878,
+ "grad_norm": 0.016526037827134132,
+ "learning_rate": 0.0006,
+ "loss": 4.374139785766602,
+ "step": 5586
+ },
+ {
+ "epoch": 77.60087336244541,
+ "grad_norm": 0.01648339442908764,
+ "learning_rate": 0.0006,
+ "loss": 4.392425060272217,
+ "step": 5587
+ },
+ {
+ "epoch": 77.61484716157206,
+ "grad_norm": 0.015269213356077671,
+ "learning_rate": 0.0006,
+ "loss": 4.33986759185791,
+ "step": 5588
+ },
+ {
+ "epoch": 77.62882096069869,
+ "grad_norm": 0.01690507121384144,
+ "learning_rate": 0.0006,
+ "loss": 4.307476043701172,
+ "step": 5589
+ },
+ {
+ "epoch": 77.64279475982532,
+ "grad_norm": 0.016438685357570648,
+ "learning_rate": 0.0006,
+ "loss": 4.2406206130981445,
+ "step": 5590
+ },
+ {
+ "epoch": 77.65676855895197,
+ "grad_norm": 0.0157600287348032,
+ "learning_rate": 0.0006,
+ "loss": 4.314070224761963,
+ "step": 5591
+ },
+ {
+ "epoch": 77.6707423580786,
+ "grad_norm": 0.015127947553992271,
+ "learning_rate": 0.0006,
+ "loss": 4.22475528717041,
+ "step": 5592
+ },
+ {
+ "epoch": 77.68471615720524,
+ "grad_norm": 0.016632677987217903,
+ "learning_rate": 0.0006,
+ "loss": 4.26954984664917,
+ "step": 5593
+ },
+ {
+ "epoch": 77.69868995633188,
+ "grad_norm": 0.013150460086762905,
+ "learning_rate": 0.0006,
+ "loss": 4.369131088256836,
+ "step": 5594
+ },
+ {
+ "epoch": 77.71266375545852,
+ "grad_norm": 0.014513005502521992,
+ "learning_rate": 0.0006,
+ "loss": 4.3248796463012695,
+ "step": 5595
+ },
+ {
+ "epoch": 77.72663755458515,
+ "grad_norm": 0.014090991578996181,
+ "learning_rate": 0.0006,
+ "loss": 4.268381118774414,
+ "step": 5596
+ },
+ {
+ "epoch": 77.74061135371178,
+ "grad_norm": 0.01633097417652607,
+ "learning_rate": 0.0006,
+ "loss": 4.238439083099365,
+ "step": 5597
+ },
+ {
+ "epoch": 77.75458515283843,
+ "grad_norm": 0.01710933819413185,
+ "learning_rate": 0.0006,
+ "loss": 4.399697780609131,
+ "step": 5598
+ },
+ {
+ "epoch": 77.76855895196506,
+ "grad_norm": 0.019251767545938492,
+ "learning_rate": 0.0006,
+ "loss": 4.290450096130371,
+ "step": 5599
+ },
+ {
+ "epoch": 77.78253275109171,
+ "grad_norm": 0.01849171705543995,
+ "learning_rate": 0.0006,
+ "loss": 4.267734527587891,
+ "step": 5600
+ },
+ {
+ "epoch": 77.79650655021834,
+ "grad_norm": 0.014173115603625774,
+ "learning_rate": 0.0006,
+ "loss": 4.288799285888672,
+ "step": 5601
+ },
+ {
+ "epoch": 77.81048034934499,
+ "grad_norm": 0.014572718180716038,
+ "learning_rate": 0.0006,
+ "loss": 4.3790435791015625,
+ "step": 5602
+ },
+ {
+ "epoch": 77.82445414847162,
+ "grad_norm": 0.016375336796045303,
+ "learning_rate": 0.0006,
+ "loss": 4.231725692749023,
+ "step": 5603
+ },
+ {
+ "epoch": 77.83842794759825,
+ "grad_norm": 0.017326701432466507,
+ "learning_rate": 0.0006,
+ "loss": 4.211495399475098,
+ "step": 5604
+ },
+ {
+ "epoch": 77.8524017467249,
+ "grad_norm": 0.016399390995502472,
+ "learning_rate": 0.0006,
+ "loss": 4.386699199676514,
+ "step": 5605
+ },
+ {
+ "epoch": 77.86637554585153,
+ "grad_norm": 0.018269989639520645,
+ "learning_rate": 0.0006,
+ "loss": 4.254856109619141,
+ "step": 5606
+ },
+ {
+ "epoch": 77.88034934497817,
+ "grad_norm": 0.018443435430526733,
+ "learning_rate": 0.0006,
+ "loss": 4.391961097717285,
+ "step": 5607
+ },
+ {
+ "epoch": 77.8943231441048,
+ "grad_norm": 0.01784636080265045,
+ "learning_rate": 0.0006,
+ "loss": 4.273881912231445,
+ "step": 5608
+ },
+ {
+ "epoch": 77.90829694323143,
+ "grad_norm": 0.01782994158565998,
+ "learning_rate": 0.0006,
+ "loss": 4.26833963394165,
+ "step": 5609
+ },
+ {
+ "epoch": 77.92227074235808,
+ "grad_norm": 0.016260741278529167,
+ "learning_rate": 0.0006,
+ "loss": 4.306807994842529,
+ "step": 5610
+ },
+ {
+ "epoch": 77.93624454148471,
+ "grad_norm": 0.015482903458178043,
+ "learning_rate": 0.0006,
+ "loss": 4.187826633453369,
+ "step": 5611
+ },
+ {
+ "epoch": 77.95021834061136,
+ "grad_norm": 0.015193257480859756,
+ "learning_rate": 0.0006,
+ "loss": 4.286060333251953,
+ "step": 5612
+ },
+ {
+ "epoch": 77.96419213973799,
+ "grad_norm": 0.015110115520656109,
+ "learning_rate": 0.0006,
+ "loss": 4.2828192710876465,
+ "step": 5613
+ },
+ {
+ "epoch": 77.97816593886463,
+ "grad_norm": 0.015973486006259918,
+ "learning_rate": 0.0006,
+ "loss": 4.317939758300781,
+ "step": 5614
+ },
+ {
+ "epoch": 77.99213973799127,
+ "grad_norm": 0.017363594844937325,
+ "learning_rate": 0.0006,
+ "loss": 4.287721633911133,
+ "step": 5615
+ },
+ {
+ "epoch": 78.0,
+ "grad_norm": 0.015277642756700516,
+ "learning_rate": 0.0006,
+ "loss": 4.356618881225586,
+ "step": 5616
+ },
+ {
+ "epoch": 78.0,
+ "eval_loss": 4.642678737640381,
+ "eval_runtime": 56.7335,
+ "eval_samples_per_second": 43.043,
+ "eval_steps_per_second": 1.357,
+ "step": 5616
+ },
+ {
+ "epoch": 78.01397379912663,
+ "grad_norm": 0.016794051975011826,
+ "learning_rate": 0.0006,
+ "loss": 4.279930591583252,
+ "step": 5617
+ },
+ {
+ "epoch": 78.02794759825328,
+ "grad_norm": 0.01632673107087612,
+ "learning_rate": 0.0006,
+ "loss": 4.235943794250488,
+ "step": 5618
+ },
+ {
+ "epoch": 78.04192139737991,
+ "grad_norm": 0.01437876746058464,
+ "learning_rate": 0.0006,
+ "loss": 4.2542243003845215,
+ "step": 5619
+ },
+ {
+ "epoch": 78.05589519650655,
+ "grad_norm": 0.014993119053542614,
+ "learning_rate": 0.0006,
+ "loss": 4.262046813964844,
+ "step": 5620
+ },
+ {
+ "epoch": 78.06986899563319,
+ "grad_norm": 0.015721119940280914,
+ "learning_rate": 0.0006,
+ "loss": 4.173436164855957,
+ "step": 5621
+ },
+ {
+ "epoch": 78.08384279475983,
+ "grad_norm": 0.01334085687994957,
+ "learning_rate": 0.0006,
+ "loss": 4.322175979614258,
+ "step": 5622
+ },
+ {
+ "epoch": 78.09781659388646,
+ "grad_norm": 0.015561388805508614,
+ "learning_rate": 0.0006,
+ "loss": 4.242125988006592,
+ "step": 5623
+ },
+ {
+ "epoch": 78.1117903930131,
+ "grad_norm": 0.014418602921068668,
+ "learning_rate": 0.0006,
+ "loss": 4.266186714172363,
+ "step": 5624
+ },
+ {
+ "epoch": 78.12576419213974,
+ "grad_norm": 0.013739909045398235,
+ "learning_rate": 0.0006,
+ "loss": 4.231633186340332,
+ "step": 5625
+ },
+ {
+ "epoch": 78.13973799126637,
+ "grad_norm": 0.01402067206799984,
+ "learning_rate": 0.0006,
+ "loss": 4.188406944274902,
+ "step": 5626
+ },
+ {
+ "epoch": 78.15371179039302,
+ "grad_norm": 0.013639481738209724,
+ "learning_rate": 0.0006,
+ "loss": 4.1586594581604,
+ "step": 5627
+ },
+ {
+ "epoch": 78.16768558951965,
+ "grad_norm": 0.012858159840106964,
+ "learning_rate": 0.0006,
+ "loss": 4.360448360443115,
+ "step": 5628
+ },
+ {
+ "epoch": 78.18165938864628,
+ "grad_norm": 0.013089739717543125,
+ "learning_rate": 0.0006,
+ "loss": 4.344371795654297,
+ "step": 5629
+ },
+ {
+ "epoch": 78.19563318777293,
+ "grad_norm": 0.01382314320653677,
+ "learning_rate": 0.0006,
+ "loss": 4.20378303527832,
+ "step": 5630
+ },
+ {
+ "epoch": 78.20960698689956,
+ "grad_norm": 0.014206619001924992,
+ "learning_rate": 0.0006,
+ "loss": 4.239891529083252,
+ "step": 5631
+ },
+ {
+ "epoch": 78.2235807860262,
+ "grad_norm": 0.01472168043255806,
+ "learning_rate": 0.0006,
+ "loss": 4.282382488250732,
+ "step": 5632
+ },
+ {
+ "epoch": 78.23755458515284,
+ "grad_norm": 0.014237025752663612,
+ "learning_rate": 0.0006,
+ "loss": 4.241545677185059,
+ "step": 5633
+ },
+ {
+ "epoch": 78.25152838427948,
+ "grad_norm": 0.013231472112238407,
+ "learning_rate": 0.0006,
+ "loss": 4.184896469116211,
+ "step": 5634
+ },
+ {
+ "epoch": 78.26550218340611,
+ "grad_norm": 0.013538471423089504,
+ "learning_rate": 0.0006,
+ "loss": 4.325181007385254,
+ "step": 5635
+ },
+ {
+ "epoch": 78.27947598253274,
+ "grad_norm": 0.013541961088776588,
+ "learning_rate": 0.0006,
+ "loss": 4.368722915649414,
+ "step": 5636
+ },
+ {
+ "epoch": 78.29344978165939,
+ "grad_norm": 0.013446723110973835,
+ "learning_rate": 0.0006,
+ "loss": 4.301422595977783,
+ "step": 5637
+ },
+ {
+ "epoch": 78.30742358078602,
+ "grad_norm": 0.014753809198737144,
+ "learning_rate": 0.0006,
+ "loss": 4.259127616882324,
+ "step": 5638
+ },
+ {
+ "epoch": 78.32139737991267,
+ "grad_norm": 0.013256008736789227,
+ "learning_rate": 0.0006,
+ "loss": 4.284032821655273,
+ "step": 5639
+ },
+ {
+ "epoch": 78.3353711790393,
+ "grad_norm": 0.014272629283368587,
+ "learning_rate": 0.0006,
+ "loss": 4.234989166259766,
+ "step": 5640
+ },
+ {
+ "epoch": 78.34934497816595,
+ "grad_norm": 0.014756826683878899,
+ "learning_rate": 0.0006,
+ "loss": 4.143499374389648,
+ "step": 5641
+ },
+ {
+ "epoch": 78.36331877729258,
+ "grad_norm": 0.016549205407500267,
+ "learning_rate": 0.0006,
+ "loss": 4.311034202575684,
+ "step": 5642
+ },
+ {
+ "epoch": 78.37729257641921,
+ "grad_norm": 0.017735283821821213,
+ "learning_rate": 0.0006,
+ "loss": 4.246767520904541,
+ "step": 5643
+ },
+ {
+ "epoch": 78.39126637554585,
+ "grad_norm": 0.017370613291859627,
+ "learning_rate": 0.0006,
+ "loss": 4.2460832595825195,
+ "step": 5644
+ },
+ {
+ "epoch": 78.40524017467249,
+ "grad_norm": 0.014642846770584583,
+ "learning_rate": 0.0006,
+ "loss": 4.303811550140381,
+ "step": 5645
+ },
+ {
+ "epoch": 78.41921397379913,
+ "grad_norm": 0.013445726595818996,
+ "learning_rate": 0.0006,
+ "loss": 4.25921630859375,
+ "step": 5646
+ },
+ {
+ "epoch": 78.43318777292576,
+ "grad_norm": 0.014254065230488777,
+ "learning_rate": 0.0006,
+ "loss": 4.129446506500244,
+ "step": 5647
+ },
+ {
+ "epoch": 78.44716157205241,
+ "grad_norm": 0.01883019506931305,
+ "learning_rate": 0.0006,
+ "loss": 4.191716194152832,
+ "step": 5648
+ },
+ {
+ "epoch": 78.46113537117904,
+ "grad_norm": 0.0204437468200922,
+ "learning_rate": 0.0006,
+ "loss": 4.333085536956787,
+ "step": 5649
+ },
+ {
+ "epoch": 78.47510917030567,
+ "grad_norm": 0.01652318798005581,
+ "learning_rate": 0.0006,
+ "loss": 4.135462760925293,
+ "step": 5650
+ },
+ {
+ "epoch": 78.48908296943232,
+ "grad_norm": 0.012753551825881004,
+ "learning_rate": 0.0006,
+ "loss": 4.195051193237305,
+ "step": 5651
+ },
+ {
+ "epoch": 78.50305676855895,
+ "grad_norm": 0.01499118935316801,
+ "learning_rate": 0.0006,
+ "loss": 4.157092094421387,
+ "step": 5652
+ },
+ {
+ "epoch": 78.5170305676856,
+ "grad_norm": 0.015383945778012276,
+ "learning_rate": 0.0006,
+ "loss": 4.253818035125732,
+ "step": 5653
+ },
+ {
+ "epoch": 78.53100436681223,
+ "grad_norm": 0.015721332281827927,
+ "learning_rate": 0.0006,
+ "loss": 4.228775978088379,
+ "step": 5654
+ },
+ {
+ "epoch": 78.54497816593886,
+ "grad_norm": 0.014595463871955872,
+ "learning_rate": 0.0006,
+ "loss": 4.135197639465332,
+ "step": 5655
+ },
+ {
+ "epoch": 78.5589519650655,
+ "grad_norm": 0.015059644356369972,
+ "learning_rate": 0.0006,
+ "loss": 4.255614280700684,
+ "step": 5656
+ },
+ {
+ "epoch": 78.57292576419214,
+ "grad_norm": 0.013833344914019108,
+ "learning_rate": 0.0006,
+ "loss": 4.301250457763672,
+ "step": 5657
+ },
+ {
+ "epoch": 78.58689956331878,
+ "grad_norm": 0.013665376231074333,
+ "learning_rate": 0.0006,
+ "loss": 4.1734232902526855,
+ "step": 5658
+ },
+ {
+ "epoch": 78.60087336244541,
+ "grad_norm": 0.01539086177945137,
+ "learning_rate": 0.0006,
+ "loss": 4.221459865570068,
+ "step": 5659
+ },
+ {
+ "epoch": 78.61484716157206,
+ "grad_norm": 0.013313917443156242,
+ "learning_rate": 0.0006,
+ "loss": 4.255619049072266,
+ "step": 5660
+ },
+ {
+ "epoch": 78.62882096069869,
+ "grad_norm": 0.01447741687297821,
+ "learning_rate": 0.0006,
+ "loss": 4.298219680786133,
+ "step": 5661
+ },
+ {
+ "epoch": 78.64279475982532,
+ "grad_norm": 0.012796915136277676,
+ "learning_rate": 0.0006,
+ "loss": 4.2542405128479,
+ "step": 5662
+ },
+ {
+ "epoch": 78.65676855895197,
+ "grad_norm": 0.014058141969144344,
+ "learning_rate": 0.0006,
+ "loss": 4.284477233886719,
+ "step": 5663
+ },
+ {
+ "epoch": 78.6707423580786,
+ "grad_norm": 0.012705069966614246,
+ "learning_rate": 0.0006,
+ "loss": 4.2826666831970215,
+ "step": 5664
+ },
+ {
+ "epoch": 78.68471615720524,
+ "grad_norm": 0.013245890848338604,
+ "learning_rate": 0.0006,
+ "loss": 4.371612548828125,
+ "step": 5665
+ },
+ {
+ "epoch": 78.69868995633188,
+ "grad_norm": 0.014192677102982998,
+ "learning_rate": 0.0006,
+ "loss": 4.267344951629639,
+ "step": 5666
+ },
+ {
+ "epoch": 78.71266375545852,
+ "grad_norm": 0.015346777625381947,
+ "learning_rate": 0.0006,
+ "loss": 4.222797393798828,
+ "step": 5667
+ },
+ {
+ "epoch": 78.72663755458515,
+ "grad_norm": 0.014068621210753918,
+ "learning_rate": 0.0006,
+ "loss": 4.172964572906494,
+ "step": 5668
+ },
+ {
+ "epoch": 78.74061135371178,
+ "grad_norm": 0.014364945702254772,
+ "learning_rate": 0.0006,
+ "loss": 4.259319305419922,
+ "step": 5669
+ },
+ {
+ "epoch": 78.75458515283843,
+ "grad_norm": 0.017337387427687645,
+ "learning_rate": 0.0006,
+ "loss": 4.29533576965332,
+ "step": 5670
+ },
+ {
+ "epoch": 78.76855895196506,
+ "grad_norm": 0.017416715621948242,
+ "learning_rate": 0.0006,
+ "loss": 4.340364456176758,
+ "step": 5671
+ },
+ {
+ "epoch": 78.78253275109171,
+ "grad_norm": 0.015925846993923187,
+ "learning_rate": 0.0006,
+ "loss": 4.177454948425293,
+ "step": 5672
+ },
+ {
+ "epoch": 78.79650655021834,
+ "grad_norm": 0.013130443170666695,
+ "learning_rate": 0.0006,
+ "loss": 4.277454853057861,
+ "step": 5673
+ },
+ {
+ "epoch": 78.81048034934499,
+ "grad_norm": 0.014628687873482704,
+ "learning_rate": 0.0006,
+ "loss": 4.273757457733154,
+ "step": 5674
+ },
+ {
+ "epoch": 78.82445414847162,
+ "grad_norm": 0.01601659320294857,
+ "learning_rate": 0.0006,
+ "loss": 4.281774520874023,
+ "step": 5675
+ },
+ {
+ "epoch": 78.83842794759825,
+ "grad_norm": 0.015443485230207443,
+ "learning_rate": 0.0006,
+ "loss": 4.199937343597412,
+ "step": 5676
+ },
+ {
+ "epoch": 78.8524017467249,
+ "grad_norm": 0.013934246264398098,
+ "learning_rate": 0.0006,
+ "loss": 4.331369876861572,
+ "step": 5677
+ },
+ {
+ "epoch": 78.86637554585153,
+ "grad_norm": 0.01572083681821823,
+ "learning_rate": 0.0006,
+ "loss": 4.2828474044799805,
+ "step": 5678
+ },
+ {
+ "epoch": 78.88034934497817,
+ "grad_norm": 0.013444120064377785,
+ "learning_rate": 0.0006,
+ "loss": 4.183859825134277,
+ "step": 5679
+ },
+ {
+ "epoch": 78.8943231441048,
+ "grad_norm": 0.01591254211962223,
+ "learning_rate": 0.0006,
+ "loss": 4.295256614685059,
+ "step": 5680
+ },
+ {
+ "epoch": 78.90829694323143,
+ "grad_norm": 0.017556684091687202,
+ "learning_rate": 0.0006,
+ "loss": 4.132272720336914,
+ "step": 5681
+ },
+ {
+ "epoch": 78.92227074235808,
+ "grad_norm": 0.018184322863817215,
+ "learning_rate": 0.0006,
+ "loss": 4.251728057861328,
+ "step": 5682
+ },
+ {
+ "epoch": 78.93624454148471,
+ "grad_norm": 0.016207285225391388,
+ "learning_rate": 0.0006,
+ "loss": 4.205544471740723,
+ "step": 5683
+ },
+ {
+ "epoch": 78.95021834061136,
+ "grad_norm": 0.014478376135230064,
+ "learning_rate": 0.0006,
+ "loss": 4.182462215423584,
+ "step": 5684
+ },
+ {
+ "epoch": 78.96419213973799,
+ "grad_norm": 0.014255235902965069,
+ "learning_rate": 0.0006,
+ "loss": 4.253366947174072,
+ "step": 5685
+ },
+ {
+ "epoch": 78.97816593886463,
+ "grad_norm": 0.01524396613240242,
+ "learning_rate": 0.0006,
+ "loss": 4.221786022186279,
+ "step": 5686
+ },
+ {
+ "epoch": 78.99213973799127,
+ "grad_norm": 0.015057693235576153,
+ "learning_rate": 0.0006,
+ "loss": 4.208454608917236,
+ "step": 5687
+ },
+ {
+ "epoch": 79.0,
+ "grad_norm": 0.015478022396564484,
+ "learning_rate": 0.0006,
+ "loss": 4.336383819580078,
+ "step": 5688
+ },
+ {
+ "epoch": 79.0,
+ "eval_loss": 4.641749382019043,
+ "eval_runtime": 56.8167,
+ "eval_samples_per_second": 42.98,
+ "eval_steps_per_second": 1.355,
+ "step": 5688
+ },
+ {
+ "epoch": 79.01397379912663,
+ "grad_norm": 0.014143316075205803,
+ "learning_rate": 0.0006,
+ "loss": 4.260974884033203,
+ "step": 5689
+ },
+ {
+ "epoch": 79.02794759825328,
+ "grad_norm": 0.013120883144438267,
+ "learning_rate": 0.0006,
+ "loss": 4.139886856079102,
+ "step": 5690
+ },
+ {
+ "epoch": 79.04192139737991,
+ "grad_norm": 0.015019661746919155,
+ "learning_rate": 0.0006,
+ "loss": 4.267004013061523,
+ "step": 5691
+ },
+ {
+ "epoch": 79.05589519650655,
+ "grad_norm": 0.014891284517943859,
+ "learning_rate": 0.0006,
+ "loss": 4.205019950866699,
+ "step": 5692
+ },
+ {
+ "epoch": 79.06986899563319,
+ "grad_norm": 0.015527728945016861,
+ "learning_rate": 0.0006,
+ "loss": 4.1513566970825195,
+ "step": 5693
+ },
+ {
+ "epoch": 79.08384279475983,
+ "grad_norm": 0.017527181655168533,
+ "learning_rate": 0.0006,
+ "loss": 4.148828983306885,
+ "step": 5694
+ },
+ {
+ "epoch": 79.09781659388646,
+ "grad_norm": 0.01632206328213215,
+ "learning_rate": 0.0006,
+ "loss": 4.195789813995361,
+ "step": 5695
+ },
+ {
+ "epoch": 79.1117903930131,
+ "grad_norm": 0.01590503193438053,
+ "learning_rate": 0.0006,
+ "loss": 4.199155330657959,
+ "step": 5696
+ },
+ {
+ "epoch": 79.12576419213974,
+ "grad_norm": 0.016489997506141663,
+ "learning_rate": 0.0006,
+ "loss": 4.155154705047607,
+ "step": 5697
+ },
+ {
+ "epoch": 79.13973799126637,
+ "grad_norm": 0.0161368977278471,
+ "learning_rate": 0.0006,
+ "loss": 4.164384841918945,
+ "step": 5698
+ },
+ {
+ "epoch": 79.15371179039302,
+ "grad_norm": 0.018574809655547142,
+ "learning_rate": 0.0006,
+ "loss": 4.1898193359375,
+ "step": 5699
+ },
+ {
+ "epoch": 79.16768558951965,
+ "grad_norm": 0.018987983465194702,
+ "learning_rate": 0.0006,
+ "loss": 4.219719409942627,
+ "step": 5700
+ },
+ {
+ "epoch": 79.18165938864628,
+ "grad_norm": 0.01668606884777546,
+ "learning_rate": 0.0006,
+ "loss": 4.152734756469727,
+ "step": 5701
+ },
+ {
+ "epoch": 79.19563318777293,
+ "grad_norm": 0.014046299271285534,
+ "learning_rate": 0.0006,
+ "loss": 4.152964115142822,
+ "step": 5702
+ },
+ {
+ "epoch": 79.20960698689956,
+ "grad_norm": 0.01628032699227333,
+ "learning_rate": 0.0006,
+ "loss": 4.299479961395264,
+ "step": 5703
+ },
+ {
+ "epoch": 79.2235807860262,
+ "grad_norm": 0.017998795956373215,
+ "learning_rate": 0.0006,
+ "loss": 4.231203079223633,
+ "step": 5704
+ },
+ {
+ "epoch": 79.23755458515284,
+ "grad_norm": 0.014799647964537144,
+ "learning_rate": 0.0006,
+ "loss": 4.3530192375183105,
+ "step": 5705
+ },
+ {
+ "epoch": 79.25152838427948,
+ "grad_norm": 0.014640678651630878,
+ "learning_rate": 0.0006,
+ "loss": 4.287322998046875,
+ "step": 5706
+ },
+ {
+ "epoch": 79.26550218340611,
+ "grad_norm": 0.0172616895288229,
+ "learning_rate": 0.0006,
+ "loss": 4.148009777069092,
+ "step": 5707
+ },
+ {
+ "epoch": 79.27947598253274,
+ "grad_norm": 0.01811790093779564,
+ "learning_rate": 0.0006,
+ "loss": 4.269433975219727,
+ "step": 5708
+ },
+ {
+ "epoch": 79.29344978165939,
+ "grad_norm": 0.01764000579714775,
+ "learning_rate": 0.0006,
+ "loss": 4.292638778686523,
+ "step": 5709
+ },
+ {
+ "epoch": 79.30742358078602,
+ "grad_norm": 0.018030935898423195,
+ "learning_rate": 0.0006,
+ "loss": 4.267889976501465,
+ "step": 5710
+ },
+ {
+ "epoch": 79.32139737991267,
+ "grad_norm": 0.017860891297459602,
+ "learning_rate": 0.0006,
+ "loss": 4.211110591888428,
+ "step": 5711
+ },
+ {
+ "epoch": 79.3353711790393,
+ "grad_norm": 0.017295073717832565,
+ "learning_rate": 0.0006,
+ "loss": 4.224645614624023,
+ "step": 5712
+ },
+ {
+ "epoch": 79.34934497816595,
+ "grad_norm": 0.015891101211309433,
+ "learning_rate": 0.0006,
+ "loss": 4.224797248840332,
+ "step": 5713
+ },
+ {
+ "epoch": 79.36331877729258,
+ "grad_norm": 0.014680417254567146,
+ "learning_rate": 0.0006,
+ "loss": 4.182242393493652,
+ "step": 5714
+ },
+ {
+ "epoch": 79.37729257641921,
+ "grad_norm": 0.016621140763163567,
+ "learning_rate": 0.0006,
+ "loss": 4.186334609985352,
+ "step": 5715
+ },
+ {
+ "epoch": 79.39126637554585,
+ "grad_norm": 0.016207249835133553,
+ "learning_rate": 0.0006,
+ "loss": 4.277074813842773,
+ "step": 5716
+ },
+ {
+ "epoch": 79.40524017467249,
+ "grad_norm": 0.015375478193163872,
+ "learning_rate": 0.0006,
+ "loss": 4.154317855834961,
+ "step": 5717
+ },
+ {
+ "epoch": 79.41921397379913,
+ "grad_norm": 0.013453228399157524,
+ "learning_rate": 0.0006,
+ "loss": 4.265270233154297,
+ "step": 5718
+ },
+ {
+ "epoch": 79.43318777292576,
+ "grad_norm": 0.014312694780528545,
+ "learning_rate": 0.0006,
+ "loss": 4.232220649719238,
+ "step": 5719
+ },
+ {
+ "epoch": 79.44716157205241,
+ "grad_norm": 0.014759554527699947,
+ "learning_rate": 0.0006,
+ "loss": 4.272275924682617,
+ "step": 5720
+ },
+ {
+ "epoch": 79.46113537117904,
+ "grad_norm": 0.015885505825281143,
+ "learning_rate": 0.0006,
+ "loss": 4.2229719161987305,
+ "step": 5721
+ },
+ {
+ "epoch": 79.47510917030567,
+ "grad_norm": 0.016516581177711487,
+ "learning_rate": 0.0006,
+ "loss": 4.200624465942383,
+ "step": 5722
+ },
+ {
+ "epoch": 79.48908296943232,
+ "grad_norm": 0.018254555761814117,
+ "learning_rate": 0.0006,
+ "loss": 4.19648551940918,
+ "step": 5723
+ },
+ {
+ "epoch": 79.50305676855895,
+ "grad_norm": 0.01467090006917715,
+ "learning_rate": 0.0006,
+ "loss": 4.135190486907959,
+ "step": 5724
+ },
+ {
+ "epoch": 79.5170305676856,
+ "grad_norm": 0.013474797829985619,
+ "learning_rate": 0.0006,
+ "loss": 4.311361312866211,
+ "step": 5725
+ },
+ {
+ "epoch": 79.53100436681223,
+ "grad_norm": 0.01742764748632908,
+ "learning_rate": 0.0006,
+ "loss": 4.1845903396606445,
+ "step": 5726
+ },
+ {
+ "epoch": 79.54497816593886,
+ "grad_norm": 0.018786882981657982,
+ "learning_rate": 0.0006,
+ "loss": 4.163942337036133,
+ "step": 5727
+ },
+ {
+ "epoch": 79.5589519650655,
+ "grad_norm": 0.017831219360232353,
+ "learning_rate": 0.0006,
+ "loss": 4.236685276031494,
+ "step": 5728
+ },
+ {
+ "epoch": 79.57292576419214,
+ "grad_norm": 0.016334986314177513,
+ "learning_rate": 0.0006,
+ "loss": 4.255853652954102,
+ "step": 5729
+ },
+ {
+ "epoch": 79.58689956331878,
+ "grad_norm": 0.0138487434014678,
+ "learning_rate": 0.0006,
+ "loss": 4.227874279022217,
+ "step": 5730
+ },
+ {
+ "epoch": 79.60087336244541,
+ "grad_norm": 0.01531760673969984,
+ "learning_rate": 0.0006,
+ "loss": 4.221432685852051,
+ "step": 5731
+ },
+ {
+ "epoch": 79.61484716157206,
+ "grad_norm": 0.01603361964225769,
+ "learning_rate": 0.0006,
+ "loss": 4.278326511383057,
+ "step": 5732
+ },
+ {
+ "epoch": 79.62882096069869,
+ "grad_norm": 0.013149895705282688,
+ "learning_rate": 0.0006,
+ "loss": 4.2474517822265625,
+ "step": 5733
+ },
+ {
+ "epoch": 79.64279475982532,
+ "grad_norm": 0.013428786769509315,
+ "learning_rate": 0.0006,
+ "loss": 4.1701178550720215,
+ "step": 5734
+ },
+ {
+ "epoch": 79.65676855895197,
+ "grad_norm": 0.014305585995316505,
+ "learning_rate": 0.0006,
+ "loss": 4.250778675079346,
+ "step": 5735
+ },
+ {
+ "epoch": 79.6707423580786,
+ "grad_norm": 0.014544566161930561,
+ "learning_rate": 0.0006,
+ "loss": 4.243236541748047,
+ "step": 5736
+ },
+ {
+ "epoch": 79.68471615720524,
+ "grad_norm": 0.013524414040148258,
+ "learning_rate": 0.0006,
+ "loss": 4.255457878112793,
+ "step": 5737
+ },
+ {
+ "epoch": 79.69868995633188,
+ "grad_norm": 0.014750641770660877,
+ "learning_rate": 0.0006,
+ "loss": 4.202816486358643,
+ "step": 5738
+ },
+ {
+ "epoch": 79.71266375545852,
+ "grad_norm": 0.015090773813426495,
+ "learning_rate": 0.0006,
+ "loss": 4.3053178787231445,
+ "step": 5739
+ },
+ {
+ "epoch": 79.72663755458515,
+ "grad_norm": 0.01606169529259205,
+ "learning_rate": 0.0006,
+ "loss": 4.224017143249512,
+ "step": 5740
+ },
+ {
+ "epoch": 79.74061135371178,
+ "grad_norm": 0.016303520649671555,
+ "learning_rate": 0.0006,
+ "loss": 4.190760612487793,
+ "step": 5741
+ },
+ {
+ "epoch": 79.75458515283843,
+ "grad_norm": 0.01465672068297863,
+ "learning_rate": 0.0006,
+ "loss": 4.24506950378418,
+ "step": 5742
+ },
+ {
+ "epoch": 79.76855895196506,
+ "grad_norm": 0.012616301886737347,
+ "learning_rate": 0.0006,
+ "loss": 4.174039840698242,
+ "step": 5743
+ },
+ {
+ "epoch": 79.78253275109171,
+ "grad_norm": 0.01358284242451191,
+ "learning_rate": 0.0006,
+ "loss": 4.236380577087402,
+ "step": 5744
+ },
+ {
+ "epoch": 79.79650655021834,
+ "grad_norm": 0.01557177770882845,
+ "learning_rate": 0.0006,
+ "loss": 4.102962493896484,
+ "step": 5745
+ },
+ {
+ "epoch": 79.81048034934499,
+ "grad_norm": 0.015584317035973072,
+ "learning_rate": 0.0006,
+ "loss": 4.242603302001953,
+ "step": 5746
+ },
+ {
+ "epoch": 79.82445414847162,
+ "grad_norm": 0.012597334571182728,
+ "learning_rate": 0.0006,
+ "loss": 4.202791690826416,
+ "step": 5747
+ },
+ {
+ "epoch": 79.83842794759825,
+ "grad_norm": 0.013722898438572884,
+ "learning_rate": 0.0006,
+ "loss": 4.204328536987305,
+ "step": 5748
+ },
+ {
+ "epoch": 79.8524017467249,
+ "grad_norm": 0.014669225551187992,
+ "learning_rate": 0.0006,
+ "loss": 4.164331912994385,
+ "step": 5749
+ },
+ {
+ "epoch": 79.86637554585153,
+ "grad_norm": 0.015012740157544613,
+ "learning_rate": 0.0006,
+ "loss": 4.260565280914307,
+ "step": 5750
+ },
+ {
+ "epoch": 79.88034934497817,
+ "grad_norm": 0.016026340425014496,
+ "learning_rate": 0.0006,
+ "loss": 4.299723148345947,
+ "step": 5751
+ },
+ {
+ "epoch": 79.8943231441048,
+ "grad_norm": 0.01626049168407917,
+ "learning_rate": 0.0006,
+ "loss": 4.207252502441406,
+ "step": 5752
+ },
+ {
+ "epoch": 79.90829694323143,
+ "grad_norm": 0.015415354631841183,
+ "learning_rate": 0.0006,
+ "loss": 4.151934623718262,
+ "step": 5753
+ },
+ {
+ "epoch": 79.92227074235808,
+ "grad_norm": 0.014158242382109165,
+ "learning_rate": 0.0006,
+ "loss": 4.3461995124816895,
+ "step": 5754
+ },
+ {
+ "epoch": 79.93624454148471,
+ "grad_norm": 0.01585628278553486,
+ "learning_rate": 0.0006,
+ "loss": 4.334761619567871,
+ "step": 5755
+ },
+ {
+ "epoch": 79.95021834061136,
+ "grad_norm": 0.013710759580135345,
+ "learning_rate": 0.0006,
+ "loss": 4.180401802062988,
+ "step": 5756
+ },
+ {
+ "epoch": 79.96419213973799,
+ "grad_norm": 0.014159204438328743,
+ "learning_rate": 0.0006,
+ "loss": 4.322331428527832,
+ "step": 5757
+ },
+ {
+ "epoch": 79.97816593886463,
+ "grad_norm": 0.015153195708990097,
+ "learning_rate": 0.0006,
+ "loss": 4.287478923797607,
+ "step": 5758
+ },
+ {
+ "epoch": 79.99213973799127,
+ "grad_norm": 0.014100227504968643,
+ "learning_rate": 0.0006,
+ "loss": 4.231807708740234,
+ "step": 5759
+ },
+ {
+ "epoch": 80.0,
+ "grad_norm": 0.015248682349920273,
+ "learning_rate": 0.0006,
+ "loss": 4.261106014251709,
+ "step": 5760
+ },
+ {
+ "epoch": 80.0,
+ "eval_loss": 4.610264301300049,
+ "eval_runtime": 56.514,
+ "eval_samples_per_second": 43.211,
+ "eval_steps_per_second": 1.362,
+ "step": 5760
+ },
+ {
+ "epoch": 80.01397379912663,
+ "grad_norm": 0.014730271883308887,
+ "learning_rate": 0.0006,
+ "loss": 4.108556747436523,
+ "step": 5761
+ },
+ {
+ "epoch": 80.02794759825328,
+ "grad_norm": 0.014575716108083725,
+ "learning_rate": 0.0006,
+ "loss": 4.169658660888672,
+ "step": 5762
+ },
+ {
+ "epoch": 80.04192139737991,
+ "grad_norm": 0.014706958085298538,
+ "learning_rate": 0.0006,
+ "loss": 4.252662658691406,
+ "step": 5763
+ },
+ {
+ "epoch": 80.05589519650655,
+ "grad_norm": 0.01561224739998579,
+ "learning_rate": 0.0006,
+ "loss": 4.193892955780029,
+ "step": 5764
+ },
+ {
+ "epoch": 80.06986899563319,
+ "grad_norm": 0.014702324755489826,
+ "learning_rate": 0.0006,
+ "loss": 4.1889238357543945,
+ "step": 5765
+ },
+ {
+ "epoch": 80.08384279475983,
+ "grad_norm": 0.015565741807222366,
+ "learning_rate": 0.0006,
+ "loss": 4.243549823760986,
+ "step": 5766
+ },
+ {
+ "epoch": 80.09781659388646,
+ "grad_norm": 0.01647181063890457,
+ "learning_rate": 0.0006,
+ "loss": 4.275446891784668,
+ "step": 5767
+ },
+ {
+ "epoch": 80.1117903930131,
+ "grad_norm": 0.018177716061472893,
+ "learning_rate": 0.0006,
+ "loss": 4.12040376663208,
+ "step": 5768
+ },
+ {
+ "epoch": 80.12576419213974,
+ "grad_norm": 0.01745191030204296,
+ "learning_rate": 0.0006,
+ "loss": 4.140371322631836,
+ "step": 5769
+ },
+ {
+ "epoch": 80.13973799126637,
+ "grad_norm": 0.01568993367254734,
+ "learning_rate": 0.0006,
+ "loss": 4.258280277252197,
+ "step": 5770
+ },
+ {
+ "epoch": 80.15371179039302,
+ "grad_norm": 0.01636597141623497,
+ "learning_rate": 0.0006,
+ "loss": 4.179052352905273,
+ "step": 5771
+ },
+ {
+ "epoch": 80.16768558951965,
+ "grad_norm": 0.016839100047945976,
+ "learning_rate": 0.0006,
+ "loss": 4.127636432647705,
+ "step": 5772
+ },
+ {
+ "epoch": 80.18165938864628,
+ "grad_norm": 0.015471016988158226,
+ "learning_rate": 0.0006,
+ "loss": 4.181325912475586,
+ "step": 5773
+ },
+ {
+ "epoch": 80.19563318777293,
+ "grad_norm": 0.015836089849472046,
+ "learning_rate": 0.0006,
+ "loss": 4.142168998718262,
+ "step": 5774
+ },
+ {
+ "epoch": 80.20960698689956,
+ "grad_norm": 0.014782954007387161,
+ "learning_rate": 0.0006,
+ "loss": 4.308799743652344,
+ "step": 5775
+ },
+ {
+ "epoch": 80.2235807860262,
+ "grad_norm": 0.014040893875062466,
+ "learning_rate": 0.0006,
+ "loss": 4.216936111450195,
+ "step": 5776
+ },
+ {
+ "epoch": 80.23755458515284,
+ "grad_norm": 0.01420280709862709,
+ "learning_rate": 0.0006,
+ "loss": 4.235230445861816,
+ "step": 5777
+ },
+ {
+ "epoch": 80.25152838427948,
+ "grad_norm": 0.013341007754206657,
+ "learning_rate": 0.0006,
+ "loss": 4.369109153747559,
+ "step": 5778
+ },
+ {
+ "epoch": 80.26550218340611,
+ "grad_norm": 0.014208327978849411,
+ "learning_rate": 0.0006,
+ "loss": 4.219112396240234,
+ "step": 5779
+ },
+ {
+ "epoch": 80.27947598253274,
+ "grad_norm": 0.013825052417814732,
+ "learning_rate": 0.0006,
+ "loss": 4.1983489990234375,
+ "step": 5780
+ },
+ {
+ "epoch": 80.29344978165939,
+ "grad_norm": 0.012916515581309795,
+ "learning_rate": 0.0006,
+ "loss": 4.204582691192627,
+ "step": 5781
+ },
+ {
+ "epoch": 80.30742358078602,
+ "grad_norm": 0.013071243651211262,
+ "learning_rate": 0.0006,
+ "loss": 4.17643928527832,
+ "step": 5782
+ },
+ {
+ "epoch": 80.32139737991267,
+ "grad_norm": 0.015507174655795097,
+ "learning_rate": 0.0006,
+ "loss": 4.172976016998291,
+ "step": 5783
+ },
+ {
+ "epoch": 80.3353711790393,
+ "grad_norm": 0.014967243187129498,
+ "learning_rate": 0.0006,
+ "loss": 4.141168594360352,
+ "step": 5784
+ },
+ {
+ "epoch": 80.34934497816595,
+ "grad_norm": 0.015629924833774567,
+ "learning_rate": 0.0006,
+ "loss": 4.253526210784912,
+ "step": 5785
+ },
+ {
+ "epoch": 80.36331877729258,
+ "grad_norm": 0.01607954502105713,
+ "learning_rate": 0.0006,
+ "loss": 4.249395370483398,
+ "step": 5786
+ },
+ {
+ "epoch": 80.37729257641921,
+ "grad_norm": 0.01500706560909748,
+ "learning_rate": 0.0006,
+ "loss": 4.279012680053711,
+ "step": 5787
+ },
+ {
+ "epoch": 80.39126637554585,
+ "grad_norm": 0.014343900606036186,
+ "learning_rate": 0.0006,
+ "loss": 4.268509864807129,
+ "step": 5788
+ },
+ {
+ "epoch": 80.40524017467249,
+ "grad_norm": 0.014682197012007236,
+ "learning_rate": 0.0006,
+ "loss": 4.206334114074707,
+ "step": 5789
+ },
+ {
+ "epoch": 80.41921397379913,
+ "grad_norm": 0.014989163726568222,
+ "learning_rate": 0.0006,
+ "loss": 4.143601417541504,
+ "step": 5790
+ },
+ {
+ "epoch": 80.43318777292576,
+ "grad_norm": 0.015001608058810234,
+ "learning_rate": 0.0006,
+ "loss": 4.164111137390137,
+ "step": 5791
+ },
+ {
+ "epoch": 80.44716157205241,
+ "grad_norm": 0.015222722664475441,
+ "learning_rate": 0.0006,
+ "loss": 4.2369914054870605,
+ "step": 5792
+ },
+ {
+ "epoch": 80.46113537117904,
+ "grad_norm": 0.014126183465123177,
+ "learning_rate": 0.0006,
+ "loss": 4.2209930419921875,
+ "step": 5793
+ },
+ {
+ "epoch": 80.47510917030567,
+ "grad_norm": 0.013065959326922894,
+ "learning_rate": 0.0006,
+ "loss": 4.231184959411621,
+ "step": 5794
+ },
+ {
+ "epoch": 80.48908296943232,
+ "grad_norm": 0.013833962380886078,
+ "learning_rate": 0.0006,
+ "loss": 4.226434707641602,
+ "step": 5795
+ },
+ {
+ "epoch": 80.50305676855895,
+ "grad_norm": 0.01387125626206398,
+ "learning_rate": 0.0006,
+ "loss": 4.241328716278076,
+ "step": 5796
+ },
+ {
+ "epoch": 80.5170305676856,
+ "grad_norm": 0.014145063236355782,
+ "learning_rate": 0.0006,
+ "loss": 4.2268805503845215,
+ "step": 5797
+ },
+ {
+ "epoch": 80.53100436681223,
+ "grad_norm": 0.012990240007638931,
+ "learning_rate": 0.0006,
+ "loss": 4.1718339920043945,
+ "step": 5798
+ },
+ {
+ "epoch": 80.54497816593886,
+ "grad_norm": 0.014185411855578423,
+ "learning_rate": 0.0006,
+ "loss": 4.232397556304932,
+ "step": 5799
+ },
+ {
+ "epoch": 80.5589519650655,
+ "grad_norm": 0.014260523952543736,
+ "learning_rate": 0.0006,
+ "loss": 4.282734394073486,
+ "step": 5800
+ },
+ {
+ "epoch": 80.57292576419214,
+ "grad_norm": 0.012542120181024075,
+ "learning_rate": 0.0006,
+ "loss": 4.134551525115967,
+ "step": 5801
+ },
+ {
+ "epoch": 80.58689956331878,
+ "grad_norm": 0.013078657910227776,
+ "learning_rate": 0.0006,
+ "loss": 4.229815483093262,
+ "step": 5802
+ },
+ {
+ "epoch": 80.60087336244541,
+ "grad_norm": 0.012821056880056858,
+ "learning_rate": 0.0006,
+ "loss": 4.208278179168701,
+ "step": 5803
+ },
+ {
+ "epoch": 80.61484716157206,
+ "grad_norm": 0.012964180670678616,
+ "learning_rate": 0.0006,
+ "loss": 4.188477039337158,
+ "step": 5804
+ },
+ {
+ "epoch": 80.62882096069869,
+ "grad_norm": 0.01631746254861355,
+ "learning_rate": 0.0006,
+ "loss": 4.3214497566223145,
+ "step": 5805
+ },
+ {
+ "epoch": 80.64279475982532,
+ "grad_norm": 0.01726640947163105,
+ "learning_rate": 0.0006,
+ "loss": 4.279359817504883,
+ "step": 5806
+ },
+ {
+ "epoch": 80.65676855895197,
+ "grad_norm": 0.018778258934617043,
+ "learning_rate": 0.0006,
+ "loss": 4.294047832489014,
+ "step": 5807
+ },
+ {
+ "epoch": 80.6707423580786,
+ "grad_norm": 0.01796831749379635,
+ "learning_rate": 0.0006,
+ "loss": 4.314162254333496,
+ "step": 5808
+ },
+ {
+ "epoch": 80.68471615720524,
+ "grad_norm": 0.016495538875460625,
+ "learning_rate": 0.0006,
+ "loss": 4.273871421813965,
+ "step": 5809
+ },
+ {
+ "epoch": 80.69868995633188,
+ "grad_norm": 0.013590306043624878,
+ "learning_rate": 0.0006,
+ "loss": 4.14961051940918,
+ "step": 5810
+ },
+ {
+ "epoch": 80.71266375545852,
+ "grad_norm": 0.014670426957309246,
+ "learning_rate": 0.0006,
+ "loss": 4.211853981018066,
+ "step": 5811
+ },
+ {
+ "epoch": 80.72663755458515,
+ "grad_norm": 0.01641341857612133,
+ "learning_rate": 0.0006,
+ "loss": 4.211948871612549,
+ "step": 5812
+ },
+ {
+ "epoch": 80.74061135371178,
+ "grad_norm": 0.016365470364689827,
+ "learning_rate": 0.0006,
+ "loss": 4.200774192810059,
+ "step": 5813
+ },
+ {
+ "epoch": 80.75458515283843,
+ "grad_norm": 0.01613970287144184,
+ "learning_rate": 0.0006,
+ "loss": 4.288322448730469,
+ "step": 5814
+ },
+ {
+ "epoch": 80.76855895196506,
+ "grad_norm": 0.016742777079343796,
+ "learning_rate": 0.0006,
+ "loss": 4.23801326751709,
+ "step": 5815
+ },
+ {
+ "epoch": 80.78253275109171,
+ "grad_norm": 0.015351069159805775,
+ "learning_rate": 0.0006,
+ "loss": 4.263832092285156,
+ "step": 5816
+ },
+ {
+ "epoch": 80.79650655021834,
+ "grad_norm": 0.013435336761176586,
+ "learning_rate": 0.0006,
+ "loss": 4.160857200622559,
+ "step": 5817
+ },
+ {
+ "epoch": 80.81048034934499,
+ "grad_norm": 0.01474886853247881,
+ "learning_rate": 0.0006,
+ "loss": 4.216025352478027,
+ "step": 5818
+ },
+ {
+ "epoch": 80.82445414847162,
+ "grad_norm": 0.01586935855448246,
+ "learning_rate": 0.0006,
+ "loss": 4.256402969360352,
+ "step": 5819
+ },
+ {
+ "epoch": 80.83842794759825,
+ "grad_norm": 0.014596572145819664,
+ "learning_rate": 0.0006,
+ "loss": 4.142584800720215,
+ "step": 5820
+ },
+ {
+ "epoch": 80.8524017467249,
+ "grad_norm": 0.01501593366265297,
+ "learning_rate": 0.0006,
+ "loss": 4.250462055206299,
+ "step": 5821
+ },
+ {
+ "epoch": 80.86637554585153,
+ "grad_norm": 0.014773511327803135,
+ "learning_rate": 0.0006,
+ "loss": 4.253305912017822,
+ "step": 5822
+ },
+ {
+ "epoch": 80.88034934497817,
+ "grad_norm": 0.013246326707303524,
+ "learning_rate": 0.0006,
+ "loss": 4.170881271362305,
+ "step": 5823
+ },
+ {
+ "epoch": 80.8943231441048,
+ "grad_norm": 0.013013756833970547,
+ "learning_rate": 0.0006,
+ "loss": 4.264721870422363,
+ "step": 5824
+ },
+ {
+ "epoch": 80.90829694323143,
+ "grad_norm": 0.013155903667211533,
+ "learning_rate": 0.0006,
+ "loss": 4.274667739868164,
+ "step": 5825
+ },
+ {
+ "epoch": 80.92227074235808,
+ "grad_norm": 0.01448952779173851,
+ "learning_rate": 0.0006,
+ "loss": 4.24777889251709,
+ "step": 5826
+ },
+ {
+ "epoch": 80.93624454148471,
+ "grad_norm": 0.01415231078863144,
+ "learning_rate": 0.0006,
+ "loss": 4.22014856338501,
+ "step": 5827
+ },
+ {
+ "epoch": 80.95021834061136,
+ "grad_norm": 0.014796273782849312,
+ "learning_rate": 0.0006,
+ "loss": 4.22036600112915,
+ "step": 5828
+ },
+ {
+ "epoch": 80.96419213973799,
+ "grad_norm": 0.013005326502025127,
+ "learning_rate": 0.0006,
+ "loss": 4.276123046875,
+ "step": 5829
+ },
+ {
+ "epoch": 80.97816593886463,
+ "grad_norm": 0.012851177714765072,
+ "learning_rate": 0.0006,
+ "loss": 4.248575687408447,
+ "step": 5830
+ },
+ {
+ "epoch": 80.99213973799127,
+ "grad_norm": 0.013740002177655697,
+ "learning_rate": 0.0006,
+ "loss": 4.051140308380127,
+ "step": 5831
+ },
+ {
+ "epoch": 81.0,
+ "grad_norm": 0.015296104364097118,
+ "learning_rate": 0.0006,
+ "loss": 4.251155853271484,
+ "step": 5832
+ },
+ {
+ "epoch": 81.0,
+ "eval_loss": 4.594127655029297,
+ "eval_runtime": 56.0959,
+ "eval_samples_per_second": 43.533,
+ "eval_steps_per_second": 1.373,
+ "step": 5832
+ },
+ {
+ "epoch": 81.01397379912663,
+ "grad_norm": 0.015804558992385864,
+ "learning_rate": 0.0006,
+ "loss": 4.152894020080566,
+ "step": 5833
+ },
+ {
+ "epoch": 81.02794759825328,
+ "grad_norm": 0.017000854015350342,
+ "learning_rate": 0.0006,
+ "loss": 4.266067981719971,
+ "step": 5834
+ },
+ {
+ "epoch": 81.04192139737991,
+ "grad_norm": 0.016351008787751198,
+ "learning_rate": 0.0006,
+ "loss": 4.192703723907471,
+ "step": 5835
+ },
+ {
+ "epoch": 81.05589519650655,
+ "grad_norm": 0.014703895896673203,
+ "learning_rate": 0.0006,
+ "loss": 4.23554801940918,
+ "step": 5836
+ },
+ {
+ "epoch": 81.06986899563319,
+ "grad_norm": 0.01619596965610981,
+ "learning_rate": 0.0006,
+ "loss": 4.233000755310059,
+ "step": 5837
+ },
+ {
+ "epoch": 81.08384279475983,
+ "grad_norm": 0.01815493032336235,
+ "learning_rate": 0.0006,
+ "loss": 4.200490951538086,
+ "step": 5838
+ },
+ {
+ "epoch": 81.09781659388646,
+ "grad_norm": 0.018305106088519096,
+ "learning_rate": 0.0006,
+ "loss": 4.154242515563965,
+ "step": 5839
+ },
+ {
+ "epoch": 81.1117903930131,
+ "grad_norm": 0.0200856514275074,
+ "learning_rate": 0.0006,
+ "loss": 4.25820779800415,
+ "step": 5840
+ },
+ {
+ "epoch": 81.12576419213974,
+ "grad_norm": 0.02101440541446209,
+ "learning_rate": 0.0006,
+ "loss": 4.168209075927734,
+ "step": 5841
+ },
+ {
+ "epoch": 81.13973799126637,
+ "grad_norm": 0.022829681634902954,
+ "learning_rate": 0.0006,
+ "loss": 4.25560998916626,
+ "step": 5842
+ },
+ {
+ "epoch": 81.15371179039302,
+ "grad_norm": 0.0195082426071167,
+ "learning_rate": 0.0006,
+ "loss": 4.2035627365112305,
+ "step": 5843
+ },
+ {
+ "epoch": 81.16768558951965,
+ "grad_norm": 0.01716248318552971,
+ "learning_rate": 0.0006,
+ "loss": 4.11322021484375,
+ "step": 5844
+ },
+ {
+ "epoch": 81.18165938864628,
+ "grad_norm": 0.02034664712846279,
+ "learning_rate": 0.0006,
+ "loss": 4.179389953613281,
+ "step": 5845
+ },
+ {
+ "epoch": 81.19563318777293,
+ "grad_norm": 0.023578710854053497,
+ "learning_rate": 0.0006,
+ "loss": 4.265695571899414,
+ "step": 5846
+ },
+ {
+ "epoch": 81.20960698689956,
+ "grad_norm": 0.019244182854890823,
+ "learning_rate": 0.0006,
+ "loss": 4.292043209075928,
+ "step": 5847
+ },
+ {
+ "epoch": 81.2235807860262,
+ "grad_norm": 0.01577291637659073,
+ "learning_rate": 0.0006,
+ "loss": 4.250776290893555,
+ "step": 5848
+ },
+ {
+ "epoch": 81.23755458515284,
+ "grad_norm": 0.01617090217769146,
+ "learning_rate": 0.0006,
+ "loss": 4.193098068237305,
+ "step": 5849
+ },
+ {
+ "epoch": 81.25152838427948,
+ "grad_norm": 0.016090184450149536,
+ "learning_rate": 0.0006,
+ "loss": 4.112357139587402,
+ "step": 5850
+ },
+ {
+ "epoch": 81.26550218340611,
+ "grad_norm": 0.01441365759819746,
+ "learning_rate": 0.0006,
+ "loss": 4.114147186279297,
+ "step": 5851
+ },
+ {
+ "epoch": 81.27947598253274,
+ "grad_norm": 0.014122162945568562,
+ "learning_rate": 0.0006,
+ "loss": 4.276175022125244,
+ "step": 5852
+ },
+ {
+ "epoch": 81.29344978165939,
+ "grad_norm": 0.014859487302601337,
+ "learning_rate": 0.0006,
+ "loss": 4.094699859619141,
+ "step": 5853
+ },
+ {
+ "epoch": 81.30742358078602,
+ "grad_norm": 0.013624267652630806,
+ "learning_rate": 0.0006,
+ "loss": 4.2648773193359375,
+ "step": 5854
+ },
+ {
+ "epoch": 81.32139737991267,
+ "grad_norm": 0.01417913381010294,
+ "learning_rate": 0.0006,
+ "loss": 4.228567123413086,
+ "step": 5855
+ },
+ {
+ "epoch": 81.3353711790393,
+ "grad_norm": 0.014560805633664131,
+ "learning_rate": 0.0006,
+ "loss": 4.148930549621582,
+ "step": 5856
+ },
+ {
+ "epoch": 81.34934497816595,
+ "grad_norm": 0.013662154786288738,
+ "learning_rate": 0.0006,
+ "loss": 4.192107677459717,
+ "step": 5857
+ },
+ {
+ "epoch": 81.36331877729258,
+ "grad_norm": 0.013434980995953083,
+ "learning_rate": 0.0006,
+ "loss": 4.162052154541016,
+ "step": 5858
+ },
+ {
+ "epoch": 81.37729257641921,
+ "grad_norm": 0.01300742756575346,
+ "learning_rate": 0.0006,
+ "loss": 4.262385368347168,
+ "step": 5859
+ },
+ {
+ "epoch": 81.39126637554585,
+ "grad_norm": 0.012544351629912853,
+ "learning_rate": 0.0006,
+ "loss": 4.215371131896973,
+ "step": 5860
+ },
+ {
+ "epoch": 81.40524017467249,
+ "grad_norm": 0.013663645833730698,
+ "learning_rate": 0.0006,
+ "loss": 4.247312545776367,
+ "step": 5861
+ },
+ {
+ "epoch": 81.41921397379913,
+ "grad_norm": 0.013127296231687069,
+ "learning_rate": 0.0006,
+ "loss": 4.087685585021973,
+ "step": 5862
+ },
+ {
+ "epoch": 81.43318777292576,
+ "grad_norm": 0.014081938192248344,
+ "learning_rate": 0.0006,
+ "loss": 4.272648811340332,
+ "step": 5863
+ },
+ {
+ "epoch": 81.44716157205241,
+ "grad_norm": 0.015470641665160656,
+ "learning_rate": 0.0006,
+ "loss": 4.278578758239746,
+ "step": 5864
+ },
+ {
+ "epoch": 81.46113537117904,
+ "grad_norm": 0.016524532809853554,
+ "learning_rate": 0.0006,
+ "loss": 4.118268013000488,
+ "step": 5865
+ },
+ {
+ "epoch": 81.47510917030567,
+ "grad_norm": 0.016242722049355507,
+ "learning_rate": 0.0006,
+ "loss": 4.182405471801758,
+ "step": 5866
+ },
+ {
+ "epoch": 81.48908296943232,
+ "grad_norm": 0.01557017583400011,
+ "learning_rate": 0.0006,
+ "loss": 4.229862213134766,
+ "step": 5867
+ },
+ {
+ "epoch": 81.50305676855895,
+ "grad_norm": 0.01562870666384697,
+ "learning_rate": 0.0006,
+ "loss": 4.27754020690918,
+ "step": 5868
+ },
+ {
+ "epoch": 81.5170305676856,
+ "grad_norm": 0.018606074154376984,
+ "learning_rate": 0.0006,
+ "loss": 4.148681640625,
+ "step": 5869
+ },
+ {
+ "epoch": 81.53100436681223,
+ "grad_norm": 0.017740730196237564,
+ "learning_rate": 0.0006,
+ "loss": 4.321932792663574,
+ "step": 5870
+ },
+ {
+ "epoch": 81.54497816593886,
+ "grad_norm": 0.014517677947878838,
+ "learning_rate": 0.0006,
+ "loss": 4.22862434387207,
+ "step": 5871
+ },
+ {
+ "epoch": 81.5589519650655,
+ "grad_norm": 0.01389714702963829,
+ "learning_rate": 0.0006,
+ "loss": 4.198000907897949,
+ "step": 5872
+ },
+ {
+ "epoch": 81.57292576419214,
+ "grad_norm": 0.015348682180047035,
+ "learning_rate": 0.0006,
+ "loss": 4.293485164642334,
+ "step": 5873
+ },
+ {
+ "epoch": 81.58689956331878,
+ "grad_norm": 0.015965791419148445,
+ "learning_rate": 0.0006,
+ "loss": 4.23481559753418,
+ "step": 5874
+ },
+ {
+ "epoch": 81.60087336244541,
+ "grad_norm": 0.015662726014852524,
+ "learning_rate": 0.0006,
+ "loss": 4.158148765563965,
+ "step": 5875
+ },
+ {
+ "epoch": 81.61484716157206,
+ "grad_norm": 0.014402080327272415,
+ "learning_rate": 0.0006,
+ "loss": 4.132723808288574,
+ "step": 5876
+ },
+ {
+ "epoch": 81.62882096069869,
+ "grad_norm": 0.016846125945448875,
+ "learning_rate": 0.0006,
+ "loss": 4.1654462814331055,
+ "step": 5877
+ },
+ {
+ "epoch": 81.64279475982532,
+ "grad_norm": 0.0196097232401371,
+ "learning_rate": 0.0006,
+ "loss": 4.228450298309326,
+ "step": 5878
+ },
+ {
+ "epoch": 81.65676855895197,
+ "grad_norm": 0.016005294397473335,
+ "learning_rate": 0.0006,
+ "loss": 4.230222225189209,
+ "step": 5879
+ },
+ {
+ "epoch": 81.6707423580786,
+ "grad_norm": 0.014685059897601604,
+ "learning_rate": 0.0006,
+ "loss": 4.1367387771606445,
+ "step": 5880
+ },
+ {
+ "epoch": 81.68471615720524,
+ "grad_norm": 0.01647144928574562,
+ "learning_rate": 0.0006,
+ "loss": 4.096675395965576,
+ "step": 5881
+ },
+ {
+ "epoch": 81.69868995633188,
+ "grad_norm": 0.015978610143065453,
+ "learning_rate": 0.0006,
+ "loss": 4.232292175292969,
+ "step": 5882
+ },
+ {
+ "epoch": 81.71266375545852,
+ "grad_norm": 0.01510288193821907,
+ "learning_rate": 0.0006,
+ "loss": 4.274591445922852,
+ "step": 5883
+ },
+ {
+ "epoch": 81.72663755458515,
+ "grad_norm": 0.014942476525902748,
+ "learning_rate": 0.0006,
+ "loss": 4.161482810974121,
+ "step": 5884
+ },
+ {
+ "epoch": 81.74061135371178,
+ "grad_norm": 0.014595299027860165,
+ "learning_rate": 0.0006,
+ "loss": 4.204291343688965,
+ "step": 5885
+ },
+ {
+ "epoch": 81.75458515283843,
+ "grad_norm": 0.014952429570257664,
+ "learning_rate": 0.0006,
+ "loss": 4.242499351501465,
+ "step": 5886
+ },
+ {
+ "epoch": 81.76855895196506,
+ "grad_norm": 0.014761554077267647,
+ "learning_rate": 0.0006,
+ "loss": 4.20393180847168,
+ "step": 5887
+ },
+ {
+ "epoch": 81.78253275109171,
+ "grad_norm": 0.014749647118151188,
+ "learning_rate": 0.0006,
+ "loss": 4.227565765380859,
+ "step": 5888
+ },
+ {
+ "epoch": 81.79650655021834,
+ "grad_norm": 0.014384493231773376,
+ "learning_rate": 0.0006,
+ "loss": 4.282228946685791,
+ "step": 5889
+ },
+ {
+ "epoch": 81.81048034934499,
+ "grad_norm": 0.01260330155491829,
+ "learning_rate": 0.0006,
+ "loss": 4.2882866859436035,
+ "step": 5890
+ },
+ {
+ "epoch": 81.82445414847162,
+ "grad_norm": 0.014055311679840088,
+ "learning_rate": 0.0006,
+ "loss": 4.22351598739624,
+ "step": 5891
+ },
+ {
+ "epoch": 81.83842794759825,
+ "grad_norm": 0.013674065470695496,
+ "learning_rate": 0.0006,
+ "loss": 4.240716934204102,
+ "step": 5892
+ },
+ {
+ "epoch": 81.8524017467249,
+ "grad_norm": 0.01225196197628975,
+ "learning_rate": 0.0006,
+ "loss": 4.2958598136901855,
+ "step": 5893
+ },
+ {
+ "epoch": 81.86637554585153,
+ "grad_norm": 0.013538680970668793,
+ "learning_rate": 0.0006,
+ "loss": 4.24342679977417,
+ "step": 5894
+ },
+ {
+ "epoch": 81.88034934497817,
+ "grad_norm": 0.013151394203305244,
+ "learning_rate": 0.0006,
+ "loss": 4.081999778747559,
+ "step": 5895
+ },
+ {
+ "epoch": 81.8943231441048,
+ "grad_norm": 0.014260678552091122,
+ "learning_rate": 0.0006,
+ "loss": 4.156423568725586,
+ "step": 5896
+ },
+ {
+ "epoch": 81.90829694323143,
+ "grad_norm": 0.01329776830971241,
+ "learning_rate": 0.0006,
+ "loss": 4.1713995933532715,
+ "step": 5897
+ },
+ {
+ "epoch": 81.92227074235808,
+ "grad_norm": 0.014623276889324188,
+ "learning_rate": 0.0006,
+ "loss": 4.244891166687012,
+ "step": 5898
+ },
+ {
+ "epoch": 81.93624454148471,
+ "grad_norm": 0.015760324895381927,
+ "learning_rate": 0.0006,
+ "loss": 4.076528549194336,
+ "step": 5899
+ },
+ {
+ "epoch": 81.95021834061136,
+ "grad_norm": 0.014331766404211521,
+ "learning_rate": 0.0006,
+ "loss": 4.283037185668945,
+ "step": 5900
+ },
+ {
+ "epoch": 81.96419213973799,
+ "grad_norm": 0.012383842840790749,
+ "learning_rate": 0.0006,
+ "loss": 4.117308616638184,
+ "step": 5901
+ },
+ {
+ "epoch": 81.97816593886463,
+ "grad_norm": 0.01250660140067339,
+ "learning_rate": 0.0006,
+ "loss": 4.241250991821289,
+ "step": 5902
+ },
+ {
+ "epoch": 81.99213973799127,
+ "grad_norm": 0.012276900932192802,
+ "learning_rate": 0.0006,
+ "loss": 4.141091823577881,
+ "step": 5903
+ },
+ {
+ "epoch": 82.0,
+ "grad_norm": 0.01588737778365612,
+ "learning_rate": 0.0006,
+ "loss": 4.065532684326172,
+ "step": 5904
+ },
+ {
+ "epoch": 82.0,
+ "eval_loss": 4.612304210662842,
+ "eval_runtime": 57.5411,
+ "eval_samples_per_second": 42.439,
+ "eval_steps_per_second": 1.338,
+ "step": 5904
+ },
+ {
+ "epoch": 82.01397379912663,
+ "grad_norm": 0.017433544620871544,
+ "learning_rate": 0.0006,
+ "loss": 4.202497482299805,
+ "step": 5905
+ },
+ {
+ "epoch": 82.02794759825328,
+ "grad_norm": 0.017260415479540825,
+ "learning_rate": 0.0006,
+ "loss": 4.099949359893799,
+ "step": 5906
+ },
+ {
+ "epoch": 82.04192139737991,
+ "grad_norm": 0.01577908545732498,
+ "learning_rate": 0.0006,
+ "loss": 4.17806339263916,
+ "step": 5907
+ },
+ {
+ "epoch": 82.05589519650655,
+ "grad_norm": 0.016093818470835686,
+ "learning_rate": 0.0006,
+ "loss": 4.219491004943848,
+ "step": 5908
+ },
+ {
+ "epoch": 82.06986899563319,
+ "grad_norm": 0.015426302328705788,
+ "learning_rate": 0.0006,
+ "loss": 4.207982063293457,
+ "step": 5909
+ },
+ {
+ "epoch": 82.08384279475983,
+ "grad_norm": 0.01605442352592945,
+ "learning_rate": 0.0006,
+ "loss": 4.265942573547363,
+ "step": 5910
+ },
+ {
+ "epoch": 82.09781659388646,
+ "grad_norm": 0.01787225715816021,
+ "learning_rate": 0.0006,
+ "loss": 4.151535511016846,
+ "step": 5911
+ },
+ {
+ "epoch": 82.1117903930131,
+ "grad_norm": 0.01750164106488228,
+ "learning_rate": 0.0006,
+ "loss": 4.133655071258545,
+ "step": 5912
+ },
+ {
+ "epoch": 82.12576419213974,
+ "grad_norm": 0.01602957211434841,
+ "learning_rate": 0.0006,
+ "loss": 4.176972389221191,
+ "step": 5913
+ },
+ {
+ "epoch": 82.13973799126637,
+ "grad_norm": 0.01553812064230442,
+ "learning_rate": 0.0006,
+ "loss": 4.125692367553711,
+ "step": 5914
+ },
+ {
+ "epoch": 82.15371179039302,
+ "grad_norm": 0.013978508301079273,
+ "learning_rate": 0.0006,
+ "loss": 4.207271575927734,
+ "step": 5915
+ },
+ {
+ "epoch": 82.16768558951965,
+ "grad_norm": 0.013561118394136429,
+ "learning_rate": 0.0006,
+ "loss": 4.238930702209473,
+ "step": 5916
+ },
+ {
+ "epoch": 82.18165938864628,
+ "grad_norm": 0.013285507448017597,
+ "learning_rate": 0.0006,
+ "loss": 4.129464626312256,
+ "step": 5917
+ },
+ {
+ "epoch": 82.19563318777293,
+ "grad_norm": 0.013903839513659477,
+ "learning_rate": 0.0006,
+ "loss": 4.214109420776367,
+ "step": 5918
+ },
+ {
+ "epoch": 82.20960698689956,
+ "grad_norm": 0.014426837675273418,
+ "learning_rate": 0.0006,
+ "loss": 4.143196105957031,
+ "step": 5919
+ },
+ {
+ "epoch": 82.2235807860262,
+ "grad_norm": 0.015347165986895561,
+ "learning_rate": 0.0006,
+ "loss": 4.235439300537109,
+ "step": 5920
+ },
+ {
+ "epoch": 82.23755458515284,
+ "grad_norm": 0.014266039244830608,
+ "learning_rate": 0.0006,
+ "loss": 3.991617202758789,
+ "step": 5921
+ },
+ {
+ "epoch": 82.25152838427948,
+ "grad_norm": 0.013101593591272831,
+ "learning_rate": 0.0006,
+ "loss": 4.266017913818359,
+ "step": 5922
+ },
+ {
+ "epoch": 82.26550218340611,
+ "grad_norm": 0.013741881586611271,
+ "learning_rate": 0.0006,
+ "loss": 4.224879741668701,
+ "step": 5923
+ },
+ {
+ "epoch": 82.27947598253274,
+ "grad_norm": 0.014206945896148682,
+ "learning_rate": 0.0006,
+ "loss": 4.102272987365723,
+ "step": 5924
+ },
+ {
+ "epoch": 82.29344978165939,
+ "grad_norm": 0.013798723928630352,
+ "learning_rate": 0.0006,
+ "loss": 4.30418586730957,
+ "step": 5925
+ },
+ {
+ "epoch": 82.30742358078602,
+ "grad_norm": 0.015470122918486595,
+ "learning_rate": 0.0006,
+ "loss": 4.2513322830200195,
+ "step": 5926
+ },
+ {
+ "epoch": 82.32139737991267,
+ "grad_norm": 0.01655442826449871,
+ "learning_rate": 0.0006,
+ "loss": 4.223002910614014,
+ "step": 5927
+ },
+ {
+ "epoch": 82.3353711790393,
+ "grad_norm": 0.017170004546642303,
+ "learning_rate": 0.0006,
+ "loss": 4.253904819488525,
+ "step": 5928
+ },
+ {
+ "epoch": 82.34934497816595,
+ "grad_norm": 0.017595812678337097,
+ "learning_rate": 0.0006,
+ "loss": 4.184927940368652,
+ "step": 5929
+ },
+ {
+ "epoch": 82.36331877729258,
+ "grad_norm": 0.015073256567120552,
+ "learning_rate": 0.0006,
+ "loss": 4.202239036560059,
+ "step": 5930
+ },
+ {
+ "epoch": 82.37729257641921,
+ "grad_norm": 0.014310941100120544,
+ "learning_rate": 0.0006,
+ "loss": 4.124934673309326,
+ "step": 5931
+ },
+ {
+ "epoch": 82.39126637554585,
+ "grad_norm": 0.013671821914613247,
+ "learning_rate": 0.0006,
+ "loss": 4.138919353485107,
+ "step": 5932
+ },
+ {
+ "epoch": 82.40524017467249,
+ "grad_norm": 0.014103314839303493,
+ "learning_rate": 0.0006,
+ "loss": 4.211289405822754,
+ "step": 5933
+ },
+ {
+ "epoch": 82.41921397379913,
+ "grad_norm": 0.014272668398916721,
+ "learning_rate": 0.0006,
+ "loss": 4.06290340423584,
+ "step": 5934
+ },
+ {
+ "epoch": 82.43318777292576,
+ "grad_norm": 0.01387016475200653,
+ "learning_rate": 0.0006,
+ "loss": 4.101334571838379,
+ "step": 5935
+ },
+ {
+ "epoch": 82.44716157205241,
+ "grad_norm": 0.013318683952093124,
+ "learning_rate": 0.0006,
+ "loss": 4.206409931182861,
+ "step": 5936
+ },
+ {
+ "epoch": 82.46113537117904,
+ "grad_norm": 0.013781268149614334,
+ "learning_rate": 0.0006,
+ "loss": 4.189474105834961,
+ "step": 5937
+ },
+ {
+ "epoch": 82.47510917030567,
+ "grad_norm": 0.013195851817727089,
+ "learning_rate": 0.0006,
+ "loss": 4.276607036590576,
+ "step": 5938
+ },
+ {
+ "epoch": 82.48908296943232,
+ "grad_norm": 0.014358163811266422,
+ "learning_rate": 0.0006,
+ "loss": 4.203262805938721,
+ "step": 5939
+ },
+ {
+ "epoch": 82.50305676855895,
+ "grad_norm": 0.01403636671602726,
+ "learning_rate": 0.0006,
+ "loss": 4.079258441925049,
+ "step": 5940
+ },
+ {
+ "epoch": 82.5170305676856,
+ "grad_norm": 0.013272535055875778,
+ "learning_rate": 0.0006,
+ "loss": 4.19197940826416,
+ "step": 5941
+ },
+ {
+ "epoch": 82.53100436681223,
+ "grad_norm": 0.013787264935672283,
+ "learning_rate": 0.0006,
+ "loss": 4.12995719909668,
+ "step": 5942
+ },
+ {
+ "epoch": 82.54497816593886,
+ "grad_norm": 0.015020057559013367,
+ "learning_rate": 0.0006,
+ "loss": 4.148890495300293,
+ "step": 5943
+ },
+ {
+ "epoch": 82.5589519650655,
+ "grad_norm": 0.014199157245457172,
+ "learning_rate": 0.0006,
+ "loss": 4.201516151428223,
+ "step": 5944
+ },
+ {
+ "epoch": 82.57292576419214,
+ "grad_norm": 0.013827123679220676,
+ "learning_rate": 0.0006,
+ "loss": 4.240894317626953,
+ "step": 5945
+ },
+ {
+ "epoch": 82.58689956331878,
+ "grad_norm": 0.014767560176551342,
+ "learning_rate": 0.0006,
+ "loss": 4.173320770263672,
+ "step": 5946
+ },
+ {
+ "epoch": 82.60087336244541,
+ "grad_norm": 0.01612078957259655,
+ "learning_rate": 0.0006,
+ "loss": 4.266448020935059,
+ "step": 5947
+ },
+ {
+ "epoch": 82.61484716157206,
+ "grad_norm": 0.016706543043255806,
+ "learning_rate": 0.0006,
+ "loss": 4.229398250579834,
+ "step": 5948
+ },
+ {
+ "epoch": 82.62882096069869,
+ "grad_norm": 0.015630796551704407,
+ "learning_rate": 0.0006,
+ "loss": 4.181282997131348,
+ "step": 5949
+ },
+ {
+ "epoch": 82.64279475982532,
+ "grad_norm": 0.013103844597935677,
+ "learning_rate": 0.0006,
+ "loss": 4.309617042541504,
+ "step": 5950
+ },
+ {
+ "epoch": 82.65676855895197,
+ "grad_norm": 0.016136541962623596,
+ "learning_rate": 0.0006,
+ "loss": 4.224159240722656,
+ "step": 5951
+ },
+ {
+ "epoch": 82.6707423580786,
+ "grad_norm": 0.016160814091563225,
+ "learning_rate": 0.0006,
+ "loss": 4.267970085144043,
+ "step": 5952
+ },
+ {
+ "epoch": 82.68471615720524,
+ "grad_norm": 0.01672680303454399,
+ "learning_rate": 0.0006,
+ "loss": 4.230506896972656,
+ "step": 5953
+ },
+ {
+ "epoch": 82.69868995633188,
+ "grad_norm": 0.0159586351364851,
+ "learning_rate": 0.0006,
+ "loss": 4.159542083740234,
+ "step": 5954
+ },
+ {
+ "epoch": 82.71266375545852,
+ "grad_norm": 0.01461219135671854,
+ "learning_rate": 0.0006,
+ "loss": 4.27863073348999,
+ "step": 5955
+ },
+ {
+ "epoch": 82.72663755458515,
+ "grad_norm": 0.015088613145053387,
+ "learning_rate": 0.0006,
+ "loss": 4.272418975830078,
+ "step": 5956
+ },
+ {
+ "epoch": 82.74061135371178,
+ "grad_norm": 0.01594499684870243,
+ "learning_rate": 0.0006,
+ "loss": 4.271326065063477,
+ "step": 5957
+ },
+ {
+ "epoch": 82.75458515283843,
+ "grad_norm": 0.016648447141051292,
+ "learning_rate": 0.0006,
+ "loss": 4.141478538513184,
+ "step": 5958
+ },
+ {
+ "epoch": 82.76855895196506,
+ "grad_norm": 0.01540316641330719,
+ "learning_rate": 0.0006,
+ "loss": 4.111814498901367,
+ "step": 5959
+ },
+ {
+ "epoch": 82.78253275109171,
+ "grad_norm": 0.01634364202618599,
+ "learning_rate": 0.0006,
+ "loss": 4.3113112449646,
+ "step": 5960
+ },
+ {
+ "epoch": 82.79650655021834,
+ "grad_norm": 0.014768977649509907,
+ "learning_rate": 0.0006,
+ "loss": 4.184217929840088,
+ "step": 5961
+ },
+ {
+ "epoch": 82.81048034934499,
+ "grad_norm": 0.013453755527734756,
+ "learning_rate": 0.0006,
+ "loss": 4.247729301452637,
+ "step": 5962
+ },
+ {
+ "epoch": 82.82445414847162,
+ "grad_norm": 0.014731907285749912,
+ "learning_rate": 0.0006,
+ "loss": 4.352839469909668,
+ "step": 5963
+ },
+ {
+ "epoch": 82.83842794759825,
+ "grad_norm": 0.016047345474362373,
+ "learning_rate": 0.0006,
+ "loss": 4.257283687591553,
+ "step": 5964
+ },
+ {
+ "epoch": 82.8524017467249,
+ "grad_norm": 0.017141170799732208,
+ "learning_rate": 0.0006,
+ "loss": 4.22850227355957,
+ "step": 5965
+ },
+ {
+ "epoch": 82.86637554585153,
+ "grad_norm": 0.01370396837592125,
+ "learning_rate": 0.0006,
+ "loss": 4.1169257164001465,
+ "step": 5966
+ },
+ {
+ "epoch": 82.88034934497817,
+ "grad_norm": 0.012060081586241722,
+ "learning_rate": 0.0006,
+ "loss": 4.195242404937744,
+ "step": 5967
+ },
+ {
+ "epoch": 82.8943231441048,
+ "grad_norm": 0.015528378076851368,
+ "learning_rate": 0.0006,
+ "loss": 4.294939994812012,
+ "step": 5968
+ },
+ {
+ "epoch": 82.90829694323143,
+ "grad_norm": 0.018151041120290756,
+ "learning_rate": 0.0006,
+ "loss": 4.137192726135254,
+ "step": 5969
+ },
+ {
+ "epoch": 82.92227074235808,
+ "grad_norm": 0.017895735800266266,
+ "learning_rate": 0.0006,
+ "loss": 4.174036026000977,
+ "step": 5970
+ },
+ {
+ "epoch": 82.93624454148471,
+ "grad_norm": 0.015610386617481709,
+ "learning_rate": 0.0006,
+ "loss": 4.230136871337891,
+ "step": 5971
+ },
+ {
+ "epoch": 82.95021834061136,
+ "grad_norm": 0.01339928898960352,
+ "learning_rate": 0.0006,
+ "loss": 4.199074745178223,
+ "step": 5972
+ },
+ {
+ "epoch": 82.96419213973799,
+ "grad_norm": 0.015454214997589588,
+ "learning_rate": 0.0006,
+ "loss": 4.216109752655029,
+ "step": 5973
+ },
+ {
+ "epoch": 82.97816593886463,
+ "grad_norm": 0.014978132210671902,
+ "learning_rate": 0.0006,
+ "loss": 4.131511688232422,
+ "step": 5974
+ },
+ {
+ "epoch": 82.99213973799127,
+ "grad_norm": 0.015611122362315655,
+ "learning_rate": 0.0006,
+ "loss": 4.246518135070801,
+ "step": 5975
+ },
+ {
+ "epoch": 83.0,
+ "grad_norm": 0.016682764515280724,
+ "learning_rate": 0.0006,
+ "loss": 4.110680103302002,
+ "step": 5976
+ },
+ {
+ "epoch": 83.0,
+ "eval_loss": 4.641707897186279,
+ "eval_runtime": 57.2799,
+ "eval_samples_per_second": 42.633,
+ "eval_steps_per_second": 1.344,
+ "step": 5976
+ },
+ {
+ "epoch": 83.01397379912663,
+ "grad_norm": 0.015007982030510902,
+ "learning_rate": 0.0006,
+ "loss": 4.100912094116211,
+ "step": 5977
+ },
+ {
+ "epoch": 83.02794759825328,
+ "grad_norm": 0.017609436064958572,
+ "learning_rate": 0.0006,
+ "loss": 4.233399868011475,
+ "step": 5978
+ },
+ {
+ "epoch": 83.04192139737991,
+ "grad_norm": 0.019817272201180458,
+ "learning_rate": 0.0006,
+ "loss": 4.191173076629639,
+ "step": 5979
+ },
+ {
+ "epoch": 83.05589519650655,
+ "grad_norm": 0.017752140760421753,
+ "learning_rate": 0.0006,
+ "loss": 4.044437885284424,
+ "step": 5980
+ },
+ {
+ "epoch": 83.06986899563319,
+ "grad_norm": 0.015482760034501553,
+ "learning_rate": 0.0006,
+ "loss": 4.319931983947754,
+ "step": 5981
+ },
+ {
+ "epoch": 83.08384279475983,
+ "grad_norm": 0.015153988264501095,
+ "learning_rate": 0.0006,
+ "loss": 4.095277309417725,
+ "step": 5982
+ },
+ {
+ "epoch": 83.09781659388646,
+ "grad_norm": 0.017025282606482506,
+ "learning_rate": 0.0006,
+ "loss": 4.235110759735107,
+ "step": 5983
+ },
+ {
+ "epoch": 83.1117903930131,
+ "grad_norm": 0.01566586270928383,
+ "learning_rate": 0.0006,
+ "loss": 4.1934685707092285,
+ "step": 5984
+ },
+ {
+ "epoch": 83.12576419213974,
+ "grad_norm": 0.016419263556599617,
+ "learning_rate": 0.0006,
+ "loss": 4.234419822692871,
+ "step": 5985
+ },
+ {
+ "epoch": 83.13973799126637,
+ "grad_norm": 0.017132675275206566,
+ "learning_rate": 0.0006,
+ "loss": 4.221447944641113,
+ "step": 5986
+ },
+ {
+ "epoch": 83.15371179039302,
+ "grad_norm": 0.016138629987835884,
+ "learning_rate": 0.0006,
+ "loss": 4.199830055236816,
+ "step": 5987
+ },
+ {
+ "epoch": 83.16768558951965,
+ "grad_norm": 0.016210174188017845,
+ "learning_rate": 0.0006,
+ "loss": 4.081717491149902,
+ "step": 5988
+ },
+ {
+ "epoch": 83.18165938864628,
+ "grad_norm": 0.016055308282375336,
+ "learning_rate": 0.0006,
+ "loss": 4.247655391693115,
+ "step": 5989
+ },
+ {
+ "epoch": 83.19563318777293,
+ "grad_norm": 0.016132952645421028,
+ "learning_rate": 0.0006,
+ "loss": 4.262535572052002,
+ "step": 5990
+ },
+ {
+ "epoch": 83.20960698689956,
+ "grad_norm": 0.01779935695230961,
+ "learning_rate": 0.0006,
+ "loss": 4.153120040893555,
+ "step": 5991
+ },
+ {
+ "epoch": 83.2235807860262,
+ "grad_norm": 0.018360646441578865,
+ "learning_rate": 0.0006,
+ "loss": 4.232644081115723,
+ "step": 5992
+ },
+ {
+ "epoch": 83.23755458515284,
+ "grad_norm": 0.019629675894975662,
+ "learning_rate": 0.0006,
+ "loss": 3.9758734703063965,
+ "step": 5993
+ },
+ {
+ "epoch": 83.25152838427948,
+ "grad_norm": 0.017819996923208237,
+ "learning_rate": 0.0006,
+ "loss": 4.196932315826416,
+ "step": 5994
+ },
+ {
+ "epoch": 83.26550218340611,
+ "grad_norm": 0.01557681243866682,
+ "learning_rate": 0.0006,
+ "loss": 4.122654914855957,
+ "step": 5995
+ },
+ {
+ "epoch": 83.27947598253274,
+ "grad_norm": 0.017858853563666344,
+ "learning_rate": 0.0006,
+ "loss": 4.185330867767334,
+ "step": 5996
+ },
+ {
+ "epoch": 83.29344978165939,
+ "grad_norm": 0.018065953627228737,
+ "learning_rate": 0.0006,
+ "loss": 4.218471050262451,
+ "step": 5997
+ },
+ {
+ "epoch": 83.30742358078602,
+ "grad_norm": 0.017247939482331276,
+ "learning_rate": 0.0006,
+ "loss": 4.127076148986816,
+ "step": 5998
+ },
+ {
+ "epoch": 83.32139737991267,
+ "grad_norm": 0.017047952860593796,
+ "learning_rate": 0.0006,
+ "loss": 4.131356716156006,
+ "step": 5999
+ },
+ {
+ "epoch": 83.3353711790393,
+ "grad_norm": 0.015905629843473434,
+ "learning_rate": 0.0006,
+ "loss": 4.2239789962768555,
+ "step": 6000
+ },
+ {
+ "epoch": 83.34934497816595,
+ "grad_norm": 0.016522347927093506,
+ "learning_rate": 0.0006,
+ "loss": 4.054897785186768,
+ "step": 6001
+ },
+ {
+ "epoch": 83.36331877729258,
+ "grad_norm": 0.01622571051120758,
+ "learning_rate": 0.0006,
+ "loss": 4.284210681915283,
+ "step": 6002
+ },
+ {
+ "epoch": 83.37729257641921,
+ "grad_norm": 0.01640467159450054,
+ "learning_rate": 0.0006,
+ "loss": 4.268840789794922,
+ "step": 6003
+ },
+ {
+ "epoch": 83.39126637554585,
+ "grad_norm": 0.015698859468102455,
+ "learning_rate": 0.0006,
+ "loss": 4.179793357849121,
+ "step": 6004
+ },
+ {
+ "epoch": 83.40524017467249,
+ "grad_norm": 0.016694528982043266,
+ "learning_rate": 0.0006,
+ "loss": 4.132297515869141,
+ "step": 6005
+ },
+ {
+ "epoch": 83.41921397379913,
+ "grad_norm": 0.015345900319516659,
+ "learning_rate": 0.0006,
+ "loss": 4.090163230895996,
+ "step": 6006
+ },
+ {
+ "epoch": 83.43318777292576,
+ "grad_norm": 0.014722758904099464,
+ "learning_rate": 0.0006,
+ "loss": 4.179166793823242,
+ "step": 6007
+ },
+ {
+ "epoch": 83.44716157205241,
+ "grad_norm": 0.014884887263178825,
+ "learning_rate": 0.0006,
+ "loss": 4.137659072875977,
+ "step": 6008
+ },
+ {
+ "epoch": 83.46113537117904,
+ "grad_norm": 0.014041902497410774,
+ "learning_rate": 0.0006,
+ "loss": 4.1202073097229,
+ "step": 6009
+ },
+ {
+ "epoch": 83.47510917030567,
+ "grad_norm": 0.012799333781003952,
+ "learning_rate": 0.0006,
+ "loss": 4.092607498168945,
+ "step": 6010
+ },
+ {
+ "epoch": 83.48908296943232,
+ "grad_norm": 0.013272707350552082,
+ "learning_rate": 0.0006,
+ "loss": 4.132715225219727,
+ "step": 6011
+ },
+ {
+ "epoch": 83.50305676855895,
+ "grad_norm": 0.012934371829032898,
+ "learning_rate": 0.0006,
+ "loss": 4.230278968811035,
+ "step": 6012
+ },
+ {
+ "epoch": 83.5170305676856,
+ "grad_norm": 0.0120006762444973,
+ "learning_rate": 0.0006,
+ "loss": 4.140068531036377,
+ "step": 6013
+ },
+ {
+ "epoch": 83.53100436681223,
+ "grad_norm": 0.013616700656712055,
+ "learning_rate": 0.0006,
+ "loss": 4.20107364654541,
+ "step": 6014
+ },
+ {
+ "epoch": 83.54497816593886,
+ "grad_norm": 0.013957851566374302,
+ "learning_rate": 0.0006,
+ "loss": 4.124945163726807,
+ "step": 6015
+ },
+ {
+ "epoch": 83.5589519650655,
+ "grad_norm": 0.012762055732309818,
+ "learning_rate": 0.0006,
+ "loss": 4.022052764892578,
+ "step": 6016
+ },
+ {
+ "epoch": 83.57292576419214,
+ "grad_norm": 0.013080576434731483,
+ "learning_rate": 0.0006,
+ "loss": 4.103252410888672,
+ "step": 6017
+ },
+ {
+ "epoch": 83.58689956331878,
+ "grad_norm": 0.013957500457763672,
+ "learning_rate": 0.0006,
+ "loss": 4.235089302062988,
+ "step": 6018
+ },
+ {
+ "epoch": 83.60087336244541,
+ "grad_norm": 0.014707164838910103,
+ "learning_rate": 0.0006,
+ "loss": 4.191570281982422,
+ "step": 6019
+ },
+ {
+ "epoch": 83.61484716157206,
+ "grad_norm": 0.012807869352400303,
+ "learning_rate": 0.0006,
+ "loss": 4.294543266296387,
+ "step": 6020
+ },
+ {
+ "epoch": 83.62882096069869,
+ "grad_norm": 0.012027994729578495,
+ "learning_rate": 0.0006,
+ "loss": 4.13849401473999,
+ "step": 6021
+ },
+ {
+ "epoch": 83.64279475982532,
+ "grad_norm": 0.012147724628448486,
+ "learning_rate": 0.0006,
+ "loss": 4.127481937408447,
+ "step": 6022
+ },
+ {
+ "epoch": 83.65676855895197,
+ "grad_norm": 0.013241392560303211,
+ "learning_rate": 0.0006,
+ "loss": 4.218972682952881,
+ "step": 6023
+ },
+ {
+ "epoch": 83.6707423580786,
+ "grad_norm": 0.014071648940443993,
+ "learning_rate": 0.0006,
+ "loss": 4.096724033355713,
+ "step": 6024
+ },
+ {
+ "epoch": 83.68471615720524,
+ "grad_norm": 0.015212797559797764,
+ "learning_rate": 0.0006,
+ "loss": 4.27097225189209,
+ "step": 6025
+ },
+ {
+ "epoch": 83.69868995633188,
+ "grad_norm": 0.017647385597229004,
+ "learning_rate": 0.0006,
+ "loss": 4.209835529327393,
+ "step": 6026
+ },
+ {
+ "epoch": 83.71266375545852,
+ "grad_norm": 0.01709647662937641,
+ "learning_rate": 0.0006,
+ "loss": 4.042106628417969,
+ "step": 6027
+ },
+ {
+ "epoch": 83.72663755458515,
+ "grad_norm": 0.014530927874147892,
+ "learning_rate": 0.0006,
+ "loss": 4.181731700897217,
+ "step": 6028
+ },
+ {
+ "epoch": 83.74061135371178,
+ "grad_norm": 0.015356711111962795,
+ "learning_rate": 0.0006,
+ "loss": 4.314183235168457,
+ "step": 6029
+ },
+ {
+ "epoch": 83.75458515283843,
+ "grad_norm": 0.015500029549002647,
+ "learning_rate": 0.0006,
+ "loss": 4.06776762008667,
+ "step": 6030
+ },
+ {
+ "epoch": 83.76855895196506,
+ "grad_norm": 0.015351332724094391,
+ "learning_rate": 0.0006,
+ "loss": 4.185141563415527,
+ "step": 6031
+ },
+ {
+ "epoch": 83.78253275109171,
+ "grad_norm": 0.016678154468536377,
+ "learning_rate": 0.0006,
+ "loss": 4.179529666900635,
+ "step": 6032
+ },
+ {
+ "epoch": 83.79650655021834,
+ "grad_norm": 0.0156265702098608,
+ "learning_rate": 0.0006,
+ "loss": 4.238584041595459,
+ "step": 6033
+ },
+ {
+ "epoch": 83.81048034934499,
+ "grad_norm": 0.016418596729636192,
+ "learning_rate": 0.0006,
+ "loss": 4.1593523025512695,
+ "step": 6034
+ },
+ {
+ "epoch": 83.82445414847162,
+ "grad_norm": 0.016113586723804474,
+ "learning_rate": 0.0006,
+ "loss": 4.157470226287842,
+ "step": 6035
+ },
+ {
+ "epoch": 83.83842794759825,
+ "grad_norm": 0.014434042386710644,
+ "learning_rate": 0.0006,
+ "loss": 4.196226119995117,
+ "step": 6036
+ },
+ {
+ "epoch": 83.8524017467249,
+ "grad_norm": 0.012647042982280254,
+ "learning_rate": 0.0006,
+ "loss": 4.130589485168457,
+ "step": 6037
+ },
+ {
+ "epoch": 83.86637554585153,
+ "grad_norm": 0.014412877149879932,
+ "learning_rate": 0.0006,
+ "loss": 4.154845237731934,
+ "step": 6038
+ },
+ {
+ "epoch": 83.88034934497817,
+ "grad_norm": 0.014222027733922005,
+ "learning_rate": 0.0006,
+ "loss": 4.189724922180176,
+ "step": 6039
+ },
+ {
+ "epoch": 83.8943231441048,
+ "grad_norm": 0.013084255158901215,
+ "learning_rate": 0.0006,
+ "loss": 4.303338527679443,
+ "step": 6040
+ },
+ {
+ "epoch": 83.90829694323143,
+ "grad_norm": 0.012659232132136822,
+ "learning_rate": 0.0006,
+ "loss": 4.08192253112793,
+ "step": 6041
+ },
+ {
+ "epoch": 83.92227074235808,
+ "grad_norm": 0.01246276218444109,
+ "learning_rate": 0.0006,
+ "loss": 4.283324718475342,
+ "step": 6042
+ },
+ {
+ "epoch": 83.93624454148471,
+ "grad_norm": 0.01340118795633316,
+ "learning_rate": 0.0006,
+ "loss": 4.117908477783203,
+ "step": 6043
+ },
+ {
+ "epoch": 83.95021834061136,
+ "grad_norm": 0.014496722258627415,
+ "learning_rate": 0.0006,
+ "loss": 4.110318660736084,
+ "step": 6044
+ },
+ {
+ "epoch": 83.96419213973799,
+ "grad_norm": 0.013348298147320747,
+ "learning_rate": 0.0006,
+ "loss": 4.185139179229736,
+ "step": 6045
+ },
+ {
+ "epoch": 83.97816593886463,
+ "grad_norm": 0.013294277712702751,
+ "learning_rate": 0.0006,
+ "loss": 4.271645545959473,
+ "step": 6046
+ },
+ {
+ "epoch": 83.99213973799127,
+ "grad_norm": 0.014431594870984554,
+ "learning_rate": 0.0006,
+ "loss": 4.161383152008057,
+ "step": 6047
+ },
+ {
+ "epoch": 84.0,
+ "grad_norm": 0.016268927603960037,
+ "learning_rate": 0.0006,
+ "loss": 4.178324222564697,
+ "step": 6048
+ },
+ {
+ "epoch": 84.0,
+ "eval_loss": 4.6307549476623535,
+ "eval_runtime": 56.8521,
+ "eval_samples_per_second": 42.954,
+ "eval_steps_per_second": 1.354,
+ "step": 6048
+ },
+ {
+ "epoch": 84.01397379912663,
+ "grad_norm": 0.014864908531308174,
+ "learning_rate": 0.0006,
+ "loss": 4.121300220489502,
+ "step": 6049
+ },
+ {
+ "epoch": 84.02794759825328,
+ "grad_norm": 0.014863367192447186,
+ "learning_rate": 0.0006,
+ "loss": 4.1650543212890625,
+ "step": 6050
+ },
+ {
+ "epoch": 84.04192139737991,
+ "grad_norm": 0.014379646629095078,
+ "learning_rate": 0.0006,
+ "loss": 4.0201239585876465,
+ "step": 6051
+ },
+ {
+ "epoch": 84.05589519650655,
+ "grad_norm": 0.014871201477944851,
+ "learning_rate": 0.0006,
+ "loss": 4.0820698738098145,
+ "step": 6052
+ },
+ {
+ "epoch": 84.06986899563319,
+ "grad_norm": 0.013085179962217808,
+ "learning_rate": 0.0006,
+ "loss": 4.223984718322754,
+ "step": 6053
+ },
+ {
+ "epoch": 84.08384279475983,
+ "grad_norm": 0.012485949322581291,
+ "learning_rate": 0.0006,
+ "loss": 4.182206153869629,
+ "step": 6054
+ },
+ {
+ "epoch": 84.09781659388646,
+ "grad_norm": 0.013653291389346123,
+ "learning_rate": 0.0006,
+ "loss": 4.142557144165039,
+ "step": 6055
+ },
+ {
+ "epoch": 84.1117903930131,
+ "grad_norm": 0.013930532149970531,
+ "learning_rate": 0.0006,
+ "loss": 4.13972282409668,
+ "step": 6056
+ },
+ {
+ "epoch": 84.12576419213974,
+ "grad_norm": 0.01297379657626152,
+ "learning_rate": 0.0006,
+ "loss": 4.144664764404297,
+ "step": 6057
+ },
+ {
+ "epoch": 84.13973799126637,
+ "grad_norm": 0.013210492208600044,
+ "learning_rate": 0.0006,
+ "loss": 4.253764629364014,
+ "step": 6058
+ },
+ {
+ "epoch": 84.15371179039302,
+ "grad_norm": 0.014650176279246807,
+ "learning_rate": 0.0006,
+ "loss": 4.246866226196289,
+ "step": 6059
+ },
+ {
+ "epoch": 84.16768558951965,
+ "grad_norm": 0.01482292078435421,
+ "learning_rate": 0.0006,
+ "loss": 4.140187740325928,
+ "step": 6060
+ },
+ {
+ "epoch": 84.18165938864628,
+ "grad_norm": 0.01592552848160267,
+ "learning_rate": 0.0006,
+ "loss": 4.170705318450928,
+ "step": 6061
+ },
+ {
+ "epoch": 84.19563318777293,
+ "grad_norm": 0.017152156680822372,
+ "learning_rate": 0.0006,
+ "loss": 4.126092433929443,
+ "step": 6062
+ },
+ {
+ "epoch": 84.20960698689956,
+ "grad_norm": 0.015880905091762543,
+ "learning_rate": 0.0006,
+ "loss": 4.161938190460205,
+ "step": 6063
+ },
+ {
+ "epoch": 84.2235807860262,
+ "grad_norm": 0.015451471321284771,
+ "learning_rate": 0.0006,
+ "loss": 4.263461112976074,
+ "step": 6064
+ },
+ {
+ "epoch": 84.23755458515284,
+ "grad_norm": 0.014365996234118938,
+ "learning_rate": 0.0006,
+ "loss": 4.14488410949707,
+ "step": 6065
+ },
+ {
+ "epoch": 84.25152838427948,
+ "grad_norm": 0.016364749521017075,
+ "learning_rate": 0.0006,
+ "loss": 4.112489700317383,
+ "step": 6066
+ },
+ {
+ "epoch": 84.26550218340611,
+ "grad_norm": 0.01791754551231861,
+ "learning_rate": 0.0006,
+ "loss": 4.133513450622559,
+ "step": 6067
+ },
+ {
+ "epoch": 84.27947598253274,
+ "grad_norm": 0.016609208658337593,
+ "learning_rate": 0.0006,
+ "loss": 4.201829433441162,
+ "step": 6068
+ },
+ {
+ "epoch": 84.29344978165939,
+ "grad_norm": 0.017587820068001747,
+ "learning_rate": 0.0006,
+ "loss": 4.203151702880859,
+ "step": 6069
+ },
+ {
+ "epoch": 84.30742358078602,
+ "grad_norm": 0.017047742381691933,
+ "learning_rate": 0.0006,
+ "loss": 4.077008247375488,
+ "step": 6070
+ },
+ {
+ "epoch": 84.32139737991267,
+ "grad_norm": 0.01524646207690239,
+ "learning_rate": 0.0006,
+ "loss": 4.198038101196289,
+ "step": 6071
+ },
+ {
+ "epoch": 84.3353711790393,
+ "grad_norm": 0.014196421951055527,
+ "learning_rate": 0.0006,
+ "loss": 4.164245128631592,
+ "step": 6072
+ },
+ {
+ "epoch": 84.34934497816595,
+ "grad_norm": 0.016291776672005653,
+ "learning_rate": 0.0006,
+ "loss": 4.150152206420898,
+ "step": 6073
+ },
+ {
+ "epoch": 84.36331877729258,
+ "grad_norm": 0.015627920627593994,
+ "learning_rate": 0.0006,
+ "loss": 4.110006332397461,
+ "step": 6074
+ },
+ {
+ "epoch": 84.37729257641921,
+ "grad_norm": 0.013251487165689468,
+ "learning_rate": 0.0006,
+ "loss": 4.101463317871094,
+ "step": 6075
+ },
+ {
+ "epoch": 84.39126637554585,
+ "grad_norm": 0.013601605780422688,
+ "learning_rate": 0.0006,
+ "loss": 4.191105842590332,
+ "step": 6076
+ },
+ {
+ "epoch": 84.40524017467249,
+ "grad_norm": 0.014878741465508938,
+ "learning_rate": 0.0006,
+ "loss": 4.153046131134033,
+ "step": 6077
+ },
+ {
+ "epoch": 84.41921397379913,
+ "grad_norm": 0.014783897437155247,
+ "learning_rate": 0.0006,
+ "loss": 4.362000465393066,
+ "step": 6078
+ },
+ {
+ "epoch": 84.43318777292576,
+ "grad_norm": 0.014221332035958767,
+ "learning_rate": 0.0006,
+ "loss": 4.144364833831787,
+ "step": 6079
+ },
+ {
+ "epoch": 84.44716157205241,
+ "grad_norm": 0.014911034144461155,
+ "learning_rate": 0.0006,
+ "loss": 4.088857650756836,
+ "step": 6080
+ },
+ {
+ "epoch": 84.46113537117904,
+ "grad_norm": 0.014924770221114159,
+ "learning_rate": 0.0006,
+ "loss": 4.1677446365356445,
+ "step": 6081
+ },
+ {
+ "epoch": 84.47510917030567,
+ "grad_norm": 0.014394409023225307,
+ "learning_rate": 0.0006,
+ "loss": 4.087103843688965,
+ "step": 6082
+ },
+ {
+ "epoch": 84.48908296943232,
+ "grad_norm": 0.013926676474511623,
+ "learning_rate": 0.0006,
+ "loss": 3.9972851276397705,
+ "step": 6083
+ },
+ {
+ "epoch": 84.50305676855895,
+ "grad_norm": 0.013069954700767994,
+ "learning_rate": 0.0006,
+ "loss": 3.998237133026123,
+ "step": 6084
+ },
+ {
+ "epoch": 84.5170305676856,
+ "grad_norm": 0.012748011387884617,
+ "learning_rate": 0.0006,
+ "loss": 4.195858955383301,
+ "step": 6085
+ },
+ {
+ "epoch": 84.53100436681223,
+ "grad_norm": 0.012278077192604542,
+ "learning_rate": 0.0006,
+ "loss": 4.102165222167969,
+ "step": 6086
+ },
+ {
+ "epoch": 84.54497816593886,
+ "grad_norm": 0.014259163290262222,
+ "learning_rate": 0.0006,
+ "loss": 4.092750549316406,
+ "step": 6087
+ },
+ {
+ "epoch": 84.5589519650655,
+ "grad_norm": 0.014553683809936047,
+ "learning_rate": 0.0006,
+ "loss": 4.1412763595581055,
+ "step": 6088
+ },
+ {
+ "epoch": 84.57292576419214,
+ "grad_norm": 0.013605243526399136,
+ "learning_rate": 0.0006,
+ "loss": 4.192173480987549,
+ "step": 6089
+ },
+ {
+ "epoch": 84.58689956331878,
+ "grad_norm": 0.014259358868002892,
+ "learning_rate": 0.0006,
+ "loss": 4.356374740600586,
+ "step": 6090
+ },
+ {
+ "epoch": 84.60087336244541,
+ "grad_norm": 0.014166025444865227,
+ "learning_rate": 0.0006,
+ "loss": 4.165982246398926,
+ "step": 6091
+ },
+ {
+ "epoch": 84.61484716157206,
+ "grad_norm": 0.012198803946375847,
+ "learning_rate": 0.0006,
+ "loss": 4.287566184997559,
+ "step": 6092
+ },
+ {
+ "epoch": 84.62882096069869,
+ "grad_norm": 0.013794238679111004,
+ "learning_rate": 0.0006,
+ "loss": 4.199882984161377,
+ "step": 6093
+ },
+ {
+ "epoch": 84.64279475982532,
+ "grad_norm": 0.014757642522454262,
+ "learning_rate": 0.0006,
+ "loss": 4.116573333740234,
+ "step": 6094
+ },
+ {
+ "epoch": 84.65676855895197,
+ "grad_norm": 0.014591689221560955,
+ "learning_rate": 0.0006,
+ "loss": 4.240896224975586,
+ "step": 6095
+ },
+ {
+ "epoch": 84.6707423580786,
+ "grad_norm": 0.015217579901218414,
+ "learning_rate": 0.0006,
+ "loss": 4.257142066955566,
+ "step": 6096
+ },
+ {
+ "epoch": 84.68471615720524,
+ "grad_norm": 0.015371371991932392,
+ "learning_rate": 0.0006,
+ "loss": 4.181037425994873,
+ "step": 6097
+ },
+ {
+ "epoch": 84.69868995633188,
+ "grad_norm": 0.014412983320653439,
+ "learning_rate": 0.0006,
+ "loss": 4.2207794189453125,
+ "step": 6098
+ },
+ {
+ "epoch": 84.71266375545852,
+ "grad_norm": 0.013773671351373196,
+ "learning_rate": 0.0006,
+ "loss": 4.211123466491699,
+ "step": 6099
+ },
+ {
+ "epoch": 84.72663755458515,
+ "grad_norm": 0.012838046066462994,
+ "learning_rate": 0.0006,
+ "loss": 4.175459384918213,
+ "step": 6100
+ },
+ {
+ "epoch": 84.74061135371178,
+ "grad_norm": 0.014094929210841656,
+ "learning_rate": 0.0006,
+ "loss": 4.1724958419799805,
+ "step": 6101
+ },
+ {
+ "epoch": 84.75458515283843,
+ "grad_norm": 0.014573262073099613,
+ "learning_rate": 0.0006,
+ "loss": 4.15885066986084,
+ "step": 6102
+ },
+ {
+ "epoch": 84.76855895196506,
+ "grad_norm": 0.011964819394052029,
+ "learning_rate": 0.0006,
+ "loss": 3.9804892539978027,
+ "step": 6103
+ },
+ {
+ "epoch": 84.78253275109171,
+ "grad_norm": 0.014140539802610874,
+ "learning_rate": 0.0006,
+ "loss": 4.203708648681641,
+ "step": 6104
+ },
+ {
+ "epoch": 84.79650655021834,
+ "grad_norm": 0.016618940979242325,
+ "learning_rate": 0.0006,
+ "loss": 4.147684097290039,
+ "step": 6105
+ },
+ {
+ "epoch": 84.81048034934499,
+ "grad_norm": 0.015094385482370853,
+ "learning_rate": 0.0006,
+ "loss": 4.392223834991455,
+ "step": 6106
+ },
+ {
+ "epoch": 84.82445414847162,
+ "grad_norm": 0.014386717230081558,
+ "learning_rate": 0.0006,
+ "loss": 4.253332138061523,
+ "step": 6107
+ },
+ {
+ "epoch": 84.83842794759825,
+ "grad_norm": 0.013877755962312222,
+ "learning_rate": 0.0006,
+ "loss": 4.075056076049805,
+ "step": 6108
+ },
+ {
+ "epoch": 84.8524017467249,
+ "grad_norm": 0.014062155969440937,
+ "learning_rate": 0.0006,
+ "loss": 4.072048664093018,
+ "step": 6109
+ },
+ {
+ "epoch": 84.86637554585153,
+ "grad_norm": 0.014445619657635689,
+ "learning_rate": 0.0006,
+ "loss": 4.151832580566406,
+ "step": 6110
+ },
+ {
+ "epoch": 84.88034934497817,
+ "grad_norm": 0.015116363763809204,
+ "learning_rate": 0.0006,
+ "loss": 4.088900566101074,
+ "step": 6111
+ },
+ {
+ "epoch": 84.8943231441048,
+ "grad_norm": 0.014425474219024181,
+ "learning_rate": 0.0006,
+ "loss": 4.200483322143555,
+ "step": 6112
+ },
+ {
+ "epoch": 84.90829694323143,
+ "grad_norm": 0.01449573878198862,
+ "learning_rate": 0.0006,
+ "loss": 4.239742279052734,
+ "step": 6113
+ },
+ {
+ "epoch": 84.92227074235808,
+ "grad_norm": 0.015698261559009552,
+ "learning_rate": 0.0006,
+ "loss": 4.197391510009766,
+ "step": 6114
+ },
+ {
+ "epoch": 84.93624454148471,
+ "grad_norm": 0.014545733109116554,
+ "learning_rate": 0.0006,
+ "loss": 4.035508155822754,
+ "step": 6115
+ },
+ {
+ "epoch": 84.95021834061136,
+ "grad_norm": 0.012796483002603054,
+ "learning_rate": 0.0006,
+ "loss": 4.195248126983643,
+ "step": 6116
+ },
+ {
+ "epoch": 84.96419213973799,
+ "grad_norm": 0.01264485064893961,
+ "learning_rate": 0.0006,
+ "loss": 4.158118724822998,
+ "step": 6117
+ },
+ {
+ "epoch": 84.97816593886463,
+ "grad_norm": 0.013695276342332363,
+ "learning_rate": 0.0006,
+ "loss": 4.255288124084473,
+ "step": 6118
+ },
+ {
+ "epoch": 84.99213973799127,
+ "grad_norm": 0.013459564186632633,
+ "learning_rate": 0.0006,
+ "loss": 4.204586505889893,
+ "step": 6119
+ },
+ {
+ "epoch": 85.0,
+ "grad_norm": 0.014312652871012688,
+ "learning_rate": 0.0006,
+ "loss": 4.167879104614258,
+ "step": 6120
+ },
+ {
+ "epoch": 85.0,
+ "eval_loss": 4.608152389526367,
+ "eval_runtime": 57.1903,
+ "eval_samples_per_second": 42.7,
+ "eval_steps_per_second": 1.346,
+ "step": 6120
+ },
+ {
+ "epoch": 85.01397379912663,
+ "grad_norm": 0.013918368145823479,
+ "learning_rate": 0.0006,
+ "loss": 4.129991054534912,
+ "step": 6121
+ },
+ {
+ "epoch": 85.02794759825328,
+ "grad_norm": 0.01604384370148182,
+ "learning_rate": 0.0006,
+ "loss": 4.233926773071289,
+ "step": 6122
+ },
+ {
+ "epoch": 85.04192139737991,
+ "grad_norm": 0.016565652564167976,
+ "learning_rate": 0.0006,
+ "loss": 4.129604339599609,
+ "step": 6123
+ },
+ {
+ "epoch": 85.05589519650655,
+ "grad_norm": 0.0149751678109169,
+ "learning_rate": 0.0006,
+ "loss": 4.165839195251465,
+ "step": 6124
+ },
+ {
+ "epoch": 85.06986899563319,
+ "grad_norm": 0.013812461867928505,
+ "learning_rate": 0.0006,
+ "loss": 4.153792858123779,
+ "step": 6125
+ },
+ {
+ "epoch": 85.08384279475983,
+ "grad_norm": 0.014432745985686779,
+ "learning_rate": 0.0006,
+ "loss": 4.170895576477051,
+ "step": 6126
+ },
+ {
+ "epoch": 85.09781659388646,
+ "grad_norm": 0.015208868309855461,
+ "learning_rate": 0.0006,
+ "loss": 4.10639762878418,
+ "step": 6127
+ },
+ {
+ "epoch": 85.1117903930131,
+ "grad_norm": 0.013209272176027298,
+ "learning_rate": 0.0006,
+ "loss": 4.200364112854004,
+ "step": 6128
+ },
+ {
+ "epoch": 85.12576419213974,
+ "grad_norm": 0.014807002618908882,
+ "learning_rate": 0.0006,
+ "loss": 4.172666549682617,
+ "step": 6129
+ },
+ {
+ "epoch": 85.13973799126637,
+ "grad_norm": 0.015662945806980133,
+ "learning_rate": 0.0006,
+ "loss": 4.145219326019287,
+ "step": 6130
+ },
+ {
+ "epoch": 85.15371179039302,
+ "grad_norm": 0.014598497189581394,
+ "learning_rate": 0.0006,
+ "loss": 4.275618553161621,
+ "step": 6131
+ },
+ {
+ "epoch": 85.16768558951965,
+ "grad_norm": 0.013778576627373695,
+ "learning_rate": 0.0006,
+ "loss": 4.204716682434082,
+ "step": 6132
+ },
+ {
+ "epoch": 85.18165938864628,
+ "grad_norm": 0.015762265771627426,
+ "learning_rate": 0.0006,
+ "loss": 4.145524024963379,
+ "step": 6133
+ },
+ {
+ "epoch": 85.19563318777293,
+ "grad_norm": 0.015579195693135262,
+ "learning_rate": 0.0006,
+ "loss": 4.150625228881836,
+ "step": 6134
+ },
+ {
+ "epoch": 85.20960698689956,
+ "grad_norm": 0.016641216352581978,
+ "learning_rate": 0.0006,
+ "loss": 4.141819953918457,
+ "step": 6135
+ },
+ {
+ "epoch": 85.2235807860262,
+ "grad_norm": 0.01632271707057953,
+ "learning_rate": 0.0006,
+ "loss": 4.030620574951172,
+ "step": 6136
+ },
+ {
+ "epoch": 85.23755458515284,
+ "grad_norm": 0.015564918518066406,
+ "learning_rate": 0.0006,
+ "loss": 4.147753715515137,
+ "step": 6137
+ },
+ {
+ "epoch": 85.25152838427948,
+ "grad_norm": 0.0161129217594862,
+ "learning_rate": 0.0006,
+ "loss": 4.128524303436279,
+ "step": 6138
+ },
+ {
+ "epoch": 85.26550218340611,
+ "grad_norm": 0.017766550183296204,
+ "learning_rate": 0.0006,
+ "loss": 4.134856224060059,
+ "step": 6139
+ },
+ {
+ "epoch": 85.27947598253274,
+ "grad_norm": 0.01869054324924946,
+ "learning_rate": 0.0006,
+ "loss": 4.231565475463867,
+ "step": 6140
+ },
+ {
+ "epoch": 85.29344978165939,
+ "grad_norm": 0.01917010359466076,
+ "learning_rate": 0.0006,
+ "loss": 4.21175479888916,
+ "step": 6141
+ },
+ {
+ "epoch": 85.30742358078602,
+ "grad_norm": 0.018116485327482224,
+ "learning_rate": 0.0006,
+ "loss": 4.122644424438477,
+ "step": 6142
+ },
+ {
+ "epoch": 85.32139737991267,
+ "grad_norm": 0.015949107706546783,
+ "learning_rate": 0.0006,
+ "loss": 4.135458946228027,
+ "step": 6143
+ },
+ {
+ "epoch": 85.3353711790393,
+ "grad_norm": 0.014783683232963085,
+ "learning_rate": 0.0006,
+ "loss": 4.118994235992432,
+ "step": 6144
+ },
+ {
+ "epoch": 85.34934497816595,
+ "grad_norm": 0.015406692400574684,
+ "learning_rate": 0.0006,
+ "loss": 4.157268524169922,
+ "step": 6145
+ },
+ {
+ "epoch": 85.36331877729258,
+ "grad_norm": 0.013893205672502518,
+ "learning_rate": 0.0006,
+ "loss": 4.056166648864746,
+ "step": 6146
+ },
+ {
+ "epoch": 85.37729257641921,
+ "grad_norm": 0.014380923472344875,
+ "learning_rate": 0.0006,
+ "loss": 4.206905364990234,
+ "step": 6147
+ },
+ {
+ "epoch": 85.39126637554585,
+ "grad_norm": 0.015669258311390877,
+ "learning_rate": 0.0006,
+ "loss": 4.213881492614746,
+ "step": 6148
+ },
+ {
+ "epoch": 85.40524017467249,
+ "grad_norm": 0.01553428452461958,
+ "learning_rate": 0.0006,
+ "loss": 4.1814374923706055,
+ "step": 6149
+ },
+ {
+ "epoch": 85.41921397379913,
+ "grad_norm": 0.01697700284421444,
+ "learning_rate": 0.0006,
+ "loss": 4.083939552307129,
+ "step": 6150
+ },
+ {
+ "epoch": 85.43318777292576,
+ "grad_norm": 0.016639070585370064,
+ "learning_rate": 0.0006,
+ "loss": 4.279311180114746,
+ "step": 6151
+ },
+ {
+ "epoch": 85.44716157205241,
+ "grad_norm": 0.016285281628370285,
+ "learning_rate": 0.0006,
+ "loss": 4.244845390319824,
+ "step": 6152
+ },
+ {
+ "epoch": 85.46113537117904,
+ "grad_norm": 0.016071317717432976,
+ "learning_rate": 0.0006,
+ "loss": 4.089814186096191,
+ "step": 6153
+ },
+ {
+ "epoch": 85.47510917030567,
+ "grad_norm": 0.01640418916940689,
+ "learning_rate": 0.0006,
+ "loss": 4.168088912963867,
+ "step": 6154
+ },
+ {
+ "epoch": 85.48908296943232,
+ "grad_norm": 0.014781414531171322,
+ "learning_rate": 0.0006,
+ "loss": 4.196092128753662,
+ "step": 6155
+ },
+ {
+ "epoch": 85.50305676855895,
+ "grad_norm": 0.013393213972449303,
+ "learning_rate": 0.0006,
+ "loss": 4.193429946899414,
+ "step": 6156
+ },
+ {
+ "epoch": 85.5170305676856,
+ "grad_norm": 0.01467406190931797,
+ "learning_rate": 0.0006,
+ "loss": 4.216714859008789,
+ "step": 6157
+ },
+ {
+ "epoch": 85.53100436681223,
+ "grad_norm": 0.014950765296816826,
+ "learning_rate": 0.0006,
+ "loss": 4.093854904174805,
+ "step": 6158
+ },
+ {
+ "epoch": 85.54497816593886,
+ "grad_norm": 0.014655301347374916,
+ "learning_rate": 0.0006,
+ "loss": 4.09483528137207,
+ "step": 6159
+ },
+ {
+ "epoch": 85.5589519650655,
+ "grad_norm": 0.01425520796328783,
+ "learning_rate": 0.0006,
+ "loss": 4.090396881103516,
+ "step": 6160
+ },
+ {
+ "epoch": 85.57292576419214,
+ "grad_norm": 0.01403278298676014,
+ "learning_rate": 0.0006,
+ "loss": 4.088724136352539,
+ "step": 6161
+ },
+ {
+ "epoch": 85.58689956331878,
+ "grad_norm": 0.01364967692643404,
+ "learning_rate": 0.0006,
+ "loss": 4.095558166503906,
+ "step": 6162
+ },
+ {
+ "epoch": 85.60087336244541,
+ "grad_norm": 0.013085267506539822,
+ "learning_rate": 0.0006,
+ "loss": 4.163962364196777,
+ "step": 6163
+ },
+ {
+ "epoch": 85.61484716157206,
+ "grad_norm": 0.0136906448751688,
+ "learning_rate": 0.0006,
+ "loss": 4.197134017944336,
+ "step": 6164
+ },
+ {
+ "epoch": 85.62882096069869,
+ "grad_norm": 0.013472451828420162,
+ "learning_rate": 0.0006,
+ "loss": 4.1578369140625,
+ "step": 6165
+ },
+ {
+ "epoch": 85.64279475982532,
+ "grad_norm": 0.012329773046076298,
+ "learning_rate": 0.0006,
+ "loss": 4.231907844543457,
+ "step": 6166
+ },
+ {
+ "epoch": 85.65676855895197,
+ "grad_norm": 0.011977200396358967,
+ "learning_rate": 0.0006,
+ "loss": 4.1552958488464355,
+ "step": 6167
+ },
+ {
+ "epoch": 85.6707423580786,
+ "grad_norm": 0.013049956411123276,
+ "learning_rate": 0.0006,
+ "loss": 4.17852783203125,
+ "step": 6168
+ },
+ {
+ "epoch": 85.68471615720524,
+ "grad_norm": 0.011440741829574108,
+ "learning_rate": 0.0006,
+ "loss": 4.144646644592285,
+ "step": 6169
+ },
+ {
+ "epoch": 85.69868995633188,
+ "grad_norm": 0.012472936883568764,
+ "learning_rate": 0.0006,
+ "loss": 4.2306413650512695,
+ "step": 6170
+ },
+ {
+ "epoch": 85.71266375545852,
+ "grad_norm": 0.013015356846153736,
+ "learning_rate": 0.0006,
+ "loss": 4.193439960479736,
+ "step": 6171
+ },
+ {
+ "epoch": 85.72663755458515,
+ "grad_norm": 0.013832319527864456,
+ "learning_rate": 0.0006,
+ "loss": 4.18073034286499,
+ "step": 6172
+ },
+ {
+ "epoch": 85.74061135371178,
+ "grad_norm": 0.014613982290029526,
+ "learning_rate": 0.0006,
+ "loss": 4.223531246185303,
+ "step": 6173
+ },
+ {
+ "epoch": 85.75458515283843,
+ "grad_norm": 0.014209083281457424,
+ "learning_rate": 0.0006,
+ "loss": 4.142666816711426,
+ "step": 6174
+ },
+ {
+ "epoch": 85.76855895196506,
+ "grad_norm": 0.013396669179201126,
+ "learning_rate": 0.0006,
+ "loss": 4.185786247253418,
+ "step": 6175
+ },
+ {
+ "epoch": 85.78253275109171,
+ "grad_norm": 0.01364929135888815,
+ "learning_rate": 0.0006,
+ "loss": 4.125226974487305,
+ "step": 6176
+ },
+ {
+ "epoch": 85.79650655021834,
+ "grad_norm": 0.012561003677546978,
+ "learning_rate": 0.0006,
+ "loss": 4.084519386291504,
+ "step": 6177
+ },
+ {
+ "epoch": 85.81048034934499,
+ "grad_norm": 0.013508481904864311,
+ "learning_rate": 0.0006,
+ "loss": 4.193173885345459,
+ "step": 6178
+ },
+ {
+ "epoch": 85.82445414847162,
+ "grad_norm": 0.014154613018035889,
+ "learning_rate": 0.0006,
+ "loss": 4.257272720336914,
+ "step": 6179
+ },
+ {
+ "epoch": 85.83842794759825,
+ "grad_norm": 0.012478531338274479,
+ "learning_rate": 0.0006,
+ "loss": 4.161380767822266,
+ "step": 6180
+ },
+ {
+ "epoch": 85.8524017467249,
+ "grad_norm": 0.013468451797962189,
+ "learning_rate": 0.0006,
+ "loss": 4.121395111083984,
+ "step": 6181
+ },
+ {
+ "epoch": 85.86637554585153,
+ "grad_norm": 0.015155194327235222,
+ "learning_rate": 0.0006,
+ "loss": 4.242815971374512,
+ "step": 6182
+ },
+ {
+ "epoch": 85.88034934497817,
+ "grad_norm": 0.016223272308707237,
+ "learning_rate": 0.0006,
+ "loss": 4.176027297973633,
+ "step": 6183
+ },
+ {
+ "epoch": 85.8943231441048,
+ "grad_norm": 0.013741295784711838,
+ "learning_rate": 0.0006,
+ "loss": 4.131257057189941,
+ "step": 6184
+ },
+ {
+ "epoch": 85.90829694323143,
+ "grad_norm": 0.014982067979872227,
+ "learning_rate": 0.0006,
+ "loss": 4.237896919250488,
+ "step": 6185
+ },
+ {
+ "epoch": 85.92227074235808,
+ "grad_norm": 0.018197322264313698,
+ "learning_rate": 0.0006,
+ "loss": 4.154962539672852,
+ "step": 6186
+ },
+ {
+ "epoch": 85.93624454148471,
+ "grad_norm": 0.01971200481057167,
+ "learning_rate": 0.0006,
+ "loss": 4.16819429397583,
+ "step": 6187
+ },
+ {
+ "epoch": 85.95021834061136,
+ "grad_norm": 0.018008720129728317,
+ "learning_rate": 0.0006,
+ "loss": 4.146164417266846,
+ "step": 6188
+ },
+ {
+ "epoch": 85.96419213973799,
+ "grad_norm": 0.015859264880418777,
+ "learning_rate": 0.0006,
+ "loss": 4.208099365234375,
+ "step": 6189
+ },
+ {
+ "epoch": 85.97816593886463,
+ "grad_norm": 0.013846252113580704,
+ "learning_rate": 0.0006,
+ "loss": 4.1881184577941895,
+ "step": 6190
+ },
+ {
+ "epoch": 85.99213973799127,
+ "grad_norm": 0.018641939386725426,
+ "learning_rate": 0.0006,
+ "loss": 4.127632141113281,
+ "step": 6191
+ },
+ {
+ "epoch": 86.0,
+ "grad_norm": 0.02385227382183075,
+ "learning_rate": 0.0006,
+ "loss": 4.281900405883789,
+ "step": 6192
+ },
+ {
+ "epoch": 86.0,
+ "eval_loss": 4.625458717346191,
+ "eval_runtime": 57.0634,
+ "eval_samples_per_second": 42.795,
+ "eval_steps_per_second": 1.349,
+ "step": 6192
+ },
+ {
+ "epoch": 86.01397379912663,
+ "grad_norm": 0.019032124429941177,
+ "learning_rate": 0.0006,
+ "loss": 4.2163190841674805,
+ "step": 6193
+ },
+ {
+ "epoch": 86.02794759825328,
+ "grad_norm": 0.018049925565719604,
+ "learning_rate": 0.0006,
+ "loss": 4.1534037590026855,
+ "step": 6194
+ },
+ {
+ "epoch": 86.04192139737991,
+ "grad_norm": 0.019116360694169998,
+ "learning_rate": 0.0006,
+ "loss": 4.212411403656006,
+ "step": 6195
+ },
+ {
+ "epoch": 86.05589519650655,
+ "grad_norm": 0.018967336043715477,
+ "learning_rate": 0.0006,
+ "loss": 4.190918922424316,
+ "step": 6196
+ },
+ {
+ "epoch": 86.06986899563319,
+ "grad_norm": 0.01945611648261547,
+ "learning_rate": 0.0006,
+ "loss": 4.18459415435791,
+ "step": 6197
+ },
+ {
+ "epoch": 86.08384279475983,
+ "grad_norm": 0.018104076385498047,
+ "learning_rate": 0.0006,
+ "loss": 4.0942816734313965,
+ "step": 6198
+ },
+ {
+ "epoch": 86.09781659388646,
+ "grad_norm": 0.01815992221236229,
+ "learning_rate": 0.0006,
+ "loss": 4.011547565460205,
+ "step": 6199
+ },
+ {
+ "epoch": 86.1117903930131,
+ "grad_norm": 0.017574431374669075,
+ "learning_rate": 0.0006,
+ "loss": 4.051593780517578,
+ "step": 6200
+ },
+ {
+ "epoch": 86.12576419213974,
+ "grad_norm": 0.01516508124768734,
+ "learning_rate": 0.0006,
+ "loss": 4.239673614501953,
+ "step": 6201
+ },
+ {
+ "epoch": 86.13973799126637,
+ "grad_norm": 0.015689702704548836,
+ "learning_rate": 0.0006,
+ "loss": 4.1085734367370605,
+ "step": 6202
+ },
+ {
+ "epoch": 86.15371179039302,
+ "grad_norm": 0.01596900261938572,
+ "learning_rate": 0.0006,
+ "loss": 4.1509599685668945,
+ "step": 6203
+ },
+ {
+ "epoch": 86.16768558951965,
+ "grad_norm": 0.01550096832215786,
+ "learning_rate": 0.0006,
+ "loss": 4.160238265991211,
+ "step": 6204
+ },
+ {
+ "epoch": 86.18165938864628,
+ "grad_norm": 0.014601529575884342,
+ "learning_rate": 0.0006,
+ "loss": 4.029219150543213,
+ "step": 6205
+ },
+ {
+ "epoch": 86.19563318777293,
+ "grad_norm": 0.014006740413606167,
+ "learning_rate": 0.0006,
+ "loss": 4.117501258850098,
+ "step": 6206
+ },
+ {
+ "epoch": 86.20960698689956,
+ "grad_norm": 0.013191957026720047,
+ "learning_rate": 0.0006,
+ "loss": 4.252002716064453,
+ "step": 6207
+ },
+ {
+ "epoch": 86.2235807860262,
+ "grad_norm": 0.013031814247369766,
+ "learning_rate": 0.0006,
+ "loss": 4.209517478942871,
+ "step": 6208
+ },
+ {
+ "epoch": 86.23755458515284,
+ "grad_norm": 0.013605714775621891,
+ "learning_rate": 0.0006,
+ "loss": 4.176309585571289,
+ "step": 6209
+ },
+ {
+ "epoch": 86.25152838427948,
+ "grad_norm": 0.012753087095916271,
+ "learning_rate": 0.0006,
+ "loss": 4.097990036010742,
+ "step": 6210
+ },
+ {
+ "epoch": 86.26550218340611,
+ "grad_norm": 0.012940628454089165,
+ "learning_rate": 0.0006,
+ "loss": 4.085589408874512,
+ "step": 6211
+ },
+ {
+ "epoch": 86.27947598253274,
+ "grad_norm": 0.014286619611084461,
+ "learning_rate": 0.0006,
+ "loss": 4.199261665344238,
+ "step": 6212
+ },
+ {
+ "epoch": 86.29344978165939,
+ "grad_norm": 0.014643525704741478,
+ "learning_rate": 0.0006,
+ "loss": 4.23718786239624,
+ "step": 6213
+ },
+ {
+ "epoch": 86.30742358078602,
+ "grad_norm": 0.015182147733867168,
+ "learning_rate": 0.0006,
+ "loss": 4.219855308532715,
+ "step": 6214
+ },
+ {
+ "epoch": 86.32139737991267,
+ "grad_norm": 0.01530102826654911,
+ "learning_rate": 0.0006,
+ "loss": 4.190094947814941,
+ "step": 6215
+ },
+ {
+ "epoch": 86.3353711790393,
+ "grad_norm": 0.015492603182792664,
+ "learning_rate": 0.0006,
+ "loss": 4.153780460357666,
+ "step": 6216
+ },
+ {
+ "epoch": 86.34934497816595,
+ "grad_norm": 0.015339907258749008,
+ "learning_rate": 0.0006,
+ "loss": 4.182865619659424,
+ "step": 6217
+ },
+ {
+ "epoch": 86.36331877729258,
+ "grad_norm": 0.015576880425214767,
+ "learning_rate": 0.0006,
+ "loss": 4.196417331695557,
+ "step": 6218
+ },
+ {
+ "epoch": 86.37729257641921,
+ "grad_norm": 0.014805737882852554,
+ "learning_rate": 0.0006,
+ "loss": 4.265018463134766,
+ "step": 6219
+ },
+ {
+ "epoch": 86.39126637554585,
+ "grad_norm": 0.01597665622830391,
+ "learning_rate": 0.0006,
+ "loss": 4.2864837646484375,
+ "step": 6220
+ },
+ {
+ "epoch": 86.40524017467249,
+ "grad_norm": 0.01693643443286419,
+ "learning_rate": 0.0006,
+ "loss": 4.118192672729492,
+ "step": 6221
+ },
+ {
+ "epoch": 86.41921397379913,
+ "grad_norm": 0.016648951917886734,
+ "learning_rate": 0.0006,
+ "loss": 4.23712682723999,
+ "step": 6222
+ },
+ {
+ "epoch": 86.43318777292576,
+ "grad_norm": 0.014809766784310341,
+ "learning_rate": 0.0006,
+ "loss": 4.251391410827637,
+ "step": 6223
+ },
+ {
+ "epoch": 86.44716157205241,
+ "grad_norm": 0.014333373866975307,
+ "learning_rate": 0.0006,
+ "loss": 4.17243766784668,
+ "step": 6224
+ },
+ {
+ "epoch": 86.46113537117904,
+ "grad_norm": 0.013912711292505264,
+ "learning_rate": 0.0006,
+ "loss": 4.250836372375488,
+ "step": 6225
+ },
+ {
+ "epoch": 86.47510917030567,
+ "grad_norm": 0.014821341261267662,
+ "learning_rate": 0.0006,
+ "loss": 4.073995590209961,
+ "step": 6226
+ },
+ {
+ "epoch": 86.48908296943232,
+ "grad_norm": 0.016634922474622726,
+ "learning_rate": 0.0006,
+ "loss": 4.155867576599121,
+ "step": 6227
+ },
+ {
+ "epoch": 86.50305676855895,
+ "grad_norm": 0.01748695969581604,
+ "learning_rate": 0.0006,
+ "loss": 4.139344692230225,
+ "step": 6228
+ },
+ {
+ "epoch": 86.5170305676856,
+ "grad_norm": 0.01569363661110401,
+ "learning_rate": 0.0006,
+ "loss": 4.23923397064209,
+ "step": 6229
+ },
+ {
+ "epoch": 86.53100436681223,
+ "grad_norm": 0.014037121087312698,
+ "learning_rate": 0.0006,
+ "loss": 4.200334072113037,
+ "step": 6230
+ },
+ {
+ "epoch": 86.54497816593886,
+ "grad_norm": 0.01401279866695404,
+ "learning_rate": 0.0006,
+ "loss": 4.035246849060059,
+ "step": 6231
+ },
+ {
+ "epoch": 86.5589519650655,
+ "grad_norm": 0.014603731222450733,
+ "learning_rate": 0.0006,
+ "loss": 4.141385555267334,
+ "step": 6232
+ },
+ {
+ "epoch": 86.57292576419214,
+ "grad_norm": 0.015113704837858677,
+ "learning_rate": 0.0006,
+ "loss": 4.130824089050293,
+ "step": 6233
+ },
+ {
+ "epoch": 86.58689956331878,
+ "grad_norm": 0.01588696800172329,
+ "learning_rate": 0.0006,
+ "loss": 4.05495023727417,
+ "step": 6234
+ },
+ {
+ "epoch": 86.60087336244541,
+ "grad_norm": 0.01593640074133873,
+ "learning_rate": 0.0006,
+ "loss": 4.22935152053833,
+ "step": 6235
+ },
+ {
+ "epoch": 86.61484716157206,
+ "grad_norm": 0.014292445965111256,
+ "learning_rate": 0.0006,
+ "loss": 4.309839248657227,
+ "step": 6236
+ },
+ {
+ "epoch": 86.62882096069869,
+ "grad_norm": 0.015471610240638256,
+ "learning_rate": 0.0006,
+ "loss": 4.268278121948242,
+ "step": 6237
+ },
+ {
+ "epoch": 86.64279475982532,
+ "grad_norm": 0.016643544659018517,
+ "learning_rate": 0.0006,
+ "loss": 4.135109901428223,
+ "step": 6238
+ },
+ {
+ "epoch": 86.65676855895197,
+ "grad_norm": 0.014733029529452324,
+ "learning_rate": 0.0006,
+ "loss": 4.060197830200195,
+ "step": 6239
+ },
+ {
+ "epoch": 86.6707423580786,
+ "grad_norm": 0.014869216829538345,
+ "learning_rate": 0.0006,
+ "loss": 4.126917362213135,
+ "step": 6240
+ },
+ {
+ "epoch": 86.68471615720524,
+ "grad_norm": 0.01608203910291195,
+ "learning_rate": 0.0006,
+ "loss": 4.183716773986816,
+ "step": 6241
+ },
+ {
+ "epoch": 86.69868995633188,
+ "grad_norm": 0.016178175806999207,
+ "learning_rate": 0.0006,
+ "loss": 4.118875503540039,
+ "step": 6242
+ },
+ {
+ "epoch": 86.71266375545852,
+ "grad_norm": 0.014576597139239311,
+ "learning_rate": 0.0006,
+ "loss": 4.261541366577148,
+ "step": 6243
+ },
+ {
+ "epoch": 86.72663755458515,
+ "grad_norm": 0.014466420747339725,
+ "learning_rate": 0.0006,
+ "loss": 4.307284355163574,
+ "step": 6244
+ },
+ {
+ "epoch": 86.74061135371178,
+ "grad_norm": 0.014042644761502743,
+ "learning_rate": 0.0006,
+ "loss": 4.2729997634887695,
+ "step": 6245
+ },
+ {
+ "epoch": 86.75458515283843,
+ "grad_norm": 0.014589524827897549,
+ "learning_rate": 0.0006,
+ "loss": 4.072981834411621,
+ "step": 6246
+ },
+ {
+ "epoch": 86.76855895196506,
+ "grad_norm": 0.018765483051538467,
+ "learning_rate": 0.0006,
+ "loss": 4.264769554138184,
+ "step": 6247
+ },
+ {
+ "epoch": 86.78253275109171,
+ "grad_norm": 0.019397731870412827,
+ "learning_rate": 0.0006,
+ "loss": 4.183379173278809,
+ "step": 6248
+ },
+ {
+ "epoch": 86.79650655021834,
+ "grad_norm": 0.017525402829051018,
+ "learning_rate": 0.0006,
+ "loss": 4.1494245529174805,
+ "step": 6249
+ },
+ {
+ "epoch": 86.81048034934499,
+ "grad_norm": 0.016310200095176697,
+ "learning_rate": 0.0006,
+ "loss": 4.149104118347168,
+ "step": 6250
+ },
+ {
+ "epoch": 86.82445414847162,
+ "grad_norm": 0.01506423857063055,
+ "learning_rate": 0.0006,
+ "loss": 4.22091817855835,
+ "step": 6251
+ },
+ {
+ "epoch": 86.83842794759825,
+ "grad_norm": 0.014200284145772457,
+ "learning_rate": 0.0006,
+ "loss": 4.177569389343262,
+ "step": 6252
+ },
+ {
+ "epoch": 86.8524017467249,
+ "grad_norm": 0.015016552060842514,
+ "learning_rate": 0.0006,
+ "loss": 4.089859962463379,
+ "step": 6253
+ },
+ {
+ "epoch": 86.86637554585153,
+ "grad_norm": 0.016054971143603325,
+ "learning_rate": 0.0006,
+ "loss": 4.157166004180908,
+ "step": 6254
+ },
+ {
+ "epoch": 86.88034934497817,
+ "grad_norm": 0.014388521201908588,
+ "learning_rate": 0.0006,
+ "loss": 4.1437578201293945,
+ "step": 6255
+ },
+ {
+ "epoch": 86.8943231441048,
+ "grad_norm": 0.015705369412899017,
+ "learning_rate": 0.0006,
+ "loss": 4.045689582824707,
+ "step": 6256
+ },
+ {
+ "epoch": 86.90829694323143,
+ "grad_norm": 0.014192395843565464,
+ "learning_rate": 0.0006,
+ "loss": 4.098793029785156,
+ "step": 6257
+ },
+ {
+ "epoch": 86.92227074235808,
+ "grad_norm": 0.013751034624874592,
+ "learning_rate": 0.0006,
+ "loss": 3.989920139312744,
+ "step": 6258
+ },
+ {
+ "epoch": 86.93624454148471,
+ "grad_norm": 0.01582193188369274,
+ "learning_rate": 0.0006,
+ "loss": 4.283449172973633,
+ "step": 6259
+ },
+ {
+ "epoch": 86.95021834061136,
+ "grad_norm": 0.01476013008505106,
+ "learning_rate": 0.0006,
+ "loss": 4.221401691436768,
+ "step": 6260
+ },
+ {
+ "epoch": 86.96419213973799,
+ "grad_norm": 0.01346337515860796,
+ "learning_rate": 0.0006,
+ "loss": 4.05896520614624,
+ "step": 6261
+ },
+ {
+ "epoch": 86.97816593886463,
+ "grad_norm": 0.012899307534098625,
+ "learning_rate": 0.0006,
+ "loss": 4.159560203552246,
+ "step": 6262
+ },
+ {
+ "epoch": 86.99213973799127,
+ "grad_norm": 0.013150574639439583,
+ "learning_rate": 0.0006,
+ "loss": 4.217330455780029,
+ "step": 6263
+ },
+ {
+ "epoch": 87.0,
+ "grad_norm": 0.014940370805561543,
+ "learning_rate": 0.0006,
+ "loss": 4.1611833572387695,
+ "step": 6264
+ },
+ {
+ "epoch": 87.0,
+ "eval_loss": 4.666043758392334,
+ "eval_runtime": 56.5978,
+ "eval_samples_per_second": 43.147,
+ "eval_steps_per_second": 1.36,
+ "step": 6264
+ },
+ {
+ "epoch": 87.01397379912663,
+ "grad_norm": 0.013721180148422718,
+ "learning_rate": 0.0006,
+ "loss": 4.142454147338867,
+ "step": 6265
+ },
+ {
+ "epoch": 87.02794759825328,
+ "grad_norm": 0.014805255457758904,
+ "learning_rate": 0.0006,
+ "loss": 4.121677875518799,
+ "step": 6266
+ },
+ {
+ "epoch": 87.04192139737991,
+ "grad_norm": 0.015576514415442944,
+ "learning_rate": 0.0006,
+ "loss": 4.1188435554504395,
+ "step": 6267
+ },
+ {
+ "epoch": 87.05589519650655,
+ "grad_norm": 0.01605502888560295,
+ "learning_rate": 0.0006,
+ "loss": 4.255428314208984,
+ "step": 6268
+ },
+ {
+ "epoch": 87.06986899563319,
+ "grad_norm": 0.014933343045413494,
+ "learning_rate": 0.0006,
+ "loss": 4.122596740722656,
+ "step": 6269
+ },
+ {
+ "epoch": 87.08384279475983,
+ "grad_norm": 0.014630894176661968,
+ "learning_rate": 0.0006,
+ "loss": 4.148405075073242,
+ "step": 6270
+ },
+ {
+ "epoch": 87.09781659388646,
+ "grad_norm": 0.014211163856089115,
+ "learning_rate": 0.0006,
+ "loss": 4.058030128479004,
+ "step": 6271
+ },
+ {
+ "epoch": 87.1117903930131,
+ "grad_norm": 0.014060168527066708,
+ "learning_rate": 0.0006,
+ "loss": 4.146586894989014,
+ "step": 6272
+ },
+ {
+ "epoch": 87.12576419213974,
+ "grad_norm": 0.014740100130438805,
+ "learning_rate": 0.0006,
+ "loss": 4.077739715576172,
+ "step": 6273
+ },
+ {
+ "epoch": 87.13973799126637,
+ "grad_norm": 0.013970664702355862,
+ "learning_rate": 0.0006,
+ "loss": 4.038252830505371,
+ "step": 6274
+ },
+ {
+ "epoch": 87.15371179039302,
+ "grad_norm": 0.014399951323866844,
+ "learning_rate": 0.0006,
+ "loss": 4.143700122833252,
+ "step": 6275
+ },
+ {
+ "epoch": 87.16768558951965,
+ "grad_norm": 0.01496410183608532,
+ "learning_rate": 0.0006,
+ "loss": 4.105609893798828,
+ "step": 6276
+ },
+ {
+ "epoch": 87.18165938864628,
+ "grad_norm": 0.014287355355918407,
+ "learning_rate": 0.0006,
+ "loss": 4.173645973205566,
+ "step": 6277
+ },
+ {
+ "epoch": 87.19563318777293,
+ "grad_norm": 0.0141292130574584,
+ "learning_rate": 0.0006,
+ "loss": 4.068511486053467,
+ "step": 6278
+ },
+ {
+ "epoch": 87.20960698689956,
+ "grad_norm": 0.01393523346632719,
+ "learning_rate": 0.0006,
+ "loss": 4.060687065124512,
+ "step": 6279
+ },
+ {
+ "epoch": 87.2235807860262,
+ "grad_norm": 0.01431021187454462,
+ "learning_rate": 0.0006,
+ "loss": 4.219209671020508,
+ "step": 6280
+ },
+ {
+ "epoch": 87.23755458515284,
+ "grad_norm": 0.014856027439236641,
+ "learning_rate": 0.0006,
+ "loss": 4.146749973297119,
+ "step": 6281
+ },
+ {
+ "epoch": 87.25152838427948,
+ "grad_norm": 0.013677539303898811,
+ "learning_rate": 0.0006,
+ "loss": 3.9859797954559326,
+ "step": 6282
+ },
+ {
+ "epoch": 87.26550218340611,
+ "grad_norm": 0.014078967273235321,
+ "learning_rate": 0.0006,
+ "loss": 4.129668235778809,
+ "step": 6283
+ },
+ {
+ "epoch": 87.27947598253274,
+ "grad_norm": 0.015564394183456898,
+ "learning_rate": 0.0006,
+ "loss": 4.0007829666137695,
+ "step": 6284
+ },
+ {
+ "epoch": 87.29344978165939,
+ "grad_norm": 0.015077749267220497,
+ "learning_rate": 0.0006,
+ "loss": 4.112264156341553,
+ "step": 6285
+ },
+ {
+ "epoch": 87.30742358078602,
+ "grad_norm": 0.015341218560934067,
+ "learning_rate": 0.0006,
+ "loss": 4.112249374389648,
+ "step": 6286
+ },
+ {
+ "epoch": 87.32139737991267,
+ "grad_norm": 0.016399366781115532,
+ "learning_rate": 0.0006,
+ "loss": 4.176196098327637,
+ "step": 6287
+ },
+ {
+ "epoch": 87.3353711790393,
+ "grad_norm": 0.01599763333797455,
+ "learning_rate": 0.0006,
+ "loss": 4.163441181182861,
+ "step": 6288
+ },
+ {
+ "epoch": 87.34934497816595,
+ "grad_norm": 0.015439776703715324,
+ "learning_rate": 0.0006,
+ "loss": 4.11902379989624,
+ "step": 6289
+ },
+ {
+ "epoch": 87.36331877729258,
+ "grad_norm": 0.015909943729639053,
+ "learning_rate": 0.0006,
+ "loss": 4.16621732711792,
+ "step": 6290
+ },
+ {
+ "epoch": 87.37729257641921,
+ "grad_norm": 0.016710078343749046,
+ "learning_rate": 0.0006,
+ "loss": 4.044960975646973,
+ "step": 6291
+ },
+ {
+ "epoch": 87.39126637554585,
+ "grad_norm": 0.015883933752775192,
+ "learning_rate": 0.0006,
+ "loss": 4.065184593200684,
+ "step": 6292
+ },
+ {
+ "epoch": 87.40524017467249,
+ "grad_norm": 0.013912747614085674,
+ "learning_rate": 0.0006,
+ "loss": 4.102450370788574,
+ "step": 6293
+ },
+ {
+ "epoch": 87.41921397379913,
+ "grad_norm": 0.013170420192182064,
+ "learning_rate": 0.0006,
+ "loss": 4.234364032745361,
+ "step": 6294
+ },
+ {
+ "epoch": 87.43318777292576,
+ "grad_norm": 0.014043302275240421,
+ "learning_rate": 0.0006,
+ "loss": 4.129565238952637,
+ "step": 6295
+ },
+ {
+ "epoch": 87.44716157205241,
+ "grad_norm": 0.015010139904916286,
+ "learning_rate": 0.0006,
+ "loss": 4.166678428649902,
+ "step": 6296
+ },
+ {
+ "epoch": 87.46113537117904,
+ "grad_norm": 0.014721784740686417,
+ "learning_rate": 0.0006,
+ "loss": 4.182618618011475,
+ "step": 6297
+ },
+ {
+ "epoch": 87.47510917030567,
+ "grad_norm": 0.015713468194007874,
+ "learning_rate": 0.0006,
+ "loss": 4.103855133056641,
+ "step": 6298
+ },
+ {
+ "epoch": 87.48908296943232,
+ "grad_norm": 0.017204247415065765,
+ "learning_rate": 0.0006,
+ "loss": 4.140937328338623,
+ "step": 6299
+ },
+ {
+ "epoch": 87.50305676855895,
+ "grad_norm": 0.016856668516993523,
+ "learning_rate": 0.0006,
+ "loss": 4.147859573364258,
+ "step": 6300
+ },
+ {
+ "epoch": 87.5170305676856,
+ "grad_norm": 0.017205018550157547,
+ "learning_rate": 0.0006,
+ "loss": 4.1667327880859375,
+ "step": 6301
+ },
+ {
+ "epoch": 87.53100436681223,
+ "grad_norm": 0.014821539632976055,
+ "learning_rate": 0.0006,
+ "loss": 4.089935302734375,
+ "step": 6302
+ },
+ {
+ "epoch": 87.54497816593886,
+ "grad_norm": 0.01446484588086605,
+ "learning_rate": 0.0006,
+ "loss": 4.243657112121582,
+ "step": 6303
+ },
+ {
+ "epoch": 87.5589519650655,
+ "grad_norm": 0.014374484308063984,
+ "learning_rate": 0.0006,
+ "loss": 4.14356803894043,
+ "step": 6304
+ },
+ {
+ "epoch": 87.57292576419214,
+ "grad_norm": 0.01420109998434782,
+ "learning_rate": 0.0006,
+ "loss": 4.1273651123046875,
+ "step": 6305
+ },
+ {
+ "epoch": 87.58689956331878,
+ "grad_norm": 0.014810388907790184,
+ "learning_rate": 0.0006,
+ "loss": 4.222290515899658,
+ "step": 6306
+ },
+ {
+ "epoch": 87.60087336244541,
+ "grad_norm": 0.015826983377337456,
+ "learning_rate": 0.0006,
+ "loss": 4.137996196746826,
+ "step": 6307
+ },
+ {
+ "epoch": 87.61484716157206,
+ "grad_norm": 0.014457454904913902,
+ "learning_rate": 0.0006,
+ "loss": 4.0780744552612305,
+ "step": 6308
+ },
+ {
+ "epoch": 87.62882096069869,
+ "grad_norm": 0.014458604156970978,
+ "learning_rate": 0.0006,
+ "loss": 4.1287922859191895,
+ "step": 6309
+ },
+ {
+ "epoch": 87.64279475982532,
+ "grad_norm": 0.015282071195542812,
+ "learning_rate": 0.0006,
+ "loss": 4.189508438110352,
+ "step": 6310
+ },
+ {
+ "epoch": 87.65676855895197,
+ "grad_norm": 0.017536642029881477,
+ "learning_rate": 0.0006,
+ "loss": 4.070173263549805,
+ "step": 6311
+ },
+ {
+ "epoch": 87.6707423580786,
+ "grad_norm": 0.017303580418229103,
+ "learning_rate": 0.0006,
+ "loss": 4.119133949279785,
+ "step": 6312
+ },
+ {
+ "epoch": 87.68471615720524,
+ "grad_norm": 0.016148580238223076,
+ "learning_rate": 0.0006,
+ "loss": 3.994354486465454,
+ "step": 6313
+ },
+ {
+ "epoch": 87.69868995633188,
+ "grad_norm": 0.01706942543387413,
+ "learning_rate": 0.0006,
+ "loss": 4.183113098144531,
+ "step": 6314
+ },
+ {
+ "epoch": 87.71266375545852,
+ "grad_norm": 0.016971010714769363,
+ "learning_rate": 0.0006,
+ "loss": 4.104086875915527,
+ "step": 6315
+ },
+ {
+ "epoch": 87.72663755458515,
+ "grad_norm": 0.015516847372055054,
+ "learning_rate": 0.0006,
+ "loss": 4.2409772872924805,
+ "step": 6316
+ },
+ {
+ "epoch": 87.74061135371178,
+ "grad_norm": 0.015352269634604454,
+ "learning_rate": 0.0006,
+ "loss": 4.026427268981934,
+ "step": 6317
+ },
+ {
+ "epoch": 87.75458515283843,
+ "grad_norm": 0.017702415585517883,
+ "learning_rate": 0.0006,
+ "loss": 4.177840232849121,
+ "step": 6318
+ },
+ {
+ "epoch": 87.76855895196506,
+ "grad_norm": 0.01629541628062725,
+ "learning_rate": 0.0006,
+ "loss": 4.229039192199707,
+ "step": 6319
+ },
+ {
+ "epoch": 87.78253275109171,
+ "grad_norm": 0.01503711100667715,
+ "learning_rate": 0.0006,
+ "loss": 4.214350700378418,
+ "step": 6320
+ },
+ {
+ "epoch": 87.79650655021834,
+ "grad_norm": 0.014913883060216904,
+ "learning_rate": 0.0006,
+ "loss": 4.168880939483643,
+ "step": 6321
+ },
+ {
+ "epoch": 87.81048034934499,
+ "grad_norm": 0.014513909816741943,
+ "learning_rate": 0.0006,
+ "loss": 4.162267684936523,
+ "step": 6322
+ },
+ {
+ "epoch": 87.82445414847162,
+ "grad_norm": 0.014635873027145863,
+ "learning_rate": 0.0006,
+ "loss": 4.220067977905273,
+ "step": 6323
+ },
+ {
+ "epoch": 87.83842794759825,
+ "grad_norm": 0.017102796584367752,
+ "learning_rate": 0.0006,
+ "loss": 4.1493377685546875,
+ "step": 6324
+ },
+ {
+ "epoch": 87.8524017467249,
+ "grad_norm": 0.015533884055912495,
+ "learning_rate": 0.0006,
+ "loss": 4.0646209716796875,
+ "step": 6325
+ },
+ {
+ "epoch": 87.86637554585153,
+ "grad_norm": 0.014617550186812878,
+ "learning_rate": 0.0006,
+ "loss": 4.1390533447265625,
+ "step": 6326
+ },
+ {
+ "epoch": 87.88034934497817,
+ "grad_norm": 0.014142967760562897,
+ "learning_rate": 0.0006,
+ "loss": 4.133934020996094,
+ "step": 6327
+ },
+ {
+ "epoch": 87.8943231441048,
+ "grad_norm": 0.013182433322072029,
+ "learning_rate": 0.0006,
+ "loss": 4.125327110290527,
+ "step": 6328
+ },
+ {
+ "epoch": 87.90829694323143,
+ "grad_norm": 0.013928229920566082,
+ "learning_rate": 0.0006,
+ "loss": 4.142777442932129,
+ "step": 6329
+ },
+ {
+ "epoch": 87.92227074235808,
+ "grad_norm": 0.015337013639509678,
+ "learning_rate": 0.0006,
+ "loss": 4.140210151672363,
+ "step": 6330
+ },
+ {
+ "epoch": 87.93624454148471,
+ "grad_norm": 0.014873536303639412,
+ "learning_rate": 0.0006,
+ "loss": 4.102636337280273,
+ "step": 6331
+ },
+ {
+ "epoch": 87.95021834061136,
+ "grad_norm": 0.01387431938201189,
+ "learning_rate": 0.0006,
+ "loss": 4.149059295654297,
+ "step": 6332
+ },
+ {
+ "epoch": 87.96419213973799,
+ "grad_norm": 0.013234943151473999,
+ "learning_rate": 0.0006,
+ "loss": 4.1564459800720215,
+ "step": 6333
+ },
+ {
+ "epoch": 87.97816593886463,
+ "grad_norm": 0.013781017623841763,
+ "learning_rate": 0.0006,
+ "loss": 4.189839839935303,
+ "step": 6334
+ },
+ {
+ "epoch": 87.99213973799127,
+ "grad_norm": 0.013733026571571827,
+ "learning_rate": 0.0006,
+ "loss": 4.29277229309082,
+ "step": 6335
+ },
+ {
+ "epoch": 88.0,
+ "grad_norm": 0.014752327464520931,
+ "learning_rate": 0.0006,
+ "loss": 4.163867950439453,
+ "step": 6336
+ },
+ {
+ "epoch": 88.0,
+ "eval_loss": 4.535566329956055,
+ "eval_runtime": 56.4677,
+ "eval_samples_per_second": 43.246,
+ "eval_steps_per_second": 1.364,
+ "step": 6336
+ },
+ {
+ "epoch": 88.01397379912663,
+ "grad_norm": 0.013958732597529888,
+ "learning_rate": 0.0006,
+ "loss": 4.200681209564209,
+ "step": 6337
+ },
+ {
+ "epoch": 88.02794759825328,
+ "grad_norm": 0.015283545479178429,
+ "learning_rate": 0.0006,
+ "loss": 4.1736016273498535,
+ "step": 6338
+ },
+ {
+ "epoch": 88.04192139737991,
+ "grad_norm": 0.016192059963941574,
+ "learning_rate": 0.0006,
+ "loss": 4.040589809417725,
+ "step": 6339
+ },
+ {
+ "epoch": 88.05589519650655,
+ "grad_norm": 0.018096154555678368,
+ "learning_rate": 0.0006,
+ "loss": 4.146696090698242,
+ "step": 6340
+ },
+ {
+ "epoch": 88.06986899563319,
+ "grad_norm": 0.01797664538025856,
+ "learning_rate": 0.0006,
+ "loss": 4.007248878479004,
+ "step": 6341
+ },
+ {
+ "epoch": 88.08384279475983,
+ "grad_norm": 0.016275612637400627,
+ "learning_rate": 0.0006,
+ "loss": 4.193206787109375,
+ "step": 6342
+ },
+ {
+ "epoch": 88.09781659388646,
+ "grad_norm": 0.014210258610546589,
+ "learning_rate": 0.0006,
+ "loss": 4.064548492431641,
+ "step": 6343
+ },
+ {
+ "epoch": 88.1117903930131,
+ "grad_norm": 0.016085151582956314,
+ "learning_rate": 0.0006,
+ "loss": 4.003668785095215,
+ "step": 6344
+ },
+ {
+ "epoch": 88.12576419213974,
+ "grad_norm": 0.01930483989417553,
+ "learning_rate": 0.0006,
+ "loss": 4.236578941345215,
+ "step": 6345
+ },
+ {
+ "epoch": 88.13973799126637,
+ "grad_norm": 0.017140906304121017,
+ "learning_rate": 0.0006,
+ "loss": 3.9593005180358887,
+ "step": 6346
+ },
+ {
+ "epoch": 88.15371179039302,
+ "grad_norm": 0.01521560549736023,
+ "learning_rate": 0.0006,
+ "loss": 4.137644290924072,
+ "step": 6347
+ },
+ {
+ "epoch": 88.16768558951965,
+ "grad_norm": 0.015788141638040543,
+ "learning_rate": 0.0006,
+ "loss": 4.124845027923584,
+ "step": 6348
+ },
+ {
+ "epoch": 88.18165938864628,
+ "grad_norm": 0.01784597896039486,
+ "learning_rate": 0.0006,
+ "loss": 4.139750957489014,
+ "step": 6349
+ },
+ {
+ "epoch": 88.19563318777293,
+ "grad_norm": 0.016266558319330215,
+ "learning_rate": 0.0006,
+ "loss": 4.074851989746094,
+ "step": 6350
+ },
+ {
+ "epoch": 88.20960698689956,
+ "grad_norm": 0.014024483971297741,
+ "learning_rate": 0.0006,
+ "loss": 4.164689064025879,
+ "step": 6351
+ },
+ {
+ "epoch": 88.2235807860262,
+ "grad_norm": 0.015694420784711838,
+ "learning_rate": 0.0006,
+ "loss": 4.219979286193848,
+ "step": 6352
+ },
+ {
+ "epoch": 88.23755458515284,
+ "grad_norm": 0.014897859655320644,
+ "learning_rate": 0.0006,
+ "loss": 4.050738334655762,
+ "step": 6353
+ },
+ {
+ "epoch": 88.25152838427948,
+ "grad_norm": 0.014827374368906021,
+ "learning_rate": 0.0006,
+ "loss": 4.137135028839111,
+ "step": 6354
+ },
+ {
+ "epoch": 88.26550218340611,
+ "grad_norm": 0.01454879529774189,
+ "learning_rate": 0.0006,
+ "loss": 3.9980897903442383,
+ "step": 6355
+ },
+ {
+ "epoch": 88.27947598253274,
+ "grad_norm": 0.015623947605490685,
+ "learning_rate": 0.0006,
+ "loss": 4.226551055908203,
+ "step": 6356
+ },
+ {
+ "epoch": 88.29344978165939,
+ "grad_norm": 0.015527035109698772,
+ "learning_rate": 0.0006,
+ "loss": 4.11053466796875,
+ "step": 6357
+ },
+ {
+ "epoch": 88.30742358078602,
+ "grad_norm": 0.015242574736475945,
+ "learning_rate": 0.0006,
+ "loss": 4.105966567993164,
+ "step": 6358
+ },
+ {
+ "epoch": 88.32139737991267,
+ "grad_norm": 0.01375538855791092,
+ "learning_rate": 0.0006,
+ "loss": 4.148453712463379,
+ "step": 6359
+ },
+ {
+ "epoch": 88.3353711790393,
+ "grad_norm": 0.01595226116478443,
+ "learning_rate": 0.0006,
+ "loss": 4.095836162567139,
+ "step": 6360
+ },
+ {
+ "epoch": 88.34934497816595,
+ "grad_norm": 0.015252740122377872,
+ "learning_rate": 0.0006,
+ "loss": 4.136335849761963,
+ "step": 6361
+ },
+ {
+ "epoch": 88.36331877729258,
+ "grad_norm": 0.014405866153538227,
+ "learning_rate": 0.0006,
+ "loss": 4.084691047668457,
+ "step": 6362
+ },
+ {
+ "epoch": 88.37729257641921,
+ "grad_norm": 0.014755490235984325,
+ "learning_rate": 0.0006,
+ "loss": 4.095475673675537,
+ "step": 6363
+ },
+ {
+ "epoch": 88.39126637554585,
+ "grad_norm": 0.015339708887040615,
+ "learning_rate": 0.0006,
+ "loss": 4.163136959075928,
+ "step": 6364
+ },
+ {
+ "epoch": 88.40524017467249,
+ "grad_norm": 0.014667164534330368,
+ "learning_rate": 0.0006,
+ "loss": 4.062265396118164,
+ "step": 6365
+ },
+ {
+ "epoch": 88.41921397379913,
+ "grad_norm": 0.014588423073291779,
+ "learning_rate": 0.0006,
+ "loss": 4.13359260559082,
+ "step": 6366
+ },
+ {
+ "epoch": 88.43318777292576,
+ "grad_norm": 0.014005377888679504,
+ "learning_rate": 0.0006,
+ "loss": 4.219711780548096,
+ "step": 6367
+ },
+ {
+ "epoch": 88.44716157205241,
+ "grad_norm": 0.012769930064678192,
+ "learning_rate": 0.0006,
+ "loss": 4.114658355712891,
+ "step": 6368
+ },
+ {
+ "epoch": 88.46113537117904,
+ "grad_norm": 0.013632693327963352,
+ "learning_rate": 0.0006,
+ "loss": 4.0675225257873535,
+ "step": 6369
+ },
+ {
+ "epoch": 88.47510917030567,
+ "grad_norm": 0.013829804956912994,
+ "learning_rate": 0.0006,
+ "loss": 4.1125969886779785,
+ "step": 6370
+ },
+ {
+ "epoch": 88.48908296943232,
+ "grad_norm": 0.013882109895348549,
+ "learning_rate": 0.0006,
+ "loss": 4.1627092361450195,
+ "step": 6371
+ },
+ {
+ "epoch": 88.50305676855895,
+ "grad_norm": 0.014023828320205212,
+ "learning_rate": 0.0006,
+ "loss": 4.01038932800293,
+ "step": 6372
+ },
+ {
+ "epoch": 88.5170305676856,
+ "grad_norm": 0.014840706251561642,
+ "learning_rate": 0.0006,
+ "loss": 4.117877960205078,
+ "step": 6373
+ },
+ {
+ "epoch": 88.53100436681223,
+ "grad_norm": 0.014165930449962616,
+ "learning_rate": 0.0006,
+ "loss": 4.1864237785339355,
+ "step": 6374
+ },
+ {
+ "epoch": 88.54497816593886,
+ "grad_norm": 0.014390201307833195,
+ "learning_rate": 0.0006,
+ "loss": 4.1044602394104,
+ "step": 6375
+ },
+ {
+ "epoch": 88.5589519650655,
+ "grad_norm": 0.015037395060062408,
+ "learning_rate": 0.0006,
+ "loss": 4.101107120513916,
+ "step": 6376
+ },
+ {
+ "epoch": 88.57292576419214,
+ "grad_norm": 0.01671452447772026,
+ "learning_rate": 0.0006,
+ "loss": 4.170581817626953,
+ "step": 6377
+ },
+ {
+ "epoch": 88.58689956331878,
+ "grad_norm": 0.01993076130747795,
+ "learning_rate": 0.0006,
+ "loss": 4.261412620544434,
+ "step": 6378
+ },
+ {
+ "epoch": 88.60087336244541,
+ "grad_norm": 0.023721016943454742,
+ "learning_rate": 0.0006,
+ "loss": 4.199363708496094,
+ "step": 6379
+ },
+ {
+ "epoch": 88.61484716157206,
+ "grad_norm": 0.023726163432002068,
+ "learning_rate": 0.0006,
+ "loss": 4.061285018920898,
+ "step": 6380
+ },
+ {
+ "epoch": 88.62882096069869,
+ "grad_norm": 0.018352877348661423,
+ "learning_rate": 0.0006,
+ "loss": 4.0721282958984375,
+ "step": 6381
+ },
+ {
+ "epoch": 88.64279475982532,
+ "grad_norm": 0.01679830253124237,
+ "learning_rate": 0.0006,
+ "loss": 4.1717939376831055,
+ "step": 6382
+ },
+ {
+ "epoch": 88.65676855895197,
+ "grad_norm": 0.017193637788295746,
+ "learning_rate": 0.0006,
+ "loss": 4.230783939361572,
+ "step": 6383
+ },
+ {
+ "epoch": 88.6707423580786,
+ "grad_norm": 0.01581464521586895,
+ "learning_rate": 0.0006,
+ "loss": 4.029906272888184,
+ "step": 6384
+ },
+ {
+ "epoch": 88.68471615720524,
+ "grad_norm": 0.014049704186618328,
+ "learning_rate": 0.0006,
+ "loss": 4.192776679992676,
+ "step": 6385
+ },
+ {
+ "epoch": 88.69868995633188,
+ "grad_norm": 0.017214450985193253,
+ "learning_rate": 0.0006,
+ "loss": 4.248916149139404,
+ "step": 6386
+ },
+ {
+ "epoch": 88.71266375545852,
+ "grad_norm": 0.0173103716224432,
+ "learning_rate": 0.0006,
+ "loss": 4.124347686767578,
+ "step": 6387
+ },
+ {
+ "epoch": 88.72663755458515,
+ "grad_norm": 0.015745092183351517,
+ "learning_rate": 0.0006,
+ "loss": 4.192878246307373,
+ "step": 6388
+ },
+ {
+ "epoch": 88.74061135371178,
+ "grad_norm": 0.013750772923231125,
+ "learning_rate": 0.0006,
+ "loss": 4.1564483642578125,
+ "step": 6389
+ },
+ {
+ "epoch": 88.75458515283843,
+ "grad_norm": 0.014969858340919018,
+ "learning_rate": 0.0006,
+ "loss": 4.0980424880981445,
+ "step": 6390
+ },
+ {
+ "epoch": 88.76855895196506,
+ "grad_norm": 0.01287173479795456,
+ "learning_rate": 0.0006,
+ "loss": 4.145712852478027,
+ "step": 6391
+ },
+ {
+ "epoch": 88.78253275109171,
+ "grad_norm": 0.012716300785541534,
+ "learning_rate": 0.0006,
+ "loss": 4.214226722717285,
+ "step": 6392
+ },
+ {
+ "epoch": 88.79650655021834,
+ "grad_norm": 0.013855204917490482,
+ "learning_rate": 0.0006,
+ "loss": 4.1667256355285645,
+ "step": 6393
+ },
+ {
+ "epoch": 88.81048034934499,
+ "grad_norm": 0.013650864362716675,
+ "learning_rate": 0.0006,
+ "loss": 4.161942481994629,
+ "step": 6394
+ },
+ {
+ "epoch": 88.82445414847162,
+ "grad_norm": 0.013780126348137856,
+ "learning_rate": 0.0006,
+ "loss": 4.087825775146484,
+ "step": 6395
+ },
+ {
+ "epoch": 88.83842794759825,
+ "grad_norm": 0.014029948972165585,
+ "learning_rate": 0.0006,
+ "loss": 4.142813205718994,
+ "step": 6396
+ },
+ {
+ "epoch": 88.8524017467249,
+ "grad_norm": 0.01322692260146141,
+ "learning_rate": 0.0006,
+ "loss": 4.106776714324951,
+ "step": 6397
+ },
+ {
+ "epoch": 88.86637554585153,
+ "grad_norm": 0.013522760011255741,
+ "learning_rate": 0.0006,
+ "loss": 4.090996265411377,
+ "step": 6398
+ },
+ {
+ "epoch": 88.88034934497817,
+ "grad_norm": 0.014474684372544289,
+ "learning_rate": 0.0006,
+ "loss": 4.124624252319336,
+ "step": 6399
+ },
+ {
+ "epoch": 88.8943231441048,
+ "grad_norm": 0.014490422792732716,
+ "learning_rate": 0.0006,
+ "loss": 4.065431594848633,
+ "step": 6400
+ },
+ {
+ "epoch": 88.90829694323143,
+ "grad_norm": 0.015379710122942924,
+ "learning_rate": 0.0006,
+ "loss": 4.207939624786377,
+ "step": 6401
+ },
+ {
+ "epoch": 88.92227074235808,
+ "grad_norm": 0.015914618968963623,
+ "learning_rate": 0.0006,
+ "loss": 4.041152000427246,
+ "step": 6402
+ },
+ {
+ "epoch": 88.93624454148471,
+ "grad_norm": 0.01480048056691885,
+ "learning_rate": 0.0006,
+ "loss": 4.138754844665527,
+ "step": 6403
+ },
+ {
+ "epoch": 88.95021834061136,
+ "grad_norm": 0.013734851032495499,
+ "learning_rate": 0.0006,
+ "loss": 4.25323486328125,
+ "step": 6404
+ },
+ {
+ "epoch": 88.96419213973799,
+ "grad_norm": 0.01452311035245657,
+ "learning_rate": 0.0006,
+ "loss": 4.115228176116943,
+ "step": 6405
+ },
+ {
+ "epoch": 88.97816593886463,
+ "grad_norm": 0.016197681427001953,
+ "learning_rate": 0.0006,
+ "loss": 4.099531173706055,
+ "step": 6406
+ },
+ {
+ "epoch": 88.99213973799127,
+ "grad_norm": 0.017169523984193802,
+ "learning_rate": 0.0006,
+ "loss": 4.182981491088867,
+ "step": 6407
+ },
+ {
+ "epoch": 89.0,
+ "grad_norm": 0.017855646088719368,
+ "learning_rate": 0.0006,
+ "loss": 4.1470136642456055,
+ "step": 6408
+ },
+ {
+ "epoch": 89.0,
+ "eval_loss": 4.6035614013671875,
+ "eval_runtime": 57.0625,
+ "eval_samples_per_second": 42.795,
+ "eval_steps_per_second": 1.349,
+ "step": 6408
+ },
+ {
+ "epoch": 89.01397379912663,
+ "grad_norm": 0.014834101311862469,
+ "learning_rate": 0.0006,
+ "loss": 4.159243583679199,
+ "step": 6409
+ },
+ {
+ "epoch": 89.02794759825328,
+ "grad_norm": 0.015316772274672985,
+ "learning_rate": 0.0006,
+ "loss": 4.0709123611450195,
+ "step": 6410
+ },
+ {
+ "epoch": 89.04192139737991,
+ "grad_norm": 0.01576591096818447,
+ "learning_rate": 0.0006,
+ "loss": 4.0921196937561035,
+ "step": 6411
+ },
+ {
+ "epoch": 89.05589519650655,
+ "grad_norm": 0.015948045998811722,
+ "learning_rate": 0.0006,
+ "loss": 4.181877136230469,
+ "step": 6412
+ },
+ {
+ "epoch": 89.06986899563319,
+ "grad_norm": 0.016612380743026733,
+ "learning_rate": 0.0006,
+ "loss": 4.081249237060547,
+ "step": 6413
+ },
+ {
+ "epoch": 89.08384279475983,
+ "grad_norm": 0.017455685883760452,
+ "learning_rate": 0.0006,
+ "loss": 4.175990104675293,
+ "step": 6414
+ },
+ {
+ "epoch": 89.09781659388646,
+ "grad_norm": 0.015040521509945393,
+ "learning_rate": 0.0006,
+ "loss": 4.121378421783447,
+ "step": 6415
+ },
+ {
+ "epoch": 89.1117903930131,
+ "grad_norm": 0.014513209462165833,
+ "learning_rate": 0.0006,
+ "loss": 3.996107816696167,
+ "step": 6416
+ },
+ {
+ "epoch": 89.12576419213974,
+ "grad_norm": 0.015455638989806175,
+ "learning_rate": 0.0006,
+ "loss": 4.073675155639648,
+ "step": 6417
+ },
+ {
+ "epoch": 89.13973799126637,
+ "grad_norm": 0.015468129888176918,
+ "learning_rate": 0.0006,
+ "loss": 4.130242824554443,
+ "step": 6418
+ },
+ {
+ "epoch": 89.15371179039302,
+ "grad_norm": 0.015284297987818718,
+ "learning_rate": 0.0006,
+ "loss": 4.18555212020874,
+ "step": 6419
+ },
+ {
+ "epoch": 89.16768558951965,
+ "grad_norm": 0.01476874016225338,
+ "learning_rate": 0.0006,
+ "loss": 4.134175777435303,
+ "step": 6420
+ },
+ {
+ "epoch": 89.18165938864628,
+ "grad_norm": 0.015836201608181,
+ "learning_rate": 0.0006,
+ "loss": 4.094529151916504,
+ "step": 6421
+ },
+ {
+ "epoch": 89.19563318777293,
+ "grad_norm": 0.0161928478628397,
+ "learning_rate": 0.0006,
+ "loss": 4.12647819519043,
+ "step": 6422
+ },
+ {
+ "epoch": 89.20960698689956,
+ "grad_norm": 0.015680814161896706,
+ "learning_rate": 0.0006,
+ "loss": 3.9620392322540283,
+ "step": 6423
+ },
+ {
+ "epoch": 89.2235807860262,
+ "grad_norm": 0.01584681309759617,
+ "learning_rate": 0.0006,
+ "loss": 4.049304008483887,
+ "step": 6424
+ },
+ {
+ "epoch": 89.23755458515284,
+ "grad_norm": 0.01452907919883728,
+ "learning_rate": 0.0006,
+ "loss": 4.064027786254883,
+ "step": 6425
+ },
+ {
+ "epoch": 89.25152838427948,
+ "grad_norm": 0.014002552255988121,
+ "learning_rate": 0.0006,
+ "loss": 4.151862621307373,
+ "step": 6426
+ },
+ {
+ "epoch": 89.26550218340611,
+ "grad_norm": 0.014658625237643719,
+ "learning_rate": 0.0006,
+ "loss": 4.1672210693359375,
+ "step": 6427
+ },
+ {
+ "epoch": 89.27947598253274,
+ "grad_norm": 0.015021766535937786,
+ "learning_rate": 0.0006,
+ "loss": 3.9742586612701416,
+ "step": 6428
+ },
+ {
+ "epoch": 89.29344978165939,
+ "grad_norm": 0.014681321568787098,
+ "learning_rate": 0.0006,
+ "loss": 4.024159908294678,
+ "step": 6429
+ },
+ {
+ "epoch": 89.30742358078602,
+ "grad_norm": 0.014161549508571625,
+ "learning_rate": 0.0006,
+ "loss": 4.105640888214111,
+ "step": 6430
+ },
+ {
+ "epoch": 89.32139737991267,
+ "grad_norm": 0.014801453799009323,
+ "learning_rate": 0.0006,
+ "loss": 4.123049736022949,
+ "step": 6431
+ },
+ {
+ "epoch": 89.3353711790393,
+ "grad_norm": 0.014307181350886822,
+ "learning_rate": 0.0006,
+ "loss": 4.086954593658447,
+ "step": 6432
+ },
+ {
+ "epoch": 89.34934497816595,
+ "grad_norm": 0.013299369253218174,
+ "learning_rate": 0.0006,
+ "loss": 4.109533786773682,
+ "step": 6433
+ },
+ {
+ "epoch": 89.36331877729258,
+ "grad_norm": 0.014388669282197952,
+ "learning_rate": 0.0006,
+ "loss": 4.222370147705078,
+ "step": 6434
+ },
+ {
+ "epoch": 89.37729257641921,
+ "grad_norm": 0.014501879923045635,
+ "learning_rate": 0.0006,
+ "loss": 4.224178314208984,
+ "step": 6435
+ },
+ {
+ "epoch": 89.39126637554585,
+ "grad_norm": 0.014109021984040737,
+ "learning_rate": 0.0006,
+ "loss": 4.112702369689941,
+ "step": 6436
+ },
+ {
+ "epoch": 89.40524017467249,
+ "grad_norm": 0.013521002605557442,
+ "learning_rate": 0.0006,
+ "loss": 4.010124683380127,
+ "step": 6437
+ },
+ {
+ "epoch": 89.41921397379913,
+ "grad_norm": 0.013288144953548908,
+ "learning_rate": 0.0006,
+ "loss": 4.127592086791992,
+ "step": 6438
+ },
+ {
+ "epoch": 89.43318777292576,
+ "grad_norm": 0.013882486149668694,
+ "learning_rate": 0.0006,
+ "loss": 4.031871318817139,
+ "step": 6439
+ },
+ {
+ "epoch": 89.44716157205241,
+ "grad_norm": 0.01289335172623396,
+ "learning_rate": 0.0006,
+ "loss": 4.0757551193237305,
+ "step": 6440
+ },
+ {
+ "epoch": 89.46113537117904,
+ "grad_norm": 0.01356302760541439,
+ "learning_rate": 0.0006,
+ "loss": 4.075678825378418,
+ "step": 6441
+ },
+ {
+ "epoch": 89.47510917030567,
+ "grad_norm": 0.01447434164583683,
+ "learning_rate": 0.0006,
+ "loss": 4.093453407287598,
+ "step": 6442
+ },
+ {
+ "epoch": 89.48908296943232,
+ "grad_norm": 0.014443481341004372,
+ "learning_rate": 0.0006,
+ "loss": 4.075372219085693,
+ "step": 6443
+ },
+ {
+ "epoch": 89.50305676855895,
+ "grad_norm": 0.01314218994230032,
+ "learning_rate": 0.0006,
+ "loss": 4.171533584594727,
+ "step": 6444
+ },
+ {
+ "epoch": 89.5170305676856,
+ "grad_norm": 0.01348346658051014,
+ "learning_rate": 0.0006,
+ "loss": 4.227236270904541,
+ "step": 6445
+ },
+ {
+ "epoch": 89.53100436681223,
+ "grad_norm": 0.013370679691433907,
+ "learning_rate": 0.0006,
+ "loss": 3.973062515258789,
+ "step": 6446
+ },
+ {
+ "epoch": 89.54497816593886,
+ "grad_norm": 0.01410642359405756,
+ "learning_rate": 0.0006,
+ "loss": 4.231175422668457,
+ "step": 6447
+ },
+ {
+ "epoch": 89.5589519650655,
+ "grad_norm": 0.01419337373226881,
+ "learning_rate": 0.0006,
+ "loss": 4.145868301391602,
+ "step": 6448
+ },
+ {
+ "epoch": 89.57292576419214,
+ "grad_norm": 0.013691053725779057,
+ "learning_rate": 0.0006,
+ "loss": 4.175131797790527,
+ "step": 6449
+ },
+ {
+ "epoch": 89.58689956331878,
+ "grad_norm": 0.013714492321014404,
+ "learning_rate": 0.0006,
+ "loss": 4.251893997192383,
+ "step": 6450
+ },
+ {
+ "epoch": 89.60087336244541,
+ "grad_norm": 0.01393197476863861,
+ "learning_rate": 0.0006,
+ "loss": 4.047079086303711,
+ "step": 6451
+ },
+ {
+ "epoch": 89.61484716157206,
+ "grad_norm": 0.013957309536635876,
+ "learning_rate": 0.0006,
+ "loss": 4.128530502319336,
+ "step": 6452
+ },
+ {
+ "epoch": 89.62882096069869,
+ "grad_norm": 0.013346714898943901,
+ "learning_rate": 0.0006,
+ "loss": 4.2296929359436035,
+ "step": 6453
+ },
+ {
+ "epoch": 89.64279475982532,
+ "grad_norm": 0.014243829064071178,
+ "learning_rate": 0.0006,
+ "loss": 4.128233909606934,
+ "step": 6454
+ },
+ {
+ "epoch": 89.65676855895197,
+ "grad_norm": 0.01606106385588646,
+ "learning_rate": 0.0006,
+ "loss": 4.202202320098877,
+ "step": 6455
+ },
+ {
+ "epoch": 89.6707423580786,
+ "grad_norm": 0.016074348241090775,
+ "learning_rate": 0.0006,
+ "loss": 4.215323448181152,
+ "step": 6456
+ },
+ {
+ "epoch": 89.68471615720524,
+ "grad_norm": 0.01611054316163063,
+ "learning_rate": 0.0006,
+ "loss": 4.096758842468262,
+ "step": 6457
+ },
+ {
+ "epoch": 89.69868995633188,
+ "grad_norm": 0.013788217678666115,
+ "learning_rate": 0.0006,
+ "loss": 4.098560810089111,
+ "step": 6458
+ },
+ {
+ "epoch": 89.71266375545852,
+ "grad_norm": 0.014408969320356846,
+ "learning_rate": 0.0006,
+ "loss": 4.107398509979248,
+ "step": 6459
+ },
+ {
+ "epoch": 89.72663755458515,
+ "grad_norm": 0.014765267260372639,
+ "learning_rate": 0.0006,
+ "loss": 4.057718276977539,
+ "step": 6460
+ },
+ {
+ "epoch": 89.74061135371178,
+ "grad_norm": 0.017395280301570892,
+ "learning_rate": 0.0006,
+ "loss": 4.100795269012451,
+ "step": 6461
+ },
+ {
+ "epoch": 89.75458515283843,
+ "grad_norm": 0.016821149736642838,
+ "learning_rate": 0.0006,
+ "loss": 4.153731346130371,
+ "step": 6462
+ },
+ {
+ "epoch": 89.76855895196506,
+ "grad_norm": 0.015633124858140945,
+ "learning_rate": 0.0006,
+ "loss": 4.061585903167725,
+ "step": 6463
+ },
+ {
+ "epoch": 89.78253275109171,
+ "grad_norm": 0.01451497245579958,
+ "learning_rate": 0.0006,
+ "loss": 4.1624884605407715,
+ "step": 6464
+ },
+ {
+ "epoch": 89.79650655021834,
+ "grad_norm": 0.013868392445147038,
+ "learning_rate": 0.0006,
+ "loss": 4.178536415100098,
+ "step": 6465
+ },
+ {
+ "epoch": 89.81048034934499,
+ "grad_norm": 0.014259875752031803,
+ "learning_rate": 0.0006,
+ "loss": 4.073114395141602,
+ "step": 6466
+ },
+ {
+ "epoch": 89.82445414847162,
+ "grad_norm": 0.013651594519615173,
+ "learning_rate": 0.0006,
+ "loss": 4.050265312194824,
+ "step": 6467
+ },
+ {
+ "epoch": 89.83842794759825,
+ "grad_norm": 0.014063458889722824,
+ "learning_rate": 0.0006,
+ "loss": 4.2014689445495605,
+ "step": 6468
+ },
+ {
+ "epoch": 89.8524017467249,
+ "grad_norm": 0.014179578050971031,
+ "learning_rate": 0.0006,
+ "loss": 4.110550880432129,
+ "step": 6469
+ },
+ {
+ "epoch": 89.86637554585153,
+ "grad_norm": 0.013616854324936867,
+ "learning_rate": 0.0006,
+ "loss": 4.1776580810546875,
+ "step": 6470
+ },
+ {
+ "epoch": 89.88034934497817,
+ "grad_norm": 0.014351330697536469,
+ "learning_rate": 0.0006,
+ "loss": 4.058526992797852,
+ "step": 6471
+ },
+ {
+ "epoch": 89.8943231441048,
+ "grad_norm": 0.0141244325786829,
+ "learning_rate": 0.0006,
+ "loss": 4.068673610687256,
+ "step": 6472
+ },
+ {
+ "epoch": 89.90829694323143,
+ "grad_norm": 0.01552131213247776,
+ "learning_rate": 0.0006,
+ "loss": 4.086787223815918,
+ "step": 6473
+ },
+ {
+ "epoch": 89.92227074235808,
+ "grad_norm": 0.014614894054830074,
+ "learning_rate": 0.0006,
+ "loss": 4.2190656661987305,
+ "step": 6474
+ },
+ {
+ "epoch": 89.93624454148471,
+ "grad_norm": 0.014369758777320385,
+ "learning_rate": 0.0006,
+ "loss": 4.117021560668945,
+ "step": 6475
+ },
+ {
+ "epoch": 89.95021834061136,
+ "grad_norm": 0.015352046117186546,
+ "learning_rate": 0.0006,
+ "loss": 4.167664527893066,
+ "step": 6476
+ },
+ {
+ "epoch": 89.96419213973799,
+ "grad_norm": 0.014314834959805012,
+ "learning_rate": 0.0006,
+ "loss": 4.160182952880859,
+ "step": 6477
+ },
+ {
+ "epoch": 89.97816593886463,
+ "grad_norm": 0.013511341996490955,
+ "learning_rate": 0.0006,
+ "loss": 4.096034049987793,
+ "step": 6478
+ },
+ {
+ "epoch": 89.99213973799127,
+ "grad_norm": 0.01308482326567173,
+ "learning_rate": 0.0006,
+ "loss": 4.062689781188965,
+ "step": 6479
+ },
+ {
+ "epoch": 90.0,
+ "grad_norm": 0.01442806702107191,
+ "learning_rate": 0.0006,
+ "loss": 4.1955885887146,
+ "step": 6480
+ },
+ {
+ "epoch": 90.0,
+ "eval_loss": 4.644347667694092,
+ "eval_runtime": 56.8677,
+ "eval_samples_per_second": 42.942,
+ "eval_steps_per_second": 1.354,
+ "step": 6480
+ },
+ {
+ "epoch": 90.01397379912663,
+ "grad_norm": 0.014839007519185543,
+ "learning_rate": 0.0006,
+ "loss": 4.171868324279785,
+ "step": 6481
+ },
+ {
+ "epoch": 90.02794759825328,
+ "grad_norm": 0.01476900465786457,
+ "learning_rate": 0.0006,
+ "loss": 4.119754791259766,
+ "step": 6482
+ },
+ {
+ "epoch": 90.04192139737991,
+ "grad_norm": 0.015179301612079144,
+ "learning_rate": 0.0006,
+ "loss": 4.06740140914917,
+ "step": 6483
+ },
+ {
+ "epoch": 90.05589519650655,
+ "grad_norm": 0.015552980825304985,
+ "learning_rate": 0.0006,
+ "loss": 4.112268447875977,
+ "step": 6484
+ },
+ {
+ "epoch": 90.06986899563319,
+ "grad_norm": 0.014930234290659428,
+ "learning_rate": 0.0006,
+ "loss": 4.071866989135742,
+ "step": 6485
+ },
+ {
+ "epoch": 90.08384279475983,
+ "grad_norm": 0.014124148525297642,
+ "learning_rate": 0.0006,
+ "loss": 4.013731956481934,
+ "step": 6486
+ },
+ {
+ "epoch": 90.09781659388646,
+ "grad_norm": 0.014204252511262894,
+ "learning_rate": 0.0006,
+ "loss": 4.133620738983154,
+ "step": 6487
+ },
+ {
+ "epoch": 90.1117903930131,
+ "grad_norm": 0.014845543541014194,
+ "learning_rate": 0.0006,
+ "loss": 4.122956275939941,
+ "step": 6488
+ },
+ {
+ "epoch": 90.12576419213974,
+ "grad_norm": 0.0146744754165411,
+ "learning_rate": 0.0006,
+ "loss": 4.087121963500977,
+ "step": 6489
+ },
+ {
+ "epoch": 90.13973799126637,
+ "grad_norm": 0.016007190570235252,
+ "learning_rate": 0.0006,
+ "loss": 3.98740816116333,
+ "step": 6490
+ },
+ {
+ "epoch": 90.15371179039302,
+ "grad_norm": 0.018072661012411118,
+ "learning_rate": 0.0006,
+ "loss": 4.1155009269714355,
+ "step": 6491
+ },
+ {
+ "epoch": 90.16768558951965,
+ "grad_norm": 0.016868166625499725,
+ "learning_rate": 0.0006,
+ "loss": 4.123727321624756,
+ "step": 6492
+ },
+ {
+ "epoch": 90.18165938864628,
+ "grad_norm": 0.017091726884245872,
+ "learning_rate": 0.0006,
+ "loss": 4.1405229568481445,
+ "step": 6493
+ },
+ {
+ "epoch": 90.19563318777293,
+ "grad_norm": 0.018329855054616928,
+ "learning_rate": 0.0006,
+ "loss": 4.093900680541992,
+ "step": 6494
+ },
+ {
+ "epoch": 90.20960698689956,
+ "grad_norm": 0.018906457349658012,
+ "learning_rate": 0.0006,
+ "loss": 4.0552849769592285,
+ "step": 6495
+ },
+ {
+ "epoch": 90.2235807860262,
+ "grad_norm": 0.017275869846343994,
+ "learning_rate": 0.0006,
+ "loss": 4.1310625076293945,
+ "step": 6496
+ },
+ {
+ "epoch": 90.23755458515284,
+ "grad_norm": 0.015029897913336754,
+ "learning_rate": 0.0006,
+ "loss": 4.100166320800781,
+ "step": 6497
+ },
+ {
+ "epoch": 90.25152838427948,
+ "grad_norm": 0.014891199767589569,
+ "learning_rate": 0.0006,
+ "loss": 4.1314802169799805,
+ "step": 6498
+ },
+ {
+ "epoch": 90.26550218340611,
+ "grad_norm": 0.014877690002322197,
+ "learning_rate": 0.0006,
+ "loss": 4.076098442077637,
+ "step": 6499
+ },
+ {
+ "epoch": 90.27947598253274,
+ "grad_norm": 0.014670864678919315,
+ "learning_rate": 0.0006,
+ "loss": 4.066137313842773,
+ "step": 6500
+ },
+ {
+ "epoch": 90.29344978165939,
+ "grad_norm": 0.01444234699010849,
+ "learning_rate": 0.0006,
+ "loss": 4.013591766357422,
+ "step": 6501
+ },
+ {
+ "epoch": 90.30742358078602,
+ "grad_norm": 0.01339263841509819,
+ "learning_rate": 0.0006,
+ "loss": 4.117959499359131,
+ "step": 6502
+ },
+ {
+ "epoch": 90.32139737991267,
+ "grad_norm": 0.014753810130059719,
+ "learning_rate": 0.0006,
+ "loss": 4.042947769165039,
+ "step": 6503
+ },
+ {
+ "epoch": 90.3353711790393,
+ "grad_norm": 0.014947964809834957,
+ "learning_rate": 0.0006,
+ "loss": 4.112810134887695,
+ "step": 6504
+ },
+ {
+ "epoch": 90.34934497816595,
+ "grad_norm": 0.015155353583395481,
+ "learning_rate": 0.0006,
+ "loss": 4.213214874267578,
+ "step": 6505
+ },
+ {
+ "epoch": 90.36331877729258,
+ "grad_norm": 0.015502216294407845,
+ "learning_rate": 0.0006,
+ "loss": 4.127619743347168,
+ "step": 6506
+ },
+ {
+ "epoch": 90.37729257641921,
+ "grad_norm": 0.015377781353890896,
+ "learning_rate": 0.0006,
+ "loss": 4.135862350463867,
+ "step": 6507
+ },
+ {
+ "epoch": 90.39126637554585,
+ "grad_norm": 0.01662006787955761,
+ "learning_rate": 0.0006,
+ "loss": 4.076877593994141,
+ "step": 6508
+ },
+ {
+ "epoch": 90.40524017467249,
+ "grad_norm": 0.016324395313858986,
+ "learning_rate": 0.0006,
+ "loss": 4.214895725250244,
+ "step": 6509
+ },
+ {
+ "epoch": 90.41921397379913,
+ "grad_norm": 0.016099024564027786,
+ "learning_rate": 0.0006,
+ "loss": 4.139717102050781,
+ "step": 6510
+ },
+ {
+ "epoch": 90.43318777292576,
+ "grad_norm": 0.014403503388166428,
+ "learning_rate": 0.0006,
+ "loss": 4.040343284606934,
+ "step": 6511
+ },
+ {
+ "epoch": 90.44716157205241,
+ "grad_norm": 0.01413416862487793,
+ "learning_rate": 0.0006,
+ "loss": 4.087702751159668,
+ "step": 6512
+ },
+ {
+ "epoch": 90.46113537117904,
+ "grad_norm": 0.01603989489376545,
+ "learning_rate": 0.0006,
+ "loss": 4.173558712005615,
+ "step": 6513
+ },
+ {
+ "epoch": 90.47510917030567,
+ "grad_norm": 0.016846206039190292,
+ "learning_rate": 0.0006,
+ "loss": 4.1955461502075195,
+ "step": 6514
+ },
+ {
+ "epoch": 90.48908296943232,
+ "grad_norm": 0.01652812585234642,
+ "learning_rate": 0.0006,
+ "loss": 4.061838150024414,
+ "step": 6515
+ },
+ {
+ "epoch": 90.50305676855895,
+ "grad_norm": 0.015100528486073017,
+ "learning_rate": 0.0006,
+ "loss": 4.041808605194092,
+ "step": 6516
+ },
+ {
+ "epoch": 90.5170305676856,
+ "grad_norm": 0.013867822475731373,
+ "learning_rate": 0.0006,
+ "loss": 4.12662410736084,
+ "step": 6517
+ },
+ {
+ "epoch": 90.53100436681223,
+ "grad_norm": 0.013684107922017574,
+ "learning_rate": 0.0006,
+ "loss": 4.1254048347473145,
+ "step": 6518
+ },
+ {
+ "epoch": 90.54497816593886,
+ "grad_norm": 0.014200013130903244,
+ "learning_rate": 0.0006,
+ "loss": 4.183216094970703,
+ "step": 6519
+ },
+ {
+ "epoch": 90.5589519650655,
+ "grad_norm": 0.013707022182643414,
+ "learning_rate": 0.0006,
+ "loss": 4.07176399230957,
+ "step": 6520
+ },
+ {
+ "epoch": 90.57292576419214,
+ "grad_norm": 0.013996582478284836,
+ "learning_rate": 0.0006,
+ "loss": 4.006575584411621,
+ "step": 6521
+ },
+ {
+ "epoch": 90.58689956331878,
+ "grad_norm": 0.014886210672557354,
+ "learning_rate": 0.0006,
+ "loss": 4.118956565856934,
+ "step": 6522
+ },
+ {
+ "epoch": 90.60087336244541,
+ "grad_norm": 0.014022685587406158,
+ "learning_rate": 0.0006,
+ "loss": 3.96986722946167,
+ "step": 6523
+ },
+ {
+ "epoch": 90.61484716157206,
+ "grad_norm": 0.014547634869813919,
+ "learning_rate": 0.0006,
+ "loss": 4.122886657714844,
+ "step": 6524
+ },
+ {
+ "epoch": 90.62882096069869,
+ "grad_norm": 0.013682851567864418,
+ "learning_rate": 0.0006,
+ "loss": 3.988292694091797,
+ "step": 6525
+ },
+ {
+ "epoch": 90.64279475982532,
+ "grad_norm": 0.014016421511769295,
+ "learning_rate": 0.0006,
+ "loss": 4.110270977020264,
+ "step": 6526
+ },
+ {
+ "epoch": 90.65676855895197,
+ "grad_norm": 0.013392233289778233,
+ "learning_rate": 0.0006,
+ "loss": 4.065059661865234,
+ "step": 6527
+ },
+ {
+ "epoch": 90.6707423580786,
+ "grad_norm": 0.012107795104384422,
+ "learning_rate": 0.0006,
+ "loss": 4.034260272979736,
+ "step": 6528
+ },
+ {
+ "epoch": 90.68471615720524,
+ "grad_norm": 0.013198604807257652,
+ "learning_rate": 0.0006,
+ "loss": 4.162083625793457,
+ "step": 6529
+ },
+ {
+ "epoch": 90.69868995633188,
+ "grad_norm": 0.012380579486489296,
+ "learning_rate": 0.0006,
+ "loss": 4.040456295013428,
+ "step": 6530
+ },
+ {
+ "epoch": 90.71266375545852,
+ "grad_norm": 0.012784135527908802,
+ "learning_rate": 0.0006,
+ "loss": 4.0823798179626465,
+ "step": 6531
+ },
+ {
+ "epoch": 90.72663755458515,
+ "grad_norm": 0.012920220382511616,
+ "learning_rate": 0.0006,
+ "loss": 4.181064128875732,
+ "step": 6532
+ },
+ {
+ "epoch": 90.74061135371178,
+ "grad_norm": 0.01338814664632082,
+ "learning_rate": 0.0006,
+ "loss": 4.111660003662109,
+ "step": 6533
+ },
+ {
+ "epoch": 90.75458515283843,
+ "grad_norm": 0.013628040440380573,
+ "learning_rate": 0.0006,
+ "loss": 4.237852573394775,
+ "step": 6534
+ },
+ {
+ "epoch": 90.76855895196506,
+ "grad_norm": 0.013384854421019554,
+ "learning_rate": 0.0006,
+ "loss": 4.121862888336182,
+ "step": 6535
+ },
+ {
+ "epoch": 90.78253275109171,
+ "grad_norm": 0.014726929366588593,
+ "learning_rate": 0.0006,
+ "loss": 4.1695990562438965,
+ "step": 6536
+ },
+ {
+ "epoch": 90.79650655021834,
+ "grad_norm": 0.0161006897687912,
+ "learning_rate": 0.0006,
+ "loss": 4.17794132232666,
+ "step": 6537
+ },
+ {
+ "epoch": 90.81048034934499,
+ "grad_norm": 0.016761884093284607,
+ "learning_rate": 0.0006,
+ "loss": 4.093764781951904,
+ "step": 6538
+ },
+ {
+ "epoch": 90.82445414847162,
+ "grad_norm": 0.017246825620532036,
+ "learning_rate": 0.0006,
+ "loss": 3.9382331371307373,
+ "step": 6539
+ },
+ {
+ "epoch": 90.83842794759825,
+ "grad_norm": 0.018141290172934532,
+ "learning_rate": 0.0006,
+ "loss": 3.973609209060669,
+ "step": 6540
+ },
+ {
+ "epoch": 90.8524017467249,
+ "grad_norm": 0.018129250034689903,
+ "learning_rate": 0.0006,
+ "loss": 4.227414131164551,
+ "step": 6541
+ },
+ {
+ "epoch": 90.86637554585153,
+ "grad_norm": 0.018234573304653168,
+ "learning_rate": 0.0006,
+ "loss": 4.151870250701904,
+ "step": 6542
+ },
+ {
+ "epoch": 90.88034934497817,
+ "grad_norm": 0.01560172438621521,
+ "learning_rate": 0.0006,
+ "loss": 4.127397537231445,
+ "step": 6543
+ },
+ {
+ "epoch": 90.8943231441048,
+ "grad_norm": 0.014367681927978992,
+ "learning_rate": 0.0006,
+ "loss": 4.17820930480957,
+ "step": 6544
+ },
+ {
+ "epoch": 90.90829694323143,
+ "grad_norm": 0.016071319580078125,
+ "learning_rate": 0.0006,
+ "loss": 4.059854507446289,
+ "step": 6545
+ },
+ {
+ "epoch": 90.92227074235808,
+ "grad_norm": 0.01589169353246689,
+ "learning_rate": 0.0006,
+ "loss": 4.282748222351074,
+ "step": 6546
+ },
+ {
+ "epoch": 90.93624454148471,
+ "grad_norm": 0.013639772310853004,
+ "learning_rate": 0.0006,
+ "loss": 4.147368907928467,
+ "step": 6547
+ },
+ {
+ "epoch": 90.95021834061136,
+ "grad_norm": 0.014248427003622055,
+ "learning_rate": 0.0006,
+ "loss": 4.240248680114746,
+ "step": 6548
+ },
+ {
+ "epoch": 90.96419213973799,
+ "grad_norm": 0.014600320719182491,
+ "learning_rate": 0.0006,
+ "loss": 4.122123718261719,
+ "step": 6549
+ },
+ {
+ "epoch": 90.97816593886463,
+ "grad_norm": 0.014371675439178944,
+ "learning_rate": 0.0006,
+ "loss": 4.013302803039551,
+ "step": 6550
+ },
+ {
+ "epoch": 90.99213973799127,
+ "grad_norm": 0.013133526779711246,
+ "learning_rate": 0.0006,
+ "loss": 4.097136497497559,
+ "step": 6551
+ },
+ {
+ "epoch": 91.0,
+ "grad_norm": 0.015599353238940239,
+ "learning_rate": 0.0006,
+ "loss": 4.198554039001465,
+ "step": 6552
+ },
+ {
+ "epoch": 91.0,
+ "eval_loss": 4.603907585144043,
+ "eval_runtime": 56.9418,
+ "eval_samples_per_second": 42.886,
+ "eval_steps_per_second": 1.352,
+ "step": 6552
+ },
+ {
+ "epoch": 91.01397379912663,
+ "grad_norm": 0.015917908400297165,
+ "learning_rate": 0.0006,
+ "loss": 4.133877754211426,
+ "step": 6553
+ },
+ {
+ "epoch": 91.02794759825328,
+ "grad_norm": 0.015485800802707672,
+ "learning_rate": 0.0006,
+ "loss": 4.111617088317871,
+ "step": 6554
+ },
+ {
+ "epoch": 91.04192139737991,
+ "grad_norm": 0.012969603762030602,
+ "learning_rate": 0.0006,
+ "loss": 4.064828395843506,
+ "step": 6555
+ },
+ {
+ "epoch": 91.05589519650655,
+ "grad_norm": 0.014482097700238228,
+ "learning_rate": 0.0006,
+ "loss": 4.101710319519043,
+ "step": 6556
+ },
+ {
+ "epoch": 91.06986899563319,
+ "grad_norm": 0.017242740839719772,
+ "learning_rate": 0.0006,
+ "loss": 4.015342712402344,
+ "step": 6557
+ },
+ {
+ "epoch": 91.08384279475983,
+ "grad_norm": 0.017100440338253975,
+ "learning_rate": 0.0006,
+ "loss": 3.9801602363586426,
+ "step": 6558
+ },
+ {
+ "epoch": 91.09781659388646,
+ "grad_norm": 0.015969131141901016,
+ "learning_rate": 0.0006,
+ "loss": 4.049958229064941,
+ "step": 6559
+ },
+ {
+ "epoch": 91.1117903930131,
+ "grad_norm": 0.016018927097320557,
+ "learning_rate": 0.0006,
+ "loss": 4.094139099121094,
+ "step": 6560
+ },
+ {
+ "epoch": 91.12576419213974,
+ "grad_norm": 0.018131926655769348,
+ "learning_rate": 0.0006,
+ "loss": 3.9499974250793457,
+ "step": 6561
+ },
+ {
+ "epoch": 91.13973799126637,
+ "grad_norm": 0.019425220787525177,
+ "learning_rate": 0.0006,
+ "loss": 3.9910552501678467,
+ "step": 6562
+ },
+ {
+ "epoch": 91.15371179039302,
+ "grad_norm": 0.01908663846552372,
+ "learning_rate": 0.0006,
+ "loss": 4.032179832458496,
+ "step": 6563
+ },
+ {
+ "epoch": 91.16768558951965,
+ "grad_norm": 0.01871207356452942,
+ "learning_rate": 0.0006,
+ "loss": 4.112380027770996,
+ "step": 6564
+ },
+ {
+ "epoch": 91.18165938864628,
+ "grad_norm": 0.020867953076958656,
+ "learning_rate": 0.0006,
+ "loss": 4.06105899810791,
+ "step": 6565
+ },
+ {
+ "epoch": 91.19563318777293,
+ "grad_norm": 0.019053412601351738,
+ "learning_rate": 0.0006,
+ "loss": 3.992504596710205,
+ "step": 6566
+ },
+ {
+ "epoch": 91.20960698689956,
+ "grad_norm": 0.016084402799606323,
+ "learning_rate": 0.0006,
+ "loss": 4.100559234619141,
+ "step": 6567
+ },
+ {
+ "epoch": 91.2235807860262,
+ "grad_norm": 0.016857590526342392,
+ "learning_rate": 0.0006,
+ "loss": 4.1567277908325195,
+ "step": 6568
+ },
+ {
+ "epoch": 91.23755458515284,
+ "grad_norm": 0.01702957972884178,
+ "learning_rate": 0.0006,
+ "loss": 4.234771728515625,
+ "step": 6569
+ },
+ {
+ "epoch": 91.25152838427948,
+ "grad_norm": 0.016869494691491127,
+ "learning_rate": 0.0006,
+ "loss": 3.9931387901306152,
+ "step": 6570
+ },
+ {
+ "epoch": 91.26550218340611,
+ "grad_norm": 0.01696581020951271,
+ "learning_rate": 0.0006,
+ "loss": 4.184908866882324,
+ "step": 6571
+ },
+ {
+ "epoch": 91.27947598253274,
+ "grad_norm": 0.014982126653194427,
+ "learning_rate": 0.0006,
+ "loss": 4.036668300628662,
+ "step": 6572
+ },
+ {
+ "epoch": 91.29344978165939,
+ "grad_norm": 0.014363301917910576,
+ "learning_rate": 0.0006,
+ "loss": 4.074010848999023,
+ "step": 6573
+ },
+ {
+ "epoch": 91.30742358078602,
+ "grad_norm": 0.015492337755858898,
+ "learning_rate": 0.0006,
+ "loss": 4.127626419067383,
+ "step": 6574
+ },
+ {
+ "epoch": 91.32139737991267,
+ "grad_norm": 0.015375050716102123,
+ "learning_rate": 0.0006,
+ "loss": 4.159229755401611,
+ "step": 6575
+ },
+ {
+ "epoch": 91.3353711790393,
+ "grad_norm": 0.015068565495312214,
+ "learning_rate": 0.0006,
+ "loss": 4.172623634338379,
+ "step": 6576
+ },
+ {
+ "epoch": 91.34934497816595,
+ "grad_norm": 0.0140256742015481,
+ "learning_rate": 0.0006,
+ "loss": 4.207065105438232,
+ "step": 6577
+ },
+ {
+ "epoch": 91.36331877729258,
+ "grad_norm": 0.014220152981579304,
+ "learning_rate": 0.0006,
+ "loss": 4.131444931030273,
+ "step": 6578
+ },
+ {
+ "epoch": 91.37729257641921,
+ "grad_norm": 0.0143149234354496,
+ "learning_rate": 0.0006,
+ "loss": 4.065218448638916,
+ "step": 6579
+ },
+ {
+ "epoch": 91.39126637554585,
+ "grad_norm": 0.013149659149348736,
+ "learning_rate": 0.0006,
+ "loss": 4.106743335723877,
+ "step": 6580
+ },
+ {
+ "epoch": 91.40524017467249,
+ "grad_norm": 0.014550330117344856,
+ "learning_rate": 0.0006,
+ "loss": 4.123722553253174,
+ "step": 6581
+ },
+ {
+ "epoch": 91.41921397379913,
+ "grad_norm": 0.014611710794270039,
+ "learning_rate": 0.0006,
+ "loss": 4.079610824584961,
+ "step": 6582
+ },
+ {
+ "epoch": 91.43318777292576,
+ "grad_norm": 0.013369769789278507,
+ "learning_rate": 0.0006,
+ "loss": 4.133187770843506,
+ "step": 6583
+ },
+ {
+ "epoch": 91.44716157205241,
+ "grad_norm": 0.013243545778095722,
+ "learning_rate": 0.0006,
+ "loss": 4.042607307434082,
+ "step": 6584
+ },
+ {
+ "epoch": 91.46113537117904,
+ "grad_norm": 0.013388266786932945,
+ "learning_rate": 0.0006,
+ "loss": 4.14863395690918,
+ "step": 6585
+ },
+ {
+ "epoch": 91.47510917030567,
+ "grad_norm": 0.014333273284137249,
+ "learning_rate": 0.0006,
+ "loss": 4.109699726104736,
+ "step": 6586
+ },
+ {
+ "epoch": 91.48908296943232,
+ "grad_norm": 0.012742757797241211,
+ "learning_rate": 0.0006,
+ "loss": 4.126582145690918,
+ "step": 6587
+ },
+ {
+ "epoch": 91.50305676855895,
+ "grad_norm": 0.012880304828286171,
+ "learning_rate": 0.0006,
+ "loss": 4.205300331115723,
+ "step": 6588
+ },
+ {
+ "epoch": 91.5170305676856,
+ "grad_norm": 0.013896043412387371,
+ "learning_rate": 0.0006,
+ "loss": 4.132510185241699,
+ "step": 6589
+ },
+ {
+ "epoch": 91.53100436681223,
+ "grad_norm": 0.014361281879246235,
+ "learning_rate": 0.0006,
+ "loss": 4.1487932205200195,
+ "step": 6590
+ },
+ {
+ "epoch": 91.54497816593886,
+ "grad_norm": 0.015137741342186928,
+ "learning_rate": 0.0006,
+ "loss": 4.1564531326293945,
+ "step": 6591
+ },
+ {
+ "epoch": 91.5589519650655,
+ "grad_norm": 0.014660682529211044,
+ "learning_rate": 0.0006,
+ "loss": 4.056064128875732,
+ "step": 6592
+ },
+ {
+ "epoch": 91.57292576419214,
+ "grad_norm": 0.014888476580381393,
+ "learning_rate": 0.0006,
+ "loss": 4.0633344650268555,
+ "step": 6593
+ },
+ {
+ "epoch": 91.58689956331878,
+ "grad_norm": 0.012162433005869389,
+ "learning_rate": 0.0006,
+ "loss": 4.026293754577637,
+ "step": 6594
+ },
+ {
+ "epoch": 91.60087336244541,
+ "grad_norm": 0.012892271392047405,
+ "learning_rate": 0.0006,
+ "loss": 4.267086029052734,
+ "step": 6595
+ },
+ {
+ "epoch": 91.61484716157206,
+ "grad_norm": 0.01450759544968605,
+ "learning_rate": 0.0006,
+ "loss": 4.149359703063965,
+ "step": 6596
+ },
+ {
+ "epoch": 91.62882096069869,
+ "grad_norm": 0.01419843453913927,
+ "learning_rate": 0.0006,
+ "loss": 4.094183921813965,
+ "step": 6597
+ },
+ {
+ "epoch": 91.64279475982532,
+ "grad_norm": 0.014401616528630257,
+ "learning_rate": 0.0006,
+ "loss": 4.125172138214111,
+ "step": 6598
+ },
+ {
+ "epoch": 91.65676855895197,
+ "grad_norm": 0.013945513404905796,
+ "learning_rate": 0.0006,
+ "loss": 4.146623134613037,
+ "step": 6599
+ },
+ {
+ "epoch": 91.6707423580786,
+ "grad_norm": 0.013516047969460487,
+ "learning_rate": 0.0006,
+ "loss": 4.160526275634766,
+ "step": 6600
+ },
+ {
+ "epoch": 91.68471615720524,
+ "grad_norm": 0.013262378051877022,
+ "learning_rate": 0.0006,
+ "loss": 4.074923038482666,
+ "step": 6601
+ },
+ {
+ "epoch": 91.69868995633188,
+ "grad_norm": 0.014365669339895248,
+ "learning_rate": 0.0006,
+ "loss": 4.118094444274902,
+ "step": 6602
+ },
+ {
+ "epoch": 91.71266375545852,
+ "grad_norm": 0.016713734716176987,
+ "learning_rate": 0.0006,
+ "loss": 4.021799087524414,
+ "step": 6603
+ },
+ {
+ "epoch": 91.72663755458515,
+ "grad_norm": 0.01988680474460125,
+ "learning_rate": 0.0006,
+ "loss": 4.238436222076416,
+ "step": 6604
+ },
+ {
+ "epoch": 91.74061135371178,
+ "grad_norm": 0.022749733179807663,
+ "learning_rate": 0.0006,
+ "loss": 4.010470390319824,
+ "step": 6605
+ },
+ {
+ "epoch": 91.75458515283843,
+ "grad_norm": 0.022533461451530457,
+ "learning_rate": 0.0006,
+ "loss": 4.152201175689697,
+ "step": 6606
+ },
+ {
+ "epoch": 91.76855895196506,
+ "grad_norm": 0.02021082676947117,
+ "learning_rate": 0.0006,
+ "loss": 4.120572090148926,
+ "step": 6607
+ },
+ {
+ "epoch": 91.78253275109171,
+ "grad_norm": 0.016298197209835052,
+ "learning_rate": 0.0006,
+ "loss": 4.115815162658691,
+ "step": 6608
+ },
+ {
+ "epoch": 91.79650655021834,
+ "grad_norm": 0.017562786117196083,
+ "learning_rate": 0.0006,
+ "loss": 4.109724044799805,
+ "step": 6609
+ },
+ {
+ "epoch": 91.81048034934499,
+ "grad_norm": 0.018153153359889984,
+ "learning_rate": 0.0006,
+ "loss": 4.143759727478027,
+ "step": 6610
+ },
+ {
+ "epoch": 91.82445414847162,
+ "grad_norm": 0.014762306585907936,
+ "learning_rate": 0.0006,
+ "loss": 4.25596809387207,
+ "step": 6611
+ },
+ {
+ "epoch": 91.83842794759825,
+ "grad_norm": 0.015074343420565128,
+ "learning_rate": 0.0006,
+ "loss": 4.207723140716553,
+ "step": 6612
+ },
+ {
+ "epoch": 91.8524017467249,
+ "grad_norm": 0.015028579160571098,
+ "learning_rate": 0.0006,
+ "loss": 4.01950216293335,
+ "step": 6613
+ },
+ {
+ "epoch": 91.86637554585153,
+ "grad_norm": 0.013996795751154423,
+ "learning_rate": 0.0006,
+ "loss": 4.132997989654541,
+ "step": 6614
+ },
+ {
+ "epoch": 91.88034934497817,
+ "grad_norm": 0.014978236518800259,
+ "learning_rate": 0.0006,
+ "loss": 4.132402420043945,
+ "step": 6615
+ },
+ {
+ "epoch": 91.8943231441048,
+ "grad_norm": 0.013111517764627934,
+ "learning_rate": 0.0006,
+ "loss": 4.1223649978637695,
+ "step": 6616
+ },
+ {
+ "epoch": 91.90829694323143,
+ "grad_norm": 0.013872234150767326,
+ "learning_rate": 0.0006,
+ "loss": 4.096590042114258,
+ "step": 6617
+ },
+ {
+ "epoch": 91.92227074235808,
+ "grad_norm": 0.013642470352351665,
+ "learning_rate": 0.0006,
+ "loss": 4.236227035522461,
+ "step": 6618
+ },
+ {
+ "epoch": 91.93624454148471,
+ "grad_norm": 0.014684873633086681,
+ "learning_rate": 0.0006,
+ "loss": 4.037961959838867,
+ "step": 6619
+ },
+ {
+ "epoch": 91.95021834061136,
+ "grad_norm": 0.014498206786811352,
+ "learning_rate": 0.0006,
+ "loss": 4.043325424194336,
+ "step": 6620
+ },
+ {
+ "epoch": 91.96419213973799,
+ "grad_norm": 0.012768656946718693,
+ "learning_rate": 0.0006,
+ "loss": 4.138803482055664,
+ "step": 6621
+ },
+ {
+ "epoch": 91.97816593886463,
+ "grad_norm": 0.012155905365943909,
+ "learning_rate": 0.0006,
+ "loss": 4.163424968719482,
+ "step": 6622
+ },
+ {
+ "epoch": 91.99213973799127,
+ "grad_norm": 0.012584930285811424,
+ "learning_rate": 0.0006,
+ "loss": 4.086629867553711,
+ "step": 6623
+ },
+ {
+ "epoch": 92.0,
+ "grad_norm": 0.01578151248395443,
+ "learning_rate": 0.0006,
+ "loss": 4.108386993408203,
+ "step": 6624
+ },
+ {
+ "epoch": 92.0,
+ "eval_loss": 4.541166305541992,
+ "eval_runtime": 56.871,
+ "eval_samples_per_second": 42.939,
+ "eval_steps_per_second": 1.354,
+ "step": 6624
+ },
+ {
+ "epoch": 92.01397379912663,
+ "grad_norm": 0.016901545226573944,
+ "learning_rate": 0.0006,
+ "loss": 4.043694496154785,
+ "step": 6625
+ },
+ {
+ "epoch": 92.02794759825328,
+ "grad_norm": 0.016299039125442505,
+ "learning_rate": 0.0006,
+ "loss": 4.100504398345947,
+ "step": 6626
+ },
+ {
+ "epoch": 92.04192139737991,
+ "grad_norm": 0.013988777063786983,
+ "learning_rate": 0.0006,
+ "loss": 4.063943862915039,
+ "step": 6627
+ },
+ {
+ "epoch": 92.05589519650655,
+ "grad_norm": 0.01343096699565649,
+ "learning_rate": 0.0006,
+ "loss": 4.0321197509765625,
+ "step": 6628
+ },
+ {
+ "epoch": 92.06986899563319,
+ "grad_norm": 0.014343147166073322,
+ "learning_rate": 0.0006,
+ "loss": 4.095499038696289,
+ "step": 6629
+ },
+ {
+ "epoch": 92.08384279475983,
+ "grad_norm": 0.014989756047725677,
+ "learning_rate": 0.0006,
+ "loss": 4.122310638427734,
+ "step": 6630
+ },
+ {
+ "epoch": 92.09781659388646,
+ "grad_norm": 0.013929174281656742,
+ "learning_rate": 0.0006,
+ "loss": 4.02870512008667,
+ "step": 6631
+ },
+ {
+ "epoch": 92.1117903930131,
+ "grad_norm": 0.014139998704195023,
+ "learning_rate": 0.0006,
+ "loss": 4.180675506591797,
+ "step": 6632
+ },
+ {
+ "epoch": 92.12576419213974,
+ "grad_norm": 0.012920474633574486,
+ "learning_rate": 0.0006,
+ "loss": 4.111686706542969,
+ "step": 6633
+ },
+ {
+ "epoch": 92.13973799126637,
+ "grad_norm": 0.01291937381029129,
+ "learning_rate": 0.0006,
+ "loss": 3.958092212677002,
+ "step": 6634
+ },
+ {
+ "epoch": 92.15371179039302,
+ "grad_norm": 0.013250474818050861,
+ "learning_rate": 0.0006,
+ "loss": 4.075105667114258,
+ "step": 6635
+ },
+ {
+ "epoch": 92.16768558951965,
+ "grad_norm": 0.01321406476199627,
+ "learning_rate": 0.0006,
+ "loss": 4.072322368621826,
+ "step": 6636
+ },
+ {
+ "epoch": 92.18165938864628,
+ "grad_norm": 0.014002865180373192,
+ "learning_rate": 0.0006,
+ "loss": 4.061572074890137,
+ "step": 6637
+ },
+ {
+ "epoch": 92.19563318777293,
+ "grad_norm": 0.013876105658710003,
+ "learning_rate": 0.0006,
+ "loss": 4.099213600158691,
+ "step": 6638
+ },
+ {
+ "epoch": 92.20960698689956,
+ "grad_norm": 0.01464459951967001,
+ "learning_rate": 0.0006,
+ "loss": 4.1573638916015625,
+ "step": 6639
+ },
+ {
+ "epoch": 92.2235807860262,
+ "grad_norm": 0.015354706905782223,
+ "learning_rate": 0.0006,
+ "loss": 4.078714370727539,
+ "step": 6640
+ },
+ {
+ "epoch": 92.23755458515284,
+ "grad_norm": 0.015265689231455326,
+ "learning_rate": 0.0006,
+ "loss": 4.036652565002441,
+ "step": 6641
+ },
+ {
+ "epoch": 92.25152838427948,
+ "grad_norm": 0.016140855848789215,
+ "learning_rate": 0.0006,
+ "loss": 4.120541095733643,
+ "step": 6642
+ },
+ {
+ "epoch": 92.26550218340611,
+ "grad_norm": 0.017044052481651306,
+ "learning_rate": 0.0006,
+ "loss": 4.015849590301514,
+ "step": 6643
+ },
+ {
+ "epoch": 92.27947598253274,
+ "grad_norm": 0.01469327136874199,
+ "learning_rate": 0.0006,
+ "loss": 4.137286186218262,
+ "step": 6644
+ },
+ {
+ "epoch": 92.29344978165939,
+ "grad_norm": 0.014596942812204361,
+ "learning_rate": 0.0006,
+ "loss": 4.11309814453125,
+ "step": 6645
+ },
+ {
+ "epoch": 92.30742358078602,
+ "grad_norm": 0.014671559445559978,
+ "learning_rate": 0.0006,
+ "loss": 4.087950229644775,
+ "step": 6646
+ },
+ {
+ "epoch": 92.32139737991267,
+ "grad_norm": 0.015040507540106773,
+ "learning_rate": 0.0006,
+ "loss": 3.967053174972534,
+ "step": 6647
+ },
+ {
+ "epoch": 92.3353711790393,
+ "grad_norm": 0.01417640782892704,
+ "learning_rate": 0.0006,
+ "loss": 3.990346908569336,
+ "step": 6648
+ },
+ {
+ "epoch": 92.34934497816595,
+ "grad_norm": 0.013855187222361565,
+ "learning_rate": 0.0006,
+ "loss": 3.965132474899292,
+ "step": 6649
+ },
+ {
+ "epoch": 92.36331877729258,
+ "grad_norm": 0.014291216619312763,
+ "learning_rate": 0.0006,
+ "loss": 4.161717891693115,
+ "step": 6650
+ },
+ {
+ "epoch": 92.37729257641921,
+ "grad_norm": 0.014148853719234467,
+ "learning_rate": 0.0006,
+ "loss": 4.019951820373535,
+ "step": 6651
+ },
+ {
+ "epoch": 92.39126637554585,
+ "grad_norm": 0.01414303295314312,
+ "learning_rate": 0.0006,
+ "loss": 4.125976085662842,
+ "step": 6652
+ },
+ {
+ "epoch": 92.40524017467249,
+ "grad_norm": 0.01358176488429308,
+ "learning_rate": 0.0006,
+ "loss": 4.088022232055664,
+ "step": 6653
+ },
+ {
+ "epoch": 92.41921397379913,
+ "grad_norm": 0.014174354262650013,
+ "learning_rate": 0.0006,
+ "loss": 4.1137518882751465,
+ "step": 6654
+ },
+ {
+ "epoch": 92.43318777292576,
+ "grad_norm": 0.013756301254034042,
+ "learning_rate": 0.0006,
+ "loss": 4.0088725090026855,
+ "step": 6655
+ },
+ {
+ "epoch": 92.44716157205241,
+ "grad_norm": 0.014257770963013172,
+ "learning_rate": 0.0006,
+ "loss": 4.05248498916626,
+ "step": 6656
+ },
+ {
+ "epoch": 92.46113537117904,
+ "grad_norm": 0.015319934114813805,
+ "learning_rate": 0.0006,
+ "loss": 4.1375732421875,
+ "step": 6657
+ },
+ {
+ "epoch": 92.47510917030567,
+ "grad_norm": 0.014655457809567451,
+ "learning_rate": 0.0006,
+ "loss": 4.172857284545898,
+ "step": 6658
+ },
+ {
+ "epoch": 92.48908296943232,
+ "grad_norm": 0.014981928281486034,
+ "learning_rate": 0.0006,
+ "loss": 4.064846038818359,
+ "step": 6659
+ },
+ {
+ "epoch": 92.50305676855895,
+ "grad_norm": 0.015197387896478176,
+ "learning_rate": 0.0006,
+ "loss": 4.119397163391113,
+ "step": 6660
+ },
+ {
+ "epoch": 92.5170305676856,
+ "grad_norm": 0.016060935333371162,
+ "learning_rate": 0.0006,
+ "loss": 4.09212064743042,
+ "step": 6661
+ },
+ {
+ "epoch": 92.53100436681223,
+ "grad_norm": 0.014884199015796185,
+ "learning_rate": 0.0006,
+ "loss": 4.027483940124512,
+ "step": 6662
+ },
+ {
+ "epoch": 92.54497816593886,
+ "grad_norm": 0.01562194712460041,
+ "learning_rate": 0.0006,
+ "loss": 3.9722909927368164,
+ "step": 6663
+ },
+ {
+ "epoch": 92.5589519650655,
+ "grad_norm": 0.014801043085753918,
+ "learning_rate": 0.0006,
+ "loss": 4.048130989074707,
+ "step": 6664
+ },
+ {
+ "epoch": 92.57292576419214,
+ "grad_norm": 0.013954494148492813,
+ "learning_rate": 0.0006,
+ "loss": 4.050351619720459,
+ "step": 6665
+ },
+ {
+ "epoch": 92.58689956331878,
+ "grad_norm": 0.013292429968714714,
+ "learning_rate": 0.0006,
+ "loss": 4.245860576629639,
+ "step": 6666
+ },
+ {
+ "epoch": 92.60087336244541,
+ "grad_norm": 0.015737095847725868,
+ "learning_rate": 0.0006,
+ "loss": 4.154323101043701,
+ "step": 6667
+ },
+ {
+ "epoch": 92.61484716157206,
+ "grad_norm": 0.018242565914988518,
+ "learning_rate": 0.0006,
+ "loss": 4.118710517883301,
+ "step": 6668
+ },
+ {
+ "epoch": 92.62882096069869,
+ "grad_norm": 0.017666636034846306,
+ "learning_rate": 0.0006,
+ "loss": 4.1013875007629395,
+ "step": 6669
+ },
+ {
+ "epoch": 92.64279475982532,
+ "grad_norm": 0.015017223544418812,
+ "learning_rate": 0.0006,
+ "loss": 4.084001064300537,
+ "step": 6670
+ },
+ {
+ "epoch": 92.65676855895197,
+ "grad_norm": 0.01359238475561142,
+ "learning_rate": 0.0006,
+ "loss": 4.091456413269043,
+ "step": 6671
+ },
+ {
+ "epoch": 92.6707423580786,
+ "grad_norm": 0.014524881727993488,
+ "learning_rate": 0.0006,
+ "loss": 4.121245384216309,
+ "step": 6672
+ },
+ {
+ "epoch": 92.68471615720524,
+ "grad_norm": 0.016214480623602867,
+ "learning_rate": 0.0006,
+ "loss": 4.133184432983398,
+ "step": 6673
+ },
+ {
+ "epoch": 92.69868995633188,
+ "grad_norm": 0.0176643505692482,
+ "learning_rate": 0.0006,
+ "loss": 4.142827987670898,
+ "step": 6674
+ },
+ {
+ "epoch": 92.71266375545852,
+ "grad_norm": 0.017004093155264854,
+ "learning_rate": 0.0006,
+ "loss": 4.067666530609131,
+ "step": 6675
+ },
+ {
+ "epoch": 92.72663755458515,
+ "grad_norm": 0.014077397994697094,
+ "learning_rate": 0.0006,
+ "loss": 4.076465129852295,
+ "step": 6676
+ },
+ {
+ "epoch": 92.74061135371178,
+ "grad_norm": 0.015905627980828285,
+ "learning_rate": 0.0006,
+ "loss": 4.083726406097412,
+ "step": 6677
+ },
+ {
+ "epoch": 92.75458515283843,
+ "grad_norm": 0.016467567533254623,
+ "learning_rate": 0.0006,
+ "loss": 4.241217136383057,
+ "step": 6678
+ },
+ {
+ "epoch": 92.76855895196506,
+ "grad_norm": 0.01541806198656559,
+ "learning_rate": 0.0006,
+ "loss": 4.05365514755249,
+ "step": 6679
+ },
+ {
+ "epoch": 92.78253275109171,
+ "grad_norm": 0.015028242021799088,
+ "learning_rate": 0.0006,
+ "loss": 4.029950141906738,
+ "step": 6680
+ },
+ {
+ "epoch": 92.79650655021834,
+ "grad_norm": 0.014366233721375465,
+ "learning_rate": 0.0006,
+ "loss": 4.181772232055664,
+ "step": 6681
+ },
+ {
+ "epoch": 92.81048034934499,
+ "grad_norm": 0.014313179068267345,
+ "learning_rate": 0.0006,
+ "loss": 4.152658462524414,
+ "step": 6682
+ },
+ {
+ "epoch": 92.82445414847162,
+ "grad_norm": 0.016504302620887756,
+ "learning_rate": 0.0006,
+ "loss": 4.080532073974609,
+ "step": 6683
+ },
+ {
+ "epoch": 92.83842794759825,
+ "grad_norm": 0.01685677096247673,
+ "learning_rate": 0.0006,
+ "loss": 4.194268226623535,
+ "step": 6684
+ },
+ {
+ "epoch": 92.8524017467249,
+ "grad_norm": 0.018684150651097298,
+ "learning_rate": 0.0006,
+ "loss": 4.218497276306152,
+ "step": 6685
+ },
+ {
+ "epoch": 92.86637554585153,
+ "grad_norm": 0.01880732551217079,
+ "learning_rate": 0.0006,
+ "loss": 4.133942604064941,
+ "step": 6686
+ },
+ {
+ "epoch": 92.88034934497817,
+ "grad_norm": 0.016504917293787003,
+ "learning_rate": 0.0006,
+ "loss": 4.139945030212402,
+ "step": 6687
+ },
+ {
+ "epoch": 92.8943231441048,
+ "grad_norm": 0.016495356336236,
+ "learning_rate": 0.0006,
+ "loss": 4.098918914794922,
+ "step": 6688
+ },
+ {
+ "epoch": 92.90829694323143,
+ "grad_norm": 0.0170587245374918,
+ "learning_rate": 0.0006,
+ "loss": 4.025818824768066,
+ "step": 6689
+ },
+ {
+ "epoch": 92.92227074235808,
+ "grad_norm": 0.01718820631504059,
+ "learning_rate": 0.0006,
+ "loss": 4.125217437744141,
+ "step": 6690
+ },
+ {
+ "epoch": 92.93624454148471,
+ "grad_norm": 0.015789354220032692,
+ "learning_rate": 0.0006,
+ "loss": 4.150068283081055,
+ "step": 6691
+ },
+ {
+ "epoch": 92.95021834061136,
+ "grad_norm": 0.01587480492889881,
+ "learning_rate": 0.0006,
+ "loss": 4.194794654846191,
+ "step": 6692
+ },
+ {
+ "epoch": 92.96419213973799,
+ "grad_norm": 0.015669209882616997,
+ "learning_rate": 0.0006,
+ "loss": 4.0550336837768555,
+ "step": 6693
+ },
+ {
+ "epoch": 92.97816593886463,
+ "grad_norm": 0.01519355084747076,
+ "learning_rate": 0.0006,
+ "loss": 4.114232063293457,
+ "step": 6694
+ },
+ {
+ "epoch": 92.99213973799127,
+ "grad_norm": 0.014596261084079742,
+ "learning_rate": 0.0006,
+ "loss": 3.992065906524658,
+ "step": 6695
+ },
+ {
+ "epoch": 93.0,
+ "grad_norm": 0.016378438100218773,
+ "learning_rate": 0.0006,
+ "loss": 4.093693733215332,
+ "step": 6696
+ },
+ {
+ "epoch": 93.0,
+ "eval_loss": 4.541905879974365,
+ "eval_runtime": 57.0177,
+ "eval_samples_per_second": 42.829,
+ "eval_steps_per_second": 1.35,
+ "step": 6696
+ },
+ {
+ "epoch": 93.01397379912663,
+ "grad_norm": 0.01593143492937088,
+ "learning_rate": 0.0006,
+ "loss": 4.052624225616455,
+ "step": 6697
+ },
+ {
+ "epoch": 93.02794759825328,
+ "grad_norm": 0.016086837276816368,
+ "learning_rate": 0.0006,
+ "loss": 4.130306243896484,
+ "step": 6698
+ },
+ {
+ "epoch": 93.04192139737991,
+ "grad_norm": 0.016004323959350586,
+ "learning_rate": 0.0006,
+ "loss": 3.965893268585205,
+ "step": 6699
+ },
+ {
+ "epoch": 93.05589519650655,
+ "grad_norm": 0.017667284235358238,
+ "learning_rate": 0.0006,
+ "loss": 4.095634460449219,
+ "step": 6700
+ },
+ {
+ "epoch": 93.06986899563319,
+ "grad_norm": 0.01949942484498024,
+ "learning_rate": 0.0006,
+ "loss": 3.966041326522827,
+ "step": 6701
+ },
+ {
+ "epoch": 93.08384279475983,
+ "grad_norm": 0.021898971870541573,
+ "learning_rate": 0.0006,
+ "loss": 4.010378837585449,
+ "step": 6702
+ },
+ {
+ "epoch": 93.09781659388646,
+ "grad_norm": 0.019985264167189598,
+ "learning_rate": 0.0006,
+ "loss": 4.061577320098877,
+ "step": 6703
+ },
+ {
+ "epoch": 93.1117903930131,
+ "grad_norm": 0.016470855101943016,
+ "learning_rate": 0.0006,
+ "loss": 4.0601677894592285,
+ "step": 6704
+ },
+ {
+ "epoch": 93.12576419213974,
+ "grad_norm": 0.015600379556417465,
+ "learning_rate": 0.0006,
+ "loss": 4.0860700607299805,
+ "step": 6705
+ },
+ {
+ "epoch": 93.13973799126637,
+ "grad_norm": 0.017514333128929138,
+ "learning_rate": 0.0006,
+ "loss": 4.240778923034668,
+ "step": 6706
+ },
+ {
+ "epoch": 93.15371179039302,
+ "grad_norm": 0.01780967228114605,
+ "learning_rate": 0.0006,
+ "loss": 4.076499938964844,
+ "step": 6707
+ },
+ {
+ "epoch": 93.16768558951965,
+ "grad_norm": 0.016370438039302826,
+ "learning_rate": 0.0006,
+ "loss": 4.0677618980407715,
+ "step": 6708
+ },
+ {
+ "epoch": 93.18165938864628,
+ "grad_norm": 0.015205491334199905,
+ "learning_rate": 0.0006,
+ "loss": 4.004945755004883,
+ "step": 6709
+ },
+ {
+ "epoch": 93.19563318777293,
+ "grad_norm": 0.01537297572940588,
+ "learning_rate": 0.0006,
+ "loss": 4.121147155761719,
+ "step": 6710
+ },
+ {
+ "epoch": 93.20960698689956,
+ "grad_norm": 0.017343711107969284,
+ "learning_rate": 0.0006,
+ "loss": 4.103215217590332,
+ "step": 6711
+ },
+ {
+ "epoch": 93.2235807860262,
+ "grad_norm": 0.01618131436407566,
+ "learning_rate": 0.0006,
+ "loss": 4.083741664886475,
+ "step": 6712
+ },
+ {
+ "epoch": 93.23755458515284,
+ "grad_norm": 0.013340235687792301,
+ "learning_rate": 0.0006,
+ "loss": 4.055032730102539,
+ "step": 6713
+ },
+ {
+ "epoch": 93.25152838427948,
+ "grad_norm": 0.014084120281040668,
+ "learning_rate": 0.0006,
+ "loss": 4.005610942840576,
+ "step": 6714
+ },
+ {
+ "epoch": 93.26550218340611,
+ "grad_norm": 0.016104530543088913,
+ "learning_rate": 0.0006,
+ "loss": 4.091764450073242,
+ "step": 6715
+ },
+ {
+ "epoch": 93.27947598253274,
+ "grad_norm": 0.015232563950121403,
+ "learning_rate": 0.0006,
+ "loss": 4.105735778808594,
+ "step": 6716
+ },
+ {
+ "epoch": 93.29344978165939,
+ "grad_norm": 0.014003508724272251,
+ "learning_rate": 0.0006,
+ "loss": 4.05783224105835,
+ "step": 6717
+ },
+ {
+ "epoch": 93.30742358078602,
+ "grad_norm": 0.01614833064377308,
+ "learning_rate": 0.0006,
+ "loss": 4.124567031860352,
+ "step": 6718
+ },
+ {
+ "epoch": 93.32139737991267,
+ "grad_norm": 0.015010425820946693,
+ "learning_rate": 0.0006,
+ "loss": 4.07656192779541,
+ "step": 6719
+ },
+ {
+ "epoch": 93.3353711790393,
+ "grad_norm": 0.01444012951105833,
+ "learning_rate": 0.0006,
+ "loss": 4.122415542602539,
+ "step": 6720
+ },
+ {
+ "epoch": 93.34934497816595,
+ "grad_norm": 0.015715785324573517,
+ "learning_rate": 0.0006,
+ "loss": 4.070283889770508,
+ "step": 6721
+ },
+ {
+ "epoch": 93.36331877729258,
+ "grad_norm": 0.014976450242102146,
+ "learning_rate": 0.0006,
+ "loss": 4.087043762207031,
+ "step": 6722
+ },
+ {
+ "epoch": 93.37729257641921,
+ "grad_norm": 0.014868560247123241,
+ "learning_rate": 0.0006,
+ "loss": 3.9671802520751953,
+ "step": 6723
+ },
+ {
+ "epoch": 93.39126637554585,
+ "grad_norm": 0.013518013060092926,
+ "learning_rate": 0.0006,
+ "loss": 4.136781692504883,
+ "step": 6724
+ },
+ {
+ "epoch": 93.40524017467249,
+ "grad_norm": 0.013498986139893532,
+ "learning_rate": 0.0006,
+ "loss": 3.982309103012085,
+ "step": 6725
+ },
+ {
+ "epoch": 93.41921397379913,
+ "grad_norm": 0.013707910664379597,
+ "learning_rate": 0.0006,
+ "loss": 4.181885719299316,
+ "step": 6726
+ },
+ {
+ "epoch": 93.43318777292576,
+ "grad_norm": 0.013913207687437534,
+ "learning_rate": 0.0006,
+ "loss": 3.980070114135742,
+ "step": 6727
+ },
+ {
+ "epoch": 93.44716157205241,
+ "grad_norm": 0.013169893994927406,
+ "learning_rate": 0.0006,
+ "loss": 4.023380279541016,
+ "step": 6728
+ },
+ {
+ "epoch": 93.46113537117904,
+ "grad_norm": 0.014451324008405209,
+ "learning_rate": 0.0006,
+ "loss": 4.038372039794922,
+ "step": 6729
+ },
+ {
+ "epoch": 93.47510917030567,
+ "grad_norm": 0.014401541091501713,
+ "learning_rate": 0.0006,
+ "loss": 4.069009304046631,
+ "step": 6730
+ },
+ {
+ "epoch": 93.48908296943232,
+ "grad_norm": 0.013971361331641674,
+ "learning_rate": 0.0006,
+ "loss": 4.105923652648926,
+ "step": 6731
+ },
+ {
+ "epoch": 93.50305676855895,
+ "grad_norm": 0.013192536309361458,
+ "learning_rate": 0.0006,
+ "loss": 4.07828950881958,
+ "step": 6732
+ },
+ {
+ "epoch": 93.5170305676856,
+ "grad_norm": 0.013581741601228714,
+ "learning_rate": 0.0006,
+ "loss": 4.079203128814697,
+ "step": 6733
+ },
+ {
+ "epoch": 93.53100436681223,
+ "grad_norm": 0.01508194301277399,
+ "learning_rate": 0.0006,
+ "loss": 4.084749221801758,
+ "step": 6734
+ },
+ {
+ "epoch": 93.54497816593886,
+ "grad_norm": 0.0156723503023386,
+ "learning_rate": 0.0006,
+ "loss": 4.070216178894043,
+ "step": 6735
+ },
+ {
+ "epoch": 93.5589519650655,
+ "grad_norm": 0.014996538870036602,
+ "learning_rate": 0.0006,
+ "loss": 4.111318588256836,
+ "step": 6736
+ },
+ {
+ "epoch": 93.57292576419214,
+ "grad_norm": 0.012815269641578197,
+ "learning_rate": 0.0006,
+ "loss": 4.130414962768555,
+ "step": 6737
+ },
+ {
+ "epoch": 93.58689956331878,
+ "grad_norm": 0.012887579388916492,
+ "learning_rate": 0.0006,
+ "loss": 4.128446102142334,
+ "step": 6738
+ },
+ {
+ "epoch": 93.60087336244541,
+ "grad_norm": 0.013632671907544136,
+ "learning_rate": 0.0006,
+ "loss": 4.20977258682251,
+ "step": 6739
+ },
+ {
+ "epoch": 93.61484716157206,
+ "grad_norm": 0.014318895526230335,
+ "learning_rate": 0.0006,
+ "loss": 4.005739212036133,
+ "step": 6740
+ },
+ {
+ "epoch": 93.62882096069869,
+ "grad_norm": 0.014160127379000187,
+ "learning_rate": 0.0006,
+ "loss": 4.122544288635254,
+ "step": 6741
+ },
+ {
+ "epoch": 93.64279475982532,
+ "grad_norm": 0.01308423187583685,
+ "learning_rate": 0.0006,
+ "loss": 4.108774662017822,
+ "step": 6742
+ },
+ {
+ "epoch": 93.65676855895197,
+ "grad_norm": 0.014293737709522247,
+ "learning_rate": 0.0006,
+ "loss": 3.983156681060791,
+ "step": 6743
+ },
+ {
+ "epoch": 93.6707423580786,
+ "grad_norm": 0.015155135653913021,
+ "learning_rate": 0.0006,
+ "loss": 4.006117820739746,
+ "step": 6744
+ },
+ {
+ "epoch": 93.68471615720524,
+ "grad_norm": 0.014391839504241943,
+ "learning_rate": 0.0006,
+ "loss": 4.1658935546875,
+ "step": 6745
+ },
+ {
+ "epoch": 93.69868995633188,
+ "grad_norm": 0.014924845658242702,
+ "learning_rate": 0.0006,
+ "loss": 4.090868949890137,
+ "step": 6746
+ },
+ {
+ "epoch": 93.71266375545852,
+ "grad_norm": 0.014786627143621445,
+ "learning_rate": 0.0006,
+ "loss": 4.121716499328613,
+ "step": 6747
+ },
+ {
+ "epoch": 93.72663755458515,
+ "grad_norm": 0.016948824748396873,
+ "learning_rate": 0.0006,
+ "loss": 4.213179111480713,
+ "step": 6748
+ },
+ {
+ "epoch": 93.74061135371178,
+ "grad_norm": 0.01732620596885681,
+ "learning_rate": 0.0006,
+ "loss": 4.079215049743652,
+ "step": 6749
+ },
+ {
+ "epoch": 93.75458515283843,
+ "grad_norm": 0.016298124566674232,
+ "learning_rate": 0.0006,
+ "loss": 4.096126079559326,
+ "step": 6750
+ },
+ {
+ "epoch": 93.76855895196506,
+ "grad_norm": 0.015744732692837715,
+ "learning_rate": 0.0006,
+ "loss": 4.046300411224365,
+ "step": 6751
+ },
+ {
+ "epoch": 93.78253275109171,
+ "grad_norm": 0.016330812126398087,
+ "learning_rate": 0.0006,
+ "loss": 4.098011016845703,
+ "step": 6752
+ },
+ {
+ "epoch": 93.79650655021834,
+ "grad_norm": 0.01710600219666958,
+ "learning_rate": 0.0006,
+ "loss": 4.031402587890625,
+ "step": 6753
+ },
+ {
+ "epoch": 93.81048034934499,
+ "grad_norm": 0.01561805047094822,
+ "learning_rate": 0.0006,
+ "loss": 4.023388862609863,
+ "step": 6754
+ },
+ {
+ "epoch": 93.82445414847162,
+ "grad_norm": 0.013637243770062923,
+ "learning_rate": 0.0006,
+ "loss": 4.143238067626953,
+ "step": 6755
+ },
+ {
+ "epoch": 93.83842794759825,
+ "grad_norm": 0.013088075444102287,
+ "learning_rate": 0.0006,
+ "loss": 3.9083704948425293,
+ "step": 6756
+ },
+ {
+ "epoch": 93.8524017467249,
+ "grad_norm": 0.015893537551164627,
+ "learning_rate": 0.0006,
+ "loss": 4.102263450622559,
+ "step": 6757
+ },
+ {
+ "epoch": 93.86637554585153,
+ "grad_norm": 0.015213928185403347,
+ "learning_rate": 0.0006,
+ "loss": 4.111978530883789,
+ "step": 6758
+ },
+ {
+ "epoch": 93.88034934497817,
+ "grad_norm": 0.014543982222676277,
+ "learning_rate": 0.0006,
+ "loss": 4.05889892578125,
+ "step": 6759
+ },
+ {
+ "epoch": 93.8943231441048,
+ "grad_norm": 0.015104752965271473,
+ "learning_rate": 0.0006,
+ "loss": 4.136166572570801,
+ "step": 6760
+ },
+ {
+ "epoch": 93.90829694323143,
+ "grad_norm": 0.01505881268531084,
+ "learning_rate": 0.0006,
+ "loss": 4.061883449554443,
+ "step": 6761
+ },
+ {
+ "epoch": 93.92227074235808,
+ "grad_norm": 0.015248455107212067,
+ "learning_rate": 0.0006,
+ "loss": 4.109863758087158,
+ "step": 6762
+ },
+ {
+ "epoch": 93.93624454148471,
+ "grad_norm": 0.013651185669004917,
+ "learning_rate": 0.0006,
+ "loss": 4.061549663543701,
+ "step": 6763
+ },
+ {
+ "epoch": 93.95021834061136,
+ "grad_norm": 0.013806282542645931,
+ "learning_rate": 0.0006,
+ "loss": 4.205403804779053,
+ "step": 6764
+ },
+ {
+ "epoch": 93.96419213973799,
+ "grad_norm": 0.01538451761007309,
+ "learning_rate": 0.0006,
+ "loss": 4.14816951751709,
+ "step": 6765
+ },
+ {
+ "epoch": 93.97816593886463,
+ "grad_norm": 0.01602427288889885,
+ "learning_rate": 0.0006,
+ "loss": 4.000303268432617,
+ "step": 6766
+ },
+ {
+ "epoch": 93.99213973799127,
+ "grad_norm": 0.014558793976902962,
+ "learning_rate": 0.0006,
+ "loss": 4.096562385559082,
+ "step": 6767
+ },
+ {
+ "epoch": 94.0,
+ "grad_norm": 0.01613018475472927,
+ "learning_rate": 0.0006,
+ "loss": 4.262835502624512,
+ "step": 6768
+ },
+ {
+ "epoch": 94.0,
+ "eval_loss": 4.532126426696777,
+ "eval_runtime": 56.7592,
+ "eval_samples_per_second": 43.024,
+ "eval_steps_per_second": 1.357,
+ "step": 6768
+ },
+ {
+ "epoch": 94.01397379912663,
+ "grad_norm": 0.01614535041153431,
+ "learning_rate": 0.0006,
+ "loss": 4.073431491851807,
+ "step": 6769
+ },
+ {
+ "epoch": 94.02794759825328,
+ "grad_norm": 0.015986818820238113,
+ "learning_rate": 0.0006,
+ "loss": 4.070001602172852,
+ "step": 6770
+ },
+ {
+ "epoch": 94.04192139737991,
+ "grad_norm": 0.015678714960813522,
+ "learning_rate": 0.0006,
+ "loss": 4.06920051574707,
+ "step": 6771
+ },
+ {
+ "epoch": 94.05589519650655,
+ "grad_norm": 0.016593851149082184,
+ "learning_rate": 0.0006,
+ "loss": 4.095780849456787,
+ "step": 6772
+ },
+ {
+ "epoch": 94.06986899563319,
+ "grad_norm": 0.01780948042869568,
+ "learning_rate": 0.0006,
+ "loss": 4.092916488647461,
+ "step": 6773
+ },
+ {
+ "epoch": 94.08384279475983,
+ "grad_norm": 0.01755528897047043,
+ "learning_rate": 0.0006,
+ "loss": 3.9833455085754395,
+ "step": 6774
+ },
+ {
+ "epoch": 94.09781659388646,
+ "grad_norm": 0.01708311215043068,
+ "learning_rate": 0.0006,
+ "loss": 4.053336143493652,
+ "step": 6775
+ },
+ {
+ "epoch": 94.1117903930131,
+ "grad_norm": 0.017362922430038452,
+ "learning_rate": 0.0006,
+ "loss": 4.014343738555908,
+ "step": 6776
+ },
+ {
+ "epoch": 94.12576419213974,
+ "grad_norm": 0.01824064739048481,
+ "learning_rate": 0.0006,
+ "loss": 4.1215691566467285,
+ "step": 6777
+ },
+ {
+ "epoch": 94.13973799126637,
+ "grad_norm": 0.017743906006217003,
+ "learning_rate": 0.0006,
+ "loss": 4.091215133666992,
+ "step": 6778
+ },
+ {
+ "epoch": 94.15371179039302,
+ "grad_norm": 0.016585713252425194,
+ "learning_rate": 0.0006,
+ "loss": 4.013662338256836,
+ "step": 6779
+ },
+ {
+ "epoch": 94.16768558951965,
+ "grad_norm": 0.017397213727235794,
+ "learning_rate": 0.0006,
+ "loss": 4.070602893829346,
+ "step": 6780
+ },
+ {
+ "epoch": 94.18165938864628,
+ "grad_norm": 0.01733497530221939,
+ "learning_rate": 0.0006,
+ "loss": 4.1304426193237305,
+ "step": 6781
+ },
+ {
+ "epoch": 94.19563318777293,
+ "grad_norm": 0.015215300023555756,
+ "learning_rate": 0.0006,
+ "loss": 3.965014696121216,
+ "step": 6782
+ },
+ {
+ "epoch": 94.20960698689956,
+ "grad_norm": 0.015908699482679367,
+ "learning_rate": 0.0006,
+ "loss": 3.924964427947998,
+ "step": 6783
+ },
+ {
+ "epoch": 94.2235807860262,
+ "grad_norm": 0.014044968411326408,
+ "learning_rate": 0.0006,
+ "loss": 4.08673095703125,
+ "step": 6784
+ },
+ {
+ "epoch": 94.23755458515284,
+ "grad_norm": 0.01375142764300108,
+ "learning_rate": 0.0006,
+ "loss": 4.011466979980469,
+ "step": 6785
+ },
+ {
+ "epoch": 94.25152838427948,
+ "grad_norm": 0.015539257787168026,
+ "learning_rate": 0.0006,
+ "loss": 4.09976863861084,
+ "step": 6786
+ },
+ {
+ "epoch": 94.26550218340611,
+ "grad_norm": 0.014303367584943771,
+ "learning_rate": 0.0006,
+ "loss": 4.108369827270508,
+ "step": 6787
+ },
+ {
+ "epoch": 94.27947598253274,
+ "grad_norm": 0.01281928364187479,
+ "learning_rate": 0.0006,
+ "loss": 4.073737144470215,
+ "step": 6788
+ },
+ {
+ "epoch": 94.29344978165939,
+ "grad_norm": 0.013859959319233894,
+ "learning_rate": 0.0006,
+ "loss": 4.054823875427246,
+ "step": 6789
+ },
+ {
+ "epoch": 94.30742358078602,
+ "grad_norm": 0.013326860964298248,
+ "learning_rate": 0.0006,
+ "loss": 4.133251190185547,
+ "step": 6790
+ },
+ {
+ "epoch": 94.32139737991267,
+ "grad_norm": 0.013923076912760735,
+ "learning_rate": 0.0006,
+ "loss": 4.190706729888916,
+ "step": 6791
+ },
+ {
+ "epoch": 94.3353711790393,
+ "grad_norm": 0.01420847699046135,
+ "learning_rate": 0.0006,
+ "loss": 3.9336605072021484,
+ "step": 6792
+ },
+ {
+ "epoch": 94.34934497816595,
+ "grad_norm": 0.014607555232942104,
+ "learning_rate": 0.0006,
+ "loss": 4.128399848937988,
+ "step": 6793
+ },
+ {
+ "epoch": 94.36331877729258,
+ "grad_norm": 0.012862961739301682,
+ "learning_rate": 0.0006,
+ "loss": 4.0132527351379395,
+ "step": 6794
+ },
+ {
+ "epoch": 94.37729257641921,
+ "grad_norm": 0.012530217878520489,
+ "learning_rate": 0.0006,
+ "loss": 4.0384368896484375,
+ "step": 6795
+ },
+ {
+ "epoch": 94.39126637554585,
+ "grad_norm": 0.01286802627146244,
+ "learning_rate": 0.0006,
+ "loss": 3.95570707321167,
+ "step": 6796
+ },
+ {
+ "epoch": 94.40524017467249,
+ "grad_norm": 0.013450917787849903,
+ "learning_rate": 0.0006,
+ "loss": 4.016484260559082,
+ "step": 6797
+ },
+ {
+ "epoch": 94.41921397379913,
+ "grad_norm": 0.01393423043191433,
+ "learning_rate": 0.0006,
+ "loss": 4.124959468841553,
+ "step": 6798
+ },
+ {
+ "epoch": 94.43318777292576,
+ "grad_norm": 0.013077184557914734,
+ "learning_rate": 0.0006,
+ "loss": 4.029163837432861,
+ "step": 6799
+ },
+ {
+ "epoch": 94.44716157205241,
+ "grad_norm": 0.013577334582805634,
+ "learning_rate": 0.0006,
+ "loss": 4.22929573059082,
+ "step": 6800
+ },
+ {
+ "epoch": 94.46113537117904,
+ "grad_norm": 0.013930312357842922,
+ "learning_rate": 0.0006,
+ "loss": 4.024711608886719,
+ "step": 6801
+ },
+ {
+ "epoch": 94.47510917030567,
+ "grad_norm": 0.013746250420808792,
+ "learning_rate": 0.0006,
+ "loss": 3.9766428470611572,
+ "step": 6802
+ },
+ {
+ "epoch": 94.48908296943232,
+ "grad_norm": 0.013682983815670013,
+ "learning_rate": 0.0006,
+ "loss": 4.072981834411621,
+ "step": 6803
+ },
+ {
+ "epoch": 94.50305676855895,
+ "grad_norm": 0.01473842654377222,
+ "learning_rate": 0.0006,
+ "loss": 4.197623252868652,
+ "step": 6804
+ },
+ {
+ "epoch": 94.5170305676856,
+ "grad_norm": 0.014133758842945099,
+ "learning_rate": 0.0006,
+ "loss": 4.055549621582031,
+ "step": 6805
+ },
+ {
+ "epoch": 94.53100436681223,
+ "grad_norm": 0.0139645179733634,
+ "learning_rate": 0.0006,
+ "loss": 4.130135536193848,
+ "step": 6806
+ },
+ {
+ "epoch": 94.54497816593886,
+ "grad_norm": 0.013973365537822247,
+ "learning_rate": 0.0006,
+ "loss": 4.006230354309082,
+ "step": 6807
+ },
+ {
+ "epoch": 94.5589519650655,
+ "grad_norm": 0.014227126725018024,
+ "learning_rate": 0.0006,
+ "loss": 4.130459785461426,
+ "step": 6808
+ },
+ {
+ "epoch": 94.57292576419214,
+ "grad_norm": 0.013129506260156631,
+ "learning_rate": 0.0006,
+ "loss": 4.014974594116211,
+ "step": 6809
+ },
+ {
+ "epoch": 94.58689956331878,
+ "grad_norm": 0.013015891425311565,
+ "learning_rate": 0.0006,
+ "loss": 4.161370277404785,
+ "step": 6810
+ },
+ {
+ "epoch": 94.60087336244541,
+ "grad_norm": 0.013957147486507893,
+ "learning_rate": 0.0006,
+ "loss": 4.09519100189209,
+ "step": 6811
+ },
+ {
+ "epoch": 94.61484716157206,
+ "grad_norm": 0.015494572930037975,
+ "learning_rate": 0.0006,
+ "loss": 4.029949188232422,
+ "step": 6812
+ },
+ {
+ "epoch": 94.62882096069869,
+ "grad_norm": 0.016732869669795036,
+ "learning_rate": 0.0006,
+ "loss": 4.135669708251953,
+ "step": 6813
+ },
+ {
+ "epoch": 94.64279475982532,
+ "grad_norm": 0.01769322343170643,
+ "learning_rate": 0.0006,
+ "loss": 4.109872817993164,
+ "step": 6814
+ },
+ {
+ "epoch": 94.65676855895197,
+ "grad_norm": 0.015932537615299225,
+ "learning_rate": 0.0006,
+ "loss": 4.014595031738281,
+ "step": 6815
+ },
+ {
+ "epoch": 94.6707423580786,
+ "grad_norm": 0.015152990818023682,
+ "learning_rate": 0.0006,
+ "loss": 4.1983442306518555,
+ "step": 6816
+ },
+ {
+ "epoch": 94.68471615720524,
+ "grad_norm": 0.01590685173869133,
+ "learning_rate": 0.0006,
+ "loss": 4.119250774383545,
+ "step": 6817
+ },
+ {
+ "epoch": 94.69868995633188,
+ "grad_norm": 0.016018465161323547,
+ "learning_rate": 0.0006,
+ "loss": 4.048069953918457,
+ "step": 6818
+ },
+ {
+ "epoch": 94.71266375545852,
+ "grad_norm": 0.016621652990579605,
+ "learning_rate": 0.0006,
+ "loss": 4.043731689453125,
+ "step": 6819
+ },
+ {
+ "epoch": 94.72663755458515,
+ "grad_norm": 0.014535024762153625,
+ "learning_rate": 0.0006,
+ "loss": 4.1886820793151855,
+ "step": 6820
+ },
+ {
+ "epoch": 94.74061135371178,
+ "grad_norm": 0.014192107133567333,
+ "learning_rate": 0.0006,
+ "loss": 4.103938102722168,
+ "step": 6821
+ },
+ {
+ "epoch": 94.75458515283843,
+ "grad_norm": 0.0166022852063179,
+ "learning_rate": 0.0006,
+ "loss": 4.017818450927734,
+ "step": 6822
+ },
+ {
+ "epoch": 94.76855895196506,
+ "grad_norm": 0.016247155144810677,
+ "learning_rate": 0.0006,
+ "loss": 4.0038981437683105,
+ "step": 6823
+ },
+ {
+ "epoch": 94.78253275109171,
+ "grad_norm": 0.01593470573425293,
+ "learning_rate": 0.0006,
+ "loss": 4.043969631195068,
+ "step": 6824
+ },
+ {
+ "epoch": 94.79650655021834,
+ "grad_norm": 0.014334345236420631,
+ "learning_rate": 0.0006,
+ "loss": 4.097829341888428,
+ "step": 6825
+ },
+ {
+ "epoch": 94.81048034934499,
+ "grad_norm": 0.015581986866891384,
+ "learning_rate": 0.0006,
+ "loss": 4.148984432220459,
+ "step": 6826
+ },
+ {
+ "epoch": 94.82445414847162,
+ "grad_norm": 0.017519738525152206,
+ "learning_rate": 0.0006,
+ "loss": 4.245623588562012,
+ "step": 6827
+ },
+ {
+ "epoch": 94.83842794759825,
+ "grad_norm": 0.015155277214944363,
+ "learning_rate": 0.0006,
+ "loss": 4.031019687652588,
+ "step": 6828
+ },
+ {
+ "epoch": 94.8524017467249,
+ "grad_norm": 0.014969222247600555,
+ "learning_rate": 0.0006,
+ "loss": 4.036084175109863,
+ "step": 6829
+ },
+ {
+ "epoch": 94.86637554585153,
+ "grad_norm": 0.015987908467650414,
+ "learning_rate": 0.0006,
+ "loss": 4.08396053314209,
+ "step": 6830
+ },
+ {
+ "epoch": 94.88034934497817,
+ "grad_norm": 0.016324151307344437,
+ "learning_rate": 0.0006,
+ "loss": 4.036787033081055,
+ "step": 6831
+ },
+ {
+ "epoch": 94.8943231441048,
+ "grad_norm": 0.015957487747073174,
+ "learning_rate": 0.0006,
+ "loss": 4.147607803344727,
+ "step": 6832
+ },
+ {
+ "epoch": 94.90829694323143,
+ "grad_norm": 0.015286967158317566,
+ "learning_rate": 0.0006,
+ "loss": 4.184615135192871,
+ "step": 6833
+ },
+ {
+ "epoch": 94.92227074235808,
+ "grad_norm": 0.01338377594947815,
+ "learning_rate": 0.0006,
+ "loss": 4.014981269836426,
+ "step": 6834
+ },
+ {
+ "epoch": 94.93624454148471,
+ "grad_norm": 0.013997633941471577,
+ "learning_rate": 0.0006,
+ "loss": 4.021933078765869,
+ "step": 6835
+ },
+ {
+ "epoch": 94.95021834061136,
+ "grad_norm": 0.014572402462363243,
+ "learning_rate": 0.0006,
+ "loss": 4.0326948165893555,
+ "step": 6836
+ },
+ {
+ "epoch": 94.96419213973799,
+ "grad_norm": 0.014379586093127728,
+ "learning_rate": 0.0006,
+ "loss": 4.060697555541992,
+ "step": 6837
+ },
+ {
+ "epoch": 94.97816593886463,
+ "grad_norm": 0.013513042591512203,
+ "learning_rate": 0.0006,
+ "loss": 4.05123233795166,
+ "step": 6838
+ },
+ {
+ "epoch": 94.99213973799127,
+ "grad_norm": 0.012904108501970768,
+ "learning_rate": 0.0006,
+ "loss": 4.020852565765381,
+ "step": 6839
+ },
+ {
+ "epoch": 95.0,
+ "grad_norm": 0.014364209026098251,
+ "learning_rate": 0.0006,
+ "loss": 4.080837249755859,
+ "step": 6840
+ },
+ {
+ "epoch": 95.0,
+ "eval_loss": 4.608776569366455,
+ "eval_runtime": 56.6516,
+ "eval_samples_per_second": 43.106,
+ "eval_steps_per_second": 1.359,
+ "step": 6840
+ },
+ {
+ "epoch": 95.01397379912663,
+ "grad_norm": 0.01357510220259428,
+ "learning_rate": 0.0006,
+ "loss": 4.105312347412109,
+ "step": 6841
+ },
+ {
+ "epoch": 95.02794759825328,
+ "grad_norm": 0.014248257502913475,
+ "learning_rate": 0.0006,
+ "loss": 3.971667766571045,
+ "step": 6842
+ },
+ {
+ "epoch": 95.04192139737991,
+ "grad_norm": 0.014773483388125896,
+ "learning_rate": 0.0006,
+ "loss": 4.03553581237793,
+ "step": 6843
+ },
+ {
+ "epoch": 95.05589519650655,
+ "grad_norm": 0.014766111969947815,
+ "learning_rate": 0.0006,
+ "loss": 4.074707508087158,
+ "step": 6844
+ },
+ {
+ "epoch": 95.06986899563319,
+ "grad_norm": 0.014778018929064274,
+ "learning_rate": 0.0006,
+ "loss": 4.1083855628967285,
+ "step": 6845
+ },
+ {
+ "epoch": 95.08384279475983,
+ "grad_norm": 0.014466686174273491,
+ "learning_rate": 0.0006,
+ "loss": 4.054587364196777,
+ "step": 6846
+ },
+ {
+ "epoch": 95.09781659388646,
+ "grad_norm": 0.015195904299616814,
+ "learning_rate": 0.0006,
+ "loss": 4.142580032348633,
+ "step": 6847
+ },
+ {
+ "epoch": 95.1117903930131,
+ "grad_norm": 0.01530557032674551,
+ "learning_rate": 0.0006,
+ "loss": 4.062134265899658,
+ "step": 6848
+ },
+ {
+ "epoch": 95.12576419213974,
+ "grad_norm": 0.015203768387436867,
+ "learning_rate": 0.0006,
+ "loss": 4.082962989807129,
+ "step": 6849
+ },
+ {
+ "epoch": 95.13973799126637,
+ "grad_norm": 0.014906628057360649,
+ "learning_rate": 0.0006,
+ "loss": 4.033127784729004,
+ "step": 6850
+ },
+ {
+ "epoch": 95.15371179039302,
+ "grad_norm": 0.01553534809499979,
+ "learning_rate": 0.0006,
+ "loss": 4.051743030548096,
+ "step": 6851
+ },
+ {
+ "epoch": 95.16768558951965,
+ "grad_norm": 0.01607976108789444,
+ "learning_rate": 0.0006,
+ "loss": 4.038073539733887,
+ "step": 6852
+ },
+ {
+ "epoch": 95.18165938864628,
+ "grad_norm": 0.015776360407471657,
+ "learning_rate": 0.0006,
+ "loss": 4.013091087341309,
+ "step": 6853
+ },
+ {
+ "epoch": 95.19563318777293,
+ "grad_norm": 0.01623626798391342,
+ "learning_rate": 0.0006,
+ "loss": 4.12293815612793,
+ "step": 6854
+ },
+ {
+ "epoch": 95.20960698689956,
+ "grad_norm": 0.016887828707695007,
+ "learning_rate": 0.0006,
+ "loss": 4.120452404022217,
+ "step": 6855
+ },
+ {
+ "epoch": 95.2235807860262,
+ "grad_norm": 0.01561128068715334,
+ "learning_rate": 0.0006,
+ "loss": 4.071322441101074,
+ "step": 6856
+ },
+ {
+ "epoch": 95.23755458515284,
+ "grad_norm": 0.015679724514484406,
+ "learning_rate": 0.0006,
+ "loss": 4.065072059631348,
+ "step": 6857
+ },
+ {
+ "epoch": 95.25152838427948,
+ "grad_norm": 0.016506962478160858,
+ "learning_rate": 0.0006,
+ "loss": 4.0667924880981445,
+ "step": 6858
+ },
+ {
+ "epoch": 95.26550218340611,
+ "grad_norm": 0.015340839512646198,
+ "learning_rate": 0.0006,
+ "loss": 4.14047908782959,
+ "step": 6859
+ },
+ {
+ "epoch": 95.27947598253274,
+ "grad_norm": 0.013935413211584091,
+ "learning_rate": 0.0006,
+ "loss": 3.9688191413879395,
+ "step": 6860
+ },
+ {
+ "epoch": 95.29344978165939,
+ "grad_norm": 0.015223405323922634,
+ "learning_rate": 0.0006,
+ "loss": 4.101613521575928,
+ "step": 6861
+ },
+ {
+ "epoch": 95.30742358078602,
+ "grad_norm": 0.01562546379864216,
+ "learning_rate": 0.0006,
+ "loss": 4.128626823425293,
+ "step": 6862
+ },
+ {
+ "epoch": 95.32139737991267,
+ "grad_norm": 0.016159312799572945,
+ "learning_rate": 0.0006,
+ "loss": 4.080667018890381,
+ "step": 6863
+ },
+ {
+ "epoch": 95.3353711790393,
+ "grad_norm": 0.01724511943757534,
+ "learning_rate": 0.0006,
+ "loss": 4.08463191986084,
+ "step": 6864
+ },
+ {
+ "epoch": 95.34934497816595,
+ "grad_norm": 0.01592276245355606,
+ "learning_rate": 0.0006,
+ "loss": 4.127612113952637,
+ "step": 6865
+ },
+ {
+ "epoch": 95.36331877729258,
+ "grad_norm": 0.01744082197546959,
+ "learning_rate": 0.0006,
+ "loss": 4.034480094909668,
+ "step": 6866
+ },
+ {
+ "epoch": 95.37729257641921,
+ "grad_norm": 0.01774069108068943,
+ "learning_rate": 0.0006,
+ "loss": 4.091196537017822,
+ "step": 6867
+ },
+ {
+ "epoch": 95.39126637554585,
+ "grad_norm": 0.01653675176203251,
+ "learning_rate": 0.0006,
+ "loss": 4.032937526702881,
+ "step": 6868
+ },
+ {
+ "epoch": 95.40524017467249,
+ "grad_norm": 0.014933439902961254,
+ "learning_rate": 0.0006,
+ "loss": 3.9386508464813232,
+ "step": 6869
+ },
+ {
+ "epoch": 95.41921397379913,
+ "grad_norm": 0.014743257313966751,
+ "learning_rate": 0.0006,
+ "loss": 4.126462459564209,
+ "step": 6870
+ },
+ {
+ "epoch": 95.43318777292576,
+ "grad_norm": 0.014614643529057503,
+ "learning_rate": 0.0006,
+ "loss": 4.093838214874268,
+ "step": 6871
+ },
+ {
+ "epoch": 95.44716157205241,
+ "grad_norm": 0.01665084809064865,
+ "learning_rate": 0.0006,
+ "loss": 4.1471967697143555,
+ "step": 6872
+ },
+ {
+ "epoch": 95.46113537117904,
+ "grad_norm": 0.018056001514196396,
+ "learning_rate": 0.0006,
+ "loss": 4.1388349533081055,
+ "step": 6873
+ },
+ {
+ "epoch": 95.47510917030567,
+ "grad_norm": 0.02129282057285309,
+ "learning_rate": 0.0006,
+ "loss": 4.130726337432861,
+ "step": 6874
+ },
+ {
+ "epoch": 95.48908296943232,
+ "grad_norm": 0.022237276658415794,
+ "learning_rate": 0.0006,
+ "loss": 4.10750675201416,
+ "step": 6875
+ },
+ {
+ "epoch": 95.50305676855895,
+ "grad_norm": 0.021106960251927376,
+ "learning_rate": 0.0006,
+ "loss": 4.136451721191406,
+ "step": 6876
+ },
+ {
+ "epoch": 95.5170305676856,
+ "grad_norm": 0.019107872620224953,
+ "learning_rate": 0.0006,
+ "loss": 4.139644145965576,
+ "step": 6877
+ },
+ {
+ "epoch": 95.53100436681223,
+ "grad_norm": 0.018290145322680473,
+ "learning_rate": 0.0006,
+ "loss": 4.161884784698486,
+ "step": 6878
+ },
+ {
+ "epoch": 95.54497816593886,
+ "grad_norm": 0.018161306157708168,
+ "learning_rate": 0.0006,
+ "loss": 4.1252336502075195,
+ "step": 6879
+ },
+ {
+ "epoch": 95.5589519650655,
+ "grad_norm": 0.018432756885886192,
+ "learning_rate": 0.0006,
+ "loss": 4.0635986328125,
+ "step": 6880
+ },
+ {
+ "epoch": 95.57292576419214,
+ "grad_norm": 0.016592558473348618,
+ "learning_rate": 0.0006,
+ "loss": 4.037112236022949,
+ "step": 6881
+ },
+ {
+ "epoch": 95.58689956331878,
+ "grad_norm": 0.015283452346920967,
+ "learning_rate": 0.0006,
+ "loss": 4.1964826583862305,
+ "step": 6882
+ },
+ {
+ "epoch": 95.60087336244541,
+ "grad_norm": 0.015539180487394333,
+ "learning_rate": 0.0006,
+ "loss": 4.110917568206787,
+ "step": 6883
+ },
+ {
+ "epoch": 95.61484716157206,
+ "grad_norm": 0.016508113592863083,
+ "learning_rate": 0.0006,
+ "loss": 3.9705517292022705,
+ "step": 6884
+ },
+ {
+ "epoch": 95.62882096069869,
+ "grad_norm": 0.016635406762361526,
+ "learning_rate": 0.0006,
+ "loss": 4.025636196136475,
+ "step": 6885
+ },
+ {
+ "epoch": 95.64279475982532,
+ "grad_norm": 0.01688900962471962,
+ "learning_rate": 0.0006,
+ "loss": 4.01814079284668,
+ "step": 6886
+ },
+ {
+ "epoch": 95.65676855895197,
+ "grad_norm": 0.01614062674343586,
+ "learning_rate": 0.0006,
+ "loss": 4.098041534423828,
+ "step": 6887
+ },
+ {
+ "epoch": 95.6707423580786,
+ "grad_norm": 0.015679053962230682,
+ "learning_rate": 0.0006,
+ "loss": 4.077049255371094,
+ "step": 6888
+ },
+ {
+ "epoch": 95.68471615720524,
+ "grad_norm": 0.016617605462670326,
+ "learning_rate": 0.0006,
+ "loss": 4.10363245010376,
+ "step": 6889
+ },
+ {
+ "epoch": 95.69868995633188,
+ "grad_norm": 0.015966709703207016,
+ "learning_rate": 0.0006,
+ "loss": 4.100924491882324,
+ "step": 6890
+ },
+ {
+ "epoch": 95.71266375545852,
+ "grad_norm": 0.015768567100167274,
+ "learning_rate": 0.0006,
+ "loss": 4.063976287841797,
+ "step": 6891
+ },
+ {
+ "epoch": 95.72663755458515,
+ "grad_norm": 0.015800341963768005,
+ "learning_rate": 0.0006,
+ "loss": 4.008409023284912,
+ "step": 6892
+ },
+ {
+ "epoch": 95.74061135371178,
+ "grad_norm": 0.014288571663200855,
+ "learning_rate": 0.0006,
+ "loss": 4.139591217041016,
+ "step": 6893
+ },
+ {
+ "epoch": 95.75458515283843,
+ "grad_norm": 0.013743448071181774,
+ "learning_rate": 0.0006,
+ "loss": 3.967352867126465,
+ "step": 6894
+ },
+ {
+ "epoch": 95.76855895196506,
+ "grad_norm": 0.01387989055365324,
+ "learning_rate": 0.0006,
+ "loss": 4.127554893493652,
+ "step": 6895
+ },
+ {
+ "epoch": 95.78253275109171,
+ "grad_norm": 0.013910802081227303,
+ "learning_rate": 0.0006,
+ "loss": 4.130941390991211,
+ "step": 6896
+ },
+ {
+ "epoch": 95.79650655021834,
+ "grad_norm": 0.013936805538833141,
+ "learning_rate": 0.0006,
+ "loss": 4.138744831085205,
+ "step": 6897
+ },
+ {
+ "epoch": 95.81048034934499,
+ "grad_norm": 0.015135723166167736,
+ "learning_rate": 0.0006,
+ "loss": 4.131860256195068,
+ "step": 6898
+ },
+ {
+ "epoch": 95.82445414847162,
+ "grad_norm": 0.014921358786523342,
+ "learning_rate": 0.0006,
+ "loss": 4.1949462890625,
+ "step": 6899
+ },
+ {
+ "epoch": 95.83842794759825,
+ "grad_norm": 0.01401488296687603,
+ "learning_rate": 0.0006,
+ "loss": 4.181267738342285,
+ "step": 6900
+ },
+ {
+ "epoch": 95.8524017467249,
+ "grad_norm": 0.013691626489162445,
+ "learning_rate": 0.0006,
+ "loss": 4.113667964935303,
+ "step": 6901
+ },
+ {
+ "epoch": 95.86637554585153,
+ "grad_norm": 0.013890047557651997,
+ "learning_rate": 0.0006,
+ "loss": 4.163067817687988,
+ "step": 6902
+ },
+ {
+ "epoch": 95.88034934497817,
+ "grad_norm": 0.01385503076016903,
+ "learning_rate": 0.0006,
+ "loss": 4.076153755187988,
+ "step": 6903
+ },
+ {
+ "epoch": 95.8943231441048,
+ "grad_norm": 0.01383926346898079,
+ "learning_rate": 0.0006,
+ "loss": 4.102423667907715,
+ "step": 6904
+ },
+ {
+ "epoch": 95.90829694323143,
+ "grad_norm": 0.015961717814207077,
+ "learning_rate": 0.0006,
+ "loss": 4.1543779373168945,
+ "step": 6905
+ },
+ {
+ "epoch": 95.92227074235808,
+ "grad_norm": 0.016718363389372826,
+ "learning_rate": 0.0006,
+ "loss": 3.991607189178467,
+ "step": 6906
+ },
+ {
+ "epoch": 95.93624454148471,
+ "grad_norm": 0.01585792563855648,
+ "learning_rate": 0.0006,
+ "loss": 4.048587799072266,
+ "step": 6907
+ },
+ {
+ "epoch": 95.95021834061136,
+ "grad_norm": 0.015311713330447674,
+ "learning_rate": 0.0006,
+ "loss": 4.1070451736450195,
+ "step": 6908
+ },
+ {
+ "epoch": 95.96419213973799,
+ "grad_norm": 0.014521263539791107,
+ "learning_rate": 0.0006,
+ "loss": 4.087587356567383,
+ "step": 6909
+ },
+ {
+ "epoch": 95.97816593886463,
+ "grad_norm": 0.01433157455176115,
+ "learning_rate": 0.0006,
+ "loss": 4.139904499053955,
+ "step": 6910
+ },
+ {
+ "epoch": 95.99213973799127,
+ "grad_norm": 0.014475514180958271,
+ "learning_rate": 0.0006,
+ "loss": 4.182525634765625,
+ "step": 6911
+ },
+ {
+ "epoch": 96.0,
+ "grad_norm": 0.015245620161294937,
+ "learning_rate": 0.0006,
+ "loss": 4.1199870109558105,
+ "step": 6912
+ },
+ {
+ "epoch": 96.0,
+ "eval_loss": 4.58721399307251,
+ "eval_runtime": 56.395,
+ "eval_samples_per_second": 43.302,
+ "eval_steps_per_second": 1.365,
+ "step": 6912
+ },
+ {
+ "epoch": 96.01397379912663,
+ "grad_norm": 0.014325885102152824,
+ "learning_rate": 0.0006,
+ "loss": 4.025214195251465,
+ "step": 6913
+ },
+ {
+ "epoch": 96.02794759825328,
+ "grad_norm": 0.01678297482430935,
+ "learning_rate": 0.0006,
+ "loss": 3.998842716217041,
+ "step": 6914
+ },
+ {
+ "epoch": 96.04192139737991,
+ "grad_norm": 0.017044762149453163,
+ "learning_rate": 0.0006,
+ "loss": 4.122270584106445,
+ "step": 6915
+ },
+ {
+ "epoch": 96.05589519650655,
+ "grad_norm": 0.017583075910806656,
+ "learning_rate": 0.0006,
+ "loss": 3.9782891273498535,
+ "step": 6916
+ },
+ {
+ "epoch": 96.06986899563319,
+ "grad_norm": 0.0179787315428257,
+ "learning_rate": 0.0006,
+ "loss": 4.030402660369873,
+ "step": 6917
+ },
+ {
+ "epoch": 96.08384279475983,
+ "grad_norm": 0.0159553661942482,
+ "learning_rate": 0.0006,
+ "loss": 3.949040651321411,
+ "step": 6918
+ },
+ {
+ "epoch": 96.09781659388646,
+ "grad_norm": 0.01554026547819376,
+ "learning_rate": 0.0006,
+ "loss": 4.0429534912109375,
+ "step": 6919
+ },
+ {
+ "epoch": 96.1117903930131,
+ "grad_norm": 0.016099359840154648,
+ "learning_rate": 0.0006,
+ "loss": 4.037537097930908,
+ "step": 6920
+ },
+ {
+ "epoch": 96.12576419213974,
+ "grad_norm": 0.01676846854388714,
+ "learning_rate": 0.0006,
+ "loss": 4.022023677825928,
+ "step": 6921
+ },
+ {
+ "epoch": 96.13973799126637,
+ "grad_norm": 0.019358212128281593,
+ "learning_rate": 0.0006,
+ "loss": 4.048679351806641,
+ "step": 6922
+ },
+ {
+ "epoch": 96.15371179039302,
+ "grad_norm": 0.020390303805470467,
+ "learning_rate": 0.0006,
+ "loss": 4.144132614135742,
+ "step": 6923
+ },
+ {
+ "epoch": 96.16768558951965,
+ "grad_norm": 0.01884407177567482,
+ "learning_rate": 0.0006,
+ "loss": 4.128715515136719,
+ "step": 6924
+ },
+ {
+ "epoch": 96.18165938864628,
+ "grad_norm": 0.015098451636731625,
+ "learning_rate": 0.0006,
+ "loss": 4.128741264343262,
+ "step": 6925
+ },
+ {
+ "epoch": 96.19563318777293,
+ "grad_norm": 0.0148141635581851,
+ "learning_rate": 0.0006,
+ "loss": 4.054587364196777,
+ "step": 6926
+ },
+ {
+ "epoch": 96.20960698689956,
+ "grad_norm": 0.015594304539263248,
+ "learning_rate": 0.0006,
+ "loss": 4.073890209197998,
+ "step": 6927
+ },
+ {
+ "epoch": 96.2235807860262,
+ "grad_norm": 0.0139383003115654,
+ "learning_rate": 0.0006,
+ "loss": 4.115614891052246,
+ "step": 6928
+ },
+ {
+ "epoch": 96.23755458515284,
+ "grad_norm": 0.0150460135191679,
+ "learning_rate": 0.0006,
+ "loss": 4.040738105773926,
+ "step": 6929
+ },
+ {
+ "epoch": 96.25152838427948,
+ "grad_norm": 0.013879122212529182,
+ "learning_rate": 0.0006,
+ "loss": 3.9292469024658203,
+ "step": 6930
+ },
+ {
+ "epoch": 96.26550218340611,
+ "grad_norm": 0.015546537935733795,
+ "learning_rate": 0.0006,
+ "loss": 4.111361503601074,
+ "step": 6931
+ },
+ {
+ "epoch": 96.27947598253274,
+ "grad_norm": 0.016172925010323524,
+ "learning_rate": 0.0006,
+ "loss": 4.041029930114746,
+ "step": 6932
+ },
+ {
+ "epoch": 96.29344978165939,
+ "grad_norm": 0.015113193541765213,
+ "learning_rate": 0.0006,
+ "loss": 4.157950401306152,
+ "step": 6933
+ },
+ {
+ "epoch": 96.30742358078602,
+ "grad_norm": 0.014663247391581535,
+ "learning_rate": 0.0006,
+ "loss": 4.020952224731445,
+ "step": 6934
+ },
+ {
+ "epoch": 96.32139737991267,
+ "grad_norm": 0.015092858113348484,
+ "learning_rate": 0.0006,
+ "loss": 4.065241813659668,
+ "step": 6935
+ },
+ {
+ "epoch": 96.3353711790393,
+ "grad_norm": 0.013199460692703724,
+ "learning_rate": 0.0006,
+ "loss": 4.062674522399902,
+ "step": 6936
+ },
+ {
+ "epoch": 96.34934497816595,
+ "grad_norm": 0.015065675601363182,
+ "learning_rate": 0.0006,
+ "loss": 4.038748741149902,
+ "step": 6937
+ },
+ {
+ "epoch": 96.36331877729258,
+ "grad_norm": 0.014270043931901455,
+ "learning_rate": 0.0006,
+ "loss": 4.073092460632324,
+ "step": 6938
+ },
+ {
+ "epoch": 96.37729257641921,
+ "grad_norm": 0.01423644833266735,
+ "learning_rate": 0.0006,
+ "loss": 4.238361835479736,
+ "step": 6939
+ },
+ {
+ "epoch": 96.39126637554585,
+ "grad_norm": 0.01380224246531725,
+ "learning_rate": 0.0006,
+ "loss": 4.137866973876953,
+ "step": 6940
+ },
+ {
+ "epoch": 96.40524017467249,
+ "grad_norm": 0.012464135885238647,
+ "learning_rate": 0.0006,
+ "loss": 3.9527339935302734,
+ "step": 6941
+ },
+ {
+ "epoch": 96.41921397379913,
+ "grad_norm": 0.013094968162477016,
+ "learning_rate": 0.0006,
+ "loss": 4.106327056884766,
+ "step": 6942
+ },
+ {
+ "epoch": 96.43318777292576,
+ "grad_norm": 0.013918030075728893,
+ "learning_rate": 0.0006,
+ "loss": 4.072292327880859,
+ "step": 6943
+ },
+ {
+ "epoch": 96.44716157205241,
+ "grad_norm": 0.013835827820003033,
+ "learning_rate": 0.0006,
+ "loss": 4.094313621520996,
+ "step": 6944
+ },
+ {
+ "epoch": 96.46113537117904,
+ "grad_norm": 0.013890027068555355,
+ "learning_rate": 0.0006,
+ "loss": 3.947615623474121,
+ "step": 6945
+ },
+ {
+ "epoch": 96.47510917030567,
+ "grad_norm": 0.012371144257485867,
+ "learning_rate": 0.0006,
+ "loss": 4.038928985595703,
+ "step": 6946
+ },
+ {
+ "epoch": 96.48908296943232,
+ "grad_norm": 0.012236416339874268,
+ "learning_rate": 0.0006,
+ "loss": 3.9835925102233887,
+ "step": 6947
+ },
+ {
+ "epoch": 96.50305676855895,
+ "grad_norm": 0.012866619043052197,
+ "learning_rate": 0.0006,
+ "loss": 4.076041221618652,
+ "step": 6948
+ },
+ {
+ "epoch": 96.5170305676856,
+ "grad_norm": 0.013425248675048351,
+ "learning_rate": 0.0006,
+ "loss": 4.086307525634766,
+ "step": 6949
+ },
+ {
+ "epoch": 96.53100436681223,
+ "grad_norm": 0.014234314672648907,
+ "learning_rate": 0.0006,
+ "loss": 4.087897300720215,
+ "step": 6950
+ },
+ {
+ "epoch": 96.54497816593886,
+ "grad_norm": 0.013670518063008785,
+ "learning_rate": 0.0006,
+ "loss": 4.088400840759277,
+ "step": 6951
+ },
+ {
+ "epoch": 96.5589519650655,
+ "grad_norm": 0.013920796103775501,
+ "learning_rate": 0.0006,
+ "loss": 4.026619911193848,
+ "step": 6952
+ },
+ {
+ "epoch": 96.57292576419214,
+ "grad_norm": 0.013115637935698032,
+ "learning_rate": 0.0006,
+ "loss": 4.141090393066406,
+ "step": 6953
+ },
+ {
+ "epoch": 96.58689956331878,
+ "grad_norm": 0.013466455042362213,
+ "learning_rate": 0.0006,
+ "loss": 4.048084259033203,
+ "step": 6954
+ },
+ {
+ "epoch": 96.60087336244541,
+ "grad_norm": 0.013038012199103832,
+ "learning_rate": 0.0006,
+ "loss": 4.034649848937988,
+ "step": 6955
+ },
+ {
+ "epoch": 96.61484716157206,
+ "grad_norm": 0.012926512397825718,
+ "learning_rate": 0.0006,
+ "loss": 4.138445854187012,
+ "step": 6956
+ },
+ {
+ "epoch": 96.62882096069869,
+ "grad_norm": 0.013271111994981766,
+ "learning_rate": 0.0006,
+ "loss": 4.240598201751709,
+ "step": 6957
+ },
+ {
+ "epoch": 96.64279475982532,
+ "grad_norm": 0.013092855922877789,
+ "learning_rate": 0.0006,
+ "loss": 4.091870307922363,
+ "step": 6958
+ },
+ {
+ "epoch": 96.65676855895197,
+ "grad_norm": 0.013304130174219608,
+ "learning_rate": 0.0006,
+ "loss": 4.048708915710449,
+ "step": 6959
+ },
+ {
+ "epoch": 96.6707423580786,
+ "grad_norm": 0.01289574895054102,
+ "learning_rate": 0.0006,
+ "loss": 4.083846092224121,
+ "step": 6960
+ },
+ {
+ "epoch": 96.68471615720524,
+ "grad_norm": 0.012587146833539009,
+ "learning_rate": 0.0006,
+ "loss": 4.1233015060424805,
+ "step": 6961
+ },
+ {
+ "epoch": 96.69868995633188,
+ "grad_norm": 0.012312786653637886,
+ "learning_rate": 0.0006,
+ "loss": 4.0938262939453125,
+ "step": 6962
+ },
+ {
+ "epoch": 96.71266375545852,
+ "grad_norm": 0.014632516540586948,
+ "learning_rate": 0.0006,
+ "loss": 4.0226545333862305,
+ "step": 6963
+ },
+ {
+ "epoch": 96.72663755458515,
+ "grad_norm": 0.015110369771718979,
+ "learning_rate": 0.0006,
+ "loss": 4.052829742431641,
+ "step": 6964
+ },
+ {
+ "epoch": 96.74061135371178,
+ "grad_norm": 0.015997016802430153,
+ "learning_rate": 0.0006,
+ "loss": 4.126044750213623,
+ "step": 6965
+ },
+ {
+ "epoch": 96.75458515283843,
+ "grad_norm": 0.016428956761956215,
+ "learning_rate": 0.0006,
+ "loss": 4.126011848449707,
+ "step": 6966
+ },
+ {
+ "epoch": 96.76855895196506,
+ "grad_norm": 0.01638922467827797,
+ "learning_rate": 0.0006,
+ "loss": 4.1207685470581055,
+ "step": 6967
+ },
+ {
+ "epoch": 96.78253275109171,
+ "grad_norm": 0.01387663371860981,
+ "learning_rate": 0.0006,
+ "loss": 3.9403412342071533,
+ "step": 6968
+ },
+ {
+ "epoch": 96.79650655021834,
+ "grad_norm": 0.014120660722255707,
+ "learning_rate": 0.0006,
+ "loss": 4.202868461608887,
+ "step": 6969
+ },
+ {
+ "epoch": 96.81048034934499,
+ "grad_norm": 0.014854361303150654,
+ "learning_rate": 0.0006,
+ "loss": 4.015185832977295,
+ "step": 6970
+ },
+ {
+ "epoch": 96.82445414847162,
+ "grad_norm": 0.014359727501869202,
+ "learning_rate": 0.0006,
+ "loss": 4.119946479797363,
+ "step": 6971
+ },
+ {
+ "epoch": 96.83842794759825,
+ "grad_norm": 0.01564539223909378,
+ "learning_rate": 0.0006,
+ "loss": 3.9208037853240967,
+ "step": 6972
+ },
+ {
+ "epoch": 96.8524017467249,
+ "grad_norm": 0.015814676880836487,
+ "learning_rate": 0.0006,
+ "loss": 4.056029319763184,
+ "step": 6973
+ },
+ {
+ "epoch": 96.86637554585153,
+ "grad_norm": 0.016342822462320328,
+ "learning_rate": 0.0006,
+ "loss": 4.160576343536377,
+ "step": 6974
+ },
+ {
+ "epoch": 96.88034934497817,
+ "grad_norm": 0.016163306310772896,
+ "learning_rate": 0.0006,
+ "loss": 3.94950008392334,
+ "step": 6975
+ },
+ {
+ "epoch": 96.8943231441048,
+ "grad_norm": 0.018308300524950027,
+ "learning_rate": 0.0006,
+ "loss": 4.096719741821289,
+ "step": 6976
+ },
+ {
+ "epoch": 96.90829694323143,
+ "grad_norm": 0.01985356956720352,
+ "learning_rate": 0.0006,
+ "loss": 4.052276611328125,
+ "step": 6977
+ },
+ {
+ "epoch": 96.92227074235808,
+ "grad_norm": 0.01949184387922287,
+ "learning_rate": 0.0006,
+ "loss": 4.077932357788086,
+ "step": 6978
+ },
+ {
+ "epoch": 96.93624454148471,
+ "grad_norm": 0.018964678049087524,
+ "learning_rate": 0.0006,
+ "loss": 4.019469738006592,
+ "step": 6979
+ },
+ {
+ "epoch": 96.95021834061136,
+ "grad_norm": 0.019365238025784492,
+ "learning_rate": 0.0006,
+ "loss": 4.032814979553223,
+ "step": 6980
+ },
+ {
+ "epoch": 96.96419213973799,
+ "grad_norm": 0.01766037568449974,
+ "learning_rate": 0.0006,
+ "loss": 4.061336517333984,
+ "step": 6981
+ },
+ {
+ "epoch": 96.97816593886463,
+ "grad_norm": 0.018002403900027275,
+ "learning_rate": 0.0006,
+ "loss": 4.054131507873535,
+ "step": 6982
+ },
+ {
+ "epoch": 96.99213973799127,
+ "grad_norm": 0.01888604648411274,
+ "learning_rate": 0.0006,
+ "loss": 3.9295921325683594,
+ "step": 6983
+ },
+ {
+ "epoch": 97.0,
+ "grad_norm": 0.01609404757618904,
+ "learning_rate": 0.0006,
+ "loss": 4.037106513977051,
+ "step": 6984
+ },
+ {
+ "epoch": 97.0,
+ "eval_loss": 4.5986433029174805,
+ "eval_runtime": 56.5822,
+ "eval_samples_per_second": 43.158,
+ "eval_steps_per_second": 1.361,
+ "step": 6984
+ },
+ {
+ "epoch": 97.01397379912663,
+ "grad_norm": 0.015459981746971607,
+ "learning_rate": 0.0006,
+ "loss": 4.06088924407959,
+ "step": 6985
+ },
+ {
+ "epoch": 97.02794759825328,
+ "grad_norm": 0.016333693638443947,
+ "learning_rate": 0.0006,
+ "loss": 3.9525675773620605,
+ "step": 6986
+ },
+ {
+ "epoch": 97.04192139737991,
+ "grad_norm": 0.0157504603266716,
+ "learning_rate": 0.0006,
+ "loss": 4.143257141113281,
+ "step": 6987
+ },
+ {
+ "epoch": 97.05589519650655,
+ "grad_norm": 0.015252048149704933,
+ "learning_rate": 0.0006,
+ "loss": 4.158165454864502,
+ "step": 6988
+ },
+ {
+ "epoch": 97.06986899563319,
+ "grad_norm": 0.015041683800518513,
+ "learning_rate": 0.0006,
+ "loss": 4.030884742736816,
+ "step": 6989
+ },
+ {
+ "epoch": 97.08384279475983,
+ "grad_norm": 0.014693699777126312,
+ "learning_rate": 0.0006,
+ "loss": 3.8566339015960693,
+ "step": 6990
+ },
+ {
+ "epoch": 97.09781659388646,
+ "grad_norm": 0.0141681469976902,
+ "learning_rate": 0.0006,
+ "loss": 4.001909255981445,
+ "step": 6991
+ },
+ {
+ "epoch": 97.1117903930131,
+ "grad_norm": 0.013231917284429073,
+ "learning_rate": 0.0006,
+ "loss": 3.995041608810425,
+ "step": 6992
+ },
+ {
+ "epoch": 97.12576419213974,
+ "grad_norm": 0.014617119915783405,
+ "learning_rate": 0.0006,
+ "loss": 4.012053489685059,
+ "step": 6993
+ },
+ {
+ "epoch": 97.13973799126637,
+ "grad_norm": 0.014983531087636948,
+ "learning_rate": 0.0006,
+ "loss": 4.084125518798828,
+ "step": 6994
+ },
+ {
+ "epoch": 97.15371179039302,
+ "grad_norm": 0.013970102183520794,
+ "learning_rate": 0.0006,
+ "loss": 4.001402378082275,
+ "step": 6995
+ },
+ {
+ "epoch": 97.16768558951965,
+ "grad_norm": 0.0134900426492095,
+ "learning_rate": 0.0006,
+ "loss": 4.007713794708252,
+ "step": 6996
+ },
+ {
+ "epoch": 97.18165938864628,
+ "grad_norm": 0.013806418515741825,
+ "learning_rate": 0.0006,
+ "loss": 3.9524781703948975,
+ "step": 6997
+ },
+ {
+ "epoch": 97.19563318777293,
+ "grad_norm": 0.014304264448583126,
+ "learning_rate": 0.0006,
+ "loss": 3.9666526317596436,
+ "step": 6998
+ },
+ {
+ "epoch": 97.20960698689956,
+ "grad_norm": 0.016392510384321213,
+ "learning_rate": 0.0006,
+ "loss": 3.9915683269500732,
+ "step": 6999
+ },
+ {
+ "epoch": 97.2235807860262,
+ "grad_norm": 0.017044387757778168,
+ "learning_rate": 0.0006,
+ "loss": 4.11865234375,
+ "step": 7000
+ },
+ {
+ "epoch": 97.23755458515284,
+ "grad_norm": 0.016325082629919052,
+ "learning_rate": 0.0006,
+ "loss": 4.0636820793151855,
+ "step": 7001
+ },
+ {
+ "epoch": 97.25152838427948,
+ "grad_norm": 0.014634544961154461,
+ "learning_rate": 0.0006,
+ "loss": 3.967127799987793,
+ "step": 7002
+ },
+ {
+ "epoch": 97.26550218340611,
+ "grad_norm": 0.014043385162949562,
+ "learning_rate": 0.0006,
+ "loss": 4.072741985321045,
+ "step": 7003
+ },
+ {
+ "epoch": 97.27947598253274,
+ "grad_norm": 0.01376431342214346,
+ "learning_rate": 0.0006,
+ "loss": 4.07547664642334,
+ "step": 7004
+ },
+ {
+ "epoch": 97.29344978165939,
+ "grad_norm": 0.013959003612399101,
+ "learning_rate": 0.0006,
+ "loss": 3.9862523078918457,
+ "step": 7005
+ },
+ {
+ "epoch": 97.30742358078602,
+ "grad_norm": 0.01420508325099945,
+ "learning_rate": 0.0006,
+ "loss": 4.085540294647217,
+ "step": 7006
+ },
+ {
+ "epoch": 97.32139737991267,
+ "grad_norm": 0.01613876223564148,
+ "learning_rate": 0.0006,
+ "loss": 4.025545120239258,
+ "step": 7007
+ },
+ {
+ "epoch": 97.3353711790393,
+ "grad_norm": 0.017000367864966393,
+ "learning_rate": 0.0006,
+ "loss": 4.047660827636719,
+ "step": 7008
+ },
+ {
+ "epoch": 97.34934497816595,
+ "grad_norm": 0.01646680012345314,
+ "learning_rate": 0.0006,
+ "loss": 4.087928771972656,
+ "step": 7009
+ },
+ {
+ "epoch": 97.36331877729258,
+ "grad_norm": 0.015728263184428215,
+ "learning_rate": 0.0006,
+ "loss": 4.022225856781006,
+ "step": 7010
+ },
+ {
+ "epoch": 97.37729257641921,
+ "grad_norm": 0.015262200497090816,
+ "learning_rate": 0.0006,
+ "loss": 3.944697618484497,
+ "step": 7011
+ },
+ {
+ "epoch": 97.39126637554585,
+ "grad_norm": 0.016452079638838768,
+ "learning_rate": 0.0006,
+ "loss": 3.957644462585449,
+ "step": 7012
+ },
+ {
+ "epoch": 97.40524017467249,
+ "grad_norm": 0.018871447071433067,
+ "learning_rate": 0.0006,
+ "loss": 4.017784118652344,
+ "step": 7013
+ },
+ {
+ "epoch": 97.41921397379913,
+ "grad_norm": 0.01902632787823677,
+ "learning_rate": 0.0006,
+ "loss": 3.973320484161377,
+ "step": 7014
+ },
+ {
+ "epoch": 97.43318777292576,
+ "grad_norm": 0.016632359474897385,
+ "learning_rate": 0.0006,
+ "loss": 4.047779083251953,
+ "step": 7015
+ },
+ {
+ "epoch": 97.44716157205241,
+ "grad_norm": 0.01676122285425663,
+ "learning_rate": 0.0006,
+ "loss": 4.149074554443359,
+ "step": 7016
+ },
+ {
+ "epoch": 97.46113537117904,
+ "grad_norm": 0.01764233037829399,
+ "learning_rate": 0.0006,
+ "loss": 3.92502760887146,
+ "step": 7017
+ },
+ {
+ "epoch": 97.47510917030567,
+ "grad_norm": 0.01836075820028782,
+ "learning_rate": 0.0006,
+ "loss": 4.173796653747559,
+ "step": 7018
+ },
+ {
+ "epoch": 97.48908296943232,
+ "grad_norm": 0.017320366576313972,
+ "learning_rate": 0.0006,
+ "loss": 4.007115364074707,
+ "step": 7019
+ },
+ {
+ "epoch": 97.50305676855895,
+ "grad_norm": 0.015976276248693466,
+ "learning_rate": 0.0006,
+ "loss": 4.165112495422363,
+ "step": 7020
+ },
+ {
+ "epoch": 97.5170305676856,
+ "grad_norm": 0.0173528753221035,
+ "learning_rate": 0.0006,
+ "loss": 4.158927917480469,
+ "step": 7021
+ },
+ {
+ "epoch": 97.53100436681223,
+ "grad_norm": 0.01881110668182373,
+ "learning_rate": 0.0006,
+ "loss": 4.069684982299805,
+ "step": 7022
+ },
+ {
+ "epoch": 97.54497816593886,
+ "grad_norm": 0.01849045790731907,
+ "learning_rate": 0.0006,
+ "loss": 4.085758209228516,
+ "step": 7023
+ },
+ {
+ "epoch": 97.5589519650655,
+ "grad_norm": 0.016648925840854645,
+ "learning_rate": 0.0006,
+ "loss": 4.047055721282959,
+ "step": 7024
+ },
+ {
+ "epoch": 97.57292576419214,
+ "grad_norm": 0.014700878411531448,
+ "learning_rate": 0.0006,
+ "loss": 4.0177507400512695,
+ "step": 7025
+ },
+ {
+ "epoch": 97.58689956331878,
+ "grad_norm": 0.014601817354559898,
+ "learning_rate": 0.0006,
+ "loss": 4.070394039154053,
+ "step": 7026
+ },
+ {
+ "epoch": 97.60087336244541,
+ "grad_norm": 0.016391467303037643,
+ "learning_rate": 0.0006,
+ "loss": 3.957401990890503,
+ "step": 7027
+ },
+ {
+ "epoch": 97.61484716157206,
+ "grad_norm": 0.015401630662381649,
+ "learning_rate": 0.0006,
+ "loss": 4.040555000305176,
+ "step": 7028
+ },
+ {
+ "epoch": 97.62882096069869,
+ "grad_norm": 0.014783227816224098,
+ "learning_rate": 0.0006,
+ "loss": 4.039798259735107,
+ "step": 7029
+ },
+ {
+ "epoch": 97.64279475982532,
+ "grad_norm": 0.014704087749123573,
+ "learning_rate": 0.0006,
+ "loss": 4.059834003448486,
+ "step": 7030
+ },
+ {
+ "epoch": 97.65676855895197,
+ "grad_norm": 0.01375004556030035,
+ "learning_rate": 0.0006,
+ "loss": 4.084839344024658,
+ "step": 7031
+ },
+ {
+ "epoch": 97.6707423580786,
+ "grad_norm": 0.01403744611889124,
+ "learning_rate": 0.0006,
+ "loss": 4.060206890106201,
+ "step": 7032
+ },
+ {
+ "epoch": 97.68471615720524,
+ "grad_norm": 0.012827585451304913,
+ "learning_rate": 0.0006,
+ "loss": 4.187256813049316,
+ "step": 7033
+ },
+ {
+ "epoch": 97.69868995633188,
+ "grad_norm": 0.013072148896753788,
+ "learning_rate": 0.0006,
+ "loss": 4.136262893676758,
+ "step": 7034
+ },
+ {
+ "epoch": 97.71266375545852,
+ "grad_norm": 0.013496107421815395,
+ "learning_rate": 0.0006,
+ "loss": 3.9419479370117188,
+ "step": 7035
+ },
+ {
+ "epoch": 97.72663755458515,
+ "grad_norm": 0.014445106498897076,
+ "learning_rate": 0.0006,
+ "loss": 3.9494845867156982,
+ "step": 7036
+ },
+ {
+ "epoch": 97.74061135371178,
+ "grad_norm": 0.014507179148495197,
+ "learning_rate": 0.0006,
+ "loss": 4.05567741394043,
+ "step": 7037
+ },
+ {
+ "epoch": 97.75458515283843,
+ "grad_norm": 0.015422643162310123,
+ "learning_rate": 0.0006,
+ "loss": 4.103999137878418,
+ "step": 7038
+ },
+ {
+ "epoch": 97.76855895196506,
+ "grad_norm": 0.01515049859881401,
+ "learning_rate": 0.0006,
+ "loss": 4.094231605529785,
+ "step": 7039
+ },
+ {
+ "epoch": 97.78253275109171,
+ "grad_norm": 0.013850145973265171,
+ "learning_rate": 0.0006,
+ "loss": 4.084870338439941,
+ "step": 7040
+ },
+ {
+ "epoch": 97.79650655021834,
+ "grad_norm": 0.013332060538232327,
+ "learning_rate": 0.0006,
+ "loss": 4.193096160888672,
+ "step": 7041
+ },
+ {
+ "epoch": 97.81048034934499,
+ "grad_norm": 0.014778176322579384,
+ "learning_rate": 0.0006,
+ "loss": 4.0440216064453125,
+ "step": 7042
+ },
+ {
+ "epoch": 97.82445414847162,
+ "grad_norm": 0.013498589396476746,
+ "learning_rate": 0.0006,
+ "loss": 4.088800430297852,
+ "step": 7043
+ },
+ {
+ "epoch": 97.83842794759825,
+ "grad_norm": 0.014520619064569473,
+ "learning_rate": 0.0006,
+ "loss": 4.156583786010742,
+ "step": 7044
+ },
+ {
+ "epoch": 97.8524017467249,
+ "grad_norm": 0.014858689159154892,
+ "learning_rate": 0.0006,
+ "loss": 4.113174915313721,
+ "step": 7045
+ },
+ {
+ "epoch": 97.86637554585153,
+ "grad_norm": 0.015516536310315132,
+ "learning_rate": 0.0006,
+ "loss": 4.07958984375,
+ "step": 7046
+ },
+ {
+ "epoch": 97.88034934497817,
+ "grad_norm": 0.016619030386209488,
+ "learning_rate": 0.0006,
+ "loss": 4.03786563873291,
+ "step": 7047
+ },
+ {
+ "epoch": 97.8943231441048,
+ "grad_norm": 0.015789631754159927,
+ "learning_rate": 0.0006,
+ "loss": 4.065053462982178,
+ "step": 7048
+ },
+ {
+ "epoch": 97.90829694323143,
+ "grad_norm": 0.015848442912101746,
+ "learning_rate": 0.0006,
+ "loss": 4.183546543121338,
+ "step": 7049
+ },
+ {
+ "epoch": 97.92227074235808,
+ "grad_norm": 0.014621895737946033,
+ "learning_rate": 0.0006,
+ "loss": 4.024744987487793,
+ "step": 7050
+ },
+ {
+ "epoch": 97.93624454148471,
+ "grad_norm": 0.015793966129422188,
+ "learning_rate": 0.0006,
+ "loss": 3.9715819358825684,
+ "step": 7051
+ },
+ {
+ "epoch": 97.95021834061136,
+ "grad_norm": 0.014637742191553116,
+ "learning_rate": 0.0006,
+ "loss": 4.061223983764648,
+ "step": 7052
+ },
+ {
+ "epoch": 97.96419213973799,
+ "grad_norm": 0.0143470773473382,
+ "learning_rate": 0.0006,
+ "loss": 4.00929069519043,
+ "step": 7053
+ },
+ {
+ "epoch": 97.97816593886463,
+ "grad_norm": 0.014167622663080692,
+ "learning_rate": 0.0006,
+ "loss": 3.982048749923706,
+ "step": 7054
+ },
+ {
+ "epoch": 97.99213973799127,
+ "grad_norm": 0.01527823694050312,
+ "learning_rate": 0.0006,
+ "loss": 4.063419818878174,
+ "step": 7055
+ },
+ {
+ "epoch": 98.0,
+ "grad_norm": 0.015410988591611385,
+ "learning_rate": 0.0006,
+ "loss": 4.058195114135742,
+ "step": 7056
+ },
+ {
+ "epoch": 98.0,
+ "eval_loss": 4.5309038162231445,
+ "eval_runtime": 56.6931,
+ "eval_samples_per_second": 43.074,
+ "eval_steps_per_second": 1.358,
+ "step": 7056
+ },
+ {
+ "epoch": 98.01397379912663,
+ "grad_norm": 0.013760345987975597,
+ "learning_rate": 0.0006,
+ "loss": 4.059067249298096,
+ "step": 7057
+ },
+ {
+ "epoch": 98.02794759825328,
+ "grad_norm": 0.013799279928207397,
+ "learning_rate": 0.0006,
+ "loss": 3.975419044494629,
+ "step": 7058
+ },
+ {
+ "epoch": 98.04192139737991,
+ "grad_norm": 0.014793423935770988,
+ "learning_rate": 0.0006,
+ "loss": 4.163051605224609,
+ "step": 7059
+ },
+ {
+ "epoch": 98.05589519650655,
+ "grad_norm": 0.015820559114217758,
+ "learning_rate": 0.0006,
+ "loss": 4.032471656799316,
+ "step": 7060
+ },
+ {
+ "epoch": 98.06986899563319,
+ "grad_norm": 0.014252204447984695,
+ "learning_rate": 0.0006,
+ "loss": 4.038850784301758,
+ "step": 7061
+ },
+ {
+ "epoch": 98.08384279475983,
+ "grad_norm": 0.01272337045520544,
+ "learning_rate": 0.0006,
+ "loss": 4.054554462432861,
+ "step": 7062
+ },
+ {
+ "epoch": 98.09781659388646,
+ "grad_norm": 0.01498495228588581,
+ "learning_rate": 0.0006,
+ "loss": 4.009291648864746,
+ "step": 7063
+ },
+ {
+ "epoch": 98.1117903930131,
+ "grad_norm": 0.015501809306442738,
+ "learning_rate": 0.0006,
+ "loss": 3.999412775039673,
+ "step": 7064
+ },
+ {
+ "epoch": 98.12576419213974,
+ "grad_norm": 0.015108918771147728,
+ "learning_rate": 0.0006,
+ "loss": 4.015168190002441,
+ "step": 7065
+ },
+ {
+ "epoch": 98.13973799126637,
+ "grad_norm": 0.01489692460745573,
+ "learning_rate": 0.0006,
+ "loss": 4.163906097412109,
+ "step": 7066
+ },
+ {
+ "epoch": 98.15371179039302,
+ "grad_norm": 0.014246871694922447,
+ "learning_rate": 0.0006,
+ "loss": 3.993173122406006,
+ "step": 7067
+ },
+ {
+ "epoch": 98.16768558951965,
+ "grad_norm": 0.015163703821599483,
+ "learning_rate": 0.0006,
+ "loss": 3.908207893371582,
+ "step": 7068
+ },
+ {
+ "epoch": 98.18165938864628,
+ "grad_norm": 0.0141439288854599,
+ "learning_rate": 0.0006,
+ "loss": 4.064990997314453,
+ "step": 7069
+ },
+ {
+ "epoch": 98.19563318777293,
+ "grad_norm": 0.013251710683107376,
+ "learning_rate": 0.0006,
+ "loss": 4.042545318603516,
+ "step": 7070
+ },
+ {
+ "epoch": 98.20960698689956,
+ "grad_norm": 0.01428674440830946,
+ "learning_rate": 0.0006,
+ "loss": 4.041232585906982,
+ "step": 7071
+ },
+ {
+ "epoch": 98.2235807860262,
+ "grad_norm": 0.012818179093301296,
+ "learning_rate": 0.0006,
+ "loss": 3.9160139560699463,
+ "step": 7072
+ },
+ {
+ "epoch": 98.23755458515284,
+ "grad_norm": 0.012890368700027466,
+ "learning_rate": 0.0006,
+ "loss": 3.9687235355377197,
+ "step": 7073
+ },
+ {
+ "epoch": 98.25152838427948,
+ "grad_norm": 0.014132961630821228,
+ "learning_rate": 0.0006,
+ "loss": 4.066800117492676,
+ "step": 7074
+ },
+ {
+ "epoch": 98.26550218340611,
+ "grad_norm": 0.013544383458793163,
+ "learning_rate": 0.0006,
+ "loss": 4.111777305603027,
+ "step": 7075
+ },
+ {
+ "epoch": 98.27947598253274,
+ "grad_norm": 0.014414685778319836,
+ "learning_rate": 0.0006,
+ "loss": 4.109633445739746,
+ "step": 7076
+ },
+ {
+ "epoch": 98.29344978165939,
+ "grad_norm": 0.015201127156615257,
+ "learning_rate": 0.0006,
+ "loss": 4.033087253570557,
+ "step": 7077
+ },
+ {
+ "epoch": 98.30742358078602,
+ "grad_norm": 0.016936076804995537,
+ "learning_rate": 0.0006,
+ "loss": 4.10897970199585,
+ "step": 7078
+ },
+ {
+ "epoch": 98.32139737991267,
+ "grad_norm": 0.01556481048464775,
+ "learning_rate": 0.0006,
+ "loss": 3.968942165374756,
+ "step": 7079
+ },
+ {
+ "epoch": 98.3353711790393,
+ "grad_norm": 0.015971971675753593,
+ "learning_rate": 0.0006,
+ "loss": 4.024805068969727,
+ "step": 7080
+ },
+ {
+ "epoch": 98.34934497816595,
+ "grad_norm": 0.016742313280701637,
+ "learning_rate": 0.0006,
+ "loss": 4.037862777709961,
+ "step": 7081
+ },
+ {
+ "epoch": 98.36331877729258,
+ "grad_norm": 0.015753688290715218,
+ "learning_rate": 0.0006,
+ "loss": 4.015701770782471,
+ "step": 7082
+ },
+ {
+ "epoch": 98.37729257641921,
+ "grad_norm": 0.014211287721991539,
+ "learning_rate": 0.0006,
+ "loss": 4.068171501159668,
+ "step": 7083
+ },
+ {
+ "epoch": 98.39126637554585,
+ "grad_norm": 0.013903875835239887,
+ "learning_rate": 0.0006,
+ "loss": 4.052516937255859,
+ "step": 7084
+ },
+ {
+ "epoch": 98.40524017467249,
+ "grad_norm": 0.015590954571962357,
+ "learning_rate": 0.0006,
+ "loss": 4.056125164031982,
+ "step": 7085
+ },
+ {
+ "epoch": 98.41921397379913,
+ "grad_norm": 0.016314011067152023,
+ "learning_rate": 0.0006,
+ "loss": 4.014211177825928,
+ "step": 7086
+ },
+ {
+ "epoch": 98.43318777292576,
+ "grad_norm": 0.017208732664585114,
+ "learning_rate": 0.0006,
+ "loss": 4.0615339279174805,
+ "step": 7087
+ },
+ {
+ "epoch": 98.44716157205241,
+ "grad_norm": 0.017213109880685806,
+ "learning_rate": 0.0006,
+ "loss": 3.9580445289611816,
+ "step": 7088
+ },
+ {
+ "epoch": 98.46113537117904,
+ "grad_norm": 0.016858434304594994,
+ "learning_rate": 0.0006,
+ "loss": 3.9932594299316406,
+ "step": 7089
+ },
+ {
+ "epoch": 98.47510917030567,
+ "grad_norm": 0.014381436631083488,
+ "learning_rate": 0.0006,
+ "loss": 4.159034729003906,
+ "step": 7090
+ },
+ {
+ "epoch": 98.48908296943232,
+ "grad_norm": 0.013713826425373554,
+ "learning_rate": 0.0006,
+ "loss": 4.055055141448975,
+ "step": 7091
+ },
+ {
+ "epoch": 98.50305676855895,
+ "grad_norm": 0.014764957129955292,
+ "learning_rate": 0.0006,
+ "loss": 4.059102535247803,
+ "step": 7092
+ },
+ {
+ "epoch": 98.5170305676856,
+ "grad_norm": 0.014522310346364975,
+ "learning_rate": 0.0006,
+ "loss": 4.093504905700684,
+ "step": 7093
+ },
+ {
+ "epoch": 98.53100436681223,
+ "grad_norm": 0.013484988361597061,
+ "learning_rate": 0.0006,
+ "loss": 4.079945087432861,
+ "step": 7094
+ },
+ {
+ "epoch": 98.54497816593886,
+ "grad_norm": 0.013231529854238033,
+ "learning_rate": 0.0006,
+ "loss": 4.046164512634277,
+ "step": 7095
+ },
+ {
+ "epoch": 98.5589519650655,
+ "grad_norm": 0.01392320916056633,
+ "learning_rate": 0.0006,
+ "loss": 3.975592613220215,
+ "step": 7096
+ },
+ {
+ "epoch": 98.57292576419214,
+ "grad_norm": 0.012323490343987942,
+ "learning_rate": 0.0006,
+ "loss": 3.9167721271514893,
+ "step": 7097
+ },
+ {
+ "epoch": 98.58689956331878,
+ "grad_norm": 0.012565829791128635,
+ "learning_rate": 0.0006,
+ "loss": 4.085109710693359,
+ "step": 7098
+ },
+ {
+ "epoch": 98.60087336244541,
+ "grad_norm": 0.013498359359800816,
+ "learning_rate": 0.0006,
+ "loss": 4.102974891662598,
+ "step": 7099
+ },
+ {
+ "epoch": 98.61484716157206,
+ "grad_norm": 0.014182298444211483,
+ "learning_rate": 0.0006,
+ "loss": 4.041601657867432,
+ "step": 7100
+ },
+ {
+ "epoch": 98.62882096069869,
+ "grad_norm": 0.013925517909228802,
+ "learning_rate": 0.0006,
+ "loss": 4.115788459777832,
+ "step": 7101
+ },
+ {
+ "epoch": 98.64279475982532,
+ "grad_norm": 0.03483536094427109,
+ "learning_rate": 0.0006,
+ "loss": 4.105683326721191,
+ "step": 7102
+ },
+ {
+ "epoch": 98.65676855895197,
+ "grad_norm": 0.014278477989137173,
+ "learning_rate": 0.0006,
+ "loss": 4.04194974899292,
+ "step": 7103
+ },
+ {
+ "epoch": 98.6707423580786,
+ "grad_norm": 0.015309838578104973,
+ "learning_rate": 0.0006,
+ "loss": 4.094045162200928,
+ "step": 7104
+ },
+ {
+ "epoch": 98.68471615720524,
+ "grad_norm": 0.015831857919692993,
+ "learning_rate": 0.0006,
+ "loss": 3.9989206790924072,
+ "step": 7105
+ },
+ {
+ "epoch": 98.69868995633188,
+ "grad_norm": 0.014071783982217312,
+ "learning_rate": 0.0006,
+ "loss": 4.0480852127075195,
+ "step": 7106
+ },
+ {
+ "epoch": 98.71266375545852,
+ "grad_norm": 0.014099884778261185,
+ "learning_rate": 0.0006,
+ "loss": 4.0519633293151855,
+ "step": 7107
+ },
+ {
+ "epoch": 98.72663755458515,
+ "grad_norm": 0.0143047496676445,
+ "learning_rate": 0.0006,
+ "loss": 3.9782276153564453,
+ "step": 7108
+ },
+ {
+ "epoch": 98.74061135371178,
+ "grad_norm": 0.013741862028837204,
+ "learning_rate": 0.0006,
+ "loss": 4.035103797912598,
+ "step": 7109
+ },
+ {
+ "epoch": 98.75458515283843,
+ "grad_norm": 0.014014648273587227,
+ "learning_rate": 0.0006,
+ "loss": 4.09281063079834,
+ "step": 7110
+ },
+ {
+ "epoch": 98.76855895196506,
+ "grad_norm": 0.013909446075558662,
+ "learning_rate": 0.0006,
+ "loss": 4.033784866333008,
+ "step": 7111
+ },
+ {
+ "epoch": 98.78253275109171,
+ "grad_norm": 0.013188562355935574,
+ "learning_rate": 0.0006,
+ "loss": 4.095016956329346,
+ "step": 7112
+ },
+ {
+ "epoch": 98.79650655021834,
+ "grad_norm": 0.013292612507939339,
+ "learning_rate": 0.0006,
+ "loss": 4.000730514526367,
+ "step": 7113
+ },
+ {
+ "epoch": 98.81048034934499,
+ "grad_norm": 0.013189040124416351,
+ "learning_rate": 0.0006,
+ "loss": 4.087883949279785,
+ "step": 7114
+ },
+ {
+ "epoch": 98.82445414847162,
+ "grad_norm": 0.013953711837530136,
+ "learning_rate": 0.0006,
+ "loss": 4.051388740539551,
+ "step": 7115
+ },
+ {
+ "epoch": 98.83842794759825,
+ "grad_norm": 0.014403220266103745,
+ "learning_rate": 0.0006,
+ "loss": 3.9786248207092285,
+ "step": 7116
+ },
+ {
+ "epoch": 98.8524017467249,
+ "grad_norm": 0.013747540302574635,
+ "learning_rate": 0.0006,
+ "loss": 3.9938254356384277,
+ "step": 7117
+ },
+ {
+ "epoch": 98.86637554585153,
+ "grad_norm": 0.014865634962916374,
+ "learning_rate": 0.0006,
+ "loss": 4.075680255889893,
+ "step": 7118
+ },
+ {
+ "epoch": 98.88034934497817,
+ "grad_norm": 0.015241672284901142,
+ "learning_rate": 0.0006,
+ "loss": 3.955733299255371,
+ "step": 7119
+ },
+ {
+ "epoch": 98.8943231441048,
+ "grad_norm": 0.013307510875165462,
+ "learning_rate": 0.0006,
+ "loss": 3.94956374168396,
+ "step": 7120
+ },
+ {
+ "epoch": 98.90829694323143,
+ "grad_norm": 0.013500694185495377,
+ "learning_rate": 0.0006,
+ "loss": 4.120021343231201,
+ "step": 7121
+ },
+ {
+ "epoch": 98.92227074235808,
+ "grad_norm": 0.015430806204676628,
+ "learning_rate": 0.0006,
+ "loss": 4.108362674713135,
+ "step": 7122
+ },
+ {
+ "epoch": 98.93624454148471,
+ "grad_norm": 0.015447405166924,
+ "learning_rate": 0.0006,
+ "loss": 4.040914058685303,
+ "step": 7123
+ },
+ {
+ "epoch": 98.95021834061136,
+ "grad_norm": 0.014040131121873856,
+ "learning_rate": 0.0006,
+ "loss": 4.086688995361328,
+ "step": 7124
+ },
+ {
+ "epoch": 98.96419213973799,
+ "grad_norm": 0.0136969443410635,
+ "learning_rate": 0.0006,
+ "loss": 4.083828926086426,
+ "step": 7125
+ },
+ {
+ "epoch": 98.97816593886463,
+ "grad_norm": 0.01413224171847105,
+ "learning_rate": 0.0006,
+ "loss": 4.142146110534668,
+ "step": 7126
+ },
+ {
+ "epoch": 98.99213973799127,
+ "grad_norm": 0.012932452373206615,
+ "learning_rate": 0.0006,
+ "loss": 4.1193342208862305,
+ "step": 7127
+ },
+ {
+ "epoch": 99.0,
+ "grad_norm": 0.01586783304810524,
+ "learning_rate": 0.0006,
+ "loss": 4.191825866699219,
+ "step": 7128
+ },
+ {
+ "epoch": 99.0,
+ "eval_loss": 4.550753116607666,
+ "eval_runtime": 56.4315,
+ "eval_samples_per_second": 43.274,
+ "eval_steps_per_second": 1.364,
+ "step": 7128
+ },
+ {
+ "epoch": 99.01397379912663,
+ "grad_norm": 0.016847660765051842,
+ "learning_rate": 0.0006,
+ "loss": 3.9874870777130127,
+ "step": 7129
+ },
+ {
+ "epoch": 99.02794759825328,
+ "grad_norm": 0.016426295042037964,
+ "learning_rate": 0.0006,
+ "loss": 4.108767032623291,
+ "step": 7130
+ },
+ {
+ "epoch": 99.04192139737991,
+ "grad_norm": 0.01687243953347206,
+ "learning_rate": 0.0006,
+ "loss": 3.9658308029174805,
+ "step": 7131
+ },
+ {
+ "epoch": 99.05589519650655,
+ "grad_norm": 0.016468465328216553,
+ "learning_rate": 0.0006,
+ "loss": 3.91381573677063,
+ "step": 7132
+ },
+ {
+ "epoch": 99.06986899563319,
+ "grad_norm": 0.013777337968349457,
+ "learning_rate": 0.0006,
+ "loss": 3.9960741996765137,
+ "step": 7133
+ },
+ {
+ "epoch": 99.08384279475983,
+ "grad_norm": 0.014061779715120792,
+ "learning_rate": 0.0006,
+ "loss": 4.069479942321777,
+ "step": 7134
+ },
+ {
+ "epoch": 99.09781659388646,
+ "grad_norm": 0.01513410173356533,
+ "learning_rate": 0.0006,
+ "loss": 4.145164489746094,
+ "step": 7135
+ },
+ {
+ "epoch": 99.1117903930131,
+ "grad_norm": 0.01567850261926651,
+ "learning_rate": 0.0006,
+ "loss": 4.016868591308594,
+ "step": 7136
+ },
+ {
+ "epoch": 99.12576419213974,
+ "grad_norm": 0.01662704534828663,
+ "learning_rate": 0.0006,
+ "loss": 3.966663360595703,
+ "step": 7137
+ },
+ {
+ "epoch": 99.13973799126637,
+ "grad_norm": 0.01690712384879589,
+ "learning_rate": 0.0006,
+ "loss": 4.003656387329102,
+ "step": 7138
+ },
+ {
+ "epoch": 99.15371179039302,
+ "grad_norm": 0.01541030965745449,
+ "learning_rate": 0.0006,
+ "loss": 3.9990856647491455,
+ "step": 7139
+ },
+ {
+ "epoch": 99.16768558951965,
+ "grad_norm": 0.014438044279813766,
+ "learning_rate": 0.0006,
+ "loss": 4.089632034301758,
+ "step": 7140
+ },
+ {
+ "epoch": 99.18165938864628,
+ "grad_norm": 0.01470417995005846,
+ "learning_rate": 0.0006,
+ "loss": 4.018681049346924,
+ "step": 7141
+ },
+ {
+ "epoch": 99.19563318777293,
+ "grad_norm": 0.015567616559565067,
+ "learning_rate": 0.0006,
+ "loss": 4.067337989807129,
+ "step": 7142
+ },
+ {
+ "epoch": 99.20960698689956,
+ "grad_norm": 0.01784040965139866,
+ "learning_rate": 0.0006,
+ "loss": 4.083293914794922,
+ "step": 7143
+ },
+ {
+ "epoch": 99.2235807860262,
+ "grad_norm": 0.017351210117340088,
+ "learning_rate": 0.0006,
+ "loss": 4.115224361419678,
+ "step": 7144
+ },
+ {
+ "epoch": 99.23755458515284,
+ "grad_norm": 0.014561209827661514,
+ "learning_rate": 0.0006,
+ "loss": 4.1001057624816895,
+ "step": 7145
+ },
+ {
+ "epoch": 99.25152838427948,
+ "grad_norm": 0.013986383564770222,
+ "learning_rate": 0.0006,
+ "loss": 3.9758031368255615,
+ "step": 7146
+ },
+ {
+ "epoch": 99.26550218340611,
+ "grad_norm": 0.014053978025913239,
+ "learning_rate": 0.0006,
+ "loss": 4.1420369148254395,
+ "step": 7147
+ },
+ {
+ "epoch": 99.27947598253274,
+ "grad_norm": 0.014878278598189354,
+ "learning_rate": 0.0006,
+ "loss": 4.018189430236816,
+ "step": 7148
+ },
+ {
+ "epoch": 99.29344978165939,
+ "grad_norm": 0.014634845778346062,
+ "learning_rate": 0.0006,
+ "loss": 4.062045574188232,
+ "step": 7149
+ },
+ {
+ "epoch": 99.30742358078602,
+ "grad_norm": 0.014713970944285393,
+ "learning_rate": 0.0006,
+ "loss": 4.017163276672363,
+ "step": 7150
+ },
+ {
+ "epoch": 99.32139737991267,
+ "grad_norm": 0.01589985005557537,
+ "learning_rate": 0.0006,
+ "loss": 4.093014717102051,
+ "step": 7151
+ },
+ {
+ "epoch": 99.3353711790393,
+ "grad_norm": 0.016232378780841827,
+ "learning_rate": 0.0006,
+ "loss": 4.029753684997559,
+ "step": 7152
+ },
+ {
+ "epoch": 99.34934497816595,
+ "grad_norm": 0.016340898349881172,
+ "learning_rate": 0.0006,
+ "loss": 4.056434631347656,
+ "step": 7153
+ },
+ {
+ "epoch": 99.36331877729258,
+ "grad_norm": 0.01587427407503128,
+ "learning_rate": 0.0006,
+ "loss": 3.932020425796509,
+ "step": 7154
+ },
+ {
+ "epoch": 99.37729257641921,
+ "grad_norm": 0.014977093786001205,
+ "learning_rate": 0.0006,
+ "loss": 4.015125751495361,
+ "step": 7155
+ },
+ {
+ "epoch": 99.39126637554585,
+ "grad_norm": 0.014813411049544811,
+ "learning_rate": 0.0006,
+ "loss": 4.088035583496094,
+ "step": 7156
+ },
+ {
+ "epoch": 99.40524017467249,
+ "grad_norm": 0.014407774433493614,
+ "learning_rate": 0.0006,
+ "loss": 3.9979054927825928,
+ "step": 7157
+ },
+ {
+ "epoch": 99.41921397379913,
+ "grad_norm": 0.013969994150102139,
+ "learning_rate": 0.0006,
+ "loss": 4.180194854736328,
+ "step": 7158
+ },
+ {
+ "epoch": 99.43318777292576,
+ "grad_norm": 0.0159455519169569,
+ "learning_rate": 0.0006,
+ "loss": 4.049341201782227,
+ "step": 7159
+ },
+ {
+ "epoch": 99.44716157205241,
+ "grad_norm": 0.016426580026745796,
+ "learning_rate": 0.0006,
+ "loss": 4.01624059677124,
+ "step": 7160
+ },
+ {
+ "epoch": 99.46113537117904,
+ "grad_norm": 0.014921349473297596,
+ "learning_rate": 0.0006,
+ "loss": 4.003138542175293,
+ "step": 7161
+ },
+ {
+ "epoch": 99.47510917030567,
+ "grad_norm": 0.014607159420847893,
+ "learning_rate": 0.0006,
+ "loss": 3.993896484375,
+ "step": 7162
+ },
+ {
+ "epoch": 99.48908296943232,
+ "grad_norm": 0.014638705179095268,
+ "learning_rate": 0.0006,
+ "loss": 4.015478134155273,
+ "step": 7163
+ },
+ {
+ "epoch": 99.50305676855895,
+ "grad_norm": 0.01346661988645792,
+ "learning_rate": 0.0006,
+ "loss": 3.9999020099639893,
+ "step": 7164
+ },
+ {
+ "epoch": 99.5170305676856,
+ "grad_norm": 0.013515726663172245,
+ "learning_rate": 0.0006,
+ "loss": 4.045872688293457,
+ "step": 7165
+ },
+ {
+ "epoch": 99.53100436681223,
+ "grad_norm": 0.013773965649306774,
+ "learning_rate": 0.0006,
+ "loss": 4.094198703765869,
+ "step": 7166
+ },
+ {
+ "epoch": 99.54497816593886,
+ "grad_norm": 0.014282289892435074,
+ "learning_rate": 0.0006,
+ "loss": 4.091470718383789,
+ "step": 7167
+ },
+ {
+ "epoch": 99.5589519650655,
+ "grad_norm": 0.015272347256541252,
+ "learning_rate": 0.0006,
+ "loss": 4.026535987854004,
+ "step": 7168
+ },
+ {
+ "epoch": 99.57292576419214,
+ "grad_norm": 0.015618950128555298,
+ "learning_rate": 0.0006,
+ "loss": 4.006448745727539,
+ "step": 7169
+ },
+ {
+ "epoch": 99.58689956331878,
+ "grad_norm": 0.015366379171609879,
+ "learning_rate": 0.0006,
+ "loss": 4.025970458984375,
+ "step": 7170
+ },
+ {
+ "epoch": 99.60087336244541,
+ "grad_norm": 0.013261756859719753,
+ "learning_rate": 0.0006,
+ "loss": 3.980401039123535,
+ "step": 7171
+ },
+ {
+ "epoch": 99.61484716157206,
+ "grad_norm": 0.013438165187835693,
+ "learning_rate": 0.0006,
+ "loss": 4.0106096267700195,
+ "step": 7172
+ },
+ {
+ "epoch": 99.62882096069869,
+ "grad_norm": 0.015591824427247047,
+ "learning_rate": 0.0006,
+ "loss": 4.0458784103393555,
+ "step": 7173
+ },
+ {
+ "epoch": 99.64279475982532,
+ "grad_norm": 0.014863507822155952,
+ "learning_rate": 0.0006,
+ "loss": 4.014740943908691,
+ "step": 7174
+ },
+ {
+ "epoch": 99.65676855895197,
+ "grad_norm": 0.012508790008723736,
+ "learning_rate": 0.0006,
+ "loss": 4.022995948791504,
+ "step": 7175
+ },
+ {
+ "epoch": 99.6707423580786,
+ "grad_norm": 0.013107656501233578,
+ "learning_rate": 0.0006,
+ "loss": 4.085544109344482,
+ "step": 7176
+ },
+ {
+ "epoch": 99.68471615720524,
+ "grad_norm": 0.013550374656915665,
+ "learning_rate": 0.0006,
+ "loss": 4.1239237785339355,
+ "step": 7177
+ },
+ {
+ "epoch": 99.69868995633188,
+ "grad_norm": 0.012947923503816128,
+ "learning_rate": 0.0006,
+ "loss": 4.01287317276001,
+ "step": 7178
+ },
+ {
+ "epoch": 99.71266375545852,
+ "grad_norm": 0.015236291103065014,
+ "learning_rate": 0.0006,
+ "loss": 4.075140476226807,
+ "step": 7179
+ },
+ {
+ "epoch": 99.72663755458515,
+ "grad_norm": 0.014862542040646076,
+ "learning_rate": 0.0006,
+ "loss": 4.111439228057861,
+ "step": 7180
+ },
+ {
+ "epoch": 99.74061135371178,
+ "grad_norm": 0.013883531093597412,
+ "learning_rate": 0.0006,
+ "loss": 3.9860856533050537,
+ "step": 7181
+ },
+ {
+ "epoch": 99.75458515283843,
+ "grad_norm": 0.01507870014756918,
+ "learning_rate": 0.0006,
+ "loss": 4.096504211425781,
+ "step": 7182
+ },
+ {
+ "epoch": 99.76855895196506,
+ "grad_norm": 0.015378627926111221,
+ "learning_rate": 0.0006,
+ "loss": 4.034113883972168,
+ "step": 7183
+ },
+ {
+ "epoch": 99.78253275109171,
+ "grad_norm": 0.013696327805519104,
+ "learning_rate": 0.0006,
+ "loss": 4.026954650878906,
+ "step": 7184
+ },
+ {
+ "epoch": 99.79650655021834,
+ "grad_norm": 0.013519738800823689,
+ "learning_rate": 0.0006,
+ "loss": 4.040035247802734,
+ "step": 7185
+ },
+ {
+ "epoch": 99.81048034934499,
+ "grad_norm": 0.015677500516176224,
+ "learning_rate": 0.0006,
+ "loss": 4.057739734649658,
+ "step": 7186
+ },
+ {
+ "epoch": 99.82445414847162,
+ "grad_norm": 0.014739335514605045,
+ "learning_rate": 0.0006,
+ "loss": 4.022522449493408,
+ "step": 7187
+ },
+ {
+ "epoch": 99.83842794759825,
+ "grad_norm": 0.014079970307648182,
+ "learning_rate": 0.0006,
+ "loss": 3.9286317825317383,
+ "step": 7188
+ },
+ {
+ "epoch": 99.8524017467249,
+ "grad_norm": 0.014899874106049538,
+ "learning_rate": 0.0006,
+ "loss": 4.081925868988037,
+ "step": 7189
+ },
+ {
+ "epoch": 99.86637554585153,
+ "grad_norm": 0.014767617918550968,
+ "learning_rate": 0.0006,
+ "loss": 4.029382705688477,
+ "step": 7190
+ },
+ {
+ "epoch": 99.88034934497817,
+ "grad_norm": 0.014914393424987793,
+ "learning_rate": 0.0006,
+ "loss": 4.056460857391357,
+ "step": 7191
+ },
+ {
+ "epoch": 99.8943231441048,
+ "grad_norm": 0.01720450073480606,
+ "learning_rate": 0.0006,
+ "loss": 4.0184478759765625,
+ "step": 7192
+ },
+ {
+ "epoch": 99.90829694323143,
+ "grad_norm": 0.016545087099075317,
+ "learning_rate": 0.0006,
+ "loss": 3.9638831615448,
+ "step": 7193
+ },
+ {
+ "epoch": 99.92227074235808,
+ "grad_norm": 0.014489716850221157,
+ "learning_rate": 0.0006,
+ "loss": 4.0267653465271,
+ "step": 7194
+ },
+ {
+ "epoch": 99.93624454148471,
+ "grad_norm": 0.01438206061720848,
+ "learning_rate": 0.0006,
+ "loss": 4.212850570678711,
+ "step": 7195
+ },
+ {
+ "epoch": 99.95021834061136,
+ "grad_norm": 0.017636865377426147,
+ "learning_rate": 0.0006,
+ "loss": 3.942308187484741,
+ "step": 7196
+ },
+ {
+ "epoch": 99.96419213973799,
+ "grad_norm": 0.019075388088822365,
+ "learning_rate": 0.0006,
+ "loss": 4.026660919189453,
+ "step": 7197
+ },
+ {
+ "epoch": 99.97816593886463,
+ "grad_norm": 0.016569023951888084,
+ "learning_rate": 0.0006,
+ "loss": 4.018840789794922,
+ "step": 7198
+ },
+ {
+ "epoch": 99.99213973799127,
+ "grad_norm": 0.014842798933386803,
+ "learning_rate": 0.0006,
+ "loss": 4.041904449462891,
+ "step": 7199
+ },
+ {
+ "epoch": 100.0,
+ "grad_norm": 0.0167684406042099,
+ "learning_rate": 0.0006,
+ "loss": 3.9079346656799316,
+ "step": 7200
+ },
+ {
+ "epoch": 100.0,
+ "eval_loss": 4.585540771484375,
+ "eval_runtime": 56.5095,
+ "eval_samples_per_second": 43.214,
+ "eval_steps_per_second": 1.363,
+ "step": 7200
+ },
+ {
+ "epoch": 100.01397379912663,
+ "grad_norm": 0.016783876344561577,
+ "learning_rate": 0.0006,
+ "loss": 3.998840808868408,
+ "step": 7201
+ },
+ {
+ "epoch": 100.02794759825328,
+ "grad_norm": 0.015306975692510605,
+ "learning_rate": 0.0006,
+ "loss": 4.161236763000488,
+ "step": 7202
+ },
+ {
+ "epoch": 100.04192139737991,
+ "grad_norm": 0.016536490991711617,
+ "learning_rate": 0.0006,
+ "loss": 4.044086933135986,
+ "step": 7203
+ },
+ {
+ "epoch": 100.05589519650655,
+ "grad_norm": 0.016115620732307434,
+ "learning_rate": 0.0006,
+ "loss": 3.9210472106933594,
+ "step": 7204
+ },
+ {
+ "epoch": 100.06986899563319,
+ "grad_norm": 0.014647277072072029,
+ "learning_rate": 0.0006,
+ "loss": 3.9738388061523438,
+ "step": 7205
+ },
+ {
+ "epoch": 100.08384279475983,
+ "grad_norm": 0.013562940992414951,
+ "learning_rate": 0.0006,
+ "loss": 4.093536376953125,
+ "step": 7206
+ },
+ {
+ "epoch": 100.09781659388646,
+ "grad_norm": 0.013523460365831852,
+ "learning_rate": 0.0006,
+ "loss": 3.985286235809326,
+ "step": 7207
+ },
+ {
+ "epoch": 100.1117903930131,
+ "grad_norm": 0.01463227067142725,
+ "learning_rate": 0.0006,
+ "loss": 3.8212950229644775,
+ "step": 7208
+ },
+ {
+ "epoch": 100.12576419213974,
+ "grad_norm": 0.015067536383867264,
+ "learning_rate": 0.0006,
+ "loss": 4.050893783569336,
+ "step": 7209
+ },
+ {
+ "epoch": 100.13973799126637,
+ "grad_norm": 0.015035216696560383,
+ "learning_rate": 0.0006,
+ "loss": 4.108796119689941,
+ "step": 7210
+ },
+ {
+ "epoch": 100.15371179039302,
+ "grad_norm": 0.01694556325674057,
+ "learning_rate": 0.0006,
+ "loss": 4.001124382019043,
+ "step": 7211
+ },
+ {
+ "epoch": 100.16768558951965,
+ "grad_norm": 0.01810646802186966,
+ "learning_rate": 0.0006,
+ "loss": 4.122286319732666,
+ "step": 7212
+ },
+ {
+ "epoch": 100.18165938864628,
+ "grad_norm": 0.019156092777848244,
+ "learning_rate": 0.0006,
+ "loss": 4.087690353393555,
+ "step": 7213
+ },
+ {
+ "epoch": 100.19563318777293,
+ "grad_norm": 0.018056577071547508,
+ "learning_rate": 0.0006,
+ "loss": 4.135015487670898,
+ "step": 7214
+ },
+ {
+ "epoch": 100.20960698689956,
+ "grad_norm": 0.017466023564338684,
+ "learning_rate": 0.0006,
+ "loss": 4.028027534484863,
+ "step": 7215
+ },
+ {
+ "epoch": 100.2235807860262,
+ "grad_norm": 0.017239956185221672,
+ "learning_rate": 0.0006,
+ "loss": 4.128888130187988,
+ "step": 7216
+ },
+ {
+ "epoch": 100.23755458515284,
+ "grad_norm": 0.016968458890914917,
+ "learning_rate": 0.0006,
+ "loss": 3.998007297515869,
+ "step": 7217
+ },
+ {
+ "epoch": 100.25152838427948,
+ "grad_norm": 0.016718817874789238,
+ "learning_rate": 0.0006,
+ "loss": 4.125210762023926,
+ "step": 7218
+ },
+ {
+ "epoch": 100.26550218340611,
+ "grad_norm": 0.0164450965821743,
+ "learning_rate": 0.0006,
+ "loss": 4.000813007354736,
+ "step": 7219
+ },
+ {
+ "epoch": 100.27947598253274,
+ "grad_norm": 0.01655936799943447,
+ "learning_rate": 0.0006,
+ "loss": 3.9809513092041016,
+ "step": 7220
+ },
+ {
+ "epoch": 100.29344978165939,
+ "grad_norm": 0.015745731070637703,
+ "learning_rate": 0.0006,
+ "loss": 4.089069366455078,
+ "step": 7221
+ },
+ {
+ "epoch": 100.30742358078602,
+ "grad_norm": 0.01796283759176731,
+ "learning_rate": 0.0006,
+ "loss": 4.049221992492676,
+ "step": 7222
+ },
+ {
+ "epoch": 100.32139737991267,
+ "grad_norm": 0.018306732177734375,
+ "learning_rate": 0.0006,
+ "loss": 4.043508052825928,
+ "step": 7223
+ },
+ {
+ "epoch": 100.3353711790393,
+ "grad_norm": 0.016871308907866478,
+ "learning_rate": 0.0006,
+ "loss": 4.051634311676025,
+ "step": 7224
+ },
+ {
+ "epoch": 100.34934497816595,
+ "grad_norm": 0.016082527115941048,
+ "learning_rate": 0.0006,
+ "loss": 3.905829668045044,
+ "step": 7225
+ },
+ {
+ "epoch": 100.36331877729258,
+ "grad_norm": 0.01602019928395748,
+ "learning_rate": 0.0006,
+ "loss": 4.039409637451172,
+ "step": 7226
+ },
+ {
+ "epoch": 100.37729257641921,
+ "grad_norm": 0.01674795150756836,
+ "learning_rate": 0.0006,
+ "loss": 3.9757080078125,
+ "step": 7227
+ },
+ {
+ "epoch": 100.39126637554585,
+ "grad_norm": 0.014572949148714542,
+ "learning_rate": 0.0006,
+ "loss": 4.073253631591797,
+ "step": 7228
+ },
+ {
+ "epoch": 100.40524017467249,
+ "grad_norm": 0.015504667535424232,
+ "learning_rate": 0.0006,
+ "loss": 4.122897624969482,
+ "step": 7229
+ },
+ {
+ "epoch": 100.41921397379913,
+ "grad_norm": 0.0163279939442873,
+ "learning_rate": 0.0006,
+ "loss": 4.1488823890686035,
+ "step": 7230
+ },
+ {
+ "epoch": 100.43318777292576,
+ "grad_norm": 0.016203004866838455,
+ "learning_rate": 0.0006,
+ "loss": 4.099796295166016,
+ "step": 7231
+ },
+ {
+ "epoch": 100.44716157205241,
+ "grad_norm": 0.014418032020330429,
+ "learning_rate": 0.0006,
+ "loss": 3.994591474533081,
+ "step": 7232
+ },
+ {
+ "epoch": 100.46113537117904,
+ "grad_norm": 0.014775440096855164,
+ "learning_rate": 0.0006,
+ "loss": 4.07340145111084,
+ "step": 7233
+ },
+ {
+ "epoch": 100.47510917030567,
+ "grad_norm": 0.015920625999569893,
+ "learning_rate": 0.0006,
+ "loss": 3.985908269882202,
+ "step": 7234
+ },
+ {
+ "epoch": 100.48908296943232,
+ "grad_norm": 0.015350693836808205,
+ "learning_rate": 0.0006,
+ "loss": 4.007932662963867,
+ "step": 7235
+ },
+ {
+ "epoch": 100.50305676855895,
+ "grad_norm": 0.014701326377689838,
+ "learning_rate": 0.0006,
+ "loss": 4.151021957397461,
+ "step": 7236
+ },
+ {
+ "epoch": 100.5170305676856,
+ "grad_norm": 0.013621610589325428,
+ "learning_rate": 0.0006,
+ "loss": 3.988795280456543,
+ "step": 7237
+ },
+ {
+ "epoch": 100.53100436681223,
+ "grad_norm": 0.014496153220534325,
+ "learning_rate": 0.0006,
+ "loss": 4.019025802612305,
+ "step": 7238
+ },
+ {
+ "epoch": 100.54497816593886,
+ "grad_norm": 0.013380124233663082,
+ "learning_rate": 0.0006,
+ "loss": 4.090785980224609,
+ "step": 7239
+ },
+ {
+ "epoch": 100.5589519650655,
+ "grad_norm": 0.01378310564905405,
+ "learning_rate": 0.0006,
+ "loss": 4.158342361450195,
+ "step": 7240
+ },
+ {
+ "epoch": 100.57292576419214,
+ "grad_norm": 0.01402236521244049,
+ "learning_rate": 0.0006,
+ "loss": 4.108238220214844,
+ "step": 7241
+ },
+ {
+ "epoch": 100.58689956331878,
+ "grad_norm": 0.014660182408988476,
+ "learning_rate": 0.0006,
+ "loss": 4.052890300750732,
+ "step": 7242
+ },
+ {
+ "epoch": 100.60087336244541,
+ "grad_norm": 0.01419265940785408,
+ "learning_rate": 0.0006,
+ "loss": 4.048383712768555,
+ "step": 7243
+ },
+ {
+ "epoch": 100.61484716157206,
+ "grad_norm": 0.014536422677338123,
+ "learning_rate": 0.0006,
+ "loss": 3.938699960708618,
+ "step": 7244
+ },
+ {
+ "epoch": 100.62882096069869,
+ "grad_norm": 0.013167395256459713,
+ "learning_rate": 0.0006,
+ "loss": 4.13656759262085,
+ "step": 7245
+ },
+ {
+ "epoch": 100.64279475982532,
+ "grad_norm": 0.013563545420765877,
+ "learning_rate": 0.0006,
+ "loss": 4.013095855712891,
+ "step": 7246
+ },
+ {
+ "epoch": 100.65676855895197,
+ "grad_norm": 0.013953451067209244,
+ "learning_rate": 0.0006,
+ "loss": 4.091291427612305,
+ "step": 7247
+ },
+ {
+ "epoch": 100.6707423580786,
+ "grad_norm": 0.014537926763296127,
+ "learning_rate": 0.0006,
+ "loss": 4.041815757751465,
+ "step": 7248
+ },
+ {
+ "epoch": 100.68471615720524,
+ "grad_norm": 0.014321094378829002,
+ "learning_rate": 0.0006,
+ "loss": 4.033531665802002,
+ "step": 7249
+ },
+ {
+ "epoch": 100.69868995633188,
+ "grad_norm": 0.013944404199719429,
+ "learning_rate": 0.0006,
+ "loss": 4.119833946228027,
+ "step": 7250
+ },
+ {
+ "epoch": 100.71266375545852,
+ "grad_norm": 0.01371681783348322,
+ "learning_rate": 0.0006,
+ "loss": 3.965501546859741,
+ "step": 7251
+ },
+ {
+ "epoch": 100.72663755458515,
+ "grad_norm": 0.012983391061425209,
+ "learning_rate": 0.0006,
+ "loss": 3.987842559814453,
+ "step": 7252
+ },
+ {
+ "epoch": 100.74061135371178,
+ "grad_norm": 0.013693883083760738,
+ "learning_rate": 0.0006,
+ "loss": 4.066164493560791,
+ "step": 7253
+ },
+ {
+ "epoch": 100.75458515283843,
+ "grad_norm": 0.014391372911632061,
+ "learning_rate": 0.0006,
+ "loss": 4.095022201538086,
+ "step": 7254
+ },
+ {
+ "epoch": 100.76855895196506,
+ "grad_norm": 0.013853080570697784,
+ "learning_rate": 0.0006,
+ "loss": 4.099021911621094,
+ "step": 7255
+ },
+ {
+ "epoch": 100.78253275109171,
+ "grad_norm": 0.014468333683907986,
+ "learning_rate": 0.0006,
+ "loss": 4.040180206298828,
+ "step": 7256
+ },
+ {
+ "epoch": 100.79650655021834,
+ "grad_norm": 0.015436802990734577,
+ "learning_rate": 0.0006,
+ "loss": 4.0907087326049805,
+ "step": 7257
+ },
+ {
+ "epoch": 100.81048034934499,
+ "grad_norm": 0.014863028191030025,
+ "learning_rate": 0.0006,
+ "loss": 4.075537204742432,
+ "step": 7258
+ },
+ {
+ "epoch": 100.82445414847162,
+ "grad_norm": 0.014109591953456402,
+ "learning_rate": 0.0006,
+ "loss": 3.922210693359375,
+ "step": 7259
+ },
+ {
+ "epoch": 100.83842794759825,
+ "grad_norm": 0.014673606492578983,
+ "learning_rate": 0.0006,
+ "loss": 4.013545036315918,
+ "step": 7260
+ },
+ {
+ "epoch": 100.8524017467249,
+ "grad_norm": 0.01422286219894886,
+ "learning_rate": 0.0006,
+ "loss": 3.926832675933838,
+ "step": 7261
+ },
+ {
+ "epoch": 100.86637554585153,
+ "grad_norm": 0.013322828337550163,
+ "learning_rate": 0.0006,
+ "loss": 4.032454490661621,
+ "step": 7262
+ },
+ {
+ "epoch": 100.88034934497817,
+ "grad_norm": 0.014322307892143726,
+ "learning_rate": 0.0006,
+ "loss": 4.034879684448242,
+ "step": 7263
+ },
+ {
+ "epoch": 100.8943231441048,
+ "grad_norm": 0.013952548615634441,
+ "learning_rate": 0.0006,
+ "loss": 4.092174530029297,
+ "step": 7264
+ },
+ {
+ "epoch": 100.90829694323143,
+ "grad_norm": 0.01494069304317236,
+ "learning_rate": 0.0006,
+ "loss": 4.051094055175781,
+ "step": 7265
+ },
+ {
+ "epoch": 100.92227074235808,
+ "grad_norm": 0.01506600808352232,
+ "learning_rate": 0.0006,
+ "loss": 4.150339603424072,
+ "step": 7266
+ },
+ {
+ "epoch": 100.93624454148471,
+ "grad_norm": 0.015410789288580418,
+ "learning_rate": 0.0006,
+ "loss": 4.029681205749512,
+ "step": 7267
+ },
+ {
+ "epoch": 100.95021834061136,
+ "grad_norm": 0.01526960451155901,
+ "learning_rate": 0.0006,
+ "loss": 3.9946770668029785,
+ "step": 7268
+ },
+ {
+ "epoch": 100.96419213973799,
+ "grad_norm": 0.01581135205924511,
+ "learning_rate": 0.0006,
+ "loss": 4.050585746765137,
+ "step": 7269
+ },
+ {
+ "epoch": 100.97816593886463,
+ "grad_norm": 0.015924792736768723,
+ "learning_rate": 0.0006,
+ "loss": 4.005122184753418,
+ "step": 7270
+ },
+ {
+ "epoch": 100.99213973799127,
+ "grad_norm": 0.01638958603143692,
+ "learning_rate": 0.0006,
+ "loss": 4.074165344238281,
+ "step": 7271
+ },
+ {
+ "epoch": 101.0,
+ "grad_norm": 0.016363589093089104,
+ "learning_rate": 0.0006,
+ "loss": 4.004770278930664,
+ "step": 7272
+ },
+ {
+ "epoch": 101.0,
+ "eval_loss": 4.521492958068848,
+ "eval_runtime": 56.7059,
+ "eval_samples_per_second": 43.064,
+ "eval_steps_per_second": 1.358,
+ "step": 7272
+ },
+ {
+ "epoch": 101.01397379912663,
+ "grad_norm": 0.013911865651607513,
+ "learning_rate": 0.0006,
+ "loss": 4.073930263519287,
+ "step": 7273
+ },
+ {
+ "epoch": 101.02794759825328,
+ "grad_norm": 0.014139825478196144,
+ "learning_rate": 0.0006,
+ "loss": 4.0691142082214355,
+ "step": 7274
+ },
+ {
+ "epoch": 101.04192139737991,
+ "grad_norm": 0.015765748918056488,
+ "learning_rate": 0.0006,
+ "loss": 4.112308502197266,
+ "step": 7275
+ },
+ {
+ "epoch": 101.05589519650655,
+ "grad_norm": 0.016765642911195755,
+ "learning_rate": 0.0006,
+ "loss": 3.9341437816619873,
+ "step": 7276
+ },
+ {
+ "epoch": 101.06986899563319,
+ "grad_norm": 0.016253530979156494,
+ "learning_rate": 0.0006,
+ "loss": 3.9987592697143555,
+ "step": 7277
+ },
+ {
+ "epoch": 101.08384279475983,
+ "grad_norm": 0.015670686960220337,
+ "learning_rate": 0.0006,
+ "loss": 4.0460357666015625,
+ "step": 7278
+ },
+ {
+ "epoch": 101.09781659388646,
+ "grad_norm": 0.01603814959526062,
+ "learning_rate": 0.0006,
+ "loss": 4.005774974822998,
+ "step": 7279
+ },
+ {
+ "epoch": 101.1117903930131,
+ "grad_norm": 0.016335027292370796,
+ "learning_rate": 0.0006,
+ "loss": 3.9924376010894775,
+ "step": 7280
+ },
+ {
+ "epoch": 101.12576419213974,
+ "grad_norm": 0.017239267006516457,
+ "learning_rate": 0.0006,
+ "loss": 4.034348487854004,
+ "step": 7281
+ },
+ {
+ "epoch": 101.13973799126637,
+ "grad_norm": 0.016026441007852554,
+ "learning_rate": 0.0006,
+ "loss": 4.054862022399902,
+ "step": 7282
+ },
+ {
+ "epoch": 101.15371179039302,
+ "grad_norm": 0.015502813272178173,
+ "learning_rate": 0.0006,
+ "loss": 3.974360466003418,
+ "step": 7283
+ },
+ {
+ "epoch": 101.16768558951965,
+ "grad_norm": 0.01737619936466217,
+ "learning_rate": 0.0006,
+ "loss": 3.991811752319336,
+ "step": 7284
+ },
+ {
+ "epoch": 101.18165938864628,
+ "grad_norm": 0.01761457324028015,
+ "learning_rate": 0.0006,
+ "loss": 4.021451473236084,
+ "step": 7285
+ },
+ {
+ "epoch": 101.19563318777293,
+ "grad_norm": 0.017094692215323448,
+ "learning_rate": 0.0006,
+ "loss": 4.00285530090332,
+ "step": 7286
+ },
+ {
+ "epoch": 101.20960698689956,
+ "grad_norm": 0.0163987148553133,
+ "learning_rate": 0.0006,
+ "loss": 4.061247825622559,
+ "step": 7287
+ },
+ {
+ "epoch": 101.2235807860262,
+ "grad_norm": 0.015596904791891575,
+ "learning_rate": 0.0006,
+ "loss": 3.971940040588379,
+ "step": 7288
+ },
+ {
+ "epoch": 101.23755458515284,
+ "grad_norm": 0.015181072056293488,
+ "learning_rate": 0.0006,
+ "loss": 4.127270698547363,
+ "step": 7289
+ },
+ {
+ "epoch": 101.25152838427948,
+ "grad_norm": 0.014719299040734768,
+ "learning_rate": 0.0006,
+ "loss": 4.095149993896484,
+ "step": 7290
+ },
+ {
+ "epoch": 101.26550218340611,
+ "grad_norm": 0.01476318296045065,
+ "learning_rate": 0.0006,
+ "loss": 4.091351509094238,
+ "step": 7291
+ },
+ {
+ "epoch": 101.27947598253274,
+ "grad_norm": 0.014008177444338799,
+ "learning_rate": 0.0006,
+ "loss": 3.9897727966308594,
+ "step": 7292
+ },
+ {
+ "epoch": 101.29344978165939,
+ "grad_norm": 0.012773942202329636,
+ "learning_rate": 0.0006,
+ "loss": 3.986448287963867,
+ "step": 7293
+ },
+ {
+ "epoch": 101.30742358078602,
+ "grad_norm": 0.013246177695691586,
+ "learning_rate": 0.0006,
+ "loss": 4.105729579925537,
+ "step": 7294
+ },
+ {
+ "epoch": 101.32139737991267,
+ "grad_norm": 0.013527474366128445,
+ "learning_rate": 0.0006,
+ "loss": 4.012033939361572,
+ "step": 7295
+ },
+ {
+ "epoch": 101.3353711790393,
+ "grad_norm": 0.013510984368622303,
+ "learning_rate": 0.0006,
+ "loss": 3.9836957454681396,
+ "step": 7296
+ },
+ {
+ "epoch": 101.34934497816595,
+ "grad_norm": 0.01403056737035513,
+ "learning_rate": 0.0006,
+ "loss": 4.100725173950195,
+ "step": 7297
+ },
+ {
+ "epoch": 101.36331877729258,
+ "grad_norm": 0.013127653859555721,
+ "learning_rate": 0.0006,
+ "loss": 3.977855682373047,
+ "step": 7298
+ },
+ {
+ "epoch": 101.37729257641921,
+ "grad_norm": 0.015479992143809795,
+ "learning_rate": 0.0006,
+ "loss": 3.97141170501709,
+ "step": 7299
+ },
+ {
+ "epoch": 101.39126637554585,
+ "grad_norm": 0.01583872176706791,
+ "learning_rate": 0.0006,
+ "loss": 4.0838518142700195,
+ "step": 7300
+ },
+ {
+ "epoch": 101.40524017467249,
+ "grad_norm": 0.015457033179700375,
+ "learning_rate": 0.0006,
+ "loss": 4.061662673950195,
+ "step": 7301
+ },
+ {
+ "epoch": 101.41921397379913,
+ "grad_norm": 0.01576954312622547,
+ "learning_rate": 0.0006,
+ "loss": 4.069742202758789,
+ "step": 7302
+ },
+ {
+ "epoch": 101.43318777292576,
+ "grad_norm": 0.015752265229821205,
+ "learning_rate": 0.0006,
+ "loss": 3.8633975982666016,
+ "step": 7303
+ },
+ {
+ "epoch": 101.44716157205241,
+ "grad_norm": 0.015041268430650234,
+ "learning_rate": 0.0006,
+ "loss": 4.118044853210449,
+ "step": 7304
+ },
+ {
+ "epoch": 101.46113537117904,
+ "grad_norm": 0.015786072239279747,
+ "learning_rate": 0.0006,
+ "loss": 4.060590744018555,
+ "step": 7305
+ },
+ {
+ "epoch": 101.47510917030567,
+ "grad_norm": 0.01664225198328495,
+ "learning_rate": 0.0006,
+ "loss": 3.9695353507995605,
+ "step": 7306
+ },
+ {
+ "epoch": 101.48908296943232,
+ "grad_norm": 0.016616834327578545,
+ "learning_rate": 0.0006,
+ "loss": 4.1170454025268555,
+ "step": 7307
+ },
+ {
+ "epoch": 101.50305676855895,
+ "grad_norm": 0.01654205098748207,
+ "learning_rate": 0.0006,
+ "loss": 3.999567985534668,
+ "step": 7308
+ },
+ {
+ "epoch": 101.5170305676856,
+ "grad_norm": 0.01749691553413868,
+ "learning_rate": 0.0006,
+ "loss": 4.1020917892456055,
+ "step": 7309
+ },
+ {
+ "epoch": 101.53100436681223,
+ "grad_norm": 0.01645602658390999,
+ "learning_rate": 0.0006,
+ "loss": 4.070579528808594,
+ "step": 7310
+ },
+ {
+ "epoch": 101.54497816593886,
+ "grad_norm": 0.015536080114543438,
+ "learning_rate": 0.0006,
+ "loss": 4.013237953186035,
+ "step": 7311
+ },
+ {
+ "epoch": 101.5589519650655,
+ "grad_norm": 0.016780823469161987,
+ "learning_rate": 0.0006,
+ "loss": 4.027227401733398,
+ "step": 7312
+ },
+ {
+ "epoch": 101.57292576419214,
+ "grad_norm": 0.017896618694067,
+ "learning_rate": 0.0006,
+ "loss": 3.95709228515625,
+ "step": 7313
+ },
+ {
+ "epoch": 101.58689956331878,
+ "grad_norm": 0.01725279912352562,
+ "learning_rate": 0.0006,
+ "loss": 3.9525325298309326,
+ "step": 7314
+ },
+ {
+ "epoch": 101.60087336244541,
+ "grad_norm": 0.015343503095209599,
+ "learning_rate": 0.0006,
+ "loss": 4.032679557800293,
+ "step": 7315
+ },
+ {
+ "epoch": 101.61484716157206,
+ "grad_norm": 0.015583384782075882,
+ "learning_rate": 0.0006,
+ "loss": 3.9139232635498047,
+ "step": 7316
+ },
+ {
+ "epoch": 101.62882096069869,
+ "grad_norm": 0.01616024412214756,
+ "learning_rate": 0.0006,
+ "loss": 3.984429121017456,
+ "step": 7317
+ },
+ {
+ "epoch": 101.64279475982532,
+ "grad_norm": 0.015488795936107635,
+ "learning_rate": 0.0006,
+ "loss": 4.001983165740967,
+ "step": 7318
+ },
+ {
+ "epoch": 101.65676855895197,
+ "grad_norm": 0.015609036199748516,
+ "learning_rate": 0.0006,
+ "loss": 4.029224395751953,
+ "step": 7319
+ },
+ {
+ "epoch": 101.6707423580786,
+ "grad_norm": 0.01512704137712717,
+ "learning_rate": 0.0006,
+ "loss": 4.138855934143066,
+ "step": 7320
+ },
+ {
+ "epoch": 101.68471615720524,
+ "grad_norm": 0.014452820643782616,
+ "learning_rate": 0.0006,
+ "loss": 4.014325141906738,
+ "step": 7321
+ },
+ {
+ "epoch": 101.69868995633188,
+ "grad_norm": 0.013438846915960312,
+ "learning_rate": 0.0006,
+ "loss": 4.038142204284668,
+ "step": 7322
+ },
+ {
+ "epoch": 101.71266375545852,
+ "grad_norm": 0.014854568056762218,
+ "learning_rate": 0.0006,
+ "loss": 4.005454063415527,
+ "step": 7323
+ },
+ {
+ "epoch": 101.72663755458515,
+ "grad_norm": 0.01406457182019949,
+ "learning_rate": 0.0006,
+ "loss": 4.163865089416504,
+ "step": 7324
+ },
+ {
+ "epoch": 101.74061135371178,
+ "grad_norm": 0.01440865732729435,
+ "learning_rate": 0.0006,
+ "loss": 4.050951957702637,
+ "step": 7325
+ },
+ {
+ "epoch": 101.75458515283843,
+ "grad_norm": 0.013415827415883541,
+ "learning_rate": 0.0006,
+ "loss": 4.061484336853027,
+ "step": 7326
+ },
+ {
+ "epoch": 101.76855895196506,
+ "grad_norm": 0.012345070950686932,
+ "learning_rate": 0.0006,
+ "loss": 4.113116264343262,
+ "step": 7327
+ },
+ {
+ "epoch": 101.78253275109171,
+ "grad_norm": 0.01455951202660799,
+ "learning_rate": 0.0006,
+ "loss": 4.1283650398254395,
+ "step": 7328
+ },
+ {
+ "epoch": 101.79650655021834,
+ "grad_norm": 0.015864135697484016,
+ "learning_rate": 0.0006,
+ "loss": 4.084014892578125,
+ "step": 7329
+ },
+ {
+ "epoch": 101.81048034934499,
+ "grad_norm": 0.015105332247912884,
+ "learning_rate": 0.0006,
+ "loss": 4.136471748352051,
+ "step": 7330
+ },
+ {
+ "epoch": 101.82445414847162,
+ "grad_norm": 0.014927559532225132,
+ "learning_rate": 0.0006,
+ "loss": 4.039514541625977,
+ "step": 7331
+ },
+ {
+ "epoch": 101.83842794759825,
+ "grad_norm": 0.014508268795907497,
+ "learning_rate": 0.0006,
+ "loss": 4.052594184875488,
+ "step": 7332
+ },
+ {
+ "epoch": 101.8524017467249,
+ "grad_norm": 0.01513114757835865,
+ "learning_rate": 0.0006,
+ "loss": 4.084187984466553,
+ "step": 7333
+ },
+ {
+ "epoch": 101.86637554585153,
+ "grad_norm": 0.014415159821510315,
+ "learning_rate": 0.0006,
+ "loss": 3.8839468955993652,
+ "step": 7334
+ },
+ {
+ "epoch": 101.88034934497817,
+ "grad_norm": 0.013508095405995846,
+ "learning_rate": 0.0006,
+ "loss": 4.091848850250244,
+ "step": 7335
+ },
+ {
+ "epoch": 101.8943231441048,
+ "grad_norm": 0.013396956957876682,
+ "learning_rate": 0.0006,
+ "loss": 3.983945369720459,
+ "step": 7336
+ },
+ {
+ "epoch": 101.90829694323143,
+ "grad_norm": 0.013772518374025822,
+ "learning_rate": 0.0006,
+ "loss": 4.029302597045898,
+ "step": 7337
+ },
+ {
+ "epoch": 101.92227074235808,
+ "grad_norm": 0.01479576900601387,
+ "learning_rate": 0.0006,
+ "loss": 3.8791067600250244,
+ "step": 7338
+ },
+ {
+ "epoch": 101.93624454148471,
+ "grad_norm": 0.014177192002534866,
+ "learning_rate": 0.0006,
+ "loss": 4.044843673706055,
+ "step": 7339
+ },
+ {
+ "epoch": 101.95021834061136,
+ "grad_norm": 0.014353828504681587,
+ "learning_rate": 0.0006,
+ "loss": 4.112372398376465,
+ "step": 7340
+ },
+ {
+ "epoch": 101.96419213973799,
+ "grad_norm": 0.013797440566122532,
+ "learning_rate": 0.0006,
+ "loss": 4.2181396484375,
+ "step": 7341
+ },
+ {
+ "epoch": 101.97816593886463,
+ "grad_norm": 0.013296752236783504,
+ "learning_rate": 0.0006,
+ "loss": 4.015590190887451,
+ "step": 7342
+ },
+ {
+ "epoch": 101.99213973799127,
+ "grad_norm": 0.014963190071284771,
+ "learning_rate": 0.0006,
+ "loss": 3.9683375358581543,
+ "step": 7343
+ },
+ {
+ "epoch": 102.0,
+ "grad_norm": 0.016375407576560974,
+ "learning_rate": 0.0006,
+ "loss": 4.058782577514648,
+ "step": 7344
+ },
+ {
+ "epoch": 102.0,
+ "eval_loss": 4.608489036560059,
+ "eval_runtime": 57.1021,
+ "eval_samples_per_second": 42.765,
+ "eval_steps_per_second": 1.348,
+ "step": 7344
+ },
+ {
+ "epoch": 102.01397379912663,
+ "grad_norm": 0.015222231857478619,
+ "learning_rate": 0.0006,
+ "loss": 3.9541208744049072,
+ "step": 7345
+ },
+ {
+ "epoch": 102.02794759825328,
+ "grad_norm": 0.015121257863938808,
+ "learning_rate": 0.0006,
+ "loss": 4.031475067138672,
+ "step": 7346
+ },
+ {
+ "epoch": 102.04192139737991,
+ "grad_norm": 0.015098890289664268,
+ "learning_rate": 0.0006,
+ "loss": 4.025700569152832,
+ "step": 7347
+ },
+ {
+ "epoch": 102.05589519650655,
+ "grad_norm": 0.016747184097766876,
+ "learning_rate": 0.0006,
+ "loss": 4.026192665100098,
+ "step": 7348
+ },
+ {
+ "epoch": 102.06986899563319,
+ "grad_norm": 0.015377930365502834,
+ "learning_rate": 0.0006,
+ "loss": 4.099573612213135,
+ "step": 7349
+ },
+ {
+ "epoch": 102.08384279475983,
+ "grad_norm": 0.014129774644970894,
+ "learning_rate": 0.0006,
+ "loss": 3.9992661476135254,
+ "step": 7350
+ },
+ {
+ "epoch": 102.09781659388646,
+ "grad_norm": 0.016497403383255005,
+ "learning_rate": 0.0006,
+ "loss": 4.038721561431885,
+ "step": 7351
+ },
+ {
+ "epoch": 102.1117903930131,
+ "grad_norm": 0.01801230199635029,
+ "learning_rate": 0.0006,
+ "loss": 4.024539947509766,
+ "step": 7352
+ },
+ {
+ "epoch": 102.12576419213974,
+ "grad_norm": 0.016215801239013672,
+ "learning_rate": 0.0006,
+ "loss": 3.990530490875244,
+ "step": 7353
+ },
+ {
+ "epoch": 102.13973799126637,
+ "grad_norm": 0.015095847658813,
+ "learning_rate": 0.0006,
+ "loss": 4.039212226867676,
+ "step": 7354
+ },
+ {
+ "epoch": 102.15371179039302,
+ "grad_norm": 0.014441692270338535,
+ "learning_rate": 0.0006,
+ "loss": 4.033042907714844,
+ "step": 7355
+ },
+ {
+ "epoch": 102.16768558951965,
+ "grad_norm": 0.015459827147424221,
+ "learning_rate": 0.0006,
+ "loss": 3.965804100036621,
+ "step": 7356
+ },
+ {
+ "epoch": 102.18165938864628,
+ "grad_norm": 0.01581769622862339,
+ "learning_rate": 0.0006,
+ "loss": 3.9772019386291504,
+ "step": 7357
+ },
+ {
+ "epoch": 102.19563318777293,
+ "grad_norm": 0.014575488865375519,
+ "learning_rate": 0.0006,
+ "loss": 4.005141735076904,
+ "step": 7358
+ },
+ {
+ "epoch": 102.20960698689956,
+ "grad_norm": 0.01389391254633665,
+ "learning_rate": 0.0006,
+ "loss": 3.950014114379883,
+ "step": 7359
+ },
+ {
+ "epoch": 102.2235807860262,
+ "grad_norm": 0.013461007736623287,
+ "learning_rate": 0.0006,
+ "loss": 4.047972679138184,
+ "step": 7360
+ },
+ {
+ "epoch": 102.23755458515284,
+ "grad_norm": 0.015122898854315281,
+ "learning_rate": 0.0006,
+ "loss": 4.06884241104126,
+ "step": 7361
+ },
+ {
+ "epoch": 102.25152838427948,
+ "grad_norm": 0.018046488985419273,
+ "learning_rate": 0.0006,
+ "loss": 3.831460952758789,
+ "step": 7362
+ },
+ {
+ "epoch": 102.26550218340611,
+ "grad_norm": 0.01921425573527813,
+ "learning_rate": 0.0006,
+ "loss": 3.982016086578369,
+ "step": 7363
+ },
+ {
+ "epoch": 102.27947598253274,
+ "grad_norm": 0.021769993007183075,
+ "learning_rate": 0.0006,
+ "loss": 4.113831520080566,
+ "step": 7364
+ },
+ {
+ "epoch": 102.29344978165939,
+ "grad_norm": 0.02255893312394619,
+ "learning_rate": 0.0006,
+ "loss": 3.9648423194885254,
+ "step": 7365
+ },
+ {
+ "epoch": 102.30742358078602,
+ "grad_norm": 0.01851275935769081,
+ "learning_rate": 0.0006,
+ "loss": 3.963744640350342,
+ "step": 7366
+ },
+ {
+ "epoch": 102.32139737991267,
+ "grad_norm": 0.01630641333758831,
+ "learning_rate": 0.0006,
+ "loss": 3.893911600112915,
+ "step": 7367
+ },
+ {
+ "epoch": 102.3353711790393,
+ "grad_norm": 0.01931816339492798,
+ "learning_rate": 0.0006,
+ "loss": 3.9658846855163574,
+ "step": 7368
+ },
+ {
+ "epoch": 102.34934497816595,
+ "grad_norm": 0.019413312897086143,
+ "learning_rate": 0.0006,
+ "loss": 4.00821590423584,
+ "step": 7369
+ },
+ {
+ "epoch": 102.36331877729258,
+ "grad_norm": 0.01617329567670822,
+ "learning_rate": 0.0006,
+ "loss": 4.049162864685059,
+ "step": 7370
+ },
+ {
+ "epoch": 102.37729257641921,
+ "grad_norm": 0.015705779194831848,
+ "learning_rate": 0.0006,
+ "loss": 3.9896233081817627,
+ "step": 7371
+ },
+ {
+ "epoch": 102.39126637554585,
+ "grad_norm": 0.015586525201797485,
+ "learning_rate": 0.0006,
+ "loss": 3.974888801574707,
+ "step": 7372
+ },
+ {
+ "epoch": 102.40524017467249,
+ "grad_norm": 0.015439536422491074,
+ "learning_rate": 0.0006,
+ "loss": 3.9923813343048096,
+ "step": 7373
+ },
+ {
+ "epoch": 102.41921397379913,
+ "grad_norm": 0.015753930434584618,
+ "learning_rate": 0.0006,
+ "loss": 3.986898899078369,
+ "step": 7374
+ },
+ {
+ "epoch": 102.43318777292576,
+ "grad_norm": 0.014091573655605316,
+ "learning_rate": 0.0006,
+ "loss": 4.128416061401367,
+ "step": 7375
+ },
+ {
+ "epoch": 102.44716157205241,
+ "grad_norm": 0.01459917426109314,
+ "learning_rate": 0.0006,
+ "loss": 3.974200487136841,
+ "step": 7376
+ },
+ {
+ "epoch": 102.46113537117904,
+ "grad_norm": 0.014099135994911194,
+ "learning_rate": 0.0006,
+ "loss": 4.034481525421143,
+ "step": 7377
+ },
+ {
+ "epoch": 102.47510917030567,
+ "grad_norm": 0.013483141548931599,
+ "learning_rate": 0.0006,
+ "loss": 3.901756763458252,
+ "step": 7378
+ },
+ {
+ "epoch": 102.48908296943232,
+ "grad_norm": 0.013499923050403595,
+ "learning_rate": 0.0006,
+ "loss": 3.965855598449707,
+ "step": 7379
+ },
+ {
+ "epoch": 102.50305676855895,
+ "grad_norm": 0.01356455497443676,
+ "learning_rate": 0.0006,
+ "loss": 3.8891913890838623,
+ "step": 7380
+ },
+ {
+ "epoch": 102.5170305676856,
+ "grad_norm": 0.013884859159588814,
+ "learning_rate": 0.0006,
+ "loss": 4.171500205993652,
+ "step": 7381
+ },
+ {
+ "epoch": 102.53100436681223,
+ "grad_norm": 0.014240458607673645,
+ "learning_rate": 0.0006,
+ "loss": 4.021063804626465,
+ "step": 7382
+ },
+ {
+ "epoch": 102.54497816593886,
+ "grad_norm": 0.015500220470130444,
+ "learning_rate": 0.0006,
+ "loss": 4.073713302612305,
+ "step": 7383
+ },
+ {
+ "epoch": 102.5589519650655,
+ "grad_norm": 0.01415219809859991,
+ "learning_rate": 0.0006,
+ "loss": 4.095281600952148,
+ "step": 7384
+ },
+ {
+ "epoch": 102.57292576419214,
+ "grad_norm": 0.01237252727150917,
+ "learning_rate": 0.0006,
+ "loss": 3.929661750793457,
+ "step": 7385
+ },
+ {
+ "epoch": 102.58689956331878,
+ "grad_norm": 0.013517644256353378,
+ "learning_rate": 0.0006,
+ "loss": 4.0342583656311035,
+ "step": 7386
+ },
+ {
+ "epoch": 102.60087336244541,
+ "grad_norm": 0.013656767085194588,
+ "learning_rate": 0.0006,
+ "loss": 4.05634880065918,
+ "step": 7387
+ },
+ {
+ "epoch": 102.61484716157206,
+ "grad_norm": 0.01358469296246767,
+ "learning_rate": 0.0006,
+ "loss": 4.069542407989502,
+ "step": 7388
+ },
+ {
+ "epoch": 102.62882096069869,
+ "grad_norm": 0.013493471778929234,
+ "learning_rate": 0.0006,
+ "loss": 3.9387717247009277,
+ "step": 7389
+ },
+ {
+ "epoch": 102.64279475982532,
+ "grad_norm": 0.015176896937191486,
+ "learning_rate": 0.0006,
+ "loss": 3.9376473426818848,
+ "step": 7390
+ },
+ {
+ "epoch": 102.65676855895197,
+ "grad_norm": 0.013986237347126007,
+ "learning_rate": 0.0006,
+ "loss": 3.96488618850708,
+ "step": 7391
+ },
+ {
+ "epoch": 102.6707423580786,
+ "grad_norm": 0.013385794125497341,
+ "learning_rate": 0.0006,
+ "loss": 4.048396110534668,
+ "step": 7392
+ },
+ {
+ "epoch": 102.68471615720524,
+ "grad_norm": 0.012358390726149082,
+ "learning_rate": 0.0006,
+ "loss": 3.950759172439575,
+ "step": 7393
+ },
+ {
+ "epoch": 102.69868995633188,
+ "grad_norm": 0.01348735112696886,
+ "learning_rate": 0.0006,
+ "loss": 4.1605658531188965,
+ "step": 7394
+ },
+ {
+ "epoch": 102.71266375545852,
+ "grad_norm": 0.01473618671298027,
+ "learning_rate": 0.0006,
+ "loss": 4.0965046882629395,
+ "step": 7395
+ },
+ {
+ "epoch": 102.72663755458515,
+ "grad_norm": 0.014261185191571712,
+ "learning_rate": 0.0006,
+ "loss": 4.105083465576172,
+ "step": 7396
+ },
+ {
+ "epoch": 102.74061135371178,
+ "grad_norm": 0.013883501291275024,
+ "learning_rate": 0.0006,
+ "loss": 4.028477668762207,
+ "step": 7397
+ },
+ {
+ "epoch": 102.75458515283843,
+ "grad_norm": 0.0138440802693367,
+ "learning_rate": 0.0006,
+ "loss": 4.05599308013916,
+ "step": 7398
+ },
+ {
+ "epoch": 102.76855895196506,
+ "grad_norm": 0.014218976721167564,
+ "learning_rate": 0.0006,
+ "loss": 4.154707908630371,
+ "step": 7399
+ },
+ {
+ "epoch": 102.78253275109171,
+ "grad_norm": 0.013328758999705315,
+ "learning_rate": 0.0006,
+ "loss": 4.060786247253418,
+ "step": 7400
+ },
+ {
+ "epoch": 102.79650655021834,
+ "grad_norm": 0.013351088389754295,
+ "learning_rate": 0.0006,
+ "loss": 4.045477867126465,
+ "step": 7401
+ },
+ {
+ "epoch": 102.81048034934499,
+ "grad_norm": 0.014380485750734806,
+ "learning_rate": 0.0006,
+ "loss": 4.032599449157715,
+ "step": 7402
+ },
+ {
+ "epoch": 102.82445414847162,
+ "grad_norm": 0.014046687632799149,
+ "learning_rate": 0.0006,
+ "loss": 4.006680488586426,
+ "step": 7403
+ },
+ {
+ "epoch": 102.83842794759825,
+ "grad_norm": 0.013186605647206306,
+ "learning_rate": 0.0006,
+ "loss": 4.0323991775512695,
+ "step": 7404
+ },
+ {
+ "epoch": 102.8524017467249,
+ "grad_norm": 0.013620814308524132,
+ "learning_rate": 0.0006,
+ "loss": 3.9950883388519287,
+ "step": 7405
+ },
+ {
+ "epoch": 102.86637554585153,
+ "grad_norm": 0.013619142584502697,
+ "learning_rate": 0.0006,
+ "loss": 4.052000999450684,
+ "step": 7406
+ },
+ {
+ "epoch": 102.88034934497817,
+ "grad_norm": 0.012630553916096687,
+ "learning_rate": 0.0006,
+ "loss": 4.104743957519531,
+ "step": 7407
+ },
+ {
+ "epoch": 102.8943231441048,
+ "grad_norm": 0.013503247871994972,
+ "learning_rate": 0.0006,
+ "loss": 4.055751800537109,
+ "step": 7408
+ },
+ {
+ "epoch": 102.90829694323143,
+ "grad_norm": 0.01509852521121502,
+ "learning_rate": 0.0006,
+ "loss": 4.064550399780273,
+ "step": 7409
+ },
+ {
+ "epoch": 102.92227074235808,
+ "grad_norm": 0.015878785401582718,
+ "learning_rate": 0.0006,
+ "loss": 4.0609917640686035,
+ "step": 7410
+ },
+ {
+ "epoch": 102.93624454148471,
+ "grad_norm": 0.015079273842275143,
+ "learning_rate": 0.0006,
+ "loss": 3.9825186729431152,
+ "step": 7411
+ },
+ {
+ "epoch": 102.95021834061136,
+ "grad_norm": 0.014393322169780731,
+ "learning_rate": 0.0006,
+ "loss": 4.089018821716309,
+ "step": 7412
+ },
+ {
+ "epoch": 102.96419213973799,
+ "grad_norm": 0.01369137316942215,
+ "learning_rate": 0.0006,
+ "loss": 4.038545608520508,
+ "step": 7413
+ },
+ {
+ "epoch": 102.97816593886463,
+ "grad_norm": 0.013304094783961773,
+ "learning_rate": 0.0006,
+ "loss": 4.075267314910889,
+ "step": 7414
+ },
+ {
+ "epoch": 102.99213973799127,
+ "grad_norm": 0.01388571783900261,
+ "learning_rate": 0.0006,
+ "loss": 4.054773330688477,
+ "step": 7415
+ },
+ {
+ "epoch": 103.0,
+ "grad_norm": 0.015209192410111427,
+ "learning_rate": 0.0006,
+ "loss": 3.9846372604370117,
+ "step": 7416
+ },
+ {
+ "epoch": 103.0,
+ "eval_loss": 4.649291515350342,
+ "eval_runtime": 56.474,
+ "eval_samples_per_second": 43.241,
+ "eval_steps_per_second": 1.363,
+ "step": 7416
+ },
+ {
+ "epoch": 103.01397379912663,
+ "grad_norm": 0.014897244982421398,
+ "learning_rate": 0.0006,
+ "loss": 3.905099868774414,
+ "step": 7417
+ },
+ {
+ "epoch": 103.02794759825328,
+ "grad_norm": 0.015420001931488514,
+ "learning_rate": 0.0006,
+ "loss": 3.985867500305176,
+ "step": 7418
+ },
+ {
+ "epoch": 103.04192139737991,
+ "grad_norm": 0.016167795285582542,
+ "learning_rate": 0.0006,
+ "loss": 3.8834733963012695,
+ "step": 7419
+ },
+ {
+ "epoch": 103.05589519650655,
+ "grad_norm": 0.017051100730895996,
+ "learning_rate": 0.0006,
+ "loss": 4.044392108917236,
+ "step": 7420
+ },
+ {
+ "epoch": 103.06986899563319,
+ "grad_norm": 0.016036640852689743,
+ "learning_rate": 0.0006,
+ "loss": 4.0110883712768555,
+ "step": 7421
+ },
+ {
+ "epoch": 103.08384279475983,
+ "grad_norm": 0.016148710623383522,
+ "learning_rate": 0.0006,
+ "loss": 4.0064544677734375,
+ "step": 7422
+ },
+ {
+ "epoch": 103.09781659388646,
+ "grad_norm": 0.014666100963950157,
+ "learning_rate": 0.0006,
+ "loss": 3.9640684127807617,
+ "step": 7423
+ },
+ {
+ "epoch": 103.1117903930131,
+ "grad_norm": 0.015464609488844872,
+ "learning_rate": 0.0006,
+ "loss": 4.069262504577637,
+ "step": 7424
+ },
+ {
+ "epoch": 103.12576419213974,
+ "grad_norm": 0.01731940731406212,
+ "learning_rate": 0.0006,
+ "loss": 3.9752140045166016,
+ "step": 7425
+ },
+ {
+ "epoch": 103.13973799126637,
+ "grad_norm": 0.017115836963057518,
+ "learning_rate": 0.0006,
+ "loss": 3.9323995113372803,
+ "step": 7426
+ },
+ {
+ "epoch": 103.15371179039302,
+ "grad_norm": 0.016439078375697136,
+ "learning_rate": 0.0006,
+ "loss": 4.088135719299316,
+ "step": 7427
+ },
+ {
+ "epoch": 103.16768558951965,
+ "grad_norm": 0.015544523485004902,
+ "learning_rate": 0.0006,
+ "loss": 4.152291297912598,
+ "step": 7428
+ },
+ {
+ "epoch": 103.18165938864628,
+ "grad_norm": 0.016065534204244614,
+ "learning_rate": 0.0006,
+ "loss": 4.069518566131592,
+ "step": 7429
+ },
+ {
+ "epoch": 103.19563318777293,
+ "grad_norm": 0.015838216990232468,
+ "learning_rate": 0.0006,
+ "loss": 3.9768013954162598,
+ "step": 7430
+ },
+ {
+ "epoch": 103.20960698689956,
+ "grad_norm": 0.01694624498486519,
+ "learning_rate": 0.0006,
+ "loss": 3.974421977996826,
+ "step": 7431
+ },
+ {
+ "epoch": 103.2235807860262,
+ "grad_norm": 0.017309164628386497,
+ "learning_rate": 0.0006,
+ "loss": 4.101884841918945,
+ "step": 7432
+ },
+ {
+ "epoch": 103.23755458515284,
+ "grad_norm": 0.017330938950181007,
+ "learning_rate": 0.0006,
+ "loss": 4.018990516662598,
+ "step": 7433
+ },
+ {
+ "epoch": 103.25152838427948,
+ "grad_norm": 0.01596875861287117,
+ "learning_rate": 0.0006,
+ "loss": 4.1121063232421875,
+ "step": 7434
+ },
+ {
+ "epoch": 103.26550218340611,
+ "grad_norm": 0.017698034644126892,
+ "learning_rate": 0.0006,
+ "loss": 4.090383052825928,
+ "step": 7435
+ },
+ {
+ "epoch": 103.27947598253274,
+ "grad_norm": 0.018080011010169983,
+ "learning_rate": 0.0006,
+ "loss": 3.980351448059082,
+ "step": 7436
+ },
+ {
+ "epoch": 103.29344978165939,
+ "grad_norm": 0.01784810796380043,
+ "learning_rate": 0.0006,
+ "loss": 3.836942672729492,
+ "step": 7437
+ },
+ {
+ "epoch": 103.30742358078602,
+ "grad_norm": 0.01571204699575901,
+ "learning_rate": 0.0006,
+ "loss": 3.9809398651123047,
+ "step": 7438
+ },
+ {
+ "epoch": 103.32139737991267,
+ "grad_norm": 0.015144702047109604,
+ "learning_rate": 0.0006,
+ "loss": 4.122274875640869,
+ "step": 7439
+ },
+ {
+ "epoch": 103.3353711790393,
+ "grad_norm": 0.017243649810552597,
+ "learning_rate": 0.0006,
+ "loss": 3.9569320678710938,
+ "step": 7440
+ },
+ {
+ "epoch": 103.34934497816595,
+ "grad_norm": 0.017546052113175392,
+ "learning_rate": 0.0006,
+ "loss": 3.8998186588287354,
+ "step": 7441
+ },
+ {
+ "epoch": 103.36331877729258,
+ "grad_norm": 0.016121823340654373,
+ "learning_rate": 0.0006,
+ "loss": 4.154199600219727,
+ "step": 7442
+ },
+ {
+ "epoch": 103.37729257641921,
+ "grad_norm": 0.01464556809514761,
+ "learning_rate": 0.0006,
+ "loss": 3.8896102905273438,
+ "step": 7443
+ },
+ {
+ "epoch": 103.39126637554585,
+ "grad_norm": 0.015656189993023872,
+ "learning_rate": 0.0006,
+ "loss": 3.9351630210876465,
+ "step": 7444
+ },
+ {
+ "epoch": 103.40524017467249,
+ "grad_norm": 0.01604314334690571,
+ "learning_rate": 0.0006,
+ "loss": 4.0186238288879395,
+ "step": 7445
+ },
+ {
+ "epoch": 103.41921397379913,
+ "grad_norm": 0.014529184438288212,
+ "learning_rate": 0.0006,
+ "loss": 4.003365516662598,
+ "step": 7446
+ },
+ {
+ "epoch": 103.43318777292576,
+ "grad_norm": 0.015263797715306282,
+ "learning_rate": 0.0006,
+ "loss": 4.097224712371826,
+ "step": 7447
+ },
+ {
+ "epoch": 103.44716157205241,
+ "grad_norm": 0.015169554390013218,
+ "learning_rate": 0.0006,
+ "loss": 4.023651123046875,
+ "step": 7448
+ },
+ {
+ "epoch": 103.46113537117904,
+ "grad_norm": 0.014693229459226131,
+ "learning_rate": 0.0006,
+ "loss": 4.095487117767334,
+ "step": 7449
+ },
+ {
+ "epoch": 103.47510917030567,
+ "grad_norm": 0.015125681646168232,
+ "learning_rate": 0.0006,
+ "loss": 3.9849324226379395,
+ "step": 7450
+ },
+ {
+ "epoch": 103.48908296943232,
+ "grad_norm": 0.014218835160136223,
+ "learning_rate": 0.0006,
+ "loss": 4.104394912719727,
+ "step": 7451
+ },
+ {
+ "epoch": 103.50305676855895,
+ "grad_norm": 0.014997185207903385,
+ "learning_rate": 0.0006,
+ "loss": 3.9534459114074707,
+ "step": 7452
+ },
+ {
+ "epoch": 103.5170305676856,
+ "grad_norm": 0.014991429634392262,
+ "learning_rate": 0.0006,
+ "loss": 4.198554992675781,
+ "step": 7453
+ },
+ {
+ "epoch": 103.53100436681223,
+ "grad_norm": 0.013290980830788612,
+ "learning_rate": 0.0006,
+ "loss": 3.857856273651123,
+ "step": 7454
+ },
+ {
+ "epoch": 103.54497816593886,
+ "grad_norm": 0.013200514949858189,
+ "learning_rate": 0.0006,
+ "loss": 3.96994948387146,
+ "step": 7455
+ },
+ {
+ "epoch": 103.5589519650655,
+ "grad_norm": 0.01466651912778616,
+ "learning_rate": 0.0006,
+ "loss": 4.026296138763428,
+ "step": 7456
+ },
+ {
+ "epoch": 103.57292576419214,
+ "grad_norm": 0.015408058650791645,
+ "learning_rate": 0.0006,
+ "loss": 3.9223692417144775,
+ "step": 7457
+ },
+ {
+ "epoch": 103.58689956331878,
+ "grad_norm": 0.014617101289331913,
+ "learning_rate": 0.0006,
+ "loss": 3.9703640937805176,
+ "step": 7458
+ },
+ {
+ "epoch": 103.60087336244541,
+ "grad_norm": 0.013879700563848019,
+ "learning_rate": 0.0006,
+ "loss": 3.9978384971618652,
+ "step": 7459
+ },
+ {
+ "epoch": 103.61484716157206,
+ "grad_norm": 0.0155031718313694,
+ "learning_rate": 0.0006,
+ "loss": 4.004831314086914,
+ "step": 7460
+ },
+ {
+ "epoch": 103.62882096069869,
+ "grad_norm": 0.01537798810750246,
+ "learning_rate": 0.0006,
+ "loss": 4.076343059539795,
+ "step": 7461
+ },
+ {
+ "epoch": 103.64279475982532,
+ "grad_norm": 0.012790355831384659,
+ "learning_rate": 0.0006,
+ "loss": 4.096382141113281,
+ "step": 7462
+ },
+ {
+ "epoch": 103.65676855895197,
+ "grad_norm": 0.012812363915145397,
+ "learning_rate": 0.0006,
+ "loss": 4.065099716186523,
+ "step": 7463
+ },
+ {
+ "epoch": 103.6707423580786,
+ "grad_norm": 0.01436334103345871,
+ "learning_rate": 0.0006,
+ "loss": 3.9989516735076904,
+ "step": 7464
+ },
+ {
+ "epoch": 103.68471615720524,
+ "grad_norm": 0.014180350117385387,
+ "learning_rate": 0.0006,
+ "loss": 4.0170698165893555,
+ "step": 7465
+ },
+ {
+ "epoch": 103.69868995633188,
+ "grad_norm": 0.01401323638856411,
+ "learning_rate": 0.0006,
+ "loss": 4.02116584777832,
+ "step": 7466
+ },
+ {
+ "epoch": 103.71266375545852,
+ "grad_norm": 0.01357064489275217,
+ "learning_rate": 0.0006,
+ "loss": 4.083914279937744,
+ "step": 7467
+ },
+ {
+ "epoch": 103.72663755458515,
+ "grad_norm": 0.013478172942996025,
+ "learning_rate": 0.0006,
+ "loss": 3.946953773498535,
+ "step": 7468
+ },
+ {
+ "epoch": 103.74061135371178,
+ "grad_norm": 0.014460090547800064,
+ "learning_rate": 0.0006,
+ "loss": 3.8642897605895996,
+ "step": 7469
+ },
+ {
+ "epoch": 103.75458515283843,
+ "grad_norm": 0.015068473294377327,
+ "learning_rate": 0.0006,
+ "loss": 4.096778869628906,
+ "step": 7470
+ },
+ {
+ "epoch": 103.76855895196506,
+ "grad_norm": 0.01593821309506893,
+ "learning_rate": 0.0006,
+ "loss": 3.9700300693511963,
+ "step": 7471
+ },
+ {
+ "epoch": 103.78253275109171,
+ "grad_norm": 0.016736770048737526,
+ "learning_rate": 0.0006,
+ "loss": 4.045576572418213,
+ "step": 7472
+ },
+ {
+ "epoch": 103.79650655021834,
+ "grad_norm": 0.01562635414302349,
+ "learning_rate": 0.0006,
+ "loss": 4.108773231506348,
+ "step": 7473
+ },
+ {
+ "epoch": 103.81048034934499,
+ "grad_norm": 0.014363548718392849,
+ "learning_rate": 0.0006,
+ "loss": 4.199751853942871,
+ "step": 7474
+ },
+ {
+ "epoch": 103.82445414847162,
+ "grad_norm": 0.013952836394309998,
+ "learning_rate": 0.0006,
+ "loss": 4.087684631347656,
+ "step": 7475
+ },
+ {
+ "epoch": 103.83842794759825,
+ "grad_norm": 0.015203139744699001,
+ "learning_rate": 0.0006,
+ "loss": 4.094961166381836,
+ "step": 7476
+ },
+ {
+ "epoch": 103.8524017467249,
+ "grad_norm": 0.015558234415948391,
+ "learning_rate": 0.0006,
+ "loss": 4.041770935058594,
+ "step": 7477
+ },
+ {
+ "epoch": 103.86637554585153,
+ "grad_norm": 0.01485239528119564,
+ "learning_rate": 0.0006,
+ "loss": 3.9825830459594727,
+ "step": 7478
+ },
+ {
+ "epoch": 103.88034934497817,
+ "grad_norm": 0.013139205053448677,
+ "learning_rate": 0.0006,
+ "loss": 4.084344863891602,
+ "step": 7479
+ },
+ {
+ "epoch": 103.8943231441048,
+ "grad_norm": 0.013366639614105225,
+ "learning_rate": 0.0006,
+ "loss": 4.10720682144165,
+ "step": 7480
+ },
+ {
+ "epoch": 103.90829694323143,
+ "grad_norm": 0.012898124754428864,
+ "learning_rate": 0.0006,
+ "loss": 4.066727161407471,
+ "step": 7481
+ },
+ {
+ "epoch": 103.92227074235808,
+ "grad_norm": 0.012501404620707035,
+ "learning_rate": 0.0006,
+ "loss": 4.064421653747559,
+ "step": 7482
+ },
+ {
+ "epoch": 103.93624454148471,
+ "grad_norm": 0.013129777275025845,
+ "learning_rate": 0.0006,
+ "loss": 3.908412456512451,
+ "step": 7483
+ },
+ {
+ "epoch": 103.95021834061136,
+ "grad_norm": 0.01346709206700325,
+ "learning_rate": 0.0006,
+ "loss": 4.070006847381592,
+ "step": 7484
+ },
+ {
+ "epoch": 103.96419213973799,
+ "grad_norm": 0.013403614051640034,
+ "learning_rate": 0.0006,
+ "loss": 3.978532314300537,
+ "step": 7485
+ },
+ {
+ "epoch": 103.97816593886463,
+ "grad_norm": 0.013506206683814526,
+ "learning_rate": 0.0006,
+ "loss": 4.025320053100586,
+ "step": 7486
+ },
+ {
+ "epoch": 103.99213973799127,
+ "grad_norm": 0.015543299727141857,
+ "learning_rate": 0.0006,
+ "loss": 4.066288948059082,
+ "step": 7487
+ },
+ {
+ "epoch": 104.0,
+ "grad_norm": 0.017104074358940125,
+ "learning_rate": 0.0006,
+ "loss": 3.9784560203552246,
+ "step": 7488
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 28800,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 400,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": true,
+ "should_log": false,
+ "should_save": false,
+ "should_training_stop": false
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 3.192397895785513e+18,
+ "train_batch_size": 8,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-7488/training_args.bin b/runs/l2r50-i2-fulle-lm/checkpoint-7488/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..cad232831c222cc7f8b938b8bbe42f61e7d89c3a
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-7488/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:f7c40e3bfa2f5ac1450ca1243e3f2a895d44a06ed5995e6aefd97f78d229d8c8
+size 4792
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-864/chat_template.jinja b/runs/l2r50-i2-fulle-lm/checkpoint-864/chat_template.jinja
new file mode 100644
index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-864/chat_template.jinja
@@ -0,0 +1,85 @@
+{%- if tools %}
+ {{- '<|im_start|>system\n' }}
+ {%- if messages[0].role == 'system' %}
+ {{- messages[0].content + '\n\n' }}
+ {%- endif %}
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }}
+ {%- for tool in tools %}
+ {{- "\n" }}
+ {{- tool | tojson }}
+ {%- endfor %}
+ {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }}
+{%- else %}
+ {%- if messages[0].role == 'system' %}
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
+ {%- endif %}
+{%- endif %}
+{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
+{%- for message in messages[::-1] %}
+ {%- set index = (messages|length - 1) - loop.index0 %}
+ {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %}
+ {%- set ns.multi_step_tool = false %}
+ {%- set ns.last_query_index = index %}
+ {%- endif %}
+{%- endfor %}
+{%- for message in messages %}
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
+ {%- elif message.role == "assistant" %}
+ {%- set content = message.content %}
+ {%- set reasoning_content = '' %}
+ {%- if message.reasoning_content is defined and message.reasoning_content is not none %}
+ {%- set reasoning_content = message.reasoning_content %}
+ {%- else %}
+ {%- if '' in message.content %}
+ {%- set content = message.content.split('')[-1].lstrip('\n') %}
+ {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %}
+ {%- endif %}
+ {%- endif %}
+ {%- if loop.index0 > ns.last_query_index %}
+ {%- if loop.last or (not loop.last and reasoning_content) %}
+ {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- if message.tool_calls %}
+ {%- for tool_call in message.tool_calls %}
+ {%- if (loop.first and content) or (not loop.first) %}
+ {{- '\n' }}
+ {%- endif %}
+ {%- if tool_call.function %}
+ {%- set tool_call = tool_call.function %}
+ {%- endif %}
+ {{- '\n{"name": "' }}
+ {{- tool_call.name }}
+ {{- '", "arguments": ' }}
+ {%- if tool_call.arguments is string %}
+ {{- tool_call.arguments }}
+ {%- else %}
+ {{- tool_call.arguments | tojson }}
+ {%- endif %}
+ {{- '}\n' }}
+ {%- endfor %}
+ {%- endif %}
+ {{- '<|im_end|>\n' }}
+ {%- elif message.role == "tool" %}
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
+ {{- '<|im_start|>user' }}
+ {%- endif %}
+ {{- '\n\n' }}
+ {{- message.content }}
+ {{- '\n' }}
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
+ {{- '<|im_end|>\n' }}
+ {%- endif %}
+ {%- endif %}
+{%- endfor %}
+{%- if add_generation_prompt %}
+ {{- '<|im_start|>assistant\n' }}
+ {%- if enable_thinking is defined and enable_thinking is false %}
+ {{- '\n\n\n\n' }}
+ {%- endif %}
+{%- endif %}
\ No newline at end of file
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-864/config.json b/runs/l2r50-i2-fulle-lm/checkpoint-864/config.json
new file mode 100644
index 0000000000000000000000000000000000000000..22f250953c6ce8fc91ee4411856dd73041a415e2
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-864/config.json
@@ -0,0 +1,32 @@
+{
+ "architectures": [
+ "LlamaForCausalLM"
+ ],
+ "attention_bias": false,
+ "attention_dropout": 0.0,
+ "bos_token_id": null,
+ "dtype": "float32",
+ "eos_token_id": 151645,
+ "head_dim": 128,
+ "hidden_act": "silu",
+ "hidden_size": 512,
+ "initializer_range": 0.02,
+ "intermediate_size": 1536,
+ "max_position_embeddings": 2048,
+ "mlp_bias": false,
+ "model_type": "llama",
+ "num_attention_heads": 4,
+ "num_hidden_layers": 20,
+ "num_key_value_heads": 4,
+ "pad_token_id": 151645,
+ "pretraining_tp": 1,
+ "rms_norm_eps": 1e-06,
+ "rope_parameters": {
+ "rope_theta": 10000.0,
+ "rope_type": "default"
+ },
+ "tie_word_embeddings": true,
+ "transformers_version": "5.5.0",
+ "use_cache": false,
+ "vocab_size": 151673
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-864/generation_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-864/generation_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-864/generation_config.json
@@ -0,0 +1,11 @@
+{
+ "_from_model_config": true,
+ "eos_token_id": [
+ 151645
+ ],
+ "output_attentions": false,
+ "output_hidden_states": false,
+ "pad_token_id": 151645,
+ "transformers_version": "5.5.0",
+ "use_cache": true
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-864/model.safetensors b/runs/l2r50-i2-fulle-lm/checkpoint-864/model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..54b5c83669af15f0b9140b26ae8ae51ec68dca54
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-864/model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:179bf5635dd10044802d1dc44ed24b19194380f3efc90306146f615333929f19
+size 583360328
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-864/optimizer.pt b/runs/l2r50-i2-fulle-lm/checkpoint-864/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..d1f8285526f273614069da6f6ec8b57e2be1aa31
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-864/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:8e1c6b2105f2225f0b40ba61fdf70d02a365d80a1338b8755155116ec1401037
+size 1166833530
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-864/rng_state_0.pth b/runs/l2r50-i2-fulle-lm/checkpoint-864/rng_state_0.pth
new file mode 100644
index 0000000000000000000000000000000000000000..2850db5a8b3c23d091b3fd3a0338fa0da162e9cb
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-864/rng_state_0.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:d88b5132c373442240592255aff89e931e66ca54a5309ba0106d24acc1c940e8
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-864/rng_state_1.pth b/runs/l2r50-i2-fulle-lm/checkpoint-864/rng_state_1.pth
new file mode 100644
index 0000000000000000000000000000000000000000..01e3351fc2ba528e4ec4ec330aa331839cf9147b
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-864/rng_state_1.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:f969ec38f9da2e2770ade6f42ea9a45cc7aea3a552040d097648e88ad7651bc0
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-864/rng_state_2.pth b/runs/l2r50-i2-fulle-lm/checkpoint-864/rng_state_2.pth
new file mode 100644
index 0000000000000000000000000000000000000000..e3400b288bec9024746b50e3d0efc784072fe544
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-864/rng_state_2.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:b28227375ca73bdf9d25dbde42e64d4ea3bea06c551f80c1045cbb59913998a9
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-864/rng_state_3.pth b/runs/l2r50-i2-fulle-lm/checkpoint-864/rng_state_3.pth
new file mode 100644
index 0000000000000000000000000000000000000000..0784197f679ee63333ba951dc2ea04a15f9a85df
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-864/rng_state_3.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:89301a7e872e726f45ed2ce4fe0342d3a8e335a1ba8b8f8575869beec7fb5f36
+size 15024
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-864/scheduler.pt b/runs/l2r50-i2-fulle-lm/checkpoint-864/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..f21fc027a37dccb66b77a4752de227c557d8ff04
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-864/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:b1a1dda88112968a1342ac518f362b4f5e138437f0b670c818ad8493b91ca49d
+size 1064
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-864/tokenizer.json b/runs/l2r50-i2-fulle-lm/checkpoint-864/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..53786016f070e11184813f3c83101dc5e93e392e
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-864/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:c3c63fda1646a555448e8f03662542e04b11d103721319e4d9b267d7da03ceb8
+size 11424463
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-864/tokenizer_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-864/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..98e2fb9b2cc54785dc9d871b8e503f8b8f2defe7
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-864/tokenizer_config.json
@@ -0,0 +1,20 @@
+{
+ "add_prefix_space": false,
+ "backend": "tokenizers",
+ "bos_token": null,
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "<|im_end|>",
+ "errors": "replace",
+ "extra_special_tokens": [
+ "<|l2r_pred|>",
+ "<|r2l_pred|>",
+ "<|next_1_pred|>",
+ "<|next_2_pred|>"
+ ],
+ "is_local": false,
+ "model_max_length": 131072,
+ "pad_token": "<|im_end|>",
+ "split_special_tokens": false,
+ "tokenizer_class": "Qwen2Tokenizer",
+ "unk_token": null
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-864/trainer_state.json b/runs/l2r50-i2-fulle-lm/checkpoint-864/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..31ae1d5dacebb36361cb567c4f9d884367b89410
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-864/trainer_state.json
@@ -0,0 +1,6170 @@
+{
+ "best_global_step": null,
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 12.0,
+ "eval_steps": 500,
+ "global_step": 864,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.013973799126637555,
+ "grad_norm": 0.1739506870508194,
+ "learning_rate": 0.0,
+ "loss": 11.99299430847168,
+ "step": 1
+ },
+ {
+ "epoch": 0.02794759825327511,
+ "grad_norm": 0.17216597497463226,
+ "learning_rate": 5.999999999999999e-06,
+ "loss": 11.988704681396484,
+ "step": 2
+ },
+ {
+ "epoch": 0.04192139737991266,
+ "grad_norm": 0.17459715902805328,
+ "learning_rate": 1.1999999999999999e-05,
+ "loss": 11.945581436157227,
+ "step": 3
+ },
+ {
+ "epoch": 0.05589519650655022,
+ "grad_norm": 0.17577414214611053,
+ "learning_rate": 1.7999999999999997e-05,
+ "loss": 11.859772682189941,
+ "step": 4
+ },
+ {
+ "epoch": 0.06986899563318777,
+ "grad_norm": 0.16163651645183563,
+ "learning_rate": 2.3999999999999997e-05,
+ "loss": 11.744415283203125,
+ "step": 5
+ },
+ {
+ "epoch": 0.08384279475982533,
+ "grad_norm": 0.14409932494163513,
+ "learning_rate": 2.9999999999999997e-05,
+ "loss": 11.618099212646484,
+ "step": 6
+ },
+ {
+ "epoch": 0.09781659388646288,
+ "grad_norm": 0.13079848885536194,
+ "learning_rate": 3.5999999999999994e-05,
+ "loss": 11.505935668945312,
+ "step": 7
+ },
+ {
+ "epoch": 0.11179039301310044,
+ "grad_norm": 0.11964613944292068,
+ "learning_rate": 4.2e-05,
+ "loss": 11.416532516479492,
+ "step": 8
+ },
+ {
+ "epoch": 0.125764192139738,
+ "grad_norm": 0.11345142871141434,
+ "learning_rate": 4.7999999999999994e-05,
+ "loss": 11.33206558227539,
+ "step": 9
+ },
+ {
+ "epoch": 0.13973799126637554,
+ "grad_norm": 0.11033952981233597,
+ "learning_rate": 5.399999999999999e-05,
+ "loss": 11.259052276611328,
+ "step": 10
+ },
+ {
+ "epoch": 0.1537117903930131,
+ "grad_norm": 0.10776489973068237,
+ "learning_rate": 5.9999999999999995e-05,
+ "loss": 11.203384399414062,
+ "step": 11
+ },
+ {
+ "epoch": 0.16768558951965065,
+ "grad_norm": 0.10721946507692337,
+ "learning_rate": 6.599999999999999e-05,
+ "loss": 11.153295516967773,
+ "step": 12
+ },
+ {
+ "epoch": 0.18165938864628822,
+ "grad_norm": 0.10673430562019348,
+ "learning_rate": 7.199999999999999e-05,
+ "loss": 11.110200881958008,
+ "step": 13
+ },
+ {
+ "epoch": 0.19563318777292577,
+ "grad_norm": 0.10630254447460175,
+ "learning_rate": 7.8e-05,
+ "loss": 11.067477226257324,
+ "step": 14
+ },
+ {
+ "epoch": 0.2096069868995633,
+ "grad_norm": 0.10631615668535233,
+ "learning_rate": 8.4e-05,
+ "loss": 11.022941589355469,
+ "step": 15
+ },
+ {
+ "epoch": 0.22358078602620088,
+ "grad_norm": 0.10609443485736847,
+ "learning_rate": 8.999999999999999e-05,
+ "loss": 10.970746040344238,
+ "step": 16
+ },
+ {
+ "epoch": 0.23755458515283842,
+ "grad_norm": 0.10648232698440552,
+ "learning_rate": 9.599999999999999e-05,
+ "loss": 10.91267204284668,
+ "step": 17
+ },
+ {
+ "epoch": 0.251528384279476,
+ "grad_norm": 0.10584589838981628,
+ "learning_rate": 0.000102,
+ "loss": 10.854547500610352,
+ "step": 18
+ },
+ {
+ "epoch": 0.26550218340611353,
+ "grad_norm": 0.10591720044612885,
+ "learning_rate": 0.00010799999999999998,
+ "loss": 10.788917541503906,
+ "step": 19
+ },
+ {
+ "epoch": 0.2794759825327511,
+ "grad_norm": 0.10611672699451447,
+ "learning_rate": 0.00011399999999999999,
+ "loss": 10.717556953430176,
+ "step": 20
+ },
+ {
+ "epoch": 0.2934497816593886,
+ "grad_norm": 0.10455359518527985,
+ "learning_rate": 0.00011999999999999999,
+ "loss": 10.650186538696289,
+ "step": 21
+ },
+ {
+ "epoch": 0.3074235807860262,
+ "grad_norm": 0.1047525629401207,
+ "learning_rate": 0.00012599999999999997,
+ "loss": 10.56808090209961,
+ "step": 22
+ },
+ {
+ "epoch": 0.32139737991266376,
+ "grad_norm": 0.10511605441570282,
+ "learning_rate": 0.00013199999999999998,
+ "loss": 10.484028816223145,
+ "step": 23
+ },
+ {
+ "epoch": 0.3353711790393013,
+ "grad_norm": 0.10467177629470825,
+ "learning_rate": 0.000138,
+ "loss": 10.399947166442871,
+ "step": 24
+ },
+ {
+ "epoch": 0.34934497816593885,
+ "grad_norm": 0.10432136058807373,
+ "learning_rate": 0.00014399999999999998,
+ "loss": 10.312714576721191,
+ "step": 25
+ },
+ {
+ "epoch": 0.36331877729257644,
+ "grad_norm": 0.10527262091636658,
+ "learning_rate": 0.00015,
+ "loss": 10.212261199951172,
+ "step": 26
+ },
+ {
+ "epoch": 0.377292576419214,
+ "grad_norm": 0.103610560297966,
+ "learning_rate": 0.000156,
+ "loss": 10.132034301757812,
+ "step": 27
+ },
+ {
+ "epoch": 0.39126637554585153,
+ "grad_norm": 0.10553059726953506,
+ "learning_rate": 0.000162,
+ "loss": 10.020434379577637,
+ "step": 28
+ },
+ {
+ "epoch": 0.4052401746724891,
+ "grad_norm": 0.1037299782037735,
+ "learning_rate": 0.000168,
+ "loss": 9.934466361999512,
+ "step": 29
+ },
+ {
+ "epoch": 0.4192139737991266,
+ "grad_norm": 0.1042739674448967,
+ "learning_rate": 0.00017399999999999997,
+ "loss": 9.822539329528809,
+ "step": 30
+ },
+ {
+ "epoch": 0.4331877729257642,
+ "grad_norm": 0.10300865024328232,
+ "learning_rate": 0.00017999999999999998,
+ "loss": 9.734994888305664,
+ "step": 31
+ },
+ {
+ "epoch": 0.44716157205240176,
+ "grad_norm": 0.10291654616594315,
+ "learning_rate": 0.000186,
+ "loss": 9.624505996704102,
+ "step": 32
+ },
+ {
+ "epoch": 0.4611353711790393,
+ "grad_norm": 0.10210851579904556,
+ "learning_rate": 0.00019199999999999998,
+ "loss": 9.518205642700195,
+ "step": 33
+ },
+ {
+ "epoch": 0.47510917030567684,
+ "grad_norm": 0.10222529619932175,
+ "learning_rate": 0.000198,
+ "loss": 9.412222862243652,
+ "step": 34
+ },
+ {
+ "epoch": 0.4890829694323144,
+ "grad_norm": 0.09992019087076187,
+ "learning_rate": 0.000204,
+ "loss": 9.326566696166992,
+ "step": 35
+ },
+ {
+ "epoch": 0.503056768558952,
+ "grad_norm": 0.09931084513664246,
+ "learning_rate": 0.00020999999999999998,
+ "loss": 9.210111618041992,
+ "step": 36
+ },
+ {
+ "epoch": 0.5170305676855895,
+ "grad_norm": 0.09928256273269653,
+ "learning_rate": 0.00021599999999999996,
+ "loss": 9.100493431091309,
+ "step": 37
+ },
+ {
+ "epoch": 0.5310043668122271,
+ "grad_norm": 0.09843816608190536,
+ "learning_rate": 0.00022199999999999998,
+ "loss": 8.996543884277344,
+ "step": 38
+ },
+ {
+ "epoch": 0.5449781659388646,
+ "grad_norm": 0.09825374186038971,
+ "learning_rate": 0.00022799999999999999,
+ "loss": 8.888991355895996,
+ "step": 39
+ },
+ {
+ "epoch": 0.5589519650655022,
+ "grad_norm": 0.09602385759353638,
+ "learning_rate": 0.000234,
+ "loss": 8.802907943725586,
+ "step": 40
+ },
+ {
+ "epoch": 0.5729257641921397,
+ "grad_norm": 0.09414064139127731,
+ "learning_rate": 0.00023999999999999998,
+ "loss": 8.715423583984375,
+ "step": 41
+ },
+ {
+ "epoch": 0.5868995633187772,
+ "grad_norm": 0.09273893386125565,
+ "learning_rate": 0.00024599999999999996,
+ "loss": 8.61374282836914,
+ "step": 42
+ },
+ {
+ "epoch": 0.6008733624454149,
+ "grad_norm": 0.09079599380493164,
+ "learning_rate": 0.00025199999999999995,
+ "loss": 8.502904891967773,
+ "step": 43
+ },
+ {
+ "epoch": 0.6148471615720524,
+ "grad_norm": 0.08605825901031494,
+ "learning_rate": 0.000258,
+ "loss": 8.453008651733398,
+ "step": 44
+ },
+ {
+ "epoch": 0.62882096069869,
+ "grad_norm": 0.08338996022939682,
+ "learning_rate": 0.00026399999999999997,
+ "loss": 8.365163803100586,
+ "step": 45
+ },
+ {
+ "epoch": 0.6427947598253275,
+ "grad_norm": 0.08150867372751236,
+ "learning_rate": 0.00027,
+ "loss": 8.285372734069824,
+ "step": 46
+ },
+ {
+ "epoch": 0.6567685589519651,
+ "grad_norm": 0.07794807851314545,
+ "learning_rate": 0.000276,
+ "loss": 8.185811996459961,
+ "step": 47
+ },
+ {
+ "epoch": 0.6707423580786026,
+ "grad_norm": 0.07592720538377762,
+ "learning_rate": 0.00028199999999999997,
+ "loss": 8.098482131958008,
+ "step": 48
+ },
+ {
+ "epoch": 0.6847161572052402,
+ "grad_norm": 0.0709897130727768,
+ "learning_rate": 0.00028799999999999995,
+ "loss": 8.042091369628906,
+ "step": 49
+ },
+ {
+ "epoch": 0.6986899563318777,
+ "grad_norm": 0.06582305580377579,
+ "learning_rate": 0.000294,
+ "loss": 7.972844123840332,
+ "step": 50
+ },
+ {
+ "epoch": 0.7126637554585152,
+ "grad_norm": 0.059452351182699203,
+ "learning_rate": 0.0003,
+ "loss": 7.924579620361328,
+ "step": 51
+ },
+ {
+ "epoch": 0.7266375545851529,
+ "grad_norm": 0.052426718175411224,
+ "learning_rate": 0.00030599999999999996,
+ "loss": 7.876336097717285,
+ "step": 52
+ },
+ {
+ "epoch": 0.7406113537117904,
+ "grad_norm": 0.04768150672316551,
+ "learning_rate": 0.000312,
+ "loss": 7.8102707862854,
+ "step": 53
+ },
+ {
+ "epoch": 0.754585152838428,
+ "grad_norm": 0.03970075398683548,
+ "learning_rate": 0.000318,
+ "loss": 7.791913986206055,
+ "step": 54
+ },
+ {
+ "epoch": 0.7685589519650655,
+ "grad_norm": 0.03559752553701401,
+ "learning_rate": 0.000324,
+ "loss": 7.756960868835449,
+ "step": 55
+ },
+ {
+ "epoch": 0.7825327510917031,
+ "grad_norm": 0.029321353882551193,
+ "learning_rate": 0.00033,
+ "loss": 7.724878787994385,
+ "step": 56
+ },
+ {
+ "epoch": 0.7965065502183406,
+ "grad_norm": 0.023485427722334862,
+ "learning_rate": 0.000336,
+ "loss": 7.72553014755249,
+ "step": 57
+ },
+ {
+ "epoch": 0.8104803493449781,
+ "grad_norm": 0.019319474697113037,
+ "learning_rate": 0.00034199999999999996,
+ "loss": 7.709493160247803,
+ "step": 58
+ },
+ {
+ "epoch": 0.8244541484716157,
+ "grad_norm": 0.016705691814422607,
+ "learning_rate": 0.00034799999999999995,
+ "loss": 7.6873297691345215,
+ "step": 59
+ },
+ {
+ "epoch": 0.8384279475982532,
+ "grad_norm": 0.017667734995484352,
+ "learning_rate": 0.00035399999999999993,
+ "loss": 7.677059173583984,
+ "step": 60
+ },
+ {
+ "epoch": 0.8524017467248908,
+ "grad_norm": 0.01671992614865303,
+ "learning_rate": 0.00035999999999999997,
+ "loss": 7.643512725830078,
+ "step": 61
+ },
+ {
+ "epoch": 0.8663755458515284,
+ "grad_norm": 0.012288172729313374,
+ "learning_rate": 0.00036599999999999995,
+ "loss": 7.658343315124512,
+ "step": 62
+ },
+ {
+ "epoch": 0.880349344978166,
+ "grad_norm": 0.012004776857793331,
+ "learning_rate": 0.000372,
+ "loss": 7.649759292602539,
+ "step": 63
+ },
+ {
+ "epoch": 0.8943231441048035,
+ "grad_norm": 0.012487445957958698,
+ "learning_rate": 0.00037799999999999997,
+ "loss": 7.631503105163574,
+ "step": 64
+ },
+ {
+ "epoch": 0.9082969432314411,
+ "grad_norm": 0.01489241048693657,
+ "learning_rate": 0.00038399999999999996,
+ "loss": 7.647110462188721,
+ "step": 65
+ },
+ {
+ "epoch": 0.9222707423580786,
+ "grad_norm": 0.01428980939090252,
+ "learning_rate": 0.00039,
+ "loss": 7.636983871459961,
+ "step": 66
+ },
+ {
+ "epoch": 0.9362445414847161,
+ "grad_norm": 0.010711144655942917,
+ "learning_rate": 0.000396,
+ "loss": 7.621130466461182,
+ "step": 67
+ },
+ {
+ "epoch": 0.9502183406113537,
+ "grad_norm": 0.010178886353969574,
+ "learning_rate": 0.000402,
+ "loss": 7.593924522399902,
+ "step": 68
+ },
+ {
+ "epoch": 0.9641921397379912,
+ "grad_norm": 0.010604874230921268,
+ "learning_rate": 0.000408,
+ "loss": 7.6015496253967285,
+ "step": 69
+ },
+ {
+ "epoch": 0.9781659388646288,
+ "grad_norm": 0.009751426056027412,
+ "learning_rate": 0.0004139999999999999,
+ "loss": 7.571964263916016,
+ "step": 70
+ },
+ {
+ "epoch": 0.9921397379912664,
+ "grad_norm": 0.009816481731832027,
+ "learning_rate": 0.00041999999999999996,
+ "loss": 7.57344913482666,
+ "step": 71
+ },
+ {
+ "epoch": 1.0,
+ "grad_norm": 0.008665637113153934,
+ "learning_rate": 0.00042599999999999995,
+ "loss": 7.579050064086914,
+ "step": 72
+ },
+ {
+ "epoch": 1.0,
+ "eval_loss": 7.581204891204834,
+ "eval_runtime": 59.0091,
+ "eval_samples_per_second": 41.383,
+ "eval_steps_per_second": 1.305,
+ "step": 72
+ },
+ {
+ "epoch": 1.0139737991266375,
+ "grad_norm": 0.009684201329946518,
+ "learning_rate": 0.00043199999999999993,
+ "loss": 7.568594932556152,
+ "step": 73
+ },
+ {
+ "epoch": 1.027947598253275,
+ "grad_norm": 0.025783710181713104,
+ "learning_rate": 0.00043799999999999997,
+ "loss": 7.578948497772217,
+ "step": 74
+ },
+ {
+ "epoch": 1.0419213973799126,
+ "grad_norm": 0.0327819399535656,
+ "learning_rate": 0.00044399999999999995,
+ "loss": 7.548173904418945,
+ "step": 75
+ },
+ {
+ "epoch": 1.0558951965065502,
+ "grad_norm": 0.021604498848319054,
+ "learning_rate": 0.00045,
+ "loss": 7.552953720092773,
+ "step": 76
+ },
+ {
+ "epoch": 1.0698689956331877,
+ "grad_norm": 0.010230105370283127,
+ "learning_rate": 0.00045599999999999997,
+ "loss": 7.5213212966918945,
+ "step": 77
+ },
+ {
+ "epoch": 1.0838427947598253,
+ "grad_norm": 0.023315589874982834,
+ "learning_rate": 0.00046199999999999995,
+ "loss": 7.501640319824219,
+ "step": 78
+ },
+ {
+ "epoch": 1.0978165938864628,
+ "grad_norm": 0.01109944935888052,
+ "learning_rate": 0.000468,
+ "loss": 7.539519786834717,
+ "step": 79
+ },
+ {
+ "epoch": 1.1117903930131003,
+ "grad_norm": 0.026093758642673492,
+ "learning_rate": 0.000474,
+ "loss": 7.525928974151611,
+ "step": 80
+ },
+ {
+ "epoch": 1.125764192139738,
+ "grad_norm": 0.03274895250797272,
+ "learning_rate": 0.00047999999999999996,
+ "loss": 7.535602569580078,
+ "step": 81
+ },
+ {
+ "epoch": 1.1397379912663754,
+ "grad_norm": 0.01543602254241705,
+ "learning_rate": 0.000486,
+ "loss": 7.513683795928955,
+ "step": 82
+ },
+ {
+ "epoch": 1.1537117903930132,
+ "grad_norm": 0.028377249836921692,
+ "learning_rate": 0.0004919999999999999,
+ "loss": 7.49384880065918,
+ "step": 83
+ },
+ {
+ "epoch": 1.1676855895196507,
+ "grad_norm": 0.04121481627225876,
+ "learning_rate": 0.000498,
+ "loss": 7.488447189331055,
+ "step": 84
+ },
+ {
+ "epoch": 1.1816593886462883,
+ "grad_norm": 0.013282302767038345,
+ "learning_rate": 0.0005039999999999999,
+ "loss": 7.463010787963867,
+ "step": 85
+ },
+ {
+ "epoch": 1.1956331877729258,
+ "grad_norm": 0.03200364485383034,
+ "learning_rate": 0.0005099999999999999,
+ "loss": 7.473515033721924,
+ "step": 86
+ },
+ {
+ "epoch": 1.2096069868995634,
+ "grad_norm": 0.020768029615283012,
+ "learning_rate": 0.000516,
+ "loss": 7.464688777923584,
+ "step": 87
+ },
+ {
+ "epoch": 1.223580786026201,
+ "grad_norm": 0.021715788170695305,
+ "learning_rate": 0.000522,
+ "loss": 7.4835028648376465,
+ "step": 88
+ },
+ {
+ "epoch": 1.2375545851528384,
+ "grad_norm": 0.026255754753947258,
+ "learning_rate": 0.0005279999999999999,
+ "loss": 7.443392753601074,
+ "step": 89
+ },
+ {
+ "epoch": 1.251528384279476,
+ "grad_norm": 0.022325951606035233,
+ "learning_rate": 0.000534,
+ "loss": 7.457373142242432,
+ "step": 90
+ },
+ {
+ "epoch": 1.2655021834061135,
+ "grad_norm": 0.009270387701690197,
+ "learning_rate": 0.00054,
+ "loss": 7.455986976623535,
+ "step": 91
+ },
+ {
+ "epoch": 1.279475982532751,
+ "grad_norm": 0.02369544468820095,
+ "learning_rate": 0.0005459999999999999,
+ "loss": 7.442216873168945,
+ "step": 92
+ },
+ {
+ "epoch": 1.2934497816593886,
+ "grad_norm": 0.008474506437778473,
+ "learning_rate": 0.000552,
+ "loss": 7.4294915199279785,
+ "step": 93
+ },
+ {
+ "epoch": 1.3074235807860262,
+ "grad_norm": 0.017043249681591988,
+ "learning_rate": 0.000558,
+ "loss": 7.406932830810547,
+ "step": 94
+ },
+ {
+ "epoch": 1.3213973799126637,
+ "grad_norm": 0.011813412420451641,
+ "learning_rate": 0.0005639999999999999,
+ "loss": 7.415900230407715,
+ "step": 95
+ },
+ {
+ "epoch": 1.3353711790393012,
+ "grad_norm": 0.01086883433163166,
+ "learning_rate": 0.00057,
+ "loss": 7.452998638153076,
+ "step": 96
+ },
+ {
+ "epoch": 1.3493449781659388,
+ "grad_norm": 0.014700951054692268,
+ "learning_rate": 0.0005759999999999999,
+ "loss": 7.410249710083008,
+ "step": 97
+ },
+ {
+ "epoch": 1.3633187772925766,
+ "grad_norm": 0.014512976631522179,
+ "learning_rate": 0.0005819999999999999,
+ "loss": 7.429132461547852,
+ "step": 98
+ },
+ {
+ "epoch": 1.3772925764192139,
+ "grad_norm": 0.012519920244812965,
+ "learning_rate": 0.000588,
+ "loss": 7.430817604064941,
+ "step": 99
+ },
+ {
+ "epoch": 1.3912663755458516,
+ "grad_norm": 0.007132664322853088,
+ "learning_rate": 0.0005939999999999999,
+ "loss": 7.405234336853027,
+ "step": 100
+ },
+ {
+ "epoch": 1.405240174672489,
+ "grad_norm": 0.019489388912916183,
+ "learning_rate": 0.0006,
+ "loss": 7.427791595458984,
+ "step": 101
+ },
+ {
+ "epoch": 1.4192139737991267,
+ "grad_norm": 0.027252664789557457,
+ "learning_rate": 0.0006,
+ "loss": 7.41761589050293,
+ "step": 102
+ },
+ {
+ "epoch": 1.4331877729257643,
+ "grad_norm": 0.02501637488603592,
+ "learning_rate": 0.0006,
+ "loss": 7.416735649108887,
+ "step": 103
+ },
+ {
+ "epoch": 1.4471615720524018,
+ "grad_norm": 0.00989772193133831,
+ "learning_rate": 0.0006,
+ "loss": 7.3756632804870605,
+ "step": 104
+ },
+ {
+ "epoch": 1.4611353711790394,
+ "grad_norm": 0.01622949168086052,
+ "learning_rate": 0.0006,
+ "loss": 7.400811672210693,
+ "step": 105
+ },
+ {
+ "epoch": 1.475109170305677,
+ "grad_norm": 0.012407763861119747,
+ "learning_rate": 0.0006,
+ "loss": 7.37675666809082,
+ "step": 106
+ },
+ {
+ "epoch": 1.4890829694323144,
+ "grad_norm": 0.010353565216064453,
+ "learning_rate": 0.0006,
+ "loss": 7.4013142585754395,
+ "step": 107
+ },
+ {
+ "epoch": 1.503056768558952,
+ "grad_norm": 0.0120553532615304,
+ "learning_rate": 0.0006,
+ "loss": 7.379873275756836,
+ "step": 108
+ },
+ {
+ "epoch": 1.5170305676855895,
+ "grad_norm": 0.01610225811600685,
+ "learning_rate": 0.0006,
+ "loss": 7.371631145477295,
+ "step": 109
+ },
+ {
+ "epoch": 1.531004366812227,
+ "grad_norm": 0.01159879844635725,
+ "learning_rate": 0.0006,
+ "loss": 7.389007568359375,
+ "step": 110
+ },
+ {
+ "epoch": 1.5449781659388646,
+ "grad_norm": 0.008013816550374031,
+ "learning_rate": 0.0006,
+ "loss": 7.395920753479004,
+ "step": 111
+ },
+ {
+ "epoch": 1.5589519650655022,
+ "grad_norm": 0.016334809362888336,
+ "learning_rate": 0.0006,
+ "loss": 7.393121719360352,
+ "step": 112
+ },
+ {
+ "epoch": 1.5729257641921397,
+ "grad_norm": 0.018821721896529198,
+ "learning_rate": 0.0006,
+ "loss": 7.367543697357178,
+ "step": 113
+ },
+ {
+ "epoch": 1.5868995633187772,
+ "grad_norm": 0.014088994823396206,
+ "learning_rate": 0.0006,
+ "loss": 7.385537624359131,
+ "step": 114
+ },
+ {
+ "epoch": 1.600873362445415,
+ "grad_norm": 0.005864050704985857,
+ "learning_rate": 0.0006,
+ "loss": 7.365346908569336,
+ "step": 115
+ },
+ {
+ "epoch": 1.6148471615720523,
+ "grad_norm": 0.011745430529117584,
+ "learning_rate": 0.0006,
+ "loss": 7.390227317810059,
+ "step": 116
+ },
+ {
+ "epoch": 1.62882096069869,
+ "grad_norm": 0.017205573618412018,
+ "learning_rate": 0.0006,
+ "loss": 7.3558549880981445,
+ "step": 117
+ },
+ {
+ "epoch": 1.6427947598253274,
+ "grad_norm": 0.021256346255540848,
+ "learning_rate": 0.0006,
+ "loss": 7.377346992492676,
+ "step": 118
+ },
+ {
+ "epoch": 1.6567685589519652,
+ "grad_norm": 0.019335204735398293,
+ "learning_rate": 0.0006,
+ "loss": 7.346194267272949,
+ "step": 119
+ },
+ {
+ "epoch": 1.6707423580786025,
+ "grad_norm": 0.008134003728628159,
+ "learning_rate": 0.0006,
+ "loss": 7.3063578605651855,
+ "step": 120
+ },
+ {
+ "epoch": 1.6847161572052403,
+ "grad_norm": 0.017613831907510757,
+ "learning_rate": 0.0006,
+ "loss": 7.338171005249023,
+ "step": 121
+ },
+ {
+ "epoch": 1.6986899563318776,
+ "grad_norm": 0.00890734139829874,
+ "learning_rate": 0.0006,
+ "loss": 7.355384826660156,
+ "step": 122
+ },
+ {
+ "epoch": 1.7126637554585153,
+ "grad_norm": 0.013448155485093594,
+ "learning_rate": 0.0006,
+ "loss": 7.344296455383301,
+ "step": 123
+ },
+ {
+ "epoch": 1.726637554585153,
+ "grad_norm": 0.009241633117198944,
+ "learning_rate": 0.0006,
+ "loss": 7.3454060554504395,
+ "step": 124
+ },
+ {
+ "epoch": 1.7406113537117904,
+ "grad_norm": 0.018561311066150665,
+ "learning_rate": 0.0006,
+ "loss": 7.32348108291626,
+ "step": 125
+ },
+ {
+ "epoch": 1.754585152838428,
+ "grad_norm": 0.025667704641819,
+ "learning_rate": 0.0006,
+ "loss": 7.356142520904541,
+ "step": 126
+ },
+ {
+ "epoch": 1.7685589519650655,
+ "grad_norm": 0.036232151091098785,
+ "learning_rate": 0.0006,
+ "loss": 7.363539695739746,
+ "step": 127
+ },
+ {
+ "epoch": 1.782532751091703,
+ "grad_norm": 0.059483714401721954,
+ "learning_rate": 0.0006,
+ "loss": 7.402331352233887,
+ "step": 128
+ },
+ {
+ "epoch": 1.7965065502183406,
+ "grad_norm": 0.05365009233355522,
+ "learning_rate": 0.0006,
+ "loss": 7.40457820892334,
+ "step": 129
+ },
+ {
+ "epoch": 1.8104803493449781,
+ "grad_norm": 0.02738415263593197,
+ "learning_rate": 0.0006,
+ "loss": 7.3756208419799805,
+ "step": 130
+ },
+ {
+ "epoch": 1.8244541484716157,
+ "grad_norm": 0.035017091780900955,
+ "learning_rate": 0.0006,
+ "loss": 7.379263877868652,
+ "step": 131
+ },
+ {
+ "epoch": 1.8384279475982532,
+ "grad_norm": 0.02819863148033619,
+ "learning_rate": 0.0006,
+ "loss": 7.327673435211182,
+ "step": 132
+ },
+ {
+ "epoch": 1.8524017467248908,
+ "grad_norm": 0.017754238098859787,
+ "learning_rate": 0.0006,
+ "loss": 7.333420276641846,
+ "step": 133
+ },
+ {
+ "epoch": 1.8663755458515285,
+ "grad_norm": 0.024685733020305634,
+ "learning_rate": 0.0006,
+ "loss": 7.34493350982666,
+ "step": 134
+ },
+ {
+ "epoch": 1.8803493449781659,
+ "grad_norm": 0.019426532089710236,
+ "learning_rate": 0.0006,
+ "loss": 7.328448295593262,
+ "step": 135
+ },
+ {
+ "epoch": 1.8943231441048036,
+ "grad_norm": 0.02106391079723835,
+ "learning_rate": 0.0006,
+ "loss": 7.327917098999023,
+ "step": 136
+ },
+ {
+ "epoch": 1.908296943231441,
+ "grad_norm": 0.02056879736483097,
+ "learning_rate": 0.0006,
+ "loss": 7.3311991691589355,
+ "step": 137
+ },
+ {
+ "epoch": 1.9222707423580787,
+ "grad_norm": 0.016865141689777374,
+ "learning_rate": 0.0006,
+ "loss": 7.312766075134277,
+ "step": 138
+ },
+ {
+ "epoch": 1.936244541484716,
+ "grad_norm": 0.017549855634570122,
+ "learning_rate": 0.0006,
+ "loss": 7.31191349029541,
+ "step": 139
+ },
+ {
+ "epoch": 1.9502183406113538,
+ "grad_norm": 0.015930157154798508,
+ "learning_rate": 0.0006,
+ "loss": 7.330901622772217,
+ "step": 140
+ },
+ {
+ "epoch": 1.9641921397379911,
+ "grad_norm": 0.013283558189868927,
+ "learning_rate": 0.0006,
+ "loss": 7.298318386077881,
+ "step": 141
+ },
+ {
+ "epoch": 1.9781659388646289,
+ "grad_norm": 0.012023469433188438,
+ "learning_rate": 0.0006,
+ "loss": 7.282337188720703,
+ "step": 142
+ },
+ {
+ "epoch": 1.9921397379912664,
+ "grad_norm": 0.01459511648863554,
+ "learning_rate": 0.0006,
+ "loss": 7.282924175262451,
+ "step": 143
+ },
+ {
+ "epoch": 2.0,
+ "grad_norm": 0.017474235966801643,
+ "learning_rate": 0.0006,
+ "loss": 7.254333972930908,
+ "step": 144
+ },
+ {
+ "epoch": 2.0,
+ "eval_loss": 7.318708896636963,
+ "eval_runtime": 59.6555,
+ "eval_samples_per_second": 40.935,
+ "eval_steps_per_second": 1.291,
+ "step": 144
+ },
+ {
+ "epoch": 2.0139737991266378,
+ "grad_norm": 0.019757507368922234,
+ "learning_rate": 0.0006,
+ "loss": 7.30288553237915,
+ "step": 145
+ },
+ {
+ "epoch": 2.027947598253275,
+ "grad_norm": 0.014571248553693295,
+ "learning_rate": 0.0006,
+ "loss": 7.274681568145752,
+ "step": 146
+ },
+ {
+ "epoch": 2.041921397379913,
+ "grad_norm": 0.015744198113679886,
+ "learning_rate": 0.0006,
+ "loss": 7.278632640838623,
+ "step": 147
+ },
+ {
+ "epoch": 2.05589519650655,
+ "grad_norm": 0.028048858046531677,
+ "learning_rate": 0.0006,
+ "loss": 7.26558780670166,
+ "step": 148
+ },
+ {
+ "epoch": 2.069868995633188,
+ "grad_norm": 0.06547307223081589,
+ "learning_rate": 0.0006,
+ "loss": 7.32567834854126,
+ "step": 149
+ },
+ {
+ "epoch": 2.0838427947598253,
+ "grad_norm": 0.05969763547182083,
+ "learning_rate": 0.0006,
+ "loss": 7.395836353302002,
+ "step": 150
+ },
+ {
+ "epoch": 2.097816593886463,
+ "grad_norm": 0.032574281096458435,
+ "learning_rate": 0.0006,
+ "loss": 7.318150997161865,
+ "step": 151
+ },
+ {
+ "epoch": 2.1117903930131003,
+ "grad_norm": 0.029452962800860405,
+ "learning_rate": 0.0006,
+ "loss": 7.308066368103027,
+ "step": 152
+ },
+ {
+ "epoch": 2.125764192139738,
+ "grad_norm": 0.03161991387605667,
+ "learning_rate": 0.0006,
+ "loss": 7.306290626525879,
+ "step": 153
+ },
+ {
+ "epoch": 2.1397379912663754,
+ "grad_norm": 0.027096295729279518,
+ "learning_rate": 0.0006,
+ "loss": 7.297796249389648,
+ "step": 154
+ },
+ {
+ "epoch": 2.153711790393013,
+ "grad_norm": 0.028657056391239166,
+ "learning_rate": 0.0006,
+ "loss": 7.317424774169922,
+ "step": 155
+ },
+ {
+ "epoch": 2.1676855895196505,
+ "grad_norm": 0.02046995982527733,
+ "learning_rate": 0.0006,
+ "loss": 7.297986030578613,
+ "step": 156
+ },
+ {
+ "epoch": 2.1816593886462883,
+ "grad_norm": 0.02220369316637516,
+ "learning_rate": 0.0006,
+ "loss": 7.268237590789795,
+ "step": 157
+ },
+ {
+ "epoch": 2.1956331877729256,
+ "grad_norm": 0.018381357192993164,
+ "learning_rate": 0.0006,
+ "loss": 7.259415626525879,
+ "step": 158
+ },
+ {
+ "epoch": 2.2096069868995634,
+ "grad_norm": 0.0204413291066885,
+ "learning_rate": 0.0006,
+ "loss": 7.258173942565918,
+ "step": 159
+ },
+ {
+ "epoch": 2.2235807860262007,
+ "grad_norm": 0.015874288976192474,
+ "learning_rate": 0.0006,
+ "loss": 7.281834602355957,
+ "step": 160
+ },
+ {
+ "epoch": 2.2375545851528384,
+ "grad_norm": 0.01634068600833416,
+ "learning_rate": 0.0006,
+ "loss": 7.2469868659973145,
+ "step": 161
+ },
+ {
+ "epoch": 2.251528384279476,
+ "grad_norm": 0.01728747971355915,
+ "learning_rate": 0.0006,
+ "loss": 7.280995845794678,
+ "step": 162
+ },
+ {
+ "epoch": 2.2655021834061135,
+ "grad_norm": 0.016341188922524452,
+ "learning_rate": 0.0006,
+ "loss": 7.273357391357422,
+ "step": 163
+ },
+ {
+ "epoch": 2.279475982532751,
+ "grad_norm": 0.016125807538628578,
+ "learning_rate": 0.0006,
+ "loss": 7.273510932922363,
+ "step": 164
+ },
+ {
+ "epoch": 2.2934497816593886,
+ "grad_norm": 0.01629287749528885,
+ "learning_rate": 0.0006,
+ "loss": 7.267029762268066,
+ "step": 165
+ },
+ {
+ "epoch": 2.3074235807860264,
+ "grad_norm": 0.015296131372451782,
+ "learning_rate": 0.0006,
+ "loss": 7.211989402770996,
+ "step": 166
+ },
+ {
+ "epoch": 2.3213973799126637,
+ "grad_norm": 0.01709570363163948,
+ "learning_rate": 0.0006,
+ "loss": 7.242851734161377,
+ "step": 167
+ },
+ {
+ "epoch": 2.3353711790393015,
+ "grad_norm": 0.012552388943731785,
+ "learning_rate": 0.0006,
+ "loss": 7.241696834564209,
+ "step": 168
+ },
+ {
+ "epoch": 2.349344978165939,
+ "grad_norm": 0.014571291394531727,
+ "learning_rate": 0.0006,
+ "loss": 7.270681381225586,
+ "step": 169
+ },
+ {
+ "epoch": 2.3633187772925766,
+ "grad_norm": 0.013389437459409237,
+ "learning_rate": 0.0006,
+ "loss": 7.191600799560547,
+ "step": 170
+ },
+ {
+ "epoch": 2.377292576419214,
+ "grad_norm": 0.011895314790308475,
+ "learning_rate": 0.0006,
+ "loss": 7.213610649108887,
+ "step": 171
+ },
+ {
+ "epoch": 2.3912663755458516,
+ "grad_norm": 0.01345642190426588,
+ "learning_rate": 0.0006,
+ "loss": 7.193787574768066,
+ "step": 172
+ },
+ {
+ "epoch": 2.405240174672489,
+ "grad_norm": 0.0091794328764081,
+ "learning_rate": 0.0006,
+ "loss": 7.233364105224609,
+ "step": 173
+ },
+ {
+ "epoch": 2.4192139737991267,
+ "grad_norm": 0.013767086900770664,
+ "learning_rate": 0.0006,
+ "loss": 7.217272758483887,
+ "step": 174
+ },
+ {
+ "epoch": 2.433187772925764,
+ "grad_norm": 0.015134960412979126,
+ "learning_rate": 0.0006,
+ "loss": 7.192791938781738,
+ "step": 175
+ },
+ {
+ "epoch": 2.447161572052402,
+ "grad_norm": 0.01790648326277733,
+ "learning_rate": 0.0006,
+ "loss": 7.229467391967773,
+ "step": 176
+ },
+ {
+ "epoch": 2.461135371179039,
+ "grad_norm": 0.015362377278506756,
+ "learning_rate": 0.0006,
+ "loss": 7.180346488952637,
+ "step": 177
+ },
+ {
+ "epoch": 2.475109170305677,
+ "grad_norm": 0.010505498386919498,
+ "learning_rate": 0.0006,
+ "loss": 7.188833236694336,
+ "step": 178
+ },
+ {
+ "epoch": 2.489082969432314,
+ "grad_norm": 0.009129747748374939,
+ "learning_rate": 0.0006,
+ "loss": 7.203153133392334,
+ "step": 179
+ },
+ {
+ "epoch": 2.503056768558952,
+ "grad_norm": 0.01147883478552103,
+ "learning_rate": 0.0006,
+ "loss": 7.163824558258057,
+ "step": 180
+ },
+ {
+ "epoch": 2.5170305676855893,
+ "grad_norm": 0.011338942684233189,
+ "learning_rate": 0.0006,
+ "loss": 7.16287899017334,
+ "step": 181
+ },
+ {
+ "epoch": 2.531004366812227,
+ "grad_norm": 0.013277675025165081,
+ "learning_rate": 0.0006,
+ "loss": 7.155905723571777,
+ "step": 182
+ },
+ {
+ "epoch": 2.544978165938865,
+ "grad_norm": 0.015269347466528416,
+ "learning_rate": 0.0006,
+ "loss": 7.189371109008789,
+ "step": 183
+ },
+ {
+ "epoch": 2.558951965065502,
+ "grad_norm": 0.017905596643686295,
+ "learning_rate": 0.0006,
+ "loss": 7.177547931671143,
+ "step": 184
+ },
+ {
+ "epoch": 2.5729257641921395,
+ "grad_norm": 0.024431610479950905,
+ "learning_rate": 0.0006,
+ "loss": 7.192169189453125,
+ "step": 185
+ },
+ {
+ "epoch": 2.5868995633187772,
+ "grad_norm": 0.03656961768865585,
+ "learning_rate": 0.0006,
+ "loss": 7.124387741088867,
+ "step": 186
+ },
+ {
+ "epoch": 2.600873362445415,
+ "grad_norm": 0.04337048903107643,
+ "learning_rate": 0.0006,
+ "loss": 7.140231609344482,
+ "step": 187
+ },
+ {
+ "epoch": 2.6148471615720523,
+ "grad_norm": 0.04047228768467903,
+ "learning_rate": 0.0006,
+ "loss": 7.171927452087402,
+ "step": 188
+ },
+ {
+ "epoch": 2.62882096069869,
+ "grad_norm": 0.07825770974159241,
+ "learning_rate": 0.0006,
+ "loss": 7.207631587982178,
+ "step": 189
+ },
+ {
+ "epoch": 2.6427947598253274,
+ "grad_norm": 0.07118063420057297,
+ "learning_rate": 0.0006,
+ "loss": 7.209488391876221,
+ "step": 190
+ },
+ {
+ "epoch": 2.656768558951965,
+ "grad_norm": 0.05153293535113335,
+ "learning_rate": 0.0006,
+ "loss": 7.14793062210083,
+ "step": 191
+ },
+ {
+ "epoch": 2.6707423580786025,
+ "grad_norm": 0.03318839520215988,
+ "learning_rate": 0.0006,
+ "loss": 7.163010597229004,
+ "step": 192
+ },
+ {
+ "epoch": 2.6847161572052403,
+ "grad_norm": 0.029333539307117462,
+ "learning_rate": 0.0006,
+ "loss": 7.159955978393555,
+ "step": 193
+ },
+ {
+ "epoch": 2.6986899563318776,
+ "grad_norm": 0.029405880719423294,
+ "learning_rate": 0.0006,
+ "loss": 7.136178493499756,
+ "step": 194
+ },
+ {
+ "epoch": 2.7126637554585153,
+ "grad_norm": 0.037815988063812256,
+ "learning_rate": 0.0006,
+ "loss": 7.183186054229736,
+ "step": 195
+ },
+ {
+ "epoch": 2.726637554585153,
+ "grad_norm": 0.020307250320911407,
+ "learning_rate": 0.0006,
+ "loss": 7.133164882659912,
+ "step": 196
+ },
+ {
+ "epoch": 2.7406113537117904,
+ "grad_norm": 0.026048069819808006,
+ "learning_rate": 0.0006,
+ "loss": 7.124977111816406,
+ "step": 197
+ },
+ {
+ "epoch": 2.7545851528384278,
+ "grad_norm": 0.027316724881529808,
+ "learning_rate": 0.0006,
+ "loss": 7.07265567779541,
+ "step": 198
+ },
+ {
+ "epoch": 2.7685589519650655,
+ "grad_norm": 0.021629009395837784,
+ "learning_rate": 0.0006,
+ "loss": 7.150020599365234,
+ "step": 199
+ },
+ {
+ "epoch": 2.7825327510917033,
+ "grad_norm": 0.01605929434299469,
+ "learning_rate": 0.0006,
+ "loss": 7.10919189453125,
+ "step": 200
+ },
+ {
+ "epoch": 2.7965065502183406,
+ "grad_norm": 0.020800089463591576,
+ "learning_rate": 0.0006,
+ "loss": 7.132045269012451,
+ "step": 201
+ },
+ {
+ "epoch": 2.810480349344978,
+ "grad_norm": 0.019225774332880974,
+ "learning_rate": 0.0006,
+ "loss": 7.128364562988281,
+ "step": 202
+ },
+ {
+ "epoch": 2.8244541484716157,
+ "grad_norm": 0.01561372634023428,
+ "learning_rate": 0.0006,
+ "loss": 7.083669662475586,
+ "step": 203
+ },
+ {
+ "epoch": 2.8384279475982535,
+ "grad_norm": 0.021969519555568695,
+ "learning_rate": 0.0006,
+ "loss": 7.050841331481934,
+ "step": 204
+ },
+ {
+ "epoch": 2.8524017467248908,
+ "grad_norm": 0.021453695371747017,
+ "learning_rate": 0.0006,
+ "loss": 7.097085475921631,
+ "step": 205
+ },
+ {
+ "epoch": 2.8663755458515285,
+ "grad_norm": 0.022836022078990936,
+ "learning_rate": 0.0006,
+ "loss": 7.10405969619751,
+ "step": 206
+ },
+ {
+ "epoch": 2.880349344978166,
+ "grad_norm": 0.02430851384997368,
+ "learning_rate": 0.0006,
+ "loss": 7.045181751251221,
+ "step": 207
+ },
+ {
+ "epoch": 2.8943231441048036,
+ "grad_norm": 0.020166993141174316,
+ "learning_rate": 0.0006,
+ "loss": 7.068183898925781,
+ "step": 208
+ },
+ {
+ "epoch": 2.908296943231441,
+ "grad_norm": 0.0118505684658885,
+ "learning_rate": 0.0006,
+ "loss": 7.07132625579834,
+ "step": 209
+ },
+ {
+ "epoch": 2.9222707423580787,
+ "grad_norm": 0.023772427812218666,
+ "learning_rate": 0.0006,
+ "loss": 7.055238723754883,
+ "step": 210
+ },
+ {
+ "epoch": 2.936244541484716,
+ "grad_norm": 0.02350712940096855,
+ "learning_rate": 0.0006,
+ "loss": 7.055898666381836,
+ "step": 211
+ },
+ {
+ "epoch": 2.950218340611354,
+ "grad_norm": 0.017244907096028328,
+ "learning_rate": 0.0006,
+ "loss": 7.0099711418151855,
+ "step": 212
+ },
+ {
+ "epoch": 2.964192139737991,
+ "grad_norm": 0.021565575152635574,
+ "learning_rate": 0.0006,
+ "loss": 6.991751670837402,
+ "step": 213
+ },
+ {
+ "epoch": 2.978165938864629,
+ "grad_norm": 0.03241860866546631,
+ "learning_rate": 0.0006,
+ "loss": 7.005980491638184,
+ "step": 214
+ },
+ {
+ "epoch": 2.992139737991266,
+ "grad_norm": 0.04613726586103439,
+ "learning_rate": 0.0006,
+ "loss": 7.027859687805176,
+ "step": 215
+ },
+ {
+ "epoch": 3.0,
+ "grad_norm": 0.06404894590377808,
+ "learning_rate": 0.0006,
+ "loss": 7.123725891113281,
+ "step": 216
+ },
+ {
+ "epoch": 3.0,
+ "eval_loss": 7.070647716522217,
+ "eval_runtime": 58.5801,
+ "eval_samples_per_second": 41.687,
+ "eval_steps_per_second": 1.314,
+ "step": 216
+ },
+ {
+ "epoch": 3.0139737991266378,
+ "grad_norm": 0.05542432889342308,
+ "learning_rate": 0.0006,
+ "loss": 7.078299522399902,
+ "step": 217
+ },
+ {
+ "epoch": 3.027947598253275,
+ "grad_norm": 0.026484526693820953,
+ "learning_rate": 0.0006,
+ "loss": 6.961159706115723,
+ "step": 218
+ },
+ {
+ "epoch": 3.041921397379913,
+ "grad_norm": 0.028276391327381134,
+ "learning_rate": 0.0006,
+ "loss": 7.019961833953857,
+ "step": 219
+ },
+ {
+ "epoch": 3.05589519650655,
+ "grad_norm": 0.024486595764756203,
+ "learning_rate": 0.0006,
+ "loss": 7.039677619934082,
+ "step": 220
+ },
+ {
+ "epoch": 3.069868995633188,
+ "grad_norm": 0.017620893195271492,
+ "learning_rate": 0.0006,
+ "loss": 6.9270782470703125,
+ "step": 221
+ },
+ {
+ "epoch": 3.0838427947598253,
+ "grad_norm": 0.02547302283346653,
+ "learning_rate": 0.0006,
+ "loss": 6.924233436584473,
+ "step": 222
+ },
+ {
+ "epoch": 3.097816593886463,
+ "grad_norm": 0.02539009042084217,
+ "learning_rate": 0.0006,
+ "loss": 6.973979949951172,
+ "step": 223
+ },
+ {
+ "epoch": 3.1117903930131003,
+ "grad_norm": 0.03256227448582649,
+ "learning_rate": 0.0006,
+ "loss": 6.933725357055664,
+ "step": 224
+ },
+ {
+ "epoch": 3.125764192139738,
+ "grad_norm": 0.04651800915598869,
+ "learning_rate": 0.0006,
+ "loss": 6.951044082641602,
+ "step": 225
+ },
+ {
+ "epoch": 3.1397379912663754,
+ "grad_norm": 0.04822975769639015,
+ "learning_rate": 0.0006,
+ "loss": 6.9614410400390625,
+ "step": 226
+ },
+ {
+ "epoch": 3.153711790393013,
+ "grad_norm": 0.031165020540356636,
+ "learning_rate": 0.0006,
+ "loss": 6.91193151473999,
+ "step": 227
+ },
+ {
+ "epoch": 3.1676855895196505,
+ "grad_norm": 0.029754292219877243,
+ "learning_rate": 0.0006,
+ "loss": 6.91802978515625,
+ "step": 228
+ },
+ {
+ "epoch": 3.1816593886462883,
+ "grad_norm": 0.02193152718245983,
+ "learning_rate": 0.0006,
+ "loss": 6.928090572357178,
+ "step": 229
+ },
+ {
+ "epoch": 3.1956331877729256,
+ "grad_norm": 0.02428470179438591,
+ "learning_rate": 0.0006,
+ "loss": 6.90396785736084,
+ "step": 230
+ },
+ {
+ "epoch": 3.2096069868995634,
+ "grad_norm": 0.023374345153570175,
+ "learning_rate": 0.0006,
+ "loss": 6.885858535766602,
+ "step": 231
+ },
+ {
+ "epoch": 3.2235807860262007,
+ "grad_norm": 0.023405000567436218,
+ "learning_rate": 0.0006,
+ "loss": 6.8855767250061035,
+ "step": 232
+ },
+ {
+ "epoch": 3.2375545851528384,
+ "grad_norm": 0.017277134582400322,
+ "learning_rate": 0.0006,
+ "loss": 6.901449203491211,
+ "step": 233
+ },
+ {
+ "epoch": 3.251528384279476,
+ "grad_norm": 0.019496750086545944,
+ "learning_rate": 0.0006,
+ "loss": 6.885687828063965,
+ "step": 234
+ },
+ {
+ "epoch": 3.2655021834061135,
+ "grad_norm": 0.016893276944756508,
+ "learning_rate": 0.0006,
+ "loss": 6.872369766235352,
+ "step": 235
+ },
+ {
+ "epoch": 3.279475982532751,
+ "grad_norm": 0.02058715932071209,
+ "learning_rate": 0.0006,
+ "loss": 6.823280334472656,
+ "step": 236
+ },
+ {
+ "epoch": 3.2934497816593886,
+ "grad_norm": 0.015530755743384361,
+ "learning_rate": 0.0006,
+ "loss": 6.860499858856201,
+ "step": 237
+ },
+ {
+ "epoch": 3.3074235807860264,
+ "grad_norm": 0.01934926211833954,
+ "learning_rate": 0.0006,
+ "loss": 6.817448616027832,
+ "step": 238
+ },
+ {
+ "epoch": 3.3213973799126637,
+ "grad_norm": 0.017413552850484848,
+ "learning_rate": 0.0006,
+ "loss": 6.881955146789551,
+ "step": 239
+ },
+ {
+ "epoch": 3.3353711790393015,
+ "grad_norm": 0.026478875428438187,
+ "learning_rate": 0.0006,
+ "loss": 6.827404499053955,
+ "step": 240
+ },
+ {
+ "epoch": 3.349344978165939,
+ "grad_norm": 0.032611701637506485,
+ "learning_rate": 0.0006,
+ "loss": 6.78285551071167,
+ "step": 241
+ },
+ {
+ "epoch": 3.3633187772925766,
+ "grad_norm": 0.041900306940078735,
+ "learning_rate": 0.0006,
+ "loss": 6.783053874969482,
+ "step": 242
+ },
+ {
+ "epoch": 3.377292576419214,
+ "grad_norm": 0.04149497672915459,
+ "learning_rate": 0.0006,
+ "loss": 6.745220184326172,
+ "step": 243
+ },
+ {
+ "epoch": 3.3912663755458516,
+ "grad_norm": 0.023949826136231422,
+ "learning_rate": 0.0006,
+ "loss": 6.841533184051514,
+ "step": 244
+ },
+ {
+ "epoch": 3.405240174672489,
+ "grad_norm": 0.028601842001080513,
+ "learning_rate": 0.0006,
+ "loss": 6.852852821350098,
+ "step": 245
+ },
+ {
+ "epoch": 3.4192139737991267,
+ "grad_norm": 0.027365155518054962,
+ "learning_rate": 0.0006,
+ "loss": 6.74376106262207,
+ "step": 246
+ },
+ {
+ "epoch": 3.433187772925764,
+ "grad_norm": 0.02856568619608879,
+ "learning_rate": 0.0006,
+ "loss": 6.74906063079834,
+ "step": 247
+ },
+ {
+ "epoch": 3.447161572052402,
+ "grad_norm": 0.019172416999936104,
+ "learning_rate": 0.0006,
+ "loss": 6.809206008911133,
+ "step": 248
+ },
+ {
+ "epoch": 3.461135371179039,
+ "grad_norm": 0.02161746844649315,
+ "learning_rate": 0.0006,
+ "loss": 6.789888381958008,
+ "step": 249
+ },
+ {
+ "epoch": 3.475109170305677,
+ "grad_norm": 0.020273666828870773,
+ "learning_rate": 0.0006,
+ "loss": 6.73512077331543,
+ "step": 250
+ },
+ {
+ "epoch": 3.489082969432314,
+ "grad_norm": 0.022329283878207207,
+ "learning_rate": 0.0006,
+ "loss": 6.826148509979248,
+ "step": 251
+ },
+ {
+ "epoch": 3.503056768558952,
+ "grad_norm": 0.02284284122288227,
+ "learning_rate": 0.0006,
+ "loss": 6.748521327972412,
+ "step": 252
+ },
+ {
+ "epoch": 3.5170305676855893,
+ "grad_norm": 0.026368247345089912,
+ "learning_rate": 0.0006,
+ "loss": 6.645879745483398,
+ "step": 253
+ },
+ {
+ "epoch": 3.531004366812227,
+ "grad_norm": 0.03627244383096695,
+ "learning_rate": 0.0006,
+ "loss": 6.666064739227295,
+ "step": 254
+ },
+ {
+ "epoch": 3.544978165938865,
+ "grad_norm": 0.04140935093164444,
+ "learning_rate": 0.0006,
+ "loss": 6.712275981903076,
+ "step": 255
+ },
+ {
+ "epoch": 3.558951965065502,
+ "grad_norm": 0.0390462800860405,
+ "learning_rate": 0.0006,
+ "loss": 6.665900230407715,
+ "step": 256
+ },
+ {
+ "epoch": 3.5729257641921395,
+ "grad_norm": 0.05178583785891533,
+ "learning_rate": 0.0006,
+ "loss": 6.753298759460449,
+ "step": 257
+ },
+ {
+ "epoch": 3.5868995633187772,
+ "grad_norm": 0.054328061640262604,
+ "learning_rate": 0.0006,
+ "loss": 6.732633113861084,
+ "step": 258
+ },
+ {
+ "epoch": 3.600873362445415,
+ "grad_norm": 0.04590460658073425,
+ "learning_rate": 0.0006,
+ "loss": 6.65250301361084,
+ "step": 259
+ },
+ {
+ "epoch": 3.6148471615720523,
+ "grad_norm": 0.044331666082143784,
+ "learning_rate": 0.0006,
+ "loss": 6.68654203414917,
+ "step": 260
+ },
+ {
+ "epoch": 3.62882096069869,
+ "grad_norm": 0.05308730527758598,
+ "learning_rate": 0.0006,
+ "loss": 6.642482757568359,
+ "step": 261
+ },
+ {
+ "epoch": 3.6427947598253274,
+ "grad_norm": 0.03999016433954239,
+ "learning_rate": 0.0006,
+ "loss": 6.625227451324463,
+ "step": 262
+ },
+ {
+ "epoch": 3.656768558951965,
+ "grad_norm": 0.04088086634874344,
+ "learning_rate": 0.0006,
+ "loss": 6.67623233795166,
+ "step": 263
+ },
+ {
+ "epoch": 3.6707423580786025,
+ "grad_norm": 0.03615636005997658,
+ "learning_rate": 0.0006,
+ "loss": 6.6869001388549805,
+ "step": 264
+ },
+ {
+ "epoch": 3.6847161572052403,
+ "grad_norm": 0.02817360684275627,
+ "learning_rate": 0.0006,
+ "loss": 6.674601078033447,
+ "step": 265
+ },
+ {
+ "epoch": 3.6986899563318776,
+ "grad_norm": 0.032586123794317245,
+ "learning_rate": 0.0006,
+ "loss": 6.612994194030762,
+ "step": 266
+ },
+ {
+ "epoch": 3.7126637554585153,
+ "grad_norm": 0.031117867678403854,
+ "learning_rate": 0.0006,
+ "loss": 6.6244354248046875,
+ "step": 267
+ },
+ {
+ "epoch": 3.726637554585153,
+ "grad_norm": 0.028330687433481216,
+ "learning_rate": 0.0006,
+ "loss": 6.669557094573975,
+ "step": 268
+ },
+ {
+ "epoch": 3.7406113537117904,
+ "grad_norm": 0.021489202976226807,
+ "learning_rate": 0.0006,
+ "loss": 6.632035732269287,
+ "step": 269
+ },
+ {
+ "epoch": 3.7545851528384278,
+ "grad_norm": 0.024269424378871918,
+ "learning_rate": 0.0006,
+ "loss": 6.660867691040039,
+ "step": 270
+ },
+ {
+ "epoch": 3.7685589519650655,
+ "grad_norm": 0.02237873338162899,
+ "learning_rate": 0.0006,
+ "loss": 6.602587699890137,
+ "step": 271
+ },
+ {
+ "epoch": 3.7825327510917033,
+ "grad_norm": 0.026503929868340492,
+ "learning_rate": 0.0006,
+ "loss": 6.599200248718262,
+ "step": 272
+ },
+ {
+ "epoch": 3.7965065502183406,
+ "grad_norm": 0.03687124326825142,
+ "learning_rate": 0.0006,
+ "loss": 6.567296504974365,
+ "step": 273
+ },
+ {
+ "epoch": 3.810480349344978,
+ "grad_norm": 0.04708952456712723,
+ "learning_rate": 0.0006,
+ "loss": 6.621265411376953,
+ "step": 274
+ },
+ {
+ "epoch": 3.8244541484716157,
+ "grad_norm": 0.041846614331007004,
+ "learning_rate": 0.0006,
+ "loss": 6.550536155700684,
+ "step": 275
+ },
+ {
+ "epoch": 3.8384279475982535,
+ "grad_norm": 0.031249675899744034,
+ "learning_rate": 0.0006,
+ "loss": 6.54278564453125,
+ "step": 276
+ },
+ {
+ "epoch": 3.8524017467248908,
+ "grad_norm": 0.023082256317138672,
+ "learning_rate": 0.0006,
+ "loss": 6.544787406921387,
+ "step": 277
+ },
+ {
+ "epoch": 3.8663755458515285,
+ "grad_norm": 0.020995570346713066,
+ "learning_rate": 0.0006,
+ "loss": 6.518143653869629,
+ "step": 278
+ },
+ {
+ "epoch": 3.880349344978166,
+ "grad_norm": 0.020839324221014977,
+ "learning_rate": 0.0006,
+ "loss": 6.546667098999023,
+ "step": 279
+ },
+ {
+ "epoch": 3.8943231441048036,
+ "grad_norm": 0.018377432599663734,
+ "learning_rate": 0.0006,
+ "loss": 6.478307723999023,
+ "step": 280
+ },
+ {
+ "epoch": 3.908296943231441,
+ "grad_norm": 0.021713724359869957,
+ "learning_rate": 0.0006,
+ "loss": 6.485574245452881,
+ "step": 281
+ },
+ {
+ "epoch": 3.9222707423580787,
+ "grad_norm": 0.01359301246702671,
+ "learning_rate": 0.0006,
+ "loss": 6.570268630981445,
+ "step": 282
+ },
+ {
+ "epoch": 3.936244541484716,
+ "grad_norm": 0.016233332455158234,
+ "learning_rate": 0.0006,
+ "loss": 6.57150936126709,
+ "step": 283
+ },
+ {
+ "epoch": 3.950218340611354,
+ "grad_norm": 0.020687608048319817,
+ "learning_rate": 0.0006,
+ "loss": 6.527956008911133,
+ "step": 284
+ },
+ {
+ "epoch": 3.964192139737991,
+ "grad_norm": 0.02032964862883091,
+ "learning_rate": 0.0006,
+ "loss": 6.468169689178467,
+ "step": 285
+ },
+ {
+ "epoch": 3.978165938864629,
+ "grad_norm": 0.016676288098096848,
+ "learning_rate": 0.0006,
+ "loss": 6.448418617248535,
+ "step": 286
+ },
+ {
+ "epoch": 3.992139737991266,
+ "grad_norm": 0.0168539360165596,
+ "learning_rate": 0.0006,
+ "loss": 6.461180686950684,
+ "step": 287
+ },
+ {
+ "epoch": 4.0,
+ "grad_norm": 0.02666233666241169,
+ "learning_rate": 0.0006,
+ "loss": 6.344893932342529,
+ "step": 288
+ },
+ {
+ "epoch": 4.0,
+ "eval_loss": 6.5384979248046875,
+ "eval_runtime": 58.814,
+ "eval_samples_per_second": 41.521,
+ "eval_steps_per_second": 1.309,
+ "step": 288
+ },
+ {
+ "epoch": 4.013973799126638,
+ "grad_norm": 0.04154708981513977,
+ "learning_rate": 0.0006,
+ "loss": 6.41855525970459,
+ "step": 289
+ },
+ {
+ "epoch": 4.0279475982532755,
+ "grad_norm": 0.04895668104290962,
+ "learning_rate": 0.0006,
+ "loss": 6.499302864074707,
+ "step": 290
+ },
+ {
+ "epoch": 4.041921397379912,
+ "grad_norm": 0.034464482218027115,
+ "learning_rate": 0.0006,
+ "loss": 6.438611030578613,
+ "step": 291
+ },
+ {
+ "epoch": 4.05589519650655,
+ "grad_norm": 0.06845773756504059,
+ "learning_rate": 0.0006,
+ "loss": 6.538361072540283,
+ "step": 292
+ },
+ {
+ "epoch": 4.069868995633188,
+ "grad_norm": 0.053768858313560486,
+ "learning_rate": 0.0006,
+ "loss": 6.527524948120117,
+ "step": 293
+ },
+ {
+ "epoch": 4.083842794759826,
+ "grad_norm": 0.039037931710481644,
+ "learning_rate": 0.0006,
+ "loss": 6.503847122192383,
+ "step": 294
+ },
+ {
+ "epoch": 4.097816593886463,
+ "grad_norm": 0.03962196782231331,
+ "learning_rate": 0.0006,
+ "loss": 6.475178241729736,
+ "step": 295
+ },
+ {
+ "epoch": 4.1117903930131,
+ "grad_norm": 0.049101535230875015,
+ "learning_rate": 0.0006,
+ "loss": 6.474587917327881,
+ "step": 296
+ },
+ {
+ "epoch": 4.125764192139738,
+ "grad_norm": 0.02830282226204872,
+ "learning_rate": 0.0006,
+ "loss": 6.45065975189209,
+ "step": 297
+ },
+ {
+ "epoch": 4.139737991266376,
+ "grad_norm": 0.023153482005000114,
+ "learning_rate": 0.0006,
+ "loss": 6.5218706130981445,
+ "step": 298
+ },
+ {
+ "epoch": 4.153711790393013,
+ "grad_norm": 0.025064002722501755,
+ "learning_rate": 0.0006,
+ "loss": 6.456475257873535,
+ "step": 299
+ },
+ {
+ "epoch": 4.1676855895196505,
+ "grad_norm": 0.02357092685997486,
+ "learning_rate": 0.0006,
+ "loss": 6.485952377319336,
+ "step": 300
+ },
+ {
+ "epoch": 4.181659388646288,
+ "grad_norm": 0.021832305938005447,
+ "learning_rate": 0.0006,
+ "loss": 6.425506591796875,
+ "step": 301
+ },
+ {
+ "epoch": 4.195633187772926,
+ "grad_norm": 0.0227900929749012,
+ "learning_rate": 0.0006,
+ "loss": 6.439155578613281,
+ "step": 302
+ },
+ {
+ "epoch": 4.209606986899563,
+ "grad_norm": 0.015961607918143272,
+ "learning_rate": 0.0006,
+ "loss": 6.460862636566162,
+ "step": 303
+ },
+ {
+ "epoch": 4.223580786026201,
+ "grad_norm": 0.017876390367746353,
+ "learning_rate": 0.0006,
+ "loss": 6.44564151763916,
+ "step": 304
+ },
+ {
+ "epoch": 4.2375545851528384,
+ "grad_norm": 0.013703204691410065,
+ "learning_rate": 0.0006,
+ "loss": 6.418414115905762,
+ "step": 305
+ },
+ {
+ "epoch": 4.251528384279476,
+ "grad_norm": 0.015240306034684181,
+ "learning_rate": 0.0006,
+ "loss": 6.411238670349121,
+ "step": 306
+ },
+ {
+ "epoch": 4.265502183406113,
+ "grad_norm": 0.01587662845849991,
+ "learning_rate": 0.0006,
+ "loss": 6.395642280578613,
+ "step": 307
+ },
+ {
+ "epoch": 4.279475982532751,
+ "grad_norm": 0.014650700613856316,
+ "learning_rate": 0.0006,
+ "loss": 6.302793025970459,
+ "step": 308
+ },
+ {
+ "epoch": 4.293449781659389,
+ "grad_norm": 0.016533225774765015,
+ "learning_rate": 0.0006,
+ "loss": 6.380588054656982,
+ "step": 309
+ },
+ {
+ "epoch": 4.307423580786026,
+ "grad_norm": 0.019963741302490234,
+ "learning_rate": 0.0006,
+ "loss": 6.345336437225342,
+ "step": 310
+ },
+ {
+ "epoch": 4.321397379912664,
+ "grad_norm": 0.02237936295568943,
+ "learning_rate": 0.0006,
+ "loss": 6.387770652770996,
+ "step": 311
+ },
+ {
+ "epoch": 4.335371179039301,
+ "grad_norm": 0.026658328250050545,
+ "learning_rate": 0.0006,
+ "loss": 6.323662757873535,
+ "step": 312
+ },
+ {
+ "epoch": 4.349344978165939,
+ "grad_norm": 0.02852284163236618,
+ "learning_rate": 0.0006,
+ "loss": 6.382204055786133,
+ "step": 313
+ },
+ {
+ "epoch": 4.3633187772925766,
+ "grad_norm": 0.028815090656280518,
+ "learning_rate": 0.0006,
+ "loss": 6.402390480041504,
+ "step": 314
+ },
+ {
+ "epoch": 4.377292576419214,
+ "grad_norm": 0.029393529519438744,
+ "learning_rate": 0.0006,
+ "loss": 6.280278205871582,
+ "step": 315
+ },
+ {
+ "epoch": 4.391266375545851,
+ "grad_norm": 0.02186041697859764,
+ "learning_rate": 0.0006,
+ "loss": 6.302707672119141,
+ "step": 316
+ },
+ {
+ "epoch": 4.405240174672489,
+ "grad_norm": 0.02071218006312847,
+ "learning_rate": 0.0006,
+ "loss": 6.3365020751953125,
+ "step": 317
+ },
+ {
+ "epoch": 4.419213973799127,
+ "grad_norm": 0.02779117226600647,
+ "learning_rate": 0.0006,
+ "loss": 6.252157211303711,
+ "step": 318
+ },
+ {
+ "epoch": 4.4331877729257645,
+ "grad_norm": 0.03470654785633087,
+ "learning_rate": 0.0006,
+ "loss": 6.319070339202881,
+ "step": 319
+ },
+ {
+ "epoch": 4.447161572052401,
+ "grad_norm": 0.037067804485559464,
+ "learning_rate": 0.0006,
+ "loss": 6.314116477966309,
+ "step": 320
+ },
+ {
+ "epoch": 4.461135371179039,
+ "grad_norm": 0.032182756811380386,
+ "learning_rate": 0.0006,
+ "loss": 6.267298698425293,
+ "step": 321
+ },
+ {
+ "epoch": 4.475109170305677,
+ "grad_norm": 0.026305217295885086,
+ "learning_rate": 0.0006,
+ "loss": 6.331688404083252,
+ "step": 322
+ },
+ {
+ "epoch": 4.489082969432315,
+ "grad_norm": 0.027228357270359993,
+ "learning_rate": 0.0006,
+ "loss": 6.31139612197876,
+ "step": 323
+ },
+ {
+ "epoch": 4.503056768558952,
+ "grad_norm": 0.025781169533729553,
+ "learning_rate": 0.0006,
+ "loss": 6.194684982299805,
+ "step": 324
+ },
+ {
+ "epoch": 4.517030567685589,
+ "grad_norm": 0.029186977073550224,
+ "learning_rate": 0.0006,
+ "loss": 6.31306266784668,
+ "step": 325
+ },
+ {
+ "epoch": 4.531004366812227,
+ "grad_norm": 0.023608777672052383,
+ "learning_rate": 0.0006,
+ "loss": 6.285243034362793,
+ "step": 326
+ },
+ {
+ "epoch": 4.544978165938865,
+ "grad_norm": 0.02267594076693058,
+ "learning_rate": 0.0006,
+ "loss": 6.30342960357666,
+ "step": 327
+ },
+ {
+ "epoch": 4.558951965065502,
+ "grad_norm": 0.01926310732960701,
+ "learning_rate": 0.0006,
+ "loss": 6.280606269836426,
+ "step": 328
+ },
+ {
+ "epoch": 4.5729257641921395,
+ "grad_norm": 0.026484837755560875,
+ "learning_rate": 0.0006,
+ "loss": 6.237980365753174,
+ "step": 329
+ },
+ {
+ "epoch": 4.586899563318777,
+ "grad_norm": 0.026228854432702065,
+ "learning_rate": 0.0006,
+ "loss": 6.3018293380737305,
+ "step": 330
+ },
+ {
+ "epoch": 4.600873362445415,
+ "grad_norm": 0.022096967324614525,
+ "learning_rate": 0.0006,
+ "loss": 6.253863334655762,
+ "step": 331
+ },
+ {
+ "epoch": 4.614847161572053,
+ "grad_norm": 0.027223890647292137,
+ "learning_rate": 0.0006,
+ "loss": 6.166882038116455,
+ "step": 332
+ },
+ {
+ "epoch": 4.62882096069869,
+ "grad_norm": 0.03160124272108078,
+ "learning_rate": 0.0006,
+ "loss": 6.230309963226318,
+ "step": 333
+ },
+ {
+ "epoch": 4.642794759825327,
+ "grad_norm": 0.04110539332032204,
+ "learning_rate": 0.0006,
+ "loss": 6.2334818840026855,
+ "step": 334
+ },
+ {
+ "epoch": 4.656768558951965,
+ "grad_norm": 0.04355933889746666,
+ "learning_rate": 0.0006,
+ "loss": 6.160956382751465,
+ "step": 335
+ },
+ {
+ "epoch": 4.670742358078603,
+ "grad_norm": 0.03611086308956146,
+ "learning_rate": 0.0006,
+ "loss": 6.23399543762207,
+ "step": 336
+ },
+ {
+ "epoch": 4.68471615720524,
+ "grad_norm": 0.051811400800943375,
+ "learning_rate": 0.0006,
+ "loss": 6.241855144500732,
+ "step": 337
+ },
+ {
+ "epoch": 4.698689956331878,
+ "grad_norm": 0.057434193789958954,
+ "learning_rate": 0.0006,
+ "loss": 6.220993995666504,
+ "step": 338
+ },
+ {
+ "epoch": 4.712663755458515,
+ "grad_norm": 0.04420953616499901,
+ "learning_rate": 0.0006,
+ "loss": 6.300461292266846,
+ "step": 339
+ },
+ {
+ "epoch": 4.726637554585153,
+ "grad_norm": 0.03515457361936569,
+ "learning_rate": 0.0006,
+ "loss": 6.269349098205566,
+ "step": 340
+ },
+ {
+ "epoch": 4.74061135371179,
+ "grad_norm": 0.03222600743174553,
+ "learning_rate": 0.0006,
+ "loss": 6.248613357543945,
+ "step": 341
+ },
+ {
+ "epoch": 4.754585152838428,
+ "grad_norm": 0.0384046845138073,
+ "learning_rate": 0.0006,
+ "loss": 6.266979217529297,
+ "step": 342
+ },
+ {
+ "epoch": 4.7685589519650655,
+ "grad_norm": 0.0514867827296257,
+ "learning_rate": 0.0006,
+ "loss": 6.234103202819824,
+ "step": 343
+ },
+ {
+ "epoch": 4.782532751091703,
+ "grad_norm": 0.050747793167829514,
+ "learning_rate": 0.0006,
+ "loss": 6.322582244873047,
+ "step": 344
+ },
+ {
+ "epoch": 4.796506550218341,
+ "grad_norm": 0.040878988802433014,
+ "learning_rate": 0.0006,
+ "loss": 6.212509632110596,
+ "step": 345
+ },
+ {
+ "epoch": 4.810480349344978,
+ "grad_norm": 0.034655820578336716,
+ "learning_rate": 0.0006,
+ "loss": 6.265879154205322,
+ "step": 346
+ },
+ {
+ "epoch": 4.824454148471616,
+ "grad_norm": 0.04202224686741829,
+ "learning_rate": 0.0006,
+ "loss": 6.2896409034729,
+ "step": 347
+ },
+ {
+ "epoch": 4.8384279475982535,
+ "grad_norm": 0.03632277995347977,
+ "learning_rate": 0.0006,
+ "loss": 6.253917217254639,
+ "step": 348
+ },
+ {
+ "epoch": 4.85240174672489,
+ "grad_norm": 0.028645765036344528,
+ "learning_rate": 0.0006,
+ "loss": 6.207965850830078,
+ "step": 349
+ },
+ {
+ "epoch": 4.866375545851528,
+ "grad_norm": 0.025305170565843582,
+ "learning_rate": 0.0006,
+ "loss": 6.168068885803223,
+ "step": 350
+ },
+ {
+ "epoch": 4.880349344978166,
+ "grad_norm": 0.022137803956866264,
+ "learning_rate": 0.0006,
+ "loss": 6.235283374786377,
+ "step": 351
+ },
+ {
+ "epoch": 4.894323144104804,
+ "grad_norm": 0.01814538985490799,
+ "learning_rate": 0.0006,
+ "loss": 6.259641647338867,
+ "step": 352
+ },
+ {
+ "epoch": 4.908296943231441,
+ "grad_norm": 0.018465086817741394,
+ "learning_rate": 0.0006,
+ "loss": 6.111738681793213,
+ "step": 353
+ },
+ {
+ "epoch": 4.922270742358078,
+ "grad_norm": 0.018638933077454567,
+ "learning_rate": 0.0006,
+ "loss": 6.225642204284668,
+ "step": 354
+ },
+ {
+ "epoch": 4.936244541484716,
+ "grad_norm": 0.016461260616779327,
+ "learning_rate": 0.0006,
+ "loss": 6.140647888183594,
+ "step": 355
+ },
+ {
+ "epoch": 4.950218340611354,
+ "grad_norm": 0.015458385460078716,
+ "learning_rate": 0.0006,
+ "loss": 6.155224800109863,
+ "step": 356
+ },
+ {
+ "epoch": 4.964192139737992,
+ "grad_norm": 0.014934048056602478,
+ "learning_rate": 0.0006,
+ "loss": 6.101555824279785,
+ "step": 357
+ },
+ {
+ "epoch": 4.978165938864628,
+ "grad_norm": 0.013979545794427395,
+ "learning_rate": 0.0006,
+ "loss": 6.161718845367432,
+ "step": 358
+ },
+ {
+ "epoch": 4.992139737991266,
+ "grad_norm": 0.01640394888818264,
+ "learning_rate": 0.0006,
+ "loss": 6.055768966674805,
+ "step": 359
+ },
+ {
+ "epoch": 5.0,
+ "grad_norm": 0.015135680325329304,
+ "learning_rate": 0.0006,
+ "loss": 6.169695854187012,
+ "step": 360
+ },
+ {
+ "epoch": 5.0,
+ "eval_loss": 6.16408109664917,
+ "eval_runtime": 59.3649,
+ "eval_samples_per_second": 41.135,
+ "eval_steps_per_second": 1.297,
+ "step": 360
+ },
+ {
+ "epoch": 5.013973799126638,
+ "grad_norm": 0.012849048711359501,
+ "learning_rate": 0.0006,
+ "loss": 6.055830001831055,
+ "step": 361
+ },
+ {
+ "epoch": 5.0279475982532755,
+ "grad_norm": 0.018169576302170753,
+ "learning_rate": 0.0006,
+ "loss": 6.082655429840088,
+ "step": 362
+ },
+ {
+ "epoch": 5.041921397379912,
+ "grad_norm": 0.024293430149555206,
+ "learning_rate": 0.0006,
+ "loss": 6.109032154083252,
+ "step": 363
+ },
+ {
+ "epoch": 5.05589519650655,
+ "grad_norm": 0.03266787901520729,
+ "learning_rate": 0.0006,
+ "loss": 6.030613899230957,
+ "step": 364
+ },
+ {
+ "epoch": 5.069868995633188,
+ "grad_norm": 0.03572266176342964,
+ "learning_rate": 0.0006,
+ "loss": 6.050390720367432,
+ "step": 365
+ },
+ {
+ "epoch": 5.083842794759826,
+ "grad_norm": 0.032625213265419006,
+ "learning_rate": 0.0006,
+ "loss": 6.169313430786133,
+ "step": 366
+ },
+ {
+ "epoch": 5.097816593886463,
+ "grad_norm": 0.04051872715353966,
+ "learning_rate": 0.0006,
+ "loss": 6.10398006439209,
+ "step": 367
+ },
+ {
+ "epoch": 5.1117903930131,
+ "grad_norm": 0.0351213738322258,
+ "learning_rate": 0.0006,
+ "loss": 6.121973514556885,
+ "step": 368
+ },
+ {
+ "epoch": 5.125764192139738,
+ "grad_norm": 0.03523759916424751,
+ "learning_rate": 0.0006,
+ "loss": 6.0626654624938965,
+ "step": 369
+ },
+ {
+ "epoch": 5.139737991266376,
+ "grad_norm": 0.02706182189285755,
+ "learning_rate": 0.0006,
+ "loss": 6.049354076385498,
+ "step": 370
+ },
+ {
+ "epoch": 5.153711790393013,
+ "grad_norm": 0.024212589487433434,
+ "learning_rate": 0.0006,
+ "loss": 6.034826278686523,
+ "step": 371
+ },
+ {
+ "epoch": 5.1676855895196505,
+ "grad_norm": 0.024798082187771797,
+ "learning_rate": 0.0006,
+ "loss": 6.004058837890625,
+ "step": 372
+ },
+ {
+ "epoch": 5.181659388646288,
+ "grad_norm": 0.02325567416846752,
+ "learning_rate": 0.0006,
+ "loss": 5.986461639404297,
+ "step": 373
+ },
+ {
+ "epoch": 5.195633187772926,
+ "grad_norm": 0.019060570746660233,
+ "learning_rate": 0.0006,
+ "loss": 6.059736251831055,
+ "step": 374
+ },
+ {
+ "epoch": 5.209606986899563,
+ "grad_norm": 0.016822976991534233,
+ "learning_rate": 0.0006,
+ "loss": 6.072957515716553,
+ "step": 375
+ },
+ {
+ "epoch": 5.223580786026201,
+ "grad_norm": 0.017218658700585365,
+ "learning_rate": 0.0006,
+ "loss": 6.031739234924316,
+ "step": 376
+ },
+ {
+ "epoch": 5.2375545851528384,
+ "grad_norm": 0.014889383688569069,
+ "learning_rate": 0.0006,
+ "loss": 6.025674343109131,
+ "step": 377
+ },
+ {
+ "epoch": 5.251528384279476,
+ "grad_norm": 0.01708107627928257,
+ "learning_rate": 0.0006,
+ "loss": 6.090451240539551,
+ "step": 378
+ },
+ {
+ "epoch": 5.265502183406113,
+ "grad_norm": 0.020260317251086235,
+ "learning_rate": 0.0006,
+ "loss": 6.094257354736328,
+ "step": 379
+ },
+ {
+ "epoch": 5.279475982532751,
+ "grad_norm": 0.020110400393605232,
+ "learning_rate": 0.0006,
+ "loss": 6.021188735961914,
+ "step": 380
+ },
+ {
+ "epoch": 5.293449781659389,
+ "grad_norm": 0.02027830481529236,
+ "learning_rate": 0.0006,
+ "loss": 6.022156238555908,
+ "step": 381
+ },
+ {
+ "epoch": 5.307423580786026,
+ "grad_norm": 0.022747062146663666,
+ "learning_rate": 0.0006,
+ "loss": 6.011836051940918,
+ "step": 382
+ },
+ {
+ "epoch": 5.321397379912664,
+ "grad_norm": 0.01990230567753315,
+ "learning_rate": 0.0006,
+ "loss": 6.003054618835449,
+ "step": 383
+ },
+ {
+ "epoch": 5.335371179039301,
+ "grad_norm": 0.01596238650381565,
+ "learning_rate": 0.0006,
+ "loss": 5.991410732269287,
+ "step": 384
+ },
+ {
+ "epoch": 5.349344978165939,
+ "grad_norm": 0.015847105532884598,
+ "learning_rate": 0.0006,
+ "loss": 6.100622177124023,
+ "step": 385
+ },
+ {
+ "epoch": 5.3633187772925766,
+ "grad_norm": 0.016179397702217102,
+ "learning_rate": 0.0006,
+ "loss": 6.032659530639648,
+ "step": 386
+ },
+ {
+ "epoch": 5.377292576419214,
+ "grad_norm": 0.018256602808833122,
+ "learning_rate": 0.0006,
+ "loss": 5.961983680725098,
+ "step": 387
+ },
+ {
+ "epoch": 5.391266375545851,
+ "grad_norm": 0.02805912122130394,
+ "learning_rate": 0.0006,
+ "loss": 5.983541965484619,
+ "step": 388
+ },
+ {
+ "epoch": 5.405240174672489,
+ "grad_norm": 0.039082255214452744,
+ "learning_rate": 0.0006,
+ "loss": 6.021474361419678,
+ "step": 389
+ },
+ {
+ "epoch": 5.419213973799127,
+ "grad_norm": 0.036757659167051315,
+ "learning_rate": 0.0006,
+ "loss": 6.005046367645264,
+ "step": 390
+ },
+ {
+ "epoch": 5.4331877729257645,
+ "grad_norm": 0.035277366638183594,
+ "learning_rate": 0.0006,
+ "loss": 6.036296844482422,
+ "step": 391
+ },
+ {
+ "epoch": 5.447161572052401,
+ "grad_norm": 0.034404635429382324,
+ "learning_rate": 0.0006,
+ "loss": 5.994539260864258,
+ "step": 392
+ },
+ {
+ "epoch": 5.461135371179039,
+ "grad_norm": 0.0377206988632679,
+ "learning_rate": 0.0006,
+ "loss": 5.941534519195557,
+ "step": 393
+ },
+ {
+ "epoch": 5.475109170305677,
+ "grad_norm": 0.0389922633767128,
+ "learning_rate": 0.0006,
+ "loss": 5.983644485473633,
+ "step": 394
+ },
+ {
+ "epoch": 5.489082969432315,
+ "grad_norm": 0.04176778718829155,
+ "learning_rate": 0.0006,
+ "loss": 6.030922889709473,
+ "step": 395
+ },
+ {
+ "epoch": 5.503056768558952,
+ "grad_norm": 0.0337153784930706,
+ "learning_rate": 0.0006,
+ "loss": 5.961367130279541,
+ "step": 396
+ },
+ {
+ "epoch": 5.517030567685589,
+ "grad_norm": 0.036116816103458405,
+ "learning_rate": 0.0006,
+ "loss": 5.819280624389648,
+ "step": 397
+ },
+ {
+ "epoch": 5.531004366812227,
+ "grad_norm": 0.030724041163921356,
+ "learning_rate": 0.0006,
+ "loss": 5.906380653381348,
+ "step": 398
+ },
+ {
+ "epoch": 5.544978165938865,
+ "grad_norm": 0.030264202505350113,
+ "learning_rate": 0.0006,
+ "loss": 5.975833415985107,
+ "step": 399
+ },
+ {
+ "epoch": 5.558951965065502,
+ "grad_norm": 0.032096318900585175,
+ "learning_rate": 0.0006,
+ "loss": 6.001348972320557,
+ "step": 400
+ },
+ {
+ "epoch": 5.5729257641921395,
+ "grad_norm": 0.030579356476664543,
+ "learning_rate": 0.0006,
+ "loss": 5.928768634796143,
+ "step": 401
+ },
+ {
+ "epoch": 5.586899563318777,
+ "grad_norm": 0.028241973370313644,
+ "learning_rate": 0.0006,
+ "loss": 5.92582893371582,
+ "step": 402
+ },
+ {
+ "epoch": 5.600873362445415,
+ "grad_norm": 0.023086953908205032,
+ "learning_rate": 0.0006,
+ "loss": 5.892926216125488,
+ "step": 403
+ },
+ {
+ "epoch": 5.614847161572053,
+ "grad_norm": 0.02547992393374443,
+ "learning_rate": 0.0006,
+ "loss": 5.903195381164551,
+ "step": 404
+ },
+ {
+ "epoch": 5.62882096069869,
+ "grad_norm": 0.023089831694960594,
+ "learning_rate": 0.0006,
+ "loss": 5.912033557891846,
+ "step": 405
+ },
+ {
+ "epoch": 5.642794759825327,
+ "grad_norm": 0.025446368381381035,
+ "learning_rate": 0.0006,
+ "loss": 5.926138401031494,
+ "step": 406
+ },
+ {
+ "epoch": 5.656768558951965,
+ "grad_norm": 0.03748747706413269,
+ "learning_rate": 0.0006,
+ "loss": 5.913451194763184,
+ "step": 407
+ },
+ {
+ "epoch": 5.670742358078603,
+ "grad_norm": 0.03493810072541237,
+ "learning_rate": 0.0006,
+ "loss": 5.97633695602417,
+ "step": 408
+ },
+ {
+ "epoch": 5.68471615720524,
+ "grad_norm": 0.023131275549530983,
+ "learning_rate": 0.0006,
+ "loss": 5.899571418762207,
+ "step": 409
+ },
+ {
+ "epoch": 5.698689956331878,
+ "grad_norm": 0.02986014075577259,
+ "learning_rate": 0.0006,
+ "loss": 5.894903182983398,
+ "step": 410
+ },
+ {
+ "epoch": 5.712663755458515,
+ "grad_norm": 0.030171144753694534,
+ "learning_rate": 0.0006,
+ "loss": 5.876595497131348,
+ "step": 411
+ },
+ {
+ "epoch": 5.726637554585153,
+ "grad_norm": 0.029377546161413193,
+ "learning_rate": 0.0006,
+ "loss": 5.911637306213379,
+ "step": 412
+ },
+ {
+ "epoch": 5.74061135371179,
+ "grad_norm": 0.029829490929841995,
+ "learning_rate": 0.0006,
+ "loss": 5.819629669189453,
+ "step": 413
+ },
+ {
+ "epoch": 5.754585152838428,
+ "grad_norm": 0.024139299988746643,
+ "learning_rate": 0.0006,
+ "loss": 5.81728458404541,
+ "step": 414
+ },
+ {
+ "epoch": 5.7685589519650655,
+ "grad_norm": 0.021453432738780975,
+ "learning_rate": 0.0006,
+ "loss": 5.987326145172119,
+ "step": 415
+ },
+ {
+ "epoch": 5.782532751091703,
+ "grad_norm": 0.017942246049642563,
+ "learning_rate": 0.0006,
+ "loss": 5.893008232116699,
+ "step": 416
+ },
+ {
+ "epoch": 5.796506550218341,
+ "grad_norm": 0.019724637269973755,
+ "learning_rate": 0.0006,
+ "loss": 5.912441253662109,
+ "step": 417
+ },
+ {
+ "epoch": 5.810480349344978,
+ "grad_norm": 0.018597912043333054,
+ "learning_rate": 0.0006,
+ "loss": 5.860394477844238,
+ "step": 418
+ },
+ {
+ "epoch": 5.824454148471616,
+ "grad_norm": 0.016232412308454514,
+ "learning_rate": 0.0006,
+ "loss": 5.94991397857666,
+ "step": 419
+ },
+ {
+ "epoch": 5.8384279475982535,
+ "grad_norm": 0.01667204685509205,
+ "learning_rate": 0.0006,
+ "loss": 5.8506364822387695,
+ "step": 420
+ },
+ {
+ "epoch": 5.85240174672489,
+ "grad_norm": 0.014232151210308075,
+ "learning_rate": 0.0006,
+ "loss": 5.928424835205078,
+ "step": 421
+ },
+ {
+ "epoch": 5.866375545851528,
+ "grad_norm": 0.01570476032793522,
+ "learning_rate": 0.0006,
+ "loss": 5.953692436218262,
+ "step": 422
+ },
+ {
+ "epoch": 5.880349344978166,
+ "grad_norm": 0.01586179807782173,
+ "learning_rate": 0.0006,
+ "loss": 5.820014476776123,
+ "step": 423
+ },
+ {
+ "epoch": 5.894323144104804,
+ "grad_norm": 0.014729145914316177,
+ "learning_rate": 0.0006,
+ "loss": 5.860154628753662,
+ "step": 424
+ },
+ {
+ "epoch": 5.908296943231441,
+ "grad_norm": 0.013111459091305733,
+ "learning_rate": 0.0006,
+ "loss": 5.85988187789917,
+ "step": 425
+ },
+ {
+ "epoch": 5.922270742358078,
+ "grad_norm": 0.012937922962009907,
+ "learning_rate": 0.0006,
+ "loss": 5.756265640258789,
+ "step": 426
+ },
+ {
+ "epoch": 5.936244541484716,
+ "grad_norm": 0.01377453189343214,
+ "learning_rate": 0.0006,
+ "loss": 5.8333048820495605,
+ "step": 427
+ },
+ {
+ "epoch": 5.950218340611354,
+ "grad_norm": 0.014054795727133751,
+ "learning_rate": 0.0006,
+ "loss": 5.803333759307861,
+ "step": 428
+ },
+ {
+ "epoch": 5.964192139737992,
+ "grad_norm": 0.01674959622323513,
+ "learning_rate": 0.0006,
+ "loss": 5.8718485832214355,
+ "step": 429
+ },
+ {
+ "epoch": 5.978165938864628,
+ "grad_norm": 0.018733017146587372,
+ "learning_rate": 0.0006,
+ "loss": 5.875979900360107,
+ "step": 430
+ },
+ {
+ "epoch": 5.992139737991266,
+ "grad_norm": 0.02088983915746212,
+ "learning_rate": 0.0006,
+ "loss": 5.763634204864502,
+ "step": 431
+ },
+ {
+ "epoch": 6.0,
+ "grad_norm": 0.026705985888838768,
+ "learning_rate": 0.0006,
+ "loss": 5.6922478675842285,
+ "step": 432
+ },
+ {
+ "epoch": 6.0,
+ "eval_loss": 5.847958087921143,
+ "eval_runtime": 58.5451,
+ "eval_samples_per_second": 41.711,
+ "eval_steps_per_second": 1.315,
+ "step": 432
+ },
+ {
+ "epoch": 6.013973799126638,
+ "grad_norm": 0.03181544691324234,
+ "learning_rate": 0.0006,
+ "loss": 5.786368370056152,
+ "step": 433
+ },
+ {
+ "epoch": 6.0279475982532755,
+ "grad_norm": 0.03238112851977348,
+ "learning_rate": 0.0006,
+ "loss": 5.810310363769531,
+ "step": 434
+ },
+ {
+ "epoch": 6.041921397379912,
+ "grad_norm": 0.03916756808757782,
+ "learning_rate": 0.0006,
+ "loss": 5.7761993408203125,
+ "step": 435
+ },
+ {
+ "epoch": 6.05589519650655,
+ "grad_norm": 0.044009730219841,
+ "learning_rate": 0.0006,
+ "loss": 5.898112773895264,
+ "step": 436
+ },
+ {
+ "epoch": 6.069868995633188,
+ "grad_norm": 0.045235246419906616,
+ "learning_rate": 0.0006,
+ "loss": 5.876371383666992,
+ "step": 437
+ },
+ {
+ "epoch": 6.083842794759826,
+ "grad_norm": 0.050218384712934494,
+ "learning_rate": 0.0006,
+ "loss": 5.805103302001953,
+ "step": 438
+ },
+ {
+ "epoch": 6.097816593886463,
+ "grad_norm": 0.05444490164518356,
+ "learning_rate": 0.0006,
+ "loss": 5.826424598693848,
+ "step": 439
+ },
+ {
+ "epoch": 6.1117903930131,
+ "grad_norm": 0.04890982061624527,
+ "learning_rate": 0.0006,
+ "loss": 5.7735395431518555,
+ "step": 440
+ },
+ {
+ "epoch": 6.125764192139738,
+ "grad_norm": 0.05478639155626297,
+ "learning_rate": 0.0006,
+ "loss": 5.823677062988281,
+ "step": 441
+ },
+ {
+ "epoch": 6.139737991266376,
+ "grad_norm": 0.057562313973903656,
+ "learning_rate": 0.0006,
+ "loss": 5.889334678649902,
+ "step": 442
+ },
+ {
+ "epoch": 6.153711790393013,
+ "grad_norm": 0.06447703391313553,
+ "learning_rate": 0.0006,
+ "loss": 5.84621524810791,
+ "step": 443
+ },
+ {
+ "epoch": 6.1676855895196505,
+ "grad_norm": 0.048861872404813766,
+ "learning_rate": 0.0006,
+ "loss": 5.931595802307129,
+ "step": 444
+ },
+ {
+ "epoch": 6.181659388646288,
+ "grad_norm": 0.05521339178085327,
+ "learning_rate": 0.0006,
+ "loss": 5.8229780197143555,
+ "step": 445
+ },
+ {
+ "epoch": 6.195633187772926,
+ "grad_norm": 0.053666990250349045,
+ "learning_rate": 0.0006,
+ "loss": 5.879191875457764,
+ "step": 446
+ },
+ {
+ "epoch": 6.209606986899563,
+ "grad_norm": 0.0451025515794754,
+ "learning_rate": 0.0006,
+ "loss": 5.801628112792969,
+ "step": 447
+ },
+ {
+ "epoch": 6.223580786026201,
+ "grad_norm": 0.04291655868291855,
+ "learning_rate": 0.0006,
+ "loss": 5.914680480957031,
+ "step": 448
+ },
+ {
+ "epoch": 6.2375545851528384,
+ "grad_norm": 0.035102490335702896,
+ "learning_rate": 0.0006,
+ "loss": 5.843682765960693,
+ "step": 449
+ },
+ {
+ "epoch": 6.251528384279476,
+ "grad_norm": 0.03403995931148529,
+ "learning_rate": 0.0006,
+ "loss": 5.827154159545898,
+ "step": 450
+ },
+ {
+ "epoch": 6.265502183406113,
+ "grad_norm": 0.03444375470280647,
+ "learning_rate": 0.0006,
+ "loss": 5.800136566162109,
+ "step": 451
+ },
+ {
+ "epoch": 6.279475982532751,
+ "grad_norm": 0.03165159001946449,
+ "learning_rate": 0.0006,
+ "loss": 5.906252384185791,
+ "step": 452
+ },
+ {
+ "epoch": 6.293449781659389,
+ "grad_norm": 0.026153093203902245,
+ "learning_rate": 0.0006,
+ "loss": 5.7423200607299805,
+ "step": 453
+ },
+ {
+ "epoch": 6.307423580786026,
+ "grad_norm": 0.024357657879590988,
+ "learning_rate": 0.0006,
+ "loss": 5.701364517211914,
+ "step": 454
+ },
+ {
+ "epoch": 6.321397379912664,
+ "grad_norm": 0.021508049219846725,
+ "learning_rate": 0.0006,
+ "loss": 5.8328094482421875,
+ "step": 455
+ },
+ {
+ "epoch": 6.335371179039301,
+ "grad_norm": 0.017313921824097633,
+ "learning_rate": 0.0006,
+ "loss": 5.792145252227783,
+ "step": 456
+ },
+ {
+ "epoch": 6.349344978165939,
+ "grad_norm": 0.018563397228717804,
+ "learning_rate": 0.0006,
+ "loss": 5.732672214508057,
+ "step": 457
+ },
+ {
+ "epoch": 6.3633187772925766,
+ "grad_norm": 0.016568679362535477,
+ "learning_rate": 0.0006,
+ "loss": 5.81948184967041,
+ "step": 458
+ },
+ {
+ "epoch": 6.377292576419214,
+ "grad_norm": 0.014133770950138569,
+ "learning_rate": 0.0006,
+ "loss": 5.734982490539551,
+ "step": 459
+ },
+ {
+ "epoch": 6.391266375545851,
+ "grad_norm": 0.014114447869360447,
+ "learning_rate": 0.0006,
+ "loss": 5.70073127746582,
+ "step": 460
+ },
+ {
+ "epoch": 6.405240174672489,
+ "grad_norm": 0.012907393276691437,
+ "learning_rate": 0.0006,
+ "loss": 5.758626937866211,
+ "step": 461
+ },
+ {
+ "epoch": 6.419213973799127,
+ "grad_norm": 0.012673444114625454,
+ "learning_rate": 0.0006,
+ "loss": 5.775138854980469,
+ "step": 462
+ },
+ {
+ "epoch": 6.4331877729257645,
+ "grad_norm": 0.01311410591006279,
+ "learning_rate": 0.0006,
+ "loss": 5.70920991897583,
+ "step": 463
+ },
+ {
+ "epoch": 6.447161572052401,
+ "grad_norm": 0.012935544364154339,
+ "learning_rate": 0.0006,
+ "loss": 5.781479835510254,
+ "step": 464
+ },
+ {
+ "epoch": 6.461135371179039,
+ "grad_norm": 0.01439738366752863,
+ "learning_rate": 0.0006,
+ "loss": 5.691315650939941,
+ "step": 465
+ },
+ {
+ "epoch": 6.475109170305677,
+ "grad_norm": 0.01205186452716589,
+ "learning_rate": 0.0006,
+ "loss": 5.662499904632568,
+ "step": 466
+ },
+ {
+ "epoch": 6.489082969432315,
+ "grad_norm": 0.011575652286410332,
+ "learning_rate": 0.0006,
+ "loss": 5.6271772384643555,
+ "step": 467
+ },
+ {
+ "epoch": 6.503056768558952,
+ "grad_norm": 0.011672502383589745,
+ "learning_rate": 0.0006,
+ "loss": 5.761662006378174,
+ "step": 468
+ },
+ {
+ "epoch": 6.517030567685589,
+ "grad_norm": 0.011541751213371754,
+ "learning_rate": 0.0006,
+ "loss": 5.762078285217285,
+ "step": 469
+ },
+ {
+ "epoch": 6.531004366812227,
+ "grad_norm": 0.011396365240216255,
+ "learning_rate": 0.0006,
+ "loss": 5.67873477935791,
+ "step": 470
+ },
+ {
+ "epoch": 6.544978165938865,
+ "grad_norm": 0.010688499547541142,
+ "learning_rate": 0.0006,
+ "loss": 5.717051982879639,
+ "step": 471
+ },
+ {
+ "epoch": 6.558951965065502,
+ "grad_norm": 0.012224317528307438,
+ "learning_rate": 0.0006,
+ "loss": 5.707948684692383,
+ "step": 472
+ },
+ {
+ "epoch": 6.5729257641921395,
+ "grad_norm": 0.011856825090944767,
+ "learning_rate": 0.0006,
+ "loss": 5.70878791809082,
+ "step": 473
+ },
+ {
+ "epoch": 6.586899563318777,
+ "grad_norm": 0.01199064590036869,
+ "learning_rate": 0.0006,
+ "loss": 5.708697319030762,
+ "step": 474
+ },
+ {
+ "epoch": 6.600873362445415,
+ "grad_norm": 0.01219446212053299,
+ "learning_rate": 0.0006,
+ "loss": 5.6061577796936035,
+ "step": 475
+ },
+ {
+ "epoch": 6.614847161572053,
+ "grad_norm": 0.013397484086453915,
+ "learning_rate": 0.0006,
+ "loss": 5.624789714813232,
+ "step": 476
+ },
+ {
+ "epoch": 6.62882096069869,
+ "grad_norm": 0.01483136136084795,
+ "learning_rate": 0.0006,
+ "loss": 5.721141338348389,
+ "step": 477
+ },
+ {
+ "epoch": 6.642794759825327,
+ "grad_norm": 0.0194488987326622,
+ "learning_rate": 0.0006,
+ "loss": 5.604279518127441,
+ "step": 478
+ },
+ {
+ "epoch": 6.656768558951965,
+ "grad_norm": 0.02079896256327629,
+ "learning_rate": 0.0006,
+ "loss": 5.586322784423828,
+ "step": 479
+ },
+ {
+ "epoch": 6.670742358078603,
+ "grad_norm": 0.018996229395270348,
+ "learning_rate": 0.0006,
+ "loss": 5.5987772941589355,
+ "step": 480
+ },
+ {
+ "epoch": 6.68471615720524,
+ "grad_norm": 0.015779603272676468,
+ "learning_rate": 0.0006,
+ "loss": 5.588602066040039,
+ "step": 481
+ },
+ {
+ "epoch": 6.698689956331878,
+ "grad_norm": 0.015322973020374775,
+ "learning_rate": 0.0006,
+ "loss": 5.686467170715332,
+ "step": 482
+ },
+ {
+ "epoch": 6.712663755458515,
+ "grad_norm": 0.015282213687896729,
+ "learning_rate": 0.0006,
+ "loss": 5.664676666259766,
+ "step": 483
+ },
+ {
+ "epoch": 6.726637554585153,
+ "grad_norm": 0.016589272767305374,
+ "learning_rate": 0.0006,
+ "loss": 5.666739463806152,
+ "step": 484
+ },
+ {
+ "epoch": 6.74061135371179,
+ "grad_norm": 0.01811421848833561,
+ "learning_rate": 0.0006,
+ "loss": 5.647593021392822,
+ "step": 485
+ },
+ {
+ "epoch": 6.754585152838428,
+ "grad_norm": 0.019652029499411583,
+ "learning_rate": 0.0006,
+ "loss": 5.628726959228516,
+ "step": 486
+ },
+ {
+ "epoch": 6.7685589519650655,
+ "grad_norm": 0.02118665538728237,
+ "learning_rate": 0.0006,
+ "loss": 5.6580705642700195,
+ "step": 487
+ },
+ {
+ "epoch": 6.782532751091703,
+ "grad_norm": 0.02237832546234131,
+ "learning_rate": 0.0006,
+ "loss": 5.635931968688965,
+ "step": 488
+ },
+ {
+ "epoch": 6.796506550218341,
+ "grad_norm": 0.023897631093859673,
+ "learning_rate": 0.0006,
+ "loss": 5.517949104309082,
+ "step": 489
+ },
+ {
+ "epoch": 6.810480349344978,
+ "grad_norm": 0.02442227490246296,
+ "learning_rate": 0.0006,
+ "loss": 5.600021839141846,
+ "step": 490
+ },
+ {
+ "epoch": 6.824454148471616,
+ "grad_norm": 0.024675287306308746,
+ "learning_rate": 0.0006,
+ "loss": 5.6639909744262695,
+ "step": 491
+ },
+ {
+ "epoch": 6.8384279475982535,
+ "grad_norm": 0.030372392386198044,
+ "learning_rate": 0.0006,
+ "loss": 5.582888603210449,
+ "step": 492
+ },
+ {
+ "epoch": 6.85240174672489,
+ "grad_norm": 0.03237690031528473,
+ "learning_rate": 0.0006,
+ "loss": 5.573896884918213,
+ "step": 493
+ },
+ {
+ "epoch": 6.866375545851528,
+ "grad_norm": 0.03511589393019676,
+ "learning_rate": 0.0006,
+ "loss": 5.565878868103027,
+ "step": 494
+ },
+ {
+ "epoch": 6.880349344978166,
+ "grad_norm": 0.03883938118815422,
+ "learning_rate": 0.0006,
+ "loss": 5.58540678024292,
+ "step": 495
+ },
+ {
+ "epoch": 6.894323144104804,
+ "grad_norm": 0.04175502806901932,
+ "learning_rate": 0.0006,
+ "loss": 5.6095194816589355,
+ "step": 496
+ },
+ {
+ "epoch": 6.908296943231441,
+ "grad_norm": 0.04013441503047943,
+ "learning_rate": 0.0006,
+ "loss": 5.654401779174805,
+ "step": 497
+ },
+ {
+ "epoch": 6.922270742358078,
+ "grad_norm": 0.04618770629167557,
+ "learning_rate": 0.0006,
+ "loss": 5.6483473777771,
+ "step": 498
+ },
+ {
+ "epoch": 6.936244541484716,
+ "grad_norm": 0.04507308453321457,
+ "learning_rate": 0.0006,
+ "loss": 5.5856218338012695,
+ "step": 499
+ },
+ {
+ "epoch": 6.950218340611354,
+ "grad_norm": 0.03151383996009827,
+ "learning_rate": 0.0006,
+ "loss": 5.594086647033691,
+ "step": 500
+ },
+ {
+ "epoch": 6.964192139737992,
+ "grad_norm": 0.026883797720074654,
+ "learning_rate": 0.0006,
+ "loss": 5.607676029205322,
+ "step": 501
+ },
+ {
+ "epoch": 6.978165938864628,
+ "grad_norm": 0.02981836162507534,
+ "learning_rate": 0.0006,
+ "loss": 5.563666820526123,
+ "step": 502
+ },
+ {
+ "epoch": 6.992139737991266,
+ "grad_norm": 0.03233477845788002,
+ "learning_rate": 0.0006,
+ "loss": 5.667543411254883,
+ "step": 503
+ },
+ {
+ "epoch": 7.0,
+ "grad_norm": 0.029927456751465797,
+ "learning_rate": 0.0006,
+ "loss": 5.733482837677002,
+ "step": 504
+ },
+ {
+ "epoch": 7.0,
+ "eval_loss": 5.679076671600342,
+ "eval_runtime": 59.5965,
+ "eval_samples_per_second": 40.976,
+ "eval_steps_per_second": 1.292,
+ "step": 504
+ },
+ {
+ "epoch": 7.013973799126638,
+ "grad_norm": 0.027081597596406937,
+ "learning_rate": 0.0006,
+ "loss": 5.634487152099609,
+ "step": 505
+ },
+ {
+ "epoch": 7.0279475982532755,
+ "grad_norm": 0.031959421932697296,
+ "learning_rate": 0.0006,
+ "loss": 5.643294811248779,
+ "step": 506
+ },
+ {
+ "epoch": 7.041921397379912,
+ "grad_norm": 0.02802063524723053,
+ "learning_rate": 0.0006,
+ "loss": 5.573239326477051,
+ "step": 507
+ },
+ {
+ "epoch": 7.05589519650655,
+ "grad_norm": 0.032579705119132996,
+ "learning_rate": 0.0006,
+ "loss": 5.627150058746338,
+ "step": 508
+ },
+ {
+ "epoch": 7.069868995633188,
+ "grad_norm": 0.02503928542137146,
+ "learning_rate": 0.0006,
+ "loss": 5.6045308113098145,
+ "step": 509
+ },
+ {
+ "epoch": 7.083842794759826,
+ "grad_norm": 0.02421317622065544,
+ "learning_rate": 0.0006,
+ "loss": 5.522153377532959,
+ "step": 510
+ },
+ {
+ "epoch": 7.097816593886463,
+ "grad_norm": 0.021983426064252853,
+ "learning_rate": 0.0006,
+ "loss": 5.632939338684082,
+ "step": 511
+ },
+ {
+ "epoch": 7.1117903930131,
+ "grad_norm": 0.021933183073997498,
+ "learning_rate": 0.0006,
+ "loss": 5.592185974121094,
+ "step": 512
+ },
+ {
+ "epoch": 7.125764192139738,
+ "grad_norm": 0.02287031151354313,
+ "learning_rate": 0.0006,
+ "loss": 5.613700866699219,
+ "step": 513
+ },
+ {
+ "epoch": 7.139737991266376,
+ "grad_norm": 0.02281602844595909,
+ "learning_rate": 0.0006,
+ "loss": 5.551383018493652,
+ "step": 514
+ },
+ {
+ "epoch": 7.153711790393013,
+ "grad_norm": 0.021050360053777695,
+ "learning_rate": 0.0006,
+ "loss": 5.55275821685791,
+ "step": 515
+ },
+ {
+ "epoch": 7.1676855895196505,
+ "grad_norm": 0.018299926072359085,
+ "learning_rate": 0.0006,
+ "loss": 5.528225421905518,
+ "step": 516
+ },
+ {
+ "epoch": 7.181659388646288,
+ "grad_norm": 0.02024853602051735,
+ "learning_rate": 0.0006,
+ "loss": 5.522453784942627,
+ "step": 517
+ },
+ {
+ "epoch": 7.195633187772926,
+ "grad_norm": 0.02135239914059639,
+ "learning_rate": 0.0006,
+ "loss": 5.527522563934326,
+ "step": 518
+ },
+ {
+ "epoch": 7.209606986899563,
+ "grad_norm": 0.023394184187054634,
+ "learning_rate": 0.0006,
+ "loss": 5.4535017013549805,
+ "step": 519
+ },
+ {
+ "epoch": 7.223580786026201,
+ "grad_norm": 0.025952080264687538,
+ "learning_rate": 0.0006,
+ "loss": 5.5374674797058105,
+ "step": 520
+ },
+ {
+ "epoch": 7.2375545851528384,
+ "grad_norm": 0.022287718951702118,
+ "learning_rate": 0.0006,
+ "loss": 5.493753433227539,
+ "step": 521
+ },
+ {
+ "epoch": 7.251528384279476,
+ "grad_norm": 0.02018447406589985,
+ "learning_rate": 0.0006,
+ "loss": 5.510575294494629,
+ "step": 522
+ },
+ {
+ "epoch": 7.265502183406113,
+ "grad_norm": 0.02050507813692093,
+ "learning_rate": 0.0006,
+ "loss": 5.466026782989502,
+ "step": 523
+ },
+ {
+ "epoch": 7.279475982532751,
+ "grad_norm": 0.023688802495598793,
+ "learning_rate": 0.0006,
+ "loss": 5.485815525054932,
+ "step": 524
+ },
+ {
+ "epoch": 7.293449781659389,
+ "grad_norm": 0.021078843623399734,
+ "learning_rate": 0.0006,
+ "loss": 5.415581703186035,
+ "step": 525
+ },
+ {
+ "epoch": 7.307423580786026,
+ "grad_norm": 0.018266011029481888,
+ "learning_rate": 0.0006,
+ "loss": 5.402815818786621,
+ "step": 526
+ },
+ {
+ "epoch": 7.321397379912664,
+ "grad_norm": 0.019693493843078613,
+ "learning_rate": 0.0006,
+ "loss": 5.504674911499023,
+ "step": 527
+ },
+ {
+ "epoch": 7.335371179039301,
+ "grad_norm": 0.02031373605132103,
+ "learning_rate": 0.0006,
+ "loss": 5.503837585449219,
+ "step": 528
+ },
+ {
+ "epoch": 7.349344978165939,
+ "grad_norm": 0.017636626958847046,
+ "learning_rate": 0.0006,
+ "loss": 5.530580997467041,
+ "step": 529
+ },
+ {
+ "epoch": 7.3633187772925766,
+ "grad_norm": 0.01787244901061058,
+ "learning_rate": 0.0006,
+ "loss": 5.450218677520752,
+ "step": 530
+ },
+ {
+ "epoch": 7.377292576419214,
+ "grad_norm": 0.0170669574290514,
+ "learning_rate": 0.0006,
+ "loss": 5.481570243835449,
+ "step": 531
+ },
+ {
+ "epoch": 7.391266375545851,
+ "grad_norm": 0.01802165061235428,
+ "learning_rate": 0.0006,
+ "loss": 5.386394500732422,
+ "step": 532
+ },
+ {
+ "epoch": 7.405240174672489,
+ "grad_norm": 0.022949472069740295,
+ "learning_rate": 0.0006,
+ "loss": 5.424929618835449,
+ "step": 533
+ },
+ {
+ "epoch": 7.419213973799127,
+ "grad_norm": 0.034213390201330185,
+ "learning_rate": 0.0006,
+ "loss": 5.346663475036621,
+ "step": 534
+ },
+ {
+ "epoch": 7.4331877729257645,
+ "grad_norm": 0.03847593814134598,
+ "learning_rate": 0.0006,
+ "loss": 5.418482780456543,
+ "step": 535
+ },
+ {
+ "epoch": 7.447161572052401,
+ "grad_norm": 0.029393676668405533,
+ "learning_rate": 0.0006,
+ "loss": 5.456090927124023,
+ "step": 536
+ },
+ {
+ "epoch": 7.461135371179039,
+ "grad_norm": 0.03388667106628418,
+ "learning_rate": 0.0006,
+ "loss": 5.572983741760254,
+ "step": 537
+ },
+ {
+ "epoch": 7.475109170305677,
+ "grad_norm": 0.039690613746643066,
+ "learning_rate": 0.0006,
+ "loss": 5.414067268371582,
+ "step": 538
+ },
+ {
+ "epoch": 7.489082969432315,
+ "grad_norm": 0.03630776330828667,
+ "learning_rate": 0.0006,
+ "loss": 5.523739814758301,
+ "step": 539
+ },
+ {
+ "epoch": 7.503056768558952,
+ "grad_norm": 0.03356870263814926,
+ "learning_rate": 0.0006,
+ "loss": 5.444767951965332,
+ "step": 540
+ },
+ {
+ "epoch": 7.517030567685589,
+ "grad_norm": 0.030812304466962814,
+ "learning_rate": 0.0006,
+ "loss": 5.461244106292725,
+ "step": 541
+ },
+ {
+ "epoch": 7.531004366812227,
+ "grad_norm": 0.03319217637181282,
+ "learning_rate": 0.0006,
+ "loss": 5.502161026000977,
+ "step": 542
+ },
+ {
+ "epoch": 7.544978165938865,
+ "grad_norm": 0.032997481524944305,
+ "learning_rate": 0.0006,
+ "loss": 5.4162278175354,
+ "step": 543
+ },
+ {
+ "epoch": 7.558951965065502,
+ "grad_norm": 0.03364962339401245,
+ "learning_rate": 0.0006,
+ "loss": 5.415736198425293,
+ "step": 544
+ },
+ {
+ "epoch": 7.5729257641921395,
+ "grad_norm": 0.0344221405684948,
+ "learning_rate": 0.0006,
+ "loss": 5.541967868804932,
+ "step": 545
+ },
+ {
+ "epoch": 7.586899563318777,
+ "grad_norm": 0.029609503224492073,
+ "learning_rate": 0.0006,
+ "loss": 5.440614223480225,
+ "step": 546
+ },
+ {
+ "epoch": 7.600873362445415,
+ "grad_norm": 0.029055926948785782,
+ "learning_rate": 0.0006,
+ "loss": 5.462865352630615,
+ "step": 547
+ },
+ {
+ "epoch": 7.614847161572053,
+ "grad_norm": 0.02658848837018013,
+ "learning_rate": 0.0006,
+ "loss": 5.444467544555664,
+ "step": 548
+ },
+ {
+ "epoch": 7.62882096069869,
+ "grad_norm": 0.026276404038071632,
+ "learning_rate": 0.0006,
+ "loss": 5.462278842926025,
+ "step": 549
+ },
+ {
+ "epoch": 7.642794759825327,
+ "grad_norm": 0.0286889486014843,
+ "learning_rate": 0.0006,
+ "loss": 5.4849534034729,
+ "step": 550
+ },
+ {
+ "epoch": 7.656768558951965,
+ "grad_norm": 0.026849817484617233,
+ "learning_rate": 0.0006,
+ "loss": 5.431473731994629,
+ "step": 551
+ },
+ {
+ "epoch": 7.670742358078603,
+ "grad_norm": 0.02312396466732025,
+ "learning_rate": 0.0006,
+ "loss": 5.45462703704834,
+ "step": 552
+ },
+ {
+ "epoch": 7.68471615720524,
+ "grad_norm": 0.026197485625743866,
+ "learning_rate": 0.0006,
+ "loss": 5.452552318572998,
+ "step": 553
+ },
+ {
+ "epoch": 7.698689956331878,
+ "grad_norm": 0.02747255191206932,
+ "learning_rate": 0.0006,
+ "loss": 5.447169780731201,
+ "step": 554
+ },
+ {
+ "epoch": 7.712663755458515,
+ "grad_norm": 0.028123432770371437,
+ "learning_rate": 0.0006,
+ "loss": 5.394400596618652,
+ "step": 555
+ },
+ {
+ "epoch": 7.726637554585153,
+ "grad_norm": 0.02599870041012764,
+ "learning_rate": 0.0006,
+ "loss": 5.398341178894043,
+ "step": 556
+ },
+ {
+ "epoch": 7.74061135371179,
+ "grad_norm": 0.022171657532453537,
+ "learning_rate": 0.0006,
+ "loss": 5.368820667266846,
+ "step": 557
+ },
+ {
+ "epoch": 7.754585152838428,
+ "grad_norm": 0.022309480234980583,
+ "learning_rate": 0.0006,
+ "loss": 5.416772365570068,
+ "step": 558
+ },
+ {
+ "epoch": 7.7685589519650655,
+ "grad_norm": 0.024074165150523186,
+ "learning_rate": 0.0006,
+ "loss": 5.359002113342285,
+ "step": 559
+ },
+ {
+ "epoch": 7.782532751091703,
+ "grad_norm": 0.02636653557419777,
+ "learning_rate": 0.0006,
+ "loss": 5.484368324279785,
+ "step": 560
+ },
+ {
+ "epoch": 7.796506550218341,
+ "grad_norm": 0.02203752100467682,
+ "learning_rate": 0.0006,
+ "loss": 5.375498294830322,
+ "step": 561
+ },
+ {
+ "epoch": 7.810480349344978,
+ "grad_norm": 0.019924819469451904,
+ "learning_rate": 0.0006,
+ "loss": 5.334723472595215,
+ "step": 562
+ },
+ {
+ "epoch": 7.824454148471616,
+ "grad_norm": 0.018755966797471046,
+ "learning_rate": 0.0006,
+ "loss": 5.348781108856201,
+ "step": 563
+ },
+ {
+ "epoch": 7.8384279475982535,
+ "grad_norm": 0.016747845336794853,
+ "learning_rate": 0.0006,
+ "loss": 5.433079719543457,
+ "step": 564
+ },
+ {
+ "epoch": 7.85240174672489,
+ "grad_norm": 0.0170395877212286,
+ "learning_rate": 0.0006,
+ "loss": 5.346794128417969,
+ "step": 565
+ },
+ {
+ "epoch": 7.866375545851528,
+ "grad_norm": 0.016074176877737045,
+ "learning_rate": 0.0006,
+ "loss": 5.3277740478515625,
+ "step": 566
+ },
+ {
+ "epoch": 7.880349344978166,
+ "grad_norm": 0.014119806699454784,
+ "learning_rate": 0.0006,
+ "loss": 5.483551979064941,
+ "step": 567
+ },
+ {
+ "epoch": 7.894323144104804,
+ "grad_norm": 0.015270394273102283,
+ "learning_rate": 0.0006,
+ "loss": 5.378519535064697,
+ "step": 568
+ },
+ {
+ "epoch": 7.908296943231441,
+ "grad_norm": 0.015677539631724358,
+ "learning_rate": 0.0006,
+ "loss": 5.290210247039795,
+ "step": 569
+ },
+ {
+ "epoch": 7.922270742358078,
+ "grad_norm": 0.015930423513054848,
+ "learning_rate": 0.0006,
+ "loss": 5.427360534667969,
+ "step": 570
+ },
+ {
+ "epoch": 7.936244541484716,
+ "grad_norm": 0.016641564667224884,
+ "learning_rate": 0.0006,
+ "loss": 5.301599979400635,
+ "step": 571
+ },
+ {
+ "epoch": 7.950218340611354,
+ "grad_norm": 0.018293552100658417,
+ "learning_rate": 0.0006,
+ "loss": 5.277889251708984,
+ "step": 572
+ },
+ {
+ "epoch": 7.964192139737992,
+ "grad_norm": 0.017618799582123756,
+ "learning_rate": 0.0006,
+ "loss": 5.297194480895996,
+ "step": 573
+ },
+ {
+ "epoch": 7.978165938864628,
+ "grad_norm": 0.015034242533147335,
+ "learning_rate": 0.0006,
+ "loss": 5.42428731918335,
+ "step": 574
+ },
+ {
+ "epoch": 7.992139737991266,
+ "grad_norm": 0.01617511734366417,
+ "learning_rate": 0.0006,
+ "loss": 5.420044898986816,
+ "step": 575
+ },
+ {
+ "epoch": 8.0,
+ "grad_norm": 0.01837257295846939,
+ "learning_rate": 0.0006,
+ "loss": 5.240387916564941,
+ "step": 576
+ },
+ {
+ "epoch": 8.0,
+ "eval_loss": 5.423932075500488,
+ "eval_runtime": 58.0841,
+ "eval_samples_per_second": 42.042,
+ "eval_steps_per_second": 1.326,
+ "step": 576
+ },
+ {
+ "epoch": 8.013973799126637,
+ "grad_norm": 0.020953809842467308,
+ "learning_rate": 0.0006,
+ "loss": 5.4320783615112305,
+ "step": 577
+ },
+ {
+ "epoch": 8.027947598253276,
+ "grad_norm": 0.0227745920419693,
+ "learning_rate": 0.0006,
+ "loss": 5.310197830200195,
+ "step": 578
+ },
+ {
+ "epoch": 8.041921397379912,
+ "grad_norm": 0.021372603252530098,
+ "learning_rate": 0.0006,
+ "loss": 5.339468002319336,
+ "step": 579
+ },
+ {
+ "epoch": 8.055895196506551,
+ "grad_norm": 0.022401731461286545,
+ "learning_rate": 0.0006,
+ "loss": 5.279829978942871,
+ "step": 580
+ },
+ {
+ "epoch": 8.069868995633188,
+ "grad_norm": 0.023942379280924797,
+ "learning_rate": 0.0006,
+ "loss": 5.372323989868164,
+ "step": 581
+ },
+ {
+ "epoch": 8.083842794759825,
+ "grad_norm": 0.024380534887313843,
+ "learning_rate": 0.0006,
+ "loss": 5.253486633300781,
+ "step": 582
+ },
+ {
+ "epoch": 8.097816593886463,
+ "grad_norm": 0.03246625140309334,
+ "learning_rate": 0.0006,
+ "loss": 5.305774211883545,
+ "step": 583
+ },
+ {
+ "epoch": 8.1117903930131,
+ "grad_norm": 0.030399736016988754,
+ "learning_rate": 0.0006,
+ "loss": 5.370976448059082,
+ "step": 584
+ },
+ {
+ "epoch": 8.125764192139737,
+ "grad_norm": 0.027148913592100143,
+ "learning_rate": 0.0006,
+ "loss": 5.368010520935059,
+ "step": 585
+ },
+ {
+ "epoch": 8.139737991266376,
+ "grad_norm": 0.02925540879368782,
+ "learning_rate": 0.0006,
+ "loss": 5.303013801574707,
+ "step": 586
+ },
+ {
+ "epoch": 8.153711790393013,
+ "grad_norm": 0.02915453538298607,
+ "learning_rate": 0.0006,
+ "loss": 5.329667091369629,
+ "step": 587
+ },
+ {
+ "epoch": 8.167685589519651,
+ "grad_norm": 0.031732626259326935,
+ "learning_rate": 0.0006,
+ "loss": 5.3634934425354,
+ "step": 588
+ },
+ {
+ "epoch": 8.181659388646288,
+ "grad_norm": 0.036166731268167496,
+ "learning_rate": 0.0006,
+ "loss": 5.223980903625488,
+ "step": 589
+ },
+ {
+ "epoch": 8.195633187772925,
+ "grad_norm": 0.035654108971357346,
+ "learning_rate": 0.0006,
+ "loss": 5.361059188842773,
+ "step": 590
+ },
+ {
+ "epoch": 8.209606986899564,
+ "grad_norm": 0.0365324430167675,
+ "learning_rate": 0.0006,
+ "loss": 5.336151599884033,
+ "step": 591
+ },
+ {
+ "epoch": 8.2235807860262,
+ "grad_norm": 0.03578880429267883,
+ "learning_rate": 0.0006,
+ "loss": 5.434076309204102,
+ "step": 592
+ },
+ {
+ "epoch": 8.237554585152838,
+ "grad_norm": 0.035734012722969055,
+ "learning_rate": 0.0006,
+ "loss": 5.330893516540527,
+ "step": 593
+ },
+ {
+ "epoch": 8.251528384279476,
+ "grad_norm": 0.03196857124567032,
+ "learning_rate": 0.0006,
+ "loss": 5.306009769439697,
+ "step": 594
+ },
+ {
+ "epoch": 8.265502183406113,
+ "grad_norm": 0.032042086124420166,
+ "learning_rate": 0.0006,
+ "loss": 5.343267917633057,
+ "step": 595
+ },
+ {
+ "epoch": 8.279475982532752,
+ "grad_norm": 0.03160746395587921,
+ "learning_rate": 0.0006,
+ "loss": 5.2353715896606445,
+ "step": 596
+ },
+ {
+ "epoch": 8.293449781659389,
+ "grad_norm": 0.026689818128943443,
+ "learning_rate": 0.0006,
+ "loss": 5.254042625427246,
+ "step": 597
+ },
+ {
+ "epoch": 8.307423580786025,
+ "grad_norm": 0.02880188450217247,
+ "learning_rate": 0.0006,
+ "loss": 5.326833248138428,
+ "step": 598
+ },
+ {
+ "epoch": 8.321397379912664,
+ "grad_norm": 0.027516381815075874,
+ "learning_rate": 0.0006,
+ "loss": 5.342336177825928,
+ "step": 599
+ },
+ {
+ "epoch": 8.335371179039301,
+ "grad_norm": 0.027875030413269997,
+ "learning_rate": 0.0006,
+ "loss": 5.200719833374023,
+ "step": 600
+ },
+ {
+ "epoch": 8.34934497816594,
+ "grad_norm": 0.0268265288323164,
+ "learning_rate": 0.0006,
+ "loss": 5.249449729919434,
+ "step": 601
+ },
+ {
+ "epoch": 8.363318777292577,
+ "grad_norm": 0.024824608117341995,
+ "learning_rate": 0.0006,
+ "loss": 5.318952560424805,
+ "step": 602
+ },
+ {
+ "epoch": 8.377292576419213,
+ "grad_norm": 0.022990627214312553,
+ "learning_rate": 0.0006,
+ "loss": 5.244993209838867,
+ "step": 603
+ },
+ {
+ "epoch": 8.391266375545852,
+ "grad_norm": 0.022804604843258858,
+ "learning_rate": 0.0006,
+ "loss": 5.3236846923828125,
+ "step": 604
+ },
+ {
+ "epoch": 8.405240174672489,
+ "grad_norm": 0.02419872209429741,
+ "learning_rate": 0.0006,
+ "loss": 5.255486488342285,
+ "step": 605
+ },
+ {
+ "epoch": 8.419213973799126,
+ "grad_norm": 0.021952059119939804,
+ "learning_rate": 0.0006,
+ "loss": 5.3511552810668945,
+ "step": 606
+ },
+ {
+ "epoch": 8.433187772925764,
+ "grad_norm": 0.022375497967004776,
+ "learning_rate": 0.0006,
+ "loss": 5.294790267944336,
+ "step": 607
+ },
+ {
+ "epoch": 8.447161572052401,
+ "grad_norm": 0.019242815673351288,
+ "learning_rate": 0.0006,
+ "loss": 5.252618789672852,
+ "step": 608
+ },
+ {
+ "epoch": 8.46113537117904,
+ "grad_norm": 0.0173486340790987,
+ "learning_rate": 0.0006,
+ "loss": 5.30147647857666,
+ "step": 609
+ },
+ {
+ "epoch": 8.475109170305677,
+ "grad_norm": 0.01612604409456253,
+ "learning_rate": 0.0006,
+ "loss": 5.23960542678833,
+ "step": 610
+ },
+ {
+ "epoch": 8.489082969432314,
+ "grad_norm": 0.014293976128101349,
+ "learning_rate": 0.0006,
+ "loss": 5.119932651519775,
+ "step": 611
+ },
+ {
+ "epoch": 8.503056768558952,
+ "grad_norm": 0.014453536830842495,
+ "learning_rate": 0.0006,
+ "loss": 5.166794776916504,
+ "step": 612
+ },
+ {
+ "epoch": 8.51703056768559,
+ "grad_norm": 0.014685769565403461,
+ "learning_rate": 0.0006,
+ "loss": 5.264822959899902,
+ "step": 613
+ },
+ {
+ "epoch": 8.531004366812226,
+ "grad_norm": 0.01427740603685379,
+ "learning_rate": 0.0006,
+ "loss": 5.278433799743652,
+ "step": 614
+ },
+ {
+ "epoch": 8.544978165938865,
+ "grad_norm": 0.014864951372146606,
+ "learning_rate": 0.0006,
+ "loss": 5.214512348175049,
+ "step": 615
+ },
+ {
+ "epoch": 8.558951965065502,
+ "grad_norm": 0.01565164513885975,
+ "learning_rate": 0.0006,
+ "loss": 5.183066368103027,
+ "step": 616
+ },
+ {
+ "epoch": 8.57292576419214,
+ "grad_norm": 0.01779816672205925,
+ "learning_rate": 0.0006,
+ "loss": 5.163185119628906,
+ "step": 617
+ },
+ {
+ "epoch": 8.586899563318777,
+ "grad_norm": 0.017254827544093132,
+ "learning_rate": 0.0006,
+ "loss": 5.243555068969727,
+ "step": 618
+ },
+ {
+ "epoch": 8.600873362445414,
+ "grad_norm": 0.01650584116578102,
+ "learning_rate": 0.0006,
+ "loss": 5.234884738922119,
+ "step": 619
+ },
+ {
+ "epoch": 8.614847161572053,
+ "grad_norm": 0.017021115869283676,
+ "learning_rate": 0.0006,
+ "loss": 5.12271785736084,
+ "step": 620
+ },
+ {
+ "epoch": 8.62882096069869,
+ "grad_norm": 0.01773759163916111,
+ "learning_rate": 0.0006,
+ "loss": 5.242153167724609,
+ "step": 621
+ },
+ {
+ "epoch": 8.642794759825328,
+ "grad_norm": 0.015679042786359787,
+ "learning_rate": 0.0006,
+ "loss": 5.170779228210449,
+ "step": 622
+ },
+ {
+ "epoch": 8.656768558951965,
+ "grad_norm": 0.013760549947619438,
+ "learning_rate": 0.0006,
+ "loss": 5.238644599914551,
+ "step": 623
+ },
+ {
+ "epoch": 8.670742358078602,
+ "grad_norm": 0.014571522362530231,
+ "learning_rate": 0.0006,
+ "loss": 5.177056312561035,
+ "step": 624
+ },
+ {
+ "epoch": 8.68471615720524,
+ "grad_norm": 0.016209015622735023,
+ "learning_rate": 0.0006,
+ "loss": 5.163750648498535,
+ "step": 625
+ },
+ {
+ "epoch": 8.698689956331878,
+ "grad_norm": 0.016813941299915314,
+ "learning_rate": 0.0006,
+ "loss": 5.183428764343262,
+ "step": 626
+ },
+ {
+ "epoch": 8.712663755458514,
+ "grad_norm": 0.020985357463359833,
+ "learning_rate": 0.0006,
+ "loss": 5.229465007781982,
+ "step": 627
+ },
+ {
+ "epoch": 8.726637554585153,
+ "grad_norm": 0.02679632417857647,
+ "learning_rate": 0.0006,
+ "loss": 5.257368087768555,
+ "step": 628
+ },
+ {
+ "epoch": 8.74061135371179,
+ "grad_norm": 0.025756070390343666,
+ "learning_rate": 0.0006,
+ "loss": 5.253736972808838,
+ "step": 629
+ },
+ {
+ "epoch": 8.754585152838429,
+ "grad_norm": 0.02643490768969059,
+ "learning_rate": 0.0006,
+ "loss": 5.228633403778076,
+ "step": 630
+ },
+ {
+ "epoch": 8.768558951965066,
+ "grad_norm": 0.03303210437297821,
+ "learning_rate": 0.0006,
+ "loss": 5.163540363311768,
+ "step": 631
+ },
+ {
+ "epoch": 8.782532751091702,
+ "grad_norm": 0.030432431027293205,
+ "learning_rate": 0.0006,
+ "loss": 5.243169784545898,
+ "step": 632
+ },
+ {
+ "epoch": 8.796506550218341,
+ "grad_norm": 0.028631288558244705,
+ "learning_rate": 0.0006,
+ "loss": 5.166398048400879,
+ "step": 633
+ },
+ {
+ "epoch": 8.810480349344978,
+ "grad_norm": 0.026188310235738754,
+ "learning_rate": 0.0006,
+ "loss": 5.128122329711914,
+ "step": 634
+ },
+ {
+ "epoch": 8.824454148471617,
+ "grad_norm": 0.025308528915047646,
+ "learning_rate": 0.0006,
+ "loss": 5.211284637451172,
+ "step": 635
+ },
+ {
+ "epoch": 8.838427947598253,
+ "grad_norm": 0.020657729357481003,
+ "learning_rate": 0.0006,
+ "loss": 5.161575794219971,
+ "step": 636
+ },
+ {
+ "epoch": 8.85240174672489,
+ "grad_norm": 0.021901234984397888,
+ "learning_rate": 0.0006,
+ "loss": 5.20050048828125,
+ "step": 637
+ },
+ {
+ "epoch": 8.866375545851529,
+ "grad_norm": 0.021219318732619286,
+ "learning_rate": 0.0006,
+ "loss": 5.211699485778809,
+ "step": 638
+ },
+ {
+ "epoch": 8.880349344978166,
+ "grad_norm": 0.0200974028557539,
+ "learning_rate": 0.0006,
+ "loss": 5.092503547668457,
+ "step": 639
+ },
+ {
+ "epoch": 8.894323144104803,
+ "grad_norm": 0.023804882541298866,
+ "learning_rate": 0.0006,
+ "loss": 5.216068267822266,
+ "step": 640
+ },
+ {
+ "epoch": 8.908296943231441,
+ "grad_norm": 0.026088010519742966,
+ "learning_rate": 0.0006,
+ "loss": 5.155592918395996,
+ "step": 641
+ },
+ {
+ "epoch": 8.922270742358078,
+ "grad_norm": 0.02601276896893978,
+ "learning_rate": 0.0006,
+ "loss": 5.197238922119141,
+ "step": 642
+ },
+ {
+ "epoch": 8.936244541484717,
+ "grad_norm": 0.020387211814522743,
+ "learning_rate": 0.0006,
+ "loss": 5.1955437660217285,
+ "step": 643
+ },
+ {
+ "epoch": 8.950218340611354,
+ "grad_norm": 0.019214622676372528,
+ "learning_rate": 0.0006,
+ "loss": 5.206346035003662,
+ "step": 644
+ },
+ {
+ "epoch": 8.96419213973799,
+ "grad_norm": 0.019674314185976982,
+ "learning_rate": 0.0006,
+ "loss": 5.219857215881348,
+ "step": 645
+ },
+ {
+ "epoch": 8.97816593886463,
+ "grad_norm": 0.020318234339356422,
+ "learning_rate": 0.0006,
+ "loss": 5.191132545471191,
+ "step": 646
+ },
+ {
+ "epoch": 8.992139737991266,
+ "grad_norm": 0.021428676322102547,
+ "learning_rate": 0.0006,
+ "loss": 5.214695930480957,
+ "step": 647
+ },
+ {
+ "epoch": 9.0,
+ "grad_norm": 0.023235054686665535,
+ "learning_rate": 0.0006,
+ "loss": 5.229091644287109,
+ "step": 648
+ },
+ {
+ "epoch": 9.0,
+ "eval_loss": 5.289593696594238,
+ "eval_runtime": 56.9256,
+ "eval_samples_per_second": 42.898,
+ "eval_steps_per_second": 1.353,
+ "step": 648
+ },
+ {
+ "epoch": 9.013973799126637,
+ "grad_norm": 0.02320289984345436,
+ "learning_rate": 0.0006,
+ "loss": 4.983669281005859,
+ "step": 649
+ },
+ {
+ "epoch": 9.027947598253276,
+ "grad_norm": 0.018521282821893692,
+ "learning_rate": 0.0006,
+ "loss": 5.200575828552246,
+ "step": 650
+ },
+ {
+ "epoch": 9.041921397379912,
+ "grad_norm": 0.0180169939994812,
+ "learning_rate": 0.0006,
+ "loss": 5.148033618927002,
+ "step": 651
+ },
+ {
+ "epoch": 9.055895196506551,
+ "grad_norm": 0.019969860091805458,
+ "learning_rate": 0.0006,
+ "loss": 5.099125385284424,
+ "step": 652
+ },
+ {
+ "epoch": 9.069868995633188,
+ "grad_norm": 0.01643305830657482,
+ "learning_rate": 0.0006,
+ "loss": 5.092347145080566,
+ "step": 653
+ },
+ {
+ "epoch": 9.083842794759825,
+ "grad_norm": 0.016022363677620888,
+ "learning_rate": 0.0006,
+ "loss": 5.129279613494873,
+ "step": 654
+ },
+ {
+ "epoch": 9.097816593886463,
+ "grad_norm": 0.01657041162252426,
+ "learning_rate": 0.0006,
+ "loss": 5.159562587738037,
+ "step": 655
+ },
+ {
+ "epoch": 9.1117903930131,
+ "grad_norm": 0.017924228683114052,
+ "learning_rate": 0.0006,
+ "loss": 5.165497779846191,
+ "step": 656
+ },
+ {
+ "epoch": 9.125764192139737,
+ "grad_norm": 0.021592361852526665,
+ "learning_rate": 0.0006,
+ "loss": 5.204775810241699,
+ "step": 657
+ },
+ {
+ "epoch": 9.139737991266376,
+ "grad_norm": 0.026924120262265205,
+ "learning_rate": 0.0006,
+ "loss": 5.128296852111816,
+ "step": 658
+ },
+ {
+ "epoch": 9.153711790393013,
+ "grad_norm": 0.023156899958848953,
+ "learning_rate": 0.0006,
+ "loss": 5.1255340576171875,
+ "step": 659
+ },
+ {
+ "epoch": 9.167685589519651,
+ "grad_norm": 0.020670367404818535,
+ "learning_rate": 0.0006,
+ "loss": 5.128479957580566,
+ "step": 660
+ },
+ {
+ "epoch": 9.181659388646288,
+ "grad_norm": 0.025447173044085503,
+ "learning_rate": 0.0006,
+ "loss": 5.154821872711182,
+ "step": 661
+ },
+ {
+ "epoch": 9.195633187772925,
+ "grad_norm": 0.027940068393945694,
+ "learning_rate": 0.0006,
+ "loss": 5.143466472625732,
+ "step": 662
+ },
+ {
+ "epoch": 9.209606986899564,
+ "grad_norm": 0.026847844943404198,
+ "learning_rate": 0.0006,
+ "loss": 5.045665740966797,
+ "step": 663
+ },
+ {
+ "epoch": 9.2235807860262,
+ "grad_norm": 0.02340601570904255,
+ "learning_rate": 0.0006,
+ "loss": 5.21933126449585,
+ "step": 664
+ },
+ {
+ "epoch": 9.237554585152838,
+ "grad_norm": 0.02340371161699295,
+ "learning_rate": 0.0006,
+ "loss": 4.942949295043945,
+ "step": 665
+ },
+ {
+ "epoch": 9.251528384279476,
+ "grad_norm": 0.02221992425620556,
+ "learning_rate": 0.0006,
+ "loss": 5.197099208831787,
+ "step": 666
+ },
+ {
+ "epoch": 9.265502183406113,
+ "grad_norm": 0.023508677259087563,
+ "learning_rate": 0.0006,
+ "loss": 5.022897720336914,
+ "step": 667
+ },
+ {
+ "epoch": 9.279475982532752,
+ "grad_norm": 0.026120394468307495,
+ "learning_rate": 0.0006,
+ "loss": 5.087725639343262,
+ "step": 668
+ },
+ {
+ "epoch": 9.293449781659389,
+ "grad_norm": 0.027273228392004967,
+ "learning_rate": 0.0006,
+ "loss": 5.090963840484619,
+ "step": 669
+ },
+ {
+ "epoch": 9.307423580786025,
+ "grad_norm": 0.03241586685180664,
+ "learning_rate": 0.0006,
+ "loss": 5.047842979431152,
+ "step": 670
+ },
+ {
+ "epoch": 9.321397379912664,
+ "grad_norm": 0.030245667323470116,
+ "learning_rate": 0.0006,
+ "loss": 5.104803085327148,
+ "step": 671
+ },
+ {
+ "epoch": 9.335371179039301,
+ "grad_norm": 0.027698364108800888,
+ "learning_rate": 0.0006,
+ "loss": 5.0644097328186035,
+ "step": 672
+ },
+ {
+ "epoch": 9.34934497816594,
+ "grad_norm": 0.029692554846405983,
+ "learning_rate": 0.0006,
+ "loss": 5.111942291259766,
+ "step": 673
+ },
+ {
+ "epoch": 9.363318777292577,
+ "grad_norm": 0.03328656405210495,
+ "learning_rate": 0.0006,
+ "loss": 5.1574554443359375,
+ "step": 674
+ },
+ {
+ "epoch": 9.377292576419213,
+ "grad_norm": 0.031938180327415466,
+ "learning_rate": 0.0006,
+ "loss": 5.190371990203857,
+ "step": 675
+ },
+ {
+ "epoch": 9.391266375545852,
+ "grad_norm": 0.024858945980668068,
+ "learning_rate": 0.0006,
+ "loss": 5.121420860290527,
+ "step": 676
+ },
+ {
+ "epoch": 9.405240174672489,
+ "grad_norm": 0.023033203557133675,
+ "learning_rate": 0.0006,
+ "loss": 5.0974345207214355,
+ "step": 677
+ },
+ {
+ "epoch": 9.419213973799126,
+ "grad_norm": 0.022083545103669167,
+ "learning_rate": 0.0006,
+ "loss": 5.046319961547852,
+ "step": 678
+ },
+ {
+ "epoch": 9.433187772925764,
+ "grad_norm": 0.0228437427431345,
+ "learning_rate": 0.0006,
+ "loss": 5.119963645935059,
+ "step": 679
+ },
+ {
+ "epoch": 9.447161572052401,
+ "grad_norm": 0.0242351982742548,
+ "learning_rate": 0.0006,
+ "loss": 5.0444560050964355,
+ "step": 680
+ },
+ {
+ "epoch": 9.46113537117904,
+ "grad_norm": 0.02401185780763626,
+ "learning_rate": 0.0006,
+ "loss": 5.063238620758057,
+ "step": 681
+ },
+ {
+ "epoch": 9.475109170305677,
+ "grad_norm": 0.02560959942638874,
+ "learning_rate": 0.0006,
+ "loss": 5.035224914550781,
+ "step": 682
+ },
+ {
+ "epoch": 9.489082969432314,
+ "grad_norm": 0.024985626339912415,
+ "learning_rate": 0.0006,
+ "loss": 5.049284934997559,
+ "step": 683
+ },
+ {
+ "epoch": 9.503056768558952,
+ "grad_norm": 0.023178115487098694,
+ "learning_rate": 0.0006,
+ "loss": 5.2266621589660645,
+ "step": 684
+ },
+ {
+ "epoch": 9.51703056768559,
+ "grad_norm": 0.0215512253344059,
+ "learning_rate": 0.0006,
+ "loss": 5.1875715255737305,
+ "step": 685
+ },
+ {
+ "epoch": 9.531004366812226,
+ "grad_norm": 0.01662837713956833,
+ "learning_rate": 0.0006,
+ "loss": 5.137012481689453,
+ "step": 686
+ },
+ {
+ "epoch": 9.544978165938865,
+ "grad_norm": 0.01952764391899109,
+ "learning_rate": 0.0006,
+ "loss": 5.072500228881836,
+ "step": 687
+ },
+ {
+ "epoch": 9.558951965065502,
+ "grad_norm": 0.017521077767014503,
+ "learning_rate": 0.0006,
+ "loss": 5.045711994171143,
+ "step": 688
+ },
+ {
+ "epoch": 9.57292576419214,
+ "grad_norm": 0.017355265095829964,
+ "learning_rate": 0.0006,
+ "loss": 5.02881383895874,
+ "step": 689
+ },
+ {
+ "epoch": 9.586899563318777,
+ "grad_norm": 0.015587719157338142,
+ "learning_rate": 0.0006,
+ "loss": 5.132787704467773,
+ "step": 690
+ },
+ {
+ "epoch": 9.600873362445414,
+ "grad_norm": 0.017469845712184906,
+ "learning_rate": 0.0006,
+ "loss": 5.105447292327881,
+ "step": 691
+ },
+ {
+ "epoch": 9.614847161572053,
+ "grad_norm": 0.019630100578069687,
+ "learning_rate": 0.0006,
+ "loss": 5.032464981079102,
+ "step": 692
+ },
+ {
+ "epoch": 9.62882096069869,
+ "grad_norm": 0.01784994639456272,
+ "learning_rate": 0.0006,
+ "loss": 5.081345558166504,
+ "step": 693
+ },
+ {
+ "epoch": 9.642794759825328,
+ "grad_norm": 0.01872754842042923,
+ "learning_rate": 0.0006,
+ "loss": 4.9772114753723145,
+ "step": 694
+ },
+ {
+ "epoch": 9.656768558951965,
+ "grad_norm": 0.01999032311141491,
+ "learning_rate": 0.0006,
+ "loss": 5.089740753173828,
+ "step": 695
+ },
+ {
+ "epoch": 9.670742358078602,
+ "grad_norm": 0.01710767112672329,
+ "learning_rate": 0.0006,
+ "loss": 4.9869890213012695,
+ "step": 696
+ },
+ {
+ "epoch": 9.68471615720524,
+ "grad_norm": 0.015144161880016327,
+ "learning_rate": 0.0006,
+ "loss": 5.022564888000488,
+ "step": 697
+ },
+ {
+ "epoch": 9.698689956331878,
+ "grad_norm": 0.01618543453514576,
+ "learning_rate": 0.0006,
+ "loss": 5.035921573638916,
+ "step": 698
+ },
+ {
+ "epoch": 9.712663755458514,
+ "grad_norm": 0.014653387479484081,
+ "learning_rate": 0.0006,
+ "loss": 4.914456367492676,
+ "step": 699
+ },
+ {
+ "epoch": 9.726637554585153,
+ "grad_norm": 0.013434977270662785,
+ "learning_rate": 0.0006,
+ "loss": 5.063764572143555,
+ "step": 700
+ },
+ {
+ "epoch": 9.74061135371179,
+ "grad_norm": 0.015128864906728268,
+ "learning_rate": 0.0006,
+ "loss": 5.052367687225342,
+ "step": 701
+ },
+ {
+ "epoch": 9.754585152838429,
+ "grad_norm": 0.014210768043994904,
+ "learning_rate": 0.0006,
+ "loss": 5.016489028930664,
+ "step": 702
+ },
+ {
+ "epoch": 9.768558951965066,
+ "grad_norm": 0.015191680751740932,
+ "learning_rate": 0.0006,
+ "loss": 4.990176200866699,
+ "step": 703
+ },
+ {
+ "epoch": 9.782532751091702,
+ "grad_norm": 0.018285127356648445,
+ "learning_rate": 0.0006,
+ "loss": 5.015164375305176,
+ "step": 704
+ },
+ {
+ "epoch": 9.796506550218341,
+ "grad_norm": 0.0204079058021307,
+ "learning_rate": 0.0006,
+ "loss": 5.091142654418945,
+ "step": 705
+ },
+ {
+ "epoch": 9.810480349344978,
+ "grad_norm": 0.01980586163699627,
+ "learning_rate": 0.0006,
+ "loss": 5.03350830078125,
+ "step": 706
+ },
+ {
+ "epoch": 9.824454148471617,
+ "grad_norm": 0.02048729732632637,
+ "learning_rate": 0.0006,
+ "loss": 5.03611421585083,
+ "step": 707
+ },
+ {
+ "epoch": 9.838427947598253,
+ "grad_norm": 0.02058163844048977,
+ "learning_rate": 0.0006,
+ "loss": 4.9613447189331055,
+ "step": 708
+ },
+ {
+ "epoch": 9.85240174672489,
+ "grad_norm": 0.018004924058914185,
+ "learning_rate": 0.0006,
+ "loss": 4.966541767120361,
+ "step": 709
+ },
+ {
+ "epoch": 9.866375545851529,
+ "grad_norm": 0.0198152307420969,
+ "learning_rate": 0.0006,
+ "loss": 5.01750373840332,
+ "step": 710
+ },
+ {
+ "epoch": 9.880349344978166,
+ "grad_norm": 0.018532969057559967,
+ "learning_rate": 0.0006,
+ "loss": 5.097424507141113,
+ "step": 711
+ },
+ {
+ "epoch": 9.894323144104803,
+ "grad_norm": 0.018939971923828125,
+ "learning_rate": 0.0006,
+ "loss": 5.177223205566406,
+ "step": 712
+ },
+ {
+ "epoch": 9.908296943231441,
+ "grad_norm": 0.017969170585274696,
+ "learning_rate": 0.0006,
+ "loss": 5.118013858795166,
+ "step": 713
+ },
+ {
+ "epoch": 9.922270742358078,
+ "grad_norm": 0.018021270632743835,
+ "learning_rate": 0.0006,
+ "loss": 5.034787178039551,
+ "step": 714
+ },
+ {
+ "epoch": 9.936244541484717,
+ "grad_norm": 0.017009710893034935,
+ "learning_rate": 0.0006,
+ "loss": 5.003920555114746,
+ "step": 715
+ },
+ {
+ "epoch": 9.950218340611354,
+ "grad_norm": 0.017722049728035927,
+ "learning_rate": 0.0006,
+ "loss": 5.031939506530762,
+ "step": 716
+ },
+ {
+ "epoch": 9.96419213973799,
+ "grad_norm": 0.017134087160229683,
+ "learning_rate": 0.0006,
+ "loss": 5.093474388122559,
+ "step": 717
+ },
+ {
+ "epoch": 9.97816593886463,
+ "grad_norm": 0.018511613830924034,
+ "learning_rate": 0.0006,
+ "loss": 5.078604698181152,
+ "step": 718
+ },
+ {
+ "epoch": 9.992139737991266,
+ "grad_norm": 0.021206334233283997,
+ "learning_rate": 0.0006,
+ "loss": 5.052572727203369,
+ "step": 719
+ },
+ {
+ "epoch": 10.0,
+ "grad_norm": 0.02239062264561653,
+ "learning_rate": 0.0006,
+ "loss": 5.063547134399414,
+ "step": 720
+ },
+ {
+ "epoch": 10.0,
+ "eval_loss": 5.173044681549072,
+ "eval_runtime": 56.7287,
+ "eval_samples_per_second": 43.047,
+ "eval_steps_per_second": 1.357,
+ "step": 720
+ },
+ {
+ "epoch": 10.013973799126637,
+ "grad_norm": 0.020757969468832016,
+ "learning_rate": 0.0006,
+ "loss": 5.045648574829102,
+ "step": 721
+ },
+ {
+ "epoch": 10.027947598253276,
+ "grad_norm": 0.022697702050209045,
+ "learning_rate": 0.0006,
+ "loss": 4.993833065032959,
+ "step": 722
+ },
+ {
+ "epoch": 10.041921397379912,
+ "grad_norm": 0.02434331737458706,
+ "learning_rate": 0.0006,
+ "loss": 5.073219299316406,
+ "step": 723
+ },
+ {
+ "epoch": 10.055895196506551,
+ "grad_norm": 0.025499660521745682,
+ "learning_rate": 0.0006,
+ "loss": 5.019072532653809,
+ "step": 724
+ },
+ {
+ "epoch": 10.069868995633188,
+ "grad_norm": 0.029281053692102432,
+ "learning_rate": 0.0006,
+ "loss": 4.959851264953613,
+ "step": 725
+ },
+ {
+ "epoch": 10.083842794759825,
+ "grad_norm": 0.033878110349178314,
+ "learning_rate": 0.0006,
+ "loss": 5.132717609405518,
+ "step": 726
+ },
+ {
+ "epoch": 10.097816593886463,
+ "grad_norm": 0.03385764732956886,
+ "learning_rate": 0.0006,
+ "loss": 5.0482892990112305,
+ "step": 727
+ },
+ {
+ "epoch": 10.1117903930131,
+ "grad_norm": 0.03466648980975151,
+ "learning_rate": 0.0006,
+ "loss": 4.946855545043945,
+ "step": 728
+ },
+ {
+ "epoch": 10.125764192139737,
+ "grad_norm": 0.028546305373311043,
+ "learning_rate": 0.0006,
+ "loss": 5.027774333953857,
+ "step": 729
+ },
+ {
+ "epoch": 10.139737991266376,
+ "grad_norm": 0.030575869604945183,
+ "learning_rate": 0.0006,
+ "loss": 4.9768476486206055,
+ "step": 730
+ },
+ {
+ "epoch": 10.153711790393013,
+ "grad_norm": 0.03614303469657898,
+ "learning_rate": 0.0006,
+ "loss": 5.078113555908203,
+ "step": 731
+ },
+ {
+ "epoch": 10.167685589519651,
+ "grad_norm": 0.040991660207509995,
+ "learning_rate": 0.0006,
+ "loss": 5.033185958862305,
+ "step": 732
+ },
+ {
+ "epoch": 10.181659388646288,
+ "grad_norm": 0.04551699757575989,
+ "learning_rate": 0.0006,
+ "loss": 5.0820746421813965,
+ "step": 733
+ },
+ {
+ "epoch": 10.195633187772925,
+ "grad_norm": 0.040193989872932434,
+ "learning_rate": 0.0006,
+ "loss": 5.059700965881348,
+ "step": 734
+ },
+ {
+ "epoch": 10.209606986899564,
+ "grad_norm": 0.035851605236530304,
+ "learning_rate": 0.0006,
+ "loss": 5.142467498779297,
+ "step": 735
+ },
+ {
+ "epoch": 10.2235807860262,
+ "grad_norm": 0.034200169146060944,
+ "learning_rate": 0.0006,
+ "loss": 5.072312355041504,
+ "step": 736
+ },
+ {
+ "epoch": 10.237554585152838,
+ "grad_norm": 0.033621110022068024,
+ "learning_rate": 0.0006,
+ "loss": 5.08037805557251,
+ "step": 737
+ },
+ {
+ "epoch": 10.251528384279476,
+ "grad_norm": 0.037372369319200516,
+ "learning_rate": 0.0006,
+ "loss": 4.998425483703613,
+ "step": 738
+ },
+ {
+ "epoch": 10.265502183406113,
+ "grad_norm": 0.037044707685709,
+ "learning_rate": 0.0006,
+ "loss": 4.923993110656738,
+ "step": 739
+ },
+ {
+ "epoch": 10.279475982532752,
+ "grad_norm": 0.030454808846116066,
+ "learning_rate": 0.0006,
+ "loss": 5.0889997482299805,
+ "step": 740
+ },
+ {
+ "epoch": 10.293449781659389,
+ "grad_norm": 0.030969638377428055,
+ "learning_rate": 0.0006,
+ "loss": 4.984112739562988,
+ "step": 741
+ },
+ {
+ "epoch": 10.307423580786025,
+ "grad_norm": 0.0276983380317688,
+ "learning_rate": 0.0006,
+ "loss": 4.955412864685059,
+ "step": 742
+ },
+ {
+ "epoch": 10.321397379912664,
+ "grad_norm": 0.02716052532196045,
+ "learning_rate": 0.0006,
+ "loss": 5.0098876953125,
+ "step": 743
+ },
+ {
+ "epoch": 10.335371179039301,
+ "grad_norm": 0.02403552085161209,
+ "learning_rate": 0.0006,
+ "loss": 5.165160655975342,
+ "step": 744
+ },
+ {
+ "epoch": 10.34934497816594,
+ "grad_norm": 0.0257862601429224,
+ "learning_rate": 0.0006,
+ "loss": 5.001547813415527,
+ "step": 745
+ },
+ {
+ "epoch": 10.363318777292577,
+ "grad_norm": 0.02300378680229187,
+ "learning_rate": 0.0006,
+ "loss": 5.038888931274414,
+ "step": 746
+ },
+ {
+ "epoch": 10.377292576419213,
+ "grad_norm": 0.019766854122281075,
+ "learning_rate": 0.0006,
+ "loss": 4.924291610717773,
+ "step": 747
+ },
+ {
+ "epoch": 10.391266375545852,
+ "grad_norm": 0.01888395845890045,
+ "learning_rate": 0.0006,
+ "loss": 5.059971332550049,
+ "step": 748
+ },
+ {
+ "epoch": 10.405240174672489,
+ "grad_norm": 0.015577499754726887,
+ "learning_rate": 0.0006,
+ "loss": 4.987686634063721,
+ "step": 749
+ },
+ {
+ "epoch": 10.419213973799126,
+ "grad_norm": 0.016012346372008324,
+ "learning_rate": 0.0006,
+ "loss": 5.015157699584961,
+ "step": 750
+ },
+ {
+ "epoch": 10.433187772925764,
+ "grad_norm": 0.016277998685836792,
+ "learning_rate": 0.0006,
+ "loss": 5.004931449890137,
+ "step": 751
+ },
+ {
+ "epoch": 10.447161572052401,
+ "grad_norm": 0.017504561692476273,
+ "learning_rate": 0.0006,
+ "loss": 4.902827739715576,
+ "step": 752
+ },
+ {
+ "epoch": 10.46113537117904,
+ "grad_norm": 0.01569022797048092,
+ "learning_rate": 0.0006,
+ "loss": 4.954863548278809,
+ "step": 753
+ },
+ {
+ "epoch": 10.475109170305677,
+ "grad_norm": 0.014114422723650932,
+ "learning_rate": 0.0006,
+ "loss": 5.00656795501709,
+ "step": 754
+ },
+ {
+ "epoch": 10.489082969432314,
+ "grad_norm": 0.01573832333087921,
+ "learning_rate": 0.0006,
+ "loss": 4.978353023529053,
+ "step": 755
+ },
+ {
+ "epoch": 10.503056768558952,
+ "grad_norm": 0.016480036079883575,
+ "learning_rate": 0.0006,
+ "loss": 4.953327655792236,
+ "step": 756
+ },
+ {
+ "epoch": 10.51703056768559,
+ "grad_norm": 0.014008025638759136,
+ "learning_rate": 0.0006,
+ "loss": 5.050840377807617,
+ "step": 757
+ },
+ {
+ "epoch": 10.531004366812226,
+ "grad_norm": 0.013147998601198196,
+ "learning_rate": 0.0006,
+ "loss": 4.974964141845703,
+ "step": 758
+ },
+ {
+ "epoch": 10.544978165938865,
+ "grad_norm": 0.013807603158056736,
+ "learning_rate": 0.0006,
+ "loss": 4.927907943725586,
+ "step": 759
+ },
+ {
+ "epoch": 10.558951965065502,
+ "grad_norm": 0.013555224984884262,
+ "learning_rate": 0.0006,
+ "loss": 4.980690956115723,
+ "step": 760
+ },
+ {
+ "epoch": 10.57292576419214,
+ "grad_norm": 0.013045408762991428,
+ "learning_rate": 0.0006,
+ "loss": 4.9436116218566895,
+ "step": 761
+ },
+ {
+ "epoch": 10.586899563318777,
+ "grad_norm": 0.012858862057328224,
+ "learning_rate": 0.0006,
+ "loss": 4.9595770835876465,
+ "step": 762
+ },
+ {
+ "epoch": 10.600873362445414,
+ "grad_norm": 0.01468253880739212,
+ "learning_rate": 0.0006,
+ "loss": 4.90725040435791,
+ "step": 763
+ },
+ {
+ "epoch": 10.614847161572053,
+ "grad_norm": 0.013224679045379162,
+ "learning_rate": 0.0006,
+ "loss": 4.877615928649902,
+ "step": 764
+ },
+ {
+ "epoch": 10.62882096069869,
+ "grad_norm": 0.013555348850786686,
+ "learning_rate": 0.0006,
+ "loss": 4.9456610679626465,
+ "step": 765
+ },
+ {
+ "epoch": 10.642794759825328,
+ "grad_norm": 0.014387160539627075,
+ "learning_rate": 0.0006,
+ "loss": 4.9233717918396,
+ "step": 766
+ },
+ {
+ "epoch": 10.656768558951965,
+ "grad_norm": 0.014354088343679905,
+ "learning_rate": 0.0006,
+ "loss": 4.983511924743652,
+ "step": 767
+ },
+ {
+ "epoch": 10.670742358078602,
+ "grad_norm": 0.01332700252532959,
+ "learning_rate": 0.0006,
+ "loss": 4.934416770935059,
+ "step": 768
+ },
+ {
+ "epoch": 10.68471615720524,
+ "grad_norm": 0.011543313041329384,
+ "learning_rate": 0.0006,
+ "loss": 4.96388053894043,
+ "step": 769
+ },
+ {
+ "epoch": 10.698689956331878,
+ "grad_norm": 0.012330878525972366,
+ "learning_rate": 0.0006,
+ "loss": 4.946396350860596,
+ "step": 770
+ },
+ {
+ "epoch": 10.712663755458514,
+ "grad_norm": 0.01379779726266861,
+ "learning_rate": 0.0006,
+ "loss": 4.981019020080566,
+ "step": 771
+ },
+ {
+ "epoch": 10.726637554585153,
+ "grad_norm": 0.014612431637942791,
+ "learning_rate": 0.0006,
+ "loss": 4.894567966461182,
+ "step": 772
+ },
+ {
+ "epoch": 10.74061135371179,
+ "grad_norm": 0.013062899932265282,
+ "learning_rate": 0.0006,
+ "loss": 5.033038139343262,
+ "step": 773
+ },
+ {
+ "epoch": 10.754585152838429,
+ "grad_norm": 0.013898049481213093,
+ "learning_rate": 0.0006,
+ "loss": 4.823636054992676,
+ "step": 774
+ },
+ {
+ "epoch": 10.768558951965066,
+ "grad_norm": 0.01502019353210926,
+ "learning_rate": 0.0006,
+ "loss": 5.020156383514404,
+ "step": 775
+ },
+ {
+ "epoch": 10.782532751091702,
+ "grad_norm": 0.013810204342007637,
+ "learning_rate": 0.0006,
+ "loss": 4.935647010803223,
+ "step": 776
+ },
+ {
+ "epoch": 10.796506550218341,
+ "grad_norm": 0.013616513460874557,
+ "learning_rate": 0.0006,
+ "loss": 4.901486396789551,
+ "step": 777
+ },
+ {
+ "epoch": 10.810480349344978,
+ "grad_norm": 0.013601024635136127,
+ "learning_rate": 0.0006,
+ "loss": 4.931873321533203,
+ "step": 778
+ },
+ {
+ "epoch": 10.824454148471617,
+ "grad_norm": 0.013668350875377655,
+ "learning_rate": 0.0006,
+ "loss": 4.937256813049316,
+ "step": 779
+ },
+ {
+ "epoch": 10.838427947598253,
+ "grad_norm": 0.01581823080778122,
+ "learning_rate": 0.0006,
+ "loss": 4.9611406326293945,
+ "step": 780
+ },
+ {
+ "epoch": 10.85240174672489,
+ "grad_norm": 0.019478484988212585,
+ "learning_rate": 0.0006,
+ "loss": 4.9212236404418945,
+ "step": 781
+ },
+ {
+ "epoch": 10.866375545851529,
+ "grad_norm": 0.026921333745121956,
+ "learning_rate": 0.0006,
+ "loss": 4.88501501083374,
+ "step": 782
+ },
+ {
+ "epoch": 10.880349344978166,
+ "grad_norm": 0.03155914694070816,
+ "learning_rate": 0.0006,
+ "loss": 4.989967346191406,
+ "step": 783
+ },
+ {
+ "epoch": 10.894323144104803,
+ "grad_norm": 0.029689429327845573,
+ "learning_rate": 0.0006,
+ "loss": 4.936000823974609,
+ "step": 784
+ },
+ {
+ "epoch": 10.908296943231441,
+ "grad_norm": 0.026626063510775566,
+ "learning_rate": 0.0006,
+ "loss": 4.984785079956055,
+ "step": 785
+ },
+ {
+ "epoch": 10.922270742358078,
+ "grad_norm": 0.026168212294578552,
+ "learning_rate": 0.0006,
+ "loss": 4.908282279968262,
+ "step": 786
+ },
+ {
+ "epoch": 10.936244541484717,
+ "grad_norm": 0.02828158251941204,
+ "learning_rate": 0.0006,
+ "loss": 4.926390647888184,
+ "step": 787
+ },
+ {
+ "epoch": 10.950218340611354,
+ "grad_norm": 0.02649078331887722,
+ "learning_rate": 0.0006,
+ "loss": 5.00400972366333,
+ "step": 788
+ },
+ {
+ "epoch": 10.96419213973799,
+ "grad_norm": 0.0256856270134449,
+ "learning_rate": 0.0006,
+ "loss": 5.0317487716674805,
+ "step": 789
+ },
+ {
+ "epoch": 10.97816593886463,
+ "grad_norm": 0.024728883057832718,
+ "learning_rate": 0.0006,
+ "loss": 4.9299635887146,
+ "step": 790
+ },
+ {
+ "epoch": 10.992139737991266,
+ "grad_norm": 0.026241116225719452,
+ "learning_rate": 0.0006,
+ "loss": 4.947319984436035,
+ "step": 791
+ },
+ {
+ "epoch": 11.0,
+ "grad_norm": 0.023526743054389954,
+ "learning_rate": 0.0006,
+ "loss": 4.94429349899292,
+ "step": 792
+ },
+ {
+ "epoch": 11.0,
+ "eval_loss": 5.104345798492432,
+ "eval_runtime": 57.0955,
+ "eval_samples_per_second": 42.77,
+ "eval_steps_per_second": 1.349,
+ "step": 792
+ },
+ {
+ "epoch": 11.013973799126637,
+ "grad_norm": 0.0216788612306118,
+ "learning_rate": 0.0006,
+ "loss": 4.881950855255127,
+ "step": 793
+ },
+ {
+ "epoch": 11.027947598253276,
+ "grad_norm": 0.01830463856458664,
+ "learning_rate": 0.0006,
+ "loss": 4.971016883850098,
+ "step": 794
+ },
+ {
+ "epoch": 11.041921397379912,
+ "grad_norm": 0.02204521745443344,
+ "learning_rate": 0.0006,
+ "loss": 4.894623756408691,
+ "step": 795
+ },
+ {
+ "epoch": 11.055895196506551,
+ "grad_norm": 0.0210255216807127,
+ "learning_rate": 0.0006,
+ "loss": 4.834009170532227,
+ "step": 796
+ },
+ {
+ "epoch": 11.069868995633188,
+ "grad_norm": 0.019797448068857193,
+ "learning_rate": 0.0006,
+ "loss": 5.086714744567871,
+ "step": 797
+ },
+ {
+ "epoch": 11.083842794759825,
+ "grad_norm": 0.02036641724407673,
+ "learning_rate": 0.0006,
+ "loss": 4.920056343078613,
+ "step": 798
+ },
+ {
+ "epoch": 11.097816593886463,
+ "grad_norm": 0.01966066285967827,
+ "learning_rate": 0.0006,
+ "loss": 4.879859447479248,
+ "step": 799
+ },
+ {
+ "epoch": 11.1117903930131,
+ "grad_norm": 0.0202149897813797,
+ "learning_rate": 0.0006,
+ "loss": 4.933640480041504,
+ "step": 800
+ },
+ {
+ "epoch": 11.125764192139737,
+ "grad_norm": 0.01893039606511593,
+ "learning_rate": 0.0006,
+ "loss": 4.901970863342285,
+ "step": 801
+ },
+ {
+ "epoch": 11.139737991266376,
+ "grad_norm": 0.016040155664086342,
+ "learning_rate": 0.0006,
+ "loss": 4.905491828918457,
+ "step": 802
+ },
+ {
+ "epoch": 11.153711790393013,
+ "grad_norm": 0.01720350608229637,
+ "learning_rate": 0.0006,
+ "loss": 4.798816680908203,
+ "step": 803
+ },
+ {
+ "epoch": 11.167685589519651,
+ "grad_norm": 0.017002852633595467,
+ "learning_rate": 0.0006,
+ "loss": 4.851859092712402,
+ "step": 804
+ },
+ {
+ "epoch": 11.181659388646288,
+ "grad_norm": 0.015594680793583393,
+ "learning_rate": 0.0006,
+ "loss": 4.90328311920166,
+ "step": 805
+ },
+ {
+ "epoch": 11.195633187772925,
+ "grad_norm": 0.014904462732374668,
+ "learning_rate": 0.0006,
+ "loss": 4.904215335845947,
+ "step": 806
+ },
+ {
+ "epoch": 11.209606986899564,
+ "grad_norm": 0.01366228237748146,
+ "learning_rate": 0.0006,
+ "loss": 4.9813055992126465,
+ "step": 807
+ },
+ {
+ "epoch": 11.2235807860262,
+ "grad_norm": 0.014719638973474503,
+ "learning_rate": 0.0006,
+ "loss": 4.927306175231934,
+ "step": 808
+ },
+ {
+ "epoch": 11.237554585152838,
+ "grad_norm": 0.015016036108136177,
+ "learning_rate": 0.0006,
+ "loss": 4.874215126037598,
+ "step": 809
+ },
+ {
+ "epoch": 11.251528384279476,
+ "grad_norm": 0.014925424009561539,
+ "learning_rate": 0.0006,
+ "loss": 4.838257789611816,
+ "step": 810
+ },
+ {
+ "epoch": 11.265502183406113,
+ "grad_norm": 0.014335823245346546,
+ "learning_rate": 0.0006,
+ "loss": 4.940608024597168,
+ "step": 811
+ },
+ {
+ "epoch": 11.279475982532752,
+ "grad_norm": 0.014987512491643429,
+ "learning_rate": 0.0006,
+ "loss": 4.831531524658203,
+ "step": 812
+ },
+ {
+ "epoch": 11.293449781659389,
+ "grad_norm": 0.016087457537651062,
+ "learning_rate": 0.0006,
+ "loss": 4.863227367401123,
+ "step": 813
+ },
+ {
+ "epoch": 11.307423580786025,
+ "grad_norm": 0.01671300269663334,
+ "learning_rate": 0.0006,
+ "loss": 4.8535661697387695,
+ "step": 814
+ },
+ {
+ "epoch": 11.321397379912664,
+ "grad_norm": 0.015229142270982265,
+ "learning_rate": 0.0006,
+ "loss": 4.948635101318359,
+ "step": 815
+ },
+ {
+ "epoch": 11.335371179039301,
+ "grad_norm": 0.01532587967813015,
+ "learning_rate": 0.0006,
+ "loss": 4.9035797119140625,
+ "step": 816
+ },
+ {
+ "epoch": 11.34934497816594,
+ "grad_norm": 0.015715476125478745,
+ "learning_rate": 0.0006,
+ "loss": 4.894709587097168,
+ "step": 817
+ },
+ {
+ "epoch": 11.363318777292577,
+ "grad_norm": 0.014156977646052837,
+ "learning_rate": 0.0006,
+ "loss": 4.960206985473633,
+ "step": 818
+ },
+ {
+ "epoch": 11.377292576419213,
+ "grad_norm": 0.017076566815376282,
+ "learning_rate": 0.0006,
+ "loss": 5.005850791931152,
+ "step": 819
+ },
+ {
+ "epoch": 11.391266375545852,
+ "grad_norm": 0.0172084029763937,
+ "learning_rate": 0.0006,
+ "loss": 4.903355598449707,
+ "step": 820
+ },
+ {
+ "epoch": 11.405240174672489,
+ "grad_norm": 0.0180258359760046,
+ "learning_rate": 0.0006,
+ "loss": 4.842667579650879,
+ "step": 821
+ },
+ {
+ "epoch": 11.419213973799126,
+ "grad_norm": 0.022130120545625687,
+ "learning_rate": 0.0006,
+ "loss": 4.799970626831055,
+ "step": 822
+ },
+ {
+ "epoch": 11.433187772925764,
+ "grad_norm": 0.02384897693991661,
+ "learning_rate": 0.0006,
+ "loss": 4.860746383666992,
+ "step": 823
+ },
+ {
+ "epoch": 11.447161572052401,
+ "grad_norm": 0.021881572902202606,
+ "learning_rate": 0.0006,
+ "loss": 4.918107032775879,
+ "step": 824
+ },
+ {
+ "epoch": 11.46113537117904,
+ "grad_norm": 0.023085080087184906,
+ "learning_rate": 0.0006,
+ "loss": 4.878946304321289,
+ "step": 825
+ },
+ {
+ "epoch": 11.475109170305677,
+ "grad_norm": 0.02291012555360794,
+ "learning_rate": 0.0006,
+ "loss": 4.958887100219727,
+ "step": 826
+ },
+ {
+ "epoch": 11.489082969432314,
+ "grad_norm": 0.0199968870729208,
+ "learning_rate": 0.0006,
+ "loss": 4.80955696105957,
+ "step": 827
+ },
+ {
+ "epoch": 11.503056768558952,
+ "grad_norm": 0.018702229484915733,
+ "learning_rate": 0.0006,
+ "loss": 5.0276408195495605,
+ "step": 828
+ },
+ {
+ "epoch": 11.51703056768559,
+ "grad_norm": 0.017787311226129532,
+ "learning_rate": 0.0006,
+ "loss": 4.96018123626709,
+ "step": 829
+ },
+ {
+ "epoch": 11.531004366812226,
+ "grad_norm": 0.01704501546919346,
+ "learning_rate": 0.0006,
+ "loss": 4.914553165435791,
+ "step": 830
+ },
+ {
+ "epoch": 11.544978165938865,
+ "grad_norm": 0.017519677057862282,
+ "learning_rate": 0.0006,
+ "loss": 4.798098087310791,
+ "step": 831
+ },
+ {
+ "epoch": 11.558951965065502,
+ "grad_norm": 0.018511991947889328,
+ "learning_rate": 0.0006,
+ "loss": 4.817785263061523,
+ "step": 832
+ },
+ {
+ "epoch": 11.57292576419214,
+ "grad_norm": 0.01876644790172577,
+ "learning_rate": 0.0006,
+ "loss": 4.818995475769043,
+ "step": 833
+ },
+ {
+ "epoch": 11.586899563318777,
+ "grad_norm": 0.01753568835556507,
+ "learning_rate": 0.0006,
+ "loss": 4.811938285827637,
+ "step": 834
+ },
+ {
+ "epoch": 11.600873362445414,
+ "grad_norm": 0.017046531662344933,
+ "learning_rate": 0.0006,
+ "loss": 4.914674282073975,
+ "step": 835
+ },
+ {
+ "epoch": 11.614847161572053,
+ "grad_norm": 0.01802094094455242,
+ "learning_rate": 0.0006,
+ "loss": 4.993284225463867,
+ "step": 836
+ },
+ {
+ "epoch": 11.62882096069869,
+ "grad_norm": 0.019493764266371727,
+ "learning_rate": 0.0006,
+ "loss": 4.895628452301025,
+ "step": 837
+ },
+ {
+ "epoch": 11.642794759825328,
+ "grad_norm": 0.019461210817098618,
+ "learning_rate": 0.0006,
+ "loss": 4.869269371032715,
+ "step": 838
+ },
+ {
+ "epoch": 11.656768558951965,
+ "grad_norm": 0.01981567218899727,
+ "learning_rate": 0.0006,
+ "loss": 4.863052845001221,
+ "step": 839
+ },
+ {
+ "epoch": 11.670742358078602,
+ "grad_norm": 0.019894491881132126,
+ "learning_rate": 0.0006,
+ "loss": 4.8775787353515625,
+ "step": 840
+ },
+ {
+ "epoch": 11.68471615720524,
+ "grad_norm": 0.01875786855816841,
+ "learning_rate": 0.0006,
+ "loss": 4.817763328552246,
+ "step": 841
+ },
+ {
+ "epoch": 11.698689956331878,
+ "grad_norm": 0.020926695317029953,
+ "learning_rate": 0.0006,
+ "loss": 4.7470526695251465,
+ "step": 842
+ },
+ {
+ "epoch": 11.712663755458514,
+ "grad_norm": 0.021477103233337402,
+ "learning_rate": 0.0006,
+ "loss": 4.762682914733887,
+ "step": 843
+ },
+ {
+ "epoch": 11.726637554585153,
+ "grad_norm": 0.020628711208701134,
+ "learning_rate": 0.0006,
+ "loss": 4.807985782623291,
+ "step": 844
+ },
+ {
+ "epoch": 11.74061135371179,
+ "grad_norm": 0.021489612758159637,
+ "learning_rate": 0.0006,
+ "loss": 4.843091011047363,
+ "step": 845
+ },
+ {
+ "epoch": 11.754585152838429,
+ "grad_norm": 0.021194491535425186,
+ "learning_rate": 0.0006,
+ "loss": 4.8300251960754395,
+ "step": 846
+ },
+ {
+ "epoch": 11.768558951965066,
+ "grad_norm": 0.01833350397646427,
+ "learning_rate": 0.0006,
+ "loss": 5.000202178955078,
+ "step": 847
+ },
+ {
+ "epoch": 11.782532751091702,
+ "grad_norm": 0.018096931278705597,
+ "learning_rate": 0.0006,
+ "loss": 4.880648136138916,
+ "step": 848
+ },
+ {
+ "epoch": 11.796506550218341,
+ "grad_norm": 0.0197161752730608,
+ "learning_rate": 0.0006,
+ "loss": 4.788316249847412,
+ "step": 849
+ },
+ {
+ "epoch": 11.810480349344978,
+ "grad_norm": 0.0203181654214859,
+ "learning_rate": 0.0006,
+ "loss": 4.870944976806641,
+ "step": 850
+ },
+ {
+ "epoch": 11.824454148471617,
+ "grad_norm": 0.01908455230295658,
+ "learning_rate": 0.0006,
+ "loss": 4.890144348144531,
+ "step": 851
+ },
+ {
+ "epoch": 11.838427947598253,
+ "grad_norm": 0.021281801164150238,
+ "learning_rate": 0.0006,
+ "loss": 4.89578914642334,
+ "step": 852
+ },
+ {
+ "epoch": 11.85240174672489,
+ "grad_norm": 0.02124273031949997,
+ "learning_rate": 0.0006,
+ "loss": 4.883749008178711,
+ "step": 853
+ },
+ {
+ "epoch": 11.866375545851529,
+ "grad_norm": 0.020969398319721222,
+ "learning_rate": 0.0006,
+ "loss": 4.809142589569092,
+ "step": 854
+ },
+ {
+ "epoch": 11.880349344978166,
+ "grad_norm": 0.01967657171189785,
+ "learning_rate": 0.0006,
+ "loss": 4.821864604949951,
+ "step": 855
+ },
+ {
+ "epoch": 11.894323144104803,
+ "grad_norm": 0.020887333899736404,
+ "learning_rate": 0.0006,
+ "loss": 4.917985916137695,
+ "step": 856
+ },
+ {
+ "epoch": 11.908296943231441,
+ "grad_norm": 0.019203083589673042,
+ "learning_rate": 0.0006,
+ "loss": 4.824960708618164,
+ "step": 857
+ },
+ {
+ "epoch": 11.922270742358078,
+ "grad_norm": 0.020162401720881462,
+ "learning_rate": 0.0006,
+ "loss": 4.772617816925049,
+ "step": 858
+ },
+ {
+ "epoch": 11.936244541484717,
+ "grad_norm": 0.02251487225294113,
+ "learning_rate": 0.0006,
+ "loss": 4.735418319702148,
+ "step": 859
+ },
+ {
+ "epoch": 11.950218340611354,
+ "grad_norm": 0.020313527435064316,
+ "learning_rate": 0.0006,
+ "loss": 4.939233779907227,
+ "step": 860
+ },
+ {
+ "epoch": 11.96419213973799,
+ "grad_norm": 0.021635599434375763,
+ "learning_rate": 0.0006,
+ "loss": 4.861807346343994,
+ "step": 861
+ },
+ {
+ "epoch": 11.97816593886463,
+ "grad_norm": 0.02285311557352543,
+ "learning_rate": 0.0006,
+ "loss": 4.932732582092285,
+ "step": 862
+ },
+ {
+ "epoch": 11.992139737991266,
+ "grad_norm": 0.02072535641491413,
+ "learning_rate": 0.0006,
+ "loss": 4.796767711639404,
+ "step": 863
+ },
+ {
+ "epoch": 12.0,
+ "grad_norm": 0.020588304847478867,
+ "learning_rate": 0.0006,
+ "loss": 4.997537612915039,
+ "step": 864
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 28800,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 400,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": true,
+ "should_log": false,
+ "should_save": false,
+ "should_training_stop": false
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 3.683536033598669e+17,
+ "train_batch_size": 8,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-864/training_args.bin b/runs/l2r50-i2-fulle-lm/checkpoint-864/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..cad232831c222cc7f8b938b8bbe42f61e7d89c3a
--- /dev/null
+++ b/runs/l2r50-i2-fulle-lm/checkpoint-864/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:f7c40e3bfa2f5ac1450ca1243e3f2a895d44a06ed5995e6aefd97f78d229d8c8
+size 4792
diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1152/chat_template.jinja b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1152/chat_template.jinja
new file mode 100644
index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230
--- /dev/null
+++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1152/chat_template.jinja
@@ -0,0 +1,85 @@
+{%- if tools %}
+ {{- '<|im_start|>system\n' }}
+ {%- if messages[0].role == 'system' %}
+ {{- messages[0].content + '\n\n' }}
+ {%- endif %}
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }}
+ {%- for tool in tools %}
+ {{- "\n" }}
+ {{- tool | tojson }}
+ {%- endfor %}
+ {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }}
+{%- else %}
+ {%- if messages[0].role == 'system' %}
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
+ {%- endif %}
+{%- endif %}
+{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
+{%- for message in messages[::-1] %}
+ {%- set index = (messages|length - 1) - loop.index0 %}
+ {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %}
+ {%- set ns.multi_step_tool = false %}
+ {%- set ns.last_query_index = index %}
+ {%- endif %}
+{%- endfor %}
+{%- for message in messages %}
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
+ {%- elif message.role == "assistant" %}
+ {%- set content = message.content %}
+ {%- set reasoning_content = '' %}
+ {%- if message.reasoning_content is defined and message.reasoning_content is not none %}
+ {%- set reasoning_content = message.reasoning_content %}
+ {%- else %}
+ {%- if '' in message.content %}
+ {%- set content = message.content.split('')[-1].lstrip('\n') %}
+ {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %}
+ {%- endif %}
+ {%- endif %}
+ {%- if loop.index0 > ns.last_query_index %}
+ {%- if loop.last or (not loop.last and reasoning_content) %}
+ {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- if message.tool_calls %}
+ {%- for tool_call in message.tool_calls %}
+ {%- if (loop.first and content) or (not loop.first) %}
+ {{- '\n' }}
+ {%- endif %}
+ {%- if tool_call.function %}
+ {%- set tool_call = tool_call.function %}
+ {%- endif %}
+ {{- '\n{"name": "' }}
+ {{- tool_call.name }}
+ {{- '", "arguments": ' }}
+ {%- if tool_call.arguments is string %}
+ {{- tool_call.arguments }}
+ {%- else %}
+ {{- tool_call.arguments | tojson }}
+ {%- endif %}
+ {{- '}\n' }}
+ {%- endfor %}
+ {%- endif %}
+ {{- '<|im_end|>\n' }}
+ {%- elif message.role == "tool" %}
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
+ {{- '<|im_start|>user' }}
+ {%- endif %}
+ {{- '\n\n' }}
+ {{- message.content }}
+ {{- '\n' }}
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
+ {{- '<|im_end|>\n' }}
+ {%- endif %}
+ {%- endif %}
+{%- endfor %}
+{%- if add_generation_prompt %}
+ {{- '<|im_start|>assistant\n' }}
+ {%- if enable_thinking is defined and enable_thinking is false %}
+ {{- '\n\n\n\n' }}
+ {%- endif %}
+{%- endif %}
\ No newline at end of file
diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1152/config.json b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1152/config.json
new file mode 100644
index 0000000000000000000000000000000000000000..175ee3431793018f23f74824f69554b21afcbd4f
--- /dev/null
+++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1152/config.json
@@ -0,0 +1,32 @@
+{
+ "architectures": [
+ "LlamaForCausalLM"
+ ],
+ "attention_bias": false,
+ "attention_dropout": 0.0,
+ "bos_token_id": null,
+ "dtype": "float32",
+ "eos_token_id": 151645,
+ "head_dim": 128,
+ "hidden_act": "silu",
+ "hidden_size": 512,
+ "initializer_range": 0.02,
+ "intermediate_size": 1536,
+ "max_position_embeddings": 2048,
+ "mlp_bias": false,
+ "model_type": "llama",
+ "num_attention_heads": 4,
+ "num_hidden_layers": 20,
+ "num_key_value_heads": 4,
+ "pad_token_id": 151645,
+ "pretraining_tp": 1,
+ "rms_norm_eps": 1e-06,
+ "rope_parameters": {
+ "rope_theta": 10000.0,
+ "rope_type": "default"
+ },
+ "tie_word_embeddings": true,
+ "transformers_version": "5.5.0",
+ "use_cache": false,
+ "vocab_size": 151674
+}
diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1152/generation_config.json b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1152/generation_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3
--- /dev/null
+++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1152/generation_config.json
@@ -0,0 +1,11 @@
+{
+ "_from_model_config": true,
+ "eos_token_id": [
+ 151645
+ ],
+ "output_attentions": false,
+ "output_hidden_states": false,
+ "pad_token_id": 151645,
+ "transformers_version": "5.5.0",
+ "use_cache": true
+}
diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1152/model.safetensors b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1152/model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..26beed4f25e71b90d2eaf6a618a168a86d6b91f0
--- /dev/null
+++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1152/model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:ffe1e08b4029f005c849c88b50a316524a347d2e07c34a943a74a14cad3689ae
+size 583362376
diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1152/optimizer.pt b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1152/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..040ccbd15db63769ecb441ec45ced302b25f3ff9
--- /dev/null
+++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1152/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:c4af92d4f36cd57db42725af078d57ed2acd8f24e0b3cffe329864d680577c40
+size 1166837626
diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1152/rng_state_0.pth b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1152/rng_state_0.pth
new file mode 100644
index 0000000000000000000000000000000000000000..e2c723a27d937a5234fdd83c600d5494d2b4920c
--- /dev/null
+++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1152/rng_state_0.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:ce1873f29596200b93f59d5a85747bfece46572d1a0840b7d8fe0df166b999af
+size 15024
diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1152/rng_state_1.pth b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1152/rng_state_1.pth
new file mode 100644
index 0000000000000000000000000000000000000000..625005d6c88f4c74f804d5cccc7d953809f2e5f3
--- /dev/null
+++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1152/rng_state_1.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:692b8a3b6c6eb68decf07e62e78d5cb6d2f573938e34d6ffe16394958ddf9d90
+size 15024
diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1152/rng_state_2.pth b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1152/rng_state_2.pth
new file mode 100644
index 0000000000000000000000000000000000000000..0cbbf074a1f543e1364e32d0b7c6822f36d1bb3c
--- /dev/null
+++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1152/rng_state_2.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:35e67c56d39aed6d994eecdadecbd71140c0e5d738c1d60c5f25335076198259
+size 15024
diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1152/rng_state_3.pth b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1152/rng_state_3.pth
new file mode 100644
index 0000000000000000000000000000000000000000..b5195c5c2b47ccc1182a69bb7e6216b99388af1f
--- /dev/null
+++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1152/rng_state_3.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:340b9fc530c021cafb0ea29114f89f46734f8bbee4420f160a23216252c1f127
+size 15024
diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1152/scheduler.pt b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1152/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..630fbc73d70608487940c0c34e65caee42415237
--- /dev/null
+++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1152/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:6a67914066d1aebda73626cb9d7d4cbc46ee7df332e5ae85a86cc87be8e519c8
+size 1064
diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1152/tokenizer.json b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1152/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..9ea28d9eb325c36bcab4c1a08fc93fd598929f08
--- /dev/null
+++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1152/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:60e99075d371ecbb9f1dde97363f612d45bf0bf88f1a2006d9bfcf1939333c14
+size 11424648
diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1152/tokenizer_config.json b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1152/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..cadd997fc26040d23ad6677fcb39d8d18b01f55e
--- /dev/null
+++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1152/tokenizer_config.json
@@ -0,0 +1,21 @@
+{
+ "add_prefix_space": false,
+ "backend": "tokenizers",
+ "bos_token": null,
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "<|im_end|>",
+ "errors": "replace",
+ "extra_special_tokens": [
+ "<|l2r_pred|>",
+ "<|r2l_pred|>",
+ "<|next_1_pred|>",
+ "<|next_2_pred|>",
+ "<|mask|>"
+ ],
+ "is_local": false,
+ "model_max_length": 131072,
+ "pad_token": "<|im_end|>",
+ "split_special_tokens": false,
+ "tokenizer_class": "Qwen2Tokenizer",
+ "unk_token": null
+}
diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1152/trainer_state.json b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1152/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..ea35b40b0f3087e742fa7ab8b4bb7ecb3fbe4950
--- /dev/null
+++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1152/trainer_state.json
@@ -0,0 +1,8218 @@
+{
+ "best_global_step": null,
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 16.0,
+ "eval_steps": 500,
+ "global_step": 1152,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.013973799126637555,
+ "grad_norm": 0.5768774151802063,
+ "learning_rate": 0.0,
+ "loss": 12.050328254699707,
+ "step": 1
+ },
+ {
+ "epoch": 0.02794759825327511,
+ "grad_norm": 0.5784570574760437,
+ "learning_rate": 5.999999999999999e-06,
+ "loss": 12.049251556396484,
+ "step": 2
+ },
+ {
+ "epoch": 0.04192139737991266,
+ "grad_norm": 0.5607187151908875,
+ "learning_rate": 1.1999999999999999e-05,
+ "loss": 11.921792030334473,
+ "step": 3
+ },
+ {
+ "epoch": 0.05589519650655022,
+ "grad_norm": 0.46466919779777527,
+ "learning_rate": 1.7999999999999997e-05,
+ "loss": 11.686628341674805,
+ "step": 4
+ },
+ {
+ "epoch": 0.06986899563318777,
+ "grad_norm": 0.38012945652008057,
+ "learning_rate": 2.3999999999999997e-05,
+ "loss": 11.432823181152344,
+ "step": 5
+ },
+ {
+ "epoch": 0.08384279475982533,
+ "grad_norm": 0.33215776085853577,
+ "learning_rate": 2.9999999999999997e-05,
+ "loss": 11.207542419433594,
+ "step": 6
+ },
+ {
+ "epoch": 0.09781659388646288,
+ "grad_norm": 0.2928776741027832,
+ "learning_rate": 3.5999999999999994e-05,
+ "loss": 11.000097274780273,
+ "step": 7
+ },
+ {
+ "epoch": 0.11179039301310044,
+ "grad_norm": 0.26018962264060974,
+ "learning_rate": 4.2e-05,
+ "loss": 10.825748443603516,
+ "step": 8
+ },
+ {
+ "epoch": 0.125764192139738,
+ "grad_norm": 0.23717083036899567,
+ "learning_rate": 4.7999999999999994e-05,
+ "loss": 10.677122116088867,
+ "step": 9
+ },
+ {
+ "epoch": 0.13973799126637554,
+ "grad_norm": 0.21641355752944946,
+ "learning_rate": 5.399999999999999e-05,
+ "loss": 10.569148063659668,
+ "step": 10
+ },
+ {
+ "epoch": 0.1537117903930131,
+ "grad_norm": 0.1982956826686859,
+ "learning_rate": 5.9999999999999995e-05,
+ "loss": 10.49653434753418,
+ "step": 11
+ },
+ {
+ "epoch": 0.16768558951965065,
+ "grad_norm": 0.18819282948970795,
+ "learning_rate": 6.599999999999999e-05,
+ "loss": 10.432951927185059,
+ "step": 12
+ },
+ {
+ "epoch": 0.18165938864628822,
+ "grad_norm": 0.18322618305683136,
+ "learning_rate": 7.199999999999999e-05,
+ "loss": 10.387115478515625,
+ "step": 13
+ },
+ {
+ "epoch": 0.19563318777292577,
+ "grad_norm": 0.18399226665496826,
+ "learning_rate": 7.8e-05,
+ "loss": 10.33381462097168,
+ "step": 14
+ },
+ {
+ "epoch": 0.2096069868995633,
+ "grad_norm": 0.1833190619945526,
+ "learning_rate": 8.4e-05,
+ "loss": 10.301689147949219,
+ "step": 15
+ },
+ {
+ "epoch": 0.22358078602620088,
+ "grad_norm": 0.18490834534168243,
+ "learning_rate": 8.999999999999999e-05,
+ "loss": 10.240968704223633,
+ "step": 16
+ },
+ {
+ "epoch": 0.23755458515283842,
+ "grad_norm": 0.18358023464679718,
+ "learning_rate": 9.599999999999999e-05,
+ "loss": 10.194717407226562,
+ "step": 17
+ },
+ {
+ "epoch": 0.251528384279476,
+ "grad_norm": 0.1801682710647583,
+ "learning_rate": 0.000102,
+ "loss": 10.14897632598877,
+ "step": 18
+ },
+ {
+ "epoch": 0.26550218340611353,
+ "grad_norm": 0.17899440228939056,
+ "learning_rate": 0.00010799999999999998,
+ "loss": 10.078763961791992,
+ "step": 19
+ },
+ {
+ "epoch": 0.2794759825327511,
+ "grad_norm": 0.17666497826576233,
+ "learning_rate": 0.00011399999999999999,
+ "loss": 10.01095962524414,
+ "step": 20
+ },
+ {
+ "epoch": 0.2934497816593886,
+ "grad_norm": 0.17196154594421387,
+ "learning_rate": 0.00011999999999999999,
+ "loss": 9.951203346252441,
+ "step": 21
+ },
+ {
+ "epoch": 0.3074235807860262,
+ "grad_norm": 0.169203519821167,
+ "learning_rate": 0.00012599999999999997,
+ "loss": 9.871726036071777,
+ "step": 22
+ },
+ {
+ "epoch": 0.32139737991266376,
+ "grad_norm": 0.16500306129455566,
+ "learning_rate": 0.00013199999999999998,
+ "loss": 9.803646087646484,
+ "step": 23
+ },
+ {
+ "epoch": 0.3353711790393013,
+ "grad_norm": 0.16128124296665192,
+ "learning_rate": 0.000138,
+ "loss": 9.728487014770508,
+ "step": 24
+ },
+ {
+ "epoch": 0.34934497816593885,
+ "grad_norm": 0.15904058516025543,
+ "learning_rate": 0.00014399999999999998,
+ "loss": 9.642473220825195,
+ "step": 25
+ },
+ {
+ "epoch": 0.36331877729257644,
+ "grad_norm": 0.15560561418533325,
+ "learning_rate": 0.00015,
+ "loss": 9.56414794921875,
+ "step": 26
+ },
+ {
+ "epoch": 0.377292576419214,
+ "grad_norm": 0.15278612077236176,
+ "learning_rate": 0.000156,
+ "loss": 9.484230041503906,
+ "step": 27
+ },
+ {
+ "epoch": 0.39126637554585153,
+ "grad_norm": 0.15263237059116364,
+ "learning_rate": 0.000162,
+ "loss": 9.383644104003906,
+ "step": 28
+ },
+ {
+ "epoch": 0.4052401746724891,
+ "grad_norm": 0.14945024251937866,
+ "learning_rate": 0.000168,
+ "loss": 9.29840087890625,
+ "step": 29
+ },
+ {
+ "epoch": 0.4192139737991266,
+ "grad_norm": 0.14614646136760712,
+ "learning_rate": 0.00017399999999999997,
+ "loss": 9.208931922912598,
+ "step": 30
+ },
+ {
+ "epoch": 0.4331877729257642,
+ "grad_norm": 0.14295098185539246,
+ "learning_rate": 0.00017999999999999998,
+ "loss": 9.122323989868164,
+ "step": 31
+ },
+ {
+ "epoch": 0.44716157205240176,
+ "grad_norm": 0.13868345320224762,
+ "learning_rate": 0.000186,
+ "loss": 9.03034496307373,
+ "step": 32
+ },
+ {
+ "epoch": 0.4611353711790393,
+ "grad_norm": 0.135064035654068,
+ "learning_rate": 0.00019199999999999998,
+ "loss": 8.92818832397461,
+ "step": 33
+ },
+ {
+ "epoch": 0.47510917030567684,
+ "grad_norm": 0.13097241520881653,
+ "learning_rate": 0.000198,
+ "loss": 8.834627151489258,
+ "step": 34
+ },
+ {
+ "epoch": 0.4890829694323144,
+ "grad_norm": 0.12478776276111603,
+ "learning_rate": 0.000204,
+ "loss": 8.755172729492188,
+ "step": 35
+ },
+ {
+ "epoch": 0.503056768558952,
+ "grad_norm": 0.1214982345700264,
+ "learning_rate": 0.00020999999999999998,
+ "loss": 8.647891998291016,
+ "step": 36
+ },
+ {
+ "epoch": 0.5170305676855895,
+ "grad_norm": 0.11958328634500504,
+ "learning_rate": 0.00021599999999999996,
+ "loss": 8.543274879455566,
+ "step": 37
+ },
+ {
+ "epoch": 0.5310043668122271,
+ "grad_norm": 0.11680915951728821,
+ "learning_rate": 0.00022199999999999998,
+ "loss": 8.44540023803711,
+ "step": 38
+ },
+ {
+ "epoch": 0.5449781659388646,
+ "grad_norm": 0.11491474509239197,
+ "learning_rate": 0.00022799999999999999,
+ "loss": 8.349921226501465,
+ "step": 39
+ },
+ {
+ "epoch": 0.5589519650655022,
+ "grad_norm": 0.11160755902528763,
+ "learning_rate": 0.000234,
+ "loss": 8.26960563659668,
+ "step": 40
+ },
+ {
+ "epoch": 0.5729257641921397,
+ "grad_norm": 0.10840637236833572,
+ "learning_rate": 0.00023999999999999998,
+ "loss": 8.192136764526367,
+ "step": 41
+ },
+ {
+ "epoch": 0.5868995633187772,
+ "grad_norm": 0.10585298389196396,
+ "learning_rate": 0.00024599999999999996,
+ "loss": 8.09546947479248,
+ "step": 42
+ },
+ {
+ "epoch": 0.6008733624454149,
+ "grad_norm": 0.1032542735338211,
+ "learning_rate": 0.00025199999999999995,
+ "loss": 7.991090774536133,
+ "step": 43
+ },
+ {
+ "epoch": 0.6148471615720524,
+ "grad_norm": 0.0986943170428276,
+ "learning_rate": 0.000258,
+ "loss": 7.92384147644043,
+ "step": 44
+ },
+ {
+ "epoch": 0.62882096069869,
+ "grad_norm": 0.09320762753486633,
+ "learning_rate": 0.00026399999999999997,
+ "loss": 7.8478193283081055,
+ "step": 45
+ },
+ {
+ "epoch": 0.6427947598253275,
+ "grad_norm": 0.09168332070112228,
+ "learning_rate": 0.00027,
+ "loss": 7.76185417175293,
+ "step": 46
+ },
+ {
+ "epoch": 0.6567685589519651,
+ "grad_norm": 0.08832395076751709,
+ "learning_rate": 0.000276,
+ "loss": 7.679399490356445,
+ "step": 47
+ },
+ {
+ "epoch": 0.6707423580786026,
+ "grad_norm": 0.08816400915384293,
+ "learning_rate": 0.00028199999999999997,
+ "loss": 7.5789642333984375,
+ "step": 48
+ },
+ {
+ "epoch": 0.6847161572052402,
+ "grad_norm": 0.0861717164516449,
+ "learning_rate": 0.00028799999999999995,
+ "loss": 7.519214630126953,
+ "step": 49
+ },
+ {
+ "epoch": 0.6986899563318777,
+ "grad_norm": 0.08103746175765991,
+ "learning_rate": 0.000294,
+ "loss": 7.4475321769714355,
+ "step": 50
+ },
+ {
+ "epoch": 0.7126637554585152,
+ "grad_norm": 0.07545104622840881,
+ "learning_rate": 0.0003,
+ "loss": 7.401182174682617,
+ "step": 51
+ },
+ {
+ "epoch": 0.7266375545851529,
+ "grad_norm": 0.07042575627565384,
+ "learning_rate": 0.00030599999999999996,
+ "loss": 7.340173244476318,
+ "step": 52
+ },
+ {
+ "epoch": 0.7406113537117904,
+ "grad_norm": 0.06737572699785233,
+ "learning_rate": 0.000312,
+ "loss": 7.265697002410889,
+ "step": 53
+ },
+ {
+ "epoch": 0.754585152838428,
+ "grad_norm": 0.06151146441698074,
+ "learning_rate": 0.000318,
+ "loss": 7.231061935424805,
+ "step": 54
+ },
+ {
+ "epoch": 0.7685589519650655,
+ "grad_norm": 0.05638071522116661,
+ "learning_rate": 0.000324,
+ "loss": 7.195466995239258,
+ "step": 55
+ },
+ {
+ "epoch": 0.7825327510917031,
+ "grad_norm": 0.05084463208913803,
+ "learning_rate": 0.00033,
+ "loss": 7.137603759765625,
+ "step": 56
+ },
+ {
+ "epoch": 0.7965065502183406,
+ "grad_norm": 0.04293261095881462,
+ "learning_rate": 0.000336,
+ "loss": 7.115458965301514,
+ "step": 57
+ },
+ {
+ "epoch": 0.8104803493449781,
+ "grad_norm": 0.036064863204956055,
+ "learning_rate": 0.00034199999999999996,
+ "loss": 7.089038848876953,
+ "step": 58
+ },
+ {
+ "epoch": 0.8244541484716157,
+ "grad_norm": 0.029026754200458527,
+ "learning_rate": 0.00034799999999999995,
+ "loss": 7.0584306716918945,
+ "step": 59
+ },
+ {
+ "epoch": 0.8384279475982532,
+ "grad_norm": 0.02286517433822155,
+ "learning_rate": 0.00035399999999999993,
+ "loss": 7.049450874328613,
+ "step": 60
+ },
+ {
+ "epoch": 0.8524017467248908,
+ "grad_norm": 0.018878433853387833,
+ "learning_rate": 0.00035999999999999997,
+ "loss": 7.008824348449707,
+ "step": 61
+ },
+ {
+ "epoch": 0.8663755458515284,
+ "grad_norm": 0.010773813351988792,
+ "learning_rate": 0.00036599999999999995,
+ "loss": 7.024829864501953,
+ "step": 62
+ },
+ {
+ "epoch": 0.880349344978166,
+ "grad_norm": 0.008794732391834259,
+ "learning_rate": 0.000372,
+ "loss": 7.018399238586426,
+ "step": 63
+ },
+ {
+ "epoch": 0.8943231441048035,
+ "grad_norm": 0.011709796264767647,
+ "learning_rate": 0.00037799999999999997,
+ "loss": 7.0101752281188965,
+ "step": 64
+ },
+ {
+ "epoch": 0.9082969432314411,
+ "grad_norm": 0.012123258784413338,
+ "learning_rate": 0.00038399999999999996,
+ "loss": 7.024317741394043,
+ "step": 65
+ },
+ {
+ "epoch": 0.9222707423580786,
+ "grad_norm": 0.01530320756137371,
+ "learning_rate": 0.00039,
+ "loss": 7.027738571166992,
+ "step": 66
+ },
+ {
+ "epoch": 0.9362445414847161,
+ "grad_norm": 0.01890002004802227,
+ "learning_rate": 0.000396,
+ "loss": 7.024099349975586,
+ "step": 67
+ },
+ {
+ "epoch": 0.9502183406113537,
+ "grad_norm": 0.020684150978922844,
+ "learning_rate": 0.000402,
+ "loss": 7.007943153381348,
+ "step": 68
+ },
+ {
+ "epoch": 0.9641921397379912,
+ "grad_norm": 0.019826196134090424,
+ "learning_rate": 0.000408,
+ "loss": 7.006874084472656,
+ "step": 69
+ },
+ {
+ "epoch": 0.9781659388646288,
+ "grad_norm": 0.019218673929572105,
+ "learning_rate": 0.0004139999999999999,
+ "loss": 7.002897262573242,
+ "step": 70
+ },
+ {
+ "epoch": 0.9921397379912664,
+ "grad_norm": 0.013856411911547184,
+ "learning_rate": 0.00041999999999999996,
+ "loss": 6.993999481201172,
+ "step": 71
+ },
+ {
+ "epoch": 1.0,
+ "grad_norm": 0.010390003211796284,
+ "learning_rate": 0.00042599999999999995,
+ "loss": 7.000683784484863,
+ "step": 72
+ },
+ {
+ "epoch": 1.0,
+ "eval_loss": 7.91140079498291,
+ "eval_runtime": 57.7925,
+ "eval_samples_per_second": 42.255,
+ "eval_steps_per_second": 1.332,
+ "step": 72
+ },
+ {
+ "epoch": 1.0139737991266375,
+ "grad_norm": 0.013802163302898407,
+ "learning_rate": 0.00043199999999999993,
+ "loss": 6.998257637023926,
+ "step": 73
+ },
+ {
+ "epoch": 1.027947598253275,
+ "grad_norm": 0.010109687224030495,
+ "learning_rate": 0.00043799999999999997,
+ "loss": 7.002542018890381,
+ "step": 74
+ },
+ {
+ "epoch": 1.0419213973799126,
+ "grad_norm": 0.008019139990210533,
+ "learning_rate": 0.00044399999999999995,
+ "loss": 6.985941410064697,
+ "step": 75
+ },
+ {
+ "epoch": 1.0558951965065502,
+ "grad_norm": 0.0069672465324401855,
+ "learning_rate": 0.00045,
+ "loss": 6.99245023727417,
+ "step": 76
+ },
+ {
+ "epoch": 1.0698689956331877,
+ "grad_norm": 0.00714076729491353,
+ "learning_rate": 0.00045599999999999997,
+ "loss": 6.959296226501465,
+ "step": 77
+ },
+ {
+ "epoch": 1.0838427947598253,
+ "grad_norm": 0.008408435620367527,
+ "learning_rate": 0.00046199999999999995,
+ "loss": 6.954504489898682,
+ "step": 78
+ },
+ {
+ "epoch": 1.0978165938864628,
+ "grad_norm": 0.008778786286711693,
+ "learning_rate": 0.000468,
+ "loss": 7.0007004737854,
+ "step": 79
+ },
+ {
+ "epoch": 1.1117903930131003,
+ "grad_norm": 0.007753407116979361,
+ "learning_rate": 0.000474,
+ "loss": 6.98183536529541,
+ "step": 80
+ },
+ {
+ "epoch": 1.125764192139738,
+ "grad_norm": 0.01087260153144598,
+ "learning_rate": 0.00047999999999999996,
+ "loss": 6.991066932678223,
+ "step": 81
+ },
+ {
+ "epoch": 1.1397379912663754,
+ "grad_norm": 0.006122548598796129,
+ "learning_rate": 0.000486,
+ "loss": 6.978824138641357,
+ "step": 82
+ },
+ {
+ "epoch": 1.1537117903930132,
+ "grad_norm": 0.0067383465357124805,
+ "learning_rate": 0.0004919999999999999,
+ "loss": 6.961703777313232,
+ "step": 83
+ },
+ {
+ "epoch": 1.1676855895196507,
+ "grad_norm": 0.007922791875898838,
+ "learning_rate": 0.000498,
+ "loss": 6.954860687255859,
+ "step": 84
+ },
+ {
+ "epoch": 1.1816593886462883,
+ "grad_norm": 0.00818922370672226,
+ "learning_rate": 0.0005039999999999999,
+ "loss": 6.945961952209473,
+ "step": 85
+ },
+ {
+ "epoch": 1.1956331877729258,
+ "grad_norm": 0.006791461259126663,
+ "learning_rate": 0.0005099999999999999,
+ "loss": 6.968680381774902,
+ "step": 86
+ },
+ {
+ "epoch": 1.2096069868995634,
+ "grad_norm": 0.005394163075834513,
+ "learning_rate": 0.000516,
+ "loss": 6.956404685974121,
+ "step": 87
+ },
+ {
+ "epoch": 1.223580786026201,
+ "grad_norm": 0.004500220064073801,
+ "learning_rate": 0.000522,
+ "loss": 6.96998405456543,
+ "step": 88
+ },
+ {
+ "epoch": 1.2375545851528384,
+ "grad_norm": 0.008699574507772923,
+ "learning_rate": 0.0005279999999999999,
+ "loss": 6.955585479736328,
+ "step": 89
+ },
+ {
+ "epoch": 1.251528384279476,
+ "grad_norm": 0.005889365915209055,
+ "learning_rate": 0.000534,
+ "loss": 6.9708051681518555,
+ "step": 90
+ },
+ {
+ "epoch": 1.2655021834061135,
+ "grad_norm": 0.009646810591220856,
+ "learning_rate": 0.00054,
+ "loss": 6.962608814239502,
+ "step": 91
+ },
+ {
+ "epoch": 1.279475982532751,
+ "grad_norm": 0.004517171997576952,
+ "learning_rate": 0.0005459999999999999,
+ "loss": 6.957627296447754,
+ "step": 92
+ },
+ {
+ "epoch": 1.2934497816593886,
+ "grad_norm": 0.006339677143841982,
+ "learning_rate": 0.000552,
+ "loss": 6.963141441345215,
+ "step": 93
+ },
+ {
+ "epoch": 1.3074235807860262,
+ "grad_norm": 0.005387383047491312,
+ "learning_rate": 0.000558,
+ "loss": 6.934465408325195,
+ "step": 94
+ },
+ {
+ "epoch": 1.3213973799126637,
+ "grad_norm": 0.005252422299236059,
+ "learning_rate": 0.0005639999999999999,
+ "loss": 6.956108093261719,
+ "step": 95
+ },
+ {
+ "epoch": 1.3353711790393012,
+ "grad_norm": 0.008699414320290089,
+ "learning_rate": 0.00057,
+ "loss": 6.973369598388672,
+ "step": 96
+ },
+ {
+ "epoch": 1.3493449781659388,
+ "grad_norm": 0.003947409335523844,
+ "learning_rate": 0.0005759999999999999,
+ "loss": 6.950225830078125,
+ "step": 97
+ },
+ {
+ "epoch": 1.3633187772925766,
+ "grad_norm": 0.00768681475892663,
+ "learning_rate": 0.0005819999999999999,
+ "loss": 6.964739799499512,
+ "step": 98
+ },
+ {
+ "epoch": 1.3772925764192139,
+ "grad_norm": 0.006563829258084297,
+ "learning_rate": 0.000588,
+ "loss": 6.972421169281006,
+ "step": 99
+ },
+ {
+ "epoch": 1.3912663755458516,
+ "grad_norm": 0.007606555242091417,
+ "learning_rate": 0.0005939999999999999,
+ "loss": 6.958213806152344,
+ "step": 100
+ },
+ {
+ "epoch": 1.405240174672489,
+ "grad_norm": 0.006250880658626556,
+ "learning_rate": 0.0006,
+ "loss": 6.962296485900879,
+ "step": 101
+ },
+ {
+ "epoch": 1.4192139737991267,
+ "grad_norm": 0.0073033408261835575,
+ "learning_rate": 0.0006,
+ "loss": 6.969705104827881,
+ "step": 102
+ },
+ {
+ "epoch": 1.4331877729257643,
+ "grad_norm": 0.0054510910995304585,
+ "learning_rate": 0.0006,
+ "loss": 6.9652862548828125,
+ "step": 103
+ },
+ {
+ "epoch": 1.4471615720524018,
+ "grad_norm": 0.004895673133432865,
+ "learning_rate": 0.0006,
+ "loss": 6.956023693084717,
+ "step": 104
+ },
+ {
+ "epoch": 1.4611353711790394,
+ "grad_norm": 0.005306515377014875,
+ "learning_rate": 0.0006,
+ "loss": 6.963231086730957,
+ "step": 105
+ },
+ {
+ "epoch": 1.475109170305677,
+ "grad_norm": 0.004742146469652653,
+ "learning_rate": 0.0006,
+ "loss": 6.9582061767578125,
+ "step": 106
+ },
+ {
+ "epoch": 1.4890829694323144,
+ "grad_norm": 0.004220000468194485,
+ "learning_rate": 0.0006,
+ "loss": 6.972961902618408,
+ "step": 107
+ },
+ {
+ "epoch": 1.503056768558952,
+ "grad_norm": 0.0042153168469667435,
+ "learning_rate": 0.0006,
+ "loss": 6.961165904998779,
+ "step": 108
+ },
+ {
+ "epoch": 1.5170305676855895,
+ "grad_norm": 0.005165040958672762,
+ "learning_rate": 0.0006,
+ "loss": 6.95018196105957,
+ "step": 109
+ },
+ {
+ "epoch": 1.531004366812227,
+ "grad_norm": 0.003930266480892897,
+ "learning_rate": 0.0006,
+ "loss": 6.9651689529418945,
+ "step": 110
+ },
+ {
+ "epoch": 1.5449781659388646,
+ "grad_norm": 0.004465071018785238,
+ "learning_rate": 0.0006,
+ "loss": 6.960553169250488,
+ "step": 111
+ },
+ {
+ "epoch": 1.5589519650655022,
+ "grad_norm": 0.004888612311333418,
+ "learning_rate": 0.0006,
+ "loss": 6.967258930206299,
+ "step": 112
+ },
+ {
+ "epoch": 1.5729257641921397,
+ "grad_norm": 0.007497863378375769,
+ "learning_rate": 0.0006,
+ "loss": 6.954545021057129,
+ "step": 113
+ },
+ {
+ "epoch": 1.5868995633187772,
+ "grad_norm": 0.005659648682922125,
+ "learning_rate": 0.0006,
+ "loss": 6.962299823760986,
+ "step": 114
+ },
+ {
+ "epoch": 1.600873362445415,
+ "grad_norm": 0.006641396787017584,
+ "learning_rate": 0.0006,
+ "loss": 6.952983856201172,
+ "step": 115
+ },
+ {
+ "epoch": 1.6148471615720523,
+ "grad_norm": 0.009715928696095943,
+ "learning_rate": 0.0006,
+ "loss": 6.958622932434082,
+ "step": 116
+ },
+ {
+ "epoch": 1.62882096069869,
+ "grad_norm": 0.012789523229002953,
+ "learning_rate": 0.0006,
+ "loss": 6.932043075561523,
+ "step": 117
+ },
+ {
+ "epoch": 1.6427947598253274,
+ "grad_norm": 0.021673237904906273,
+ "learning_rate": 0.0006,
+ "loss": 6.931290149688721,
+ "step": 118
+ },
+ {
+ "epoch": 1.6567685589519652,
+ "grad_norm": 0.16024991869926453,
+ "learning_rate": 0.0006,
+ "loss": 6.979497909545898,
+ "step": 119
+ },
+ {
+ "epoch": 1.6707423580786025,
+ "grad_norm": 0.2434523105621338,
+ "learning_rate": 0.0006,
+ "loss": 7.185218811035156,
+ "step": 120
+ },
+ {
+ "epoch": 1.6847161572052403,
+ "grad_norm": 0.39252421259880066,
+ "learning_rate": 0.0006,
+ "loss": 7.137226581573486,
+ "step": 121
+ },
+ {
+ "epoch": 1.6986899563318776,
+ "grad_norm": 0.10565312206745148,
+ "learning_rate": 0.0006,
+ "loss": 6.984494209289551,
+ "step": 122
+ },
+ {
+ "epoch": 1.7126637554585153,
+ "grad_norm": 0.1033431813120842,
+ "learning_rate": 0.0006,
+ "loss": 7.001287937164307,
+ "step": 123
+ },
+ {
+ "epoch": 1.726637554585153,
+ "grad_norm": 0.14875490963459015,
+ "learning_rate": 0.0006,
+ "loss": 7.051436901092529,
+ "step": 124
+ },
+ {
+ "epoch": 1.7406113537117904,
+ "grad_norm": 0.1374562531709671,
+ "learning_rate": 0.0006,
+ "loss": 7.0293121337890625,
+ "step": 125
+ },
+ {
+ "epoch": 1.754585152838428,
+ "grad_norm": 0.0910211056470871,
+ "learning_rate": 0.0006,
+ "loss": 7.007458686828613,
+ "step": 126
+ },
+ {
+ "epoch": 1.7685589519650655,
+ "grad_norm": 0.021633487194776535,
+ "learning_rate": 0.0006,
+ "loss": 6.996816635131836,
+ "step": 127
+ },
+ {
+ "epoch": 1.782532751091703,
+ "grad_norm": 0.08692754060029984,
+ "learning_rate": 0.0006,
+ "loss": 6.998541831970215,
+ "step": 128
+ },
+ {
+ "epoch": 1.7965065502183406,
+ "grad_norm": 0.12613292038440704,
+ "learning_rate": 0.0006,
+ "loss": 6.9970703125,
+ "step": 129
+ },
+ {
+ "epoch": 1.8104803493449781,
+ "grad_norm": 0.09462810307741165,
+ "learning_rate": 0.0006,
+ "loss": 6.997045040130615,
+ "step": 130
+ },
+ {
+ "epoch": 1.8244541484716157,
+ "grad_norm": 0.03534962609410286,
+ "learning_rate": 0.0006,
+ "loss": 6.9738450050354,
+ "step": 131
+ },
+ {
+ "epoch": 1.8384279475982532,
+ "grad_norm": 0.033839885145425797,
+ "learning_rate": 0.0006,
+ "loss": 6.946001052856445,
+ "step": 132
+ },
+ {
+ "epoch": 1.8524017467248908,
+ "grad_norm": 0.06263269484043121,
+ "learning_rate": 0.0006,
+ "loss": 6.968981742858887,
+ "step": 133
+ },
+ {
+ "epoch": 1.8663755458515285,
+ "grad_norm": 0.07117997109889984,
+ "learning_rate": 0.0006,
+ "loss": 6.981535911560059,
+ "step": 134
+ },
+ {
+ "epoch": 1.8803493449781659,
+ "grad_norm": 0.06336277723312378,
+ "learning_rate": 0.0006,
+ "loss": 6.976946830749512,
+ "step": 135
+ },
+ {
+ "epoch": 1.8943231441048036,
+ "grad_norm": 0.047467734664678574,
+ "learning_rate": 0.0006,
+ "loss": 6.961673736572266,
+ "step": 136
+ },
+ {
+ "epoch": 1.908296943231441,
+ "grad_norm": 0.018787868320941925,
+ "learning_rate": 0.0006,
+ "loss": 6.948919296264648,
+ "step": 137
+ },
+ {
+ "epoch": 1.9222707423580787,
+ "grad_norm": 0.025909408926963806,
+ "learning_rate": 0.0006,
+ "loss": 6.976190567016602,
+ "step": 138
+ },
+ {
+ "epoch": 1.936244541484716,
+ "grad_norm": 0.052842333912849426,
+ "learning_rate": 0.0006,
+ "loss": 6.971207618713379,
+ "step": 139
+ },
+ {
+ "epoch": 1.9502183406113538,
+ "grad_norm": 0.06421888619661331,
+ "learning_rate": 0.0006,
+ "loss": 6.979109764099121,
+ "step": 140
+ },
+ {
+ "epoch": 1.9641921397379911,
+ "grad_norm": 0.055596910417079926,
+ "learning_rate": 0.0006,
+ "loss": 6.972118854522705,
+ "step": 141
+ },
+ {
+ "epoch": 1.9781659388646289,
+ "grad_norm": 0.03267091140151024,
+ "learning_rate": 0.0006,
+ "loss": 6.947260856628418,
+ "step": 142
+ },
+ {
+ "epoch": 1.9921397379912664,
+ "grad_norm": 0.008032613433897495,
+ "learning_rate": 0.0006,
+ "loss": 6.936816215515137,
+ "step": 143
+ },
+ {
+ "epoch": 2.0,
+ "grad_norm": 0.020184025168418884,
+ "learning_rate": 0.0006,
+ "loss": 6.929848670959473,
+ "step": 144
+ },
+ {
+ "epoch": 2.0,
+ "eval_loss": 7.836523056030273,
+ "eval_runtime": 58.5622,
+ "eval_samples_per_second": 41.699,
+ "eval_steps_per_second": 1.315,
+ "step": 144
+ },
+ {
+ "epoch": 2.0139737991266378,
+ "grad_norm": 0.031980376690626144,
+ "learning_rate": 0.0006,
+ "loss": 6.968081474304199,
+ "step": 145
+ },
+ {
+ "epoch": 2.027947598253275,
+ "grad_norm": 0.03802533075213432,
+ "learning_rate": 0.0006,
+ "loss": 6.95728874206543,
+ "step": 146
+ },
+ {
+ "epoch": 2.041921397379913,
+ "grad_norm": 0.03796226158738136,
+ "learning_rate": 0.0006,
+ "loss": 6.96516752243042,
+ "step": 147
+ },
+ {
+ "epoch": 2.05589519650655,
+ "grad_norm": 0.02621951326727867,
+ "learning_rate": 0.0006,
+ "loss": 6.956058502197266,
+ "step": 148
+ },
+ {
+ "epoch": 2.069868995633188,
+ "grad_norm": 0.013292718678712845,
+ "learning_rate": 0.0006,
+ "loss": 6.9783935546875,
+ "step": 149
+ },
+ {
+ "epoch": 2.0838427947598253,
+ "grad_norm": 0.009446968324482441,
+ "learning_rate": 0.0006,
+ "loss": 6.96873664855957,
+ "step": 150
+ },
+ {
+ "epoch": 2.097816593886463,
+ "grad_norm": 0.021703092381358147,
+ "learning_rate": 0.0006,
+ "loss": 6.959234237670898,
+ "step": 151
+ },
+ {
+ "epoch": 2.1117903930131003,
+ "grad_norm": 0.027765290811657906,
+ "learning_rate": 0.0006,
+ "loss": 6.968524932861328,
+ "step": 152
+ },
+ {
+ "epoch": 2.125764192139738,
+ "grad_norm": 0.028761165216565132,
+ "learning_rate": 0.0006,
+ "loss": 6.954620838165283,
+ "step": 153
+ },
+ {
+ "epoch": 2.1397379912663754,
+ "grad_norm": 0.019729383289813995,
+ "learning_rate": 0.0006,
+ "loss": 6.949009895324707,
+ "step": 154
+ },
+ {
+ "epoch": 2.153711790393013,
+ "grad_norm": 0.008990089409053326,
+ "learning_rate": 0.0006,
+ "loss": 6.956226348876953,
+ "step": 155
+ },
+ {
+ "epoch": 2.1676855895196505,
+ "grad_norm": 0.00549644511193037,
+ "learning_rate": 0.0006,
+ "loss": 6.957143783569336,
+ "step": 156
+ },
+ {
+ "epoch": 2.1816593886462883,
+ "grad_norm": 0.014436670579016209,
+ "learning_rate": 0.0006,
+ "loss": 6.954257011413574,
+ "step": 157
+ },
+ {
+ "epoch": 2.1956331877729256,
+ "grad_norm": 0.01990494504570961,
+ "learning_rate": 0.0006,
+ "loss": 6.942131519317627,
+ "step": 158
+ },
+ {
+ "epoch": 2.2096069868995634,
+ "grad_norm": 0.019155776128172874,
+ "learning_rate": 0.0006,
+ "loss": 6.946969985961914,
+ "step": 159
+ },
+ {
+ "epoch": 2.2235807860262007,
+ "grad_norm": 0.014970783144235611,
+ "learning_rate": 0.0006,
+ "loss": 6.957938194274902,
+ "step": 160
+ },
+ {
+ "epoch": 2.2375545851528384,
+ "grad_norm": 0.009875659830868244,
+ "learning_rate": 0.0006,
+ "loss": 6.953309059143066,
+ "step": 161
+ },
+ {
+ "epoch": 2.251528384279476,
+ "grad_norm": 0.004792087245732546,
+ "learning_rate": 0.0006,
+ "loss": 6.975146770477295,
+ "step": 162
+ },
+ {
+ "epoch": 2.2655021834061135,
+ "grad_norm": 0.009936527349054813,
+ "learning_rate": 0.0006,
+ "loss": 6.961674690246582,
+ "step": 163
+ },
+ {
+ "epoch": 2.279475982532751,
+ "grad_norm": 0.015100127086043358,
+ "learning_rate": 0.0006,
+ "loss": 6.968759536743164,
+ "step": 164
+ },
+ {
+ "epoch": 2.2934497816593886,
+ "grad_norm": 0.01449542585760355,
+ "learning_rate": 0.0006,
+ "loss": 6.966510772705078,
+ "step": 165
+ },
+ {
+ "epoch": 2.3074235807860264,
+ "grad_norm": 0.01073537115007639,
+ "learning_rate": 0.0006,
+ "loss": 6.936373710632324,
+ "step": 166
+ },
+ {
+ "epoch": 2.3213973799126637,
+ "grad_norm": 0.006766272243112326,
+ "learning_rate": 0.0006,
+ "loss": 6.96397590637207,
+ "step": 167
+ },
+ {
+ "epoch": 2.3353711790393015,
+ "grad_norm": 0.005055439658463001,
+ "learning_rate": 0.0006,
+ "loss": 6.961050987243652,
+ "step": 168
+ },
+ {
+ "epoch": 2.349344978165939,
+ "grad_norm": 0.006252622697502375,
+ "learning_rate": 0.0006,
+ "loss": 6.980506420135498,
+ "step": 169
+ },
+ {
+ "epoch": 2.3633187772925766,
+ "grad_norm": 0.008084346540272236,
+ "learning_rate": 0.0006,
+ "loss": 6.9346604347229,
+ "step": 170
+ },
+ {
+ "epoch": 2.377292576419214,
+ "grad_norm": 0.010260380804538727,
+ "learning_rate": 0.0006,
+ "loss": 6.964118957519531,
+ "step": 171
+ },
+ {
+ "epoch": 2.3912663755458516,
+ "grad_norm": 0.010697875171899796,
+ "learning_rate": 0.0006,
+ "loss": 6.941572189331055,
+ "step": 172
+ },
+ {
+ "epoch": 2.405240174672489,
+ "grad_norm": 0.007338830269873142,
+ "learning_rate": 0.0006,
+ "loss": 6.950471878051758,
+ "step": 173
+ },
+ {
+ "epoch": 2.4192139737991267,
+ "grad_norm": 0.005931832827627659,
+ "learning_rate": 0.0006,
+ "loss": 6.962935924530029,
+ "step": 174
+ },
+ {
+ "epoch": 2.433187772925764,
+ "grad_norm": 0.0061243255622684956,
+ "learning_rate": 0.0006,
+ "loss": 6.946418285369873,
+ "step": 175
+ },
+ {
+ "epoch": 2.447161572052402,
+ "grad_norm": 0.007983328774571419,
+ "learning_rate": 0.0006,
+ "loss": 6.959758758544922,
+ "step": 176
+ },
+ {
+ "epoch": 2.461135371179039,
+ "grad_norm": 0.010470672510564327,
+ "learning_rate": 0.0006,
+ "loss": 6.93851375579834,
+ "step": 177
+ },
+ {
+ "epoch": 2.475109170305677,
+ "grad_norm": 0.011147388257086277,
+ "learning_rate": 0.0006,
+ "loss": 6.93258810043335,
+ "step": 178
+ },
+ {
+ "epoch": 2.489082969432314,
+ "grad_norm": 0.01058544497936964,
+ "learning_rate": 0.0006,
+ "loss": 6.932259559631348,
+ "step": 179
+ },
+ {
+ "epoch": 2.503056768558952,
+ "grad_norm": 0.016850745305418968,
+ "learning_rate": 0.0006,
+ "loss": 6.924542427062988,
+ "step": 180
+ },
+ {
+ "epoch": 2.5170305676855893,
+ "grad_norm": 0.023101547732949257,
+ "learning_rate": 0.0006,
+ "loss": 6.914878845214844,
+ "step": 181
+ },
+ {
+ "epoch": 2.531004366812227,
+ "grad_norm": 0.015795735642313957,
+ "learning_rate": 0.0006,
+ "loss": 6.922432899475098,
+ "step": 182
+ },
+ {
+ "epoch": 2.544978165938865,
+ "grad_norm": 0.008821426890790462,
+ "learning_rate": 0.0006,
+ "loss": 6.924562454223633,
+ "step": 183
+ },
+ {
+ "epoch": 2.558951965065502,
+ "grad_norm": 0.012502548284828663,
+ "learning_rate": 0.0006,
+ "loss": 6.914718151092529,
+ "step": 184
+ },
+ {
+ "epoch": 2.5729257641921395,
+ "grad_norm": 0.011545066721737385,
+ "learning_rate": 0.0006,
+ "loss": 6.915705680847168,
+ "step": 185
+ },
+ {
+ "epoch": 2.5868995633187772,
+ "grad_norm": 0.022190893068909645,
+ "learning_rate": 0.0006,
+ "loss": 6.88851261138916,
+ "step": 186
+ },
+ {
+ "epoch": 2.600873362445415,
+ "grad_norm": 0.023395851254463196,
+ "learning_rate": 0.0006,
+ "loss": 6.877655029296875,
+ "step": 187
+ },
+ {
+ "epoch": 2.6148471615720523,
+ "grad_norm": 0.015601156279444695,
+ "learning_rate": 0.0006,
+ "loss": 6.889644145965576,
+ "step": 188
+ },
+ {
+ "epoch": 2.62882096069869,
+ "grad_norm": 0.01362934336066246,
+ "learning_rate": 0.0006,
+ "loss": 6.891547203063965,
+ "step": 189
+ },
+ {
+ "epoch": 2.6427947598253274,
+ "grad_norm": 0.012305490672588348,
+ "learning_rate": 0.0006,
+ "loss": 6.890357971191406,
+ "step": 190
+ },
+ {
+ "epoch": 2.656768558951965,
+ "grad_norm": 0.01623358018696308,
+ "learning_rate": 0.0006,
+ "loss": 6.849443435668945,
+ "step": 191
+ },
+ {
+ "epoch": 2.6707423580786025,
+ "grad_norm": 0.018063291907310486,
+ "learning_rate": 0.0006,
+ "loss": 6.842372894287109,
+ "step": 192
+ },
+ {
+ "epoch": 2.6847161572052403,
+ "grad_norm": 0.010179434902966022,
+ "learning_rate": 0.0006,
+ "loss": 6.870055198669434,
+ "step": 193
+ },
+ {
+ "epoch": 2.6986899563318776,
+ "grad_norm": 0.009986722841858864,
+ "learning_rate": 0.0006,
+ "loss": 6.859061241149902,
+ "step": 194
+ },
+ {
+ "epoch": 2.7126637554585153,
+ "grad_norm": 0.01117654051631689,
+ "learning_rate": 0.0006,
+ "loss": 6.861199378967285,
+ "step": 195
+ },
+ {
+ "epoch": 2.726637554585153,
+ "grad_norm": 0.008551654405891895,
+ "learning_rate": 0.0006,
+ "loss": 6.850255012512207,
+ "step": 196
+ },
+ {
+ "epoch": 2.7406113537117904,
+ "grad_norm": 0.018904326483607292,
+ "learning_rate": 0.0006,
+ "loss": 6.848639965057373,
+ "step": 197
+ },
+ {
+ "epoch": 2.7545851528384278,
+ "grad_norm": 0.02080724574625492,
+ "learning_rate": 0.0006,
+ "loss": 6.80456018447876,
+ "step": 198
+ },
+ {
+ "epoch": 2.7685589519650655,
+ "grad_norm": 0.012010873295366764,
+ "learning_rate": 0.0006,
+ "loss": 6.848775863647461,
+ "step": 199
+ },
+ {
+ "epoch": 2.7825327510917033,
+ "grad_norm": 0.008113056421279907,
+ "learning_rate": 0.0006,
+ "loss": 6.849387168884277,
+ "step": 200
+ },
+ {
+ "epoch": 2.7965065502183406,
+ "grad_norm": 0.012912592850625515,
+ "learning_rate": 0.0006,
+ "loss": 6.832240104675293,
+ "step": 201
+ },
+ {
+ "epoch": 2.810480349344978,
+ "grad_norm": 0.01146865077316761,
+ "learning_rate": 0.0006,
+ "loss": 6.850383281707764,
+ "step": 202
+ },
+ {
+ "epoch": 2.8244541484716157,
+ "grad_norm": 0.0171892661601305,
+ "learning_rate": 0.0006,
+ "loss": 6.839577674865723,
+ "step": 203
+ },
+ {
+ "epoch": 2.8384279475982535,
+ "grad_norm": 0.012083170004189014,
+ "learning_rate": 0.0006,
+ "loss": 6.818614482879639,
+ "step": 204
+ },
+ {
+ "epoch": 2.8524017467248908,
+ "grad_norm": 0.0062555656768381596,
+ "learning_rate": 0.0006,
+ "loss": 6.831291198730469,
+ "step": 205
+ },
+ {
+ "epoch": 2.8663755458515285,
+ "grad_norm": 0.010102466680109501,
+ "learning_rate": 0.0006,
+ "loss": 6.846522808074951,
+ "step": 206
+ },
+ {
+ "epoch": 2.880349344978166,
+ "grad_norm": 0.009042087942361832,
+ "learning_rate": 0.0006,
+ "loss": 6.822072982788086,
+ "step": 207
+ },
+ {
+ "epoch": 2.8943231441048036,
+ "grad_norm": 0.011133556254208088,
+ "learning_rate": 0.0006,
+ "loss": 6.811252593994141,
+ "step": 208
+ },
+ {
+ "epoch": 2.908296943231441,
+ "grad_norm": 0.00990577507764101,
+ "learning_rate": 0.0006,
+ "loss": 6.814435958862305,
+ "step": 209
+ },
+ {
+ "epoch": 2.9222707423580787,
+ "grad_norm": 0.006605073343962431,
+ "learning_rate": 0.0006,
+ "loss": 6.831300735473633,
+ "step": 210
+ },
+ {
+ "epoch": 2.936244541484716,
+ "grad_norm": 0.007831175811588764,
+ "learning_rate": 0.0006,
+ "loss": 6.841738700866699,
+ "step": 211
+ },
+ {
+ "epoch": 2.950218340611354,
+ "grad_norm": 0.005536255426704884,
+ "learning_rate": 0.0006,
+ "loss": 6.8043670654296875,
+ "step": 212
+ },
+ {
+ "epoch": 2.964192139737991,
+ "grad_norm": 0.006532667670398951,
+ "learning_rate": 0.0006,
+ "loss": 6.8091840744018555,
+ "step": 213
+ },
+ {
+ "epoch": 2.978165938864629,
+ "grad_norm": 0.006674565840512514,
+ "learning_rate": 0.0006,
+ "loss": 6.811327934265137,
+ "step": 214
+ },
+ {
+ "epoch": 2.992139737991266,
+ "grad_norm": 0.01150344219058752,
+ "learning_rate": 0.0006,
+ "loss": 6.810762405395508,
+ "step": 215
+ },
+ {
+ "epoch": 3.0,
+ "grad_norm": 0.009015528485178947,
+ "learning_rate": 0.0006,
+ "loss": 6.836431980133057,
+ "step": 216
+ },
+ {
+ "epoch": 3.0,
+ "eval_loss": 7.718832969665527,
+ "eval_runtime": 58.6615,
+ "eval_samples_per_second": 41.629,
+ "eval_steps_per_second": 1.313,
+ "step": 216
+ },
+ {
+ "epoch": 3.0139737991266378,
+ "grad_norm": 0.009791829623281956,
+ "learning_rate": 0.0006,
+ "loss": 6.8102593421936035,
+ "step": 217
+ },
+ {
+ "epoch": 3.027947598253275,
+ "grad_norm": 0.008970947004854679,
+ "learning_rate": 0.0006,
+ "loss": 6.792663097381592,
+ "step": 218
+ },
+ {
+ "epoch": 3.041921397379913,
+ "grad_norm": 0.013689392246305943,
+ "learning_rate": 0.0006,
+ "loss": 6.815651893615723,
+ "step": 219
+ },
+ {
+ "epoch": 3.05589519650655,
+ "grad_norm": 0.031499531120061874,
+ "learning_rate": 0.0006,
+ "loss": 6.809841156005859,
+ "step": 220
+ },
+ {
+ "epoch": 3.069868995633188,
+ "grad_norm": 0.03203094005584717,
+ "learning_rate": 0.0006,
+ "loss": 6.779417037963867,
+ "step": 221
+ },
+ {
+ "epoch": 3.0838427947598253,
+ "grad_norm": 0.007947996258735657,
+ "learning_rate": 0.0006,
+ "loss": 6.773656845092773,
+ "step": 222
+ },
+ {
+ "epoch": 3.097816593886463,
+ "grad_norm": 0.03152519091963768,
+ "learning_rate": 0.0006,
+ "loss": 6.804083824157715,
+ "step": 223
+ },
+ {
+ "epoch": 3.1117903930131003,
+ "grad_norm": 0.014937258325517178,
+ "learning_rate": 0.0006,
+ "loss": 6.792213439941406,
+ "step": 224
+ },
+ {
+ "epoch": 3.125764192139738,
+ "grad_norm": 0.020961705595254898,
+ "learning_rate": 0.0006,
+ "loss": 6.8092803955078125,
+ "step": 225
+ },
+ {
+ "epoch": 3.1397379912663754,
+ "grad_norm": 0.009152988903224468,
+ "learning_rate": 0.0006,
+ "loss": 6.7737603187561035,
+ "step": 226
+ },
+ {
+ "epoch": 3.153711790393013,
+ "grad_norm": 0.02770620957016945,
+ "learning_rate": 0.0006,
+ "loss": 6.767014503479004,
+ "step": 227
+ },
+ {
+ "epoch": 3.1676855895196505,
+ "grad_norm": 0.01165871974080801,
+ "learning_rate": 0.0006,
+ "loss": 6.760884761810303,
+ "step": 228
+ },
+ {
+ "epoch": 3.1816593886462883,
+ "grad_norm": 0.01910376362502575,
+ "learning_rate": 0.0006,
+ "loss": 6.775722980499268,
+ "step": 229
+ },
+ {
+ "epoch": 3.1956331877729256,
+ "grad_norm": 0.010413747280836105,
+ "learning_rate": 0.0006,
+ "loss": 6.779816627502441,
+ "step": 230
+ },
+ {
+ "epoch": 3.2096069868995634,
+ "grad_norm": 0.01918417401611805,
+ "learning_rate": 0.0006,
+ "loss": 6.760381698608398,
+ "step": 231
+ },
+ {
+ "epoch": 3.2235807860262007,
+ "grad_norm": 0.017960334196686745,
+ "learning_rate": 0.0006,
+ "loss": 6.761198997497559,
+ "step": 232
+ },
+ {
+ "epoch": 3.2375545851528384,
+ "grad_norm": 0.016546795144677162,
+ "learning_rate": 0.0006,
+ "loss": 6.742448806762695,
+ "step": 233
+ },
+ {
+ "epoch": 3.251528384279476,
+ "grad_norm": 0.01889597810804844,
+ "learning_rate": 0.0006,
+ "loss": 6.758237838745117,
+ "step": 234
+ },
+ {
+ "epoch": 3.2655021834061135,
+ "grad_norm": 0.009611992165446281,
+ "learning_rate": 0.0006,
+ "loss": 6.75614070892334,
+ "step": 235
+ },
+ {
+ "epoch": 3.279475982532751,
+ "grad_norm": 0.02034934051334858,
+ "learning_rate": 0.0006,
+ "loss": 6.736965179443359,
+ "step": 236
+ },
+ {
+ "epoch": 3.2934497816593886,
+ "grad_norm": 0.010740065947175026,
+ "learning_rate": 0.0006,
+ "loss": 6.746947288513184,
+ "step": 237
+ },
+ {
+ "epoch": 3.3074235807860264,
+ "grad_norm": 0.018383340910077095,
+ "learning_rate": 0.0006,
+ "loss": 6.747332572937012,
+ "step": 238
+ },
+ {
+ "epoch": 3.3213973799126637,
+ "grad_norm": 0.012362911365926266,
+ "learning_rate": 0.0006,
+ "loss": 6.767669200897217,
+ "step": 239
+ },
+ {
+ "epoch": 3.3353711790393015,
+ "grad_norm": 0.01127463299781084,
+ "learning_rate": 0.0006,
+ "loss": 6.769808292388916,
+ "step": 240
+ },
+ {
+ "epoch": 3.349344978165939,
+ "grad_norm": 0.017693769186735153,
+ "learning_rate": 0.0006,
+ "loss": 6.733658790588379,
+ "step": 241
+ },
+ {
+ "epoch": 3.3633187772925766,
+ "grad_norm": 0.006420376244932413,
+ "learning_rate": 0.0006,
+ "loss": 6.74934720993042,
+ "step": 242
+ },
+ {
+ "epoch": 3.377292576419214,
+ "grad_norm": 0.014020364731550217,
+ "learning_rate": 0.0006,
+ "loss": 6.728157043457031,
+ "step": 243
+ },
+ {
+ "epoch": 3.3912663755458516,
+ "grad_norm": 0.008018582127988338,
+ "learning_rate": 0.0006,
+ "loss": 6.730726718902588,
+ "step": 244
+ },
+ {
+ "epoch": 3.405240174672489,
+ "grad_norm": 0.00782222580164671,
+ "learning_rate": 0.0006,
+ "loss": 6.7499895095825195,
+ "step": 245
+ },
+ {
+ "epoch": 3.4192139737991267,
+ "grad_norm": 0.008278409950435162,
+ "learning_rate": 0.0006,
+ "loss": 6.712803363800049,
+ "step": 246
+ },
+ {
+ "epoch": 3.433187772925764,
+ "grad_norm": 0.009826759807765484,
+ "learning_rate": 0.0006,
+ "loss": 6.742262840270996,
+ "step": 247
+ },
+ {
+ "epoch": 3.447161572052402,
+ "grad_norm": 0.009488224983215332,
+ "learning_rate": 0.0006,
+ "loss": 6.739130020141602,
+ "step": 248
+ },
+ {
+ "epoch": 3.461135371179039,
+ "grad_norm": 0.007667903788387775,
+ "learning_rate": 0.0006,
+ "loss": 6.747515678405762,
+ "step": 249
+ },
+ {
+ "epoch": 3.475109170305677,
+ "grad_norm": 0.0073090000078082085,
+ "learning_rate": 0.0006,
+ "loss": 6.740870475769043,
+ "step": 250
+ },
+ {
+ "epoch": 3.489082969432314,
+ "grad_norm": 0.010058708488941193,
+ "learning_rate": 0.0006,
+ "loss": 6.777563095092773,
+ "step": 251
+ },
+ {
+ "epoch": 3.503056768558952,
+ "grad_norm": 0.020676063373684883,
+ "learning_rate": 0.0006,
+ "loss": 6.708603858947754,
+ "step": 252
+ },
+ {
+ "epoch": 3.5170305676855893,
+ "grad_norm": 0.027210544794797897,
+ "learning_rate": 0.0006,
+ "loss": 6.694097995758057,
+ "step": 253
+ },
+ {
+ "epoch": 3.531004366812227,
+ "grad_norm": 0.038053520023822784,
+ "learning_rate": 0.0006,
+ "loss": 6.697017669677734,
+ "step": 254
+ },
+ {
+ "epoch": 3.544978165938865,
+ "grad_norm": 0.04684312269091606,
+ "learning_rate": 0.0006,
+ "loss": 6.729883193969727,
+ "step": 255
+ },
+ {
+ "epoch": 3.558951965065502,
+ "grad_norm": 0.03673060983419418,
+ "learning_rate": 0.0006,
+ "loss": 6.743753433227539,
+ "step": 256
+ },
+ {
+ "epoch": 3.5729257641921395,
+ "grad_norm": 0.006882632151246071,
+ "learning_rate": 0.0006,
+ "loss": 6.71405029296875,
+ "step": 257
+ },
+ {
+ "epoch": 3.5868995633187772,
+ "grad_norm": 0.03588201105594635,
+ "learning_rate": 0.0006,
+ "loss": 6.712467193603516,
+ "step": 258
+ },
+ {
+ "epoch": 3.600873362445415,
+ "grad_norm": 0.029415199533104897,
+ "learning_rate": 0.0006,
+ "loss": 6.715634822845459,
+ "step": 259
+ },
+ {
+ "epoch": 3.6148471615720523,
+ "grad_norm": 0.011425039730966091,
+ "learning_rate": 0.0006,
+ "loss": 6.697334289550781,
+ "step": 260
+ },
+ {
+ "epoch": 3.62882096069869,
+ "grad_norm": 0.0297263003885746,
+ "learning_rate": 0.0006,
+ "loss": 6.724848747253418,
+ "step": 261
+ },
+ {
+ "epoch": 3.6427947598253274,
+ "grad_norm": 0.007282314356416464,
+ "learning_rate": 0.0006,
+ "loss": 6.702306270599365,
+ "step": 262
+ },
+ {
+ "epoch": 3.656768558951965,
+ "grad_norm": 0.02673051692545414,
+ "learning_rate": 0.0006,
+ "loss": 6.72357177734375,
+ "step": 263
+ },
+ {
+ "epoch": 3.6707423580786025,
+ "grad_norm": 0.016174081712961197,
+ "learning_rate": 0.0006,
+ "loss": 6.730698585510254,
+ "step": 264
+ },
+ {
+ "epoch": 3.6847161572052403,
+ "grad_norm": 0.0182416420429945,
+ "learning_rate": 0.0006,
+ "loss": 6.717386245727539,
+ "step": 265
+ },
+ {
+ "epoch": 3.6986899563318776,
+ "grad_norm": 0.023250866681337357,
+ "learning_rate": 0.0006,
+ "loss": 6.697832107543945,
+ "step": 266
+ },
+ {
+ "epoch": 3.7126637554585153,
+ "grad_norm": 0.007539753336459398,
+ "learning_rate": 0.0006,
+ "loss": 6.714348793029785,
+ "step": 267
+ },
+ {
+ "epoch": 3.726637554585153,
+ "grad_norm": 0.030086630955338478,
+ "learning_rate": 0.0006,
+ "loss": 6.724481582641602,
+ "step": 268
+ },
+ {
+ "epoch": 3.7406113537117904,
+ "grad_norm": 0.012784913182258606,
+ "learning_rate": 0.0006,
+ "loss": 6.703491687774658,
+ "step": 269
+ },
+ {
+ "epoch": 3.7545851528384278,
+ "grad_norm": 0.02042769268155098,
+ "learning_rate": 0.0006,
+ "loss": 6.721139907836914,
+ "step": 270
+ },
+ {
+ "epoch": 3.7685589519650655,
+ "grad_norm": 0.023557329550385475,
+ "learning_rate": 0.0006,
+ "loss": 6.728122711181641,
+ "step": 271
+ },
+ {
+ "epoch": 3.7825327510917033,
+ "grad_norm": 0.007716621737927198,
+ "learning_rate": 0.0006,
+ "loss": 6.70085334777832,
+ "step": 272
+ },
+ {
+ "epoch": 3.7965065502183406,
+ "grad_norm": 0.022772278636693954,
+ "learning_rate": 0.0006,
+ "loss": 6.712882995605469,
+ "step": 273
+ },
+ {
+ "epoch": 3.810480349344978,
+ "grad_norm": 0.007928933948278427,
+ "learning_rate": 0.0006,
+ "loss": 6.698320388793945,
+ "step": 274
+ },
+ {
+ "epoch": 3.8244541484716157,
+ "grad_norm": 0.022355245426297188,
+ "learning_rate": 0.0006,
+ "loss": 6.687088489532471,
+ "step": 275
+ },
+ {
+ "epoch": 3.8384279475982535,
+ "grad_norm": 0.02037850394845009,
+ "learning_rate": 0.0006,
+ "loss": 6.676856994628906,
+ "step": 276
+ },
+ {
+ "epoch": 3.8524017467248908,
+ "grad_norm": 0.012290362268686295,
+ "learning_rate": 0.0006,
+ "loss": 6.682883262634277,
+ "step": 277
+ },
+ {
+ "epoch": 3.8663755458515285,
+ "grad_norm": 0.022331275045871735,
+ "learning_rate": 0.0006,
+ "loss": 6.694752216339111,
+ "step": 278
+ },
+ {
+ "epoch": 3.880349344978166,
+ "grad_norm": 0.007527181413024664,
+ "learning_rate": 0.0006,
+ "loss": 6.686394691467285,
+ "step": 279
+ },
+ {
+ "epoch": 3.8943231441048036,
+ "grad_norm": 0.02159409038722515,
+ "learning_rate": 0.0006,
+ "loss": 6.697943687438965,
+ "step": 280
+ },
+ {
+ "epoch": 3.908296943231441,
+ "grad_norm": 0.015054763294756413,
+ "learning_rate": 0.0006,
+ "loss": 6.677796363830566,
+ "step": 281
+ },
+ {
+ "epoch": 3.9222707423580787,
+ "grad_norm": 0.01466019731014967,
+ "learning_rate": 0.0006,
+ "loss": 6.697715759277344,
+ "step": 282
+ },
+ {
+ "epoch": 3.936244541484716,
+ "grad_norm": 0.01875465176999569,
+ "learning_rate": 0.0006,
+ "loss": 6.700228691101074,
+ "step": 283
+ },
+ {
+ "epoch": 3.950218340611354,
+ "grad_norm": 0.00778991961851716,
+ "learning_rate": 0.0006,
+ "loss": 6.69331693649292,
+ "step": 284
+ },
+ {
+ "epoch": 3.964192139737991,
+ "grad_norm": 0.020182441920042038,
+ "learning_rate": 0.0006,
+ "loss": 6.672651290893555,
+ "step": 285
+ },
+ {
+ "epoch": 3.978165938864629,
+ "grad_norm": 0.01240516733378172,
+ "learning_rate": 0.0006,
+ "loss": 6.688860893249512,
+ "step": 286
+ },
+ {
+ "epoch": 3.992139737991266,
+ "grad_norm": 0.011218412779271603,
+ "learning_rate": 0.0006,
+ "loss": 6.669534683227539,
+ "step": 287
+ },
+ {
+ "epoch": 4.0,
+ "grad_norm": 0.0141852255910635,
+ "learning_rate": 0.0006,
+ "loss": 6.690206527709961,
+ "step": 288
+ },
+ {
+ "epoch": 4.0,
+ "eval_loss": 7.537620544433594,
+ "eval_runtime": 58.4398,
+ "eval_samples_per_second": 41.787,
+ "eval_steps_per_second": 1.318,
+ "step": 288
+ },
+ {
+ "epoch": 4.013973799126638,
+ "grad_norm": 0.008106146939098835,
+ "learning_rate": 0.0006,
+ "loss": 6.688591003417969,
+ "step": 289
+ },
+ {
+ "epoch": 4.0279475982532755,
+ "grad_norm": 0.014588244259357452,
+ "learning_rate": 0.0006,
+ "loss": 6.684961318969727,
+ "step": 290
+ },
+ {
+ "epoch": 4.041921397379912,
+ "grad_norm": 0.011985368095338345,
+ "learning_rate": 0.0006,
+ "loss": 6.672794342041016,
+ "step": 291
+ },
+ {
+ "epoch": 4.05589519650655,
+ "grad_norm": 0.014632102102041245,
+ "learning_rate": 0.0006,
+ "loss": 6.702736854553223,
+ "step": 292
+ },
+ {
+ "epoch": 4.069868995633188,
+ "grad_norm": 0.011243713088333607,
+ "learning_rate": 0.0006,
+ "loss": 6.658792495727539,
+ "step": 293
+ },
+ {
+ "epoch": 4.083842794759826,
+ "grad_norm": 0.010519202798604965,
+ "learning_rate": 0.0006,
+ "loss": 6.662495136260986,
+ "step": 294
+ },
+ {
+ "epoch": 4.097816593886463,
+ "grad_norm": 0.007318088319152594,
+ "learning_rate": 0.0006,
+ "loss": 6.675954818725586,
+ "step": 295
+ },
+ {
+ "epoch": 4.1117903930131,
+ "grad_norm": 0.011712398380041122,
+ "learning_rate": 0.0006,
+ "loss": 6.673973083496094,
+ "step": 296
+ },
+ {
+ "epoch": 4.125764192139738,
+ "grad_norm": 0.006726635154336691,
+ "learning_rate": 0.0006,
+ "loss": 6.656833171844482,
+ "step": 297
+ },
+ {
+ "epoch": 4.139737991266376,
+ "grad_norm": 0.005466983187943697,
+ "learning_rate": 0.0006,
+ "loss": 6.675468921661377,
+ "step": 298
+ },
+ {
+ "epoch": 4.153711790393013,
+ "grad_norm": 0.007163621485233307,
+ "learning_rate": 0.0006,
+ "loss": 6.6862874031066895,
+ "step": 299
+ },
+ {
+ "epoch": 4.1676855895196505,
+ "grad_norm": 0.007626629434525967,
+ "learning_rate": 0.0006,
+ "loss": 6.654741287231445,
+ "step": 300
+ },
+ {
+ "epoch": 4.181659388646288,
+ "grad_norm": 0.0066330647096037865,
+ "learning_rate": 0.0006,
+ "loss": 6.68114709854126,
+ "step": 301
+ },
+ {
+ "epoch": 4.195633187772926,
+ "grad_norm": 0.007363482844084501,
+ "learning_rate": 0.0006,
+ "loss": 6.683980941772461,
+ "step": 302
+ },
+ {
+ "epoch": 4.209606986899563,
+ "grad_norm": 0.00971233006566763,
+ "learning_rate": 0.0006,
+ "loss": 6.669700622558594,
+ "step": 303
+ },
+ {
+ "epoch": 4.223580786026201,
+ "grad_norm": 0.013185256160795689,
+ "learning_rate": 0.0006,
+ "loss": 6.686490535736084,
+ "step": 304
+ },
+ {
+ "epoch": 4.2375545851528384,
+ "grad_norm": 0.019116027280688286,
+ "learning_rate": 0.0006,
+ "loss": 6.645479202270508,
+ "step": 305
+ },
+ {
+ "epoch": 4.251528384279476,
+ "grad_norm": 0.0328814722597599,
+ "learning_rate": 0.0006,
+ "loss": 6.68080997467041,
+ "step": 306
+ },
+ {
+ "epoch": 4.265502183406113,
+ "grad_norm": 0.04482287913560867,
+ "learning_rate": 0.0006,
+ "loss": 6.698016166687012,
+ "step": 307
+ },
+ {
+ "epoch": 4.279475982532751,
+ "grad_norm": 0.03053313121199608,
+ "learning_rate": 0.0006,
+ "loss": 6.642240524291992,
+ "step": 308
+ },
+ {
+ "epoch": 4.293449781659389,
+ "grad_norm": 0.043351784348487854,
+ "learning_rate": 0.0006,
+ "loss": 6.662343978881836,
+ "step": 309
+ },
+ {
+ "epoch": 4.307423580786026,
+ "grad_norm": 0.03165886923670769,
+ "learning_rate": 0.0006,
+ "loss": 6.6291961669921875,
+ "step": 310
+ },
+ {
+ "epoch": 4.321397379912664,
+ "grad_norm": 0.028601787984371185,
+ "learning_rate": 0.0006,
+ "loss": 6.667544364929199,
+ "step": 311
+ },
+ {
+ "epoch": 4.335371179039301,
+ "grad_norm": 0.019558480009436607,
+ "learning_rate": 0.0006,
+ "loss": 6.641025543212891,
+ "step": 312
+ },
+ {
+ "epoch": 4.349344978165939,
+ "grad_norm": 0.039851363748311996,
+ "learning_rate": 0.0006,
+ "loss": 6.665952205657959,
+ "step": 313
+ },
+ {
+ "epoch": 4.3633187772925766,
+ "grad_norm": 0.018417146056890488,
+ "learning_rate": 0.0006,
+ "loss": 6.657257556915283,
+ "step": 314
+ },
+ {
+ "epoch": 4.377292576419214,
+ "grad_norm": 0.02894527092576027,
+ "learning_rate": 0.0006,
+ "loss": 6.656288146972656,
+ "step": 315
+ },
+ {
+ "epoch": 4.391266375545851,
+ "grad_norm": 0.01910966821014881,
+ "learning_rate": 0.0006,
+ "loss": 6.688958168029785,
+ "step": 316
+ },
+ {
+ "epoch": 4.405240174672489,
+ "grad_norm": 0.01668412610888481,
+ "learning_rate": 0.0006,
+ "loss": 6.64876127243042,
+ "step": 317
+ },
+ {
+ "epoch": 4.419213973799127,
+ "grad_norm": 0.02935275249183178,
+ "learning_rate": 0.0006,
+ "loss": 6.626096725463867,
+ "step": 318
+ },
+ {
+ "epoch": 4.4331877729257645,
+ "grad_norm": 0.017997972667217255,
+ "learning_rate": 0.0006,
+ "loss": 6.640386581420898,
+ "step": 319
+ },
+ {
+ "epoch": 4.447161572052401,
+ "grad_norm": 0.01779189519584179,
+ "learning_rate": 0.0006,
+ "loss": 6.641165256500244,
+ "step": 320
+ },
+ {
+ "epoch": 4.461135371179039,
+ "grad_norm": 0.013901712372899055,
+ "learning_rate": 0.0006,
+ "loss": 6.651627063751221,
+ "step": 321
+ },
+ {
+ "epoch": 4.475109170305677,
+ "grad_norm": 0.013211199082434177,
+ "learning_rate": 0.0006,
+ "loss": 6.630064964294434,
+ "step": 322
+ },
+ {
+ "epoch": 4.489082969432315,
+ "grad_norm": 0.02324107475578785,
+ "learning_rate": 0.0006,
+ "loss": 6.670443058013916,
+ "step": 323
+ },
+ {
+ "epoch": 4.503056768558952,
+ "grad_norm": 0.019452504813671112,
+ "learning_rate": 0.0006,
+ "loss": 6.651266098022461,
+ "step": 324
+ },
+ {
+ "epoch": 4.517030567685589,
+ "grad_norm": 0.017943572252988815,
+ "learning_rate": 0.0006,
+ "loss": 6.657943248748779,
+ "step": 325
+ },
+ {
+ "epoch": 4.531004366812227,
+ "grad_norm": 0.012136237695813179,
+ "learning_rate": 0.0006,
+ "loss": 6.640082359313965,
+ "step": 326
+ },
+ {
+ "epoch": 4.544978165938865,
+ "grad_norm": 0.011544011533260345,
+ "learning_rate": 0.0006,
+ "loss": 6.6186933517456055,
+ "step": 327
+ },
+ {
+ "epoch": 4.558951965065502,
+ "grad_norm": 0.01569426618516445,
+ "learning_rate": 0.0006,
+ "loss": 6.640767574310303,
+ "step": 328
+ },
+ {
+ "epoch": 4.5729257641921395,
+ "grad_norm": 0.027668394148349762,
+ "learning_rate": 0.0006,
+ "loss": 6.651412487030029,
+ "step": 329
+ },
+ {
+ "epoch": 4.586899563318777,
+ "grad_norm": 0.04073230177164078,
+ "learning_rate": 0.0006,
+ "loss": 6.636565208435059,
+ "step": 330
+ },
+ {
+ "epoch": 4.600873362445415,
+ "grad_norm": 0.04942672327160835,
+ "learning_rate": 0.0006,
+ "loss": 6.667994499206543,
+ "step": 331
+ },
+ {
+ "epoch": 4.614847161572053,
+ "grad_norm": 0.037568598985672,
+ "learning_rate": 0.0006,
+ "loss": 6.632008075714111,
+ "step": 332
+ },
+ {
+ "epoch": 4.62882096069869,
+ "grad_norm": 0.009880400262773037,
+ "learning_rate": 0.0006,
+ "loss": 6.6504058837890625,
+ "step": 333
+ },
+ {
+ "epoch": 4.642794759825327,
+ "grad_norm": 0.033384259790182114,
+ "learning_rate": 0.0006,
+ "loss": 6.63852596282959,
+ "step": 334
+ },
+ {
+ "epoch": 4.656768558951965,
+ "grad_norm": 0.035640884190797806,
+ "learning_rate": 0.0006,
+ "loss": 6.627986431121826,
+ "step": 335
+ },
+ {
+ "epoch": 4.670742358078603,
+ "grad_norm": 0.013637864030897617,
+ "learning_rate": 0.0006,
+ "loss": 6.624503135681152,
+ "step": 336
+ },
+ {
+ "epoch": 4.68471615720524,
+ "grad_norm": 0.025304675102233887,
+ "learning_rate": 0.0006,
+ "loss": 6.612077236175537,
+ "step": 337
+ },
+ {
+ "epoch": 4.698689956331878,
+ "grad_norm": 0.03377160802483559,
+ "learning_rate": 0.0006,
+ "loss": 6.608938217163086,
+ "step": 338
+ },
+ {
+ "epoch": 4.712663755458515,
+ "grad_norm": 0.019803153350949287,
+ "learning_rate": 0.0006,
+ "loss": 6.636343955993652,
+ "step": 339
+ },
+ {
+ "epoch": 4.726637554585153,
+ "grad_norm": 0.018928129225969315,
+ "learning_rate": 0.0006,
+ "loss": 6.636613845825195,
+ "step": 340
+ },
+ {
+ "epoch": 4.74061135371179,
+ "grad_norm": 0.023068996146321297,
+ "learning_rate": 0.0006,
+ "loss": 6.6258440017700195,
+ "step": 341
+ },
+ {
+ "epoch": 4.754585152838428,
+ "grad_norm": 0.016849957406520844,
+ "learning_rate": 0.0006,
+ "loss": 6.638483047485352,
+ "step": 342
+ },
+ {
+ "epoch": 4.7685589519650655,
+ "grad_norm": 0.021430544555187225,
+ "learning_rate": 0.0006,
+ "loss": 6.619649410247803,
+ "step": 343
+ },
+ {
+ "epoch": 4.782532751091703,
+ "grad_norm": 0.024053368717432022,
+ "learning_rate": 0.0006,
+ "loss": 6.639214515686035,
+ "step": 344
+ },
+ {
+ "epoch": 4.796506550218341,
+ "grad_norm": 0.011926693841814995,
+ "learning_rate": 0.0006,
+ "loss": 6.604836463928223,
+ "step": 345
+ },
+ {
+ "epoch": 4.810480349344978,
+ "grad_norm": 0.01961611397564411,
+ "learning_rate": 0.0006,
+ "loss": 6.624864101409912,
+ "step": 346
+ },
+ {
+ "epoch": 4.824454148471616,
+ "grad_norm": 0.027946067973971367,
+ "learning_rate": 0.0006,
+ "loss": 6.621284484863281,
+ "step": 347
+ },
+ {
+ "epoch": 4.8384279475982535,
+ "grad_norm": 0.02108592912554741,
+ "learning_rate": 0.0006,
+ "loss": 6.5982465744018555,
+ "step": 348
+ },
+ {
+ "epoch": 4.85240174672489,
+ "grad_norm": 0.017272260040044785,
+ "learning_rate": 0.0006,
+ "loss": 6.626142501831055,
+ "step": 349
+ },
+ {
+ "epoch": 4.866375545851528,
+ "grad_norm": 0.015654582530260086,
+ "learning_rate": 0.0006,
+ "loss": 6.621735572814941,
+ "step": 350
+ },
+ {
+ "epoch": 4.880349344978166,
+ "grad_norm": 0.015797540545463562,
+ "learning_rate": 0.0006,
+ "loss": 6.6016411781311035,
+ "step": 351
+ },
+ {
+ "epoch": 4.894323144104804,
+ "grad_norm": 0.028488904237747192,
+ "learning_rate": 0.0006,
+ "loss": 6.598413467407227,
+ "step": 352
+ },
+ {
+ "epoch": 4.908296943231441,
+ "grad_norm": 0.03411734476685524,
+ "learning_rate": 0.0006,
+ "loss": 6.597321033477783,
+ "step": 353
+ },
+ {
+ "epoch": 4.922270742358078,
+ "grad_norm": 0.027188263833522797,
+ "learning_rate": 0.0006,
+ "loss": 6.630616188049316,
+ "step": 354
+ },
+ {
+ "epoch": 4.936244541484716,
+ "grad_norm": 0.03411437198519707,
+ "learning_rate": 0.0006,
+ "loss": 6.57647705078125,
+ "step": 355
+ },
+ {
+ "epoch": 4.950218340611354,
+ "grad_norm": 0.019907327368855476,
+ "learning_rate": 0.0006,
+ "loss": 6.579545974731445,
+ "step": 356
+ },
+ {
+ "epoch": 4.964192139737992,
+ "grad_norm": 0.02485930174589157,
+ "learning_rate": 0.0006,
+ "loss": 6.591073036193848,
+ "step": 357
+ },
+ {
+ "epoch": 4.978165938864628,
+ "grad_norm": 0.026106052100658417,
+ "learning_rate": 0.0006,
+ "loss": 6.580729961395264,
+ "step": 358
+ },
+ {
+ "epoch": 4.992139737991266,
+ "grad_norm": 0.029401391744613647,
+ "learning_rate": 0.0006,
+ "loss": 6.5460076332092285,
+ "step": 359
+ },
+ {
+ "epoch": 5.0,
+ "grad_norm": 0.032869379967451096,
+ "learning_rate": 0.0006,
+ "loss": 6.583278179168701,
+ "step": 360
+ },
+ {
+ "epoch": 5.0,
+ "eval_loss": 7.416288375854492,
+ "eval_runtime": 60.4177,
+ "eval_samples_per_second": 40.419,
+ "eval_steps_per_second": 1.274,
+ "step": 360
+ },
+ {
+ "epoch": 5.013973799126638,
+ "grad_norm": 0.04862433671951294,
+ "learning_rate": 0.0006,
+ "loss": 6.58473014831543,
+ "step": 361
+ },
+ {
+ "epoch": 5.0279475982532755,
+ "grad_norm": 0.06636223942041397,
+ "learning_rate": 0.0006,
+ "loss": 6.621153354644775,
+ "step": 362
+ },
+ {
+ "epoch": 5.041921397379912,
+ "grad_norm": 0.05888264626264572,
+ "learning_rate": 0.0006,
+ "loss": 6.621196269989014,
+ "step": 363
+ },
+ {
+ "epoch": 5.05589519650655,
+ "grad_norm": 0.05331697314977646,
+ "learning_rate": 0.0006,
+ "loss": 6.600763320922852,
+ "step": 364
+ },
+ {
+ "epoch": 5.069868995633188,
+ "grad_norm": 0.03947671875357628,
+ "learning_rate": 0.0006,
+ "loss": 6.600671768188477,
+ "step": 365
+ },
+ {
+ "epoch": 5.083842794759826,
+ "grad_norm": 0.04045343026518822,
+ "learning_rate": 0.0006,
+ "loss": 6.63529109954834,
+ "step": 366
+ },
+ {
+ "epoch": 5.097816593886463,
+ "grad_norm": 0.029766036197543144,
+ "learning_rate": 0.0006,
+ "loss": 6.612868309020996,
+ "step": 367
+ },
+ {
+ "epoch": 5.1117903930131,
+ "grad_norm": 0.03416327014565468,
+ "learning_rate": 0.0006,
+ "loss": 6.582240581512451,
+ "step": 368
+ },
+ {
+ "epoch": 5.125764192139738,
+ "grad_norm": 0.03458515554666519,
+ "learning_rate": 0.0006,
+ "loss": 6.604381561279297,
+ "step": 369
+ },
+ {
+ "epoch": 5.139737991266376,
+ "grad_norm": 0.024099089205265045,
+ "learning_rate": 0.0006,
+ "loss": 6.597355365753174,
+ "step": 370
+ },
+ {
+ "epoch": 5.153711790393013,
+ "grad_norm": 0.02694438211619854,
+ "learning_rate": 0.0006,
+ "loss": 6.597512245178223,
+ "step": 371
+ },
+ {
+ "epoch": 5.1676855895196505,
+ "grad_norm": 0.024201856926083565,
+ "learning_rate": 0.0006,
+ "loss": 6.564685821533203,
+ "step": 372
+ },
+ {
+ "epoch": 5.181659388646288,
+ "grad_norm": 0.015959380194544792,
+ "learning_rate": 0.0006,
+ "loss": 6.582263946533203,
+ "step": 373
+ },
+ {
+ "epoch": 5.195633187772926,
+ "grad_norm": 0.025690171867609024,
+ "learning_rate": 0.0006,
+ "loss": 6.601057052612305,
+ "step": 374
+ },
+ {
+ "epoch": 5.209606986899563,
+ "grad_norm": 0.014497052878141403,
+ "learning_rate": 0.0006,
+ "loss": 6.584796905517578,
+ "step": 375
+ },
+ {
+ "epoch": 5.223580786026201,
+ "grad_norm": 0.019870297983288765,
+ "learning_rate": 0.0006,
+ "loss": 6.574010372161865,
+ "step": 376
+ },
+ {
+ "epoch": 5.2375545851528384,
+ "grad_norm": 0.01807512901723385,
+ "learning_rate": 0.0006,
+ "loss": 6.558596134185791,
+ "step": 377
+ },
+ {
+ "epoch": 5.251528384279476,
+ "grad_norm": 0.014525193721055984,
+ "learning_rate": 0.0006,
+ "loss": 6.571803092956543,
+ "step": 378
+ },
+ {
+ "epoch": 5.265502183406113,
+ "grad_norm": 0.016113435849547386,
+ "learning_rate": 0.0006,
+ "loss": 6.568970680236816,
+ "step": 379
+ },
+ {
+ "epoch": 5.279475982532751,
+ "grad_norm": 0.020724017173051834,
+ "learning_rate": 0.0006,
+ "loss": 6.5669403076171875,
+ "step": 380
+ },
+ {
+ "epoch": 5.293449781659389,
+ "grad_norm": 0.013963675126433372,
+ "learning_rate": 0.0006,
+ "loss": 6.586382865905762,
+ "step": 381
+ },
+ {
+ "epoch": 5.307423580786026,
+ "grad_norm": 0.01621115580201149,
+ "learning_rate": 0.0006,
+ "loss": 6.539546966552734,
+ "step": 382
+ },
+ {
+ "epoch": 5.321397379912664,
+ "grad_norm": 0.014925207942724228,
+ "learning_rate": 0.0006,
+ "loss": 6.531795501708984,
+ "step": 383
+ },
+ {
+ "epoch": 5.335371179039301,
+ "grad_norm": 0.015454350039362907,
+ "learning_rate": 0.0006,
+ "loss": 6.575699806213379,
+ "step": 384
+ },
+ {
+ "epoch": 5.349344978165939,
+ "grad_norm": 0.01623542234301567,
+ "learning_rate": 0.0006,
+ "loss": 6.582373142242432,
+ "step": 385
+ },
+ {
+ "epoch": 5.3633187772925766,
+ "grad_norm": 0.015452183783054352,
+ "learning_rate": 0.0006,
+ "loss": 6.576451301574707,
+ "step": 386
+ },
+ {
+ "epoch": 5.377292576419214,
+ "grad_norm": 0.021102432161569595,
+ "learning_rate": 0.0006,
+ "loss": 6.543055534362793,
+ "step": 387
+ },
+ {
+ "epoch": 5.391266375545851,
+ "grad_norm": 0.009827060624957085,
+ "learning_rate": 0.0006,
+ "loss": 6.526939392089844,
+ "step": 388
+ },
+ {
+ "epoch": 5.405240174672489,
+ "grad_norm": 0.013735136948525906,
+ "learning_rate": 0.0006,
+ "loss": 6.538825988769531,
+ "step": 389
+ },
+ {
+ "epoch": 5.419213973799127,
+ "grad_norm": 0.02073509432375431,
+ "learning_rate": 0.0006,
+ "loss": 6.521315574645996,
+ "step": 390
+ },
+ {
+ "epoch": 5.4331877729257645,
+ "grad_norm": 0.03110959194600582,
+ "learning_rate": 0.0006,
+ "loss": 6.561807632446289,
+ "step": 391
+ },
+ {
+ "epoch": 5.447161572052401,
+ "grad_norm": 0.045058365911245346,
+ "learning_rate": 0.0006,
+ "loss": 6.526515960693359,
+ "step": 392
+ },
+ {
+ "epoch": 5.461135371179039,
+ "grad_norm": 0.06359834969043732,
+ "learning_rate": 0.0006,
+ "loss": 6.554903030395508,
+ "step": 393
+ },
+ {
+ "epoch": 5.475109170305677,
+ "grad_norm": 0.06406767666339874,
+ "learning_rate": 0.0006,
+ "loss": 6.576272010803223,
+ "step": 394
+ },
+ {
+ "epoch": 5.489082969432315,
+ "grad_norm": 0.03403699770569801,
+ "learning_rate": 0.0006,
+ "loss": 6.5306243896484375,
+ "step": 395
+ },
+ {
+ "epoch": 5.503056768558952,
+ "grad_norm": 0.036109618842601776,
+ "learning_rate": 0.0006,
+ "loss": 6.54176139831543,
+ "step": 396
+ },
+ {
+ "epoch": 5.517030567685589,
+ "grad_norm": 0.03478201478719711,
+ "learning_rate": 0.0006,
+ "loss": 6.530734062194824,
+ "step": 397
+ },
+ {
+ "epoch": 5.531004366812227,
+ "grad_norm": 0.03401544317603111,
+ "learning_rate": 0.0006,
+ "loss": 6.557794094085693,
+ "step": 398
+ },
+ {
+ "epoch": 5.544978165938865,
+ "grad_norm": 0.031629662960767746,
+ "learning_rate": 0.0006,
+ "loss": 6.564887046813965,
+ "step": 399
+ },
+ {
+ "epoch": 5.558951965065502,
+ "grad_norm": 0.02305024117231369,
+ "learning_rate": 0.0006,
+ "loss": 6.547412872314453,
+ "step": 400
+ },
+ {
+ "epoch": 5.5729257641921395,
+ "grad_norm": 0.030852800235152245,
+ "learning_rate": 0.0006,
+ "loss": 6.495370388031006,
+ "step": 401
+ },
+ {
+ "epoch": 5.586899563318777,
+ "grad_norm": 0.03368550166487694,
+ "learning_rate": 0.0006,
+ "loss": 6.495980262756348,
+ "step": 402
+ },
+ {
+ "epoch": 5.600873362445415,
+ "grad_norm": 0.020976359024643898,
+ "learning_rate": 0.0006,
+ "loss": 6.509410858154297,
+ "step": 403
+ },
+ {
+ "epoch": 5.614847161572053,
+ "grad_norm": 0.019652143120765686,
+ "learning_rate": 0.0006,
+ "loss": 6.493232727050781,
+ "step": 404
+ },
+ {
+ "epoch": 5.62882096069869,
+ "grad_norm": 0.024320121854543686,
+ "learning_rate": 0.0006,
+ "loss": 6.521531581878662,
+ "step": 405
+ },
+ {
+ "epoch": 5.642794759825327,
+ "grad_norm": 0.016611427068710327,
+ "learning_rate": 0.0006,
+ "loss": 6.507271766662598,
+ "step": 406
+ },
+ {
+ "epoch": 5.656768558951965,
+ "grad_norm": 0.01635856367647648,
+ "learning_rate": 0.0006,
+ "loss": 6.492038726806641,
+ "step": 407
+ },
+ {
+ "epoch": 5.670742358078603,
+ "grad_norm": 0.017642095685005188,
+ "learning_rate": 0.0006,
+ "loss": 6.500349998474121,
+ "step": 408
+ },
+ {
+ "epoch": 5.68471615720524,
+ "grad_norm": 0.01679178513586521,
+ "learning_rate": 0.0006,
+ "loss": 6.500361442565918,
+ "step": 409
+ },
+ {
+ "epoch": 5.698689956331878,
+ "grad_norm": 0.011776307597756386,
+ "learning_rate": 0.0006,
+ "loss": 6.482490539550781,
+ "step": 410
+ },
+ {
+ "epoch": 5.712663755458515,
+ "grad_norm": 0.0176856629550457,
+ "learning_rate": 0.0006,
+ "loss": 6.48488712310791,
+ "step": 411
+ },
+ {
+ "epoch": 5.726637554585153,
+ "grad_norm": 0.015193155966699123,
+ "learning_rate": 0.0006,
+ "loss": 6.459545612335205,
+ "step": 412
+ },
+ {
+ "epoch": 5.74061135371179,
+ "grad_norm": 0.018623212352395058,
+ "learning_rate": 0.0006,
+ "loss": 6.478870391845703,
+ "step": 413
+ },
+ {
+ "epoch": 5.754585152838428,
+ "grad_norm": 0.012841006740927696,
+ "learning_rate": 0.0006,
+ "loss": 6.474418640136719,
+ "step": 414
+ },
+ {
+ "epoch": 5.7685589519650655,
+ "grad_norm": 0.015540511347353458,
+ "learning_rate": 0.0006,
+ "loss": 6.4706525802612305,
+ "step": 415
+ },
+ {
+ "epoch": 5.782532751091703,
+ "grad_norm": 0.022695740684866905,
+ "learning_rate": 0.0006,
+ "loss": 6.455738067626953,
+ "step": 416
+ },
+ {
+ "epoch": 5.796506550218341,
+ "grad_norm": 0.026930129155516624,
+ "learning_rate": 0.0006,
+ "loss": 6.458016872406006,
+ "step": 417
+ },
+ {
+ "epoch": 5.810480349344978,
+ "grad_norm": 0.03555797412991524,
+ "learning_rate": 0.0006,
+ "loss": 6.480044364929199,
+ "step": 418
+ },
+ {
+ "epoch": 5.824454148471616,
+ "grad_norm": 0.05024682730436325,
+ "learning_rate": 0.0006,
+ "loss": 6.454564571380615,
+ "step": 419
+ },
+ {
+ "epoch": 5.8384279475982535,
+ "grad_norm": 0.06616255640983582,
+ "learning_rate": 0.0006,
+ "loss": 6.433927536010742,
+ "step": 420
+ },
+ {
+ "epoch": 5.85240174672489,
+ "grad_norm": 0.059065770357847214,
+ "learning_rate": 0.0006,
+ "loss": 6.508362770080566,
+ "step": 421
+ },
+ {
+ "epoch": 5.866375545851528,
+ "grad_norm": 0.026689983904361725,
+ "learning_rate": 0.0006,
+ "loss": 6.464297294616699,
+ "step": 422
+ },
+ {
+ "epoch": 5.880349344978166,
+ "grad_norm": 0.04616398364305496,
+ "learning_rate": 0.0006,
+ "loss": 6.45448112487793,
+ "step": 423
+ },
+ {
+ "epoch": 5.894323144104804,
+ "grad_norm": 0.04980356618762016,
+ "learning_rate": 0.0006,
+ "loss": 6.4696574211120605,
+ "step": 424
+ },
+ {
+ "epoch": 5.908296943231441,
+ "grad_norm": 0.0551983006298542,
+ "learning_rate": 0.0006,
+ "loss": 6.419000625610352,
+ "step": 425
+ },
+ {
+ "epoch": 5.922270742358078,
+ "grad_norm": 0.06988243758678436,
+ "learning_rate": 0.0006,
+ "loss": 6.442890167236328,
+ "step": 426
+ },
+ {
+ "epoch": 5.936244541484716,
+ "grad_norm": 0.06040226295590401,
+ "learning_rate": 0.0006,
+ "loss": 6.513258934020996,
+ "step": 427
+ },
+ {
+ "epoch": 5.950218340611354,
+ "grad_norm": 0.04635024070739746,
+ "learning_rate": 0.0006,
+ "loss": 6.4710283279418945,
+ "step": 428
+ },
+ {
+ "epoch": 5.964192139737992,
+ "grad_norm": 0.04762702435255051,
+ "learning_rate": 0.0006,
+ "loss": 6.469327926635742,
+ "step": 429
+ },
+ {
+ "epoch": 5.978165938864628,
+ "grad_norm": 0.04857536032795906,
+ "learning_rate": 0.0006,
+ "loss": 6.426105499267578,
+ "step": 430
+ },
+ {
+ "epoch": 5.992139737991266,
+ "grad_norm": 0.04412693530321121,
+ "learning_rate": 0.0006,
+ "loss": 6.43179988861084,
+ "step": 431
+ },
+ {
+ "epoch": 6.0,
+ "grad_norm": 0.026129672303795815,
+ "learning_rate": 0.0006,
+ "loss": 6.437918663024902,
+ "step": 432
+ },
+ {
+ "epoch": 6.0,
+ "eval_loss": 7.219249248504639,
+ "eval_runtime": 58.6837,
+ "eval_samples_per_second": 41.613,
+ "eval_steps_per_second": 1.312,
+ "step": 432
+ },
+ {
+ "epoch": 6.013973799126638,
+ "grad_norm": 0.03329874947667122,
+ "learning_rate": 0.0006,
+ "loss": 6.438179969787598,
+ "step": 433
+ },
+ {
+ "epoch": 6.0279475982532755,
+ "grad_norm": 0.03723187372088432,
+ "learning_rate": 0.0006,
+ "loss": 6.394242286682129,
+ "step": 434
+ },
+ {
+ "epoch": 6.041921397379912,
+ "grad_norm": 0.014625726267695427,
+ "learning_rate": 0.0006,
+ "loss": 6.412118911743164,
+ "step": 435
+ },
+ {
+ "epoch": 6.05589519650655,
+ "grad_norm": 0.02917475439608097,
+ "learning_rate": 0.0006,
+ "loss": 6.453164577484131,
+ "step": 436
+ },
+ {
+ "epoch": 6.069868995633188,
+ "grad_norm": 0.01811208389699459,
+ "learning_rate": 0.0006,
+ "loss": 6.396003246307373,
+ "step": 437
+ },
+ {
+ "epoch": 6.083842794759826,
+ "grad_norm": 0.027235351502895355,
+ "learning_rate": 0.0006,
+ "loss": 6.402976036071777,
+ "step": 438
+ },
+ {
+ "epoch": 6.097816593886463,
+ "grad_norm": 0.02835794910788536,
+ "learning_rate": 0.0006,
+ "loss": 6.380568504333496,
+ "step": 439
+ },
+ {
+ "epoch": 6.1117903930131,
+ "grad_norm": 0.020868154242634773,
+ "learning_rate": 0.0006,
+ "loss": 6.38240909576416,
+ "step": 440
+ },
+ {
+ "epoch": 6.125764192139738,
+ "grad_norm": 0.02113114856183529,
+ "learning_rate": 0.0006,
+ "loss": 6.377391815185547,
+ "step": 441
+ },
+ {
+ "epoch": 6.139737991266376,
+ "grad_norm": 0.018894119188189507,
+ "learning_rate": 0.0006,
+ "loss": 6.354569435119629,
+ "step": 442
+ },
+ {
+ "epoch": 6.153711790393013,
+ "grad_norm": 0.01820225641131401,
+ "learning_rate": 0.0006,
+ "loss": 6.392882347106934,
+ "step": 443
+ },
+ {
+ "epoch": 6.1676855895196505,
+ "grad_norm": 0.017510728910565376,
+ "learning_rate": 0.0006,
+ "loss": 6.364162445068359,
+ "step": 444
+ },
+ {
+ "epoch": 6.181659388646288,
+ "grad_norm": 0.017893686890602112,
+ "learning_rate": 0.0006,
+ "loss": 6.373456954956055,
+ "step": 445
+ },
+ {
+ "epoch": 6.195633187772926,
+ "grad_norm": 0.013382102362811565,
+ "learning_rate": 0.0006,
+ "loss": 6.3714399337768555,
+ "step": 446
+ },
+ {
+ "epoch": 6.209606986899563,
+ "grad_norm": 0.022540833801031113,
+ "learning_rate": 0.0006,
+ "loss": 6.348170757293701,
+ "step": 447
+ },
+ {
+ "epoch": 6.223580786026201,
+ "grad_norm": 0.02110104262828827,
+ "learning_rate": 0.0006,
+ "loss": 6.413206100463867,
+ "step": 448
+ },
+ {
+ "epoch": 6.2375545851528384,
+ "grad_norm": 0.014498349279165268,
+ "learning_rate": 0.0006,
+ "loss": 6.361384391784668,
+ "step": 449
+ },
+ {
+ "epoch": 6.251528384279476,
+ "grad_norm": 0.01819637045264244,
+ "learning_rate": 0.0006,
+ "loss": 6.315275192260742,
+ "step": 450
+ },
+ {
+ "epoch": 6.265502183406113,
+ "grad_norm": 0.0195972491055727,
+ "learning_rate": 0.0006,
+ "loss": 6.321994781494141,
+ "step": 451
+ },
+ {
+ "epoch": 6.279475982532751,
+ "grad_norm": 0.018822453916072845,
+ "learning_rate": 0.0006,
+ "loss": 6.345219612121582,
+ "step": 452
+ },
+ {
+ "epoch": 6.293449781659389,
+ "grad_norm": 0.019247712567448616,
+ "learning_rate": 0.0006,
+ "loss": 6.348073959350586,
+ "step": 453
+ },
+ {
+ "epoch": 6.307423580786026,
+ "grad_norm": 0.03175375238060951,
+ "learning_rate": 0.0006,
+ "loss": 6.300344944000244,
+ "step": 454
+ },
+ {
+ "epoch": 6.321397379912664,
+ "grad_norm": 0.06326444447040558,
+ "learning_rate": 0.0006,
+ "loss": 6.392579078674316,
+ "step": 455
+ },
+ {
+ "epoch": 6.335371179039301,
+ "grad_norm": 0.09667232632637024,
+ "learning_rate": 0.0006,
+ "loss": 6.36024284362793,
+ "step": 456
+ },
+ {
+ "epoch": 6.349344978165939,
+ "grad_norm": 0.05774446204304695,
+ "learning_rate": 0.0006,
+ "loss": 6.365338325500488,
+ "step": 457
+ },
+ {
+ "epoch": 6.3633187772925766,
+ "grad_norm": 0.08793356269598007,
+ "learning_rate": 0.0006,
+ "loss": 6.43284273147583,
+ "step": 458
+ },
+ {
+ "epoch": 6.377292576419214,
+ "grad_norm": 0.0640595331788063,
+ "learning_rate": 0.0006,
+ "loss": 6.342428207397461,
+ "step": 459
+ },
+ {
+ "epoch": 6.391266375545851,
+ "grad_norm": 0.0447838194668293,
+ "learning_rate": 0.0006,
+ "loss": 6.367971897125244,
+ "step": 460
+ },
+ {
+ "epoch": 6.405240174672489,
+ "grad_norm": 0.04768967255949974,
+ "learning_rate": 0.0006,
+ "loss": 6.35027551651001,
+ "step": 461
+ },
+ {
+ "epoch": 6.419213973799127,
+ "grad_norm": 0.027981851249933243,
+ "learning_rate": 0.0006,
+ "loss": 6.33723258972168,
+ "step": 462
+ },
+ {
+ "epoch": 6.4331877729257645,
+ "grad_norm": 0.03479623422026634,
+ "learning_rate": 0.0006,
+ "loss": 6.36590051651001,
+ "step": 463
+ },
+ {
+ "epoch": 6.447161572052401,
+ "grad_norm": 0.020846687257289886,
+ "learning_rate": 0.0006,
+ "loss": 6.315143585205078,
+ "step": 464
+ },
+ {
+ "epoch": 6.461135371179039,
+ "grad_norm": 0.0312928780913353,
+ "learning_rate": 0.0006,
+ "loss": 6.340150833129883,
+ "step": 465
+ },
+ {
+ "epoch": 6.475109170305677,
+ "grad_norm": 0.020593881607055664,
+ "learning_rate": 0.0006,
+ "loss": 6.348175048828125,
+ "step": 466
+ },
+ {
+ "epoch": 6.489082969432315,
+ "grad_norm": 0.022010236978530884,
+ "learning_rate": 0.0006,
+ "loss": 6.318607330322266,
+ "step": 467
+ },
+ {
+ "epoch": 6.503056768558952,
+ "grad_norm": 0.018427059054374695,
+ "learning_rate": 0.0006,
+ "loss": 6.331822395324707,
+ "step": 468
+ },
+ {
+ "epoch": 6.517030567685589,
+ "grad_norm": 0.019964968785643578,
+ "learning_rate": 0.0006,
+ "loss": 6.287320137023926,
+ "step": 469
+ },
+ {
+ "epoch": 6.531004366812227,
+ "grad_norm": 0.02130270190536976,
+ "learning_rate": 0.0006,
+ "loss": 6.353344917297363,
+ "step": 470
+ },
+ {
+ "epoch": 6.544978165938865,
+ "grad_norm": 0.01616514101624489,
+ "learning_rate": 0.0006,
+ "loss": 6.314052581787109,
+ "step": 471
+ },
+ {
+ "epoch": 6.558951965065502,
+ "grad_norm": 0.016639474779367447,
+ "learning_rate": 0.0006,
+ "loss": 6.320351600646973,
+ "step": 472
+ },
+ {
+ "epoch": 6.5729257641921395,
+ "grad_norm": 0.017620692029595375,
+ "learning_rate": 0.0006,
+ "loss": 6.311459064483643,
+ "step": 473
+ },
+ {
+ "epoch": 6.586899563318777,
+ "grad_norm": 0.013542844913899899,
+ "learning_rate": 0.0006,
+ "loss": 6.324368476867676,
+ "step": 474
+ },
+ {
+ "epoch": 6.600873362445415,
+ "grad_norm": 0.01308678649365902,
+ "learning_rate": 0.0006,
+ "loss": 6.307419776916504,
+ "step": 475
+ },
+ {
+ "epoch": 6.614847161572053,
+ "grad_norm": 0.014672392047941685,
+ "learning_rate": 0.0006,
+ "loss": 6.2785420417785645,
+ "step": 476
+ },
+ {
+ "epoch": 6.62882096069869,
+ "grad_norm": 0.01399585697799921,
+ "learning_rate": 0.0006,
+ "loss": 6.258619785308838,
+ "step": 477
+ },
+ {
+ "epoch": 6.642794759825327,
+ "grad_norm": 0.01330904383212328,
+ "learning_rate": 0.0006,
+ "loss": 6.299595832824707,
+ "step": 478
+ },
+ {
+ "epoch": 6.656768558951965,
+ "grad_norm": 0.016613377258181572,
+ "learning_rate": 0.0006,
+ "loss": 6.2900309562683105,
+ "step": 479
+ },
+ {
+ "epoch": 6.670742358078603,
+ "grad_norm": 0.015076331794261932,
+ "learning_rate": 0.0006,
+ "loss": 6.231028079986572,
+ "step": 480
+ },
+ {
+ "epoch": 6.68471615720524,
+ "grad_norm": 0.01589728519320488,
+ "learning_rate": 0.0006,
+ "loss": 6.257922172546387,
+ "step": 481
+ },
+ {
+ "epoch": 6.698689956331878,
+ "grad_norm": 0.030972328037023544,
+ "learning_rate": 0.0006,
+ "loss": 6.285238265991211,
+ "step": 482
+ },
+ {
+ "epoch": 6.712663755458515,
+ "grad_norm": 0.05048783868551254,
+ "learning_rate": 0.0006,
+ "loss": 6.287471771240234,
+ "step": 483
+ },
+ {
+ "epoch": 6.726637554585153,
+ "grad_norm": 0.07390714436769485,
+ "learning_rate": 0.0006,
+ "loss": 6.284293174743652,
+ "step": 484
+ },
+ {
+ "epoch": 6.74061135371179,
+ "grad_norm": 0.07166943699121475,
+ "learning_rate": 0.0006,
+ "loss": 6.290704727172852,
+ "step": 485
+ },
+ {
+ "epoch": 6.754585152838428,
+ "grad_norm": 0.028792452067136765,
+ "learning_rate": 0.0006,
+ "loss": 6.243415832519531,
+ "step": 486
+ },
+ {
+ "epoch": 6.7685589519650655,
+ "grad_norm": 0.04822041466832161,
+ "learning_rate": 0.0006,
+ "loss": 6.278229713439941,
+ "step": 487
+ },
+ {
+ "epoch": 6.782532751091703,
+ "grad_norm": 0.036688923835754395,
+ "learning_rate": 0.0006,
+ "loss": 6.275590896606445,
+ "step": 488
+ },
+ {
+ "epoch": 6.796506550218341,
+ "grad_norm": 0.03679114952683449,
+ "learning_rate": 0.0006,
+ "loss": 6.236509323120117,
+ "step": 489
+ },
+ {
+ "epoch": 6.810480349344978,
+ "grad_norm": 0.023249102756381035,
+ "learning_rate": 0.0006,
+ "loss": 6.230788230895996,
+ "step": 490
+ },
+ {
+ "epoch": 6.824454148471616,
+ "grad_norm": 0.02682286500930786,
+ "learning_rate": 0.0006,
+ "loss": 6.278890609741211,
+ "step": 491
+ },
+ {
+ "epoch": 6.8384279475982535,
+ "grad_norm": 0.02202044241130352,
+ "learning_rate": 0.0006,
+ "loss": 6.234531402587891,
+ "step": 492
+ },
+ {
+ "epoch": 6.85240174672489,
+ "grad_norm": 0.02545459009706974,
+ "learning_rate": 0.0006,
+ "loss": 6.271790981292725,
+ "step": 493
+ },
+ {
+ "epoch": 6.866375545851528,
+ "grad_norm": 0.02012932486832142,
+ "learning_rate": 0.0006,
+ "loss": 6.247462749481201,
+ "step": 494
+ },
+ {
+ "epoch": 6.880349344978166,
+ "grad_norm": 0.02216474711894989,
+ "learning_rate": 0.0006,
+ "loss": 6.26396369934082,
+ "step": 495
+ },
+ {
+ "epoch": 6.894323144104804,
+ "grad_norm": 0.016078704968094826,
+ "learning_rate": 0.0006,
+ "loss": 6.228609085083008,
+ "step": 496
+ },
+ {
+ "epoch": 6.908296943231441,
+ "grad_norm": 0.023021478205919266,
+ "learning_rate": 0.0006,
+ "loss": 6.197244167327881,
+ "step": 497
+ },
+ {
+ "epoch": 6.922270742358078,
+ "grad_norm": 0.024766333401203156,
+ "learning_rate": 0.0006,
+ "loss": 6.181090354919434,
+ "step": 498
+ },
+ {
+ "epoch": 6.936244541484716,
+ "grad_norm": 0.025581466034054756,
+ "learning_rate": 0.0006,
+ "loss": 6.201817512512207,
+ "step": 499
+ },
+ {
+ "epoch": 6.950218340611354,
+ "grad_norm": 0.016964981332421303,
+ "learning_rate": 0.0006,
+ "loss": 6.201486587524414,
+ "step": 500
+ },
+ {
+ "epoch": 6.964192139737992,
+ "grad_norm": 0.023957721889019012,
+ "learning_rate": 0.0006,
+ "loss": 6.185701847076416,
+ "step": 501
+ },
+ {
+ "epoch": 6.978165938864628,
+ "grad_norm": 0.029775245115160942,
+ "learning_rate": 0.0006,
+ "loss": 6.242060661315918,
+ "step": 502
+ },
+ {
+ "epoch": 6.992139737991266,
+ "grad_norm": 0.04007191210985184,
+ "learning_rate": 0.0006,
+ "loss": 6.19931697845459,
+ "step": 503
+ },
+ {
+ "epoch": 7.0,
+ "grad_norm": 0.04882276430726051,
+ "learning_rate": 0.0006,
+ "loss": 6.1880598068237305,
+ "step": 504
+ },
+ {
+ "epoch": 7.0,
+ "eval_loss": 7.018043518066406,
+ "eval_runtime": 59.8815,
+ "eval_samples_per_second": 40.781,
+ "eval_steps_per_second": 1.286,
+ "step": 504
+ },
+ {
+ "epoch": 7.013973799126638,
+ "grad_norm": 0.07144106924533844,
+ "learning_rate": 0.0006,
+ "loss": 6.172799110412598,
+ "step": 505
+ },
+ {
+ "epoch": 7.0279475982532755,
+ "grad_norm": 0.06712841987609863,
+ "learning_rate": 0.0006,
+ "loss": 6.228682518005371,
+ "step": 506
+ },
+ {
+ "epoch": 7.041921397379912,
+ "grad_norm": 0.04405239224433899,
+ "learning_rate": 0.0006,
+ "loss": 6.1827898025512695,
+ "step": 507
+ },
+ {
+ "epoch": 7.05589519650655,
+ "grad_norm": 0.03418995440006256,
+ "learning_rate": 0.0006,
+ "loss": 6.200876235961914,
+ "step": 508
+ },
+ {
+ "epoch": 7.069868995633188,
+ "grad_norm": 0.04002382606267929,
+ "learning_rate": 0.0006,
+ "loss": 6.1799211502075195,
+ "step": 509
+ },
+ {
+ "epoch": 7.083842794759826,
+ "grad_norm": 0.029748762026429176,
+ "learning_rate": 0.0006,
+ "loss": 6.125422477722168,
+ "step": 510
+ },
+ {
+ "epoch": 7.097816593886463,
+ "grad_norm": 0.03408357501029968,
+ "learning_rate": 0.0006,
+ "loss": 6.203458309173584,
+ "step": 511
+ },
+ {
+ "epoch": 7.1117903930131,
+ "grad_norm": 0.02227279357612133,
+ "learning_rate": 0.0006,
+ "loss": 6.135439872741699,
+ "step": 512
+ },
+ {
+ "epoch": 7.125764192139738,
+ "grad_norm": 0.02727862447500229,
+ "learning_rate": 0.0006,
+ "loss": 6.202380180358887,
+ "step": 513
+ },
+ {
+ "epoch": 7.139737991266376,
+ "grad_norm": 0.026019075885415077,
+ "learning_rate": 0.0006,
+ "loss": 6.201343536376953,
+ "step": 514
+ },
+ {
+ "epoch": 7.153711790393013,
+ "grad_norm": 0.027227282524108887,
+ "learning_rate": 0.0006,
+ "loss": 6.164059638977051,
+ "step": 515
+ },
+ {
+ "epoch": 7.1676855895196505,
+ "grad_norm": 0.027649980038404465,
+ "learning_rate": 0.0006,
+ "loss": 6.135448932647705,
+ "step": 516
+ },
+ {
+ "epoch": 7.181659388646288,
+ "grad_norm": 0.02168039232492447,
+ "learning_rate": 0.0006,
+ "loss": 6.154831886291504,
+ "step": 517
+ },
+ {
+ "epoch": 7.195633187772926,
+ "grad_norm": 0.02101467363536358,
+ "learning_rate": 0.0006,
+ "loss": 6.128695487976074,
+ "step": 518
+ },
+ {
+ "epoch": 7.209606986899563,
+ "grad_norm": 0.018344245851039886,
+ "learning_rate": 0.0006,
+ "loss": 6.136524200439453,
+ "step": 519
+ },
+ {
+ "epoch": 7.223580786026201,
+ "grad_norm": 0.020323453471064568,
+ "learning_rate": 0.0006,
+ "loss": 6.1206488609313965,
+ "step": 520
+ },
+ {
+ "epoch": 7.2375545851528384,
+ "grad_norm": 0.01750849559903145,
+ "learning_rate": 0.0006,
+ "loss": 6.128170967102051,
+ "step": 521
+ },
+ {
+ "epoch": 7.251528384279476,
+ "grad_norm": 0.017778221517801285,
+ "learning_rate": 0.0006,
+ "loss": 6.150458812713623,
+ "step": 522
+ },
+ {
+ "epoch": 7.265502183406113,
+ "grad_norm": 0.01863938942551613,
+ "learning_rate": 0.0006,
+ "loss": 6.100106716156006,
+ "step": 523
+ },
+ {
+ "epoch": 7.279475982532751,
+ "grad_norm": 0.020196080207824707,
+ "learning_rate": 0.0006,
+ "loss": 6.149346351623535,
+ "step": 524
+ },
+ {
+ "epoch": 7.293449781659389,
+ "grad_norm": 0.03203890845179558,
+ "learning_rate": 0.0006,
+ "loss": 6.124648094177246,
+ "step": 525
+ },
+ {
+ "epoch": 7.307423580786026,
+ "grad_norm": 0.05325170233845711,
+ "learning_rate": 0.0006,
+ "loss": 6.115618705749512,
+ "step": 526
+ },
+ {
+ "epoch": 7.321397379912664,
+ "grad_norm": 0.05888345465064049,
+ "learning_rate": 0.0006,
+ "loss": 6.165221214294434,
+ "step": 527
+ },
+ {
+ "epoch": 7.335371179039301,
+ "grad_norm": 0.026873115450143814,
+ "learning_rate": 0.0006,
+ "loss": 6.039590835571289,
+ "step": 528
+ },
+ {
+ "epoch": 7.349344978165939,
+ "grad_norm": 0.06304649263620377,
+ "learning_rate": 0.0006,
+ "loss": 6.131900787353516,
+ "step": 529
+ },
+ {
+ "epoch": 7.3633187772925766,
+ "grad_norm": 0.08567557483911514,
+ "learning_rate": 0.0006,
+ "loss": 6.218114852905273,
+ "step": 530
+ },
+ {
+ "epoch": 7.377292576419214,
+ "grad_norm": 0.08868218213319778,
+ "learning_rate": 0.0006,
+ "loss": 6.216706275939941,
+ "step": 531
+ },
+ {
+ "epoch": 7.391266375545851,
+ "grad_norm": 0.03955601155757904,
+ "learning_rate": 0.0006,
+ "loss": 6.1566925048828125,
+ "step": 532
+ },
+ {
+ "epoch": 7.405240174672489,
+ "grad_norm": 0.04979498311877251,
+ "learning_rate": 0.0006,
+ "loss": 6.19327449798584,
+ "step": 533
+ },
+ {
+ "epoch": 7.419213973799127,
+ "grad_norm": 0.03253026679158211,
+ "learning_rate": 0.0006,
+ "loss": 6.133993148803711,
+ "step": 534
+ },
+ {
+ "epoch": 7.4331877729257645,
+ "grad_norm": 0.03016652725636959,
+ "learning_rate": 0.0006,
+ "loss": 6.174962997436523,
+ "step": 535
+ },
+ {
+ "epoch": 7.447161572052401,
+ "grad_norm": 0.03159862384200096,
+ "learning_rate": 0.0006,
+ "loss": 6.082124710083008,
+ "step": 536
+ },
+ {
+ "epoch": 7.461135371179039,
+ "grad_norm": 0.027673959732055664,
+ "learning_rate": 0.0006,
+ "loss": 6.111873149871826,
+ "step": 537
+ },
+ {
+ "epoch": 7.475109170305677,
+ "grad_norm": 0.022789714857935905,
+ "learning_rate": 0.0006,
+ "loss": 6.129437446594238,
+ "step": 538
+ },
+ {
+ "epoch": 7.489082969432315,
+ "grad_norm": 0.030061012133955956,
+ "learning_rate": 0.0006,
+ "loss": 6.126370429992676,
+ "step": 539
+ },
+ {
+ "epoch": 7.503056768558952,
+ "grad_norm": 0.019007068127393723,
+ "learning_rate": 0.0006,
+ "loss": 6.07887077331543,
+ "step": 540
+ },
+ {
+ "epoch": 7.517030567685589,
+ "grad_norm": 0.01897653006017208,
+ "learning_rate": 0.0006,
+ "loss": 6.096936225891113,
+ "step": 541
+ },
+ {
+ "epoch": 7.531004366812227,
+ "grad_norm": 0.024659698829054832,
+ "learning_rate": 0.0006,
+ "loss": 6.0933380126953125,
+ "step": 542
+ },
+ {
+ "epoch": 7.544978165938865,
+ "grad_norm": 0.021924549713730812,
+ "learning_rate": 0.0006,
+ "loss": 6.105311870574951,
+ "step": 543
+ },
+ {
+ "epoch": 7.558951965065502,
+ "grad_norm": 0.018817558884620667,
+ "learning_rate": 0.0006,
+ "loss": 6.058895111083984,
+ "step": 544
+ },
+ {
+ "epoch": 7.5729257641921395,
+ "grad_norm": 0.02477840520441532,
+ "learning_rate": 0.0006,
+ "loss": 6.038928031921387,
+ "step": 545
+ },
+ {
+ "epoch": 7.586899563318777,
+ "grad_norm": 0.018968632444739342,
+ "learning_rate": 0.0006,
+ "loss": 6.1368937492370605,
+ "step": 546
+ },
+ {
+ "epoch": 7.600873362445415,
+ "grad_norm": 0.02409801445901394,
+ "learning_rate": 0.0006,
+ "loss": 6.110324859619141,
+ "step": 547
+ },
+ {
+ "epoch": 7.614847161572053,
+ "grad_norm": 0.017595326527953148,
+ "learning_rate": 0.0006,
+ "loss": 6.074368476867676,
+ "step": 548
+ },
+ {
+ "epoch": 7.62882096069869,
+ "grad_norm": 0.01875406503677368,
+ "learning_rate": 0.0006,
+ "loss": 6.121023654937744,
+ "step": 549
+ },
+ {
+ "epoch": 7.642794759825327,
+ "grad_norm": 0.018781628459692,
+ "learning_rate": 0.0006,
+ "loss": 6.064914703369141,
+ "step": 550
+ },
+ {
+ "epoch": 7.656768558951965,
+ "grad_norm": 0.017003126442432404,
+ "learning_rate": 0.0006,
+ "loss": 6.039608001708984,
+ "step": 551
+ },
+ {
+ "epoch": 7.670742358078603,
+ "grad_norm": 0.017665037885308266,
+ "learning_rate": 0.0006,
+ "loss": 6.053671836853027,
+ "step": 552
+ },
+ {
+ "epoch": 7.68471615720524,
+ "grad_norm": 0.02066265605390072,
+ "learning_rate": 0.0006,
+ "loss": 6.090536117553711,
+ "step": 553
+ },
+ {
+ "epoch": 7.698689956331878,
+ "grad_norm": 0.02888769656419754,
+ "learning_rate": 0.0006,
+ "loss": 5.9858903884887695,
+ "step": 554
+ },
+ {
+ "epoch": 7.712663755458515,
+ "grad_norm": 0.03744485229253769,
+ "learning_rate": 0.0006,
+ "loss": 6.086264610290527,
+ "step": 555
+ },
+ {
+ "epoch": 7.726637554585153,
+ "grad_norm": 0.04719431325793266,
+ "learning_rate": 0.0006,
+ "loss": 6.09747314453125,
+ "step": 556
+ },
+ {
+ "epoch": 7.74061135371179,
+ "grad_norm": 0.044963907450437546,
+ "learning_rate": 0.0006,
+ "loss": 6.091272354125977,
+ "step": 557
+ },
+ {
+ "epoch": 7.754585152838428,
+ "grad_norm": 0.03810194507241249,
+ "learning_rate": 0.0006,
+ "loss": 6.084904193878174,
+ "step": 558
+ },
+ {
+ "epoch": 7.7685589519650655,
+ "grad_norm": 0.04626769199967384,
+ "learning_rate": 0.0006,
+ "loss": 6.017384052276611,
+ "step": 559
+ },
+ {
+ "epoch": 7.782532751091703,
+ "grad_norm": 0.03459552302956581,
+ "learning_rate": 0.0006,
+ "loss": 6.111314296722412,
+ "step": 560
+ },
+ {
+ "epoch": 7.796506550218341,
+ "grad_norm": 0.027385666966438293,
+ "learning_rate": 0.0006,
+ "loss": 6.002509117126465,
+ "step": 561
+ },
+ {
+ "epoch": 7.810480349344978,
+ "grad_norm": 0.03417652100324631,
+ "learning_rate": 0.0006,
+ "loss": 6.0421061515808105,
+ "step": 562
+ },
+ {
+ "epoch": 7.824454148471616,
+ "grad_norm": 0.04415450245141983,
+ "learning_rate": 0.0006,
+ "loss": 6.035305023193359,
+ "step": 563
+ },
+ {
+ "epoch": 7.8384279475982535,
+ "grad_norm": 0.03259635716676712,
+ "learning_rate": 0.0006,
+ "loss": 6.025840759277344,
+ "step": 564
+ },
+ {
+ "epoch": 7.85240174672489,
+ "grad_norm": 0.024941129609942436,
+ "learning_rate": 0.0006,
+ "loss": 6.043416976928711,
+ "step": 565
+ },
+ {
+ "epoch": 7.866375545851528,
+ "grad_norm": 0.02434500865638256,
+ "learning_rate": 0.0006,
+ "loss": 6.002052307128906,
+ "step": 566
+ },
+ {
+ "epoch": 7.880349344978166,
+ "grad_norm": 0.01876584254205227,
+ "learning_rate": 0.0006,
+ "loss": 6.045035362243652,
+ "step": 567
+ },
+ {
+ "epoch": 7.894323144104804,
+ "grad_norm": 0.02254367619752884,
+ "learning_rate": 0.0006,
+ "loss": 5.972503662109375,
+ "step": 568
+ },
+ {
+ "epoch": 7.908296943231441,
+ "grad_norm": 0.024504603818058968,
+ "learning_rate": 0.0006,
+ "loss": 5.992846488952637,
+ "step": 569
+ },
+ {
+ "epoch": 7.922270742358078,
+ "grad_norm": 0.02317827194929123,
+ "learning_rate": 0.0006,
+ "loss": 6.007697582244873,
+ "step": 570
+ },
+ {
+ "epoch": 7.936244541484716,
+ "grad_norm": 0.02225821278989315,
+ "learning_rate": 0.0006,
+ "loss": 6.01463508605957,
+ "step": 571
+ },
+ {
+ "epoch": 7.950218340611354,
+ "grad_norm": 0.01811717264354229,
+ "learning_rate": 0.0006,
+ "loss": 5.995022773742676,
+ "step": 572
+ },
+ {
+ "epoch": 7.964192139737992,
+ "grad_norm": 0.019589431583881378,
+ "learning_rate": 0.0006,
+ "loss": 5.990633964538574,
+ "step": 573
+ },
+ {
+ "epoch": 7.978165938864628,
+ "grad_norm": 0.015422016382217407,
+ "learning_rate": 0.0006,
+ "loss": 6.011263847351074,
+ "step": 574
+ },
+ {
+ "epoch": 7.992139737991266,
+ "grad_norm": 0.023298880085349083,
+ "learning_rate": 0.0006,
+ "loss": 5.9954938888549805,
+ "step": 575
+ },
+ {
+ "epoch": 8.0,
+ "grad_norm": 0.022172044962644577,
+ "learning_rate": 0.0006,
+ "loss": 5.95679235458374,
+ "step": 576
+ },
+ {
+ "epoch": 8.0,
+ "eval_loss": 6.75809383392334,
+ "eval_runtime": 72.8926,
+ "eval_samples_per_second": 33.501,
+ "eval_steps_per_second": 1.056,
+ "step": 576
+ },
+ {
+ "epoch": 8.013973799126637,
+ "grad_norm": 0.029082221910357475,
+ "learning_rate": 0.0006,
+ "loss": 6.018675804138184,
+ "step": 577
+ },
+ {
+ "epoch": 8.027947598253276,
+ "grad_norm": 0.03959495946764946,
+ "learning_rate": 0.0006,
+ "loss": 5.93310546875,
+ "step": 578
+ },
+ {
+ "epoch": 8.041921397379912,
+ "grad_norm": 0.043756261467933655,
+ "learning_rate": 0.0006,
+ "loss": 5.965944290161133,
+ "step": 579
+ },
+ {
+ "epoch": 8.055895196506551,
+ "grad_norm": 0.04996631667017937,
+ "learning_rate": 0.0006,
+ "loss": 6.0381059646606445,
+ "step": 580
+ },
+ {
+ "epoch": 8.069868995633188,
+ "grad_norm": 0.05721999332308769,
+ "learning_rate": 0.0006,
+ "loss": 5.968381881713867,
+ "step": 581
+ },
+ {
+ "epoch": 8.083842794759825,
+ "grad_norm": 0.07594525814056396,
+ "learning_rate": 0.0006,
+ "loss": 6.027225494384766,
+ "step": 582
+ },
+ {
+ "epoch": 8.097816593886463,
+ "grad_norm": 0.06861685961484909,
+ "learning_rate": 0.0006,
+ "loss": 6.035175323486328,
+ "step": 583
+ },
+ {
+ "epoch": 8.1117903930131,
+ "grad_norm": 0.04156472906470299,
+ "learning_rate": 0.0006,
+ "loss": 5.9500017166137695,
+ "step": 584
+ },
+ {
+ "epoch": 8.125764192139737,
+ "grad_norm": 0.045399412512779236,
+ "learning_rate": 0.0006,
+ "loss": 5.9953389167785645,
+ "step": 585
+ },
+ {
+ "epoch": 8.139737991266376,
+ "grad_norm": 0.04399585723876953,
+ "learning_rate": 0.0006,
+ "loss": 5.986051559448242,
+ "step": 586
+ },
+ {
+ "epoch": 8.153711790393013,
+ "grad_norm": 0.05764509737491608,
+ "learning_rate": 0.0006,
+ "loss": 6.005392074584961,
+ "step": 587
+ },
+ {
+ "epoch": 8.167685589519651,
+ "grad_norm": 0.060871850699186325,
+ "learning_rate": 0.0006,
+ "loss": 6.01304817199707,
+ "step": 588
+ },
+ {
+ "epoch": 8.181659388646288,
+ "grad_norm": 0.03381241112947464,
+ "learning_rate": 0.0006,
+ "loss": 5.977019309997559,
+ "step": 589
+ },
+ {
+ "epoch": 8.195633187772925,
+ "grad_norm": 0.03384104371070862,
+ "learning_rate": 0.0006,
+ "loss": 5.988002777099609,
+ "step": 590
+ },
+ {
+ "epoch": 8.209606986899564,
+ "grad_norm": 0.03120286948978901,
+ "learning_rate": 0.0006,
+ "loss": 5.968227386474609,
+ "step": 591
+ },
+ {
+ "epoch": 8.2235807860262,
+ "grad_norm": 0.026532653719186783,
+ "learning_rate": 0.0006,
+ "loss": 5.936810493469238,
+ "step": 592
+ },
+ {
+ "epoch": 8.237554585152838,
+ "grad_norm": 0.024956922978162766,
+ "learning_rate": 0.0006,
+ "loss": 5.9348015785217285,
+ "step": 593
+ },
+ {
+ "epoch": 8.251528384279476,
+ "grad_norm": 0.025926144793629646,
+ "learning_rate": 0.0006,
+ "loss": 5.952630519866943,
+ "step": 594
+ },
+ {
+ "epoch": 8.265502183406113,
+ "grad_norm": 0.02893793396651745,
+ "learning_rate": 0.0006,
+ "loss": 5.9569854736328125,
+ "step": 595
+ },
+ {
+ "epoch": 8.279475982532752,
+ "grad_norm": 0.0320001058280468,
+ "learning_rate": 0.0006,
+ "loss": 5.917050838470459,
+ "step": 596
+ },
+ {
+ "epoch": 8.293449781659389,
+ "grad_norm": 0.03925463184714317,
+ "learning_rate": 0.0006,
+ "loss": 5.9925127029418945,
+ "step": 597
+ },
+ {
+ "epoch": 8.307423580786025,
+ "grad_norm": 0.05096640810370445,
+ "learning_rate": 0.0006,
+ "loss": 5.926358222961426,
+ "step": 598
+ },
+ {
+ "epoch": 8.321397379912664,
+ "grad_norm": 0.041637953370809555,
+ "learning_rate": 0.0006,
+ "loss": 5.954983711242676,
+ "step": 599
+ },
+ {
+ "epoch": 8.335371179039301,
+ "grad_norm": 0.03138125315308571,
+ "learning_rate": 0.0006,
+ "loss": 5.893692970275879,
+ "step": 600
+ },
+ {
+ "epoch": 8.34934497816594,
+ "grad_norm": 0.029299303889274597,
+ "learning_rate": 0.0006,
+ "loss": 5.94670295715332,
+ "step": 601
+ },
+ {
+ "epoch": 8.363318777292577,
+ "grad_norm": 0.023969192057847977,
+ "learning_rate": 0.0006,
+ "loss": 5.925658702850342,
+ "step": 602
+ },
+ {
+ "epoch": 8.377292576419213,
+ "grad_norm": 0.025406932458281517,
+ "learning_rate": 0.0006,
+ "loss": 5.966875076293945,
+ "step": 603
+ },
+ {
+ "epoch": 8.391266375545852,
+ "grad_norm": 0.024309059605002403,
+ "learning_rate": 0.0006,
+ "loss": 5.932581424713135,
+ "step": 604
+ },
+ {
+ "epoch": 8.405240174672489,
+ "grad_norm": 0.017237074673175812,
+ "learning_rate": 0.0006,
+ "loss": 5.898823261260986,
+ "step": 605
+ },
+ {
+ "epoch": 8.419213973799126,
+ "grad_norm": 0.025594931095838547,
+ "learning_rate": 0.0006,
+ "loss": 5.9285993576049805,
+ "step": 606
+ },
+ {
+ "epoch": 8.433187772925764,
+ "grad_norm": 0.025184160098433495,
+ "learning_rate": 0.0006,
+ "loss": 5.915966510772705,
+ "step": 607
+ },
+ {
+ "epoch": 8.447161572052401,
+ "grad_norm": 0.019004398956894875,
+ "learning_rate": 0.0006,
+ "loss": 5.893395900726318,
+ "step": 608
+ },
+ {
+ "epoch": 8.46113537117904,
+ "grad_norm": 0.021802261471748352,
+ "learning_rate": 0.0006,
+ "loss": 5.953311443328857,
+ "step": 609
+ },
+ {
+ "epoch": 8.475109170305677,
+ "grad_norm": 0.02202655002474785,
+ "learning_rate": 0.0006,
+ "loss": 5.953000545501709,
+ "step": 610
+ },
+ {
+ "epoch": 8.489082969432314,
+ "grad_norm": 0.02067241445183754,
+ "learning_rate": 0.0006,
+ "loss": 5.869247913360596,
+ "step": 611
+ },
+ {
+ "epoch": 8.503056768558952,
+ "grad_norm": 0.015430927276611328,
+ "learning_rate": 0.0006,
+ "loss": 5.855883598327637,
+ "step": 612
+ },
+ {
+ "epoch": 8.51703056768559,
+ "grad_norm": 0.019917506724596024,
+ "learning_rate": 0.0006,
+ "loss": 5.9284539222717285,
+ "step": 613
+ },
+ {
+ "epoch": 8.531004366812226,
+ "grad_norm": 0.024291323497891426,
+ "learning_rate": 0.0006,
+ "loss": 5.973604202270508,
+ "step": 614
+ },
+ {
+ "epoch": 8.544978165938865,
+ "grad_norm": 0.022133296355605125,
+ "learning_rate": 0.0006,
+ "loss": 5.85693359375,
+ "step": 615
+ },
+ {
+ "epoch": 8.558951965065502,
+ "grad_norm": 0.0254366435110569,
+ "learning_rate": 0.0006,
+ "loss": 5.939558982849121,
+ "step": 616
+ },
+ {
+ "epoch": 8.57292576419214,
+ "grad_norm": 0.030134694650769234,
+ "learning_rate": 0.0006,
+ "loss": 5.892992973327637,
+ "step": 617
+ },
+ {
+ "epoch": 8.586899563318777,
+ "grad_norm": 0.03845192864537239,
+ "learning_rate": 0.0006,
+ "loss": 5.89303731918335,
+ "step": 618
+ },
+ {
+ "epoch": 8.600873362445414,
+ "grad_norm": 0.042409464716911316,
+ "learning_rate": 0.0006,
+ "loss": 5.914068222045898,
+ "step": 619
+ },
+ {
+ "epoch": 8.614847161572053,
+ "grad_norm": 0.0454271174967289,
+ "learning_rate": 0.0006,
+ "loss": 5.853579044342041,
+ "step": 620
+ },
+ {
+ "epoch": 8.62882096069869,
+ "grad_norm": 0.04934931918978691,
+ "learning_rate": 0.0006,
+ "loss": 5.881861686706543,
+ "step": 621
+ },
+ {
+ "epoch": 8.642794759825328,
+ "grad_norm": 0.0519695058465004,
+ "learning_rate": 0.0006,
+ "loss": 5.872620582580566,
+ "step": 622
+ },
+ {
+ "epoch": 8.656768558951965,
+ "grad_norm": 0.06172563508152962,
+ "learning_rate": 0.0006,
+ "loss": 5.938283443450928,
+ "step": 623
+ },
+ {
+ "epoch": 8.670742358078602,
+ "grad_norm": 0.08416368812322617,
+ "learning_rate": 0.0006,
+ "loss": 5.943403720855713,
+ "step": 624
+ },
+ {
+ "epoch": 8.68471615720524,
+ "grad_norm": 0.07311230152845383,
+ "learning_rate": 0.0006,
+ "loss": 5.954029083251953,
+ "step": 625
+ },
+ {
+ "epoch": 8.698689956331878,
+ "grad_norm": 0.09851036965847015,
+ "learning_rate": 0.0006,
+ "loss": 5.992746829986572,
+ "step": 626
+ },
+ {
+ "epoch": 8.712663755458514,
+ "grad_norm": 0.07521767914295197,
+ "learning_rate": 0.0006,
+ "loss": 5.956884384155273,
+ "step": 627
+ },
+ {
+ "epoch": 8.726637554585153,
+ "grad_norm": 0.05720978602766991,
+ "learning_rate": 0.0006,
+ "loss": 5.911205291748047,
+ "step": 628
+ },
+ {
+ "epoch": 8.74061135371179,
+ "grad_norm": 0.05360936373472214,
+ "learning_rate": 0.0006,
+ "loss": 5.9734787940979,
+ "step": 629
+ },
+ {
+ "epoch": 8.754585152838429,
+ "grad_norm": 0.035826463252305984,
+ "learning_rate": 0.0006,
+ "loss": 5.95751953125,
+ "step": 630
+ },
+ {
+ "epoch": 8.768558951965066,
+ "grad_norm": 0.030137093737721443,
+ "learning_rate": 0.0006,
+ "loss": 5.91911506652832,
+ "step": 631
+ },
+ {
+ "epoch": 8.782532751091702,
+ "grad_norm": 0.030039150267839432,
+ "learning_rate": 0.0006,
+ "loss": 5.910290718078613,
+ "step": 632
+ },
+ {
+ "epoch": 8.796506550218341,
+ "grad_norm": 0.03349437564611435,
+ "learning_rate": 0.0006,
+ "loss": 5.904623985290527,
+ "step": 633
+ },
+ {
+ "epoch": 8.810480349344978,
+ "grad_norm": 0.02211189828813076,
+ "learning_rate": 0.0006,
+ "loss": 5.909668445587158,
+ "step": 634
+ },
+ {
+ "epoch": 8.824454148471617,
+ "grad_norm": 0.022601254284381866,
+ "learning_rate": 0.0006,
+ "loss": 5.942885875701904,
+ "step": 635
+ },
+ {
+ "epoch": 8.838427947598253,
+ "grad_norm": 0.02197270467877388,
+ "learning_rate": 0.0006,
+ "loss": 5.867103576660156,
+ "step": 636
+ },
+ {
+ "epoch": 8.85240174672489,
+ "grad_norm": 0.022765018045902252,
+ "learning_rate": 0.0006,
+ "loss": 5.894528865814209,
+ "step": 637
+ },
+ {
+ "epoch": 8.866375545851529,
+ "grad_norm": 0.021181151270866394,
+ "learning_rate": 0.0006,
+ "loss": 5.896910667419434,
+ "step": 638
+ },
+ {
+ "epoch": 8.880349344978166,
+ "grad_norm": 0.01878993771970272,
+ "learning_rate": 0.0006,
+ "loss": 5.860502243041992,
+ "step": 639
+ },
+ {
+ "epoch": 8.894323144104803,
+ "grad_norm": 0.014679803512990475,
+ "learning_rate": 0.0006,
+ "loss": 5.8725266456604,
+ "step": 640
+ },
+ {
+ "epoch": 8.908296943231441,
+ "grad_norm": 0.017919695004820824,
+ "learning_rate": 0.0006,
+ "loss": 5.88902473449707,
+ "step": 641
+ },
+ {
+ "epoch": 8.922270742358078,
+ "grad_norm": 0.01437009871006012,
+ "learning_rate": 0.0006,
+ "loss": 5.868186950683594,
+ "step": 642
+ },
+ {
+ "epoch": 8.936244541484717,
+ "grad_norm": 0.014293290674686432,
+ "learning_rate": 0.0006,
+ "loss": 5.833364963531494,
+ "step": 643
+ },
+ {
+ "epoch": 8.950218340611354,
+ "grad_norm": 0.015440084971487522,
+ "learning_rate": 0.0006,
+ "loss": 5.860260009765625,
+ "step": 644
+ },
+ {
+ "epoch": 8.96419213973799,
+ "grad_norm": 0.014292514882981777,
+ "learning_rate": 0.0006,
+ "loss": 5.844939231872559,
+ "step": 645
+ },
+ {
+ "epoch": 8.97816593886463,
+ "grad_norm": 0.015702251344919205,
+ "learning_rate": 0.0006,
+ "loss": 5.825929164886475,
+ "step": 646
+ },
+ {
+ "epoch": 8.992139737991266,
+ "grad_norm": 0.013372798450291157,
+ "learning_rate": 0.0006,
+ "loss": 5.83062744140625,
+ "step": 647
+ },
+ {
+ "epoch": 9.0,
+ "grad_norm": 0.013794616796076298,
+ "learning_rate": 0.0006,
+ "loss": 5.7897820472717285,
+ "step": 648
+ },
+ {
+ "epoch": 9.0,
+ "eval_loss": 6.600830078125,
+ "eval_runtime": 71.4316,
+ "eval_samples_per_second": 34.187,
+ "eval_steps_per_second": 1.078,
+ "step": 648
+ },
+ {
+ "epoch": 9.013973799126637,
+ "grad_norm": 0.01262663770467043,
+ "learning_rate": 0.0006,
+ "loss": 5.802824020385742,
+ "step": 649
+ },
+ {
+ "epoch": 9.027947598253276,
+ "grad_norm": 0.016375157982110977,
+ "learning_rate": 0.0006,
+ "loss": 5.8018317222595215,
+ "step": 650
+ },
+ {
+ "epoch": 9.041921397379912,
+ "grad_norm": 0.010541851632297039,
+ "learning_rate": 0.0006,
+ "loss": 5.848208427429199,
+ "step": 651
+ },
+ {
+ "epoch": 9.055895196506551,
+ "grad_norm": 0.011726257391273975,
+ "learning_rate": 0.0006,
+ "loss": 5.771823883056641,
+ "step": 652
+ },
+ {
+ "epoch": 9.069868995633188,
+ "grad_norm": 0.01124380249530077,
+ "learning_rate": 0.0006,
+ "loss": 5.773855209350586,
+ "step": 653
+ },
+ {
+ "epoch": 9.083842794759825,
+ "grad_norm": 0.010801885277032852,
+ "learning_rate": 0.0006,
+ "loss": 5.847390651702881,
+ "step": 654
+ },
+ {
+ "epoch": 9.097816593886463,
+ "grad_norm": 0.013018375262618065,
+ "learning_rate": 0.0006,
+ "loss": 5.786786079406738,
+ "step": 655
+ },
+ {
+ "epoch": 9.1117903930131,
+ "grad_norm": 0.014233003370463848,
+ "learning_rate": 0.0006,
+ "loss": 5.7913713455200195,
+ "step": 656
+ },
+ {
+ "epoch": 9.125764192139737,
+ "grad_norm": 0.01465831883251667,
+ "learning_rate": 0.0006,
+ "loss": 5.821170330047607,
+ "step": 657
+ },
+ {
+ "epoch": 9.139737991266376,
+ "grad_norm": 0.017383430153131485,
+ "learning_rate": 0.0006,
+ "loss": 5.811469078063965,
+ "step": 658
+ },
+ {
+ "epoch": 9.153711790393013,
+ "grad_norm": 0.020328007638454437,
+ "learning_rate": 0.0006,
+ "loss": 5.718976020812988,
+ "step": 659
+ },
+ {
+ "epoch": 9.167685589519651,
+ "grad_norm": 0.01776740700006485,
+ "learning_rate": 0.0006,
+ "loss": 5.80593729019165,
+ "step": 660
+ },
+ {
+ "epoch": 9.181659388646288,
+ "grad_norm": 0.017713800072669983,
+ "learning_rate": 0.0006,
+ "loss": 5.744930267333984,
+ "step": 661
+ },
+ {
+ "epoch": 9.195633187772925,
+ "grad_norm": 0.028774067759513855,
+ "learning_rate": 0.0006,
+ "loss": 5.85262393951416,
+ "step": 662
+ },
+ {
+ "epoch": 9.209606986899564,
+ "grad_norm": 0.05153049901127815,
+ "learning_rate": 0.0006,
+ "loss": 5.739948272705078,
+ "step": 663
+ },
+ {
+ "epoch": 9.2235807860262,
+ "grad_norm": 0.06882233172655106,
+ "learning_rate": 0.0006,
+ "loss": 5.844270706176758,
+ "step": 664
+ },
+ {
+ "epoch": 9.237554585152838,
+ "grad_norm": 0.0420759879052639,
+ "learning_rate": 0.0006,
+ "loss": 5.79060173034668,
+ "step": 665
+ },
+ {
+ "epoch": 9.251528384279476,
+ "grad_norm": 0.05598323419690132,
+ "learning_rate": 0.0006,
+ "loss": 5.833735466003418,
+ "step": 666
+ },
+ {
+ "epoch": 9.265502183406113,
+ "grad_norm": 0.05011717975139618,
+ "learning_rate": 0.0006,
+ "loss": 5.840887069702148,
+ "step": 667
+ },
+ {
+ "epoch": 9.279475982532752,
+ "grad_norm": 0.040782492607831955,
+ "learning_rate": 0.0006,
+ "loss": 5.7905449867248535,
+ "step": 668
+ },
+ {
+ "epoch": 9.293449781659389,
+ "grad_norm": 0.030804185196757317,
+ "learning_rate": 0.0006,
+ "loss": 5.802691459655762,
+ "step": 669
+ },
+ {
+ "epoch": 9.307423580786025,
+ "grad_norm": 0.032001759856939316,
+ "learning_rate": 0.0006,
+ "loss": 5.726612091064453,
+ "step": 670
+ },
+ {
+ "epoch": 9.321397379912664,
+ "grad_norm": 0.0295492485165596,
+ "learning_rate": 0.0006,
+ "loss": 5.798826694488525,
+ "step": 671
+ },
+ {
+ "epoch": 9.335371179039301,
+ "grad_norm": 0.034406546503305435,
+ "learning_rate": 0.0006,
+ "loss": 5.720219135284424,
+ "step": 672
+ },
+ {
+ "epoch": 9.34934497816594,
+ "grad_norm": 0.02627669647336006,
+ "learning_rate": 0.0006,
+ "loss": 5.723702430725098,
+ "step": 673
+ },
+ {
+ "epoch": 9.363318777292577,
+ "grad_norm": 0.0282363910228014,
+ "learning_rate": 0.0006,
+ "loss": 5.818911552429199,
+ "step": 674
+ },
+ {
+ "epoch": 9.377292576419213,
+ "grad_norm": 0.017894433811306953,
+ "learning_rate": 0.0006,
+ "loss": 5.772168159484863,
+ "step": 675
+ },
+ {
+ "epoch": 9.391266375545852,
+ "grad_norm": 0.019017303362488747,
+ "learning_rate": 0.0006,
+ "loss": 5.774587631225586,
+ "step": 676
+ },
+ {
+ "epoch": 9.405240174672489,
+ "grad_norm": 0.022159283980727196,
+ "learning_rate": 0.0006,
+ "loss": 5.812951564788818,
+ "step": 677
+ },
+ {
+ "epoch": 9.419213973799126,
+ "grad_norm": 0.02198116108775139,
+ "learning_rate": 0.0006,
+ "loss": 5.808734893798828,
+ "step": 678
+ },
+ {
+ "epoch": 9.433187772925764,
+ "grad_norm": 0.017741188406944275,
+ "learning_rate": 0.0006,
+ "loss": 5.776142120361328,
+ "step": 679
+ },
+ {
+ "epoch": 9.447161572052401,
+ "grad_norm": 0.017197400331497192,
+ "learning_rate": 0.0006,
+ "loss": 5.7471208572387695,
+ "step": 680
+ },
+ {
+ "epoch": 9.46113537117904,
+ "grad_norm": 0.01752566732466221,
+ "learning_rate": 0.0006,
+ "loss": 5.780961990356445,
+ "step": 681
+ },
+ {
+ "epoch": 9.475109170305677,
+ "grad_norm": 0.017961280420422554,
+ "learning_rate": 0.0006,
+ "loss": 5.713686943054199,
+ "step": 682
+ },
+ {
+ "epoch": 9.489082969432314,
+ "grad_norm": 0.02227521874010563,
+ "learning_rate": 0.0006,
+ "loss": 5.717774868011475,
+ "step": 683
+ },
+ {
+ "epoch": 9.503056768558952,
+ "grad_norm": 0.024075504392385483,
+ "learning_rate": 0.0006,
+ "loss": 5.715015411376953,
+ "step": 684
+ },
+ {
+ "epoch": 9.51703056768559,
+ "grad_norm": 0.029195033013820648,
+ "learning_rate": 0.0006,
+ "loss": 5.659659385681152,
+ "step": 685
+ },
+ {
+ "epoch": 9.531004366812226,
+ "grad_norm": 0.02959120273590088,
+ "learning_rate": 0.0006,
+ "loss": 5.78981351852417,
+ "step": 686
+ },
+ {
+ "epoch": 9.544978165938865,
+ "grad_norm": 0.021854368969798088,
+ "learning_rate": 0.0006,
+ "loss": 5.750844955444336,
+ "step": 687
+ },
+ {
+ "epoch": 9.558951965065502,
+ "grad_norm": 0.017120081931352615,
+ "learning_rate": 0.0006,
+ "loss": 5.672736167907715,
+ "step": 688
+ },
+ {
+ "epoch": 9.57292576419214,
+ "grad_norm": 0.020712822675704956,
+ "learning_rate": 0.0006,
+ "loss": 5.708046913146973,
+ "step": 689
+ },
+ {
+ "epoch": 9.586899563318777,
+ "grad_norm": 0.019186170771718025,
+ "learning_rate": 0.0006,
+ "loss": 5.71742057800293,
+ "step": 690
+ },
+ {
+ "epoch": 9.600873362445414,
+ "grad_norm": 0.01985814794898033,
+ "learning_rate": 0.0006,
+ "loss": 5.735311508178711,
+ "step": 691
+ },
+ {
+ "epoch": 9.614847161572053,
+ "grad_norm": 0.01965234987437725,
+ "learning_rate": 0.0006,
+ "loss": 5.728727340698242,
+ "step": 692
+ },
+ {
+ "epoch": 9.62882096069869,
+ "grad_norm": 0.019827688112854958,
+ "learning_rate": 0.0006,
+ "loss": 5.743068695068359,
+ "step": 693
+ },
+ {
+ "epoch": 9.642794759825328,
+ "grad_norm": 0.025500580668449402,
+ "learning_rate": 0.0006,
+ "loss": 5.76585578918457,
+ "step": 694
+ },
+ {
+ "epoch": 9.656768558951965,
+ "grad_norm": 0.03701791539788246,
+ "learning_rate": 0.0006,
+ "loss": 5.728367805480957,
+ "step": 695
+ },
+ {
+ "epoch": 9.670742358078602,
+ "grad_norm": 0.04173659533262253,
+ "learning_rate": 0.0006,
+ "loss": 5.683089256286621,
+ "step": 696
+ },
+ {
+ "epoch": 9.68471615720524,
+ "grad_norm": 0.031390558928251266,
+ "learning_rate": 0.0006,
+ "loss": 5.740921974182129,
+ "step": 697
+ },
+ {
+ "epoch": 9.698689956331878,
+ "grad_norm": 0.03970158472657204,
+ "learning_rate": 0.0006,
+ "loss": 5.730632781982422,
+ "step": 698
+ },
+ {
+ "epoch": 9.712663755458514,
+ "grad_norm": 0.03865141049027443,
+ "learning_rate": 0.0006,
+ "loss": 5.686657905578613,
+ "step": 699
+ },
+ {
+ "epoch": 9.726637554585153,
+ "grad_norm": 0.04856107011437416,
+ "learning_rate": 0.0006,
+ "loss": 5.657513618469238,
+ "step": 700
+ },
+ {
+ "epoch": 9.74061135371179,
+ "grad_norm": 0.04016133397817612,
+ "learning_rate": 0.0006,
+ "loss": 5.754552841186523,
+ "step": 701
+ },
+ {
+ "epoch": 9.754585152838429,
+ "grad_norm": 0.03961458429694176,
+ "learning_rate": 0.0006,
+ "loss": 5.672088623046875,
+ "step": 702
+ },
+ {
+ "epoch": 9.768558951965066,
+ "grad_norm": 0.03403536602854729,
+ "learning_rate": 0.0006,
+ "loss": 5.736673355102539,
+ "step": 703
+ },
+ {
+ "epoch": 9.782532751091702,
+ "grad_norm": 0.03766333684325218,
+ "learning_rate": 0.0006,
+ "loss": 5.716219902038574,
+ "step": 704
+ },
+ {
+ "epoch": 9.796506550218341,
+ "grad_norm": 0.04085501655936241,
+ "learning_rate": 0.0006,
+ "loss": 5.678958415985107,
+ "step": 705
+ },
+ {
+ "epoch": 9.810480349344978,
+ "grad_norm": 0.036888182163238525,
+ "learning_rate": 0.0006,
+ "loss": 5.6990461349487305,
+ "step": 706
+ },
+ {
+ "epoch": 9.824454148471617,
+ "grad_norm": 0.0492791011929512,
+ "learning_rate": 0.0006,
+ "loss": 5.78053092956543,
+ "step": 707
+ },
+ {
+ "epoch": 9.838427947598253,
+ "grad_norm": 0.05572429671883583,
+ "learning_rate": 0.0006,
+ "loss": 5.7543535232543945,
+ "step": 708
+ },
+ {
+ "epoch": 9.85240174672489,
+ "grad_norm": 0.047000352293252945,
+ "learning_rate": 0.0006,
+ "loss": 5.7174272537231445,
+ "step": 709
+ },
+ {
+ "epoch": 9.866375545851529,
+ "grad_norm": 0.058798179030418396,
+ "learning_rate": 0.0006,
+ "loss": 5.793713569641113,
+ "step": 710
+ },
+ {
+ "epoch": 9.880349344978166,
+ "grad_norm": 0.046992287039756775,
+ "learning_rate": 0.0006,
+ "loss": 5.784457683563232,
+ "step": 711
+ },
+ {
+ "epoch": 9.894323144104803,
+ "grad_norm": 0.04382370039820671,
+ "learning_rate": 0.0006,
+ "loss": 5.793731212615967,
+ "step": 712
+ },
+ {
+ "epoch": 9.908296943231441,
+ "grad_norm": 0.040616173297166824,
+ "learning_rate": 0.0006,
+ "loss": 5.69599723815918,
+ "step": 713
+ },
+ {
+ "epoch": 9.922270742358078,
+ "grad_norm": 0.03446170687675476,
+ "learning_rate": 0.0006,
+ "loss": 5.694674968719482,
+ "step": 714
+ },
+ {
+ "epoch": 9.936244541484717,
+ "grad_norm": 0.029936550185084343,
+ "learning_rate": 0.0006,
+ "loss": 5.6704487800598145,
+ "step": 715
+ },
+ {
+ "epoch": 9.950218340611354,
+ "grad_norm": 0.02224753610789776,
+ "learning_rate": 0.0006,
+ "loss": 5.728894233703613,
+ "step": 716
+ },
+ {
+ "epoch": 9.96419213973799,
+ "grad_norm": 0.024346651509404182,
+ "learning_rate": 0.0006,
+ "loss": 5.772412300109863,
+ "step": 717
+ },
+ {
+ "epoch": 9.97816593886463,
+ "grad_norm": 0.022997068241238594,
+ "learning_rate": 0.0006,
+ "loss": 5.645310401916504,
+ "step": 718
+ },
+ {
+ "epoch": 9.992139737991266,
+ "grad_norm": 0.024918686598539352,
+ "learning_rate": 0.0006,
+ "loss": 5.671009063720703,
+ "step": 719
+ },
+ {
+ "epoch": 10.0,
+ "grad_norm": 0.020067067816853523,
+ "learning_rate": 0.0006,
+ "loss": 5.637441635131836,
+ "step": 720
+ },
+ {
+ "epoch": 10.0,
+ "eval_loss": 6.387807846069336,
+ "eval_runtime": 72.3372,
+ "eval_samples_per_second": 33.759,
+ "eval_steps_per_second": 1.064,
+ "step": 720
+ },
+ {
+ "epoch": 10.013973799126637,
+ "grad_norm": 0.02015046589076519,
+ "learning_rate": 0.0006,
+ "loss": 5.598219871520996,
+ "step": 721
+ },
+ {
+ "epoch": 10.027947598253276,
+ "grad_norm": 0.017398346215486526,
+ "learning_rate": 0.0006,
+ "loss": 5.66834831237793,
+ "step": 722
+ },
+ {
+ "epoch": 10.041921397379912,
+ "grad_norm": 0.016871167346835136,
+ "learning_rate": 0.0006,
+ "loss": 5.673916816711426,
+ "step": 723
+ },
+ {
+ "epoch": 10.055895196506551,
+ "grad_norm": 0.01695060543715954,
+ "learning_rate": 0.0006,
+ "loss": 5.646677017211914,
+ "step": 724
+ },
+ {
+ "epoch": 10.069868995633188,
+ "grad_norm": 0.019298404455184937,
+ "learning_rate": 0.0006,
+ "loss": 5.663064479827881,
+ "step": 725
+ },
+ {
+ "epoch": 10.083842794759825,
+ "grad_norm": 0.018699361011385918,
+ "learning_rate": 0.0006,
+ "loss": 5.615242958068848,
+ "step": 726
+ },
+ {
+ "epoch": 10.097816593886463,
+ "grad_norm": 0.019567854702472687,
+ "learning_rate": 0.0006,
+ "loss": 5.5521240234375,
+ "step": 727
+ },
+ {
+ "epoch": 10.1117903930131,
+ "grad_norm": 0.018347466364502907,
+ "learning_rate": 0.0006,
+ "loss": 5.5929107666015625,
+ "step": 728
+ },
+ {
+ "epoch": 10.125764192139737,
+ "grad_norm": 0.020350053906440735,
+ "learning_rate": 0.0006,
+ "loss": 5.6380109786987305,
+ "step": 729
+ },
+ {
+ "epoch": 10.139737991266376,
+ "grad_norm": 0.016790535300970078,
+ "learning_rate": 0.0006,
+ "loss": 5.62994384765625,
+ "step": 730
+ },
+ {
+ "epoch": 10.153711790393013,
+ "grad_norm": 0.01413649506866932,
+ "learning_rate": 0.0006,
+ "loss": 5.6301727294921875,
+ "step": 731
+ },
+ {
+ "epoch": 10.167685589519651,
+ "grad_norm": 0.013762339949607849,
+ "learning_rate": 0.0006,
+ "loss": 5.584138870239258,
+ "step": 732
+ },
+ {
+ "epoch": 10.181659388646288,
+ "grad_norm": 0.014790690504014492,
+ "learning_rate": 0.0006,
+ "loss": 5.666123390197754,
+ "step": 733
+ },
+ {
+ "epoch": 10.195633187772925,
+ "grad_norm": 0.015471262857317924,
+ "learning_rate": 0.0006,
+ "loss": 5.587042808532715,
+ "step": 734
+ },
+ {
+ "epoch": 10.209606986899564,
+ "grad_norm": 0.015804501250386238,
+ "learning_rate": 0.0006,
+ "loss": 5.627497673034668,
+ "step": 735
+ },
+ {
+ "epoch": 10.2235807860262,
+ "grad_norm": 0.017627960070967674,
+ "learning_rate": 0.0006,
+ "loss": 5.556890487670898,
+ "step": 736
+ },
+ {
+ "epoch": 10.237554585152838,
+ "grad_norm": 0.02074616774916649,
+ "learning_rate": 0.0006,
+ "loss": 5.586511135101318,
+ "step": 737
+ },
+ {
+ "epoch": 10.251528384279476,
+ "grad_norm": 0.026389935985207558,
+ "learning_rate": 0.0006,
+ "loss": 5.598520278930664,
+ "step": 738
+ },
+ {
+ "epoch": 10.265502183406113,
+ "grad_norm": 0.03736288473010063,
+ "learning_rate": 0.0006,
+ "loss": 5.575788497924805,
+ "step": 739
+ },
+ {
+ "epoch": 10.279475982532752,
+ "grad_norm": 0.03523678705096245,
+ "learning_rate": 0.0006,
+ "loss": 5.586930751800537,
+ "step": 740
+ },
+ {
+ "epoch": 10.293449781659389,
+ "grad_norm": 0.024266984313726425,
+ "learning_rate": 0.0006,
+ "loss": 5.596694469451904,
+ "step": 741
+ },
+ {
+ "epoch": 10.307423580786025,
+ "grad_norm": 0.033640433102846146,
+ "learning_rate": 0.0006,
+ "loss": 5.536456108093262,
+ "step": 742
+ },
+ {
+ "epoch": 10.321397379912664,
+ "grad_norm": 0.04689764603972435,
+ "learning_rate": 0.0006,
+ "loss": 5.61761999130249,
+ "step": 743
+ },
+ {
+ "epoch": 10.335371179039301,
+ "grad_norm": 0.04910685867071152,
+ "learning_rate": 0.0006,
+ "loss": 5.6298418045043945,
+ "step": 744
+ },
+ {
+ "epoch": 10.34934497816594,
+ "grad_norm": 0.03152550011873245,
+ "learning_rate": 0.0006,
+ "loss": 5.639697074890137,
+ "step": 745
+ },
+ {
+ "epoch": 10.363318777292577,
+ "grad_norm": 0.037753574550151825,
+ "learning_rate": 0.0006,
+ "loss": 5.697381019592285,
+ "step": 746
+ },
+ {
+ "epoch": 10.377292576419213,
+ "grad_norm": 0.04877670109272003,
+ "learning_rate": 0.0006,
+ "loss": 5.605184555053711,
+ "step": 747
+ },
+ {
+ "epoch": 10.391266375545852,
+ "grad_norm": 0.04228494316339493,
+ "learning_rate": 0.0006,
+ "loss": 5.624238967895508,
+ "step": 748
+ },
+ {
+ "epoch": 10.405240174672489,
+ "grad_norm": 0.03575790300965309,
+ "learning_rate": 0.0006,
+ "loss": 5.6037983894348145,
+ "step": 749
+ },
+ {
+ "epoch": 10.419213973799126,
+ "grad_norm": 0.036425549536943436,
+ "learning_rate": 0.0006,
+ "loss": 5.62905216217041,
+ "step": 750
+ },
+ {
+ "epoch": 10.433187772925764,
+ "grad_norm": 0.04001355543732643,
+ "learning_rate": 0.0006,
+ "loss": 5.581656455993652,
+ "step": 751
+ },
+ {
+ "epoch": 10.447161572052401,
+ "grad_norm": 0.05040939524769783,
+ "learning_rate": 0.0006,
+ "loss": 5.621358871459961,
+ "step": 752
+ },
+ {
+ "epoch": 10.46113537117904,
+ "grad_norm": 0.04140249267220497,
+ "learning_rate": 0.0006,
+ "loss": 5.668188095092773,
+ "step": 753
+ },
+ {
+ "epoch": 10.475109170305677,
+ "grad_norm": 0.033116403967142105,
+ "learning_rate": 0.0006,
+ "loss": 5.580477714538574,
+ "step": 754
+ },
+ {
+ "epoch": 10.489082969432314,
+ "grad_norm": 0.03864264488220215,
+ "learning_rate": 0.0006,
+ "loss": 5.539153099060059,
+ "step": 755
+ },
+ {
+ "epoch": 10.503056768558952,
+ "grad_norm": 0.03026588261127472,
+ "learning_rate": 0.0006,
+ "loss": 5.619814872741699,
+ "step": 756
+ },
+ {
+ "epoch": 10.51703056768559,
+ "grad_norm": 0.02381783165037632,
+ "learning_rate": 0.0006,
+ "loss": 5.64422607421875,
+ "step": 757
+ },
+ {
+ "epoch": 10.531004366812226,
+ "grad_norm": 0.02427830547094345,
+ "learning_rate": 0.0006,
+ "loss": 5.659873008728027,
+ "step": 758
+ },
+ {
+ "epoch": 10.544978165938865,
+ "grad_norm": 0.021709900349378586,
+ "learning_rate": 0.0006,
+ "loss": 5.590245723724365,
+ "step": 759
+ },
+ {
+ "epoch": 10.558951965065502,
+ "grad_norm": 0.022812914103269577,
+ "learning_rate": 0.0006,
+ "loss": 5.610434532165527,
+ "step": 760
+ },
+ {
+ "epoch": 10.57292576419214,
+ "grad_norm": 0.019250890240073204,
+ "learning_rate": 0.0006,
+ "loss": 5.588078498840332,
+ "step": 761
+ },
+ {
+ "epoch": 10.586899563318777,
+ "grad_norm": 0.021995501592755318,
+ "learning_rate": 0.0006,
+ "loss": 5.541754245758057,
+ "step": 762
+ },
+ {
+ "epoch": 10.600873362445414,
+ "grad_norm": 0.01477806642651558,
+ "learning_rate": 0.0006,
+ "loss": 5.549173355102539,
+ "step": 763
+ },
+ {
+ "epoch": 10.614847161572053,
+ "grad_norm": 0.016680821776390076,
+ "learning_rate": 0.0006,
+ "loss": 5.58221435546875,
+ "step": 764
+ },
+ {
+ "epoch": 10.62882096069869,
+ "grad_norm": 0.013848318718373775,
+ "learning_rate": 0.0006,
+ "loss": 5.5505876541137695,
+ "step": 765
+ },
+ {
+ "epoch": 10.642794759825328,
+ "grad_norm": 0.016387000679969788,
+ "learning_rate": 0.0006,
+ "loss": 5.570460319519043,
+ "step": 766
+ },
+ {
+ "epoch": 10.656768558951965,
+ "grad_norm": 0.016496233642101288,
+ "learning_rate": 0.0006,
+ "loss": 5.513058662414551,
+ "step": 767
+ },
+ {
+ "epoch": 10.670742358078602,
+ "grad_norm": 0.01675555482506752,
+ "learning_rate": 0.0006,
+ "loss": 5.559998989105225,
+ "step": 768
+ },
+ {
+ "epoch": 10.68471615720524,
+ "grad_norm": 0.016939565539360046,
+ "learning_rate": 0.0006,
+ "loss": 5.56727409362793,
+ "step": 769
+ },
+ {
+ "epoch": 10.698689956331878,
+ "grad_norm": 0.01679506152868271,
+ "learning_rate": 0.0006,
+ "loss": 5.603643417358398,
+ "step": 770
+ },
+ {
+ "epoch": 10.712663755458514,
+ "grad_norm": 0.01656349189579487,
+ "learning_rate": 0.0006,
+ "loss": 5.54292631149292,
+ "step": 771
+ },
+ {
+ "epoch": 10.726637554585153,
+ "grad_norm": 0.01975332945585251,
+ "learning_rate": 0.0006,
+ "loss": 5.5225677490234375,
+ "step": 772
+ },
+ {
+ "epoch": 10.74061135371179,
+ "grad_norm": 0.02329389750957489,
+ "learning_rate": 0.0006,
+ "loss": 5.537827491760254,
+ "step": 773
+ },
+ {
+ "epoch": 10.754585152838429,
+ "grad_norm": 0.026616891846060753,
+ "learning_rate": 0.0006,
+ "loss": 5.500011920928955,
+ "step": 774
+ },
+ {
+ "epoch": 10.768558951965066,
+ "grad_norm": 0.025511575862765312,
+ "learning_rate": 0.0006,
+ "loss": 5.5316619873046875,
+ "step": 775
+ },
+ {
+ "epoch": 10.782532751091702,
+ "grad_norm": 0.025009512901306152,
+ "learning_rate": 0.0006,
+ "loss": 5.468493938446045,
+ "step": 776
+ },
+ {
+ "epoch": 10.796506550218341,
+ "grad_norm": 0.02946239896118641,
+ "learning_rate": 0.0006,
+ "loss": 5.5116753578186035,
+ "step": 777
+ },
+ {
+ "epoch": 10.810480349344978,
+ "grad_norm": 0.036607250571250916,
+ "learning_rate": 0.0006,
+ "loss": 5.518636226654053,
+ "step": 778
+ },
+ {
+ "epoch": 10.824454148471617,
+ "grad_norm": 0.04347814992070198,
+ "learning_rate": 0.0006,
+ "loss": 5.572815895080566,
+ "step": 779
+ },
+ {
+ "epoch": 10.838427947598253,
+ "grad_norm": 0.034463658928871155,
+ "learning_rate": 0.0006,
+ "loss": 5.520634651184082,
+ "step": 780
+ },
+ {
+ "epoch": 10.85240174672489,
+ "grad_norm": 0.03486703336238861,
+ "learning_rate": 0.0006,
+ "loss": 5.545441150665283,
+ "step": 781
+ },
+ {
+ "epoch": 10.866375545851529,
+ "grad_norm": 0.04476739838719368,
+ "learning_rate": 0.0006,
+ "loss": 5.554450035095215,
+ "step": 782
+ },
+ {
+ "epoch": 10.880349344978166,
+ "grad_norm": 0.048745233565568924,
+ "learning_rate": 0.0006,
+ "loss": 5.554961204528809,
+ "step": 783
+ },
+ {
+ "epoch": 10.894323144104803,
+ "grad_norm": 0.04636877775192261,
+ "learning_rate": 0.0006,
+ "loss": 5.574655055999756,
+ "step": 784
+ },
+ {
+ "epoch": 10.908296943231441,
+ "grad_norm": 0.032686877995729446,
+ "learning_rate": 0.0006,
+ "loss": 5.5734639167785645,
+ "step": 785
+ },
+ {
+ "epoch": 10.922270742358078,
+ "grad_norm": 0.03075302764773369,
+ "learning_rate": 0.0006,
+ "loss": 5.487662315368652,
+ "step": 786
+ },
+ {
+ "epoch": 10.936244541484717,
+ "grad_norm": 0.028860175982117653,
+ "learning_rate": 0.0006,
+ "loss": 5.519798755645752,
+ "step": 787
+ },
+ {
+ "epoch": 10.950218340611354,
+ "grad_norm": 0.028599515557289124,
+ "learning_rate": 0.0006,
+ "loss": 5.558349609375,
+ "step": 788
+ },
+ {
+ "epoch": 10.96419213973799,
+ "grad_norm": 0.02384260669350624,
+ "learning_rate": 0.0006,
+ "loss": 5.520661354064941,
+ "step": 789
+ },
+ {
+ "epoch": 10.97816593886463,
+ "grad_norm": 0.023180771619081497,
+ "learning_rate": 0.0006,
+ "loss": 5.496896743774414,
+ "step": 790
+ },
+ {
+ "epoch": 10.992139737991266,
+ "grad_norm": 0.021167006343603134,
+ "learning_rate": 0.0006,
+ "loss": 5.517192363739014,
+ "step": 791
+ },
+ {
+ "epoch": 11.0,
+ "grad_norm": 0.02069013938307762,
+ "learning_rate": 0.0006,
+ "loss": 5.447656154632568,
+ "step": 792
+ },
+ {
+ "epoch": 11.0,
+ "eval_loss": 6.161121845245361,
+ "eval_runtime": 71.5972,
+ "eval_samples_per_second": 34.107,
+ "eval_steps_per_second": 1.075,
+ "step": 792
+ },
+ {
+ "epoch": 11.013973799126637,
+ "grad_norm": 0.022983292117714882,
+ "learning_rate": 0.0006,
+ "loss": 5.441346645355225,
+ "step": 793
+ },
+ {
+ "epoch": 11.027947598253276,
+ "grad_norm": 0.022833511233329773,
+ "learning_rate": 0.0006,
+ "loss": 5.45123815536499,
+ "step": 794
+ },
+ {
+ "epoch": 11.041921397379912,
+ "grad_norm": 0.020015383139252663,
+ "learning_rate": 0.0006,
+ "loss": 5.430389404296875,
+ "step": 795
+ },
+ {
+ "epoch": 11.055895196506551,
+ "grad_norm": 0.016957225278019905,
+ "learning_rate": 0.0006,
+ "loss": 5.473883628845215,
+ "step": 796
+ },
+ {
+ "epoch": 11.069868995633188,
+ "grad_norm": 0.02090892568230629,
+ "learning_rate": 0.0006,
+ "loss": 5.467441558837891,
+ "step": 797
+ },
+ {
+ "epoch": 11.083842794759825,
+ "grad_norm": 0.022806301712989807,
+ "learning_rate": 0.0006,
+ "loss": 5.5453081130981445,
+ "step": 798
+ },
+ {
+ "epoch": 11.097816593886463,
+ "grad_norm": 0.022816946730017662,
+ "learning_rate": 0.0006,
+ "loss": 5.426213264465332,
+ "step": 799
+ },
+ {
+ "epoch": 11.1117903930131,
+ "grad_norm": 0.020761270076036453,
+ "learning_rate": 0.0006,
+ "loss": 5.464138507843018,
+ "step": 800
+ },
+ {
+ "epoch": 11.125764192139737,
+ "grad_norm": 0.01982618309557438,
+ "learning_rate": 0.0006,
+ "loss": 5.428033828735352,
+ "step": 801
+ },
+ {
+ "epoch": 11.139737991266376,
+ "grad_norm": 0.021953724324703217,
+ "learning_rate": 0.0006,
+ "loss": 5.456646919250488,
+ "step": 802
+ },
+ {
+ "epoch": 11.153711790393013,
+ "grad_norm": 0.01871841959655285,
+ "learning_rate": 0.0006,
+ "loss": 5.508403778076172,
+ "step": 803
+ },
+ {
+ "epoch": 11.167685589519651,
+ "grad_norm": 0.01810353808104992,
+ "learning_rate": 0.0006,
+ "loss": 5.442957878112793,
+ "step": 804
+ },
+ {
+ "epoch": 11.181659388646288,
+ "grad_norm": 0.017609186470508575,
+ "learning_rate": 0.0006,
+ "loss": 5.4774699211120605,
+ "step": 805
+ },
+ {
+ "epoch": 11.195633187772925,
+ "grad_norm": 0.01931103691458702,
+ "learning_rate": 0.0006,
+ "loss": 5.429748058319092,
+ "step": 806
+ },
+ {
+ "epoch": 11.209606986899564,
+ "grad_norm": 0.019401391968131065,
+ "learning_rate": 0.0006,
+ "loss": 5.466834545135498,
+ "step": 807
+ },
+ {
+ "epoch": 11.2235807860262,
+ "grad_norm": 0.018862580880522728,
+ "learning_rate": 0.0006,
+ "loss": 5.431701183319092,
+ "step": 808
+ },
+ {
+ "epoch": 11.237554585152838,
+ "grad_norm": 0.01796964928507805,
+ "learning_rate": 0.0006,
+ "loss": 5.388613700866699,
+ "step": 809
+ },
+ {
+ "epoch": 11.251528384279476,
+ "grad_norm": 0.021166296675801277,
+ "learning_rate": 0.0006,
+ "loss": 5.454118251800537,
+ "step": 810
+ },
+ {
+ "epoch": 11.265502183406113,
+ "grad_norm": 0.028568729758262634,
+ "learning_rate": 0.0006,
+ "loss": 5.510260581970215,
+ "step": 811
+ },
+ {
+ "epoch": 11.279475982532752,
+ "grad_norm": 0.03902891278266907,
+ "learning_rate": 0.0006,
+ "loss": 5.386396408081055,
+ "step": 812
+ },
+ {
+ "epoch": 11.293449781659389,
+ "grad_norm": 0.039114370942115784,
+ "learning_rate": 0.0006,
+ "loss": 5.457526206970215,
+ "step": 813
+ },
+ {
+ "epoch": 11.307423580786025,
+ "grad_norm": 0.03560924902558327,
+ "learning_rate": 0.0006,
+ "loss": 5.4149980545043945,
+ "step": 814
+ },
+ {
+ "epoch": 11.321397379912664,
+ "grad_norm": 0.03745810315012932,
+ "learning_rate": 0.0006,
+ "loss": 5.464084625244141,
+ "step": 815
+ },
+ {
+ "epoch": 11.335371179039301,
+ "grad_norm": 0.03383171558380127,
+ "learning_rate": 0.0006,
+ "loss": 5.45137882232666,
+ "step": 816
+ },
+ {
+ "epoch": 11.34934497816594,
+ "grad_norm": 0.027873128652572632,
+ "learning_rate": 0.0006,
+ "loss": 5.485784530639648,
+ "step": 817
+ },
+ {
+ "epoch": 11.363318777292577,
+ "grad_norm": 0.02697703056037426,
+ "learning_rate": 0.0006,
+ "loss": 5.459152698516846,
+ "step": 818
+ },
+ {
+ "epoch": 11.377292576419213,
+ "grad_norm": 0.02646288089454174,
+ "learning_rate": 0.0006,
+ "loss": 5.433281421661377,
+ "step": 819
+ },
+ {
+ "epoch": 11.391266375545852,
+ "grad_norm": 0.02655179798603058,
+ "learning_rate": 0.0006,
+ "loss": 5.456627368927002,
+ "step": 820
+ },
+ {
+ "epoch": 11.405240174672489,
+ "grad_norm": 0.02633797563612461,
+ "learning_rate": 0.0006,
+ "loss": 5.484803676605225,
+ "step": 821
+ },
+ {
+ "epoch": 11.419213973799126,
+ "grad_norm": 0.031187085434794426,
+ "learning_rate": 0.0006,
+ "loss": 5.362293243408203,
+ "step": 822
+ },
+ {
+ "epoch": 11.433187772925764,
+ "grad_norm": 0.02971349097788334,
+ "learning_rate": 0.0006,
+ "loss": 5.4369611740112305,
+ "step": 823
+ },
+ {
+ "epoch": 11.447161572052401,
+ "grad_norm": 0.02996021695435047,
+ "learning_rate": 0.0006,
+ "loss": 5.419068813323975,
+ "step": 824
+ },
+ {
+ "epoch": 11.46113537117904,
+ "grad_norm": 0.032519739121198654,
+ "learning_rate": 0.0006,
+ "loss": 5.422189712524414,
+ "step": 825
+ },
+ {
+ "epoch": 11.475109170305677,
+ "grad_norm": 0.03403490409255028,
+ "learning_rate": 0.0006,
+ "loss": 5.470074653625488,
+ "step": 826
+ },
+ {
+ "epoch": 11.489082969432314,
+ "grad_norm": 0.036989886313676834,
+ "learning_rate": 0.0006,
+ "loss": 5.460174560546875,
+ "step": 827
+ },
+ {
+ "epoch": 11.503056768558952,
+ "grad_norm": 0.03427419438958168,
+ "learning_rate": 0.0006,
+ "loss": 5.495920658111572,
+ "step": 828
+ },
+ {
+ "epoch": 11.51703056768559,
+ "grad_norm": 0.02519620768725872,
+ "learning_rate": 0.0006,
+ "loss": 5.474429130554199,
+ "step": 829
+ },
+ {
+ "epoch": 11.531004366812226,
+ "grad_norm": 0.02473096363246441,
+ "learning_rate": 0.0006,
+ "loss": 5.395847320556641,
+ "step": 830
+ },
+ {
+ "epoch": 11.544978165938865,
+ "grad_norm": 0.02792714536190033,
+ "learning_rate": 0.0006,
+ "loss": 5.376262664794922,
+ "step": 831
+ },
+ {
+ "epoch": 11.558951965065502,
+ "grad_norm": 0.024172596633434296,
+ "learning_rate": 0.0006,
+ "loss": 5.459731101989746,
+ "step": 832
+ },
+ {
+ "epoch": 11.57292576419214,
+ "grad_norm": 0.02304821088910103,
+ "learning_rate": 0.0006,
+ "loss": 5.479475498199463,
+ "step": 833
+ },
+ {
+ "epoch": 11.586899563318777,
+ "grad_norm": 0.02342340536415577,
+ "learning_rate": 0.0006,
+ "loss": 5.387955665588379,
+ "step": 834
+ },
+ {
+ "epoch": 11.600873362445414,
+ "grad_norm": 0.021962853148579597,
+ "learning_rate": 0.0006,
+ "loss": 5.385697364807129,
+ "step": 835
+ },
+ {
+ "epoch": 11.614847161572053,
+ "grad_norm": 0.019193042069673538,
+ "learning_rate": 0.0006,
+ "loss": 5.431601047515869,
+ "step": 836
+ },
+ {
+ "epoch": 11.62882096069869,
+ "grad_norm": 0.021378040313720703,
+ "learning_rate": 0.0006,
+ "loss": 5.3070969581604,
+ "step": 837
+ },
+ {
+ "epoch": 11.642794759825328,
+ "grad_norm": 0.021641355007886887,
+ "learning_rate": 0.0006,
+ "loss": 5.360584735870361,
+ "step": 838
+ },
+ {
+ "epoch": 11.656768558951965,
+ "grad_norm": 0.019899170845746994,
+ "learning_rate": 0.0006,
+ "loss": 5.398155212402344,
+ "step": 839
+ },
+ {
+ "epoch": 11.670742358078602,
+ "grad_norm": 0.02298869751393795,
+ "learning_rate": 0.0006,
+ "loss": 5.410561561584473,
+ "step": 840
+ },
+ {
+ "epoch": 11.68471615720524,
+ "grad_norm": 0.02815551869571209,
+ "learning_rate": 0.0006,
+ "loss": 5.298604965209961,
+ "step": 841
+ },
+ {
+ "epoch": 11.698689956331878,
+ "grad_norm": 0.03041219152510166,
+ "learning_rate": 0.0006,
+ "loss": 5.425869941711426,
+ "step": 842
+ },
+ {
+ "epoch": 11.712663755458514,
+ "grad_norm": 0.025710400193929672,
+ "learning_rate": 0.0006,
+ "loss": 5.344316482543945,
+ "step": 843
+ },
+ {
+ "epoch": 11.726637554585153,
+ "grad_norm": 0.018438996747136116,
+ "learning_rate": 0.0006,
+ "loss": 5.379913330078125,
+ "step": 844
+ },
+ {
+ "epoch": 11.74061135371179,
+ "grad_norm": 0.016683703288435936,
+ "learning_rate": 0.0006,
+ "loss": 5.338386058807373,
+ "step": 845
+ },
+ {
+ "epoch": 11.754585152838429,
+ "grad_norm": 0.017252853140234947,
+ "learning_rate": 0.0006,
+ "loss": 5.360256195068359,
+ "step": 846
+ },
+ {
+ "epoch": 11.768558951965066,
+ "grad_norm": 0.019606001675128937,
+ "learning_rate": 0.0006,
+ "loss": 5.401415824890137,
+ "step": 847
+ },
+ {
+ "epoch": 11.782532751091702,
+ "grad_norm": 0.016299517825245857,
+ "learning_rate": 0.0006,
+ "loss": 5.39270544052124,
+ "step": 848
+ },
+ {
+ "epoch": 11.796506550218341,
+ "grad_norm": 0.016657982021570206,
+ "learning_rate": 0.0006,
+ "loss": 5.359323978424072,
+ "step": 849
+ },
+ {
+ "epoch": 11.810480349344978,
+ "grad_norm": 0.017462270334362984,
+ "learning_rate": 0.0006,
+ "loss": 5.453655242919922,
+ "step": 850
+ },
+ {
+ "epoch": 11.824454148471617,
+ "grad_norm": 0.01723778434097767,
+ "learning_rate": 0.0006,
+ "loss": 5.454940319061279,
+ "step": 851
+ },
+ {
+ "epoch": 11.838427947598253,
+ "grad_norm": 0.019071755930781364,
+ "learning_rate": 0.0006,
+ "loss": 5.355914115905762,
+ "step": 852
+ },
+ {
+ "epoch": 11.85240174672489,
+ "grad_norm": 0.02169782482087612,
+ "learning_rate": 0.0006,
+ "loss": 5.396806716918945,
+ "step": 853
+ },
+ {
+ "epoch": 11.866375545851529,
+ "grad_norm": 0.025049131363630295,
+ "learning_rate": 0.0006,
+ "loss": 5.347422122955322,
+ "step": 854
+ },
+ {
+ "epoch": 11.880349344978166,
+ "grad_norm": 0.024977540597319603,
+ "learning_rate": 0.0006,
+ "loss": 5.32526969909668,
+ "step": 855
+ },
+ {
+ "epoch": 11.894323144104803,
+ "grad_norm": 0.024223418906331062,
+ "learning_rate": 0.0006,
+ "loss": 5.372445106506348,
+ "step": 856
+ },
+ {
+ "epoch": 11.908296943231441,
+ "grad_norm": 0.02519913762807846,
+ "learning_rate": 0.0006,
+ "loss": 5.280451774597168,
+ "step": 857
+ },
+ {
+ "epoch": 11.922270742358078,
+ "grad_norm": 0.0255630724132061,
+ "learning_rate": 0.0006,
+ "loss": 5.462911605834961,
+ "step": 858
+ },
+ {
+ "epoch": 11.936244541484717,
+ "grad_norm": 0.022740041837096214,
+ "learning_rate": 0.0006,
+ "loss": 5.319089412689209,
+ "step": 859
+ },
+ {
+ "epoch": 11.950218340611354,
+ "grad_norm": 0.02038649097084999,
+ "learning_rate": 0.0006,
+ "loss": 5.37426233291626,
+ "step": 860
+ },
+ {
+ "epoch": 11.96419213973799,
+ "grad_norm": 0.020509786903858185,
+ "learning_rate": 0.0006,
+ "loss": 5.337924957275391,
+ "step": 861
+ },
+ {
+ "epoch": 11.97816593886463,
+ "grad_norm": 0.021707357838749886,
+ "learning_rate": 0.0006,
+ "loss": 5.374853134155273,
+ "step": 862
+ },
+ {
+ "epoch": 11.992139737991266,
+ "grad_norm": 0.02714136242866516,
+ "learning_rate": 0.0006,
+ "loss": 5.2982282638549805,
+ "step": 863
+ },
+ {
+ "epoch": 12.0,
+ "grad_norm": 0.027913225814700127,
+ "learning_rate": 0.0006,
+ "loss": 5.406044960021973,
+ "step": 864
+ },
+ {
+ "epoch": 12.0,
+ "eval_loss": 5.948610305786133,
+ "eval_runtime": 75.79,
+ "eval_samples_per_second": 32.221,
+ "eval_steps_per_second": 1.016,
+ "step": 864
+ },
+ {
+ "epoch": 12.013973799126637,
+ "grad_norm": 0.023761706426739693,
+ "learning_rate": 0.0006,
+ "loss": 5.378547191619873,
+ "step": 865
+ },
+ {
+ "epoch": 12.027947598253276,
+ "grad_norm": 0.026866815984249115,
+ "learning_rate": 0.0006,
+ "loss": 5.356947898864746,
+ "step": 866
+ },
+ {
+ "epoch": 12.041921397379912,
+ "grad_norm": 0.03335791453719139,
+ "learning_rate": 0.0006,
+ "loss": 5.264157772064209,
+ "step": 867
+ },
+ {
+ "epoch": 12.055895196506551,
+ "grad_norm": 0.03153866529464722,
+ "learning_rate": 0.0006,
+ "loss": 5.363075256347656,
+ "step": 868
+ },
+ {
+ "epoch": 12.069868995633188,
+ "grad_norm": 0.03270367905497551,
+ "learning_rate": 0.0006,
+ "loss": 5.298289775848389,
+ "step": 869
+ },
+ {
+ "epoch": 12.083842794759825,
+ "grad_norm": 0.030872846022248268,
+ "learning_rate": 0.0006,
+ "loss": 5.324321269989014,
+ "step": 870
+ },
+ {
+ "epoch": 12.097816593886463,
+ "grad_norm": 0.028438350185751915,
+ "learning_rate": 0.0006,
+ "loss": 5.3512115478515625,
+ "step": 871
+ },
+ {
+ "epoch": 12.1117903930131,
+ "grad_norm": 0.030846193432807922,
+ "learning_rate": 0.0006,
+ "loss": 5.304851531982422,
+ "step": 872
+ },
+ {
+ "epoch": 12.125764192139737,
+ "grad_norm": 0.03879433870315552,
+ "learning_rate": 0.0006,
+ "loss": 5.289634704589844,
+ "step": 873
+ },
+ {
+ "epoch": 12.139737991266376,
+ "grad_norm": 0.03665967285633087,
+ "learning_rate": 0.0006,
+ "loss": 5.352825164794922,
+ "step": 874
+ },
+ {
+ "epoch": 12.153711790393013,
+ "grad_norm": 0.03454619646072388,
+ "learning_rate": 0.0006,
+ "loss": 5.346499443054199,
+ "step": 875
+ },
+ {
+ "epoch": 12.167685589519651,
+ "grad_norm": 0.043245580047369,
+ "learning_rate": 0.0006,
+ "loss": 5.344263076782227,
+ "step": 876
+ },
+ {
+ "epoch": 12.181659388646288,
+ "grad_norm": 0.04450082778930664,
+ "learning_rate": 0.0006,
+ "loss": 5.447438716888428,
+ "step": 877
+ },
+ {
+ "epoch": 12.195633187772925,
+ "grad_norm": 0.04980820417404175,
+ "learning_rate": 0.0006,
+ "loss": 5.364802360534668,
+ "step": 878
+ },
+ {
+ "epoch": 12.209606986899564,
+ "grad_norm": 0.04757903143763542,
+ "learning_rate": 0.0006,
+ "loss": 5.363414287567139,
+ "step": 879
+ },
+ {
+ "epoch": 12.2235807860262,
+ "grad_norm": 0.048647016286849976,
+ "learning_rate": 0.0006,
+ "loss": 5.379604339599609,
+ "step": 880
+ },
+ {
+ "epoch": 12.237554585152838,
+ "grad_norm": 0.04497060924768448,
+ "learning_rate": 0.0006,
+ "loss": 5.354029178619385,
+ "step": 881
+ },
+ {
+ "epoch": 12.251528384279476,
+ "grad_norm": 0.032642051577568054,
+ "learning_rate": 0.0006,
+ "loss": 5.335702419281006,
+ "step": 882
+ },
+ {
+ "epoch": 12.265502183406113,
+ "grad_norm": 0.037766437977552414,
+ "learning_rate": 0.0006,
+ "loss": 5.331804275512695,
+ "step": 883
+ },
+ {
+ "epoch": 12.279475982532752,
+ "grad_norm": 0.030145324766635895,
+ "learning_rate": 0.0006,
+ "loss": 5.327176094055176,
+ "step": 884
+ },
+ {
+ "epoch": 12.293449781659389,
+ "grad_norm": 0.032248858362436295,
+ "learning_rate": 0.0006,
+ "loss": 5.2410712242126465,
+ "step": 885
+ },
+ {
+ "epoch": 12.307423580786025,
+ "grad_norm": 0.026990503072738647,
+ "learning_rate": 0.0006,
+ "loss": 5.369958877563477,
+ "step": 886
+ },
+ {
+ "epoch": 12.321397379912664,
+ "grad_norm": 0.023413540795445442,
+ "learning_rate": 0.0006,
+ "loss": 5.322418212890625,
+ "step": 887
+ },
+ {
+ "epoch": 12.335371179039301,
+ "grad_norm": 0.025248292833566666,
+ "learning_rate": 0.0006,
+ "loss": 5.300156116485596,
+ "step": 888
+ },
+ {
+ "epoch": 12.34934497816594,
+ "grad_norm": 0.021037699654698372,
+ "learning_rate": 0.0006,
+ "loss": 5.26978063583374,
+ "step": 889
+ },
+ {
+ "epoch": 12.363318777292577,
+ "grad_norm": 0.0155200669541955,
+ "learning_rate": 0.0006,
+ "loss": 5.297061920166016,
+ "step": 890
+ },
+ {
+ "epoch": 12.377292576419213,
+ "grad_norm": 0.017867842689156532,
+ "learning_rate": 0.0006,
+ "loss": 5.312634468078613,
+ "step": 891
+ },
+ {
+ "epoch": 12.391266375545852,
+ "grad_norm": 0.01675534062087536,
+ "learning_rate": 0.0006,
+ "loss": 5.290635108947754,
+ "step": 892
+ },
+ {
+ "epoch": 12.405240174672489,
+ "grad_norm": 0.015546303242444992,
+ "learning_rate": 0.0006,
+ "loss": 5.289699554443359,
+ "step": 893
+ },
+ {
+ "epoch": 12.419213973799126,
+ "grad_norm": 0.01612856611609459,
+ "learning_rate": 0.0006,
+ "loss": 5.312010765075684,
+ "step": 894
+ },
+ {
+ "epoch": 12.433187772925764,
+ "grad_norm": 0.015225349925458431,
+ "learning_rate": 0.0006,
+ "loss": 5.255838871002197,
+ "step": 895
+ },
+ {
+ "epoch": 12.447161572052401,
+ "grad_norm": 0.015317161567509174,
+ "learning_rate": 0.0006,
+ "loss": 5.234170436859131,
+ "step": 896
+ },
+ {
+ "epoch": 12.46113537117904,
+ "grad_norm": 0.014820792712271214,
+ "learning_rate": 0.0006,
+ "loss": 5.255383014678955,
+ "step": 897
+ },
+ {
+ "epoch": 12.475109170305677,
+ "grad_norm": 0.01343776099383831,
+ "learning_rate": 0.0006,
+ "loss": 5.256904602050781,
+ "step": 898
+ },
+ {
+ "epoch": 12.489082969432314,
+ "grad_norm": 0.01418706402182579,
+ "learning_rate": 0.0006,
+ "loss": 5.272562026977539,
+ "step": 899
+ },
+ {
+ "epoch": 12.503056768558952,
+ "grad_norm": 0.013406263664364815,
+ "learning_rate": 0.0006,
+ "loss": 5.17042350769043,
+ "step": 900
+ },
+ {
+ "epoch": 12.51703056768559,
+ "grad_norm": 0.01331746019423008,
+ "learning_rate": 0.0006,
+ "loss": 5.267889022827148,
+ "step": 901
+ },
+ {
+ "epoch": 12.531004366812226,
+ "grad_norm": 0.014625098556280136,
+ "learning_rate": 0.0006,
+ "loss": 5.281431198120117,
+ "step": 902
+ },
+ {
+ "epoch": 12.544978165938865,
+ "grad_norm": 0.01346531230956316,
+ "learning_rate": 0.0006,
+ "loss": 5.286074638366699,
+ "step": 903
+ },
+ {
+ "epoch": 12.558951965065502,
+ "grad_norm": 0.01303934771567583,
+ "learning_rate": 0.0006,
+ "loss": 5.344804763793945,
+ "step": 904
+ },
+ {
+ "epoch": 12.57292576419214,
+ "grad_norm": 0.014180841855704784,
+ "learning_rate": 0.0006,
+ "loss": 5.252955436706543,
+ "step": 905
+ },
+ {
+ "epoch": 12.586899563318777,
+ "grad_norm": 0.014279021881520748,
+ "learning_rate": 0.0006,
+ "loss": 5.283501625061035,
+ "step": 906
+ },
+ {
+ "epoch": 12.600873362445414,
+ "grad_norm": 0.013922203332185745,
+ "learning_rate": 0.0006,
+ "loss": 5.225350379943848,
+ "step": 907
+ },
+ {
+ "epoch": 12.614847161572053,
+ "grad_norm": 0.014055478386580944,
+ "learning_rate": 0.0006,
+ "loss": 5.283530235290527,
+ "step": 908
+ },
+ {
+ "epoch": 12.62882096069869,
+ "grad_norm": 0.013623543083667755,
+ "learning_rate": 0.0006,
+ "loss": 5.204877853393555,
+ "step": 909
+ },
+ {
+ "epoch": 12.642794759825328,
+ "grad_norm": 0.014378432184457779,
+ "learning_rate": 0.0006,
+ "loss": 5.275735378265381,
+ "step": 910
+ },
+ {
+ "epoch": 12.656768558951965,
+ "grad_norm": 0.017884068191051483,
+ "learning_rate": 0.0006,
+ "loss": 5.256485939025879,
+ "step": 911
+ },
+ {
+ "epoch": 12.670742358078602,
+ "grad_norm": 0.0203024260699749,
+ "learning_rate": 0.0006,
+ "loss": 5.292258262634277,
+ "step": 912
+ },
+ {
+ "epoch": 12.68471615720524,
+ "grad_norm": 0.021196166053414345,
+ "learning_rate": 0.0006,
+ "loss": 5.30052375793457,
+ "step": 913
+ },
+ {
+ "epoch": 12.698689956331878,
+ "grad_norm": 0.02249213121831417,
+ "learning_rate": 0.0006,
+ "loss": 5.236924648284912,
+ "step": 914
+ },
+ {
+ "epoch": 12.712663755458514,
+ "grad_norm": 0.019980570301413536,
+ "learning_rate": 0.0006,
+ "loss": 5.290444374084473,
+ "step": 915
+ },
+ {
+ "epoch": 12.726637554585153,
+ "grad_norm": 0.022030364722013474,
+ "learning_rate": 0.0006,
+ "loss": 5.287881851196289,
+ "step": 916
+ },
+ {
+ "epoch": 12.74061135371179,
+ "grad_norm": 0.02438683621585369,
+ "learning_rate": 0.0006,
+ "loss": 5.230462074279785,
+ "step": 917
+ },
+ {
+ "epoch": 12.754585152838429,
+ "grad_norm": 0.026895932853221893,
+ "learning_rate": 0.0006,
+ "loss": 5.232667446136475,
+ "step": 918
+ },
+ {
+ "epoch": 12.768558951965066,
+ "grad_norm": 0.02861972153186798,
+ "learning_rate": 0.0006,
+ "loss": 5.205913066864014,
+ "step": 919
+ },
+ {
+ "epoch": 12.782532751091702,
+ "grad_norm": 0.033064406365156174,
+ "learning_rate": 0.0006,
+ "loss": 5.242166042327881,
+ "step": 920
+ },
+ {
+ "epoch": 12.796506550218341,
+ "grad_norm": 0.030850905925035477,
+ "learning_rate": 0.0006,
+ "loss": 5.2503581047058105,
+ "step": 921
+ },
+ {
+ "epoch": 12.810480349344978,
+ "grad_norm": 0.03139309585094452,
+ "learning_rate": 0.0006,
+ "loss": 5.222465515136719,
+ "step": 922
+ },
+ {
+ "epoch": 12.824454148471617,
+ "grad_norm": 0.035760655999183655,
+ "learning_rate": 0.0006,
+ "loss": 5.279786586761475,
+ "step": 923
+ },
+ {
+ "epoch": 12.838427947598253,
+ "grad_norm": 0.036571063101291656,
+ "learning_rate": 0.0006,
+ "loss": 5.237204551696777,
+ "step": 924
+ },
+ {
+ "epoch": 12.85240174672489,
+ "grad_norm": 0.03467971459031105,
+ "learning_rate": 0.0006,
+ "loss": 5.210154056549072,
+ "step": 925
+ },
+ {
+ "epoch": 12.866375545851529,
+ "grad_norm": 0.053154416382312775,
+ "learning_rate": 0.0006,
+ "loss": 5.295984268188477,
+ "step": 926
+ },
+ {
+ "epoch": 12.880349344978166,
+ "grad_norm": 0.05374281480908394,
+ "learning_rate": 0.0006,
+ "loss": 5.293900012969971,
+ "step": 927
+ },
+ {
+ "epoch": 12.894323144104803,
+ "grad_norm": 0.05851082503795624,
+ "learning_rate": 0.0006,
+ "loss": 5.3758931159973145,
+ "step": 928
+ },
+ {
+ "epoch": 12.908296943231441,
+ "grad_norm": 0.05145686864852905,
+ "learning_rate": 0.0006,
+ "loss": 5.309441566467285,
+ "step": 929
+ },
+ {
+ "epoch": 12.922270742358078,
+ "grad_norm": 0.04484057426452637,
+ "learning_rate": 0.0006,
+ "loss": 5.297204494476318,
+ "step": 930
+ },
+ {
+ "epoch": 12.936244541484717,
+ "grad_norm": 0.05348053574562073,
+ "learning_rate": 0.0006,
+ "loss": 5.278589248657227,
+ "step": 931
+ },
+ {
+ "epoch": 12.950218340611354,
+ "grad_norm": 0.05677935481071472,
+ "learning_rate": 0.0006,
+ "loss": 5.290632247924805,
+ "step": 932
+ },
+ {
+ "epoch": 12.96419213973799,
+ "grad_norm": 0.05244941636919975,
+ "learning_rate": 0.0006,
+ "loss": 5.353776454925537,
+ "step": 933
+ },
+ {
+ "epoch": 12.97816593886463,
+ "grad_norm": 0.03900328651070595,
+ "learning_rate": 0.0006,
+ "loss": 5.284013748168945,
+ "step": 934
+ },
+ {
+ "epoch": 12.992139737991266,
+ "grad_norm": 0.03373100981116295,
+ "learning_rate": 0.0006,
+ "loss": 5.3313751220703125,
+ "step": 935
+ },
+ {
+ "epoch": 13.0,
+ "grad_norm": 0.03497432917356491,
+ "learning_rate": 0.0006,
+ "loss": 5.395255088806152,
+ "step": 936
+ },
+ {
+ "epoch": 13.0,
+ "eval_loss": 5.799665451049805,
+ "eval_runtime": 81.3525,
+ "eval_samples_per_second": 30.018,
+ "eval_steps_per_second": 0.946,
+ "step": 936
+ },
+ {
+ "epoch": 13.013973799126637,
+ "grad_norm": 0.039509695023298264,
+ "learning_rate": 0.0006,
+ "loss": 5.298876762390137,
+ "step": 937
+ },
+ {
+ "epoch": 13.027947598253276,
+ "grad_norm": 0.034599535167217255,
+ "learning_rate": 0.0006,
+ "loss": 5.264098167419434,
+ "step": 938
+ },
+ {
+ "epoch": 13.041921397379912,
+ "grad_norm": 0.029622144997119904,
+ "learning_rate": 0.0006,
+ "loss": 5.21397066116333,
+ "step": 939
+ },
+ {
+ "epoch": 13.055895196506551,
+ "grad_norm": 0.02309993840754032,
+ "learning_rate": 0.0006,
+ "loss": 5.227992057800293,
+ "step": 940
+ },
+ {
+ "epoch": 13.069868995633188,
+ "grad_norm": 0.024507572874426842,
+ "learning_rate": 0.0006,
+ "loss": 5.192324161529541,
+ "step": 941
+ },
+ {
+ "epoch": 13.083842794759825,
+ "grad_norm": 0.018214358016848564,
+ "learning_rate": 0.0006,
+ "loss": 5.23871374130249,
+ "step": 942
+ },
+ {
+ "epoch": 13.097816593886463,
+ "grad_norm": 0.01937488280236721,
+ "learning_rate": 0.0006,
+ "loss": 5.227265357971191,
+ "step": 943
+ },
+ {
+ "epoch": 13.1117903930131,
+ "grad_norm": 0.017225295305252075,
+ "learning_rate": 0.0006,
+ "loss": 5.232108116149902,
+ "step": 944
+ },
+ {
+ "epoch": 13.125764192139737,
+ "grad_norm": 0.016314376145601273,
+ "learning_rate": 0.0006,
+ "loss": 5.2288360595703125,
+ "step": 945
+ },
+ {
+ "epoch": 13.139737991266376,
+ "grad_norm": 0.014701297506690025,
+ "learning_rate": 0.0006,
+ "loss": 5.297238349914551,
+ "step": 946
+ },
+ {
+ "epoch": 13.153711790393013,
+ "grad_norm": 0.015610086731612682,
+ "learning_rate": 0.0006,
+ "loss": 5.148407936096191,
+ "step": 947
+ },
+ {
+ "epoch": 13.167685589519651,
+ "grad_norm": 0.014056755229830742,
+ "learning_rate": 0.0006,
+ "loss": 5.208695411682129,
+ "step": 948
+ },
+ {
+ "epoch": 13.181659388646288,
+ "grad_norm": 0.012802829965949059,
+ "learning_rate": 0.0006,
+ "loss": 5.229208946228027,
+ "step": 949
+ },
+ {
+ "epoch": 13.195633187772925,
+ "grad_norm": 0.011958962306380272,
+ "learning_rate": 0.0006,
+ "loss": 5.201716423034668,
+ "step": 950
+ },
+ {
+ "epoch": 13.209606986899564,
+ "grad_norm": 0.012841588817536831,
+ "learning_rate": 0.0006,
+ "loss": 5.240131378173828,
+ "step": 951
+ },
+ {
+ "epoch": 13.2235807860262,
+ "grad_norm": 0.014653137885034084,
+ "learning_rate": 0.0006,
+ "loss": 5.146340370178223,
+ "step": 952
+ },
+ {
+ "epoch": 13.237554585152838,
+ "grad_norm": 0.010446185246109962,
+ "learning_rate": 0.0006,
+ "loss": 5.21809720993042,
+ "step": 953
+ },
+ {
+ "epoch": 13.251528384279476,
+ "grad_norm": 0.01092249620705843,
+ "learning_rate": 0.0006,
+ "loss": 5.135411739349365,
+ "step": 954
+ },
+ {
+ "epoch": 13.265502183406113,
+ "grad_norm": 0.010561129078269005,
+ "learning_rate": 0.0006,
+ "loss": 5.194492340087891,
+ "step": 955
+ },
+ {
+ "epoch": 13.279475982532752,
+ "grad_norm": 0.010118645615875721,
+ "learning_rate": 0.0006,
+ "loss": 5.235012531280518,
+ "step": 956
+ },
+ {
+ "epoch": 13.293449781659389,
+ "grad_norm": 0.01081101130694151,
+ "learning_rate": 0.0006,
+ "loss": 5.162468910217285,
+ "step": 957
+ },
+ {
+ "epoch": 13.307423580786025,
+ "grad_norm": 0.010701699182391167,
+ "learning_rate": 0.0006,
+ "loss": 5.217264175415039,
+ "step": 958
+ },
+ {
+ "epoch": 13.321397379912664,
+ "grad_norm": 0.010495852679014206,
+ "learning_rate": 0.0006,
+ "loss": 5.202086925506592,
+ "step": 959
+ },
+ {
+ "epoch": 13.335371179039301,
+ "grad_norm": 0.009838009253144264,
+ "learning_rate": 0.0006,
+ "loss": 5.201089859008789,
+ "step": 960
+ },
+ {
+ "epoch": 13.34934497816594,
+ "grad_norm": 0.011280231177806854,
+ "learning_rate": 0.0006,
+ "loss": 5.1775336265563965,
+ "step": 961
+ },
+ {
+ "epoch": 13.363318777292577,
+ "grad_norm": 0.011244936846196651,
+ "learning_rate": 0.0006,
+ "loss": 5.185998439788818,
+ "step": 962
+ },
+ {
+ "epoch": 13.377292576419213,
+ "grad_norm": 0.01153804361820221,
+ "learning_rate": 0.0006,
+ "loss": 5.132739067077637,
+ "step": 963
+ },
+ {
+ "epoch": 13.391266375545852,
+ "grad_norm": 0.010817309841513634,
+ "learning_rate": 0.0006,
+ "loss": 5.268172264099121,
+ "step": 964
+ },
+ {
+ "epoch": 13.405240174672489,
+ "grad_norm": 0.011870342306792736,
+ "learning_rate": 0.0006,
+ "loss": 5.154515266418457,
+ "step": 965
+ },
+ {
+ "epoch": 13.419213973799126,
+ "grad_norm": 0.011373576708137989,
+ "learning_rate": 0.0006,
+ "loss": 5.160059928894043,
+ "step": 966
+ },
+ {
+ "epoch": 13.433187772925764,
+ "grad_norm": 0.01116643100976944,
+ "learning_rate": 0.0006,
+ "loss": 5.079656600952148,
+ "step": 967
+ },
+ {
+ "epoch": 13.447161572052401,
+ "grad_norm": 0.01202121190726757,
+ "learning_rate": 0.0006,
+ "loss": 5.205666542053223,
+ "step": 968
+ },
+ {
+ "epoch": 13.46113537117904,
+ "grad_norm": 0.013683444820344448,
+ "learning_rate": 0.0006,
+ "loss": 5.168977737426758,
+ "step": 969
+ },
+ {
+ "epoch": 13.475109170305677,
+ "grad_norm": 0.012892362661659718,
+ "learning_rate": 0.0006,
+ "loss": 5.168679237365723,
+ "step": 970
+ },
+ {
+ "epoch": 13.489082969432314,
+ "grad_norm": 0.01123902853578329,
+ "learning_rate": 0.0006,
+ "loss": 5.143795013427734,
+ "step": 971
+ },
+ {
+ "epoch": 13.503056768558952,
+ "grad_norm": 0.010333497077226639,
+ "learning_rate": 0.0006,
+ "loss": 5.170592784881592,
+ "step": 972
+ },
+ {
+ "epoch": 13.51703056768559,
+ "grad_norm": 0.011278851889073849,
+ "learning_rate": 0.0006,
+ "loss": 5.162395477294922,
+ "step": 973
+ },
+ {
+ "epoch": 13.531004366812226,
+ "grad_norm": 0.013309221714735031,
+ "learning_rate": 0.0006,
+ "loss": 5.0712103843688965,
+ "step": 974
+ },
+ {
+ "epoch": 13.544978165938865,
+ "grad_norm": 0.014573273248970509,
+ "learning_rate": 0.0006,
+ "loss": 5.098300933837891,
+ "step": 975
+ },
+ {
+ "epoch": 13.558951965065502,
+ "grad_norm": 0.01755373552441597,
+ "learning_rate": 0.0006,
+ "loss": 5.081066131591797,
+ "step": 976
+ },
+ {
+ "epoch": 13.57292576419214,
+ "grad_norm": 0.02018735185265541,
+ "learning_rate": 0.0006,
+ "loss": 5.173007965087891,
+ "step": 977
+ },
+ {
+ "epoch": 13.586899563318777,
+ "grad_norm": 0.023258069530129433,
+ "learning_rate": 0.0006,
+ "loss": 5.107330322265625,
+ "step": 978
+ },
+ {
+ "epoch": 13.600873362445414,
+ "grad_norm": 0.0250930804759264,
+ "learning_rate": 0.0006,
+ "loss": 5.193347930908203,
+ "step": 979
+ },
+ {
+ "epoch": 13.614847161572053,
+ "grad_norm": 0.02333216182887554,
+ "learning_rate": 0.0006,
+ "loss": 5.160326957702637,
+ "step": 980
+ },
+ {
+ "epoch": 13.62882096069869,
+ "grad_norm": 0.019765766337513924,
+ "learning_rate": 0.0006,
+ "loss": 5.05100679397583,
+ "step": 981
+ },
+ {
+ "epoch": 13.642794759825328,
+ "grad_norm": 0.020395414903759956,
+ "learning_rate": 0.0006,
+ "loss": 5.157297611236572,
+ "step": 982
+ },
+ {
+ "epoch": 13.656768558951965,
+ "grad_norm": 0.019553566351532936,
+ "learning_rate": 0.0006,
+ "loss": 5.158632755279541,
+ "step": 983
+ },
+ {
+ "epoch": 13.670742358078602,
+ "grad_norm": 0.018838031217455864,
+ "learning_rate": 0.0006,
+ "loss": 5.167074203491211,
+ "step": 984
+ },
+ {
+ "epoch": 13.68471615720524,
+ "grad_norm": 0.020884843543171883,
+ "learning_rate": 0.0006,
+ "loss": 5.238006114959717,
+ "step": 985
+ },
+ {
+ "epoch": 13.698689956331878,
+ "grad_norm": 0.021462196484208107,
+ "learning_rate": 0.0006,
+ "loss": 5.1703901290893555,
+ "step": 986
+ },
+ {
+ "epoch": 13.712663755458514,
+ "grad_norm": 0.021178999915719032,
+ "learning_rate": 0.0006,
+ "loss": 5.109465599060059,
+ "step": 987
+ },
+ {
+ "epoch": 13.726637554585153,
+ "grad_norm": 0.022374693304300308,
+ "learning_rate": 0.0006,
+ "loss": 5.195014953613281,
+ "step": 988
+ },
+ {
+ "epoch": 13.74061135371179,
+ "grad_norm": 0.021914934739470482,
+ "learning_rate": 0.0006,
+ "loss": 5.15154504776001,
+ "step": 989
+ },
+ {
+ "epoch": 13.754585152838429,
+ "grad_norm": 0.017769858241081238,
+ "learning_rate": 0.0006,
+ "loss": 5.119076251983643,
+ "step": 990
+ },
+ {
+ "epoch": 13.768558951965066,
+ "grad_norm": 0.02018590457737446,
+ "learning_rate": 0.0006,
+ "loss": 5.087010860443115,
+ "step": 991
+ },
+ {
+ "epoch": 13.782532751091702,
+ "grad_norm": 0.023143045604228973,
+ "learning_rate": 0.0006,
+ "loss": 5.087374687194824,
+ "step": 992
+ },
+ {
+ "epoch": 13.796506550218341,
+ "grad_norm": 0.02214638516306877,
+ "learning_rate": 0.0006,
+ "loss": 5.043274879455566,
+ "step": 993
+ },
+ {
+ "epoch": 13.810480349344978,
+ "grad_norm": 0.024159753695130348,
+ "learning_rate": 0.0006,
+ "loss": 5.11695671081543,
+ "step": 994
+ },
+ {
+ "epoch": 13.824454148471617,
+ "grad_norm": 0.023281913250684738,
+ "learning_rate": 0.0006,
+ "loss": 5.159364700317383,
+ "step": 995
+ },
+ {
+ "epoch": 13.838427947598253,
+ "grad_norm": 0.02846658229827881,
+ "learning_rate": 0.0006,
+ "loss": 5.0652570724487305,
+ "step": 996
+ },
+ {
+ "epoch": 13.85240174672489,
+ "grad_norm": 0.03165476396679878,
+ "learning_rate": 0.0006,
+ "loss": 5.136387825012207,
+ "step": 997
+ },
+ {
+ "epoch": 13.866375545851529,
+ "grad_norm": 0.02582591027021408,
+ "learning_rate": 0.0006,
+ "loss": 5.059978485107422,
+ "step": 998
+ },
+ {
+ "epoch": 13.880349344978166,
+ "grad_norm": 0.025306783616542816,
+ "learning_rate": 0.0006,
+ "loss": 5.123497486114502,
+ "step": 999
+ },
+ {
+ "epoch": 13.894323144104803,
+ "grad_norm": 0.02598070539534092,
+ "learning_rate": 0.0006,
+ "loss": 5.155566215515137,
+ "step": 1000
+ },
+ {
+ "epoch": 13.908296943231441,
+ "grad_norm": 0.026244031265378,
+ "learning_rate": 0.0006,
+ "loss": 5.0871381759643555,
+ "step": 1001
+ },
+ {
+ "epoch": 13.922270742358078,
+ "grad_norm": 0.025219639763236046,
+ "learning_rate": 0.0006,
+ "loss": 5.102408409118652,
+ "step": 1002
+ },
+ {
+ "epoch": 13.936244541484717,
+ "grad_norm": 0.023916568607091904,
+ "learning_rate": 0.0006,
+ "loss": 5.131472587585449,
+ "step": 1003
+ },
+ {
+ "epoch": 13.950218340611354,
+ "grad_norm": 0.019619153812527657,
+ "learning_rate": 0.0006,
+ "loss": 5.14764404296875,
+ "step": 1004
+ },
+ {
+ "epoch": 13.96419213973799,
+ "grad_norm": 0.019982660189270973,
+ "learning_rate": 0.0006,
+ "loss": 5.0941972732543945,
+ "step": 1005
+ },
+ {
+ "epoch": 13.97816593886463,
+ "grad_norm": 0.018386343494057655,
+ "learning_rate": 0.0006,
+ "loss": 5.138663291931152,
+ "step": 1006
+ },
+ {
+ "epoch": 13.992139737991266,
+ "grad_norm": 0.021332457661628723,
+ "learning_rate": 0.0006,
+ "loss": 5.141696453094482,
+ "step": 1007
+ },
+ {
+ "epoch": 14.0,
+ "grad_norm": 0.021462492644786835,
+ "learning_rate": 0.0006,
+ "loss": 5.049698352813721,
+ "step": 1008
+ },
+ {
+ "epoch": 14.0,
+ "eval_loss": 5.544618129730225,
+ "eval_runtime": 72.1254,
+ "eval_samples_per_second": 33.858,
+ "eval_steps_per_second": 1.068,
+ "step": 1008
+ },
+ {
+ "epoch": 14.013973799126637,
+ "grad_norm": 0.02191290818154812,
+ "learning_rate": 0.0006,
+ "loss": 5.0831732749938965,
+ "step": 1009
+ },
+ {
+ "epoch": 14.027947598253276,
+ "grad_norm": 0.02411566860973835,
+ "learning_rate": 0.0006,
+ "loss": 5.116334915161133,
+ "step": 1010
+ },
+ {
+ "epoch": 14.041921397379912,
+ "grad_norm": 0.027553394436836243,
+ "learning_rate": 0.0006,
+ "loss": 5.050045013427734,
+ "step": 1011
+ },
+ {
+ "epoch": 14.055895196506551,
+ "grad_norm": 0.02357514388859272,
+ "learning_rate": 0.0006,
+ "loss": 5.091012001037598,
+ "step": 1012
+ },
+ {
+ "epoch": 14.069868995633188,
+ "grad_norm": 0.021199429407715797,
+ "learning_rate": 0.0006,
+ "loss": 5.104509353637695,
+ "step": 1013
+ },
+ {
+ "epoch": 14.083842794759825,
+ "grad_norm": 0.019254926592111588,
+ "learning_rate": 0.0006,
+ "loss": 5.141340255737305,
+ "step": 1014
+ },
+ {
+ "epoch": 14.097816593886463,
+ "grad_norm": 0.01811135932803154,
+ "learning_rate": 0.0006,
+ "loss": 5.046824932098389,
+ "step": 1015
+ },
+ {
+ "epoch": 14.1117903930131,
+ "grad_norm": 0.017739299684762955,
+ "learning_rate": 0.0006,
+ "loss": 5.045833587646484,
+ "step": 1016
+ },
+ {
+ "epoch": 14.125764192139737,
+ "grad_norm": 0.018534071743488312,
+ "learning_rate": 0.0006,
+ "loss": 4.9618401527404785,
+ "step": 1017
+ },
+ {
+ "epoch": 14.139737991266376,
+ "grad_norm": 0.016690023243427277,
+ "learning_rate": 0.0006,
+ "loss": 5.165715217590332,
+ "step": 1018
+ },
+ {
+ "epoch": 14.153711790393013,
+ "grad_norm": 0.018630729988217354,
+ "learning_rate": 0.0006,
+ "loss": 5.082975387573242,
+ "step": 1019
+ },
+ {
+ "epoch": 14.167685589519651,
+ "grad_norm": 0.01884593442082405,
+ "learning_rate": 0.0006,
+ "loss": 5.048741817474365,
+ "step": 1020
+ },
+ {
+ "epoch": 14.181659388646288,
+ "grad_norm": 0.018500948324799538,
+ "learning_rate": 0.0006,
+ "loss": 5.105315685272217,
+ "step": 1021
+ },
+ {
+ "epoch": 14.195633187772925,
+ "grad_norm": 0.020263809710741043,
+ "learning_rate": 0.0006,
+ "loss": 5.087735176086426,
+ "step": 1022
+ },
+ {
+ "epoch": 14.209606986899564,
+ "grad_norm": 0.019454646855592728,
+ "learning_rate": 0.0006,
+ "loss": 5.0284810066223145,
+ "step": 1023
+ },
+ {
+ "epoch": 14.2235807860262,
+ "grad_norm": 0.01877632364630699,
+ "learning_rate": 0.0006,
+ "loss": 5.0442609786987305,
+ "step": 1024
+ },
+ {
+ "epoch": 14.237554585152838,
+ "grad_norm": 0.020227789878845215,
+ "learning_rate": 0.0006,
+ "loss": 5.081351280212402,
+ "step": 1025
+ },
+ {
+ "epoch": 14.251528384279476,
+ "grad_norm": 0.018256092444062233,
+ "learning_rate": 0.0006,
+ "loss": 5.016656875610352,
+ "step": 1026
+ },
+ {
+ "epoch": 14.265502183406113,
+ "grad_norm": 0.01916426233947277,
+ "learning_rate": 0.0006,
+ "loss": 5.034103870391846,
+ "step": 1027
+ },
+ {
+ "epoch": 14.279475982532752,
+ "grad_norm": 0.017990969121456146,
+ "learning_rate": 0.0006,
+ "loss": 5.039344310760498,
+ "step": 1028
+ },
+ {
+ "epoch": 14.293449781659389,
+ "grad_norm": 0.018346179276704788,
+ "learning_rate": 0.0006,
+ "loss": 5.017010688781738,
+ "step": 1029
+ },
+ {
+ "epoch": 14.307423580786025,
+ "grad_norm": 0.01584712415933609,
+ "learning_rate": 0.0006,
+ "loss": 4.992602348327637,
+ "step": 1030
+ },
+ {
+ "epoch": 14.321397379912664,
+ "grad_norm": 0.014064064249396324,
+ "learning_rate": 0.0006,
+ "loss": 5.020243167877197,
+ "step": 1031
+ },
+ {
+ "epoch": 14.335371179039301,
+ "grad_norm": 0.014658658765256405,
+ "learning_rate": 0.0006,
+ "loss": 4.995723724365234,
+ "step": 1032
+ },
+ {
+ "epoch": 14.34934497816594,
+ "grad_norm": 0.015613174997270107,
+ "learning_rate": 0.0006,
+ "loss": 5.0587310791015625,
+ "step": 1033
+ },
+ {
+ "epoch": 14.363318777292577,
+ "grad_norm": 0.016534067690372467,
+ "learning_rate": 0.0006,
+ "loss": 4.943386077880859,
+ "step": 1034
+ },
+ {
+ "epoch": 14.377292576419213,
+ "grad_norm": 0.016116850078105927,
+ "learning_rate": 0.0006,
+ "loss": 5.117921829223633,
+ "step": 1035
+ },
+ {
+ "epoch": 14.391266375545852,
+ "grad_norm": 0.014806007035076618,
+ "learning_rate": 0.0006,
+ "loss": 5.077454566955566,
+ "step": 1036
+ },
+ {
+ "epoch": 14.405240174672489,
+ "grad_norm": 0.013742697425186634,
+ "learning_rate": 0.0006,
+ "loss": 5.059109687805176,
+ "step": 1037
+ },
+ {
+ "epoch": 14.419213973799126,
+ "grad_norm": 0.01445091050118208,
+ "learning_rate": 0.0006,
+ "loss": 4.97079610824585,
+ "step": 1038
+ },
+ {
+ "epoch": 14.433187772925764,
+ "grad_norm": 0.014255733229219913,
+ "learning_rate": 0.0006,
+ "loss": 5.056376934051514,
+ "step": 1039
+ },
+ {
+ "epoch": 14.447161572052401,
+ "grad_norm": 0.014356110244989395,
+ "learning_rate": 0.0006,
+ "loss": 5.095856666564941,
+ "step": 1040
+ },
+ {
+ "epoch": 14.46113537117904,
+ "grad_norm": 0.015831749886274338,
+ "learning_rate": 0.0006,
+ "loss": 5.063546180725098,
+ "step": 1041
+ },
+ {
+ "epoch": 14.475109170305677,
+ "grad_norm": 0.01640896685421467,
+ "learning_rate": 0.0006,
+ "loss": 5.038229942321777,
+ "step": 1042
+ },
+ {
+ "epoch": 14.489082969432314,
+ "grad_norm": 0.017240721732378006,
+ "learning_rate": 0.0006,
+ "loss": 4.989873886108398,
+ "step": 1043
+ },
+ {
+ "epoch": 14.503056768558952,
+ "grad_norm": 0.020683523267507553,
+ "learning_rate": 0.0006,
+ "loss": 4.945237159729004,
+ "step": 1044
+ },
+ {
+ "epoch": 14.51703056768559,
+ "grad_norm": 0.02303830161690712,
+ "learning_rate": 0.0006,
+ "loss": 5.039238929748535,
+ "step": 1045
+ },
+ {
+ "epoch": 14.531004366812226,
+ "grad_norm": 0.02372933179140091,
+ "learning_rate": 0.0006,
+ "loss": 5.005778789520264,
+ "step": 1046
+ },
+ {
+ "epoch": 14.544978165938865,
+ "grad_norm": 0.02596980705857277,
+ "learning_rate": 0.0006,
+ "loss": 5.047043800354004,
+ "step": 1047
+ },
+ {
+ "epoch": 14.558951965065502,
+ "grad_norm": 0.03045373223721981,
+ "learning_rate": 0.0006,
+ "loss": 5.07199764251709,
+ "step": 1048
+ },
+ {
+ "epoch": 14.57292576419214,
+ "grad_norm": 0.033439598977565765,
+ "learning_rate": 0.0006,
+ "loss": 5.008430480957031,
+ "step": 1049
+ },
+ {
+ "epoch": 14.586899563318777,
+ "grad_norm": 0.027381205931305885,
+ "learning_rate": 0.0006,
+ "loss": 5.10188102722168,
+ "step": 1050
+ },
+ {
+ "epoch": 14.600873362445414,
+ "grad_norm": 0.029322044923901558,
+ "learning_rate": 0.0006,
+ "loss": 5.0966997146606445,
+ "step": 1051
+ },
+ {
+ "epoch": 14.614847161572053,
+ "grad_norm": 0.02821085974574089,
+ "learning_rate": 0.0006,
+ "loss": 4.9426422119140625,
+ "step": 1052
+ },
+ {
+ "epoch": 14.62882096069869,
+ "grad_norm": 0.03356366604566574,
+ "learning_rate": 0.0006,
+ "loss": 5.083784580230713,
+ "step": 1053
+ },
+ {
+ "epoch": 14.642794759825328,
+ "grad_norm": 0.03138979524374008,
+ "learning_rate": 0.0006,
+ "loss": 5.063078880310059,
+ "step": 1054
+ },
+ {
+ "epoch": 14.656768558951965,
+ "grad_norm": 0.02739642933011055,
+ "learning_rate": 0.0006,
+ "loss": 5.0485687255859375,
+ "step": 1055
+ },
+ {
+ "epoch": 14.670742358078602,
+ "grad_norm": 0.026842772960662842,
+ "learning_rate": 0.0006,
+ "loss": 5.040291786193848,
+ "step": 1056
+ },
+ {
+ "epoch": 14.68471615720524,
+ "grad_norm": 0.03270867466926575,
+ "learning_rate": 0.0006,
+ "loss": 5.118686199188232,
+ "step": 1057
+ },
+ {
+ "epoch": 14.698689956331878,
+ "grad_norm": 0.027592353522777557,
+ "learning_rate": 0.0006,
+ "loss": 5.05806303024292,
+ "step": 1058
+ },
+ {
+ "epoch": 14.712663755458514,
+ "grad_norm": 0.029291553422808647,
+ "learning_rate": 0.0006,
+ "loss": 5.048781394958496,
+ "step": 1059
+ },
+ {
+ "epoch": 14.726637554585153,
+ "grad_norm": 0.027100825682282448,
+ "learning_rate": 0.0006,
+ "loss": 5.09072208404541,
+ "step": 1060
+ },
+ {
+ "epoch": 14.74061135371179,
+ "grad_norm": 0.02333410270512104,
+ "learning_rate": 0.0006,
+ "loss": 5.0193939208984375,
+ "step": 1061
+ },
+ {
+ "epoch": 14.754585152838429,
+ "grad_norm": 0.022513234987854958,
+ "learning_rate": 0.0006,
+ "loss": 4.96424674987793,
+ "step": 1062
+ },
+ {
+ "epoch": 14.768558951965066,
+ "grad_norm": 0.01899099536240101,
+ "learning_rate": 0.0006,
+ "loss": 4.982175350189209,
+ "step": 1063
+ },
+ {
+ "epoch": 14.782532751091702,
+ "grad_norm": 0.0192430317401886,
+ "learning_rate": 0.0006,
+ "loss": 5.094186782836914,
+ "step": 1064
+ },
+ {
+ "epoch": 14.796506550218341,
+ "grad_norm": 0.02133013680577278,
+ "learning_rate": 0.0006,
+ "loss": 5.04017448425293,
+ "step": 1065
+ },
+ {
+ "epoch": 14.810480349344978,
+ "grad_norm": 0.019499894231557846,
+ "learning_rate": 0.0006,
+ "loss": 5.058502197265625,
+ "step": 1066
+ },
+ {
+ "epoch": 14.824454148471617,
+ "grad_norm": 0.016153644770383835,
+ "learning_rate": 0.0006,
+ "loss": 5.006505012512207,
+ "step": 1067
+ },
+ {
+ "epoch": 14.838427947598253,
+ "grad_norm": 0.016342492774128914,
+ "learning_rate": 0.0006,
+ "loss": 4.998578071594238,
+ "step": 1068
+ },
+ {
+ "epoch": 14.85240174672489,
+ "grad_norm": 0.014268498867750168,
+ "learning_rate": 0.0006,
+ "loss": 5.099554538726807,
+ "step": 1069
+ },
+ {
+ "epoch": 14.866375545851529,
+ "grad_norm": 0.014293930493295193,
+ "learning_rate": 0.0006,
+ "loss": 4.929508686065674,
+ "step": 1070
+ },
+ {
+ "epoch": 14.880349344978166,
+ "grad_norm": 0.013110974803566933,
+ "learning_rate": 0.0006,
+ "loss": 4.981067657470703,
+ "step": 1071
+ },
+ {
+ "epoch": 14.894323144104803,
+ "grad_norm": 0.014049272052943707,
+ "learning_rate": 0.0006,
+ "loss": 4.97393274307251,
+ "step": 1072
+ },
+ {
+ "epoch": 14.908296943231441,
+ "grad_norm": 0.011994240805506706,
+ "learning_rate": 0.0006,
+ "loss": 4.9473419189453125,
+ "step": 1073
+ },
+ {
+ "epoch": 14.922270742358078,
+ "grad_norm": 0.013252398930490017,
+ "learning_rate": 0.0006,
+ "loss": 4.965734481811523,
+ "step": 1074
+ },
+ {
+ "epoch": 14.936244541484717,
+ "grad_norm": 0.013359254226088524,
+ "learning_rate": 0.0006,
+ "loss": 4.968218803405762,
+ "step": 1075
+ },
+ {
+ "epoch": 14.950218340611354,
+ "grad_norm": 0.01238918025046587,
+ "learning_rate": 0.0006,
+ "loss": 5.07440710067749,
+ "step": 1076
+ },
+ {
+ "epoch": 14.96419213973799,
+ "grad_norm": 0.014492994174361229,
+ "learning_rate": 0.0006,
+ "loss": 5.0300493240356445,
+ "step": 1077
+ },
+ {
+ "epoch": 14.97816593886463,
+ "grad_norm": 0.01550526823848486,
+ "learning_rate": 0.0006,
+ "loss": 5.031158924102783,
+ "step": 1078
+ },
+ {
+ "epoch": 14.992139737991266,
+ "grad_norm": 0.016029849648475647,
+ "learning_rate": 0.0006,
+ "loss": 5.000946521759033,
+ "step": 1079
+ },
+ {
+ "epoch": 15.0,
+ "grad_norm": 0.0141118373721838,
+ "learning_rate": 0.0006,
+ "loss": 5.079067230224609,
+ "step": 1080
+ },
+ {
+ "epoch": 15.0,
+ "eval_loss": 5.463352203369141,
+ "eval_runtime": 79.3035,
+ "eval_samples_per_second": 30.793,
+ "eval_steps_per_second": 0.971,
+ "step": 1080
+ },
+ {
+ "epoch": 15.013973799126637,
+ "grad_norm": 0.015567110851407051,
+ "learning_rate": 0.0006,
+ "loss": 5.006353855133057,
+ "step": 1081
+ },
+ {
+ "epoch": 15.027947598253276,
+ "grad_norm": 0.017325397580862045,
+ "learning_rate": 0.0006,
+ "loss": 5.022473335266113,
+ "step": 1082
+ },
+ {
+ "epoch": 15.041921397379912,
+ "grad_norm": 0.018784036859869957,
+ "learning_rate": 0.0006,
+ "loss": 4.922494888305664,
+ "step": 1083
+ },
+ {
+ "epoch": 15.055895196506551,
+ "grad_norm": 0.020597755908966064,
+ "learning_rate": 0.0006,
+ "loss": 5.045244216918945,
+ "step": 1084
+ },
+ {
+ "epoch": 15.069868995633188,
+ "grad_norm": 0.023631447926163673,
+ "learning_rate": 0.0006,
+ "loss": 4.916990280151367,
+ "step": 1085
+ },
+ {
+ "epoch": 15.083842794759825,
+ "grad_norm": 0.02467048540711403,
+ "learning_rate": 0.0006,
+ "loss": 5.0631208419799805,
+ "step": 1086
+ },
+ {
+ "epoch": 15.097816593886463,
+ "grad_norm": 0.024088503792881966,
+ "learning_rate": 0.0006,
+ "loss": 4.917226791381836,
+ "step": 1087
+ },
+ {
+ "epoch": 15.1117903930131,
+ "grad_norm": 0.022297589108347893,
+ "learning_rate": 0.0006,
+ "loss": 4.940986156463623,
+ "step": 1088
+ },
+ {
+ "epoch": 15.125764192139737,
+ "grad_norm": 0.02313072420656681,
+ "learning_rate": 0.0006,
+ "loss": 5.013306617736816,
+ "step": 1089
+ },
+ {
+ "epoch": 15.139737991266376,
+ "grad_norm": 0.028538871556520462,
+ "learning_rate": 0.0006,
+ "loss": 4.99459171295166,
+ "step": 1090
+ },
+ {
+ "epoch": 15.153711790393013,
+ "grad_norm": 0.029122058302164078,
+ "learning_rate": 0.0006,
+ "loss": 5.000114440917969,
+ "step": 1091
+ },
+ {
+ "epoch": 15.167685589519651,
+ "grad_norm": 0.022197917103767395,
+ "learning_rate": 0.0006,
+ "loss": 4.945086479187012,
+ "step": 1092
+ },
+ {
+ "epoch": 15.181659388646288,
+ "grad_norm": 0.021084975451231003,
+ "learning_rate": 0.0006,
+ "loss": 4.920244216918945,
+ "step": 1093
+ },
+ {
+ "epoch": 15.195633187772925,
+ "grad_norm": 0.022474950179457664,
+ "learning_rate": 0.0006,
+ "loss": 5.005110740661621,
+ "step": 1094
+ },
+ {
+ "epoch": 15.209606986899564,
+ "grad_norm": 0.02403559722006321,
+ "learning_rate": 0.0006,
+ "loss": 5.04328727722168,
+ "step": 1095
+ },
+ {
+ "epoch": 15.2235807860262,
+ "grad_norm": 0.026955733075737953,
+ "learning_rate": 0.0006,
+ "loss": 4.962291717529297,
+ "step": 1096
+ },
+ {
+ "epoch": 15.237554585152838,
+ "grad_norm": 0.024654097855091095,
+ "learning_rate": 0.0006,
+ "loss": 5.0805840492248535,
+ "step": 1097
+ },
+ {
+ "epoch": 15.251528384279476,
+ "grad_norm": 0.01992264948785305,
+ "learning_rate": 0.0006,
+ "loss": 5.044548034667969,
+ "step": 1098
+ },
+ {
+ "epoch": 15.265502183406113,
+ "grad_norm": 0.01696479320526123,
+ "learning_rate": 0.0006,
+ "loss": 4.912850856781006,
+ "step": 1099
+ },
+ {
+ "epoch": 15.279475982532752,
+ "grad_norm": 0.017840469256043434,
+ "learning_rate": 0.0006,
+ "loss": 5.0556793212890625,
+ "step": 1100
+ },
+ {
+ "epoch": 15.293449781659389,
+ "grad_norm": 0.01623854599893093,
+ "learning_rate": 0.0006,
+ "loss": 5.040976524353027,
+ "step": 1101
+ },
+ {
+ "epoch": 15.307423580786025,
+ "grad_norm": 0.016552211716771126,
+ "learning_rate": 0.0006,
+ "loss": 4.971366882324219,
+ "step": 1102
+ },
+ {
+ "epoch": 15.321397379912664,
+ "grad_norm": 0.015316464938223362,
+ "learning_rate": 0.0006,
+ "loss": 4.952664375305176,
+ "step": 1103
+ },
+ {
+ "epoch": 15.335371179039301,
+ "grad_norm": 0.01712590828537941,
+ "learning_rate": 0.0006,
+ "loss": 4.969407081604004,
+ "step": 1104
+ },
+ {
+ "epoch": 15.34934497816594,
+ "grad_norm": 0.015214908868074417,
+ "learning_rate": 0.0006,
+ "loss": 4.9463911056518555,
+ "step": 1105
+ },
+ {
+ "epoch": 15.363318777292577,
+ "grad_norm": 0.014866896905004978,
+ "learning_rate": 0.0006,
+ "loss": 5.071913242340088,
+ "step": 1106
+ },
+ {
+ "epoch": 15.377292576419213,
+ "grad_norm": 0.01533361803740263,
+ "learning_rate": 0.0006,
+ "loss": 5.0071916580200195,
+ "step": 1107
+ },
+ {
+ "epoch": 15.391266375545852,
+ "grad_norm": 0.014022483490407467,
+ "learning_rate": 0.0006,
+ "loss": 4.954067707061768,
+ "step": 1108
+ },
+ {
+ "epoch": 15.405240174672489,
+ "grad_norm": 0.015019413083791733,
+ "learning_rate": 0.0006,
+ "loss": 4.9905900955200195,
+ "step": 1109
+ },
+ {
+ "epoch": 15.419213973799126,
+ "grad_norm": 0.015474580228328705,
+ "learning_rate": 0.0006,
+ "loss": 4.999631881713867,
+ "step": 1110
+ },
+ {
+ "epoch": 15.433187772925764,
+ "grad_norm": 0.020479638129472733,
+ "learning_rate": 0.0006,
+ "loss": 5.0200700759887695,
+ "step": 1111
+ },
+ {
+ "epoch": 15.447161572052401,
+ "grad_norm": 0.0237763449549675,
+ "learning_rate": 0.0006,
+ "loss": 5.008729457855225,
+ "step": 1112
+ },
+ {
+ "epoch": 15.46113537117904,
+ "grad_norm": 0.022637002170085907,
+ "learning_rate": 0.0006,
+ "loss": 4.940751075744629,
+ "step": 1113
+ },
+ {
+ "epoch": 15.475109170305677,
+ "grad_norm": 0.019718417897820473,
+ "learning_rate": 0.0006,
+ "loss": 4.9788737297058105,
+ "step": 1114
+ },
+ {
+ "epoch": 15.489082969432314,
+ "grad_norm": 0.023200126364827156,
+ "learning_rate": 0.0006,
+ "loss": 4.906790256500244,
+ "step": 1115
+ },
+ {
+ "epoch": 15.503056768558952,
+ "grad_norm": 0.025976872071623802,
+ "learning_rate": 0.0006,
+ "loss": 5.002409934997559,
+ "step": 1116
+ },
+ {
+ "epoch": 15.51703056768559,
+ "grad_norm": 0.023092452436685562,
+ "learning_rate": 0.0006,
+ "loss": 4.90887451171875,
+ "step": 1117
+ },
+ {
+ "epoch": 15.531004366812226,
+ "grad_norm": 0.01955891214311123,
+ "learning_rate": 0.0006,
+ "loss": 4.995199203491211,
+ "step": 1118
+ },
+ {
+ "epoch": 15.544978165938865,
+ "grad_norm": 0.02019096165895462,
+ "learning_rate": 0.0006,
+ "loss": 4.908695697784424,
+ "step": 1119
+ },
+ {
+ "epoch": 15.558951965065502,
+ "grad_norm": 0.019376836717128754,
+ "learning_rate": 0.0006,
+ "loss": 4.872652053833008,
+ "step": 1120
+ },
+ {
+ "epoch": 15.57292576419214,
+ "grad_norm": 0.019068755209445953,
+ "learning_rate": 0.0006,
+ "loss": 4.952585220336914,
+ "step": 1121
+ },
+ {
+ "epoch": 15.586899563318777,
+ "grad_norm": 0.016856160014867783,
+ "learning_rate": 0.0006,
+ "loss": 4.927958965301514,
+ "step": 1122
+ },
+ {
+ "epoch": 15.600873362445414,
+ "grad_norm": 0.017544621601700783,
+ "learning_rate": 0.0006,
+ "loss": 4.922074794769287,
+ "step": 1123
+ },
+ {
+ "epoch": 15.614847161572053,
+ "grad_norm": 0.01814766600728035,
+ "learning_rate": 0.0006,
+ "loss": 4.884115219116211,
+ "step": 1124
+ },
+ {
+ "epoch": 15.62882096069869,
+ "grad_norm": 0.01756724901497364,
+ "learning_rate": 0.0006,
+ "loss": 4.969221115112305,
+ "step": 1125
+ },
+ {
+ "epoch": 15.642794759825328,
+ "grad_norm": 0.016586478799581528,
+ "learning_rate": 0.0006,
+ "loss": 4.897098541259766,
+ "step": 1126
+ },
+ {
+ "epoch": 15.656768558951965,
+ "grad_norm": 0.0165987778455019,
+ "learning_rate": 0.0006,
+ "loss": 4.8751044273376465,
+ "step": 1127
+ },
+ {
+ "epoch": 15.670742358078602,
+ "grad_norm": 0.019033856689929962,
+ "learning_rate": 0.0006,
+ "loss": 4.894170761108398,
+ "step": 1128
+ },
+ {
+ "epoch": 15.68471615720524,
+ "grad_norm": 0.016953010112047195,
+ "learning_rate": 0.0006,
+ "loss": 4.918987274169922,
+ "step": 1129
+ },
+ {
+ "epoch": 15.698689956331878,
+ "grad_norm": 0.016293920576572418,
+ "learning_rate": 0.0006,
+ "loss": 4.949659824371338,
+ "step": 1130
+ },
+ {
+ "epoch": 15.712663755458514,
+ "grad_norm": 0.019263453781604767,
+ "learning_rate": 0.0006,
+ "loss": 4.923409461975098,
+ "step": 1131
+ },
+ {
+ "epoch": 15.726637554585153,
+ "grad_norm": 0.01789483241736889,
+ "learning_rate": 0.0006,
+ "loss": 4.894615173339844,
+ "step": 1132
+ },
+ {
+ "epoch": 15.74061135371179,
+ "grad_norm": 0.01610061153769493,
+ "learning_rate": 0.0006,
+ "loss": 4.9311604499816895,
+ "step": 1133
+ },
+ {
+ "epoch": 15.754585152838429,
+ "grad_norm": 0.01674770377576351,
+ "learning_rate": 0.0006,
+ "loss": 4.938290596008301,
+ "step": 1134
+ },
+ {
+ "epoch": 15.768558951965066,
+ "grad_norm": 0.01579914800822735,
+ "learning_rate": 0.0006,
+ "loss": 4.936288833618164,
+ "step": 1135
+ },
+ {
+ "epoch": 15.782532751091702,
+ "grad_norm": 0.015871549025177956,
+ "learning_rate": 0.0006,
+ "loss": 4.970236301422119,
+ "step": 1136
+ },
+ {
+ "epoch": 15.796506550218341,
+ "grad_norm": 0.014079514890909195,
+ "learning_rate": 0.0006,
+ "loss": 4.935883045196533,
+ "step": 1137
+ },
+ {
+ "epoch": 15.810480349344978,
+ "grad_norm": 0.015348718501627445,
+ "learning_rate": 0.0006,
+ "loss": 4.919315814971924,
+ "step": 1138
+ },
+ {
+ "epoch": 15.824454148471617,
+ "grad_norm": 0.0189143568277359,
+ "learning_rate": 0.0006,
+ "loss": 4.972513198852539,
+ "step": 1139
+ },
+ {
+ "epoch": 15.838427947598253,
+ "grad_norm": 0.021286798641085625,
+ "learning_rate": 0.0006,
+ "loss": 4.89121675491333,
+ "step": 1140
+ },
+ {
+ "epoch": 15.85240174672489,
+ "grad_norm": 0.0212328489869833,
+ "learning_rate": 0.0006,
+ "loss": 4.895394802093506,
+ "step": 1141
+ },
+ {
+ "epoch": 15.866375545851529,
+ "grad_norm": 0.020312150940299034,
+ "learning_rate": 0.0006,
+ "loss": 4.943729400634766,
+ "step": 1142
+ },
+ {
+ "epoch": 15.880349344978166,
+ "grad_norm": 0.02121325396001339,
+ "learning_rate": 0.0006,
+ "loss": 4.9322638511657715,
+ "step": 1143
+ },
+ {
+ "epoch": 15.894323144104803,
+ "grad_norm": 0.02030770666897297,
+ "learning_rate": 0.0006,
+ "loss": 4.8969550132751465,
+ "step": 1144
+ },
+ {
+ "epoch": 15.908296943231441,
+ "grad_norm": 0.02408582717180252,
+ "learning_rate": 0.0006,
+ "loss": 4.94023323059082,
+ "step": 1145
+ },
+ {
+ "epoch": 15.922270742358078,
+ "grad_norm": 0.025230303406715393,
+ "learning_rate": 0.0006,
+ "loss": 4.964077472686768,
+ "step": 1146
+ },
+ {
+ "epoch": 15.936244541484717,
+ "grad_norm": 0.025903481990098953,
+ "learning_rate": 0.0006,
+ "loss": 4.9319024085998535,
+ "step": 1147
+ },
+ {
+ "epoch": 15.950218340611354,
+ "grad_norm": 0.02517629973590374,
+ "learning_rate": 0.0006,
+ "loss": 4.9937262535095215,
+ "step": 1148
+ },
+ {
+ "epoch": 15.96419213973799,
+ "grad_norm": 0.023874718695878983,
+ "learning_rate": 0.0006,
+ "loss": 4.916127681732178,
+ "step": 1149
+ },
+ {
+ "epoch": 15.97816593886463,
+ "grad_norm": 0.02531718835234642,
+ "learning_rate": 0.0006,
+ "loss": 4.949875831604004,
+ "step": 1150
+ },
+ {
+ "epoch": 15.992139737991266,
+ "grad_norm": 0.026585623621940613,
+ "learning_rate": 0.0006,
+ "loss": 4.883786201477051,
+ "step": 1151
+ },
+ {
+ "epoch": 16.0,
+ "grad_norm": 0.029191339388489723,
+ "learning_rate": 0.0006,
+ "loss": 4.955547332763672,
+ "step": 1152
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 28800,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 400,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": true,
+ "should_log": false,
+ "should_save": false,
+ "should_training_stop": false
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 4.9113813781315584e+17,
+ "train_batch_size": 8,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1152/training_args.bin b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1152/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..c799bcf4051249aabf4c62fff9d38801df7bc2b5
--- /dev/null
+++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1152/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:40fb25263250c08c3479a6f0ac9fdf1f196e2a3b08a4eb10384786ff9b1903ef
+size 4792
diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1440/chat_template.jinja b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1440/chat_template.jinja
new file mode 100644
index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230
--- /dev/null
+++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1440/chat_template.jinja
@@ -0,0 +1,85 @@
+{%- if tools %}
+ {{- '<|im_start|>system\n' }}
+ {%- if messages[0].role == 'system' %}
+ {{- messages[0].content + '\n\n' }}
+ {%- endif %}
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }}
+ {%- for tool in tools %}
+ {{- "\n" }}
+ {{- tool | tojson }}
+ {%- endfor %}
+ {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }}
+{%- else %}
+ {%- if messages[0].role == 'system' %}
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
+ {%- endif %}
+{%- endif %}
+{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
+{%- for message in messages[::-1] %}
+ {%- set index = (messages|length - 1) - loop.index0 %}
+ {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %}
+ {%- set ns.multi_step_tool = false %}
+ {%- set ns.last_query_index = index %}
+ {%- endif %}
+{%- endfor %}
+{%- for message in messages %}
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
+ {%- elif message.role == "assistant" %}
+ {%- set content = message.content %}
+ {%- set reasoning_content = '' %}
+ {%- if message.reasoning_content is defined and message.reasoning_content is not none %}
+ {%- set reasoning_content = message.reasoning_content %}
+ {%- else %}
+ {%- if '' in message.content %}
+ {%- set content = message.content.split('')[-1].lstrip('\n') %}
+ {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %}
+ {%- endif %}
+ {%- endif %}
+ {%- if loop.index0 > ns.last_query_index %}
+ {%- if loop.last or (not loop.last and reasoning_content) %}
+ {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- if message.tool_calls %}
+ {%- for tool_call in message.tool_calls %}
+ {%- if (loop.first and content) or (not loop.first) %}
+ {{- '\n' }}
+ {%- endif %}
+ {%- if tool_call.function %}
+ {%- set tool_call = tool_call.function %}
+ {%- endif %}
+ {{- '\n{"name": "' }}
+ {{- tool_call.name }}
+ {{- '", "arguments": ' }}
+ {%- if tool_call.arguments is string %}
+ {{- tool_call.arguments }}
+ {%- else %}
+ {{- tool_call.arguments | tojson }}
+ {%- endif %}
+ {{- '}\n' }}
+ {%- endfor %}
+ {%- endif %}
+ {{- '<|im_end|>\n' }}
+ {%- elif message.role == "tool" %}
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
+ {{- '<|im_start|>user' }}
+ {%- endif %}
+ {{- '\n\n' }}
+ {{- message.content }}
+ {{- '\n' }}
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
+ {{- '<|im_end|>\n' }}
+ {%- endif %}
+ {%- endif %}
+{%- endfor %}
+{%- if add_generation_prompt %}
+ {{- '<|im_start|>assistant\n' }}
+ {%- if enable_thinking is defined and enable_thinking is false %}
+ {{- '\n\n\n\n' }}
+ {%- endif %}
+{%- endif %}
\ No newline at end of file
diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1440/config.json b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1440/config.json
new file mode 100644
index 0000000000000000000000000000000000000000..175ee3431793018f23f74824f69554b21afcbd4f
--- /dev/null
+++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1440/config.json
@@ -0,0 +1,32 @@
+{
+ "architectures": [
+ "LlamaForCausalLM"
+ ],
+ "attention_bias": false,
+ "attention_dropout": 0.0,
+ "bos_token_id": null,
+ "dtype": "float32",
+ "eos_token_id": 151645,
+ "head_dim": 128,
+ "hidden_act": "silu",
+ "hidden_size": 512,
+ "initializer_range": 0.02,
+ "intermediate_size": 1536,
+ "max_position_embeddings": 2048,
+ "mlp_bias": false,
+ "model_type": "llama",
+ "num_attention_heads": 4,
+ "num_hidden_layers": 20,
+ "num_key_value_heads": 4,
+ "pad_token_id": 151645,
+ "pretraining_tp": 1,
+ "rms_norm_eps": 1e-06,
+ "rope_parameters": {
+ "rope_theta": 10000.0,
+ "rope_type": "default"
+ },
+ "tie_word_embeddings": true,
+ "transformers_version": "5.5.0",
+ "use_cache": false,
+ "vocab_size": 151674
+}
diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1440/generation_config.json b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1440/generation_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3
--- /dev/null
+++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1440/generation_config.json
@@ -0,0 +1,11 @@
+{
+ "_from_model_config": true,
+ "eos_token_id": [
+ 151645
+ ],
+ "output_attentions": false,
+ "output_hidden_states": false,
+ "pad_token_id": 151645,
+ "transformers_version": "5.5.0",
+ "use_cache": true
+}
diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1440/model.safetensors b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1440/model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..0e36a9e703f0be14f7d30556367636bd1d91e1c5
--- /dev/null
+++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1440/model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:7ab162fa96483f968ae7ded2cf2b482442f54e0da91166a0b597716364e7c064
+size 583362376
diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1440/optimizer.pt b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1440/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..a1eb75fd2c66e5adb2d7586ade67c4886d8f6a99
--- /dev/null
+++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1440/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:80e0b92eff1988e3f1b7e6cc77dcc94de424b81f7d783bcaa3d266463e964d8f
+size 1166837626
diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1440/rng_state_0.pth b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1440/rng_state_0.pth
new file mode 100644
index 0000000000000000000000000000000000000000..b64929d57263eff2fb995a23b08eba35f6f93010
--- /dev/null
+++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1440/rng_state_0.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:63caf9dac92083e4136aebe27ed5f14192e67e2b7c1be28ddb92c9378e30cbec
+size 15024
diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1440/rng_state_1.pth b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1440/rng_state_1.pth
new file mode 100644
index 0000000000000000000000000000000000000000..4c39af5c7e36ba4e1c7db044f83e67f098dcb23a
--- /dev/null
+++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1440/rng_state_1.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:865832ee592ced5b17b41bdfc0ab0aa506de831285ad28234b98fa4dcfae8116
+size 15024
diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1440/rng_state_2.pth b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1440/rng_state_2.pth
new file mode 100644
index 0000000000000000000000000000000000000000..607d15463fe50c2ab955ee306939c8c8e89b2d37
--- /dev/null
+++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1440/rng_state_2.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:7cc13de805c6c2ec23231d41507998e68cb38eb124f7e877be15ee1e574b7c57
+size 15024
diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1440/rng_state_3.pth b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1440/rng_state_3.pth
new file mode 100644
index 0000000000000000000000000000000000000000..dab9765a6a5ab1ac06d19e29f8f7d3e1c80aaedd
--- /dev/null
+++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1440/rng_state_3.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:654b8fd45d1880a532d0df9f8d24dd859c9a20985b2bc449c6d9ebcffb476338
+size 15024
diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1440/scheduler.pt b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1440/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..0d34fe10b069b50b5fecbd4a0affbd8f653b7b9b
--- /dev/null
+++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1440/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:8a84bbc84bb0dd489a2d72fd441cc0b8482e9e06d7cde3ac44b50a389883ce86
+size 1064
diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1440/tokenizer.json b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1440/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..9ea28d9eb325c36bcab4c1a08fc93fd598929f08
--- /dev/null
+++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1440/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:60e99075d371ecbb9f1dde97363f612d45bf0bf88f1a2006d9bfcf1939333c14
+size 11424648
diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1440/tokenizer_config.json b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1440/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..cadd997fc26040d23ad6677fcb39d8d18b01f55e
--- /dev/null
+++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1440/tokenizer_config.json
@@ -0,0 +1,21 @@
+{
+ "add_prefix_space": false,
+ "backend": "tokenizers",
+ "bos_token": null,
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "<|im_end|>",
+ "errors": "replace",
+ "extra_special_tokens": [
+ "<|l2r_pred|>",
+ "<|r2l_pred|>",
+ "<|next_1_pred|>",
+ "<|next_2_pred|>",
+ "<|mask|>"
+ ],
+ "is_local": false,
+ "model_max_length": 131072,
+ "pad_token": "<|im_end|>",
+ "split_special_tokens": false,
+ "tokenizer_class": "Qwen2Tokenizer",
+ "unk_token": null
+}
diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1440/trainer_state.json b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1440/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..61823ff1b5d55f798d5b62cbe1d5f3533f3ec8a1
--- /dev/null
+++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1440/trainer_state.json
@@ -0,0 +1,10266 @@
+{
+ "best_global_step": null,
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 20.0,
+ "eval_steps": 500,
+ "global_step": 1440,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.013973799126637555,
+ "grad_norm": 0.5768774151802063,
+ "learning_rate": 0.0,
+ "loss": 12.050328254699707,
+ "step": 1
+ },
+ {
+ "epoch": 0.02794759825327511,
+ "grad_norm": 0.5784570574760437,
+ "learning_rate": 5.999999999999999e-06,
+ "loss": 12.049251556396484,
+ "step": 2
+ },
+ {
+ "epoch": 0.04192139737991266,
+ "grad_norm": 0.5607187151908875,
+ "learning_rate": 1.1999999999999999e-05,
+ "loss": 11.921792030334473,
+ "step": 3
+ },
+ {
+ "epoch": 0.05589519650655022,
+ "grad_norm": 0.46466919779777527,
+ "learning_rate": 1.7999999999999997e-05,
+ "loss": 11.686628341674805,
+ "step": 4
+ },
+ {
+ "epoch": 0.06986899563318777,
+ "grad_norm": 0.38012945652008057,
+ "learning_rate": 2.3999999999999997e-05,
+ "loss": 11.432823181152344,
+ "step": 5
+ },
+ {
+ "epoch": 0.08384279475982533,
+ "grad_norm": 0.33215776085853577,
+ "learning_rate": 2.9999999999999997e-05,
+ "loss": 11.207542419433594,
+ "step": 6
+ },
+ {
+ "epoch": 0.09781659388646288,
+ "grad_norm": 0.2928776741027832,
+ "learning_rate": 3.5999999999999994e-05,
+ "loss": 11.000097274780273,
+ "step": 7
+ },
+ {
+ "epoch": 0.11179039301310044,
+ "grad_norm": 0.26018962264060974,
+ "learning_rate": 4.2e-05,
+ "loss": 10.825748443603516,
+ "step": 8
+ },
+ {
+ "epoch": 0.125764192139738,
+ "grad_norm": 0.23717083036899567,
+ "learning_rate": 4.7999999999999994e-05,
+ "loss": 10.677122116088867,
+ "step": 9
+ },
+ {
+ "epoch": 0.13973799126637554,
+ "grad_norm": 0.21641355752944946,
+ "learning_rate": 5.399999999999999e-05,
+ "loss": 10.569148063659668,
+ "step": 10
+ },
+ {
+ "epoch": 0.1537117903930131,
+ "grad_norm": 0.1982956826686859,
+ "learning_rate": 5.9999999999999995e-05,
+ "loss": 10.49653434753418,
+ "step": 11
+ },
+ {
+ "epoch": 0.16768558951965065,
+ "grad_norm": 0.18819282948970795,
+ "learning_rate": 6.599999999999999e-05,
+ "loss": 10.432951927185059,
+ "step": 12
+ },
+ {
+ "epoch": 0.18165938864628822,
+ "grad_norm": 0.18322618305683136,
+ "learning_rate": 7.199999999999999e-05,
+ "loss": 10.387115478515625,
+ "step": 13
+ },
+ {
+ "epoch": 0.19563318777292577,
+ "grad_norm": 0.18399226665496826,
+ "learning_rate": 7.8e-05,
+ "loss": 10.33381462097168,
+ "step": 14
+ },
+ {
+ "epoch": 0.2096069868995633,
+ "grad_norm": 0.1833190619945526,
+ "learning_rate": 8.4e-05,
+ "loss": 10.301689147949219,
+ "step": 15
+ },
+ {
+ "epoch": 0.22358078602620088,
+ "grad_norm": 0.18490834534168243,
+ "learning_rate": 8.999999999999999e-05,
+ "loss": 10.240968704223633,
+ "step": 16
+ },
+ {
+ "epoch": 0.23755458515283842,
+ "grad_norm": 0.18358023464679718,
+ "learning_rate": 9.599999999999999e-05,
+ "loss": 10.194717407226562,
+ "step": 17
+ },
+ {
+ "epoch": 0.251528384279476,
+ "grad_norm": 0.1801682710647583,
+ "learning_rate": 0.000102,
+ "loss": 10.14897632598877,
+ "step": 18
+ },
+ {
+ "epoch": 0.26550218340611353,
+ "grad_norm": 0.17899440228939056,
+ "learning_rate": 0.00010799999999999998,
+ "loss": 10.078763961791992,
+ "step": 19
+ },
+ {
+ "epoch": 0.2794759825327511,
+ "grad_norm": 0.17666497826576233,
+ "learning_rate": 0.00011399999999999999,
+ "loss": 10.01095962524414,
+ "step": 20
+ },
+ {
+ "epoch": 0.2934497816593886,
+ "grad_norm": 0.17196154594421387,
+ "learning_rate": 0.00011999999999999999,
+ "loss": 9.951203346252441,
+ "step": 21
+ },
+ {
+ "epoch": 0.3074235807860262,
+ "grad_norm": 0.169203519821167,
+ "learning_rate": 0.00012599999999999997,
+ "loss": 9.871726036071777,
+ "step": 22
+ },
+ {
+ "epoch": 0.32139737991266376,
+ "grad_norm": 0.16500306129455566,
+ "learning_rate": 0.00013199999999999998,
+ "loss": 9.803646087646484,
+ "step": 23
+ },
+ {
+ "epoch": 0.3353711790393013,
+ "grad_norm": 0.16128124296665192,
+ "learning_rate": 0.000138,
+ "loss": 9.728487014770508,
+ "step": 24
+ },
+ {
+ "epoch": 0.34934497816593885,
+ "grad_norm": 0.15904058516025543,
+ "learning_rate": 0.00014399999999999998,
+ "loss": 9.642473220825195,
+ "step": 25
+ },
+ {
+ "epoch": 0.36331877729257644,
+ "grad_norm": 0.15560561418533325,
+ "learning_rate": 0.00015,
+ "loss": 9.56414794921875,
+ "step": 26
+ },
+ {
+ "epoch": 0.377292576419214,
+ "grad_norm": 0.15278612077236176,
+ "learning_rate": 0.000156,
+ "loss": 9.484230041503906,
+ "step": 27
+ },
+ {
+ "epoch": 0.39126637554585153,
+ "grad_norm": 0.15263237059116364,
+ "learning_rate": 0.000162,
+ "loss": 9.383644104003906,
+ "step": 28
+ },
+ {
+ "epoch": 0.4052401746724891,
+ "grad_norm": 0.14945024251937866,
+ "learning_rate": 0.000168,
+ "loss": 9.29840087890625,
+ "step": 29
+ },
+ {
+ "epoch": 0.4192139737991266,
+ "grad_norm": 0.14614646136760712,
+ "learning_rate": 0.00017399999999999997,
+ "loss": 9.208931922912598,
+ "step": 30
+ },
+ {
+ "epoch": 0.4331877729257642,
+ "grad_norm": 0.14295098185539246,
+ "learning_rate": 0.00017999999999999998,
+ "loss": 9.122323989868164,
+ "step": 31
+ },
+ {
+ "epoch": 0.44716157205240176,
+ "grad_norm": 0.13868345320224762,
+ "learning_rate": 0.000186,
+ "loss": 9.03034496307373,
+ "step": 32
+ },
+ {
+ "epoch": 0.4611353711790393,
+ "grad_norm": 0.135064035654068,
+ "learning_rate": 0.00019199999999999998,
+ "loss": 8.92818832397461,
+ "step": 33
+ },
+ {
+ "epoch": 0.47510917030567684,
+ "grad_norm": 0.13097241520881653,
+ "learning_rate": 0.000198,
+ "loss": 8.834627151489258,
+ "step": 34
+ },
+ {
+ "epoch": 0.4890829694323144,
+ "grad_norm": 0.12478776276111603,
+ "learning_rate": 0.000204,
+ "loss": 8.755172729492188,
+ "step": 35
+ },
+ {
+ "epoch": 0.503056768558952,
+ "grad_norm": 0.1214982345700264,
+ "learning_rate": 0.00020999999999999998,
+ "loss": 8.647891998291016,
+ "step": 36
+ },
+ {
+ "epoch": 0.5170305676855895,
+ "grad_norm": 0.11958328634500504,
+ "learning_rate": 0.00021599999999999996,
+ "loss": 8.543274879455566,
+ "step": 37
+ },
+ {
+ "epoch": 0.5310043668122271,
+ "grad_norm": 0.11680915951728821,
+ "learning_rate": 0.00022199999999999998,
+ "loss": 8.44540023803711,
+ "step": 38
+ },
+ {
+ "epoch": 0.5449781659388646,
+ "grad_norm": 0.11491474509239197,
+ "learning_rate": 0.00022799999999999999,
+ "loss": 8.349921226501465,
+ "step": 39
+ },
+ {
+ "epoch": 0.5589519650655022,
+ "grad_norm": 0.11160755902528763,
+ "learning_rate": 0.000234,
+ "loss": 8.26960563659668,
+ "step": 40
+ },
+ {
+ "epoch": 0.5729257641921397,
+ "grad_norm": 0.10840637236833572,
+ "learning_rate": 0.00023999999999999998,
+ "loss": 8.192136764526367,
+ "step": 41
+ },
+ {
+ "epoch": 0.5868995633187772,
+ "grad_norm": 0.10585298389196396,
+ "learning_rate": 0.00024599999999999996,
+ "loss": 8.09546947479248,
+ "step": 42
+ },
+ {
+ "epoch": 0.6008733624454149,
+ "grad_norm": 0.1032542735338211,
+ "learning_rate": 0.00025199999999999995,
+ "loss": 7.991090774536133,
+ "step": 43
+ },
+ {
+ "epoch": 0.6148471615720524,
+ "grad_norm": 0.0986943170428276,
+ "learning_rate": 0.000258,
+ "loss": 7.92384147644043,
+ "step": 44
+ },
+ {
+ "epoch": 0.62882096069869,
+ "grad_norm": 0.09320762753486633,
+ "learning_rate": 0.00026399999999999997,
+ "loss": 7.8478193283081055,
+ "step": 45
+ },
+ {
+ "epoch": 0.6427947598253275,
+ "grad_norm": 0.09168332070112228,
+ "learning_rate": 0.00027,
+ "loss": 7.76185417175293,
+ "step": 46
+ },
+ {
+ "epoch": 0.6567685589519651,
+ "grad_norm": 0.08832395076751709,
+ "learning_rate": 0.000276,
+ "loss": 7.679399490356445,
+ "step": 47
+ },
+ {
+ "epoch": 0.6707423580786026,
+ "grad_norm": 0.08816400915384293,
+ "learning_rate": 0.00028199999999999997,
+ "loss": 7.5789642333984375,
+ "step": 48
+ },
+ {
+ "epoch": 0.6847161572052402,
+ "grad_norm": 0.0861717164516449,
+ "learning_rate": 0.00028799999999999995,
+ "loss": 7.519214630126953,
+ "step": 49
+ },
+ {
+ "epoch": 0.6986899563318777,
+ "grad_norm": 0.08103746175765991,
+ "learning_rate": 0.000294,
+ "loss": 7.4475321769714355,
+ "step": 50
+ },
+ {
+ "epoch": 0.7126637554585152,
+ "grad_norm": 0.07545104622840881,
+ "learning_rate": 0.0003,
+ "loss": 7.401182174682617,
+ "step": 51
+ },
+ {
+ "epoch": 0.7266375545851529,
+ "grad_norm": 0.07042575627565384,
+ "learning_rate": 0.00030599999999999996,
+ "loss": 7.340173244476318,
+ "step": 52
+ },
+ {
+ "epoch": 0.7406113537117904,
+ "grad_norm": 0.06737572699785233,
+ "learning_rate": 0.000312,
+ "loss": 7.265697002410889,
+ "step": 53
+ },
+ {
+ "epoch": 0.754585152838428,
+ "grad_norm": 0.06151146441698074,
+ "learning_rate": 0.000318,
+ "loss": 7.231061935424805,
+ "step": 54
+ },
+ {
+ "epoch": 0.7685589519650655,
+ "grad_norm": 0.05638071522116661,
+ "learning_rate": 0.000324,
+ "loss": 7.195466995239258,
+ "step": 55
+ },
+ {
+ "epoch": 0.7825327510917031,
+ "grad_norm": 0.05084463208913803,
+ "learning_rate": 0.00033,
+ "loss": 7.137603759765625,
+ "step": 56
+ },
+ {
+ "epoch": 0.7965065502183406,
+ "grad_norm": 0.04293261095881462,
+ "learning_rate": 0.000336,
+ "loss": 7.115458965301514,
+ "step": 57
+ },
+ {
+ "epoch": 0.8104803493449781,
+ "grad_norm": 0.036064863204956055,
+ "learning_rate": 0.00034199999999999996,
+ "loss": 7.089038848876953,
+ "step": 58
+ },
+ {
+ "epoch": 0.8244541484716157,
+ "grad_norm": 0.029026754200458527,
+ "learning_rate": 0.00034799999999999995,
+ "loss": 7.0584306716918945,
+ "step": 59
+ },
+ {
+ "epoch": 0.8384279475982532,
+ "grad_norm": 0.02286517433822155,
+ "learning_rate": 0.00035399999999999993,
+ "loss": 7.049450874328613,
+ "step": 60
+ },
+ {
+ "epoch": 0.8524017467248908,
+ "grad_norm": 0.018878433853387833,
+ "learning_rate": 0.00035999999999999997,
+ "loss": 7.008824348449707,
+ "step": 61
+ },
+ {
+ "epoch": 0.8663755458515284,
+ "grad_norm": 0.010773813351988792,
+ "learning_rate": 0.00036599999999999995,
+ "loss": 7.024829864501953,
+ "step": 62
+ },
+ {
+ "epoch": 0.880349344978166,
+ "grad_norm": 0.008794732391834259,
+ "learning_rate": 0.000372,
+ "loss": 7.018399238586426,
+ "step": 63
+ },
+ {
+ "epoch": 0.8943231441048035,
+ "grad_norm": 0.011709796264767647,
+ "learning_rate": 0.00037799999999999997,
+ "loss": 7.0101752281188965,
+ "step": 64
+ },
+ {
+ "epoch": 0.9082969432314411,
+ "grad_norm": 0.012123258784413338,
+ "learning_rate": 0.00038399999999999996,
+ "loss": 7.024317741394043,
+ "step": 65
+ },
+ {
+ "epoch": 0.9222707423580786,
+ "grad_norm": 0.01530320756137371,
+ "learning_rate": 0.00039,
+ "loss": 7.027738571166992,
+ "step": 66
+ },
+ {
+ "epoch": 0.9362445414847161,
+ "grad_norm": 0.01890002004802227,
+ "learning_rate": 0.000396,
+ "loss": 7.024099349975586,
+ "step": 67
+ },
+ {
+ "epoch": 0.9502183406113537,
+ "grad_norm": 0.020684150978922844,
+ "learning_rate": 0.000402,
+ "loss": 7.007943153381348,
+ "step": 68
+ },
+ {
+ "epoch": 0.9641921397379912,
+ "grad_norm": 0.019826196134090424,
+ "learning_rate": 0.000408,
+ "loss": 7.006874084472656,
+ "step": 69
+ },
+ {
+ "epoch": 0.9781659388646288,
+ "grad_norm": 0.019218673929572105,
+ "learning_rate": 0.0004139999999999999,
+ "loss": 7.002897262573242,
+ "step": 70
+ },
+ {
+ "epoch": 0.9921397379912664,
+ "grad_norm": 0.013856411911547184,
+ "learning_rate": 0.00041999999999999996,
+ "loss": 6.993999481201172,
+ "step": 71
+ },
+ {
+ "epoch": 1.0,
+ "grad_norm": 0.010390003211796284,
+ "learning_rate": 0.00042599999999999995,
+ "loss": 7.000683784484863,
+ "step": 72
+ },
+ {
+ "epoch": 1.0,
+ "eval_loss": 7.91140079498291,
+ "eval_runtime": 57.7925,
+ "eval_samples_per_second": 42.255,
+ "eval_steps_per_second": 1.332,
+ "step": 72
+ },
+ {
+ "epoch": 1.0139737991266375,
+ "grad_norm": 0.013802163302898407,
+ "learning_rate": 0.00043199999999999993,
+ "loss": 6.998257637023926,
+ "step": 73
+ },
+ {
+ "epoch": 1.027947598253275,
+ "grad_norm": 0.010109687224030495,
+ "learning_rate": 0.00043799999999999997,
+ "loss": 7.002542018890381,
+ "step": 74
+ },
+ {
+ "epoch": 1.0419213973799126,
+ "grad_norm": 0.008019139990210533,
+ "learning_rate": 0.00044399999999999995,
+ "loss": 6.985941410064697,
+ "step": 75
+ },
+ {
+ "epoch": 1.0558951965065502,
+ "grad_norm": 0.0069672465324401855,
+ "learning_rate": 0.00045,
+ "loss": 6.99245023727417,
+ "step": 76
+ },
+ {
+ "epoch": 1.0698689956331877,
+ "grad_norm": 0.00714076729491353,
+ "learning_rate": 0.00045599999999999997,
+ "loss": 6.959296226501465,
+ "step": 77
+ },
+ {
+ "epoch": 1.0838427947598253,
+ "grad_norm": 0.008408435620367527,
+ "learning_rate": 0.00046199999999999995,
+ "loss": 6.954504489898682,
+ "step": 78
+ },
+ {
+ "epoch": 1.0978165938864628,
+ "grad_norm": 0.008778786286711693,
+ "learning_rate": 0.000468,
+ "loss": 7.0007004737854,
+ "step": 79
+ },
+ {
+ "epoch": 1.1117903930131003,
+ "grad_norm": 0.007753407116979361,
+ "learning_rate": 0.000474,
+ "loss": 6.98183536529541,
+ "step": 80
+ },
+ {
+ "epoch": 1.125764192139738,
+ "grad_norm": 0.01087260153144598,
+ "learning_rate": 0.00047999999999999996,
+ "loss": 6.991066932678223,
+ "step": 81
+ },
+ {
+ "epoch": 1.1397379912663754,
+ "grad_norm": 0.006122548598796129,
+ "learning_rate": 0.000486,
+ "loss": 6.978824138641357,
+ "step": 82
+ },
+ {
+ "epoch": 1.1537117903930132,
+ "grad_norm": 0.0067383465357124805,
+ "learning_rate": 0.0004919999999999999,
+ "loss": 6.961703777313232,
+ "step": 83
+ },
+ {
+ "epoch": 1.1676855895196507,
+ "grad_norm": 0.007922791875898838,
+ "learning_rate": 0.000498,
+ "loss": 6.954860687255859,
+ "step": 84
+ },
+ {
+ "epoch": 1.1816593886462883,
+ "grad_norm": 0.00818922370672226,
+ "learning_rate": 0.0005039999999999999,
+ "loss": 6.945961952209473,
+ "step": 85
+ },
+ {
+ "epoch": 1.1956331877729258,
+ "grad_norm": 0.006791461259126663,
+ "learning_rate": 0.0005099999999999999,
+ "loss": 6.968680381774902,
+ "step": 86
+ },
+ {
+ "epoch": 1.2096069868995634,
+ "grad_norm": 0.005394163075834513,
+ "learning_rate": 0.000516,
+ "loss": 6.956404685974121,
+ "step": 87
+ },
+ {
+ "epoch": 1.223580786026201,
+ "grad_norm": 0.004500220064073801,
+ "learning_rate": 0.000522,
+ "loss": 6.96998405456543,
+ "step": 88
+ },
+ {
+ "epoch": 1.2375545851528384,
+ "grad_norm": 0.008699574507772923,
+ "learning_rate": 0.0005279999999999999,
+ "loss": 6.955585479736328,
+ "step": 89
+ },
+ {
+ "epoch": 1.251528384279476,
+ "grad_norm": 0.005889365915209055,
+ "learning_rate": 0.000534,
+ "loss": 6.9708051681518555,
+ "step": 90
+ },
+ {
+ "epoch": 1.2655021834061135,
+ "grad_norm": 0.009646810591220856,
+ "learning_rate": 0.00054,
+ "loss": 6.962608814239502,
+ "step": 91
+ },
+ {
+ "epoch": 1.279475982532751,
+ "grad_norm": 0.004517171997576952,
+ "learning_rate": 0.0005459999999999999,
+ "loss": 6.957627296447754,
+ "step": 92
+ },
+ {
+ "epoch": 1.2934497816593886,
+ "grad_norm": 0.006339677143841982,
+ "learning_rate": 0.000552,
+ "loss": 6.963141441345215,
+ "step": 93
+ },
+ {
+ "epoch": 1.3074235807860262,
+ "grad_norm": 0.005387383047491312,
+ "learning_rate": 0.000558,
+ "loss": 6.934465408325195,
+ "step": 94
+ },
+ {
+ "epoch": 1.3213973799126637,
+ "grad_norm": 0.005252422299236059,
+ "learning_rate": 0.0005639999999999999,
+ "loss": 6.956108093261719,
+ "step": 95
+ },
+ {
+ "epoch": 1.3353711790393012,
+ "grad_norm": 0.008699414320290089,
+ "learning_rate": 0.00057,
+ "loss": 6.973369598388672,
+ "step": 96
+ },
+ {
+ "epoch": 1.3493449781659388,
+ "grad_norm": 0.003947409335523844,
+ "learning_rate": 0.0005759999999999999,
+ "loss": 6.950225830078125,
+ "step": 97
+ },
+ {
+ "epoch": 1.3633187772925766,
+ "grad_norm": 0.00768681475892663,
+ "learning_rate": 0.0005819999999999999,
+ "loss": 6.964739799499512,
+ "step": 98
+ },
+ {
+ "epoch": 1.3772925764192139,
+ "grad_norm": 0.006563829258084297,
+ "learning_rate": 0.000588,
+ "loss": 6.972421169281006,
+ "step": 99
+ },
+ {
+ "epoch": 1.3912663755458516,
+ "grad_norm": 0.007606555242091417,
+ "learning_rate": 0.0005939999999999999,
+ "loss": 6.958213806152344,
+ "step": 100
+ },
+ {
+ "epoch": 1.405240174672489,
+ "grad_norm": 0.006250880658626556,
+ "learning_rate": 0.0006,
+ "loss": 6.962296485900879,
+ "step": 101
+ },
+ {
+ "epoch": 1.4192139737991267,
+ "grad_norm": 0.0073033408261835575,
+ "learning_rate": 0.0006,
+ "loss": 6.969705104827881,
+ "step": 102
+ },
+ {
+ "epoch": 1.4331877729257643,
+ "grad_norm": 0.0054510910995304585,
+ "learning_rate": 0.0006,
+ "loss": 6.9652862548828125,
+ "step": 103
+ },
+ {
+ "epoch": 1.4471615720524018,
+ "grad_norm": 0.004895673133432865,
+ "learning_rate": 0.0006,
+ "loss": 6.956023693084717,
+ "step": 104
+ },
+ {
+ "epoch": 1.4611353711790394,
+ "grad_norm": 0.005306515377014875,
+ "learning_rate": 0.0006,
+ "loss": 6.963231086730957,
+ "step": 105
+ },
+ {
+ "epoch": 1.475109170305677,
+ "grad_norm": 0.004742146469652653,
+ "learning_rate": 0.0006,
+ "loss": 6.9582061767578125,
+ "step": 106
+ },
+ {
+ "epoch": 1.4890829694323144,
+ "grad_norm": 0.004220000468194485,
+ "learning_rate": 0.0006,
+ "loss": 6.972961902618408,
+ "step": 107
+ },
+ {
+ "epoch": 1.503056768558952,
+ "grad_norm": 0.0042153168469667435,
+ "learning_rate": 0.0006,
+ "loss": 6.961165904998779,
+ "step": 108
+ },
+ {
+ "epoch": 1.5170305676855895,
+ "grad_norm": 0.005165040958672762,
+ "learning_rate": 0.0006,
+ "loss": 6.95018196105957,
+ "step": 109
+ },
+ {
+ "epoch": 1.531004366812227,
+ "grad_norm": 0.003930266480892897,
+ "learning_rate": 0.0006,
+ "loss": 6.9651689529418945,
+ "step": 110
+ },
+ {
+ "epoch": 1.5449781659388646,
+ "grad_norm": 0.004465071018785238,
+ "learning_rate": 0.0006,
+ "loss": 6.960553169250488,
+ "step": 111
+ },
+ {
+ "epoch": 1.5589519650655022,
+ "grad_norm": 0.004888612311333418,
+ "learning_rate": 0.0006,
+ "loss": 6.967258930206299,
+ "step": 112
+ },
+ {
+ "epoch": 1.5729257641921397,
+ "grad_norm": 0.007497863378375769,
+ "learning_rate": 0.0006,
+ "loss": 6.954545021057129,
+ "step": 113
+ },
+ {
+ "epoch": 1.5868995633187772,
+ "grad_norm": 0.005659648682922125,
+ "learning_rate": 0.0006,
+ "loss": 6.962299823760986,
+ "step": 114
+ },
+ {
+ "epoch": 1.600873362445415,
+ "grad_norm": 0.006641396787017584,
+ "learning_rate": 0.0006,
+ "loss": 6.952983856201172,
+ "step": 115
+ },
+ {
+ "epoch": 1.6148471615720523,
+ "grad_norm": 0.009715928696095943,
+ "learning_rate": 0.0006,
+ "loss": 6.958622932434082,
+ "step": 116
+ },
+ {
+ "epoch": 1.62882096069869,
+ "grad_norm": 0.012789523229002953,
+ "learning_rate": 0.0006,
+ "loss": 6.932043075561523,
+ "step": 117
+ },
+ {
+ "epoch": 1.6427947598253274,
+ "grad_norm": 0.021673237904906273,
+ "learning_rate": 0.0006,
+ "loss": 6.931290149688721,
+ "step": 118
+ },
+ {
+ "epoch": 1.6567685589519652,
+ "grad_norm": 0.16024991869926453,
+ "learning_rate": 0.0006,
+ "loss": 6.979497909545898,
+ "step": 119
+ },
+ {
+ "epoch": 1.6707423580786025,
+ "grad_norm": 0.2434523105621338,
+ "learning_rate": 0.0006,
+ "loss": 7.185218811035156,
+ "step": 120
+ },
+ {
+ "epoch": 1.6847161572052403,
+ "grad_norm": 0.39252421259880066,
+ "learning_rate": 0.0006,
+ "loss": 7.137226581573486,
+ "step": 121
+ },
+ {
+ "epoch": 1.6986899563318776,
+ "grad_norm": 0.10565312206745148,
+ "learning_rate": 0.0006,
+ "loss": 6.984494209289551,
+ "step": 122
+ },
+ {
+ "epoch": 1.7126637554585153,
+ "grad_norm": 0.1033431813120842,
+ "learning_rate": 0.0006,
+ "loss": 7.001287937164307,
+ "step": 123
+ },
+ {
+ "epoch": 1.726637554585153,
+ "grad_norm": 0.14875490963459015,
+ "learning_rate": 0.0006,
+ "loss": 7.051436901092529,
+ "step": 124
+ },
+ {
+ "epoch": 1.7406113537117904,
+ "grad_norm": 0.1374562531709671,
+ "learning_rate": 0.0006,
+ "loss": 7.0293121337890625,
+ "step": 125
+ },
+ {
+ "epoch": 1.754585152838428,
+ "grad_norm": 0.0910211056470871,
+ "learning_rate": 0.0006,
+ "loss": 7.007458686828613,
+ "step": 126
+ },
+ {
+ "epoch": 1.7685589519650655,
+ "grad_norm": 0.021633487194776535,
+ "learning_rate": 0.0006,
+ "loss": 6.996816635131836,
+ "step": 127
+ },
+ {
+ "epoch": 1.782532751091703,
+ "grad_norm": 0.08692754060029984,
+ "learning_rate": 0.0006,
+ "loss": 6.998541831970215,
+ "step": 128
+ },
+ {
+ "epoch": 1.7965065502183406,
+ "grad_norm": 0.12613292038440704,
+ "learning_rate": 0.0006,
+ "loss": 6.9970703125,
+ "step": 129
+ },
+ {
+ "epoch": 1.8104803493449781,
+ "grad_norm": 0.09462810307741165,
+ "learning_rate": 0.0006,
+ "loss": 6.997045040130615,
+ "step": 130
+ },
+ {
+ "epoch": 1.8244541484716157,
+ "grad_norm": 0.03534962609410286,
+ "learning_rate": 0.0006,
+ "loss": 6.9738450050354,
+ "step": 131
+ },
+ {
+ "epoch": 1.8384279475982532,
+ "grad_norm": 0.033839885145425797,
+ "learning_rate": 0.0006,
+ "loss": 6.946001052856445,
+ "step": 132
+ },
+ {
+ "epoch": 1.8524017467248908,
+ "grad_norm": 0.06263269484043121,
+ "learning_rate": 0.0006,
+ "loss": 6.968981742858887,
+ "step": 133
+ },
+ {
+ "epoch": 1.8663755458515285,
+ "grad_norm": 0.07117997109889984,
+ "learning_rate": 0.0006,
+ "loss": 6.981535911560059,
+ "step": 134
+ },
+ {
+ "epoch": 1.8803493449781659,
+ "grad_norm": 0.06336277723312378,
+ "learning_rate": 0.0006,
+ "loss": 6.976946830749512,
+ "step": 135
+ },
+ {
+ "epoch": 1.8943231441048036,
+ "grad_norm": 0.047467734664678574,
+ "learning_rate": 0.0006,
+ "loss": 6.961673736572266,
+ "step": 136
+ },
+ {
+ "epoch": 1.908296943231441,
+ "grad_norm": 0.018787868320941925,
+ "learning_rate": 0.0006,
+ "loss": 6.948919296264648,
+ "step": 137
+ },
+ {
+ "epoch": 1.9222707423580787,
+ "grad_norm": 0.025909408926963806,
+ "learning_rate": 0.0006,
+ "loss": 6.976190567016602,
+ "step": 138
+ },
+ {
+ "epoch": 1.936244541484716,
+ "grad_norm": 0.052842333912849426,
+ "learning_rate": 0.0006,
+ "loss": 6.971207618713379,
+ "step": 139
+ },
+ {
+ "epoch": 1.9502183406113538,
+ "grad_norm": 0.06421888619661331,
+ "learning_rate": 0.0006,
+ "loss": 6.979109764099121,
+ "step": 140
+ },
+ {
+ "epoch": 1.9641921397379911,
+ "grad_norm": 0.055596910417079926,
+ "learning_rate": 0.0006,
+ "loss": 6.972118854522705,
+ "step": 141
+ },
+ {
+ "epoch": 1.9781659388646289,
+ "grad_norm": 0.03267091140151024,
+ "learning_rate": 0.0006,
+ "loss": 6.947260856628418,
+ "step": 142
+ },
+ {
+ "epoch": 1.9921397379912664,
+ "grad_norm": 0.008032613433897495,
+ "learning_rate": 0.0006,
+ "loss": 6.936816215515137,
+ "step": 143
+ },
+ {
+ "epoch": 2.0,
+ "grad_norm": 0.020184025168418884,
+ "learning_rate": 0.0006,
+ "loss": 6.929848670959473,
+ "step": 144
+ },
+ {
+ "epoch": 2.0,
+ "eval_loss": 7.836523056030273,
+ "eval_runtime": 58.5622,
+ "eval_samples_per_second": 41.699,
+ "eval_steps_per_second": 1.315,
+ "step": 144
+ },
+ {
+ "epoch": 2.0139737991266378,
+ "grad_norm": 0.031980376690626144,
+ "learning_rate": 0.0006,
+ "loss": 6.968081474304199,
+ "step": 145
+ },
+ {
+ "epoch": 2.027947598253275,
+ "grad_norm": 0.03802533075213432,
+ "learning_rate": 0.0006,
+ "loss": 6.95728874206543,
+ "step": 146
+ },
+ {
+ "epoch": 2.041921397379913,
+ "grad_norm": 0.03796226158738136,
+ "learning_rate": 0.0006,
+ "loss": 6.96516752243042,
+ "step": 147
+ },
+ {
+ "epoch": 2.05589519650655,
+ "grad_norm": 0.02621951326727867,
+ "learning_rate": 0.0006,
+ "loss": 6.956058502197266,
+ "step": 148
+ },
+ {
+ "epoch": 2.069868995633188,
+ "grad_norm": 0.013292718678712845,
+ "learning_rate": 0.0006,
+ "loss": 6.9783935546875,
+ "step": 149
+ },
+ {
+ "epoch": 2.0838427947598253,
+ "grad_norm": 0.009446968324482441,
+ "learning_rate": 0.0006,
+ "loss": 6.96873664855957,
+ "step": 150
+ },
+ {
+ "epoch": 2.097816593886463,
+ "grad_norm": 0.021703092381358147,
+ "learning_rate": 0.0006,
+ "loss": 6.959234237670898,
+ "step": 151
+ },
+ {
+ "epoch": 2.1117903930131003,
+ "grad_norm": 0.027765290811657906,
+ "learning_rate": 0.0006,
+ "loss": 6.968524932861328,
+ "step": 152
+ },
+ {
+ "epoch": 2.125764192139738,
+ "grad_norm": 0.028761165216565132,
+ "learning_rate": 0.0006,
+ "loss": 6.954620838165283,
+ "step": 153
+ },
+ {
+ "epoch": 2.1397379912663754,
+ "grad_norm": 0.019729383289813995,
+ "learning_rate": 0.0006,
+ "loss": 6.949009895324707,
+ "step": 154
+ },
+ {
+ "epoch": 2.153711790393013,
+ "grad_norm": 0.008990089409053326,
+ "learning_rate": 0.0006,
+ "loss": 6.956226348876953,
+ "step": 155
+ },
+ {
+ "epoch": 2.1676855895196505,
+ "grad_norm": 0.00549644511193037,
+ "learning_rate": 0.0006,
+ "loss": 6.957143783569336,
+ "step": 156
+ },
+ {
+ "epoch": 2.1816593886462883,
+ "grad_norm": 0.014436670579016209,
+ "learning_rate": 0.0006,
+ "loss": 6.954257011413574,
+ "step": 157
+ },
+ {
+ "epoch": 2.1956331877729256,
+ "grad_norm": 0.01990494504570961,
+ "learning_rate": 0.0006,
+ "loss": 6.942131519317627,
+ "step": 158
+ },
+ {
+ "epoch": 2.2096069868995634,
+ "grad_norm": 0.019155776128172874,
+ "learning_rate": 0.0006,
+ "loss": 6.946969985961914,
+ "step": 159
+ },
+ {
+ "epoch": 2.2235807860262007,
+ "grad_norm": 0.014970783144235611,
+ "learning_rate": 0.0006,
+ "loss": 6.957938194274902,
+ "step": 160
+ },
+ {
+ "epoch": 2.2375545851528384,
+ "grad_norm": 0.009875659830868244,
+ "learning_rate": 0.0006,
+ "loss": 6.953309059143066,
+ "step": 161
+ },
+ {
+ "epoch": 2.251528384279476,
+ "grad_norm": 0.004792087245732546,
+ "learning_rate": 0.0006,
+ "loss": 6.975146770477295,
+ "step": 162
+ },
+ {
+ "epoch": 2.2655021834061135,
+ "grad_norm": 0.009936527349054813,
+ "learning_rate": 0.0006,
+ "loss": 6.961674690246582,
+ "step": 163
+ },
+ {
+ "epoch": 2.279475982532751,
+ "grad_norm": 0.015100127086043358,
+ "learning_rate": 0.0006,
+ "loss": 6.968759536743164,
+ "step": 164
+ },
+ {
+ "epoch": 2.2934497816593886,
+ "grad_norm": 0.01449542585760355,
+ "learning_rate": 0.0006,
+ "loss": 6.966510772705078,
+ "step": 165
+ },
+ {
+ "epoch": 2.3074235807860264,
+ "grad_norm": 0.01073537115007639,
+ "learning_rate": 0.0006,
+ "loss": 6.936373710632324,
+ "step": 166
+ },
+ {
+ "epoch": 2.3213973799126637,
+ "grad_norm": 0.006766272243112326,
+ "learning_rate": 0.0006,
+ "loss": 6.96397590637207,
+ "step": 167
+ },
+ {
+ "epoch": 2.3353711790393015,
+ "grad_norm": 0.005055439658463001,
+ "learning_rate": 0.0006,
+ "loss": 6.961050987243652,
+ "step": 168
+ },
+ {
+ "epoch": 2.349344978165939,
+ "grad_norm": 0.006252622697502375,
+ "learning_rate": 0.0006,
+ "loss": 6.980506420135498,
+ "step": 169
+ },
+ {
+ "epoch": 2.3633187772925766,
+ "grad_norm": 0.008084346540272236,
+ "learning_rate": 0.0006,
+ "loss": 6.9346604347229,
+ "step": 170
+ },
+ {
+ "epoch": 2.377292576419214,
+ "grad_norm": 0.010260380804538727,
+ "learning_rate": 0.0006,
+ "loss": 6.964118957519531,
+ "step": 171
+ },
+ {
+ "epoch": 2.3912663755458516,
+ "grad_norm": 0.010697875171899796,
+ "learning_rate": 0.0006,
+ "loss": 6.941572189331055,
+ "step": 172
+ },
+ {
+ "epoch": 2.405240174672489,
+ "grad_norm": 0.007338830269873142,
+ "learning_rate": 0.0006,
+ "loss": 6.950471878051758,
+ "step": 173
+ },
+ {
+ "epoch": 2.4192139737991267,
+ "grad_norm": 0.005931832827627659,
+ "learning_rate": 0.0006,
+ "loss": 6.962935924530029,
+ "step": 174
+ },
+ {
+ "epoch": 2.433187772925764,
+ "grad_norm": 0.0061243255622684956,
+ "learning_rate": 0.0006,
+ "loss": 6.946418285369873,
+ "step": 175
+ },
+ {
+ "epoch": 2.447161572052402,
+ "grad_norm": 0.007983328774571419,
+ "learning_rate": 0.0006,
+ "loss": 6.959758758544922,
+ "step": 176
+ },
+ {
+ "epoch": 2.461135371179039,
+ "grad_norm": 0.010470672510564327,
+ "learning_rate": 0.0006,
+ "loss": 6.93851375579834,
+ "step": 177
+ },
+ {
+ "epoch": 2.475109170305677,
+ "grad_norm": 0.011147388257086277,
+ "learning_rate": 0.0006,
+ "loss": 6.93258810043335,
+ "step": 178
+ },
+ {
+ "epoch": 2.489082969432314,
+ "grad_norm": 0.01058544497936964,
+ "learning_rate": 0.0006,
+ "loss": 6.932259559631348,
+ "step": 179
+ },
+ {
+ "epoch": 2.503056768558952,
+ "grad_norm": 0.016850745305418968,
+ "learning_rate": 0.0006,
+ "loss": 6.924542427062988,
+ "step": 180
+ },
+ {
+ "epoch": 2.5170305676855893,
+ "grad_norm": 0.023101547732949257,
+ "learning_rate": 0.0006,
+ "loss": 6.914878845214844,
+ "step": 181
+ },
+ {
+ "epoch": 2.531004366812227,
+ "grad_norm": 0.015795735642313957,
+ "learning_rate": 0.0006,
+ "loss": 6.922432899475098,
+ "step": 182
+ },
+ {
+ "epoch": 2.544978165938865,
+ "grad_norm": 0.008821426890790462,
+ "learning_rate": 0.0006,
+ "loss": 6.924562454223633,
+ "step": 183
+ },
+ {
+ "epoch": 2.558951965065502,
+ "grad_norm": 0.012502548284828663,
+ "learning_rate": 0.0006,
+ "loss": 6.914718151092529,
+ "step": 184
+ },
+ {
+ "epoch": 2.5729257641921395,
+ "grad_norm": 0.011545066721737385,
+ "learning_rate": 0.0006,
+ "loss": 6.915705680847168,
+ "step": 185
+ },
+ {
+ "epoch": 2.5868995633187772,
+ "grad_norm": 0.022190893068909645,
+ "learning_rate": 0.0006,
+ "loss": 6.88851261138916,
+ "step": 186
+ },
+ {
+ "epoch": 2.600873362445415,
+ "grad_norm": 0.023395851254463196,
+ "learning_rate": 0.0006,
+ "loss": 6.877655029296875,
+ "step": 187
+ },
+ {
+ "epoch": 2.6148471615720523,
+ "grad_norm": 0.015601156279444695,
+ "learning_rate": 0.0006,
+ "loss": 6.889644145965576,
+ "step": 188
+ },
+ {
+ "epoch": 2.62882096069869,
+ "grad_norm": 0.01362934336066246,
+ "learning_rate": 0.0006,
+ "loss": 6.891547203063965,
+ "step": 189
+ },
+ {
+ "epoch": 2.6427947598253274,
+ "grad_norm": 0.012305490672588348,
+ "learning_rate": 0.0006,
+ "loss": 6.890357971191406,
+ "step": 190
+ },
+ {
+ "epoch": 2.656768558951965,
+ "grad_norm": 0.01623358018696308,
+ "learning_rate": 0.0006,
+ "loss": 6.849443435668945,
+ "step": 191
+ },
+ {
+ "epoch": 2.6707423580786025,
+ "grad_norm": 0.018063291907310486,
+ "learning_rate": 0.0006,
+ "loss": 6.842372894287109,
+ "step": 192
+ },
+ {
+ "epoch": 2.6847161572052403,
+ "grad_norm": 0.010179434902966022,
+ "learning_rate": 0.0006,
+ "loss": 6.870055198669434,
+ "step": 193
+ },
+ {
+ "epoch": 2.6986899563318776,
+ "grad_norm": 0.009986722841858864,
+ "learning_rate": 0.0006,
+ "loss": 6.859061241149902,
+ "step": 194
+ },
+ {
+ "epoch": 2.7126637554585153,
+ "grad_norm": 0.01117654051631689,
+ "learning_rate": 0.0006,
+ "loss": 6.861199378967285,
+ "step": 195
+ },
+ {
+ "epoch": 2.726637554585153,
+ "grad_norm": 0.008551654405891895,
+ "learning_rate": 0.0006,
+ "loss": 6.850255012512207,
+ "step": 196
+ },
+ {
+ "epoch": 2.7406113537117904,
+ "grad_norm": 0.018904326483607292,
+ "learning_rate": 0.0006,
+ "loss": 6.848639965057373,
+ "step": 197
+ },
+ {
+ "epoch": 2.7545851528384278,
+ "grad_norm": 0.02080724574625492,
+ "learning_rate": 0.0006,
+ "loss": 6.80456018447876,
+ "step": 198
+ },
+ {
+ "epoch": 2.7685589519650655,
+ "grad_norm": 0.012010873295366764,
+ "learning_rate": 0.0006,
+ "loss": 6.848775863647461,
+ "step": 199
+ },
+ {
+ "epoch": 2.7825327510917033,
+ "grad_norm": 0.008113056421279907,
+ "learning_rate": 0.0006,
+ "loss": 6.849387168884277,
+ "step": 200
+ },
+ {
+ "epoch": 2.7965065502183406,
+ "grad_norm": 0.012912592850625515,
+ "learning_rate": 0.0006,
+ "loss": 6.832240104675293,
+ "step": 201
+ },
+ {
+ "epoch": 2.810480349344978,
+ "grad_norm": 0.01146865077316761,
+ "learning_rate": 0.0006,
+ "loss": 6.850383281707764,
+ "step": 202
+ },
+ {
+ "epoch": 2.8244541484716157,
+ "grad_norm": 0.0171892661601305,
+ "learning_rate": 0.0006,
+ "loss": 6.839577674865723,
+ "step": 203
+ },
+ {
+ "epoch": 2.8384279475982535,
+ "grad_norm": 0.012083170004189014,
+ "learning_rate": 0.0006,
+ "loss": 6.818614482879639,
+ "step": 204
+ },
+ {
+ "epoch": 2.8524017467248908,
+ "grad_norm": 0.0062555656768381596,
+ "learning_rate": 0.0006,
+ "loss": 6.831291198730469,
+ "step": 205
+ },
+ {
+ "epoch": 2.8663755458515285,
+ "grad_norm": 0.010102466680109501,
+ "learning_rate": 0.0006,
+ "loss": 6.846522808074951,
+ "step": 206
+ },
+ {
+ "epoch": 2.880349344978166,
+ "grad_norm": 0.009042087942361832,
+ "learning_rate": 0.0006,
+ "loss": 6.822072982788086,
+ "step": 207
+ },
+ {
+ "epoch": 2.8943231441048036,
+ "grad_norm": 0.011133556254208088,
+ "learning_rate": 0.0006,
+ "loss": 6.811252593994141,
+ "step": 208
+ },
+ {
+ "epoch": 2.908296943231441,
+ "grad_norm": 0.00990577507764101,
+ "learning_rate": 0.0006,
+ "loss": 6.814435958862305,
+ "step": 209
+ },
+ {
+ "epoch": 2.9222707423580787,
+ "grad_norm": 0.006605073343962431,
+ "learning_rate": 0.0006,
+ "loss": 6.831300735473633,
+ "step": 210
+ },
+ {
+ "epoch": 2.936244541484716,
+ "grad_norm": 0.007831175811588764,
+ "learning_rate": 0.0006,
+ "loss": 6.841738700866699,
+ "step": 211
+ },
+ {
+ "epoch": 2.950218340611354,
+ "grad_norm": 0.005536255426704884,
+ "learning_rate": 0.0006,
+ "loss": 6.8043670654296875,
+ "step": 212
+ },
+ {
+ "epoch": 2.964192139737991,
+ "grad_norm": 0.006532667670398951,
+ "learning_rate": 0.0006,
+ "loss": 6.8091840744018555,
+ "step": 213
+ },
+ {
+ "epoch": 2.978165938864629,
+ "grad_norm": 0.006674565840512514,
+ "learning_rate": 0.0006,
+ "loss": 6.811327934265137,
+ "step": 214
+ },
+ {
+ "epoch": 2.992139737991266,
+ "grad_norm": 0.01150344219058752,
+ "learning_rate": 0.0006,
+ "loss": 6.810762405395508,
+ "step": 215
+ },
+ {
+ "epoch": 3.0,
+ "grad_norm": 0.009015528485178947,
+ "learning_rate": 0.0006,
+ "loss": 6.836431980133057,
+ "step": 216
+ },
+ {
+ "epoch": 3.0,
+ "eval_loss": 7.718832969665527,
+ "eval_runtime": 58.6615,
+ "eval_samples_per_second": 41.629,
+ "eval_steps_per_second": 1.313,
+ "step": 216
+ },
+ {
+ "epoch": 3.0139737991266378,
+ "grad_norm": 0.009791829623281956,
+ "learning_rate": 0.0006,
+ "loss": 6.8102593421936035,
+ "step": 217
+ },
+ {
+ "epoch": 3.027947598253275,
+ "grad_norm": 0.008970947004854679,
+ "learning_rate": 0.0006,
+ "loss": 6.792663097381592,
+ "step": 218
+ },
+ {
+ "epoch": 3.041921397379913,
+ "grad_norm": 0.013689392246305943,
+ "learning_rate": 0.0006,
+ "loss": 6.815651893615723,
+ "step": 219
+ },
+ {
+ "epoch": 3.05589519650655,
+ "grad_norm": 0.031499531120061874,
+ "learning_rate": 0.0006,
+ "loss": 6.809841156005859,
+ "step": 220
+ },
+ {
+ "epoch": 3.069868995633188,
+ "grad_norm": 0.03203094005584717,
+ "learning_rate": 0.0006,
+ "loss": 6.779417037963867,
+ "step": 221
+ },
+ {
+ "epoch": 3.0838427947598253,
+ "grad_norm": 0.007947996258735657,
+ "learning_rate": 0.0006,
+ "loss": 6.773656845092773,
+ "step": 222
+ },
+ {
+ "epoch": 3.097816593886463,
+ "grad_norm": 0.03152519091963768,
+ "learning_rate": 0.0006,
+ "loss": 6.804083824157715,
+ "step": 223
+ },
+ {
+ "epoch": 3.1117903930131003,
+ "grad_norm": 0.014937258325517178,
+ "learning_rate": 0.0006,
+ "loss": 6.792213439941406,
+ "step": 224
+ },
+ {
+ "epoch": 3.125764192139738,
+ "grad_norm": 0.020961705595254898,
+ "learning_rate": 0.0006,
+ "loss": 6.8092803955078125,
+ "step": 225
+ },
+ {
+ "epoch": 3.1397379912663754,
+ "grad_norm": 0.009152988903224468,
+ "learning_rate": 0.0006,
+ "loss": 6.7737603187561035,
+ "step": 226
+ },
+ {
+ "epoch": 3.153711790393013,
+ "grad_norm": 0.02770620957016945,
+ "learning_rate": 0.0006,
+ "loss": 6.767014503479004,
+ "step": 227
+ },
+ {
+ "epoch": 3.1676855895196505,
+ "grad_norm": 0.01165871974080801,
+ "learning_rate": 0.0006,
+ "loss": 6.760884761810303,
+ "step": 228
+ },
+ {
+ "epoch": 3.1816593886462883,
+ "grad_norm": 0.01910376362502575,
+ "learning_rate": 0.0006,
+ "loss": 6.775722980499268,
+ "step": 229
+ },
+ {
+ "epoch": 3.1956331877729256,
+ "grad_norm": 0.010413747280836105,
+ "learning_rate": 0.0006,
+ "loss": 6.779816627502441,
+ "step": 230
+ },
+ {
+ "epoch": 3.2096069868995634,
+ "grad_norm": 0.01918417401611805,
+ "learning_rate": 0.0006,
+ "loss": 6.760381698608398,
+ "step": 231
+ },
+ {
+ "epoch": 3.2235807860262007,
+ "grad_norm": 0.017960334196686745,
+ "learning_rate": 0.0006,
+ "loss": 6.761198997497559,
+ "step": 232
+ },
+ {
+ "epoch": 3.2375545851528384,
+ "grad_norm": 0.016546795144677162,
+ "learning_rate": 0.0006,
+ "loss": 6.742448806762695,
+ "step": 233
+ },
+ {
+ "epoch": 3.251528384279476,
+ "grad_norm": 0.01889597810804844,
+ "learning_rate": 0.0006,
+ "loss": 6.758237838745117,
+ "step": 234
+ },
+ {
+ "epoch": 3.2655021834061135,
+ "grad_norm": 0.009611992165446281,
+ "learning_rate": 0.0006,
+ "loss": 6.75614070892334,
+ "step": 235
+ },
+ {
+ "epoch": 3.279475982532751,
+ "grad_norm": 0.02034934051334858,
+ "learning_rate": 0.0006,
+ "loss": 6.736965179443359,
+ "step": 236
+ },
+ {
+ "epoch": 3.2934497816593886,
+ "grad_norm": 0.010740065947175026,
+ "learning_rate": 0.0006,
+ "loss": 6.746947288513184,
+ "step": 237
+ },
+ {
+ "epoch": 3.3074235807860264,
+ "grad_norm": 0.018383340910077095,
+ "learning_rate": 0.0006,
+ "loss": 6.747332572937012,
+ "step": 238
+ },
+ {
+ "epoch": 3.3213973799126637,
+ "grad_norm": 0.012362911365926266,
+ "learning_rate": 0.0006,
+ "loss": 6.767669200897217,
+ "step": 239
+ },
+ {
+ "epoch": 3.3353711790393015,
+ "grad_norm": 0.01127463299781084,
+ "learning_rate": 0.0006,
+ "loss": 6.769808292388916,
+ "step": 240
+ },
+ {
+ "epoch": 3.349344978165939,
+ "grad_norm": 0.017693769186735153,
+ "learning_rate": 0.0006,
+ "loss": 6.733658790588379,
+ "step": 241
+ },
+ {
+ "epoch": 3.3633187772925766,
+ "grad_norm": 0.006420376244932413,
+ "learning_rate": 0.0006,
+ "loss": 6.74934720993042,
+ "step": 242
+ },
+ {
+ "epoch": 3.377292576419214,
+ "grad_norm": 0.014020364731550217,
+ "learning_rate": 0.0006,
+ "loss": 6.728157043457031,
+ "step": 243
+ },
+ {
+ "epoch": 3.3912663755458516,
+ "grad_norm": 0.008018582127988338,
+ "learning_rate": 0.0006,
+ "loss": 6.730726718902588,
+ "step": 244
+ },
+ {
+ "epoch": 3.405240174672489,
+ "grad_norm": 0.00782222580164671,
+ "learning_rate": 0.0006,
+ "loss": 6.7499895095825195,
+ "step": 245
+ },
+ {
+ "epoch": 3.4192139737991267,
+ "grad_norm": 0.008278409950435162,
+ "learning_rate": 0.0006,
+ "loss": 6.712803363800049,
+ "step": 246
+ },
+ {
+ "epoch": 3.433187772925764,
+ "grad_norm": 0.009826759807765484,
+ "learning_rate": 0.0006,
+ "loss": 6.742262840270996,
+ "step": 247
+ },
+ {
+ "epoch": 3.447161572052402,
+ "grad_norm": 0.009488224983215332,
+ "learning_rate": 0.0006,
+ "loss": 6.739130020141602,
+ "step": 248
+ },
+ {
+ "epoch": 3.461135371179039,
+ "grad_norm": 0.007667903788387775,
+ "learning_rate": 0.0006,
+ "loss": 6.747515678405762,
+ "step": 249
+ },
+ {
+ "epoch": 3.475109170305677,
+ "grad_norm": 0.0073090000078082085,
+ "learning_rate": 0.0006,
+ "loss": 6.740870475769043,
+ "step": 250
+ },
+ {
+ "epoch": 3.489082969432314,
+ "grad_norm": 0.010058708488941193,
+ "learning_rate": 0.0006,
+ "loss": 6.777563095092773,
+ "step": 251
+ },
+ {
+ "epoch": 3.503056768558952,
+ "grad_norm": 0.020676063373684883,
+ "learning_rate": 0.0006,
+ "loss": 6.708603858947754,
+ "step": 252
+ },
+ {
+ "epoch": 3.5170305676855893,
+ "grad_norm": 0.027210544794797897,
+ "learning_rate": 0.0006,
+ "loss": 6.694097995758057,
+ "step": 253
+ },
+ {
+ "epoch": 3.531004366812227,
+ "grad_norm": 0.038053520023822784,
+ "learning_rate": 0.0006,
+ "loss": 6.697017669677734,
+ "step": 254
+ },
+ {
+ "epoch": 3.544978165938865,
+ "grad_norm": 0.04684312269091606,
+ "learning_rate": 0.0006,
+ "loss": 6.729883193969727,
+ "step": 255
+ },
+ {
+ "epoch": 3.558951965065502,
+ "grad_norm": 0.03673060983419418,
+ "learning_rate": 0.0006,
+ "loss": 6.743753433227539,
+ "step": 256
+ },
+ {
+ "epoch": 3.5729257641921395,
+ "grad_norm": 0.006882632151246071,
+ "learning_rate": 0.0006,
+ "loss": 6.71405029296875,
+ "step": 257
+ },
+ {
+ "epoch": 3.5868995633187772,
+ "grad_norm": 0.03588201105594635,
+ "learning_rate": 0.0006,
+ "loss": 6.712467193603516,
+ "step": 258
+ },
+ {
+ "epoch": 3.600873362445415,
+ "grad_norm": 0.029415199533104897,
+ "learning_rate": 0.0006,
+ "loss": 6.715634822845459,
+ "step": 259
+ },
+ {
+ "epoch": 3.6148471615720523,
+ "grad_norm": 0.011425039730966091,
+ "learning_rate": 0.0006,
+ "loss": 6.697334289550781,
+ "step": 260
+ },
+ {
+ "epoch": 3.62882096069869,
+ "grad_norm": 0.0297263003885746,
+ "learning_rate": 0.0006,
+ "loss": 6.724848747253418,
+ "step": 261
+ },
+ {
+ "epoch": 3.6427947598253274,
+ "grad_norm": 0.007282314356416464,
+ "learning_rate": 0.0006,
+ "loss": 6.702306270599365,
+ "step": 262
+ },
+ {
+ "epoch": 3.656768558951965,
+ "grad_norm": 0.02673051692545414,
+ "learning_rate": 0.0006,
+ "loss": 6.72357177734375,
+ "step": 263
+ },
+ {
+ "epoch": 3.6707423580786025,
+ "grad_norm": 0.016174081712961197,
+ "learning_rate": 0.0006,
+ "loss": 6.730698585510254,
+ "step": 264
+ },
+ {
+ "epoch": 3.6847161572052403,
+ "grad_norm": 0.0182416420429945,
+ "learning_rate": 0.0006,
+ "loss": 6.717386245727539,
+ "step": 265
+ },
+ {
+ "epoch": 3.6986899563318776,
+ "grad_norm": 0.023250866681337357,
+ "learning_rate": 0.0006,
+ "loss": 6.697832107543945,
+ "step": 266
+ },
+ {
+ "epoch": 3.7126637554585153,
+ "grad_norm": 0.007539753336459398,
+ "learning_rate": 0.0006,
+ "loss": 6.714348793029785,
+ "step": 267
+ },
+ {
+ "epoch": 3.726637554585153,
+ "grad_norm": 0.030086630955338478,
+ "learning_rate": 0.0006,
+ "loss": 6.724481582641602,
+ "step": 268
+ },
+ {
+ "epoch": 3.7406113537117904,
+ "grad_norm": 0.012784913182258606,
+ "learning_rate": 0.0006,
+ "loss": 6.703491687774658,
+ "step": 269
+ },
+ {
+ "epoch": 3.7545851528384278,
+ "grad_norm": 0.02042769268155098,
+ "learning_rate": 0.0006,
+ "loss": 6.721139907836914,
+ "step": 270
+ },
+ {
+ "epoch": 3.7685589519650655,
+ "grad_norm": 0.023557329550385475,
+ "learning_rate": 0.0006,
+ "loss": 6.728122711181641,
+ "step": 271
+ },
+ {
+ "epoch": 3.7825327510917033,
+ "grad_norm": 0.007716621737927198,
+ "learning_rate": 0.0006,
+ "loss": 6.70085334777832,
+ "step": 272
+ },
+ {
+ "epoch": 3.7965065502183406,
+ "grad_norm": 0.022772278636693954,
+ "learning_rate": 0.0006,
+ "loss": 6.712882995605469,
+ "step": 273
+ },
+ {
+ "epoch": 3.810480349344978,
+ "grad_norm": 0.007928933948278427,
+ "learning_rate": 0.0006,
+ "loss": 6.698320388793945,
+ "step": 274
+ },
+ {
+ "epoch": 3.8244541484716157,
+ "grad_norm": 0.022355245426297188,
+ "learning_rate": 0.0006,
+ "loss": 6.687088489532471,
+ "step": 275
+ },
+ {
+ "epoch": 3.8384279475982535,
+ "grad_norm": 0.02037850394845009,
+ "learning_rate": 0.0006,
+ "loss": 6.676856994628906,
+ "step": 276
+ },
+ {
+ "epoch": 3.8524017467248908,
+ "grad_norm": 0.012290362268686295,
+ "learning_rate": 0.0006,
+ "loss": 6.682883262634277,
+ "step": 277
+ },
+ {
+ "epoch": 3.8663755458515285,
+ "grad_norm": 0.022331275045871735,
+ "learning_rate": 0.0006,
+ "loss": 6.694752216339111,
+ "step": 278
+ },
+ {
+ "epoch": 3.880349344978166,
+ "grad_norm": 0.007527181413024664,
+ "learning_rate": 0.0006,
+ "loss": 6.686394691467285,
+ "step": 279
+ },
+ {
+ "epoch": 3.8943231441048036,
+ "grad_norm": 0.02159409038722515,
+ "learning_rate": 0.0006,
+ "loss": 6.697943687438965,
+ "step": 280
+ },
+ {
+ "epoch": 3.908296943231441,
+ "grad_norm": 0.015054763294756413,
+ "learning_rate": 0.0006,
+ "loss": 6.677796363830566,
+ "step": 281
+ },
+ {
+ "epoch": 3.9222707423580787,
+ "grad_norm": 0.01466019731014967,
+ "learning_rate": 0.0006,
+ "loss": 6.697715759277344,
+ "step": 282
+ },
+ {
+ "epoch": 3.936244541484716,
+ "grad_norm": 0.01875465176999569,
+ "learning_rate": 0.0006,
+ "loss": 6.700228691101074,
+ "step": 283
+ },
+ {
+ "epoch": 3.950218340611354,
+ "grad_norm": 0.00778991961851716,
+ "learning_rate": 0.0006,
+ "loss": 6.69331693649292,
+ "step": 284
+ },
+ {
+ "epoch": 3.964192139737991,
+ "grad_norm": 0.020182441920042038,
+ "learning_rate": 0.0006,
+ "loss": 6.672651290893555,
+ "step": 285
+ },
+ {
+ "epoch": 3.978165938864629,
+ "grad_norm": 0.01240516733378172,
+ "learning_rate": 0.0006,
+ "loss": 6.688860893249512,
+ "step": 286
+ },
+ {
+ "epoch": 3.992139737991266,
+ "grad_norm": 0.011218412779271603,
+ "learning_rate": 0.0006,
+ "loss": 6.669534683227539,
+ "step": 287
+ },
+ {
+ "epoch": 4.0,
+ "grad_norm": 0.0141852255910635,
+ "learning_rate": 0.0006,
+ "loss": 6.690206527709961,
+ "step": 288
+ },
+ {
+ "epoch": 4.0,
+ "eval_loss": 7.537620544433594,
+ "eval_runtime": 58.4398,
+ "eval_samples_per_second": 41.787,
+ "eval_steps_per_second": 1.318,
+ "step": 288
+ },
+ {
+ "epoch": 4.013973799126638,
+ "grad_norm": 0.008106146939098835,
+ "learning_rate": 0.0006,
+ "loss": 6.688591003417969,
+ "step": 289
+ },
+ {
+ "epoch": 4.0279475982532755,
+ "grad_norm": 0.014588244259357452,
+ "learning_rate": 0.0006,
+ "loss": 6.684961318969727,
+ "step": 290
+ },
+ {
+ "epoch": 4.041921397379912,
+ "grad_norm": 0.011985368095338345,
+ "learning_rate": 0.0006,
+ "loss": 6.672794342041016,
+ "step": 291
+ },
+ {
+ "epoch": 4.05589519650655,
+ "grad_norm": 0.014632102102041245,
+ "learning_rate": 0.0006,
+ "loss": 6.702736854553223,
+ "step": 292
+ },
+ {
+ "epoch": 4.069868995633188,
+ "grad_norm": 0.011243713088333607,
+ "learning_rate": 0.0006,
+ "loss": 6.658792495727539,
+ "step": 293
+ },
+ {
+ "epoch": 4.083842794759826,
+ "grad_norm": 0.010519202798604965,
+ "learning_rate": 0.0006,
+ "loss": 6.662495136260986,
+ "step": 294
+ },
+ {
+ "epoch": 4.097816593886463,
+ "grad_norm": 0.007318088319152594,
+ "learning_rate": 0.0006,
+ "loss": 6.675954818725586,
+ "step": 295
+ },
+ {
+ "epoch": 4.1117903930131,
+ "grad_norm": 0.011712398380041122,
+ "learning_rate": 0.0006,
+ "loss": 6.673973083496094,
+ "step": 296
+ },
+ {
+ "epoch": 4.125764192139738,
+ "grad_norm": 0.006726635154336691,
+ "learning_rate": 0.0006,
+ "loss": 6.656833171844482,
+ "step": 297
+ },
+ {
+ "epoch": 4.139737991266376,
+ "grad_norm": 0.005466983187943697,
+ "learning_rate": 0.0006,
+ "loss": 6.675468921661377,
+ "step": 298
+ },
+ {
+ "epoch": 4.153711790393013,
+ "grad_norm": 0.007163621485233307,
+ "learning_rate": 0.0006,
+ "loss": 6.6862874031066895,
+ "step": 299
+ },
+ {
+ "epoch": 4.1676855895196505,
+ "grad_norm": 0.007626629434525967,
+ "learning_rate": 0.0006,
+ "loss": 6.654741287231445,
+ "step": 300
+ },
+ {
+ "epoch": 4.181659388646288,
+ "grad_norm": 0.0066330647096037865,
+ "learning_rate": 0.0006,
+ "loss": 6.68114709854126,
+ "step": 301
+ },
+ {
+ "epoch": 4.195633187772926,
+ "grad_norm": 0.007363482844084501,
+ "learning_rate": 0.0006,
+ "loss": 6.683980941772461,
+ "step": 302
+ },
+ {
+ "epoch": 4.209606986899563,
+ "grad_norm": 0.00971233006566763,
+ "learning_rate": 0.0006,
+ "loss": 6.669700622558594,
+ "step": 303
+ },
+ {
+ "epoch": 4.223580786026201,
+ "grad_norm": 0.013185256160795689,
+ "learning_rate": 0.0006,
+ "loss": 6.686490535736084,
+ "step": 304
+ },
+ {
+ "epoch": 4.2375545851528384,
+ "grad_norm": 0.019116027280688286,
+ "learning_rate": 0.0006,
+ "loss": 6.645479202270508,
+ "step": 305
+ },
+ {
+ "epoch": 4.251528384279476,
+ "grad_norm": 0.0328814722597599,
+ "learning_rate": 0.0006,
+ "loss": 6.68080997467041,
+ "step": 306
+ },
+ {
+ "epoch": 4.265502183406113,
+ "grad_norm": 0.04482287913560867,
+ "learning_rate": 0.0006,
+ "loss": 6.698016166687012,
+ "step": 307
+ },
+ {
+ "epoch": 4.279475982532751,
+ "grad_norm": 0.03053313121199608,
+ "learning_rate": 0.0006,
+ "loss": 6.642240524291992,
+ "step": 308
+ },
+ {
+ "epoch": 4.293449781659389,
+ "grad_norm": 0.043351784348487854,
+ "learning_rate": 0.0006,
+ "loss": 6.662343978881836,
+ "step": 309
+ },
+ {
+ "epoch": 4.307423580786026,
+ "grad_norm": 0.03165886923670769,
+ "learning_rate": 0.0006,
+ "loss": 6.6291961669921875,
+ "step": 310
+ },
+ {
+ "epoch": 4.321397379912664,
+ "grad_norm": 0.028601787984371185,
+ "learning_rate": 0.0006,
+ "loss": 6.667544364929199,
+ "step": 311
+ },
+ {
+ "epoch": 4.335371179039301,
+ "grad_norm": 0.019558480009436607,
+ "learning_rate": 0.0006,
+ "loss": 6.641025543212891,
+ "step": 312
+ },
+ {
+ "epoch": 4.349344978165939,
+ "grad_norm": 0.039851363748311996,
+ "learning_rate": 0.0006,
+ "loss": 6.665952205657959,
+ "step": 313
+ },
+ {
+ "epoch": 4.3633187772925766,
+ "grad_norm": 0.018417146056890488,
+ "learning_rate": 0.0006,
+ "loss": 6.657257556915283,
+ "step": 314
+ },
+ {
+ "epoch": 4.377292576419214,
+ "grad_norm": 0.02894527092576027,
+ "learning_rate": 0.0006,
+ "loss": 6.656288146972656,
+ "step": 315
+ },
+ {
+ "epoch": 4.391266375545851,
+ "grad_norm": 0.01910966821014881,
+ "learning_rate": 0.0006,
+ "loss": 6.688958168029785,
+ "step": 316
+ },
+ {
+ "epoch": 4.405240174672489,
+ "grad_norm": 0.01668412610888481,
+ "learning_rate": 0.0006,
+ "loss": 6.64876127243042,
+ "step": 317
+ },
+ {
+ "epoch": 4.419213973799127,
+ "grad_norm": 0.02935275249183178,
+ "learning_rate": 0.0006,
+ "loss": 6.626096725463867,
+ "step": 318
+ },
+ {
+ "epoch": 4.4331877729257645,
+ "grad_norm": 0.017997972667217255,
+ "learning_rate": 0.0006,
+ "loss": 6.640386581420898,
+ "step": 319
+ },
+ {
+ "epoch": 4.447161572052401,
+ "grad_norm": 0.01779189519584179,
+ "learning_rate": 0.0006,
+ "loss": 6.641165256500244,
+ "step": 320
+ },
+ {
+ "epoch": 4.461135371179039,
+ "grad_norm": 0.013901712372899055,
+ "learning_rate": 0.0006,
+ "loss": 6.651627063751221,
+ "step": 321
+ },
+ {
+ "epoch": 4.475109170305677,
+ "grad_norm": 0.013211199082434177,
+ "learning_rate": 0.0006,
+ "loss": 6.630064964294434,
+ "step": 322
+ },
+ {
+ "epoch": 4.489082969432315,
+ "grad_norm": 0.02324107475578785,
+ "learning_rate": 0.0006,
+ "loss": 6.670443058013916,
+ "step": 323
+ },
+ {
+ "epoch": 4.503056768558952,
+ "grad_norm": 0.019452504813671112,
+ "learning_rate": 0.0006,
+ "loss": 6.651266098022461,
+ "step": 324
+ },
+ {
+ "epoch": 4.517030567685589,
+ "grad_norm": 0.017943572252988815,
+ "learning_rate": 0.0006,
+ "loss": 6.657943248748779,
+ "step": 325
+ },
+ {
+ "epoch": 4.531004366812227,
+ "grad_norm": 0.012136237695813179,
+ "learning_rate": 0.0006,
+ "loss": 6.640082359313965,
+ "step": 326
+ },
+ {
+ "epoch": 4.544978165938865,
+ "grad_norm": 0.011544011533260345,
+ "learning_rate": 0.0006,
+ "loss": 6.6186933517456055,
+ "step": 327
+ },
+ {
+ "epoch": 4.558951965065502,
+ "grad_norm": 0.01569426618516445,
+ "learning_rate": 0.0006,
+ "loss": 6.640767574310303,
+ "step": 328
+ },
+ {
+ "epoch": 4.5729257641921395,
+ "grad_norm": 0.027668394148349762,
+ "learning_rate": 0.0006,
+ "loss": 6.651412487030029,
+ "step": 329
+ },
+ {
+ "epoch": 4.586899563318777,
+ "grad_norm": 0.04073230177164078,
+ "learning_rate": 0.0006,
+ "loss": 6.636565208435059,
+ "step": 330
+ },
+ {
+ "epoch": 4.600873362445415,
+ "grad_norm": 0.04942672327160835,
+ "learning_rate": 0.0006,
+ "loss": 6.667994499206543,
+ "step": 331
+ },
+ {
+ "epoch": 4.614847161572053,
+ "grad_norm": 0.037568598985672,
+ "learning_rate": 0.0006,
+ "loss": 6.632008075714111,
+ "step": 332
+ },
+ {
+ "epoch": 4.62882096069869,
+ "grad_norm": 0.009880400262773037,
+ "learning_rate": 0.0006,
+ "loss": 6.6504058837890625,
+ "step": 333
+ },
+ {
+ "epoch": 4.642794759825327,
+ "grad_norm": 0.033384259790182114,
+ "learning_rate": 0.0006,
+ "loss": 6.63852596282959,
+ "step": 334
+ },
+ {
+ "epoch": 4.656768558951965,
+ "grad_norm": 0.035640884190797806,
+ "learning_rate": 0.0006,
+ "loss": 6.627986431121826,
+ "step": 335
+ },
+ {
+ "epoch": 4.670742358078603,
+ "grad_norm": 0.013637864030897617,
+ "learning_rate": 0.0006,
+ "loss": 6.624503135681152,
+ "step": 336
+ },
+ {
+ "epoch": 4.68471615720524,
+ "grad_norm": 0.025304675102233887,
+ "learning_rate": 0.0006,
+ "loss": 6.612077236175537,
+ "step": 337
+ },
+ {
+ "epoch": 4.698689956331878,
+ "grad_norm": 0.03377160802483559,
+ "learning_rate": 0.0006,
+ "loss": 6.608938217163086,
+ "step": 338
+ },
+ {
+ "epoch": 4.712663755458515,
+ "grad_norm": 0.019803153350949287,
+ "learning_rate": 0.0006,
+ "loss": 6.636343955993652,
+ "step": 339
+ },
+ {
+ "epoch": 4.726637554585153,
+ "grad_norm": 0.018928129225969315,
+ "learning_rate": 0.0006,
+ "loss": 6.636613845825195,
+ "step": 340
+ },
+ {
+ "epoch": 4.74061135371179,
+ "grad_norm": 0.023068996146321297,
+ "learning_rate": 0.0006,
+ "loss": 6.6258440017700195,
+ "step": 341
+ },
+ {
+ "epoch": 4.754585152838428,
+ "grad_norm": 0.016849957406520844,
+ "learning_rate": 0.0006,
+ "loss": 6.638483047485352,
+ "step": 342
+ },
+ {
+ "epoch": 4.7685589519650655,
+ "grad_norm": 0.021430544555187225,
+ "learning_rate": 0.0006,
+ "loss": 6.619649410247803,
+ "step": 343
+ },
+ {
+ "epoch": 4.782532751091703,
+ "grad_norm": 0.024053368717432022,
+ "learning_rate": 0.0006,
+ "loss": 6.639214515686035,
+ "step": 344
+ },
+ {
+ "epoch": 4.796506550218341,
+ "grad_norm": 0.011926693841814995,
+ "learning_rate": 0.0006,
+ "loss": 6.604836463928223,
+ "step": 345
+ },
+ {
+ "epoch": 4.810480349344978,
+ "grad_norm": 0.01961611397564411,
+ "learning_rate": 0.0006,
+ "loss": 6.624864101409912,
+ "step": 346
+ },
+ {
+ "epoch": 4.824454148471616,
+ "grad_norm": 0.027946067973971367,
+ "learning_rate": 0.0006,
+ "loss": 6.621284484863281,
+ "step": 347
+ },
+ {
+ "epoch": 4.8384279475982535,
+ "grad_norm": 0.02108592912554741,
+ "learning_rate": 0.0006,
+ "loss": 6.5982465744018555,
+ "step": 348
+ },
+ {
+ "epoch": 4.85240174672489,
+ "grad_norm": 0.017272260040044785,
+ "learning_rate": 0.0006,
+ "loss": 6.626142501831055,
+ "step": 349
+ },
+ {
+ "epoch": 4.866375545851528,
+ "grad_norm": 0.015654582530260086,
+ "learning_rate": 0.0006,
+ "loss": 6.621735572814941,
+ "step": 350
+ },
+ {
+ "epoch": 4.880349344978166,
+ "grad_norm": 0.015797540545463562,
+ "learning_rate": 0.0006,
+ "loss": 6.6016411781311035,
+ "step": 351
+ },
+ {
+ "epoch": 4.894323144104804,
+ "grad_norm": 0.028488904237747192,
+ "learning_rate": 0.0006,
+ "loss": 6.598413467407227,
+ "step": 352
+ },
+ {
+ "epoch": 4.908296943231441,
+ "grad_norm": 0.03411734476685524,
+ "learning_rate": 0.0006,
+ "loss": 6.597321033477783,
+ "step": 353
+ },
+ {
+ "epoch": 4.922270742358078,
+ "grad_norm": 0.027188263833522797,
+ "learning_rate": 0.0006,
+ "loss": 6.630616188049316,
+ "step": 354
+ },
+ {
+ "epoch": 4.936244541484716,
+ "grad_norm": 0.03411437198519707,
+ "learning_rate": 0.0006,
+ "loss": 6.57647705078125,
+ "step": 355
+ },
+ {
+ "epoch": 4.950218340611354,
+ "grad_norm": 0.019907327368855476,
+ "learning_rate": 0.0006,
+ "loss": 6.579545974731445,
+ "step": 356
+ },
+ {
+ "epoch": 4.964192139737992,
+ "grad_norm": 0.02485930174589157,
+ "learning_rate": 0.0006,
+ "loss": 6.591073036193848,
+ "step": 357
+ },
+ {
+ "epoch": 4.978165938864628,
+ "grad_norm": 0.026106052100658417,
+ "learning_rate": 0.0006,
+ "loss": 6.580729961395264,
+ "step": 358
+ },
+ {
+ "epoch": 4.992139737991266,
+ "grad_norm": 0.029401391744613647,
+ "learning_rate": 0.0006,
+ "loss": 6.5460076332092285,
+ "step": 359
+ },
+ {
+ "epoch": 5.0,
+ "grad_norm": 0.032869379967451096,
+ "learning_rate": 0.0006,
+ "loss": 6.583278179168701,
+ "step": 360
+ },
+ {
+ "epoch": 5.0,
+ "eval_loss": 7.416288375854492,
+ "eval_runtime": 60.4177,
+ "eval_samples_per_second": 40.419,
+ "eval_steps_per_second": 1.274,
+ "step": 360
+ },
+ {
+ "epoch": 5.013973799126638,
+ "grad_norm": 0.04862433671951294,
+ "learning_rate": 0.0006,
+ "loss": 6.58473014831543,
+ "step": 361
+ },
+ {
+ "epoch": 5.0279475982532755,
+ "grad_norm": 0.06636223942041397,
+ "learning_rate": 0.0006,
+ "loss": 6.621153354644775,
+ "step": 362
+ },
+ {
+ "epoch": 5.041921397379912,
+ "grad_norm": 0.05888264626264572,
+ "learning_rate": 0.0006,
+ "loss": 6.621196269989014,
+ "step": 363
+ },
+ {
+ "epoch": 5.05589519650655,
+ "grad_norm": 0.05331697314977646,
+ "learning_rate": 0.0006,
+ "loss": 6.600763320922852,
+ "step": 364
+ },
+ {
+ "epoch": 5.069868995633188,
+ "grad_norm": 0.03947671875357628,
+ "learning_rate": 0.0006,
+ "loss": 6.600671768188477,
+ "step": 365
+ },
+ {
+ "epoch": 5.083842794759826,
+ "grad_norm": 0.04045343026518822,
+ "learning_rate": 0.0006,
+ "loss": 6.63529109954834,
+ "step": 366
+ },
+ {
+ "epoch": 5.097816593886463,
+ "grad_norm": 0.029766036197543144,
+ "learning_rate": 0.0006,
+ "loss": 6.612868309020996,
+ "step": 367
+ },
+ {
+ "epoch": 5.1117903930131,
+ "grad_norm": 0.03416327014565468,
+ "learning_rate": 0.0006,
+ "loss": 6.582240581512451,
+ "step": 368
+ },
+ {
+ "epoch": 5.125764192139738,
+ "grad_norm": 0.03458515554666519,
+ "learning_rate": 0.0006,
+ "loss": 6.604381561279297,
+ "step": 369
+ },
+ {
+ "epoch": 5.139737991266376,
+ "grad_norm": 0.024099089205265045,
+ "learning_rate": 0.0006,
+ "loss": 6.597355365753174,
+ "step": 370
+ },
+ {
+ "epoch": 5.153711790393013,
+ "grad_norm": 0.02694438211619854,
+ "learning_rate": 0.0006,
+ "loss": 6.597512245178223,
+ "step": 371
+ },
+ {
+ "epoch": 5.1676855895196505,
+ "grad_norm": 0.024201856926083565,
+ "learning_rate": 0.0006,
+ "loss": 6.564685821533203,
+ "step": 372
+ },
+ {
+ "epoch": 5.181659388646288,
+ "grad_norm": 0.015959380194544792,
+ "learning_rate": 0.0006,
+ "loss": 6.582263946533203,
+ "step": 373
+ },
+ {
+ "epoch": 5.195633187772926,
+ "grad_norm": 0.025690171867609024,
+ "learning_rate": 0.0006,
+ "loss": 6.601057052612305,
+ "step": 374
+ },
+ {
+ "epoch": 5.209606986899563,
+ "grad_norm": 0.014497052878141403,
+ "learning_rate": 0.0006,
+ "loss": 6.584796905517578,
+ "step": 375
+ },
+ {
+ "epoch": 5.223580786026201,
+ "grad_norm": 0.019870297983288765,
+ "learning_rate": 0.0006,
+ "loss": 6.574010372161865,
+ "step": 376
+ },
+ {
+ "epoch": 5.2375545851528384,
+ "grad_norm": 0.01807512901723385,
+ "learning_rate": 0.0006,
+ "loss": 6.558596134185791,
+ "step": 377
+ },
+ {
+ "epoch": 5.251528384279476,
+ "grad_norm": 0.014525193721055984,
+ "learning_rate": 0.0006,
+ "loss": 6.571803092956543,
+ "step": 378
+ },
+ {
+ "epoch": 5.265502183406113,
+ "grad_norm": 0.016113435849547386,
+ "learning_rate": 0.0006,
+ "loss": 6.568970680236816,
+ "step": 379
+ },
+ {
+ "epoch": 5.279475982532751,
+ "grad_norm": 0.020724017173051834,
+ "learning_rate": 0.0006,
+ "loss": 6.5669403076171875,
+ "step": 380
+ },
+ {
+ "epoch": 5.293449781659389,
+ "grad_norm": 0.013963675126433372,
+ "learning_rate": 0.0006,
+ "loss": 6.586382865905762,
+ "step": 381
+ },
+ {
+ "epoch": 5.307423580786026,
+ "grad_norm": 0.01621115580201149,
+ "learning_rate": 0.0006,
+ "loss": 6.539546966552734,
+ "step": 382
+ },
+ {
+ "epoch": 5.321397379912664,
+ "grad_norm": 0.014925207942724228,
+ "learning_rate": 0.0006,
+ "loss": 6.531795501708984,
+ "step": 383
+ },
+ {
+ "epoch": 5.335371179039301,
+ "grad_norm": 0.015454350039362907,
+ "learning_rate": 0.0006,
+ "loss": 6.575699806213379,
+ "step": 384
+ },
+ {
+ "epoch": 5.349344978165939,
+ "grad_norm": 0.01623542234301567,
+ "learning_rate": 0.0006,
+ "loss": 6.582373142242432,
+ "step": 385
+ },
+ {
+ "epoch": 5.3633187772925766,
+ "grad_norm": 0.015452183783054352,
+ "learning_rate": 0.0006,
+ "loss": 6.576451301574707,
+ "step": 386
+ },
+ {
+ "epoch": 5.377292576419214,
+ "grad_norm": 0.021102432161569595,
+ "learning_rate": 0.0006,
+ "loss": 6.543055534362793,
+ "step": 387
+ },
+ {
+ "epoch": 5.391266375545851,
+ "grad_norm": 0.009827060624957085,
+ "learning_rate": 0.0006,
+ "loss": 6.526939392089844,
+ "step": 388
+ },
+ {
+ "epoch": 5.405240174672489,
+ "grad_norm": 0.013735136948525906,
+ "learning_rate": 0.0006,
+ "loss": 6.538825988769531,
+ "step": 389
+ },
+ {
+ "epoch": 5.419213973799127,
+ "grad_norm": 0.02073509432375431,
+ "learning_rate": 0.0006,
+ "loss": 6.521315574645996,
+ "step": 390
+ },
+ {
+ "epoch": 5.4331877729257645,
+ "grad_norm": 0.03110959194600582,
+ "learning_rate": 0.0006,
+ "loss": 6.561807632446289,
+ "step": 391
+ },
+ {
+ "epoch": 5.447161572052401,
+ "grad_norm": 0.045058365911245346,
+ "learning_rate": 0.0006,
+ "loss": 6.526515960693359,
+ "step": 392
+ },
+ {
+ "epoch": 5.461135371179039,
+ "grad_norm": 0.06359834969043732,
+ "learning_rate": 0.0006,
+ "loss": 6.554903030395508,
+ "step": 393
+ },
+ {
+ "epoch": 5.475109170305677,
+ "grad_norm": 0.06406767666339874,
+ "learning_rate": 0.0006,
+ "loss": 6.576272010803223,
+ "step": 394
+ },
+ {
+ "epoch": 5.489082969432315,
+ "grad_norm": 0.03403699770569801,
+ "learning_rate": 0.0006,
+ "loss": 6.5306243896484375,
+ "step": 395
+ },
+ {
+ "epoch": 5.503056768558952,
+ "grad_norm": 0.036109618842601776,
+ "learning_rate": 0.0006,
+ "loss": 6.54176139831543,
+ "step": 396
+ },
+ {
+ "epoch": 5.517030567685589,
+ "grad_norm": 0.03478201478719711,
+ "learning_rate": 0.0006,
+ "loss": 6.530734062194824,
+ "step": 397
+ },
+ {
+ "epoch": 5.531004366812227,
+ "grad_norm": 0.03401544317603111,
+ "learning_rate": 0.0006,
+ "loss": 6.557794094085693,
+ "step": 398
+ },
+ {
+ "epoch": 5.544978165938865,
+ "grad_norm": 0.031629662960767746,
+ "learning_rate": 0.0006,
+ "loss": 6.564887046813965,
+ "step": 399
+ },
+ {
+ "epoch": 5.558951965065502,
+ "grad_norm": 0.02305024117231369,
+ "learning_rate": 0.0006,
+ "loss": 6.547412872314453,
+ "step": 400
+ },
+ {
+ "epoch": 5.5729257641921395,
+ "grad_norm": 0.030852800235152245,
+ "learning_rate": 0.0006,
+ "loss": 6.495370388031006,
+ "step": 401
+ },
+ {
+ "epoch": 5.586899563318777,
+ "grad_norm": 0.03368550166487694,
+ "learning_rate": 0.0006,
+ "loss": 6.495980262756348,
+ "step": 402
+ },
+ {
+ "epoch": 5.600873362445415,
+ "grad_norm": 0.020976359024643898,
+ "learning_rate": 0.0006,
+ "loss": 6.509410858154297,
+ "step": 403
+ },
+ {
+ "epoch": 5.614847161572053,
+ "grad_norm": 0.019652143120765686,
+ "learning_rate": 0.0006,
+ "loss": 6.493232727050781,
+ "step": 404
+ },
+ {
+ "epoch": 5.62882096069869,
+ "grad_norm": 0.024320121854543686,
+ "learning_rate": 0.0006,
+ "loss": 6.521531581878662,
+ "step": 405
+ },
+ {
+ "epoch": 5.642794759825327,
+ "grad_norm": 0.016611427068710327,
+ "learning_rate": 0.0006,
+ "loss": 6.507271766662598,
+ "step": 406
+ },
+ {
+ "epoch": 5.656768558951965,
+ "grad_norm": 0.01635856367647648,
+ "learning_rate": 0.0006,
+ "loss": 6.492038726806641,
+ "step": 407
+ },
+ {
+ "epoch": 5.670742358078603,
+ "grad_norm": 0.017642095685005188,
+ "learning_rate": 0.0006,
+ "loss": 6.500349998474121,
+ "step": 408
+ },
+ {
+ "epoch": 5.68471615720524,
+ "grad_norm": 0.01679178513586521,
+ "learning_rate": 0.0006,
+ "loss": 6.500361442565918,
+ "step": 409
+ },
+ {
+ "epoch": 5.698689956331878,
+ "grad_norm": 0.011776307597756386,
+ "learning_rate": 0.0006,
+ "loss": 6.482490539550781,
+ "step": 410
+ },
+ {
+ "epoch": 5.712663755458515,
+ "grad_norm": 0.0176856629550457,
+ "learning_rate": 0.0006,
+ "loss": 6.48488712310791,
+ "step": 411
+ },
+ {
+ "epoch": 5.726637554585153,
+ "grad_norm": 0.015193155966699123,
+ "learning_rate": 0.0006,
+ "loss": 6.459545612335205,
+ "step": 412
+ },
+ {
+ "epoch": 5.74061135371179,
+ "grad_norm": 0.018623212352395058,
+ "learning_rate": 0.0006,
+ "loss": 6.478870391845703,
+ "step": 413
+ },
+ {
+ "epoch": 5.754585152838428,
+ "grad_norm": 0.012841006740927696,
+ "learning_rate": 0.0006,
+ "loss": 6.474418640136719,
+ "step": 414
+ },
+ {
+ "epoch": 5.7685589519650655,
+ "grad_norm": 0.015540511347353458,
+ "learning_rate": 0.0006,
+ "loss": 6.4706525802612305,
+ "step": 415
+ },
+ {
+ "epoch": 5.782532751091703,
+ "grad_norm": 0.022695740684866905,
+ "learning_rate": 0.0006,
+ "loss": 6.455738067626953,
+ "step": 416
+ },
+ {
+ "epoch": 5.796506550218341,
+ "grad_norm": 0.026930129155516624,
+ "learning_rate": 0.0006,
+ "loss": 6.458016872406006,
+ "step": 417
+ },
+ {
+ "epoch": 5.810480349344978,
+ "grad_norm": 0.03555797412991524,
+ "learning_rate": 0.0006,
+ "loss": 6.480044364929199,
+ "step": 418
+ },
+ {
+ "epoch": 5.824454148471616,
+ "grad_norm": 0.05024682730436325,
+ "learning_rate": 0.0006,
+ "loss": 6.454564571380615,
+ "step": 419
+ },
+ {
+ "epoch": 5.8384279475982535,
+ "grad_norm": 0.06616255640983582,
+ "learning_rate": 0.0006,
+ "loss": 6.433927536010742,
+ "step": 420
+ },
+ {
+ "epoch": 5.85240174672489,
+ "grad_norm": 0.059065770357847214,
+ "learning_rate": 0.0006,
+ "loss": 6.508362770080566,
+ "step": 421
+ },
+ {
+ "epoch": 5.866375545851528,
+ "grad_norm": 0.026689983904361725,
+ "learning_rate": 0.0006,
+ "loss": 6.464297294616699,
+ "step": 422
+ },
+ {
+ "epoch": 5.880349344978166,
+ "grad_norm": 0.04616398364305496,
+ "learning_rate": 0.0006,
+ "loss": 6.45448112487793,
+ "step": 423
+ },
+ {
+ "epoch": 5.894323144104804,
+ "grad_norm": 0.04980356618762016,
+ "learning_rate": 0.0006,
+ "loss": 6.4696574211120605,
+ "step": 424
+ },
+ {
+ "epoch": 5.908296943231441,
+ "grad_norm": 0.0551983006298542,
+ "learning_rate": 0.0006,
+ "loss": 6.419000625610352,
+ "step": 425
+ },
+ {
+ "epoch": 5.922270742358078,
+ "grad_norm": 0.06988243758678436,
+ "learning_rate": 0.0006,
+ "loss": 6.442890167236328,
+ "step": 426
+ },
+ {
+ "epoch": 5.936244541484716,
+ "grad_norm": 0.06040226295590401,
+ "learning_rate": 0.0006,
+ "loss": 6.513258934020996,
+ "step": 427
+ },
+ {
+ "epoch": 5.950218340611354,
+ "grad_norm": 0.04635024070739746,
+ "learning_rate": 0.0006,
+ "loss": 6.4710283279418945,
+ "step": 428
+ },
+ {
+ "epoch": 5.964192139737992,
+ "grad_norm": 0.04762702435255051,
+ "learning_rate": 0.0006,
+ "loss": 6.469327926635742,
+ "step": 429
+ },
+ {
+ "epoch": 5.978165938864628,
+ "grad_norm": 0.04857536032795906,
+ "learning_rate": 0.0006,
+ "loss": 6.426105499267578,
+ "step": 430
+ },
+ {
+ "epoch": 5.992139737991266,
+ "grad_norm": 0.04412693530321121,
+ "learning_rate": 0.0006,
+ "loss": 6.43179988861084,
+ "step": 431
+ },
+ {
+ "epoch": 6.0,
+ "grad_norm": 0.026129672303795815,
+ "learning_rate": 0.0006,
+ "loss": 6.437918663024902,
+ "step": 432
+ },
+ {
+ "epoch": 6.0,
+ "eval_loss": 7.219249248504639,
+ "eval_runtime": 58.6837,
+ "eval_samples_per_second": 41.613,
+ "eval_steps_per_second": 1.312,
+ "step": 432
+ },
+ {
+ "epoch": 6.013973799126638,
+ "grad_norm": 0.03329874947667122,
+ "learning_rate": 0.0006,
+ "loss": 6.438179969787598,
+ "step": 433
+ },
+ {
+ "epoch": 6.0279475982532755,
+ "grad_norm": 0.03723187372088432,
+ "learning_rate": 0.0006,
+ "loss": 6.394242286682129,
+ "step": 434
+ },
+ {
+ "epoch": 6.041921397379912,
+ "grad_norm": 0.014625726267695427,
+ "learning_rate": 0.0006,
+ "loss": 6.412118911743164,
+ "step": 435
+ },
+ {
+ "epoch": 6.05589519650655,
+ "grad_norm": 0.02917475439608097,
+ "learning_rate": 0.0006,
+ "loss": 6.453164577484131,
+ "step": 436
+ },
+ {
+ "epoch": 6.069868995633188,
+ "grad_norm": 0.01811208389699459,
+ "learning_rate": 0.0006,
+ "loss": 6.396003246307373,
+ "step": 437
+ },
+ {
+ "epoch": 6.083842794759826,
+ "grad_norm": 0.027235351502895355,
+ "learning_rate": 0.0006,
+ "loss": 6.402976036071777,
+ "step": 438
+ },
+ {
+ "epoch": 6.097816593886463,
+ "grad_norm": 0.02835794910788536,
+ "learning_rate": 0.0006,
+ "loss": 6.380568504333496,
+ "step": 439
+ },
+ {
+ "epoch": 6.1117903930131,
+ "grad_norm": 0.020868154242634773,
+ "learning_rate": 0.0006,
+ "loss": 6.38240909576416,
+ "step": 440
+ },
+ {
+ "epoch": 6.125764192139738,
+ "grad_norm": 0.02113114856183529,
+ "learning_rate": 0.0006,
+ "loss": 6.377391815185547,
+ "step": 441
+ },
+ {
+ "epoch": 6.139737991266376,
+ "grad_norm": 0.018894119188189507,
+ "learning_rate": 0.0006,
+ "loss": 6.354569435119629,
+ "step": 442
+ },
+ {
+ "epoch": 6.153711790393013,
+ "grad_norm": 0.01820225641131401,
+ "learning_rate": 0.0006,
+ "loss": 6.392882347106934,
+ "step": 443
+ },
+ {
+ "epoch": 6.1676855895196505,
+ "grad_norm": 0.017510728910565376,
+ "learning_rate": 0.0006,
+ "loss": 6.364162445068359,
+ "step": 444
+ },
+ {
+ "epoch": 6.181659388646288,
+ "grad_norm": 0.017893686890602112,
+ "learning_rate": 0.0006,
+ "loss": 6.373456954956055,
+ "step": 445
+ },
+ {
+ "epoch": 6.195633187772926,
+ "grad_norm": 0.013382102362811565,
+ "learning_rate": 0.0006,
+ "loss": 6.3714399337768555,
+ "step": 446
+ },
+ {
+ "epoch": 6.209606986899563,
+ "grad_norm": 0.022540833801031113,
+ "learning_rate": 0.0006,
+ "loss": 6.348170757293701,
+ "step": 447
+ },
+ {
+ "epoch": 6.223580786026201,
+ "grad_norm": 0.02110104262828827,
+ "learning_rate": 0.0006,
+ "loss": 6.413206100463867,
+ "step": 448
+ },
+ {
+ "epoch": 6.2375545851528384,
+ "grad_norm": 0.014498349279165268,
+ "learning_rate": 0.0006,
+ "loss": 6.361384391784668,
+ "step": 449
+ },
+ {
+ "epoch": 6.251528384279476,
+ "grad_norm": 0.01819637045264244,
+ "learning_rate": 0.0006,
+ "loss": 6.315275192260742,
+ "step": 450
+ },
+ {
+ "epoch": 6.265502183406113,
+ "grad_norm": 0.0195972491055727,
+ "learning_rate": 0.0006,
+ "loss": 6.321994781494141,
+ "step": 451
+ },
+ {
+ "epoch": 6.279475982532751,
+ "grad_norm": 0.018822453916072845,
+ "learning_rate": 0.0006,
+ "loss": 6.345219612121582,
+ "step": 452
+ },
+ {
+ "epoch": 6.293449781659389,
+ "grad_norm": 0.019247712567448616,
+ "learning_rate": 0.0006,
+ "loss": 6.348073959350586,
+ "step": 453
+ },
+ {
+ "epoch": 6.307423580786026,
+ "grad_norm": 0.03175375238060951,
+ "learning_rate": 0.0006,
+ "loss": 6.300344944000244,
+ "step": 454
+ },
+ {
+ "epoch": 6.321397379912664,
+ "grad_norm": 0.06326444447040558,
+ "learning_rate": 0.0006,
+ "loss": 6.392579078674316,
+ "step": 455
+ },
+ {
+ "epoch": 6.335371179039301,
+ "grad_norm": 0.09667232632637024,
+ "learning_rate": 0.0006,
+ "loss": 6.36024284362793,
+ "step": 456
+ },
+ {
+ "epoch": 6.349344978165939,
+ "grad_norm": 0.05774446204304695,
+ "learning_rate": 0.0006,
+ "loss": 6.365338325500488,
+ "step": 457
+ },
+ {
+ "epoch": 6.3633187772925766,
+ "grad_norm": 0.08793356269598007,
+ "learning_rate": 0.0006,
+ "loss": 6.43284273147583,
+ "step": 458
+ },
+ {
+ "epoch": 6.377292576419214,
+ "grad_norm": 0.0640595331788063,
+ "learning_rate": 0.0006,
+ "loss": 6.342428207397461,
+ "step": 459
+ },
+ {
+ "epoch": 6.391266375545851,
+ "grad_norm": 0.0447838194668293,
+ "learning_rate": 0.0006,
+ "loss": 6.367971897125244,
+ "step": 460
+ },
+ {
+ "epoch": 6.405240174672489,
+ "grad_norm": 0.04768967255949974,
+ "learning_rate": 0.0006,
+ "loss": 6.35027551651001,
+ "step": 461
+ },
+ {
+ "epoch": 6.419213973799127,
+ "grad_norm": 0.027981851249933243,
+ "learning_rate": 0.0006,
+ "loss": 6.33723258972168,
+ "step": 462
+ },
+ {
+ "epoch": 6.4331877729257645,
+ "grad_norm": 0.03479623422026634,
+ "learning_rate": 0.0006,
+ "loss": 6.36590051651001,
+ "step": 463
+ },
+ {
+ "epoch": 6.447161572052401,
+ "grad_norm": 0.020846687257289886,
+ "learning_rate": 0.0006,
+ "loss": 6.315143585205078,
+ "step": 464
+ },
+ {
+ "epoch": 6.461135371179039,
+ "grad_norm": 0.0312928780913353,
+ "learning_rate": 0.0006,
+ "loss": 6.340150833129883,
+ "step": 465
+ },
+ {
+ "epoch": 6.475109170305677,
+ "grad_norm": 0.020593881607055664,
+ "learning_rate": 0.0006,
+ "loss": 6.348175048828125,
+ "step": 466
+ },
+ {
+ "epoch": 6.489082969432315,
+ "grad_norm": 0.022010236978530884,
+ "learning_rate": 0.0006,
+ "loss": 6.318607330322266,
+ "step": 467
+ },
+ {
+ "epoch": 6.503056768558952,
+ "grad_norm": 0.018427059054374695,
+ "learning_rate": 0.0006,
+ "loss": 6.331822395324707,
+ "step": 468
+ },
+ {
+ "epoch": 6.517030567685589,
+ "grad_norm": 0.019964968785643578,
+ "learning_rate": 0.0006,
+ "loss": 6.287320137023926,
+ "step": 469
+ },
+ {
+ "epoch": 6.531004366812227,
+ "grad_norm": 0.02130270190536976,
+ "learning_rate": 0.0006,
+ "loss": 6.353344917297363,
+ "step": 470
+ },
+ {
+ "epoch": 6.544978165938865,
+ "grad_norm": 0.01616514101624489,
+ "learning_rate": 0.0006,
+ "loss": 6.314052581787109,
+ "step": 471
+ },
+ {
+ "epoch": 6.558951965065502,
+ "grad_norm": 0.016639474779367447,
+ "learning_rate": 0.0006,
+ "loss": 6.320351600646973,
+ "step": 472
+ },
+ {
+ "epoch": 6.5729257641921395,
+ "grad_norm": 0.017620692029595375,
+ "learning_rate": 0.0006,
+ "loss": 6.311459064483643,
+ "step": 473
+ },
+ {
+ "epoch": 6.586899563318777,
+ "grad_norm": 0.013542844913899899,
+ "learning_rate": 0.0006,
+ "loss": 6.324368476867676,
+ "step": 474
+ },
+ {
+ "epoch": 6.600873362445415,
+ "grad_norm": 0.01308678649365902,
+ "learning_rate": 0.0006,
+ "loss": 6.307419776916504,
+ "step": 475
+ },
+ {
+ "epoch": 6.614847161572053,
+ "grad_norm": 0.014672392047941685,
+ "learning_rate": 0.0006,
+ "loss": 6.2785420417785645,
+ "step": 476
+ },
+ {
+ "epoch": 6.62882096069869,
+ "grad_norm": 0.01399585697799921,
+ "learning_rate": 0.0006,
+ "loss": 6.258619785308838,
+ "step": 477
+ },
+ {
+ "epoch": 6.642794759825327,
+ "grad_norm": 0.01330904383212328,
+ "learning_rate": 0.0006,
+ "loss": 6.299595832824707,
+ "step": 478
+ },
+ {
+ "epoch": 6.656768558951965,
+ "grad_norm": 0.016613377258181572,
+ "learning_rate": 0.0006,
+ "loss": 6.2900309562683105,
+ "step": 479
+ },
+ {
+ "epoch": 6.670742358078603,
+ "grad_norm": 0.015076331794261932,
+ "learning_rate": 0.0006,
+ "loss": 6.231028079986572,
+ "step": 480
+ },
+ {
+ "epoch": 6.68471615720524,
+ "grad_norm": 0.01589728519320488,
+ "learning_rate": 0.0006,
+ "loss": 6.257922172546387,
+ "step": 481
+ },
+ {
+ "epoch": 6.698689956331878,
+ "grad_norm": 0.030972328037023544,
+ "learning_rate": 0.0006,
+ "loss": 6.285238265991211,
+ "step": 482
+ },
+ {
+ "epoch": 6.712663755458515,
+ "grad_norm": 0.05048783868551254,
+ "learning_rate": 0.0006,
+ "loss": 6.287471771240234,
+ "step": 483
+ },
+ {
+ "epoch": 6.726637554585153,
+ "grad_norm": 0.07390714436769485,
+ "learning_rate": 0.0006,
+ "loss": 6.284293174743652,
+ "step": 484
+ },
+ {
+ "epoch": 6.74061135371179,
+ "grad_norm": 0.07166943699121475,
+ "learning_rate": 0.0006,
+ "loss": 6.290704727172852,
+ "step": 485
+ },
+ {
+ "epoch": 6.754585152838428,
+ "grad_norm": 0.028792452067136765,
+ "learning_rate": 0.0006,
+ "loss": 6.243415832519531,
+ "step": 486
+ },
+ {
+ "epoch": 6.7685589519650655,
+ "grad_norm": 0.04822041466832161,
+ "learning_rate": 0.0006,
+ "loss": 6.278229713439941,
+ "step": 487
+ },
+ {
+ "epoch": 6.782532751091703,
+ "grad_norm": 0.036688923835754395,
+ "learning_rate": 0.0006,
+ "loss": 6.275590896606445,
+ "step": 488
+ },
+ {
+ "epoch": 6.796506550218341,
+ "grad_norm": 0.03679114952683449,
+ "learning_rate": 0.0006,
+ "loss": 6.236509323120117,
+ "step": 489
+ },
+ {
+ "epoch": 6.810480349344978,
+ "grad_norm": 0.023249102756381035,
+ "learning_rate": 0.0006,
+ "loss": 6.230788230895996,
+ "step": 490
+ },
+ {
+ "epoch": 6.824454148471616,
+ "grad_norm": 0.02682286500930786,
+ "learning_rate": 0.0006,
+ "loss": 6.278890609741211,
+ "step": 491
+ },
+ {
+ "epoch": 6.8384279475982535,
+ "grad_norm": 0.02202044241130352,
+ "learning_rate": 0.0006,
+ "loss": 6.234531402587891,
+ "step": 492
+ },
+ {
+ "epoch": 6.85240174672489,
+ "grad_norm": 0.02545459009706974,
+ "learning_rate": 0.0006,
+ "loss": 6.271790981292725,
+ "step": 493
+ },
+ {
+ "epoch": 6.866375545851528,
+ "grad_norm": 0.02012932486832142,
+ "learning_rate": 0.0006,
+ "loss": 6.247462749481201,
+ "step": 494
+ },
+ {
+ "epoch": 6.880349344978166,
+ "grad_norm": 0.02216474711894989,
+ "learning_rate": 0.0006,
+ "loss": 6.26396369934082,
+ "step": 495
+ },
+ {
+ "epoch": 6.894323144104804,
+ "grad_norm": 0.016078704968094826,
+ "learning_rate": 0.0006,
+ "loss": 6.228609085083008,
+ "step": 496
+ },
+ {
+ "epoch": 6.908296943231441,
+ "grad_norm": 0.023021478205919266,
+ "learning_rate": 0.0006,
+ "loss": 6.197244167327881,
+ "step": 497
+ },
+ {
+ "epoch": 6.922270742358078,
+ "grad_norm": 0.024766333401203156,
+ "learning_rate": 0.0006,
+ "loss": 6.181090354919434,
+ "step": 498
+ },
+ {
+ "epoch": 6.936244541484716,
+ "grad_norm": 0.025581466034054756,
+ "learning_rate": 0.0006,
+ "loss": 6.201817512512207,
+ "step": 499
+ },
+ {
+ "epoch": 6.950218340611354,
+ "grad_norm": 0.016964981332421303,
+ "learning_rate": 0.0006,
+ "loss": 6.201486587524414,
+ "step": 500
+ },
+ {
+ "epoch": 6.964192139737992,
+ "grad_norm": 0.023957721889019012,
+ "learning_rate": 0.0006,
+ "loss": 6.185701847076416,
+ "step": 501
+ },
+ {
+ "epoch": 6.978165938864628,
+ "grad_norm": 0.029775245115160942,
+ "learning_rate": 0.0006,
+ "loss": 6.242060661315918,
+ "step": 502
+ },
+ {
+ "epoch": 6.992139737991266,
+ "grad_norm": 0.04007191210985184,
+ "learning_rate": 0.0006,
+ "loss": 6.19931697845459,
+ "step": 503
+ },
+ {
+ "epoch": 7.0,
+ "grad_norm": 0.04882276430726051,
+ "learning_rate": 0.0006,
+ "loss": 6.1880598068237305,
+ "step": 504
+ },
+ {
+ "epoch": 7.0,
+ "eval_loss": 7.018043518066406,
+ "eval_runtime": 59.8815,
+ "eval_samples_per_second": 40.781,
+ "eval_steps_per_second": 1.286,
+ "step": 504
+ },
+ {
+ "epoch": 7.013973799126638,
+ "grad_norm": 0.07144106924533844,
+ "learning_rate": 0.0006,
+ "loss": 6.172799110412598,
+ "step": 505
+ },
+ {
+ "epoch": 7.0279475982532755,
+ "grad_norm": 0.06712841987609863,
+ "learning_rate": 0.0006,
+ "loss": 6.228682518005371,
+ "step": 506
+ },
+ {
+ "epoch": 7.041921397379912,
+ "grad_norm": 0.04405239224433899,
+ "learning_rate": 0.0006,
+ "loss": 6.1827898025512695,
+ "step": 507
+ },
+ {
+ "epoch": 7.05589519650655,
+ "grad_norm": 0.03418995440006256,
+ "learning_rate": 0.0006,
+ "loss": 6.200876235961914,
+ "step": 508
+ },
+ {
+ "epoch": 7.069868995633188,
+ "grad_norm": 0.04002382606267929,
+ "learning_rate": 0.0006,
+ "loss": 6.1799211502075195,
+ "step": 509
+ },
+ {
+ "epoch": 7.083842794759826,
+ "grad_norm": 0.029748762026429176,
+ "learning_rate": 0.0006,
+ "loss": 6.125422477722168,
+ "step": 510
+ },
+ {
+ "epoch": 7.097816593886463,
+ "grad_norm": 0.03408357501029968,
+ "learning_rate": 0.0006,
+ "loss": 6.203458309173584,
+ "step": 511
+ },
+ {
+ "epoch": 7.1117903930131,
+ "grad_norm": 0.02227279357612133,
+ "learning_rate": 0.0006,
+ "loss": 6.135439872741699,
+ "step": 512
+ },
+ {
+ "epoch": 7.125764192139738,
+ "grad_norm": 0.02727862447500229,
+ "learning_rate": 0.0006,
+ "loss": 6.202380180358887,
+ "step": 513
+ },
+ {
+ "epoch": 7.139737991266376,
+ "grad_norm": 0.026019075885415077,
+ "learning_rate": 0.0006,
+ "loss": 6.201343536376953,
+ "step": 514
+ },
+ {
+ "epoch": 7.153711790393013,
+ "grad_norm": 0.027227282524108887,
+ "learning_rate": 0.0006,
+ "loss": 6.164059638977051,
+ "step": 515
+ },
+ {
+ "epoch": 7.1676855895196505,
+ "grad_norm": 0.027649980038404465,
+ "learning_rate": 0.0006,
+ "loss": 6.135448932647705,
+ "step": 516
+ },
+ {
+ "epoch": 7.181659388646288,
+ "grad_norm": 0.02168039232492447,
+ "learning_rate": 0.0006,
+ "loss": 6.154831886291504,
+ "step": 517
+ },
+ {
+ "epoch": 7.195633187772926,
+ "grad_norm": 0.02101467363536358,
+ "learning_rate": 0.0006,
+ "loss": 6.128695487976074,
+ "step": 518
+ },
+ {
+ "epoch": 7.209606986899563,
+ "grad_norm": 0.018344245851039886,
+ "learning_rate": 0.0006,
+ "loss": 6.136524200439453,
+ "step": 519
+ },
+ {
+ "epoch": 7.223580786026201,
+ "grad_norm": 0.020323453471064568,
+ "learning_rate": 0.0006,
+ "loss": 6.1206488609313965,
+ "step": 520
+ },
+ {
+ "epoch": 7.2375545851528384,
+ "grad_norm": 0.01750849559903145,
+ "learning_rate": 0.0006,
+ "loss": 6.128170967102051,
+ "step": 521
+ },
+ {
+ "epoch": 7.251528384279476,
+ "grad_norm": 0.017778221517801285,
+ "learning_rate": 0.0006,
+ "loss": 6.150458812713623,
+ "step": 522
+ },
+ {
+ "epoch": 7.265502183406113,
+ "grad_norm": 0.01863938942551613,
+ "learning_rate": 0.0006,
+ "loss": 6.100106716156006,
+ "step": 523
+ },
+ {
+ "epoch": 7.279475982532751,
+ "grad_norm": 0.020196080207824707,
+ "learning_rate": 0.0006,
+ "loss": 6.149346351623535,
+ "step": 524
+ },
+ {
+ "epoch": 7.293449781659389,
+ "grad_norm": 0.03203890845179558,
+ "learning_rate": 0.0006,
+ "loss": 6.124648094177246,
+ "step": 525
+ },
+ {
+ "epoch": 7.307423580786026,
+ "grad_norm": 0.05325170233845711,
+ "learning_rate": 0.0006,
+ "loss": 6.115618705749512,
+ "step": 526
+ },
+ {
+ "epoch": 7.321397379912664,
+ "grad_norm": 0.05888345465064049,
+ "learning_rate": 0.0006,
+ "loss": 6.165221214294434,
+ "step": 527
+ },
+ {
+ "epoch": 7.335371179039301,
+ "grad_norm": 0.026873115450143814,
+ "learning_rate": 0.0006,
+ "loss": 6.039590835571289,
+ "step": 528
+ },
+ {
+ "epoch": 7.349344978165939,
+ "grad_norm": 0.06304649263620377,
+ "learning_rate": 0.0006,
+ "loss": 6.131900787353516,
+ "step": 529
+ },
+ {
+ "epoch": 7.3633187772925766,
+ "grad_norm": 0.08567557483911514,
+ "learning_rate": 0.0006,
+ "loss": 6.218114852905273,
+ "step": 530
+ },
+ {
+ "epoch": 7.377292576419214,
+ "grad_norm": 0.08868218213319778,
+ "learning_rate": 0.0006,
+ "loss": 6.216706275939941,
+ "step": 531
+ },
+ {
+ "epoch": 7.391266375545851,
+ "grad_norm": 0.03955601155757904,
+ "learning_rate": 0.0006,
+ "loss": 6.1566925048828125,
+ "step": 532
+ },
+ {
+ "epoch": 7.405240174672489,
+ "grad_norm": 0.04979498311877251,
+ "learning_rate": 0.0006,
+ "loss": 6.19327449798584,
+ "step": 533
+ },
+ {
+ "epoch": 7.419213973799127,
+ "grad_norm": 0.03253026679158211,
+ "learning_rate": 0.0006,
+ "loss": 6.133993148803711,
+ "step": 534
+ },
+ {
+ "epoch": 7.4331877729257645,
+ "grad_norm": 0.03016652725636959,
+ "learning_rate": 0.0006,
+ "loss": 6.174962997436523,
+ "step": 535
+ },
+ {
+ "epoch": 7.447161572052401,
+ "grad_norm": 0.03159862384200096,
+ "learning_rate": 0.0006,
+ "loss": 6.082124710083008,
+ "step": 536
+ },
+ {
+ "epoch": 7.461135371179039,
+ "grad_norm": 0.027673959732055664,
+ "learning_rate": 0.0006,
+ "loss": 6.111873149871826,
+ "step": 537
+ },
+ {
+ "epoch": 7.475109170305677,
+ "grad_norm": 0.022789714857935905,
+ "learning_rate": 0.0006,
+ "loss": 6.129437446594238,
+ "step": 538
+ },
+ {
+ "epoch": 7.489082969432315,
+ "grad_norm": 0.030061012133955956,
+ "learning_rate": 0.0006,
+ "loss": 6.126370429992676,
+ "step": 539
+ },
+ {
+ "epoch": 7.503056768558952,
+ "grad_norm": 0.019007068127393723,
+ "learning_rate": 0.0006,
+ "loss": 6.07887077331543,
+ "step": 540
+ },
+ {
+ "epoch": 7.517030567685589,
+ "grad_norm": 0.01897653006017208,
+ "learning_rate": 0.0006,
+ "loss": 6.096936225891113,
+ "step": 541
+ },
+ {
+ "epoch": 7.531004366812227,
+ "grad_norm": 0.024659698829054832,
+ "learning_rate": 0.0006,
+ "loss": 6.0933380126953125,
+ "step": 542
+ },
+ {
+ "epoch": 7.544978165938865,
+ "grad_norm": 0.021924549713730812,
+ "learning_rate": 0.0006,
+ "loss": 6.105311870574951,
+ "step": 543
+ },
+ {
+ "epoch": 7.558951965065502,
+ "grad_norm": 0.018817558884620667,
+ "learning_rate": 0.0006,
+ "loss": 6.058895111083984,
+ "step": 544
+ },
+ {
+ "epoch": 7.5729257641921395,
+ "grad_norm": 0.02477840520441532,
+ "learning_rate": 0.0006,
+ "loss": 6.038928031921387,
+ "step": 545
+ },
+ {
+ "epoch": 7.586899563318777,
+ "grad_norm": 0.018968632444739342,
+ "learning_rate": 0.0006,
+ "loss": 6.1368937492370605,
+ "step": 546
+ },
+ {
+ "epoch": 7.600873362445415,
+ "grad_norm": 0.02409801445901394,
+ "learning_rate": 0.0006,
+ "loss": 6.110324859619141,
+ "step": 547
+ },
+ {
+ "epoch": 7.614847161572053,
+ "grad_norm": 0.017595326527953148,
+ "learning_rate": 0.0006,
+ "loss": 6.074368476867676,
+ "step": 548
+ },
+ {
+ "epoch": 7.62882096069869,
+ "grad_norm": 0.01875406503677368,
+ "learning_rate": 0.0006,
+ "loss": 6.121023654937744,
+ "step": 549
+ },
+ {
+ "epoch": 7.642794759825327,
+ "grad_norm": 0.018781628459692,
+ "learning_rate": 0.0006,
+ "loss": 6.064914703369141,
+ "step": 550
+ },
+ {
+ "epoch": 7.656768558951965,
+ "grad_norm": 0.017003126442432404,
+ "learning_rate": 0.0006,
+ "loss": 6.039608001708984,
+ "step": 551
+ },
+ {
+ "epoch": 7.670742358078603,
+ "grad_norm": 0.017665037885308266,
+ "learning_rate": 0.0006,
+ "loss": 6.053671836853027,
+ "step": 552
+ },
+ {
+ "epoch": 7.68471615720524,
+ "grad_norm": 0.02066265605390072,
+ "learning_rate": 0.0006,
+ "loss": 6.090536117553711,
+ "step": 553
+ },
+ {
+ "epoch": 7.698689956331878,
+ "grad_norm": 0.02888769656419754,
+ "learning_rate": 0.0006,
+ "loss": 5.9858903884887695,
+ "step": 554
+ },
+ {
+ "epoch": 7.712663755458515,
+ "grad_norm": 0.03744485229253769,
+ "learning_rate": 0.0006,
+ "loss": 6.086264610290527,
+ "step": 555
+ },
+ {
+ "epoch": 7.726637554585153,
+ "grad_norm": 0.04719431325793266,
+ "learning_rate": 0.0006,
+ "loss": 6.09747314453125,
+ "step": 556
+ },
+ {
+ "epoch": 7.74061135371179,
+ "grad_norm": 0.044963907450437546,
+ "learning_rate": 0.0006,
+ "loss": 6.091272354125977,
+ "step": 557
+ },
+ {
+ "epoch": 7.754585152838428,
+ "grad_norm": 0.03810194507241249,
+ "learning_rate": 0.0006,
+ "loss": 6.084904193878174,
+ "step": 558
+ },
+ {
+ "epoch": 7.7685589519650655,
+ "grad_norm": 0.04626769199967384,
+ "learning_rate": 0.0006,
+ "loss": 6.017384052276611,
+ "step": 559
+ },
+ {
+ "epoch": 7.782532751091703,
+ "grad_norm": 0.03459552302956581,
+ "learning_rate": 0.0006,
+ "loss": 6.111314296722412,
+ "step": 560
+ },
+ {
+ "epoch": 7.796506550218341,
+ "grad_norm": 0.027385666966438293,
+ "learning_rate": 0.0006,
+ "loss": 6.002509117126465,
+ "step": 561
+ },
+ {
+ "epoch": 7.810480349344978,
+ "grad_norm": 0.03417652100324631,
+ "learning_rate": 0.0006,
+ "loss": 6.0421061515808105,
+ "step": 562
+ },
+ {
+ "epoch": 7.824454148471616,
+ "grad_norm": 0.04415450245141983,
+ "learning_rate": 0.0006,
+ "loss": 6.035305023193359,
+ "step": 563
+ },
+ {
+ "epoch": 7.8384279475982535,
+ "grad_norm": 0.03259635716676712,
+ "learning_rate": 0.0006,
+ "loss": 6.025840759277344,
+ "step": 564
+ },
+ {
+ "epoch": 7.85240174672489,
+ "grad_norm": 0.024941129609942436,
+ "learning_rate": 0.0006,
+ "loss": 6.043416976928711,
+ "step": 565
+ },
+ {
+ "epoch": 7.866375545851528,
+ "grad_norm": 0.02434500865638256,
+ "learning_rate": 0.0006,
+ "loss": 6.002052307128906,
+ "step": 566
+ },
+ {
+ "epoch": 7.880349344978166,
+ "grad_norm": 0.01876584254205227,
+ "learning_rate": 0.0006,
+ "loss": 6.045035362243652,
+ "step": 567
+ },
+ {
+ "epoch": 7.894323144104804,
+ "grad_norm": 0.02254367619752884,
+ "learning_rate": 0.0006,
+ "loss": 5.972503662109375,
+ "step": 568
+ },
+ {
+ "epoch": 7.908296943231441,
+ "grad_norm": 0.024504603818058968,
+ "learning_rate": 0.0006,
+ "loss": 5.992846488952637,
+ "step": 569
+ },
+ {
+ "epoch": 7.922270742358078,
+ "grad_norm": 0.02317827194929123,
+ "learning_rate": 0.0006,
+ "loss": 6.007697582244873,
+ "step": 570
+ },
+ {
+ "epoch": 7.936244541484716,
+ "grad_norm": 0.02225821278989315,
+ "learning_rate": 0.0006,
+ "loss": 6.01463508605957,
+ "step": 571
+ },
+ {
+ "epoch": 7.950218340611354,
+ "grad_norm": 0.01811717264354229,
+ "learning_rate": 0.0006,
+ "loss": 5.995022773742676,
+ "step": 572
+ },
+ {
+ "epoch": 7.964192139737992,
+ "grad_norm": 0.019589431583881378,
+ "learning_rate": 0.0006,
+ "loss": 5.990633964538574,
+ "step": 573
+ },
+ {
+ "epoch": 7.978165938864628,
+ "grad_norm": 0.015422016382217407,
+ "learning_rate": 0.0006,
+ "loss": 6.011263847351074,
+ "step": 574
+ },
+ {
+ "epoch": 7.992139737991266,
+ "grad_norm": 0.023298880085349083,
+ "learning_rate": 0.0006,
+ "loss": 5.9954938888549805,
+ "step": 575
+ },
+ {
+ "epoch": 8.0,
+ "grad_norm": 0.022172044962644577,
+ "learning_rate": 0.0006,
+ "loss": 5.95679235458374,
+ "step": 576
+ },
+ {
+ "epoch": 8.0,
+ "eval_loss": 6.75809383392334,
+ "eval_runtime": 72.8926,
+ "eval_samples_per_second": 33.501,
+ "eval_steps_per_second": 1.056,
+ "step": 576
+ },
+ {
+ "epoch": 8.013973799126637,
+ "grad_norm": 0.029082221910357475,
+ "learning_rate": 0.0006,
+ "loss": 6.018675804138184,
+ "step": 577
+ },
+ {
+ "epoch": 8.027947598253276,
+ "grad_norm": 0.03959495946764946,
+ "learning_rate": 0.0006,
+ "loss": 5.93310546875,
+ "step": 578
+ },
+ {
+ "epoch": 8.041921397379912,
+ "grad_norm": 0.043756261467933655,
+ "learning_rate": 0.0006,
+ "loss": 5.965944290161133,
+ "step": 579
+ },
+ {
+ "epoch": 8.055895196506551,
+ "grad_norm": 0.04996631667017937,
+ "learning_rate": 0.0006,
+ "loss": 6.0381059646606445,
+ "step": 580
+ },
+ {
+ "epoch": 8.069868995633188,
+ "grad_norm": 0.05721999332308769,
+ "learning_rate": 0.0006,
+ "loss": 5.968381881713867,
+ "step": 581
+ },
+ {
+ "epoch": 8.083842794759825,
+ "grad_norm": 0.07594525814056396,
+ "learning_rate": 0.0006,
+ "loss": 6.027225494384766,
+ "step": 582
+ },
+ {
+ "epoch": 8.097816593886463,
+ "grad_norm": 0.06861685961484909,
+ "learning_rate": 0.0006,
+ "loss": 6.035175323486328,
+ "step": 583
+ },
+ {
+ "epoch": 8.1117903930131,
+ "grad_norm": 0.04156472906470299,
+ "learning_rate": 0.0006,
+ "loss": 5.9500017166137695,
+ "step": 584
+ },
+ {
+ "epoch": 8.125764192139737,
+ "grad_norm": 0.045399412512779236,
+ "learning_rate": 0.0006,
+ "loss": 5.9953389167785645,
+ "step": 585
+ },
+ {
+ "epoch": 8.139737991266376,
+ "grad_norm": 0.04399585723876953,
+ "learning_rate": 0.0006,
+ "loss": 5.986051559448242,
+ "step": 586
+ },
+ {
+ "epoch": 8.153711790393013,
+ "grad_norm": 0.05764509737491608,
+ "learning_rate": 0.0006,
+ "loss": 6.005392074584961,
+ "step": 587
+ },
+ {
+ "epoch": 8.167685589519651,
+ "grad_norm": 0.060871850699186325,
+ "learning_rate": 0.0006,
+ "loss": 6.01304817199707,
+ "step": 588
+ },
+ {
+ "epoch": 8.181659388646288,
+ "grad_norm": 0.03381241112947464,
+ "learning_rate": 0.0006,
+ "loss": 5.977019309997559,
+ "step": 589
+ },
+ {
+ "epoch": 8.195633187772925,
+ "grad_norm": 0.03384104371070862,
+ "learning_rate": 0.0006,
+ "loss": 5.988002777099609,
+ "step": 590
+ },
+ {
+ "epoch": 8.209606986899564,
+ "grad_norm": 0.03120286948978901,
+ "learning_rate": 0.0006,
+ "loss": 5.968227386474609,
+ "step": 591
+ },
+ {
+ "epoch": 8.2235807860262,
+ "grad_norm": 0.026532653719186783,
+ "learning_rate": 0.0006,
+ "loss": 5.936810493469238,
+ "step": 592
+ },
+ {
+ "epoch": 8.237554585152838,
+ "grad_norm": 0.024956922978162766,
+ "learning_rate": 0.0006,
+ "loss": 5.9348015785217285,
+ "step": 593
+ },
+ {
+ "epoch": 8.251528384279476,
+ "grad_norm": 0.025926144793629646,
+ "learning_rate": 0.0006,
+ "loss": 5.952630519866943,
+ "step": 594
+ },
+ {
+ "epoch": 8.265502183406113,
+ "grad_norm": 0.02893793396651745,
+ "learning_rate": 0.0006,
+ "loss": 5.9569854736328125,
+ "step": 595
+ },
+ {
+ "epoch": 8.279475982532752,
+ "grad_norm": 0.0320001058280468,
+ "learning_rate": 0.0006,
+ "loss": 5.917050838470459,
+ "step": 596
+ },
+ {
+ "epoch": 8.293449781659389,
+ "grad_norm": 0.03925463184714317,
+ "learning_rate": 0.0006,
+ "loss": 5.9925127029418945,
+ "step": 597
+ },
+ {
+ "epoch": 8.307423580786025,
+ "grad_norm": 0.05096640810370445,
+ "learning_rate": 0.0006,
+ "loss": 5.926358222961426,
+ "step": 598
+ },
+ {
+ "epoch": 8.321397379912664,
+ "grad_norm": 0.041637953370809555,
+ "learning_rate": 0.0006,
+ "loss": 5.954983711242676,
+ "step": 599
+ },
+ {
+ "epoch": 8.335371179039301,
+ "grad_norm": 0.03138125315308571,
+ "learning_rate": 0.0006,
+ "loss": 5.893692970275879,
+ "step": 600
+ },
+ {
+ "epoch": 8.34934497816594,
+ "grad_norm": 0.029299303889274597,
+ "learning_rate": 0.0006,
+ "loss": 5.94670295715332,
+ "step": 601
+ },
+ {
+ "epoch": 8.363318777292577,
+ "grad_norm": 0.023969192057847977,
+ "learning_rate": 0.0006,
+ "loss": 5.925658702850342,
+ "step": 602
+ },
+ {
+ "epoch": 8.377292576419213,
+ "grad_norm": 0.025406932458281517,
+ "learning_rate": 0.0006,
+ "loss": 5.966875076293945,
+ "step": 603
+ },
+ {
+ "epoch": 8.391266375545852,
+ "grad_norm": 0.024309059605002403,
+ "learning_rate": 0.0006,
+ "loss": 5.932581424713135,
+ "step": 604
+ },
+ {
+ "epoch": 8.405240174672489,
+ "grad_norm": 0.017237074673175812,
+ "learning_rate": 0.0006,
+ "loss": 5.898823261260986,
+ "step": 605
+ },
+ {
+ "epoch": 8.419213973799126,
+ "grad_norm": 0.025594931095838547,
+ "learning_rate": 0.0006,
+ "loss": 5.9285993576049805,
+ "step": 606
+ },
+ {
+ "epoch": 8.433187772925764,
+ "grad_norm": 0.025184160098433495,
+ "learning_rate": 0.0006,
+ "loss": 5.915966510772705,
+ "step": 607
+ },
+ {
+ "epoch": 8.447161572052401,
+ "grad_norm": 0.019004398956894875,
+ "learning_rate": 0.0006,
+ "loss": 5.893395900726318,
+ "step": 608
+ },
+ {
+ "epoch": 8.46113537117904,
+ "grad_norm": 0.021802261471748352,
+ "learning_rate": 0.0006,
+ "loss": 5.953311443328857,
+ "step": 609
+ },
+ {
+ "epoch": 8.475109170305677,
+ "grad_norm": 0.02202655002474785,
+ "learning_rate": 0.0006,
+ "loss": 5.953000545501709,
+ "step": 610
+ },
+ {
+ "epoch": 8.489082969432314,
+ "grad_norm": 0.02067241445183754,
+ "learning_rate": 0.0006,
+ "loss": 5.869247913360596,
+ "step": 611
+ },
+ {
+ "epoch": 8.503056768558952,
+ "grad_norm": 0.015430927276611328,
+ "learning_rate": 0.0006,
+ "loss": 5.855883598327637,
+ "step": 612
+ },
+ {
+ "epoch": 8.51703056768559,
+ "grad_norm": 0.019917506724596024,
+ "learning_rate": 0.0006,
+ "loss": 5.9284539222717285,
+ "step": 613
+ },
+ {
+ "epoch": 8.531004366812226,
+ "grad_norm": 0.024291323497891426,
+ "learning_rate": 0.0006,
+ "loss": 5.973604202270508,
+ "step": 614
+ },
+ {
+ "epoch": 8.544978165938865,
+ "grad_norm": 0.022133296355605125,
+ "learning_rate": 0.0006,
+ "loss": 5.85693359375,
+ "step": 615
+ },
+ {
+ "epoch": 8.558951965065502,
+ "grad_norm": 0.0254366435110569,
+ "learning_rate": 0.0006,
+ "loss": 5.939558982849121,
+ "step": 616
+ },
+ {
+ "epoch": 8.57292576419214,
+ "grad_norm": 0.030134694650769234,
+ "learning_rate": 0.0006,
+ "loss": 5.892992973327637,
+ "step": 617
+ },
+ {
+ "epoch": 8.586899563318777,
+ "grad_norm": 0.03845192864537239,
+ "learning_rate": 0.0006,
+ "loss": 5.89303731918335,
+ "step": 618
+ },
+ {
+ "epoch": 8.600873362445414,
+ "grad_norm": 0.042409464716911316,
+ "learning_rate": 0.0006,
+ "loss": 5.914068222045898,
+ "step": 619
+ },
+ {
+ "epoch": 8.614847161572053,
+ "grad_norm": 0.0454271174967289,
+ "learning_rate": 0.0006,
+ "loss": 5.853579044342041,
+ "step": 620
+ },
+ {
+ "epoch": 8.62882096069869,
+ "grad_norm": 0.04934931918978691,
+ "learning_rate": 0.0006,
+ "loss": 5.881861686706543,
+ "step": 621
+ },
+ {
+ "epoch": 8.642794759825328,
+ "grad_norm": 0.0519695058465004,
+ "learning_rate": 0.0006,
+ "loss": 5.872620582580566,
+ "step": 622
+ },
+ {
+ "epoch": 8.656768558951965,
+ "grad_norm": 0.06172563508152962,
+ "learning_rate": 0.0006,
+ "loss": 5.938283443450928,
+ "step": 623
+ },
+ {
+ "epoch": 8.670742358078602,
+ "grad_norm": 0.08416368812322617,
+ "learning_rate": 0.0006,
+ "loss": 5.943403720855713,
+ "step": 624
+ },
+ {
+ "epoch": 8.68471615720524,
+ "grad_norm": 0.07311230152845383,
+ "learning_rate": 0.0006,
+ "loss": 5.954029083251953,
+ "step": 625
+ },
+ {
+ "epoch": 8.698689956331878,
+ "grad_norm": 0.09851036965847015,
+ "learning_rate": 0.0006,
+ "loss": 5.992746829986572,
+ "step": 626
+ },
+ {
+ "epoch": 8.712663755458514,
+ "grad_norm": 0.07521767914295197,
+ "learning_rate": 0.0006,
+ "loss": 5.956884384155273,
+ "step": 627
+ },
+ {
+ "epoch": 8.726637554585153,
+ "grad_norm": 0.05720978602766991,
+ "learning_rate": 0.0006,
+ "loss": 5.911205291748047,
+ "step": 628
+ },
+ {
+ "epoch": 8.74061135371179,
+ "grad_norm": 0.05360936373472214,
+ "learning_rate": 0.0006,
+ "loss": 5.9734787940979,
+ "step": 629
+ },
+ {
+ "epoch": 8.754585152838429,
+ "grad_norm": 0.035826463252305984,
+ "learning_rate": 0.0006,
+ "loss": 5.95751953125,
+ "step": 630
+ },
+ {
+ "epoch": 8.768558951965066,
+ "grad_norm": 0.030137093737721443,
+ "learning_rate": 0.0006,
+ "loss": 5.91911506652832,
+ "step": 631
+ },
+ {
+ "epoch": 8.782532751091702,
+ "grad_norm": 0.030039150267839432,
+ "learning_rate": 0.0006,
+ "loss": 5.910290718078613,
+ "step": 632
+ },
+ {
+ "epoch": 8.796506550218341,
+ "grad_norm": 0.03349437564611435,
+ "learning_rate": 0.0006,
+ "loss": 5.904623985290527,
+ "step": 633
+ },
+ {
+ "epoch": 8.810480349344978,
+ "grad_norm": 0.02211189828813076,
+ "learning_rate": 0.0006,
+ "loss": 5.909668445587158,
+ "step": 634
+ },
+ {
+ "epoch": 8.824454148471617,
+ "grad_norm": 0.022601254284381866,
+ "learning_rate": 0.0006,
+ "loss": 5.942885875701904,
+ "step": 635
+ },
+ {
+ "epoch": 8.838427947598253,
+ "grad_norm": 0.02197270467877388,
+ "learning_rate": 0.0006,
+ "loss": 5.867103576660156,
+ "step": 636
+ },
+ {
+ "epoch": 8.85240174672489,
+ "grad_norm": 0.022765018045902252,
+ "learning_rate": 0.0006,
+ "loss": 5.894528865814209,
+ "step": 637
+ },
+ {
+ "epoch": 8.866375545851529,
+ "grad_norm": 0.021181151270866394,
+ "learning_rate": 0.0006,
+ "loss": 5.896910667419434,
+ "step": 638
+ },
+ {
+ "epoch": 8.880349344978166,
+ "grad_norm": 0.01878993771970272,
+ "learning_rate": 0.0006,
+ "loss": 5.860502243041992,
+ "step": 639
+ },
+ {
+ "epoch": 8.894323144104803,
+ "grad_norm": 0.014679803512990475,
+ "learning_rate": 0.0006,
+ "loss": 5.8725266456604,
+ "step": 640
+ },
+ {
+ "epoch": 8.908296943231441,
+ "grad_norm": 0.017919695004820824,
+ "learning_rate": 0.0006,
+ "loss": 5.88902473449707,
+ "step": 641
+ },
+ {
+ "epoch": 8.922270742358078,
+ "grad_norm": 0.01437009871006012,
+ "learning_rate": 0.0006,
+ "loss": 5.868186950683594,
+ "step": 642
+ },
+ {
+ "epoch": 8.936244541484717,
+ "grad_norm": 0.014293290674686432,
+ "learning_rate": 0.0006,
+ "loss": 5.833364963531494,
+ "step": 643
+ },
+ {
+ "epoch": 8.950218340611354,
+ "grad_norm": 0.015440084971487522,
+ "learning_rate": 0.0006,
+ "loss": 5.860260009765625,
+ "step": 644
+ },
+ {
+ "epoch": 8.96419213973799,
+ "grad_norm": 0.014292514882981777,
+ "learning_rate": 0.0006,
+ "loss": 5.844939231872559,
+ "step": 645
+ },
+ {
+ "epoch": 8.97816593886463,
+ "grad_norm": 0.015702251344919205,
+ "learning_rate": 0.0006,
+ "loss": 5.825929164886475,
+ "step": 646
+ },
+ {
+ "epoch": 8.992139737991266,
+ "grad_norm": 0.013372798450291157,
+ "learning_rate": 0.0006,
+ "loss": 5.83062744140625,
+ "step": 647
+ },
+ {
+ "epoch": 9.0,
+ "grad_norm": 0.013794616796076298,
+ "learning_rate": 0.0006,
+ "loss": 5.7897820472717285,
+ "step": 648
+ },
+ {
+ "epoch": 9.0,
+ "eval_loss": 6.600830078125,
+ "eval_runtime": 71.4316,
+ "eval_samples_per_second": 34.187,
+ "eval_steps_per_second": 1.078,
+ "step": 648
+ },
+ {
+ "epoch": 9.013973799126637,
+ "grad_norm": 0.01262663770467043,
+ "learning_rate": 0.0006,
+ "loss": 5.802824020385742,
+ "step": 649
+ },
+ {
+ "epoch": 9.027947598253276,
+ "grad_norm": 0.016375157982110977,
+ "learning_rate": 0.0006,
+ "loss": 5.8018317222595215,
+ "step": 650
+ },
+ {
+ "epoch": 9.041921397379912,
+ "grad_norm": 0.010541851632297039,
+ "learning_rate": 0.0006,
+ "loss": 5.848208427429199,
+ "step": 651
+ },
+ {
+ "epoch": 9.055895196506551,
+ "grad_norm": 0.011726257391273975,
+ "learning_rate": 0.0006,
+ "loss": 5.771823883056641,
+ "step": 652
+ },
+ {
+ "epoch": 9.069868995633188,
+ "grad_norm": 0.01124380249530077,
+ "learning_rate": 0.0006,
+ "loss": 5.773855209350586,
+ "step": 653
+ },
+ {
+ "epoch": 9.083842794759825,
+ "grad_norm": 0.010801885277032852,
+ "learning_rate": 0.0006,
+ "loss": 5.847390651702881,
+ "step": 654
+ },
+ {
+ "epoch": 9.097816593886463,
+ "grad_norm": 0.013018375262618065,
+ "learning_rate": 0.0006,
+ "loss": 5.786786079406738,
+ "step": 655
+ },
+ {
+ "epoch": 9.1117903930131,
+ "grad_norm": 0.014233003370463848,
+ "learning_rate": 0.0006,
+ "loss": 5.7913713455200195,
+ "step": 656
+ },
+ {
+ "epoch": 9.125764192139737,
+ "grad_norm": 0.01465831883251667,
+ "learning_rate": 0.0006,
+ "loss": 5.821170330047607,
+ "step": 657
+ },
+ {
+ "epoch": 9.139737991266376,
+ "grad_norm": 0.017383430153131485,
+ "learning_rate": 0.0006,
+ "loss": 5.811469078063965,
+ "step": 658
+ },
+ {
+ "epoch": 9.153711790393013,
+ "grad_norm": 0.020328007638454437,
+ "learning_rate": 0.0006,
+ "loss": 5.718976020812988,
+ "step": 659
+ },
+ {
+ "epoch": 9.167685589519651,
+ "grad_norm": 0.01776740700006485,
+ "learning_rate": 0.0006,
+ "loss": 5.80593729019165,
+ "step": 660
+ },
+ {
+ "epoch": 9.181659388646288,
+ "grad_norm": 0.017713800072669983,
+ "learning_rate": 0.0006,
+ "loss": 5.744930267333984,
+ "step": 661
+ },
+ {
+ "epoch": 9.195633187772925,
+ "grad_norm": 0.028774067759513855,
+ "learning_rate": 0.0006,
+ "loss": 5.85262393951416,
+ "step": 662
+ },
+ {
+ "epoch": 9.209606986899564,
+ "grad_norm": 0.05153049901127815,
+ "learning_rate": 0.0006,
+ "loss": 5.739948272705078,
+ "step": 663
+ },
+ {
+ "epoch": 9.2235807860262,
+ "grad_norm": 0.06882233172655106,
+ "learning_rate": 0.0006,
+ "loss": 5.844270706176758,
+ "step": 664
+ },
+ {
+ "epoch": 9.237554585152838,
+ "grad_norm": 0.0420759879052639,
+ "learning_rate": 0.0006,
+ "loss": 5.79060173034668,
+ "step": 665
+ },
+ {
+ "epoch": 9.251528384279476,
+ "grad_norm": 0.05598323419690132,
+ "learning_rate": 0.0006,
+ "loss": 5.833735466003418,
+ "step": 666
+ },
+ {
+ "epoch": 9.265502183406113,
+ "grad_norm": 0.05011717975139618,
+ "learning_rate": 0.0006,
+ "loss": 5.840887069702148,
+ "step": 667
+ },
+ {
+ "epoch": 9.279475982532752,
+ "grad_norm": 0.040782492607831955,
+ "learning_rate": 0.0006,
+ "loss": 5.7905449867248535,
+ "step": 668
+ },
+ {
+ "epoch": 9.293449781659389,
+ "grad_norm": 0.030804185196757317,
+ "learning_rate": 0.0006,
+ "loss": 5.802691459655762,
+ "step": 669
+ },
+ {
+ "epoch": 9.307423580786025,
+ "grad_norm": 0.032001759856939316,
+ "learning_rate": 0.0006,
+ "loss": 5.726612091064453,
+ "step": 670
+ },
+ {
+ "epoch": 9.321397379912664,
+ "grad_norm": 0.0295492485165596,
+ "learning_rate": 0.0006,
+ "loss": 5.798826694488525,
+ "step": 671
+ },
+ {
+ "epoch": 9.335371179039301,
+ "grad_norm": 0.034406546503305435,
+ "learning_rate": 0.0006,
+ "loss": 5.720219135284424,
+ "step": 672
+ },
+ {
+ "epoch": 9.34934497816594,
+ "grad_norm": 0.02627669647336006,
+ "learning_rate": 0.0006,
+ "loss": 5.723702430725098,
+ "step": 673
+ },
+ {
+ "epoch": 9.363318777292577,
+ "grad_norm": 0.0282363910228014,
+ "learning_rate": 0.0006,
+ "loss": 5.818911552429199,
+ "step": 674
+ },
+ {
+ "epoch": 9.377292576419213,
+ "grad_norm": 0.017894433811306953,
+ "learning_rate": 0.0006,
+ "loss": 5.772168159484863,
+ "step": 675
+ },
+ {
+ "epoch": 9.391266375545852,
+ "grad_norm": 0.019017303362488747,
+ "learning_rate": 0.0006,
+ "loss": 5.774587631225586,
+ "step": 676
+ },
+ {
+ "epoch": 9.405240174672489,
+ "grad_norm": 0.022159283980727196,
+ "learning_rate": 0.0006,
+ "loss": 5.812951564788818,
+ "step": 677
+ },
+ {
+ "epoch": 9.419213973799126,
+ "grad_norm": 0.02198116108775139,
+ "learning_rate": 0.0006,
+ "loss": 5.808734893798828,
+ "step": 678
+ },
+ {
+ "epoch": 9.433187772925764,
+ "grad_norm": 0.017741188406944275,
+ "learning_rate": 0.0006,
+ "loss": 5.776142120361328,
+ "step": 679
+ },
+ {
+ "epoch": 9.447161572052401,
+ "grad_norm": 0.017197400331497192,
+ "learning_rate": 0.0006,
+ "loss": 5.7471208572387695,
+ "step": 680
+ },
+ {
+ "epoch": 9.46113537117904,
+ "grad_norm": 0.01752566732466221,
+ "learning_rate": 0.0006,
+ "loss": 5.780961990356445,
+ "step": 681
+ },
+ {
+ "epoch": 9.475109170305677,
+ "grad_norm": 0.017961280420422554,
+ "learning_rate": 0.0006,
+ "loss": 5.713686943054199,
+ "step": 682
+ },
+ {
+ "epoch": 9.489082969432314,
+ "grad_norm": 0.02227521874010563,
+ "learning_rate": 0.0006,
+ "loss": 5.717774868011475,
+ "step": 683
+ },
+ {
+ "epoch": 9.503056768558952,
+ "grad_norm": 0.024075504392385483,
+ "learning_rate": 0.0006,
+ "loss": 5.715015411376953,
+ "step": 684
+ },
+ {
+ "epoch": 9.51703056768559,
+ "grad_norm": 0.029195033013820648,
+ "learning_rate": 0.0006,
+ "loss": 5.659659385681152,
+ "step": 685
+ },
+ {
+ "epoch": 9.531004366812226,
+ "grad_norm": 0.02959120273590088,
+ "learning_rate": 0.0006,
+ "loss": 5.78981351852417,
+ "step": 686
+ },
+ {
+ "epoch": 9.544978165938865,
+ "grad_norm": 0.021854368969798088,
+ "learning_rate": 0.0006,
+ "loss": 5.750844955444336,
+ "step": 687
+ },
+ {
+ "epoch": 9.558951965065502,
+ "grad_norm": 0.017120081931352615,
+ "learning_rate": 0.0006,
+ "loss": 5.672736167907715,
+ "step": 688
+ },
+ {
+ "epoch": 9.57292576419214,
+ "grad_norm": 0.020712822675704956,
+ "learning_rate": 0.0006,
+ "loss": 5.708046913146973,
+ "step": 689
+ },
+ {
+ "epoch": 9.586899563318777,
+ "grad_norm": 0.019186170771718025,
+ "learning_rate": 0.0006,
+ "loss": 5.71742057800293,
+ "step": 690
+ },
+ {
+ "epoch": 9.600873362445414,
+ "grad_norm": 0.01985814794898033,
+ "learning_rate": 0.0006,
+ "loss": 5.735311508178711,
+ "step": 691
+ },
+ {
+ "epoch": 9.614847161572053,
+ "grad_norm": 0.01965234987437725,
+ "learning_rate": 0.0006,
+ "loss": 5.728727340698242,
+ "step": 692
+ },
+ {
+ "epoch": 9.62882096069869,
+ "grad_norm": 0.019827688112854958,
+ "learning_rate": 0.0006,
+ "loss": 5.743068695068359,
+ "step": 693
+ },
+ {
+ "epoch": 9.642794759825328,
+ "grad_norm": 0.025500580668449402,
+ "learning_rate": 0.0006,
+ "loss": 5.76585578918457,
+ "step": 694
+ },
+ {
+ "epoch": 9.656768558951965,
+ "grad_norm": 0.03701791539788246,
+ "learning_rate": 0.0006,
+ "loss": 5.728367805480957,
+ "step": 695
+ },
+ {
+ "epoch": 9.670742358078602,
+ "grad_norm": 0.04173659533262253,
+ "learning_rate": 0.0006,
+ "loss": 5.683089256286621,
+ "step": 696
+ },
+ {
+ "epoch": 9.68471615720524,
+ "grad_norm": 0.031390558928251266,
+ "learning_rate": 0.0006,
+ "loss": 5.740921974182129,
+ "step": 697
+ },
+ {
+ "epoch": 9.698689956331878,
+ "grad_norm": 0.03970158472657204,
+ "learning_rate": 0.0006,
+ "loss": 5.730632781982422,
+ "step": 698
+ },
+ {
+ "epoch": 9.712663755458514,
+ "grad_norm": 0.03865141049027443,
+ "learning_rate": 0.0006,
+ "loss": 5.686657905578613,
+ "step": 699
+ },
+ {
+ "epoch": 9.726637554585153,
+ "grad_norm": 0.04856107011437416,
+ "learning_rate": 0.0006,
+ "loss": 5.657513618469238,
+ "step": 700
+ },
+ {
+ "epoch": 9.74061135371179,
+ "grad_norm": 0.04016133397817612,
+ "learning_rate": 0.0006,
+ "loss": 5.754552841186523,
+ "step": 701
+ },
+ {
+ "epoch": 9.754585152838429,
+ "grad_norm": 0.03961458429694176,
+ "learning_rate": 0.0006,
+ "loss": 5.672088623046875,
+ "step": 702
+ },
+ {
+ "epoch": 9.768558951965066,
+ "grad_norm": 0.03403536602854729,
+ "learning_rate": 0.0006,
+ "loss": 5.736673355102539,
+ "step": 703
+ },
+ {
+ "epoch": 9.782532751091702,
+ "grad_norm": 0.03766333684325218,
+ "learning_rate": 0.0006,
+ "loss": 5.716219902038574,
+ "step": 704
+ },
+ {
+ "epoch": 9.796506550218341,
+ "grad_norm": 0.04085501655936241,
+ "learning_rate": 0.0006,
+ "loss": 5.678958415985107,
+ "step": 705
+ },
+ {
+ "epoch": 9.810480349344978,
+ "grad_norm": 0.036888182163238525,
+ "learning_rate": 0.0006,
+ "loss": 5.6990461349487305,
+ "step": 706
+ },
+ {
+ "epoch": 9.824454148471617,
+ "grad_norm": 0.0492791011929512,
+ "learning_rate": 0.0006,
+ "loss": 5.78053092956543,
+ "step": 707
+ },
+ {
+ "epoch": 9.838427947598253,
+ "grad_norm": 0.05572429671883583,
+ "learning_rate": 0.0006,
+ "loss": 5.7543535232543945,
+ "step": 708
+ },
+ {
+ "epoch": 9.85240174672489,
+ "grad_norm": 0.047000352293252945,
+ "learning_rate": 0.0006,
+ "loss": 5.7174272537231445,
+ "step": 709
+ },
+ {
+ "epoch": 9.866375545851529,
+ "grad_norm": 0.058798179030418396,
+ "learning_rate": 0.0006,
+ "loss": 5.793713569641113,
+ "step": 710
+ },
+ {
+ "epoch": 9.880349344978166,
+ "grad_norm": 0.046992287039756775,
+ "learning_rate": 0.0006,
+ "loss": 5.784457683563232,
+ "step": 711
+ },
+ {
+ "epoch": 9.894323144104803,
+ "grad_norm": 0.04382370039820671,
+ "learning_rate": 0.0006,
+ "loss": 5.793731212615967,
+ "step": 712
+ },
+ {
+ "epoch": 9.908296943231441,
+ "grad_norm": 0.040616173297166824,
+ "learning_rate": 0.0006,
+ "loss": 5.69599723815918,
+ "step": 713
+ },
+ {
+ "epoch": 9.922270742358078,
+ "grad_norm": 0.03446170687675476,
+ "learning_rate": 0.0006,
+ "loss": 5.694674968719482,
+ "step": 714
+ },
+ {
+ "epoch": 9.936244541484717,
+ "grad_norm": 0.029936550185084343,
+ "learning_rate": 0.0006,
+ "loss": 5.6704487800598145,
+ "step": 715
+ },
+ {
+ "epoch": 9.950218340611354,
+ "grad_norm": 0.02224753610789776,
+ "learning_rate": 0.0006,
+ "loss": 5.728894233703613,
+ "step": 716
+ },
+ {
+ "epoch": 9.96419213973799,
+ "grad_norm": 0.024346651509404182,
+ "learning_rate": 0.0006,
+ "loss": 5.772412300109863,
+ "step": 717
+ },
+ {
+ "epoch": 9.97816593886463,
+ "grad_norm": 0.022997068241238594,
+ "learning_rate": 0.0006,
+ "loss": 5.645310401916504,
+ "step": 718
+ },
+ {
+ "epoch": 9.992139737991266,
+ "grad_norm": 0.024918686598539352,
+ "learning_rate": 0.0006,
+ "loss": 5.671009063720703,
+ "step": 719
+ },
+ {
+ "epoch": 10.0,
+ "grad_norm": 0.020067067816853523,
+ "learning_rate": 0.0006,
+ "loss": 5.637441635131836,
+ "step": 720
+ },
+ {
+ "epoch": 10.0,
+ "eval_loss": 6.387807846069336,
+ "eval_runtime": 72.3372,
+ "eval_samples_per_second": 33.759,
+ "eval_steps_per_second": 1.064,
+ "step": 720
+ },
+ {
+ "epoch": 10.013973799126637,
+ "grad_norm": 0.02015046589076519,
+ "learning_rate": 0.0006,
+ "loss": 5.598219871520996,
+ "step": 721
+ },
+ {
+ "epoch": 10.027947598253276,
+ "grad_norm": 0.017398346215486526,
+ "learning_rate": 0.0006,
+ "loss": 5.66834831237793,
+ "step": 722
+ },
+ {
+ "epoch": 10.041921397379912,
+ "grad_norm": 0.016871167346835136,
+ "learning_rate": 0.0006,
+ "loss": 5.673916816711426,
+ "step": 723
+ },
+ {
+ "epoch": 10.055895196506551,
+ "grad_norm": 0.01695060543715954,
+ "learning_rate": 0.0006,
+ "loss": 5.646677017211914,
+ "step": 724
+ },
+ {
+ "epoch": 10.069868995633188,
+ "grad_norm": 0.019298404455184937,
+ "learning_rate": 0.0006,
+ "loss": 5.663064479827881,
+ "step": 725
+ },
+ {
+ "epoch": 10.083842794759825,
+ "grad_norm": 0.018699361011385918,
+ "learning_rate": 0.0006,
+ "loss": 5.615242958068848,
+ "step": 726
+ },
+ {
+ "epoch": 10.097816593886463,
+ "grad_norm": 0.019567854702472687,
+ "learning_rate": 0.0006,
+ "loss": 5.5521240234375,
+ "step": 727
+ },
+ {
+ "epoch": 10.1117903930131,
+ "grad_norm": 0.018347466364502907,
+ "learning_rate": 0.0006,
+ "loss": 5.5929107666015625,
+ "step": 728
+ },
+ {
+ "epoch": 10.125764192139737,
+ "grad_norm": 0.020350053906440735,
+ "learning_rate": 0.0006,
+ "loss": 5.6380109786987305,
+ "step": 729
+ },
+ {
+ "epoch": 10.139737991266376,
+ "grad_norm": 0.016790535300970078,
+ "learning_rate": 0.0006,
+ "loss": 5.62994384765625,
+ "step": 730
+ },
+ {
+ "epoch": 10.153711790393013,
+ "grad_norm": 0.01413649506866932,
+ "learning_rate": 0.0006,
+ "loss": 5.6301727294921875,
+ "step": 731
+ },
+ {
+ "epoch": 10.167685589519651,
+ "grad_norm": 0.013762339949607849,
+ "learning_rate": 0.0006,
+ "loss": 5.584138870239258,
+ "step": 732
+ },
+ {
+ "epoch": 10.181659388646288,
+ "grad_norm": 0.014790690504014492,
+ "learning_rate": 0.0006,
+ "loss": 5.666123390197754,
+ "step": 733
+ },
+ {
+ "epoch": 10.195633187772925,
+ "grad_norm": 0.015471262857317924,
+ "learning_rate": 0.0006,
+ "loss": 5.587042808532715,
+ "step": 734
+ },
+ {
+ "epoch": 10.209606986899564,
+ "grad_norm": 0.015804501250386238,
+ "learning_rate": 0.0006,
+ "loss": 5.627497673034668,
+ "step": 735
+ },
+ {
+ "epoch": 10.2235807860262,
+ "grad_norm": 0.017627960070967674,
+ "learning_rate": 0.0006,
+ "loss": 5.556890487670898,
+ "step": 736
+ },
+ {
+ "epoch": 10.237554585152838,
+ "grad_norm": 0.02074616774916649,
+ "learning_rate": 0.0006,
+ "loss": 5.586511135101318,
+ "step": 737
+ },
+ {
+ "epoch": 10.251528384279476,
+ "grad_norm": 0.026389935985207558,
+ "learning_rate": 0.0006,
+ "loss": 5.598520278930664,
+ "step": 738
+ },
+ {
+ "epoch": 10.265502183406113,
+ "grad_norm": 0.03736288473010063,
+ "learning_rate": 0.0006,
+ "loss": 5.575788497924805,
+ "step": 739
+ },
+ {
+ "epoch": 10.279475982532752,
+ "grad_norm": 0.03523678705096245,
+ "learning_rate": 0.0006,
+ "loss": 5.586930751800537,
+ "step": 740
+ },
+ {
+ "epoch": 10.293449781659389,
+ "grad_norm": 0.024266984313726425,
+ "learning_rate": 0.0006,
+ "loss": 5.596694469451904,
+ "step": 741
+ },
+ {
+ "epoch": 10.307423580786025,
+ "grad_norm": 0.033640433102846146,
+ "learning_rate": 0.0006,
+ "loss": 5.536456108093262,
+ "step": 742
+ },
+ {
+ "epoch": 10.321397379912664,
+ "grad_norm": 0.04689764603972435,
+ "learning_rate": 0.0006,
+ "loss": 5.61761999130249,
+ "step": 743
+ },
+ {
+ "epoch": 10.335371179039301,
+ "grad_norm": 0.04910685867071152,
+ "learning_rate": 0.0006,
+ "loss": 5.6298418045043945,
+ "step": 744
+ },
+ {
+ "epoch": 10.34934497816594,
+ "grad_norm": 0.03152550011873245,
+ "learning_rate": 0.0006,
+ "loss": 5.639697074890137,
+ "step": 745
+ },
+ {
+ "epoch": 10.363318777292577,
+ "grad_norm": 0.037753574550151825,
+ "learning_rate": 0.0006,
+ "loss": 5.697381019592285,
+ "step": 746
+ },
+ {
+ "epoch": 10.377292576419213,
+ "grad_norm": 0.04877670109272003,
+ "learning_rate": 0.0006,
+ "loss": 5.605184555053711,
+ "step": 747
+ },
+ {
+ "epoch": 10.391266375545852,
+ "grad_norm": 0.04228494316339493,
+ "learning_rate": 0.0006,
+ "loss": 5.624238967895508,
+ "step": 748
+ },
+ {
+ "epoch": 10.405240174672489,
+ "grad_norm": 0.03575790300965309,
+ "learning_rate": 0.0006,
+ "loss": 5.6037983894348145,
+ "step": 749
+ },
+ {
+ "epoch": 10.419213973799126,
+ "grad_norm": 0.036425549536943436,
+ "learning_rate": 0.0006,
+ "loss": 5.62905216217041,
+ "step": 750
+ },
+ {
+ "epoch": 10.433187772925764,
+ "grad_norm": 0.04001355543732643,
+ "learning_rate": 0.0006,
+ "loss": 5.581656455993652,
+ "step": 751
+ },
+ {
+ "epoch": 10.447161572052401,
+ "grad_norm": 0.05040939524769783,
+ "learning_rate": 0.0006,
+ "loss": 5.621358871459961,
+ "step": 752
+ },
+ {
+ "epoch": 10.46113537117904,
+ "grad_norm": 0.04140249267220497,
+ "learning_rate": 0.0006,
+ "loss": 5.668188095092773,
+ "step": 753
+ },
+ {
+ "epoch": 10.475109170305677,
+ "grad_norm": 0.033116403967142105,
+ "learning_rate": 0.0006,
+ "loss": 5.580477714538574,
+ "step": 754
+ },
+ {
+ "epoch": 10.489082969432314,
+ "grad_norm": 0.03864264488220215,
+ "learning_rate": 0.0006,
+ "loss": 5.539153099060059,
+ "step": 755
+ },
+ {
+ "epoch": 10.503056768558952,
+ "grad_norm": 0.03026588261127472,
+ "learning_rate": 0.0006,
+ "loss": 5.619814872741699,
+ "step": 756
+ },
+ {
+ "epoch": 10.51703056768559,
+ "grad_norm": 0.02381783165037632,
+ "learning_rate": 0.0006,
+ "loss": 5.64422607421875,
+ "step": 757
+ },
+ {
+ "epoch": 10.531004366812226,
+ "grad_norm": 0.02427830547094345,
+ "learning_rate": 0.0006,
+ "loss": 5.659873008728027,
+ "step": 758
+ },
+ {
+ "epoch": 10.544978165938865,
+ "grad_norm": 0.021709900349378586,
+ "learning_rate": 0.0006,
+ "loss": 5.590245723724365,
+ "step": 759
+ },
+ {
+ "epoch": 10.558951965065502,
+ "grad_norm": 0.022812914103269577,
+ "learning_rate": 0.0006,
+ "loss": 5.610434532165527,
+ "step": 760
+ },
+ {
+ "epoch": 10.57292576419214,
+ "grad_norm": 0.019250890240073204,
+ "learning_rate": 0.0006,
+ "loss": 5.588078498840332,
+ "step": 761
+ },
+ {
+ "epoch": 10.586899563318777,
+ "grad_norm": 0.021995501592755318,
+ "learning_rate": 0.0006,
+ "loss": 5.541754245758057,
+ "step": 762
+ },
+ {
+ "epoch": 10.600873362445414,
+ "grad_norm": 0.01477806642651558,
+ "learning_rate": 0.0006,
+ "loss": 5.549173355102539,
+ "step": 763
+ },
+ {
+ "epoch": 10.614847161572053,
+ "grad_norm": 0.016680821776390076,
+ "learning_rate": 0.0006,
+ "loss": 5.58221435546875,
+ "step": 764
+ },
+ {
+ "epoch": 10.62882096069869,
+ "grad_norm": 0.013848318718373775,
+ "learning_rate": 0.0006,
+ "loss": 5.5505876541137695,
+ "step": 765
+ },
+ {
+ "epoch": 10.642794759825328,
+ "grad_norm": 0.016387000679969788,
+ "learning_rate": 0.0006,
+ "loss": 5.570460319519043,
+ "step": 766
+ },
+ {
+ "epoch": 10.656768558951965,
+ "grad_norm": 0.016496233642101288,
+ "learning_rate": 0.0006,
+ "loss": 5.513058662414551,
+ "step": 767
+ },
+ {
+ "epoch": 10.670742358078602,
+ "grad_norm": 0.01675555482506752,
+ "learning_rate": 0.0006,
+ "loss": 5.559998989105225,
+ "step": 768
+ },
+ {
+ "epoch": 10.68471615720524,
+ "grad_norm": 0.016939565539360046,
+ "learning_rate": 0.0006,
+ "loss": 5.56727409362793,
+ "step": 769
+ },
+ {
+ "epoch": 10.698689956331878,
+ "grad_norm": 0.01679506152868271,
+ "learning_rate": 0.0006,
+ "loss": 5.603643417358398,
+ "step": 770
+ },
+ {
+ "epoch": 10.712663755458514,
+ "grad_norm": 0.01656349189579487,
+ "learning_rate": 0.0006,
+ "loss": 5.54292631149292,
+ "step": 771
+ },
+ {
+ "epoch": 10.726637554585153,
+ "grad_norm": 0.01975332945585251,
+ "learning_rate": 0.0006,
+ "loss": 5.5225677490234375,
+ "step": 772
+ },
+ {
+ "epoch": 10.74061135371179,
+ "grad_norm": 0.02329389750957489,
+ "learning_rate": 0.0006,
+ "loss": 5.537827491760254,
+ "step": 773
+ },
+ {
+ "epoch": 10.754585152838429,
+ "grad_norm": 0.026616891846060753,
+ "learning_rate": 0.0006,
+ "loss": 5.500011920928955,
+ "step": 774
+ },
+ {
+ "epoch": 10.768558951965066,
+ "grad_norm": 0.025511575862765312,
+ "learning_rate": 0.0006,
+ "loss": 5.5316619873046875,
+ "step": 775
+ },
+ {
+ "epoch": 10.782532751091702,
+ "grad_norm": 0.025009512901306152,
+ "learning_rate": 0.0006,
+ "loss": 5.468493938446045,
+ "step": 776
+ },
+ {
+ "epoch": 10.796506550218341,
+ "grad_norm": 0.02946239896118641,
+ "learning_rate": 0.0006,
+ "loss": 5.5116753578186035,
+ "step": 777
+ },
+ {
+ "epoch": 10.810480349344978,
+ "grad_norm": 0.036607250571250916,
+ "learning_rate": 0.0006,
+ "loss": 5.518636226654053,
+ "step": 778
+ },
+ {
+ "epoch": 10.824454148471617,
+ "grad_norm": 0.04347814992070198,
+ "learning_rate": 0.0006,
+ "loss": 5.572815895080566,
+ "step": 779
+ },
+ {
+ "epoch": 10.838427947598253,
+ "grad_norm": 0.034463658928871155,
+ "learning_rate": 0.0006,
+ "loss": 5.520634651184082,
+ "step": 780
+ },
+ {
+ "epoch": 10.85240174672489,
+ "grad_norm": 0.03486703336238861,
+ "learning_rate": 0.0006,
+ "loss": 5.545441150665283,
+ "step": 781
+ },
+ {
+ "epoch": 10.866375545851529,
+ "grad_norm": 0.04476739838719368,
+ "learning_rate": 0.0006,
+ "loss": 5.554450035095215,
+ "step": 782
+ },
+ {
+ "epoch": 10.880349344978166,
+ "grad_norm": 0.048745233565568924,
+ "learning_rate": 0.0006,
+ "loss": 5.554961204528809,
+ "step": 783
+ },
+ {
+ "epoch": 10.894323144104803,
+ "grad_norm": 0.04636877775192261,
+ "learning_rate": 0.0006,
+ "loss": 5.574655055999756,
+ "step": 784
+ },
+ {
+ "epoch": 10.908296943231441,
+ "grad_norm": 0.032686877995729446,
+ "learning_rate": 0.0006,
+ "loss": 5.5734639167785645,
+ "step": 785
+ },
+ {
+ "epoch": 10.922270742358078,
+ "grad_norm": 0.03075302764773369,
+ "learning_rate": 0.0006,
+ "loss": 5.487662315368652,
+ "step": 786
+ },
+ {
+ "epoch": 10.936244541484717,
+ "grad_norm": 0.028860175982117653,
+ "learning_rate": 0.0006,
+ "loss": 5.519798755645752,
+ "step": 787
+ },
+ {
+ "epoch": 10.950218340611354,
+ "grad_norm": 0.028599515557289124,
+ "learning_rate": 0.0006,
+ "loss": 5.558349609375,
+ "step": 788
+ },
+ {
+ "epoch": 10.96419213973799,
+ "grad_norm": 0.02384260669350624,
+ "learning_rate": 0.0006,
+ "loss": 5.520661354064941,
+ "step": 789
+ },
+ {
+ "epoch": 10.97816593886463,
+ "grad_norm": 0.023180771619081497,
+ "learning_rate": 0.0006,
+ "loss": 5.496896743774414,
+ "step": 790
+ },
+ {
+ "epoch": 10.992139737991266,
+ "grad_norm": 0.021167006343603134,
+ "learning_rate": 0.0006,
+ "loss": 5.517192363739014,
+ "step": 791
+ },
+ {
+ "epoch": 11.0,
+ "grad_norm": 0.02069013938307762,
+ "learning_rate": 0.0006,
+ "loss": 5.447656154632568,
+ "step": 792
+ },
+ {
+ "epoch": 11.0,
+ "eval_loss": 6.161121845245361,
+ "eval_runtime": 71.5972,
+ "eval_samples_per_second": 34.107,
+ "eval_steps_per_second": 1.075,
+ "step": 792
+ },
+ {
+ "epoch": 11.013973799126637,
+ "grad_norm": 0.022983292117714882,
+ "learning_rate": 0.0006,
+ "loss": 5.441346645355225,
+ "step": 793
+ },
+ {
+ "epoch": 11.027947598253276,
+ "grad_norm": 0.022833511233329773,
+ "learning_rate": 0.0006,
+ "loss": 5.45123815536499,
+ "step": 794
+ },
+ {
+ "epoch": 11.041921397379912,
+ "grad_norm": 0.020015383139252663,
+ "learning_rate": 0.0006,
+ "loss": 5.430389404296875,
+ "step": 795
+ },
+ {
+ "epoch": 11.055895196506551,
+ "grad_norm": 0.016957225278019905,
+ "learning_rate": 0.0006,
+ "loss": 5.473883628845215,
+ "step": 796
+ },
+ {
+ "epoch": 11.069868995633188,
+ "grad_norm": 0.02090892568230629,
+ "learning_rate": 0.0006,
+ "loss": 5.467441558837891,
+ "step": 797
+ },
+ {
+ "epoch": 11.083842794759825,
+ "grad_norm": 0.022806301712989807,
+ "learning_rate": 0.0006,
+ "loss": 5.5453081130981445,
+ "step": 798
+ },
+ {
+ "epoch": 11.097816593886463,
+ "grad_norm": 0.022816946730017662,
+ "learning_rate": 0.0006,
+ "loss": 5.426213264465332,
+ "step": 799
+ },
+ {
+ "epoch": 11.1117903930131,
+ "grad_norm": 0.020761270076036453,
+ "learning_rate": 0.0006,
+ "loss": 5.464138507843018,
+ "step": 800
+ },
+ {
+ "epoch": 11.125764192139737,
+ "grad_norm": 0.01982618309557438,
+ "learning_rate": 0.0006,
+ "loss": 5.428033828735352,
+ "step": 801
+ },
+ {
+ "epoch": 11.139737991266376,
+ "grad_norm": 0.021953724324703217,
+ "learning_rate": 0.0006,
+ "loss": 5.456646919250488,
+ "step": 802
+ },
+ {
+ "epoch": 11.153711790393013,
+ "grad_norm": 0.01871841959655285,
+ "learning_rate": 0.0006,
+ "loss": 5.508403778076172,
+ "step": 803
+ },
+ {
+ "epoch": 11.167685589519651,
+ "grad_norm": 0.01810353808104992,
+ "learning_rate": 0.0006,
+ "loss": 5.442957878112793,
+ "step": 804
+ },
+ {
+ "epoch": 11.181659388646288,
+ "grad_norm": 0.017609186470508575,
+ "learning_rate": 0.0006,
+ "loss": 5.4774699211120605,
+ "step": 805
+ },
+ {
+ "epoch": 11.195633187772925,
+ "grad_norm": 0.01931103691458702,
+ "learning_rate": 0.0006,
+ "loss": 5.429748058319092,
+ "step": 806
+ },
+ {
+ "epoch": 11.209606986899564,
+ "grad_norm": 0.019401391968131065,
+ "learning_rate": 0.0006,
+ "loss": 5.466834545135498,
+ "step": 807
+ },
+ {
+ "epoch": 11.2235807860262,
+ "grad_norm": 0.018862580880522728,
+ "learning_rate": 0.0006,
+ "loss": 5.431701183319092,
+ "step": 808
+ },
+ {
+ "epoch": 11.237554585152838,
+ "grad_norm": 0.01796964928507805,
+ "learning_rate": 0.0006,
+ "loss": 5.388613700866699,
+ "step": 809
+ },
+ {
+ "epoch": 11.251528384279476,
+ "grad_norm": 0.021166296675801277,
+ "learning_rate": 0.0006,
+ "loss": 5.454118251800537,
+ "step": 810
+ },
+ {
+ "epoch": 11.265502183406113,
+ "grad_norm": 0.028568729758262634,
+ "learning_rate": 0.0006,
+ "loss": 5.510260581970215,
+ "step": 811
+ },
+ {
+ "epoch": 11.279475982532752,
+ "grad_norm": 0.03902891278266907,
+ "learning_rate": 0.0006,
+ "loss": 5.386396408081055,
+ "step": 812
+ },
+ {
+ "epoch": 11.293449781659389,
+ "grad_norm": 0.039114370942115784,
+ "learning_rate": 0.0006,
+ "loss": 5.457526206970215,
+ "step": 813
+ },
+ {
+ "epoch": 11.307423580786025,
+ "grad_norm": 0.03560924902558327,
+ "learning_rate": 0.0006,
+ "loss": 5.4149980545043945,
+ "step": 814
+ },
+ {
+ "epoch": 11.321397379912664,
+ "grad_norm": 0.03745810315012932,
+ "learning_rate": 0.0006,
+ "loss": 5.464084625244141,
+ "step": 815
+ },
+ {
+ "epoch": 11.335371179039301,
+ "grad_norm": 0.03383171558380127,
+ "learning_rate": 0.0006,
+ "loss": 5.45137882232666,
+ "step": 816
+ },
+ {
+ "epoch": 11.34934497816594,
+ "grad_norm": 0.027873128652572632,
+ "learning_rate": 0.0006,
+ "loss": 5.485784530639648,
+ "step": 817
+ },
+ {
+ "epoch": 11.363318777292577,
+ "grad_norm": 0.02697703056037426,
+ "learning_rate": 0.0006,
+ "loss": 5.459152698516846,
+ "step": 818
+ },
+ {
+ "epoch": 11.377292576419213,
+ "grad_norm": 0.02646288089454174,
+ "learning_rate": 0.0006,
+ "loss": 5.433281421661377,
+ "step": 819
+ },
+ {
+ "epoch": 11.391266375545852,
+ "grad_norm": 0.02655179798603058,
+ "learning_rate": 0.0006,
+ "loss": 5.456627368927002,
+ "step": 820
+ },
+ {
+ "epoch": 11.405240174672489,
+ "grad_norm": 0.02633797563612461,
+ "learning_rate": 0.0006,
+ "loss": 5.484803676605225,
+ "step": 821
+ },
+ {
+ "epoch": 11.419213973799126,
+ "grad_norm": 0.031187085434794426,
+ "learning_rate": 0.0006,
+ "loss": 5.362293243408203,
+ "step": 822
+ },
+ {
+ "epoch": 11.433187772925764,
+ "grad_norm": 0.02971349097788334,
+ "learning_rate": 0.0006,
+ "loss": 5.4369611740112305,
+ "step": 823
+ },
+ {
+ "epoch": 11.447161572052401,
+ "grad_norm": 0.02996021695435047,
+ "learning_rate": 0.0006,
+ "loss": 5.419068813323975,
+ "step": 824
+ },
+ {
+ "epoch": 11.46113537117904,
+ "grad_norm": 0.032519739121198654,
+ "learning_rate": 0.0006,
+ "loss": 5.422189712524414,
+ "step": 825
+ },
+ {
+ "epoch": 11.475109170305677,
+ "grad_norm": 0.03403490409255028,
+ "learning_rate": 0.0006,
+ "loss": 5.470074653625488,
+ "step": 826
+ },
+ {
+ "epoch": 11.489082969432314,
+ "grad_norm": 0.036989886313676834,
+ "learning_rate": 0.0006,
+ "loss": 5.460174560546875,
+ "step": 827
+ },
+ {
+ "epoch": 11.503056768558952,
+ "grad_norm": 0.03427419438958168,
+ "learning_rate": 0.0006,
+ "loss": 5.495920658111572,
+ "step": 828
+ },
+ {
+ "epoch": 11.51703056768559,
+ "grad_norm": 0.02519620768725872,
+ "learning_rate": 0.0006,
+ "loss": 5.474429130554199,
+ "step": 829
+ },
+ {
+ "epoch": 11.531004366812226,
+ "grad_norm": 0.02473096363246441,
+ "learning_rate": 0.0006,
+ "loss": 5.395847320556641,
+ "step": 830
+ },
+ {
+ "epoch": 11.544978165938865,
+ "grad_norm": 0.02792714536190033,
+ "learning_rate": 0.0006,
+ "loss": 5.376262664794922,
+ "step": 831
+ },
+ {
+ "epoch": 11.558951965065502,
+ "grad_norm": 0.024172596633434296,
+ "learning_rate": 0.0006,
+ "loss": 5.459731101989746,
+ "step": 832
+ },
+ {
+ "epoch": 11.57292576419214,
+ "grad_norm": 0.02304821088910103,
+ "learning_rate": 0.0006,
+ "loss": 5.479475498199463,
+ "step": 833
+ },
+ {
+ "epoch": 11.586899563318777,
+ "grad_norm": 0.02342340536415577,
+ "learning_rate": 0.0006,
+ "loss": 5.387955665588379,
+ "step": 834
+ },
+ {
+ "epoch": 11.600873362445414,
+ "grad_norm": 0.021962853148579597,
+ "learning_rate": 0.0006,
+ "loss": 5.385697364807129,
+ "step": 835
+ },
+ {
+ "epoch": 11.614847161572053,
+ "grad_norm": 0.019193042069673538,
+ "learning_rate": 0.0006,
+ "loss": 5.431601047515869,
+ "step": 836
+ },
+ {
+ "epoch": 11.62882096069869,
+ "grad_norm": 0.021378040313720703,
+ "learning_rate": 0.0006,
+ "loss": 5.3070969581604,
+ "step": 837
+ },
+ {
+ "epoch": 11.642794759825328,
+ "grad_norm": 0.021641355007886887,
+ "learning_rate": 0.0006,
+ "loss": 5.360584735870361,
+ "step": 838
+ },
+ {
+ "epoch": 11.656768558951965,
+ "grad_norm": 0.019899170845746994,
+ "learning_rate": 0.0006,
+ "loss": 5.398155212402344,
+ "step": 839
+ },
+ {
+ "epoch": 11.670742358078602,
+ "grad_norm": 0.02298869751393795,
+ "learning_rate": 0.0006,
+ "loss": 5.410561561584473,
+ "step": 840
+ },
+ {
+ "epoch": 11.68471615720524,
+ "grad_norm": 0.02815551869571209,
+ "learning_rate": 0.0006,
+ "loss": 5.298604965209961,
+ "step": 841
+ },
+ {
+ "epoch": 11.698689956331878,
+ "grad_norm": 0.03041219152510166,
+ "learning_rate": 0.0006,
+ "loss": 5.425869941711426,
+ "step": 842
+ },
+ {
+ "epoch": 11.712663755458514,
+ "grad_norm": 0.025710400193929672,
+ "learning_rate": 0.0006,
+ "loss": 5.344316482543945,
+ "step": 843
+ },
+ {
+ "epoch": 11.726637554585153,
+ "grad_norm": 0.018438996747136116,
+ "learning_rate": 0.0006,
+ "loss": 5.379913330078125,
+ "step": 844
+ },
+ {
+ "epoch": 11.74061135371179,
+ "grad_norm": 0.016683703288435936,
+ "learning_rate": 0.0006,
+ "loss": 5.338386058807373,
+ "step": 845
+ },
+ {
+ "epoch": 11.754585152838429,
+ "grad_norm": 0.017252853140234947,
+ "learning_rate": 0.0006,
+ "loss": 5.360256195068359,
+ "step": 846
+ },
+ {
+ "epoch": 11.768558951965066,
+ "grad_norm": 0.019606001675128937,
+ "learning_rate": 0.0006,
+ "loss": 5.401415824890137,
+ "step": 847
+ },
+ {
+ "epoch": 11.782532751091702,
+ "grad_norm": 0.016299517825245857,
+ "learning_rate": 0.0006,
+ "loss": 5.39270544052124,
+ "step": 848
+ },
+ {
+ "epoch": 11.796506550218341,
+ "grad_norm": 0.016657982021570206,
+ "learning_rate": 0.0006,
+ "loss": 5.359323978424072,
+ "step": 849
+ },
+ {
+ "epoch": 11.810480349344978,
+ "grad_norm": 0.017462270334362984,
+ "learning_rate": 0.0006,
+ "loss": 5.453655242919922,
+ "step": 850
+ },
+ {
+ "epoch": 11.824454148471617,
+ "grad_norm": 0.01723778434097767,
+ "learning_rate": 0.0006,
+ "loss": 5.454940319061279,
+ "step": 851
+ },
+ {
+ "epoch": 11.838427947598253,
+ "grad_norm": 0.019071755930781364,
+ "learning_rate": 0.0006,
+ "loss": 5.355914115905762,
+ "step": 852
+ },
+ {
+ "epoch": 11.85240174672489,
+ "grad_norm": 0.02169782482087612,
+ "learning_rate": 0.0006,
+ "loss": 5.396806716918945,
+ "step": 853
+ },
+ {
+ "epoch": 11.866375545851529,
+ "grad_norm": 0.025049131363630295,
+ "learning_rate": 0.0006,
+ "loss": 5.347422122955322,
+ "step": 854
+ },
+ {
+ "epoch": 11.880349344978166,
+ "grad_norm": 0.024977540597319603,
+ "learning_rate": 0.0006,
+ "loss": 5.32526969909668,
+ "step": 855
+ },
+ {
+ "epoch": 11.894323144104803,
+ "grad_norm": 0.024223418906331062,
+ "learning_rate": 0.0006,
+ "loss": 5.372445106506348,
+ "step": 856
+ },
+ {
+ "epoch": 11.908296943231441,
+ "grad_norm": 0.02519913762807846,
+ "learning_rate": 0.0006,
+ "loss": 5.280451774597168,
+ "step": 857
+ },
+ {
+ "epoch": 11.922270742358078,
+ "grad_norm": 0.0255630724132061,
+ "learning_rate": 0.0006,
+ "loss": 5.462911605834961,
+ "step": 858
+ },
+ {
+ "epoch": 11.936244541484717,
+ "grad_norm": 0.022740041837096214,
+ "learning_rate": 0.0006,
+ "loss": 5.319089412689209,
+ "step": 859
+ },
+ {
+ "epoch": 11.950218340611354,
+ "grad_norm": 0.02038649097084999,
+ "learning_rate": 0.0006,
+ "loss": 5.37426233291626,
+ "step": 860
+ },
+ {
+ "epoch": 11.96419213973799,
+ "grad_norm": 0.020509786903858185,
+ "learning_rate": 0.0006,
+ "loss": 5.337924957275391,
+ "step": 861
+ },
+ {
+ "epoch": 11.97816593886463,
+ "grad_norm": 0.021707357838749886,
+ "learning_rate": 0.0006,
+ "loss": 5.374853134155273,
+ "step": 862
+ },
+ {
+ "epoch": 11.992139737991266,
+ "grad_norm": 0.02714136242866516,
+ "learning_rate": 0.0006,
+ "loss": 5.2982282638549805,
+ "step": 863
+ },
+ {
+ "epoch": 12.0,
+ "grad_norm": 0.027913225814700127,
+ "learning_rate": 0.0006,
+ "loss": 5.406044960021973,
+ "step": 864
+ },
+ {
+ "epoch": 12.0,
+ "eval_loss": 5.948610305786133,
+ "eval_runtime": 75.79,
+ "eval_samples_per_second": 32.221,
+ "eval_steps_per_second": 1.016,
+ "step": 864
+ },
+ {
+ "epoch": 12.013973799126637,
+ "grad_norm": 0.023761706426739693,
+ "learning_rate": 0.0006,
+ "loss": 5.378547191619873,
+ "step": 865
+ },
+ {
+ "epoch": 12.027947598253276,
+ "grad_norm": 0.026866815984249115,
+ "learning_rate": 0.0006,
+ "loss": 5.356947898864746,
+ "step": 866
+ },
+ {
+ "epoch": 12.041921397379912,
+ "grad_norm": 0.03335791453719139,
+ "learning_rate": 0.0006,
+ "loss": 5.264157772064209,
+ "step": 867
+ },
+ {
+ "epoch": 12.055895196506551,
+ "grad_norm": 0.03153866529464722,
+ "learning_rate": 0.0006,
+ "loss": 5.363075256347656,
+ "step": 868
+ },
+ {
+ "epoch": 12.069868995633188,
+ "grad_norm": 0.03270367905497551,
+ "learning_rate": 0.0006,
+ "loss": 5.298289775848389,
+ "step": 869
+ },
+ {
+ "epoch": 12.083842794759825,
+ "grad_norm": 0.030872846022248268,
+ "learning_rate": 0.0006,
+ "loss": 5.324321269989014,
+ "step": 870
+ },
+ {
+ "epoch": 12.097816593886463,
+ "grad_norm": 0.028438350185751915,
+ "learning_rate": 0.0006,
+ "loss": 5.3512115478515625,
+ "step": 871
+ },
+ {
+ "epoch": 12.1117903930131,
+ "grad_norm": 0.030846193432807922,
+ "learning_rate": 0.0006,
+ "loss": 5.304851531982422,
+ "step": 872
+ },
+ {
+ "epoch": 12.125764192139737,
+ "grad_norm": 0.03879433870315552,
+ "learning_rate": 0.0006,
+ "loss": 5.289634704589844,
+ "step": 873
+ },
+ {
+ "epoch": 12.139737991266376,
+ "grad_norm": 0.03665967285633087,
+ "learning_rate": 0.0006,
+ "loss": 5.352825164794922,
+ "step": 874
+ },
+ {
+ "epoch": 12.153711790393013,
+ "grad_norm": 0.03454619646072388,
+ "learning_rate": 0.0006,
+ "loss": 5.346499443054199,
+ "step": 875
+ },
+ {
+ "epoch": 12.167685589519651,
+ "grad_norm": 0.043245580047369,
+ "learning_rate": 0.0006,
+ "loss": 5.344263076782227,
+ "step": 876
+ },
+ {
+ "epoch": 12.181659388646288,
+ "grad_norm": 0.04450082778930664,
+ "learning_rate": 0.0006,
+ "loss": 5.447438716888428,
+ "step": 877
+ },
+ {
+ "epoch": 12.195633187772925,
+ "grad_norm": 0.04980820417404175,
+ "learning_rate": 0.0006,
+ "loss": 5.364802360534668,
+ "step": 878
+ },
+ {
+ "epoch": 12.209606986899564,
+ "grad_norm": 0.04757903143763542,
+ "learning_rate": 0.0006,
+ "loss": 5.363414287567139,
+ "step": 879
+ },
+ {
+ "epoch": 12.2235807860262,
+ "grad_norm": 0.048647016286849976,
+ "learning_rate": 0.0006,
+ "loss": 5.379604339599609,
+ "step": 880
+ },
+ {
+ "epoch": 12.237554585152838,
+ "grad_norm": 0.04497060924768448,
+ "learning_rate": 0.0006,
+ "loss": 5.354029178619385,
+ "step": 881
+ },
+ {
+ "epoch": 12.251528384279476,
+ "grad_norm": 0.032642051577568054,
+ "learning_rate": 0.0006,
+ "loss": 5.335702419281006,
+ "step": 882
+ },
+ {
+ "epoch": 12.265502183406113,
+ "grad_norm": 0.037766437977552414,
+ "learning_rate": 0.0006,
+ "loss": 5.331804275512695,
+ "step": 883
+ },
+ {
+ "epoch": 12.279475982532752,
+ "grad_norm": 0.030145324766635895,
+ "learning_rate": 0.0006,
+ "loss": 5.327176094055176,
+ "step": 884
+ },
+ {
+ "epoch": 12.293449781659389,
+ "grad_norm": 0.032248858362436295,
+ "learning_rate": 0.0006,
+ "loss": 5.2410712242126465,
+ "step": 885
+ },
+ {
+ "epoch": 12.307423580786025,
+ "grad_norm": 0.026990503072738647,
+ "learning_rate": 0.0006,
+ "loss": 5.369958877563477,
+ "step": 886
+ },
+ {
+ "epoch": 12.321397379912664,
+ "grad_norm": 0.023413540795445442,
+ "learning_rate": 0.0006,
+ "loss": 5.322418212890625,
+ "step": 887
+ },
+ {
+ "epoch": 12.335371179039301,
+ "grad_norm": 0.025248292833566666,
+ "learning_rate": 0.0006,
+ "loss": 5.300156116485596,
+ "step": 888
+ },
+ {
+ "epoch": 12.34934497816594,
+ "grad_norm": 0.021037699654698372,
+ "learning_rate": 0.0006,
+ "loss": 5.26978063583374,
+ "step": 889
+ },
+ {
+ "epoch": 12.363318777292577,
+ "grad_norm": 0.0155200669541955,
+ "learning_rate": 0.0006,
+ "loss": 5.297061920166016,
+ "step": 890
+ },
+ {
+ "epoch": 12.377292576419213,
+ "grad_norm": 0.017867842689156532,
+ "learning_rate": 0.0006,
+ "loss": 5.312634468078613,
+ "step": 891
+ },
+ {
+ "epoch": 12.391266375545852,
+ "grad_norm": 0.01675534062087536,
+ "learning_rate": 0.0006,
+ "loss": 5.290635108947754,
+ "step": 892
+ },
+ {
+ "epoch": 12.405240174672489,
+ "grad_norm": 0.015546303242444992,
+ "learning_rate": 0.0006,
+ "loss": 5.289699554443359,
+ "step": 893
+ },
+ {
+ "epoch": 12.419213973799126,
+ "grad_norm": 0.01612856611609459,
+ "learning_rate": 0.0006,
+ "loss": 5.312010765075684,
+ "step": 894
+ },
+ {
+ "epoch": 12.433187772925764,
+ "grad_norm": 0.015225349925458431,
+ "learning_rate": 0.0006,
+ "loss": 5.255838871002197,
+ "step": 895
+ },
+ {
+ "epoch": 12.447161572052401,
+ "grad_norm": 0.015317161567509174,
+ "learning_rate": 0.0006,
+ "loss": 5.234170436859131,
+ "step": 896
+ },
+ {
+ "epoch": 12.46113537117904,
+ "grad_norm": 0.014820792712271214,
+ "learning_rate": 0.0006,
+ "loss": 5.255383014678955,
+ "step": 897
+ },
+ {
+ "epoch": 12.475109170305677,
+ "grad_norm": 0.01343776099383831,
+ "learning_rate": 0.0006,
+ "loss": 5.256904602050781,
+ "step": 898
+ },
+ {
+ "epoch": 12.489082969432314,
+ "grad_norm": 0.01418706402182579,
+ "learning_rate": 0.0006,
+ "loss": 5.272562026977539,
+ "step": 899
+ },
+ {
+ "epoch": 12.503056768558952,
+ "grad_norm": 0.013406263664364815,
+ "learning_rate": 0.0006,
+ "loss": 5.17042350769043,
+ "step": 900
+ },
+ {
+ "epoch": 12.51703056768559,
+ "grad_norm": 0.01331746019423008,
+ "learning_rate": 0.0006,
+ "loss": 5.267889022827148,
+ "step": 901
+ },
+ {
+ "epoch": 12.531004366812226,
+ "grad_norm": 0.014625098556280136,
+ "learning_rate": 0.0006,
+ "loss": 5.281431198120117,
+ "step": 902
+ },
+ {
+ "epoch": 12.544978165938865,
+ "grad_norm": 0.01346531230956316,
+ "learning_rate": 0.0006,
+ "loss": 5.286074638366699,
+ "step": 903
+ },
+ {
+ "epoch": 12.558951965065502,
+ "grad_norm": 0.01303934771567583,
+ "learning_rate": 0.0006,
+ "loss": 5.344804763793945,
+ "step": 904
+ },
+ {
+ "epoch": 12.57292576419214,
+ "grad_norm": 0.014180841855704784,
+ "learning_rate": 0.0006,
+ "loss": 5.252955436706543,
+ "step": 905
+ },
+ {
+ "epoch": 12.586899563318777,
+ "grad_norm": 0.014279021881520748,
+ "learning_rate": 0.0006,
+ "loss": 5.283501625061035,
+ "step": 906
+ },
+ {
+ "epoch": 12.600873362445414,
+ "grad_norm": 0.013922203332185745,
+ "learning_rate": 0.0006,
+ "loss": 5.225350379943848,
+ "step": 907
+ },
+ {
+ "epoch": 12.614847161572053,
+ "grad_norm": 0.014055478386580944,
+ "learning_rate": 0.0006,
+ "loss": 5.283530235290527,
+ "step": 908
+ },
+ {
+ "epoch": 12.62882096069869,
+ "grad_norm": 0.013623543083667755,
+ "learning_rate": 0.0006,
+ "loss": 5.204877853393555,
+ "step": 909
+ },
+ {
+ "epoch": 12.642794759825328,
+ "grad_norm": 0.014378432184457779,
+ "learning_rate": 0.0006,
+ "loss": 5.275735378265381,
+ "step": 910
+ },
+ {
+ "epoch": 12.656768558951965,
+ "grad_norm": 0.017884068191051483,
+ "learning_rate": 0.0006,
+ "loss": 5.256485939025879,
+ "step": 911
+ },
+ {
+ "epoch": 12.670742358078602,
+ "grad_norm": 0.0203024260699749,
+ "learning_rate": 0.0006,
+ "loss": 5.292258262634277,
+ "step": 912
+ },
+ {
+ "epoch": 12.68471615720524,
+ "grad_norm": 0.021196166053414345,
+ "learning_rate": 0.0006,
+ "loss": 5.30052375793457,
+ "step": 913
+ },
+ {
+ "epoch": 12.698689956331878,
+ "grad_norm": 0.02249213121831417,
+ "learning_rate": 0.0006,
+ "loss": 5.236924648284912,
+ "step": 914
+ },
+ {
+ "epoch": 12.712663755458514,
+ "grad_norm": 0.019980570301413536,
+ "learning_rate": 0.0006,
+ "loss": 5.290444374084473,
+ "step": 915
+ },
+ {
+ "epoch": 12.726637554585153,
+ "grad_norm": 0.022030364722013474,
+ "learning_rate": 0.0006,
+ "loss": 5.287881851196289,
+ "step": 916
+ },
+ {
+ "epoch": 12.74061135371179,
+ "grad_norm": 0.02438683621585369,
+ "learning_rate": 0.0006,
+ "loss": 5.230462074279785,
+ "step": 917
+ },
+ {
+ "epoch": 12.754585152838429,
+ "grad_norm": 0.026895932853221893,
+ "learning_rate": 0.0006,
+ "loss": 5.232667446136475,
+ "step": 918
+ },
+ {
+ "epoch": 12.768558951965066,
+ "grad_norm": 0.02861972153186798,
+ "learning_rate": 0.0006,
+ "loss": 5.205913066864014,
+ "step": 919
+ },
+ {
+ "epoch": 12.782532751091702,
+ "grad_norm": 0.033064406365156174,
+ "learning_rate": 0.0006,
+ "loss": 5.242166042327881,
+ "step": 920
+ },
+ {
+ "epoch": 12.796506550218341,
+ "grad_norm": 0.030850905925035477,
+ "learning_rate": 0.0006,
+ "loss": 5.2503581047058105,
+ "step": 921
+ },
+ {
+ "epoch": 12.810480349344978,
+ "grad_norm": 0.03139309585094452,
+ "learning_rate": 0.0006,
+ "loss": 5.222465515136719,
+ "step": 922
+ },
+ {
+ "epoch": 12.824454148471617,
+ "grad_norm": 0.035760655999183655,
+ "learning_rate": 0.0006,
+ "loss": 5.279786586761475,
+ "step": 923
+ },
+ {
+ "epoch": 12.838427947598253,
+ "grad_norm": 0.036571063101291656,
+ "learning_rate": 0.0006,
+ "loss": 5.237204551696777,
+ "step": 924
+ },
+ {
+ "epoch": 12.85240174672489,
+ "grad_norm": 0.03467971459031105,
+ "learning_rate": 0.0006,
+ "loss": 5.210154056549072,
+ "step": 925
+ },
+ {
+ "epoch": 12.866375545851529,
+ "grad_norm": 0.053154416382312775,
+ "learning_rate": 0.0006,
+ "loss": 5.295984268188477,
+ "step": 926
+ },
+ {
+ "epoch": 12.880349344978166,
+ "grad_norm": 0.05374281480908394,
+ "learning_rate": 0.0006,
+ "loss": 5.293900012969971,
+ "step": 927
+ },
+ {
+ "epoch": 12.894323144104803,
+ "grad_norm": 0.05851082503795624,
+ "learning_rate": 0.0006,
+ "loss": 5.3758931159973145,
+ "step": 928
+ },
+ {
+ "epoch": 12.908296943231441,
+ "grad_norm": 0.05145686864852905,
+ "learning_rate": 0.0006,
+ "loss": 5.309441566467285,
+ "step": 929
+ },
+ {
+ "epoch": 12.922270742358078,
+ "grad_norm": 0.04484057426452637,
+ "learning_rate": 0.0006,
+ "loss": 5.297204494476318,
+ "step": 930
+ },
+ {
+ "epoch": 12.936244541484717,
+ "grad_norm": 0.05348053574562073,
+ "learning_rate": 0.0006,
+ "loss": 5.278589248657227,
+ "step": 931
+ },
+ {
+ "epoch": 12.950218340611354,
+ "grad_norm": 0.05677935481071472,
+ "learning_rate": 0.0006,
+ "loss": 5.290632247924805,
+ "step": 932
+ },
+ {
+ "epoch": 12.96419213973799,
+ "grad_norm": 0.05244941636919975,
+ "learning_rate": 0.0006,
+ "loss": 5.353776454925537,
+ "step": 933
+ },
+ {
+ "epoch": 12.97816593886463,
+ "grad_norm": 0.03900328651070595,
+ "learning_rate": 0.0006,
+ "loss": 5.284013748168945,
+ "step": 934
+ },
+ {
+ "epoch": 12.992139737991266,
+ "grad_norm": 0.03373100981116295,
+ "learning_rate": 0.0006,
+ "loss": 5.3313751220703125,
+ "step": 935
+ },
+ {
+ "epoch": 13.0,
+ "grad_norm": 0.03497432917356491,
+ "learning_rate": 0.0006,
+ "loss": 5.395255088806152,
+ "step": 936
+ },
+ {
+ "epoch": 13.0,
+ "eval_loss": 5.799665451049805,
+ "eval_runtime": 81.3525,
+ "eval_samples_per_second": 30.018,
+ "eval_steps_per_second": 0.946,
+ "step": 936
+ },
+ {
+ "epoch": 13.013973799126637,
+ "grad_norm": 0.039509695023298264,
+ "learning_rate": 0.0006,
+ "loss": 5.298876762390137,
+ "step": 937
+ },
+ {
+ "epoch": 13.027947598253276,
+ "grad_norm": 0.034599535167217255,
+ "learning_rate": 0.0006,
+ "loss": 5.264098167419434,
+ "step": 938
+ },
+ {
+ "epoch": 13.041921397379912,
+ "grad_norm": 0.029622144997119904,
+ "learning_rate": 0.0006,
+ "loss": 5.21397066116333,
+ "step": 939
+ },
+ {
+ "epoch": 13.055895196506551,
+ "grad_norm": 0.02309993840754032,
+ "learning_rate": 0.0006,
+ "loss": 5.227992057800293,
+ "step": 940
+ },
+ {
+ "epoch": 13.069868995633188,
+ "grad_norm": 0.024507572874426842,
+ "learning_rate": 0.0006,
+ "loss": 5.192324161529541,
+ "step": 941
+ },
+ {
+ "epoch": 13.083842794759825,
+ "grad_norm": 0.018214358016848564,
+ "learning_rate": 0.0006,
+ "loss": 5.23871374130249,
+ "step": 942
+ },
+ {
+ "epoch": 13.097816593886463,
+ "grad_norm": 0.01937488280236721,
+ "learning_rate": 0.0006,
+ "loss": 5.227265357971191,
+ "step": 943
+ },
+ {
+ "epoch": 13.1117903930131,
+ "grad_norm": 0.017225295305252075,
+ "learning_rate": 0.0006,
+ "loss": 5.232108116149902,
+ "step": 944
+ },
+ {
+ "epoch": 13.125764192139737,
+ "grad_norm": 0.016314376145601273,
+ "learning_rate": 0.0006,
+ "loss": 5.2288360595703125,
+ "step": 945
+ },
+ {
+ "epoch": 13.139737991266376,
+ "grad_norm": 0.014701297506690025,
+ "learning_rate": 0.0006,
+ "loss": 5.297238349914551,
+ "step": 946
+ },
+ {
+ "epoch": 13.153711790393013,
+ "grad_norm": 0.015610086731612682,
+ "learning_rate": 0.0006,
+ "loss": 5.148407936096191,
+ "step": 947
+ },
+ {
+ "epoch": 13.167685589519651,
+ "grad_norm": 0.014056755229830742,
+ "learning_rate": 0.0006,
+ "loss": 5.208695411682129,
+ "step": 948
+ },
+ {
+ "epoch": 13.181659388646288,
+ "grad_norm": 0.012802829965949059,
+ "learning_rate": 0.0006,
+ "loss": 5.229208946228027,
+ "step": 949
+ },
+ {
+ "epoch": 13.195633187772925,
+ "grad_norm": 0.011958962306380272,
+ "learning_rate": 0.0006,
+ "loss": 5.201716423034668,
+ "step": 950
+ },
+ {
+ "epoch": 13.209606986899564,
+ "grad_norm": 0.012841588817536831,
+ "learning_rate": 0.0006,
+ "loss": 5.240131378173828,
+ "step": 951
+ },
+ {
+ "epoch": 13.2235807860262,
+ "grad_norm": 0.014653137885034084,
+ "learning_rate": 0.0006,
+ "loss": 5.146340370178223,
+ "step": 952
+ },
+ {
+ "epoch": 13.237554585152838,
+ "grad_norm": 0.010446185246109962,
+ "learning_rate": 0.0006,
+ "loss": 5.21809720993042,
+ "step": 953
+ },
+ {
+ "epoch": 13.251528384279476,
+ "grad_norm": 0.01092249620705843,
+ "learning_rate": 0.0006,
+ "loss": 5.135411739349365,
+ "step": 954
+ },
+ {
+ "epoch": 13.265502183406113,
+ "grad_norm": 0.010561129078269005,
+ "learning_rate": 0.0006,
+ "loss": 5.194492340087891,
+ "step": 955
+ },
+ {
+ "epoch": 13.279475982532752,
+ "grad_norm": 0.010118645615875721,
+ "learning_rate": 0.0006,
+ "loss": 5.235012531280518,
+ "step": 956
+ },
+ {
+ "epoch": 13.293449781659389,
+ "grad_norm": 0.01081101130694151,
+ "learning_rate": 0.0006,
+ "loss": 5.162468910217285,
+ "step": 957
+ },
+ {
+ "epoch": 13.307423580786025,
+ "grad_norm": 0.010701699182391167,
+ "learning_rate": 0.0006,
+ "loss": 5.217264175415039,
+ "step": 958
+ },
+ {
+ "epoch": 13.321397379912664,
+ "grad_norm": 0.010495852679014206,
+ "learning_rate": 0.0006,
+ "loss": 5.202086925506592,
+ "step": 959
+ },
+ {
+ "epoch": 13.335371179039301,
+ "grad_norm": 0.009838009253144264,
+ "learning_rate": 0.0006,
+ "loss": 5.201089859008789,
+ "step": 960
+ },
+ {
+ "epoch": 13.34934497816594,
+ "grad_norm": 0.011280231177806854,
+ "learning_rate": 0.0006,
+ "loss": 5.1775336265563965,
+ "step": 961
+ },
+ {
+ "epoch": 13.363318777292577,
+ "grad_norm": 0.011244936846196651,
+ "learning_rate": 0.0006,
+ "loss": 5.185998439788818,
+ "step": 962
+ },
+ {
+ "epoch": 13.377292576419213,
+ "grad_norm": 0.01153804361820221,
+ "learning_rate": 0.0006,
+ "loss": 5.132739067077637,
+ "step": 963
+ },
+ {
+ "epoch": 13.391266375545852,
+ "grad_norm": 0.010817309841513634,
+ "learning_rate": 0.0006,
+ "loss": 5.268172264099121,
+ "step": 964
+ },
+ {
+ "epoch": 13.405240174672489,
+ "grad_norm": 0.011870342306792736,
+ "learning_rate": 0.0006,
+ "loss": 5.154515266418457,
+ "step": 965
+ },
+ {
+ "epoch": 13.419213973799126,
+ "grad_norm": 0.011373576708137989,
+ "learning_rate": 0.0006,
+ "loss": 5.160059928894043,
+ "step": 966
+ },
+ {
+ "epoch": 13.433187772925764,
+ "grad_norm": 0.01116643100976944,
+ "learning_rate": 0.0006,
+ "loss": 5.079656600952148,
+ "step": 967
+ },
+ {
+ "epoch": 13.447161572052401,
+ "grad_norm": 0.01202121190726757,
+ "learning_rate": 0.0006,
+ "loss": 5.205666542053223,
+ "step": 968
+ },
+ {
+ "epoch": 13.46113537117904,
+ "grad_norm": 0.013683444820344448,
+ "learning_rate": 0.0006,
+ "loss": 5.168977737426758,
+ "step": 969
+ },
+ {
+ "epoch": 13.475109170305677,
+ "grad_norm": 0.012892362661659718,
+ "learning_rate": 0.0006,
+ "loss": 5.168679237365723,
+ "step": 970
+ },
+ {
+ "epoch": 13.489082969432314,
+ "grad_norm": 0.01123902853578329,
+ "learning_rate": 0.0006,
+ "loss": 5.143795013427734,
+ "step": 971
+ },
+ {
+ "epoch": 13.503056768558952,
+ "grad_norm": 0.010333497077226639,
+ "learning_rate": 0.0006,
+ "loss": 5.170592784881592,
+ "step": 972
+ },
+ {
+ "epoch": 13.51703056768559,
+ "grad_norm": 0.011278851889073849,
+ "learning_rate": 0.0006,
+ "loss": 5.162395477294922,
+ "step": 973
+ },
+ {
+ "epoch": 13.531004366812226,
+ "grad_norm": 0.013309221714735031,
+ "learning_rate": 0.0006,
+ "loss": 5.0712103843688965,
+ "step": 974
+ },
+ {
+ "epoch": 13.544978165938865,
+ "grad_norm": 0.014573273248970509,
+ "learning_rate": 0.0006,
+ "loss": 5.098300933837891,
+ "step": 975
+ },
+ {
+ "epoch": 13.558951965065502,
+ "grad_norm": 0.01755373552441597,
+ "learning_rate": 0.0006,
+ "loss": 5.081066131591797,
+ "step": 976
+ },
+ {
+ "epoch": 13.57292576419214,
+ "grad_norm": 0.02018735185265541,
+ "learning_rate": 0.0006,
+ "loss": 5.173007965087891,
+ "step": 977
+ },
+ {
+ "epoch": 13.586899563318777,
+ "grad_norm": 0.023258069530129433,
+ "learning_rate": 0.0006,
+ "loss": 5.107330322265625,
+ "step": 978
+ },
+ {
+ "epoch": 13.600873362445414,
+ "grad_norm": 0.0250930804759264,
+ "learning_rate": 0.0006,
+ "loss": 5.193347930908203,
+ "step": 979
+ },
+ {
+ "epoch": 13.614847161572053,
+ "grad_norm": 0.02333216182887554,
+ "learning_rate": 0.0006,
+ "loss": 5.160326957702637,
+ "step": 980
+ },
+ {
+ "epoch": 13.62882096069869,
+ "grad_norm": 0.019765766337513924,
+ "learning_rate": 0.0006,
+ "loss": 5.05100679397583,
+ "step": 981
+ },
+ {
+ "epoch": 13.642794759825328,
+ "grad_norm": 0.020395414903759956,
+ "learning_rate": 0.0006,
+ "loss": 5.157297611236572,
+ "step": 982
+ },
+ {
+ "epoch": 13.656768558951965,
+ "grad_norm": 0.019553566351532936,
+ "learning_rate": 0.0006,
+ "loss": 5.158632755279541,
+ "step": 983
+ },
+ {
+ "epoch": 13.670742358078602,
+ "grad_norm": 0.018838031217455864,
+ "learning_rate": 0.0006,
+ "loss": 5.167074203491211,
+ "step": 984
+ },
+ {
+ "epoch": 13.68471615720524,
+ "grad_norm": 0.020884843543171883,
+ "learning_rate": 0.0006,
+ "loss": 5.238006114959717,
+ "step": 985
+ },
+ {
+ "epoch": 13.698689956331878,
+ "grad_norm": 0.021462196484208107,
+ "learning_rate": 0.0006,
+ "loss": 5.1703901290893555,
+ "step": 986
+ },
+ {
+ "epoch": 13.712663755458514,
+ "grad_norm": 0.021178999915719032,
+ "learning_rate": 0.0006,
+ "loss": 5.109465599060059,
+ "step": 987
+ },
+ {
+ "epoch": 13.726637554585153,
+ "grad_norm": 0.022374693304300308,
+ "learning_rate": 0.0006,
+ "loss": 5.195014953613281,
+ "step": 988
+ },
+ {
+ "epoch": 13.74061135371179,
+ "grad_norm": 0.021914934739470482,
+ "learning_rate": 0.0006,
+ "loss": 5.15154504776001,
+ "step": 989
+ },
+ {
+ "epoch": 13.754585152838429,
+ "grad_norm": 0.017769858241081238,
+ "learning_rate": 0.0006,
+ "loss": 5.119076251983643,
+ "step": 990
+ },
+ {
+ "epoch": 13.768558951965066,
+ "grad_norm": 0.02018590457737446,
+ "learning_rate": 0.0006,
+ "loss": 5.087010860443115,
+ "step": 991
+ },
+ {
+ "epoch": 13.782532751091702,
+ "grad_norm": 0.023143045604228973,
+ "learning_rate": 0.0006,
+ "loss": 5.087374687194824,
+ "step": 992
+ },
+ {
+ "epoch": 13.796506550218341,
+ "grad_norm": 0.02214638516306877,
+ "learning_rate": 0.0006,
+ "loss": 5.043274879455566,
+ "step": 993
+ },
+ {
+ "epoch": 13.810480349344978,
+ "grad_norm": 0.024159753695130348,
+ "learning_rate": 0.0006,
+ "loss": 5.11695671081543,
+ "step": 994
+ },
+ {
+ "epoch": 13.824454148471617,
+ "grad_norm": 0.023281913250684738,
+ "learning_rate": 0.0006,
+ "loss": 5.159364700317383,
+ "step": 995
+ },
+ {
+ "epoch": 13.838427947598253,
+ "grad_norm": 0.02846658229827881,
+ "learning_rate": 0.0006,
+ "loss": 5.0652570724487305,
+ "step": 996
+ },
+ {
+ "epoch": 13.85240174672489,
+ "grad_norm": 0.03165476396679878,
+ "learning_rate": 0.0006,
+ "loss": 5.136387825012207,
+ "step": 997
+ },
+ {
+ "epoch": 13.866375545851529,
+ "grad_norm": 0.02582591027021408,
+ "learning_rate": 0.0006,
+ "loss": 5.059978485107422,
+ "step": 998
+ },
+ {
+ "epoch": 13.880349344978166,
+ "grad_norm": 0.025306783616542816,
+ "learning_rate": 0.0006,
+ "loss": 5.123497486114502,
+ "step": 999
+ },
+ {
+ "epoch": 13.894323144104803,
+ "grad_norm": 0.02598070539534092,
+ "learning_rate": 0.0006,
+ "loss": 5.155566215515137,
+ "step": 1000
+ },
+ {
+ "epoch": 13.908296943231441,
+ "grad_norm": 0.026244031265378,
+ "learning_rate": 0.0006,
+ "loss": 5.0871381759643555,
+ "step": 1001
+ },
+ {
+ "epoch": 13.922270742358078,
+ "grad_norm": 0.025219639763236046,
+ "learning_rate": 0.0006,
+ "loss": 5.102408409118652,
+ "step": 1002
+ },
+ {
+ "epoch": 13.936244541484717,
+ "grad_norm": 0.023916568607091904,
+ "learning_rate": 0.0006,
+ "loss": 5.131472587585449,
+ "step": 1003
+ },
+ {
+ "epoch": 13.950218340611354,
+ "grad_norm": 0.019619153812527657,
+ "learning_rate": 0.0006,
+ "loss": 5.14764404296875,
+ "step": 1004
+ },
+ {
+ "epoch": 13.96419213973799,
+ "grad_norm": 0.019982660189270973,
+ "learning_rate": 0.0006,
+ "loss": 5.0941972732543945,
+ "step": 1005
+ },
+ {
+ "epoch": 13.97816593886463,
+ "grad_norm": 0.018386343494057655,
+ "learning_rate": 0.0006,
+ "loss": 5.138663291931152,
+ "step": 1006
+ },
+ {
+ "epoch": 13.992139737991266,
+ "grad_norm": 0.021332457661628723,
+ "learning_rate": 0.0006,
+ "loss": 5.141696453094482,
+ "step": 1007
+ },
+ {
+ "epoch": 14.0,
+ "grad_norm": 0.021462492644786835,
+ "learning_rate": 0.0006,
+ "loss": 5.049698352813721,
+ "step": 1008
+ },
+ {
+ "epoch": 14.0,
+ "eval_loss": 5.544618129730225,
+ "eval_runtime": 72.1254,
+ "eval_samples_per_second": 33.858,
+ "eval_steps_per_second": 1.068,
+ "step": 1008
+ },
+ {
+ "epoch": 14.013973799126637,
+ "grad_norm": 0.02191290818154812,
+ "learning_rate": 0.0006,
+ "loss": 5.0831732749938965,
+ "step": 1009
+ },
+ {
+ "epoch": 14.027947598253276,
+ "grad_norm": 0.02411566860973835,
+ "learning_rate": 0.0006,
+ "loss": 5.116334915161133,
+ "step": 1010
+ },
+ {
+ "epoch": 14.041921397379912,
+ "grad_norm": 0.027553394436836243,
+ "learning_rate": 0.0006,
+ "loss": 5.050045013427734,
+ "step": 1011
+ },
+ {
+ "epoch": 14.055895196506551,
+ "grad_norm": 0.02357514388859272,
+ "learning_rate": 0.0006,
+ "loss": 5.091012001037598,
+ "step": 1012
+ },
+ {
+ "epoch": 14.069868995633188,
+ "grad_norm": 0.021199429407715797,
+ "learning_rate": 0.0006,
+ "loss": 5.104509353637695,
+ "step": 1013
+ },
+ {
+ "epoch": 14.083842794759825,
+ "grad_norm": 0.019254926592111588,
+ "learning_rate": 0.0006,
+ "loss": 5.141340255737305,
+ "step": 1014
+ },
+ {
+ "epoch": 14.097816593886463,
+ "grad_norm": 0.01811135932803154,
+ "learning_rate": 0.0006,
+ "loss": 5.046824932098389,
+ "step": 1015
+ },
+ {
+ "epoch": 14.1117903930131,
+ "grad_norm": 0.017739299684762955,
+ "learning_rate": 0.0006,
+ "loss": 5.045833587646484,
+ "step": 1016
+ },
+ {
+ "epoch": 14.125764192139737,
+ "grad_norm": 0.018534071743488312,
+ "learning_rate": 0.0006,
+ "loss": 4.9618401527404785,
+ "step": 1017
+ },
+ {
+ "epoch": 14.139737991266376,
+ "grad_norm": 0.016690023243427277,
+ "learning_rate": 0.0006,
+ "loss": 5.165715217590332,
+ "step": 1018
+ },
+ {
+ "epoch": 14.153711790393013,
+ "grad_norm": 0.018630729988217354,
+ "learning_rate": 0.0006,
+ "loss": 5.082975387573242,
+ "step": 1019
+ },
+ {
+ "epoch": 14.167685589519651,
+ "grad_norm": 0.01884593442082405,
+ "learning_rate": 0.0006,
+ "loss": 5.048741817474365,
+ "step": 1020
+ },
+ {
+ "epoch": 14.181659388646288,
+ "grad_norm": 0.018500948324799538,
+ "learning_rate": 0.0006,
+ "loss": 5.105315685272217,
+ "step": 1021
+ },
+ {
+ "epoch": 14.195633187772925,
+ "grad_norm": 0.020263809710741043,
+ "learning_rate": 0.0006,
+ "loss": 5.087735176086426,
+ "step": 1022
+ },
+ {
+ "epoch": 14.209606986899564,
+ "grad_norm": 0.019454646855592728,
+ "learning_rate": 0.0006,
+ "loss": 5.0284810066223145,
+ "step": 1023
+ },
+ {
+ "epoch": 14.2235807860262,
+ "grad_norm": 0.01877632364630699,
+ "learning_rate": 0.0006,
+ "loss": 5.0442609786987305,
+ "step": 1024
+ },
+ {
+ "epoch": 14.237554585152838,
+ "grad_norm": 0.020227789878845215,
+ "learning_rate": 0.0006,
+ "loss": 5.081351280212402,
+ "step": 1025
+ },
+ {
+ "epoch": 14.251528384279476,
+ "grad_norm": 0.018256092444062233,
+ "learning_rate": 0.0006,
+ "loss": 5.016656875610352,
+ "step": 1026
+ },
+ {
+ "epoch": 14.265502183406113,
+ "grad_norm": 0.01916426233947277,
+ "learning_rate": 0.0006,
+ "loss": 5.034103870391846,
+ "step": 1027
+ },
+ {
+ "epoch": 14.279475982532752,
+ "grad_norm": 0.017990969121456146,
+ "learning_rate": 0.0006,
+ "loss": 5.039344310760498,
+ "step": 1028
+ },
+ {
+ "epoch": 14.293449781659389,
+ "grad_norm": 0.018346179276704788,
+ "learning_rate": 0.0006,
+ "loss": 5.017010688781738,
+ "step": 1029
+ },
+ {
+ "epoch": 14.307423580786025,
+ "grad_norm": 0.01584712415933609,
+ "learning_rate": 0.0006,
+ "loss": 4.992602348327637,
+ "step": 1030
+ },
+ {
+ "epoch": 14.321397379912664,
+ "grad_norm": 0.014064064249396324,
+ "learning_rate": 0.0006,
+ "loss": 5.020243167877197,
+ "step": 1031
+ },
+ {
+ "epoch": 14.335371179039301,
+ "grad_norm": 0.014658658765256405,
+ "learning_rate": 0.0006,
+ "loss": 4.995723724365234,
+ "step": 1032
+ },
+ {
+ "epoch": 14.34934497816594,
+ "grad_norm": 0.015613174997270107,
+ "learning_rate": 0.0006,
+ "loss": 5.0587310791015625,
+ "step": 1033
+ },
+ {
+ "epoch": 14.363318777292577,
+ "grad_norm": 0.016534067690372467,
+ "learning_rate": 0.0006,
+ "loss": 4.943386077880859,
+ "step": 1034
+ },
+ {
+ "epoch": 14.377292576419213,
+ "grad_norm": 0.016116850078105927,
+ "learning_rate": 0.0006,
+ "loss": 5.117921829223633,
+ "step": 1035
+ },
+ {
+ "epoch": 14.391266375545852,
+ "grad_norm": 0.014806007035076618,
+ "learning_rate": 0.0006,
+ "loss": 5.077454566955566,
+ "step": 1036
+ },
+ {
+ "epoch": 14.405240174672489,
+ "grad_norm": 0.013742697425186634,
+ "learning_rate": 0.0006,
+ "loss": 5.059109687805176,
+ "step": 1037
+ },
+ {
+ "epoch": 14.419213973799126,
+ "grad_norm": 0.01445091050118208,
+ "learning_rate": 0.0006,
+ "loss": 4.97079610824585,
+ "step": 1038
+ },
+ {
+ "epoch": 14.433187772925764,
+ "grad_norm": 0.014255733229219913,
+ "learning_rate": 0.0006,
+ "loss": 5.056376934051514,
+ "step": 1039
+ },
+ {
+ "epoch": 14.447161572052401,
+ "grad_norm": 0.014356110244989395,
+ "learning_rate": 0.0006,
+ "loss": 5.095856666564941,
+ "step": 1040
+ },
+ {
+ "epoch": 14.46113537117904,
+ "grad_norm": 0.015831749886274338,
+ "learning_rate": 0.0006,
+ "loss": 5.063546180725098,
+ "step": 1041
+ },
+ {
+ "epoch": 14.475109170305677,
+ "grad_norm": 0.01640896685421467,
+ "learning_rate": 0.0006,
+ "loss": 5.038229942321777,
+ "step": 1042
+ },
+ {
+ "epoch": 14.489082969432314,
+ "grad_norm": 0.017240721732378006,
+ "learning_rate": 0.0006,
+ "loss": 4.989873886108398,
+ "step": 1043
+ },
+ {
+ "epoch": 14.503056768558952,
+ "grad_norm": 0.020683523267507553,
+ "learning_rate": 0.0006,
+ "loss": 4.945237159729004,
+ "step": 1044
+ },
+ {
+ "epoch": 14.51703056768559,
+ "grad_norm": 0.02303830161690712,
+ "learning_rate": 0.0006,
+ "loss": 5.039238929748535,
+ "step": 1045
+ },
+ {
+ "epoch": 14.531004366812226,
+ "grad_norm": 0.02372933179140091,
+ "learning_rate": 0.0006,
+ "loss": 5.005778789520264,
+ "step": 1046
+ },
+ {
+ "epoch": 14.544978165938865,
+ "grad_norm": 0.02596980705857277,
+ "learning_rate": 0.0006,
+ "loss": 5.047043800354004,
+ "step": 1047
+ },
+ {
+ "epoch": 14.558951965065502,
+ "grad_norm": 0.03045373223721981,
+ "learning_rate": 0.0006,
+ "loss": 5.07199764251709,
+ "step": 1048
+ },
+ {
+ "epoch": 14.57292576419214,
+ "grad_norm": 0.033439598977565765,
+ "learning_rate": 0.0006,
+ "loss": 5.008430480957031,
+ "step": 1049
+ },
+ {
+ "epoch": 14.586899563318777,
+ "grad_norm": 0.027381205931305885,
+ "learning_rate": 0.0006,
+ "loss": 5.10188102722168,
+ "step": 1050
+ },
+ {
+ "epoch": 14.600873362445414,
+ "grad_norm": 0.029322044923901558,
+ "learning_rate": 0.0006,
+ "loss": 5.0966997146606445,
+ "step": 1051
+ },
+ {
+ "epoch": 14.614847161572053,
+ "grad_norm": 0.02821085974574089,
+ "learning_rate": 0.0006,
+ "loss": 4.9426422119140625,
+ "step": 1052
+ },
+ {
+ "epoch": 14.62882096069869,
+ "grad_norm": 0.03356366604566574,
+ "learning_rate": 0.0006,
+ "loss": 5.083784580230713,
+ "step": 1053
+ },
+ {
+ "epoch": 14.642794759825328,
+ "grad_norm": 0.03138979524374008,
+ "learning_rate": 0.0006,
+ "loss": 5.063078880310059,
+ "step": 1054
+ },
+ {
+ "epoch": 14.656768558951965,
+ "grad_norm": 0.02739642933011055,
+ "learning_rate": 0.0006,
+ "loss": 5.0485687255859375,
+ "step": 1055
+ },
+ {
+ "epoch": 14.670742358078602,
+ "grad_norm": 0.026842772960662842,
+ "learning_rate": 0.0006,
+ "loss": 5.040291786193848,
+ "step": 1056
+ },
+ {
+ "epoch": 14.68471615720524,
+ "grad_norm": 0.03270867466926575,
+ "learning_rate": 0.0006,
+ "loss": 5.118686199188232,
+ "step": 1057
+ },
+ {
+ "epoch": 14.698689956331878,
+ "grad_norm": 0.027592353522777557,
+ "learning_rate": 0.0006,
+ "loss": 5.05806303024292,
+ "step": 1058
+ },
+ {
+ "epoch": 14.712663755458514,
+ "grad_norm": 0.029291553422808647,
+ "learning_rate": 0.0006,
+ "loss": 5.048781394958496,
+ "step": 1059
+ },
+ {
+ "epoch": 14.726637554585153,
+ "grad_norm": 0.027100825682282448,
+ "learning_rate": 0.0006,
+ "loss": 5.09072208404541,
+ "step": 1060
+ },
+ {
+ "epoch": 14.74061135371179,
+ "grad_norm": 0.02333410270512104,
+ "learning_rate": 0.0006,
+ "loss": 5.0193939208984375,
+ "step": 1061
+ },
+ {
+ "epoch": 14.754585152838429,
+ "grad_norm": 0.022513234987854958,
+ "learning_rate": 0.0006,
+ "loss": 4.96424674987793,
+ "step": 1062
+ },
+ {
+ "epoch": 14.768558951965066,
+ "grad_norm": 0.01899099536240101,
+ "learning_rate": 0.0006,
+ "loss": 4.982175350189209,
+ "step": 1063
+ },
+ {
+ "epoch": 14.782532751091702,
+ "grad_norm": 0.0192430317401886,
+ "learning_rate": 0.0006,
+ "loss": 5.094186782836914,
+ "step": 1064
+ },
+ {
+ "epoch": 14.796506550218341,
+ "grad_norm": 0.02133013680577278,
+ "learning_rate": 0.0006,
+ "loss": 5.04017448425293,
+ "step": 1065
+ },
+ {
+ "epoch": 14.810480349344978,
+ "grad_norm": 0.019499894231557846,
+ "learning_rate": 0.0006,
+ "loss": 5.058502197265625,
+ "step": 1066
+ },
+ {
+ "epoch": 14.824454148471617,
+ "grad_norm": 0.016153644770383835,
+ "learning_rate": 0.0006,
+ "loss": 5.006505012512207,
+ "step": 1067
+ },
+ {
+ "epoch": 14.838427947598253,
+ "grad_norm": 0.016342492774128914,
+ "learning_rate": 0.0006,
+ "loss": 4.998578071594238,
+ "step": 1068
+ },
+ {
+ "epoch": 14.85240174672489,
+ "grad_norm": 0.014268498867750168,
+ "learning_rate": 0.0006,
+ "loss": 5.099554538726807,
+ "step": 1069
+ },
+ {
+ "epoch": 14.866375545851529,
+ "grad_norm": 0.014293930493295193,
+ "learning_rate": 0.0006,
+ "loss": 4.929508686065674,
+ "step": 1070
+ },
+ {
+ "epoch": 14.880349344978166,
+ "grad_norm": 0.013110974803566933,
+ "learning_rate": 0.0006,
+ "loss": 4.981067657470703,
+ "step": 1071
+ },
+ {
+ "epoch": 14.894323144104803,
+ "grad_norm": 0.014049272052943707,
+ "learning_rate": 0.0006,
+ "loss": 4.97393274307251,
+ "step": 1072
+ },
+ {
+ "epoch": 14.908296943231441,
+ "grad_norm": 0.011994240805506706,
+ "learning_rate": 0.0006,
+ "loss": 4.9473419189453125,
+ "step": 1073
+ },
+ {
+ "epoch": 14.922270742358078,
+ "grad_norm": 0.013252398930490017,
+ "learning_rate": 0.0006,
+ "loss": 4.965734481811523,
+ "step": 1074
+ },
+ {
+ "epoch": 14.936244541484717,
+ "grad_norm": 0.013359254226088524,
+ "learning_rate": 0.0006,
+ "loss": 4.968218803405762,
+ "step": 1075
+ },
+ {
+ "epoch": 14.950218340611354,
+ "grad_norm": 0.01238918025046587,
+ "learning_rate": 0.0006,
+ "loss": 5.07440710067749,
+ "step": 1076
+ },
+ {
+ "epoch": 14.96419213973799,
+ "grad_norm": 0.014492994174361229,
+ "learning_rate": 0.0006,
+ "loss": 5.0300493240356445,
+ "step": 1077
+ },
+ {
+ "epoch": 14.97816593886463,
+ "grad_norm": 0.01550526823848486,
+ "learning_rate": 0.0006,
+ "loss": 5.031158924102783,
+ "step": 1078
+ },
+ {
+ "epoch": 14.992139737991266,
+ "grad_norm": 0.016029849648475647,
+ "learning_rate": 0.0006,
+ "loss": 5.000946521759033,
+ "step": 1079
+ },
+ {
+ "epoch": 15.0,
+ "grad_norm": 0.0141118373721838,
+ "learning_rate": 0.0006,
+ "loss": 5.079067230224609,
+ "step": 1080
+ },
+ {
+ "epoch": 15.0,
+ "eval_loss": 5.463352203369141,
+ "eval_runtime": 79.3035,
+ "eval_samples_per_second": 30.793,
+ "eval_steps_per_second": 0.971,
+ "step": 1080
+ },
+ {
+ "epoch": 15.013973799126637,
+ "grad_norm": 0.015567110851407051,
+ "learning_rate": 0.0006,
+ "loss": 5.006353855133057,
+ "step": 1081
+ },
+ {
+ "epoch": 15.027947598253276,
+ "grad_norm": 0.017325397580862045,
+ "learning_rate": 0.0006,
+ "loss": 5.022473335266113,
+ "step": 1082
+ },
+ {
+ "epoch": 15.041921397379912,
+ "grad_norm": 0.018784036859869957,
+ "learning_rate": 0.0006,
+ "loss": 4.922494888305664,
+ "step": 1083
+ },
+ {
+ "epoch": 15.055895196506551,
+ "grad_norm": 0.020597755908966064,
+ "learning_rate": 0.0006,
+ "loss": 5.045244216918945,
+ "step": 1084
+ },
+ {
+ "epoch": 15.069868995633188,
+ "grad_norm": 0.023631447926163673,
+ "learning_rate": 0.0006,
+ "loss": 4.916990280151367,
+ "step": 1085
+ },
+ {
+ "epoch": 15.083842794759825,
+ "grad_norm": 0.02467048540711403,
+ "learning_rate": 0.0006,
+ "loss": 5.0631208419799805,
+ "step": 1086
+ },
+ {
+ "epoch": 15.097816593886463,
+ "grad_norm": 0.024088503792881966,
+ "learning_rate": 0.0006,
+ "loss": 4.917226791381836,
+ "step": 1087
+ },
+ {
+ "epoch": 15.1117903930131,
+ "grad_norm": 0.022297589108347893,
+ "learning_rate": 0.0006,
+ "loss": 4.940986156463623,
+ "step": 1088
+ },
+ {
+ "epoch": 15.125764192139737,
+ "grad_norm": 0.02313072420656681,
+ "learning_rate": 0.0006,
+ "loss": 5.013306617736816,
+ "step": 1089
+ },
+ {
+ "epoch": 15.139737991266376,
+ "grad_norm": 0.028538871556520462,
+ "learning_rate": 0.0006,
+ "loss": 4.99459171295166,
+ "step": 1090
+ },
+ {
+ "epoch": 15.153711790393013,
+ "grad_norm": 0.029122058302164078,
+ "learning_rate": 0.0006,
+ "loss": 5.000114440917969,
+ "step": 1091
+ },
+ {
+ "epoch": 15.167685589519651,
+ "grad_norm": 0.022197917103767395,
+ "learning_rate": 0.0006,
+ "loss": 4.945086479187012,
+ "step": 1092
+ },
+ {
+ "epoch": 15.181659388646288,
+ "grad_norm": 0.021084975451231003,
+ "learning_rate": 0.0006,
+ "loss": 4.920244216918945,
+ "step": 1093
+ },
+ {
+ "epoch": 15.195633187772925,
+ "grad_norm": 0.022474950179457664,
+ "learning_rate": 0.0006,
+ "loss": 5.005110740661621,
+ "step": 1094
+ },
+ {
+ "epoch": 15.209606986899564,
+ "grad_norm": 0.02403559722006321,
+ "learning_rate": 0.0006,
+ "loss": 5.04328727722168,
+ "step": 1095
+ },
+ {
+ "epoch": 15.2235807860262,
+ "grad_norm": 0.026955733075737953,
+ "learning_rate": 0.0006,
+ "loss": 4.962291717529297,
+ "step": 1096
+ },
+ {
+ "epoch": 15.237554585152838,
+ "grad_norm": 0.024654097855091095,
+ "learning_rate": 0.0006,
+ "loss": 5.0805840492248535,
+ "step": 1097
+ },
+ {
+ "epoch": 15.251528384279476,
+ "grad_norm": 0.01992264948785305,
+ "learning_rate": 0.0006,
+ "loss": 5.044548034667969,
+ "step": 1098
+ },
+ {
+ "epoch": 15.265502183406113,
+ "grad_norm": 0.01696479320526123,
+ "learning_rate": 0.0006,
+ "loss": 4.912850856781006,
+ "step": 1099
+ },
+ {
+ "epoch": 15.279475982532752,
+ "grad_norm": 0.017840469256043434,
+ "learning_rate": 0.0006,
+ "loss": 5.0556793212890625,
+ "step": 1100
+ },
+ {
+ "epoch": 15.293449781659389,
+ "grad_norm": 0.01623854599893093,
+ "learning_rate": 0.0006,
+ "loss": 5.040976524353027,
+ "step": 1101
+ },
+ {
+ "epoch": 15.307423580786025,
+ "grad_norm": 0.016552211716771126,
+ "learning_rate": 0.0006,
+ "loss": 4.971366882324219,
+ "step": 1102
+ },
+ {
+ "epoch": 15.321397379912664,
+ "grad_norm": 0.015316464938223362,
+ "learning_rate": 0.0006,
+ "loss": 4.952664375305176,
+ "step": 1103
+ },
+ {
+ "epoch": 15.335371179039301,
+ "grad_norm": 0.01712590828537941,
+ "learning_rate": 0.0006,
+ "loss": 4.969407081604004,
+ "step": 1104
+ },
+ {
+ "epoch": 15.34934497816594,
+ "grad_norm": 0.015214908868074417,
+ "learning_rate": 0.0006,
+ "loss": 4.9463911056518555,
+ "step": 1105
+ },
+ {
+ "epoch": 15.363318777292577,
+ "grad_norm": 0.014866896905004978,
+ "learning_rate": 0.0006,
+ "loss": 5.071913242340088,
+ "step": 1106
+ },
+ {
+ "epoch": 15.377292576419213,
+ "grad_norm": 0.01533361803740263,
+ "learning_rate": 0.0006,
+ "loss": 5.0071916580200195,
+ "step": 1107
+ },
+ {
+ "epoch": 15.391266375545852,
+ "grad_norm": 0.014022483490407467,
+ "learning_rate": 0.0006,
+ "loss": 4.954067707061768,
+ "step": 1108
+ },
+ {
+ "epoch": 15.405240174672489,
+ "grad_norm": 0.015019413083791733,
+ "learning_rate": 0.0006,
+ "loss": 4.9905900955200195,
+ "step": 1109
+ },
+ {
+ "epoch": 15.419213973799126,
+ "grad_norm": 0.015474580228328705,
+ "learning_rate": 0.0006,
+ "loss": 4.999631881713867,
+ "step": 1110
+ },
+ {
+ "epoch": 15.433187772925764,
+ "grad_norm": 0.020479638129472733,
+ "learning_rate": 0.0006,
+ "loss": 5.0200700759887695,
+ "step": 1111
+ },
+ {
+ "epoch": 15.447161572052401,
+ "grad_norm": 0.0237763449549675,
+ "learning_rate": 0.0006,
+ "loss": 5.008729457855225,
+ "step": 1112
+ },
+ {
+ "epoch": 15.46113537117904,
+ "grad_norm": 0.022637002170085907,
+ "learning_rate": 0.0006,
+ "loss": 4.940751075744629,
+ "step": 1113
+ },
+ {
+ "epoch": 15.475109170305677,
+ "grad_norm": 0.019718417897820473,
+ "learning_rate": 0.0006,
+ "loss": 4.9788737297058105,
+ "step": 1114
+ },
+ {
+ "epoch": 15.489082969432314,
+ "grad_norm": 0.023200126364827156,
+ "learning_rate": 0.0006,
+ "loss": 4.906790256500244,
+ "step": 1115
+ },
+ {
+ "epoch": 15.503056768558952,
+ "grad_norm": 0.025976872071623802,
+ "learning_rate": 0.0006,
+ "loss": 5.002409934997559,
+ "step": 1116
+ },
+ {
+ "epoch": 15.51703056768559,
+ "grad_norm": 0.023092452436685562,
+ "learning_rate": 0.0006,
+ "loss": 4.90887451171875,
+ "step": 1117
+ },
+ {
+ "epoch": 15.531004366812226,
+ "grad_norm": 0.01955891214311123,
+ "learning_rate": 0.0006,
+ "loss": 4.995199203491211,
+ "step": 1118
+ },
+ {
+ "epoch": 15.544978165938865,
+ "grad_norm": 0.02019096165895462,
+ "learning_rate": 0.0006,
+ "loss": 4.908695697784424,
+ "step": 1119
+ },
+ {
+ "epoch": 15.558951965065502,
+ "grad_norm": 0.019376836717128754,
+ "learning_rate": 0.0006,
+ "loss": 4.872652053833008,
+ "step": 1120
+ },
+ {
+ "epoch": 15.57292576419214,
+ "grad_norm": 0.019068755209445953,
+ "learning_rate": 0.0006,
+ "loss": 4.952585220336914,
+ "step": 1121
+ },
+ {
+ "epoch": 15.586899563318777,
+ "grad_norm": 0.016856160014867783,
+ "learning_rate": 0.0006,
+ "loss": 4.927958965301514,
+ "step": 1122
+ },
+ {
+ "epoch": 15.600873362445414,
+ "grad_norm": 0.017544621601700783,
+ "learning_rate": 0.0006,
+ "loss": 4.922074794769287,
+ "step": 1123
+ },
+ {
+ "epoch": 15.614847161572053,
+ "grad_norm": 0.01814766600728035,
+ "learning_rate": 0.0006,
+ "loss": 4.884115219116211,
+ "step": 1124
+ },
+ {
+ "epoch": 15.62882096069869,
+ "grad_norm": 0.01756724901497364,
+ "learning_rate": 0.0006,
+ "loss": 4.969221115112305,
+ "step": 1125
+ },
+ {
+ "epoch": 15.642794759825328,
+ "grad_norm": 0.016586478799581528,
+ "learning_rate": 0.0006,
+ "loss": 4.897098541259766,
+ "step": 1126
+ },
+ {
+ "epoch": 15.656768558951965,
+ "grad_norm": 0.0165987778455019,
+ "learning_rate": 0.0006,
+ "loss": 4.8751044273376465,
+ "step": 1127
+ },
+ {
+ "epoch": 15.670742358078602,
+ "grad_norm": 0.019033856689929962,
+ "learning_rate": 0.0006,
+ "loss": 4.894170761108398,
+ "step": 1128
+ },
+ {
+ "epoch": 15.68471615720524,
+ "grad_norm": 0.016953010112047195,
+ "learning_rate": 0.0006,
+ "loss": 4.918987274169922,
+ "step": 1129
+ },
+ {
+ "epoch": 15.698689956331878,
+ "grad_norm": 0.016293920576572418,
+ "learning_rate": 0.0006,
+ "loss": 4.949659824371338,
+ "step": 1130
+ },
+ {
+ "epoch": 15.712663755458514,
+ "grad_norm": 0.019263453781604767,
+ "learning_rate": 0.0006,
+ "loss": 4.923409461975098,
+ "step": 1131
+ },
+ {
+ "epoch": 15.726637554585153,
+ "grad_norm": 0.01789483241736889,
+ "learning_rate": 0.0006,
+ "loss": 4.894615173339844,
+ "step": 1132
+ },
+ {
+ "epoch": 15.74061135371179,
+ "grad_norm": 0.01610061153769493,
+ "learning_rate": 0.0006,
+ "loss": 4.9311604499816895,
+ "step": 1133
+ },
+ {
+ "epoch": 15.754585152838429,
+ "grad_norm": 0.01674770377576351,
+ "learning_rate": 0.0006,
+ "loss": 4.938290596008301,
+ "step": 1134
+ },
+ {
+ "epoch": 15.768558951965066,
+ "grad_norm": 0.01579914800822735,
+ "learning_rate": 0.0006,
+ "loss": 4.936288833618164,
+ "step": 1135
+ },
+ {
+ "epoch": 15.782532751091702,
+ "grad_norm": 0.015871549025177956,
+ "learning_rate": 0.0006,
+ "loss": 4.970236301422119,
+ "step": 1136
+ },
+ {
+ "epoch": 15.796506550218341,
+ "grad_norm": 0.014079514890909195,
+ "learning_rate": 0.0006,
+ "loss": 4.935883045196533,
+ "step": 1137
+ },
+ {
+ "epoch": 15.810480349344978,
+ "grad_norm": 0.015348718501627445,
+ "learning_rate": 0.0006,
+ "loss": 4.919315814971924,
+ "step": 1138
+ },
+ {
+ "epoch": 15.824454148471617,
+ "grad_norm": 0.0189143568277359,
+ "learning_rate": 0.0006,
+ "loss": 4.972513198852539,
+ "step": 1139
+ },
+ {
+ "epoch": 15.838427947598253,
+ "grad_norm": 0.021286798641085625,
+ "learning_rate": 0.0006,
+ "loss": 4.89121675491333,
+ "step": 1140
+ },
+ {
+ "epoch": 15.85240174672489,
+ "grad_norm": 0.0212328489869833,
+ "learning_rate": 0.0006,
+ "loss": 4.895394802093506,
+ "step": 1141
+ },
+ {
+ "epoch": 15.866375545851529,
+ "grad_norm": 0.020312150940299034,
+ "learning_rate": 0.0006,
+ "loss": 4.943729400634766,
+ "step": 1142
+ },
+ {
+ "epoch": 15.880349344978166,
+ "grad_norm": 0.02121325396001339,
+ "learning_rate": 0.0006,
+ "loss": 4.9322638511657715,
+ "step": 1143
+ },
+ {
+ "epoch": 15.894323144104803,
+ "grad_norm": 0.02030770666897297,
+ "learning_rate": 0.0006,
+ "loss": 4.8969550132751465,
+ "step": 1144
+ },
+ {
+ "epoch": 15.908296943231441,
+ "grad_norm": 0.02408582717180252,
+ "learning_rate": 0.0006,
+ "loss": 4.94023323059082,
+ "step": 1145
+ },
+ {
+ "epoch": 15.922270742358078,
+ "grad_norm": 0.025230303406715393,
+ "learning_rate": 0.0006,
+ "loss": 4.964077472686768,
+ "step": 1146
+ },
+ {
+ "epoch": 15.936244541484717,
+ "grad_norm": 0.025903481990098953,
+ "learning_rate": 0.0006,
+ "loss": 4.9319024085998535,
+ "step": 1147
+ },
+ {
+ "epoch": 15.950218340611354,
+ "grad_norm": 0.02517629973590374,
+ "learning_rate": 0.0006,
+ "loss": 4.9937262535095215,
+ "step": 1148
+ },
+ {
+ "epoch": 15.96419213973799,
+ "grad_norm": 0.023874718695878983,
+ "learning_rate": 0.0006,
+ "loss": 4.916127681732178,
+ "step": 1149
+ },
+ {
+ "epoch": 15.97816593886463,
+ "grad_norm": 0.02531718835234642,
+ "learning_rate": 0.0006,
+ "loss": 4.949875831604004,
+ "step": 1150
+ },
+ {
+ "epoch": 15.992139737991266,
+ "grad_norm": 0.026585623621940613,
+ "learning_rate": 0.0006,
+ "loss": 4.883786201477051,
+ "step": 1151
+ },
+ {
+ "epoch": 16.0,
+ "grad_norm": 0.029191339388489723,
+ "learning_rate": 0.0006,
+ "loss": 4.955547332763672,
+ "step": 1152
+ },
+ {
+ "epoch": 16.0,
+ "eval_loss": 5.376297473907471,
+ "eval_runtime": 57.4412,
+ "eval_samples_per_second": 42.513,
+ "eval_steps_per_second": 1.341,
+ "step": 1152
+ },
+ {
+ "epoch": 16.01397379912664,
+ "grad_norm": 0.020493783056735992,
+ "learning_rate": 0.0006,
+ "loss": 4.887005805969238,
+ "step": 1153
+ },
+ {
+ "epoch": 16.027947598253274,
+ "grad_norm": 0.022943608462810516,
+ "learning_rate": 0.0006,
+ "loss": 4.9878644943237305,
+ "step": 1154
+ },
+ {
+ "epoch": 16.041921397379912,
+ "grad_norm": 0.026248518377542496,
+ "learning_rate": 0.0006,
+ "loss": 4.929098129272461,
+ "step": 1155
+ },
+ {
+ "epoch": 16.05589519650655,
+ "grad_norm": 0.027000896632671356,
+ "learning_rate": 0.0006,
+ "loss": 4.97839879989624,
+ "step": 1156
+ },
+ {
+ "epoch": 16.069868995633186,
+ "grad_norm": 0.026088254526257515,
+ "learning_rate": 0.0006,
+ "loss": 5.037898540496826,
+ "step": 1157
+ },
+ {
+ "epoch": 16.083842794759825,
+ "grad_norm": 0.027397966012358665,
+ "learning_rate": 0.0006,
+ "loss": 4.8592939376831055,
+ "step": 1158
+ },
+ {
+ "epoch": 16.097816593886463,
+ "grad_norm": 0.024916915223002434,
+ "learning_rate": 0.0006,
+ "loss": 4.908574104309082,
+ "step": 1159
+ },
+ {
+ "epoch": 16.111790393013102,
+ "grad_norm": 0.026325512677431107,
+ "learning_rate": 0.0006,
+ "loss": 4.8777313232421875,
+ "step": 1160
+ },
+ {
+ "epoch": 16.125764192139737,
+ "grad_norm": 0.02321736514568329,
+ "learning_rate": 0.0006,
+ "loss": 4.950716972351074,
+ "step": 1161
+ },
+ {
+ "epoch": 16.139737991266376,
+ "grad_norm": 0.020332302898168564,
+ "learning_rate": 0.0006,
+ "loss": 4.934875965118408,
+ "step": 1162
+ },
+ {
+ "epoch": 16.153711790393015,
+ "grad_norm": 0.02042531967163086,
+ "learning_rate": 0.0006,
+ "loss": 4.937285423278809,
+ "step": 1163
+ },
+ {
+ "epoch": 16.16768558951965,
+ "grad_norm": 0.018892595544457436,
+ "learning_rate": 0.0006,
+ "loss": 4.909456253051758,
+ "step": 1164
+ },
+ {
+ "epoch": 16.18165938864629,
+ "grad_norm": 0.018573295325040817,
+ "learning_rate": 0.0006,
+ "loss": 4.886369228363037,
+ "step": 1165
+ },
+ {
+ "epoch": 16.195633187772927,
+ "grad_norm": 0.019991718232631683,
+ "learning_rate": 0.0006,
+ "loss": 4.923654556274414,
+ "step": 1166
+ },
+ {
+ "epoch": 16.209606986899562,
+ "grad_norm": 0.015044108964502811,
+ "learning_rate": 0.0006,
+ "loss": 4.983735084533691,
+ "step": 1167
+ },
+ {
+ "epoch": 16.2235807860262,
+ "grad_norm": 0.016339287161827087,
+ "learning_rate": 0.0006,
+ "loss": 4.840668678283691,
+ "step": 1168
+ },
+ {
+ "epoch": 16.23755458515284,
+ "grad_norm": 0.015570246614515781,
+ "learning_rate": 0.0006,
+ "loss": 4.88963508605957,
+ "step": 1169
+ },
+ {
+ "epoch": 16.251528384279474,
+ "grad_norm": 0.014452921226620674,
+ "learning_rate": 0.0006,
+ "loss": 4.931870460510254,
+ "step": 1170
+ },
+ {
+ "epoch": 16.265502183406113,
+ "grad_norm": 0.014817588031291962,
+ "learning_rate": 0.0006,
+ "loss": 4.893659591674805,
+ "step": 1171
+ },
+ {
+ "epoch": 16.27947598253275,
+ "grad_norm": 0.013385407626628876,
+ "learning_rate": 0.0006,
+ "loss": 4.899247646331787,
+ "step": 1172
+ },
+ {
+ "epoch": 16.29344978165939,
+ "grad_norm": 0.01410810649394989,
+ "learning_rate": 0.0006,
+ "loss": 4.9700541496276855,
+ "step": 1173
+ },
+ {
+ "epoch": 16.307423580786025,
+ "grad_norm": 0.015247197821736336,
+ "learning_rate": 0.0006,
+ "loss": 4.945037841796875,
+ "step": 1174
+ },
+ {
+ "epoch": 16.321397379912664,
+ "grad_norm": 0.015484587289392948,
+ "learning_rate": 0.0006,
+ "loss": 4.924195289611816,
+ "step": 1175
+ },
+ {
+ "epoch": 16.335371179039303,
+ "grad_norm": 0.014522048644721508,
+ "learning_rate": 0.0006,
+ "loss": 4.89838981628418,
+ "step": 1176
+ },
+ {
+ "epoch": 16.349344978165938,
+ "grad_norm": 0.015260211192071438,
+ "learning_rate": 0.0006,
+ "loss": 4.845786094665527,
+ "step": 1177
+ },
+ {
+ "epoch": 16.363318777292577,
+ "grad_norm": 0.015550734475255013,
+ "learning_rate": 0.0006,
+ "loss": 4.893673896789551,
+ "step": 1178
+ },
+ {
+ "epoch": 16.377292576419215,
+ "grad_norm": 0.013600634410977364,
+ "learning_rate": 0.0006,
+ "loss": 4.906546592712402,
+ "step": 1179
+ },
+ {
+ "epoch": 16.39126637554585,
+ "grad_norm": 0.013318133540451527,
+ "learning_rate": 0.0006,
+ "loss": 4.897005081176758,
+ "step": 1180
+ },
+ {
+ "epoch": 16.40524017467249,
+ "grad_norm": 0.013978407718241215,
+ "learning_rate": 0.0006,
+ "loss": 4.850296974182129,
+ "step": 1181
+ },
+ {
+ "epoch": 16.419213973799128,
+ "grad_norm": 0.014549889601767063,
+ "learning_rate": 0.0006,
+ "loss": 4.874753952026367,
+ "step": 1182
+ },
+ {
+ "epoch": 16.433187772925763,
+ "grad_norm": 0.018115175887942314,
+ "learning_rate": 0.0006,
+ "loss": 4.818711280822754,
+ "step": 1183
+ },
+ {
+ "epoch": 16.4471615720524,
+ "grad_norm": 0.018728114664554596,
+ "learning_rate": 0.0006,
+ "loss": 4.882294654846191,
+ "step": 1184
+ },
+ {
+ "epoch": 16.46113537117904,
+ "grad_norm": 0.018031353130936623,
+ "learning_rate": 0.0006,
+ "loss": 4.828677177429199,
+ "step": 1185
+ },
+ {
+ "epoch": 16.475109170305675,
+ "grad_norm": 0.017293527722358704,
+ "learning_rate": 0.0006,
+ "loss": 4.872529983520508,
+ "step": 1186
+ },
+ {
+ "epoch": 16.489082969432314,
+ "grad_norm": 0.015710802748799324,
+ "learning_rate": 0.0006,
+ "loss": 4.921950817108154,
+ "step": 1187
+ },
+ {
+ "epoch": 16.503056768558952,
+ "grad_norm": 0.017632456496357918,
+ "learning_rate": 0.0006,
+ "loss": 4.907940864562988,
+ "step": 1188
+ },
+ {
+ "epoch": 16.51703056768559,
+ "grad_norm": 0.018707087263464928,
+ "learning_rate": 0.0006,
+ "loss": 4.797819137573242,
+ "step": 1189
+ },
+ {
+ "epoch": 16.531004366812226,
+ "grad_norm": 0.019626259803771973,
+ "learning_rate": 0.0006,
+ "loss": 4.795708656311035,
+ "step": 1190
+ },
+ {
+ "epoch": 16.544978165938865,
+ "grad_norm": 0.018569709733128548,
+ "learning_rate": 0.0006,
+ "loss": 4.876678466796875,
+ "step": 1191
+ },
+ {
+ "epoch": 16.558951965065503,
+ "grad_norm": 0.01720404624938965,
+ "learning_rate": 0.0006,
+ "loss": 4.922236442565918,
+ "step": 1192
+ },
+ {
+ "epoch": 16.57292576419214,
+ "grad_norm": 0.018940523266792297,
+ "learning_rate": 0.0006,
+ "loss": 4.892475605010986,
+ "step": 1193
+ },
+ {
+ "epoch": 16.586899563318777,
+ "grad_norm": 0.020524146035313606,
+ "learning_rate": 0.0006,
+ "loss": 4.814975738525391,
+ "step": 1194
+ },
+ {
+ "epoch": 16.600873362445416,
+ "grad_norm": 0.017521994188427925,
+ "learning_rate": 0.0006,
+ "loss": 4.912031173706055,
+ "step": 1195
+ },
+ {
+ "epoch": 16.61484716157205,
+ "grad_norm": 0.019237685948610306,
+ "learning_rate": 0.0006,
+ "loss": 4.8619184494018555,
+ "step": 1196
+ },
+ {
+ "epoch": 16.62882096069869,
+ "grad_norm": 0.018941357731819153,
+ "learning_rate": 0.0006,
+ "loss": 4.983427047729492,
+ "step": 1197
+ },
+ {
+ "epoch": 16.64279475982533,
+ "grad_norm": 0.01969471573829651,
+ "learning_rate": 0.0006,
+ "loss": 4.877185344696045,
+ "step": 1198
+ },
+ {
+ "epoch": 16.656768558951963,
+ "grad_norm": 0.01886482909321785,
+ "learning_rate": 0.0006,
+ "loss": 4.858532905578613,
+ "step": 1199
+ },
+ {
+ "epoch": 16.670742358078602,
+ "grad_norm": 0.018075549975037575,
+ "learning_rate": 0.0006,
+ "loss": 4.85374641418457,
+ "step": 1200
+ },
+ {
+ "epoch": 16.68471615720524,
+ "grad_norm": 0.018726196140050888,
+ "learning_rate": 0.0006,
+ "loss": 4.9164719581604,
+ "step": 1201
+ },
+ {
+ "epoch": 16.69868995633188,
+ "grad_norm": 0.020012663677334785,
+ "learning_rate": 0.0006,
+ "loss": 4.790283203125,
+ "step": 1202
+ },
+ {
+ "epoch": 16.712663755458514,
+ "grad_norm": 0.01965046115219593,
+ "learning_rate": 0.0006,
+ "loss": 4.833796501159668,
+ "step": 1203
+ },
+ {
+ "epoch": 16.726637554585153,
+ "grad_norm": 0.020102867856621742,
+ "learning_rate": 0.0006,
+ "loss": 4.835939407348633,
+ "step": 1204
+ },
+ {
+ "epoch": 16.74061135371179,
+ "grad_norm": 0.01858644187450409,
+ "learning_rate": 0.0006,
+ "loss": 4.819847106933594,
+ "step": 1205
+ },
+ {
+ "epoch": 16.754585152838427,
+ "grad_norm": 0.02036326937377453,
+ "learning_rate": 0.0006,
+ "loss": 4.9040985107421875,
+ "step": 1206
+ },
+ {
+ "epoch": 16.768558951965066,
+ "grad_norm": 0.023322949185967445,
+ "learning_rate": 0.0006,
+ "loss": 4.824136734008789,
+ "step": 1207
+ },
+ {
+ "epoch": 16.782532751091704,
+ "grad_norm": 0.023923879489302635,
+ "learning_rate": 0.0006,
+ "loss": 4.91016960144043,
+ "step": 1208
+ },
+ {
+ "epoch": 16.79650655021834,
+ "grad_norm": 0.02464689128100872,
+ "learning_rate": 0.0006,
+ "loss": 4.949073791503906,
+ "step": 1209
+ },
+ {
+ "epoch": 16.810480349344978,
+ "grad_norm": 0.02039799466729164,
+ "learning_rate": 0.0006,
+ "loss": 4.93691349029541,
+ "step": 1210
+ },
+ {
+ "epoch": 16.824454148471617,
+ "grad_norm": 0.019222574308514595,
+ "learning_rate": 0.0006,
+ "loss": 4.936672210693359,
+ "step": 1211
+ },
+ {
+ "epoch": 16.83842794759825,
+ "grad_norm": 0.017282437533140182,
+ "learning_rate": 0.0006,
+ "loss": 4.818748950958252,
+ "step": 1212
+ },
+ {
+ "epoch": 16.85240174672489,
+ "grad_norm": 0.021188177168369293,
+ "learning_rate": 0.0006,
+ "loss": 4.880117416381836,
+ "step": 1213
+ },
+ {
+ "epoch": 16.86637554585153,
+ "grad_norm": 0.027654651552438736,
+ "learning_rate": 0.0006,
+ "loss": 4.827604293823242,
+ "step": 1214
+ },
+ {
+ "epoch": 16.880349344978168,
+ "grad_norm": 0.028346989303827286,
+ "learning_rate": 0.0006,
+ "loss": 4.872309684753418,
+ "step": 1215
+ },
+ {
+ "epoch": 16.894323144104803,
+ "grad_norm": 0.02339531108736992,
+ "learning_rate": 0.0006,
+ "loss": 4.87053918838501,
+ "step": 1216
+ },
+ {
+ "epoch": 16.90829694323144,
+ "grad_norm": 0.02051672339439392,
+ "learning_rate": 0.0006,
+ "loss": 4.901670455932617,
+ "step": 1217
+ },
+ {
+ "epoch": 16.92227074235808,
+ "grad_norm": 0.02038448490202427,
+ "learning_rate": 0.0006,
+ "loss": 4.822744846343994,
+ "step": 1218
+ },
+ {
+ "epoch": 16.936244541484715,
+ "grad_norm": 0.0207635797560215,
+ "learning_rate": 0.0006,
+ "loss": 4.83148717880249,
+ "step": 1219
+ },
+ {
+ "epoch": 16.950218340611354,
+ "grad_norm": 0.020679432898759842,
+ "learning_rate": 0.0006,
+ "loss": 4.870957374572754,
+ "step": 1220
+ },
+ {
+ "epoch": 16.964192139737992,
+ "grad_norm": 0.019351216033101082,
+ "learning_rate": 0.0006,
+ "loss": 4.956622123718262,
+ "step": 1221
+ },
+ {
+ "epoch": 16.978165938864628,
+ "grad_norm": 0.018666930496692657,
+ "learning_rate": 0.0006,
+ "loss": 4.853818893432617,
+ "step": 1222
+ },
+ {
+ "epoch": 16.992139737991266,
+ "grad_norm": 0.019687574356794357,
+ "learning_rate": 0.0006,
+ "loss": 4.827330589294434,
+ "step": 1223
+ },
+ {
+ "epoch": 17.0,
+ "grad_norm": 0.020736118778586388,
+ "learning_rate": 0.0006,
+ "loss": 4.856723785400391,
+ "step": 1224
+ },
+ {
+ "epoch": 17.0,
+ "eval_loss": 5.262957572937012,
+ "eval_runtime": 59.9518,
+ "eval_samples_per_second": 40.733,
+ "eval_steps_per_second": 1.284,
+ "step": 1224
+ },
+ {
+ "epoch": 17.01397379912664,
+ "grad_norm": 0.019920889288187027,
+ "learning_rate": 0.0006,
+ "loss": 4.925846099853516,
+ "step": 1225
+ },
+ {
+ "epoch": 17.027947598253274,
+ "grad_norm": 0.020640799775719643,
+ "learning_rate": 0.0006,
+ "loss": 4.83962345123291,
+ "step": 1226
+ },
+ {
+ "epoch": 17.041921397379912,
+ "grad_norm": 0.019742414355278015,
+ "learning_rate": 0.0006,
+ "loss": 4.935937881469727,
+ "step": 1227
+ },
+ {
+ "epoch": 17.05589519650655,
+ "grad_norm": 0.018709300085902214,
+ "learning_rate": 0.0006,
+ "loss": 4.866799831390381,
+ "step": 1228
+ },
+ {
+ "epoch": 17.069868995633186,
+ "grad_norm": 0.01504162885248661,
+ "learning_rate": 0.0006,
+ "loss": 4.810856819152832,
+ "step": 1229
+ },
+ {
+ "epoch": 17.083842794759825,
+ "grad_norm": 0.014874175190925598,
+ "learning_rate": 0.0006,
+ "loss": 4.881101608276367,
+ "step": 1230
+ },
+ {
+ "epoch": 17.097816593886463,
+ "grad_norm": 0.017353734001517296,
+ "learning_rate": 0.0006,
+ "loss": 4.846688747406006,
+ "step": 1231
+ },
+ {
+ "epoch": 17.111790393013102,
+ "grad_norm": 0.017002031207084656,
+ "learning_rate": 0.0006,
+ "loss": 4.807414531707764,
+ "step": 1232
+ },
+ {
+ "epoch": 17.125764192139737,
+ "grad_norm": 0.016849525272846222,
+ "learning_rate": 0.0006,
+ "loss": 4.811706066131592,
+ "step": 1233
+ },
+ {
+ "epoch": 17.139737991266376,
+ "grad_norm": 0.017146047204732895,
+ "learning_rate": 0.0006,
+ "loss": 4.811017990112305,
+ "step": 1234
+ },
+ {
+ "epoch": 17.153711790393015,
+ "grad_norm": 0.018497949466109276,
+ "learning_rate": 0.0006,
+ "loss": 4.8794403076171875,
+ "step": 1235
+ },
+ {
+ "epoch": 17.16768558951965,
+ "grad_norm": 0.02204900048673153,
+ "learning_rate": 0.0006,
+ "loss": 4.917919635772705,
+ "step": 1236
+ },
+ {
+ "epoch": 17.18165938864629,
+ "grad_norm": 0.02386346273124218,
+ "learning_rate": 0.0006,
+ "loss": 4.941521644592285,
+ "step": 1237
+ },
+ {
+ "epoch": 17.195633187772927,
+ "grad_norm": 0.019803019240498543,
+ "learning_rate": 0.0006,
+ "loss": 4.772615432739258,
+ "step": 1238
+ },
+ {
+ "epoch": 17.209606986899562,
+ "grad_norm": 0.01547364704310894,
+ "learning_rate": 0.0006,
+ "loss": 4.817747592926025,
+ "step": 1239
+ },
+ {
+ "epoch": 17.2235807860262,
+ "grad_norm": 0.015355406329035759,
+ "learning_rate": 0.0006,
+ "loss": 4.9097771644592285,
+ "step": 1240
+ },
+ {
+ "epoch": 17.23755458515284,
+ "grad_norm": 0.013763652183115482,
+ "learning_rate": 0.0006,
+ "loss": 4.790132522583008,
+ "step": 1241
+ },
+ {
+ "epoch": 17.251528384279474,
+ "grad_norm": 0.015364146791398525,
+ "learning_rate": 0.0006,
+ "loss": 4.876467704772949,
+ "step": 1242
+ },
+ {
+ "epoch": 17.265502183406113,
+ "grad_norm": 0.01718062348663807,
+ "learning_rate": 0.0006,
+ "loss": 4.871622562408447,
+ "step": 1243
+ },
+ {
+ "epoch": 17.27947598253275,
+ "grad_norm": 0.01469398569315672,
+ "learning_rate": 0.0006,
+ "loss": 4.829773426055908,
+ "step": 1244
+ },
+ {
+ "epoch": 17.29344978165939,
+ "grad_norm": 0.015336665324866772,
+ "learning_rate": 0.0006,
+ "loss": 4.886260032653809,
+ "step": 1245
+ },
+ {
+ "epoch": 17.307423580786025,
+ "grad_norm": 0.01615484617650509,
+ "learning_rate": 0.0006,
+ "loss": 4.840301990509033,
+ "step": 1246
+ },
+ {
+ "epoch": 17.321397379912664,
+ "grad_norm": 0.017155757173895836,
+ "learning_rate": 0.0006,
+ "loss": 4.7739787101745605,
+ "step": 1247
+ },
+ {
+ "epoch": 17.335371179039303,
+ "grad_norm": 0.016230205073952675,
+ "learning_rate": 0.0006,
+ "loss": 4.790524005889893,
+ "step": 1248
+ },
+ {
+ "epoch": 17.349344978165938,
+ "grad_norm": 0.016484249383211136,
+ "learning_rate": 0.0006,
+ "loss": 4.948139190673828,
+ "step": 1249
+ },
+ {
+ "epoch": 17.363318777292577,
+ "grad_norm": 0.016610991209745407,
+ "learning_rate": 0.0006,
+ "loss": 4.872422218322754,
+ "step": 1250
+ },
+ {
+ "epoch": 17.377292576419215,
+ "grad_norm": 0.017981581389904022,
+ "learning_rate": 0.0006,
+ "loss": 4.873099327087402,
+ "step": 1251
+ },
+ {
+ "epoch": 17.39126637554585,
+ "grad_norm": 0.020764052867889404,
+ "learning_rate": 0.0006,
+ "loss": 4.840051174163818,
+ "step": 1252
+ },
+ {
+ "epoch": 17.40524017467249,
+ "grad_norm": 0.02078232914209366,
+ "learning_rate": 0.0006,
+ "loss": 4.776272773742676,
+ "step": 1253
+ },
+ {
+ "epoch": 17.419213973799128,
+ "grad_norm": 0.020214475691318512,
+ "learning_rate": 0.0006,
+ "loss": 4.83798360824585,
+ "step": 1254
+ },
+ {
+ "epoch": 17.433187772925763,
+ "grad_norm": 0.02398747019469738,
+ "learning_rate": 0.0006,
+ "loss": 4.770004749298096,
+ "step": 1255
+ },
+ {
+ "epoch": 17.4471615720524,
+ "grad_norm": 0.024379247799515724,
+ "learning_rate": 0.0006,
+ "loss": 4.905134201049805,
+ "step": 1256
+ },
+ {
+ "epoch": 17.46113537117904,
+ "grad_norm": 0.020821094512939453,
+ "learning_rate": 0.0006,
+ "loss": 4.821808815002441,
+ "step": 1257
+ },
+ {
+ "epoch": 17.475109170305675,
+ "grad_norm": 0.018303627148270607,
+ "learning_rate": 0.0006,
+ "loss": 4.7170562744140625,
+ "step": 1258
+ },
+ {
+ "epoch": 17.489082969432314,
+ "grad_norm": 0.01970406249165535,
+ "learning_rate": 0.0006,
+ "loss": 4.872300148010254,
+ "step": 1259
+ },
+ {
+ "epoch": 17.503056768558952,
+ "grad_norm": 0.01975054107606411,
+ "learning_rate": 0.0006,
+ "loss": 4.850522994995117,
+ "step": 1260
+ },
+ {
+ "epoch": 17.51703056768559,
+ "grad_norm": 0.018307320773601532,
+ "learning_rate": 0.0006,
+ "loss": 4.80927038192749,
+ "step": 1261
+ },
+ {
+ "epoch": 17.531004366812226,
+ "grad_norm": 0.0156140411272645,
+ "learning_rate": 0.0006,
+ "loss": 4.816516399383545,
+ "step": 1262
+ },
+ {
+ "epoch": 17.544978165938865,
+ "grad_norm": 0.0153433782979846,
+ "learning_rate": 0.0006,
+ "loss": 4.719747066497803,
+ "step": 1263
+ },
+ {
+ "epoch": 17.558951965065503,
+ "grad_norm": 0.014209619723260403,
+ "learning_rate": 0.0006,
+ "loss": 4.838152885437012,
+ "step": 1264
+ },
+ {
+ "epoch": 17.57292576419214,
+ "grad_norm": 0.015228115022182465,
+ "learning_rate": 0.0006,
+ "loss": 4.744102954864502,
+ "step": 1265
+ },
+ {
+ "epoch": 17.586899563318777,
+ "grad_norm": 0.015567339025437832,
+ "learning_rate": 0.0006,
+ "loss": 4.789331436157227,
+ "step": 1266
+ },
+ {
+ "epoch": 17.600873362445416,
+ "grad_norm": 0.01851458102464676,
+ "learning_rate": 0.0006,
+ "loss": 4.847858905792236,
+ "step": 1267
+ },
+ {
+ "epoch": 17.61484716157205,
+ "grad_norm": 0.01986544393002987,
+ "learning_rate": 0.0006,
+ "loss": 4.826117992401123,
+ "step": 1268
+ },
+ {
+ "epoch": 17.62882096069869,
+ "grad_norm": 0.0173768550157547,
+ "learning_rate": 0.0006,
+ "loss": 4.838115215301514,
+ "step": 1269
+ },
+ {
+ "epoch": 17.64279475982533,
+ "grad_norm": 0.01780487783253193,
+ "learning_rate": 0.0006,
+ "loss": 4.820503234863281,
+ "step": 1270
+ },
+ {
+ "epoch": 17.656768558951963,
+ "grad_norm": 0.018898125737905502,
+ "learning_rate": 0.0006,
+ "loss": 4.8488359451293945,
+ "step": 1271
+ },
+ {
+ "epoch": 17.670742358078602,
+ "grad_norm": 0.016740918159484863,
+ "learning_rate": 0.0006,
+ "loss": 4.815840244293213,
+ "step": 1272
+ },
+ {
+ "epoch": 17.68471615720524,
+ "grad_norm": 0.017316583544015884,
+ "learning_rate": 0.0006,
+ "loss": 4.787709712982178,
+ "step": 1273
+ },
+ {
+ "epoch": 17.69868995633188,
+ "grad_norm": 0.01631874032318592,
+ "learning_rate": 0.0006,
+ "loss": 4.807087421417236,
+ "step": 1274
+ },
+ {
+ "epoch": 17.712663755458514,
+ "grad_norm": 0.018026666715741158,
+ "learning_rate": 0.0006,
+ "loss": 4.810066223144531,
+ "step": 1275
+ },
+ {
+ "epoch": 17.726637554585153,
+ "grad_norm": 0.01884373649954796,
+ "learning_rate": 0.0006,
+ "loss": 4.852358818054199,
+ "step": 1276
+ },
+ {
+ "epoch": 17.74061135371179,
+ "grad_norm": 0.01915278099477291,
+ "learning_rate": 0.0006,
+ "loss": 4.850739002227783,
+ "step": 1277
+ },
+ {
+ "epoch": 17.754585152838427,
+ "grad_norm": 0.01981119066476822,
+ "learning_rate": 0.0006,
+ "loss": 4.794268608093262,
+ "step": 1278
+ },
+ {
+ "epoch": 17.768558951965066,
+ "grad_norm": 0.020773326978087425,
+ "learning_rate": 0.0006,
+ "loss": 4.789695739746094,
+ "step": 1279
+ },
+ {
+ "epoch": 17.782532751091704,
+ "grad_norm": 0.0198501106351614,
+ "learning_rate": 0.0006,
+ "loss": 4.812196731567383,
+ "step": 1280
+ },
+ {
+ "epoch": 17.79650655021834,
+ "grad_norm": 0.016126036643981934,
+ "learning_rate": 0.0006,
+ "loss": 4.8933563232421875,
+ "step": 1281
+ },
+ {
+ "epoch": 17.810480349344978,
+ "grad_norm": 0.015466060489416122,
+ "learning_rate": 0.0006,
+ "loss": 4.765222549438477,
+ "step": 1282
+ },
+ {
+ "epoch": 17.824454148471617,
+ "grad_norm": 0.01579921506345272,
+ "learning_rate": 0.0006,
+ "loss": 4.859963893890381,
+ "step": 1283
+ },
+ {
+ "epoch": 17.83842794759825,
+ "grad_norm": 0.015828445553779602,
+ "learning_rate": 0.0006,
+ "loss": 4.8465094566345215,
+ "step": 1284
+ },
+ {
+ "epoch": 17.85240174672489,
+ "grad_norm": 0.01661691628396511,
+ "learning_rate": 0.0006,
+ "loss": 4.766627311706543,
+ "step": 1285
+ },
+ {
+ "epoch": 17.86637554585153,
+ "grad_norm": 0.01876002363860607,
+ "learning_rate": 0.0006,
+ "loss": 4.792072296142578,
+ "step": 1286
+ },
+ {
+ "epoch": 17.880349344978168,
+ "grad_norm": 0.01939656026661396,
+ "learning_rate": 0.0006,
+ "loss": 4.8140668869018555,
+ "step": 1287
+ },
+ {
+ "epoch": 17.894323144104803,
+ "grad_norm": 0.019839171320199966,
+ "learning_rate": 0.0006,
+ "loss": 4.909534454345703,
+ "step": 1288
+ },
+ {
+ "epoch": 17.90829694323144,
+ "grad_norm": 0.020329177379608154,
+ "learning_rate": 0.0006,
+ "loss": 4.803993225097656,
+ "step": 1289
+ },
+ {
+ "epoch": 17.92227074235808,
+ "grad_norm": 0.019034743309020996,
+ "learning_rate": 0.0006,
+ "loss": 4.822991847991943,
+ "step": 1290
+ },
+ {
+ "epoch": 17.936244541484715,
+ "grad_norm": 0.018672214820981026,
+ "learning_rate": 0.0006,
+ "loss": 4.771214485168457,
+ "step": 1291
+ },
+ {
+ "epoch": 17.950218340611354,
+ "grad_norm": 0.019761107861995697,
+ "learning_rate": 0.0006,
+ "loss": 4.84083366394043,
+ "step": 1292
+ },
+ {
+ "epoch": 17.964192139737992,
+ "grad_norm": 0.018283944576978683,
+ "learning_rate": 0.0006,
+ "loss": 4.782181739807129,
+ "step": 1293
+ },
+ {
+ "epoch": 17.978165938864628,
+ "grad_norm": 0.02122645638883114,
+ "learning_rate": 0.0006,
+ "loss": 4.7942962646484375,
+ "step": 1294
+ },
+ {
+ "epoch": 17.992139737991266,
+ "grad_norm": 0.02146247588098049,
+ "learning_rate": 0.0006,
+ "loss": 4.7972211837768555,
+ "step": 1295
+ },
+ {
+ "epoch": 18.0,
+ "grad_norm": 0.022740444168448448,
+ "learning_rate": 0.0006,
+ "loss": 4.778811931610107,
+ "step": 1296
+ },
+ {
+ "epoch": 18.0,
+ "eval_loss": 5.211768627166748,
+ "eval_runtime": 58.8095,
+ "eval_samples_per_second": 41.524,
+ "eval_steps_per_second": 1.309,
+ "step": 1296
+ },
+ {
+ "epoch": 18.01397379912664,
+ "grad_norm": 0.02356165461242199,
+ "learning_rate": 0.0006,
+ "loss": 4.815979480743408,
+ "step": 1297
+ },
+ {
+ "epoch": 18.027947598253274,
+ "grad_norm": 0.024365399032831192,
+ "learning_rate": 0.0006,
+ "loss": 4.731906890869141,
+ "step": 1298
+ },
+ {
+ "epoch": 18.041921397379912,
+ "grad_norm": 0.02496817521750927,
+ "learning_rate": 0.0006,
+ "loss": 4.798555850982666,
+ "step": 1299
+ },
+ {
+ "epoch": 18.05589519650655,
+ "grad_norm": 0.023278620094060898,
+ "learning_rate": 0.0006,
+ "loss": 4.764307975769043,
+ "step": 1300
+ },
+ {
+ "epoch": 18.069868995633186,
+ "grad_norm": 0.020955104380846024,
+ "learning_rate": 0.0006,
+ "loss": 4.766234874725342,
+ "step": 1301
+ },
+ {
+ "epoch": 18.083842794759825,
+ "grad_norm": 0.021213296800851822,
+ "learning_rate": 0.0006,
+ "loss": 4.856618881225586,
+ "step": 1302
+ },
+ {
+ "epoch": 18.097816593886463,
+ "grad_norm": 0.022735338658094406,
+ "learning_rate": 0.0006,
+ "loss": 4.853489875793457,
+ "step": 1303
+ },
+ {
+ "epoch": 18.111790393013102,
+ "grad_norm": 0.024209095165133476,
+ "learning_rate": 0.0006,
+ "loss": 4.858719825744629,
+ "step": 1304
+ },
+ {
+ "epoch": 18.125764192139737,
+ "grad_norm": 0.025767242535948753,
+ "learning_rate": 0.0006,
+ "loss": 4.871748447418213,
+ "step": 1305
+ },
+ {
+ "epoch": 18.139737991266376,
+ "grad_norm": 0.021460559219121933,
+ "learning_rate": 0.0006,
+ "loss": 4.686347961425781,
+ "step": 1306
+ },
+ {
+ "epoch": 18.153711790393015,
+ "grad_norm": 0.019937612116336823,
+ "learning_rate": 0.0006,
+ "loss": 4.792636871337891,
+ "step": 1307
+ },
+ {
+ "epoch": 18.16768558951965,
+ "grad_norm": 0.022846097126603127,
+ "learning_rate": 0.0006,
+ "loss": 4.818110466003418,
+ "step": 1308
+ },
+ {
+ "epoch": 18.18165938864629,
+ "grad_norm": 0.020246420055627823,
+ "learning_rate": 0.0006,
+ "loss": 4.855318546295166,
+ "step": 1309
+ },
+ {
+ "epoch": 18.195633187772927,
+ "grad_norm": 0.017798585817217827,
+ "learning_rate": 0.0006,
+ "loss": 4.776525497436523,
+ "step": 1310
+ },
+ {
+ "epoch": 18.209606986899562,
+ "grad_norm": 0.01727782003581524,
+ "learning_rate": 0.0006,
+ "loss": 4.850035667419434,
+ "step": 1311
+ },
+ {
+ "epoch": 18.2235807860262,
+ "grad_norm": 0.01917205937206745,
+ "learning_rate": 0.0006,
+ "loss": 4.787570953369141,
+ "step": 1312
+ },
+ {
+ "epoch": 18.23755458515284,
+ "grad_norm": 0.020757799968123436,
+ "learning_rate": 0.0006,
+ "loss": 4.797884941101074,
+ "step": 1313
+ },
+ {
+ "epoch": 18.251528384279474,
+ "grad_norm": 0.019872257485985756,
+ "learning_rate": 0.0006,
+ "loss": 4.823526382446289,
+ "step": 1314
+ },
+ {
+ "epoch": 18.265502183406113,
+ "grad_norm": 0.017858969047665596,
+ "learning_rate": 0.0006,
+ "loss": 4.774180889129639,
+ "step": 1315
+ },
+ {
+ "epoch": 18.27947598253275,
+ "grad_norm": 0.01774270087480545,
+ "learning_rate": 0.0006,
+ "loss": 4.739023208618164,
+ "step": 1316
+ },
+ {
+ "epoch": 18.29344978165939,
+ "grad_norm": 0.018475241959095,
+ "learning_rate": 0.0006,
+ "loss": 4.8812456130981445,
+ "step": 1317
+ },
+ {
+ "epoch": 18.307423580786025,
+ "grad_norm": 0.018330633640289307,
+ "learning_rate": 0.0006,
+ "loss": 4.738748073577881,
+ "step": 1318
+ },
+ {
+ "epoch": 18.321397379912664,
+ "grad_norm": 0.016155779361724854,
+ "learning_rate": 0.0006,
+ "loss": 4.81126594543457,
+ "step": 1319
+ },
+ {
+ "epoch": 18.335371179039303,
+ "grad_norm": 0.016643118113279343,
+ "learning_rate": 0.0006,
+ "loss": 4.796675682067871,
+ "step": 1320
+ },
+ {
+ "epoch": 18.349344978165938,
+ "grad_norm": 0.0177546925842762,
+ "learning_rate": 0.0006,
+ "loss": 4.7574028968811035,
+ "step": 1321
+ },
+ {
+ "epoch": 18.363318777292577,
+ "grad_norm": 0.017025742679834366,
+ "learning_rate": 0.0006,
+ "loss": 4.737842559814453,
+ "step": 1322
+ },
+ {
+ "epoch": 18.377292576419215,
+ "grad_norm": 0.01645711250603199,
+ "learning_rate": 0.0006,
+ "loss": 4.762307643890381,
+ "step": 1323
+ },
+ {
+ "epoch": 18.39126637554585,
+ "grad_norm": 0.015202434733510017,
+ "learning_rate": 0.0006,
+ "loss": 4.844311237335205,
+ "step": 1324
+ },
+ {
+ "epoch": 18.40524017467249,
+ "grad_norm": 0.013969679363071918,
+ "learning_rate": 0.0006,
+ "loss": 4.757219314575195,
+ "step": 1325
+ },
+ {
+ "epoch": 18.419213973799128,
+ "grad_norm": 0.013256058096885681,
+ "learning_rate": 0.0006,
+ "loss": 4.748269081115723,
+ "step": 1326
+ },
+ {
+ "epoch": 18.433187772925763,
+ "grad_norm": 0.014061232097446918,
+ "learning_rate": 0.0006,
+ "loss": 4.796285152435303,
+ "step": 1327
+ },
+ {
+ "epoch": 18.4471615720524,
+ "grad_norm": 0.014178436249494553,
+ "learning_rate": 0.0006,
+ "loss": 4.773000717163086,
+ "step": 1328
+ },
+ {
+ "epoch": 18.46113537117904,
+ "grad_norm": 0.015295447781682014,
+ "learning_rate": 0.0006,
+ "loss": 4.785943031311035,
+ "step": 1329
+ },
+ {
+ "epoch": 18.475109170305675,
+ "grad_norm": 0.014845133759081364,
+ "learning_rate": 0.0006,
+ "loss": 4.831053733825684,
+ "step": 1330
+ },
+ {
+ "epoch": 18.489082969432314,
+ "grad_norm": 0.015798132866621017,
+ "learning_rate": 0.0006,
+ "loss": 4.87584114074707,
+ "step": 1331
+ },
+ {
+ "epoch": 18.503056768558952,
+ "grad_norm": 0.017797362059354782,
+ "learning_rate": 0.0006,
+ "loss": 4.769850254058838,
+ "step": 1332
+ },
+ {
+ "epoch": 18.51703056768559,
+ "grad_norm": 0.01941198669373989,
+ "learning_rate": 0.0006,
+ "loss": 4.71127986907959,
+ "step": 1333
+ },
+ {
+ "epoch": 18.531004366812226,
+ "grad_norm": 0.01806589961051941,
+ "learning_rate": 0.0006,
+ "loss": 4.718900203704834,
+ "step": 1334
+ },
+ {
+ "epoch": 18.544978165938865,
+ "grad_norm": 0.01684599742293358,
+ "learning_rate": 0.0006,
+ "loss": 4.77104377746582,
+ "step": 1335
+ },
+ {
+ "epoch": 18.558951965065503,
+ "grad_norm": 0.017405332997441292,
+ "learning_rate": 0.0006,
+ "loss": 4.751084327697754,
+ "step": 1336
+ },
+ {
+ "epoch": 18.57292576419214,
+ "grad_norm": 0.017482969909906387,
+ "learning_rate": 0.0006,
+ "loss": 4.771002769470215,
+ "step": 1337
+ },
+ {
+ "epoch": 18.586899563318777,
+ "grad_norm": 0.018441764637827873,
+ "learning_rate": 0.0006,
+ "loss": 4.77370548248291,
+ "step": 1338
+ },
+ {
+ "epoch": 18.600873362445416,
+ "grad_norm": 0.019556425511837006,
+ "learning_rate": 0.0006,
+ "loss": 4.791923999786377,
+ "step": 1339
+ },
+ {
+ "epoch": 18.61484716157205,
+ "grad_norm": 0.022014815360307693,
+ "learning_rate": 0.0006,
+ "loss": 4.7738189697265625,
+ "step": 1340
+ },
+ {
+ "epoch": 18.62882096069869,
+ "grad_norm": 0.0217057466506958,
+ "learning_rate": 0.0006,
+ "loss": 4.738826274871826,
+ "step": 1341
+ },
+ {
+ "epoch": 18.64279475982533,
+ "grad_norm": 0.01779630407691002,
+ "learning_rate": 0.0006,
+ "loss": 4.771679401397705,
+ "step": 1342
+ },
+ {
+ "epoch": 18.656768558951963,
+ "grad_norm": 0.017392629757523537,
+ "learning_rate": 0.0006,
+ "loss": 4.796988487243652,
+ "step": 1343
+ },
+ {
+ "epoch": 18.670742358078602,
+ "grad_norm": 0.016696074977517128,
+ "learning_rate": 0.0006,
+ "loss": 4.720436096191406,
+ "step": 1344
+ },
+ {
+ "epoch": 18.68471615720524,
+ "grad_norm": 0.01689007878303528,
+ "learning_rate": 0.0006,
+ "loss": 4.798285961151123,
+ "step": 1345
+ },
+ {
+ "epoch": 18.69868995633188,
+ "grad_norm": 0.018068408593535423,
+ "learning_rate": 0.0006,
+ "loss": 4.764993190765381,
+ "step": 1346
+ },
+ {
+ "epoch": 18.712663755458514,
+ "grad_norm": 0.01882725954055786,
+ "learning_rate": 0.0006,
+ "loss": 4.751399040222168,
+ "step": 1347
+ },
+ {
+ "epoch": 18.726637554585153,
+ "grad_norm": 0.01765439659357071,
+ "learning_rate": 0.0006,
+ "loss": 4.8288493156433105,
+ "step": 1348
+ },
+ {
+ "epoch": 18.74061135371179,
+ "grad_norm": 0.018233707174658775,
+ "learning_rate": 0.0006,
+ "loss": 4.762392997741699,
+ "step": 1349
+ },
+ {
+ "epoch": 18.754585152838427,
+ "grad_norm": 0.019315950572490692,
+ "learning_rate": 0.0006,
+ "loss": 4.775703430175781,
+ "step": 1350
+ },
+ {
+ "epoch": 18.768558951965066,
+ "grad_norm": 0.01821810007095337,
+ "learning_rate": 0.0006,
+ "loss": 4.8006181716918945,
+ "step": 1351
+ },
+ {
+ "epoch": 18.782532751091704,
+ "grad_norm": 0.018199540674686432,
+ "learning_rate": 0.0006,
+ "loss": 4.774545192718506,
+ "step": 1352
+ },
+ {
+ "epoch": 18.79650655021834,
+ "grad_norm": 0.019126253202557564,
+ "learning_rate": 0.0006,
+ "loss": 4.796529769897461,
+ "step": 1353
+ },
+ {
+ "epoch": 18.810480349344978,
+ "grad_norm": 0.019279757514595985,
+ "learning_rate": 0.0006,
+ "loss": 4.7565107345581055,
+ "step": 1354
+ },
+ {
+ "epoch": 18.824454148471617,
+ "grad_norm": 0.017493106424808502,
+ "learning_rate": 0.0006,
+ "loss": 4.74436616897583,
+ "step": 1355
+ },
+ {
+ "epoch": 18.83842794759825,
+ "grad_norm": 0.01618535816669464,
+ "learning_rate": 0.0006,
+ "loss": 4.8202362060546875,
+ "step": 1356
+ },
+ {
+ "epoch": 18.85240174672489,
+ "grad_norm": 0.014561207965016365,
+ "learning_rate": 0.0006,
+ "loss": 4.8702802658081055,
+ "step": 1357
+ },
+ {
+ "epoch": 18.86637554585153,
+ "grad_norm": 0.014859725721180439,
+ "learning_rate": 0.0006,
+ "loss": 4.740710735321045,
+ "step": 1358
+ },
+ {
+ "epoch": 18.880349344978168,
+ "grad_norm": 0.014770046807825565,
+ "learning_rate": 0.0006,
+ "loss": 4.726039409637451,
+ "step": 1359
+ },
+ {
+ "epoch": 18.894323144104803,
+ "grad_norm": 0.01728186383843422,
+ "learning_rate": 0.0006,
+ "loss": 4.751114368438721,
+ "step": 1360
+ },
+ {
+ "epoch": 18.90829694323144,
+ "grad_norm": 0.016778577119112015,
+ "learning_rate": 0.0006,
+ "loss": 4.841277122497559,
+ "step": 1361
+ },
+ {
+ "epoch": 18.92227074235808,
+ "grad_norm": 0.017361296340823174,
+ "learning_rate": 0.0006,
+ "loss": 4.753442764282227,
+ "step": 1362
+ },
+ {
+ "epoch": 18.936244541484715,
+ "grad_norm": 0.020100396126508713,
+ "learning_rate": 0.0006,
+ "loss": 4.657037734985352,
+ "step": 1363
+ },
+ {
+ "epoch": 18.950218340611354,
+ "grad_norm": 0.024198055267333984,
+ "learning_rate": 0.0006,
+ "loss": 4.868254661560059,
+ "step": 1364
+ },
+ {
+ "epoch": 18.964192139737992,
+ "grad_norm": 0.024044830352067947,
+ "learning_rate": 0.0006,
+ "loss": 4.760613918304443,
+ "step": 1365
+ },
+ {
+ "epoch": 18.978165938864628,
+ "grad_norm": 0.02042238786816597,
+ "learning_rate": 0.0006,
+ "loss": 4.729440212249756,
+ "step": 1366
+ },
+ {
+ "epoch": 18.992139737991266,
+ "grad_norm": 0.02100338228046894,
+ "learning_rate": 0.0006,
+ "loss": 4.751463413238525,
+ "step": 1367
+ },
+ {
+ "epoch": 19.0,
+ "grad_norm": 0.020707951858639717,
+ "learning_rate": 0.0006,
+ "loss": 4.8115129470825195,
+ "step": 1368
+ },
+ {
+ "epoch": 19.0,
+ "eval_loss": 5.2392497062683105,
+ "eval_runtime": 61.5697,
+ "eval_samples_per_second": 39.662,
+ "eval_steps_per_second": 1.251,
+ "step": 1368
+ },
+ {
+ "epoch": 19.01397379912664,
+ "grad_norm": 0.020968902856111526,
+ "learning_rate": 0.0006,
+ "loss": 4.749300479888916,
+ "step": 1369
+ },
+ {
+ "epoch": 19.027947598253274,
+ "grad_norm": 0.019791511818766594,
+ "learning_rate": 0.0006,
+ "loss": 4.752042770385742,
+ "step": 1370
+ },
+ {
+ "epoch": 19.041921397379912,
+ "grad_norm": 0.01768038608133793,
+ "learning_rate": 0.0006,
+ "loss": 4.804799556732178,
+ "step": 1371
+ },
+ {
+ "epoch": 19.05589519650655,
+ "grad_norm": 0.015827463939785957,
+ "learning_rate": 0.0006,
+ "loss": 4.746969699859619,
+ "step": 1372
+ },
+ {
+ "epoch": 19.069868995633186,
+ "grad_norm": 0.01784376986324787,
+ "learning_rate": 0.0006,
+ "loss": 4.709347724914551,
+ "step": 1373
+ },
+ {
+ "epoch": 19.083842794759825,
+ "grad_norm": 0.017368916422128677,
+ "learning_rate": 0.0006,
+ "loss": 4.701427459716797,
+ "step": 1374
+ },
+ {
+ "epoch": 19.097816593886463,
+ "grad_norm": 0.017079368233680725,
+ "learning_rate": 0.0006,
+ "loss": 4.740375518798828,
+ "step": 1375
+ },
+ {
+ "epoch": 19.111790393013102,
+ "grad_norm": 0.01675686053931713,
+ "learning_rate": 0.0006,
+ "loss": 4.7559404373168945,
+ "step": 1376
+ },
+ {
+ "epoch": 19.125764192139737,
+ "grad_norm": 0.016627594828605652,
+ "learning_rate": 0.0006,
+ "loss": 4.686899185180664,
+ "step": 1377
+ },
+ {
+ "epoch": 19.139737991266376,
+ "grad_norm": 0.01687782071530819,
+ "learning_rate": 0.0006,
+ "loss": 4.694512367248535,
+ "step": 1378
+ },
+ {
+ "epoch": 19.153711790393015,
+ "grad_norm": 0.017982905730605125,
+ "learning_rate": 0.0006,
+ "loss": 4.771855354309082,
+ "step": 1379
+ },
+ {
+ "epoch": 19.16768558951965,
+ "grad_norm": 0.018535513430833817,
+ "learning_rate": 0.0006,
+ "loss": 4.67527437210083,
+ "step": 1380
+ },
+ {
+ "epoch": 19.18165938864629,
+ "grad_norm": 0.018792208284139633,
+ "learning_rate": 0.0006,
+ "loss": 4.776930332183838,
+ "step": 1381
+ },
+ {
+ "epoch": 19.195633187772927,
+ "grad_norm": 0.019609590992331505,
+ "learning_rate": 0.0006,
+ "loss": 4.740622520446777,
+ "step": 1382
+ },
+ {
+ "epoch": 19.209606986899562,
+ "grad_norm": 0.020893512293696404,
+ "learning_rate": 0.0006,
+ "loss": 4.761295795440674,
+ "step": 1383
+ },
+ {
+ "epoch": 19.2235807860262,
+ "grad_norm": 0.018166005611419678,
+ "learning_rate": 0.0006,
+ "loss": 4.760861396789551,
+ "step": 1384
+ },
+ {
+ "epoch": 19.23755458515284,
+ "grad_norm": 0.016187280416488647,
+ "learning_rate": 0.0006,
+ "loss": 4.709270477294922,
+ "step": 1385
+ },
+ {
+ "epoch": 19.251528384279474,
+ "grad_norm": 0.014516904018819332,
+ "learning_rate": 0.0006,
+ "loss": 4.706027984619141,
+ "step": 1386
+ },
+ {
+ "epoch": 19.265502183406113,
+ "grad_norm": 0.014672615565359592,
+ "learning_rate": 0.0006,
+ "loss": 4.764942169189453,
+ "step": 1387
+ },
+ {
+ "epoch": 19.27947598253275,
+ "grad_norm": 0.015825409442186356,
+ "learning_rate": 0.0006,
+ "loss": 4.862574100494385,
+ "step": 1388
+ },
+ {
+ "epoch": 19.29344978165939,
+ "grad_norm": 0.01423126831650734,
+ "learning_rate": 0.0006,
+ "loss": 4.721527576446533,
+ "step": 1389
+ },
+ {
+ "epoch": 19.307423580786025,
+ "grad_norm": 0.015230957418680191,
+ "learning_rate": 0.0006,
+ "loss": 4.789156913757324,
+ "step": 1390
+ },
+ {
+ "epoch": 19.321397379912664,
+ "grad_norm": 0.014533719047904015,
+ "learning_rate": 0.0006,
+ "loss": 4.727848052978516,
+ "step": 1391
+ },
+ {
+ "epoch": 19.335371179039303,
+ "grad_norm": 0.015545960515737534,
+ "learning_rate": 0.0006,
+ "loss": 4.67042350769043,
+ "step": 1392
+ },
+ {
+ "epoch": 19.349344978165938,
+ "grad_norm": 0.015026269480586052,
+ "learning_rate": 0.0006,
+ "loss": 4.8094048500061035,
+ "step": 1393
+ },
+ {
+ "epoch": 19.363318777292577,
+ "grad_norm": 0.016231181100010872,
+ "learning_rate": 0.0006,
+ "loss": 4.743175983428955,
+ "step": 1394
+ },
+ {
+ "epoch": 19.377292576419215,
+ "grad_norm": 0.0172134917229414,
+ "learning_rate": 0.0006,
+ "loss": 4.633797645568848,
+ "step": 1395
+ },
+ {
+ "epoch": 19.39126637554585,
+ "grad_norm": 0.016263330355286598,
+ "learning_rate": 0.0006,
+ "loss": 4.642855644226074,
+ "step": 1396
+ },
+ {
+ "epoch": 19.40524017467249,
+ "grad_norm": 0.014997457154095173,
+ "learning_rate": 0.0006,
+ "loss": 4.6688408851623535,
+ "step": 1397
+ },
+ {
+ "epoch": 19.419213973799128,
+ "grad_norm": 0.015129785053431988,
+ "learning_rate": 0.0006,
+ "loss": 4.673038482666016,
+ "step": 1398
+ },
+ {
+ "epoch": 19.433187772925763,
+ "grad_norm": 0.017703495919704437,
+ "learning_rate": 0.0006,
+ "loss": 4.699586868286133,
+ "step": 1399
+ },
+ {
+ "epoch": 19.4471615720524,
+ "grad_norm": 0.017693769186735153,
+ "learning_rate": 0.0006,
+ "loss": 4.741053104400635,
+ "step": 1400
+ },
+ {
+ "epoch": 19.46113537117904,
+ "grad_norm": 0.018453532829880714,
+ "learning_rate": 0.0006,
+ "loss": 4.696603775024414,
+ "step": 1401
+ },
+ {
+ "epoch": 19.475109170305675,
+ "grad_norm": 0.018508777022361755,
+ "learning_rate": 0.0006,
+ "loss": 4.652768135070801,
+ "step": 1402
+ },
+ {
+ "epoch": 19.489082969432314,
+ "grad_norm": 0.016347555443644524,
+ "learning_rate": 0.0006,
+ "loss": 4.696932792663574,
+ "step": 1403
+ },
+ {
+ "epoch": 19.503056768558952,
+ "grad_norm": 0.016488511115312576,
+ "learning_rate": 0.0006,
+ "loss": 4.668985843658447,
+ "step": 1404
+ },
+ {
+ "epoch": 19.51703056768559,
+ "grad_norm": 0.016894331201910973,
+ "learning_rate": 0.0006,
+ "loss": 4.712052345275879,
+ "step": 1405
+ },
+ {
+ "epoch": 19.531004366812226,
+ "grad_norm": 0.01606992818415165,
+ "learning_rate": 0.0006,
+ "loss": 4.809332847595215,
+ "step": 1406
+ },
+ {
+ "epoch": 19.544978165938865,
+ "grad_norm": 0.01614546775817871,
+ "learning_rate": 0.0006,
+ "loss": 4.718560218811035,
+ "step": 1407
+ },
+ {
+ "epoch": 19.558951965065503,
+ "grad_norm": 0.015790050849318504,
+ "learning_rate": 0.0006,
+ "loss": 4.7086992263793945,
+ "step": 1408
+ },
+ {
+ "epoch": 19.57292576419214,
+ "grad_norm": 0.015570463612675667,
+ "learning_rate": 0.0006,
+ "loss": 4.701510906219482,
+ "step": 1409
+ },
+ {
+ "epoch": 19.586899563318777,
+ "grad_norm": 0.015313868410885334,
+ "learning_rate": 0.0006,
+ "loss": 4.6717400550842285,
+ "step": 1410
+ },
+ {
+ "epoch": 19.600873362445416,
+ "grad_norm": 0.016266439110040665,
+ "learning_rate": 0.0006,
+ "loss": 4.708248138427734,
+ "step": 1411
+ },
+ {
+ "epoch": 19.61484716157205,
+ "grad_norm": 0.01676693931221962,
+ "learning_rate": 0.0006,
+ "loss": 4.708803176879883,
+ "step": 1412
+ },
+ {
+ "epoch": 19.62882096069869,
+ "grad_norm": 0.016014447435736656,
+ "learning_rate": 0.0006,
+ "loss": 4.7899603843688965,
+ "step": 1413
+ },
+ {
+ "epoch": 19.64279475982533,
+ "grad_norm": 0.01589064858853817,
+ "learning_rate": 0.0006,
+ "loss": 4.683867454528809,
+ "step": 1414
+ },
+ {
+ "epoch": 19.656768558951963,
+ "grad_norm": 0.01659962348639965,
+ "learning_rate": 0.0006,
+ "loss": 4.789435863494873,
+ "step": 1415
+ },
+ {
+ "epoch": 19.670742358078602,
+ "grad_norm": 0.01895272172987461,
+ "learning_rate": 0.0006,
+ "loss": 4.753860950469971,
+ "step": 1416
+ },
+ {
+ "epoch": 19.68471615720524,
+ "grad_norm": 0.019914697855710983,
+ "learning_rate": 0.0006,
+ "loss": 4.701515197753906,
+ "step": 1417
+ },
+ {
+ "epoch": 19.69868995633188,
+ "grad_norm": 0.021606286987662315,
+ "learning_rate": 0.0006,
+ "loss": 4.754340171813965,
+ "step": 1418
+ },
+ {
+ "epoch": 19.712663755458514,
+ "grad_norm": 0.02217685803771019,
+ "learning_rate": 0.0006,
+ "loss": 4.73709774017334,
+ "step": 1419
+ },
+ {
+ "epoch": 19.726637554585153,
+ "grad_norm": 0.020211679860949516,
+ "learning_rate": 0.0006,
+ "loss": 4.6904473304748535,
+ "step": 1420
+ },
+ {
+ "epoch": 19.74061135371179,
+ "grad_norm": 0.020696299150586128,
+ "learning_rate": 0.0006,
+ "loss": 4.718563079833984,
+ "step": 1421
+ },
+ {
+ "epoch": 19.754585152838427,
+ "grad_norm": 0.02091088518500328,
+ "learning_rate": 0.0006,
+ "loss": 4.803199291229248,
+ "step": 1422
+ },
+ {
+ "epoch": 19.768558951965066,
+ "grad_norm": 0.020736388862133026,
+ "learning_rate": 0.0006,
+ "loss": 4.735616207122803,
+ "step": 1423
+ },
+ {
+ "epoch": 19.782532751091704,
+ "grad_norm": 0.018370863050222397,
+ "learning_rate": 0.0006,
+ "loss": 4.6523332595825195,
+ "step": 1424
+ },
+ {
+ "epoch": 19.79650655021834,
+ "grad_norm": 0.01750843971967697,
+ "learning_rate": 0.0006,
+ "loss": 4.715566635131836,
+ "step": 1425
+ },
+ {
+ "epoch": 19.810480349344978,
+ "grad_norm": 0.018078286200761795,
+ "learning_rate": 0.0006,
+ "loss": 4.734353065490723,
+ "step": 1426
+ },
+ {
+ "epoch": 19.824454148471617,
+ "grad_norm": 0.018023716285824776,
+ "learning_rate": 0.0006,
+ "loss": 4.652746200561523,
+ "step": 1427
+ },
+ {
+ "epoch": 19.83842794759825,
+ "grad_norm": 0.018733404576778412,
+ "learning_rate": 0.0006,
+ "loss": 4.7124176025390625,
+ "step": 1428
+ },
+ {
+ "epoch": 19.85240174672489,
+ "grad_norm": 0.018829114735126495,
+ "learning_rate": 0.0006,
+ "loss": 4.720004081726074,
+ "step": 1429
+ },
+ {
+ "epoch": 19.86637554585153,
+ "grad_norm": 0.019517114385962486,
+ "learning_rate": 0.0006,
+ "loss": 4.7428741455078125,
+ "step": 1430
+ },
+ {
+ "epoch": 19.880349344978168,
+ "grad_norm": 0.01994675025343895,
+ "learning_rate": 0.0006,
+ "loss": 4.717165946960449,
+ "step": 1431
+ },
+ {
+ "epoch": 19.894323144104803,
+ "grad_norm": 0.01924729160964489,
+ "learning_rate": 0.0006,
+ "loss": 4.749120712280273,
+ "step": 1432
+ },
+ {
+ "epoch": 19.90829694323144,
+ "grad_norm": 0.018079813569784164,
+ "learning_rate": 0.0006,
+ "loss": 4.794156551361084,
+ "step": 1433
+ },
+ {
+ "epoch": 19.92227074235808,
+ "grad_norm": 0.017745722085237503,
+ "learning_rate": 0.0006,
+ "loss": 4.7152910232543945,
+ "step": 1434
+ },
+ {
+ "epoch": 19.936244541484715,
+ "grad_norm": 0.020233964547514915,
+ "learning_rate": 0.0006,
+ "loss": 4.7771477699279785,
+ "step": 1435
+ },
+ {
+ "epoch": 19.950218340611354,
+ "grad_norm": 0.020702997222542763,
+ "learning_rate": 0.0006,
+ "loss": 4.7343339920043945,
+ "step": 1436
+ },
+ {
+ "epoch": 19.964192139737992,
+ "grad_norm": 0.021028192713856697,
+ "learning_rate": 0.0006,
+ "loss": 4.7011847496032715,
+ "step": 1437
+ },
+ {
+ "epoch": 19.978165938864628,
+ "grad_norm": 0.018680263310670853,
+ "learning_rate": 0.0006,
+ "loss": 4.796342849731445,
+ "step": 1438
+ },
+ {
+ "epoch": 19.992139737991266,
+ "grad_norm": 0.017404882237315178,
+ "learning_rate": 0.0006,
+ "loss": 4.678404808044434,
+ "step": 1439
+ },
+ {
+ "epoch": 20.0,
+ "grad_norm": 0.017616212368011475,
+ "learning_rate": 0.0006,
+ "loss": 4.792126655578613,
+ "step": 1440
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 28800,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 400,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": true,
+ "should_log": false,
+ "should_save": false,
+ "should_training_stop": false
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 6.139226722664448e+17,
+ "train_batch_size": 8,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1440/training_args.bin b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1440/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..c799bcf4051249aabf4c62fff9d38801df7bc2b5
--- /dev/null
+++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1440/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:40fb25263250c08c3479a6f0ac9fdf1f196e2a3b08a4eb10384786ff9b1903ef
+size 4792
diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-288/chat_template.jinja b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-288/chat_template.jinja
new file mode 100644
index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230
--- /dev/null
+++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-288/chat_template.jinja
@@ -0,0 +1,85 @@
+{%- if tools %}
+ {{- '<|im_start|>system\n' }}
+ {%- if messages[0].role == 'system' %}
+ {{- messages[0].content + '\n\n' }}
+ {%- endif %}
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }}
+ {%- for tool in tools %}
+ {{- "\n" }}
+ {{- tool | tojson }}
+ {%- endfor %}
+ {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }}
+{%- else %}
+ {%- if messages[0].role == 'system' %}
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
+ {%- endif %}
+{%- endif %}
+{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
+{%- for message in messages[::-1] %}
+ {%- set index = (messages|length - 1) - loop.index0 %}
+ {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %}
+ {%- set ns.multi_step_tool = false %}
+ {%- set ns.last_query_index = index %}
+ {%- endif %}
+{%- endfor %}
+{%- for message in messages %}
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
+ {%- elif message.role == "assistant" %}
+ {%- set content = message.content %}
+ {%- set reasoning_content = '' %}
+ {%- if message.reasoning_content is defined and message.reasoning_content is not none %}
+ {%- set reasoning_content = message.reasoning_content %}
+ {%- else %}
+ {%- if '' in message.content %}
+ {%- set content = message.content.split('')[-1].lstrip('\n') %}
+ {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %}
+ {%- endif %}
+ {%- endif %}
+ {%- if loop.index0 > ns.last_query_index %}
+ {%- if loop.last or (not loop.last and reasoning_content) %}
+ {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- if message.tool_calls %}
+ {%- for tool_call in message.tool_calls %}
+ {%- if (loop.first and content) or (not loop.first) %}
+ {{- '\n' }}
+ {%- endif %}
+ {%- if tool_call.function %}
+ {%- set tool_call = tool_call.function %}
+ {%- endif %}
+ {{- '\n{"name": "' }}
+ {{- tool_call.name }}
+ {{- '", "arguments": ' }}
+ {%- if tool_call.arguments is string %}
+ {{- tool_call.arguments }}
+ {%- else %}
+ {{- tool_call.arguments | tojson }}
+ {%- endif %}
+ {{- '}\n' }}
+ {%- endfor %}
+ {%- endif %}
+ {{- '<|im_end|>\n' }}
+ {%- elif message.role == "tool" %}
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
+ {{- '<|im_start|>user' }}
+ {%- endif %}
+ {{- '\n\n' }}
+ {{- message.content }}
+ {{- '\n' }}
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
+ {{- '<|im_end|>\n' }}
+ {%- endif %}
+ {%- endif %}
+{%- endfor %}
+{%- if add_generation_prompt %}
+ {{- '<|im_start|>assistant\n' }}
+ {%- if enable_thinking is defined and enable_thinking is false %}
+ {{- '\n\n\n\n' }}
+ {%- endif %}
+{%- endif %}
\ No newline at end of file
diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-288/config.json b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-288/config.json
new file mode 100644
index 0000000000000000000000000000000000000000..175ee3431793018f23f74824f69554b21afcbd4f
--- /dev/null
+++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-288/config.json
@@ -0,0 +1,32 @@
+{
+ "architectures": [
+ "LlamaForCausalLM"
+ ],
+ "attention_bias": false,
+ "attention_dropout": 0.0,
+ "bos_token_id": null,
+ "dtype": "float32",
+ "eos_token_id": 151645,
+ "head_dim": 128,
+ "hidden_act": "silu",
+ "hidden_size": 512,
+ "initializer_range": 0.02,
+ "intermediate_size": 1536,
+ "max_position_embeddings": 2048,
+ "mlp_bias": false,
+ "model_type": "llama",
+ "num_attention_heads": 4,
+ "num_hidden_layers": 20,
+ "num_key_value_heads": 4,
+ "pad_token_id": 151645,
+ "pretraining_tp": 1,
+ "rms_norm_eps": 1e-06,
+ "rope_parameters": {
+ "rope_theta": 10000.0,
+ "rope_type": "default"
+ },
+ "tie_word_embeddings": true,
+ "transformers_version": "5.5.0",
+ "use_cache": false,
+ "vocab_size": 151674
+}
diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-288/generation_config.json b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-288/generation_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3
--- /dev/null
+++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-288/generation_config.json
@@ -0,0 +1,11 @@
+{
+ "_from_model_config": true,
+ "eos_token_id": [
+ 151645
+ ],
+ "output_attentions": false,
+ "output_hidden_states": false,
+ "pad_token_id": 151645,
+ "transformers_version": "5.5.0",
+ "use_cache": true
+}
diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-288/model.safetensors b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-288/model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..1aed3be13b8db54755855a33bd4e4da31a650c8b
--- /dev/null
+++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-288/model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:ecb01abdbbdb51cecfa2cf09d771518c1906251227edb20fdbfe400a9ddf08fe
+size 583362376
diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-288/optimizer.pt b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-288/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..5aebb2357fb3cc1180e6bd78ef51e9c2157663ec
--- /dev/null
+++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-288/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:945d4f54df19bbbb716cfcb2dd5351ce119b256582670b18b3f3bf825f213d2b
+size 1166837626
diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-288/rng_state_0.pth b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-288/rng_state_0.pth
new file mode 100644
index 0000000000000000000000000000000000000000..81d2bd9decff40322863d3c4253a6f90b727f3ce
--- /dev/null
+++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-288/rng_state_0.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:f123a34234139b64f0d7a1a8066d6c6e69ecc4d10c733547345f9db86126081c
+size 15024
diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-288/rng_state_1.pth b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-288/rng_state_1.pth
new file mode 100644
index 0000000000000000000000000000000000000000..51368e431549a274df92d5a7d67ffef092809d98
--- /dev/null
+++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-288/rng_state_1.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:f1e2b3f9a2385ece5e2513f9f80c9c776fe9b56f608574d5f628c540850f3d4d
+size 15024
diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-288/rng_state_2.pth b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-288/rng_state_2.pth
new file mode 100644
index 0000000000000000000000000000000000000000..77d0825438ee7f93a226a8383ff7b82b66e7d59a
--- /dev/null
+++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-288/rng_state_2.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:775f48d24f2da0edd9623b436cc53a297a5f382317102f871a0af9b07b45a983
+size 15024
diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-288/rng_state_3.pth b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-288/rng_state_3.pth
new file mode 100644
index 0000000000000000000000000000000000000000..6b101452eb1e045f4259258fdac0f1336b38a5f5
--- /dev/null
+++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-288/rng_state_3.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:b255c26ded3d367afe0151ee7ddf729f4b29ddc8696bcfb44058cf024acb072b
+size 15024
diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-288/scheduler.pt b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-288/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..e7f2e4bf5abc860ae45f2d5dbda8115542dd3611
--- /dev/null
+++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-288/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:4d1eab1d6f570939dc1e4b9d10383b7354ef244100fb2b34ff0ac8b58642b63c
+size 1064
diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-288/tokenizer.json b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-288/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..9ea28d9eb325c36bcab4c1a08fc93fd598929f08
--- /dev/null
+++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-288/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:60e99075d371ecbb9f1dde97363f612d45bf0bf88f1a2006d9bfcf1939333c14
+size 11424648
diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-288/tokenizer_config.json b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-288/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..cadd997fc26040d23ad6677fcb39d8d18b01f55e
--- /dev/null
+++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-288/tokenizer_config.json
@@ -0,0 +1,21 @@
+{
+ "add_prefix_space": false,
+ "backend": "tokenizers",
+ "bos_token": null,
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "<|im_end|>",
+ "errors": "replace",
+ "extra_special_tokens": [
+ "<|l2r_pred|>",
+ "<|r2l_pred|>",
+ "<|next_1_pred|>",
+ "<|next_2_pred|>",
+ "<|mask|>"
+ ],
+ "is_local": false,
+ "model_max_length": 131072,
+ "pad_token": "<|im_end|>",
+ "split_special_tokens": false,
+ "tokenizer_class": "Qwen2Tokenizer",
+ "unk_token": null
+}
diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-288/trainer_state.json b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-288/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..c539125b4bac0682d1fbedc1a03928bdb3624066
--- /dev/null
+++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-288/trainer_state.json
@@ -0,0 +1,2074 @@
+{
+ "best_global_step": null,
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 4.0,
+ "eval_steps": 500,
+ "global_step": 288,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.013973799126637555,
+ "grad_norm": 0.5768774151802063,
+ "learning_rate": 0.0,
+ "loss": 12.050328254699707,
+ "step": 1
+ },
+ {
+ "epoch": 0.02794759825327511,
+ "grad_norm": 0.5784570574760437,
+ "learning_rate": 5.999999999999999e-06,
+ "loss": 12.049251556396484,
+ "step": 2
+ },
+ {
+ "epoch": 0.04192139737991266,
+ "grad_norm": 0.5607187151908875,
+ "learning_rate": 1.1999999999999999e-05,
+ "loss": 11.921792030334473,
+ "step": 3
+ },
+ {
+ "epoch": 0.05589519650655022,
+ "grad_norm": 0.46466919779777527,
+ "learning_rate": 1.7999999999999997e-05,
+ "loss": 11.686628341674805,
+ "step": 4
+ },
+ {
+ "epoch": 0.06986899563318777,
+ "grad_norm": 0.38012945652008057,
+ "learning_rate": 2.3999999999999997e-05,
+ "loss": 11.432823181152344,
+ "step": 5
+ },
+ {
+ "epoch": 0.08384279475982533,
+ "grad_norm": 0.33215776085853577,
+ "learning_rate": 2.9999999999999997e-05,
+ "loss": 11.207542419433594,
+ "step": 6
+ },
+ {
+ "epoch": 0.09781659388646288,
+ "grad_norm": 0.2928776741027832,
+ "learning_rate": 3.5999999999999994e-05,
+ "loss": 11.000097274780273,
+ "step": 7
+ },
+ {
+ "epoch": 0.11179039301310044,
+ "grad_norm": 0.26018962264060974,
+ "learning_rate": 4.2e-05,
+ "loss": 10.825748443603516,
+ "step": 8
+ },
+ {
+ "epoch": 0.125764192139738,
+ "grad_norm": 0.23717083036899567,
+ "learning_rate": 4.7999999999999994e-05,
+ "loss": 10.677122116088867,
+ "step": 9
+ },
+ {
+ "epoch": 0.13973799126637554,
+ "grad_norm": 0.21641355752944946,
+ "learning_rate": 5.399999999999999e-05,
+ "loss": 10.569148063659668,
+ "step": 10
+ },
+ {
+ "epoch": 0.1537117903930131,
+ "grad_norm": 0.1982956826686859,
+ "learning_rate": 5.9999999999999995e-05,
+ "loss": 10.49653434753418,
+ "step": 11
+ },
+ {
+ "epoch": 0.16768558951965065,
+ "grad_norm": 0.18819282948970795,
+ "learning_rate": 6.599999999999999e-05,
+ "loss": 10.432951927185059,
+ "step": 12
+ },
+ {
+ "epoch": 0.18165938864628822,
+ "grad_norm": 0.18322618305683136,
+ "learning_rate": 7.199999999999999e-05,
+ "loss": 10.387115478515625,
+ "step": 13
+ },
+ {
+ "epoch": 0.19563318777292577,
+ "grad_norm": 0.18399226665496826,
+ "learning_rate": 7.8e-05,
+ "loss": 10.33381462097168,
+ "step": 14
+ },
+ {
+ "epoch": 0.2096069868995633,
+ "grad_norm": 0.1833190619945526,
+ "learning_rate": 8.4e-05,
+ "loss": 10.301689147949219,
+ "step": 15
+ },
+ {
+ "epoch": 0.22358078602620088,
+ "grad_norm": 0.18490834534168243,
+ "learning_rate": 8.999999999999999e-05,
+ "loss": 10.240968704223633,
+ "step": 16
+ },
+ {
+ "epoch": 0.23755458515283842,
+ "grad_norm": 0.18358023464679718,
+ "learning_rate": 9.599999999999999e-05,
+ "loss": 10.194717407226562,
+ "step": 17
+ },
+ {
+ "epoch": 0.251528384279476,
+ "grad_norm": 0.1801682710647583,
+ "learning_rate": 0.000102,
+ "loss": 10.14897632598877,
+ "step": 18
+ },
+ {
+ "epoch": 0.26550218340611353,
+ "grad_norm": 0.17899440228939056,
+ "learning_rate": 0.00010799999999999998,
+ "loss": 10.078763961791992,
+ "step": 19
+ },
+ {
+ "epoch": 0.2794759825327511,
+ "grad_norm": 0.17666497826576233,
+ "learning_rate": 0.00011399999999999999,
+ "loss": 10.01095962524414,
+ "step": 20
+ },
+ {
+ "epoch": 0.2934497816593886,
+ "grad_norm": 0.17196154594421387,
+ "learning_rate": 0.00011999999999999999,
+ "loss": 9.951203346252441,
+ "step": 21
+ },
+ {
+ "epoch": 0.3074235807860262,
+ "grad_norm": 0.169203519821167,
+ "learning_rate": 0.00012599999999999997,
+ "loss": 9.871726036071777,
+ "step": 22
+ },
+ {
+ "epoch": 0.32139737991266376,
+ "grad_norm": 0.16500306129455566,
+ "learning_rate": 0.00013199999999999998,
+ "loss": 9.803646087646484,
+ "step": 23
+ },
+ {
+ "epoch": 0.3353711790393013,
+ "grad_norm": 0.16128124296665192,
+ "learning_rate": 0.000138,
+ "loss": 9.728487014770508,
+ "step": 24
+ },
+ {
+ "epoch": 0.34934497816593885,
+ "grad_norm": 0.15904058516025543,
+ "learning_rate": 0.00014399999999999998,
+ "loss": 9.642473220825195,
+ "step": 25
+ },
+ {
+ "epoch": 0.36331877729257644,
+ "grad_norm": 0.15560561418533325,
+ "learning_rate": 0.00015,
+ "loss": 9.56414794921875,
+ "step": 26
+ },
+ {
+ "epoch": 0.377292576419214,
+ "grad_norm": 0.15278612077236176,
+ "learning_rate": 0.000156,
+ "loss": 9.484230041503906,
+ "step": 27
+ },
+ {
+ "epoch": 0.39126637554585153,
+ "grad_norm": 0.15263237059116364,
+ "learning_rate": 0.000162,
+ "loss": 9.383644104003906,
+ "step": 28
+ },
+ {
+ "epoch": 0.4052401746724891,
+ "grad_norm": 0.14945024251937866,
+ "learning_rate": 0.000168,
+ "loss": 9.29840087890625,
+ "step": 29
+ },
+ {
+ "epoch": 0.4192139737991266,
+ "grad_norm": 0.14614646136760712,
+ "learning_rate": 0.00017399999999999997,
+ "loss": 9.208931922912598,
+ "step": 30
+ },
+ {
+ "epoch": 0.4331877729257642,
+ "grad_norm": 0.14295098185539246,
+ "learning_rate": 0.00017999999999999998,
+ "loss": 9.122323989868164,
+ "step": 31
+ },
+ {
+ "epoch": 0.44716157205240176,
+ "grad_norm": 0.13868345320224762,
+ "learning_rate": 0.000186,
+ "loss": 9.03034496307373,
+ "step": 32
+ },
+ {
+ "epoch": 0.4611353711790393,
+ "grad_norm": 0.135064035654068,
+ "learning_rate": 0.00019199999999999998,
+ "loss": 8.92818832397461,
+ "step": 33
+ },
+ {
+ "epoch": 0.47510917030567684,
+ "grad_norm": 0.13097241520881653,
+ "learning_rate": 0.000198,
+ "loss": 8.834627151489258,
+ "step": 34
+ },
+ {
+ "epoch": 0.4890829694323144,
+ "grad_norm": 0.12478776276111603,
+ "learning_rate": 0.000204,
+ "loss": 8.755172729492188,
+ "step": 35
+ },
+ {
+ "epoch": 0.503056768558952,
+ "grad_norm": 0.1214982345700264,
+ "learning_rate": 0.00020999999999999998,
+ "loss": 8.647891998291016,
+ "step": 36
+ },
+ {
+ "epoch": 0.5170305676855895,
+ "grad_norm": 0.11958328634500504,
+ "learning_rate": 0.00021599999999999996,
+ "loss": 8.543274879455566,
+ "step": 37
+ },
+ {
+ "epoch": 0.5310043668122271,
+ "grad_norm": 0.11680915951728821,
+ "learning_rate": 0.00022199999999999998,
+ "loss": 8.44540023803711,
+ "step": 38
+ },
+ {
+ "epoch": 0.5449781659388646,
+ "grad_norm": 0.11491474509239197,
+ "learning_rate": 0.00022799999999999999,
+ "loss": 8.349921226501465,
+ "step": 39
+ },
+ {
+ "epoch": 0.5589519650655022,
+ "grad_norm": 0.11160755902528763,
+ "learning_rate": 0.000234,
+ "loss": 8.26960563659668,
+ "step": 40
+ },
+ {
+ "epoch": 0.5729257641921397,
+ "grad_norm": 0.10840637236833572,
+ "learning_rate": 0.00023999999999999998,
+ "loss": 8.192136764526367,
+ "step": 41
+ },
+ {
+ "epoch": 0.5868995633187772,
+ "grad_norm": 0.10585298389196396,
+ "learning_rate": 0.00024599999999999996,
+ "loss": 8.09546947479248,
+ "step": 42
+ },
+ {
+ "epoch": 0.6008733624454149,
+ "grad_norm": 0.1032542735338211,
+ "learning_rate": 0.00025199999999999995,
+ "loss": 7.991090774536133,
+ "step": 43
+ },
+ {
+ "epoch": 0.6148471615720524,
+ "grad_norm": 0.0986943170428276,
+ "learning_rate": 0.000258,
+ "loss": 7.92384147644043,
+ "step": 44
+ },
+ {
+ "epoch": 0.62882096069869,
+ "grad_norm": 0.09320762753486633,
+ "learning_rate": 0.00026399999999999997,
+ "loss": 7.8478193283081055,
+ "step": 45
+ },
+ {
+ "epoch": 0.6427947598253275,
+ "grad_norm": 0.09168332070112228,
+ "learning_rate": 0.00027,
+ "loss": 7.76185417175293,
+ "step": 46
+ },
+ {
+ "epoch": 0.6567685589519651,
+ "grad_norm": 0.08832395076751709,
+ "learning_rate": 0.000276,
+ "loss": 7.679399490356445,
+ "step": 47
+ },
+ {
+ "epoch": 0.6707423580786026,
+ "grad_norm": 0.08816400915384293,
+ "learning_rate": 0.00028199999999999997,
+ "loss": 7.5789642333984375,
+ "step": 48
+ },
+ {
+ "epoch": 0.6847161572052402,
+ "grad_norm": 0.0861717164516449,
+ "learning_rate": 0.00028799999999999995,
+ "loss": 7.519214630126953,
+ "step": 49
+ },
+ {
+ "epoch": 0.6986899563318777,
+ "grad_norm": 0.08103746175765991,
+ "learning_rate": 0.000294,
+ "loss": 7.4475321769714355,
+ "step": 50
+ },
+ {
+ "epoch": 0.7126637554585152,
+ "grad_norm": 0.07545104622840881,
+ "learning_rate": 0.0003,
+ "loss": 7.401182174682617,
+ "step": 51
+ },
+ {
+ "epoch": 0.7266375545851529,
+ "grad_norm": 0.07042575627565384,
+ "learning_rate": 0.00030599999999999996,
+ "loss": 7.340173244476318,
+ "step": 52
+ },
+ {
+ "epoch": 0.7406113537117904,
+ "grad_norm": 0.06737572699785233,
+ "learning_rate": 0.000312,
+ "loss": 7.265697002410889,
+ "step": 53
+ },
+ {
+ "epoch": 0.754585152838428,
+ "grad_norm": 0.06151146441698074,
+ "learning_rate": 0.000318,
+ "loss": 7.231061935424805,
+ "step": 54
+ },
+ {
+ "epoch": 0.7685589519650655,
+ "grad_norm": 0.05638071522116661,
+ "learning_rate": 0.000324,
+ "loss": 7.195466995239258,
+ "step": 55
+ },
+ {
+ "epoch": 0.7825327510917031,
+ "grad_norm": 0.05084463208913803,
+ "learning_rate": 0.00033,
+ "loss": 7.137603759765625,
+ "step": 56
+ },
+ {
+ "epoch": 0.7965065502183406,
+ "grad_norm": 0.04293261095881462,
+ "learning_rate": 0.000336,
+ "loss": 7.115458965301514,
+ "step": 57
+ },
+ {
+ "epoch": 0.8104803493449781,
+ "grad_norm": 0.036064863204956055,
+ "learning_rate": 0.00034199999999999996,
+ "loss": 7.089038848876953,
+ "step": 58
+ },
+ {
+ "epoch": 0.8244541484716157,
+ "grad_norm": 0.029026754200458527,
+ "learning_rate": 0.00034799999999999995,
+ "loss": 7.0584306716918945,
+ "step": 59
+ },
+ {
+ "epoch": 0.8384279475982532,
+ "grad_norm": 0.02286517433822155,
+ "learning_rate": 0.00035399999999999993,
+ "loss": 7.049450874328613,
+ "step": 60
+ },
+ {
+ "epoch": 0.8524017467248908,
+ "grad_norm": 0.018878433853387833,
+ "learning_rate": 0.00035999999999999997,
+ "loss": 7.008824348449707,
+ "step": 61
+ },
+ {
+ "epoch": 0.8663755458515284,
+ "grad_norm": 0.010773813351988792,
+ "learning_rate": 0.00036599999999999995,
+ "loss": 7.024829864501953,
+ "step": 62
+ },
+ {
+ "epoch": 0.880349344978166,
+ "grad_norm": 0.008794732391834259,
+ "learning_rate": 0.000372,
+ "loss": 7.018399238586426,
+ "step": 63
+ },
+ {
+ "epoch": 0.8943231441048035,
+ "grad_norm": 0.011709796264767647,
+ "learning_rate": 0.00037799999999999997,
+ "loss": 7.0101752281188965,
+ "step": 64
+ },
+ {
+ "epoch": 0.9082969432314411,
+ "grad_norm": 0.012123258784413338,
+ "learning_rate": 0.00038399999999999996,
+ "loss": 7.024317741394043,
+ "step": 65
+ },
+ {
+ "epoch": 0.9222707423580786,
+ "grad_norm": 0.01530320756137371,
+ "learning_rate": 0.00039,
+ "loss": 7.027738571166992,
+ "step": 66
+ },
+ {
+ "epoch": 0.9362445414847161,
+ "grad_norm": 0.01890002004802227,
+ "learning_rate": 0.000396,
+ "loss": 7.024099349975586,
+ "step": 67
+ },
+ {
+ "epoch": 0.9502183406113537,
+ "grad_norm": 0.020684150978922844,
+ "learning_rate": 0.000402,
+ "loss": 7.007943153381348,
+ "step": 68
+ },
+ {
+ "epoch": 0.9641921397379912,
+ "grad_norm": 0.019826196134090424,
+ "learning_rate": 0.000408,
+ "loss": 7.006874084472656,
+ "step": 69
+ },
+ {
+ "epoch": 0.9781659388646288,
+ "grad_norm": 0.019218673929572105,
+ "learning_rate": 0.0004139999999999999,
+ "loss": 7.002897262573242,
+ "step": 70
+ },
+ {
+ "epoch": 0.9921397379912664,
+ "grad_norm": 0.013856411911547184,
+ "learning_rate": 0.00041999999999999996,
+ "loss": 6.993999481201172,
+ "step": 71
+ },
+ {
+ "epoch": 1.0,
+ "grad_norm": 0.010390003211796284,
+ "learning_rate": 0.00042599999999999995,
+ "loss": 7.000683784484863,
+ "step": 72
+ },
+ {
+ "epoch": 1.0,
+ "eval_loss": 7.91140079498291,
+ "eval_runtime": 57.7925,
+ "eval_samples_per_second": 42.255,
+ "eval_steps_per_second": 1.332,
+ "step": 72
+ },
+ {
+ "epoch": 1.0139737991266375,
+ "grad_norm": 0.013802163302898407,
+ "learning_rate": 0.00043199999999999993,
+ "loss": 6.998257637023926,
+ "step": 73
+ },
+ {
+ "epoch": 1.027947598253275,
+ "grad_norm": 0.010109687224030495,
+ "learning_rate": 0.00043799999999999997,
+ "loss": 7.002542018890381,
+ "step": 74
+ },
+ {
+ "epoch": 1.0419213973799126,
+ "grad_norm": 0.008019139990210533,
+ "learning_rate": 0.00044399999999999995,
+ "loss": 6.985941410064697,
+ "step": 75
+ },
+ {
+ "epoch": 1.0558951965065502,
+ "grad_norm": 0.0069672465324401855,
+ "learning_rate": 0.00045,
+ "loss": 6.99245023727417,
+ "step": 76
+ },
+ {
+ "epoch": 1.0698689956331877,
+ "grad_norm": 0.00714076729491353,
+ "learning_rate": 0.00045599999999999997,
+ "loss": 6.959296226501465,
+ "step": 77
+ },
+ {
+ "epoch": 1.0838427947598253,
+ "grad_norm": 0.008408435620367527,
+ "learning_rate": 0.00046199999999999995,
+ "loss": 6.954504489898682,
+ "step": 78
+ },
+ {
+ "epoch": 1.0978165938864628,
+ "grad_norm": 0.008778786286711693,
+ "learning_rate": 0.000468,
+ "loss": 7.0007004737854,
+ "step": 79
+ },
+ {
+ "epoch": 1.1117903930131003,
+ "grad_norm": 0.007753407116979361,
+ "learning_rate": 0.000474,
+ "loss": 6.98183536529541,
+ "step": 80
+ },
+ {
+ "epoch": 1.125764192139738,
+ "grad_norm": 0.01087260153144598,
+ "learning_rate": 0.00047999999999999996,
+ "loss": 6.991066932678223,
+ "step": 81
+ },
+ {
+ "epoch": 1.1397379912663754,
+ "grad_norm": 0.006122548598796129,
+ "learning_rate": 0.000486,
+ "loss": 6.978824138641357,
+ "step": 82
+ },
+ {
+ "epoch": 1.1537117903930132,
+ "grad_norm": 0.0067383465357124805,
+ "learning_rate": 0.0004919999999999999,
+ "loss": 6.961703777313232,
+ "step": 83
+ },
+ {
+ "epoch": 1.1676855895196507,
+ "grad_norm": 0.007922791875898838,
+ "learning_rate": 0.000498,
+ "loss": 6.954860687255859,
+ "step": 84
+ },
+ {
+ "epoch": 1.1816593886462883,
+ "grad_norm": 0.00818922370672226,
+ "learning_rate": 0.0005039999999999999,
+ "loss": 6.945961952209473,
+ "step": 85
+ },
+ {
+ "epoch": 1.1956331877729258,
+ "grad_norm": 0.006791461259126663,
+ "learning_rate": 0.0005099999999999999,
+ "loss": 6.968680381774902,
+ "step": 86
+ },
+ {
+ "epoch": 1.2096069868995634,
+ "grad_norm": 0.005394163075834513,
+ "learning_rate": 0.000516,
+ "loss": 6.956404685974121,
+ "step": 87
+ },
+ {
+ "epoch": 1.223580786026201,
+ "grad_norm": 0.004500220064073801,
+ "learning_rate": 0.000522,
+ "loss": 6.96998405456543,
+ "step": 88
+ },
+ {
+ "epoch": 1.2375545851528384,
+ "grad_norm": 0.008699574507772923,
+ "learning_rate": 0.0005279999999999999,
+ "loss": 6.955585479736328,
+ "step": 89
+ },
+ {
+ "epoch": 1.251528384279476,
+ "grad_norm": 0.005889365915209055,
+ "learning_rate": 0.000534,
+ "loss": 6.9708051681518555,
+ "step": 90
+ },
+ {
+ "epoch": 1.2655021834061135,
+ "grad_norm": 0.009646810591220856,
+ "learning_rate": 0.00054,
+ "loss": 6.962608814239502,
+ "step": 91
+ },
+ {
+ "epoch": 1.279475982532751,
+ "grad_norm": 0.004517171997576952,
+ "learning_rate": 0.0005459999999999999,
+ "loss": 6.957627296447754,
+ "step": 92
+ },
+ {
+ "epoch": 1.2934497816593886,
+ "grad_norm": 0.006339677143841982,
+ "learning_rate": 0.000552,
+ "loss": 6.963141441345215,
+ "step": 93
+ },
+ {
+ "epoch": 1.3074235807860262,
+ "grad_norm": 0.005387383047491312,
+ "learning_rate": 0.000558,
+ "loss": 6.934465408325195,
+ "step": 94
+ },
+ {
+ "epoch": 1.3213973799126637,
+ "grad_norm": 0.005252422299236059,
+ "learning_rate": 0.0005639999999999999,
+ "loss": 6.956108093261719,
+ "step": 95
+ },
+ {
+ "epoch": 1.3353711790393012,
+ "grad_norm": 0.008699414320290089,
+ "learning_rate": 0.00057,
+ "loss": 6.973369598388672,
+ "step": 96
+ },
+ {
+ "epoch": 1.3493449781659388,
+ "grad_norm": 0.003947409335523844,
+ "learning_rate": 0.0005759999999999999,
+ "loss": 6.950225830078125,
+ "step": 97
+ },
+ {
+ "epoch": 1.3633187772925766,
+ "grad_norm": 0.00768681475892663,
+ "learning_rate": 0.0005819999999999999,
+ "loss": 6.964739799499512,
+ "step": 98
+ },
+ {
+ "epoch": 1.3772925764192139,
+ "grad_norm": 0.006563829258084297,
+ "learning_rate": 0.000588,
+ "loss": 6.972421169281006,
+ "step": 99
+ },
+ {
+ "epoch": 1.3912663755458516,
+ "grad_norm": 0.007606555242091417,
+ "learning_rate": 0.0005939999999999999,
+ "loss": 6.958213806152344,
+ "step": 100
+ },
+ {
+ "epoch": 1.405240174672489,
+ "grad_norm": 0.006250880658626556,
+ "learning_rate": 0.0006,
+ "loss": 6.962296485900879,
+ "step": 101
+ },
+ {
+ "epoch": 1.4192139737991267,
+ "grad_norm": 0.0073033408261835575,
+ "learning_rate": 0.0006,
+ "loss": 6.969705104827881,
+ "step": 102
+ },
+ {
+ "epoch": 1.4331877729257643,
+ "grad_norm": 0.0054510910995304585,
+ "learning_rate": 0.0006,
+ "loss": 6.9652862548828125,
+ "step": 103
+ },
+ {
+ "epoch": 1.4471615720524018,
+ "grad_norm": 0.004895673133432865,
+ "learning_rate": 0.0006,
+ "loss": 6.956023693084717,
+ "step": 104
+ },
+ {
+ "epoch": 1.4611353711790394,
+ "grad_norm": 0.005306515377014875,
+ "learning_rate": 0.0006,
+ "loss": 6.963231086730957,
+ "step": 105
+ },
+ {
+ "epoch": 1.475109170305677,
+ "grad_norm": 0.004742146469652653,
+ "learning_rate": 0.0006,
+ "loss": 6.9582061767578125,
+ "step": 106
+ },
+ {
+ "epoch": 1.4890829694323144,
+ "grad_norm": 0.004220000468194485,
+ "learning_rate": 0.0006,
+ "loss": 6.972961902618408,
+ "step": 107
+ },
+ {
+ "epoch": 1.503056768558952,
+ "grad_norm": 0.0042153168469667435,
+ "learning_rate": 0.0006,
+ "loss": 6.961165904998779,
+ "step": 108
+ },
+ {
+ "epoch": 1.5170305676855895,
+ "grad_norm": 0.005165040958672762,
+ "learning_rate": 0.0006,
+ "loss": 6.95018196105957,
+ "step": 109
+ },
+ {
+ "epoch": 1.531004366812227,
+ "grad_norm": 0.003930266480892897,
+ "learning_rate": 0.0006,
+ "loss": 6.9651689529418945,
+ "step": 110
+ },
+ {
+ "epoch": 1.5449781659388646,
+ "grad_norm": 0.004465071018785238,
+ "learning_rate": 0.0006,
+ "loss": 6.960553169250488,
+ "step": 111
+ },
+ {
+ "epoch": 1.5589519650655022,
+ "grad_norm": 0.004888612311333418,
+ "learning_rate": 0.0006,
+ "loss": 6.967258930206299,
+ "step": 112
+ },
+ {
+ "epoch": 1.5729257641921397,
+ "grad_norm": 0.007497863378375769,
+ "learning_rate": 0.0006,
+ "loss": 6.954545021057129,
+ "step": 113
+ },
+ {
+ "epoch": 1.5868995633187772,
+ "grad_norm": 0.005659648682922125,
+ "learning_rate": 0.0006,
+ "loss": 6.962299823760986,
+ "step": 114
+ },
+ {
+ "epoch": 1.600873362445415,
+ "grad_norm": 0.006641396787017584,
+ "learning_rate": 0.0006,
+ "loss": 6.952983856201172,
+ "step": 115
+ },
+ {
+ "epoch": 1.6148471615720523,
+ "grad_norm": 0.009715928696095943,
+ "learning_rate": 0.0006,
+ "loss": 6.958622932434082,
+ "step": 116
+ },
+ {
+ "epoch": 1.62882096069869,
+ "grad_norm": 0.012789523229002953,
+ "learning_rate": 0.0006,
+ "loss": 6.932043075561523,
+ "step": 117
+ },
+ {
+ "epoch": 1.6427947598253274,
+ "grad_norm": 0.021673237904906273,
+ "learning_rate": 0.0006,
+ "loss": 6.931290149688721,
+ "step": 118
+ },
+ {
+ "epoch": 1.6567685589519652,
+ "grad_norm": 0.16024991869926453,
+ "learning_rate": 0.0006,
+ "loss": 6.979497909545898,
+ "step": 119
+ },
+ {
+ "epoch": 1.6707423580786025,
+ "grad_norm": 0.2434523105621338,
+ "learning_rate": 0.0006,
+ "loss": 7.185218811035156,
+ "step": 120
+ },
+ {
+ "epoch": 1.6847161572052403,
+ "grad_norm": 0.39252421259880066,
+ "learning_rate": 0.0006,
+ "loss": 7.137226581573486,
+ "step": 121
+ },
+ {
+ "epoch": 1.6986899563318776,
+ "grad_norm": 0.10565312206745148,
+ "learning_rate": 0.0006,
+ "loss": 6.984494209289551,
+ "step": 122
+ },
+ {
+ "epoch": 1.7126637554585153,
+ "grad_norm": 0.1033431813120842,
+ "learning_rate": 0.0006,
+ "loss": 7.001287937164307,
+ "step": 123
+ },
+ {
+ "epoch": 1.726637554585153,
+ "grad_norm": 0.14875490963459015,
+ "learning_rate": 0.0006,
+ "loss": 7.051436901092529,
+ "step": 124
+ },
+ {
+ "epoch": 1.7406113537117904,
+ "grad_norm": 0.1374562531709671,
+ "learning_rate": 0.0006,
+ "loss": 7.0293121337890625,
+ "step": 125
+ },
+ {
+ "epoch": 1.754585152838428,
+ "grad_norm": 0.0910211056470871,
+ "learning_rate": 0.0006,
+ "loss": 7.007458686828613,
+ "step": 126
+ },
+ {
+ "epoch": 1.7685589519650655,
+ "grad_norm": 0.021633487194776535,
+ "learning_rate": 0.0006,
+ "loss": 6.996816635131836,
+ "step": 127
+ },
+ {
+ "epoch": 1.782532751091703,
+ "grad_norm": 0.08692754060029984,
+ "learning_rate": 0.0006,
+ "loss": 6.998541831970215,
+ "step": 128
+ },
+ {
+ "epoch": 1.7965065502183406,
+ "grad_norm": 0.12613292038440704,
+ "learning_rate": 0.0006,
+ "loss": 6.9970703125,
+ "step": 129
+ },
+ {
+ "epoch": 1.8104803493449781,
+ "grad_norm": 0.09462810307741165,
+ "learning_rate": 0.0006,
+ "loss": 6.997045040130615,
+ "step": 130
+ },
+ {
+ "epoch": 1.8244541484716157,
+ "grad_norm": 0.03534962609410286,
+ "learning_rate": 0.0006,
+ "loss": 6.9738450050354,
+ "step": 131
+ },
+ {
+ "epoch": 1.8384279475982532,
+ "grad_norm": 0.033839885145425797,
+ "learning_rate": 0.0006,
+ "loss": 6.946001052856445,
+ "step": 132
+ },
+ {
+ "epoch": 1.8524017467248908,
+ "grad_norm": 0.06263269484043121,
+ "learning_rate": 0.0006,
+ "loss": 6.968981742858887,
+ "step": 133
+ },
+ {
+ "epoch": 1.8663755458515285,
+ "grad_norm": 0.07117997109889984,
+ "learning_rate": 0.0006,
+ "loss": 6.981535911560059,
+ "step": 134
+ },
+ {
+ "epoch": 1.8803493449781659,
+ "grad_norm": 0.06336277723312378,
+ "learning_rate": 0.0006,
+ "loss": 6.976946830749512,
+ "step": 135
+ },
+ {
+ "epoch": 1.8943231441048036,
+ "grad_norm": 0.047467734664678574,
+ "learning_rate": 0.0006,
+ "loss": 6.961673736572266,
+ "step": 136
+ },
+ {
+ "epoch": 1.908296943231441,
+ "grad_norm": 0.018787868320941925,
+ "learning_rate": 0.0006,
+ "loss": 6.948919296264648,
+ "step": 137
+ },
+ {
+ "epoch": 1.9222707423580787,
+ "grad_norm": 0.025909408926963806,
+ "learning_rate": 0.0006,
+ "loss": 6.976190567016602,
+ "step": 138
+ },
+ {
+ "epoch": 1.936244541484716,
+ "grad_norm": 0.052842333912849426,
+ "learning_rate": 0.0006,
+ "loss": 6.971207618713379,
+ "step": 139
+ },
+ {
+ "epoch": 1.9502183406113538,
+ "grad_norm": 0.06421888619661331,
+ "learning_rate": 0.0006,
+ "loss": 6.979109764099121,
+ "step": 140
+ },
+ {
+ "epoch": 1.9641921397379911,
+ "grad_norm": 0.055596910417079926,
+ "learning_rate": 0.0006,
+ "loss": 6.972118854522705,
+ "step": 141
+ },
+ {
+ "epoch": 1.9781659388646289,
+ "grad_norm": 0.03267091140151024,
+ "learning_rate": 0.0006,
+ "loss": 6.947260856628418,
+ "step": 142
+ },
+ {
+ "epoch": 1.9921397379912664,
+ "grad_norm": 0.008032613433897495,
+ "learning_rate": 0.0006,
+ "loss": 6.936816215515137,
+ "step": 143
+ },
+ {
+ "epoch": 2.0,
+ "grad_norm": 0.020184025168418884,
+ "learning_rate": 0.0006,
+ "loss": 6.929848670959473,
+ "step": 144
+ },
+ {
+ "epoch": 2.0,
+ "eval_loss": 7.836523056030273,
+ "eval_runtime": 58.5622,
+ "eval_samples_per_second": 41.699,
+ "eval_steps_per_second": 1.315,
+ "step": 144
+ },
+ {
+ "epoch": 2.0139737991266378,
+ "grad_norm": 0.031980376690626144,
+ "learning_rate": 0.0006,
+ "loss": 6.968081474304199,
+ "step": 145
+ },
+ {
+ "epoch": 2.027947598253275,
+ "grad_norm": 0.03802533075213432,
+ "learning_rate": 0.0006,
+ "loss": 6.95728874206543,
+ "step": 146
+ },
+ {
+ "epoch": 2.041921397379913,
+ "grad_norm": 0.03796226158738136,
+ "learning_rate": 0.0006,
+ "loss": 6.96516752243042,
+ "step": 147
+ },
+ {
+ "epoch": 2.05589519650655,
+ "grad_norm": 0.02621951326727867,
+ "learning_rate": 0.0006,
+ "loss": 6.956058502197266,
+ "step": 148
+ },
+ {
+ "epoch": 2.069868995633188,
+ "grad_norm": 0.013292718678712845,
+ "learning_rate": 0.0006,
+ "loss": 6.9783935546875,
+ "step": 149
+ },
+ {
+ "epoch": 2.0838427947598253,
+ "grad_norm": 0.009446968324482441,
+ "learning_rate": 0.0006,
+ "loss": 6.96873664855957,
+ "step": 150
+ },
+ {
+ "epoch": 2.097816593886463,
+ "grad_norm": 0.021703092381358147,
+ "learning_rate": 0.0006,
+ "loss": 6.959234237670898,
+ "step": 151
+ },
+ {
+ "epoch": 2.1117903930131003,
+ "grad_norm": 0.027765290811657906,
+ "learning_rate": 0.0006,
+ "loss": 6.968524932861328,
+ "step": 152
+ },
+ {
+ "epoch": 2.125764192139738,
+ "grad_norm": 0.028761165216565132,
+ "learning_rate": 0.0006,
+ "loss": 6.954620838165283,
+ "step": 153
+ },
+ {
+ "epoch": 2.1397379912663754,
+ "grad_norm": 0.019729383289813995,
+ "learning_rate": 0.0006,
+ "loss": 6.949009895324707,
+ "step": 154
+ },
+ {
+ "epoch": 2.153711790393013,
+ "grad_norm": 0.008990089409053326,
+ "learning_rate": 0.0006,
+ "loss": 6.956226348876953,
+ "step": 155
+ },
+ {
+ "epoch": 2.1676855895196505,
+ "grad_norm": 0.00549644511193037,
+ "learning_rate": 0.0006,
+ "loss": 6.957143783569336,
+ "step": 156
+ },
+ {
+ "epoch": 2.1816593886462883,
+ "grad_norm": 0.014436670579016209,
+ "learning_rate": 0.0006,
+ "loss": 6.954257011413574,
+ "step": 157
+ },
+ {
+ "epoch": 2.1956331877729256,
+ "grad_norm": 0.01990494504570961,
+ "learning_rate": 0.0006,
+ "loss": 6.942131519317627,
+ "step": 158
+ },
+ {
+ "epoch": 2.2096069868995634,
+ "grad_norm": 0.019155776128172874,
+ "learning_rate": 0.0006,
+ "loss": 6.946969985961914,
+ "step": 159
+ },
+ {
+ "epoch": 2.2235807860262007,
+ "grad_norm": 0.014970783144235611,
+ "learning_rate": 0.0006,
+ "loss": 6.957938194274902,
+ "step": 160
+ },
+ {
+ "epoch": 2.2375545851528384,
+ "grad_norm": 0.009875659830868244,
+ "learning_rate": 0.0006,
+ "loss": 6.953309059143066,
+ "step": 161
+ },
+ {
+ "epoch": 2.251528384279476,
+ "grad_norm": 0.004792087245732546,
+ "learning_rate": 0.0006,
+ "loss": 6.975146770477295,
+ "step": 162
+ },
+ {
+ "epoch": 2.2655021834061135,
+ "grad_norm": 0.009936527349054813,
+ "learning_rate": 0.0006,
+ "loss": 6.961674690246582,
+ "step": 163
+ },
+ {
+ "epoch": 2.279475982532751,
+ "grad_norm": 0.015100127086043358,
+ "learning_rate": 0.0006,
+ "loss": 6.968759536743164,
+ "step": 164
+ },
+ {
+ "epoch": 2.2934497816593886,
+ "grad_norm": 0.01449542585760355,
+ "learning_rate": 0.0006,
+ "loss": 6.966510772705078,
+ "step": 165
+ },
+ {
+ "epoch": 2.3074235807860264,
+ "grad_norm": 0.01073537115007639,
+ "learning_rate": 0.0006,
+ "loss": 6.936373710632324,
+ "step": 166
+ },
+ {
+ "epoch": 2.3213973799126637,
+ "grad_norm": 0.006766272243112326,
+ "learning_rate": 0.0006,
+ "loss": 6.96397590637207,
+ "step": 167
+ },
+ {
+ "epoch": 2.3353711790393015,
+ "grad_norm": 0.005055439658463001,
+ "learning_rate": 0.0006,
+ "loss": 6.961050987243652,
+ "step": 168
+ },
+ {
+ "epoch": 2.349344978165939,
+ "grad_norm": 0.006252622697502375,
+ "learning_rate": 0.0006,
+ "loss": 6.980506420135498,
+ "step": 169
+ },
+ {
+ "epoch": 2.3633187772925766,
+ "grad_norm": 0.008084346540272236,
+ "learning_rate": 0.0006,
+ "loss": 6.9346604347229,
+ "step": 170
+ },
+ {
+ "epoch": 2.377292576419214,
+ "grad_norm": 0.010260380804538727,
+ "learning_rate": 0.0006,
+ "loss": 6.964118957519531,
+ "step": 171
+ },
+ {
+ "epoch": 2.3912663755458516,
+ "grad_norm": 0.010697875171899796,
+ "learning_rate": 0.0006,
+ "loss": 6.941572189331055,
+ "step": 172
+ },
+ {
+ "epoch": 2.405240174672489,
+ "grad_norm": 0.007338830269873142,
+ "learning_rate": 0.0006,
+ "loss": 6.950471878051758,
+ "step": 173
+ },
+ {
+ "epoch": 2.4192139737991267,
+ "grad_norm": 0.005931832827627659,
+ "learning_rate": 0.0006,
+ "loss": 6.962935924530029,
+ "step": 174
+ },
+ {
+ "epoch": 2.433187772925764,
+ "grad_norm": 0.0061243255622684956,
+ "learning_rate": 0.0006,
+ "loss": 6.946418285369873,
+ "step": 175
+ },
+ {
+ "epoch": 2.447161572052402,
+ "grad_norm": 0.007983328774571419,
+ "learning_rate": 0.0006,
+ "loss": 6.959758758544922,
+ "step": 176
+ },
+ {
+ "epoch": 2.461135371179039,
+ "grad_norm": 0.010470672510564327,
+ "learning_rate": 0.0006,
+ "loss": 6.93851375579834,
+ "step": 177
+ },
+ {
+ "epoch": 2.475109170305677,
+ "grad_norm": 0.011147388257086277,
+ "learning_rate": 0.0006,
+ "loss": 6.93258810043335,
+ "step": 178
+ },
+ {
+ "epoch": 2.489082969432314,
+ "grad_norm": 0.01058544497936964,
+ "learning_rate": 0.0006,
+ "loss": 6.932259559631348,
+ "step": 179
+ },
+ {
+ "epoch": 2.503056768558952,
+ "grad_norm": 0.016850745305418968,
+ "learning_rate": 0.0006,
+ "loss": 6.924542427062988,
+ "step": 180
+ },
+ {
+ "epoch": 2.5170305676855893,
+ "grad_norm": 0.023101547732949257,
+ "learning_rate": 0.0006,
+ "loss": 6.914878845214844,
+ "step": 181
+ },
+ {
+ "epoch": 2.531004366812227,
+ "grad_norm": 0.015795735642313957,
+ "learning_rate": 0.0006,
+ "loss": 6.922432899475098,
+ "step": 182
+ },
+ {
+ "epoch": 2.544978165938865,
+ "grad_norm": 0.008821426890790462,
+ "learning_rate": 0.0006,
+ "loss": 6.924562454223633,
+ "step": 183
+ },
+ {
+ "epoch": 2.558951965065502,
+ "grad_norm": 0.012502548284828663,
+ "learning_rate": 0.0006,
+ "loss": 6.914718151092529,
+ "step": 184
+ },
+ {
+ "epoch": 2.5729257641921395,
+ "grad_norm": 0.011545066721737385,
+ "learning_rate": 0.0006,
+ "loss": 6.915705680847168,
+ "step": 185
+ },
+ {
+ "epoch": 2.5868995633187772,
+ "grad_norm": 0.022190893068909645,
+ "learning_rate": 0.0006,
+ "loss": 6.88851261138916,
+ "step": 186
+ },
+ {
+ "epoch": 2.600873362445415,
+ "grad_norm": 0.023395851254463196,
+ "learning_rate": 0.0006,
+ "loss": 6.877655029296875,
+ "step": 187
+ },
+ {
+ "epoch": 2.6148471615720523,
+ "grad_norm": 0.015601156279444695,
+ "learning_rate": 0.0006,
+ "loss": 6.889644145965576,
+ "step": 188
+ },
+ {
+ "epoch": 2.62882096069869,
+ "grad_norm": 0.01362934336066246,
+ "learning_rate": 0.0006,
+ "loss": 6.891547203063965,
+ "step": 189
+ },
+ {
+ "epoch": 2.6427947598253274,
+ "grad_norm": 0.012305490672588348,
+ "learning_rate": 0.0006,
+ "loss": 6.890357971191406,
+ "step": 190
+ },
+ {
+ "epoch": 2.656768558951965,
+ "grad_norm": 0.01623358018696308,
+ "learning_rate": 0.0006,
+ "loss": 6.849443435668945,
+ "step": 191
+ },
+ {
+ "epoch": 2.6707423580786025,
+ "grad_norm": 0.018063291907310486,
+ "learning_rate": 0.0006,
+ "loss": 6.842372894287109,
+ "step": 192
+ },
+ {
+ "epoch": 2.6847161572052403,
+ "grad_norm": 0.010179434902966022,
+ "learning_rate": 0.0006,
+ "loss": 6.870055198669434,
+ "step": 193
+ },
+ {
+ "epoch": 2.6986899563318776,
+ "grad_norm": 0.009986722841858864,
+ "learning_rate": 0.0006,
+ "loss": 6.859061241149902,
+ "step": 194
+ },
+ {
+ "epoch": 2.7126637554585153,
+ "grad_norm": 0.01117654051631689,
+ "learning_rate": 0.0006,
+ "loss": 6.861199378967285,
+ "step": 195
+ },
+ {
+ "epoch": 2.726637554585153,
+ "grad_norm": 0.008551654405891895,
+ "learning_rate": 0.0006,
+ "loss": 6.850255012512207,
+ "step": 196
+ },
+ {
+ "epoch": 2.7406113537117904,
+ "grad_norm": 0.018904326483607292,
+ "learning_rate": 0.0006,
+ "loss": 6.848639965057373,
+ "step": 197
+ },
+ {
+ "epoch": 2.7545851528384278,
+ "grad_norm": 0.02080724574625492,
+ "learning_rate": 0.0006,
+ "loss": 6.80456018447876,
+ "step": 198
+ },
+ {
+ "epoch": 2.7685589519650655,
+ "grad_norm": 0.012010873295366764,
+ "learning_rate": 0.0006,
+ "loss": 6.848775863647461,
+ "step": 199
+ },
+ {
+ "epoch": 2.7825327510917033,
+ "grad_norm": 0.008113056421279907,
+ "learning_rate": 0.0006,
+ "loss": 6.849387168884277,
+ "step": 200
+ },
+ {
+ "epoch": 2.7965065502183406,
+ "grad_norm": 0.012912592850625515,
+ "learning_rate": 0.0006,
+ "loss": 6.832240104675293,
+ "step": 201
+ },
+ {
+ "epoch": 2.810480349344978,
+ "grad_norm": 0.01146865077316761,
+ "learning_rate": 0.0006,
+ "loss": 6.850383281707764,
+ "step": 202
+ },
+ {
+ "epoch": 2.8244541484716157,
+ "grad_norm": 0.0171892661601305,
+ "learning_rate": 0.0006,
+ "loss": 6.839577674865723,
+ "step": 203
+ },
+ {
+ "epoch": 2.8384279475982535,
+ "grad_norm": 0.012083170004189014,
+ "learning_rate": 0.0006,
+ "loss": 6.818614482879639,
+ "step": 204
+ },
+ {
+ "epoch": 2.8524017467248908,
+ "grad_norm": 0.0062555656768381596,
+ "learning_rate": 0.0006,
+ "loss": 6.831291198730469,
+ "step": 205
+ },
+ {
+ "epoch": 2.8663755458515285,
+ "grad_norm": 0.010102466680109501,
+ "learning_rate": 0.0006,
+ "loss": 6.846522808074951,
+ "step": 206
+ },
+ {
+ "epoch": 2.880349344978166,
+ "grad_norm": 0.009042087942361832,
+ "learning_rate": 0.0006,
+ "loss": 6.822072982788086,
+ "step": 207
+ },
+ {
+ "epoch": 2.8943231441048036,
+ "grad_norm": 0.011133556254208088,
+ "learning_rate": 0.0006,
+ "loss": 6.811252593994141,
+ "step": 208
+ },
+ {
+ "epoch": 2.908296943231441,
+ "grad_norm": 0.00990577507764101,
+ "learning_rate": 0.0006,
+ "loss": 6.814435958862305,
+ "step": 209
+ },
+ {
+ "epoch": 2.9222707423580787,
+ "grad_norm": 0.006605073343962431,
+ "learning_rate": 0.0006,
+ "loss": 6.831300735473633,
+ "step": 210
+ },
+ {
+ "epoch": 2.936244541484716,
+ "grad_norm": 0.007831175811588764,
+ "learning_rate": 0.0006,
+ "loss": 6.841738700866699,
+ "step": 211
+ },
+ {
+ "epoch": 2.950218340611354,
+ "grad_norm": 0.005536255426704884,
+ "learning_rate": 0.0006,
+ "loss": 6.8043670654296875,
+ "step": 212
+ },
+ {
+ "epoch": 2.964192139737991,
+ "grad_norm": 0.006532667670398951,
+ "learning_rate": 0.0006,
+ "loss": 6.8091840744018555,
+ "step": 213
+ },
+ {
+ "epoch": 2.978165938864629,
+ "grad_norm": 0.006674565840512514,
+ "learning_rate": 0.0006,
+ "loss": 6.811327934265137,
+ "step": 214
+ },
+ {
+ "epoch": 2.992139737991266,
+ "grad_norm": 0.01150344219058752,
+ "learning_rate": 0.0006,
+ "loss": 6.810762405395508,
+ "step": 215
+ },
+ {
+ "epoch": 3.0,
+ "grad_norm": 0.009015528485178947,
+ "learning_rate": 0.0006,
+ "loss": 6.836431980133057,
+ "step": 216
+ },
+ {
+ "epoch": 3.0,
+ "eval_loss": 7.718832969665527,
+ "eval_runtime": 58.6615,
+ "eval_samples_per_second": 41.629,
+ "eval_steps_per_second": 1.313,
+ "step": 216
+ },
+ {
+ "epoch": 3.0139737991266378,
+ "grad_norm": 0.009791829623281956,
+ "learning_rate": 0.0006,
+ "loss": 6.8102593421936035,
+ "step": 217
+ },
+ {
+ "epoch": 3.027947598253275,
+ "grad_norm": 0.008970947004854679,
+ "learning_rate": 0.0006,
+ "loss": 6.792663097381592,
+ "step": 218
+ },
+ {
+ "epoch": 3.041921397379913,
+ "grad_norm": 0.013689392246305943,
+ "learning_rate": 0.0006,
+ "loss": 6.815651893615723,
+ "step": 219
+ },
+ {
+ "epoch": 3.05589519650655,
+ "grad_norm": 0.031499531120061874,
+ "learning_rate": 0.0006,
+ "loss": 6.809841156005859,
+ "step": 220
+ },
+ {
+ "epoch": 3.069868995633188,
+ "grad_norm": 0.03203094005584717,
+ "learning_rate": 0.0006,
+ "loss": 6.779417037963867,
+ "step": 221
+ },
+ {
+ "epoch": 3.0838427947598253,
+ "grad_norm": 0.007947996258735657,
+ "learning_rate": 0.0006,
+ "loss": 6.773656845092773,
+ "step": 222
+ },
+ {
+ "epoch": 3.097816593886463,
+ "grad_norm": 0.03152519091963768,
+ "learning_rate": 0.0006,
+ "loss": 6.804083824157715,
+ "step": 223
+ },
+ {
+ "epoch": 3.1117903930131003,
+ "grad_norm": 0.014937258325517178,
+ "learning_rate": 0.0006,
+ "loss": 6.792213439941406,
+ "step": 224
+ },
+ {
+ "epoch": 3.125764192139738,
+ "grad_norm": 0.020961705595254898,
+ "learning_rate": 0.0006,
+ "loss": 6.8092803955078125,
+ "step": 225
+ },
+ {
+ "epoch": 3.1397379912663754,
+ "grad_norm": 0.009152988903224468,
+ "learning_rate": 0.0006,
+ "loss": 6.7737603187561035,
+ "step": 226
+ },
+ {
+ "epoch": 3.153711790393013,
+ "grad_norm": 0.02770620957016945,
+ "learning_rate": 0.0006,
+ "loss": 6.767014503479004,
+ "step": 227
+ },
+ {
+ "epoch": 3.1676855895196505,
+ "grad_norm": 0.01165871974080801,
+ "learning_rate": 0.0006,
+ "loss": 6.760884761810303,
+ "step": 228
+ },
+ {
+ "epoch": 3.1816593886462883,
+ "grad_norm": 0.01910376362502575,
+ "learning_rate": 0.0006,
+ "loss": 6.775722980499268,
+ "step": 229
+ },
+ {
+ "epoch": 3.1956331877729256,
+ "grad_norm": 0.010413747280836105,
+ "learning_rate": 0.0006,
+ "loss": 6.779816627502441,
+ "step": 230
+ },
+ {
+ "epoch": 3.2096069868995634,
+ "grad_norm": 0.01918417401611805,
+ "learning_rate": 0.0006,
+ "loss": 6.760381698608398,
+ "step": 231
+ },
+ {
+ "epoch": 3.2235807860262007,
+ "grad_norm": 0.017960334196686745,
+ "learning_rate": 0.0006,
+ "loss": 6.761198997497559,
+ "step": 232
+ },
+ {
+ "epoch": 3.2375545851528384,
+ "grad_norm": 0.016546795144677162,
+ "learning_rate": 0.0006,
+ "loss": 6.742448806762695,
+ "step": 233
+ },
+ {
+ "epoch": 3.251528384279476,
+ "grad_norm": 0.01889597810804844,
+ "learning_rate": 0.0006,
+ "loss": 6.758237838745117,
+ "step": 234
+ },
+ {
+ "epoch": 3.2655021834061135,
+ "grad_norm": 0.009611992165446281,
+ "learning_rate": 0.0006,
+ "loss": 6.75614070892334,
+ "step": 235
+ },
+ {
+ "epoch": 3.279475982532751,
+ "grad_norm": 0.02034934051334858,
+ "learning_rate": 0.0006,
+ "loss": 6.736965179443359,
+ "step": 236
+ },
+ {
+ "epoch": 3.2934497816593886,
+ "grad_norm": 0.010740065947175026,
+ "learning_rate": 0.0006,
+ "loss": 6.746947288513184,
+ "step": 237
+ },
+ {
+ "epoch": 3.3074235807860264,
+ "grad_norm": 0.018383340910077095,
+ "learning_rate": 0.0006,
+ "loss": 6.747332572937012,
+ "step": 238
+ },
+ {
+ "epoch": 3.3213973799126637,
+ "grad_norm": 0.012362911365926266,
+ "learning_rate": 0.0006,
+ "loss": 6.767669200897217,
+ "step": 239
+ },
+ {
+ "epoch": 3.3353711790393015,
+ "grad_norm": 0.01127463299781084,
+ "learning_rate": 0.0006,
+ "loss": 6.769808292388916,
+ "step": 240
+ },
+ {
+ "epoch": 3.349344978165939,
+ "grad_norm": 0.017693769186735153,
+ "learning_rate": 0.0006,
+ "loss": 6.733658790588379,
+ "step": 241
+ },
+ {
+ "epoch": 3.3633187772925766,
+ "grad_norm": 0.006420376244932413,
+ "learning_rate": 0.0006,
+ "loss": 6.74934720993042,
+ "step": 242
+ },
+ {
+ "epoch": 3.377292576419214,
+ "grad_norm": 0.014020364731550217,
+ "learning_rate": 0.0006,
+ "loss": 6.728157043457031,
+ "step": 243
+ },
+ {
+ "epoch": 3.3912663755458516,
+ "grad_norm": 0.008018582127988338,
+ "learning_rate": 0.0006,
+ "loss": 6.730726718902588,
+ "step": 244
+ },
+ {
+ "epoch": 3.405240174672489,
+ "grad_norm": 0.00782222580164671,
+ "learning_rate": 0.0006,
+ "loss": 6.7499895095825195,
+ "step": 245
+ },
+ {
+ "epoch": 3.4192139737991267,
+ "grad_norm": 0.008278409950435162,
+ "learning_rate": 0.0006,
+ "loss": 6.712803363800049,
+ "step": 246
+ },
+ {
+ "epoch": 3.433187772925764,
+ "grad_norm": 0.009826759807765484,
+ "learning_rate": 0.0006,
+ "loss": 6.742262840270996,
+ "step": 247
+ },
+ {
+ "epoch": 3.447161572052402,
+ "grad_norm": 0.009488224983215332,
+ "learning_rate": 0.0006,
+ "loss": 6.739130020141602,
+ "step": 248
+ },
+ {
+ "epoch": 3.461135371179039,
+ "grad_norm": 0.007667903788387775,
+ "learning_rate": 0.0006,
+ "loss": 6.747515678405762,
+ "step": 249
+ },
+ {
+ "epoch": 3.475109170305677,
+ "grad_norm": 0.0073090000078082085,
+ "learning_rate": 0.0006,
+ "loss": 6.740870475769043,
+ "step": 250
+ },
+ {
+ "epoch": 3.489082969432314,
+ "grad_norm": 0.010058708488941193,
+ "learning_rate": 0.0006,
+ "loss": 6.777563095092773,
+ "step": 251
+ },
+ {
+ "epoch": 3.503056768558952,
+ "grad_norm": 0.020676063373684883,
+ "learning_rate": 0.0006,
+ "loss": 6.708603858947754,
+ "step": 252
+ },
+ {
+ "epoch": 3.5170305676855893,
+ "grad_norm": 0.027210544794797897,
+ "learning_rate": 0.0006,
+ "loss": 6.694097995758057,
+ "step": 253
+ },
+ {
+ "epoch": 3.531004366812227,
+ "grad_norm": 0.038053520023822784,
+ "learning_rate": 0.0006,
+ "loss": 6.697017669677734,
+ "step": 254
+ },
+ {
+ "epoch": 3.544978165938865,
+ "grad_norm": 0.04684312269091606,
+ "learning_rate": 0.0006,
+ "loss": 6.729883193969727,
+ "step": 255
+ },
+ {
+ "epoch": 3.558951965065502,
+ "grad_norm": 0.03673060983419418,
+ "learning_rate": 0.0006,
+ "loss": 6.743753433227539,
+ "step": 256
+ },
+ {
+ "epoch": 3.5729257641921395,
+ "grad_norm": 0.006882632151246071,
+ "learning_rate": 0.0006,
+ "loss": 6.71405029296875,
+ "step": 257
+ },
+ {
+ "epoch": 3.5868995633187772,
+ "grad_norm": 0.03588201105594635,
+ "learning_rate": 0.0006,
+ "loss": 6.712467193603516,
+ "step": 258
+ },
+ {
+ "epoch": 3.600873362445415,
+ "grad_norm": 0.029415199533104897,
+ "learning_rate": 0.0006,
+ "loss": 6.715634822845459,
+ "step": 259
+ },
+ {
+ "epoch": 3.6148471615720523,
+ "grad_norm": 0.011425039730966091,
+ "learning_rate": 0.0006,
+ "loss": 6.697334289550781,
+ "step": 260
+ },
+ {
+ "epoch": 3.62882096069869,
+ "grad_norm": 0.0297263003885746,
+ "learning_rate": 0.0006,
+ "loss": 6.724848747253418,
+ "step": 261
+ },
+ {
+ "epoch": 3.6427947598253274,
+ "grad_norm": 0.007282314356416464,
+ "learning_rate": 0.0006,
+ "loss": 6.702306270599365,
+ "step": 262
+ },
+ {
+ "epoch": 3.656768558951965,
+ "grad_norm": 0.02673051692545414,
+ "learning_rate": 0.0006,
+ "loss": 6.72357177734375,
+ "step": 263
+ },
+ {
+ "epoch": 3.6707423580786025,
+ "grad_norm": 0.016174081712961197,
+ "learning_rate": 0.0006,
+ "loss": 6.730698585510254,
+ "step": 264
+ },
+ {
+ "epoch": 3.6847161572052403,
+ "grad_norm": 0.0182416420429945,
+ "learning_rate": 0.0006,
+ "loss": 6.717386245727539,
+ "step": 265
+ },
+ {
+ "epoch": 3.6986899563318776,
+ "grad_norm": 0.023250866681337357,
+ "learning_rate": 0.0006,
+ "loss": 6.697832107543945,
+ "step": 266
+ },
+ {
+ "epoch": 3.7126637554585153,
+ "grad_norm": 0.007539753336459398,
+ "learning_rate": 0.0006,
+ "loss": 6.714348793029785,
+ "step": 267
+ },
+ {
+ "epoch": 3.726637554585153,
+ "grad_norm": 0.030086630955338478,
+ "learning_rate": 0.0006,
+ "loss": 6.724481582641602,
+ "step": 268
+ },
+ {
+ "epoch": 3.7406113537117904,
+ "grad_norm": 0.012784913182258606,
+ "learning_rate": 0.0006,
+ "loss": 6.703491687774658,
+ "step": 269
+ },
+ {
+ "epoch": 3.7545851528384278,
+ "grad_norm": 0.02042769268155098,
+ "learning_rate": 0.0006,
+ "loss": 6.721139907836914,
+ "step": 270
+ },
+ {
+ "epoch": 3.7685589519650655,
+ "grad_norm": 0.023557329550385475,
+ "learning_rate": 0.0006,
+ "loss": 6.728122711181641,
+ "step": 271
+ },
+ {
+ "epoch": 3.7825327510917033,
+ "grad_norm": 0.007716621737927198,
+ "learning_rate": 0.0006,
+ "loss": 6.70085334777832,
+ "step": 272
+ },
+ {
+ "epoch": 3.7965065502183406,
+ "grad_norm": 0.022772278636693954,
+ "learning_rate": 0.0006,
+ "loss": 6.712882995605469,
+ "step": 273
+ },
+ {
+ "epoch": 3.810480349344978,
+ "grad_norm": 0.007928933948278427,
+ "learning_rate": 0.0006,
+ "loss": 6.698320388793945,
+ "step": 274
+ },
+ {
+ "epoch": 3.8244541484716157,
+ "grad_norm": 0.022355245426297188,
+ "learning_rate": 0.0006,
+ "loss": 6.687088489532471,
+ "step": 275
+ },
+ {
+ "epoch": 3.8384279475982535,
+ "grad_norm": 0.02037850394845009,
+ "learning_rate": 0.0006,
+ "loss": 6.676856994628906,
+ "step": 276
+ },
+ {
+ "epoch": 3.8524017467248908,
+ "grad_norm": 0.012290362268686295,
+ "learning_rate": 0.0006,
+ "loss": 6.682883262634277,
+ "step": 277
+ },
+ {
+ "epoch": 3.8663755458515285,
+ "grad_norm": 0.022331275045871735,
+ "learning_rate": 0.0006,
+ "loss": 6.694752216339111,
+ "step": 278
+ },
+ {
+ "epoch": 3.880349344978166,
+ "grad_norm": 0.007527181413024664,
+ "learning_rate": 0.0006,
+ "loss": 6.686394691467285,
+ "step": 279
+ },
+ {
+ "epoch": 3.8943231441048036,
+ "grad_norm": 0.02159409038722515,
+ "learning_rate": 0.0006,
+ "loss": 6.697943687438965,
+ "step": 280
+ },
+ {
+ "epoch": 3.908296943231441,
+ "grad_norm": 0.015054763294756413,
+ "learning_rate": 0.0006,
+ "loss": 6.677796363830566,
+ "step": 281
+ },
+ {
+ "epoch": 3.9222707423580787,
+ "grad_norm": 0.01466019731014967,
+ "learning_rate": 0.0006,
+ "loss": 6.697715759277344,
+ "step": 282
+ },
+ {
+ "epoch": 3.936244541484716,
+ "grad_norm": 0.01875465176999569,
+ "learning_rate": 0.0006,
+ "loss": 6.700228691101074,
+ "step": 283
+ },
+ {
+ "epoch": 3.950218340611354,
+ "grad_norm": 0.00778991961851716,
+ "learning_rate": 0.0006,
+ "loss": 6.69331693649292,
+ "step": 284
+ },
+ {
+ "epoch": 3.964192139737991,
+ "grad_norm": 0.020182441920042038,
+ "learning_rate": 0.0006,
+ "loss": 6.672651290893555,
+ "step": 285
+ },
+ {
+ "epoch": 3.978165938864629,
+ "grad_norm": 0.01240516733378172,
+ "learning_rate": 0.0006,
+ "loss": 6.688860893249512,
+ "step": 286
+ },
+ {
+ "epoch": 3.992139737991266,
+ "grad_norm": 0.011218412779271603,
+ "learning_rate": 0.0006,
+ "loss": 6.669534683227539,
+ "step": 287
+ },
+ {
+ "epoch": 4.0,
+ "grad_norm": 0.0141852255910635,
+ "learning_rate": 0.0006,
+ "loss": 6.690206527709961,
+ "step": 288
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 28800,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 400,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": true,
+ "should_log": false,
+ "should_save": false,
+ "should_training_stop": false
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 1.2278453445328896e+17,
+ "train_batch_size": 8,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-288/training_args.bin b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-288/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..c799bcf4051249aabf4c62fff9d38801df7bc2b5
--- /dev/null
+++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-288/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:40fb25263250c08c3479a6f0ac9fdf1f196e2a3b08a4eb10384786ff9b1903ef
+size 4792
diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-576/chat_template.jinja b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-576/chat_template.jinja
new file mode 100644
index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230
--- /dev/null
+++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-576/chat_template.jinja
@@ -0,0 +1,85 @@
+{%- if tools %}
+ {{- '<|im_start|>system\n' }}
+ {%- if messages[0].role == 'system' %}
+ {{- messages[0].content + '\n\n' }}
+ {%- endif %}
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }}
+ {%- for tool in tools %}
+ {{- "\n" }}
+ {{- tool | tojson }}
+ {%- endfor %}
+ {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }}
+{%- else %}
+ {%- if messages[0].role == 'system' %}
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
+ {%- endif %}
+{%- endif %}
+{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
+{%- for message in messages[::-1] %}
+ {%- set index = (messages|length - 1) - loop.index0 %}
+ {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %}
+ {%- set ns.multi_step_tool = false %}
+ {%- set ns.last_query_index = index %}
+ {%- endif %}
+{%- endfor %}
+{%- for message in messages %}
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
+ {%- elif message.role == "assistant" %}
+ {%- set content = message.content %}
+ {%- set reasoning_content = '' %}
+ {%- if message.reasoning_content is defined and message.reasoning_content is not none %}
+ {%- set reasoning_content = message.reasoning_content %}
+ {%- else %}
+ {%- if '' in message.content %}
+ {%- set content = message.content.split('')[-1].lstrip('\n') %}
+ {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %}
+ {%- endif %}
+ {%- endif %}
+ {%- if loop.index0 > ns.last_query_index %}
+ {%- if loop.last or (not loop.last and reasoning_content) %}
+ {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- if message.tool_calls %}
+ {%- for tool_call in message.tool_calls %}
+ {%- if (loop.first and content) or (not loop.first) %}
+ {{- '\n' }}
+ {%- endif %}
+ {%- if tool_call.function %}
+ {%- set tool_call = tool_call.function %}
+ {%- endif %}
+ {{- '\n{"name": "' }}
+ {{- tool_call.name }}
+ {{- '", "arguments": ' }}
+ {%- if tool_call.arguments is string %}
+ {{- tool_call.arguments }}
+ {%- else %}
+ {{- tool_call.arguments | tojson }}
+ {%- endif %}
+ {{- '}\n' }}
+ {%- endfor %}
+ {%- endif %}
+ {{- '<|im_end|>\n' }}
+ {%- elif message.role == "tool" %}
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
+ {{- '<|im_start|>user' }}
+ {%- endif %}
+ {{- '\n\n' }}
+ {{- message.content }}
+ {{- '\n' }}
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
+ {{- '<|im_end|>\n' }}
+ {%- endif %}
+ {%- endif %}
+{%- endfor %}
+{%- if add_generation_prompt %}
+ {{- '<|im_start|>assistant\n' }}
+ {%- if enable_thinking is defined and enable_thinking is false %}
+ {{- '\n\n\n\n' }}
+ {%- endif %}
+{%- endif %}
\ No newline at end of file
diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-576/config.json b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-576/config.json
new file mode 100644
index 0000000000000000000000000000000000000000..175ee3431793018f23f74824f69554b21afcbd4f
--- /dev/null
+++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-576/config.json
@@ -0,0 +1,32 @@
+{
+ "architectures": [
+ "LlamaForCausalLM"
+ ],
+ "attention_bias": false,
+ "attention_dropout": 0.0,
+ "bos_token_id": null,
+ "dtype": "float32",
+ "eos_token_id": 151645,
+ "head_dim": 128,
+ "hidden_act": "silu",
+ "hidden_size": 512,
+ "initializer_range": 0.02,
+ "intermediate_size": 1536,
+ "max_position_embeddings": 2048,
+ "mlp_bias": false,
+ "model_type": "llama",
+ "num_attention_heads": 4,
+ "num_hidden_layers": 20,
+ "num_key_value_heads": 4,
+ "pad_token_id": 151645,
+ "pretraining_tp": 1,
+ "rms_norm_eps": 1e-06,
+ "rope_parameters": {
+ "rope_theta": 10000.0,
+ "rope_type": "default"
+ },
+ "tie_word_embeddings": true,
+ "transformers_version": "5.5.0",
+ "use_cache": false,
+ "vocab_size": 151674
+}
diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-576/generation_config.json b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-576/generation_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3
--- /dev/null
+++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-576/generation_config.json
@@ -0,0 +1,11 @@
+{
+ "_from_model_config": true,
+ "eos_token_id": [
+ 151645
+ ],
+ "output_attentions": false,
+ "output_hidden_states": false,
+ "pad_token_id": 151645,
+ "transformers_version": "5.5.0",
+ "use_cache": true
+}
diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-576/model.safetensors b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-576/model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..0bf8f0e1914ebcf20f10ca6ef1d21e17f4e2fa9e
--- /dev/null
+++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-576/model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:ea4335cd4b1048a8590c5723e07a74c2f88f7104372810a660a78b3e8a83153c
+size 583362376
diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-576/optimizer.pt b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-576/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..77daaeb55ca47e42103f5232718b3c16a7060a25
--- /dev/null
+++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-576/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:736618660f2f79bdb5b703a4de951220a9791c054c8458a3aad796bb4e4970f6
+size 1166837626
diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-576/rng_state_0.pth b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-576/rng_state_0.pth
new file mode 100644
index 0000000000000000000000000000000000000000..ac9172a69fff7bb676f49c9906b9314a59fdad2d
--- /dev/null
+++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-576/rng_state_0.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:c48c5ec5c9bb4c521c495daafd6b39a9c355c1b6424c2e78c7796e18f5fdd3fb
+size 15024
diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-576/rng_state_1.pth b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-576/rng_state_1.pth
new file mode 100644
index 0000000000000000000000000000000000000000..7f003aa3c9fce222cf1ebce0837bf5b92cde59e3
--- /dev/null
+++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-576/rng_state_1.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:2e8a93a2b7133c2fed28a7b289f0e7b9c9545ff71cee4aae92941f10b916a20f
+size 15024
diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-576/rng_state_2.pth b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-576/rng_state_2.pth
new file mode 100644
index 0000000000000000000000000000000000000000..86120748c38cd9ddec772def3d227c8f1512792b
--- /dev/null
+++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-576/rng_state_2.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:cb930d5328a582540de3500ea4a3237c706f2698967029d50f2025f2c69fbf69
+size 15024
diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-576/rng_state_3.pth b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-576/rng_state_3.pth
new file mode 100644
index 0000000000000000000000000000000000000000..c35ff8c6ec57f5a32994c021011ed89dc2fb1637
--- /dev/null
+++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-576/rng_state_3.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:a71e181c472366aba18ce125f3793f840051870bcbc1acd0fa9ef87d62644fc2
+size 15024
diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-576/scheduler.pt b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-576/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..dbc454a52c5d1edb241846f7cbf86268ced2c2ff
--- /dev/null
+++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-576/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:5f28b8022dab4a12fbbad2bccde532c0f7c60e469bde0bf66d40f7d043ac1727
+size 1064
diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-576/tokenizer.json b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-576/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..9ea28d9eb325c36bcab4c1a08fc93fd598929f08
--- /dev/null
+++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-576/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:60e99075d371ecbb9f1dde97363f612d45bf0bf88f1a2006d9bfcf1939333c14
+size 11424648
diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-576/tokenizer_config.json b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-576/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..cadd997fc26040d23ad6677fcb39d8d18b01f55e
--- /dev/null
+++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-576/tokenizer_config.json
@@ -0,0 +1,21 @@
+{
+ "add_prefix_space": false,
+ "backend": "tokenizers",
+ "bos_token": null,
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "<|im_end|>",
+ "errors": "replace",
+ "extra_special_tokens": [
+ "<|l2r_pred|>",
+ "<|r2l_pred|>",
+ "<|next_1_pred|>",
+ "<|next_2_pred|>",
+ "<|mask|>"
+ ],
+ "is_local": false,
+ "model_max_length": 131072,
+ "pad_token": "<|im_end|>",
+ "split_special_tokens": false,
+ "tokenizer_class": "Qwen2Tokenizer",
+ "unk_token": null
+}
diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-576/trainer_state.json b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-576/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..e15036434503c93bf7f8c05c5b04b17aa1ee6e5a
--- /dev/null
+++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-576/trainer_state.json
@@ -0,0 +1,4122 @@
+{
+ "best_global_step": null,
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 8.0,
+ "eval_steps": 500,
+ "global_step": 576,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.013973799126637555,
+ "grad_norm": 0.5768774151802063,
+ "learning_rate": 0.0,
+ "loss": 12.050328254699707,
+ "step": 1
+ },
+ {
+ "epoch": 0.02794759825327511,
+ "grad_norm": 0.5784570574760437,
+ "learning_rate": 5.999999999999999e-06,
+ "loss": 12.049251556396484,
+ "step": 2
+ },
+ {
+ "epoch": 0.04192139737991266,
+ "grad_norm": 0.5607187151908875,
+ "learning_rate": 1.1999999999999999e-05,
+ "loss": 11.921792030334473,
+ "step": 3
+ },
+ {
+ "epoch": 0.05589519650655022,
+ "grad_norm": 0.46466919779777527,
+ "learning_rate": 1.7999999999999997e-05,
+ "loss": 11.686628341674805,
+ "step": 4
+ },
+ {
+ "epoch": 0.06986899563318777,
+ "grad_norm": 0.38012945652008057,
+ "learning_rate": 2.3999999999999997e-05,
+ "loss": 11.432823181152344,
+ "step": 5
+ },
+ {
+ "epoch": 0.08384279475982533,
+ "grad_norm": 0.33215776085853577,
+ "learning_rate": 2.9999999999999997e-05,
+ "loss": 11.207542419433594,
+ "step": 6
+ },
+ {
+ "epoch": 0.09781659388646288,
+ "grad_norm": 0.2928776741027832,
+ "learning_rate": 3.5999999999999994e-05,
+ "loss": 11.000097274780273,
+ "step": 7
+ },
+ {
+ "epoch": 0.11179039301310044,
+ "grad_norm": 0.26018962264060974,
+ "learning_rate": 4.2e-05,
+ "loss": 10.825748443603516,
+ "step": 8
+ },
+ {
+ "epoch": 0.125764192139738,
+ "grad_norm": 0.23717083036899567,
+ "learning_rate": 4.7999999999999994e-05,
+ "loss": 10.677122116088867,
+ "step": 9
+ },
+ {
+ "epoch": 0.13973799126637554,
+ "grad_norm": 0.21641355752944946,
+ "learning_rate": 5.399999999999999e-05,
+ "loss": 10.569148063659668,
+ "step": 10
+ },
+ {
+ "epoch": 0.1537117903930131,
+ "grad_norm": 0.1982956826686859,
+ "learning_rate": 5.9999999999999995e-05,
+ "loss": 10.49653434753418,
+ "step": 11
+ },
+ {
+ "epoch": 0.16768558951965065,
+ "grad_norm": 0.18819282948970795,
+ "learning_rate": 6.599999999999999e-05,
+ "loss": 10.432951927185059,
+ "step": 12
+ },
+ {
+ "epoch": 0.18165938864628822,
+ "grad_norm": 0.18322618305683136,
+ "learning_rate": 7.199999999999999e-05,
+ "loss": 10.387115478515625,
+ "step": 13
+ },
+ {
+ "epoch": 0.19563318777292577,
+ "grad_norm": 0.18399226665496826,
+ "learning_rate": 7.8e-05,
+ "loss": 10.33381462097168,
+ "step": 14
+ },
+ {
+ "epoch": 0.2096069868995633,
+ "grad_norm": 0.1833190619945526,
+ "learning_rate": 8.4e-05,
+ "loss": 10.301689147949219,
+ "step": 15
+ },
+ {
+ "epoch": 0.22358078602620088,
+ "grad_norm": 0.18490834534168243,
+ "learning_rate": 8.999999999999999e-05,
+ "loss": 10.240968704223633,
+ "step": 16
+ },
+ {
+ "epoch": 0.23755458515283842,
+ "grad_norm": 0.18358023464679718,
+ "learning_rate": 9.599999999999999e-05,
+ "loss": 10.194717407226562,
+ "step": 17
+ },
+ {
+ "epoch": 0.251528384279476,
+ "grad_norm": 0.1801682710647583,
+ "learning_rate": 0.000102,
+ "loss": 10.14897632598877,
+ "step": 18
+ },
+ {
+ "epoch": 0.26550218340611353,
+ "grad_norm": 0.17899440228939056,
+ "learning_rate": 0.00010799999999999998,
+ "loss": 10.078763961791992,
+ "step": 19
+ },
+ {
+ "epoch": 0.2794759825327511,
+ "grad_norm": 0.17666497826576233,
+ "learning_rate": 0.00011399999999999999,
+ "loss": 10.01095962524414,
+ "step": 20
+ },
+ {
+ "epoch": 0.2934497816593886,
+ "grad_norm": 0.17196154594421387,
+ "learning_rate": 0.00011999999999999999,
+ "loss": 9.951203346252441,
+ "step": 21
+ },
+ {
+ "epoch": 0.3074235807860262,
+ "grad_norm": 0.169203519821167,
+ "learning_rate": 0.00012599999999999997,
+ "loss": 9.871726036071777,
+ "step": 22
+ },
+ {
+ "epoch": 0.32139737991266376,
+ "grad_norm": 0.16500306129455566,
+ "learning_rate": 0.00013199999999999998,
+ "loss": 9.803646087646484,
+ "step": 23
+ },
+ {
+ "epoch": 0.3353711790393013,
+ "grad_norm": 0.16128124296665192,
+ "learning_rate": 0.000138,
+ "loss": 9.728487014770508,
+ "step": 24
+ },
+ {
+ "epoch": 0.34934497816593885,
+ "grad_norm": 0.15904058516025543,
+ "learning_rate": 0.00014399999999999998,
+ "loss": 9.642473220825195,
+ "step": 25
+ },
+ {
+ "epoch": 0.36331877729257644,
+ "grad_norm": 0.15560561418533325,
+ "learning_rate": 0.00015,
+ "loss": 9.56414794921875,
+ "step": 26
+ },
+ {
+ "epoch": 0.377292576419214,
+ "grad_norm": 0.15278612077236176,
+ "learning_rate": 0.000156,
+ "loss": 9.484230041503906,
+ "step": 27
+ },
+ {
+ "epoch": 0.39126637554585153,
+ "grad_norm": 0.15263237059116364,
+ "learning_rate": 0.000162,
+ "loss": 9.383644104003906,
+ "step": 28
+ },
+ {
+ "epoch": 0.4052401746724891,
+ "grad_norm": 0.14945024251937866,
+ "learning_rate": 0.000168,
+ "loss": 9.29840087890625,
+ "step": 29
+ },
+ {
+ "epoch": 0.4192139737991266,
+ "grad_norm": 0.14614646136760712,
+ "learning_rate": 0.00017399999999999997,
+ "loss": 9.208931922912598,
+ "step": 30
+ },
+ {
+ "epoch": 0.4331877729257642,
+ "grad_norm": 0.14295098185539246,
+ "learning_rate": 0.00017999999999999998,
+ "loss": 9.122323989868164,
+ "step": 31
+ },
+ {
+ "epoch": 0.44716157205240176,
+ "grad_norm": 0.13868345320224762,
+ "learning_rate": 0.000186,
+ "loss": 9.03034496307373,
+ "step": 32
+ },
+ {
+ "epoch": 0.4611353711790393,
+ "grad_norm": 0.135064035654068,
+ "learning_rate": 0.00019199999999999998,
+ "loss": 8.92818832397461,
+ "step": 33
+ },
+ {
+ "epoch": 0.47510917030567684,
+ "grad_norm": 0.13097241520881653,
+ "learning_rate": 0.000198,
+ "loss": 8.834627151489258,
+ "step": 34
+ },
+ {
+ "epoch": 0.4890829694323144,
+ "grad_norm": 0.12478776276111603,
+ "learning_rate": 0.000204,
+ "loss": 8.755172729492188,
+ "step": 35
+ },
+ {
+ "epoch": 0.503056768558952,
+ "grad_norm": 0.1214982345700264,
+ "learning_rate": 0.00020999999999999998,
+ "loss": 8.647891998291016,
+ "step": 36
+ },
+ {
+ "epoch": 0.5170305676855895,
+ "grad_norm": 0.11958328634500504,
+ "learning_rate": 0.00021599999999999996,
+ "loss": 8.543274879455566,
+ "step": 37
+ },
+ {
+ "epoch": 0.5310043668122271,
+ "grad_norm": 0.11680915951728821,
+ "learning_rate": 0.00022199999999999998,
+ "loss": 8.44540023803711,
+ "step": 38
+ },
+ {
+ "epoch": 0.5449781659388646,
+ "grad_norm": 0.11491474509239197,
+ "learning_rate": 0.00022799999999999999,
+ "loss": 8.349921226501465,
+ "step": 39
+ },
+ {
+ "epoch": 0.5589519650655022,
+ "grad_norm": 0.11160755902528763,
+ "learning_rate": 0.000234,
+ "loss": 8.26960563659668,
+ "step": 40
+ },
+ {
+ "epoch": 0.5729257641921397,
+ "grad_norm": 0.10840637236833572,
+ "learning_rate": 0.00023999999999999998,
+ "loss": 8.192136764526367,
+ "step": 41
+ },
+ {
+ "epoch": 0.5868995633187772,
+ "grad_norm": 0.10585298389196396,
+ "learning_rate": 0.00024599999999999996,
+ "loss": 8.09546947479248,
+ "step": 42
+ },
+ {
+ "epoch": 0.6008733624454149,
+ "grad_norm": 0.1032542735338211,
+ "learning_rate": 0.00025199999999999995,
+ "loss": 7.991090774536133,
+ "step": 43
+ },
+ {
+ "epoch": 0.6148471615720524,
+ "grad_norm": 0.0986943170428276,
+ "learning_rate": 0.000258,
+ "loss": 7.92384147644043,
+ "step": 44
+ },
+ {
+ "epoch": 0.62882096069869,
+ "grad_norm": 0.09320762753486633,
+ "learning_rate": 0.00026399999999999997,
+ "loss": 7.8478193283081055,
+ "step": 45
+ },
+ {
+ "epoch": 0.6427947598253275,
+ "grad_norm": 0.09168332070112228,
+ "learning_rate": 0.00027,
+ "loss": 7.76185417175293,
+ "step": 46
+ },
+ {
+ "epoch": 0.6567685589519651,
+ "grad_norm": 0.08832395076751709,
+ "learning_rate": 0.000276,
+ "loss": 7.679399490356445,
+ "step": 47
+ },
+ {
+ "epoch": 0.6707423580786026,
+ "grad_norm": 0.08816400915384293,
+ "learning_rate": 0.00028199999999999997,
+ "loss": 7.5789642333984375,
+ "step": 48
+ },
+ {
+ "epoch": 0.6847161572052402,
+ "grad_norm": 0.0861717164516449,
+ "learning_rate": 0.00028799999999999995,
+ "loss": 7.519214630126953,
+ "step": 49
+ },
+ {
+ "epoch": 0.6986899563318777,
+ "grad_norm": 0.08103746175765991,
+ "learning_rate": 0.000294,
+ "loss": 7.4475321769714355,
+ "step": 50
+ },
+ {
+ "epoch": 0.7126637554585152,
+ "grad_norm": 0.07545104622840881,
+ "learning_rate": 0.0003,
+ "loss": 7.401182174682617,
+ "step": 51
+ },
+ {
+ "epoch": 0.7266375545851529,
+ "grad_norm": 0.07042575627565384,
+ "learning_rate": 0.00030599999999999996,
+ "loss": 7.340173244476318,
+ "step": 52
+ },
+ {
+ "epoch": 0.7406113537117904,
+ "grad_norm": 0.06737572699785233,
+ "learning_rate": 0.000312,
+ "loss": 7.265697002410889,
+ "step": 53
+ },
+ {
+ "epoch": 0.754585152838428,
+ "grad_norm": 0.06151146441698074,
+ "learning_rate": 0.000318,
+ "loss": 7.231061935424805,
+ "step": 54
+ },
+ {
+ "epoch": 0.7685589519650655,
+ "grad_norm": 0.05638071522116661,
+ "learning_rate": 0.000324,
+ "loss": 7.195466995239258,
+ "step": 55
+ },
+ {
+ "epoch": 0.7825327510917031,
+ "grad_norm": 0.05084463208913803,
+ "learning_rate": 0.00033,
+ "loss": 7.137603759765625,
+ "step": 56
+ },
+ {
+ "epoch": 0.7965065502183406,
+ "grad_norm": 0.04293261095881462,
+ "learning_rate": 0.000336,
+ "loss": 7.115458965301514,
+ "step": 57
+ },
+ {
+ "epoch": 0.8104803493449781,
+ "grad_norm": 0.036064863204956055,
+ "learning_rate": 0.00034199999999999996,
+ "loss": 7.089038848876953,
+ "step": 58
+ },
+ {
+ "epoch": 0.8244541484716157,
+ "grad_norm": 0.029026754200458527,
+ "learning_rate": 0.00034799999999999995,
+ "loss": 7.0584306716918945,
+ "step": 59
+ },
+ {
+ "epoch": 0.8384279475982532,
+ "grad_norm": 0.02286517433822155,
+ "learning_rate": 0.00035399999999999993,
+ "loss": 7.049450874328613,
+ "step": 60
+ },
+ {
+ "epoch": 0.8524017467248908,
+ "grad_norm": 0.018878433853387833,
+ "learning_rate": 0.00035999999999999997,
+ "loss": 7.008824348449707,
+ "step": 61
+ },
+ {
+ "epoch": 0.8663755458515284,
+ "grad_norm": 0.010773813351988792,
+ "learning_rate": 0.00036599999999999995,
+ "loss": 7.024829864501953,
+ "step": 62
+ },
+ {
+ "epoch": 0.880349344978166,
+ "grad_norm": 0.008794732391834259,
+ "learning_rate": 0.000372,
+ "loss": 7.018399238586426,
+ "step": 63
+ },
+ {
+ "epoch": 0.8943231441048035,
+ "grad_norm": 0.011709796264767647,
+ "learning_rate": 0.00037799999999999997,
+ "loss": 7.0101752281188965,
+ "step": 64
+ },
+ {
+ "epoch": 0.9082969432314411,
+ "grad_norm": 0.012123258784413338,
+ "learning_rate": 0.00038399999999999996,
+ "loss": 7.024317741394043,
+ "step": 65
+ },
+ {
+ "epoch": 0.9222707423580786,
+ "grad_norm": 0.01530320756137371,
+ "learning_rate": 0.00039,
+ "loss": 7.027738571166992,
+ "step": 66
+ },
+ {
+ "epoch": 0.9362445414847161,
+ "grad_norm": 0.01890002004802227,
+ "learning_rate": 0.000396,
+ "loss": 7.024099349975586,
+ "step": 67
+ },
+ {
+ "epoch": 0.9502183406113537,
+ "grad_norm": 0.020684150978922844,
+ "learning_rate": 0.000402,
+ "loss": 7.007943153381348,
+ "step": 68
+ },
+ {
+ "epoch": 0.9641921397379912,
+ "grad_norm": 0.019826196134090424,
+ "learning_rate": 0.000408,
+ "loss": 7.006874084472656,
+ "step": 69
+ },
+ {
+ "epoch": 0.9781659388646288,
+ "grad_norm": 0.019218673929572105,
+ "learning_rate": 0.0004139999999999999,
+ "loss": 7.002897262573242,
+ "step": 70
+ },
+ {
+ "epoch": 0.9921397379912664,
+ "grad_norm": 0.013856411911547184,
+ "learning_rate": 0.00041999999999999996,
+ "loss": 6.993999481201172,
+ "step": 71
+ },
+ {
+ "epoch": 1.0,
+ "grad_norm": 0.010390003211796284,
+ "learning_rate": 0.00042599999999999995,
+ "loss": 7.000683784484863,
+ "step": 72
+ },
+ {
+ "epoch": 1.0,
+ "eval_loss": 7.91140079498291,
+ "eval_runtime": 57.7925,
+ "eval_samples_per_second": 42.255,
+ "eval_steps_per_second": 1.332,
+ "step": 72
+ },
+ {
+ "epoch": 1.0139737991266375,
+ "grad_norm": 0.013802163302898407,
+ "learning_rate": 0.00043199999999999993,
+ "loss": 6.998257637023926,
+ "step": 73
+ },
+ {
+ "epoch": 1.027947598253275,
+ "grad_norm": 0.010109687224030495,
+ "learning_rate": 0.00043799999999999997,
+ "loss": 7.002542018890381,
+ "step": 74
+ },
+ {
+ "epoch": 1.0419213973799126,
+ "grad_norm": 0.008019139990210533,
+ "learning_rate": 0.00044399999999999995,
+ "loss": 6.985941410064697,
+ "step": 75
+ },
+ {
+ "epoch": 1.0558951965065502,
+ "grad_norm": 0.0069672465324401855,
+ "learning_rate": 0.00045,
+ "loss": 6.99245023727417,
+ "step": 76
+ },
+ {
+ "epoch": 1.0698689956331877,
+ "grad_norm": 0.00714076729491353,
+ "learning_rate": 0.00045599999999999997,
+ "loss": 6.959296226501465,
+ "step": 77
+ },
+ {
+ "epoch": 1.0838427947598253,
+ "grad_norm": 0.008408435620367527,
+ "learning_rate": 0.00046199999999999995,
+ "loss": 6.954504489898682,
+ "step": 78
+ },
+ {
+ "epoch": 1.0978165938864628,
+ "grad_norm": 0.008778786286711693,
+ "learning_rate": 0.000468,
+ "loss": 7.0007004737854,
+ "step": 79
+ },
+ {
+ "epoch": 1.1117903930131003,
+ "grad_norm": 0.007753407116979361,
+ "learning_rate": 0.000474,
+ "loss": 6.98183536529541,
+ "step": 80
+ },
+ {
+ "epoch": 1.125764192139738,
+ "grad_norm": 0.01087260153144598,
+ "learning_rate": 0.00047999999999999996,
+ "loss": 6.991066932678223,
+ "step": 81
+ },
+ {
+ "epoch": 1.1397379912663754,
+ "grad_norm": 0.006122548598796129,
+ "learning_rate": 0.000486,
+ "loss": 6.978824138641357,
+ "step": 82
+ },
+ {
+ "epoch": 1.1537117903930132,
+ "grad_norm": 0.0067383465357124805,
+ "learning_rate": 0.0004919999999999999,
+ "loss": 6.961703777313232,
+ "step": 83
+ },
+ {
+ "epoch": 1.1676855895196507,
+ "grad_norm": 0.007922791875898838,
+ "learning_rate": 0.000498,
+ "loss": 6.954860687255859,
+ "step": 84
+ },
+ {
+ "epoch": 1.1816593886462883,
+ "grad_norm": 0.00818922370672226,
+ "learning_rate": 0.0005039999999999999,
+ "loss": 6.945961952209473,
+ "step": 85
+ },
+ {
+ "epoch": 1.1956331877729258,
+ "grad_norm": 0.006791461259126663,
+ "learning_rate": 0.0005099999999999999,
+ "loss": 6.968680381774902,
+ "step": 86
+ },
+ {
+ "epoch": 1.2096069868995634,
+ "grad_norm": 0.005394163075834513,
+ "learning_rate": 0.000516,
+ "loss": 6.956404685974121,
+ "step": 87
+ },
+ {
+ "epoch": 1.223580786026201,
+ "grad_norm": 0.004500220064073801,
+ "learning_rate": 0.000522,
+ "loss": 6.96998405456543,
+ "step": 88
+ },
+ {
+ "epoch": 1.2375545851528384,
+ "grad_norm": 0.008699574507772923,
+ "learning_rate": 0.0005279999999999999,
+ "loss": 6.955585479736328,
+ "step": 89
+ },
+ {
+ "epoch": 1.251528384279476,
+ "grad_norm": 0.005889365915209055,
+ "learning_rate": 0.000534,
+ "loss": 6.9708051681518555,
+ "step": 90
+ },
+ {
+ "epoch": 1.2655021834061135,
+ "grad_norm": 0.009646810591220856,
+ "learning_rate": 0.00054,
+ "loss": 6.962608814239502,
+ "step": 91
+ },
+ {
+ "epoch": 1.279475982532751,
+ "grad_norm": 0.004517171997576952,
+ "learning_rate": 0.0005459999999999999,
+ "loss": 6.957627296447754,
+ "step": 92
+ },
+ {
+ "epoch": 1.2934497816593886,
+ "grad_norm": 0.006339677143841982,
+ "learning_rate": 0.000552,
+ "loss": 6.963141441345215,
+ "step": 93
+ },
+ {
+ "epoch": 1.3074235807860262,
+ "grad_norm": 0.005387383047491312,
+ "learning_rate": 0.000558,
+ "loss": 6.934465408325195,
+ "step": 94
+ },
+ {
+ "epoch": 1.3213973799126637,
+ "grad_norm": 0.005252422299236059,
+ "learning_rate": 0.0005639999999999999,
+ "loss": 6.956108093261719,
+ "step": 95
+ },
+ {
+ "epoch": 1.3353711790393012,
+ "grad_norm": 0.008699414320290089,
+ "learning_rate": 0.00057,
+ "loss": 6.973369598388672,
+ "step": 96
+ },
+ {
+ "epoch": 1.3493449781659388,
+ "grad_norm": 0.003947409335523844,
+ "learning_rate": 0.0005759999999999999,
+ "loss": 6.950225830078125,
+ "step": 97
+ },
+ {
+ "epoch": 1.3633187772925766,
+ "grad_norm": 0.00768681475892663,
+ "learning_rate": 0.0005819999999999999,
+ "loss": 6.964739799499512,
+ "step": 98
+ },
+ {
+ "epoch": 1.3772925764192139,
+ "grad_norm": 0.006563829258084297,
+ "learning_rate": 0.000588,
+ "loss": 6.972421169281006,
+ "step": 99
+ },
+ {
+ "epoch": 1.3912663755458516,
+ "grad_norm": 0.007606555242091417,
+ "learning_rate": 0.0005939999999999999,
+ "loss": 6.958213806152344,
+ "step": 100
+ },
+ {
+ "epoch": 1.405240174672489,
+ "grad_norm": 0.006250880658626556,
+ "learning_rate": 0.0006,
+ "loss": 6.962296485900879,
+ "step": 101
+ },
+ {
+ "epoch": 1.4192139737991267,
+ "grad_norm": 0.0073033408261835575,
+ "learning_rate": 0.0006,
+ "loss": 6.969705104827881,
+ "step": 102
+ },
+ {
+ "epoch": 1.4331877729257643,
+ "grad_norm": 0.0054510910995304585,
+ "learning_rate": 0.0006,
+ "loss": 6.9652862548828125,
+ "step": 103
+ },
+ {
+ "epoch": 1.4471615720524018,
+ "grad_norm": 0.004895673133432865,
+ "learning_rate": 0.0006,
+ "loss": 6.956023693084717,
+ "step": 104
+ },
+ {
+ "epoch": 1.4611353711790394,
+ "grad_norm": 0.005306515377014875,
+ "learning_rate": 0.0006,
+ "loss": 6.963231086730957,
+ "step": 105
+ },
+ {
+ "epoch": 1.475109170305677,
+ "grad_norm": 0.004742146469652653,
+ "learning_rate": 0.0006,
+ "loss": 6.9582061767578125,
+ "step": 106
+ },
+ {
+ "epoch": 1.4890829694323144,
+ "grad_norm": 0.004220000468194485,
+ "learning_rate": 0.0006,
+ "loss": 6.972961902618408,
+ "step": 107
+ },
+ {
+ "epoch": 1.503056768558952,
+ "grad_norm": 0.0042153168469667435,
+ "learning_rate": 0.0006,
+ "loss": 6.961165904998779,
+ "step": 108
+ },
+ {
+ "epoch": 1.5170305676855895,
+ "grad_norm": 0.005165040958672762,
+ "learning_rate": 0.0006,
+ "loss": 6.95018196105957,
+ "step": 109
+ },
+ {
+ "epoch": 1.531004366812227,
+ "grad_norm": 0.003930266480892897,
+ "learning_rate": 0.0006,
+ "loss": 6.9651689529418945,
+ "step": 110
+ },
+ {
+ "epoch": 1.5449781659388646,
+ "grad_norm": 0.004465071018785238,
+ "learning_rate": 0.0006,
+ "loss": 6.960553169250488,
+ "step": 111
+ },
+ {
+ "epoch": 1.5589519650655022,
+ "grad_norm": 0.004888612311333418,
+ "learning_rate": 0.0006,
+ "loss": 6.967258930206299,
+ "step": 112
+ },
+ {
+ "epoch": 1.5729257641921397,
+ "grad_norm": 0.007497863378375769,
+ "learning_rate": 0.0006,
+ "loss": 6.954545021057129,
+ "step": 113
+ },
+ {
+ "epoch": 1.5868995633187772,
+ "grad_norm": 0.005659648682922125,
+ "learning_rate": 0.0006,
+ "loss": 6.962299823760986,
+ "step": 114
+ },
+ {
+ "epoch": 1.600873362445415,
+ "grad_norm": 0.006641396787017584,
+ "learning_rate": 0.0006,
+ "loss": 6.952983856201172,
+ "step": 115
+ },
+ {
+ "epoch": 1.6148471615720523,
+ "grad_norm": 0.009715928696095943,
+ "learning_rate": 0.0006,
+ "loss": 6.958622932434082,
+ "step": 116
+ },
+ {
+ "epoch": 1.62882096069869,
+ "grad_norm": 0.012789523229002953,
+ "learning_rate": 0.0006,
+ "loss": 6.932043075561523,
+ "step": 117
+ },
+ {
+ "epoch": 1.6427947598253274,
+ "grad_norm": 0.021673237904906273,
+ "learning_rate": 0.0006,
+ "loss": 6.931290149688721,
+ "step": 118
+ },
+ {
+ "epoch": 1.6567685589519652,
+ "grad_norm": 0.16024991869926453,
+ "learning_rate": 0.0006,
+ "loss": 6.979497909545898,
+ "step": 119
+ },
+ {
+ "epoch": 1.6707423580786025,
+ "grad_norm": 0.2434523105621338,
+ "learning_rate": 0.0006,
+ "loss": 7.185218811035156,
+ "step": 120
+ },
+ {
+ "epoch": 1.6847161572052403,
+ "grad_norm": 0.39252421259880066,
+ "learning_rate": 0.0006,
+ "loss": 7.137226581573486,
+ "step": 121
+ },
+ {
+ "epoch": 1.6986899563318776,
+ "grad_norm": 0.10565312206745148,
+ "learning_rate": 0.0006,
+ "loss": 6.984494209289551,
+ "step": 122
+ },
+ {
+ "epoch": 1.7126637554585153,
+ "grad_norm": 0.1033431813120842,
+ "learning_rate": 0.0006,
+ "loss": 7.001287937164307,
+ "step": 123
+ },
+ {
+ "epoch": 1.726637554585153,
+ "grad_norm": 0.14875490963459015,
+ "learning_rate": 0.0006,
+ "loss": 7.051436901092529,
+ "step": 124
+ },
+ {
+ "epoch": 1.7406113537117904,
+ "grad_norm": 0.1374562531709671,
+ "learning_rate": 0.0006,
+ "loss": 7.0293121337890625,
+ "step": 125
+ },
+ {
+ "epoch": 1.754585152838428,
+ "grad_norm": 0.0910211056470871,
+ "learning_rate": 0.0006,
+ "loss": 7.007458686828613,
+ "step": 126
+ },
+ {
+ "epoch": 1.7685589519650655,
+ "grad_norm": 0.021633487194776535,
+ "learning_rate": 0.0006,
+ "loss": 6.996816635131836,
+ "step": 127
+ },
+ {
+ "epoch": 1.782532751091703,
+ "grad_norm": 0.08692754060029984,
+ "learning_rate": 0.0006,
+ "loss": 6.998541831970215,
+ "step": 128
+ },
+ {
+ "epoch": 1.7965065502183406,
+ "grad_norm": 0.12613292038440704,
+ "learning_rate": 0.0006,
+ "loss": 6.9970703125,
+ "step": 129
+ },
+ {
+ "epoch": 1.8104803493449781,
+ "grad_norm": 0.09462810307741165,
+ "learning_rate": 0.0006,
+ "loss": 6.997045040130615,
+ "step": 130
+ },
+ {
+ "epoch": 1.8244541484716157,
+ "grad_norm": 0.03534962609410286,
+ "learning_rate": 0.0006,
+ "loss": 6.9738450050354,
+ "step": 131
+ },
+ {
+ "epoch": 1.8384279475982532,
+ "grad_norm": 0.033839885145425797,
+ "learning_rate": 0.0006,
+ "loss": 6.946001052856445,
+ "step": 132
+ },
+ {
+ "epoch": 1.8524017467248908,
+ "grad_norm": 0.06263269484043121,
+ "learning_rate": 0.0006,
+ "loss": 6.968981742858887,
+ "step": 133
+ },
+ {
+ "epoch": 1.8663755458515285,
+ "grad_norm": 0.07117997109889984,
+ "learning_rate": 0.0006,
+ "loss": 6.981535911560059,
+ "step": 134
+ },
+ {
+ "epoch": 1.8803493449781659,
+ "grad_norm": 0.06336277723312378,
+ "learning_rate": 0.0006,
+ "loss": 6.976946830749512,
+ "step": 135
+ },
+ {
+ "epoch": 1.8943231441048036,
+ "grad_norm": 0.047467734664678574,
+ "learning_rate": 0.0006,
+ "loss": 6.961673736572266,
+ "step": 136
+ },
+ {
+ "epoch": 1.908296943231441,
+ "grad_norm": 0.018787868320941925,
+ "learning_rate": 0.0006,
+ "loss": 6.948919296264648,
+ "step": 137
+ },
+ {
+ "epoch": 1.9222707423580787,
+ "grad_norm": 0.025909408926963806,
+ "learning_rate": 0.0006,
+ "loss": 6.976190567016602,
+ "step": 138
+ },
+ {
+ "epoch": 1.936244541484716,
+ "grad_norm": 0.052842333912849426,
+ "learning_rate": 0.0006,
+ "loss": 6.971207618713379,
+ "step": 139
+ },
+ {
+ "epoch": 1.9502183406113538,
+ "grad_norm": 0.06421888619661331,
+ "learning_rate": 0.0006,
+ "loss": 6.979109764099121,
+ "step": 140
+ },
+ {
+ "epoch": 1.9641921397379911,
+ "grad_norm": 0.055596910417079926,
+ "learning_rate": 0.0006,
+ "loss": 6.972118854522705,
+ "step": 141
+ },
+ {
+ "epoch": 1.9781659388646289,
+ "grad_norm": 0.03267091140151024,
+ "learning_rate": 0.0006,
+ "loss": 6.947260856628418,
+ "step": 142
+ },
+ {
+ "epoch": 1.9921397379912664,
+ "grad_norm": 0.008032613433897495,
+ "learning_rate": 0.0006,
+ "loss": 6.936816215515137,
+ "step": 143
+ },
+ {
+ "epoch": 2.0,
+ "grad_norm": 0.020184025168418884,
+ "learning_rate": 0.0006,
+ "loss": 6.929848670959473,
+ "step": 144
+ },
+ {
+ "epoch": 2.0,
+ "eval_loss": 7.836523056030273,
+ "eval_runtime": 58.5622,
+ "eval_samples_per_second": 41.699,
+ "eval_steps_per_second": 1.315,
+ "step": 144
+ },
+ {
+ "epoch": 2.0139737991266378,
+ "grad_norm": 0.031980376690626144,
+ "learning_rate": 0.0006,
+ "loss": 6.968081474304199,
+ "step": 145
+ },
+ {
+ "epoch": 2.027947598253275,
+ "grad_norm": 0.03802533075213432,
+ "learning_rate": 0.0006,
+ "loss": 6.95728874206543,
+ "step": 146
+ },
+ {
+ "epoch": 2.041921397379913,
+ "grad_norm": 0.03796226158738136,
+ "learning_rate": 0.0006,
+ "loss": 6.96516752243042,
+ "step": 147
+ },
+ {
+ "epoch": 2.05589519650655,
+ "grad_norm": 0.02621951326727867,
+ "learning_rate": 0.0006,
+ "loss": 6.956058502197266,
+ "step": 148
+ },
+ {
+ "epoch": 2.069868995633188,
+ "grad_norm": 0.013292718678712845,
+ "learning_rate": 0.0006,
+ "loss": 6.9783935546875,
+ "step": 149
+ },
+ {
+ "epoch": 2.0838427947598253,
+ "grad_norm": 0.009446968324482441,
+ "learning_rate": 0.0006,
+ "loss": 6.96873664855957,
+ "step": 150
+ },
+ {
+ "epoch": 2.097816593886463,
+ "grad_norm": 0.021703092381358147,
+ "learning_rate": 0.0006,
+ "loss": 6.959234237670898,
+ "step": 151
+ },
+ {
+ "epoch": 2.1117903930131003,
+ "grad_norm": 0.027765290811657906,
+ "learning_rate": 0.0006,
+ "loss": 6.968524932861328,
+ "step": 152
+ },
+ {
+ "epoch": 2.125764192139738,
+ "grad_norm": 0.028761165216565132,
+ "learning_rate": 0.0006,
+ "loss": 6.954620838165283,
+ "step": 153
+ },
+ {
+ "epoch": 2.1397379912663754,
+ "grad_norm": 0.019729383289813995,
+ "learning_rate": 0.0006,
+ "loss": 6.949009895324707,
+ "step": 154
+ },
+ {
+ "epoch": 2.153711790393013,
+ "grad_norm": 0.008990089409053326,
+ "learning_rate": 0.0006,
+ "loss": 6.956226348876953,
+ "step": 155
+ },
+ {
+ "epoch": 2.1676855895196505,
+ "grad_norm": 0.00549644511193037,
+ "learning_rate": 0.0006,
+ "loss": 6.957143783569336,
+ "step": 156
+ },
+ {
+ "epoch": 2.1816593886462883,
+ "grad_norm": 0.014436670579016209,
+ "learning_rate": 0.0006,
+ "loss": 6.954257011413574,
+ "step": 157
+ },
+ {
+ "epoch": 2.1956331877729256,
+ "grad_norm": 0.01990494504570961,
+ "learning_rate": 0.0006,
+ "loss": 6.942131519317627,
+ "step": 158
+ },
+ {
+ "epoch": 2.2096069868995634,
+ "grad_norm": 0.019155776128172874,
+ "learning_rate": 0.0006,
+ "loss": 6.946969985961914,
+ "step": 159
+ },
+ {
+ "epoch": 2.2235807860262007,
+ "grad_norm": 0.014970783144235611,
+ "learning_rate": 0.0006,
+ "loss": 6.957938194274902,
+ "step": 160
+ },
+ {
+ "epoch": 2.2375545851528384,
+ "grad_norm": 0.009875659830868244,
+ "learning_rate": 0.0006,
+ "loss": 6.953309059143066,
+ "step": 161
+ },
+ {
+ "epoch": 2.251528384279476,
+ "grad_norm": 0.004792087245732546,
+ "learning_rate": 0.0006,
+ "loss": 6.975146770477295,
+ "step": 162
+ },
+ {
+ "epoch": 2.2655021834061135,
+ "grad_norm": 0.009936527349054813,
+ "learning_rate": 0.0006,
+ "loss": 6.961674690246582,
+ "step": 163
+ },
+ {
+ "epoch": 2.279475982532751,
+ "grad_norm": 0.015100127086043358,
+ "learning_rate": 0.0006,
+ "loss": 6.968759536743164,
+ "step": 164
+ },
+ {
+ "epoch": 2.2934497816593886,
+ "grad_norm": 0.01449542585760355,
+ "learning_rate": 0.0006,
+ "loss": 6.966510772705078,
+ "step": 165
+ },
+ {
+ "epoch": 2.3074235807860264,
+ "grad_norm": 0.01073537115007639,
+ "learning_rate": 0.0006,
+ "loss": 6.936373710632324,
+ "step": 166
+ },
+ {
+ "epoch": 2.3213973799126637,
+ "grad_norm": 0.006766272243112326,
+ "learning_rate": 0.0006,
+ "loss": 6.96397590637207,
+ "step": 167
+ },
+ {
+ "epoch": 2.3353711790393015,
+ "grad_norm": 0.005055439658463001,
+ "learning_rate": 0.0006,
+ "loss": 6.961050987243652,
+ "step": 168
+ },
+ {
+ "epoch": 2.349344978165939,
+ "grad_norm": 0.006252622697502375,
+ "learning_rate": 0.0006,
+ "loss": 6.980506420135498,
+ "step": 169
+ },
+ {
+ "epoch": 2.3633187772925766,
+ "grad_norm": 0.008084346540272236,
+ "learning_rate": 0.0006,
+ "loss": 6.9346604347229,
+ "step": 170
+ },
+ {
+ "epoch": 2.377292576419214,
+ "grad_norm": 0.010260380804538727,
+ "learning_rate": 0.0006,
+ "loss": 6.964118957519531,
+ "step": 171
+ },
+ {
+ "epoch": 2.3912663755458516,
+ "grad_norm": 0.010697875171899796,
+ "learning_rate": 0.0006,
+ "loss": 6.941572189331055,
+ "step": 172
+ },
+ {
+ "epoch": 2.405240174672489,
+ "grad_norm": 0.007338830269873142,
+ "learning_rate": 0.0006,
+ "loss": 6.950471878051758,
+ "step": 173
+ },
+ {
+ "epoch": 2.4192139737991267,
+ "grad_norm": 0.005931832827627659,
+ "learning_rate": 0.0006,
+ "loss": 6.962935924530029,
+ "step": 174
+ },
+ {
+ "epoch": 2.433187772925764,
+ "grad_norm": 0.0061243255622684956,
+ "learning_rate": 0.0006,
+ "loss": 6.946418285369873,
+ "step": 175
+ },
+ {
+ "epoch": 2.447161572052402,
+ "grad_norm": 0.007983328774571419,
+ "learning_rate": 0.0006,
+ "loss": 6.959758758544922,
+ "step": 176
+ },
+ {
+ "epoch": 2.461135371179039,
+ "grad_norm": 0.010470672510564327,
+ "learning_rate": 0.0006,
+ "loss": 6.93851375579834,
+ "step": 177
+ },
+ {
+ "epoch": 2.475109170305677,
+ "grad_norm": 0.011147388257086277,
+ "learning_rate": 0.0006,
+ "loss": 6.93258810043335,
+ "step": 178
+ },
+ {
+ "epoch": 2.489082969432314,
+ "grad_norm": 0.01058544497936964,
+ "learning_rate": 0.0006,
+ "loss": 6.932259559631348,
+ "step": 179
+ },
+ {
+ "epoch": 2.503056768558952,
+ "grad_norm": 0.016850745305418968,
+ "learning_rate": 0.0006,
+ "loss": 6.924542427062988,
+ "step": 180
+ },
+ {
+ "epoch": 2.5170305676855893,
+ "grad_norm": 0.023101547732949257,
+ "learning_rate": 0.0006,
+ "loss": 6.914878845214844,
+ "step": 181
+ },
+ {
+ "epoch": 2.531004366812227,
+ "grad_norm": 0.015795735642313957,
+ "learning_rate": 0.0006,
+ "loss": 6.922432899475098,
+ "step": 182
+ },
+ {
+ "epoch": 2.544978165938865,
+ "grad_norm": 0.008821426890790462,
+ "learning_rate": 0.0006,
+ "loss": 6.924562454223633,
+ "step": 183
+ },
+ {
+ "epoch": 2.558951965065502,
+ "grad_norm": 0.012502548284828663,
+ "learning_rate": 0.0006,
+ "loss": 6.914718151092529,
+ "step": 184
+ },
+ {
+ "epoch": 2.5729257641921395,
+ "grad_norm": 0.011545066721737385,
+ "learning_rate": 0.0006,
+ "loss": 6.915705680847168,
+ "step": 185
+ },
+ {
+ "epoch": 2.5868995633187772,
+ "grad_norm": 0.022190893068909645,
+ "learning_rate": 0.0006,
+ "loss": 6.88851261138916,
+ "step": 186
+ },
+ {
+ "epoch": 2.600873362445415,
+ "grad_norm": 0.023395851254463196,
+ "learning_rate": 0.0006,
+ "loss": 6.877655029296875,
+ "step": 187
+ },
+ {
+ "epoch": 2.6148471615720523,
+ "grad_norm": 0.015601156279444695,
+ "learning_rate": 0.0006,
+ "loss": 6.889644145965576,
+ "step": 188
+ },
+ {
+ "epoch": 2.62882096069869,
+ "grad_norm": 0.01362934336066246,
+ "learning_rate": 0.0006,
+ "loss": 6.891547203063965,
+ "step": 189
+ },
+ {
+ "epoch": 2.6427947598253274,
+ "grad_norm": 0.012305490672588348,
+ "learning_rate": 0.0006,
+ "loss": 6.890357971191406,
+ "step": 190
+ },
+ {
+ "epoch": 2.656768558951965,
+ "grad_norm": 0.01623358018696308,
+ "learning_rate": 0.0006,
+ "loss": 6.849443435668945,
+ "step": 191
+ },
+ {
+ "epoch": 2.6707423580786025,
+ "grad_norm": 0.018063291907310486,
+ "learning_rate": 0.0006,
+ "loss": 6.842372894287109,
+ "step": 192
+ },
+ {
+ "epoch": 2.6847161572052403,
+ "grad_norm": 0.010179434902966022,
+ "learning_rate": 0.0006,
+ "loss": 6.870055198669434,
+ "step": 193
+ },
+ {
+ "epoch": 2.6986899563318776,
+ "grad_norm": 0.009986722841858864,
+ "learning_rate": 0.0006,
+ "loss": 6.859061241149902,
+ "step": 194
+ },
+ {
+ "epoch": 2.7126637554585153,
+ "grad_norm": 0.01117654051631689,
+ "learning_rate": 0.0006,
+ "loss": 6.861199378967285,
+ "step": 195
+ },
+ {
+ "epoch": 2.726637554585153,
+ "grad_norm": 0.008551654405891895,
+ "learning_rate": 0.0006,
+ "loss": 6.850255012512207,
+ "step": 196
+ },
+ {
+ "epoch": 2.7406113537117904,
+ "grad_norm": 0.018904326483607292,
+ "learning_rate": 0.0006,
+ "loss": 6.848639965057373,
+ "step": 197
+ },
+ {
+ "epoch": 2.7545851528384278,
+ "grad_norm": 0.02080724574625492,
+ "learning_rate": 0.0006,
+ "loss": 6.80456018447876,
+ "step": 198
+ },
+ {
+ "epoch": 2.7685589519650655,
+ "grad_norm": 0.012010873295366764,
+ "learning_rate": 0.0006,
+ "loss": 6.848775863647461,
+ "step": 199
+ },
+ {
+ "epoch": 2.7825327510917033,
+ "grad_norm": 0.008113056421279907,
+ "learning_rate": 0.0006,
+ "loss": 6.849387168884277,
+ "step": 200
+ },
+ {
+ "epoch": 2.7965065502183406,
+ "grad_norm": 0.012912592850625515,
+ "learning_rate": 0.0006,
+ "loss": 6.832240104675293,
+ "step": 201
+ },
+ {
+ "epoch": 2.810480349344978,
+ "grad_norm": 0.01146865077316761,
+ "learning_rate": 0.0006,
+ "loss": 6.850383281707764,
+ "step": 202
+ },
+ {
+ "epoch": 2.8244541484716157,
+ "grad_norm": 0.0171892661601305,
+ "learning_rate": 0.0006,
+ "loss": 6.839577674865723,
+ "step": 203
+ },
+ {
+ "epoch": 2.8384279475982535,
+ "grad_norm": 0.012083170004189014,
+ "learning_rate": 0.0006,
+ "loss": 6.818614482879639,
+ "step": 204
+ },
+ {
+ "epoch": 2.8524017467248908,
+ "grad_norm": 0.0062555656768381596,
+ "learning_rate": 0.0006,
+ "loss": 6.831291198730469,
+ "step": 205
+ },
+ {
+ "epoch": 2.8663755458515285,
+ "grad_norm": 0.010102466680109501,
+ "learning_rate": 0.0006,
+ "loss": 6.846522808074951,
+ "step": 206
+ },
+ {
+ "epoch": 2.880349344978166,
+ "grad_norm": 0.009042087942361832,
+ "learning_rate": 0.0006,
+ "loss": 6.822072982788086,
+ "step": 207
+ },
+ {
+ "epoch": 2.8943231441048036,
+ "grad_norm": 0.011133556254208088,
+ "learning_rate": 0.0006,
+ "loss": 6.811252593994141,
+ "step": 208
+ },
+ {
+ "epoch": 2.908296943231441,
+ "grad_norm": 0.00990577507764101,
+ "learning_rate": 0.0006,
+ "loss": 6.814435958862305,
+ "step": 209
+ },
+ {
+ "epoch": 2.9222707423580787,
+ "grad_norm": 0.006605073343962431,
+ "learning_rate": 0.0006,
+ "loss": 6.831300735473633,
+ "step": 210
+ },
+ {
+ "epoch": 2.936244541484716,
+ "grad_norm": 0.007831175811588764,
+ "learning_rate": 0.0006,
+ "loss": 6.841738700866699,
+ "step": 211
+ },
+ {
+ "epoch": 2.950218340611354,
+ "grad_norm": 0.005536255426704884,
+ "learning_rate": 0.0006,
+ "loss": 6.8043670654296875,
+ "step": 212
+ },
+ {
+ "epoch": 2.964192139737991,
+ "grad_norm": 0.006532667670398951,
+ "learning_rate": 0.0006,
+ "loss": 6.8091840744018555,
+ "step": 213
+ },
+ {
+ "epoch": 2.978165938864629,
+ "grad_norm": 0.006674565840512514,
+ "learning_rate": 0.0006,
+ "loss": 6.811327934265137,
+ "step": 214
+ },
+ {
+ "epoch": 2.992139737991266,
+ "grad_norm": 0.01150344219058752,
+ "learning_rate": 0.0006,
+ "loss": 6.810762405395508,
+ "step": 215
+ },
+ {
+ "epoch": 3.0,
+ "grad_norm": 0.009015528485178947,
+ "learning_rate": 0.0006,
+ "loss": 6.836431980133057,
+ "step": 216
+ },
+ {
+ "epoch": 3.0,
+ "eval_loss": 7.718832969665527,
+ "eval_runtime": 58.6615,
+ "eval_samples_per_second": 41.629,
+ "eval_steps_per_second": 1.313,
+ "step": 216
+ },
+ {
+ "epoch": 3.0139737991266378,
+ "grad_norm": 0.009791829623281956,
+ "learning_rate": 0.0006,
+ "loss": 6.8102593421936035,
+ "step": 217
+ },
+ {
+ "epoch": 3.027947598253275,
+ "grad_norm": 0.008970947004854679,
+ "learning_rate": 0.0006,
+ "loss": 6.792663097381592,
+ "step": 218
+ },
+ {
+ "epoch": 3.041921397379913,
+ "grad_norm": 0.013689392246305943,
+ "learning_rate": 0.0006,
+ "loss": 6.815651893615723,
+ "step": 219
+ },
+ {
+ "epoch": 3.05589519650655,
+ "grad_norm": 0.031499531120061874,
+ "learning_rate": 0.0006,
+ "loss": 6.809841156005859,
+ "step": 220
+ },
+ {
+ "epoch": 3.069868995633188,
+ "grad_norm": 0.03203094005584717,
+ "learning_rate": 0.0006,
+ "loss": 6.779417037963867,
+ "step": 221
+ },
+ {
+ "epoch": 3.0838427947598253,
+ "grad_norm": 0.007947996258735657,
+ "learning_rate": 0.0006,
+ "loss": 6.773656845092773,
+ "step": 222
+ },
+ {
+ "epoch": 3.097816593886463,
+ "grad_norm": 0.03152519091963768,
+ "learning_rate": 0.0006,
+ "loss": 6.804083824157715,
+ "step": 223
+ },
+ {
+ "epoch": 3.1117903930131003,
+ "grad_norm": 0.014937258325517178,
+ "learning_rate": 0.0006,
+ "loss": 6.792213439941406,
+ "step": 224
+ },
+ {
+ "epoch": 3.125764192139738,
+ "grad_norm": 0.020961705595254898,
+ "learning_rate": 0.0006,
+ "loss": 6.8092803955078125,
+ "step": 225
+ },
+ {
+ "epoch": 3.1397379912663754,
+ "grad_norm": 0.009152988903224468,
+ "learning_rate": 0.0006,
+ "loss": 6.7737603187561035,
+ "step": 226
+ },
+ {
+ "epoch": 3.153711790393013,
+ "grad_norm": 0.02770620957016945,
+ "learning_rate": 0.0006,
+ "loss": 6.767014503479004,
+ "step": 227
+ },
+ {
+ "epoch": 3.1676855895196505,
+ "grad_norm": 0.01165871974080801,
+ "learning_rate": 0.0006,
+ "loss": 6.760884761810303,
+ "step": 228
+ },
+ {
+ "epoch": 3.1816593886462883,
+ "grad_norm": 0.01910376362502575,
+ "learning_rate": 0.0006,
+ "loss": 6.775722980499268,
+ "step": 229
+ },
+ {
+ "epoch": 3.1956331877729256,
+ "grad_norm": 0.010413747280836105,
+ "learning_rate": 0.0006,
+ "loss": 6.779816627502441,
+ "step": 230
+ },
+ {
+ "epoch": 3.2096069868995634,
+ "grad_norm": 0.01918417401611805,
+ "learning_rate": 0.0006,
+ "loss": 6.760381698608398,
+ "step": 231
+ },
+ {
+ "epoch": 3.2235807860262007,
+ "grad_norm": 0.017960334196686745,
+ "learning_rate": 0.0006,
+ "loss": 6.761198997497559,
+ "step": 232
+ },
+ {
+ "epoch": 3.2375545851528384,
+ "grad_norm": 0.016546795144677162,
+ "learning_rate": 0.0006,
+ "loss": 6.742448806762695,
+ "step": 233
+ },
+ {
+ "epoch": 3.251528384279476,
+ "grad_norm": 0.01889597810804844,
+ "learning_rate": 0.0006,
+ "loss": 6.758237838745117,
+ "step": 234
+ },
+ {
+ "epoch": 3.2655021834061135,
+ "grad_norm": 0.009611992165446281,
+ "learning_rate": 0.0006,
+ "loss": 6.75614070892334,
+ "step": 235
+ },
+ {
+ "epoch": 3.279475982532751,
+ "grad_norm": 0.02034934051334858,
+ "learning_rate": 0.0006,
+ "loss": 6.736965179443359,
+ "step": 236
+ },
+ {
+ "epoch": 3.2934497816593886,
+ "grad_norm": 0.010740065947175026,
+ "learning_rate": 0.0006,
+ "loss": 6.746947288513184,
+ "step": 237
+ },
+ {
+ "epoch": 3.3074235807860264,
+ "grad_norm": 0.018383340910077095,
+ "learning_rate": 0.0006,
+ "loss": 6.747332572937012,
+ "step": 238
+ },
+ {
+ "epoch": 3.3213973799126637,
+ "grad_norm": 0.012362911365926266,
+ "learning_rate": 0.0006,
+ "loss": 6.767669200897217,
+ "step": 239
+ },
+ {
+ "epoch": 3.3353711790393015,
+ "grad_norm": 0.01127463299781084,
+ "learning_rate": 0.0006,
+ "loss": 6.769808292388916,
+ "step": 240
+ },
+ {
+ "epoch": 3.349344978165939,
+ "grad_norm": 0.017693769186735153,
+ "learning_rate": 0.0006,
+ "loss": 6.733658790588379,
+ "step": 241
+ },
+ {
+ "epoch": 3.3633187772925766,
+ "grad_norm": 0.006420376244932413,
+ "learning_rate": 0.0006,
+ "loss": 6.74934720993042,
+ "step": 242
+ },
+ {
+ "epoch": 3.377292576419214,
+ "grad_norm": 0.014020364731550217,
+ "learning_rate": 0.0006,
+ "loss": 6.728157043457031,
+ "step": 243
+ },
+ {
+ "epoch": 3.3912663755458516,
+ "grad_norm": 0.008018582127988338,
+ "learning_rate": 0.0006,
+ "loss": 6.730726718902588,
+ "step": 244
+ },
+ {
+ "epoch": 3.405240174672489,
+ "grad_norm": 0.00782222580164671,
+ "learning_rate": 0.0006,
+ "loss": 6.7499895095825195,
+ "step": 245
+ },
+ {
+ "epoch": 3.4192139737991267,
+ "grad_norm": 0.008278409950435162,
+ "learning_rate": 0.0006,
+ "loss": 6.712803363800049,
+ "step": 246
+ },
+ {
+ "epoch": 3.433187772925764,
+ "grad_norm": 0.009826759807765484,
+ "learning_rate": 0.0006,
+ "loss": 6.742262840270996,
+ "step": 247
+ },
+ {
+ "epoch": 3.447161572052402,
+ "grad_norm": 0.009488224983215332,
+ "learning_rate": 0.0006,
+ "loss": 6.739130020141602,
+ "step": 248
+ },
+ {
+ "epoch": 3.461135371179039,
+ "grad_norm": 0.007667903788387775,
+ "learning_rate": 0.0006,
+ "loss": 6.747515678405762,
+ "step": 249
+ },
+ {
+ "epoch": 3.475109170305677,
+ "grad_norm": 0.0073090000078082085,
+ "learning_rate": 0.0006,
+ "loss": 6.740870475769043,
+ "step": 250
+ },
+ {
+ "epoch": 3.489082969432314,
+ "grad_norm": 0.010058708488941193,
+ "learning_rate": 0.0006,
+ "loss": 6.777563095092773,
+ "step": 251
+ },
+ {
+ "epoch": 3.503056768558952,
+ "grad_norm": 0.020676063373684883,
+ "learning_rate": 0.0006,
+ "loss": 6.708603858947754,
+ "step": 252
+ },
+ {
+ "epoch": 3.5170305676855893,
+ "grad_norm": 0.027210544794797897,
+ "learning_rate": 0.0006,
+ "loss": 6.694097995758057,
+ "step": 253
+ },
+ {
+ "epoch": 3.531004366812227,
+ "grad_norm": 0.038053520023822784,
+ "learning_rate": 0.0006,
+ "loss": 6.697017669677734,
+ "step": 254
+ },
+ {
+ "epoch": 3.544978165938865,
+ "grad_norm": 0.04684312269091606,
+ "learning_rate": 0.0006,
+ "loss": 6.729883193969727,
+ "step": 255
+ },
+ {
+ "epoch": 3.558951965065502,
+ "grad_norm": 0.03673060983419418,
+ "learning_rate": 0.0006,
+ "loss": 6.743753433227539,
+ "step": 256
+ },
+ {
+ "epoch": 3.5729257641921395,
+ "grad_norm": 0.006882632151246071,
+ "learning_rate": 0.0006,
+ "loss": 6.71405029296875,
+ "step": 257
+ },
+ {
+ "epoch": 3.5868995633187772,
+ "grad_norm": 0.03588201105594635,
+ "learning_rate": 0.0006,
+ "loss": 6.712467193603516,
+ "step": 258
+ },
+ {
+ "epoch": 3.600873362445415,
+ "grad_norm": 0.029415199533104897,
+ "learning_rate": 0.0006,
+ "loss": 6.715634822845459,
+ "step": 259
+ },
+ {
+ "epoch": 3.6148471615720523,
+ "grad_norm": 0.011425039730966091,
+ "learning_rate": 0.0006,
+ "loss": 6.697334289550781,
+ "step": 260
+ },
+ {
+ "epoch": 3.62882096069869,
+ "grad_norm": 0.0297263003885746,
+ "learning_rate": 0.0006,
+ "loss": 6.724848747253418,
+ "step": 261
+ },
+ {
+ "epoch": 3.6427947598253274,
+ "grad_norm": 0.007282314356416464,
+ "learning_rate": 0.0006,
+ "loss": 6.702306270599365,
+ "step": 262
+ },
+ {
+ "epoch": 3.656768558951965,
+ "grad_norm": 0.02673051692545414,
+ "learning_rate": 0.0006,
+ "loss": 6.72357177734375,
+ "step": 263
+ },
+ {
+ "epoch": 3.6707423580786025,
+ "grad_norm": 0.016174081712961197,
+ "learning_rate": 0.0006,
+ "loss": 6.730698585510254,
+ "step": 264
+ },
+ {
+ "epoch": 3.6847161572052403,
+ "grad_norm": 0.0182416420429945,
+ "learning_rate": 0.0006,
+ "loss": 6.717386245727539,
+ "step": 265
+ },
+ {
+ "epoch": 3.6986899563318776,
+ "grad_norm": 0.023250866681337357,
+ "learning_rate": 0.0006,
+ "loss": 6.697832107543945,
+ "step": 266
+ },
+ {
+ "epoch": 3.7126637554585153,
+ "grad_norm": 0.007539753336459398,
+ "learning_rate": 0.0006,
+ "loss": 6.714348793029785,
+ "step": 267
+ },
+ {
+ "epoch": 3.726637554585153,
+ "grad_norm": 0.030086630955338478,
+ "learning_rate": 0.0006,
+ "loss": 6.724481582641602,
+ "step": 268
+ },
+ {
+ "epoch": 3.7406113537117904,
+ "grad_norm": 0.012784913182258606,
+ "learning_rate": 0.0006,
+ "loss": 6.703491687774658,
+ "step": 269
+ },
+ {
+ "epoch": 3.7545851528384278,
+ "grad_norm": 0.02042769268155098,
+ "learning_rate": 0.0006,
+ "loss": 6.721139907836914,
+ "step": 270
+ },
+ {
+ "epoch": 3.7685589519650655,
+ "grad_norm": 0.023557329550385475,
+ "learning_rate": 0.0006,
+ "loss": 6.728122711181641,
+ "step": 271
+ },
+ {
+ "epoch": 3.7825327510917033,
+ "grad_norm": 0.007716621737927198,
+ "learning_rate": 0.0006,
+ "loss": 6.70085334777832,
+ "step": 272
+ },
+ {
+ "epoch": 3.7965065502183406,
+ "grad_norm": 0.022772278636693954,
+ "learning_rate": 0.0006,
+ "loss": 6.712882995605469,
+ "step": 273
+ },
+ {
+ "epoch": 3.810480349344978,
+ "grad_norm": 0.007928933948278427,
+ "learning_rate": 0.0006,
+ "loss": 6.698320388793945,
+ "step": 274
+ },
+ {
+ "epoch": 3.8244541484716157,
+ "grad_norm": 0.022355245426297188,
+ "learning_rate": 0.0006,
+ "loss": 6.687088489532471,
+ "step": 275
+ },
+ {
+ "epoch": 3.8384279475982535,
+ "grad_norm": 0.02037850394845009,
+ "learning_rate": 0.0006,
+ "loss": 6.676856994628906,
+ "step": 276
+ },
+ {
+ "epoch": 3.8524017467248908,
+ "grad_norm": 0.012290362268686295,
+ "learning_rate": 0.0006,
+ "loss": 6.682883262634277,
+ "step": 277
+ },
+ {
+ "epoch": 3.8663755458515285,
+ "grad_norm": 0.022331275045871735,
+ "learning_rate": 0.0006,
+ "loss": 6.694752216339111,
+ "step": 278
+ },
+ {
+ "epoch": 3.880349344978166,
+ "grad_norm": 0.007527181413024664,
+ "learning_rate": 0.0006,
+ "loss": 6.686394691467285,
+ "step": 279
+ },
+ {
+ "epoch": 3.8943231441048036,
+ "grad_norm": 0.02159409038722515,
+ "learning_rate": 0.0006,
+ "loss": 6.697943687438965,
+ "step": 280
+ },
+ {
+ "epoch": 3.908296943231441,
+ "grad_norm": 0.015054763294756413,
+ "learning_rate": 0.0006,
+ "loss": 6.677796363830566,
+ "step": 281
+ },
+ {
+ "epoch": 3.9222707423580787,
+ "grad_norm": 0.01466019731014967,
+ "learning_rate": 0.0006,
+ "loss": 6.697715759277344,
+ "step": 282
+ },
+ {
+ "epoch": 3.936244541484716,
+ "grad_norm": 0.01875465176999569,
+ "learning_rate": 0.0006,
+ "loss": 6.700228691101074,
+ "step": 283
+ },
+ {
+ "epoch": 3.950218340611354,
+ "grad_norm": 0.00778991961851716,
+ "learning_rate": 0.0006,
+ "loss": 6.69331693649292,
+ "step": 284
+ },
+ {
+ "epoch": 3.964192139737991,
+ "grad_norm": 0.020182441920042038,
+ "learning_rate": 0.0006,
+ "loss": 6.672651290893555,
+ "step": 285
+ },
+ {
+ "epoch": 3.978165938864629,
+ "grad_norm": 0.01240516733378172,
+ "learning_rate": 0.0006,
+ "loss": 6.688860893249512,
+ "step": 286
+ },
+ {
+ "epoch": 3.992139737991266,
+ "grad_norm": 0.011218412779271603,
+ "learning_rate": 0.0006,
+ "loss": 6.669534683227539,
+ "step": 287
+ },
+ {
+ "epoch": 4.0,
+ "grad_norm": 0.0141852255910635,
+ "learning_rate": 0.0006,
+ "loss": 6.690206527709961,
+ "step": 288
+ },
+ {
+ "epoch": 4.0,
+ "eval_loss": 7.537620544433594,
+ "eval_runtime": 58.4398,
+ "eval_samples_per_second": 41.787,
+ "eval_steps_per_second": 1.318,
+ "step": 288
+ },
+ {
+ "epoch": 4.013973799126638,
+ "grad_norm": 0.008106146939098835,
+ "learning_rate": 0.0006,
+ "loss": 6.688591003417969,
+ "step": 289
+ },
+ {
+ "epoch": 4.0279475982532755,
+ "grad_norm": 0.014588244259357452,
+ "learning_rate": 0.0006,
+ "loss": 6.684961318969727,
+ "step": 290
+ },
+ {
+ "epoch": 4.041921397379912,
+ "grad_norm": 0.011985368095338345,
+ "learning_rate": 0.0006,
+ "loss": 6.672794342041016,
+ "step": 291
+ },
+ {
+ "epoch": 4.05589519650655,
+ "grad_norm": 0.014632102102041245,
+ "learning_rate": 0.0006,
+ "loss": 6.702736854553223,
+ "step": 292
+ },
+ {
+ "epoch": 4.069868995633188,
+ "grad_norm": 0.011243713088333607,
+ "learning_rate": 0.0006,
+ "loss": 6.658792495727539,
+ "step": 293
+ },
+ {
+ "epoch": 4.083842794759826,
+ "grad_norm": 0.010519202798604965,
+ "learning_rate": 0.0006,
+ "loss": 6.662495136260986,
+ "step": 294
+ },
+ {
+ "epoch": 4.097816593886463,
+ "grad_norm": 0.007318088319152594,
+ "learning_rate": 0.0006,
+ "loss": 6.675954818725586,
+ "step": 295
+ },
+ {
+ "epoch": 4.1117903930131,
+ "grad_norm": 0.011712398380041122,
+ "learning_rate": 0.0006,
+ "loss": 6.673973083496094,
+ "step": 296
+ },
+ {
+ "epoch": 4.125764192139738,
+ "grad_norm": 0.006726635154336691,
+ "learning_rate": 0.0006,
+ "loss": 6.656833171844482,
+ "step": 297
+ },
+ {
+ "epoch": 4.139737991266376,
+ "grad_norm": 0.005466983187943697,
+ "learning_rate": 0.0006,
+ "loss": 6.675468921661377,
+ "step": 298
+ },
+ {
+ "epoch": 4.153711790393013,
+ "grad_norm": 0.007163621485233307,
+ "learning_rate": 0.0006,
+ "loss": 6.6862874031066895,
+ "step": 299
+ },
+ {
+ "epoch": 4.1676855895196505,
+ "grad_norm": 0.007626629434525967,
+ "learning_rate": 0.0006,
+ "loss": 6.654741287231445,
+ "step": 300
+ },
+ {
+ "epoch": 4.181659388646288,
+ "grad_norm": 0.0066330647096037865,
+ "learning_rate": 0.0006,
+ "loss": 6.68114709854126,
+ "step": 301
+ },
+ {
+ "epoch": 4.195633187772926,
+ "grad_norm": 0.007363482844084501,
+ "learning_rate": 0.0006,
+ "loss": 6.683980941772461,
+ "step": 302
+ },
+ {
+ "epoch": 4.209606986899563,
+ "grad_norm": 0.00971233006566763,
+ "learning_rate": 0.0006,
+ "loss": 6.669700622558594,
+ "step": 303
+ },
+ {
+ "epoch": 4.223580786026201,
+ "grad_norm": 0.013185256160795689,
+ "learning_rate": 0.0006,
+ "loss": 6.686490535736084,
+ "step": 304
+ },
+ {
+ "epoch": 4.2375545851528384,
+ "grad_norm": 0.019116027280688286,
+ "learning_rate": 0.0006,
+ "loss": 6.645479202270508,
+ "step": 305
+ },
+ {
+ "epoch": 4.251528384279476,
+ "grad_norm": 0.0328814722597599,
+ "learning_rate": 0.0006,
+ "loss": 6.68080997467041,
+ "step": 306
+ },
+ {
+ "epoch": 4.265502183406113,
+ "grad_norm": 0.04482287913560867,
+ "learning_rate": 0.0006,
+ "loss": 6.698016166687012,
+ "step": 307
+ },
+ {
+ "epoch": 4.279475982532751,
+ "grad_norm": 0.03053313121199608,
+ "learning_rate": 0.0006,
+ "loss": 6.642240524291992,
+ "step": 308
+ },
+ {
+ "epoch": 4.293449781659389,
+ "grad_norm": 0.043351784348487854,
+ "learning_rate": 0.0006,
+ "loss": 6.662343978881836,
+ "step": 309
+ },
+ {
+ "epoch": 4.307423580786026,
+ "grad_norm": 0.03165886923670769,
+ "learning_rate": 0.0006,
+ "loss": 6.6291961669921875,
+ "step": 310
+ },
+ {
+ "epoch": 4.321397379912664,
+ "grad_norm": 0.028601787984371185,
+ "learning_rate": 0.0006,
+ "loss": 6.667544364929199,
+ "step": 311
+ },
+ {
+ "epoch": 4.335371179039301,
+ "grad_norm": 0.019558480009436607,
+ "learning_rate": 0.0006,
+ "loss": 6.641025543212891,
+ "step": 312
+ },
+ {
+ "epoch": 4.349344978165939,
+ "grad_norm": 0.039851363748311996,
+ "learning_rate": 0.0006,
+ "loss": 6.665952205657959,
+ "step": 313
+ },
+ {
+ "epoch": 4.3633187772925766,
+ "grad_norm": 0.018417146056890488,
+ "learning_rate": 0.0006,
+ "loss": 6.657257556915283,
+ "step": 314
+ },
+ {
+ "epoch": 4.377292576419214,
+ "grad_norm": 0.02894527092576027,
+ "learning_rate": 0.0006,
+ "loss": 6.656288146972656,
+ "step": 315
+ },
+ {
+ "epoch": 4.391266375545851,
+ "grad_norm": 0.01910966821014881,
+ "learning_rate": 0.0006,
+ "loss": 6.688958168029785,
+ "step": 316
+ },
+ {
+ "epoch": 4.405240174672489,
+ "grad_norm": 0.01668412610888481,
+ "learning_rate": 0.0006,
+ "loss": 6.64876127243042,
+ "step": 317
+ },
+ {
+ "epoch": 4.419213973799127,
+ "grad_norm": 0.02935275249183178,
+ "learning_rate": 0.0006,
+ "loss": 6.626096725463867,
+ "step": 318
+ },
+ {
+ "epoch": 4.4331877729257645,
+ "grad_norm": 0.017997972667217255,
+ "learning_rate": 0.0006,
+ "loss": 6.640386581420898,
+ "step": 319
+ },
+ {
+ "epoch": 4.447161572052401,
+ "grad_norm": 0.01779189519584179,
+ "learning_rate": 0.0006,
+ "loss": 6.641165256500244,
+ "step": 320
+ },
+ {
+ "epoch": 4.461135371179039,
+ "grad_norm": 0.013901712372899055,
+ "learning_rate": 0.0006,
+ "loss": 6.651627063751221,
+ "step": 321
+ },
+ {
+ "epoch": 4.475109170305677,
+ "grad_norm": 0.013211199082434177,
+ "learning_rate": 0.0006,
+ "loss": 6.630064964294434,
+ "step": 322
+ },
+ {
+ "epoch": 4.489082969432315,
+ "grad_norm": 0.02324107475578785,
+ "learning_rate": 0.0006,
+ "loss": 6.670443058013916,
+ "step": 323
+ },
+ {
+ "epoch": 4.503056768558952,
+ "grad_norm": 0.019452504813671112,
+ "learning_rate": 0.0006,
+ "loss": 6.651266098022461,
+ "step": 324
+ },
+ {
+ "epoch": 4.517030567685589,
+ "grad_norm": 0.017943572252988815,
+ "learning_rate": 0.0006,
+ "loss": 6.657943248748779,
+ "step": 325
+ },
+ {
+ "epoch": 4.531004366812227,
+ "grad_norm": 0.012136237695813179,
+ "learning_rate": 0.0006,
+ "loss": 6.640082359313965,
+ "step": 326
+ },
+ {
+ "epoch": 4.544978165938865,
+ "grad_norm": 0.011544011533260345,
+ "learning_rate": 0.0006,
+ "loss": 6.6186933517456055,
+ "step": 327
+ },
+ {
+ "epoch": 4.558951965065502,
+ "grad_norm": 0.01569426618516445,
+ "learning_rate": 0.0006,
+ "loss": 6.640767574310303,
+ "step": 328
+ },
+ {
+ "epoch": 4.5729257641921395,
+ "grad_norm": 0.027668394148349762,
+ "learning_rate": 0.0006,
+ "loss": 6.651412487030029,
+ "step": 329
+ },
+ {
+ "epoch": 4.586899563318777,
+ "grad_norm": 0.04073230177164078,
+ "learning_rate": 0.0006,
+ "loss": 6.636565208435059,
+ "step": 330
+ },
+ {
+ "epoch": 4.600873362445415,
+ "grad_norm": 0.04942672327160835,
+ "learning_rate": 0.0006,
+ "loss": 6.667994499206543,
+ "step": 331
+ },
+ {
+ "epoch": 4.614847161572053,
+ "grad_norm": 0.037568598985672,
+ "learning_rate": 0.0006,
+ "loss": 6.632008075714111,
+ "step": 332
+ },
+ {
+ "epoch": 4.62882096069869,
+ "grad_norm": 0.009880400262773037,
+ "learning_rate": 0.0006,
+ "loss": 6.6504058837890625,
+ "step": 333
+ },
+ {
+ "epoch": 4.642794759825327,
+ "grad_norm": 0.033384259790182114,
+ "learning_rate": 0.0006,
+ "loss": 6.63852596282959,
+ "step": 334
+ },
+ {
+ "epoch": 4.656768558951965,
+ "grad_norm": 0.035640884190797806,
+ "learning_rate": 0.0006,
+ "loss": 6.627986431121826,
+ "step": 335
+ },
+ {
+ "epoch": 4.670742358078603,
+ "grad_norm": 0.013637864030897617,
+ "learning_rate": 0.0006,
+ "loss": 6.624503135681152,
+ "step": 336
+ },
+ {
+ "epoch": 4.68471615720524,
+ "grad_norm": 0.025304675102233887,
+ "learning_rate": 0.0006,
+ "loss": 6.612077236175537,
+ "step": 337
+ },
+ {
+ "epoch": 4.698689956331878,
+ "grad_norm": 0.03377160802483559,
+ "learning_rate": 0.0006,
+ "loss": 6.608938217163086,
+ "step": 338
+ },
+ {
+ "epoch": 4.712663755458515,
+ "grad_norm": 0.019803153350949287,
+ "learning_rate": 0.0006,
+ "loss": 6.636343955993652,
+ "step": 339
+ },
+ {
+ "epoch": 4.726637554585153,
+ "grad_norm": 0.018928129225969315,
+ "learning_rate": 0.0006,
+ "loss": 6.636613845825195,
+ "step": 340
+ },
+ {
+ "epoch": 4.74061135371179,
+ "grad_norm": 0.023068996146321297,
+ "learning_rate": 0.0006,
+ "loss": 6.6258440017700195,
+ "step": 341
+ },
+ {
+ "epoch": 4.754585152838428,
+ "grad_norm": 0.016849957406520844,
+ "learning_rate": 0.0006,
+ "loss": 6.638483047485352,
+ "step": 342
+ },
+ {
+ "epoch": 4.7685589519650655,
+ "grad_norm": 0.021430544555187225,
+ "learning_rate": 0.0006,
+ "loss": 6.619649410247803,
+ "step": 343
+ },
+ {
+ "epoch": 4.782532751091703,
+ "grad_norm": 0.024053368717432022,
+ "learning_rate": 0.0006,
+ "loss": 6.639214515686035,
+ "step": 344
+ },
+ {
+ "epoch": 4.796506550218341,
+ "grad_norm": 0.011926693841814995,
+ "learning_rate": 0.0006,
+ "loss": 6.604836463928223,
+ "step": 345
+ },
+ {
+ "epoch": 4.810480349344978,
+ "grad_norm": 0.01961611397564411,
+ "learning_rate": 0.0006,
+ "loss": 6.624864101409912,
+ "step": 346
+ },
+ {
+ "epoch": 4.824454148471616,
+ "grad_norm": 0.027946067973971367,
+ "learning_rate": 0.0006,
+ "loss": 6.621284484863281,
+ "step": 347
+ },
+ {
+ "epoch": 4.8384279475982535,
+ "grad_norm": 0.02108592912554741,
+ "learning_rate": 0.0006,
+ "loss": 6.5982465744018555,
+ "step": 348
+ },
+ {
+ "epoch": 4.85240174672489,
+ "grad_norm": 0.017272260040044785,
+ "learning_rate": 0.0006,
+ "loss": 6.626142501831055,
+ "step": 349
+ },
+ {
+ "epoch": 4.866375545851528,
+ "grad_norm": 0.015654582530260086,
+ "learning_rate": 0.0006,
+ "loss": 6.621735572814941,
+ "step": 350
+ },
+ {
+ "epoch": 4.880349344978166,
+ "grad_norm": 0.015797540545463562,
+ "learning_rate": 0.0006,
+ "loss": 6.6016411781311035,
+ "step": 351
+ },
+ {
+ "epoch": 4.894323144104804,
+ "grad_norm": 0.028488904237747192,
+ "learning_rate": 0.0006,
+ "loss": 6.598413467407227,
+ "step": 352
+ },
+ {
+ "epoch": 4.908296943231441,
+ "grad_norm": 0.03411734476685524,
+ "learning_rate": 0.0006,
+ "loss": 6.597321033477783,
+ "step": 353
+ },
+ {
+ "epoch": 4.922270742358078,
+ "grad_norm": 0.027188263833522797,
+ "learning_rate": 0.0006,
+ "loss": 6.630616188049316,
+ "step": 354
+ },
+ {
+ "epoch": 4.936244541484716,
+ "grad_norm": 0.03411437198519707,
+ "learning_rate": 0.0006,
+ "loss": 6.57647705078125,
+ "step": 355
+ },
+ {
+ "epoch": 4.950218340611354,
+ "grad_norm": 0.019907327368855476,
+ "learning_rate": 0.0006,
+ "loss": 6.579545974731445,
+ "step": 356
+ },
+ {
+ "epoch": 4.964192139737992,
+ "grad_norm": 0.02485930174589157,
+ "learning_rate": 0.0006,
+ "loss": 6.591073036193848,
+ "step": 357
+ },
+ {
+ "epoch": 4.978165938864628,
+ "grad_norm": 0.026106052100658417,
+ "learning_rate": 0.0006,
+ "loss": 6.580729961395264,
+ "step": 358
+ },
+ {
+ "epoch": 4.992139737991266,
+ "grad_norm": 0.029401391744613647,
+ "learning_rate": 0.0006,
+ "loss": 6.5460076332092285,
+ "step": 359
+ },
+ {
+ "epoch": 5.0,
+ "grad_norm": 0.032869379967451096,
+ "learning_rate": 0.0006,
+ "loss": 6.583278179168701,
+ "step": 360
+ },
+ {
+ "epoch": 5.0,
+ "eval_loss": 7.416288375854492,
+ "eval_runtime": 60.4177,
+ "eval_samples_per_second": 40.419,
+ "eval_steps_per_second": 1.274,
+ "step": 360
+ },
+ {
+ "epoch": 5.013973799126638,
+ "grad_norm": 0.04862433671951294,
+ "learning_rate": 0.0006,
+ "loss": 6.58473014831543,
+ "step": 361
+ },
+ {
+ "epoch": 5.0279475982532755,
+ "grad_norm": 0.06636223942041397,
+ "learning_rate": 0.0006,
+ "loss": 6.621153354644775,
+ "step": 362
+ },
+ {
+ "epoch": 5.041921397379912,
+ "grad_norm": 0.05888264626264572,
+ "learning_rate": 0.0006,
+ "loss": 6.621196269989014,
+ "step": 363
+ },
+ {
+ "epoch": 5.05589519650655,
+ "grad_norm": 0.05331697314977646,
+ "learning_rate": 0.0006,
+ "loss": 6.600763320922852,
+ "step": 364
+ },
+ {
+ "epoch": 5.069868995633188,
+ "grad_norm": 0.03947671875357628,
+ "learning_rate": 0.0006,
+ "loss": 6.600671768188477,
+ "step": 365
+ },
+ {
+ "epoch": 5.083842794759826,
+ "grad_norm": 0.04045343026518822,
+ "learning_rate": 0.0006,
+ "loss": 6.63529109954834,
+ "step": 366
+ },
+ {
+ "epoch": 5.097816593886463,
+ "grad_norm": 0.029766036197543144,
+ "learning_rate": 0.0006,
+ "loss": 6.612868309020996,
+ "step": 367
+ },
+ {
+ "epoch": 5.1117903930131,
+ "grad_norm": 0.03416327014565468,
+ "learning_rate": 0.0006,
+ "loss": 6.582240581512451,
+ "step": 368
+ },
+ {
+ "epoch": 5.125764192139738,
+ "grad_norm": 0.03458515554666519,
+ "learning_rate": 0.0006,
+ "loss": 6.604381561279297,
+ "step": 369
+ },
+ {
+ "epoch": 5.139737991266376,
+ "grad_norm": 0.024099089205265045,
+ "learning_rate": 0.0006,
+ "loss": 6.597355365753174,
+ "step": 370
+ },
+ {
+ "epoch": 5.153711790393013,
+ "grad_norm": 0.02694438211619854,
+ "learning_rate": 0.0006,
+ "loss": 6.597512245178223,
+ "step": 371
+ },
+ {
+ "epoch": 5.1676855895196505,
+ "grad_norm": 0.024201856926083565,
+ "learning_rate": 0.0006,
+ "loss": 6.564685821533203,
+ "step": 372
+ },
+ {
+ "epoch": 5.181659388646288,
+ "grad_norm": 0.015959380194544792,
+ "learning_rate": 0.0006,
+ "loss": 6.582263946533203,
+ "step": 373
+ },
+ {
+ "epoch": 5.195633187772926,
+ "grad_norm": 0.025690171867609024,
+ "learning_rate": 0.0006,
+ "loss": 6.601057052612305,
+ "step": 374
+ },
+ {
+ "epoch": 5.209606986899563,
+ "grad_norm": 0.014497052878141403,
+ "learning_rate": 0.0006,
+ "loss": 6.584796905517578,
+ "step": 375
+ },
+ {
+ "epoch": 5.223580786026201,
+ "grad_norm": 0.019870297983288765,
+ "learning_rate": 0.0006,
+ "loss": 6.574010372161865,
+ "step": 376
+ },
+ {
+ "epoch": 5.2375545851528384,
+ "grad_norm": 0.01807512901723385,
+ "learning_rate": 0.0006,
+ "loss": 6.558596134185791,
+ "step": 377
+ },
+ {
+ "epoch": 5.251528384279476,
+ "grad_norm": 0.014525193721055984,
+ "learning_rate": 0.0006,
+ "loss": 6.571803092956543,
+ "step": 378
+ },
+ {
+ "epoch": 5.265502183406113,
+ "grad_norm": 0.016113435849547386,
+ "learning_rate": 0.0006,
+ "loss": 6.568970680236816,
+ "step": 379
+ },
+ {
+ "epoch": 5.279475982532751,
+ "grad_norm": 0.020724017173051834,
+ "learning_rate": 0.0006,
+ "loss": 6.5669403076171875,
+ "step": 380
+ },
+ {
+ "epoch": 5.293449781659389,
+ "grad_norm": 0.013963675126433372,
+ "learning_rate": 0.0006,
+ "loss": 6.586382865905762,
+ "step": 381
+ },
+ {
+ "epoch": 5.307423580786026,
+ "grad_norm": 0.01621115580201149,
+ "learning_rate": 0.0006,
+ "loss": 6.539546966552734,
+ "step": 382
+ },
+ {
+ "epoch": 5.321397379912664,
+ "grad_norm": 0.014925207942724228,
+ "learning_rate": 0.0006,
+ "loss": 6.531795501708984,
+ "step": 383
+ },
+ {
+ "epoch": 5.335371179039301,
+ "grad_norm": 0.015454350039362907,
+ "learning_rate": 0.0006,
+ "loss": 6.575699806213379,
+ "step": 384
+ },
+ {
+ "epoch": 5.349344978165939,
+ "grad_norm": 0.01623542234301567,
+ "learning_rate": 0.0006,
+ "loss": 6.582373142242432,
+ "step": 385
+ },
+ {
+ "epoch": 5.3633187772925766,
+ "grad_norm": 0.015452183783054352,
+ "learning_rate": 0.0006,
+ "loss": 6.576451301574707,
+ "step": 386
+ },
+ {
+ "epoch": 5.377292576419214,
+ "grad_norm": 0.021102432161569595,
+ "learning_rate": 0.0006,
+ "loss": 6.543055534362793,
+ "step": 387
+ },
+ {
+ "epoch": 5.391266375545851,
+ "grad_norm": 0.009827060624957085,
+ "learning_rate": 0.0006,
+ "loss": 6.526939392089844,
+ "step": 388
+ },
+ {
+ "epoch": 5.405240174672489,
+ "grad_norm": 0.013735136948525906,
+ "learning_rate": 0.0006,
+ "loss": 6.538825988769531,
+ "step": 389
+ },
+ {
+ "epoch": 5.419213973799127,
+ "grad_norm": 0.02073509432375431,
+ "learning_rate": 0.0006,
+ "loss": 6.521315574645996,
+ "step": 390
+ },
+ {
+ "epoch": 5.4331877729257645,
+ "grad_norm": 0.03110959194600582,
+ "learning_rate": 0.0006,
+ "loss": 6.561807632446289,
+ "step": 391
+ },
+ {
+ "epoch": 5.447161572052401,
+ "grad_norm": 0.045058365911245346,
+ "learning_rate": 0.0006,
+ "loss": 6.526515960693359,
+ "step": 392
+ },
+ {
+ "epoch": 5.461135371179039,
+ "grad_norm": 0.06359834969043732,
+ "learning_rate": 0.0006,
+ "loss": 6.554903030395508,
+ "step": 393
+ },
+ {
+ "epoch": 5.475109170305677,
+ "grad_norm": 0.06406767666339874,
+ "learning_rate": 0.0006,
+ "loss": 6.576272010803223,
+ "step": 394
+ },
+ {
+ "epoch": 5.489082969432315,
+ "grad_norm": 0.03403699770569801,
+ "learning_rate": 0.0006,
+ "loss": 6.5306243896484375,
+ "step": 395
+ },
+ {
+ "epoch": 5.503056768558952,
+ "grad_norm": 0.036109618842601776,
+ "learning_rate": 0.0006,
+ "loss": 6.54176139831543,
+ "step": 396
+ },
+ {
+ "epoch": 5.517030567685589,
+ "grad_norm": 0.03478201478719711,
+ "learning_rate": 0.0006,
+ "loss": 6.530734062194824,
+ "step": 397
+ },
+ {
+ "epoch": 5.531004366812227,
+ "grad_norm": 0.03401544317603111,
+ "learning_rate": 0.0006,
+ "loss": 6.557794094085693,
+ "step": 398
+ },
+ {
+ "epoch": 5.544978165938865,
+ "grad_norm": 0.031629662960767746,
+ "learning_rate": 0.0006,
+ "loss": 6.564887046813965,
+ "step": 399
+ },
+ {
+ "epoch": 5.558951965065502,
+ "grad_norm": 0.02305024117231369,
+ "learning_rate": 0.0006,
+ "loss": 6.547412872314453,
+ "step": 400
+ },
+ {
+ "epoch": 5.5729257641921395,
+ "grad_norm": 0.030852800235152245,
+ "learning_rate": 0.0006,
+ "loss": 6.495370388031006,
+ "step": 401
+ },
+ {
+ "epoch": 5.586899563318777,
+ "grad_norm": 0.03368550166487694,
+ "learning_rate": 0.0006,
+ "loss": 6.495980262756348,
+ "step": 402
+ },
+ {
+ "epoch": 5.600873362445415,
+ "grad_norm": 0.020976359024643898,
+ "learning_rate": 0.0006,
+ "loss": 6.509410858154297,
+ "step": 403
+ },
+ {
+ "epoch": 5.614847161572053,
+ "grad_norm": 0.019652143120765686,
+ "learning_rate": 0.0006,
+ "loss": 6.493232727050781,
+ "step": 404
+ },
+ {
+ "epoch": 5.62882096069869,
+ "grad_norm": 0.024320121854543686,
+ "learning_rate": 0.0006,
+ "loss": 6.521531581878662,
+ "step": 405
+ },
+ {
+ "epoch": 5.642794759825327,
+ "grad_norm": 0.016611427068710327,
+ "learning_rate": 0.0006,
+ "loss": 6.507271766662598,
+ "step": 406
+ },
+ {
+ "epoch": 5.656768558951965,
+ "grad_norm": 0.01635856367647648,
+ "learning_rate": 0.0006,
+ "loss": 6.492038726806641,
+ "step": 407
+ },
+ {
+ "epoch": 5.670742358078603,
+ "grad_norm": 0.017642095685005188,
+ "learning_rate": 0.0006,
+ "loss": 6.500349998474121,
+ "step": 408
+ },
+ {
+ "epoch": 5.68471615720524,
+ "grad_norm": 0.01679178513586521,
+ "learning_rate": 0.0006,
+ "loss": 6.500361442565918,
+ "step": 409
+ },
+ {
+ "epoch": 5.698689956331878,
+ "grad_norm": 0.011776307597756386,
+ "learning_rate": 0.0006,
+ "loss": 6.482490539550781,
+ "step": 410
+ },
+ {
+ "epoch": 5.712663755458515,
+ "grad_norm": 0.0176856629550457,
+ "learning_rate": 0.0006,
+ "loss": 6.48488712310791,
+ "step": 411
+ },
+ {
+ "epoch": 5.726637554585153,
+ "grad_norm": 0.015193155966699123,
+ "learning_rate": 0.0006,
+ "loss": 6.459545612335205,
+ "step": 412
+ },
+ {
+ "epoch": 5.74061135371179,
+ "grad_norm": 0.018623212352395058,
+ "learning_rate": 0.0006,
+ "loss": 6.478870391845703,
+ "step": 413
+ },
+ {
+ "epoch": 5.754585152838428,
+ "grad_norm": 0.012841006740927696,
+ "learning_rate": 0.0006,
+ "loss": 6.474418640136719,
+ "step": 414
+ },
+ {
+ "epoch": 5.7685589519650655,
+ "grad_norm": 0.015540511347353458,
+ "learning_rate": 0.0006,
+ "loss": 6.4706525802612305,
+ "step": 415
+ },
+ {
+ "epoch": 5.782532751091703,
+ "grad_norm": 0.022695740684866905,
+ "learning_rate": 0.0006,
+ "loss": 6.455738067626953,
+ "step": 416
+ },
+ {
+ "epoch": 5.796506550218341,
+ "grad_norm": 0.026930129155516624,
+ "learning_rate": 0.0006,
+ "loss": 6.458016872406006,
+ "step": 417
+ },
+ {
+ "epoch": 5.810480349344978,
+ "grad_norm": 0.03555797412991524,
+ "learning_rate": 0.0006,
+ "loss": 6.480044364929199,
+ "step": 418
+ },
+ {
+ "epoch": 5.824454148471616,
+ "grad_norm": 0.05024682730436325,
+ "learning_rate": 0.0006,
+ "loss": 6.454564571380615,
+ "step": 419
+ },
+ {
+ "epoch": 5.8384279475982535,
+ "grad_norm": 0.06616255640983582,
+ "learning_rate": 0.0006,
+ "loss": 6.433927536010742,
+ "step": 420
+ },
+ {
+ "epoch": 5.85240174672489,
+ "grad_norm": 0.059065770357847214,
+ "learning_rate": 0.0006,
+ "loss": 6.508362770080566,
+ "step": 421
+ },
+ {
+ "epoch": 5.866375545851528,
+ "grad_norm": 0.026689983904361725,
+ "learning_rate": 0.0006,
+ "loss": 6.464297294616699,
+ "step": 422
+ },
+ {
+ "epoch": 5.880349344978166,
+ "grad_norm": 0.04616398364305496,
+ "learning_rate": 0.0006,
+ "loss": 6.45448112487793,
+ "step": 423
+ },
+ {
+ "epoch": 5.894323144104804,
+ "grad_norm": 0.04980356618762016,
+ "learning_rate": 0.0006,
+ "loss": 6.4696574211120605,
+ "step": 424
+ },
+ {
+ "epoch": 5.908296943231441,
+ "grad_norm": 0.0551983006298542,
+ "learning_rate": 0.0006,
+ "loss": 6.419000625610352,
+ "step": 425
+ },
+ {
+ "epoch": 5.922270742358078,
+ "grad_norm": 0.06988243758678436,
+ "learning_rate": 0.0006,
+ "loss": 6.442890167236328,
+ "step": 426
+ },
+ {
+ "epoch": 5.936244541484716,
+ "grad_norm": 0.06040226295590401,
+ "learning_rate": 0.0006,
+ "loss": 6.513258934020996,
+ "step": 427
+ },
+ {
+ "epoch": 5.950218340611354,
+ "grad_norm": 0.04635024070739746,
+ "learning_rate": 0.0006,
+ "loss": 6.4710283279418945,
+ "step": 428
+ },
+ {
+ "epoch": 5.964192139737992,
+ "grad_norm": 0.04762702435255051,
+ "learning_rate": 0.0006,
+ "loss": 6.469327926635742,
+ "step": 429
+ },
+ {
+ "epoch": 5.978165938864628,
+ "grad_norm": 0.04857536032795906,
+ "learning_rate": 0.0006,
+ "loss": 6.426105499267578,
+ "step": 430
+ },
+ {
+ "epoch": 5.992139737991266,
+ "grad_norm": 0.04412693530321121,
+ "learning_rate": 0.0006,
+ "loss": 6.43179988861084,
+ "step": 431
+ },
+ {
+ "epoch": 6.0,
+ "grad_norm": 0.026129672303795815,
+ "learning_rate": 0.0006,
+ "loss": 6.437918663024902,
+ "step": 432
+ },
+ {
+ "epoch": 6.0,
+ "eval_loss": 7.219249248504639,
+ "eval_runtime": 58.6837,
+ "eval_samples_per_second": 41.613,
+ "eval_steps_per_second": 1.312,
+ "step": 432
+ },
+ {
+ "epoch": 6.013973799126638,
+ "grad_norm": 0.03329874947667122,
+ "learning_rate": 0.0006,
+ "loss": 6.438179969787598,
+ "step": 433
+ },
+ {
+ "epoch": 6.0279475982532755,
+ "grad_norm": 0.03723187372088432,
+ "learning_rate": 0.0006,
+ "loss": 6.394242286682129,
+ "step": 434
+ },
+ {
+ "epoch": 6.041921397379912,
+ "grad_norm": 0.014625726267695427,
+ "learning_rate": 0.0006,
+ "loss": 6.412118911743164,
+ "step": 435
+ },
+ {
+ "epoch": 6.05589519650655,
+ "grad_norm": 0.02917475439608097,
+ "learning_rate": 0.0006,
+ "loss": 6.453164577484131,
+ "step": 436
+ },
+ {
+ "epoch": 6.069868995633188,
+ "grad_norm": 0.01811208389699459,
+ "learning_rate": 0.0006,
+ "loss": 6.396003246307373,
+ "step": 437
+ },
+ {
+ "epoch": 6.083842794759826,
+ "grad_norm": 0.027235351502895355,
+ "learning_rate": 0.0006,
+ "loss": 6.402976036071777,
+ "step": 438
+ },
+ {
+ "epoch": 6.097816593886463,
+ "grad_norm": 0.02835794910788536,
+ "learning_rate": 0.0006,
+ "loss": 6.380568504333496,
+ "step": 439
+ },
+ {
+ "epoch": 6.1117903930131,
+ "grad_norm": 0.020868154242634773,
+ "learning_rate": 0.0006,
+ "loss": 6.38240909576416,
+ "step": 440
+ },
+ {
+ "epoch": 6.125764192139738,
+ "grad_norm": 0.02113114856183529,
+ "learning_rate": 0.0006,
+ "loss": 6.377391815185547,
+ "step": 441
+ },
+ {
+ "epoch": 6.139737991266376,
+ "grad_norm": 0.018894119188189507,
+ "learning_rate": 0.0006,
+ "loss": 6.354569435119629,
+ "step": 442
+ },
+ {
+ "epoch": 6.153711790393013,
+ "grad_norm": 0.01820225641131401,
+ "learning_rate": 0.0006,
+ "loss": 6.392882347106934,
+ "step": 443
+ },
+ {
+ "epoch": 6.1676855895196505,
+ "grad_norm": 0.017510728910565376,
+ "learning_rate": 0.0006,
+ "loss": 6.364162445068359,
+ "step": 444
+ },
+ {
+ "epoch": 6.181659388646288,
+ "grad_norm": 0.017893686890602112,
+ "learning_rate": 0.0006,
+ "loss": 6.373456954956055,
+ "step": 445
+ },
+ {
+ "epoch": 6.195633187772926,
+ "grad_norm": 0.013382102362811565,
+ "learning_rate": 0.0006,
+ "loss": 6.3714399337768555,
+ "step": 446
+ },
+ {
+ "epoch": 6.209606986899563,
+ "grad_norm": 0.022540833801031113,
+ "learning_rate": 0.0006,
+ "loss": 6.348170757293701,
+ "step": 447
+ },
+ {
+ "epoch": 6.223580786026201,
+ "grad_norm": 0.02110104262828827,
+ "learning_rate": 0.0006,
+ "loss": 6.413206100463867,
+ "step": 448
+ },
+ {
+ "epoch": 6.2375545851528384,
+ "grad_norm": 0.014498349279165268,
+ "learning_rate": 0.0006,
+ "loss": 6.361384391784668,
+ "step": 449
+ },
+ {
+ "epoch": 6.251528384279476,
+ "grad_norm": 0.01819637045264244,
+ "learning_rate": 0.0006,
+ "loss": 6.315275192260742,
+ "step": 450
+ },
+ {
+ "epoch": 6.265502183406113,
+ "grad_norm": 0.0195972491055727,
+ "learning_rate": 0.0006,
+ "loss": 6.321994781494141,
+ "step": 451
+ },
+ {
+ "epoch": 6.279475982532751,
+ "grad_norm": 0.018822453916072845,
+ "learning_rate": 0.0006,
+ "loss": 6.345219612121582,
+ "step": 452
+ },
+ {
+ "epoch": 6.293449781659389,
+ "grad_norm": 0.019247712567448616,
+ "learning_rate": 0.0006,
+ "loss": 6.348073959350586,
+ "step": 453
+ },
+ {
+ "epoch": 6.307423580786026,
+ "grad_norm": 0.03175375238060951,
+ "learning_rate": 0.0006,
+ "loss": 6.300344944000244,
+ "step": 454
+ },
+ {
+ "epoch": 6.321397379912664,
+ "grad_norm": 0.06326444447040558,
+ "learning_rate": 0.0006,
+ "loss": 6.392579078674316,
+ "step": 455
+ },
+ {
+ "epoch": 6.335371179039301,
+ "grad_norm": 0.09667232632637024,
+ "learning_rate": 0.0006,
+ "loss": 6.36024284362793,
+ "step": 456
+ },
+ {
+ "epoch": 6.349344978165939,
+ "grad_norm": 0.05774446204304695,
+ "learning_rate": 0.0006,
+ "loss": 6.365338325500488,
+ "step": 457
+ },
+ {
+ "epoch": 6.3633187772925766,
+ "grad_norm": 0.08793356269598007,
+ "learning_rate": 0.0006,
+ "loss": 6.43284273147583,
+ "step": 458
+ },
+ {
+ "epoch": 6.377292576419214,
+ "grad_norm": 0.0640595331788063,
+ "learning_rate": 0.0006,
+ "loss": 6.342428207397461,
+ "step": 459
+ },
+ {
+ "epoch": 6.391266375545851,
+ "grad_norm": 0.0447838194668293,
+ "learning_rate": 0.0006,
+ "loss": 6.367971897125244,
+ "step": 460
+ },
+ {
+ "epoch": 6.405240174672489,
+ "grad_norm": 0.04768967255949974,
+ "learning_rate": 0.0006,
+ "loss": 6.35027551651001,
+ "step": 461
+ },
+ {
+ "epoch": 6.419213973799127,
+ "grad_norm": 0.027981851249933243,
+ "learning_rate": 0.0006,
+ "loss": 6.33723258972168,
+ "step": 462
+ },
+ {
+ "epoch": 6.4331877729257645,
+ "grad_norm": 0.03479623422026634,
+ "learning_rate": 0.0006,
+ "loss": 6.36590051651001,
+ "step": 463
+ },
+ {
+ "epoch": 6.447161572052401,
+ "grad_norm": 0.020846687257289886,
+ "learning_rate": 0.0006,
+ "loss": 6.315143585205078,
+ "step": 464
+ },
+ {
+ "epoch": 6.461135371179039,
+ "grad_norm": 0.0312928780913353,
+ "learning_rate": 0.0006,
+ "loss": 6.340150833129883,
+ "step": 465
+ },
+ {
+ "epoch": 6.475109170305677,
+ "grad_norm": 0.020593881607055664,
+ "learning_rate": 0.0006,
+ "loss": 6.348175048828125,
+ "step": 466
+ },
+ {
+ "epoch": 6.489082969432315,
+ "grad_norm": 0.022010236978530884,
+ "learning_rate": 0.0006,
+ "loss": 6.318607330322266,
+ "step": 467
+ },
+ {
+ "epoch": 6.503056768558952,
+ "grad_norm": 0.018427059054374695,
+ "learning_rate": 0.0006,
+ "loss": 6.331822395324707,
+ "step": 468
+ },
+ {
+ "epoch": 6.517030567685589,
+ "grad_norm": 0.019964968785643578,
+ "learning_rate": 0.0006,
+ "loss": 6.287320137023926,
+ "step": 469
+ },
+ {
+ "epoch": 6.531004366812227,
+ "grad_norm": 0.02130270190536976,
+ "learning_rate": 0.0006,
+ "loss": 6.353344917297363,
+ "step": 470
+ },
+ {
+ "epoch": 6.544978165938865,
+ "grad_norm": 0.01616514101624489,
+ "learning_rate": 0.0006,
+ "loss": 6.314052581787109,
+ "step": 471
+ },
+ {
+ "epoch": 6.558951965065502,
+ "grad_norm": 0.016639474779367447,
+ "learning_rate": 0.0006,
+ "loss": 6.320351600646973,
+ "step": 472
+ },
+ {
+ "epoch": 6.5729257641921395,
+ "grad_norm": 0.017620692029595375,
+ "learning_rate": 0.0006,
+ "loss": 6.311459064483643,
+ "step": 473
+ },
+ {
+ "epoch": 6.586899563318777,
+ "grad_norm": 0.013542844913899899,
+ "learning_rate": 0.0006,
+ "loss": 6.324368476867676,
+ "step": 474
+ },
+ {
+ "epoch": 6.600873362445415,
+ "grad_norm": 0.01308678649365902,
+ "learning_rate": 0.0006,
+ "loss": 6.307419776916504,
+ "step": 475
+ },
+ {
+ "epoch": 6.614847161572053,
+ "grad_norm": 0.014672392047941685,
+ "learning_rate": 0.0006,
+ "loss": 6.2785420417785645,
+ "step": 476
+ },
+ {
+ "epoch": 6.62882096069869,
+ "grad_norm": 0.01399585697799921,
+ "learning_rate": 0.0006,
+ "loss": 6.258619785308838,
+ "step": 477
+ },
+ {
+ "epoch": 6.642794759825327,
+ "grad_norm": 0.01330904383212328,
+ "learning_rate": 0.0006,
+ "loss": 6.299595832824707,
+ "step": 478
+ },
+ {
+ "epoch": 6.656768558951965,
+ "grad_norm": 0.016613377258181572,
+ "learning_rate": 0.0006,
+ "loss": 6.2900309562683105,
+ "step": 479
+ },
+ {
+ "epoch": 6.670742358078603,
+ "grad_norm": 0.015076331794261932,
+ "learning_rate": 0.0006,
+ "loss": 6.231028079986572,
+ "step": 480
+ },
+ {
+ "epoch": 6.68471615720524,
+ "grad_norm": 0.01589728519320488,
+ "learning_rate": 0.0006,
+ "loss": 6.257922172546387,
+ "step": 481
+ },
+ {
+ "epoch": 6.698689956331878,
+ "grad_norm": 0.030972328037023544,
+ "learning_rate": 0.0006,
+ "loss": 6.285238265991211,
+ "step": 482
+ },
+ {
+ "epoch": 6.712663755458515,
+ "grad_norm": 0.05048783868551254,
+ "learning_rate": 0.0006,
+ "loss": 6.287471771240234,
+ "step": 483
+ },
+ {
+ "epoch": 6.726637554585153,
+ "grad_norm": 0.07390714436769485,
+ "learning_rate": 0.0006,
+ "loss": 6.284293174743652,
+ "step": 484
+ },
+ {
+ "epoch": 6.74061135371179,
+ "grad_norm": 0.07166943699121475,
+ "learning_rate": 0.0006,
+ "loss": 6.290704727172852,
+ "step": 485
+ },
+ {
+ "epoch": 6.754585152838428,
+ "grad_norm": 0.028792452067136765,
+ "learning_rate": 0.0006,
+ "loss": 6.243415832519531,
+ "step": 486
+ },
+ {
+ "epoch": 6.7685589519650655,
+ "grad_norm": 0.04822041466832161,
+ "learning_rate": 0.0006,
+ "loss": 6.278229713439941,
+ "step": 487
+ },
+ {
+ "epoch": 6.782532751091703,
+ "grad_norm": 0.036688923835754395,
+ "learning_rate": 0.0006,
+ "loss": 6.275590896606445,
+ "step": 488
+ },
+ {
+ "epoch": 6.796506550218341,
+ "grad_norm": 0.03679114952683449,
+ "learning_rate": 0.0006,
+ "loss": 6.236509323120117,
+ "step": 489
+ },
+ {
+ "epoch": 6.810480349344978,
+ "grad_norm": 0.023249102756381035,
+ "learning_rate": 0.0006,
+ "loss": 6.230788230895996,
+ "step": 490
+ },
+ {
+ "epoch": 6.824454148471616,
+ "grad_norm": 0.02682286500930786,
+ "learning_rate": 0.0006,
+ "loss": 6.278890609741211,
+ "step": 491
+ },
+ {
+ "epoch": 6.8384279475982535,
+ "grad_norm": 0.02202044241130352,
+ "learning_rate": 0.0006,
+ "loss": 6.234531402587891,
+ "step": 492
+ },
+ {
+ "epoch": 6.85240174672489,
+ "grad_norm": 0.02545459009706974,
+ "learning_rate": 0.0006,
+ "loss": 6.271790981292725,
+ "step": 493
+ },
+ {
+ "epoch": 6.866375545851528,
+ "grad_norm": 0.02012932486832142,
+ "learning_rate": 0.0006,
+ "loss": 6.247462749481201,
+ "step": 494
+ },
+ {
+ "epoch": 6.880349344978166,
+ "grad_norm": 0.02216474711894989,
+ "learning_rate": 0.0006,
+ "loss": 6.26396369934082,
+ "step": 495
+ },
+ {
+ "epoch": 6.894323144104804,
+ "grad_norm": 0.016078704968094826,
+ "learning_rate": 0.0006,
+ "loss": 6.228609085083008,
+ "step": 496
+ },
+ {
+ "epoch": 6.908296943231441,
+ "grad_norm": 0.023021478205919266,
+ "learning_rate": 0.0006,
+ "loss": 6.197244167327881,
+ "step": 497
+ },
+ {
+ "epoch": 6.922270742358078,
+ "grad_norm": 0.024766333401203156,
+ "learning_rate": 0.0006,
+ "loss": 6.181090354919434,
+ "step": 498
+ },
+ {
+ "epoch": 6.936244541484716,
+ "grad_norm": 0.025581466034054756,
+ "learning_rate": 0.0006,
+ "loss": 6.201817512512207,
+ "step": 499
+ },
+ {
+ "epoch": 6.950218340611354,
+ "grad_norm": 0.016964981332421303,
+ "learning_rate": 0.0006,
+ "loss": 6.201486587524414,
+ "step": 500
+ },
+ {
+ "epoch": 6.964192139737992,
+ "grad_norm": 0.023957721889019012,
+ "learning_rate": 0.0006,
+ "loss": 6.185701847076416,
+ "step": 501
+ },
+ {
+ "epoch": 6.978165938864628,
+ "grad_norm": 0.029775245115160942,
+ "learning_rate": 0.0006,
+ "loss": 6.242060661315918,
+ "step": 502
+ },
+ {
+ "epoch": 6.992139737991266,
+ "grad_norm": 0.04007191210985184,
+ "learning_rate": 0.0006,
+ "loss": 6.19931697845459,
+ "step": 503
+ },
+ {
+ "epoch": 7.0,
+ "grad_norm": 0.04882276430726051,
+ "learning_rate": 0.0006,
+ "loss": 6.1880598068237305,
+ "step": 504
+ },
+ {
+ "epoch": 7.0,
+ "eval_loss": 7.018043518066406,
+ "eval_runtime": 59.8815,
+ "eval_samples_per_second": 40.781,
+ "eval_steps_per_second": 1.286,
+ "step": 504
+ },
+ {
+ "epoch": 7.013973799126638,
+ "grad_norm": 0.07144106924533844,
+ "learning_rate": 0.0006,
+ "loss": 6.172799110412598,
+ "step": 505
+ },
+ {
+ "epoch": 7.0279475982532755,
+ "grad_norm": 0.06712841987609863,
+ "learning_rate": 0.0006,
+ "loss": 6.228682518005371,
+ "step": 506
+ },
+ {
+ "epoch": 7.041921397379912,
+ "grad_norm": 0.04405239224433899,
+ "learning_rate": 0.0006,
+ "loss": 6.1827898025512695,
+ "step": 507
+ },
+ {
+ "epoch": 7.05589519650655,
+ "grad_norm": 0.03418995440006256,
+ "learning_rate": 0.0006,
+ "loss": 6.200876235961914,
+ "step": 508
+ },
+ {
+ "epoch": 7.069868995633188,
+ "grad_norm": 0.04002382606267929,
+ "learning_rate": 0.0006,
+ "loss": 6.1799211502075195,
+ "step": 509
+ },
+ {
+ "epoch": 7.083842794759826,
+ "grad_norm": 0.029748762026429176,
+ "learning_rate": 0.0006,
+ "loss": 6.125422477722168,
+ "step": 510
+ },
+ {
+ "epoch": 7.097816593886463,
+ "grad_norm": 0.03408357501029968,
+ "learning_rate": 0.0006,
+ "loss": 6.203458309173584,
+ "step": 511
+ },
+ {
+ "epoch": 7.1117903930131,
+ "grad_norm": 0.02227279357612133,
+ "learning_rate": 0.0006,
+ "loss": 6.135439872741699,
+ "step": 512
+ },
+ {
+ "epoch": 7.125764192139738,
+ "grad_norm": 0.02727862447500229,
+ "learning_rate": 0.0006,
+ "loss": 6.202380180358887,
+ "step": 513
+ },
+ {
+ "epoch": 7.139737991266376,
+ "grad_norm": 0.026019075885415077,
+ "learning_rate": 0.0006,
+ "loss": 6.201343536376953,
+ "step": 514
+ },
+ {
+ "epoch": 7.153711790393013,
+ "grad_norm": 0.027227282524108887,
+ "learning_rate": 0.0006,
+ "loss": 6.164059638977051,
+ "step": 515
+ },
+ {
+ "epoch": 7.1676855895196505,
+ "grad_norm": 0.027649980038404465,
+ "learning_rate": 0.0006,
+ "loss": 6.135448932647705,
+ "step": 516
+ },
+ {
+ "epoch": 7.181659388646288,
+ "grad_norm": 0.02168039232492447,
+ "learning_rate": 0.0006,
+ "loss": 6.154831886291504,
+ "step": 517
+ },
+ {
+ "epoch": 7.195633187772926,
+ "grad_norm": 0.02101467363536358,
+ "learning_rate": 0.0006,
+ "loss": 6.128695487976074,
+ "step": 518
+ },
+ {
+ "epoch": 7.209606986899563,
+ "grad_norm": 0.018344245851039886,
+ "learning_rate": 0.0006,
+ "loss": 6.136524200439453,
+ "step": 519
+ },
+ {
+ "epoch": 7.223580786026201,
+ "grad_norm": 0.020323453471064568,
+ "learning_rate": 0.0006,
+ "loss": 6.1206488609313965,
+ "step": 520
+ },
+ {
+ "epoch": 7.2375545851528384,
+ "grad_norm": 0.01750849559903145,
+ "learning_rate": 0.0006,
+ "loss": 6.128170967102051,
+ "step": 521
+ },
+ {
+ "epoch": 7.251528384279476,
+ "grad_norm": 0.017778221517801285,
+ "learning_rate": 0.0006,
+ "loss": 6.150458812713623,
+ "step": 522
+ },
+ {
+ "epoch": 7.265502183406113,
+ "grad_norm": 0.01863938942551613,
+ "learning_rate": 0.0006,
+ "loss": 6.100106716156006,
+ "step": 523
+ },
+ {
+ "epoch": 7.279475982532751,
+ "grad_norm": 0.020196080207824707,
+ "learning_rate": 0.0006,
+ "loss": 6.149346351623535,
+ "step": 524
+ },
+ {
+ "epoch": 7.293449781659389,
+ "grad_norm": 0.03203890845179558,
+ "learning_rate": 0.0006,
+ "loss": 6.124648094177246,
+ "step": 525
+ },
+ {
+ "epoch": 7.307423580786026,
+ "grad_norm": 0.05325170233845711,
+ "learning_rate": 0.0006,
+ "loss": 6.115618705749512,
+ "step": 526
+ },
+ {
+ "epoch": 7.321397379912664,
+ "grad_norm": 0.05888345465064049,
+ "learning_rate": 0.0006,
+ "loss": 6.165221214294434,
+ "step": 527
+ },
+ {
+ "epoch": 7.335371179039301,
+ "grad_norm": 0.026873115450143814,
+ "learning_rate": 0.0006,
+ "loss": 6.039590835571289,
+ "step": 528
+ },
+ {
+ "epoch": 7.349344978165939,
+ "grad_norm": 0.06304649263620377,
+ "learning_rate": 0.0006,
+ "loss": 6.131900787353516,
+ "step": 529
+ },
+ {
+ "epoch": 7.3633187772925766,
+ "grad_norm": 0.08567557483911514,
+ "learning_rate": 0.0006,
+ "loss": 6.218114852905273,
+ "step": 530
+ },
+ {
+ "epoch": 7.377292576419214,
+ "grad_norm": 0.08868218213319778,
+ "learning_rate": 0.0006,
+ "loss": 6.216706275939941,
+ "step": 531
+ },
+ {
+ "epoch": 7.391266375545851,
+ "grad_norm": 0.03955601155757904,
+ "learning_rate": 0.0006,
+ "loss": 6.1566925048828125,
+ "step": 532
+ },
+ {
+ "epoch": 7.405240174672489,
+ "grad_norm": 0.04979498311877251,
+ "learning_rate": 0.0006,
+ "loss": 6.19327449798584,
+ "step": 533
+ },
+ {
+ "epoch": 7.419213973799127,
+ "grad_norm": 0.03253026679158211,
+ "learning_rate": 0.0006,
+ "loss": 6.133993148803711,
+ "step": 534
+ },
+ {
+ "epoch": 7.4331877729257645,
+ "grad_norm": 0.03016652725636959,
+ "learning_rate": 0.0006,
+ "loss": 6.174962997436523,
+ "step": 535
+ },
+ {
+ "epoch": 7.447161572052401,
+ "grad_norm": 0.03159862384200096,
+ "learning_rate": 0.0006,
+ "loss": 6.082124710083008,
+ "step": 536
+ },
+ {
+ "epoch": 7.461135371179039,
+ "grad_norm": 0.027673959732055664,
+ "learning_rate": 0.0006,
+ "loss": 6.111873149871826,
+ "step": 537
+ },
+ {
+ "epoch": 7.475109170305677,
+ "grad_norm": 0.022789714857935905,
+ "learning_rate": 0.0006,
+ "loss": 6.129437446594238,
+ "step": 538
+ },
+ {
+ "epoch": 7.489082969432315,
+ "grad_norm": 0.030061012133955956,
+ "learning_rate": 0.0006,
+ "loss": 6.126370429992676,
+ "step": 539
+ },
+ {
+ "epoch": 7.503056768558952,
+ "grad_norm": 0.019007068127393723,
+ "learning_rate": 0.0006,
+ "loss": 6.07887077331543,
+ "step": 540
+ },
+ {
+ "epoch": 7.517030567685589,
+ "grad_norm": 0.01897653006017208,
+ "learning_rate": 0.0006,
+ "loss": 6.096936225891113,
+ "step": 541
+ },
+ {
+ "epoch": 7.531004366812227,
+ "grad_norm": 0.024659698829054832,
+ "learning_rate": 0.0006,
+ "loss": 6.0933380126953125,
+ "step": 542
+ },
+ {
+ "epoch": 7.544978165938865,
+ "grad_norm": 0.021924549713730812,
+ "learning_rate": 0.0006,
+ "loss": 6.105311870574951,
+ "step": 543
+ },
+ {
+ "epoch": 7.558951965065502,
+ "grad_norm": 0.018817558884620667,
+ "learning_rate": 0.0006,
+ "loss": 6.058895111083984,
+ "step": 544
+ },
+ {
+ "epoch": 7.5729257641921395,
+ "grad_norm": 0.02477840520441532,
+ "learning_rate": 0.0006,
+ "loss": 6.038928031921387,
+ "step": 545
+ },
+ {
+ "epoch": 7.586899563318777,
+ "grad_norm": 0.018968632444739342,
+ "learning_rate": 0.0006,
+ "loss": 6.1368937492370605,
+ "step": 546
+ },
+ {
+ "epoch": 7.600873362445415,
+ "grad_norm": 0.02409801445901394,
+ "learning_rate": 0.0006,
+ "loss": 6.110324859619141,
+ "step": 547
+ },
+ {
+ "epoch": 7.614847161572053,
+ "grad_norm": 0.017595326527953148,
+ "learning_rate": 0.0006,
+ "loss": 6.074368476867676,
+ "step": 548
+ },
+ {
+ "epoch": 7.62882096069869,
+ "grad_norm": 0.01875406503677368,
+ "learning_rate": 0.0006,
+ "loss": 6.121023654937744,
+ "step": 549
+ },
+ {
+ "epoch": 7.642794759825327,
+ "grad_norm": 0.018781628459692,
+ "learning_rate": 0.0006,
+ "loss": 6.064914703369141,
+ "step": 550
+ },
+ {
+ "epoch": 7.656768558951965,
+ "grad_norm": 0.017003126442432404,
+ "learning_rate": 0.0006,
+ "loss": 6.039608001708984,
+ "step": 551
+ },
+ {
+ "epoch": 7.670742358078603,
+ "grad_norm": 0.017665037885308266,
+ "learning_rate": 0.0006,
+ "loss": 6.053671836853027,
+ "step": 552
+ },
+ {
+ "epoch": 7.68471615720524,
+ "grad_norm": 0.02066265605390072,
+ "learning_rate": 0.0006,
+ "loss": 6.090536117553711,
+ "step": 553
+ },
+ {
+ "epoch": 7.698689956331878,
+ "grad_norm": 0.02888769656419754,
+ "learning_rate": 0.0006,
+ "loss": 5.9858903884887695,
+ "step": 554
+ },
+ {
+ "epoch": 7.712663755458515,
+ "grad_norm": 0.03744485229253769,
+ "learning_rate": 0.0006,
+ "loss": 6.086264610290527,
+ "step": 555
+ },
+ {
+ "epoch": 7.726637554585153,
+ "grad_norm": 0.04719431325793266,
+ "learning_rate": 0.0006,
+ "loss": 6.09747314453125,
+ "step": 556
+ },
+ {
+ "epoch": 7.74061135371179,
+ "grad_norm": 0.044963907450437546,
+ "learning_rate": 0.0006,
+ "loss": 6.091272354125977,
+ "step": 557
+ },
+ {
+ "epoch": 7.754585152838428,
+ "grad_norm": 0.03810194507241249,
+ "learning_rate": 0.0006,
+ "loss": 6.084904193878174,
+ "step": 558
+ },
+ {
+ "epoch": 7.7685589519650655,
+ "grad_norm": 0.04626769199967384,
+ "learning_rate": 0.0006,
+ "loss": 6.017384052276611,
+ "step": 559
+ },
+ {
+ "epoch": 7.782532751091703,
+ "grad_norm": 0.03459552302956581,
+ "learning_rate": 0.0006,
+ "loss": 6.111314296722412,
+ "step": 560
+ },
+ {
+ "epoch": 7.796506550218341,
+ "grad_norm": 0.027385666966438293,
+ "learning_rate": 0.0006,
+ "loss": 6.002509117126465,
+ "step": 561
+ },
+ {
+ "epoch": 7.810480349344978,
+ "grad_norm": 0.03417652100324631,
+ "learning_rate": 0.0006,
+ "loss": 6.0421061515808105,
+ "step": 562
+ },
+ {
+ "epoch": 7.824454148471616,
+ "grad_norm": 0.04415450245141983,
+ "learning_rate": 0.0006,
+ "loss": 6.035305023193359,
+ "step": 563
+ },
+ {
+ "epoch": 7.8384279475982535,
+ "grad_norm": 0.03259635716676712,
+ "learning_rate": 0.0006,
+ "loss": 6.025840759277344,
+ "step": 564
+ },
+ {
+ "epoch": 7.85240174672489,
+ "grad_norm": 0.024941129609942436,
+ "learning_rate": 0.0006,
+ "loss": 6.043416976928711,
+ "step": 565
+ },
+ {
+ "epoch": 7.866375545851528,
+ "grad_norm": 0.02434500865638256,
+ "learning_rate": 0.0006,
+ "loss": 6.002052307128906,
+ "step": 566
+ },
+ {
+ "epoch": 7.880349344978166,
+ "grad_norm": 0.01876584254205227,
+ "learning_rate": 0.0006,
+ "loss": 6.045035362243652,
+ "step": 567
+ },
+ {
+ "epoch": 7.894323144104804,
+ "grad_norm": 0.02254367619752884,
+ "learning_rate": 0.0006,
+ "loss": 5.972503662109375,
+ "step": 568
+ },
+ {
+ "epoch": 7.908296943231441,
+ "grad_norm": 0.024504603818058968,
+ "learning_rate": 0.0006,
+ "loss": 5.992846488952637,
+ "step": 569
+ },
+ {
+ "epoch": 7.922270742358078,
+ "grad_norm": 0.02317827194929123,
+ "learning_rate": 0.0006,
+ "loss": 6.007697582244873,
+ "step": 570
+ },
+ {
+ "epoch": 7.936244541484716,
+ "grad_norm": 0.02225821278989315,
+ "learning_rate": 0.0006,
+ "loss": 6.01463508605957,
+ "step": 571
+ },
+ {
+ "epoch": 7.950218340611354,
+ "grad_norm": 0.01811717264354229,
+ "learning_rate": 0.0006,
+ "loss": 5.995022773742676,
+ "step": 572
+ },
+ {
+ "epoch": 7.964192139737992,
+ "grad_norm": 0.019589431583881378,
+ "learning_rate": 0.0006,
+ "loss": 5.990633964538574,
+ "step": 573
+ },
+ {
+ "epoch": 7.978165938864628,
+ "grad_norm": 0.015422016382217407,
+ "learning_rate": 0.0006,
+ "loss": 6.011263847351074,
+ "step": 574
+ },
+ {
+ "epoch": 7.992139737991266,
+ "grad_norm": 0.023298880085349083,
+ "learning_rate": 0.0006,
+ "loss": 5.9954938888549805,
+ "step": 575
+ },
+ {
+ "epoch": 8.0,
+ "grad_norm": 0.022172044962644577,
+ "learning_rate": 0.0006,
+ "loss": 5.95679235458374,
+ "step": 576
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 28800,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 400,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": true,
+ "should_log": false,
+ "should_save": false,
+ "should_training_stop": false
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 2.4556906890657792e+17,
+ "train_batch_size": 8,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-576/training_args.bin b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-576/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..c799bcf4051249aabf4c62fff9d38801df7bc2b5
--- /dev/null
+++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-576/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:40fb25263250c08c3479a6f0ac9fdf1f196e2a3b08a4eb10384786ff9b1903ef
+size 4792
diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-72/chat_template.jinja b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-72/chat_template.jinja
new file mode 100644
index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230
--- /dev/null
+++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-72/chat_template.jinja
@@ -0,0 +1,85 @@
+{%- if tools %}
+ {{- '<|im_start|>system\n' }}
+ {%- if messages[0].role == 'system' %}
+ {{- messages[0].content + '\n\n' }}
+ {%- endif %}
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }}
+ {%- for tool in tools %}
+ {{- "\n" }}
+ {{- tool | tojson }}
+ {%- endfor %}
+ {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }}
+{%- else %}
+ {%- if messages[0].role == 'system' %}
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
+ {%- endif %}
+{%- endif %}
+{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
+{%- for message in messages[::-1] %}
+ {%- set index = (messages|length - 1) - loop.index0 %}
+ {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %}
+ {%- set ns.multi_step_tool = false %}
+ {%- set ns.last_query_index = index %}
+ {%- endif %}
+{%- endfor %}
+{%- for message in messages %}
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
+ {%- elif message.role == "assistant" %}
+ {%- set content = message.content %}
+ {%- set reasoning_content = '' %}
+ {%- if message.reasoning_content is defined and message.reasoning_content is not none %}
+ {%- set reasoning_content = message.reasoning_content %}
+ {%- else %}
+ {%- if '' in message.content %}
+ {%- set content = message.content.split('')[-1].lstrip('\n') %}
+ {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %}
+ {%- endif %}
+ {%- endif %}
+ {%- if loop.index0 > ns.last_query_index %}
+ {%- if loop.last or (not loop.last and reasoning_content) %}
+ {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- if message.tool_calls %}
+ {%- for tool_call in message.tool_calls %}
+ {%- if (loop.first and content) or (not loop.first) %}
+ {{- '\n' }}
+ {%- endif %}
+ {%- if tool_call.function %}
+ {%- set tool_call = tool_call.function %}
+ {%- endif %}
+ {{- '\n{"name": "' }}
+ {{- tool_call.name }}
+ {{- '", "arguments": ' }}
+ {%- if tool_call.arguments is string %}
+ {{- tool_call.arguments }}
+ {%- else %}
+ {{- tool_call.arguments | tojson }}
+ {%- endif %}
+ {{- '}\n' }}
+ {%- endfor %}
+ {%- endif %}
+ {{- '<|im_end|>\n' }}
+ {%- elif message.role == "tool" %}
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
+ {{- '<|im_start|>user' }}
+ {%- endif %}
+ {{- '\n\n' }}
+ {{- message.content }}
+ {{- '\n' }}
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
+ {{- '<|im_end|>\n' }}
+ {%- endif %}
+ {%- endif %}
+{%- endfor %}
+{%- if add_generation_prompt %}
+ {{- '<|im_start|>assistant\n' }}
+ {%- if enable_thinking is defined and enable_thinking is false %}
+ {{- '\n\n\n\n' }}
+ {%- endif %}
+{%- endif %}
\ No newline at end of file
diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-72/config.json b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-72/config.json
new file mode 100644
index 0000000000000000000000000000000000000000..175ee3431793018f23f74824f69554b21afcbd4f
--- /dev/null
+++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-72/config.json
@@ -0,0 +1,32 @@
+{
+ "architectures": [
+ "LlamaForCausalLM"
+ ],
+ "attention_bias": false,
+ "attention_dropout": 0.0,
+ "bos_token_id": null,
+ "dtype": "float32",
+ "eos_token_id": 151645,
+ "head_dim": 128,
+ "hidden_act": "silu",
+ "hidden_size": 512,
+ "initializer_range": 0.02,
+ "intermediate_size": 1536,
+ "max_position_embeddings": 2048,
+ "mlp_bias": false,
+ "model_type": "llama",
+ "num_attention_heads": 4,
+ "num_hidden_layers": 20,
+ "num_key_value_heads": 4,
+ "pad_token_id": 151645,
+ "pretraining_tp": 1,
+ "rms_norm_eps": 1e-06,
+ "rope_parameters": {
+ "rope_theta": 10000.0,
+ "rope_type": "default"
+ },
+ "tie_word_embeddings": true,
+ "transformers_version": "5.5.0",
+ "use_cache": false,
+ "vocab_size": 151674
+}
diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-72/generation_config.json b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-72/generation_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3
--- /dev/null
+++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-72/generation_config.json
@@ -0,0 +1,11 @@
+{
+ "_from_model_config": true,
+ "eos_token_id": [
+ 151645
+ ],
+ "output_attentions": false,
+ "output_hidden_states": false,
+ "pad_token_id": 151645,
+ "transformers_version": "5.5.0",
+ "use_cache": true
+}
diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-72/model.safetensors b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-72/model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..b853dbdc80b4273e5d550f1e2dbd877a098bb903
--- /dev/null
+++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-72/model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:0e34000ad5b58ca0564625863e22c69e39127b37d86835e20118d4fd08824445
+size 583362376
diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-72/optimizer.pt b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-72/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..6f03aa4ff10b813a3e1c00c87d636b2965e1eec4
--- /dev/null
+++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-72/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:ed7b28dab7d301258603fe180514d85e1b9a3b9f82f0a8c925caeda756dc10a7
+size 1166837626
diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-72/rng_state_0.pth b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-72/rng_state_0.pth
new file mode 100644
index 0000000000000000000000000000000000000000..35b54c5c64a65f3977bb4b4ec2b748cd17b20f52
--- /dev/null
+++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-72/rng_state_0.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:7e5d12f33f319393c8dafffd6b14ce4cfa3aeb46c8c3760b273eddbb8152d1c7
+size 15024
diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-72/rng_state_1.pth b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-72/rng_state_1.pth
new file mode 100644
index 0000000000000000000000000000000000000000..aa9143a4e362d6c0ee326f19cdd26dcecaf7e6a3
--- /dev/null
+++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-72/rng_state_1.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:ce7ffd7ac75644c8f6fc42f002287b7b0c3c585e58668b8936b9fddaed2f8cc5
+size 15024
diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-72/rng_state_2.pth b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-72/rng_state_2.pth
new file mode 100644
index 0000000000000000000000000000000000000000..d65c7a733824cbda65cbe53f302ffcca01bce8c6
--- /dev/null
+++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-72/rng_state_2.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:cc504e473f67f378865973402a0109aa8223a3b343d77a078ed4b4bd0df27098
+size 15024
diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-72/rng_state_3.pth b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-72/rng_state_3.pth
new file mode 100644
index 0000000000000000000000000000000000000000..476c1ccbff619bfca532c36bdd72e59d0543bfd6
--- /dev/null
+++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-72/rng_state_3.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:afa0376616f56e7a706215e7b0a97a0be812db73d881b1ab615ea8260bc9350f
+size 15024
diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-72/scheduler.pt b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-72/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..43cf114590f1088fd438450226c42885d843da41
--- /dev/null
+++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-72/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:878ec53f5c2a524fc37af578b94a9025923faefea55f621be8e5643a79915d0f
+size 1064
diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-72/tokenizer.json b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-72/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..9ea28d9eb325c36bcab4c1a08fc93fd598929f08
--- /dev/null
+++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-72/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:60e99075d371ecbb9f1dde97363f612d45bf0bf88f1a2006d9bfcf1939333c14
+size 11424648
diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-72/tokenizer_config.json b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-72/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..cadd997fc26040d23ad6677fcb39d8d18b01f55e
--- /dev/null
+++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-72/tokenizer_config.json
@@ -0,0 +1,21 @@
+{
+ "add_prefix_space": false,
+ "backend": "tokenizers",
+ "bos_token": null,
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "<|im_end|>",
+ "errors": "replace",
+ "extra_special_tokens": [
+ "<|l2r_pred|>",
+ "<|r2l_pred|>",
+ "<|next_1_pred|>",
+ "<|next_2_pred|>",
+ "<|mask|>"
+ ],
+ "is_local": false,
+ "model_max_length": 131072,
+ "pad_token": "<|im_end|>",
+ "split_special_tokens": false,
+ "tokenizer_class": "Qwen2Tokenizer",
+ "unk_token": null
+}
diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-72/trainer_state.json b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-72/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..34f68e1a9ca8b8c4c94fbde71d1947f973c6be4c
--- /dev/null
+++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-72/trainer_state.json
@@ -0,0 +1,538 @@
+{
+ "best_global_step": null,
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 1.0,
+ "eval_steps": 500,
+ "global_step": 72,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.013973799126637555,
+ "grad_norm": 0.5768774151802063,
+ "learning_rate": 0.0,
+ "loss": 12.050328254699707,
+ "step": 1
+ },
+ {
+ "epoch": 0.02794759825327511,
+ "grad_norm": 0.5784570574760437,
+ "learning_rate": 5.999999999999999e-06,
+ "loss": 12.049251556396484,
+ "step": 2
+ },
+ {
+ "epoch": 0.04192139737991266,
+ "grad_norm": 0.5607187151908875,
+ "learning_rate": 1.1999999999999999e-05,
+ "loss": 11.921792030334473,
+ "step": 3
+ },
+ {
+ "epoch": 0.05589519650655022,
+ "grad_norm": 0.46466919779777527,
+ "learning_rate": 1.7999999999999997e-05,
+ "loss": 11.686628341674805,
+ "step": 4
+ },
+ {
+ "epoch": 0.06986899563318777,
+ "grad_norm": 0.38012945652008057,
+ "learning_rate": 2.3999999999999997e-05,
+ "loss": 11.432823181152344,
+ "step": 5
+ },
+ {
+ "epoch": 0.08384279475982533,
+ "grad_norm": 0.33215776085853577,
+ "learning_rate": 2.9999999999999997e-05,
+ "loss": 11.207542419433594,
+ "step": 6
+ },
+ {
+ "epoch": 0.09781659388646288,
+ "grad_norm": 0.2928776741027832,
+ "learning_rate": 3.5999999999999994e-05,
+ "loss": 11.000097274780273,
+ "step": 7
+ },
+ {
+ "epoch": 0.11179039301310044,
+ "grad_norm": 0.26018962264060974,
+ "learning_rate": 4.2e-05,
+ "loss": 10.825748443603516,
+ "step": 8
+ },
+ {
+ "epoch": 0.125764192139738,
+ "grad_norm": 0.23717083036899567,
+ "learning_rate": 4.7999999999999994e-05,
+ "loss": 10.677122116088867,
+ "step": 9
+ },
+ {
+ "epoch": 0.13973799126637554,
+ "grad_norm": 0.21641355752944946,
+ "learning_rate": 5.399999999999999e-05,
+ "loss": 10.569148063659668,
+ "step": 10
+ },
+ {
+ "epoch": 0.1537117903930131,
+ "grad_norm": 0.1982956826686859,
+ "learning_rate": 5.9999999999999995e-05,
+ "loss": 10.49653434753418,
+ "step": 11
+ },
+ {
+ "epoch": 0.16768558951965065,
+ "grad_norm": 0.18819282948970795,
+ "learning_rate": 6.599999999999999e-05,
+ "loss": 10.432951927185059,
+ "step": 12
+ },
+ {
+ "epoch": 0.18165938864628822,
+ "grad_norm": 0.18322618305683136,
+ "learning_rate": 7.199999999999999e-05,
+ "loss": 10.387115478515625,
+ "step": 13
+ },
+ {
+ "epoch": 0.19563318777292577,
+ "grad_norm": 0.18399226665496826,
+ "learning_rate": 7.8e-05,
+ "loss": 10.33381462097168,
+ "step": 14
+ },
+ {
+ "epoch": 0.2096069868995633,
+ "grad_norm": 0.1833190619945526,
+ "learning_rate": 8.4e-05,
+ "loss": 10.301689147949219,
+ "step": 15
+ },
+ {
+ "epoch": 0.22358078602620088,
+ "grad_norm": 0.18490834534168243,
+ "learning_rate": 8.999999999999999e-05,
+ "loss": 10.240968704223633,
+ "step": 16
+ },
+ {
+ "epoch": 0.23755458515283842,
+ "grad_norm": 0.18358023464679718,
+ "learning_rate": 9.599999999999999e-05,
+ "loss": 10.194717407226562,
+ "step": 17
+ },
+ {
+ "epoch": 0.251528384279476,
+ "grad_norm": 0.1801682710647583,
+ "learning_rate": 0.000102,
+ "loss": 10.14897632598877,
+ "step": 18
+ },
+ {
+ "epoch": 0.26550218340611353,
+ "grad_norm": 0.17899440228939056,
+ "learning_rate": 0.00010799999999999998,
+ "loss": 10.078763961791992,
+ "step": 19
+ },
+ {
+ "epoch": 0.2794759825327511,
+ "grad_norm": 0.17666497826576233,
+ "learning_rate": 0.00011399999999999999,
+ "loss": 10.01095962524414,
+ "step": 20
+ },
+ {
+ "epoch": 0.2934497816593886,
+ "grad_norm": 0.17196154594421387,
+ "learning_rate": 0.00011999999999999999,
+ "loss": 9.951203346252441,
+ "step": 21
+ },
+ {
+ "epoch": 0.3074235807860262,
+ "grad_norm": 0.169203519821167,
+ "learning_rate": 0.00012599999999999997,
+ "loss": 9.871726036071777,
+ "step": 22
+ },
+ {
+ "epoch": 0.32139737991266376,
+ "grad_norm": 0.16500306129455566,
+ "learning_rate": 0.00013199999999999998,
+ "loss": 9.803646087646484,
+ "step": 23
+ },
+ {
+ "epoch": 0.3353711790393013,
+ "grad_norm": 0.16128124296665192,
+ "learning_rate": 0.000138,
+ "loss": 9.728487014770508,
+ "step": 24
+ },
+ {
+ "epoch": 0.34934497816593885,
+ "grad_norm": 0.15904058516025543,
+ "learning_rate": 0.00014399999999999998,
+ "loss": 9.642473220825195,
+ "step": 25
+ },
+ {
+ "epoch": 0.36331877729257644,
+ "grad_norm": 0.15560561418533325,
+ "learning_rate": 0.00015,
+ "loss": 9.56414794921875,
+ "step": 26
+ },
+ {
+ "epoch": 0.377292576419214,
+ "grad_norm": 0.15278612077236176,
+ "learning_rate": 0.000156,
+ "loss": 9.484230041503906,
+ "step": 27
+ },
+ {
+ "epoch": 0.39126637554585153,
+ "grad_norm": 0.15263237059116364,
+ "learning_rate": 0.000162,
+ "loss": 9.383644104003906,
+ "step": 28
+ },
+ {
+ "epoch": 0.4052401746724891,
+ "grad_norm": 0.14945024251937866,
+ "learning_rate": 0.000168,
+ "loss": 9.29840087890625,
+ "step": 29
+ },
+ {
+ "epoch": 0.4192139737991266,
+ "grad_norm": 0.14614646136760712,
+ "learning_rate": 0.00017399999999999997,
+ "loss": 9.208931922912598,
+ "step": 30
+ },
+ {
+ "epoch": 0.4331877729257642,
+ "grad_norm": 0.14295098185539246,
+ "learning_rate": 0.00017999999999999998,
+ "loss": 9.122323989868164,
+ "step": 31
+ },
+ {
+ "epoch": 0.44716157205240176,
+ "grad_norm": 0.13868345320224762,
+ "learning_rate": 0.000186,
+ "loss": 9.03034496307373,
+ "step": 32
+ },
+ {
+ "epoch": 0.4611353711790393,
+ "grad_norm": 0.135064035654068,
+ "learning_rate": 0.00019199999999999998,
+ "loss": 8.92818832397461,
+ "step": 33
+ },
+ {
+ "epoch": 0.47510917030567684,
+ "grad_norm": 0.13097241520881653,
+ "learning_rate": 0.000198,
+ "loss": 8.834627151489258,
+ "step": 34
+ },
+ {
+ "epoch": 0.4890829694323144,
+ "grad_norm": 0.12478776276111603,
+ "learning_rate": 0.000204,
+ "loss": 8.755172729492188,
+ "step": 35
+ },
+ {
+ "epoch": 0.503056768558952,
+ "grad_norm": 0.1214982345700264,
+ "learning_rate": 0.00020999999999999998,
+ "loss": 8.647891998291016,
+ "step": 36
+ },
+ {
+ "epoch": 0.5170305676855895,
+ "grad_norm": 0.11958328634500504,
+ "learning_rate": 0.00021599999999999996,
+ "loss": 8.543274879455566,
+ "step": 37
+ },
+ {
+ "epoch": 0.5310043668122271,
+ "grad_norm": 0.11680915951728821,
+ "learning_rate": 0.00022199999999999998,
+ "loss": 8.44540023803711,
+ "step": 38
+ },
+ {
+ "epoch": 0.5449781659388646,
+ "grad_norm": 0.11491474509239197,
+ "learning_rate": 0.00022799999999999999,
+ "loss": 8.349921226501465,
+ "step": 39
+ },
+ {
+ "epoch": 0.5589519650655022,
+ "grad_norm": 0.11160755902528763,
+ "learning_rate": 0.000234,
+ "loss": 8.26960563659668,
+ "step": 40
+ },
+ {
+ "epoch": 0.5729257641921397,
+ "grad_norm": 0.10840637236833572,
+ "learning_rate": 0.00023999999999999998,
+ "loss": 8.192136764526367,
+ "step": 41
+ },
+ {
+ "epoch": 0.5868995633187772,
+ "grad_norm": 0.10585298389196396,
+ "learning_rate": 0.00024599999999999996,
+ "loss": 8.09546947479248,
+ "step": 42
+ },
+ {
+ "epoch": 0.6008733624454149,
+ "grad_norm": 0.1032542735338211,
+ "learning_rate": 0.00025199999999999995,
+ "loss": 7.991090774536133,
+ "step": 43
+ },
+ {
+ "epoch": 0.6148471615720524,
+ "grad_norm": 0.0986943170428276,
+ "learning_rate": 0.000258,
+ "loss": 7.92384147644043,
+ "step": 44
+ },
+ {
+ "epoch": 0.62882096069869,
+ "grad_norm": 0.09320762753486633,
+ "learning_rate": 0.00026399999999999997,
+ "loss": 7.8478193283081055,
+ "step": 45
+ },
+ {
+ "epoch": 0.6427947598253275,
+ "grad_norm": 0.09168332070112228,
+ "learning_rate": 0.00027,
+ "loss": 7.76185417175293,
+ "step": 46
+ },
+ {
+ "epoch": 0.6567685589519651,
+ "grad_norm": 0.08832395076751709,
+ "learning_rate": 0.000276,
+ "loss": 7.679399490356445,
+ "step": 47
+ },
+ {
+ "epoch": 0.6707423580786026,
+ "grad_norm": 0.08816400915384293,
+ "learning_rate": 0.00028199999999999997,
+ "loss": 7.5789642333984375,
+ "step": 48
+ },
+ {
+ "epoch": 0.6847161572052402,
+ "grad_norm": 0.0861717164516449,
+ "learning_rate": 0.00028799999999999995,
+ "loss": 7.519214630126953,
+ "step": 49
+ },
+ {
+ "epoch": 0.6986899563318777,
+ "grad_norm": 0.08103746175765991,
+ "learning_rate": 0.000294,
+ "loss": 7.4475321769714355,
+ "step": 50
+ },
+ {
+ "epoch": 0.7126637554585152,
+ "grad_norm": 0.07545104622840881,
+ "learning_rate": 0.0003,
+ "loss": 7.401182174682617,
+ "step": 51
+ },
+ {
+ "epoch": 0.7266375545851529,
+ "grad_norm": 0.07042575627565384,
+ "learning_rate": 0.00030599999999999996,
+ "loss": 7.340173244476318,
+ "step": 52
+ },
+ {
+ "epoch": 0.7406113537117904,
+ "grad_norm": 0.06737572699785233,
+ "learning_rate": 0.000312,
+ "loss": 7.265697002410889,
+ "step": 53
+ },
+ {
+ "epoch": 0.754585152838428,
+ "grad_norm": 0.06151146441698074,
+ "learning_rate": 0.000318,
+ "loss": 7.231061935424805,
+ "step": 54
+ },
+ {
+ "epoch": 0.7685589519650655,
+ "grad_norm": 0.05638071522116661,
+ "learning_rate": 0.000324,
+ "loss": 7.195466995239258,
+ "step": 55
+ },
+ {
+ "epoch": 0.7825327510917031,
+ "grad_norm": 0.05084463208913803,
+ "learning_rate": 0.00033,
+ "loss": 7.137603759765625,
+ "step": 56
+ },
+ {
+ "epoch": 0.7965065502183406,
+ "grad_norm": 0.04293261095881462,
+ "learning_rate": 0.000336,
+ "loss": 7.115458965301514,
+ "step": 57
+ },
+ {
+ "epoch": 0.8104803493449781,
+ "grad_norm": 0.036064863204956055,
+ "learning_rate": 0.00034199999999999996,
+ "loss": 7.089038848876953,
+ "step": 58
+ },
+ {
+ "epoch": 0.8244541484716157,
+ "grad_norm": 0.029026754200458527,
+ "learning_rate": 0.00034799999999999995,
+ "loss": 7.0584306716918945,
+ "step": 59
+ },
+ {
+ "epoch": 0.8384279475982532,
+ "grad_norm": 0.02286517433822155,
+ "learning_rate": 0.00035399999999999993,
+ "loss": 7.049450874328613,
+ "step": 60
+ },
+ {
+ "epoch": 0.8524017467248908,
+ "grad_norm": 0.018878433853387833,
+ "learning_rate": 0.00035999999999999997,
+ "loss": 7.008824348449707,
+ "step": 61
+ },
+ {
+ "epoch": 0.8663755458515284,
+ "grad_norm": 0.010773813351988792,
+ "learning_rate": 0.00036599999999999995,
+ "loss": 7.024829864501953,
+ "step": 62
+ },
+ {
+ "epoch": 0.880349344978166,
+ "grad_norm": 0.008794732391834259,
+ "learning_rate": 0.000372,
+ "loss": 7.018399238586426,
+ "step": 63
+ },
+ {
+ "epoch": 0.8943231441048035,
+ "grad_norm": 0.011709796264767647,
+ "learning_rate": 0.00037799999999999997,
+ "loss": 7.0101752281188965,
+ "step": 64
+ },
+ {
+ "epoch": 0.9082969432314411,
+ "grad_norm": 0.012123258784413338,
+ "learning_rate": 0.00038399999999999996,
+ "loss": 7.024317741394043,
+ "step": 65
+ },
+ {
+ "epoch": 0.9222707423580786,
+ "grad_norm": 0.01530320756137371,
+ "learning_rate": 0.00039,
+ "loss": 7.027738571166992,
+ "step": 66
+ },
+ {
+ "epoch": 0.9362445414847161,
+ "grad_norm": 0.01890002004802227,
+ "learning_rate": 0.000396,
+ "loss": 7.024099349975586,
+ "step": 67
+ },
+ {
+ "epoch": 0.9502183406113537,
+ "grad_norm": 0.020684150978922844,
+ "learning_rate": 0.000402,
+ "loss": 7.007943153381348,
+ "step": 68
+ },
+ {
+ "epoch": 0.9641921397379912,
+ "grad_norm": 0.019826196134090424,
+ "learning_rate": 0.000408,
+ "loss": 7.006874084472656,
+ "step": 69
+ },
+ {
+ "epoch": 0.9781659388646288,
+ "grad_norm": 0.019218673929572105,
+ "learning_rate": 0.0004139999999999999,
+ "loss": 7.002897262573242,
+ "step": 70
+ },
+ {
+ "epoch": 0.9921397379912664,
+ "grad_norm": 0.013856411911547184,
+ "learning_rate": 0.00041999999999999996,
+ "loss": 6.993999481201172,
+ "step": 71
+ },
+ {
+ "epoch": 1.0,
+ "grad_norm": 0.010390003211796284,
+ "learning_rate": 0.00042599999999999995,
+ "loss": 7.000683784484863,
+ "step": 72
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 28800,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 400,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": true,
+ "should_log": false,
+ "should_save": false,
+ "should_training_stop": false
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 3.069613361332224e+16,
+ "train_batch_size": 8,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-72/training_args.bin b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-72/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..c799bcf4051249aabf4c62fff9d38801df7bc2b5
--- /dev/null
+++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-72/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:40fb25263250c08c3479a6f0ac9fdf1f196e2a3b08a4eb10384786ff9b1903ef
+size 4792
diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-864/chat_template.jinja b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-864/chat_template.jinja
new file mode 100644
index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230
--- /dev/null
+++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-864/chat_template.jinja
@@ -0,0 +1,85 @@
+{%- if tools %}
+ {{- '<|im_start|>system\n' }}
+ {%- if messages[0].role == 'system' %}
+ {{- messages[0].content + '\n\n' }}
+ {%- endif %}
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }}
+ {%- for tool in tools %}
+ {{- "\n" }}
+ {{- tool | tojson }}
+ {%- endfor %}
+ {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }}
+{%- else %}
+ {%- if messages[0].role == 'system' %}
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
+ {%- endif %}
+{%- endif %}
+{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
+{%- for message in messages[::-1] %}
+ {%- set index = (messages|length - 1) - loop.index0 %}
+ {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %}
+ {%- set ns.multi_step_tool = false %}
+ {%- set ns.last_query_index = index %}
+ {%- endif %}
+{%- endfor %}
+{%- for message in messages %}
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
+ {%- elif message.role == "assistant" %}
+ {%- set content = message.content %}
+ {%- set reasoning_content = '' %}
+ {%- if message.reasoning_content is defined and message.reasoning_content is not none %}
+ {%- set reasoning_content = message.reasoning_content %}
+ {%- else %}
+ {%- if '' in message.content %}
+ {%- set content = message.content.split('')[-1].lstrip('\n') %}
+ {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %}
+ {%- endif %}
+ {%- endif %}
+ {%- if loop.index0 > ns.last_query_index %}
+ {%- if loop.last or (not loop.last and reasoning_content) %}
+ {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- if message.tool_calls %}
+ {%- for tool_call in message.tool_calls %}
+ {%- if (loop.first and content) or (not loop.first) %}
+ {{- '\n' }}
+ {%- endif %}
+ {%- if tool_call.function %}
+ {%- set tool_call = tool_call.function %}
+ {%- endif %}
+ {{- '\n{"name": "' }}
+ {{- tool_call.name }}
+ {{- '", "arguments": ' }}
+ {%- if tool_call.arguments is string %}
+ {{- tool_call.arguments }}
+ {%- else %}
+ {{- tool_call.arguments | tojson }}
+ {%- endif %}
+ {{- '}\n' }}
+ {%- endfor %}
+ {%- endif %}
+ {{- '<|im_end|>\n' }}
+ {%- elif message.role == "tool" %}
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
+ {{- '<|im_start|>user' }}
+ {%- endif %}
+ {{- '\n\n' }}
+ {{- message.content }}
+ {{- '\n' }}
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
+ {{- '<|im_end|>\n' }}
+ {%- endif %}
+ {%- endif %}
+{%- endfor %}
+{%- if add_generation_prompt %}
+ {{- '<|im_start|>assistant\n' }}
+ {%- if enable_thinking is defined and enable_thinking is false %}
+ {{- '\n\n\n\n' }}
+ {%- endif %}
+{%- endif %}
\ No newline at end of file
diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-864/config.json b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-864/config.json
new file mode 100644
index 0000000000000000000000000000000000000000..175ee3431793018f23f74824f69554b21afcbd4f
--- /dev/null
+++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-864/config.json
@@ -0,0 +1,32 @@
+{
+ "architectures": [
+ "LlamaForCausalLM"
+ ],
+ "attention_bias": false,
+ "attention_dropout": 0.0,
+ "bos_token_id": null,
+ "dtype": "float32",
+ "eos_token_id": 151645,
+ "head_dim": 128,
+ "hidden_act": "silu",
+ "hidden_size": 512,
+ "initializer_range": 0.02,
+ "intermediate_size": 1536,
+ "max_position_embeddings": 2048,
+ "mlp_bias": false,
+ "model_type": "llama",
+ "num_attention_heads": 4,
+ "num_hidden_layers": 20,
+ "num_key_value_heads": 4,
+ "pad_token_id": 151645,
+ "pretraining_tp": 1,
+ "rms_norm_eps": 1e-06,
+ "rope_parameters": {
+ "rope_theta": 10000.0,
+ "rope_type": "default"
+ },
+ "tie_word_embeddings": true,
+ "transformers_version": "5.5.0",
+ "use_cache": false,
+ "vocab_size": 151674
+}
diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-864/generation_config.json b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-864/generation_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3
--- /dev/null
+++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-864/generation_config.json
@@ -0,0 +1,11 @@
+{
+ "_from_model_config": true,
+ "eos_token_id": [
+ 151645
+ ],
+ "output_attentions": false,
+ "output_hidden_states": false,
+ "pad_token_id": 151645,
+ "transformers_version": "5.5.0",
+ "use_cache": true
+}
diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-864/model.safetensors b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-864/model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..c4db0d8d056f67e0168a10695de929a8ce130a69
--- /dev/null
+++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-864/model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:1a15b1846227dee8f858b94836547f08520546232e3813c085490d2ba5c6d725
+size 583362376
diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-864/optimizer.pt b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-864/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..e8f771496f9d9ade0a40bf0d852c47e2b4551c06
--- /dev/null
+++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-864/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:1659a40ad5174425855429c58a7e46d8cc0ef902b4bbf178f1a54296d8b98fbb
+size 1166837626
diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-864/rng_state_0.pth b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-864/rng_state_0.pth
new file mode 100644
index 0000000000000000000000000000000000000000..870ed7cdb8744b07d1ff1f9112387c96a63c2807
--- /dev/null
+++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-864/rng_state_0.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:dad714e377efb4a1838314db7e924ef1cd99f199b53380b22d6c44b28966149b
+size 15024
diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-864/rng_state_1.pth b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-864/rng_state_1.pth
new file mode 100644
index 0000000000000000000000000000000000000000..5925e0f1a7b68861f2b2cc205a9eff31aa6648fb
--- /dev/null
+++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-864/rng_state_1.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:d0999d19c49c17c3519518d23620eff41a1653e67643d57a5770816881df7313
+size 15024
diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-864/rng_state_2.pth b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-864/rng_state_2.pth
new file mode 100644
index 0000000000000000000000000000000000000000..8fd148eaedcb574a10f491ff8aadebd0bcdaabbe
--- /dev/null
+++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-864/rng_state_2.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:016c5f9e311d3225842840635607ff7f8e43814071a963a3c3f1559b8937b7c9
+size 15024
diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-864/rng_state_3.pth b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-864/rng_state_3.pth
new file mode 100644
index 0000000000000000000000000000000000000000..94829b15fcfd51a5cc4108f3610ac808956477b6
--- /dev/null
+++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-864/rng_state_3.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:5539b9e658812aa3870e38193d2b0bdc5b6664af0e769e6a604f2a69a04a9c9d
+size 15024
diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-864/scheduler.pt b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-864/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..f21fc027a37dccb66b77a4752de227c557d8ff04
--- /dev/null
+++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-864/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:b1a1dda88112968a1342ac518f362b4f5e138437f0b670c818ad8493b91ca49d
+size 1064
diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-864/tokenizer.json b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-864/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..9ea28d9eb325c36bcab4c1a08fc93fd598929f08
--- /dev/null
+++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-864/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:60e99075d371ecbb9f1dde97363f612d45bf0bf88f1a2006d9bfcf1939333c14
+size 11424648
diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-864/tokenizer_config.json b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-864/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..cadd997fc26040d23ad6677fcb39d8d18b01f55e
--- /dev/null
+++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-864/tokenizer_config.json
@@ -0,0 +1,21 @@
+{
+ "add_prefix_space": false,
+ "backend": "tokenizers",
+ "bos_token": null,
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "<|im_end|>",
+ "errors": "replace",
+ "extra_special_tokens": [
+ "<|l2r_pred|>",
+ "<|r2l_pred|>",
+ "<|next_1_pred|>",
+ "<|next_2_pred|>",
+ "<|mask|>"
+ ],
+ "is_local": false,
+ "model_max_length": 131072,
+ "pad_token": "<|im_end|>",
+ "split_special_tokens": false,
+ "tokenizer_class": "Qwen2Tokenizer",
+ "unk_token": null
+}
diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-864/trainer_state.json b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-864/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..fb8c37d335462658260ed45e6615de9045152ec1
--- /dev/null
+++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-864/trainer_state.json
@@ -0,0 +1,6170 @@
+{
+ "best_global_step": null,
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 12.0,
+ "eval_steps": 500,
+ "global_step": 864,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.013973799126637555,
+ "grad_norm": 0.5768774151802063,
+ "learning_rate": 0.0,
+ "loss": 12.050328254699707,
+ "step": 1
+ },
+ {
+ "epoch": 0.02794759825327511,
+ "grad_norm": 0.5784570574760437,
+ "learning_rate": 5.999999999999999e-06,
+ "loss": 12.049251556396484,
+ "step": 2
+ },
+ {
+ "epoch": 0.04192139737991266,
+ "grad_norm": 0.5607187151908875,
+ "learning_rate": 1.1999999999999999e-05,
+ "loss": 11.921792030334473,
+ "step": 3
+ },
+ {
+ "epoch": 0.05589519650655022,
+ "grad_norm": 0.46466919779777527,
+ "learning_rate": 1.7999999999999997e-05,
+ "loss": 11.686628341674805,
+ "step": 4
+ },
+ {
+ "epoch": 0.06986899563318777,
+ "grad_norm": 0.38012945652008057,
+ "learning_rate": 2.3999999999999997e-05,
+ "loss": 11.432823181152344,
+ "step": 5
+ },
+ {
+ "epoch": 0.08384279475982533,
+ "grad_norm": 0.33215776085853577,
+ "learning_rate": 2.9999999999999997e-05,
+ "loss": 11.207542419433594,
+ "step": 6
+ },
+ {
+ "epoch": 0.09781659388646288,
+ "grad_norm": 0.2928776741027832,
+ "learning_rate": 3.5999999999999994e-05,
+ "loss": 11.000097274780273,
+ "step": 7
+ },
+ {
+ "epoch": 0.11179039301310044,
+ "grad_norm": 0.26018962264060974,
+ "learning_rate": 4.2e-05,
+ "loss": 10.825748443603516,
+ "step": 8
+ },
+ {
+ "epoch": 0.125764192139738,
+ "grad_norm": 0.23717083036899567,
+ "learning_rate": 4.7999999999999994e-05,
+ "loss": 10.677122116088867,
+ "step": 9
+ },
+ {
+ "epoch": 0.13973799126637554,
+ "grad_norm": 0.21641355752944946,
+ "learning_rate": 5.399999999999999e-05,
+ "loss": 10.569148063659668,
+ "step": 10
+ },
+ {
+ "epoch": 0.1537117903930131,
+ "grad_norm": 0.1982956826686859,
+ "learning_rate": 5.9999999999999995e-05,
+ "loss": 10.49653434753418,
+ "step": 11
+ },
+ {
+ "epoch": 0.16768558951965065,
+ "grad_norm": 0.18819282948970795,
+ "learning_rate": 6.599999999999999e-05,
+ "loss": 10.432951927185059,
+ "step": 12
+ },
+ {
+ "epoch": 0.18165938864628822,
+ "grad_norm": 0.18322618305683136,
+ "learning_rate": 7.199999999999999e-05,
+ "loss": 10.387115478515625,
+ "step": 13
+ },
+ {
+ "epoch": 0.19563318777292577,
+ "grad_norm": 0.18399226665496826,
+ "learning_rate": 7.8e-05,
+ "loss": 10.33381462097168,
+ "step": 14
+ },
+ {
+ "epoch": 0.2096069868995633,
+ "grad_norm": 0.1833190619945526,
+ "learning_rate": 8.4e-05,
+ "loss": 10.301689147949219,
+ "step": 15
+ },
+ {
+ "epoch": 0.22358078602620088,
+ "grad_norm": 0.18490834534168243,
+ "learning_rate": 8.999999999999999e-05,
+ "loss": 10.240968704223633,
+ "step": 16
+ },
+ {
+ "epoch": 0.23755458515283842,
+ "grad_norm": 0.18358023464679718,
+ "learning_rate": 9.599999999999999e-05,
+ "loss": 10.194717407226562,
+ "step": 17
+ },
+ {
+ "epoch": 0.251528384279476,
+ "grad_norm": 0.1801682710647583,
+ "learning_rate": 0.000102,
+ "loss": 10.14897632598877,
+ "step": 18
+ },
+ {
+ "epoch": 0.26550218340611353,
+ "grad_norm": 0.17899440228939056,
+ "learning_rate": 0.00010799999999999998,
+ "loss": 10.078763961791992,
+ "step": 19
+ },
+ {
+ "epoch": 0.2794759825327511,
+ "grad_norm": 0.17666497826576233,
+ "learning_rate": 0.00011399999999999999,
+ "loss": 10.01095962524414,
+ "step": 20
+ },
+ {
+ "epoch": 0.2934497816593886,
+ "grad_norm": 0.17196154594421387,
+ "learning_rate": 0.00011999999999999999,
+ "loss": 9.951203346252441,
+ "step": 21
+ },
+ {
+ "epoch": 0.3074235807860262,
+ "grad_norm": 0.169203519821167,
+ "learning_rate": 0.00012599999999999997,
+ "loss": 9.871726036071777,
+ "step": 22
+ },
+ {
+ "epoch": 0.32139737991266376,
+ "grad_norm": 0.16500306129455566,
+ "learning_rate": 0.00013199999999999998,
+ "loss": 9.803646087646484,
+ "step": 23
+ },
+ {
+ "epoch": 0.3353711790393013,
+ "grad_norm": 0.16128124296665192,
+ "learning_rate": 0.000138,
+ "loss": 9.728487014770508,
+ "step": 24
+ },
+ {
+ "epoch": 0.34934497816593885,
+ "grad_norm": 0.15904058516025543,
+ "learning_rate": 0.00014399999999999998,
+ "loss": 9.642473220825195,
+ "step": 25
+ },
+ {
+ "epoch": 0.36331877729257644,
+ "grad_norm": 0.15560561418533325,
+ "learning_rate": 0.00015,
+ "loss": 9.56414794921875,
+ "step": 26
+ },
+ {
+ "epoch": 0.377292576419214,
+ "grad_norm": 0.15278612077236176,
+ "learning_rate": 0.000156,
+ "loss": 9.484230041503906,
+ "step": 27
+ },
+ {
+ "epoch": 0.39126637554585153,
+ "grad_norm": 0.15263237059116364,
+ "learning_rate": 0.000162,
+ "loss": 9.383644104003906,
+ "step": 28
+ },
+ {
+ "epoch": 0.4052401746724891,
+ "grad_norm": 0.14945024251937866,
+ "learning_rate": 0.000168,
+ "loss": 9.29840087890625,
+ "step": 29
+ },
+ {
+ "epoch": 0.4192139737991266,
+ "grad_norm": 0.14614646136760712,
+ "learning_rate": 0.00017399999999999997,
+ "loss": 9.208931922912598,
+ "step": 30
+ },
+ {
+ "epoch": 0.4331877729257642,
+ "grad_norm": 0.14295098185539246,
+ "learning_rate": 0.00017999999999999998,
+ "loss": 9.122323989868164,
+ "step": 31
+ },
+ {
+ "epoch": 0.44716157205240176,
+ "grad_norm": 0.13868345320224762,
+ "learning_rate": 0.000186,
+ "loss": 9.03034496307373,
+ "step": 32
+ },
+ {
+ "epoch": 0.4611353711790393,
+ "grad_norm": 0.135064035654068,
+ "learning_rate": 0.00019199999999999998,
+ "loss": 8.92818832397461,
+ "step": 33
+ },
+ {
+ "epoch": 0.47510917030567684,
+ "grad_norm": 0.13097241520881653,
+ "learning_rate": 0.000198,
+ "loss": 8.834627151489258,
+ "step": 34
+ },
+ {
+ "epoch": 0.4890829694323144,
+ "grad_norm": 0.12478776276111603,
+ "learning_rate": 0.000204,
+ "loss": 8.755172729492188,
+ "step": 35
+ },
+ {
+ "epoch": 0.503056768558952,
+ "grad_norm": 0.1214982345700264,
+ "learning_rate": 0.00020999999999999998,
+ "loss": 8.647891998291016,
+ "step": 36
+ },
+ {
+ "epoch": 0.5170305676855895,
+ "grad_norm": 0.11958328634500504,
+ "learning_rate": 0.00021599999999999996,
+ "loss": 8.543274879455566,
+ "step": 37
+ },
+ {
+ "epoch": 0.5310043668122271,
+ "grad_norm": 0.11680915951728821,
+ "learning_rate": 0.00022199999999999998,
+ "loss": 8.44540023803711,
+ "step": 38
+ },
+ {
+ "epoch": 0.5449781659388646,
+ "grad_norm": 0.11491474509239197,
+ "learning_rate": 0.00022799999999999999,
+ "loss": 8.349921226501465,
+ "step": 39
+ },
+ {
+ "epoch": 0.5589519650655022,
+ "grad_norm": 0.11160755902528763,
+ "learning_rate": 0.000234,
+ "loss": 8.26960563659668,
+ "step": 40
+ },
+ {
+ "epoch": 0.5729257641921397,
+ "grad_norm": 0.10840637236833572,
+ "learning_rate": 0.00023999999999999998,
+ "loss": 8.192136764526367,
+ "step": 41
+ },
+ {
+ "epoch": 0.5868995633187772,
+ "grad_norm": 0.10585298389196396,
+ "learning_rate": 0.00024599999999999996,
+ "loss": 8.09546947479248,
+ "step": 42
+ },
+ {
+ "epoch": 0.6008733624454149,
+ "grad_norm": 0.1032542735338211,
+ "learning_rate": 0.00025199999999999995,
+ "loss": 7.991090774536133,
+ "step": 43
+ },
+ {
+ "epoch": 0.6148471615720524,
+ "grad_norm": 0.0986943170428276,
+ "learning_rate": 0.000258,
+ "loss": 7.92384147644043,
+ "step": 44
+ },
+ {
+ "epoch": 0.62882096069869,
+ "grad_norm": 0.09320762753486633,
+ "learning_rate": 0.00026399999999999997,
+ "loss": 7.8478193283081055,
+ "step": 45
+ },
+ {
+ "epoch": 0.6427947598253275,
+ "grad_norm": 0.09168332070112228,
+ "learning_rate": 0.00027,
+ "loss": 7.76185417175293,
+ "step": 46
+ },
+ {
+ "epoch": 0.6567685589519651,
+ "grad_norm": 0.08832395076751709,
+ "learning_rate": 0.000276,
+ "loss": 7.679399490356445,
+ "step": 47
+ },
+ {
+ "epoch": 0.6707423580786026,
+ "grad_norm": 0.08816400915384293,
+ "learning_rate": 0.00028199999999999997,
+ "loss": 7.5789642333984375,
+ "step": 48
+ },
+ {
+ "epoch": 0.6847161572052402,
+ "grad_norm": 0.0861717164516449,
+ "learning_rate": 0.00028799999999999995,
+ "loss": 7.519214630126953,
+ "step": 49
+ },
+ {
+ "epoch": 0.6986899563318777,
+ "grad_norm": 0.08103746175765991,
+ "learning_rate": 0.000294,
+ "loss": 7.4475321769714355,
+ "step": 50
+ },
+ {
+ "epoch": 0.7126637554585152,
+ "grad_norm": 0.07545104622840881,
+ "learning_rate": 0.0003,
+ "loss": 7.401182174682617,
+ "step": 51
+ },
+ {
+ "epoch": 0.7266375545851529,
+ "grad_norm": 0.07042575627565384,
+ "learning_rate": 0.00030599999999999996,
+ "loss": 7.340173244476318,
+ "step": 52
+ },
+ {
+ "epoch": 0.7406113537117904,
+ "grad_norm": 0.06737572699785233,
+ "learning_rate": 0.000312,
+ "loss": 7.265697002410889,
+ "step": 53
+ },
+ {
+ "epoch": 0.754585152838428,
+ "grad_norm": 0.06151146441698074,
+ "learning_rate": 0.000318,
+ "loss": 7.231061935424805,
+ "step": 54
+ },
+ {
+ "epoch": 0.7685589519650655,
+ "grad_norm": 0.05638071522116661,
+ "learning_rate": 0.000324,
+ "loss": 7.195466995239258,
+ "step": 55
+ },
+ {
+ "epoch": 0.7825327510917031,
+ "grad_norm": 0.05084463208913803,
+ "learning_rate": 0.00033,
+ "loss": 7.137603759765625,
+ "step": 56
+ },
+ {
+ "epoch": 0.7965065502183406,
+ "grad_norm": 0.04293261095881462,
+ "learning_rate": 0.000336,
+ "loss": 7.115458965301514,
+ "step": 57
+ },
+ {
+ "epoch": 0.8104803493449781,
+ "grad_norm": 0.036064863204956055,
+ "learning_rate": 0.00034199999999999996,
+ "loss": 7.089038848876953,
+ "step": 58
+ },
+ {
+ "epoch": 0.8244541484716157,
+ "grad_norm": 0.029026754200458527,
+ "learning_rate": 0.00034799999999999995,
+ "loss": 7.0584306716918945,
+ "step": 59
+ },
+ {
+ "epoch": 0.8384279475982532,
+ "grad_norm": 0.02286517433822155,
+ "learning_rate": 0.00035399999999999993,
+ "loss": 7.049450874328613,
+ "step": 60
+ },
+ {
+ "epoch": 0.8524017467248908,
+ "grad_norm": 0.018878433853387833,
+ "learning_rate": 0.00035999999999999997,
+ "loss": 7.008824348449707,
+ "step": 61
+ },
+ {
+ "epoch": 0.8663755458515284,
+ "grad_norm": 0.010773813351988792,
+ "learning_rate": 0.00036599999999999995,
+ "loss": 7.024829864501953,
+ "step": 62
+ },
+ {
+ "epoch": 0.880349344978166,
+ "grad_norm": 0.008794732391834259,
+ "learning_rate": 0.000372,
+ "loss": 7.018399238586426,
+ "step": 63
+ },
+ {
+ "epoch": 0.8943231441048035,
+ "grad_norm": 0.011709796264767647,
+ "learning_rate": 0.00037799999999999997,
+ "loss": 7.0101752281188965,
+ "step": 64
+ },
+ {
+ "epoch": 0.9082969432314411,
+ "grad_norm": 0.012123258784413338,
+ "learning_rate": 0.00038399999999999996,
+ "loss": 7.024317741394043,
+ "step": 65
+ },
+ {
+ "epoch": 0.9222707423580786,
+ "grad_norm": 0.01530320756137371,
+ "learning_rate": 0.00039,
+ "loss": 7.027738571166992,
+ "step": 66
+ },
+ {
+ "epoch": 0.9362445414847161,
+ "grad_norm": 0.01890002004802227,
+ "learning_rate": 0.000396,
+ "loss": 7.024099349975586,
+ "step": 67
+ },
+ {
+ "epoch": 0.9502183406113537,
+ "grad_norm": 0.020684150978922844,
+ "learning_rate": 0.000402,
+ "loss": 7.007943153381348,
+ "step": 68
+ },
+ {
+ "epoch": 0.9641921397379912,
+ "grad_norm": 0.019826196134090424,
+ "learning_rate": 0.000408,
+ "loss": 7.006874084472656,
+ "step": 69
+ },
+ {
+ "epoch": 0.9781659388646288,
+ "grad_norm": 0.019218673929572105,
+ "learning_rate": 0.0004139999999999999,
+ "loss": 7.002897262573242,
+ "step": 70
+ },
+ {
+ "epoch": 0.9921397379912664,
+ "grad_norm": 0.013856411911547184,
+ "learning_rate": 0.00041999999999999996,
+ "loss": 6.993999481201172,
+ "step": 71
+ },
+ {
+ "epoch": 1.0,
+ "grad_norm": 0.010390003211796284,
+ "learning_rate": 0.00042599999999999995,
+ "loss": 7.000683784484863,
+ "step": 72
+ },
+ {
+ "epoch": 1.0,
+ "eval_loss": 7.91140079498291,
+ "eval_runtime": 57.7925,
+ "eval_samples_per_second": 42.255,
+ "eval_steps_per_second": 1.332,
+ "step": 72
+ },
+ {
+ "epoch": 1.0139737991266375,
+ "grad_norm": 0.013802163302898407,
+ "learning_rate": 0.00043199999999999993,
+ "loss": 6.998257637023926,
+ "step": 73
+ },
+ {
+ "epoch": 1.027947598253275,
+ "grad_norm": 0.010109687224030495,
+ "learning_rate": 0.00043799999999999997,
+ "loss": 7.002542018890381,
+ "step": 74
+ },
+ {
+ "epoch": 1.0419213973799126,
+ "grad_norm": 0.008019139990210533,
+ "learning_rate": 0.00044399999999999995,
+ "loss": 6.985941410064697,
+ "step": 75
+ },
+ {
+ "epoch": 1.0558951965065502,
+ "grad_norm": 0.0069672465324401855,
+ "learning_rate": 0.00045,
+ "loss": 6.99245023727417,
+ "step": 76
+ },
+ {
+ "epoch": 1.0698689956331877,
+ "grad_norm": 0.00714076729491353,
+ "learning_rate": 0.00045599999999999997,
+ "loss": 6.959296226501465,
+ "step": 77
+ },
+ {
+ "epoch": 1.0838427947598253,
+ "grad_norm": 0.008408435620367527,
+ "learning_rate": 0.00046199999999999995,
+ "loss": 6.954504489898682,
+ "step": 78
+ },
+ {
+ "epoch": 1.0978165938864628,
+ "grad_norm": 0.008778786286711693,
+ "learning_rate": 0.000468,
+ "loss": 7.0007004737854,
+ "step": 79
+ },
+ {
+ "epoch": 1.1117903930131003,
+ "grad_norm": 0.007753407116979361,
+ "learning_rate": 0.000474,
+ "loss": 6.98183536529541,
+ "step": 80
+ },
+ {
+ "epoch": 1.125764192139738,
+ "grad_norm": 0.01087260153144598,
+ "learning_rate": 0.00047999999999999996,
+ "loss": 6.991066932678223,
+ "step": 81
+ },
+ {
+ "epoch": 1.1397379912663754,
+ "grad_norm": 0.006122548598796129,
+ "learning_rate": 0.000486,
+ "loss": 6.978824138641357,
+ "step": 82
+ },
+ {
+ "epoch": 1.1537117903930132,
+ "grad_norm": 0.0067383465357124805,
+ "learning_rate": 0.0004919999999999999,
+ "loss": 6.961703777313232,
+ "step": 83
+ },
+ {
+ "epoch": 1.1676855895196507,
+ "grad_norm": 0.007922791875898838,
+ "learning_rate": 0.000498,
+ "loss": 6.954860687255859,
+ "step": 84
+ },
+ {
+ "epoch": 1.1816593886462883,
+ "grad_norm": 0.00818922370672226,
+ "learning_rate": 0.0005039999999999999,
+ "loss": 6.945961952209473,
+ "step": 85
+ },
+ {
+ "epoch": 1.1956331877729258,
+ "grad_norm": 0.006791461259126663,
+ "learning_rate": 0.0005099999999999999,
+ "loss": 6.968680381774902,
+ "step": 86
+ },
+ {
+ "epoch": 1.2096069868995634,
+ "grad_norm": 0.005394163075834513,
+ "learning_rate": 0.000516,
+ "loss": 6.956404685974121,
+ "step": 87
+ },
+ {
+ "epoch": 1.223580786026201,
+ "grad_norm": 0.004500220064073801,
+ "learning_rate": 0.000522,
+ "loss": 6.96998405456543,
+ "step": 88
+ },
+ {
+ "epoch": 1.2375545851528384,
+ "grad_norm": 0.008699574507772923,
+ "learning_rate": 0.0005279999999999999,
+ "loss": 6.955585479736328,
+ "step": 89
+ },
+ {
+ "epoch": 1.251528384279476,
+ "grad_norm": 0.005889365915209055,
+ "learning_rate": 0.000534,
+ "loss": 6.9708051681518555,
+ "step": 90
+ },
+ {
+ "epoch": 1.2655021834061135,
+ "grad_norm": 0.009646810591220856,
+ "learning_rate": 0.00054,
+ "loss": 6.962608814239502,
+ "step": 91
+ },
+ {
+ "epoch": 1.279475982532751,
+ "grad_norm": 0.004517171997576952,
+ "learning_rate": 0.0005459999999999999,
+ "loss": 6.957627296447754,
+ "step": 92
+ },
+ {
+ "epoch": 1.2934497816593886,
+ "grad_norm": 0.006339677143841982,
+ "learning_rate": 0.000552,
+ "loss": 6.963141441345215,
+ "step": 93
+ },
+ {
+ "epoch": 1.3074235807860262,
+ "grad_norm": 0.005387383047491312,
+ "learning_rate": 0.000558,
+ "loss": 6.934465408325195,
+ "step": 94
+ },
+ {
+ "epoch": 1.3213973799126637,
+ "grad_norm": 0.005252422299236059,
+ "learning_rate": 0.0005639999999999999,
+ "loss": 6.956108093261719,
+ "step": 95
+ },
+ {
+ "epoch": 1.3353711790393012,
+ "grad_norm": 0.008699414320290089,
+ "learning_rate": 0.00057,
+ "loss": 6.973369598388672,
+ "step": 96
+ },
+ {
+ "epoch": 1.3493449781659388,
+ "grad_norm": 0.003947409335523844,
+ "learning_rate": 0.0005759999999999999,
+ "loss": 6.950225830078125,
+ "step": 97
+ },
+ {
+ "epoch": 1.3633187772925766,
+ "grad_norm": 0.00768681475892663,
+ "learning_rate": 0.0005819999999999999,
+ "loss": 6.964739799499512,
+ "step": 98
+ },
+ {
+ "epoch": 1.3772925764192139,
+ "grad_norm": 0.006563829258084297,
+ "learning_rate": 0.000588,
+ "loss": 6.972421169281006,
+ "step": 99
+ },
+ {
+ "epoch": 1.3912663755458516,
+ "grad_norm": 0.007606555242091417,
+ "learning_rate": 0.0005939999999999999,
+ "loss": 6.958213806152344,
+ "step": 100
+ },
+ {
+ "epoch": 1.405240174672489,
+ "grad_norm": 0.006250880658626556,
+ "learning_rate": 0.0006,
+ "loss": 6.962296485900879,
+ "step": 101
+ },
+ {
+ "epoch": 1.4192139737991267,
+ "grad_norm": 0.0073033408261835575,
+ "learning_rate": 0.0006,
+ "loss": 6.969705104827881,
+ "step": 102
+ },
+ {
+ "epoch": 1.4331877729257643,
+ "grad_norm": 0.0054510910995304585,
+ "learning_rate": 0.0006,
+ "loss": 6.9652862548828125,
+ "step": 103
+ },
+ {
+ "epoch": 1.4471615720524018,
+ "grad_norm": 0.004895673133432865,
+ "learning_rate": 0.0006,
+ "loss": 6.956023693084717,
+ "step": 104
+ },
+ {
+ "epoch": 1.4611353711790394,
+ "grad_norm": 0.005306515377014875,
+ "learning_rate": 0.0006,
+ "loss": 6.963231086730957,
+ "step": 105
+ },
+ {
+ "epoch": 1.475109170305677,
+ "grad_norm": 0.004742146469652653,
+ "learning_rate": 0.0006,
+ "loss": 6.9582061767578125,
+ "step": 106
+ },
+ {
+ "epoch": 1.4890829694323144,
+ "grad_norm": 0.004220000468194485,
+ "learning_rate": 0.0006,
+ "loss": 6.972961902618408,
+ "step": 107
+ },
+ {
+ "epoch": 1.503056768558952,
+ "grad_norm": 0.0042153168469667435,
+ "learning_rate": 0.0006,
+ "loss": 6.961165904998779,
+ "step": 108
+ },
+ {
+ "epoch": 1.5170305676855895,
+ "grad_norm": 0.005165040958672762,
+ "learning_rate": 0.0006,
+ "loss": 6.95018196105957,
+ "step": 109
+ },
+ {
+ "epoch": 1.531004366812227,
+ "grad_norm": 0.003930266480892897,
+ "learning_rate": 0.0006,
+ "loss": 6.9651689529418945,
+ "step": 110
+ },
+ {
+ "epoch": 1.5449781659388646,
+ "grad_norm": 0.004465071018785238,
+ "learning_rate": 0.0006,
+ "loss": 6.960553169250488,
+ "step": 111
+ },
+ {
+ "epoch": 1.5589519650655022,
+ "grad_norm": 0.004888612311333418,
+ "learning_rate": 0.0006,
+ "loss": 6.967258930206299,
+ "step": 112
+ },
+ {
+ "epoch": 1.5729257641921397,
+ "grad_norm": 0.007497863378375769,
+ "learning_rate": 0.0006,
+ "loss": 6.954545021057129,
+ "step": 113
+ },
+ {
+ "epoch": 1.5868995633187772,
+ "grad_norm": 0.005659648682922125,
+ "learning_rate": 0.0006,
+ "loss": 6.962299823760986,
+ "step": 114
+ },
+ {
+ "epoch": 1.600873362445415,
+ "grad_norm": 0.006641396787017584,
+ "learning_rate": 0.0006,
+ "loss": 6.952983856201172,
+ "step": 115
+ },
+ {
+ "epoch": 1.6148471615720523,
+ "grad_norm": 0.009715928696095943,
+ "learning_rate": 0.0006,
+ "loss": 6.958622932434082,
+ "step": 116
+ },
+ {
+ "epoch": 1.62882096069869,
+ "grad_norm": 0.012789523229002953,
+ "learning_rate": 0.0006,
+ "loss": 6.932043075561523,
+ "step": 117
+ },
+ {
+ "epoch": 1.6427947598253274,
+ "grad_norm": 0.021673237904906273,
+ "learning_rate": 0.0006,
+ "loss": 6.931290149688721,
+ "step": 118
+ },
+ {
+ "epoch": 1.6567685589519652,
+ "grad_norm": 0.16024991869926453,
+ "learning_rate": 0.0006,
+ "loss": 6.979497909545898,
+ "step": 119
+ },
+ {
+ "epoch": 1.6707423580786025,
+ "grad_norm": 0.2434523105621338,
+ "learning_rate": 0.0006,
+ "loss": 7.185218811035156,
+ "step": 120
+ },
+ {
+ "epoch": 1.6847161572052403,
+ "grad_norm": 0.39252421259880066,
+ "learning_rate": 0.0006,
+ "loss": 7.137226581573486,
+ "step": 121
+ },
+ {
+ "epoch": 1.6986899563318776,
+ "grad_norm": 0.10565312206745148,
+ "learning_rate": 0.0006,
+ "loss": 6.984494209289551,
+ "step": 122
+ },
+ {
+ "epoch": 1.7126637554585153,
+ "grad_norm": 0.1033431813120842,
+ "learning_rate": 0.0006,
+ "loss": 7.001287937164307,
+ "step": 123
+ },
+ {
+ "epoch": 1.726637554585153,
+ "grad_norm": 0.14875490963459015,
+ "learning_rate": 0.0006,
+ "loss": 7.051436901092529,
+ "step": 124
+ },
+ {
+ "epoch": 1.7406113537117904,
+ "grad_norm": 0.1374562531709671,
+ "learning_rate": 0.0006,
+ "loss": 7.0293121337890625,
+ "step": 125
+ },
+ {
+ "epoch": 1.754585152838428,
+ "grad_norm": 0.0910211056470871,
+ "learning_rate": 0.0006,
+ "loss": 7.007458686828613,
+ "step": 126
+ },
+ {
+ "epoch": 1.7685589519650655,
+ "grad_norm": 0.021633487194776535,
+ "learning_rate": 0.0006,
+ "loss": 6.996816635131836,
+ "step": 127
+ },
+ {
+ "epoch": 1.782532751091703,
+ "grad_norm": 0.08692754060029984,
+ "learning_rate": 0.0006,
+ "loss": 6.998541831970215,
+ "step": 128
+ },
+ {
+ "epoch": 1.7965065502183406,
+ "grad_norm": 0.12613292038440704,
+ "learning_rate": 0.0006,
+ "loss": 6.9970703125,
+ "step": 129
+ },
+ {
+ "epoch": 1.8104803493449781,
+ "grad_norm": 0.09462810307741165,
+ "learning_rate": 0.0006,
+ "loss": 6.997045040130615,
+ "step": 130
+ },
+ {
+ "epoch": 1.8244541484716157,
+ "grad_norm": 0.03534962609410286,
+ "learning_rate": 0.0006,
+ "loss": 6.9738450050354,
+ "step": 131
+ },
+ {
+ "epoch": 1.8384279475982532,
+ "grad_norm": 0.033839885145425797,
+ "learning_rate": 0.0006,
+ "loss": 6.946001052856445,
+ "step": 132
+ },
+ {
+ "epoch": 1.8524017467248908,
+ "grad_norm": 0.06263269484043121,
+ "learning_rate": 0.0006,
+ "loss": 6.968981742858887,
+ "step": 133
+ },
+ {
+ "epoch": 1.8663755458515285,
+ "grad_norm": 0.07117997109889984,
+ "learning_rate": 0.0006,
+ "loss": 6.981535911560059,
+ "step": 134
+ },
+ {
+ "epoch": 1.8803493449781659,
+ "grad_norm": 0.06336277723312378,
+ "learning_rate": 0.0006,
+ "loss": 6.976946830749512,
+ "step": 135
+ },
+ {
+ "epoch": 1.8943231441048036,
+ "grad_norm": 0.047467734664678574,
+ "learning_rate": 0.0006,
+ "loss": 6.961673736572266,
+ "step": 136
+ },
+ {
+ "epoch": 1.908296943231441,
+ "grad_norm": 0.018787868320941925,
+ "learning_rate": 0.0006,
+ "loss": 6.948919296264648,
+ "step": 137
+ },
+ {
+ "epoch": 1.9222707423580787,
+ "grad_norm": 0.025909408926963806,
+ "learning_rate": 0.0006,
+ "loss": 6.976190567016602,
+ "step": 138
+ },
+ {
+ "epoch": 1.936244541484716,
+ "grad_norm": 0.052842333912849426,
+ "learning_rate": 0.0006,
+ "loss": 6.971207618713379,
+ "step": 139
+ },
+ {
+ "epoch": 1.9502183406113538,
+ "grad_norm": 0.06421888619661331,
+ "learning_rate": 0.0006,
+ "loss": 6.979109764099121,
+ "step": 140
+ },
+ {
+ "epoch": 1.9641921397379911,
+ "grad_norm": 0.055596910417079926,
+ "learning_rate": 0.0006,
+ "loss": 6.972118854522705,
+ "step": 141
+ },
+ {
+ "epoch": 1.9781659388646289,
+ "grad_norm": 0.03267091140151024,
+ "learning_rate": 0.0006,
+ "loss": 6.947260856628418,
+ "step": 142
+ },
+ {
+ "epoch": 1.9921397379912664,
+ "grad_norm": 0.008032613433897495,
+ "learning_rate": 0.0006,
+ "loss": 6.936816215515137,
+ "step": 143
+ },
+ {
+ "epoch": 2.0,
+ "grad_norm": 0.020184025168418884,
+ "learning_rate": 0.0006,
+ "loss": 6.929848670959473,
+ "step": 144
+ },
+ {
+ "epoch": 2.0,
+ "eval_loss": 7.836523056030273,
+ "eval_runtime": 58.5622,
+ "eval_samples_per_second": 41.699,
+ "eval_steps_per_second": 1.315,
+ "step": 144
+ },
+ {
+ "epoch": 2.0139737991266378,
+ "grad_norm": 0.031980376690626144,
+ "learning_rate": 0.0006,
+ "loss": 6.968081474304199,
+ "step": 145
+ },
+ {
+ "epoch": 2.027947598253275,
+ "grad_norm": 0.03802533075213432,
+ "learning_rate": 0.0006,
+ "loss": 6.95728874206543,
+ "step": 146
+ },
+ {
+ "epoch": 2.041921397379913,
+ "grad_norm": 0.03796226158738136,
+ "learning_rate": 0.0006,
+ "loss": 6.96516752243042,
+ "step": 147
+ },
+ {
+ "epoch": 2.05589519650655,
+ "grad_norm": 0.02621951326727867,
+ "learning_rate": 0.0006,
+ "loss": 6.956058502197266,
+ "step": 148
+ },
+ {
+ "epoch": 2.069868995633188,
+ "grad_norm": 0.013292718678712845,
+ "learning_rate": 0.0006,
+ "loss": 6.9783935546875,
+ "step": 149
+ },
+ {
+ "epoch": 2.0838427947598253,
+ "grad_norm": 0.009446968324482441,
+ "learning_rate": 0.0006,
+ "loss": 6.96873664855957,
+ "step": 150
+ },
+ {
+ "epoch": 2.097816593886463,
+ "grad_norm": 0.021703092381358147,
+ "learning_rate": 0.0006,
+ "loss": 6.959234237670898,
+ "step": 151
+ },
+ {
+ "epoch": 2.1117903930131003,
+ "grad_norm": 0.027765290811657906,
+ "learning_rate": 0.0006,
+ "loss": 6.968524932861328,
+ "step": 152
+ },
+ {
+ "epoch": 2.125764192139738,
+ "grad_norm": 0.028761165216565132,
+ "learning_rate": 0.0006,
+ "loss": 6.954620838165283,
+ "step": 153
+ },
+ {
+ "epoch": 2.1397379912663754,
+ "grad_norm": 0.019729383289813995,
+ "learning_rate": 0.0006,
+ "loss": 6.949009895324707,
+ "step": 154
+ },
+ {
+ "epoch": 2.153711790393013,
+ "grad_norm": 0.008990089409053326,
+ "learning_rate": 0.0006,
+ "loss": 6.956226348876953,
+ "step": 155
+ },
+ {
+ "epoch": 2.1676855895196505,
+ "grad_norm": 0.00549644511193037,
+ "learning_rate": 0.0006,
+ "loss": 6.957143783569336,
+ "step": 156
+ },
+ {
+ "epoch": 2.1816593886462883,
+ "grad_norm": 0.014436670579016209,
+ "learning_rate": 0.0006,
+ "loss": 6.954257011413574,
+ "step": 157
+ },
+ {
+ "epoch": 2.1956331877729256,
+ "grad_norm": 0.01990494504570961,
+ "learning_rate": 0.0006,
+ "loss": 6.942131519317627,
+ "step": 158
+ },
+ {
+ "epoch": 2.2096069868995634,
+ "grad_norm": 0.019155776128172874,
+ "learning_rate": 0.0006,
+ "loss": 6.946969985961914,
+ "step": 159
+ },
+ {
+ "epoch": 2.2235807860262007,
+ "grad_norm": 0.014970783144235611,
+ "learning_rate": 0.0006,
+ "loss": 6.957938194274902,
+ "step": 160
+ },
+ {
+ "epoch": 2.2375545851528384,
+ "grad_norm": 0.009875659830868244,
+ "learning_rate": 0.0006,
+ "loss": 6.953309059143066,
+ "step": 161
+ },
+ {
+ "epoch": 2.251528384279476,
+ "grad_norm": 0.004792087245732546,
+ "learning_rate": 0.0006,
+ "loss": 6.975146770477295,
+ "step": 162
+ },
+ {
+ "epoch": 2.2655021834061135,
+ "grad_norm": 0.009936527349054813,
+ "learning_rate": 0.0006,
+ "loss": 6.961674690246582,
+ "step": 163
+ },
+ {
+ "epoch": 2.279475982532751,
+ "grad_norm": 0.015100127086043358,
+ "learning_rate": 0.0006,
+ "loss": 6.968759536743164,
+ "step": 164
+ },
+ {
+ "epoch": 2.2934497816593886,
+ "grad_norm": 0.01449542585760355,
+ "learning_rate": 0.0006,
+ "loss": 6.966510772705078,
+ "step": 165
+ },
+ {
+ "epoch": 2.3074235807860264,
+ "grad_norm": 0.01073537115007639,
+ "learning_rate": 0.0006,
+ "loss": 6.936373710632324,
+ "step": 166
+ },
+ {
+ "epoch": 2.3213973799126637,
+ "grad_norm": 0.006766272243112326,
+ "learning_rate": 0.0006,
+ "loss": 6.96397590637207,
+ "step": 167
+ },
+ {
+ "epoch": 2.3353711790393015,
+ "grad_norm": 0.005055439658463001,
+ "learning_rate": 0.0006,
+ "loss": 6.961050987243652,
+ "step": 168
+ },
+ {
+ "epoch": 2.349344978165939,
+ "grad_norm": 0.006252622697502375,
+ "learning_rate": 0.0006,
+ "loss": 6.980506420135498,
+ "step": 169
+ },
+ {
+ "epoch": 2.3633187772925766,
+ "grad_norm": 0.008084346540272236,
+ "learning_rate": 0.0006,
+ "loss": 6.9346604347229,
+ "step": 170
+ },
+ {
+ "epoch": 2.377292576419214,
+ "grad_norm": 0.010260380804538727,
+ "learning_rate": 0.0006,
+ "loss": 6.964118957519531,
+ "step": 171
+ },
+ {
+ "epoch": 2.3912663755458516,
+ "grad_norm": 0.010697875171899796,
+ "learning_rate": 0.0006,
+ "loss": 6.941572189331055,
+ "step": 172
+ },
+ {
+ "epoch": 2.405240174672489,
+ "grad_norm": 0.007338830269873142,
+ "learning_rate": 0.0006,
+ "loss": 6.950471878051758,
+ "step": 173
+ },
+ {
+ "epoch": 2.4192139737991267,
+ "grad_norm": 0.005931832827627659,
+ "learning_rate": 0.0006,
+ "loss": 6.962935924530029,
+ "step": 174
+ },
+ {
+ "epoch": 2.433187772925764,
+ "grad_norm": 0.0061243255622684956,
+ "learning_rate": 0.0006,
+ "loss": 6.946418285369873,
+ "step": 175
+ },
+ {
+ "epoch": 2.447161572052402,
+ "grad_norm": 0.007983328774571419,
+ "learning_rate": 0.0006,
+ "loss": 6.959758758544922,
+ "step": 176
+ },
+ {
+ "epoch": 2.461135371179039,
+ "grad_norm": 0.010470672510564327,
+ "learning_rate": 0.0006,
+ "loss": 6.93851375579834,
+ "step": 177
+ },
+ {
+ "epoch": 2.475109170305677,
+ "grad_norm": 0.011147388257086277,
+ "learning_rate": 0.0006,
+ "loss": 6.93258810043335,
+ "step": 178
+ },
+ {
+ "epoch": 2.489082969432314,
+ "grad_norm": 0.01058544497936964,
+ "learning_rate": 0.0006,
+ "loss": 6.932259559631348,
+ "step": 179
+ },
+ {
+ "epoch": 2.503056768558952,
+ "grad_norm": 0.016850745305418968,
+ "learning_rate": 0.0006,
+ "loss": 6.924542427062988,
+ "step": 180
+ },
+ {
+ "epoch": 2.5170305676855893,
+ "grad_norm": 0.023101547732949257,
+ "learning_rate": 0.0006,
+ "loss": 6.914878845214844,
+ "step": 181
+ },
+ {
+ "epoch": 2.531004366812227,
+ "grad_norm": 0.015795735642313957,
+ "learning_rate": 0.0006,
+ "loss": 6.922432899475098,
+ "step": 182
+ },
+ {
+ "epoch": 2.544978165938865,
+ "grad_norm": 0.008821426890790462,
+ "learning_rate": 0.0006,
+ "loss": 6.924562454223633,
+ "step": 183
+ },
+ {
+ "epoch": 2.558951965065502,
+ "grad_norm": 0.012502548284828663,
+ "learning_rate": 0.0006,
+ "loss": 6.914718151092529,
+ "step": 184
+ },
+ {
+ "epoch": 2.5729257641921395,
+ "grad_norm": 0.011545066721737385,
+ "learning_rate": 0.0006,
+ "loss": 6.915705680847168,
+ "step": 185
+ },
+ {
+ "epoch": 2.5868995633187772,
+ "grad_norm": 0.022190893068909645,
+ "learning_rate": 0.0006,
+ "loss": 6.88851261138916,
+ "step": 186
+ },
+ {
+ "epoch": 2.600873362445415,
+ "grad_norm": 0.023395851254463196,
+ "learning_rate": 0.0006,
+ "loss": 6.877655029296875,
+ "step": 187
+ },
+ {
+ "epoch": 2.6148471615720523,
+ "grad_norm": 0.015601156279444695,
+ "learning_rate": 0.0006,
+ "loss": 6.889644145965576,
+ "step": 188
+ },
+ {
+ "epoch": 2.62882096069869,
+ "grad_norm": 0.01362934336066246,
+ "learning_rate": 0.0006,
+ "loss": 6.891547203063965,
+ "step": 189
+ },
+ {
+ "epoch": 2.6427947598253274,
+ "grad_norm": 0.012305490672588348,
+ "learning_rate": 0.0006,
+ "loss": 6.890357971191406,
+ "step": 190
+ },
+ {
+ "epoch": 2.656768558951965,
+ "grad_norm": 0.01623358018696308,
+ "learning_rate": 0.0006,
+ "loss": 6.849443435668945,
+ "step": 191
+ },
+ {
+ "epoch": 2.6707423580786025,
+ "grad_norm": 0.018063291907310486,
+ "learning_rate": 0.0006,
+ "loss": 6.842372894287109,
+ "step": 192
+ },
+ {
+ "epoch": 2.6847161572052403,
+ "grad_norm": 0.010179434902966022,
+ "learning_rate": 0.0006,
+ "loss": 6.870055198669434,
+ "step": 193
+ },
+ {
+ "epoch": 2.6986899563318776,
+ "grad_norm": 0.009986722841858864,
+ "learning_rate": 0.0006,
+ "loss": 6.859061241149902,
+ "step": 194
+ },
+ {
+ "epoch": 2.7126637554585153,
+ "grad_norm": 0.01117654051631689,
+ "learning_rate": 0.0006,
+ "loss": 6.861199378967285,
+ "step": 195
+ },
+ {
+ "epoch": 2.726637554585153,
+ "grad_norm": 0.008551654405891895,
+ "learning_rate": 0.0006,
+ "loss": 6.850255012512207,
+ "step": 196
+ },
+ {
+ "epoch": 2.7406113537117904,
+ "grad_norm": 0.018904326483607292,
+ "learning_rate": 0.0006,
+ "loss": 6.848639965057373,
+ "step": 197
+ },
+ {
+ "epoch": 2.7545851528384278,
+ "grad_norm": 0.02080724574625492,
+ "learning_rate": 0.0006,
+ "loss": 6.80456018447876,
+ "step": 198
+ },
+ {
+ "epoch": 2.7685589519650655,
+ "grad_norm": 0.012010873295366764,
+ "learning_rate": 0.0006,
+ "loss": 6.848775863647461,
+ "step": 199
+ },
+ {
+ "epoch": 2.7825327510917033,
+ "grad_norm": 0.008113056421279907,
+ "learning_rate": 0.0006,
+ "loss": 6.849387168884277,
+ "step": 200
+ },
+ {
+ "epoch": 2.7965065502183406,
+ "grad_norm": 0.012912592850625515,
+ "learning_rate": 0.0006,
+ "loss": 6.832240104675293,
+ "step": 201
+ },
+ {
+ "epoch": 2.810480349344978,
+ "grad_norm": 0.01146865077316761,
+ "learning_rate": 0.0006,
+ "loss": 6.850383281707764,
+ "step": 202
+ },
+ {
+ "epoch": 2.8244541484716157,
+ "grad_norm": 0.0171892661601305,
+ "learning_rate": 0.0006,
+ "loss": 6.839577674865723,
+ "step": 203
+ },
+ {
+ "epoch": 2.8384279475982535,
+ "grad_norm": 0.012083170004189014,
+ "learning_rate": 0.0006,
+ "loss": 6.818614482879639,
+ "step": 204
+ },
+ {
+ "epoch": 2.8524017467248908,
+ "grad_norm": 0.0062555656768381596,
+ "learning_rate": 0.0006,
+ "loss": 6.831291198730469,
+ "step": 205
+ },
+ {
+ "epoch": 2.8663755458515285,
+ "grad_norm": 0.010102466680109501,
+ "learning_rate": 0.0006,
+ "loss": 6.846522808074951,
+ "step": 206
+ },
+ {
+ "epoch": 2.880349344978166,
+ "grad_norm": 0.009042087942361832,
+ "learning_rate": 0.0006,
+ "loss": 6.822072982788086,
+ "step": 207
+ },
+ {
+ "epoch": 2.8943231441048036,
+ "grad_norm": 0.011133556254208088,
+ "learning_rate": 0.0006,
+ "loss": 6.811252593994141,
+ "step": 208
+ },
+ {
+ "epoch": 2.908296943231441,
+ "grad_norm": 0.00990577507764101,
+ "learning_rate": 0.0006,
+ "loss": 6.814435958862305,
+ "step": 209
+ },
+ {
+ "epoch": 2.9222707423580787,
+ "grad_norm": 0.006605073343962431,
+ "learning_rate": 0.0006,
+ "loss": 6.831300735473633,
+ "step": 210
+ },
+ {
+ "epoch": 2.936244541484716,
+ "grad_norm": 0.007831175811588764,
+ "learning_rate": 0.0006,
+ "loss": 6.841738700866699,
+ "step": 211
+ },
+ {
+ "epoch": 2.950218340611354,
+ "grad_norm": 0.005536255426704884,
+ "learning_rate": 0.0006,
+ "loss": 6.8043670654296875,
+ "step": 212
+ },
+ {
+ "epoch": 2.964192139737991,
+ "grad_norm": 0.006532667670398951,
+ "learning_rate": 0.0006,
+ "loss": 6.8091840744018555,
+ "step": 213
+ },
+ {
+ "epoch": 2.978165938864629,
+ "grad_norm": 0.006674565840512514,
+ "learning_rate": 0.0006,
+ "loss": 6.811327934265137,
+ "step": 214
+ },
+ {
+ "epoch": 2.992139737991266,
+ "grad_norm": 0.01150344219058752,
+ "learning_rate": 0.0006,
+ "loss": 6.810762405395508,
+ "step": 215
+ },
+ {
+ "epoch": 3.0,
+ "grad_norm": 0.009015528485178947,
+ "learning_rate": 0.0006,
+ "loss": 6.836431980133057,
+ "step": 216
+ },
+ {
+ "epoch": 3.0,
+ "eval_loss": 7.718832969665527,
+ "eval_runtime": 58.6615,
+ "eval_samples_per_second": 41.629,
+ "eval_steps_per_second": 1.313,
+ "step": 216
+ },
+ {
+ "epoch": 3.0139737991266378,
+ "grad_norm": 0.009791829623281956,
+ "learning_rate": 0.0006,
+ "loss": 6.8102593421936035,
+ "step": 217
+ },
+ {
+ "epoch": 3.027947598253275,
+ "grad_norm": 0.008970947004854679,
+ "learning_rate": 0.0006,
+ "loss": 6.792663097381592,
+ "step": 218
+ },
+ {
+ "epoch": 3.041921397379913,
+ "grad_norm": 0.013689392246305943,
+ "learning_rate": 0.0006,
+ "loss": 6.815651893615723,
+ "step": 219
+ },
+ {
+ "epoch": 3.05589519650655,
+ "grad_norm": 0.031499531120061874,
+ "learning_rate": 0.0006,
+ "loss": 6.809841156005859,
+ "step": 220
+ },
+ {
+ "epoch": 3.069868995633188,
+ "grad_norm": 0.03203094005584717,
+ "learning_rate": 0.0006,
+ "loss": 6.779417037963867,
+ "step": 221
+ },
+ {
+ "epoch": 3.0838427947598253,
+ "grad_norm": 0.007947996258735657,
+ "learning_rate": 0.0006,
+ "loss": 6.773656845092773,
+ "step": 222
+ },
+ {
+ "epoch": 3.097816593886463,
+ "grad_norm": 0.03152519091963768,
+ "learning_rate": 0.0006,
+ "loss": 6.804083824157715,
+ "step": 223
+ },
+ {
+ "epoch": 3.1117903930131003,
+ "grad_norm": 0.014937258325517178,
+ "learning_rate": 0.0006,
+ "loss": 6.792213439941406,
+ "step": 224
+ },
+ {
+ "epoch": 3.125764192139738,
+ "grad_norm": 0.020961705595254898,
+ "learning_rate": 0.0006,
+ "loss": 6.8092803955078125,
+ "step": 225
+ },
+ {
+ "epoch": 3.1397379912663754,
+ "grad_norm": 0.009152988903224468,
+ "learning_rate": 0.0006,
+ "loss": 6.7737603187561035,
+ "step": 226
+ },
+ {
+ "epoch": 3.153711790393013,
+ "grad_norm": 0.02770620957016945,
+ "learning_rate": 0.0006,
+ "loss": 6.767014503479004,
+ "step": 227
+ },
+ {
+ "epoch": 3.1676855895196505,
+ "grad_norm": 0.01165871974080801,
+ "learning_rate": 0.0006,
+ "loss": 6.760884761810303,
+ "step": 228
+ },
+ {
+ "epoch": 3.1816593886462883,
+ "grad_norm": 0.01910376362502575,
+ "learning_rate": 0.0006,
+ "loss": 6.775722980499268,
+ "step": 229
+ },
+ {
+ "epoch": 3.1956331877729256,
+ "grad_norm": 0.010413747280836105,
+ "learning_rate": 0.0006,
+ "loss": 6.779816627502441,
+ "step": 230
+ },
+ {
+ "epoch": 3.2096069868995634,
+ "grad_norm": 0.01918417401611805,
+ "learning_rate": 0.0006,
+ "loss": 6.760381698608398,
+ "step": 231
+ },
+ {
+ "epoch": 3.2235807860262007,
+ "grad_norm": 0.017960334196686745,
+ "learning_rate": 0.0006,
+ "loss": 6.761198997497559,
+ "step": 232
+ },
+ {
+ "epoch": 3.2375545851528384,
+ "grad_norm": 0.016546795144677162,
+ "learning_rate": 0.0006,
+ "loss": 6.742448806762695,
+ "step": 233
+ },
+ {
+ "epoch": 3.251528384279476,
+ "grad_norm": 0.01889597810804844,
+ "learning_rate": 0.0006,
+ "loss": 6.758237838745117,
+ "step": 234
+ },
+ {
+ "epoch": 3.2655021834061135,
+ "grad_norm": 0.009611992165446281,
+ "learning_rate": 0.0006,
+ "loss": 6.75614070892334,
+ "step": 235
+ },
+ {
+ "epoch": 3.279475982532751,
+ "grad_norm": 0.02034934051334858,
+ "learning_rate": 0.0006,
+ "loss": 6.736965179443359,
+ "step": 236
+ },
+ {
+ "epoch": 3.2934497816593886,
+ "grad_norm": 0.010740065947175026,
+ "learning_rate": 0.0006,
+ "loss": 6.746947288513184,
+ "step": 237
+ },
+ {
+ "epoch": 3.3074235807860264,
+ "grad_norm": 0.018383340910077095,
+ "learning_rate": 0.0006,
+ "loss": 6.747332572937012,
+ "step": 238
+ },
+ {
+ "epoch": 3.3213973799126637,
+ "grad_norm": 0.012362911365926266,
+ "learning_rate": 0.0006,
+ "loss": 6.767669200897217,
+ "step": 239
+ },
+ {
+ "epoch": 3.3353711790393015,
+ "grad_norm": 0.01127463299781084,
+ "learning_rate": 0.0006,
+ "loss": 6.769808292388916,
+ "step": 240
+ },
+ {
+ "epoch": 3.349344978165939,
+ "grad_norm": 0.017693769186735153,
+ "learning_rate": 0.0006,
+ "loss": 6.733658790588379,
+ "step": 241
+ },
+ {
+ "epoch": 3.3633187772925766,
+ "grad_norm": 0.006420376244932413,
+ "learning_rate": 0.0006,
+ "loss": 6.74934720993042,
+ "step": 242
+ },
+ {
+ "epoch": 3.377292576419214,
+ "grad_norm": 0.014020364731550217,
+ "learning_rate": 0.0006,
+ "loss": 6.728157043457031,
+ "step": 243
+ },
+ {
+ "epoch": 3.3912663755458516,
+ "grad_norm": 0.008018582127988338,
+ "learning_rate": 0.0006,
+ "loss": 6.730726718902588,
+ "step": 244
+ },
+ {
+ "epoch": 3.405240174672489,
+ "grad_norm": 0.00782222580164671,
+ "learning_rate": 0.0006,
+ "loss": 6.7499895095825195,
+ "step": 245
+ },
+ {
+ "epoch": 3.4192139737991267,
+ "grad_norm": 0.008278409950435162,
+ "learning_rate": 0.0006,
+ "loss": 6.712803363800049,
+ "step": 246
+ },
+ {
+ "epoch": 3.433187772925764,
+ "grad_norm": 0.009826759807765484,
+ "learning_rate": 0.0006,
+ "loss": 6.742262840270996,
+ "step": 247
+ },
+ {
+ "epoch": 3.447161572052402,
+ "grad_norm": 0.009488224983215332,
+ "learning_rate": 0.0006,
+ "loss": 6.739130020141602,
+ "step": 248
+ },
+ {
+ "epoch": 3.461135371179039,
+ "grad_norm": 0.007667903788387775,
+ "learning_rate": 0.0006,
+ "loss": 6.747515678405762,
+ "step": 249
+ },
+ {
+ "epoch": 3.475109170305677,
+ "grad_norm": 0.0073090000078082085,
+ "learning_rate": 0.0006,
+ "loss": 6.740870475769043,
+ "step": 250
+ },
+ {
+ "epoch": 3.489082969432314,
+ "grad_norm": 0.010058708488941193,
+ "learning_rate": 0.0006,
+ "loss": 6.777563095092773,
+ "step": 251
+ },
+ {
+ "epoch": 3.503056768558952,
+ "grad_norm": 0.020676063373684883,
+ "learning_rate": 0.0006,
+ "loss": 6.708603858947754,
+ "step": 252
+ },
+ {
+ "epoch": 3.5170305676855893,
+ "grad_norm": 0.027210544794797897,
+ "learning_rate": 0.0006,
+ "loss": 6.694097995758057,
+ "step": 253
+ },
+ {
+ "epoch": 3.531004366812227,
+ "grad_norm": 0.038053520023822784,
+ "learning_rate": 0.0006,
+ "loss": 6.697017669677734,
+ "step": 254
+ },
+ {
+ "epoch": 3.544978165938865,
+ "grad_norm": 0.04684312269091606,
+ "learning_rate": 0.0006,
+ "loss": 6.729883193969727,
+ "step": 255
+ },
+ {
+ "epoch": 3.558951965065502,
+ "grad_norm": 0.03673060983419418,
+ "learning_rate": 0.0006,
+ "loss": 6.743753433227539,
+ "step": 256
+ },
+ {
+ "epoch": 3.5729257641921395,
+ "grad_norm": 0.006882632151246071,
+ "learning_rate": 0.0006,
+ "loss": 6.71405029296875,
+ "step": 257
+ },
+ {
+ "epoch": 3.5868995633187772,
+ "grad_norm": 0.03588201105594635,
+ "learning_rate": 0.0006,
+ "loss": 6.712467193603516,
+ "step": 258
+ },
+ {
+ "epoch": 3.600873362445415,
+ "grad_norm": 0.029415199533104897,
+ "learning_rate": 0.0006,
+ "loss": 6.715634822845459,
+ "step": 259
+ },
+ {
+ "epoch": 3.6148471615720523,
+ "grad_norm": 0.011425039730966091,
+ "learning_rate": 0.0006,
+ "loss": 6.697334289550781,
+ "step": 260
+ },
+ {
+ "epoch": 3.62882096069869,
+ "grad_norm": 0.0297263003885746,
+ "learning_rate": 0.0006,
+ "loss": 6.724848747253418,
+ "step": 261
+ },
+ {
+ "epoch": 3.6427947598253274,
+ "grad_norm": 0.007282314356416464,
+ "learning_rate": 0.0006,
+ "loss": 6.702306270599365,
+ "step": 262
+ },
+ {
+ "epoch": 3.656768558951965,
+ "grad_norm": 0.02673051692545414,
+ "learning_rate": 0.0006,
+ "loss": 6.72357177734375,
+ "step": 263
+ },
+ {
+ "epoch": 3.6707423580786025,
+ "grad_norm": 0.016174081712961197,
+ "learning_rate": 0.0006,
+ "loss": 6.730698585510254,
+ "step": 264
+ },
+ {
+ "epoch": 3.6847161572052403,
+ "grad_norm": 0.0182416420429945,
+ "learning_rate": 0.0006,
+ "loss": 6.717386245727539,
+ "step": 265
+ },
+ {
+ "epoch": 3.6986899563318776,
+ "grad_norm": 0.023250866681337357,
+ "learning_rate": 0.0006,
+ "loss": 6.697832107543945,
+ "step": 266
+ },
+ {
+ "epoch": 3.7126637554585153,
+ "grad_norm": 0.007539753336459398,
+ "learning_rate": 0.0006,
+ "loss": 6.714348793029785,
+ "step": 267
+ },
+ {
+ "epoch": 3.726637554585153,
+ "grad_norm": 0.030086630955338478,
+ "learning_rate": 0.0006,
+ "loss": 6.724481582641602,
+ "step": 268
+ },
+ {
+ "epoch": 3.7406113537117904,
+ "grad_norm": 0.012784913182258606,
+ "learning_rate": 0.0006,
+ "loss": 6.703491687774658,
+ "step": 269
+ },
+ {
+ "epoch": 3.7545851528384278,
+ "grad_norm": 0.02042769268155098,
+ "learning_rate": 0.0006,
+ "loss": 6.721139907836914,
+ "step": 270
+ },
+ {
+ "epoch": 3.7685589519650655,
+ "grad_norm": 0.023557329550385475,
+ "learning_rate": 0.0006,
+ "loss": 6.728122711181641,
+ "step": 271
+ },
+ {
+ "epoch": 3.7825327510917033,
+ "grad_norm": 0.007716621737927198,
+ "learning_rate": 0.0006,
+ "loss": 6.70085334777832,
+ "step": 272
+ },
+ {
+ "epoch": 3.7965065502183406,
+ "grad_norm": 0.022772278636693954,
+ "learning_rate": 0.0006,
+ "loss": 6.712882995605469,
+ "step": 273
+ },
+ {
+ "epoch": 3.810480349344978,
+ "grad_norm": 0.007928933948278427,
+ "learning_rate": 0.0006,
+ "loss": 6.698320388793945,
+ "step": 274
+ },
+ {
+ "epoch": 3.8244541484716157,
+ "grad_norm": 0.022355245426297188,
+ "learning_rate": 0.0006,
+ "loss": 6.687088489532471,
+ "step": 275
+ },
+ {
+ "epoch": 3.8384279475982535,
+ "grad_norm": 0.02037850394845009,
+ "learning_rate": 0.0006,
+ "loss": 6.676856994628906,
+ "step": 276
+ },
+ {
+ "epoch": 3.8524017467248908,
+ "grad_norm": 0.012290362268686295,
+ "learning_rate": 0.0006,
+ "loss": 6.682883262634277,
+ "step": 277
+ },
+ {
+ "epoch": 3.8663755458515285,
+ "grad_norm": 0.022331275045871735,
+ "learning_rate": 0.0006,
+ "loss": 6.694752216339111,
+ "step": 278
+ },
+ {
+ "epoch": 3.880349344978166,
+ "grad_norm": 0.007527181413024664,
+ "learning_rate": 0.0006,
+ "loss": 6.686394691467285,
+ "step": 279
+ },
+ {
+ "epoch": 3.8943231441048036,
+ "grad_norm": 0.02159409038722515,
+ "learning_rate": 0.0006,
+ "loss": 6.697943687438965,
+ "step": 280
+ },
+ {
+ "epoch": 3.908296943231441,
+ "grad_norm": 0.015054763294756413,
+ "learning_rate": 0.0006,
+ "loss": 6.677796363830566,
+ "step": 281
+ },
+ {
+ "epoch": 3.9222707423580787,
+ "grad_norm": 0.01466019731014967,
+ "learning_rate": 0.0006,
+ "loss": 6.697715759277344,
+ "step": 282
+ },
+ {
+ "epoch": 3.936244541484716,
+ "grad_norm": 0.01875465176999569,
+ "learning_rate": 0.0006,
+ "loss": 6.700228691101074,
+ "step": 283
+ },
+ {
+ "epoch": 3.950218340611354,
+ "grad_norm": 0.00778991961851716,
+ "learning_rate": 0.0006,
+ "loss": 6.69331693649292,
+ "step": 284
+ },
+ {
+ "epoch": 3.964192139737991,
+ "grad_norm": 0.020182441920042038,
+ "learning_rate": 0.0006,
+ "loss": 6.672651290893555,
+ "step": 285
+ },
+ {
+ "epoch": 3.978165938864629,
+ "grad_norm": 0.01240516733378172,
+ "learning_rate": 0.0006,
+ "loss": 6.688860893249512,
+ "step": 286
+ },
+ {
+ "epoch": 3.992139737991266,
+ "grad_norm": 0.011218412779271603,
+ "learning_rate": 0.0006,
+ "loss": 6.669534683227539,
+ "step": 287
+ },
+ {
+ "epoch": 4.0,
+ "grad_norm": 0.0141852255910635,
+ "learning_rate": 0.0006,
+ "loss": 6.690206527709961,
+ "step": 288
+ },
+ {
+ "epoch": 4.0,
+ "eval_loss": 7.537620544433594,
+ "eval_runtime": 58.4398,
+ "eval_samples_per_second": 41.787,
+ "eval_steps_per_second": 1.318,
+ "step": 288
+ },
+ {
+ "epoch": 4.013973799126638,
+ "grad_norm": 0.008106146939098835,
+ "learning_rate": 0.0006,
+ "loss": 6.688591003417969,
+ "step": 289
+ },
+ {
+ "epoch": 4.0279475982532755,
+ "grad_norm": 0.014588244259357452,
+ "learning_rate": 0.0006,
+ "loss": 6.684961318969727,
+ "step": 290
+ },
+ {
+ "epoch": 4.041921397379912,
+ "grad_norm": 0.011985368095338345,
+ "learning_rate": 0.0006,
+ "loss": 6.672794342041016,
+ "step": 291
+ },
+ {
+ "epoch": 4.05589519650655,
+ "grad_norm": 0.014632102102041245,
+ "learning_rate": 0.0006,
+ "loss": 6.702736854553223,
+ "step": 292
+ },
+ {
+ "epoch": 4.069868995633188,
+ "grad_norm": 0.011243713088333607,
+ "learning_rate": 0.0006,
+ "loss": 6.658792495727539,
+ "step": 293
+ },
+ {
+ "epoch": 4.083842794759826,
+ "grad_norm": 0.010519202798604965,
+ "learning_rate": 0.0006,
+ "loss": 6.662495136260986,
+ "step": 294
+ },
+ {
+ "epoch": 4.097816593886463,
+ "grad_norm": 0.007318088319152594,
+ "learning_rate": 0.0006,
+ "loss": 6.675954818725586,
+ "step": 295
+ },
+ {
+ "epoch": 4.1117903930131,
+ "grad_norm": 0.011712398380041122,
+ "learning_rate": 0.0006,
+ "loss": 6.673973083496094,
+ "step": 296
+ },
+ {
+ "epoch": 4.125764192139738,
+ "grad_norm": 0.006726635154336691,
+ "learning_rate": 0.0006,
+ "loss": 6.656833171844482,
+ "step": 297
+ },
+ {
+ "epoch": 4.139737991266376,
+ "grad_norm": 0.005466983187943697,
+ "learning_rate": 0.0006,
+ "loss": 6.675468921661377,
+ "step": 298
+ },
+ {
+ "epoch": 4.153711790393013,
+ "grad_norm": 0.007163621485233307,
+ "learning_rate": 0.0006,
+ "loss": 6.6862874031066895,
+ "step": 299
+ },
+ {
+ "epoch": 4.1676855895196505,
+ "grad_norm": 0.007626629434525967,
+ "learning_rate": 0.0006,
+ "loss": 6.654741287231445,
+ "step": 300
+ },
+ {
+ "epoch": 4.181659388646288,
+ "grad_norm": 0.0066330647096037865,
+ "learning_rate": 0.0006,
+ "loss": 6.68114709854126,
+ "step": 301
+ },
+ {
+ "epoch": 4.195633187772926,
+ "grad_norm": 0.007363482844084501,
+ "learning_rate": 0.0006,
+ "loss": 6.683980941772461,
+ "step": 302
+ },
+ {
+ "epoch": 4.209606986899563,
+ "grad_norm": 0.00971233006566763,
+ "learning_rate": 0.0006,
+ "loss": 6.669700622558594,
+ "step": 303
+ },
+ {
+ "epoch": 4.223580786026201,
+ "grad_norm": 0.013185256160795689,
+ "learning_rate": 0.0006,
+ "loss": 6.686490535736084,
+ "step": 304
+ },
+ {
+ "epoch": 4.2375545851528384,
+ "grad_norm": 0.019116027280688286,
+ "learning_rate": 0.0006,
+ "loss": 6.645479202270508,
+ "step": 305
+ },
+ {
+ "epoch": 4.251528384279476,
+ "grad_norm": 0.0328814722597599,
+ "learning_rate": 0.0006,
+ "loss": 6.68080997467041,
+ "step": 306
+ },
+ {
+ "epoch": 4.265502183406113,
+ "grad_norm": 0.04482287913560867,
+ "learning_rate": 0.0006,
+ "loss": 6.698016166687012,
+ "step": 307
+ },
+ {
+ "epoch": 4.279475982532751,
+ "grad_norm": 0.03053313121199608,
+ "learning_rate": 0.0006,
+ "loss": 6.642240524291992,
+ "step": 308
+ },
+ {
+ "epoch": 4.293449781659389,
+ "grad_norm": 0.043351784348487854,
+ "learning_rate": 0.0006,
+ "loss": 6.662343978881836,
+ "step": 309
+ },
+ {
+ "epoch": 4.307423580786026,
+ "grad_norm": 0.03165886923670769,
+ "learning_rate": 0.0006,
+ "loss": 6.6291961669921875,
+ "step": 310
+ },
+ {
+ "epoch": 4.321397379912664,
+ "grad_norm": 0.028601787984371185,
+ "learning_rate": 0.0006,
+ "loss": 6.667544364929199,
+ "step": 311
+ },
+ {
+ "epoch": 4.335371179039301,
+ "grad_norm": 0.019558480009436607,
+ "learning_rate": 0.0006,
+ "loss": 6.641025543212891,
+ "step": 312
+ },
+ {
+ "epoch": 4.349344978165939,
+ "grad_norm": 0.039851363748311996,
+ "learning_rate": 0.0006,
+ "loss": 6.665952205657959,
+ "step": 313
+ },
+ {
+ "epoch": 4.3633187772925766,
+ "grad_norm": 0.018417146056890488,
+ "learning_rate": 0.0006,
+ "loss": 6.657257556915283,
+ "step": 314
+ },
+ {
+ "epoch": 4.377292576419214,
+ "grad_norm": 0.02894527092576027,
+ "learning_rate": 0.0006,
+ "loss": 6.656288146972656,
+ "step": 315
+ },
+ {
+ "epoch": 4.391266375545851,
+ "grad_norm": 0.01910966821014881,
+ "learning_rate": 0.0006,
+ "loss": 6.688958168029785,
+ "step": 316
+ },
+ {
+ "epoch": 4.405240174672489,
+ "grad_norm": 0.01668412610888481,
+ "learning_rate": 0.0006,
+ "loss": 6.64876127243042,
+ "step": 317
+ },
+ {
+ "epoch": 4.419213973799127,
+ "grad_norm": 0.02935275249183178,
+ "learning_rate": 0.0006,
+ "loss": 6.626096725463867,
+ "step": 318
+ },
+ {
+ "epoch": 4.4331877729257645,
+ "grad_norm": 0.017997972667217255,
+ "learning_rate": 0.0006,
+ "loss": 6.640386581420898,
+ "step": 319
+ },
+ {
+ "epoch": 4.447161572052401,
+ "grad_norm": 0.01779189519584179,
+ "learning_rate": 0.0006,
+ "loss": 6.641165256500244,
+ "step": 320
+ },
+ {
+ "epoch": 4.461135371179039,
+ "grad_norm": 0.013901712372899055,
+ "learning_rate": 0.0006,
+ "loss": 6.651627063751221,
+ "step": 321
+ },
+ {
+ "epoch": 4.475109170305677,
+ "grad_norm": 0.013211199082434177,
+ "learning_rate": 0.0006,
+ "loss": 6.630064964294434,
+ "step": 322
+ },
+ {
+ "epoch": 4.489082969432315,
+ "grad_norm": 0.02324107475578785,
+ "learning_rate": 0.0006,
+ "loss": 6.670443058013916,
+ "step": 323
+ },
+ {
+ "epoch": 4.503056768558952,
+ "grad_norm": 0.019452504813671112,
+ "learning_rate": 0.0006,
+ "loss": 6.651266098022461,
+ "step": 324
+ },
+ {
+ "epoch": 4.517030567685589,
+ "grad_norm": 0.017943572252988815,
+ "learning_rate": 0.0006,
+ "loss": 6.657943248748779,
+ "step": 325
+ },
+ {
+ "epoch": 4.531004366812227,
+ "grad_norm": 0.012136237695813179,
+ "learning_rate": 0.0006,
+ "loss": 6.640082359313965,
+ "step": 326
+ },
+ {
+ "epoch": 4.544978165938865,
+ "grad_norm": 0.011544011533260345,
+ "learning_rate": 0.0006,
+ "loss": 6.6186933517456055,
+ "step": 327
+ },
+ {
+ "epoch": 4.558951965065502,
+ "grad_norm": 0.01569426618516445,
+ "learning_rate": 0.0006,
+ "loss": 6.640767574310303,
+ "step": 328
+ },
+ {
+ "epoch": 4.5729257641921395,
+ "grad_norm": 0.027668394148349762,
+ "learning_rate": 0.0006,
+ "loss": 6.651412487030029,
+ "step": 329
+ },
+ {
+ "epoch": 4.586899563318777,
+ "grad_norm": 0.04073230177164078,
+ "learning_rate": 0.0006,
+ "loss": 6.636565208435059,
+ "step": 330
+ },
+ {
+ "epoch": 4.600873362445415,
+ "grad_norm": 0.04942672327160835,
+ "learning_rate": 0.0006,
+ "loss": 6.667994499206543,
+ "step": 331
+ },
+ {
+ "epoch": 4.614847161572053,
+ "grad_norm": 0.037568598985672,
+ "learning_rate": 0.0006,
+ "loss": 6.632008075714111,
+ "step": 332
+ },
+ {
+ "epoch": 4.62882096069869,
+ "grad_norm": 0.009880400262773037,
+ "learning_rate": 0.0006,
+ "loss": 6.6504058837890625,
+ "step": 333
+ },
+ {
+ "epoch": 4.642794759825327,
+ "grad_norm": 0.033384259790182114,
+ "learning_rate": 0.0006,
+ "loss": 6.63852596282959,
+ "step": 334
+ },
+ {
+ "epoch": 4.656768558951965,
+ "grad_norm": 0.035640884190797806,
+ "learning_rate": 0.0006,
+ "loss": 6.627986431121826,
+ "step": 335
+ },
+ {
+ "epoch": 4.670742358078603,
+ "grad_norm": 0.013637864030897617,
+ "learning_rate": 0.0006,
+ "loss": 6.624503135681152,
+ "step": 336
+ },
+ {
+ "epoch": 4.68471615720524,
+ "grad_norm": 0.025304675102233887,
+ "learning_rate": 0.0006,
+ "loss": 6.612077236175537,
+ "step": 337
+ },
+ {
+ "epoch": 4.698689956331878,
+ "grad_norm": 0.03377160802483559,
+ "learning_rate": 0.0006,
+ "loss": 6.608938217163086,
+ "step": 338
+ },
+ {
+ "epoch": 4.712663755458515,
+ "grad_norm": 0.019803153350949287,
+ "learning_rate": 0.0006,
+ "loss": 6.636343955993652,
+ "step": 339
+ },
+ {
+ "epoch": 4.726637554585153,
+ "grad_norm": 0.018928129225969315,
+ "learning_rate": 0.0006,
+ "loss": 6.636613845825195,
+ "step": 340
+ },
+ {
+ "epoch": 4.74061135371179,
+ "grad_norm": 0.023068996146321297,
+ "learning_rate": 0.0006,
+ "loss": 6.6258440017700195,
+ "step": 341
+ },
+ {
+ "epoch": 4.754585152838428,
+ "grad_norm": 0.016849957406520844,
+ "learning_rate": 0.0006,
+ "loss": 6.638483047485352,
+ "step": 342
+ },
+ {
+ "epoch": 4.7685589519650655,
+ "grad_norm": 0.021430544555187225,
+ "learning_rate": 0.0006,
+ "loss": 6.619649410247803,
+ "step": 343
+ },
+ {
+ "epoch": 4.782532751091703,
+ "grad_norm": 0.024053368717432022,
+ "learning_rate": 0.0006,
+ "loss": 6.639214515686035,
+ "step": 344
+ },
+ {
+ "epoch": 4.796506550218341,
+ "grad_norm": 0.011926693841814995,
+ "learning_rate": 0.0006,
+ "loss": 6.604836463928223,
+ "step": 345
+ },
+ {
+ "epoch": 4.810480349344978,
+ "grad_norm": 0.01961611397564411,
+ "learning_rate": 0.0006,
+ "loss": 6.624864101409912,
+ "step": 346
+ },
+ {
+ "epoch": 4.824454148471616,
+ "grad_norm": 0.027946067973971367,
+ "learning_rate": 0.0006,
+ "loss": 6.621284484863281,
+ "step": 347
+ },
+ {
+ "epoch": 4.8384279475982535,
+ "grad_norm": 0.02108592912554741,
+ "learning_rate": 0.0006,
+ "loss": 6.5982465744018555,
+ "step": 348
+ },
+ {
+ "epoch": 4.85240174672489,
+ "grad_norm": 0.017272260040044785,
+ "learning_rate": 0.0006,
+ "loss": 6.626142501831055,
+ "step": 349
+ },
+ {
+ "epoch": 4.866375545851528,
+ "grad_norm": 0.015654582530260086,
+ "learning_rate": 0.0006,
+ "loss": 6.621735572814941,
+ "step": 350
+ },
+ {
+ "epoch": 4.880349344978166,
+ "grad_norm": 0.015797540545463562,
+ "learning_rate": 0.0006,
+ "loss": 6.6016411781311035,
+ "step": 351
+ },
+ {
+ "epoch": 4.894323144104804,
+ "grad_norm": 0.028488904237747192,
+ "learning_rate": 0.0006,
+ "loss": 6.598413467407227,
+ "step": 352
+ },
+ {
+ "epoch": 4.908296943231441,
+ "grad_norm": 0.03411734476685524,
+ "learning_rate": 0.0006,
+ "loss": 6.597321033477783,
+ "step": 353
+ },
+ {
+ "epoch": 4.922270742358078,
+ "grad_norm": 0.027188263833522797,
+ "learning_rate": 0.0006,
+ "loss": 6.630616188049316,
+ "step": 354
+ },
+ {
+ "epoch": 4.936244541484716,
+ "grad_norm": 0.03411437198519707,
+ "learning_rate": 0.0006,
+ "loss": 6.57647705078125,
+ "step": 355
+ },
+ {
+ "epoch": 4.950218340611354,
+ "grad_norm": 0.019907327368855476,
+ "learning_rate": 0.0006,
+ "loss": 6.579545974731445,
+ "step": 356
+ },
+ {
+ "epoch": 4.964192139737992,
+ "grad_norm": 0.02485930174589157,
+ "learning_rate": 0.0006,
+ "loss": 6.591073036193848,
+ "step": 357
+ },
+ {
+ "epoch": 4.978165938864628,
+ "grad_norm": 0.026106052100658417,
+ "learning_rate": 0.0006,
+ "loss": 6.580729961395264,
+ "step": 358
+ },
+ {
+ "epoch": 4.992139737991266,
+ "grad_norm": 0.029401391744613647,
+ "learning_rate": 0.0006,
+ "loss": 6.5460076332092285,
+ "step": 359
+ },
+ {
+ "epoch": 5.0,
+ "grad_norm": 0.032869379967451096,
+ "learning_rate": 0.0006,
+ "loss": 6.583278179168701,
+ "step": 360
+ },
+ {
+ "epoch": 5.0,
+ "eval_loss": 7.416288375854492,
+ "eval_runtime": 60.4177,
+ "eval_samples_per_second": 40.419,
+ "eval_steps_per_second": 1.274,
+ "step": 360
+ },
+ {
+ "epoch": 5.013973799126638,
+ "grad_norm": 0.04862433671951294,
+ "learning_rate": 0.0006,
+ "loss": 6.58473014831543,
+ "step": 361
+ },
+ {
+ "epoch": 5.0279475982532755,
+ "grad_norm": 0.06636223942041397,
+ "learning_rate": 0.0006,
+ "loss": 6.621153354644775,
+ "step": 362
+ },
+ {
+ "epoch": 5.041921397379912,
+ "grad_norm": 0.05888264626264572,
+ "learning_rate": 0.0006,
+ "loss": 6.621196269989014,
+ "step": 363
+ },
+ {
+ "epoch": 5.05589519650655,
+ "grad_norm": 0.05331697314977646,
+ "learning_rate": 0.0006,
+ "loss": 6.600763320922852,
+ "step": 364
+ },
+ {
+ "epoch": 5.069868995633188,
+ "grad_norm": 0.03947671875357628,
+ "learning_rate": 0.0006,
+ "loss": 6.600671768188477,
+ "step": 365
+ },
+ {
+ "epoch": 5.083842794759826,
+ "grad_norm": 0.04045343026518822,
+ "learning_rate": 0.0006,
+ "loss": 6.63529109954834,
+ "step": 366
+ },
+ {
+ "epoch": 5.097816593886463,
+ "grad_norm": 0.029766036197543144,
+ "learning_rate": 0.0006,
+ "loss": 6.612868309020996,
+ "step": 367
+ },
+ {
+ "epoch": 5.1117903930131,
+ "grad_norm": 0.03416327014565468,
+ "learning_rate": 0.0006,
+ "loss": 6.582240581512451,
+ "step": 368
+ },
+ {
+ "epoch": 5.125764192139738,
+ "grad_norm": 0.03458515554666519,
+ "learning_rate": 0.0006,
+ "loss": 6.604381561279297,
+ "step": 369
+ },
+ {
+ "epoch": 5.139737991266376,
+ "grad_norm": 0.024099089205265045,
+ "learning_rate": 0.0006,
+ "loss": 6.597355365753174,
+ "step": 370
+ },
+ {
+ "epoch": 5.153711790393013,
+ "grad_norm": 0.02694438211619854,
+ "learning_rate": 0.0006,
+ "loss": 6.597512245178223,
+ "step": 371
+ },
+ {
+ "epoch": 5.1676855895196505,
+ "grad_norm": 0.024201856926083565,
+ "learning_rate": 0.0006,
+ "loss": 6.564685821533203,
+ "step": 372
+ },
+ {
+ "epoch": 5.181659388646288,
+ "grad_norm": 0.015959380194544792,
+ "learning_rate": 0.0006,
+ "loss": 6.582263946533203,
+ "step": 373
+ },
+ {
+ "epoch": 5.195633187772926,
+ "grad_norm": 0.025690171867609024,
+ "learning_rate": 0.0006,
+ "loss": 6.601057052612305,
+ "step": 374
+ },
+ {
+ "epoch": 5.209606986899563,
+ "grad_norm": 0.014497052878141403,
+ "learning_rate": 0.0006,
+ "loss": 6.584796905517578,
+ "step": 375
+ },
+ {
+ "epoch": 5.223580786026201,
+ "grad_norm": 0.019870297983288765,
+ "learning_rate": 0.0006,
+ "loss": 6.574010372161865,
+ "step": 376
+ },
+ {
+ "epoch": 5.2375545851528384,
+ "grad_norm": 0.01807512901723385,
+ "learning_rate": 0.0006,
+ "loss": 6.558596134185791,
+ "step": 377
+ },
+ {
+ "epoch": 5.251528384279476,
+ "grad_norm": 0.014525193721055984,
+ "learning_rate": 0.0006,
+ "loss": 6.571803092956543,
+ "step": 378
+ },
+ {
+ "epoch": 5.265502183406113,
+ "grad_norm": 0.016113435849547386,
+ "learning_rate": 0.0006,
+ "loss": 6.568970680236816,
+ "step": 379
+ },
+ {
+ "epoch": 5.279475982532751,
+ "grad_norm": 0.020724017173051834,
+ "learning_rate": 0.0006,
+ "loss": 6.5669403076171875,
+ "step": 380
+ },
+ {
+ "epoch": 5.293449781659389,
+ "grad_norm": 0.013963675126433372,
+ "learning_rate": 0.0006,
+ "loss": 6.586382865905762,
+ "step": 381
+ },
+ {
+ "epoch": 5.307423580786026,
+ "grad_norm": 0.01621115580201149,
+ "learning_rate": 0.0006,
+ "loss": 6.539546966552734,
+ "step": 382
+ },
+ {
+ "epoch": 5.321397379912664,
+ "grad_norm": 0.014925207942724228,
+ "learning_rate": 0.0006,
+ "loss": 6.531795501708984,
+ "step": 383
+ },
+ {
+ "epoch": 5.335371179039301,
+ "grad_norm": 0.015454350039362907,
+ "learning_rate": 0.0006,
+ "loss": 6.575699806213379,
+ "step": 384
+ },
+ {
+ "epoch": 5.349344978165939,
+ "grad_norm": 0.01623542234301567,
+ "learning_rate": 0.0006,
+ "loss": 6.582373142242432,
+ "step": 385
+ },
+ {
+ "epoch": 5.3633187772925766,
+ "grad_norm": 0.015452183783054352,
+ "learning_rate": 0.0006,
+ "loss": 6.576451301574707,
+ "step": 386
+ },
+ {
+ "epoch": 5.377292576419214,
+ "grad_norm": 0.021102432161569595,
+ "learning_rate": 0.0006,
+ "loss": 6.543055534362793,
+ "step": 387
+ },
+ {
+ "epoch": 5.391266375545851,
+ "grad_norm": 0.009827060624957085,
+ "learning_rate": 0.0006,
+ "loss": 6.526939392089844,
+ "step": 388
+ },
+ {
+ "epoch": 5.405240174672489,
+ "grad_norm": 0.013735136948525906,
+ "learning_rate": 0.0006,
+ "loss": 6.538825988769531,
+ "step": 389
+ },
+ {
+ "epoch": 5.419213973799127,
+ "grad_norm": 0.02073509432375431,
+ "learning_rate": 0.0006,
+ "loss": 6.521315574645996,
+ "step": 390
+ },
+ {
+ "epoch": 5.4331877729257645,
+ "grad_norm": 0.03110959194600582,
+ "learning_rate": 0.0006,
+ "loss": 6.561807632446289,
+ "step": 391
+ },
+ {
+ "epoch": 5.447161572052401,
+ "grad_norm": 0.045058365911245346,
+ "learning_rate": 0.0006,
+ "loss": 6.526515960693359,
+ "step": 392
+ },
+ {
+ "epoch": 5.461135371179039,
+ "grad_norm": 0.06359834969043732,
+ "learning_rate": 0.0006,
+ "loss": 6.554903030395508,
+ "step": 393
+ },
+ {
+ "epoch": 5.475109170305677,
+ "grad_norm": 0.06406767666339874,
+ "learning_rate": 0.0006,
+ "loss": 6.576272010803223,
+ "step": 394
+ },
+ {
+ "epoch": 5.489082969432315,
+ "grad_norm": 0.03403699770569801,
+ "learning_rate": 0.0006,
+ "loss": 6.5306243896484375,
+ "step": 395
+ },
+ {
+ "epoch": 5.503056768558952,
+ "grad_norm": 0.036109618842601776,
+ "learning_rate": 0.0006,
+ "loss": 6.54176139831543,
+ "step": 396
+ },
+ {
+ "epoch": 5.517030567685589,
+ "grad_norm": 0.03478201478719711,
+ "learning_rate": 0.0006,
+ "loss": 6.530734062194824,
+ "step": 397
+ },
+ {
+ "epoch": 5.531004366812227,
+ "grad_norm": 0.03401544317603111,
+ "learning_rate": 0.0006,
+ "loss": 6.557794094085693,
+ "step": 398
+ },
+ {
+ "epoch": 5.544978165938865,
+ "grad_norm": 0.031629662960767746,
+ "learning_rate": 0.0006,
+ "loss": 6.564887046813965,
+ "step": 399
+ },
+ {
+ "epoch": 5.558951965065502,
+ "grad_norm": 0.02305024117231369,
+ "learning_rate": 0.0006,
+ "loss": 6.547412872314453,
+ "step": 400
+ },
+ {
+ "epoch": 5.5729257641921395,
+ "grad_norm": 0.030852800235152245,
+ "learning_rate": 0.0006,
+ "loss": 6.495370388031006,
+ "step": 401
+ },
+ {
+ "epoch": 5.586899563318777,
+ "grad_norm": 0.03368550166487694,
+ "learning_rate": 0.0006,
+ "loss": 6.495980262756348,
+ "step": 402
+ },
+ {
+ "epoch": 5.600873362445415,
+ "grad_norm": 0.020976359024643898,
+ "learning_rate": 0.0006,
+ "loss": 6.509410858154297,
+ "step": 403
+ },
+ {
+ "epoch": 5.614847161572053,
+ "grad_norm": 0.019652143120765686,
+ "learning_rate": 0.0006,
+ "loss": 6.493232727050781,
+ "step": 404
+ },
+ {
+ "epoch": 5.62882096069869,
+ "grad_norm": 0.024320121854543686,
+ "learning_rate": 0.0006,
+ "loss": 6.521531581878662,
+ "step": 405
+ },
+ {
+ "epoch": 5.642794759825327,
+ "grad_norm": 0.016611427068710327,
+ "learning_rate": 0.0006,
+ "loss": 6.507271766662598,
+ "step": 406
+ },
+ {
+ "epoch": 5.656768558951965,
+ "grad_norm": 0.01635856367647648,
+ "learning_rate": 0.0006,
+ "loss": 6.492038726806641,
+ "step": 407
+ },
+ {
+ "epoch": 5.670742358078603,
+ "grad_norm": 0.017642095685005188,
+ "learning_rate": 0.0006,
+ "loss": 6.500349998474121,
+ "step": 408
+ },
+ {
+ "epoch": 5.68471615720524,
+ "grad_norm": 0.01679178513586521,
+ "learning_rate": 0.0006,
+ "loss": 6.500361442565918,
+ "step": 409
+ },
+ {
+ "epoch": 5.698689956331878,
+ "grad_norm": 0.011776307597756386,
+ "learning_rate": 0.0006,
+ "loss": 6.482490539550781,
+ "step": 410
+ },
+ {
+ "epoch": 5.712663755458515,
+ "grad_norm": 0.0176856629550457,
+ "learning_rate": 0.0006,
+ "loss": 6.48488712310791,
+ "step": 411
+ },
+ {
+ "epoch": 5.726637554585153,
+ "grad_norm": 0.015193155966699123,
+ "learning_rate": 0.0006,
+ "loss": 6.459545612335205,
+ "step": 412
+ },
+ {
+ "epoch": 5.74061135371179,
+ "grad_norm": 0.018623212352395058,
+ "learning_rate": 0.0006,
+ "loss": 6.478870391845703,
+ "step": 413
+ },
+ {
+ "epoch": 5.754585152838428,
+ "grad_norm": 0.012841006740927696,
+ "learning_rate": 0.0006,
+ "loss": 6.474418640136719,
+ "step": 414
+ },
+ {
+ "epoch": 5.7685589519650655,
+ "grad_norm": 0.015540511347353458,
+ "learning_rate": 0.0006,
+ "loss": 6.4706525802612305,
+ "step": 415
+ },
+ {
+ "epoch": 5.782532751091703,
+ "grad_norm": 0.022695740684866905,
+ "learning_rate": 0.0006,
+ "loss": 6.455738067626953,
+ "step": 416
+ },
+ {
+ "epoch": 5.796506550218341,
+ "grad_norm": 0.026930129155516624,
+ "learning_rate": 0.0006,
+ "loss": 6.458016872406006,
+ "step": 417
+ },
+ {
+ "epoch": 5.810480349344978,
+ "grad_norm": 0.03555797412991524,
+ "learning_rate": 0.0006,
+ "loss": 6.480044364929199,
+ "step": 418
+ },
+ {
+ "epoch": 5.824454148471616,
+ "grad_norm": 0.05024682730436325,
+ "learning_rate": 0.0006,
+ "loss": 6.454564571380615,
+ "step": 419
+ },
+ {
+ "epoch": 5.8384279475982535,
+ "grad_norm": 0.06616255640983582,
+ "learning_rate": 0.0006,
+ "loss": 6.433927536010742,
+ "step": 420
+ },
+ {
+ "epoch": 5.85240174672489,
+ "grad_norm": 0.059065770357847214,
+ "learning_rate": 0.0006,
+ "loss": 6.508362770080566,
+ "step": 421
+ },
+ {
+ "epoch": 5.866375545851528,
+ "grad_norm": 0.026689983904361725,
+ "learning_rate": 0.0006,
+ "loss": 6.464297294616699,
+ "step": 422
+ },
+ {
+ "epoch": 5.880349344978166,
+ "grad_norm": 0.04616398364305496,
+ "learning_rate": 0.0006,
+ "loss": 6.45448112487793,
+ "step": 423
+ },
+ {
+ "epoch": 5.894323144104804,
+ "grad_norm": 0.04980356618762016,
+ "learning_rate": 0.0006,
+ "loss": 6.4696574211120605,
+ "step": 424
+ },
+ {
+ "epoch": 5.908296943231441,
+ "grad_norm": 0.0551983006298542,
+ "learning_rate": 0.0006,
+ "loss": 6.419000625610352,
+ "step": 425
+ },
+ {
+ "epoch": 5.922270742358078,
+ "grad_norm": 0.06988243758678436,
+ "learning_rate": 0.0006,
+ "loss": 6.442890167236328,
+ "step": 426
+ },
+ {
+ "epoch": 5.936244541484716,
+ "grad_norm": 0.06040226295590401,
+ "learning_rate": 0.0006,
+ "loss": 6.513258934020996,
+ "step": 427
+ },
+ {
+ "epoch": 5.950218340611354,
+ "grad_norm": 0.04635024070739746,
+ "learning_rate": 0.0006,
+ "loss": 6.4710283279418945,
+ "step": 428
+ },
+ {
+ "epoch": 5.964192139737992,
+ "grad_norm": 0.04762702435255051,
+ "learning_rate": 0.0006,
+ "loss": 6.469327926635742,
+ "step": 429
+ },
+ {
+ "epoch": 5.978165938864628,
+ "grad_norm": 0.04857536032795906,
+ "learning_rate": 0.0006,
+ "loss": 6.426105499267578,
+ "step": 430
+ },
+ {
+ "epoch": 5.992139737991266,
+ "grad_norm": 0.04412693530321121,
+ "learning_rate": 0.0006,
+ "loss": 6.43179988861084,
+ "step": 431
+ },
+ {
+ "epoch": 6.0,
+ "grad_norm": 0.026129672303795815,
+ "learning_rate": 0.0006,
+ "loss": 6.437918663024902,
+ "step": 432
+ },
+ {
+ "epoch": 6.0,
+ "eval_loss": 7.219249248504639,
+ "eval_runtime": 58.6837,
+ "eval_samples_per_second": 41.613,
+ "eval_steps_per_second": 1.312,
+ "step": 432
+ },
+ {
+ "epoch": 6.013973799126638,
+ "grad_norm": 0.03329874947667122,
+ "learning_rate": 0.0006,
+ "loss": 6.438179969787598,
+ "step": 433
+ },
+ {
+ "epoch": 6.0279475982532755,
+ "grad_norm": 0.03723187372088432,
+ "learning_rate": 0.0006,
+ "loss": 6.394242286682129,
+ "step": 434
+ },
+ {
+ "epoch": 6.041921397379912,
+ "grad_norm": 0.014625726267695427,
+ "learning_rate": 0.0006,
+ "loss": 6.412118911743164,
+ "step": 435
+ },
+ {
+ "epoch": 6.05589519650655,
+ "grad_norm": 0.02917475439608097,
+ "learning_rate": 0.0006,
+ "loss": 6.453164577484131,
+ "step": 436
+ },
+ {
+ "epoch": 6.069868995633188,
+ "grad_norm": 0.01811208389699459,
+ "learning_rate": 0.0006,
+ "loss": 6.396003246307373,
+ "step": 437
+ },
+ {
+ "epoch": 6.083842794759826,
+ "grad_norm": 0.027235351502895355,
+ "learning_rate": 0.0006,
+ "loss": 6.402976036071777,
+ "step": 438
+ },
+ {
+ "epoch": 6.097816593886463,
+ "grad_norm": 0.02835794910788536,
+ "learning_rate": 0.0006,
+ "loss": 6.380568504333496,
+ "step": 439
+ },
+ {
+ "epoch": 6.1117903930131,
+ "grad_norm": 0.020868154242634773,
+ "learning_rate": 0.0006,
+ "loss": 6.38240909576416,
+ "step": 440
+ },
+ {
+ "epoch": 6.125764192139738,
+ "grad_norm": 0.02113114856183529,
+ "learning_rate": 0.0006,
+ "loss": 6.377391815185547,
+ "step": 441
+ },
+ {
+ "epoch": 6.139737991266376,
+ "grad_norm": 0.018894119188189507,
+ "learning_rate": 0.0006,
+ "loss": 6.354569435119629,
+ "step": 442
+ },
+ {
+ "epoch": 6.153711790393013,
+ "grad_norm": 0.01820225641131401,
+ "learning_rate": 0.0006,
+ "loss": 6.392882347106934,
+ "step": 443
+ },
+ {
+ "epoch": 6.1676855895196505,
+ "grad_norm": 0.017510728910565376,
+ "learning_rate": 0.0006,
+ "loss": 6.364162445068359,
+ "step": 444
+ },
+ {
+ "epoch": 6.181659388646288,
+ "grad_norm": 0.017893686890602112,
+ "learning_rate": 0.0006,
+ "loss": 6.373456954956055,
+ "step": 445
+ },
+ {
+ "epoch": 6.195633187772926,
+ "grad_norm": 0.013382102362811565,
+ "learning_rate": 0.0006,
+ "loss": 6.3714399337768555,
+ "step": 446
+ },
+ {
+ "epoch": 6.209606986899563,
+ "grad_norm": 0.022540833801031113,
+ "learning_rate": 0.0006,
+ "loss": 6.348170757293701,
+ "step": 447
+ },
+ {
+ "epoch": 6.223580786026201,
+ "grad_norm": 0.02110104262828827,
+ "learning_rate": 0.0006,
+ "loss": 6.413206100463867,
+ "step": 448
+ },
+ {
+ "epoch": 6.2375545851528384,
+ "grad_norm": 0.014498349279165268,
+ "learning_rate": 0.0006,
+ "loss": 6.361384391784668,
+ "step": 449
+ },
+ {
+ "epoch": 6.251528384279476,
+ "grad_norm": 0.01819637045264244,
+ "learning_rate": 0.0006,
+ "loss": 6.315275192260742,
+ "step": 450
+ },
+ {
+ "epoch": 6.265502183406113,
+ "grad_norm": 0.0195972491055727,
+ "learning_rate": 0.0006,
+ "loss": 6.321994781494141,
+ "step": 451
+ },
+ {
+ "epoch": 6.279475982532751,
+ "grad_norm": 0.018822453916072845,
+ "learning_rate": 0.0006,
+ "loss": 6.345219612121582,
+ "step": 452
+ },
+ {
+ "epoch": 6.293449781659389,
+ "grad_norm": 0.019247712567448616,
+ "learning_rate": 0.0006,
+ "loss": 6.348073959350586,
+ "step": 453
+ },
+ {
+ "epoch": 6.307423580786026,
+ "grad_norm": 0.03175375238060951,
+ "learning_rate": 0.0006,
+ "loss": 6.300344944000244,
+ "step": 454
+ },
+ {
+ "epoch": 6.321397379912664,
+ "grad_norm": 0.06326444447040558,
+ "learning_rate": 0.0006,
+ "loss": 6.392579078674316,
+ "step": 455
+ },
+ {
+ "epoch": 6.335371179039301,
+ "grad_norm": 0.09667232632637024,
+ "learning_rate": 0.0006,
+ "loss": 6.36024284362793,
+ "step": 456
+ },
+ {
+ "epoch": 6.349344978165939,
+ "grad_norm": 0.05774446204304695,
+ "learning_rate": 0.0006,
+ "loss": 6.365338325500488,
+ "step": 457
+ },
+ {
+ "epoch": 6.3633187772925766,
+ "grad_norm": 0.08793356269598007,
+ "learning_rate": 0.0006,
+ "loss": 6.43284273147583,
+ "step": 458
+ },
+ {
+ "epoch": 6.377292576419214,
+ "grad_norm": 0.0640595331788063,
+ "learning_rate": 0.0006,
+ "loss": 6.342428207397461,
+ "step": 459
+ },
+ {
+ "epoch": 6.391266375545851,
+ "grad_norm": 0.0447838194668293,
+ "learning_rate": 0.0006,
+ "loss": 6.367971897125244,
+ "step": 460
+ },
+ {
+ "epoch": 6.405240174672489,
+ "grad_norm": 0.04768967255949974,
+ "learning_rate": 0.0006,
+ "loss": 6.35027551651001,
+ "step": 461
+ },
+ {
+ "epoch": 6.419213973799127,
+ "grad_norm": 0.027981851249933243,
+ "learning_rate": 0.0006,
+ "loss": 6.33723258972168,
+ "step": 462
+ },
+ {
+ "epoch": 6.4331877729257645,
+ "grad_norm": 0.03479623422026634,
+ "learning_rate": 0.0006,
+ "loss": 6.36590051651001,
+ "step": 463
+ },
+ {
+ "epoch": 6.447161572052401,
+ "grad_norm": 0.020846687257289886,
+ "learning_rate": 0.0006,
+ "loss": 6.315143585205078,
+ "step": 464
+ },
+ {
+ "epoch": 6.461135371179039,
+ "grad_norm": 0.0312928780913353,
+ "learning_rate": 0.0006,
+ "loss": 6.340150833129883,
+ "step": 465
+ },
+ {
+ "epoch": 6.475109170305677,
+ "grad_norm": 0.020593881607055664,
+ "learning_rate": 0.0006,
+ "loss": 6.348175048828125,
+ "step": 466
+ },
+ {
+ "epoch": 6.489082969432315,
+ "grad_norm": 0.022010236978530884,
+ "learning_rate": 0.0006,
+ "loss": 6.318607330322266,
+ "step": 467
+ },
+ {
+ "epoch": 6.503056768558952,
+ "grad_norm": 0.018427059054374695,
+ "learning_rate": 0.0006,
+ "loss": 6.331822395324707,
+ "step": 468
+ },
+ {
+ "epoch": 6.517030567685589,
+ "grad_norm": 0.019964968785643578,
+ "learning_rate": 0.0006,
+ "loss": 6.287320137023926,
+ "step": 469
+ },
+ {
+ "epoch": 6.531004366812227,
+ "grad_norm": 0.02130270190536976,
+ "learning_rate": 0.0006,
+ "loss": 6.353344917297363,
+ "step": 470
+ },
+ {
+ "epoch": 6.544978165938865,
+ "grad_norm": 0.01616514101624489,
+ "learning_rate": 0.0006,
+ "loss": 6.314052581787109,
+ "step": 471
+ },
+ {
+ "epoch": 6.558951965065502,
+ "grad_norm": 0.016639474779367447,
+ "learning_rate": 0.0006,
+ "loss": 6.320351600646973,
+ "step": 472
+ },
+ {
+ "epoch": 6.5729257641921395,
+ "grad_norm": 0.017620692029595375,
+ "learning_rate": 0.0006,
+ "loss": 6.311459064483643,
+ "step": 473
+ },
+ {
+ "epoch": 6.586899563318777,
+ "grad_norm": 0.013542844913899899,
+ "learning_rate": 0.0006,
+ "loss": 6.324368476867676,
+ "step": 474
+ },
+ {
+ "epoch": 6.600873362445415,
+ "grad_norm": 0.01308678649365902,
+ "learning_rate": 0.0006,
+ "loss": 6.307419776916504,
+ "step": 475
+ },
+ {
+ "epoch": 6.614847161572053,
+ "grad_norm": 0.014672392047941685,
+ "learning_rate": 0.0006,
+ "loss": 6.2785420417785645,
+ "step": 476
+ },
+ {
+ "epoch": 6.62882096069869,
+ "grad_norm": 0.01399585697799921,
+ "learning_rate": 0.0006,
+ "loss": 6.258619785308838,
+ "step": 477
+ },
+ {
+ "epoch": 6.642794759825327,
+ "grad_norm": 0.01330904383212328,
+ "learning_rate": 0.0006,
+ "loss": 6.299595832824707,
+ "step": 478
+ },
+ {
+ "epoch": 6.656768558951965,
+ "grad_norm": 0.016613377258181572,
+ "learning_rate": 0.0006,
+ "loss": 6.2900309562683105,
+ "step": 479
+ },
+ {
+ "epoch": 6.670742358078603,
+ "grad_norm": 0.015076331794261932,
+ "learning_rate": 0.0006,
+ "loss": 6.231028079986572,
+ "step": 480
+ },
+ {
+ "epoch": 6.68471615720524,
+ "grad_norm": 0.01589728519320488,
+ "learning_rate": 0.0006,
+ "loss": 6.257922172546387,
+ "step": 481
+ },
+ {
+ "epoch": 6.698689956331878,
+ "grad_norm": 0.030972328037023544,
+ "learning_rate": 0.0006,
+ "loss": 6.285238265991211,
+ "step": 482
+ },
+ {
+ "epoch": 6.712663755458515,
+ "grad_norm": 0.05048783868551254,
+ "learning_rate": 0.0006,
+ "loss": 6.287471771240234,
+ "step": 483
+ },
+ {
+ "epoch": 6.726637554585153,
+ "grad_norm": 0.07390714436769485,
+ "learning_rate": 0.0006,
+ "loss": 6.284293174743652,
+ "step": 484
+ },
+ {
+ "epoch": 6.74061135371179,
+ "grad_norm": 0.07166943699121475,
+ "learning_rate": 0.0006,
+ "loss": 6.290704727172852,
+ "step": 485
+ },
+ {
+ "epoch": 6.754585152838428,
+ "grad_norm": 0.028792452067136765,
+ "learning_rate": 0.0006,
+ "loss": 6.243415832519531,
+ "step": 486
+ },
+ {
+ "epoch": 6.7685589519650655,
+ "grad_norm": 0.04822041466832161,
+ "learning_rate": 0.0006,
+ "loss": 6.278229713439941,
+ "step": 487
+ },
+ {
+ "epoch": 6.782532751091703,
+ "grad_norm": 0.036688923835754395,
+ "learning_rate": 0.0006,
+ "loss": 6.275590896606445,
+ "step": 488
+ },
+ {
+ "epoch": 6.796506550218341,
+ "grad_norm": 0.03679114952683449,
+ "learning_rate": 0.0006,
+ "loss": 6.236509323120117,
+ "step": 489
+ },
+ {
+ "epoch": 6.810480349344978,
+ "grad_norm": 0.023249102756381035,
+ "learning_rate": 0.0006,
+ "loss": 6.230788230895996,
+ "step": 490
+ },
+ {
+ "epoch": 6.824454148471616,
+ "grad_norm": 0.02682286500930786,
+ "learning_rate": 0.0006,
+ "loss": 6.278890609741211,
+ "step": 491
+ },
+ {
+ "epoch": 6.8384279475982535,
+ "grad_norm": 0.02202044241130352,
+ "learning_rate": 0.0006,
+ "loss": 6.234531402587891,
+ "step": 492
+ },
+ {
+ "epoch": 6.85240174672489,
+ "grad_norm": 0.02545459009706974,
+ "learning_rate": 0.0006,
+ "loss": 6.271790981292725,
+ "step": 493
+ },
+ {
+ "epoch": 6.866375545851528,
+ "grad_norm": 0.02012932486832142,
+ "learning_rate": 0.0006,
+ "loss": 6.247462749481201,
+ "step": 494
+ },
+ {
+ "epoch": 6.880349344978166,
+ "grad_norm": 0.02216474711894989,
+ "learning_rate": 0.0006,
+ "loss": 6.26396369934082,
+ "step": 495
+ },
+ {
+ "epoch": 6.894323144104804,
+ "grad_norm": 0.016078704968094826,
+ "learning_rate": 0.0006,
+ "loss": 6.228609085083008,
+ "step": 496
+ },
+ {
+ "epoch": 6.908296943231441,
+ "grad_norm": 0.023021478205919266,
+ "learning_rate": 0.0006,
+ "loss": 6.197244167327881,
+ "step": 497
+ },
+ {
+ "epoch": 6.922270742358078,
+ "grad_norm": 0.024766333401203156,
+ "learning_rate": 0.0006,
+ "loss": 6.181090354919434,
+ "step": 498
+ },
+ {
+ "epoch": 6.936244541484716,
+ "grad_norm": 0.025581466034054756,
+ "learning_rate": 0.0006,
+ "loss": 6.201817512512207,
+ "step": 499
+ },
+ {
+ "epoch": 6.950218340611354,
+ "grad_norm": 0.016964981332421303,
+ "learning_rate": 0.0006,
+ "loss": 6.201486587524414,
+ "step": 500
+ },
+ {
+ "epoch": 6.964192139737992,
+ "grad_norm": 0.023957721889019012,
+ "learning_rate": 0.0006,
+ "loss": 6.185701847076416,
+ "step": 501
+ },
+ {
+ "epoch": 6.978165938864628,
+ "grad_norm": 0.029775245115160942,
+ "learning_rate": 0.0006,
+ "loss": 6.242060661315918,
+ "step": 502
+ },
+ {
+ "epoch": 6.992139737991266,
+ "grad_norm": 0.04007191210985184,
+ "learning_rate": 0.0006,
+ "loss": 6.19931697845459,
+ "step": 503
+ },
+ {
+ "epoch": 7.0,
+ "grad_norm": 0.04882276430726051,
+ "learning_rate": 0.0006,
+ "loss": 6.1880598068237305,
+ "step": 504
+ },
+ {
+ "epoch": 7.0,
+ "eval_loss": 7.018043518066406,
+ "eval_runtime": 59.8815,
+ "eval_samples_per_second": 40.781,
+ "eval_steps_per_second": 1.286,
+ "step": 504
+ },
+ {
+ "epoch": 7.013973799126638,
+ "grad_norm": 0.07144106924533844,
+ "learning_rate": 0.0006,
+ "loss": 6.172799110412598,
+ "step": 505
+ },
+ {
+ "epoch": 7.0279475982532755,
+ "grad_norm": 0.06712841987609863,
+ "learning_rate": 0.0006,
+ "loss": 6.228682518005371,
+ "step": 506
+ },
+ {
+ "epoch": 7.041921397379912,
+ "grad_norm": 0.04405239224433899,
+ "learning_rate": 0.0006,
+ "loss": 6.1827898025512695,
+ "step": 507
+ },
+ {
+ "epoch": 7.05589519650655,
+ "grad_norm": 0.03418995440006256,
+ "learning_rate": 0.0006,
+ "loss": 6.200876235961914,
+ "step": 508
+ },
+ {
+ "epoch": 7.069868995633188,
+ "grad_norm": 0.04002382606267929,
+ "learning_rate": 0.0006,
+ "loss": 6.1799211502075195,
+ "step": 509
+ },
+ {
+ "epoch": 7.083842794759826,
+ "grad_norm": 0.029748762026429176,
+ "learning_rate": 0.0006,
+ "loss": 6.125422477722168,
+ "step": 510
+ },
+ {
+ "epoch": 7.097816593886463,
+ "grad_norm": 0.03408357501029968,
+ "learning_rate": 0.0006,
+ "loss": 6.203458309173584,
+ "step": 511
+ },
+ {
+ "epoch": 7.1117903930131,
+ "grad_norm": 0.02227279357612133,
+ "learning_rate": 0.0006,
+ "loss": 6.135439872741699,
+ "step": 512
+ },
+ {
+ "epoch": 7.125764192139738,
+ "grad_norm": 0.02727862447500229,
+ "learning_rate": 0.0006,
+ "loss": 6.202380180358887,
+ "step": 513
+ },
+ {
+ "epoch": 7.139737991266376,
+ "grad_norm": 0.026019075885415077,
+ "learning_rate": 0.0006,
+ "loss": 6.201343536376953,
+ "step": 514
+ },
+ {
+ "epoch": 7.153711790393013,
+ "grad_norm": 0.027227282524108887,
+ "learning_rate": 0.0006,
+ "loss": 6.164059638977051,
+ "step": 515
+ },
+ {
+ "epoch": 7.1676855895196505,
+ "grad_norm": 0.027649980038404465,
+ "learning_rate": 0.0006,
+ "loss": 6.135448932647705,
+ "step": 516
+ },
+ {
+ "epoch": 7.181659388646288,
+ "grad_norm": 0.02168039232492447,
+ "learning_rate": 0.0006,
+ "loss": 6.154831886291504,
+ "step": 517
+ },
+ {
+ "epoch": 7.195633187772926,
+ "grad_norm": 0.02101467363536358,
+ "learning_rate": 0.0006,
+ "loss": 6.128695487976074,
+ "step": 518
+ },
+ {
+ "epoch": 7.209606986899563,
+ "grad_norm": 0.018344245851039886,
+ "learning_rate": 0.0006,
+ "loss": 6.136524200439453,
+ "step": 519
+ },
+ {
+ "epoch": 7.223580786026201,
+ "grad_norm": 0.020323453471064568,
+ "learning_rate": 0.0006,
+ "loss": 6.1206488609313965,
+ "step": 520
+ },
+ {
+ "epoch": 7.2375545851528384,
+ "grad_norm": 0.01750849559903145,
+ "learning_rate": 0.0006,
+ "loss": 6.128170967102051,
+ "step": 521
+ },
+ {
+ "epoch": 7.251528384279476,
+ "grad_norm": 0.017778221517801285,
+ "learning_rate": 0.0006,
+ "loss": 6.150458812713623,
+ "step": 522
+ },
+ {
+ "epoch": 7.265502183406113,
+ "grad_norm": 0.01863938942551613,
+ "learning_rate": 0.0006,
+ "loss": 6.100106716156006,
+ "step": 523
+ },
+ {
+ "epoch": 7.279475982532751,
+ "grad_norm": 0.020196080207824707,
+ "learning_rate": 0.0006,
+ "loss": 6.149346351623535,
+ "step": 524
+ },
+ {
+ "epoch": 7.293449781659389,
+ "grad_norm": 0.03203890845179558,
+ "learning_rate": 0.0006,
+ "loss": 6.124648094177246,
+ "step": 525
+ },
+ {
+ "epoch": 7.307423580786026,
+ "grad_norm": 0.05325170233845711,
+ "learning_rate": 0.0006,
+ "loss": 6.115618705749512,
+ "step": 526
+ },
+ {
+ "epoch": 7.321397379912664,
+ "grad_norm": 0.05888345465064049,
+ "learning_rate": 0.0006,
+ "loss": 6.165221214294434,
+ "step": 527
+ },
+ {
+ "epoch": 7.335371179039301,
+ "grad_norm": 0.026873115450143814,
+ "learning_rate": 0.0006,
+ "loss": 6.039590835571289,
+ "step": 528
+ },
+ {
+ "epoch": 7.349344978165939,
+ "grad_norm": 0.06304649263620377,
+ "learning_rate": 0.0006,
+ "loss": 6.131900787353516,
+ "step": 529
+ },
+ {
+ "epoch": 7.3633187772925766,
+ "grad_norm": 0.08567557483911514,
+ "learning_rate": 0.0006,
+ "loss": 6.218114852905273,
+ "step": 530
+ },
+ {
+ "epoch": 7.377292576419214,
+ "grad_norm": 0.08868218213319778,
+ "learning_rate": 0.0006,
+ "loss": 6.216706275939941,
+ "step": 531
+ },
+ {
+ "epoch": 7.391266375545851,
+ "grad_norm": 0.03955601155757904,
+ "learning_rate": 0.0006,
+ "loss": 6.1566925048828125,
+ "step": 532
+ },
+ {
+ "epoch": 7.405240174672489,
+ "grad_norm": 0.04979498311877251,
+ "learning_rate": 0.0006,
+ "loss": 6.19327449798584,
+ "step": 533
+ },
+ {
+ "epoch": 7.419213973799127,
+ "grad_norm": 0.03253026679158211,
+ "learning_rate": 0.0006,
+ "loss": 6.133993148803711,
+ "step": 534
+ },
+ {
+ "epoch": 7.4331877729257645,
+ "grad_norm": 0.03016652725636959,
+ "learning_rate": 0.0006,
+ "loss": 6.174962997436523,
+ "step": 535
+ },
+ {
+ "epoch": 7.447161572052401,
+ "grad_norm": 0.03159862384200096,
+ "learning_rate": 0.0006,
+ "loss": 6.082124710083008,
+ "step": 536
+ },
+ {
+ "epoch": 7.461135371179039,
+ "grad_norm": 0.027673959732055664,
+ "learning_rate": 0.0006,
+ "loss": 6.111873149871826,
+ "step": 537
+ },
+ {
+ "epoch": 7.475109170305677,
+ "grad_norm": 0.022789714857935905,
+ "learning_rate": 0.0006,
+ "loss": 6.129437446594238,
+ "step": 538
+ },
+ {
+ "epoch": 7.489082969432315,
+ "grad_norm": 0.030061012133955956,
+ "learning_rate": 0.0006,
+ "loss": 6.126370429992676,
+ "step": 539
+ },
+ {
+ "epoch": 7.503056768558952,
+ "grad_norm": 0.019007068127393723,
+ "learning_rate": 0.0006,
+ "loss": 6.07887077331543,
+ "step": 540
+ },
+ {
+ "epoch": 7.517030567685589,
+ "grad_norm": 0.01897653006017208,
+ "learning_rate": 0.0006,
+ "loss": 6.096936225891113,
+ "step": 541
+ },
+ {
+ "epoch": 7.531004366812227,
+ "grad_norm": 0.024659698829054832,
+ "learning_rate": 0.0006,
+ "loss": 6.0933380126953125,
+ "step": 542
+ },
+ {
+ "epoch": 7.544978165938865,
+ "grad_norm": 0.021924549713730812,
+ "learning_rate": 0.0006,
+ "loss": 6.105311870574951,
+ "step": 543
+ },
+ {
+ "epoch": 7.558951965065502,
+ "grad_norm": 0.018817558884620667,
+ "learning_rate": 0.0006,
+ "loss": 6.058895111083984,
+ "step": 544
+ },
+ {
+ "epoch": 7.5729257641921395,
+ "grad_norm": 0.02477840520441532,
+ "learning_rate": 0.0006,
+ "loss": 6.038928031921387,
+ "step": 545
+ },
+ {
+ "epoch": 7.586899563318777,
+ "grad_norm": 0.018968632444739342,
+ "learning_rate": 0.0006,
+ "loss": 6.1368937492370605,
+ "step": 546
+ },
+ {
+ "epoch": 7.600873362445415,
+ "grad_norm": 0.02409801445901394,
+ "learning_rate": 0.0006,
+ "loss": 6.110324859619141,
+ "step": 547
+ },
+ {
+ "epoch": 7.614847161572053,
+ "grad_norm": 0.017595326527953148,
+ "learning_rate": 0.0006,
+ "loss": 6.074368476867676,
+ "step": 548
+ },
+ {
+ "epoch": 7.62882096069869,
+ "grad_norm": 0.01875406503677368,
+ "learning_rate": 0.0006,
+ "loss": 6.121023654937744,
+ "step": 549
+ },
+ {
+ "epoch": 7.642794759825327,
+ "grad_norm": 0.018781628459692,
+ "learning_rate": 0.0006,
+ "loss": 6.064914703369141,
+ "step": 550
+ },
+ {
+ "epoch": 7.656768558951965,
+ "grad_norm": 0.017003126442432404,
+ "learning_rate": 0.0006,
+ "loss": 6.039608001708984,
+ "step": 551
+ },
+ {
+ "epoch": 7.670742358078603,
+ "grad_norm": 0.017665037885308266,
+ "learning_rate": 0.0006,
+ "loss": 6.053671836853027,
+ "step": 552
+ },
+ {
+ "epoch": 7.68471615720524,
+ "grad_norm": 0.02066265605390072,
+ "learning_rate": 0.0006,
+ "loss": 6.090536117553711,
+ "step": 553
+ },
+ {
+ "epoch": 7.698689956331878,
+ "grad_norm": 0.02888769656419754,
+ "learning_rate": 0.0006,
+ "loss": 5.9858903884887695,
+ "step": 554
+ },
+ {
+ "epoch": 7.712663755458515,
+ "grad_norm": 0.03744485229253769,
+ "learning_rate": 0.0006,
+ "loss": 6.086264610290527,
+ "step": 555
+ },
+ {
+ "epoch": 7.726637554585153,
+ "grad_norm": 0.04719431325793266,
+ "learning_rate": 0.0006,
+ "loss": 6.09747314453125,
+ "step": 556
+ },
+ {
+ "epoch": 7.74061135371179,
+ "grad_norm": 0.044963907450437546,
+ "learning_rate": 0.0006,
+ "loss": 6.091272354125977,
+ "step": 557
+ },
+ {
+ "epoch": 7.754585152838428,
+ "grad_norm": 0.03810194507241249,
+ "learning_rate": 0.0006,
+ "loss": 6.084904193878174,
+ "step": 558
+ },
+ {
+ "epoch": 7.7685589519650655,
+ "grad_norm": 0.04626769199967384,
+ "learning_rate": 0.0006,
+ "loss": 6.017384052276611,
+ "step": 559
+ },
+ {
+ "epoch": 7.782532751091703,
+ "grad_norm": 0.03459552302956581,
+ "learning_rate": 0.0006,
+ "loss": 6.111314296722412,
+ "step": 560
+ },
+ {
+ "epoch": 7.796506550218341,
+ "grad_norm": 0.027385666966438293,
+ "learning_rate": 0.0006,
+ "loss": 6.002509117126465,
+ "step": 561
+ },
+ {
+ "epoch": 7.810480349344978,
+ "grad_norm": 0.03417652100324631,
+ "learning_rate": 0.0006,
+ "loss": 6.0421061515808105,
+ "step": 562
+ },
+ {
+ "epoch": 7.824454148471616,
+ "grad_norm": 0.04415450245141983,
+ "learning_rate": 0.0006,
+ "loss": 6.035305023193359,
+ "step": 563
+ },
+ {
+ "epoch": 7.8384279475982535,
+ "grad_norm": 0.03259635716676712,
+ "learning_rate": 0.0006,
+ "loss": 6.025840759277344,
+ "step": 564
+ },
+ {
+ "epoch": 7.85240174672489,
+ "grad_norm": 0.024941129609942436,
+ "learning_rate": 0.0006,
+ "loss": 6.043416976928711,
+ "step": 565
+ },
+ {
+ "epoch": 7.866375545851528,
+ "grad_norm": 0.02434500865638256,
+ "learning_rate": 0.0006,
+ "loss": 6.002052307128906,
+ "step": 566
+ },
+ {
+ "epoch": 7.880349344978166,
+ "grad_norm": 0.01876584254205227,
+ "learning_rate": 0.0006,
+ "loss": 6.045035362243652,
+ "step": 567
+ },
+ {
+ "epoch": 7.894323144104804,
+ "grad_norm": 0.02254367619752884,
+ "learning_rate": 0.0006,
+ "loss": 5.972503662109375,
+ "step": 568
+ },
+ {
+ "epoch": 7.908296943231441,
+ "grad_norm": 0.024504603818058968,
+ "learning_rate": 0.0006,
+ "loss": 5.992846488952637,
+ "step": 569
+ },
+ {
+ "epoch": 7.922270742358078,
+ "grad_norm": 0.02317827194929123,
+ "learning_rate": 0.0006,
+ "loss": 6.007697582244873,
+ "step": 570
+ },
+ {
+ "epoch": 7.936244541484716,
+ "grad_norm": 0.02225821278989315,
+ "learning_rate": 0.0006,
+ "loss": 6.01463508605957,
+ "step": 571
+ },
+ {
+ "epoch": 7.950218340611354,
+ "grad_norm": 0.01811717264354229,
+ "learning_rate": 0.0006,
+ "loss": 5.995022773742676,
+ "step": 572
+ },
+ {
+ "epoch": 7.964192139737992,
+ "grad_norm": 0.019589431583881378,
+ "learning_rate": 0.0006,
+ "loss": 5.990633964538574,
+ "step": 573
+ },
+ {
+ "epoch": 7.978165938864628,
+ "grad_norm": 0.015422016382217407,
+ "learning_rate": 0.0006,
+ "loss": 6.011263847351074,
+ "step": 574
+ },
+ {
+ "epoch": 7.992139737991266,
+ "grad_norm": 0.023298880085349083,
+ "learning_rate": 0.0006,
+ "loss": 5.9954938888549805,
+ "step": 575
+ },
+ {
+ "epoch": 8.0,
+ "grad_norm": 0.022172044962644577,
+ "learning_rate": 0.0006,
+ "loss": 5.95679235458374,
+ "step": 576
+ },
+ {
+ "epoch": 8.0,
+ "eval_loss": 6.75809383392334,
+ "eval_runtime": 72.8926,
+ "eval_samples_per_second": 33.501,
+ "eval_steps_per_second": 1.056,
+ "step": 576
+ },
+ {
+ "epoch": 8.013973799126637,
+ "grad_norm": 0.029082221910357475,
+ "learning_rate": 0.0006,
+ "loss": 6.018675804138184,
+ "step": 577
+ },
+ {
+ "epoch": 8.027947598253276,
+ "grad_norm": 0.03959495946764946,
+ "learning_rate": 0.0006,
+ "loss": 5.93310546875,
+ "step": 578
+ },
+ {
+ "epoch": 8.041921397379912,
+ "grad_norm": 0.043756261467933655,
+ "learning_rate": 0.0006,
+ "loss": 5.965944290161133,
+ "step": 579
+ },
+ {
+ "epoch": 8.055895196506551,
+ "grad_norm": 0.04996631667017937,
+ "learning_rate": 0.0006,
+ "loss": 6.0381059646606445,
+ "step": 580
+ },
+ {
+ "epoch": 8.069868995633188,
+ "grad_norm": 0.05721999332308769,
+ "learning_rate": 0.0006,
+ "loss": 5.968381881713867,
+ "step": 581
+ },
+ {
+ "epoch": 8.083842794759825,
+ "grad_norm": 0.07594525814056396,
+ "learning_rate": 0.0006,
+ "loss": 6.027225494384766,
+ "step": 582
+ },
+ {
+ "epoch": 8.097816593886463,
+ "grad_norm": 0.06861685961484909,
+ "learning_rate": 0.0006,
+ "loss": 6.035175323486328,
+ "step": 583
+ },
+ {
+ "epoch": 8.1117903930131,
+ "grad_norm": 0.04156472906470299,
+ "learning_rate": 0.0006,
+ "loss": 5.9500017166137695,
+ "step": 584
+ },
+ {
+ "epoch": 8.125764192139737,
+ "grad_norm": 0.045399412512779236,
+ "learning_rate": 0.0006,
+ "loss": 5.9953389167785645,
+ "step": 585
+ },
+ {
+ "epoch": 8.139737991266376,
+ "grad_norm": 0.04399585723876953,
+ "learning_rate": 0.0006,
+ "loss": 5.986051559448242,
+ "step": 586
+ },
+ {
+ "epoch": 8.153711790393013,
+ "grad_norm": 0.05764509737491608,
+ "learning_rate": 0.0006,
+ "loss": 6.005392074584961,
+ "step": 587
+ },
+ {
+ "epoch": 8.167685589519651,
+ "grad_norm": 0.060871850699186325,
+ "learning_rate": 0.0006,
+ "loss": 6.01304817199707,
+ "step": 588
+ },
+ {
+ "epoch": 8.181659388646288,
+ "grad_norm": 0.03381241112947464,
+ "learning_rate": 0.0006,
+ "loss": 5.977019309997559,
+ "step": 589
+ },
+ {
+ "epoch": 8.195633187772925,
+ "grad_norm": 0.03384104371070862,
+ "learning_rate": 0.0006,
+ "loss": 5.988002777099609,
+ "step": 590
+ },
+ {
+ "epoch": 8.209606986899564,
+ "grad_norm": 0.03120286948978901,
+ "learning_rate": 0.0006,
+ "loss": 5.968227386474609,
+ "step": 591
+ },
+ {
+ "epoch": 8.2235807860262,
+ "grad_norm": 0.026532653719186783,
+ "learning_rate": 0.0006,
+ "loss": 5.936810493469238,
+ "step": 592
+ },
+ {
+ "epoch": 8.237554585152838,
+ "grad_norm": 0.024956922978162766,
+ "learning_rate": 0.0006,
+ "loss": 5.9348015785217285,
+ "step": 593
+ },
+ {
+ "epoch": 8.251528384279476,
+ "grad_norm": 0.025926144793629646,
+ "learning_rate": 0.0006,
+ "loss": 5.952630519866943,
+ "step": 594
+ },
+ {
+ "epoch": 8.265502183406113,
+ "grad_norm": 0.02893793396651745,
+ "learning_rate": 0.0006,
+ "loss": 5.9569854736328125,
+ "step": 595
+ },
+ {
+ "epoch": 8.279475982532752,
+ "grad_norm": 0.0320001058280468,
+ "learning_rate": 0.0006,
+ "loss": 5.917050838470459,
+ "step": 596
+ },
+ {
+ "epoch": 8.293449781659389,
+ "grad_norm": 0.03925463184714317,
+ "learning_rate": 0.0006,
+ "loss": 5.9925127029418945,
+ "step": 597
+ },
+ {
+ "epoch": 8.307423580786025,
+ "grad_norm": 0.05096640810370445,
+ "learning_rate": 0.0006,
+ "loss": 5.926358222961426,
+ "step": 598
+ },
+ {
+ "epoch": 8.321397379912664,
+ "grad_norm": 0.041637953370809555,
+ "learning_rate": 0.0006,
+ "loss": 5.954983711242676,
+ "step": 599
+ },
+ {
+ "epoch": 8.335371179039301,
+ "grad_norm": 0.03138125315308571,
+ "learning_rate": 0.0006,
+ "loss": 5.893692970275879,
+ "step": 600
+ },
+ {
+ "epoch": 8.34934497816594,
+ "grad_norm": 0.029299303889274597,
+ "learning_rate": 0.0006,
+ "loss": 5.94670295715332,
+ "step": 601
+ },
+ {
+ "epoch": 8.363318777292577,
+ "grad_norm": 0.023969192057847977,
+ "learning_rate": 0.0006,
+ "loss": 5.925658702850342,
+ "step": 602
+ },
+ {
+ "epoch": 8.377292576419213,
+ "grad_norm": 0.025406932458281517,
+ "learning_rate": 0.0006,
+ "loss": 5.966875076293945,
+ "step": 603
+ },
+ {
+ "epoch": 8.391266375545852,
+ "grad_norm": 0.024309059605002403,
+ "learning_rate": 0.0006,
+ "loss": 5.932581424713135,
+ "step": 604
+ },
+ {
+ "epoch": 8.405240174672489,
+ "grad_norm": 0.017237074673175812,
+ "learning_rate": 0.0006,
+ "loss": 5.898823261260986,
+ "step": 605
+ },
+ {
+ "epoch": 8.419213973799126,
+ "grad_norm": 0.025594931095838547,
+ "learning_rate": 0.0006,
+ "loss": 5.9285993576049805,
+ "step": 606
+ },
+ {
+ "epoch": 8.433187772925764,
+ "grad_norm": 0.025184160098433495,
+ "learning_rate": 0.0006,
+ "loss": 5.915966510772705,
+ "step": 607
+ },
+ {
+ "epoch": 8.447161572052401,
+ "grad_norm": 0.019004398956894875,
+ "learning_rate": 0.0006,
+ "loss": 5.893395900726318,
+ "step": 608
+ },
+ {
+ "epoch": 8.46113537117904,
+ "grad_norm": 0.021802261471748352,
+ "learning_rate": 0.0006,
+ "loss": 5.953311443328857,
+ "step": 609
+ },
+ {
+ "epoch": 8.475109170305677,
+ "grad_norm": 0.02202655002474785,
+ "learning_rate": 0.0006,
+ "loss": 5.953000545501709,
+ "step": 610
+ },
+ {
+ "epoch": 8.489082969432314,
+ "grad_norm": 0.02067241445183754,
+ "learning_rate": 0.0006,
+ "loss": 5.869247913360596,
+ "step": 611
+ },
+ {
+ "epoch": 8.503056768558952,
+ "grad_norm": 0.015430927276611328,
+ "learning_rate": 0.0006,
+ "loss": 5.855883598327637,
+ "step": 612
+ },
+ {
+ "epoch": 8.51703056768559,
+ "grad_norm": 0.019917506724596024,
+ "learning_rate": 0.0006,
+ "loss": 5.9284539222717285,
+ "step": 613
+ },
+ {
+ "epoch": 8.531004366812226,
+ "grad_norm": 0.024291323497891426,
+ "learning_rate": 0.0006,
+ "loss": 5.973604202270508,
+ "step": 614
+ },
+ {
+ "epoch": 8.544978165938865,
+ "grad_norm": 0.022133296355605125,
+ "learning_rate": 0.0006,
+ "loss": 5.85693359375,
+ "step": 615
+ },
+ {
+ "epoch": 8.558951965065502,
+ "grad_norm": 0.0254366435110569,
+ "learning_rate": 0.0006,
+ "loss": 5.939558982849121,
+ "step": 616
+ },
+ {
+ "epoch": 8.57292576419214,
+ "grad_norm": 0.030134694650769234,
+ "learning_rate": 0.0006,
+ "loss": 5.892992973327637,
+ "step": 617
+ },
+ {
+ "epoch": 8.586899563318777,
+ "grad_norm": 0.03845192864537239,
+ "learning_rate": 0.0006,
+ "loss": 5.89303731918335,
+ "step": 618
+ },
+ {
+ "epoch": 8.600873362445414,
+ "grad_norm": 0.042409464716911316,
+ "learning_rate": 0.0006,
+ "loss": 5.914068222045898,
+ "step": 619
+ },
+ {
+ "epoch": 8.614847161572053,
+ "grad_norm": 0.0454271174967289,
+ "learning_rate": 0.0006,
+ "loss": 5.853579044342041,
+ "step": 620
+ },
+ {
+ "epoch": 8.62882096069869,
+ "grad_norm": 0.04934931918978691,
+ "learning_rate": 0.0006,
+ "loss": 5.881861686706543,
+ "step": 621
+ },
+ {
+ "epoch": 8.642794759825328,
+ "grad_norm": 0.0519695058465004,
+ "learning_rate": 0.0006,
+ "loss": 5.872620582580566,
+ "step": 622
+ },
+ {
+ "epoch": 8.656768558951965,
+ "grad_norm": 0.06172563508152962,
+ "learning_rate": 0.0006,
+ "loss": 5.938283443450928,
+ "step": 623
+ },
+ {
+ "epoch": 8.670742358078602,
+ "grad_norm": 0.08416368812322617,
+ "learning_rate": 0.0006,
+ "loss": 5.943403720855713,
+ "step": 624
+ },
+ {
+ "epoch": 8.68471615720524,
+ "grad_norm": 0.07311230152845383,
+ "learning_rate": 0.0006,
+ "loss": 5.954029083251953,
+ "step": 625
+ },
+ {
+ "epoch": 8.698689956331878,
+ "grad_norm": 0.09851036965847015,
+ "learning_rate": 0.0006,
+ "loss": 5.992746829986572,
+ "step": 626
+ },
+ {
+ "epoch": 8.712663755458514,
+ "grad_norm": 0.07521767914295197,
+ "learning_rate": 0.0006,
+ "loss": 5.956884384155273,
+ "step": 627
+ },
+ {
+ "epoch": 8.726637554585153,
+ "grad_norm": 0.05720978602766991,
+ "learning_rate": 0.0006,
+ "loss": 5.911205291748047,
+ "step": 628
+ },
+ {
+ "epoch": 8.74061135371179,
+ "grad_norm": 0.05360936373472214,
+ "learning_rate": 0.0006,
+ "loss": 5.9734787940979,
+ "step": 629
+ },
+ {
+ "epoch": 8.754585152838429,
+ "grad_norm": 0.035826463252305984,
+ "learning_rate": 0.0006,
+ "loss": 5.95751953125,
+ "step": 630
+ },
+ {
+ "epoch": 8.768558951965066,
+ "grad_norm": 0.030137093737721443,
+ "learning_rate": 0.0006,
+ "loss": 5.91911506652832,
+ "step": 631
+ },
+ {
+ "epoch": 8.782532751091702,
+ "grad_norm": 0.030039150267839432,
+ "learning_rate": 0.0006,
+ "loss": 5.910290718078613,
+ "step": 632
+ },
+ {
+ "epoch": 8.796506550218341,
+ "grad_norm": 0.03349437564611435,
+ "learning_rate": 0.0006,
+ "loss": 5.904623985290527,
+ "step": 633
+ },
+ {
+ "epoch": 8.810480349344978,
+ "grad_norm": 0.02211189828813076,
+ "learning_rate": 0.0006,
+ "loss": 5.909668445587158,
+ "step": 634
+ },
+ {
+ "epoch": 8.824454148471617,
+ "grad_norm": 0.022601254284381866,
+ "learning_rate": 0.0006,
+ "loss": 5.942885875701904,
+ "step": 635
+ },
+ {
+ "epoch": 8.838427947598253,
+ "grad_norm": 0.02197270467877388,
+ "learning_rate": 0.0006,
+ "loss": 5.867103576660156,
+ "step": 636
+ },
+ {
+ "epoch": 8.85240174672489,
+ "grad_norm": 0.022765018045902252,
+ "learning_rate": 0.0006,
+ "loss": 5.894528865814209,
+ "step": 637
+ },
+ {
+ "epoch": 8.866375545851529,
+ "grad_norm": 0.021181151270866394,
+ "learning_rate": 0.0006,
+ "loss": 5.896910667419434,
+ "step": 638
+ },
+ {
+ "epoch": 8.880349344978166,
+ "grad_norm": 0.01878993771970272,
+ "learning_rate": 0.0006,
+ "loss": 5.860502243041992,
+ "step": 639
+ },
+ {
+ "epoch": 8.894323144104803,
+ "grad_norm": 0.014679803512990475,
+ "learning_rate": 0.0006,
+ "loss": 5.8725266456604,
+ "step": 640
+ },
+ {
+ "epoch": 8.908296943231441,
+ "grad_norm": 0.017919695004820824,
+ "learning_rate": 0.0006,
+ "loss": 5.88902473449707,
+ "step": 641
+ },
+ {
+ "epoch": 8.922270742358078,
+ "grad_norm": 0.01437009871006012,
+ "learning_rate": 0.0006,
+ "loss": 5.868186950683594,
+ "step": 642
+ },
+ {
+ "epoch": 8.936244541484717,
+ "grad_norm": 0.014293290674686432,
+ "learning_rate": 0.0006,
+ "loss": 5.833364963531494,
+ "step": 643
+ },
+ {
+ "epoch": 8.950218340611354,
+ "grad_norm": 0.015440084971487522,
+ "learning_rate": 0.0006,
+ "loss": 5.860260009765625,
+ "step": 644
+ },
+ {
+ "epoch": 8.96419213973799,
+ "grad_norm": 0.014292514882981777,
+ "learning_rate": 0.0006,
+ "loss": 5.844939231872559,
+ "step": 645
+ },
+ {
+ "epoch": 8.97816593886463,
+ "grad_norm": 0.015702251344919205,
+ "learning_rate": 0.0006,
+ "loss": 5.825929164886475,
+ "step": 646
+ },
+ {
+ "epoch": 8.992139737991266,
+ "grad_norm": 0.013372798450291157,
+ "learning_rate": 0.0006,
+ "loss": 5.83062744140625,
+ "step": 647
+ },
+ {
+ "epoch": 9.0,
+ "grad_norm": 0.013794616796076298,
+ "learning_rate": 0.0006,
+ "loss": 5.7897820472717285,
+ "step": 648
+ },
+ {
+ "epoch": 9.0,
+ "eval_loss": 6.600830078125,
+ "eval_runtime": 71.4316,
+ "eval_samples_per_second": 34.187,
+ "eval_steps_per_second": 1.078,
+ "step": 648
+ },
+ {
+ "epoch": 9.013973799126637,
+ "grad_norm": 0.01262663770467043,
+ "learning_rate": 0.0006,
+ "loss": 5.802824020385742,
+ "step": 649
+ },
+ {
+ "epoch": 9.027947598253276,
+ "grad_norm": 0.016375157982110977,
+ "learning_rate": 0.0006,
+ "loss": 5.8018317222595215,
+ "step": 650
+ },
+ {
+ "epoch": 9.041921397379912,
+ "grad_norm": 0.010541851632297039,
+ "learning_rate": 0.0006,
+ "loss": 5.848208427429199,
+ "step": 651
+ },
+ {
+ "epoch": 9.055895196506551,
+ "grad_norm": 0.011726257391273975,
+ "learning_rate": 0.0006,
+ "loss": 5.771823883056641,
+ "step": 652
+ },
+ {
+ "epoch": 9.069868995633188,
+ "grad_norm": 0.01124380249530077,
+ "learning_rate": 0.0006,
+ "loss": 5.773855209350586,
+ "step": 653
+ },
+ {
+ "epoch": 9.083842794759825,
+ "grad_norm": 0.010801885277032852,
+ "learning_rate": 0.0006,
+ "loss": 5.847390651702881,
+ "step": 654
+ },
+ {
+ "epoch": 9.097816593886463,
+ "grad_norm": 0.013018375262618065,
+ "learning_rate": 0.0006,
+ "loss": 5.786786079406738,
+ "step": 655
+ },
+ {
+ "epoch": 9.1117903930131,
+ "grad_norm": 0.014233003370463848,
+ "learning_rate": 0.0006,
+ "loss": 5.7913713455200195,
+ "step": 656
+ },
+ {
+ "epoch": 9.125764192139737,
+ "grad_norm": 0.01465831883251667,
+ "learning_rate": 0.0006,
+ "loss": 5.821170330047607,
+ "step": 657
+ },
+ {
+ "epoch": 9.139737991266376,
+ "grad_norm": 0.017383430153131485,
+ "learning_rate": 0.0006,
+ "loss": 5.811469078063965,
+ "step": 658
+ },
+ {
+ "epoch": 9.153711790393013,
+ "grad_norm": 0.020328007638454437,
+ "learning_rate": 0.0006,
+ "loss": 5.718976020812988,
+ "step": 659
+ },
+ {
+ "epoch": 9.167685589519651,
+ "grad_norm": 0.01776740700006485,
+ "learning_rate": 0.0006,
+ "loss": 5.80593729019165,
+ "step": 660
+ },
+ {
+ "epoch": 9.181659388646288,
+ "grad_norm": 0.017713800072669983,
+ "learning_rate": 0.0006,
+ "loss": 5.744930267333984,
+ "step": 661
+ },
+ {
+ "epoch": 9.195633187772925,
+ "grad_norm": 0.028774067759513855,
+ "learning_rate": 0.0006,
+ "loss": 5.85262393951416,
+ "step": 662
+ },
+ {
+ "epoch": 9.209606986899564,
+ "grad_norm": 0.05153049901127815,
+ "learning_rate": 0.0006,
+ "loss": 5.739948272705078,
+ "step": 663
+ },
+ {
+ "epoch": 9.2235807860262,
+ "grad_norm": 0.06882233172655106,
+ "learning_rate": 0.0006,
+ "loss": 5.844270706176758,
+ "step": 664
+ },
+ {
+ "epoch": 9.237554585152838,
+ "grad_norm": 0.0420759879052639,
+ "learning_rate": 0.0006,
+ "loss": 5.79060173034668,
+ "step": 665
+ },
+ {
+ "epoch": 9.251528384279476,
+ "grad_norm": 0.05598323419690132,
+ "learning_rate": 0.0006,
+ "loss": 5.833735466003418,
+ "step": 666
+ },
+ {
+ "epoch": 9.265502183406113,
+ "grad_norm": 0.05011717975139618,
+ "learning_rate": 0.0006,
+ "loss": 5.840887069702148,
+ "step": 667
+ },
+ {
+ "epoch": 9.279475982532752,
+ "grad_norm": 0.040782492607831955,
+ "learning_rate": 0.0006,
+ "loss": 5.7905449867248535,
+ "step": 668
+ },
+ {
+ "epoch": 9.293449781659389,
+ "grad_norm": 0.030804185196757317,
+ "learning_rate": 0.0006,
+ "loss": 5.802691459655762,
+ "step": 669
+ },
+ {
+ "epoch": 9.307423580786025,
+ "grad_norm": 0.032001759856939316,
+ "learning_rate": 0.0006,
+ "loss": 5.726612091064453,
+ "step": 670
+ },
+ {
+ "epoch": 9.321397379912664,
+ "grad_norm": 0.0295492485165596,
+ "learning_rate": 0.0006,
+ "loss": 5.798826694488525,
+ "step": 671
+ },
+ {
+ "epoch": 9.335371179039301,
+ "grad_norm": 0.034406546503305435,
+ "learning_rate": 0.0006,
+ "loss": 5.720219135284424,
+ "step": 672
+ },
+ {
+ "epoch": 9.34934497816594,
+ "grad_norm": 0.02627669647336006,
+ "learning_rate": 0.0006,
+ "loss": 5.723702430725098,
+ "step": 673
+ },
+ {
+ "epoch": 9.363318777292577,
+ "grad_norm": 0.0282363910228014,
+ "learning_rate": 0.0006,
+ "loss": 5.818911552429199,
+ "step": 674
+ },
+ {
+ "epoch": 9.377292576419213,
+ "grad_norm": 0.017894433811306953,
+ "learning_rate": 0.0006,
+ "loss": 5.772168159484863,
+ "step": 675
+ },
+ {
+ "epoch": 9.391266375545852,
+ "grad_norm": 0.019017303362488747,
+ "learning_rate": 0.0006,
+ "loss": 5.774587631225586,
+ "step": 676
+ },
+ {
+ "epoch": 9.405240174672489,
+ "grad_norm": 0.022159283980727196,
+ "learning_rate": 0.0006,
+ "loss": 5.812951564788818,
+ "step": 677
+ },
+ {
+ "epoch": 9.419213973799126,
+ "grad_norm": 0.02198116108775139,
+ "learning_rate": 0.0006,
+ "loss": 5.808734893798828,
+ "step": 678
+ },
+ {
+ "epoch": 9.433187772925764,
+ "grad_norm": 0.017741188406944275,
+ "learning_rate": 0.0006,
+ "loss": 5.776142120361328,
+ "step": 679
+ },
+ {
+ "epoch": 9.447161572052401,
+ "grad_norm": 0.017197400331497192,
+ "learning_rate": 0.0006,
+ "loss": 5.7471208572387695,
+ "step": 680
+ },
+ {
+ "epoch": 9.46113537117904,
+ "grad_norm": 0.01752566732466221,
+ "learning_rate": 0.0006,
+ "loss": 5.780961990356445,
+ "step": 681
+ },
+ {
+ "epoch": 9.475109170305677,
+ "grad_norm": 0.017961280420422554,
+ "learning_rate": 0.0006,
+ "loss": 5.713686943054199,
+ "step": 682
+ },
+ {
+ "epoch": 9.489082969432314,
+ "grad_norm": 0.02227521874010563,
+ "learning_rate": 0.0006,
+ "loss": 5.717774868011475,
+ "step": 683
+ },
+ {
+ "epoch": 9.503056768558952,
+ "grad_norm": 0.024075504392385483,
+ "learning_rate": 0.0006,
+ "loss": 5.715015411376953,
+ "step": 684
+ },
+ {
+ "epoch": 9.51703056768559,
+ "grad_norm": 0.029195033013820648,
+ "learning_rate": 0.0006,
+ "loss": 5.659659385681152,
+ "step": 685
+ },
+ {
+ "epoch": 9.531004366812226,
+ "grad_norm": 0.02959120273590088,
+ "learning_rate": 0.0006,
+ "loss": 5.78981351852417,
+ "step": 686
+ },
+ {
+ "epoch": 9.544978165938865,
+ "grad_norm": 0.021854368969798088,
+ "learning_rate": 0.0006,
+ "loss": 5.750844955444336,
+ "step": 687
+ },
+ {
+ "epoch": 9.558951965065502,
+ "grad_norm": 0.017120081931352615,
+ "learning_rate": 0.0006,
+ "loss": 5.672736167907715,
+ "step": 688
+ },
+ {
+ "epoch": 9.57292576419214,
+ "grad_norm": 0.020712822675704956,
+ "learning_rate": 0.0006,
+ "loss": 5.708046913146973,
+ "step": 689
+ },
+ {
+ "epoch": 9.586899563318777,
+ "grad_norm": 0.019186170771718025,
+ "learning_rate": 0.0006,
+ "loss": 5.71742057800293,
+ "step": 690
+ },
+ {
+ "epoch": 9.600873362445414,
+ "grad_norm": 0.01985814794898033,
+ "learning_rate": 0.0006,
+ "loss": 5.735311508178711,
+ "step": 691
+ },
+ {
+ "epoch": 9.614847161572053,
+ "grad_norm": 0.01965234987437725,
+ "learning_rate": 0.0006,
+ "loss": 5.728727340698242,
+ "step": 692
+ },
+ {
+ "epoch": 9.62882096069869,
+ "grad_norm": 0.019827688112854958,
+ "learning_rate": 0.0006,
+ "loss": 5.743068695068359,
+ "step": 693
+ },
+ {
+ "epoch": 9.642794759825328,
+ "grad_norm": 0.025500580668449402,
+ "learning_rate": 0.0006,
+ "loss": 5.76585578918457,
+ "step": 694
+ },
+ {
+ "epoch": 9.656768558951965,
+ "grad_norm": 0.03701791539788246,
+ "learning_rate": 0.0006,
+ "loss": 5.728367805480957,
+ "step": 695
+ },
+ {
+ "epoch": 9.670742358078602,
+ "grad_norm": 0.04173659533262253,
+ "learning_rate": 0.0006,
+ "loss": 5.683089256286621,
+ "step": 696
+ },
+ {
+ "epoch": 9.68471615720524,
+ "grad_norm": 0.031390558928251266,
+ "learning_rate": 0.0006,
+ "loss": 5.740921974182129,
+ "step": 697
+ },
+ {
+ "epoch": 9.698689956331878,
+ "grad_norm": 0.03970158472657204,
+ "learning_rate": 0.0006,
+ "loss": 5.730632781982422,
+ "step": 698
+ },
+ {
+ "epoch": 9.712663755458514,
+ "grad_norm": 0.03865141049027443,
+ "learning_rate": 0.0006,
+ "loss": 5.686657905578613,
+ "step": 699
+ },
+ {
+ "epoch": 9.726637554585153,
+ "grad_norm": 0.04856107011437416,
+ "learning_rate": 0.0006,
+ "loss": 5.657513618469238,
+ "step": 700
+ },
+ {
+ "epoch": 9.74061135371179,
+ "grad_norm": 0.04016133397817612,
+ "learning_rate": 0.0006,
+ "loss": 5.754552841186523,
+ "step": 701
+ },
+ {
+ "epoch": 9.754585152838429,
+ "grad_norm": 0.03961458429694176,
+ "learning_rate": 0.0006,
+ "loss": 5.672088623046875,
+ "step": 702
+ },
+ {
+ "epoch": 9.768558951965066,
+ "grad_norm": 0.03403536602854729,
+ "learning_rate": 0.0006,
+ "loss": 5.736673355102539,
+ "step": 703
+ },
+ {
+ "epoch": 9.782532751091702,
+ "grad_norm": 0.03766333684325218,
+ "learning_rate": 0.0006,
+ "loss": 5.716219902038574,
+ "step": 704
+ },
+ {
+ "epoch": 9.796506550218341,
+ "grad_norm": 0.04085501655936241,
+ "learning_rate": 0.0006,
+ "loss": 5.678958415985107,
+ "step": 705
+ },
+ {
+ "epoch": 9.810480349344978,
+ "grad_norm": 0.036888182163238525,
+ "learning_rate": 0.0006,
+ "loss": 5.6990461349487305,
+ "step": 706
+ },
+ {
+ "epoch": 9.824454148471617,
+ "grad_norm": 0.0492791011929512,
+ "learning_rate": 0.0006,
+ "loss": 5.78053092956543,
+ "step": 707
+ },
+ {
+ "epoch": 9.838427947598253,
+ "grad_norm": 0.05572429671883583,
+ "learning_rate": 0.0006,
+ "loss": 5.7543535232543945,
+ "step": 708
+ },
+ {
+ "epoch": 9.85240174672489,
+ "grad_norm": 0.047000352293252945,
+ "learning_rate": 0.0006,
+ "loss": 5.7174272537231445,
+ "step": 709
+ },
+ {
+ "epoch": 9.866375545851529,
+ "grad_norm": 0.058798179030418396,
+ "learning_rate": 0.0006,
+ "loss": 5.793713569641113,
+ "step": 710
+ },
+ {
+ "epoch": 9.880349344978166,
+ "grad_norm": 0.046992287039756775,
+ "learning_rate": 0.0006,
+ "loss": 5.784457683563232,
+ "step": 711
+ },
+ {
+ "epoch": 9.894323144104803,
+ "grad_norm": 0.04382370039820671,
+ "learning_rate": 0.0006,
+ "loss": 5.793731212615967,
+ "step": 712
+ },
+ {
+ "epoch": 9.908296943231441,
+ "grad_norm": 0.040616173297166824,
+ "learning_rate": 0.0006,
+ "loss": 5.69599723815918,
+ "step": 713
+ },
+ {
+ "epoch": 9.922270742358078,
+ "grad_norm": 0.03446170687675476,
+ "learning_rate": 0.0006,
+ "loss": 5.694674968719482,
+ "step": 714
+ },
+ {
+ "epoch": 9.936244541484717,
+ "grad_norm": 0.029936550185084343,
+ "learning_rate": 0.0006,
+ "loss": 5.6704487800598145,
+ "step": 715
+ },
+ {
+ "epoch": 9.950218340611354,
+ "grad_norm": 0.02224753610789776,
+ "learning_rate": 0.0006,
+ "loss": 5.728894233703613,
+ "step": 716
+ },
+ {
+ "epoch": 9.96419213973799,
+ "grad_norm": 0.024346651509404182,
+ "learning_rate": 0.0006,
+ "loss": 5.772412300109863,
+ "step": 717
+ },
+ {
+ "epoch": 9.97816593886463,
+ "grad_norm": 0.022997068241238594,
+ "learning_rate": 0.0006,
+ "loss": 5.645310401916504,
+ "step": 718
+ },
+ {
+ "epoch": 9.992139737991266,
+ "grad_norm": 0.024918686598539352,
+ "learning_rate": 0.0006,
+ "loss": 5.671009063720703,
+ "step": 719
+ },
+ {
+ "epoch": 10.0,
+ "grad_norm": 0.020067067816853523,
+ "learning_rate": 0.0006,
+ "loss": 5.637441635131836,
+ "step": 720
+ },
+ {
+ "epoch": 10.0,
+ "eval_loss": 6.387807846069336,
+ "eval_runtime": 72.3372,
+ "eval_samples_per_second": 33.759,
+ "eval_steps_per_second": 1.064,
+ "step": 720
+ },
+ {
+ "epoch": 10.013973799126637,
+ "grad_norm": 0.02015046589076519,
+ "learning_rate": 0.0006,
+ "loss": 5.598219871520996,
+ "step": 721
+ },
+ {
+ "epoch": 10.027947598253276,
+ "grad_norm": 0.017398346215486526,
+ "learning_rate": 0.0006,
+ "loss": 5.66834831237793,
+ "step": 722
+ },
+ {
+ "epoch": 10.041921397379912,
+ "grad_norm": 0.016871167346835136,
+ "learning_rate": 0.0006,
+ "loss": 5.673916816711426,
+ "step": 723
+ },
+ {
+ "epoch": 10.055895196506551,
+ "grad_norm": 0.01695060543715954,
+ "learning_rate": 0.0006,
+ "loss": 5.646677017211914,
+ "step": 724
+ },
+ {
+ "epoch": 10.069868995633188,
+ "grad_norm": 0.019298404455184937,
+ "learning_rate": 0.0006,
+ "loss": 5.663064479827881,
+ "step": 725
+ },
+ {
+ "epoch": 10.083842794759825,
+ "grad_norm": 0.018699361011385918,
+ "learning_rate": 0.0006,
+ "loss": 5.615242958068848,
+ "step": 726
+ },
+ {
+ "epoch": 10.097816593886463,
+ "grad_norm": 0.019567854702472687,
+ "learning_rate": 0.0006,
+ "loss": 5.5521240234375,
+ "step": 727
+ },
+ {
+ "epoch": 10.1117903930131,
+ "grad_norm": 0.018347466364502907,
+ "learning_rate": 0.0006,
+ "loss": 5.5929107666015625,
+ "step": 728
+ },
+ {
+ "epoch": 10.125764192139737,
+ "grad_norm": 0.020350053906440735,
+ "learning_rate": 0.0006,
+ "loss": 5.6380109786987305,
+ "step": 729
+ },
+ {
+ "epoch": 10.139737991266376,
+ "grad_norm": 0.016790535300970078,
+ "learning_rate": 0.0006,
+ "loss": 5.62994384765625,
+ "step": 730
+ },
+ {
+ "epoch": 10.153711790393013,
+ "grad_norm": 0.01413649506866932,
+ "learning_rate": 0.0006,
+ "loss": 5.6301727294921875,
+ "step": 731
+ },
+ {
+ "epoch": 10.167685589519651,
+ "grad_norm": 0.013762339949607849,
+ "learning_rate": 0.0006,
+ "loss": 5.584138870239258,
+ "step": 732
+ },
+ {
+ "epoch": 10.181659388646288,
+ "grad_norm": 0.014790690504014492,
+ "learning_rate": 0.0006,
+ "loss": 5.666123390197754,
+ "step": 733
+ },
+ {
+ "epoch": 10.195633187772925,
+ "grad_norm": 0.015471262857317924,
+ "learning_rate": 0.0006,
+ "loss": 5.587042808532715,
+ "step": 734
+ },
+ {
+ "epoch": 10.209606986899564,
+ "grad_norm": 0.015804501250386238,
+ "learning_rate": 0.0006,
+ "loss": 5.627497673034668,
+ "step": 735
+ },
+ {
+ "epoch": 10.2235807860262,
+ "grad_norm": 0.017627960070967674,
+ "learning_rate": 0.0006,
+ "loss": 5.556890487670898,
+ "step": 736
+ },
+ {
+ "epoch": 10.237554585152838,
+ "grad_norm": 0.02074616774916649,
+ "learning_rate": 0.0006,
+ "loss": 5.586511135101318,
+ "step": 737
+ },
+ {
+ "epoch": 10.251528384279476,
+ "grad_norm": 0.026389935985207558,
+ "learning_rate": 0.0006,
+ "loss": 5.598520278930664,
+ "step": 738
+ },
+ {
+ "epoch": 10.265502183406113,
+ "grad_norm": 0.03736288473010063,
+ "learning_rate": 0.0006,
+ "loss": 5.575788497924805,
+ "step": 739
+ },
+ {
+ "epoch": 10.279475982532752,
+ "grad_norm": 0.03523678705096245,
+ "learning_rate": 0.0006,
+ "loss": 5.586930751800537,
+ "step": 740
+ },
+ {
+ "epoch": 10.293449781659389,
+ "grad_norm": 0.024266984313726425,
+ "learning_rate": 0.0006,
+ "loss": 5.596694469451904,
+ "step": 741
+ },
+ {
+ "epoch": 10.307423580786025,
+ "grad_norm": 0.033640433102846146,
+ "learning_rate": 0.0006,
+ "loss": 5.536456108093262,
+ "step": 742
+ },
+ {
+ "epoch": 10.321397379912664,
+ "grad_norm": 0.04689764603972435,
+ "learning_rate": 0.0006,
+ "loss": 5.61761999130249,
+ "step": 743
+ },
+ {
+ "epoch": 10.335371179039301,
+ "grad_norm": 0.04910685867071152,
+ "learning_rate": 0.0006,
+ "loss": 5.6298418045043945,
+ "step": 744
+ },
+ {
+ "epoch": 10.34934497816594,
+ "grad_norm": 0.03152550011873245,
+ "learning_rate": 0.0006,
+ "loss": 5.639697074890137,
+ "step": 745
+ },
+ {
+ "epoch": 10.363318777292577,
+ "grad_norm": 0.037753574550151825,
+ "learning_rate": 0.0006,
+ "loss": 5.697381019592285,
+ "step": 746
+ },
+ {
+ "epoch": 10.377292576419213,
+ "grad_norm": 0.04877670109272003,
+ "learning_rate": 0.0006,
+ "loss": 5.605184555053711,
+ "step": 747
+ },
+ {
+ "epoch": 10.391266375545852,
+ "grad_norm": 0.04228494316339493,
+ "learning_rate": 0.0006,
+ "loss": 5.624238967895508,
+ "step": 748
+ },
+ {
+ "epoch": 10.405240174672489,
+ "grad_norm": 0.03575790300965309,
+ "learning_rate": 0.0006,
+ "loss": 5.6037983894348145,
+ "step": 749
+ },
+ {
+ "epoch": 10.419213973799126,
+ "grad_norm": 0.036425549536943436,
+ "learning_rate": 0.0006,
+ "loss": 5.62905216217041,
+ "step": 750
+ },
+ {
+ "epoch": 10.433187772925764,
+ "grad_norm": 0.04001355543732643,
+ "learning_rate": 0.0006,
+ "loss": 5.581656455993652,
+ "step": 751
+ },
+ {
+ "epoch": 10.447161572052401,
+ "grad_norm": 0.05040939524769783,
+ "learning_rate": 0.0006,
+ "loss": 5.621358871459961,
+ "step": 752
+ },
+ {
+ "epoch": 10.46113537117904,
+ "grad_norm": 0.04140249267220497,
+ "learning_rate": 0.0006,
+ "loss": 5.668188095092773,
+ "step": 753
+ },
+ {
+ "epoch": 10.475109170305677,
+ "grad_norm": 0.033116403967142105,
+ "learning_rate": 0.0006,
+ "loss": 5.580477714538574,
+ "step": 754
+ },
+ {
+ "epoch": 10.489082969432314,
+ "grad_norm": 0.03864264488220215,
+ "learning_rate": 0.0006,
+ "loss": 5.539153099060059,
+ "step": 755
+ },
+ {
+ "epoch": 10.503056768558952,
+ "grad_norm": 0.03026588261127472,
+ "learning_rate": 0.0006,
+ "loss": 5.619814872741699,
+ "step": 756
+ },
+ {
+ "epoch": 10.51703056768559,
+ "grad_norm": 0.02381783165037632,
+ "learning_rate": 0.0006,
+ "loss": 5.64422607421875,
+ "step": 757
+ },
+ {
+ "epoch": 10.531004366812226,
+ "grad_norm": 0.02427830547094345,
+ "learning_rate": 0.0006,
+ "loss": 5.659873008728027,
+ "step": 758
+ },
+ {
+ "epoch": 10.544978165938865,
+ "grad_norm": 0.021709900349378586,
+ "learning_rate": 0.0006,
+ "loss": 5.590245723724365,
+ "step": 759
+ },
+ {
+ "epoch": 10.558951965065502,
+ "grad_norm": 0.022812914103269577,
+ "learning_rate": 0.0006,
+ "loss": 5.610434532165527,
+ "step": 760
+ },
+ {
+ "epoch": 10.57292576419214,
+ "grad_norm": 0.019250890240073204,
+ "learning_rate": 0.0006,
+ "loss": 5.588078498840332,
+ "step": 761
+ },
+ {
+ "epoch": 10.586899563318777,
+ "grad_norm": 0.021995501592755318,
+ "learning_rate": 0.0006,
+ "loss": 5.541754245758057,
+ "step": 762
+ },
+ {
+ "epoch": 10.600873362445414,
+ "grad_norm": 0.01477806642651558,
+ "learning_rate": 0.0006,
+ "loss": 5.549173355102539,
+ "step": 763
+ },
+ {
+ "epoch": 10.614847161572053,
+ "grad_norm": 0.016680821776390076,
+ "learning_rate": 0.0006,
+ "loss": 5.58221435546875,
+ "step": 764
+ },
+ {
+ "epoch": 10.62882096069869,
+ "grad_norm": 0.013848318718373775,
+ "learning_rate": 0.0006,
+ "loss": 5.5505876541137695,
+ "step": 765
+ },
+ {
+ "epoch": 10.642794759825328,
+ "grad_norm": 0.016387000679969788,
+ "learning_rate": 0.0006,
+ "loss": 5.570460319519043,
+ "step": 766
+ },
+ {
+ "epoch": 10.656768558951965,
+ "grad_norm": 0.016496233642101288,
+ "learning_rate": 0.0006,
+ "loss": 5.513058662414551,
+ "step": 767
+ },
+ {
+ "epoch": 10.670742358078602,
+ "grad_norm": 0.01675555482506752,
+ "learning_rate": 0.0006,
+ "loss": 5.559998989105225,
+ "step": 768
+ },
+ {
+ "epoch": 10.68471615720524,
+ "grad_norm": 0.016939565539360046,
+ "learning_rate": 0.0006,
+ "loss": 5.56727409362793,
+ "step": 769
+ },
+ {
+ "epoch": 10.698689956331878,
+ "grad_norm": 0.01679506152868271,
+ "learning_rate": 0.0006,
+ "loss": 5.603643417358398,
+ "step": 770
+ },
+ {
+ "epoch": 10.712663755458514,
+ "grad_norm": 0.01656349189579487,
+ "learning_rate": 0.0006,
+ "loss": 5.54292631149292,
+ "step": 771
+ },
+ {
+ "epoch": 10.726637554585153,
+ "grad_norm": 0.01975332945585251,
+ "learning_rate": 0.0006,
+ "loss": 5.5225677490234375,
+ "step": 772
+ },
+ {
+ "epoch": 10.74061135371179,
+ "grad_norm": 0.02329389750957489,
+ "learning_rate": 0.0006,
+ "loss": 5.537827491760254,
+ "step": 773
+ },
+ {
+ "epoch": 10.754585152838429,
+ "grad_norm": 0.026616891846060753,
+ "learning_rate": 0.0006,
+ "loss": 5.500011920928955,
+ "step": 774
+ },
+ {
+ "epoch": 10.768558951965066,
+ "grad_norm": 0.025511575862765312,
+ "learning_rate": 0.0006,
+ "loss": 5.5316619873046875,
+ "step": 775
+ },
+ {
+ "epoch": 10.782532751091702,
+ "grad_norm": 0.025009512901306152,
+ "learning_rate": 0.0006,
+ "loss": 5.468493938446045,
+ "step": 776
+ },
+ {
+ "epoch": 10.796506550218341,
+ "grad_norm": 0.02946239896118641,
+ "learning_rate": 0.0006,
+ "loss": 5.5116753578186035,
+ "step": 777
+ },
+ {
+ "epoch": 10.810480349344978,
+ "grad_norm": 0.036607250571250916,
+ "learning_rate": 0.0006,
+ "loss": 5.518636226654053,
+ "step": 778
+ },
+ {
+ "epoch": 10.824454148471617,
+ "grad_norm": 0.04347814992070198,
+ "learning_rate": 0.0006,
+ "loss": 5.572815895080566,
+ "step": 779
+ },
+ {
+ "epoch": 10.838427947598253,
+ "grad_norm": 0.034463658928871155,
+ "learning_rate": 0.0006,
+ "loss": 5.520634651184082,
+ "step": 780
+ },
+ {
+ "epoch": 10.85240174672489,
+ "grad_norm": 0.03486703336238861,
+ "learning_rate": 0.0006,
+ "loss": 5.545441150665283,
+ "step": 781
+ },
+ {
+ "epoch": 10.866375545851529,
+ "grad_norm": 0.04476739838719368,
+ "learning_rate": 0.0006,
+ "loss": 5.554450035095215,
+ "step": 782
+ },
+ {
+ "epoch": 10.880349344978166,
+ "grad_norm": 0.048745233565568924,
+ "learning_rate": 0.0006,
+ "loss": 5.554961204528809,
+ "step": 783
+ },
+ {
+ "epoch": 10.894323144104803,
+ "grad_norm": 0.04636877775192261,
+ "learning_rate": 0.0006,
+ "loss": 5.574655055999756,
+ "step": 784
+ },
+ {
+ "epoch": 10.908296943231441,
+ "grad_norm": 0.032686877995729446,
+ "learning_rate": 0.0006,
+ "loss": 5.5734639167785645,
+ "step": 785
+ },
+ {
+ "epoch": 10.922270742358078,
+ "grad_norm": 0.03075302764773369,
+ "learning_rate": 0.0006,
+ "loss": 5.487662315368652,
+ "step": 786
+ },
+ {
+ "epoch": 10.936244541484717,
+ "grad_norm": 0.028860175982117653,
+ "learning_rate": 0.0006,
+ "loss": 5.519798755645752,
+ "step": 787
+ },
+ {
+ "epoch": 10.950218340611354,
+ "grad_norm": 0.028599515557289124,
+ "learning_rate": 0.0006,
+ "loss": 5.558349609375,
+ "step": 788
+ },
+ {
+ "epoch": 10.96419213973799,
+ "grad_norm": 0.02384260669350624,
+ "learning_rate": 0.0006,
+ "loss": 5.520661354064941,
+ "step": 789
+ },
+ {
+ "epoch": 10.97816593886463,
+ "grad_norm": 0.023180771619081497,
+ "learning_rate": 0.0006,
+ "loss": 5.496896743774414,
+ "step": 790
+ },
+ {
+ "epoch": 10.992139737991266,
+ "grad_norm": 0.021167006343603134,
+ "learning_rate": 0.0006,
+ "loss": 5.517192363739014,
+ "step": 791
+ },
+ {
+ "epoch": 11.0,
+ "grad_norm": 0.02069013938307762,
+ "learning_rate": 0.0006,
+ "loss": 5.447656154632568,
+ "step": 792
+ },
+ {
+ "epoch": 11.0,
+ "eval_loss": 6.161121845245361,
+ "eval_runtime": 71.5972,
+ "eval_samples_per_second": 34.107,
+ "eval_steps_per_second": 1.075,
+ "step": 792
+ },
+ {
+ "epoch": 11.013973799126637,
+ "grad_norm": 0.022983292117714882,
+ "learning_rate": 0.0006,
+ "loss": 5.441346645355225,
+ "step": 793
+ },
+ {
+ "epoch": 11.027947598253276,
+ "grad_norm": 0.022833511233329773,
+ "learning_rate": 0.0006,
+ "loss": 5.45123815536499,
+ "step": 794
+ },
+ {
+ "epoch": 11.041921397379912,
+ "grad_norm": 0.020015383139252663,
+ "learning_rate": 0.0006,
+ "loss": 5.430389404296875,
+ "step": 795
+ },
+ {
+ "epoch": 11.055895196506551,
+ "grad_norm": 0.016957225278019905,
+ "learning_rate": 0.0006,
+ "loss": 5.473883628845215,
+ "step": 796
+ },
+ {
+ "epoch": 11.069868995633188,
+ "grad_norm": 0.02090892568230629,
+ "learning_rate": 0.0006,
+ "loss": 5.467441558837891,
+ "step": 797
+ },
+ {
+ "epoch": 11.083842794759825,
+ "grad_norm": 0.022806301712989807,
+ "learning_rate": 0.0006,
+ "loss": 5.5453081130981445,
+ "step": 798
+ },
+ {
+ "epoch": 11.097816593886463,
+ "grad_norm": 0.022816946730017662,
+ "learning_rate": 0.0006,
+ "loss": 5.426213264465332,
+ "step": 799
+ },
+ {
+ "epoch": 11.1117903930131,
+ "grad_norm": 0.020761270076036453,
+ "learning_rate": 0.0006,
+ "loss": 5.464138507843018,
+ "step": 800
+ },
+ {
+ "epoch": 11.125764192139737,
+ "grad_norm": 0.01982618309557438,
+ "learning_rate": 0.0006,
+ "loss": 5.428033828735352,
+ "step": 801
+ },
+ {
+ "epoch": 11.139737991266376,
+ "grad_norm": 0.021953724324703217,
+ "learning_rate": 0.0006,
+ "loss": 5.456646919250488,
+ "step": 802
+ },
+ {
+ "epoch": 11.153711790393013,
+ "grad_norm": 0.01871841959655285,
+ "learning_rate": 0.0006,
+ "loss": 5.508403778076172,
+ "step": 803
+ },
+ {
+ "epoch": 11.167685589519651,
+ "grad_norm": 0.01810353808104992,
+ "learning_rate": 0.0006,
+ "loss": 5.442957878112793,
+ "step": 804
+ },
+ {
+ "epoch": 11.181659388646288,
+ "grad_norm": 0.017609186470508575,
+ "learning_rate": 0.0006,
+ "loss": 5.4774699211120605,
+ "step": 805
+ },
+ {
+ "epoch": 11.195633187772925,
+ "grad_norm": 0.01931103691458702,
+ "learning_rate": 0.0006,
+ "loss": 5.429748058319092,
+ "step": 806
+ },
+ {
+ "epoch": 11.209606986899564,
+ "grad_norm": 0.019401391968131065,
+ "learning_rate": 0.0006,
+ "loss": 5.466834545135498,
+ "step": 807
+ },
+ {
+ "epoch": 11.2235807860262,
+ "grad_norm": 0.018862580880522728,
+ "learning_rate": 0.0006,
+ "loss": 5.431701183319092,
+ "step": 808
+ },
+ {
+ "epoch": 11.237554585152838,
+ "grad_norm": 0.01796964928507805,
+ "learning_rate": 0.0006,
+ "loss": 5.388613700866699,
+ "step": 809
+ },
+ {
+ "epoch": 11.251528384279476,
+ "grad_norm": 0.021166296675801277,
+ "learning_rate": 0.0006,
+ "loss": 5.454118251800537,
+ "step": 810
+ },
+ {
+ "epoch": 11.265502183406113,
+ "grad_norm": 0.028568729758262634,
+ "learning_rate": 0.0006,
+ "loss": 5.510260581970215,
+ "step": 811
+ },
+ {
+ "epoch": 11.279475982532752,
+ "grad_norm": 0.03902891278266907,
+ "learning_rate": 0.0006,
+ "loss": 5.386396408081055,
+ "step": 812
+ },
+ {
+ "epoch": 11.293449781659389,
+ "grad_norm": 0.039114370942115784,
+ "learning_rate": 0.0006,
+ "loss": 5.457526206970215,
+ "step": 813
+ },
+ {
+ "epoch": 11.307423580786025,
+ "grad_norm": 0.03560924902558327,
+ "learning_rate": 0.0006,
+ "loss": 5.4149980545043945,
+ "step": 814
+ },
+ {
+ "epoch": 11.321397379912664,
+ "grad_norm": 0.03745810315012932,
+ "learning_rate": 0.0006,
+ "loss": 5.464084625244141,
+ "step": 815
+ },
+ {
+ "epoch": 11.335371179039301,
+ "grad_norm": 0.03383171558380127,
+ "learning_rate": 0.0006,
+ "loss": 5.45137882232666,
+ "step": 816
+ },
+ {
+ "epoch": 11.34934497816594,
+ "grad_norm": 0.027873128652572632,
+ "learning_rate": 0.0006,
+ "loss": 5.485784530639648,
+ "step": 817
+ },
+ {
+ "epoch": 11.363318777292577,
+ "grad_norm": 0.02697703056037426,
+ "learning_rate": 0.0006,
+ "loss": 5.459152698516846,
+ "step": 818
+ },
+ {
+ "epoch": 11.377292576419213,
+ "grad_norm": 0.02646288089454174,
+ "learning_rate": 0.0006,
+ "loss": 5.433281421661377,
+ "step": 819
+ },
+ {
+ "epoch": 11.391266375545852,
+ "grad_norm": 0.02655179798603058,
+ "learning_rate": 0.0006,
+ "loss": 5.456627368927002,
+ "step": 820
+ },
+ {
+ "epoch": 11.405240174672489,
+ "grad_norm": 0.02633797563612461,
+ "learning_rate": 0.0006,
+ "loss": 5.484803676605225,
+ "step": 821
+ },
+ {
+ "epoch": 11.419213973799126,
+ "grad_norm": 0.031187085434794426,
+ "learning_rate": 0.0006,
+ "loss": 5.362293243408203,
+ "step": 822
+ },
+ {
+ "epoch": 11.433187772925764,
+ "grad_norm": 0.02971349097788334,
+ "learning_rate": 0.0006,
+ "loss": 5.4369611740112305,
+ "step": 823
+ },
+ {
+ "epoch": 11.447161572052401,
+ "grad_norm": 0.02996021695435047,
+ "learning_rate": 0.0006,
+ "loss": 5.419068813323975,
+ "step": 824
+ },
+ {
+ "epoch": 11.46113537117904,
+ "grad_norm": 0.032519739121198654,
+ "learning_rate": 0.0006,
+ "loss": 5.422189712524414,
+ "step": 825
+ },
+ {
+ "epoch": 11.475109170305677,
+ "grad_norm": 0.03403490409255028,
+ "learning_rate": 0.0006,
+ "loss": 5.470074653625488,
+ "step": 826
+ },
+ {
+ "epoch": 11.489082969432314,
+ "grad_norm": 0.036989886313676834,
+ "learning_rate": 0.0006,
+ "loss": 5.460174560546875,
+ "step": 827
+ },
+ {
+ "epoch": 11.503056768558952,
+ "grad_norm": 0.03427419438958168,
+ "learning_rate": 0.0006,
+ "loss": 5.495920658111572,
+ "step": 828
+ },
+ {
+ "epoch": 11.51703056768559,
+ "grad_norm": 0.02519620768725872,
+ "learning_rate": 0.0006,
+ "loss": 5.474429130554199,
+ "step": 829
+ },
+ {
+ "epoch": 11.531004366812226,
+ "grad_norm": 0.02473096363246441,
+ "learning_rate": 0.0006,
+ "loss": 5.395847320556641,
+ "step": 830
+ },
+ {
+ "epoch": 11.544978165938865,
+ "grad_norm": 0.02792714536190033,
+ "learning_rate": 0.0006,
+ "loss": 5.376262664794922,
+ "step": 831
+ },
+ {
+ "epoch": 11.558951965065502,
+ "grad_norm": 0.024172596633434296,
+ "learning_rate": 0.0006,
+ "loss": 5.459731101989746,
+ "step": 832
+ },
+ {
+ "epoch": 11.57292576419214,
+ "grad_norm": 0.02304821088910103,
+ "learning_rate": 0.0006,
+ "loss": 5.479475498199463,
+ "step": 833
+ },
+ {
+ "epoch": 11.586899563318777,
+ "grad_norm": 0.02342340536415577,
+ "learning_rate": 0.0006,
+ "loss": 5.387955665588379,
+ "step": 834
+ },
+ {
+ "epoch": 11.600873362445414,
+ "grad_norm": 0.021962853148579597,
+ "learning_rate": 0.0006,
+ "loss": 5.385697364807129,
+ "step": 835
+ },
+ {
+ "epoch": 11.614847161572053,
+ "grad_norm": 0.019193042069673538,
+ "learning_rate": 0.0006,
+ "loss": 5.431601047515869,
+ "step": 836
+ },
+ {
+ "epoch": 11.62882096069869,
+ "grad_norm": 0.021378040313720703,
+ "learning_rate": 0.0006,
+ "loss": 5.3070969581604,
+ "step": 837
+ },
+ {
+ "epoch": 11.642794759825328,
+ "grad_norm": 0.021641355007886887,
+ "learning_rate": 0.0006,
+ "loss": 5.360584735870361,
+ "step": 838
+ },
+ {
+ "epoch": 11.656768558951965,
+ "grad_norm": 0.019899170845746994,
+ "learning_rate": 0.0006,
+ "loss": 5.398155212402344,
+ "step": 839
+ },
+ {
+ "epoch": 11.670742358078602,
+ "grad_norm": 0.02298869751393795,
+ "learning_rate": 0.0006,
+ "loss": 5.410561561584473,
+ "step": 840
+ },
+ {
+ "epoch": 11.68471615720524,
+ "grad_norm": 0.02815551869571209,
+ "learning_rate": 0.0006,
+ "loss": 5.298604965209961,
+ "step": 841
+ },
+ {
+ "epoch": 11.698689956331878,
+ "grad_norm": 0.03041219152510166,
+ "learning_rate": 0.0006,
+ "loss": 5.425869941711426,
+ "step": 842
+ },
+ {
+ "epoch": 11.712663755458514,
+ "grad_norm": 0.025710400193929672,
+ "learning_rate": 0.0006,
+ "loss": 5.344316482543945,
+ "step": 843
+ },
+ {
+ "epoch": 11.726637554585153,
+ "grad_norm": 0.018438996747136116,
+ "learning_rate": 0.0006,
+ "loss": 5.379913330078125,
+ "step": 844
+ },
+ {
+ "epoch": 11.74061135371179,
+ "grad_norm": 0.016683703288435936,
+ "learning_rate": 0.0006,
+ "loss": 5.338386058807373,
+ "step": 845
+ },
+ {
+ "epoch": 11.754585152838429,
+ "grad_norm": 0.017252853140234947,
+ "learning_rate": 0.0006,
+ "loss": 5.360256195068359,
+ "step": 846
+ },
+ {
+ "epoch": 11.768558951965066,
+ "grad_norm": 0.019606001675128937,
+ "learning_rate": 0.0006,
+ "loss": 5.401415824890137,
+ "step": 847
+ },
+ {
+ "epoch": 11.782532751091702,
+ "grad_norm": 0.016299517825245857,
+ "learning_rate": 0.0006,
+ "loss": 5.39270544052124,
+ "step": 848
+ },
+ {
+ "epoch": 11.796506550218341,
+ "grad_norm": 0.016657982021570206,
+ "learning_rate": 0.0006,
+ "loss": 5.359323978424072,
+ "step": 849
+ },
+ {
+ "epoch": 11.810480349344978,
+ "grad_norm": 0.017462270334362984,
+ "learning_rate": 0.0006,
+ "loss": 5.453655242919922,
+ "step": 850
+ },
+ {
+ "epoch": 11.824454148471617,
+ "grad_norm": 0.01723778434097767,
+ "learning_rate": 0.0006,
+ "loss": 5.454940319061279,
+ "step": 851
+ },
+ {
+ "epoch": 11.838427947598253,
+ "grad_norm": 0.019071755930781364,
+ "learning_rate": 0.0006,
+ "loss": 5.355914115905762,
+ "step": 852
+ },
+ {
+ "epoch": 11.85240174672489,
+ "grad_norm": 0.02169782482087612,
+ "learning_rate": 0.0006,
+ "loss": 5.396806716918945,
+ "step": 853
+ },
+ {
+ "epoch": 11.866375545851529,
+ "grad_norm": 0.025049131363630295,
+ "learning_rate": 0.0006,
+ "loss": 5.347422122955322,
+ "step": 854
+ },
+ {
+ "epoch": 11.880349344978166,
+ "grad_norm": 0.024977540597319603,
+ "learning_rate": 0.0006,
+ "loss": 5.32526969909668,
+ "step": 855
+ },
+ {
+ "epoch": 11.894323144104803,
+ "grad_norm": 0.024223418906331062,
+ "learning_rate": 0.0006,
+ "loss": 5.372445106506348,
+ "step": 856
+ },
+ {
+ "epoch": 11.908296943231441,
+ "grad_norm": 0.02519913762807846,
+ "learning_rate": 0.0006,
+ "loss": 5.280451774597168,
+ "step": 857
+ },
+ {
+ "epoch": 11.922270742358078,
+ "grad_norm": 0.0255630724132061,
+ "learning_rate": 0.0006,
+ "loss": 5.462911605834961,
+ "step": 858
+ },
+ {
+ "epoch": 11.936244541484717,
+ "grad_norm": 0.022740041837096214,
+ "learning_rate": 0.0006,
+ "loss": 5.319089412689209,
+ "step": 859
+ },
+ {
+ "epoch": 11.950218340611354,
+ "grad_norm": 0.02038649097084999,
+ "learning_rate": 0.0006,
+ "loss": 5.37426233291626,
+ "step": 860
+ },
+ {
+ "epoch": 11.96419213973799,
+ "grad_norm": 0.020509786903858185,
+ "learning_rate": 0.0006,
+ "loss": 5.337924957275391,
+ "step": 861
+ },
+ {
+ "epoch": 11.97816593886463,
+ "grad_norm": 0.021707357838749886,
+ "learning_rate": 0.0006,
+ "loss": 5.374853134155273,
+ "step": 862
+ },
+ {
+ "epoch": 11.992139737991266,
+ "grad_norm": 0.02714136242866516,
+ "learning_rate": 0.0006,
+ "loss": 5.2982282638549805,
+ "step": 863
+ },
+ {
+ "epoch": 12.0,
+ "grad_norm": 0.027913225814700127,
+ "learning_rate": 0.0006,
+ "loss": 5.406044960021973,
+ "step": 864
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 28800,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 400,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": true,
+ "should_log": false,
+ "should_save": false,
+ "should_training_stop": false
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 3.683536033598669e+17,
+ "train_batch_size": 8,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-864/training_args.bin b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-864/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..c799bcf4051249aabf4c62fff9d38801df7bc2b5
--- /dev/null
+++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-864/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:40fb25263250c08c3479a6f0ac9fdf1f196e2a3b08a4eb10384786ff9b1903ef
+size 4792